mirror of
git://nv-tegra.nvidia.com/linux-nvgpu.git
synced 2025-12-22 17:36:20 +03:00
Historically, nvgpu has supported a struct gk20a_dmabuf_priv and associated it with a dmabuf instance. This was aided by Nvmap's dma_buf_set_drv_data() and dma_buf_get_drvdata() APIs. gk20a_dmabuf_priv is used to store Comptag IDs i.e. (1 per 64 kb) as well as can store the dmabuf attachments to avoid multiple attach/detach calls. dma_buf_set_drv_data() allows Nvgpu to associate an instance of struct gk20a_dmabuf_priv with the instance of the dmabuf and also provide a release callback to delete the instance when the last reference to the dmabuf is put. Nvmap accomplishes this by modifying the struct dma_buf_ops definition to include the set_drv_data and get_drv_data callbacks in the kernel code. The above approach won't work for upstream Kstable and Nvmap plans to remove these APIs for upcoming newer downstream kernels as well. In order to implement the same functionality without depending on Nvmap, Nvgpu will implement a release chaining mechanism. Dmabuf's 'ops' pointer points to a constant struct and hence a whole copy of the ops is made followed by altering the new copy's release pointer. struct gk20a_dmabuf_priv stores the new copy and the dmabuf's 'ops' is changed to point to this. This allows Nvgpu to retrieve the corresponding gk20a_dmabuf_priv instance using container_of. Nvgpu's custom release callback will invoke the original release callback of the dmabuf's producer as a last step, thus completing the full circle. In case, the driver is removed, Nvgpu restores the dmabuf's 'ops' back to the original state. In order to accomplish this, every instance of a struct nvgpu_os_linux maintains a linkedlist of the gk20a_dma_buf instances. During the driver removal, this linkedlist is traversed and the corresponding dmabuf's 'ops' pointer is put back to its original state followed by freeing of this instance. Nvgpu is a producer of dmabuf's for vidmem and needs a way to check whether the given dmabuf belongs to itself. Its no longer reliable to depend on a comparision of the 'ops' pointer. Instead dmabuf_export_info() allows a name to be set by the exporter and this can be used to compare with a memory location that belongs to Nvgpu. Similarly for sysmem dmabufs, Nvmap makes a similar change in the way it identifies whether a dmabuf belongs to itself. Removed NVGPU_DMABUF_HAS_DRVDATA and moved to a unified mechanism for both downstream as well as upstream kernel. Some of the other changes in this file include the following. 1) Deletion of dmabuf.c and moving its contents over to dmabuf_priv.c 2) Replacing gk20a_mm_pin_has_drvdata with nvgpu_mm_pin_privdata and vice-versa for unpin. Bug 2878569 Change-Id: Icf8e79b05a25ad5a85f478c3ee0fc1eb7747e22d Signed-off-by: Debarshi Dutta <ddutta@nvidia.com> Reviewed-on: https://git-master.nvidia.com/r/c/linux-nvgpu/+/2341001 Reviewed-by: automaticguardword <automaticguardword@nvidia.com> Reviewed-by: Puneet Saxena <puneets@nvidia.com> Reviewed-by: Konsta Holtta <kholtta@nvidia.com> Reviewed-by: Vijayakumar Subbu <vsubbu@nvidia.com> Reviewed-by: mobile promotions <svcmobile_promotions@nvidia.com> GVS: Gerrit_Virtual_Submit Tested-by: mobile promotions <svcmobile_promotions@nvidia.com>
849 lines
19 KiB
C
849 lines
19 KiB
C
/*
|
|
* Copyright (c) 2016-2020, NVIDIA CORPORATION. All rights reserved.
|
|
*
|
|
* This program is free software; you can redistribute it and/or modify it
|
|
* under the terms and conditions of the GNU General Public License,
|
|
* version 2, as published by the Free Software Foundation.
|
|
*
|
|
* This program is distributed in the hope it will be useful, but WITHOUT
|
|
* ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
|
|
* FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for
|
|
* more details.
|
|
*
|
|
* You should have received a copy of the GNU General Public License
|
|
* along with this program. If not, see <http://www.gnu.org/licenses/>.
|
|
*/
|
|
|
|
#include <linux/pci.h>
|
|
#include <linux/interrupt.h>
|
|
#include <linux/pm_runtime.h>
|
|
#include <linux/of_platform.h>
|
|
#include <linux/of_address.h>
|
|
|
|
#include <nvgpu/nvhost.h>
|
|
#include <nvgpu/nvgpu_common.h>
|
|
#include <nvgpu/kmem.h>
|
|
#include <nvgpu/enabled.h>
|
|
#include <nvgpu/nvlink_probe.h>
|
|
#include <nvgpu/soc.h>
|
|
#include <nvgpu/sim.h>
|
|
#include <nvgpu/gk20a.h>
|
|
#include <nvgpu/string.h>
|
|
#include <nvgpu/gr/gr.h>
|
|
#include <nvgpu/nvgpu_init.h>
|
|
|
|
#include "nvlink.h"
|
|
#include "module.h"
|
|
#include "intr.h"
|
|
#include "sysfs.h"
|
|
#include "os_linux.h"
|
|
#include "platform_gk20a.h"
|
|
|
|
#include "pci.h"
|
|
#include "pci_power.h"
|
|
|
|
#include "driver_common.h"
|
|
#include "dmabuf_priv.h"
|
|
|
|
#define BOOT_GPC2CLK_MHZ 2581U
|
|
#define PCI_INTERFACE_NAME "card-%s%%s"
|
|
|
|
static int nvgpu_pci_tegra_probe(struct device *dev)
|
|
{
|
|
return 0;
|
|
}
|
|
|
|
static int nvgpu_pci_tegra_remove(struct device *dev)
|
|
{
|
|
return 0;
|
|
}
|
|
|
|
static bool nvgpu_pci_tegra_is_railgated(struct device *pdev)
|
|
{
|
|
return false;
|
|
}
|
|
|
|
static long nvgpu_pci_clk_round_rate(struct device *dev, unsigned long rate)
|
|
{
|
|
long ret = (long)rate;
|
|
|
|
if (rate == UINT_MAX)
|
|
ret = BOOT_GPC2CLK_MHZ * 1000000UL;
|
|
|
|
return ret;
|
|
}
|
|
|
|
static struct gk20a_platform nvgpu_pci_device[] = {
|
|
/* SKU 0x1ebf */
|
|
{
|
|
/* ptimer src frequency in hz */
|
|
.ptimer_src_freq = 31250000,
|
|
|
|
.probe = nvgpu_pci_tegra_probe,
|
|
.remove = nvgpu_pci_tegra_remove,
|
|
|
|
/* power management configuration */
|
|
.railgate_delay_init = 500,
|
|
.can_railgate_init = false,
|
|
.can_elpg_init = false,
|
|
.enable_elpg = false,
|
|
.enable_elcg = false,
|
|
.enable_slcg = false,
|
|
.enable_blcg = false,
|
|
.enable_mscg = false,
|
|
.can_slcg = false,
|
|
.can_blcg = false,
|
|
.can_elcg = false,
|
|
|
|
.disable_aspm = true,
|
|
.pstate = true,
|
|
|
|
/* power management callbacks */
|
|
.is_railgated = nvgpu_pci_tegra_is_railgated,
|
|
.clk_round_rate = nvgpu_pci_clk_round_rate,
|
|
|
|
/*
|
|
* WAR: PCIE X1 is very slow, set to very high value till nvlink is up
|
|
*/
|
|
.ch_wdt_init_limit_ms = 30000,
|
|
|
|
.unify_address_spaces = true,
|
|
.honors_aperture = true,
|
|
.dma_mask = DMA_BIT_MASK(40),
|
|
.hardcode_sw_threshold = false,
|
|
.unified_memory = false,
|
|
},
|
|
|
|
/* 0x1eba, 0x1efa, 0x1ebb, 0x1efb */
|
|
/* 0x1eae, 0x1eaf (internal chip SKUs) */
|
|
{
|
|
/* ptimer src frequency in hz */
|
|
.ptimer_src_freq = 31250000,
|
|
|
|
.probe = nvgpu_pci_tegra_probe,
|
|
.remove = nvgpu_pci_tegra_remove,
|
|
|
|
/* power management configuration */
|
|
.railgate_delay_init = 500,
|
|
.can_railgate_init = false,
|
|
.can_pci_gc_off = false,
|
|
.can_elpg_init = false,
|
|
.enable_elpg = false,
|
|
.enable_elcg = false,
|
|
.enable_slcg = true,
|
|
.enable_blcg = true,
|
|
.enable_mscg = false,
|
|
.can_slcg = true,
|
|
.can_blcg = true,
|
|
.can_elcg = false,
|
|
|
|
.disable_aspm = true,
|
|
.pstate = true,
|
|
|
|
/* power management callbacks */
|
|
.is_railgated = nvgpu_pci_tegra_is_railgated,
|
|
.clk_round_rate = nvgpu_pci_clk_round_rate,
|
|
|
|
.ch_wdt_init_limit_ms = 7000,
|
|
|
|
.unify_address_spaces = true,
|
|
.honors_aperture = true,
|
|
.dma_mask = DMA_BIT_MASK(40),
|
|
.hardcode_sw_threshold = false,
|
|
#ifdef CONFIG_TEGRA_GK20A_NVHOST
|
|
.has_syncpoints = true,
|
|
#endif
|
|
},
|
|
|
|
/* 0x1eb0 (RTX 5000 : TU104 based) */
|
|
{
|
|
/* ptimer src frequency in hz */
|
|
.ptimer_src_freq = 31250000,
|
|
|
|
.probe = nvgpu_pci_tegra_probe,
|
|
.remove = nvgpu_pci_tegra_remove,
|
|
|
|
/* power management configuration */
|
|
.railgate_delay_init = 500,
|
|
.can_railgate_init = false,
|
|
.can_pci_gc_off = false,
|
|
.can_elpg_init = false,
|
|
.enable_elpg = false,
|
|
.enable_elcg = false,
|
|
.enable_slcg = true,
|
|
.enable_blcg = true,
|
|
.enable_mscg = false,
|
|
.can_slcg = true,
|
|
.can_blcg = true,
|
|
.can_elcg = false,
|
|
|
|
.disable_aspm = true,
|
|
|
|
/* power management callbacks */
|
|
.is_railgated = nvgpu_pci_tegra_is_railgated,
|
|
.clk_round_rate = nvgpu_pci_clk_round_rate,
|
|
|
|
.ch_wdt_init_limit_ms = 7000,
|
|
|
|
.unify_address_spaces = true,
|
|
.honors_aperture = true,
|
|
.dma_mask = DMA_BIT_MASK(40),
|
|
.hardcode_sw_threshold = false,
|
|
#ifdef CONFIG_TEGRA_GK20A_NVHOST
|
|
.has_syncpoints = true,
|
|
#endif
|
|
},
|
|
/* PG209 */
|
|
{
|
|
/* ptimer src frequency in hz */
|
|
.ptimer_src_freq = 31250000,
|
|
|
|
.probe = nvgpu_pci_tegra_probe,
|
|
.remove = nvgpu_pci_tegra_remove,
|
|
|
|
/* power management configuration */
|
|
.railgate_delay_init = 500,
|
|
.can_railgate_init = false,
|
|
.can_pci_gc_off = false,
|
|
.can_elpg_init = false,
|
|
.enable_elpg = false,
|
|
.enable_elcg = false,
|
|
.enable_slcg = true,
|
|
.enable_blcg = true,
|
|
.enable_mscg = false,
|
|
.can_slcg = true,
|
|
.can_blcg = true,
|
|
.can_elcg = false,
|
|
|
|
.disable_aspm = true,
|
|
.pstate = true,
|
|
|
|
/* power management callbacks */
|
|
.is_railgated = nvgpu_pci_tegra_is_railgated,
|
|
.clk_round_rate = nvgpu_pci_clk_round_rate,
|
|
|
|
.ch_wdt_init_limit_ms = 7000,
|
|
|
|
.unify_address_spaces = true,
|
|
.honors_aperture = true,
|
|
.dma_mask = DMA_BIT_MASK(40),
|
|
.hardcode_sw_threshold = false,
|
|
#ifdef CONFIG_TEGRA_GK20A_NVHOST
|
|
.has_syncpoints = true,
|
|
#endif
|
|
},
|
|
|
|
};
|
|
|
|
#define PCI_DEVICE_INDEX(driver_data) ((driver_data) & 0x0000FFFFU)
|
|
#define PCI_DEVICE_FLAGS(driver_data) ((driver_data) & 0xFFFF0000U)
|
|
|
|
#define PCI_DEVICE_F_INTERNAL_CHIP_SKU BIT(31)
|
|
#define PCI_DEVICE_F_FUSA_CHIP_SKU BIT(30)
|
|
|
|
static struct pci_device_id nvgpu_pci_table[] = {
|
|
{
|
|
PCI_DEVICE(PCI_VENDOR_ID_NVIDIA, 0x1ebf),
|
|
.class = PCI_BASE_CLASS_DISPLAY << 16,
|
|
.class_mask = 0xff << 16,
|
|
.driver_data = 0,
|
|
},
|
|
{
|
|
PCI_DEVICE(PCI_VENDOR_ID_NVIDIA, 0x1eba),
|
|
.class = PCI_BASE_CLASS_DISPLAY << 16,
|
|
.class_mask = 0xff << 16,
|
|
.driver_data = 1,
|
|
},
|
|
{
|
|
PCI_DEVICE(PCI_VENDOR_ID_NVIDIA, 0x1eb0),
|
|
.class = PCI_BASE_CLASS_DISPLAY << 16,
|
|
.class_mask = 0xff << 16,
|
|
.driver_data = 2,
|
|
},
|
|
{
|
|
PCI_DEVICE(PCI_VENDOR_ID_NVIDIA, 0x1efa),
|
|
.class = PCI_BASE_CLASS_DISPLAY << 16,
|
|
.class_mask = 0xff << 16,
|
|
.driver_data = 1,
|
|
},
|
|
{
|
|
PCI_DEVICE(PCI_VENDOR_ID_NVIDIA, 0x1ebb),
|
|
.class = PCI_BASE_CLASS_DISPLAY << 16,
|
|
.class_mask = 0xff << 16,
|
|
.driver_data = 1,
|
|
},
|
|
{
|
|
PCI_DEVICE(PCI_VENDOR_ID_NVIDIA, 0x1efb),
|
|
.class = PCI_BASE_CLASS_DISPLAY << 16,
|
|
.class_mask = 0xff << 16,
|
|
.driver_data = 1,
|
|
},
|
|
{
|
|
PCI_DEVICE(PCI_VENDOR_ID_NVIDIA, 0x1eae),
|
|
.class = PCI_BASE_CLASS_DISPLAY << 16,
|
|
.class_mask = 0xff << 16,
|
|
.driver_data = 1 | PCI_DEVICE_F_INTERNAL_CHIP_SKU,
|
|
},
|
|
{
|
|
PCI_DEVICE(PCI_VENDOR_ID_NVIDIA, 0x1eaf),
|
|
.class = PCI_BASE_CLASS_DISPLAY << 16,
|
|
.class_mask = 0xff << 16,
|
|
.driver_data = 1 | PCI_DEVICE_F_INTERNAL_CHIP_SKU,
|
|
},
|
|
{
|
|
/* TU104-QS SKU*/
|
|
PCI_DEVICE(PCI_VENDOR_ID_NVIDIA, 0x1ebc),
|
|
.class = PCI_BASE_CLASS_DISPLAY << 16,
|
|
.class_mask = 0xff << 16,
|
|
.driver_data = 1 | PCI_DEVICE_F_FUSA_CHIP_SKU,
|
|
},
|
|
{
|
|
/* TU104-QS SKU*/
|
|
PCI_DEVICE(PCI_VENDOR_ID_NVIDIA, 0x1efc),
|
|
.class = PCI_BASE_CLASS_DISPLAY << 16,
|
|
.class_mask = 0xff << 16,
|
|
.driver_data = 1 | PCI_DEVICE_F_FUSA_CHIP_SKU,
|
|
},
|
|
{
|
|
/* PG209 SKU*/
|
|
PCI_DEVICE(PCI_VENDOR_ID_NVIDIA, 0x20b0),
|
|
.class = PCI_BASE_CLASS_DISPLAY << 16,
|
|
.class_mask = 0xff << 16,
|
|
.driver_data = 3,
|
|
},
|
|
|
|
{}
|
|
};
|
|
|
|
static irqreturn_t nvgpu_pci_isr(int irq, void *dev_id)
|
|
{
|
|
struct gk20a *g = dev_id;
|
|
irqreturn_t ret_stall;
|
|
irqreturn_t ret_nonstall;
|
|
|
|
ret_stall = nvgpu_intr_stall(g);
|
|
ret_nonstall = nvgpu_intr_nonstall(g);
|
|
|
|
#if defined(CONFIG_PCI_MSI)
|
|
/* Send MSI EOI */
|
|
if (g->ops.xve.rearm_msi && g->msi_enabled)
|
|
g->ops.xve.rearm_msi(g);
|
|
#endif
|
|
|
|
return (ret_stall == IRQ_NONE) ? ret_nonstall : IRQ_WAKE_THREAD;
|
|
}
|
|
|
|
static irqreturn_t nvgpu_pci_intr_thread(int irq, void *dev_id)
|
|
{
|
|
struct gk20a *g = dev_id;
|
|
|
|
return nvgpu_intr_thread_stall(g);
|
|
}
|
|
|
|
static int nvgpu_pci_init_support(struct pci_dev *pdev)
|
|
{
|
|
int err = 0;
|
|
struct gk20a *g = get_gk20a(&pdev->dev);
|
|
struct nvgpu_os_linux *l = nvgpu_os_linux_from_gk20a(g);
|
|
struct device *dev = &pdev->dev;
|
|
|
|
l->regs = nvgpu_devm_ioremap(dev, pci_resource_start(pdev, 0),
|
|
pci_resource_len(pdev, 0));
|
|
if (IS_ERR(l->regs)) {
|
|
nvgpu_err(g, "failed to remap gk20a registers");
|
|
err = PTR_ERR(l->regs);
|
|
goto fail;
|
|
}
|
|
|
|
l->regs_bus_addr = pci_resource_start(pdev, 0);
|
|
if (!l->regs_bus_addr) {
|
|
nvgpu_err(g, "failed to read register bus offset");
|
|
err = -ENODEV;
|
|
goto fail;
|
|
}
|
|
|
|
l->bar1 = nvgpu_devm_ioremap(dev, pci_resource_start(pdev, 1),
|
|
pci_resource_len(pdev, 1));
|
|
if (IS_ERR(l->bar1)) {
|
|
nvgpu_err(g, "failed to remap gk20a bar1");
|
|
err = PTR_ERR(l->bar1);
|
|
goto fail;
|
|
}
|
|
|
|
err = nvgpu_init_sim_support_linux_pci(g);
|
|
if (err)
|
|
goto fail;
|
|
err = nvgpu_init_sim_support_pci(g);
|
|
if (err)
|
|
goto fail_sim;
|
|
|
|
err = nvgpu_gr_alloc(g);
|
|
if (err != 0) {
|
|
nvgpu_err(g, "couldn't allocate gr memory");
|
|
goto fail_sim;
|
|
}
|
|
|
|
return 0;
|
|
|
|
fail_sim:
|
|
nvgpu_remove_sim_support_linux_pci(g);
|
|
fail:
|
|
if (l->regs)
|
|
l->regs = NULL;
|
|
|
|
if (l->bar1)
|
|
l->bar1 = NULL;
|
|
|
|
return err;
|
|
}
|
|
|
|
static char *nvgpu_pci_devnode(struct device *dev, umode_t *mode)
|
|
{
|
|
if (mode)
|
|
*mode = S_IRUGO | S_IWUGO;
|
|
return kasprintf(GFP_KERNEL, "nvgpu-pci/%s", dev_name(dev));
|
|
}
|
|
|
|
static struct class nvgpu_pci_class = {
|
|
.owner = THIS_MODULE,
|
|
.name = "nvidia-pci-gpu",
|
|
.devnode = nvgpu_pci_devnode,
|
|
};
|
|
|
|
#ifdef CONFIG_PM
|
|
static int nvgpu_pci_pm_runtime_resume(struct device *dev)
|
|
{
|
|
return gk20a_pm_finalize_poweron(dev);
|
|
}
|
|
|
|
static int nvgpu_pci_pm_runtime_suspend(struct device *dev)
|
|
{
|
|
return 0;
|
|
}
|
|
|
|
static int nvgpu_pci_pm_resume(struct device *dev)
|
|
{
|
|
return gk20a_pm_finalize_poweron(dev);
|
|
}
|
|
|
|
static int nvgpu_pci_pm_suspend(struct device *dev)
|
|
{
|
|
return 0;
|
|
}
|
|
|
|
static const struct dev_pm_ops nvgpu_pci_pm_ops = {
|
|
.runtime_resume = nvgpu_pci_pm_runtime_resume,
|
|
.runtime_suspend = nvgpu_pci_pm_runtime_suspend,
|
|
.resume = nvgpu_pci_pm_resume,
|
|
.suspend = nvgpu_pci_pm_suspend,
|
|
};
|
|
#endif
|
|
|
|
static int nvgpu_pci_pm_init(struct device *dev)
|
|
{
|
|
#ifdef CONFIG_PM
|
|
struct gk20a *g = get_gk20a(dev);
|
|
|
|
if (!nvgpu_is_enabled(g, NVGPU_CAN_RAILGATE)) {
|
|
pm_runtime_disable(dev);
|
|
} else {
|
|
if (g->railgate_delay)
|
|
pm_runtime_set_autosuspend_delay(dev,
|
|
g->railgate_delay);
|
|
|
|
/*
|
|
* set gpu dev's use_autosuspend flag to allow
|
|
* runtime power management of GPU
|
|
*/
|
|
pm_runtime_use_autosuspend(dev);
|
|
|
|
/*
|
|
* runtime PM for PCI devices is forbidden
|
|
* by default, so unblock RTPM of GPU
|
|
*/
|
|
pm_runtime_put_noidle(dev);
|
|
pm_runtime_allow(dev);
|
|
}
|
|
#endif
|
|
return 0;
|
|
}
|
|
|
|
static int nvgpu_pci_pm_deinit(struct device *dev)
|
|
{
|
|
#ifdef CONFIG_PM
|
|
struct gk20a *g = get_gk20a(dev);
|
|
|
|
if (!nvgpu_is_enabled(g, NVGPU_CAN_RAILGATE))
|
|
pm_runtime_enable(dev);
|
|
else
|
|
pm_runtime_forbid(dev);
|
|
#endif
|
|
return 0;
|
|
}
|
|
|
|
static int nvgpu_get_dt_clock_limit(struct gk20a *g, u16 *gpuclk_clkmhz)
|
|
{
|
|
struct device_node *np;
|
|
u32 gpuclk_dt_cap = 0U;
|
|
|
|
np = of_find_node_by_name(NULL, "nvgpu");
|
|
if (!np) {
|
|
return -ENOENT;
|
|
}
|
|
|
|
if (of_property_read_u32(np, "dgpuclk-max-mhz", &gpuclk_dt_cap)) {
|
|
nvgpu_info(g, "dgpuclk-max-mhz not defined,"
|
|
"P-state will be used");
|
|
}
|
|
|
|
*gpuclk_clkmhz = (u16)gpuclk_dt_cap;
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int nvgpu_pci_probe(struct pci_dev *pdev,
|
|
const struct pci_device_id *pent)
|
|
{
|
|
struct gk20a_platform *platform = NULL;
|
|
struct nvgpu_os_linux *l;
|
|
struct gk20a *g;
|
|
int err;
|
|
char nodefmt[64];
|
|
struct device_node *np;
|
|
u32 device_index = PCI_DEVICE_INDEX(pent->driver_data);
|
|
u32 device_flags = PCI_DEVICE_FLAGS(pent->driver_data);
|
|
|
|
/* make sure driver_data is a sane index */
|
|
if (device_index >= sizeof(nvgpu_pci_device) /
|
|
sizeof(nvgpu_pci_device[0])) {
|
|
return -EINVAL;
|
|
}
|
|
|
|
l = kzalloc(sizeof(*l), GFP_KERNEL);
|
|
if (!l) {
|
|
dev_err(&pdev->dev, "couldn't allocate gk20a support");
|
|
return -ENOMEM;
|
|
}
|
|
|
|
hash_init(l->ecc_sysfs_stats_htable);
|
|
|
|
g = &l->g;
|
|
|
|
g->log_mask = NVGPU_DEFAULT_DBG_MASK;
|
|
|
|
nvgpu_init_gk20a(g);
|
|
|
|
nvgpu_kmem_init(g);
|
|
|
|
/* Allocate memory to hold platform data*/
|
|
platform = (struct gk20a_platform *)nvgpu_kzalloc( g,
|
|
sizeof(struct gk20a_platform));
|
|
if (!platform) {
|
|
dev_err(&pdev->dev, "couldn't allocate platform data");
|
|
err = -ENOMEM;
|
|
goto err_free_l;
|
|
}
|
|
|
|
/* copy detected device data to allocated platform space*/
|
|
nvgpu_memcpy((u8 *)platform,
|
|
(u8 *)&nvgpu_pci_device[device_index],
|
|
sizeof(struct gk20a_platform));
|
|
|
|
pci_set_drvdata(pdev, platform);
|
|
|
|
err = nvgpu_init_enabled_flags(g);
|
|
if (err)
|
|
goto err_free_platform;
|
|
|
|
platform->g = g;
|
|
l->dev = &pdev->dev;
|
|
|
|
np = nvgpu_get_node(g);
|
|
if (of_dma_is_coherent(np)) {
|
|
nvgpu_set_enabled(g, NVGPU_USE_COHERENT_SYSMEM, true);
|
|
nvgpu_set_enabled(g, NVGPU_SUPPORT_IO_COHERENCE, true);
|
|
}
|
|
|
|
err = pci_enable_device(pdev);
|
|
if (err)
|
|
goto err_free_platform;
|
|
pci_set_master(pdev);
|
|
|
|
g->pci_vendor_id = pdev->vendor;
|
|
g->pci_device_id = pdev->device;
|
|
g->pci_subsystem_vendor_id = pdev->subsystem_vendor;
|
|
g->pci_subsystem_device_id = pdev->subsystem_device;
|
|
g->pci_class = (pdev->class >> 8) & 0xFFFFU; // we only want base/sub
|
|
g->pci_revision = pdev->revision;
|
|
|
|
if ((device_flags & PCI_DEVICE_F_INTERNAL_CHIP_SKU) != 0U) {
|
|
nvgpu_err(g, "internal chip SKU %08x detected",
|
|
g->pci_device_id);
|
|
nvgpu_err(g, "replace board, or use at your own risks");
|
|
}
|
|
|
|
if ((device_flags & PCI_DEVICE_F_FUSA_CHIP_SKU) != 0U) {
|
|
g->is_fusa_sku = true;
|
|
}
|
|
|
|
g->ina3221_dcb_index = platform->ina3221_dcb_index;
|
|
g->ina3221_i2c_address = platform->ina3221_i2c_address;
|
|
g->ina3221_i2c_port = platform->ina3221_i2c_port;
|
|
g->hardcode_sw_threshold = platform->hardcode_sw_threshold;
|
|
|
|
#if defined(CONFIG_PCI_MSI)
|
|
err = pci_enable_msi(pdev);
|
|
if (err) {
|
|
nvgpu_err(g,
|
|
"MSI could not be enabled, falling back to legacy");
|
|
g->msi_enabled = false;
|
|
} else
|
|
g->msi_enabled = true;
|
|
#endif
|
|
|
|
g->mc.irq_stall = pdev->irq;
|
|
g->mc.irq_nonstall = pdev->irq;
|
|
if ((int)g->mc.irq_stall < 0) {
|
|
err = -ENXIO;
|
|
goto err_disable_msi;
|
|
}
|
|
|
|
/* Number of stall interrupt line = 1 (for dgpu <= tu10x) */
|
|
g->mc.irq_stall_count = 1;
|
|
|
|
g->mc.irq_stall_lines[0] = g->mc.irq_stall;
|
|
|
|
err = devm_request_threaded_irq(&pdev->dev,
|
|
g->mc.irq_stall,
|
|
nvgpu_pci_isr,
|
|
nvgpu_pci_intr_thread,
|
|
#if defined(CONFIG_PCI_MSI)
|
|
g->msi_enabled ? 0 :
|
|
#endif
|
|
IRQF_SHARED, "nvgpu", g);
|
|
if (err) {
|
|
nvgpu_err(g,
|
|
"failed to request irq @ %d", g->mc.irq_stall);
|
|
goto err_disable_msi;
|
|
}
|
|
disable_irq(g->mc.irq_stall);
|
|
|
|
err = nvgpu_pci_init_support(pdev);
|
|
if (err)
|
|
goto err_free_irq;
|
|
|
|
if (strchr(dev_name(&pdev->dev), '%')) {
|
|
nvgpu_err(g, "illegal character in device name");
|
|
err = -EINVAL;
|
|
goto err_free_irq;
|
|
}
|
|
|
|
(void) snprintf(nodefmt, sizeof(nodefmt),
|
|
PCI_INTERFACE_NAME, dev_name(&pdev->dev));
|
|
|
|
err = nvgpu_probe(g, "gpu_pci", nodefmt, &nvgpu_pci_class);
|
|
if (err)
|
|
goto err_free_irq;
|
|
|
|
err = nvgpu_pci_pm_init(&pdev->dev);
|
|
if (err) {
|
|
nvgpu_err(g, "pm init failed");
|
|
goto err_free_irq;
|
|
}
|
|
|
|
err = nvgpu_nvlink_probe(g);
|
|
/*
|
|
* ENODEV is a legal error which means there is no NVLINK
|
|
* any other error is fatal
|
|
*/
|
|
if (err) {
|
|
if (err != -ENODEV) {
|
|
nvgpu_err(g, "fatal error probing nvlink, bailing out");
|
|
goto err_free_irq;
|
|
}
|
|
nvgpu_set_enabled(g, NVGPU_SUPPORT_NVLINK, false);
|
|
}
|
|
|
|
#ifdef CONFIG_TEGRA_GK20A_NVHOST
|
|
err = nvgpu_nvhost_syncpt_init(g);
|
|
if (err) {
|
|
if (err != -ENOSYS) {
|
|
nvgpu_err(g, "syncpt init failed");
|
|
goto err_free_irq;
|
|
}
|
|
}
|
|
#endif
|
|
|
|
err = nvgpu_get_dt_clock_limit(g, &g->dgpu_max_clk);
|
|
if (err != 0) {
|
|
nvgpu_info(g, "Missing nvgpu node");
|
|
}
|
|
|
|
err = nvgpu_pci_add_pci_power(pdev);
|
|
if (err) {
|
|
nvgpu_err(g, "add pci power failed (%d).", err);
|
|
goto err_free_irq;
|
|
}
|
|
|
|
nvgpu_mutex_init(&l->dmabuf_priv_list_lock);
|
|
nvgpu_init_list_node(&l->dmabuf_priv_list);
|
|
|
|
return 0;
|
|
|
|
err_free_irq:
|
|
nvgpu_free_irq(g);
|
|
nvgpu_gr_free(g);
|
|
err_disable_msi:
|
|
#if defined(CONFIG_PCI_MSI)
|
|
if (g->msi_enabled)
|
|
pci_disable_msi(pdev);
|
|
#endif
|
|
err_free_platform:
|
|
nvgpu_kfree(g, platform);
|
|
err_free_l:
|
|
kfree(l);
|
|
return err;
|
|
}
|
|
|
|
static void nvgpu_thermal_deinit(struct gk20a *g)
|
|
{
|
|
struct nvgpu_os_linux *l = nvgpu_os_linux_from_gk20a(g);
|
|
struct device *dev = dev_from_gk20a(g);
|
|
|
|
devm_free_irq(dev, l->thermal_alert.therm_alert_irq, g);
|
|
if (l->thermal_alert.workqueue != NULL) {
|
|
cancel_work_sync(&l->thermal_alert.work);
|
|
destroy_workqueue(l->thermal_alert.workqueue);
|
|
l->thermal_alert.workqueue = NULL;
|
|
}
|
|
}
|
|
|
|
static void nvgpu_pci_remove(struct pci_dev *pdev)
|
|
{
|
|
struct gk20a *g = get_gk20a(&pdev->dev);
|
|
struct device *dev = dev_from_gk20a(g);
|
|
int err;
|
|
struct nvgpu_os_linux *l = nvgpu_os_linux_from_gk20a(g);
|
|
|
|
/* no support yet for unbind if DGPU is in VGPU mode */
|
|
if (gk20a_gpu_is_virtual(dev))
|
|
return;
|
|
|
|
gk20a_dma_buf_priv_list_clear(l);
|
|
nvgpu_mutex_destroy(&l->dmabuf_priv_list_lock);
|
|
|
|
err = nvgpu_pci_clear_pci_power(dev_name(dev));
|
|
WARN(err, "gpu failed to clear pci power");
|
|
|
|
err = nvgpu_nvlink_deinit(g);
|
|
/* ENODEV is a legal error if there is no NVLINK */
|
|
if (err != -ENODEV) {
|
|
WARN(err, "gpu failed to remove nvlink");
|
|
}
|
|
|
|
gk20a_driver_start_unload(g);
|
|
|
|
if (nvgpu_is_enabled(g, NVGPU_SUPPORT_DGPU_THERMAL_ALERT) &&
|
|
nvgpu_platform_is_silicon(g)) {
|
|
nvgpu_thermal_deinit(g);
|
|
}
|
|
|
|
err = nvgpu_quiesce(g);
|
|
/* TODO: handle failure to idle */
|
|
WARN(err, "gpu failed to idle during driver removal");
|
|
|
|
nvgpu_free_irq(g);
|
|
|
|
nvgpu_remove(dev, &nvgpu_pci_class);
|
|
|
|
#if defined(CONFIG_PCI_MSI)
|
|
if (g->msi_enabled)
|
|
pci_disable_msi(pdev);
|
|
else {
|
|
/* IRQ does not need to be enabled in MSI as the line is not
|
|
* shared
|
|
*/
|
|
enable_irq(g->mc.irq_stall);
|
|
}
|
|
#endif
|
|
nvgpu_pci_pm_deinit(&pdev->dev);
|
|
|
|
/* free allocated platform data space */
|
|
gk20a_get_platform(&pdev->dev)->g = NULL;
|
|
nvgpu_kfree(g, gk20a_get_platform(&pdev->dev));
|
|
|
|
nvgpu_put(g);
|
|
}
|
|
|
|
void nvgpu_pci_shutdown(struct pci_dev *pdev)
|
|
{
|
|
struct gk20a *g = get_gk20a(&pdev->dev);
|
|
struct device *dev = dev_from_gk20a(g);
|
|
int err;
|
|
|
|
nvgpu_info(g, "shutting down");
|
|
|
|
/* no support yet if DGPU is in VGPU mode */
|
|
if (gk20a_gpu_is_virtual(dev))
|
|
return;
|
|
|
|
if (!nvgpu_is_powered_on(g)) {
|
|
return;
|
|
}
|
|
|
|
if (is_nvgpu_gpu_state_valid(g)) {
|
|
err = nvgpu_nvlink_deinit(g);
|
|
/* ENODEV is a legal error if there is no NVLINK */
|
|
if (err != -ENODEV) {
|
|
WARN(err, "gpu failed to remove nvlink");
|
|
}
|
|
} else
|
|
nvgpu_err(g, "skipped nvlink deinit");
|
|
|
|
nvgpu_info(g, "shut down complete");
|
|
}
|
|
|
|
static struct pci_driver nvgpu_pci_driver = {
|
|
.name = "nvgpu",
|
|
.id_table = nvgpu_pci_table,
|
|
.probe = nvgpu_pci_probe,
|
|
.remove = nvgpu_pci_remove,
|
|
.shutdown = nvgpu_pci_shutdown,
|
|
#ifdef CONFIG_PM
|
|
.driver.pm = &nvgpu_pci_pm_ops,
|
|
#endif
|
|
};
|
|
|
|
int __init nvgpu_pci_init(void)
|
|
{
|
|
int ret;
|
|
|
|
ret = class_register(&nvgpu_pci_class);
|
|
if (ret)
|
|
return ret;
|
|
|
|
ret = pci_register_driver(&nvgpu_pci_driver);
|
|
if (ret)
|
|
goto driver_fail;
|
|
|
|
ret = nvgpu_pci_power_init(&nvgpu_pci_driver);
|
|
if (ret)
|
|
goto power_init_fail;
|
|
|
|
return 0;
|
|
|
|
power_init_fail:
|
|
pci_unregister_driver(&nvgpu_pci_driver);
|
|
driver_fail:
|
|
class_unregister(&nvgpu_pci_class);
|
|
return ret;
|
|
}
|
|
|
|
void __exit nvgpu_pci_exit(void)
|
|
{
|
|
nvgpu_pci_power_exit(&nvgpu_pci_driver);
|
|
pci_unregister_driver(&nvgpu_pci_driver);
|
|
class_unregister(&nvgpu_pci_class);
|
|
nvgpu_pci_power_cleanup();
|
|
}
|