mirror of
git://nv-tegra.nvidia.com/linux-nvgpu.git
synced 2025-12-22 09:12:24 +03:00
MIG mode does not support graphics, ELPG, and use cases like TPC floorsweeping. Skip all such initialization functions in common.gr unit if MIG mode is enabled. Set can_elpg to false if MIG mode is enabled. Jira NVGPU-5648 Change-Id: I03656dc6289e49a21ec7783430db9c8564c6bf1f Signed-off-by: Deepak Nibade <dnibade@nvidia.com> Reviewed-on: https://git-master.nvidia.com/r/c/linux-nvgpu/+/2411741 Reviewed-by: automaticguardword <automaticguardword@nvidia.com> Reviewed-by: svc-mobile-coverity <svc-mobile-coverity@nvidia.com> Reviewed-by: svc-mobile-cert <svc-mobile-cert@nvidia.com> Reviewed-by: Lakshmanan M <lm@nvidia.com> Reviewed-by: mobile promotions <svcmobile_promotions@nvidia.com> GVS: Gerrit_Virtual_Submit Tested-by: mobile promotions <svcmobile_promotions@nvidia.com>
336 lines
9.2 KiB
C
336 lines
9.2 KiB
C
/*
|
|
* Copyright (c) 2016-2020, NVIDIA CORPORATION. All rights reserved.
|
|
*
|
|
* This program is free software; you can redistribute it and/or modify it
|
|
* under the terms and conditions of the GNU General Public License,
|
|
* version 2, as published by the Free Software Foundation.
|
|
*
|
|
* This program is distributed in the hope it will be useful, but WITHOUT
|
|
* ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
|
|
* FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for
|
|
* more details.
|
|
*
|
|
* You should have received a copy of the GNU General Public License
|
|
* along with this program. If not, see <http://www.gnu.org/licenses/>.
|
|
*/
|
|
|
|
#include <linux/reboot.h>
|
|
#include <linux/dma-mapping.h>
|
|
#include <linux/mm.h>
|
|
#include <linux/slab.h>
|
|
#include <uapi/linux/nvgpu.h>
|
|
|
|
#include <nvgpu/defaults.h>
|
|
#include <nvgpu/kmem.h>
|
|
#include <nvgpu/nvgpu_common.h>
|
|
#include <nvgpu/soc.h>
|
|
#include <nvgpu/bug.h>
|
|
#include <nvgpu/enabled.h>
|
|
#include <nvgpu/debug.h>
|
|
#include <nvgpu/sizes.h>
|
|
#include <nvgpu/gk20a.h>
|
|
#include <nvgpu/regops.h>
|
|
#include <nvgpu/tsg.h>
|
|
#include <nvgpu/gr/gr.h>
|
|
|
|
#include "platform_gk20a.h"
|
|
#include "module.h"
|
|
#include "os_linux.h"
|
|
#include "sysfs.h"
|
|
#include "ioctl.h"
|
|
|
|
#define EMC3D_DEFAULT_RATIO 750
|
|
|
|
void nvgpu_kernel_restart(void *cmd)
|
|
{
|
|
kernel_restart(cmd);
|
|
}
|
|
|
|
static void nvgpu_init_vars(struct gk20a *g)
|
|
{
|
|
struct nvgpu_os_linux *l = nvgpu_os_linux_from_gk20a(g);
|
|
struct device *dev = dev_from_gk20a(g);
|
|
struct gk20a_platform *platform = dev_get_drvdata(dev);
|
|
|
|
nvgpu_cond_init(&g->mc.sw_irq_stall_last_handled_cond);
|
|
nvgpu_cond_init(&g->mc.sw_irq_nonstall_last_handled_cond);
|
|
|
|
init_rwsem(&l->busy_lock);
|
|
nvgpu_rwsem_init(&g->deterministic_busy);
|
|
|
|
nvgpu_spinlock_init(&g->mc.enable_lock);
|
|
|
|
nvgpu_spinlock_init(&g->power_spinlock);
|
|
|
|
nvgpu_spinlock_init(&g->mc.intr_lock);
|
|
|
|
nvgpu_mutex_init(&platform->railgate_lock);
|
|
nvgpu_mutex_init(&g->dbg_sessions_lock);
|
|
nvgpu_mutex_init(&g->client_lock);
|
|
nvgpu_mutex_init(&g->power_lock);
|
|
nvgpu_mutex_init(&g->tpc_pg_lock);
|
|
nvgpu_mutex_init(&g->clk_arb_enable_lock);
|
|
nvgpu_mutex_init(&g->cg_pg_lock);
|
|
#if defined(CONFIG_NVGPU_CYCLESTATS)
|
|
nvgpu_mutex_init(&g->cs_lock);
|
|
#endif
|
|
|
|
/* Init the clock req count to 0 */
|
|
nvgpu_atomic_set(&g->clk_arb_global_nr, 0);
|
|
|
|
nvgpu_mutex_init(&l->ctrl.privs_lock);
|
|
nvgpu_init_list_node(&l->ctrl.privs);
|
|
|
|
l->regs_saved = l->regs;
|
|
l->bar1_saved = l->bar1;
|
|
|
|
g->emc3d_ratio = EMC3D_DEFAULT_RATIO;
|
|
|
|
/* Set DMA parameters to allow larger sgt lists */
|
|
dev->dma_parms = &l->dma_parms;
|
|
dma_set_max_seg_size(dev, UINT_MAX);
|
|
|
|
/*
|
|
* A default of 16GB is the largest supported DMA size that is
|
|
* acceptable to all currently supported Tegra SoCs.
|
|
*/
|
|
if (!platform->dma_mask)
|
|
platform->dma_mask = DMA_BIT_MASK(34);
|
|
|
|
dma_set_mask(dev, platform->dma_mask);
|
|
dma_set_coherent_mask(dev, platform->dma_mask);
|
|
dma_set_seg_boundary(dev, platform->dma_mask);
|
|
|
|
nvgpu_init_list_node(&g->profiler_objects);
|
|
|
|
nvgpu_init_list_node(&g->boardobj_head);
|
|
nvgpu_init_list_node(&g->boardobjgrp_head);
|
|
|
|
nvgpu_set_enabled(g, NVGPU_HAS_SYNCPOINTS, platform->has_syncpoints);
|
|
}
|
|
|
|
static void nvgpu_init_max_comptag(struct gk20a *g)
|
|
{
|
|
#if LINUX_VERSION_CODE >= KERNEL_VERSION(5, 0, 0)
|
|
nvgpu_log_info(g, "total ram pages : %lu", totalram_pages());
|
|
#else
|
|
nvgpu_log_info(g, "total ram pages : %lu", totalram_pages);
|
|
#endif
|
|
g->max_comptag_mem = totalram_size_in_mb;
|
|
}
|
|
|
|
static void nvgpu_init_timeout(struct gk20a *g)
|
|
{
|
|
struct gk20a_platform *platform = dev_get_drvdata(dev_from_gk20a(g));
|
|
|
|
g->timeouts_disabled_by_user = false;
|
|
nvgpu_atomic_set(&g->timeouts_disabled_refcount, 0);
|
|
|
|
if (nvgpu_platform_is_silicon(g)) {
|
|
g->poll_timeout_default = NVGPU_DEFAULT_POLL_TIMEOUT_MS;
|
|
} else if (nvgpu_platform_is_fpga(g)) {
|
|
g->poll_timeout_default = NVGPU_DEFAULT_FPGA_TIMEOUT_MS;
|
|
} else {
|
|
g->poll_timeout_default = (u32)ULONG_MAX;
|
|
}
|
|
g->ch_wdt_init_limit_ms = platform->ch_wdt_init_limit_ms;
|
|
g->ctxsw_timeout_period_ms = CTXSW_TIMEOUT_PERIOD_MS;
|
|
}
|
|
|
|
static void nvgpu_init_timeslice(struct gk20a *g)
|
|
{
|
|
g->runlist_interleave = true;
|
|
|
|
g->tsg_timeslice_low_priority_us =
|
|
NVGPU_TSG_TIMESLICE_LOW_PRIORITY_US;
|
|
g->tsg_timeslice_medium_priority_us =
|
|
NVGPU_TSG_TIMESLICE_MEDIUM_PRIORITY_US;
|
|
g->tsg_timeslice_high_priority_us =
|
|
NVGPU_TSG_TIMESLICE_HIGH_PRIORITY_US;
|
|
|
|
g->tsg_timeslice_min_us = NVGPU_TSG_TIMESLICE_MIN_US;
|
|
g->tsg_timeslice_max_us = NVGPU_TSG_TIMESLICE_MAX_US;
|
|
}
|
|
|
|
static void nvgpu_init_pm_vars(struct gk20a *g)
|
|
{
|
|
struct gk20a_platform *platform = dev_get_drvdata(dev_from_gk20a(g));
|
|
u32 i = 0;
|
|
|
|
/*
|
|
* Set up initial power settings. For non-slicon platforms, disable
|
|
* power features and for silicon platforms, read from platform data
|
|
*/
|
|
g->slcg_enabled =
|
|
nvgpu_platform_is_silicon(g) ? platform->enable_slcg : false;
|
|
g->blcg_enabled =
|
|
nvgpu_platform_is_silicon(g) ? platform->enable_blcg : false;
|
|
g->elcg_enabled =
|
|
nvgpu_platform_is_silicon(g) ? platform->enable_elcg : false;
|
|
|
|
nvgpu_set_enabled(g, NVGPU_GPU_CAN_ELCG,
|
|
nvgpu_platform_is_silicon(g) ? platform->can_elcg : false);
|
|
nvgpu_set_enabled(g, NVGPU_GPU_CAN_SLCG,
|
|
nvgpu_platform_is_silicon(g) ? platform->can_slcg : false);
|
|
nvgpu_set_enabled(g, NVGPU_GPU_CAN_BLCG,
|
|
nvgpu_platform_is_silicon(g) ? platform->can_blcg : false);
|
|
|
|
g->aggressive_sync_destroy_thresh = platform->aggressive_sync_destroy_thresh;
|
|
#ifdef CONFIG_NVGPU_SUPPORT_CDE
|
|
g->has_cde = platform->has_cde;
|
|
#endif
|
|
g->ptimer_src_freq = platform->ptimer_src_freq;
|
|
|
|
nvgpu_set_enabled(g, NVGPU_CAN_RAILGATE, platform->can_railgate_init);
|
|
g->can_tpc_powergate = platform->can_tpc_powergate;
|
|
|
|
for (i = 0; i < MAX_TPC_PG_CONFIGS; i++)
|
|
g->valid_tpc_mask[i] = platform->valid_tpc_mask[i];
|
|
|
|
g->ldiv_slowdown_factor = platform->ldiv_slowdown_factor_init;
|
|
/* if default delay is not set, set default delay to 500msec */
|
|
if (platform->railgate_delay_init)
|
|
g->railgate_delay = platform->railgate_delay_init;
|
|
else
|
|
g->railgate_delay = NVGPU_DEFAULT_RAILGATE_IDLE_TIMEOUT;
|
|
|
|
g->support_ls_pmu = support_gk20a_pmu(dev_from_gk20a(g));
|
|
|
|
if (g->support_ls_pmu) {
|
|
g->elpg_enabled =
|
|
nvgpu_platform_is_silicon(g) ? platform->enable_elpg : false;
|
|
g->aelpg_enabled =
|
|
nvgpu_platform_is_silicon(g) ? platform->enable_aelpg : false;
|
|
g->mscg_enabled =
|
|
nvgpu_platform_is_silicon(g) ? platform->enable_mscg : false;
|
|
g->can_elpg =
|
|
nvgpu_platform_is_silicon(g) ? platform->can_elpg_init : false;
|
|
if (nvgpu_is_enabled(g, NVGPU_SUPPORT_MIG)) {
|
|
g->can_elpg = false;
|
|
}
|
|
|
|
nvgpu_set_enabled(g, NVGPU_PMU_PERFMON, platform->enable_perfmon);
|
|
}
|
|
|
|
nvgpu_set_enabled(g, NVGPU_SUPPORT_ASPM, !platform->disable_aspm);
|
|
#ifdef CONFIG_NVGPU_SIM
|
|
if (nvgpu_is_enabled(g, NVGPU_IS_FMODEL)) {
|
|
nvgpu_set_enabled(g, NVGPU_PMU_PSTATE, false);
|
|
} else
|
|
#endif
|
|
{
|
|
nvgpu_set_enabled(g, NVGPU_PMU_PSTATE, platform->pstate);
|
|
}
|
|
}
|
|
|
|
static void nvgpu_init_vbios_vars(struct gk20a *g)
|
|
{
|
|
struct gk20a_platform *platform = dev_get_drvdata(dev_from_gk20a(g));
|
|
|
|
nvgpu_set_enabled(g, NVGPU_PMU_RUN_PREOS, platform->run_preos);
|
|
}
|
|
|
|
static void nvgpu_init_ltc_vars(struct gk20a *g)
|
|
{
|
|
struct gk20a_platform *platform = dev_get_drvdata(dev_from_gk20a(g));
|
|
|
|
g->ltc_streamid = platform->ltc_streamid;
|
|
}
|
|
|
|
static void nvgpu_init_mm_vars(struct gk20a *g)
|
|
{
|
|
struct gk20a_platform *platform = dev_get_drvdata(dev_from_gk20a(g));
|
|
|
|
g->mm.disable_bigpage = platform->disable_bigpage;
|
|
nvgpu_set_enabled(g, NVGPU_MM_HONORS_APERTURE,
|
|
platform->honors_aperture);
|
|
nvgpu_set_enabled(g, NVGPU_MM_UNIFIED_MEMORY,
|
|
platform->unified_memory);
|
|
nvgpu_set_enabled(g, NVGPU_MM_UNIFY_ADDRESS_SPACES,
|
|
platform->unify_address_spaces);
|
|
nvgpu_set_enabled(g, NVGPU_MM_FORCE_128K_PMU_VM,
|
|
platform->force_128K_pmu_vm);
|
|
|
|
nvgpu_mutex_init(&g->mm.tlb_lock);
|
|
nvgpu_mutex_init(&g->mm.priv_lock);
|
|
}
|
|
|
|
int nvgpu_probe(struct gk20a *g,
|
|
const char *debugfs_symlink,
|
|
const char *interface_name,
|
|
struct class *device_class)
|
|
{
|
|
struct device *dev = dev_from_gk20a(g);
|
|
struct gk20a_platform *platform = dev_get_drvdata(dev);
|
|
int err = 0;
|
|
|
|
nvgpu_init_vars(g);
|
|
nvgpu_init_max_comptag(g);
|
|
nvgpu_init_timeout(g);
|
|
nvgpu_init_timeslice(g);
|
|
nvgpu_init_pm_vars(g);
|
|
nvgpu_init_vbios_vars(g);
|
|
nvgpu_init_ltc_vars(g);
|
|
err = nvgpu_init_soc_vars(g);
|
|
if (err) {
|
|
nvgpu_err(g, "init soc vars failed");
|
|
return err;
|
|
}
|
|
|
|
/* Initialize the platform interface. */
|
|
err = platform->probe(dev);
|
|
if (err) {
|
|
if (err == -EPROBE_DEFER)
|
|
nvgpu_info(g, "platform probe failed");
|
|
else
|
|
nvgpu_err(g, "platform probe failed");
|
|
return err;
|
|
}
|
|
|
|
nvgpu_init_mm_vars(g);
|
|
|
|
/* platform probe can defer do user init only if probe succeeds */
|
|
err = gk20a_user_init(dev, interface_name, device_class);
|
|
if (err)
|
|
return err;
|
|
|
|
if (platform->late_probe) {
|
|
err = platform->late_probe(dev);
|
|
if (err) {
|
|
nvgpu_err(g, "late probe failed");
|
|
return err;
|
|
}
|
|
}
|
|
|
|
nvgpu_create_sysfs(dev);
|
|
gk20a_debug_init(g, debugfs_symlink);
|
|
|
|
#ifdef CONFIG_NVGPU_DEBUGGER
|
|
g->dbg_regops_tmp_buf = nvgpu_kzalloc(g, SZ_4K);
|
|
if (!g->dbg_regops_tmp_buf) {
|
|
nvgpu_err(g, "couldn't allocate regops tmp buf");
|
|
return -ENOMEM;
|
|
}
|
|
g->dbg_regops_tmp_buf_ops =
|
|
SZ_4K / sizeof(g->dbg_regops_tmp_buf[0]);
|
|
#endif
|
|
|
|
g->remove_support = gk20a_remove_support;
|
|
|
|
nvgpu_ref_init(&g->refcount);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static void nvgpu_free_gk20a(struct gk20a *g)
|
|
{
|
|
struct nvgpu_os_linux *l = nvgpu_os_linux_from_gk20a(g);
|
|
|
|
kfree(l);
|
|
}
|
|
|
|
void nvgpu_init_gk20a(struct gk20a *g)
|
|
{
|
|
g->gfree = nvgpu_free_gk20a;
|
|
}
|