mirror of
git://nv-tegra.nvidia.com/linux-nvgpu.git
synced 2025-12-22 17:36:20 +03:00
The CONFIG_NVGPU_NEXT config is no longer required now that ga10b and ga100 sources have been collapsed. However, the ga100, ga10b sources are not safety certified, so mark them as NON_FUSA by replacing CONFIG_NVGPU_NEXT with CONFIG_NVGPU_NON_FUSA. Move CONFIG_NVGPU_MIG to Makefile.linux.config and enable MIG support by default on standard build. Jira NVGPU-4771 Change-Id: Idc5861fe71d9d510766cf242c6858e2faf97d7d0 Signed-off-by: Antony Clince Alex <aalex@nvidia.com> Reviewed-on: https://git-master.nvidia.com/r/c/linux-nvgpu/+/2547092 Tested-by: mobile promotions <svcmobile_promotions@nvidia.com> Reviewed-by: mobile promotions <svcmobile_promotions@nvidia.com>
711 lines
15 KiB
C
711 lines
15 KiB
C
/*
|
|
* Copyright (c) 2017-2021, NVIDIA CORPORATION. All rights reserved.
|
|
*
|
|
* Permission is hereby granted, free of charge, to any person obtaining a
|
|
* copy of this software and associated documentation files (the "Software"),
|
|
* to deal in the Software without restriction, including without limitation
|
|
* the rights to use, copy, modify, merge, publish, distribute, sublicense,
|
|
* and/or sell copies of the Software, and to permit persons to whom the
|
|
* Software is furnished to do so, subject to the following conditions:
|
|
*
|
|
* The above copyright notice and this permission notice shall be included in
|
|
* all copies or substantial portions of the Software.
|
|
*
|
|
* THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
|
|
* IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
|
|
* FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
|
|
* THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
|
|
* LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING
|
|
* FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER
|
|
* DEALINGS IN THE SOFTWARE.
|
|
*/
|
|
|
|
#include <nvgpu/mm.h>
|
|
#include <nvgpu/vm.h>
|
|
#include <nvgpu/dma.h>
|
|
#include <nvgpu/vm_area.h>
|
|
#include <nvgpu/acr.h>
|
|
#include <nvgpu/gmmu.h>
|
|
#include <nvgpu/vidmem.h>
|
|
#include <nvgpu/semaphore.h>
|
|
#include <nvgpu/pramin.h>
|
|
#include <nvgpu/enabled.h>
|
|
#include <nvgpu/errata.h>
|
|
#include <nvgpu/ce_app.h>
|
|
#include <nvgpu/gk20a.h>
|
|
#include <nvgpu/engines.h>
|
|
#include <nvgpu/static_analysis.h>
|
|
#include <nvgpu/power_features/cg.h>
|
|
|
|
int nvgpu_mm_suspend(struct gk20a *g)
|
|
{
|
|
int err;
|
|
|
|
nvgpu_log_info(g, "MM suspend running...");
|
|
|
|
#ifdef CONFIG_NVGPU_DGPU
|
|
nvgpu_vidmem_thread_pause_sync(&g->mm);
|
|
#endif
|
|
|
|
#ifdef CONFIG_NVGPU_COMPRESSION
|
|
g->ops.mm.cache.cbc_clean(g);
|
|
#endif
|
|
err = g->ops.mm.cache.l2_flush(g, false);
|
|
if (err != 0) {
|
|
nvgpu_err(g, "l2_flush failed");
|
|
return err;
|
|
}
|
|
|
|
if (g->ops.fb.intr.disable != NULL) {
|
|
g->ops.fb.intr.disable(g);
|
|
}
|
|
|
|
if (g->ops.mm.mmu_fault.disable_hw != NULL) {
|
|
g->ops.mm.mmu_fault.disable_hw(g);
|
|
}
|
|
|
|
nvgpu_log_info(g, "MM suspend done!");
|
|
|
|
return err;
|
|
}
|
|
|
|
u64 nvgpu_inst_block_addr(struct gk20a *g, struct nvgpu_mem *inst_block)
|
|
{
|
|
if (nvgpu_is_enabled(g, NVGPU_SUPPORT_NVLINK)) {
|
|
return nvgpu_mem_get_phys_addr(g, inst_block);
|
|
} else {
|
|
return nvgpu_mem_get_addr(g, inst_block);
|
|
}
|
|
}
|
|
|
|
u32 nvgpu_inst_block_ptr(struct gk20a *g, struct nvgpu_mem *inst_block)
|
|
{
|
|
u64 addr = nvgpu_inst_block_addr(g, inst_block) >>
|
|
g->ops.ramin.base_shift();
|
|
|
|
nvgpu_assert(u64_hi32(addr) == 0U);
|
|
return u64_lo32(addr);
|
|
}
|
|
|
|
void nvgpu_free_inst_block(struct gk20a *g, struct nvgpu_mem *inst_block)
|
|
{
|
|
if (nvgpu_mem_is_valid(inst_block)) {
|
|
nvgpu_dma_free(g, inst_block);
|
|
}
|
|
}
|
|
|
|
int nvgpu_alloc_inst_block(struct gk20a *g, struct nvgpu_mem *inst_block)
|
|
{
|
|
int err;
|
|
|
|
nvgpu_log_fn(g, " ");
|
|
|
|
err = nvgpu_dma_alloc(g, g->ops.ramin.alloc_size(), inst_block);
|
|
if (err != 0) {
|
|
nvgpu_err(g, "%s: memory allocation failed", __func__);
|
|
return err;
|
|
}
|
|
|
|
nvgpu_log_fn(g, "done");
|
|
return 0;
|
|
}
|
|
|
|
static int nvgpu_alloc_sysmem_flush(struct gk20a *g)
|
|
{
|
|
return nvgpu_dma_alloc_sys(g, SZ_4K, &g->mm.sysmem_flush);
|
|
}
|
|
|
|
#ifdef CONFIG_NVGPU_DGPU
|
|
static void nvgpu_remove_mm_ce_support(struct mm_gk20a *mm)
|
|
{
|
|
struct gk20a *g = gk20a_from_mm(mm);
|
|
|
|
if (mm->vidmem.ce_ctx_id != NVGPU_CE_INVAL_CTX_ID) {
|
|
nvgpu_ce_app_delete_context(g, mm->vidmem.ce_ctx_id);
|
|
}
|
|
mm->vidmem.ce_ctx_id = NVGPU_CE_INVAL_CTX_ID;
|
|
|
|
nvgpu_vm_put(mm->ce.vm);
|
|
}
|
|
#endif
|
|
|
|
static void nvgpu_remove_mm_support(struct mm_gk20a *mm)
|
|
{
|
|
struct gk20a *g = gk20a_from_mm(mm);
|
|
|
|
nvgpu_dma_free(g, &mm->mmu_wr_mem);
|
|
nvgpu_dma_free(g, &mm->mmu_rd_mem);
|
|
|
|
#if defined(CONFIG_NVGPU_NON_FUSA)
|
|
if (nvgpu_fb_vab_teardown_hal(g) != 0) {
|
|
nvgpu_err(g, "failed to teardown VAB");
|
|
}
|
|
|
|
#endif
|
|
|
|
if (g->ops.mm.mmu_fault.info_mem_destroy != NULL) {
|
|
g->ops.mm.mmu_fault.info_mem_destroy(g);
|
|
}
|
|
|
|
if (g->ops.mm.remove_bar2_vm != NULL) {
|
|
g->ops.mm.remove_bar2_vm(g);
|
|
}
|
|
|
|
nvgpu_free_inst_block(g, &mm->bar1.inst_block);
|
|
nvgpu_vm_put(mm->bar1.vm);
|
|
|
|
nvgpu_free_inst_block(g, &mm->pmu.inst_block);
|
|
nvgpu_free_inst_block(g, &mm->hwpm.inst_block);
|
|
nvgpu_vm_put(mm->pmu.vm);
|
|
|
|
if (nvgpu_is_enabled(g, NVGPU_SUPPORT_SEC2_VM)) {
|
|
nvgpu_free_inst_block(g, &mm->sec2.inst_block);
|
|
nvgpu_vm_put(mm->sec2.vm);
|
|
}
|
|
|
|
if (nvgpu_is_enabled(g, NVGPU_SUPPORT_GSP_VM)) {
|
|
nvgpu_free_inst_block(g, &mm->gsp.inst_block);
|
|
nvgpu_vm_put(mm->gsp.vm);
|
|
}
|
|
|
|
if (g->has_cde) {
|
|
nvgpu_vm_put(mm->cde.vm);
|
|
}
|
|
|
|
#ifdef CONFIG_NVGPU_SW_SEMAPHORE
|
|
nvgpu_semaphore_sea_destroy(g);
|
|
#endif
|
|
#ifdef CONFIG_NVGPU_DGPU
|
|
nvgpu_vidmem_destroy(g);
|
|
|
|
if (nvgpu_is_errata_present(g, NVGPU_ERRATA_INIT_PDB_CACHE)) {
|
|
g->ops.ramin.deinit_pdb_cache_errata(g);
|
|
}
|
|
#endif
|
|
nvgpu_pd_cache_fini(g);
|
|
}
|
|
|
|
/* pmu vm, share channel_vm interfaces */
|
|
static int nvgpu_init_system_vm(struct mm_gk20a *mm)
|
|
{
|
|
int err;
|
|
struct gk20a *g = gk20a_from_mm(mm);
|
|
struct nvgpu_mem *inst_block = &mm->pmu.inst_block;
|
|
u32 big_page_size = g->ops.mm.gmmu.get_default_big_page_size();
|
|
u64 low_hole, aperture_size;
|
|
|
|
/*
|
|
* For some reason the maxwell PMU code is dependent on the large page
|
|
* size. No reason AFAICT for this. Probably a bug somewhere.
|
|
*/
|
|
if (nvgpu_is_errata_present(g, NVGPU_ERRATA_MM_FORCE_128K_PMU_VM)) {
|
|
big_page_size = nvgpu_safe_cast_u64_to_u32(SZ_128K);
|
|
}
|
|
|
|
/*
|
|
* No user region - so we will pass that as zero sized.
|
|
*/
|
|
low_hole = SZ_4K * 16UL;
|
|
aperture_size = GK20A_PMU_VA_SIZE;
|
|
|
|
mm->pmu.aperture_size = GK20A_PMU_VA_SIZE;
|
|
nvgpu_log_info(g, "pmu vm size = 0x%x", mm->pmu.aperture_size);
|
|
|
|
mm->pmu.vm = nvgpu_vm_init(g, big_page_size,
|
|
low_hole,
|
|
0ULL,
|
|
nvgpu_safe_sub_u64(aperture_size, low_hole),
|
|
0ULL,
|
|
true,
|
|
false,
|
|
false,
|
|
"system");
|
|
if (mm->pmu.vm == NULL) {
|
|
return -ENOMEM;
|
|
}
|
|
|
|
err = nvgpu_alloc_inst_block(g, inst_block);
|
|
if (err != 0) {
|
|
goto clean_up_vm;
|
|
}
|
|
g->ops.mm.init_inst_block(inst_block, mm->pmu.vm, big_page_size);
|
|
|
|
return 0;
|
|
|
|
clean_up_vm:
|
|
nvgpu_vm_put(mm->pmu.vm);
|
|
return err;
|
|
}
|
|
|
|
static int nvgpu_init_hwpm(struct mm_gk20a *mm)
|
|
{
|
|
int err;
|
|
struct gk20a *g = gk20a_from_mm(mm);
|
|
struct nvgpu_mem *inst_block = &mm->hwpm.inst_block;
|
|
|
|
err = nvgpu_alloc_inst_block(g, inst_block);
|
|
if (err != 0) {
|
|
return err;
|
|
}
|
|
g->ops.mm.init_inst_block(inst_block, mm->pmu.vm, 0);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int nvgpu_init_cde_vm(struct mm_gk20a *mm)
|
|
{
|
|
struct gk20a *g = gk20a_from_mm(mm);
|
|
u64 user_size, kernel_size;
|
|
u32 big_page_size = g->ops.mm.gmmu.get_default_big_page_size();
|
|
|
|
g->ops.mm.get_default_va_sizes(NULL, &user_size, &kernel_size);
|
|
|
|
mm->cde.vm = nvgpu_vm_init(g, big_page_size,
|
|
U64(big_page_size) << U64(10),
|
|
nvgpu_safe_sub_u64(user_size,
|
|
U64(big_page_size) << U64(10)),
|
|
kernel_size,
|
|
0ULL,
|
|
false, false, false, "cde");
|
|
if (mm->cde.vm == NULL) {
|
|
return -ENOMEM;
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
static int nvgpu_init_ce_vm(struct mm_gk20a *mm)
|
|
{
|
|
struct gk20a *g = gk20a_from_mm(mm);
|
|
u64 user_size, kernel_size;
|
|
u32 big_page_size = g->ops.mm.gmmu.get_default_big_page_size();
|
|
|
|
g->ops.mm.get_default_va_sizes(NULL, &user_size, &kernel_size);
|
|
|
|
mm->ce.vm = nvgpu_vm_init(g, big_page_size,
|
|
U64(big_page_size) << U64(10),
|
|
nvgpu_safe_sub_u64(user_size,
|
|
U64(big_page_size) << U64(10)),
|
|
kernel_size,
|
|
0ULL,
|
|
false, false, false, "ce");
|
|
if (mm->ce.vm == NULL) {
|
|
return -ENOMEM;
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
static int nvgpu_init_mmu_debug(struct mm_gk20a *mm)
|
|
{
|
|
struct gk20a *g = gk20a_from_mm(mm);
|
|
int err;
|
|
|
|
if (!nvgpu_mem_is_valid(&mm->mmu_wr_mem)) {
|
|
err = nvgpu_dma_alloc_sys(g, SZ_4K, &mm->mmu_wr_mem);
|
|
if (err != 0) {
|
|
goto err;
|
|
}
|
|
}
|
|
|
|
if (!nvgpu_mem_is_valid(&mm->mmu_rd_mem)) {
|
|
err = nvgpu_dma_alloc_sys(g, SZ_4K, &mm->mmu_rd_mem);
|
|
if (err != 0) {
|
|
goto err_free_wr_mem;
|
|
}
|
|
}
|
|
return 0;
|
|
|
|
err_free_wr_mem:
|
|
nvgpu_dma_free(g, &mm->mmu_wr_mem);
|
|
err:
|
|
return -ENOMEM;
|
|
}
|
|
|
|
#if defined(CONFIG_NVGPU_DGPU)
|
|
void nvgpu_init_mm_ce_context(struct gk20a *g)
|
|
{
|
|
if (g->mm.vidmem.size > 0U &&
|
|
(g->mm.vidmem.ce_ctx_id == NVGPU_CE_INVAL_CTX_ID)) {
|
|
g->mm.vidmem.ce_ctx_id =
|
|
nvgpu_ce_app_create_context(g,
|
|
nvgpu_engine_get_fast_ce_runlist_id(g),
|
|
-1,
|
|
-1);
|
|
|
|
if (g->mm.vidmem.ce_ctx_id == NVGPU_CE_INVAL_CTX_ID) {
|
|
nvgpu_err(g,
|
|
"Failed to allocate CE context for vidmem page clearing support");
|
|
}
|
|
}
|
|
}
|
|
#endif
|
|
|
|
static int nvgpu_init_bar1_vm(struct mm_gk20a *mm)
|
|
{
|
|
int err;
|
|
struct gk20a *g = gk20a_from_mm(mm);
|
|
struct nvgpu_mem *inst_block = &mm->bar1.inst_block;
|
|
u32 big_page_size = g->ops.mm.gmmu.get_default_big_page_size();
|
|
|
|
mm->bar1.aperture_size = bar1_aperture_size_mb_gk20a() << 20;
|
|
nvgpu_log_info(g, "bar1 vm size = 0x%x", mm->bar1.aperture_size);
|
|
mm->bar1.vm = nvgpu_vm_init(g,
|
|
big_page_size,
|
|
SZ_64K,
|
|
0ULL,
|
|
nvgpu_safe_sub_u64(mm->bar1.aperture_size, SZ_64K),
|
|
0ULL,
|
|
true, false, false,
|
|
"bar1");
|
|
if (mm->bar1.vm == NULL) {
|
|
return -ENOMEM;
|
|
}
|
|
|
|
err = nvgpu_alloc_inst_block(g, inst_block);
|
|
if (err != 0) {
|
|
goto clean_up_vm;
|
|
}
|
|
g->ops.mm.init_inst_block(inst_block, mm->bar1.vm, big_page_size);
|
|
|
|
return 0;
|
|
|
|
clean_up_vm:
|
|
nvgpu_vm_put(mm->bar1.vm);
|
|
return err;
|
|
}
|
|
|
|
static int nvgpu_init_engine_ucode_vm(struct gk20a *g,
|
|
struct engine_ucode *ucode, const char *address_space_name)
|
|
{
|
|
int err;
|
|
struct nvgpu_mem *inst_block = &ucode->inst_block;
|
|
u32 big_page_size = g->ops.mm.gmmu.get_default_big_page_size();
|
|
|
|
/* ucode aperture size is 32MB */
|
|
ucode->aperture_size = U32(32) << 20U;
|
|
nvgpu_log_info(g, "%s vm size = 0x%x", address_space_name,
|
|
ucode->aperture_size);
|
|
|
|
ucode->vm = nvgpu_vm_init(g, big_page_size, SZ_4K,
|
|
0ULL, nvgpu_safe_sub_u64(ucode->aperture_size, SZ_4K), 0ULL,
|
|
false, false, false,
|
|
address_space_name);
|
|
if (ucode->vm == NULL) {
|
|
return -ENOMEM;
|
|
}
|
|
|
|
/* allocate instance mem for engine ucode */
|
|
err = nvgpu_alloc_inst_block(g, inst_block);
|
|
if (err != 0) {
|
|
goto clean_up_va;
|
|
}
|
|
|
|
g->ops.mm.init_inst_block(inst_block, ucode->vm, big_page_size);
|
|
|
|
return 0;
|
|
|
|
clean_up_va:
|
|
nvgpu_vm_put(ucode->vm);
|
|
return err;
|
|
}
|
|
|
|
static int nvgpu_init_mm_setup_bar(struct gk20a *g)
|
|
{
|
|
struct mm_gk20a *mm = &g->mm;
|
|
int err;
|
|
|
|
err = nvgpu_init_bar1_vm(mm);
|
|
if (err != 0) {
|
|
return err;
|
|
}
|
|
|
|
if (g->ops.mm.init_bar2_vm != NULL) {
|
|
err = g->ops.mm.init_bar2_vm(g);
|
|
if (err != 0) {
|
|
return err;
|
|
}
|
|
}
|
|
err = nvgpu_init_system_vm(mm);
|
|
if (err != 0) {
|
|
return err;
|
|
}
|
|
|
|
err = nvgpu_init_hwpm(mm);
|
|
if (err != 0) {
|
|
return err;
|
|
}
|
|
|
|
return err;
|
|
}
|
|
|
|
static int nvgpu_init_mm_setup_vm(struct gk20a *g)
|
|
{
|
|
struct mm_gk20a *mm = &g->mm;
|
|
int err;
|
|
|
|
if (nvgpu_is_enabled(g, NVGPU_SUPPORT_SEC2_VM)) {
|
|
err = nvgpu_init_engine_ucode_vm(g, &mm->sec2, "sec2");
|
|
if (err != 0) {
|
|
return err;
|
|
}
|
|
}
|
|
|
|
if (nvgpu_is_enabled(g, NVGPU_SUPPORT_GSP_VM)) {
|
|
err = nvgpu_init_engine_ucode_vm(g, &mm->gsp, "gsp");
|
|
if (err != 0) {
|
|
return err;
|
|
}
|
|
}
|
|
|
|
if (g->has_cde) {
|
|
err = nvgpu_init_cde_vm(mm);
|
|
if (err != 0) {
|
|
return err;
|
|
}
|
|
}
|
|
|
|
err = nvgpu_init_ce_vm(mm);
|
|
if (err != 0) {
|
|
return err;
|
|
}
|
|
|
|
return err;
|
|
}
|
|
|
|
static int nvgpu_init_mm_components(struct gk20a *g)
|
|
{
|
|
int err = 0;
|
|
struct mm_gk20a *mm = &g->mm;
|
|
|
|
err = nvgpu_alloc_sysmem_flush(g);
|
|
if (err != 0) {
|
|
return err;
|
|
}
|
|
|
|
err = nvgpu_init_mm_setup_bar(g);
|
|
if (err != 0) {
|
|
return err;
|
|
}
|
|
|
|
err = nvgpu_init_mm_setup_vm(g);
|
|
if (err != 0) {
|
|
return err;
|
|
}
|
|
|
|
err = nvgpu_init_mmu_debug(mm);
|
|
if (err != 0) {
|
|
return err;
|
|
}
|
|
|
|
/*
|
|
* Some chips support replayable MMU faults. For such chips make sure
|
|
* SW is initialized.
|
|
*/
|
|
if (g->ops.mm.mmu_fault.setup_sw != NULL) {
|
|
err = g->ops.mm.mmu_fault.setup_sw(g);
|
|
if (err != 0) {
|
|
return err;
|
|
}
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int nvgpu_init_mm_setup_sw(struct gk20a *g)
|
|
{
|
|
struct mm_gk20a *mm = &g->mm;
|
|
int err = 0;
|
|
|
|
if (mm->sw_ready) {
|
|
nvgpu_log_info(g, "skip init");
|
|
return 0;
|
|
}
|
|
|
|
mm->g = g;
|
|
nvgpu_mutex_init(&mm->l2_op_lock);
|
|
|
|
/*TBD: make channel vm size configurable */
|
|
g->ops.mm.get_default_va_sizes(NULL, &mm->channel.user_size,
|
|
&mm->channel.kernel_size);
|
|
|
|
nvgpu_log_info(g, "channel vm size: user %uMB kernel %uMB",
|
|
nvgpu_safe_cast_u64_to_u32(mm->channel.user_size >> U64(20)),
|
|
nvgpu_safe_cast_u64_to_u32(mm->channel.kernel_size >> U64(20)));
|
|
|
|
#ifdef CONFIG_NVGPU_DGPU
|
|
mm->vidmem.ce_ctx_id = NVGPU_CE_INVAL_CTX_ID;
|
|
|
|
nvgpu_init_pramin(mm);
|
|
|
|
err = nvgpu_vidmem_init(mm);
|
|
if (err != 0) {
|
|
return err;
|
|
}
|
|
|
|
/*
|
|
* this requires fixed allocations in vidmem which must be
|
|
* allocated before all other buffers
|
|
*/
|
|
|
|
if (!nvgpu_is_enabled(g, NVGPU_MM_UNIFIED_MEMORY) &&
|
|
nvgpu_is_enabled(g, NVGPU_SEC_PRIVSECURITY)) {
|
|
err = nvgpu_acr_alloc_blob_prerequisite(g, g->acr, 0);
|
|
if (err != 0) {
|
|
return err;
|
|
}
|
|
}
|
|
#endif
|
|
|
|
err = nvgpu_init_mm_components(g);
|
|
if (err != 0) {
|
|
return err;
|
|
}
|
|
|
|
if ((g->ops.fb.ecc.init != NULL) && !g->ecc.initialized) {
|
|
err = g->ops.fb.ecc.init(g);
|
|
if (err != 0) {
|
|
return err;
|
|
}
|
|
}
|
|
|
|
#if defined(CONFIG_NVGPU_NON_FUSA)
|
|
if (nvgpu_fb_vab_init_hal(g) != 0) {
|
|
nvgpu_err(g, "failed to init VAB");
|
|
}
|
|
#endif
|
|
|
|
mm->remove_support = nvgpu_remove_mm_support;
|
|
#ifdef CONFIG_NVGPU_DGPU
|
|
mm->remove_ce_support = nvgpu_remove_mm_ce_support;
|
|
#endif
|
|
|
|
mm->sw_ready = true;
|
|
|
|
return 0;
|
|
}
|
|
|
|
#ifdef CONFIG_NVGPU_DGPU
|
|
static int nvgpu_init_mm_pdb_cache_errata(struct gk20a *g)
|
|
{
|
|
int err;
|
|
|
|
if (nvgpu_is_errata_present(g, NVGPU_ERRATA_INIT_PDB_CACHE)) {
|
|
err = g->ops.ramin.init_pdb_cache_errata(g);
|
|
if (err != 0) {
|
|
return err;
|
|
}
|
|
}
|
|
|
|
if (nvgpu_is_errata_present(g, NVGPU_ERRATA_FB_PDB_CACHE)) {
|
|
err = g->ops.fb.apply_pdb_cache_errata(g);
|
|
if (err != 0) {
|
|
return err;
|
|
}
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
#endif
|
|
|
|
/*
|
|
* Called through the HAL to handle vGPU: the vGPU doesn't have HW to initialize
|
|
* here.
|
|
*/
|
|
int nvgpu_mm_setup_hw(struct gk20a *g)
|
|
{
|
|
struct mm_gk20a *mm = &g->mm;
|
|
int err;
|
|
|
|
nvgpu_log_fn(g, " ");
|
|
|
|
if (g->ops.fb.set_mmu_page_size != NULL) {
|
|
g->ops.fb.set_mmu_page_size(g);
|
|
}
|
|
|
|
#ifdef CONFIG_NVGPU_COMPRESSION
|
|
if (g->ops.fb.set_use_full_comp_tag_line != NULL) {
|
|
mm->use_full_comp_tag_line =
|
|
g->ops.fb.set_use_full_comp_tag_line(g);
|
|
}
|
|
#endif
|
|
|
|
g->ops.fb.init_hw(g);
|
|
|
|
if (g->ops.bus.bar1_bind != NULL) {
|
|
err = g->ops.bus.bar1_bind(g, &mm->bar1.inst_block);
|
|
if (err != 0) {
|
|
return err;
|
|
}
|
|
}
|
|
|
|
if (g->ops.bus.bar2_bind != NULL) {
|
|
err = g->ops.bus.bar2_bind(g, &mm->bar2.inst_block);
|
|
if (err != 0) {
|
|
return err;
|
|
}
|
|
}
|
|
|
|
if ((g->ops.mm.cache.fb_flush(g) != 0) ||
|
|
(g->ops.mm.cache.fb_flush(g) != 0)) {
|
|
return -EBUSY;
|
|
}
|
|
|
|
if (g->ops.mm.mmu_fault.setup_hw != NULL) {
|
|
g->ops.mm.mmu_fault.setup_hw(g);
|
|
}
|
|
|
|
nvgpu_log_fn(g, "done");
|
|
return 0;
|
|
}
|
|
|
|
int nvgpu_init_mm_support(struct gk20a *g)
|
|
{
|
|
int err;
|
|
|
|
#ifdef CONFIG_NVGPU_DGPU
|
|
err = nvgpu_init_mm_pdb_cache_errata(g);
|
|
if (err != 0) {
|
|
return err;
|
|
}
|
|
#endif
|
|
|
|
err = nvgpu_init_mm_setup_sw(g);
|
|
if (err != 0) {
|
|
return err;
|
|
}
|
|
|
|
if (g->ops.mm.setup_hw != NULL) {
|
|
err = g->ops.mm.setup_hw(g);
|
|
}
|
|
|
|
return err;
|
|
}
|
|
|
|
u32 nvgpu_mm_get_default_big_page_size(struct gk20a *g)
|
|
{
|
|
u32 big_page_size;
|
|
|
|
big_page_size = g->ops.mm.gmmu.get_default_big_page_size();
|
|
|
|
if (g->mm.disable_bigpage) {
|
|
big_page_size = 0;
|
|
}
|
|
|
|
return big_page_size;
|
|
}
|
|
|
|
u32 nvgpu_mm_get_available_big_page_sizes(struct gk20a *g)
|
|
{
|
|
u32 available_big_page_sizes = 0;
|
|
|
|
if (g->mm.disable_bigpage) {
|
|
return available_big_page_sizes;
|
|
}
|
|
|
|
available_big_page_sizes = g->ops.mm.gmmu.get_default_big_page_size();
|
|
if (g->ops.mm.gmmu.get_big_page_sizes != NULL) {
|
|
available_big_page_sizes |= g->ops.mm.gmmu.get_big_page_sizes();
|
|
}
|
|
|
|
return available_big_page_sizes;
|
|
}
|