mirror of
git://nv-tegra.nvidia.com/linux-nvgpu.git
synced 2025-12-23 01:50:07 +03:00
IRQs can get triggered during nvgpu power-on due to MMU fault, invalid PRIV ring or bus access etc. Handlers for those IRQs can't access the full state related to the IRQ unless nvgpu is fully powered on. In order to let the IRQ handlers know about the nvgpu power-on state gk20a.power_on_state variable has to be protected through spinlock to avoid the deadlock due to usage of earlier power_lock mutex. Further the IRQs need to be disabled on local CPU while updating the power state variable hence use spin_lock_irqsave and spin_unlock_- irqrestore APIs for protecting the access. JIRA NVGPU-1592 Change-Id: If5d1b5e2617ad90a68faa56ff47f62bb3f0b232b Signed-off-by: Sagar Kamble <skamble@nvidia.com> Reviewed-on: https://git-master.nvidia.com/r/2203860 Reviewed-by: mobile promotions <svcmobile_promotions@nvidia.com> Tested-by: mobile promotions <svcmobile_promotions@nvidia.com>
317 lines
8.5 KiB
C
317 lines
8.5 KiB
C
/*
|
|
* Copyright (c) 2018-2019, NVIDIA CORPORATION. All rights reserved.
|
|
*
|
|
* Permission is hereby granted, free of charge, to any person obtaining a
|
|
* copy of this software and associated documentation files (the "Software"),
|
|
* to deal in the Software without restriction, including without limitation
|
|
* the rights to use, copy, modify, merge, publish, distribute, sublicense,
|
|
* and/or sell copies of the Software, and to permit persons to whom the
|
|
* Software is furnished to do so, subject to the following conditions:
|
|
*
|
|
* The above copyright notice and this permission notice shall be included in
|
|
* all copies or substantial portions of the Software.
|
|
*
|
|
* THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
|
|
* IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
|
|
* FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
|
|
* THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
|
|
* LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING
|
|
* FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER
|
|
* DEALINGS IN THE SOFTWARE.
|
|
*/
|
|
|
|
#ifdef CONFIG_NVGPU_TRACE
|
|
#include <trace/events/gk20a.h>
|
|
#endif
|
|
|
|
#include <nvgpu/log.h>
|
|
#include <nvgpu/types.h>
|
|
#include <nvgpu/timers.h>
|
|
#include <nvgpu/nvgpu_mem.h>
|
|
#include <nvgpu/io.h>
|
|
#include <nvgpu/utils.h>
|
|
#include <nvgpu/gk20a.h>
|
|
#include <nvgpu/bug.h>
|
|
#include <nvgpu/nvgpu_init.h>
|
|
|
|
#include "hal/fb/fb_gv11b.h"
|
|
#include "hal/fb/fb_gv100.h"
|
|
#include "hal/mc/mc_tu104.h"
|
|
|
|
#include "hal/func/func_tu104.h"
|
|
|
|
#include "fb_tu104.h"
|
|
|
|
#include "nvgpu/hw/tu104/hw_fb_tu104.h"
|
|
#include "nvgpu/hw/tu104/hw_func_tu104.h"
|
|
|
|
int fb_tu104_tlb_invalidate(struct gk20a *g, struct nvgpu_mem *pdb)
|
|
{
|
|
struct nvgpu_timeout timeout;
|
|
u32 addr_lo;
|
|
u32 data;
|
|
int err = 0;
|
|
|
|
nvgpu_log_fn(g, " ");
|
|
|
|
/*
|
|
* pagetables are considered sw states which are preserved after
|
|
* prepare_poweroff. When gk20a deinit releases those pagetables,
|
|
* common code in vm unmap path calls tlb invalidate that touches
|
|
* hw. Use the power_on flag to skip tlb invalidation when gpu
|
|
* power is turned off
|
|
*/
|
|
if (nvgpu_is_powered_off(g)) {
|
|
return 0;
|
|
}
|
|
|
|
addr_lo = u64_lo32(nvgpu_mem_get_addr(g, pdb) >> 12);
|
|
|
|
err = nvgpu_timeout_init(g, &timeout, 1000, NVGPU_TIMER_RETRY_TIMER);
|
|
if (err != 0) {
|
|
return err;
|
|
}
|
|
|
|
nvgpu_mutex_acquire(&g->mm.tlb_lock);
|
|
|
|
#ifdef CONFIG_NVGPU_TRACE
|
|
trace_gk20a_mm_tlb_invalidate(g->name);
|
|
#endif
|
|
|
|
nvgpu_func_writel(g, func_priv_mmu_invalidate_pdb_r(),
|
|
fb_mmu_invalidate_pdb_addr_f(addr_lo) |
|
|
nvgpu_aperture_mask(g, pdb,
|
|
fb_mmu_invalidate_pdb_aperture_sys_mem_f(),
|
|
fb_mmu_invalidate_pdb_aperture_sys_mem_f(),
|
|
fb_mmu_invalidate_pdb_aperture_vid_mem_f()));
|
|
|
|
nvgpu_func_writel(g, func_priv_mmu_invalidate_r(),
|
|
fb_mmu_invalidate_all_va_true_f() |
|
|
fb_mmu_invalidate_trigger_true_f());
|
|
|
|
do {
|
|
data = nvgpu_func_readl(g,
|
|
func_priv_mmu_invalidate_r());
|
|
if (fb_mmu_invalidate_trigger_v(data) !=
|
|
fb_mmu_invalidate_trigger_true_v()) {
|
|
break;
|
|
}
|
|
nvgpu_udelay(2);
|
|
} while (nvgpu_timeout_expired_msg(&timeout,
|
|
"wait mmu invalidate") == 0);
|
|
|
|
#ifdef CONFIG_NVGPU_TRACE
|
|
trace_gk20a_mm_tlb_invalidate_done(g->name);
|
|
#endif
|
|
|
|
nvgpu_mutex_release(&g->mm.tlb_lock);
|
|
return err;
|
|
}
|
|
|
|
#ifdef CONFIG_NVGPU_COMPRESSION
|
|
void tu104_fb_cbc_configure(struct gk20a *g, struct nvgpu_cbc *cbc)
|
|
{
|
|
u64 base_divisor;
|
|
u64 compbit_store_base;
|
|
u64 compbit_store_pa;
|
|
u64 cbc_start_addr, cbc_end_addr;
|
|
u64 cbc_top;
|
|
u64 cbc_top_size;
|
|
u32 cbc_max;
|
|
|
|
compbit_store_pa = nvgpu_mem_get_addr(g, &cbc->compbit_store.mem);
|
|
base_divisor = g->ops.cbc.get_base_divisor(g);
|
|
compbit_store_base = DIV_ROUND_UP(compbit_store_pa, base_divisor);
|
|
|
|
cbc_start_addr = (u64)nvgpu_ltc_get_ltc_count(g) *
|
|
(compbit_store_base <<
|
|
fb_mmu_cbc_base_address_alignment_shift_v());
|
|
cbc_end_addr = cbc_start_addr + cbc->compbit_backing_size;
|
|
|
|
cbc_top = (cbc_end_addr / nvgpu_ltc_get_ltc_count(g)) >>
|
|
fb_mmu_cbc_base_address_alignment_shift_v();
|
|
cbc_top_size = u64_lo32(cbc_top) - compbit_store_base;
|
|
|
|
nvgpu_assert(cbc_top_size < U64(U32_MAX));
|
|
nvgpu_writel(g, fb_mmu_cbc_top_r(),
|
|
fb_mmu_cbc_top_size_f(U32(cbc_top_size)));
|
|
|
|
cbc_max = nvgpu_readl(g, fb_mmu_cbc_max_r());
|
|
cbc_max = set_field(cbc_max,
|
|
fb_mmu_cbc_max_comptagline_m(),
|
|
fb_mmu_cbc_max_comptagline_f(cbc->max_comptag_lines));
|
|
nvgpu_writel(g, fb_mmu_cbc_max_r(), cbc_max);
|
|
|
|
nvgpu_assert(compbit_store_base < U64(U32_MAX));
|
|
nvgpu_writel(g, fb_mmu_cbc_base_r(),
|
|
fb_mmu_cbc_base_address_f(U32(compbit_store_base)));
|
|
|
|
nvgpu_log(g, gpu_dbg_info | gpu_dbg_map_v | gpu_dbg_pte,
|
|
"compbit base.pa: 0x%x,%08x cbc_base:0x%llx\n",
|
|
(u32)(compbit_store_pa >> 32),
|
|
(u32)(compbit_store_pa & 0xffffffffU),
|
|
compbit_store_base);
|
|
|
|
cbc->compbit_store.base_hw = compbit_store_base;
|
|
|
|
}
|
|
#endif
|
|
|
|
static int tu104_fb_wait_mmu_bind(struct gk20a *g)
|
|
{
|
|
struct nvgpu_timeout timeout;
|
|
u32 val;
|
|
int err;
|
|
|
|
err = nvgpu_timeout_init(g, &timeout, 1000, NVGPU_TIMER_RETRY_TIMER);
|
|
if (err != 0) {
|
|
return err;
|
|
}
|
|
|
|
do {
|
|
val = nvgpu_readl(g, fb_mmu_bind_r());
|
|
if ((val & fb_mmu_bind_trigger_true_f()) !=
|
|
fb_mmu_bind_trigger_true_f()) {
|
|
return 0;
|
|
}
|
|
nvgpu_udelay(2);
|
|
} while (nvgpu_timeout_expired_msg(&timeout, "mmu bind timedout") == 0);
|
|
|
|
return -ETIMEDOUT;
|
|
}
|
|
|
|
int tu104_fb_apply_pdb_cache_war(struct gk20a *g)
|
|
{
|
|
u64 inst_blk_base_addr;
|
|
u32 inst_blk_addr;
|
|
u32 i;
|
|
int err;
|
|
|
|
if (!nvgpu_mem_is_valid(&g->pdb_cache_war_mem)) {
|
|
return -EINVAL;
|
|
}
|
|
|
|
inst_blk_base_addr = nvgpu_mem_get_addr(g, &g->pdb_cache_war_mem);
|
|
|
|
/* Bind 256 instance blocks to unused engine ID 0x0 */
|
|
for (i = 0U; i < 256U; i++) {
|
|
inst_blk_addr = u64_lo32((inst_blk_base_addr +
|
|
(U64(i) * U64(PAGE_SIZE)))
|
|
>> fb_mmu_bind_imb_addr_alignment_v());
|
|
|
|
nvgpu_writel(g, fb_mmu_bind_imb_r(),
|
|
fb_mmu_bind_imb_addr_f(inst_blk_addr) |
|
|
nvgpu_aperture_mask(g, &g->pdb_cache_war_mem,
|
|
fb_mmu_bind_imb_aperture_sys_mem_nc_f(),
|
|
fb_mmu_bind_imb_aperture_sys_mem_c_f(),
|
|
fb_mmu_bind_imb_aperture_vid_mem_f()));
|
|
|
|
nvgpu_writel(g, fb_mmu_bind_r(),
|
|
fb_mmu_bind_engine_id_f(0x0U) |
|
|
fb_mmu_bind_trigger_true_f());
|
|
|
|
err = tu104_fb_wait_mmu_bind(g);
|
|
if (err != 0) {
|
|
return err;
|
|
}
|
|
}
|
|
|
|
/* first unbind */
|
|
nvgpu_writel(g, fb_mmu_bind_imb_r(),
|
|
fb_mmu_bind_imb_aperture_f(0x1U) |
|
|
fb_mmu_bind_imb_addr_f(0x0U));
|
|
|
|
nvgpu_writel(g, fb_mmu_bind_r(),
|
|
fb_mmu_bind_engine_id_f(0x0U) |
|
|
fb_mmu_bind_trigger_true_f());
|
|
|
|
err = tu104_fb_wait_mmu_bind(g);
|
|
if (err != 0) {
|
|
return err;
|
|
}
|
|
|
|
/* second unbind */
|
|
nvgpu_writel(g, fb_mmu_bind_r(),
|
|
fb_mmu_bind_engine_id_f(0x0U) |
|
|
fb_mmu_bind_trigger_true_f());
|
|
|
|
err = tu104_fb_wait_mmu_bind(g);
|
|
if (err != 0) {
|
|
return err;
|
|
}
|
|
|
|
/* Bind 257th (last) instance block that reserves PDB cache entry 255 */
|
|
inst_blk_addr = u64_lo32((inst_blk_base_addr + (256ULL * U64(PAGE_SIZE)))
|
|
>> U64(fb_mmu_bind_imb_addr_alignment_v()));
|
|
|
|
nvgpu_writel(g, fb_mmu_bind_imb_r(),
|
|
fb_mmu_bind_imb_addr_f(inst_blk_addr) |
|
|
nvgpu_aperture_mask(g, &g->pdb_cache_war_mem,
|
|
fb_mmu_bind_imb_aperture_sys_mem_nc_f(),
|
|
fb_mmu_bind_imb_aperture_sys_mem_c_f(),
|
|
fb_mmu_bind_imb_aperture_vid_mem_f()));
|
|
|
|
nvgpu_writel(g, fb_mmu_bind_r(),
|
|
fb_mmu_bind_engine_id_f(0x0U) |
|
|
fb_mmu_bind_trigger_true_f());
|
|
|
|
err = tu104_fb_wait_mmu_bind(g);
|
|
if (err != 0) {
|
|
return err;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
#ifdef CONFIG_NVGPU_DGPU
|
|
size_t tu104_fb_get_vidmem_size(struct gk20a *g)
|
|
{
|
|
u32 range = gk20a_readl(g, fb_mmu_local_memory_range_r());
|
|
u32 mag = fb_mmu_local_memory_range_lower_mag_v(range);
|
|
u32 scale = fb_mmu_local_memory_range_lower_scale_v(range);
|
|
u32 ecc = fb_mmu_local_memory_range_ecc_mode_v(range);
|
|
size_t bytes = ((size_t)mag << scale) * SZ_1M;
|
|
|
|
#ifdef CONFIG_NVGPU_SIM
|
|
if (nvgpu_is_enabled(g, NVGPU_IS_FMODEL) && (bytes == 0)) {
|
|
/* 192 MB */
|
|
bytes = 192*1024*1024;
|
|
}
|
|
#endif
|
|
|
|
if (ecc != 0U) {
|
|
bytes = bytes / 16U * 15U;
|
|
}
|
|
|
|
return bytes;
|
|
}
|
|
#endif
|
|
|
|
int tu104_fb_enable_nvlink(struct gk20a *g)
|
|
{
|
|
int ret = 0;
|
|
u32 data;
|
|
|
|
nvgpu_log(g, gpu_dbg_nvlink|gpu_dbg_info, "enabling nvlink");
|
|
|
|
ret = gv100_fb_enable_nvlink(g);
|
|
if (ret != 0) {
|
|
return ret;
|
|
}
|
|
|
|
/* NV_PFB_PRI_MMU_CTRL_ATOMIC_CAPABILITY_SYS_NCOH_MODE to L2 */
|
|
data = nvgpu_readl(g, fb_mmu_ctrl_r());
|
|
data = set_field(data, fb_mmu_ctrl_atomic_capability_sys_ncoh_mode_m(),
|
|
fb_mmu_ctrl_atomic_capability_sys_ncoh_mode_l2_f());
|
|
nvgpu_writel(g, fb_mmu_ctrl_r(), data);
|
|
|
|
/* NV_PFB_FBHUB_NUM_ACTIVE_LTCS_HUB_SYS_NCOH_ATOMIC_MODE to USE_READ */
|
|
data = nvgpu_readl(g, fb_fbhub_num_active_ltcs_r());
|
|
data = set_field(data,
|
|
fb_fbhub_num_active_ltcs_hub_sys_ncoh_atomic_mode_m(),
|
|
fb_fbhub_num_active_ltcs_hub_sys_ncoh_atomic_mode_use_read_f());
|
|
nvgpu_writel(g, fb_fbhub_num_active_ltcs_r(), data);
|
|
|
|
return ret;
|
|
}
|