mirror of
git://nv-tegra.nvidia.com/linux-nvgpu.git
synced 2025-12-23 18:16:01 +03:00
GM20B changes in PMU boot sequence to support booting in HS mode and LS mode Bug 1509680 Change-Id: I2832eda0efe17dd5e3a8f11dd06e7d4da267be70 Signed-off-by: Supriya <ssharatkumar@nvidia.com> Reviewed-on: http://git-master/r/423140 Reviewed-by: Automatic_Commit_Validation_User Reviewed-by: Vijayakumar Subbu <vsubbu@nvidia.com> Reviewed-by: Seshendra Gadagottu <sgadagottu@nvidia.com> Tested-by: Seshendra Gadagottu <sgadagottu@nvidia.com> GVS: Gerrit_Virtual_Submit Reviewed-by: Shridhar Rasal <srasal@nvidia.com> Reviewed-by: Deepak Nibade <dnibade@nvidia.com> Reviewed-by: Bharat Nihalani <bnihalani@nvidia.com>
272 lines
6.6 KiB
C
272 lines
6.6 KiB
C
/*
|
|
* GM20B MMU
|
|
*
|
|
* Copyright (c) 2014, NVIDIA CORPORATION. All rights reserved.
|
|
*
|
|
* This program is free software; you can redistribute it and/or modify it
|
|
* under the terms and conditions of the GNU General Public License,
|
|
* version 2, as published by the Free Software Foundation.
|
|
*
|
|
* This program is distributed in the hope it will be useful, but WITHOUT
|
|
* ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
|
|
* FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for
|
|
* more details.
|
|
*/
|
|
|
|
#include <linux/pm_runtime.h>
|
|
#include "gk20a/gk20a.h"
|
|
#include "mm_gm20b.h"
|
|
#include "hw_gmmu_gm20b.h"
|
|
#include "hw_fb_gm20b.h"
|
|
|
|
static const u32 gmmu_page_sizes[gmmu_nr_page_sizes] = { SZ_4K, SZ_128K };
|
|
static const u32 gmmu_page_shifts[gmmu_nr_page_sizes] = { 12, 17 };
|
|
static const u64 gmmu_page_offset_masks[gmmu_nr_page_sizes] = { 0xfffLL,
|
|
0x1ffffLL };
|
|
static const u64 gmmu_page_masks[gmmu_nr_page_sizes] = { ~0xfffLL, ~0x1ffffLL };
|
|
|
|
static int allocate_gmmu_ptes_sparse(struct vm_gk20a *vm,
|
|
enum gmmu_pgsz_gk20a pgsz_idx,
|
|
u64 first_vaddr, u64 last_vaddr)
|
|
{
|
|
int err;
|
|
u32 pte_lo, pte_hi;
|
|
u32 pde_lo, pde_hi;
|
|
u32 pte_w[2] = {0, 0}; /* invalid pte */
|
|
u64 addr = 0;
|
|
u32 pte_cur;
|
|
void *pte_kv_cur;
|
|
struct page_table_gk20a *pte;
|
|
|
|
gk20a_dbg_fn("");
|
|
|
|
pde_range_from_vaddr_range(vm, first_vaddr, last_vaddr,
|
|
&pde_lo, &pde_hi);
|
|
|
|
gk20a_dbg(gpu_dbg_pte, "size_idx=%d, pde_lo=%d, pde_hi=%d",
|
|
pgsz_idx, pde_lo, pde_hi);
|
|
|
|
/* Expect ptes of the same pde */
|
|
BUG_ON(pde_lo != pde_hi);
|
|
|
|
pte = vm->pdes.ptes[pgsz_idx] + pde_lo;
|
|
|
|
pte_lo = pte_index_from_vaddr(vm, first_vaddr, pgsz_idx);
|
|
pte_hi = pte_index_from_vaddr(vm, last_vaddr, pgsz_idx);
|
|
|
|
/* get cpu access to the ptes */
|
|
err = map_gmmu_pages(pte->ref, pte->sgt, &pte_kv_cur, pte->size);
|
|
if (err)
|
|
goto fail;
|
|
|
|
gk20a_dbg(gpu_dbg_pte, "pte_lo=%d, pte_hi=%d", pte_lo, pte_hi);
|
|
for (pte_cur = pte_lo; pte_cur <= pte_hi; pte_cur++) {
|
|
pte_w[0] = gmmu_pte_valid_false_f();
|
|
pte_w[1] = gmmu_pte_vol_true_f();
|
|
|
|
gk20a_dbg(gpu_dbg_pte,
|
|
"pte_cur=%d addr=%llx refs=%d"
|
|
" [0x%08x,0x%08x]",
|
|
pte_cur, addr,
|
|
pte->ref_cnt, pte_w[1], pte_w[0]);
|
|
|
|
gk20a_mem_wr32(pte_kv_cur + pte_cur*8, 0, pte_w[0]);
|
|
gk20a_mem_wr32(pte_kv_cur + pte_cur*8, 1, pte_w[1]);
|
|
}
|
|
|
|
unmap_gmmu_pages(pte->ref, pte->sgt, pte_kv_cur);
|
|
|
|
smp_mb();
|
|
vm->tlb_dirty = true;
|
|
gk20a_dbg_fn("set tlb dirty");
|
|
|
|
return 0;
|
|
fail:
|
|
return err;
|
|
|
|
}
|
|
|
|
static void allocate_gmmu_pde_sparse(struct vm_gk20a *vm, u32 i)
|
|
{
|
|
bool small_valid, big_valid;
|
|
u64 pte_addr[2] = {0, 0};
|
|
struct page_table_gk20a *small_pte =
|
|
vm->pdes.ptes[gmmu_page_size_small] + i;
|
|
struct page_table_gk20a *big_pte =
|
|
vm->pdes.ptes[gmmu_page_size_big] + i;
|
|
u32 pde_v[2] = {0, 0};
|
|
u32 *pde;
|
|
|
|
gk20a_dbg_fn("");
|
|
|
|
small_valid = small_pte && small_pte->ref;
|
|
big_valid = big_pte && big_pte->ref;
|
|
|
|
if (small_valid)
|
|
pte_addr[gmmu_page_size_small] =
|
|
gk20a_mm_iova_addr(small_pte->sgt->sgl);
|
|
if (big_valid)
|
|
pte_addr[gmmu_page_size_big] =
|
|
gk20a_mm_iova_addr(big_pte->sgt->sgl);
|
|
|
|
pde_v[0] = gmmu_pde_size_full_f();
|
|
pde_v[0] |= gmmu_pde_aperture_big_invalid_f();
|
|
pde_v[1] |= gmmu_pde_aperture_small_invalid_f() |
|
|
gmmu_pde_vol_big_true_f();
|
|
|
|
pde = pde_from_index(vm, i);
|
|
|
|
gk20a_mem_wr32(pde, 0, pde_v[0]);
|
|
gk20a_mem_wr32(pde, 1, pde_v[1]);
|
|
|
|
smp_mb();
|
|
|
|
FLUSH_CPU_DCACHE(pde,
|
|
sg_phys(vm->pdes.sgt->sgl) + (i*gmmu_pde__size_v()),
|
|
sizeof(u32)*2);
|
|
|
|
gk20a_mm_l2_invalidate(vm->mm->g);
|
|
|
|
gk20a_dbg(gpu_dbg_pte, "pde:%d = 0x%x,0x%08x\n", i, pde_v[1], pde_v[0]);
|
|
|
|
vm->tlb_dirty = true;
|
|
}
|
|
|
|
static bool gm20b_vm_is_pde_in_range(struct vm_gk20a *vm, u64 vaddr_lo,
|
|
u64 vaddr_hi, u32 pde)
|
|
{
|
|
u64 pde_vaddr_lo, pde_vaddr_hi;
|
|
|
|
gk20a_dbg_fn("");
|
|
|
|
pde_vaddr_lo = (u64)pde << vm->mm->pde_stride_shift;
|
|
pde_vaddr_hi = pde_vaddr_lo |
|
|
((0x1UL << (vm->mm->pde_stride_shift)) - 1);
|
|
|
|
return ((vaddr_lo <= pde_vaddr_lo) && (vaddr_hi) >= pde_vaddr_hi);
|
|
}
|
|
|
|
static int gm20b_vm_put_sparse(struct vm_gk20a *vm, u64 vaddr,
|
|
u32 num_pages, u32 pgsz_idx)
|
|
{
|
|
struct mm_gk20a *mm = vm->mm;
|
|
u32 pgsz = gmmu_page_sizes[pgsz_idx];
|
|
u32 pde_shift = vm->mm->pde_stride_shift;
|
|
u64 vaddr_hi;
|
|
u64 vaddr_pde_start;
|
|
u32 i;
|
|
u32 pde_lo, pde_hi;
|
|
int err;
|
|
|
|
gk20a_dbg_fn("");
|
|
|
|
vaddr_hi = vaddr + pgsz * num_pages - 1;
|
|
pde_range_from_vaddr_range(vm,
|
|
vaddr,
|
|
vaddr_hi,
|
|
&pde_lo, &pde_hi);
|
|
|
|
gk20a_dbg_info("vaddr: 0x%llx, vaddr_hi: 0x%llx, pde_lo: 0x%x, "
|
|
"pde_hi: 0x%x, pgsz: %d, pde_stride_shift: %d",
|
|
vaddr, vaddr_hi, pde_lo, pde_hi,
|
|
vm->mm->pde_stride_shift, pgsz);
|
|
|
|
for (i = pde_lo; i <= pde_hi; i++) {
|
|
/* Mark all ptes as sparse. */
|
|
err = validate_gmmu_page_table_gk20a_locked(vm, i,
|
|
pgsz_idx);
|
|
if (err) {
|
|
gk20a_err(dev_from_vm(vm),
|
|
"failed to validate page table %d: %d",
|
|
i, err);
|
|
goto fail;
|
|
}
|
|
|
|
if (gm20b_vm_is_pde_in_range(vm, vaddr, vaddr_hi, i)) {
|
|
/* entire pde is marked as sparse */
|
|
vaddr_pde_start = (u64)i << pde_shift;
|
|
allocate_gmmu_ptes_sparse(vm, pgsz_idx,
|
|
vaddr_pde_start,
|
|
PDE_ADDR_END(vaddr_pde_start,
|
|
pde_shift));
|
|
} else {
|
|
/* Check leading and trailing spaces which doesn't fit
|
|
* into entire pde. */
|
|
if (pde_lo == pde_hi)
|
|
allocate_gmmu_ptes_sparse(vm, pgsz_idx, vaddr,
|
|
vaddr_hi);
|
|
else if (i == pde_lo)
|
|
allocate_gmmu_ptes_sparse(vm, pgsz_idx, vaddr,
|
|
PDE_ADDR_END(vaddr, pde_shift));
|
|
else
|
|
allocate_gmmu_ptes_sparse(vm, pgsz_idx,
|
|
PDE_ADDR_START(vaddr_hi, pde_shift),
|
|
vaddr_hi);
|
|
}
|
|
}
|
|
|
|
gk20a_mm_l2_flush(mm->g, true);
|
|
|
|
return 0;
|
|
|
|
fail:
|
|
WARN_ON(1);
|
|
|
|
return err;
|
|
}
|
|
|
|
static int gm20b_mm_mmu_vpr_info_fetch_wait(struct gk20a *g,
|
|
const unsigned int msec)
|
|
{
|
|
unsigned long timeout;
|
|
|
|
if (tegra_platform_is_silicon())
|
|
timeout = jiffies + msecs_to_jiffies(msec);
|
|
else
|
|
timeout = msecs_to_jiffies(msec);
|
|
|
|
while (1) {
|
|
u32 val;
|
|
val = gk20a_readl(g, fb_mmu_vpr_info_r());
|
|
if (fb_mmu_vpr_info_fetch_v(val) ==
|
|
fb_mmu_vpr_info_fetch_false_v())
|
|
break;
|
|
if (tegra_platform_is_silicon()) {
|
|
if (WARN_ON(time_after(jiffies, timeout)))
|
|
return -ETIME;
|
|
} else if (--timeout == 0)
|
|
return -ETIME;
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
int gm20b_mm_mmu_vpr_info_fetch(struct gk20a *g)
|
|
{
|
|
int ret = 0;
|
|
|
|
gk20a_busy_noresume(g->dev);
|
|
#ifdef CONFIG_PM_RUNTIME
|
|
if (!pm_runtime_active(&g->dev->dev))
|
|
goto fail;
|
|
#endif
|
|
|
|
if (gm20b_mm_mmu_vpr_info_fetch_wait(g, VPR_INFO_FETCH_WAIT)) {
|
|
ret = -ETIME;
|
|
goto fail;
|
|
}
|
|
|
|
gk20a_writel(g, fb_mmu_vpr_info_r(),
|
|
fb_mmu_vpr_info_fetch_true_v());
|
|
|
|
ret = gm20b_mm_mmu_vpr_info_fetch_wait(g, VPR_INFO_FETCH_WAIT);
|
|
|
|
fail:
|
|
gk20a_idle(g->dev);
|
|
return ret;
|
|
}
|
|
|
|
void gm20b_init_mm(struct gpu_ops *gops)
|
|
{
|
|
gops->mm.set_sparse = gm20b_vm_put_sparse;
|
|
}
|