mirror of
git://nv-tegra.nvidia.com/linux-nvgpu.git
synced 2025-12-23 09:57:08 +03:00
The basic nvgpu_mem_sgl implementation provides support for OS specific scatter-gather list implementations by simply copying them node by node. This is inefficient, taking extra time and memory. This patch implements an nvgpu_mem_sgt struct to act as a header which is inserted at the front of any scatter- gather list implementation. This labels every struct with a set of ops which can be used to interact with the attached scatter gather list. Since nvgpu common code only has to interact with these function pointers, any sgl implementation can be used. Initialization only requires the allocation of a single struct, removing the need to copy or iterate through the sgl being converted. Jira NVGPU-186 Change-Id: I2994f804a4a4cc141b702e987e9081d8560ba2e8 Signed-off-by: Sunny He <suhe@nvidia.com> Reviewed-on: https://git-master.nvidia.com/r/1541426 Reviewed-by: mobile promotions <svcmobile_promotions@nvidia.com> Tested-by: mobile promotions <svcmobile_promotions@nvidia.com>
67 lines
2.0 KiB
C
67 lines
2.0 KiB
C
/*
|
|
* Copyright (c) 2017, NVIDIA CORPORATION. All rights reserved.
|
|
*
|
|
* This program is free software; you can redistribute it and/or modify it
|
|
* under the terms and conditions of the GNU General Public License,
|
|
* version 2, as published by the Free Software Foundation.
|
|
*
|
|
* This program is distributed in the hope it will be useful, but WITHOUT
|
|
* ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
|
|
* FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for
|
|
* more details.
|
|
*
|
|
* You should have received a copy of the GNU General Public License
|
|
* along with this program. If not, see <http://www.gnu.org/licenses/>.
|
|
*/
|
|
|
|
#include <nvgpu/page_allocator.h>
|
|
#include <nvgpu/bug.h>
|
|
|
|
#include "gk20a/gk20a.h"
|
|
#include "gk20a/mm_gk20a.h"
|
|
#include "gk20a/pramin_gk20a.h"
|
|
|
|
#include <nvgpu/hw/gk20a/hw_bus_gk20a.h>
|
|
#include <nvgpu/hw/gk20a/hw_pram_gk20a.h>
|
|
|
|
/* WARNING: returns pramin_window_lock taken, complement with pramin_exit() */
|
|
u32 gk20a_pramin_enter(struct gk20a *g, struct nvgpu_mem *mem,
|
|
struct nvgpu_sgt *sgt, void *sgl, u32 w)
|
|
{
|
|
u64 bufbase = nvgpu_sgt_get_phys(sgt, sgl);
|
|
u64 addr = bufbase + w * sizeof(u32);
|
|
u32 hi = (u32)((addr & ~(u64)0xfffff)
|
|
>> bus_bar0_window_target_bar0_window_base_shift_v());
|
|
u32 lo = (u32)(addr & 0xfffff);
|
|
u32 win = nvgpu_aperture_mask(g, mem,
|
|
bus_bar0_window_target_sys_mem_noncoherent_f(),
|
|
bus_bar0_window_target_vid_mem_f()) |
|
|
bus_bar0_window_base_f(hi);
|
|
|
|
gk20a_dbg(gpu_dbg_mem,
|
|
"0x%08x:%08x begin for %p,%p at [%llx,%llx] (sz %llx)",
|
|
hi, lo, mem, sgl, bufbase,
|
|
bufbase + nvgpu_sgt_get_phys(sgt, sgl),
|
|
nvgpu_sgt_get_length(sgt, sgl));
|
|
|
|
WARN_ON(!bufbase);
|
|
|
|
nvgpu_spinlock_acquire(&g->mm.pramin_window_lock);
|
|
|
|
if (g->mm.pramin_window != win) {
|
|
gk20a_writel(g, bus_bar0_window_r(), win);
|
|
gk20a_readl(g, bus_bar0_window_r());
|
|
g->mm.pramin_window = win;
|
|
}
|
|
|
|
return lo;
|
|
}
|
|
|
|
void gk20a_pramin_exit(struct gk20a *g, struct nvgpu_mem *mem,
|
|
void *sgl)
|
|
{
|
|
gk20a_dbg(gpu_dbg_mem, "end for %p,%p", mem, sgl);
|
|
|
|
nvgpu_spinlock_release(&g->mm.pramin_window_lock);
|
|
}
|