mirror of
git://nv-tegra.nvidia.com/linux-nvgpu.git
synced 2025-12-24 10:34:43 +03:00
gpu: nvgpu: sim: make ring buffer independent of PAGE_SIZE
The simulator ring buffer DMA interface supports buffers of the following sizes: 4, 8, 12 and 16K. At present, it is configured to 4K and it happens to match with the kernel PAGE_SIZE, which is used to wrap back the GET/PUT pointers once 4K is reached. However, this is not always true; for instance, take 64K pages. Hence, replace PAGE_SIZE with SIM_BFR_SIZE. Introduce macro NVGPU_CPU_PAGE_SIZE which aliases to PAGE_SIZE and replace latter with former. Bug 200658101 Jira NVGPU-6018 Change-Id: I83cc62b87291734015c51f3e5a98173549e065de Signed-off-by: Antony Clince Alex <aalex@nvidia.com> Reviewed-on: https://git-master.nvidia.com/r/c/linux-nvgpu/+/2420728 Tested-by: mobile promotions <svcmobile_promotions@nvidia.com> Reviewed-by: mobile promotions <svcmobile_promotions@nvidia.com>
This commit is contained in:
committed by
Alex Waterman
parent
09857ecd91
commit
c36752fe3d
@@ -37,7 +37,7 @@ int nvgpu_userd_init_slabs(struct gk20a *g)
|
||||
|
||||
nvgpu_mutex_init(&f->userd_mutex);
|
||||
|
||||
f->num_channels_per_slab = PAGE_SIZE / g->ops.userd.entry_size(g);
|
||||
f->num_channels_per_slab = NVGPU_CPU_PAGE_SIZE / g->ops.userd.entry_size(g);
|
||||
f->num_userd_slabs =
|
||||
DIV_ROUND_UP(f->num_channels, f->num_channels_per_slab);
|
||||
|
||||
@@ -88,7 +88,7 @@ int nvgpu_userd_init_channel(struct gk20a *g, struct nvgpu_channel *c)
|
||||
|
||||
nvgpu_mutex_acquire(&f->userd_mutex);
|
||||
if (!nvgpu_mem_is_valid(mem)) {
|
||||
err = nvgpu_dma_alloc_sys(g, PAGE_SIZE, mem);
|
||||
err = nvgpu_dma_alloc_sys(g, NVGPU_CPU_PAGE_SIZE, mem);
|
||||
if (err != 0) {
|
||||
nvgpu_err(g, "userd allocation failed, err=%d", err);
|
||||
goto done;
|
||||
@@ -96,7 +96,7 @@ int nvgpu_userd_init_channel(struct gk20a *g, struct nvgpu_channel *c)
|
||||
|
||||
if (g->ops.mm.is_bar1_supported(g)) {
|
||||
mem->gpu_va = g->ops.mm.bar1_map_userd(g, mem,
|
||||
slab * PAGE_SIZE);
|
||||
slab * NVGPU_CPU_PAGE_SIZE);
|
||||
}
|
||||
}
|
||||
c->userd_mem = mem;
|
||||
@@ -128,9 +128,9 @@ int nvgpu_userd_setup_sw(struct gk20a *g)
|
||||
}
|
||||
|
||||
size = f->num_channels * g->ops.userd.entry_size(g);
|
||||
num_pages = DIV_ROUND_UP(size, PAGE_SIZE);
|
||||
num_pages = DIV_ROUND_UP(size, NVGPU_CPU_PAGE_SIZE);
|
||||
err = nvgpu_vm_area_alloc(g->mm.bar1.vm,
|
||||
num_pages, PAGE_SIZE, &f->userd_gpu_va, 0);
|
||||
num_pages, NVGPU_CPU_PAGE_SIZE, &f->userd_gpu_va, 0);
|
||||
if (err != 0) {
|
||||
nvgpu_err(g, "userd gpu va allocation failed, err=%d", err);
|
||||
goto clean_up;
|
||||
|
||||
Reference in New Issue
Block a user