diff --git a/drivers/gpu/nvgpu/common/mm/vm.c b/drivers/gpu/nvgpu/common/mm/vm.c index 24cf845da..05ec0dd45 100644 --- a/drivers/gpu/nvgpu/common/mm/vm.c +++ b/drivers/gpu/nvgpu/common/mm/vm.c @@ -701,7 +701,7 @@ static void nvgpu_remove_mapped_buf(struct vm_gk20a *vm, vm->num_user_mapped_buffers--; } -struct nvgpu_mapped_buf *__nvgpu_vm_find_mapped_buf( +struct nvgpu_mapped_buf *nvgpu_vm_find_mapped_buf( struct vm_gk20a *vm, u64 addr) { struct nvgpu_rbtree_node *node = NULL; @@ -715,7 +715,7 @@ struct nvgpu_mapped_buf *__nvgpu_vm_find_mapped_buf( return mapped_buffer_from_rbtree_node(node); } -struct nvgpu_mapped_buf *__nvgpu_vm_find_mapped_buf_range( +struct nvgpu_mapped_buf *nvgpu_vm_find_mapped_buf_range( struct vm_gk20a *vm, u64 addr) { struct nvgpu_rbtree_node *node = NULL; @@ -729,7 +729,7 @@ struct nvgpu_mapped_buf *__nvgpu_vm_find_mapped_buf_range( return mapped_buffer_from_rbtree_node(node); } -struct nvgpu_mapped_buf *__nvgpu_vm_find_mapped_buf_less_than( +struct nvgpu_mapped_buf *nvgpu_vm_find_mapped_buf_less_than( struct vm_gk20a *vm, u64 addr) { struct nvgpu_rbtree_node *node = NULL; @@ -1210,7 +1210,7 @@ void nvgpu_vm_unmap(struct vm_gk20a *vm, u64 offset, nvgpu_mutex_acquire(&vm->update_gmmu_lock); - mapped_buffer = __nvgpu_vm_find_mapped_buf(vm, offset); + mapped_buffer = nvgpu_vm_find_mapped_buf(vm, offset); if (mapped_buffer == NULL) { goto done; } diff --git a/drivers/gpu/nvgpu/common/mm/vm_area.c b/drivers/gpu/nvgpu/common/mm/vm_area.c index 3b617448f..9ca2f0de2 100644 --- a/drivers/gpu/nvgpu/common/mm/vm_area.c +++ b/drivers/gpu/nvgpu/common/mm/vm_area.c @@ -80,7 +80,7 @@ int nvgpu_vm_area_validate_buffer(struct vm_gk20a *vm, /* check that this mapping does not collide with existing * mappings by checking the buffer with the highest GPU VA * that is less than our buffer end */ - buffer = __nvgpu_vm_find_mapped_buf_less_than( + buffer = nvgpu_vm_find_mapped_buf_less_than( vm, map_addr + map_size); if (buffer != NULL && buffer->addr + buffer->size > map_addr) { nvgpu_warn(g, "overlapping buffer map requested"); diff --git a/drivers/gpu/nvgpu/include/nvgpu/vm.h b/drivers/gpu/nvgpu/include/nvgpu/vm.h index dfe8c3f9c..12147d9c6 100644 --- a/drivers/gpu/nvgpu/include/nvgpu/vm.h +++ b/drivers/gpu/nvgpu/include/nvgpu/vm.h @@ -282,11 +282,11 @@ u64 nvgpu_os_buf_get_size(struct nvgpu_os_buffer *os_buf); /* * These all require the VM update lock to be held. */ -struct nvgpu_mapped_buf *__nvgpu_vm_find_mapped_buf( +struct nvgpu_mapped_buf *nvgpu_vm_find_mapped_buf( struct vm_gk20a *vm, u64 addr); -struct nvgpu_mapped_buf *__nvgpu_vm_find_mapped_buf_range( +struct nvgpu_mapped_buf *nvgpu_vm_find_mapped_buf_range( struct vm_gk20a *vm, u64 addr); -struct nvgpu_mapped_buf *__nvgpu_vm_find_mapped_buf_less_than( +struct nvgpu_mapped_buf *nvgpu_vm_find_mapped_buf_less_than( struct vm_gk20a *vm, u64 addr); int nvgpu_insert_mapped_buf(struct vm_gk20a *vm, diff --git a/drivers/gpu/nvgpu/os/linux/cde.c b/drivers/gpu/nvgpu/os/linux/cde.c index df80bbd3e..2670e1f92 100644 --- a/drivers/gpu/nvgpu/os/linux/cde.c +++ b/drivers/gpu/nvgpu/os/linux/cde.c @@ -68,7 +68,7 @@ static dma_addr_t gpuva_to_iova_base(struct vm_gk20a *vm, u64 gpu_vaddr) struct gk20a *g = gk20a_from_vm(vm); nvgpu_mutex_acquire(&vm->update_gmmu_lock); - buffer = __nvgpu_vm_find_mapped_buf(vm, gpu_vaddr); + buffer = nvgpu_vm_find_mapped_buf(vm, gpu_vaddr); if (buffer) addr = nvgpu_mem_get_addr_sgl(g, buffer->os_priv.sgt->sgl); nvgpu_mutex_release(&vm->update_gmmu_lock); diff --git a/drivers/gpu/nvgpu/os/linux/vm.c b/drivers/gpu/nvgpu/os/linux/vm.c index 97dd06717..5ceff59f1 100644 --- a/drivers/gpu/nvgpu/os/linux/vm.c +++ b/drivers/gpu/nvgpu/os/linux/vm.c @@ -62,7 +62,7 @@ static u32 nvgpu_vm_translate_linux_flags(struct gk20a *g, u32 flags) return core_flags; } -static struct nvgpu_mapped_buf *__nvgpu_vm_find_mapped_buf_reverse( +static struct nvgpu_mapped_buf *nvgpu_vm_find_mapped_buf_reverse( struct vm_gk20a *vm, struct dma_buf *dmabuf, u32 kind) { struct nvgpu_rbtree_node *node = NULL; @@ -95,7 +95,7 @@ int nvgpu_vm_find_buf(struct vm_gk20a *vm, u64 gpu_va, nvgpu_mutex_acquire(&vm->update_gmmu_lock); - mapped_buffer = __nvgpu_vm_find_mapped_buf_range(vm, gpu_va); + mapped_buffer = nvgpu_vm_find_mapped_buf_range(vm, gpu_va); if (!mapped_buffer) { nvgpu_mutex_release(&vm->update_gmmu_lock); return -EINVAL; @@ -129,7 +129,7 @@ struct nvgpu_mapped_buf *nvgpu_vm_find_mapping(struct vm_gk20a *vm, struct nvgpu_mapped_buf *mapped_buffer = NULL; if (flags & NVGPU_VM_MAP_FIXED_OFFSET) { - mapped_buffer = __nvgpu_vm_find_mapped_buf(vm, map_addr); + mapped_buffer = nvgpu_vm_find_mapped_buf(vm, map_addr); if (!mapped_buffer) return NULL; @@ -138,9 +138,9 @@ struct nvgpu_mapped_buf *nvgpu_vm_find_mapping(struct vm_gk20a *vm, return NULL; } else { mapped_buffer = - __nvgpu_vm_find_mapped_buf_reverse(vm, - os_buf->dmabuf, - kind); + nvgpu_vm_find_mapped_buf_reverse(vm, + os_buf->dmabuf, + kind); if (!mapped_buffer) return NULL; }