Files
linux-nvgpu/drivers/gpu/nvgpu/os/posix/posix-dma.c
Sagar Kamble a16cc2dde3 gpu: nvgpu: compile out vidmem from safety build
Safety build does not support vidmem. This patch compiles out vidmem
related changes - vidmem, dma alloc, cbc/acr/pmu alloc based on
vidmem and corresponding tests like pramin, page allocator &
gmmu_map_unmap_vidmem..
As vidmem is applicable only in case of DGPUs the code is compiled
out using CONFIG_NVGPU_DGPU.

JIRA NVGPU-3524

Change-Id: Ic623801112484ffc071195e828ab9f290f945d4d
Signed-off-by: Sagar Kamble <skamble@nvidia.com>
Reviewed-on: https://git-master.nvidia.com/r/2132773
GVS: Gerrit_Virtual_Submit
Reviewed-by: Vaibhav Kachore <vkachore@nvidia.com>
Reviewed-by: Deepak Nibade <dnibade@nvidia.com>
Reviewed-by: mobile promotions <svcmobile_promotions@nvidia.com>
Tested-by: mobile promotions <svcmobile_promotions@nvidia.com>
2019-06-25 04:37:08 -07:00

244 lines
6.3 KiB
C

/*
* Copyright (c) 2018-2019, NVIDIA CORPORATION. All rights reserved.
*
* Permission is hereby granted, free of charge, to any person obtaining a
* copy of this software and associated documentation files (the "Software"),
* to deal in the Software without restriction, including without limitation
* the rights to use, copy, modify, merge, publish, distribute, sublicense,
* and/or sell copies of the Software, and to permit persons to whom the
* Software is furnished to do so, subject to the following conditions:
*
* The above copyright notice and this permission notice shall be included in
* all copies or substantial portions of the Software.
*
* THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
* IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
* FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
* THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
* LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING
* FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER
* DEALINGS IN THE SOFTWARE.
*/
#include <stdlib.h>
#include <nvgpu/mm.h>
#include <nvgpu/vm.h>
#include <nvgpu/bug.h>
#include <nvgpu/dma.h>
#include <nvgpu/gmmu.h>
#include <nvgpu/kmem.h>
#include <nvgpu/nvgpu_mem.h>
#include <nvgpu/nvgpu_sgt.h>
#include <nvgpu/page_allocator.h>
#include <nvgpu/enabled.h>
#include <nvgpu/vidmem.h>
#include <nvgpu/posix/posix-fault-injection.h>
#include <nvgpu/posix/posix_vidmem.h>
#include "os_posix.h"
_Thread_local struct nvgpu_posix_fault_inj dma_fi = {
.enabled = false,
.counter = 0U,
};
struct nvgpu_posix_fault_inj *nvgpu_dma_alloc_get_fault_injection(void)
{
return &dma_fi;
}
/*
* In userspace vidmem vs sysmem is just a difference in what is placed in the
* aperture field.
*/
static int __nvgpu_do_dma_alloc(struct gk20a *g, unsigned long flags,
size_t size, struct nvgpu_mem *mem,
enum nvgpu_aperture ap)
{
void *memory;
#ifdef NVGPU_UNITTEST_FAULT_INJECTION_ENABLEMENT
if (nvgpu_posix_fault_injection_handle_call(&dma_fi)) {
return -ENOMEM;
}
#endif
memory = malloc(PAGE_ALIGN(size));
if (memory == NULL) {
return -ENOMEM;
}
memset(memory, 0, PAGE_ALIGN(size));
mem->cpu_va = memory;
mem->aperture = ap;
mem->size = size;
mem->aligned_size = PAGE_ALIGN(size);
mem->gpu_va = 0ULL;
mem->skip_wmb = true;
#ifdef CONFIG_NVGPU_DGPU
mem->vidmem_alloc = NULL;
mem->allocator = NULL;
#endif
return 0;
}
bool nvgpu_iommuable(struct gk20a *g)
{
struct nvgpu_os_posix *p = nvgpu_os_posix_from_gk20a(g);
return p->mm_is_iommuable;
}
int nvgpu_dma_alloc_flags_sys(struct gk20a *g, unsigned long flags,
size_t size, struct nvgpu_mem *mem)
{
/* note: fault injection handled in common function */
return __nvgpu_do_dma_alloc(g, flags, size, mem, APERTURE_SYSMEM);
}
#ifdef CONFIG_NVGPU_DGPU
static u64 __nvgpu_dma_alloc(struct nvgpu_allocator *allocator, u64 at,
size_t size)
{
u64 addr = 0;
if (at)
addr = nvgpu_alloc_fixed(allocator, at, size, 0);
else
addr = nvgpu_alloc(allocator, size);
return addr;
}
static size_t mock_fb_get_vidmem_size(struct gk20a *g)
{
return SZ_4G;
}
/* In userspace, vidmem requires only a few fields populated */
int nvgpu_dma_alloc_flags_vid_at(struct gk20a *g, unsigned long flags,
size_t size, struct nvgpu_mem *mem, u64 at)
{
u64 addr;
int err;
#ifdef NVGPU_UNITTEST_FAULT_INJECTION_ENABLEMENT
if (nvgpu_posix_fault_injection_handle_call(&dma_fi)) {
return -ENOMEM;
}
#endif
g->ops.fb.get_vidmem_size = mock_fb_get_vidmem_size;
nvgpu_set_enabled(g, NVGPU_MM_UNIFIED_MEMORY, false);
/* Initialize nvgpu_allocators */
err = nvgpu_vidmem_init(&g->mm);
if (err != 0) {
nvgpu_err(g, "vidmem init failed with err=%d", err);
return err;
}
struct nvgpu_allocator *vidmem_alloc = g->mm.vidmem.cleared ?
&g->mm.vidmem.allocator :
&g->mm.vidmem.bootstrap_allocator;
if (nvgpu_mem_is_valid(mem)) {
nvgpu_warn(g, "memory leak !!");
WARN_ON(1);
}
mem->size = size;
size = PAGE_ALIGN(size);
if (!nvgpu_alloc_initialized(&g->mm.vidmem.allocator)) {
err = -ENOSYS;
nvgpu_err(g, "nvgpu alloc not initialized\n");
goto dma_err;
}
/*
* Our own allocator doesn't have any flags yet, and we can't
* kernel-map these, so require explicit flags.
*/
WARN_ON(flags != NVGPU_DMA_NO_KERNEL_MAPPING);
addr = __nvgpu_dma_alloc(vidmem_alloc, at, size);
if (!addr) {
/*
* If memory is known to be freed soon, let the user know that
* it may be available after a while.
*/
err = -ENOMEM;
nvgpu_err(g, " not initialized\n");
goto dma_err;
}
if (at)
mem->mem_flags |= NVGPU_MEM_FLAG_FIXED;
/* POSIX doesn't have sg_table struct. Allocate memory for nvgpu_sgt */
mem->priv.sgt = nvgpu_kzalloc(g, sizeof(struct nvgpu_sgt));
if (!mem->priv.sgt) {
err = -ENOMEM;
goto fail_sgtfree;
}
/* Allocate memory for sgl */
mem->priv.sgt->sgl = (struct nvgpu_sgl *)
nvgpu_kzalloc(g, sizeof(struct nvgpu_mem_sgl));
if (mem->priv.sgt->sgl == NULL) {
nvgpu_err(g, "sgl allocation failed\n");
err = -ENOMEM;
goto fail_sglfree;
}
nvgpu_vidmem_set_page_alloc(
(struct nvgpu_mem_sgl *)mem->priv.sgt->sgl, addr);
mem->aligned_size = size;
mem->aperture = APERTURE_VIDMEM;
mem->vidmem_alloc = (struct nvgpu_page_alloc *)(uintptr_t)addr;
mem->allocator = vidmem_alloc;
return 0;
fail_sglfree:
nvgpu_kfree(g, mem->priv.sgt);
fail_sgtfree:
nvgpu_free(&g->mm.vidmem.allocator, addr);
mem->size = 0;
dma_err:
nvgpu_vidmem_destroy(g);
nvgpu_cond_destroy(&g->mm.vidmem.clearing_thread_cond);
nvgpu_thread_stop_graceful(&g->mm.vidmem.clearing_thread, NULL, NULL);
return err;
}
void nvgpu_dma_free_vid(struct gk20a *g, struct nvgpu_mem *mem)
{
nvgpu_memset(g, mem, 0, 0, mem->aligned_size);
nvgpu_free(mem->allocator, (u64)nvgpu_vidmem_get_page_alloc(
(struct nvgpu_mem_sgl *)mem->priv.sgt->sgl));
nvgpu_kfree(g, mem->priv.sgt);
mem->size = 0;
mem->aligned_size = 0;
mem->aperture = APERTURE_INVALID;
nvgpu_vidmem_destroy(g);
nvgpu_cond_destroy(&g->mm.vidmem.clearing_thread_cond);
nvgpu_thread_stop_graceful(&g->mm.vidmem.clearing_thread, NULL, NULL);
}
#endif
void nvgpu_dma_free_sys(struct gk20a *g, struct nvgpu_mem *mem)
{
if (!(mem->mem_flags & NVGPU_MEM_FLAG_SHADOW_COPY)) {
free(mem->cpu_va);
}
(void) memset(mem, 0, sizeof(*mem));
}