Files
linux-nvgpu/drivers/gpu/nvgpu/hal/fifo/userd_gk20a.c
Debarshi Dutta 17486ec1f6 gpu: nvgpu: rename tsg_gk20a and channel_gk20a structs
rename struct tsg_gk20a to struct nvgpu_tsg and rename struct
channel_gk20a to struct nvgpu_channel

Jira NVGPU-3248

Change-Id: I2a227347d249f9eea59223d82f09eae23dfc1306
Signed-off-by: Debarshi Dutta <ddutta@nvidia.com>
Reviewed-on: https://git-master.nvidia.com/r/2112424
GVS: Gerrit_Virtual_Submit
Reviewed-by: Vijayakumar Subbu <vsubbu@nvidia.com>
Reviewed-by: mobile promotions <svcmobile_promotions@nvidia.com>
Tested-by: mobile promotions <svcmobile_promotions@nvidia.com>
2019-05-06 02:56:53 -07:00

90 lines
3.1 KiB
C

/*
* GK20A USERD
*
* Copyright (c) 2011-2019, NVIDIA CORPORATION. All rights reserved.
*
* Permission is hereby granted, free of charge, to any person obtaining a
* copy of this software and associated documentation files (the "Software"),
* to deal in the Software without restriction, including without limitation
* the rights to use, copy, modify, merge, publish, distribute, sublicense,
* and/or sell copies of the Software, and to permit persons to whom the
* Software is furnished to do so, subject to the following conditions:
*
* The above copyright notice and this permission notice shall be included in
* all copies or substantial portions of the Software.
*
* THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
* IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
* FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
* THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
* LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING
* FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER
* DEALINGS IN THE SOFTWARE.
*/
#include <nvgpu/bug.h>
#include <nvgpu/channel.h>
#include <nvgpu/gk20a.h>
#include <nvgpu/io.h>
#include <nvgpu/nvgpu_mem.h>
#include "userd_gk20a.h"
#include <nvgpu/hw/gk20a/hw_ram_gk20a.h>
void gk20a_userd_init_mem(struct gk20a *g, struct nvgpu_channel *c)
{
struct nvgpu_mem *mem = c->userd_mem;
u32 offset = c->userd_offset / U32(sizeof(u32));
nvgpu_log_fn(g, " ");
nvgpu_mem_wr32(g, mem, offset + ram_userd_put_w(), 0);
nvgpu_mem_wr32(g, mem, offset + ram_userd_get_w(), 0);
nvgpu_mem_wr32(g, mem, offset + ram_userd_ref_w(), 0);
nvgpu_mem_wr32(g, mem, offset + ram_userd_put_hi_w(), 0);
nvgpu_mem_wr32(g, mem, offset + ram_userd_gp_top_level_get_w(), 0);
nvgpu_mem_wr32(g, mem, offset + ram_userd_gp_top_level_get_hi_w(), 0);
nvgpu_mem_wr32(g, mem, offset + ram_userd_get_hi_w(), 0);
nvgpu_mem_wr32(g, mem, offset + ram_userd_gp_get_w(), 0);
nvgpu_mem_wr32(g, mem, offset + ram_userd_gp_put_w(), 0);
}
u32 gk20a_userd_gp_get(struct gk20a *g, struct nvgpu_channel *c)
{
u64 userd_gpu_va = gk20a_channel_userd_gpu_va(c);
u64 addr = userd_gpu_va + sizeof(u32) * ram_userd_gp_get_w();
BUG_ON(u64_hi32(addr) != 0U);
return nvgpu_bar1_readl(g, (u32)addr);
}
u64 gk20a_userd_pb_get(struct gk20a *g, struct nvgpu_channel *c)
{
u64 userd_gpu_va = gk20a_channel_userd_gpu_va(c);
u64 lo_addr = userd_gpu_va + sizeof(u32) * ram_userd_get_w();
u64 hi_addr = userd_gpu_va + sizeof(u32) * ram_userd_get_hi_w();
u32 lo, hi;
BUG_ON((u64_hi32(lo_addr) != 0U) || (u64_hi32(hi_addr) != 0U));
lo = nvgpu_bar1_readl(g, (u32)lo_addr);
hi = nvgpu_bar1_readl(g, (u32)hi_addr);
return ((u64)hi << 32) | lo;
}
void gk20a_userd_gp_put(struct gk20a *g, struct nvgpu_channel *c)
{
u64 userd_gpu_va = gk20a_channel_userd_gpu_va(c);
u64 addr = userd_gpu_va + sizeof(u32) * ram_userd_gp_put_w();
BUG_ON(u64_hi32(addr) != 0U);
nvgpu_bar1_writel(g, (u32)addr, c->gpfifo.put);
}
u32 gk20a_userd_entry_size(struct gk20a *g)
{
return BIT32(ram_userd_base_shift_v());
}