gpu: nvgpu: remove dependency on linux header for sim_gk20a*

This patch removes linux dependencies from sim_gk20a.h under
gk20a/sim_gk20a.h. The following changes are made in this patch.

1) Created a linux based structure sim_gk20a_linux that contains a
common sim_gk20a struct inside it. The common struct sim_gk20a doesn't
contain any linux specific structs.
2) The common struct sim_gk20a contains an added function pointer which
is used to invoke gk20a_sim_esc_readl() method.
3) sim_gk20a.c is moved to nvgpu/common/linux along with a new header
sim_gk20a.h that contains the definition of struct sim_gk20a_linux.
4) struct gk20a now contains a pointer of sim_gk20a instead of the
entire object. The memory for this struct is allocated and  initialized during
gk20a_init_support() and freed during invocation of
gk20_remove_support().
5) We first obtain the pointer for struct sim_gk20a_linux from the
pointer of sim_gk20a using the container_of method in order to work on
the struct.

JIRA NVGPU-386

Change-Id: Ic82b8702642377f82694577a53c3ca0b9c1bb2ab
Signed-off-by: Debarshi Dutta <ddutta@nvidia.com>
Reviewed-on: https://git-master.nvidia.com/r/1603073
GVS: Gerrit_Virtual_Submit
Reviewed-by: Deepak Nibade <dnibade@nvidia.com>
Reviewed-by: Terje Bergstrom <tbergstrom@nvidia.com>
Reviewed-by: mobile promotions <svcmobile_promotions@nvidia.com>
Tested-by: mobile promotions <svcmobile_promotions@nvidia.com>
This commit is contained in:
Debarshi Dutta
2017-11-22 13:20:19 +05:30
committed by mobile promotions
parent 830d3f10ca
commit 312f6c2c5f
7 changed files with 268 additions and 204 deletions

View File

@@ -48,6 +48,7 @@ nvgpu-y := \
common/linux/sched.o \ common/linux/sched.o \
common/linux/channel.o \ common/linux/channel.o \
common/linux/ce2.o \ common/linux/ce2.o \
common/linux/sim_gk20a.o \
common/mm/nvgpu_allocator.o \ common/mm/nvgpu_allocator.o \
common/mm/bitmap_allocator.o \ common/mm/bitmap_allocator.o \
common/mm/buddy_allocator.o \ common/mm/buddy_allocator.o \
@@ -99,7 +100,6 @@ nvgpu-y := \
gk20a/tsg_gk20a.o \ gk20a/tsg_gk20a.o \
gk20a/fecs_trace_gk20a.o \ gk20a/fecs_trace_gk20a.o \
gk20a/mc_gk20a.o \ gk20a/mc_gk20a.o \
gk20a/sim_gk20a.o \
gm20b/hal_gm20b.o \ gm20b/hal_gm20b.o \
gm20b/bus_gm20b.o \ gm20b/bus_gm20b.o \
gm20b/ltc_gm20b.o \ gm20b/ltc_gm20b.o \

View File

@@ -47,6 +47,7 @@
#include "intr.h" #include "intr.h"
#include "cde.h" #include "cde.h"
#include "ioctl.h" #include "ioctl.h"
#include "sim_gk20a.h"
#ifdef CONFIG_TEGRA_19x_GPU #ifdef CONFIG_TEGRA_19x_GPU
#include "nvgpu_gpuid_t19x.h" #include "nvgpu_gpuid_t19x.h"
#ifdef CONFIG_TEGRA_GR_VIRTUALIZATION #ifdef CONFIG_TEGRA_GR_VIRTUALIZATION
@@ -637,11 +638,10 @@ void gk20a_remove_support(struct gk20a *g)
if (g->mm.remove_support) if (g->mm.remove_support)
g->mm.remove_support(&g->mm); g->mm.remove_support(&g->mm);
if (g->sim.remove_support) if (g->sim->remove_support)
g->sim.remove_support(&g->sim); g->sim->remove_support(g->sim);
/* free mappings to registers, etc */ /* free mappings to registers, etc */
if (l->regs) { if (l->regs) {
iounmap(l->regs); iounmap(l->regs);
l->regs = NULL; l->regs = NULL;
@@ -661,6 +661,11 @@ static int gk20a_init_support(struct platform_device *dev)
int err = 0; int err = 0;
struct gk20a *g = get_gk20a(&dev->dev); struct gk20a *g = get_gk20a(&dev->dev);
struct nvgpu_os_linux *l = nvgpu_os_linux_from_gk20a(g); struct nvgpu_os_linux *l = nvgpu_os_linux_from_gk20a(g);
struct sim_gk20a_linux *sim_linux = nvgpu_kzalloc(g, sizeof(*sim_linux));
if (!sim_linux)
goto fail;
g->sim = &sim_linux->sim;
tegra_register_idle_unidle(gk20a_do_idle, gk20a_do_unidle, g); tegra_register_idle_unidle(gk20a_do_idle, gk20a_do_unidle, g);
@@ -681,13 +686,13 @@ static int gk20a_init_support(struct platform_device *dev)
} }
if (nvgpu_platform_is_simulation(g)) { if (nvgpu_platform_is_simulation(g)) {
g->sim.g = g; g->sim->g = g;
g->sim.regs = gk20a_ioremap_resource(dev, sim_linux->regs = gk20a_ioremap_resource(dev,
GK20A_SIM_IORESOURCE_MEM, GK20A_SIM_IORESOURCE_MEM,
&g->sim.reg_mem); &sim_linux->reg_mem);
if (IS_ERR(g->sim.regs)) { if (IS_ERR(sim_linux->regs)) {
nvgpu_err(g, "failed to remap gk20a sim regs"); nvgpu_err(g, "failed to remap gk20a sim regs");
err = PTR_ERR(g->sim.regs); err = PTR_ERR(sim_linux->regs);
goto fail; goto fail;
} }
@@ -703,6 +708,8 @@ static int gk20a_init_support(struct platform_device *dev)
return 0; return 0;
fail: fail:
nvgpu_kfree(g, sim_linux);
g->sim = NULL;
return err; return err;
} }

View File

@@ -1,23 +1,17 @@
/* /*
* Copyright (c) 2017, NVIDIA CORPORATION. All rights reserved. * Copyright (c) 2017, NVIDIA CORPORATION. All rights reserved.
* *
* Permission is hereby granted, free of charge, to any person obtaining a * This program is free software; you can redistribute it and/or modify it
* copy of this software and associated documentation files (the "Software"), * under the terms and conditions of the GNU General Public License,
* to deal in the Software without restriction, including without limitation * version 2, as published by the Free Software Foundation.
* the rights to use, copy, modify, merge, publish, distribute, sublicense,
* and/or sell copies of the Software, and to permit persons to whom the
* Software is furnished to do so, subject to the following conditions:
* *
* The above copyright notice and this permission notice shall be included in * This program is distributed in the hope it will be useful, but WITHOUT
* all copies or substantial portions of the Software. * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
* FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for
* more details.
* *
* THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR * You should have received a copy of the GNU General Public License
* IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, * along with this program. If not, see <http://www.gnu.org/licenses/>.
* FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
* THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
* LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING
* FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER
* DEALINGS IN THE SOFTWARE.
*/ */
#include <linux/io.h> #include <linux/io.h>
@@ -27,18 +21,19 @@
#include <nvgpu/log.h> #include <nvgpu/log.h>
#include <nvgpu/linux/vm.h> #include <nvgpu/linux/vm.h>
#include "gk20a.h" #include "gk20a/gk20a.h"
#include "sim_gk20a.h"
#include <nvgpu/hw/gk20a/hw_sim_gk20a.h> #include <nvgpu/hw/gk20a/hw_sim_gk20a.h>
static inline void sim_writel(struct gk20a *g, u32 r, u32 v) static inline void sim_writel(struct sim_gk20a_linux *sim_linux, u32 r, u32 v)
{ {
writel(v, g->sim.regs + r); writel(v, sim_linux->regs + r);
} }
static inline u32 sim_readl(struct gk20a *g, u32 r) static inline u32 sim_readl(struct sim_gk20a_linux *sim_linux, u32 r)
{ {
return readl(g->sim.regs + r); return readl(sim_linux->regs + r);
} }
static void kunmap_and_free_iopage(void **kvaddr, struct page **page) static void kunmap_and_free_iopage(void **kvaddr, struct page **page)
@@ -55,23 +50,36 @@ static void kunmap_and_free_iopage(void **kvaddr, struct page **page)
static void gk20a_free_sim_support(struct gk20a *g) static void gk20a_free_sim_support(struct gk20a *g)
{ {
struct sim_gk20a_linux *sim_linux =
container_of(g->sim, struct sim_gk20a_linux, sim);
/* free sim mappings, bfrs */ /* free sim mappings, bfrs */
kunmap_and_free_iopage(&g->sim.send_bfr.kvaddr, kunmap_and_free_iopage(&sim_linux->send_bfr.kvaddr,
&g->sim.send_bfr.page); &sim_linux->send_bfr.page);
kunmap_and_free_iopage(&g->sim.recv_bfr.kvaddr, kunmap_and_free_iopage(&sim_linux->recv_bfr.kvaddr,
&g->sim.recv_bfr.page); &sim_linux->recv_bfr.page);
kunmap_and_free_iopage(&g->sim.msg_bfr.kvaddr, kunmap_and_free_iopage(&sim_linux->msg_bfr.kvaddr,
&g->sim.msg_bfr.page); &sim_linux->msg_bfr.page);
} }
static void gk20a_remove_sim_support(struct sim_gk20a *s) static void gk20a_remove_sim_support(struct sim_gk20a *s)
{ {
struct gk20a *g = s->g; struct gk20a *g = s->g;
if (g->sim.regs) struct sim_gk20a_linux *sim_linux =
sim_writel(g, sim_config_r(), sim_config_mode_disabled_v()); container_of(g->sim, struct sim_gk20a_linux, sim);
if (sim_linux->regs)
sim_writel(sim_linux, sim_config_r(), sim_config_mode_disabled_v());
gk20a_free_sim_support(g); gk20a_free_sim_support(g);
if (sim_linux->regs) {
iounmap(sim_linux->regs);
sim_linux->regs = NULL;
}
nvgpu_kfree(g, sim_linux);
g->sim = NULL;
} }
static int alloc_and_kmap_iopage(struct gk20a *g, static int alloc_and_kmap_iopage(struct gk20a *g,
@@ -103,72 +111,6 @@ static int alloc_and_kmap_iopage(struct gk20a *g,
} }
int gk20a_init_sim_support(struct gk20a *g)
{
int err = 0;
u64 phys;
/* allocate sim event/msg buffers */
err = alloc_and_kmap_iopage(g, &g->sim.send_bfr.kvaddr,
&g->sim.send_bfr.phys,
&g->sim.send_bfr.page);
err = err || alloc_and_kmap_iopage(g, &g->sim.recv_bfr.kvaddr,
&g->sim.recv_bfr.phys,
&g->sim.recv_bfr.page);
err = err || alloc_and_kmap_iopage(g, &g->sim.msg_bfr.kvaddr,
&g->sim.msg_bfr.phys,
&g->sim.msg_bfr.page);
if (!(g->sim.send_bfr.kvaddr && g->sim.recv_bfr.kvaddr &&
g->sim.msg_bfr.kvaddr)) {
nvgpu_err(g, "couldn't allocate all sim buffers");
goto fail;
}
/*mark send ring invalid*/
sim_writel(g, sim_send_ring_r(), sim_send_ring_status_invalid_f());
/*read get pointer and make equal to put*/
g->sim.send_ring_put = sim_readl(g, sim_send_get_r());
sim_writel(g, sim_send_put_r(), g->sim.send_ring_put);
/*write send ring address and make it valid*/
phys = g->sim.send_bfr.phys;
sim_writel(g, sim_send_ring_hi_r(),
sim_send_ring_hi_addr_f(u64_hi32(phys)));
sim_writel(g, sim_send_ring_r(),
sim_send_ring_status_valid_f() |
sim_send_ring_target_phys_pci_coherent_f() |
sim_send_ring_size_4kb_f() |
sim_send_ring_addr_lo_f(phys >> PAGE_SHIFT));
/*repeat for recv ring (but swap put,get as roles are opposite) */
sim_writel(g, sim_recv_ring_r(), sim_recv_ring_status_invalid_f());
/*read put pointer and make equal to get*/
g->sim.recv_ring_get = sim_readl(g, sim_recv_put_r());
sim_writel(g, sim_recv_get_r(), g->sim.recv_ring_get);
/*write send ring address and make it valid*/
phys = g->sim.recv_bfr.phys;
sim_writel(g, sim_recv_ring_hi_r(),
sim_recv_ring_hi_addr_f(u64_hi32(phys)));
sim_writel(g, sim_recv_ring_r(),
sim_recv_ring_status_valid_f() |
sim_recv_ring_target_phys_pci_coherent_f() |
sim_recv_ring_size_4kb_f() |
sim_recv_ring_addr_lo_f(phys >> PAGE_SHIFT));
g->sim.remove_support = gk20a_remove_sim_support;
return 0;
fail:
gk20a_free_sim_support(g);
return err;
}
static inline u32 sim_msg_header_size(void) static inline u32 sim_msg_header_size(void)
{ {
return 24;/*TBD: fix the header to gt this from NV_VGPU_MSG_HEADER*/ return 24;/*TBD: fix the header to gt this from NV_VGPU_MSG_HEADER*/
@@ -176,7 +118,9 @@ static inline u32 sim_msg_header_size(void)
static inline u32 *sim_msg_bfr(struct gk20a *g, u32 byte_offset) static inline u32 *sim_msg_bfr(struct gk20a *g, u32 byte_offset)
{ {
return (u32 *)(g->sim.msg_bfr.kvaddr + byte_offset); struct sim_gk20a_linux *sim_linux =
container_of(g->sim, struct sim_gk20a_linux, sim);
return (u32 *)(sim_linux->msg_bfr.kvaddr + byte_offset);
} }
static inline u32 *sim_msg_hdr(struct gk20a *g, u32 byte_offset) static inline u32 *sim_msg_hdr(struct gk20a *g, u32 byte_offset)
@@ -192,7 +136,7 @@ static inline u32 *sim_msg_param(struct gk20a *g, u32 byte_offset)
static inline void sim_write_hdr(struct gk20a *g, u32 func, u32 size) static inline void sim_write_hdr(struct gk20a *g, u32 func, u32 size)
{ {
/*memset(g->sim.msg_bfr.kvaddr,0,min(PAGE_SIZE,size));*/ /*memset(g->sim->msg_bfr.kvaddr,0,min(PAGE_SIZE,size));*/
*sim_msg_hdr(g, sim_msg_signature_r()) = sim_msg_signature_valid_v(); *sim_msg_hdr(g, sim_msg_signature_r()) = sim_msg_signature_valid_v();
*sim_msg_hdr(g, sim_msg_result_r()) = sim_msg_result_rpc_pending_v(); *sim_msg_hdr(g, sim_msg_result_r()) = sim_msg_result_rpc_pending_v();
*sim_msg_hdr(g, sim_msg_spare_r()) = sim_msg_spare__init_v(); *sim_msg_hdr(g, sim_msg_spare_r()) = sim_msg_spare__init_v();
@@ -207,62 +151,70 @@ static inline u32 sim_escape_read_hdr_size(void)
static u32 *sim_send_ring_bfr(struct gk20a *g, u32 byte_offset) static u32 *sim_send_ring_bfr(struct gk20a *g, u32 byte_offset)
{ {
return (u32 *)(g->sim.send_bfr.kvaddr + byte_offset); struct sim_gk20a_linux *sim_linux =
container_of(g->sim, struct sim_gk20a_linux, sim);
return (u32 *)(sim_linux->send_bfr.kvaddr + byte_offset);
} }
static int rpc_send_message(struct gk20a *g) static int rpc_send_message(struct gk20a *g)
{ {
/* calculations done in units of u32s */ /* calculations done in units of u32s */
u32 send_base = sim_send_put_pointer_v(g->sim.send_ring_put) * 2; u32 send_base = sim_send_put_pointer_v(g->sim->send_ring_put) * 2;
u32 dma_offset = send_base + sim_dma_r()/sizeof(u32); u32 dma_offset = send_base + sim_dma_r()/sizeof(u32);
u32 dma_hi_offset = send_base + sim_dma_hi_r()/sizeof(u32); u32 dma_hi_offset = send_base + sim_dma_hi_r()/sizeof(u32);
struct sim_gk20a_linux *sim_linux =
container_of(g->sim, struct sim_gk20a_linux, sim);
*sim_send_ring_bfr(g, dma_offset*sizeof(u32)) = *sim_send_ring_bfr(g, dma_offset*sizeof(u32)) =
sim_dma_target_phys_pci_coherent_f() | sim_dma_target_phys_pci_coherent_f() |
sim_dma_status_valid_f() | sim_dma_status_valid_f() |
sim_dma_size_4kb_f() | sim_dma_size_4kb_f() |
sim_dma_addr_lo_f(g->sim.msg_bfr.phys >> PAGE_SHIFT); sim_dma_addr_lo_f(sim_linux->msg_bfr.phys >> PAGE_SHIFT);
*sim_send_ring_bfr(g, dma_hi_offset*sizeof(u32)) = *sim_send_ring_bfr(g, dma_hi_offset*sizeof(u32)) =
u64_hi32(g->sim.msg_bfr.phys); u64_hi32(sim_linux->msg_bfr.phys);
*sim_msg_hdr(g, sim_msg_sequence_r()) = g->sim.sequence_base++; *sim_msg_hdr(g, sim_msg_sequence_r()) = g->sim->sequence_base++;
g->sim.send_ring_put = (g->sim.send_ring_put + 2 * sizeof(u32)) % g->sim->send_ring_put = (g->sim->send_ring_put + 2 * sizeof(u32)) %
PAGE_SIZE; PAGE_SIZE;
__cpuc_flush_dcache_area(g->sim.msg_bfr.kvaddr, PAGE_SIZE); __cpuc_flush_dcache_area(sim_linux->msg_bfr.kvaddr, PAGE_SIZE);
__cpuc_flush_dcache_area(g->sim.send_bfr.kvaddr, PAGE_SIZE); __cpuc_flush_dcache_area(sim_linux->send_bfr.kvaddr, PAGE_SIZE);
__cpuc_flush_dcache_area(g->sim.recv_bfr.kvaddr, PAGE_SIZE); __cpuc_flush_dcache_area(sim_linux->recv_bfr.kvaddr, PAGE_SIZE);
/* Update the put pointer. This will trap into the host. */ /* Update the put pointer. This will trap into the host. */
sim_writel(g, sim_send_put_r(), g->sim.send_ring_put); sim_writel(sim_linux, sim_send_put_r(), g->sim->send_ring_put);
return 0; return 0;
} }
static inline u32 *sim_recv_ring_bfr(struct gk20a *g, u32 byte_offset) static inline u32 *sim_recv_ring_bfr(struct gk20a *g, u32 byte_offset)
{ {
return (u32 *)(g->sim.recv_bfr.kvaddr + byte_offset); struct sim_gk20a_linux *sim_linux =
container_of(g->sim, struct sim_gk20a_linux, sim);
return (u32 *)(sim_linux->recv_bfr.kvaddr + byte_offset);
} }
static int rpc_recv_poll(struct gk20a *g) static int rpc_recv_poll(struct gk20a *g)
{ {
u64 recv_phys_addr; u64 recv_phys_addr;
struct sim_gk20a_linux *sim_linux =
container_of(g->sim, struct sim_gk20a_linux, sim);
/* XXX This read is not required (?) */ /* XXX This read is not required (?) */
/*pVGpu->recv_ring_get = VGPU_REG_RD32(pGpu, NV_VGPU_RECV_GET);*/ /*pVGpu->recv_ring_get = VGPU_REG_RD32(pGpu, NV_VGPU_RECV_GET);*/
/* Poll the recv ring get pointer in an infinite loop*/ /* Poll the recv ring get pointer in an infinite loop*/
do { do {
g->sim.recv_ring_put = sim_readl(g, sim_recv_put_r()); g->sim->recv_ring_put = sim_readl(sim_linux, sim_recv_put_r());
} while (g->sim.recv_ring_put == g->sim.recv_ring_get); } while (g->sim->recv_ring_put == g->sim->recv_ring_get);
/* process all replies */ /* process all replies */
while (g->sim.recv_ring_put != g->sim.recv_ring_get) { while (g->sim->recv_ring_put != g->sim->recv_ring_get) {
/* these are in u32 offsets*/ /* these are in u32 offsets*/
u32 dma_lo_offset = u32 dma_lo_offset =
sim_recv_put_pointer_v(g->sim.recv_ring_get)*2 + 0; sim_recv_put_pointer_v(g->sim->recv_ring_get)*2 + 0;
u32 dma_hi_offset = dma_lo_offset + 1; u32 dma_hi_offset = dma_lo_offset + 1;
u32 recv_phys_addr_lo = sim_dma_addr_lo_v( u32 recv_phys_addr_lo = sim_dma_addr_lo_v(
*sim_recv_ring_bfr(g, dma_lo_offset*4)); *sim_recv_ring_bfr(g, dma_lo_offset*4));
@@ -272,23 +224,23 @@ static int rpc_recv_poll(struct gk20a *g)
recv_phys_addr = (u64)recv_phys_addr_hi << 32 | recv_phys_addr = (u64)recv_phys_addr_hi << 32 |
(u64)recv_phys_addr_lo << PAGE_SHIFT; (u64)recv_phys_addr_lo << PAGE_SHIFT;
if (recv_phys_addr != g->sim.msg_bfr.phys) { if (recv_phys_addr != sim_linux->msg_bfr.phys) {
nvgpu_err(g, "%s Error in RPC reply", nvgpu_err(g, "%s Error in RPC reply",
__func__); __func__);
return -1; return -1;
} }
/* Update GET pointer */ /* Update GET pointer */
g->sim.recv_ring_get = (g->sim.recv_ring_get + 2*sizeof(u32)) % g->sim->recv_ring_get = (g->sim->recv_ring_get + 2*sizeof(u32)) %
PAGE_SIZE; PAGE_SIZE;
__cpuc_flush_dcache_area(g->sim.msg_bfr.kvaddr, PAGE_SIZE); __cpuc_flush_dcache_area(sim_linux->msg_bfr.kvaddr, PAGE_SIZE);
__cpuc_flush_dcache_area(g->sim.send_bfr.kvaddr, PAGE_SIZE); __cpuc_flush_dcache_area(sim_linux->send_bfr.kvaddr, PAGE_SIZE);
__cpuc_flush_dcache_area(g->sim.recv_bfr.kvaddr, PAGE_SIZE); __cpuc_flush_dcache_area(sim_linux->recv_bfr.kvaddr, PAGE_SIZE);
sim_writel(g, sim_recv_get_r(), g->sim.recv_ring_get); sim_writel(sim_linux, sim_recv_get_r(), g->sim->recv_ring_get);
g->sim.recv_ring_put = sim_readl(g, sim_recv_put_r()); g->sim->recv_ring_put = sim_readl(sim_linux, sim_recv_put_r());
} }
return 0; return 0;
@@ -321,7 +273,7 @@ static int issue_rpc_and_wait(struct gk20a *g)
return 0; return 0;
} }
int gk20a_sim_esc_readl(struct gk20a *g, char *path, u32 index, u32 *data) static int gk20a_sim_esc_readl(struct gk20a *g, char *path, u32 index, u32 *data)
{ {
int err; int err;
size_t pathlen = strlen(path); size_t pathlen = strlen(path);
@@ -341,3 +293,73 @@ int gk20a_sim_esc_readl(struct gk20a *g, char *path, u32 index, u32 *data)
memcpy(data, sim_msg_param(g, data_offset), sizeof(u32)); memcpy(data, sim_msg_param(g, data_offset), sizeof(u32));
return err; return err;
} }
int gk20a_init_sim_support(struct gk20a *g)
{
int err = 0;
u64 phys;
struct sim_gk20a_linux *sim_linux =
container_of(g->sim, struct sim_gk20a_linux, sim);
/* allocate sim event/msg buffers */
err = alloc_and_kmap_iopage(g, &sim_linux->send_bfr.kvaddr,
&sim_linux->send_bfr.phys,
&sim_linux->send_bfr.page);
err = err || alloc_and_kmap_iopage(g, &sim_linux->recv_bfr.kvaddr,
&sim_linux->recv_bfr.phys,
&sim_linux->recv_bfr.page);
err = err || alloc_and_kmap_iopage(g, &sim_linux->msg_bfr.kvaddr,
&sim_linux->msg_bfr.phys,
&sim_linux->msg_bfr.page);
if (!(sim_linux->send_bfr.kvaddr && sim_linux->recv_bfr.kvaddr &&
sim_linux->msg_bfr.kvaddr)) {
nvgpu_err(g, "couldn't allocate all sim buffers");
goto fail;
}
/*mark send ring invalid*/
sim_writel(sim_linux, sim_send_ring_r(), sim_send_ring_status_invalid_f());
/*read get pointer and make equal to put*/
g->sim->send_ring_put = sim_readl(sim_linux, sim_send_get_r());
sim_writel(sim_linux, sim_send_put_r(), g->sim->send_ring_put);
/*write send ring address and make it valid*/
phys = sim_linux->send_bfr.phys;
sim_writel(sim_linux, sim_send_ring_hi_r(),
sim_send_ring_hi_addr_f(u64_hi32(phys)));
sim_writel(sim_linux, sim_send_ring_r(),
sim_send_ring_status_valid_f() |
sim_send_ring_target_phys_pci_coherent_f() |
sim_send_ring_size_4kb_f() |
sim_send_ring_addr_lo_f(phys >> PAGE_SHIFT));
/*repeat for recv ring (but swap put,get as roles are opposite) */
sim_writel(sim_linux, sim_recv_ring_r(), sim_recv_ring_status_invalid_f());
/*read put pointer and make equal to get*/
g->sim->recv_ring_get = sim_readl(sim_linux, sim_recv_put_r());
sim_writel(sim_linux, sim_recv_get_r(), g->sim->recv_ring_get);
/*write send ring address and make it valid*/
phys = sim_linux->recv_bfr.phys;
sim_writel(sim_linux, sim_recv_ring_hi_r(),
sim_recv_ring_hi_addr_f(u64_hi32(phys)));
sim_writel(sim_linux, sim_recv_ring_r(),
sim_recv_ring_status_valid_f() |
sim_recv_ring_target_phys_pci_coherent_f() |
sim_recv_ring_size_4kb_f() |
sim_recv_ring_addr_lo_f(phys >> PAGE_SHIFT));
g->sim->remove_support = gk20a_remove_sim_support;
g->sim->esc_readl = gk20a_sim_esc_readl;
return 0;
fail:
gk20a_free_sim_support(g);
return err;
}

View File

@@ -0,0 +1,39 @@
/*
* drivers/video/tegra/host/gk20a/sim_gk20a.h
*
* GK20A sim support
*
* Copyright (c) 2017, NVIDIA CORPORATION. All rights reserved.
*
* This program is free software; you can redistribute it and/or modify it
* under the terms and conditions of the GNU General Public License,
* version 2, as published by the Free Software Foundation.
*
* This program is distributed in the hope it will be useful, but WITHOUT
* ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
* FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for
* more details.
*
* You should have received a copy of the GNU General Public License
* along with this program. If not, see <http://www.gnu.org/licenses/>.
*/
#ifndef __SIM_GK20A_LINUX_H__
#define __SIM_GK20A_LINUX_H__
#include "gk20a/sim_gk20a.h"
struct sim_gk20a_linux {
struct sim_gk20a sim;
struct resource *reg_mem;
void __iomem *regs;
struct {
struct page *page;
void *kvaddr;
u64 phys;
} send_bfr, recv_bfr, msg_bfr;
};
int gk20a_init_sim_support(struct gk20a *g);
#endif

View File

@@ -1127,7 +1127,7 @@ struct gk20a {
struct clk_gk20a clk; struct clk_gk20a clk;
struct fifo_gk20a fifo; struct fifo_gk20a fifo;
struct gr_gk20a gr; struct gr_gk20a gr;
struct sim_gk20a sim; struct sim_gk20a *sim;
struct mm_gk20a mm; struct mm_gk20a mm;
struct nvgpu_pmu pmu; struct nvgpu_pmu pmu;
struct acr_desc acr; struct acr_desc acr;

View File

@@ -41,47 +41,52 @@ int gr_gk20a_init_ctx_vars_sim(struct gk20a *g, struct gr_gk20a *gr)
g->gr.ctx_vars.dynamic = true; g->gr.ctx_vars.dynamic = true;
g->gr.netlist = GR_NETLIST_DYNAMIC; g->gr.netlist = GR_NETLIST_DYNAMIC;
if(!g->sim->esc_readl) {
nvgpu_err(g, "Invalid pointer to query function.");
goto fail;
}
/* query sizes and counts */ /* query sizes and counts */
gk20a_sim_esc_readl(g, "GRCTX_UCODE_INST_FECS_COUNT", 0, g->sim->esc_readl(g, "GRCTX_UCODE_INST_FECS_COUNT", 0,
&g->gr.ctx_vars.ucode.fecs.inst.count); &g->gr.ctx_vars.ucode.fecs.inst.count);
gk20a_sim_esc_readl(g, "GRCTX_UCODE_DATA_FECS_COUNT", 0, g->sim->esc_readl(g, "GRCTX_UCODE_DATA_FECS_COUNT", 0,
&g->gr.ctx_vars.ucode.fecs.data.count); &g->gr.ctx_vars.ucode.fecs.data.count);
gk20a_sim_esc_readl(g, "GRCTX_UCODE_INST_GPCCS_COUNT", 0, g->sim->esc_readl(g, "GRCTX_UCODE_INST_GPCCS_COUNT", 0,
&g->gr.ctx_vars.ucode.gpccs.inst.count); &g->gr.ctx_vars.ucode.gpccs.inst.count);
gk20a_sim_esc_readl(g, "GRCTX_UCODE_DATA_GPCCS_COUNT", 0, g->sim->esc_readl(g, "GRCTX_UCODE_DATA_GPCCS_COUNT", 0,
&g->gr.ctx_vars.ucode.gpccs.data.count); &g->gr.ctx_vars.ucode.gpccs.data.count);
gk20a_sim_esc_readl(g, "GRCTX_ALL_CTX_TOTAL_WORDS", 0, &temp); g->sim->esc_readl(g, "GRCTX_ALL_CTX_TOTAL_WORDS", 0, &temp);
g->gr.ctx_vars.buffer_size = temp << 2; g->gr.ctx_vars.buffer_size = temp << 2;
gk20a_sim_esc_readl(g, "GRCTX_SW_BUNDLE_INIT_SIZE", 0, g->sim->esc_readl(g, "GRCTX_SW_BUNDLE_INIT_SIZE", 0,
&g->gr.ctx_vars.sw_bundle_init.count); &g->gr.ctx_vars.sw_bundle_init.count);
gk20a_sim_esc_readl(g, "GRCTX_SW_METHOD_INIT_SIZE", 0, g->sim->esc_readl(g, "GRCTX_SW_METHOD_INIT_SIZE", 0,
&g->gr.ctx_vars.sw_method_init.count); &g->gr.ctx_vars.sw_method_init.count);
gk20a_sim_esc_readl(g, "GRCTX_SW_CTX_LOAD_SIZE", 0, g->sim->esc_readl(g, "GRCTX_SW_CTX_LOAD_SIZE", 0,
&g->gr.ctx_vars.sw_ctx_load.count); &g->gr.ctx_vars.sw_ctx_load.count);
gk20a_sim_esc_readl(g, "GRCTX_SW_VEID_BUNDLE_INIT_SIZE", 0, g->sim->esc_readl(g, "GRCTX_SW_VEID_BUNDLE_INIT_SIZE", 0,
&g->gr.ctx_vars.sw_veid_bundle_init.count); &g->gr.ctx_vars.sw_veid_bundle_init.count);
gk20a_sim_esc_readl(g, "GRCTX_NONCTXSW_REG_SIZE", 0, g->sim->esc_readl(g, "GRCTX_NONCTXSW_REG_SIZE", 0,
&g->gr.ctx_vars.sw_non_ctx_load.count); &g->gr.ctx_vars.sw_non_ctx_load.count);
gk20a_sim_esc_readl(g, "GRCTX_REG_LIST_SYS_COUNT", 0, g->sim->esc_readl(g, "GRCTX_REG_LIST_SYS_COUNT", 0,
&g->gr.ctx_vars.ctxsw_regs.sys.count); &g->gr.ctx_vars.ctxsw_regs.sys.count);
gk20a_sim_esc_readl(g, "GRCTX_REG_LIST_GPC_COUNT", 0, g->sim->esc_readl(g, "GRCTX_REG_LIST_GPC_COUNT", 0,
&g->gr.ctx_vars.ctxsw_regs.gpc.count); &g->gr.ctx_vars.ctxsw_regs.gpc.count);
gk20a_sim_esc_readl(g, "GRCTX_REG_LIST_TPC_COUNT", 0, g->sim->esc_readl(g, "GRCTX_REG_LIST_TPC_COUNT", 0,
&g->gr.ctx_vars.ctxsw_regs.tpc.count); &g->gr.ctx_vars.ctxsw_regs.tpc.count);
gk20a_sim_esc_readl(g, "GRCTX_REG_LIST_ZCULL_GPC_COUNT", 0, g->sim->esc_readl(g, "GRCTX_REG_LIST_ZCULL_GPC_COUNT", 0,
&g->gr.ctx_vars.ctxsw_regs.zcull_gpc.count); &g->gr.ctx_vars.ctxsw_regs.zcull_gpc.count);
gk20a_sim_esc_readl(g, "GRCTX_REG_LIST_PM_SYS_COUNT", 0, g->sim->esc_readl(g, "GRCTX_REG_LIST_PM_SYS_COUNT", 0,
&g->gr.ctx_vars.ctxsw_regs.pm_sys.count); &g->gr.ctx_vars.ctxsw_regs.pm_sys.count);
gk20a_sim_esc_readl(g, "GRCTX_REG_LIST_PM_GPC_COUNT", 0, g->sim->esc_readl(g, "GRCTX_REG_LIST_PM_GPC_COUNT", 0,
&g->gr.ctx_vars.ctxsw_regs.pm_gpc.count); &g->gr.ctx_vars.ctxsw_regs.pm_gpc.count);
gk20a_sim_esc_readl(g, "GRCTX_REG_LIST_PM_TPC_COUNT", 0, g->sim->esc_readl(g, "GRCTX_REG_LIST_PM_TPC_COUNT", 0,
&g->gr.ctx_vars.ctxsw_regs.pm_tpc.count); &g->gr.ctx_vars.ctxsw_regs.pm_tpc.count);
gk20a_sim_esc_readl(g, "GRCTX_REG_LIST_PPC_COUNT", 0, g->sim->esc_readl(g, "GRCTX_REG_LIST_PPC_COUNT", 0,
&g->gr.ctx_vars.ctxsw_regs.ppc.count); &g->gr.ctx_vars.ctxsw_regs.ppc.count);
gk20a_sim_esc_readl(g, "GRCTX_REG_LIST_ETPC_COUNT", 0, g->sim->esc_readl(g, "GRCTX_REG_LIST_ETPC_COUNT", 0,
&g->gr.ctx_vars.ctxsw_regs.etpc.count); &g->gr.ctx_vars.ctxsw_regs.etpc.count);
gk20a_sim_esc_readl(g, "GRCTX_REG_LIST_PPC_COUNT", 0, g->sim->esc_readl(g, "GRCTX_REG_LIST_PPC_COUNT", 0,
&g->gr.ctx_vars.ctxsw_regs.ppc.count); &g->gr.ctx_vars.ctxsw_regs.ppc.count);
err |= !alloc_u32_list_gk20a(g, &g->gr.ctx_vars.ucode.fecs.inst); err |= !alloc_u32_list_gk20a(g, &g->gr.ctx_vars.ucode.fecs.inst);
@@ -107,152 +112,152 @@ int gr_gk20a_init_ctx_vars_sim(struct gk20a *g, struct gr_gk20a *gr)
goto fail; goto fail;
for (i = 0; i < g->gr.ctx_vars.ucode.fecs.inst.count; i++) for (i = 0; i < g->gr.ctx_vars.ucode.fecs.inst.count; i++)
gk20a_sim_esc_readl(g, "GRCTX_UCODE_INST_FECS", g->sim->esc_readl(g, "GRCTX_UCODE_INST_FECS",
i, &g->gr.ctx_vars.ucode.fecs.inst.l[i]); i, &g->gr.ctx_vars.ucode.fecs.inst.l[i]);
for (i = 0; i < g->gr.ctx_vars.ucode.fecs.data.count; i++) for (i = 0; i < g->gr.ctx_vars.ucode.fecs.data.count; i++)
gk20a_sim_esc_readl(g, "GRCTX_UCODE_DATA_FECS", g->sim->esc_readl(g, "GRCTX_UCODE_DATA_FECS",
i, &g->gr.ctx_vars.ucode.fecs.data.l[i]); i, &g->gr.ctx_vars.ucode.fecs.data.l[i]);
for (i = 0; i < g->gr.ctx_vars.ucode.gpccs.inst.count; i++) for (i = 0; i < g->gr.ctx_vars.ucode.gpccs.inst.count; i++)
gk20a_sim_esc_readl(g, "GRCTX_UCODE_INST_GPCCS", g->sim->esc_readl(g, "GRCTX_UCODE_INST_GPCCS",
i, &g->gr.ctx_vars.ucode.gpccs.inst.l[i]); i, &g->gr.ctx_vars.ucode.gpccs.inst.l[i]);
for (i = 0; i < g->gr.ctx_vars.ucode.gpccs.data.count; i++) for (i = 0; i < g->gr.ctx_vars.ucode.gpccs.data.count; i++)
gk20a_sim_esc_readl(g, "GRCTX_UCODE_DATA_GPCCS", g->sim->esc_readl(g, "GRCTX_UCODE_DATA_GPCCS",
i, &g->gr.ctx_vars.ucode.gpccs.data.l[i]); i, &g->gr.ctx_vars.ucode.gpccs.data.l[i]);
for (i = 0; i < g->gr.ctx_vars.sw_bundle_init.count; i++) { for (i = 0; i < g->gr.ctx_vars.sw_bundle_init.count; i++) {
struct av_gk20a *l = g->gr.ctx_vars.sw_bundle_init.l; struct av_gk20a *l = g->gr.ctx_vars.sw_bundle_init.l;
gk20a_sim_esc_readl(g, "GRCTX_SW_BUNDLE_INIT:ADDR", g->sim->esc_readl(g, "GRCTX_SW_BUNDLE_INIT:ADDR",
i, &l[i].addr); i, &l[i].addr);
gk20a_sim_esc_readl(g, "GRCTX_SW_BUNDLE_INIT:VALUE", g->sim->esc_readl(g, "GRCTX_SW_BUNDLE_INIT:VALUE",
i, &l[i].value); i, &l[i].value);
} }
for (i = 0; i < g->gr.ctx_vars.sw_method_init.count; i++) { for (i = 0; i < g->gr.ctx_vars.sw_method_init.count; i++) {
struct av_gk20a *l = g->gr.ctx_vars.sw_method_init.l; struct av_gk20a *l = g->gr.ctx_vars.sw_method_init.l;
gk20a_sim_esc_readl(g, "GRCTX_SW_METHOD_INIT:ADDR", g->sim->esc_readl(g, "GRCTX_SW_METHOD_INIT:ADDR",
i, &l[i].addr); i, &l[i].addr);
gk20a_sim_esc_readl(g, "GRCTX_SW_METHOD_INIT:VALUE", g->sim->esc_readl(g, "GRCTX_SW_METHOD_INIT:VALUE",
i, &l[i].value); i, &l[i].value);
} }
for (i = 0; i < g->gr.ctx_vars.sw_ctx_load.count; i++) { for (i = 0; i < g->gr.ctx_vars.sw_ctx_load.count; i++) {
struct aiv_gk20a *l = g->gr.ctx_vars.sw_ctx_load.l; struct aiv_gk20a *l = g->gr.ctx_vars.sw_ctx_load.l;
gk20a_sim_esc_readl(g, "GRCTX_SW_CTX_LOAD:ADDR", g->sim->esc_readl(g, "GRCTX_SW_CTX_LOAD:ADDR",
i, &l[i].addr); i, &l[i].addr);
gk20a_sim_esc_readl(g, "GRCTX_SW_CTX_LOAD:INDEX", g->sim->esc_readl(g, "GRCTX_SW_CTX_LOAD:INDEX",
i, &l[i].index); i, &l[i].index);
gk20a_sim_esc_readl(g, "GRCTX_SW_CTX_LOAD:VALUE", g->sim->esc_readl(g, "GRCTX_SW_CTX_LOAD:VALUE",
i, &l[i].value); i, &l[i].value);
} }
for (i = 0; i < g->gr.ctx_vars.sw_non_ctx_load.count; i++) { for (i = 0; i < g->gr.ctx_vars.sw_non_ctx_load.count; i++) {
struct av_gk20a *l = g->gr.ctx_vars.sw_non_ctx_load.l; struct av_gk20a *l = g->gr.ctx_vars.sw_non_ctx_load.l;
gk20a_sim_esc_readl(g, "GRCTX_NONCTXSW_REG:REG", g->sim->esc_readl(g, "GRCTX_NONCTXSW_REG:REG",
i, &l[i].addr); i, &l[i].addr);
gk20a_sim_esc_readl(g, "GRCTX_NONCTXSW_REG:VALUE", g->sim->esc_readl(g, "GRCTX_NONCTXSW_REG:VALUE",
i, &l[i].value); i, &l[i].value);
} }
for (i = 0; i < g->gr.ctx_vars.sw_veid_bundle_init.count; i++) { for (i = 0; i < g->gr.ctx_vars.sw_veid_bundle_init.count; i++) {
struct av_gk20a *l = g->gr.ctx_vars.sw_veid_bundle_init.l; struct av_gk20a *l = g->gr.ctx_vars.sw_veid_bundle_init.l;
gk20a_sim_esc_readl(g, "GRCTX_SW_VEID_BUNDLE_INIT:ADDR", g->sim->esc_readl(g, "GRCTX_SW_VEID_BUNDLE_INIT:ADDR",
i, &l[i].addr); i, &l[i].addr);
gk20a_sim_esc_readl(g, "GRCTX_SW_VEID_BUNDLE_INIT:VALUE", g->sim->esc_readl(g, "GRCTX_SW_VEID_BUNDLE_INIT:VALUE",
i, &l[i].value); i, &l[i].value);
} }
for (i = 0; i < g->gr.ctx_vars.ctxsw_regs.sys.count; i++) { for (i = 0; i < g->gr.ctx_vars.ctxsw_regs.sys.count; i++) {
struct aiv_gk20a *l = g->gr.ctx_vars.ctxsw_regs.sys.l; struct aiv_gk20a *l = g->gr.ctx_vars.ctxsw_regs.sys.l;
gk20a_sim_esc_readl(g, "GRCTX_REG_LIST_SYS:ADDR", g->sim->esc_readl(g, "GRCTX_REG_LIST_SYS:ADDR",
i, &l[i].addr); i, &l[i].addr);
gk20a_sim_esc_readl(g, "GRCTX_REG_LIST_SYS:INDEX", g->sim->esc_readl(g, "GRCTX_REG_LIST_SYS:INDEX",
i, &l[i].index); i, &l[i].index);
gk20a_sim_esc_readl(g, "GRCTX_REG_LIST_SYS:VALUE", g->sim->esc_readl(g, "GRCTX_REG_LIST_SYS:VALUE",
i, &l[i].value); i, &l[i].value);
} }
for (i = 0; i < g->gr.ctx_vars.ctxsw_regs.gpc.count; i++) { for (i = 0; i < g->gr.ctx_vars.ctxsw_regs.gpc.count; i++) {
struct aiv_gk20a *l = g->gr.ctx_vars.ctxsw_regs.gpc.l; struct aiv_gk20a *l = g->gr.ctx_vars.ctxsw_regs.gpc.l;
gk20a_sim_esc_readl(g, "GRCTX_REG_LIST_GPC:ADDR", g->sim->esc_readl(g, "GRCTX_REG_LIST_GPC:ADDR",
i, &l[i].addr); i, &l[i].addr);
gk20a_sim_esc_readl(g, "GRCTX_REG_LIST_GPC:INDEX", g->sim->esc_readl(g, "GRCTX_REG_LIST_GPC:INDEX",
i, &l[i].index); i, &l[i].index);
gk20a_sim_esc_readl(g, "GRCTX_REG_LIST_GPC:VALUE", g->sim->esc_readl(g, "GRCTX_REG_LIST_GPC:VALUE",
i, &l[i].value); i, &l[i].value);
} }
for (i = 0; i < g->gr.ctx_vars.ctxsw_regs.tpc.count; i++) { for (i = 0; i < g->gr.ctx_vars.ctxsw_regs.tpc.count; i++) {
struct aiv_gk20a *l = g->gr.ctx_vars.ctxsw_regs.tpc.l; struct aiv_gk20a *l = g->gr.ctx_vars.ctxsw_regs.tpc.l;
gk20a_sim_esc_readl(g, "GRCTX_REG_LIST_TPC:ADDR", g->sim->esc_readl(g, "GRCTX_REG_LIST_TPC:ADDR",
i, &l[i].addr); i, &l[i].addr);
gk20a_sim_esc_readl(g, "GRCTX_REG_LIST_TPC:INDEX", g->sim->esc_readl(g, "GRCTX_REG_LIST_TPC:INDEX",
i, &l[i].index); i, &l[i].index);
gk20a_sim_esc_readl(g, "GRCTX_REG_LIST_TPC:VALUE", g->sim->esc_readl(g, "GRCTX_REG_LIST_TPC:VALUE",
i, &l[i].value); i, &l[i].value);
} }
for (i = 0; i < g->gr.ctx_vars.ctxsw_regs.ppc.count; i++) { for (i = 0; i < g->gr.ctx_vars.ctxsw_regs.ppc.count; i++) {
struct aiv_gk20a *l = g->gr.ctx_vars.ctxsw_regs.ppc.l; struct aiv_gk20a *l = g->gr.ctx_vars.ctxsw_regs.ppc.l;
gk20a_sim_esc_readl(g, "GRCTX_REG_LIST_PPC:ADDR", g->sim->esc_readl(g, "GRCTX_REG_LIST_PPC:ADDR",
i, &l[i].addr); i, &l[i].addr);
gk20a_sim_esc_readl(g, "GRCTX_REG_LIST_PPC:INDEX", g->sim->esc_readl(g, "GRCTX_REG_LIST_PPC:INDEX",
i, &l[i].index); i, &l[i].index);
gk20a_sim_esc_readl(g, "GRCTX_REG_LIST_PPC:VALUE", g->sim->esc_readl(g, "GRCTX_REG_LIST_PPC:VALUE",
i, &l[i].value); i, &l[i].value);
} }
for (i = 0; i < g->gr.ctx_vars.ctxsw_regs.zcull_gpc.count; i++) { for (i = 0; i < g->gr.ctx_vars.ctxsw_regs.zcull_gpc.count; i++) {
struct aiv_gk20a *l = g->gr.ctx_vars.ctxsw_regs.zcull_gpc.l; struct aiv_gk20a *l = g->gr.ctx_vars.ctxsw_regs.zcull_gpc.l;
gk20a_sim_esc_readl(g, "GRCTX_REG_LIST_ZCULL_GPC:ADDR", g->sim->esc_readl(g, "GRCTX_REG_LIST_ZCULL_GPC:ADDR",
i, &l[i].addr); i, &l[i].addr);
gk20a_sim_esc_readl(g, "GRCTX_REG_LIST_ZCULL_GPC:INDEX", g->sim->esc_readl(g, "GRCTX_REG_LIST_ZCULL_GPC:INDEX",
i, &l[i].index); i, &l[i].index);
gk20a_sim_esc_readl(g, "GRCTX_REG_LIST_ZCULL_GPC:VALUE", g->sim->esc_readl(g, "GRCTX_REG_LIST_ZCULL_GPC:VALUE",
i, &l[i].value); i, &l[i].value);
} }
for (i = 0; i < g->gr.ctx_vars.ctxsw_regs.pm_sys.count; i++) { for (i = 0; i < g->gr.ctx_vars.ctxsw_regs.pm_sys.count; i++) {
struct aiv_gk20a *l = g->gr.ctx_vars.ctxsw_regs.pm_sys.l; struct aiv_gk20a *l = g->gr.ctx_vars.ctxsw_regs.pm_sys.l;
gk20a_sim_esc_readl(g, "GRCTX_REG_LIST_PM_SYS:ADDR", g->sim->esc_readl(g, "GRCTX_REG_LIST_PM_SYS:ADDR",
i, &l[i].addr); i, &l[i].addr);
gk20a_sim_esc_readl(g, "GRCTX_REG_LIST_PM_SYS:INDEX", g->sim->esc_readl(g, "GRCTX_REG_LIST_PM_SYS:INDEX",
i, &l[i].index); i, &l[i].index);
gk20a_sim_esc_readl(g, "GRCTX_REG_LIST_PM_SYS:VALUE", g->sim->esc_readl(g, "GRCTX_REG_LIST_PM_SYS:VALUE",
i, &l[i].value); i, &l[i].value);
} }
for (i = 0; i < g->gr.ctx_vars.ctxsw_regs.pm_gpc.count; i++) { for (i = 0; i < g->gr.ctx_vars.ctxsw_regs.pm_gpc.count; i++) {
struct aiv_gk20a *l = g->gr.ctx_vars.ctxsw_regs.pm_gpc.l; struct aiv_gk20a *l = g->gr.ctx_vars.ctxsw_regs.pm_gpc.l;
gk20a_sim_esc_readl(g, "GRCTX_REG_LIST_PM_GPC:ADDR", g->sim->esc_readl(g, "GRCTX_REG_LIST_PM_GPC:ADDR",
i, &l[i].addr); i, &l[i].addr);
gk20a_sim_esc_readl(g, "GRCTX_REG_LIST_PM_GPC:INDEX", g->sim->esc_readl(g, "GRCTX_REG_LIST_PM_GPC:INDEX",
i, &l[i].index); i, &l[i].index);
gk20a_sim_esc_readl(g, "GRCTX_REG_LIST_PM_GPC:VALUE", g->sim->esc_readl(g, "GRCTX_REG_LIST_PM_GPC:VALUE",
i, &l[i].value); i, &l[i].value);
} }
for (i = 0; i < g->gr.ctx_vars.ctxsw_regs.pm_tpc.count; i++) { for (i = 0; i < g->gr.ctx_vars.ctxsw_regs.pm_tpc.count; i++) {
struct aiv_gk20a *l = g->gr.ctx_vars.ctxsw_regs.pm_tpc.l; struct aiv_gk20a *l = g->gr.ctx_vars.ctxsw_regs.pm_tpc.l;
gk20a_sim_esc_readl(g, "GRCTX_REG_LIST_PM_TPC:ADDR", g->sim->esc_readl(g, "GRCTX_REG_LIST_PM_TPC:ADDR",
i, &l[i].addr); i, &l[i].addr);
gk20a_sim_esc_readl(g, "GRCTX_REG_LIST_PM_TPC:INDEX", g->sim->esc_readl(g, "GRCTX_REG_LIST_PM_TPC:INDEX",
i, &l[i].index); i, &l[i].index);
gk20a_sim_esc_readl(g, "GRCTX_REG_LIST_PM_TPC:VALUE", g->sim->esc_readl(g, "GRCTX_REG_LIST_PM_TPC:VALUE",
i, &l[i].value); i, &l[i].value);
} }
gk20a_dbg(gpu_dbg_info | gpu_dbg_fn, "query GRCTX_REG_LIST_ETPC"); gk20a_dbg(gpu_dbg_info | gpu_dbg_fn, "query GRCTX_REG_LIST_ETPC");
for (i = 0; i < g->gr.ctx_vars.ctxsw_regs.etpc.count; i++) { for (i = 0; i < g->gr.ctx_vars.ctxsw_regs.etpc.count; i++) {
struct aiv_gk20a *l = g->gr.ctx_vars.ctxsw_regs.etpc.l; struct aiv_gk20a *l = g->gr.ctx_vars.ctxsw_regs.etpc.l;
gk20a_sim_esc_readl(g, "GRCTX_REG_LIST_ETPC:ADDR", g->sim->esc_readl(g, "GRCTX_REG_LIST_ETPC:ADDR",
i, &l[i].addr); i, &l[i].addr);
gk20a_sim_esc_readl(g, "GRCTX_REG_LIST_ETPC:INDEX", g->sim->esc_readl(g, "GRCTX_REG_LIST_ETPC:INDEX",
i, &l[i].index); i, &l[i].index);
gk20a_sim_esc_readl(g, "GRCTX_REG_LIST_ETPC:VALUE", g->sim->esc_readl(g, "GRCTX_REG_LIST_ETPC:VALUE",
i, &l[i].value); i, &l[i].value);
gk20a_dbg(gpu_dbg_info | gpu_dbg_fn, gk20a_dbg(gpu_dbg_info | gpu_dbg_fn,
"addr:0x%#08x index:0x%08x value:0x%08x", "addr:0x%#08x index:0x%08x value:0x%08x",
@@ -261,7 +266,7 @@ int gr_gk20a_init_ctx_vars_sim(struct gk20a *g, struct gr_gk20a *gr)
g->gr.ctx_vars.valid = true; g->gr.ctx_vars.valid = true;
gk20a_sim_esc_readl(g, "GRCTX_GEN_CTX_REGS_BASE_INDEX", 0, g->sim->esc_readl(g, "GRCTX_GEN_CTX_REGS_BASE_INDEX", 0,
&g->gr.ctx_vars.regs_base_index); &g->gr.ctx_vars.regs_base_index);
gk20a_dbg(gpu_dbg_info | gpu_dbg_fn, "finished querying grctx info from chiplib"); gk20a_dbg(gpu_dbg_info | gpu_dbg_fn, "finished querying grctx info from chiplib");

View File

@@ -3,7 +3,7 @@
* *
* GK20A sim support * GK20A sim support
* *
* Copyright (c) 2013, NVIDIA CORPORATION. All rights reserved. * Copyright (c) 2013-2017, NVIDIA CORPORATION. All rights reserved.
* *
* Permission is hereby granted, free of charge, to any person obtaining a * Permission is hereby granted, free of charge, to any person obtaining a
* copy of this software and associated documentation files (the "Software"), * copy of this software and associated documentation files (the "Software"),
@@ -27,25 +27,16 @@
#define __SIM_GK20A_H__ #define __SIM_GK20A_H__
struct gk20a; struct gk20a;
struct platform_device;
struct sim_gk20a { struct sim_gk20a {
struct gk20a *g; struct gk20a *g;
struct resource *reg_mem;
void __iomem *regs;
struct {
struct page *page;
void *kvaddr;
u64 phys;
} send_bfr, recv_bfr, msg_bfr;
u32 send_ring_put; u32 send_ring_put;
u32 recv_ring_get; u32 recv_ring_get;
u32 recv_ring_put; u32 recv_ring_put;
u32 sequence_base; u32 sequence_base;
void (*remove_support)(struct sim_gk20a *); void (*remove_support)(struct sim_gk20a *);
int (*esc_readl)(
struct gk20a *g, char *path, u32 index, u32 *data);
}; };
int gk20a_init_sim_support(struct gk20a *g);
int gk20a_sim_esc_readl(struct gk20a *g, char *path, u32 index, u32 *data);
#endif /*__SIM_GK20A_H__*/ #endif /*__SIM_GK20A_H__*/