mirror of
git://nv-tegra.nvidia.com/linux-nvgpu.git
synced 2025-12-24 02:22:34 +03:00
gpu: nvgpu: Rename gk20a_mem_* functions
Rename the functions used for mem_desc access to nvgpu_mem_*. JIRA NVGPU-12 Change-Id: I5a1180c9a08d33c3dfc361ce8579c3c767fa5656 Signed-off-by: Alex Waterman <alexw@nvidia.com> Reviewed-on: http://git-master/r/1326193 Reviewed-by: mobile promotions <svcmobile_promotions@nvidia.com> Tested-by: mobile promotions <svcmobile_promotions@nvidia.com>
This commit is contained in:
committed by
mobile promotions
parent
2766420dfb
commit
633d331ae2
@@ -117,20 +117,20 @@ static int channel_gv11b_setup_ramfc(struct channel_gk20a *c,
|
|||||||
|
|
||||||
gk20a_dbg_fn("");
|
gk20a_dbg_fn("");
|
||||||
|
|
||||||
gk20a_memset(g, mem, 0, 0, ram_fc_size_val_v());
|
nvgpu_memset(g, mem, 0, 0, ram_fc_size_val_v());
|
||||||
|
|
||||||
gk20a_mem_wr32(g, mem, ram_fc_gp_base_w(),
|
nvgpu_mem_wr32(g, mem, ram_fc_gp_base_w(),
|
||||||
pbdma_gp_base_offset_f(
|
pbdma_gp_base_offset_f(
|
||||||
u64_lo32(gpfifo_base >> pbdma_gp_base_rsvd_s())));
|
u64_lo32(gpfifo_base >> pbdma_gp_base_rsvd_s())));
|
||||||
|
|
||||||
gk20a_mem_wr32(g, mem, ram_fc_gp_base_hi_w(),
|
nvgpu_mem_wr32(g, mem, ram_fc_gp_base_hi_w(),
|
||||||
pbdma_gp_base_hi_offset_f(u64_hi32(gpfifo_base)) |
|
pbdma_gp_base_hi_offset_f(u64_hi32(gpfifo_base)) |
|
||||||
pbdma_gp_base_hi_limit2_f(ilog2(gpfifo_entries)));
|
pbdma_gp_base_hi_limit2_f(ilog2(gpfifo_entries)));
|
||||||
|
|
||||||
gk20a_mem_wr32(g, mem, ram_fc_signature_w(),
|
nvgpu_mem_wr32(g, mem, ram_fc_signature_w(),
|
||||||
c->g->ops.fifo.get_pbdma_signature(c->g));
|
c->g->ops.fifo.get_pbdma_signature(c->g));
|
||||||
|
|
||||||
gk20a_mem_wr32(g, mem, ram_fc_pb_header_w(),
|
nvgpu_mem_wr32(g, mem, ram_fc_pb_header_w(),
|
||||||
pbdma_pb_header_priv_user_f() |
|
pbdma_pb_header_priv_user_f() |
|
||||||
pbdma_pb_header_method_zero_f() |
|
pbdma_pb_header_method_zero_f() |
|
||||||
pbdma_pb_header_subchannel_zero_f() |
|
pbdma_pb_header_subchannel_zero_f() |
|
||||||
@@ -138,44 +138,44 @@ static int channel_gv11b_setup_ramfc(struct channel_gk20a *c,
|
|||||||
pbdma_pb_header_first_true_f() |
|
pbdma_pb_header_first_true_f() |
|
||||||
pbdma_pb_header_type_inc_f());
|
pbdma_pb_header_type_inc_f());
|
||||||
|
|
||||||
gk20a_mem_wr32(g, mem, ram_fc_subdevice_w(),
|
nvgpu_mem_wr32(g, mem, ram_fc_subdevice_w(),
|
||||||
pbdma_subdevice_id_f(PBDMA_SUBDEVICE_ID) |
|
pbdma_subdevice_id_f(PBDMA_SUBDEVICE_ID) |
|
||||||
pbdma_subdevice_status_active_f() |
|
pbdma_subdevice_status_active_f() |
|
||||||
pbdma_subdevice_channel_dma_enable_f());
|
pbdma_subdevice_channel_dma_enable_f());
|
||||||
|
|
||||||
gk20a_mem_wr32(g, mem, ram_fc_target_w(),
|
nvgpu_mem_wr32(g, mem, ram_fc_target_w(),
|
||||||
pbdma_target_eng_ctx_valid_true_f() |
|
pbdma_target_eng_ctx_valid_true_f() |
|
||||||
pbdma_target_ce_ctx_valid_true_f() |
|
pbdma_target_ce_ctx_valid_true_f() |
|
||||||
pbdma_target_engine_sw_f());
|
pbdma_target_engine_sw_f());
|
||||||
|
|
||||||
gk20a_mem_wr32(g, mem, ram_fc_acquire_w(),
|
nvgpu_mem_wr32(g, mem, ram_fc_acquire_w(),
|
||||||
g->ops.fifo.pbdma_acquire_val(acquire_timeout));
|
g->ops.fifo.pbdma_acquire_val(acquire_timeout));
|
||||||
|
|
||||||
gk20a_mem_wr32(g, mem, ram_fc_runlist_timeslice_w(),
|
nvgpu_mem_wr32(g, mem, ram_fc_runlist_timeslice_w(),
|
||||||
pbdma_runlist_timeslice_timeout_128_f() |
|
pbdma_runlist_timeslice_timeout_128_f() |
|
||||||
pbdma_runlist_timeslice_timescale_3_f() |
|
pbdma_runlist_timeslice_timescale_3_f() |
|
||||||
pbdma_runlist_timeslice_enable_true_f());
|
pbdma_runlist_timeslice_enable_true_f());
|
||||||
|
|
||||||
|
|
||||||
gk20a_mem_wr32(g, mem, ram_fc_chid_w(), ram_fc_chid_id_f(c->hw_chid));
|
nvgpu_mem_wr32(g, mem, ram_fc_chid_w(), ram_fc_chid_id_f(c->hw_chid));
|
||||||
|
|
||||||
/* Until full subcontext is supported, always use VEID0 */
|
/* Until full subcontext is supported, always use VEID0 */
|
||||||
gk20a_mem_wr32(g, mem, ram_fc_set_channel_info_w(),
|
nvgpu_mem_wr32(g, mem, ram_fc_set_channel_info_w(),
|
||||||
pbdma_set_channel_info_scg_type_graphics_compute0_f() |
|
pbdma_set_channel_info_scg_type_graphics_compute0_f() |
|
||||||
pbdma_set_channel_info_veid_f(CHANNEL_INFO_VEID0));
|
pbdma_set_channel_info_veid_f(CHANNEL_INFO_VEID0));
|
||||||
|
|
||||||
if (c->is_privileged_channel) {
|
if (c->is_privileged_channel) {
|
||||||
/* Set privilege level for channel */
|
/* Set privilege level for channel */
|
||||||
gk20a_mem_wr32(g, mem, ram_fc_config_w(),
|
nvgpu_mem_wr32(g, mem, ram_fc_config_w(),
|
||||||
pbdma_config_auth_level_privileged_f());
|
pbdma_config_auth_level_privileged_f());
|
||||||
|
|
||||||
gk20a_fifo_setup_ramfc_for_privileged_channel(c);
|
gk20a_fifo_setup_ramfc_for_privileged_channel(c);
|
||||||
}
|
}
|
||||||
|
|
||||||
/* Enable userd writeback */
|
/* Enable userd writeback */
|
||||||
data = gk20a_mem_rd32(g, mem, ram_fc_config_w());
|
data = nvgpu_mem_rd32(g, mem, ram_fc_config_w());
|
||||||
data = data | pbdma_config_userd_writeback_enable_f();
|
data = data | pbdma_config_userd_writeback_enable_f();
|
||||||
gk20a_mem_wr32(g, mem, ram_fc_config_w(),data);
|
nvgpu_mem_wr32(g, mem, ram_fc_config_w(),data);
|
||||||
|
|
||||||
gv11b_userd_writeback_config(g);
|
gv11b_userd_writeback_config(g);
|
||||||
|
|
||||||
@@ -196,7 +196,7 @@ static u32 gv11b_userd_gp_get(struct gk20a *g, struct channel_gk20a *c)
|
|||||||
struct mem_desc *userd_mem = &g->fifo.userd;
|
struct mem_desc *userd_mem = &g->fifo.userd;
|
||||||
u32 offset = c->hw_chid * (g->fifo.userd_entry_size / sizeof(u32));
|
u32 offset = c->hw_chid * (g->fifo.userd_entry_size / sizeof(u32));
|
||||||
|
|
||||||
return gk20a_mem_rd32(g, userd_mem,
|
return nvgpu_mem_rd32(g, userd_mem,
|
||||||
offset + ram_userd_gp_get_w());
|
offset + ram_userd_gp_get_w());
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -205,7 +205,7 @@ static void gv11b_userd_gp_put(struct gk20a *g, struct channel_gk20a *c)
|
|||||||
struct mem_desc *userd_mem = &g->fifo.userd;
|
struct mem_desc *userd_mem = &g->fifo.userd;
|
||||||
u32 offset = c->hw_chid * (g->fifo.userd_entry_size / sizeof(u32));
|
u32 offset = c->hw_chid * (g->fifo.userd_entry_size / sizeof(u32));
|
||||||
|
|
||||||
gk20a_mem_wr32(g, userd_mem, offset + ram_userd_gp_put_w(),
|
nvgpu_mem_wr32(g, userd_mem, offset + ram_userd_gp_put_w(),
|
||||||
c->gpfifo.put);
|
c->gpfifo.put);
|
||||||
/* commit everything to cpu */
|
/* commit everything to cpu */
|
||||||
smp_mb();
|
smp_mb();
|
||||||
|
|||||||
@@ -780,36 +780,36 @@ static void dump_ctx_switch_stats(struct gk20a *g, struct vm_gk20a *vm,
|
|||||||
{
|
{
|
||||||
struct mem_desc *mem = &gr_ctx->mem;
|
struct mem_desc *mem = &gr_ctx->mem;
|
||||||
|
|
||||||
if (gk20a_mem_begin(g, mem)) {
|
if (nvgpu_mem_begin(g, mem)) {
|
||||||
WARN_ON("Cannot map context");
|
WARN_ON("Cannot map context");
|
||||||
return;
|
return;
|
||||||
}
|
}
|
||||||
gk20a_err(dev_from_gk20a(g), "ctxsw_prog_main_image_magic_value_o : %x (expect %x)\n",
|
gk20a_err(dev_from_gk20a(g), "ctxsw_prog_main_image_magic_value_o : %x (expect %x)\n",
|
||||||
gk20a_mem_rd(g, mem,
|
nvgpu_mem_rd(g, mem,
|
||||||
ctxsw_prog_main_image_magic_value_o()),
|
ctxsw_prog_main_image_magic_value_o()),
|
||||||
ctxsw_prog_main_image_magic_value_v_value_v());
|
ctxsw_prog_main_image_magic_value_v_value_v());
|
||||||
|
|
||||||
|
|
||||||
gk20a_err(dev_from_gk20a(g), "NUM_SAVE_OPERATIONS : %d\n",
|
gk20a_err(dev_from_gk20a(g), "NUM_SAVE_OPERATIONS : %d\n",
|
||||||
gk20a_mem_rd(g, mem,
|
nvgpu_mem_rd(g, mem,
|
||||||
ctxsw_prog_main_image_num_save_ops_o()));
|
ctxsw_prog_main_image_num_save_ops_o()));
|
||||||
gk20a_err(dev_from_gk20a(g), "WFI_SAVE_OPERATIONS : %d\n",
|
gk20a_err(dev_from_gk20a(g), "WFI_SAVE_OPERATIONS : %d\n",
|
||||||
gk20a_mem_rd(g, mem,
|
nvgpu_mem_rd(g, mem,
|
||||||
ctxsw_prog_main_image_num_wfi_save_ops_o()));
|
ctxsw_prog_main_image_num_wfi_save_ops_o()));
|
||||||
gk20a_err(dev_from_gk20a(g), "CTA_SAVE_OPERATIONS : %d\n",
|
gk20a_err(dev_from_gk20a(g), "CTA_SAVE_OPERATIONS : %d\n",
|
||||||
gk20a_mem_rd(g, mem,
|
nvgpu_mem_rd(g, mem,
|
||||||
ctxsw_prog_main_image_num_cta_save_ops_o()));
|
ctxsw_prog_main_image_num_cta_save_ops_o()));
|
||||||
gk20a_err(dev_from_gk20a(g), "GFXP_SAVE_OPERATIONS : %d\n",
|
gk20a_err(dev_from_gk20a(g), "GFXP_SAVE_OPERATIONS : %d\n",
|
||||||
gk20a_mem_rd(g, mem,
|
nvgpu_mem_rd(g, mem,
|
||||||
ctxsw_prog_main_image_num_gfxp_save_ops_o()));
|
ctxsw_prog_main_image_num_gfxp_save_ops_o()));
|
||||||
gk20a_err(dev_from_gk20a(g), "CILP_SAVE_OPERATIONS : %d\n",
|
gk20a_err(dev_from_gk20a(g), "CILP_SAVE_OPERATIONS : %d\n",
|
||||||
gk20a_mem_rd(g, mem,
|
nvgpu_mem_rd(g, mem,
|
||||||
ctxsw_prog_main_image_num_cilp_save_ops_o()));
|
ctxsw_prog_main_image_num_cilp_save_ops_o()));
|
||||||
gk20a_err(dev_from_gk20a(g),
|
gk20a_err(dev_from_gk20a(g),
|
||||||
"image gfx preemption option (GFXP is 1) %x\n",
|
"image gfx preemption option (GFXP is 1) %x\n",
|
||||||
gk20a_mem_rd(g, mem,
|
nvgpu_mem_rd(g, mem,
|
||||||
ctxsw_prog_main_image_graphics_preemption_options_o()));
|
ctxsw_prog_main_image_graphics_preemption_options_o()));
|
||||||
gk20a_mem_end(g, mem);
|
nvgpu_mem_end(g, mem);
|
||||||
}
|
}
|
||||||
|
|
||||||
static void gr_gv11b_free_gr_ctx(struct gk20a *g, struct vm_gk20a *vm,
|
static void gr_gv11b_free_gr_ctx(struct gk20a *g, struct vm_gk20a *vm,
|
||||||
@@ -847,13 +847,13 @@ static void gr_gv11b_update_ctxsw_preemption_mode(struct gk20a *g,
|
|||||||
|
|
||||||
if (gr_ctx->graphics_preempt_mode == NVGPU_GRAPHICS_PREEMPTION_MODE_GFXP) {
|
if (gr_ctx->graphics_preempt_mode == NVGPU_GRAPHICS_PREEMPTION_MODE_GFXP) {
|
||||||
gk20a_dbg_info("GfxP: %x", gfxp_preempt_option);
|
gk20a_dbg_info("GfxP: %x", gfxp_preempt_option);
|
||||||
gk20a_mem_wr(g, mem, ctxsw_prog_main_image_graphics_preemption_options_o(),
|
nvgpu_mem_wr(g, mem, ctxsw_prog_main_image_graphics_preemption_options_o(),
|
||||||
gfxp_preempt_option);
|
gfxp_preempt_option);
|
||||||
}
|
}
|
||||||
|
|
||||||
if (gr_ctx->compute_preempt_mode == NVGPU_COMPUTE_PREEMPTION_MODE_CILP) {
|
if (gr_ctx->compute_preempt_mode == NVGPU_COMPUTE_PREEMPTION_MODE_CILP) {
|
||||||
gk20a_dbg_info("CILP: %x", cilp_preempt_option);
|
gk20a_dbg_info("CILP: %x", cilp_preempt_option);
|
||||||
gk20a_mem_wr(g, mem, ctxsw_prog_main_image_compute_preemption_options_o(),
|
nvgpu_mem_wr(g, mem, ctxsw_prog_main_image_compute_preemption_options_o(),
|
||||||
cilp_preempt_option);
|
cilp_preempt_option);
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -862,7 +862,7 @@ static void gr_gv11b_update_ctxsw_preemption_mode(struct gk20a *g,
|
|||||||
u32 size;
|
u32 size;
|
||||||
u32 cbes_reserve;
|
u32 cbes_reserve;
|
||||||
|
|
||||||
gk20a_mem_wr(g, mem, ctxsw_prog_main_image_full_preemption_ptr_o(),
|
nvgpu_mem_wr(g, mem, ctxsw_prog_main_image_full_preemption_ptr_o(),
|
||||||
gr_ctx->t18x.preempt_ctxsw_buffer.gpu_va >> 8);
|
gr_ctx->t18x.preempt_ctxsw_buffer.gpu_va >> 8);
|
||||||
|
|
||||||
err = gr_gk20a_ctx_patch_write_begin(g, ch_ctx);
|
err = gr_gk20a_ctx_patch_write_begin(g, ch_ctx);
|
||||||
@@ -1858,12 +1858,12 @@ static int gr_gv11b_commit_inst(struct channel_gk20a *c, u64 gpu_va)
|
|||||||
addr_hi = u64_hi32(ctx->mem.gpu_va);
|
addr_hi = u64_hi32(ctx->mem.gpu_va);
|
||||||
|
|
||||||
/* point this address to engine_wfi_ptr */
|
/* point this address to engine_wfi_ptr */
|
||||||
gk20a_mem_wr32(c->g, &c->inst_block, ram_in_engine_wfi_target_w(),
|
nvgpu_mem_wr32(c->g, &c->inst_block, ram_in_engine_wfi_target_w(),
|
||||||
ram_in_engine_cs_wfi_v() |
|
ram_in_engine_cs_wfi_v() |
|
||||||
ram_in_engine_wfi_mode_f(ram_in_engine_wfi_mode_virtual_v()) |
|
ram_in_engine_wfi_mode_f(ram_in_engine_wfi_mode_virtual_v()) |
|
||||||
ram_in_engine_wfi_ptr_lo_f(addr_lo));
|
ram_in_engine_wfi_ptr_lo_f(addr_lo));
|
||||||
|
|
||||||
gk20a_mem_wr32(c->g, &c->inst_block, ram_in_engine_wfi_ptr_hi_w(),
|
nvgpu_mem_wr32(c->g, &c->inst_block, ram_in_engine_wfi_ptr_hi_w(),
|
||||||
ram_in_engine_wfi_ptr_hi_f(addr_hi));
|
ram_in_engine_wfi_ptr_hi_f(addr_hi));
|
||||||
|
|
||||||
return 0;
|
return 0;
|
||||||
@@ -1930,20 +1930,20 @@ static void gv11b_restore_context_header(struct gk20a *g,
|
|||||||
u32 va_lo, va_hi;
|
u32 va_lo, va_hi;
|
||||||
struct gr_gk20a *gr = &g->gr;
|
struct gr_gk20a *gr = &g->gr;
|
||||||
|
|
||||||
va_hi = gk20a_mem_rd(g, ctxheader,
|
va_hi = nvgpu_mem_rd(g, ctxheader,
|
||||||
ctxsw_prog_main_image_context_buffer_ptr_hi_o());
|
ctxsw_prog_main_image_context_buffer_ptr_hi_o());
|
||||||
va_lo = gk20a_mem_rd(g, ctxheader,
|
va_lo = nvgpu_mem_rd(g, ctxheader,
|
||||||
ctxsw_prog_main_image_context_buffer_ptr_o());
|
ctxsw_prog_main_image_context_buffer_ptr_o());
|
||||||
gk20a_mem_wr_n(g, ctxheader, 0,
|
nvgpu_mem_wr_n(g, ctxheader, 0,
|
||||||
gr->ctx_vars.local_golden_image,
|
gr->ctx_vars.local_golden_image,
|
||||||
gr->ctx_vars.golden_image_size);
|
gr->ctx_vars.golden_image_size);
|
||||||
gk20a_mem_wr(g, ctxheader,
|
nvgpu_mem_wr(g, ctxheader,
|
||||||
ctxsw_prog_main_image_context_buffer_ptr_hi_o(), va_hi);
|
ctxsw_prog_main_image_context_buffer_ptr_hi_o(), va_hi);
|
||||||
gk20a_mem_wr(g, ctxheader,
|
nvgpu_mem_wr(g, ctxheader,
|
||||||
ctxsw_prog_main_image_context_buffer_ptr_o(), va_lo);
|
ctxsw_prog_main_image_context_buffer_ptr_o(), va_lo);
|
||||||
gk20a_mem_wr(g, ctxheader,
|
nvgpu_mem_wr(g, ctxheader,
|
||||||
ctxsw_prog_main_image_num_restore_ops_o(), 0);
|
ctxsw_prog_main_image_num_restore_ops_o(), 0);
|
||||||
gk20a_mem_wr(g, ctxheader,
|
nvgpu_mem_wr(g, ctxheader,
|
||||||
ctxsw_prog_main_image_num_save_ops_o(), 0);
|
ctxsw_prog_main_image_num_save_ops_o(), 0);
|
||||||
}
|
}
|
||||||
static void gr_gv11b_write_zcull_ptr(struct gk20a *g,
|
static void gr_gv11b_write_zcull_ptr(struct gk20a *g,
|
||||||
@@ -1954,9 +1954,9 @@ static void gr_gv11b_write_zcull_ptr(struct gk20a *g,
|
|||||||
gpu_va = gpu_va >> 8;
|
gpu_va = gpu_va >> 8;
|
||||||
va_lo = u64_lo32(gpu_va);
|
va_lo = u64_lo32(gpu_va);
|
||||||
va_hi = u64_hi32(gpu_va);
|
va_hi = u64_hi32(gpu_va);
|
||||||
gk20a_mem_wr(g, mem,
|
nvgpu_mem_wr(g, mem,
|
||||||
ctxsw_prog_main_image_zcull_ptr_o(), va_lo);
|
ctxsw_prog_main_image_zcull_ptr_o(), va_lo);
|
||||||
gk20a_mem_wr(g, mem,
|
nvgpu_mem_wr(g, mem,
|
||||||
ctxsw_prog_main_image_zcull_ptr_hi_o(), va_hi);
|
ctxsw_prog_main_image_zcull_ptr_hi_o(), va_hi);
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -1969,9 +1969,9 @@ static void gr_gv11b_write_pm_ptr(struct gk20a *g,
|
|||||||
gpu_va = gpu_va >> 8;
|
gpu_va = gpu_va >> 8;
|
||||||
va_lo = u64_lo32(gpu_va);
|
va_lo = u64_lo32(gpu_va);
|
||||||
va_hi = u64_hi32(gpu_va);
|
va_hi = u64_hi32(gpu_va);
|
||||||
gk20a_mem_wr(g, mem,
|
nvgpu_mem_wr(g, mem,
|
||||||
ctxsw_prog_main_image_pm_ptr_o(), va_lo);
|
ctxsw_prog_main_image_pm_ptr_o(), va_lo);
|
||||||
gk20a_mem_wr(g, mem,
|
nvgpu_mem_wr(g, mem,
|
||||||
ctxsw_prog_main_image_pm_ptr_hi_o(), va_hi);
|
ctxsw_prog_main_image_pm_ptr_hi_o(), va_hi);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|||||||
@@ -73,11 +73,11 @@ int gv11b_alloc_subctx_header(struct channel_gk20a *c)
|
|||||||
return -ENOMEM;
|
return -ENOMEM;
|
||||||
}
|
}
|
||||||
/* Now clear the buffer */
|
/* Now clear the buffer */
|
||||||
if (gk20a_mem_begin(g, &ctx->mem))
|
if (nvgpu_mem_begin(g, &ctx->mem))
|
||||||
return -ENOMEM;
|
return -ENOMEM;
|
||||||
|
|
||||||
gk20a_memset(g, &ctx->mem, 0, 0, ctx->mem.size);
|
nvgpu_memset(g, &ctx->mem, 0, 0, ctx->mem.size);
|
||||||
gk20a_mem_end(g, &ctx->mem);
|
nvgpu_mem_end(g, &ctx->mem);
|
||||||
|
|
||||||
gv11b_init_subcontext_pdb(c, &c->inst_block);
|
gv11b_init_subcontext_pdb(c, &c->inst_block);
|
||||||
|
|
||||||
@@ -111,14 +111,14 @@ static void gv11b_init_subcontext_pdb(struct channel_gk20a *c,
|
|||||||
ram_in_sc_page_dir_base_lo_0_f(pdb_addr_lo);
|
ram_in_sc_page_dir_base_lo_0_f(pdb_addr_lo);
|
||||||
lo = ram_in_sc_page_dir_base_vol_0_w();
|
lo = ram_in_sc_page_dir_base_vol_0_w();
|
||||||
hi = ram_in_sc_page_dir_base_hi_0_w();
|
hi = ram_in_sc_page_dir_base_hi_0_w();
|
||||||
gk20a_mem_wr32(g, inst_block, lo, format_word);
|
nvgpu_mem_wr32(g, inst_block, lo, format_word);
|
||||||
gk20a_mem_wr32(g, inst_block, hi, pdb_addr_hi);
|
nvgpu_mem_wr32(g, inst_block, hi, pdb_addr_hi);
|
||||||
|
|
||||||
/* make subcontext0 address space to valid */
|
/* make subcontext0 address space to valid */
|
||||||
/* TODO fix proper hw register definations */
|
/* TODO fix proper hw register definations */
|
||||||
gk20a_mem_wr32(g, inst_block, 166, 0x1);
|
nvgpu_mem_wr32(g, inst_block, 166, 0x1);
|
||||||
gk20a_mem_wr32(g, inst_block, 167, 0);
|
nvgpu_mem_wr32(g, inst_block, 167, 0);
|
||||||
gk20a_mem_wr32(g, inst_block, ram_in_engine_wfi_veid_w(),
|
nvgpu_mem_wr32(g, inst_block, ram_in_engine_wfi_veid_w(),
|
||||||
ram_in_engine_wfi_veid_f(0));
|
ram_in_engine_wfi_veid_f(0));
|
||||||
|
|
||||||
}
|
}
|
||||||
@@ -136,13 +136,13 @@ int gv11b_update_subctx_header(struct channel_gk20a *c, u64 gpu_va)
|
|||||||
|
|
||||||
gr_mem = &ctx->mem;
|
gr_mem = &ctx->mem;
|
||||||
g->ops.mm.l2_flush(g, true);
|
g->ops.mm.l2_flush(g, true);
|
||||||
if (gk20a_mem_begin(g, gr_mem))
|
if (nvgpu_mem_begin(g, gr_mem))
|
||||||
return -ENOMEM;
|
return -ENOMEM;
|
||||||
|
|
||||||
gk20a_mem_wr(g, gr_mem,
|
nvgpu_mem_wr(g, gr_mem,
|
||||||
ctxsw_prog_main_image_context_buffer_ptr_hi_o(), addr_hi);
|
ctxsw_prog_main_image_context_buffer_ptr_hi_o(), addr_hi);
|
||||||
gk20a_mem_wr(g, gr_mem,
|
nvgpu_mem_wr(g, gr_mem,
|
||||||
ctxsw_prog_main_image_context_buffer_ptr_o(), addr_lo);
|
ctxsw_prog_main_image_context_buffer_ptr_o(), addr_lo);
|
||||||
gk20a_mem_end(g, gr_mem);
|
nvgpu_mem_end(g, gr_mem);
|
||||||
return ret;
|
return ret;
|
||||||
}
|
}
|
||||||
|
|||||||
Reference in New Issue
Block a user