mirror of
git://nv-tegra.nvidia.com/linux-nvgpu.git
synced 2025-12-24 02:22:34 +03:00
gpu: nvgpu: changes related to preemption
Added function pointers to check chip specific valid gfx class and compute class. Also added function pointer to update ctx header with preemption buffer pointers. Bug 200292090 Change-Id: I8119ee082e2abb67186a8ac07088f8db7f410ba1 Signed-off-by: seshendra Gadagottu <sgadagottu@nvidia.com> Reviewed-on: http://git-master/r/1293502 Reviewed-by: mobile promotions <svcmobile_promotions@nvidia.com> Tested-by: mobile promotions <svcmobile_promotions@nvidia.com>
This commit is contained in:
committed by
mobile promotions
parent
8b981f3c64
commit
2baab4e49e
@@ -194,6 +194,8 @@ struct gpu_ops {
|
||||
void (*set_circular_buffer_size)(struct gk20a *g, u32 data);
|
||||
void (*enable_hww_exceptions)(struct gk20a *g);
|
||||
bool (*is_valid_class)(struct gk20a *g, u32 class_num);
|
||||
bool (*is_valid_gfx_class)(struct gk20a *g, u32 class_num);
|
||||
bool (*is_valid_compute_class)(struct gk20a *g, u32 class_num);
|
||||
void (*get_sm_dsm_perf_regs)(struct gk20a *g,
|
||||
u32 *num_sm_dsm_perf_regs,
|
||||
u32 **sm_dsm_perf_regs,
|
||||
@@ -354,6 +356,8 @@ struct gpu_ops {
|
||||
struct nvgpu_mem *mem, u64 gpu_va);
|
||||
void (*write_pm_ptr)(struct gk20a *g,
|
||||
struct nvgpu_mem *mem, u64 gpu_va);
|
||||
void (*write_preemption_ptr)(struct gk20a *g,
|
||||
struct nvgpu_mem *mem, u64 gpu_va);
|
||||
void (*init_elcg_mode)(struct gk20a *g, u32 mode, u32 engine);
|
||||
void (*load_tpc_mask)(struct gk20a *g);
|
||||
int (*inval_icache)(struct gk20a *g, struct channel_gk20a *ch);
|
||||
|
||||
@@ -3113,6 +3113,22 @@ static bool gr_gk20a_is_valid_class(struct gk20a *g, u32 class_num)
|
||||
return valid;
|
||||
}
|
||||
|
||||
static bool gr_gk20a_is_valid_gfx_class(struct gk20a *g, u32 class_num)
|
||||
{
|
||||
if (class_num == KEPLER_C)
|
||||
return true;
|
||||
else
|
||||
return false;
|
||||
}
|
||||
|
||||
static bool gr_gk20a_is_valid_compute_class(struct gk20a *g, u32 class_num)
|
||||
{
|
||||
if (class_num == KEPLER_COMPUTE_A)
|
||||
return true;
|
||||
else
|
||||
return false;
|
||||
}
|
||||
|
||||
int gk20a_alloc_obj_ctx(struct channel_gk20a *c,
|
||||
struct nvgpu_alloc_obj_ctx_args *args)
|
||||
{
|
||||
@@ -9462,6 +9478,8 @@ void gk20a_init_gr_ops(struct gpu_ops *gops)
|
||||
gk20a_gr_set_alpha_circular_buffer_size;
|
||||
gops->gr.enable_hww_exceptions = gr_gk20a_enable_hww_exceptions;
|
||||
gops->gr.is_valid_class = gr_gk20a_is_valid_class;
|
||||
gops->gr.is_valid_gfx_class = gr_gk20a_is_valid_gfx_class;
|
||||
gops->gr.is_valid_compute_class = gr_gk20a_is_valid_compute_class;
|
||||
gops->gr.get_sm_dsm_perf_regs = gr_gk20a_get_sm_dsm_perf_regs;
|
||||
gops->gr.get_sm_dsm_perf_ctrl_regs = gr_gk20a_get_sm_dsm_perf_ctrl_regs;
|
||||
gops->gr.init_fs_state = gr_gk20a_init_fs_state;
|
||||
|
||||
@@ -474,6 +474,23 @@ static bool gr_gm20b_is_valid_class(struct gk20a *g, u32 class_num)
|
||||
return valid;
|
||||
}
|
||||
|
||||
static bool gr_gm20b_is_valid_gfx_class(struct gk20a *g, u32 class_num)
|
||||
{
|
||||
if (class_num == MAXWELL_B)
|
||||
return true;
|
||||
else
|
||||
return false;
|
||||
}
|
||||
|
||||
static bool gr_gm20b_is_valid_compute_class(struct gk20a *g, u32 class_num)
|
||||
{
|
||||
if (class_num == MAXWELL_COMPUTE_B)
|
||||
return true;
|
||||
else
|
||||
return false;
|
||||
}
|
||||
|
||||
|
||||
/* Following are the blocks of registers that the ucode
|
||||
stores in the extended region.*/
|
||||
/* == ctxsw_extended_sm_dsm_perf_counter_register_stride_v() ? */
|
||||
@@ -1527,6 +1544,8 @@ void gm20b_init_gr(struct gpu_ops *gops)
|
||||
gops->gr.set_circular_buffer_size = gr_gm20b_set_circular_buffer_size;
|
||||
gops->gr.enable_hww_exceptions = gr_gk20a_enable_hww_exceptions;
|
||||
gops->gr.is_valid_class = gr_gm20b_is_valid_class;
|
||||
gops->gr.is_valid_gfx_class = gr_gm20b_is_valid_gfx_class;
|
||||
gops->gr.is_valid_compute_class = gr_gm20b_is_valid_compute_class;
|
||||
gops->gr.get_sm_dsm_perf_regs = gr_gm20b_get_sm_dsm_perf_regs;
|
||||
gops->gr.get_sm_dsm_perf_ctrl_regs = gr_gm20b_get_sm_dsm_perf_ctrl_regs;
|
||||
gops->gr.init_fs_state = gr_gm20b_init_fs_state;
|
||||
|
||||
@@ -65,6 +65,23 @@ static bool gr_gp10b_is_valid_class(struct gk20a *g, u32 class_num)
|
||||
return valid;
|
||||
}
|
||||
|
||||
static bool gr_gp10b_is_valid_gfx_class(struct gk20a *g, u32 class_num)
|
||||
{
|
||||
if (class_num == PASCAL_A || class_num == MAXWELL_B)
|
||||
return true;
|
||||
else
|
||||
return false;
|
||||
}
|
||||
|
||||
static bool gr_gp10b_is_valid_compute_class(struct gk20a *g, u32 class_num)
|
||||
{
|
||||
if (class_num == PASCAL_COMPUTE_A || class_num == MAXWELL_COMPUTE_B)
|
||||
return true;
|
||||
else
|
||||
return false;
|
||||
}
|
||||
|
||||
|
||||
static void gr_gp10b_sm_lrf_ecc_overcount_war(int single_err,
|
||||
u32 sed_status,
|
||||
u32 ded_status,
|
||||
@@ -869,10 +886,11 @@ static int gr_gp10b_set_ctxsw_preemption_mode(struct gk20a *g,
|
||||
{
|
||||
int err = 0;
|
||||
|
||||
if (class == PASCAL_A && g->gr.t18x.ctx_vars.force_preemption_gfxp)
|
||||
if (g->ops.gr.is_valid_gfx_class(g, class) &&
|
||||
g->gr.t18x.ctx_vars.force_preemption_gfxp)
|
||||
graphics_preempt_mode = NVGPU_GRAPHICS_PREEMPTION_MODE_GFXP;
|
||||
|
||||
if (class == PASCAL_COMPUTE_A &&
|
||||
if (g->ops.gr.is_valid_compute_class(g, class) &&
|
||||
g->gr.t18x.ctx_vars.force_preemption_cilp)
|
||||
compute_preempt_mode = NVGPU_COMPUTE_PREEMPTION_MODE_CILP;
|
||||
|
||||
@@ -959,7 +977,8 @@ static int gr_gp10b_set_ctxsw_preemption_mode(struct gk20a *g,
|
||||
break;
|
||||
}
|
||||
|
||||
if (class == PASCAL_COMPUTE_A || class == PASCAL_A) {
|
||||
if (g->ops.gr.is_valid_compute_class(g, class) ||
|
||||
g->ops.gr.is_valid_gfx_class(g, class)) {
|
||||
switch (compute_preempt_mode) {
|
||||
case NVGPU_COMPUTE_PREEMPTION_MODE_WFI:
|
||||
case NVGPU_COMPUTE_PREEMPTION_MODE_CTA:
|
||||
@@ -1141,9 +1160,9 @@ static void gr_gp10b_update_ctxsw_preemption_mode(struct gk20a *g,
|
||||
u32 size;
|
||||
u32 cbes_reserve;
|
||||
|
||||
nvgpu_mem_wr(g, mem,
|
||||
ctxsw_prog_main_image_full_preemption_ptr_o(),
|
||||
gr_ctx->t18x.preempt_ctxsw_buffer.gpu_va >> 8);
|
||||
if (g->ops.gr.write_preemption_ptr)
|
||||
g->ops.gr.write_preemption_ptr(g, mem,
|
||||
gr_ctx->t18x.preempt_ctxsw_buffer.gpu_va);
|
||||
|
||||
err = gr_gk20a_ctx_patch_write_begin(g, ch_ctx);
|
||||
if (err) {
|
||||
@@ -2110,6 +2129,8 @@ static int gr_gp10b_set_preemption_mode(struct channel_gk20a *ch,
|
||||
struct tsg_gk20a *tsg;
|
||||
struct vm_gk20a *vm;
|
||||
struct nvgpu_mem *mem = &gr_ctx->mem;
|
||||
struct ctx_header_desc *ctx = &ch->ch_ctx.ctx_header;
|
||||
struct nvgpu_mem *ctxheader = &ctx->mem;
|
||||
u32 class;
|
||||
int err = 0;
|
||||
|
||||
@@ -2156,6 +2177,9 @@ static int gr_gp10b_set_preemption_mode(struct channel_gk20a *ch,
|
||||
if (nvgpu_mem_begin(g, mem))
|
||||
return -ENOMEM;
|
||||
|
||||
if (nvgpu_mem_begin(g, ctxheader))
|
||||
goto unamp_ctx_header;
|
||||
|
||||
err = gk20a_disable_channel_tsg(g, ch);
|
||||
if (err)
|
||||
goto unmap_ctx;
|
||||
@@ -2165,7 +2189,12 @@ static int gr_gp10b_set_preemption_mode(struct channel_gk20a *ch,
|
||||
goto enable_ch;
|
||||
|
||||
if (g->ops.gr.update_ctxsw_preemption_mode) {
|
||||
g->ops.gr.update_ctxsw_preemption_mode(ch->g, ch_ctx, mem);
|
||||
if (ctxheader->gpu_va)
|
||||
g->ops.gr.update_ctxsw_preemption_mode(ch->g,
|
||||
ch_ctx, ctxheader);
|
||||
else
|
||||
g->ops.gr.update_ctxsw_preemption_mode(ch->g,
|
||||
ch_ctx, mem);
|
||||
|
||||
err = gr_gk20a_ctx_patch_write_begin(g, ch_ctx);
|
||||
if (err) {
|
||||
@@ -2179,6 +2208,8 @@ static int gr_gp10b_set_preemption_mode(struct channel_gk20a *ch,
|
||||
enable_ch:
|
||||
gk20a_enable_channel_tsg(g, ch);
|
||||
unmap_ctx:
|
||||
nvgpu_mem_end(g, ctxheader);
|
||||
unamp_ctx_header:
|
||||
nvgpu_mem_end(g, mem);
|
||||
|
||||
return err;
|
||||
@@ -2262,14 +2293,28 @@ static int gr_gp10b_init_preemption_state(struct gk20a *g)
|
||||
return 0;
|
||||
}
|
||||
|
||||
static void gr_gp10b_write_preemption_ptr(struct gk20a *g,
|
||||
struct nvgpu_mem *mem, u64 gpu_va)
|
||||
{
|
||||
u32 va = u64_lo32(gpu_va >> 8);
|
||||
|
||||
nvgpu_mem_wr(g, mem,
|
||||
ctxsw_prog_main_image_full_preemption_ptr_o(), va);
|
||||
|
||||
}
|
||||
|
||||
|
||||
void gp10b_init_gr(struct gpu_ops *gops)
|
||||
{
|
||||
gm20b_init_gr(gops);
|
||||
gops->gr.init_fs_state = gr_gp10b_init_fs_state;
|
||||
gops->gr.init_preemption_state = gr_gp10b_init_preemption_state;
|
||||
gops->gr.is_valid_class = gr_gp10b_is_valid_class;
|
||||
gops->gr.is_valid_gfx_class = gr_gp10b_is_valid_gfx_class;
|
||||
gops->gr.is_valid_compute_class = gr_gp10b_is_valid_compute_class;
|
||||
gops->gr.commit_global_cb_manager = gr_gp10b_commit_global_cb_manager;
|
||||
gops->gr.commit_global_pagepool = gr_gp10b_commit_global_pagepool;
|
||||
gops->gr.write_preemption_ptr = gr_gp10b_write_preemption_ptr;
|
||||
gops->gr.add_zbc_color = gr_gp10b_add_zbc_color;
|
||||
gops->gr.add_zbc_depth = gr_gp10b_add_zbc_depth;
|
||||
gops->gr.pagepool_default_size = gr_gp10b_pagepool_default_size;
|
||||
|
||||
Reference in New Issue
Block a user