mirror of
git://nv-tegra.nvidia.com/linux-nvgpu.git
synced 2025-12-23 01:50:07 +03:00
Only certain combination of channels of GFX/Compute object classes can be assigned to particular pbdma and/or VEID. CILP can be enabled only in certain configs. Implement checks for the configurations verified during alloc_obj_ctx and/or setting preemption mode. Bug 3677982 Change-Id: Ie7026cbb240819c1727b3736ed34044d7138d3cd Signed-off-by: Sagar Kamble <skamble@nvidia.com> Reviewed-on: https://git-master.nvidia.com/r/c/linux-nvgpu/+/2719995 Reviewed-by: Ankur Kishore <ankkishore@nvidia.com> GVS: Gerrit_Virtual_Submit <buildbot_gerritrpt@nvidia.com>
1349 lines
36 KiB
C
1349 lines
36 KiB
C
/*
|
|
* Copyright (c) 2019-2022, NVIDIA CORPORATION. All rights reserved.
|
|
*
|
|
* Permission is hereby granted, free of charge, to any person obtaining a
|
|
* copy of this software and associated documentation files (the "Software"),
|
|
* to deal in the Software without restriction, including without limitation
|
|
* the rights to use, copy, modify, merge, publish, distribute, sublicense,
|
|
* and/or sell copies of the Software, and to permit persons to whom the
|
|
* Software is furnished to do so, subject to the following conditions:
|
|
*
|
|
* The above copyright notice and this permission notice shall be included in
|
|
* all copies or substantial portions of the Software.
|
|
*
|
|
* THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
|
|
* IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
|
|
* FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
|
|
* THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
|
|
* LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING
|
|
* FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER
|
|
* DEALINGS IN THE SOFTWARE.
|
|
*/
|
|
|
|
#include <nvgpu/class.h>
|
|
#include <nvgpu/gk20a.h>
|
|
#include <nvgpu/log.h>
|
|
#include <nvgpu/io.h>
|
|
#include <nvgpu/mm.h>
|
|
#ifdef CONFIG_NVGPU_POWER_PG
|
|
#include <nvgpu/pmu/pmu_pg.h>
|
|
#include <nvgpu/power_features/pg.h>
|
|
#endif
|
|
#include <nvgpu/gr/ctx.h>
|
|
#include <nvgpu/gr/ctx_mappings.h>
|
|
#include <nvgpu/tsg_subctx.h>
|
|
#include <nvgpu/gr/setup.h>
|
|
#include <nvgpu/gr/subctx.h>
|
|
#include <nvgpu/gr/global_ctx.h>
|
|
#include <nvgpu/gr/obj_ctx.h>
|
|
#include <nvgpu/gr/config.h>
|
|
#include <nvgpu/gr/ctx.h>
|
|
#include <nvgpu/netlist.h>
|
|
#include <nvgpu/gr/gr_falcon.h>
|
|
#include <nvgpu/gr/fs_state.h>
|
|
#include <nvgpu/gr/gr_utils.h>
|
|
#include <nvgpu/engines.h>
|
|
#include <nvgpu/nvgpu_init.h>
|
|
#include <nvgpu/power_features/cg.h>
|
|
#include <nvgpu/static_analysis.h>
|
|
#include <nvgpu/tsg.h>
|
|
#include <nvgpu/vm.h>
|
|
|
|
#include "obj_ctx_priv.h"
|
|
|
|
void nvgpu_gr_obj_ctx_commit_inst_gpu_va(struct gk20a *g,
|
|
struct nvgpu_mem *inst_block, u64 gpu_va)
|
|
{
|
|
g->ops.ramin.set_gr_ptr(g, inst_block, gpu_va);
|
|
}
|
|
|
|
#ifdef CONFIG_NVGPU_DEBUGGER
|
|
static void nvgpu_gr_obj_ctx_set_pm_ctx_gpu_va(struct gk20a *g,
|
|
struct nvgpu_gr_ctx *gr_ctx, struct nvgpu_tsg_subctx *tsg_subctx)
|
|
{
|
|
struct nvgpu_gr_subctx *subctx;
|
|
bool set_pm_ctx_gpu_va;
|
|
|
|
nvgpu_log(g, gpu_dbg_fn | gpu_dbg_gr, " ");
|
|
|
|
set_pm_ctx_gpu_va = nvgpu_gr_ctx_get_pm_ctx_pm_mode(gr_ctx) !=
|
|
g->ops.gr.ctxsw_prog.hw_get_pm_mode_no_ctxsw();
|
|
|
|
if (nvgpu_is_enabled(g, NVGPU_SUPPORT_TSG_SUBCONTEXTS)) {
|
|
subctx = nvgpu_tsg_subctx_get_gr_subctx(tsg_subctx);
|
|
nvgpu_gr_subctx_set_hwpm_ptr(g, subctx,
|
|
set_pm_ctx_gpu_va);
|
|
} else {
|
|
nvgpu_gr_ctx_set_hwpm_ptr(g, gr_ctx, set_pm_ctx_gpu_va);
|
|
}
|
|
|
|
nvgpu_log(g, gpu_dbg_fn | gpu_dbg_gr, "done");
|
|
}
|
|
#endif
|
|
|
|
void nvgpu_gr_obj_ctx_commit_inst(struct gk20a *g, struct nvgpu_mem *inst_block,
|
|
struct nvgpu_gr_ctx *gr_ctx, struct nvgpu_tsg_subctx *tsg_subctx,
|
|
struct nvgpu_gr_ctx_mappings *mappings)
|
|
{
|
|
struct nvgpu_gr_subctx *subctx;
|
|
struct nvgpu_mem *ctxheader;
|
|
u64 gpu_va;
|
|
|
|
nvgpu_log(g, gpu_dbg_fn | gpu_dbg_gr, " ");
|
|
|
|
#ifdef CONFIG_NVGPU_DEBUGGER
|
|
nvgpu_gr_obj_ctx_set_pm_ctx_gpu_va(g, gr_ctx, tsg_subctx);
|
|
#endif
|
|
|
|
if (nvgpu_is_enabled(g, NVGPU_SUPPORT_TSG_SUBCONTEXTS)) {
|
|
subctx = nvgpu_tsg_subctx_get_gr_subctx(tsg_subctx);
|
|
nvgpu_gr_subctx_load_ctx_header(g, subctx, gr_ctx, mappings);
|
|
|
|
ctxheader = nvgpu_gr_subctx_get_ctx_header(subctx);
|
|
nvgpu_gr_obj_ctx_commit_inst_gpu_va(g, inst_block,
|
|
ctxheader->gpu_va);
|
|
} else {
|
|
gpu_va = nvgpu_gr_ctx_mappings_get_ctx_va(mappings, NVGPU_GR_CTX_CTX);
|
|
nvgpu_gr_obj_ctx_commit_inst_gpu_va(g, inst_block, gpu_va);
|
|
}
|
|
|
|
nvgpu_log(g, gpu_dbg_fn | gpu_dbg_gr, "done");
|
|
}
|
|
|
|
#if defined(CONFIG_NVGPU_GFXP) || defined(CONFIG_NVGPU_CILP)
|
|
static void nvgpu_gr_obj_ctx_init_ctxsw_preemption_mode(struct gk20a *g,
|
|
struct nvgpu_gr_ctx_desc *gr_ctx_desc,
|
|
u32 class_num, u32 flags,
|
|
u32 *graphics_preempt_mode, u32 *compute_preempt_mode)
|
|
{
|
|
u32 default_graphics_preempt_mode = 0U;
|
|
u32 default_compute_preempt_mode = 0U;
|
|
|
|
g->ops.gr.init.get_default_preemption_modes(
|
|
&default_graphics_preempt_mode,
|
|
&default_compute_preempt_mode);
|
|
|
|
#ifdef CONFIG_NVGPU_GFXP
|
|
if ((flags & NVGPU_OBJ_CTX_FLAGS_SUPPORT_GFXP) != 0U) {
|
|
*graphics_preempt_mode = NVGPU_PREEMPTION_MODE_GRAPHICS_GFXP;
|
|
}
|
|
|
|
if (g->ops.gpu_class.is_valid_gfx(class_num) &&
|
|
nvgpu_gr_ctx_desc_force_preemption_gfxp(gr_ctx_desc)) {
|
|
*graphics_preempt_mode = NVGPU_PREEMPTION_MODE_GRAPHICS_GFXP;
|
|
}
|
|
#endif
|
|
|
|
#ifdef CONFIG_NVGPU_CILP
|
|
if ((flags & NVGPU_OBJ_CTX_FLAGS_SUPPORT_CILP) != 0U) {
|
|
*compute_preempt_mode = NVGPU_PREEMPTION_MODE_COMPUTE_CILP;
|
|
}
|
|
|
|
if (g->ops.gpu_class.is_valid_compute(class_num) &&
|
|
nvgpu_gr_ctx_desc_force_preemption_cilp(gr_ctx_desc)) {
|
|
*compute_preempt_mode = NVGPU_PREEMPTION_MODE_COMPUTE_CILP;
|
|
}
|
|
#endif
|
|
|
|
if (*compute_preempt_mode == 0U) {
|
|
*compute_preempt_mode = default_compute_preempt_mode;
|
|
}
|
|
|
|
if (*graphics_preempt_mode == 0U) {
|
|
*graphics_preempt_mode = default_graphics_preempt_mode;
|
|
}
|
|
}
|
|
|
|
static int nvgpu_gr_obj_ctx_init_ctxsw_preemption(struct gk20a *g,
|
|
struct nvgpu_channel *ch,
|
|
struct nvgpu_gr_config *config, struct nvgpu_gr_ctx_desc *gr_ctx_desc,
|
|
struct nvgpu_gr_ctx *gr_ctx,
|
|
u32 class_num, u32 flags)
|
|
{
|
|
int err;
|
|
u32 graphics_preempt_mode = 0U;
|
|
u32 compute_preempt_mode = 0U;
|
|
|
|
nvgpu_log(g, gpu_dbg_fn | gpu_dbg_gr, " ");
|
|
|
|
/* Skip for engines other than GR */
|
|
if (!g->ops.gpu_class.is_valid_compute(class_num) &&
|
|
!g->ops.gpu_class.is_valid_gfx(class_num)) {
|
|
return 0;
|
|
}
|
|
|
|
if (nvgpu_gr_ctx_get_ctx_initialized(gr_ctx)) {
|
|
#ifdef CONFIG_NVGPU_GFXP
|
|
if ((flags & NVGPU_OBJ_CTX_FLAGS_SUPPORT_GFXP) != 0U) {
|
|
graphics_preempt_mode = NVGPU_PREEMPTION_MODE_GRAPHICS_GFXP;
|
|
}
|
|
#endif
|
|
|
|
#ifdef CONFIG_NVGPU_CILP
|
|
if ((flags & NVGPU_OBJ_CTX_FLAGS_SUPPORT_CILP) != 0U) {
|
|
compute_preempt_mode = NVGPU_PREEMPTION_MODE_COMPUTE_CILP;
|
|
}
|
|
#endif
|
|
if (nvgpu_gr_setup_validate_preemption_mode(&graphics_preempt_mode,
|
|
&compute_preempt_mode, gr_ctx) == false) {
|
|
return 0;
|
|
}
|
|
} else {
|
|
nvgpu_gr_obj_ctx_init_ctxsw_preemption_mode(g, gr_ctx_desc,
|
|
class_num, flags, &graphics_preempt_mode,
|
|
&compute_preempt_mode);
|
|
}
|
|
|
|
err = nvgpu_gr_obj_ctx_set_ctxsw_preemption_mode(g, ch, config,
|
|
gr_ctx_desc, gr_ctx, class_num, graphics_preempt_mode,
|
|
compute_preempt_mode);
|
|
if (err != 0) {
|
|
nvgpu_err(g, "set_ctxsw_preemption_mode failed");
|
|
return err;
|
|
}
|
|
|
|
nvgpu_log(g, gpu_dbg_fn | gpu_dbg_gr, "done");
|
|
|
|
return 0;
|
|
}
|
|
#endif
|
|
|
|
#ifdef CONFIG_NVGPU_GRAPHICS
|
|
static int nvgpu_gr_obj_ctx_set_graphics_preemption_mode(struct gk20a *g,
|
|
struct nvgpu_gr_config *config, struct nvgpu_gr_ctx_desc *gr_ctx_desc,
|
|
struct nvgpu_gr_ctx *gr_ctx,
|
|
u32 graphics_preempt_mode)
|
|
{
|
|
int err = 0;
|
|
|
|
(void)config;
|
|
(void)gr_ctx_desc;
|
|
|
|
/* set preemption modes */
|
|
switch (graphics_preempt_mode) {
|
|
#ifdef CONFIG_NVGPU_GFXP
|
|
case NVGPU_PREEMPTION_MODE_GRAPHICS_GFXP:
|
|
{
|
|
u32 rtv_cb_size;
|
|
u32 spill_size = g->ops.gr.init.get_ctx_spill_size(g);
|
|
u32 pagepool_size = g->ops.gr.init.get_ctx_pagepool_size(g);
|
|
u32 betacb_size = g->ops.gr.init.get_ctx_betacb_size(g);
|
|
u32 attrib_cb_size =
|
|
g->ops.gr.init.get_ctx_attrib_cb_size(g, betacb_size,
|
|
nvgpu_gr_config_get_tpc_count(config),
|
|
nvgpu_gr_config_get_max_tpc_count(config));
|
|
|
|
nvgpu_log_info(g, "gfxp context spill_size=%d", spill_size);
|
|
nvgpu_log_info(g, "gfxp context pagepool_size=%d", pagepool_size);
|
|
nvgpu_log_info(g, "gfxp context attrib_cb_size=%d",
|
|
attrib_cb_size);
|
|
|
|
if ((nvgpu_gr_ctx_get_size(gr_ctx_desc,
|
|
NVGPU_GR_CTX_PREEMPT_CTXSW) == 0U) ||
|
|
(spill_size == 0U) || (attrib_cb_size == 0U) ||
|
|
(pagepool_size == 0U)) {
|
|
return -EINVAL;
|
|
}
|
|
|
|
nvgpu_gr_ctx_set_size(gr_ctx_desc,
|
|
NVGPU_GR_CTX_SPILL_CTXSW, spill_size);
|
|
nvgpu_gr_ctx_set_size(gr_ctx_desc,
|
|
NVGPU_GR_CTX_BETACB_CTXSW, attrib_cb_size);
|
|
nvgpu_gr_ctx_set_size(gr_ctx_desc,
|
|
NVGPU_GR_CTX_PAGEPOOL_CTXSW, pagepool_size);
|
|
|
|
if (g->ops.gr.init.get_gfxp_rtv_cb_size != NULL) {
|
|
rtv_cb_size = g->ops.gr.init.get_gfxp_rtv_cb_size(g);
|
|
nvgpu_gr_ctx_set_size(gr_ctx_desc,
|
|
NVGPU_GR_CTX_GFXP_RTVCB_CTXSW, rtv_cb_size);
|
|
}
|
|
|
|
nvgpu_gr_ctx_init_graphics_preemption_mode(gr_ctx,
|
|
graphics_preempt_mode);
|
|
break;
|
|
}
|
|
#endif
|
|
case NVGPU_PREEMPTION_MODE_GRAPHICS_WFI:
|
|
nvgpu_gr_ctx_init_graphics_preemption_mode(gr_ctx,
|
|
graphics_preempt_mode);
|
|
break;
|
|
|
|
default:
|
|
nvgpu_log_info(g, "graphics_preempt_mode=%u",
|
|
graphics_preempt_mode);
|
|
break;
|
|
}
|
|
|
|
return err;
|
|
}
|
|
#endif
|
|
|
|
static int nvgpu_gr_obj_ctx_set_compute_preemption_mode(struct gk20a *g,
|
|
struct nvgpu_gr_ctx *gr_ctx, u32 class_num, u32 compute_preempt_mode)
|
|
{
|
|
|
|
if (g->ops.gpu_class.is_valid_compute(class_num)
|
|
#ifdef CONFIG_NVGPU_GRAPHICS
|
|
|| g->ops.gpu_class.is_valid_gfx(class_num)
|
|
#endif
|
|
) {
|
|
nvgpu_gr_ctx_init_compute_preemption_mode(gr_ctx,
|
|
compute_preempt_mode);
|
|
return 0;
|
|
} else {
|
|
return -EINVAL;
|
|
}
|
|
|
|
}
|
|
|
|
int nvgpu_gr_obj_ctx_set_ctxsw_preemption_mode(struct gk20a *g,
|
|
struct nvgpu_channel *ch,
|
|
struct nvgpu_gr_config *config, struct nvgpu_gr_ctx_desc *gr_ctx_desc,
|
|
struct nvgpu_gr_ctx *gr_ctx, u32 class_num,
|
|
u32 graphics_preempt_mode, u32 compute_preempt_mode)
|
|
{
|
|
int err = 0;
|
|
|
|
/* check for invalid combinations */
|
|
if (nvgpu_gr_ctx_check_valid_preemption_mode(g, ch, gr_ctx,
|
|
graphics_preempt_mode, compute_preempt_mode) == false) {
|
|
err = -EINVAL;
|
|
goto fail;
|
|
}
|
|
|
|
nvgpu_log(g, gpu_dbg_gr, "graphics_preempt_mode=%u compute_preempt_mode=%u",
|
|
graphics_preempt_mode, compute_preempt_mode);
|
|
|
|
#ifdef CONFIG_NVGPU_GRAPHICS
|
|
err = nvgpu_gr_obj_ctx_set_graphics_preemption_mode(g, config,
|
|
gr_ctx_desc, gr_ctx, graphics_preempt_mode);
|
|
|
|
if (err != 0) {
|
|
goto fail;
|
|
}
|
|
#endif
|
|
|
|
err = nvgpu_gr_obj_ctx_set_compute_preemption_mode(g, gr_ctx,
|
|
class_num, compute_preempt_mode);
|
|
|
|
fail:
|
|
return err;
|
|
}
|
|
|
|
#ifdef CONFIG_NVGPU_GFXP
|
|
static void nvgpu_gr_obj_ctx_commit_veid0_preemption_buffers(struct gk20a *g,
|
|
struct nvgpu_gr_config *config,
|
|
struct nvgpu_gr_ctx *gr_ctx,
|
|
struct nvgpu_gr_ctx_mappings *mappings)
|
|
{
|
|
u64 addr;
|
|
u32 size;
|
|
struct nvgpu_mem *mem;
|
|
|
|
addr = nvgpu_gr_ctx_mappings_get_ctx_va(mappings, NVGPU_GR_CTX_BETACB_CTXSW);
|
|
g->ops.gr.init.commit_global_attrib_cb(g, gr_ctx, mappings,
|
|
nvgpu_gr_config_get_tpc_count(config),
|
|
nvgpu_gr_config_get_max_tpc_count(config), addr,
|
|
true);
|
|
|
|
mem = nvgpu_gr_ctx_get_ctx_mem(gr_ctx, NVGPU_GR_CTX_PAGEPOOL_CTXSW);
|
|
addr = nvgpu_gr_ctx_mappings_get_ctx_va(mappings, NVGPU_GR_CTX_PAGEPOOL_CTXSW);
|
|
nvgpu_assert(mem->size <= U32_MAX);
|
|
size = (u32)mem->size;
|
|
|
|
g->ops.gr.init.commit_global_pagepool(g, gr_ctx, addr, size,
|
|
true, false);
|
|
|
|
mem = nvgpu_gr_ctx_get_ctx_mem(gr_ctx, NVGPU_GR_CTX_SPILL_CTXSW);
|
|
addr = nvgpu_gr_ctx_mappings_get_ctx_va(mappings, NVGPU_GR_CTX_SPILL_CTXSW);
|
|
nvgpu_assert(mem->size <= U32_MAX);
|
|
size = (u32)mem->size;
|
|
|
|
g->ops.gr.init.commit_ctxsw_spill(g, gr_ctx, addr, size, true);
|
|
}
|
|
#endif
|
|
|
|
bool nvgpu_gr_obj_ctx_is_gfx_engine(struct gk20a *g, struct nvgpu_tsg_subctx *subctx)
|
|
{
|
|
if (nvgpu_is_enabled(g, NVGPU_SUPPORT_TSG_SUBCONTEXTS)) {
|
|
if (!nvgpu_is_enabled(g, NVGPU_SUPPORT_MIG) &&
|
|
nvgpu_tsg_subctx_get_id(subctx) == CHANNEL_INFO_VEID0) {
|
|
return true;
|
|
}
|
|
} else if (!nvgpu_is_enabled(g, NVGPU_SUPPORT_MIG)) {
|
|
return true;
|
|
}
|
|
|
|
return false;
|
|
}
|
|
|
|
void nvgpu_gr_obj_ctx_update_ctxsw_preemption_mode(struct gk20a *g,
|
|
struct nvgpu_gr_config *config,
|
|
struct nvgpu_gr_ctx *gr_ctx, struct nvgpu_tsg_subctx *subctx,
|
|
struct nvgpu_gr_ctx_mappings *mappings)
|
|
{
|
|
(void)config;
|
|
(void)subctx;
|
|
(void)mappings;
|
|
|
|
nvgpu_log(g, gpu_dbg_fn | gpu_dbg_gr, " ");
|
|
|
|
nvgpu_gr_ctx_set_preemption_modes(g, gr_ctx);
|
|
|
|
#ifdef CONFIG_NVGPU_GFXP
|
|
if (!nvgpu_is_enabled(g, NVGPU_SUPPORT_PREEMPTION_GFXP)) {
|
|
goto done;
|
|
}
|
|
|
|
if (!nvgpu_mem_is_valid(
|
|
nvgpu_gr_ctx_get_ctx_mem(gr_ctx,
|
|
NVGPU_GR_CTX_PREEMPT_CTXSW))) {
|
|
goto done;
|
|
}
|
|
|
|
/*
|
|
* Commit NVGPU_GR_CTX_PREEMPT_CTXSW gpu va for all subcontexts
|
|
* considering VEID0 gpu va when subcontexts are enabled.
|
|
*/
|
|
if (nvgpu_is_enabled(g, NVGPU_SUPPORT_TSG_SUBCONTEXTS)) {
|
|
nvgpu_tsg_subctxs_set_preemption_buffer_va(subctx);
|
|
} else {
|
|
nvgpu_gr_ctx_set_preemption_buffer_va(g, gr_ctx);
|
|
}
|
|
|
|
if (!nvgpu_gr_obj_ctx_is_gfx_engine(g, subctx)) {
|
|
goto done;
|
|
}
|
|
|
|
if (nvgpu_gr_obj_ctx_preempt_buffers_patched(gr_ctx)) {
|
|
goto done;
|
|
}
|
|
|
|
nvgpu_gr_obj_ctx_set_preempt_buffers_patched(gr_ctx, true);
|
|
|
|
/*
|
|
* Commit other preemption buffers only for VEID0 when subcontexts are
|
|
* enabled. Commit always when subcontext are disabled.
|
|
*/
|
|
nvgpu_gr_ctx_patch_write_begin(g, gr_ctx, true);
|
|
|
|
nvgpu_gr_obj_ctx_commit_veid0_preemption_buffers(g, config,
|
|
gr_ctx, mappings);
|
|
|
|
g->ops.gr.init.commit_cbes_reserve(g, gr_ctx, true);
|
|
|
|
if (g->ops.gr.init.gfxp_wfi_timeout != NULL) {
|
|
g->ops.gr.init.gfxp_wfi_timeout(g, gr_ctx, true);
|
|
}
|
|
|
|
if (g->ops.gr.init.commit_gfxp_rtv_cb != NULL) {
|
|
g->ops.gr.init.commit_gfxp_rtv_cb(g, gr_ctx, mappings, true);
|
|
}
|
|
|
|
nvgpu_gr_ctx_patch_write_end(g, gr_ctx, true);
|
|
|
|
done:
|
|
#endif
|
|
nvgpu_log(g, gpu_dbg_fn | gpu_dbg_gr, "done");
|
|
}
|
|
|
|
void nvgpu_gr_obj_ctx_commit_global_ctx_buffers(struct gk20a *g,
|
|
struct nvgpu_gr_global_ctx_buffer_desc *global_ctx_buffer,
|
|
struct nvgpu_gr_config *config,
|
|
struct nvgpu_gr_ctx *gr_ctx,
|
|
struct nvgpu_tsg_subctx *subctx,
|
|
struct nvgpu_gr_ctx_mappings *mappings,
|
|
bool patch)
|
|
{
|
|
u64 addr;
|
|
u32 size;
|
|
|
|
nvgpu_log(g, gpu_dbg_fn | gpu_dbg_gr, " ");
|
|
|
|
if (patch) {
|
|
nvgpu_gr_ctx_patch_write_begin(g, gr_ctx, true);
|
|
}
|
|
|
|
/*
|
|
* MIG supports only compute class.
|
|
* Skip BUNDLE_CB, PAGEPOOL, ATTRIBUTE_CB and RTV_CB
|
|
* if 2D/3D/I2M classes(graphics) are not supported.
|
|
*/
|
|
if (nvgpu_gr_obj_ctx_is_gfx_engine(g, subctx)) {
|
|
if (patch && nvgpu_gr_obj_ctx_global_ctx_buffers_patched(gr_ctx)) {
|
|
goto commit_sm_id;
|
|
}
|
|
|
|
/* global pagepool buffer */
|
|
addr = nvgpu_gr_ctx_mappings_get_global_ctx_va(mappings,
|
|
NVGPU_GR_GLOBAL_CTX_PAGEPOOL_VA);
|
|
size = nvgpu_safe_cast_u64_to_u32(nvgpu_gr_global_ctx_get_size(
|
|
global_ctx_buffer,
|
|
NVGPU_GR_GLOBAL_CTX_PAGEPOOL));
|
|
|
|
g->ops.gr.init.commit_global_pagepool(g, gr_ctx, addr, size,
|
|
patch, true);
|
|
|
|
/* global bundle cb */
|
|
addr = nvgpu_gr_ctx_mappings_get_global_ctx_va(mappings,
|
|
NVGPU_GR_GLOBAL_CTX_CIRCULAR_VA);
|
|
size = nvgpu_safe_cast_u64_to_u32(
|
|
g->ops.gr.init.get_bundle_cb_default_size(g));
|
|
|
|
g->ops.gr.init.commit_global_bundle_cb(g, gr_ctx, addr, size,
|
|
patch);
|
|
|
|
/* global attrib cb */
|
|
addr = nvgpu_gr_ctx_mappings_get_global_ctx_va(mappings,
|
|
NVGPU_GR_GLOBAL_CTX_ATTRIBUTE_VA);
|
|
|
|
g->ops.gr.init.commit_global_attrib_cb(g, gr_ctx, mappings,
|
|
nvgpu_gr_config_get_tpc_count(config),
|
|
nvgpu_gr_config_get_max_tpc_count(config), addr, patch);
|
|
|
|
g->ops.gr.init.commit_global_cb_manager(g, config, gr_ctx,
|
|
patch);
|
|
|
|
#ifdef CONFIG_NVGPU_GRAPHICS
|
|
if (g->ops.gr.init.commit_rtv_cb != NULL) {
|
|
/* RTV circular buffer */
|
|
addr = nvgpu_gr_ctx_mappings_get_global_ctx_va(mappings,
|
|
NVGPU_GR_GLOBAL_CTX_RTV_CIRCULAR_BUFFER_VA);
|
|
|
|
g->ops.gr.init.commit_rtv_cb(g, addr, gr_ctx, patch);
|
|
}
|
|
#endif
|
|
|
|
if (patch) {
|
|
nvgpu_gr_obj_ctx_set_global_ctx_buffers_patched(gr_ctx,
|
|
true);
|
|
}
|
|
}
|
|
|
|
commit_sm_id:
|
|
if (patch && nvgpu_gr_ctx_get_ctx_initialized(gr_ctx)) {
|
|
goto out;
|
|
}
|
|
|
|
#ifdef CONFIG_NVGPU_SM_DIVERSITY
|
|
if ((nvgpu_is_enabled(g, NVGPU_SUPPORT_SM_DIVERSITY)) &&
|
|
(nvgpu_gr_ctx_get_sm_diversity_config(gr_ctx) !=
|
|
NVGPU_DEFAULT_SM_DIVERSITY_CONFIG) &&
|
|
(g->ops.gr.init.commit_sm_id_programming != NULL)) {
|
|
int err;
|
|
|
|
err = g->ops.gr.init.commit_sm_id_programming(
|
|
g, config, gr_ctx, patch);
|
|
if (err != 0) {
|
|
nvgpu_err(g,
|
|
"commit_sm_id_programming failed err=%d", err);
|
|
}
|
|
}
|
|
#endif
|
|
|
|
#ifdef CONFIG_NVGPU_GRAPHICS
|
|
if (g->ops.gr.init.commit_rops_crop_override != NULL) {
|
|
g->ops.gr.init.commit_rops_crop_override(g, gr_ctx, patch);
|
|
}
|
|
#endif
|
|
|
|
out:
|
|
if (patch) {
|
|
nvgpu_gr_ctx_patch_write_end(g, gr_ctx, true);
|
|
}
|
|
|
|
nvgpu_log(g, gpu_dbg_fn | gpu_dbg_gr, "done");
|
|
}
|
|
|
|
static int nvgpu_gr_obj_ctx_alloc_sw_bundle(struct gk20a *g)
|
|
{
|
|
int err = 0;
|
|
struct netlist_av_list *sw_bundle_init =
|
|
nvgpu_netlist_get_sw_bundle_init_av_list(g);
|
|
struct netlist_av_list *sw_veid_bundle_init =
|
|
nvgpu_netlist_get_sw_veid_bundle_init_av_list(g);
|
|
struct netlist_av64_list *sw_bundle64_init =
|
|
nvgpu_netlist_get_sw_bundle64_init_av64_list(g);
|
|
|
|
/* enable pipe mode override */
|
|
g->ops.gr.init.pipe_mode_override(g, true);
|
|
|
|
/* load bundle init */
|
|
err = g->ops.gr.init.load_sw_bundle_init(g, sw_bundle_init);
|
|
if (err != 0) {
|
|
goto error;
|
|
}
|
|
|
|
if (g->ops.gr.init.load_sw_veid_bundle != NULL) {
|
|
err = g->ops.gr.init.load_sw_veid_bundle(g,
|
|
sw_veid_bundle_init);
|
|
if (err != 0) {
|
|
goto error;
|
|
}
|
|
}
|
|
|
|
if (g->ops.gr.init.load_sw_bundle64 != NULL) {
|
|
err = g->ops.gr.init.load_sw_bundle64(g, sw_bundle64_init);
|
|
if (err != 0) {
|
|
goto error;
|
|
}
|
|
}
|
|
|
|
/* disable pipe mode override */
|
|
g->ops.gr.init.pipe_mode_override(g, false);
|
|
|
|
err = g->ops.gr.init.wait_idle(g);
|
|
|
|
return err;
|
|
|
|
error:
|
|
/* in case of error skip waiting for GR idle - just restore state */
|
|
g->ops.gr.init.pipe_mode_override(g, false);
|
|
|
|
return err;
|
|
}
|
|
|
|
static int nvgpu_gr_obj_ctx_init_hw_state(struct gk20a *g,
|
|
struct nvgpu_mem *inst_block)
|
|
{
|
|
int err = 0;
|
|
u32 data;
|
|
u32 i;
|
|
struct netlist_aiv_list *sw_ctx_load =
|
|
nvgpu_netlist_get_sw_ctx_load_aiv_list(g);
|
|
|
|
nvgpu_log(g, gpu_dbg_gr, " ");
|
|
|
|
err = g->ops.gr.init.fe_pwr_mode_force_on(g, true);
|
|
if (err != 0) {
|
|
goto clean_up;
|
|
}
|
|
|
|
g->ops.gr.init.override_context_reset(g);
|
|
|
|
err = g->ops.gr.init.fe_pwr_mode_force_on(g, false);
|
|
if (err != 0) {
|
|
goto clean_up;
|
|
}
|
|
|
|
data = g->ops.gr.falcon.get_fecs_current_ctx_data(g, inst_block);
|
|
err = g->ops.gr.falcon.ctrl_ctxsw(g,
|
|
NVGPU_GR_FALCON_METHOD_ADDRESS_BIND_PTR, data, NULL);
|
|
if (err != 0) {
|
|
goto clean_up;
|
|
}
|
|
|
|
err = g->ops.gr.init.wait_idle(g);
|
|
|
|
/* load ctx init */
|
|
nvgpu_log_info(g, "begin: netlist: sw_ctx_load: register writes");
|
|
for (i = 0U; i < sw_ctx_load->count; i++) {
|
|
#ifdef CONFIG_NVGPU_MIG
|
|
if ((nvgpu_is_enabled(g, NVGPU_SUPPORT_MIG)) &&
|
|
(g->ops.gr.init.is_allowed_reg != NULL) &&
|
|
(!(g->ops.gr.init.is_allowed_reg(g,
|
|
sw_ctx_load->l[i].addr)))) {
|
|
nvgpu_log(g, gpu_dbg_mig | gpu_dbg_gr,
|
|
"(MIG) Skip graphics ctx load reg "
|
|
"index[%u] addr[%x] value[%x] ",
|
|
i, sw_ctx_load->l[i].addr,
|
|
sw_ctx_load->l[i].value);
|
|
continue;
|
|
}
|
|
#endif
|
|
nvgpu_writel(g, sw_ctx_load->l[i].addr,
|
|
sw_ctx_load->l[i].value);
|
|
}
|
|
nvgpu_log_info(g, "end: netlist: sw_ctx_load: register writes");
|
|
|
|
nvgpu_log_info(g, "configure sm_hww_esr_report mask after sw_ctx_load");
|
|
g->ops.gr.intr.set_hww_esr_report_mask(g);
|
|
|
|
#ifdef CONFIG_NVGPU_GFXP
|
|
if (g->ops.gr.init.preemption_state != NULL) {
|
|
err = g->ops.gr.init.preemption_state(g);
|
|
if (err != 0) {
|
|
goto clean_up;
|
|
}
|
|
}
|
|
#endif
|
|
|
|
nvgpu_cg_blcg_gr_load_enable(g);
|
|
|
|
err = g->ops.gr.init.wait_idle(g);
|
|
|
|
clean_up:
|
|
if (err == 0) {
|
|
nvgpu_log(g, gpu_dbg_gr, "done");
|
|
}
|
|
return err;
|
|
}
|
|
|
|
static int nvgpu_gr_obj_ctx_commit_hw_state(struct gk20a *g,
|
|
struct nvgpu_gr_global_ctx_buffer_desc *global_ctx_buffer,
|
|
struct nvgpu_gr_config *config, struct nvgpu_gr_ctx *gr_ctx,
|
|
struct nvgpu_tsg_subctx *subctx, struct nvgpu_gr_ctx_mappings *mappings)
|
|
{
|
|
int err = 0;
|
|
struct netlist_av_list *sw_method_init =
|
|
nvgpu_netlist_get_sw_method_init_av_list(g);
|
|
#ifdef CONFIG_NVGPU_GR_GOLDEN_CTX_VERIFICATION
|
|
struct netlist_av_list *sw_bundle_init =
|
|
nvgpu_netlist_get_sw_bundle_init_av_list(g);
|
|
#endif
|
|
|
|
nvgpu_log(g, gpu_dbg_gr, " ");
|
|
|
|
/* disable fe_go_idle */
|
|
g->ops.gr.init.fe_go_idle_timeout(g, false);
|
|
|
|
nvgpu_gr_obj_ctx_commit_global_ctx_buffers(g, global_ctx_buffer,
|
|
config, gr_ctx, subctx, mappings, false);
|
|
|
|
if (!nvgpu_is_enabled(g, NVGPU_SUPPORT_MIG)) {
|
|
/* override a few ctx state registers */
|
|
g->ops.gr.init.commit_global_timeslice(g);
|
|
}
|
|
|
|
/* floorsweep anything left */
|
|
err = nvgpu_gr_fs_state_init(g, config);
|
|
if (err != 0) {
|
|
goto restore_fe_go_idle;
|
|
}
|
|
|
|
err = g->ops.gr.init.wait_idle(g);
|
|
if (err != 0) {
|
|
goto restore_fe_go_idle;
|
|
}
|
|
if (g->ops.gr.init.auto_go_idle != NULL) {
|
|
g->ops.gr.init.auto_go_idle(g, false);
|
|
}
|
|
err = nvgpu_gr_obj_ctx_alloc_sw_bundle(g);
|
|
if (err != 0) {
|
|
goto restore_fe_go_idle;
|
|
}
|
|
|
|
if (g->ops.gr.init.auto_go_idle != NULL) {
|
|
g->ops.gr.init.auto_go_idle(g, true);
|
|
}
|
|
|
|
/* restore fe_go_idle */
|
|
g->ops.gr.init.fe_go_idle_timeout(g, true);
|
|
|
|
/* load method init */
|
|
g->ops.gr.init.load_method_init(g, sw_method_init);
|
|
|
|
#ifdef CONFIG_NVGPU_GR_GOLDEN_CTX_VERIFICATION
|
|
/* restore stats bundle data through mme shadow methods */
|
|
if (g->ops.gr.init.restore_stats_counter_bundle_data != NULL) {
|
|
g->ops.gr.init.restore_stats_counter_bundle_data(g,
|
|
sw_bundle_init);
|
|
}
|
|
#endif
|
|
|
|
err = g->ops.gr.init.wait_idle(g);
|
|
if (err != 0) {
|
|
goto clean_up;
|
|
}
|
|
|
|
nvgpu_log(g, gpu_dbg_gr, "done");
|
|
return 0;
|
|
|
|
restore_fe_go_idle:
|
|
/* restore fe_go_idle */
|
|
g->ops.gr.init.fe_go_idle_timeout(g, true);
|
|
if (g->ops.gr.init.auto_go_idle != NULL) {
|
|
g->ops.gr.init.auto_go_idle(g, true);
|
|
}
|
|
|
|
clean_up:
|
|
return err;
|
|
}
|
|
|
|
static int nvgpu_gr_obj_ctx_save_golden_ctx(struct gk20a *g,
|
|
struct nvgpu_gr_obj_ctx_golden_image *golden_image,
|
|
struct nvgpu_gr_ctx *gr_ctx, struct nvgpu_mem *inst_block)
|
|
{
|
|
int err = 0;
|
|
struct nvgpu_mem *gr_mem;
|
|
u64 size;
|
|
u32 data;
|
|
|
|
nvgpu_log(g, gpu_dbg_gr, " ");
|
|
|
|
gr_mem = nvgpu_gr_ctx_get_ctx_mem(gr_ctx, NVGPU_GR_CTX_CTX);
|
|
size = nvgpu_gr_obj_ctx_get_golden_image_size(golden_image);
|
|
|
|
#ifdef CONFIG_NVGPU_GR_GOLDEN_CTX_VERIFICATION
|
|
/*
|
|
* Save ctx data before first golden context save. Restore same data
|
|
* before second golden context save. This temporary copy is
|
|
* saved in local_golden_image_temp.
|
|
*/
|
|
nvgpu_gr_global_ctx_init_local_golden_image(g,
|
|
golden_image->local_golden_image_copy, gr_mem, size);
|
|
#endif
|
|
|
|
data = g->ops.gr.falcon.get_fecs_current_ctx_data(g, inst_block);
|
|
err = g->ops.gr.falcon.ctrl_ctxsw(g,
|
|
NVGPU_GR_FALCON_METHOD_GOLDEN_IMAGE_SAVE, data, NULL);
|
|
if (err != 0) {
|
|
goto clean_up;
|
|
}
|
|
|
|
nvgpu_gr_global_ctx_init_local_golden_image(g,
|
|
golden_image->local_golden_image, gr_mem, size);
|
|
|
|
#ifdef CONFIG_NVGPU_GR_GOLDEN_CTX_VERIFICATION
|
|
/* Before second golden context save restore to before known state */
|
|
nvgpu_gr_global_ctx_load_local_golden_image(g,
|
|
golden_image->local_golden_image_copy, gr_mem);
|
|
|
|
/* Initiate second golden context save */
|
|
data = g->ops.gr.falcon.get_fecs_current_ctx_data(g, inst_block);
|
|
err = g->ops.gr.falcon.ctrl_ctxsw(g,
|
|
NVGPU_GR_FALCON_METHOD_GOLDEN_IMAGE_SAVE, data, NULL);
|
|
if (err != 0) {
|
|
goto clean_up;
|
|
}
|
|
|
|
/* Copy the data to local buffer */
|
|
nvgpu_gr_global_ctx_init_local_golden_image(g,
|
|
golden_image->local_golden_image_copy, gr_mem, size);
|
|
|
|
/* Compare two golden context images */
|
|
if (!nvgpu_gr_global_ctx_compare_golden_images(g,
|
|
nvgpu_mem_is_sysmem(gr_mem),
|
|
golden_image->local_golden_image,
|
|
golden_image->local_golden_image_copy,
|
|
size)) {
|
|
nvgpu_err(g, "golden context mismatch");
|
|
err = -ENOMEM;
|
|
}
|
|
|
|
/* free temporary copy now */
|
|
nvgpu_gr_global_ctx_deinit_local_golden_image(g,
|
|
golden_image->local_golden_image_copy);
|
|
golden_image->local_golden_image_copy = NULL;
|
|
#endif
|
|
|
|
clean_up:
|
|
if (err == 0) {
|
|
nvgpu_log(g, gpu_dbg_gr, "golden image saved with size = %llu", size);
|
|
}
|
|
return err;
|
|
}
|
|
|
|
/*
|
|
* init global golden image from a fresh gr_ctx in channel ctx.
|
|
* save a copy in local_golden_image.
|
|
*/
|
|
int nvgpu_gr_obj_ctx_alloc_golden_ctx_image(struct gk20a *g,
|
|
struct nvgpu_gr_obj_ctx_golden_image *golden_image,
|
|
struct nvgpu_gr_global_ctx_buffer_desc *global_ctx_buffer,
|
|
struct nvgpu_gr_config *config,
|
|
struct nvgpu_gr_ctx *gr_ctx,
|
|
struct nvgpu_tsg_subctx *subctx,
|
|
struct nvgpu_gr_ctx_mappings *mappings,
|
|
struct nvgpu_mem *inst_block)
|
|
{
|
|
int err = 0;
|
|
|
|
nvgpu_log(g, gpu_dbg_fn | gpu_dbg_gr, " ");
|
|
|
|
/*
|
|
* golden ctx is global to all channels. Although only the first
|
|
* channel initializes golden image, driver needs to prevent multiple
|
|
* channels from initializing golden ctx at the same time
|
|
*/
|
|
if (golden_image->ready) {
|
|
nvgpu_log(g, gpu_dbg_gr, "golden image already saved");
|
|
goto clean_up;
|
|
}
|
|
|
|
err = nvgpu_gr_obj_ctx_init_hw_state(g, inst_block);
|
|
if (err != 0) {
|
|
goto clean_up;
|
|
}
|
|
|
|
err = nvgpu_gr_obj_ctx_commit_hw_state(g, global_ctx_buffer,
|
|
config, gr_ctx, subctx, mappings);
|
|
if (err != 0) {
|
|
goto clean_up;
|
|
}
|
|
|
|
#ifdef CONFIG_NVGPU_GRAPHICS
|
|
if (nvgpu_gr_obj_ctx_is_gfx_engine(g, subctx)) {
|
|
err = nvgpu_gr_ctx_init_zcull(g, gr_ctx);
|
|
if (err != 0) {
|
|
goto clean_up;
|
|
}
|
|
}
|
|
#endif
|
|
|
|
err = nvgpu_gr_obj_ctx_save_golden_ctx(g, golden_image,
|
|
gr_ctx, inst_block);
|
|
if (err != 0) {
|
|
goto clean_up;
|
|
}
|
|
|
|
/*
|
|
* Read and save register init values that need to be configured
|
|
* differently for graphics contexts.
|
|
* Updated values are written to the context in
|
|
* gops.gr.init.set_default_gfx_regs().
|
|
*/
|
|
if (g->ops.gr.init.capture_gfx_regs != NULL) {
|
|
g->ops.gr.init.capture_gfx_regs(g, &golden_image->gfx_regs);
|
|
}
|
|
|
|
golden_image->ready = true;
|
|
#ifdef CONFIG_NVGPU_POWER_PG
|
|
nvgpu_pmu_set_golden_image_initialized(g, GOLDEN_IMG_READY);
|
|
#endif
|
|
g->ops.gr.falcon.set_current_ctx_invalid(g);
|
|
|
|
clean_up:
|
|
if (err != 0) {
|
|
nvgpu_err(g, "fail");
|
|
} else {
|
|
nvgpu_log(g, gpu_dbg_fn | gpu_dbg_gr, "done");
|
|
}
|
|
|
|
return err;
|
|
}
|
|
|
|
static void nvgpu_gr_obj_ctx_gr_ctx_set_size(struct gk20a *g,
|
|
struct nvgpu_gr_obj_ctx_golden_image *golden_image,
|
|
struct nvgpu_gr_ctx_desc *gr_ctx_desc)
|
|
{
|
|
u64 size;
|
|
|
|
size = nvgpu_gr_obj_ctx_get_golden_image_size(golden_image);
|
|
nvgpu_gr_ctx_set_size(gr_ctx_desc, NVGPU_GR_CTX_CTX,
|
|
nvgpu_safe_cast_u64_to_u32(size));
|
|
|
|
nvgpu_log(g, gpu_dbg_gr, "gr_ctx size = %llu", size);
|
|
}
|
|
|
|
static void nvgpu_gr_obj_ctx_patch_ctx_set_size(struct gk20a *g,
|
|
struct nvgpu_gr_config *config,
|
|
struct nvgpu_gr_ctx_desc *gr_ctx_desc)
|
|
{
|
|
u32 size;
|
|
|
|
size = nvgpu_safe_mult_u32(
|
|
g->ops.gr.init.get_patch_slots(g, config),
|
|
PATCH_CTX_SLOTS_REQUIRED_PER_ENTRY);
|
|
nvgpu_gr_ctx_set_size(gr_ctx_desc, NVGPU_GR_CTX_PATCH_CTX, size);
|
|
|
|
nvgpu_log(g, gpu_dbg_gr, "patch_ctx size = %u", size);
|
|
}
|
|
|
|
static int nvgpu_gr_obj_ctx_alloc_buffers(struct gk20a *g,
|
|
struct nvgpu_channel *ch,
|
|
struct nvgpu_gr_obj_ctx_golden_image *golden_image,
|
|
struct nvgpu_gr_ctx_desc *gr_ctx_desc,
|
|
struct nvgpu_gr_config *config,
|
|
struct nvgpu_gr_ctx *gr_ctx,
|
|
u32 class_num, u32 flags)
|
|
{
|
|
int err;
|
|
|
|
(void)ch;
|
|
(void)class_num;
|
|
(void)flags;
|
|
|
|
nvgpu_log(g, gpu_dbg_gr, " ");
|
|
|
|
if (!nvgpu_gr_ctx_get_ctx_initialized(gr_ctx)) {
|
|
nvgpu_gr_obj_ctx_gr_ctx_set_size(g, golden_image, gr_ctx_desc);
|
|
nvgpu_gr_obj_ctx_patch_ctx_set_size(g, config, gr_ctx_desc);
|
|
nvgpu_gr_ctx_set_patch_ctx_data_count(gr_ctx, 0);
|
|
}
|
|
|
|
#if defined(CONFIG_NVGPU_GFXP) || defined(CONFIG_NVGPU_CILP)
|
|
err = nvgpu_gr_obj_ctx_init_ctxsw_preemption(g, ch, config,
|
|
gr_ctx_desc, gr_ctx, class_num, flags);
|
|
if (err != 0) {
|
|
nvgpu_err(g, "fail to init preemption mode");
|
|
return err;
|
|
}
|
|
#endif
|
|
|
|
err = nvgpu_gr_ctx_alloc_ctx_buffers(g, gr_ctx_desc, gr_ctx);
|
|
if (err != 0) {
|
|
nvgpu_err(g, "fail to allocate ctx buffers");
|
|
return err;
|
|
}
|
|
|
|
nvgpu_log(g, gpu_dbg_gr, "done");
|
|
|
|
return err;
|
|
}
|
|
|
|
int nvgpu_gr_obj_ctx_init_golden_context_image(struct gk20a *g)
|
|
{
|
|
struct nvgpu_gr_obj_ctx_golden_image *golden_image =
|
|
nvgpu_gr_get_golden_image_ptr(g);
|
|
struct nvgpu_setup_bind_args setup_bind_args;
|
|
struct nvgpu_channel *veid0_ch;
|
|
u64 user_size, kernel_size;
|
|
struct nvgpu_tsg *tsg;
|
|
struct vm_gk20a *vm;
|
|
u32 big_page_size;
|
|
u32 obj_class;
|
|
int err = 0;
|
|
|
|
err = gk20a_busy(g);
|
|
if (err != 0) {
|
|
nvgpu_err(g, "failed to power on, %d", err);
|
|
return err;
|
|
}
|
|
|
|
nvgpu_mutex_acquire(&golden_image->ctx_mutex);
|
|
|
|
if (golden_image->ready) {
|
|
goto out;
|
|
}
|
|
|
|
big_page_size = g->ops.mm.gmmu.get_default_big_page_size();
|
|
|
|
/* allocate a tsg */
|
|
tsg = nvgpu_tsg_open(g, 0);
|
|
if (tsg == NULL) {
|
|
nvgpu_err(g, "tsg not available");
|
|
err = -ENOMEM;
|
|
goto out;
|
|
}
|
|
|
|
/* allocate a VM */
|
|
g->ops.mm.get_default_va_sizes(NULL, &user_size, &kernel_size);
|
|
vm = nvgpu_vm_init(g, big_page_size,
|
|
U64(big_page_size) << U64(10),
|
|
nvgpu_safe_sub_u64(user_size,
|
|
U64(big_page_size) << U64(10)),
|
|
kernel_size,
|
|
0ULL,
|
|
false, false, false, "golden_context");
|
|
if (vm == NULL) {
|
|
nvgpu_err(g, "vm init failed");
|
|
err = -ENOMEM;
|
|
goto out_release_tsg;
|
|
}
|
|
|
|
/* allocate veid0 channel by specifying GR runlist id */
|
|
veid0_ch = nvgpu_channel_open_new(g, nvgpu_engine_get_gr_runlist_id(g),
|
|
true, 0, 0);
|
|
if (veid0_ch == NULL) {
|
|
nvgpu_err(g, "channel not available");
|
|
err = -ENOMEM;
|
|
goto out_release_vm;
|
|
}
|
|
|
|
veid0_ch->golden_ctx_init_ch = true;
|
|
|
|
/* bind the channel to the vm */
|
|
err = g->ops.mm.vm_bind_channel(vm, veid0_ch);
|
|
if (err != 0) {
|
|
nvgpu_err(g, "could not bind vm");
|
|
goto out_release_ch;
|
|
}
|
|
|
|
/* bind the channel to the tsg */
|
|
err = nvgpu_tsg_bind_channel(tsg, veid0_ch);
|
|
if (err != 0) {
|
|
nvgpu_err(g, "unable to bind to tsg");
|
|
goto out_release_ch;
|
|
}
|
|
|
|
setup_bind_args.num_gpfifo_entries = 1024;
|
|
setup_bind_args.num_inflight_jobs = 0;
|
|
#ifdef CONFIG_NVGPU_KERNEL_MODE_SUBMIT
|
|
setup_bind_args.flags = 0;
|
|
#else
|
|
/*
|
|
* Usermode gpfifo and userd buffers are just allocated here but they
|
|
* are not used for submitting any work. Since these buffers are
|
|
* nvgpu allocated ones, we don't specify userd_dmabuf_fd and
|
|
* gpfifo_dmabuf_fd here.
|
|
*/
|
|
setup_bind_args.flags = NVGPU_SETUP_BIND_FLAGS_USERMODE_SUPPORT;
|
|
#endif
|
|
err = nvgpu_channel_setup_bind(veid0_ch, &setup_bind_args);
|
|
if (err != 0) {
|
|
nvgpu_err(g, "unable to setup and bind channel");
|
|
goto out_release_ch;
|
|
}
|
|
|
|
#ifdef CONFIG_NVGPU_HAL_NON_FUSA
|
|
obj_class = MAXWELL_B;
|
|
#else
|
|
obj_class = VOLTA_A;
|
|
#endif
|
|
|
|
/* allocate obj_ctx to initialize golden image */
|
|
err = g->ops.gr.setup.alloc_obj_ctx(veid0_ch, obj_class, 0U);
|
|
if (err != 0) {
|
|
nvgpu_err(g, "unable to alloc obj_ctx");
|
|
goto out_release_ch;
|
|
}
|
|
|
|
/* This state update is needed for vGPU case */
|
|
golden_image->ready = true;
|
|
|
|
nvgpu_log(g, gpu_dbg_gr, "Golden context image initialized!");
|
|
|
|
out_release_ch:
|
|
nvgpu_channel_close(veid0_ch);
|
|
out_release_vm:
|
|
nvgpu_vm_put(vm);
|
|
out_release_tsg:
|
|
nvgpu_ref_put(&tsg->refcount, nvgpu_tsg_release);
|
|
out:
|
|
nvgpu_mutex_release(&golden_image->ctx_mutex);
|
|
gk20a_idle(g);
|
|
return err;
|
|
}
|
|
|
|
static int nvgpu_gr_obj_ctx_load_golden_image(struct gk20a *g,
|
|
struct nvgpu_gr_obj_ctx_golden_image *golden_image,
|
|
struct nvgpu_gr_global_ctx_buffer_desc *global_ctx_buffer,
|
|
struct nvgpu_gr_config *config,
|
|
struct nvgpu_gr_ctx *gr_ctx,
|
|
struct nvgpu_tsg_subctx *subctx,
|
|
struct nvgpu_gr_ctx_mappings *mappings,
|
|
struct nvgpu_mem *inst_block,
|
|
bool cde)
|
|
{
|
|
int err;
|
|
|
|
/* init golden image */
|
|
err = nvgpu_gr_obj_ctx_alloc_golden_ctx_image(g, golden_image,
|
|
global_ctx_buffer, config, gr_ctx, subctx,
|
|
mappings, inst_block);
|
|
if (err != 0) {
|
|
nvgpu_err(g, "fail to init golden ctx image");
|
|
return err;
|
|
}
|
|
|
|
#ifdef CONFIG_NVGPU_POWER_PG
|
|
/* Re-enable ELPG now that golden image has been initialized.
|
|
* The PMU PG init code may already have tried to enable elpg, but
|
|
* would not have been able to complete this action since the golden
|
|
* image hadn't been initialized yet, so do this now.
|
|
*/
|
|
err = nvgpu_pmu_reenable_elpg(g);
|
|
if (err != 0) {
|
|
nvgpu_err(g, "fail to re-enable elpg");
|
|
return err;
|
|
}
|
|
#endif
|
|
|
|
/* load golden image */
|
|
nvgpu_gr_ctx_load_golden_ctx_image(g, gr_ctx, mappings,
|
|
golden_image->local_golden_image, cde);
|
|
|
|
return 0;
|
|
}
|
|
|
|
int nvgpu_gr_obj_ctx_alloc(struct gk20a *g,
|
|
struct nvgpu_gr_obj_ctx_golden_image *golden_image,
|
|
struct nvgpu_gr_global_ctx_buffer_desc *global_ctx_buffer,
|
|
struct nvgpu_gr_ctx_desc *gr_ctx_desc,
|
|
struct nvgpu_gr_config *config,
|
|
struct nvgpu_gr_ctx *gr_ctx,
|
|
struct nvgpu_channel *c,
|
|
struct nvgpu_tsg_subctx *subctx,
|
|
struct nvgpu_gr_ctx_mappings *mappings,
|
|
struct nvgpu_mem *inst_block,
|
|
u32 class_num, u32 flags,
|
|
bool cde, bool vpr)
|
|
{
|
|
int err = 0;
|
|
|
|
nvgpu_log(g, gpu_dbg_fn | gpu_dbg_gr, " ");
|
|
|
|
err = nvgpu_gr_obj_ctx_alloc_buffers(g, c, golden_image, gr_ctx_desc,
|
|
config, gr_ctx, class_num, flags);
|
|
if (err != 0) {
|
|
nvgpu_err(g, "failed to alloc ctx buffers");
|
|
goto out;
|
|
}
|
|
|
|
if (!nvgpu_gr_ctx_get_ctx_initialized(gr_ctx)) {
|
|
nvgpu_gr_ctx_init_ctx_buffers_mapping_flags(g, gr_ctx);
|
|
}
|
|
|
|
err = nvgpu_gr_ctx_mappings_map_gr_ctx_buffers(g, gr_ctx, subctx,
|
|
global_ctx_buffer, mappings, vpr);
|
|
if (err != 0) {
|
|
nvgpu_err(g, "failed to map ctx buffers");
|
|
goto out;
|
|
}
|
|
|
|
nvgpu_gr_obj_ctx_commit_global_ctx_buffers(g, global_ctx_buffer,
|
|
config, gr_ctx, subctx, mappings, true);
|
|
|
|
/* commit gr ctx buffer */
|
|
nvgpu_gr_obj_ctx_commit_inst(g, inst_block, gr_ctx, subctx, mappings);
|
|
|
|
if (!nvgpu_gr_ctx_get_ctx_initialized(gr_ctx)) {
|
|
err = nvgpu_gr_obj_ctx_load_golden_image(g, golden_image,
|
|
global_ctx_buffer, config, gr_ctx, subctx,
|
|
mappings, inst_block, cde);
|
|
if (err != 0) {
|
|
nvgpu_err(g, "fail to load golden ctx image");
|
|
goto out;
|
|
}
|
|
}
|
|
|
|
nvgpu_gr_obj_ctx_update_ctxsw_preemption_mode(g, config, gr_ctx,
|
|
subctx, mappings);
|
|
|
|
#ifndef CONFIG_NVGPU_NON_FUSA
|
|
if (g->ops.gpu_class.is_valid_compute(class_num) &&
|
|
(g->ops.gr.init.set_default_compute_regs != NULL) &&
|
|
(!nvgpu_gr_obj_ctx_default_compute_regs_patched(gr_ctx))) {
|
|
g->ops.gr.init.set_default_compute_regs(g, gr_ctx);
|
|
nvgpu_gr_obj_ctx_set_default_compute_regs_patched(gr_ctx, true);
|
|
}
|
|
|
|
if ((g->ops.ltc.set_default_l2_max_ways_evict_last != NULL) &&
|
|
(!nvgpu_gr_ctx_get_ctx_initialized(gr_ctx))) {
|
|
g->ops.ltc.set_default_l2_max_ways_evict_last(g, gr_ctx);
|
|
}
|
|
#endif
|
|
|
|
#ifdef CONFIG_NVGPU_NON_FUSA
|
|
if ((g->ops.gr.init.enable_mme_config_ptimer != NULL) &&
|
|
(!nvgpu_gr_ctx_get_ctx_initialized(gr_ctx))) {
|
|
|
|
err = nvgpu_pg_elpg_protected_call(g,
|
|
g->ops.gr.init.enable_mme_config_ptimer(g, gr_ctx));
|
|
|
|
if (err != 0) {
|
|
nvgpu_err(g, "fail to enable mme_config_ptimer");
|
|
goto out;
|
|
}
|
|
}
|
|
#endif
|
|
/*
|
|
* Register init values are saved in
|
|
* gops.gr.init.capture_gfx_regs(). Update and set the values as
|
|
* required for graphics contexts.
|
|
*/
|
|
if (g->ops.gpu_class.is_valid_gfx(class_num) &&
|
|
(g->ops.gr.init.set_default_gfx_regs != NULL) &&
|
|
(!nvgpu_gr_obj_ctx_default_gfx_regs_patched(gr_ctx))) {
|
|
g->ops.gr.init.set_default_gfx_regs(g, gr_ctx, &golden_image->gfx_regs);
|
|
nvgpu_gr_obj_ctx_set_default_gfx_regs_patched(gr_ctx, true);
|
|
}
|
|
|
|
nvgpu_log(g, gpu_dbg_fn | gpu_dbg_gr, "done");
|
|
return 0;
|
|
out:
|
|
/*
|
|
* 1. gr_ctx, patch_ctx and global ctx buffer mapping
|
|
* can be reused so no need to release them.
|
|
* 2. golden image init and load is a one time thing so if
|
|
* they pass, no need to undo.
|
|
*/
|
|
nvgpu_err(g, "fail");
|
|
return err;
|
|
}
|
|
|
|
void nvgpu_gr_obj_ctx_set_golden_image_size(
|
|
struct nvgpu_gr_obj_ctx_golden_image *golden_image,
|
|
size_t size)
|
|
{
|
|
golden_image->size = size;
|
|
}
|
|
|
|
size_t nvgpu_gr_obj_ctx_get_golden_image_size(
|
|
struct nvgpu_gr_obj_ctx_golden_image *golden_image)
|
|
{
|
|
return golden_image->size;
|
|
}
|
|
|
|
#ifdef CONFIG_NVGPU_DEBUGGER
|
|
u32 *nvgpu_gr_obj_ctx_get_local_golden_image_ptr(
|
|
struct nvgpu_gr_obj_ctx_golden_image *golden_image)
|
|
{
|
|
return nvgpu_gr_global_ctx_get_local_golden_image_ptr(
|
|
golden_image->local_golden_image);
|
|
}
|
|
#endif
|
|
|
|
bool nvgpu_gr_obj_ctx_is_golden_image_ready(
|
|
struct nvgpu_gr_obj_ctx_golden_image *golden_image)
|
|
{
|
|
bool ready;
|
|
|
|
nvgpu_mutex_acquire(&golden_image->ctx_mutex);
|
|
ready = golden_image->ready;
|
|
nvgpu_mutex_release(&golden_image->ctx_mutex);
|
|
|
|
return ready;
|
|
}
|
|
|
|
int nvgpu_gr_obj_ctx_init(struct gk20a *g,
|
|
struct nvgpu_gr_obj_ctx_golden_image **gr_golden_image, u32 size)
|
|
{
|
|
struct nvgpu_gr_obj_ctx_golden_image *golden_image;
|
|
int err;
|
|
|
|
nvgpu_log(g, gpu_dbg_gr, "size = %u", size);
|
|
|
|
golden_image = nvgpu_kzalloc(g, sizeof(*golden_image));
|
|
if (golden_image == NULL) {
|
|
return -ENOMEM;
|
|
}
|
|
|
|
nvgpu_gr_obj_ctx_set_golden_image_size(golden_image, size);
|
|
|
|
nvgpu_mutex_init(&golden_image->ctx_mutex);
|
|
|
|
err = nvgpu_gr_global_ctx_alloc_local_golden_image(g,
|
|
&golden_image->local_golden_image, size);
|
|
if (err != 0) {
|
|
nvgpu_kfree(g, golden_image);
|
|
return err;
|
|
}
|
|
|
|
#ifdef CONFIG_NVGPU_GR_GOLDEN_CTX_VERIFICATION
|
|
err = nvgpu_gr_global_ctx_alloc_local_golden_image(g,
|
|
&golden_image->local_golden_image_copy, size);
|
|
if (err != 0) {
|
|
nvgpu_gr_global_ctx_deinit_local_golden_image(g,
|
|
golden_image->local_golden_image);
|
|
nvgpu_kfree(g, golden_image);
|
|
return err;
|
|
}
|
|
#endif
|
|
|
|
*gr_golden_image = golden_image;
|
|
|
|
return 0;
|
|
}
|
|
|
|
void nvgpu_gr_obj_ctx_deinit(struct gk20a *g,
|
|
struct nvgpu_gr_obj_ctx_golden_image *golden_image)
|
|
{
|
|
if (golden_image == NULL) {
|
|
return;
|
|
}
|
|
|
|
if (golden_image->local_golden_image != NULL) {
|
|
nvgpu_gr_global_ctx_deinit_local_golden_image(g,
|
|
golden_image->local_golden_image);
|
|
golden_image->local_golden_image = NULL;
|
|
}
|
|
#ifdef CONFIG_NVGPU_POWER_PG
|
|
nvgpu_pmu_set_golden_image_initialized(g, GOLDEN_IMG_NOT_READY);
|
|
#endif
|
|
golden_image->ready = false;
|
|
nvgpu_kfree(g, golden_image);
|
|
}
|
|
|