mirror of
git://nv-tegra.nvidia.com/linux-nvgpu.git
synced 2025-12-22 17:36:20 +03:00
gpu: nvgpu: configure CWD sm id regs before ucode load
GPCCS ucode is expecting the SM ID programmed in GPM and CWD registers to be in sync. So create a hal called gr.init.load_sm_id_config() for the sm_id programming for the CWD registers and invoke them before the ucode load. Initialize this hal only for the required GPUs. JIRA NVGPU-9757 Change-Id: Ib0984fd6326c37e0c2a06123041032575a23ec04 Signed-off-by: Ramalingam C <ramalingamc@nvidia.com> Reviewed-on: https://git-master.nvidia.com/r/c/linux-nvgpu/+/2864999 Reviewed-by: svc-mobile-coverity <svc-mobile-coverity@nvidia.com> Reviewed-by: svc-mobile-cert <svc-mobile-cert@nvidia.com> Reviewed-by: Seema Khowala <seemaj@nvidia.com> Reviewed-by: Rajesh Devaraj <rdevaraj@nvidia.com> GVS: Gerrit_Virtual_Submit <buildbot_gerritrpt@nvidia.com>
This commit is contained in:
committed by
mobile promotions
parent
3ff15b69de
commit
af48120169
@@ -1,5 +1,5 @@
|
||||
/*
|
||||
* Copyright (c) 2019-2021, NVIDIA CORPORATION. All rights reserved.
|
||||
* Copyright (c) 2019-2023, NVIDIA CORPORATION. All rights reserved.
|
||||
*
|
||||
* Permission is hereby granted, free of charge, to any person obtaining a
|
||||
* copy of this software and associated documentation files (the "Software"),
|
||||
@@ -28,7 +28,7 @@
|
||||
#include <nvgpu/gr/gr_instances.h>
|
||||
#include <nvgpu/grmgr.h>
|
||||
|
||||
static int gr_load_sm_id_config(struct gk20a *g, struct nvgpu_gr_config *config)
|
||||
int nvgpu_gr_load_sm_id_config(struct gk20a *g, struct nvgpu_gr_config *config)
|
||||
{
|
||||
int err;
|
||||
u32 *tpc_sm_id;
|
||||
@@ -130,7 +130,7 @@ int nvgpu_gr_fs_state_init(struct gk20a *g, struct nvgpu_gr_config *config)
|
||||
g->ops.gr.init.gr_load_tpc_mask(g, config);
|
||||
}
|
||||
|
||||
err = gr_load_sm_id_config(g, config);
|
||||
err = nvgpu_gr_load_sm_id_config(g, config);
|
||||
if (err != 0) {
|
||||
nvgpu_err(g, "load_smid_config failed err=%d", err);
|
||||
}
|
||||
|
||||
@@ -818,6 +818,14 @@ static int gr_init_sm_id_config_early(struct gk20a *g, struct nvgpu_gr *gr)
|
||||
}
|
||||
}
|
||||
|
||||
if (g->ops.gr.init.load_sm_id_config != NULL) {
|
||||
err = g->ops.gr.init.load_sm_id_config(g, gr->config);
|
||||
if (err != 0) {
|
||||
nvgpu_err(g, "load_sm_id_config failed err=%d", err);
|
||||
return err;
|
||||
}
|
||||
}
|
||||
|
||||
return 0;
|
||||
}
|
||||
|
||||
|
||||
@@ -935,6 +935,8 @@ struct gops_gr_init {
|
||||
int (*reset_gpcs)(struct gk20a *g);
|
||||
int (*sm_id_config_early)(struct gk20a *g,
|
||||
struct nvgpu_gr_config *config);
|
||||
int (*load_sm_id_config)(struct gk20a *g,
|
||||
struct nvgpu_gr_config *config);
|
||||
void (*set_sm_l1tag_surface_collector)(struct gk20a *g);
|
||||
/** @endcond */
|
||||
};
|
||||
|
||||
@@ -1,5 +1,5 @@
|
||||
/*
|
||||
* Copyright (c) 2019-2021, NVIDIA CORPORATION. All rights reserved.
|
||||
* Copyright (c) 2019-2023, NVIDIA CORPORATION. All rights reserved.
|
||||
*
|
||||
* Permission is hereby granted, free of charge, to any person obtaining a
|
||||
* copy of this software and associated documentation files (the "Software"),
|
||||
@@ -53,6 +53,7 @@ struct nvgpu_gr_config;
|
||||
int nvgpu_gr_fs_state_init(struct gk20a *g, struct nvgpu_gr_config *config);
|
||||
/** @cond DOXYGEN_SHOULD_SKIP_THIS */
|
||||
int nvgpu_gr_init_sm_id_early_config(struct gk20a *g, struct nvgpu_gr_config *config);
|
||||
int nvgpu_gr_load_sm_id_config(struct gk20a *g, struct nvgpu_gr_config *config);
|
||||
/** @endcond DOXYGEN_SHOULD_SKIP_THIS */
|
||||
|
||||
#endif /* NVGPU_GR_FS_STATE_H */
|
||||
|
||||
Reference in New Issue
Block a user