gpu: nvgpu: add cg and pg function

Add new power/clock gating functions that can be called by
other units.

New clock_gating functions will reside in cg.c under
common/power_features/cg unit.

New power gating functions will reside in pg.c under
common/power_features/pg unit.

Use nvgpu_pg_elpg_disable and nvgpu_pg_elpg_enable to disable/enable
elpg and also in gr_gk20a_elpg_protected macro to access gr registers.

Add cg_pg_lock to make elpg_enabled, elcg_enabled, blcg_enabled
and slcg_enabled thread safe.

JIRA NVGPU-2014

Change-Id: I00d124c2ee16242c9a3ef82e7620fbb7f1297aff
Signed-off-by: Seema Khowala <seemaj@nvidia.com>
Reviewed-on: https://git-master.nvidia.com/r/2025493
Signed-off-by: Debarshi Dutta <ddutta@nvidia.com>
(cherry-picked from c905858565 in
dev-kernel)
Reviewed-on: https://git-master.nvidia.com/r/2108406
GVS: Gerrit_Virtual_Submit
Reviewed-by: Bibek Basu <bbasu@nvidia.com>
Reviewed-by: mobile promotions <svcmobile_promotions@nvidia.com>
Tested-by: mobile promotions <svcmobile_promotions@nvidia.com>
This commit is contained in:
Debarshi Dutta
2019-04-30 13:54:08 +05:30
committed by mobile promotions
parent f495f52c70
commit c81cc032c4
24 changed files with 984 additions and 354 deletions

View File

@@ -51,8 +51,10 @@ nvgpu-y += common/bus/bus_gk20a.o \
common/mc/mc_gp10b.o \ common/mc/mc_gp10b.o \
common/mc/mc_gv11b.o \ common/mc/mc_gv11b.o \
common/mc/mc_gv100.o \ common/mc/mc_gv100.o \
common/sync/channel_sync.o common/sync/channel_sync.o \
common/power_features/power_features.o \
common/power_features/cg/cg.o \
common/power_features/pg/pg.o
# Linux specific parts of nvgpu. # Linux specific parts of nvgpu.
nvgpu-y += \ nvgpu-y += \
os/linux/os_ops.o \ os/linux/os_ops.o \

View File

@@ -116,6 +116,9 @@ srcs := os/posix/nvgpu.c \
common/mc/mc_gp10b.c \ common/mc/mc_gp10b.c \
common/mc/mc_gv11b.c \ common/mc/mc_gv11b.c \
common/mc/mc_gv100.c \ common/mc/mc_gv100.c \
common/power_features/power_features.c \
common/power_features/cg/cg.c \
common/power_features/pg/pg.c \
boardobj/boardobj.c \ boardobj/boardobj.c \
boardobj/boardobjgrp.c \ boardobj/boardobjgrp.c \
boardobj/boardobjgrpmask.c \ boardobj/boardobjgrpmask.c \

View File

@@ -30,6 +30,7 @@
#include <nvgpu/pramin.h> #include <nvgpu/pramin.h>
#include <nvgpu/enabled.h> #include <nvgpu/enabled.h>
#include <nvgpu/gk20a.h> #include <nvgpu/gk20a.h>
#include <nvgpu/power_features/cg.h>
/* /*
* Attempt to find a reserved memory area to determine PTE size for the passed * Attempt to find a reserved memory area to determine PTE size for the passed
@@ -349,22 +350,9 @@ static int nvgpu_init_mm_reset_enable_hw(struct gk20a *g)
g->ops.mc.fb_reset(g); g->ops.mc.fb_reset(g);
} }
if (g->ops.clock_gating.slcg_fb_load_gating_prod) { nvgpu_cg_slcg_fb_ltc_load_enable(g);
g->ops.clock_gating.slcg_fb_load_gating_prod(g,
g->slcg_enabled); nvgpu_cg_blcg_fb_ltc_load_enable(g);
}
if (g->ops.clock_gating.slcg_ltc_load_gating_prod) {
g->ops.clock_gating.slcg_ltc_load_gating_prod(g,
g->slcg_enabled);
}
if (g->ops.clock_gating.blcg_fb_load_gating_prod) {
g->ops.clock_gating.blcg_fb_load_gating_prod(g,
g->blcg_enabled);
}
if (g->ops.clock_gating.blcg_ltc_load_gating_prod) {
g->ops.clock_gating.blcg_ltc_load_gating_prod(g,
g->blcg_enabled);
}
if (g->ops.fb.init_fs_state) { if (g->ops.fb.init_fs_state) {
g->ops.fb.init_fs_state(g); g->ops.fb.init_fs_state(g);

View File

@@ -30,6 +30,8 @@
#include <nvgpu/bug.h> #include <nvgpu/bug.h>
#include <nvgpu/utils.h> #include <nvgpu/utils.h>
#include <nvgpu/gk20a.h> #include <nvgpu/gk20a.h>
#include <nvgpu/power_features/cg.h>
static int nvgpu_pg_init_task(void *arg); static int nvgpu_pg_init_task(void *arg);
@@ -44,15 +46,9 @@ static int pmu_enable_hw(struct nvgpu_pmu *pmu, bool enable)
/* bring PMU falcon/engine out of reset */ /* bring PMU falcon/engine out of reset */
g->ops.pmu.reset_engine(g, true); g->ops.pmu.reset_engine(g, true);
if (g->ops.clock_gating.slcg_pmu_load_gating_prod) { nvgpu_cg_slcg_pmu_load_enable(g);
g->ops.clock_gating.slcg_pmu_load_gating_prod(g,
g->slcg_enabled);
}
if (g->ops.clock_gating.blcg_pmu_load_gating_prod) { nvgpu_cg_blcg_pmu_load_enable(g);
g->ops.clock_gating.blcg_pmu_load_gating_prod(g,
g->blcg_enabled);
}
if (nvgpu_flcn_mem_scrub_wait(pmu->flcn)) { if (nvgpu_flcn_mem_scrub_wait(pmu->flcn)) {
/* keep PMU falcon/engine in reset /* keep PMU falcon/engine in reset
@@ -446,7 +442,7 @@ static void pmu_setup_hw_enable_elpg(struct gk20a *g)
g->ops.gr.pmu_save_zbc(g, 0xf); g->ops.gr.pmu_save_zbc(g, 0xf);
} }
if (g->elpg_enabled) { if (g->can_elpg && g->elpg_enabled) {
/* Init reg with prod values*/ /* Init reg with prod values*/
if (g->ops.pmu.pmu_setup_elpg) { if (g->ops.pmu.pmu_setup_elpg) {
g->ops.pmu.pmu_setup_elpg(g); g->ops.pmu.pmu_setup_elpg(g);

View File

@@ -0,0 +1,566 @@
/*
* Copyright (c) 2019, NVIDIA CORPORATION. All rights reserved.
*
* Permission is hereby granted, free of charge, to any person obtaining a
* copy of this software and associated documentation files (the "Software"),
* to deal in the Software without restriction, including without limitation
* the rights to use, copy, modify, merge, publish, distribute, sublicense,
* and/or sell copies of the Software, and to permit persons to whom the
* Software is furnished to do so, subject to the following conditions:
*
* The above copyright notice and this permission notice shall be included in
* all copies or substantial portions of the Software.
*
* THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
* IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
* FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
* THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
* LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING
* FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER
* DEALINGS IN THE SOFTWARE.
*/
#include <nvgpu/gk20a.h>
#include <nvgpu/enabled.h>
#include <nvgpu/power_features/cg.h>
static void nvgpu_cg_set_mode(struct gk20a *g, int cgmode, int mode_config)
{
u32 engine_idx;
u32 active_engine_id = 0;
struct fifo_engine_info_gk20a *engine_info = NULL;
struct fifo_gk20a *f = &g->fifo;
nvgpu_log_fn(g, " ");
for (engine_idx = 0; engine_idx < f->num_engines; ++engine_idx) {
active_engine_id = f->active_engines_list[engine_idx];
engine_info = &f->engine_info[active_engine_id];
/* gr_engine supports both BLCG and ELCG */
if ((cgmode == BLCG_MODE) && (engine_info->engine_enum ==
ENGINE_GR_GK20A)) {
g->ops.therm.init_blcg_mode(g, (u32)mode_config,
active_engine_id);
break;
} else if (cgmode == ELCG_MODE) {
g->ops.therm.init_elcg_mode(g, (u32)mode_config,
active_engine_id);
} else {
nvgpu_err(g, "invalid cg mode %d, config %d for "
"act_eng_id %d",
cgmode, mode_config, active_engine_id);
}
}
}
void nvgpu_cg_elcg_enable(struct gk20a *g)
{
nvgpu_log_fn(g, " ");
if (!nvgpu_is_enabled(g, NVGPU_GPU_CAN_ELCG)) {
return;
}
nvgpu_mutex_acquire(&g->cg_pg_lock);
if (g->elcg_enabled) {
nvgpu_cg_set_mode(g, ELCG_MODE, ELCG_AUTO);
}
nvgpu_mutex_release(&g->cg_pg_lock);
}
void nvgpu_cg_elcg_disable(struct gk20a *g)
{
nvgpu_log_fn(g, " ");
if (!nvgpu_is_enabled(g, NVGPU_GPU_CAN_ELCG)) {
return;
}
nvgpu_mutex_acquire(&g->cg_pg_lock);
if (g->elcg_enabled) {
nvgpu_cg_set_mode(g, ELCG_MODE, ELCG_RUN);
}
nvgpu_mutex_release(&g->cg_pg_lock);
}
void nvgpu_cg_blcg_mode_enable(struct gk20a *g)
{
nvgpu_log_fn(g, " ");
if (!nvgpu_is_enabled(g, NVGPU_GPU_CAN_BLCG)) {
return;
}
nvgpu_mutex_acquire(&g->cg_pg_lock);
if (g->blcg_enabled) {
nvgpu_cg_set_mode(g, BLCG_MODE, BLCG_AUTO);
}
nvgpu_mutex_release(&g->cg_pg_lock);
}
void nvgpu_cg_blcg_mode_disable(struct gk20a *g)
{
nvgpu_log_fn(g, " ");
if (!nvgpu_is_enabled(g, NVGPU_GPU_CAN_BLCG)) {
return;
}
nvgpu_mutex_acquire(&g->cg_pg_lock);
if (g->blcg_enabled) {
nvgpu_cg_set_mode(g, BLCG_MODE, BLCG_RUN);
}
nvgpu_mutex_release(&g->cg_pg_lock);
}
void nvgpu_cg_blcg_fb_ltc_load_enable(struct gk20a *g)
{
nvgpu_log_fn(g, " ");
if (!nvgpu_is_enabled(g, NVGPU_GPU_CAN_BLCG)) {
return;
}
nvgpu_mutex_acquire(&g->cg_pg_lock);
if (!g->blcg_enabled) {
goto done;
}
if (g->ops.clock_gating.blcg_fb_load_gating_prod != NULL) {
g->ops.clock_gating.blcg_fb_load_gating_prod(g, true);
}
if (g->ops.clock_gating.blcg_ltc_load_gating_prod != NULL) {
g->ops.clock_gating.blcg_ltc_load_gating_prod(g, true);
}
done:
nvgpu_mutex_release(&g->cg_pg_lock);
}
void nvgpu_cg_blcg_fifo_load_enable(struct gk20a *g)
{
nvgpu_log_fn(g, " ");
if (!nvgpu_is_enabled(g, NVGPU_GPU_CAN_BLCG)) {
return;
}
nvgpu_mutex_acquire(&g->cg_pg_lock);
if (!g->blcg_enabled) {
goto done;
}
if (g->ops.clock_gating.blcg_fifo_load_gating_prod != NULL) {
g->ops.clock_gating.blcg_fifo_load_gating_prod(g, true);
}
done:
nvgpu_mutex_release(&g->cg_pg_lock);
}
void nvgpu_cg_blcg_pmu_load_enable(struct gk20a *g)
{
nvgpu_log_fn(g, " ");
if (!nvgpu_is_enabled(g, NVGPU_GPU_CAN_BLCG)) {
return;
}
nvgpu_mutex_acquire(&g->cg_pg_lock);
if (!g->blcg_enabled) {
goto done;
}
if (g->ops.clock_gating.blcg_pmu_load_gating_prod != NULL) {
g->ops.clock_gating.blcg_pmu_load_gating_prod(g, true);
}
done:
nvgpu_mutex_release(&g->cg_pg_lock);
}
void nvgpu_cg_blcg_ce_load_enable(struct gk20a *g)
{
nvgpu_log_fn(g, " ");
if (!nvgpu_is_enabled(g, NVGPU_GPU_CAN_BLCG)) {
return;
}
nvgpu_mutex_acquire(&g->cg_pg_lock);
if (!g->blcg_enabled) {
goto done;
}
if (g->ops.clock_gating.blcg_ce_load_gating_prod != NULL) {
g->ops.clock_gating.blcg_ce_load_gating_prod(g, true);
}
done:
nvgpu_mutex_release(&g->cg_pg_lock);
}
void nvgpu_cg_blcg_gr_load_enable(struct gk20a *g)
{
nvgpu_log_fn(g, " ");
if (!nvgpu_is_enabled(g, NVGPU_GPU_CAN_BLCG)) {
return;
}
nvgpu_mutex_acquire(&g->cg_pg_lock);
if (!g->blcg_enabled) {
goto done;
}
if (g->ops.clock_gating.blcg_gr_load_gating_prod != NULL) {
g->ops.clock_gating.blcg_gr_load_gating_prod(g, true);
}
done:
nvgpu_mutex_release(&g->cg_pg_lock);
}
void nvgpu_cg_slcg_fb_ltc_load_enable(struct gk20a *g)
{
nvgpu_log_fn(g, " ");
if (!nvgpu_is_enabled(g, NVGPU_GPU_CAN_SLCG)) {
return;
}
nvgpu_mutex_acquire(&g->cg_pg_lock);
if (!g->slcg_enabled) {
goto done;
}
if (g->ops.clock_gating.slcg_fb_load_gating_prod != NULL) {
g->ops.clock_gating.slcg_fb_load_gating_prod(g, true);
}
if (g->ops.clock_gating.slcg_ltc_load_gating_prod != NULL) {
g->ops.clock_gating.slcg_ltc_load_gating_prod(g, true);
}
done:
nvgpu_mutex_release(&g->cg_pg_lock);
}
void nvgpu_cg_slcg_priring_load_enable(struct gk20a *g)
{
nvgpu_log_fn(g, " ");
if (!nvgpu_is_enabled(g, NVGPU_GPU_CAN_SLCG)) {
return;
}
nvgpu_mutex_acquire(&g->cg_pg_lock);
if (!g->slcg_enabled) {
goto done;
}
if (g->ops.clock_gating.slcg_priring_load_gating_prod != NULL) {
g->ops.clock_gating.slcg_priring_load_gating_prod(g, true);
}
done:
nvgpu_mutex_release(&g->cg_pg_lock);
}
void nvgpu_cg_slcg_gr_perf_ltc_load_enable(struct gk20a *g)
{
nvgpu_log_fn(g, " ");
if (!nvgpu_is_enabled(g, NVGPU_GPU_CAN_SLCG)) {
return;
}
nvgpu_mutex_acquire(&g->cg_pg_lock);
if (!g->slcg_enabled) {
goto done;
}
if (g->ops.clock_gating.slcg_ltc_load_gating_prod != NULL) {
g->ops.clock_gating.slcg_ltc_load_gating_prod(g, true);
}
if (g->ops.clock_gating.slcg_perf_load_gating_prod != NULL) {
g->ops.clock_gating.slcg_perf_load_gating_prod(g, true);
}
if (g->ops.clock_gating.slcg_gr_load_gating_prod != NULL) {
g->ops.clock_gating.slcg_gr_load_gating_prod(g, true);
}
done:
nvgpu_mutex_release(&g->cg_pg_lock);
}
void nvgpu_cg_slcg_gr_perf_ltc_load_disable(struct gk20a *g)
{
nvgpu_log_fn(g, " ");
if (!nvgpu_is_enabled(g, NVGPU_GPU_CAN_SLCG)) {
return;
}
nvgpu_mutex_acquire(&g->cg_pg_lock);
if (!g->slcg_enabled) {
goto done;
}
if (g->ops.clock_gating.slcg_gr_load_gating_prod != NULL) {
g->ops.clock_gating.slcg_gr_load_gating_prod(g, false);
}
if (g->ops.clock_gating.slcg_perf_load_gating_prod != NULL) {
g->ops.clock_gating.slcg_perf_load_gating_prod(g, false);
}
if (g->ops.clock_gating.slcg_ltc_load_gating_prod != NULL) {
g->ops.clock_gating.slcg_ltc_load_gating_prod(g, false);
}
done:
nvgpu_mutex_release(&g->cg_pg_lock);
}
void nvgpu_cg_slcg_fifo_load_enable(struct gk20a *g)
{
nvgpu_log_fn(g, " ");
if (!nvgpu_is_enabled(g, NVGPU_GPU_CAN_SLCG)) {
return;
}
nvgpu_mutex_acquire(&g->cg_pg_lock);
if (!g->slcg_enabled) {
goto done;
}
if (g->ops.clock_gating.slcg_fifo_load_gating_prod != NULL) {
g->ops.clock_gating.slcg_fifo_load_gating_prod(g, true);
}
done:
nvgpu_mutex_release(&g->cg_pg_lock);
}
void nvgpu_cg_slcg_pmu_load_enable(struct gk20a *g)
{
nvgpu_log_fn(g, " ");
if (!nvgpu_is_enabled(g, NVGPU_GPU_CAN_SLCG)) {
return;
}
nvgpu_mutex_acquire(&g->cg_pg_lock);
if (!g->slcg_enabled) {
goto done;
}
if (g->ops.clock_gating.slcg_pmu_load_gating_prod != NULL) {
g->ops.clock_gating.slcg_pmu_load_gating_prod(g, true);
}
done:
nvgpu_mutex_release(&g->cg_pg_lock);
}
void nvgpu_cg_slcg_ce2_load_enable(struct gk20a *g)
{
nvgpu_log_fn(g, " ");
if (!nvgpu_is_enabled(g, NVGPU_GPU_CAN_SLCG)) {
return;
}
nvgpu_mutex_acquire(&g->cg_pg_lock);
if (!g->slcg_enabled) {
goto done;
}
if (g->ops.clock_gating.slcg_ce2_load_gating_prod != NULL) {
g->ops.clock_gating.slcg_ce2_load_gating_prod(g, true);
}
done:
nvgpu_mutex_release(&g->cg_pg_lock);
}
void nvgpu_cg_init_gr_load_gating_prod(struct gk20a *g)
{
nvgpu_log_fn(g, " ");
nvgpu_mutex_acquire(&g->cg_pg_lock);
if (!nvgpu_is_enabled(g, NVGPU_GPU_CAN_SLCG)) {
goto check_can_blcg;
}
if (!g->slcg_enabled) {
goto check_can_blcg;
}
if (g->ops.clock_gating.slcg_bus_load_gating_prod != NULL) {
g->ops.clock_gating.slcg_bus_load_gating_prod(g, true);
}
if (g->ops.clock_gating.slcg_chiplet_load_gating_prod != NULL) {
g->ops.clock_gating.slcg_chiplet_load_gating_prod(g, true);
}
if (g->ops.clock_gating.slcg_gr_load_gating_prod != NULL) {
g->ops.clock_gating.slcg_gr_load_gating_prod(g, true);
}
if (g->ops.clock_gating.slcg_ctxsw_firmware_load_gating_prod != NULL) {
g->ops.clock_gating.slcg_ctxsw_firmware_load_gating_prod(g,
true);
}
if (g->ops.clock_gating.slcg_perf_load_gating_prod != NULL) {
g->ops.clock_gating.slcg_perf_load_gating_prod(g, true);
}
if (g->ops.clock_gating.slcg_xbar_load_gating_prod != NULL) {
g->ops.clock_gating.slcg_xbar_load_gating_prod(g, true);
}
check_can_blcg:
if (!nvgpu_is_enabled(g, NVGPU_GPU_CAN_BLCG)) {
goto pg_gr_load;
}
if (!g->blcg_enabled) {
goto pg_gr_load;
}
if (g->ops.clock_gating.blcg_bus_load_gating_prod != NULL) {
g->ops.clock_gating.blcg_bus_load_gating_prod(g, true);
}
if (g->ops.clock_gating.blcg_gr_load_gating_prod != NULL) {
g->ops.clock_gating.blcg_gr_load_gating_prod(g, true);
}
if (g->ops.clock_gating.blcg_ctxsw_firmware_load_gating_prod != NULL) {
g->ops.clock_gating.blcg_ctxsw_firmware_load_gating_prod(g,
true);
}
if (g->ops.clock_gating.blcg_xbar_load_gating_prod != NULL) {
g->ops.clock_gating.blcg_xbar_load_gating_prod(g, true);
}
pg_gr_load:
if (g->ops.clock_gating.pg_gr_load_gating_prod != NULL) {
g->ops.clock_gating.pg_gr_load_gating_prod(g, true);
}
nvgpu_mutex_release(&g->cg_pg_lock);
}
void nvgpu_cg_elcg_set_elcg_enabled(struct gk20a *g, bool enable)
{
nvgpu_log_fn(g, " ");
if (!nvgpu_is_enabled(g, NVGPU_GPU_CAN_ELCG)) {
return;
}
nvgpu_mutex_release(&g->cg_pg_lock);
if (enable) {
if (!g->elcg_enabled) {
g->elcg_enabled = true;
nvgpu_cg_set_mode(g, ELCG_MODE, ELCG_AUTO);
}
} else {
if (g->elcg_enabled) {
g->elcg_enabled = false;
nvgpu_cg_set_mode(g, ELCG_MODE, ELCG_RUN);
}
}
nvgpu_mutex_release(&g->cg_pg_lock);
}
void nvgpu_cg_blcg_set_blcg_enabled(struct gk20a *g, bool enable)
{
bool load = false;
nvgpu_log_fn(g, " ");
if (!nvgpu_is_enabled(g, NVGPU_GPU_CAN_BLCG)) {
return;
}
nvgpu_mutex_acquire(&g->cg_pg_lock);
if (enable) {
if (!g->blcg_enabled) {
load = true;
g->blcg_enabled = true;
}
} else {
if (g->blcg_enabled) {
load = true;
g->blcg_enabled = false;
}
}
if (!load ) {
goto done;
}
if (g->ops.clock_gating.blcg_bus_load_gating_prod != NULL) {
g->ops.clock_gating.blcg_bus_load_gating_prod(g, enable);
}
if (g->ops.clock_gating.blcg_ce_load_gating_prod != NULL) {
g->ops.clock_gating.blcg_ce_load_gating_prod(g, enable);
}
if (g->ops.clock_gating.blcg_ctxsw_firmware_load_gating_prod != NULL) {
g->ops.clock_gating.blcg_ctxsw_firmware_load_gating_prod(g,
enable);
}
if (g->ops.clock_gating.blcg_fb_load_gating_prod != NULL) {
g->ops.clock_gating.blcg_fb_load_gating_prod(g, enable);
}
if (g->ops.clock_gating.blcg_fifo_load_gating_prod != NULL) {
g->ops.clock_gating.blcg_fifo_load_gating_prod(g, enable);
}
if (g->ops.clock_gating.blcg_gr_load_gating_prod != NULL) {
g->ops.clock_gating.blcg_gr_load_gating_prod(g, enable);
}
if (g->ops.clock_gating.blcg_ltc_load_gating_prod != NULL) {
g->ops.clock_gating.blcg_ltc_load_gating_prod(g, enable);
}
if (g->ops.clock_gating.blcg_pmu_load_gating_prod != NULL) {
g->ops.clock_gating.blcg_pmu_load_gating_prod(g, enable);
}
if (g->ops.clock_gating.blcg_xbar_load_gating_prod != NULL) {
g->ops.clock_gating.blcg_xbar_load_gating_prod(g, enable);
}
done:
nvgpu_mutex_release(&g->cg_pg_lock);
}
void nvgpu_cg_slcg_set_slcg_enabled(struct gk20a *g, bool enable)
{
bool load = false;
nvgpu_log_fn(g, " ");
if (!nvgpu_is_enabled(g, NVGPU_GPU_CAN_SLCG)) {
return;
}
nvgpu_mutex_acquire(&g->cg_pg_lock);
if (enable) {
if (!g->slcg_enabled) {
load = true;
g->slcg_enabled = true;
}
} else {
if (g->slcg_enabled) {
load = true;
g->slcg_enabled = false;
}
}
if (!load ) {
goto done;
}
if (g->ops.clock_gating.slcg_bus_load_gating_prod != NULL) {
g->ops.clock_gating.slcg_bus_load_gating_prod(g, enable);
}
if (g->ops.clock_gating.slcg_ce2_load_gating_prod != NULL) {
g->ops.clock_gating.slcg_ce2_load_gating_prod(g, enable);
}
if (g->ops.clock_gating.slcg_chiplet_load_gating_prod != NULL) {
g->ops.clock_gating.slcg_chiplet_load_gating_prod(g, enable);
}
if (g->ops.clock_gating.slcg_ctxsw_firmware_load_gating_prod !=
NULL) {
g->ops.clock_gating.slcg_ctxsw_firmware_load_gating_prod(g,
enable);
}
if (g->ops.clock_gating.slcg_fb_load_gating_prod != NULL) {
g->ops.clock_gating.slcg_fb_load_gating_prod(g, enable);
}
if (g->ops.clock_gating.slcg_fifo_load_gating_prod != NULL) {
g->ops.clock_gating.slcg_fifo_load_gating_prod(g, enable);
}
if (g->ops.clock_gating.slcg_gr_load_gating_prod != NULL) {
g->ops.clock_gating.slcg_gr_load_gating_prod(g, enable);
}
if (g->ops.clock_gating.slcg_ltc_load_gating_prod != NULL) {
g->ops.clock_gating.slcg_ltc_load_gating_prod(g, enable);
}
if (g->ops.clock_gating.slcg_perf_load_gating_prod != NULL) {
g->ops.clock_gating.slcg_perf_load_gating_prod(g, enable);
}
if (g->ops.clock_gating.slcg_priring_load_gating_prod != NULL) {
g->ops.clock_gating.slcg_priring_load_gating_prod(g,
enable);
}
if (g->ops.clock_gating.slcg_pmu_load_gating_prod != NULL) {
g->ops.clock_gating.slcg_pmu_load_gating_prod(g, enable);
}
if (g->ops.clock_gating.slcg_xbar_load_gating_prod != NULL) {
g->ops.clock_gating.slcg_xbar_load_gating_prod(g, enable);
}
done:
nvgpu_mutex_release(&g->cg_pg_lock);
}

View File

@@ -0,0 +1,106 @@
/*
* Copyright (c) 2019, NVIDIA CORPORATION. All rights reserved.
*
* Permission is hereby granted, free of charge, to any person obtaining a
* copy of this software and associated documentation files (the "Software"),
* to deal in the Software without restriction, including without limitation
* the rights to use, copy, modify, merge, publish, distribute, sublicense,
* and/or sell copies of the Software, and to permit persons to whom the
* Software is furnished to do so, subject to the following conditions:
*
* The above copyright notice and this permission notice shall be included in
* all copies or substantial portions of the Software.
*
* THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
* IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
* FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
* THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
* LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING
* FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER
* DEALINGS IN THE SOFTWARE.
*/
#include <nvgpu/gk20a.h>
#include <nvgpu/pmu.h>
#include <nvgpu/power_features/pg.h>
bool nvgpu_pg_elpg_is_enabled(struct gk20a *g)
{
bool elpg_enabled;
nvgpu_log_fn(g, " ");
nvgpu_mutex_acquire(&g->cg_pg_lock);
elpg_enabled = g->elpg_enabled;
nvgpu_mutex_release(&g->cg_pg_lock);
return elpg_enabled;
}
int nvgpu_pg_elpg_enable(struct gk20a *g)
{
int err = 0;
nvgpu_log_fn(g, " ");
if (!g->can_elpg) {
return 0;
}
nvgpu_mutex_acquire(&g->cg_pg_lock);
if (g->elpg_enabled) {
err = nvgpu_pmu_pg_global_enable(g, true);
}
nvgpu_mutex_release(&g->cg_pg_lock);
return err;
}
int nvgpu_pg_elpg_disable(struct gk20a *g)
{
int err = 0;
nvgpu_log_fn(g, " ");
if (!g->can_elpg) {
return 0;
}
nvgpu_mutex_acquire(&g->cg_pg_lock);
if (g->elpg_enabled) {
err = nvgpu_pmu_pg_global_enable(g, false);
}
nvgpu_mutex_release(&g->cg_pg_lock);
return err;
}
int nvgpu_pg_elpg_set_elpg_enabled(struct gk20a *g, bool enable)
{
int err = 0;
bool change_mode = false;
nvgpu_log_fn(g, " ");
if (!g->can_elpg) {
return 0;
}
nvgpu_mutex_acquire(&g->cg_pg_lock);
if (enable) {
if (!g->elpg_enabled) {
change_mode = true;
g->elpg_enabled = true;
}
} else {
if (g->elpg_enabled) {
change_mode = true;
g->elpg_enabled = false;
}
}
if (!change_mode) {
goto done;
}
err = nvgpu_pmu_pg_global_enable(g, enable);
done:
nvgpu_mutex_release(&g->cg_pg_lock);
return err;
}

View File

@@ -0,0 +1,66 @@
/*
* Copyright (c) 2019, NVIDIA CORPORATION. All rights reserved.
*
* Permission is hereby granted, free of charge, to any person obtaining a
* copy of this software and associated documentation files (the "Software"),
* to deal in the Software without restriction, including without limitation
* the rights to use, copy, modify, merge, publish, distribute, sublicense,
* and/or sell copies of the Software, and to permit persons to whom the
* Software is furnished to do so, subject to the following conditions:
*
* The above copyright notice and this permission notice shall be included in
* all copies or substantial portions of the Software.
*
* THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
* IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
* FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
* THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
* LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING
* FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER
* DEALINGS IN THE SOFTWARE.
*/
#include <nvgpu/gk20a.h>
#include <nvgpu/power_features/cg.h>
#include <nvgpu/power_features/pg.h>
#include <nvgpu/power_features/power_features.h>
int nvgpu_cg_pg_disable(struct gk20a *g)
{
int err = 0;
nvgpu_log_fn(g, " ");
/* disable elpg before clock gating */
err = nvgpu_pg_elpg_disable(g);
if (err != 0) {
nvgpu_err(g, "failed to set disable elpg");
}
nvgpu_cg_slcg_gr_perf_ltc_load_disable(g);
nvgpu_cg_blcg_mode_disable(g);
nvgpu_cg_elcg_disable(g);
return err;
}
int nvgpu_cg_pg_enable(struct gk20a *g)
{
int err = 0;
nvgpu_log_fn(g, " ");
nvgpu_cg_elcg_enable(g);
nvgpu_cg_blcg_mode_enable(g);
nvgpu_cg_slcg_gr_perf_ltc_load_enable(g);
err = nvgpu_pg_elpg_enable(g);
if (err != 0) {
nvgpu_err(g, "failed to set enable elpg");
}
return err;
}

View File

@@ -26,6 +26,7 @@
#include <nvgpu/enabled.h> #include <nvgpu/enabled.h>
#include <nvgpu/io.h> #include <nvgpu/io.h>
#include <nvgpu/utils.h> #include <nvgpu/utils.h>
#include <nvgpu/power_features/cg.h>
#include "priv_ring_gm20b.h" #include "priv_ring_gm20b.h"
@@ -41,10 +42,7 @@ void gm20b_priv_ring_enable(struct gk20a *g)
nvgpu_log(g, gpu_dbg_info, "enabling priv ring"); nvgpu_log(g, gpu_dbg_info, "enabling priv ring");
if (g->ops.clock_gating.slcg_priring_load_gating_prod) { nvgpu_cg_slcg_priring_load_enable(g);
g->ops.clock_gating.slcg_priring_load_gating_prod(g,
g->slcg_enabled);
}
gk20a_writel(g,pri_ringmaster_command_r(), gk20a_writel(g,pri_ringmaster_command_r(),
0x4); 0x4);

View File

@@ -30,6 +30,7 @@
#include <nvgpu/io.h> #include <nvgpu/io.h>
#include <nvgpu/utils.h> #include <nvgpu/utils.h>
#include <nvgpu/channel.h> #include <nvgpu/channel.h>
#include <nvgpu/power_features/cg.h>
#include "gk20a.h" #include "gk20a.h"
#include "gk20a/fence_gk20a.h" #include "gk20a/fence_gk20a.h"
@@ -339,14 +340,9 @@ int gk20a_init_ce_support(struct gk20a *g)
g->ops.mc.reset(g, ce_reset_mask); g->ops.mc.reset(g, ce_reset_mask);
if (g->ops.clock_gating.slcg_ce2_load_gating_prod) { nvgpu_cg_slcg_ce2_load_enable(g);
g->ops.clock_gating.slcg_ce2_load_gating_prod(g,
g->slcg_enabled); nvgpu_cg_blcg_ce_load_enable(g);
}
if (g->ops.clock_gating.blcg_ce_load_gating_prod) {
g->ops.clock_gating.blcg_ce_load_gating_prod(g,
g->blcg_enabled);
}
if (ce_app->initialised) { if (ce_app->initialised) {
/* assume this happen during poweron/poweroff GPU sequence */ /* assume this happen during poweron/poweroff GPU sequence */

View File

@@ -1,7 +1,7 @@
/* /*
* Tegra GK20A GPU Debugger/Profiler Driver * Tegra GK20A GPU Debugger/Profiler Driver
* *
* Copyright (c) 2013-2018, NVIDIA CORPORATION. All rights reserved. * Copyright (c) 2013-2019, NVIDIA CORPORATION. All rights reserved.
* *
* Permission is hereby granted, free of charge, to any person obtaining a * Permission is hereby granted, free of charge, to any person obtaining a
* copy of this software and associated documentation files (the "Software"), * copy of this software and associated documentation files (the "Software"),
@@ -32,6 +32,7 @@
#include <nvgpu/utils.h> #include <nvgpu/utils.h>
#include <nvgpu/channel.h> #include <nvgpu/channel.h>
#include <nvgpu/unit.h> #include <nvgpu/unit.h>
#include <nvgpu/power_features/power_features.h>
#include "gk20a.h" #include "gk20a.h"
#include "gr_gk20a.h" #include "gr_gk20a.h"
@@ -234,60 +235,28 @@ int dbg_set_powergate(struct dbg_session_gk20a *dbg_s, bool disable_powergate)
return err; return err;
} }
/*do elpg disable before clock gating */ err = nvgpu_cg_pg_disable(g);
nvgpu_pmu_pg_global_enable(g, false);
if (g->ops.clock_gating.slcg_gr_load_gating_prod) { if (err == 0) {
g->ops.clock_gating.slcg_gr_load_gating_prod(g, dbg_s->is_pg_disabled = true;
false); nvgpu_log(g, gpu_dbg_gpu_dbg | gpu_dbg_fn,
"pg disabled");
} }
if (g->ops.clock_gating.slcg_perf_load_gating_prod) {
g->ops.clock_gating.slcg_perf_load_gating_prod(g,
false);
}
if (g->ops.clock_gating.slcg_ltc_load_gating_prod) {
g->ops.clock_gating.slcg_ltc_load_gating_prod(g,
false);
}
gr_gk20a_init_cg_mode(g, BLCG_MODE, BLCG_RUN);
gr_gk20a_init_cg_mode(g, ELCG_MODE, ELCG_RUN);
dbg_s->is_pg_disabled = true;
} else { } else {
/* restore (can) powergate, clk state */ /* restore (can) powergate, clk state */
/* release pending exceptions to fault/be handled as usual */ /* release pending exceptions to fault/be handled as usual */
/*TBD: ordering of these? */ /*TBD: ordering of these? */
if (g->elcg_enabled) { err = nvgpu_cg_pg_enable(g);
gr_gk20a_init_cg_mode(g, ELCG_MODE, ELCG_AUTO);
}
if (g->blcg_enabled) { nvgpu_log(g, gpu_dbg_gpu_dbg | gpu_dbg_fn, "module idle");
gr_gk20a_init_cg_mode(g, BLCG_MODE, BLCG_AUTO);
}
if (g->slcg_enabled) {
if (g->ops.clock_gating.slcg_ltc_load_gating_prod) {
g->ops.clock_gating.slcg_ltc_load_gating_prod(g,
g->slcg_enabled);
}
if (g->ops.clock_gating.slcg_perf_load_gating_prod) {
g->ops.clock_gating.slcg_perf_load_gating_prod(g,
g->slcg_enabled);
}
if (g->ops.clock_gating.slcg_gr_load_gating_prod) {
g->ops.clock_gating.slcg_gr_load_gating_prod(g,
g->slcg_enabled);
}
}
nvgpu_pmu_pg_global_enable(g, true);
nvgpu_log(g, gpu_dbg_gpu_dbg | gpu_dbg_fn,
"module idle");
gk20a_idle(g); gk20a_idle(g);
dbg_s->is_pg_disabled = false; if (err == 0) {
dbg_s->is_pg_disabled = false;
nvgpu_log(g, gpu_dbg_gpu_dbg | gpu_dbg_fn,
"pg enabled");
}
} }
nvgpu_log(g, gpu_dbg_fn|gpu_dbg_gpu_dbg, "%s powergate mode = %s done", nvgpu_log(g, gpu_dbg_fn|gpu_dbg_gpu_dbg, "%s powergate mode = %s done",

View File

@@ -45,6 +45,8 @@
#include <nvgpu/utils.h> #include <nvgpu/utils.h>
#include <nvgpu/channel.h> #include <nvgpu/channel.h>
#include <nvgpu/unit.h> #include <nvgpu/unit.h>
#include <nvgpu/power_features/power_features.h>
#include <nvgpu/power_features/cg.h>
#include "gk20a.h" #include "gk20a.h"
#include "mm_gk20a.h" #include "mm_gk20a.h"
@@ -824,14 +826,9 @@ int gk20a_init_fifo_reset_enable_hw(struct gk20a *g)
/* enable pmc pfifo */ /* enable pmc pfifo */
g->ops.mc.reset(g, g->ops.mc.reset_mask(g, NVGPU_UNIT_FIFO)); g->ops.mc.reset(g, g->ops.mc.reset_mask(g, NVGPU_UNIT_FIFO));
if (g->ops.clock_gating.slcg_fifo_load_gating_prod) { nvgpu_cg_slcg_fifo_load_enable(g);
g->ops.clock_gating.slcg_fifo_load_gating_prod(g,
g->slcg_enabled); nvgpu_cg_blcg_fifo_load_enable(g);
}
if (g->ops.clock_gating.blcg_fifo_load_gating_prod) {
g->ops.clock_gating.blcg_fifo_load_gating_prod(g,
g->blcg_enabled);
}
timeout = gk20a_readl(g, fifo_fb_timeout_r()); timeout = gk20a_readl(g, fifo_fb_timeout_r());
timeout = set_field(timeout, fifo_fb_timeout_period_m(), timeout = set_field(timeout, fifo_fb_timeout_period_m(),
@@ -1361,8 +1358,8 @@ void gk20a_fifo_reset_engine(struct gk20a *g, u32 engine_id)
} }
if (engine_enum == ENGINE_GR_GK20A) { if (engine_enum == ENGINE_GR_GK20A) {
if (g->support_pmu && g->can_elpg) { if (g->support_pmu) {
if (nvgpu_pmu_disable_elpg(g)) { if (nvgpu_pg_elpg_disable(g) != 0 ) {
nvgpu_err(g, "failed to set disable elpg"); nvgpu_err(g, "failed to set disable elpg");
} }
} }
@@ -1391,8 +1388,10 @@ void gk20a_fifo_reset_engine(struct gk20a *g, u32 engine_id)
"HALT gr pipe not supported and " "HALT gr pipe not supported and "
"gr cannot be reset without halting gr pipe"); "gr cannot be reset without halting gr pipe");
} }
if (g->support_pmu && g->can_elpg) { if (g->support_pmu) {
nvgpu_pmu_enable_elpg(g); if (nvgpu_pg_elpg_enable(g) != 0 ) {
nvgpu_err(g, "failed to set enable elpg");
}
} }
} }
if ((engine_enum == ENGINE_GRCE_GK20A) || if ((engine_enum == ENGINE_GRCE_GK20A) ||
@@ -1638,25 +1637,11 @@ static bool gk20a_fifo_handle_mmu_fault_locked(
g->fifo.deferred_reset_pending = false; g->fifo.deferred_reset_pending = false;
/* Disable power management */ /* Disable power management */
if (g->support_pmu && g->can_elpg) { if (g->support_pmu) {
if (nvgpu_pmu_disable_elpg(g)) { if (nvgpu_cg_pg_disable(g) != 0) {
nvgpu_err(g, "failed to set disable elpg"); nvgpu_warn(g, "fail to disable power mgmt");
} }
} }
if (g->ops.clock_gating.slcg_gr_load_gating_prod) {
g->ops.clock_gating.slcg_gr_load_gating_prod(g,
false);
}
if (g->ops.clock_gating.slcg_perf_load_gating_prod) {
g->ops.clock_gating.slcg_perf_load_gating_prod(g,
false);
}
if (g->ops.clock_gating.slcg_ltc_load_gating_prod) {
g->ops.clock_gating.slcg_ltc_load_gating_prod(g,
false);
}
gr_gk20a_init_cg_mode(g, ELCG_MODE, ELCG_RUN);
/* Disable fifo access */ /* Disable fifo access */
grfifo_ctl = gk20a_readl(g, gr_gpfifo_ctl_r()); grfifo_ctl = gk20a_readl(g, gr_gpfifo_ctl_r());
@@ -1842,8 +1827,10 @@ static bool gk20a_fifo_handle_mmu_fault_locked(
gr_gpfifo_ctl_semaphore_access_enabled_f()); gr_gpfifo_ctl_semaphore_access_enabled_f());
/* It is safe to enable ELPG again. */ /* It is safe to enable ELPG again. */
if (g->support_pmu && g->can_elpg) { if (g->support_pmu) {
nvgpu_pmu_enable_elpg(g); if (nvgpu_cg_pg_enable(g) != 0) {
nvgpu_warn(g, "fail to enable power mgmt");
}
} }
return verbose; return verbose;

View File

@@ -43,6 +43,8 @@
#include <nvgpu/utils.h> #include <nvgpu/utils.h>
#include <nvgpu/channel.h> #include <nvgpu/channel.h>
#include <nvgpu/unit.h> #include <nvgpu/unit.h>
#include <nvgpu/power_features/pg.h>
#include <nvgpu/power_features/cg.h>
#include "gk20a.h" #include "gk20a.h"
#include "gr_gk20a.h" #include "gr_gk20a.h"
@@ -91,8 +93,6 @@ static void gr_gk20a_free_channel_patch_ctx(struct gk20a *g,
/* golden ctx image */ /* golden ctx image */
static int gr_gk20a_init_golden_ctx_image(struct gk20a *g, static int gr_gk20a_init_golden_ctx_image(struct gk20a *g,
struct channel_gk20a *c); struct channel_gk20a *c);
/*elcg init */
static void gr_gk20a_enable_elcg(struct gk20a *g);
int gr_gk20a_get_ctx_id(struct gk20a *g, int gr_gk20a_get_ctx_id(struct gk20a *g,
struct channel_gk20a *c, struct channel_gk20a *c,
@@ -4227,33 +4227,6 @@ int gk20a_gr_zbc_set_table(struct gk20a *g, struct gr_gk20a *gr,
gr_gk20a_add_zbc(g, gr, zbc_val)); gr_gk20a_add_zbc(g, gr, zbc_val));
} }
void gr_gk20a_init_cg_mode(struct gk20a *g, u32 cgmode, u32 mode_config)
{
u32 engine_idx;
u32 active_engine_id = 0;
struct fifo_engine_info_gk20a *engine_info = NULL;
struct fifo_gk20a *f = &g->fifo;
for (engine_idx = 0; engine_idx < f->num_engines; ++engine_idx) {
active_engine_id = f->active_engines_list[engine_idx];
engine_info = &f->engine_info[active_engine_id];
/* gr_engine supports both BLCG and ELCG */
if ((cgmode == BLCG_MODE) &&
(engine_info->engine_enum == ENGINE_GR_GK20A)) {
g->ops.therm.init_blcg_mode(g, mode_config, active_engine_id);
break;
} else if (cgmode == ELCG_MODE) {
g->ops.therm.init_elcg_mode(g, mode_config,
active_engine_id);
} else {
nvgpu_err(g, "invalid cg mode %d, config %d for "
"act_eng_id %d",
cgmode, mode_config, active_engine_id);
}
}
}
void gr_gk20a_program_zcull_mapping(struct gk20a *g, u32 zcull_num_entries, void gr_gk20a_program_zcull_mapping(struct gk20a *g, u32 zcull_num_entries,
u32 *zcull_map_tiles) u32 *zcull_map_tiles)
{ {
@@ -4655,60 +4628,6 @@ out:
return err; return err;
} }
static void gr_gk20a_load_gating_prod(struct gk20a *g)
{
nvgpu_log_fn(g, " ");
/* slcg prod values */
if (g->ops.clock_gating.slcg_bus_load_gating_prod) {
g->ops.clock_gating.slcg_bus_load_gating_prod(g,
g->slcg_enabled);
}
if (g->ops.clock_gating.slcg_chiplet_load_gating_prod) {
g->ops.clock_gating.slcg_chiplet_load_gating_prod(g,
g->slcg_enabled);
}
if (g->ops.clock_gating.slcg_gr_load_gating_prod) {
g->ops.clock_gating.slcg_gr_load_gating_prod(g,
g->slcg_enabled);
}
if (g->ops.clock_gating.slcg_ctxsw_firmware_load_gating_prod) {
g->ops.clock_gating.slcg_ctxsw_firmware_load_gating_prod(g,
g->slcg_enabled);
}
if (g->ops.clock_gating.slcg_perf_load_gating_prod) {
g->ops.clock_gating.slcg_perf_load_gating_prod(g,
g->slcg_enabled);
}
if (g->ops.clock_gating.slcg_xbar_load_gating_prod) {
g->ops.clock_gating.slcg_xbar_load_gating_prod(g,
g->slcg_enabled);
}
/* blcg prod values */
if (g->ops.clock_gating.blcg_bus_load_gating_prod) {
g->ops.clock_gating.blcg_bus_load_gating_prod(g,
g->blcg_enabled);
}
if (g->ops.clock_gating.blcg_gr_load_gating_prod) {
g->ops.clock_gating.blcg_gr_load_gating_prod(g,
g->blcg_enabled);
}
if (g->ops.clock_gating.blcg_ctxsw_firmware_load_gating_prod) {
g->ops.clock_gating.blcg_ctxsw_firmware_load_gating_prod(g,
g->blcg_enabled);
}
if (g->ops.clock_gating.blcg_xbar_load_gating_prod) {
g->ops.clock_gating.blcg_xbar_load_gating_prod(g,
g->blcg_enabled);
}
if (g->ops.clock_gating.pg_gr_load_gating_prod) {
g->ops.clock_gating.pg_gr_load_gating_prod(g, true);
}
nvgpu_log_fn(g, "done");
}
static int gk20a_init_gr_prepare(struct gk20a *g) static int gk20a_init_gr_prepare(struct gk20a *g)
{ {
u32 err = 0; u32 err = 0;
@@ -4718,10 +4637,10 @@ static int gk20a_init_gr_prepare(struct gk20a *g)
g->ops.mc.reset_mask(g, NVGPU_UNIT_BLG) | g->ops.mc.reset_mask(g, NVGPU_UNIT_BLG) |
g->ops.mc.reset_mask(g, NVGPU_UNIT_PERFMON)); g->ops.mc.reset_mask(g, NVGPU_UNIT_PERFMON));
gr_gk20a_load_gating_prod(g); nvgpu_cg_init_gr_load_gating_prod(g);
/* Disable elcg until it gets enabled later in the init*/ /* Disable elcg until it gets enabled later in the init*/
gr_gk20a_init_cg_mode(g, ELCG_MODE, ELCG_RUN); nvgpu_cg_elcg_disable(g);
/* enable fifo access */ /* enable fifo access */
gk20a_writel(g, gr_gpfifo_ctl_r(), gk20a_writel(g, gr_gpfifo_ctl_r(),
@@ -5041,7 +4960,7 @@ int gk20a_init_gr_support(struct gk20a *g)
} }
} }
gr_gk20a_enable_elcg(g); nvgpu_cg_elcg_enable(g);
/* GR is inialized, signal possible waiters */ /* GR is inialized, signal possible waiters */
g->gr.initialized = true; g->gr.initialized = true;
nvgpu_cond_signal(&g->gr.init_wq); nvgpu_cond_signal(&g->gr.init_wq);
@@ -5128,15 +5047,6 @@ int gk20a_enable_gr_hw(struct gk20a *g)
return 0; return 0;
} }
static void gr_gk20a_enable_elcg(struct gk20a *g)
{
if (g->elcg_enabled) {
gr_gk20a_init_cg_mode(g, ELCG_MODE, ELCG_AUTO);
} else {
gr_gk20a_init_cg_mode(g, ELCG_MODE, ELCG_RUN);
}
}
int gk20a_gr_reset(struct gk20a *g) int gk20a_gr_reset(struct gk20a *g)
{ {
int err; int err;
@@ -5193,8 +5103,8 @@ int gk20a_gr_reset(struct gk20a *g)
return err; return err;
} }
gr_gk20a_load_gating_prod(g); nvgpu_cg_init_gr_load_gating_prod(g);
gr_gk20a_enable_elcg(g); nvgpu_cg_elcg_enable(g);
return err; return err;
} }

View File

@@ -28,6 +28,7 @@
#include "gr_ctx_gk20a.h" #include "gr_ctx_gk20a.h"
#include "mm_gk20a.h" #include "mm_gk20a.h"
#include <nvgpu/power_features/pg.h>
#include <nvgpu/comptags.h> #include <nvgpu/comptags.h>
#include <nvgpu/cond.h> #include <nvgpu/cond.h>
@@ -598,16 +599,16 @@ u32 gk20a_gr_get_sm_no_lock_down_hww_global_esr_mask(struct gk20a *g);
#define gr_gk20a_elpg_protected_call(g, func) \ #define gr_gk20a_elpg_protected_call(g, func) \
({ \ ({ \
int err = 0; \ int err = 0; \
if ((g->support_pmu) && (g->elpg_enabled)) {\ if (g->support_pmu) {\
err = nvgpu_pmu_disable_elpg(g); \ err = nvgpu_pg_elpg_disable(g);\
if (err != 0) {\ if (err != 0) {\
nvgpu_pmu_enable_elpg(g); \ err = nvgpu_pg_elpg_enable(g); \
} \ } \
} \ } \
if (err == 0) { \ if (err == 0) { \
err = func; \ err = func; \
if ((g->support_pmu) && (g->elpg_enabled)) {\ if (g->support_pmu) {\
nvgpu_pmu_enable_elpg(g); \ (void)nvgpu_pg_elpg_enable(g); \
} \ } \
} \ } \
err; \ err; \

View File

@@ -106,7 +106,7 @@ int gm20b_pmu_setup_elpg(struct gk20a *g)
nvgpu_log_fn(g, " "); nvgpu_log_fn(g, " ");
if (g->elpg_enabled) { if (g->can_elpg && g->elpg_enabled) {
reg_writes = ((sizeof(_pginitseq_gm20b) / reg_writes = ((sizeof(_pginitseq_gm20b) /
sizeof((_pginitseq_gm20b)[0]))); sizeof((_pginitseq_gm20b)[0])));
/* Initialize registers with production values*/ /* Initialize registers with production values*/

View File

@@ -282,7 +282,7 @@ int gp10b_pmu_setup_elpg(struct gk20a *g)
nvgpu_log_fn(g, " "); nvgpu_log_fn(g, " ");
if (g->elpg_enabled) { if (g->can_elpg && g->elpg_enabled) {
reg_writes = ((sizeof(_pginitseq_gp10b) / reg_writes = ((sizeof(_pginitseq_gp10b) /
sizeof((_pginitseq_gp10b)[0]))); sizeof((_pginitseq_gp10b)[0])));
/* Initialize registers with production values*/ /* Initialize registers with production values*/

View File

@@ -43,6 +43,8 @@
#include <nvgpu/gk20a.h> #include <nvgpu/gk20a.h>
#include <nvgpu/channel.h> #include <nvgpu/channel.h>
#include <nvgpu/unit.h> #include <nvgpu/unit.h>
#include <nvgpu/power_features/cg.h>
#include <nvgpu/power_features/power_features.h>
#include "gk20a/fifo_gk20a.h" #include "gk20a/fifo_gk20a.h"
@@ -1095,25 +1097,11 @@ void gv11b_fifo_teardown_ch_tsg(struct gk20a *g, u32 act_eng_bitmask,
g->fifo.deferred_reset_pending = false; g->fifo.deferred_reset_pending = false;
/* Disable power management */ /* Disable power management */
if (g->support_pmu && g->elpg_enabled) { if (g->support_pmu) {
if (nvgpu_pmu_disable_elpg(g)) { if (nvgpu_cg_pg_disable(g) != 0) {
nvgpu_err(g, "failed to set disable elpg"); nvgpu_warn(g, "fail to disable power mgmt");
} }
} }
if (g->ops.clock_gating.slcg_gr_load_gating_prod) {
g->ops.clock_gating.slcg_gr_load_gating_prod(g,
false);
}
if (g->ops.clock_gating.slcg_perf_load_gating_prod) {
g->ops.clock_gating.slcg_perf_load_gating_prod(g,
false);
}
if (g->ops.clock_gating.slcg_ltc_load_gating_prod) {
g->ops.clock_gating.slcg_ltc_load_gating_prod(g,
false);
}
gr_gk20a_init_cg_mode(g, ELCG_MODE, ELCG_RUN);
if (rc_type == RC_TYPE_MMU_FAULT) { if (rc_type == RC_TYPE_MMU_FAULT) {
gk20a_debug_dump(g); gk20a_debug_dump(g);
@@ -1220,8 +1208,10 @@ void gv11b_fifo_teardown_ch_tsg(struct gk20a *g, u32 act_eng_bitmask,
gk20a_fifo_set_runlist_state(g, runlists_mask, RUNLIST_ENABLED); gk20a_fifo_set_runlist_state(g, runlists_mask, RUNLIST_ENABLED);
/* It is safe to enable ELPG again. */ /* It is safe to enable ELPG again. */
if (g->support_pmu && g->elpg_enabled) { if (g->support_pmu) {
nvgpu_pmu_enable_elpg(g); if (nvgpu_cg_pg_enable(g) != 0) {
nvgpu_warn(g, "fail to enable power mgmt");
}
} }
g->ops.fifo.teardown_unmask_intr(g); g->ops.fifo.teardown_unmask_intr(g);
@@ -1312,18 +1302,11 @@ int gv11b_init_fifo_reset_enable_hw(struct gk20a *g)
/* enable pmc pfifo */ /* enable pmc pfifo */
g->ops.mc.reset(g, g->ops.mc.reset_mask(g, NVGPU_UNIT_FIFO)); g->ops.mc.reset(g, g->ops.mc.reset_mask(g, NVGPU_UNIT_FIFO));
if (g->ops.clock_gating.slcg_ce2_load_gating_prod) { nvgpu_cg_slcg_ce2_load_enable(g);
g->ops.clock_gating.slcg_ce2_load_gating_prod(g,
g->slcg_enabled); nvgpu_cg_slcg_fifo_load_enable(g);
}
if (g->ops.clock_gating.slcg_fifo_load_gating_prod) { nvgpu_cg_blcg_fifo_load_enable(g);
g->ops.clock_gating.slcg_fifo_load_gating_prod(g,
g->slcg_enabled);
}
if (g->ops.clock_gating.blcg_fifo_load_gating_prod) {
g->ops.clock_gating.blcg_fifo_load_gating_prod(g,
g->blcg_enabled);
}
timeout = gk20a_readl(g, fifo_fb_timeout_r()); timeout = gk20a_readl(g, fifo_fb_timeout_r());
nvgpu_log_info(g, "fifo_fb_timeout reg val = 0x%08x", timeout); nvgpu_log_info(g, "fifo_fb_timeout reg val = 0x%08x", timeout);

View File

@@ -124,7 +124,7 @@ int gv11b_pmu_setup_elpg(struct gk20a *g)
nvgpu_log_fn(g, " "); nvgpu_log_fn(g, " ");
if (g->elpg_enabled) { if (g->can_elpg && g->elpg_enabled) {
reg_writes = ((sizeof(_pginitseq_gv11b) / reg_writes = ((sizeof(_pginitseq_gv11b) /
sizeof((_pginitseq_gv11b)[0]))); sizeof((_pginitseq_gv11b)[0])));
/* Initialize registers with production values*/ /* Initialize registers with production values*/

View File

@@ -1490,6 +1490,7 @@ struct gk20a {
u32 max_timeslice_us; u32 max_timeslice_us;
bool runlist_interleave; bool runlist_interleave;
struct nvgpu_mutex cg_pg_lock;
bool slcg_enabled; bool slcg_enabled;
bool blcg_enabled; bool blcg_enabled;
bool elcg_enabled; bool elcg_enabled;

View File

@@ -0,0 +1,55 @@
/*
* Copyright (c) 2019, NVIDIA CORPORATION. All rights reserved.
*
* Permission is hereby granted, free of charge, to any person obtaining a
* copy of this software and associated documentation files (the "Software"),
* to deal in the Software without restriction, including without limitation
* the rights to use, copy, modify, merge, publish, distribute, sublicense,
* and/or sell copies of the Software, and to permit persons to whom the
* Software is furnished to do so, subject to the following conditions:
*
* The above copyright notice and this permission notice shall be included in
* all copies or substantial portions of the Software.
*
* THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
* IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
* FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
* THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
* LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING
* FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER
* DEALINGS IN THE SOFTWARE.
*/
#ifndef NVGPU_POWER_FEATURES_CG_H
#define NVGPU_POWER_FEATURES_CG_H
#include <nvgpu/types.h>
struct gk20a;
struct fifo_gk20a;
void nvgpu_cg_init_gr_load_gating_prod(struct gk20a *g);
void nvgpu_cg_elcg_enable(struct gk20a *g);
void nvgpu_cg_elcg_disable(struct gk20a *g);
void nvgpu_cg_elcg_set_elcg_enabled(struct gk20a *g, bool enable);
void nvgpu_cg_blcg_mode_enable(struct gk20a *g);
void nvgpu_cg_blcg_mode_disable(struct gk20a *g);
void nvgpu_cg_blcg_fb_ltc_load_enable(struct gk20a *g);
void nvgpu_cg_blcg_fifo_load_enable(struct gk20a *g);
void nvgpu_cg_blcg_pmu_load_enable(struct gk20a *g);
void nvgpu_cg_blcg_ce_load_enable(struct gk20a *g);
void nvgpu_cg_blcg_gr_load_enable(struct gk20a *g);
void nvgpu_cg_blcg_set_blcg_enabled(struct gk20a *g, bool enable);
void nvgpu_cg_slcg_gr_perf_ltc_load_enable(struct gk20a *g);
void nvgpu_cg_slcg_gr_perf_ltc_load_disable(struct gk20a *g);
void nvgpu_cg_slcg_fb_ltc_load_enable(struct gk20a *g);
void nvgpu_cg_slcg_priring_load_enable(struct gk20a *g);
void nvgpu_cg_slcg_fifo_load_enable(struct gk20a *g);
void nvgpu_cg_slcg_pmu_load_enable(struct gk20a *g);
void nvgpu_cg_slcg_ce2_load_enable(struct gk20a *g);
void nvgpu_cg_slcg_set_slcg_enabled(struct gk20a *g, bool enable);
#endif /*NVGPU_POWER_FEATURES_CG_H*/

View File

@@ -0,0 +1,36 @@
/*
* Copyright (c) 2019, NVIDIA CORPORATION. All rights reserved.
*
* Permission is hereby granted, free of charge, to any person obtaining a
* copy of this software and associated documentation files (the "Software"),
* to deal in the Software without restriction, including without limitation
* the rights to use, copy, modify, merge, publish, distribute, sublicense,
* and/or sell copies of the Software, and to permit persons to whom the
* Software is furnished to do so, subject to the following conditions:
*
* The above copyright notice and this permission notice shall be included in
* all copies or substantial portions of the Software.
*
* THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
* IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
* FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
* THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
* LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING
* FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER
* DEALINGS IN THE SOFTWARE.
*/
#ifndef NVGPU_POWER_FEATURES_PG_H
#define NVGPU_POWER_FEATURES_PG_H
#include <nvgpu/types.h>
struct gk20a;
int nvgpu_pg_elpg_disable(struct gk20a *g);
int nvgpu_pg_elpg_enable(struct gk20a *g);
bool nvgpu_pg_elpg_is_enabled(struct gk20a *g);
int nvgpu_pg_elpg_set_elpg_enabled(struct gk20a *g, bool enable);
#endif /*NVGPU_POWER_FEATURES_PG_H*/

View File

@@ -0,0 +1,34 @@
/*
* Copyright (c) 2019, NVIDIA CORPORATION. All rights reserved.
*
* Permission is hereby granted, free of charge, to any person obtaining a
* copy of this software and associated documentation files (the "Software"),
* to deal in the Software without restriction, including without limitation
* the rights to use, copy, modify, merge, publish, distribute, sublicense,
* and/or sell copies of the Software, and to permit persons to whom the
* Software is furnished to do so, subject to the following conditions:
*
* The above copyright notice and this permission notice shall be included in
* all copies or substantial portions of the Software.
*
* THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
* IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
* FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
* THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
* LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING
* FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER
* DEALINGS IN THE SOFTWARE.
*/
#ifndef NVGPU_POWER_FEATURES_H
#define NVGPU_POWER_FEATURES_H
#include <nvgpu/types.h>
struct gk20a;
int nvgpu_cg_pg_disable(struct gk20a *g);
int nvgpu_cg_pg_enable(struct gk20a *g);
#endif /*NVGPU_POWER_FEATURES_H*/

View File

@@ -65,6 +65,7 @@ static void nvgpu_init_vars(struct gk20a *g)
nvgpu_mutex_init(&g->ctxsw_disable_lock); nvgpu_mutex_init(&g->ctxsw_disable_lock);
nvgpu_mutex_init(&g->tpc_pg_lock); nvgpu_mutex_init(&g->tpc_pg_lock);
nvgpu_mutex_init(&g->clk_arb_enable_lock); nvgpu_mutex_init(&g->clk_arb_enable_lock);
nvgpu_mutex_init(&g->cg_pg_lock);
/* Init the clock req count to 0 */ /* Init the clock req count to 0 */
nvgpu_atomic_set(&g->clk_arb_global_nr, 0); nvgpu_atomic_set(&g->clk_arb_global_nr, 0);

View File

@@ -21,6 +21,8 @@
#include <nvgpu/kmem.h> #include <nvgpu/kmem.h>
#include <nvgpu/nvhost.h> #include <nvgpu/nvhost.h>
#include <nvgpu/ptimer.h> #include <nvgpu/ptimer.h>
#include <nvgpu/power_features/cg.h>
#include <nvgpu/power_features/pg.h>
#include "os_linux.h" #include "os_linux.h"
#include "sysfs.h" #include "sysfs.h"
@@ -49,16 +51,14 @@ static ssize_t elcg_enable_store(struct device *dev,
return err; return err;
if (val) { if (val) {
g->elcg_enabled = true; nvgpu_cg_elcg_set_elcg_enabled(g, true);
gr_gk20a_init_cg_mode(g, ELCG_MODE, ELCG_AUTO);
} else { } else {
g->elcg_enabled = false; nvgpu_cg_elcg_set_elcg_enabled(g, false);
gr_gk20a_init_cg_mode(g, ELCG_MODE, ELCG_RUN);
} }
gk20a_idle(g); gk20a_idle(g);
nvgpu_info(g, "ELCG is %s.", g->elcg_enabled ? "enabled" : nvgpu_info(g, "ELCG is %s.", val ? "enabled" :
"disabled"); "disabled");
return count; return count;
@@ -84,45 +84,19 @@ static ssize_t blcg_enable_store(struct device *dev,
if (kstrtoul(buf, 10, &val) < 0) if (kstrtoul(buf, 10, &val) < 0)
return -EINVAL; return -EINVAL;
if (val)
g->blcg_enabled = true;
else
g->blcg_enabled = false;
err = gk20a_busy(g); err = gk20a_busy(g);
if (err) if (err)
return err; return err;
if (g->ops.clock_gating.blcg_bus_load_gating_prod) if (val) {
g->ops.clock_gating.blcg_bus_load_gating_prod(g, nvgpu_cg_blcg_set_blcg_enabled(g, true);
g->blcg_enabled); } else {
if (g->ops.clock_gating.blcg_ce_load_gating_prod) nvgpu_cg_blcg_set_blcg_enabled(g, false);
g->ops.clock_gating.blcg_ce_load_gating_prod(g, }
g->blcg_enabled);
if (g->ops.clock_gating.blcg_ctxsw_firmware_load_gating_prod)
g->ops.clock_gating.blcg_ctxsw_firmware_load_gating_prod(g,
g->blcg_enabled);
if (g->ops.clock_gating.blcg_fb_load_gating_prod)
g->ops.clock_gating.blcg_fb_load_gating_prod(g,
g->blcg_enabled);
if (g->ops.clock_gating.blcg_fifo_load_gating_prod)
g->ops.clock_gating.blcg_fifo_load_gating_prod(g,
g->blcg_enabled);
if (g->ops.clock_gating.blcg_gr_load_gating_prod)
g->ops.clock_gating.blcg_gr_load_gating_prod(g,
g->blcg_enabled);
if (g->ops.clock_gating.blcg_ltc_load_gating_prod)
g->ops.clock_gating.blcg_ltc_load_gating_prod(g,
g->blcg_enabled);
if (g->ops.clock_gating.blcg_pmu_load_gating_prod)
g->ops.clock_gating.blcg_pmu_load_gating_prod(g,
g->blcg_enabled);
if (g->ops.clock_gating.blcg_xbar_load_gating_prod)
g->ops.clock_gating.blcg_xbar_load_gating_prod(g,
g->blcg_enabled);
gk20a_idle(g); gk20a_idle(g);
nvgpu_info(g, "BLCG is %s.", g->blcg_enabled ? "enabled" : nvgpu_info(g, "BLCG is %s.", val ? "enabled" :
"disabled"); "disabled");
return count; return count;
@@ -149,59 +123,25 @@ static ssize_t slcg_enable_store(struct device *dev,
if (kstrtoul(buf, 10, &val) < 0) if (kstrtoul(buf, 10, &val) < 0)
return -EINVAL; return -EINVAL;
if (val) err = gk20a_busy(g);
g->slcg_enabled = true; if (err) {
else return err;
g->slcg_enabled = false; }
if (val) {
nvgpu_cg_slcg_set_slcg_enabled(g, true);
} else {
nvgpu_cg_slcg_set_slcg_enabled(g, false);
}
/* /*
* TODO: slcg_therm_load_gating is not enabled anywhere during * TODO: slcg_therm_load_gating is not enabled anywhere during
* init. Therefore, it would be incongruous to add it here. Once * init. Therefore, it would be incongruous to add it here. Once
* it is added to init, we should add it here too. * it is added to init, we should add it here too.
*/ */
err = gk20a_busy(g);
if (err)
return err;
if (g->ops.clock_gating.slcg_bus_load_gating_prod)
g->ops.clock_gating.slcg_bus_load_gating_prod(g,
g->slcg_enabled);
if (g->ops.clock_gating.slcg_ce2_load_gating_prod)
g->ops.clock_gating.slcg_ce2_load_gating_prod(g,
g->slcg_enabled);
if (g->ops.clock_gating.slcg_chiplet_load_gating_prod)
g->ops.clock_gating.slcg_chiplet_load_gating_prod(g,
g->slcg_enabled);
if (g->ops.clock_gating.slcg_ctxsw_firmware_load_gating_prod)
g->ops.clock_gating.slcg_ctxsw_firmware_load_gating_prod(g,
g->slcg_enabled);
if (g->ops.clock_gating.slcg_fb_load_gating_prod)
g->ops.clock_gating.slcg_fb_load_gating_prod(g,
g->slcg_enabled);
if (g->ops.clock_gating.slcg_fifo_load_gating_prod)
g->ops.clock_gating.slcg_fifo_load_gating_prod(g,
g->slcg_enabled);
if (g->ops.clock_gating.slcg_gr_load_gating_prod)
g->ops.clock_gating.slcg_gr_load_gating_prod(g,
g->slcg_enabled);
if (g->ops.clock_gating.slcg_ltc_load_gating_prod)
g->ops.clock_gating.slcg_ltc_load_gating_prod(g,
g->slcg_enabled);
if (g->ops.clock_gating.slcg_perf_load_gating_prod)
g->ops.clock_gating.slcg_perf_load_gating_prod(g,
g->slcg_enabled);
if (g->ops.clock_gating.slcg_priring_load_gating_prod)
g->ops.clock_gating.slcg_priring_load_gating_prod(g,
g->slcg_enabled);
if (g->ops.clock_gating.slcg_pmu_load_gating_prod)
g->ops.clock_gating.slcg_pmu_load_gating_prod(g,
g->slcg_enabled);
if (g->ops.clock_gating.slcg_xbar_load_gating_prod)
g->ops.clock_gating.slcg_xbar_load_gating_prod(g,
g->slcg_enabled);
gk20a_idle(g); gk20a_idle(g);
nvgpu_info(g, "SLCG is %s.", g->slcg_enabled ? "enabled" : nvgpu_info(g, "SLCG is %s.", val ? "enabled" :
"disabled"); "disabled");
return count; return count;
@@ -474,7 +414,7 @@ static ssize_t elpg_enable_store(struct device *dev,
return -EINVAL; return -EINVAL;
if (!g->power_on) { if (!g->power_on) {
g->elpg_enabled = val ? true : false; return -EINVAL;
} else { } else {
err = gk20a_busy(g); err = gk20a_busy(g);
if (err) if (err)
@@ -483,25 +423,14 @@ static ssize_t elpg_enable_store(struct device *dev,
* Since elpg is refcounted, we should not unnecessarily call * Since elpg is refcounted, we should not unnecessarily call
* enable/disable if it is already so. * enable/disable if it is already so.
*/ */
if (val && !g->elpg_enabled) { if (val != 0) {
g->elpg_enabled = true; nvgpu_pg_elpg_set_elpg_enabled(g, true);
nvgpu_pmu_pg_global_enable(g, true); } else {
nvgpu_pg_elpg_set_elpg_enabled(g, false);
} else if (!val && g->elpg_enabled) {
if (g->ops.pmu.pmu_pg_engines_feature_list &&
g->ops.pmu.pmu_pg_engines_feature_list(g,
PMU_PG_ELPG_ENGINE_ID_GRAPHICS) !=
NVGPU_PMU_GR_FEATURE_MASK_POWER_GATING) {
nvgpu_pmu_pg_global_enable(g, false);
g->elpg_enabled = false;
} else {
g->elpg_enabled = false;
nvgpu_pmu_pg_global_enable(g, false);
}
} }
gk20a_idle(g); gk20a_idle(g);
} }
nvgpu_info(g, "ELPG is %s.", g->elpg_enabled ? "enabled" : nvgpu_info(g, "ELPG is %s.", val ? "enabled" :
"disabled"); "disabled");
return count; return count;
@@ -512,7 +441,8 @@ static ssize_t elpg_enable_read(struct device *dev,
{ {
struct gk20a *g = get_gk20a(dev); struct gk20a *g = get_gk20a(dev);
return snprintf(buf, PAGE_SIZE, "%d\n", g->elpg_enabled ? 1 : 0); return snprintf(buf, PAGE_SIZE, "%d\n",
nvgpu_pg_elpg_is_enabled(g) ? 1 : 0);
} }
static DEVICE_ATTR(elpg_enable, ROOTRW, elpg_enable_read, elpg_enable_store); static DEVICE_ATTR(elpg_enable, ROOTRW, elpg_enable_read, elpg_enable_store);
@@ -610,8 +540,9 @@ static ssize_t mscg_enable_store(struct device *dev,
/* make status visible */ /* make status visible */
smp_mb(); smp_mb();
g->mscg_enabled = false; g->mscg_enabled = false;
if (g->elpg_enabled) if (nvgpu_pg_elpg_is_enabled(g)) {
nvgpu_pmu_pg_global_enable(g, true); nvgpu_pg_elpg_enable(g);
}
} }
g->mscg_enabled = false; g->mscg_enabled = false;
} }

View File

@@ -69,6 +69,7 @@ static void vgpu_init_vars(struct gk20a *g, struct gk20a_platform *platform)
nvgpu_mutex_init(&g->power_lock); nvgpu_mutex_init(&g->power_lock);
nvgpu_mutex_init(&g->ctxsw_disable_lock); nvgpu_mutex_init(&g->ctxsw_disable_lock);
nvgpu_mutex_init(&g->clk_arb_enable_lock); nvgpu_mutex_init(&g->clk_arb_enable_lock);
nvgpu_mutex_init(&g->cg_pg_lock);
nvgpu_mutex_init(&priv->vgpu_clk_get_freq_lock); nvgpu_mutex_init(&priv->vgpu_clk_get_freq_lock);