mirror of
git://nv-tegra.nvidia.com/linux-nvgpu.git
synced 2025-12-22 17:36:20 +03:00
gpu: nvgpu: add cg and pg function
Add new power/clock gating functions that can be called by
other units.
New clock_gating functions will reside in cg.c under
common/power_features/cg unit.
New power gating functions will reside in pg.c under
common/power_features/pg unit.
Use nvgpu_pg_elpg_disable and nvgpu_pg_elpg_enable to disable/enable
elpg and also in gr_gk20a_elpg_protected macro to access gr registers.
Add cg_pg_lock to make elpg_enabled, elcg_enabled, blcg_enabled
and slcg_enabled thread safe.
JIRA NVGPU-2014
Change-Id: I00d124c2ee16242c9a3ef82e7620fbb7f1297aff
Signed-off-by: Seema Khowala <seemaj@nvidia.com>
Reviewed-on: https://git-master.nvidia.com/r/2025493
Signed-off-by: Debarshi Dutta <ddutta@nvidia.com>
(cherry-picked from c905858565 in
dev-kernel)
Reviewed-on: https://git-master.nvidia.com/r/2108406
GVS: Gerrit_Virtual_Submit
Reviewed-by: Bibek Basu <bbasu@nvidia.com>
Reviewed-by: mobile promotions <svcmobile_promotions@nvidia.com>
Tested-by: mobile promotions <svcmobile_promotions@nvidia.com>
This commit is contained in:
committed by
mobile promotions
parent
f495f52c70
commit
c81cc032c4
@@ -51,8 +51,10 @@ nvgpu-y += common/bus/bus_gk20a.o \
|
||||
common/mc/mc_gp10b.o \
|
||||
common/mc/mc_gv11b.o \
|
||||
common/mc/mc_gv100.o \
|
||||
common/sync/channel_sync.o
|
||||
|
||||
common/sync/channel_sync.o \
|
||||
common/power_features/power_features.o \
|
||||
common/power_features/cg/cg.o \
|
||||
common/power_features/pg/pg.o
|
||||
# Linux specific parts of nvgpu.
|
||||
nvgpu-y += \
|
||||
os/linux/os_ops.o \
|
||||
|
||||
@@ -116,6 +116,9 @@ srcs := os/posix/nvgpu.c \
|
||||
common/mc/mc_gp10b.c \
|
||||
common/mc/mc_gv11b.c \
|
||||
common/mc/mc_gv100.c \
|
||||
common/power_features/power_features.c \
|
||||
common/power_features/cg/cg.c \
|
||||
common/power_features/pg/pg.c \
|
||||
boardobj/boardobj.c \
|
||||
boardobj/boardobjgrp.c \
|
||||
boardobj/boardobjgrpmask.c \
|
||||
|
||||
@@ -30,6 +30,7 @@
|
||||
#include <nvgpu/pramin.h>
|
||||
#include <nvgpu/enabled.h>
|
||||
#include <nvgpu/gk20a.h>
|
||||
#include <nvgpu/power_features/cg.h>
|
||||
|
||||
/*
|
||||
* Attempt to find a reserved memory area to determine PTE size for the passed
|
||||
@@ -349,22 +350,9 @@ static int nvgpu_init_mm_reset_enable_hw(struct gk20a *g)
|
||||
g->ops.mc.fb_reset(g);
|
||||
}
|
||||
|
||||
if (g->ops.clock_gating.slcg_fb_load_gating_prod) {
|
||||
g->ops.clock_gating.slcg_fb_load_gating_prod(g,
|
||||
g->slcg_enabled);
|
||||
}
|
||||
if (g->ops.clock_gating.slcg_ltc_load_gating_prod) {
|
||||
g->ops.clock_gating.slcg_ltc_load_gating_prod(g,
|
||||
g->slcg_enabled);
|
||||
}
|
||||
if (g->ops.clock_gating.blcg_fb_load_gating_prod) {
|
||||
g->ops.clock_gating.blcg_fb_load_gating_prod(g,
|
||||
g->blcg_enabled);
|
||||
}
|
||||
if (g->ops.clock_gating.blcg_ltc_load_gating_prod) {
|
||||
g->ops.clock_gating.blcg_ltc_load_gating_prod(g,
|
||||
g->blcg_enabled);
|
||||
}
|
||||
nvgpu_cg_slcg_fb_ltc_load_enable(g);
|
||||
|
||||
nvgpu_cg_blcg_fb_ltc_load_enable(g);
|
||||
|
||||
if (g->ops.fb.init_fs_state) {
|
||||
g->ops.fb.init_fs_state(g);
|
||||
|
||||
@@ -30,6 +30,8 @@
|
||||
#include <nvgpu/bug.h>
|
||||
#include <nvgpu/utils.h>
|
||||
#include <nvgpu/gk20a.h>
|
||||
#include <nvgpu/power_features/cg.h>
|
||||
|
||||
|
||||
static int nvgpu_pg_init_task(void *arg);
|
||||
|
||||
@@ -44,15 +46,9 @@ static int pmu_enable_hw(struct nvgpu_pmu *pmu, bool enable)
|
||||
/* bring PMU falcon/engine out of reset */
|
||||
g->ops.pmu.reset_engine(g, true);
|
||||
|
||||
if (g->ops.clock_gating.slcg_pmu_load_gating_prod) {
|
||||
g->ops.clock_gating.slcg_pmu_load_gating_prod(g,
|
||||
g->slcg_enabled);
|
||||
}
|
||||
nvgpu_cg_slcg_pmu_load_enable(g);
|
||||
|
||||
if (g->ops.clock_gating.blcg_pmu_load_gating_prod) {
|
||||
g->ops.clock_gating.blcg_pmu_load_gating_prod(g,
|
||||
g->blcg_enabled);
|
||||
}
|
||||
nvgpu_cg_blcg_pmu_load_enable(g);
|
||||
|
||||
if (nvgpu_flcn_mem_scrub_wait(pmu->flcn)) {
|
||||
/* keep PMU falcon/engine in reset
|
||||
@@ -446,7 +442,7 @@ static void pmu_setup_hw_enable_elpg(struct gk20a *g)
|
||||
g->ops.gr.pmu_save_zbc(g, 0xf);
|
||||
}
|
||||
|
||||
if (g->elpg_enabled) {
|
||||
if (g->can_elpg && g->elpg_enabled) {
|
||||
/* Init reg with prod values*/
|
||||
if (g->ops.pmu.pmu_setup_elpg) {
|
||||
g->ops.pmu.pmu_setup_elpg(g);
|
||||
|
||||
566
drivers/gpu/nvgpu/common/power_features/cg/cg.c
Normal file
566
drivers/gpu/nvgpu/common/power_features/cg/cg.c
Normal file
@@ -0,0 +1,566 @@
|
||||
/*
|
||||
* Copyright (c) 2019, NVIDIA CORPORATION. All rights reserved.
|
||||
*
|
||||
* Permission is hereby granted, free of charge, to any person obtaining a
|
||||
* copy of this software and associated documentation files (the "Software"),
|
||||
* to deal in the Software without restriction, including without limitation
|
||||
* the rights to use, copy, modify, merge, publish, distribute, sublicense,
|
||||
* and/or sell copies of the Software, and to permit persons to whom the
|
||||
* Software is furnished to do so, subject to the following conditions:
|
||||
*
|
||||
* The above copyright notice and this permission notice shall be included in
|
||||
* all copies or substantial portions of the Software.
|
||||
*
|
||||
* THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
|
||||
* IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
|
||||
* FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
|
||||
* THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
|
||||
* LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING
|
||||
* FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER
|
||||
* DEALINGS IN THE SOFTWARE.
|
||||
*/
|
||||
|
||||
#include <nvgpu/gk20a.h>
|
||||
#include <nvgpu/enabled.h>
|
||||
#include <nvgpu/power_features/cg.h>
|
||||
|
||||
static void nvgpu_cg_set_mode(struct gk20a *g, int cgmode, int mode_config)
|
||||
{
|
||||
u32 engine_idx;
|
||||
u32 active_engine_id = 0;
|
||||
struct fifo_engine_info_gk20a *engine_info = NULL;
|
||||
struct fifo_gk20a *f = &g->fifo;
|
||||
|
||||
nvgpu_log_fn(g, " ");
|
||||
|
||||
for (engine_idx = 0; engine_idx < f->num_engines; ++engine_idx) {
|
||||
active_engine_id = f->active_engines_list[engine_idx];
|
||||
engine_info = &f->engine_info[active_engine_id];
|
||||
|
||||
/* gr_engine supports both BLCG and ELCG */
|
||||
if ((cgmode == BLCG_MODE) && (engine_info->engine_enum ==
|
||||
ENGINE_GR_GK20A)) {
|
||||
g->ops.therm.init_blcg_mode(g, (u32)mode_config,
|
||||
active_engine_id);
|
||||
break;
|
||||
} else if (cgmode == ELCG_MODE) {
|
||||
g->ops.therm.init_elcg_mode(g, (u32)mode_config,
|
||||
active_engine_id);
|
||||
} else {
|
||||
nvgpu_err(g, "invalid cg mode %d, config %d for "
|
||||
"act_eng_id %d",
|
||||
cgmode, mode_config, active_engine_id);
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
void nvgpu_cg_elcg_enable(struct gk20a *g)
|
||||
{
|
||||
nvgpu_log_fn(g, " ");
|
||||
|
||||
if (!nvgpu_is_enabled(g, NVGPU_GPU_CAN_ELCG)) {
|
||||
return;
|
||||
}
|
||||
|
||||
nvgpu_mutex_acquire(&g->cg_pg_lock);
|
||||
if (g->elcg_enabled) {
|
||||
nvgpu_cg_set_mode(g, ELCG_MODE, ELCG_AUTO);
|
||||
}
|
||||
nvgpu_mutex_release(&g->cg_pg_lock);
|
||||
}
|
||||
|
||||
void nvgpu_cg_elcg_disable(struct gk20a *g)
|
||||
{
|
||||
nvgpu_log_fn(g, " ");
|
||||
|
||||
if (!nvgpu_is_enabled(g, NVGPU_GPU_CAN_ELCG)) {
|
||||
return;
|
||||
}
|
||||
|
||||
nvgpu_mutex_acquire(&g->cg_pg_lock);
|
||||
if (g->elcg_enabled) {
|
||||
nvgpu_cg_set_mode(g, ELCG_MODE, ELCG_RUN);
|
||||
}
|
||||
nvgpu_mutex_release(&g->cg_pg_lock);
|
||||
|
||||
}
|
||||
|
||||
void nvgpu_cg_blcg_mode_enable(struct gk20a *g)
|
||||
{
|
||||
nvgpu_log_fn(g, " ");
|
||||
|
||||
if (!nvgpu_is_enabled(g, NVGPU_GPU_CAN_BLCG)) {
|
||||
return;
|
||||
}
|
||||
|
||||
nvgpu_mutex_acquire(&g->cg_pg_lock);
|
||||
if (g->blcg_enabled) {
|
||||
nvgpu_cg_set_mode(g, BLCG_MODE, BLCG_AUTO);
|
||||
}
|
||||
nvgpu_mutex_release(&g->cg_pg_lock);
|
||||
|
||||
}
|
||||
|
||||
void nvgpu_cg_blcg_mode_disable(struct gk20a *g)
|
||||
{
|
||||
nvgpu_log_fn(g, " ");
|
||||
|
||||
if (!nvgpu_is_enabled(g, NVGPU_GPU_CAN_BLCG)) {
|
||||
return;
|
||||
}
|
||||
|
||||
nvgpu_mutex_acquire(&g->cg_pg_lock);
|
||||
if (g->blcg_enabled) {
|
||||
nvgpu_cg_set_mode(g, BLCG_MODE, BLCG_RUN);
|
||||
}
|
||||
nvgpu_mutex_release(&g->cg_pg_lock);
|
||||
|
||||
|
||||
}
|
||||
|
||||
void nvgpu_cg_blcg_fb_ltc_load_enable(struct gk20a *g)
|
||||
{
|
||||
nvgpu_log_fn(g, " ");
|
||||
|
||||
if (!nvgpu_is_enabled(g, NVGPU_GPU_CAN_BLCG)) {
|
||||
return;
|
||||
}
|
||||
nvgpu_mutex_acquire(&g->cg_pg_lock);
|
||||
if (!g->blcg_enabled) {
|
||||
goto done;
|
||||
}
|
||||
if (g->ops.clock_gating.blcg_fb_load_gating_prod != NULL) {
|
||||
g->ops.clock_gating.blcg_fb_load_gating_prod(g, true);
|
||||
}
|
||||
if (g->ops.clock_gating.blcg_ltc_load_gating_prod != NULL) {
|
||||
g->ops.clock_gating.blcg_ltc_load_gating_prod(g, true);
|
||||
}
|
||||
done:
|
||||
nvgpu_mutex_release(&g->cg_pg_lock);
|
||||
}
|
||||
|
||||
void nvgpu_cg_blcg_fifo_load_enable(struct gk20a *g)
|
||||
{
|
||||
nvgpu_log_fn(g, " ");
|
||||
|
||||
if (!nvgpu_is_enabled(g, NVGPU_GPU_CAN_BLCG)) {
|
||||
return;
|
||||
}
|
||||
nvgpu_mutex_acquire(&g->cg_pg_lock);
|
||||
if (!g->blcg_enabled) {
|
||||
goto done;
|
||||
}
|
||||
if (g->ops.clock_gating.blcg_fifo_load_gating_prod != NULL) {
|
||||
g->ops.clock_gating.blcg_fifo_load_gating_prod(g, true);
|
||||
}
|
||||
done:
|
||||
nvgpu_mutex_release(&g->cg_pg_lock);
|
||||
}
|
||||
|
||||
void nvgpu_cg_blcg_pmu_load_enable(struct gk20a *g)
|
||||
{
|
||||
nvgpu_log_fn(g, " ");
|
||||
|
||||
if (!nvgpu_is_enabled(g, NVGPU_GPU_CAN_BLCG)) {
|
||||
return;
|
||||
}
|
||||
nvgpu_mutex_acquire(&g->cg_pg_lock);
|
||||
if (!g->blcg_enabled) {
|
||||
goto done;
|
||||
}
|
||||
if (g->ops.clock_gating.blcg_pmu_load_gating_prod != NULL) {
|
||||
g->ops.clock_gating.blcg_pmu_load_gating_prod(g, true);
|
||||
}
|
||||
done:
|
||||
nvgpu_mutex_release(&g->cg_pg_lock);
|
||||
}
|
||||
|
||||
void nvgpu_cg_blcg_ce_load_enable(struct gk20a *g)
|
||||
{
|
||||
nvgpu_log_fn(g, " ");
|
||||
|
||||
if (!nvgpu_is_enabled(g, NVGPU_GPU_CAN_BLCG)) {
|
||||
return;
|
||||
}
|
||||
nvgpu_mutex_acquire(&g->cg_pg_lock);
|
||||
if (!g->blcg_enabled) {
|
||||
goto done;
|
||||
}
|
||||
if (g->ops.clock_gating.blcg_ce_load_gating_prod != NULL) {
|
||||
g->ops.clock_gating.blcg_ce_load_gating_prod(g, true);
|
||||
}
|
||||
done:
|
||||
nvgpu_mutex_release(&g->cg_pg_lock);
|
||||
}
|
||||
|
||||
void nvgpu_cg_blcg_gr_load_enable(struct gk20a *g)
|
||||
{
|
||||
nvgpu_log_fn(g, " ");
|
||||
|
||||
if (!nvgpu_is_enabled(g, NVGPU_GPU_CAN_BLCG)) {
|
||||
return;
|
||||
}
|
||||
nvgpu_mutex_acquire(&g->cg_pg_lock);
|
||||
if (!g->blcg_enabled) {
|
||||
goto done;
|
||||
}
|
||||
if (g->ops.clock_gating.blcg_gr_load_gating_prod != NULL) {
|
||||
g->ops.clock_gating.blcg_gr_load_gating_prod(g, true);
|
||||
}
|
||||
done:
|
||||
nvgpu_mutex_release(&g->cg_pg_lock);
|
||||
}
|
||||
|
||||
void nvgpu_cg_slcg_fb_ltc_load_enable(struct gk20a *g)
|
||||
{
|
||||
nvgpu_log_fn(g, " ");
|
||||
|
||||
if (!nvgpu_is_enabled(g, NVGPU_GPU_CAN_SLCG)) {
|
||||
return;
|
||||
}
|
||||
nvgpu_mutex_acquire(&g->cg_pg_lock);
|
||||
if (!g->slcg_enabled) {
|
||||
goto done;
|
||||
}
|
||||
if (g->ops.clock_gating.slcg_fb_load_gating_prod != NULL) {
|
||||
g->ops.clock_gating.slcg_fb_load_gating_prod(g, true);
|
||||
}
|
||||
if (g->ops.clock_gating.slcg_ltc_load_gating_prod != NULL) {
|
||||
g->ops.clock_gating.slcg_ltc_load_gating_prod(g, true);
|
||||
}
|
||||
done:
|
||||
nvgpu_mutex_release(&g->cg_pg_lock);
|
||||
}
|
||||
|
||||
void nvgpu_cg_slcg_priring_load_enable(struct gk20a *g)
|
||||
{
|
||||
nvgpu_log_fn(g, " ");
|
||||
|
||||
if (!nvgpu_is_enabled(g, NVGPU_GPU_CAN_SLCG)) {
|
||||
return;
|
||||
}
|
||||
nvgpu_mutex_acquire(&g->cg_pg_lock);
|
||||
if (!g->slcg_enabled) {
|
||||
goto done;
|
||||
}
|
||||
if (g->ops.clock_gating.slcg_priring_load_gating_prod != NULL) {
|
||||
g->ops.clock_gating.slcg_priring_load_gating_prod(g, true);
|
||||
}
|
||||
done:
|
||||
nvgpu_mutex_release(&g->cg_pg_lock);
|
||||
}
|
||||
|
||||
void nvgpu_cg_slcg_gr_perf_ltc_load_enable(struct gk20a *g)
|
||||
{
|
||||
nvgpu_log_fn(g, " ");
|
||||
|
||||
if (!nvgpu_is_enabled(g, NVGPU_GPU_CAN_SLCG)) {
|
||||
return;
|
||||
}
|
||||
nvgpu_mutex_acquire(&g->cg_pg_lock);
|
||||
if (!g->slcg_enabled) {
|
||||
goto done;
|
||||
}
|
||||
if (g->ops.clock_gating.slcg_ltc_load_gating_prod != NULL) {
|
||||
g->ops.clock_gating.slcg_ltc_load_gating_prod(g, true);
|
||||
}
|
||||
if (g->ops.clock_gating.slcg_perf_load_gating_prod != NULL) {
|
||||
g->ops.clock_gating.slcg_perf_load_gating_prod(g, true);
|
||||
}
|
||||
if (g->ops.clock_gating.slcg_gr_load_gating_prod != NULL) {
|
||||
g->ops.clock_gating.slcg_gr_load_gating_prod(g, true);
|
||||
}
|
||||
done:
|
||||
nvgpu_mutex_release(&g->cg_pg_lock);
|
||||
}
|
||||
|
||||
void nvgpu_cg_slcg_gr_perf_ltc_load_disable(struct gk20a *g)
|
||||
{
|
||||
nvgpu_log_fn(g, " ");
|
||||
|
||||
if (!nvgpu_is_enabled(g, NVGPU_GPU_CAN_SLCG)) {
|
||||
return;
|
||||
}
|
||||
nvgpu_mutex_acquire(&g->cg_pg_lock);
|
||||
if (!g->slcg_enabled) {
|
||||
goto done;
|
||||
}
|
||||
if (g->ops.clock_gating.slcg_gr_load_gating_prod != NULL) {
|
||||
g->ops.clock_gating.slcg_gr_load_gating_prod(g, false);
|
||||
}
|
||||
if (g->ops.clock_gating.slcg_perf_load_gating_prod != NULL) {
|
||||
g->ops.clock_gating.slcg_perf_load_gating_prod(g, false);
|
||||
}
|
||||
if (g->ops.clock_gating.slcg_ltc_load_gating_prod != NULL) {
|
||||
g->ops.clock_gating.slcg_ltc_load_gating_prod(g, false);
|
||||
}
|
||||
done:
|
||||
nvgpu_mutex_release(&g->cg_pg_lock);
|
||||
}
|
||||
|
||||
void nvgpu_cg_slcg_fifo_load_enable(struct gk20a *g)
|
||||
{
|
||||
nvgpu_log_fn(g, " ");
|
||||
|
||||
if (!nvgpu_is_enabled(g, NVGPU_GPU_CAN_SLCG)) {
|
||||
return;
|
||||
}
|
||||
nvgpu_mutex_acquire(&g->cg_pg_lock);
|
||||
if (!g->slcg_enabled) {
|
||||
goto done;
|
||||
}
|
||||
if (g->ops.clock_gating.slcg_fifo_load_gating_prod != NULL) {
|
||||
g->ops.clock_gating.slcg_fifo_load_gating_prod(g, true);
|
||||
}
|
||||
done:
|
||||
nvgpu_mutex_release(&g->cg_pg_lock);
|
||||
}
|
||||
|
||||
void nvgpu_cg_slcg_pmu_load_enable(struct gk20a *g)
|
||||
{
|
||||
nvgpu_log_fn(g, " ");
|
||||
|
||||
if (!nvgpu_is_enabled(g, NVGPU_GPU_CAN_SLCG)) {
|
||||
return;
|
||||
}
|
||||
nvgpu_mutex_acquire(&g->cg_pg_lock);
|
||||
if (!g->slcg_enabled) {
|
||||
goto done;
|
||||
}
|
||||
if (g->ops.clock_gating.slcg_pmu_load_gating_prod != NULL) {
|
||||
g->ops.clock_gating.slcg_pmu_load_gating_prod(g, true);
|
||||
}
|
||||
done:
|
||||
nvgpu_mutex_release(&g->cg_pg_lock);
|
||||
}
|
||||
|
||||
void nvgpu_cg_slcg_ce2_load_enable(struct gk20a *g)
|
||||
{
|
||||
nvgpu_log_fn(g, " ");
|
||||
|
||||
if (!nvgpu_is_enabled(g, NVGPU_GPU_CAN_SLCG)) {
|
||||
return;
|
||||
}
|
||||
nvgpu_mutex_acquire(&g->cg_pg_lock);
|
||||
if (!g->slcg_enabled) {
|
||||
goto done;
|
||||
}
|
||||
if (g->ops.clock_gating.slcg_ce2_load_gating_prod != NULL) {
|
||||
g->ops.clock_gating.slcg_ce2_load_gating_prod(g, true);
|
||||
}
|
||||
done:
|
||||
nvgpu_mutex_release(&g->cg_pg_lock);
|
||||
}
|
||||
|
||||
void nvgpu_cg_init_gr_load_gating_prod(struct gk20a *g)
|
||||
{
|
||||
nvgpu_log_fn(g, " ");
|
||||
|
||||
nvgpu_mutex_acquire(&g->cg_pg_lock);
|
||||
|
||||
if (!nvgpu_is_enabled(g, NVGPU_GPU_CAN_SLCG)) {
|
||||
goto check_can_blcg;
|
||||
}
|
||||
if (!g->slcg_enabled) {
|
||||
goto check_can_blcg;
|
||||
}
|
||||
|
||||
if (g->ops.clock_gating.slcg_bus_load_gating_prod != NULL) {
|
||||
g->ops.clock_gating.slcg_bus_load_gating_prod(g, true);
|
||||
}
|
||||
if (g->ops.clock_gating.slcg_chiplet_load_gating_prod != NULL) {
|
||||
g->ops.clock_gating.slcg_chiplet_load_gating_prod(g, true);
|
||||
}
|
||||
if (g->ops.clock_gating.slcg_gr_load_gating_prod != NULL) {
|
||||
g->ops.clock_gating.slcg_gr_load_gating_prod(g, true);
|
||||
}
|
||||
if (g->ops.clock_gating.slcg_ctxsw_firmware_load_gating_prod != NULL) {
|
||||
g->ops.clock_gating.slcg_ctxsw_firmware_load_gating_prod(g,
|
||||
true);
|
||||
}
|
||||
if (g->ops.clock_gating.slcg_perf_load_gating_prod != NULL) {
|
||||
g->ops.clock_gating.slcg_perf_load_gating_prod(g, true);
|
||||
}
|
||||
if (g->ops.clock_gating.slcg_xbar_load_gating_prod != NULL) {
|
||||
g->ops.clock_gating.slcg_xbar_load_gating_prod(g, true);
|
||||
}
|
||||
|
||||
check_can_blcg:
|
||||
if (!nvgpu_is_enabled(g, NVGPU_GPU_CAN_BLCG)) {
|
||||
goto pg_gr_load;
|
||||
}
|
||||
if (!g->blcg_enabled) {
|
||||
goto pg_gr_load;
|
||||
}
|
||||
if (g->ops.clock_gating.blcg_bus_load_gating_prod != NULL) {
|
||||
g->ops.clock_gating.blcg_bus_load_gating_prod(g, true);
|
||||
}
|
||||
if (g->ops.clock_gating.blcg_gr_load_gating_prod != NULL) {
|
||||
g->ops.clock_gating.blcg_gr_load_gating_prod(g, true);
|
||||
}
|
||||
if (g->ops.clock_gating.blcg_ctxsw_firmware_load_gating_prod != NULL) {
|
||||
g->ops.clock_gating.blcg_ctxsw_firmware_load_gating_prod(g,
|
||||
true);
|
||||
}
|
||||
if (g->ops.clock_gating.blcg_xbar_load_gating_prod != NULL) {
|
||||
g->ops.clock_gating.blcg_xbar_load_gating_prod(g, true);
|
||||
}
|
||||
pg_gr_load:
|
||||
if (g->ops.clock_gating.pg_gr_load_gating_prod != NULL) {
|
||||
g->ops.clock_gating.pg_gr_load_gating_prod(g, true);
|
||||
}
|
||||
|
||||
nvgpu_mutex_release(&g->cg_pg_lock);
|
||||
}
|
||||
|
||||
void nvgpu_cg_elcg_set_elcg_enabled(struct gk20a *g, bool enable)
|
||||
{
|
||||
nvgpu_log_fn(g, " ");
|
||||
|
||||
if (!nvgpu_is_enabled(g, NVGPU_GPU_CAN_ELCG)) {
|
||||
return;
|
||||
}
|
||||
|
||||
nvgpu_mutex_release(&g->cg_pg_lock);
|
||||
if (enable) {
|
||||
if (!g->elcg_enabled) {
|
||||
g->elcg_enabled = true;
|
||||
nvgpu_cg_set_mode(g, ELCG_MODE, ELCG_AUTO);
|
||||
}
|
||||
} else {
|
||||
if (g->elcg_enabled) {
|
||||
g->elcg_enabled = false;
|
||||
nvgpu_cg_set_mode(g, ELCG_MODE, ELCG_RUN);
|
||||
}
|
||||
}
|
||||
nvgpu_mutex_release(&g->cg_pg_lock);
|
||||
}
|
||||
|
||||
void nvgpu_cg_blcg_set_blcg_enabled(struct gk20a *g, bool enable)
|
||||
{
|
||||
bool load = false;
|
||||
|
||||
nvgpu_log_fn(g, " ");
|
||||
|
||||
if (!nvgpu_is_enabled(g, NVGPU_GPU_CAN_BLCG)) {
|
||||
return;
|
||||
}
|
||||
|
||||
nvgpu_mutex_acquire(&g->cg_pg_lock);
|
||||
if (enable) {
|
||||
if (!g->blcg_enabled) {
|
||||
load = true;
|
||||
g->blcg_enabled = true;
|
||||
}
|
||||
} else {
|
||||
if (g->blcg_enabled) {
|
||||
load = true;
|
||||
g->blcg_enabled = false;
|
||||
}
|
||||
}
|
||||
if (!load ) {
|
||||
goto done;
|
||||
}
|
||||
|
||||
if (g->ops.clock_gating.blcg_bus_load_gating_prod != NULL) {
|
||||
g->ops.clock_gating.blcg_bus_load_gating_prod(g, enable);
|
||||
}
|
||||
if (g->ops.clock_gating.blcg_ce_load_gating_prod != NULL) {
|
||||
g->ops.clock_gating.blcg_ce_load_gating_prod(g, enable);
|
||||
}
|
||||
if (g->ops.clock_gating.blcg_ctxsw_firmware_load_gating_prod != NULL) {
|
||||
g->ops.clock_gating.blcg_ctxsw_firmware_load_gating_prod(g,
|
||||
enable);
|
||||
}
|
||||
if (g->ops.clock_gating.blcg_fb_load_gating_prod != NULL) {
|
||||
g->ops.clock_gating.blcg_fb_load_gating_prod(g, enable);
|
||||
}
|
||||
if (g->ops.clock_gating.blcg_fifo_load_gating_prod != NULL) {
|
||||
g->ops.clock_gating.blcg_fifo_load_gating_prod(g, enable);
|
||||
}
|
||||
if (g->ops.clock_gating.blcg_gr_load_gating_prod != NULL) {
|
||||
g->ops.clock_gating.blcg_gr_load_gating_prod(g, enable);
|
||||
}
|
||||
if (g->ops.clock_gating.blcg_ltc_load_gating_prod != NULL) {
|
||||
g->ops.clock_gating.blcg_ltc_load_gating_prod(g, enable);
|
||||
}
|
||||
if (g->ops.clock_gating.blcg_pmu_load_gating_prod != NULL) {
|
||||
g->ops.clock_gating.blcg_pmu_load_gating_prod(g, enable);
|
||||
}
|
||||
if (g->ops.clock_gating.blcg_xbar_load_gating_prod != NULL) {
|
||||
g->ops.clock_gating.blcg_xbar_load_gating_prod(g, enable);
|
||||
}
|
||||
|
||||
done:
|
||||
nvgpu_mutex_release(&g->cg_pg_lock);
|
||||
}
|
||||
|
||||
void nvgpu_cg_slcg_set_slcg_enabled(struct gk20a *g, bool enable)
|
||||
{
|
||||
bool load = false;
|
||||
|
||||
nvgpu_log_fn(g, " ");
|
||||
|
||||
if (!nvgpu_is_enabled(g, NVGPU_GPU_CAN_SLCG)) {
|
||||
return;
|
||||
}
|
||||
|
||||
nvgpu_mutex_acquire(&g->cg_pg_lock);
|
||||
if (enable) {
|
||||
if (!g->slcg_enabled) {
|
||||
load = true;
|
||||
g->slcg_enabled = true;
|
||||
}
|
||||
} else {
|
||||
if (g->slcg_enabled) {
|
||||
load = true;
|
||||
g->slcg_enabled = false;
|
||||
}
|
||||
}
|
||||
if (!load ) {
|
||||
goto done;
|
||||
}
|
||||
|
||||
if (g->ops.clock_gating.slcg_bus_load_gating_prod != NULL) {
|
||||
g->ops.clock_gating.slcg_bus_load_gating_prod(g, enable);
|
||||
}
|
||||
if (g->ops.clock_gating.slcg_ce2_load_gating_prod != NULL) {
|
||||
g->ops.clock_gating.slcg_ce2_load_gating_prod(g, enable);
|
||||
}
|
||||
if (g->ops.clock_gating.slcg_chiplet_load_gating_prod != NULL) {
|
||||
g->ops.clock_gating.slcg_chiplet_load_gating_prod(g, enable);
|
||||
}
|
||||
if (g->ops.clock_gating.slcg_ctxsw_firmware_load_gating_prod !=
|
||||
NULL) {
|
||||
g->ops.clock_gating.slcg_ctxsw_firmware_load_gating_prod(g,
|
||||
enable);
|
||||
}
|
||||
if (g->ops.clock_gating.slcg_fb_load_gating_prod != NULL) {
|
||||
g->ops.clock_gating.slcg_fb_load_gating_prod(g, enable);
|
||||
}
|
||||
if (g->ops.clock_gating.slcg_fifo_load_gating_prod != NULL) {
|
||||
g->ops.clock_gating.slcg_fifo_load_gating_prod(g, enable);
|
||||
}
|
||||
if (g->ops.clock_gating.slcg_gr_load_gating_prod != NULL) {
|
||||
g->ops.clock_gating.slcg_gr_load_gating_prod(g, enable);
|
||||
}
|
||||
if (g->ops.clock_gating.slcg_ltc_load_gating_prod != NULL) {
|
||||
g->ops.clock_gating.slcg_ltc_load_gating_prod(g, enable);
|
||||
}
|
||||
if (g->ops.clock_gating.slcg_perf_load_gating_prod != NULL) {
|
||||
g->ops.clock_gating.slcg_perf_load_gating_prod(g, enable);
|
||||
}
|
||||
if (g->ops.clock_gating.slcg_priring_load_gating_prod != NULL) {
|
||||
g->ops.clock_gating.slcg_priring_load_gating_prod(g,
|
||||
enable);
|
||||
}
|
||||
if (g->ops.clock_gating.slcg_pmu_load_gating_prod != NULL) {
|
||||
g->ops.clock_gating.slcg_pmu_load_gating_prod(g, enable);
|
||||
}
|
||||
if (g->ops.clock_gating.slcg_xbar_load_gating_prod != NULL) {
|
||||
g->ops.clock_gating.slcg_xbar_load_gating_prod(g, enable);
|
||||
}
|
||||
|
||||
done:
|
||||
nvgpu_mutex_release(&g->cg_pg_lock);
|
||||
}
|
||||
106
drivers/gpu/nvgpu/common/power_features/pg/pg.c
Normal file
106
drivers/gpu/nvgpu/common/power_features/pg/pg.c
Normal file
@@ -0,0 +1,106 @@
|
||||
/*
|
||||
* Copyright (c) 2019, NVIDIA CORPORATION. All rights reserved.
|
||||
*
|
||||
* Permission is hereby granted, free of charge, to any person obtaining a
|
||||
* copy of this software and associated documentation files (the "Software"),
|
||||
* to deal in the Software without restriction, including without limitation
|
||||
* the rights to use, copy, modify, merge, publish, distribute, sublicense,
|
||||
* and/or sell copies of the Software, and to permit persons to whom the
|
||||
* Software is furnished to do so, subject to the following conditions:
|
||||
*
|
||||
* The above copyright notice and this permission notice shall be included in
|
||||
* all copies or substantial portions of the Software.
|
||||
*
|
||||
* THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
|
||||
* IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
|
||||
* FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
|
||||
* THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
|
||||
* LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING
|
||||
* FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER
|
||||
* DEALINGS IN THE SOFTWARE.
|
||||
*/
|
||||
|
||||
#include <nvgpu/gk20a.h>
|
||||
#include <nvgpu/pmu.h>
|
||||
#include <nvgpu/power_features/pg.h>
|
||||
|
||||
bool nvgpu_pg_elpg_is_enabled(struct gk20a *g)
|
||||
{
|
||||
bool elpg_enabled;
|
||||
|
||||
nvgpu_log_fn(g, " ");
|
||||
|
||||
nvgpu_mutex_acquire(&g->cg_pg_lock);
|
||||
elpg_enabled = g->elpg_enabled;
|
||||
nvgpu_mutex_release(&g->cg_pg_lock);
|
||||
return elpg_enabled;
|
||||
}
|
||||
|
||||
int nvgpu_pg_elpg_enable(struct gk20a *g)
|
||||
{
|
||||
int err = 0;
|
||||
|
||||
nvgpu_log_fn(g, " ");
|
||||
|
||||
if (!g->can_elpg) {
|
||||
return 0;
|
||||
}
|
||||
|
||||
nvgpu_mutex_acquire(&g->cg_pg_lock);
|
||||
if (g->elpg_enabled) {
|
||||
err = nvgpu_pmu_pg_global_enable(g, true);
|
||||
}
|
||||
nvgpu_mutex_release(&g->cg_pg_lock);
|
||||
return err;
|
||||
}
|
||||
|
||||
int nvgpu_pg_elpg_disable(struct gk20a *g)
|
||||
{
|
||||
int err = 0;
|
||||
|
||||
nvgpu_log_fn(g, " ");
|
||||
|
||||
if (!g->can_elpg) {
|
||||
return 0;
|
||||
}
|
||||
|
||||
nvgpu_mutex_acquire(&g->cg_pg_lock);
|
||||
if (g->elpg_enabled) {
|
||||
err = nvgpu_pmu_pg_global_enable(g, false);
|
||||
}
|
||||
nvgpu_mutex_release(&g->cg_pg_lock);
|
||||
return err;
|
||||
}
|
||||
|
||||
int nvgpu_pg_elpg_set_elpg_enabled(struct gk20a *g, bool enable)
|
||||
{
|
||||
int err = 0;
|
||||
bool change_mode = false;
|
||||
|
||||
nvgpu_log_fn(g, " ");
|
||||
|
||||
if (!g->can_elpg) {
|
||||
return 0;
|
||||
}
|
||||
|
||||
nvgpu_mutex_acquire(&g->cg_pg_lock);
|
||||
if (enable) {
|
||||
if (!g->elpg_enabled) {
|
||||
change_mode = true;
|
||||
g->elpg_enabled = true;
|
||||
}
|
||||
} else {
|
||||
if (g->elpg_enabled) {
|
||||
change_mode = true;
|
||||
g->elpg_enabled = false;
|
||||
}
|
||||
}
|
||||
if (!change_mode) {
|
||||
goto done;
|
||||
}
|
||||
|
||||
err = nvgpu_pmu_pg_global_enable(g, enable);
|
||||
done:
|
||||
nvgpu_mutex_release(&g->cg_pg_lock);
|
||||
return err;
|
||||
}
|
||||
66
drivers/gpu/nvgpu/common/power_features/power_features.c
Normal file
66
drivers/gpu/nvgpu/common/power_features/power_features.c
Normal file
@@ -0,0 +1,66 @@
|
||||
/*
|
||||
* Copyright (c) 2019, NVIDIA CORPORATION. All rights reserved.
|
||||
*
|
||||
* Permission is hereby granted, free of charge, to any person obtaining a
|
||||
* copy of this software and associated documentation files (the "Software"),
|
||||
* to deal in the Software without restriction, including without limitation
|
||||
* the rights to use, copy, modify, merge, publish, distribute, sublicense,
|
||||
* and/or sell copies of the Software, and to permit persons to whom the
|
||||
* Software is furnished to do so, subject to the following conditions:
|
||||
*
|
||||
* The above copyright notice and this permission notice shall be included in
|
||||
* all copies or substantial portions of the Software.
|
||||
*
|
||||
* THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
|
||||
* IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
|
||||
* FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
|
||||
* THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
|
||||
* LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING
|
||||
* FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER
|
||||
* DEALINGS IN THE SOFTWARE.
|
||||
*/
|
||||
|
||||
#include <nvgpu/gk20a.h>
|
||||
#include <nvgpu/power_features/cg.h>
|
||||
#include <nvgpu/power_features/pg.h>
|
||||
#include <nvgpu/power_features/power_features.h>
|
||||
|
||||
int nvgpu_cg_pg_disable(struct gk20a *g)
|
||||
{
|
||||
int err = 0;
|
||||
|
||||
nvgpu_log_fn(g, " ");
|
||||
|
||||
/* disable elpg before clock gating */
|
||||
err = nvgpu_pg_elpg_disable(g);
|
||||
if (err != 0) {
|
||||
nvgpu_err(g, "failed to set disable elpg");
|
||||
}
|
||||
nvgpu_cg_slcg_gr_perf_ltc_load_disable(g);
|
||||
|
||||
nvgpu_cg_blcg_mode_disable(g);
|
||||
|
||||
nvgpu_cg_elcg_disable(g);
|
||||
|
||||
return err;
|
||||
}
|
||||
|
||||
int nvgpu_cg_pg_enable(struct gk20a *g)
|
||||
{
|
||||
int err = 0;
|
||||
|
||||
nvgpu_log_fn(g, " ");
|
||||
|
||||
nvgpu_cg_elcg_enable(g);
|
||||
|
||||
nvgpu_cg_blcg_mode_enable(g);
|
||||
|
||||
nvgpu_cg_slcg_gr_perf_ltc_load_enable(g);
|
||||
|
||||
err = nvgpu_pg_elpg_enable(g);
|
||||
if (err != 0) {
|
||||
nvgpu_err(g, "failed to set enable elpg");
|
||||
}
|
||||
|
||||
return err;
|
||||
}
|
||||
@@ -26,6 +26,7 @@
|
||||
#include <nvgpu/enabled.h>
|
||||
#include <nvgpu/io.h>
|
||||
#include <nvgpu/utils.h>
|
||||
#include <nvgpu/power_features/cg.h>
|
||||
|
||||
#include "priv_ring_gm20b.h"
|
||||
|
||||
@@ -41,10 +42,7 @@ void gm20b_priv_ring_enable(struct gk20a *g)
|
||||
|
||||
nvgpu_log(g, gpu_dbg_info, "enabling priv ring");
|
||||
|
||||
if (g->ops.clock_gating.slcg_priring_load_gating_prod) {
|
||||
g->ops.clock_gating.slcg_priring_load_gating_prod(g,
|
||||
g->slcg_enabled);
|
||||
}
|
||||
nvgpu_cg_slcg_priring_load_enable(g);
|
||||
|
||||
gk20a_writel(g,pri_ringmaster_command_r(),
|
||||
0x4);
|
||||
|
||||
@@ -30,6 +30,7 @@
|
||||
#include <nvgpu/io.h>
|
||||
#include <nvgpu/utils.h>
|
||||
#include <nvgpu/channel.h>
|
||||
#include <nvgpu/power_features/cg.h>
|
||||
|
||||
#include "gk20a.h"
|
||||
#include "gk20a/fence_gk20a.h"
|
||||
@@ -339,14 +340,9 @@ int gk20a_init_ce_support(struct gk20a *g)
|
||||
|
||||
g->ops.mc.reset(g, ce_reset_mask);
|
||||
|
||||
if (g->ops.clock_gating.slcg_ce2_load_gating_prod) {
|
||||
g->ops.clock_gating.slcg_ce2_load_gating_prod(g,
|
||||
g->slcg_enabled);
|
||||
}
|
||||
if (g->ops.clock_gating.blcg_ce_load_gating_prod) {
|
||||
g->ops.clock_gating.blcg_ce_load_gating_prod(g,
|
||||
g->blcg_enabled);
|
||||
}
|
||||
nvgpu_cg_slcg_ce2_load_enable(g);
|
||||
|
||||
nvgpu_cg_blcg_ce_load_enable(g);
|
||||
|
||||
if (ce_app->initialised) {
|
||||
/* assume this happen during poweron/poweroff GPU sequence */
|
||||
|
||||
@@ -1,7 +1,7 @@
|
||||
/*
|
||||
* Tegra GK20A GPU Debugger/Profiler Driver
|
||||
*
|
||||
* Copyright (c) 2013-2018, NVIDIA CORPORATION. All rights reserved.
|
||||
* Copyright (c) 2013-2019, NVIDIA CORPORATION. All rights reserved.
|
||||
*
|
||||
* Permission is hereby granted, free of charge, to any person obtaining a
|
||||
* copy of this software and associated documentation files (the "Software"),
|
||||
@@ -32,6 +32,7 @@
|
||||
#include <nvgpu/utils.h>
|
||||
#include <nvgpu/channel.h>
|
||||
#include <nvgpu/unit.h>
|
||||
#include <nvgpu/power_features/power_features.h>
|
||||
|
||||
#include "gk20a.h"
|
||||
#include "gr_gk20a.h"
|
||||
@@ -234,60 +235,28 @@ int dbg_set_powergate(struct dbg_session_gk20a *dbg_s, bool disable_powergate)
|
||||
return err;
|
||||
}
|
||||
|
||||
/*do elpg disable before clock gating */
|
||||
nvgpu_pmu_pg_global_enable(g, false);
|
||||
|
||||
if (g->ops.clock_gating.slcg_gr_load_gating_prod) {
|
||||
g->ops.clock_gating.slcg_gr_load_gating_prod(g,
|
||||
false);
|
||||
}
|
||||
if (g->ops.clock_gating.slcg_perf_load_gating_prod) {
|
||||
g->ops.clock_gating.slcg_perf_load_gating_prod(g,
|
||||
false);
|
||||
}
|
||||
if (g->ops.clock_gating.slcg_ltc_load_gating_prod) {
|
||||
g->ops.clock_gating.slcg_ltc_load_gating_prod(g,
|
||||
false);
|
||||
}
|
||||
|
||||
gr_gk20a_init_cg_mode(g, BLCG_MODE, BLCG_RUN);
|
||||
gr_gk20a_init_cg_mode(g, ELCG_MODE, ELCG_RUN);
|
||||
err = nvgpu_cg_pg_disable(g);
|
||||
|
||||
if (err == 0) {
|
||||
dbg_s->is_pg_disabled = true;
|
||||
nvgpu_log(g, gpu_dbg_gpu_dbg | gpu_dbg_fn,
|
||||
"pg disabled");
|
||||
}
|
||||
} else {
|
||||
/* restore (can) powergate, clk state */
|
||||
/* release pending exceptions to fault/be handled as usual */
|
||||
/*TBD: ordering of these? */
|
||||
|
||||
if (g->elcg_enabled) {
|
||||
gr_gk20a_init_cg_mode(g, ELCG_MODE, ELCG_AUTO);
|
||||
}
|
||||
err = nvgpu_cg_pg_enable(g);
|
||||
|
||||
if (g->blcg_enabled) {
|
||||
gr_gk20a_init_cg_mode(g, BLCG_MODE, BLCG_AUTO);
|
||||
}
|
||||
|
||||
if (g->slcg_enabled) {
|
||||
if (g->ops.clock_gating.slcg_ltc_load_gating_prod) {
|
||||
g->ops.clock_gating.slcg_ltc_load_gating_prod(g,
|
||||
g->slcg_enabled);
|
||||
}
|
||||
if (g->ops.clock_gating.slcg_perf_load_gating_prod) {
|
||||
g->ops.clock_gating.slcg_perf_load_gating_prod(g,
|
||||
g->slcg_enabled);
|
||||
}
|
||||
if (g->ops.clock_gating.slcg_gr_load_gating_prod) {
|
||||
g->ops.clock_gating.slcg_gr_load_gating_prod(g,
|
||||
g->slcg_enabled);
|
||||
}
|
||||
}
|
||||
nvgpu_pmu_pg_global_enable(g, true);
|
||||
|
||||
nvgpu_log(g, gpu_dbg_gpu_dbg | gpu_dbg_fn,
|
||||
"module idle");
|
||||
nvgpu_log(g, gpu_dbg_gpu_dbg | gpu_dbg_fn, "module idle");
|
||||
gk20a_idle(g);
|
||||
|
||||
if (err == 0) {
|
||||
dbg_s->is_pg_disabled = false;
|
||||
nvgpu_log(g, gpu_dbg_gpu_dbg | gpu_dbg_fn,
|
||||
"pg enabled");
|
||||
}
|
||||
}
|
||||
|
||||
nvgpu_log(g, gpu_dbg_fn|gpu_dbg_gpu_dbg, "%s powergate mode = %s done",
|
||||
|
||||
@@ -45,6 +45,8 @@
|
||||
#include <nvgpu/utils.h>
|
||||
#include <nvgpu/channel.h>
|
||||
#include <nvgpu/unit.h>
|
||||
#include <nvgpu/power_features/power_features.h>
|
||||
#include <nvgpu/power_features/cg.h>
|
||||
|
||||
#include "gk20a.h"
|
||||
#include "mm_gk20a.h"
|
||||
@@ -824,14 +826,9 @@ int gk20a_init_fifo_reset_enable_hw(struct gk20a *g)
|
||||
/* enable pmc pfifo */
|
||||
g->ops.mc.reset(g, g->ops.mc.reset_mask(g, NVGPU_UNIT_FIFO));
|
||||
|
||||
if (g->ops.clock_gating.slcg_fifo_load_gating_prod) {
|
||||
g->ops.clock_gating.slcg_fifo_load_gating_prod(g,
|
||||
g->slcg_enabled);
|
||||
}
|
||||
if (g->ops.clock_gating.blcg_fifo_load_gating_prod) {
|
||||
g->ops.clock_gating.blcg_fifo_load_gating_prod(g,
|
||||
g->blcg_enabled);
|
||||
}
|
||||
nvgpu_cg_slcg_fifo_load_enable(g);
|
||||
|
||||
nvgpu_cg_blcg_fifo_load_enable(g);
|
||||
|
||||
timeout = gk20a_readl(g, fifo_fb_timeout_r());
|
||||
timeout = set_field(timeout, fifo_fb_timeout_period_m(),
|
||||
@@ -1361,8 +1358,8 @@ void gk20a_fifo_reset_engine(struct gk20a *g, u32 engine_id)
|
||||
}
|
||||
|
||||
if (engine_enum == ENGINE_GR_GK20A) {
|
||||
if (g->support_pmu && g->can_elpg) {
|
||||
if (nvgpu_pmu_disable_elpg(g)) {
|
||||
if (g->support_pmu) {
|
||||
if (nvgpu_pg_elpg_disable(g) != 0 ) {
|
||||
nvgpu_err(g, "failed to set disable elpg");
|
||||
}
|
||||
}
|
||||
@@ -1391,8 +1388,10 @@ void gk20a_fifo_reset_engine(struct gk20a *g, u32 engine_id)
|
||||
"HALT gr pipe not supported and "
|
||||
"gr cannot be reset without halting gr pipe");
|
||||
}
|
||||
if (g->support_pmu && g->can_elpg) {
|
||||
nvgpu_pmu_enable_elpg(g);
|
||||
if (g->support_pmu) {
|
||||
if (nvgpu_pg_elpg_enable(g) != 0 ) {
|
||||
nvgpu_err(g, "failed to set enable elpg");
|
||||
}
|
||||
}
|
||||
}
|
||||
if ((engine_enum == ENGINE_GRCE_GK20A) ||
|
||||
@@ -1638,25 +1637,11 @@ static bool gk20a_fifo_handle_mmu_fault_locked(
|
||||
g->fifo.deferred_reset_pending = false;
|
||||
|
||||
/* Disable power management */
|
||||
if (g->support_pmu && g->can_elpg) {
|
||||
if (nvgpu_pmu_disable_elpg(g)) {
|
||||
nvgpu_err(g, "failed to set disable elpg");
|
||||
if (g->support_pmu) {
|
||||
if (nvgpu_cg_pg_disable(g) != 0) {
|
||||
nvgpu_warn(g, "fail to disable power mgmt");
|
||||
}
|
||||
}
|
||||
if (g->ops.clock_gating.slcg_gr_load_gating_prod) {
|
||||
g->ops.clock_gating.slcg_gr_load_gating_prod(g,
|
||||
false);
|
||||
}
|
||||
if (g->ops.clock_gating.slcg_perf_load_gating_prod) {
|
||||
g->ops.clock_gating.slcg_perf_load_gating_prod(g,
|
||||
false);
|
||||
}
|
||||
if (g->ops.clock_gating.slcg_ltc_load_gating_prod) {
|
||||
g->ops.clock_gating.slcg_ltc_load_gating_prod(g,
|
||||
false);
|
||||
}
|
||||
|
||||
gr_gk20a_init_cg_mode(g, ELCG_MODE, ELCG_RUN);
|
||||
|
||||
/* Disable fifo access */
|
||||
grfifo_ctl = gk20a_readl(g, gr_gpfifo_ctl_r());
|
||||
@@ -1842,8 +1827,10 @@ static bool gk20a_fifo_handle_mmu_fault_locked(
|
||||
gr_gpfifo_ctl_semaphore_access_enabled_f());
|
||||
|
||||
/* It is safe to enable ELPG again. */
|
||||
if (g->support_pmu && g->can_elpg) {
|
||||
nvgpu_pmu_enable_elpg(g);
|
||||
if (g->support_pmu) {
|
||||
if (nvgpu_cg_pg_enable(g) != 0) {
|
||||
nvgpu_warn(g, "fail to enable power mgmt");
|
||||
}
|
||||
}
|
||||
|
||||
return verbose;
|
||||
|
||||
@@ -43,6 +43,8 @@
|
||||
#include <nvgpu/utils.h>
|
||||
#include <nvgpu/channel.h>
|
||||
#include <nvgpu/unit.h>
|
||||
#include <nvgpu/power_features/pg.h>
|
||||
#include <nvgpu/power_features/cg.h>
|
||||
|
||||
#include "gk20a.h"
|
||||
#include "gr_gk20a.h"
|
||||
@@ -91,8 +93,6 @@ static void gr_gk20a_free_channel_patch_ctx(struct gk20a *g,
|
||||
/* golden ctx image */
|
||||
static int gr_gk20a_init_golden_ctx_image(struct gk20a *g,
|
||||
struct channel_gk20a *c);
|
||||
/*elcg init */
|
||||
static void gr_gk20a_enable_elcg(struct gk20a *g);
|
||||
|
||||
int gr_gk20a_get_ctx_id(struct gk20a *g,
|
||||
struct channel_gk20a *c,
|
||||
@@ -4227,33 +4227,6 @@ int gk20a_gr_zbc_set_table(struct gk20a *g, struct gr_gk20a *gr,
|
||||
gr_gk20a_add_zbc(g, gr, zbc_val));
|
||||
}
|
||||
|
||||
void gr_gk20a_init_cg_mode(struct gk20a *g, u32 cgmode, u32 mode_config)
|
||||
{
|
||||
u32 engine_idx;
|
||||
u32 active_engine_id = 0;
|
||||
struct fifo_engine_info_gk20a *engine_info = NULL;
|
||||
struct fifo_gk20a *f = &g->fifo;
|
||||
|
||||
for (engine_idx = 0; engine_idx < f->num_engines; ++engine_idx) {
|
||||
active_engine_id = f->active_engines_list[engine_idx];
|
||||
engine_info = &f->engine_info[active_engine_id];
|
||||
|
||||
/* gr_engine supports both BLCG and ELCG */
|
||||
if ((cgmode == BLCG_MODE) &&
|
||||
(engine_info->engine_enum == ENGINE_GR_GK20A)) {
|
||||
g->ops.therm.init_blcg_mode(g, mode_config, active_engine_id);
|
||||
break;
|
||||
} else if (cgmode == ELCG_MODE) {
|
||||
g->ops.therm.init_elcg_mode(g, mode_config,
|
||||
active_engine_id);
|
||||
} else {
|
||||
nvgpu_err(g, "invalid cg mode %d, config %d for "
|
||||
"act_eng_id %d",
|
||||
cgmode, mode_config, active_engine_id);
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
void gr_gk20a_program_zcull_mapping(struct gk20a *g, u32 zcull_num_entries,
|
||||
u32 *zcull_map_tiles)
|
||||
{
|
||||
@@ -4655,60 +4628,6 @@ out:
|
||||
return err;
|
||||
}
|
||||
|
||||
static void gr_gk20a_load_gating_prod(struct gk20a *g)
|
||||
{
|
||||
nvgpu_log_fn(g, " ");
|
||||
|
||||
/* slcg prod values */
|
||||
if (g->ops.clock_gating.slcg_bus_load_gating_prod) {
|
||||
g->ops.clock_gating.slcg_bus_load_gating_prod(g,
|
||||
g->slcg_enabled);
|
||||
}
|
||||
if (g->ops.clock_gating.slcg_chiplet_load_gating_prod) {
|
||||
g->ops.clock_gating.slcg_chiplet_load_gating_prod(g,
|
||||
g->slcg_enabled);
|
||||
}
|
||||
if (g->ops.clock_gating.slcg_gr_load_gating_prod) {
|
||||
g->ops.clock_gating.slcg_gr_load_gating_prod(g,
|
||||
g->slcg_enabled);
|
||||
}
|
||||
if (g->ops.clock_gating.slcg_ctxsw_firmware_load_gating_prod) {
|
||||
g->ops.clock_gating.slcg_ctxsw_firmware_load_gating_prod(g,
|
||||
g->slcg_enabled);
|
||||
}
|
||||
if (g->ops.clock_gating.slcg_perf_load_gating_prod) {
|
||||
g->ops.clock_gating.slcg_perf_load_gating_prod(g,
|
||||
g->slcg_enabled);
|
||||
}
|
||||
if (g->ops.clock_gating.slcg_xbar_load_gating_prod) {
|
||||
g->ops.clock_gating.slcg_xbar_load_gating_prod(g,
|
||||
g->slcg_enabled);
|
||||
}
|
||||
|
||||
/* blcg prod values */
|
||||
if (g->ops.clock_gating.blcg_bus_load_gating_prod) {
|
||||
g->ops.clock_gating.blcg_bus_load_gating_prod(g,
|
||||
g->blcg_enabled);
|
||||
}
|
||||
if (g->ops.clock_gating.blcg_gr_load_gating_prod) {
|
||||
g->ops.clock_gating.blcg_gr_load_gating_prod(g,
|
||||
g->blcg_enabled);
|
||||
}
|
||||
if (g->ops.clock_gating.blcg_ctxsw_firmware_load_gating_prod) {
|
||||
g->ops.clock_gating.blcg_ctxsw_firmware_load_gating_prod(g,
|
||||
g->blcg_enabled);
|
||||
}
|
||||
if (g->ops.clock_gating.blcg_xbar_load_gating_prod) {
|
||||
g->ops.clock_gating.blcg_xbar_load_gating_prod(g,
|
||||
g->blcg_enabled);
|
||||
}
|
||||
if (g->ops.clock_gating.pg_gr_load_gating_prod) {
|
||||
g->ops.clock_gating.pg_gr_load_gating_prod(g, true);
|
||||
}
|
||||
|
||||
nvgpu_log_fn(g, "done");
|
||||
}
|
||||
|
||||
static int gk20a_init_gr_prepare(struct gk20a *g)
|
||||
{
|
||||
u32 err = 0;
|
||||
@@ -4718,10 +4637,10 @@ static int gk20a_init_gr_prepare(struct gk20a *g)
|
||||
g->ops.mc.reset_mask(g, NVGPU_UNIT_BLG) |
|
||||
g->ops.mc.reset_mask(g, NVGPU_UNIT_PERFMON));
|
||||
|
||||
gr_gk20a_load_gating_prod(g);
|
||||
nvgpu_cg_init_gr_load_gating_prod(g);
|
||||
|
||||
/* Disable elcg until it gets enabled later in the init*/
|
||||
gr_gk20a_init_cg_mode(g, ELCG_MODE, ELCG_RUN);
|
||||
nvgpu_cg_elcg_disable(g);
|
||||
|
||||
/* enable fifo access */
|
||||
gk20a_writel(g, gr_gpfifo_ctl_r(),
|
||||
@@ -5041,7 +4960,7 @@ int gk20a_init_gr_support(struct gk20a *g)
|
||||
}
|
||||
}
|
||||
|
||||
gr_gk20a_enable_elcg(g);
|
||||
nvgpu_cg_elcg_enable(g);
|
||||
/* GR is inialized, signal possible waiters */
|
||||
g->gr.initialized = true;
|
||||
nvgpu_cond_signal(&g->gr.init_wq);
|
||||
@@ -5128,15 +5047,6 @@ int gk20a_enable_gr_hw(struct gk20a *g)
|
||||
return 0;
|
||||
}
|
||||
|
||||
static void gr_gk20a_enable_elcg(struct gk20a *g)
|
||||
{
|
||||
if (g->elcg_enabled) {
|
||||
gr_gk20a_init_cg_mode(g, ELCG_MODE, ELCG_AUTO);
|
||||
} else {
|
||||
gr_gk20a_init_cg_mode(g, ELCG_MODE, ELCG_RUN);
|
||||
}
|
||||
}
|
||||
|
||||
int gk20a_gr_reset(struct gk20a *g)
|
||||
{
|
||||
int err;
|
||||
@@ -5193,8 +5103,8 @@ int gk20a_gr_reset(struct gk20a *g)
|
||||
return err;
|
||||
}
|
||||
|
||||
gr_gk20a_load_gating_prod(g);
|
||||
gr_gk20a_enable_elcg(g);
|
||||
nvgpu_cg_init_gr_load_gating_prod(g);
|
||||
nvgpu_cg_elcg_enable(g);
|
||||
|
||||
return err;
|
||||
}
|
||||
|
||||
@@ -28,6 +28,7 @@
|
||||
|
||||
#include "gr_ctx_gk20a.h"
|
||||
#include "mm_gk20a.h"
|
||||
#include <nvgpu/power_features/pg.h>
|
||||
|
||||
#include <nvgpu/comptags.h>
|
||||
#include <nvgpu/cond.h>
|
||||
@@ -598,16 +599,16 @@ u32 gk20a_gr_get_sm_no_lock_down_hww_global_esr_mask(struct gk20a *g);
|
||||
#define gr_gk20a_elpg_protected_call(g, func) \
|
||||
({ \
|
||||
int err = 0; \
|
||||
if ((g->support_pmu) && (g->elpg_enabled)) {\
|
||||
err = nvgpu_pmu_disable_elpg(g); \
|
||||
if (g->support_pmu) {\
|
||||
err = nvgpu_pg_elpg_disable(g);\
|
||||
if (err != 0) {\
|
||||
nvgpu_pmu_enable_elpg(g); \
|
||||
err = nvgpu_pg_elpg_enable(g); \
|
||||
} \
|
||||
} \
|
||||
if (err == 0) { \
|
||||
err = func; \
|
||||
if ((g->support_pmu) && (g->elpg_enabled)) {\
|
||||
nvgpu_pmu_enable_elpg(g); \
|
||||
if (g->support_pmu) {\
|
||||
(void)nvgpu_pg_elpg_enable(g); \
|
||||
} \
|
||||
} \
|
||||
err; \
|
||||
|
||||
@@ -106,7 +106,7 @@ int gm20b_pmu_setup_elpg(struct gk20a *g)
|
||||
|
||||
nvgpu_log_fn(g, " ");
|
||||
|
||||
if (g->elpg_enabled) {
|
||||
if (g->can_elpg && g->elpg_enabled) {
|
||||
reg_writes = ((sizeof(_pginitseq_gm20b) /
|
||||
sizeof((_pginitseq_gm20b)[0])));
|
||||
/* Initialize registers with production values*/
|
||||
|
||||
@@ -282,7 +282,7 @@ int gp10b_pmu_setup_elpg(struct gk20a *g)
|
||||
|
||||
nvgpu_log_fn(g, " ");
|
||||
|
||||
if (g->elpg_enabled) {
|
||||
if (g->can_elpg && g->elpg_enabled) {
|
||||
reg_writes = ((sizeof(_pginitseq_gp10b) /
|
||||
sizeof((_pginitseq_gp10b)[0])));
|
||||
/* Initialize registers with production values*/
|
||||
|
||||
@@ -43,6 +43,8 @@
|
||||
#include <nvgpu/gk20a.h>
|
||||
#include <nvgpu/channel.h>
|
||||
#include <nvgpu/unit.h>
|
||||
#include <nvgpu/power_features/cg.h>
|
||||
#include <nvgpu/power_features/power_features.h>
|
||||
|
||||
#include "gk20a/fifo_gk20a.h"
|
||||
|
||||
@@ -1095,25 +1097,11 @@ void gv11b_fifo_teardown_ch_tsg(struct gk20a *g, u32 act_eng_bitmask,
|
||||
g->fifo.deferred_reset_pending = false;
|
||||
|
||||
/* Disable power management */
|
||||
if (g->support_pmu && g->elpg_enabled) {
|
||||
if (nvgpu_pmu_disable_elpg(g)) {
|
||||
nvgpu_err(g, "failed to set disable elpg");
|
||||
if (g->support_pmu) {
|
||||
if (nvgpu_cg_pg_disable(g) != 0) {
|
||||
nvgpu_warn(g, "fail to disable power mgmt");
|
||||
}
|
||||
}
|
||||
if (g->ops.clock_gating.slcg_gr_load_gating_prod) {
|
||||
g->ops.clock_gating.slcg_gr_load_gating_prod(g,
|
||||
false);
|
||||
}
|
||||
if (g->ops.clock_gating.slcg_perf_load_gating_prod) {
|
||||
g->ops.clock_gating.slcg_perf_load_gating_prod(g,
|
||||
false);
|
||||
}
|
||||
if (g->ops.clock_gating.slcg_ltc_load_gating_prod) {
|
||||
g->ops.clock_gating.slcg_ltc_load_gating_prod(g,
|
||||
false);
|
||||
}
|
||||
|
||||
gr_gk20a_init_cg_mode(g, ELCG_MODE, ELCG_RUN);
|
||||
|
||||
if (rc_type == RC_TYPE_MMU_FAULT) {
|
||||
gk20a_debug_dump(g);
|
||||
@@ -1220,8 +1208,10 @@ void gv11b_fifo_teardown_ch_tsg(struct gk20a *g, u32 act_eng_bitmask,
|
||||
gk20a_fifo_set_runlist_state(g, runlists_mask, RUNLIST_ENABLED);
|
||||
|
||||
/* It is safe to enable ELPG again. */
|
||||
if (g->support_pmu && g->elpg_enabled) {
|
||||
nvgpu_pmu_enable_elpg(g);
|
||||
if (g->support_pmu) {
|
||||
if (nvgpu_cg_pg_enable(g) != 0) {
|
||||
nvgpu_warn(g, "fail to enable power mgmt");
|
||||
}
|
||||
}
|
||||
|
||||
g->ops.fifo.teardown_unmask_intr(g);
|
||||
@@ -1312,18 +1302,11 @@ int gv11b_init_fifo_reset_enable_hw(struct gk20a *g)
|
||||
/* enable pmc pfifo */
|
||||
g->ops.mc.reset(g, g->ops.mc.reset_mask(g, NVGPU_UNIT_FIFO));
|
||||
|
||||
if (g->ops.clock_gating.slcg_ce2_load_gating_prod) {
|
||||
g->ops.clock_gating.slcg_ce2_load_gating_prod(g,
|
||||
g->slcg_enabled);
|
||||
}
|
||||
if (g->ops.clock_gating.slcg_fifo_load_gating_prod) {
|
||||
g->ops.clock_gating.slcg_fifo_load_gating_prod(g,
|
||||
g->slcg_enabled);
|
||||
}
|
||||
if (g->ops.clock_gating.blcg_fifo_load_gating_prod) {
|
||||
g->ops.clock_gating.blcg_fifo_load_gating_prod(g,
|
||||
g->blcg_enabled);
|
||||
}
|
||||
nvgpu_cg_slcg_ce2_load_enable(g);
|
||||
|
||||
nvgpu_cg_slcg_fifo_load_enable(g);
|
||||
|
||||
nvgpu_cg_blcg_fifo_load_enable(g);
|
||||
|
||||
timeout = gk20a_readl(g, fifo_fb_timeout_r());
|
||||
nvgpu_log_info(g, "fifo_fb_timeout reg val = 0x%08x", timeout);
|
||||
|
||||
@@ -124,7 +124,7 @@ int gv11b_pmu_setup_elpg(struct gk20a *g)
|
||||
|
||||
nvgpu_log_fn(g, " ");
|
||||
|
||||
if (g->elpg_enabled) {
|
||||
if (g->can_elpg && g->elpg_enabled) {
|
||||
reg_writes = ((sizeof(_pginitseq_gv11b) /
|
||||
sizeof((_pginitseq_gv11b)[0])));
|
||||
/* Initialize registers with production values*/
|
||||
|
||||
@@ -1490,6 +1490,7 @@ struct gk20a {
|
||||
u32 max_timeslice_us;
|
||||
bool runlist_interleave;
|
||||
|
||||
struct nvgpu_mutex cg_pg_lock;
|
||||
bool slcg_enabled;
|
||||
bool blcg_enabled;
|
||||
bool elcg_enabled;
|
||||
|
||||
55
drivers/gpu/nvgpu/include/nvgpu/power_features/cg.h
Normal file
55
drivers/gpu/nvgpu/include/nvgpu/power_features/cg.h
Normal file
@@ -0,0 +1,55 @@
|
||||
/*
|
||||
* Copyright (c) 2019, NVIDIA CORPORATION. All rights reserved.
|
||||
*
|
||||
* Permission is hereby granted, free of charge, to any person obtaining a
|
||||
* copy of this software and associated documentation files (the "Software"),
|
||||
* to deal in the Software without restriction, including without limitation
|
||||
* the rights to use, copy, modify, merge, publish, distribute, sublicense,
|
||||
* and/or sell copies of the Software, and to permit persons to whom the
|
||||
* Software is furnished to do so, subject to the following conditions:
|
||||
*
|
||||
* The above copyright notice and this permission notice shall be included in
|
||||
* all copies or substantial portions of the Software.
|
||||
*
|
||||
* THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
|
||||
* IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
|
||||
* FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
|
||||
* THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
|
||||
* LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING
|
||||
* FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER
|
||||
* DEALINGS IN THE SOFTWARE.
|
||||
*/
|
||||
|
||||
|
||||
#ifndef NVGPU_POWER_FEATURES_CG_H
|
||||
#define NVGPU_POWER_FEATURES_CG_H
|
||||
|
||||
#include <nvgpu/types.h>
|
||||
|
||||
struct gk20a;
|
||||
struct fifo_gk20a;
|
||||
|
||||
void nvgpu_cg_init_gr_load_gating_prod(struct gk20a *g);
|
||||
void nvgpu_cg_elcg_enable(struct gk20a *g);
|
||||
void nvgpu_cg_elcg_disable(struct gk20a *g);
|
||||
void nvgpu_cg_elcg_set_elcg_enabled(struct gk20a *g, bool enable);
|
||||
|
||||
void nvgpu_cg_blcg_mode_enable(struct gk20a *g);
|
||||
void nvgpu_cg_blcg_mode_disable(struct gk20a *g);
|
||||
void nvgpu_cg_blcg_fb_ltc_load_enable(struct gk20a *g);
|
||||
void nvgpu_cg_blcg_fifo_load_enable(struct gk20a *g);
|
||||
void nvgpu_cg_blcg_pmu_load_enable(struct gk20a *g);
|
||||
void nvgpu_cg_blcg_ce_load_enable(struct gk20a *g);
|
||||
void nvgpu_cg_blcg_gr_load_enable(struct gk20a *g);
|
||||
void nvgpu_cg_blcg_set_blcg_enabled(struct gk20a *g, bool enable);
|
||||
|
||||
void nvgpu_cg_slcg_gr_perf_ltc_load_enable(struct gk20a *g);
|
||||
void nvgpu_cg_slcg_gr_perf_ltc_load_disable(struct gk20a *g);
|
||||
void nvgpu_cg_slcg_fb_ltc_load_enable(struct gk20a *g);
|
||||
void nvgpu_cg_slcg_priring_load_enable(struct gk20a *g);
|
||||
void nvgpu_cg_slcg_fifo_load_enable(struct gk20a *g);
|
||||
void nvgpu_cg_slcg_pmu_load_enable(struct gk20a *g);
|
||||
void nvgpu_cg_slcg_ce2_load_enable(struct gk20a *g);
|
||||
void nvgpu_cg_slcg_set_slcg_enabled(struct gk20a *g, bool enable);
|
||||
|
||||
#endif /*NVGPU_POWER_FEATURES_CG_H*/
|
||||
36
drivers/gpu/nvgpu/include/nvgpu/power_features/pg.h
Normal file
36
drivers/gpu/nvgpu/include/nvgpu/power_features/pg.h
Normal file
@@ -0,0 +1,36 @@
|
||||
/*
|
||||
* Copyright (c) 2019, NVIDIA CORPORATION. All rights reserved.
|
||||
*
|
||||
* Permission is hereby granted, free of charge, to any person obtaining a
|
||||
* copy of this software and associated documentation files (the "Software"),
|
||||
* to deal in the Software without restriction, including without limitation
|
||||
* the rights to use, copy, modify, merge, publish, distribute, sublicense,
|
||||
* and/or sell copies of the Software, and to permit persons to whom the
|
||||
* Software is furnished to do so, subject to the following conditions:
|
||||
*
|
||||
* The above copyright notice and this permission notice shall be included in
|
||||
* all copies or substantial portions of the Software.
|
||||
*
|
||||
* THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
|
||||
* IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
|
||||
* FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
|
||||
* THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
|
||||
* LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING
|
||||
* FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER
|
||||
* DEALINGS IN THE SOFTWARE.
|
||||
*/
|
||||
|
||||
|
||||
#ifndef NVGPU_POWER_FEATURES_PG_H
|
||||
#define NVGPU_POWER_FEATURES_PG_H
|
||||
|
||||
#include <nvgpu/types.h>
|
||||
|
||||
struct gk20a;
|
||||
|
||||
int nvgpu_pg_elpg_disable(struct gk20a *g);
|
||||
int nvgpu_pg_elpg_enable(struct gk20a *g);
|
||||
bool nvgpu_pg_elpg_is_enabled(struct gk20a *g);
|
||||
int nvgpu_pg_elpg_set_elpg_enabled(struct gk20a *g, bool enable);
|
||||
|
||||
#endif /*NVGPU_POWER_FEATURES_PG_H*/
|
||||
@@ -0,0 +1,34 @@
|
||||
/*
|
||||
* Copyright (c) 2019, NVIDIA CORPORATION. All rights reserved.
|
||||
*
|
||||
* Permission is hereby granted, free of charge, to any person obtaining a
|
||||
* copy of this software and associated documentation files (the "Software"),
|
||||
* to deal in the Software without restriction, including without limitation
|
||||
* the rights to use, copy, modify, merge, publish, distribute, sublicense,
|
||||
* and/or sell copies of the Software, and to permit persons to whom the
|
||||
* Software is furnished to do so, subject to the following conditions:
|
||||
*
|
||||
* The above copyright notice and this permission notice shall be included in
|
||||
* all copies or substantial portions of the Software.
|
||||
*
|
||||
* THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
|
||||
* IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
|
||||
* FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
|
||||
* THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
|
||||
* LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING
|
||||
* FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER
|
||||
* DEALINGS IN THE SOFTWARE.
|
||||
*/
|
||||
|
||||
|
||||
#ifndef NVGPU_POWER_FEATURES_H
|
||||
#define NVGPU_POWER_FEATURES_H
|
||||
|
||||
#include <nvgpu/types.h>
|
||||
|
||||
struct gk20a;
|
||||
|
||||
int nvgpu_cg_pg_disable(struct gk20a *g);
|
||||
int nvgpu_cg_pg_enable(struct gk20a *g);
|
||||
|
||||
#endif /*NVGPU_POWER_FEATURES_H*/
|
||||
@@ -65,6 +65,7 @@ static void nvgpu_init_vars(struct gk20a *g)
|
||||
nvgpu_mutex_init(&g->ctxsw_disable_lock);
|
||||
nvgpu_mutex_init(&g->tpc_pg_lock);
|
||||
nvgpu_mutex_init(&g->clk_arb_enable_lock);
|
||||
nvgpu_mutex_init(&g->cg_pg_lock);
|
||||
|
||||
/* Init the clock req count to 0 */
|
||||
nvgpu_atomic_set(&g->clk_arb_global_nr, 0);
|
||||
|
||||
@@ -21,6 +21,8 @@
|
||||
#include <nvgpu/kmem.h>
|
||||
#include <nvgpu/nvhost.h>
|
||||
#include <nvgpu/ptimer.h>
|
||||
#include <nvgpu/power_features/cg.h>
|
||||
#include <nvgpu/power_features/pg.h>
|
||||
|
||||
#include "os_linux.h"
|
||||
#include "sysfs.h"
|
||||
@@ -49,16 +51,14 @@ static ssize_t elcg_enable_store(struct device *dev,
|
||||
return err;
|
||||
|
||||
if (val) {
|
||||
g->elcg_enabled = true;
|
||||
gr_gk20a_init_cg_mode(g, ELCG_MODE, ELCG_AUTO);
|
||||
nvgpu_cg_elcg_set_elcg_enabled(g, true);
|
||||
} else {
|
||||
g->elcg_enabled = false;
|
||||
gr_gk20a_init_cg_mode(g, ELCG_MODE, ELCG_RUN);
|
||||
nvgpu_cg_elcg_set_elcg_enabled(g, false);
|
||||
}
|
||||
|
||||
gk20a_idle(g);
|
||||
|
||||
nvgpu_info(g, "ELCG is %s.", g->elcg_enabled ? "enabled" :
|
||||
nvgpu_info(g, "ELCG is %s.", val ? "enabled" :
|
||||
"disabled");
|
||||
|
||||
return count;
|
||||
@@ -84,45 +84,19 @@ static ssize_t blcg_enable_store(struct device *dev,
|
||||
if (kstrtoul(buf, 10, &val) < 0)
|
||||
return -EINVAL;
|
||||
|
||||
if (val)
|
||||
g->blcg_enabled = true;
|
||||
else
|
||||
g->blcg_enabled = false;
|
||||
|
||||
err = gk20a_busy(g);
|
||||
if (err)
|
||||
return err;
|
||||
|
||||
if (g->ops.clock_gating.blcg_bus_load_gating_prod)
|
||||
g->ops.clock_gating.blcg_bus_load_gating_prod(g,
|
||||
g->blcg_enabled);
|
||||
if (g->ops.clock_gating.blcg_ce_load_gating_prod)
|
||||
g->ops.clock_gating.blcg_ce_load_gating_prod(g,
|
||||
g->blcg_enabled);
|
||||
if (g->ops.clock_gating.blcg_ctxsw_firmware_load_gating_prod)
|
||||
g->ops.clock_gating.blcg_ctxsw_firmware_load_gating_prod(g,
|
||||
g->blcg_enabled);
|
||||
if (g->ops.clock_gating.blcg_fb_load_gating_prod)
|
||||
g->ops.clock_gating.blcg_fb_load_gating_prod(g,
|
||||
g->blcg_enabled);
|
||||
if (g->ops.clock_gating.blcg_fifo_load_gating_prod)
|
||||
g->ops.clock_gating.blcg_fifo_load_gating_prod(g,
|
||||
g->blcg_enabled);
|
||||
if (g->ops.clock_gating.blcg_gr_load_gating_prod)
|
||||
g->ops.clock_gating.blcg_gr_load_gating_prod(g,
|
||||
g->blcg_enabled);
|
||||
if (g->ops.clock_gating.blcg_ltc_load_gating_prod)
|
||||
g->ops.clock_gating.blcg_ltc_load_gating_prod(g,
|
||||
g->blcg_enabled);
|
||||
if (g->ops.clock_gating.blcg_pmu_load_gating_prod)
|
||||
g->ops.clock_gating.blcg_pmu_load_gating_prod(g,
|
||||
g->blcg_enabled);
|
||||
if (g->ops.clock_gating.blcg_xbar_load_gating_prod)
|
||||
g->ops.clock_gating.blcg_xbar_load_gating_prod(g,
|
||||
g->blcg_enabled);
|
||||
if (val) {
|
||||
nvgpu_cg_blcg_set_blcg_enabled(g, true);
|
||||
} else {
|
||||
nvgpu_cg_blcg_set_blcg_enabled(g, false);
|
||||
}
|
||||
|
||||
gk20a_idle(g);
|
||||
|
||||
nvgpu_info(g, "BLCG is %s.", g->blcg_enabled ? "enabled" :
|
||||
nvgpu_info(g, "BLCG is %s.", val ? "enabled" :
|
||||
"disabled");
|
||||
|
||||
return count;
|
||||
@@ -149,59 +123,25 @@ static ssize_t slcg_enable_store(struct device *dev,
|
||||
if (kstrtoul(buf, 10, &val) < 0)
|
||||
return -EINVAL;
|
||||
|
||||
if (val)
|
||||
g->slcg_enabled = true;
|
||||
else
|
||||
g->slcg_enabled = false;
|
||||
err = gk20a_busy(g);
|
||||
if (err) {
|
||||
return err;
|
||||
}
|
||||
|
||||
if (val) {
|
||||
nvgpu_cg_slcg_set_slcg_enabled(g, true);
|
||||
} else {
|
||||
nvgpu_cg_slcg_set_slcg_enabled(g, false);
|
||||
}
|
||||
|
||||
/*
|
||||
* TODO: slcg_therm_load_gating is not enabled anywhere during
|
||||
* init. Therefore, it would be incongruous to add it here. Once
|
||||
* it is added to init, we should add it here too.
|
||||
*/
|
||||
err = gk20a_busy(g);
|
||||
if (err)
|
||||
return err;
|
||||
|
||||
if (g->ops.clock_gating.slcg_bus_load_gating_prod)
|
||||
g->ops.clock_gating.slcg_bus_load_gating_prod(g,
|
||||
g->slcg_enabled);
|
||||
if (g->ops.clock_gating.slcg_ce2_load_gating_prod)
|
||||
g->ops.clock_gating.slcg_ce2_load_gating_prod(g,
|
||||
g->slcg_enabled);
|
||||
if (g->ops.clock_gating.slcg_chiplet_load_gating_prod)
|
||||
g->ops.clock_gating.slcg_chiplet_load_gating_prod(g,
|
||||
g->slcg_enabled);
|
||||
if (g->ops.clock_gating.slcg_ctxsw_firmware_load_gating_prod)
|
||||
g->ops.clock_gating.slcg_ctxsw_firmware_load_gating_prod(g,
|
||||
g->slcg_enabled);
|
||||
if (g->ops.clock_gating.slcg_fb_load_gating_prod)
|
||||
g->ops.clock_gating.slcg_fb_load_gating_prod(g,
|
||||
g->slcg_enabled);
|
||||
if (g->ops.clock_gating.slcg_fifo_load_gating_prod)
|
||||
g->ops.clock_gating.slcg_fifo_load_gating_prod(g,
|
||||
g->slcg_enabled);
|
||||
if (g->ops.clock_gating.slcg_gr_load_gating_prod)
|
||||
g->ops.clock_gating.slcg_gr_load_gating_prod(g,
|
||||
g->slcg_enabled);
|
||||
if (g->ops.clock_gating.slcg_ltc_load_gating_prod)
|
||||
g->ops.clock_gating.slcg_ltc_load_gating_prod(g,
|
||||
g->slcg_enabled);
|
||||
if (g->ops.clock_gating.slcg_perf_load_gating_prod)
|
||||
g->ops.clock_gating.slcg_perf_load_gating_prod(g,
|
||||
g->slcg_enabled);
|
||||
if (g->ops.clock_gating.slcg_priring_load_gating_prod)
|
||||
g->ops.clock_gating.slcg_priring_load_gating_prod(g,
|
||||
g->slcg_enabled);
|
||||
if (g->ops.clock_gating.slcg_pmu_load_gating_prod)
|
||||
g->ops.clock_gating.slcg_pmu_load_gating_prod(g,
|
||||
g->slcg_enabled);
|
||||
if (g->ops.clock_gating.slcg_xbar_load_gating_prod)
|
||||
g->ops.clock_gating.slcg_xbar_load_gating_prod(g,
|
||||
g->slcg_enabled);
|
||||
gk20a_idle(g);
|
||||
|
||||
nvgpu_info(g, "SLCG is %s.", g->slcg_enabled ? "enabled" :
|
||||
nvgpu_info(g, "SLCG is %s.", val ? "enabled" :
|
||||
"disabled");
|
||||
|
||||
return count;
|
||||
@@ -474,7 +414,7 @@ static ssize_t elpg_enable_store(struct device *dev,
|
||||
return -EINVAL;
|
||||
|
||||
if (!g->power_on) {
|
||||
g->elpg_enabled = val ? true : false;
|
||||
return -EINVAL;
|
||||
} else {
|
||||
err = gk20a_busy(g);
|
||||
if (err)
|
||||
@@ -483,25 +423,14 @@ static ssize_t elpg_enable_store(struct device *dev,
|
||||
* Since elpg is refcounted, we should not unnecessarily call
|
||||
* enable/disable if it is already so.
|
||||
*/
|
||||
if (val && !g->elpg_enabled) {
|
||||
g->elpg_enabled = true;
|
||||
nvgpu_pmu_pg_global_enable(g, true);
|
||||
|
||||
} else if (!val && g->elpg_enabled) {
|
||||
if (g->ops.pmu.pmu_pg_engines_feature_list &&
|
||||
g->ops.pmu.pmu_pg_engines_feature_list(g,
|
||||
PMU_PG_ELPG_ENGINE_ID_GRAPHICS) !=
|
||||
NVGPU_PMU_GR_FEATURE_MASK_POWER_GATING) {
|
||||
nvgpu_pmu_pg_global_enable(g, false);
|
||||
g->elpg_enabled = false;
|
||||
if (val != 0) {
|
||||
nvgpu_pg_elpg_set_elpg_enabled(g, true);
|
||||
} else {
|
||||
g->elpg_enabled = false;
|
||||
nvgpu_pmu_pg_global_enable(g, false);
|
||||
}
|
||||
nvgpu_pg_elpg_set_elpg_enabled(g, false);
|
||||
}
|
||||
gk20a_idle(g);
|
||||
}
|
||||
nvgpu_info(g, "ELPG is %s.", g->elpg_enabled ? "enabled" :
|
||||
nvgpu_info(g, "ELPG is %s.", val ? "enabled" :
|
||||
"disabled");
|
||||
|
||||
return count;
|
||||
@@ -512,7 +441,8 @@ static ssize_t elpg_enable_read(struct device *dev,
|
||||
{
|
||||
struct gk20a *g = get_gk20a(dev);
|
||||
|
||||
return snprintf(buf, PAGE_SIZE, "%d\n", g->elpg_enabled ? 1 : 0);
|
||||
return snprintf(buf, PAGE_SIZE, "%d\n",
|
||||
nvgpu_pg_elpg_is_enabled(g) ? 1 : 0);
|
||||
}
|
||||
|
||||
static DEVICE_ATTR(elpg_enable, ROOTRW, elpg_enable_read, elpg_enable_store);
|
||||
@@ -610,8 +540,9 @@ static ssize_t mscg_enable_store(struct device *dev,
|
||||
/* make status visible */
|
||||
smp_mb();
|
||||
g->mscg_enabled = false;
|
||||
if (g->elpg_enabled)
|
||||
nvgpu_pmu_pg_global_enable(g, true);
|
||||
if (nvgpu_pg_elpg_is_enabled(g)) {
|
||||
nvgpu_pg_elpg_enable(g);
|
||||
}
|
||||
}
|
||||
g->mscg_enabled = false;
|
||||
}
|
||||
|
||||
@@ -69,6 +69,7 @@ static void vgpu_init_vars(struct gk20a *g, struct gk20a_platform *platform)
|
||||
nvgpu_mutex_init(&g->power_lock);
|
||||
nvgpu_mutex_init(&g->ctxsw_disable_lock);
|
||||
nvgpu_mutex_init(&g->clk_arb_enable_lock);
|
||||
nvgpu_mutex_init(&g->cg_pg_lock);
|
||||
|
||||
nvgpu_mutex_init(&priv->vgpu_clk_get_freq_lock);
|
||||
|
||||
|
||||
Reference in New Issue
Block a user