Files
linux-nvgpu/drivers/gpu/nvgpu/common/pmu/lpwr/lpwr.c
Mahantesh Kumbar 4bb9b0b987 gpu: nvgpu: use support_ls_pmu flag to check LS PMU support
Currently PMU support enable check is done with multiple
methods which added complexity to know status of PMU
support.

Changed to replace multiple methods with support_pmu
flag to know the PMU support, support_pmu will be updated
at init stage based on platform/chip specific settings
to know the PMU support status.

Cleaned up support_pmu flag check with platform specific
PMU members in multiple places & moved check to
public functions

JIRA NVGPU-173

Change-Id: Ief2c64250d1f78e3b054203be56499e4d1d9b046
Signed-off-by: Mahantesh Kumbar <mkumbar@nvidia.com>
Reviewed-on: https://git-master.nvidia.com/r/2024024
Reviewed-by: svc-mobile-coverity <svc-mobile-coverity@nvidia.com>
Reviewed-by: svc-mobile-misra <svc-mobile-misra@nvidia.com>
Reviewed-by: svc-misra-checker <svc-misra-checker@nvidia.com>
GVS: Gerrit_Virtual_Submit
Reviewed-by: Vijayakumar Subbu <vsubbu@nvidia.com>
Reviewed-by: mobile promotions <svcmobile_promotions@nvidia.com>
Tested-by: mobile promotions <svcmobile_promotions@nvidia.com>
2019-03-04 03:33:16 -08:00

456 lines
11 KiB
C

/*
* Copyright (c) 2016-2019, NVIDIA CORPORATION. All rights reserved.
*
* Permission is hereby granted, free of charge, to any person obtaining a
* copy of this software and associated documentation files (the "Software"),
* to deal in the Software without restriction, including without limitation
* the rights to use, copy, modify, merge, publish, distribute, sublicense,
* and/or sell copies of the Software, and to permit persons to whom the
* Software is furnished to do so, subject to the following conditions:
*
* The above copyright notice and this permission notice shall be included in
* all copies or substantial portions of the Software.
*
* THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
* IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
* FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
* THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
* LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING
* FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER
* DEALINGS IN THE SOFTWARE.
*/
#include <nvgpu/bios.h>
#include <nvgpu/pmu.h>
#include <nvgpu/clk_arb.h>
#include <nvgpu/gk20a.h>
#include <nvgpu/string.h>
#include <nvgpu/pmu/pstate.h>
#include <nvgpu/pmu/perf.h>
#include "gp106/bios_gp106.h"
#include "lpwr.h"
static int get_lpwr_idx_table(struct gk20a *g)
{
u8 *lpwr_idx_table_ptr;
u8 *entry_addr;
u32 idx;
struct nvgpu_lpwr_bios_idx_data *pidx_data =
&g->perf_pmu->lpwr.lwpr_bios_data.idx;
struct nvgpu_bios_lpwr_idx_table_1x_header header = { 0 };
struct nvgpu_bios_lpwr_idx_table_1x_entry entry = { 0 };
lpwr_idx_table_ptr = (u8 *)nvgpu_bios_get_perf_table_ptrs(g,
g->bios.perf_token, LOWPOWER_TABLE);
if (lpwr_idx_table_ptr == NULL) {
return -EINVAL;
}
nvgpu_memcpy((u8 *)&header, lpwr_idx_table_ptr,
sizeof(struct nvgpu_bios_lpwr_idx_table_1x_header));
if (header.entry_count >= LPWR_VBIOS_IDX_ENTRY_COUNT_MAX) {
return -EINVAL;
}
pidx_data->base_sampling_period = (u16)header.base_sampling_period;
/* Parse the LPWR Index Table entries.*/
for (idx = 0; idx < header.entry_count; idx++) {
entry_addr = lpwr_idx_table_ptr + header.header_size +
(idx * header.entry_size);
nvgpu_memcpy((u8 *)&entry, entry_addr,
sizeof(struct nvgpu_bios_lpwr_idx_table_1x_entry));
pidx_data->entry[idx].pcie_idx = entry.pcie_idx;
pidx_data->entry[idx].gr_idx = entry.gr_idx;
pidx_data->entry[idx].ms_idx = entry.ms_idx;
pidx_data->entry[idx].di_idx = entry.di_idx;
pidx_data->entry[idx].gc6_idx = entry.gc6_idx;
}
return 0;
}
static int get_lpwr_gr_table(struct gk20a *g)
{
u8 *lpwr_gr_table_ptr;
u8 *entry_addr;
u32 idx;
struct nvgpu_lpwr_bios_gr_data *pgr_data =
&g->perf_pmu->lpwr.lwpr_bios_data.gr;
struct nvgpu_bios_lpwr_gr_table_1x_header header = { 0 };
struct nvgpu_bios_lpwr_gr_table_1x_entry entry = { 0 };
lpwr_gr_table_ptr = (u8 *)nvgpu_bios_get_perf_table_ptrs(g,
g->bios.perf_token, LOWPOWER_GR_TABLE);
if (lpwr_gr_table_ptr == NULL) {
return -EINVAL;
}
nvgpu_memcpy((u8 *)&header, lpwr_gr_table_ptr,
sizeof(struct nvgpu_bios_lpwr_gr_table_1x_header));
/* Parse the LPWR Index Table entries.*/
for (idx = 0; idx < header.entry_count; idx++) {
entry_addr = lpwr_gr_table_ptr + header.header_size +
(idx * header.entry_size);
nvgpu_memcpy((u8 *)&entry, entry_addr,
sizeof(struct nvgpu_bios_lpwr_gr_table_1x_entry));
if (BIOS_GET_FIELD(bool, entry.feautre_mask,
NV_VBIOS_LPWR_MS_FEATURE_MASK_MS)) {
pgr_data->entry[idx].gr_enabled = true;
pgr_data->entry[idx].feature_mask =
NVGPU_PMU_GR_FEATURE_MASK_ALL;
if (!BIOS_GET_FIELD(bool, entry.feautre_mask,
NV_VBIOS_LPWR_GR_FEATURE_MASK_GR_RPPG)) {
pgr_data->entry[idx].feature_mask &=
~NVGPU_PMU_GR_FEATURE_MASK_RPPG;
}
}
}
return 0;
}
static int get_lpwr_ms_table(struct gk20a *g)
{
u8 *lpwr_ms_table_ptr;
u8 *entry_addr;
u32 idx;
struct nvgpu_lpwr_bios_ms_data *pms_data =
&g->perf_pmu->lpwr.lwpr_bios_data.ms;
struct nvgpu_bios_lpwr_ms_table_1x_header header = { 0 };
struct nvgpu_bios_lpwr_ms_table_1x_entry entry = { 0 };
lpwr_ms_table_ptr = (u8 *)nvgpu_bios_get_perf_table_ptrs(g,
g->bios.perf_token, LOWPOWER_MS_TABLE);
if (lpwr_ms_table_ptr == NULL) {
return -EINVAL;
}
nvgpu_memcpy((u8 *)&header, lpwr_ms_table_ptr,
sizeof(struct nvgpu_bios_lpwr_ms_table_1x_header));
if (header.entry_count >= LPWR_VBIOS_MS_ENTRY_COUNT_MAX) {
return -EINVAL;
}
pms_data->default_entry_idx = (u8)header.default_entry_idx;
pms_data->idle_threshold_us = U32(header.idle_threshold_us) * U32(10);
/* Parse the LPWR MS Table entries.*/
for (idx = 0; idx < header.entry_count; idx++) {
entry_addr = lpwr_ms_table_ptr + header.header_size +
(idx * header.entry_size);
nvgpu_memcpy((u8 *)&entry, entry_addr,
sizeof(struct nvgpu_bios_lpwr_ms_table_1x_entry));
if (BIOS_GET_FIELD(bool, entry.feautre_mask,
NV_VBIOS_LPWR_MS_FEATURE_MASK_MS)) {
pms_data->entry[idx].ms_enabled = true;
pms_data->entry[idx].feature_mask =
NVGPU_PMU_MS_FEATURE_MASK_ALL;
if (!BIOS_GET_FIELD(bool, entry.feautre_mask,
NV_VBIOS_LPWR_MS_FEATURE_MASK_MS_CLOCK_GATING)) {
pms_data->entry[idx].feature_mask &=
~NVGPU_PMU_MS_FEATURE_MASK_CLOCK_GATING;
}
if (!BIOS_GET_FIELD(bool, entry.feautre_mask,
NV_VBIOS_LPWR_MS_FEATURE_MASK_MS_SWASR)) {
pms_data->entry[idx].feature_mask &=
~NVGPU_PMU_MS_FEATURE_MASK_SW_ASR;
}
if (!BIOS_GET_FIELD(bool, entry.feautre_mask,
NV_VBIOS_LPWR_MS_FEATURE_MASK_MS_RPPG)) {
pms_data->entry[idx].feature_mask &=
~NVGPU_PMU_MS_FEATURE_MASK_RPPG;
}
}
pms_data->entry[idx].dynamic_current_logic =
entry.dynamic_current_logic;
pms_data->entry[idx].dynamic_current_sram =
entry.dynamic_current_sram;
}
return 0;
}
int nvgpu_lpwr_pg_setup(struct gk20a *g)
{
int err = 0;
nvgpu_log_fn(g, " ");
err = get_lpwr_gr_table(g);
if (err != 0) {
return err;
}
err = get_lpwr_ms_table(g);
if (err != 0) {
return err;
}
err = get_lpwr_idx_table(g);
return err;
}
static void nvgpu_pmu_handle_param_lpwr_msg(struct gk20a *g,
struct pmu_msg *msg, void *param,
u32 handle, u32 status)
{
u32 *ack_status = param;
nvgpu_log_fn(g, " ");
if (status != 0U) {
nvgpu_err(g, "LWPR PARAM cmd aborted");
return;
}
*ack_status = 1;
nvgpu_pmu_dbg(g, "lpwr-param is acknowledged from PMU %x",
msg->msg.pg.msg_type);
}
int nvgpu_lwpr_mclk_change(struct gk20a *g, u32 pstate)
{
struct pmu_cmd cmd;
u32 seq;
int status = 0;
u32 payload = NV_PMU_PG_PARAM_MCLK_CHANGE_MS_SWASR_ENABLED;
struct clk_set_info *pstate_info;
u32 ack_status = 0;
nvgpu_log_fn(g, " ");
pstate_info = pstate_get_clk_set_info(g, pstate,
CLKWHICH_MCLK);
if (pstate_info == NULL) {
return -EINVAL;
}
if (pstate_info->max_mhz >
MAX_SWASR_MCLK_FREQ_WITHOUT_WR_TRAINING_MAXWELL_MHZ) {
payload |=
NV_PMU_PG_PARAM_MCLK_CHANGE_GDDR5_WR_TRAINING_ENABLED;
}
if (payload != g->perf_pmu->lpwr.mclk_change_cache) {
size_t tmp_size = PMU_CMD_HDR_SIZE +
sizeof(struct pmu_pg_cmd_mclk_change);
g->perf_pmu->lpwr.mclk_change_cache = payload;
cmd.hdr.unit_id = PMU_UNIT_PG;
nvgpu_assert(tmp_size <= U8_MAX);
cmd.hdr.size = (u8)(tmp_size);
cmd.cmd.pg.mclk_change.cmd_type =
PMU_PG_CMD_ID_PG_PARAM;
cmd.cmd.pg.mclk_change.cmd_id =
PMU_PG_PARAM_CMD_MCLK_CHANGE;
cmd.cmd.pg.mclk_change.data = payload;
nvgpu_pmu_dbg(g, "cmd post MS PMU_PG_PARAM_CMD_MCLK_CHANGE");
status = nvgpu_pmu_cmd_post(g, &cmd, NULL, NULL,
PMU_COMMAND_QUEUE_HPQ,
nvgpu_pmu_handle_param_lpwr_msg, &ack_status, &seq);
pmu_wait_message_cond(&g->pmu, gk20a_get_gr_idle_timeout(g),
&ack_status, 1);
if (ack_status == 0U) {
status = -EINVAL;
nvgpu_err(g, "MCLK-CHANGE ACK failed");
}
}
return status;
}
int nvgpu_lpwr_post_init(struct gk20a *g)
{
struct pmu_cmd cmd;
u32 seq;
int status = 0;
u32 ack_status = 0;
size_t tmp_size = PMU_CMD_HDR_SIZE +
sizeof(struct pmu_pg_cmd_post_init_param);
(void) memset(&cmd, 0, sizeof(struct pmu_cmd));
cmd.hdr.unit_id = PMU_UNIT_PG;
nvgpu_assert(tmp_size <= U8_MAX);
cmd.hdr.size = (u8)tmp_size;
cmd.cmd.pg.post_init.cmd_type =
PMU_PG_CMD_ID_PG_PARAM;
cmd.cmd.pg.post_init.cmd_id =
PMU_PG_PARAM_CMD_POST_INIT;
nvgpu_pmu_dbg(g, "cmd post post-init PMU_PG_PARAM_CMD_POST_INIT");
status = nvgpu_pmu_cmd_post(g, &cmd, NULL, NULL,
PMU_COMMAND_QUEUE_LPQ,
nvgpu_pmu_handle_param_lpwr_msg, &ack_status, &seq);
pmu_wait_message_cond(&g->pmu, gk20a_get_gr_idle_timeout(g),
&ack_status, 1);
if (ack_status == 0U) {
status = -EINVAL;
nvgpu_err(g, "post-init ack failed");
}
return status;
}
bool nvgpu_lpwr_is_mscg_supported(struct gk20a *g, u32 pstate_num)
{
struct nvgpu_lpwr_bios_ms_data *pms_data =
&g->perf_pmu->lpwr.lwpr_bios_data.ms;
struct nvgpu_lpwr_bios_idx_data *pidx_data =
&g->perf_pmu->lpwr.lwpr_bios_data.idx;
struct pstate *pstate = pstate_find(g, pstate_num);
u32 ms_idx;
nvgpu_log_fn(g, " ");
if (pstate == NULL) {
return false;
}
ms_idx = pidx_data->entry[pstate->lpwr_entry_idx].ms_idx;
if (pms_data->entry[ms_idx].ms_enabled) {
return true;
} else {
return false;
}
}
bool nvgpu_lpwr_is_rppg_supported(struct gk20a *g, u32 pstate_num)
{
struct nvgpu_lpwr_bios_gr_data *pgr_data =
&g->perf_pmu->lpwr.lwpr_bios_data.gr;
struct nvgpu_lpwr_bios_idx_data *pidx_data =
&g->perf_pmu->lpwr.lwpr_bios_data.idx;
struct pstate *pstate = pstate_find(g, pstate_num);
u32 idx;
nvgpu_log_fn(g, " ");
if (pstate == NULL) {
return false;
}
idx = pidx_data->entry[pstate->lpwr_entry_idx].gr_idx;
if (pgr_data->entry[idx].gr_enabled) {
return true;
} else {
return false;
}
}
int nvgpu_lpwr_enable_pg(struct gk20a *g, bool pstate_lock)
{
struct nvgpu_pmu *pmu = &g->pmu;
int status = 0;
bool is_mscg_supported = false;
bool is_rppg_supported = false;
u32 present_pstate = 0;
nvgpu_log_fn(g, " ");
if (pstate_lock) {
nvgpu_clk_arb_pstate_change_lock(g, true);
}
nvgpu_mutex_acquire(&pmu->pg_mutex);
present_pstate = nvgpu_clk_arb_get_current_pstate(g);
is_mscg_supported = nvgpu_lpwr_is_mscg_supported(g,
present_pstate);
if (is_mscg_supported && g->mscg_enabled) {
if (pmu->mscg_stat == 0U) {
pmu->mscg_stat = PMU_MSCG_ENABLED;
}
}
is_rppg_supported = nvgpu_lpwr_is_rppg_supported(g,
present_pstate);
if (is_rppg_supported) {
if (g->can_elpg) {
status = nvgpu_pmu_enable_elpg(g);
}
}
nvgpu_mutex_release(&pmu->pg_mutex);
if (pstate_lock) {
nvgpu_clk_arb_pstate_change_lock(g, false);
}
return status;
}
int nvgpu_lpwr_disable_pg(struct gk20a *g, bool pstate_lock)
{
struct nvgpu_pmu *pmu = &g->pmu;
int status = 0;
bool is_mscg_supported = false;
bool is_rppg_supported = false;
u32 present_pstate = 0;
nvgpu_log_fn(g, " ");
if (pstate_lock) {
nvgpu_clk_arb_pstate_change_lock(g, true);
}
nvgpu_mutex_acquire(&pmu->pg_mutex);
present_pstate = nvgpu_clk_arb_get_current_pstate(g);
is_rppg_supported = nvgpu_lpwr_is_rppg_supported(g,
present_pstate);
if (is_rppg_supported) {
if (g->elpg_enabled) {
status = nvgpu_pmu_disable_elpg(g);
if (status != 0) {
goto exit_unlock;
}
}
}
is_mscg_supported = nvgpu_lpwr_is_mscg_supported(g,
present_pstate);
if (is_mscg_supported && g->mscg_enabled) {
if (pmu->mscg_stat != 0U) {
pmu->mscg_stat = PMU_MSCG_DISABLED;
}
}
exit_unlock:
nvgpu_mutex_release(&pmu->pg_mutex);
if (pstate_lock) {
nvgpu_clk_arb_pstate_change_lock(g, false);
}
nvgpu_log_fn(g, "done");
return status;
}