mirror of
git://nv-tegra.nvidia.com/linux-nvgpu.git
synced 2025-12-23 01:50:07 +03:00
For every copy_back enabled PMU cmd sent by NVGPU we are waiting for PMU response but not checking for timeout error. This will result in copying invalid data which causes errors. Implemented timeout check and return error if timedout. NVBUG-200530426 Change-Id: I32eba16eeb6f7a56724329ab6d85fae062c6fa3f Signed-off-by: rmylavarapu <rmylavarapu@nvidia.com> Reviewed-on: https://git-master.nvidia.com/r/c/linux-nvgpu/+/2258947 Tested-by: mobile promotions <svcmobile_promotions@nvidia.com> Reviewed-by: mobile promotions <svcmobile_promotions@nvidia.com>
346 lines
11 KiB
C
346 lines
11 KiB
C
/*
|
|
* Copyright (c) 2016-2020, NVIDIA CORPORATION. All rights reserved.
|
|
*
|
|
* Permission is hereby granted, free of charge, to any person obtaining a
|
|
* copy of this software and associated documentation files (the "Software"),
|
|
* to deal in the Software without restriction, including without limitation
|
|
* the rights to use, copy, modify, merge, publish, distribute, sublicense,
|
|
* and/or sell copies of the Software, and to permit persons to whom the
|
|
* Software is furnished to do so, subject to the following conditions:
|
|
*
|
|
* The above copyright notice and this permission notice shall be included in
|
|
* all copies or substantial portions of the Software.
|
|
*
|
|
* THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
|
|
* IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
|
|
* FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
|
|
* THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
|
|
* LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING
|
|
* FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER
|
|
* DEALINGS IN THE SOFTWARE.
|
|
*/
|
|
|
|
#include <nvgpu/pmu.h>
|
|
#include <nvgpu/pmu/pmuif/nvgpu_cmdif.h>
|
|
#include <nvgpu/gk20a.h>
|
|
#include <nvgpu/timers.h>
|
|
#include <nvgpu/boardobj.h>
|
|
#include <nvgpu/boardobjgrp_e32.h>
|
|
#include <nvgpu/pmu/clk/clk.h>
|
|
#include <nvgpu/pmu/perf.h>
|
|
#include <nvgpu/pmu/cmd.h>
|
|
#include <nvgpu/pmu/super_surface.h>
|
|
#include <nvgpu/pmu/pmu_pstate.h>
|
|
#include <nvgpu/pmu/perf.h>
|
|
|
|
#include "change_seq.h"
|
|
#include "perf.h"
|
|
|
|
static int perf_change_seq_sw_setup_super(struct gk20a *g,
|
|
struct change_seq *p_change_seq)
|
|
{
|
|
int status = 0;
|
|
|
|
nvgpu_log_fn(g, " ");
|
|
|
|
/* Initialize parameters */
|
|
p_change_seq->client_lock_mask = 0;
|
|
|
|
p_change_seq->version = CTRL_PERF_CHANGE_SEQ_VERSION_35;
|
|
|
|
status = nvgpu_boardobjgrpmask_init(
|
|
&p_change_seq->clk_domains_exclusion_mask.super,
|
|
32U, ((void*)0));
|
|
if (status != 0) {
|
|
nvgpu_err(g, "clk_domains_exclusion_mask failed to init %d",
|
|
status);
|
|
goto perf_change_seq_sw_setup_super_exit;
|
|
}
|
|
|
|
status = nvgpu_boardobjgrpmask_init(
|
|
&p_change_seq->clk_domains_inclusion_mask.super,
|
|
32U, ((void*)0));
|
|
if (status != 0) {
|
|
nvgpu_err(g, "clk_domains_inclusion_mask failed to init %d",
|
|
status);
|
|
goto perf_change_seq_sw_setup_super_exit;
|
|
}
|
|
|
|
perf_change_seq_sw_setup_super_exit:
|
|
return status;
|
|
}
|
|
|
|
int perf_change_seq_sw_setup(struct gk20a *g)
|
|
{
|
|
struct change_seq_pmu *perf_change_seq_pmu =
|
|
&(g->pmu->perf_pmu->changeseq_pmu);
|
|
int status = 0;
|
|
|
|
nvgpu_log_fn(g, " ");
|
|
|
|
(void) memset(perf_change_seq_pmu, 0,
|
|
sizeof(struct change_seq_pmu));
|
|
|
|
status = perf_change_seq_sw_setup_super(g, &perf_change_seq_pmu->super);
|
|
if (status != 0) {
|
|
goto exit;
|
|
}
|
|
|
|
perf_change_seq_pmu->super.b_enabled_pmu_support = true;
|
|
/*exclude MCLK, may not be needed as MCLK is already fixed */
|
|
perf_change_seq_pmu->super.clk_domains_exclusion_mask.super.data[0]
|
|
= 0x04U;
|
|
perf_change_seq_pmu->b_vf_point_check_ignore = false;
|
|
perf_change_seq_pmu->b_lock = false;
|
|
perf_change_seq_pmu->cpu_step_id_mask = 0;
|
|
perf_change_seq_pmu->cpu_adverised_step_id_mask = 0;
|
|
perf_change_seq_pmu->change_state = 0U;
|
|
|
|
exit:
|
|
return status;
|
|
}
|
|
|
|
static void build_change_seq_boot (struct gk20a *g)
|
|
{
|
|
struct nvgpu_pmu *pmu = g->pmu;
|
|
struct change_seq_pmu *perf_change_seq_pmu =
|
|
&(g->pmu->perf_pmu->changeseq_pmu);
|
|
struct nvgpu_clk_domain *pdomain;
|
|
struct clk_set_info *p0_info;
|
|
struct change_seq_pmu_script *script_last =
|
|
&perf_change_seq_pmu->script_last;
|
|
u8 i = 0;
|
|
|
|
nvgpu_log_fn(g, " ");
|
|
|
|
script_last->super_surface_offset =
|
|
nvgpu_pmu_get_ss_member_set_offset(g, pmu,
|
|
NV_PMU_SUPER_SURFACE_MEMBER_CHANGE_SEQ_GRP) +
|
|
(u32)(sizeof(struct perf_change_seq_pmu_script) *
|
|
SEQ_SCRIPT_LAST);
|
|
|
|
nvgpu_mem_rd_n(g, nvgpu_pmu_super_surface_mem(g,
|
|
pmu, pmu->super_surface),
|
|
script_last->super_surface_offset,
|
|
&script_last->buf,
|
|
(u32) sizeof(struct perf_change_seq_pmu_script));
|
|
|
|
script_last->buf.change.data.flags = CTRL_PERF_CHANGE_SEQ_CHANGE_NONE;
|
|
|
|
BOARDOBJGRP_FOR_EACH(&(g->pmu->clk_pmu->clk_domainobjs->super.super),
|
|
struct nvgpu_clk_domain *, pdomain, i) {
|
|
|
|
p0_info = nvgpu_pmu_perf_pstate_get_clk_set_info(g,
|
|
CTRL_PERF_PSTATE_P0, pdomain->domain);
|
|
|
|
script_last->buf.change.data.clk_list.clk_domains[i].clk_domain =
|
|
pdomain->api_domain;
|
|
|
|
script_last->buf.change.data.clk_list.clk_domains[i].clk_freq_khz =
|
|
p0_info->nominal_mhz * 1000U;
|
|
|
|
/* VBIOS always boots with FFR*/
|
|
script_last->buf.change.data.clk_list.clk_domains[i].regime_id =
|
|
CTRL_CLK_FLL_REGIME_ID_FFR;
|
|
|
|
script_last->buf.change.data.clk_list.num_domains++;
|
|
|
|
nvgpu_pmu_dbg(g, "Domain %x, Nom Freq = %d Max Freq =%d, regime %d",
|
|
pdomain->api_domain,p0_info->nominal_mhz, p0_info->max_mhz,
|
|
CTRL_CLK_FLL_REGIME_ID_FFR);
|
|
}
|
|
|
|
nvgpu_pmu_dbg(g,"Total domains = %d\n",
|
|
script_last->buf.change.data.clk_list.num_domains);
|
|
|
|
/* Assume everything is P0 - Need to find the index for P0 */
|
|
script_last->buf.change.data.pstate_index =
|
|
nvgpu_get_pstate_entry_idx(g, CTRL_PERF_PSTATE_P0);
|
|
|
|
nvgpu_mem_wr_n(g, nvgpu_pmu_super_surface_mem(g,
|
|
pmu, pmu->super_surface),
|
|
script_last->super_surface_offset,
|
|
&script_last->buf,
|
|
(u32) sizeof(struct perf_change_seq_pmu_script));
|
|
|
|
return;
|
|
}
|
|
|
|
int perf_change_seq_pmu_setup(struct gk20a *g)
|
|
{
|
|
struct nv_pmu_rpc_perf_change_seq_info_get info_get;
|
|
struct nv_pmu_rpc_perf_change_seq_info_set info_set;
|
|
struct nvgpu_pmu *pmu = g->pmu;
|
|
struct change_seq_pmu *perf_change_seq_pmu =
|
|
&(g->pmu->perf_pmu->changeseq_pmu);
|
|
int status;
|
|
|
|
/* Do this till we enable performance table */
|
|
build_change_seq_boot(g);
|
|
|
|
(void) memset(&info_get, 0,
|
|
sizeof(struct nv_pmu_rpc_perf_change_seq_info_get));
|
|
(void) memset(&info_set, 0,
|
|
sizeof(struct nv_pmu_rpc_perf_change_seq_info_set));
|
|
|
|
PMU_RPC_EXECUTE_CPB(status, pmu, PERF, CHANGE_SEQ_INFO_GET, &info_get, 0);
|
|
if (status != 0) {
|
|
nvgpu_err(g,
|
|
"Failed to execute Change Seq GET RPC status=0x%x",
|
|
status);
|
|
goto perf_change_seq_pmu_setup_exit;
|
|
}
|
|
|
|
info_set.info_set.super.version = perf_change_seq_pmu->super.version;
|
|
|
|
status = nvgpu_boardobjgrpmask_export(
|
|
&perf_change_seq_pmu->super.clk_domains_exclusion_mask.super,
|
|
perf_change_seq_pmu->
|
|
super.clk_domains_exclusion_mask.super.bitcount,
|
|
&info_set.info_set.super.clk_domains_exclusion_mask.super);
|
|
if ( status != 0 ) {
|
|
nvgpu_err(g, "Could not export clkdomains exclusion mask");
|
|
goto perf_change_seq_pmu_setup_exit;
|
|
}
|
|
|
|
status = nvgpu_boardobjgrpmask_export(
|
|
&perf_change_seq_pmu->super.clk_domains_inclusion_mask.super,
|
|
perf_change_seq_pmu->
|
|
super.clk_domains_inclusion_mask.super.bitcount,
|
|
&info_set.info_set.super.clk_domains_inclusion_mask.super);
|
|
if ( status != 0 ) {
|
|
nvgpu_err(g, "Could not export clkdomains inclusion mask");
|
|
goto perf_change_seq_pmu_setup_exit;
|
|
}
|
|
|
|
info_set.info_set.b_vf_point_check_ignore =
|
|
perf_change_seq_pmu->b_vf_point_check_ignore;
|
|
info_set.info_set.cpu_step_id_mask =
|
|
perf_change_seq_pmu->cpu_step_id_mask;
|
|
info_set.info_set.b_lock =
|
|
perf_change_seq_pmu->b_lock;
|
|
|
|
perf_change_seq_pmu->script_last.super_surface_offset =
|
|
nvgpu_pmu_get_ss_member_set_offset(g, pmu,
|
|
NV_PMU_SUPER_SURFACE_MEMBER_CHANGE_SEQ_GRP) +
|
|
(u32)(sizeof(struct perf_change_seq_pmu_script) *
|
|
SEQ_SCRIPT_LAST);
|
|
|
|
nvgpu_mem_rd_n(g, nvgpu_pmu_super_surface_mem(g,
|
|
pmu, pmu->super_surface),
|
|
perf_change_seq_pmu->script_last.super_surface_offset,
|
|
&perf_change_seq_pmu->script_last.buf,
|
|
(u32) sizeof(struct perf_change_seq_pmu_script));
|
|
|
|
/* Assume everything is P0 - Need to find the index for P0 */
|
|
perf_change_seq_pmu->script_last.buf.change.data.pstate_index =
|
|
nvgpu_get_pstate_entry_idx(g, CTRL_PERF_PSTATE_P0);;
|
|
|
|
nvgpu_mem_wr_n(g, nvgpu_pmu_super_surface_mem(g,
|
|
pmu, pmu->super_surface),
|
|
perf_change_seq_pmu->script_last.super_surface_offset,
|
|
&perf_change_seq_pmu->script_last.buf,
|
|
(u32) sizeof(struct perf_change_seq_pmu_script));
|
|
|
|
/* Continue with PMU setup, assume FB map is done */
|
|
PMU_RPC_EXECUTE_CPB(status, pmu, PERF, CHANGE_SEQ_INFO_SET, &info_set, 0);
|
|
if (status != 0) {
|
|
nvgpu_err(g,
|
|
"Failed to execute Change Seq SET RPC status=0x%x",
|
|
status);
|
|
goto perf_change_seq_pmu_setup_exit;
|
|
}
|
|
|
|
perf_change_seq_pmu_setup_exit:
|
|
return status;
|
|
}
|
|
|
|
int nvgpu_pmu_perf_changeseq_set_clks(struct gk20a *g,
|
|
struct nvgpu_clk_slave_freq *vf_point)
|
|
{
|
|
struct nvgpu_pmu *pmu = g->pmu;
|
|
struct nv_pmu_rpc_perf_change_seq_queue_change rpc;
|
|
struct ctrl_perf_change_seq_change_input change_input;
|
|
struct change_seq_pmu *change_seq_pmu = &g->pmu->perf_pmu->changeseq_pmu;
|
|
int status = 0;
|
|
u8 gpcclk_domain = 0U;
|
|
u32 gpcclk_voltuv = 0U, gpcclk_clkmhz = 0U;
|
|
u32 vmin_uv = 0U, vmax_uv = 0U;
|
|
u32 vmargin_uv = 0U, fmargin_mhz = 0U;
|
|
|
|
(void) memset(&change_input, 0,
|
|
sizeof(struct ctrl_perf_change_seq_change_input));
|
|
|
|
g->pmu->clk_pmu->set_p0_clks(g, &gpcclk_domain, &gpcclk_clkmhz,
|
|
vf_point, &change_input);
|
|
|
|
change_input.pstate_index =
|
|
nvgpu_get_pstate_entry_idx(g, CTRL_PERF_PSTATE_P0);
|
|
change_input.flags = (u32)CTRL_PERF_CHANGE_SEQ_CHANGE_FORCE;
|
|
change_input.vf_points_cache_counter = 0xFFFFFFFFU;
|
|
|
|
status = nvgpu_pmu_perf_vfe_get_freq_margin(g, &fmargin_mhz);
|
|
if (status != 0) {
|
|
nvgpu_err(g, "Failed to fetch Fmargin status=0x%x", status);
|
|
return status;
|
|
}
|
|
|
|
gpcclk_clkmhz += fmargin_mhz;
|
|
status = nvgpu_pmu_clk_domain_freq_to_volt(g, gpcclk_domain,
|
|
&gpcclk_clkmhz, &gpcclk_voltuv, CTRL_VOLT_DOMAIN_LOGIC);
|
|
|
|
status = nvgpu_pmu_perf_vfe_get_volt_margin(g, &vmargin_uv);
|
|
if (status != 0) {
|
|
nvgpu_err(g, "Failed to fetch Vmargin status=0x%x", status);
|
|
return status;
|
|
}
|
|
|
|
gpcclk_voltuv += vmargin_uv;
|
|
status = nvgpu_pmu_volt_get_vmin_vmax_ps35(g, &vmin_uv, &vmax_uv);
|
|
if (status != 0) {
|
|
nvgpu_pmu_dbg(g, "Get vmin,vmax failed, proceeding with "
|
|
"freq_to_volt value");
|
|
}
|
|
if ((status == 0) && (vmin_uv > gpcclk_voltuv)) {
|
|
gpcclk_voltuv = vmin_uv;
|
|
nvgpu_log_fn(g, "Vmin is higher than evaluated Volt");
|
|
}
|
|
|
|
if (gpcclk_voltuv > vmax_uv) {
|
|
nvgpu_err(g, "Error: Requested voltage is more than chip max");
|
|
return -EINVAL;
|
|
}
|
|
|
|
change_input.volt[0].voltage_uv = gpcclk_voltuv;
|
|
change_input.volt[0].voltage_min_noise_unaware_uv = gpcclk_voltuv;
|
|
change_input.volt_rails_mask.super.data[0] = 1U;
|
|
|
|
/* RPC to PMU to queue to execute change sequence request*/
|
|
(void) memset(&rpc, 0,
|
|
sizeof(struct nv_pmu_rpc_perf_change_seq_queue_change));
|
|
rpc.change = change_input;
|
|
rpc.change.pstate_index =
|
|
nvgpu_get_pstate_entry_idx(g, CTRL_PERF_PSTATE_P0);
|
|
change_seq_pmu->change_state = 0U;
|
|
change_seq_pmu->start_time = nvgpu_current_time_us();
|
|
PMU_RPC_EXECUTE_CPB(status, pmu, PERF,
|
|
CHANGE_SEQ_QUEUE_CHANGE, &rpc, 0);
|
|
if (status != 0) {
|
|
nvgpu_err(g, "Failed to execute Change Seq RPC status=0x%x",
|
|
status);
|
|
}
|
|
|
|
/* Wait for sync change to complete. */
|
|
if ((rpc.change.flags & CTRL_PERF_CHANGE_SEQ_CHANGE_ASYNC) == 0U) {
|
|
/* wait till RPC execute in PMU & ACK */
|
|
if (nvgpu_pmu_wait_fw_ack_status(g, pmu,
|
|
nvgpu_get_poll_timeout(g),
|
|
&change_seq_pmu->change_state, 1U) != 0) {
|
|
nvgpu_err(g, "PMU wait timeout expired.");
|
|
status = -ETIMEDOUT;
|
|
}
|
|
}
|
|
change_seq_pmu->stop_time = nvgpu_current_time_us();
|
|
return status;
|
|
}
|