mirror of
git://nv-tegra.nvidia.com/linux-nvgpu.git
synced 2025-12-23 18:16:01 +03:00
MISRA Rule-15.6 requires that all loop bodies must be enclosed in braces including single statement loop bodies. This patch fix the MISRA violations due to single statement loop bodies without braces by adding them. JIRA NVGPU-989 Change-Id: If79f56f92b94d0114477b66a6f654ac16ee8ea27 Signed-off-by: Srirangan <smadhavan@nvidia.com> Reviewed-on: https://git-master.nvidia.com/r/1791194 Reviewed-by: Adeel Raza <araza@nvidia.com> GVS: Gerrit_Virtual_Submit Reviewed-by: Alex Waterman <alexw@nvidia.com> Reviewed-by: mobile promotions <svcmobile_promotions@nvidia.com> Tested-by: mobile promotions <svcmobile_promotions@nvidia.com>
296 lines
9.6 KiB
C
296 lines
9.6 KiB
C
/*
|
|
* Copyright (c) 2017-2018, NVIDIA CORPORATION. All rights reserved.
|
|
*
|
|
* Permission is hereby granted, free of charge, to any person obtaining a
|
|
* copy of this software and associated documentation files (the "Software"),
|
|
* to deal in the Software without restriction, including without limitation
|
|
* the rights to use, copy, modify, merge, publish, distribute, sublicense,
|
|
* and/or sell copies of the Software, and to permit persons to whom the
|
|
* Software is furnished to do so, subject to the following conditions:
|
|
*
|
|
* The above copyright notice and this permission notice shall be included in
|
|
* all copies or substantial portions of the Software.
|
|
*
|
|
* THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
|
|
* IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
|
|
* FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
|
|
* THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
|
|
* LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING
|
|
* FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER
|
|
* DEALINGS IN THE SOFTWARE.
|
|
*/
|
|
|
|
#include <nvgpu/types.h>
|
|
#include <nvgpu/dma.h>
|
|
#include <nvgpu/gmmu.h>
|
|
#include <nvgpu/timers.h>
|
|
#include <nvgpu/nvgpu_common.h>
|
|
#include <nvgpu/kmem.h>
|
|
#include <nvgpu/nvgpu_mem.h>
|
|
#include <nvgpu/firmware.h>
|
|
#include <nvgpu/mm.h>
|
|
#include <nvgpu/acr/nvgpu_acr.h>
|
|
#include <nvgpu/enabled.h>
|
|
#include <nvgpu/io.h>
|
|
|
|
#include "gk20a/gk20a.h"
|
|
#include "acr_gv11b.h"
|
|
#include "pmu_gv11b.h"
|
|
#include "gk20a/pmu_gk20a.h"
|
|
#include "gm20b/mm_gm20b.h"
|
|
#include "gm20b/acr_gm20b.h"
|
|
#include "gp106/acr_gp106.h"
|
|
|
|
#include <nvgpu/hw/gv11b/hw_pwr_gv11b.h>
|
|
|
|
/*Defines*/
|
|
#define gv11b_dbg_pmu(g, fmt, arg...) \
|
|
nvgpu_log(g, gpu_dbg_pmu, fmt, ##arg)
|
|
|
|
static void flcn64_set_dma(struct falc_u64 *dma_addr, u64 value)
|
|
{
|
|
dma_addr->lo |= u64_lo32(value);
|
|
dma_addr->hi |= u64_hi32(value);
|
|
}
|
|
/*Externs*/
|
|
|
|
/*Forwards*/
|
|
|
|
int gv11b_alloc_blob_space(struct gk20a *g,
|
|
size_t size, struct nvgpu_mem *mem)
|
|
{
|
|
int err;
|
|
|
|
gv11b_dbg_pmu(g, "alloc blob space: NVGPU_DMA_FORCE_CONTIGUOUS");
|
|
err = nvgpu_dma_alloc_flags_sys(g, NVGPU_DMA_FORCE_CONTIGUOUS,
|
|
size, mem);
|
|
|
|
return err;
|
|
}
|
|
|
|
/*Loads ACR bin to FB mem and bootstraps PMU with bootloader code
|
|
* start and end are addresses of ucode blob in non-WPR region*/
|
|
int gv11b_bootstrap_hs_flcn(struct gk20a *g)
|
|
{
|
|
struct mm_gk20a *mm = &g->mm;
|
|
struct vm_gk20a *vm = mm->pmu.vm;
|
|
int err = 0;
|
|
u64 *acr_dmem;
|
|
u32 img_size_in_bytes = 0;
|
|
u32 status, size, index;
|
|
u64 start;
|
|
struct acr_desc *acr = &g->acr;
|
|
struct nvgpu_firmware *acr_fw = acr->acr_fw;
|
|
struct flcn_bl_dmem_desc_v1 *bl_dmem_desc = &acr->bl_dmem_desc_v1;
|
|
u32 *acr_ucode_header_t210_load;
|
|
u32 *acr_ucode_data_t210_load;
|
|
|
|
start = nvgpu_mem_get_addr(g, &acr->ucode_blob);
|
|
size = acr->ucode_blob.size;
|
|
|
|
gv11b_dbg_pmu(g, "acr ucode blob start %llx\n", start);
|
|
gv11b_dbg_pmu(g, "acr ucode blob size %x\n", size);
|
|
|
|
gv11b_dbg_pmu(g, " ");
|
|
|
|
if (!acr_fw) {
|
|
/*First time init case*/
|
|
acr_fw = nvgpu_request_firmware(g,
|
|
GM20B_HSBIN_PMU_UCODE_IMAGE, 0);
|
|
if (!acr_fw) {
|
|
nvgpu_err(g, "pmu ucode get fail");
|
|
return -ENOENT;
|
|
}
|
|
acr->acr_fw = acr_fw;
|
|
acr->hsbin_hdr = (struct bin_hdr *)acr_fw->data;
|
|
acr->fw_hdr = (struct acr_fw_header *)(acr_fw->data +
|
|
acr->hsbin_hdr->header_offset);
|
|
acr_ucode_data_t210_load = (u32 *)(acr_fw->data +
|
|
acr->hsbin_hdr->data_offset);
|
|
acr_ucode_header_t210_load = (u32 *)(acr_fw->data +
|
|
acr->fw_hdr->hdr_offset);
|
|
img_size_in_bytes = ALIGN((acr->hsbin_hdr->data_size), 256);
|
|
|
|
gv11b_dbg_pmu(g, "sig dbg offset %u\n",
|
|
acr->fw_hdr->sig_dbg_offset);
|
|
gv11b_dbg_pmu(g, "sig dbg size %u\n", acr->fw_hdr->sig_dbg_size);
|
|
gv11b_dbg_pmu(g, "sig prod offset %u\n",
|
|
acr->fw_hdr->sig_prod_offset);
|
|
gv11b_dbg_pmu(g, "sig prod size %u\n",
|
|
acr->fw_hdr->sig_prod_size);
|
|
gv11b_dbg_pmu(g, "patch loc %u\n", acr->fw_hdr->patch_loc);
|
|
gv11b_dbg_pmu(g, "patch sig %u\n", acr->fw_hdr->patch_sig);
|
|
gv11b_dbg_pmu(g, "header offset %u\n", acr->fw_hdr->hdr_offset);
|
|
gv11b_dbg_pmu(g, "header size %u\n", acr->fw_hdr->hdr_size);
|
|
|
|
/* Lets patch the signatures first.. */
|
|
if (acr_ucode_patch_sig(g, acr_ucode_data_t210_load,
|
|
(u32 *)(acr_fw->data +
|
|
acr->fw_hdr->sig_prod_offset),
|
|
(u32 *)(acr_fw->data +
|
|
acr->fw_hdr->sig_dbg_offset),
|
|
(u32 *)(acr_fw->data +
|
|
acr->fw_hdr->patch_loc),
|
|
(u32 *)(acr_fw->data +
|
|
acr->fw_hdr->patch_sig)) < 0) {
|
|
nvgpu_err(g, "patch signatures fail");
|
|
err = -1;
|
|
goto err_release_acr_fw;
|
|
}
|
|
err = nvgpu_dma_alloc_map_sys(vm, img_size_in_bytes,
|
|
&acr->acr_ucode);
|
|
if (err) {
|
|
err = -ENOMEM;
|
|
goto err_release_acr_fw;
|
|
}
|
|
|
|
for (index = 0; index < 9; index++) {
|
|
gv11b_dbg_pmu(g, "acr_ucode_header_t210_load %u\n",
|
|
acr_ucode_header_t210_load[index]);
|
|
}
|
|
|
|
acr_dmem = (u64 *)
|
|
&(((u8 *)acr_ucode_data_t210_load)[
|
|
acr_ucode_header_t210_load[2]]);
|
|
acr->acr_dmem_desc_v1 = (struct flcn_acr_desc_v1 *)((u8 *)(
|
|
acr->acr_ucode.cpu_va) + acr_ucode_header_t210_load[2]);
|
|
((struct flcn_acr_desc_v1 *)acr_dmem)->nonwpr_ucode_blob_start =
|
|
(start);
|
|
((struct flcn_acr_desc_v1 *)acr_dmem)->nonwpr_ucode_blob_size =
|
|
size;
|
|
((struct flcn_acr_desc_v1 *)acr_dmem)->regions.no_regions = 2;
|
|
((struct flcn_acr_desc_v1 *)acr_dmem)->wpr_offset = 0;
|
|
|
|
nvgpu_mem_wr_n(g, &acr->acr_ucode, 0,
|
|
acr_ucode_data_t210_load, img_size_in_bytes);
|
|
/*
|
|
* In order to execute this binary, we will be using
|
|
* a bootloader which will load this image into PMU IMEM/DMEM.
|
|
* Fill up the bootloader descriptor for PMU HAL to use..
|
|
* TODO: Use standard descriptor which the generic bootloader is
|
|
* checked in.
|
|
*/
|
|
bl_dmem_desc->signature[0] = 0;
|
|
bl_dmem_desc->signature[1] = 0;
|
|
bl_dmem_desc->signature[2] = 0;
|
|
bl_dmem_desc->signature[3] = 0;
|
|
bl_dmem_desc->ctx_dma = GK20A_PMU_DMAIDX_VIRT;
|
|
flcn64_set_dma(&bl_dmem_desc->code_dma_base,
|
|
acr->acr_ucode.gpu_va);
|
|
bl_dmem_desc->non_sec_code_off = acr_ucode_header_t210_load[0];
|
|
bl_dmem_desc->non_sec_code_size = acr_ucode_header_t210_load[1];
|
|
bl_dmem_desc->sec_code_off = acr_ucode_header_t210_load[5];
|
|
bl_dmem_desc->sec_code_size = acr_ucode_header_t210_load[6];
|
|
bl_dmem_desc->code_entry_point = 0; /* Start at 0th offset */
|
|
flcn64_set_dma(&bl_dmem_desc->data_dma_base,
|
|
acr->acr_ucode.gpu_va +
|
|
acr_ucode_header_t210_load[2]);
|
|
bl_dmem_desc->data_size = acr_ucode_header_t210_load[3];
|
|
} else
|
|
acr->acr_dmem_desc_v1->nonwpr_ucode_blob_size = 0;
|
|
status = pmu_exec_gen_bl(g, bl_dmem_desc, 1);
|
|
if (status != 0) {
|
|
err = status;
|
|
goto err_free_ucode_map;
|
|
}
|
|
|
|
return 0;
|
|
err_free_ucode_map:
|
|
nvgpu_dma_unmap_free(vm, &acr->acr_ucode);
|
|
err_release_acr_fw:
|
|
nvgpu_release_firmware(g, acr_fw);
|
|
acr->acr_fw = NULL;
|
|
|
|
return err;
|
|
}
|
|
|
|
static int bl_bootstrap(struct nvgpu_pmu *pmu,
|
|
struct flcn_bl_dmem_desc_v1 *pbl_desc, u32 bl_sz)
|
|
{
|
|
struct gk20a *g = gk20a_from_pmu(pmu);
|
|
struct mm_gk20a *mm = &g->mm;
|
|
struct nvgpu_falcon_bl_info bl_info;
|
|
|
|
nvgpu_log_fn(g, " ");
|
|
|
|
gk20a_writel(g, pwr_falcon_itfen_r(),
|
|
gk20a_readl(g, pwr_falcon_itfen_r()) |
|
|
pwr_falcon_itfen_ctxen_enable_f());
|
|
gk20a_writel(g, pwr_pmu_new_instblk_r(),
|
|
pwr_pmu_new_instblk_ptr_f(
|
|
nvgpu_inst_block_addr(g, &mm->pmu.inst_block) >> 12) |
|
|
pwr_pmu_new_instblk_valid_f(1) |
|
|
(nvgpu_is_enabled(g, NVGPU_USE_COHERENT_SYSMEM) ?
|
|
pwr_pmu_new_instblk_target_sys_coh_f() :
|
|
pwr_pmu_new_instblk_target_sys_ncoh_f())) ;
|
|
|
|
bl_info.bl_src = g->acr.hsbl_ucode.cpu_va;
|
|
bl_info.bl_desc = (u8 *)pbl_desc;
|
|
bl_info.bl_desc_size = sizeof(struct flcn_bl_dmem_desc_v1);
|
|
bl_info.bl_size = bl_sz;
|
|
bl_info.bl_start_tag = g->acr.pmu_hsbl_desc->bl_start_tag;
|
|
nvgpu_flcn_bl_bootstrap(&g->pmu_flcn, &bl_info);
|
|
|
|
return 0;
|
|
}
|
|
|
|
void gv11b_setup_apertures(struct gk20a *g)
|
|
{
|
|
struct mm_gk20a *mm = &g->mm;
|
|
struct nvgpu_mem *inst_block = &mm->pmu.inst_block;
|
|
|
|
/* setup apertures - virtual */
|
|
gk20a_writel(g, pwr_fbif_transcfg_r(GK20A_PMU_DMAIDX_UCODE),
|
|
pwr_fbif_transcfg_mem_type_physical_f() |
|
|
nvgpu_aperture_mask(g, inst_block,
|
|
pwr_fbif_transcfg_target_noncoherent_sysmem_f(),
|
|
pwr_fbif_transcfg_target_coherent_sysmem_f(),
|
|
pwr_fbif_transcfg_target_local_fb_f()));
|
|
gk20a_writel(g, pwr_fbif_transcfg_r(GK20A_PMU_DMAIDX_VIRT),
|
|
pwr_fbif_transcfg_mem_type_virtual_f());
|
|
/* setup apertures - physical */
|
|
gk20a_writel(g, pwr_fbif_transcfg_r(GK20A_PMU_DMAIDX_PHYS_VID),
|
|
pwr_fbif_transcfg_mem_type_physical_f() |
|
|
nvgpu_aperture_mask(g, inst_block,
|
|
pwr_fbif_transcfg_target_noncoherent_sysmem_f(),
|
|
pwr_fbif_transcfg_target_coherent_sysmem_f(),
|
|
pwr_fbif_transcfg_target_local_fb_f()));
|
|
gk20a_writel(g, pwr_fbif_transcfg_r(GK20A_PMU_DMAIDX_PHYS_SYS_COH),
|
|
pwr_fbif_transcfg_mem_type_physical_f() |
|
|
pwr_fbif_transcfg_target_coherent_sysmem_f());
|
|
gk20a_writel(g, pwr_fbif_transcfg_r(GK20A_PMU_DMAIDX_PHYS_SYS_NCOH),
|
|
pwr_fbif_transcfg_mem_type_physical_f() |
|
|
pwr_fbif_transcfg_target_noncoherent_sysmem_f());
|
|
}
|
|
|
|
int gv11b_init_pmu_setup_hw1(struct gk20a *g,
|
|
void *desc, u32 bl_sz)
|
|
{
|
|
struct nvgpu_pmu *pmu = &g->pmu;
|
|
int err;
|
|
|
|
nvgpu_log_fn(g, " ");
|
|
|
|
nvgpu_mutex_acquire(&pmu->isr_mutex);
|
|
nvgpu_flcn_reset(pmu->flcn);
|
|
pmu->isr_enabled = true;
|
|
nvgpu_mutex_release(&pmu->isr_mutex);
|
|
|
|
if (g->ops.pmu.setup_apertures)
|
|
g->ops.pmu.setup_apertures(g);
|
|
if (g->ops.pmu.update_lspmu_cmdline_args)
|
|
g->ops.pmu.update_lspmu_cmdline_args(g);
|
|
|
|
/*disable irqs for hs falcon booting as we will poll for halt*/
|
|
nvgpu_mutex_acquire(&pmu->isr_mutex);
|
|
pmu_enable_irq(pmu, false);
|
|
pmu->isr_enabled = false;
|
|
nvgpu_mutex_release(&pmu->isr_mutex);
|
|
/*Clearing mailbox register used to reflect capabilities*/
|
|
gk20a_writel(g, pwr_falcon_mailbox1_r(), 0);
|
|
err = bl_bootstrap(pmu, desc, bl_sz);
|
|
if (err)
|
|
return err;
|
|
return 0;
|
|
}
|