mirror of
git://nv-tegra.nvidia.com/linux-nvgpu.git
synced 2025-12-22 09:12:24 +03:00
gpu: nvgpu: gsp units separation
Separated gsp unit into three unit: - GSP unit which holds the core functionality of GSP RISCV core, bootstrap, interrupt, etc. - GSP Scheduler to hold the cmd/msg management, IPC, etc. - GSP Test to hold stress test ucode specific support. NVGPU-7492 Signed-off-by: Ramesh Mylavarapu <rmylavarapu@nvidia.com> Change-Id: I12340dc776d610502f28c8574843afc7481c0871 Reviewed-on: https://git-master.nvidia.com/r/c/linux-nvgpu/+/2660619 Tested-by: mobile promotions <svcmobile_promotions@nvidia.com> Reviewed-by: mobile promotions <svcmobile_promotions@nvidia.com>
This commit is contained in:
committed by
mobile promotions
parent
14ed75e857
commit
9302b2efee
@@ -315,22 +315,33 @@ gsp:
|
||||
owner: Ramesh M
|
||||
gpu: igpu
|
||||
sources: [ common/gsp/gsp_init.c,
|
||||
common/gsp/gsp_priv.h,
|
||||
common/gsp/gsp_bootstrap.c,
|
||||
common/gsp/gsp_bootstrap.h,
|
||||
common/gsp/gsp_test.c,
|
||||
include/nvgpu/gsp.h,
|
||||
include/nvgpu/gsp/gsp_test.h,
|
||||
common/gsp/ipc/gsp_seq.c,
|
||||
common/gsp/ipc/gsp_seq.h,
|
||||
common/gsp/ipc/gsp_queue.c,
|
||||
common/gsp/ipc/gsp_queue.h,
|
||||
common/gsp/ipc/gsp_cmd.c,
|
||||
common/gsp/ipc/gsp_cmd.h,
|
||||
common/gsp/ipc/gsp_msg.c,
|
||||
common/gsp/ipc/gsp_msg.h,
|
||||
include/nvgpu/gsp.h ]
|
||||
|
||||
gsp_sched:
|
||||
safe: no
|
||||
owner: Ramesh M
|
||||
gpu: igpu
|
||||
sources: [ common/gsp_scheduler/ipc/gsp_seq.c,
|
||||
common/gsp_scheduler/ipc/gsp_seq.h,
|
||||
common/gsp_scheduler/ipc/gsp_queue.c,
|
||||
common/gsp_scheduler/ipc/gsp_queue.h,
|
||||
common/gsp_scheduler/ipc/gsp_cmd.c,
|
||||
common/gsp_scheduler/ipc/gsp_cmd.h,
|
||||
common/gsp_scheduler/ipc/gsp_msg.c,
|
||||
common/gsp_scheduler/ipc/gsp_msg.h,
|
||||
common/gsp_scheduler/gsp_scheduler.c,
|
||||
common/gsp_scheduler/gsp_scheduler.h,
|
||||
include/nvgpu/gsp_sched.h ]
|
||||
|
||||
gsp_test:
|
||||
safe: no
|
||||
owner: Ramesh M
|
||||
gpu: igpu
|
||||
sources: [ include/nvgpu/gsp/gsp_test.h,
|
||||
common/gsp_test/gsp_test.c,
|
||||
common/gsp_test/gsp_test.h ]
|
||||
|
||||
engine_queues:
|
||||
owner: Sagar K
|
||||
children:
|
||||
|
||||
@@ -423,15 +423,16 @@ ifeq ($(CONFIG_NVGPU_GSP_SCHEDULER),y)
|
||||
nvgpu-$(CONFIG_NVGPU_GSP_SCHEDULER) += \
|
||||
common/gsp/gsp_init.o \
|
||||
common/gsp/gsp_bootstrap.o \
|
||||
common/gsp/ipc/gsp_seq.o \
|
||||
common/gsp/ipc/gsp_queue.o \
|
||||
common/gsp/ipc/gsp_cmd.o \
|
||||
common/gsp/ipc/gsp_msg.o
|
||||
common/gsp_scheduler/ipc/gsp_seq.o \
|
||||
common/gsp_scheduler/ipc/gsp_queue.o \
|
||||
common/gsp_scheduler/ipc/gsp_cmd.o \
|
||||
common/gsp_scheduler/ipc/gsp_msg.o \
|
||||
common/gsp_scheduler/gsp_scheduler.o
|
||||
endif
|
||||
|
||||
ifeq ($(CONFIG_NVGPU_GSP_STRESS_TEST),y)
|
||||
nvgpu-$(CONFIG_NVGPU_GSP_STRESS_TEST) += \
|
||||
common/gsp/gsp_test.o
|
||||
common/gsp_test/gsp_test.o
|
||||
endif
|
||||
|
||||
# Linux specific parts of nvgpu.
|
||||
|
||||
@@ -301,8 +301,8 @@ CONFIG_NVGPU_MIG := 1
|
||||
NVGPU_COMMON_CFLAGS += -DCONFIG_NVGPU_MIG
|
||||
|
||||
# Enable gsp scheduler for normal build
|
||||
CONFIG_NVGPU_GSP_SCHEDULER......:= 1
|
||||
NVGPU_COMMON_CFLAGS.............+= -DCONFIG_NVGPU_GSP_SCHEDULER
|
||||
CONFIG_NVGPU_GSP_SCHEDULER := 1
|
||||
NVGPU_COMMON_CFLAGS += -DCONFIG_NVGPU_GSP_SCHEDULER
|
||||
|
||||
# Enable GSP stress test
|
||||
ifeq ($(CONFIG_NVGPU_GSP_SCHEDULER),1)
|
||||
|
||||
@@ -181,10 +181,15 @@ srcs += common/device.c \
|
||||
ifeq ($(CONFIG_NVGPU_GSP_SCHEDULER),1)
|
||||
srcs += common/gsp/gsp_init.c \
|
||||
common/gsp/gsp_bootstrap.c \
|
||||
common/gsp/ipc/gsp_seq.c \
|
||||
common/gsp/ipc/gsp_queue.c \
|
||||
common/gsp/ipc/gsp_cmd.c \
|
||||
common/gsp/ipc/gsp_msg.c
|
||||
common/gsp_scheduler/ipc/gsp_seq.c \
|
||||
common/gsp_scheduler/ipc/gsp_queue.c \
|
||||
common/gsp_scheduler/ipc/gsp_cmd.c \
|
||||
common/gsp_scheduler/ipc/gsp_msg.c \
|
||||
common/gsp_scheduler/gsp_scheduler.c
|
||||
endif
|
||||
|
||||
ifeq ($(CONFIG_NVGPU_GSP_STRESS_TEST),1)
|
||||
srcs += common/gsp_test/gsp_test.c
|
||||
endif
|
||||
|
||||
# Source files below are functionaly safe (FuSa) and must always be included.
|
||||
|
||||
@@ -1,5 +1,5 @@
|
||||
/*
|
||||
* Copyright (c) 2021, NVIDIA CORPORATION. All rights reserved.
|
||||
* Copyright (c) 2021-2022, NVIDIA CORPORATION. All rights reserved.
|
||||
*
|
||||
* Permission is hereby granted, free of charge, to any person obtaining a
|
||||
* copy of this software and associated documentation files (the "Software"),
|
||||
@@ -28,25 +28,8 @@
|
||||
#include <nvgpu/firmware.h>
|
||||
#include <nvgpu/io.h>
|
||||
#include <nvgpu/gsp.h>
|
||||
|
||||
#include "gsp_priv.h"
|
||||
#include "gsp_bootstrap.h"
|
||||
|
||||
#define GSP_WAIT_TIME_MS 10000U
|
||||
#define GSP_DBG_RISCV_FW_MANIFEST "sample-gsp.manifest.encrypt.bin.out.bin"
|
||||
#define GSP_DBG_RISCV_FW_CODE "sample-gsp.text.encrypt.bin"
|
||||
#define GSP_DBG_RISCV_FW_DATA "sample-gsp.data.encrypt.bin"
|
||||
|
||||
#ifdef CONFIG_NVGPU_GSP_STRESS_TEST
|
||||
#define GSPDBG_RISCV_STRESS_TEST_FW_MANIFEST "gsp-stress.manifest.encrypt.bin.out.bin"
|
||||
#define GSPDBG_RISCV_STRESS_TEST_FW_CODE "gsp-stress.text.encrypt.bin"
|
||||
#define GSPDBG_RISCV_STRESS_TEST_FW_DATA "gsp-stress.data.encrypt.bin"
|
||||
|
||||
#define GSPPROD_RISCV_STRESS_TEST_FW_MANIFEST "gsp-stress.manifest.encrypt.bin.out.bin.prod"
|
||||
#define GSPPROD_RISCV_STRESS_TEST_FW_CODE "gsp-stress.text.encrypt.bin.prod"
|
||||
#define GSPPROD_RISCV_STRESS_TEST_FW_DATA "gsp-stress.data.encrypt.bin.prod"
|
||||
|
||||
#define GSP_STRESS_TEST_MAILBOX_PASS 0xAAAAAAAA
|
||||
#include <nvgpu/gsp/gsp_test.h>
|
||||
#endif
|
||||
|
||||
static void gsp_release_firmware(struct gk20a *g, struct nvgpu_gsp *gsp)
|
||||
@@ -64,86 +47,45 @@ static void gsp_release_firmware(struct gk20a *g, struct nvgpu_gsp *gsp)
|
||||
}
|
||||
}
|
||||
|
||||
static int gsp_read_firmware(struct gk20a *g, struct gsp_fw *gsp_ucode)
|
||||
static int gsp_read_firmware(struct gk20a *g, struct nvgpu_gsp *gsp,
|
||||
struct gsp_fw *gsp_ucode)
|
||||
{
|
||||
const char *gsp_code_name;
|
||||
const char *gsp_data_name;
|
||||
const char *gsp_manifest_name;
|
||||
const char *code_name = gsp_ucode->code_name;
|
||||
const char *data_name = gsp_ucode->data_name;
|
||||
const char *manifest_name = gsp_ucode->manifest_name;
|
||||
|
||||
nvgpu_log_fn(g, " ");
|
||||
|
||||
#ifdef CONFIG_NVGPU_GSP_STRESS_TEST
|
||||
if (g->gsp->gsp_test.load_stress_test) {
|
||||
/*
|
||||
* TODO Switch to GSP specific register
|
||||
*/
|
||||
if (g->ops.pmu.is_debug_mode_enabled(g)) {
|
||||
gsp_code_name = GSPDBG_RISCV_STRESS_TEST_FW_CODE;
|
||||
gsp_data_name = GSPDBG_RISCV_STRESS_TEST_FW_DATA;
|
||||
gsp_manifest_name = GSPDBG_RISCV_STRESS_TEST_FW_MANIFEST;
|
||||
} else {
|
||||
gsp_code_name = GSPPROD_RISCV_STRESS_TEST_FW_CODE;
|
||||
gsp_data_name = GSPPROD_RISCV_STRESS_TEST_FW_DATA;
|
||||
gsp_manifest_name = GSPPROD_RISCV_STRESS_TEST_FW_MANIFEST;
|
||||
}
|
||||
} else
|
||||
#endif
|
||||
{
|
||||
gsp_code_name = GSP_DBG_RISCV_FW_CODE;
|
||||
gsp_data_name = GSP_DBG_RISCV_FW_DATA;
|
||||
gsp_manifest_name = GSP_DBG_RISCV_FW_MANIFEST;
|
||||
}
|
||||
|
||||
gsp_ucode->manifest = nvgpu_request_firmware(g,
|
||||
gsp_manifest_name, NVGPU_REQUEST_FIRMWARE_NO_WARN);
|
||||
manifest_name, NVGPU_REQUEST_FIRMWARE_NO_WARN);
|
||||
if (gsp_ucode->manifest == NULL) {
|
||||
nvgpu_err(g, "%s ucode get failed", gsp_manifest_name);
|
||||
nvgpu_err(g, "%s ucode get failed", manifest_name);
|
||||
goto fw_release;
|
||||
}
|
||||
|
||||
gsp_ucode->code = nvgpu_request_firmware(g,
|
||||
gsp_code_name, NVGPU_REQUEST_FIRMWARE_NO_WARN);
|
||||
code_name, NVGPU_REQUEST_FIRMWARE_NO_WARN);
|
||||
if (gsp_ucode->code == NULL) {
|
||||
nvgpu_err(g, "%s ucode get failed", gsp_code_name);
|
||||
nvgpu_err(g, "%s ucode get failed", code_name);
|
||||
goto fw_release;
|
||||
}
|
||||
|
||||
gsp_ucode->data = nvgpu_request_firmware(g,
|
||||
gsp_data_name, NVGPU_REQUEST_FIRMWARE_NO_WARN);
|
||||
data_name, NVGPU_REQUEST_FIRMWARE_NO_WARN);
|
||||
if (gsp_ucode->data == NULL) {
|
||||
nvgpu_err(g, "%s ucode get failed", gsp_data_name);
|
||||
nvgpu_err(g, "%s ucode get failed", data_name);
|
||||
goto fw_release;
|
||||
}
|
||||
|
||||
return 0;
|
||||
|
||||
fw_release:
|
||||
gsp_release_firmware(g, g->gsp);
|
||||
gsp_release_firmware(g, gsp);
|
||||
return -ENOENT;
|
||||
}
|
||||
|
||||
#ifdef CONFIG_NVGPU_GSP_STRESS_TEST
|
||||
static void gsp_write_test_sysmem_addr(struct nvgpu_gsp *gsp)
|
||||
{
|
||||
struct gk20a *g;
|
||||
struct nvgpu_falcon *flcn;
|
||||
u64 sysmem_addr;
|
||||
|
||||
g = gsp->g;
|
||||
flcn = gsp->gsp_flcn;
|
||||
|
||||
sysmem_addr = nvgpu_mem_get_addr(g, &gsp->gsp_test.gsp_test_sysmem_block);
|
||||
|
||||
nvgpu_falcon_mailbox_write(flcn, FALCON_MAILBOX_0, u64_lo32(sysmem_addr));
|
||||
|
||||
nvgpu_falcon_mailbox_write(flcn, FALCON_MAILBOX_1, u64_hi32(sysmem_addr));
|
||||
|
||||
}
|
||||
#endif
|
||||
|
||||
static int gsp_ucode_load_and_bootstrap(struct gk20a *g,
|
||||
struct nvgpu_falcon *flcn,
|
||||
struct gsp_fw *gsp_ucode)
|
||||
struct nvgpu_falcon *flcn, struct gsp_fw *gsp_ucode)
|
||||
{
|
||||
u32 dmem_size = 0U;
|
||||
u32 code_size = gsp_ucode->code->size;
|
||||
@@ -197,8 +139,9 @@ static int gsp_ucode_load_and_bootstrap(struct gk20a *g,
|
||||
* Update the address of the allocated sysmem block in the
|
||||
* mailbox register for stress test.
|
||||
*/
|
||||
if (g->gsp->gsp_test.load_stress_test)
|
||||
gsp_write_test_sysmem_addr(g->gsp);
|
||||
if (nvgpu_gsp_get_stress_test_load(g)) {
|
||||
nvgpu_gsp_write_test_sysmem_addr(g);
|
||||
}
|
||||
#endif
|
||||
|
||||
g->ops.falcon.bootstrap(flcn, 0x0);
|
||||
@@ -240,32 +183,23 @@ exit:
|
||||
return -1;
|
||||
}
|
||||
|
||||
static int gsp_wait_for_mailbox_update(struct nvgpu_gsp *gsp,
|
||||
u32 mailbox_index, signed int timeoutms)
|
||||
int nvgpu_gsp_wait_for_mailbox_update(struct nvgpu_gsp *gsp,
|
||||
u32 mailbox_index, u32 exp_value, signed int timeoutms)
|
||||
{
|
||||
u32 mail_box_data = 0;
|
||||
u32 pass_val = 0;
|
||||
struct nvgpu_falcon *flcn = gsp->gsp_flcn;
|
||||
|
||||
nvgpu_log_fn(flcn->g, " ");
|
||||
|
||||
#ifdef CONFIG_NVGPU_GSP_STRESS_TEST
|
||||
if (gsp->gsp_test.load_stress_test) {
|
||||
pass_val = GSP_STRESS_TEST_MAILBOX_PASS;
|
||||
}
|
||||
#endif
|
||||
|
||||
do {
|
||||
mail_box_data = flcn->g->ops.falcon.mailbox_read(
|
||||
flcn, mailbox_index);
|
||||
if (mail_box_data != 0U) {
|
||||
if ((pass_val == 0U) || (mail_box_data == pass_val)) {
|
||||
if (mail_box_data == exp_value) {
|
||||
nvgpu_info(flcn->g,
|
||||
"gsp mailbox-0 updated successful with 0x%x",
|
||||
mail_box_data);
|
||||
break;
|
||||
}
|
||||
}
|
||||
|
||||
if (timeoutms <= 0) {
|
||||
nvgpu_err(flcn->g, "gsp mailbox check timedout");
|
||||
@@ -280,7 +214,7 @@ static int gsp_wait_for_mailbox_update(struct nvgpu_gsp *gsp,
|
||||
return 0;
|
||||
}
|
||||
|
||||
int gsp_bootstrap_ns(struct gk20a *g, struct nvgpu_gsp *gsp)
|
||||
int nvgpu_gsp_bootstrap_ns(struct gk20a *g, struct nvgpu_gsp *gsp)
|
||||
{
|
||||
int err = 0;
|
||||
struct gsp_fw *gsp_ucode = &gsp->gsp_ucode;
|
||||
@@ -288,46 +222,34 @@ int gsp_bootstrap_ns(struct gk20a *g, struct nvgpu_gsp *gsp)
|
||||
|
||||
nvgpu_log_fn(g, " ");
|
||||
|
||||
err = gsp_read_firmware(g, gsp_ucode);
|
||||
err = gsp_read_firmware(g, gsp, gsp_ucode);
|
||||
if (err != 0) {
|
||||
nvgpu_err(g, "gsp firmware reading failed");
|
||||
goto exit;
|
||||
}
|
||||
|
||||
/* core reset */
|
||||
err = nvgpu_falcon_reset(gsp->gsp_flcn);
|
||||
err = nvgpu_falcon_reset(flcn);
|
||||
if (err != 0) {
|
||||
nvgpu_err(g, "gsp core reset failed err=%d", err);
|
||||
goto exit;
|
||||
}
|
||||
|
||||
/* Enable required interrupts support and isr */
|
||||
nvgpu_gsp_isr_support(g, true);
|
||||
nvgpu_gsp_isr_support(g, gsp, true);
|
||||
|
||||
/* setup falcon apertures */
|
||||
if (flcn->flcn_engine_dep_ops.setup_bootstrap_config != NULL) {
|
||||
flcn->flcn_engine_dep_ops.setup_bootstrap_config(flcn->g);
|
||||
}
|
||||
|
||||
err = gsp_ucode_load_and_bootstrap(g, gsp->gsp_flcn, gsp_ucode);
|
||||
err = gsp_ucode_load_and_bootstrap(g, flcn, gsp_ucode);
|
||||
if (err != 0) {
|
||||
nvgpu_err(g, "gsp load and bootstrap failed");
|
||||
goto exit;
|
||||
}
|
||||
|
||||
err = gsp_check_for_brom_completion(gsp->gsp_flcn, GSP_WAIT_TIME_MS);
|
||||
err = gsp_check_for_brom_completion(flcn, GSP_WAIT_TIME_MS);
|
||||
if (err != 0) {
|
||||
nvgpu_err(g, "gsp BROM failed");
|
||||
goto exit;
|
||||
}
|
||||
|
||||
/* wait for mailbox-0 update with non-zero value */
|
||||
err = gsp_wait_for_mailbox_update(gsp, 0x0, GSP_WAIT_TIME_MS);
|
||||
if (err != 0) {
|
||||
nvgpu_err(g, "gsp ucode failed to update mailbox-0");
|
||||
}
|
||||
|
||||
exit:
|
||||
gsp_release_firmware(g, g->gsp);
|
||||
gsp_release_firmware(g, gsp);
|
||||
return err;
|
||||
}
|
||||
|
||||
@@ -1,5 +1,5 @@
|
||||
/*
|
||||
* Copyright (c) 2021, NVIDIA CORPORATION. All rights reserved.
|
||||
* Copyright (c) 2021-2022, NVIDIA CORPORATION. All rights reserved.
|
||||
*
|
||||
* Permission is hereby granted, free of charge, to any person obtaining a
|
||||
* copy of this software and associated documentation files (the "Software"),
|
||||
@@ -26,103 +26,48 @@
|
||||
#include <nvgpu/log.h>
|
||||
#include <nvgpu/gsp.h>
|
||||
#ifdef CONFIG_NVGPU_GSP_STRESS_TEST
|
||||
#include <nvgpu/dma.h>
|
||||
#include <nvgpu/gsp/gsp_test.h>
|
||||
#endif
|
||||
#ifdef CONFIG_NVGPU_GSP_SCHEDULER
|
||||
#include <nvgpu/gsp_sched.h>
|
||||
#endif
|
||||
|
||||
#include "ipc/gsp_seq.h"
|
||||
#include "ipc/gsp_queue.h"
|
||||
#include "gsp_priv.h"
|
||||
#include "gsp_bootstrap.h"
|
||||
|
||||
void nvgpu_gsp_isr_support(struct gk20a *g, bool enable)
|
||||
void nvgpu_gsp_isr_support(struct gk20a *g, struct nvgpu_gsp *gsp, bool enable)
|
||||
{
|
||||
nvgpu_log_fn(g, " ");
|
||||
|
||||
/* Enable irq*/
|
||||
nvgpu_mutex_acquire(&g->gsp->isr_mutex);
|
||||
nvgpu_mutex_acquire(&gsp->isr_mutex);
|
||||
if (g->ops.gsp.enable_irq != NULL) {
|
||||
g->ops.gsp.enable_irq(g, enable);
|
||||
}
|
||||
g->gsp->isr_enabled = enable;
|
||||
nvgpu_mutex_release(&g->gsp->isr_mutex);
|
||||
gsp->isr_enabled = enable;
|
||||
nvgpu_mutex_release(&gsp->isr_mutex);
|
||||
}
|
||||
|
||||
void nvgpu_gsp_suspend(struct gk20a *g)
|
||||
void nvgpu_gsp_suspend(struct gk20a *g, struct nvgpu_gsp *gsp)
|
||||
{
|
||||
nvgpu_gsp_isr_support(g, false);
|
||||
}
|
||||
nvgpu_gsp_isr_support(g, gsp, false);
|
||||
|
||||
void nvgpu_gsp_sw_deinit(struct gk20a *g)
|
||||
{
|
||||
if (g->gsp != NULL) {
|
||||
|
||||
nvgpu_mutex_destroy(&g->gsp->isr_mutex);
|
||||
#ifdef CONFIG_NVGPU_FALCON_DEBUG
|
||||
nvgpu_falcon_dbg_buf_destroy(g->gsp->gsp_flcn);
|
||||
nvgpu_falcon_dbg_error_print_enable(gsp->gsp_flcn, false);
|
||||
#endif
|
||||
#ifdef CONFIG_NVGPU_GSP_STRESS_TEST
|
||||
nvgpu_dma_free(g, &g->gsp->gsp_test.gsp_test_sysmem_block);
|
||||
#endif
|
||||
nvgpu_gsp_sequences_free(g, g->gsp->sequences);
|
||||
|
||||
nvgpu_gsp_queues_free(g, g->gsp->queues);
|
||||
|
||||
g->gsp->gsp_ready = false;
|
||||
|
||||
nvgpu_kfree(g, g->gsp);
|
||||
g->gsp = NULL;
|
||||
}
|
||||
}
|
||||
|
||||
int nvgpu_gsp_sw_init(struct gk20a *g)
|
||||
void nvgpu_gsp_sw_deinit(struct gk20a *g, struct nvgpu_gsp *gsp)
|
||||
{
|
||||
int err = 0;
|
||||
struct nvgpu_gsp *gsp;
|
||||
|
||||
nvgpu_log_fn(g, " ");
|
||||
|
||||
if (g->gsp != NULL) {
|
||||
/*
|
||||
* Recovery/unrailgate case, we do not need to do gsp init as
|
||||
* gsp is set during cold boot & doesn't execute gsp clean up as
|
||||
* part of power off sequence, so reuse to perform faster boot.
|
||||
*/
|
||||
#ifdef CONFIG_NVGPU_GSP_STRESS_TEST
|
||||
nvgpu_gsp_stress_test_bootstrap(g, false);
|
||||
#endif
|
||||
return err;
|
||||
}
|
||||
|
||||
/* Init struct holding the gsp software state */
|
||||
g->gsp = (struct nvgpu_gsp *)nvgpu_kzalloc(g, sizeof(struct nvgpu_gsp));
|
||||
if (g->gsp == NULL) {
|
||||
err = -ENOMEM;
|
||||
goto exit;
|
||||
}
|
||||
|
||||
gsp = g->gsp;
|
||||
|
||||
gsp->g = g;
|
||||
|
||||
/* gsp falcon software state */
|
||||
gsp->gsp_flcn = &g->gsp_flcn;
|
||||
|
||||
/* Init isr mutex */
|
||||
nvgpu_mutex_init(&gsp->isr_mutex);
|
||||
|
||||
err = nvgpu_gsp_sequences_init(g, g->gsp);
|
||||
if (err != 0) {
|
||||
nvgpu_err(g, "GSP sequences init failed");
|
||||
if (gsp != NULL) {
|
||||
nvgpu_mutex_destroy(&gsp->isr_mutex);
|
||||
nvgpu_kfree(g, g->gsp);
|
||||
g->gsp = NULL;
|
||||
#ifdef CONFIG_NVGPU_FALCON_DEBUG
|
||||
nvgpu_falcon_dbg_buf_destroy(gsp->gsp_flcn);
|
||||
#endif
|
||||
|
||||
nvgpu_kfree(g, gsp);
|
||||
gsp = NULL;
|
||||
}
|
||||
}
|
||||
|
||||
exit:
|
||||
return err;
|
||||
}
|
||||
|
||||
int nvgpu_gsp_bootstrap(struct gk20a *g)
|
||||
int nvgpu_gsp_debug_buf_init(struct gk20a *g, u32 queue_no, u32 buffer_size)
|
||||
{
|
||||
int err = 0;
|
||||
|
||||
@@ -133,146 +78,47 @@ int nvgpu_gsp_bootstrap(struct gk20a *g)
|
||||
if ((g->ops.gsp.gsp_get_queue_head != NULL) &&
|
||||
(g->ops.gsp.gsp_get_queue_tail != NULL)) {
|
||||
err = nvgpu_falcon_dbg_buf_init(
|
||||
g->gsp->gsp_flcn, GSP_DMESG_BUFFER_SIZE,
|
||||
g->ops.gsp.gsp_get_queue_head(GSP_DEBUG_BUFFER_QUEUE),
|
||||
g->ops.gsp.gsp_get_queue_tail(GSP_DEBUG_BUFFER_QUEUE));
|
||||
&g->gsp_flcn, buffer_size,
|
||||
g->ops.gsp.gsp_get_queue_head(queue_no),
|
||||
g->ops.gsp.gsp_get_queue_tail(queue_no));
|
||||
if (err != 0) {
|
||||
nvgpu_err(g, "GSP debug init failed");
|
||||
goto de_init;
|
||||
}
|
||||
}
|
||||
#endif
|
||||
|
||||
err = gsp_bootstrap_ns(g, g->gsp);
|
||||
if (err != 0) {
|
||||
nvgpu_err(g, "GSP bootstrap failed");
|
||||
goto de_init;
|
||||
}
|
||||
|
||||
return err;
|
||||
de_init:
|
||||
nvgpu_gsp_sw_deinit(g);
|
||||
return err;
|
||||
}
|
||||
|
||||
void nvgpu_gsp_isr_mutex_aquire(struct gk20a *g)
|
||||
void nvgpu_gsp_isr_mutex_acquire(struct gk20a *g, struct nvgpu_gsp *gsp)
|
||||
{
|
||||
struct nvgpu_gsp *gsp = g->gsp;
|
||||
|
||||
nvgpu_mutex_acquire(&gsp->isr_mutex);
|
||||
}
|
||||
|
||||
void nvgpu_gsp_isr_mutex_release(struct gk20a *g)
|
||||
void nvgpu_gsp_isr_mutex_release(struct gk20a *g, struct nvgpu_gsp *gsp)
|
||||
{
|
||||
struct nvgpu_gsp *gsp = g->gsp;
|
||||
|
||||
nvgpu_mutex_release(&gsp->isr_mutex);
|
||||
}
|
||||
|
||||
bool nvgpu_gsp_is_isr_enable(struct gk20a *g)
|
||||
bool nvgpu_gsp_is_isr_enable(struct gk20a *g, struct nvgpu_gsp *gsp)
|
||||
{
|
||||
struct nvgpu_gsp *gsp = g->gsp;
|
||||
|
||||
return gsp->isr_enabled;
|
||||
}
|
||||
|
||||
struct nvgpu_falcon *nvgpu_gsp_falcon_instance(struct gk20a *g)
|
||||
{
|
||||
struct nvgpu_gsp *gsp = g->gsp;
|
||||
|
||||
return gsp->gsp_flcn;
|
||||
return &g->gsp_flcn;
|
||||
}
|
||||
|
||||
void nvgpu_gsp_isr(struct gk20a *g)
|
||||
{
|
||||
#ifdef CONFIG_NVGPU_GSP_STRESS_TEST
|
||||
int nvgpu_gsp_stress_test_bootstrap(struct gk20a *g, bool start)
|
||||
{
|
||||
int err = 0;
|
||||
struct nvgpu_gsp *gsp;
|
||||
|
||||
nvgpu_log_fn(g, " ");
|
||||
|
||||
gsp = g->gsp;
|
||||
|
||||
if (gsp == NULL) {
|
||||
nvgpu_err(g, "GSP not initialized");
|
||||
err = -EFAULT;
|
||||
goto exit;
|
||||
}
|
||||
|
||||
if (!start && !(gsp->gsp_test.load_stress_test))
|
||||
return err;
|
||||
|
||||
if (start) {
|
||||
err = nvgpu_dma_alloc_flags_sys(g,
|
||||
NVGPU_DMA_PHYSICALLY_ADDRESSED,
|
||||
SZ_64K,
|
||||
&g->gsp->gsp_test.gsp_test_sysmem_block);
|
||||
if (err != 0) {
|
||||
nvgpu_err(g, "GSP test memory alloc failed");
|
||||
goto exit;
|
||||
}
|
||||
}
|
||||
|
||||
gsp->gsp_test.load_stress_test = true;
|
||||
|
||||
err = nvgpu_gsp_bootstrap(g);
|
||||
if (err != 0) {
|
||||
nvgpu_err(g, "GSP bootstrap failed for stress test");
|
||||
goto exit;
|
||||
}
|
||||
|
||||
if (gsp->gsp_test.enable_stress_test) {
|
||||
nvgpu_info(g, "Restarting GSP stress test");
|
||||
nvgpu_falcon_mailbox_write(gsp->gsp_flcn, FALCON_MAILBOX_1, 0xFFFFFFFF);
|
||||
}
|
||||
|
||||
return err;
|
||||
|
||||
exit:
|
||||
gsp->gsp_test.load_stress_test = false;
|
||||
|
||||
return err;
|
||||
}
|
||||
|
||||
int nvgpu_gsp_stress_test_halt(struct gk20a *g, bool restart)
|
||||
{
|
||||
int err = 0;
|
||||
struct nvgpu_gsp *gsp;
|
||||
|
||||
nvgpu_log_fn(g, " ");
|
||||
|
||||
gsp = g->gsp;
|
||||
|
||||
if ((gsp == NULL)) {
|
||||
nvgpu_info(g, "GSP not initialized");
|
||||
goto exit;
|
||||
}
|
||||
|
||||
if (restart && (gsp->gsp_test.load_stress_test == false)) {
|
||||
nvgpu_info(g, "GSP stress test not loaded ");
|
||||
goto exit;
|
||||
}
|
||||
|
||||
err = nvgpu_falcon_reset(gsp->gsp_flcn);
|
||||
if (err != 0) {
|
||||
nvgpu_err(g, "gsp reset failed err=%d", err);
|
||||
goto exit;
|
||||
}
|
||||
|
||||
if (!restart) {
|
||||
gsp->gsp_test.load_stress_test = false;
|
||||
nvgpu_dma_free(g, &g->gsp->gsp_test.gsp_test_sysmem_block);
|
||||
}
|
||||
|
||||
exit:
|
||||
return err;
|
||||
}
|
||||
|
||||
bool nvgpu_gsp_is_stress_test(struct gk20a *g)
|
||||
{
|
||||
if (g->gsp->gsp_test.load_stress_test)
|
||||
return true;
|
||||
else
|
||||
return false;
|
||||
if (nvgpu_gsp_get_stress_test_load(g)) {
|
||||
nvgpu_gsp_stest_isr(g);
|
||||
return;
|
||||
}
|
||||
#endif
|
||||
#ifdef CONFIG_NVGPU_GSP_SCHEDULER
|
||||
nvgpu_gsp_sched_isr(g);
|
||||
#endif
|
||||
return;
|
||||
}
|
||||
|
||||
@@ -1,114 +0,0 @@
|
||||
// SPDX-License-Identifier: GPL-2.0-or-later
|
||||
/*
|
||||
* GSP Test Functions
|
||||
*
|
||||
* Copyright (c) 2021, NVIDIA Corporation. All rights reserved.
|
||||
*
|
||||
* This program is free software; you can redistribute it and/or modify it
|
||||
* under the terms and conditions of the GNU General Public License,
|
||||
* version 2, as published by the Free Software Foundation.
|
||||
*
|
||||
* This program is distributed in the hope it will be useful, but WITHOUT
|
||||
* ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
|
||||
* FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for
|
||||
* more details.
|
||||
*
|
||||
* You should have received a copy of the GNU General Public License
|
||||
* along with this program. If not, see <http://www.gnu.org/licenses/>.
|
||||
*/
|
||||
#include <nvgpu/firmware.h>
|
||||
#include <nvgpu/falcon.h>
|
||||
#include <nvgpu/gk20a.h>
|
||||
#include <nvgpu/log.h>
|
||||
#include <nvgpu/gsp.h>
|
||||
#include <nvgpu/gsp/gsp_test.h>
|
||||
|
||||
#include "gsp_priv.h"
|
||||
#include "gsp_bootstrap.h"
|
||||
|
||||
u32 nvgpu_gsp_get_current_iteration(struct gk20a *g)
|
||||
{
|
||||
u32 data = 0;
|
||||
struct nvgpu_gsp *gsp = g->gsp;
|
||||
|
||||
nvgpu_log_fn(g, " ");
|
||||
|
||||
data = nvgpu_falcon_mailbox_read(gsp->gsp_flcn, FALCON_MAILBOX_1);
|
||||
|
||||
return data;
|
||||
}
|
||||
|
||||
u32 nvgpu_gsp_get_current_test(struct gk20a *g)
|
||||
{
|
||||
u32 data = 0;
|
||||
struct nvgpu_gsp *gsp = g->gsp;
|
||||
|
||||
nvgpu_log_fn(g, " ");
|
||||
|
||||
data = nvgpu_falcon_mailbox_read(gsp->gsp_flcn, FALCON_MAILBOX_0);
|
||||
|
||||
return data;
|
||||
}
|
||||
|
||||
bool nvgpu_gsp_get_test_fail_status(struct gk20a *g)
|
||||
{
|
||||
struct nvgpu_gsp *gsp = g->gsp;
|
||||
|
||||
return gsp->gsp_test.stress_test_fail_status;
|
||||
}
|
||||
|
||||
bool nvgpu_gsp_get_stress_test_start(struct gk20a *g)
|
||||
{
|
||||
struct nvgpu_gsp *gsp = g->gsp;
|
||||
|
||||
return gsp->gsp_test.enable_stress_test;
|
||||
}
|
||||
|
||||
bool nvgpu_gsp_get_stress_test_load(struct gk20a *g)
|
||||
{
|
||||
struct nvgpu_gsp *gsp = g->gsp;
|
||||
|
||||
if (gsp == NULL)
|
||||
return false;
|
||||
|
||||
return gsp->gsp_test.load_stress_test;
|
||||
}
|
||||
|
||||
void nvgpu_gsp_set_test_fail_status(struct gk20a *g, bool val)
|
||||
{
|
||||
struct nvgpu_gsp *gsp = g->gsp;
|
||||
|
||||
gsp->gsp_test.stress_test_fail_status = val;
|
||||
}
|
||||
|
||||
int nvgpu_gsp_set_stress_test_start(struct gk20a *g, bool flag)
|
||||
{
|
||||
int err = 0;
|
||||
struct nvgpu_gsp *gsp = g->gsp;
|
||||
|
||||
nvgpu_log_fn(g, " ");
|
||||
|
||||
if (flag) {
|
||||
nvgpu_info(g, "Enabling GSP test");
|
||||
nvgpu_falcon_mailbox_write(gsp->gsp_flcn, FALCON_MAILBOX_1, 0xFFFFFFFF);
|
||||
} else {
|
||||
nvgpu_info(g, "Halting GSP test");
|
||||
nvgpu_gsp_stress_test_halt(g, false);
|
||||
}
|
||||
|
||||
gsp->gsp_test.enable_stress_test = flag;
|
||||
|
||||
return err;
|
||||
}
|
||||
|
||||
int nvgpu_gsp_set_stress_test_load(struct gk20a *g, bool flag)
|
||||
{
|
||||
int err = 0;
|
||||
|
||||
nvgpu_log_fn(g, " ");
|
||||
|
||||
if (flag)
|
||||
err = nvgpu_gsp_stress_test_bootstrap(g, flag);
|
||||
|
||||
return err;
|
||||
}
|
||||
184
drivers/gpu/nvgpu/common/gsp_scheduler/gsp_scheduler.c
Normal file
184
drivers/gpu/nvgpu/common/gsp_scheduler/gsp_scheduler.c
Normal file
@@ -0,0 +1,184 @@
|
||||
/*
|
||||
* Copyright (c) 2022, NVIDIA CORPORATION. All rights reserved.
|
||||
*
|
||||
* Permission is hereby granted, free of charge, to any person obtaining a
|
||||
* copy of this software and associated documentation files (the "Software"),
|
||||
* to deal in the Software without restriction, including without limitation
|
||||
* the rights to use, copy, modify, merge, publish, distribute, sublicense,
|
||||
* and/or sell copies of the Software, and to permit persons to whom the
|
||||
* Software is furnished to do so, subject to the following conditions:
|
||||
*
|
||||
* The above copyright notice and this permission notice shall be included in
|
||||
* all copies or substantial portions of the Software.
|
||||
*
|
||||
* THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
|
||||
* IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
|
||||
* FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
|
||||
* THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
|
||||
* LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING
|
||||
* FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER
|
||||
* DEALINGS IN THE SOFTWARE.
|
||||
*/
|
||||
|
||||
#include <nvgpu/gk20a.h>
|
||||
#include <nvgpu/string.h>
|
||||
#include <nvgpu/kmem.h>
|
||||
#include <nvgpu/log.h>
|
||||
#include <nvgpu/errno.h>
|
||||
#include <nvgpu/gsp.h>
|
||||
#include <nvgpu/gsp_sched.h>
|
||||
|
||||
#include "gsp_scheduler.h"
|
||||
#include "ipc/gsp_seq.h"
|
||||
#include "ipc/gsp_queue.h"
|
||||
|
||||
static void gsp_sched_get_file_names(struct gk20a *g, struct gsp_fw *gsp_ucode)
|
||||
{
|
||||
nvgpu_log_fn(g, " ");
|
||||
|
||||
gsp_ucode->code_name = GSP_DBG_RISCV_FW_CODE;
|
||||
gsp_ucode->data_name = GSP_DBG_RISCV_FW_DATA;
|
||||
gsp_ucode->manifest_name = GSP_DBG_RISCV_FW_MANIFEST;
|
||||
}
|
||||
|
||||
void nvgpu_gsp_sched_suspend(struct gk20a *g, struct nvgpu_gsp_sched *gsp_sched)
|
||||
{
|
||||
struct nvgpu_gsp *gsp = gsp_sched->gsp;
|
||||
|
||||
nvgpu_log_fn(g, " ");
|
||||
|
||||
if (gsp == NULL) {
|
||||
nvgpu_info(g, "GSP not initialized");
|
||||
return;
|
||||
}
|
||||
|
||||
gsp_sched->gsp_ready = false;
|
||||
nvgpu_gsp_suspend(g, gsp);
|
||||
}
|
||||
|
||||
static void gsp_sched_deinit(struct gk20a *g, struct nvgpu_gsp_sched *gsp_sched)
|
||||
{
|
||||
gsp_sched->gsp_ready = false;
|
||||
|
||||
nvgpu_kfree(g, gsp_sched);
|
||||
gsp_sched = NULL;
|
||||
}
|
||||
|
||||
void nvgpu_gsp_sched_sw_deinit(struct gk20a *g)
|
||||
{
|
||||
struct nvgpu_gsp_sched *gsp_sched = g->gsp_sched;
|
||||
|
||||
nvgpu_log_fn(g, " ");
|
||||
|
||||
if (gsp_sched == NULL) {
|
||||
return;
|
||||
}
|
||||
|
||||
if (gsp_sched->gsp != NULL) {
|
||||
nvgpu_gsp_sw_deinit(g, gsp_sched->gsp);
|
||||
}
|
||||
|
||||
if (gsp_sched->sequences != NULL) {
|
||||
nvgpu_gsp_sequences_free(g, gsp_sched->sequences);
|
||||
}
|
||||
|
||||
if (gsp_sched->queues != NULL) {
|
||||
nvgpu_gsp_queues_free(g, gsp_sched->queues);
|
||||
}
|
||||
|
||||
if (gsp_sched != NULL) {
|
||||
gsp_sched_deinit(g, gsp_sched);
|
||||
}
|
||||
}
|
||||
|
||||
int nvgpu_gsp_sched_sw_init(struct gk20a *g)
|
||||
{
|
||||
int err = 0;
|
||||
struct nvgpu_gsp_sched *gsp_sched;
|
||||
struct nvgpu_gsp *gsp;
|
||||
|
||||
nvgpu_log_fn(g, " ");
|
||||
|
||||
if (g->gsp_sched != NULL) {
|
||||
/*
|
||||
* Recovery/unrailgate case, we do not need to do gsp_sched init as
|
||||
* gsp_sched is set during cold boot & doesn't execute gsp_sched clean
|
||||
* up as part of power off sequence, so reuse to perform faster boot.
|
||||
*/
|
||||
return err;
|
||||
}
|
||||
|
||||
/* Init struct holding the gsp sched software state */
|
||||
g->gsp_sched = (struct nvgpu_gsp_sched *)
|
||||
nvgpu_kzalloc(g, sizeof(struct nvgpu_gsp_sched));
|
||||
if (g->gsp_sched == NULL) {
|
||||
err = -ENOMEM;
|
||||
goto de_init;
|
||||
}
|
||||
|
||||
/* Init struct holding the gsp software state */
|
||||
g->gsp_sched->gsp = (struct nvgpu_gsp *)
|
||||
nvgpu_kzalloc(g, sizeof(struct nvgpu_gsp));
|
||||
if (g->gsp_sched->gsp == NULL) {
|
||||
err = -ENOMEM;
|
||||
goto de_init;
|
||||
}
|
||||
|
||||
gsp_sched = g->gsp_sched;
|
||||
gsp = g->gsp_sched->gsp;
|
||||
|
||||
/* gsp falcon software state */
|
||||
gsp->gsp_flcn = &g->gsp_flcn;
|
||||
gsp->g = g;
|
||||
|
||||
/* Init isr mutex */
|
||||
nvgpu_mutex_init(&gsp->isr_mutex);
|
||||
|
||||
err = nvgpu_gsp_sequences_init(g, gsp_sched);
|
||||
if (err != 0) {
|
||||
nvgpu_err(g, "GSP sequences init failed");
|
||||
goto de_init;
|
||||
}
|
||||
|
||||
nvgpu_log_fn(g, " Done ");
|
||||
return err;
|
||||
de_init:
|
||||
nvgpu_gsp_sched_sw_deinit(g);
|
||||
return err;
|
||||
}
|
||||
|
||||
int nvgpu_gsp_sched_bootstrap_ns(struct gk20a *g)
|
||||
{
|
||||
struct nvgpu_gsp_sched *gsp_sched = g->gsp_sched;
|
||||
int status = 0;
|
||||
|
||||
#ifdef CONFIG_NVGPU_FALCON_DEBUG
|
||||
status = nvgpu_gsp_debug_buf_init(g, GSP_SCHED_DEBUG_BUFFER_QUEUE,
|
||||
GSP_SCHED_DMESG_BUFFER_SIZE);
|
||||
if (status != 0) {
|
||||
nvgpu_err(g, "GSP sched debug buf init failed");
|
||||
goto de_init;
|
||||
}
|
||||
#endif
|
||||
|
||||
/* Get ucode file names */
|
||||
gsp_sched_get_file_names(g, &gsp_sched->gsp->gsp_ucode);
|
||||
|
||||
status = nvgpu_gsp_bootstrap_ns(g, gsp_sched->gsp);
|
||||
if (status != 0) {
|
||||
nvgpu_err(g, " GSP sched bootstrap failed ");
|
||||
goto de_init;
|
||||
}
|
||||
|
||||
return status;
|
||||
de_init:
|
||||
nvgpu_gsp_sched_sw_deinit(g);
|
||||
return status;
|
||||
}
|
||||
|
||||
void nvgpu_gsp_sched_isr(struct gk20a *g)
|
||||
{
|
||||
struct nvgpu_gsp *gsp = g->gsp_sched->gsp;
|
||||
|
||||
g->ops.gsp.gsp_isr(g, gsp);
|
||||
}
|
||||
49
drivers/gpu/nvgpu/common/gsp_scheduler/gsp_scheduler.h
Normal file
49
drivers/gpu/nvgpu/common/gsp_scheduler/gsp_scheduler.h
Normal file
@@ -0,0 +1,49 @@
|
||||
/*
|
||||
* Copyright (c) 2022, NVIDIA CORPORATION. All rights reserved.
|
||||
*
|
||||
* Permission is hereby granted, free of charge, to any person obtaining a
|
||||
* copy of this software and associated documentation files (the "Software"),
|
||||
* to deal in the Software without restriction, including without limitation
|
||||
* the rights to use, copy, modify, merge, publish, distribute, sublicense,
|
||||
* and/or sell copies of the Software, and to permit persons to whom the
|
||||
* Software is furnished to do so, subject to the following conditions:
|
||||
*
|
||||
* The above copyright notice and this permission notice shall be included in
|
||||
* all copies or substantial portions of the Software.
|
||||
*
|
||||
* THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
|
||||
* IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
|
||||
* FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
|
||||
* THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
|
||||
* LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING
|
||||
* FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER
|
||||
* DEALINGS IN THE SOFTWARE.
|
||||
*/
|
||||
|
||||
#ifndef GSP_SCHEDULER_H
|
||||
#define GSP_SCHEDULER_H
|
||||
|
||||
#define GSP_SCHED_DEBUG_BUFFER_QUEUE 3U
|
||||
#define GSP_SCHED_DMESG_BUFFER_SIZE 0x1000U
|
||||
|
||||
#define GSP_QUEUE_NUM 2U
|
||||
|
||||
#define GSP_DBG_RISCV_FW_MANIFEST "sample-gsp.manifest.encrypt.bin.out.bin"
|
||||
#define GSP_DBG_RISCV_FW_CODE "sample-gsp.text.encrypt.bin"
|
||||
#define GSP_DBG_RISCV_FW_DATA "sample-gsp.data.encrypt.bin"
|
||||
|
||||
/* GSP descriptor's */
|
||||
struct nvgpu_gsp_sched {
|
||||
struct nvgpu_gsp *gsp;
|
||||
|
||||
struct gsp_sequences *sequences;
|
||||
|
||||
struct nvgpu_engine_mem_queue *queues[GSP_QUEUE_NUM];
|
||||
|
||||
u32 command_ack;
|
||||
|
||||
/* set to true once init received */
|
||||
bool gsp_ready;
|
||||
};
|
||||
|
||||
#endif /* GSP_SCHEDULER_H */
|
||||
@@ -1,5 +1,5 @@
|
||||
/*
|
||||
* Copyright (c) 2021, NVIDIA CORPORATION. All rights reserved.
|
||||
* Copyright (c) 2021-2022, NVIDIA CORPORATION. All rights reserved.
|
||||
*
|
||||
* Permission is hereby granted, free of charge, to any person obtaining a
|
||||
* copy of this software and associated documentation files (the "Software"),
|
||||
@@ -25,7 +25,7 @@
|
||||
#include <nvgpu/log.h>
|
||||
#include <nvgpu/gsp.h>
|
||||
|
||||
#include "../gsp_priv.h"
|
||||
#include "../gsp_scheduler.h"
|
||||
#include "gsp_seq.h"
|
||||
#include "gsp_queue.h"
|
||||
#include "gsp_cmd.h"
|
||||
@@ -35,10 +35,10 @@ u8 gsp_unit_id_is_valid(u8 id)
|
||||
return (id < NV_GSP_UNIT_END);
|
||||
}
|
||||
|
||||
static bool gsp_validate_cmd(struct nvgpu_gsp *gsp,
|
||||
static bool gsp_validate_cmd(struct nvgpu_gsp_sched *gsp_sched,
|
||||
struct nv_flcn_cmd_gsp *cmd, u32 queue_id)
|
||||
{
|
||||
struct gk20a *g = gsp->g;
|
||||
struct gk20a *g = gsp_sched->gsp->g;
|
||||
u32 queue_size;
|
||||
|
||||
if (queue_id != GSP_NV_CMDQ_LOG_ID) {
|
||||
@@ -49,7 +49,7 @@ static bool gsp_validate_cmd(struct nvgpu_gsp *gsp,
|
||||
goto invalid_cmd;
|
||||
}
|
||||
|
||||
queue_size = nvgpu_gsp_queue_get_size(gsp->queues, queue_id);
|
||||
queue_size = nvgpu_gsp_queue_get_size(gsp_sched->queues, queue_id);
|
||||
|
||||
if (cmd->hdr.size > (queue_size >> 1)) {
|
||||
goto invalid_cmd;
|
||||
@@ -69,12 +69,13 @@ invalid_cmd:
|
||||
return false;
|
||||
}
|
||||
|
||||
static int gsp_write_cmd(struct nvgpu_gsp *gsp,
|
||||
static int gsp_write_cmd(struct nvgpu_gsp_sched *gsp_sched,
|
||||
struct nv_flcn_cmd_gsp *cmd, u32 queue_id,
|
||||
u32 timeout_ms)
|
||||
{
|
||||
struct nvgpu_timeout timeout;
|
||||
struct gk20a *g = gsp->g;
|
||||
struct gk20a *g = gsp_sched->gsp->g;
|
||||
struct nvgpu_gsp *gsp = gsp_sched->gsp;
|
||||
int err;
|
||||
|
||||
nvgpu_log_fn(g, " ");
|
||||
@@ -82,7 +83,7 @@ static int gsp_write_cmd(struct nvgpu_gsp *gsp,
|
||||
nvgpu_timeout_init_cpu_timer(g, &timeout, timeout_ms);
|
||||
|
||||
do {
|
||||
err = nvgpu_gsp_queue_push(gsp->queues, queue_id, gsp->gsp_flcn,
|
||||
err = nvgpu_gsp_queue_push(gsp_sched->queues, queue_id, gsp->gsp_flcn,
|
||||
cmd, cmd->hdr.size);
|
||||
if ((err == -EAGAIN) &&
|
||||
(nvgpu_timeout_expired(&timeout) == 0)) {
|
||||
@@ -103,7 +104,7 @@ int nvgpu_gsp_cmd_post(struct gk20a *g, struct nv_flcn_cmd_gsp *cmd,
|
||||
u32 queue_id, gsp_callback callback,
|
||||
void *cb_param, u32 timeout)
|
||||
{
|
||||
struct nvgpu_gsp *gsp = g->gsp;
|
||||
struct nvgpu_gsp_sched *gsp_sched = g->gsp_sched;
|
||||
struct gsp_sequence *seq = NULL;
|
||||
int err = 0;
|
||||
|
||||
@@ -114,13 +115,13 @@ int nvgpu_gsp_cmd_post(struct gk20a *g, struct nv_flcn_cmd_gsp *cmd,
|
||||
}
|
||||
|
||||
/* Sanity check the command input. */
|
||||
if (!gsp_validate_cmd(gsp, cmd, queue_id)) {
|
||||
if (!gsp_validate_cmd(gsp_sched, cmd, queue_id)) {
|
||||
err = -EINVAL;
|
||||
goto exit;
|
||||
}
|
||||
|
||||
/* Attempt to reserve a sequence for this command. */
|
||||
err = nvgpu_gsp_seq_acquire(g, gsp->sequences, &seq,
|
||||
err = nvgpu_gsp_seq_acquire(g, gsp_sched->sequences, &seq,
|
||||
callback, cb_param);
|
||||
if (err != 0) {
|
||||
goto exit;
|
||||
@@ -134,9 +135,9 @@ int nvgpu_gsp_cmd_post(struct gk20a *g, struct nv_flcn_cmd_gsp *cmd,
|
||||
|
||||
nvgpu_gsp_seq_set_state(seq, GSP_SEQ_STATE_USED);
|
||||
|
||||
err = gsp_write_cmd(gsp, cmd, queue_id, timeout);
|
||||
err = gsp_write_cmd(gsp_sched, cmd, queue_id, timeout);
|
||||
if (err != 0) {
|
||||
gsp_seq_release(gsp->sequences, seq);
|
||||
gsp_seq_release(gsp_sched->sequences, seq);
|
||||
}
|
||||
|
||||
exit:
|
||||
@@ -1,5 +1,5 @@
|
||||
/*
|
||||
* Copyright (c) 2021, NVIDIA CORPORATION. All rights reserved.
|
||||
* Copyright (c) 2021-2022, NVIDIA CORPORATION. All rights reserved.
|
||||
*
|
||||
* Permission is hereby granted, free of charge, to any person obtaining a
|
||||
* copy of this software and associated documentation files (the "Software"),
|
||||
@@ -1,5 +1,5 @@
|
||||
/*
|
||||
* Copyright (c) 2021, NVIDIA CORPORATION. All rights reserved.
|
||||
* Copyright (c) 2021-2022, NVIDIA CORPORATION. All rights reserved.
|
||||
*
|
||||
* Permission is hereby granted, free of charge, to any person obtaining a
|
||||
* copy of this software and associated documentation files (the "Software"),
|
||||
@@ -24,22 +24,22 @@
|
||||
#include <nvgpu/gk20a.h>
|
||||
#include <nvgpu/gsp.h>
|
||||
|
||||
#include "../gsp_priv.h"
|
||||
#include "../gsp_scheduler.h"
|
||||
#include "gsp_queue.h"
|
||||
#include "gsp_msg.h"
|
||||
#include "gsp_seq.h"
|
||||
|
||||
/* Message/Event request handlers */
|
||||
static int gsp_response_handle(struct nvgpu_gsp *gsp,
|
||||
static int gsp_response_handle(struct nvgpu_gsp_sched *gsp_sched,
|
||||
struct nv_flcn_msg_gsp *msg)
|
||||
{
|
||||
struct gk20a *g = gsp->g;
|
||||
struct gk20a *g = gsp_sched->gsp->g;
|
||||
|
||||
return nvgpu_gsp_seq_response_handle(g, gsp->sequences,
|
||||
return nvgpu_gsp_seq_response_handle(g, gsp_sched->sequences,
|
||||
msg, msg->hdr.seq_id);
|
||||
}
|
||||
|
||||
static int gsp_handle_event(struct nvgpu_gsp *gsp,
|
||||
static int gsp_handle_event(struct nvgpu_gsp_sched *gsp_sched,
|
||||
struct nv_flcn_msg_gsp *msg)
|
||||
{
|
||||
int err = 0;
|
||||
@@ -52,29 +52,29 @@ static int gsp_handle_event(struct nvgpu_gsp *gsp,
|
||||
return err;
|
||||
}
|
||||
|
||||
static bool gsp_read_message(struct nvgpu_gsp *gsp,
|
||||
static bool gsp_read_message(struct nvgpu_gsp_sched *gsp_sched,
|
||||
u32 queue_id, struct nv_flcn_msg_gsp *msg, int *status)
|
||||
{
|
||||
struct gk20a *g = gsp->g;
|
||||
struct gk20a *g = gsp_sched->gsp->g;
|
||||
u32 read_size;
|
||||
int err;
|
||||
|
||||
*status = 0U;
|
||||
|
||||
if (nvgpu_gsp_queue_is_empty(gsp->queues, queue_id)) {
|
||||
if (nvgpu_gsp_queue_is_empty(gsp_sched->queues, queue_id)) {
|
||||
return false;
|
||||
}
|
||||
|
||||
if (!nvgpu_gsp_queue_read(g, gsp->queues, queue_id,
|
||||
gsp->gsp_flcn, &msg->hdr,
|
||||
if (!nvgpu_gsp_queue_read(g, gsp_sched->queues, queue_id,
|
||||
gsp_sched->gsp->gsp_flcn, &msg->hdr,
|
||||
GSP_MSG_HDR_SIZE, status)) {
|
||||
nvgpu_err(g, "fail to read msg from queue %d", queue_id);
|
||||
goto clean_up;
|
||||
}
|
||||
|
||||
if (msg->hdr.unit_id == NV_GSP_UNIT_REWIND) {
|
||||
err = nvgpu_gsp_queue_rewind(gsp->gsp_flcn,
|
||||
gsp->queues, queue_id);
|
||||
err = nvgpu_gsp_queue_rewind(gsp_sched->gsp->gsp_flcn,
|
||||
gsp_sched->queues, queue_id);
|
||||
if (err != 0) {
|
||||
nvgpu_err(g, "fail to rewind queue %d", queue_id);
|
||||
*status = err;
|
||||
@@ -82,8 +82,8 @@ static bool gsp_read_message(struct nvgpu_gsp *gsp,
|
||||
}
|
||||
|
||||
/* read again after rewind */
|
||||
if (!nvgpu_gsp_queue_read(g, gsp->queues, queue_id,
|
||||
gsp->gsp_flcn, &msg->hdr,
|
||||
if (!nvgpu_gsp_queue_read(g, gsp_sched->queues, queue_id,
|
||||
gsp_sched->gsp->gsp_flcn, &msg->hdr,
|
||||
GSP_MSG_HDR_SIZE, status)) {
|
||||
nvgpu_err(g, "fail to read msg from queue %d",
|
||||
queue_id);
|
||||
@@ -100,8 +100,8 @@ static bool gsp_read_message(struct nvgpu_gsp *gsp,
|
||||
|
||||
if (msg->hdr.size > GSP_MSG_HDR_SIZE) {
|
||||
read_size = msg->hdr.size - GSP_MSG_HDR_SIZE;
|
||||
if (!nvgpu_gsp_queue_read(g, gsp->queues, queue_id,
|
||||
gsp->gsp_flcn, &msg->msg,
|
||||
if (!nvgpu_gsp_queue_read(g, gsp_sched->queues, queue_id,
|
||||
gsp_sched->gsp->gsp_flcn, &msg->msg,
|
||||
read_size, status)) {
|
||||
nvgpu_err(g, "fail to read msg from queue %d",
|
||||
queue_id);
|
||||
@@ -115,17 +115,18 @@ clean_up:
|
||||
return false;
|
||||
}
|
||||
|
||||
static int gsp_process_init_msg(struct nvgpu_gsp *gsp,
|
||||
static int gsp_process_init_msg(struct nvgpu_gsp_sched *gsp_sched,
|
||||
struct nv_flcn_msg_gsp *msg)
|
||||
{
|
||||
struct gk20a *g = gsp->g;
|
||||
struct gk20a *g = gsp_sched->gsp->g;
|
||||
struct nvgpu_gsp *gsp = gsp_sched->gsp;
|
||||
struct gsp_init_msg_gsp_init *gsp_init;
|
||||
u32 tail = 0;
|
||||
int err = 0;
|
||||
|
||||
g->ops.gsp.msgq_tail(g, gsp, &tail, QUEUE_GET);
|
||||
|
||||
err = nvgpu_falcon_copy_from_emem(gsp->gsp_flcn, tail,
|
||||
err = nvgpu_falcon_copy_from_emem(gsp_sched->gsp->gsp_flcn, tail,
|
||||
(u8 *)&msg->hdr, GSP_MSG_HDR_SIZE, 0U);
|
||||
if (err != 0) {
|
||||
goto exit;
|
||||
@@ -137,7 +138,7 @@ static int gsp_process_init_msg(struct nvgpu_gsp *gsp,
|
||||
goto exit;
|
||||
}
|
||||
|
||||
err = nvgpu_falcon_copy_from_emem(gsp->gsp_flcn, tail + GSP_MSG_HDR_SIZE,
|
||||
err = nvgpu_falcon_copy_from_emem(gsp_sched->gsp->gsp_flcn, tail + GSP_MSG_HDR_SIZE,
|
||||
(u8 *)&msg->msg, msg->hdr.size - GSP_MSG_HDR_SIZE, 0U);
|
||||
if (err != 0) {
|
||||
goto exit;
|
||||
@@ -154,12 +155,12 @@ static int gsp_process_init_msg(struct nvgpu_gsp *gsp,
|
||||
|
||||
gsp_init = &msg->msg.init.gsp_init;
|
||||
|
||||
err = nvgpu_gsp_queues_init(g, gsp->queues, gsp_init);
|
||||
err = nvgpu_gsp_queues_init(g, gsp_sched->queues, gsp_init);
|
||||
if (err != 0) {
|
||||
return err;
|
||||
}
|
||||
|
||||
gsp->gsp_ready = true;
|
||||
gsp_sched->gsp_ready = true;
|
||||
|
||||
exit:
|
||||
return err;
|
||||
@@ -167,20 +168,20 @@ exit:
|
||||
|
||||
int nvgpu_gsp_process_message(struct gk20a *g)
|
||||
{
|
||||
struct nvgpu_gsp *gsp = g->gsp;
|
||||
struct nvgpu_gsp_sched *gsp_sched = g->gsp_sched;
|
||||
struct nv_flcn_msg_gsp msg;
|
||||
bool read_msg;
|
||||
int status = 0;
|
||||
|
||||
nvgpu_log_fn(g, " ");
|
||||
|
||||
if (unlikely(!gsp->gsp_ready)) {
|
||||
status = gsp_process_init_msg(gsp, &msg);
|
||||
if (unlikely(!gsp_sched->gsp_ready)) {
|
||||
status = gsp_process_init_msg(gsp_sched, &msg);
|
||||
goto exit;
|
||||
}
|
||||
|
||||
do {
|
||||
read_msg = gsp_read_message(gsp,
|
||||
read_msg = gsp_read_message(gsp_sched,
|
||||
GSP_NV_MSGQ_LOG_ID, &msg, &status);
|
||||
if (read_msg == false) {
|
||||
break;
|
||||
@@ -195,12 +196,12 @@ int nvgpu_gsp_process_message(struct gk20a *g)
|
||||
msg.hdr.ctrl_flags &= ~GSP_CMD_FLAGS_MASK;
|
||||
|
||||
if (msg.hdr.ctrl_flags == GSP_CMD_FLAGS_EVENT) {
|
||||
gsp_handle_event(gsp, &msg);
|
||||
gsp_handle_event(gsp_sched, &msg);
|
||||
} else {
|
||||
gsp_response_handle(gsp, &msg);
|
||||
gsp_response_handle(gsp_sched, &msg);
|
||||
}
|
||||
|
||||
if (!nvgpu_gsp_queue_is_empty(gsp->queues,
|
||||
if (!nvgpu_gsp_queue_is_empty(gsp_sched->queues,
|
||||
GSP_NV_MSGQ_LOG_ID)) {
|
||||
g->ops.gsp.set_msg_intr(g);
|
||||
}
|
||||
@@ -210,10 +211,9 @@ exit:
|
||||
return status;
|
||||
}
|
||||
|
||||
int nvgpu_gsp_wait_message_cond(struct nvgpu_gsp *gsp, u32 timeout_ms,
|
||||
int nvgpu_gsp_wait_message_cond(struct gk20a *g, u32 timeout_ms,
|
||||
void *var, u8 val)
|
||||
{
|
||||
struct gk20a *g = gsp->g;
|
||||
struct nvgpu_timeout timeout;
|
||||
u32 delay = POLL_DELAY_MIN_US;
|
||||
|
||||
@@ -224,8 +224,6 @@ int nvgpu_gsp_wait_message_cond(struct nvgpu_gsp *gsp, u32 timeout_ms,
|
||||
return 0;
|
||||
}
|
||||
|
||||
g->ops.gsp.gsp_isr(g);
|
||||
|
||||
nvgpu_usleep_range(delay, delay * 2U);
|
||||
delay = min_t(u32, delay << 1U, POLL_DELAY_MAX_US);
|
||||
} while (nvgpu_timeout_expired(&timeout) == 0);
|
||||
@@ -1,5 +1,5 @@
|
||||
/*
|
||||
* Copyright (c) 2021, NVIDIA CORPORATION. All rights reserved.
|
||||
* Copyright (c) 2021-2022, NVIDIA CORPORATION. All rights reserved.
|
||||
*
|
||||
* Permission is hereby granted, free of charge, to any person obtaining a
|
||||
* copy of this software and associated documentation files (the "Software"),
|
||||
@@ -27,6 +27,7 @@
|
||||
#include <nvgpu/gsp.h>
|
||||
|
||||
#include "gsp_cmd.h"
|
||||
#include "../gsp_scheduler.h"
|
||||
|
||||
struct nvgpu_gsp;
|
||||
|
||||
@@ -68,7 +69,7 @@ struct nv_flcn_msg_gsp {
|
||||
} msg;
|
||||
};
|
||||
|
||||
int nvgpu_gsp_wait_message_cond(struct nvgpu_gsp *gsp, u32 timeout_ms,
|
||||
int nvgpu_gsp_wait_message_cond(struct gk20a *g, u32 timeout_ms,
|
||||
void *var, u8 val);
|
||||
|
||||
#endif /* NVGPU_GSP_MSG_H */
|
||||
@@ -1,5 +1,5 @@
|
||||
/*
|
||||
* Copyright (c) 2021, NVIDIA CORPORATION. All rights reserved.
|
||||
* Copyright (c) 2021-2022, NVIDIA CORPORATION. All rights reserved.
|
||||
*
|
||||
* Permission is hereby granted, free of charge, to any person obtaining a
|
||||
* copy of this software and associated documentation files (the "Software"),
|
||||
@@ -26,7 +26,7 @@
|
||||
#include <nvgpu/log.h>
|
||||
#include <nvgpu/gsp.h>
|
||||
|
||||
#include "../gsp_priv.h"
|
||||
#include "../gsp_scheduler.h"
|
||||
#include "gsp_queue.h"
|
||||
#include "gsp_msg.h"
|
||||
|
||||
@@ -1,5 +1,5 @@
|
||||
/*
|
||||
* Copyright (c) 2021, NVIDIA CORPORATION. All rights reserved.
|
||||
* Copyright (c) 2021-2022, NVIDIA CORPORATION. All rights reserved.
|
||||
*
|
||||
* Permission is hereby granted, free of charge, to any person obtaining a
|
||||
* copy of this software and associated documentation files (the "Software"),
|
||||
@@ -1,5 +1,5 @@
|
||||
/*
|
||||
* Copyright (c) 2021, NVIDIA CORPORATION. All rights reserved.
|
||||
* Copyright (c) 2021-2022, NVIDIA CORPORATION. All rights reserved.
|
||||
*
|
||||
* Permission is hereby granted, free of charge, to any person obtaining a
|
||||
* copy of this software and associated documentation files (the "Software"),
|
||||
@@ -26,7 +26,7 @@
|
||||
#include <nvgpu/errno.h>
|
||||
#include <nvgpu/gsp.h>
|
||||
|
||||
#include "../gsp_priv.h"
|
||||
#include "../gsp_scheduler.h"
|
||||
#include "gsp_seq.h"
|
||||
|
||||
static void gsp_sequences_init(struct gk20a *g,
|
||||
@@ -47,7 +47,7 @@ static void gsp_sequences_init(struct gk20a *g,
|
||||
}
|
||||
}
|
||||
|
||||
int nvgpu_gsp_sequences_init(struct gk20a *g, struct nvgpu_gsp *gsp)
|
||||
int nvgpu_gsp_sequences_init(struct gk20a *g, struct nvgpu_gsp_sched *gsp_sched)
|
||||
{
|
||||
int err = 0;
|
||||
struct gsp_sequences *seqs;
|
||||
@@ -68,8 +68,8 @@ int nvgpu_gsp_sequences_init(struct gk20a *g, struct nvgpu_gsp *gsp)
|
||||
return -ENOMEM;
|
||||
}
|
||||
|
||||
gsp->sequences = seqs;
|
||||
gsp->sequences->seq = seqs->seq;
|
||||
gsp_sched->sequences = seqs;
|
||||
gsp_sched->sequences->seq = seqs->seq;
|
||||
|
||||
nvgpu_mutex_init(&seqs->gsp_seq_lock);
|
||||
|
||||
@@ -1,5 +1,5 @@
|
||||
/*
|
||||
* Copyright (c) 2021, NVIDIA CORPORATION. All rights reserved.
|
||||
* Copyright (c) 2021-2022, NVIDIA CORPORATION. All rights reserved.
|
||||
*
|
||||
* Permission is hereby granted, free of charge, to any person obtaining a
|
||||
* copy of this software and associated documentation files (the "Software"),
|
||||
@@ -56,7 +56,7 @@ struct gsp_sequences {
|
||||
struct nvgpu_mutex gsp_seq_lock;
|
||||
};
|
||||
|
||||
int nvgpu_gsp_sequences_init(struct gk20a *g, struct nvgpu_gsp *gsp);
|
||||
int nvgpu_gsp_sequences_init(struct gk20a *g, struct nvgpu_gsp_sched *gsp_sched);
|
||||
void nvgpu_gsp_sequences_free(struct gk20a *g,
|
||||
struct gsp_sequences *sequences);
|
||||
int nvgpu_gsp_seq_acquire(struct gk20a *g,
|
||||
343
drivers/gpu/nvgpu/common/gsp_test/gsp_test.c
Normal file
343
drivers/gpu/nvgpu/common/gsp_test/gsp_test.c
Normal file
@@ -0,0 +1,343 @@
|
||||
// SPDX-License-Identifier: GPL-2.0-or-later
|
||||
/*
|
||||
* GSP Test Functions
|
||||
*
|
||||
* Copyright (c) 2021-2022, NVIDIA Corporation. All rights reserved.
|
||||
*
|
||||
* This program is free software; you can redistribute it and/or modify it
|
||||
* under the terms and conditions of the GNU General Public License,
|
||||
* version 2, as published by the Free Software Foundation.
|
||||
*
|
||||
* This program is distributed in the hope it will be useful, but WITHOUT
|
||||
* ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
|
||||
* FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for
|
||||
* more details.
|
||||
*
|
||||
* You should have received a copy of the GNU General Public License
|
||||
* along with this program. If not, see <http://www.gnu.org/licenses/>.
|
||||
*/
|
||||
#include <nvgpu/firmware.h>
|
||||
#include <nvgpu/falcon.h>
|
||||
#include <nvgpu/dma.h>
|
||||
#include <nvgpu/gk20a.h>
|
||||
#include <nvgpu/log.h>
|
||||
#include <nvgpu/gsp.h>
|
||||
#include <nvgpu/gsp/gsp_test.h>
|
||||
|
||||
#include "gsp_test.h"
|
||||
|
||||
u32 nvgpu_gsp_get_current_iteration(struct gk20a *g)
|
||||
{
|
||||
u32 data = 0;
|
||||
struct nvgpu_gsp *gsp = g->gsp_stest->gsp;
|
||||
|
||||
nvgpu_log_fn(g, " ");
|
||||
|
||||
data = nvgpu_falcon_mailbox_read(gsp->gsp_flcn, FALCON_MAILBOX_1);
|
||||
|
||||
return data;
|
||||
}
|
||||
|
||||
u32 nvgpu_gsp_get_current_test(struct gk20a *g)
|
||||
{
|
||||
u32 data = 0;
|
||||
struct nvgpu_gsp *gsp = g->gsp_stest->gsp;
|
||||
|
||||
nvgpu_log_fn(g, " ");
|
||||
|
||||
data = nvgpu_falcon_mailbox_read(gsp->gsp_flcn, FALCON_MAILBOX_0);
|
||||
|
||||
return data;
|
||||
}
|
||||
|
||||
bool nvgpu_gsp_get_test_fail_status(struct gk20a *g)
|
||||
{
|
||||
struct nvgpu_gsp_test *gsp_stest = g->gsp_stest;
|
||||
|
||||
return gsp_stest->gsp_test.stress_test_fail_status;
|
||||
}
|
||||
|
||||
bool nvgpu_gsp_get_stress_test_start(struct gk20a *g)
|
||||
{
|
||||
struct nvgpu_gsp_test *gsp_stest = g->gsp_stest;
|
||||
|
||||
return gsp_stest->gsp_test.enable_stress_test;
|
||||
}
|
||||
|
||||
bool nvgpu_gsp_get_stress_test_load(struct gk20a *g)
|
||||
{
|
||||
struct nvgpu_gsp_test *gsp_stest = g->gsp_stest;
|
||||
|
||||
if (gsp_stest == NULL)
|
||||
return false;
|
||||
|
||||
return gsp_stest->gsp_test.load_stress_test;
|
||||
}
|
||||
|
||||
void nvgpu_gsp_set_test_fail_status(struct gk20a *g, bool val)
|
||||
{
|
||||
struct nvgpu_gsp_test *gsp_stest = g->gsp_stest;
|
||||
|
||||
gsp_stest->gsp_test.stress_test_fail_status = val;
|
||||
}
|
||||
|
||||
int nvgpu_gsp_set_stress_test_start(struct gk20a *g, bool flag)
|
||||
{
|
||||
int err = 0;
|
||||
struct nvgpu_gsp *gsp = g->gsp_stest->gsp;
|
||||
struct nvgpu_gsp_test *gsp_stest = g->gsp_stest;
|
||||
|
||||
nvgpu_log_fn(g, " ");
|
||||
|
||||
if (flag) {
|
||||
nvgpu_info(g, "Enabling GSP test");
|
||||
nvgpu_falcon_mailbox_write(gsp->gsp_flcn, FALCON_MAILBOX_1, 0xFFFFFFFF);
|
||||
} else {
|
||||
nvgpu_info(g, "Halting GSP test");
|
||||
nvgpu_gsp_stress_test_halt(g, false);
|
||||
}
|
||||
|
||||
gsp_stest->gsp_test.enable_stress_test = flag;
|
||||
|
||||
return err;
|
||||
}
|
||||
|
||||
int nvgpu_gsp_set_stress_test_load(struct gk20a *g, bool flag)
|
||||
{
|
||||
int err = 0;
|
||||
|
||||
nvgpu_log_fn(g, " ");
|
||||
|
||||
if (flag)
|
||||
err = nvgpu_gsp_stress_test_bootstrap(g, flag);
|
||||
|
||||
return err;
|
||||
}
|
||||
|
||||
static void gsp_test_get_file_names(struct gk20a *g, struct gsp_fw *gsp_ucode)
|
||||
{
|
||||
/*
|
||||
* TODO Switch to GSP specific register
|
||||
*/
|
||||
if (g->ops.pmu.is_debug_mode_enabled(g)) {
|
||||
gsp_ucode->code_name = GSPDBG_RISCV_STRESS_TEST_FW_CODE;
|
||||
gsp_ucode->data_name = GSPDBG_RISCV_STRESS_TEST_FW_DATA;
|
||||
gsp_ucode->manifest_name = GSPDBG_RISCV_STRESS_TEST_FW_MANIFEST;
|
||||
} else {
|
||||
gsp_ucode->code_name = GSPPROD_RISCV_STRESS_TEST_FW_CODE;
|
||||
gsp_ucode->data_name = GSPPROD_RISCV_STRESS_TEST_FW_DATA;
|
||||
gsp_ucode->manifest_name = GSPPROD_RISCV_STRESS_TEST_FW_MANIFEST;
|
||||
}
|
||||
}
|
||||
|
||||
void nvgpu_gsp_write_test_sysmem_addr(struct gk20a *g)
|
||||
{
|
||||
struct nvgpu_gsp *gsp = g->gsp_stest->gsp;
|
||||
struct nvgpu_falcon *flcn;
|
||||
u64 sysmem_addr;
|
||||
struct nvgpu_gsp_test *gsp_stest;
|
||||
|
||||
flcn = gsp->gsp_flcn;
|
||||
gsp_stest = g->gsp_stest;
|
||||
|
||||
sysmem_addr = nvgpu_mem_get_addr(g, &gsp_stest->gsp_test.gsp_test_sysmem_block);
|
||||
|
||||
nvgpu_falcon_mailbox_write(flcn, FALCON_MAILBOX_0, u64_lo32(sysmem_addr));
|
||||
|
||||
nvgpu_falcon_mailbox_write(flcn, FALCON_MAILBOX_1, u64_hi32(sysmem_addr));
|
||||
}
|
||||
|
||||
int nvgpu_gsp_stress_test_bootstrap(struct gk20a *g, bool start)
|
||||
{
|
||||
int err = 0;
|
||||
struct nvgpu_gsp *gsp = g->gsp_stest->gsp;
|
||||
struct nvgpu_gsp_test *gsp_stest = g->gsp_stest;
|
||||
|
||||
nvgpu_log_fn(g, " ");
|
||||
|
||||
if (gsp_stest == NULL) {
|
||||
nvgpu_err(g, "GSP not initialized");
|
||||
err = -EFAULT;
|
||||
goto exit;
|
||||
}
|
||||
|
||||
if (!start && !(gsp_stest->gsp_test.load_stress_test))
|
||||
return err;
|
||||
|
||||
if (start) {
|
||||
err = nvgpu_dma_alloc_flags_sys(g,
|
||||
NVGPU_DMA_PHYSICALLY_ADDRESSED,
|
||||
SZ_64K,
|
||||
&g->gsp_stest->gsp_test.gsp_test_sysmem_block);
|
||||
if (err != 0) {
|
||||
nvgpu_err(g, "GSP test memory alloc failed");
|
||||
goto exit;
|
||||
}
|
||||
}
|
||||
|
||||
gsp_stest->gsp_test.load_stress_test = true;
|
||||
|
||||
#ifdef CONFIG_NVGPU_FALCON_DEBUG
|
||||
err = nvgpu_gsp_debug_buf_init(g, GSP_TEST_DEBUG_BUFFER_QUEUE,
|
||||
GSP_TEST_DMESG_BUFFER_SIZE);
|
||||
if (err != 0) {
|
||||
nvgpu_err(g, "GSP sched debug buf init failed");
|
||||
goto exit;
|
||||
}
|
||||
#endif
|
||||
|
||||
gsp_test_get_file_names(g, &gsp->gsp_ucode);
|
||||
|
||||
err = nvgpu_gsp_bootstrap_ns(g, gsp);
|
||||
if (err != 0) {
|
||||
nvgpu_err(g, "GSP bootstrap failed for stress test");
|
||||
goto exit;
|
||||
}
|
||||
|
||||
/* wait for mailbox-0 update with non-zero value */
|
||||
err = nvgpu_gsp_wait_for_mailbox_update(gsp, 0x0,
|
||||
GSP_STRESS_TEST_MAILBOX_PASS, GSP_WAIT_TIME_MS);
|
||||
if (err != 0) {
|
||||
nvgpu_err(g, "gsp ucode failed to update mailbox-0");
|
||||
}
|
||||
|
||||
if (gsp_stest->gsp_test.enable_stress_test) {
|
||||
nvgpu_info(g, "Restarting GSP stress test");
|
||||
nvgpu_falcon_mailbox_write(gsp->gsp_flcn, FALCON_MAILBOX_1, 0xFFFFFFFF);
|
||||
}
|
||||
|
||||
return err;
|
||||
|
||||
exit:
|
||||
gsp_stest->gsp_test.load_stress_test = false;
|
||||
return err;
|
||||
}
|
||||
|
||||
int nvgpu_gsp_stress_test_halt(struct gk20a *g, bool restart)
|
||||
{
|
||||
int err = 0;
|
||||
struct nvgpu_gsp *gsp = g->gsp_stest->gsp;
|
||||
struct nvgpu_gsp_test *gsp_stest = g->gsp_stest;
|
||||
|
||||
|
||||
nvgpu_log_fn(g, " ");
|
||||
|
||||
if (gsp == NULL) {
|
||||
nvgpu_info(g, "GSP not initialized");
|
||||
goto exit;
|
||||
}
|
||||
|
||||
nvgpu_gsp_suspend(g, gsp);
|
||||
|
||||
if (restart && (gsp_stest->gsp_test.load_stress_test == false)) {
|
||||
nvgpu_info(g, "GSP stress test not loaded ");
|
||||
goto exit;
|
||||
}
|
||||
|
||||
err = nvgpu_falcon_reset(gsp->gsp_flcn);
|
||||
if (err != 0) {
|
||||
nvgpu_err(g, "gsp reset failed err=%d", err);
|
||||
goto exit;
|
||||
}
|
||||
|
||||
if (!restart) {
|
||||
gsp_stest->gsp_test.load_stress_test = false;
|
||||
nvgpu_dma_free(g, &gsp_stest->gsp_test.gsp_test_sysmem_block);
|
||||
}
|
||||
|
||||
exit:
|
||||
return err;
|
||||
}
|
||||
|
||||
bool nvgpu_gsp_is_stress_test(struct gk20a *g)
|
||||
{
|
||||
if (g->gsp_stest->gsp_test.load_stress_test)
|
||||
return true;
|
||||
else
|
||||
return false;
|
||||
}
|
||||
|
||||
static void gsp_test_sw_deinit(struct gk20a *g, struct nvgpu_gsp_test *gsp_stest)
|
||||
{
|
||||
nvgpu_dma_free(g, &gsp_stest->gsp_test.gsp_test_sysmem_block);
|
||||
|
||||
nvgpu_kfree(g, gsp_stest);
|
||||
gsp_stest = NULL;
|
||||
}
|
||||
|
||||
void nvgpu_gsp_test_sw_deinit(struct gk20a *g)
|
||||
{
|
||||
struct nvgpu_gsp_test *gsp_stest = g->gsp_stest;
|
||||
nvgpu_log_fn(g, " ");
|
||||
|
||||
if (gsp_stest == NULL) {
|
||||
nvgpu_info(g, "GSP stest not initialized");
|
||||
return;
|
||||
}
|
||||
|
||||
if (gsp_stest->gsp != NULL) {
|
||||
nvgpu_gsp_sw_deinit(g, gsp_stest->gsp);
|
||||
}
|
||||
|
||||
if (gsp_stest != NULL) {
|
||||
gsp_test_sw_deinit(g, gsp_stest);
|
||||
}
|
||||
}
|
||||
|
||||
int nvgpu_gsp_stress_test_sw_init(struct gk20a *g)
|
||||
{
|
||||
int err = 0;
|
||||
struct nvgpu_gsp_test *gsp_stest;
|
||||
struct nvgpu_gsp *gsp;
|
||||
|
||||
nvgpu_log_fn(g, " ");
|
||||
|
||||
if (g->gsp_stest != NULL) {
|
||||
/*
|
||||
* Recovery/unrailgate case, we do not need to do gsp_stest init as
|
||||
* gsp_stest is set during cold boot & doesn't execute gsp_stest clean
|
||||
* up as part of power off sequence, so reuse to perform faster boot.
|
||||
*/
|
||||
nvgpu_gsp_stress_test_bootstrap(g, false);
|
||||
return err;
|
||||
}
|
||||
|
||||
/* Init struct holding the gsp_stest software state */
|
||||
g->gsp_stest = (struct nvgpu_gsp_test *)
|
||||
nvgpu_kzalloc(g, sizeof(struct nvgpu_gsp_test));
|
||||
if (g->gsp_stest == NULL) {
|
||||
err = -ENOMEM;
|
||||
goto de_init;
|
||||
}
|
||||
|
||||
/* Init struct holding the gsp software state */
|
||||
g->gsp_stest->gsp = (struct nvgpu_gsp *)
|
||||
nvgpu_kzalloc(g, sizeof(struct nvgpu_gsp));
|
||||
if (g->gsp_stest->gsp == NULL) {
|
||||
err = -ENOMEM;
|
||||
goto de_init;
|
||||
}
|
||||
|
||||
gsp_stest = g->gsp_stest;
|
||||
gsp = g->gsp_stest->gsp;
|
||||
|
||||
/* gsp falcon software state */
|
||||
gsp->gsp_flcn = &g->gsp_flcn;
|
||||
gsp->g = g;
|
||||
|
||||
/* Init isr mutex */
|
||||
nvgpu_mutex_init(&gsp->isr_mutex);
|
||||
|
||||
nvgpu_log_fn(g, " Done ");
|
||||
return err;
|
||||
de_init:
|
||||
nvgpu_gsp_test_sw_deinit(g);
|
||||
return err;
|
||||
}
|
||||
|
||||
void nvgpu_gsp_stest_isr(struct gk20a *g)
|
||||
{
|
||||
struct nvgpu_gsp *gsp = g->gsp_stest->gsp;
|
||||
|
||||
g->ops.gsp.gsp_isr(g, gsp);
|
||||
}
|
||||
@@ -1,5 +1,5 @@
|
||||
/*
|
||||
* Copyright (c) 2021, NVIDIA CORPORATION. All rights reserved.
|
||||
* Copyright (c) 2022, NVIDIA CORPORATION. All rights reserved.
|
||||
*
|
||||
* Permission is hereby granted, free of charge, to any person obtaining a
|
||||
* copy of this software and associated documentation files (the "Software"),
|
||||
@@ -20,25 +20,22 @@
|
||||
* DEALINGS IN THE SOFTWARE.
|
||||
*/
|
||||
|
||||
#ifndef NVGPU_GSP_PRIV
|
||||
#define NVGPU_GSP_PRIV
|
||||
#ifndef GSP_TEST_H
|
||||
#define GSP_TEST_H
|
||||
|
||||
#include <nvgpu/lock.h>
|
||||
#include <nvgpu/nvgpu_mem.h>
|
||||
#define GSP_TEST_DEBUG_BUFFER_QUEUE 3U
|
||||
#define GSP_TEST_DMESG_BUFFER_SIZE 0xC00U
|
||||
|
||||
#define GSP_DEBUG_BUFFER_QUEUE 3U
|
||||
#define GSP_DMESG_BUFFER_SIZE 0xC00U
|
||||
#define GSPDBG_RISCV_STRESS_TEST_FW_MANIFEST "gsp-stress.manifest.encrypt.bin.out.bin"
|
||||
#define GSPDBG_RISCV_STRESS_TEST_FW_CODE "gsp-stress.text.encrypt.bin"
|
||||
#define GSPDBG_RISCV_STRESS_TEST_FW_DATA "gsp-stress.data.encrypt.bin"
|
||||
|
||||
#define GSP_QUEUE_NUM 2U
|
||||
#define GSPPROD_RISCV_STRESS_TEST_FW_MANIFEST "gsp-stress.manifest.encrypt.bin.out.bin.prod"
|
||||
#define GSPPROD_RISCV_STRESS_TEST_FW_CODE "gsp-stress.text.encrypt.bin.prod"
|
||||
#define GSPPROD_RISCV_STRESS_TEST_FW_DATA "gsp-stress.data.encrypt.bin.prod"
|
||||
|
||||
struct gsp_fw {
|
||||
/* gsp ucode */
|
||||
struct nvgpu_firmware *code;
|
||||
struct nvgpu_firmware *data;
|
||||
struct nvgpu_firmware *manifest;
|
||||
};
|
||||
#define GSP_STRESS_TEST_MAILBOX_PASS 0xAAAAAAAA
|
||||
|
||||
#ifdef CONFIG_NVGPU_GSP_STRESS_TEST
|
||||
struct gsp_stress_test {
|
||||
bool load_stress_test;
|
||||
bool enable_stress_test;
|
||||
@@ -47,28 +44,11 @@ struct gsp_stress_test {
|
||||
u32 test_name;
|
||||
struct nvgpu_mem gsp_test_sysmem_block;
|
||||
};
|
||||
#endif
|
||||
|
||||
/* GSP descriptor's */
|
||||
struct nvgpu_gsp {
|
||||
struct gk20a *g;
|
||||
|
||||
struct gsp_fw gsp_ucode;
|
||||
struct nvgpu_falcon *gsp_flcn;
|
||||
|
||||
bool isr_enabled;
|
||||
struct nvgpu_mutex isr_mutex;
|
||||
|
||||
struct gsp_sequences *sequences;
|
||||
|
||||
struct nvgpu_engine_mem_queue *queues[GSP_QUEUE_NUM];
|
||||
|
||||
u32 command_ack;
|
||||
|
||||
/* set to true once init received */
|
||||
bool gsp_ready;
|
||||
|
||||
#ifdef CONFIG_NVGPU_GSP_STRESS_TEST
|
||||
struct nvgpu_gsp_test {
|
||||
struct nvgpu_gsp *gsp;
|
||||
struct gsp_stress_test gsp_test;
|
||||
#endif
|
||||
};
|
||||
#endif /* NVGPU_GSP_PRIV */
|
||||
|
||||
#endif /* GSP_TEST_H */
|
||||
@@ -44,6 +44,10 @@
|
||||
#include <nvgpu/power_features/cg.h>
|
||||
#ifdef CONFIG_NVGPU_GSP_SCHEDULER
|
||||
#include <nvgpu/gsp.h>
|
||||
#include <nvgpu/gsp_sched.h>
|
||||
#endif
|
||||
#ifdef CONFIG_NVGPU_GSP_STRESS_TEST
|
||||
#include <nvgpu/gsp/gsp_test.h>
|
||||
#endif
|
||||
#include <nvgpu/pm_reservation.h>
|
||||
#include <nvgpu/netlist.h>
|
||||
@@ -346,6 +350,7 @@ int nvgpu_prepare_poweroff(struct gk20a *g)
|
||||
ret = tmp_ret;
|
||||
}
|
||||
|
||||
#ifndef CONFIG_NVGPU_DGPU
|
||||
#ifdef CONFIG_NVGPU_GSP_STRESS_TEST
|
||||
tmp_ret = nvgpu_gsp_stress_test_halt(g, true);
|
||||
if (tmp_ret != 0) {
|
||||
@@ -353,8 +358,9 @@ int nvgpu_prepare_poweroff(struct gk20a *g)
|
||||
nvgpu_err(g, "Failed to halt GSP stress test");
|
||||
}
|
||||
#endif
|
||||
#if defined(CONFIG_NVGPU_GSP_SCHEDULER)
|
||||
nvgpu_gsp_suspend(g);
|
||||
#ifdef CONFIG_NVGPU_GSP_SCHEDULER
|
||||
nvgpu_gsp_sched_suspend(g, g->gsp_sched);
|
||||
#endif
|
||||
#endif
|
||||
|
||||
nvgpu_falcons_sw_free(g);
|
||||
@@ -948,9 +954,14 @@ int nvgpu_finalize_poweron(struct gk20a *g)
|
||||
#endif
|
||||
NVGPU_INIT_TABLE_ENTRY(g->ops.channel.resume_all_serviceable_ch,
|
||||
NO_FLAG),
|
||||
#if defined(CONFIG_NVGPU_GSP_SCHEDULER) || defined(CONFIG_NVGPU_GSP_STRESS_TEST)
|
||||
#ifndef CONFIG_NVGPU_DGPU
|
||||
#ifdef CONFIG_NVGPU_GSP_SCHEDULER
|
||||
/* Init gsp ops */
|
||||
NVGPU_INIT_TABLE_ENTRY(&nvgpu_gsp_sw_init, NO_FLAG),
|
||||
NVGPU_INIT_TABLE_ENTRY(&nvgpu_gsp_sched_sw_init, NO_FLAG),
|
||||
#endif
|
||||
#ifdef CONFIG_NVGPU_GSP_STRESS_TEST
|
||||
NVGPU_INIT_TABLE_ENTRY(&nvgpu_gsp_stress_test_sw_init, NO_FLAG),
|
||||
#endif
|
||||
#endif
|
||||
};
|
||||
size_t i;
|
||||
|
||||
@@ -1,5 +1,5 @@
|
||||
/*
|
||||
* Copyright (c) 2021, NVIDIA CORPORATION. All rights reserved.
|
||||
* Copyright (c) 2021-2022, NVIDIA CORPORATION. All rights reserved.
|
||||
*
|
||||
* Permission is hereby granted, free of charge, to any person obtaining a
|
||||
* copy of this software and associated documentation files (the "Software"),
|
||||
@@ -177,7 +177,7 @@ void ga10b_gsp_handle_interrupts(struct gk20a *g, u32 intr)
|
||||
}
|
||||
}
|
||||
|
||||
void ga10b_gsp_isr(struct gk20a *g)
|
||||
void ga10b_gsp_isr(struct gk20a *g, struct nvgpu_gsp *gsp)
|
||||
{
|
||||
u32 intr = 0U;
|
||||
u32 mask = 0U;
|
||||
@@ -189,8 +189,8 @@ void ga10b_gsp_isr(struct gk20a *g)
|
||||
return;
|
||||
}
|
||||
|
||||
nvgpu_gsp_isr_mutex_aquire(g);
|
||||
if (!nvgpu_gsp_is_isr_enable(g)) {
|
||||
nvgpu_gsp_isr_mutex_acquire(g, gsp);
|
||||
if (!nvgpu_gsp_is_isr_enable(g, gsp)) {
|
||||
goto exit;
|
||||
}
|
||||
|
||||
@@ -212,69 +212,21 @@ void ga10b_gsp_isr(struct gk20a *g)
|
||||
ga10b_gsp_handle_interrupts(g, intr);
|
||||
|
||||
exit:
|
||||
nvgpu_gsp_isr_mutex_release(g);
|
||||
}
|
||||
|
||||
static void ga10b_riscv_set_irq(struct gk20a *g, bool set_irq,
|
||||
u32 intr_mask, u32 intr_dest)
|
||||
{
|
||||
if (set_irq) {
|
||||
gk20a_writel(g, pgsp_riscv_irqmset_r(), intr_mask);
|
||||
gk20a_writel(g, pgsp_riscv_irqdest_r(), intr_dest);
|
||||
} else {
|
||||
gk20a_writel(g, pgsp_riscv_irqmclr_r(), 0xffffffffU);
|
||||
}
|
||||
nvgpu_gsp_isr_mutex_release(g, gsp);
|
||||
}
|
||||
|
||||
void ga10b_gsp_enable_irq(struct gk20a *g, bool enable)
|
||||
{
|
||||
u32 intr_mask;
|
||||
u32 intr_dest;
|
||||
bool skip_priv = false;
|
||||
|
||||
nvgpu_log_fn(g, " ");
|
||||
|
||||
#ifdef CONFIG_NVGPU_GSP_STRESS_TEST
|
||||
if (nvgpu_gsp_is_stress_test(g))
|
||||
skip_priv = true;
|
||||
#endif
|
||||
|
||||
/* clear before setting required irq */
|
||||
if ((!skip_priv) || (!enable))
|
||||
ga10b_riscv_set_irq(g, false, 0x0, 0x0);
|
||||
|
||||
nvgpu_cic_mon_intr_stall_unit_config(g,
|
||||
NVGPU_CIC_INTR_UNIT_GSP, NVGPU_CIC_INTR_DISABLE);
|
||||
|
||||
if (enable) {
|
||||
if (!skip_priv) {
|
||||
/* dest 0=falcon, 1=host; level 0=irq0, 1=irq1 */
|
||||
intr_dest = pgsp_riscv_irqdest_gptmr_f(0) |
|
||||
pgsp_riscv_irqdest_wdtmr_f(1) |
|
||||
pgsp_riscv_irqdest_mthd_f(0) |
|
||||
pgsp_riscv_irqdest_ctxsw_f(0) |
|
||||
pgsp_riscv_irqdest_halt_f(1) |
|
||||
pgsp_riscv_irqdest_exterr_f(0) |
|
||||
pgsp_riscv_irqdest_swgen0_f(1) |
|
||||
pgsp_riscv_irqdest_swgen1_f(1) |
|
||||
pgsp_riscv_irqdest_ext_f(0xff);
|
||||
|
||||
/* 0=disable, 1=enable */
|
||||
intr_mask = pgsp_riscv_irqmset_gptmr_f(1) |
|
||||
pgsp_riscv_irqmset_wdtmr_f(1) |
|
||||
pgsp_riscv_irqmset_mthd_f(0) |
|
||||
pgsp_riscv_irqmset_ctxsw_f(0) |
|
||||
pgsp_riscv_irqmset_halt_f(1) |
|
||||
pgsp_riscv_irqmset_exterr_f(1) |
|
||||
pgsp_riscv_irqmset_swgen0_f(1) |
|
||||
pgsp_riscv_irqmset_swgen1_f(1);
|
||||
|
||||
/* set required irq */
|
||||
ga10b_riscv_set_irq(g, true, intr_mask, intr_dest);
|
||||
}
|
||||
|
||||
nvgpu_cic_mon_intr_stall_unit_config(g,
|
||||
NVGPU_CIC_INTR_UNIT_GSP, NVGPU_CIC_INTR_ENABLE);
|
||||
|
||||
/* Configuring RISCV interrupts is expected to be done inside firmware */
|
||||
}
|
||||
}
|
||||
|
||||
|
||||
@@ -1,5 +1,5 @@
|
||||
/*
|
||||
* Copyright (c) 2021, NVIDIA CORPORATION. All rights reserved.
|
||||
* Copyright (c) 2021-2022, NVIDIA CORPORATION. All rights reserved.
|
||||
*
|
||||
* Permission is hereby granted, free of charge, to any person obtaining a
|
||||
* copy of this software and associated documentation files (the "Software"),
|
||||
@@ -47,7 +47,7 @@ int ga10b_gsp_flcn_copy_from_emem(struct gk20a *g,
|
||||
|
||||
/* interrupt */
|
||||
void ga10b_gsp_enable_irq(struct gk20a *g, bool enable);
|
||||
void ga10b_gsp_isr(struct gk20a *g);
|
||||
void ga10b_gsp_isr(struct gk20a *g, struct nvgpu_gsp *gsp);
|
||||
void ga10b_gsp_set_msg_intr(struct gk20a *g);
|
||||
#endif /* CONFIG_NVGPU_GSP_SCHEDULER */
|
||||
#endif /* GSP_GA10B_H */
|
||||
|
||||
@@ -30,6 +30,9 @@
|
||||
#include <nvgpu/cic_mon.h>
|
||||
#include <nvgpu/power_features/pg.h>
|
||||
#include <nvgpu/gr/gr_instances.h>
|
||||
#ifdef CONFIG_NVGPU_GSP_SCHEDULER
|
||||
#include <nvgpu/gsp.h>
|
||||
#endif
|
||||
|
||||
#include "mc_intr_ga10b.h"
|
||||
|
||||
@@ -703,7 +706,7 @@ static void ga10b_intr_isr_stall_host2soc_2(struct gk20a *g)
|
||||
&unit_subtree_mask) == true) {
|
||||
handled_subtree_mask |= unit_subtree_mask;
|
||||
ga10b_intr_subtree_clear(g, subtree, unit_subtree_mask);
|
||||
g->ops.gsp.gsp_isr(g);
|
||||
nvgpu_gsp_isr(g);
|
||||
}
|
||||
#endif /* CONFIG_NVGPU_GSP_SCHEDULER */
|
||||
|
||||
|
||||
@@ -110,7 +110,10 @@ struct clk_domains_mon_status_params;
|
||||
struct nvgpu_cic_mon;
|
||||
struct nvgpu_cic_rm;
|
||||
#ifdef CONFIG_NVGPU_GSP_SCHEDULER
|
||||
struct nvgpu_gsp;
|
||||
struct nvgpu_gsp_sched;
|
||||
#endif
|
||||
#ifdef CONFIG_NVGPU_GSP_STRESS_TEST
|
||||
struct nvgpu_gsp_test;
|
||||
#endif
|
||||
|
||||
enum nvgpu_flush_op;
|
||||
@@ -496,7 +499,10 @@ struct gk20a {
|
||||
struct nvgpu_acr *acr;
|
||||
#ifdef CONFIG_NVGPU_GSP_SCHEDULER
|
||||
/** Pointer to struct maintaining GSP unit's software state. */
|
||||
struct nvgpu_gsp *gsp;
|
||||
struct nvgpu_gsp_sched *gsp_sched;
|
||||
#endif
|
||||
#ifdef CONFIG_NVGPU_GSP_STRESS_TEST
|
||||
struct nvgpu_gsp_test *gsp_stest;
|
||||
#endif
|
||||
/** Top level struct maintaining ECC unit's software state. */
|
||||
struct nvgpu_ecc ecc;
|
||||
|
||||
@@ -1,5 +1,5 @@
|
||||
/*
|
||||
* Copyright (c) 2020-2021, NVIDIA CORPORATION. All rights reserved.
|
||||
* Copyright (c) 2020-2022, NVIDIA CORPORATION. All rights reserved.
|
||||
*
|
||||
* Permission is hereby granted, free of charge, to any person obtaining a
|
||||
* copy of this software and associated documentation files (the "Software"),
|
||||
@@ -27,6 +27,8 @@
|
||||
struct gk20a;
|
||||
struct nvgpu_gsp;
|
||||
|
||||
#define GSP_WAIT_TIME_MS 10000U
|
||||
|
||||
struct gops_gsp {
|
||||
u32 (*falcon_base_addr)(void);
|
||||
u32 (*falcon2_base_addr)(void);
|
||||
@@ -50,7 +52,7 @@ struct gops_gsp {
|
||||
void (*msgq_tail)(struct gk20a *g, struct nvgpu_gsp *gsp,
|
||||
u32 *tail, bool set);
|
||||
void (*enable_irq)(struct gk20a *g, bool enable);
|
||||
void (*gsp_isr)(struct gk20a *g);
|
||||
void (*gsp_isr)(struct gk20a *g, struct nvgpu_gsp *gsp);
|
||||
void (*set_msg_intr)(struct gk20a *g);
|
||||
#endif /* CONFIG_NVGPU_GSP_SCHEDULER */
|
||||
};
|
||||
|
||||
@@ -1,5 +1,5 @@
|
||||
/*
|
||||
* Copyright (c) 2021, NVIDIA CORPORATION. All rights reserved.
|
||||
* Copyright (c) 2021-2022, NVIDIA CORPORATION. All rights reserved.
|
||||
*
|
||||
* Permission is hereby granted, free of charge, to any person obtaining a
|
||||
* copy of this software and associated documentation files (the "Software"),
|
||||
@@ -22,23 +22,45 @@
|
||||
|
||||
#ifndef NVGPU_GSP
|
||||
#define NVGPU_GSP
|
||||
#include <nvgpu/lock.h>
|
||||
#include <nvgpu/nvgpu_mem.h>
|
||||
struct gk20a;
|
||||
struct nvgpu_gsp;
|
||||
|
||||
int nvgpu_gsp_sw_init(struct gk20a *g);
|
||||
int nvgpu_gsp_bootstrap(struct gk20a *g);
|
||||
void nvgpu_gsp_suspend(struct gk20a *g);
|
||||
void nvgpu_gsp_sw_deinit(struct gk20a *g);
|
||||
void nvgpu_gsp_isr_support(struct gk20a *g, bool enable);
|
||||
void nvgpu_gsp_isr_mutex_aquire(struct gk20a *g);
|
||||
void nvgpu_gsp_isr_mutex_release(struct gk20a *g);
|
||||
bool nvgpu_gsp_is_isr_enable(struct gk20a *g);
|
||||
struct gsp_fw {
|
||||
/* gsp ucode name */
|
||||
char *code_name;
|
||||
char *data_name;
|
||||
char *manifest_name;
|
||||
|
||||
/* gsp ucode */
|
||||
struct nvgpu_firmware *code;
|
||||
struct nvgpu_firmware *data;
|
||||
struct nvgpu_firmware *manifest;
|
||||
};
|
||||
|
||||
/* GSP descriptor's */
|
||||
struct nvgpu_gsp {
|
||||
struct gk20a *g;
|
||||
|
||||
struct gsp_fw gsp_ucode;
|
||||
struct nvgpu_falcon *gsp_flcn;
|
||||
|
||||
bool isr_enabled;
|
||||
struct nvgpu_mutex isr_mutex;
|
||||
};
|
||||
|
||||
int nvgpu_gsp_debug_buf_init(struct gk20a *g, u32 queue_no, u32 buffer_size);
|
||||
void nvgpu_gsp_suspend(struct gk20a *g, struct nvgpu_gsp *gsp);
|
||||
void nvgpu_gsp_sw_deinit(struct gk20a *g, struct nvgpu_gsp *gsp);
|
||||
void nvgpu_gsp_isr_mutex_acquire(struct gk20a *g, struct nvgpu_gsp *gsp);
|
||||
void nvgpu_gsp_isr_mutex_release(struct gk20a *g, struct nvgpu_gsp *gsp);
|
||||
bool nvgpu_gsp_is_isr_enable(struct gk20a *g, struct nvgpu_gsp *gsp);
|
||||
u32 nvgpu_gsp_get_last_cmd_id(struct gk20a *g);
|
||||
struct nvgpu_falcon *nvgpu_gsp_falcon_instance(struct gk20a *g);
|
||||
int nvgpu_gsp_process_message(struct gk20a *g);
|
||||
#ifdef CONFIG_NVGPU_GSP_STRESS_TEST
|
||||
int nvgpu_gsp_stress_test_bootstrap(struct gk20a *g, bool start);
|
||||
int nvgpu_gsp_stress_test_halt(struct gk20a *g, bool restart);
|
||||
bool nvgpu_gsp_is_stress_test(struct gk20a *g);
|
||||
#endif
|
||||
int nvgpu_gsp_wait_for_mailbox_update(struct nvgpu_gsp *gsp,
|
||||
u32 mailbox_index, u32 exp_value, signed int timeoutms);
|
||||
int nvgpu_gsp_bootstrap_ns(struct gk20a *g, struct nvgpu_gsp *gsp);
|
||||
void nvgpu_gsp_isr(struct gk20a *g);
|
||||
void nvgpu_gsp_isr_support(struct gk20a *g, struct nvgpu_gsp *gsp, bool enable);
|
||||
#endif /* NVGPU_GSP */
|
||||
|
||||
@@ -2,7 +2,7 @@
|
||||
/*
|
||||
* GSP Tests
|
||||
*
|
||||
* Copyright (c) 2021, NVIDIA Corporation. All rights reserved.
|
||||
* Copyright (c) 2021-2022, NVIDIA Corporation. All rights reserved.
|
||||
*
|
||||
* This program is free software; you can redistribute it and/or modify it
|
||||
* under the terms and conditions of the GNU General Public License,
|
||||
@@ -30,4 +30,11 @@ bool nvgpu_gsp_get_stress_test_start(struct gk20a *g);
|
||||
int nvgpu_gsp_set_stress_test_start(struct gk20a *g, bool flag);
|
||||
bool nvgpu_gsp_get_stress_test_load(struct gk20a *g);
|
||||
int nvgpu_gsp_set_stress_test_load(struct gk20a *g, bool flag);
|
||||
int nvgpu_gsp_stress_test_bootstrap(struct gk20a *g, bool start);
|
||||
int nvgpu_gsp_stress_test_halt(struct gk20a *g, bool restart);
|
||||
bool nvgpu_gsp_is_stress_test(struct gk20a *g);
|
||||
int nvgpu_gsp_stress_test_sw_init(struct gk20a *g);
|
||||
void nvgpu_gsp_test_sw_deinit(struct gk20a *g);
|
||||
void nvgpu_gsp_write_test_sysmem_addr(struct gk20a *g);
|
||||
void nvgpu_gsp_stest_isr(struct gk20a *g);
|
||||
#endif /* NVGPU_GSP_TEST */
|
||||
|
||||
@@ -1,5 +1,5 @@
|
||||
/*
|
||||
* Copyright (c) 2021, NVIDIA CORPORATION. All rights reserved.
|
||||
* Copyright (c) 2022, NVIDIA CORPORATION. All rights reserved.
|
||||
*
|
||||
* Permission is hereby granted, free of charge, to any person obtaining a
|
||||
* copy of this software and associated documentation files (the "Software"),
|
||||
@@ -20,12 +20,14 @@
|
||||
* DEALINGS IN THE SOFTWARE.
|
||||
*/
|
||||
|
||||
#ifndef NVGPU_GSP_BOOTSTRAP
|
||||
#define NVGPU_GSP_BOOTSTRAP
|
||||
struct nvgpu_gsp;
|
||||
#ifndef GSP_SCHED_H
|
||||
#define GSP_SCHED_H
|
||||
struct gk20a;
|
||||
struct nvgpu_gsp_sched;
|
||||
|
||||
#define GSP_UCODE_SIZE_MAX (256U * 1024U)
|
||||
|
||||
int gsp_bootstrap_ns(struct gk20a *g, struct nvgpu_gsp *gsp);
|
||||
|
||||
#endif /* NVGPU_GSP_BOOTSTRAP */
|
||||
int nvgpu_gsp_sched_bootstrap_ns(struct gk20a *g);
|
||||
int nvgpu_gsp_sched_sw_init(struct gk20a *g);
|
||||
void nvgpu_gsp_sched_sw_deinit(struct gk20a *g);
|
||||
void nvgpu_gsp_sched_suspend(struct gk20a *g, struct nvgpu_gsp_sched *gsp_sched);
|
||||
void nvgpu_gsp_sched_isr(struct gk20a *g);
|
||||
#endif /* GSP_SCHED_H */
|
||||
@@ -93,6 +93,10 @@
|
||||
|
||||
#ifdef CONFIG_NVGPU_GSP_SCHEDULER
|
||||
#include "nvgpu/gsp.h"
|
||||
#include "nvgpu/gsp_sched.h"
|
||||
#endif
|
||||
#ifdef CONFIG_NVGPU_GSP_STRESS_TEST
|
||||
#include "nvgpu/gsp/gsp_test.h"
|
||||
#endif
|
||||
|
||||
#ifdef CONFIG_NVGPU_SUPPORT_CDE
|
||||
@@ -1063,8 +1067,13 @@ void gk20a_remove_support(struct gk20a *g)
|
||||
nvgpu_free_cyclestats_snapshot_data(g);
|
||||
#endif
|
||||
|
||||
#ifndef CONFIG_NVGPU_DGPU
|
||||
#ifdef CONFIG_NVGPU_GSP_SCHEDULER
|
||||
nvgpu_gsp_sw_deinit(g);
|
||||
nvgpu_gsp_sched_sw_deinit(g);
|
||||
#endif
|
||||
#ifdef CONFIG_NVGPU_GSP_STRESS_TEST
|
||||
nvgpu_gsp_test_sw_deinit(g);
|
||||
#endif
|
||||
#endif
|
||||
|
||||
nvgpu_fbp_remove_support(g);
|
||||
|
||||
Reference in New Issue
Block a user