mirror of
git://nv-tegra.nvidia.com/linux-nvgpu.git
synced 2025-12-22 17:36:20 +03:00
- DeviceGetTpcExceptionEnMask test calls ioctl NVGPU_GPU_IOCTL_GET_TPC_EXCEPTION_EN_STATUS which reads register gr_gpc0_tpc0_tpccs_tpc_exception_en_r(). This causes IDLE_SNAP and further disengages ELPG. - Add elpg protected call for the tpc_enabled_exceptions HAL. Bug 3522086 Change-Id: I137ac2c643c693b596b6ce3e879da9c786ee3a85 Signed-off-by: Divya Singhatwaria <dsinghatwari@nvidia.com> Reviewed-on: https://git-master.nvidia.com/r/c/linux-nvgpu/+/2674509 Reviewed-by: svc-mobile-coverity <svc-mobile-coverity@nvidia.com> Reviewed-by: Antony Clince Alex <aalex@nvidia.com> Reviewed-by: Seema Khowala <seemaj@nvidia.com> GVS: Gerrit_Virtual_Submit
2771 lines
70 KiB
C
2771 lines
70 KiB
C
/*
|
|
* Copyright (c) 2011-2022, NVIDIA Corporation. All rights reserved.
|
|
*
|
|
* This program is free software; you can redistribute it and/or modify it
|
|
* under the terms and conditions of the GNU General Public License,
|
|
* version 2, as published by the Free Software Foundation.
|
|
*
|
|
* This program is distributed in the hope it will be useful, but WITHOUT
|
|
* ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
|
|
* FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for
|
|
* more details.
|
|
*
|
|
* You should have received a copy of the GNU General Public License
|
|
* along with this program. If not, see <http://www.gnu.org/licenses/>.
|
|
*/
|
|
|
|
#include <linux/uaccess.h>
|
|
#include <linux/cdev.h>
|
|
#include <linux/file.h>
|
|
#include <linux/anon_inodes.h>
|
|
#include <linux/fs.h>
|
|
#include <uapi/linux/nvgpu.h>
|
|
#include <nvgpu/pmu/clk/clk.h>
|
|
|
|
#include <nvgpu/bitops.h>
|
|
#include <nvgpu/comptags.h>
|
|
#include <nvgpu/kmem.h>
|
|
#include <nvgpu/nvhost.h>
|
|
#include <nvgpu/bug.h>
|
|
#include <nvgpu/ptimer.h>
|
|
#include <nvgpu/vidmem.h>
|
|
#include <nvgpu/log.h>
|
|
#include <nvgpu/enabled.h>
|
|
#include <nvgpu/sizes.h>
|
|
#include <nvgpu/list.h>
|
|
#include <nvgpu/fbp.h>
|
|
#include <nvgpu/clk_arb.h>
|
|
#include <nvgpu/gk20a.h>
|
|
#include <nvgpu/engines.h>
|
|
#include <nvgpu/device.h>
|
|
#include <nvgpu/gr/config.h>
|
|
#ifdef CONFIG_NVGPU_GRAPHICS
|
|
#include <nvgpu/gr/zbc.h>
|
|
#include <nvgpu/gr/zcull.h>
|
|
#endif
|
|
#include <nvgpu/gr/gr.h>
|
|
#include <nvgpu/gr/gr_utils.h>
|
|
#include <nvgpu/gr/gr_instances.h>
|
|
#include <nvgpu/gr/warpstate.h>
|
|
#include <nvgpu/channel.h>
|
|
#include <nvgpu/pmu/pmgr.h>
|
|
#include <nvgpu/pmu/therm.h>
|
|
#include <nvgpu/power_features/pg.h>
|
|
#include <nvgpu/fence.h>
|
|
#include <nvgpu/channel_sync_syncpt.h>
|
|
#include <nvgpu/soc.h>
|
|
#include <nvgpu/nvgpu_init.h>
|
|
#include <nvgpu/user_fence.h>
|
|
#include <nvgpu/nvgpu_init.h>
|
|
#include <nvgpu/grmgr.h>
|
|
#include <nvgpu/string.h>
|
|
|
|
#include "ioctl_ctrl.h"
|
|
#include "ioctl_dbg.h"
|
|
#include "ioctl_as.h"
|
|
#include "ioctl_tsg.h"
|
|
#include "ioctl_channel.h"
|
|
#include "ioctl.h"
|
|
|
|
#include "dmabuf_priv.h"
|
|
#include "platform_gk20a.h"
|
|
#include "os_linux.h"
|
|
#include "channel.h"
|
|
#include "dmabuf_vidmem.h"
|
|
#include "fecs_trace_linux.h"
|
|
|
|
#define HZ_TO_MHZ(a) ((a > 0xF414F9CD7ULL) ? 0xffff : (a >> 32) ? \
|
|
(u32) ((a * 0x10C8ULL) >> 32) : (u16) ((u32) a/MHZ))
|
|
#define MHZ_TO_HZ(a) ((u64)a * MHZ)
|
|
|
|
extern const struct file_operations gk20a_as_ops;
|
|
extern const struct file_operations gk20a_tsg_ops;
|
|
|
|
struct gk20a_ctrl_priv {
|
|
struct device *dev;
|
|
struct gk20a *g;
|
|
struct nvgpu_clk_session *clk_session;
|
|
struct nvgpu_cdev *cdev;
|
|
|
|
struct nvgpu_list_node list;
|
|
struct {
|
|
struct vm_area_struct *vma;
|
|
bool vma_mapped;
|
|
} usermode_vma;
|
|
};
|
|
|
|
static inline struct gk20a_ctrl_priv *
|
|
gk20a_ctrl_priv_from_list(struct nvgpu_list_node *node)
|
|
{
|
|
return (struct gk20a_ctrl_priv *)
|
|
((uintptr_t)node - offsetof(struct gk20a_ctrl_priv, list));
|
|
}
|
|
|
|
static u32 gk20a_as_translate_as_alloc_flags(struct gk20a *g, u32 flags)
|
|
{
|
|
u32 core_flags = 0;
|
|
|
|
if (flags & NVGPU_GPU_IOCTL_ALLOC_AS_FLAGS_USERSPACE_MANAGED)
|
|
core_flags |= NVGPU_AS_ALLOC_USERSPACE_MANAGED;
|
|
if (flags & NVGPU_GPU_IOCTL_ALLOC_AS_FLAGS_UNIFIED_VA)
|
|
core_flags |= NVGPU_AS_ALLOC_UNIFIED_VA;
|
|
|
|
return core_flags;
|
|
}
|
|
|
|
int gk20a_ctrl_dev_open(struct inode *inode, struct file *filp)
|
|
{
|
|
struct nvgpu_os_linux *l;
|
|
struct gk20a *g;
|
|
struct gk20a_ctrl_priv *priv;
|
|
int err = 0;
|
|
struct nvgpu_cdev *cdev;
|
|
|
|
cdev = container_of(inode->i_cdev, struct nvgpu_cdev, cdev);
|
|
g = nvgpu_get_gk20a_from_cdev(cdev);
|
|
|
|
g = nvgpu_get(g);
|
|
if (!g)
|
|
return -ENODEV;
|
|
|
|
l = nvgpu_os_linux_from_gk20a(g);
|
|
|
|
nvgpu_log_fn(g, " ");
|
|
|
|
priv = nvgpu_kzalloc(g, sizeof(struct gk20a_ctrl_priv));
|
|
if (!priv) {
|
|
err = -ENOMEM;
|
|
goto free_ref;
|
|
}
|
|
filp->private_data = priv;
|
|
priv->dev = dev_from_gk20a(g);
|
|
priv->cdev = cdev;
|
|
/*
|
|
* We dont close the arbiter fd's after driver teardown to support
|
|
* GPU_LOST events, so we store g here, instead of dereferencing the
|
|
* dev structure on teardown
|
|
*/
|
|
priv->g = g;
|
|
|
|
if (!g->sw_ready) {
|
|
err = gk20a_busy(g);
|
|
if (err)
|
|
goto free_ref;
|
|
gk20a_idle(g);
|
|
}
|
|
|
|
if (nvgpu_is_enabled(g, NVGPU_CLK_ARB_ENABLED)) {
|
|
err = nvgpu_clk_arb_init_session(g, &priv->clk_session);
|
|
}
|
|
free_ref:
|
|
if (err != 0) {
|
|
nvgpu_put(g);
|
|
if (priv)
|
|
nvgpu_kfree(g, priv);
|
|
} else {
|
|
nvgpu_mutex_acquire(&l->ctrl_privs_lock);
|
|
nvgpu_list_add(&priv->list, &l->ctrl_privs);
|
|
nvgpu_mutex_release(&l->ctrl_privs_lock);
|
|
}
|
|
|
|
return err;
|
|
}
|
|
int gk20a_ctrl_dev_release(struct inode *inode, struct file *filp)
|
|
{
|
|
struct gk20a_ctrl_priv *priv = filp->private_data;
|
|
struct gk20a *g = priv->g;
|
|
struct nvgpu_os_linux *l = nvgpu_os_linux_from_gk20a(g);
|
|
|
|
nvgpu_log_fn(g, " ");
|
|
|
|
nvgpu_mutex_acquire(&l->ctrl_privs_lock);
|
|
nvgpu_list_del(&priv->list);
|
|
nvgpu_mutex_release(&l->ctrl_privs_lock);
|
|
|
|
if (priv->clk_session)
|
|
nvgpu_clk_arb_release_session(g, priv->clk_session);
|
|
|
|
nvgpu_put(g);
|
|
nvgpu_kfree(g, priv);
|
|
|
|
return 0;
|
|
}
|
|
|
|
struct nvgpu_flags_mapping {
|
|
u64 ioctl_flag;
|
|
int enabled_flag;
|
|
};
|
|
|
|
static struct nvgpu_flags_mapping flags_mapping[] = {
|
|
{NVGPU_GPU_FLAGS_CAN_RAILGATE,
|
|
NVGPU_CAN_RAILGATE},
|
|
{NVGPU_GPU_FLAGS_HAS_SYNCPOINTS,
|
|
NVGPU_HAS_SYNCPOINTS},
|
|
{NVGPU_GPU_FLAGS_SUPPORT_PARTIAL_MAPPINGS,
|
|
NVGPU_SUPPORT_PARTIAL_MAPPINGS},
|
|
{NVGPU_GPU_FLAGS_SUPPORT_SPARSE_ALLOCS,
|
|
NVGPU_SUPPORT_SPARSE_ALLOCS},
|
|
{NVGPU_GPU_FLAGS_SUPPORT_SYNC_FENCE_FDS,
|
|
NVGPU_SUPPORT_SYNC_FENCE_FDS},
|
|
{NVGPU_GPU_FLAGS_SUPPORT_CYCLE_STATS,
|
|
NVGPU_SUPPORT_CYCLE_STATS},
|
|
{NVGPU_GPU_FLAGS_SUPPORT_CYCLE_STATS_SNAPSHOT,
|
|
NVGPU_SUPPORT_CYCLE_STATS_SNAPSHOT},
|
|
{NVGPU_GPU_FLAGS_SUPPORT_USERSPACE_MANAGED_AS,
|
|
NVGPU_SUPPORT_USERSPACE_MANAGED_AS},
|
|
{NVGPU_GPU_FLAGS_SUPPORT_TSG,
|
|
NVGPU_SUPPORT_TSG},
|
|
{NVGPU_GPU_FLAGS_SUPPORT_CLOCK_CONTROLS,
|
|
NVGPU_SUPPORT_CLOCK_CONTROLS},
|
|
{NVGPU_GPU_FLAGS_SUPPORT_GET_VOLTAGE,
|
|
NVGPU_SUPPORT_GET_VOLTAGE},
|
|
{NVGPU_GPU_FLAGS_SUPPORT_GET_CURRENT,
|
|
NVGPU_SUPPORT_GET_CURRENT},
|
|
{NVGPU_GPU_FLAGS_SUPPORT_GET_POWER,
|
|
NVGPU_SUPPORT_GET_POWER},
|
|
{NVGPU_GPU_FLAGS_SUPPORT_GET_TEMPERATURE,
|
|
NVGPU_SUPPORT_GET_TEMPERATURE},
|
|
{NVGPU_GPU_FLAGS_SUPPORT_SET_THERM_ALERT_LIMIT,
|
|
NVGPU_SUPPORT_SET_THERM_ALERT_LIMIT},
|
|
{NVGPU_GPU_FLAGS_SUPPORT_DEVICE_EVENTS,
|
|
NVGPU_SUPPORT_DEVICE_EVENTS},
|
|
{NVGPU_GPU_FLAGS_SUPPORT_FECS_CTXSW_TRACE,
|
|
NVGPU_SUPPORT_FECS_CTXSW_TRACE},
|
|
{NVGPU_GPU_FLAGS_SUPPORT_DETERMINISTIC_SUBMIT_NO_JOBTRACKING,
|
|
NVGPU_SUPPORT_DETERMINISTIC_SUBMIT_NO_JOBTRACKING},
|
|
{NVGPU_GPU_FLAGS_SUPPORT_DETERMINISTIC_SUBMIT_FULL,
|
|
NVGPU_SUPPORT_DETERMINISTIC_SUBMIT_FULL},
|
|
{NVGPU_GPU_FLAGS_SUPPORT_DETERMINISTIC_OPTS,
|
|
NVGPU_SUPPORT_DETERMINISTIC_OPTS},
|
|
{NVGPU_GPU_FLAGS_SUPPORT_SYNCPOINT_ADDRESS,
|
|
NVGPU_SUPPORT_SYNCPOINT_ADDRESS},
|
|
{NVGPU_GPU_FLAGS_SUPPORT_USER_SYNCPOINT,
|
|
NVGPU_SUPPORT_USER_SYNCPOINT},
|
|
{NVGPU_GPU_FLAGS_SUPPORT_USERMODE_SUBMIT,
|
|
NVGPU_SUPPORT_USERMODE_SUBMIT},
|
|
{NVGPU_GPU_FLAGS_SUPPORT_IO_COHERENCE,
|
|
NVGPU_SUPPORT_IO_COHERENCE},
|
|
{NVGPU_GPU_FLAGS_SUPPORT_RESCHEDULE_RUNLIST,
|
|
NVGPU_SUPPORT_RESCHEDULE_RUNLIST},
|
|
{NVGPU_GPU_FLAGS_SUPPORT_MAP_DIRECT_KIND_CTRL,
|
|
NVGPU_SUPPORT_MAP_DIRECT_KIND_CTRL},
|
|
{NVGPU_GPU_FLAGS_ECC_ENABLED_SM_LRF,
|
|
NVGPU_ECC_ENABLED_SM_LRF},
|
|
{NVGPU_GPU_FLAGS_ECC_ENABLED_SM_SHM,
|
|
NVGPU_ECC_ENABLED_SM_SHM},
|
|
{NVGPU_GPU_FLAGS_ECC_ENABLED_TEX,
|
|
NVGPU_ECC_ENABLED_TEX},
|
|
{NVGPU_GPU_FLAGS_ECC_ENABLED_LTC,
|
|
NVGPU_ECC_ENABLED_LTC},
|
|
{NVGPU_GPU_FLAGS_SUPPORT_TSG_SUBCONTEXTS,
|
|
NVGPU_SUPPORT_TSG_SUBCONTEXTS},
|
|
{NVGPU_GPU_FLAGS_SUPPORT_SCG,
|
|
NVGPU_SUPPORT_SCG},
|
|
{NVGPU_GPU_FLAGS_SUPPORT_VPR,
|
|
NVGPU_SUPPORT_VPR},
|
|
{NVGPU_GPU_FLAGS_DRIVER_REDUCED_PROFILE,
|
|
NVGPU_DRIVER_REDUCED_PROFILE},
|
|
{NVGPU_GPU_FLAGS_SUPPORT_SET_CTX_MMU_DEBUG_MODE,
|
|
NVGPU_SUPPORT_SET_CTX_MMU_DEBUG_MODE},
|
|
{NVGPU_GPU_FLAGS_SUPPORT_FAULT_RECOVERY,
|
|
NVGPU_SUPPORT_FAULT_RECOVERY},
|
|
{NVGPU_GPU_FLAGS_SUPPORT_MAPPING_MODIFY,
|
|
NVGPU_SUPPORT_MAPPING_MODIFY},
|
|
{NVGPU_GPU_FLAGS_SUPPORT_REMAP,
|
|
NVGPU_SUPPORT_REMAP},
|
|
{NVGPU_GPU_FLAGS_SUPPORT_COMPRESSION,
|
|
NVGPU_SUPPORT_COMPRESSION},
|
|
{NVGPU_GPU_FLAGS_SUPPORT_SM_TTU,
|
|
NVGPU_SUPPORT_SM_TTU},
|
|
{NVGPU_GPU_FLAGS_SUPPORT_POST_L2_COMPRESSION,
|
|
NVGPU_SUPPORT_POST_L2_COMPRESSION},
|
|
{NVGPU_GPU_FLAGS_SUPPORT_MAP_ACCESS_TYPE,
|
|
NVGPU_SUPPORT_MAP_ACCESS_TYPE},
|
|
{NVGPU_GPU_FLAGS_SUPPORT_2D,
|
|
NVGPU_SUPPORT_2D},
|
|
{NVGPU_GPU_FLAGS_SUPPORT_3D,
|
|
NVGPU_SUPPORT_3D},
|
|
{NVGPU_GPU_FLAGS_SUPPORT_COMPUTE,
|
|
NVGPU_SUPPORT_COMPUTE},
|
|
{NVGPU_GPU_FLAGS_SUPPORT_I2M,
|
|
NVGPU_SUPPORT_I2M},
|
|
{NVGPU_GPU_FLAGS_SUPPORT_ZBC,
|
|
NVGPU_SUPPORT_ZBC},
|
|
{NVGPU_GPU_FLAGS_SUPPORT_PROFILER_V2_DEVICE,
|
|
NVGPU_SUPPORT_PROFILER_V2_DEVICE},
|
|
{NVGPU_GPU_FLAGS_SUPPORT_PROFILER_V2_CONTEXT,
|
|
NVGPU_SUPPORT_PROFILER_V2_CONTEXT},
|
|
{NVGPU_GPU_FLAGS_SUPPORT_SMPC_GLOBAL_MODE,
|
|
NVGPU_SUPPORT_SMPC_GLOBAL_MODE},
|
|
{NVGPU_GPU_FLAGS_SUPPORT_GET_GR_CONTEXT,
|
|
NVGPU_SUPPORT_GET_GR_CONTEXT},
|
|
{NVGPU_GPU_FLAGS_L2_MAX_WAYS_EVICT_LAST_ENABLED,
|
|
NVGPU_L2_MAX_WAYS_EVICT_LAST_ENABLED},
|
|
{NVGPU_GPU_FLAGS_SUPPORT_VAB,
|
|
NVGPU_SUPPORT_VAB_ENABLED},
|
|
{NVGPU_GPU_FLAGS_SUPPORT_BUFFER_METADATA,
|
|
NVGPU_SUPPORT_BUFFER_METADATA},
|
|
{NVGPU_GPU_FLAGS_SUPPORT_NVS,
|
|
NVGPU_SUPPORT_NVS},
|
|
};
|
|
|
|
static u64 nvgpu_ctrl_ioctl_gpu_characteristics_flags(struct gk20a *g)
|
|
{
|
|
unsigned int i;
|
|
u64 ioctl_flags = 0;
|
|
|
|
for (i = 0; i < sizeof(flags_mapping)/sizeof(*flags_mapping); i++) {
|
|
if (nvgpu_is_enabled(g, flags_mapping[i].enabled_flag))
|
|
ioctl_flags |= flags_mapping[i].ioctl_flag;
|
|
}
|
|
|
|
if (!capable(CAP_SYS_NICE)) {
|
|
ioctl_flags &= ~NVGPU_GPU_FLAGS_SUPPORT_RESCHEDULE_RUNLIST;
|
|
}
|
|
|
|
return ioctl_flags;
|
|
}
|
|
|
|
static void nvgpu_set_preemption_mode_flags(struct gk20a *g,
|
|
struct nvgpu_gpu_characteristics *gpu)
|
|
{
|
|
u32 graphics_preemption_mode_flags = 0U;
|
|
u32 compute_preemption_mode_flags = 0U;
|
|
u32 default_graphics_preempt_mode = 0U;
|
|
u32 default_compute_preempt_mode = 0U;
|
|
|
|
g->ops.gr.init.get_supported__preemption_modes(
|
|
&graphics_preemption_mode_flags,
|
|
&compute_preemption_mode_flags);
|
|
g->ops.gr.init.get_default_preemption_modes(
|
|
&default_graphics_preempt_mode,
|
|
&default_compute_preempt_mode);
|
|
|
|
gpu->graphics_preemption_mode_flags =
|
|
nvgpu_get_ioctl_graphics_preempt_mode_flags(
|
|
graphics_preemption_mode_flags);
|
|
gpu->compute_preemption_mode_flags =
|
|
nvgpu_get_ioctl_compute_preempt_mode_flags(
|
|
compute_preemption_mode_flags);
|
|
|
|
gpu->default_graphics_preempt_mode =
|
|
nvgpu_get_ioctl_graphics_preempt_mode(
|
|
default_graphics_preempt_mode);
|
|
gpu->default_compute_preempt_mode =
|
|
nvgpu_get_ioctl_compute_preempt_mode(
|
|
default_compute_preempt_mode);
|
|
}
|
|
|
|
static long gk20a_ctrl_ioctl_gpu_characteristics(
|
|
struct gk20a *g, u32 gpu_instance_id, struct nvgpu_gr_config *gr_config,
|
|
struct nvgpu_gpu_get_characteristics *request)
|
|
{
|
|
struct nvgpu_gpu_characteristics gpu;
|
|
long err = 0;
|
|
struct nvgpu_gpu_instance *gpu_instance;
|
|
|
|
u32 gr_instance_id = nvgpu_grmgr_get_gr_instance_id(g, gpu_instance_id);
|
|
|
|
if (gk20a_busy(g)) {
|
|
nvgpu_err(g, "failed to power on gpu");
|
|
return -EINVAL;
|
|
}
|
|
|
|
(void) memset(&gpu, 0, sizeof(gpu));
|
|
gpu_instance = &g->mig.gpu_instance[gpu_instance_id];
|
|
|
|
gpu.L2_cache_size = g->ops.ltc.determine_L2_size_bytes(g);
|
|
gpu.on_board_video_memory_size = 0; /* integrated GPU */
|
|
|
|
gpu.num_gpc = nvgpu_gr_config_get_gpc_count(gr_config);
|
|
gpu.max_gpc_count = nvgpu_gr_config_get_max_gpc_count(gr_config);
|
|
/* Convert logical to physical masks */
|
|
gpu.gpc_mask = nvgpu_grmgr_get_gr_physical_gpc_mask(g, gr_instance_id);
|
|
|
|
nvgpu_log(g, gpu_dbg_fn | gpu_dbg_gpu_dbg,
|
|
"GR Instance ID = %u, physical gpc_mask = 0x%08X, logical gpc_mask = 0x%08X",
|
|
gr_instance_id, gpu.gpc_mask, nvgpu_grmgr_get_gr_logical_gpc_mask(g,
|
|
gr_instance_id));
|
|
|
|
gpu.num_tpc_per_gpc = nvgpu_gr_config_get_max_tpc_per_gpc_count(gr_config);
|
|
|
|
gpu.num_ppc_per_gpc = nvgpu_gr_config_get_pe_count_per_gpc(gr_config);
|
|
|
|
gpu.max_veid_count_per_tsg =
|
|
gpu_instance->gr_syspipe.max_veid_count_per_tsg;
|
|
|
|
gpu.bus_type = NVGPU_GPU_BUS_TYPE_AXI; /* always AXI for now */
|
|
|
|
#ifdef CONFIG_NVGPU_COMPRESSION
|
|
if (nvgpu_is_enabled(g, NVGPU_SUPPORT_COMPRESSION)) {
|
|
gpu.compression_page_size = g->ops.fb.compression_page_size(g);
|
|
gpu.gr_compbit_store_base_hw = g->cbc->compbit_store.base_hw;
|
|
gpu.gr_gobs_per_comptagline_per_slice =
|
|
g->cbc->gobs_per_comptagline_per_slice;
|
|
gpu.cbc_comptags_per_line = g->cbc->comptags_per_cacheline;
|
|
}
|
|
#endif
|
|
|
|
if (!nvgpu_is_enabled(g, NVGPU_SUPPORT_MIG) ||
|
|
(gpu_instance_id != 0U) ||
|
|
(!nvgpu_grmgr_is_multi_gr_enabled(g))) {
|
|
gpu.flags = nvgpu_ctrl_ioctl_gpu_characteristics_flags(g);
|
|
nvgpu_set_preemption_mode_flags(g, &gpu);
|
|
} else {
|
|
gpu.flags = (NVGPU_GPU_FLAGS_SUPPORT_CLOCK_CONTROLS |
|
|
NVGPU_GPU_FLAGS_SUPPORT_GET_VOLTAGE |
|
|
NVGPU_GPU_FLAGS_SUPPORT_GET_CURRENT |
|
|
NVGPU_GPU_FLAGS_SUPPORT_GET_POWER |
|
|
NVGPU_GPU_FLAGS_SUPPORT_GET_TEMPERATURE |
|
|
NVGPU_GPU_FLAGS_SUPPORT_SET_THERM_ALERT_LIMIT |
|
|
NVGPU_GPU_FLAGS_SUPPORT_DEVICE_EVENTS |
|
|
NVGPU_GPU_FLAGS_SUPPORT_SM_TTU |
|
|
NVGPU_GPU_FLAGS_SUPPORT_PROFILER_V2_DEVICE |
|
|
NVGPU_GPU_FLAGS_SUPPORT_PROFILER_V2_CONTEXT |
|
|
NVGPU_GPU_FLAGS_SUPPORT_SMPC_GLOBAL_MODE);
|
|
}
|
|
|
|
gpu.arch = g->params.gpu_arch;
|
|
gpu.impl = g->params.gpu_impl;
|
|
gpu.rev = g->params.gpu_rev;
|
|
gpu.reg_ops_limit = NVGPU_IOCTL_DBG_REG_OPS_LIMIT;
|
|
gpu.map_buffer_batch_limit = nvgpu_is_enabled(g, NVGPU_SUPPORT_MAP_BUFFER_BATCH) ?
|
|
NVGPU_IOCTL_AS_MAP_BUFFER_BATCH_LIMIT : 0;
|
|
|
|
if (nvgpu_is_enabled(g, NVGPU_SUPPORT_MIG)) {
|
|
if (gpu_instance_id != 0U) {
|
|
gpu.compute_class = g->ops.get_litter_value(g, GPU_LIT_COMPUTE_CLASS);
|
|
gpu.gpfifo_class = g->ops.get_litter_value(g, GPU_LIT_GPFIFO_CLASS);
|
|
gpu.dma_copy_class =
|
|
g->ops.get_litter_value(g, GPU_LIT_DMA_COPY_CLASS);
|
|
}
|
|
} else {
|
|
#ifdef CONFIG_NVGPU_GRAPHICS
|
|
gpu.twod_class = g->ops.get_litter_value(g, GPU_LIT_TWOD_CLASS);
|
|
gpu.threed_class = g->ops.get_litter_value(g, GPU_LIT_THREED_CLASS);
|
|
#endif
|
|
gpu.compute_class = g->ops.get_litter_value(g, GPU_LIT_COMPUTE_CLASS);
|
|
gpu.gpfifo_class = g->ops.get_litter_value(g, GPU_LIT_GPFIFO_CLASS);
|
|
gpu.inline_to_memory_class =
|
|
g->ops.get_litter_value(g, GPU_LIT_I2M_CLASS);
|
|
gpu.dma_copy_class =
|
|
g->ops.get_litter_value(g, GPU_LIT_DMA_COPY_CLASS);
|
|
}
|
|
|
|
#ifdef CONFIG_NVGPU_DGPU
|
|
gpu.vbios_version = nvgpu_bios_get_vbios_version(g);
|
|
gpu.vbios_oem_version = nvgpu_bios_get_vbios_oem_version(g);
|
|
#else
|
|
gpu.vbios_version = 0;
|
|
gpu.vbios_oem_version = 0;
|
|
#endif
|
|
gpu.big_page_size = nvgpu_mm_get_default_big_page_size(g);
|
|
gpu.pde_coverage_bit_count =
|
|
g->ops.mm.gmmu.get_mmu_levels(g,
|
|
gpu.big_page_size)[0].lo_bit[0];
|
|
gpu.available_big_page_sizes = nvgpu_mm_get_available_big_page_sizes(g);
|
|
|
|
gpu.sm_arch_sm_version = g->params.sm_arch_sm_version;
|
|
gpu.sm_arch_spa_version = g->params.sm_arch_spa_version;
|
|
gpu.sm_arch_warp_count = g->params.sm_arch_warp_count;
|
|
|
|
gpu.max_css_buffer_size = g->ops.css.get_max_buffer_size(g);;
|
|
gpu.max_ctxsw_ring_buffer_size = GK20A_CTXSW_TRACE_MAX_VM_RING_SIZE;
|
|
|
|
gpu.gpu_ioctl_nr_last = NVGPU_GPU_IOCTL_LAST;
|
|
gpu.tsg_ioctl_nr_last = NVGPU_TSG_IOCTL_LAST;
|
|
gpu.dbg_gpu_ioctl_nr_last = NVGPU_DBG_GPU_IOCTL_LAST;
|
|
gpu.ioctl_channel_nr_last = NVGPU_IOCTL_CHANNEL_LAST;
|
|
gpu.as_ioctl_nr_last = NVGPU_AS_IOCTL_LAST;
|
|
gpu.event_ioctl_nr_last = NVGPU_EVENT_IOCTL_LAST;
|
|
gpu.ctxsw_ioctl_nr_last = NVGPU_CTXSW_IOCTL_LAST;
|
|
gpu.prof_ioctl_nr_last = NVGPU_PROFILER_IOCTL_LAST;
|
|
gpu.nvs_ioctl_nr_last = NVGPU_NVS_IOCTL_LAST;
|
|
gpu.gpu_va_bit_count = 40;
|
|
gpu.max_dbg_tsg_timeslice = g->tsg_dbg_timeslice_max_us;
|
|
|
|
strlcpy(gpu.chipname, g->name, sizeof(gpu.chipname));
|
|
gpu.max_fbps_count = nvgpu_grmgr_get_max_fbps_count(g);
|
|
gpu.fbp_en_mask = nvgpu_grmgr_get_fbp_en_mask(g, gpu_instance_id);
|
|
gpu.max_ltc_per_fbp = g->ops.top.get_max_ltc_per_fbp(g);
|
|
gpu.max_lts_per_ltc = g->ops.top.get_max_lts_per_ltc(g);
|
|
gpu.num_ltc = nvgpu_ltc_get_ltc_count(g);
|
|
gpu.lts_per_ltc = nvgpu_ltc_get_slices_per_ltc(g);
|
|
gpu.cbc_cache_line_size = nvgpu_ltc_get_cacheline_size(g);
|
|
|
|
/*
|
|
* TODO : Need to replace with proper HAL.
|
|
*/
|
|
if (g->pci_device_id != (u16)0) {
|
|
/* All nvgpu supported dGPUs have 64 bit FBIO channel
|
|
* So number of Sub partition per FBPA is always 0x2.
|
|
* Half FBPA (32BIT channel mode) enablement
|
|
* (1 sub partition per FBPA) is disabled for tegra dGPUs.
|
|
*/
|
|
gpu.num_sub_partition_per_fbpa = 0x2;
|
|
} else {
|
|
/*
|
|
* All iGPUs don't have real FBPA/FBSP units at all.
|
|
* So num_sub_partition_per_fbpa should be 0 for iGPUs.
|
|
*/
|
|
gpu.num_sub_partition_per_fbpa = 0x00;
|
|
}
|
|
|
|
if ((g->ops.clk.get_maxrate) && nvgpu_platform_is_silicon(g)) {
|
|
gpu.max_freq = g->ops.clk.get_maxrate(g,
|
|
CTRL_CLK_DOMAIN_GPCCLK);
|
|
}
|
|
|
|
#ifdef CONFIG_NVGPU_DGPU
|
|
gpu.local_video_memory_size = g->mm.vidmem.size;
|
|
#endif
|
|
|
|
gpu.pci_vendor_id = g->pci_vendor_id;
|
|
gpu.pci_device_id = g->pci_device_id;
|
|
gpu.pci_subsystem_vendor_id = g->pci_subsystem_vendor_id;
|
|
gpu.pci_subsystem_device_id = g->pci_subsystem_device_id;
|
|
gpu.pci_class = g->pci_class;
|
|
gpu.pci_revision = g->pci_revision;
|
|
|
|
gpu.per_device_identifier = g->per_device_identifier;
|
|
|
|
gpu.gpu_instance_id = gpu_instance->gpu_instance_id;
|
|
gpu.gr_instance_id = gpu_instance->gr_syspipe.gr_syspipe_id;
|
|
|
|
gpu.max_gpfifo_entries = rounddown_pow_of_two(U32_MAX /
|
|
nvgpu_get_gpfifo_entry_size());
|
|
|
|
if (request->gpu_characteristics_buf_size > 0) {
|
|
size_t write_size = sizeof(gpu);
|
|
|
|
nvgpu_speculation_barrier();
|
|
if (write_size > request->gpu_characteristics_buf_size)
|
|
write_size = request->gpu_characteristics_buf_size;
|
|
|
|
err = copy_to_user((void __user *)(uintptr_t)
|
|
request->gpu_characteristics_buf_addr,
|
|
&gpu, write_size);
|
|
}
|
|
|
|
if (err == 0)
|
|
request->gpu_characteristics_buf_size = sizeof(gpu);
|
|
|
|
gk20a_idle(g);
|
|
|
|
return err;
|
|
}
|
|
|
|
static int gk20a_ctrl_prepare_compressible_read(
|
|
struct gk20a *g,
|
|
struct nvgpu_gpu_prepare_compressible_read_args *args)
|
|
{
|
|
int ret = -ENOSYS;
|
|
|
|
#ifdef CONFIG_NVGPU_SUPPORT_CDE
|
|
struct nvgpu_os_linux *l = nvgpu_os_linux_from_gk20a(g);
|
|
struct nvgpu_channel_fence fence;
|
|
struct nvgpu_user_fence fence_out = nvgpu_user_fence_init();
|
|
int submit_flags = nvgpu_submit_gpfifo_user_flags_to_common_flags(
|
|
args->submit_flags);
|
|
int fd = -1;
|
|
|
|
fence.id = args->fence.syncpt_id;
|
|
fence.value = args->fence.syncpt_value;
|
|
|
|
/* Try and allocate an fd here*/
|
|
if ((submit_flags & NVGPU_SUBMIT_FLAGS_FENCE_GET)
|
|
&& (submit_flags & NVGPU_SUBMIT_FLAGS_SYNC_FENCE)) {
|
|
fd = get_unused_fd_flags(O_RDWR);
|
|
if (fd < 0)
|
|
return fd;
|
|
}
|
|
|
|
ret = gk20a_prepare_compressible_read(l, args->handle,
|
|
args->request_compbits, args->offset,
|
|
args->compbits_hoffset, args->compbits_voffset,
|
|
args->scatterbuffer_offset,
|
|
args->width, args->height, args->block_height_log2,
|
|
submit_flags, &fence, &args->valid_compbits,
|
|
&args->zbc_color, &fence_out);
|
|
|
|
if (ret) {
|
|
if (fd != -1)
|
|
put_unused_fd(fd);
|
|
return ret;
|
|
}
|
|
|
|
/*
|
|
* Convert fence_out, if any, to something we can pass back to user
|
|
* space. Even if successful, the fence may not exist if there was
|
|
* nothing to be done (no compbits requested); that's not an error.
|
|
*/
|
|
if (submit_flags & NVGPU_SUBMIT_FLAGS_FENCE_GET) {
|
|
if (submit_flags & NVGPU_SUBMIT_FLAGS_SYNC_FENCE) {
|
|
if (nvgpu_os_fence_is_initialized(&fence_out.os_fence)) {
|
|
ret = fence_out.os_fence.ops->install_fence(
|
|
&fence_out.os_fence, fd);
|
|
if (ret) {
|
|
put_unused_fd(fd);
|
|
fd = -1;
|
|
}
|
|
} else {
|
|
put_unused_fd(fd);
|
|
fd = -1;
|
|
}
|
|
args->fence.fd = fd;
|
|
} else {
|
|
args->fence.syncpt_id = fence_out.syncpt_id;
|
|
args->fence.syncpt_value = fence_out.syncpt_value;
|
|
}
|
|
nvgpu_user_fence_release(&fence_out);
|
|
}
|
|
#endif
|
|
|
|
return ret;
|
|
}
|
|
|
|
static int gk20a_ctrl_mark_compressible_write(
|
|
struct gk20a *g,
|
|
struct nvgpu_gpu_mark_compressible_write_args *args)
|
|
{
|
|
int ret = -ENOSYS;
|
|
|
|
#ifdef CONFIG_NVGPU_SUPPORT_CDE
|
|
ret = gk20a_mark_compressible_write(g, args->handle,
|
|
args->valid_compbits, args->offset, args->zbc_color);
|
|
#endif
|
|
|
|
return ret;
|
|
}
|
|
|
|
static int gk20a_ctrl_alloc_as(
|
|
struct gk20a *g,
|
|
struct nvgpu_alloc_as_args *args)
|
|
{
|
|
struct gk20a_as_share *as_share;
|
|
int err;
|
|
int fd;
|
|
struct file *file;
|
|
char name[64];
|
|
|
|
err = get_unused_fd_flags(O_RDWR);
|
|
if (err < 0)
|
|
return err;
|
|
fd = err;
|
|
|
|
(void) snprintf(name, sizeof(name), "nvhost-%s-fd%d", g->name, fd);
|
|
|
|
err = gk20a_as_alloc_share(g, args->big_page_size,
|
|
gk20a_as_translate_as_alloc_flags(g,
|
|
args->flags),
|
|
args->va_range_start,
|
|
args->va_range_end,
|
|
args->va_range_split,
|
|
&as_share);
|
|
if (err)
|
|
goto clean_up;
|
|
|
|
file = anon_inode_getfile(name, &gk20a_as_ops, as_share, O_RDWR);
|
|
if (IS_ERR(file)) {
|
|
err = PTR_ERR(file);
|
|
goto clean_up_as;
|
|
}
|
|
|
|
fd_install(fd, file);
|
|
|
|
args->as_fd = fd;
|
|
return 0;
|
|
|
|
clean_up_as:
|
|
gk20a_as_release_share(as_share);
|
|
clean_up:
|
|
put_unused_fd(fd);
|
|
return err;
|
|
}
|
|
|
|
static int gk20a_ctrl_open_tsg(struct gk20a *g, struct nvgpu_cdev *cdev,
|
|
struct nvgpu_gpu_open_tsg_args *args)
|
|
{
|
|
int err;
|
|
int fd;
|
|
struct file *file;
|
|
char name[64];
|
|
|
|
err = get_unused_fd_flags(O_RDWR);
|
|
if (err < 0)
|
|
return err;
|
|
fd = err;
|
|
|
|
(void) snprintf(name, sizeof(name), "nvgpu-%s-tsg%d", g->name, fd);
|
|
|
|
file = anon_inode_getfile(name, &gk20a_tsg_ops, NULL, O_RDWR);
|
|
if (IS_ERR(file)) {
|
|
err = PTR_ERR(file);
|
|
goto clean_up;
|
|
}
|
|
|
|
err = nvgpu_ioctl_tsg_open(g, cdev, file);
|
|
if (err)
|
|
goto clean_up_file;
|
|
|
|
fd_install(fd, file);
|
|
args->tsg_fd = fd;
|
|
return 0;
|
|
|
|
clean_up_file:
|
|
fput(file);
|
|
clean_up:
|
|
put_unused_fd(fd);
|
|
return err;
|
|
}
|
|
|
|
static int gk20a_ctrl_get_tpc_masks(struct gk20a *g, struct nvgpu_gr_config *gr_config,
|
|
struct nvgpu_gpu_get_tpc_masks_args *args)
|
|
{
|
|
int err = 0;
|
|
const u32 gpc_tpc_mask_size = sizeof(u32) *
|
|
nvgpu_gr_config_get_max_gpc_count(gr_config);
|
|
|
|
if (args->mask_buf_size > 0) {
|
|
size_t write_size = gpc_tpc_mask_size;
|
|
|
|
nvgpu_speculation_barrier();
|
|
if (write_size > args->mask_buf_size)
|
|
write_size = args->mask_buf_size;
|
|
|
|
err = copy_to_user((void __user *)(uintptr_t)
|
|
args->mask_buf_addr,
|
|
nvgpu_gr_config_get_gpc_tpc_mask_physical_base(gr_config),
|
|
write_size);
|
|
}
|
|
|
|
if (err == 0)
|
|
args->mask_buf_size = gpc_tpc_mask_size;
|
|
|
|
return err;
|
|
}
|
|
|
|
static int gk20a_ctrl_get_fbp_l2_masks(
|
|
struct gk20a *g, u32 gpu_instance_id,
|
|
struct nvgpu_gpu_get_fbp_l2_masks_args *args)
|
|
{
|
|
int err = 0;
|
|
const u32 fbp_l2_mask_size = sizeof(u32) *
|
|
nvgpu_grmgr_get_max_fbps_count(g);
|
|
u32 *fbp_l2_en_mask =
|
|
nvgpu_grmgr_get_fbp_l2_en_mask(g, gpu_instance_id);
|
|
|
|
if (args->mask_buf_size > 0) {
|
|
size_t write_size = fbp_l2_mask_size;
|
|
|
|
nvgpu_speculation_barrier();
|
|
if (write_size > args->mask_buf_size)
|
|
write_size = args->mask_buf_size;
|
|
|
|
err = copy_to_user((void __user *)(uintptr_t)
|
|
args->mask_buf_addr,
|
|
fbp_l2_en_mask, write_size);
|
|
}
|
|
|
|
if (err == 0)
|
|
args->mask_buf_size = fbp_l2_mask_size;
|
|
|
|
return err;
|
|
}
|
|
|
|
static int nvgpu_gpu_ioctl_l2_fb_ops(struct gk20a *g,
|
|
struct nvgpu_gpu_l2_fb_args *args)
|
|
{
|
|
int err = 0;
|
|
|
|
if ((!args->l2_flush && !args->fb_flush) ||
|
|
(!args->l2_flush && args->l2_invalidate))
|
|
return -EINVAL;
|
|
|
|
/* In case of railgating enabled, exit if nvgpu is powered off */
|
|
if (nvgpu_is_enabled(g, NVGPU_CAN_RAILGATE) && nvgpu_is_powered_off(g)) {
|
|
return 0;
|
|
}
|
|
|
|
err = gk20a_busy(g);
|
|
if (err != 0) {
|
|
nvgpu_err(g, "failed to take power ref");
|
|
return err;
|
|
}
|
|
|
|
if (args->l2_flush) {
|
|
err = nvgpu_pg_elpg_ms_protected_call(g,
|
|
g->ops.mm.cache.l2_flush(g, args->l2_invalidate ?
|
|
true : false));
|
|
if (err != 0) {
|
|
nvgpu_err(g, "l2_flush failed");
|
|
goto out;
|
|
}
|
|
}
|
|
|
|
if (args->fb_flush) {
|
|
err = g->ops.mm.cache.fb_flush(g);
|
|
if (err != 0) {
|
|
nvgpu_err(g, "mm.cache.fb_flush() failed err=%d", err);
|
|
goto out;
|
|
}
|
|
}
|
|
|
|
out:
|
|
gk20a_idle(g);
|
|
|
|
return err;
|
|
}
|
|
|
|
static int nvgpu_gpu_ioctl_set_mmu_debug_mode(
|
|
struct gk20a *g,
|
|
struct nvgpu_gpu_mmu_debug_mode_args *args)
|
|
{
|
|
if (gk20a_busy(g)) {
|
|
nvgpu_err(g, "failed to power on gpu");
|
|
return -EINVAL;
|
|
}
|
|
|
|
nvgpu_mutex_acquire(&g->dbg_sessions_lock);
|
|
g->ops.fb.set_debug_mode(g, args->state == 1);
|
|
nvgpu_mutex_release(&g->dbg_sessions_lock);
|
|
|
|
gk20a_idle(g);
|
|
return 0;
|
|
}
|
|
|
|
static int nvgpu_gpu_ioctl_set_debug_mode(
|
|
struct gk20a *g,
|
|
struct nvgpu_gpu_sm_debug_mode_args *args,
|
|
u32 gr_instance_id)
|
|
{
|
|
struct nvgpu_channel *ch;
|
|
int err;
|
|
|
|
ch = nvgpu_channel_get_from_file(args->channel_fd);
|
|
if (!ch)
|
|
return -EINVAL;
|
|
|
|
nvgpu_mutex_acquire(&g->dbg_sessions_lock);
|
|
if (g->ops.gr.set_sm_debug_mode)
|
|
err = nvgpu_gr_exec_with_err_for_instance(g, gr_instance_id,
|
|
g->ops.gr.set_sm_debug_mode(g, ch,
|
|
args->sms, !!args->enable));
|
|
else
|
|
err = -ENOSYS;
|
|
nvgpu_mutex_release(&g->dbg_sessions_lock);
|
|
|
|
nvgpu_channel_put(ch);
|
|
return err;
|
|
}
|
|
|
|
static int nvgpu_gpu_ioctl_trigger_suspend(struct gk20a *g, u32 gr_instance_id)
|
|
{
|
|
int err;
|
|
|
|
err = gk20a_busy(g);
|
|
if (err)
|
|
return err;
|
|
|
|
nvgpu_mutex_acquire(&g->dbg_sessions_lock);
|
|
if (g->ops.gr.trigger_suspend != NULL) {
|
|
err = nvgpu_gr_exec_with_err_for_instance(g, gr_instance_id,
|
|
g->ops.gr.trigger_suspend(g));
|
|
} else {
|
|
err = -ENOSYS;
|
|
}
|
|
nvgpu_mutex_release(&g->dbg_sessions_lock);
|
|
|
|
gk20a_idle(g);
|
|
|
|
return err;
|
|
}
|
|
|
|
static int nvgpu_gpu_ioctl_wait_for_pause(struct gk20a *g,
|
|
struct nvgpu_gpu_wait_pause_args *args, u32 gr_instance_id)
|
|
{
|
|
int err;
|
|
struct warpstate *ioctl_w_state;
|
|
struct nvgpu_warpstate *w_state = NULL;
|
|
u32 ioctl_size, size, sm_id, no_of_sm;
|
|
struct nvgpu_gr_config *gr_config =
|
|
nvgpu_gr_get_gr_instance_config_ptr(g, gr_instance_id);
|
|
|
|
no_of_sm = nvgpu_gr_config_get_no_of_sm(gr_config);
|
|
|
|
ioctl_size = no_of_sm * sizeof(struct warpstate);
|
|
ioctl_w_state = nvgpu_kzalloc(g, ioctl_size);
|
|
if (!ioctl_w_state) {
|
|
return -ENOMEM;
|
|
}
|
|
|
|
size = no_of_sm * sizeof(struct nvgpu_warpstate);
|
|
w_state = nvgpu_kzalloc(g, size);
|
|
if (!w_state) {
|
|
err = -ENOMEM;
|
|
goto out_free;
|
|
}
|
|
|
|
err = gk20a_busy(g);
|
|
if (err) {
|
|
goto out_free;
|
|
}
|
|
|
|
nvgpu_mutex_acquire(&g->dbg_sessions_lock);
|
|
if (g->ops.gr.wait_for_pause != NULL) {
|
|
err = nvgpu_gr_exec_with_err_for_instance(g, gr_instance_id,
|
|
g->ops.gr.wait_for_pause(g, w_state));
|
|
|
|
for (sm_id = 0; sm_id < no_of_sm; sm_id++) {
|
|
ioctl_w_state[sm_id].valid_warps[0] =
|
|
w_state[sm_id].valid_warps[0];
|
|
ioctl_w_state[sm_id].valid_warps[1] =
|
|
w_state[sm_id].valid_warps[1];
|
|
ioctl_w_state[sm_id].trapped_warps[0] =
|
|
w_state[sm_id].trapped_warps[0];
|
|
ioctl_w_state[sm_id].trapped_warps[1] =
|
|
w_state[sm_id].trapped_warps[1];
|
|
ioctl_w_state[sm_id].paused_warps[0] =
|
|
w_state[sm_id].paused_warps[0];
|
|
ioctl_w_state[sm_id].paused_warps[1] =
|
|
w_state[sm_id].paused_warps[1];
|
|
}
|
|
/* Copy to user space - pointed by "args->pwarpstate" */
|
|
if (copy_to_user((void __user *)(uintptr_t)args->pwarpstate,
|
|
w_state, ioctl_size)) {
|
|
nvgpu_log_fn(g, "copy_to_user failed!");
|
|
err = -EFAULT;
|
|
}
|
|
} else {
|
|
err = -ENOSYS;
|
|
}
|
|
|
|
nvgpu_mutex_release(&g->dbg_sessions_lock);
|
|
|
|
gk20a_idle(g);
|
|
|
|
out_free:
|
|
nvgpu_kfree(g, w_state);
|
|
nvgpu_kfree(g, ioctl_w_state);
|
|
|
|
return err;
|
|
}
|
|
|
|
static int nvgpu_gpu_ioctl_resume_from_pause(struct gk20a *g, u32 gr_instance_id)
|
|
{
|
|
int err;
|
|
|
|
err = gk20a_busy(g);
|
|
if (err)
|
|
return err;
|
|
|
|
nvgpu_mutex_acquire(&g->dbg_sessions_lock);
|
|
if (g->ops.gr.resume_from_pause != NULL) {
|
|
err = nvgpu_gr_exec_with_err_for_instance(g, gr_instance_id,
|
|
g->ops.gr.resume_from_pause(g));
|
|
} else {
|
|
err = -ENOSYS;
|
|
}
|
|
nvgpu_mutex_release(&g->dbg_sessions_lock);
|
|
|
|
gk20a_idle(g);
|
|
|
|
return err;
|
|
}
|
|
|
|
static int nvgpu_gpu_ioctl_clear_sm_errors(struct gk20a *g, u32 gr_instance_id)
|
|
{
|
|
int err;
|
|
|
|
if (g->ops.gr.clear_sm_errors == NULL) {
|
|
return -ENOSYS;
|
|
}
|
|
|
|
err = gk20a_busy(g);
|
|
if (err)
|
|
return err;
|
|
|
|
err = nvgpu_gr_exec_with_err_for_instance(g, gr_instance_id,
|
|
g->ops.gr.clear_sm_errors(g));
|
|
|
|
gk20a_idle(g);
|
|
|
|
return err;
|
|
}
|
|
|
|
static int nvgpu_gpu_ioctl_has_any_exception(
|
|
struct gk20a *g,
|
|
struct nvgpu_gpu_tpc_exception_en_status_args *args)
|
|
{
|
|
u64 tpc_exception_en;
|
|
|
|
if (g->ops.gr.intr.tpc_enabled_exceptions == NULL) {
|
|
return -ENOSYS;
|
|
}
|
|
|
|
nvgpu_mutex_acquire(&g->dbg_sessions_lock);
|
|
tpc_exception_en = nvgpu_pg_elpg_protected_call(g,
|
|
g->ops.gr.intr.tpc_enabled_exceptions(g));
|
|
nvgpu_mutex_release(&g->dbg_sessions_lock);
|
|
|
|
args->tpc_exception_en_sm_mask = tpc_exception_en;
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int gk20a_ctrl_get_num_vsms(struct gk20a *g,
|
|
struct nvgpu_gr_config *gr_config, struct nvgpu_gpu_num_vsms *args)
|
|
{
|
|
args->num_vsms = nvgpu_gr_config_get_no_of_sm(gr_config);
|
|
return 0;
|
|
}
|
|
|
|
static int gk20a_ctrl_vsm_mapping(struct gk20a *g,
|
|
struct nvgpu_gr_config *gr_config, struct nvgpu_gpu_vsms_mapping *args)
|
|
{
|
|
int err = 0;
|
|
u32 no_of_sm = nvgpu_gr_config_get_no_of_sm(gr_config);
|
|
size_t write_size = no_of_sm *
|
|
sizeof(struct nvgpu_gpu_vsms_mapping_entry);
|
|
struct nvgpu_gpu_vsms_mapping_entry *vsms_buf;
|
|
u32 i;
|
|
|
|
vsms_buf = nvgpu_kzalloc(g, write_size);
|
|
if (vsms_buf == NULL)
|
|
return -ENOMEM;
|
|
|
|
for (i = 0; i < no_of_sm; i++) {
|
|
struct nvgpu_sm_info *sm_info =
|
|
nvgpu_gr_config_get_sm_info(gr_config, i);
|
|
|
|
vsms_buf[i].gpc_index =
|
|
nvgpu_gr_config_get_sm_info_gpc_index(sm_info);
|
|
if (g->ops.gr.init.get_nonpes_aware_tpc)
|
|
vsms_buf[i].tpc_index =
|
|
g->ops.gr.init.get_nonpes_aware_tpc(g,
|
|
nvgpu_gr_config_get_sm_info_gpc_index(sm_info),
|
|
nvgpu_gr_config_get_sm_info_tpc_index(sm_info),
|
|
gr_config);
|
|
else
|
|
vsms_buf[i].tpc_index =
|
|
nvgpu_gr_config_get_sm_info_tpc_index(sm_info);
|
|
}
|
|
|
|
err = copy_to_user((void __user *)(uintptr_t)
|
|
args->vsms_map_buf_addr,
|
|
vsms_buf, write_size);
|
|
nvgpu_kfree(g, vsms_buf);
|
|
|
|
return err;
|
|
}
|
|
|
|
static int nvgpu_gpu_get_cpu_time_correlation_info(
|
|
struct gk20a *g,
|
|
struct nvgpu_gpu_get_cpu_time_correlation_info_args *args)
|
|
{
|
|
struct nvgpu_cpu_time_correlation_sample *samples;
|
|
int err;
|
|
u32 i;
|
|
|
|
if (args->count > NVGPU_GPU_GET_CPU_TIME_CORRELATION_INFO_MAX_COUNT ||
|
|
args->source_id != NVGPU_GPU_GET_CPU_TIME_CORRELATION_INFO_SRC_ID_TSC)
|
|
return -EINVAL;
|
|
|
|
samples = nvgpu_kzalloc(g, args->count *
|
|
sizeof(struct nvgpu_cpu_time_correlation_sample));
|
|
if (!samples) {
|
|
return -ENOMEM;
|
|
}
|
|
|
|
err = g->ops.ptimer.get_timestamps_zipper(g,
|
|
args->source_id, args->count, samples);
|
|
if (!err) {
|
|
for (i = 0; i < args->count; i++) {
|
|
args->samples[i].cpu_timestamp = samples[i].cpu_timestamp;
|
|
args->samples[i].gpu_timestamp = samples[i].gpu_timestamp;
|
|
}
|
|
}
|
|
|
|
nvgpu_kfree(g, samples);
|
|
|
|
return err;
|
|
}
|
|
|
|
static int nvgpu_gpu_get_gpu_time(
|
|
struct gk20a *g,
|
|
struct nvgpu_gpu_get_gpu_time_args *args)
|
|
{
|
|
u64 time;
|
|
int err;
|
|
|
|
err = gk20a_busy(g);
|
|
if (err)
|
|
return err;
|
|
|
|
err = g->ops.ptimer.read_ptimer(g, &time);
|
|
if (!err)
|
|
args->gpu_timestamp = time;
|
|
|
|
gk20a_idle(g);
|
|
return err;
|
|
}
|
|
|
|
static void nvgpu_gpu_fetch_engine_info_item(struct gk20a *g,
|
|
struct nvgpu_gpu_get_engine_info_item *dst_info,
|
|
const struct nvgpu_device *dev, u32 dev_inst_id, u32 gr_runlist_id)
|
|
{
|
|
(void) memset(dst_info, 0, sizeof(*dst_info));
|
|
|
|
if (nvgpu_device_is_graphics(g, dev)) {
|
|
dst_info->engine_id = NVGPU_GPU_ENGINE_ID_GR;
|
|
} else if (nvgpu_device_is_ce(g, dev)) {
|
|
/*
|
|
* There's two types of CE userpsace is interested in:
|
|
* ASYNC_CEs which are copy engines with their own
|
|
* runlists and GRCEs which are CEs that share a runlist
|
|
* with GR.
|
|
*/
|
|
if (dev->runlist_id == gr_runlist_id) {
|
|
dst_info->engine_id = NVGPU_GPU_ENGINE_ID_GR_COPY;
|
|
} else {
|
|
dst_info->engine_id = NVGPU_GPU_ENGINE_ID_ASYNC_COPY;
|
|
}
|
|
}
|
|
|
|
dst_info->engine_instance = dev_inst_id;
|
|
dst_info->runlist_id = dev->runlist_id;
|
|
}
|
|
|
|
static int nvgpu_gpu_get_engine_info(
|
|
struct gk20a *g,
|
|
struct nvgpu_gpu_get_engine_info_args *args)
|
|
{
|
|
int err = 0;
|
|
u32 report_index = 0;
|
|
u32 i;
|
|
const struct nvgpu_device *gr_dev;
|
|
const u32 max_buffer_engines = args->engine_info_buf_size /
|
|
sizeof(struct nvgpu_gpu_get_engine_info_item);
|
|
struct nvgpu_gpu_get_engine_info_item __user *dst_item_list =
|
|
(void __user *)(uintptr_t)args->engine_info_buf_addr;
|
|
|
|
gr_dev = nvgpu_device_get(g, NVGPU_DEVTYPE_GRAPHICS, 0);
|
|
nvgpu_assert(gr_dev != NULL);
|
|
|
|
for (i = 0; i < g->fifo.num_engines; i++) {
|
|
const struct nvgpu_device *dev = g->fifo.active_engines[i];
|
|
struct nvgpu_gpu_get_engine_info_item dst_info;
|
|
|
|
nvgpu_gpu_fetch_engine_info_item(g, &dst_info, dev,
|
|
dev->inst_id, gr_dev->runlist_id);
|
|
|
|
if (report_index < max_buffer_engines) {
|
|
err = copy_to_user(&dst_item_list[report_index],
|
|
&dst_info, sizeof(dst_info));
|
|
if (err)
|
|
goto clean_up;
|
|
}
|
|
|
|
++report_index;
|
|
}
|
|
|
|
args->engine_info_buf_size =
|
|
report_index * sizeof(struct nvgpu_gpu_get_engine_info_item);
|
|
|
|
clean_up:
|
|
return err;
|
|
}
|
|
|
|
static int nvgpu_gpu_get_gpu_instance_engine_info(
|
|
struct gk20a *g, u32 gpu_instance_id,
|
|
struct nvgpu_gpu_get_engine_info_args *args)
|
|
{
|
|
int err = 0;
|
|
u32 report_index = 0U;
|
|
u32 i;
|
|
const struct nvgpu_device *gr_dev;
|
|
const u32 max_buffer_engines = args->engine_info_buf_size /
|
|
sizeof(struct nvgpu_gpu_get_engine_info_item);
|
|
struct nvgpu_gpu_get_engine_info_item __user *dst_item_list =
|
|
(void __user *)(uintptr_t)args->engine_info_buf_addr;
|
|
struct nvgpu_gpu_get_engine_info_item dst_info;
|
|
struct nvgpu_gpu_instance *gpu_instance =
|
|
&g->mig.gpu_instance[gpu_instance_id];
|
|
|
|
gr_dev = gpu_instance->gr_syspipe.gr_dev;
|
|
nvgpu_assert(gr_dev != NULL);
|
|
|
|
nvgpu_gpu_fetch_engine_info_item(g, &dst_info, gr_dev, 0U, gr_dev->runlist_id);
|
|
|
|
if (report_index < max_buffer_engines) {
|
|
err = copy_to_user(&dst_item_list[report_index],
|
|
&dst_info, sizeof(dst_info));
|
|
if (err)
|
|
goto clean_up;
|
|
}
|
|
|
|
++report_index;
|
|
|
|
for (i = 0U; i < gpu_instance->num_lce; i++) {
|
|
const struct nvgpu_device *dev = gpu_instance->lce_devs[i];
|
|
|
|
nvgpu_gpu_fetch_engine_info_item(g, &dst_info, dev, i, gr_dev->runlist_id);
|
|
|
|
if (report_index < max_buffer_engines) {
|
|
err = copy_to_user(&dst_item_list[report_index],
|
|
&dst_info, sizeof(dst_info));
|
|
if (err)
|
|
goto clean_up;
|
|
}
|
|
|
|
++report_index;
|
|
}
|
|
|
|
args->engine_info_buf_size =
|
|
report_index * sizeof(struct nvgpu_gpu_get_engine_info_item);
|
|
|
|
clean_up:
|
|
return err;
|
|
}
|
|
|
|
#ifdef CONFIG_NVGPU_DGPU
|
|
static int nvgpu_gpu_alloc_vidmem(struct gk20a *g,
|
|
struct nvgpu_gpu_alloc_vidmem_args *args)
|
|
{
|
|
u32 align = args->in.alignment ? args->in.alignment : SZ_4K;
|
|
int fd;
|
|
|
|
nvgpu_log_fn(g, " ");
|
|
|
|
if (args->in.flags & NVGPU_GPU_ALLOC_VIDMEM_FLAG_CPU_MASK) {
|
|
nvgpu_warn(g,
|
|
"Allocating vidmem with FLAG_CPU_MASK is not yet supported");
|
|
return -EINVAL;
|
|
}
|
|
|
|
if (args->in.flags & NVGPU_GPU_ALLOC_VIDMEM_FLAG_VPR) {
|
|
nvgpu_warn(g,
|
|
"Allocating vidmem with FLAG_VPR is not yet supported");
|
|
return -EINVAL;
|
|
}
|
|
|
|
if (args->in.size & (SZ_4K - 1))
|
|
return -EINVAL;
|
|
|
|
if (!args->in.size)
|
|
return -EINVAL;
|
|
|
|
if (align & (align - 1))
|
|
return -EINVAL;
|
|
|
|
if (align > roundup_pow_of_two(args->in.size)) {
|
|
/* log this special case, buddy allocator detail */
|
|
nvgpu_warn(g,
|
|
"alignment larger than buffer size rounded up to power of 2 is not supported");
|
|
return -EINVAL;
|
|
}
|
|
|
|
fd = nvgpu_vidmem_export_linux(g, args->in.size);
|
|
if (fd < 0)
|
|
return fd;
|
|
|
|
args->out.dmabuf_fd = fd;
|
|
|
|
nvgpu_log_fn(g, "done, fd=%d", fd);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int nvgpu_gpu_get_memory_state(struct gk20a *g,
|
|
struct nvgpu_gpu_get_memory_state_args *args)
|
|
{
|
|
int err;
|
|
|
|
nvgpu_log_fn(g, " ");
|
|
|
|
if (args->reserved[0] || args->reserved[1] ||
|
|
args->reserved[2] || args->reserved[3])
|
|
return -EINVAL;
|
|
|
|
err = nvgpu_vidmem_get_space(g, &args->total_free_bytes);
|
|
|
|
nvgpu_log_fn(g, "done, err=%d, bytes=%lld", err, args->total_free_bytes);
|
|
|
|
return err;
|
|
}
|
|
#endif
|
|
|
|
static u32 nvgpu_gpu_convert_clk_domain(u32 clk_domain)
|
|
{
|
|
u32 domain = 0;
|
|
|
|
if (clk_domain == NVGPU_GPU_CLK_DOMAIN_MCLK)
|
|
domain = NVGPU_CLK_DOMAIN_MCLK;
|
|
else if (clk_domain == NVGPU_GPU_CLK_DOMAIN_GPCCLK)
|
|
domain = NVGPU_CLK_DOMAIN_GPCCLK;
|
|
else
|
|
domain = NVGPU_CLK_DOMAIN_MAX + 1;
|
|
|
|
return domain;
|
|
}
|
|
|
|
static int nvgpu_gpu_clk_get_vf_points(struct gk20a *g,
|
|
struct gk20a_ctrl_priv *priv,
|
|
struct nvgpu_gpu_clk_vf_points_args *args)
|
|
{
|
|
struct nvgpu_gpu_clk_vf_point clk_point;
|
|
struct nvgpu_gpu_clk_vf_point __user *entry;
|
|
struct nvgpu_clk_session *session = priv->clk_session;
|
|
u32 clk_domains = 0;
|
|
int err;
|
|
u16 last_mhz;
|
|
u16 *fpoints;
|
|
u32 i;
|
|
u32 max_points = 0;
|
|
u32 num_points = 0;
|
|
u16 min_mhz;
|
|
u16 max_mhz;
|
|
|
|
nvgpu_log_fn(g, " ");
|
|
|
|
if (!session || args->flags)
|
|
return -EINVAL;
|
|
|
|
clk_domains = nvgpu_clk_arb_get_arbiter_clk_domains(g);
|
|
args->num_entries = 0;
|
|
|
|
if (!nvgpu_clk_arb_is_valid_domain(g,
|
|
nvgpu_gpu_convert_clk_domain(args->clk_domain)))
|
|
return -EINVAL;
|
|
|
|
err = nvgpu_clk_arb_get_arbiter_clk_f_points(g,
|
|
nvgpu_gpu_convert_clk_domain(args->clk_domain),
|
|
&max_points, NULL);
|
|
if (err)
|
|
return err;
|
|
|
|
if (!args->max_entries) {
|
|
args->max_entries = max_points;
|
|
return 0;
|
|
}
|
|
|
|
if (args->max_entries < max_points)
|
|
return -EINVAL;
|
|
|
|
err = nvgpu_clk_arb_get_arbiter_clk_range(g,
|
|
nvgpu_gpu_convert_clk_domain(args->clk_domain),
|
|
&min_mhz, &max_mhz);
|
|
if (err)
|
|
return err;
|
|
|
|
fpoints = nvgpu_kcalloc(g, max_points, sizeof(u16));
|
|
if (!fpoints)
|
|
return -ENOMEM;
|
|
|
|
err = nvgpu_clk_arb_get_arbiter_clk_f_points(g,
|
|
nvgpu_gpu_convert_clk_domain(args->clk_domain),
|
|
&max_points, fpoints);
|
|
if (err)
|
|
goto fail;
|
|
|
|
entry = (struct nvgpu_gpu_clk_vf_point __user *)
|
|
(uintptr_t)args->clk_vf_point_entries;
|
|
|
|
last_mhz = 0;
|
|
num_points = 0;
|
|
for (i = 0; (i < max_points) && !err; i++) {
|
|
|
|
/* filter out duplicate frequencies */
|
|
if (fpoints[i] == last_mhz)
|
|
continue;
|
|
|
|
/* filter out out-of-range frequencies */
|
|
if ((fpoints[i] < min_mhz) || (fpoints[i] > max_mhz))
|
|
continue;
|
|
|
|
last_mhz = fpoints[i];
|
|
clk_point.freq_hz = MHZ_TO_HZ(fpoints[i]);
|
|
|
|
err = copy_to_user((void __user *)entry, &clk_point,
|
|
sizeof(clk_point));
|
|
|
|
num_points++;
|
|
entry++;
|
|
}
|
|
|
|
args->num_entries = num_points;
|
|
|
|
fail:
|
|
nvgpu_kfree(g, fpoints);
|
|
return err;
|
|
}
|
|
|
|
static int nvgpu_gpu_clk_get_range(struct gk20a *g,
|
|
struct gk20a_ctrl_priv *priv,
|
|
struct nvgpu_gpu_clk_range_args *args)
|
|
{
|
|
struct nvgpu_gpu_clk_range clk_range;
|
|
struct nvgpu_gpu_clk_range __user *entry;
|
|
struct nvgpu_clk_session *session = priv->clk_session;
|
|
|
|
u32 clk_domains = 0;
|
|
u32 num_domains;
|
|
u32 num_entries;
|
|
u32 i;
|
|
int bit;
|
|
int err;
|
|
u16 min_mhz, max_mhz;
|
|
|
|
nvgpu_log_fn(g, " ");
|
|
|
|
if (!session)
|
|
return -EINVAL;
|
|
|
|
clk_domains = nvgpu_clk_arb_get_arbiter_clk_domains(g);
|
|
num_domains = hweight_long(clk_domains);
|
|
|
|
if (!args->flags) {
|
|
if (!args->num_entries) {
|
|
args->num_entries = num_domains;
|
|
return 0;
|
|
}
|
|
|
|
if (args->num_entries < num_domains)
|
|
return -EINVAL;
|
|
|
|
args->num_entries = 0;
|
|
num_entries = num_domains;
|
|
|
|
} else {
|
|
if (args->flags != NVGPU_GPU_CLK_FLAG_SPECIFIC_DOMAINS)
|
|
return -EINVAL;
|
|
|
|
num_entries = args->num_entries;
|
|
if (num_entries > num_domains)
|
|
return -EINVAL;
|
|
}
|
|
|
|
entry = (struct nvgpu_gpu_clk_range __user *)
|
|
(uintptr_t)args->clk_range_entries;
|
|
|
|
for (i = 0; i < num_entries; i++, entry++) {
|
|
|
|
if (args->flags == NVGPU_GPU_CLK_FLAG_SPECIFIC_DOMAINS) {
|
|
if (copy_from_user(&clk_range, (void __user *)entry,
|
|
sizeof(clk_range)))
|
|
return -EFAULT;
|
|
} else {
|
|
bit = ffs(clk_domains) - 1;
|
|
clk_range.clk_domain = bit;
|
|
clk_domains &= ~BIT(bit);
|
|
}
|
|
|
|
clk_range.flags = 0;
|
|
err = nvgpu_clk_arb_get_arbiter_clk_range(g,
|
|
nvgpu_gpu_convert_clk_domain(clk_range.clk_domain),
|
|
&min_mhz, &max_mhz);
|
|
clk_range.min_hz = MHZ_TO_HZ(min_mhz);
|
|
clk_range.max_hz = MHZ_TO_HZ(max_mhz);
|
|
|
|
if (err)
|
|
return err;
|
|
|
|
err = copy_to_user(entry, &clk_range, sizeof(clk_range));
|
|
if (err)
|
|
return -EFAULT;
|
|
}
|
|
|
|
args->num_entries = num_entries;
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int nvgpu_gpu_clk_set_info(struct gk20a *g,
|
|
struct gk20a_ctrl_priv *priv,
|
|
struct nvgpu_gpu_clk_set_info_args *args)
|
|
{
|
|
struct nvgpu_gpu_clk_info clk_info;
|
|
struct nvgpu_gpu_clk_info __user *entry;
|
|
struct nvgpu_clk_session *session = priv->clk_session;
|
|
|
|
int fd;
|
|
u32 clk_domains = 0;
|
|
u16 freq_mhz;
|
|
int i;
|
|
int ret;
|
|
|
|
nvgpu_log_fn(g, " ");
|
|
|
|
if (!session || args->flags)
|
|
return -EINVAL;
|
|
|
|
clk_domains = nvgpu_clk_arb_get_arbiter_clk_domains(g);
|
|
if (!clk_domains)
|
|
return -EINVAL;
|
|
|
|
entry = (struct nvgpu_gpu_clk_info __user *)
|
|
(uintptr_t)args->clk_info_entries;
|
|
|
|
for (i = 0; i < args->num_entries; i++, entry++) {
|
|
|
|
if (copy_from_user(&clk_info, entry, sizeof(clk_info)))
|
|
return -EFAULT;
|
|
|
|
if (!nvgpu_clk_arb_is_valid_domain(g,
|
|
nvgpu_gpu_convert_clk_domain(clk_info.clk_domain)))
|
|
return -EINVAL;
|
|
}
|
|
nvgpu_speculation_barrier();
|
|
|
|
entry = (struct nvgpu_gpu_clk_info __user *)
|
|
(uintptr_t)args->clk_info_entries;
|
|
|
|
ret = nvgpu_clk_arb_install_request_fd(g, session, &fd);
|
|
if (ret < 0)
|
|
return ret;
|
|
|
|
for (i = 0; i < args->num_entries; i++, entry++) {
|
|
|
|
if (copy_from_user(&clk_info, (void __user *)entry,
|
|
sizeof(clk_info)))
|
|
return -EFAULT;
|
|
freq_mhz = HZ_TO_MHZ(clk_info.freq_hz);
|
|
|
|
nvgpu_clk_arb_set_session_target_mhz(session, fd,
|
|
nvgpu_gpu_convert_clk_domain(clk_info.clk_domain), freq_mhz);
|
|
}
|
|
|
|
nvgpu_speculation_barrier();
|
|
ret = nvgpu_clk_arb_commit_request_fd(g, session, fd);
|
|
if (ret < 0)
|
|
return ret;
|
|
|
|
args->completion_fd = fd;
|
|
|
|
return ret;
|
|
}
|
|
|
|
static int nvgpu_gpu_clk_get_info(struct gk20a *g,
|
|
struct gk20a_ctrl_priv *priv,
|
|
struct nvgpu_gpu_clk_get_info_args *args)
|
|
{
|
|
struct nvgpu_gpu_clk_info clk_info;
|
|
struct nvgpu_gpu_clk_info __user *entry;
|
|
struct nvgpu_clk_session *session = priv->clk_session;
|
|
u32 clk_domains = 0;
|
|
u32 num_domains;
|
|
u32 num_entries;
|
|
u32 i;
|
|
u16 freq_mhz;
|
|
int err;
|
|
int bit;
|
|
|
|
nvgpu_log_fn(g, " ");
|
|
|
|
if (!session)
|
|
return -EINVAL;
|
|
|
|
clk_domains = nvgpu_clk_arb_get_arbiter_clk_domains(g);
|
|
num_domains = hweight_long(clk_domains);
|
|
|
|
if (!args->flags) {
|
|
if (!args->num_entries) {
|
|
args->num_entries = num_domains;
|
|
return 0;
|
|
}
|
|
|
|
if (args->num_entries < num_domains)
|
|
return -EINVAL;
|
|
|
|
args->num_entries = 0;
|
|
num_entries = num_domains;
|
|
|
|
} else {
|
|
if (args->flags != NVGPU_GPU_CLK_FLAG_SPECIFIC_DOMAINS)
|
|
return -EINVAL;
|
|
|
|
num_entries = args->num_entries;
|
|
if (num_entries > num_domains * 3)
|
|
return -EINVAL;
|
|
}
|
|
|
|
entry = (struct nvgpu_gpu_clk_info __user *)
|
|
(uintptr_t)args->clk_info_entries;
|
|
|
|
for (i = 0; i < num_entries; i++, entry++) {
|
|
|
|
if (args->flags == NVGPU_GPU_CLK_FLAG_SPECIFIC_DOMAINS) {
|
|
if (copy_from_user(&clk_info, (void __user *)entry,
|
|
sizeof(clk_info)))
|
|
return -EFAULT;
|
|
} else {
|
|
bit = ffs(clk_domains) - 1;
|
|
clk_info.clk_domain = bit;
|
|
clk_domains &= ~BIT(bit);
|
|
clk_info.clk_type = args->clk_type;
|
|
}
|
|
|
|
nvgpu_speculation_barrier();
|
|
switch (clk_info.clk_type) {
|
|
case NVGPU_GPU_CLK_TYPE_TARGET:
|
|
err = nvgpu_clk_arb_get_session_target_mhz(session,
|
|
nvgpu_gpu_convert_clk_domain(clk_info.clk_domain),
|
|
&freq_mhz);
|
|
break;
|
|
case NVGPU_GPU_CLK_TYPE_ACTUAL:
|
|
err = nvgpu_clk_arb_get_arbiter_actual_mhz(g,
|
|
nvgpu_gpu_convert_clk_domain(clk_info.clk_domain),
|
|
&freq_mhz);
|
|
break;
|
|
case NVGPU_GPU_CLK_TYPE_EFFECTIVE:
|
|
err = nvgpu_clk_arb_get_arbiter_effective_mhz(g,
|
|
nvgpu_gpu_convert_clk_domain(clk_info.clk_domain),
|
|
&freq_mhz);
|
|
break;
|
|
default:
|
|
freq_mhz = 0;
|
|
err = -EINVAL;
|
|
break;
|
|
}
|
|
if (err)
|
|
return err;
|
|
|
|
clk_info.flags = 0;
|
|
clk_info.freq_hz = MHZ_TO_HZ(freq_mhz);
|
|
|
|
err = copy_to_user((void __user *)entry, &clk_info,
|
|
sizeof(clk_info));
|
|
if (err)
|
|
return -EFAULT;
|
|
}
|
|
|
|
nvgpu_speculation_barrier();
|
|
args->num_entries = num_entries;
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int nvgpu_gpu_get_event_fd(struct gk20a *g,
|
|
struct gk20a_ctrl_priv *priv,
|
|
struct nvgpu_gpu_get_event_fd_args *args)
|
|
{
|
|
struct nvgpu_clk_session *session = priv->clk_session;
|
|
|
|
nvgpu_log_fn(g, " ");
|
|
|
|
if (!nvgpu_is_enabled(g, NVGPU_SUPPORT_DEVICE_EVENTS))
|
|
return -EINVAL;
|
|
|
|
if (!session)
|
|
return -EINVAL;
|
|
|
|
return nvgpu_clk_arb_install_event_fd(g, session, &args->event_fd,
|
|
args->flags);
|
|
}
|
|
|
|
static int nvgpu_gpu_get_voltage(struct gk20a *g,
|
|
struct nvgpu_gpu_get_voltage_args *args)
|
|
{
|
|
int err = -EINVAL;
|
|
|
|
nvgpu_log_fn(g, " ");
|
|
|
|
if (args->reserved)
|
|
return -EINVAL;
|
|
|
|
if (!nvgpu_is_enabled(g, NVGPU_SUPPORT_GET_VOLTAGE))
|
|
return -EINVAL;
|
|
|
|
err = gk20a_busy(g);
|
|
if (err)
|
|
return err;
|
|
|
|
nvgpu_speculation_barrier();
|
|
|
|
err = nvgpu_pmu_volt_get_curr_volt_ps35(g, &args->voltage);
|
|
if (err) {
|
|
return err;
|
|
}
|
|
|
|
gk20a_idle(g);
|
|
|
|
return err;
|
|
}
|
|
|
|
static int nvgpu_gpu_get_current(struct gk20a *g,
|
|
struct nvgpu_gpu_get_current_args *args)
|
|
{
|
|
int err;
|
|
|
|
nvgpu_log_fn(g, " ");
|
|
|
|
if (args->reserved[0] || args->reserved[1] || args->reserved[2])
|
|
return -EINVAL;
|
|
|
|
if (!nvgpu_is_enabled(g, NVGPU_SUPPORT_GET_CURRENT))
|
|
return -EINVAL;
|
|
|
|
err = gk20a_busy(g);
|
|
if (err)
|
|
return err;
|
|
|
|
err = pmgr_pwr_devices_get_current(g, &args->currnt);
|
|
|
|
gk20a_idle(g);
|
|
|
|
return err;
|
|
}
|
|
|
|
static int nvgpu_gpu_get_power(struct gk20a *g,
|
|
struct nvgpu_gpu_get_power_args *args)
|
|
{
|
|
int err;
|
|
|
|
nvgpu_log_fn(g, " ");
|
|
|
|
if (args->reserved[0] || args->reserved[1] || args->reserved[2])
|
|
return -EINVAL;
|
|
|
|
if (!nvgpu_is_enabled(g, NVGPU_SUPPORT_GET_POWER))
|
|
return -EINVAL;
|
|
|
|
err = gk20a_busy(g);
|
|
if (err)
|
|
return err;
|
|
|
|
err = pmgr_pwr_devices_get_power(g, &args->power);
|
|
|
|
gk20a_idle(g);
|
|
|
|
return err;
|
|
}
|
|
|
|
static int nvgpu_gpu_get_temperature(struct gk20a *g,
|
|
struct nvgpu_gpu_get_temperature_args *args)
|
|
{
|
|
int err;
|
|
u32 temp_f24_8;
|
|
|
|
nvgpu_log_fn(g, " ");
|
|
|
|
#ifdef CONFIG_NVGPU_SIM
|
|
if (nvgpu_is_enabled(g, NVGPU_IS_FMODEL)) {
|
|
return 0;
|
|
}
|
|
#endif
|
|
|
|
if (args->reserved[0] || args->reserved[1] || args->reserved[2])
|
|
return -EINVAL;
|
|
|
|
if (!nvgpu_is_enabled(g, NVGPU_SUPPORT_GET_TEMPERATURE))
|
|
return -EINVAL;
|
|
|
|
err = gk20a_busy(g);
|
|
if (err)
|
|
return err;
|
|
|
|
/*
|
|
* If PSTATE is enabled, temp value is taken from THERM_GET_STATUS.
|
|
* If PSTATE is disable, temp value is read from NV_THERM_I2CS_SENSOR_00
|
|
* register value.
|
|
*/
|
|
if (nvgpu_is_enabled(g, NVGPU_PMU_PSTATE)) {
|
|
err = nvgpu_pmu_therm_channel_get_curr_temp(g, &temp_f24_8);
|
|
if (err) {
|
|
nvgpu_err(g, "pmu therm channel get status failed");
|
|
return err;
|
|
}
|
|
} else {
|
|
if (!g->ops.therm.get_internal_sensor_curr_temp) {
|
|
nvgpu_err(g, "reading NV_THERM_I2CS_SENSOR_00 not enabled");
|
|
return -EINVAL;
|
|
}
|
|
|
|
g->ops.therm.get_internal_sensor_curr_temp(g, &temp_f24_8);
|
|
}
|
|
|
|
gk20a_idle(g);
|
|
|
|
args->temp_f24_8 = (s32)temp_f24_8;
|
|
|
|
return err;
|
|
}
|
|
|
|
static int nvgpu_gpu_set_therm_alert_limit(struct gk20a *g,
|
|
struct nvgpu_gpu_set_therm_alert_limit_args *args)
|
|
{
|
|
int err;
|
|
|
|
nvgpu_log_fn(g, " ");
|
|
|
|
if (args->reserved[0] || args->reserved[1] || args->reserved[2])
|
|
return -EINVAL;
|
|
|
|
if (!g->ops.therm.configure_therm_alert)
|
|
return -EINVAL;
|
|
|
|
err = gk20a_busy(g);
|
|
if (err)
|
|
return err;
|
|
|
|
err = g->ops.therm.configure_therm_alert(g, args->temp_f24_8);
|
|
|
|
gk20a_idle(g);
|
|
|
|
return err;
|
|
}
|
|
|
|
static int nvgpu_gpu_set_deterministic_ch_railgate(struct nvgpu_channel *ch,
|
|
u32 flags)
|
|
{
|
|
int err = 0;
|
|
bool allow;
|
|
bool disallow;
|
|
|
|
allow = flags &
|
|
NVGPU_GPU_SET_DETERMINISTIC_OPTS_FLAGS_ALLOW_RAILGATING;
|
|
|
|
disallow = flags &
|
|
NVGPU_GPU_SET_DETERMINISTIC_OPTS_FLAGS_DISALLOW_RAILGATING;
|
|
|
|
/* Can't be both at the same time */
|
|
if (allow && disallow)
|
|
return -EINVAL;
|
|
|
|
/* Nothing to do */
|
|
if (!allow && !disallow)
|
|
return 0;
|
|
|
|
/*
|
|
* Moving into explicit idle or back from it? A call that doesn't
|
|
* change the status is a no-op.
|
|
*/
|
|
if (!ch->deterministic_railgate_allowed &&
|
|
allow) {
|
|
gk20a_idle(ch->g);
|
|
} else if (ch->deterministic_railgate_allowed &&
|
|
!allow) {
|
|
err = gk20a_busy(ch->g);
|
|
if (err) {
|
|
nvgpu_warn(ch->g,
|
|
"cannot busy to restore deterministic ch");
|
|
return err;
|
|
}
|
|
}
|
|
ch->deterministic_railgate_allowed = allow;
|
|
|
|
return err;
|
|
}
|
|
|
|
#ifdef CONFIG_NVGPU_DETERMINISTIC_CHANNELS
|
|
static int nvgpu_gpu_set_deterministic_ch(struct nvgpu_channel *ch, u32 flags)
|
|
{
|
|
if (!ch->deterministic)
|
|
return -EINVAL;
|
|
|
|
return nvgpu_gpu_set_deterministic_ch_railgate(ch, flags);
|
|
}
|
|
|
|
static int nvgpu_gpu_set_deterministic_opts(struct gk20a *g,
|
|
struct nvgpu_gpu_set_deterministic_opts_args *args)
|
|
{
|
|
int __user *user_channels;
|
|
u32 i = 0;
|
|
int err = 0;
|
|
|
|
nvgpu_log_fn(g, " ");
|
|
|
|
user_channels = (int __user *)(uintptr_t)args->channels;
|
|
|
|
/* Upper limit; prevent holding deterministic_busy for long */
|
|
if (args->num_channels > g->fifo.num_channels) {
|
|
err = -EINVAL;
|
|
goto out;
|
|
}
|
|
|
|
/* Trivial sanity check first */
|
|
#if LINUX_VERSION_CODE >= KERNEL_VERSION(5, 0, 0)
|
|
if (!access_ok(user_channels,
|
|
args->num_channels * sizeof(int))) {
|
|
#else
|
|
if (!access_ok(VERIFY_READ, user_channels,
|
|
args->num_channels * sizeof(int))) {
|
|
#endif
|
|
err = -EFAULT;
|
|
goto out;
|
|
}
|
|
|
|
nvgpu_rwsem_down_read(&g->deterministic_busy);
|
|
|
|
/* note: we exit at the first failure */
|
|
for (; i < args->num_channels; i++) {
|
|
int ch_fd = 0;
|
|
struct nvgpu_channel *ch;
|
|
|
|
if (copy_from_user(&ch_fd, &user_channels[i], sizeof(int))) {
|
|
/* User raced with above access_ok */
|
|
err = -EFAULT;
|
|
break;
|
|
}
|
|
|
|
ch = nvgpu_channel_get_from_file(ch_fd);
|
|
if (!ch) {
|
|
err = -EINVAL;
|
|
break;
|
|
}
|
|
|
|
err = nvgpu_gpu_set_deterministic_ch(ch, args->flags);
|
|
|
|
nvgpu_channel_put(ch);
|
|
|
|
if (err)
|
|
break;
|
|
}
|
|
|
|
nvgpu_speculation_barrier();
|
|
nvgpu_rwsem_up_read(&g->deterministic_busy);
|
|
|
|
out:
|
|
args->num_channels = i;
|
|
return err;
|
|
}
|
|
#endif
|
|
|
|
static int nvgpu_gpu_ioctl_get_buffer_info(struct gk20a *g,
|
|
struct nvgpu_gpu_get_buffer_info_args *args)
|
|
{
|
|
u64 user_metadata_addr = args->in.metadata_addr;
|
|
u32 in_metadata_size = args->in.metadata_size;
|
|
struct gk20a_dmabuf_priv *priv = NULL;
|
|
s32 dmabuf_fd = args->in.dmabuf_fd;
|
|
struct dma_buf *dmabuf;
|
|
int err = 0;
|
|
|
|
nvgpu_log_fn(g, " ");
|
|
|
|
if (!nvgpu_is_enabled(g, NVGPU_SUPPORT_BUFFER_METADATA)) {
|
|
nvgpu_err(g, "Buffer metadata not supported");
|
|
return -EINVAL;
|
|
}
|
|
|
|
args->out.metadata_size = 0;
|
|
args->out.flags = 0;
|
|
args->out.size = 0;
|
|
|
|
dmabuf = dma_buf_get(dmabuf_fd);
|
|
if (IS_ERR(dmabuf)) {
|
|
nvgpu_warn(g, "%s: fd %d is not a dmabuf",
|
|
__func__, dmabuf_fd);
|
|
return PTR_ERR(dmabuf);
|
|
}
|
|
|
|
args->out.size = dmabuf->size;
|
|
|
|
priv = gk20a_dma_buf_get_drvdata(dmabuf, dev_from_gk20a(g));
|
|
if (!priv) {
|
|
nvgpu_log_info(g, "Buffer metadata not allocated");
|
|
goto out;
|
|
}
|
|
|
|
nvgpu_mutex_acquire(&priv->lock);
|
|
|
|
if (in_metadata_size > 0) {
|
|
size_t write_size = priv->metadata_blob_size;
|
|
|
|
nvgpu_speculation_barrier();
|
|
|
|
if (write_size > in_metadata_size) {
|
|
write_size = in_metadata_size;
|
|
}
|
|
|
|
if (copy_to_user((void __user *)(uintptr_t)
|
|
user_metadata_addr,
|
|
priv->metadata_blob, write_size)) {
|
|
nvgpu_err(g, "metadata blob copy failed");
|
|
err = -EFAULT;
|
|
goto out_priv_unlock;
|
|
}
|
|
}
|
|
|
|
args->out.metadata_size = priv->metadata_blob_size;
|
|
|
|
if (priv->registered) {
|
|
args->out.flags |=
|
|
NVGPU_GPU_BUFFER_INFO_FLAGS_METADATA_REGISTERED;
|
|
}
|
|
|
|
#ifdef CONFIG_NVGPU_COMPRESSION
|
|
if (nvgpu_is_enabled(g, NVGPU_SUPPORT_COMPRESSION) &&
|
|
priv->comptags.enabled) {
|
|
args->out.flags |=
|
|
NVGPU_GPU_BUFFER_INFO_FLAGS_COMPTAGS_ALLOCATED;
|
|
}
|
|
#endif
|
|
|
|
if (priv->mutable_metadata) {
|
|
args->out.flags |=
|
|
NVGPU_GPU_BUFFER_INFO_FLAGS_MUTABLE_METADATA;
|
|
}
|
|
|
|
nvgpu_log_info(g, "buffer info: fd: %d, flags %llx, size %llu",
|
|
dmabuf_fd, args->out.flags, args->out.size);
|
|
|
|
out_priv_unlock:
|
|
nvgpu_mutex_release(&priv->lock);
|
|
out:
|
|
dma_buf_put(dmabuf);
|
|
return err;
|
|
}
|
|
|
|
#ifdef CONFIG_NVGPU_COMPRESSION
|
|
static int nvgpu_handle_comptags_control(struct gk20a *g,
|
|
struct dma_buf *dmabuf,
|
|
struct gk20a_dmabuf_priv *priv,
|
|
u8 comptags_alloc_control)
|
|
{
|
|
struct nvgpu_os_buffer os_buf = {0};
|
|
int err = 0;
|
|
|
|
if (!nvgpu_is_enabled(g, NVGPU_SUPPORT_COMPRESSION)) {
|
|
if (comptags_alloc_control == NVGPU_GPU_COMPTAGS_ALLOC_REQUIRED) {
|
|
nvgpu_err(g, "Comptags allocation (required) failed. Compression disabled.");
|
|
return -EINVAL;
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
if (comptags_alloc_control == NVGPU_GPU_COMPTAGS_ALLOC_NONE) {
|
|
if (priv->comptags.allocated) {
|
|
/*
|
|
* Just mark the comptags as disabled. Comptags will be
|
|
* freed on freeing the buffer.
|
|
*/
|
|
priv->comptags.enabled = false;
|
|
nvgpu_log_info(g, "Comptags disabled.");
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
/* Allocate the comptags if requested/required. */
|
|
if (priv->comptags.allocated) {
|
|
priv->comptags.enabled = priv->comptags.lines > 0;
|
|
if (priv->comptags.enabled) {
|
|
nvgpu_log_info(g, "Comptags enabled.");
|
|
return 0;
|
|
} else {
|
|
if (comptags_alloc_control ==
|
|
NVGPU_GPU_COMPTAGS_ALLOC_REQUIRED) {
|
|
nvgpu_err(g,
|
|
"Previous allocation has failed, could not enable comptags (required)");
|
|
return -ENOMEM;
|
|
} else {
|
|
nvgpu_log_info(g,
|
|
"Previous allocation has failed, could not enable comptags (requested)");
|
|
return 0;
|
|
}
|
|
}
|
|
}
|
|
|
|
os_buf.dmabuf = dmabuf;
|
|
os_buf.dev = dev_from_gk20a(g);
|
|
|
|
err = gk20a_alloc_comptags(g, &os_buf, &g->cbc->comp_tags);
|
|
if (err != 0) {
|
|
if (comptags_alloc_control ==
|
|
NVGPU_GPU_COMPTAGS_ALLOC_REQUIRED) {
|
|
nvgpu_err(g, "Comptags allocation (required) failed (%d)",
|
|
err);
|
|
} else {
|
|
nvgpu_err(g, "Comptags allocation (requested) failed (%d)",
|
|
err);
|
|
err = 0;
|
|
}
|
|
}
|
|
|
|
return err;
|
|
}
|
|
#endif
|
|
|
|
static int nvgpu_gpu_ioctl_register_buffer(struct gk20a *g,
|
|
struct nvgpu_gpu_register_buffer_args *args)
|
|
{
|
|
struct gk20a_dmabuf_priv *priv = NULL;
|
|
bool mutable_metadata = false;
|
|
bool modify_metadata = false;
|
|
struct dma_buf *dmabuf;
|
|
u8 *blob_copy = NULL;
|
|
int err = 0;
|
|
|
|
nvgpu_log_fn(g, " ");
|
|
|
|
if (!nvgpu_is_enabled(g, NVGPU_SUPPORT_BUFFER_METADATA)) {
|
|
nvgpu_err(g, "Buffer metadata not supported");
|
|
return -EINVAL;
|
|
}
|
|
|
|
if (args->metadata_size > NVGPU_GPU_REGISTER_BUFFER_METADATA_MAX_SIZE) {
|
|
nvgpu_err(g, "Invalid metadata blob size");
|
|
return -EINVAL;
|
|
}
|
|
|
|
if (args->comptags_alloc_control > NVGPU_GPU_COMPTAGS_ALLOC_REQUIRED) {
|
|
nvgpu_err(g, "Invalid comptags_alloc_control");
|
|
return -EINVAL;
|
|
}
|
|
|
|
nvgpu_log_info(g, "dmabuf_fd: %d, comptags control: %u, metadata size: %u, flags: %u",
|
|
args->dmabuf_fd, args->comptags_alloc_control,
|
|
args->metadata_size, args->flags);
|
|
|
|
mutable_metadata = (args->flags & NVGPU_GPU_REGISTER_BUFFER_FLAGS_MUTABLE) != 0;
|
|
modify_metadata = (args->flags & NVGPU_GPU_REGISTER_BUFFER_FLAGS_MODIFY) != 0;
|
|
|
|
dmabuf = dma_buf_get(args->dmabuf_fd);
|
|
if (IS_ERR(dmabuf)) {
|
|
nvgpu_warn(g, "%s: fd %d is not a dmabuf",
|
|
__func__, args->dmabuf_fd);
|
|
return PTR_ERR(dmabuf);
|
|
}
|
|
|
|
/*
|
|
* Allocate or get the buffer metadata state.
|
|
*/
|
|
err = gk20a_dmabuf_alloc_or_get_drvdata(
|
|
dmabuf, dev_from_gk20a(g), &priv);
|
|
if (err != 0) {
|
|
nvgpu_err(g, "Error allocating buffer metadata %d", err);
|
|
goto out;
|
|
}
|
|
|
|
nvgpu_mutex_acquire(&priv->lock);
|
|
|
|
/* Check for valid buffer metadata re-registration */
|
|
if (priv->registered) {
|
|
if (!modify_metadata) {
|
|
nvgpu_err(g, "attempt to modify buffer metadata without NVGPU_GPU_REGISTER_BUFFER_FLAGS_MODIFY");
|
|
err = -EINVAL;
|
|
goto out_priv_unlock;
|
|
} else if (!priv->mutable_metadata) {
|
|
nvgpu_err(g, "attempt to redefine immutable metadata");
|
|
err = -EINVAL;
|
|
goto out_priv_unlock;
|
|
}
|
|
}
|
|
|
|
/* Allocate memory for the metadata blob */
|
|
blob_copy = nvgpu_kzalloc(g, args->metadata_size);
|
|
if (!blob_copy) {
|
|
nvgpu_err(g, "Error allocating memory for blob");
|
|
err = -ENOMEM;
|
|
goto out_priv_unlock;
|
|
}
|
|
|
|
/* Copy the metadata blob */
|
|
if (copy_from_user(blob_copy,
|
|
(void __user *) args->metadata_addr,
|
|
args->metadata_size)) {
|
|
err = -EFAULT;
|
|
nvgpu_err(g, "Error copying buffer metadata blob");
|
|
goto out_priv_unlock;
|
|
}
|
|
|
|
#ifdef CONFIG_NVGPU_COMPRESSION
|
|
/* Comptags allocation */
|
|
err = nvgpu_handle_comptags_control(g, dmabuf, priv,
|
|
args->comptags_alloc_control);
|
|
if (err != 0) {
|
|
nvgpu_err(g, "Comptags alloc control failed %d", err);
|
|
goto out_priv_unlock;
|
|
}
|
|
#endif
|
|
|
|
/* All done, update metadata blob */
|
|
nvgpu_kfree(g, priv->metadata_blob);
|
|
|
|
priv->metadata_blob = blob_copy;
|
|
priv->metadata_blob_size = args->metadata_size;
|
|
blob_copy = NULL;
|
|
|
|
/* Mark registered and update mutability */
|
|
priv->registered = true;
|
|
priv->mutable_metadata = mutable_metadata;
|
|
|
|
/* Output variables */
|
|
args->flags = 0;
|
|
|
|
#ifdef CONFIG_NVGPU_COMPRESSION
|
|
if (nvgpu_is_enabled(g, NVGPU_SUPPORT_COMPRESSION) &&
|
|
priv->comptags.enabled) {
|
|
args->flags |=
|
|
NVGPU_GPU_REGISTER_BUFFER_FLAGS_COMPTAGS_ALLOCATED;
|
|
}
|
|
#endif
|
|
|
|
nvgpu_log_info(g, "buffer registered: mutable: %s, metadata size: %u, flags: 0x%8x",
|
|
priv->mutable_metadata ? "yes" : "no", priv->metadata_blob_size,
|
|
args->flags);
|
|
|
|
out_priv_unlock:
|
|
nvgpu_mutex_release(&priv->lock);
|
|
out:
|
|
dma_buf_put(dmabuf);
|
|
nvgpu_kfree(g, blob_copy);
|
|
|
|
return err;
|
|
}
|
|
|
|
long gk20a_ctrl_dev_ioctl(struct file *filp, unsigned int cmd, unsigned long arg)
|
|
{
|
|
struct gk20a_ctrl_priv *priv = filp->private_data;
|
|
struct gk20a *g = priv->g;
|
|
u8 buf[NVGPU_GPU_IOCTL_MAX_ARG_SIZE];
|
|
u32 gpu_instance_id, gr_instance_id;
|
|
struct nvgpu_gr_config *gr_config;
|
|
#ifdef CONFIG_NVGPU_GRAPHICS
|
|
struct nvgpu_gpu_zcull_get_ctx_size_args *get_ctx_size_args;
|
|
struct nvgpu_gpu_zcull_get_info_args *get_info_args;
|
|
struct nvgpu_gr_zcull_info *zcull_info;
|
|
struct nvgpu_gr_zcull *gr_zcull = nvgpu_gr_get_zcull_ptr(g);
|
|
struct nvgpu_gr_zbc *gr_zbc = nvgpu_gr_get_zbc_ptr(g);
|
|
struct nvgpu_gr_zbc_entry *zbc_val;
|
|
struct nvgpu_gr_zbc_query_params *zbc_tbl;
|
|
struct nvgpu_gpu_zbc_set_table_args *set_table_args;
|
|
struct nvgpu_gpu_zbc_query_table_args *query_table_args;
|
|
u32 i;
|
|
#endif /* CONFIG_NVGPU_GRAPHICS */
|
|
int err = 0;
|
|
|
|
nvgpu_log_fn(g, "start %d", _IOC_NR(cmd));
|
|
|
|
if ((_IOC_TYPE(cmd) != NVGPU_GPU_IOCTL_MAGIC) ||
|
|
(_IOC_NR(cmd) == 0) ||
|
|
(_IOC_NR(cmd) > NVGPU_GPU_IOCTL_LAST) ||
|
|
(_IOC_SIZE(cmd) > NVGPU_GPU_IOCTL_MAX_ARG_SIZE))
|
|
return -EINVAL;
|
|
|
|
(void) memset(buf, 0, sizeof(buf));
|
|
if (_IOC_DIR(cmd) & _IOC_WRITE) {
|
|
if (copy_from_user(buf, (void __user *)arg, _IOC_SIZE(cmd)))
|
|
return -EFAULT;
|
|
}
|
|
|
|
if (!g->sw_ready) {
|
|
err = gk20a_busy(g);
|
|
if (err)
|
|
return err;
|
|
|
|
gk20a_idle(g);
|
|
}
|
|
|
|
gpu_instance_id = nvgpu_get_gpu_instance_id_from_cdev(g, priv->cdev);
|
|
nvgpu_assert(gpu_instance_id < g->mig.num_gpu_instances);
|
|
|
|
gr_instance_id = nvgpu_grmgr_get_gr_instance_id(g, gpu_instance_id);
|
|
nvgpu_assert(gr_instance_id < g->num_gr_instances);
|
|
|
|
gr_config = nvgpu_gr_get_gpu_instance_config_ptr(g, gpu_instance_id);
|
|
|
|
nvgpu_speculation_barrier();
|
|
switch (cmd) {
|
|
#ifdef CONFIG_NVGPU_GRAPHICS
|
|
case NVGPU_GPU_IOCTL_ZCULL_GET_CTX_SIZE:
|
|
if (gr_zcull == NULL)
|
|
return -ENODEV;
|
|
|
|
get_ctx_size_args = (struct nvgpu_gpu_zcull_get_ctx_size_args *)buf;
|
|
|
|
get_ctx_size_args->size = nvgpu_gr_get_ctxsw_zcull_size(g, gr_zcull);
|
|
|
|
break;
|
|
case NVGPU_GPU_IOCTL_ZCULL_GET_INFO:
|
|
if (gr_zcull == NULL)
|
|
return -ENODEV;
|
|
|
|
get_info_args = (struct nvgpu_gpu_zcull_get_info_args *)buf;
|
|
|
|
(void) memset(get_info_args, 0,
|
|
sizeof(struct nvgpu_gpu_zcull_get_info_args));
|
|
|
|
zcull_info = nvgpu_kzalloc(g, sizeof(*zcull_info));
|
|
if (zcull_info == NULL)
|
|
return -ENOMEM;
|
|
|
|
err = g->ops.gr.zcull.get_zcull_info(g, gr_config,
|
|
gr_zcull, zcull_info);
|
|
if (err) {
|
|
nvgpu_kfree(g, zcull_info);
|
|
break;
|
|
}
|
|
|
|
get_info_args->width_align_pixels = zcull_info->width_align_pixels;
|
|
get_info_args->height_align_pixels = zcull_info->height_align_pixels;
|
|
get_info_args->pixel_squares_by_aliquots = zcull_info->pixel_squares_by_aliquots;
|
|
get_info_args->aliquot_total = zcull_info->aliquot_total;
|
|
get_info_args->region_byte_multiplier = zcull_info->region_byte_multiplier;
|
|
get_info_args->region_header_size = zcull_info->region_header_size;
|
|
get_info_args->subregion_header_size = zcull_info->subregion_header_size;
|
|
get_info_args->subregion_width_align_pixels = zcull_info->subregion_width_align_pixels;
|
|
get_info_args->subregion_height_align_pixels = zcull_info->subregion_height_align_pixels;
|
|
get_info_args->subregion_count = zcull_info->subregion_count;
|
|
|
|
nvgpu_kfree(g, zcull_info);
|
|
break;
|
|
case NVGPU_GPU_IOCTL_ZBC_SET_TABLE:
|
|
if (!nvgpu_is_enabled(g, NVGPU_SUPPORT_ZBC))
|
|
return -ENODEV;
|
|
|
|
set_table_args = (struct nvgpu_gpu_zbc_set_table_args *)buf;
|
|
|
|
zbc_val = nvgpu_gr_zbc_entry_alloc(g);
|
|
if (zbc_val == NULL)
|
|
return -ENOMEM;
|
|
|
|
nvgpu_gr_zbc_set_entry_format(zbc_val, set_table_args->format);
|
|
nvgpu_gr_zbc_set_entry_type(zbc_val, set_table_args->type);
|
|
|
|
nvgpu_speculation_barrier();
|
|
switch (nvgpu_gr_zbc_get_entry_type(zbc_val)) {
|
|
case NVGPU_GR_ZBC_TYPE_COLOR:
|
|
for (i = 0U; i < NVGPU_GR_ZBC_COLOR_VALUE_SIZE; i++) {
|
|
nvgpu_gr_zbc_set_entry_color_ds(zbc_val, i,
|
|
set_table_args->color_ds[i]);
|
|
nvgpu_gr_zbc_set_entry_color_l2(zbc_val, i,
|
|
set_table_args->color_l2[i]);
|
|
}
|
|
break;
|
|
case NVGPU_GR_ZBC_TYPE_DEPTH:
|
|
nvgpu_gr_zbc_set_entry_depth(zbc_val,
|
|
set_table_args->depth);
|
|
break;
|
|
case NVGPU_GR_ZBC_TYPE_STENCIL:
|
|
nvgpu_gr_zbc_set_entry_stencil(zbc_val,
|
|
set_table_args->stencil);
|
|
break;
|
|
default:
|
|
err = -EINVAL;
|
|
}
|
|
|
|
if (!err) {
|
|
err = gk20a_busy(g);
|
|
if (!err) {
|
|
err = g->ops.gr.zbc.set_table(g, gr_zbc,
|
|
zbc_val);
|
|
gk20a_idle(g);
|
|
}
|
|
}
|
|
|
|
if (zbc_val)
|
|
nvgpu_gr_zbc_entry_free(g, zbc_val);
|
|
break;
|
|
case NVGPU_GPU_IOCTL_ZBC_QUERY_TABLE:
|
|
if (!nvgpu_is_enabled(g, NVGPU_SUPPORT_ZBC))
|
|
return -ENODEV;
|
|
|
|
query_table_args = (struct nvgpu_gpu_zbc_query_table_args *)buf;
|
|
|
|
zbc_tbl = nvgpu_kzalloc(g, sizeof(struct nvgpu_gr_zbc_query_params));
|
|
if (zbc_tbl == NULL)
|
|
return -ENOMEM;
|
|
|
|
zbc_tbl->type = query_table_args->type;
|
|
zbc_tbl->index_size = query_table_args->index_size;
|
|
|
|
err = g->ops.gr.zbc.query_table(g, gr_zbc, zbc_tbl);
|
|
|
|
if (!err) {
|
|
switch (zbc_tbl->type) {
|
|
case NVGPU_GR_ZBC_TYPE_COLOR:
|
|
for (i = 0U; i < NVGPU_GR_ZBC_COLOR_VALUE_SIZE; i++) {
|
|
query_table_args->color_ds[i] = zbc_tbl->color_ds[i];
|
|
query_table_args->color_l2[i] = zbc_tbl->color_l2[i];
|
|
}
|
|
break;
|
|
case NVGPU_GR_ZBC_TYPE_DEPTH:
|
|
query_table_args->depth = zbc_tbl->depth;
|
|
break;
|
|
case NVGPU_GR_ZBC_TYPE_STENCIL:
|
|
query_table_args->stencil = zbc_tbl->stencil;
|
|
break;
|
|
case NVGPU_GR_ZBC_TYPE_INVALID:
|
|
query_table_args->index_size = zbc_tbl->index_size;
|
|
break;
|
|
default:
|
|
err = -EINVAL;
|
|
}
|
|
if (!err) {
|
|
query_table_args->format = zbc_tbl->format;
|
|
query_table_args->ref_cnt = zbc_tbl->ref_cnt;
|
|
}
|
|
}
|
|
|
|
if (zbc_tbl)
|
|
nvgpu_kfree(g, zbc_tbl);
|
|
break;
|
|
#endif /* CONFIG_NVGPU_GRAPHICS */
|
|
case NVGPU_GPU_IOCTL_GET_CHARACTERISTICS:
|
|
err = gk20a_ctrl_ioctl_gpu_characteristics(g, gpu_instance_id, gr_config,
|
|
(struct nvgpu_gpu_get_characteristics *)buf);
|
|
break;
|
|
case NVGPU_GPU_IOCTL_PREPARE_COMPRESSIBLE_READ:
|
|
err = gk20a_ctrl_prepare_compressible_read(g,
|
|
(struct nvgpu_gpu_prepare_compressible_read_args *)buf);
|
|
break;
|
|
case NVGPU_GPU_IOCTL_MARK_COMPRESSIBLE_WRITE:
|
|
err = gk20a_ctrl_mark_compressible_write(g,
|
|
(struct nvgpu_gpu_mark_compressible_write_args *)buf);
|
|
break;
|
|
case NVGPU_GPU_IOCTL_ALLOC_AS:
|
|
err = gk20a_ctrl_alloc_as(g,
|
|
(struct nvgpu_alloc_as_args *)buf);
|
|
break;
|
|
case NVGPU_GPU_IOCTL_OPEN_TSG:
|
|
err = gk20a_ctrl_open_tsg(g, priv->cdev,
|
|
(struct nvgpu_gpu_open_tsg_args *)buf);
|
|
break;
|
|
case NVGPU_GPU_IOCTL_GET_TPC_MASKS:
|
|
err = gk20a_ctrl_get_tpc_masks(g, gr_config,
|
|
(struct nvgpu_gpu_get_tpc_masks_args *)buf);
|
|
break;
|
|
case NVGPU_GPU_IOCTL_GET_FBP_L2_MASKS:
|
|
err = gk20a_ctrl_get_fbp_l2_masks(g, gpu_instance_id,
|
|
(struct nvgpu_gpu_get_fbp_l2_masks_args *)buf);
|
|
break;
|
|
case NVGPU_GPU_IOCTL_OPEN_CHANNEL:
|
|
/* this arg type here, but ..gpu_open_channel_args in nvgpu.h
|
|
* for consistency - they are the same */
|
|
err = gk20a_channel_open_ioctl(g, priv->cdev,
|
|
(struct nvgpu_channel_open_args *)buf);
|
|
break;
|
|
case NVGPU_GPU_IOCTL_FLUSH_L2:
|
|
err = nvgpu_gpu_ioctl_l2_fb_ops(g,
|
|
(struct nvgpu_gpu_l2_fb_args *)buf);
|
|
break;
|
|
|
|
case NVGPU_GPU_IOCTL_SET_MMUDEBUG_MODE:
|
|
err = nvgpu_gpu_ioctl_set_mmu_debug_mode(g,
|
|
(struct nvgpu_gpu_mmu_debug_mode_args *)buf);
|
|
break;
|
|
|
|
case NVGPU_GPU_IOCTL_SET_SM_DEBUG_MODE:
|
|
err = nvgpu_pg_elpg_protected_call(g,
|
|
nvgpu_gpu_ioctl_set_debug_mode(g,
|
|
(struct nvgpu_gpu_sm_debug_mode_args *)buf,
|
|
gr_instance_id));
|
|
break;
|
|
|
|
case NVGPU_GPU_IOCTL_TRIGGER_SUSPEND:
|
|
err = nvgpu_pg_elpg_protected_call(g,
|
|
nvgpu_gpu_ioctl_trigger_suspend(g, gr_instance_id));
|
|
break;
|
|
|
|
case NVGPU_GPU_IOCTL_WAIT_FOR_PAUSE:
|
|
err = nvgpu_pg_elpg_protected_call(g,
|
|
nvgpu_gpu_ioctl_wait_for_pause(g,
|
|
(struct nvgpu_gpu_wait_pause_args *)buf,
|
|
gr_instance_id));
|
|
break;
|
|
|
|
case NVGPU_GPU_IOCTL_RESUME_FROM_PAUSE:
|
|
err = nvgpu_pg_elpg_protected_call(g,
|
|
nvgpu_gpu_ioctl_resume_from_pause(g, gr_instance_id));
|
|
break;
|
|
|
|
case NVGPU_GPU_IOCTL_CLEAR_SM_ERRORS:
|
|
err = nvgpu_pg_elpg_protected_call(g,
|
|
nvgpu_gpu_ioctl_clear_sm_errors(g, gr_instance_id));
|
|
break;
|
|
|
|
case NVGPU_GPU_IOCTL_GET_TPC_EXCEPTION_EN_STATUS:
|
|
err = nvgpu_gpu_ioctl_has_any_exception(g,
|
|
(struct nvgpu_gpu_tpc_exception_en_status_args *)buf);
|
|
break;
|
|
|
|
case NVGPU_GPU_IOCTL_NUM_VSMS:
|
|
err = gk20a_ctrl_get_num_vsms(g, gr_config,
|
|
(struct nvgpu_gpu_num_vsms *)buf);
|
|
break;
|
|
case NVGPU_GPU_IOCTL_VSMS_MAPPING:
|
|
err = gk20a_ctrl_vsm_mapping(g, gr_config,
|
|
(struct nvgpu_gpu_vsms_mapping *)buf);
|
|
break;
|
|
|
|
case NVGPU_GPU_IOCTL_GET_CPU_TIME_CORRELATION_INFO:
|
|
err = nvgpu_gpu_get_cpu_time_correlation_info(g,
|
|
(struct nvgpu_gpu_get_cpu_time_correlation_info_args *)buf);
|
|
break;
|
|
|
|
case NVGPU_GPU_IOCTL_GET_GPU_TIME:
|
|
err = nvgpu_gpu_get_gpu_time(g,
|
|
(struct nvgpu_gpu_get_gpu_time_args *)buf);
|
|
break;
|
|
|
|
case NVGPU_GPU_IOCTL_GET_ENGINE_INFO:
|
|
if (nvgpu_is_enabled(g, NVGPU_SUPPORT_MIG) &&
|
|
((gpu_instance_id != 0U) ||
|
|
(!nvgpu_grmgr_is_multi_gr_enabled(g)))) {
|
|
err = nvgpu_gpu_get_gpu_instance_engine_info(g, gpu_instance_id,
|
|
(struct nvgpu_gpu_get_engine_info_args *)buf);
|
|
} else {
|
|
err = nvgpu_gpu_get_engine_info(g,
|
|
(struct nvgpu_gpu_get_engine_info_args *)buf);
|
|
}
|
|
break;
|
|
|
|
#ifdef CONFIG_NVGPU_DGPU
|
|
case NVGPU_GPU_IOCTL_ALLOC_VIDMEM:
|
|
err = nvgpu_gpu_alloc_vidmem(g,
|
|
(struct nvgpu_gpu_alloc_vidmem_args *)buf);
|
|
break;
|
|
|
|
case NVGPU_GPU_IOCTL_GET_MEMORY_STATE:
|
|
err = nvgpu_gpu_get_memory_state(g,
|
|
(struct nvgpu_gpu_get_memory_state_args *)buf);
|
|
break;
|
|
#endif
|
|
|
|
case NVGPU_GPU_IOCTL_CLK_GET_RANGE:
|
|
err = nvgpu_gpu_clk_get_range(g, priv,
|
|
(struct nvgpu_gpu_clk_range_args *)buf);
|
|
break;
|
|
|
|
case NVGPU_GPU_IOCTL_CLK_GET_VF_POINTS:
|
|
err = nvgpu_gpu_clk_get_vf_points(g, priv,
|
|
(struct nvgpu_gpu_clk_vf_points_args *)buf);
|
|
break;
|
|
|
|
case NVGPU_GPU_IOCTL_CLK_SET_INFO:
|
|
err = nvgpu_gpu_clk_set_info(g, priv,
|
|
(struct nvgpu_gpu_clk_set_info_args *)buf);
|
|
break;
|
|
|
|
case NVGPU_GPU_IOCTL_CLK_GET_INFO:
|
|
err = nvgpu_gpu_clk_get_info(g, priv,
|
|
(struct nvgpu_gpu_clk_get_info_args *)buf);
|
|
break;
|
|
|
|
case NVGPU_GPU_IOCTL_GET_EVENT_FD:
|
|
err = nvgpu_gpu_get_event_fd(g, priv,
|
|
(struct nvgpu_gpu_get_event_fd_args *)buf);
|
|
break;
|
|
|
|
case NVGPU_GPU_IOCTL_GET_VOLTAGE:
|
|
err = nvgpu_gpu_get_voltage(g,
|
|
(struct nvgpu_gpu_get_voltage_args *)buf);
|
|
break;
|
|
|
|
case NVGPU_GPU_IOCTL_GET_CURRENT:
|
|
err = nvgpu_gpu_get_current(g,
|
|
(struct nvgpu_gpu_get_current_args *)buf);
|
|
break;
|
|
|
|
case NVGPU_GPU_IOCTL_GET_POWER:
|
|
err = nvgpu_gpu_get_power(g,
|
|
(struct nvgpu_gpu_get_power_args *)buf);
|
|
break;
|
|
|
|
case NVGPU_GPU_IOCTL_GET_TEMPERATURE:
|
|
err = nvgpu_gpu_get_temperature(g,
|
|
(struct nvgpu_gpu_get_temperature_args *)buf);
|
|
break;
|
|
|
|
case NVGPU_GPU_IOCTL_SET_THERM_ALERT_LIMIT:
|
|
err = nvgpu_gpu_set_therm_alert_limit(g,
|
|
(struct nvgpu_gpu_set_therm_alert_limit_args *)buf);
|
|
break;
|
|
|
|
case NVGPU_GPU_IOCTL_SET_DETERMINISTIC_OPTS:
|
|
err = nvgpu_gpu_set_deterministic_opts(g,
|
|
(struct nvgpu_gpu_set_deterministic_opts_args *)buf);
|
|
break;
|
|
|
|
case NVGPU_GPU_IOCTL_REGISTER_BUFFER:
|
|
err = nvgpu_gpu_ioctl_register_buffer(g,
|
|
(struct nvgpu_gpu_register_buffer_args *)buf);
|
|
break;
|
|
|
|
case NVGPU_GPU_IOCTL_GET_BUFFER_INFO:
|
|
err = nvgpu_gpu_ioctl_get_buffer_info(g,
|
|
(struct nvgpu_gpu_get_buffer_info_args *)buf);
|
|
break;
|
|
|
|
default:
|
|
nvgpu_log_info(g, "unrecognized gpu ioctl cmd: 0x%x", cmd);
|
|
err = -ENOTTY;
|
|
break;
|
|
}
|
|
|
|
if ((err == 0) && (_IOC_DIR(cmd) & _IOC_READ))
|
|
err = copy_to_user((void __user *)arg, buf, _IOC_SIZE(cmd));
|
|
|
|
return err;
|
|
}
|
|
|
|
static void usermode_vma_close(struct vm_area_struct *vma)
|
|
{
|
|
struct gk20a_ctrl_priv *priv = vma->vm_private_data;
|
|
struct gk20a *g = priv->g;
|
|
struct nvgpu_os_linux *l = nvgpu_os_linux_from_gk20a(g);
|
|
|
|
nvgpu_mutex_acquire(&l->ctrl_privs_lock);
|
|
priv->usermode_vma.vma = NULL;
|
|
priv->usermode_vma.vma_mapped = false;
|
|
nvgpu_mutex_release(&l->ctrl_privs_lock);
|
|
}
|
|
|
|
struct vm_operations_struct usermode_vma_ops = {
|
|
/* no .open - we use VM_DONTCOPY and don't support fork */
|
|
.close = usermode_vma_close,
|
|
};
|
|
|
|
int gk20a_ctrl_dev_mmap(struct file *filp, struct vm_area_struct *vma)
|
|
{
|
|
struct gk20a_ctrl_priv *priv = filp->private_data;
|
|
struct gk20a *g = priv->g;
|
|
struct nvgpu_os_linux *l = nvgpu_os_linux_from_gk20a(g);
|
|
int err;
|
|
|
|
if (g->ops.usermode.base == NULL)
|
|
return -ENOSYS;
|
|
|
|
if (priv->usermode_vma.vma != NULL)
|
|
return -EBUSY;
|
|
|
|
if (vma->vm_end - vma->vm_start > SZ_64K)
|
|
return -EINVAL;
|
|
|
|
if (vma->vm_pgoff != 0UL)
|
|
return -EINVAL;
|
|
|
|
/* Sync with poweron/poweroff, and require valid regs */
|
|
err = gk20a_busy(g);
|
|
if (err) {
|
|
return err;
|
|
}
|
|
|
|
nvgpu_mutex_acquire(&l->ctrl_privs_lock);
|
|
|
|
vma->vm_flags |= VM_IO | VM_DONTCOPY | VM_DONTEXPAND | VM_NORESERVE |
|
|
VM_DONTDUMP | VM_PFNMAP;
|
|
vma->vm_ops = &usermode_vma_ops;
|
|
vma->vm_page_prot = pgprot_noncached(vma->vm_page_prot);
|
|
|
|
err = io_remap_pfn_range(vma, vma->vm_start,
|
|
g->usermode_regs_bus_addr >> PAGE_SHIFT,
|
|
vma->vm_end - vma->vm_start, vma->vm_page_prot);
|
|
if (!err) {
|
|
priv->usermode_vma.vma = vma;
|
|
vma->vm_private_data = priv;
|
|
priv->usermode_vma.vma_mapped = true;
|
|
}
|
|
nvgpu_mutex_release(&l->ctrl_privs_lock);
|
|
|
|
gk20a_idle(g);
|
|
|
|
return err;
|
|
}
|
|
|
|
static int alter_usermode_mapping(struct gk20a *g,
|
|
struct gk20a_ctrl_priv *priv,
|
|
bool poweroff)
|
|
{
|
|
struct vm_area_struct *vma = priv->usermode_vma.vma;
|
|
bool vma_mapped = priv->usermode_vma.vma_mapped;
|
|
int err = 0;
|
|
|
|
if (!vma) {
|
|
/* Nothing to do - no mmap called */
|
|
return 0;
|
|
}
|
|
|
|
/*
|
|
* This is a no-op for the below cases
|
|
* a) poweroff and !vma_mapped - > do nothing as no map exists
|
|
* b) !poweroff and vmap_mapped -> do nothing as already mapped
|
|
*/
|
|
if (poweroff != vma_mapped) {
|
|
return 0;
|
|
}
|
|
|
|
/*
|
|
* We use trylock due to lock inversion: we need to acquire
|
|
* mmap_lock while holding ctrl_privs_lock. usermode_vma_close
|
|
* does it in reverse order. Trylock is a way to avoid deadlock.
|
|
*/
|
|
#if LINUX_VERSION_CODE >= KERNEL_VERSION(5, 8, 0)
|
|
if (!mmap_write_trylock(vma->vm_mm)) {
|
|
#else
|
|
if (!down_write_trylock(&vma->vm_mm->mmap_sem)) {
|
|
#endif
|
|
return -EBUSY;
|
|
}
|
|
|
|
if (poweroff) {
|
|
#if LINUX_VERSION_CODE >= KERNEL_VERSION(4, 18, 0)
|
|
zap_vma_ptes(vma, vma->vm_start, vma->vm_end - vma->vm_start);
|
|
err = 0;
|
|
#else
|
|
err = zap_vma_ptes(vma, vma->vm_start,
|
|
vma->vm_end - vma->vm_start);
|
|
#endif
|
|
if (err == 0) {
|
|
priv->usermode_vma.vma_mapped = false;
|
|
} else {
|
|
nvgpu_err(g, "can't remove usermode mapping");
|
|
}
|
|
} else {
|
|
err = io_remap_pfn_range(vma, vma->vm_start,
|
|
g->usermode_regs_bus_addr >> PAGE_SHIFT,
|
|
vma->vm_end - vma->vm_start, vma->vm_page_prot);
|
|
if (err != 0) {
|
|
nvgpu_err(g, "can't restore usermode mapping");
|
|
} else {
|
|
priv->usermode_vma.vma_mapped = true;
|
|
}
|
|
}
|
|
|
|
#if LINUX_VERSION_CODE >= KERNEL_VERSION(5, 8, 0)
|
|
mmap_write_unlock(vma->vm_mm);
|
|
#else
|
|
up_write(&vma->vm_mm->mmap_sem);
|
|
#endif
|
|
|
|
return err;
|
|
}
|
|
|
|
static void alter_usermode_mappings(struct gk20a *g, bool poweroff)
|
|
{
|
|
struct gk20a_ctrl_priv *priv;
|
|
struct nvgpu_os_linux *l = nvgpu_os_linux_from_gk20a(g);
|
|
int err = 0;
|
|
|
|
do {
|
|
nvgpu_mutex_acquire(&l->ctrl_privs_lock);
|
|
nvgpu_list_for_each_entry(priv, &l->ctrl_privs,
|
|
gk20a_ctrl_priv, list) {
|
|
err = alter_usermode_mapping(g, priv, poweroff);
|
|
if (err != 0) {
|
|
break;
|
|
}
|
|
}
|
|
nvgpu_mutex_release(&l->ctrl_privs_lock);
|
|
|
|
if (err == -EBUSY) {
|
|
nvgpu_log_info(g, "ctrl_privs_lock lock contended. retry altering usermode mappings");
|
|
nvgpu_udelay(10);
|
|
} else if (err != 0) {
|
|
nvgpu_err(g, "can't alter usermode mapping. err = %d", err);
|
|
}
|
|
} while (err == -EBUSY);
|
|
}
|
|
|
|
void nvgpu_hide_usermode_for_poweroff(struct gk20a *g)
|
|
{
|
|
alter_usermode_mappings(g, true);
|
|
}
|
|
|
|
void nvgpu_restore_usermode_for_poweron(struct gk20a *g)
|
|
{
|
|
alter_usermode_mappings(g, false);
|
|
}
|