gpu: nvgpu: add cycle stats to debugger node

Add NVGPU_DBG_GPU_IOCTL_CYCLE_STATS to debugger node, to
install/uninstall a buffer for cycle stats.

Add NVGPU_DBG_GPU_IOCTL_CYCLE_STATS_SNAPSHOT to debugger
node, to attach/flush/detach a buffer for Mode-E streamout.

Those ioctls will apply to the first channel in the debug session.

Bug 2660206
Bug 200464613

Change-Id: I0b96d9a07c016690140292fa5886fda545697ee6
Signed-off-by: Thomas Fleury <tfleury@nvidia.com>
Reviewed-on: https://git-master.nvidia.com/r/2002060
(cherry picked from commit 90b0bf98ac)
Signed-off-by: Gagan Grover <ggrover@nvidia.com>
Reviewed-on: https://git-master.nvidia.com/r/c/linux-nvgpu/+/2092008
Reviewed-by: automaticguardword <automaticguardword@nvidia.com>
Reviewed-by: Phoenix Jung <pjung@nvidia.com>
Reviewed-by: mobile promotions <svcmobile_promotions@nvidia.com>
GVS: Gerrit_Virtual_Submit
Tested-by: Peter Daifuku <pdaifuku@nvidia.com>
Tested-by: mobile promotions <svcmobile_promotions@nvidia.com>
This commit is contained in:
Thomas Fleury
2018-12-21 10:04:27 +05:30
committed by mobile promotions
parent e878686302
commit 5ecc45b5e7
4 changed files with 138 additions and 13 deletions

View File

@@ -1,7 +1,7 @@
/* /*
* GK20A Graphics channel * GK20A Graphics channel
* *
* Copyright (c) 2011-2018, NVIDIA CORPORATION. All rights reserved. * Copyright (c) 2011-2020, NVIDIA CORPORATION. All rights reserved.
* *
* This program is free software; you can redistribute it and/or modify it * This program is free software; you can redistribute it and/or modify it
* under the terms and conditions of the GNU General Public License, * under the terms and conditions of the GNU General Public License,
@@ -128,8 +128,7 @@ void gk20a_channel_free_cycle_stats_buffer(struct channel_gk20a *ch)
nvgpu_mutex_release(&ch->cyclestate.cyclestate_buffer_mutex); nvgpu_mutex_release(&ch->cyclestate.cyclestate_buffer_mutex);
} }
static int gk20a_channel_cycle_stats(struct channel_gk20a *ch, int gk20a_channel_cycle_stats(struct channel_gk20a *ch, int dmabuf_fd)
struct nvgpu_cycle_stats_args *args)
{ {
struct dma_buf *dmabuf; struct dma_buf *dmabuf;
void *virtual_address; void *virtual_address;
@@ -139,10 +138,10 @@ static int gk20a_channel_cycle_stats(struct channel_gk20a *ch,
if (!nvgpu_is_enabled(ch->g, NVGPU_SUPPORT_CYCLE_STATS)) if (!nvgpu_is_enabled(ch->g, NVGPU_SUPPORT_CYCLE_STATS))
return -ENOSYS; return -ENOSYS;
if (args->dmabuf_fd && !priv->cyclestate_buffer_handler) { if (dmabuf_fd && !priv->cyclestate_buffer_handler) {
/* set up new cyclestats buffer */ /* set up new cyclestats buffer */
dmabuf = dma_buf_get(args->dmabuf_fd); dmabuf = dma_buf_get(dmabuf_fd);
if (IS_ERR(dmabuf)) if (IS_ERR(dmabuf))
return PTR_ERR(dmabuf); return PTR_ERR(dmabuf);
virtual_address = dma_buf_vmap(dmabuf); virtual_address = dma_buf_vmap(dmabuf);
@@ -154,12 +153,12 @@ static int gk20a_channel_cycle_stats(struct channel_gk20a *ch,
ch->cyclestate.cyclestate_buffer_size = dmabuf->size; ch->cyclestate.cyclestate_buffer_size = dmabuf->size;
return 0; return 0;
} else if (!args->dmabuf_fd && priv->cyclestate_buffer_handler) { } else if (!dmabuf_fd && priv->cyclestate_buffer_handler) {
gk20a_channel_free_cycle_stats_buffer(ch); gk20a_channel_free_cycle_stats_buffer(ch);
return 0; return 0;
} else if (!args->dmabuf_fd && !priv->cyclestate_buffer_handler) { } else if (!dmabuf_fd && !priv->cyclestate_buffer_handler) {
/* no requst from GL */ /* no request from GL */
return 0; return 0;
} else { } else {
@@ -168,7 +167,7 @@ static int gk20a_channel_cycle_stats(struct channel_gk20a *ch,
} }
} }
static int gk20a_flush_cycle_stats_snapshot(struct channel_gk20a *ch) int gk20a_flush_cycle_stats_snapshot(struct channel_gk20a *ch)
{ {
int ret; int ret;
@@ -182,7 +181,7 @@ static int gk20a_flush_cycle_stats_snapshot(struct channel_gk20a *ch)
return ret; return ret;
} }
static int gk20a_attach_cycle_stats_snapshot(struct channel_gk20a *ch, int gk20a_attach_cycle_stats_snapshot(struct channel_gk20a *ch,
u32 dmabuf_fd, u32 dmabuf_fd,
u32 perfmon_id_count, u32 perfmon_id_count,
u32 *perfmon_id_start) u32 *perfmon_id_start)
@@ -1280,7 +1279,7 @@ long gk20a_channel_ioctl(struct file *filp,
break; break;
} }
err = gk20a_channel_cycle_stats(ch, err = gk20a_channel_cycle_stats(ch,
(struct nvgpu_cycle_stats_args *)buf); ((struct nvgpu_cycle_stats_args *)buf)->dmabuf_fd);
gk20a_idle(ch->g); gk20a_idle(ch->g);
break; break;
#endif #endif

View File

@@ -1,5 +1,5 @@
/* /*
* Copyright (c) 2017, NVIDIA CORPORATION. All rights reserved. * Copyright (c) 2017-2020, NVIDIA CORPORATION. All rights reserved.
* *
* This program is free software; you can redistribute it and/or modify it * This program is free software; you can redistribute it and/or modify it
* under the terms and conditions of the GNU General Public License, * under the terms and conditions of the GNU General Public License,
@@ -36,9 +36,16 @@ long gk20a_channel_ioctl(struct file *filp,
int gk20a_channel_open_ioctl(struct gk20a *g, int gk20a_channel_open_ioctl(struct gk20a *g,
struct nvgpu_channel_open_args *args); struct nvgpu_channel_open_args *args);
int gk20a_channel_free_cycle_stats_snapshot(struct channel_gk20a *ch); int gk20a_channel_cycle_stats(struct channel_gk20a *ch, int dmabuf_fd);
void gk20a_channel_free_cycle_stats_buffer(struct channel_gk20a *ch); void gk20a_channel_free_cycle_stats_buffer(struct channel_gk20a *ch);
int gk20a_attach_cycle_stats_snapshot(struct channel_gk20a *ch,
u32 dmabuf_fd,
u32 perfmon_id_count,
u32 *perfmon_id_start);
int gk20a_flush_cycle_stats_snapshot(struct channel_gk20a *ch);
int gk20a_channel_free_cycle_stats_snapshot(struct channel_gk20a *ch);
extern const struct file_operations gk20a_channel_ops; extern const struct file_operations gk20a_channel_ops;
u32 nvgpu_get_common_runlist_level(u32 level); u32 nvgpu_get_common_runlist_level(u32 level);

View File

@@ -42,6 +42,7 @@
#include "os_linux.h" #include "os_linux.h"
#include "platform_gk20a.h" #include "platform_gk20a.h"
#include "ioctl_dbg.h" #include "ioctl_dbg.h"
#include "ioctl_channel.h"
#include "dmabuf_vidmem.h" #include "dmabuf_vidmem.h"
struct dbg_session_gk20a_linux { struct dbg_session_gk20a_linux {
@@ -1935,6 +1936,87 @@ static int nvgpu_dbg_gpu_set_sm_exception_type_mask(
return err; return err;
} }
#if defined(CONFIG_GK20A_CYCLE_STATS)
static int nvgpu_dbg_gpu_cycle_stats(struct dbg_session_gk20a *dbg_s,
struct nvgpu_dbg_gpu_cycle_stats_args *args)
{
struct channel_gk20a *ch = NULL;
int err;
ch = nvgpu_dbg_gpu_get_session_channel(dbg_s);
if (ch == NULL) {
return -EINVAL;
}
err = gk20a_busy(ch->g);
if (err != 0) {
return err;
}
err = gk20a_channel_cycle_stats(ch, args->dmabuf_fd);
gk20a_idle(ch->g);
return err;
}
static int nvgpu_dbg_gpu_cycle_stats_snapshot(struct dbg_session_gk20a *dbg_s,
struct nvgpu_dbg_gpu_cycle_stats_snapshot_args *args)
{
struct channel_gk20a *ch = NULL;
int err;
if (!args->dmabuf_fd) {
return -EINVAL;
}
nvgpu_speculation_barrier();
ch = nvgpu_dbg_gpu_get_session_channel(dbg_s);
if (ch == NULL) {
return -EINVAL;
}
/* is it allowed to handle calls for current GPU? */
if (!nvgpu_is_enabled(ch->g, NVGPU_SUPPORT_CYCLE_STATS_SNAPSHOT)) {
return -ENOSYS;
}
err = gk20a_busy(ch->g);
if (err != 0) {
return err;
}
/* handle the command (most frequent cases first) */
switch (args->cmd) {
case NVGPU_DBG_GPU_IOCTL_CYCLE_STATS_SNAPSHOT_CMD_FLUSH:
err = gk20a_flush_cycle_stats_snapshot(ch);
args->extra = 0;
break;
case NVGPU_DBG_GPU_IOCTL_CYCLE_STATS_SNAPSHOT_CMD_ATTACH:
err = gk20a_attach_cycle_stats_snapshot(ch,
args->dmabuf_fd,
args->extra,
&args->extra);
break;
case NVGPU_DBG_GPU_IOCTL_CYCLE_STATS_SNAPSHOT_CMD_DETACH:
err = gk20a_channel_free_cycle_stats_snapshot(ch);
args->extra = 0;
break;
default:
pr_err("cyclestats: unknown command %u\n", args->cmd);
err = -EINVAL;
break;
}
gk20a_idle(ch->g);
return err;
}
#endif
int gk20a_dbg_gpu_dev_open(struct inode *inode, struct file *filp) int gk20a_dbg_gpu_dev_open(struct inode *inode, struct file *filp)
{ {
struct nvgpu_os_linux *l = container_of(inode->i_cdev, struct nvgpu_os_linux *l = container_of(inode->i_cdev,
@@ -2096,6 +2178,18 @@ long gk20a_dbg_gpu_dev_ioctl(struct file *filp, unsigned int cmd,
(struct nvgpu_dbg_gpu_set_ctx_mmu_debug_mode_args *)buf); (struct nvgpu_dbg_gpu_set_ctx_mmu_debug_mode_args *)buf);
break; break;
#ifdef CONFIG_GK20A_CYCLE_STATS
case NVGPU_DBG_GPU_IOCTL_CYCLE_STATS:
err = nvgpu_dbg_gpu_cycle_stats(dbg_s,
(struct nvgpu_dbg_gpu_cycle_stats_args *)buf);
break;
case NVGPU_DBG_GPU_IOCTL_CYCLE_STATS_SNAPSHOT:
err = nvgpu_dbg_gpu_cycle_stats_snapshot(dbg_s,
(struct nvgpu_dbg_gpu_cycle_stats_snapshot_args *)buf);
break;
#endif
default: default:
nvgpu_err(g, nvgpu_err(g,
"unrecognized dbg gpu ioctl cmd: 0x%x", "unrecognized dbg gpu ioctl cmd: 0x%x",

View File

@@ -1416,6 +1416,31 @@ struct nvgpu_dbg_gpu_set_sm_exception_type_mask_args {
_IOW(NVGPU_DBG_GPU_IOCTL_MAGIC, 23, \ _IOW(NVGPU_DBG_GPU_IOCTL_MAGIC, 23, \
struct nvgpu_dbg_gpu_set_sm_exception_type_mask_args) struct nvgpu_dbg_gpu_set_sm_exception_type_mask_args)
struct nvgpu_dbg_gpu_cycle_stats_args {
__u32 dmabuf_fd;
__u32 reserved;
};
#define NVGPU_DBG_GPU_IOCTL_CYCLE_STATS \
_IOWR(NVGPU_DBG_GPU_IOCTL_MAGIC, 24, struct nvgpu_dbg_gpu_cycle_stats_args)
/* cycle stats snapshot buffer support for mode E */
struct nvgpu_dbg_gpu_cycle_stats_snapshot_args {
__u32 cmd; /* in: command to handle */
__u32 dmabuf_fd; /* in: dma buffer handler */
__u32 extra; /* in/out: extra payload e.g.*/
/* counter/start perfmon */
__u32 reserved;
};
/* valid commands to control cycle stats shared buffer */
#define NVGPU_DBG_GPU_IOCTL_CYCLE_STATS_SNAPSHOT_CMD_FLUSH 0
#define NVGPU_DBG_GPU_IOCTL_CYCLE_STATS_SNAPSHOT_CMD_ATTACH 1
#define NVGPU_DBG_GPU_IOCTL_CYCLE_STATS_SNAPSHOT_CMD_DETACH 2
#define NVGPU_DBG_GPU_IOCTL_CYCLE_STATS_SNAPSHOT \
_IOWR(NVGPU_DBG_GPU_IOCTL_MAGIC, 25, struct nvgpu_dbg_gpu_cycle_stats_snapshot_args)
/* MMU Debug Mode */ /* MMU Debug Mode */
#define NVGPU_DBG_GPU_CTX_MMU_DEBUG_MODE_DISABLED 0 #define NVGPU_DBG_GPU_CTX_MMU_DEBUG_MODE_DISABLED 0
#define NVGPU_DBG_GPU_CTX_MMU_DEBUG_MODE_ENABLED 1 #define NVGPU_DBG_GPU_CTX_MMU_DEBUG_MODE_ENABLED 1