mirror of
git://nv-tegra.nvidia.com/linux-nv-oot.git
synced 2025-12-22 17:25:35 +03:00
Use devm_of_icc_get() in rtcpu base and debug driver. CONFIG_TEGRA_T23X_GRHOST, CONFIG_TEGRA_ISOMGR, CONFIG_TEGRA_BWMGR are not enabled in K5.15 so cleanup tegra_camera_platform driver. Don't read memory-bw from rtcpu dt, set it to max during probe. Bug 3997304 Bug 4311411 Change-Id: Ib76b3338749de5d33e13cd518375a6b55dd17f5b Signed-off-by: Ankur Pawar <ankurp@nvidia.com> Reviewed-on: https://git-master.nvidia.com/r/c/linux-nv-oot/+/2899021 Reviewed-by: Jonathan Hunter <jonathanh@nvidia.com> Reviewed-by: Johnny Liu <johnliu@nvidia.com> Reviewed-by: Laxman Dewangan <ldewangan@nvidia.com> GVS: Gerrit_Virtual_Submit <buildbot_gerritrpt@nvidia.com>
1951 lines
49 KiB
C
1951 lines
49 KiB
C
// SPDX-License-Identifier: GPL-2.0
|
|
// Copyright (c) 2022-2023, NVIDIA CORPORATION & AFFILIATES. All rights reserved.
|
|
|
|
#include "soc/tegra/camrtc-dbg-messages.h"
|
|
|
|
#include <linux/debugfs.h>
|
|
#if IS_ENABLED(CONFIG_INTERCONNECT)
|
|
#include <linux/interconnect.h>
|
|
#include <dt-bindings/interconnect/tegra_icc_id.h>
|
|
#endif
|
|
#include <linux/io.h>
|
|
#include <linux/iommu.h>
|
|
#include <linux/kernel.h>
|
|
#include <linux/module.h>
|
|
#include <linux/of.h>
|
|
#include <linux/of_address.h>
|
|
#include <linux/of_platform.h>
|
|
#include <linux/of_reserved_mem.h>
|
|
#include <linux/platform_device.h>
|
|
#include <linux/pm_runtime.h>
|
|
#include <linux/sched.h>
|
|
#include <linux/sched/signal.h>
|
|
#include <linux/sched/clock.h>
|
|
#include <linux/slab.h>
|
|
#include <linux/string.h>
|
|
#include <linux/tegra-camera-rtcpu.h>
|
|
#include <soc/tegra/ivc_ext.h>
|
|
#include <linux/tegra-ivc-bus.h>
|
|
#include <linux/platform/tegra/common.h>
|
|
#include <soc/tegra/fuse.h>
|
|
|
|
#define CAMRTC_TEST_CAM_DEVICES 4
|
|
|
|
struct camrtc_test_device {
|
|
/* device handle */
|
|
struct device *dev;
|
|
/* device iova for the memory in context */
|
|
dma_addr_t dev_iova;
|
|
};
|
|
|
|
struct camrtc_test_mem {
|
|
/* access id in memory array */
|
|
u32 index;
|
|
/* occupied memory size */
|
|
size_t used;
|
|
/* total size */
|
|
size_t size;
|
|
/* CPU address */
|
|
void *ptr;
|
|
/* Physical base address, offsets valid for first page only */
|
|
phys_addr_t phys_addr;
|
|
/* base iova for device used for allocation */
|
|
dma_addr_t iova;
|
|
/* device index */
|
|
u32 dev_index;
|
|
/* metadata for all the devices using this memory */
|
|
struct camrtc_test_device devices[CAMRTC_TEST_CAM_DEVICES];
|
|
};
|
|
|
|
struct camrtc_falcon_coverage {
|
|
u8 id;
|
|
bool enabled;
|
|
struct camrtc_test_mem mem;
|
|
struct sg_table sgt;
|
|
u64 falc_iova;
|
|
struct tegra_ivc_channel *ch;
|
|
struct device *mem_dev;
|
|
struct device *falcon_dev;
|
|
};
|
|
|
|
struct camrtc_debug {
|
|
struct tegra_ivc_channel *channel;
|
|
struct mutex mutex;
|
|
struct dentry *root;
|
|
wait_queue_head_t waitq;
|
|
struct {
|
|
u32 completion_timeout;
|
|
u32 mods_case;
|
|
u32 mods_loops;
|
|
u32 mods_dma_channels;
|
|
char *test_case;
|
|
size_t test_case_size;
|
|
u32 test_timeout;
|
|
u32 test_bw;
|
|
} parameters;
|
|
struct camrtc_falcon_coverage vi_falc_coverage;
|
|
struct camrtc_falcon_coverage isp_falc_coverage;
|
|
struct icc_path *icc_path;
|
|
struct camrtc_test_mem mem[CAMRTC_DBG_NUM_MEM_TEST_MEM];
|
|
struct device *mem_devices[CAMRTC_TEST_CAM_DEVICES];
|
|
struct ast_regset {
|
|
struct debugfs_regset32 common, region[8];
|
|
} ast_regsets[2];
|
|
};
|
|
|
|
#define NV(x) "nvidia," #x
|
|
#define FALCON_COVERAGE_MEM_SIZE (1024 * 128) /* 128kB */
|
|
|
|
struct camrtc_dbgfs_rmem {
|
|
/* reserved memory base address */
|
|
phys_addr_t base_address;
|
|
/* reserved memory size */
|
|
unsigned long total_size;
|
|
/* if reserved memory enabled */
|
|
bool enabled;
|
|
/* memory contexts */
|
|
struct camrtc_rmem_ctx {
|
|
phys_addr_t address;
|
|
unsigned long size;
|
|
} mem_ctxs[CAMRTC_DBG_NUM_MEM_TEST_MEM];
|
|
};
|
|
|
|
static struct camrtc_dbgfs_rmem _camdbg_rmem;
|
|
|
|
static int __init camrtc_dbgfs_rmem_init(struct reserved_mem *rmem)
|
|
{
|
|
int i;
|
|
phys_addr_t curr_address = rmem->base;
|
|
unsigned long ctx_size = rmem->size/CAMRTC_DBG_NUM_MEM_TEST_MEM;
|
|
|
|
_camdbg_rmem.base_address = rmem->base;
|
|
_camdbg_rmem.total_size = rmem->size;
|
|
|
|
for (i = 0; i < CAMRTC_DBG_NUM_MEM_TEST_MEM; i++) {
|
|
_camdbg_rmem.mem_ctxs[i].address = curr_address;
|
|
_camdbg_rmem.mem_ctxs[i].size = ctx_size;
|
|
curr_address += ctx_size;
|
|
}
|
|
|
|
_camdbg_rmem.enabled = true;
|
|
|
|
return 0;
|
|
}
|
|
|
|
RESERVEDMEM_OF_DECLARE(tegra_cam_rtcpu,
|
|
"nvidia,camdbg_carveout", camrtc_dbgfs_rmem_init);
|
|
|
|
/* Get a camera-rtcpu device */
|
|
static struct device *camrtc_get_device(struct tegra_ivc_channel *ch)
|
|
{
|
|
if (unlikely(ch == NULL))
|
|
return NULL;
|
|
|
|
BUG_ON(ch->dev.parent == NULL);
|
|
BUG_ON(ch->dev.parent->parent == NULL);
|
|
|
|
return ch->dev.parent->parent;
|
|
}
|
|
|
|
#define INIT_OPEN_FOPS(_open) { \
|
|
.open = _open, \
|
|
.read = seq_read, \
|
|
.llseek = seq_lseek, \
|
|
.release = single_release \
|
|
}
|
|
|
|
#define DEFINE_SEQ_FOPS(_fops_, _show_) \
|
|
static int _fops_ ## _open(struct inode *inode, struct file *file) \
|
|
{ \
|
|
return single_open(file, _show_, inode->i_private); \
|
|
} \
|
|
static const struct file_operations _fops_ = INIT_OPEN_FOPS(_fops_ ## _open)
|
|
|
|
static int camrtc_show_version(struct seq_file *file, void *data)
|
|
{
|
|
struct tegra_ivc_channel *ch = file->private;
|
|
struct device *rce_dev = camrtc_get_device(ch);
|
|
char version[TEGRA_CAMRTC_VERSION_LEN];
|
|
|
|
tegra_camrtc_print_version(rce_dev, version, sizeof(version));
|
|
|
|
seq_puts(file, version);
|
|
seq_puts(file, "\n");
|
|
|
|
return 0;
|
|
}
|
|
|
|
DEFINE_SEQ_FOPS(camrtc_dbgfs_fops_version, camrtc_show_version);
|
|
|
|
static int camrtc_show_reboot(struct seq_file *file, void *data)
|
|
{
|
|
struct tegra_ivc_channel *ch = file->private;
|
|
struct device *rce_dev = camrtc_get_device(ch);
|
|
int ret = 0;
|
|
|
|
/* Make rtcpu online */
|
|
ret = tegra_ivc_channel_runtime_get(ch);
|
|
if (ret < 0)
|
|
goto error;
|
|
|
|
ret = tegra_camrtc_reboot(rce_dev);
|
|
if (ret)
|
|
goto error;
|
|
|
|
seq_puts(file, "0\n");
|
|
|
|
error:
|
|
tegra_ivc_channel_runtime_put(ch);
|
|
return ret;
|
|
}
|
|
|
|
DEFINE_SEQ_FOPS(camrtc_dbgfs_fops_reboot, camrtc_show_reboot);
|
|
|
|
static void camrtc_debug_notify(struct tegra_ivc_channel *ch)
|
|
{
|
|
struct camrtc_debug *crd = tegra_ivc_channel_get_drvdata(ch);
|
|
|
|
wake_up_all(&crd->waitq);
|
|
}
|
|
|
|
static int camrtc_show_forced_reset_restore(struct seq_file *file, void *data)
|
|
{
|
|
struct tegra_ivc_channel *ch = file->private;
|
|
struct device *rce_dev = camrtc_get_device(ch);
|
|
int ret = 0;
|
|
|
|
/* Make rtcpu online */
|
|
ret = tegra_ivc_channel_runtime_get(ch);
|
|
if (ret < 0)
|
|
goto error;
|
|
|
|
ret = tegra_camrtc_restore(rce_dev);
|
|
if (ret)
|
|
goto error;
|
|
|
|
seq_puts(file, "0\n");
|
|
|
|
error:
|
|
tegra_ivc_channel_runtime_put(ch);
|
|
return ret;
|
|
}
|
|
|
|
DEFINE_SEQ_FOPS(camrtc_dbgfs_fops_forced_reset_restore,
|
|
camrtc_show_forced_reset_restore);
|
|
|
|
static int camrtc_ivc_dbg_full_frame_xact(
|
|
struct tegra_ivc_channel *ch,
|
|
struct camrtc_dbg_request *req,
|
|
size_t req_size,
|
|
struct camrtc_dbg_response *resp,
|
|
size_t resp_size,
|
|
long timeout)
|
|
{
|
|
struct camrtc_debug *crd = tegra_ivc_channel_get_drvdata(ch);
|
|
int ret;
|
|
|
|
if (req == NULL || resp == NULL)
|
|
return -ENOMEM;
|
|
|
|
if (timeout == 0)
|
|
timeout = crd->parameters.completion_timeout;
|
|
|
|
timeout = msecs_to_jiffies(timeout);
|
|
|
|
ret = mutex_lock_interruptible(&crd->mutex);
|
|
if (ret)
|
|
return ret;
|
|
|
|
ret = tegra_ivc_channel_runtime_get(ch);
|
|
if (ret < 0)
|
|
goto unlock;
|
|
|
|
if ((!tegra_ivc_channel_online_check(ch))) {
|
|
ret = -ECONNRESET;
|
|
goto out;
|
|
}
|
|
|
|
while (tegra_ivc_can_read(&ch->ivc)) {
|
|
tegra_ivc_read_advance(&ch->ivc);
|
|
dev_warn(&ch->dev, "stray response\n");
|
|
}
|
|
|
|
timeout = wait_event_interruptible_timeout(crd->waitq,
|
|
tegra_ivc_channel_has_been_reset(ch) ||
|
|
tegra_ivc_can_write(&ch->ivc), timeout);
|
|
if (timeout <= 0) {
|
|
ret = timeout ?: -ETIMEDOUT;
|
|
goto out;
|
|
}
|
|
if (tegra_ivc_channel_has_been_reset(ch)) {
|
|
ret = -ECONNRESET;
|
|
goto out;
|
|
}
|
|
|
|
ret = tegra_ivc_write(&ch->ivc, NULL, req, req_size);
|
|
if (ret < 0) {
|
|
dev_err(&ch->dev, "IVC write error: %d\n", ret);
|
|
goto out;
|
|
}
|
|
|
|
for (;;) {
|
|
timeout = wait_event_interruptible_timeout(crd->waitq,
|
|
tegra_ivc_channel_has_been_reset(ch) ||
|
|
tegra_ivc_can_read(&ch->ivc),
|
|
timeout);
|
|
if (timeout <= 0) {
|
|
ret = timeout ?: -ETIMEDOUT;
|
|
break;
|
|
}
|
|
if (tegra_ivc_channel_has_been_reset(ch)) {
|
|
ret = -ECONNRESET;
|
|
break;
|
|
}
|
|
|
|
dev_dbg(&ch->dev, "rx msg\n");
|
|
|
|
ret = tegra_ivc_read_peek(&ch->ivc, NULL, resp, 0, resp_size);
|
|
if (ret < 0) {
|
|
dev_err(&ch->dev, "IVC read error: %d\n", ret);
|
|
break;
|
|
}
|
|
|
|
tegra_ivc_read_advance(&ch->ivc);
|
|
|
|
if (resp->resp_type == req->req_type) {
|
|
ret = 0;
|
|
break;
|
|
}
|
|
|
|
dev_err(&ch->dev, "unexpected response\n");
|
|
}
|
|
|
|
out:
|
|
tegra_ivc_channel_runtime_put(ch);
|
|
unlock:
|
|
mutex_unlock(&crd->mutex);
|
|
return ret;
|
|
}
|
|
|
|
static inline int camrtc_ivc_dbg_xact(
|
|
struct tegra_ivc_channel *ch,
|
|
struct camrtc_dbg_request *req,
|
|
struct camrtc_dbg_response *resp,
|
|
long timeout)
|
|
{
|
|
return camrtc_ivc_dbg_full_frame_xact(ch, req, sizeof(*req),
|
|
resp, sizeof(*resp),
|
|
timeout);
|
|
}
|
|
|
|
static int camrtc_show_ping(struct seq_file *file, void *data)
|
|
{
|
|
struct tegra_ivc_channel *ch = file->private;
|
|
struct camrtc_dbg_request req = {
|
|
.req_type = CAMRTC_REQ_PING,
|
|
};
|
|
struct camrtc_dbg_response resp;
|
|
u64 sent, recv, tsc;
|
|
int ret;
|
|
|
|
sent = sched_clock();
|
|
req.data.ping_data.ts_req = sent;
|
|
|
|
ret = camrtc_ivc_dbg_xact(ch, &req, &resp, 0);
|
|
if (ret)
|
|
return ret;
|
|
|
|
recv = sched_clock();
|
|
tsc = resp.data.ping_data.ts_resp;
|
|
seq_printf(file,
|
|
"roundtrip=%llu.%03llu us "
|
|
"(sent=%llu.%09llu recv=%llu.%09llu)\n",
|
|
(recv - sent) / 1000, (recv - sent) % 1000,
|
|
sent / 1000000000, sent % 1000000000,
|
|
recv / 1000000000, recv % 1000000000);
|
|
seq_printf(file,
|
|
"rtcpu tsc=%llu.%09llu offset=%llu.%09llu\n",
|
|
tsc / (1000000000 / 32), tsc % (1000000000 / 32),
|
|
(tsc * 32ULL - sent) / 1000000000,
|
|
(tsc * 32ULL - sent) % 1000000000);
|
|
seq_printf(file, "%.*s\n",
|
|
(int)sizeof(resp.data.ping_data.data),
|
|
(char *)resp.data.ping_data.data);
|
|
|
|
return 0;
|
|
}
|
|
|
|
DEFINE_SEQ_FOPS(camrtc_dbgfs_fops_ping, camrtc_show_ping);
|
|
|
|
static int camrtc_show_sm_ping(struct seq_file *file, void *data)
|
|
{
|
|
struct tegra_ivc_channel *ch = file->private;
|
|
struct device *camrtc = camrtc_get_device(ch);
|
|
u64 sent, recv;
|
|
int err;
|
|
|
|
err = tegra_ivc_channel_runtime_get(ch);
|
|
if (err < 0)
|
|
return err;
|
|
|
|
sent = sched_clock();
|
|
|
|
err = tegra_camrtc_ping(camrtc, (uint32_t)sent & 0xffffffU, 0);
|
|
if (err < 0)
|
|
goto error;
|
|
|
|
recv = sched_clock();
|
|
err = 0;
|
|
|
|
seq_printf(file,
|
|
"roundtrip=%llu.%03llu us "
|
|
"(sent=%llu.%09llu recv=%llu.%09llu)\n",
|
|
(recv - sent) / 1000, (recv - sent) % 1000,
|
|
sent / 1000000000, sent % 1000000000,
|
|
recv / 1000000000, recv % 1000000000);
|
|
|
|
error:
|
|
tegra_ivc_channel_runtime_put(ch);
|
|
|
|
return err;
|
|
}
|
|
|
|
DEFINE_SEQ_FOPS(camrtc_dbgfs_fops_sm_ping, camrtc_show_sm_ping);
|
|
|
|
static int camrtc_dbgfs_show_loglevel(void *data, u64 *val)
|
|
{
|
|
struct tegra_ivc_channel *ch = data;
|
|
struct camrtc_dbg_request req = {
|
|
.req_type = CAMRTC_REQ_GET_LOGLEVEL,
|
|
};
|
|
struct camrtc_dbg_response resp;
|
|
int ret;
|
|
|
|
ret = camrtc_ivc_dbg_xact(ch, &req, &resp, 0);
|
|
if (ret)
|
|
return ret;
|
|
|
|
if (resp.status != CAMRTC_STATUS_OK)
|
|
return -EPROTO;
|
|
|
|
*val = resp.data.log_data.level;
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int camrtc_dbgfs_store_loglevel(void *data, u64 val)
|
|
{
|
|
struct tegra_ivc_channel *ch = data;
|
|
struct camrtc_dbg_request req = {
|
|
.req_type = CAMRTC_REQ_SET_LOGLEVEL,
|
|
};
|
|
struct camrtc_dbg_response resp;
|
|
int ret;
|
|
|
|
if ((u32)val != val)
|
|
return -EINVAL;
|
|
|
|
req.data.log_data.level = val;
|
|
|
|
ret = camrtc_ivc_dbg_xact(ch, &req, &resp, 0);
|
|
if (ret)
|
|
return ret;
|
|
|
|
if (resp.status == CAMRTC_STATUS_INVALID_PARAM)
|
|
return -EINVAL;
|
|
else if (resp.status != CAMRTC_STATUS_OK)
|
|
return -EPROTO;
|
|
|
|
return 0;
|
|
}
|
|
|
|
DEFINE_SIMPLE_ATTRIBUTE(camrtc_dbgfs_fops_loglevel,
|
|
camrtc_dbgfs_show_loglevel,
|
|
camrtc_dbgfs_store_loglevel,
|
|
"%lld\n");
|
|
|
|
static int camrtc_show_mods_result(struct seq_file *file, void *data)
|
|
{
|
|
struct tegra_ivc_channel *ch = file->private;
|
|
struct camrtc_debug *crd = tegra_ivc_channel_get_drvdata(ch);
|
|
struct camrtc_dbg_request req = {
|
|
.req_type = CAMRTC_REQ_MODS_TEST,
|
|
};
|
|
struct camrtc_dbg_response resp;
|
|
int ret;
|
|
unsigned long timeout = crd->parameters.completion_timeout;
|
|
u32 loops = crd->parameters.mods_loops;
|
|
|
|
req.data.mods_data.mods_case = crd->parameters.mods_case;
|
|
req.data.mods_data.mods_loops = loops;
|
|
req.data.mods_data.mods_dma_channels = crd->parameters.mods_dma_channels;
|
|
|
|
ret = camrtc_ivc_dbg_xact(ch, &req, &resp, loops * timeout);
|
|
if (ret == 0)
|
|
seq_printf(file, "mods=%u\n", resp.status);
|
|
|
|
return ret;
|
|
}
|
|
|
|
DEFINE_SEQ_FOPS(camrtc_dbgfs_fops_mods_result, camrtc_show_mods_result);
|
|
|
|
static int camrtc_dbgfs_show_freertos_state(struct seq_file *file, void *data)
|
|
{
|
|
struct tegra_ivc_channel *ch = file->private;
|
|
struct camrtc_dbg_request req = {
|
|
.req_type = CAMRTC_REQ_RTOS_STATE,
|
|
};
|
|
struct camrtc_dbg_response resp;
|
|
int ret = 0;
|
|
|
|
ret = camrtc_ivc_dbg_xact(ch, &req, &resp, 0);
|
|
if (ret == 0) {
|
|
seq_printf(file, "%.*s",
|
|
(int) sizeof(resp.data.rtos_state_data.rtos_state),
|
|
resp.data.rtos_state_data.rtos_state);
|
|
}
|
|
|
|
return ret;
|
|
}
|
|
|
|
DEFINE_SEQ_FOPS(camrtc_dbgfs_fops_freertos_state,
|
|
camrtc_dbgfs_show_freertos_state);
|
|
|
|
static int camrtc_dbgfs_show_memstat(struct seq_file *file, void *data)
|
|
{
|
|
struct tegra_ivc_channel *ch = file->private;
|
|
struct camrtc_dbg_request req = {
|
|
.req_type = CAMRTC_REQ_GET_MEM_USAGE,
|
|
};
|
|
struct camrtc_dbg_response resp;
|
|
int ret = 0;
|
|
|
|
ret = camrtc_ivc_dbg_xact(ch, &req, &resp, 0);
|
|
if (ret == 0) {
|
|
const struct camrtc_dbg_mem_usage *m = &resp.data.mem_usage;
|
|
uint32_t total = m->text + m->bss + m->data + m->heap +
|
|
m->stack + m->free_mem;
|
|
|
|
#define K(x) (((x) + 1023) / 1024)
|
|
seq_printf(file, "%7s %7s %7s %7s %7s %7s %7s\n",
|
|
"text", "bss", "data", "heap", "sys", "free", "TOTAL");
|
|
seq_printf(file, "%7u\t%7u\t%7u\t%7u\t%7u\t%7u\t%7u\n",
|
|
m->text, m->bss, m->data, m->heap, m->stack, m->free_mem, total);
|
|
seq_printf(file, "%7u\t%7u\t%7u\t%7u\t%7u\t%7u\t%7u (in kilobytes)\n",
|
|
K(m->text), K(m->bss), K(m->data), K(m->heap),
|
|
K(m->stack), K(m->free_mem), K(total));
|
|
}
|
|
|
|
return ret;
|
|
}
|
|
|
|
DEFINE_SEQ_FOPS(camrtc_dbgfs_fops_memstat, camrtc_dbgfs_show_memstat);
|
|
|
|
static int camrtc_dbgfs_show_irqstat(struct seq_file *file, void *data)
|
|
{
|
|
int ret = -ENOMSG;
|
|
#ifdef CAMRTC_REQ_GET_IRQ_STAT
|
|
struct tegra_ivc_channel *ch = file->private;
|
|
struct camrtc_dbg_request req = {
|
|
.req_type = CAMRTC_REQ_GET_IRQ_STAT,
|
|
};
|
|
void *mem = kzalloc(ch->ivc.frame_size, GFP_KERNEL | __GFP_ZERO);
|
|
struct camrtc_dbg_response *resp = mem;
|
|
const struct camrtc_dbg_irq_stat *stat = &resp->data.irq_stat;
|
|
uint32_t i;
|
|
uint32_t max_runtime = 0;
|
|
|
|
ret = camrtc_ivc_dbg_full_frame_xact(ch, &req, sizeof(req),
|
|
resp, ch->ivc.frame_size, 0);
|
|
if (ret != 0)
|
|
goto done;
|
|
|
|
seq_printf(file, "Irq#\tCount\tRuntime\tMax rt\tName\n");
|
|
|
|
for (i = 0; i < stat->n_irq; i++) {
|
|
seq_printf(file, "%u\t%u\t%llu\t%u\t%.*s\n",
|
|
stat->irqs[i].irq_num,
|
|
stat->irqs[i].num_called,
|
|
stat->irqs[i].runtime,
|
|
stat->irqs[i].max_runtime,
|
|
(int)sizeof(stat->irqs[i].name), stat->irqs[i].name);
|
|
|
|
if (max_runtime < stat->irqs[i].max_runtime)
|
|
max_runtime = stat->irqs[i].max_runtime;
|
|
}
|
|
|
|
seq_printf(file, "-\t%llu\t%llu\t%u\t%s\n", stat->total_called,
|
|
stat->total_runtime, max_runtime, "total");
|
|
|
|
done:
|
|
kfree(mem);
|
|
#endif
|
|
return ret;
|
|
}
|
|
|
|
DEFINE_SEQ_FOPS(camrtc_dbgfs_fops_irqstat, camrtc_dbgfs_show_irqstat);
|
|
|
|
static size_t camrtc_dbgfs_get_max_test_size(
|
|
const struct tegra_ivc_channel *ch)
|
|
{
|
|
return ch->ivc.frame_size - offsetof(struct camrtc_dbg_request,
|
|
data.run_mem_test_data.data);
|
|
}
|
|
|
|
static ssize_t camrtc_dbgfs_read_test_case(struct file *file,
|
|
char __user *buf, size_t count, loff_t *f_pos)
|
|
{
|
|
struct tegra_ivc_channel *ch = file->f_inode->i_private;
|
|
struct camrtc_debug *crd = tegra_ivc_channel_get_drvdata(ch);
|
|
|
|
return simple_read_from_buffer(buf, count, f_pos,
|
|
crd->parameters.test_case,
|
|
crd->parameters.test_case_size);
|
|
}
|
|
|
|
static ssize_t camrtc_dbgfs_write_test_case(struct file *file,
|
|
const char __user *buf, size_t count, loff_t *f_pos)
|
|
{
|
|
struct tegra_ivc_channel *ch = file->f_inode->i_private;
|
|
struct camrtc_debug *crd = tegra_ivc_channel_get_drvdata(ch);
|
|
char *test_case = crd->parameters.test_case;
|
|
size_t max_size = camrtc_dbgfs_get_max_test_size(ch);
|
|
int i;
|
|
ssize_t ret;
|
|
|
|
ret = simple_write_to_buffer(test_case, max_size, f_pos, buf, count);
|
|
|
|
if (ret >= 0)
|
|
crd->parameters.test_case_size = *f_pos;
|
|
|
|
/* Mark input buffers empty */
|
|
for (i = 0; i < ARRAY_SIZE(crd->mem); i++)
|
|
crd->mem[i].used = 0;
|
|
|
|
return ret;
|
|
}
|
|
|
|
static const struct file_operations camrtc_dbgfs_fops_test_case = {
|
|
.read = camrtc_dbgfs_read_test_case,
|
|
.write = camrtc_dbgfs_write_test_case,
|
|
};
|
|
|
|
static struct device *camrtc_dbgfs_memory_dev(
|
|
const struct camrtc_debug *crd)
|
|
{
|
|
/*
|
|
* If VI misses stage-1 SMMU translation, the allocations need
|
|
* to be contiguous. Just allocate everything through VI and
|
|
* map it to other contexts separately.
|
|
*/
|
|
if (crd->mem_devices[1] != NULL)
|
|
return crd->mem_devices[1];
|
|
else
|
|
return crd->mem_devices[0];
|
|
}
|
|
|
|
static ssize_t camrtc_dbgfs_read_test_mem(struct file *file,
|
|
char __user *buf, size_t count, loff_t *f_pos)
|
|
{
|
|
struct camrtc_test_mem *mem = file->f_inode->i_private;
|
|
|
|
return simple_read_from_buffer(buf, count, f_pos, mem->ptr, mem->used);
|
|
}
|
|
|
|
static ssize_t camrtc_dbgfs_write_test_mem(struct file *file,
|
|
const char __user *buf, size_t count, loff_t *f_pos)
|
|
{
|
|
struct camrtc_test_mem *mem = file->f_inode->i_private;
|
|
struct camrtc_debug *crd = container_of(
|
|
mem, struct camrtc_debug, mem[mem->index]);
|
|
struct device *mem_dev = camrtc_dbgfs_memory_dev(crd);
|
|
struct iommu_domain *domain = iommu_get_domain_for_dev(mem_dev);
|
|
ssize_t ret;
|
|
|
|
if ((*f_pos + count) > mem->size) {
|
|
if (_camdbg_rmem.enabled) {
|
|
size_t size = round_up(*f_pos + count, 64 * 1024);
|
|
void *ptr = phys_to_virt(
|
|
_camdbg_rmem.mem_ctxs[mem->index].address);
|
|
unsigned long rmem_size =
|
|
_camdbg_rmem.mem_ctxs[mem->index].size;
|
|
|
|
if (size > rmem_size) {
|
|
pr_err("%s: not enough memory\n", __func__);
|
|
return -ENOMEM;
|
|
}
|
|
|
|
if (mem->ptr)
|
|
dma_unmap_single(mem_dev, mem->iova, mem->size,
|
|
DMA_BIDIRECTIONAL);
|
|
|
|
/* same addr, no overwrite concern */
|
|
mem->ptr = ptr;
|
|
mem->size = size;
|
|
|
|
mem->iova = dma_map_single(mem_dev, mem->ptr,
|
|
mem->size, DMA_BIDIRECTIONAL);
|
|
if (dma_mapping_error(mem_dev, mem->iova)) {
|
|
pr_err("%s: dma map failed\n", __func__);
|
|
return -ENOMEM;
|
|
}
|
|
} else {
|
|
size_t size = round_up(*f_pos + count, 64 * 1024);
|
|
dma_addr_t iova;
|
|
void *ptr = dma_alloc_coherent(mem_dev, size, &iova,
|
|
GFP_KERNEL | __GFP_ZERO);
|
|
if (ptr == NULL)
|
|
return -ENOMEM;
|
|
if (mem->ptr) {
|
|
memcpy(ptr, mem->ptr, mem->used);
|
|
dma_free_coherent(mem_dev, mem->size, mem->ptr,
|
|
mem->iova);
|
|
}
|
|
mem->ptr = ptr;
|
|
mem->size = size;
|
|
mem->iova = iova;
|
|
}
|
|
|
|
/* If mem_dev is not connected to SMMU, the iova is physical */
|
|
if (domain)
|
|
mem->phys_addr = iommu_iova_to_phys(domain, mem->iova);
|
|
else
|
|
mem->phys_addr = mem->iova;
|
|
}
|
|
|
|
ret = simple_write_to_buffer(mem->ptr, mem->size, f_pos, buf, count);
|
|
|
|
if (ret >= 0) {
|
|
mem->used = *f_pos;
|
|
|
|
if (mem->used == 0 && mem->ptr != NULL) {
|
|
if (_camdbg_rmem.enabled)
|
|
dma_unmap_single(mem_dev, mem->iova, mem->size,
|
|
DMA_BIDIRECTIONAL);
|
|
else
|
|
dma_free_coherent(mem_dev, mem->size, mem->ptr,
|
|
mem->iova);
|
|
|
|
memset(mem, 0, sizeof(*mem));
|
|
}
|
|
}
|
|
|
|
return ret;
|
|
}
|
|
|
|
static const struct file_operations camrtc_dbgfs_fops_test_mem = {
|
|
.read = camrtc_dbgfs_read_test_mem,
|
|
.write = camrtc_dbgfs_write_test_mem,
|
|
};
|
|
|
|
#define BUILD_BUG_ON_MISMATCH(s1, f1, s2, f2) \
|
|
BUILD_BUG_ON(offsetof(s1, data.f1) != offsetof(s2, data.f2))
|
|
|
|
static int camrtc_test_run_and_show_result(struct seq_file *file,
|
|
struct camrtc_dbg_request *req,
|
|
struct camrtc_dbg_response *resp,
|
|
size_t data_offset)
|
|
{
|
|
struct tegra_ivc_channel *ch = file->private;
|
|
struct camrtc_debug *crd = tegra_ivc_channel_get_drvdata(ch);
|
|
const char *test_case = crd->parameters.test_case;
|
|
size_t test_case_size = crd->parameters.test_case_size;
|
|
unsigned long timeout = crd->parameters.test_timeout;
|
|
uint64_t ns;
|
|
size_t req_size = ch->ivc.frame_size;
|
|
size_t resp_size = ch->ivc.frame_size;
|
|
int ret;
|
|
const char *result = (const void *)resp + data_offset;
|
|
size_t result_size = resp_size - data_offset;
|
|
const char *nul;
|
|
|
|
if (WARN_ON(test_case_size > camrtc_dbgfs_get_max_test_size(ch)))
|
|
test_case_size = camrtc_dbgfs_get_max_test_size(ch);
|
|
|
|
memcpy((char *)req + data_offset, test_case, test_case_size);
|
|
|
|
/* Timeout is in ms, run_test_data.timeout in ns */
|
|
if (timeout > 40)
|
|
ns = 1000000ULL * (timeout - 20);
|
|
else
|
|
ns = 1000000ULL * (timeout / 2);
|
|
|
|
BUILD_BUG_ON_MISMATCH(
|
|
struct camrtc_dbg_request, run_mem_test_data.timeout,
|
|
struct camrtc_dbg_request, run_test_data.timeout);
|
|
|
|
ret = tegra_ivc_channel_runtime_get(ch);
|
|
if (ret < 0)
|
|
return ret;
|
|
|
|
req->data.run_test_data.timeout = ns;
|
|
|
|
ret = camrtc_ivc_dbg_full_frame_xact(ch, req, req_size,
|
|
resp, resp_size, timeout);
|
|
|
|
tegra_camrtc_flush_trace(camrtc_get_device(ch));
|
|
|
|
if (ret < 0) {
|
|
if (ret != -ECONNRESET) {
|
|
dev_info(&ch->dev, "rebooting after a failed test run");
|
|
(void)tegra_camrtc_reboot(camrtc_get_device(ch));
|
|
}
|
|
goto runtime_put;
|
|
}
|
|
|
|
BUILD_BUG_ON_MISMATCH(
|
|
struct camrtc_dbg_response, run_mem_test_data.timeout,
|
|
struct camrtc_dbg_response, run_test_data.timeout);
|
|
|
|
ns = resp->data.run_test_data.timeout;
|
|
|
|
seq_printf(file, "result=%u runtime=%llu.%06llu ms\n\n",
|
|
resp->status, ns / 1000000, ns % 1000000);
|
|
|
|
nul = memchr(result, '\0', result_size);
|
|
if (nul)
|
|
seq_write(file, result, nul - result);
|
|
else
|
|
seq_write(file, result, result_size);
|
|
|
|
runtime_put:
|
|
tegra_ivc_channel_runtime_put(ch);
|
|
|
|
return ret;
|
|
}
|
|
|
|
static void camrtc_run_rmem_unmap_all(struct camrtc_debug *crd,
|
|
struct camrtc_test_mem *mem, bool all)
|
|
{
|
|
int i;
|
|
struct device *mem_dev = camrtc_dbgfs_memory_dev(crd);
|
|
|
|
/* Nothing to unmap */
|
|
if (mem->ptr == NULL)
|
|
return;
|
|
|
|
for (i = 0; i < mem->dev_index; i++) {
|
|
struct device *dev = mem->devices[i].dev;
|
|
dma_addr_t iova = mem->devices[i].dev_iova;
|
|
|
|
if (dev == NULL)
|
|
break;
|
|
|
|
/* keep mem_dev mapped unless forced */
|
|
if (!all && (dev == mem_dev))
|
|
continue;
|
|
|
|
dma_unmap_single(dev, iova,
|
|
mem->size, DMA_BIDIRECTIONAL);
|
|
}
|
|
}
|
|
|
|
#define INT_MAX ((int)(~0U >> 1))
|
|
|
|
static int camrtc_run_mem_map(struct tegra_ivc_channel *ch,
|
|
struct device *mem_dev,
|
|
struct device *dev,
|
|
struct sg_table *sgt,
|
|
struct camrtc_test_mem *mem,
|
|
uint64_t *return_iova)
|
|
{
|
|
int ret = 0;
|
|
|
|
*return_iova = 0ULL;
|
|
|
|
if (dev == NULL)
|
|
return 0;
|
|
|
|
if (mem->dev_index >= CAMRTC_TEST_CAM_DEVICES) {
|
|
pr_err("%s: device list exhausted\n", __func__);
|
|
return -ENOMEM;
|
|
}
|
|
|
|
if (mem_dev == dev) {
|
|
*return_iova = mem->iova;
|
|
dma_sync_single_for_device(dev, mem->iova, mem->size,
|
|
DMA_BIDIRECTIONAL);
|
|
goto done;
|
|
}
|
|
|
|
if (_camdbg_rmem.enabled) {
|
|
*return_iova = dma_map_single(dev, mem->ptr,
|
|
mem->size, DMA_BIDIRECTIONAL);
|
|
if (dma_mapping_error(dev, *return_iova)) {
|
|
pr_err("%s: dma map failed\n", __func__);
|
|
*return_iova = 0ULL;
|
|
return -ENOMEM;
|
|
}
|
|
dma_sync_single_for_device(dev, mem->iova, mem->size,
|
|
DMA_BIDIRECTIONAL);
|
|
} else {
|
|
ret = dma_get_sgtable(dev, sgt, mem->ptr, mem->iova, mem->size);
|
|
if (ret < 0) {
|
|
dev_err(&ch->dev, "dma_get_sgtable for %s failed\n",
|
|
dev_name(dev));
|
|
return ret;
|
|
}
|
|
|
|
if (!dma_map_sg(dev, sgt->sgl, sgt->orig_nents,
|
|
DMA_BIDIRECTIONAL)) {
|
|
dev_err(&ch->dev, "failed to map %s mem at 0x%llx\n",
|
|
dev_name(dev), (u64)mem->iova);
|
|
sg_free_table(sgt);
|
|
ret = -ENXIO;
|
|
}
|
|
|
|
*return_iova = sgt->sgl->dma_address;
|
|
if (sgt->nents <= INT_MAX)
|
|
dma_sync_sg_for_device(dev, sgt->sgl, (int)sgt->nents, DMA_BIDIRECTIONAL);
|
|
else
|
|
ret = -EINVAL;
|
|
}
|
|
|
|
done:
|
|
mem->devices[mem->dev_index].dev = dev;
|
|
mem->devices[mem->dev_index++].dev_iova = *return_iova;
|
|
|
|
return ret;
|
|
}
|
|
|
|
static void camrtc_membw_set(struct camrtc_debug *crd, u32 bw)
|
|
{
|
|
int ret;
|
|
|
|
if (crd->icc_path) {
|
|
ret = icc_set_bw(crd->icc_path, 0, bw);
|
|
if (ret)
|
|
dev_err(crd->mem_devices[0],
|
|
"set icc bw [%u] failed: %d\n", bw, ret);
|
|
else
|
|
dev_dbg(crd->mem_devices[0], "requested icc bw %u\n", bw);
|
|
}
|
|
}
|
|
|
|
static int camrtc_run_mem_test(struct seq_file *file,
|
|
struct camrtc_dbg_request *req,
|
|
struct camrtc_dbg_response *resp)
|
|
{
|
|
struct tegra_ivc_channel *ch = file->private;
|
|
struct camrtc_debug *crd = tegra_ivc_channel_get_drvdata(ch);
|
|
struct camrtc_dbg_test_mem *testmem;
|
|
size_t i;
|
|
int ret = 0;
|
|
struct device *mem_dev = camrtc_dbgfs_memory_dev(crd);
|
|
struct device *rce_dev = crd->mem_devices[0];
|
|
struct sg_table rce_sgt[ARRAY_SIZE(crd->mem)];
|
|
struct device *vi_dev = crd->mem_devices[1];
|
|
struct sg_table vi_sgt[ARRAY_SIZE(crd->mem)];
|
|
struct device *isp_dev = crd->mem_devices[2];
|
|
struct sg_table isp_sgt[ARRAY_SIZE(crd->mem)];
|
|
struct device *vi2_dev = crd->mem_devices[3];
|
|
struct sg_table vi2_sgt[ARRAY_SIZE(crd->mem)];
|
|
struct camrtc_test_mem *mem0 = &crd->mem[0];
|
|
|
|
memset(rce_sgt, 0, sizeof(rce_sgt));
|
|
memset(vi_sgt, 0, sizeof(vi_sgt));
|
|
memset(isp_sgt, 0, sizeof(isp_sgt));
|
|
memset(vi2_sgt, 0, sizeof(vi2_sgt));
|
|
|
|
req->req_type = CAMRTC_REQ_RUN_MEM_TEST;
|
|
|
|
/* Allocate 6MB scratch memory in mem0 by default */
|
|
if (!mem0->used) {
|
|
const size_t size = 6U << 20U; /* 6 MB */
|
|
dma_addr_t iova;
|
|
void *ptr;
|
|
struct iommu_domain *domain = iommu_get_domain_for_dev(mem_dev);
|
|
|
|
if (mem0->ptr) {
|
|
if (_camdbg_rmem.enabled)
|
|
camrtc_run_rmem_unmap_all(crd, mem0, true);
|
|
else
|
|
dma_free_coherent(mem_dev, mem0->size,
|
|
mem0->ptr, mem0->iova);
|
|
|
|
memset(mem0, 0, sizeof(*mem0));
|
|
}
|
|
|
|
if (_camdbg_rmem.enabled) {
|
|
if (_camdbg_rmem.mem_ctxs[0].size < size) {
|
|
pr_err(
|
|
"%s: mem [%lu] < req size [%lu]\n",
|
|
__func__, _camdbg_rmem.mem_ctxs[0].size,
|
|
size);
|
|
return -ENOMEM;
|
|
}
|
|
|
|
ptr = phys_to_virt(_camdbg_rmem.mem_ctxs[0].address);
|
|
|
|
iova = dma_map_single(mem_dev, ptr, size,
|
|
DMA_BIDIRECTIONAL);
|
|
if (dma_mapping_error(mem_dev, iova)) {
|
|
pr_err("%s: dma map failed\n", __func__);
|
|
return -ENOMEM;
|
|
}
|
|
} else {
|
|
ptr = dma_alloc_coherent(mem_dev, size, &iova,
|
|
GFP_KERNEL | __GFP_ZERO);
|
|
if (ptr == NULL)
|
|
return -ENOMEM;
|
|
}
|
|
|
|
mem0->ptr = ptr;
|
|
mem0->size = size;
|
|
|
|
/* If mem_dev is not connected to SMMU, the iova is physical */
|
|
if (domain)
|
|
mem0->phys_addr = iommu_iova_to_phys(domain, iova);
|
|
else
|
|
mem0->phys_addr = iova;
|
|
|
|
mem0->iova = iova;
|
|
mem0->used = size;
|
|
}
|
|
|
|
camrtc_membw_set(crd, crd->parameters.test_bw);
|
|
|
|
for (i = 0; i < ARRAY_SIZE(crd->mem); i++) {
|
|
struct camrtc_test_mem *mem = &crd->mem[i];
|
|
|
|
if (mem->used == 0)
|
|
continue;
|
|
|
|
testmem = &req->data.run_mem_test_data.mem[i];
|
|
|
|
testmem->size = mem->used;
|
|
testmem->page_size = PAGE_SIZE;
|
|
testmem->phys_addr = mem->phys_addr;
|
|
|
|
ret = camrtc_run_mem_map(ch, mem_dev, rce_dev, &rce_sgt[i], mem,
|
|
&testmem->rtcpu_iova);
|
|
if (ret < 0)
|
|
goto unmap;
|
|
|
|
ret = camrtc_run_mem_map(ch, mem_dev, vi_dev, &vi_sgt[i], mem,
|
|
&testmem->vi_iova);
|
|
if (ret < 0)
|
|
goto unmap;
|
|
|
|
ret = camrtc_run_mem_map(ch, mem_dev, isp_dev, &isp_sgt[i], mem,
|
|
&testmem->isp_iova);
|
|
if (ret < 0)
|
|
goto unmap;
|
|
|
|
ret = camrtc_run_mem_map(ch, mem_dev, vi2_dev, &vi2_sgt[i], mem,
|
|
&testmem->vi2_iova);
|
|
if (ret < 0)
|
|
goto unmap;
|
|
|
|
}
|
|
|
|
BUILD_BUG_ON_MISMATCH(
|
|
struct camrtc_dbg_request, run_mem_test_data.data,
|
|
struct camrtc_dbg_response, run_mem_test_data.data);
|
|
|
|
ret = camrtc_test_run_and_show_result(file, req, resp,
|
|
offsetof(struct camrtc_dbg_response,
|
|
data.run_mem_test_data.data));
|
|
if (ret < 0)
|
|
goto unmap;
|
|
|
|
for (i = 0; i < ARRAY_SIZE(crd->mem); i++) {
|
|
struct camrtc_test_mem *mem = &crd->mem[i];
|
|
|
|
if (mem->size == 0)
|
|
continue;
|
|
|
|
testmem = &resp->data.run_mem_test_data.mem[i];
|
|
if (!WARN_ON(testmem->size > mem->size))
|
|
mem->used = testmem->size;
|
|
|
|
if (_camdbg_rmem.enabled)
|
|
dma_sync_single_for_cpu(mem_dev, mem->iova, mem->used,
|
|
DMA_BIDIRECTIONAL);
|
|
else
|
|
dma_sync_sg_for_cpu(mem_dev, vi_sgt[i].sgl,
|
|
vi_sgt[i].nents, DMA_BIDIRECTIONAL);
|
|
}
|
|
|
|
unmap:
|
|
camrtc_membw_set(crd, 0);
|
|
|
|
for (i = 0; i < ARRAY_SIZE(vi_sgt); i++) {
|
|
if (rce_sgt[i].sgl) {
|
|
dma_unmap_sg(rce_dev, rce_sgt[i].sgl,
|
|
rce_sgt[i].orig_nents, DMA_BIDIRECTIONAL);
|
|
sg_free_table(&rce_sgt[i]);
|
|
}
|
|
if (vi_sgt[i].sgl) {
|
|
dma_unmap_sg(vi_dev, vi_sgt[i].sgl,
|
|
vi_sgt[i].orig_nents, DMA_BIDIRECTIONAL);
|
|
sg_free_table(&vi_sgt[i]);
|
|
}
|
|
if (isp_sgt[i].sgl) {
|
|
dma_unmap_sg(isp_dev, isp_sgt[i].sgl,
|
|
isp_sgt[i].orig_nents, DMA_BIDIRECTIONAL);
|
|
sg_free_table(&isp_sgt[i]);
|
|
}
|
|
if (vi2_sgt[i].sgl) {
|
|
dma_unmap_sg(vi2_dev, vi2_sgt[i].sgl,
|
|
vi2_sgt[i].orig_nents, DMA_BIDIRECTIONAL);
|
|
sg_free_table(&vi2_sgt[i]);
|
|
}
|
|
}
|
|
|
|
if (_camdbg_rmem.enabled) {
|
|
for (i = 0; i < ARRAY_SIZE(crd->mem); i++) {
|
|
struct camrtc_test_mem *mem = &crd->mem[i];
|
|
camrtc_run_rmem_unmap_all(crd, mem, false);
|
|
}
|
|
}
|
|
|
|
/* Reset mapping info, memory can still be used by cpu tests */
|
|
for (i = 0; i < ARRAY_SIZE(crd->mem); i++) {
|
|
crd->mem[i].dev_index = 0U;
|
|
memset(&crd->mem[i].devices, 0,
|
|
(ARRAY_SIZE(crd->mem[i].devices) *
|
|
sizeof(struct camrtc_test_device)));
|
|
}
|
|
|
|
return ret;
|
|
}
|
|
|
|
static int camrtc_dbgfs_show_test_result(struct seq_file *file, void *data)
|
|
{
|
|
struct tegra_ivc_channel *ch = file->private;
|
|
void *mem = kzalloc(2 * ch->ivc.frame_size, GFP_KERNEL | __GFP_ZERO);
|
|
struct camrtc_dbg_request *req = mem;
|
|
struct camrtc_dbg_response *resp = mem + ch->ivc.frame_size;
|
|
int ret;
|
|
|
|
if (mem == NULL)
|
|
return -ENOMEM;
|
|
|
|
ret = camrtc_run_mem_test(file, req, resp);
|
|
kfree(mem);
|
|
|
|
return ret;
|
|
}
|
|
|
|
DEFINE_SEQ_FOPS(camrtc_dbgfs_fops_test_result, camrtc_dbgfs_show_test_result);
|
|
|
|
static int camrtc_dbgfs_show_test_list(struct seq_file *file, void *data)
|
|
{
|
|
struct tegra_ivc_channel *ch = file->private;
|
|
struct camrtc_dbg_request req = {
|
|
.req_type = CAMRTC_REQ_RUN_TEST,
|
|
};
|
|
struct camrtc_dbg_response *resp;
|
|
int ret;
|
|
|
|
resp = kzalloc(ch->ivc.frame_size, GFP_KERNEL | __GFP_ZERO);
|
|
if (resp == NULL)
|
|
return -ENOMEM;
|
|
|
|
memset(req.data.run_test_data.data, 0,
|
|
sizeof(req.data.run_test_data.data));
|
|
strcpy(req.data.run_test_data.data, "list\n");
|
|
|
|
ret = camrtc_ivc_dbg_full_frame_xact(ch, &req, sizeof(req),
|
|
resp, ch->ivc.frame_size, 0);
|
|
if (ret == 0 && resp->status == CAMRTC_STATUS_OK) {
|
|
char const *list = (char const *)resp->data.run_test_data.data;
|
|
size_t textsize = ch->ivc.frame_size -
|
|
offsetof(struct camrtc_dbg_response,
|
|
data.run_test_data.data);
|
|
size_t i;
|
|
|
|
/* Remove first line */
|
|
for (i = 0; i < textsize; i++)
|
|
if (list[i] == '\n')
|
|
break;
|
|
for (; i < textsize; i++)
|
|
if (list[i] != '\n' && list[i] != '\r')
|
|
break;
|
|
|
|
seq_printf(file, "%.*s", (int)(textsize - i), list + i);
|
|
}
|
|
|
|
kfree(resp);
|
|
|
|
return ret;
|
|
}
|
|
|
|
DEFINE_SEQ_FOPS(camrtc_dbgfs_fops_test_list, camrtc_dbgfs_show_test_list);
|
|
|
|
static int camrtc_coverage_msg(struct camrtc_falcon_coverage *cov,
|
|
struct camrtc_dbg_response *resp,
|
|
bool flush, bool reset)
|
|
{
|
|
struct camrtc_dbg_request req = {
|
|
.req_type = CAMRTC_REQ_SET_FALCON_COVERAGE,
|
|
.data = {
|
|
.coverage_data = {
|
|
.falcon_id = cov->id,
|
|
.size = cov->enabled ? cov->mem.size : 0,
|
|
.iova = cov->enabled ? cov->falc_iova : 0,
|
|
.flush = flush ? 1 : 0,
|
|
.reset = reset ? 1 : 0,
|
|
},
|
|
},
|
|
};
|
|
struct tegra_ivc_channel *ch = cov->ch;
|
|
int ret;
|
|
|
|
ret = camrtc_ivc_dbg_xact(ch, &req, resp, 200);
|
|
|
|
if (ret || (resp->status != CAMRTC_STATUS_OK)) {
|
|
dev_warn(&ch->dev, "Coverage IVC error: %d, status %u, id %u\n",
|
|
ret, resp->status, cov->id);
|
|
ret = -ENODEV;
|
|
} else if (resp->data.coverage_stat.full == 1) {
|
|
ret = -EOVERFLOW;
|
|
}
|
|
|
|
return ret;
|
|
}
|
|
|
|
static bool camrtc_coverage_is_supported(struct camrtc_falcon_coverage *cov)
|
|
{
|
|
struct camrtc_dbg_response resp;
|
|
|
|
(void)camrtc_coverage_msg(cov, &resp, false, false);
|
|
|
|
return (resp.status == CAMRTC_STATUS_OK);
|
|
}
|
|
|
|
static ssize_t camrtc_read_falcon_coverage(struct file *file,
|
|
char __user *buf, size_t count, loff_t *f_pos)
|
|
{
|
|
struct camrtc_falcon_coverage *cov = file->f_inode->i_private;
|
|
struct camrtc_dbg_response resp;
|
|
ssize_t ret = 0;
|
|
|
|
if (!cov->enabled) {
|
|
ret = -ENODEV;
|
|
goto done;
|
|
}
|
|
|
|
/* In the beginning, do a flush */
|
|
if (*f_pos == 0) {
|
|
/* Flush falcon buffer */
|
|
ret = camrtc_coverage_msg(cov, &resp, true, false);
|
|
|
|
if (ret)
|
|
goto done;
|
|
|
|
cov->mem.used = resp.data.coverage_stat.bytes_written;
|
|
|
|
dma_sync_single_for_cpu(cov->mem_dev, cov->mem.iova,
|
|
cov->mem.size, DMA_BIDIRECTIONAL);
|
|
}
|
|
|
|
ret = simple_read_from_buffer(buf, count, f_pos,
|
|
cov->mem.ptr, cov->mem.used);
|
|
done:
|
|
return ret;
|
|
}
|
|
|
|
static ssize_t camrtc_write_falcon_coverage(struct file *file,
|
|
const char __user *buf, size_t count, loff_t *f_pos)
|
|
{
|
|
struct camrtc_falcon_coverage *cov = file->f_inode->i_private;
|
|
struct camrtc_dbg_response resp;
|
|
ssize_t ret = count;
|
|
|
|
if (cov->enabled) {
|
|
memset(cov->mem.ptr, 0, cov->mem.size);
|
|
if (camrtc_coverage_msg(cov, &resp, false, true))
|
|
ret = -ENODEV;
|
|
else
|
|
*f_pos += count;
|
|
} else {
|
|
ret = -ENODEV;
|
|
}
|
|
|
|
return ret;
|
|
}
|
|
static const struct file_operations camrtc_dbgfs_fops_falcon_coverage = {
|
|
.read = camrtc_read_falcon_coverage,
|
|
.write = camrtc_write_falcon_coverage,
|
|
};
|
|
|
|
static int camrtc_falcon_coverage_enable(struct camrtc_falcon_coverage *cov)
|
|
{
|
|
struct tegra_ivc_channel *ch = cov->ch;
|
|
struct device *mem_dev = cov->mem_dev;
|
|
struct device *falcon_dev = cov->falcon_dev;
|
|
struct camrtc_dbg_response resp;
|
|
int ret = 0;
|
|
|
|
if (cov->enabled)
|
|
goto done;
|
|
|
|
if (!camrtc_coverage_is_supported(cov)) {
|
|
ret = -ENODEV;
|
|
goto done;
|
|
}
|
|
|
|
cov->mem.ptr = dma_alloc_coherent(mem_dev,
|
|
FALCON_COVERAGE_MEM_SIZE,
|
|
&cov->mem.iova,
|
|
GFP_KERNEL | __GFP_ZERO);
|
|
if (cov->mem.ptr == NULL) {
|
|
dev_warn(&ch->dev,
|
|
"Failed to allocate Falcon 0x%02x coverage memory!\n",
|
|
cov->id);
|
|
goto error;
|
|
}
|
|
|
|
cov->mem.size = FALCON_COVERAGE_MEM_SIZE;
|
|
|
|
if (camrtc_run_mem_map(ch, cov->mem_dev, falcon_dev,
|
|
&cov->sgt, &cov->mem,
|
|
&cov->falc_iova)) {
|
|
dev_warn(&ch->dev,
|
|
"Failed to map Falcon 0x%02x coverage memory\n",
|
|
cov->id);
|
|
goto clean_mem;
|
|
}
|
|
|
|
/* Keep rtcpu alive when falcon coverage is in use. */
|
|
ret = tegra_ivc_channel_runtime_get(ch);
|
|
if (ret < 0)
|
|
goto clean_mem;
|
|
|
|
cov->enabled = true;
|
|
|
|
/* Sync state with rtcpu */
|
|
camrtc_coverage_msg(cov, &resp, false, false);
|
|
|
|
dev_dbg(&ch->dev, "Falcon 0x%02x code coverage enabled.\n",
|
|
cov->id);
|
|
|
|
done:
|
|
return ret;
|
|
|
|
clean_mem:
|
|
dma_free_coherent(mem_dev, cov->mem.size, cov->mem.ptr, cov->mem.iova);
|
|
memset(&cov->mem, 0, sizeof(struct camrtc_test_mem));
|
|
cov->enabled = false;
|
|
|
|
error:
|
|
return ret;
|
|
}
|
|
|
|
static void camrtc_falcon_coverage_disable(struct camrtc_falcon_coverage *cov)
|
|
{
|
|
struct tegra_ivc_channel *ch = cov->ch;
|
|
struct device *mem_dev = cov->mem_dev;
|
|
struct device *falcon_dev = cov->falcon_dev;
|
|
struct camrtc_dbg_response resp;
|
|
|
|
if (!cov->enabled)
|
|
return;
|
|
|
|
/* Disable and sync with rtpcu */
|
|
cov->enabled = false;
|
|
camrtc_coverage_msg(cov, &resp, false, false);
|
|
|
|
if (cov->sgt.sgl) {
|
|
dma_unmap_sg(falcon_dev, cov->sgt.sgl,
|
|
cov->sgt.orig_nents, DMA_BIDIRECTIONAL);
|
|
sg_free_table(&cov->sgt);
|
|
}
|
|
|
|
if (cov->mem.ptr) {
|
|
dma_free_coherent(mem_dev, cov->mem.size,
|
|
cov->mem.ptr, cov->mem.iova);
|
|
memset(&cov->mem, 0, sizeof(struct camrtc_test_mem));
|
|
}
|
|
|
|
tegra_ivc_channel_runtime_put(ch);
|
|
}
|
|
|
|
static int camrtc_dbgfs_show_coverage_enable(void *data, u64 *val)
|
|
{
|
|
struct camrtc_falcon_coverage *cov = data;
|
|
|
|
*val = cov->enabled ? 1 : 0;
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int camrtc_dbgfs_store_coverage_enable(void *data, u64 val)
|
|
{
|
|
struct camrtc_falcon_coverage *cov = data;
|
|
bool enable = (val != 0) ? true : false;
|
|
int ret = 0;
|
|
|
|
if (cov->enabled != enable) {
|
|
if (enable)
|
|
ret = camrtc_falcon_coverage_enable(cov);
|
|
else
|
|
camrtc_falcon_coverage_disable(cov);
|
|
}
|
|
|
|
return ret;
|
|
}
|
|
|
|
DEFINE_SIMPLE_ATTRIBUTE(camrtc_dbgfs_fops_coverage_enable,
|
|
camrtc_dbgfs_show_coverage_enable,
|
|
camrtc_dbgfs_store_coverage_enable,
|
|
"%lld\n");
|
|
|
|
#define TEGRA_APS_AST_CONTROL 0x0
|
|
#define TEGRA_APS_AST_STREAMID_CTL 0x20
|
|
#define TEGRA_APS_AST_REGION_0_SLAVE_BASE_LO 0x100
|
|
#define TEGRA_APS_AST_REGION_0_SLAVE_BASE_HI 0x104
|
|
#define TEGRA_APS_AST_REGION_0_MASK_LO 0x108
|
|
#define TEGRA_APS_AST_REGION_0_MASK_HI 0x10c
|
|
#define TEGRA_APS_AST_REGION_0_MASTER_BASE_LO 0x110
|
|
#define TEGRA_APS_AST_REGION_0_MASTER_BASE_HI 0x114
|
|
#define TEGRA_APS_AST_REGION_0_CONTROL 0x118
|
|
|
|
#define TEGRA_APS_AST_REGION_STRIDE 0x20
|
|
|
|
#define AST_RGN_CTRL_VM_INDEX 15
|
|
#define AST_RGN_CTRL_SNOOP BIT(2)
|
|
|
|
#define AST_ADDR_MASK64 (~0xfffULL)
|
|
|
|
struct tegra_ast_region_info {
|
|
u8 enabled;
|
|
u8 lock;
|
|
u8 snoop;
|
|
u8 non_secure;
|
|
|
|
u8 ns_passthru;
|
|
u8 carveout_id;
|
|
u8 carveout_al;
|
|
u8 vpr_rd;
|
|
|
|
u8 vpr_wr;
|
|
u8 vpr_passthru;
|
|
u8 vm_index;
|
|
u8 physical;
|
|
|
|
u8 stream_id;
|
|
u8 stream_id_enabled;
|
|
u8 pad[2];
|
|
|
|
u64 slave;
|
|
u64 mask;
|
|
u64 master;
|
|
u32 control;
|
|
};
|
|
|
|
static void tegra_ast_get_region_info(void __iomem *base,
|
|
u32 region,
|
|
struct tegra_ast_region_info *info)
|
|
{
|
|
u32 offset = region * TEGRA_APS_AST_REGION_STRIDE;
|
|
u32 vmidx, stream_id, gcontrol, control;
|
|
u64 lo, hi;
|
|
|
|
control = readl(base + TEGRA_APS_AST_REGION_0_CONTROL + offset);
|
|
info->control = control;
|
|
|
|
info->lock = (control & BIT(0)) != 0;
|
|
info->snoop = (control & BIT(2)) != 0;
|
|
info->non_secure = (control & BIT(3)) != 0;
|
|
info->ns_passthru = (control & BIT(4)) != 0;
|
|
info->carveout_id = (control >> 5) & (0x1f);
|
|
info->carveout_al = (control >> 10) & 0x3;
|
|
info->vpr_rd = (control & BIT(12)) != 0;
|
|
info->vpr_wr = (control & BIT(13)) != 0;
|
|
info->vpr_passthru = (control & BIT(14)) != 0;
|
|
vmidx = (control >> AST_RGN_CTRL_VM_INDEX) & 0xf;
|
|
info->vm_index = vmidx;
|
|
info->physical = (control & BIT(19)) != 0;
|
|
|
|
if (info->physical) {
|
|
gcontrol = readl(base + TEGRA_APS_AST_CONTROL);
|
|
info->stream_id = (gcontrol >> 22) & 0x7F;
|
|
info->stream_id_enabled = 1;
|
|
} else {
|
|
stream_id = readl(base + TEGRA_APS_AST_STREAMID_CTL +
|
|
(4 * vmidx));
|
|
info->stream_id = (stream_id >> 8) & 0xFF;
|
|
info->stream_id_enabled = (stream_id & BIT(0)) != 0;
|
|
}
|
|
|
|
lo = readl(base + TEGRA_APS_AST_REGION_0_SLAVE_BASE_LO + offset);
|
|
hi = readl(base + TEGRA_APS_AST_REGION_0_SLAVE_BASE_HI + offset);
|
|
|
|
info->slave = ((hi << 32U) + lo) & AST_ADDR_MASK64;
|
|
info->enabled = (lo & BIT(0)) != 0;
|
|
|
|
hi = readl(base + TEGRA_APS_AST_REGION_0_MASK_HI + offset);
|
|
lo = readl(base + TEGRA_APS_AST_REGION_0_MASK_LO + offset);
|
|
|
|
info->mask = ((hi << 32) + lo) | ~AST_ADDR_MASK64;
|
|
|
|
hi = readl(base + TEGRA_APS_AST_REGION_0_MASTER_BASE_HI + offset);
|
|
lo = readl(base + TEGRA_APS_AST_REGION_0_MASTER_BASE_LO + offset);
|
|
|
|
info->master = ((hi << 32U) + lo) & AST_ADDR_MASK64;
|
|
}
|
|
|
|
static void __iomem *iomap_byname(struct device *dev, const char *name)
|
|
{
|
|
int index = of_property_match_string(dev->of_node, "reg-names", name);
|
|
if (index < 0)
|
|
return IOMEM_ERR_PTR(-ENOENT);
|
|
|
|
return of_iomap(dev->of_node, index);
|
|
}
|
|
|
|
static void camrtc_dbgfs_show_ast_region(struct seq_file *file,
|
|
void __iomem *base, u32 index)
|
|
{
|
|
struct tegra_ast_region_info info;
|
|
|
|
tegra_ast_get_region_info(base, index, &info);
|
|
|
|
seq_printf(file, "ast region %u %s\n", index,
|
|
info.enabled ? "enabled" : "disabled");
|
|
|
|
if (!info.enabled)
|
|
return;
|
|
|
|
seq_printf(file,
|
|
"\tslave=0x%llx\n"
|
|
"\tmaster=0x%llx\n"
|
|
"\tsize=0x%llx\n"
|
|
"\tlock=%u snoop=%u non_secure=%u ns_passthru=%u\n"
|
|
"\tcarveout_id=%u carveout_al=%u\n"
|
|
"\tvpr_rd=%u vpr_wr=%u vpr_passthru=%u\n"
|
|
"\tvm_index=%u physical=%u\n"
|
|
"\tstream_id=%u (enabled=%u)\n",
|
|
info.slave, info.master, info.mask + 1,
|
|
info.lock, info.snoop,
|
|
info.non_secure, info.ns_passthru,
|
|
info.carveout_id, info.carveout_al,
|
|
info.vpr_rd, info.vpr_wr, info.vpr_passthru,
|
|
info.vm_index, info.physical,
|
|
info.stream_id, info.stream_id_enabled);
|
|
}
|
|
|
|
struct camrtc_dbgfs_ast_node {
|
|
struct tegra_ivc_channel *ch;
|
|
const char *name;
|
|
uint8_t mask;
|
|
};
|
|
|
|
static int camrtc_dbgfs_show_ast(struct seq_file *file,
|
|
void *data)
|
|
{
|
|
struct camrtc_dbgfs_ast_node *node = file->private;
|
|
void __iomem *ast;
|
|
int i;
|
|
|
|
ast = iomap_byname(camrtc_get_device(node->ch), node->name);
|
|
if (ast == NULL)
|
|
return -ENOMEM;
|
|
|
|
for (i = 0; i <= 7; i++) {
|
|
if (!(node->mask & BIT(i)))
|
|
continue;
|
|
|
|
camrtc_dbgfs_show_ast_region(file, ast, i);
|
|
|
|
if (node->mask & (node->mask - 1)) /* are multiple bits set? */
|
|
seq_puts(file, "\n");
|
|
}
|
|
|
|
iounmap(ast);
|
|
return 0;
|
|
}
|
|
|
|
DEFINE_SEQ_FOPS(camrtc_dbgfs_fops_ast, camrtc_dbgfs_show_ast);
|
|
|
|
static const struct debugfs_reg32 ast_common_regs[] = {
|
|
{ .name = "control", 0x0 },
|
|
{ .name = "error_status", 0x4 },
|
|
{ .name = "error_addr_lo", 0x8 },
|
|
{ .name = "error_addr_h", 0xC },
|
|
{ .name = "streamid_ctl_0", 0x20 },
|
|
{ .name = "streamid_ctl_1", 0x24 },
|
|
{ .name = "streamid_ctl_2", 0x28 },
|
|
{ .name = "streamid_ctl_3", 0x2C },
|
|
{ .name = "streamid_ctl_4", 0x30 },
|
|
{ .name = "streamid_ctl_5", 0x34 },
|
|
{ .name = "streamid_ctl_6", 0x38 },
|
|
{ .name = "streamid_ctl_7", 0x3C },
|
|
{ .name = "streamid_ctl_8", 0x40 },
|
|
{ .name = "streamid_ctl_9", 0x44 },
|
|
{ .name = "streamid_ctl_10", 0x48 },
|
|
{ .name = "streamid_ctl_11", 0x4C },
|
|
{ .name = "streamid_ctl_12", 0x50 },
|
|
{ .name = "streamid_ctl_13", 0x54 },
|
|
{ .name = "streamid_ctl_14", 0x58 },
|
|
{ .name = "streamid_ctl_15", 0x5C },
|
|
{ .name = "write_block_status", 0x60 },
|
|
{ .name = "read_block_status", 0x64 },
|
|
};
|
|
|
|
static const struct debugfs_reg32 ast_region_regs[] = {
|
|
{ .name = "slave_lo", 0x100 },
|
|
{ .name = "slave_hi", 0x104 },
|
|
{ .name = "mask_lo", 0x108 },
|
|
{ .name = "mask_hi", 0x10C },
|
|
{ .name = "master_lo", 0x110 },
|
|
{ .name = "master_hi", 0x114 },
|
|
{ .name = "control", 0x118 },
|
|
};
|
|
|
|
static int ast_regset_create_files(struct tegra_ivc_channel *ch,
|
|
struct dentry *dir,
|
|
struct ast_regset *ars,
|
|
char const *ast_name)
|
|
{
|
|
void __iomem *base;
|
|
int i;
|
|
|
|
base = iomap_byname(camrtc_get_device(ch), ast_name);
|
|
if (IS_ERR_OR_NULL(base))
|
|
return -ENOMEM;
|
|
|
|
ars->common.base = base;
|
|
ars->common.regs = ast_common_regs;
|
|
ars->common.nregs = ARRAY_SIZE(ast_common_regs);
|
|
|
|
debugfs_create_regset32("regs-common", 0444, dir, &ars->common);
|
|
|
|
for (i = 0; i < ARRAY_SIZE(ars->region); i++) {
|
|
char name[16];
|
|
|
|
snprintf(name, sizeof(name), "regs-region%u", i);
|
|
|
|
ars->region[i].base = base + i * TEGRA_APS_AST_REGION_STRIDE;
|
|
ars->region[i].regs = ast_region_regs;
|
|
ars->region[i].nregs = ARRAY_SIZE(ast_region_regs);
|
|
|
|
debugfs_create_regset32(name, 0444, dir, &ars->region[i]);
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int camrtc_debug_populate(struct tegra_ivc_channel *ch)
|
|
{
|
|
struct camrtc_debug *crd = tegra_ivc_channel_get_drvdata(ch);
|
|
struct dentry *dir;
|
|
struct dentry *coverage;
|
|
struct dentry *vi;
|
|
struct dentry *isp;
|
|
struct camrtc_dbgfs_ast_node *ast_nodes;
|
|
unsigned int i, dma, region;
|
|
char const *name = "camrtc";
|
|
|
|
of_property_read_string(ch->dev.of_node, NV(debugfs), &name);
|
|
|
|
crd->root = dir = debugfs_create_dir(name, NULL);
|
|
if (dir == NULL)
|
|
return -ENOMEM;
|
|
|
|
coverage = debugfs_create_dir("coverage", dir);
|
|
if (coverage == NULL)
|
|
goto error;
|
|
vi = debugfs_create_dir("vi", coverage);
|
|
if (vi == NULL)
|
|
goto error;
|
|
isp = debugfs_create_dir("isp", coverage);
|
|
if (isp == NULL)
|
|
goto error;
|
|
if (!debugfs_create_file("data", 0600, vi,
|
|
&crd->vi_falc_coverage,
|
|
&camrtc_dbgfs_fops_falcon_coverage))
|
|
goto error;
|
|
if (!debugfs_create_file("enable", 0600, vi,
|
|
&crd->vi_falc_coverage,
|
|
&camrtc_dbgfs_fops_coverage_enable))
|
|
goto error;
|
|
if (!debugfs_create_file("data", 0600, isp,
|
|
&crd->isp_falc_coverage,
|
|
&camrtc_dbgfs_fops_falcon_coverage))
|
|
goto error;
|
|
if (!debugfs_create_file("enable", 0600, isp,
|
|
&crd->isp_falc_coverage,
|
|
&camrtc_dbgfs_fops_coverage_enable))
|
|
goto error;
|
|
|
|
if (!debugfs_create_file("version", 0444, dir, ch,
|
|
&camrtc_dbgfs_fops_version))
|
|
goto error;
|
|
if (!debugfs_create_file("reboot", 0400, dir, ch,
|
|
&camrtc_dbgfs_fops_reboot))
|
|
goto error;
|
|
if (!debugfs_create_file("ping", 0444, dir, ch,
|
|
&camrtc_dbgfs_fops_ping))
|
|
goto error;
|
|
if (!debugfs_create_file("sm-ping", 0444, dir, ch,
|
|
&camrtc_dbgfs_fops_sm_ping))
|
|
goto error;
|
|
if (!debugfs_create_file("log-level", 0644, dir, ch,
|
|
&camrtc_dbgfs_fops_loglevel))
|
|
goto error;
|
|
|
|
debugfs_create_u32("timeout", 0644, dir,
|
|
&crd->parameters.completion_timeout);
|
|
|
|
if (!debugfs_create_file("forced-reset-restore", 0400, dir, ch,
|
|
&camrtc_dbgfs_fops_forced_reset_restore))
|
|
goto error;
|
|
|
|
if (!debugfs_create_file("irqstat", 0444, dir, ch,
|
|
&camrtc_dbgfs_fops_irqstat))
|
|
goto error;
|
|
if (!debugfs_create_file("memstat", 0444, dir, ch,
|
|
&camrtc_dbgfs_fops_memstat))
|
|
goto error;
|
|
|
|
dir = debugfs_create_dir("mods", crd->root);
|
|
if (!dir)
|
|
goto error;
|
|
|
|
debugfs_create_u32("case", 0644, dir,
|
|
&crd->parameters.mods_case);
|
|
|
|
debugfs_create_u32("loops", 0644, dir,
|
|
&crd->parameters.mods_loops);
|
|
|
|
debugfs_create_x32("dma_channels", 0644, dir,
|
|
&crd->parameters.mods_dma_channels);
|
|
|
|
if (!debugfs_create_file("result", 0400, dir, ch,
|
|
&camrtc_dbgfs_fops_mods_result))
|
|
goto error;
|
|
|
|
dir = debugfs_create_dir("rtos", crd->root);
|
|
if (!dir)
|
|
goto error;
|
|
if (!debugfs_create_file("state", 0444, dir, ch,
|
|
&camrtc_dbgfs_fops_freertos_state))
|
|
goto error;
|
|
|
|
dir = debugfs_create_dir("test", crd->root);
|
|
if (!dir)
|
|
goto error;
|
|
if (!debugfs_create_file("available", 0444, dir, ch,
|
|
&camrtc_dbgfs_fops_test_list))
|
|
goto error;
|
|
if (!debugfs_create_file("case", 0644, dir, ch,
|
|
&camrtc_dbgfs_fops_test_case))
|
|
goto error;
|
|
if (!debugfs_create_file("result", 0400, dir, ch,
|
|
&camrtc_dbgfs_fops_test_result))
|
|
goto error;
|
|
|
|
debugfs_create_u32("timeout", 0644, dir,
|
|
&crd->parameters.test_timeout);
|
|
|
|
for (i = 0; i < ARRAY_SIZE(crd->mem); i++) {
|
|
char name[8];
|
|
|
|
crd->mem[i].index = i;
|
|
snprintf(name, sizeof(name), "mem%u", i);
|
|
if (!debugfs_create_file(name, 0644, dir,
|
|
&crd->mem[i], &camrtc_dbgfs_fops_test_mem))
|
|
goto error;
|
|
}
|
|
|
|
ast_nodes = devm_kzalloc(&ch->dev, 18 * sizeof(*ast_nodes),
|
|
GFP_KERNEL);
|
|
if (unlikely(ast_nodes == NULL))
|
|
goto error;
|
|
|
|
for (dma = 0; dma <= 1; dma++) {
|
|
const char *ast_name = dma ? "ast-dma" : "ast-cpu";
|
|
|
|
dir = debugfs_create_dir(ast_name, crd->root);
|
|
if (dir == NULL)
|
|
goto error;
|
|
|
|
ast_regset_create_files(ch, dir, &crd->ast_regsets[dma],
|
|
ast_name);
|
|
|
|
ast_nodes->ch = ch;
|
|
ast_nodes->name = ast_name;
|
|
ast_nodes->mask = 0xff;
|
|
|
|
if (!debugfs_create_file("all", 0444, dir, ast_nodes,
|
|
&camrtc_dbgfs_fops_ast))
|
|
goto error;
|
|
|
|
ast_nodes++;
|
|
|
|
for (region = 0; region < 8; region++) {
|
|
char name[8];
|
|
|
|
snprintf(name, sizeof name, "%u", region);
|
|
|
|
ast_nodes->ch = ch;
|
|
ast_nodes->name = ast_name;
|
|
ast_nodes->mask = BIT(region);
|
|
|
|
if (!debugfs_create_file(name, 0444, dir, ast_nodes,
|
|
&camrtc_dbgfs_fops_ast))
|
|
goto error;
|
|
|
|
ast_nodes++;
|
|
}
|
|
}
|
|
|
|
return 0;
|
|
error:
|
|
debugfs_remove_recursive(crd->root);
|
|
return -ENOMEM;
|
|
}
|
|
|
|
static struct device *camrtc_get_linked_device(
|
|
struct device *dev, char const *name, int index)
|
|
{
|
|
struct device_node *np;
|
|
struct platform_device *pdev;
|
|
|
|
np = of_parse_phandle(dev->of_node, name, index);
|
|
if (np == NULL)
|
|
return NULL;
|
|
|
|
pdev = of_find_device_by_node(np);
|
|
of_node_put(np);
|
|
|
|
if (pdev == NULL) {
|
|
dev_warn(dev, "%s[%u] node has no device\n", name, index);
|
|
return NULL;
|
|
}
|
|
|
|
return &pdev->dev;
|
|
}
|
|
|
|
static int camrtc_debug_probe(struct tegra_ivc_channel *ch)
|
|
{
|
|
struct device *dev = &ch->dev;
|
|
struct camrtc_debug *crd;
|
|
uint32_t bw;
|
|
|
|
BUG_ON(ch->ivc.frame_size < sizeof(struct camrtc_dbg_request));
|
|
BUG_ON(ch->ivc.frame_size < sizeof(struct camrtc_dbg_response));
|
|
|
|
crd = devm_kzalloc(dev, sizeof(*crd) + ch->ivc.frame_size, GFP_KERNEL);
|
|
if (unlikely(crd == NULL))
|
|
return -ENOMEM;
|
|
|
|
crd->channel = ch;
|
|
crd->parameters.test_case = (char *)(crd + 1);
|
|
crd->parameters.mods_case = CAMRTC_MODS_TEST_BASIC;
|
|
crd->parameters.mods_loops = 20;
|
|
crd->parameters.mods_dma_channels = 0;
|
|
|
|
if (of_property_read_u32(dev->of_node,
|
|
NV(ivc-timeout),
|
|
&crd->parameters.completion_timeout))
|
|
crd->parameters.completion_timeout = 50;
|
|
|
|
if (of_property_read_u32(dev->of_node,
|
|
NV(test-timeout),
|
|
&crd->parameters.test_timeout))
|
|
crd->parameters.test_timeout = 1000;
|
|
|
|
mutex_init(&crd->mutex);
|
|
init_waitqueue_head(&crd->waitq);
|
|
|
|
tegra_ivc_channel_set_drvdata(ch, crd);
|
|
|
|
crd->mem_devices[0] = camrtc_get_linked_device(dev, NV(mem-map), 0);
|
|
crd->mem_devices[1] = camrtc_get_linked_device(dev, NV(mem-map), 1);
|
|
crd->mem_devices[2] = camrtc_get_linked_device(dev, NV(mem-map), 2);
|
|
crd->mem_devices[3] = camrtc_get_linked_device(dev, NV(mem-map), 3);
|
|
|
|
crd->vi_falc_coverage.id = CAMRTC_DBG_FALCON_ID_VI;
|
|
crd->vi_falc_coverage.mem_dev = camrtc_dbgfs_memory_dev(crd);
|
|
crd->vi_falc_coverage.falcon_dev = crd->mem_devices[1];
|
|
crd->vi_falc_coverage.ch = ch;
|
|
|
|
crd->isp_falc_coverage.id = CAMRTC_DBG_FALCON_ID_ISP;
|
|
crd->isp_falc_coverage.mem_dev = crd->mem_devices[0];
|
|
crd->isp_falc_coverage.falcon_dev = crd->mem_devices[2];
|
|
crd->isp_falc_coverage.ch = ch;
|
|
|
|
if (of_property_read_u32(dev->of_node, NV(test-bw), &bw) == 0) {
|
|
crd->parameters.test_bw = bw;
|
|
|
|
dev_dbg(dev, "using emc bw %u for tests\n", bw);
|
|
}
|
|
|
|
if (crd->mem_devices[0] == NULL) {
|
|
dev_dbg(dev, "missing %s\n", NV(mem-map));
|
|
crd->mem_devices[0] = get_device(camrtc_get_device(ch));
|
|
}
|
|
|
|
crd->icc_path = devm_of_icc_get(crd->mem_devices[0], "write");
|
|
if (IS_ERR(crd->icc_path)) {
|
|
dev_err(dev, "failed to get icc path for rtcpu, err: %ld\n",
|
|
PTR_ERR(crd->icc_path));
|
|
crd->icc_path = NULL;
|
|
}
|
|
|
|
if (camrtc_debug_populate(ch))
|
|
return -ENOMEM;
|
|
|
|
return 0;
|
|
}
|
|
|
|
static void camrtc_debug_remove(struct tegra_ivc_channel *ch)
|
|
{
|
|
struct camrtc_debug *crd = tegra_ivc_channel_get_drvdata(ch);
|
|
int i;
|
|
struct device *mem_dev = camrtc_dbgfs_memory_dev(crd);
|
|
|
|
camrtc_falcon_coverage_disable(&crd->vi_falc_coverage);
|
|
camrtc_falcon_coverage_disable(&crd->isp_falc_coverage);
|
|
|
|
for (i = 0; i < ARRAY_SIZE(crd->mem); i++) {
|
|
struct camrtc_test_mem *mem = &crd->mem[i];
|
|
|
|
if (mem->size == 0)
|
|
continue;
|
|
|
|
dma_free_coherent(mem_dev, mem->size, mem->ptr, mem->iova);
|
|
memset(mem, 0, sizeof(*mem));
|
|
}
|
|
|
|
put_device(crd->mem_devices[0]);
|
|
put_device(crd->mem_devices[1]);
|
|
put_device(crd->mem_devices[2]);
|
|
put_device(crd->mem_devices[3]);
|
|
|
|
debugfs_remove_recursive(crd->root);
|
|
}
|
|
|
|
static const struct tegra_ivc_channel_ops tegra_ivc_channel_debug_ops = {
|
|
.probe = camrtc_debug_probe,
|
|
.remove = camrtc_debug_remove,
|
|
.notify = camrtc_debug_notify,
|
|
};
|
|
|
|
static const struct of_device_id camrtc_debug_of_match[] = {
|
|
{ .compatible = "nvidia,tegra186-camera-ivc-protocol-debug" },
|
|
{ },
|
|
};
|
|
MODULE_DEVICE_TABLE(of, camrtc_debug_of_match);
|
|
|
|
static struct tegra_ivc_driver camrtc_debug_driver = {
|
|
.driver = {
|
|
.owner = THIS_MODULE,
|
|
.bus = &tegra_ivc_bus_type,
|
|
.name = "tegra-camera-rtcpu-debugfs",
|
|
.of_match_table = camrtc_debug_of_match,
|
|
},
|
|
.dev_type = &tegra_ivc_channel_type,
|
|
.ops.channel = &tegra_ivc_channel_debug_ops,
|
|
};
|
|
tegra_ivc_subsys_driver_default(camrtc_debug_driver);
|
|
|
|
MODULE_DESCRIPTION("Debug Driver for Camera RTCPU");
|
|
MODULE_AUTHOR("Pekka Pessi <ppessi@nvidia.com>");
|
|
MODULE_LICENSE("GPL v2");
|