mirror of
git://nv-tegra.nvidia.com/linux-nvgpu.git
synced 2025-12-23 18:16:01 +03:00
We were not using the engine_type field in device info, and the code did not handle chained entries properly. The code assumed that first entry is for graphics and second for CE, which is not always true. Improve the code to go through all entries of device_info, and preserve values across entries until we reach the last entry. Only last entry triggers a write to fifo engine info. There can also be multiple engines with same type, so accumulate interrupts and reset ids from all of them. As the code got fixed, now it reads the engine enum correctly from hardware. We used to compare that against CE0, but we should compare against CE2. gk20a_fifo_reset_engine() uses wrong constants - it is passed a internal numbering of engines, but it compares them against hardware engine enum. Change-Id: Ia59273921c602d2a090f7a5b1404afb0fca2532c Signed-off-by: Terje Bergstrom <tbergstrom@nvidia.com> Reviewed-on: http://git-master/r/1147746 Reviewed-by: Konsta Holtta <kholtta@nvidia.com>
228 lines
6.2 KiB
C
228 lines
6.2 KiB
C
/*
|
|
* drivers/video/tegra/host/gk20a/fifo_gk20a.h
|
|
*
|
|
* GK20A graphics fifo (gr host)
|
|
*
|
|
* Copyright (c) 2011-2016, NVIDIA CORPORATION. All rights reserved.
|
|
*
|
|
* This program is free software; you can redistribute it and/or modify it
|
|
* under the terms and conditions of the GNU General Public License,
|
|
* version 2, as published by the Free Software Foundation.
|
|
*
|
|
* This program is distributed in the hope it will be useful, but WITHOUT
|
|
* ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
|
|
* FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for
|
|
* more details.
|
|
*
|
|
* You should have received a copy of the GNU General Public License along with
|
|
* this program; if not, write to the Free Software Foundation, Inc.,
|
|
* 51 Franklin St - Fifth Floor, Boston, MA 02110-1301 USA.
|
|
*/
|
|
#ifndef __FIFO_GK20A_H__
|
|
#define __FIFO_GK20A_H__
|
|
|
|
#include "channel_gk20a.h"
|
|
#include "tsg_gk20a.h"
|
|
|
|
#define MAX_RUNLIST_BUFFERS 2
|
|
|
|
/* generally corresponds to the "pbdma" engine */
|
|
|
|
struct fifo_runlist_info_gk20a {
|
|
unsigned long *active_channels;
|
|
unsigned long *active_tsgs;
|
|
/* Each engine has its own SW and HW runlist buffer.*/
|
|
struct mem_desc mem[MAX_RUNLIST_BUFFERS];
|
|
u32 cur_buffer;
|
|
u32 total_entries;
|
|
bool stopped;
|
|
bool support_tsg;
|
|
struct mutex mutex; /* protect channel preempt and runlist upate */
|
|
};
|
|
|
|
/* so far gk20a has two engines: gr and ce2(gr_copy) */
|
|
enum {
|
|
ENGINE_GR_GK20A = 0,
|
|
ENGINE_CE2_GK20A = 1,
|
|
ENGINE_INVAL_GK20A
|
|
};
|
|
|
|
struct fifo_pbdma_exception_info_gk20a {
|
|
u32 status_r; /* raw register value from hardware */
|
|
u32 id, next_id;
|
|
u32 chan_status_v; /* raw value from hardware */
|
|
bool id_is_chid, next_id_is_chid;
|
|
bool chsw_in_progress;
|
|
};
|
|
|
|
struct fifo_engine_exception_info_gk20a {
|
|
u32 status_r; /* raw register value from hardware */
|
|
u32 id, next_id;
|
|
u32 ctx_status_v; /* raw value from hardware */
|
|
bool id_is_chid, next_id_is_chid;
|
|
bool faulted, idle, ctxsw_in_progress;
|
|
};
|
|
|
|
struct fifo_mmu_fault_info_gk20a {
|
|
u32 fault_info_v;
|
|
u32 fault_type_v;
|
|
u32 engine_subid_v;
|
|
u32 client_v;
|
|
u32 fault_hi_v;
|
|
u32 fault_lo_v;
|
|
u64 inst_ptr;
|
|
const char *fault_type_desc;
|
|
const char *engine_subid_desc;
|
|
const char *client_desc;
|
|
};
|
|
|
|
struct fifo_engine_info_gk20a {
|
|
u32 engine_id;
|
|
u32 runlist_id;
|
|
u32 intr_mask;
|
|
u32 reset_mask;
|
|
u32 pbdma_id;
|
|
struct fifo_pbdma_exception_info_gk20a pbdma_exception_info;
|
|
struct fifo_engine_exception_info_gk20a engine_exception_info;
|
|
struct fifo_mmu_fault_info_gk20a mmu_fault_info;
|
|
|
|
};
|
|
|
|
struct fifo_gk20a {
|
|
struct gk20a *g;
|
|
int num_channels;
|
|
int num_runlist_entries;
|
|
|
|
int num_pbdma;
|
|
u32 *pbdma_map;
|
|
|
|
struct fifo_engine_info_gk20a *engine_info;
|
|
u32 max_engines;
|
|
u32 num_engines;
|
|
|
|
struct fifo_runlist_info_gk20a *runlist_info;
|
|
u32 max_runlists;
|
|
|
|
struct mem_desc userd;
|
|
u32 userd_entry_size;
|
|
|
|
int used_channels;
|
|
struct channel_gk20a *channel;
|
|
/* zero-kref'd channels here */
|
|
struct list_head free_chs;
|
|
struct mutex free_chs_mutex;
|
|
struct mutex gr_reset_mutex;
|
|
|
|
struct tsg_gk20a *tsg;
|
|
struct mutex tsg_inuse_mutex;
|
|
|
|
void (*remove_support)(struct fifo_gk20a *);
|
|
bool sw_ready;
|
|
struct {
|
|
/* share info between isrs and non-isr code */
|
|
struct {
|
|
struct mutex mutex;
|
|
} isr;
|
|
struct {
|
|
u32 device_fatal_0;
|
|
u32 channel_fatal_0;
|
|
u32 restartable_0;
|
|
} pbdma;
|
|
struct {
|
|
|
|
} engine;
|
|
|
|
|
|
} intr;
|
|
|
|
u32 deferred_fault_engines;
|
|
bool deferred_reset_pending;
|
|
struct mutex deferred_reset_mutex;
|
|
};
|
|
|
|
static inline const char *gk20a_fifo_interleave_level_name(u32 interleave_level)
|
|
{
|
|
switch (interleave_level) {
|
|
case NVGPU_RUNLIST_INTERLEAVE_LEVEL_LOW:
|
|
return "LOW";
|
|
|
|
case NVGPU_RUNLIST_INTERLEAVE_LEVEL_MEDIUM:
|
|
return "MEDIUM";
|
|
|
|
case NVGPU_RUNLIST_INTERLEAVE_LEVEL_HIGH:
|
|
return "HIGH";
|
|
|
|
default:
|
|
return "?";
|
|
}
|
|
}
|
|
|
|
int gk20a_init_fifo_support(struct gk20a *g);
|
|
|
|
void gk20a_fifo_isr(struct gk20a *g);
|
|
void gk20a_fifo_nonstall_isr(struct gk20a *g);
|
|
|
|
int gk20a_fifo_preempt_channel(struct gk20a *g, u32 hw_chid);
|
|
int gk20a_fifo_preempt_tsg(struct gk20a *g, u32 tsgid);
|
|
int gk20a_fifo_preempt(struct gk20a *g, struct channel_gk20a *ch);
|
|
|
|
int gk20a_fifo_enable_engine_activity(struct gk20a *g,
|
|
struct fifo_engine_info_gk20a *eng_info);
|
|
int gk20a_fifo_enable_all_engine_activity(struct gk20a *g);
|
|
int gk20a_fifo_disable_engine_activity(struct gk20a *g,
|
|
struct fifo_engine_info_gk20a *eng_info,
|
|
bool wait_for_idle);
|
|
int gk20a_fifo_disable_all_engine_activity(struct gk20a *g,
|
|
bool wait_for_idle);
|
|
u32 gk20a_fifo_engines_on_ch(struct gk20a *g, u32 hw_chid);
|
|
|
|
int gk20a_fifo_update_runlist(struct gk20a *g, u32 engine_id, u32 hw_chid,
|
|
bool add, bool wait_for_finish);
|
|
|
|
int gk20a_fifo_suspend(struct gk20a *g);
|
|
|
|
bool gk20a_fifo_mmu_fault_pending(struct gk20a *g);
|
|
|
|
void gk20a_fifo_recover(struct gk20a *g,
|
|
u32 engine_ids, /* if zero, will be queried from HW */
|
|
u32 hw_id, /* if ~0, will be queried from HW */
|
|
bool hw_id_is_tsg, /* ignored if hw_id == ~0 */
|
|
bool id_is_known, bool verbose);
|
|
void gk20a_fifo_recover_ch(struct gk20a *g, u32 hw_chid, bool verbose);
|
|
void gk20a_fifo_recover_tsg(struct gk20a *g, u32 tsgid, bool verbose);
|
|
int gk20a_fifo_force_reset_ch(struct channel_gk20a *ch, bool verbose);
|
|
void gk20a_fifo_reset_engine(struct gk20a *g, u32 engine_id);
|
|
int gk20a_init_fifo_reset_enable_hw(struct gk20a *g);
|
|
void gk20a_init_fifo(struct gpu_ops *gops);
|
|
|
|
void fifo_gk20a_finish_mmu_fault_handling(struct gk20a *g,
|
|
unsigned long fault_id);
|
|
int gk20a_fifo_wait_engine_idle(struct gk20a *g);
|
|
u32 gk20a_fifo_engine_interrupt_mask(struct gk20a *g);
|
|
u32 gk20a_fifo_get_pbdma_signature(struct gk20a *g);
|
|
u32 gk20a_fifo_get_failing_engine_data(struct gk20a *g,
|
|
int *__id, bool *__is_tsg);
|
|
bool gk20a_fifo_set_ctx_mmu_error_tsg(struct gk20a *g,
|
|
struct tsg_gk20a *tsg);
|
|
void gk20a_fifo_abort_tsg(struct gk20a *g, u32 tsgid);
|
|
bool gk20a_fifo_set_ctx_mmu_error_ch(struct gk20a *g,
|
|
struct channel_gk20a *ch);
|
|
|
|
struct channel_gk20a *gk20a_fifo_channel_from_hw_chid(struct gk20a *g,
|
|
u32 hw_chid);
|
|
|
|
void gk20a_fifo_issue_preempt(struct gk20a *g, u32 id, bool is_tsg);
|
|
int gk20a_fifo_set_runlist_interleave(struct gk20a *g,
|
|
u32 id,
|
|
bool is_tsg,
|
|
u32 runlist_id,
|
|
u32 new_level);
|
|
|
|
void gk20a_fifo_debugfs_init(struct device *dev);
|
|
|
|
const char *gk20a_fifo_interleave_level_name(u32 interleave_level);
|
|
|
|
int gk20a_fifo_engine_enum_from_type(struct gk20a *g, u32 engine_type);
|
|
|
|
#endif /*__GR_GK20A_H__*/
|