mirror of
git://nv-tegra.nvidia.com/linux-nvgpu.git
synced 2025-12-23 09:57:08 +03:00
Protect callback races from spurious gk20a channel updates by testing if the channel update callback still exists when in the scheduled work (instead of only when scheduling the work to the queue), and by canceling the work when the channel is freed. Protect access to the callback and its data by accessing them together inside spinlock-protected regions. Bug 200051384 Change-Id: Ib4e1571c35f662195e1dec1e362df32ddc099eb3 Signed-off-by: Konsta Holtta <kholtta@nvidia.com> Reviewed-on: http://git-master/r/592026 Reviewed-by: Terje Bergstrom <tbergstrom@nvidia.com> Tested-by: Terje Bergstrom <tbergstrom@nvidia.com>
230 lines
6.3 KiB
C
230 lines
6.3 KiB
C
/*
|
|
* GK20A graphics channel
|
|
*
|
|
* Copyright (c) 2011-2014, NVIDIA CORPORATION. All rights reserved.
|
|
*
|
|
* This program is free software; you can redistribute it and/or modify it
|
|
* under the terms and conditions of the GNU General Public License,
|
|
* version 2, as published by the Free Software Foundation.
|
|
*
|
|
* This program is distributed in the hope it will be useful, but WITHOUT
|
|
* ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
|
|
* FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for
|
|
* more details.
|
|
*
|
|
* You should have received a copy of the GNU General Public License
|
|
* along with this program. If not, see <http://www.gnu.org/licenses/>.
|
|
*/
|
|
#ifndef CHANNEL_GK20A_H
|
|
#define CHANNEL_GK20A_H
|
|
|
|
#include <linux/log2.h>
|
|
#include <linux/slab.h>
|
|
#include <linux/wait.h>
|
|
#include <linux/mutex.h>
|
|
#include <uapi/linux/nvgpu.h>
|
|
#include <linux/poll.h>
|
|
#include <linux/spinlock.h>
|
|
|
|
struct gk20a;
|
|
struct gr_gk20a;
|
|
struct dbg_session_gk20a;
|
|
struct gk20a_fence;
|
|
|
|
#include "channel_sync_gk20a.h"
|
|
|
|
#include "mm_gk20a.h"
|
|
#include "gr_gk20a.h"
|
|
#include "fence_gk20a.h"
|
|
|
|
struct gpfifo {
|
|
u32 entry0;
|
|
u32 entry1;
|
|
};
|
|
|
|
struct notification {
|
|
struct {
|
|
u32 nanoseconds[2];
|
|
} timestamp;
|
|
u32 info32;
|
|
u16 info16;
|
|
u16 status;
|
|
};
|
|
|
|
struct fence {
|
|
u32 hw_chid;
|
|
u32 syncpt_val;
|
|
};
|
|
|
|
/* contexts associated with a channel */
|
|
struct channel_ctx_gk20a {
|
|
struct gr_ctx_desc *gr_ctx;
|
|
struct pm_ctx_desc pm_ctx;
|
|
struct patch_desc patch_ctx;
|
|
struct zcull_ctx_desc zcull_ctx;
|
|
u64 global_ctx_buffer_va[NR_GLOBAL_CTX_BUF_VA];
|
|
u64 global_ctx_buffer_size[NR_GLOBAL_CTX_BUF_VA];
|
|
bool global_ctx_buffer_mapped;
|
|
};
|
|
|
|
struct channel_gk20a_job {
|
|
struct mapped_buffer_node **mapped_buffers;
|
|
int num_mapped_buffers;
|
|
struct gk20a_fence *pre_fence;
|
|
struct gk20a_fence *post_fence;
|
|
struct list_head list;
|
|
};
|
|
|
|
struct channel_gk20a_poll_events {
|
|
struct mutex lock;
|
|
bool events_enabled;
|
|
int num_pending_events;
|
|
};
|
|
|
|
/* this is the priv element of struct nvhost_channel */
|
|
struct channel_gk20a {
|
|
struct gk20a *g;
|
|
bool in_use;
|
|
int hw_chid;
|
|
bool bound;
|
|
bool first_init;
|
|
bool vpr;
|
|
pid_t pid;
|
|
|
|
int tsgid;
|
|
struct list_head ch_entry; /* channel's entry in TSG */
|
|
|
|
struct list_head jobs;
|
|
struct mutex jobs_lock;
|
|
struct mutex submit_lock;
|
|
|
|
struct vm_gk20a *vm;
|
|
|
|
struct gpfifo_desc gpfifo;
|
|
|
|
struct channel_ctx_gk20a ch_ctx;
|
|
|
|
struct inst_desc inst_block;
|
|
struct mem_desc_sub ramfc;
|
|
|
|
void *userd_cpu_va;
|
|
u64 userd_iova;
|
|
u64 userd_gpu_va;
|
|
|
|
s32 num_objects;
|
|
u32 obj_class; /* we support only one obj per channel */
|
|
|
|
struct priv_cmd_queue priv_cmd_q;
|
|
|
|
wait_queue_head_t notifier_wq;
|
|
wait_queue_head_t semaphore_wq;
|
|
wait_queue_head_t submit_wq;
|
|
|
|
u32 timeout_accumulated_ms;
|
|
u32 timeout_gpfifo_get;
|
|
|
|
bool cmds_pending;
|
|
struct {
|
|
/* These fences should be accessed with submit_lock held. */
|
|
struct gk20a_fence *pre_fence;
|
|
struct gk20a_fence *post_fence;
|
|
} last_submit;
|
|
|
|
void (*remove_support)(struct channel_gk20a *);
|
|
#if defined(CONFIG_GK20A_CYCLE_STATS)
|
|
struct {
|
|
void *cyclestate_buffer;
|
|
u32 cyclestate_buffer_size;
|
|
struct dma_buf *cyclestate_buffer_handler;
|
|
struct mutex cyclestate_buffer_mutex;
|
|
} cyclestate;
|
|
#endif
|
|
struct mutex dbg_s_lock;
|
|
struct list_head dbg_s_list;
|
|
|
|
bool has_timedout;
|
|
u32 timeout_ms_max;
|
|
bool timeout_debug_dump;
|
|
|
|
struct dma_buf *error_notifier_ref;
|
|
struct nvgpu_notification *error_notifier;
|
|
void *error_notifier_va;
|
|
|
|
struct gk20a_channel_sync *sync;
|
|
|
|
#ifdef CONFIG_TEGRA_GR_VIRTUALIZATION
|
|
u64 virt_ctx;
|
|
#endif
|
|
|
|
/* event support */
|
|
struct channel_gk20a_poll_events poll_events;
|
|
|
|
/* signal channel owner via a callback, if set, in gk20a_channel_update
|
|
* via schedule_work */
|
|
void (*update_fn)(struct channel_gk20a *, void *);
|
|
void *update_fn_data;
|
|
spinlock_t update_fn_lock; /* make access to the two above atomic */
|
|
struct work_struct update_fn_work;
|
|
};
|
|
|
|
static inline bool gk20a_channel_as_bound(struct channel_gk20a *ch)
|
|
{
|
|
return !!ch->vm;
|
|
}
|
|
int channel_gk20a_commit_va(struct channel_gk20a *c);
|
|
int gk20a_init_channel_support(struct gk20a *, u32 chid);
|
|
void gk20a_free_channel(struct channel_gk20a *ch, bool finish);
|
|
bool gk20a_channel_update_and_check_timeout(struct channel_gk20a *ch,
|
|
u32 timeout_delta_ms);
|
|
void gk20a_disable_channel(struct channel_gk20a *ch,
|
|
bool wait_for_finish,
|
|
unsigned long finish_timeout);
|
|
void gk20a_channel_abort(struct channel_gk20a *ch);
|
|
int gk20a_channel_finish(struct channel_gk20a *ch, unsigned long timeout);
|
|
void gk20a_set_error_notifier(struct channel_gk20a *ch, __u32 error);
|
|
void gk20a_channel_semaphore_wakeup(struct gk20a *g);
|
|
int gk20a_channel_alloc_priv_cmdbuf(struct channel_gk20a *c, u32 size,
|
|
struct priv_cmd_entry **entry);
|
|
|
|
int gk20a_channel_suspend(struct gk20a *g);
|
|
int gk20a_channel_resume(struct gk20a *g);
|
|
|
|
/* Channel file operations */
|
|
int gk20a_channel_open(struct inode *inode, struct file *filp);
|
|
long gk20a_channel_ioctl(struct file *filp,
|
|
unsigned int cmd,
|
|
unsigned long arg);
|
|
int gk20a_channel_release(struct inode *inode, struct file *filp);
|
|
struct channel_gk20a *gk20a_get_channel_from_file(int fd);
|
|
void gk20a_channel_update(struct channel_gk20a *c, int nr_completed);
|
|
unsigned int gk20a_channel_poll(struct file *filep, poll_table *wait);
|
|
void gk20a_channel_event(struct channel_gk20a *ch);
|
|
|
|
void gk20a_init_channel(struct gpu_ops *gops);
|
|
|
|
int gk20a_wait_channel_idle(struct channel_gk20a *ch);
|
|
struct channel_gk20a *gk20a_open_new_channel(struct gk20a *g);
|
|
struct channel_gk20a *gk20a_open_new_channel_with_cb(struct gk20a *g,
|
|
void (*update_fn)(struct channel_gk20a *, void *),
|
|
void *update_fn_data);
|
|
void channel_gk20a_unbind(struct channel_gk20a *ch_gk20a);
|
|
|
|
int gk20a_submit_channel_gpfifo(struct channel_gk20a *c,
|
|
struct nvgpu_gpfifo *gpfifo,
|
|
u32 num_entries,
|
|
u32 flags,
|
|
struct nvgpu_fence *fence,
|
|
struct gk20a_fence **fence_out);
|
|
|
|
int gk20a_alloc_channel_gpfifo(struct channel_gk20a *c,
|
|
struct nvgpu_alloc_gpfifo_args *args);
|
|
|
|
void channel_gk20a_unbind(struct channel_gk20a *ch_gk20a);
|
|
void channel_gk20a_disable(struct channel_gk20a *ch);
|
|
int channel_gk20a_alloc_inst(struct gk20a *g, struct channel_gk20a *ch);
|
|
void channel_gk20a_free_inst(struct gk20a *g, struct channel_gk20a *ch);
|
|
int channel_gk20a_setup_ramfc(struct channel_gk20a *c,
|
|
u64 gpfifo_base, u32 gpfifo_entries);
|
|
void channel_gk20a_enable(struct channel_gk20a *ch);
|
|
#endif /* CHANNEL_GK20A_H */
|