Files
linux-nvgpu/drivers/gpu/nvgpu/common/sync/channel_sync.c
Seema Khowala 312f91f991 gpu: nvgpu: move fence_gk20a to common/fence
Move gk20a/fence_gk20a.c to common/fence/fence.c

Renamed
gk20a_fence_from_semaphore -> nvgpu_fence_from_semaphore
gk20a_fence_from_syncpt -> nvgpu_fence_from_syncpt
gk20a_alloc_fence_pool -> nvgpu_fence_pool_alloc
gk20a_free_fence_pool -> nvgpu_fence_pool_free
gk20a_alloc_fence -> nvgpu_fence_alloc
gk20a_init_fence -> nvgpu_fence_init
gk20a_fence_put -> nvgpu_fence_put
gk20a_fence_get -> nvgpu_fence_get
gk20a_fence_wait -> nvgpu_fence_wait
gk20a_fence_is_expired -> nvgpu_fence_is_expired
gk20a_fence_install_fd -> nvgpu_fence_install_fd
gk20a_fence_ops struct -> nvgpu_fence_ops struct
gk20a_fence struct -> nvgpu_fence_type struct

JIRA NVGPU-1982

Change-Id: Ife77b2c3c386ff4368683c78ca02f00c99cddb4b
Signed-off-by: Seema Khowala <seemaj@nvidia.com>
Reviewed-on: https://git-master.nvidia.com/r/2093002
Reviewed-by: mobile promotions <svcmobile_promotions@nvidia.com>
Tested-by: mobile promotions <svcmobile_promotions@nvidia.com>
2019-04-10 17:24:52 -07:00

119 lines
3.5 KiB
C

/*
* GK20A Channel Synchronization Abstraction
*
* Copyright (c) 2014-2019, NVIDIA CORPORATION. All rights reserved.
*
* Permission is hereby granted, free of charge, to any person obtaining a
* copy of this software and associated documentation files (the "Software"),
* to deal in the Software without restriction, including without limitation
* the rights to use, copy, modify, merge, publish, distribute, sublicense,
* and/or sell copies of the Software, and to permit persons to whom the
* Software is furnished to do so, subject to the following conditions:
*
* The above copyright notice and this permission notice shall be included in
* all copies or substantial portions of the Software.
*
* THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
* IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
* FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
* THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
* LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING
* FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER
* DEALINGS IN THE SOFTWARE.
*/
#include <nvgpu/kmem.h>
#include <nvgpu/log.h>
#include <nvgpu/atomic.h>
#include <nvgpu/bug.h>
#include <nvgpu/list.h>
#include <nvgpu/nvhost.h>
#include <nvgpu/gk20a.h>
#include <nvgpu/os_fence.h>
#include <nvgpu/os_fence_syncpts.h>
#include <nvgpu/os_fence_semas.h>
#include <nvgpu/channel.h>
#include <nvgpu/channel_sync.h>
#include <nvgpu/channel_sync_syncpt.h>
#include <nvgpu/channel_sync_semaphore.h>
#include <nvgpu/fence.h>
#include "channel_sync_priv.h"
#include "gk20a/mm_gk20a.h"
struct nvgpu_channel_sync *nvgpu_channel_sync_create(struct channel_gk20a *c,
bool user_managed)
{
if (nvgpu_has_syncpoints(c->g)) {
return nvgpu_channel_sync_syncpt_create(c, user_managed);
} else {
return nvgpu_channel_sync_semaphore_create(c, user_managed);
}
}
bool nvgpu_channel_sync_needs_os_fence_framework(struct gk20a *g)
{
return !nvgpu_has_syncpoints(g);
}
bool nvgpu_has_syncpoints(struct gk20a *g)
{
#ifdef CONFIG_TEGRA_GK20A_NVHOST
return nvgpu_is_enabled(g, NVGPU_HAS_SYNCPOINTS) &&
!g->disable_syncpoints;
#else
return false;
#endif
}
int nvgpu_channel_sync_wait_fence_fd(struct nvgpu_channel_sync *s, int fd,
struct priv_cmd_entry *entry, u32 max_wait_cmds)
{
return s->wait_fence_fd(s, fd, entry, max_wait_cmds);
}
int nvgpu_channel_sync_incr(struct nvgpu_channel_sync *s,
struct priv_cmd_entry *entry, struct nvgpu_fence_type *fence,
bool need_sync_fence, bool register_irq)
{
return s->incr(s, entry, fence, need_sync_fence, register_irq);
}
int nvgpu_channel_sync_incr_user(struct nvgpu_channel_sync *s,
int wait_fence_fd, struct priv_cmd_entry *entry,
struct nvgpu_fence_type *fence, bool wfi, bool need_sync_fence,
bool register_irq)
{
return s->incr_user(s, wait_fence_fd, entry, fence, wfi,
need_sync_fence, register_irq);
}
void nvgpu_channel_sync_set_min_eq_max(struct nvgpu_channel_sync *s)
{
s->set_min_eq_max(s);
}
void nvgpu_channel_sync_set_safe_state(struct nvgpu_channel_sync *s)
{
s->set_safe_state(s);
}
void nvgpu_channel_sync_destroy(struct nvgpu_channel_sync *sync,
bool set_safe_state)
{
if (set_safe_state) {
sync->set_safe_state(sync);
}
sync->destroy(sync);
}
void nvgpu_channel_sync_get_ref(struct nvgpu_channel_sync *s)
{
nvgpu_atomic_inc(&s->refcount);
}
bool nvgpu_channel_sync_put_ref_and_check(struct nvgpu_channel_sync *s)
{
return nvgpu_atomic_dec_and_test(&s->refcount);
}