Files
linux-nvgpu/drivers/gpu/nvgpu/gk20a/fence_gk20a.c
Debarshi Dutta d0e4dfd6ef gpu: nvgpu: sync_framework cleanups
This patch deals with cleanups meant to make things simpler for the
upcoming os abstraction patches for the sync framework. This patch
causes some substantial changes which are listed out as follows.

1) sync_timeline is moved out of gk20a_fence into struct
nvgpu_channel_linux. New function pointers are created to facilitate os
independent methods for enabling/disabling timeline and are now named
as os_fence_framework. These function pointers are located in the struct
os_channel under struct gk20a.

2) construction of the channel_sync require nvgpu_finalize_poweron_linux()
to be invoked before invocations to nvgpu_init_mm_ce_context(). Hence,
these methods are now moved away from gk20a_finalize_poweron() and
invoked after nvgpu_finalize_poweron_linux().

3) sync_fence creation is now delinked from fence construction and move
to the channel_sync_gk20a's  channel_incr methods. These sync_fences are
mainly associated with post_fences.

4) In case userspace requires the sync_fences to be constructed, we
try to obtain an fd before the gk20a_channel_submit_gpfifo() instead of
trying to do that later. This is used to avoid potential after effects
of duplicate work submission due to failure to obtain an unused fd.

JIRA NVGPU-66

Change-Id: I42a3e4e2e692a113b1b36d2b48ab107ae4444dfa
Signed-off-by: Debarshi Dutta <ddutta@nvidia.com>
Reviewed-on: https://git-master.nvidia.com/r/1678400
Reviewed-by: mobile promotions <svcmobile_promotions@nvidia.com>
Tested-by: mobile promotions <svcmobile_promotions@nvidia.com>
2018-04-22 21:04:48 -07:00

320 lines
7.5 KiB
C

/*
* Copyright (c) 2014-2018, NVIDIA CORPORATION. All rights reserved.
*
* Permission is hereby granted, free of charge, to any person obtaining a
* copy of this software and associated documentation files (the "Software"),
* to deal in the Software without restriction, including without limitation
* the rights to use, copy, modify, merge, publish, distribute, sublicense,
* and/or sell copies of the Software, and to permit persons to whom the
* Software is furnished to do so, subject to the following conditions:
*
* The above copyright notice and this permission notice shall be included in
* all copies or substantial portions of the Software.
*
* THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
* IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
* FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
* THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
* LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING
* FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER
* DEALINGS IN THE SOFTWARE.
*/
#include "fence_gk20a.h"
#ifdef CONFIG_SYNC
#include <linux/file.h>
#include <linux/fs.h>
#endif
#include <nvgpu/semaphore.h>
#include <nvgpu/kmem.h>
#include <nvgpu/soc.h>
#include <nvgpu/nvhost.h>
#include <nvgpu/barrier.h>
#include "gk20a.h"
#include "channel_gk20a.h"
#include "sync_gk20a.h"
#ifdef CONFIG_SYNC
#include "../drivers/staging/android/sync.h"
#endif
struct gk20a_fence_ops {
int (*wait)(struct gk20a_fence *, long timeout);
bool (*is_expired)(struct gk20a_fence *);
void *(*free)(struct nvgpu_ref *);
};
static void gk20a_fence_free(struct nvgpu_ref *ref)
{
struct gk20a_fence *f =
container_of(ref, struct gk20a_fence, ref);
struct gk20a *g = f->g;
#ifdef CONFIG_SYNC
if (f->os_fence)
sync_fence_put(f->os_fence);
#endif
if (f->semaphore)
nvgpu_semaphore_put(f->semaphore);
if (f->allocator) {
if (nvgpu_alloc_initialized(f->allocator))
nvgpu_free(f->allocator, (u64)(uintptr_t)f);
} else
nvgpu_kfree(g, f);
}
void gk20a_fence_put(struct gk20a_fence *f)
{
if (f)
nvgpu_ref_put(&f->ref, gk20a_fence_free);
}
struct gk20a_fence *gk20a_fence_get(struct gk20a_fence *f)
{
if (f)
nvgpu_ref_get(&f->ref);
return f;
}
inline bool gk20a_fence_is_valid(struct gk20a_fence *f)
{
bool valid = f->valid;
nvgpu_smp_rmb();
return valid;
}
int gk20a_fence_install_fd(struct gk20a_fence *f, int fd)
{
#ifdef CONFIG_SYNC
if (!f || !gk20a_fence_is_valid(f) || !f->os_fence)
return -EINVAL;
sync_fence_get(f->os_fence);
sync_fence_install(f->os_fence, fd);
return 0;
#else
return -ENODEV;
#endif
}
int gk20a_fence_wait(struct gk20a *g, struct gk20a_fence *f,
unsigned long timeout)
{
if (f && gk20a_fence_is_valid(f)) {
if (!nvgpu_platform_is_silicon(g))
timeout = MAX_SCHEDULE_TIMEOUT;
return f->ops->wait(f, timeout);
}
return 0;
}
bool gk20a_fence_is_expired(struct gk20a_fence *f)
{
if (f && gk20a_fence_is_valid(f) && f->ops)
return f->ops->is_expired(f);
else
return true;
}
int gk20a_alloc_fence_pool(struct channel_gk20a *c, unsigned int count)
{
int err;
size_t size;
struct gk20a_fence *fence_pool = NULL;
size = sizeof(struct gk20a_fence);
if (count <= UINT_MAX / size) {
size = count * size;
fence_pool = nvgpu_vzalloc(c->g, size);
}
if (!fence_pool)
return -ENOMEM;
err = nvgpu_lockless_allocator_init(c->g, &c->fence_allocator,
"fence_pool", (size_t)fence_pool, size,
sizeof(struct gk20a_fence), 0);
if (err)
goto fail;
return 0;
fail:
nvgpu_vfree(c->g, fence_pool);
return err;
}
void gk20a_free_fence_pool(struct channel_gk20a *c)
{
if (nvgpu_alloc_initialized(&c->fence_allocator)) {
void *base = (void *)(uintptr_t)
nvgpu_alloc_base(&c->fence_allocator);
nvgpu_alloc_destroy(&c->fence_allocator);
nvgpu_vfree(c->g, base);
}
}
struct gk20a_fence *gk20a_alloc_fence(struct channel_gk20a *c)
{
struct gk20a_fence *fence = NULL;
if (channel_gk20a_is_prealloc_enabled(c)) {
if (nvgpu_alloc_initialized(&c->fence_allocator)) {
fence = (struct gk20a_fence *)(uintptr_t)
nvgpu_alloc(&c->fence_allocator,
sizeof(struct gk20a_fence));
/* clear the node and reset the allocator pointer */
if (fence) {
memset(fence, 0, sizeof(*fence));
fence->allocator = &c->fence_allocator;
}
}
} else
fence = nvgpu_kzalloc(c->g, sizeof(struct gk20a_fence));
if (fence) {
nvgpu_ref_init(&fence->ref);
fence->g = c->g;
}
return fence;
}
void gk20a_init_fence(struct gk20a_fence *f,
const struct gk20a_fence_ops *ops,
struct sync_fence *os_fence)
{
if (!f)
return;
f->ops = ops;
f->syncpt_id = -1;
f->semaphore = NULL;
f->os_fence = os_fence;
}
/* Fences that are backed by GPU semaphores: */
static int nvgpu_semaphore_fence_wait(struct gk20a_fence *f, long timeout)
{
if (!nvgpu_semaphore_is_acquired(f->semaphore))
return 0;
return NVGPU_COND_WAIT_INTERRUPTIBLE(
f->semaphore_wq,
!nvgpu_semaphore_is_acquired(f->semaphore),
timeout);
}
static bool nvgpu_semaphore_fence_is_expired(struct gk20a_fence *f)
{
return !nvgpu_semaphore_is_acquired(f->semaphore);
}
static const struct gk20a_fence_ops nvgpu_semaphore_fence_ops = {
.wait = &nvgpu_semaphore_fence_wait,
.is_expired = &nvgpu_semaphore_fence_is_expired,
};
/* This function takes ownership of the semaphore as well as the os_fence */
int gk20a_fence_from_semaphore(
struct gk20a_fence *fence_out,
struct nvgpu_semaphore *semaphore,
struct nvgpu_cond *semaphore_wq,
struct sync_fence *os_fence)
{
struct gk20a_fence *f = fence_out;
gk20a_init_fence(f, &nvgpu_semaphore_fence_ops, os_fence);
if (!f)
return -EINVAL;
f->semaphore = semaphore;
f->semaphore_wq = semaphore_wq;
/* commit previous writes before setting the valid flag */
nvgpu_smp_wmb();
f->valid = true;
return 0;
}
#ifdef CONFIG_TEGRA_GK20A_NVHOST
/* Fences that are backed by host1x syncpoints: */
static int gk20a_syncpt_fence_wait(struct gk20a_fence *f, long timeout)
{
return nvgpu_nvhost_syncpt_wait_timeout_ext(
f->nvhost_dev, f->syncpt_id, f->syncpt_value,
(u32)timeout, NULL, NULL);
}
static bool gk20a_syncpt_fence_is_expired(struct gk20a_fence *f)
{
/*
* In cases we don't register a notifier, we can't expect the
* syncpt value to be updated. For this case, we force a read
* of the value from HW, and then check for expiration.
*/
if (!nvgpu_nvhost_syncpt_is_expired_ext(f->nvhost_dev, f->syncpt_id,
f->syncpt_value)) {
u32 val;
if (!nvgpu_nvhost_syncpt_read_ext_check(f->nvhost_dev,
f->syncpt_id, &val)) {
return nvgpu_nvhost_syncpt_is_expired_ext(
f->nvhost_dev,
f->syncpt_id, f->syncpt_value);
}
}
return true;
}
static const struct gk20a_fence_ops gk20a_syncpt_fence_ops = {
.wait = &gk20a_syncpt_fence_wait,
.is_expired = &gk20a_syncpt_fence_is_expired,
};
/* This function takes the ownership of the os_fence */
int gk20a_fence_from_syncpt(
struct gk20a_fence *fence_out,
struct nvgpu_nvhost_dev *nvhost_dev,
u32 id, u32 value, struct sync_fence *os_fence)
{
struct gk20a_fence *f = fence_out;
gk20a_init_fence(f, &gk20a_syncpt_fence_ops, os_fence);
if (!f)
return -EINVAL;
f->nvhost_dev = nvhost_dev;
f->syncpt_id = id;
f->syncpt_value = value;
/* commit previous writes before setting the valid flag */
nvgpu_smp_wmb();
f->valid = true;
return 0;
}
#else
int gk20a_fence_from_syncpt(
struct gk20a_fence *fence_out,
struct nvgpu_nvhost_dev *nvhost_dev,
u32 id, u32 value, struct sync_fence *os_fence)
{
return -EINVAL;
}
#endif