mirror of
git://nv-tegra.nvidia.com/linux-nvgpu.git
synced 2025-12-22 17:36:20 +03:00
Remove a global debugfs variable and instead save the allocator debugfs root node in the gk20a struct. Bug 1799159 Change-Id: If4eed34fa24775e962001e34840b334658f2321c Signed-off-by: Alex Waterman <alexw@nvidia.com> Reviewed-on: http://git-master/r/1225611 (cherry picked from commit 1908fde10bb1fb60ce898ea329f5a441a3e4297a) Reviewed-on: http://git-master/r/1242390 GVS: Gerrit_Virtual_Submit Reviewed-by: Terje Bergstrom <tbergstrom@nvidia.com>
443 lines
11 KiB
C
443 lines
11 KiB
C
/*
|
|
* Copyright (c) 2016, NVIDIA CORPORATION. All rights reserved.
|
|
*
|
|
* This program is free software; you can redistribute it and/or modify it
|
|
* under the terms and conditions of the GNU General Public License,
|
|
* version 2, as published by the Free Software Foundation.
|
|
*
|
|
* This program is distributed in the hope it will be useful, but WITHOUT
|
|
* ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
|
|
* FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for
|
|
* more details.
|
|
*
|
|
* You should have received a copy of the GNU General Public License
|
|
* along with this program. If not, see <http://www.gnu.org/licenses/>.
|
|
*/
|
|
|
|
#include <linux/kernel.h>
|
|
#include <linux/slab.h>
|
|
#include <linux/bitops.h>
|
|
|
|
#include "gk20a_allocator.h"
|
|
#include "bitmap_allocator_priv.h"
|
|
|
|
static struct kmem_cache *meta_data_cache; /* slab cache for meta data. */
|
|
static DEFINE_MUTEX(meta_data_cache_lock);
|
|
|
|
static u64 gk20a_bitmap_alloc_length(struct gk20a_allocator *a)
|
|
{
|
|
struct gk20a_bitmap_allocator *ba = a->priv;
|
|
|
|
return ba->length;
|
|
}
|
|
|
|
static u64 gk20a_bitmap_alloc_base(struct gk20a_allocator *a)
|
|
{
|
|
struct gk20a_bitmap_allocator *ba = a->priv;
|
|
|
|
return ba->base;
|
|
}
|
|
|
|
static int gk20a_bitmap_alloc_inited(struct gk20a_allocator *a)
|
|
{
|
|
struct gk20a_bitmap_allocator *ba = a->priv;
|
|
int inited = ba->inited;
|
|
|
|
rmb();
|
|
return inited;
|
|
}
|
|
|
|
static u64 gk20a_bitmap_alloc_end(struct gk20a_allocator *a)
|
|
{
|
|
struct gk20a_bitmap_allocator *ba = a->priv;
|
|
|
|
return ba->base + ba->length;
|
|
}
|
|
|
|
static u64 gk20a_bitmap_alloc_fixed(struct gk20a_allocator *__a,
|
|
u64 base, u64 len)
|
|
{
|
|
struct gk20a_bitmap_allocator *a = bitmap_allocator(__a);
|
|
u64 blks, offs, ret;
|
|
|
|
/* Compute the bit offset and make sure it's aligned to a block. */
|
|
offs = base >> a->blk_shift;
|
|
if (offs * a->blk_size != base)
|
|
return 0;
|
|
|
|
offs -= a->bit_offs;
|
|
|
|
blks = len >> a->blk_shift;
|
|
if (blks * a->blk_size != len)
|
|
blks++;
|
|
|
|
alloc_lock(__a);
|
|
|
|
/* Check if the space requested is already occupied. */
|
|
ret = bitmap_find_next_zero_area(a->bitmap, a->num_bits, offs, blks, 0);
|
|
if (ret != offs)
|
|
goto fail;
|
|
|
|
bitmap_set(a->bitmap, offs, blks);
|
|
|
|
a->bytes_alloced += blks * a->blk_size;
|
|
a->nr_fixed_allocs++;
|
|
alloc_unlock(__a);
|
|
|
|
alloc_dbg(__a, "Alloc-fixed 0x%-10llx 0x%-5llx [bits=0x%llx (%llu)]\n",
|
|
base, len, blks, blks);
|
|
return base;
|
|
|
|
fail:
|
|
alloc_unlock(__a);
|
|
alloc_dbg(__a, "Alloc-fixed failed! (0x%llx)\n", base);
|
|
return 0;
|
|
}
|
|
|
|
/*
|
|
* Two possibilities for this function: either we are freeing a fixed allocation
|
|
* or we are freeing a regular alloc but with GPU_ALLOC_NO_ALLOC_PAGE defined.
|
|
*
|
|
* Note: this function won't do much error checking. Thus you could really
|
|
* confuse the allocator if you misuse this function.
|
|
*/
|
|
static void gk20a_bitmap_free_fixed(struct gk20a_allocator *__a,
|
|
u64 base, u64 len)
|
|
{
|
|
struct gk20a_bitmap_allocator *a = bitmap_allocator(__a);
|
|
u64 blks, offs;
|
|
|
|
offs = base >> a->blk_shift;
|
|
if (WARN_ON(offs * a->blk_size != base))
|
|
return;
|
|
|
|
offs -= a->bit_offs;
|
|
|
|
blks = len >> a->blk_shift;
|
|
if (blks * a->blk_size != len)
|
|
blks++;
|
|
|
|
alloc_lock(__a);
|
|
bitmap_clear(a->bitmap, offs, blks);
|
|
a->bytes_freed += blks * a->blk_size;
|
|
alloc_unlock(__a);
|
|
|
|
alloc_dbg(__a, "Free-fixed 0x%-10llx 0x%-5llx [bits=0x%llx (%llu)]\n",
|
|
base, len, blks, blks);
|
|
}
|
|
|
|
/*
|
|
* Add the passed alloc to the tree of stored allocations.
|
|
*/
|
|
static void insert_alloc_metadata(struct gk20a_bitmap_allocator *a,
|
|
struct gk20a_bitmap_alloc *alloc)
|
|
{
|
|
struct rb_node **new = &a->allocs.rb_node;
|
|
struct rb_node *parent = NULL;
|
|
struct gk20a_bitmap_alloc *tmp;
|
|
|
|
while (*new) {
|
|
tmp = container_of(*new, struct gk20a_bitmap_alloc,
|
|
alloc_entry);
|
|
|
|
parent = *new;
|
|
if (alloc->base < tmp->base)
|
|
new = &((*new)->rb_left);
|
|
else if (alloc->base > tmp->base)
|
|
new = &((*new)->rb_right);
|
|
else {
|
|
WARN_ON("Duplicate entries in RB alloc tree!\n");
|
|
return;
|
|
}
|
|
}
|
|
|
|
rb_link_node(&alloc->alloc_entry, parent, new);
|
|
rb_insert_color(&alloc->alloc_entry, &a->allocs);
|
|
}
|
|
|
|
/*
|
|
* Find and remove meta-data from the outstanding allocations.
|
|
*/
|
|
static struct gk20a_bitmap_alloc *find_alloc_metadata(
|
|
struct gk20a_bitmap_allocator *a, u64 addr)
|
|
{
|
|
struct rb_node *node = a->allocs.rb_node;
|
|
struct gk20a_bitmap_alloc *alloc;
|
|
|
|
while (node) {
|
|
alloc = container_of(node, struct gk20a_bitmap_alloc,
|
|
alloc_entry);
|
|
|
|
if (addr < alloc->base)
|
|
node = node->rb_left;
|
|
else if (addr > alloc->base)
|
|
node = node->rb_right;
|
|
else
|
|
break;
|
|
}
|
|
|
|
if (!node)
|
|
return NULL;
|
|
|
|
rb_erase(node, &a->allocs);
|
|
|
|
return alloc;
|
|
}
|
|
|
|
/*
|
|
* Tree of alloc meta data stores the address of the alloc not the bit offset.
|
|
*/
|
|
static int __gk20a_bitmap_store_alloc(struct gk20a_bitmap_allocator *a,
|
|
u64 addr, u64 len)
|
|
{
|
|
struct gk20a_bitmap_alloc *alloc =
|
|
kmem_cache_alloc(meta_data_cache, GFP_KERNEL);
|
|
|
|
if (!alloc)
|
|
return -ENOMEM;
|
|
|
|
alloc->base = addr;
|
|
alloc->length = len;
|
|
|
|
insert_alloc_metadata(a, alloc);
|
|
|
|
return 0;
|
|
}
|
|
|
|
/*
|
|
* @len is in bytes. This routine will figure out the right number of bits to
|
|
* actually allocate. The return is the address in bytes as well.
|
|
*/
|
|
static u64 gk20a_bitmap_alloc(struct gk20a_allocator *__a, u64 len)
|
|
{
|
|
u64 blks, addr;
|
|
unsigned long offs, adjusted_offs, limit;
|
|
struct gk20a_bitmap_allocator *a = bitmap_allocator(__a);
|
|
|
|
blks = len >> a->blk_shift;
|
|
|
|
if (blks * a->blk_size != len)
|
|
blks++;
|
|
|
|
alloc_lock(__a);
|
|
|
|
/*
|
|
* First look from next_blk and onwards...
|
|
*/
|
|
offs = bitmap_find_next_zero_area(a->bitmap, a->num_bits,
|
|
a->next_blk, blks, 0);
|
|
if (offs >= a->num_bits) {
|
|
/*
|
|
* If that didn't work try the remaining area. Since there can
|
|
* be available space that spans across a->next_blk we need to
|
|
* search up to the first set bit after that.
|
|
*/
|
|
limit = find_next_bit(a->bitmap, a->num_bits, a->next_blk);
|
|
offs = bitmap_find_next_zero_area(a->bitmap, limit,
|
|
0, blks, 0);
|
|
if (offs >= a->next_blk)
|
|
goto fail;
|
|
}
|
|
|
|
bitmap_set(a->bitmap, offs, blks);
|
|
a->next_blk = offs + blks;
|
|
|
|
adjusted_offs = offs + a->bit_offs;
|
|
addr = ((u64)adjusted_offs) * a->blk_size;
|
|
|
|
/*
|
|
* Only do meta-data storage if we are allowed to allocate storage for
|
|
* that meta-data. The issue with using kmalloc() and friends is that
|
|
* in latency and success critical paths an alloc_page() call can either
|
|
* sleep for potentially a long time or, assuming GFP_ATOMIC, fail.
|
|
* Since we might not want either of these possibilities assume that the
|
|
* caller will keep what data it needs around to successfully free this
|
|
* allocation.
|
|
*/
|
|
if (!(a->flags & GPU_ALLOC_NO_ALLOC_PAGE) &&
|
|
__gk20a_bitmap_store_alloc(a, addr, blks * a->blk_size))
|
|
goto fail_reset_bitmap;
|
|
|
|
alloc_dbg(__a, "Alloc 0x%-10llx 0x%-5llx [bits=0x%llx (%llu)]\n",
|
|
addr, len, blks, blks);
|
|
|
|
a->nr_allocs++;
|
|
a->bytes_alloced += (blks * a->blk_size);
|
|
alloc_unlock(__a);
|
|
|
|
return addr;
|
|
|
|
fail_reset_bitmap:
|
|
bitmap_clear(a->bitmap, offs, blks);
|
|
fail:
|
|
a->next_blk = 0;
|
|
alloc_unlock(__a);
|
|
alloc_dbg(__a, "Alloc failed!\n");
|
|
return 0;
|
|
}
|
|
|
|
static void gk20a_bitmap_free(struct gk20a_allocator *__a, u64 addr)
|
|
{
|
|
struct gk20a_bitmap_allocator *a = bitmap_allocator(__a);
|
|
struct gk20a_bitmap_alloc *alloc = NULL;
|
|
u64 offs, adjusted_offs, blks;
|
|
|
|
alloc_lock(__a);
|
|
|
|
if (a->flags & GPU_ALLOC_NO_ALLOC_PAGE) {
|
|
WARN(1, "Using wrong free for NO_ALLOC_PAGE bitmap allocator");
|
|
goto done;
|
|
}
|
|
|
|
alloc = find_alloc_metadata(a, addr);
|
|
if (!alloc)
|
|
goto done;
|
|
|
|
/*
|
|
* Address comes from adjusted offset (i.e the bit offset with
|
|
* a->bit_offs added. So start with that and then work out the real
|
|
* offs into the bitmap.
|
|
*/
|
|
adjusted_offs = addr >> a->blk_shift;
|
|
offs = adjusted_offs - a->bit_offs;
|
|
blks = alloc->length >> a->blk_shift;
|
|
|
|
bitmap_clear(a->bitmap, offs, blks);
|
|
alloc_dbg(__a, "Free 0x%-10llx\n", addr);
|
|
|
|
a->bytes_freed += alloc->length;
|
|
|
|
done:
|
|
kfree(alloc);
|
|
alloc_unlock(__a);
|
|
}
|
|
|
|
static void gk20a_bitmap_alloc_destroy(struct gk20a_allocator *__a)
|
|
{
|
|
struct gk20a_bitmap_allocator *a = bitmap_allocator(__a);
|
|
struct gk20a_bitmap_alloc *alloc;
|
|
struct rb_node *node;
|
|
|
|
/*
|
|
* Kill any outstanding allocations.
|
|
*/
|
|
while ((node = rb_first(&a->allocs)) != NULL) {
|
|
alloc = container_of(node, struct gk20a_bitmap_alloc,
|
|
alloc_entry);
|
|
|
|
rb_erase(node, &a->allocs);
|
|
kfree(alloc);
|
|
}
|
|
|
|
kfree(a->bitmap);
|
|
kfree(a);
|
|
}
|
|
|
|
static void gk20a_bitmap_print_stats(struct gk20a_allocator *__a,
|
|
struct seq_file *s, int lock)
|
|
{
|
|
struct gk20a_bitmap_allocator *a = bitmap_allocator(__a);
|
|
|
|
__alloc_pstat(s, __a, "Bitmap allocator params:\n");
|
|
__alloc_pstat(s, __a, " start = 0x%llx\n", a->base);
|
|
__alloc_pstat(s, __a, " end = 0x%llx\n", a->base + a->length);
|
|
__alloc_pstat(s, __a, " blks = 0x%llx\n", a->num_bits);
|
|
|
|
/* Actual stats. */
|
|
__alloc_pstat(s, __a, "Stats:\n");
|
|
__alloc_pstat(s, __a, " Number allocs = 0x%llx\n", a->nr_allocs);
|
|
__alloc_pstat(s, __a, " Number fixed = 0x%llx\n", a->nr_fixed_allocs);
|
|
__alloc_pstat(s, __a, " Bytes alloced = 0x%llx\n", a->bytes_alloced);
|
|
__alloc_pstat(s, __a, " Bytes freed = 0x%llx\n", a->bytes_freed);
|
|
__alloc_pstat(s, __a, " Outstanding = 0x%llx\n",
|
|
a->bytes_alloced - a->bytes_freed);
|
|
}
|
|
|
|
static const struct gk20a_allocator_ops bitmap_ops = {
|
|
.alloc = gk20a_bitmap_alloc,
|
|
.free = gk20a_bitmap_free,
|
|
|
|
.alloc_fixed = gk20a_bitmap_alloc_fixed,
|
|
.free_fixed = gk20a_bitmap_free_fixed,
|
|
|
|
.base = gk20a_bitmap_alloc_base,
|
|
.length = gk20a_bitmap_alloc_length,
|
|
.end = gk20a_bitmap_alloc_end,
|
|
.inited = gk20a_bitmap_alloc_inited,
|
|
|
|
.fini = gk20a_bitmap_alloc_destroy,
|
|
|
|
.print_stats = gk20a_bitmap_print_stats,
|
|
};
|
|
|
|
|
|
int gk20a_bitmap_allocator_init(struct gk20a *g, struct gk20a_allocator *__a,
|
|
const char *name, u64 base, u64 length,
|
|
u64 blk_size, u64 flags)
|
|
{
|
|
int err;
|
|
struct gk20a_bitmap_allocator *a;
|
|
|
|
mutex_lock(&meta_data_cache_lock);
|
|
if (!meta_data_cache)
|
|
meta_data_cache = KMEM_CACHE(gk20a_bitmap_alloc, 0);
|
|
mutex_unlock(&meta_data_cache_lock);
|
|
|
|
if (!meta_data_cache)
|
|
return -ENOMEM;
|
|
|
|
if (WARN_ON(blk_size & (blk_size - 1)))
|
|
return -EINVAL;
|
|
|
|
/*
|
|
* blk_size must be a power-of-2; base length also need to be aligned
|
|
* to blk_size.
|
|
*/
|
|
if (blk_size & (blk_size - 1) ||
|
|
base & (blk_size - 1) || length & (blk_size - 1))
|
|
return -EINVAL;
|
|
|
|
if (base == 0) {
|
|
base = blk_size;
|
|
length -= blk_size;
|
|
}
|
|
|
|
a = kzalloc(sizeof(struct gk20a_bitmap_allocator), GFP_KERNEL);
|
|
if (!a)
|
|
return -ENOMEM;
|
|
|
|
err = __gk20a_alloc_common_init(__a, name, a, false, &bitmap_ops);
|
|
if (err)
|
|
goto fail;
|
|
|
|
a->base = base;
|
|
a->length = length;
|
|
a->blk_size = blk_size;
|
|
a->blk_shift = __ffs(a->blk_size);
|
|
a->num_bits = length >> a->blk_shift;
|
|
a->bit_offs = a->base >> a->blk_shift;
|
|
a->flags = flags;
|
|
|
|
a->bitmap = kcalloc(BITS_TO_LONGS(a->num_bits), sizeof(*a->bitmap),
|
|
GFP_KERNEL);
|
|
if (!a->bitmap)
|
|
goto fail;
|
|
|
|
wmb();
|
|
a->inited = true;
|
|
|
|
gk20a_init_alloc_debug(g, __a);
|
|
alloc_dbg(__a, "New allocator: type bitmap\n");
|
|
alloc_dbg(__a, " base 0x%llx\n", a->base);
|
|
alloc_dbg(__a, " bit_offs 0x%llx\n", a->bit_offs);
|
|
alloc_dbg(__a, " size 0x%llx\n", a->length);
|
|
alloc_dbg(__a, " blk_size 0x%llx\n", a->blk_size);
|
|
alloc_dbg(__a, " flags 0x%llx\n", a->flags);
|
|
|
|
return 0;
|
|
|
|
fail:
|
|
kfree(a);
|
|
return err;
|
|
}
|