Files
linux-nvgpu/drivers/gpu/nvgpu/common/mm/nvgpu_allocator.c
Alex Waterman d630f1d99f gpu: nvgpu: Unify the small and large page address spaces
The basic structure of this patch is to make the small page allocator
and the large page allocator into pointers (where they used to be just
structs). Then assign each of those pointers to the same actual
allocator since the buddy allocator has supported mixed page sizes
since its inception.

For the rest of the driver some changes had to be made in order to
actually support mixed pages in a single address space.

1. Unifying the allocation page size determination

   Since the allocation and map operations happen at distinct
   times both mapping and allocation of GVA space must agree
   on page size. This is because the allocation has to separate
   allocations into separate PDEs to avoid the necessity of
   supporting mixed PDEs.

   To this end a function __get_pte_size() was introduced which
   is used both by the balloc code and the core GPU MM code. It
   determines page size based only on the length of the mapping/
   allocation.

2. Fixed address allocation + page size

   Similar to regular mappings/GVA allocations fixed address
   mapping page size determination had to be modified. In the
   past the address of the mapping determined page size since
   the address space split was by address (low addresses were
   small pages, high addresses large pages). Since that is no
   longer the case the page size field in the reserve memory
   ioctl is now honored by the mapping code. When, for instance,
   CUDA makes a memory reservation it specifies small or large
   pages. When CUDA requests mappings to be made within that
   address range the page size is then looked up in the reserved
   memory struct.

   Fixed address reservations were also modified to now always
   allocate at a PDE granularity (64M or 128M depending on
   large page size. This prevents non-fixed allocations from
   ending up in the same PDE and causing kernel panics or GMMU
   faults.

3. The rest...

   The rest of the changes are just by products of the above.
   Lots of places required minor updates to use a pointer to
   the GVA allocator struct instead of the struct itself.

Lastly, this change is not truly complete. More work remains to be
done in order to fully remove the notion that there was such a thing
as separate address spaces for different page sizes. Basically after
this patch what remains is cleanup and proper documentation.

Bug 1396644
Bug 1729947

Change-Id: If51ab396a37ba16c69e434adb47edeef083dce57
Signed-off-by: Alex Waterman <alexw@nvidia.com>
Reviewed-on: http://git-master/r/1265300
GVS: Gerrit_Virtual_Submit
Reviewed-by: Terje Bergstrom <tbergstrom@nvidia.com>
2017-01-31 16:23:07 -08:00

214 lines
4.5 KiB
C

/*
* gk20a allocator
*
* Copyright (c) 2011-2016, NVIDIA CORPORATION. All rights reserved.
*
* This program is free software; you can redistribute it and/or modify it
* under the terms and conditions of the GNU General Public License,
* version 2, as published by the Free Software Foundation.
*
* This program is distributed in the hope it will be useful, but WITHOUT
* ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
* FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for
* more details.
*
* You should have received a copy of the GNU General Public License
* along with this program. If not, see <http://www.gnu.org/licenses/>.
*/
#include <linux/kernel.h>
#include <linux/slab.h>
#include <nvgpu/allocator.h>
#include "gk20a/gk20a.h"
#include "gk20a/mm_gk20a.h"
#include "gk20a/platform_gk20a.h"
u32 nvgpu_alloc_tracing_on;
u64 nvgpu_alloc_length(struct nvgpu_allocator *a)
{
if (a->ops->length)
return a->ops->length(a);
return 0;
}
u64 nvgpu_alloc_base(struct nvgpu_allocator *a)
{
if (a->ops->base)
return a->ops->base(a);
return 0;
}
u64 nvgpu_alloc_initialized(struct nvgpu_allocator *a)
{
if (!a->ops || !a->ops->inited)
return 0;
return a->ops->inited(a);
}
u64 nvgpu_alloc_end(struct nvgpu_allocator *a)
{
if (a->ops->end)
return a->ops->end(a);
return 0;
}
u64 nvgpu_alloc_space(struct nvgpu_allocator *a)
{
if (a->ops->space)
return a->ops->space(a);
return 0;
}
u64 nvgpu_alloc(struct nvgpu_allocator *a, u64 len)
{
return a->ops->alloc(a, len);
}
void nvgpu_free(struct nvgpu_allocator *a, u64 addr)
{
a->ops->free(a, addr);
}
u64 nvgpu_alloc_fixed(struct nvgpu_allocator *a, u64 base, u64 len,
u32 page_size)
{
if (a->ops->alloc_fixed)
return a->ops->alloc_fixed(a, base, len, page_size);
return 0;
}
void nvgpu_free_fixed(struct nvgpu_allocator *a, u64 base, u64 len)
{
/*
* If this operation is not defined for the allocator then just do
* nothing. The alternative would be to fall back on the regular
* free but that may be harmful in unexpected ways.
*/
if (a->ops->free_fixed)
a->ops->free_fixed(a, base, len);
}
int nvgpu_alloc_reserve_carveout(struct nvgpu_allocator *a,
struct nvgpu_alloc_carveout *co)
{
if (a->ops->reserve_carveout)
return a->ops->reserve_carveout(a, co);
return -ENODEV;
}
void nvgpu_alloc_release_carveout(struct nvgpu_allocator *a,
struct nvgpu_alloc_carveout *co)
{
if (a->ops->release_carveout)
a->ops->release_carveout(a, co);
}
void nvgpu_alloc_destroy(struct nvgpu_allocator *a)
{
a->ops->fini(a);
memset(a, 0, sizeof(*a));
}
/*
* Handle the common init stuff for a nvgpu_allocator.
*/
int __nvgpu_alloc_common_init(struct nvgpu_allocator *a,
const char *name, void *priv, bool dbg,
const struct nvgpu_allocator_ops *ops)
{
if (!ops)
return -EINVAL;
/*
* This is the bare minimum operations required for a sensible
* allocator.
*/
if (!ops->alloc || !ops->free || !ops->fini)
return -EINVAL;
a->ops = ops;
a->priv = priv;
a->debug = dbg;
mutex_init(&a->lock);
strlcpy(a->name, name, sizeof(a->name));
return 0;
}
void nvgpu_alloc_print_stats(struct nvgpu_allocator *__a,
struct seq_file *s, int lock)
{
__a->ops->print_stats(__a, s, lock);
}
#ifdef CONFIG_DEBUG_FS
static int __alloc_show(struct seq_file *s, void *unused)
{
struct nvgpu_allocator *a = s->private;
nvgpu_alloc_print_stats(a, s, 1);
return 0;
}
static int __alloc_open(struct inode *inode, struct file *file)
{
return single_open(file, __alloc_show, inode->i_private);
}
static const struct file_operations __alloc_fops = {
.open = __alloc_open,
.read = seq_read,
.llseek = seq_lseek,
.release = single_release,
};
#endif
void nvgpu_init_alloc_debug(struct gk20a *g, struct nvgpu_allocator *a)
{
#ifdef CONFIG_DEBUG_FS
if (!g->debugfs_allocators)
return;
a->debugfs_entry = debugfs_create_file(a->name, S_IRUGO,
g->debugfs_allocators,
a, &__alloc_fops);
#endif
}
void nvgpu_fini_alloc_debug(struct nvgpu_allocator *a)
{
#ifdef CONFIG_DEBUG_FS
if (!IS_ERR_OR_NULL(a->debugfs_entry))
debugfs_remove(a->debugfs_entry);
#endif
}
void nvgpu_alloc_debugfs_init(struct device *dev)
{
#ifdef CONFIG_DEBUG_FS
struct gk20a_platform *platform = dev_get_drvdata(dev);
struct dentry *gpu_root = platform->debugfs;
struct gk20a *g = get_gk20a(dev);
g->debugfs_allocators = debugfs_create_dir("allocators", gpu_root);
if (IS_ERR_OR_NULL(g->debugfs_allocators))
return;
debugfs_create_u32("tracing", 0664, g->debugfs_allocators,
&nvgpu_alloc_tracing_on);
#endif
}