mirror of
git://nv-tegra.nvidia.com/linux-nvgpu.git
synced 2025-12-22 09:12:24 +03:00
Add compatibility version to page table and dma mapping structure.
Bug 200438879
Change-Id: I04b4601f71ae2b3e75843f39f5445ecca2b16677
Signed-off-by: Preetham Chandru R <pchandru@nvidia.com>
Reviewed-on: https://git-master.nvidia.com/r/2029086
(cherry picked from commit 8bbbd09caa)
Reviewed-on: https://git-master.nvidia.com/r/2071427
Reviewed-by: Automatic_Commit_Validation_User
GVS: Gerrit_Virtual_Submit
Reviewed-by: Bibek Basu <bbasu@nvidia.com>
Reviewed-by: mobile promotions <svcmobile_promotions@nvidia.com>
Tested-by: mobile promotions <svcmobile_promotions@nvidia.com>
300 lines
7.4 KiB
C
300 lines
7.4 KiB
C
/*
|
|
* Copyright (c) 2018-2019, NVIDIA CORPORATION. All rights reserved.
|
|
*
|
|
* Permission is hereby granted, free of charge, to any person obtaining a
|
|
* copy of this software and associated documentation files (the "Software"),
|
|
* to deal in the Software without restriction, including without limitation
|
|
* the rights to use, copy, modify, merge, publish, distribute, sublicense,
|
|
* and/or sell copies of the Software, and to permit persons to whom the
|
|
* Software is furnished to do so, subject to the following conditions:
|
|
*
|
|
* The above copyright notice and this permission notice shall be included in
|
|
* all copies or substantial portions of the Software.
|
|
*
|
|
* THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
|
|
* IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
|
|
* FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
|
|
* THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
|
|
* LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING
|
|
* FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER
|
|
* DEALINGS IN THE SOFTWARE.
|
|
*/
|
|
|
|
#include <linux/slab.h>
|
|
#include <linux/nv-p2p.h>
|
|
|
|
static void nvidia_p2p_mn_release(struct mmu_notifier *mn,
|
|
struct mm_struct *mm)
|
|
{
|
|
struct nvidia_p2p_page_table *page_table = container_of(mn,
|
|
struct nvidia_p2p_page_table,
|
|
mn);
|
|
|
|
page_table->free_callback(page_table->data);
|
|
}
|
|
|
|
static void nvidia_p2p_mn_invl_range_start(struct mmu_notifier *mn,
|
|
struct mm_struct *mm, unsigned long start, unsigned long end)
|
|
{
|
|
struct nvidia_p2p_page_table *page_table = container_of(mn,
|
|
struct nvidia_p2p_page_table,
|
|
mn);
|
|
u64 vaddr = 0;
|
|
u64 size = 0;
|
|
|
|
vaddr = page_table->vaddr;
|
|
size = page_table->size;
|
|
|
|
if (vaddr >= start && vaddr <= end) {
|
|
mmu_notifier_unregister_no_release(&page_table->mn, page_table->mm);
|
|
page_table->free_callback(page_table->data);
|
|
}
|
|
}
|
|
|
|
static struct mmu_notifier_ops nvidia_p2p_mmu_ops = {
|
|
.release = nvidia_p2p_mn_release,
|
|
.invalidate_range_start = nvidia_p2p_mn_invl_range_start,
|
|
};
|
|
|
|
int nvidia_p2p_get_pages(u64 vaddr, u64 size,
|
|
struct nvidia_p2p_page_table **page_table,
|
|
void (*free_callback)(void *data), void *data)
|
|
{
|
|
int ret = 0;
|
|
int user_pages = 0;
|
|
int locked = 0;
|
|
int nr_pages = size >> PAGE_SHIFT;
|
|
struct page **pages;
|
|
|
|
if (nr_pages <= 0) {
|
|
return -EINVAL;
|
|
}
|
|
|
|
*page_table = kzalloc(sizeof(**page_table), GFP_KERNEL);
|
|
if (!*page_table) {
|
|
return -ENOMEM;
|
|
}
|
|
|
|
pages = kcalloc(nr_pages, sizeof(*pages), GFP_KERNEL);
|
|
if (!pages) {
|
|
ret = -ENOMEM;
|
|
goto free_page_table;
|
|
}
|
|
down_read(¤t->mm->mmap_sem);
|
|
locked = 1;
|
|
user_pages = get_user_pages_locked(vaddr & PAGE_MASK, nr_pages,
|
|
FOLL_WRITE | FOLL_FORCE,
|
|
pages, &locked);
|
|
up_read(¤t->mm->mmap_sem);
|
|
if (user_pages != nr_pages) {
|
|
ret = user_pages < 0 ? user_pages : -ENOMEM;
|
|
goto free_pages;
|
|
}
|
|
|
|
(*page_table)->version = NVIDIA_P2P_PAGE_TABLE_VERSION;
|
|
(*page_table)->pages = pages;
|
|
(*page_table)->entries = user_pages;
|
|
(*page_table)->page_size = NVIDIA_P2P_PAGE_SIZE_4KB;
|
|
(*page_table)->size = size;
|
|
|
|
(*page_table)->mn.ops = &nvidia_p2p_mmu_ops;
|
|
(*page_table)->mm = current->mm;
|
|
(*page_table)->free_callback = free_callback;
|
|
(*page_table)->data = data;
|
|
(*page_table)->vaddr = vaddr;
|
|
mutex_init(&(*page_table)->lock);
|
|
(*page_table)->mapped = NVIDIA_P2P_PINNED;
|
|
|
|
ret = mmu_notifier_register(&(*page_table)->mn, (*page_table)->mm);
|
|
if (ret) {
|
|
goto free_pages;
|
|
}
|
|
|
|
return 0;
|
|
free_pages:
|
|
while (--user_pages >= 0) {
|
|
put_page(pages[user_pages]);
|
|
}
|
|
kfree(pages);
|
|
free_page_table:
|
|
kfree(*page_table);
|
|
*page_table = NULL;
|
|
return ret;
|
|
}
|
|
EXPORT_SYMBOL(nvidia_p2p_get_pages);
|
|
|
|
int nvidia_p2p_put_pages(struct nvidia_p2p_page_table *page_table)
|
|
{
|
|
if (!page_table) {
|
|
return -EINVAL;
|
|
}
|
|
|
|
mmu_notifier_unregister(&page_table->mn, page_table->mm);
|
|
|
|
return 0;
|
|
}
|
|
EXPORT_SYMBOL(nvidia_p2p_put_pages);
|
|
|
|
int nvidia_p2p_free_page_table(struct nvidia_p2p_page_table *page_table)
|
|
{
|
|
int user_pages = 0;
|
|
struct page **pages = NULL;
|
|
|
|
if (!page_table) {
|
|
return 0;
|
|
}
|
|
|
|
mutex_lock(&page_table->lock);
|
|
|
|
if (page_table->mapped & NVIDIA_P2P_MAPPED) {
|
|
WARN(1, "Attempting to free unmapped pages");
|
|
}
|
|
|
|
if (page_table->mapped & NVIDIA_P2P_PINNED) {
|
|
pages = page_table->pages;
|
|
user_pages = page_table->entries;
|
|
|
|
while (--user_pages >= 0) {
|
|
put_page(pages[user_pages]);
|
|
}
|
|
|
|
kfree(pages);
|
|
page_table->mapped &= (u32)~NVIDIA_P2P_PINNED;
|
|
}
|
|
|
|
mutex_unlock(&page_table->lock);
|
|
|
|
return 0;
|
|
}
|
|
EXPORT_SYMBOL(nvidia_p2p_free_page_table);
|
|
|
|
int nvidia_p2p_dma_map_pages(struct device *dev,
|
|
struct nvidia_p2p_page_table *page_table,
|
|
struct nvidia_p2p_dma_mapping **dma_mapping,
|
|
enum dma_data_direction direction)
|
|
{
|
|
struct sg_table *sgt = NULL;
|
|
struct scatterlist *sg;
|
|
struct page **pages = NULL;
|
|
u32 nr_pages = 0;
|
|
int ret = 0;
|
|
int i, count;
|
|
|
|
if (!page_table) {
|
|
return -EINVAL;
|
|
}
|
|
|
|
mutex_lock(&page_table->lock);
|
|
|
|
pages = page_table->pages;
|
|
nr_pages = page_table->entries;
|
|
if (nr_pages <= 0) {
|
|
mutex_unlock(&page_table->lock);
|
|
return -EINVAL;
|
|
}
|
|
|
|
*dma_mapping = kzalloc(sizeof(**dma_mapping), GFP_KERNEL);
|
|
if (!*dma_mapping) {
|
|
mutex_unlock(&page_table->lock);
|
|
return -ENOMEM;
|
|
}
|
|
sgt = kzalloc(sizeof(*sgt), GFP_KERNEL);
|
|
if (!sgt) {
|
|
ret = -ENOMEM;
|
|
goto free_dma_mapping;
|
|
}
|
|
ret = sg_alloc_table_from_pages(sgt, pages,
|
|
nr_pages, 0, page_table->size, GFP_KERNEL);
|
|
if (ret) {
|
|
goto free_sgt;
|
|
}
|
|
|
|
(*dma_mapping)->version = NVIDIA_P2P_DMA_MAPPING_VERSION;
|
|
(*dma_mapping)->sgt = sgt;
|
|
(*dma_mapping)->dev = dev;
|
|
(*dma_mapping)->direction = direction;
|
|
(*dma_mapping)->page_table = page_table;
|
|
|
|
count = dma_map_sg(dev, sgt->sgl, sgt->nents, direction);
|
|
if (count < 1) {
|
|
goto free_sg_table;
|
|
}
|
|
|
|
(*dma_mapping)->entries = count;
|
|
|
|
(*dma_mapping)->hw_address = kcalloc(count, sizeof(u64), GFP_KERNEL);
|
|
if (!((*dma_mapping)->hw_address)) {
|
|
ret = -ENOMEM;
|
|
goto unmap_sg;
|
|
}
|
|
(*dma_mapping)->hw_len = kcalloc(count, sizeof(u64), GFP_KERNEL);
|
|
if (!((*dma_mapping)->hw_len)) {
|
|
ret = -ENOMEM;
|
|
goto free_hw_address;
|
|
}
|
|
|
|
for_each_sg(sgt->sgl, sg, count, i) {
|
|
(*dma_mapping)->hw_address[i] = sg_dma_address(sg);
|
|
(*dma_mapping)->hw_len[i] = sg_dma_len(sg);
|
|
}
|
|
(*dma_mapping)->page_table->mapped |= NVIDIA_P2P_MAPPED;
|
|
mutex_unlock(&page_table->lock);
|
|
|
|
return 0;
|
|
free_hw_address:
|
|
kfree((*dma_mapping)->hw_address);
|
|
unmap_sg:
|
|
dma_unmap_sg(dev, sgt->sgl,
|
|
sgt->nents, direction);
|
|
free_sg_table:
|
|
sg_free_table(sgt);
|
|
free_sgt:
|
|
kfree(sgt);
|
|
free_dma_mapping:
|
|
kfree(*dma_mapping);
|
|
*dma_mapping = NULL;
|
|
mutex_unlock(&page_table->lock);
|
|
|
|
return ret;
|
|
}
|
|
EXPORT_SYMBOL(nvidia_p2p_dma_map_pages);
|
|
|
|
int nvidia_p2p_dma_unmap_pages(struct nvidia_p2p_dma_mapping *dma_mapping)
|
|
{
|
|
struct nvidia_p2p_page_table *page_table = NULL;
|
|
|
|
if (!dma_mapping) {
|
|
return -EINVAL;
|
|
}
|
|
|
|
page_table = dma_mapping->page_table;
|
|
if (!page_table) {
|
|
return -EFAULT;
|
|
}
|
|
|
|
mutex_lock(&page_table->lock);
|
|
if (page_table->mapped & NVIDIA_P2P_MAPPED) {
|
|
kfree(dma_mapping->hw_len);
|
|
kfree(dma_mapping->hw_address);
|
|
if (dma_mapping->entries)
|
|
dma_unmap_sg(dma_mapping->dev,
|
|
dma_mapping->sgt->sgl,
|
|
dma_mapping->sgt->nents,
|
|
dma_mapping->direction);
|
|
sg_free_table(dma_mapping->sgt);
|
|
kfree(dma_mapping->sgt);
|
|
kfree(dma_mapping);
|
|
page_table->mapped &= (u32)~NVIDIA_P2P_MAPPED;
|
|
}
|
|
mutex_unlock(&page_table->lock);
|
|
|
|
return 0;
|
|
}
|
|
EXPORT_SYMBOL(nvidia_p2p_dma_unmap_pages);
|
|
|
|
int nvidia_p2p_free_dma_mapping(struct nvidia_p2p_dma_mapping *dma_mapping)
|
|
{
|
|
return nvidia_p2p_dma_unmap_pages(dma_mapping);
|
|
}
|
|
EXPORT_SYMBOL(nvidia_p2p_free_dma_mapping);
|