open-gpu-kernel-modules/kernel-open/nvidia-uvm/uvm_pmm_sysmem.c
Andy Ritger 758b4ee818
525.53
2022-11-10 08:39:33 -08:00

1461 lines
53 KiB
C

/*******************************************************************************
Copyright (c) 2017-2021 NVIDIA Corporation
Permission is hereby granted, free of charge, to any person obtaining a copy
of this software and associated documentation files (the "Software"), to
deal in the Software without restriction, including without limitation the
rights to use, copy, modify, merge, publish, distribute, sublicense, and/or
sell copies of the Software, and to permit persons to whom the Software is
furnished to do so, subject to the following conditions:
The above copyright notice and this permission notice shall be
included in all copies or substantial portions of the Software.
THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING
FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER
DEALINGS IN THE SOFTWARE.
*******************************************************************************/
#include "uvm_gpu.h"
#include "uvm_pmm_sysmem.h"
#include "uvm_kvmalloc.h"
#include "uvm_va_block.h"
#include "uvm_va_space.h"
static int uvm_cpu_chunk_allocation_sizes = UVM_CPU_CHUNK_SIZES;
module_param(uvm_cpu_chunk_allocation_sizes, uint, S_IRUGO | S_IWUSR);
MODULE_PARM_DESC(uvm_cpu_chunk_allocation_sizes, "OR'ed value of all CPU chunk allocation sizes.");
static struct kmem_cache *g_reverse_page_map_cache __read_mostly;
NV_STATUS uvm_pmm_sysmem_init(void)
{
g_reverse_page_map_cache = NV_KMEM_CACHE_CREATE("uvm_pmm_sysmem_page_reverse_map_t",
uvm_reverse_map_t);
if (!g_reverse_page_map_cache)
return NV_ERR_NO_MEMORY;
return NV_OK;
}
void uvm_pmm_sysmem_exit(void)
{
kmem_cache_destroy_safe(&g_reverse_page_map_cache);
}
NV_STATUS uvm_pmm_sysmem_mappings_init(uvm_gpu_t *gpu, uvm_pmm_sysmem_mappings_t *sysmem_mappings)
{
memset(sysmem_mappings, 0, sizeof(*sysmem_mappings));
sysmem_mappings->gpu = gpu;
uvm_mutex_init(&sysmem_mappings->reverse_map_lock, UVM_LOCK_ORDER_LEAF);
uvm_init_radix_tree_preloadable(&sysmem_mappings->reverse_map_tree);
return NV_OK;
}
void uvm_pmm_sysmem_mappings_deinit(uvm_pmm_sysmem_mappings_t *sysmem_mappings)
{
if (sysmem_mappings->gpu) {
UVM_ASSERT_MSG(radix_tree_empty(&sysmem_mappings->reverse_map_tree),
"radix_tree not empty for GPU %s\n",
uvm_gpu_name(sysmem_mappings->gpu));
}
sysmem_mappings->gpu = NULL;
}
// TODO: Bug 1995015: use a more efficient data structure for
// physically-contiguous allocations.
NV_STATUS uvm_pmm_sysmem_mappings_add_gpu_mapping(uvm_pmm_sysmem_mappings_t *sysmem_mappings,
NvU64 dma_addr,
NvU64 virt_addr,
NvU64 region_size,
uvm_va_block_t *va_block,
uvm_processor_id_t owner)
{
NV_STATUS status = NV_OK;
uvm_reverse_map_t *new_reverse_map;
NvU64 key;
const NvU64 base_key = dma_addr / PAGE_SIZE;
const NvU32 num_pages = region_size / PAGE_SIZE;
uvm_page_index_t page_index;
UVM_ASSERT(va_block);
UVM_ASSERT(!uvm_va_block_is_dead(va_block));
UVM_ASSERT(IS_ALIGNED(dma_addr, region_size));
UVM_ASSERT(IS_ALIGNED(virt_addr, region_size));
UVM_ASSERT(region_size <= UVM_VA_BLOCK_SIZE);
UVM_ASSERT(is_power_of_2(region_size));
UVM_ASSERT(uvm_va_block_contains_address(va_block, virt_addr));
UVM_ASSERT(uvm_va_block_contains_address(va_block, virt_addr + region_size - 1));
uvm_assert_mutex_locked(&va_block->lock);
if (!sysmem_mappings->gpu->parent->access_counters_supported)
return NV_OK;
new_reverse_map = nv_kmem_cache_zalloc(g_reverse_page_map_cache, NV_UVM_GFP_FLAGS);
if (!new_reverse_map)
return NV_ERR_NO_MEMORY;
page_index = uvm_va_block_cpu_page_index(va_block, virt_addr);
new_reverse_map->va_block = va_block;
new_reverse_map->region = uvm_va_block_region(page_index, page_index + num_pages);
new_reverse_map->owner = owner;
uvm_mutex_lock(&sysmem_mappings->reverse_map_lock);
for (key = base_key; key < base_key + num_pages; ++key) {
int ret = radix_tree_insert(&sysmem_mappings->reverse_map_tree, key, new_reverse_map);
if (ret != 0) {
NvU64 remove_key;
for (remove_key = base_key; remove_key < key; ++remove_key)
(void *)radix_tree_delete(&sysmem_mappings->reverse_map_tree, remove_key);
kmem_cache_free(g_reverse_page_map_cache, new_reverse_map);
status = errno_to_nv_status(ret);
break;
}
}
uvm_mutex_unlock(&sysmem_mappings->reverse_map_lock);
// The assert is added for Coverity's sake. It is equivalent to adding
// assert(num_pages > 0) before the loop. However, Coverity is not able to
// deduce that the loop has to execute at least once from num_pages > 0.
UVM_ASSERT(key != base_key || status != NV_OK);
return status;
}
static void pmm_sysmem_mappings_remove_gpu_mapping(uvm_pmm_sysmem_mappings_t *sysmem_mappings,
NvU64 dma_addr,
bool check_mapping)
{
uvm_reverse_map_t *reverse_map;
NvU64 key;
const NvU64 base_key = dma_addr / PAGE_SIZE;
if (!sysmem_mappings->gpu->parent->access_counters_supported)
return;
uvm_mutex_lock(&sysmem_mappings->reverse_map_lock);
reverse_map = radix_tree_delete(&sysmem_mappings->reverse_map_tree, base_key);
if (check_mapping)
UVM_ASSERT(reverse_map);
if (!reverse_map) {
uvm_mutex_unlock(&sysmem_mappings->reverse_map_lock);
return;
}
uvm_assert_mutex_locked(&reverse_map->va_block->lock);
for (key = base_key + 1; key < base_key + uvm_va_block_region_num_pages(reverse_map->region); ++key) {
uvm_reverse_map_t *curr_reverse_map = radix_tree_delete(&sysmem_mappings->reverse_map_tree, key);
UVM_ASSERT(curr_reverse_map == reverse_map);
}
uvm_mutex_unlock(&sysmem_mappings->reverse_map_lock);
kmem_cache_free(g_reverse_page_map_cache, reverse_map);
}
void uvm_pmm_sysmem_mappings_remove_gpu_mapping(uvm_pmm_sysmem_mappings_t *sysmem_mappings, NvU64 dma_addr)
{
pmm_sysmem_mappings_remove_gpu_mapping(sysmem_mappings, dma_addr, true);
}
void uvm_pmm_sysmem_mappings_remove_gpu_mapping_on_eviction(uvm_pmm_sysmem_mappings_t *sysmem_mappings, NvU64 dma_addr)
{
pmm_sysmem_mappings_remove_gpu_mapping(sysmem_mappings, dma_addr, false);
}
void uvm_pmm_sysmem_mappings_reparent_gpu_mapping(uvm_pmm_sysmem_mappings_t *sysmem_mappings,
NvU64 dma_addr,
uvm_va_block_t *va_block)
{
NvU64 virt_addr;
uvm_reverse_map_t *reverse_map;
const NvU64 base_key = dma_addr / PAGE_SIZE;
uvm_page_index_t new_start_page;
UVM_ASSERT(PAGE_ALIGNED(dma_addr));
UVM_ASSERT(va_block);
UVM_ASSERT(!uvm_va_block_is_dead(va_block));
if (!sysmem_mappings->gpu->parent->access_counters_supported)
return;
uvm_mutex_lock(&sysmem_mappings->reverse_map_lock);
reverse_map = radix_tree_lookup(&sysmem_mappings->reverse_map_tree, base_key);
UVM_ASSERT(reverse_map);
// Compute virt address by hand since the old VA block may be messed up
// during split
virt_addr = reverse_map->va_block->start + reverse_map->region.first * PAGE_SIZE;
new_start_page = uvm_va_block_cpu_page_index(va_block, virt_addr);
reverse_map->region = uvm_va_block_region(new_start_page,
new_start_page + uvm_va_block_region_num_pages(reverse_map->region));
reverse_map->va_block = va_block;
UVM_ASSERT(uvm_va_block_contains_address(va_block, uvm_reverse_map_start(reverse_map)));
UVM_ASSERT(uvm_va_block_contains_address(va_block, uvm_reverse_map_end(reverse_map)));
uvm_mutex_unlock(&sysmem_mappings->reverse_map_lock);
}
NV_STATUS uvm_pmm_sysmem_mappings_split_gpu_mappings(uvm_pmm_sysmem_mappings_t *sysmem_mappings,
NvU64 dma_addr,
NvU64 new_region_size)
{
uvm_reverse_map_t *orig_reverse_map;
const NvU64 base_key = dma_addr / PAGE_SIZE;
const size_t num_pages = new_region_size / PAGE_SIZE;
size_t old_num_pages;
size_t subregion, num_subregions;
uvm_reverse_map_t **new_reverse_maps;
UVM_ASSERT(IS_ALIGNED(dma_addr, new_region_size));
UVM_ASSERT(new_region_size <= UVM_VA_BLOCK_SIZE);
UVM_ASSERT(is_power_of_2(new_region_size));
if (!sysmem_mappings->gpu->parent->access_counters_supported)
return NV_OK;
uvm_mutex_lock(&sysmem_mappings->reverse_map_lock);
orig_reverse_map = radix_tree_lookup(&sysmem_mappings->reverse_map_tree, base_key);
uvm_mutex_unlock(&sysmem_mappings->reverse_map_lock);
// We can access orig_reverse_map outside the tree lock because we hold the
// VA block lock so we cannot have concurrent modifications in the tree for
// the mappings of the chunks that belong to that VA block.
UVM_ASSERT(orig_reverse_map);
UVM_ASSERT(orig_reverse_map->va_block);
uvm_assert_mutex_locked(&orig_reverse_map->va_block->lock);
old_num_pages = uvm_va_block_region_num_pages(orig_reverse_map->region);
UVM_ASSERT(num_pages < old_num_pages);
num_subregions = old_num_pages / num_pages;
new_reverse_maps = uvm_kvmalloc_zero(sizeof(*new_reverse_maps) * (num_subregions - 1));
if (!new_reverse_maps)
return NV_ERR_NO_MEMORY;
// Allocate the descriptors for the new subregions
for (subregion = 1; subregion < num_subregions; ++subregion) {
uvm_reverse_map_t *new_reverse_map = nv_kmem_cache_zalloc(g_reverse_page_map_cache, NV_UVM_GFP_FLAGS);
uvm_page_index_t page_index = orig_reverse_map->region.first + num_pages * subregion;
if (new_reverse_map == NULL) {
// On error, free the previously-created descriptors
while (--subregion != 0)
kmem_cache_free(g_reverse_page_map_cache, new_reverse_maps[subregion - 1]);
uvm_kvfree(new_reverse_maps);
return NV_ERR_NO_MEMORY;
}
new_reverse_map->va_block = orig_reverse_map->va_block;
new_reverse_map->region = uvm_va_block_region(page_index, page_index + num_pages);
new_reverse_map->owner = orig_reverse_map->owner;
new_reverse_maps[subregion - 1] = new_reverse_map;
}
uvm_mutex_lock(&sysmem_mappings->reverse_map_lock);
for (subregion = 1; subregion < num_subregions; ++subregion) {
NvU64 key;
for (key = base_key + num_pages * subregion; key < base_key + num_pages * (subregion + 1); ++key) {
void **slot = radix_tree_lookup_slot(&sysmem_mappings->reverse_map_tree, key);
UVM_ASSERT(slot);
UVM_ASSERT(radix_tree_deref_slot(slot) == orig_reverse_map);
NV_RADIX_TREE_REPLACE_SLOT(&sysmem_mappings->reverse_map_tree, slot, new_reverse_maps[subregion - 1]);
}
}
orig_reverse_map->region = uvm_va_block_region(orig_reverse_map->region.first,
orig_reverse_map->region.first + num_pages);
uvm_mutex_unlock(&sysmem_mappings->reverse_map_lock);
uvm_kvfree(new_reverse_maps);
return NV_OK;
}
void uvm_pmm_sysmem_mappings_merge_gpu_mappings(uvm_pmm_sysmem_mappings_t *sysmem_mappings,
NvU64 dma_addr,
NvU64 new_region_size)
{
uvm_reverse_map_t *first_reverse_map;
uvm_page_index_t running_page_index;
NvU64 key;
const NvU64 base_key = dma_addr / PAGE_SIZE;
const size_t num_pages = new_region_size / PAGE_SIZE;
size_t num_mapping_pages;
UVM_ASSERT(IS_ALIGNED(dma_addr, new_region_size));
UVM_ASSERT(new_region_size <= UVM_VA_BLOCK_SIZE);
UVM_ASSERT(is_power_of_2(new_region_size));
if (!sysmem_mappings->gpu->parent->access_counters_supported)
return;
uvm_mutex_lock(&sysmem_mappings->reverse_map_lock);
// Find the first mapping in the region
first_reverse_map = radix_tree_lookup(&sysmem_mappings->reverse_map_tree, base_key);
UVM_ASSERT(first_reverse_map);
num_mapping_pages = uvm_va_block_region_num_pages(first_reverse_map->region);
UVM_ASSERT(num_pages >= num_mapping_pages);
UVM_ASSERT(IS_ALIGNED(base_key, num_mapping_pages));
// The region in the tree matches the size of the merged region, just return
if (num_pages == num_mapping_pages)
goto unlock_no_update;
// Otherwise update the rest of slots to point at the same reverse map
// descriptor
key = base_key + uvm_va_block_region_num_pages(first_reverse_map->region);
running_page_index = first_reverse_map->region.outer;
while (key < base_key + num_pages) {
uvm_reverse_map_t *reverse_map = NULL;
void **slot = radix_tree_lookup_slot(&sysmem_mappings->reverse_map_tree, key);
size_t slot_index;
UVM_ASSERT(slot);
reverse_map = radix_tree_deref_slot(slot);
UVM_ASSERT(reverse_map);
UVM_ASSERT(reverse_map != first_reverse_map);
UVM_ASSERT(reverse_map->va_block == first_reverse_map->va_block);
UVM_ASSERT(uvm_id_equal(reverse_map->owner, first_reverse_map->owner));
UVM_ASSERT(reverse_map->region.first == running_page_index);
NV_RADIX_TREE_REPLACE_SLOT(&sysmem_mappings->reverse_map_tree, slot, first_reverse_map);
num_mapping_pages = uvm_va_block_region_num_pages(reverse_map->region);
UVM_ASSERT(IS_ALIGNED(key, num_mapping_pages));
UVM_ASSERT(key + num_mapping_pages <= base_key + num_pages);
for (slot_index = 1; slot_index < num_mapping_pages; ++slot_index) {
slot = radix_tree_lookup_slot(&sysmem_mappings->reverse_map_tree, key + slot_index);
UVM_ASSERT(slot);
UVM_ASSERT(reverse_map == radix_tree_deref_slot(slot));
NV_RADIX_TREE_REPLACE_SLOT(&sysmem_mappings->reverse_map_tree, slot, first_reverse_map);
}
key += num_mapping_pages;
running_page_index = reverse_map->region.outer;
kmem_cache_free(g_reverse_page_map_cache, reverse_map);
}
// Grow the first mapping to cover the whole region
first_reverse_map->region.outer = first_reverse_map->region.first + num_pages;
unlock_no_update:
uvm_mutex_unlock(&sysmem_mappings->reverse_map_lock);
}
size_t uvm_pmm_sysmem_mappings_dma_to_virt(uvm_pmm_sysmem_mappings_t *sysmem_mappings,
NvU64 dma_addr,
NvU64 region_size,
uvm_reverse_map_t *out_mappings,
size_t max_out_mappings)
{
NvU64 key;
size_t num_mappings = 0;
const NvU64 base_key = dma_addr / PAGE_SIZE;
NvU32 num_pages = region_size / PAGE_SIZE;
UVM_ASSERT(region_size >= PAGE_SIZE);
UVM_ASSERT(PAGE_ALIGNED(region_size));
UVM_ASSERT(sysmem_mappings->gpu->parent->access_counters_supported);
UVM_ASSERT(max_out_mappings > 0);
uvm_mutex_lock(&sysmem_mappings->reverse_map_lock);
key = base_key;
do {
uvm_reverse_map_t *reverse_map = radix_tree_lookup(&sysmem_mappings->reverse_map_tree, key);
if (reverse_map) {
size_t num_chunk_pages = uvm_va_block_region_num_pages(reverse_map->region);
NvU32 page_offset = key & (num_chunk_pages - 1);
NvU32 num_mapping_pages = min(num_pages, (NvU32)num_chunk_pages - page_offset);
// Sysmem mappings are removed during VA block destruction.
// Therefore, we can safely retain the VA blocks as long as they
// are in the reverse map and we hold the reverse map lock.
uvm_va_block_retain(reverse_map->va_block);
out_mappings[num_mappings] = *reverse_map;
out_mappings[num_mappings].region.first += page_offset;
out_mappings[num_mappings].region.outer = out_mappings[num_mappings].region.first + num_mapping_pages;
if (++num_mappings == max_out_mappings)
break;
num_pages -= num_mapping_pages;
key += num_mapping_pages;
}
else {
--num_pages;
++key;
}
}
while (num_pages > 0);
uvm_mutex_unlock(&sysmem_mappings->reverse_map_lock);
return num_mappings;
}
uvm_chunk_sizes_mask_t uvm_cpu_chunk_get_allocation_sizes(void)
{
return uvm_cpu_chunk_allocation_sizes & UVM_CPU_CHUNK_SIZES;
}
static void uvm_cpu_chunk_set_size(uvm_cpu_chunk_t *chunk, uvm_chunk_size_t size)
{
#if !UVM_CPU_CHUNK_SIZE_IS_PAGE_SIZE()
chunk->log2_size = ilog2(size);
#endif
}
uvm_chunk_size_t uvm_cpu_chunk_get_size(uvm_cpu_chunk_t *chunk)
{
#if UVM_CPU_CHUNK_SIZE_IS_PAGE_SIZE()
return PAGE_SIZE;
#else
return ((uvm_chunk_size_t)1) << chunk->log2_size;
#endif
}
#if UVM_CPU_CHUNK_SIZE_IS_PAGE_SIZE()
struct page *uvm_cpu_chunk_get_cpu_page(uvm_va_block_t *va_block, uvm_cpu_chunk_t *chunk, uvm_page_index_t page_index)
{
UVM_ASSERT(chunk);
return chunk;
}
void uvm_cpu_chunk_put(uvm_cpu_chunk_t *chunk)
{
UVM_ASSERT(chunk);
put_page(chunk);
}
NV_STATUS uvm_cpu_chunk_gpu_mapping_alloc(uvm_va_block_t *va_block, uvm_gpu_id_t id)
{
uvm_va_block_gpu_state_t *gpu_state = uvm_va_block_gpu_state_get(va_block, id);
size_t num_pages = uvm_va_block_num_cpu_pages(va_block);
UVM_ASSERT(gpu_state);
gpu_state->cpu_chunks_dma_addrs = uvm_kvmalloc_zero(num_pages * sizeof(gpu_state->cpu_chunks_dma_addrs[0]));
if (!gpu_state->cpu_chunks_dma_addrs)
return NV_ERR_NO_MEMORY;
return NV_OK;
}
void uvm_cpu_chunk_gpu_mapping_split(uvm_va_block_t *existing, uvm_va_block_t *new, uvm_gpu_id_t id)
{
uvm_va_block_gpu_state_t *existing_state = uvm_va_block_gpu_state_get(existing, id);
uvm_va_block_gpu_state_t *new_state = uvm_va_block_gpu_state_get(new, id);
size_t new_pages = uvm_va_block_num_cpu_pages(new);
memcpy(&new_state->cpu_chunks_dma_addrs[0],
&existing_state->cpu_chunks_dma_addrs[uvm_va_block_num_cpu_pages(existing) - new_pages],
new_pages * sizeof(new_state->cpu_chunks_dma_addrs[0]));
}
void uvm_cpu_chunk_gpu_mapping_free(uvm_va_block_t *va_block, uvm_gpu_id_t id)
{
uvm_va_block_gpu_state_t *gpu_state = uvm_va_block_gpu_state_get(va_block, id);
if (gpu_state)
uvm_kvfree(gpu_state->cpu_chunks_dma_addrs);
}
NV_STATUS uvm_cpu_chunk_set_gpu_mapping_addr(uvm_va_block_t *va_block,
uvm_page_index_t page_index,
uvm_cpu_chunk_t *chunk,
uvm_gpu_id_t id,
NvU64 dma_addr)
{
uvm_va_block_gpu_state_t *gpu_state = uvm_va_block_gpu_state_get(va_block, id);
gpu_state->cpu_chunks_dma_addrs[page_index] = dma_addr;
return NV_OK;
}
NvU64 uvm_cpu_chunk_get_gpu_mapping_addr(uvm_va_block_t *va_block,
uvm_page_index_t page_index,
uvm_cpu_chunk_t *chunk,
uvm_gpu_id_t id)
{
uvm_va_block_gpu_state_t *gpu_state = uvm_va_block_gpu_state_get(va_block, id);
return gpu_state->cpu_chunks_dma_addrs[page_index];
}
NV_STATUS uvm_cpu_chunk_insert_in_block(uvm_va_block_t *va_block, uvm_cpu_chunk_t *chunk, uvm_page_index_t page_index)
{
if (!va_block->cpu.chunks) {
va_block->cpu.chunks = (unsigned long)uvm_kvmalloc_zero(uvm_va_block_num_cpu_pages(va_block) *
sizeof(uvm_cpu_chunk_t *));
if (!va_block->cpu.chunks)
return NV_ERR_NO_MEMORY;
}
UVM_ASSERT(!uvm_page_mask_test(&va_block->cpu.allocated, page_index));
UVM_ASSERT(((uvm_cpu_chunk_t **)va_block->cpu.chunks)[page_index] == NULL);
((uvm_cpu_chunk_t **)va_block->cpu.chunks)[page_index] = chunk;
uvm_page_mask_set(&va_block->cpu.allocated, page_index);
return NV_OK;
}
void uvm_cpu_chunk_remove_from_block(uvm_va_block_t *va_block, uvm_cpu_chunk_t *chunk, uvm_page_index_t page_index)
{
UVM_ASSERT(uvm_page_mask_test(&va_block->cpu.allocated, page_index));
UVM_ASSERT(((uvm_cpu_chunk_t **)va_block->cpu.chunks)[page_index] != NULL);
((uvm_cpu_chunk_t **)va_block->cpu.chunks)[page_index] = NULL;
uvm_page_mask_clear(&va_block->cpu.allocated, page_index);
}
uvm_cpu_chunk_t *uvm_cpu_chunk_get_chunk_for_page(uvm_va_block_t *va_block, uvm_page_index_t page_index)
{
UVM_ASSERT(page_index < uvm_va_block_num_cpu_pages(va_block));
if (!uvm_page_mask_test(&va_block->cpu.allocated, page_index))
return NULL;
return ((uvm_cpu_chunk_t **)va_block->cpu.chunks)[page_index];
}
NV_STATUS uvm_cpu_chunk_alloc(uvm_va_block_t *va_block,
uvm_page_index_t page_index,
struct mm_struct *mm,
uvm_cpu_chunk_t **new_chunk)
{
uvm_cpu_chunk_t *chunk = NULL;
gfp_t alloc_flags;
NV_STATUS status;
UVM_ASSERT(!uvm_page_mask_test(&va_block->cpu.allocated, page_index));
UVM_ASSERT(new_chunk);
alloc_flags = (mm ? NV_UVM_GFP_FLAGS_ACCOUNT : NV_UVM_GFP_FLAGS) | GFP_HIGHUSER;
if (!uvm_va_block_page_resident_processors_count(va_block, page_index))
alloc_flags |= __GFP_ZERO;
chunk = alloc_pages(alloc_flags, 0);
if (!chunk)
return NV_ERR_NO_MEMORY;
if (alloc_flags & __GFP_ZERO)
SetPageDirty(chunk);
status = uvm_cpu_chunk_insert_in_block(va_block, chunk, page_index);
if (status != NV_OK) {
uvm_cpu_chunk_put(chunk);
return status;
}
*new_chunk = chunk;
return NV_OK;
}
#else
struct page *uvm_cpu_chunk_get_cpu_page(uvm_va_block_t *va_block, uvm_cpu_chunk_t *chunk, uvm_page_index_t page_index)
{
uvm_va_block_region_t chunk_region;
UVM_ASSERT(chunk);
UVM_ASSERT(chunk->page);
chunk_region = uvm_va_block_chunk_region(va_block, uvm_cpu_chunk_get_size(chunk), page_index);
return chunk->page + (page_index - chunk_region.first);
}
static NvU64 uvm_cpu_chunk_get_virt_addr(uvm_va_block_t *va_block, uvm_cpu_chunk_t *chunk)
{
UVM_ASSERT(chunk);
UVM_ASSERT(chunk->region.first < chunk->region.outer);
return uvm_va_block_cpu_page_address(va_block, chunk->region.first);
}
static void cpu_chunk_release(nv_kref_t *kref)
{
uvm_cpu_chunk_t *chunk = container_of(kref, uvm_cpu_chunk_t, refcount);
uvm_cpu_chunk_t *parent = chunk->parent;
if (uvm_processor_mask_get_gpu_count(&chunk->gpu_mappings.dma_addrs_mask) > 1)
uvm_kvfree(chunk->gpu_mappings.dynamic_entries);
if (!parent) {
uvm_assert_spinlock_unlocked(&chunk->lock);
uvm_kvfree(chunk->dirty_bitmap);
put_page(chunk->page);
}
else {
uvm_cpu_chunk_put(parent);
}
uvm_kvfree(chunk);
}
void uvm_cpu_chunk_get(uvm_cpu_chunk_t *chunk)
{
UVM_ASSERT(chunk);
nv_kref_get(&chunk->refcount);
}
void uvm_cpu_chunk_put(uvm_cpu_chunk_t *chunk)
{
UVM_ASSERT(chunk);
nv_kref_put(&chunk->refcount, cpu_chunk_release);
}
NV_STATUS uvm_cpu_chunk_gpu_mapping_alloc(uvm_va_block_t *va_block, uvm_gpu_id_t id)
{
return NV_OK;
}
void uvm_cpu_chunk_gpu_mapping_split(uvm_va_block_t *existing, uvm_va_block_t *va_block, uvm_gpu_id_t id)
{
return;
}
void uvm_cpu_chunk_gpu_mapping_free(uvm_va_block_t *va_block, uvm_gpu_id_t id)
{
return;
}
static NvU32 compute_gpu_mappings_entry_index(uvm_processor_mask_t dma_addrs_mask, uvm_gpu_id_t id)
{
uvm_processor_mask_t subset_mask;
// Compute the array index for the given GPU ID by masking off all bits
// above the id and then counting the number of bits remaining.
uvm_processor_mask_zero(&subset_mask);
bitmap_set(subset_mask.bitmap, 0, uvm_id_value(id) + 1);
uvm_processor_mask_and(&subset_mask, &dma_addrs_mask, &subset_mask);
if (uvm_processor_mask_empty(&subset_mask))
return 0;
return uvm_processor_mask_get_gpu_count(&subset_mask) - 1;
}
NV_STATUS uvm_cpu_chunk_set_gpu_mapping_addr(uvm_va_block_t *va_block,
uvm_page_index_t page_index,
uvm_cpu_chunk_t *chunk,
uvm_gpu_id_t id,
NvU64 dma_addr)
{
NvU32 num_existing_entries = uvm_processor_mask_get_gpu_count(&chunk->gpu_mappings.dma_addrs_mask);
NvU32 num_new_entries;
NvU32 array_index;
NvU64 *new_entries;
if (uvm_processor_mask_empty(&chunk->gpu_mappings.dma_addrs_mask)) {
uvm_processor_mask_set(&chunk->gpu_mappings.dma_addrs_mask, id);
chunk->gpu_mappings.static_entry = dma_addr;
return NV_OK;
}
if (uvm_processor_mask_test(&chunk->gpu_mappings.dma_addrs_mask, id)) {
if (num_existing_entries == 1) {
chunk->gpu_mappings.static_entry = dma_addr;
}
else {
array_index = compute_gpu_mappings_entry_index(chunk->gpu_mappings.dma_addrs_mask, id);
chunk->gpu_mappings.dynamic_entries[array_index] = dma_addr;
}
return NV_OK;
}
num_new_entries = num_existing_entries + 1;
if (num_existing_entries == 1) {
new_entries = uvm_kvmalloc(sizeof(*new_entries) * num_new_entries);
if (new_entries) {
uvm_processor_id_t first = uvm_processor_mask_find_first_id(&chunk->gpu_mappings.dma_addrs_mask);
if (uvm_id_value(first) < uvm_id_value(id))
new_entries[0] = chunk->gpu_mappings.static_entry;
else
new_entries[1] = chunk->gpu_mappings.static_entry;
}
}
else {
new_entries = uvm_kvrealloc(chunk->gpu_mappings.dynamic_entries,
sizeof(*new_entries) * num_new_entries);
if (new_entries) {
// Get the number of bits set below the input id.
num_existing_entries = compute_gpu_mappings_entry_index(chunk->gpu_mappings.dma_addrs_mask, id);
for (; num_existing_entries < num_new_entries - 1; num_existing_entries++)
new_entries[num_existing_entries + 1] = new_entries[num_existing_entries];
}
}
if (!new_entries)
return NV_ERR_NO_MEMORY;
chunk->gpu_mappings.dynamic_entries = new_entries;
uvm_processor_mask_set(&chunk->gpu_mappings.dma_addrs_mask, id);
array_index = compute_gpu_mappings_entry_index(chunk->gpu_mappings.dma_addrs_mask, id);
chunk->gpu_mappings.dynamic_entries[array_index] = dma_addr;
return NV_OK;
}
NvU64 uvm_cpu_chunk_get_gpu_mapping_addr(uvm_va_block_t *va_block,
uvm_page_index_t page_index,
uvm_cpu_chunk_t *chunk,
uvm_gpu_id_t id)
{
NvU64 dma_addr;
if (!uvm_processor_mask_test(&chunk->gpu_mappings.dma_addrs_mask, id))
return 0;
if (uvm_processor_mask_get_gpu_count(&chunk->gpu_mappings.dma_addrs_mask) == 1) {
dma_addr = chunk->gpu_mappings.static_entry;
}
else {
NvU32 array_index = compute_gpu_mappings_entry_index(chunk->gpu_mappings.dma_addrs_mask, id);
dma_addr = chunk->gpu_mappings.dynamic_entries[array_index];
}
return dma_addr;
}
// The bottom two bits of uvm_va_block_t::chunks is used to indicate how
// CPU chunks are stored.
//
// CPU chunk storage is handled in three different ways depending on the
// type of chunks the VA block owns. This is done to minimize the memory
// required to hold metadata.
typedef enum
{
// The uvm_va_block_t::chunk pointer points to a single 2MB
// CPU chunk.
UVM_CPU_CHUNK_STORAGE_CHUNK = 0,
// The uvm_va_block_t::chunks pointer points to an array of
// pointers to CPU chunks.
UVM_CPU_CHUNK_STORAGE_ARRAY,
// The uvm_va_block_t::chunks pointer points to a
// structure of mixed (64K and 4K) chunks.
UVM_CPU_CHUNK_STORAGE_MIXED,
UVM_CPU_CHUNK_STORAGE_COUNT,
} uvm_cpu_chunk_storage_type_t;
#define UVM_CPU_CHUNK_STORAGE_MASK 0x3
#define UVM_CPU_STORAGE_GET_PTR(block) ((void *)((block)->cpu.chunks & ~UVM_CPU_CHUNK_STORAGE_MASK))
#define UVM_CPU_STORAGE_GET_TYPE(block) \
((uvm_cpu_chunk_storage_type_t)((block)->cpu.chunks & UVM_CPU_CHUNK_STORAGE_MASK))
// The maximum number of slots in the mixed chunk mode (64K + 4K chunks) is one
// more than MAX_BIG_PAGES_PER_UVM_VA_BLOCK to account for misaligned VA blocks.
#define MAX_BIG_CPU_CHUNK_SLOTS_PER_UVM_VA_BLOCK (MAX_BIG_PAGES_PER_UVM_VA_BLOCK + 1)
#define MAX_SMALL_CHUNK_PER_BIG_SLOT (UVM_CHUNK_SIZE_64K / PAGE_SIZE)
// This structure is used when a VA block contains 64K or a mix of 64K and 4K
// CPU chunks.
// For every 64K CPU chunks, big_chunks will have its corresponding bit set
// and the corresponding index in slots will point directly to the
// uvm_cpu_chunk_t structure.
//
// For 4K CPU chunks, the corresponding bit in big_chunks will be clear and
// the element in slots will point to an array of 16 uvm_cpu_chunk_t pointers.
typedef struct {
DECLARE_BITMAP(big_chunks, MAX_BIG_CPU_CHUNK_SLOTS_PER_UVM_VA_BLOCK);
void *slots[MAX_BIG_CPU_CHUNK_SLOTS_PER_UVM_VA_BLOCK];
} uvm_cpu_chunk_storage_mixed_t;
static uvm_page_index_t compute_slot_index(uvm_va_block_t *va_block, uvm_page_index_t page_index)
{
uvm_va_block_region_t block_region = uvm_va_block_region_from_block(va_block);
size_t prefix;
uvm_page_index_t big_page_index;
if (page_index < block_region.first || page_index >= block_region.outer)
return MAX_BIG_PAGES_PER_UVM_VA_BLOCK;
prefix = (UVM_ALIGN_UP(va_block->start, UVM_CHUNK_SIZE_64K) - va_block->start) / PAGE_SIZE;
if (page_index < prefix)
return 0;
big_page_index = ((page_index - prefix) / MAX_SMALL_CHUNK_PER_BIG_SLOT) + !!prefix;
UVM_ASSERT(big_page_index < MAX_BIG_CPU_CHUNK_SLOTS_PER_UVM_VA_BLOCK);
return big_page_index;
}
static size_t compute_small_index(uvm_va_block_t *va_block, uvm_page_index_t page_index)
{
size_t prefix = (UVM_ALIGN_UP(va_block->start, UVM_CHUNK_SIZE_64K) - va_block->start) / PAGE_SIZE;
if (page_index < prefix)
return page_index;
return (page_index - prefix) % MAX_SMALL_CHUNK_PER_BIG_SLOT;
}
NV_STATUS uvm_cpu_chunk_insert_in_block(uvm_va_block_t *va_block, uvm_cpu_chunk_t *chunk, uvm_page_index_t page_index)
{
uvm_chunk_size_t chunk_size = uvm_cpu_chunk_get_size(chunk);
uvm_page_index_t big_page_index;
uvm_cpu_chunk_storage_mixed_t *mixed;
uvm_cpu_chunk_t **chunks = NULL;
// We only want to use the bottom two bits of a pointer.
BUILD_BUG_ON(UVM_CPU_CHUNK_STORAGE_COUNT > 4);
chunk->region = uvm_va_block_region(page_index, page_index + uvm_cpu_chunk_num_pages(chunk));
UVM_ASSERT(chunk->region.outer <= PAGES_PER_UVM_VA_BLOCK);
// We want to protect against two threads manipulating the VA block's CPU
// chunks at the same time. However, when a block is split, the new block's
// lock is locked without tracking. So, we can't use
// uvm_assert_mutex_locked().
UVM_ASSERT(mutex_is_locked(&va_block->lock.m));
if (!va_block->cpu.chunks) {
switch (chunk_size) {
case UVM_CHUNK_SIZE_2M:
break;
case UVM_CHUNK_SIZE_64K:
mixed = uvm_kvmalloc_zero(sizeof(*mixed));
if (!mixed)
return NV_ERR_NO_MEMORY;
va_block->cpu.chunks = (unsigned long)mixed | UVM_CPU_CHUNK_STORAGE_MIXED;
break;
case UVM_CHUNK_SIZE_4K:
chunks = uvm_kvmalloc_zero(sizeof(*chunks) * uvm_va_block_num_cpu_pages(va_block));
if (!chunks)
return NV_ERR_NO_MEMORY;
va_block->cpu.chunks = (unsigned long)chunks | UVM_CPU_CHUNK_STORAGE_ARRAY;
break;
default:
return NV_ERR_INVALID_ARGUMENT;
}
}
switch (UVM_CPU_STORAGE_GET_TYPE(va_block)) {
case UVM_CPU_CHUNK_STORAGE_CHUNK:
if (va_block->cpu.chunks)
return NV_ERR_INVALID_STATE;
UVM_ASSERT(chunk_size == UVM_CHUNK_SIZE_2M);
va_block->cpu.chunks = (unsigned long)chunk | UVM_CPU_CHUNK_STORAGE_CHUNK;
break;
case UVM_CPU_CHUNK_STORAGE_MIXED:
mixed = UVM_CPU_STORAGE_GET_PTR(va_block);
big_page_index = compute_slot_index(va_block, page_index);
UVM_ASSERT(big_page_index != MAX_BIG_PAGES_PER_UVM_VA_BLOCK);
UVM_ASSERT(compute_slot_index(va_block, page_index + uvm_cpu_chunk_num_pages(chunk) - 1) == big_page_index);
if (test_bit(big_page_index, mixed->big_chunks))
return NV_ERR_INVALID_STATE;
if (chunk_size == UVM_CHUNK_SIZE_64K) {
mixed->slots[big_page_index] = chunk;
set_bit(big_page_index, mixed->big_chunks);
}
else {
size_t slot_index;
UVM_ASSERT(chunk_size == UVM_CHUNK_SIZE_4K);
chunks = mixed->slots[big_page_index];
if (!chunks) {
chunks = uvm_kvmalloc_zero(sizeof(*chunks) * MAX_SMALL_CHUNK_PER_BIG_SLOT);
if (!chunks)
return NV_ERR_NO_MEMORY;
mixed->slots[big_page_index] = chunks;
}
slot_index = compute_small_index(va_block, page_index);
chunks[slot_index] = chunk;
}
break;
case UVM_CPU_CHUNK_STORAGE_ARRAY:
chunks = UVM_CPU_STORAGE_GET_PTR(va_block);
if (chunk_size == UVM_CHUNK_SIZE_64K) {
uvm_cpu_chunk_t **subchunks = NULL;
uvm_page_index_t sub_page_index;
mixed = uvm_kvmalloc_zero(sizeof(*mixed));
if (!mixed)
return NV_ERR_NO_MEMORY;
big_page_index = compute_slot_index(va_block, page_index);
UVM_ASSERT(big_page_index != MAX_BIG_PAGES_PER_UVM_VA_BLOCK);
UVM_ASSERT(compute_slot_index(va_block, page_index + uvm_cpu_chunk_num_pages(chunk) - 1) ==
big_page_index);
mixed->slots[big_page_index] = chunk;
set_bit(big_page_index, mixed->big_chunks);
for (sub_page_index = 0; sub_page_index < uvm_va_block_num_cpu_pages(va_block); sub_page_index++) {
uvm_cpu_chunk_t *subchunk = chunks[sub_page_index];
size_t subchunk_index = compute_small_index(va_block, sub_page_index);
if (!subchunk)
continue;
if (!subchunks || compute_slot_index(va_block, sub_page_index) != big_page_index) {
subchunks = uvm_kvmalloc_zero(sizeof(*subchunks) * MAX_SMALL_CHUNK_PER_BIG_SLOT);
if (!subchunks) {
size_t i;
for (i = 0; i < MAX_BIG_CPU_CHUNK_SLOTS_PER_UVM_VA_BLOCK; i++) {
if (!test_bit(i, mixed->big_chunks) && mixed->slots[i])
uvm_kvfree(mixed->slots[i]);
}
uvm_kvfree(mixed);
return NV_ERR_NO_MEMORY;
}
big_page_index = compute_slot_index(va_block, sub_page_index);
UVM_ASSERT(mixed->slots[big_page_index] == NULL);
mixed->slots[big_page_index] = subchunks;
}
subchunks[subchunk_index] = subchunk;
if (subchunk_index == MAX_SMALL_CHUNK_PER_BIG_SLOT - 1)
subchunks = NULL;
}
va_block->cpu.chunks = (unsigned long)mixed | UVM_CPU_CHUNK_STORAGE_MIXED;
uvm_kvfree(chunks);
}
else {
chunks[page_index] = chunk;
}
default:
break;
}
uvm_page_mask_region_fill(&va_block->cpu.allocated,
uvm_va_block_region(page_index, page_index + uvm_cpu_chunk_num_pages(chunk)));
return NV_OK;
}
void uvm_cpu_chunk_remove_from_block(uvm_va_block_t *va_block, uvm_cpu_chunk_t *chunk, uvm_page_index_t page_index)
{
uvm_cpu_chunk_storage_mixed_t *mixed;
uvm_page_index_t big_page_index;
uvm_cpu_chunk_t **chunks;
// We want to protect against two threads manipulating the VA block's CPU
// chunks at the same time. However, when a block is split, the new block's
// lock is locked without tracking. So, we can't use
// uvm_assert_mutex_locked().
UVM_ASSERT(mutex_is_locked(&va_block->lock.m));
UVM_ASSERT(va_block->cpu.chunks);
switch (UVM_CPU_STORAGE_GET_TYPE(va_block)) {
case UVM_CPU_CHUNK_STORAGE_CHUNK:
UVM_ASSERT(uvm_cpu_chunk_get_size(chunk) == UVM_CHUNK_SIZE_2M);
UVM_ASSERT(UVM_CPU_STORAGE_GET_PTR(va_block) == chunk);
va_block->cpu.chunks = 0;
break;
case UVM_CPU_CHUNK_STORAGE_MIXED:
UVM_ASSERT(uvm_cpu_chunk_get_size(chunk) != UVM_CHUNK_SIZE_2M);
mixed = UVM_CPU_STORAGE_GET_PTR(va_block);
big_page_index = compute_slot_index(va_block, page_index);
UVM_ASSERT(big_page_index != MAX_BIG_PAGES_PER_UVM_VA_BLOCK);
UVM_ASSERT(mixed->slots[big_page_index] != NULL);
if (test_bit(big_page_index, mixed->big_chunks)) {
UVM_ASSERT(uvm_cpu_chunk_get_size(chunk) == UVM_CHUNK_SIZE_64K);
UVM_ASSERT(mixed->slots[big_page_index] == chunk);
mixed->slots[big_page_index] = NULL;
clear_bit(big_page_index, mixed->big_chunks);
}
else {
size_t slot_index;
UVM_ASSERT(uvm_cpu_chunk_get_size(chunk) == UVM_CHUNK_SIZE_4K);
chunks = mixed->slots[big_page_index];
slot_index = compute_small_index(va_block, page_index);
UVM_ASSERT(chunks[slot_index] == chunk);
chunks[slot_index] = NULL;
for (slot_index = 0; slot_index < MAX_SMALL_CHUNK_PER_BIG_SLOT; slot_index++) {
if (chunks[slot_index])
break;
}
if (slot_index == MAX_SMALL_CHUNK_PER_BIG_SLOT) {
uvm_kvfree(chunks);
mixed->slots[big_page_index] = NULL;
}
}
break;
case UVM_CPU_CHUNK_STORAGE_ARRAY:
UVM_ASSERT(uvm_cpu_chunk_get_size(chunk) == UVM_CHUNK_SIZE_4K);
chunks = UVM_CPU_STORAGE_GET_PTR(va_block);
UVM_ASSERT(chunks[page_index] == chunk);
chunks[page_index] = NULL;
break;
default:
return;
};
uvm_page_mask_region_clear(&va_block->cpu.allocated, chunk->region);
if (uvm_page_mask_empty(&va_block->cpu.allocated)) {
if (UVM_CPU_STORAGE_GET_TYPE(va_block) != UVM_CPU_CHUNK_STORAGE_CHUNK)
uvm_kvfree(UVM_CPU_STORAGE_GET_PTR(va_block));
va_block->cpu.chunks = 0;
}
}
uvm_cpu_chunk_t *uvm_cpu_chunk_get_chunk_for_page(uvm_va_block_t *va_block, uvm_page_index_t page_index)
{
uvm_cpu_chunk_storage_mixed_t *mixed;
uvm_cpu_chunk_t *chunk;
uvm_cpu_chunk_t **chunks;
uvm_page_index_t big_page_index;
size_t slot_index;
if (page_index >= uvm_va_block_num_cpu_pages(va_block) || !uvm_page_mask_test(&va_block->cpu.allocated, page_index))
return NULL;
UVM_ASSERT(va_block->cpu.chunks);
switch (UVM_CPU_STORAGE_GET_TYPE(va_block)) {
case UVM_CPU_CHUNK_STORAGE_CHUNK:
return UVM_CPU_STORAGE_GET_PTR(va_block);
case UVM_CPU_CHUNK_STORAGE_MIXED:
mixed = UVM_CPU_STORAGE_GET_PTR(va_block);
big_page_index = compute_slot_index(va_block, page_index);
UVM_ASSERT(big_page_index != MAX_BIG_PAGES_PER_UVM_VA_BLOCK);
UVM_ASSERT(mixed->slots[big_page_index] != NULL);
if (test_bit(big_page_index, mixed->big_chunks))
return mixed->slots[big_page_index];
chunks = mixed->slots[big_page_index];
slot_index = compute_small_index(va_block, page_index);
chunk = chunks[slot_index];
break;
case UVM_CPU_CHUNK_STORAGE_ARRAY:
chunks = UVM_CPU_STORAGE_GET_PTR(va_block);
chunk = chunks[page_index];
break;
default:
return NULL;
}
UVM_ASSERT(chunk);
return chunk;
}
NV_STATUS uvm_cpu_chunk_alloc(uvm_va_block_t *va_block,
uvm_page_index_t page_index,
struct mm_struct *mm,
uvm_cpu_chunk_t **new_chunk)
{
uvm_va_block_test_t *block_test = uvm_va_block_get_test(va_block);
uvm_cpu_chunk_t *chunk = NULL;
NvU32 cpu_allocation_sizes;
uvm_page_mask_t zero_page_mask;
uvm_gpu_id_t id;
struct page *page = NULL;
uvm_chunk_size_t alloc_size;
uvm_va_block_region_t region;
uvm_va_space_t *va_space;
uvm_processor_mask_t uvm_lite_gpus;
gfp_t base_alloc_flags;
NV_STATUS status;
UVM_ASSERT(new_chunk);
// Limit the allocation sizes only to the ones supported.
cpu_allocation_sizes = uvm_cpu_chunk_get_allocation_sizes();
if (block_test && block_test->cpu_chunk_allocation_size_mask)
cpu_allocation_sizes &= block_test->cpu_chunk_allocation_size_mask;
// Get a mask of all the block pages that are resident somewhere.
uvm_page_mask_zero(&zero_page_mask);
for_each_id_in_mask(id, &va_block->resident)
uvm_page_mask_or(&zero_page_mask, &zero_page_mask, uvm_va_block_resident_mask_get(va_block, id));
// If the VA space has a UVM-Lite GPU registered, only PAGE_SIZE allocations
// should be used in order to avoid extra copies due to dirty compound
// pages.
va_space = uvm_va_block_get_va_space(va_block);
uvm_processor_mask_andnot(&uvm_lite_gpus, &va_space->registered_gpus, &va_space->faultable_processors);
if (!uvm_processor_mask_empty(&uvm_lite_gpus))
cpu_allocation_sizes = PAGE_SIZE;
base_alloc_flags = (mm ? NV_UVM_GFP_FLAGS_ACCOUNT : NV_UVM_GFP_FLAGS) | GFP_HIGHUSER;
// Attempt to allocate CPU pages with the largest physically contiguous
// size from the set of CPU chunk sizes that we can.
// This is accomplished by:
// 1. Aligning the CPU page address down to the allocation size.
// 2. Ensuring that the entire allocation region fits withing the VA
// block.
// 3. Ensuring that the region covered by the allocation is empty.
for_each_chunk_size_rev(alloc_size, cpu_allocation_sizes) {
NvU64 alloc_virt_addr;
uvm_page_mask_t scratch_page_mask;
uvm_page_index_t alloc_page_index;
gfp_t alloc_flags = base_alloc_flags;
if (alloc_size < PAGE_SIZE)
break;
alloc_virt_addr = UVM_ALIGN_DOWN(uvm_va_block_cpu_page_address(va_block, page_index), alloc_size);
if (!uvm_va_block_contains_address(va_block, alloc_virt_addr) ||
!uvm_va_block_contains_address(va_block, alloc_virt_addr + alloc_size - 1))
continue;
alloc_page_index = uvm_va_block_cpu_page_index(va_block, alloc_virt_addr);
region = uvm_va_block_region(alloc_page_index, alloc_page_index + (alloc_size / PAGE_SIZE));
uvm_page_mask_init_from_region(&scratch_page_mask, region, NULL);
uvm_page_mask_and(&scratch_page_mask, &va_block->cpu.allocated, &scratch_page_mask);
if (!uvm_page_mask_empty(&scratch_page_mask))
continue;
// For allocation sizes higher than PAGE_SIZE, use __GFP_NORETRY in
// order to avoid higher allocation latency from the kernel compacting
// memory to satisfy the request.
if (alloc_size > PAGE_SIZE)
alloc_flags |= __GFP_COMP | __GFP_NORETRY;
// If not all pages in the allocation region are resident somewhere,
// zero out the allocated page.
// This could be wasteful if only a few pages in high-order allocation
// need to be zero'ed out but the alternative is to map single sub-
// pages one-by-one.
if (!uvm_page_mask_region_full(&zero_page_mask, region))
alloc_flags |= __GFP_ZERO;
page = alloc_pages(alloc_flags, get_order(alloc_size));
if (page) {
if (alloc_flags & __GFP_ZERO)
SetPageDirty(page);
break;
}
}
if (!page) {
status = NV_ERR_NO_MEMORY;
goto error;
}
chunk = uvm_kvmalloc_zero(sizeof(*chunk));
if (!chunk) {
status = NV_ERR_NO_MEMORY;
goto error;
}
chunk->page = page;
uvm_cpu_chunk_set_size(chunk, alloc_size);
chunk->region = region;
nv_kref_init(&chunk->refcount);
uvm_spin_lock_init(&chunk->lock, UVM_LOCK_ORDER_LEAF);
if (alloc_size > PAGE_SIZE) {
chunk->dirty_bitmap = uvm_kvmalloc_zero(BITS_TO_LONGS(alloc_size / PAGE_SIZE) * sizeof(*chunk->dirty_bitmap));
if (!chunk->dirty_bitmap) {
status = NV_ERR_NO_MEMORY;
goto error;
}
}
status = uvm_cpu_chunk_insert_in_block(va_block, chunk, chunk->region.first);
if (status != NV_OK)
goto error;
if (new_chunk)
*new_chunk = chunk;
return NV_OK;
error:
// If chunk has been allocated, uvm_cpu_chunk_put() will release the chunk
// and the page. Otherwise, only release the page.
if (chunk)
uvm_cpu_chunk_put(chunk);
else if (page)
__free_pages(page, get_order(alloc_size));
return status;
}
NV_STATUS uvm_cpu_chunk_split(uvm_va_block_t *va_block,
uvm_cpu_chunk_t *chunk,
uvm_chunk_size_t new_size,
uvm_page_index_t page_index,
uvm_cpu_chunk_t **new_chunks)
{
NV_STATUS status = NV_OK;
uvm_cpu_chunk_t *new_chunk;
uvm_page_index_t running_page_index = page_index;
size_t num_new_chunks;
size_t num_subchunk_pages;
size_t i;
UVM_ASSERT(chunk);
UVM_ASSERT(is_power_of_2(new_size));
UVM_ASSERT(new_size < uvm_cpu_chunk_get_size(chunk));
UVM_ASSERT(new_chunks);
num_new_chunks = uvm_cpu_chunk_get_size(chunk) / new_size;
num_subchunk_pages = new_size / PAGE_SIZE;
for (i = 0; i < num_new_chunks; i++) {
uvm_page_index_t relative_page_index = running_page_index - page_index;
uvm_gpu_id_t id;
new_chunk = uvm_kvmalloc_zero(sizeof(*new_chunk));
if (!new_chunk) {
status = NV_ERR_NO_MEMORY;
goto error;
}
new_chunk->page = chunk->page + relative_page_index;
new_chunk->offset = chunk->offset + relative_page_index;
new_chunk->region = uvm_va_block_region(running_page_index, running_page_index + num_subchunk_pages);
uvm_cpu_chunk_set_size(new_chunk, new_size);
nv_kref_init(&new_chunk->refcount);
// This lock is unused for logical chunks but initialize it for
// consistency.
uvm_spin_lock_init(&new_chunk->lock, UVM_LOCK_ORDER_LEAF);
new_chunk->parent = chunk;
uvm_cpu_chunk_get(new_chunk->parent);
for_each_gpu_id(id) {
NvU64 parent_dma_addr = uvm_cpu_chunk_get_gpu_mapping_addr(va_block, running_page_index, chunk, id);
if (!parent_dma_addr)
continue;
uvm_cpu_chunk_set_gpu_mapping_addr(va_block,
relative_page_index,
new_chunk,
id,
parent_dma_addr + (relative_page_index * PAGE_SIZE));
}
new_chunks[i] = new_chunk;
running_page_index += num_subchunk_pages;
}
// Drop the original reference count on the parent (from its creation). This
// is done so the parent's reference count goes to 0 when all the children
// are released.
uvm_cpu_chunk_put(chunk);
error:
if (status != NV_OK) {
while (i--)
uvm_cpu_chunk_put(new_chunk);
}
return status;
}
NV_STATUS uvm_cpu_chunk_merge(uvm_va_block_t *va_block,
uvm_cpu_chunk_t **chunks,
size_t num_merge_chunks,
uvm_chunk_size_t merge_size,
uvm_cpu_chunk_t **merged_chunk)
{
uvm_cpu_chunk_t *parent;
uvm_chunk_size_t chunk_size;
size_t i;
UVM_ASSERT(chunks);
UVM_ASSERT(num_merge_chunks > 0);
UVM_ASSERT(merged_chunk);
parent = chunks[0]->parent;
if (!parent)
return NV_WARN_NOTHING_TO_DO;
chunk_size = uvm_cpu_chunk_get_size(chunks[0]);
UVM_ASSERT(uvm_cpu_chunk_get_size(parent) == merge_size);
UVM_ASSERT(merge_size > chunk_size);
for (i = 1; i < num_merge_chunks; i++) {
if (chunks[i]->parent != parent || uvm_cpu_chunk_get_size(chunks[i]) != chunk_size)
return NV_ERR_INVALID_ARGUMENT;
UVM_ASSERT(nv_kref_read(&chunks[i]->refcount) == 1);
}
// Take a reference on the parent chunk so it doesn't get released when all
// of the children are released below.
uvm_cpu_chunk_get(parent);
for (i = 0; i < num_merge_chunks; i++)
uvm_cpu_chunk_put(chunks[i]);
*merged_chunk = parent;
return NV_OK;
}
static uvm_cpu_chunk_t *get_parent_cpu_chunk(uvm_cpu_chunk_t *chunk)
{
UVM_ASSERT(chunk);
while (chunk->parent)
chunk = chunk->parent;
return chunk;
}
// Check the CPU PTE dirty bit and if set, clear it and fill the
// physical chunk's dirty bitmap.
static void check_cpu_dirty_flag(uvm_cpu_chunk_t *chunk, uvm_page_index_t page_index)
{
struct page *page;
UVM_ASSERT(!chunk->parent);
uvm_assert_spinlock_locked(&chunk->lock);
// Kernels prior to v4.5 used the flags within the individual pages even for
// compound pages.
page = chunk->page + page_index;
if (PageDirty(page)) {
bitmap_fill(chunk->dirty_bitmap, uvm_cpu_chunk_get_size(chunk) / PAGE_SIZE);
ClearPageDirty(page);
}
}
static uvm_cpu_chunk_t *get_parent_and_page_index(uvm_cpu_chunk_t *chunk, uvm_page_index_t *out_page_index)
{
uvm_cpu_chunk_t *parent;
uvm_page_index_t page_index;
UVM_ASSERT(chunk);
UVM_ASSERT(chunk->page);
UVM_ASSERT(out_page_index);
page_index = *out_page_index;
UVM_ASSERT(chunk->region.first <= page_index && page_index < chunk->region.outer);
page_index = chunk->offset + (page_index - chunk->region.first);
parent = get_parent_cpu_chunk(chunk);
UVM_ASSERT(page_index < uvm_cpu_chunk_get_size(parent) / PAGE_SIZE);
*out_page_index = page_index;
return parent;
}
void uvm_cpu_chunk_mark_dirty(uvm_cpu_chunk_t *chunk, uvm_page_index_t page_index)
{
uvm_cpu_chunk_t *parent;
parent = get_parent_and_page_index(chunk, &page_index);
if (uvm_cpu_chunk_get_size(parent) == PAGE_SIZE) {
SetPageDirty(parent->page);
return;
}
uvm_spin_lock(&parent->lock);
set_bit(page_index, parent->dirty_bitmap);
uvm_spin_unlock(&parent->lock);
}
void uvm_cpu_chunk_mark_clean(uvm_cpu_chunk_t *chunk, uvm_page_index_t page_index)
{
uvm_cpu_chunk_t *parent;
parent = get_parent_and_page_index(chunk, &page_index);
if (uvm_cpu_chunk_get_size(parent) == PAGE_SIZE) {
ClearPageDirty(parent->page);
return;
}
uvm_spin_lock(&parent->lock);
check_cpu_dirty_flag(parent, page_index);
clear_bit(page_index, parent->dirty_bitmap);
uvm_spin_unlock(&parent->lock);
}
bool uvm_cpu_chunk_is_dirty(uvm_cpu_chunk_t *chunk, uvm_page_index_t page_index)
{
uvm_cpu_chunk_t *parent;
bool dirty;
parent = get_parent_and_page_index(chunk, &page_index);
if (uvm_cpu_chunk_get_size(parent) == PAGE_SIZE)
return PageDirty(parent->page);
uvm_spin_lock(&parent->lock);
check_cpu_dirty_flag(parent, page_index);
dirty = test_bit(page_index, parent->dirty_bitmap);
uvm_spin_unlock(&parent->lock);
return dirty;
}
#endif // !UVM_CPU_CHUNK_SIZE_IS_PAGE_SIZE()
uvm_cpu_chunk_t *uvm_cpu_chunk_first_in_block(uvm_va_block_t *va_block, uvm_page_index_t *out_page_index)
{
uvm_cpu_chunk_t *chunk = NULL;
uvm_page_index_t page_index;
uvm_va_block_region_t block_region = uvm_va_block_region_from_block(va_block);
page_index = uvm_va_block_first_page_in_mask(block_region, &va_block->cpu.allocated);
if (page_index < block_region.outer)
chunk = uvm_cpu_chunk_get_chunk_for_page(va_block, page_index);
if (out_page_index)
*out_page_index = page_index;
return chunk;
}
uvm_cpu_chunk_t *uvm_cpu_chunk_next(uvm_va_block_t *va_block, uvm_page_index_t *previous_page_index)
{
uvm_va_block_region_t block_region;
UVM_ASSERT(va_block);
UVM_ASSERT(previous_page_index);
block_region = uvm_va_block_region_from_block(va_block);
*previous_page_index = uvm_va_block_next_page_in_mask(block_region, &va_block->cpu.allocated, *previous_page_index);
if (*previous_page_index == block_region.outer)
return NULL;
return uvm_cpu_chunk_get_chunk_for_page(va_block, *previous_page_index);
}