mirror of
https://github.com/NVIDIA/open-gpu-kernel-modules.git
synced 2024-12-12 21:08:56 +01:00
113 lines
4.3 KiB
C
113 lines
4.3 KiB
C
/*******************************************************************************
|
|
Copyright (c) 2016-2019 NVIDIA Corporation
|
|
|
|
Permission is hereby granted, free of charge, to any person obtaining a copy
|
|
of this software and associated documentation files (the "Software"), to
|
|
deal in the Software without restriction, including without limitation the
|
|
rights to use, copy, modify, merge, publish, distribute, sublicense, and/or
|
|
sell copies of the Software, and to permit persons to whom the Software is
|
|
furnished to do so, subject to the following conditions:
|
|
|
|
The above copyright notice and this permission notice shall be
|
|
included in all copies or substantial portions of the Software.
|
|
|
|
THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
|
|
IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
|
|
FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
|
|
THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
|
|
LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING
|
|
FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER
|
|
DEALINGS IN THE SOFTWARE.
|
|
|
|
*******************************************************************************/
|
|
|
|
#include "uvm_common.h"
|
|
#include "uvm_linux.h"
|
|
#include "uvm_test.h"
|
|
#include "uvm_test_ioctl.h"
|
|
#include "uvm_va_block.h"
|
|
#include "uvm_va_space.h"
|
|
#include "uvm_mmu.h"
|
|
|
|
static NV_STATUS test_chunk_index_range(NvU64 start, NvU64 size, uvm_gpu_t *gpu)
|
|
{
|
|
size_t chunk_index, last_chunk_index = 0;
|
|
uvm_chunk_size_t chunk_size, test_chunk_size;
|
|
NvU64 addr, next_addr, outer_addr;
|
|
|
|
if (fatal_signal_pending(current))
|
|
return NV_ERR_SIGNAL_PENDING;
|
|
|
|
outer_addr = start + size;
|
|
|
|
for (addr = start; addr < outer_addr;) {
|
|
uvm_page_index_t start_page_index = (uvm_page_index_t)((addr - start) / PAGE_SIZE);
|
|
|
|
chunk_index = uvm_va_block_gpu_chunk_index_range(start, size, gpu, start_page_index, &chunk_size);
|
|
if (addr == start)
|
|
TEST_CHECK_RET(chunk_index == 0);
|
|
else
|
|
TEST_CHECK_RET(chunk_index == last_chunk_index + 1);
|
|
|
|
last_chunk_index = chunk_index;
|
|
|
|
TEST_CHECK_RET(chunk_size >= PAGE_SIZE);
|
|
TEST_CHECK_RET(IS_ALIGNED(addr, chunk_size));
|
|
next_addr = addr + chunk_size;
|
|
TEST_CHECK_RET(next_addr <= outer_addr);
|
|
|
|
// Verify that this is the largest allowed chunk size
|
|
for_each_chunk_size_rev(test_chunk_size, gpu->parent->mmu_user_chunk_sizes) {
|
|
if (IS_ALIGNED(addr, test_chunk_size) && addr + test_chunk_size <= outer_addr) {
|
|
TEST_CHECK_RET(test_chunk_size == chunk_size);
|
|
break;
|
|
}
|
|
}
|
|
|
|
// Check that every page covered by this chunk returns the same values
|
|
for (addr += PAGE_SIZE; addr < next_addr; addr += PAGE_SIZE) {
|
|
uvm_page_index_t page_index = (uvm_page_index_t)((addr - start) / PAGE_SIZE);
|
|
last_chunk_index = uvm_va_block_gpu_chunk_index_range(start, size, gpu, page_index, &test_chunk_size);
|
|
TEST_CHECK_RET(last_chunk_index == chunk_index);
|
|
TEST_CHECK_RET(test_chunk_size == chunk_size);
|
|
}
|
|
}
|
|
|
|
return NV_OK;
|
|
}
|
|
|
|
static NV_STATUS test_chunk_index(uvm_gpu_t *gpu)
|
|
{
|
|
// Fake a random address aligned to a block boundary
|
|
const NvU64 start = 17 * UVM_VA_BLOCK_SIZE;
|
|
uvm_va_block_region_t region;
|
|
|
|
for (region.first = 0; region.first < PAGES_PER_UVM_VA_BLOCK; ++region.first) {
|
|
for (region.outer = region.first + 1; region.outer <= PAGES_PER_UVM_VA_BLOCK; ++region.outer) {
|
|
NV_STATUS status = test_chunk_index_range(start + region.first * PAGE_SIZE,
|
|
uvm_va_block_region_size(region),
|
|
gpu);
|
|
if (status != NV_OK)
|
|
return status;
|
|
}
|
|
}
|
|
|
|
return NV_OK;
|
|
}
|
|
|
|
NV_STATUS uvm_test_va_block(UVM_TEST_VA_BLOCK_PARAMS *params, struct file *filp)
|
|
{
|
|
uvm_va_space_t *va_space = uvm_va_space_get(filp);
|
|
uvm_gpu_t *gpu;
|
|
NV_STATUS status = NV_OK;
|
|
|
|
uvm_va_space_down_read(va_space);
|
|
|
|
for_each_va_space_gpu(gpu, va_space)
|
|
TEST_NV_CHECK_GOTO(test_chunk_index(gpu), out);
|
|
|
|
out:
|
|
uvm_va_space_up_read(va_space);
|
|
return status;
|
|
}
|