/******************************************************************************* Copyright (c) 2016-2019 NVIDIA Corporation Permission is hereby granted, free of charge, to any person obtaining a copy of this software and associated documentation files (the "Software"), to deal in the Software without restriction, including without limitation the rights to use, copy, modify, merge, publish, distribute, sublicense, and/or sell copies of the Software, and to permit persons to whom the Software is furnished to do so, subject to the following conditions: The above copyright notice and this permission notice shall be included in all copies or substantial portions of the Software. THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE. *******************************************************************************/ #include "uvm_common.h" #include "uvm_linux.h" #include "uvm_test.h" #include "uvm_test_ioctl.h" #include "uvm_va_block.h" #include "uvm_va_space.h" #include "uvm_mmu.h" static NV_STATUS test_chunk_index_range(NvU64 start, NvU64 size, uvm_gpu_t *gpu) { size_t chunk_index, last_chunk_index = 0; uvm_chunk_size_t chunk_size, test_chunk_size; NvU64 addr, next_addr, outer_addr; if (fatal_signal_pending(current)) return NV_ERR_SIGNAL_PENDING; outer_addr = start + size; for (addr = start; addr < outer_addr;) { uvm_page_index_t start_page_index = (uvm_page_index_t)((addr - start) / PAGE_SIZE); chunk_index = uvm_va_block_gpu_chunk_index_range(start, size, gpu, start_page_index, &chunk_size); if (addr == start) TEST_CHECK_RET(chunk_index == 0); else TEST_CHECK_RET(chunk_index == last_chunk_index + 1); last_chunk_index = chunk_index; TEST_CHECK_RET(chunk_size >= PAGE_SIZE); TEST_CHECK_RET(IS_ALIGNED(addr, chunk_size)); next_addr = addr + chunk_size; TEST_CHECK_RET(next_addr <= outer_addr); // Verify that this is the largest allowed chunk size for_each_chunk_size_rev(test_chunk_size, gpu->parent->mmu_user_chunk_sizes) { if (IS_ALIGNED(addr, test_chunk_size) && addr + test_chunk_size <= outer_addr) { TEST_CHECK_RET(test_chunk_size == chunk_size); break; } } // Check that every page covered by this chunk returns the same values for (addr += PAGE_SIZE; addr < next_addr; addr += PAGE_SIZE) { uvm_page_index_t page_index = (uvm_page_index_t)((addr - start) / PAGE_SIZE); last_chunk_index = uvm_va_block_gpu_chunk_index_range(start, size, gpu, page_index, &test_chunk_size); TEST_CHECK_RET(last_chunk_index == chunk_index); TEST_CHECK_RET(test_chunk_size == chunk_size); } } return NV_OK; } static NV_STATUS test_chunk_index(uvm_gpu_t *gpu) { // Fake a random address aligned to a block boundary const NvU64 start = 17 * UVM_VA_BLOCK_SIZE; uvm_va_block_region_t region; for (region.first = 0; region.first < PAGES_PER_UVM_VA_BLOCK; ++region.first) { for (region.outer = region.first + 1; region.outer <= PAGES_PER_UVM_VA_BLOCK; ++region.outer) { NV_STATUS status = test_chunk_index_range(start + region.first * PAGE_SIZE, uvm_va_block_region_size(region), gpu); if (status != NV_OK) return status; } } return NV_OK; } NV_STATUS uvm_test_va_block(UVM_TEST_VA_BLOCK_PARAMS *params, struct file *filp) { uvm_va_space_t *va_space = uvm_va_space_get(filp); uvm_gpu_t *gpu; NV_STATUS status = NV_OK; uvm_va_space_down_read(va_space); for_each_va_space_gpu(gpu, va_space) TEST_NV_CHECK_GOTO(test_chunk_index(gpu), out); out: uvm_va_space_up_read(va_space); return status; }