2022-05-09 22:18:59 +02:00
|
|
|
/*******************************************************************************
|
|
|
|
Copyright (c) 2018 NVIDIA Corporation
|
|
|
|
|
|
|
|
Permission is hereby granted, free of charge, to any person obtaining a copy
|
|
|
|
of this software and associated documentation files (the "Software"), to
|
|
|
|
deal in the Software without restriction, including without limitation the
|
|
|
|
rights to use, copy, modify, merge, publish, distribute, sublicense, and/or
|
|
|
|
sell copies of the Software, and to permit persons to whom the Software is
|
|
|
|
furnished to do so, subject to the following conditions:
|
|
|
|
|
|
|
|
The above copyright notice and this permission notice shall be
|
|
|
|
included in all copies or substantial portions of the Software.
|
|
|
|
|
|
|
|
THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
|
|
|
|
IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
|
|
|
|
FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
|
|
|
|
THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
|
|
|
|
LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING
|
|
|
|
FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER
|
|
|
|
DEALINGS IN THE SOFTWARE.
|
|
|
|
*******************************************************************************/
|
|
|
|
|
|
|
|
#include "uvm_tools.h"
|
|
|
|
#include "uvm_va_range.h"
|
|
|
|
#include "uvm_ats_faults.h"
|
|
|
|
#include "uvm_migrate_pageable.h"
|
2023-07-18 15:54:53 +02:00
|
|
|
#include <linux/mempolicy.h>
|
2022-05-09 22:18:59 +02:00
|
|
|
|
2022-11-10 17:39:33 +01:00
|
|
|
// TODO: Bug 2103669: Implement a real prefetching policy and remove or adapt
|
|
|
|
// these experimental parameters. These are intended to help guide that policy.
|
|
|
|
static unsigned int uvm_exp_perf_prefetch_ats_order_replayable = 0;
|
|
|
|
module_param(uvm_exp_perf_prefetch_ats_order_replayable, uint, 0644);
|
|
|
|
MODULE_PARM_DESC(uvm_exp_perf_prefetch_ats_order_replayable,
|
|
|
|
"Max order of pages (2^N) to prefetch on replayable ATS faults");
|
|
|
|
|
|
|
|
static unsigned int uvm_exp_perf_prefetch_ats_order_non_replayable = 0;
|
|
|
|
module_param(uvm_exp_perf_prefetch_ats_order_non_replayable, uint, 0644);
|
|
|
|
MODULE_PARM_DESC(uvm_exp_perf_prefetch_ats_order_non_replayable,
|
|
|
|
"Max order of pages (2^N) to prefetch on non-replayable ATS faults");
|
|
|
|
|
|
|
|
// Expand the fault region to the naturally-aligned region with order given by
|
|
|
|
// the module parameters, clamped to the vma containing fault_addr (if any).
|
|
|
|
// Note that this means the region contains fault_addr but may not begin at
|
|
|
|
// fault_addr.
|
2023-05-30 19:11:36 +02:00
|
|
|
static void expand_fault_region(struct vm_area_struct *vma,
|
|
|
|
NvU64 start,
|
|
|
|
size_t length,
|
2022-11-10 17:39:33 +01:00
|
|
|
uvm_fault_client_type_t client_type,
|
2023-05-30 19:11:36 +02:00
|
|
|
unsigned long *migrate_start,
|
|
|
|
unsigned long *migrate_length)
|
2022-11-10 17:39:33 +01:00
|
|
|
{
|
|
|
|
unsigned int order;
|
|
|
|
unsigned long outer, aligned_start, aligned_size;
|
|
|
|
|
2023-05-30 19:11:36 +02:00
|
|
|
*migrate_start = start;
|
|
|
|
*migrate_length = length;
|
2022-11-10 17:39:33 +01:00
|
|
|
|
|
|
|
if (client_type == UVM_FAULT_CLIENT_TYPE_HUB)
|
|
|
|
order = uvm_exp_perf_prefetch_ats_order_non_replayable;
|
|
|
|
else
|
|
|
|
order = uvm_exp_perf_prefetch_ats_order_replayable;
|
|
|
|
|
|
|
|
if (order == 0)
|
|
|
|
return;
|
|
|
|
|
2023-05-30 19:11:36 +02:00
|
|
|
UVM_ASSERT(vma);
|
2022-11-10 17:39:33 +01:00
|
|
|
UVM_ASSERT(order < BITS_PER_LONG - PAGE_SHIFT);
|
|
|
|
|
|
|
|
aligned_size = (1UL << order) * PAGE_SIZE;
|
|
|
|
|
2023-05-30 19:11:36 +02:00
|
|
|
aligned_start = start & ~(aligned_size - 1);
|
2022-11-10 17:39:33 +01:00
|
|
|
|
2023-05-30 19:11:36 +02:00
|
|
|
*migrate_start = max(vma->vm_start, aligned_start);
|
2022-11-10 17:39:33 +01:00
|
|
|
outer = min(vma->vm_end, aligned_start + aligned_size);
|
2023-05-30 19:11:36 +02:00
|
|
|
*migrate_length = outer - *migrate_start;
|
2022-11-10 17:39:33 +01:00
|
|
|
}
|
|
|
|
|
2023-05-30 19:11:36 +02:00
|
|
|
static NV_STATUS service_ats_faults(uvm_gpu_va_space_t *gpu_va_space,
|
|
|
|
struct vm_area_struct *vma,
|
|
|
|
NvU64 start,
|
|
|
|
size_t length,
|
|
|
|
uvm_fault_access_type_t access_type,
|
2023-07-18 15:54:53 +02:00
|
|
|
uvm_ats_fault_context_t *ats_context)
|
2022-05-09 22:18:59 +02:00
|
|
|
{
|
|
|
|
uvm_va_space_t *va_space = gpu_va_space->va_space;
|
|
|
|
struct mm_struct *mm = va_space->va_space_mm.mm;
|
|
|
|
bool write = (access_type >= UVM_FAULT_ACCESS_TYPE_WRITE);
|
|
|
|
NV_STATUS status;
|
2023-05-30 19:11:36 +02:00
|
|
|
NvU64 user_space_start;
|
|
|
|
NvU64 user_space_length;
|
2022-05-09 22:18:59 +02:00
|
|
|
|
|
|
|
// Request uvm_migrate_pageable() to touch the corresponding page after
|
|
|
|
// population.
|
|
|
|
// Under virtualization ATS provides two translations:
|
|
|
|
// 1) guest virtual -> guest physical
|
|
|
|
// 2) guest physical -> host physical
|
|
|
|
//
|
|
|
|
// The overall ATS translation will fault if either of those translations is
|
2023-07-18 15:54:53 +02:00
|
|
|
// invalid. The pin_user_pages() call within uvm_migrate_pageable() call
|
|
|
|
// below handles translation #1, but not #2. We don't know if we're running
|
|
|
|
// as a guest, but in case we are we can force that translation to be valid
|
|
|
|
// by touching the guest physical address from the CPU. If the translation
|
|
|
|
// is not valid then the access will cause a hypervisor fault. Note that
|
|
|
|
// dma_map_page() can't establish mappings used by GPU ATS SVA translations.
|
|
|
|
// GPU accesses to host physical addresses obtained as a result of the
|
|
|
|
// address translation request uses the CPU address space instead of the
|
|
|
|
// IOMMU address space since the translated host physical address isn't
|
|
|
|
// necessarily an IOMMU address. The only way to establish guest physical to
|
|
|
|
// host physical mapping in the CPU address space is to touch the page from
|
|
|
|
// the CPU.
|
2022-05-09 22:18:59 +02:00
|
|
|
//
|
|
|
|
// We assume that the hypervisor mappings are all VM_PFNMAP, VM_SHARED, and
|
|
|
|
// VM_WRITE, meaning that the mappings are all granted write access on any
|
|
|
|
// fault and that the kernel will never revoke them.
|
|
|
|
// drivers/vfio/pci/vfio_pci_nvlink2.c enforces this. Thus we can assume
|
|
|
|
// that a read fault is always sufficient to also enable write access on the
|
|
|
|
// guest translation.
|
|
|
|
|
|
|
|
uvm_migrate_args_t uvm_migrate_args =
|
|
|
|
{
|
2023-07-18 15:54:53 +02:00
|
|
|
.va_space = va_space,
|
|
|
|
.mm = mm,
|
|
|
|
.dst_id = ats_context->residency_id,
|
|
|
|
.dst_node_id = ats_context->residency_node,
|
|
|
|
.populate_permissions = write ? UVM_POPULATE_PERMISSIONS_WRITE : UVM_POPULATE_PERMISSIONS_ANY,
|
|
|
|
.touch = true,
|
|
|
|
.skip_mapped = true,
|
|
|
|
.populate_on_cpu_alloc_failures = true,
|
|
|
|
.user_space_start = &user_space_start,
|
|
|
|
.user_space_length = &user_space_length,
|
2022-05-09 22:18:59 +02:00
|
|
|
};
|
|
|
|
|
|
|
|
UVM_ASSERT(uvm_ats_can_service_faults(gpu_va_space, mm));
|
|
|
|
|
2023-07-18 15:54:53 +02:00
|
|
|
expand_fault_region(vma,
|
|
|
|
start,
|
|
|
|
length,
|
|
|
|
ats_context->client_type,
|
|
|
|
&uvm_migrate_args.start,
|
|
|
|
&uvm_migrate_args.length);
|
2022-11-10 17:39:33 +01:00
|
|
|
|
2022-05-09 22:18:59 +02:00
|
|
|
// We are trying to use migrate_vma API in the kernel (if it exists) to
|
|
|
|
// populate and map the faulting region on the GPU. We want to do this only
|
|
|
|
// on the first touch. That is, pages which are not already mapped. So, we
|
|
|
|
// set skip_mapped to true. For pages already mapped, this will only handle
|
|
|
|
// PTE upgrades if needed.
|
|
|
|
status = uvm_migrate_pageable(&uvm_migrate_args);
|
|
|
|
if (status == NV_WARN_NOTHING_TO_DO)
|
|
|
|
status = NV_OK;
|
|
|
|
|
|
|
|
UVM_ASSERT(status != NV_ERR_MORE_PROCESSING_REQUIRED);
|
|
|
|
|
|
|
|
return status;
|
|
|
|
}
|
|
|
|
|
2023-05-30 19:11:36 +02:00
|
|
|
static void flush_tlb_write_faults(uvm_gpu_va_space_t *gpu_va_space,
|
|
|
|
NvU64 addr,
|
|
|
|
size_t size,
|
|
|
|
uvm_fault_client_type_t client_type)
|
2022-05-09 22:18:59 +02:00
|
|
|
{
|
2023-05-30 19:11:36 +02:00
|
|
|
uvm_ats_fault_invalidate_t *ats_invalidate;
|
|
|
|
|
|
|
|
if (client_type == UVM_FAULT_CLIENT_TYPE_GPC)
|
|
|
|
ats_invalidate = &gpu_va_space->gpu->parent->fault_buffer_info.replayable.ats_invalidate;
|
|
|
|
else
|
|
|
|
ats_invalidate = &gpu_va_space->gpu->parent->fault_buffer_info.non_replayable.ats_invalidate;
|
|
|
|
|
|
|
|
if (!ats_invalidate->write_faults_in_batch) {
|
|
|
|
uvm_tlb_batch_begin(&gpu_va_space->page_tables, &ats_invalidate->write_faults_tlb_batch);
|
|
|
|
ats_invalidate->write_faults_in_batch = true;
|
|
|
|
}
|
|
|
|
|
|
|
|
uvm_tlb_batch_invalidate(&ats_invalidate->write_faults_tlb_batch, addr, size, PAGE_SIZE, UVM_MEMBAR_NONE);
|
|
|
|
}
|
2022-05-09 22:18:59 +02:00
|
|
|
|
2023-07-18 15:54:53 +02:00
|
|
|
static void ats_batch_select_residency(uvm_gpu_va_space_t *gpu_va_space,
|
|
|
|
struct vm_area_struct *vma,
|
|
|
|
uvm_ats_fault_context_t *ats_context)
|
|
|
|
{
|
|
|
|
uvm_gpu_t *gpu = gpu_va_space->gpu;
|
|
|
|
int residency = uvm_gpu_numa_node(gpu);
|
|
|
|
|
|
|
|
#if defined(NV_MEMPOLICY_HAS_UNIFIED_NODES)
|
|
|
|
struct mempolicy *vma_policy = vma_policy(vma);
|
|
|
|
unsigned short mode;
|
|
|
|
|
|
|
|
if (!vma_policy)
|
|
|
|
goto done;
|
|
|
|
|
|
|
|
mode = vma_policy->mode;
|
|
|
|
|
|
|
|
if ((mode == MPOL_BIND) || (mode == MPOL_PREFERRED_MANY) || (mode == MPOL_PREFERRED)) {
|
|
|
|
int home_node = NUMA_NO_NODE;
|
|
|
|
|
|
|
|
#if defined(NV_MEMPOLICY_HAS_HOME_NODE)
|
|
|
|
if ((mode != MPOL_PREFERRED) && (vma_policy->home_node != NUMA_NO_NODE))
|
|
|
|
home_node = vma_policy->home_node;
|
|
|
|
#endif
|
|
|
|
|
|
|
|
// Prefer home_node if set. Otherwise, prefer the faulting GPU if it's
|
|
|
|
// in the list of preferred nodes, else prefer the closest_cpu_numa_node
|
|
|
|
// to the GPU if closest_cpu_numa_node is in the list of preferred
|
|
|
|
// nodes. Fallback to the faulting GPU if all else fails.
|
|
|
|
if (home_node != NUMA_NO_NODE) {
|
|
|
|
residency = home_node;
|
|
|
|
}
|
|
|
|
else if (!node_isset(residency, vma_policy->nodes)) {
|
|
|
|
int closest_cpu_numa_node = gpu->parent->closest_cpu_numa_node;
|
|
|
|
|
|
|
|
if ((closest_cpu_numa_node != NUMA_NO_NODE) && node_isset(closest_cpu_numa_node, vma_policy->nodes))
|
|
|
|
residency = gpu->parent->closest_cpu_numa_node;
|
|
|
|
else
|
|
|
|
residency = first_node(vma_policy->nodes);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// Update gpu if residency is not the faulting gpu.
|
|
|
|
if (residency != uvm_gpu_numa_node(gpu))
|
|
|
|
gpu = uvm_va_space_find_gpu_with_memory_node_id(gpu_va_space->va_space, residency);
|
|
|
|
|
|
|
|
done:
|
|
|
|
#endif
|
|
|
|
|
|
|
|
ats_context->residency_id = gpu ? gpu->parent->id : UVM_ID_CPU;
|
|
|
|
ats_context->residency_node = residency;
|
|
|
|
}
|
|
|
|
|
2023-05-30 19:11:36 +02:00
|
|
|
NV_STATUS uvm_ats_service_faults(uvm_gpu_va_space_t *gpu_va_space,
|
|
|
|
struct vm_area_struct *vma,
|
|
|
|
NvU64 base,
|
|
|
|
uvm_ats_fault_context_t *ats_context)
|
|
|
|
{
|
|
|
|
NV_STATUS status = NV_OK;
|
|
|
|
uvm_va_block_region_t subregion;
|
|
|
|
uvm_va_block_region_t region = uvm_va_block_region(0, PAGES_PER_UVM_VA_BLOCK);
|
|
|
|
uvm_page_mask_t *read_fault_mask = &ats_context->read_fault_mask;
|
|
|
|
uvm_page_mask_t *write_fault_mask = &ats_context->write_fault_mask;
|
|
|
|
uvm_page_mask_t *faults_serviced_mask = &ats_context->faults_serviced_mask;
|
|
|
|
uvm_page_mask_t *reads_serviced_mask = &ats_context->reads_serviced_mask;
|
|
|
|
uvm_fault_client_type_t client_type = ats_context->client_type;
|
|
|
|
|
|
|
|
UVM_ASSERT(vma);
|
|
|
|
UVM_ASSERT(IS_ALIGNED(base, UVM_VA_BLOCK_SIZE));
|
2022-05-09 22:18:59 +02:00
|
|
|
UVM_ASSERT(g_uvm_global.ats.enabled);
|
2023-05-30 19:11:36 +02:00
|
|
|
UVM_ASSERT(gpu_va_space);
|
2022-05-09 22:18:59 +02:00
|
|
|
UVM_ASSERT(gpu_va_space->ats.enabled);
|
|
|
|
UVM_ASSERT(uvm_gpu_va_space_state(gpu_va_space) == UVM_GPU_VA_SPACE_STATE_ACTIVE);
|
|
|
|
|
2023-05-30 19:11:36 +02:00
|
|
|
uvm_page_mask_zero(faults_serviced_mask);
|
|
|
|
uvm_page_mask_zero(reads_serviced_mask);
|
|
|
|
|
|
|
|
if (!(vma->vm_flags & VM_READ))
|
|
|
|
return status;
|
|
|
|
|
|
|
|
if (!(vma->vm_flags & VM_WRITE)) {
|
|
|
|
// If VMA doesn't have write permissions, all write faults are fatal.
|
|
|
|
// Try servicing such faults for read iff they are also present in
|
|
|
|
// read_fault_mask. This is because for replayable faults, if there are
|
|
|
|
// pending read accesses on the same page, we have to service them
|
|
|
|
// before we can cancel the write/atomic faults. So we try with read
|
|
|
|
// fault access type even though these write faults are fatal.
|
|
|
|
if (ats_context->client_type == UVM_FAULT_CLIENT_TYPE_GPC)
|
|
|
|
uvm_page_mask_and(write_fault_mask, write_fault_mask, read_fault_mask);
|
|
|
|
else
|
|
|
|
uvm_page_mask_zero(write_fault_mask);
|
2022-05-09 22:18:59 +02:00
|
|
|
}
|
|
|
|
|
2023-07-18 15:54:53 +02:00
|
|
|
ats_batch_select_residency(gpu_va_space, vma, ats_context);
|
|
|
|
|
2023-05-30 19:11:36 +02:00
|
|
|
for_each_va_block_subregion_in_mask(subregion, write_fault_mask, region) {
|
|
|
|
NvU64 start = base + (subregion.first * PAGE_SIZE);
|
|
|
|
size_t length = uvm_va_block_region_num_pages(subregion) * PAGE_SIZE;
|
|
|
|
uvm_fault_access_type_t access_type = (vma->vm_flags & VM_WRITE) ?
|
|
|
|
UVM_FAULT_ACCESS_TYPE_WRITE :
|
|
|
|
UVM_FAULT_ACCESS_TYPE_READ;
|
|
|
|
|
|
|
|
UVM_ASSERT(start >= vma->vm_start);
|
|
|
|
UVM_ASSERT((start + length) <= vma->vm_end);
|
|
|
|
|
2023-07-18 15:54:53 +02:00
|
|
|
status = service_ats_faults(gpu_va_space, vma, start, length, access_type, ats_context);
|
2023-05-30 19:11:36 +02:00
|
|
|
if (status != NV_OK)
|
|
|
|
return status;
|
|
|
|
|
|
|
|
if (vma->vm_flags & VM_WRITE) {
|
|
|
|
uvm_page_mask_region_fill(faults_serviced_mask, subregion);
|
|
|
|
|
|
|
|
// The Linux kernel never invalidates TLB entries on mapping
|
|
|
|
// permission upgrade. This is a problem if the GPU has cached
|
|
|
|
// entries with the old permission. The GPU will re-fetch the entry
|
|
|
|
// if the PTE is invalid and page size is not 4K (this is the case
|
|
|
|
// on P9). However, if a page gets upgraded from R/O to R/W and GPU
|
|
|
|
// has the PTEs cached with R/O permissions we will enter an
|
|
|
|
// infinite loop because we just forward the fault to the Linux
|
|
|
|
// kernel and it will see that the permissions in the page table are
|
|
|
|
// correct. Therefore, we flush TLB entries on ATS write faults.
|
|
|
|
flush_tlb_write_faults(gpu_va_space, start, length, client_type);
|
2022-05-09 22:18:59 +02:00
|
|
|
}
|
|
|
|
else {
|
2023-05-30 19:11:36 +02:00
|
|
|
uvm_page_mask_region_fill(reads_serviced_mask, subregion);
|
2022-05-09 22:18:59 +02:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2023-05-30 19:11:36 +02:00
|
|
|
// Remove write faults from read_fault_mask
|
|
|
|
uvm_page_mask_andnot(read_fault_mask, read_fault_mask, write_fault_mask);
|
|
|
|
|
|
|
|
for_each_va_block_subregion_in_mask(subregion, read_fault_mask, region) {
|
|
|
|
NvU64 start = base + (subregion.first * PAGE_SIZE);
|
|
|
|
size_t length = uvm_va_block_region_num_pages(subregion) * PAGE_SIZE;
|
2023-07-18 15:54:53 +02:00
|
|
|
uvm_fault_access_type_t access_type = UVM_FAULT_ACCESS_TYPE_READ;
|
2023-05-30 19:11:36 +02:00
|
|
|
|
|
|
|
UVM_ASSERT(start >= vma->vm_start);
|
|
|
|
UVM_ASSERT((start + length) <= vma->vm_end);
|
2022-05-09 22:18:59 +02:00
|
|
|
|
2023-07-18 15:54:53 +02:00
|
|
|
status = service_ats_faults(gpu_va_space, vma, start, length, access_type, ats_context);
|
2023-05-30 19:11:36 +02:00
|
|
|
if (status != NV_OK)
|
|
|
|
return status;
|
|
|
|
|
|
|
|
uvm_page_mask_region_fill(faults_serviced_mask, subregion);
|
2022-05-09 22:18:59 +02:00
|
|
|
}
|
|
|
|
|
|
|
|
return status;
|
|
|
|
}
|
|
|
|
|
2023-05-30 19:11:36 +02:00
|
|
|
bool uvm_ats_check_in_gmmu_region(uvm_va_space_t *va_space, NvU64 address, uvm_va_range_t *next)
|
|
|
|
{
|
|
|
|
uvm_va_range_t *prev;
|
|
|
|
NvU64 gmmu_region_base = UVM_ALIGN_DOWN(address, UVM_GMMU_ATS_GRANULARITY);
|
|
|
|
|
|
|
|
UVM_ASSERT(va_space);
|
|
|
|
|
|
|
|
if (next) {
|
|
|
|
if (next->node.start <= gmmu_region_base + UVM_GMMU_ATS_GRANULARITY - 1)
|
|
|
|
return true;
|
|
|
|
|
|
|
|
prev = uvm_va_range_container(uvm_range_tree_prev(&va_space->va_range_tree, &next->node));
|
|
|
|
}
|
|
|
|
else {
|
|
|
|
// No VA range exists after address, so check the last VA range in the
|
|
|
|
// tree.
|
|
|
|
prev = uvm_va_range_container(uvm_range_tree_last(&va_space->va_range_tree));
|
|
|
|
}
|
|
|
|
|
|
|
|
return prev && (prev->node.end >= gmmu_region_base);
|
|
|
|
}
|
|
|
|
|
2022-05-09 22:18:59 +02:00
|
|
|
NV_STATUS uvm_ats_invalidate_tlbs(uvm_gpu_va_space_t *gpu_va_space,
|
|
|
|
uvm_ats_fault_invalidate_t *ats_invalidate,
|
|
|
|
uvm_tracker_t *out_tracker)
|
|
|
|
{
|
|
|
|
NV_STATUS status;
|
|
|
|
uvm_push_t push;
|
|
|
|
|
|
|
|
if (!ats_invalidate->write_faults_in_batch)
|
|
|
|
return NV_OK;
|
|
|
|
|
|
|
|
UVM_ASSERT(gpu_va_space);
|
|
|
|
UVM_ASSERT(gpu_va_space->ats.enabled);
|
|
|
|
|
|
|
|
status = uvm_push_begin(gpu_va_space->gpu->channel_manager,
|
|
|
|
UVM_CHANNEL_TYPE_MEMOPS,
|
|
|
|
&push,
|
|
|
|
"Invalidate ATS entries");
|
|
|
|
|
|
|
|
if (status == NV_OK) {
|
|
|
|
uvm_tlb_batch_end(&ats_invalidate->write_faults_tlb_batch, &push, UVM_MEMBAR_NONE);
|
|
|
|
uvm_push_end(&push);
|
|
|
|
|
|
|
|
// Add this push to the GPU's tracker so that fault replays/clears can
|
|
|
|
// wait on it
|
|
|
|
status = uvm_tracker_add_push_safe(out_tracker, &push);
|
|
|
|
}
|
|
|
|
|
|
|
|
ats_invalidate->write_faults_in_batch = false;
|
|
|
|
|
|
|
|
return status;
|
|
|
|
}
|
2023-05-30 19:11:36 +02:00
|
|
|
|