open-gpu-kernel-modules/kernel-open/nvidia-uvm/uvm_gpu_replayable_faults.c

2794 lines
118 KiB
C
Raw Normal View History

2022-05-09 22:18:59 +02:00
/*******************************************************************************
2022-11-10 17:39:33 +01:00
Copyright (c) 2015-2022 NVIDIA Corporation
2022-05-09 22:18:59 +02:00
Permission is hereby granted, free of charge, to any person obtaining a copy
of this software and associated documentation files (the "Software"), to
deal in the Software without restriction, including without limitation the
rights to use, copy, modify, merge, publish, distribute, sublicense, and/or
sell copies of the Software, and to permit persons to whom the Software is
furnished to do so, subject to the following conditions:
The above copyright notice and this permission notice shall be
included in all copies or substantial portions of the Software.
THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING
FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER
DEALINGS IN THE SOFTWARE.
*******************************************************************************/
#include "linux/sort.h"
#include "nv_uvm_interface.h"
2023-05-30 19:11:36 +02:00
#include "uvm_common.h"
2022-05-09 22:18:59 +02:00
#include "uvm_linux.h"
#include "uvm_global.h"
#include "uvm_gpu_replayable_faults.h"
#include "uvm_hal.h"
#include "uvm_kvmalloc.h"
#include "uvm_tools.h"
#include "uvm_va_block.h"
#include "uvm_va_range.h"
#include "uvm_va_space.h"
#include "uvm_va_space_mm.h"
#include "uvm_procfs.h"
#include "uvm_perf_thrashing.h"
#include "uvm_gpu_non_replayable_faults.h"
#include "uvm_ats_faults.h"
#include "uvm_test.h"
// The documentation at the beginning of uvm_gpu_non_replayable_faults.c
// provides some background for understanding replayable faults, non-replayable
// faults, and how UVM services each fault type.
#define UVM_PERF_REENABLE_PREFETCH_FAULTS_LAPSE_MSEC_DEFAULT 1000
// Lapse of time in milliseconds after which prefetch faults can be re-enabled.
// 0 means it is never disabled
static unsigned uvm_perf_reenable_prefetch_faults_lapse_msec = UVM_PERF_REENABLE_PREFETCH_FAULTS_LAPSE_MSEC_DEFAULT;
module_param(uvm_perf_reenable_prefetch_faults_lapse_msec, uint, S_IRUGO);
#define UVM_PERF_FAULT_BATCH_COUNT_MIN 1
#define UVM_PERF_FAULT_BATCH_COUNT_DEFAULT 256
// Number of entries that are fetched from the GPU fault buffer and serviced in
// batch
static unsigned uvm_perf_fault_batch_count = UVM_PERF_FAULT_BATCH_COUNT_DEFAULT;
module_param(uvm_perf_fault_batch_count, uint, S_IRUGO);
#define UVM_PERF_FAULT_REPLAY_POLICY_DEFAULT UVM_PERF_FAULT_REPLAY_POLICY_BATCH_FLUSH
// Policy that determines when to issue fault replays
static uvm_perf_fault_replay_policy_t uvm_perf_fault_replay_policy = UVM_PERF_FAULT_REPLAY_POLICY_DEFAULT;
module_param(uvm_perf_fault_replay_policy, uint, S_IRUGO);
#define UVM_PERF_FAULT_REPLAY_UPDATE_PUT_RATIO_DEFAULT 50
// Reading fault buffer GET/PUT pointers from the CPU is expensive. However,
// updating PUT before flushing the buffer helps minimizing the number of
// duplicates in the buffer as it discards faults that were not processed
// because of the batch size limit or because they arrived during servicing.
// If PUT is not updated, the replay operation will make them show up again
// in the buffer as duplicates.
//
// We keep track of the number of duplicates in each batch and we use
// UVM_GPU_BUFFER_FLUSH_MODE_UPDATE_PUT for the fault buffer flush after if the
// percentage of duplicate faults in a batch is greater than the ratio defined
// in the following module parameter. UVM_GPU_BUFFER_FLUSH_MODE_CACHED_PUT is
// used, otherwise.
static unsigned uvm_perf_fault_replay_update_put_ratio = UVM_PERF_FAULT_REPLAY_UPDATE_PUT_RATIO_DEFAULT;
module_param(uvm_perf_fault_replay_update_put_ratio, uint, S_IRUGO);
#define UVM_PERF_FAULT_MAX_BATCHES_PER_SERVICE_DEFAULT 20
#define UVM_PERF_FAULT_MAX_THROTTLE_PER_SERVICE_DEFAULT 5
// Maximum number of batches to be processed per execution of the bottom-half
static unsigned uvm_perf_fault_max_batches_per_service = UVM_PERF_FAULT_MAX_BATCHES_PER_SERVICE_DEFAULT;
module_param(uvm_perf_fault_max_batches_per_service, uint, S_IRUGO);
// Maximum number of batches with thrashing pages per execution of the bottom-half
static unsigned uvm_perf_fault_max_throttle_per_service = UVM_PERF_FAULT_MAX_THROTTLE_PER_SERVICE_DEFAULT;
module_param(uvm_perf_fault_max_throttle_per_service, uint, S_IRUGO);
static unsigned uvm_perf_fault_coalesce = 1;
module_param(uvm_perf_fault_coalesce, uint, S_IRUGO);
// This function is used for both the initial fault buffer initialization and
// the power management resume path.
static void fault_buffer_reinit_replayable_faults(uvm_parent_gpu_t *parent_gpu)
{
uvm_replayable_fault_buffer_info_t *replayable_faults = &parent_gpu->fault_buffer_info.replayable;
// Read the current get/put pointers, as this might not be the first time
// we take control of the fault buffer since the GPU was initialized,
// or since we may need to bring UVM's cached copies back in sync following
// a sleep cycle.
replayable_faults->cached_get = parent_gpu->fault_buffer_hal->read_get(parent_gpu);
replayable_faults->cached_put = parent_gpu->fault_buffer_hal->read_put(parent_gpu);
// (Re-)enable fault prefetching
if (parent_gpu->fault_buffer_info.prefetch_faults_enabled)
parent_gpu->arch_hal->enable_prefetch_faults(parent_gpu);
else
parent_gpu->arch_hal->disable_prefetch_faults(parent_gpu);
}
// There is no error handling in this function. The caller is in charge of
// calling fault_buffer_deinit_replayable_faults on failure.
static NV_STATUS fault_buffer_init_replayable_faults(uvm_parent_gpu_t *parent_gpu)
{
NV_STATUS status = NV_OK;
uvm_replayable_fault_buffer_info_t *replayable_faults = &parent_gpu->fault_buffer_info.replayable;
uvm_fault_service_batch_context_t *batch_context = &replayable_faults->batch_service_context;
UVM_ASSERT(parent_gpu->fault_buffer_info.rm_info.replayable.bufferSize %
parent_gpu->fault_buffer_hal->entry_size(parent_gpu) == 0);
replayable_faults->max_faults = parent_gpu->fault_buffer_info.rm_info.replayable.bufferSize /
parent_gpu->fault_buffer_hal->entry_size(parent_gpu);
// Check provided module parameter value
parent_gpu->fault_buffer_info.max_batch_size = max(uvm_perf_fault_batch_count,
(NvU32)UVM_PERF_FAULT_BATCH_COUNT_MIN);
parent_gpu->fault_buffer_info.max_batch_size = min(parent_gpu->fault_buffer_info.max_batch_size,
replayable_faults->max_faults);
if (parent_gpu->fault_buffer_info.max_batch_size != uvm_perf_fault_batch_count) {
pr_info("Invalid uvm_perf_fault_batch_count value on GPU %s: %u. Valid range [%u:%u] Using %u instead\n",
parent_gpu->name,
uvm_perf_fault_batch_count,
UVM_PERF_FAULT_BATCH_COUNT_MIN,
replayable_faults->max_faults,
parent_gpu->fault_buffer_info.max_batch_size);
}
batch_context->fault_cache = uvm_kvmalloc_zero(replayable_faults->max_faults * sizeof(*batch_context->fault_cache));
if (!batch_context->fault_cache)
return NV_ERR_NO_MEMORY;
// fault_cache is used to signal that the tracker was initialized.
uvm_tracker_init(&replayable_faults->replay_tracker);
batch_context->ordered_fault_cache = uvm_kvmalloc_zero(replayable_faults->max_faults *
sizeof(*batch_context->ordered_fault_cache));
if (!batch_context->ordered_fault_cache)
return NV_ERR_NO_MEMORY;
// This value must be initialized by HAL
UVM_ASSERT(replayable_faults->utlb_count > 0);
batch_context->utlbs = uvm_kvmalloc_zero(replayable_faults->utlb_count * sizeof(*batch_context->utlbs));
if (!batch_context->utlbs)
return NV_ERR_NO_MEMORY;
batch_context->max_utlb_id = 0;
status = uvm_rm_locked_call(nvUvmInterfaceOwnPageFaultIntr(parent_gpu->rm_device, NV_TRUE));
if (status != NV_OK) {
UVM_ERR_PRINT("Failed to take page fault ownership from RM: %s, GPU %s\n",
nvstatusToString(status),
parent_gpu->name);
return status;
}
replayable_faults->replay_policy = uvm_perf_fault_replay_policy < UVM_PERF_FAULT_REPLAY_POLICY_MAX?
uvm_perf_fault_replay_policy:
UVM_PERF_FAULT_REPLAY_POLICY_DEFAULT;
if (replayable_faults->replay_policy != uvm_perf_fault_replay_policy) {
pr_info("Invalid uvm_perf_fault_replay_policy value on GPU %s: %d. Using %d instead\n",
parent_gpu->name,
uvm_perf_fault_replay_policy,
replayable_faults->replay_policy);
}
replayable_faults->replay_update_put_ratio = min(uvm_perf_fault_replay_update_put_ratio, 100u);
if (replayable_faults->replay_update_put_ratio != uvm_perf_fault_replay_update_put_ratio) {
pr_info("Invalid uvm_perf_fault_replay_update_put_ratio value on GPU %s: %u. Using %u instead\n",
parent_gpu->name,
uvm_perf_fault_replay_update_put_ratio,
replayable_faults->replay_update_put_ratio);
}
// Re-enable fault prefetching just in case it was disabled in a previous run
parent_gpu->fault_buffer_info.prefetch_faults_enabled = parent_gpu->prefetch_fault_supported;
fault_buffer_reinit_replayable_faults(parent_gpu);
return NV_OK;
}
static void fault_buffer_deinit_replayable_faults(uvm_parent_gpu_t *parent_gpu)
{
uvm_replayable_fault_buffer_info_t *replayable_faults = &parent_gpu->fault_buffer_info.replayable;
uvm_fault_service_batch_context_t *batch_context = &replayable_faults->batch_service_context;
if (batch_context->fault_cache) {
UVM_ASSERT(uvm_tracker_is_empty(&replayable_faults->replay_tracker));
uvm_tracker_deinit(&replayable_faults->replay_tracker);
}
if (parent_gpu->fault_buffer_info.rm_info.faultBufferHandle) {
// Re-enable prefetch faults in case we disabled them
if (parent_gpu->prefetch_fault_supported && !parent_gpu->fault_buffer_info.prefetch_faults_enabled)
parent_gpu->arch_hal->enable_prefetch_faults(parent_gpu);
}
uvm_kvfree(batch_context->fault_cache);
uvm_kvfree(batch_context->ordered_fault_cache);
uvm_kvfree(batch_context->utlbs);
batch_context->fault_cache = NULL;
batch_context->ordered_fault_cache = NULL;
batch_context->utlbs = NULL;
}
NV_STATUS uvm_gpu_fault_buffer_init(uvm_parent_gpu_t *parent_gpu)
{
NV_STATUS status = NV_OK;
uvm_assert_mutex_locked(&g_uvm_global.global_lock);
UVM_ASSERT(parent_gpu->replayable_faults_supported);
status = uvm_rm_locked_call(nvUvmInterfaceInitFaultInfo(parent_gpu->rm_device,
&parent_gpu->fault_buffer_info.rm_info));
if (status != NV_OK) {
UVM_ERR_PRINT("Failed to init fault buffer info from RM: %s, GPU %s\n",
nvstatusToString(status),
parent_gpu->name);
// nvUvmInterfaceInitFaultInfo may leave fields in rm_info populated
// when it returns an error. Set the buffer handle to zero as it is
// used by the deinitialization logic to determine if it was correctly
// initialized.
parent_gpu->fault_buffer_info.rm_info.faultBufferHandle = 0;
goto fail;
}
status = fault_buffer_init_replayable_faults(parent_gpu);
if (status != NV_OK)
goto fail;
if (parent_gpu->non_replayable_faults_supported) {
status = uvm_gpu_fault_buffer_init_non_replayable_faults(parent_gpu);
if (status != NV_OK)
goto fail;
}
return NV_OK;
fail:
uvm_gpu_fault_buffer_deinit(parent_gpu);
return status;
}
// Reinitialize state relevant to replayable fault handling after returning
// from a power management cycle.
void uvm_gpu_fault_buffer_resume(uvm_parent_gpu_t *parent_gpu)
{
UVM_ASSERT(parent_gpu->replayable_faults_supported);
fault_buffer_reinit_replayable_faults(parent_gpu);
}
void uvm_gpu_fault_buffer_deinit(uvm_parent_gpu_t *parent_gpu)
{
NV_STATUS status = NV_OK;
uvm_assert_mutex_locked(&g_uvm_global.global_lock);
if (parent_gpu->non_replayable_faults_supported)
uvm_gpu_fault_buffer_deinit_non_replayable_faults(parent_gpu);
fault_buffer_deinit_replayable_faults(parent_gpu);
if (parent_gpu->fault_buffer_info.rm_info.faultBufferHandle) {
status = uvm_rm_locked_call(nvUvmInterfaceOwnPageFaultIntr(parent_gpu->rm_device, NV_FALSE));
UVM_ASSERT(status == NV_OK);
uvm_rm_locked_call_void(nvUvmInterfaceDestroyFaultInfo(parent_gpu->rm_device,
&parent_gpu->fault_buffer_info.rm_info));
parent_gpu->fault_buffer_info.rm_info.faultBufferHandle = 0;
}
}
2023-05-30 19:11:36 +02:00
// TODO: Bug 4098289: this function can be removed, and the calls to it replaced
// with calls to uvm_conf_computing_mode_enabled_parent, once UVM ownership is
// dictated by Confidential Computing enablement. Currently we support a
// non-production scenario in which Confidential Computing is enabled, but
// UVM still owns the replayable fault buffer.
bool uvm_parent_gpu_replayable_fault_buffer_is_uvm_owned(uvm_parent_gpu_t *parent_gpu)
{
if (uvm_conf_computing_mode_enabled_parent(parent_gpu))
return parent_gpu->fault_buffer_info.rm_info.replayable.bUvmOwnsHwFaultBuffer;
return true;
}
2022-05-09 22:18:59 +02:00
bool uvm_gpu_replayable_faults_pending(uvm_parent_gpu_t *parent_gpu)
{
uvm_replayable_fault_buffer_info_t *replayable_faults = &parent_gpu->fault_buffer_info.replayable;
UVM_ASSERT(parent_gpu->replayable_faults_supported);
// Fast path 1: we left some faults unserviced in the buffer in the last pass
if (replayable_faults->cached_get != replayable_faults->cached_put)
return true;
// Fast path 2: read the valid bit of the fault buffer entry pointed by the
// cached get pointer
if (!parent_gpu->fault_buffer_hal->entry_is_valid(parent_gpu, replayable_faults->cached_get)) {
// Slow path: read the put pointer from the GPU register via BAR0
// over PCIe
replayable_faults->cached_put = parent_gpu->fault_buffer_hal->read_put(parent_gpu);
// No interrupt pending
if (replayable_faults->cached_get == replayable_faults->cached_put)
return false;
}
return true;
}
// Push a fault cancel method on the given client. Any failure during this
// operation may lead to application hang (requiring manual Ctrl+C from the
// user) or system crash (requiring reboot).
// In that case we log an error message.
//
// gpc_id and client_id aren't used if global_cancel is true.
//
// This function acquires both the given tracker and the replay tracker
static NV_STATUS push_cancel_on_gpu(uvm_gpu_t *gpu,
uvm_gpu_phys_address_t instance_ptr,
bool global_cancel,
NvU32 gpc_id,
NvU32 client_id,
uvm_tracker_t *tracker)
{
NV_STATUS status;
uvm_push_t push;
uvm_replayable_fault_buffer_info_t *replayable_faults = &gpu->parent->fault_buffer_info.replayable;
if (global_cancel) {
status = uvm_push_begin_acquire(gpu->channel_manager,
UVM_CHANNEL_TYPE_MEMOPS,
&replayable_faults->replay_tracker,
&push,
"Cancel targeting instance_ptr {0x%llx:%s}\n",
instance_ptr.address,
uvm_aperture_string(instance_ptr.aperture));
} else {
status = uvm_push_begin_acquire(gpu->channel_manager,
UVM_CHANNEL_TYPE_MEMOPS,
&replayable_faults->replay_tracker,
&push,
"Cancel targeting instance_ptr {0x%llx:%s} gpc %u client %u\n",
instance_ptr.address,
uvm_aperture_string(instance_ptr.aperture),
gpc_id,
client_id);
}
UVM_ASSERT(status == NV_OK);
if (status != NV_OK) {
UVM_ERR_PRINT("Failed to create push and acquire replay tracker before pushing cancel: %s, GPU %s\n",
nvstatusToString(status),
uvm_gpu_name(gpu));
return status;
}
uvm_push_acquire_tracker(&push, tracker);
if (global_cancel)
gpu->parent->host_hal->cancel_faults_global(&push, instance_ptr);
else
gpu->parent->host_hal->cancel_faults_targeted(&push, instance_ptr, gpc_id, client_id);
// We don't need to put the cancel in the GPU replay tracker since we wait
// on it immediately.
status = uvm_push_end_and_wait(&push);
UVM_ASSERT(status == NV_OK);
if (status != NV_OK)
UVM_ERR_PRINT("Failed to wait for pushed cancel: %s, GPU %s\n", nvstatusToString(status), uvm_gpu_name(gpu));
uvm_tracker_clear(&replayable_faults->replay_tracker);
return status;
}
static NV_STATUS push_cancel_on_gpu_targeted(uvm_gpu_t *gpu,
uvm_gpu_phys_address_t instance_ptr,
NvU32 gpc_id,
NvU32 client_id,
uvm_tracker_t *tracker)
{
return push_cancel_on_gpu(gpu, instance_ptr, false, gpc_id, client_id, tracker);
}
static NV_STATUS push_cancel_on_gpu_global(uvm_gpu_t *gpu, uvm_gpu_phys_address_t instance_ptr, uvm_tracker_t *tracker)
{
UVM_ASSERT(!gpu->parent->smc.enabled);
return push_cancel_on_gpu(gpu, instance_ptr, true, 0, 0, tracker);
}
// Volta implements a targeted VA fault cancel that simplifies the fault cancel
// process. You only need to specify the address, type, and mmu_engine_id for
// the access to be cancelled. Caller must hold the VA space lock for the access
// to be cancelled.
static NV_STATUS cancel_fault_precise_va(uvm_gpu_t *gpu,
uvm_fault_buffer_entry_t *fault_entry,
uvm_fault_cancel_va_mode_t cancel_va_mode)
{
NV_STATUS status;
uvm_gpu_va_space_t *gpu_va_space;
uvm_gpu_phys_address_t pdb;
uvm_push_t push;
uvm_replayable_fault_buffer_info_t *replayable_faults = &gpu->parent->fault_buffer_info.replayable;
NvU64 offset;
UVM_ASSERT(gpu->parent->replayable_faults_supported);
UVM_ASSERT(fault_entry->fatal_reason != UvmEventFatalReasonInvalid);
UVM_ASSERT(!fault_entry->filtered);
gpu_va_space = uvm_gpu_va_space_get_by_parent_gpu(fault_entry->va_space, gpu->parent);
UVM_ASSERT(gpu_va_space);
pdb = uvm_page_tree_pdb(&gpu_va_space->page_tables)->addr;
// Record fatal fault event
uvm_tools_record_gpu_fatal_fault(gpu->id, fault_entry->va_space, fault_entry, fault_entry->fatal_reason);
status = uvm_push_begin_acquire(gpu->channel_manager,
UVM_CHANNEL_TYPE_MEMOPS,
&replayable_faults->replay_tracker,
&push,
"Precise cancel targeting PDB {0x%llx:%s} VA 0x%llx VEID %u with access type %s",
pdb.address,
uvm_aperture_string(pdb.aperture),
fault_entry->fault_address,
fault_entry->fault_source.ve_id,
uvm_fault_access_type_string(fault_entry->fault_access_type));
if (status != NV_OK) {
UVM_ERR_PRINT("Failed to create push and acquire replay tracker before pushing cancel: %s, GPU %s\n",
nvstatusToString(status),
uvm_gpu_name(gpu));
return status;
}
// UVM aligns fault addresses to PAGE_SIZE as it is the smallest mapping
// and coherence tracking granularity. However, the cancel method requires
// the original address (4K-aligned) reported in the packet, which is lost
// at this point. Since the access permissions are the same for the whole
// 64K page, we issue a cancel per 4K range to make sure that the HW sees
// the address reported in the packet.
for (offset = 0; offset < PAGE_SIZE; offset += UVM_PAGE_SIZE_4K) {
gpu->parent->host_hal->cancel_faults_va(&push, pdb, fault_entry, cancel_va_mode);
fault_entry->fault_address += UVM_PAGE_SIZE_4K;
}
fault_entry->fault_address = UVM_PAGE_ALIGN_DOWN(fault_entry->fault_address - 1);
// We don't need to put the cancel in the GPU replay tracker since we wait
// on it immediately.
status = uvm_push_end_and_wait(&push);
if (status != NV_OK) {
UVM_ERR_PRINT("Failed to wait for pushed VA global fault cancel: %s, GPU %s\n",
nvstatusToString(status), uvm_gpu_name(gpu));
}
uvm_tracker_clear(&replayable_faults->replay_tracker);
return status;
}
static NV_STATUS push_replay_on_gpu(uvm_gpu_t *gpu, uvm_fault_replay_type_t type, uvm_fault_service_batch_context_t *batch_context)
{
NV_STATUS status;
uvm_push_t push;
uvm_replayable_fault_buffer_info_t *replayable_faults = &gpu->parent->fault_buffer_info.replayable;
uvm_tracker_t *tracker = NULL;
if (batch_context)
tracker = &batch_context->tracker;
status = uvm_push_begin_acquire(gpu->channel_manager, UVM_CHANNEL_TYPE_MEMOPS, tracker, &push,
"Replaying faults");
if (status != NV_OK)
return status;
gpu->parent->host_hal->replay_faults(&push, type);
// Do not count REPLAY_TYPE_START_ACK_ALL's toward the replay count.
// REPLAY_TYPE_START_ACK_ALL's are issued for cancels, and the cancel
// algorithm checks to make sure that no REPLAY_TYPE_START's have been
// issued using batch_context->replays.
if (batch_context && type != UVM_FAULT_REPLAY_TYPE_START_ACK_ALL) {
uvm_tools_broadcast_replay(gpu, &push, batch_context->batch_id, UVM_FAULT_CLIENT_TYPE_GPC);
++batch_context->num_replays;
}
uvm_push_end(&push);
// Add this push to the GPU's replay_tracker so cancel can wait on it.
status = uvm_tracker_add_push_safe(&replayable_faults->replay_tracker, &push);
if (uvm_procfs_is_debug_enabled()) {
if (type == UVM_FAULT_REPLAY_TYPE_START)
++replayable_faults->stats.num_replays;
else
++replayable_faults->stats.num_replays_ack_all;
}
return status;
}
static void write_get(uvm_parent_gpu_t *parent_gpu, NvU32 get)
{
uvm_replayable_fault_buffer_info_t *replayable_faults = &parent_gpu->fault_buffer_info.replayable;
UVM_ASSERT(uvm_sem_is_locked(&parent_gpu->isr.replayable_faults.service_lock));
// Write get on the GPU only if it's changed.
if (replayable_faults->cached_get == get)
return;
replayable_faults->cached_get = get;
// Update get pointer on the GPU
parent_gpu->fault_buffer_hal->write_get(parent_gpu, get);
}
2023-05-30 19:11:36 +02:00
static NV_STATUS hw_fault_buffer_flush_locked(uvm_parent_gpu_t *parent_gpu)
{
NV_STATUS status = NV_OK;
// When Confidential Computing is enabled, GSP-RM owns the HW replayable
// fault buffer. Flushing the fault buffer implies flushing both the HW
// buffer (using a RM API), and the SW buffer accessible by UVM ("shadow"
// buffer).
//
// The HW buffer needs to be flushed first. This is because, once that
// flush completes, any faults that were present in the HW buffer when
// fault_buffer_flush_locked is called, are now either flushed from the HW
// buffer, or are present in the shadow buffer and are about to be discarded
// too.
if (!uvm_conf_computing_mode_enabled_parent(parent_gpu))
return NV_OK;
// nvUvmInterfaceFlushReplayableFaultBuffer relies on the caller to ensure
// serialization for a given GPU.
UVM_ASSERT(uvm_sem_is_locked(&parent_gpu->isr.replayable_faults.service_lock));
// Flush the HW replayable buffer owned by GSP-RM.
status = nvUvmInterfaceFlushReplayableFaultBuffer(parent_gpu->rm_device);
UVM_ASSERT(status == NV_OK);
return status;
}
2022-05-09 22:18:59 +02:00
static NV_STATUS fault_buffer_flush_locked(uvm_gpu_t *gpu,
uvm_gpu_buffer_flush_mode_t flush_mode,
uvm_fault_replay_type_t fault_replay,
uvm_fault_service_batch_context_t *batch_context)
{
NvU32 get;
NvU32 put;
uvm_spin_loop_t spin;
2023-05-30 19:11:36 +02:00
uvm_parent_gpu_t *parent_gpu = gpu->parent;
uvm_replayable_fault_buffer_info_t *replayable_faults = &parent_gpu->fault_buffer_info.replayable;
NV_STATUS status;
2022-05-09 22:18:59 +02:00
2023-05-30 19:11:36 +02:00
UVM_ASSERT(uvm_sem_is_locked(&parent_gpu->isr.replayable_faults.service_lock));
UVM_ASSERT(parent_gpu->replayable_faults_supported);
// Wait for the prior replay to flush out old fault messages
if (flush_mode == UVM_GPU_BUFFER_FLUSH_MODE_WAIT_UPDATE_PUT) {
status = uvm_tracker_wait(&replayable_faults->replay_tracker);
if (status != NV_OK)
return status;
}
2022-05-09 22:18:59 +02:00
// Read PUT pointer from the GPU if requested
2023-05-30 19:11:36 +02:00
if (flush_mode == UVM_GPU_BUFFER_FLUSH_MODE_UPDATE_PUT || flush_mode == UVM_GPU_BUFFER_FLUSH_MODE_WAIT_UPDATE_PUT) {
status = hw_fault_buffer_flush_locked(parent_gpu);
if (status != NV_OK)
return status;
replayable_faults->cached_put = parent_gpu->fault_buffer_hal->read_put(parent_gpu);
}
2022-05-09 22:18:59 +02:00
get = replayable_faults->cached_get;
put = replayable_faults->cached_put;
while (get != put) {
// Wait until valid bit is set
2023-05-30 19:11:36 +02:00
UVM_SPIN_WHILE(!parent_gpu->fault_buffer_hal->entry_is_valid(parent_gpu, get), &spin);
2022-05-09 22:18:59 +02:00
2023-05-30 19:11:36 +02:00
parent_gpu->fault_buffer_hal->entry_clear_valid(parent_gpu, get);
2022-05-09 22:18:59 +02:00
++get;
if (get == replayable_faults->max_faults)
get = 0;
}
write_get(gpu->parent, get);
// Issue fault replay
return push_replay_on_gpu(gpu, fault_replay, batch_context);
}
NV_STATUS uvm_gpu_fault_buffer_flush(uvm_gpu_t *gpu)
{
NV_STATUS status = NV_OK;
UVM_ASSERT(gpu->parent->replayable_faults_supported);
// Disables replayable fault interrupts and fault servicing
uvm_gpu_replayable_faults_isr_lock(gpu->parent);
status = fault_buffer_flush_locked(gpu,
2023-05-30 19:11:36 +02:00
UVM_GPU_BUFFER_FLUSH_MODE_WAIT_UPDATE_PUT,
2022-05-09 22:18:59 +02:00
UVM_FAULT_REPLAY_TYPE_START,
NULL);
// This will trigger the top half to start servicing faults again, if the
// replay brought any back in
uvm_gpu_replayable_faults_isr_unlock(gpu->parent);
return status;
}
static inline int cmp_fault_instance_ptr(const uvm_fault_buffer_entry_t *a,
const uvm_fault_buffer_entry_t *b)
{
int result = uvm_gpu_phys_addr_cmp(a->instance_ptr, b->instance_ptr);
// On Volta+ we need to sort by {instance_ptr + subctx_id} pair since it can
// map to a different VA space
if (result != 0)
return result;
return UVM_CMP_DEFAULT(a->fault_source.ve_id, b->fault_source.ve_id);
}
// Compare two VA spaces
static inline int cmp_va_space(const uvm_va_space_t *a, const uvm_va_space_t *b)
{
return UVM_CMP_DEFAULT(a, b);
}
// Compare two virtual addresses
static inline int cmp_addr(NvU64 a, NvU64 b)
{
return UVM_CMP_DEFAULT(a, b);
}
// Compare two fault access types
static inline int cmp_access_type(uvm_fault_access_type_t a, uvm_fault_access_type_t b)
{
UVM_ASSERT(a >= 0 && a < UVM_FAULT_ACCESS_TYPE_COUNT);
UVM_ASSERT(b >= 0 && b < UVM_FAULT_ACCESS_TYPE_COUNT);
// Check that fault access type enum values are ordered by "intrusiveness"
BUILD_BUG_ON(UVM_FAULT_ACCESS_TYPE_ATOMIC_STRONG <= UVM_FAULT_ACCESS_TYPE_ATOMIC_WEAK);
BUILD_BUG_ON(UVM_FAULT_ACCESS_TYPE_ATOMIC_WEAK <= UVM_FAULT_ACCESS_TYPE_WRITE);
BUILD_BUG_ON(UVM_FAULT_ACCESS_TYPE_WRITE <= UVM_FAULT_ACCESS_TYPE_READ);
BUILD_BUG_ON(UVM_FAULT_ACCESS_TYPE_READ <= UVM_FAULT_ACCESS_TYPE_PREFETCH);
return b - a;
}
typedef enum
{
// Fetch a batch of faults from the buffer.
FAULT_FETCH_MODE_BATCH_ALL,
// Fetch a batch of faults from the buffer. Stop at the first entry that is
// not ready yet
FAULT_FETCH_MODE_BATCH_READY,
// Fetch all faults in the buffer before PUT. Wait for all faults to become
// ready
FAULT_FETCH_MODE_ALL,
} fault_fetch_mode_t;
static void fetch_fault_buffer_merge_entry(uvm_fault_buffer_entry_t *current_entry,
uvm_fault_buffer_entry_t *last_entry)
{
UVM_ASSERT(last_entry->num_instances > 0);
++last_entry->num_instances;
uvm_fault_access_type_mask_set(&last_entry->access_type_mask, current_entry->fault_access_type);
if (current_entry->fault_access_type > last_entry->fault_access_type) {
// If the new entry has a higher access type, it becomes the
// fault to be serviced. Add the previous one to the list of instances
current_entry->access_type_mask = last_entry->access_type_mask;
current_entry->num_instances = last_entry->num_instances;
last_entry->filtered = true;
// We only merge faults from different uTLBs if the new fault has an
// access type with the same or lower level of intrusiveness.
UVM_ASSERT(current_entry->fault_source.utlb_id == last_entry->fault_source.utlb_id);
list_replace(&last_entry->merged_instances_list, &current_entry->merged_instances_list);
list_add(&last_entry->merged_instances_list, &current_entry->merged_instances_list);
}
else {
// Add the new entry to the list of instances for reporting purposes
current_entry->filtered = true;
list_add(&current_entry->merged_instances_list, &last_entry->merged_instances_list);
}
}
static bool fetch_fault_buffer_try_merge_entry(uvm_fault_buffer_entry_t *current_entry,
uvm_fault_service_batch_context_t *batch_context,
uvm_fault_utlb_info_t *current_tlb,
bool is_same_instance_ptr)
{
uvm_fault_buffer_entry_t *last_tlb_entry = current_tlb->last_fault;
uvm_fault_buffer_entry_t *last_global_entry = batch_context->last_fault;
// Check the last coalesced fault and the coalesced fault that was
// originated from this uTLB
const bool is_last_tlb_fault = current_tlb->num_pending_faults > 0 &&
cmp_fault_instance_ptr(current_entry, last_tlb_entry) == 0 &&
current_entry->fault_address == last_tlb_entry->fault_address;
// We only merge faults from different uTLBs if the new fault has an
// access type with the same or lower level of intrusiveness. This is to
// avoid having to update num_pending_faults on both uTLBs and recomputing
// last_fault.
const bool is_last_fault = is_same_instance_ptr &&
current_entry->fault_address == last_global_entry->fault_address &&
current_entry->fault_access_type <= last_global_entry->fault_access_type;
if (is_last_tlb_fault) {
fetch_fault_buffer_merge_entry(current_entry, last_tlb_entry);
if (current_entry->fault_access_type > last_tlb_entry->fault_access_type)
current_tlb->last_fault = current_entry;
return true;
}
else if (is_last_fault) {
fetch_fault_buffer_merge_entry(current_entry, last_global_entry);
if (current_entry->fault_access_type > last_global_entry->fault_access_type)
batch_context->last_fault = current_entry;
return true;
}
return false;
}
// Fetch entries from the fault buffer, decode them and store them in the batch
// context. We implement the fetch modes described above.
//
// When possible, we coalesce duplicate entries to minimize the fault handling
// overhead. Basically, we merge faults with the same instance pointer and page
// virtual address. We keep track of the last fault per uTLB to detect
// duplicates due to local reuse and the last fault in the whole batch to
// detect reuse across CTAs.
//
// We will service the first fault entry with the most "intrusive" (atomic >
// write > read > prefetch) access type*. That fault entry is called the
// "representative". The rest of filtered faults have the "filtered" flag set
// and are added to a list in the representative fault entry for reporting
// purposes. The representative fault entry also contains a mask with all the
// access types that produced a fault on the page.
//
// *We only merge faults from different uTLBs if the new fault has an access
// type with the same or lower level of intrusiveness.
//
// This optimization cannot be performed during fault cancel on Pascal GPUs
// (fetch_mode == FAULT_FETCH_MODE_ALL) since we need accurate tracking of all
// the faults in each uTLB in order to guarantee precise fault attribution.
static void fetch_fault_buffer_entries(uvm_gpu_t *gpu,
uvm_fault_service_batch_context_t *batch_context,
fault_fetch_mode_t fetch_mode)
{
NvU32 get;
NvU32 put;
NvU32 fault_index;
NvU32 num_coalesced_faults;
NvU32 utlb_id;
uvm_fault_buffer_entry_t *fault_cache;
uvm_spin_loop_t spin;
uvm_replayable_fault_buffer_info_t *replayable_faults = &gpu->parent->fault_buffer_info.replayable;
const bool in_pascal_cancel_path = (!gpu->parent->fault_cancel_va_supported && fetch_mode == FAULT_FETCH_MODE_ALL);
const bool may_filter = uvm_perf_fault_coalesce && !in_pascal_cancel_path;
UVM_ASSERT(uvm_sem_is_locked(&gpu->parent->isr.replayable_faults.service_lock));
UVM_ASSERT(gpu->parent->replayable_faults_supported);
fault_cache = batch_context->fault_cache;
get = replayable_faults->cached_get;
// Read put pointer from GPU and cache it
if (get == replayable_faults->cached_put)
replayable_faults->cached_put = gpu->parent->fault_buffer_hal->read_put(gpu->parent);
put = replayable_faults->cached_put;
batch_context->is_single_instance_ptr = true;
batch_context->last_fault = NULL;
fault_index = 0;
num_coalesced_faults = 0;
// Clear uTLB counters
for (utlb_id = 0; utlb_id <= batch_context->max_utlb_id; ++utlb_id) {
batch_context->utlbs[utlb_id].num_pending_faults = 0;
batch_context->utlbs[utlb_id].has_fatal_faults = false;
}
batch_context->max_utlb_id = 0;
if (get == put)
goto done;
// Parse until get != put and have enough space to cache.
while ((get != put) &&
(fetch_mode == FAULT_FETCH_MODE_ALL || fault_index < gpu->parent->fault_buffer_info.max_batch_size)) {
bool is_same_instance_ptr = true;
uvm_fault_buffer_entry_t *current_entry = &fault_cache[fault_index];
uvm_fault_utlb_info_t *current_tlb;
// We cannot just wait for the last entry (the one pointed by put) to
// become valid, we have to do it individually since entries can be
// written out of order
UVM_SPIN_WHILE(!gpu->parent->fault_buffer_hal->entry_is_valid(gpu->parent, get), &spin) {
// We have some entry to work on. Let's do the rest later.
if (fetch_mode != FAULT_FETCH_MODE_ALL &&
fetch_mode != FAULT_FETCH_MODE_BATCH_ALL &&
fault_index > 0)
goto done;
}
// Prevent later accesses being moved above the read of the valid bit
smp_mb__after_atomic();
// Got valid bit set. Let's cache.
gpu->parent->fault_buffer_hal->parse_entry(gpu->parent, get, current_entry);
// The GPU aligns the fault addresses to 4k, but all of our tracking is
// done in PAGE_SIZE chunks which might be larger.
current_entry->fault_address = UVM_PAGE_ALIGN_DOWN(current_entry->fault_address);
// Make sure that all fields in the entry are properly initialized
current_entry->is_fatal = (current_entry->fault_type >= UVM_FAULT_TYPE_FATAL);
if (current_entry->is_fatal) {
// Record the fatal fault event later as we need the va_space locked
current_entry->fatal_reason = UvmEventFatalReasonInvalidFaultType;
}
else {
current_entry->fatal_reason = UvmEventFatalReasonInvalid;
}
current_entry->va_space = NULL;
current_entry->filtered = false;
if (current_entry->fault_source.utlb_id > batch_context->max_utlb_id) {
UVM_ASSERT(current_entry->fault_source.utlb_id < replayable_faults->utlb_count);
batch_context->max_utlb_id = current_entry->fault_source.utlb_id;
}
current_tlb = &batch_context->utlbs[current_entry->fault_source.utlb_id];
if (fault_index > 0) {
UVM_ASSERT(batch_context->last_fault);
is_same_instance_ptr = cmp_fault_instance_ptr(current_entry, batch_context->last_fault) == 0;
// Coalesce duplicate faults when possible
if (may_filter && !current_entry->is_fatal) {
bool merged = fetch_fault_buffer_try_merge_entry(current_entry,
batch_context,
current_tlb,
is_same_instance_ptr);
if (merged)
goto next_fault;
}
}
if (batch_context->is_single_instance_ptr && !is_same_instance_ptr)
batch_context->is_single_instance_ptr = false;
current_entry->num_instances = 1;
current_entry->access_type_mask = uvm_fault_access_type_mask_bit(current_entry->fault_access_type);
INIT_LIST_HEAD(&current_entry->merged_instances_list);
++current_tlb->num_pending_faults;
current_tlb->last_fault = current_entry;
batch_context->last_fault = current_entry;
++num_coalesced_faults;
next_fault:
++fault_index;
++get;
if (get == replayable_faults->max_faults)
get = 0;
}
done:
write_get(gpu->parent, get);
batch_context->num_cached_faults = fault_index;
batch_context->num_coalesced_faults = num_coalesced_faults;
}
// Sort comparator for pointers to fault buffer entries that sorts by
// instance pointer
static int cmp_sort_fault_entry_by_instance_ptr(const void *_a, const void *_b)
{
const uvm_fault_buffer_entry_t **a = (const uvm_fault_buffer_entry_t **)_a;
const uvm_fault_buffer_entry_t **b = (const uvm_fault_buffer_entry_t **)_b;
return cmp_fault_instance_ptr(*a, *b);
}
// Sort comparator for pointers to fault buffer entries that sorts by va_space,
// fault address and fault access type
static int cmp_sort_fault_entry_by_va_space_address_access_type(const void *_a, const void *_b)
{
const uvm_fault_buffer_entry_t **a = (const uvm_fault_buffer_entry_t **)_a;
const uvm_fault_buffer_entry_t **b = (const uvm_fault_buffer_entry_t **)_b;
int result;
result = cmp_va_space((*a)->va_space, (*b)->va_space);
if (result != 0)
return result;
result = cmp_addr((*a)->fault_address, (*b)->fault_address);
if (result != 0)
return result;
return cmp_access_type((*a)->fault_access_type, (*b)->fault_access_type);
}
// Translate all instance pointers to VA spaces. Since the buffer is ordered by
// instance_ptr, we minimize the number of translations
//
// This function returns NV_WARN_MORE_PROCESSING_REQUIRED if a fault buffer
// flush occurred and executed successfully, or the error code if it failed.
// NV_OK otherwise.
static NV_STATUS translate_instance_ptrs(uvm_gpu_t *gpu,
uvm_fault_service_batch_context_t *batch_context)
{
NvU32 i;
NV_STATUS status;
for (i = 0; i < batch_context->num_coalesced_faults; ++i) {
uvm_fault_buffer_entry_t *current_entry;
current_entry = batch_context->ordered_fault_cache[i];
// If this instance pointer matches the previous instance pointer, just
// copy over the already-translated va_space and move on.
if (i != 0 && cmp_fault_instance_ptr(current_entry, batch_context->ordered_fault_cache[i - 1]) == 0) {
current_entry->va_space = batch_context->ordered_fault_cache[i - 1]->va_space;
continue;
}
status = uvm_gpu_fault_entry_to_va_space(gpu, current_entry, &current_entry->va_space);
if (status != NV_OK) {
if (status == NV_ERR_PAGE_TABLE_NOT_AVAIL) {
// The channel is valid but the subcontext is not. This can only
// happen if the subcontext is torn down before its work is
// complete while other subcontexts in the same TSG are still
// executing. This is a violation of the programming model. We
// have limited options since the VA space is gone, meaning we
// can't target the PDB for cancel even if we wanted to. So
// we'll just throw away precise attribution and cancel this
// fault using the SW method, which validates that the intended
// context (TSG) is still running so we don't cancel an innocent
// context.
UVM_ASSERT(!current_entry->va_space);
UVM_ASSERT(gpu->max_subcontexts > 0);
if (gpu->parent->smc.enabled) {
status = push_cancel_on_gpu_targeted(gpu,
current_entry->instance_ptr,
current_entry->fault_source.gpc_id,
current_entry->fault_source.client_id,
&batch_context->tracker);
}
else {
status = push_cancel_on_gpu_global(gpu, current_entry->instance_ptr, &batch_context->tracker);
}
if (status != NV_OK)
return status;
// Fall through and let the flush restart fault processing
}
else {
UVM_ASSERT(status == NV_ERR_INVALID_CHANNEL);
}
// If the channel is gone then we're looking at a stale fault entry.
// The fault must have been resolved already (serviced or
// cancelled), so we can just flush the fault buffer.
2023-05-30 19:11:36 +02:00
//
// No need to use UVM_GPU_BUFFER_FLUSH_MODE_WAIT_UPDATE_PUT since
// there was a context preemption for the entries we want to flush,
// meaning PUT must reflect them.
2022-05-09 22:18:59 +02:00
status = fault_buffer_flush_locked(gpu,
UVM_GPU_BUFFER_FLUSH_MODE_UPDATE_PUT,
UVM_FAULT_REPLAY_TYPE_START,
batch_context);
if (status != NV_OK)
return status;
return NV_WARN_MORE_PROCESSING_REQUIRED;
}
else {
UVM_ASSERT(current_entry->va_space);
}
}
return NV_OK;
}
// Fault cache preprocessing for fault coalescing
//
// This function generates an ordered view of the given fault_cache in which
// faults are sorted by VA space, fault address (aligned to 4K) and access type
// "intrusiveness". In order to minimize the number of instance_ptr to VA space
// translations we perform a first sort by instance_ptr.
//
// This function returns NV_WARN_MORE_PROCESSING_REQUIRED if a fault buffer
// flush occurred during instance_ptr translation and executed successfully, or
// the error code if it failed. NV_OK otherwise.
//
// Current scheme:
// 1) sort by instance_ptr
// 2) translate all instance_ptrs to VA spaces
// 3) sort by va_space, fault address (fault_address is page-aligned at this
// point) and access type
static NV_STATUS preprocess_fault_batch(uvm_gpu_t *gpu, uvm_fault_service_batch_context_t *batch_context)
{
NV_STATUS status;
NvU32 i, j;
uvm_fault_buffer_entry_t **ordered_fault_cache = batch_context->ordered_fault_cache;
UVM_ASSERT(batch_context->num_coalesced_faults > 0);
UVM_ASSERT(batch_context->num_cached_faults >= batch_context->num_coalesced_faults);
// Generate an ordered view of the fault cache in ordered_fault_cache.
// We sort the pointers, not the entries in fault_cache
// Initialize pointers before they are sorted. We only sort one instance per
// coalesced fault
for (i = 0, j = 0; i < batch_context->num_cached_faults; ++i) {
if (!batch_context->fault_cache[i].filtered)
ordered_fault_cache[j++] = &batch_context->fault_cache[i];
}
UVM_ASSERT(j == batch_context->num_coalesced_faults);
// 1) if the fault batch contains more than one, sort by instance_ptr
if (!batch_context->is_single_instance_ptr) {
sort(ordered_fault_cache,
batch_context->num_coalesced_faults,
sizeof(*ordered_fault_cache),
cmp_sort_fault_entry_by_instance_ptr,
NULL);
}
// 2) translate all instance_ptrs to VA spaces
status = translate_instance_ptrs(gpu, batch_context);
if (status != NV_OK)
return status;
// 3) sort by va_space, fault address (GPU already reports 4K-aligned
// address) and access type
sort(ordered_fault_cache,
batch_context->num_coalesced_faults,
sizeof(*ordered_fault_cache),
cmp_sort_fault_entry_by_va_space_address_access_type,
NULL);
return NV_OK;
}
2023-02-28 20:12:44 +01:00
static bool check_fault_entry_duplicate(const uvm_fault_buffer_entry_t *current_entry,
const uvm_fault_buffer_entry_t *previous_entry)
{
bool is_duplicate = false;
if (previous_entry) {
is_duplicate = (current_entry->va_space == previous_entry->va_space) &&
(current_entry->fault_address == previous_entry->fault_address);
}
return is_duplicate;
}
2023-05-30 19:11:36 +02:00
static void update_batch_and_notify_fault(uvm_gpu_t *gpu,
uvm_fault_service_batch_context_t *batch_context,
uvm_va_block_t *va_block,
uvm_processor_id_t preferred_location,
uvm_fault_buffer_entry_t *current_entry,
bool is_duplicate)
2023-02-28 20:12:44 +01:00
{
2023-05-30 19:11:36 +02:00
if (is_duplicate)
batch_context->num_duplicate_faults += current_entry->num_instances;
else
batch_context->num_duplicate_faults += current_entry->num_instances - 1;
2023-02-28 20:12:44 +01:00
2023-05-30 19:11:36 +02:00
uvm_perf_event_notify_gpu_fault(&current_entry->va_space->perf_events,
va_block,
gpu->id,
preferred_location,
current_entry,
batch_context->batch_id,
is_duplicate);
}
2023-02-28 20:12:44 +01:00
2023-05-30 19:11:36 +02:00
static void mark_fault_invalid_prefetch(uvm_fault_service_batch_context_t *batch_context,
uvm_fault_buffer_entry_t *fault_entry)
{
fault_entry->is_invalid_prefetch = true;
// For block faults, the following counter might be updated more than once
// for the same fault if block_context->num_retries > 0. As a result, this
// counter might be higher than the actual count. In order for this counter
// to be always accurate, block_context needs to passed down the stack from
// all callers. But since num_retries > 0 case is uncommon and imprecise
// invalid_prefetch counter doesn't affect functionality (other than
// disabling prefetching if the counter indicates lots of invalid prefetch
// faults), this is ok.
batch_context->num_invalid_prefetch_faults += fault_entry->num_instances;
2023-02-28 20:12:44 +01:00
}
2023-05-30 19:11:36 +02:00
static void mark_fault_throttled(uvm_fault_service_batch_context_t *batch_context,
uvm_fault_buffer_entry_t *fault_entry)
2023-02-28 20:12:44 +01:00
{
2023-05-30 19:11:36 +02:00
fault_entry->is_throttled = true;
batch_context->has_throttled_faults = true;
}
2023-02-28 20:12:44 +01:00
2023-05-30 19:11:36 +02:00
static void mark_fault_fatal(uvm_fault_service_batch_context_t *batch_context,
uvm_fault_buffer_entry_t *fault_entry,
UvmEventFatalReason fatal_reason,
uvm_fault_cancel_va_mode_t cancel_va_mode)
{
uvm_fault_utlb_info_t *utlb = &batch_context->utlbs[fault_entry->fault_source.utlb_id];
2023-02-28 20:12:44 +01:00
2023-05-30 19:11:36 +02:00
fault_entry->is_fatal = true;
fault_entry->fatal_reason = fatal_reason;
fault_entry->replayable.cancel_va_mode = cancel_va_mode;
2023-02-28 20:12:44 +01:00
2023-05-30 19:11:36 +02:00
utlb->has_fatal_faults = true;
batch_context->has_fatal_faults = true;
}
2023-02-28 20:12:44 +01:00
2023-05-30 19:11:36 +02:00
static void fault_entry_duplicate_flags(uvm_fault_service_batch_context_t *batch_context,
uvm_fault_buffer_entry_t *current_entry,
const uvm_fault_buffer_entry_t *previous_entry)
{
UVM_ASSERT(previous_entry);
UVM_ASSERT(check_fault_entry_duplicate(current_entry, previous_entry));
2023-02-28 20:12:44 +01:00
2023-05-30 19:11:36 +02:00
// Propagate the is_invalid_prefetch flag across all prefetch faults
// on the page
if (previous_entry->is_invalid_prefetch)
mark_fault_invalid_prefetch(batch_context, current_entry);
// If a page is throttled, all faults on the page must be skipped
if (previous_entry->is_throttled)
mark_fault_throttled(batch_context, current_entry);
2023-02-28 20:12:44 +01:00
}
2022-05-09 22:18:59 +02:00
// This function computes the maximum access type that can be serviced for the
// reported fault instances given the logical permissions of the VA range. If
// none of the fault instances can be serviced UVM_FAULT_ACCESS_TYPE_COUNT is
// returned instead.
//
// In the case that there are faults that cannot be serviced, this function
// also sets the flags required for fault cancellation. Prefetch faults do not
// need to be cancelled since they disappear on replay.
//
// The UVM driver considers two scenarios for logical permissions violation:
// - All access types are invalid. For example, when faulting from a processor
// that doesn't have access to the preferred location of a range group when it
// is not migratable. In this case all accesses to the page must be cancelled.
// - Write/atomic accesses are invalid. Basically, when trying to modify a
// read-only VA range. In this case we restrict fault cancelling to those types
// of accesses.
//
// Return values:
// - service_access_type: highest access type that can be serviced.
static uvm_fault_access_type_t check_fault_access_permissions(uvm_gpu_t *gpu,
2023-05-30 19:11:36 +02:00
uvm_fault_service_batch_context_t *batch_context,
2022-05-09 22:18:59 +02:00
uvm_va_block_t *va_block,
2023-05-30 19:11:36 +02:00
uvm_service_block_context_t *service_block_context,
2022-05-09 22:18:59 +02:00
uvm_fault_buffer_entry_t *fault_entry,
bool allow_migration)
{
NV_STATUS perm_status;
2023-05-30 19:11:36 +02:00
UvmEventFatalReason fatal_reason;
uvm_fault_cancel_va_mode_t cancel_va_mode;
uvm_fault_access_type_t ret = UVM_FAULT_ACCESS_TYPE_COUNT;
uvm_va_block_context_t *va_block_context = &service_block_context->block_context;
2022-05-09 22:18:59 +02:00
2022-11-10 17:39:33 +01:00
perm_status = uvm_va_block_check_logical_permissions(va_block,
va_block_context,
2022-05-09 22:18:59 +02:00
gpu->id,
2022-11-10 17:39:33 +01:00
uvm_va_block_cpu_page_index(va_block,
fault_entry->fault_address),
2022-05-09 22:18:59 +02:00
fault_entry->fault_access_type,
allow_migration);
if (perm_status == NV_OK)
return fault_entry->fault_access_type;
if (fault_entry->fault_access_type == UVM_FAULT_ACCESS_TYPE_PREFETCH) {
2023-05-30 19:11:36 +02:00
// Only update the count the first time since logical permissions cannot
// change while we hold the VA space lock
// TODO: Bug 1750144: That might not be true with HMM.
if (service_block_context->num_retries == 0)
mark_fault_invalid_prefetch(batch_context, fault_entry);
return ret;
2022-05-09 22:18:59 +02:00
}
// At this point we know that some fault instances cannot be serviced
2023-05-30 19:11:36 +02:00
fatal_reason = uvm_tools_status_to_fatal_fault_reason(perm_status);
2022-05-09 22:18:59 +02:00
if (fault_entry->fault_access_type > UVM_FAULT_ACCESS_TYPE_READ) {
2023-05-30 19:11:36 +02:00
cancel_va_mode = UVM_FAULT_CANCEL_VA_MODE_WRITE_AND_ATOMIC;
2022-05-09 22:18:59 +02:00
// If there are pending read accesses on the same page, we have to
// service them before we can cancel the write/atomic faults. So we
// retry with read fault access type.
if (uvm_fault_access_type_mask_test(fault_entry->access_type_mask, UVM_FAULT_ACCESS_TYPE_READ)) {
2022-11-10 17:39:33 +01:00
perm_status = uvm_va_block_check_logical_permissions(va_block,
va_block_context,
2022-05-09 22:18:59 +02:00
gpu->id,
2022-11-10 17:39:33 +01:00
uvm_va_block_cpu_page_index(va_block,
fault_entry->fault_address),
2022-05-09 22:18:59 +02:00
UVM_FAULT_ACCESS_TYPE_READ,
allow_migration);
2023-05-30 19:11:36 +02:00
if (perm_status == NV_OK) {
ret = UVM_FAULT_ACCESS_TYPE_READ;
}
else {
// Read accesses didn't succeed, cancel all faults
cancel_va_mode = UVM_FAULT_CANCEL_VA_MODE_ALL;
fatal_reason = uvm_tools_status_to_fatal_fault_reason(perm_status);
}
2022-05-09 22:18:59 +02:00
}
}
else {
2023-05-30 19:11:36 +02:00
cancel_va_mode = UVM_FAULT_CANCEL_VA_MODE_ALL;
2022-05-09 22:18:59 +02:00
}
2023-05-30 19:11:36 +02:00
mark_fault_fatal(batch_context, fault_entry, fatal_reason, cancel_va_mode);
return ret;
2022-05-09 22:18:59 +02:00
}
// We notify the fault event for all faults within the block so that the
// performance heuristics are updated. Then, all required actions for the block
// data are performed by the performance heuristics code.
//
// Fatal faults are flagged as fatal for later cancellation. Servicing is not
// interrupted on fatal faults due to insufficient permissions or invalid
// addresses.
//
// Return codes:
// - NV_OK if all faults were handled (both fatal and non-fatal)
// - NV_ERR_MORE_PROCESSING_REQUIRED if servicing needs allocation retry
// - NV_ERR_NO_MEMORY if the faults could not be serviced due to OOM
// - Any other value is a UVM-global error
2023-02-28 20:12:44 +01:00
static NV_STATUS service_fault_batch_block_locked(uvm_gpu_t *gpu,
uvm_va_block_t *va_block,
uvm_va_block_retry_t *va_block_retry,
uvm_fault_service_batch_context_t *batch_context,
NvU32 first_fault_index,
NvU32 *block_faults)
2022-05-09 22:18:59 +02:00
{
NV_STATUS status = NV_OK;
NvU32 i;
uvm_page_index_t first_page_index;
uvm_page_index_t last_page_index;
NvU32 page_fault_count = 0;
uvm_range_group_range_iter_t iter;
uvm_replayable_fault_buffer_info_t *replayable_faults = &gpu->parent->fault_buffer_info.replayable;
uvm_fault_buffer_entry_t **ordered_fault_cache = batch_context->ordered_fault_cache;
uvm_service_block_context_t *block_context = &replayable_faults->block_service_context;
uvm_va_space_t *va_space = uvm_va_block_get_va_space(va_block);
NvU64 end;
// Check that all uvm_fault_access_type_t values can fit into an NvU8
BUILD_BUG_ON(UVM_FAULT_ACCESS_TYPE_COUNT > (int)(NvU8)-1);
uvm_assert_mutex_locked(&va_block->lock);
*block_faults = 0;
first_page_index = PAGES_PER_UVM_VA_BLOCK;
last_page_index = 0;
// Initialize fault service block context
uvm_processor_mask_zero(&block_context->resident_processors);
block_context->thrashing_pin_count = 0;
block_context->read_duplicate_count = 0;
uvm_range_group_range_migratability_iter_first(va_space, va_block->start, va_block->end, &iter);
// The first entry is guaranteed to fall within this block
UVM_ASSERT(ordered_fault_cache[first_fault_index]->va_space == va_space);
UVM_ASSERT(ordered_fault_cache[first_fault_index]->fault_address >= va_block->start);
UVM_ASSERT(ordered_fault_cache[first_fault_index]->fault_address <= va_block->end);
2022-11-10 17:39:33 +01:00
if (uvm_va_block_is_hmm(va_block)) {
2022-05-09 22:18:59 +02:00
uvm_hmm_find_policy_end(va_block,
&block_context->block_context,
ordered_fault_cache[first_fault_index]->fault_address,
&end);
2022-11-10 17:39:33 +01:00
}
else {
2022-05-09 22:18:59 +02:00
block_context->block_context.policy = uvm_va_range_get_policy(va_block->va_range);
2022-11-10 17:39:33 +01:00
end = va_block->end;
}
2022-05-09 22:18:59 +02:00
// Scan the sorted array and notify the fault event for all fault entries
// in the block
for (i = first_fault_index;
i < batch_context->num_coalesced_faults &&
ordered_fault_cache[i]->va_space == va_space &&
ordered_fault_cache[i]->fault_address <= end;
++i) {
uvm_fault_buffer_entry_t *current_entry = ordered_fault_cache[i];
const uvm_fault_buffer_entry_t *previous_entry = NULL;
bool read_duplicate;
uvm_processor_id_t new_residency;
uvm_perf_thrashing_hint_t thrashing_hint;
uvm_page_index_t page_index = uvm_va_block_cpu_page_index(va_block, current_entry->fault_address);
bool is_duplicate = false;
uvm_fault_access_type_t service_access_type;
NvU32 service_access_type_mask;
UVM_ASSERT(current_entry->fault_access_type ==
uvm_fault_access_type_mask_highest(current_entry->access_type_mask));
current_entry->is_fatal = false;
current_entry->is_throttled = false;
current_entry->is_invalid_prefetch = false;
if (i > first_fault_index) {
previous_entry = ordered_fault_cache[i - 1];
2023-02-28 20:12:44 +01:00
is_duplicate = check_fault_entry_duplicate(current_entry, previous_entry);
2022-05-09 22:18:59 +02:00
}
2023-05-30 19:11:36 +02:00
// Only update counters the first time since logical permissions cannot
// change while we hold the VA space lock.
// TODO: Bug 1750144: That might not be true with HMM.
2022-05-09 22:18:59 +02:00
if (block_context->num_retries == 0) {
2023-05-30 19:11:36 +02:00
update_batch_and_notify_fault(gpu,
batch_context,
va_block,
block_context->block_context.policy->preferred_location,
current_entry,
is_duplicate);
2022-05-09 22:18:59 +02:00
}
// Service the most intrusive fault per page, only. Waive the rest
if (is_duplicate) {
2023-05-30 19:11:36 +02:00
fault_entry_duplicate_flags(batch_context, current_entry, previous_entry);
2022-05-09 22:18:59 +02:00
// The previous fault was non-fatal so the page has been already
// serviced
if (!previous_entry->is_fatal)
2023-05-30 19:11:36 +02:00
continue;
2022-05-09 22:18:59 +02:00
}
// Ensure that the migratability iterator covers the current fault
// address
while (iter.end < current_entry->fault_address)
uvm_range_group_range_migratability_iter_next(va_space, &iter, va_block->end);
UVM_ASSERT(iter.start <= current_entry->fault_address && iter.end >= current_entry->fault_address);
2022-11-10 17:39:33 +01:00
service_access_type = check_fault_access_permissions(gpu,
2023-05-30 19:11:36 +02:00
batch_context,
2022-11-10 17:39:33 +01:00
va_block,
2023-05-30 19:11:36 +02:00
block_context,
2022-11-10 17:39:33 +01:00
current_entry,
iter.migratable);
2022-05-09 22:18:59 +02:00
// Do not exit early due to logical errors such as access permission
// violation.
if (service_access_type == UVM_FAULT_ACCESS_TYPE_COUNT)
2023-05-30 19:11:36 +02:00
continue;
2022-05-09 22:18:59 +02:00
if (service_access_type != current_entry->fault_access_type) {
// Some of the fault instances cannot be serviced due to invalid
// access permissions. Recompute the access type service mask to
// service the rest.
UVM_ASSERT(service_access_type < current_entry->fault_access_type);
service_access_type_mask = uvm_fault_access_type_mask_bit(service_access_type);
}
else {
service_access_type_mask = current_entry->access_type_mask;
}
// If the GPU already has the necessary access permission, the fault
// does not need to be serviced
if (uvm_va_block_page_is_gpu_authorized(va_block,
page_index,
gpu->id,
uvm_fault_access_type_to_prot(service_access_type)))
2023-05-30 19:11:36 +02:00
continue;
2022-05-09 22:18:59 +02:00
thrashing_hint = uvm_perf_thrashing_get_hint(va_block, current_entry->fault_address, gpu->id);
if (thrashing_hint.type == UVM_PERF_THRASHING_HINT_TYPE_THROTTLE) {
// Throttling is implemented by sleeping in the fault handler on
// the CPU and by continuing to process faults on other pages on
// the GPU
2023-05-30 19:11:36 +02:00
//
// Only update the flag the first time since logical permissions
// cannot change while we hold the VA space lock.
// TODO: Bug 1750144: That might not be true with HMM.
if (block_context->num_retries == 0)
mark_fault_throttled(batch_context, current_entry);
continue;
2022-05-09 22:18:59 +02:00
}
else if (thrashing_hint.type == UVM_PERF_THRASHING_HINT_TYPE_PIN) {
if (block_context->thrashing_pin_count++ == 0)
uvm_page_mask_zero(&block_context->thrashing_pin_mask);
uvm_page_mask_set(&block_context->thrashing_pin_mask, page_index);
}
// Compute new residency and update the masks
new_residency = uvm_va_block_select_residency(va_block,
2022-11-10 17:39:33 +01:00
&block_context->block_context,
2022-05-09 22:18:59 +02:00
page_index,
gpu->id,
service_access_type_mask,
block_context->block_context.policy,
&thrashing_hint,
UVM_SERVICE_OPERATION_REPLAYABLE_FAULTS,
&read_duplicate);
if (!uvm_processor_mask_test_and_set(&block_context->resident_processors, new_residency))
uvm_page_mask_zero(&block_context->per_processor_masks[uvm_id_value(new_residency)].new_residency);
uvm_page_mask_set(&block_context->per_processor_masks[uvm_id_value(new_residency)].new_residency, page_index);
if (read_duplicate) {
if (block_context->read_duplicate_count++ == 0)
uvm_page_mask_zero(&block_context->read_duplicate_mask);
uvm_page_mask_set(&block_context->read_duplicate_mask, page_index);
}
++page_fault_count;
block_context->access_type[page_index] = service_access_type;
if (page_index < first_page_index)
first_page_index = page_index;
if (page_index > last_page_index)
last_page_index = page_index;
}
// Apply the changes computed in the fault service block context, if there
// are pages to be serviced
if (page_fault_count > 0) {
block_context->region = uvm_va_block_region(first_page_index, last_page_index + 1);
status = uvm_va_block_service_locked(gpu->id, va_block, va_block_retry, block_context);
}
*block_faults = i - first_fault_index;
++block_context->num_retries;
if (status == NV_OK && batch_context->has_fatal_faults)
status = uvm_va_block_set_cancel(va_block, &block_context->block_context, gpu);
return status;
}
// We notify the fault event for all faults within the block so that the
// performance heuristics are updated. The VA block lock is taken for the whole
// fault servicing although it might be temporarily dropped and re-taken if
// memory eviction is required.
//
// See the comments for function service_fault_batch_block_locked for
// implementation details and error codes.
2023-02-28 20:12:44 +01:00
static NV_STATUS service_fault_batch_block(uvm_gpu_t *gpu,
uvm_va_block_t *va_block,
uvm_fault_service_batch_context_t *batch_context,
NvU32 first_fault_index,
NvU32 *block_faults)
2022-05-09 22:18:59 +02:00
{
NV_STATUS status;
uvm_va_block_retry_t va_block_retry;
NV_STATUS tracker_status;
uvm_service_block_context_t *fault_block_context = &gpu->parent->fault_buffer_info.replayable.block_service_context;
fault_block_context->operation = UVM_SERVICE_OPERATION_REPLAYABLE_FAULTS;
fault_block_context->num_retries = 0;
2023-05-30 19:11:36 +02:00
if (uvm_va_block_is_hmm(va_block))
uvm_hmm_migrate_begin_wait(va_block);
2022-05-09 22:18:59 +02:00
uvm_mutex_lock(&va_block->lock);
status = UVM_VA_BLOCK_RETRY_LOCKED(va_block, &va_block_retry,
2023-02-28 20:12:44 +01:00
service_fault_batch_block_locked(gpu,
va_block,
&va_block_retry,
batch_context,
first_fault_index,
block_faults));
2022-05-09 22:18:59 +02:00
tracker_status = uvm_tracker_add_tracker_safe(&batch_context->tracker, &va_block->tracker);
uvm_mutex_unlock(&va_block->lock);
2023-05-30 19:11:36 +02:00
if (uvm_va_block_is_hmm(va_block))
uvm_hmm_migrate_finish(va_block);
2022-05-09 22:18:59 +02:00
return status == NV_OK? tracker_status: status;
}
typedef enum
{
// Use this mode when calling from the normal fault servicing path
FAULT_SERVICE_MODE_REGULAR,
// Use this mode when servicing faults from the fault cancelling algorithm.
// In this mode no replays are issued
FAULT_SERVICE_MODE_CANCEL,
} fault_service_mode_t;
2023-05-30 19:11:36 +02:00
static void service_fault_batch_fatal(uvm_gpu_t *gpu,
uvm_fault_service_batch_context_t *batch_context,
NvU32 first_fault_index,
NV_STATUS status,
uvm_fault_cancel_va_mode_t cancel_va_mode,
NvU32 *block_faults)
2022-05-09 22:18:59 +02:00
{
2023-02-28 20:12:44 +01:00
uvm_fault_buffer_entry_t *current_entry = batch_context->ordered_fault_cache[first_fault_index];
const uvm_fault_buffer_entry_t *previous_entry = first_fault_index > 0 ?
batch_context->ordered_fault_cache[first_fault_index - 1] : NULL;
bool is_duplicate = check_fault_entry_duplicate(current_entry, previous_entry);
2022-05-09 22:18:59 +02:00
2023-02-28 20:12:44 +01:00
if (is_duplicate)
2023-05-30 19:11:36 +02:00
fault_entry_duplicate_flags(batch_context, current_entry, previous_entry);
2022-05-09 22:18:59 +02:00
2023-05-30 19:11:36 +02:00
if (current_entry->fault_access_type == UVM_FAULT_ACCESS_TYPE_PREFETCH)
mark_fault_invalid_prefetch(batch_context, current_entry);
2023-02-28 20:12:44 +01:00
else
2023-05-30 19:11:36 +02:00
mark_fault_fatal(batch_context, current_entry, uvm_tools_status_to_fatal_fault_reason(status), cancel_va_mode);
2022-05-09 22:18:59 +02:00
2023-02-28 20:12:44 +01:00
(*block_faults)++;
}
2023-05-30 19:11:36 +02:00
static void service_fault_batch_fatal_notify(uvm_gpu_t *gpu,
uvm_fault_service_batch_context_t *batch_context,
NvU32 first_fault_index,
NV_STATUS status,
uvm_fault_cancel_va_mode_t cancel_va_mode,
NvU32 *block_faults)
2023-02-28 20:12:44 +01:00
{
uvm_fault_buffer_entry_t *current_entry = batch_context->ordered_fault_cache[first_fault_index];
const uvm_fault_buffer_entry_t *previous_entry = first_fault_index > 0 ?
batch_context->ordered_fault_cache[first_fault_index - 1] : NULL;
bool is_duplicate = check_fault_entry_duplicate(current_entry, previous_entry);
2023-05-30 19:11:36 +02:00
service_fault_batch_fatal(gpu, batch_context, first_fault_index, status, cancel_va_mode, block_faults);
update_batch_and_notify_fault(gpu, batch_context, NULL, UVM_ID_INVALID, current_entry, is_duplicate);
}
static NV_STATUS service_fault_batch_ats_sub_vma(uvm_gpu_va_space_t *gpu_va_space,
struct vm_area_struct *vma,
NvU64 base,
uvm_fault_service_batch_context_t *batch_context,
NvU32 fault_index_start,
NvU32 fault_index_end,
NvU32 *block_faults)
{
NvU32 i;
NV_STATUS status = NV_OK;
uvm_gpu_t *gpu = gpu_va_space->gpu;
uvm_ats_fault_context_t *ats_context = &batch_context->ats_context;
const uvm_page_mask_t *read_fault_mask = &ats_context->read_fault_mask;
const uvm_page_mask_t *write_fault_mask = &ats_context->write_fault_mask;
const uvm_page_mask_t *faults_serviced_mask = &ats_context->faults_serviced_mask;
const uvm_page_mask_t *reads_serviced_mask = &ats_context->reads_serviced_mask;
uvm_page_mask_t *tmp_mask = &ats_context->tmp_mask;
UVM_ASSERT(vma);
ats_context->client_type = UVM_FAULT_CLIENT_TYPE_GPC;
uvm_page_mask_or(tmp_mask, write_fault_mask, read_fault_mask);
2023-02-28 20:12:44 +01:00
2023-05-30 19:11:36 +02:00
status = uvm_ats_service_faults(gpu_va_space, vma, base, &batch_context->ats_context);
UVM_ASSERT(uvm_page_mask_subset(faults_serviced_mask, tmp_mask));
if ((status != NV_OK) || uvm_page_mask_equal(faults_serviced_mask, tmp_mask)) {
(*block_faults) += (fault_index_end - fault_index_start);
return status;
2022-05-09 22:18:59 +02:00
}
2023-05-30 19:11:36 +02:00
// Check faults_serviced_mask and reads_serviced_mask for precise fault
// attribution after calling the ATS servicing routine. The
// errors returned from ATS servicing routine should only be
// global errors such as OOM or ECC. uvm_gpu_service_replayable_faults()
// handles global errors by calling cancel_fault_batch(). Precise
// attribution isn't currently supported in such cases.
//
// Precise fault attribution for global errors can be handled by
// servicing one fault at a time until fault servicing encounters an
// error.
// TODO: Bug 3989244: Precise ATS fault attribution for global errors.
for (i = fault_index_start; i < fault_index_end; i++) {
uvm_page_index_t page_index;
uvm_fault_cancel_va_mode_t cancel_va_mode;
uvm_fault_buffer_entry_t *current_entry = batch_context->ordered_fault_cache[i];
uvm_fault_access_type_t access_type = current_entry->fault_access_type;
page_index = (current_entry->fault_address - base) / PAGE_SIZE;
if (uvm_page_mask_test(faults_serviced_mask, page_index)) {
(*block_faults)++;
continue;
}
if (access_type <= UVM_FAULT_ACCESS_TYPE_READ) {
cancel_va_mode = UVM_FAULT_CANCEL_VA_MODE_ALL;
}
else if (access_type >= UVM_FAULT_ACCESS_TYPE_WRITE) {
if (uvm_fault_access_type_mask_test(current_entry->access_type_mask, UVM_FAULT_ACCESS_TYPE_READ) &&
!uvm_page_mask_test(reads_serviced_mask, page_index))
cancel_va_mode = UVM_FAULT_CANCEL_VA_MODE_ALL;
else
cancel_va_mode = UVM_FAULT_CANCEL_VA_MODE_WRITE_AND_ATOMIC;
}
service_fault_batch_fatal(gpu, batch_context, i, NV_ERR_INVALID_ADDRESS, cancel_va_mode, block_faults);
2022-05-09 22:18:59 +02:00
}
2023-05-30 19:11:36 +02:00
return status;
}
2022-05-09 22:18:59 +02:00
2023-05-30 19:11:36 +02:00
static void start_new_sub_batch(NvU64 *sub_batch_base,
NvU64 address,
NvU32 *sub_batch_fault_index,
NvU32 fault_index,
uvm_ats_fault_context_t *ats_context)
{
uvm_page_mask_zero(&ats_context->read_fault_mask);
uvm_page_mask_zero(&ats_context->write_fault_mask);
2022-05-09 22:18:59 +02:00
2023-05-30 19:11:36 +02:00
*sub_batch_fault_index = fault_index;
*sub_batch_base = UVM_VA_BLOCK_ALIGN_DOWN(address);
}
static NV_STATUS service_fault_batch_ats_sub(uvm_gpu_va_space_t *gpu_va_space,
struct vm_area_struct *vma,
uvm_fault_service_batch_context_t *batch_context,
NvU32 fault_index,
NvU64 outer,
NvU32 *block_faults)
{
NV_STATUS status = NV_OK;
NvU32 i = fault_index;
NvU32 sub_batch_fault_index;
NvU64 sub_batch_base;
uvm_fault_buffer_entry_t *previous_entry = NULL;
uvm_fault_buffer_entry_t *current_entry = batch_context->ordered_fault_cache[i];
uvm_ats_fault_context_t *ats_context = &batch_context->ats_context;
uvm_page_mask_t *read_fault_mask = &ats_context->read_fault_mask;
uvm_page_mask_t *write_fault_mask = &ats_context->write_fault_mask;
uvm_gpu_t *gpu = gpu_va_space->gpu;
bool replay_per_va_block =
(gpu->parent->fault_buffer_info.replayable.replay_policy == UVM_PERF_FAULT_REPLAY_POLICY_BLOCK);
UVM_ASSERT(vma);
outer = min(outer, (NvU64) vma->vm_end);
start_new_sub_batch(&sub_batch_base, current_entry->fault_address, &sub_batch_fault_index, i, ats_context);
do {
uvm_page_index_t page_index;
NvU64 fault_address = current_entry->fault_address;
uvm_fault_access_type_t access_type = current_entry->fault_access_type;
bool is_duplicate = check_fault_entry_duplicate(current_entry, previous_entry);
i++;
update_batch_and_notify_fault(gpu_va_space->gpu,
batch_context,
NULL,
UVM_ID_INVALID,
current_entry,
is_duplicate);
// End of sub-batch. Service faults gathered so far.
if (fault_address >= (sub_batch_base + UVM_VA_BLOCK_SIZE)) {
UVM_ASSERT(!uvm_page_mask_empty(read_fault_mask) || !uvm_page_mask_empty(write_fault_mask));
status = service_fault_batch_ats_sub_vma(gpu_va_space,
vma,
sub_batch_base,
batch_context,
sub_batch_fault_index,
i - 1,
block_faults);
if (status != NV_OK || replay_per_va_block)
break;
start_new_sub_batch(&sub_batch_base, fault_address, &sub_batch_fault_index, i - 1, ats_context);
}
page_index = (fault_address - sub_batch_base) / PAGE_SIZE;
if ((access_type <= UVM_FAULT_ACCESS_TYPE_READ) ||
uvm_fault_access_type_mask_test(current_entry->access_type_mask, UVM_FAULT_ACCESS_TYPE_READ))
uvm_page_mask_set(read_fault_mask, page_index);
if (access_type >= UVM_FAULT_ACCESS_TYPE_WRITE)
uvm_page_mask_set(write_fault_mask, page_index);
previous_entry = current_entry;
current_entry = i < batch_context->num_coalesced_faults ? batch_context->ordered_fault_cache[i] : NULL;
} while (current_entry &&
(current_entry->fault_address < outer) &&
(previous_entry->va_space == current_entry->va_space));
// Service the last sub-batch.
if ((status == NV_OK) && (!uvm_page_mask_empty(read_fault_mask) || !uvm_page_mask_empty(write_fault_mask))) {
status = service_fault_batch_ats_sub_vma(gpu_va_space,
vma,
sub_batch_base,
batch_context,
sub_batch_fault_index,
i,
block_faults);
}
return status;
}
static NV_STATUS service_fault_batch_ats(uvm_gpu_va_space_t *gpu_va_space,
struct mm_struct *mm,
uvm_fault_service_batch_context_t *batch_context,
NvU32 first_fault_index,
NvU64 outer,
NvU32 *block_faults)
{
NvU32 i;
NV_STATUS status = NV_OK;
for (i = first_fault_index; i < batch_context->num_coalesced_faults;) {
uvm_fault_buffer_entry_t *current_entry = batch_context->ordered_fault_cache[i];
const uvm_fault_buffer_entry_t *previous_entry = i > first_fault_index ?
batch_context->ordered_fault_cache[i - 1] : NULL;
NvU64 fault_address = current_entry->fault_address;
struct vm_area_struct *vma;
NvU32 num_faults_before = (*block_faults);
if (previous_entry && (previous_entry->va_space != current_entry->va_space))
break;
if (fault_address >= outer)
break;
vma = find_vma_intersection(mm, fault_address, fault_address + 1);
if (!vma) {
// Since a vma wasn't found, cancel all accesses on the page since
// cancelling write and atomic accesses will not cancel pending read
// faults and this can lead to a deadlock since read faults need to
// be serviced first before cancelling write faults.
service_fault_batch_fatal_notify(gpu_va_space->gpu,
batch_context,
i,
NV_ERR_INVALID_ADDRESS,
UVM_FAULT_CANCEL_VA_MODE_ALL,
block_faults);
// Do not fail due to logical errors.
status = NV_OK;
break;
}
status = service_fault_batch_ats_sub(gpu_va_space, vma, batch_context, i, outer, block_faults);
if (status != NV_OK)
break;
i += ((*block_faults) - num_faults_before);
}
return status;
2023-02-28 20:12:44 +01:00
}
static NV_STATUS service_fault_batch_dispatch(uvm_va_space_t *va_space,
uvm_gpu_va_space_t *gpu_va_space,
uvm_fault_service_batch_context_t *batch_context,
2023-05-30 19:11:36 +02:00
NvU32 fault_index,
NvU32 *block_faults,
bool replay_per_va_block)
2023-02-28 20:12:44 +01:00
{
NV_STATUS status;
2023-05-30 19:11:36 +02:00
uvm_va_range_t *va_range = NULL;
uvm_va_range_t *va_range_next = NULL;
2023-02-28 20:12:44 +01:00
uvm_va_block_t *va_block;
uvm_gpu_t *gpu = gpu_va_space->gpu;
uvm_va_block_context_t *va_block_context =
&gpu->parent->fault_buffer_info.replayable.block_service_context.block_context;
2023-05-30 19:11:36 +02:00
uvm_fault_buffer_entry_t *current_entry = batch_context->ordered_fault_cache[fault_index];
2023-02-28 20:12:44 +01:00
struct mm_struct *mm = va_block_context->mm;
NvU64 fault_address = current_entry->fault_address;
(*block_faults) = 0;
2023-05-30 19:11:36 +02:00
va_range_next = uvm_va_space_iter_first(va_space, fault_address, ~0ULL);
if (va_range_next && (fault_address >= va_range_next->node.start)) {
UVM_ASSERT(fault_address < va_range_next->node.end);
va_range = va_range_next;
va_range_next = uvm_va_space_iter_next(va_range_next, ~0ULL);
}
2023-02-28 20:12:44 +01:00
status = uvm_va_block_find_create_in_range(va_space, va_range, fault_address, va_block_context, &va_block);
if (status == NV_OK) {
2023-05-30 19:11:36 +02:00
status = service_fault_batch_block(gpu, va_block, batch_context, fault_index, block_faults);
2022-05-09 22:18:59 +02:00
}
2023-02-28 20:12:44 +01:00
else if ((status == NV_ERR_INVALID_ADDRESS) && uvm_ats_can_service_faults(gpu_va_space, mm)) {
2023-05-30 19:11:36 +02:00
NvU64 outer = ~0ULL;
UVM_ASSERT(replay_per_va_block ==
(gpu->parent->fault_buffer_info.replayable.replay_policy == UVM_PERF_FAULT_REPLAY_POLICY_BLOCK));
// Limit outer to the minimum of next va_range.start and first
// fault_address' next UVM_GMMU_ATS_GRANULARITY alignment so that it's
// enough to check whether the first fault in this dispatch belongs to a
// GMMU region.
if (va_range_next) {
outer = min(va_range_next->node.start,
UVM_ALIGN_DOWN(fault_address + UVM_GMMU_ATS_GRANULARITY, UVM_GMMU_ATS_GRANULARITY));
}
// ATS lookups are disabled on all addresses within the same
// UVM_GMMU_ATS_GRANULARITY as existing GMMU mappings (see documentation
// in uvm_mmu.h). User mode is supposed to reserve VAs as appropriate to
// prevent any system memory allocations from falling within the NO_ATS
// range of other GMMU mappings, so this shouldn't happen during normal
// operation. However, since this scenario may lead to infinite fault
// loops, we handle it by canceling the fault.
if (uvm_ats_check_in_gmmu_region(va_space, fault_address, va_range_next)) {
service_fault_batch_fatal_notify(gpu,
batch_context,
fault_index,
NV_ERR_INVALID_ADDRESS,
UVM_FAULT_CANCEL_VA_MODE_ALL,
block_faults);
// Do not fail due to logical errors
status = NV_OK;
}
else {
status = service_fault_batch_ats(gpu_va_space, mm, batch_context, fault_index, outer, block_faults);
}
2023-02-28 20:12:44 +01:00
}
else {
2023-05-30 19:11:36 +02:00
service_fault_batch_fatal_notify(gpu,
batch_context,
fault_index,
status,
UVM_FAULT_CANCEL_VA_MODE_ALL,
block_faults);
2022-05-09 22:18:59 +02:00
2023-02-28 20:12:44 +01:00
// Do not fail due to logical errors
status = NV_OK;
}
2022-05-09 22:18:59 +02:00
return status;
}
2023-02-28 20:12:44 +01:00
// Scan the ordered view of faults and group them by different va_blocks
// (managed faults) and service faults for each va_block, in batch.
// Service non-managed faults one at a time as they are encountered during the
// scan.
2022-05-09 22:18:59 +02:00
//
// This function returns NV_WARN_MORE_PROCESSING_REQUIRED if the fault buffer
// was flushed because the needs_fault_buffer_flush flag was set on some GPU VA
// space
static NV_STATUS service_fault_batch(uvm_gpu_t *gpu,
fault_service_mode_t service_mode,
uvm_fault_service_batch_context_t *batch_context)
{
NV_STATUS status = NV_OK;
NvU32 i;
uvm_va_space_t *va_space = NULL;
uvm_gpu_va_space_t *gpu_va_space = NULL;
uvm_ats_fault_invalidate_t *ats_invalidate = &gpu->parent->fault_buffer_info.replayable.ats_invalidate;
2023-02-28 20:12:44 +01:00
struct mm_struct *mm = NULL;
2022-05-09 22:18:59 +02:00
const bool replay_per_va_block = service_mode != FAULT_SERVICE_MODE_CANCEL &&
gpu->parent->fault_buffer_info.replayable.replay_policy == UVM_PERF_FAULT_REPLAY_POLICY_BLOCK;
2023-05-30 19:11:36 +02:00
uvm_service_block_context_t *service_context =
&gpu->parent->fault_buffer_info.replayable.block_service_context;
uvm_va_block_context_t *va_block_context = &service_context->block_context;
2022-05-09 22:18:59 +02:00
UVM_ASSERT(gpu->parent->replayable_faults_supported);
ats_invalidate->write_faults_in_batch = false;
2023-05-30 19:11:36 +02:00
uvm_hmm_service_context_init(service_context);
2022-05-09 22:18:59 +02:00
for (i = 0; i < batch_context->num_coalesced_faults;) {
NvU32 block_faults;
uvm_fault_buffer_entry_t *current_entry = batch_context->ordered_fault_cache[i];
uvm_fault_utlb_info_t *utlb = &batch_context->utlbs[current_entry->fault_source.utlb_id];
UVM_ASSERT(current_entry->va_space);
if (current_entry->va_space != va_space) {
// Fault on a different va_space, drop the lock of the old one...
if (va_space != NULL) {
// TLB entries are invalidated per GPU VA space
status = uvm_ats_invalidate_tlbs(gpu_va_space, ats_invalidate, &batch_context->tracker);
if (status != NV_OK)
goto fail;
uvm_va_space_up_read(va_space);
uvm_va_space_mm_release_unlock(va_space, mm);
mm = NULL;
}
va_space = current_entry->va_space;
// ... and take the lock of the new one
// If an mm is registered with the VA space, we have to retain it
// in order to lock it before locking the VA space. It is guaranteed
// to remain valid until we release. If no mm is registered, we
// can only service managed faults, not ATS/HMM faults.
mm = uvm_va_space_mm_retain_lock(va_space);
2022-11-10 17:39:33 +01:00
va_block_context->mm = mm;
2022-05-09 22:18:59 +02:00
uvm_va_space_down_read(va_space);
gpu_va_space = uvm_gpu_va_space_get_by_parent_gpu(va_space, gpu->parent);
2023-02-28 20:12:44 +01:00
if (uvm_processor_mask_test_and_clear_atomic(&va_space->needs_fault_buffer_flush, gpu->id)) {
2022-05-09 22:18:59 +02:00
status = fault_buffer_flush_locked(gpu,
2023-05-30 19:11:36 +02:00
UVM_GPU_BUFFER_FLUSH_MODE_WAIT_UPDATE_PUT,
2022-05-09 22:18:59 +02:00
UVM_FAULT_REPLAY_TYPE_START,
batch_context);
if (status == NV_OK)
status = NV_WARN_MORE_PROCESSING_REQUIRED;
break;
}
// The case where there is no valid GPU VA space for the GPU in this
// VA space is handled next
}
// Some faults could be already fatal if they cannot be handled by
// the UVM driver
if (current_entry->is_fatal) {
++i;
batch_context->has_fatal_faults = true;
utlb->has_fatal_faults = true;
UVM_ASSERT(utlb->num_pending_faults > 0);
continue;
}
if (!uvm_processor_mask_test(&va_space->registered_gpu_va_spaces, gpu->parent->id)) {
// If there is no GPU VA space for the GPU, ignore the fault. This
// can happen if a GPU VA space is destroyed without explicitly
// freeing all memory ranges (destroying the VA range triggers a
// flush of the fault buffer) and there are stale entries in the
// buffer that got fixed by the servicing in a previous batch.
++i;
continue;
}
2023-05-30 19:11:36 +02:00
status = service_fault_batch_dispatch(va_space,
gpu_va_space,
batch_context,
i,
&block_faults,
replay_per_va_block);
2023-02-28 20:12:44 +01:00
// TODO: Bug 3900733: clean up locking in service_fault_batch().
if (status == NV_WARN_MORE_PROCESSING_REQUIRED) {
uvm_va_space_up_read(va_space);
uvm_va_space_mm_release_unlock(va_space, mm);
mm = NULL;
va_space = NULL;
2022-05-09 22:18:59 +02:00
continue;
}
2023-02-28 20:12:44 +01:00
if (status != NV_OK)
goto fail;
i += block_faults;
2022-05-09 22:18:59 +02:00
// Don't issue replays in cancel mode
2023-02-28 20:12:44 +01:00
if (replay_per_va_block && !batch_context->has_fatal_faults) {
2022-05-09 22:18:59 +02:00
status = push_replay_on_gpu(gpu, UVM_FAULT_REPLAY_TYPE_START, batch_context);
if (status != NV_OK)
goto fail;
// Increment the batch id if UVM_PERF_FAULT_REPLAY_POLICY_BLOCK
// is used, as we issue a replay after servicing each VA block
// and we can service a number of VA blocks before returning.
++batch_context->batch_id;
}
}
// Only clobber status if invalidate_status != NV_OK, since status may also
// contain NV_WARN_MORE_PROCESSING_REQUIRED.
if (va_space != NULL) {
NV_STATUS invalidate_status = uvm_ats_invalidate_tlbs(gpu_va_space, ats_invalidate, &batch_context->tracker);
if (invalidate_status != NV_OK)
status = invalidate_status;
}
fail:
if (va_space != NULL) {
uvm_va_space_up_read(va_space);
uvm_va_space_mm_release_unlock(va_space, mm);
}
return status;
}
// Tells if the given fault entry is the first one in its uTLB
static bool is_first_fault_in_utlb(uvm_fault_service_batch_context_t *batch_context, NvU32 fault_index)
{
NvU32 i;
NvU32 utlb_id = batch_context->fault_cache[fault_index].fault_source.utlb_id;
for (i = 0; i < fault_index; ++i) {
uvm_fault_buffer_entry_t *current_entry = &batch_context->fault_cache[i];
// We have found a prior fault in the same uTLB
if (current_entry->fault_source.utlb_id == utlb_id)
return false;
}
return true;
}
// Compute the number of fatal and non-fatal faults for a page in the given uTLB
static void faults_for_page_in_utlb(uvm_fault_service_batch_context_t *batch_context,
uvm_va_space_t *va_space,
NvU64 addr,
NvU32 utlb_id,
NvU32 *fatal_faults,
NvU32 *non_fatal_faults)
{
NvU32 i;
*fatal_faults = 0;
*non_fatal_faults = 0;
// Fault filtering is not allowed in the TLB-based fault cancel path
UVM_ASSERT(batch_context->num_cached_faults == batch_context->num_coalesced_faults);
for (i = 0; i < batch_context->num_cached_faults; ++i) {
uvm_fault_buffer_entry_t *current_entry = &batch_context->fault_cache[i];
if (current_entry->fault_source.utlb_id == utlb_id &&
current_entry->va_space == va_space && current_entry->fault_address == addr) {
// We have found the page
if (current_entry->is_fatal)
++(*fatal_faults);
else
++(*non_fatal_faults);
}
}
}
// Function that tells if there are addresses (reminder: they are aligned to 4K)
// with non-fatal faults only
static bool no_fatal_pages_in_utlb(uvm_fault_service_batch_context_t *batch_context,
NvU32 start_index,
NvU32 utlb_id)
{
NvU32 i;
// Fault filtering is not allowed in the TLB-based fault cancel path
UVM_ASSERT(batch_context->num_cached_faults == batch_context->num_coalesced_faults);
for (i = start_index; i < batch_context->num_cached_faults; ++i) {
uvm_fault_buffer_entry_t *current_entry = &batch_context->fault_cache[i];
if (current_entry->fault_source.utlb_id == utlb_id) {
// We have found a fault for the uTLB
NvU32 fatal_faults;
NvU32 non_fatal_faults;
faults_for_page_in_utlb(batch_context,
current_entry->va_space,
current_entry->fault_address,
utlb_id,
&fatal_faults,
&non_fatal_faults);
if (non_fatal_faults > 0 && fatal_faults == 0)
return true;
}
}
return false;
}
static void record_fatal_fault_helper(uvm_gpu_t *gpu, uvm_fault_buffer_entry_t *entry, UvmEventFatalReason reason)
{
uvm_va_space_t *va_space;
va_space = entry->va_space;
UVM_ASSERT(va_space);
uvm_va_space_down_read(va_space);
// Record fatal fault event
uvm_tools_record_gpu_fatal_fault(gpu->parent->id, va_space, entry, reason);
uvm_va_space_up_read(va_space);
}
// This function tries to find and issue a cancel for each uTLB that meets
// the requirements to guarantee precise fault attribution:
// - No new faults can arrive on the uTLB (uTLB is in lockdown)
// - The first fault in the buffer for a specific uTLB is fatal
// - There are no other addresses in the uTLB with non-fatal faults only
//
// This function and the related helpers iterate over faults as read from HW,
// not through the ordered fault view
//
// TODO: Bug 1766754
// This is very costly, although not critical for performance since we are
// cancelling.
// - Build a list with all the faults within a uTLB
// - Sort by uTLB id
static NV_STATUS try_to_cancel_utlbs(uvm_gpu_t *gpu, uvm_fault_service_batch_context_t *batch_context)
{
NvU32 i;
// Fault filtering is not allowed in the TLB-based fault cancel path
UVM_ASSERT(batch_context->num_cached_faults == batch_context->num_coalesced_faults);
for (i = 0; i < batch_context->num_cached_faults; ++i) {
uvm_fault_buffer_entry_t *current_entry = &batch_context->fault_cache[i];
uvm_fault_utlb_info_t *utlb = &batch_context->utlbs[current_entry->fault_source.utlb_id];
NvU32 gpc_id = current_entry->fault_source.gpc_id;
NvU32 utlb_id = current_entry->fault_source.utlb_id;
NvU32 client_id = current_entry->fault_source.client_id;
// Only fatal faults are considered
if (!current_entry->is_fatal)
continue;
// Only consider uTLBs in lock-down
if (!utlb->in_lockdown)
continue;
// Issue a single cancel per uTLB
if (utlb->cancelled)
continue;
if (is_first_fault_in_utlb(batch_context, i) &&
!no_fatal_pages_in_utlb(batch_context, i + 1, utlb_id)) {
NV_STATUS status;
record_fatal_fault_helper(gpu, current_entry, current_entry->fatal_reason);
status = push_cancel_on_gpu_targeted(gpu,
current_entry->instance_ptr,
gpc_id,
client_id,
&batch_context->tracker);
if (status != NV_OK)
return status;
utlb->cancelled = true;
}
}
return NV_OK;
}
static NvU32 find_fatal_fault_in_utlb(uvm_fault_service_batch_context_t *batch_context,
NvU32 utlb_id)
{
NvU32 i;
// Fault filtering is not allowed in the TLB-based fault cancel path
UVM_ASSERT(batch_context->num_cached_faults == batch_context->num_coalesced_faults);
for (i = 0; i < batch_context->num_cached_faults; ++i) {
if (batch_context->fault_cache[i].is_fatal &&
batch_context->fault_cache[i].fault_source.utlb_id == utlb_id)
return i;
}
return i;
}
static NvU32 is_fatal_fault_in_buffer(uvm_fault_service_batch_context_t *batch_context,
uvm_fault_buffer_entry_t *fault)
{
NvU32 i;
// Fault filtering is not allowed in the TLB-based fault cancel path
UVM_ASSERT(batch_context->num_cached_faults == batch_context->num_coalesced_faults);
for (i = 0; i < batch_context->num_cached_faults; ++i) {
uvm_fault_buffer_entry_t *current_entry = &batch_context->fault_cache[i];
if (cmp_fault_instance_ptr(current_entry, fault) == 0 &&
current_entry->fault_address == fault->fault_address &&
current_entry->fault_access_type == fault->fault_access_type &&
current_entry->fault_source.utlb_id == fault->fault_source.utlb_id) {
return true;
}
}
return false;
}
typedef enum
{
// Only cancel faults flagged as fatal
FAULT_CANCEL_MODE_FATAL,
// Cancel all faults in the batch unconditionally
FAULT_CANCEL_MODE_ALL,
} fault_cancel_mode_t;
// Cancel faults in the given fault service batch context. The function provides
// two different modes depending on the value of cancel_mode:
// - If cancel_mode == FAULT_CANCEL_MODE_FATAL, only faults flagged as fatal
// will be cancelled. In this case, the reason reported to tools is the one
// contained in the fault entry itself.
// - If cancel_mode == FAULT_CANCEL_MODE_ALL, all faults will be cancelled
// unconditionally. In this case, the reason reported to tools for non-fatal
// faults is the one passed to this function.
static NV_STATUS cancel_faults_precise_va(uvm_gpu_t *gpu,
uvm_fault_service_batch_context_t *batch_context,
fault_cancel_mode_t cancel_mode,
UvmEventFatalReason reason)
{
NV_STATUS status = NV_OK;
NV_STATUS fault_status;
uvm_va_space_t *va_space = NULL;
NvU32 i;
UVM_ASSERT(gpu->parent->fault_cancel_va_supported);
if (cancel_mode == FAULT_CANCEL_MODE_ALL)
UVM_ASSERT(reason != UvmEventFatalReasonInvalid);
for (i = 0; i < batch_context->num_coalesced_faults; ++i) {
uvm_fault_buffer_entry_t *current_entry = batch_context->ordered_fault_cache[i];
UVM_ASSERT(current_entry->va_space);
if (current_entry->va_space != va_space) {
// Fault on a different va_space, drop the lock of the old one...
if (va_space != NULL)
uvm_va_space_up_read(va_space);
va_space = current_entry->va_space;
// ... and take the lock of the new one
uvm_va_space_down_read(va_space);
// We don't need to check whether a buffer flush is required
// (due to VA range destruction).
// - For cancel_mode == FAULT_CANCEL_MODE_FATAL, once a fault is
// flagged as fatal we need to cancel it, even if its VA range no
// longer exists.
// - For cancel_mode == FAULT_CANCEL_MODE_ALL we don't care about
// any of this, we just want to trigger RC in RM.
}
if (!uvm_processor_mask_test(&va_space->registered_gpu_va_spaces, gpu->parent->id)) {
// If there is no GPU VA space for the GPU, ignore the fault.
// This can happen if the GPU VA did not exist in
// service_fault_batch(), or it was destroyed since then.
// This is to avoid targetting a PDB that might have been reused
// by another process.
continue;
}
// Cancel the fault
if (cancel_mode == FAULT_CANCEL_MODE_ALL || current_entry->is_fatal) {
uvm_fault_cancel_va_mode_t cancel_va_mode = current_entry->replayable.cancel_va_mode;
// If cancelling unconditionally and the fault was not fatal,
// set the cancel reason passed to this function
if (!current_entry->is_fatal) {
current_entry->fatal_reason = reason;
cancel_va_mode = UVM_FAULT_CANCEL_VA_MODE_ALL;
}
status = cancel_fault_precise_va(gpu, current_entry, cancel_va_mode);
if (status != NV_OK)
break;
}
}
if (va_space != NULL)
uvm_va_space_up_read(va_space);
// After cancelling the fatal faults, the fault buffer is flushed to remove
// any potential duplicated fault that may have been added while processing
// the faults in this batch. This flush also avoids doing unnecessary
// processing after the fatal faults have been cancelled, so all the rest
// are unlikely to remain after a replay because the context is probably in
// the process of dying.
fault_status = fault_buffer_flush_locked(gpu,
UVM_GPU_BUFFER_FLUSH_MODE_UPDATE_PUT,
UVM_FAULT_REPLAY_TYPE_START,
batch_context);
// We report the first encountered error.
if (status == NV_OK)
status = fault_status;
return status;
}
// Function called when the system has found a global error and needs to
// trigger RC in RM.
static void cancel_fault_batch_tlb(uvm_gpu_t *gpu,
uvm_fault_service_batch_context_t *batch_context,
UvmEventFatalReason reason)
{
NvU32 i;
2022-06-28 17:00:06 +02:00
for (i = 0; i < batch_context->num_coalesced_faults; ++i) {
NV_STATUS status = NV_OK;
2022-05-09 22:18:59 +02:00
uvm_fault_buffer_entry_t *current_entry;
2022-06-28 17:00:06 +02:00
uvm_fault_buffer_entry_t *coalesced_entry;
2022-05-09 22:18:59 +02:00
2022-06-28 17:00:06 +02:00
current_entry = batch_context->ordered_fault_cache[i];
2022-05-09 22:18:59 +02:00
2022-06-28 17:00:06 +02:00
// The list iteration below skips the entry used as 'head'.
// Report the 'head' entry explicitly.
uvm_va_space_down_read(current_entry->va_space);
uvm_tools_record_gpu_fatal_fault(gpu->parent->id, current_entry->va_space, current_entry, reason);
2022-05-09 22:18:59 +02:00
2022-06-28 17:00:06 +02:00
list_for_each_entry(coalesced_entry, &current_entry->merged_instances_list, merged_instances_list)
uvm_tools_record_gpu_fatal_fault(gpu->parent->id, current_entry->va_space, coalesced_entry, reason);
uvm_va_space_up_read(current_entry->va_space);
// We need to cancel each instance pointer to correctly handle faults from multiple contexts.
status = push_cancel_on_gpu_global(gpu, current_entry->instance_ptr, &batch_context->tracker);
2022-05-09 22:18:59 +02:00
if (status != NV_OK)
break;
}
}
static void cancel_fault_batch(uvm_gpu_t *gpu,
uvm_fault_service_batch_context_t *batch_context,
UvmEventFatalReason reason)
{
if (gpu->parent->fault_cancel_va_supported) {
cancel_faults_precise_va(gpu, batch_context, FAULT_CANCEL_MODE_ALL, reason);
return;
}
cancel_fault_batch_tlb(gpu, batch_context, reason);
}
// Current fault cancel algorithm
//
// 1- Disable prefetching to avoid new requests keep coming and flooding the
// buffer.
// LOOP
// 2- Record one fatal fault per uTLB to check if it shows up after the replay
// 3- Flush fault buffer (REPLAY_TYPE_START_ACK_ALL to prevent new faults from
// coming to TLBs with pending faults)
// 4- Wait for replay to finish
// 5- Fetch all faults from buffer
// 6- Check what uTLBs are in lockdown mode and can be cancelled
// 7- Preprocess faults (order per va_space, fault address, access type)
// 8- Service all non-fatal faults and mark all non-serviceable faults as fatal
// 6.1- If fatal faults are not found, we are done
// 9- Search for a uTLB which can be targeted for cancel, as described in
// try_to_cancel_utlbs. If found, cancel it.
// END LOOP
// 10- Re-enable prefetching
//
// NOTE: prefetch faults MUST NOT trigger fault cancel. We make sure that no
// prefetch faults are left in the buffer by disabling prefetching and
// flushing the fault buffer afterwards (prefetch faults are not replayed and,
// therefore, will not show up again)
static NV_STATUS cancel_faults_precise_tlb(uvm_gpu_t *gpu, uvm_fault_service_batch_context_t *batch_context)
{
NV_STATUS status;
NV_STATUS tracker_status;
uvm_replayable_fault_buffer_info_t *replayable_faults = &gpu->parent->fault_buffer_info.replayable;
bool first = true;
UVM_ASSERT(gpu->parent->replayable_faults_supported);
// 1) Disable prefetching to avoid new requests keep coming and flooding
// the buffer
if (gpu->parent->fault_buffer_info.prefetch_faults_enabled)
gpu->parent->arch_hal->disable_prefetch_faults(gpu->parent);
while (1) {
NvU32 utlb_id;
// 2) Record one fatal fault per uTLB to check if it shows up after
// the replay. This is used to handle the case in which the uTLB is
// being cancelled from behind our backs by RM. See the comment in
// step 6.
for (utlb_id = 0; utlb_id <= batch_context->max_utlb_id; ++utlb_id) {
uvm_fault_utlb_info_t *utlb = &batch_context->utlbs[utlb_id];
if (!first && utlb->has_fatal_faults) {
NvU32 idx = find_fatal_fault_in_utlb(batch_context, utlb_id);
UVM_ASSERT(idx < batch_context->num_cached_faults);
utlb->prev_fatal_fault = batch_context->fault_cache[idx];
}
else {
utlb->prev_fatal_fault.fault_address = (NvU64)-1;
}
}
first = false;
// 3) Flush fault buffer. After this call, all faults from any of the
// faulting uTLBs are before PUT. New faults from other uTLBs can keep
// arriving. Therefore, in each iteration we just try to cancel faults
// from uTLBs that contained fatal faults in the previous iterations
// and will cause the TLB to stop generating new page faults after the
2023-05-30 19:11:36 +02:00
// following replay with type UVM_FAULT_REPLAY_TYPE_START_ACK_ALL.
//
// No need to use UVM_GPU_BUFFER_FLUSH_MODE_WAIT_UPDATE_PUT since we
// don't care too much about old faults, just new faults from uTLBs
// which faulted before the replay.
2022-05-09 22:18:59 +02:00
status = fault_buffer_flush_locked(gpu,
UVM_GPU_BUFFER_FLUSH_MODE_UPDATE_PUT,
UVM_FAULT_REPLAY_TYPE_START_ACK_ALL,
batch_context);
if (status != NV_OK)
break;
// 4) Wait for replay to finish
status = uvm_tracker_wait(&replayable_faults->replay_tracker);
if (status != NV_OK)
break;
batch_context->num_invalid_prefetch_faults = 0;
batch_context->num_replays = 0;
batch_context->has_fatal_faults = false;
batch_context->has_throttled_faults = false;
// 5) Fetch all faults from buffer
fetch_fault_buffer_entries(gpu, batch_context, FAULT_FETCH_MODE_ALL);
++batch_context->batch_id;
UVM_ASSERT(batch_context->num_cached_faults == batch_context->num_coalesced_faults);
// No more faults left, we are done
if (batch_context->num_cached_faults == 0)
break;
// 6) Check what uTLBs are in lockdown mode and can be cancelled
for (utlb_id = 0; utlb_id <= batch_context->max_utlb_id; ++utlb_id) {
uvm_fault_utlb_info_t *utlb = &batch_context->utlbs[utlb_id];
utlb->in_lockdown = false;
utlb->cancelled = false;
if (utlb->prev_fatal_fault.fault_address != (NvU64)-1) {
// If a previously-reported fault shows up again we can "safely"
// assume that the uTLB that contains it is in lockdown mode
// and no new translations will show up before cancel.
// A fatal fault could only be removed behind our backs by RM
// issuing a cancel, which only happens when RM is resetting the
// engine. That means the instance pointer can't generate any
// new faults, so we won't have an ABA problem where a new
// fault arrives with the same state.
if (is_fatal_fault_in_buffer(batch_context, &utlb->prev_fatal_fault))
utlb->in_lockdown = true;
}
}
// 7) Preprocess faults
status = preprocess_fault_batch(gpu, batch_context);
if (status == NV_WARN_MORE_PROCESSING_REQUIRED)
continue;
else if (status != NV_OK)
break;
// 8) Service all non-fatal faults and mark all non-serviceable faults
// as fatal
status = service_fault_batch(gpu, FAULT_SERVICE_MODE_CANCEL, batch_context);
if (status == NV_WARN_MORE_PROCESSING_REQUIRED)
continue;
UVM_ASSERT(batch_context->num_replays == 0);
if (status == NV_ERR_NO_MEMORY)
continue;
else if (status != NV_OK)
break;
// No more fatal faults left, we are done
if (!batch_context->has_fatal_faults)
break;
// 9) Search for uTLBs that contain fatal faults and meet the
// requirements to be cancelled
try_to_cancel_utlbs(gpu, batch_context);
}
// 10) Re-enable prefetching
if (gpu->parent->fault_buffer_info.prefetch_faults_enabled)
gpu->parent->arch_hal->enable_prefetch_faults(gpu->parent);
if (status == NV_OK)
status = push_replay_on_gpu(gpu, UVM_FAULT_REPLAY_TYPE_START, batch_context);
tracker_status = uvm_tracker_wait(&batch_context->tracker);
return status == NV_OK? tracker_status: status;
}
static NV_STATUS cancel_faults_precise(uvm_gpu_t *gpu, uvm_fault_service_batch_context_t *batch_context)
{
UVM_ASSERT(batch_context->has_fatal_faults);
if (gpu->parent->fault_cancel_va_supported) {
return cancel_faults_precise_va(gpu,
batch_context,
FAULT_CANCEL_MODE_FATAL,
UvmEventFatalReasonInvalid);
}
return cancel_faults_precise_tlb(gpu, batch_context);
}
static void enable_disable_prefetch_faults(uvm_parent_gpu_t *parent_gpu, uvm_fault_service_batch_context_t *batch_context)
{
if (!parent_gpu->prefetch_fault_supported)
return;
// If more than 66% of faults are invalid prefetch accesses, disable
// prefetch faults for a while.
2023-05-30 19:11:36 +02:00
// num_invalid_prefetch_faults may be higher than the actual count. See the
// comment in mark_fault_invalid_prefetch(..).
2022-05-09 22:18:59 +02:00
// Some tests rely on this logic (and ratio) to correctly disable prefetch
// fault reporting. If the logic changes, the tests will have to be changed.
if (parent_gpu->fault_buffer_info.prefetch_faults_enabled &&
uvm_perf_reenable_prefetch_faults_lapse_msec > 0 &&
((batch_context->num_invalid_prefetch_faults * 3 > parent_gpu->fault_buffer_info.max_batch_size * 2) ||
(uvm_enable_builtin_tests &&
parent_gpu->rm_info.isSimulated &&
batch_context->num_invalid_prefetch_faults > 5))) {
uvm_gpu_disable_prefetch_faults(parent_gpu);
}
else if (!parent_gpu->fault_buffer_info.prefetch_faults_enabled) {
NvU64 lapse = NV_GETTIME() - parent_gpu->fault_buffer_info.disable_prefetch_faults_timestamp;
// Reenable prefetch faults after some time
if (lapse > ((NvU64)uvm_perf_reenable_prefetch_faults_lapse_msec * (1000 * 1000)))
uvm_gpu_enable_prefetch_faults(parent_gpu);
}
}
void uvm_gpu_service_replayable_faults(uvm_gpu_t *gpu)
{
NvU32 num_replays = 0;
NvU32 num_batches = 0;
NvU32 num_throttled = 0;
NV_STATUS status = NV_OK;
uvm_replayable_fault_buffer_info_t *replayable_faults = &gpu->parent->fault_buffer_info.replayable;
uvm_fault_service_batch_context_t *batch_context = &replayable_faults->batch_service_context;
UVM_ASSERT(gpu->parent->replayable_faults_supported);
uvm_tracker_init(&batch_context->tracker);
// Process all faults in the buffer
while (1) {
if (num_throttled >= uvm_perf_fault_max_throttle_per_service ||
num_batches >= uvm_perf_fault_max_batches_per_service) {
break;
}
batch_context->num_invalid_prefetch_faults = 0;
batch_context->num_duplicate_faults = 0;
batch_context->num_replays = 0;
batch_context->has_fatal_faults = false;
batch_context->has_throttled_faults = false;
fetch_fault_buffer_entries(gpu, batch_context, FAULT_FETCH_MODE_BATCH_READY);
if (batch_context->num_cached_faults == 0)
break;
++batch_context->batch_id;
status = preprocess_fault_batch(gpu, batch_context);
num_replays += batch_context->num_replays;
if (status == NV_WARN_MORE_PROCESSING_REQUIRED)
continue;
else if (status != NV_OK)
break;
status = service_fault_batch(gpu, FAULT_SERVICE_MODE_REGULAR, batch_context);
// We may have issued replays even if status != NV_OK if
// UVM_PERF_FAULT_REPLAY_POLICY_BLOCK is being used or the fault buffer
// was flushed
num_replays += batch_context->num_replays;
if (status == NV_WARN_MORE_PROCESSING_REQUIRED)
continue;
enable_disable_prefetch_faults(gpu->parent, batch_context);
if (status != NV_OK) {
// Unconditionally cancel all faults to trigger RC. This will not
// provide precise attribution, but this case handles global
// errors such as OOM or ECC where it's not reasonable to
// guarantee precise attribution. We ignore the return value of
// the cancel operation since this path is already returning an
// error code.
cancel_fault_batch(gpu, batch_context, uvm_tools_status_to_fatal_fault_reason(status));
break;
}
if (batch_context->has_fatal_faults) {
status = uvm_tracker_wait(&batch_context->tracker);
if (status == NV_OK)
status = cancel_faults_precise(gpu, batch_context);
break;
}
if (replayable_faults->replay_policy == UVM_PERF_FAULT_REPLAY_POLICY_BATCH) {
status = push_replay_on_gpu(gpu, UVM_FAULT_REPLAY_TYPE_START, batch_context);
if (status != NV_OK)
break;
++num_replays;
}
else if (replayable_faults->replay_policy == UVM_PERF_FAULT_REPLAY_POLICY_BATCH_FLUSH) {
uvm_gpu_buffer_flush_mode_t flush_mode = UVM_GPU_BUFFER_FLUSH_MODE_CACHED_PUT;
if (batch_context->num_duplicate_faults * 100 >
batch_context->num_cached_faults * replayable_faults->replay_update_put_ratio) {
flush_mode = UVM_GPU_BUFFER_FLUSH_MODE_UPDATE_PUT;
}
status = fault_buffer_flush_locked(gpu, flush_mode, UVM_FAULT_REPLAY_TYPE_START, batch_context);
if (status != NV_OK)
break;
++num_replays;
status = uvm_tracker_wait(&replayable_faults->replay_tracker);
if (status != NV_OK)
break;
}
if (batch_context->has_throttled_faults)
++num_throttled;
++num_batches;
}
if (status == NV_WARN_MORE_PROCESSING_REQUIRED)
status = NV_OK;
// Make sure that we issue at least one replay if no replay has been
// issued yet to avoid dropping faults that do not show up in the buffer
if ((status == NV_OK && replayable_faults->replay_policy == UVM_PERF_FAULT_REPLAY_POLICY_ONCE) ||
num_replays == 0)
status = push_replay_on_gpu(gpu, UVM_FAULT_REPLAY_TYPE_START, batch_context);
uvm_tracker_deinit(&batch_context->tracker);
if (status != NV_OK)
UVM_DBG_PRINT("Error servicing replayable faults on GPU: %s\n", uvm_gpu_name(gpu));
}
void uvm_gpu_enable_prefetch_faults(uvm_parent_gpu_t *parent_gpu)
{
UVM_ASSERT(parent_gpu->isr.replayable_faults.handling);
UVM_ASSERT(parent_gpu->prefetch_fault_supported);
if (!parent_gpu->fault_buffer_info.prefetch_faults_enabled) {
parent_gpu->arch_hal->enable_prefetch_faults(parent_gpu);
parent_gpu->fault_buffer_info.prefetch_faults_enabled = true;
}
}
void uvm_gpu_disable_prefetch_faults(uvm_parent_gpu_t *parent_gpu)
{
UVM_ASSERT(parent_gpu->isr.replayable_faults.handling);
UVM_ASSERT(parent_gpu->prefetch_fault_supported);
if (parent_gpu->fault_buffer_info.prefetch_faults_enabled) {
parent_gpu->arch_hal->disable_prefetch_faults(parent_gpu);
parent_gpu->fault_buffer_info.prefetch_faults_enabled = false;
parent_gpu->fault_buffer_info.disable_prefetch_faults_timestamp = NV_GETTIME();
}
}
const char *uvm_perf_fault_replay_policy_string(uvm_perf_fault_replay_policy_t replay_policy)
{
BUILD_BUG_ON(UVM_PERF_FAULT_REPLAY_POLICY_MAX != 4);
switch (replay_policy) {
UVM_ENUM_STRING_CASE(UVM_PERF_FAULT_REPLAY_POLICY_BLOCK);
UVM_ENUM_STRING_CASE(UVM_PERF_FAULT_REPLAY_POLICY_BATCH);
UVM_ENUM_STRING_CASE(UVM_PERF_FAULT_REPLAY_POLICY_BATCH_FLUSH);
UVM_ENUM_STRING_CASE(UVM_PERF_FAULT_REPLAY_POLICY_ONCE);
UVM_ENUM_STRING_DEFAULT();
}
}
NV_STATUS uvm_test_get_prefetch_faults_reenable_lapse(UVM_TEST_GET_PREFETCH_FAULTS_REENABLE_LAPSE_PARAMS *params,
struct file *filp)
{
params->reenable_lapse = uvm_perf_reenable_prefetch_faults_lapse_msec;
return NV_OK;
}
NV_STATUS uvm_test_set_prefetch_faults_reenable_lapse(UVM_TEST_SET_PREFETCH_FAULTS_REENABLE_LAPSE_PARAMS *params,
struct file *filp)
{
uvm_perf_reenable_prefetch_faults_lapse_msec = params->reenable_lapse;
return NV_OK;
}
NV_STATUS uvm_test_drain_replayable_faults(UVM_TEST_DRAIN_REPLAYABLE_FAULTS_PARAMS *params, struct file *filp)
{
uvm_gpu_t *gpu;
NV_STATUS status = NV_OK;
uvm_spin_loop_t spin;
bool pending = true;
uvm_va_space_t *va_space = uvm_va_space_get(filp);
gpu = uvm_va_space_retain_gpu_by_uuid(va_space, &params->gpu_uuid);
if (!gpu)
return NV_ERR_INVALID_DEVICE;
uvm_spin_loop_init(&spin);
do {
uvm_gpu_replayable_faults_isr_lock(gpu->parent);
pending = uvm_gpu_replayable_faults_pending(gpu->parent);
uvm_gpu_replayable_faults_isr_unlock(gpu->parent);
if (!pending)
break;
if (fatal_signal_pending(current)) {
status = NV_ERR_SIGNAL_PENDING;
break;
}
UVM_SPIN_LOOP(&spin);
} while (uvm_spin_loop_elapsed(&spin) < params->timeout_ns);
if (pending && status == NV_OK)
status = NV_ERR_TIMEOUT;
uvm_gpu_release(gpu);
return status;
}