mirror of
https://github.com/NVIDIA/open-gpu-kernel-modules.git
synced 2024-12-03 03:24:18 +01:00
775 lines
32 KiB
C
775 lines
32 KiB
C
/*******************************************************************************
|
|
Copyright (c) 2016-2021 NVIDIA Corporation
|
|
|
|
Permission is hereby granted, free of charge, to any person obtaining a copy
|
|
of this software and associated documentation files (the "Software"), to
|
|
deal in the Software without restriction, including without limitation the
|
|
rights to use, copy, modify, merge, publish, distribute, sublicense, and/or
|
|
sell copies of the Software, and to permit persons to whom the Software is
|
|
furnished to do so, subject to the following conditions:
|
|
|
|
The above copyright notice and this permission notice shall be
|
|
included in all copies or substantial portions of the Software.
|
|
|
|
THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
|
|
IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
|
|
FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
|
|
THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
|
|
LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING
|
|
FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER
|
|
DEALINGS IN THE SOFTWARE.
|
|
|
|
*******************************************************************************/
|
|
|
|
#include "uvm_api.h"
|
|
#include "uvm_global.h"
|
|
#include "uvm_gpu_isr.h"
|
|
#include "uvm_hal.h"
|
|
#include "uvm_gpu.h"
|
|
#include "uvm_gpu_access_counters.h"
|
|
#include "uvm_gpu_non_replayable_faults.h"
|
|
#include "uvm_thread_context.h"
|
|
|
|
// Level-based vs pulse-based interrupts
|
|
// =====================================
|
|
// Turing switches to pulse-based interrupts for replayable/non-replayable
|
|
// faults and access counter notifications. Prior GPUs use level-based
|
|
// interrupts.
|
|
//
|
|
// Level-based interrupts are rearmed automatically as long as the interrupt
|
|
// condition is set. Pulse-based interrupts, on the other hand, are
|
|
// re-triggered by clearing their interrupt line and forcing the interrupt
|
|
// condition to be re-evaluated. However, RM re-triggers all top-level
|
|
// interrupts when exiting its top half. Thus, both level-based and pulse-based
|
|
// interrupts need to be disabled at interrupt handling boundaries, in order to
|
|
// avoid interrupt storms.
|
|
//
|
|
// Moreover, in order to make sure that pulse-based interrupts are not missed,
|
|
// we need to clear the interrupt bit and force a interrupt condition
|
|
// re-evaluation after interrupts are re-enabled. In the case of replayable
|
|
// faults and access counter notifications the interrupt condition is
|
|
// re-evaluated by writing to GET. Non-replayable faults work the same way, but
|
|
// they are currently owned by RM, so UVM doesn't have to do anything.
|
|
|
|
// For use by the nv_kthread_q that is servicing the replayable fault bottom
|
|
// half, only.
|
|
static void replayable_faults_isr_bottom_half_entry(void *args);
|
|
|
|
// For use by the nv_kthread_q that is servicing the replayable fault bottom
|
|
// half, only.
|
|
static void non_replayable_faults_isr_bottom_half_entry(void *args);
|
|
|
|
// For use by the nv_kthread_q that is servicing the replayable fault bottom
|
|
// half, only.
|
|
static void access_counters_isr_bottom_half_entry(void *args);
|
|
|
|
// Increments the reference count tracking whether replayable page fault
|
|
// interrupts should be disabled. The caller is guaranteed that replayable page
|
|
// faults are disabled upon return. Interrupts might already be disabled prior
|
|
// to making this call. Each call is ref-counted, so this must be paired with a
|
|
// call to uvm_gpu_replayable_faults_intr_enable().
|
|
//
|
|
// parent_gpu->isr.interrupts_lock must be held to call this function.
|
|
static void uvm_gpu_replayable_faults_intr_disable(uvm_parent_gpu_t *parent_gpu);
|
|
|
|
// Decrements the reference count tracking whether replayable page fault
|
|
// interrupts should be disabled. Only once the count reaches 0 are the HW
|
|
// interrupts actually enabled, so this call does not guarantee that the
|
|
// interrupts have been re-enabled upon return.
|
|
//
|
|
// uvm_gpu_replayable_faults_intr_disable() must have been called prior to
|
|
// calling this function.
|
|
//
|
|
// parent_gpu->isr.interrupts_lock must be held to call this function.
|
|
static void uvm_gpu_replayable_faults_intr_enable(uvm_parent_gpu_t *parent_gpu);
|
|
|
|
static unsigned schedule_replayable_faults_handler(uvm_parent_gpu_t *parent_gpu)
|
|
{
|
|
// handling gets set to false for all handlers during removal, so quit if
|
|
// the GPU is in the process of being removed.
|
|
if (parent_gpu->isr.replayable_faults.handling) {
|
|
|
|
// Use raw call instead of UVM helper. Ownership will be recorded in the
|
|
// bottom half. See comment replayable_faults_isr_bottom_half().
|
|
if (down_trylock(&parent_gpu->isr.replayable_faults.service_lock.sem) == 0) {
|
|
if (uvm_gpu_replayable_faults_pending(parent_gpu)) {
|
|
nv_kref_get(&parent_gpu->gpu_kref);
|
|
|
|
// Interrupts need to be disabled here to avoid an interrupt
|
|
// storm
|
|
uvm_gpu_replayable_faults_intr_disable(parent_gpu);
|
|
|
|
// Schedule a bottom half, but do *not* release the GPU ISR
|
|
// lock. The bottom half releases the GPU ISR lock as part of
|
|
// its cleanup.
|
|
nv_kthread_q_schedule_q_item(&parent_gpu->isr.bottom_half_q,
|
|
&parent_gpu->isr.replayable_faults.bottom_half_q_item);
|
|
return 1;
|
|
}
|
|
else {
|
|
up(&parent_gpu->isr.replayable_faults.service_lock.sem);
|
|
}
|
|
}
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static unsigned schedule_non_replayable_faults_handler(uvm_parent_gpu_t *parent_gpu)
|
|
{
|
|
// handling gets set to false for all handlers during removal, so quit if
|
|
// the GPU is in the process of being removed.
|
|
if (parent_gpu->isr.non_replayable_faults.handling) {
|
|
// Non-replayable_faults are stored in a synchronized circular queue
|
|
// shared by RM/UVM. Therefore, we can query the number of pending
|
|
// faults. This type of faults are not replayed and since RM advances
|
|
// GET to PUT when copying the fault packets to the queue, no further
|
|
// interrupts will be triggered by the gpu and faults may stay
|
|
// unserviced. Therefore, if there is a fault in the queue, we schedule
|
|
// a bottom half unconditionally.
|
|
if (uvm_gpu_non_replayable_faults_pending(parent_gpu)) {
|
|
bool scheduled;
|
|
nv_kref_get(&parent_gpu->gpu_kref);
|
|
|
|
scheduled = nv_kthread_q_schedule_q_item(&parent_gpu->isr.bottom_half_q,
|
|
&parent_gpu->isr.non_replayable_faults.bottom_half_q_item) != 0;
|
|
|
|
// If the q_item did not get scheduled because it was already
|
|
// queued, that instance will handle the pending faults. Just
|
|
// drop the GPU kref.
|
|
if (!scheduled)
|
|
uvm_parent_gpu_kref_put(parent_gpu);
|
|
|
|
return 1;
|
|
}
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static unsigned schedule_access_counters_handler(uvm_parent_gpu_t *parent_gpu)
|
|
{
|
|
uvm_assert_spinlock_locked(&parent_gpu->isr.interrupts_lock);
|
|
|
|
if (!parent_gpu->isr.access_counters.handling_ref_count)
|
|
return 0;
|
|
|
|
if (down_trylock(&parent_gpu->isr.access_counters.service_lock.sem))
|
|
return 0;
|
|
|
|
if (!uvm_gpu_access_counters_pending(parent_gpu)) {
|
|
up(&parent_gpu->isr.access_counters.service_lock.sem);
|
|
return 0;
|
|
}
|
|
|
|
nv_kref_get(&parent_gpu->gpu_kref);
|
|
|
|
// Interrupts need to be disabled to avoid an interrupt storm
|
|
uvm_gpu_access_counters_intr_disable(parent_gpu);
|
|
|
|
nv_kthread_q_schedule_q_item(&parent_gpu->isr.bottom_half_q,
|
|
&parent_gpu->isr.access_counters.bottom_half_q_item);
|
|
|
|
return 1;
|
|
}
|
|
|
|
// This is called from RM's top-half ISR (see: the nvidia_isr() function), and UVM is given a
|
|
// chance to handle the interrupt, before most of the RM processing. UVM communicates what it
|
|
// did, back to RM, via the return code:
|
|
//
|
|
// NV_OK:
|
|
// UVM handled an interrupt.
|
|
//
|
|
// NV_WARN_MORE_PROCESSING_REQUIRED:
|
|
// UVM did not schedule a bottom half, because it was unable to get the locks it
|
|
// needed, but there is still UVM work to be done. RM will return "not handled" to the
|
|
// Linux kernel, *unless* RM handled other faults in its top half. In that case, the
|
|
// fact that UVM did not handle its interrupt is lost. However, life and interrupt
|
|
// processing continues anyway: the GPU will soon raise another interrupt, because
|
|
// that's what it does when there are replayable page faults remaining (GET != PUT in
|
|
// the fault buffer).
|
|
//
|
|
// NV_ERR_NO_INTR_PENDING:
|
|
// UVM did not find any work to do. Currently this is handled in RM in exactly the same
|
|
// way as NV_WARN_MORE_PROCESSING_REQUIRED is handled. However, the extra precision is
|
|
// available for the future. RM's interrupt handling tends to evolve as new chips and
|
|
// new interrupts get created.
|
|
|
|
static NV_STATUS uvm_isr_top_half(const NvProcessorUuid *gpu_uuid)
|
|
{
|
|
uvm_parent_gpu_t *parent_gpu;
|
|
unsigned num_handlers_scheduled = 0;
|
|
NV_STATUS status;
|
|
|
|
if (!in_interrupt() && in_atomic()) {
|
|
// Early-out if we're not in interrupt context, but memory allocations
|
|
// require GFP_ATOMIC. This happens with CONFIG_DEBUG_SHIRQ enabled,
|
|
// where the interrupt handler is called as part of its removal to make
|
|
// sure it's prepared for being called even when it's being freed.
|
|
// This breaks the assumption that the UVM driver is called in atomic
|
|
// context only in the interrupt context, which the thread context
|
|
// management relies on.
|
|
return NV_OK;
|
|
}
|
|
|
|
if (!gpu_uuid) {
|
|
// This can happen early in the main GPU driver initialization, because
|
|
// that involves testing interrupts before the GPU is fully set up.
|
|
return NV_ERR_NO_INTR_PENDING;
|
|
}
|
|
|
|
uvm_spin_lock_irqsave(&g_uvm_global.gpu_table_lock);
|
|
|
|
parent_gpu = uvm_parent_gpu_get_by_uuid_locked(gpu_uuid);
|
|
|
|
if (parent_gpu == NULL) {
|
|
uvm_spin_unlock_irqrestore(&g_uvm_global.gpu_table_lock);
|
|
return NV_ERR_NO_INTR_PENDING;
|
|
}
|
|
|
|
// We take a reference during the top half, and an additional reference for
|
|
// each scheduled bottom. References are dropped at the end of the bottom
|
|
// halves.
|
|
nv_kref_get(&parent_gpu->gpu_kref);
|
|
uvm_spin_unlock_irqrestore(&g_uvm_global.gpu_table_lock);
|
|
|
|
// Now that we got a GPU object, lock it so that it can't be removed without us noticing.
|
|
uvm_spin_lock_irqsave(&parent_gpu->isr.interrupts_lock);
|
|
|
|
++parent_gpu->isr.interrupt_count;
|
|
|
|
if (parent_gpu->isr.is_suspended) {
|
|
status = NV_ERR_NO_INTR_PENDING;
|
|
}
|
|
else {
|
|
num_handlers_scheduled += schedule_replayable_faults_handler(parent_gpu);
|
|
num_handlers_scheduled += schedule_non_replayable_faults_handler(parent_gpu);
|
|
num_handlers_scheduled += schedule_access_counters_handler(parent_gpu);
|
|
|
|
if (num_handlers_scheduled == 0)
|
|
status = NV_WARN_MORE_PROCESSING_REQUIRED;
|
|
else
|
|
status = NV_OK;
|
|
}
|
|
|
|
uvm_spin_unlock_irqrestore(&parent_gpu->isr.interrupts_lock);
|
|
|
|
uvm_parent_gpu_kref_put(parent_gpu);
|
|
|
|
return status;
|
|
}
|
|
|
|
NV_STATUS uvm_isr_top_half_entry(const NvProcessorUuid *gpu_uuid)
|
|
{
|
|
UVM_ENTRY_RET(uvm_isr_top_half(gpu_uuid));
|
|
}
|
|
|
|
static NV_STATUS init_queue_on_node(nv_kthread_q_t *queue, const char *name, int node)
|
|
{
|
|
#if UVM_THREAD_AFFINITY_SUPPORTED()
|
|
if (node != -1 && !cpumask_empty(uvm_cpumask_of_node(node))) {
|
|
NV_STATUS status;
|
|
|
|
status = errno_to_nv_status(nv_kthread_q_init_on_node(queue, name, node));
|
|
if (status != NV_OK)
|
|
return status;
|
|
|
|
return errno_to_nv_status(set_cpus_allowed_ptr(queue->q_kthread, uvm_cpumask_of_node(node)));
|
|
}
|
|
#endif
|
|
|
|
return errno_to_nv_status(nv_kthread_q_init(queue, name));
|
|
}
|
|
|
|
NV_STATUS uvm_gpu_init_isr(uvm_parent_gpu_t *parent_gpu)
|
|
{
|
|
NV_STATUS status = NV_OK;
|
|
char kthread_name[TASK_COMM_LEN + 1];
|
|
|
|
if (parent_gpu->replayable_faults_supported) {
|
|
status = uvm_gpu_fault_buffer_init(parent_gpu);
|
|
if (status != NV_OK) {
|
|
UVM_ERR_PRINT("Failed to initialize GPU fault buffer: %s, GPU: %s\n",
|
|
nvstatusToString(status),
|
|
parent_gpu->name);
|
|
return status;
|
|
}
|
|
|
|
nv_kthread_q_item_init(&parent_gpu->isr.replayable_faults.bottom_half_q_item,
|
|
replayable_faults_isr_bottom_half_entry,
|
|
parent_gpu);
|
|
|
|
parent_gpu->isr.replayable_faults.stats.cpu_exec_count =
|
|
uvm_kvmalloc_zero(sizeof(*parent_gpu->isr.replayable_faults.stats.cpu_exec_count) * num_possible_cpus());
|
|
if (!parent_gpu->isr.replayable_faults.stats.cpu_exec_count)
|
|
return NV_ERR_NO_MEMORY;
|
|
|
|
parent_gpu->isr.replayable_faults.handling = true;
|
|
|
|
snprintf(kthread_name, sizeof(kthread_name), "UVM GPU%u BH", uvm_id_value(parent_gpu->id));
|
|
status = init_queue_on_node(&parent_gpu->isr.bottom_half_q, kthread_name, parent_gpu->closest_cpu_numa_node);
|
|
if (status != NV_OK) {
|
|
UVM_ERR_PRINT("Failed in nv_kthread_q_init for bottom_half_q: %s, GPU %s\n",
|
|
nvstatusToString(status),
|
|
parent_gpu->name);
|
|
return status;
|
|
}
|
|
|
|
if (parent_gpu->non_replayable_faults_supported) {
|
|
nv_kthread_q_item_init(&parent_gpu->isr.non_replayable_faults.bottom_half_q_item,
|
|
non_replayable_faults_isr_bottom_half_entry,
|
|
parent_gpu);
|
|
|
|
parent_gpu->isr.non_replayable_faults.stats.cpu_exec_count =
|
|
uvm_kvmalloc_zero(sizeof(*parent_gpu->isr.non_replayable_faults.stats.cpu_exec_count) *
|
|
num_possible_cpus());
|
|
if (!parent_gpu->isr.non_replayable_faults.stats.cpu_exec_count)
|
|
return NV_ERR_NO_MEMORY;
|
|
|
|
parent_gpu->isr.non_replayable_faults.handling = true;
|
|
|
|
snprintf(kthread_name, sizeof(kthread_name), "UVM GPU%u KC", uvm_id_value(parent_gpu->id));
|
|
status = init_queue_on_node(&parent_gpu->isr.kill_channel_q,
|
|
kthread_name,
|
|
parent_gpu->closest_cpu_numa_node);
|
|
if (status != NV_OK) {
|
|
UVM_ERR_PRINT("Failed in nv_kthread_q_init for kill_channel_q: %s, GPU %s\n",
|
|
nvstatusToString(status),
|
|
parent_gpu->name);
|
|
return status;
|
|
}
|
|
}
|
|
|
|
if (parent_gpu->access_counters_supported) {
|
|
status = uvm_gpu_init_access_counters(parent_gpu);
|
|
if (status != NV_OK) {
|
|
UVM_ERR_PRINT("Failed to initialize GPU access counters: %s, GPU: %s\n",
|
|
nvstatusToString(status),
|
|
parent_gpu->name);
|
|
return status;
|
|
}
|
|
|
|
nv_kthread_q_item_init(&parent_gpu->isr.access_counters.bottom_half_q_item,
|
|
access_counters_isr_bottom_half_entry,
|
|
parent_gpu);
|
|
|
|
// Access counters interrupts are initially disabled. They are
|
|
// dynamically enabled when the GPU is registered on a VA space.
|
|
parent_gpu->isr.access_counters.handling_ref_count = 0;
|
|
parent_gpu->isr.access_counters.stats.cpu_exec_count =
|
|
uvm_kvmalloc_zero(sizeof(*parent_gpu->isr.access_counters.stats.cpu_exec_count) * num_possible_cpus());
|
|
if (!parent_gpu->isr.access_counters.stats.cpu_exec_count)
|
|
return NV_ERR_NO_MEMORY;
|
|
}
|
|
}
|
|
|
|
return NV_OK;
|
|
}
|
|
|
|
void uvm_gpu_flush_bottom_halves(uvm_parent_gpu_t *parent_gpu)
|
|
{
|
|
nv_kthread_q_flush(&parent_gpu->isr.bottom_half_q);
|
|
nv_kthread_q_flush(&parent_gpu->isr.kill_channel_q);
|
|
}
|
|
|
|
void uvm_gpu_disable_isr(uvm_parent_gpu_t *parent_gpu)
|
|
{
|
|
UVM_ASSERT(parent_gpu->isr.access_counters.handling_ref_count == 0);
|
|
|
|
// Now that the GPU is safely out of the global table, lock the GPU and mark
|
|
// it as no longer handling interrupts so the top half knows not to schedule
|
|
// any more bottom halves.
|
|
uvm_spin_lock_irqsave(&parent_gpu->isr.interrupts_lock);
|
|
|
|
uvm_gpu_replayable_faults_intr_disable(parent_gpu);
|
|
|
|
parent_gpu->isr.replayable_faults.was_handling = parent_gpu->isr.replayable_faults.handling;
|
|
parent_gpu->isr.non_replayable_faults.was_handling = parent_gpu->isr.non_replayable_faults.handling;
|
|
|
|
parent_gpu->isr.replayable_faults.handling = false;
|
|
parent_gpu->isr.non_replayable_faults.handling = false;
|
|
|
|
uvm_spin_unlock_irqrestore(&parent_gpu->isr.interrupts_lock);
|
|
|
|
// Flush all bottom half ISR work items and stop the nv_kthread_q that is
|
|
// servicing this GPU's bottom halves. Note that this requires that the
|
|
// bottom half never take the global lock, since we're holding it here.
|
|
//
|
|
// Note that it's safe to call nv_kthread_q_stop() even if
|
|
// nv_kthread_q_init() failed in uvm_gpu_init_isr().
|
|
nv_kthread_q_stop(&parent_gpu->isr.bottom_half_q);
|
|
nv_kthread_q_stop(&parent_gpu->isr.kill_channel_q);
|
|
}
|
|
|
|
void uvm_gpu_deinit_isr(uvm_parent_gpu_t *parent_gpu)
|
|
{
|
|
// Return ownership to RM:
|
|
if (parent_gpu->isr.replayable_faults.was_handling) {
|
|
// No user threads could have anything left on
|
|
// replayable_faults.disable_intr_ref_count since they must retain the
|
|
// GPU across uvm_gpu_replayable_faults_isr_lock/
|
|
// uvm_gpu_replayable_faults_isr_unlock. This means the
|
|
// uvm_gpu_replayable_faults_disable_intr above could only have raced
|
|
// with bottom halves.
|
|
//
|
|
// If we cleared replayable_faults.handling before the bottom half got
|
|
// to its uvm_gpu_replayable_faults_isr_unlock, when it eventually
|
|
// reached uvm_gpu_replayable_faults_isr_unlock it would have skipped
|
|
// the disable, leaving us with extra ref counts here.
|
|
//
|
|
// In any case we're guaranteed that replayable faults interrupts are
|
|
// disabled and can't get re-enabled, so we can safely ignore the ref
|
|
// count value and just clean things up.
|
|
UVM_ASSERT_MSG(parent_gpu->isr.replayable_faults.disable_intr_ref_count > 0,
|
|
"%s replayable_faults.disable_intr_ref_count: %llu\n",
|
|
parent_gpu->name,
|
|
parent_gpu->isr.replayable_faults.disable_intr_ref_count);
|
|
|
|
uvm_gpu_fault_buffer_deinit(parent_gpu);
|
|
}
|
|
|
|
if (parent_gpu->access_counters_supported) {
|
|
// It is safe to deinitialize access counters even if they have not been
|
|
// successfully initialized.
|
|
uvm_gpu_deinit_access_counters(parent_gpu);
|
|
}
|
|
|
|
uvm_kvfree(parent_gpu->isr.replayable_faults.stats.cpu_exec_count);
|
|
uvm_kvfree(parent_gpu->isr.non_replayable_faults.stats.cpu_exec_count);
|
|
uvm_kvfree(parent_gpu->isr.access_counters.stats.cpu_exec_count);
|
|
}
|
|
|
|
static uvm_gpu_t *find_first_valid_gpu(uvm_parent_gpu_t *parent_gpu)
|
|
{
|
|
uvm_global_gpu_id_t global_gpu_id = uvm_global_gpu_id_from_gpu_id(parent_gpu->id);
|
|
uvm_gpu_t *gpu;
|
|
|
|
// When SMC is enabled, there's no longer a 1:1 relationship between the
|
|
// parent and the partitions. But because all relevant interrupt paths
|
|
// are shared, as is the fault reporting logic, it's sufficient here
|
|
// to proceed with any valid uvm_gpu_t, even if the corresponding partition
|
|
// didn't cause all, or even any of the interrupts.
|
|
// The bottom half handlers will later find the appropriate partitions by
|
|
// attributing the notifications to VA spaces as necessary.
|
|
if (parent_gpu->smc.enabled) {
|
|
NvU32 sub_processor_index;
|
|
|
|
uvm_spin_lock_irqsave(&g_uvm_global.gpu_table_lock);
|
|
|
|
sub_processor_index = find_first_bit(parent_gpu->valid_gpus, UVM_ID_MAX_SUB_PROCESSORS);
|
|
|
|
if (sub_processor_index < UVM_ID_MAX_SUB_PROCESSORS) {
|
|
gpu = uvm_gpu_get(uvm_global_id_from_value(uvm_global_id_value(global_gpu_id) + sub_processor_index));
|
|
UVM_ASSERT(gpu != NULL);
|
|
}
|
|
else {
|
|
gpu = NULL;
|
|
}
|
|
|
|
uvm_spin_unlock_irqrestore(&g_uvm_global.gpu_table_lock);
|
|
}
|
|
else {
|
|
gpu = uvm_gpu_get(global_gpu_id);
|
|
UVM_ASSERT(gpu != NULL);
|
|
}
|
|
|
|
return gpu;
|
|
}
|
|
|
|
static void replayable_faults_isr_bottom_half(void *args)
|
|
{
|
|
uvm_parent_gpu_t *parent_gpu = (uvm_parent_gpu_t *)args;
|
|
uvm_gpu_t *gpu;
|
|
unsigned int cpu;
|
|
|
|
gpu = find_first_valid_gpu(parent_gpu);
|
|
if (gpu == NULL)
|
|
goto put_kref;
|
|
|
|
UVM_ASSERT(parent_gpu->replayable_faults_supported);
|
|
|
|
// Record the lock ownership
|
|
// The service_lock semaphore is taken in the top half using a raw
|
|
// semaphore call (down_trylock()). Here, the lock "ownership" is recorded,
|
|
// using a direct call to uvm_record_lock(). The pair of the two raw calls
|
|
// result in an ownership "transfer" between the top and bottom halves.
|
|
// Due to this ownership transfer, other usages of the service_lock can
|
|
// use the UVM (un)lock helpers to handle lock ownership and record keeping.
|
|
uvm_record_lock(&parent_gpu->isr.replayable_faults.service_lock, UVM_LOCK_FLAGS_MODE_SHARED);
|
|
|
|
// Multiple bottom halves for replayable faults can be running
|
|
// concurrently, but only one can be running this function for a given GPU
|
|
// since we enter with the replayable_faults.service_lock held.
|
|
cpu = get_cpu();
|
|
++parent_gpu->isr.replayable_faults.stats.bottom_half_count;
|
|
cpumask_set_cpu(cpu, &parent_gpu->isr.replayable_faults.stats.cpus_used_mask);
|
|
++parent_gpu->isr.replayable_faults.stats.cpu_exec_count[cpu];
|
|
put_cpu();
|
|
|
|
uvm_gpu_service_replayable_faults(gpu);
|
|
|
|
uvm_gpu_replayable_faults_isr_unlock(parent_gpu);
|
|
|
|
put_kref:
|
|
uvm_parent_gpu_kref_put(parent_gpu);
|
|
}
|
|
|
|
static void replayable_faults_isr_bottom_half_entry(void *args)
|
|
{
|
|
UVM_ENTRY_VOID(replayable_faults_isr_bottom_half(args));
|
|
}
|
|
|
|
static void non_replayable_faults_isr_bottom_half(void *args)
|
|
{
|
|
uvm_parent_gpu_t *parent_gpu = (uvm_parent_gpu_t *)args;
|
|
uvm_gpu_t *gpu;
|
|
unsigned int cpu;
|
|
|
|
gpu = find_first_valid_gpu(parent_gpu);
|
|
if (gpu == NULL)
|
|
goto put_kref;
|
|
|
|
UVM_ASSERT(parent_gpu->non_replayable_faults_supported);
|
|
|
|
uvm_gpu_non_replayable_faults_isr_lock(parent_gpu);
|
|
|
|
// Multiple bottom halves for non-replayable faults can be running
|
|
// concurrently, but only one can enter this section for a given GPU
|
|
// since we acquired the non_replayable_faults.service_lock
|
|
cpu = get_cpu();
|
|
++parent_gpu->isr.non_replayable_faults.stats.bottom_half_count;
|
|
cpumask_set_cpu(cpu, &parent_gpu->isr.non_replayable_faults.stats.cpus_used_mask);
|
|
++parent_gpu->isr.non_replayable_faults.stats.cpu_exec_count[cpu];
|
|
put_cpu();
|
|
|
|
uvm_gpu_service_non_replayable_fault_buffer(gpu);
|
|
|
|
uvm_gpu_non_replayable_faults_isr_unlock(parent_gpu);
|
|
|
|
put_kref:
|
|
uvm_parent_gpu_kref_put(parent_gpu);
|
|
}
|
|
|
|
static void non_replayable_faults_isr_bottom_half_entry(void *args)
|
|
{
|
|
UVM_ENTRY_VOID(non_replayable_faults_isr_bottom_half(args));
|
|
}
|
|
|
|
static void access_counters_isr_bottom_half(void *args)
|
|
{
|
|
uvm_parent_gpu_t *parent_gpu = (uvm_parent_gpu_t *)args;
|
|
uvm_gpu_t *gpu;
|
|
unsigned int cpu;
|
|
|
|
gpu = find_first_valid_gpu(parent_gpu);
|
|
if (gpu == NULL)
|
|
goto put_kref;
|
|
|
|
UVM_ASSERT(parent_gpu->access_counters_supported);
|
|
|
|
uvm_record_lock(&parent_gpu->isr.access_counters.service_lock, UVM_LOCK_FLAGS_MODE_SHARED);
|
|
|
|
// Multiple bottom halves for counter notifications can be running
|
|
// concurrently, but only one can be running this function for a given GPU
|
|
// since we enter with the access_counters_isr_lock held.
|
|
cpu = get_cpu();
|
|
++parent_gpu->isr.access_counters.stats.bottom_half_count;
|
|
cpumask_set_cpu(cpu, &parent_gpu->isr.access_counters.stats.cpus_used_mask);
|
|
++parent_gpu->isr.access_counters.stats.cpu_exec_count[cpu];
|
|
put_cpu();
|
|
|
|
uvm_gpu_service_access_counters(gpu);
|
|
|
|
uvm_gpu_access_counters_isr_unlock(parent_gpu);
|
|
|
|
put_kref:
|
|
uvm_parent_gpu_kref_put(parent_gpu);
|
|
}
|
|
|
|
static void access_counters_isr_bottom_half_entry(void *args)
|
|
{
|
|
UVM_ENTRY_VOID(access_counters_isr_bottom_half(args));
|
|
}
|
|
|
|
void uvm_gpu_replayable_faults_isr_lock(uvm_parent_gpu_t *parent_gpu)
|
|
{
|
|
UVM_ASSERT(nv_kref_read(&parent_gpu->gpu_kref) > 0);
|
|
|
|
uvm_spin_lock_irqsave(&parent_gpu->isr.interrupts_lock);
|
|
|
|
// Bump the disable ref count. This guarantees that the bottom half or
|
|
// another thread trying to take the replayable_faults.service_lock won't
|
|
// inadvertently re-enable interrupts during this locking sequence.
|
|
uvm_gpu_replayable_faults_intr_disable(parent_gpu);
|
|
|
|
uvm_spin_unlock_irqrestore(&parent_gpu->isr.interrupts_lock);
|
|
|
|
// Now that we know replayable fault interrupts can't get enabled, take the
|
|
// lock.
|
|
uvm_down(&parent_gpu->isr.replayable_faults.service_lock);
|
|
}
|
|
|
|
void uvm_gpu_replayable_faults_isr_unlock(uvm_parent_gpu_t *parent_gpu)
|
|
{
|
|
UVM_ASSERT(nv_kref_read(&parent_gpu->gpu_kref) > 0);
|
|
|
|
uvm_spin_lock_irqsave(&parent_gpu->isr.interrupts_lock);
|
|
|
|
// The following sequence is delicate:
|
|
//
|
|
// 1) Enable replayable page fault interrupts
|
|
// 2) Rearm pulse based interrupts
|
|
// 3) Unlock GPU isr.replayable_faults.service_lock (mutex)
|
|
// 4) Unlock isr.interrupts_lock (spin lock)
|
|
//
|
|
// ...because the moment that page fault interrupts are reenabled, a top
|
|
// half might start receiving them. A top-half cannot run on the core
|
|
// executing this code as interrupts are disabled as long as the
|
|
// interrupts_lock is held. If it runs on a different core, it's going to
|
|
// spin waiting for the interrupts_lock to be released by this core before
|
|
// attempting to acquire the service_lock mutex. Hence there is no risk of
|
|
// the top-half missing interrupts after they are reenabled, but before the
|
|
// service_lock mutex is released.
|
|
|
|
if (parent_gpu->isr.replayable_faults.handling) {
|
|
// Turn page fault interrupts back on, unless remove_gpu() has already removed this GPU
|
|
// from the GPU table. remove_gpu() indicates that situation by setting
|
|
// gpu->replayable_faults.handling to false.
|
|
//
|
|
// This path can only be taken from the bottom half. User threads
|
|
// calling this function must have previously retained the GPU, so they
|
|
// can't race with remove_gpu.
|
|
//
|
|
// TODO: Bug 1766600: Assert that we're in a bottom half thread, once
|
|
// that's tracked by the lock assertion code.
|
|
//
|
|
// Note that if we're in the bottom half and the GPU was removed before
|
|
// we checked replayable_faults.handling, we won't drop our interrupt
|
|
// disable ref count from the corresponding top-half call to
|
|
// uvm_gpu_replayable_faults_intr_disable. That's ok because remove_gpu
|
|
// ignores the refcount after waiting for the bottom half to finish.
|
|
uvm_gpu_replayable_faults_intr_enable(parent_gpu);
|
|
|
|
// Rearm pulse interrupts. This guarantees that the state of the pending
|
|
// interrupt is current and the top level rearm performed by RM is only
|
|
// going to trigger it if necessary. This avoids both of the possible
|
|
// bad cases:
|
|
// 1) GET != PUT but interrupt state is not pending
|
|
// This could lead to the interrupt being lost.
|
|
// 2) GET == PUT but interrupt state is pending
|
|
// This could lead to an interrupt storm as the top-half would see
|
|
// no work to be done, but the interrupt would get constantly
|
|
// retriggered by RM's top level rearm.
|
|
// clear_replayable_faults is a no-op for architectures that don't
|
|
// support pulse-based interrupts.
|
|
parent_gpu->fault_buffer_hal->clear_replayable_faults(parent_gpu,
|
|
parent_gpu->fault_buffer_info.replayable.cached_get);
|
|
}
|
|
|
|
// This unlock call has to be out-of-order unlock due to interrupts_lock
|
|
// still being held. Otherwise, it would result in a lock order violation.
|
|
uvm_up_out_of_order(&parent_gpu->isr.replayable_faults.service_lock);
|
|
|
|
uvm_spin_unlock_irqrestore(&parent_gpu->isr.interrupts_lock);
|
|
}
|
|
|
|
void uvm_gpu_non_replayable_faults_isr_lock(uvm_parent_gpu_t *parent_gpu)
|
|
{
|
|
UVM_ASSERT(nv_kref_read(&parent_gpu->gpu_kref) > 0);
|
|
|
|
uvm_down(&parent_gpu->isr.non_replayable_faults.service_lock);
|
|
}
|
|
|
|
void uvm_gpu_non_replayable_faults_isr_unlock(uvm_parent_gpu_t *parent_gpu)
|
|
{
|
|
UVM_ASSERT(nv_kref_read(&parent_gpu->gpu_kref) > 0);
|
|
|
|
uvm_up(&parent_gpu->isr.non_replayable_faults.service_lock);
|
|
}
|
|
|
|
void uvm_gpu_access_counters_isr_lock(uvm_parent_gpu_t *parent_gpu)
|
|
{
|
|
// See comments in uvm_gpu_replayable_faults_isr_lock
|
|
|
|
uvm_spin_lock_irqsave(&parent_gpu->isr.interrupts_lock);
|
|
|
|
uvm_gpu_access_counters_intr_disable(parent_gpu);
|
|
|
|
uvm_spin_unlock_irqrestore(&parent_gpu->isr.interrupts_lock);
|
|
|
|
uvm_down(&parent_gpu->isr.access_counters.service_lock);
|
|
}
|
|
|
|
void uvm_gpu_access_counters_isr_unlock(uvm_parent_gpu_t *parent_gpu)
|
|
{
|
|
UVM_ASSERT(nv_kref_read(&parent_gpu->gpu_kref) > 0);
|
|
|
|
// See comments in uvm_gpu_replayable_faults_isr_unlock
|
|
|
|
uvm_spin_lock_irqsave(&parent_gpu->isr.interrupts_lock);
|
|
|
|
uvm_gpu_access_counters_intr_enable(parent_gpu);
|
|
|
|
if (parent_gpu->isr.access_counters.handling_ref_count > 0) {
|
|
parent_gpu->access_counter_buffer_hal->clear_access_counter_notifications(parent_gpu,
|
|
parent_gpu->access_counter_buffer_info.cached_get);
|
|
}
|
|
|
|
// This unlock call has to be out-of-order unlock due to interrupts_lock
|
|
// still being held. Otherwise, it would result in a lock order violation.
|
|
uvm_up_out_of_order(&parent_gpu->isr.access_counters.service_lock);
|
|
|
|
uvm_spin_unlock_irqrestore(&parent_gpu->isr.interrupts_lock);
|
|
}
|
|
|
|
static void uvm_gpu_replayable_faults_intr_disable(uvm_parent_gpu_t *parent_gpu)
|
|
{
|
|
uvm_assert_spinlock_locked(&parent_gpu->isr.interrupts_lock);
|
|
|
|
if (parent_gpu->isr.replayable_faults.handling && parent_gpu->isr.replayable_faults.disable_intr_ref_count == 0)
|
|
parent_gpu->fault_buffer_hal->disable_replayable_faults(parent_gpu);
|
|
|
|
++parent_gpu->isr.replayable_faults.disable_intr_ref_count;
|
|
}
|
|
|
|
static void uvm_gpu_replayable_faults_intr_enable(uvm_parent_gpu_t *parent_gpu)
|
|
{
|
|
uvm_assert_spinlock_locked(&parent_gpu->isr.interrupts_lock);
|
|
UVM_ASSERT(parent_gpu->isr.replayable_faults.disable_intr_ref_count > 0);
|
|
|
|
--parent_gpu->isr.replayable_faults.disable_intr_ref_count;
|
|
if (parent_gpu->isr.replayable_faults.handling && parent_gpu->isr.replayable_faults.disable_intr_ref_count == 0)
|
|
parent_gpu->fault_buffer_hal->enable_replayable_faults(parent_gpu);
|
|
}
|
|
|
|
void uvm_gpu_access_counters_intr_disable(uvm_parent_gpu_t *parent_gpu)
|
|
{
|
|
uvm_assert_spinlock_locked(&parent_gpu->isr.interrupts_lock);
|
|
|
|
// The read of handling_ref_count could race with a write from
|
|
// gpu_access_counters_enable/disable, since here we may not hold the
|
|
// ISR lock. But those functions are invoked with the interrupt disabled
|
|
// (disable_intr_ref_count > 0), so the check always returns false when the
|
|
// race occurs
|
|
if (parent_gpu->isr.access_counters.handling_ref_count > 0 &&
|
|
parent_gpu->isr.access_counters.disable_intr_ref_count == 0) {
|
|
parent_gpu->access_counter_buffer_hal->disable_access_counter_notifications(parent_gpu);
|
|
}
|
|
|
|
++parent_gpu->isr.access_counters.disable_intr_ref_count;
|
|
}
|
|
|
|
void uvm_gpu_access_counters_intr_enable(uvm_parent_gpu_t *parent_gpu)
|
|
{
|
|
uvm_assert_spinlock_locked(&parent_gpu->isr.interrupts_lock);
|
|
UVM_ASSERT(uvm_sem_is_locked(&parent_gpu->isr.access_counters.service_lock));
|
|
UVM_ASSERT(parent_gpu->isr.access_counters.disable_intr_ref_count > 0);
|
|
|
|
--parent_gpu->isr.access_counters.disable_intr_ref_count;
|
|
|
|
if (parent_gpu->isr.access_counters.handling_ref_count > 0 &&
|
|
parent_gpu->isr.access_counters.disable_intr_ref_count == 0) {
|
|
parent_gpu->access_counter_buffer_hal->enable_access_counter_notifications(parent_gpu);
|
|
}
|
|
}
|