mirror of
https://github.com/NVIDIA/open-gpu-kernel-modules.git
synced 2024-12-02 00:24:22 +01:00
258 lines
10 KiB
C
258 lines
10 KiB
C
/*******************************************************************************
|
|
Copyright (c) 2016-2022 NVIDIA Corporation
|
|
|
|
Permission is hereby granted, free of charge, to any person obtaining a copy
|
|
of this software and associated documentation files (the "Software"), to
|
|
deal in the Software without restriction, including without limitation the
|
|
rights to use, copy, modify, merge, publish, distribute, sublicense, and/or
|
|
sell copies of the Software, and to permit persons to whom the Software is
|
|
furnished to do so, subject to the following conditions:
|
|
|
|
The above copyright notice and this permission notice shall be
|
|
included in all copies or substantial portions of the Software.
|
|
|
|
THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
|
|
IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
|
|
FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
|
|
THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
|
|
LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING
|
|
FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER
|
|
DEALINGS IN THE SOFTWARE.
|
|
|
|
*******************************************************************************/
|
|
|
|
#ifndef __UVM_USER_CHANNEL_H__
|
|
#define __UVM_USER_CHANNEL_H__
|
|
|
|
#include "uvm_forward_decl.h"
|
|
#include "uvm_va_space.h"
|
|
#include "uvm_hal_types.h"
|
|
#include "uvm_rb_tree.h"
|
|
#include "nv-kref.h"
|
|
|
|
// This structure contains the VA spaces of all the subcontexts in a TSG. It
|
|
// is stored in a per-GPU UVM RB tree and is required to perform instance_ptr
|
|
// to VA space translations when channels are registered in a subcontext,
|
|
// since SM fault/access counter notification packets may report any
|
|
// instance_ptr in the TSG.
|
|
typedef struct
|
|
{
|
|
// Number of instance pointers referencing this subcontext info descriptor
|
|
NvU32 total_refcount;
|
|
|
|
// All channels in a TSG must be bound to the same SMC Engine
|
|
NvU32 smc_engine_id;
|
|
|
|
// Array of per-subcontext information
|
|
struct
|
|
{
|
|
uvm_va_space_t *va_space;
|
|
|
|
// Number of instance pointers referencing this specific subcontext
|
|
NvU32 refcount;
|
|
} *subctxs;
|
|
|
|
// UVM RB tree node for insertion into the parent GPU's tsg_table.
|
|
uvm_rb_tree_node_t node;
|
|
} uvm_user_channel_subctx_info_t;
|
|
|
|
struct uvm_user_channel_struct
|
|
{
|
|
// Parent GPU VA space
|
|
uvm_gpu_va_space_t *gpu_va_space;
|
|
|
|
// Parent GPU. This is also available in gpu_va_space->gpu, but we need a
|
|
// separate pointer which outlives the gpu_va_space during deferred channel
|
|
// teardown.
|
|
uvm_gpu_t *gpu;
|
|
|
|
// RM handles used to register this channel. We store them for UVM-internal
|
|
// purposes to look up the uvm_user_channel_t for unregistration, rather
|
|
// than introducing a new "UVM channel handle" object for user space.
|
|
//
|
|
// DO NOT TRUST THESE VALUES AFTER UVM_REGISTER_CHANNEL. They are passed by
|
|
// user-space at channel registration time to validate the channel with RM,
|
|
// but afterwards the user could free and reallocate either of the client or
|
|
// object handles, so we can't pass them to RM trusting that they still
|
|
// represent this channel.
|
|
//
|
|
// That's ok because we never pass these handles to RM again after
|
|
// registration.
|
|
uvm_rm_user_object_t user_rm_channel;
|
|
|
|
// Type of the engine the channel is bound to
|
|
UVM_GPU_CHANNEL_ENGINE_TYPE engine_type;
|
|
|
|
// true if the channel belongs to a subcontext or false if the channel
|
|
// belongs to a regular context
|
|
bool in_subctx;
|
|
|
|
// Subcontext ID, aka VEID
|
|
NvU32 subctx_id;
|
|
|
|
struct
|
|
{
|
|
// Whether the channel belongs to a TSG or not
|
|
bool valid;
|
|
|
|
// If valid is true, tsg_id contains the ID of the TSG
|
|
NvU32 id;
|
|
|
|
// If valid is true, this is the maximum number of subcontexts in the
|
|
// TSG
|
|
NvU32 max_subctx_count;
|
|
} tsg;
|
|
|
|
// On Turing+, the CLEAR_FAULTED method requires passing a RM-provided
|
|
// handle to identify the channel.
|
|
NvU32 clear_faulted_token;
|
|
|
|
// Tracker used to aggregate clear faulted operations, needed for user
|
|
// channel removal
|
|
uvm_tracker_t clear_faulted_tracker;
|
|
|
|
// Address of the NV_CHRAM_CHANNEL register and the runlist PRI base
|
|
// register. Only valid on GPUs with
|
|
// non_replayable_faults_supported && !has_clear_faulted_channel_method.
|
|
volatile NvU32 *chram_channel_register;
|
|
volatile NvU32 *runlist_pri_base_register;
|
|
|
|
// Id of the SMC engine this channel is bound to, or zero if the GPU
|
|
// does not support SMC or it is a CE channel
|
|
NvU32 smc_engine_id;
|
|
|
|
// VEIDs are partitioned under SMC (each SMC engine owns a subrange of
|
|
// VEIDs), but the VEID reported in fault packets* is the global (although
|
|
// GFID-local) VEID. In order to compute the SMC engine-local VEID, we need
|
|
// to subtract smc_engine_ve_id_offset from the reported one.
|
|
//
|
|
// *Access counter packets already report SMC engine-local VEIDs.
|
|
NvU32 smc_engine_ve_id_offset;
|
|
|
|
// If in_subctx is true, subctx_info will point at a per-TSG data structure
|
|
// that contains the VA spaces of all the subcontexts in the TSG. This value
|
|
// is assigned in uvm_gpu_add_user_channel.
|
|
uvm_user_channel_subctx_info_t *subctx_info;
|
|
|
|
// Number of resources reported by RM. This is the size of both the
|
|
// resources and va_ranges arrays.
|
|
size_t num_resources;
|
|
|
|
// Array of all resources for this channel, shared or not. Virtual mappings
|
|
// for matching physical resources are shared across all channels in the
|
|
// same GPU VA space and TSG. Each channel will retain the mappings (VA
|
|
// ranges) it uses at channel register and will release them at
|
|
// uvm_user_channel_destroy_detached, so these physical resources outlive
|
|
// the corresponding VA ranges.
|
|
UvmGpuChannelResourceInfo *resources;
|
|
|
|
// Array of all VA ranges associated with this channel. Entry i in this
|
|
// array corresponds to resource i in the resources array above and has the
|
|
// same descriptor. uvm_user_channel_detach will drop the ref counts for
|
|
// these VA ranges, potentially destroying them.
|
|
uvm_va_range_t **va_ranges;
|
|
|
|
// Physical instance pointer. There is a 1:1 mapping between instance
|
|
// pointer and channel. GPU faults report an instance pointer, and the GPU
|
|
// fault handler converts this instance pointer into the parent
|
|
// uvm_va_space_t.
|
|
struct
|
|
{
|
|
// Physical address of the instance pointer.
|
|
uvm_gpu_phys_address_t addr;
|
|
|
|
// Node for inserting the user channel in the parent GPU instance
|
|
// pointer table. The node will be initialized as an empty UVM RB node
|
|
// on user channel creation and will transition to not empty when
|
|
// instance_ptr -> user_channel translation has been added
|
|
// to the per-GPU UVM RB tree
|
|
uvm_rb_tree_node_t node;
|
|
} instance_ptr;
|
|
|
|
// Opaque object which refers to this channel in the nvUvmInterface APIs
|
|
void *rm_retained_channel;
|
|
|
|
// Hardware runlist and channel IDs, used for debugging and fault processing
|
|
NvU32 hw_runlist_id;
|
|
NvU32 hw_channel_id;
|
|
|
|
// Node in the owning gpu_va_space's registered_channels list. Cleared once
|
|
// the channel is detached.
|
|
struct list_head list_node;
|
|
|
|
// Boolean which is set during the window between
|
|
// nvUvmInterfaceBindChannelResources and nvUvmInterfaceStopChannel. This is
|
|
// an atomic_t because multiple threads may call nvUvmInterfaceStopChannel
|
|
// and clear this concurrently.
|
|
atomic_t is_bound;
|
|
|
|
// Node for the deferred free list where this channel is stored upon being
|
|
// detached.
|
|
uvm_deferred_free_object_t deferred_free;
|
|
|
|
// Reference count for this user channel. This only protects the memory
|
|
// object itself, for use in cases when user channel needs to be accessed
|
|
// across dropping and re-acquiring the VA space lock.
|
|
nv_kref_t kref;
|
|
|
|
struct
|
|
{
|
|
bool scheduled;
|
|
|
|
nv_kthread_q_item_t kill_channel_q_item;
|
|
|
|
uvm_va_space_t *va_space;
|
|
|
|
char fault_packet[UVM_GPU_MMU_MAX_FAULT_PACKET_SIZE];
|
|
} kill_channel;
|
|
};
|
|
|
|
// Retains the user channel memory object. uvm_user_channel_destroy_detached and
|
|
// uvm_user_channel_release drop the count. This is used to keep the
|
|
// user channel object allocated when dropping and re-taking the VA space lock.
|
|
// If another thread called uvm_user_channel_detach in the meantime,
|
|
// user_channel->gpu_va_space will be NULL.
|
|
static inline void uvm_user_channel_retain(uvm_user_channel_t *user_channel)
|
|
{
|
|
nv_kref_get(&user_channel->kref);
|
|
}
|
|
|
|
// This only frees the user channel object itself, so the user channel must have
|
|
// been detached and destroyed prior to the final release.
|
|
void uvm_user_channel_release(uvm_user_channel_t *user_channel);
|
|
|
|
// User-facing APIs (uvm_api_register_channel, uvm_api_unregister_channel) are
|
|
// declared in uvm_api.h.
|
|
|
|
// First phase of user channel destroy which stops a user channel, forcibly if
|
|
// necessary. After calling this function no new GPU faults targeting this
|
|
// channel will arrive, but old faults may continue to be serviced.
|
|
//
|
|
// LOCKING: The owning VA space must be locked in read mode, not write mode.
|
|
void uvm_user_channel_stop(uvm_user_channel_t *user_channel);
|
|
|
|
// Second phase of user channel destroy which detaches the channel from the
|
|
// parent gpu_va_space and adds it to the list of pending objects to be freed.
|
|
// uvm_user_channel_stop must have already been called on this channel.
|
|
//
|
|
// All virtual mappings associated with the channel are torn down. The
|
|
// user_channel object and the instance pointer and resources it contains are
|
|
// not destroyed. The caller must use uvm_user_channel_destroy_detached to do
|
|
// that.
|
|
//
|
|
// This multi-phase approach allows the caller to drop the VA space lock and
|
|
// flush the fault buffer before removing the instance pointer. See
|
|
// uvm_gpu_destroy_detached_channels.
|
|
//
|
|
// LOCKING: The owning VA space must be locked in write mode.
|
|
void uvm_user_channel_detach(uvm_user_channel_t *user_channel, struct list_head *deferred_free_list);
|
|
|
|
// Third phase of user channel destroy which frees the user_channel object and
|
|
// releases the corresponding resources and instance pointer. The channel must
|
|
// have been detached first.
|
|
//
|
|
// LOCKING: No lock is required, but the owning GPU must be retained.
|
|
void uvm_user_channel_destroy_detached(uvm_user_channel_t *user_channel);
|
|
|
|
#endif // __UVM_USER_CHANNEL_H__
|