open-gpu-kernel-modules/kernel-open/nvidia/nv-dma.c
2022-05-27 16:40:24 -07:00

1311 lines
29 KiB
C

/*
* SPDX-FileCopyrightText: Copyright (c) 1999-2015 NVIDIA CORPORATION & AFFILIATES. All rights reserved.
* SPDX-License-Identifier: MIT
*
* Permission is hereby granted, free of charge, to any person obtaining a
* copy of this software and associated documentation files (the "Software"),
* to deal in the Software without restriction, including without limitation
* the rights to use, copy, modify, merge, publish, distribute, sublicense,
* and/or sell copies of the Software, and to permit persons to whom the
* Software is furnished to do so, subject to the following conditions:
*
* The above copyright notice and this permission notice shall be included in
* all copies or substantial portions of the Software.
*
* THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
* IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
* FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
* THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
* LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING
* FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER
* DEALINGS IN THE SOFTWARE.
*/
#define __NO_VERSION__
#include "os-interface.h"
#include "nv-linux.h"
#include "nv-reg.h"
#define NV_DMA_DEV_PRINTF(debuglevel, dma_dev, format, ... ) \
nv_printf(debuglevel, "NVRM: %s: " format, \
(((dma_dev) && ((dma_dev)->dev)) ? dev_name((dma_dev)->dev) : \
NULL), \
## __VA_ARGS__)
NvU32 nv_dma_remap_peer_mmio = NV_DMA_REMAP_PEER_MMIO_ENABLE;
NV_STATUS nv_create_dma_map_scatterlist (nv_dma_map_t *dma_map);
void nv_destroy_dma_map_scatterlist(nv_dma_map_t *dma_map);
NV_STATUS nv_map_dma_map_scatterlist (nv_dma_map_t *dma_map);
void nv_unmap_dma_map_scatterlist (nv_dma_map_t *dma_map);
static void nv_dma_unmap_contig (nv_dma_map_t *dma_map);
static void nv_dma_unmap_scatterlist (nv_dma_map_t *dma_map);
static inline NvBool nv_dma_is_addressable(
nv_dma_device_t *dma_dev,
NvU64 start,
NvU64 size
)
{
NvU64 limit = start + size - 1;
return (start >= dma_dev->addressable_range.start) &&
(limit <= dma_dev->addressable_range.limit) &&
(limit >= start);
}
static NV_STATUS nv_dma_map_contig(
nv_dma_device_t *dma_dev,
nv_dma_map_t *dma_map,
NvU64 *va
)
{
#if defined(NV_DMA_MAP_PAGE_ATTRS_PRESENT) && defined(NV_DMA_ATTR_SKIP_CPU_SYNC_PRESENT)
*va = dma_map_page_attrs(dma_map->dev, dma_map->pages[0], 0,
dma_map->page_count * PAGE_SIZE,
DMA_BIDIRECTIONAL,
(dma_map->cache_type == NV_MEMORY_UNCACHED) ?
DMA_ATTR_SKIP_CPU_SYNC : 0);
#else
*va = dma_map_page(dma_map->dev, dma_map->pages[0], 0,
dma_map->page_count * PAGE_SIZE, DMA_BIDIRECTIONAL);
#endif
if (dma_mapping_error(dma_map->dev, *va))
{
return NV_ERR_OPERATING_SYSTEM;
}
dma_map->mapping.contig.dma_addr = *va;
if (!nv_dma_is_addressable(dma_dev, *va, dma_map->page_count * PAGE_SIZE))
{
NV_DMA_DEV_PRINTF(NV_DBG_ERRORS, dma_dev,
"DMA address not in addressable range of device "
"(0x%llx-0x%llx, 0x%llx-0x%llx)\n",
*va, *va + (dma_map->page_count * PAGE_SIZE - 1),
dma_dev->addressable_range.start,
dma_dev->addressable_range.limit);
nv_dma_unmap_contig(dma_map);
return NV_ERR_INVALID_ADDRESS;
}
return NV_OK;
}
static void nv_dma_unmap_contig(nv_dma_map_t *dma_map)
{
#if defined(NV_DMA_MAP_PAGE_ATTRS_PRESENT) && defined(NV_DMA_ATTR_SKIP_CPU_SYNC_PRESENT)
dma_unmap_page_attrs(dma_map->dev, dma_map->mapping.contig.dma_addr,
dma_map->page_count * PAGE_SIZE,
DMA_BIDIRECTIONAL,
(dma_map->cache_type == NV_MEMORY_UNCACHED) ?
DMA_ATTR_SKIP_CPU_SYNC : 0);
#else
dma_unmap_page(dma_map->dev, dma_map->mapping.contig.dma_addr,
dma_map->page_count * PAGE_SIZE, DMA_BIDIRECTIONAL);
#endif
}
static void nv_fill_scatterlist
(
struct scatterlist *sgl,
struct page **pages,
unsigned int page_count
)
{
unsigned int i;
struct scatterlist *sg;
#if defined(for_each_sg)
for_each_sg(sgl, sg, page_count, i)
{
sg_set_page(sg, pages[i], PAGE_SIZE, 0);
}
#else
for (i = 0; i < page_count; i++)
{
sg = &(sgl)[i];
sg->page = pages[i];
sg->length = PAGE_SIZE;
sg->offset = 0;
}
#endif
}
NV_STATUS nv_create_dma_map_scatterlist(nv_dma_map_t *dma_map)
{
/*
* We need to split our mapping into at most 4GB - PAGE_SIZE chunks.
* The Linux kernel stores the length (and offset) of a scatter-gather
* segment as an unsigned int, so it will overflow if we try to do
* anything larger.
*/
NV_STATUS status;
nv_dma_submap_t *submap;
NvU32 i;
NvU64 allocated_size = 0;
NvU64 num_submaps = dma_map->page_count + NV_DMA_SUBMAP_MAX_PAGES - 1;
NvU64 total_size = dma_map->page_count << PAGE_SHIFT;
/*
* This turns into 64-bit division, which the ARMv7 kernel doesn't provide
* implicitly. Instead, we need to use the platform's do_div() to perform
* the division.
*/
do_div(num_submaps, NV_DMA_SUBMAP_MAX_PAGES);
WARN_ON(NvU64_HI32(num_submaps) != 0);
if (dma_map->import_sgt && (num_submaps != 1))
{
return -EINVAL;
}
dma_map->mapping.discontig.submap_count = NvU64_LO32(num_submaps);
status = os_alloc_mem((void **)&dma_map->mapping.discontig.submaps,
sizeof(nv_dma_submap_t) * dma_map->mapping.discontig.submap_count);
if (status != NV_OK)
{
return status;
}
os_mem_set((void *)dma_map->mapping.discontig.submaps, 0,
sizeof(nv_dma_submap_t) * dma_map->mapping.discontig.submap_count);
/* If we have an imported SGT, just use that directly. */
if (dma_map->import_sgt)
{
dma_map->mapping.discontig.submaps[0].page_count = dma_map->page_count;
dma_map->mapping.discontig.submaps[0].sgt = *dma_map->import_sgt;
dma_map->mapping.discontig.submaps[0].imported = NV_TRUE;
return status;
}
NV_FOR_EACH_DMA_SUBMAP(dma_map, submap, i)
{
NvU64 submap_size = NV_MIN(NV_DMA_SUBMAP_MAX_PAGES << PAGE_SHIFT,
total_size - allocated_size);
submap->page_count = (NvU32)(submap_size >> PAGE_SHIFT);
status = NV_ALLOC_DMA_SUBMAP_SCATTERLIST(dma_map, submap, i);
if (status != NV_OK)
{
submap->page_count = 0;
break;
}
#if !defined(NV_SG_ALLOC_TABLE_FROM_PAGES_PRESENT) || \
defined(NV_DOM0_KERNEL_PRESENT)
{
NvU64 page_idx = NV_DMA_SUBMAP_IDX_TO_PAGE_IDX(i);
nv_fill_scatterlist(submap->sgt.sgl,
&dma_map->pages[page_idx], submap->page_count);
}
#endif
allocated_size += submap_size;
}
WARN_ON(allocated_size != total_size);
if (status != NV_OK)
{
nv_destroy_dma_map_scatterlist(dma_map);
}
return status;
}
NV_STATUS nv_map_dma_map_scatterlist(nv_dma_map_t *dma_map)
{
NV_STATUS status = NV_OK;
nv_dma_submap_t *submap;
NvU64 i;
NV_FOR_EACH_DMA_SUBMAP(dma_map, submap, i)
{
/* Imported SGTs will have already been mapped by the exporter. */
submap->sg_map_count = submap->imported ?
submap->sgt.orig_nents :
dma_map_sg(dma_map->dev,
submap->sgt.sgl,
submap->sgt.orig_nents,
DMA_BIDIRECTIONAL);
if (submap->sg_map_count == 0)
{
status = NV_ERR_OPERATING_SYSTEM;
break;
}
}
if (status != NV_OK)
{
nv_unmap_dma_map_scatterlist(dma_map);
}
return status;
}
void nv_unmap_dma_map_scatterlist(nv_dma_map_t *dma_map)
{
nv_dma_submap_t *submap;
NvU64 i;
NV_FOR_EACH_DMA_SUBMAP(dma_map, submap, i)
{
if (submap->sg_map_count == 0)
{
break;
}
if (submap->imported)
{
/* Imported SGTs will be unmapped by the exporter. */
continue;
}
dma_unmap_sg(dma_map->dev, submap->sgt.sgl,
submap->sgt.orig_nents,
DMA_BIDIRECTIONAL);
}
}
void nv_destroy_dma_map_scatterlist(nv_dma_map_t *dma_map)
{
nv_dma_submap_t *submap;
NvU64 i;
NV_FOR_EACH_DMA_SUBMAP(dma_map, submap, i)
{
if ((submap->page_count == 0) || submap->imported)
{
break;
}
sg_free_table(&submap->sgt);
}
os_free_mem(dma_map->mapping.discontig.submaps);
}
void nv_load_dma_map_scatterlist(
nv_dma_map_t *dma_map,
NvU64 *va_array
)
{
unsigned int i, j;
struct scatterlist *sg;
nv_dma_submap_t *submap;
NvU64 sg_addr, sg_off, sg_len, k, l = 0;
NV_FOR_EACH_DMA_SUBMAP(dma_map, submap, i)
{
for_each_sg(submap->sgt.sgl, sg, submap->sg_map_count, j)
{
/*
* It is possible for pci_map_sg() to merge scatterlist entries, so
* make sure we account for that here.
*/
for (sg_addr = sg_dma_address(sg), sg_len = sg_dma_len(sg),
sg_off = 0, k = 0;
(sg_off < sg_len) && (k < submap->page_count);
sg_off += PAGE_SIZE, l++, k++)
{
va_array[l] = sg_addr + sg_off;
}
}
}
}
static NV_STATUS nv_dma_map_scatterlist(
nv_dma_device_t *dma_dev,
nv_dma_map_t *dma_map,
NvU64 *va_array
)
{
NV_STATUS status;
NvU64 i;
status = nv_create_dma_map_scatterlist(dma_map);
if (status != NV_OK)
{
NV_DMA_DEV_PRINTF(NV_DBG_ERRORS, dma_dev,
"Failed to allocate DMA mapping scatterlist!\n");
return status;
}
status = nv_map_dma_map_scatterlist(dma_map);
if (status != NV_OK)
{
NV_DMA_DEV_PRINTF(NV_DBG_ERRORS, dma_dev,
"Failed to create a DMA mapping!\n");
nv_destroy_dma_map_scatterlist(dma_map);
return status;
}
nv_load_dma_map_scatterlist(dma_map, va_array);
for (i = 0; i < dma_map->page_count; i++)
{
if (!nv_dma_is_addressable(dma_dev, va_array[i], PAGE_SIZE))
{
NV_DMA_DEV_PRINTF(NV_DBG_ERRORS, dma_dev,
"DMA address not in addressable range of device "
"(0x%llx, 0x%llx-0x%llx)\n",
va_array[i], dma_dev->addressable_range.start,
dma_dev->addressable_range.limit);
nv_dma_unmap_scatterlist(dma_map);
return NV_ERR_INVALID_ADDRESS;
}
}
return NV_OK;
}
static void nv_dma_unmap_scatterlist(nv_dma_map_t *dma_map)
{
nv_unmap_dma_map_scatterlist(dma_map);
nv_destroy_dma_map_scatterlist(dma_map);
}
static void nv_dma_nvlink_addr_compress
(
nv_dma_device_t *dma_dev,
NvU64 *va_array,
NvU64 page_count,
NvBool contig
)
{
#if defined(NVCPU_PPC64LE)
NvU64 addr = 0;
NvU64 i;
/*
* On systems that support NVLink sysmem links, apply the required address
* compression scheme when links are trained. Otherwise check that PCIe and
* NVLink DMA mappings are equivalent as per requirements of Bug 1920398.
*/
if (dma_dev->nvlink)
{
for (i = 0; i < (contig ? 1 : page_count); i++)
{
va_array[i] = nv_compress_nvlink_addr(va_array[i]);
}
return;
}
for (i = 0; i < (contig ? 1 : page_count); i++)
{
addr = nv_compress_nvlink_addr(va_array[i]);
if (WARN_ONCE(va_array[i] != addr,
"unexpected DMA address compression (0x%llx, 0x%llx)\n",
va_array[i], addr))
{
break;
}
}
#endif
}
static void nv_dma_nvlink_addr_decompress
(
nv_dma_device_t *dma_dev,
NvU64 *va_array,
NvU64 page_count,
NvBool contig
)
{
#if defined(NVCPU_PPC64LE)
NvU64 i;
if (dma_dev->nvlink)
{
for (i = 0; i < (contig ? 1 : page_count); i++)
{
va_array[i] = nv_expand_nvlink_addr(va_array[i]);
}
}
#endif
}
NV_STATUS NV_API_CALL nv_dma_map_sgt(
nv_dma_device_t *dma_dev,
NvU64 page_count,
NvU64 *va_array,
NvU32 cache_type,
void **priv
)
{
NV_STATUS status;
nv_dma_map_t *dma_map = NULL;
if (priv == NULL)
{
return NV_ERR_NOT_SUPPORTED;
}
if (page_count > os_get_num_phys_pages())
{
NV_DMA_DEV_PRINTF(NV_DBG_ERRORS, dma_dev,
"DMA mapping request too large!\n");
return NV_ERR_INVALID_REQUEST;
}
status = os_alloc_mem((void **)&dma_map, sizeof(nv_dma_map_t));
if (status != NV_OK)
{
NV_DMA_DEV_PRINTF(NV_DBG_ERRORS, dma_dev,
"Failed to allocate nv_dma_map_t!\n");
return status;
}
dma_map->dev = dma_dev->dev;
dma_map->pages = NULL;
dma_map->import_sgt = (struct sg_table *) *priv;
dma_map->page_count = page_count;
dma_map->contiguous = NV_FALSE;
dma_map->cache_type = cache_type;
dma_map->mapping.discontig.submap_count = 0;
status = nv_dma_map_scatterlist(dma_dev, dma_map, va_array);
if (status != NV_OK)
{
os_free_mem(dma_map);
}
else
{
*priv = dma_map;
nv_dma_nvlink_addr_compress(dma_dev, va_array, dma_map->page_count,
dma_map->contiguous);
}
return status;
}
NV_STATUS NV_API_CALL nv_dma_unmap_sgt(
nv_dma_device_t *dma_dev,
void **priv
)
{
nv_dma_map_t *dma_map;
if (priv == NULL)
{
return NV_ERR_NOT_SUPPORTED;
}
dma_map = *priv;
*priv = NULL;
nv_dma_unmap_scatterlist(dma_map);
os_free_mem(dma_map);
return NV_OK;
}
NV_STATUS NV_API_CALL nv_dma_map_pages(
nv_dma_device_t *dma_dev,
NvU64 page_count,
NvU64 *va_array,
NvBool contig,
NvU32 cache_type,
void **priv
)
{
NV_STATUS status;
nv_dma_map_t *dma_map = NULL;
if (priv == NULL)
{
/*
* IOMMU path has not been implemented yet to handle
* anything except a nv_dma_map_t as the priv argument.
*/
return NV_ERR_NOT_SUPPORTED;
}
if (page_count > os_get_num_phys_pages())
{
NV_DMA_DEV_PRINTF(NV_DBG_ERRORS, dma_dev,
"DMA mapping request too large!\n");
return NV_ERR_INVALID_REQUEST;
}
status = os_alloc_mem((void **)&dma_map, sizeof(nv_dma_map_t));
if (status != NV_OK)
{
NV_DMA_DEV_PRINTF(NV_DBG_ERRORS, dma_dev,
"Failed to allocate nv_dma_map_t!\n");
return status;
}
dma_map->dev = dma_dev->dev;
dma_map->pages = *priv;
dma_map->import_sgt = NULL;
dma_map->page_count = page_count;
dma_map->contiguous = contig;
dma_map->cache_type = cache_type;
if (dma_map->page_count > 1 && !dma_map->contiguous)
{
dma_map->mapping.discontig.submap_count = 0;
status = nv_dma_map_scatterlist(dma_dev, dma_map, va_array);
}
else
{
/*
* Force single-page mappings to be contiguous to avoid scatterlist
* overhead.
*/
dma_map->contiguous = NV_TRUE;
status = nv_dma_map_contig(dma_dev, dma_map, va_array);
}
if (status != NV_OK)
{
os_free_mem(dma_map);
}
else
{
*priv = dma_map;
nv_dma_nvlink_addr_compress(dma_dev, va_array, dma_map->page_count,
dma_map->contiguous);
}
return status;
}
NV_STATUS NV_API_CALL nv_dma_unmap_pages(
nv_dma_device_t *dma_dev,
NvU64 page_count,
NvU64 *va_array,
void **priv
)
{
nv_dma_map_t *dma_map;
if (priv == NULL)
{
/*
* IOMMU path has not been implemented yet to handle
* anything except a nv_dma_map_t as the priv argument.
*/
return NV_ERR_NOT_SUPPORTED;
}
dma_map = *priv;
if (page_count > os_get_num_phys_pages())
{
NV_DMA_DEV_PRINTF(NV_DBG_ERRORS, dma_dev,
"DMA unmapping request too large!\n");
return NV_ERR_INVALID_REQUEST;
}
if (page_count != dma_map->page_count)
{
NV_DMA_DEV_PRINTF(NV_DBG_WARNINGS, dma_dev,
"Requested to DMA unmap %llu pages, but there are %llu "
"in the mapping\n", page_count, dma_map->page_count);
return NV_ERR_INVALID_REQUEST;
}
*priv = dma_map->pages;
if (dma_map->contiguous)
{
nv_dma_unmap_contig(dma_map);
}
else
{
nv_dma_unmap_scatterlist(dma_map);
}
os_free_mem(dma_map);
return NV_OK;
}
/*
* Wrappers used for DMA-remapping an nv_alloc_t during transition to more
* generic interfaces.
*/
NV_STATUS NV_API_CALL nv_dma_map_alloc
(
nv_dma_device_t *dma_dev,
NvU64 page_count,
NvU64 *va_array,
NvBool contig,
void **priv
)
{
NV_STATUS status;
NvU64 i;
nv_alloc_t *at = *priv;
struct page **pages = NULL;
NvU32 cache_type = NV_MEMORY_CACHED;
NvU64 pages_size = sizeof(struct page *) * (contig ? 1 : page_count);
/* If we have an imported SGT, just use that directly. */
if (at && at->import_sgt)
{
*priv = at->import_sgt;
status = nv_dma_map_sgt(dma_dev, page_count, va_array, at->cache_type,
priv);
if (status != NV_OK)
{
*priv = at;
}
return status;
}
/*
* Convert the nv_alloc_t into a struct page * array for
* nv_dma_map_pages().
*/
status = os_alloc_mem((void **)&pages, pages_size);
if (status != NV_OK)
{
NV_DMA_DEV_PRINTF(NV_DBG_ERRORS, dma_dev,
"Failed to allocate page array for DMA mapping!\n");
return status;
}
os_mem_set(pages, 0, pages_size);
if (at != NULL)
{
WARN_ON(page_count != at->num_pages);
if (at->flags.user)
{
pages[0] = at->user_pages[0];
if (!contig)
{
for (i = 1; i < page_count; i++)
{
pages[i] = at->user_pages[i];
}
}
}
else if (at->flags.physical && contig)
{
/* Supplied pages hold physical address */
pages[0] = pfn_to_page(PFN_DOWN(va_array[0]));
}
cache_type = at->cache_type;
}
if (pages[0] == NULL)
{
pages[0] = NV_GET_PAGE_STRUCT(va_array[0]);
if (!contig)
{
for (i = 1; i < page_count; i++)
{
pages[i] = NV_GET_PAGE_STRUCT(va_array[i]);
}
}
}
*priv = pages;
status = nv_dma_map_pages(dma_dev, page_count, va_array, contig, cache_type,
priv);
if (status != NV_OK)
{
*priv = at;
os_free_mem(pages);
}
return status;
}
NV_STATUS NV_API_CALL nv_dma_unmap_alloc
(
nv_dma_device_t *dma_dev,
NvU64 page_count,
NvU64 *va_array,
void **priv
)
{
NV_STATUS status = NV_OK;
nv_dma_map_t *dma_map;
if (priv == NULL)
{
return NV_ERR_NOT_SUPPORTED;
}
dma_map = *priv;
if (!dma_map->import_sgt)
{
status = nv_dma_unmap_pages(dma_dev, page_count, va_array, priv);
if (status != NV_OK)
{
/*
* If nv_dma_unmap_pages() fails, we hit an assert condition and the
* priv argument won't be the page array we allocated in
* nv_dma_map_alloc(), so we skip the free here. But note that since
* this is an assert condition it really should never happen.
*/
return status;
}
/* Free the struct page * array allocated by nv_dma_map_alloc() */
os_free_mem(*priv);
} else {
status = nv_dma_unmap_sgt(dma_dev, priv);
}
return status;
}
static NvBool nv_dma_use_map_resource
(
nv_dma_device_t *dma_dev
)
{
if (nv_dma_remap_peer_mmio == NV_DMA_REMAP_PEER_MMIO_DISABLE)
{
return NV_FALSE;
}
#if defined(NV_DMA_MAP_RESOURCE_PRESENT)
const struct dma_map_ops *ops = get_dma_ops(dma_dev->dev);
if (ops == NULL)
{
/* On pre-5.0 kernels, if dma_map_resource() is present, then we
* assume that ops != NULL. With direct_dma handling swiotlb on 5.0+
* kernels, ops == NULL.
*/
#if defined(NV_DMA_IS_DIRECT_PRESENT)
return NV_TRUE;
#else
return NV_FALSE;
#endif
}
return (ops->map_resource != NULL);
#else
return NV_FALSE;
#endif
}
/* DMA-map a peer PCI device's BAR for peer access. */
NV_STATUS NV_API_CALL nv_dma_map_peer
(
nv_dma_device_t *dma_dev,
nv_dma_device_t *peer_dma_dev,
NvU8 bar_index,
NvU64 page_count,
NvU64 *va
)
{
struct pci_dev *peer_pci_dev = to_pci_dev(peer_dma_dev->dev);
struct resource *res;
NV_STATUS status;
if (peer_pci_dev == NULL)
{
NV_DMA_DEV_PRINTF(NV_DBG_ERRORS, peer_dma_dev,
"Not a PCI device");
return NV_ERR_INVALID_REQUEST;
}
BUG_ON(bar_index >= NV_GPU_NUM_BARS);
res = &peer_pci_dev->resource[bar_index];
if (res->start == 0)
{
NV_DMA_DEV_PRINTF(NV_DBG_ERRORS, peer_dma_dev,
"Resource %u not valid",
bar_index);
return NV_ERR_INVALID_REQUEST;
}
if ((*va < res->start) || ((*va + (page_count * PAGE_SIZE)) > res->end))
{
NV_DMA_DEV_PRINTF(NV_DBG_ERRORS, peer_dma_dev,
"Mapping requested (start = 0x%llx, page_count = 0x%llx)"
" outside of resource bounds (start = 0x%llx, end = 0x%llx)\n",
*va, page_count, res->start, res->end);
return NV_ERR_INVALID_REQUEST;
}
if (nv_dma_use_map_resource(dma_dev))
{
status = nv_dma_map_mmio(dma_dev, page_count, va);
}
else
{
/*
* Best effort - can't map through the iommu but at least try to
* convert to a bus address.
*/
NvU64 offset = *va - res->start;
*va = nv_pci_bus_address(peer_pci_dev, bar_index) + offset;
status = NV_OK;
}
return status;
}
void NV_API_CALL nv_dma_unmap_peer
(
nv_dma_device_t *dma_dev,
NvU64 page_count,
NvU64 va
)
{
if (nv_dma_use_map_resource(dma_dev))
{
nv_dma_unmap_mmio(dma_dev, page_count, va);
}
}
/* DMA-map another anonymous device's MMIO region for peer access. */
NV_STATUS NV_API_CALL nv_dma_map_mmio
(
nv_dma_device_t *dma_dev,
NvU64 page_count,
NvU64 *va
)
{
#if defined(NV_DMA_MAP_RESOURCE_PRESENT)
BUG_ON(!va);
if (nv_dma_use_map_resource(dma_dev))
{
NvU64 mmio_addr = *va;
*va = dma_map_resource(dma_dev->dev, mmio_addr, page_count * PAGE_SIZE,
DMA_BIDIRECTIONAL, 0);
if (dma_mapping_error(dma_dev->dev, *va))
{
NV_DMA_DEV_PRINTF(NV_DBG_ERRORS, dma_dev,
"Failed to DMA map MMIO range [0x%llx-0x%llx]\n",
mmio_addr, mmio_addr + page_count * PAGE_SIZE - 1);
return NV_ERR_OPERATING_SYSTEM;
}
}
else
{
/*
* If dma_map_resource is not available, pass through the source address
* without failing. Further, adjust it using the DMA start address to
* keep RM's validation schemes happy.
*/
*va = *va + dma_dev->addressable_range.start;
}
nv_dma_nvlink_addr_compress(dma_dev, va, page_count, NV_TRUE);
return NV_OK;
#else
return NV_ERR_NOT_SUPPORTED;
#endif
}
void NV_API_CALL nv_dma_unmap_mmio
(
nv_dma_device_t *dma_dev,
NvU64 page_count,
NvU64 va
)
{
#if defined(NV_DMA_MAP_RESOURCE_PRESENT)
nv_dma_nvlink_addr_decompress(dma_dev, &va, page_count, NV_TRUE);
if (nv_dma_use_map_resource(dma_dev))
{
dma_unmap_resource(dma_dev->dev, va, page_count * PAGE_SIZE,
DMA_BIDIRECTIONAL, 0);
}
#endif
}
/*
* Invalidate DMA mapping in CPU caches by "syncing" to the device.
*
* This is only implemented for ARM platforms, since other supported
* platforms are cache coherent and have not required this (we
* explicitly haven't supported SWIOTLB bounce buffering either where
* this would be needed).
*/
void NV_API_CALL nv_dma_cache_invalidate
(
nv_dma_device_t *dma_dev,
void *priv
)
{
#if defined(NVCPU_AARCH64)
nv_dma_map_t *dma_map = priv;
if (dma_map->contiguous)
{
dma_sync_single_for_device(dma_dev->dev,
dma_map->mapping.contig.dma_addr,
(size_t) PAGE_SIZE * dma_map->page_count,
DMA_FROM_DEVICE);
}
else
{
nv_dma_submap_t *submap;
NvU64 i;
NV_FOR_EACH_DMA_SUBMAP(dma_map, submap, i)
{
dma_sync_sg_for_device(dma_dev->dev,
submap->sgt.sgl,
submap->sgt.orig_nents,
DMA_FROM_DEVICE);
}
}
#endif
}
/* Enable DMA-mapping over NVLink */
void NV_API_CALL nv_dma_enable_nvlink
(
nv_dma_device_t *dma_dev
)
{
dma_dev->nvlink = NV_TRUE;
}
#if defined(NV_LINUX_DMA_BUF_H_PRESENT) && \
defined(NV_DRM_AVAILABLE) && defined(NV_DRM_DRM_GEM_H_PRESENT)
/*
* drm_gem_object_{get/put}() added by commit
* e6b62714e87c8811d5564b6a0738dcde63a51774 (2017-02-28) and
* drm_gem_object_{reference/unreference}() removed by commit
* 3e70fd160cf0b1945225eaa08dd2cb8544f21cb8 (2018-11-15).
*/
static inline void
nv_dma_gem_object_unreference_unlocked(struct drm_gem_object *gem)
{
#if defined(NV_DRM_GEM_OBJECT_GET_PRESENT)
#if defined(NV_DRM_GEM_OBJECT_PUT_UNLOCK_PRESENT)
drm_gem_object_put_unlocked(gem);
#else
drm_gem_object_put(gem);
#endif
#else
drm_gem_object_unreference_unlocked(gem);
#endif
}
static inline void
nv_dma_gem_object_reference(struct drm_gem_object *gem)
{
#if defined(NV_DRM_GEM_OBJECT_GET_PRESENT)
drm_gem_object_get(gem);
#else
drm_gem_object_reference(gem);
#endif
}
NV_STATUS NV_API_CALL nv_dma_import_sgt
(
nv_dma_device_t *dma_dev,
struct sg_table *sgt,
struct drm_gem_object *gem
)
{
if ((dma_dev == NULL) ||
(sgt == NULL) ||
(gem == NULL))
{
NV_DMA_DEV_PRINTF(NV_DBG_ERRORS, dma_dev,
"Import arguments are NULL!\n");
return NV_ERR_INVALID_ARGUMENT;
}
// Prevent the kernel module controlling GEM from being unloaded
if (!try_module_get(gem->dev->driver->fops->owner))
{
NV_DMA_DEV_PRINTF(NV_DBG_ERRORS, dma_dev,
"Couldn't reference the GEM object's owner!\n");
return NV_ERR_INVALID_DEVICE;
}
// Do nothing with SGT, it is already mapped and pinned by the exporter
nv_dma_gem_object_reference(gem);
return NV_OK;
}
void NV_API_CALL nv_dma_release_sgt
(
struct sg_table *sgt,
struct drm_gem_object *gem
)
{
if (gem == NULL)
{
return;
}
// Do nothing with SGT, it will be unmapped and unpinned by the exporter
WARN_ON(sgt == NULL);
nv_dma_gem_object_unreference_unlocked(gem);
module_put(gem->dev->driver->fops->owner);
}
#else
NV_STATUS NV_API_CALL nv_dma_import_sgt
(
nv_dma_device_t *dma_dev,
struct sg_table *sgt,
struct drm_gem_object *gem
)
{
return NV_ERR_NOT_SUPPORTED;
}
void NV_API_CALL nv_dma_release_sgt
(
struct sg_table *sgt,
struct drm_gem_object *gem
)
{
}
#endif /* NV_LINUX_DMA_BUF_H_PRESENT && NV_DRM_AVAILABLE && NV_DRM_DRM_GEM_H_PRESENT */
#if defined(NV_LINUX_DMA_BUF_H_PRESENT)
#endif /* NV_LINUX_DMA_BUF_H_PRESENT */
#ifndef IMPORT_DMABUF_FUNCTIONS_DEFINED
NV_STATUS NV_API_CALL nv_dma_import_dma_buf
(
nv_dma_device_t *dma_dev,
struct dma_buf *dma_buf,
NvU32 *size,
void **user_pages,
struct sg_table **sgt,
nv_dma_buf_t **import_priv
)
{
return NV_ERR_NOT_SUPPORTED;
}
NV_STATUS NV_API_CALL nv_dma_import_from_fd
(
nv_dma_device_t *dma_dev,
NvS32 fd,
NvU32 *size,
void **user_pages,
struct sg_table **sgt,
nv_dma_buf_t **import_priv
)
{
return NV_ERR_NOT_SUPPORTED;
}
void NV_API_CALL nv_dma_release_dma_buf
(
void *user_pages,
nv_dma_buf_t *import_priv
)
{
}
#endif /* !IMPORT_DMABUF_FUNCTIONS_DEFINED */