/* * Copyright (c) 2020, NVIDIA CORPORATION. All rights reserved. * * Permission is hereby granted, free of charge, to any person obtaining a * copy of this software and associated documentation files (the "Software"), * to deal in the Software without restriction, including without limitation * the rights to use, copy, modify, merge, publish, distribute, sublicense, * and/or sell copies of the Software, and to permit persons to whom the * Software is furnished to do so, subject to the following conditions: * * The above copyright notice and this permission notice shall be included in * all copies or substantial portions of the Software. * * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING * FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER * DEALINGS IN THE SOFTWARE. */ #include "nvidia-drm-conftest.h" #if defined(NV_DRM_AVAILABLE) #if defined(NV_DRM_DRM_PRIME_H_PRESENT) #include #endif #if defined(NV_DRM_DRMP_H_PRESENT) #include #endif #if defined(NV_DRM_DRM_DRV_H_PRESENT) #include #endif #include "nvidia-drm-gem-dma-buf.h" #include "nvidia-drm-ioctl.h" #include "linux/dma-buf.h" static inline void __nv_drm_gem_dma_buf_free(struct nv_drm_gem_object *nv_gem) { struct nv_drm_device *nv_dev = nv_gem->nv_dev; struct nv_drm_gem_dma_buf *nv_dma_buf = to_nv_dma_buf(nv_gem); #if defined(NV_DRM_ATOMIC_MODESET_AVAILABLE) if (nv_dma_buf->base.pMemory) { /* Free NvKmsKapiMemory handle associated with this gem object */ nvKms->freeMemory(nv_dev->pDevice, nv_dma_buf->base.pMemory); } #endif drm_prime_gem_destroy(&nv_gem->base, nv_dma_buf->sgt); nv_drm_free(nv_dma_buf); } static int __nv_drm_gem_dma_buf_create_mmap_offset( struct nv_drm_device *nv_dev, struct nv_drm_gem_object *nv_gem, uint64_t *offset) { (void)nv_dev; return nv_drm_gem_create_mmap_offset(nv_gem, offset); } static int __nv_drm_gem_dma_buf_mmap(struct nv_drm_gem_object *nv_gem, struct vm_area_struct *vma) { struct dma_buf_attachment *attach = nv_gem->base.import_attach; struct dma_buf *dma_buf = attach->dmabuf; struct file *old_file; int ret; /* check if buffer supports mmap */ if (!dma_buf->file->f_op->mmap) return -EINVAL; /* readjust the vma */ get_file(dma_buf->file); old_file = vma->vm_file; vma->vm_file = dma_buf->file; vma->vm_pgoff -= drm_vma_node_start(&nv_gem->base.vma_node);; ret = dma_buf->file->f_op->mmap(dma_buf->file, vma); if (ret) { /* restore old parameters on failure */ vma->vm_file = old_file; fput(dma_buf->file); } else { if (old_file) fput(old_file); } return ret; } const struct nv_drm_gem_object_funcs __nv_gem_dma_buf_ops = { .free = __nv_drm_gem_dma_buf_free, .create_mmap_offset = __nv_drm_gem_dma_buf_create_mmap_offset, .mmap = __nv_drm_gem_dma_buf_mmap, }; struct drm_gem_object* nv_drm_gem_prime_import_sg_table(struct drm_device *dev, struct dma_buf_attachment *attach, struct sg_table *sgt) { struct nv_drm_device *nv_dev = to_nv_device(dev); struct dma_buf *dma_buf = attach->dmabuf; struct nv_drm_gem_dma_buf *nv_dma_buf; struct NvKmsKapiMemory *pMemory; if ((nv_dma_buf = nv_drm_calloc(1, sizeof(*nv_dma_buf))) == NULL) { return NULL; } // dma_buf->size must be a multiple of PAGE_SIZE BUG_ON(dma_buf->size % PAGE_SIZE); pMemory = NULL; #if defined(NV_DRM_ATOMIC_MODESET_AVAILABLE) if (drm_core_check_feature(dev, DRIVER_MODESET)) { pMemory = nvKms->getSystemMemoryHandleFromDmaBuf(nv_dev->pDevice, (NvP64)(NvUPtr)dma_buf, dma_buf->size - 1); } #endif nv_drm_gem_object_init(nv_dev, &nv_dma_buf->base, &__nv_gem_dma_buf_ops, dma_buf->size, pMemory); nv_dma_buf->sgt = sgt; return &nv_dma_buf->base.base; } int nv_drm_gem_export_dmabuf_memory_ioctl(struct drm_device *dev, void *data, struct drm_file *filep) { struct nv_drm_device *nv_dev = to_nv_device(dev); struct drm_nvidia_gem_export_dmabuf_memory_params *p = data; struct nv_drm_gem_dma_buf *nv_dma_buf = NULL; int ret = 0; struct NvKmsKapiMemory *pTmpMemory = NULL; if (!drm_core_check_feature(dev, DRIVER_MODESET)) { ret = -EINVAL; goto done; } if (p->__pad != 0) { ret = -EINVAL; NV_DRM_DEV_LOG_ERR(nv_dev, "Padding fields must be zeroed"); goto done; } if ((nv_dma_buf = nv_drm_gem_object_dma_buf_lookup( dev, filep, p->handle)) == NULL) { ret = -EINVAL; NV_DRM_DEV_LOG_ERR( nv_dev, "Failed to lookup DMA-BUF GEM object for export: 0x%08x", p->handle); goto done; } #if defined(NV_DRM_ATOMIC_MODESET_AVAILABLE) if (drm_core_check_feature(dev, DRIVER_MODESET)) { if (!nv_dma_buf->base.pMemory) { /* * Get RM system memory handle from SGT - RM will take a reference * on this GEM object to prevent the DMA-BUF from being unpinned * prematurely. */ pTmpMemory = nvKms->getSystemMemoryHandleFromSgt( nv_dev->pDevice, (NvP64)(NvUPtr)nv_dma_buf->sgt, (NvP64)(NvUPtr)&nv_dma_buf->base.base, nv_dma_buf->base.base.size - 1); } } #endif if (!nv_dma_buf->base.pMemory && !pTmpMemory) { ret = -ENOMEM; NV_DRM_DEV_LOG_ERR( nv_dev, "Failed to get memory to export from DMA-BUF GEM object: 0x%08x", p->handle); goto done; } if (!nvKms->exportMemory(nv_dev->pDevice, nv_dma_buf->base.pMemory ? nv_dma_buf->base.pMemory : pTmpMemory, p->nvkms_params_ptr, p->nvkms_params_size)) { ret = -EINVAL; NV_DRM_DEV_LOG_ERR( nv_dev, "Failed to export memory from DMA-BUF GEM object: 0x%08x", p->handle); goto done; } done: if (pTmpMemory) { /* * Release reference on RM system memory to prevent circular * refcounting. Another refcount will still be held by RM FD. */ nvKms->freeMemory(nv_dev->pDevice, pTmpMemory); } if (nv_dma_buf != NULL) { nv_drm_gem_object_unreference_unlocked(&nv_dma_buf->base); } return ret; } #endif