[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <13b80e3f-b0c4-4517-bf8b-c2fc5632751f@quicinc.com>
Date: Wed, 16 Apr 2025 22:48:50 +0530
From: Akhil P Oommen <quic_akhilpo@...cinc.com>
To: Rob Clark <robdclark@...il.com>, <dri-devel@...ts.freedesktop.org>
CC: <freedreno@...ts.freedesktop.org>, <linux-arm-msm@...r.kernel.org>,
"Rob
Clark" <robdclark@...omium.org>, Sean Paul <sean@...rly.run>,
Konrad Dybcio
<konradybcio@...nel.org>,
Abhinav Kumar <quic_abhinavk@...cinc.com>,
"Dmitry
Baryshkov" <lumag@...nel.org>,
Marijn Suijten
<marijn.suijten@...ainline.org>,
David Airlie <airlied@...il.com>, Simona
Vetter <simona@...ll.ch>,
open list <linux-kernel@...r.kernel.org>
Subject: Re: [PATCH v2 08/34] drm/msm: Remove vram carveout support
On 3/19/2025 8:22 PM, Rob Clark wrote:
> From: Rob Clark <robdclark@...omium.org>
>
> It is standing in the way of drm_gpuvm / VM_BIND support. Not to
> mention frequently broken and rarely tested. And I think only needed
> for a 10yr old not quite upstream SoC (msm8974).
>
> Maybe we can add support back in later, but I'm doubtful.
There is a small clean up of "VRAM carveout" required in a debug print
in msm_gpu.c
>
> Signed-off-by: Rob Clark <robdclark@...omium.org>
> ---
> drivers/gpu/drm/msm/adreno/a2xx_gpu.c | 6 +-
> drivers/gpu/drm/msm/adreno/a3xx_gpu.c | 13 +-
> drivers/gpu/drm/msm/adreno/a4xx_gpu.c | 13 +-
> drivers/gpu/drm/msm/adreno/adreno_device.c | 4 -
> drivers/gpu/drm/msm/adreno/adreno_gpu.h | 1 -
> drivers/gpu/drm/msm/msm_drv.c | 117 +-----------------
> drivers/gpu/drm/msm/msm_drv.h | 11 --
> drivers/gpu/drm/msm/msm_gem.c | 131 ++-------------------
> drivers/gpu/drm/msm/msm_gem.h | 5 -
> drivers/gpu/drm/msm/msm_gem_submit.c | 5 -
> 10 files changed, 19 insertions(+), 287 deletions(-)
>
> diff --git a/drivers/gpu/drm/msm/adreno/a2xx_gpu.c b/drivers/gpu/drm/msm/adreno/a2xx_gpu.c
> index 5eb063ed0b46..db1aa281ce47 100644
> --- a/drivers/gpu/drm/msm/adreno/a2xx_gpu.c
> +++ b/drivers/gpu/drm/msm/adreno/a2xx_gpu.c
> @@ -553,10 +553,8 @@ struct msm_gpu *a2xx_gpu_init(struct drm_device *dev)
>
> if (!gpu->vm) {
> dev_err(dev->dev, "No memory protection without MMU\n");
> - if (!allow_vram_carveout) {
> - ret = -ENXIO;
> - goto fail;
> - }
> + ret = -ENXIO;
> + goto fail;
> }
>
> return gpu;
> diff --git a/drivers/gpu/drm/msm/adreno/a3xx_gpu.c b/drivers/gpu/drm/msm/adreno/a3xx_gpu.c
> index 434e6ededf83..49ba1ce77144 100644
> --- a/drivers/gpu/drm/msm/adreno/a3xx_gpu.c
> +++ b/drivers/gpu/drm/msm/adreno/a3xx_gpu.c
> @@ -582,18 +582,9 @@ struct msm_gpu *a3xx_gpu_init(struct drm_device *dev)
> }
>
> if (!gpu->vm) {
> - /* TODO we think it is possible to configure the GPU to
> - * restrict access to VRAM carveout. But the required
> - * registers are unknown. For now just bail out and
> - * limp along with just modesetting. If it turns out
> - * to not be possible to restrict access, then we must
> - * implement a cmdstream validator.
> - */
> DRM_DEV_ERROR(dev->dev, "No memory protection without IOMMU\n");
> - if (!allow_vram_carveout) {
> - ret = -ENXIO;
> - goto fail;
> - }
> + ret = -ENXIO;
> + goto fail;
> }
>
> icc_path = devm_of_icc_get(&pdev->dev, "gfx-mem");
> diff --git a/drivers/gpu/drm/msm/adreno/a4xx_gpu.c b/drivers/gpu/drm/msm/adreno/a4xx_gpu.c
> index 2c75debcfd84..4faf8570aec7 100644
> --- a/drivers/gpu/drm/msm/adreno/a4xx_gpu.c
> +++ b/drivers/gpu/drm/msm/adreno/a4xx_gpu.c
> @@ -696,18 +696,9 @@ struct msm_gpu *a4xx_gpu_init(struct drm_device *dev)
> adreno_gpu->uche_trap_base = 0xffff0000ffff0000ull;
>
> if (!gpu->vm) {
Is this check still required?
-Akhil
> - /* TODO we think it is possible to configure the GPU to
> - * restrict access to VRAM carveout. But the required
> - * registers are unknown. For now just bail out and
> - * limp along with just modesetting. If it turns out
> - * to not be possible to restrict access, then we must
> - * implement a cmdstream validator.
> - */
> DRM_DEV_ERROR(dev->dev, "No memory protection without IOMMU\n");
> - if (!allow_vram_carveout) {
> - ret = -ENXIO;
> - goto fail;
> - }
> + ret = -ENXIO;
> + goto fail;
> }
>
> icc_path = devm_of_icc_get(&pdev->dev, "gfx-mem");
> diff --git a/drivers/gpu/drm/msm/adreno/adreno_device.c b/drivers/gpu/drm/msm/adreno/adreno_device.c
> index f4552b8c6767..6b0390c38bff 100644
> --- a/drivers/gpu/drm/msm/adreno/adreno_device.c
> +++ b/drivers/gpu/drm/msm/adreno/adreno_device.c
> @@ -16,10 +16,6 @@ bool snapshot_debugbus = false;
> MODULE_PARM_DESC(snapshot_debugbus, "Include debugbus sections in GPU devcoredump (if not fused off)");
> module_param_named(snapshot_debugbus, snapshot_debugbus, bool, 0600);
>
> -bool allow_vram_carveout = false;
> -MODULE_PARM_DESC(allow_vram_carveout, "Allow using VRAM Carveout, in place of IOMMU");
> -module_param_named(allow_vram_carveout, allow_vram_carveout, bool, 0600);
> -
> int enable_preemption = -1;
> MODULE_PARM_DESC(enable_preemption, "Enable preemption (A7xx only) (1=on , 0=disable, -1=auto (default))");
> module_param(enable_preemption, int, 0600);
> diff --git a/drivers/gpu/drm/msm/adreno/adreno_gpu.h b/drivers/gpu/drm/msm/adreno/adreno_gpu.h
> index eaebcb108b5e..7dbe09817edc 100644
> --- a/drivers/gpu/drm/msm/adreno/adreno_gpu.h
> +++ b/drivers/gpu/drm/msm/adreno/adreno_gpu.h
> @@ -18,7 +18,6 @@
> #include "adreno_pm4.xml.h"
>
> extern bool snapshot_debugbus;
> -extern bool allow_vram_carveout;
>
> enum {
> ADRENO_FW_PM4 = 0,
> diff --git a/drivers/gpu/drm/msm/msm_drv.c b/drivers/gpu/drm/msm/msm_drv.c
> index 903abf3532e0..978f1d355b42 100644
> --- a/drivers/gpu/drm/msm/msm_drv.c
> +++ b/drivers/gpu/drm/msm/msm_drv.c
> @@ -46,12 +46,6 @@
> #define MSM_VERSION_MINOR 12
> #define MSM_VERSION_PATCHLEVEL 0
>
> -static void msm_deinit_vram(struct drm_device *ddev);
> -
> -static char *vram = "16m";
> -MODULE_PARM_DESC(vram, "Configure VRAM size (for devices without IOMMU/GPUMMU)");
> -module_param(vram, charp, 0);
> -
> bool dumpstate;
> MODULE_PARM_DESC(dumpstate, "Dump KMS state on errors");
> module_param(dumpstate, bool, 0600);
> @@ -97,8 +91,6 @@ static int msm_drm_uninit(struct device *dev)
> if (priv->kms)
> msm_drm_kms_uninit(dev);
>
> - msm_deinit_vram(ddev);
> -
> component_unbind_all(dev, ddev);
>
> ddev->dev_private = NULL;
> @@ -109,107 +101,6 @@ static int msm_drm_uninit(struct device *dev)
> return 0;
> }
>
> -bool msm_use_mmu(struct drm_device *dev)
> -{
> - struct msm_drm_private *priv = dev->dev_private;
> -
> - /*
> - * a2xx comes with its own MMU
> - * On other platforms IOMMU can be declared specified either for the
> - * MDP/DPU device or for its parent, MDSS device.
> - */
> - return priv->is_a2xx ||
> - device_iommu_mapped(dev->dev) ||
> - device_iommu_mapped(dev->dev->parent);
> -}
> -
> -static int msm_init_vram(struct drm_device *dev)
> -{
> - struct msm_drm_private *priv = dev->dev_private;
> - struct device_node *node;
> - unsigned long size = 0;
> - int ret = 0;
> -
> - /* In the device-tree world, we could have a 'memory-region'
> - * phandle, which gives us a link to our "vram". Allocating
> - * is all nicely abstracted behind the dma api, but we need
> - * to know the entire size to allocate it all in one go. There
> - * are two cases:
> - * 1) device with no IOMMU, in which case we need exclusive
> - * access to a VRAM carveout big enough for all gpu
> - * buffers
> - * 2) device with IOMMU, but where the bootloader puts up
> - * a splash screen. In this case, the VRAM carveout
> - * need only be large enough for fbdev fb. But we need
> - * exclusive access to the buffer to avoid the kernel
> - * using those pages for other purposes (which appears
> - * as corruption on screen before we have a chance to
> - * load and do initial modeset)
> - */
> -
> - node = of_parse_phandle(dev->dev->of_node, "memory-region", 0);
> - if (node) {
> - struct resource r;
> - ret = of_address_to_resource(node, 0, &r);
> - of_node_put(node);
> - if (ret)
> - return ret;
> - size = r.end - r.start + 1;
> - DRM_INFO("using VRAM carveout: %lx@%pa\n", size, &r.start);
> -
> - /* if we have no IOMMU, then we need to use carveout allocator.
> - * Grab the entire DMA chunk carved out in early startup in
> - * mach-msm:
> - */
> - } else if (!msm_use_mmu(dev)) {
> - DRM_INFO("using %s VRAM carveout\n", vram);
> - size = memparse(vram, NULL);
> - }
> -
> - if (size) {
> - unsigned long attrs = 0;
> - void *p;
> -
> - priv->vram.size = size;
> -
> - drm_mm_init(&priv->vram.mm, 0, (size >> PAGE_SHIFT) - 1);
> - spin_lock_init(&priv->vram.lock);
> -
> - attrs |= DMA_ATTR_NO_KERNEL_MAPPING;
> - attrs |= DMA_ATTR_WRITE_COMBINE;
> -
> - /* note that for no-kernel-mapping, the vaddr returned
> - * is bogus, but non-null if allocation succeeded:
> - */
> - p = dma_alloc_attrs(dev->dev, size,
> - &priv->vram.paddr, GFP_KERNEL, attrs);
> - if (!p) {
> - DRM_DEV_ERROR(dev->dev, "failed to allocate VRAM\n");
> - priv->vram.paddr = 0;
> - return -ENOMEM;
> - }
> -
> - DRM_DEV_INFO(dev->dev, "VRAM: %08x->%08x\n",
> - (uint32_t)priv->vram.paddr,
> - (uint32_t)(priv->vram.paddr + size));
> - }
> -
> - return ret;
> -}
> -
> -static void msm_deinit_vram(struct drm_device *ddev)
> -{
> - struct msm_drm_private *priv = ddev->dev_private;
> - unsigned long attrs = DMA_ATTR_NO_KERNEL_MAPPING;
> -
> - if (!priv->vram.paddr)
> - return;
> -
> - drm_mm_takedown(&priv->vram.mm);
> - dma_free_attrs(ddev->dev, priv->vram.size, NULL, priv->vram.paddr,
> - attrs);
> -}
> -
> static int msm_drm_init(struct device *dev, const struct drm_driver *drv)
> {
> struct msm_drm_private *priv = dev_get_drvdata(dev);
> @@ -256,16 +147,12 @@ static int msm_drm_init(struct device *dev, const struct drm_driver *drv)
> goto err_destroy_wq;
> }
>
> - ret = msm_init_vram(ddev);
> - if (ret)
> - goto err_destroy_wq;
> -
> dma_set_max_seg_size(dev, UINT_MAX);
>
> /* Bind all our sub-components: */
> ret = component_bind_all(dev, ddev);
> if (ret)
> - goto err_deinit_vram;
> + goto err_destroy_wq;
>
> ret = msm_gem_shrinker_init(ddev);
> if (ret)
> @@ -302,8 +189,6 @@ static int msm_drm_init(struct device *dev, const struct drm_driver *drv)
>
> return ret;
>
> -err_deinit_vram:
> - msm_deinit_vram(ddev);
> err_destroy_wq:
> destroy_workqueue(priv->wq);
> err_put_dev:
> diff --git a/drivers/gpu/drm/msm/msm_drv.h b/drivers/gpu/drm/msm/msm_drv.h
> index 0e675c9a7f83..ad509403f072 100644
> --- a/drivers/gpu/drm/msm/msm_drv.h
> +++ b/drivers/gpu/drm/msm/msm_drv.h
> @@ -183,17 +183,6 @@ struct msm_drm_private {
>
> struct msm_drm_thread event_thread[MAX_CRTCS];
>
> - /* VRAM carveout, used when no IOMMU: */
> - struct {
> - unsigned long size;
> - dma_addr_t paddr;
> - /* NOTE: mm managed at the page level, size is in # of pages
> - * and position mm_node->start is in # of pages:
> - */
> - struct drm_mm mm;
> - spinlock_t lock; /* Protects drm_mm node allocation/removal */
> - } vram;
> -
> struct notifier_block vmap_notifier;
> struct shrinker *shrinker;
>
> diff --git a/drivers/gpu/drm/msm/msm_gem.c b/drivers/gpu/drm/msm/msm_gem.c
> index 07a30d29248c..621fb4e17a2e 100644
> --- a/drivers/gpu/drm/msm/msm_gem.c
> +++ b/drivers/gpu/drm/msm/msm_gem.c
> @@ -17,24 +17,8 @@
> #include <trace/events/gpu_mem.h>
>
> #include "msm_drv.h"
> -#include "msm_fence.h"
> #include "msm_gem.h"
> #include "msm_gpu.h"
> -#include "msm_mmu.h"
> -
> -static dma_addr_t physaddr(struct drm_gem_object *obj)
> -{
> - struct msm_gem_object *msm_obj = to_msm_bo(obj);
> - struct msm_drm_private *priv = obj->dev->dev_private;
> - return (((dma_addr_t)msm_obj->vram_node->start) << PAGE_SHIFT) +
> - priv->vram.paddr;
> -}
> -
> -static bool use_pages(struct drm_gem_object *obj)
> -{
> - struct msm_gem_object *msm_obj = to_msm_bo(obj);
> - return !msm_obj->vram_node;
> -}
>
> static int pgprot = 0;
> module_param(pgprot, int, 0600);
> @@ -139,36 +123,6 @@ static void update_lru(struct drm_gem_object *obj)
> mutex_unlock(&priv->lru.lock);
> }
>
> -/* allocate pages from VRAM carveout, used when no IOMMU: */
> -static struct page **get_pages_vram(struct drm_gem_object *obj, int npages)
> -{
> - struct msm_gem_object *msm_obj = to_msm_bo(obj);
> - struct msm_drm_private *priv = obj->dev->dev_private;
> - dma_addr_t paddr;
> - struct page **p;
> - int ret, i;
> -
> - p = kvmalloc_array(npages, sizeof(struct page *), GFP_KERNEL);
> - if (!p)
> - return ERR_PTR(-ENOMEM);
> -
> - spin_lock(&priv->vram.lock);
> - ret = drm_mm_insert_node(&priv->vram.mm, msm_obj->vram_node, npages);
> - spin_unlock(&priv->vram.lock);
> - if (ret) {
> - kvfree(p);
> - return ERR_PTR(ret);
> - }
> -
> - paddr = physaddr(obj);
> - for (i = 0; i < npages; i++) {
> - p[i] = pfn_to_page(__phys_to_pfn(paddr));
> - paddr += PAGE_SIZE;
> - }
> -
> - return p;
> -}
> -
> static struct page **get_pages(struct drm_gem_object *obj)
> {
> struct msm_gem_object *msm_obj = to_msm_bo(obj);
> @@ -180,10 +134,7 @@ static struct page **get_pages(struct drm_gem_object *obj)
> struct page **p;
> int npages = obj->size >> PAGE_SHIFT;
>
> - if (use_pages(obj))
> - p = drm_gem_get_pages(obj);
> - else
> - p = get_pages_vram(obj, npages);
> + p = drm_gem_get_pages(obj);
>
> if (IS_ERR(p)) {
> DRM_DEV_ERROR(dev->dev, "could not get pages: %ld\n",
> @@ -216,18 +167,6 @@ static struct page **get_pages(struct drm_gem_object *obj)
> return msm_obj->pages;
> }
>
> -static void put_pages_vram(struct drm_gem_object *obj)
> -{
> - struct msm_gem_object *msm_obj = to_msm_bo(obj);
> - struct msm_drm_private *priv = obj->dev->dev_private;
> -
> - spin_lock(&priv->vram.lock);
> - drm_mm_remove_node(msm_obj->vram_node);
> - spin_unlock(&priv->vram.lock);
> -
> - kvfree(msm_obj->pages);
> -}
> -
> static void put_pages(struct drm_gem_object *obj)
> {
> struct msm_gem_object *msm_obj = to_msm_bo(obj);
> @@ -248,10 +187,7 @@ static void put_pages(struct drm_gem_object *obj)
>
> update_device_mem(obj->dev->dev_private, -obj->size);
>
> - if (use_pages(obj))
> - drm_gem_put_pages(obj, msm_obj->pages, true, false);
> - else
> - put_pages_vram(obj);
> + drm_gem_put_pages(obj, msm_obj->pages, true, false);
>
> msm_obj->pages = NULL;
> update_lru(obj);
> @@ -1215,19 +1151,10 @@ struct drm_gem_object *msm_gem_new(struct drm_device *dev, uint32_t size, uint32
> struct msm_drm_private *priv = dev->dev_private;
> struct msm_gem_object *msm_obj;
> struct drm_gem_object *obj = NULL;
> - bool use_vram = false;
> int ret;
>
> size = PAGE_ALIGN(size);
>
> - if (!msm_use_mmu(dev))
> - use_vram = true;
> - else if ((flags & (MSM_BO_STOLEN | MSM_BO_SCANOUT)) && priv->vram.size)
> - use_vram = true;
> -
> - if (GEM_WARN_ON(use_vram && !priv->vram.size))
> - return ERR_PTR(-EINVAL);
> -
> /* Disallow zero sized objects as they make the underlying
> * infrastructure grumpy
> */
> @@ -1240,44 +1167,16 @@ struct drm_gem_object *msm_gem_new(struct drm_device *dev, uint32_t size, uint32
>
> msm_obj = to_msm_bo(obj);
>
> - if (use_vram) {
> - struct msm_gem_vma *vma;
> - struct page **pages;
> -
> - drm_gem_private_object_init(dev, obj, size);
> -
> - msm_gem_lock(obj);
> -
> - vma = add_vma(obj, NULL);
> - msm_gem_unlock(obj);
> - if (IS_ERR(vma)) {
> - ret = PTR_ERR(vma);
> - goto fail;
> - }
> -
> - to_msm_bo(obj)->vram_node = &vma->node;
> -
> - msm_gem_lock(obj);
> - pages = get_pages(obj);
> - msm_gem_unlock(obj);
> - if (IS_ERR(pages)) {
> - ret = PTR_ERR(pages);
> - goto fail;
> - }
> -
> - vma->iova = physaddr(obj);
> - } else {
> - ret = drm_gem_object_init(dev, obj, size);
> - if (ret)
> - goto fail;
> - /*
> - * Our buffers are kept pinned, so allocating them from the
> - * MOVABLE zone is a really bad idea, and conflicts with CMA.
> - * See comments above new_inode() why this is required _and_
> - * expected if you're going to pin these pages.
> - */
> - mapping_set_gfp_mask(obj->filp->f_mapping, GFP_HIGHUSER);
> - }
> + ret = drm_gem_object_init(dev, obj, size);
> + if (ret)
> + goto fail;
> + /*
> + * Our buffers are kept pinned, so allocating them from the
> + * MOVABLE zone is a really bad idea, and conflicts with CMA.
> + * See comments above new_inode() why this is required _and_
> + * expected if you're going to pin these pages.
> + */
> + mapping_set_gfp_mask(obj->filp->f_mapping, GFP_HIGHUSER);
>
> drm_gem_lru_move_tail(&priv->lru.unbacked, obj);
>
> @@ -1305,12 +1204,6 @@ struct drm_gem_object *msm_gem_import(struct drm_device *dev,
> uint32_t size;
> int ret, npages;
>
> - /* if we don't have IOMMU, don't bother pretending we can import: */
> - if (!msm_use_mmu(dev)) {
> - DRM_DEV_ERROR(dev->dev, "cannot import without IOMMU\n");
> - return ERR_PTR(-EINVAL);
> - }
> -
> size = PAGE_ALIGN(dmabuf->size);
>
> ret = msm_gem_new_impl(dev, size, MSM_BO_WC, &obj);
> diff --git a/drivers/gpu/drm/msm/msm_gem.h b/drivers/gpu/drm/msm/msm_gem.h
> index d2f39a371373..c16b11182831 100644
> --- a/drivers/gpu/drm/msm/msm_gem.h
> +++ b/drivers/gpu/drm/msm/msm_gem.h
> @@ -102,11 +102,6 @@ struct msm_gem_object {
>
> struct list_head vmas; /* list of msm_gem_vma */
>
> - /* For physically contiguous buffers. Used when we don't have
> - * an IOMMU. Also used for stolen/splashscreen buffer.
> - */
> - struct drm_mm_node *vram_node;
> -
> char name[32]; /* Identifier to print for the debugfs files */
>
> /* userspace metadata backchannel */
> diff --git a/drivers/gpu/drm/msm/msm_gem_submit.c b/drivers/gpu/drm/msm/msm_gem_submit.c
> index 95da4714fffb..a186b7dfea35 100644
> --- a/drivers/gpu/drm/msm/msm_gem_submit.c
> +++ b/drivers/gpu/drm/msm/msm_gem_submit.c
> @@ -659,11 +659,6 @@ int msm_ioctl_gem_submit(struct drm_device *dev, void *data,
> if (args->pad)
> return -EINVAL;
>
> - if (unlikely(!ctx->vm) && !capable(CAP_SYS_RAWIO)) {
> - DRM_ERROR_RATELIMITED("IOMMU support or CAP_SYS_RAWIO required!\n");
> - return -EPERM;
> - }
> -
> /* for now, we just have 3d pipe.. eventually this would need to
> * be more clever to dispatch to appropriate gpu module:
> */
Powered by blists - more mailing lists