lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [<thread-prev] [day] [month] [year] [list]
Message-ID: <CAF6AEGvXu_JHyfwmD+VL=iF5OcP=C-n0jsqd6sXM=Fa4KMQAgg@mail.gmail.com>
Date:   Mon, 1 May 2023 14:21:31 -0700
From:   Rob Clark <robdclark@...il.com>
To:     Dmitry Baryshkov <dmitry.baryshkov@...aro.org>
Cc:     dri-devel@...ts.freedesktop.org, freedreno@...ts.freedesktop.org,
        Rob Clark <robdclark@...omium.org>,
        Abhinav Kumar <quic_abhinavk@...cinc.com>,
        Sean Paul <sean@...rly.run>, David Airlie <airlied@...il.com>,
        Daniel Vetter <daniel@...ll.ch>,
        "open list:DRM DRIVER FOR MSM ADRENO GPU" 
        <linux-arm-msm@...r.kernel.org>,
        open list <linux-kernel@...r.kernel.org>
Subject: Re: [PATCH] drm/msm: Set max segment size earlier

On Mon, May 1, 2023 at 1:56 PM Dmitry Baryshkov
<dmitry.baryshkov@...aro.org> wrote:
>
> On 01/05/2023 23:44, Rob Clark wrote:
> > From: Rob Clark <robdclark@...omium.org>
> >
> > Fixes the following splat on a6xx gen2+ (a640, a650, a660 families),
> > a6xx gen1 has smaller GMU allocations so they fit under the default
> > 64K max segment size.
> >
> >     ------------[ cut here ]------------
> >     DMA-API: msm_dpu ae01000.display-controller: mapping sg segment longer than device claims to support [len=126976] [max=65536]
> >     WARNING: CPU: 5 PID: 9 at kernel/dma/debug.c:1160 debug_dma_map_sg+0x288/0x314
> >     Modules linked in:
> >     CPU: 5 PID: 9 Comm: kworker/u16:0 Not tainted 6.3.0-rc2-debug+ #629
> >     Hardware name: Google Villager (rev1+) with LTE (DT)
> >     Workqueue: events_unbound deferred_probe_work_func
> >     pstate: 60400009 (nZCv daif +PAN -UAO -TCO -DIT -SSBS BTYPE=--)
> >     pc : debug_dma_map_sg+0x288/0x314
> >     lr : debug_dma_map_sg+0x288/0x314
> >     sp : ffffffc00809b560
> >     x29: ffffffc00809b560 x28: 0000000000000060 x27: 0000000000000000
> >     x26: 0000000000010000 x25: 0000000000000004 x24: 0000000000000004
> >     x23: ffffffffffffffff x22: ffffffdb31693cc0 x21: ffffff8080935800
> >     x20: ffffff8087417400 x19: ffffff8087a45010 x18: 0000000000000000
> >     x17: 0000000000000000 x16: 0000000000000000 x15: 0000000000010000
> >     x14: 0000000000000001 x13: ffffffffffffffff x12: ffffffffffffffff
> >     x11: 0000000000000000 x10: 000000000000000a x9 : ffffffdb2ff05e14
> >     x8 : ffffffdb31275000 x7 : ffffffdb2ff08908 x6 : 0000000000000000
> >     x5 : 0000000000000001 x4 : ffffffdb2ff08a74 x3 : ffffffdb31275008
> >     x2 : 0000000000000000 x1 : 0000000000000000 x0 : ffffff80803a9a80
> >     Call trace:
> >      debug_dma_map_sg+0x288/0x314
> >      __dma_map_sg_attrs+0x80/0xe4
> >      dma_map_sgtable+0x30/0x4c
> >      get_pages+0x1d4/0x1e4
> >      msm_gem_pin_pages_locked+0xbc/0xf8
> >      msm_gem_pin_vma_locked+0x58/0xa0
> >      msm_gem_get_and_pin_iova_range+0x98/0xac
> >      a6xx_gmu_memory_alloc+0x7c/0x128
> >      a6xx_gmu_init+0x16c/0x9b0
> >      a6xx_gpu_init+0x38c/0x3e4
> >      adreno_bind+0x214/0x264
> >      component_bind_all+0x128/0x1f8
> >      msm_drm_bind+0x2b8/0x608
> >      try_to_bring_up_aggregate_device+0x88/0x1a4
> >      __component_add+0xec/0x13c
> >      component_add+0x1c/0x28
> >      dp_display_probe+0x3f8/0x43c
> >      platform_probe+0x70/0xc4
> >      really_probe+0x148/0x280
> >      __driver_probe_device+0xc8/0xe0
> >      driver_probe_device+0x44/0x100
> >      __device_attach_driver+0x64/0xdc
> >      bus_for_each_drv+0xb0/0xd8
> >      __device_attach+0xd8/0x168
> >      device_initial_probe+0x1c/0x28
> >      bus_probe_device+0x44/0xb0
> >      deferred_probe_work_func+0xc8/0xe0
> >      process_one_work+0x2e0/0x488
> >      process_scheduled_works+0x4c/0x50
> >      worker_thread+0x218/0x274
> >      kthread+0xf0/0x100
> >      ret_from_fork+0x10/0x20
> >     irq event stamp: 293712
> >     hardirqs last  enabled at (293711): [<ffffffdb2ff0893c>] vprintk_emit+0x160/0x25c
> >     hardirqs last disabled at (293712): [<ffffffdb30b48130>] el1_dbg+0x24/0x80
> >     softirqs last  enabled at (279520): [<ffffffdb2fe10420>] __do_softirq+0x21c/0x4bc
> >     softirqs last disabled at (279515): [<ffffffdb2fe16708>] ____do_softirq+0x18/0x24
> >     ---[ end trace 0000000000000000 ]---
> >
> > Signed-off-by: Rob Clark <robdclark@...omium.org>
>
> I think this should be:
>
> Fixes: db735fc4036b ("drm/msm: Set dma maximum segment size for mdss")

yeah, or perhaps just that commit didn't fix the issue hard enough.. I
was thinking that it was introduced by memory allocations out of
hw_init() but actually this has been here the whole time (on newer
a6xx gens).  There was an internal bug about it, but somehow it didn't
get routed to me.

BR,
-R

> Reviewed-by: Dmitry Baryshkov <dmitry.baryshkov@...aro.org>
>
> > ---
> >   drivers/gpu/drm/msm/msm_drv.c | 4 ++--
> >   1 file changed, 2 insertions(+), 2 deletions(-)
> >
> > diff --git a/drivers/gpu/drm/msm/msm_drv.c b/drivers/gpu/drm/msm/msm_drv.c
> > index 3a74b5653e96..6dec1a3534f2 100644
> > --- a/drivers/gpu/drm/msm/msm_drv.c
> > +++ b/drivers/gpu/drm/msm/msm_drv.c
> > @@ -440,27 +440,27 @@ static int msm_drm_init(struct device *dev, const struct drm_driver *drv)
> >       fs_reclaim_acquire(GFP_KERNEL);
> >       might_lock(&priv->lru.lock);
> >       fs_reclaim_release(GFP_KERNEL);
> >
> >       drm_mode_config_init(ddev);
> >
> >       ret = msm_init_vram(ddev);
> >       if (ret)
> >               goto err_drm_dev_put;
> >
> > +     dma_set_max_seg_size(dev, UINT_MAX);
> > +
> >       /* Bind all our sub-components: */
> >       ret = component_bind_all(dev, ddev);
> >       if (ret)
> >               goto err_drm_dev_put;
> >
> > -     dma_set_max_seg_size(dev, UINT_MAX);
> > -
> >       msm_gem_shrinker_init(ddev);
> >
> >       if (priv->kms_init) {
> >               ret = priv->kms_init(ddev);
> >               if (ret) {
> >                       DRM_DEV_ERROR(dev, "failed to load kms\n");
> >                       priv->kms = NULL;
> >                       goto err_msm_uninit;
> >               }
> >               kms = priv->kms;
>
> --
> With best wishes
> Dmitry
>

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ