lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <20190916145356.wddnnl3kk2awmbf4@ti.com>
Date:   Mon, 16 Sep 2019 09:53:56 -0500
From:   Benoit Parrot <bparrot@...com>
To:     "Lad, Prabhakar" <prabhakar.csengg@...il.com>
CC:     Hans Verkuil <hverkuil@...all.nl>,
        linux-media <linux-media@...r.kernel.org>,
        <devicetree@...r.kernel.org>, LKML <linux-kernel@...r.kernel.org>
Subject: Re: [Patch 05/13] media: am437x-vpfe: Streamlined vb2 buffer cleanup

Lad, Prabhakar <prabhakar.csengg@...il.com> wrote on Mon [2019-Sep-16 09:00:03 +0100]:
> Hi Benoit,
> 
> Thank you for the patch.
> 
> On Mon, Sep 9, 2019 at 5:26 PM Benoit Parrot <bparrot@...com> wrote:
> >
> > Returning queued vb2 buffers back to user space is a common
> > task best handled by a helper function.
> >
> > Signed-off-by: Benoit Parrot <bparrot@...com>
> > ---
> >  drivers/media/platform/am437x/am437x-vpfe.c | 54 ++++++++++-----------
> >  1 file changed, 26 insertions(+), 28 deletions(-)
> >
> > diff --git a/drivers/media/platform/am437x/am437x-vpfe.c b/drivers/media/platform/am437x/am437x-vpfe.c
> > index 3a8ad9bdf283..52f7fc6e11dd 100644
> > --- a/drivers/media/platform/am437x/am437x-vpfe.c
> > +++ b/drivers/media/platform/am437x/am437x-vpfe.c
> > @@ -1949,6 +1949,29 @@ static void vpfe_buffer_queue(struct vb2_buffer *vb)
> >         spin_unlock_irqrestore(&vpfe->dma_queue_lock, flags);
> >  }
> >
> > +static void vpfe_return_all_buffers(struct vpfe_device *vpfe,
> > +                                   enum vb2_buffer_state state)
> > +{
> > +       struct vpfe_cap_buffer *buf, *node;
> > +       unsigned long flags;
> > +
> > +       spin_lock_irqsave(&vpfe->dma_queue_lock, flags);
> > +       list_for_each_entry_safe(buf, node, &vpfe->dma_queue, list) {
> > +               vb2_buffer_done(&buf->vb.vb2_buf, state);
> > +               list_del(&buf->list);
> > +       }
> > +
> > +       if (vpfe->cur_frm)
> > +               vb2_buffer_done(&vpfe->cur_frm->vb.vb2_buf, state);
> > +
> > +       if (vpfe->next_frm && vpfe->next_frm != vpfe->cur_frm)
> > +               vb2_buffer_done(&vpfe->next_frm->vb.vb2_buf, state);
> > +
> > +       vpfe->cur_frm = NULL;
> > +       vpfe->next_frm = NULL;
> > +       spin_unlock_irqrestore(&vpfe->dma_queue_lock, flags);
> > +}
> > +
> >  /*
> >   * vpfe_start_streaming : Starts the DMA engine for streaming
> >   * @vb: ptr to vb2_buffer
> > @@ -1957,7 +1980,6 @@ static void vpfe_buffer_queue(struct vb2_buffer *vb)
> >  static int vpfe_start_streaming(struct vb2_queue *vq, unsigned int count)
> >  {
> >         struct vpfe_device *vpfe = vb2_get_drv_priv(vq);
> > -       struct vpfe_cap_buffer *buf, *tmp;
> >         struct vpfe_subdev_info *sdinfo;
> >         unsigned long flags;
> >         unsigned long addr;
> > @@ -2003,11 +2025,8 @@ static int vpfe_start_streaming(struct vb2_queue *vq, unsigned int count)
> >         return 0;
> >
> >  err:
> > -       list_for_each_entry_safe(buf, tmp, &vpfe->dma_queue, list) {
> > -               list_del(&buf->list);
> > -               vb2_buffer_done(&buf->vb.vb2_buf, VB2_BUF_STATE_QUEUED);
> > -       }
> > -
> > +       vpfe_return_all_buffers(vpfe, VB2_BUF_STATE_QUEUED);
> > +       vpfe_pcr_enable(&vpfe->ccdc, 0);
> 
> please create a seperate patch for the above change.

You mean a separate patch just for the vpfe_pcr_enable() call?

Benoit

> 
> Cheers,
> --Prabhakar Lad
> 
> >         return ret;
> >  }
> >
> > @@ -2022,7 +2041,6 @@ static void vpfe_stop_streaming(struct vb2_queue *vq)
> >  {
> >         struct vpfe_device *vpfe = vb2_get_drv_priv(vq);
> >         struct vpfe_subdev_info *sdinfo;
> > -       unsigned long flags;
> >         int ret;
> >
> >         vpfe_pcr_enable(&vpfe->ccdc, 0);
> > @@ -2040,27 +2058,7 @@ static void vpfe_stop_streaming(struct vb2_queue *vq)
> >                 vpfe_dbg(1, vpfe, "stream off failed in subdev\n");
> >
> >         /* release all active buffers */
> > -       spin_lock_irqsave(&vpfe->dma_queue_lock, flags);
> > -       if (vpfe->cur_frm == vpfe->next_frm) {
> > -               vb2_buffer_done(&vpfe->cur_frm->vb.vb2_buf,
> > -                               VB2_BUF_STATE_ERROR);
> > -       } else {
> > -               if (vpfe->cur_frm != NULL)
> > -                       vb2_buffer_done(&vpfe->cur_frm->vb.vb2_buf,
> > -                                       VB2_BUF_STATE_ERROR);
> > -               if (vpfe->next_frm != NULL)
> > -                       vb2_buffer_done(&vpfe->next_frm->vb.vb2_buf,
> > -                                       VB2_BUF_STATE_ERROR);
> > -       }
> > -
> > -       while (!list_empty(&vpfe->dma_queue)) {
> > -               vpfe->next_frm = list_entry(vpfe->dma_queue.next,
> > -                                               struct vpfe_cap_buffer, list);
> > -               list_del(&vpfe->next_frm->list);
> > -               vb2_buffer_done(&vpfe->next_frm->vb.vb2_buf,
> > -                               VB2_BUF_STATE_ERROR);
> > -       }
> > -       spin_unlock_irqrestore(&vpfe->dma_queue_lock, flags);
> > +       vpfe_return_all_buffers(vpfe, VB2_BUF_STATE_ERROR);
> >  }
> >
> >  static int vpfe_g_pixelaspect(struct file *file, void *priv,
> > --
> > 2.17.1
> >

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ