lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Date:   Tue, 14 Nov 2017 10:06:36 -0500
From:   Neil Horman <nhorman@...driver.com>
To:     Xin Long <lucien.xin@...il.com>
Cc:     network dev <netdev@...r.kernel.org>, linux-sctp@...r.kernel.org,
        davem <davem@...emloft.net>,
        Marcelo Ricardo Leitner <marcelo.leitner@...il.com>
Subject: Re: [PATCH net] sctp: add wait_buf flag in asoc to avoid the peeloff
 and wait sndbuf race

On Tue, Nov 14, 2017 at 10:46:34PM +0800, Xin Long wrote:
> On Tue, Nov 14, 2017 at 10:31 PM, Neil Horman <nhorman@...driver.com> wrote:
> > On Mon, Nov 13, 2017 at 11:49:28PM +0800, Xin Long wrote:
> >> On Mon, Nov 13, 2017 at 11:40 PM, Xin Long <lucien.xin@...il.com> wrote:
> >> > On Mon, Nov 13, 2017 at 11:23 PM, Neil Horman <nhorman@...driver.com> wrote:
> >> >> On Mon, Nov 13, 2017 at 01:47:58PM +0800, Xin Long wrote:
> >> >>> Commit dfcb9f4f99f1 ("sctp: deny peeloff operation on asocs with threads
> >> >>> sleeping on it") fixed the race between peeloff and wait sndbuf by
> >> >>> checking waitqueue_active(&asoc->wait) in sctp_do_peeloff().
> >> >>>
> >> >>> But it actually doesn't work as even if waitqueue_active returns false
> >> >>> the waiting sndbuf thread may still not yet hold sk lock.
> >> >>>
> >> >>> This patch is to fix this by adding wait_buf flag in asoc, and setting it
> >> >>> before going the waiting loop, clearing it until the waiting loop breaks,
> >> >>> and checking it in sctp_do_peeloff instead.
> >> >>>
> >> >>> Fixes: dfcb9f4f99f1 ("sctp: deny peeloff operation on asocs with threads sleeping on it")
> >> >>> Suggested-by: Marcelo Ricardo Leitner <marcelo.leitner@...il.com>
> >> >>> Signed-off-by: Xin Long <lucien.xin@...il.com>
> >> >>> ---
> >> >>>  include/net/sctp/structs.h | 1 +
> >> >>>  net/sctp/socket.c          | 4 +++-
> >> >>>  2 files changed, 4 insertions(+), 1 deletion(-)
> >> >>>
> >> >>> diff --git a/include/net/sctp/structs.h b/include/net/sctp/structs.h
> >> >>> index 0477945..446350e 100644
> >> >>> --- a/include/net/sctp/structs.h
> >> >>> +++ b/include/net/sctp/structs.h
> >> >>> @@ -1883,6 +1883,7 @@ struct sctp_association {
> >> >>>
> >> >>>       __u8 need_ecne:1,       /* Need to send an ECNE Chunk? */
> >> >>>            temp:1,            /* Is it a temporary association? */
> >> >>> +          wait_buf:1,
> >> >>>            force_delay:1,
> >> >>>            prsctp_enable:1,
> >> >>>            reconf_enable:1;
> >> >>> diff --git a/net/sctp/socket.c b/net/sctp/socket.c
> >> >>> index 6f45d17..1b2c78c 100644
> >> >>> --- a/net/sctp/socket.c
> >> >>> +++ b/net/sctp/socket.c
> >> >>> @@ -4946,7 +4946,7 @@ int sctp_do_peeloff(struct sock *sk, sctp_assoc_t id, struct socket **sockp)
> >> >>>       /* If there is a thread waiting on more sndbuf space for
> >> >>>        * sending on this asoc, it cannot be peeled.
> >> >>>        */
> >> >>> -     if (waitqueue_active(&asoc->wait))
> >> >>> +     if (asoc->wait_buf)
> >> >>>               return -EBUSY;
> >> >>>
> >> >>>       /* An association cannot be branched off from an already peeled-off
> >> >>> @@ -7835,6 +7835,7 @@ static int sctp_wait_for_sndbuf(struct sctp_association *asoc, long *timeo_p,
> >> >>>       /* Increment the association's refcnt.  */
> >> >>>       sctp_association_hold(asoc);
> >> >>>
> >> >>> +     asoc->wait_buf = 1;
> >> >>>       /* Wait on the association specific sndbuf space. */
> >> >>>       for (;;) {
> >> >>>               prepare_to_wait_exclusive(&asoc->wait, &wait,
> >> >>> @@ -7860,6 +7861,7 @@ static int sctp_wait_for_sndbuf(struct sctp_association *asoc, long *timeo_p,
> >> >>>       }
> >> >>>
> >> >>>  out:
> >> >>> +     asoc->wait_buf = 0;
> >> >>>       finish_wait(&asoc->wait, &wait);
> >> >>>
> >> >>>       /* Release the association's refcnt.  */
> >> >>> --
> >> >>> 2.1.0
> >> >>>
> >> >>>
> >> >>
> >> >> This doesn't make much sense to me, as it appears to be prone to aliasing.  That
> >> >> is to say:
> >> >>
> >> >> a) If multiple tasks are queued waiting in sctp_wait_for_sndbuf, the first
> >> >> thread to exit that for(;;) loop will clean asoc->wait_buf, even though others
> >> >> may be waiting on it, allowing sctp_do_peeloff to continue when it shouldn't be
> >> > You're right, we talked about this before using waitqueue_active in
> >> > earlier time.
> >> > I didn't remember this somehow. Sorry.
> >> >
> >> >>
> >> >> b) In the case of a single task blocking in sct_wait_for_sendbuf, checking
> >> >> waitqueue_active is equally good, because it returns true, until such time as
> >> >> finish_wait is called anyway.
> >> > waitqueue_active can not work here, because in sctp_wait_for_sndbuf():
> >> > ...
> >> >                 release_sock(sk);
> >> >                 current_timeo = schedule_timeout(current_timeo); <-----[a]
> >> >                 lock_sock(sk);
> >> > If another thread wakes up asoc->wait, it will be removed from
> >> > this wait queue, you check DEFINE_WAIT, the callback autoremove_wake_function
> >> > will do this removal in wake_up().
> >> >
> >> > I guess we need to think about another to fix this.
> >> maybe we can use
> >> DEFINE_WAIT_FUNC(wait, woken_wake_function);
> >> instead of DEFINE_WAIT(wait) here ?
> >>
> > I'm still not sure I see the problem here.  If we have the following situation:
> > * Exec context A is executing in sctp_do_peeloff, about to check
> >   waitqueue_active()
> > * Exec context B is blocking in sctp_wait_for_sndbuf(), specifically without the
> >   socket lock held
> >
> >
> > Then, we have two possibilities:
> >
> > a) Context A executes waitqueue_active, which returns true, implying that
> > context B is still on the queue, or that some other undescribed context has
> > begun waiting on the queue.  In either case, the behavior is correct, in that
> > the peeloff is denied.
> >
> > b) Context B is woken up (and in the most pessimal case, has its waitq entry
> > removed from queue immediately, causing context B to have waitequeue_active
> > return false, allowing it to continue processing the peeloff.  Since it holds
> > the socket lock however, context B will block on the lock_sock operation until
> > such time as the peeloff completes, so you're safe.
> >
> > About the only issue that I see (and as I write this, I may be seeing what you
> > are actually trying to fix here) is that, during the period where context A is
> > sleeping in sctp_wait_for_sendbuf, with the socket lock released, it is possible
> > for an sctp_do_peeloff operation to complete, meaning that assoc->base.sk might
> > point to a new socket, allowing each context to hold an independent socket lock
> > and execute in parallel.  To combat that, I think all you really need is some
> > code in sctp_wait_for_sndbuf that looks like this:
> >
> > release_sock(sk);
> > current_timeo = schedule_timeout(current_timeo);
> > lock_sock(sk);
> >
> > if (sk != asoc->base.sk) {
> >         /* a socket peeloff occured */
> >         release_sock(sk);
> >         sk = assoc->base.sk;
> >         lock_sock(sk);
> > }
> >
> > *timeo_p = current_timeo;
> >
> >
> > Does that make sense?  This way, you lock the 'old' socket lock to ensure that
> > the peeloff operation is completed, then you check to see if the socket has
> > changed.  If it has, you migrate your socket to the new, peeled off one and
> > continue your space availability check
> Yes, you got what I'm trying to fix in this patch exactly. :-)
> and the fix you proposed above is doable, but incomplete,
> we also need to change the sk pointer in sctp_sendmsg:
> @@ -1962,7 +1962,7 @@ static int sctp_sendmsg(struct sock *sk, struct
> msghdr *msg, size_t msg_len)
> 
>         timeo = sock_sndtimeo(sk, msg->msg_flags & MSG_DONTWAIT);
>         if (!sctp_wspace(asoc)) {
> -               err = sctp_wait_for_sndbuf(asoc, &timeo, msg_len);
> +               err = sctp_wait_for_sndbuf(asoc, &timeo, msg_len, &sk);
>                 if (err) {
>                         if (err == -ESRCH) {
>                                 /* asoc is already dead; */
> @@ -7828,7 +7828,7 @@ void sctp_sock_rfree(struct sk_buff *skb)
> 
>  /* Helper function to wait for space in the sndbuf.  */
>  static int sctp_wait_for_sndbuf(struct sctp_association *asoc, long *timeo_p,
> -                               size_t msg_len)
> +                               size_t msg_len, struct sock **orig_sk)
>  {
>         struct sock *sk = asoc->base.sk;
>         int err = 0;
> @@ -7862,11 +7862,17 @@ static int sctp_wait_for_sndbuf(struct
> sctp_association *asoc, long *timeo_p,
>                 release_sock(sk);
>                 current_timeo = schedule_timeout(current_timeo);
>                 lock_sock(sk);
> +               if (sk != asoc->base.sk) {
> +                       release_sock(sk);
> +                       sk = asoc->base.sk;
> +                       lock_sock(sk);
> +               }
> 
>                 *timeo_p = current_timeo;
>         }
> 
>  out:
> +       *orig_sk = sk;
>         finish_wait(&asoc->wait, &wait);
> 
> 
> right ?
> 

Yes, that makes sense, post that as a proper commit please, I'll support that.

Neil

Powered by blists - more mailing lists