lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Date:   Thu, 9 Sep 2021 20:32:55 -0700
From:   Li Li <dualli@...omium.org>
To:     Todd Kjos <tkjos@...gle.com>
Cc:     Li Li <dualli@...gle.com>, Greg KH <gregkh@...uxfoundation.org>,
        Christian Brauner <christian@...uner.io>,
        Arve Hjønnevåg <arve@...roid.com>,
        "open list:ANDROID DRIVERS" <devel@...verdev.osuosl.org>,
        LKML <linux-kernel@...r.kernel.org>,
        Martijn Coenen <maco@...gle.com>,
        Hridya Valsaraju <hridya@...gle.com>,
        Suren Baghdasaryan <surenb@...gle.com>,
        "Joel Fernandes (Google)" <joel@...lfernandes.org>,
        Android Kernel Team <kernel-team@...roid.com>
Subject: Re: [PATCH v1 1/1] binder: fix freeze race

Hi Todd,

Thanks for reviewing the patch! Please see my reply below.

And I'll send out v2 soon addressing your concerns.

On Thu, Sep 9, 2021 at 4:54 PM Todd Kjos <tkjos@...gle.com> wrote:
>
> On Thu, Sep 9, 2021 at 4:21 PM Li Li <dualli@...omium.org> wrote:
> >
> > From: Li Li <dualli@...gle.com>
> >
> > Currently cgroup freezer is used to freeze the application threads, and
> > BINDER_FREEZE is used to freeze binder interface. There's already a
> > mechanism for BINDER_FREEZE to wait for any existing transactions to
> > drain out before actually freezing the binder interface.
> >
> > But freezing an app requires 2 steps, freezing the binder interface with
> > BINDER_FREEZEwhich and then freezing the application main threads with
> > cgroupfs. This is not an atomic operation. The following race issue
> > might happen.
> >
> > 1) Binder interface is frozen by ioctl(BINDER_FREEZE);
> > 2) Main thread initiates a new sync binder transaction;
> > 3) Main thread is frozen by "echo 1 > cgroup.freeze";
> > 4) The response reaches the frozen thread, which will unexpectedly fail.
> >
> > This patch provides a mechanism for user space freezer manager to check
> > if there's any new pending transaction happening between BINDER_FREEZE
> > and freezing main thread. If there's any, the main thread freezing
> > operation can be rolled back.
> >
> > Furthermore, the response might reach the binder driver before the
> > rollback actually happens. That will also cause failed transaction.
> >
> > As the other process doesn't wait for another response of the response,
> > the failed response transaction can be fixed by treating the response
> > transaction like an oneway/async one, allowing it to reach the frozen
> > thread. And it will be consumed when the thread gets unfrozen later.
> >
> > Bug: 198493121
>
> Please remove "Bug: " tag. Replace it with a "Fixes:" tag that cites
> the patch that introduced the race.
>
Done in v2.

> > Signed-off-by: Li Li <dualli@...gle.com>
> > ---
> >  drivers/android/binder.c          | 32 +++++++++++++++++++++++++++----
> >  drivers/android/binder_internal.h |  2 ++
> >  2 files changed, 30 insertions(+), 4 deletions(-)
> >
> > diff --git a/drivers/android/binder.c b/drivers/android/binder.c
> > index d9030cb6b1e4..f9713a97578c 100644
> > --- a/drivers/android/binder.c
> > +++ b/drivers/android/binder.c
> > @@ -3038,9 +3038,8 @@ static void binder_transaction(struct binder_proc *proc,
> >         if (reply) {
> >                 binder_enqueue_thread_work(thread, tcomplete);
> >                 binder_inner_proc_lock(target_proc);
> > -               if (target_thread->is_dead || target_proc->is_frozen) {
> > -                       return_error = target_thread->is_dead ?
> > -                               BR_DEAD_REPLY : BR_FROZEN_REPLY;
> > +               if (target_thread->is_dead) {
> > +                       return_error = BR_DEAD_REPLY;
> >                         binder_inner_proc_unlock(target_proc);
> >                         goto err_dead_proc_or_thread;
> >                 }
> > @@ -4648,6 +4647,22 @@ static int binder_ioctl_get_node_debug_info(struct binder_proc *proc,
> >         return 0;
> >  }
> >
> > +static int binder_txns_pending(struct binder_proc *proc)
> > +{
> > +       struct rb_node *n;
> > +       struct binder_thread *thread;
> > +
> > +       if (proc->outstanding_txns > 0)
> > +               return 1;
> > +
> > +       for (n = rb_first(&proc->threads); n; n = rb_next(n)) {
> > +               thread = rb_entry(n, struct binder_thread, rb_node);
> > +               if (thread->transaction_stack)
> > +                       return 1;
> > +       }
> > +       return 0;
> > +}
> > +
> >  static int binder_ioctl_freeze(struct binder_freeze_info *info,
> >                                struct binder_proc *target_proc)
> >  {
> > @@ -4682,6 +4697,14 @@ static int binder_ioctl_freeze(struct binder_freeze_info *info,
> >         if (!ret && target_proc->outstanding_txns)
> >                 ret = -EAGAIN;
> >
> > +       /* Also check pending transactions that wait for reply */
> > +       if (ret >= 0) {
> > +               binder_inner_proc_lock(target_proc);
> > +               if (binder_txns_pending(target_proc))
>
> The convention in the binder driver is to append "_ilocked" to the
> function name if the function expects the inner proc lock to be held
> (so "binder_txns_pending_ilocked(target_proc) in this case)".
>
Done in v2.

> > +                       ret = -EAGAIN;
> > +               binder_inner_proc_unlock(target_proc);
> > +       }
> > +
> >         if (ret < 0) {
> >                 binder_inner_proc_lock(target_proc);
> >                 target_proc->is_frozen = false;
> > @@ -4705,7 +4728,8 @@ static int binder_ioctl_get_freezer_info(
> >                 if (target_proc->pid == info->pid) {
> >                         found = true;
> >                         binder_inner_proc_lock(target_proc);
> > -                       info->sync_recv |= target_proc->sync_recv;
> > +                       info->sync_recv |= target_proc->sync_recv |
> > +                                       (binder_txns_pending(target_proc) << 1);
> >                         info->async_recv |= target_proc->async_recv;
> >                         binder_inner_proc_unlock(target_proc);
> >                 }
> > diff --git a/drivers/android/binder_internal.h b/drivers/android/binder_internal.h
> > index 810c0b84d3f8..402c4d4362a8 100644
> > --- a/drivers/android/binder_internal.h
> > +++ b/drivers/android/binder_internal.h
> > @@ -378,6 +378,8 @@ struct binder_ref {
> >   *                        binder transactions
> >   *                        (protected by @inner_lock)
> >   * @sync_recv:            process received sync transactions since last frozen
> > + *                        bit 0: received sync transaction after being frozen
> > + *                        bit 1: new pending sync transaction during freezing
>
> Should these bit assignments be documented in binder.h since these bit
> assignments end up being relevant in struct binder_frozen_status_info?
>
Done in v2.

>
> >   *                        (protected by @inner_lock)
> >   * @async_recv:           process received async transactions since last frozen
> >   *                        (protected by @inner_lock)
> > --
> > 2.33.0.309.g3052b89438-goog
> >

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ