lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <20200911074931.GA1714160@krava>
Date:   Fri, 11 Sep 2020 09:49:31 +0200
From:   Jiri Olsa <jolsa@...hat.com>
To:     Namhyung Kim <namhyung@...nel.org>
Cc:     Jiri Olsa <jolsa@...nel.org>,
        Peter Zijlstra <a.p.zijlstra@...llo.nl>,
        Arnaldo Carvalho de Melo <acme@...nel.org>,
        lkml <linux-kernel@...r.kernel.org>,
        Ingo Molnar <mingo@...nel.org>,
        Alexander Shishkin <alexander.shishkin@...ux.intel.com>,
        Michael Petlan <mpetlan@...hat.com>,
        Wade Mealing <wmealing@...hat.com>
Subject: Re: [PATCH] perf: Fix race in perf_mmap_close function

On Fri, Sep 11, 2020 at 12:05:10PM +0900, Namhyung Kim wrote:
> Hi Jiri,
> 
> On Thu, Sep 10, 2020 at 11:50 PM Jiri Olsa <jolsa@...hat.com> wrote:
> >
> > On Thu, Sep 10, 2020 at 10:48:02PM +0900, Namhyung Kim wrote:
> >
> > SNIP
> >
> > > >   _do_fork+0x83/0x3a0
> > > >   __do_sys_wait4+0x83/0x90
> > > >   __do_sys_clone+0x85/0xa0
> > > >   do_syscall_64+0x5b/0x1e0
> > > >   entry_SYSCALL_64_after_hwframe+0x44/0xa9
> > > >
> > > > Using atomic decrease and check instead of separated calls.
> > > > This fixes CVE-2020-14351.
> > > >
> > > > Signed-off-by: Jiri Olsa <jolsa@...nel.org>
> > > > ---
> > > >  kernel/events/core.c | 4 +---
> > > >  1 file changed, 1 insertion(+), 3 deletions(-)
> > > >
> > > > diff --git a/kernel/events/core.c b/kernel/events/core.c
> > > > index 7ed5248f0445..29313cc54d9e 100644
> > > > --- a/kernel/events/core.c
> > > > +++ b/kernel/events/core.c
> > > > @@ -5903,8 +5903,6 @@ static void perf_mmap_close(struct vm_area_struct *vma)
> > > >                 mutex_unlock(&event->mmap_mutex);
> > > >         }
> > > >
> > > > -       atomic_dec(&rb->mmap_count);
> > > > -
> > > >         if (!atomic_dec_and_mutex_lock(&event->mmap_count, &event->mmap_mutex))
> > > >                 goto out_put;
> > >
> > > But when it takes the goto, rb->mmap_count won't decrement anymore..
> >
> > event->mmap_count is per event, so if we have have race in here,
> > 2 threads can go through with each event->mmap_count reaching zero
> 
> Maybe I'm missing something.
> 
> But as far as I can see, perf_mmap_close() always decremented both
> rb->mmap_count and event->mmap_count.  But with this change,
> it seems not decrement rb->mmap_count when event->mmap_count
> doesn't go to zero, right?

ugh, that's right.. how about change below

jirka


---
diff --git a/kernel/events/core.c b/kernel/events/core.c
index 7ed5248f0445..8ab2400aef55 100644
--- a/kernel/events/core.c
+++ b/kernel/events/core.c
@@ -5868,11 +5868,11 @@ static void perf_pmu_output_stop(struct perf_event *event);
 static void perf_mmap_close(struct vm_area_struct *vma)
 {
 	struct perf_event *event = vma->vm_file->private_data;
-
 	struct perf_buffer *rb = ring_buffer_get(event);
 	struct user_struct *mmap_user = rb->mmap_user;
 	int mmap_locked = rb->mmap_locked;
 	unsigned long size = perf_data_size(rb);
+	bool detach_rest = false;
 
 	if (event->pmu->event_unmapped)
 		event->pmu->event_unmapped(event, vma->vm_mm);
@@ -5903,7 +5903,8 @@ static void perf_mmap_close(struct vm_area_struct *vma)
 		mutex_unlock(&event->mmap_mutex);
 	}
 
-	atomic_dec(&rb->mmap_count);
+	if (atomic_dec_and_test(&rb->mmap_count))
+		detach_rest = true;
 
 	if (!atomic_dec_and_mutex_lock(&event->mmap_count, &event->mmap_mutex))
 		goto out_put;
@@ -5912,7 +5913,7 @@ static void perf_mmap_close(struct vm_area_struct *vma)
 	mutex_unlock(&event->mmap_mutex);
 
 	/* If there's still other mmap()s of this buffer, we're done. */
-	if (atomic_read(&rb->mmap_count))
+	if (!detach_rest)
 		goto out_put;
 
 	/*

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ