[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <1379691574.2815.1548715576651.JavaMail.zimbra@efficios.com>
Date: Mon, 28 Jan 2019 17:46:16 -0500 (EST)
From: Mathieu Desnoyers <mathieu.desnoyers@...icios.com>
To: paulmck <paulmck@...ux.ibm.com>
Cc: Ingo Molnar <mingo@...nel.org>,
Peter Zijlstra <peterz@...radead.org>,
linux-kernel <linux-kernel@...r.kernel.org>,
linux-api <linux-api@...r.kernel.org>,
Jann Horn <jannh@...gle.com>,
Thomas Gleixner <tglx@...utronix.de>,
Andrea Parri <parri.andrea@...il.com>,
Andy Lutomirski <luto@...nel.org>,
Avi Kivity <avi@...lladb.com>,
Benjamin Herrenschmidt <benh@...nel.crashing.org>,
Boqun Feng <boqun.feng@...il.com>,
Dave Watson <davejwatson@...com>, David Sehr <sehr@...gle.com>,
"H. Peter Anvin" <hpa@...or.com>,
Linus Torvalds <torvalds@...ux-foundation.org>,
maged michael <maged.michael@...il.com>,
Michael Ellerman <mpe@...erman.id.au>,
Paul Mackerras <paulus@...ba.org>,
"Russell King, ARM Linux" <linux@...linux.org.uk>,
Will Deacon <will.deacon@....com>,
stable <stable@...r.kernel.org>
Subject: Re: [PATCH] Fix: membarrier: racy access to p->mm in
membarrier_global_expedited()
----- On Jan 28, 2019, at 5:39 PM, paulmck paulmck@...ux.ibm.com wrote:
> On Mon, Jan 28, 2019 at 05:07:07PM -0500, Mathieu Desnoyers wrote:
>> Jann Horn identified a racy access to p->mm in the global expedited
>> command of the membarrier system call.
>>
>> The suggested fix is to hold the task_lock() around the accesses to
>> p->mm and to the mm_struct membarrier_state field to guarantee the
>> existence of the mm_struct.
>>
>> Link:
>> https://lore.kernel.org/lkml/CAG48ez2G8ctF8dHS42TF37pThfr3y0RNOOYTmxvACm4u8Yu3cw@mail.gmail.com
>> Signed-off-by: Mathieu Desnoyers <mathieu.desnoyers@...icios.com>
>> Tested-by: Jann Horn <jannh@...gle.com>
>> CC: Jann Horn <jannh@...gle.com>
>> CC: Thomas Gleixner <tglx@...utronix.de>
>> CC: Peter Zijlstra (Intel) <peterz@...radead.org>
>> CC: Ingo Molnar <mingo@...nel.org>
>> CC: Andrea Parri <parri.andrea@...il.com>
>> CC: Andy Lutomirski <luto@...nel.org>
>> CC: Avi Kivity <avi@...lladb.com>
>> CC: Benjamin Herrenschmidt <benh@...nel.crashing.org>
>> CC: Boqun Feng <boqun.feng@...il.com>
>> CC: Dave Watson <davejwatson@...com>
>> CC: David Sehr <sehr@...gle.com>
>> CC: H. Peter Anvin <hpa@...or.com>
>> CC: Linus Torvalds <torvalds@...ux-foundation.org>
>> CC: Maged Michael <maged.michael@...il.com>
>> CC: Michael Ellerman <mpe@...erman.id.au>
>> CC: Paul E. McKenney <paulmck@...ux.vnet.ibm.com>
>> CC: Paul Mackerras <paulus@...ba.org>
>> CC: Russell King <linux@...linux.org.uk>
>> CC: Will Deacon <will.deacon@....com>
>> CC: stable@...r.kernel.org # v4.16+
>> CC: linux-api@...r.kernel.org
>> ---
>> kernel/sched/membarrier.c | 27 +++++++++++++++++++++------
>> 1 file changed, 21 insertions(+), 6 deletions(-)
>>
>> diff --git a/kernel/sched/membarrier.c b/kernel/sched/membarrier.c
>> index 76e0eaf4654e..305fdcc4c5f7 100644
>> --- a/kernel/sched/membarrier.c
>> +++ b/kernel/sched/membarrier.c
>> @@ -81,12 +81,27 @@ static int membarrier_global_expedited(void)
>>
>> rcu_read_lock();
>> p = task_rcu_dereference(&cpu_rq(cpu)->curr);
>> - if (p && p->mm && (atomic_read(&p->mm->membarrier_state) &
>> - MEMBARRIER_STATE_GLOBAL_EXPEDITED)) {
>> - if (!fallback)
>> - __cpumask_set_cpu(cpu, tmpmask);
>> - else
>> - smp_call_function_single(cpu, ipi_mb, NULL, 1);
>> + /*
>> + * Skip this CPU if the runqueue's current task is NULL or if
>> + * it is a kernel thread.
>> + */
>> + if (p && READ_ONCE(p->mm)) {
>> + bool mm_match;
>> +
>> + /*
>> + * Read p->mm and access membarrier_state while holding
>> + * the task lock to ensure existence of mm.
>> + */
>> + task_lock(p);
>> + mm_match = p->mm && (atomic_read(&p->mm->membarrier_state) &
>
> Are we guaranteed that this p->mm will be the same as the one loaded via
> READ_ONCE() above? Either way, wouldn't it be better to READ_ONCE() it a
> single time and use the same value everywhere?
The first "READ_ONCE()" above is _outside_ of the task_lock() critical section.
Those two accesses _can_ load two different pointers, and this is why we
need to re-read the p->mm pointer within the task_lock() critical section to
ensure existence of the mm_struct that we use.
If we move the READ_ONCE() into the task_lock(), we need to uselessly
take a lock before we can skip kernel threads.
If we lead the READ_ONCE() outside the task_lock(), then p->mm can be updated
between the READ_ONCE() and reference to the mm_struct content within the
task_lock(), which is racy and does not guarantee its existence.
Or am I missing your point ?
Thanks,
Mathieu
>
> Thanx, Paul
>
>> + MEMBARRIER_STATE_GLOBAL_EXPEDITED);
>> + task_unlock(p);
>> + if (mm_match) {
>> + if (!fallback)
>> + __cpumask_set_cpu(cpu, tmpmask);
>> + else
>> + smp_call_function_single(cpu, ipi_mb, NULL, 1);
>> + }
>> }
>> rcu_read_unlock();
>> }
>> --
>> 2.17.1
--
Mathieu Desnoyers
EfficiOS Inc.
http://www.efficios.com
Powered by blists - more mailing lists