[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <20190424134158.GC14829@fuggles.cambridge.arm.com>
Date: Wed, 24 Apr 2019 14:41:58 +0100
From: Will Deacon <will.deacon@....com>
To: Peter Zijlstra <peterz@...radead.org>
Cc: stern@...land.harvard.edu, akiyks@...il.com,
andrea.parri@...rulasolutions.com, boqun.feng@...il.com,
dlustig@...dia.com, dhowells@...hat.com, j.alglave@....ac.uk,
luc.maranget@...ia.fr, npiggin@...il.com, paulmck@...ux.ibm.com,
linux-kernel@...r.kernel.org, torvalds@...ux-foundation.org
Subject: Re: [RFC][PATCH 5/5] x86/atomic: Fix smp_mb__{before,after}_atomic()
On Wed, Apr 24, 2019 at 02:37:01PM +0200, Peter Zijlstra wrote:
> Recent probing at the Linux Kernel Memory Model uncovered a
> 'surprise'. Strongly ordered architectures where the atomic RmW
> primitive implies full memory ordering and
> smp_mb__{before,after}_atomic() are a simple barrier() (such as x86)
> fail for:
>
> *x = 1;
> atomic_inc(u);
> smp_mb__after_atomic();
> r0 = *y;
>
> Because, while the atomic_inc() implies memory order, it
> (surprisingly) does not provide a compiler barrier. This then allows
> the compiler to re-order like so:
>
> atomic_inc(u);
> *x = 1;
> smp_mb__after_atomic();
> r0 = *y;
>
> Which the CPU is then allowed to re-order (under TSO rules) like:
>
> atomic_inc(u);
> r0 = *y;
> *x = 1;
>
> And this very much was not intended. Therefore strengthen the atomic
> RmW ops to include a compiler barrier.
>
> NOTE: atomic_{or,and,xor} and the bitops already had the compiler
> barrier.
>
> Signed-off-by: Peter Zijlstra (Intel) <peterz@...radead.org>
> ---
> Documentation/atomic_t.txt | 3 +++
> arch/x86/include/asm/atomic.h | 8 ++++----
> arch/x86/include/asm/atomic64_64.h | 8 ++++----
> arch/x86/include/asm/barrier.h | 4 ++--
> 4 files changed, 13 insertions(+), 10 deletions(-)
>
> --- a/Documentation/atomic_t.txt
> +++ b/Documentation/atomic_t.txt
> @@ -194,6 +194,9 @@ These helper barriers exist because arch
> ordering on their SMP atomic primitives. For example our TSO architectures
> provide full ordered atomics and these barriers are no-ops.
>
> +NOTE: when the atomic RmW ops are fully ordered, they should also imply a
> +compiler barrier.
This is also the case for acquire and release variants, so we should
probably include those as well to avoid the implication that they don't
need the "memory" clobber.
> +
> Thus:
>
> atomic_fetch_add();
> --- a/arch/x86/include/asm/atomic.h
> +++ b/arch/x86/include/asm/atomic.h
> @@ -54,7 +54,7 @@ static __always_inline void arch_atomic_
> {
> asm volatile(LOCK_PREFIX "addl %1,%0"
> : "+m" (v->counter)
> - : "ir" (i));
> + : "ir" (i) : "memory");
> }
>
> /**
> @@ -68,7 +68,7 @@ static __always_inline void arch_atomic_
> {
> asm volatile(LOCK_PREFIX "subl %1,%0"
> : "+m" (v->counter)
> - : "ir" (i));
> + : "ir" (i) : "memory");
> }
>
> /**
> @@ -95,7 +95,7 @@ static __always_inline bool arch_atomic_
> static __always_inline void arch_atomic_inc(atomic_t *v)
> {
> asm volatile(LOCK_PREFIX "incl %0"
> - : "+m" (v->counter));
> + : "+m" (v->counter) :: "memory");
> }
> #define arch_atomic_inc arch_atomic_inc
>
> @@ -108,7 +108,7 @@ static __always_inline void arch_atomic_
> static __always_inline void arch_atomic_dec(atomic_t *v)
> {
> asm volatile(LOCK_PREFIX "decl %0"
> - : "+m" (v->counter));
> + : "+m" (v->counter) :: "memory");
> }
> #define arch_atomic_dec arch_atomic_dec
>
> --- a/arch/x86/include/asm/atomic64_64.h
> +++ b/arch/x86/include/asm/atomic64_64.h
> @@ -45,7 +45,7 @@ static __always_inline void arch_atomic6
> {
> asm volatile(LOCK_PREFIX "addq %1,%0"
> : "=m" (v->counter)
> - : "er" (i), "m" (v->counter));
> + : "er" (i), "m" (v->counter) : "memory");
> }
>
> /**
> @@ -59,7 +59,7 @@ static inline void arch_atomic64_sub(lon
> {
> asm volatile(LOCK_PREFIX "subq %1,%0"
> : "=m" (v->counter)
> - : "er" (i), "m" (v->counter));
> + : "er" (i), "m" (v->counter) : "memory");
> }
>
> /**
> @@ -87,7 +87,7 @@ static __always_inline void arch_atomic6
> {
> asm volatile(LOCK_PREFIX "incq %0"
> : "=m" (v->counter)
> - : "m" (v->counter));
> + : "m" (v->counter) : "memory");
> }
> #define arch_atomic64_inc arch_atomic64_inc
>
> @@ -101,7 +101,7 @@ static __always_inline void arch_atomic6
> {
> asm volatile(LOCK_PREFIX "decq %0"
> : "=m" (v->counter)
> - : "m" (v->counter));
> + : "m" (v->counter) : "memory");
> }
> #define arch_atomic64_dec arch_atomic64_dec
>
> --- a/arch/x86/include/asm/barrier.h
> +++ b/arch/x86/include/asm/barrier.h
> @@ -80,8 +80,8 @@ do { \
> })
>
> /* Atomic operations are already serializing on x86 */
> -#define __smp_mb__before_atomic() barrier()
> -#define __smp_mb__after_atomic() barrier()
> +#define __smp_mb__before_atomic() do { } while (0)
> +#define __smp_mb__after_atomic() do { } while (0)
I do wonder whether or not it would be cleaner to have a Kconfig option
such as ARCH_HAS_ORDERED_ATOMIC_RMW which, when selected, NOPs these
barrier macros out in the generic code. Then it's a requirement that
you have the "memory" clobber on your relaxed/non-returning atomics if
you select the option.
But that needn't hold this up, because your patch looks good to me modulo
the earlier, minor documentation nit.
Acked-by: Will Deacon <will.deacon@....com>
Will
Powered by blists - more mailing lists