[<prev] [next>] [<thread-prev] [day] [month] [year] [list]
Message-ID: <4F356FA6.4000104@redhat.com>
Date: Fri, 10 Feb 2012 11:27:34 -0800
From: Richard Henderson <rth@...hat.com>
To: Linus Torvalds <torvalds@...ux-foundation.org>
CC: Andrew MacLeod <amacleod@...hat.com>, paulmck@...ux.vnet.ibm.com,
Torvald Riegel <triegel@...hat.com>, Jan Kara <jack@...e.cz>,
LKML <linux-kernel@...r.kernel.org>, linux-ia64@...r.kernel.org,
dsterba@...e.cz, ptesarik@...e.cz, rguenther@...e.de,
gcc@....gnu.org
Subject: Re: Memory corruption due to word sharing
On 02/03/2012 12:00 PM, Linus Torvalds wrote:
> do {
> load-link %r,%m
> if (r == value)
> return 0;
> add
> } while (store-conditional %r,%m)
> return 1;
>
> and it is used to implement two *very* common (and critical)
> reference-counting use cases:
>
> - decrement ref-count locklessly, and if it hits free, take a lock
> atomically (ie "it would be a bug for anybody to ever see it decrement
> to zero while holding the lock")
>
> - increment ref-count locklessly, but if we race with the final free,
> don't touch it, because it's gone (ie "if somebody decremented it to
> zero while holding the lock, we absolutely cannot increment it again")
>
> They may sound odd, but those two operations are absolutely critical
> for most lock-less refcount management. And taking locks for reference
> counting is absolutely death to performance, and is often impossible
> anyway (the lock may be a local lock that is *inside* the structure
> that is being reference-counted, so if the refcount is zero, then you
> cannot even look at the lock!)
>
> In the above example, the load-locked -> store-conditional would
> obviously be a cmpxchg loop instead on architectures that do cmpxchg
> instead of ll/sc.
>
> Of course, it you expose some intrinsic for the whole "ll/sc" model
> (and you then turn it into cmpxchg on demand), we could literally
> open-code it.
We can't expose the ll/sc model directly, due to various target-specific
hardware constraints (e.g. no other memory references, i.e. no spills).
But the "weak" compare-exchange is sorta-kinda close.
A "weak" compare-exchange is a ll/sc pair, without the loop for restart
on sc failure. Thus a weak compare-exchange can fail for arbitrary reasons.
You do, however, get back the current value in the memory, so you can loop
back yourself and try again.
So that loop above becomes the familiar expression as for CAS:
r = *m;
do {
if (r == value)
return;
n = r + inc;
} while (!atomic_compare_exchange(m, &r, n, /*weak=*/true,
__ATOMIC_RELAXED, __ATOMIC_RELAXED));
which, unlike with the current __sync_val_compare_and_swap, does not
result in a nested loop for the ll/sc architectures.
Yes, the ll/sc architectures can technically do slightly better by writing
this as inline asm, but the gain is significantly less than before. Given
that the line must be in L1 cache already, the redundant load from M in
the ll insn should be nearly negligible.
Of course for architectures that implement cas, there is no difference
between "weak" and "strong".
r~
--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/
Powered by blists - more mailing lists