[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <20070817185639.GF8464@linux.vnet.ibm.com>
Date: Fri, 17 Aug 2007 11:56:39 -0700
From: "Paul E. McKenney" <paulmck@...ux.vnet.ibm.com>
To: Satyam Sharma <satyam@...radead.org>
Cc: Herbert Xu <herbert@...dor.apana.org.au>,
Stefan Richter <stefanr@...6.in-berlin.de>,
Paul Mackerras <paulus@...ba.org>,
Christoph Lameter <clameter@....com>,
Chris Snook <csnook@...hat.com>,
Linux Kernel Mailing List <linux-kernel@...r.kernel.org>,
linux-arch@...r.kernel.org,
Linus Torvalds <torvalds@...ux-foundation.org>,
netdev@...r.kernel.org, Andrew Morton <akpm@...ux-foundation.org>,
ak@...e.de, heiko.carstens@...ibm.com, davem@...emloft.net,
schwidefsky@...ibm.com, wensong@...ux-vs.org, horms@...ge.net.au,
wjiang@...ilience.com, cfriesen@...tel.com, zlynx@....org,
rpjday@...dspring.com, jesper.juhl@...il.com,
segher@...nel.crashing.org
Subject: Re: [PATCH 0/24] make atomic_read() behave consistently across all architectures
On Sat, Aug 18, 2007 at 12:01:38AM +0530, Satyam Sharma wrote:
>
>
> On Fri, 17 Aug 2007, Paul E. McKenney wrote:
>
> > On Fri, Aug 17, 2007 at 01:09:08PM +0530, Satyam Sharma wrote:
> > >
> > > On Thu, 16 Aug 2007, Paul E. McKenney wrote:
> > >
> > > > On Fri, Aug 17, 2007 at 07:59:02AM +0800, Herbert Xu wrote:
> > > > >
> > > > > First of all, I think this illustrates that what you want
> > > > > here has nothing to do with atomic ops. The ORDERED_WRT_IRQ
> > > > > macro occurs a lot more times in your patch than atomic
> > > > > reads/sets. So *assuming* that it was necessary at all,
> > > > > then having an ordered variant of the atomic_read/atomic_set
> > > > > ops could do just as well.
> > > >
> > > > Indeed. If I could trust atomic_read()/atomic_set() to cause the compiler
> > > > to maintain ordering, then I could just use them instead of having to
> > > > create an ORDERED_WRT_IRQ(). (Or ACCESS_ONCE(), as it is called in a
> > > > different patch.)
> > >
> > > +#define WHATEVER(x) (*(volatile typeof(x) *)&(x))
> > > [...]
> > > Also, this gives *zero* "re-ordering" guarantees that your code wants
> > > as you've explained it below) -- neither w.r.t. CPU re-ordering (which
> > > probably you don't care about) *nor* w.r.t. compiler re-ordering
> > > (which you definitely _do_ care about).
> >
> > You are correct about CPU re-ordering (and about the fact that this
> > example doesn't care about it), but not about compiler re-ordering.
> >
> > The compiler is prohibited from moving a volatile access across a sequence
> > point. One example of a sequence point is a statement boundary. Because
> > all of the volatile accesses in this code are separated by statement
> > boundaries, a conforming compiler is prohibited from reordering them.
>
> Yes, you're right, and I believe precisely this was discussed elsewhere
> as well today.
>
> But I'd call attention to what Herbert mentioned there. You're using
> ORDERED_WRT_IRQ() on stuff that is _not_ defined to be an atomic_t at all:
>
> * Member "completed" of struct rcu_ctrlblk is a long.
> * Per-cpu variable rcu_flipctr is an array of ints.
> * Members "rcu_read_lock_nesting" and "rcu_flipctr_idx" of
> struct task_struct are ints.
>
> So are you saying you're "having to use" this volatile-access macro
> because you *couldn't* declare all the above as atomic_t and thus just
> expect the right thing to happen by using the atomic ops API by default,
> because it lacks volatile access semantics (on x86)?
>
> If so, then I wonder if using the volatile access cast is really the
> best way to achieve (at least in terms of code clarity) the kind of
> re-ordering guarantees it wants there. (there could be alternative
> solutions, such as using barrier(), or that at bottom of this mail)
>
> What I mean is this: If you switch to atomic_t, and x86 switched to
> make atomic_t have "volatile" semantics by default, the statements
> would be simply a string of: atomic_inc(), atomic_add(), atomic_set(),
> and atomic_read() statements, and nothing in there that clearly makes
> it *explicit* that the code is correct (and not buggy) simply because
> of the re-ordering guarantees that the C "volatile" type-qualifier
> keyword gives us as per the standard. But now we're firmly in
> "subjective" territory, so you or anybody could legitimately disagree.
In any case, given Linus's note, it appears that atomic_read() and
atomic_set() won't consistently have volatile semantics, at least
not while the compiler generates such ugly code for volatile accesses.
So I will continue with my current approach.
In any case, I will not be using atomic_inc() or atomic_add() in this
code, as doing so would more than double the overhead, even on machines
that are the most efficient at implementing atomic operations.
> > > > Suppose I tried replacing the ORDERED_WRT_IRQ() calls with
> > > > atomic_read() and atomic_set(). Starting with __rcu_read_lock():
> > > >
> > > > o If "ORDERED_WRT_IRQ(__get_cpu_var(rcu_flipctr)[idx])++"
> > > > was ordered by the compiler after
> > > > "ORDERED_WRT_IRQ(me->rcu_read_lock_nesting) = nesting + 1", then
> > > > suppose an NMI/SMI happened after the rcu_read_lock_nesting but
> > > > before the rcu_flipctr.
> > > >
> > > > Then if there was an rcu_read_lock() in the SMI/NMI
> > > > handler (which is perfectly legal), the nested rcu_read_lock()
> > > > would believe that it could take the then-clause of the
> > > > enclosing "if" statement. But because the rcu_flipctr per-CPU
> > > > variable had not yet been incremented, an RCU updater would
> > > > be within its rights to assume that there were no RCU reads
> > > > in progress, thus possibly yanking a data structure out from
> > > > under the reader in the SMI/NMI function.
> > > >
> > > > Fatal outcome. Note that only one CPU is involved here
> > > > because these are all either per-CPU or per-task variables.
> > >
> > > Ok, so you don't care about CPU re-ordering. Still, I should let you know
> > > that your ORDERED_WRT_IRQ() -- bad name, btw -- is still buggy. What you
> > > want is a full compiler optimization barrier().
> >
> > No. See above.
>
> True, *(volatile foo *)& _will_ work for this case.
>
> But multiple calls to barrier() (granted, would invalidate all other
> optimizations also) would work as well, would it not?
They work, but are a bit slower. So they do work, but not as well.
> [ Interestingly, if you declared all those objects mentioned earlier as
> atomic_t, and x86(-64) switched to an __asm__ __volatile__ based variant
> for atomic_{read,set}_volatile(), the bugs you want to avoid would still
> be there. "volatile" the C language type-qualifier does have compiler
> re-ordering semantics you mentioned earlier, but the "volatile" that
> applies to inline asm()s gives no re-ordering guarantees. ]
Well, that certainly would be a point in favor of "volatile" over inline
asms. ;-)
> > > > o If "ORDERED_WRT_IRQ(me->rcu_read_lock_nesting) = nesting + 1"
> > > > was ordered by the compiler to follow the
> > > > "ORDERED_WRT_IRQ(me->rcu_flipctr_idx) = idx", and an NMI/SMI
> > > > happened between the two, then an __rcu_read_lock() in the NMI/SMI
> > > > would incorrectly take the "else" clause of the enclosing "if"
> > > > statement. If some other CPU flipped the rcu_ctrlblk.completed
> > > > in the meantime, then the __rcu_read_lock() would (correctly)
> > > > write the new value into rcu_flipctr_idx.
> > > >
> > > > Well and good so far. But the problem arises in
> > > > __rcu_read_unlock(), which then decrements the wrong counter.
> > > > Depending on exactly how subsequent events played out, this could
> > > > result in either prematurely ending grace periods or never-ending
> > > > grace periods, both of which are fatal outcomes.
> > > >
> > > > And the following are not needed in the current version of the
> > > > patch, but will be in a future version that either avoids disabling
> > > > irqs or that dispenses with the smp_read_barrier_depends() that I
> > > > have 99% convinced myself is unneeded:
> > > >
> > > > o nesting = ORDERED_WRT_IRQ(me->rcu_read_lock_nesting);
> > > >
> > > > o idx = ORDERED_WRT_IRQ(rcu_ctrlblk.completed) & 0x1;
> > > >
> > > > Furthermore, in that future version, irq handlers can cause the same
> > > > mischief that SMI/NMI handlers can in this version.
>
> So don't remove the local_irq_save/restore, which is well-established and
> well-understood for such cases (it doesn't help you with SMI/NMI,
> admittedly). This isn't really about RCU or per-cpu vars as such, it's
> just about racy code where you don't want to get hit by a concurrent
> interrupt (it does turn out that doing things in a _particular order_ will
> not cause fatal/buggy behaviour, but it's still a race issue, after all).
The local_irq_save/restore are something like 30% of the overhead of
these two functions, so will be looking hard at getting rid of them.
Doing so allows the scheduling-clock interrupt to get into the mix,
and also allows preemption. The goal would be to find some trick that
suppresses preemption, fends off the grace-period-computation code
invoked from the the scheduling-clock interrupt, and otherwise keeps
things on an even keel.
> > > > Next, looking at __rcu_read_unlock():
> > > >
> > > > o If "ORDERED_WRT_IRQ(me->rcu_read_lock_nesting) = nesting - 1"
> > > > was reordered by the compiler to follow the
> > > > "ORDERED_WRT_IRQ(__get_cpu_var(rcu_flipctr)[idx])--",
> > > > then if an NMI/SMI containing an rcu_read_lock() occurs between
> > > > the two, this nested rcu_read_lock() would incorrectly believe
> > > > that it was protected by an enclosing RCU read-side critical
> > > > section as described in the first reversal discussed for
> > > > __rcu_read_lock() above. Again, fatal outcome.
> > > >
> > > > This is what we have now. It is not hard to imagine situations that
> > > > interact with -both- interrupt handlers -and- other CPUs, as described
> > > > earlier.
>
> Unless somebody's going for a lockless implementation, such situations
> normally use spin_lock_irqsave() based locking (or local_irq_save for
> those who care only for current CPU) -- problem with the patch in question,
> is that you want to prevent races with concurrent SMI/NMIs as well, which
> is not something that a lot of code needs to consider.
Or that needs to resolve similar races with IRQs without disabling them.
One reason to avoid disabling IRQs is to avoid degrading scheduling
latency. In any case, I do agree that the amount of code that must
worry about this is quite small at the moment. I believe that it
will become more common, but would imagine that this belief might not
be universal. Yet, anyway. ;-)
> [ Curiously, another thread is discussing something similar also:
> http://lkml.org/lkml/2007/8/15/393 "RFC: do get_rtc_time() correctly" ]
>
> Anyway, I didn't look at the code in that patch very much in detail, but
> why couldn't you implement some kind of synchronization variable that lets
> rcu_read_lock() or rcu_read_unlock() -- when being called from inside an
> NMI or SMI handler -- know that it has concurrently interrupted an ongoing
> rcu_read_{un}lock() and so must do things differently ... (?)
Given some low-level details of the current implementation, I could
imagine manipulating rcu_read_lock_nesting on entry to and exit from
all NMI/SMI handlers, but would like to avoid that kind of architecture
dependency. I am not confident of locating all of them, for one thing...
> I'm also wondering if there's other code that's not using locking in the
> kernel that faces similar issues, and what they've done to deal with it
> (if anything). Such bugs would be subtle, and difficult to diagnose.
Agreed!
Thanx, Paul
-
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/
Powered by blists - more mailing lists