[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <Pine.LNX.4.44L0.1809121014050.1570-100000@iolanthe.rowland.org>
Date: Wed, 12 Sep 2018 10:24:37 -0400 (EDT)
From: Alan Stern <stern@...land.harvard.edu>
To: "Paul E. McKenney" <paulmck@...ux.vnet.ibm.com>
cc: Daniel Lustig <dlustig@...dia.com>,
Will Deacon <will.deacon@....com>,
Andrea Parri <andrea.parri@...rulasolutions.com>,
Andrea Parri <parri.andrea@...il.com>,
Kernel development list <linux-kernel@...r.kernel.org>,
<linux-arch@...r.kernel.org>, <mingo@...nel.org>,
<peterz@...radead.org>, <boqun.feng@...il.com>,
<npiggin@...il.com>, <dhowells@...hat.com>,
Jade Alglave <j.alglave@....ac.uk>,
Luc Maranget <luc.maranget@...ia.fr>, <akiyks@...il.com>,
Palmer Dabbelt <palmer@...ive.com>
Subject: Re: [PATCH RFC LKMM 1/7] tools/memory-model: Add extra ordering for
locks and remove it for ordinary release/acquire
On Tue, 11 Sep 2018, Paul E. McKenney wrote:
> > I think what you meant to write in the second and third sentences was
> > something more like this:
> >
> > Any code in an RCU critical section that extends beyond the
> > end of a given RCU grace period is guaranteed to see the
> > effects of all accesses which were visible to the grace
> > period's CPU before the start of the grace period. Similarly,
> > any code that follows an RCU grace period (on the grace
> > period's CPU) is guaranteed to see the effects of all accesses
> > which were visible to an RCU critical section that began
> > before the start of the grace period.
>
> That looks to me to be an improvement, other than that the "(on the
> grace period's CPU)" seems a bit restrictive -- you could for example
> have a release-acquire chain starting after the grace period, right?
The restriction was deliberate. Without the restriction people might
think it applies to any code that happens after a grace period, which
is not true (depending on what you mean by "happens after").
The business about a release-acquire chain is derivable from what I
wrote. Since the chain starts on the grace period's CPU, the effect is
guaranteed to be visible there. Then since release-acquire chains
preserve visibility, the effect is guaranteed to be visible to the code
at the end of the chain.
> > Also, the document doesn't seem to explain how Tree RCU relies on the
> > lock-ordering guarantees of raw_spin_lock_rcu_node() and friends. It
> > _says_ that these guarantees are used, but not how or where. (Unless I
> > missed something; I didn't read the document all that carefully.)
>
> The closest is this sentence: "But the only part of rcu_prepare_for_idle()
> that really matters for this discussion are lines 37–39", which
> refers to this code:
>
> 37 raw_spin_lock_rcu_node(rnp);
> 38 needwake = rcu_accelerate_cbs(rsp, rnp, rdp);
> 39 raw_spin_unlock_rcu_node(rnp);
>
> I could add a sentence explaining the importance of the
> smp_mb__after_unlock_lock() -- is that what you are getting at?
What I was really asking is for the document to explain what could go
wrong if the smp_mb__after_unlock_lock() call was omitted from
raw_spin_lock_rcu_node(). What assumptions or requirements in the Tree
RCU code might fail, and how/where in the code are these assumptions or
requirements used?
> > In any case, you should bear in mind that the lock ordering provided by
> > Peter's raw_spin_lock_rcu_node() and friends is not the same as what we
> > have been discussing for the LKMM:
> >
> > Peter's routines are meant for the case where you release
> > one lock and then acquire another (for example, locks in
> > two different levels of the RCU tree).
> >
> > The LKMM patch applies only to cases where one CPU releases
> > a lock and then that CPU or another acquires the _same_ lock
> > again.
> >
> > As another difference, the litmus test given near the start of the
> > "Tree RCU Grace Period Memory Ordering Building Blocks" section would
> > not be forbidden by the LKMM, even with RCtso locks, if it didn't use
> > raw_spin_lock_rcu_node(). This is because the litmus test is forbidden
> > only when locks are RCsc, which is what raw_spin_lock_rcu_node()
> > provides.
>
> Agreed.
>
> > So I don't see how the RCU code can be held up as an example either for
> > or against requiring locks to be RCtso.
>
> Agreed again. The use of smp_mb__after_unlock_lock() instead
> provides RCsc. But this use case is deemed sufficiently rare that
> smp_mb__after_unlock_lock() is defined within RCU.
So let me repeat the original question which started this train of
thought: Can anybody point to any actual code in the kernel that relies
on locks being RCtso, or that could be simplified if locks were
guaranteed to be RCtso?
There have been claims that such code exists somewhere in RCU and the
scheduler. Where is it?
Alan
Powered by blists - more mailing lists