[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <20140620220448.GC4615@linux.vnet.ibm.com>
Date: Fri, 20 Jun 2014 15:04:48 -0700
From: "Paul E. McKenney" <paulmck@...ux.vnet.ibm.com>
To: josh@...htriplett.org
Cc: linux-kernel@...r.kernel.org, mingo@...nel.org,
laijs@...fujitsu.com, dipankar@...ibm.com,
akpm@...ux-foundation.org, mathieu.desnoyers@...icios.com,
niv@...ibm.com, tglx@...utronix.de, peterz@...radead.org,
rostedt@...dmis.org, dhowells@...hat.com, edumazet@...gle.com,
dvhart@...ux.intel.com, fweisbec@...il.com, oleg@...hat.com,
sbw@....edu
Subject: Re: [PATCH tip/core/rcu 0/5] Fix for cond_resched performance
regression
On Fri, Jun 20, 2014 at 12:04:34PM -0700, josh@...htriplett.org wrote:
> On Fri, Jun 20, 2014 at 11:32:49AM -0700, Paul E. McKenney wrote:
> > Hello!
> >
> > This series contains changes to address the performance regressions
> > introduced by commit ac1bea85781e (Make cond_resched() report RCU
> > quiescent states), which was in turn fixing a problem where tasks looping
> > in the kernel could delay RCU grace periods. The changes in this series
> > are as follows:
> >
> > 1. Reduce the overhead of checking added to cond_resched() and friends.
> >
> > 2. Add a new cond_resched_rcu_qs() to provide RCU quiescent states
> > even if cond_resched() should stop doing so.
> >
> > 3. Add a new RCU_COND_RESCHED_QS to prevent cond_resched() from
> > reporting RCU quiescent states.
> >
> > 4. Prevent rcutorture testing from reporting spurious RCU CPU stall
> > warnings, and also to test RCU_COND_RESCHED_QS.
> >
> > 5. Provides a boot/sysfs rcutree.jiffies_till_cond_resched_qs
> > parameter to replace the magic "7".
>
> For all five patches:
>
> Reviewed-by: Josh Triplett <josh@...htriplett.org>
Thank you, added!
> Glad to see this doesn't add any overhead to rcutiny.
I suppose I should explain why that is...
First, single-CPU systems tend not to have thousands of mass-storage
devices, processes with many thousands of open files, or terabytes
of memory. Of course, in theory, a single-CPU system -could- have all
those things, but in practice thus far, they don't. Therefore, the
looping-in-the-kernel behavior that these things can cause simply don't
happen on single-CPU systems. Maybe some day they will, at which point
we can simply re-enable TREE_RCU for !SMP systems, so that those huge
single-CPU systems can use TREE_RCU, which has the needed protections.
Small embedded systems would of course still be able to benefit from
TINY_RCU.
In addition, single-CPU systems by definition have but on CPU. This
means that having a single runnable process on that CPU for tens of
seconds is much less likely, which eliminates another class of possible
indefinite-grace-period-extension bugs. In addition, the situations
where a bunch of CPUs "gang up" on a single CPU, generating endless
cleanup work for that CPU, also cannot happen on a single-CPU system.
This in turn eliminates the "grace-period extension via unending
cleanup" class of bugs.
Make sense?
Thanx, Paul
--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/
Powered by blists - more mailing lists