[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <20080916173012.GC6717@linux.vnet.ibm.com>
Date: Tue, 16 Sep 2008 10:30:12 -0700
From: "Paul E. McKenney" <paulmck@...ux.vnet.ibm.com>
To: Manfred Spraul <manfred@...orfullife.com>
Cc: linux-kernel@...r.kernel.org, cl@...ux-foundation.org,
mingo@...e.hu, akpm@...ux-foundation.org, dipankar@...ibm.com,
josht@...ux.vnet.ibm.com, schamp@....com, niv@...ibm.com,
dvhltc@...ibm.com, ego@...ibm.com, laijs@...fujitsu.com,
rostedt@...dmis.org, peterz@...radead.org, penberg@...helsinki.fi,
andi@...stfloor.org
Subject: Re: [PATCH, RFC] v4 scalable classic RCU implementation
On Tue, Sep 16, 2008 at 06:52:54PM +0200, Manfred Spraul wrote:
> Hi Paul,
Hello, Manfred!
Thank you for looking this over!
> Paul E. McKenney wrote:
>> +/*
>> + * Scan the leaf rcu_node structures, processing dyntick state for any
>> that
>> + * have not yet encountered a quiescent state, using the function
>> specified.
>> + * Returns 1 if the current grace period ends while scanning (possibly
>> + * because we made it end).
>> + */
>> +static int rcu_process_dyntick(struct rcu_state *rsp, long lastcomp,
>> + int (*f)(struct rcu_data *))
>> +{
>> + unsigned long bit;
>> + int cpu;
>> + unsigned long flags;
>> + unsigned long mask;
>> + struct rcu_node *rnp_cur = rsp->level[NUM_RCU_LVLS - 1];
>> + struct rcu_node *rnp_end = &rsp->node[NUM_RCU_NODES];
>> +
>> + for (; rnp_cur < rnp_end; rnp_cur++) {
>> + mask = 0;
>> + spin_lock_irqsave(&rnp_cur->lock, flags);
>> + if (rsp->completed != lastcomp) {
>> + spin_unlock_irqrestore(&rnp_cur->lock, flags);
>> + return 1;
>> + }
>> + if (rnp_cur->qsmask == 0) {
>> + spin_unlock_irqrestore(&rnp_cur->lock, flags);
>> + continue;
>> + }
>> + cpu = rnp_cur->grplo;
>> + bit = 1;
>> + mask = 0;
>> + for (; cpu <= rnp_cur->grphi; cpu++, bit <<= 1) {
>> + if ((rnp_cur->qsmask & bit) != 0 && f(rsp->rda[cpu]))
>> + mask |= bit;
>> + }
>>
> I'm still comparing my implementation with your code:
> - f is called once for each cpu in the system, correct?
Not necessarily. If all CPUs corresponding to this rcu_state struct
have checked in already, we don't even get to this loop -- see the
"continue" above.
> - if at least one cpu is in nohz mode, this loop will be needed for every
> grace period.
The outer loop, yes. The inner loop only for those rcu_state structs
that have at least one CPU in nohz mode.
> That means an O(NR_CPUS) loop with disabled local interrupts :-(
> Is that correct?
With the definition of "O()" being the worst-case execution time, yes.
But this worst case could only happen when the system was mostly idle,
in which case the added overhead should not be too horribly bad. If the
system was busy enough that each CPU ran at least one process during each
grace period, then this function would not be invoked in the first place.
If this does prove to be a problem in practice, I will rework
force_quiescent_state() to run incrementally. But I would rather
avoid both the added complexity and the resulting longer grace periods,
so someone needs to bring me a real-world problem before I take that
approach.
> Unfortunately, my solution is even worse:
> My rcu_irq_exit() acquires a global spinlock when called on a nohz cpus.
> A few cpus in cpu_idle, nohz, executing 50k network interrupts/sec would
> cacheline-trash that spinlock.
> I'm considering counting interrupts: if a nohz cpu executes more than a few
> interrupts/tick, then add a timer that check rcu_pending().
I tried putting a cpu_quiet() in my rcu_irq_exit() as well, and quickly
decided that this was counter-productive. ;-)
> Perhaps even wouldn't be enough: I remember that the initial unhandled irq
> detection code broke miserably on large SGI systems:
> An atomic_inc(&global_var) in the local timer interrupt (i.e.: NR_CPUS*HZ
> calls/sec) caused so severe trashing that the system wouldn't boot. IIRC
> that was with 512 cpus.
/me runs off and checks to make sure that all of my dyntick entry/exit
code restricts itself to per-CPU variables...
Yep! (Whew!!!)
Thanx, Paul
--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/
Powered by blists - more mailing lists