[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <5350459F.5010903@hp.com>
Date: Thu, 17 Apr 2014 17:20:31 -0400
From: Waiman Long <waiman.long@...com>
To: Peter Zijlstra <peterz@...radead.org>
CC: Thomas Gleixner <tglx@...utronix.de>,
Ingo Molnar <mingo@...hat.com>,
"H. Peter Anvin" <hpa@...or.com>, linux-arch@...r.kernel.org,
x86@...nel.org, linux-kernel@...r.kernel.org,
virtualization@...ts.linux-foundation.org,
xen-devel@...ts.xenproject.org, kvm@...r.kernel.org,
Paolo Bonzini <paolo.bonzini@...il.com>,
Konrad Rzeszutek Wilk <konrad.wilk@...cle.com>,
"Paul E. McKenney" <paulmck@...ux.vnet.ibm.com>,
Rik van Riel <riel@...hat.com>,
Linus Torvalds <torvalds@...ux-foundation.org>,
Raghavendra K T <raghavendra.kt@...ux.vnet.ibm.com>,
David Vrabel <david.vrabel@...rix.com>,
Oleg Nesterov <oleg@...hat.com>,
Gleb Natapov <gleb@...hat.com>,
Scott J Norton <scott.norton@...com>,
Chegu Vinod <chegu_vinod@...com>
Subject: Re: [PATCH v9 03/19] qspinlock: Add pending bit
On 04/17/2014 11:42 AM, Peter Zijlstra wrote:
> On Thu, Apr 17, 2014 at 11:03:55AM -0400, Waiman Long wrote:
>> +/**
>> + * trylock_pending - try to acquire queue spinlock using the pending bit
>> + * @lock : Pointer to queue spinlock structure
>> + * @pval : Pointer to value of the queue spinlock 32-bit word
>> + * Return: 1 if lock acquired, 0 otherwise
>> + */
>> +static inline int trylock_pending(struct qspinlock *lock, u32 *pval)
>> +{
>> + u32 old, new, val = *pval;
> I'm not thrilled about you breaking this into a separate function; the
> compiler will put it right back and now you get to have that ugly
> pointer stuff.
>
> It also makes the function control flow not match the state diagram
> anymore.
I separate it out primarily to break the pending bit logic away from the
core MCS queuing logic to make each of them easier to understand as they
are kind of independent. I fully understand that the compiler will put
them back together. As I pile on more code, the slowpath function will
grow bigger making it harder to comprehend and find out where are the
boundary between them.
I will take a look at the state diagram to see what adjustment will be
needed.
>> +
>> + /*
>> + * trylock || pending
>> + *
>> + * 0,0,0 -> 0,0,1 ; trylock
>> + * 0,0,1 -> 0,1,1 ; pending
>> + */
>> + for (;;) {
>> + /*
>> + * If we observe any contention; queue.
>> + */
>> + if (val& ~_Q_LOCKED_MASK)
>> + return 0;
>> +
>> + new = _Q_LOCKED_VAL;
>> + if (val == new)
>> + new |= _Q_PENDING_VAL;
>> +
>> + old = atomic_cmpxchg(&lock->val, val, new);
>> + if (old == val)
>> + break;
>> +
>> + *pval = val = old;
>> + }
>> +
>> + /*
>> + * we won the trylock
>> + */
>> + if (new == _Q_LOCKED_VAL)
>> + return 1;
>> +
>> + /*
>> + * we're pending, wait for the owner to go away.
>> + *
>> + * *,1,1 -> *,1,0
>> + */
>> + while ((val = atomic_read(&lock->val))& _Q_LOCKED_MASK)
>> + arch_mutex_cpu_relax();
> That was a cpu_relax().
Yes, but arch_mutex_cpu_relax() is the same as cpu_relax() for x86.
-Longman
--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/
Powered by blists - more mailing lists