[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <20110330081716.GE17523@htj.dyndns.org>
Date: Wed, 30 Mar 2011 10:17:16 +0200
From: Tejun Heo <tj@...nel.org>
To: Peter Zijlstra <peterz@...radead.org>
Cc: Ingo Molnar <mingo@...hat.com>,
Linus Torvalds <torvalds@...ux-foundation.org>,
Andrew Morton <akpm@...ux-foundation.org>,
Chris Mason <chris.mason@...cle.com>,
linux-kernel@...r.kernel.org, linux-btrfs@...r.kernel.org
Subject: Re: [PATCH 2/2] mutex: Apply adaptive spinning on mutex_trylock()
Hey, Peter.
On Tue, Mar 29, 2011 at 07:37:33PM +0200, Peter Zijlstra wrote:
> On Tue, 2011-03-29 at 19:09 +0200, Tejun Heo wrote:
> > Here's the combined patch I was planning on testing but didn't get to
> > (yet). It implements two things - hard limit on spin duration and
> > early break if the owner also is spinning on a mutex.
>
> This is going to give massive conflicts with
>
> https://lkml.org/lkml/2011/3/2/286
> https://lkml.org/lkml/2011/3/2/282
>
> which I was planning to stuff into .40
I see. Adapting shouldn't be hard. The patch is in proof-of-concept
stage anyway.
> > + * Forward progress is guaranteed regardless of locking ordering by never
> > + * spinning longer than MAX_MUTEX_SPIN_NS. This is necessary because
> > + * mutex_trylock(), which doesn't have to follow the usual locking
> > + * ordering, also uses this function.
>
> While that puts a limit on things it'll still waste time. I'd much
> rather pass an trylock argument to mutex_spin_on_owner() and then bail
> on owner also spinning.
Do we guarantee or enforce that the lock ownership can't be
transferred to a different task? If we do, the recursive spinning
detection is enough to guarantee forward progress.
> > + if (task_thread_info(rq->curr) != owner ||
> > + rq->spinning_on_mutex || need_resched() ||
> > + local_clock() > start + MAX_MUTEX_SPIN_NS) {
>
> While we did our best with making local_clock() cheap, I'm still fairly
> uncomfortable with putting it in such a tight loop.
That's one thing I didn't really understand. It seems the spinning
code tried to be light on CPU cycle usage, but we're wasting CPU
cycles there anyway. If the spinning can become smarter using some
CPU cycles, isn't that a gain? Why is the spinning code optimizing
for less CPU cycles?
Also, it would be great if you can share what you're using for locking
performance measurements. My attempts with dbench didn't end too
well. :-(
Thanks.
--
tejun
--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/
Powered by blists - more mailing lists