[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <472B3316.2070802@gmail.com>
Date: Fri, 02 Nov 2007 10:24:22 -0400
From: Gregory Haskins <gregory.haskins.ml@...il.com>
To: Linus Torvalds <torvalds@...ux-foundation.org>
CC: Gregory Haskins <gregory.haskins@...il.com>,
Nick Piggin <npiggin@...e.de>,
Linux Kernel Mailing List <linux-kernel@...r.kernel.org>,
Andi Kleen <ak@...e.de>, Ingo Molnar <mingo@...e.hu>
Subject: Re: [patch 1/4] x86: FIFO ticket spinlocks
Linus Torvalds wrote:
>
> On Thu, 1 Nov 2007, Gregory Haskins wrote:
>> I had observed this phenomenon on some 8-ways here as well, but I didn't
>> have the bandwidth to code something up. Thumbs up!
>
> Can you test under interesting loads?
Sure thing. Ill try this next week.
>
> We're interested in:
> - is the unfairness fix really noticeable (or does it just move the
> problem somewhere else, and there is no real change in behaviour)
> - what is the performance impact?
>
> In particular, unfair spinlocks have the potential to perform much better.
I see where you are going here, and I mostly agree. I think the key is
that "given equal contention, let the guy with the hottest cache win".
The problem with the current implementation is that the spinlocks have
no way to gauge the details of the contention. They can only gauge
instantaneous snapshots of state as viewed by each TSL invocation, which
effectively resets your position each time.
On the flip side, Nick's patches take the opposite extreme. If a lock
is contended, get in line. ;) This has the desirable property of
avoiding starvation. However, it will also tend to cause more bouncing
since you are virtually guaranteed not to re-win the contended lock, as
you point out next.
> Not so much because the spinlock itself acts all that differently, but
> because being unfair also fundmanetally tends to keep the data structures
> that are *protected* by the spinlock on just one CPU.
My issue here is that this behavior can also be described as precisely
part of the problem being addressed: That is, both CPUs presumably
*want/need* access to the data or they wouldn't be taking the spinlock
to begin with. So its really not a question of keeping the structures
on one cpu per se (at least, not for unbounded durations or the system
won't operate properly).
Rather, I think the key is to minimize the impact by bouncing things
intelligently. ;) I.e. If all things are equal, favor the hottest task
so the data only bounces once instead of twice. Outside of this
condition, operate strict FIFO. If we can reasonably calculate when
this optimization is possible, we will have the best of both worlds. I
have some ideas about ways to extend Nicks algorithm to support this
which I will submit ASAP.
I think the rest of what you said is very fair: Prove that it's a
problem, this concept helps, and we don't make things worse ;)
Will do, ASAP.
Regards,
-Greg
-
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/
Powered by blists - more mailing lists