[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <20130814102041.GG10849@gmail.com>
Date: Wed, 14 Aug 2013 12:20:41 +0200
From: Ingo Molnar <mingo@...nel.org>
To: Waiman Long <waiman.long@...com>
Cc: Thomas Gleixner <tglx@...utronix.de>,
Ingo Molnar <mingo@...hat.com>,
"H. Peter Anvin" <hpa@...or.com>, Arnd Bergmann <arnd@...db.de>,
linux-arch@...r.kernel.org, x86@...nel.org,
linux-kernel@...r.kernel.org,
Peter Zijlstra <peterz@...radead.org>,
Steven Rostedt <rostedt@...dmis.org>,
Andrew Morton <akpm@...ux-foundation.org>,
Greg Kroah-Hartman <gregkh@...uxfoundation.org>,
Matt Fleming <matt.fleming@...el.com>,
Michel Lespinasse <walken@...gle.com>,
Andi Kleen <andi@...stfloor.org>,
Rik van Riel <riel@...hat.com>,
"Paul E. McKenney" <paulmck@...ux.vnet.ibm.com>,
Linus Torvalds <torvalds@...ux-foundation.org>,
Raghavendra K T <raghavendra.kt@...ux.vnet.ibm.com>,
George Spelvin <linux@...izon.com>,
Harvey Harrison <harvey.harrison@...il.com>,
"Chandramouleeswaran, Aswin" <aswin@...com>,
"Norton, Scott J" <scott.norton@...com>
Subject: Re: [PATCH v3 0/3] qrwlock: Introducing a queue read/write lock
implementation
* Waiman Long <waiman.long@...com> wrote:
> On 07/31/2013 08:00 PM, Waiman Long wrote:
> >v2->v3:
> > - Make read lock stealing the default and fair rwlock an option with
> > a different initializer.
> > - In queue_read_lock_slowpath(), check irq_count() and force spinning
> > and lock stealing in interrupt context.
> > - Unify the fair and classic read-side code path, and make write-side
> > to use cmpxchg with 2 different writer states. This slows down the
> > write lock fastpath to make the read side more efficient, but is
> > still slightly faster than a spinlock.
> >
> >v1->v2:
> > - Improve lock fastpath performance.
> > - Optionally provide classic read/write lock behavior for backward
> > compatibility.
> > - Use xadd instead of cmpxchg for fair reader code path to make it
> > immute to reader contention.
> > - Run more performance testing.
> >
> >As mentioned in the LWN article http://lwn.net/Articles/364583/, the
> >classic read/write lock suffer from an unfairness problem that it is
> >possible for a stream of incoming readers to block a waiting writer
> >from getting the lock for a long time. Also, a waiting reader/writer
> >contending a rwlock in local memory will have a higher chance of
> >acquiring the lock than a reader/writer in remote node.
> >
> >This patch set introduces a queue-based read/write lock implementation
> >that can largely solve this unfairness problem if the lock owners
> >choose to use the fair variant of the lock. The queue rwlock has two
> >variants selected at initialization time - classic (with read lock
> >stealing) and fair (to both readers and writers). The classic rwlock
> >is the default.
> >
> >The read lock slowpath will check if the reader is in an interrupt
> >context. If so, it will force lock spinning and stealing without
> >waiting in a queue. This is to ensure the read lock will be granted
> >as soon as possible.
> >
> >Even the classic rwlock is fairer than the current version as there
> >is a higher chance for writers to get the lock and is fair among
> >the writers.
> >
> >The queue write lock can also be used as a replacement for ticket
> >spinlocks that are highly contended if lock size increase is not
> >an issue.
> >
> >There is no change in the interface. By just selecting the QUEUE_RWLOCK
> >config parameter during the configuration phase, the classic read/write
> >lock will be replaced by the new queue read/write lock. This will
> >made the systems more deterministic and faster in lock contention
> >situations. In uncontended cases, the queue read/write lock may be
> >a bit slower than the classic one depending on the exact mix of read
> >and write locking primitives. Given the fact that locking overhead is
> >typically a very small percentage of the total CPU time in uncontended
> >cases, there won't be any noticeable degradation in performance with
> >this replacement.
> >
> >This patch set currently provides queue read/write lock support on
> >x86 architecture only. Support for other architectures can be added
> >later on once proper testing is done.
> >
> >Signed-off-by: Waiman Long<Waiman.Long@...com>
> >
> >Waiman Long (3):
> > qrwlock: A queue read/write lock implementation
> > qrwlock x86: Enable x86 to use queue read/write lock
> > qrwlock: Enable fair queue read/write lock behavior
> >
> > arch/x86/Kconfig | 3 +
> > arch/x86/include/asm/spinlock.h | 2 +
> > arch/x86/include/asm/spinlock_types.h | 4 +
> > include/asm-generic/qrwlock.h | 239 ++++++++++++++++++++++++++++++++
> > include/linux/rwlock.h | 15 ++
> > include/linux/rwlock_types.h | 13 ++
> > lib/Kconfig | 23 +++
> > lib/Makefile | 1 +
> > lib/qrwlock.c | 242 +++++++++++++++++++++++++++++++++
> > lib/spinlock_debug.c | 19 +++
> > 10 files changed, 561 insertions(+), 0 deletions(-)
> > create mode 100644 include/asm-generic/qrwlock.h
> > create mode 100644 lib/qrwlock.c
>
> I would like to share with you a rwlock related system crash that I
> encountered during my testing with hackbench on an 80-core DL980. The
> kernel crash because of a "watchdog detected hard lockup on cpu 79". The
> crashing CPU was running "write_lock_irq(&tasklist_lock)" in
> forget_original_parent() of the exit code path when I interrupted the
> hackbench which was spawning thousands of processes. Apparently, the
> remote CPU was not able to get the lock for a sufficient long time due
> to the unfairness of the rwlock which I think my version of queue rwlock
> will be able to alleviate this issue.
>
> So far, I was not able to reproduce the crash. I will try to see if I
> could more consistently reproduce it.
Was it an actual crash/lockup, or a longish hang followed by a lock
detector splat followed by the system eventually recovering back to
working order?
Thanks,
Ingo
--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/
Powered by blists - more mailing lists