[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <Pine.LNX.4.64.1210191758570.12753@file.rdu.redhat.com>
Date: Fri, 19 Oct 2012 18:54:41 -0400 (EDT)
From: Mikulas Patocka <mpatocka@...hat.com>
To: Peter Zijlstra <peterz@...radead.org>
cc: Oleg Nesterov <oleg@...hat.com>,
"Paul E. McKenney" <paulmck@...ux.vnet.ibm.com>,
Linus Torvalds <torvalds@...ux-foundation.org>,
Ingo Molnar <mingo@...e.hu>,
Srikar Dronamraju <srikar@...ux.vnet.ibm.com>,
Ananth N Mavinakayanahalli <ananth@...ibm.com>,
Anton Arapov <anton@...hat.com>, linux-kernel@...r.kernel.org,
Thomas Gleixner <tglx@...utronix.de>
Subject: Re: [PATCH 1/2] brw_mutex: big read-write mutex
On Fri, 19 Oct 2012, Peter Zijlstra wrote:
> > Yes, I tried this approach - it involves doing LOCK instruction on read
> > lock, remembering the cpu and doing another LOCK instruction on read
> > unlock (which will hopefully be on the same CPU, so no cacheline bouncing
> > happens in the common case). It was slower than the approach without any
> > LOCK instructions (43.3 seconds seconds for the implementation with
> > per-cpu LOCKed access, 42.7 seconds for this implementation without atomic
> > instruction; the benchmark involved doing 512-byte direct-io reads and
> > writes on a ramdisk with 8 processes on 8-core machine).
>
> So why is that a problem? Surely that's already tons better then what
> you've currently got.
Percpu rw-semaphores do not improve performance at all. I put them there
to avoid performance regression, not to improve performance.
All Linux kernels have a race condition - when you change block size of a
block device and you read or write the device at the same time, a crash
may happen. This bug is there since ever. Recently, this bug started to
cause major trouble - multiple high profile business sites report crashes
because of this race condition.
You can fix this race by using a read lock around I/O paths and write lock
around block size changing, but normal rw semaphore cause cache line
bouncing when taken for read by multiple processors and I/O performance
degradation because of it is measurable.
So I put this percpu-rw-semaphore there to fix the crashes and minimize
performance impact - on x86 it doesn't take any interlocked instructions
in the read path.
I don't quite understand why are people opposing to this and what do they
want to do instead? If you pull percpu-rw-semaphores out of the kernel,
you introduce a performance regression (raw device i/o will be slower on
3.7 than on 3.6, because on 3.6 it doesn't take any lock at all and on 3.7
it takes a read lock).
So you have options:
1) don't lock i/o just like on 3.6 and previous versions - you get a fast
kernel that randomly crashes
2) lock i/o with normal rw semaphore - you get a kernel that doesn't
crash, but that is slower than previous versions
3) lock i/o with percpu rw semaphore - you get kernel that is almost as
fast as previous kernels and that doesn't crash
For the users, the option 3) is the best. The users don't care whether it
looks ugly or not, they care about correctness and performance, that's
all.
Obviously, you can improve rw semaphores by adding lockdep annotations, or
by other things (turning rcu_read_lock/sychronize_rcu into
preempt_disable/synchronize_sched, by using barrier()-synchronize_sched()
instead of smp_mb()...), but I don't see a reason why do you want to hurt
users' experience by pulling it out and reverting to state 1) or 2) and
then, two kernel cycles later, come up with percpu-rw-semaphores again.
Mikulas
--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/
Powered by blists - more mailing lists