[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <492E90BC.1090208@gmail.com>
Date: Thu, 27 Nov 2008 14:21:16 +0200
From: Török Edwin <edwintorok@...il.com>
To: Nick Piggin <npiggin@...e.de>
CC: Mike Waychison <mikew@...gle.com>, Ying Han <yinghan@...gle.com>,
Ingo Molnar <mingo@...e.hu>, linux-mm@...ck.org,
linux-kernel@...r.kernel.org, akpm <akpm@...ux-foundation.org>,
David Rientjes <rientjes@...gle.com>,
Rohit Seth <rohitseth@...gle.com>,
Hugh Dickins <hugh@...itas.com>,
Peter Zijlstra <a.p.zijlstra@...llo.nl>,
"H. Peter Anvin" <hpa@...or.com>
Subject: Re: [RFC v1][PATCH]page_fault retry with NOPAGE_RETRY
On 2008-11-27 14:03, Nick Piggin wrote:
> On Thu, Nov 27, 2008 at 01:39:52PM +0200, Török Edwin wrote:
>
>> On 2008-11-27 11:28, Mike Waychison wrote:
>>
>>> Correct. I don't recall the numbers from the pathelogical cases we
>>> were seeing, but iirc, it was on the order of 10s of seconds, likely
>>> exascerbated by slower than usual disks. I've been digging through my
>>> inbox to find numbers without much success -- we've been using a
>>> variant of this patch since 2.6.11.
>>>
>>> Török however identified mmap taking on the order of several
>>> milliseconds due to this exact problem:
>>>
>>> http://lkml.org/lkml/2008/9/12/185
>>>
>> Hi,
>>
>> Thanks for the patch. I just tested it on top of 2.6.28-rc6-tip, see
>> /proc/lock_stat output at the end.
>>
>> Running my testcase shows no significant performance difference. What am
>> I doing wrong?
>>
>
> Software may just be doing a lot of mmap/munmap activity. threads +
> mmap is never going to be pretty because it is always going to involve
> broadcasting tlb flushes to other cores... Software writers shouldn't
> be scared of using processes (possibly with some shared memory).
>
It would be interesting to compare the performance of a threaded clamd,
and of a clamd that uses multiple processes.
Distributing tasks will be a bit more tricky, since it would need to use
IPC, instead of mutexes and condition variables.
> Actually, a lot of things get faster (like malloc, or file descriptor
> operations) because locks aren't needed.
>
> Despite common perception, processes are actually much *faster* than
> threads when doing common operations like these. They are slightly slower
> sometimes with things like creation and exit, or context switching, but
> if you're doing huge numbers of those operations, then it is unlikely
> to be a performance critical app... :)
>
How about distributing tasks to a set of worked threads, is the overhead
of using IPC instead of
mutexes/cond variables acceptable?
> (end rant; sorry, that may not have been helpful to your immediate problem,
> but we need to be realistic in what complexity we are ging to add where in
> the kernel in order to speed things up. And we need to steer userspace
> away from problems that are fundamentally hard and not going to get easier
> with trends -- like virtual address activity with multiple threads)
>
I understood that mmap() is not scalable, however look at
http://lkml.org/lkml/2008/9/12/185, even fopen/fdopen does
an (anonymous) mmap internally.
That does not affect performance that much, since the overhead of a
file-backed mmap + pagefaults is higher.
Rewriting libclamav to not use mmap() would take a significant amount of
time, however I will try to avoid using mmap()
in new code (and prefer pread/read).
Also clamd is a CPU bound application [given fast enough disks ;)] and
having to wait for mmap_sem prevents it from doing "real work".
Most of the time it reads files from /tmp, that should either be in the
page cache, or (in my case) they are always in RAM (I use tmpfs).
So mmaping, and reading from these files does not involve disk I/O, yet
threads working with /tmp files still need to wait
for disk I/O to complete because it has to wait on mmap_sem (held by
another thread).
>
>
>> ...............................................................................................................................................................................................
>>
>> &sem->wait_lock: 122700
>> 126641 0.42 77.94 125372.37
>> 1779026 7368894 0.27 1099.42 3085559.16
>> ---------------
>> &sem->wait_lock 5943
>> [<ffffffff8043a768>] __up_write+0x28/0x170
>> &sem->wait_lock 8615
>> [<ffffffff805ce3ac>] __down_write_nested+0x1c/0xc0
>> &sem->wait_lock 13568
>> [<ffffffff8043a5a0>] __down_write_trylock+0x20/0x60
>> &sem->wait_lock 49377
>> [<ffffffff8043a600>] __down_read_trylock+0x20/0x60
>> ---------------
>> &sem->wait_lock 8097
>> [<ffffffff8043a5a0>] __down_write_trylock+0x20/0x60
>> &sem->wait_lock 31540
>> [<ffffffff8043a768>] __up_write+0x28/0x170
>> &sem->wait_lock 5501
>> [<ffffffff805ce3ac>] __down_write_nested+0x1c/0xc0
>> &sem->wait_lock 33342
>> [<ffffffff8043a600>] __down_read_trylock+0x20/0x60
>>
>>
>
> Interesting. I have some (ancient) patches to make rwsems more scalable
> under heavy load by reducing contention on this lock. They should really
> have been merged... Not sure how much it would help, but if you're
> interested in testing, I could dust them off.
Sure, I can test patches (preferably against 2.6.28-rc6-tip ).
Best regards,
--Edwin
--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/
Powered by blists - more mailing lists