[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <x49hbbij3m5.fsf@segfault.boston.devel.redhat.com>
Date: Fri, 04 Mar 2011 10:40:34 -0500
From: Jeff Moyer <jmoyer@...hat.com>
To: Jan Kara <jack@...e.cz>
Cc: Corrado Zoccolo <czoccolo@...il.com>,
"Alex\,Shi" <alex.shi@...el.com>,
"Li\, Shaohua" <shaohua.li@...el.com>,
Vivek Goyal <vgoyal@...hat.com>,
"tytso\@mit.edu" <tytso@....edu>,
"jaxboe\@fusionio.com" <jaxboe@...ionio.com>,
"linux-kernel\@vger.kernel.org" <linux-kernel@...r.kernel.org>,
"Chen\, Tim C" <tim.c.chen@...el.com>
Subject: Re: [performance bug] kernel building regression on 64 LCPUs machine
Jan Kara <jack@...e.cz> writes:
> Hi Jeff,
> On Wed 02-03-11 20:14:13, Jeff Moyer wrote:
>> So, the results are in. The test workload is an fs_mark process writing
>> out 64k files and fsyncing each file after it's written. Concurrently
>> with this is a fio job running a buffered sequential reader (bsr). Each
>> data point is the average of 10 runs, after throwing out the first run.
>> File system mount options are left at their defaults, which means that
>> barriers are on. The storage is an HP EVA, connected to the host via a
>> single 4Gb FC path.
> Thanks a lot for testing! BTW: fs_mark runs in a single thread or do you
> use more threads?
I use a single fs_mark thread. FWIW, I also tested just fs_mark, and
those numbers look good.
>> ext3 looks marginally better with your patches. We get better files/sec
>> AND better throughput from the buffered reader. For ext4, the results
>> are less encouraging. We see a drop in files/sec, and an increase in
>> throughput for the sequential reader. So, the fsync-ing is being
>> starved a bit more than before.
>>
>> || ext3 || ext4 ||
>> || fs_mark | fio bsr || fs_mark | fio bsr ||
>> --------++---------+---------++---------+---------||
>> vanilla || 517.535 | 178187 || 408.547 | 277130 ||
>> patched || 540.34 | 182312 || 342.813 | 294655 ||
>> ====================================================
>> %diff || +4.4% | +2.3% || -16.1% | +6.3% ||
> Interesting. I'm surprised ext3 and ext4 results differ this much. I'm more
> than happy with ext3 results since I just wanted to verify that fsync load
> doesn't degrade too much with the improved logic preferring non-fsync load
> more than we used to.
>
> I'm not so happy with ext4 results. The difference between ext3 and ext4
> might be that amount of data written by kjournald in ext3 is considerably
> larger if it ends up pushing out data (because of data=ordered mode) as
> well. With ext4, all data are written by filemap_fdatawrite() from fsync
> because of delayed allocation. And thus maybe for ext4 WRITE_SYNC_PLUG
> is hurting us with your fast storage and small amount of written data? With
> WRITE_SYNC, data would be already on it's way to storage before we get to
> wait for them...
>
> Or it could be that we really send more data in WRITE mode rather than in
> WRITE_SYNC mode with the patch on ext4 (that should be verifiable with
> blktrace). But I wonder how that could happen...
Yeah, I've collected blktrace data and I'll get to evaluating that.
Sorry, I ran out of time yesterday.
Cheers,
Jeff
--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/
Powered by blists - more mailing lists