[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <20140731130332.GB1566@thunk.org>
Date: Thu, 31 Jul 2014 09:03:32 -0400
From: Theodore Ts'o <tytso@....edu>
To: Benjamin LaHaise <bcrl@...ck.org>
Cc: Andreas Dilger <adilger@...ger.ca>,
"linux-ext4@...r.kernel.org" <linux-ext4@...r.kernel.org>
Subject: Re: ext4: first write to large ext3 filesystem takes 96 seconds
On Wed, Jul 30, 2014 at 10:49:28AM -0400, Benjamin LaHaise wrote:
> This seems like a pretty serious regression relative to ext3. Why can't
> ext4's mballoc pick better block groups to attempt allocating from based
> on the free block counts in the block group summaries?
Allocation algorithms are *always* tradeoffs. So I don't think
regression is necessarily the best way to think about things.
Unfortuntaely, your use case really doesn't work well with how we have
set things up with ext4 now. Sure, if you your specific use case is
one where you are mostly allocating 8MB files, then we can add a
special case where if you are allocating 32768 blocks, we should
search for block groups that have 32768 blocks free. And if that's
what you are asking for, we can certainly do that.
The problem is that free block counts don't work well in general. If
I see that the free block count is 2048 blocks, that doesn't tell me
the free blocks are in a contiguous single chunk of 2048 blocks, or
2048 single block items. (We do actually pay attention to free
blocks, by the way, but it's in a nuanced way.)
If the only goal you have is fast block allocation after fail over,
you can always use the VFAT block allocation --- i.e., use the first
free block in the file system. Unfortunately, it will result in a
very badly fragmented file system, as Microsoft and its users
discovered.
I'm sure that are things we could do that would make things better for
your workload (if you want to tell us in great detail exactly what the
file/block allocation patterns are for your workload), and perhaps
even better in general, but the challenge is making sure we don't
regress for other workloads --- and this includes long-term
fragmentation resistance. This is a hard problem. Kvetching about
how it's so horrible just for you isn't really helpful for solving it.
(BTW, one of the problems is that ext4_mb_normalize_request caps large
allocations so that we use the same goal length for multiple passes as
we search for good block groups. We might want to use the original
goal length --- so long as it is less than 32768 blocks --- for the
first scan, or at least for goal lengths which are powers of two. So
if your application is regularly allocating files which are exactly
8MB, there are probably some optimizations that we could apply. But
if they aren't exactly 8MB, life gets a bit trickier.)
Regards,
- Ted
--
To unsubscribe from this list: send the line "unsubscribe linux-ext4" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at http://vger.kernel.org/majordomo-info.html
Powered by blists - more mailing lists