[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <4E41A254.3030300@linux.vnet.ibm.com>
Date: Tue, 09 Aug 2011 14:10:44 -0700
From: Allison Henderson <achender@...ux.vnet.ibm.com>
To: "Ted Ts'o" <tytso@....edu>
CC: linux-ext4@...r.kernel.org
Subject: Re: [PATCH 1/1 v4] ext4: fix xfstests 75, 112, 127 punch hole failure
On 08/09/2011 09:45 AM, Ted Ts'o wrote:
> On Mon, Aug 08, 2011 at 01:54:30PM -0700, Allison Henderson wrote:
>>> Correct me if I'm wrong, but the primary problem that we need to worry
>>> about here is that the block that has just been punched out could be
>>> reused by some other inode --- but if we still have a dirty buffer
>>> head mapped to the now-punched-out-and-released-block, writeback could
>>> end up corrupting some other buffer. So in some sense, what we are
>>> effectively doing is a bforget, right?
>>
>> Hmm, well actually when I spotted this bug, I was thinking more
>> along the lines of: because the buffer is still mapped, we end up
>> reading stale data out of it. They are not supposed to be dirty
>> because we flush out all the pages before punching holes. We did
>> this with the idea of avoiding race conditions, and also it
>> simplified the code because the delayed extents are gone. But now
>> that you point it out, Im seeing that we only flush the aligned
>> pages in the hole, so maybe we need to flush the extra two non
>> aligned pages at the beginning and end.
>
> We never use the buffer to read data "out of" the buffer. We used to
> use the buffer_head as the interface to the buffer layer when reading
> or writing an entire page, but the buffer heads were never part of the
> buffer cache, so there was never any chance that you could read "stale
> data" from a mapped buffer head. These days, we use the buffer head
> as an inefficient data structure to store the mapping from the logical
> block # to a physical block #. But that's really it.
>
> But the more I look at this code, the more I think it's not quite
> right, and I think that's why you're still having a failure with test
> 127. First of all, the comment here is wrong:
>
> /*
> * Now we need to zero out the non-block-aligned data.
> * If the file space being truncated is smaller than
> * than a block, just zero out the middle
> */
>
> If blocksize != pagesize, this comment is _wrong_ in the case where
> punched region is within a single page, but larger than a block:
>
> if (first_block> last_block)
> ext4_block_zero_page_range(handle, mapping, offset, length);
>
> ext4_block_zero_page() will zero out the entire range that has been
> punched out if it is within a single page. And that is in fact a good
> and proper thing to do, even if it is larger than a block. For
> example, assume a page size of 4096, and a block size of 1024, and the
> punched-out-region begins at offset 1000 and extends for 1030 byets.
> It's larger than a block, but because it's within a single page, we
> zero out the entire range.
Hmm, for this piece here, Im not sure I quite follow you. I was pretty
sure that ext4_block_zero_page() only deals with ranges that appear with
in one block. When I modeled ext4_unmap_partial_page_buffers() after
it, I had to add a loop to get it to move over each buffer head instead
of dealing with just one. Maybe the comment should say something "If
the file space being truncated is contained in one block". And a
similar comment for the page un-mapping code too?
>
> else {
> /* zero out the head of the hole before the first block */
> block_len = first_block_offset - offset;
> if (block_len> 0)
> ext4_block_zero_page_range(handle, mapping,
> offset, block_len);
>
> /* zero out the tail of the hole after the last block */
> block_len = offset + length - last_block_offset;
> if (block_len> 0) {
> ext4_block_zero_page_range(handle, mapping,
> last_block_offset, block_len);
> }
> }
>
> OK, now make the same assumption, but the punch range is 2000 for a
> length of 6000 bytes. And assume the file is 8192 bytes. Right now
> we will only zero out the range 2000-2048, and 7168--8000, since these
> are the "tails" before the "first block" and after the "last block".
>
> But if both 4k files are mapped in the page cache, in fact we need to
> zero out the ranges 2000-4095 and 4096--8000! Why? Because we don't
> read things out of the buffer cache, we read things out of the page
> cache. Or to make this more obvious, suppose this file is mmap'ed
> into some process address space. If you don't zero out the entire
> range of bytes, then when the process reads from the mmap'ed region,
> it will see non-zero pages. It matters not a whit that the buffers
> heads are unmapped in ext4_unmap_partial_page_buffers().
>
> The reason why it's important to remove the mapped buffers is because
> if we ever call write_page(), or read_page(), we use the mapped
> buffers as a cache so we don't have to call ext4_map_blocks(). Hence,
> if we remove some blocks from the logical->physical block mapping via
> the punch() system call, we need to update the buffer_heads that may
> be attached to the page since we have to keep the cache in sync.
>
> Does that make sense?
Yes, I think this is pretty close to what is going on in test 127. :)
When I took a closer look at the code that flushes the hole, I found it
is actually flushing the entire hole (its a little misleading, I will
fix that), but beyond i_size due to a condition that matches what you
describe. So what your saying now makes a lot of sense :) So it looks
like what I need to do now is add some code to zero out the rest of the
page when i_size and the end of the hole are in the same page.
>
>>> Furthermore, the question I have is why don't have precisely the same
>>> problem with truncate? If we have a 16k page size, and we open a 15k
>>> file for writing, and then we overwrite the first 15k, and then
>>> truncate the file down to 4k. At the moment we'll zero out the last
>>> 11k of the file, and leave the buffer heads dirty and mapped; then
>>> suppose those blocks get reassigned and used before writeback happens.
>>> We're not calling ext4_unmap_partial_page_buffers() now; why isn't
>>> this a problem today? Are we just getting lucky? Why is this more of
>>> a problem with punch, such that xfstests 75, 112, 127, etc. are
>>> failing?
>>>
>>> Am I missing something here?
>>
>> That is a really good point, I'd be surprised if we've just been
>> lucky all this time. I am thinking maybe it is because of the fact
>> that i_size is already trimmed down to the new size? I do not think
>> we read beyond i_size.
>
> It's not the read(2) that I'm worried about; it's what happens if we
> call writepage() on that last block. If we have a left-over
> buffer_head which is no longer mapped, and that block has been
> repurposed, when we call writepage() we'll end up smashing potentially
> someone else's block. If that block hasn't gotten reused at the time
> of the writepage(), we'll just do some pointless I/O, but it won't
> cause any harm. I suspect we're just getting lucky....
>
Well, that does make sense, I suppose I can make a patch to flush, zero,
and unmap the buffer heads beyond i_size during a truncate, just like
how we're doing for punch hole. It would be nice to some how verify
that the bug is there though. I wonder if fsx doesnt find it because
it's busy with only one file? Or maybe we just havnt let it run long
enough yet with a 1024 block size. If the zeroing out does the trick
for 127, I will let it run over night and keep an eye out for any other
oddness.
Allison Henderson
> Regards,
>
> - Ted
--
To unsubscribe from this list: send the line "unsubscribe linux-ext4" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at http://vger.kernel.org/majordomo-info.html
Powered by blists - more mailing lists