[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <20100603154634.GC8980@think>
Date: Thu, 3 Jun 2010 11:46:34 -0400
From: Chris Mason <chris.mason@...cle.com>
To: Nick Piggin <npiggin@...e.de>
Cc: "Martin K. Petersen" <martin.petersen@...cle.com>,
James Bottomley <James.Bottomley@...e.de>,
Matthew Wilcox <matthew@....cx>,
Christof Schmitt <christof.schmitt@...ibm.com>,
Boaz Harrosh <bharrosh@...asas.com>,
linux-scsi@...r.kernel.org, linux-kernel@...r.kernel.org,
linux-fsdevel@...r.kernel.org
Subject: Re: Wrong DIF guard tag on ext2 write
On Wed, Jun 02, 2010 at 11:41:21PM +1000, Nick Piggin wrote:
> On Wed, Jun 02, 2010 at 09:17:56AM -0400, Martin K. Petersen wrote:
> > >>>>> "Nick" == Nick Piggin <npiggin@...e.de> writes:
> >
> > >> 1) filesystem changed it
> > >> 2) corruption on the wire or in the raid controller
> > >> 3) the page was corrupted while the IO layer was doing the IO.
> > >>
> > >> 1 and 2 are easy, we bounce, retry and everyone continues on with
> > >> their lives. With #3, we'll recrc and send the IO down again
> > >> thinking the data is correct when really we're writing garbage.
> > >>
> > >> How can we tell these three cases apart?
> >
> > Nick> Do we really need to handle #3? It could have happened before the
> > Nick> checksum was calculated.
> >
> > Reason #3 is one of the main reasons for having the checksum in the
> > first place. The whole premise of the data integrity extensions is that
> > the checksum is calculated in close temporal proximity to the data
> > creation. I.e. eventually in userland.
> >
> > Filesystems will inevitably have to be integrity-aware for that to work.
> > And it will be their job to keep the data pages stable during DMA.
>
> Let's just think hard about what windows can actually be closed versus
> how much effort goes in to closing them. I also prefer not to accept
> half-solutions in the kernel because they don't want to implement real
> solutions in hardware (it's pretty hard to checksum and protect all
> kernel data structures by hand).
>
> For "normal" writes into pagecache, the data can get corrupted anywhere
> from after it is generated in userspace, during the copy, while it is
> dirty in cache, and while it is being written out.
This is why the DIF/DIX spec has the idea of a crc generated in userland
when the data is generated. At any rate the basic idea is to crc early
but not often...recalculating the crc after we hand our precious memory
to the evil device driver does weaken the end-to-end integrity checks.
What I don't want to do is weaken the basic DIF/DIX structure by letting
the lower recrc stuff as they find faults. It would be fine if we had
some definitive way to say: the FS raced, just recrc, but we really
don't.
>
> Closing the while it is dirty, while it is being written back window
> still leaves a pretty big window. Also, how do you handle mmap writes?
> Write protect and checksum the destination page after every store? Or
> leave some window between when the pagecache is dirtied and when it is
> written back? So I don't know whether it's worth putting a lot of effort
> into this case.
So, changing gears to how do we protect filesystem page cache pages
instead of the generic idea of dif/dix, btrfs crcs just before writing,
which does leave a pretty big window for the page to get corrupted.
The storage layer shouldn't care or know about that though, we hand it a
crc and it makes sure data matching that crc goes to the media.
>
> If you had an interface for userspace to insert checksums to direct IO
> requests or pagecache ranges, then not only can you close the entire gap
> between userspace data generation, and writeback. But you also can
> handle mmap writes and anything else just fine: userspace knows about
> the concurrency details, so it can add the right checksum (and
> potentially fsync etc) when it's ready.
Yeah, I do agree here.
-chris
--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/
Powered by blists - more mailing lists