[<prev] [next>] [<thread-prev] [day] [month] [year] [list]
Message-ID: <20251031164701.GA27481@lst.de>
Date: Fri, 31 Oct 2025 17:47:01 +0100
From: Christoph Hellwig <hch@....de>
To: Keith Busch <kbusch@...nel.org>
Cc: Christoph Hellwig <hch@....de>, Dave Chinner <david@...morbit.com>,
Carlos Maiolino <cem@...nel.org>,
Christian Brauner <brauner@...nel.org>, Jan Kara <jack@...e.cz>,
"Martin K. Petersen" <martin.petersen@...cle.com>,
linux-kernel@...r.kernel.org, linux-xfs@...r.kernel.org,
linux-fsdevel@...r.kernel.org, linux-raid@...r.kernel.org,
linux-block@...r.kernel.org
Subject: Re: fall back from direct to buffered I/O when stable writes are
required
On Fri, Oct 31, 2025 at 09:57:35AM -0600, Keith Busch wrote:
> Not sure of any official statement to that effect, but storage in
> general always says the behavior of modifying data concurrently with
> in-flight operations on that data produces non-deterministic results.
Yes, it's pretty clear that the result in non-deterministic in what you
get. But that result still does not result in corruption, because
there is a clear boundary ( either the sector size, or for NVMe
optionally even a larger bodunary) that designates the atomicy boundary.
> An
> application with such behavior sounds like a bug to me as I can't
> imagine anyone purposefully choosing to persist data with a random
> outcome. If PI is enabled, I think they'd rather get a deterministic
> guard check error so they know they did something with undefined
> behavior.
As long as your clearly define your transaction boundaries that
non-atomicy is not a problem per se.
> It's like having reads and writes to overlapping LBA and/or memory
> ranges concurrently outstanding. There's no guaranteed result there
> either; specs just say it's the host's responsibilty to not do that.
There is no guaranteed result as in an enforced ordering. But there
is a pretty clear model that you get either the old or new at a
well defined boundary.
> The kernel doesn't stop an application from trying that on raw block
> direct-io, but I'd say that's an application bug.
If it corrupts other applications data as in the RAID case it's
pretty clearly not an application bug. It's also pretty clear that
at least some applications (qemu and other VMs) have been doings this
for 20+ years.
Powered by blists - more mailing lists