[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-Id: <20080516135814.1c973bd5.akpm@linux-foundation.org>
Date: Fri, 16 May 2008 13:58:14 -0700
From: Andrew Morton <akpm@...ux-foundation.org>
To: Eric Sandeen <sandeen@...hat.com>
Cc: linux-ext4@...r.kernel.org, linux-kernel@...r.kernel.org,
linux-fsdevel@...r.kernel.org, jamie@...reable.org
Subject: Re: [PATCH 0/4] (RESEND) ext3[34] barrier changes
On Fri, 16 May 2008 15:53:31 -0500
Eric Sandeen <sandeen@...hat.com> wrote:
> Andrew Morton wrote:
> > On Fri, 16 May 2008 14:02:46 -0500
> > Eric Sandeen <sandeen@...hat.com> wrote:
> >
> >> A collection of patches to make ext3 & 4 use barriers by
> >> default, and to call blkdev_issue_flush on fsync if they
> >> are enabled.
> >
> > Last time this came up lots of workloads slowed down by 30% so I
> > dropped the patches in horror.
>
> I actually did a bit of research and found the old thread, honestly. I
> thought this might not be a shoo-in. :) Seems worth hashing out, though.
>
> > I just don't think we can quietly go and slow everyone's machines down
> > by this much. The overhead of journalling is already pretty horrid.
>
> But if journali[zi]ng guarantees are thrown out the window by volatile
> caches on disk, why bother with the half-solution? Slower while you
> run, worthless when you lose power? Sounds like the worst of both
> worlds. (well, ok, experience shows that it's not worthless in practice...)
>
> > If we were seeing a significant number of "hey, my disk got wrecked"
> > reports which attributable to this then yes, perhaps we should change
> > the default. But I've never seen _any_, although I've seen claims that
> > others have seen reports.
>
> Hm, how would we know, really? What does it look like? It'd totally
> depend on what got lost... When do you find out? Again depends what
> you're doing, I think. I'll admit that I don't have any good evidence
> of my own. I'll go off and do some plug-pull-testing and a benchmark or
> two.
>
> But, drive caches are only getting bigger, I assume this can't help. I
> have a hard time seeing how speed at the cost of correctness is the
> right call...
Yeah, it's all so handwavy. The only thing which isn't handwavy is
that performance hit.
> > There are no happy solutions here, and I'm inclined to let this dog
> > remain asleep and continue to leave it up to distributors to decide
> > what their default should be.
> >
> > Do we know which distros are enabling barriers by default?
>
> SuSE does (via patch for ext3). Red Hat & Fedora don't, and install by
> default on lvm which won't pass barriers anyway. So maybe it's
> hypocritical to send this patch from redhat.com :)
>
> And as another "who uses barriers" datapoint, reiserfs & xfs both have
> them on by default.
>
> I suppose alternately I could send another patch to remove "remember
> that ext3/4 by default offers higher data integrity guarantees than
> most." from Documentation/filesystems/ext4.txt ;)
We could add a big scary printk at mount time and provide a document?
--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/
Powered by blists - more mailing lists