lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-Id: <200804241141.31198.chris.mason@oracle.com>
Date:	Thu, 24 Apr 2008 11:41:30 -0400
From:	Chris Mason <chris.mason@...cle.com>
To:	Christoph Hellwig <hch@...radead.org>
Cc:	Arjan van de Ven <arjan@...radead.org>,
	T David Chinner <dgc@....com>,
	Benjamin Herrenschmidt <benh@...nel.crashing.org>,
	Ingo Molnar <mingo@...e.hu>,
	Andrew Morton <akpm@...ux-foundation.org>,
	Linux Kernel Mailing List <linux-kernel@...r.kernel.org>,
	Thomas Gleixner <tglx@...utronix.de>, sandeen@...hat.com
Subject: Re: x86: 4kstacks default

On Thursday 24 April 2008, Christoph Hellwig wrote:
> On Wed, Apr 23, 2008 at 05:45:16PM -0700, Arjan van de Ven wrote:
> > THe good news is that direct reclaim is.. rare.
> > And I also doubt XFS is unique here; imagine the whole stacking thing on
> > x86-64 just the same ...
>
> It's bad news actually.  Beause it means the stack overflow happens
> totally random and hard to reproduce.   And no, XFS is not unique there,
> any filesystem with a complex enough writeback path (aka extents +
> delalloc + smart allocator) will have to use quite a lot here.  I'll be
> my 2 cent that ext4 one finished up will run into this just as likely.
>
> > I wonder if the direct reclaim path should avoid direct reclaim if the
> > stack has only X bytes left. (where the value of X is... well we can
> > figure that one out later)
>
> Actually direct reclaim should be totally avoided for complex
> filesystems.  It's horrible for the stack and for the filesystem
> writeout policy and ondisk allocation strategies.

Just as a data point, XFS isn't alone.  I run through once or twice a month 
and try to get rid of any new btrfs stack pigs, but keeping under the 4k 
stack barrier is a constant challenge.  

My storage configuration is fairly simple, if we spin the wheel of stacked IO 
devices...it won't be pretty.

Does it make more sense to kill off some brain cells on finding ways to 
dynamically increase the stack as we run out?  Or even give the robust stack 
users like xfs/btrfs a way to say: I'm pretty sure this call path is going to 
hurt, please make my stack bigger now.

We have relatively few entry points between the rest of the kernel and the FS, 
there should be some ways to compromise here.

-chris
--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ