lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <20090421142305.GB22619@redhat.com>
Date:	Tue, 21 Apr 2009 10:23:05 -0400
From:	Vivek Goyal <vgoyal@...hat.com>
To:	Andrea Righi <righi.andrea@...il.com>
Cc:	Paul Menage <menage@...gle.com>,
	Balbir Singh <balbir@...ux.vnet.ibm.com>,
	Gui Jianfeng <guijianfeng@...fujitsu.com>,
	KAMEZAWA Hiroyuki <kamezawa.hiroyu@...fujitsu.com>,
	agk@...rceware.org, akpm@...ux-foundation.org, axboe@...nel.dk,
	baramsori72@...il.com, Carl Henrik Lunde <chlunde@...g.uio.no>,
	dave@...ux.vnet.ibm.com, Divyesh Shah <dpshah@...gle.com>,
	eric.rannaud@...il.com, fernando@....ntt.co.jp,
	Hirokazu Takahashi <taka@...inux.co.jp>,
	Li Zefan <lizf@...fujitsu.com>, matt@...ehost.com,
	dradford@...ehost.com, ngupta@...gle.com, randy.dunlap@...cle.com,
	roberto@...it.it, Ryo Tsuruta <ryov@...inux.co.jp>,
	Satoshi UCHIDA <s-uchida@...jp.nec.com>,
	subrata@...ux.vnet.ibm.com, yoshikawa.takuya@....ntt.co.jp,
	Theodore Tso <tytso@....edu>,
	containers@...ts.linux-foundation.org, linux-kernel@...r.kernel.org
Subject: Re: [PATCH 1/9] io-throttle documentation

On Tue, Apr 21, 2009 at 10:37:03AM +0200, Andrea Righi wrote:
> On Mon, Apr 20, 2009 at 09:08:46PM -0400, Vivek Goyal wrote:
> > On Tue, Apr 21, 2009 at 12:05:12AM +0200, Andrea Righi wrote:
> > 
> > [..]
> > > > > > Are we not already controlling submission of request (at crude level).
> > > > > > If application is doing writeout at high rate, then it hits vm_dirty_ratio
> > > > > > hits and this application is forced to do write out and hence it is slowed
> > > > > > down and is not allowed to submit writes at high rate.
> > > > > > 
> > > > > > Just that it is not a very fair scheme right now as during right out
> > > > > > a high prio/high weight cgroup application can start writing out some
> > > > > > other cgroups' pages.
> > > > > > 
> > > > > > For this we probably need to have some combination of solutions like
> > > > > > per cgroup upper limit on dirty pages. Secondly probably if an application
> > > > > > is slowed down because of hitting vm_drity_ratio, it should try to
> > > > > > write out the inode it is dirtying first instead of picking any random
> > > > > > inode and associated pages. This will ensure that a high weight
> > > > > > application can quickly get through the write outs and see higher
> > > > > > throughput from the disk.
> > > > > 
> > > > > For the first, I submitted a patchset some months ago to provide this
> > > > > feature in the memory controller:
> > > > > 
> > > > > https://lists.linux-foundation.org/pipermail/containers/2008-September/013140.html
> > > > > 
> > > > > We focused on the best interface to use for setting the dirty pages
> > > > > limit, but we didn't finalize it. I can rework on that and repost an
> > > > > updated version. Now that we have the dirty_ratio/dirty_bytes to set the
> > > > > global limit I think we can use the same interface and the same semantic
> > > > > within the cgroup fs, something like:
> > > > > 
> > > > >   memory.dirty_ratio
> > > > >   memory.dirty_bytes
> > > > > 
> > > > > For the second point something like this should be enough to force tasks
> > > > > to write out only the inode they're actually dirtying when they hit the
> > > > > vm_dirty_ratio limit. But it should be tested carefully and may cause
> > > > > heavy performance regressions.
> > > > > 
> > > > > Signed-off-by: Andrea Righi <righi.andrea@...il.com>
> > > > > ---
> > > > >  mm/page-writeback.c |    2 +-
> > > > >  1 files changed, 1 insertions(+), 1 deletions(-)
> > > > > 
> > > > > diff --git a/mm/page-writeback.c b/mm/page-writeback.c
> > > > > index 2630937..1e07c9d 100644
> > > > > --- a/mm/page-writeback.c
> > > > > +++ b/mm/page-writeback.c
> > > > > @@ -543,7 +543,7 @@ static void balance_dirty_pages(struct address_space *mapping)
> > > > >  		 * been flushed to permanent storage.
> > > > >  		 */
> > > > >  		if (bdi_nr_reclaimable) {
> > > > > -			writeback_inodes(&wbc);
> > > > > +			sync_inode(mapping->host, &wbc);
> > > > >  			pages_written += write_chunk - wbc.nr_to_write;
> > > > >  			get_dirty_limits(&background_thresh, &dirty_thresh,
> > > > >  				       &bdi_thresh, bdi);
> > > > 
> > > > This patch seems to be helping me a bit in getting more service
> > > > differentiation between two writer dd of different weights. But strangely
> > > > it is helping only for ext3 and not ext4. Debugging is on.
> > > 
> > > Are you explicitly mounting ext3 with data=ordered?
> > 
> > Yes. Still using 29-rc8 and data=ordered was the default then.
> > 
> > I got two partitions on same disk and created one ext3 filesystem on each
> > partition (just to take journaling intereference out of two dd threads 
> > for the time being).
> > 
> > Two dd threads doing writes to each partition. 
> 
> ...and if you're using data=writeback with ext4 sync_inode() should sync
> the metadata only. If this is the case, could you check data=ordered
> also for ext4?

No, even data=ordered mode with ext4 is also not helping. It has to be
something else.

BTW, with the above patch, what happens if the address space being dirtied
does not have sufficient dirty pages to write back (more than write_chunk).
Will the process not be in loop until the number of dirty pages come down
(hopefully due to writeout by pdflush or by other processes?)

Thanks
Vivek
--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ