[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-Id: <E1In7S9-0001rv-NO@localhost>
Date: Wed, 31 Oct 2007 14:53:25 +0800
From: Fengguang Wu <wfg@...l.ustc.edu.cn>
To: Florin Iucha <florin@...ha.net>
Cc: Linux Kernel Mailing List <linux-kernel@...r.kernel.org>,
Trond Myklebust <trond.myklebust@....uio.no>,
Peter Zijlstra <peterz@...radead.org>
Subject: Re: pdflush stuck in D state with v2.6.24-rc1-192-gef49c32
On Tue, Oct 30, 2007 at 10:52:45PM -0500, Florin Iucha wrote:
> On Tue, Oct 30, 2007 at 07:02:42PM -0500, Florin Iucha wrote:
> > I have added the patches and started a linux kernel compilation, and
> > something really interesting happens. I run the build with the
> > equivalent of "make -j3" and in a separate console I am watching the
> > build with 'top'. The build consumes 98% of both CPUs. If I stop the
> > output in the build console with "Ctrl-S", one core goes to idle,
> > while the other is in 50% waiting, then goes to 75% waiting. When I
> > resume the build with "Ctrl-Q", the build starts to use both CPUs at
> > 98-99%. The NFS4 use was minimal, as I did not login with Gnome, but
> > just logged on the console. Also, the CPU that is in 75% waiting
> > state changes occasionally. 'Top' shows pdflush in D state, using
> > 5-6% of CPU.
>
> I forgot the traces:
>
> http://iucha.net/2.6.24-rc1/fw.1.gz
> http://iucha.net/2.6.24-rc1/fw.2.gz
> http://iucha.net/2.6.24-rc1/fw.3.gz
Sorry for the delay - I've been fixing our server today.
[ 263.685691] mm/page-writeback.c 655 wb_kupdate: pdflush(248) 24235 global 4593 0 0 wc _M tw 1024 sk 0
[ 263.789648] requeue_io 301: inode 4031199 size 562 at 08:07(sda7)
[ 263.789656] requeue_io 301: inode 4031231 size 329 at 08:07(sda7)
[ 263.789660] requeue_io 301: inode 4031255 size 177 at 08:07(sda7)
[ 263.789664] requeue_io 301: inode 4031268 size 94 at 08:07(sda7)
[ 263.789667] requeue_io 301: inode 4031329 size 88 at 08:07(sda7)
[ 263.789671] requeue_io 301: inode 4031351 size 74 at 08:07(sda7)
[ 263.789674] requeue_io 301: inode 4031408 size 175 at 08:07(sda7)
[ 263.789678] requeue_io 301: inode 4031413 size 129 at 08:07(sda7)
[ 263.789681] requeue_io 301: inode 4031415 size 391 at 08:07(sda7)
[ 263.789690] mm/page-writeback.c 655 wb_kupdate: pdflush(248) 24235 global 4593 0 0 wc _M tw 1024 sk 0
[ 263.890184] requeue_io 301: inode 4031199 size 562 at 08:07(sda7)
[ 263.890191] requeue_io 301: inode 4031231 size 329 at 08:07(sda7)
[ 263.890195] requeue_io 301: inode 4031255 size 177 at 08:07(sda7)
[ 263.890198] requeue_io 301: inode 4031268 size 94 at 08:07(sda7)
[ 263.890202] requeue_io 301: inode 4031329 size 88 at 08:07(sda7)
[ 263.890205] requeue_io 301: inode 4031351 size 74 at 08:07(sda7)
[ 263.890208] requeue_io 301: inode 4031408 size 175 at 08:07(sda7)
[ 263.890212] requeue_io 301: inode 4031413 size 129 at 08:07(sda7)
[ 263.890215] requeue_io 301: inode 4031415 size 391 at 08:07(sda7)
[ 263.890223] mm/page-writeback.c 655 wb_kupdate: pdflush(248) 24235 global 4593 0 0 wc _M tw 1024 sk 0
It's about sda7, not NFSv4.
Is it a Reiserfs? We have a fresh fix for it: http://lkml.org/lkml/2007/10/23/93
Thank you,
Fengguang
-
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/
Powered by blists - more mailing lists