lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [<thread-prev] [day] [month] [year] [list]
Date:	Tue, 25 Jun 2013 16:56:42 +0200
From:	Jan Kara <jack@...e.cz>
To:	Paul Gortmaker <paul.gortmaker@...driver.com>
Cc:	Jan Kara <jack@...e.cz>, linux-rt-users@...r.kernel.org,
	linux-ext4@...r.kernel.org
Subject: Re: ext4/jbd2 hangs in __jbd2_log_wait_for_space on 3.4-RT/3.6-RT

On Tue 25-06-13 09:52:33, Paul Gortmaker wrote:
> On 13-06-25 09:18 AM, Jan Kara wrote:
> > On Fri 31-05-13 14:34:12, Paul Gortmaker wrote:
> >> This problem is seen on vanilla 3.4-RT and 3.6-RT kernels. It is
> >> not clear to me whether this is an RT issue, or whether (as usual)
> >> RT has managed to shake out an issue in mainline code.  So I've
> >> looped in the ext4 list as well as the RT list, since at the
> >> moment it appears this can impact anyone using RT and ext4...
> >>
> >> What happens is that under reasonable load, the jbd2/sda1-8 thread
> >> goes D state, and then lots of regular processes follow suit, after
> >> calling __jbd2_log_wait_for_space.  As can be seen at the bottom
> >> of the sysrq-t output, j_checkpoint_mutex is implicated.  All
> >> future processes trying to do I/O to/from that filesystem go D.
> >>
> >> More testing details:
> >> Even though debug_rt_mutex_print_deadlock shows up in each stalled
> >> process backtrace, no output is seen from debug_rt_mutex_print_deadlock.
> >> There are no messages in dmesg at all, until I trigger a SysRQ-t.
> >>
> >> I've reproduced this on v3.4.42-rt57, v3.4.47-rt62, and v3.6.11.3-rt35.
> >>
> >> The two separate versions of v3.4.x are because I noticed the 3.4.47
> >> pulled in some jbd2 commits via stable, like 794446c6 "jbd2: fix race
> >> between jbd2_journal_remove_checkpoint and ->j_commit_callback". It
> >> looked promising, but having that present didn't change things.
> >>
> >> I'm using a yocto build, configured for six parallel package builds,
> >> each pkg in turn with "make -j6" to create I/O.  I've found that also
> >> running an "rm -rf" of an old build (several gigs of data) at the
> >> same time increases the probability of it.  Typically it will fail
> >> within about 15m or so.  The test box is a dell optiplex 990 with
> >> a single disk as ext4.  The box stays alive for basic sysrq operations
> >> and anything else that doesn't touch the locked filesystem.  The build
> >> halts with a static load average equal to the number of blocked D procs.
> >>
> >> I've deleted the sysrq-t output from the irrelevant sleeping processes
> >> in order to reduce the noise.  I'll keep looking at this but I'm hoping
> >> more experienced eyes on the problem will help, since it seems common
> >> to all RT users and hence of interest to everyone (I've not yet tried
> >> 3.8.x-RT, mind you.)
> >   Hum, this sounds familiar... I was already debugging this with RT kernel
> > and I also remember it was RT specific issue. Let me try to remember the
> > whole story... yes, while wandering over the traces I think I remember what
> > was the problem: In standard kernel, whenever we scheduler process out from
> > CPU, we unplug its IO queue in sched_submit_work(). However in RT kernel
> > that was not the case. So it could happen that a process has IOs queued
> > and was sent to sleep waiting for jbd2 thread to free some journal space
> > and jbd2 thread was waiting for some IO to complete - however that never
> > happened because the IO was sitting in the sleeping process' queue.
> 
> Do you have a link to that older discussion?  I did search around before
> posting, but came up empty.  I'll try and fold your description into my
> thoughts as I return to looking at it (got dragged into other things
> as of late, and haven't been spending time on this as of late...)
  So I did some more archeology and the discussion starts here:
https://lkml.org/lkml/2012/7/11/255

								Honza
-- 
Jan Kara <jack@...e.cz>
SUSE Labs, CR
--
To unsubscribe from this list: send the line "unsubscribe linux-ext4" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html

Powered by blists - more mailing lists