lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Date:	Sat, 22 Mar 2008 22:54:51 +0100 (CET)
From:	Christian Kujau <lists@...dbynature.de>
To:	ext Andrew Morton <akpm@...ux-foundation.org>
cc:	"Rafael J. Wysocki" <rjw@...k.pl>,
	LKML <linux-kernel@...r.kernel.org>, Greg KH <greg@...ah.com>,
	Tejun Heo <htejun@...il.com>,
	Kay Sievers <kay.sievers@...y.org>
Subject: Re: 2.6.25-rc6: kernel BUG at fs/sysfs/file.c:89

On Sat, 22 Mar 2008, Christian Kujau wrote:
>> If so, the below (already merged) patch should fix this crash.
>> If this patch does not fix it then please apply this debug patch:
>> http://userweb.kernel.org/~akpm/mmotm/broken-out/gregkh-driver-driver-core-debug-for-bad-dev_attr_show-return-value.patch
>> then rerun the test.
>
> Ah, sorry, I misread your post: I applied the aforementioned dm-crypt 
> patch[0] and Greg's debug patch and got quite nasty SCSI errors, leading to a 
> complete lockup. I'll try again and only apply Neil's patch...stay tuned...

Hm, this was strange: I applied Neil's patch (and the dm-crypt patch) on 
2.6.25-rc6 and I kept getting SCSI errors (and lockups) when doing "tar 
-cf - | dd of=/dev/null" - which I did to generate disk I/O.

Doing the same as a normal user, nothing bad happened (except a
"possible circular locking dependency" warning, copied below) and the box 
is still up & running for a few hours now, with a constant read of ~30MB/s
across (md-)disks (running tar, rsync). So all in all I'm very happy, 
because with these two patches applied, 2.6.25-rc seems to be usable 
again.

Of course, it'd be interesting to know where the scsi errors come from, 
but that's another story I guess...

Thanks to all involved,
Christian.

[ 4657.715881] =======================================================
[ 4657.716512] [ INFO: possible circular locking dependency detected ]
[ 4657.716895] 2.6.25-rc6 #5
[ 4657.717170] -------------------------------------------------------
[ 4657.717552] rsync/14184 is trying to acquire lock:
[ 4657.717892]  (iprune_mutex){--..}, at: [<c017b672>] shrink_icache_memory+0x72/0x220
[ 4657.718239] 
[ 4657.718239] but task is already holding lock:
[ 4657.718463]  (&(&ip->i_iolock)->mr_lock){----}, at: [<c027a766>] xfs_ilock+0x96/0xb0
[ 4657.718714] 
[ 4657.718715] which lock already depends on the new lock.
[ 4657.718716] 
[ 4657.719047] 
[ 4657.719047] the existing dependency chain (in reverse order) is:
[ 4657.719291] 
[ 4657.719292] -> #1 (&(&ip->i_iolock)->mr_lock){----}:
[ 4657.719526]        [<c01369f4>] add_lock_to_list+0x44/0xc0
[ 4657.719779]        [<c0139486>] __lock_acquire+0xc26/0x10b0
[ 4657.720142]        [<c027a766>] xfs_ilock+0x96/0xb0
[ 4657.720483]        [<c013826d>] mark_held_locks+0x3d/0x70
[ 4657.720483]        [<c013996e>] lock_acquire+0x5e/0x80
[ 4657.720483]        [<c027a766>] xfs_ilock+0x96/0xb0
[ 4657.720483]        [<c012fc81>] down_write_nested+0x41/0x60
[ 4657.720483]        [<c027a766>] xfs_ilock+0x96/0xb0
[ 4657.720483]        [<c027a766>] xfs_ilock+0x96/0xb0
[ 4657.720483]        [<c027a8fa>] xfs_ireclaim+0x1a/0x60
[ 4657.720483]        [<c02987b3>] xfs_finish_reclaim+0x53/0x1a0
[ 4657.720483]        [<c02a7c6e>] xfs_fs_clear_inode+0x5e/0x90
[ 4657.720483]        [<c017b279>] clear_inode+0xa9/0x130
[ 4657.720483]        [<c017ac40>] destroy_inode+0x20/0x40
[ 4657.720483]        [<c017b55a>] dispose_list+0x1a/0xc0
[ 4657.720483]        [<c017b7e2>] shrink_icache_memory+0x1e2/0x220
[ 4657.720483]        [<c0150211>] shrink_slab+0x101/0x160
[ 4657.720483]        [<c01507a8>] kswapd+0x298/0x3f0
[ 4657.720483]        [<c014f140>] isolate_pages_global+0x0/0x60
[ 4657.720483]        [<c012c710>] autoremove_wake_function+0x0/0x40
[ 4657.720483]        [<c01383bc>] trace_hardirqs_on+0x9c/0x110
[ 4657.720483]        [<c0150510>] kswapd+0x0/0x3f0
[ 4657.720483]        [<c012c442>] kthread+0x42/0x70
[ 4657.720483]        [<c012c400>] kthread+0x0/0x70
[ 4657.720483]        [<c0103a1f>] kernel_thread_helper+0x7/0x18
[ 4657.720483]        [<ffffffff>] 0xffffffff
[ 4657.720483] 
[ 4657.720483] -> #0 (iprune_mutex){--..}:
[ 4657.720483]        [<c0136c80>] print_circular_bug_entry+0x40/0x50
[ 4657.720483]        [<c0139287>] __lock_acquire+0xa27/0x10b0
[ 4657.720483]        [<c01389ef>] __lock_acquire+0x18f/0x10b0
[ 4657.720483]        [<c013996e>] lock_acquire+0x5e/0x80
[ 4657.720483]        [<c017b672>] shrink_icache_memory+0x72/0x220
[ 4657.720483]        [<c043e2f9>] mutex_lock_nested+0x89/0x240
[ 4657.720483]        [<c017b672>] shrink_icache_memory+0x72/0x220
[ 4657.720483]        [<c017b672>] shrink_icache_memory+0x72/0x220
[ 4657.720483]        [<c017b672>] shrink_icache_memory+0x72/0x220
[ 4657.720483]        [<c0150131>] shrink_slab+0x21/0x160
[ 4657.720483]        [<c0150211>] shrink_slab+0x101/0x160
[ 4657.720483]        [<c01503c2>] try_to_free_pages+0x152/0x230
[ 4657.720483]        [<c014f140>] isolate_pages_global+0x0/0x60
[ 4657.720483]        [<c014ba3b>] __alloc_pages+0x14b/0x370
[ 4657.720483]        [<c043fa20>] _read_unlock_irq+0x20/0x30
[ 4657.720483]        [<c01466e1>] __grab_cache_page+0x81/0xc0
[ 4657.720483]        [<c01897d6>] block_write_begin+0x76/0xe0
[ 4657.720483]        [<c029ed56>] xfs_vm_write_begin+0x46/0x50
[ 4657.720483]        [<c029f5a0>] xfs_get_blocks+0x0/0x30
[ 4657.720483]        [<c0147377>] generic_file_buffered_write+0x117/0x650
[ 4657.720483]        [<c01846c3>] __mark_inode_dirty+0x53/0x180
[ 4657.720483]        [<c043f6b9>] _spin_lock+0x29/0x40
[ 4657.720483]        [<c01846c3>] __mark_inode_dirty+0x53/0x180
[ 4657.720483]        [<c02a74ac>] xfs_write+0x7ac/0x8a0
[ 4657.720483]        [<c0174ba1>] core_sys_select+0x21/0x350
[ 4657.720483]        [<c02a339c>] xfs_file_aio_write+0x5c/0x70
[ 4657.720483]        [<c0167cd5>] do_sync_write+0xd5/0x120
[ 4657.720483]        [<c012c710>] autoremove_wake_function+0x0/0x40
[ 4657.720483]        [<c019d0b5>] dnotify_parent+0x35/0x90
[ 4657.720483]        [<c0167c00>] do_sync_write+0x0/0x120
[ 4657.720483]        [<c016854f>] vfs_write+0x9f/0x140
[ 4657.720483]        [<c0168b01>] sys_write+0x41/0x70
[ 4657.720483]        [<c0102dee>] sysenter_past_esp+0x5f/0xa5
[ 4657.720483]        [<ffffffff>] 0xffffffff
[ 4657.720483] 
[ 4657.720483] other info that might help us debug this:
[ 4657.720483] 
[ 4657.720483] 3 locks held by rsync/14184:
[ 4657.720483]  #0:  (&sb->s_type->i_mutex_key#12){--..}, at: [<c02a70f9>] xfs_write+0x3f9/0x8a0
[ 4657.720483]  #1:  (&(&ip->i_iolock)->mr_lock){----}, at: [<c027a766>] xfs_ilock+0x96/0xb0
[ 4657.720483]  #2:  (shrinker_rwsem){----}, at: [<c0150131>] shrink_slab+0x21/0x160
[ 4657.720483] 
[ 4657.720483] stack backtrace:
[ 4657.720483] Pid: 14184, comm: rsync Not tainted 2.6.25-rc6 #5
[ 4657.720483]  [<c0137502>] print_circular_bug_tail+0x72/0x80
[ 4657.720483]  [<c0139287>] __lock_acquire+0xa27/0x10b0
[ 4657.720483]  [<c01389ef>] __lock_acquire+0x18f/0x10b0
[ 4657.720483]  [<c013996e>] lock_acquire+0x5e/0x80
[ 4657.720483]  [<c017b672>] shrink_icache_memory+0x72/0x220
[ 4657.720483]  [<c043e2f9>] mutex_lock_nested+0x89/0x240
[ 4657.720483]  [<c017b672>] shrink_icache_memory+0x72/0x220
[ 4657.720483]  [<c017b672>] shrink_icache_memory+0x72/0x220
[ 4657.720483]  [<c017b672>] shrink_icache_memory+0x72/0x220
[ 4657.720483]  [<c0150131>] shrink_slab+0x21/0x160
[ 4657.720483]  [<c0150211>] shrink_slab+0x101/0x160
[ 4657.720483]  [<c01503c2>] try_to_free_pages+0x152/0x230
[ 4657.720483]  [<c014f140>] isolate_pages_global+0x0/0x60
[ 4657.720483]  [<c014ba3b>] __alloc_pages+0x14b/0x370
[ 4657.720483]  [<c043fa20>] _read_unlock_irq+0x20/0x30
[ 4657.720483]  [<c01466e1>] __grab_cache_page+0x81/0xc0
[ 4657.720483]  [<c01897d6>] block_write_begin+0x76/0xe0
[ 4657.720483]  [<c029ed56>] xfs_vm_write_begin+0x46/0x50
[ 4657.720483]  [<c029f5a0>] xfs_get_blocks+0x0/0x30
[ 4657.720483]  [<c0147377>] generic_file_buffered_write+0x117/0x650
[ 4657.720483]  [<c01846c3>] __mark_inode_dirty+0x53/0x180
[ 4657.720483]  [<c043f6b9>] _spin_lock+0x29/0x40
[ 4657.720483]  [<c01846c3>] __mark_inode_dirty+0x53/0x180
[ 4657.720483]  [<c02a74ac>] xfs_write+0x7ac/0x8a0
[ 4657.720483]  [<c0174ba1>] core_sys_select+0x21/0x350
[ 4657.720483]  [<c02a339c>] xfs_file_aio_write+0x5c/0x70
[ 4657.720483]  [<c0167cd5>] do_sync_write+0xd5/0x120
[ 4657.720483]  [<c012c710>] autoremove_wake_function+0x0/0x40
[ 4657.720483]  [<c019d0b5>] dnotify_parent+0x35/0x90
[ 4657.720483]  [<c0167c00>] do_sync_write+0x0/0x120
[ 4657.720483]  [<c016854f>] vfs_write+0x9f/0x140
[ 4657.720483]  [<c0168b01>] sys_write+0x41/0x70
[ 4657.720483]  [<c0102dee>] sysenter_past_esp+0x5f/0xa5
[ 4657.720483]  =======================

-- 
BOFH excuse #197:

I'm sorry a pentium won't do, you need an SGI to connect with us.
--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ