lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <CA+5PVA4i0xp7jqcR=yMG3BjucXfkbM+yXP7YoDHWZgbLtdej-Q@mail.gmail.com>
Date:	Thu, 23 Jan 2014 21:51:05 -0500
From:	Josh Boyer <jwboyer@...oraproject.org>
To:	Dave Chinner <david@...morbit.com>
Cc:	Ben Myers <bpm@....com>, Eric Sandeen <sandeen@...hat.com>,
	xfs@....sgi.com,
	"Linux-Kernel@...r. Kernel. Org" <linux-kernel@...r.kernel.org>
Subject: Re: XFS lockdep spew with v3.13-4156-g90804ed

On Thu, Jan 23, 2014 at 9:29 PM, Dave Chinner <david@...morbit.com> wrote:
> On Thu, Jan 23, 2014 at 08:58:56PM -0500, Josh Boyer wrote:
>> Hi All,
>>
>> I'm hitting an XFS lockdep error with Linus' tree today after the XFS
>> merge.  I wasn't hitting this with v3.13-3995-g0dc3fd0, which seems
>> to backup the "before XFS merge" claim.  Full text below:
>
> Ugh. mmap_sem/inode lock order stupidity.
>
> Looks like a false positive. Basically, it's complaining that a page
> fault can occur in getdents() syscall on a user buffer while the
> directory IO lock is held, and then complaining that a this is the
> opposite lock order for a
>>
>>
>> [  132.638044] ======================================================
>> [  132.638045] [ INFO: possible circular locking dependency detected ]
>> [  132.638047] 3.14.0-0.rc0.git7.1.fc21.x86_64 #1 Not tainted
>> [  132.638048] -------------------------------------------------------
>> [  132.638049] gnome-session/1432 is trying to acquire lock:
>> [  132.638050]  (&mm->mmap_sem){++++++}, at: [<ffffffff811b846f>] might_fault+0x
>> 5f/0xb0
>> [  132.638055]
>> but task is already holding lock:
>> [  132.638056]  (&(&ip->i_lock)->mr_lock){++++..}, at: [<ffffffffa05b3c12>] xfs_
>> ilock+0xf2/0x1c0 [xfs]
>> [  132.638076]
>> which lock already depends on the new lock.
>>
>> [  132.638077]
>> the existing dependency chain (in reverse order) is:
>> [  132.638078]
>> -> #1 (&(&ip->i_lock)->mr_lock){++++..}:
>> [  132.638080]        [<ffffffff810deaa2>] lock_acquire+0xa2/0x1d0
>> [  132.638083]        [<ffffffff8178312e>] _raw_spin_lock+0x3e/0x80
>> [  132.638085]        [<ffffffff8123c579>] __mark_inode_dirty+0x119/0x440
>> [  132.638088]        [<ffffffff812447fc>] __set_page_dirty+0x6c/0xc0
>> [  132.638090]        [<ffffffff812477e1>] mark_buffer_dirty+0x61/0x180
>> [  132.638092]        [<ffffffff81247a31>] __block_commit_write.isra.21+0x81/0xb0
>> [  132.638094]        [<ffffffff81247be6>] block_write_end+0x36/0x70
>> [  132.638096]        [<ffffffff81247c48>] generic_write_end+0x28/0x90
>> [  132.638097]        [<ffffffffa0554cab>] xfs_vm_write_end+0x2b/0x70 [xfs]
>> [  132.638104]        [<ffffffff8118c4f6>] generic_file_buffered_write+0x156/0x260
>> [  132.638107]        [<ffffffffa05651d7>] xfs_file_buffered_aio_write+0x107/0x250 [xfs]
>> [  132.638115]        [<ffffffffa05653eb>] xfs_file_aio_write+0xcb/0x130 [xfs]
>> [  132.638122]        [<ffffffff8120af8a>] do_sync_write+0x5a/0x90
>> [  132.638125]        [<ffffffff8120b74d>] vfs_write+0xbd/0x1f0
>> [  132.638126]        [<ffffffff8120c15c>] SyS_write+0x4c/0xa0
>> [  132.638128]        [<ffffffff8178db69>] system_call_fastpath+0x16/0x1b
>
> Sorry, what? That trace is taking the ip->i_vnode->i_lock
> *spinlock*, not the ip->i_lock *rwsem*. And it's most definitely not
> currently holding the ip->i_lock rwsem here. I think lockdep has
> dumped the wrong stack trace here, because it most certainly doesn't
> match the unsafe locking scenario that has been detected.

I rebooted again with the same kernel and lockdep spit out a different
stacktrace for this part.  See below.  The rest looks mostly the same,
and it spews when I log into gnome, so at least it's recreatable.

> You can't mmap directories, and so the page fault lock order being
> shown for CPU1 can't happen on a directory. False positive.
>
> *sigh*
>
> More complexity in setting up inode lock order instances is required
> so that lockdep doesn't confuse the lock ordering semantics of
> directories with regular files. As if that code to make lockdep
> happy wasn't complex enough already....

So the summary is basically: false positive with additional annotations needed?

josh

Full output of the lockdep report after I rebooted (sorry if the
formatting is off, I'm on a different machine and doing this remotely
until tomorrow.)

[  104.565930] ======================================================
[  104.565931] [ INFO: possible circular locking dependency detected ]
[  104.565933] 3.14.0-0.rc0.git7.1.fc21.x86_64 #1 Not tainted
[  104.565934] -------------------------------------------------------
[  104.565935] gnome-session/1450 is trying to acquire lock:
[  104.565936]  (&mm->mmap_sem){++++++}, at: [<ffffffff811b846f>]
might_fault+0x5f/0xb0
[  104.565942]
but task is already holding lock:
[  104.565943]  (&(&ip->i_lock)->mr_lock){++++..}, at:
[<ffffffffa05f4c12>] xfs_ilock+0xf2/0x1c0 [xfs]
[  104.565963]
which lock already depends on the new lock.

[  104.565964]
the existing dependency chain (in reverse order) is:
[  104.565965]
-> #1 (&(&ip->i_lock)->mr_lock){++++..}:
[  104.565967]        [<ffffffff810deaa2>] lock_acquire+0xa2/0x1d0
[  104.565970]        [<ffffffff810d7d27>] down_read_nested+0x57/0xa0
[  104.565972]        [<ffffffffa05f4c12>] xfs_ilock+0xf2/0x1c0 [xfs]
[  104.565983]        [<ffffffffa05f4cff>]
xfs_ilock_data_map_shared+0x1f/0x40 [xfs]
[  104.565995]        [<ffffffffa0597367>] __xfs_get_blocks+0x87/0x730 [xfs]
[  104.566002]        [<ffffffffa0597a21>] xfs_get_blocks+0x11/0x20 [xfs]
[  104.566008]        [<ffffffff812531e7>] do_mpage_readpage+0x447/0x670
[  104.566011]        [<ffffffff812534eb>] mpage_readpages+0xdb/0x130
[  104.566012]        [<ffffffffa0595ddd>] xfs_vm_readpages+0x1d/0x20 [xfs]
[  104.566019]        [<ffffffff8119b222>] __do_page_cache_readahead+0x2c2/0x360
[  104.566022]        [<ffffffff8119b9b1>] ra_submit+0x21/0x30
[  104.566023]        [<ffffffff8118ed7d>] filemap_fault+0x37d/0x420
[  104.566026]        [<ffffffff811b8a9f>] __do_fault+0x6f/0x520
[  104.566027]        [<ffffffff811bcd91>] handle_mm_fault+0x3b1/0xe60
[  104.566029]        [<ffffffff81787e94>] __do_page_fault+0x174/0x5e0
[  104.566031]        [<ffffffff8178830e>] do_page_fault+0xe/0x10
[  104.566032]        [<ffffffff81784538>] page_fault+0x28/0x30
[  104.566034]
-> #0 (&mm->mmap_sem){++++++}:
[  104.566036]        [<ffffffff810de0fc>] __lock_acquire+0x18ec/0x1aa0
[  104.566038]        [<ffffffff810deaa2>] lock_acquire+0xa2/0x1d0
[  104.566040]        [<ffffffff811b849c>] might_fault+0x8c/0xb0
[  104.566041]        [<ffffffff81220a91>] filldir+0x91/0x120
[  104.566043]        [<ffffffffa05a1f7f>]
xfs_dir2_sf_getdents+0x23f/0x2a0 [xfs]
[  104.566051]        [<ffffffffa05a23fb>] xfs_readdir+0x16b/0x1d0 [xfs]
[  104.566059]        [<ffffffffa05a483b>] xfs_file_readdir+0x2b/0x40 [xfs]
[  104.566066]        [<ffffffff812208d8>] iterate_dir+0xa8/0xe0
[  104.566068]        [<ffffffff81220d83>] SyS_getdents+0x93/0x120
[  104.566070]        [<ffffffff8178db69>] system_call_fastpath+0x16/0x1b
[  104.566072]
other info that might help us debug this:

[  104.566073]  Possible unsafe locking scenario:

[  104.566074]        CPU0                    CPU1
[  104.566075]        ----                    ----
[  104.566075]   lock(&(&ip->i_lock)->mr_lock);
[  104.566077]                                lock(&mm->mmap_sem);
[  104.566078]                                lock(&(&ip->i_lock)->mr_lock);
[  104.566079]   lock(&mm->mmap_sem);
[  104.566081]
 *** DEADLOCK ***

[  104.566082] 2 locks held by gnome-session/1450:
[  104.566083]  #0:  (&type->i_mutex_dir_key#4){+.+.+.}, at:
[<ffffffff81220892>] iterate_dir+0x62/0xe0
[  104.566087]  #1:  (&(&ip->i_lock)->mr_lock){++++..}, at:
[<ffffffffa05f4c12>] xfs_ilock+0xf2/0x1c0 [xfs]
[  104.566099]
stack backtrace:
[  104.566101] CPU: 4 PID: 1450 Comm: gnome-session Not tainted
3.14.0-0.rc0.git7.1.fc21.x86_64 #1
[  104.566102] Hardware name: Dell Inc. XPS 8300  /0Y2MRG, BIOS A06 10/17/2011
[  104.566103]  ffffffff825ba040 ffff8800b8dedc60 ffffffff8177a8c9
ffffffff825ba040
[  104.566106]  ffff8800b8dedca0 ffffffff8177616c ffff8800b8dedcf0
ffff8802fb842648
[  104.566108]  ffff8802fb841ab0 0000000000000002 0000000000000002
ffff8802fb842648
[  104.566110] Call Trace:
[  104.566113]  [<ffffffff8177a8c9>] dump_stack+0x4d/0x66
[  104.566115]  [<ffffffff8177616c>] print_circular_bug+0x201/0x20f
[  104.566117]  [<ffffffff810de0fc>] __lock_acquire+0x18ec/0x1aa0
[  104.566119]  [<ffffffff810deaa2>] lock_acquire+0xa2/0x1d0
[  104.566121]  [<ffffffff811b846f>] ? might_fault+0x5f/0xb0
[  104.566123]  [<ffffffff811b849c>] might_fault+0x8c/0xb0
[  104.566124]  [<ffffffff811b846f>] ? might_fault+0x5f/0xb0
[  104.566126]  [<ffffffff81220a91>] filldir+0x91/0x120
[  104.566134]  [<ffffffffa05a1f7f>] xfs_dir2_sf_getdents+0x23f/0x2a0 [xfs]
[  104.566145]  [<ffffffffa05f4c12>] ? xfs_ilock+0xf2/0x1c0 [xfs]
[  104.566153]  [<ffffffffa05a23fb>] xfs_readdir+0x16b/0x1d0 [xfs]
[  104.566161]  [<ffffffffa05a483b>] xfs_file_readdir+0x2b/0x40 [xfs]
[  104.566163]  [<ffffffff812208d8>] iterate_dir+0xa8/0xe0
[  104.566166]  [<ffffffff8122ca7c>] ? fget_light+0x3c/0x4f0
[  104.566168]  [<ffffffff81220d83>] SyS_getdents+0x93/0x120
[  104.566170]  [<ffffffff81220a00>] ? fillonedir+0xf0/0xf0
[  104.566172]  [<ffffffff81134ecc>] ? __audit_syscall_entry+0x9c/0xf0
[  104.566174]  [<ffffffff8178db69>] system_call_fastpath+0x16/0x1b
--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ