[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <498473F3.3050208@gmail.com>
Date: Sat, 31 Jan 2009 16:53:23 +0100
From: Andrea Righi <righi.andrea@...il.com>
To: Andrey Borzenkov <arvidjaar@...l.ru>
CC: Geert Uytterhoeven <geert@...ux-m68k.org>,
Linux Frame Buffer Device Development
<linux-fbdev-devel@...ts.sourceforge.net>,
"Antonino A. Daplas" <adaplas@...il.com>,
linux-pm@...ts.linux-foundation.org,
Linux Kernel Development <linux-kernel@...r.kernel.org>
Subject: Re: [Linux-fbdev-devel] [2.6.29-rc2] fb_mmap: circular locking dependency
on hibernation
On 2009-01-30 11:23, Andrea Righi wrote:
> On 2009-01-30 05:15, Andrey Borzenkov wrote:
>> On 29 of January 2009 12:10:11 Geert Uytterhoeven wrote:
>>> On Tue, 27 Jan 2009, Andrey Borzenkov wrote:
>>>> I do not know if this is new; I usually do not hibernate at all and
>>>> was testing another driver actually. I do not remember such issues
>>>> in 2.6.28 though. Dmesg follows; config attached.
>>>>
>>>> [11715.652498]
>>>> =======================================================
>>>> [11715.652520] [ INFO: possible circular locking dependency
>>>> detected ] [11715.652534] 2.6.29-rc2-1avb #11
>>>> [11715.652541]
>>>> -------------------------------------------------------
>>>> [11715.652550] s2disk/25519 is trying to acquire lock:
>>>> [11715.652562] (&fb_info->lock){--..}, at: [<c0223017>]
>>>> fb_mmap+0x97/0x170 [11715.652623]
>>>> [11715.652625] but task is already holding lock:
>>>> [11715.652632] (&mm->mmap_sem){----}, at: [<c010639e>]
>>>> sys_mmap2+0x8e/0xc0 [11715.652669]
>>>> [11715.652671] which lock already depends on the new lock.
>>>> [11715.652675]
>>>> [11715.652681]
>>>> [11715.652683] the existing dependency chain (in reverse order) is:
>>>> [11715.652691]
>>>> [11715.652693] -> #1 (&mm->mmap_sem){----}:
>>>> [11715.652707] [<c0144ecf>] __lock_acquire+0x129f/0x1930
>>>> [11715.652730] [<c01455bc>] lock_acquire+0x5c/0x80
>>>> [11715.652742] [<c0181e57>] might_fault+0x77/0xa0
>>>> [11715.652764] [<c0209325>] copy_from_user+0x35/0x130
>>>> [11715.652782] [<c0223ef5>] fb_ioctl+0x1d5/0x3f0
>>>> [11715.652797] [<c01a51d0>] vfs_ioctl+0x20/0x80
>>>> [11715.652832] [<c01a53e2>] do_vfs_ioctl+0x72/0x570
>>>> [11715.652845] [<c01a5919>] sys_ioctl+0x39/0x70
>>>> [11715.652857] [<c0103371>] sysenter_do_call+0x12/0x31
>>>> [11715.652870] [<ffffffff>] 0xffffffff
>>>> [11715.652973]
>>>> [11715.652975] -> #0 (&fb_info->lock){--..}:
>>>> [11715.652988] [<c014503c>] __lock_acquire+0x140c/0x1930
>>>> [11715.653000] [<c01455bc>] lock_acquire+0x5c/0x80
>>>> [11715.653011] [<c030761a>] mutex_lock_nested+0xba/0x2f0
>>>> [11715.653032] [<c0223017>] fb_mmap+0x97/0x170
>>>> [11715.653044] [<c0188996>] mmap_region+0x1d6/0x530
>>>> [11715.653067] [<c0188ec3>] do_mmap_pgoff+0x1d3/0x2f0
>>>> [11715.653079] [<c01063bd>] sys_mmap2+0xad/0xc0
>>>> [11715.653091] [<c0103371>] sysenter_do_call+0x12/0x31
>>>> [11715.653102] [<ffffffff>] 0xffffffff
>>>> [11715.653116]
>>>> [11715.653119] other info that might help us debug this:
>>>> [11715.653122]
>>>> [11715.653130] 1 lock held by s2disk/25519:
>>>> [11715.653137] #0: (&mm->mmap_sem){----}, at: [<c010639e>]
>>>> sys_mmap2+0x8e/0xc0
>>>> [11715.653156]
>>>> [11715.653158] stack backtrace:
>>>> [11715.653170] Pid: 25519, comm: s2disk Not tainted 2.6.29-rc2-1avb
>>>> #11 [11715.653179] Call Trace:
>>>> [11715.653190] [<c0306248>] ? printk+0x18/0x20
>>>> [11715.653202] [<c01437df>] print_circular_bug_tail+0xcf/0xe0
>>>> [11715.653213] [<c014503c>] __lock_acquire+0x140c/0x1930
>>>> [11715.653224] [<c014328b>] ? trace_hardirqs_on+0xb/0x10
>>>> [11715.653242] [<c0309125>] ? _spin_unlock_irqrestore+0x35/0x60
>>>> [11715.653254] [<c0140dc5>] ? lock_release_holdtime+0x35/0x210
>>>> [11715.653265] [<c01455bc>] lock_acquire+0x5c/0x80
>>>> [11715.653277] [<c0223017>] ? fb_mmap+0x97/0x170
>>>> [11715.653287] [<c030761a>] mutex_lock_nested+0xba/0x2f0
>>>> [11715.653298] [<c0223017>] ? fb_mmap+0x97/0x170
>>>> [11715.653309] [<c0223017>] ? fb_mmap+0x97/0x170
>>>> [11715.653335] [<c019514d>] ? kmem_cache_alloc+0xad/0x100
>>>> [11715.653347] [<c0223017>] fb_mmap+0x97/0x170
>>>> [11715.653358] [<c0188996>] mmap_region+0x1d6/0x530
>>>> [11715.653371] [<c0186a9d>] ?
>>>> arch_get_unmapped_area_topdown+0x10d/0x170 [11715.653383]
>>>> [<c0188ec3>] do_mmap_pgoff+0x1d3/0x2f0
>>>> [11715.653395] [<c01063bd>] sys_mmap2+0xad/0xc0
>>>> [11715.653406] [<c0103371>] sysenter_do_call+0x12/0x31
>>> Is this fixed by http://patchwork.kernel.org/patch/3259/?
>>>
>> No. It just makes stack a bit more deep :)
>
> yep! :)
>
> Probably we should simply use an atomic notifier there, instead of blocking.
> I'll do some tests and post a new fix.
No, the solution is no so trivial. Some notifier callbacks can sleep, so
the usage of a blocking notifier is mandatory. It seems we need a more
complicated solution to totally fix it. At least the previous fix avoids
the circular locking of fb_info->lock and mm->mmap_sem due to
copy_from/to_user() called with fb_info->lock held...
-Andrea
>
> Thanks,
> -Andrea
>
>> [ 408.267429] =======================================================
>> [ 408.267448] [ INFO: possible circular locking dependency detected ]
>> [ 408.267462] 2.6.29-rc3-1avb #14
>> [ 408.267469] -------------------------------------------------------
>> [ 408.267478] s2disk/5695 is trying to acquire lock:
>> [ 408.267487] (&fb_info->lock){--..}, at: [<c02235e7>]
>> fb_mmap+0x97/0x170
>> [ 408.267544]
>> [ 408.267546] but task is already holding lock:
>> [ 408.267553] (&mm->mmap_sem){----}, at: [<c010639e>]
>> sys_mmap2+0x8e/0xc0
>> [ 408.267591]
>> [ 408.267593] which lock already depends on the new lock.
>> [ 408.267596]
>> [ 408.267603]
>> [ 408.267604] the existing dependency chain (in reverse order) is:
>> [ 408.267612]
>> [ 408.267614] -> #3 (&mm->mmap_sem){----}:
>> [ 408.267628] [<c014520f>] __lock_acquire+0x129f/0x1930
>> [ 408.267650] [<c01458fc>] lock_acquire+0x5c/0x80
>> [ 408.267661] [<c0182117>] might_fault+0x77/0xa0
>> [ 408.267683] [<c0209916>] copy_to_user+0x36/0x120
>> [ 408.267699] [<c01a60f7>] filldir+0x97/0xe0
>> [ 408.267735] [<c01de339>] sysfs_readdir+0x129/0x220
>> [ 408.267751] [<c01a62a6>] vfs_readdir+0x86/0xa0
>> [ 408.267763] [<c01a63e8>] sys_getdents+0x68/0xc0
>> [ 408.267775] [<c010344a>] syscall_call+0x7/0xb
>> [ 408.267787] [<ffffffff>] 0xffffffff
>> [ 408.267862]
>> [ 408.267864] -> #2 (sysfs_mutex){--..}:
>> [ 408.267877] [<c014520f>] __lock_acquire+0x129f/0x1930
>> [ 408.267889] [<c01458fc>] lock_acquire+0x5c/0x80
>> [ 408.267899] [<c0307f3a>] mutex_lock_nested+0xba/0x2f0
>> [ 408.267921] [<c01de6cc>] sysfs_addrm_start+0x2c/0xc0
>> [ 408.267933] [<c01dec80>] create_dir+0x40/0x90
>> [ 408.267944] [<c01decfb>] sysfs_create_dir+0x2b/0x50
>> [ 408.267955] [<c02036bc>] kobject_add_internal+0xbc/0x1b0
>> [ 408.267984] [<c0203881>] kobject_add_varg+0x31/0x50
>> [ 408.267996] [<c02038fc>] kobject_add+0x2c/0x60
>> [ 408.268008] [<c0273448>] device_add+0xa8/0x550
>> [ 408.268041] [<c0273902>] device_register+0x12/0x20
>> [ 408.268053] [<c02739bb>] device_create_vargs+0xab/0xc0
>> [ 408.268065] [<c02739f8>] device_create+0x28/0x30
>> [ 408.268077] [<c026cc6d>] register_con_driver+0xed/0x130
>> [ 408.268090] [<c026e18b>] take_over_console+0x1b/0x50
>> [ 408.268102] [<c023052d>] fbcon_takeover+0x5d/0xb0
>> [ 408.268125] [<c02311a0>] fbcon_event_notify+0x820/0x900
>> [ 408.268137] [<c030c3b3>] notifier_call_chain+0x53/0xa0
>> [ 408.268161] [<c0138ab4>]
>> __blocking_notifier_call_chain+0x44/0x60
>> [ 408.268184] [<c0138aea>]
>> blocking_notifier_call_chain+0x1a/0x20
>> [ 408.268197] [<c0222f81>] fb_notifier_call_chain+0x11/0x20
>> [ 408.268209] [<c0223f08>] register_framebuffer+0x168/0x220
>> [ 408.268223] [<c0414ba2>] vesafb_probe+0x542/0x783
>> [ 408.268250] [<c02765af>] platform_drv_probe+0xf/0x20
>> [ 408.268269] [<c02757c7>] driver_probe_device+0x87/0x1b0
>> [ 408.268281] [<c0275988>] __device_attach+0x8/0x10
>> [ 408.268291] [<c0274e0b>] bus_for_each_drv+0x5b/0x80
>> [ 408.268308] [<c0275a36>] device_attach+0x76/0x80
>> [ 408.268319] [<c0274c27>] bus_attach_device+0x47/0x70
>> [ 408.268331] [<c02736c3>] device_add+0x323/0x550
>> [ 408.268342] [<c0276f85>] platform_device_add+0x175/0x1c0
>> [ 408.268358] [<c0414e7d>] vesafb_init+0x9a/0x1ec
>> [ 408.268370] [<c010111a>] do_one_initcall+0x2a/0x160
>> [ 408.268382] [<c04004d5>] kernel_init+0x83/0xd5
>> [ 408.268411] [<c0103a77>] kernel_thread_helper+0x7/0x10
>> [ 408.268423] [<ffffffff>] 0xffffffff
>> [ 408.268436]
>> [ 408.268438] -> #1 ((fb_notifier_list).rwsem){----}:
>> [ 408.268451] [<c014520f>] __lock_acquire+0x129f/0x1930
>> [ 408.268463] [<c01458fc>] lock_acquire+0x5c/0x80
>> [ 408.268474] [<c0308879>] down_read+0x49/0x90
>> [ 408.268484] [<c0138a9a>]
>> __blocking_notifier_call_chain+0x2a/0x60
>> [ 408.268497] [<c0138aea>]
>> blocking_notifier_call_chain+0x1a/0x20
>> [ 408.268509] [<c0222f81>] fb_notifier_call_chain+0x11/0x20
>> [ 408.268521] [<c022459e>] do_fb_ioctl+0x2ae/0x5d0
>> [ 408.268533] [<c02248dd>] fb_ioctl+0x1d/0x20
>> [ 408.268545] [<c01a54b0>] vfs_ioctl+0x20/0x80
>> [ 408.268558] [<c01a56c2>] do_vfs_ioctl+0x72/0x570
>> [ 408.268570] [<c01a5bf9>] sys_ioctl+0x39/0x70
>> [ 408.268581] [<c0103371>] sysenter_do_call+0x12/0x31
>> [ 408.268592] [<ffffffff>] 0xffffffff
>> [ 408.268623]
>> [ 408.268625] -> #0 (&fb_info->lock){--..}:
>> [ 408.268638] [<c014537c>] __lock_acquire+0x140c/0x1930
>> [ 408.268649] [<c01458fc>] lock_acquire+0x5c/0x80
>> [ 408.268661] [<c0307f3a>] mutex_lock_nested+0xba/0x2f0
>> [ 408.268672] [<c02235e7>] fb_mmap+0x97/0x170
>> [ 408.268684] [<c0188c76>] mmap_region+0x1d6/0x530
>> [ 408.268706] [<c01891a3>] do_mmap_pgoff+0x1d3/0x2f0
>> [ 408.268718] [<c01063bd>] sys_mmap2+0xad/0xc0
>> [ 408.268729] [<c0103371>] sysenter_do_call+0x12/0x31
>> [ 408.268740] [<ffffffff>] 0xffffffff
>> [ 408.268752]
>> [ 408.268754] other info that might help us debug this:
>> [ 408.268758]
>> [ 408.268767] 1 lock held by s2disk/5695:
>> [ 408.268773] #0: (&mm->mmap_sem){----}, at: [<c010639e>]
>> sys_mmap2+0x8e/0xc0
>> [ 408.268792]
>> [ 408.268794] stack backtrace:
>> [ 408.268805] Pid: 5695, comm: s2disk Not tainted 2.6.29-rc3-1avb #14
>> [ 408.268814] Call Trace:
>> [ 408.268825] [<c0306b68>] ? printk+0x18/0x20
>> [ 408.268836] [<c0143b1f>] print_circular_bug_tail+0xcf/0xe0
>> [ 408.268847] [<c014537c>] __lock_acquire+0x140c/0x1930
>> [ 408.268862] [<c01487f0>] ? futex_wait+0x170/0x450
>> [ 408.268881] [<c011826c>] ? try_to_wake_up+0x10c/0x120
>> [ 408.268894] [<c0309a45>] ? _spin_unlock_irqrestore+0x35/0x60
>> [ 408.268906] [<c0141105>] ? lock_release_holdtime+0x35/0x210
>> [ 408.268917] [<c01458fc>] lock_acquire+0x5c/0x80
>> [ 408.268928] [<c02235e7>] ? fb_mmap+0x97/0x170
>> [ 408.268939] [<c0307f3a>] mutex_lock_nested+0xba/0x2f0
>> [ 408.268950] [<c02235e7>] ? fb_mmap+0x97/0x170
>> [ 408.268961] [<c02235e7>] ? fb_mmap+0x97/0x170
>> [ 408.268987] [<c019542d>] ? kmem_cache_alloc+0xad/0x100
>> [ 408.268999] [<c02235e7>] fb_mmap+0x97/0x170
>> [ 408.269010] [<c0188c76>] mmap_region+0x1d6/0x530
>> [ 408.269023] [<c0186d7d>] ?
>> arch_get_unmapped_area_topdown+0x10d/0x170
>> [ 408.269035] [<c01891a3>] do_mmap_pgoff+0x1d3/0x2f0
>> [ 408.269047] [<c01063bd>] sys_mmap2+0xad/0xc0
>> [ 408.269058] [<c0103371>] sysenter_do_call+0x12/0x31
--
Andrea Righi
Jabber ID: righiandr@...ber.linux.it
gpg --keyserver pgp.mit.edu --recv-keys 77CEF397
--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/
Powered by blists - more mailing lists