[<prev] [next>] [thread-next>] [day] [month] [year] [list]
Message-Id: <1238578840.5970.178.camel@johannes.local>
Date: Wed, 01 Apr 2009 11:40:40 +0200
From: Johannes Berg <johannes@...solutions.net>
To: Jiri Kosina <jkosina@...e.cz>
Cc: Dmitry Torokhov <dmitry.torokhov@...il.com>,
linux-input <linux-input@...r.kernel.org>,
linux-kernel <linux-kernel@...r.kernel.org>,
"Rafael J. Wysocki" <rjw@...k.pl>, Oleg Nesterov <oleg@...hat.com>
Subject: Re: lockdep report at resume
On Tue, 2009-03-31 at 10:40 +0200, Jiri Kosina wrote:
> > > Could you please send me your config?
> > Sure, attached. I haven't yet tried to reproduce on .29 though, which
> > this config is for (but I haven't changed it since, only taken it
> > forward).
I've now gotten it again on 2.6.29-wl-20327-g8f2487d-dirty.
I've analysed a bit more.
Let's start from the bottom:
-> #0 (&dev->mutex){--..}:
[<ffffffff80276e87>] check_prev_add+0x57/0x770
[<ffffffff80277b96>] validate_chain+0x5f6/0x6b0
[<ffffffff8027808f>] __lock_acquire+0x43f/0xa10
[<ffffffff802786f1>] lock_acquire+0x91/0xc0
[<ffffffff805cb56c>] mutex_lock_nested+0xfc/0x390
[<ffffffff804d19e1>] input_disconnect_device+0x31/0xf0
[<ffffffff804d1aba>] input_unregister_device+0x1a/0x110
[<ffffffffa00df329>] bcm5974_disconnect+0x29/0x90 [bcm5974]
[<ffffffffa001c9ed>] usb_unbind_interface+0x6d/0x180 [usbcore]
[<ffffffff80498661>] __device_release_driver+0x81/0xc0
[<ffffffff804987c0>] device_release_driver+0x30/0x50
[<ffffffffa001ce48>] usb_driver_release_interface+0xc8/0xf0 [usbcore]
[<ffffffffa001cf99>] usb_forced_unbind_intf+0x39/0x90 [usbcore]
[<ffffffffa00121a5>] usb_reset_device+0xd5/0x220 [usbcore]
[<ffffffffa00957fa>] hid_reset+0x18a/0x280 [usbhid]
[<ffffffff8025d28d>] run_workqueue+0x10d/0x250
Here we have hid_reset being called off schedule_work. It eventually
calls into bcm5974 which will, from its usb_driver disconnect call, call
input_unregister_device(), which acquires &dev->mutex.
-> #1 (polldev_mutex){--..}:
[<ffffffff802771e7>] check_prev_add+0x3b7/0x770
[<ffffffff80277b96>] validate_chain+0x5f6/0x6b0
[<ffffffff8027808f>] __lock_acquire+0x43f/0xa10
[<ffffffff802786f1>] lock_acquire+0x91/0xc0
[<ffffffff805cb12c>] mutex_lock_interruptible_nested+0xec/0x430
[<ffffffffa0058151>] input_open_polled_device+0x21/0xd0 [input_polldev]
[<ffffffff804d1528>] input_open_device+0x98/0xc0
[<ffffffffa009f468>] evdev_open+0x1c8/0x1f0 [evdev]
[<ffffffff804d099f>] input_open_file+0x10f/0x200
[<ffffffff802e6e47>] chrdev_open+0x147/0x220
[<ffffffff802e18eb>] __dentry_open+0x11b/0x350
[<ffffffff802e1c37>] nameidata_to_filp+0x57/0x70
[<ffffffff802f0e6e>] do_filp_open+0x1fe/0x970
[<ffffffff802e16d0>] do_sys_open+0x80/0x110
[<ffffffff802e17a0>] sys_open+0x20/0x30
This is another code path -- evdev triggered here. Any input polldev
will acquire polldev_mutex within its struct input_dev->open() callback,
and thus create a dependency of &dev->mutex on polldev_mutex because
input_open_device() is called with &dev->mutex held.
-> #2 (cpu_add_remove_lock){--..}:
[<ffffffff802771e7>] check_prev_add+0x3b7/0x770
[<ffffffff80277b96>] validate_chain+0x5f6/0x6b0
[<ffffffff8027808f>] __lock_acquire+0x43f/0xa10
[<ffffffff802786f1>] lock_acquire+0x91/0xc0
[<ffffffff805cb56c>] mutex_lock_nested+0xfc/0x390
[<ffffffff8024a447>] cpu_maps_update_begin+0x17/0x20
[<ffffffff8025d7c8>] destroy_workqueue+0x38/0xb0
[<ffffffffa0058115>] input_close_polled_device+0x45/0x60 [input_polldev]
[<ffffffff804d145c>] input_close_device+0x5c/0x90
[<ffffffffa009f1e9>] evdev_release+0xa9/0xd0 [evdev]
[<ffffffff802e4e25>] __fput+0xd5/0x1e0
[<ffffffff802e4f55>] fput+0x25/0x30
[<ffffffff802e14f8>] filp_close+0x58/0x90
[<ffffffff802e15ee>] sys_close+0xbe/0x120
[<ffffffff8020bf0b>] system_call_fastpath+0x16/0x1b
[<ffffffffffffffff>] 0xffffffffffffffff
This is cute.
So input-polldev uses its own workqueue, and it's singlethread. But
destroy_workqueue must stop CPU hotplug anyway, calls
cpu_map_update_begin() which locks cpu_add_remove_lock.
-> #3 (events){--..}:
[<ffffffff802771e7>] check_prev_add+0x3b7/0x770
[<ffffffff80277b96>] validate_chain+0x5f6/0x6b0
[<ffffffff8027808f>] __lock_acquire+0x43f/0xa10
[<ffffffff802786f1>] lock_acquire+0x91/0xc0
[<ffffffff8025d722>] cleanup_workqueue_thread+0x42/0x90
[<ffffffff805bca8d>] workqueue_cpu_callback+0x9d/0x132
[<ffffffff805d1205>] notifier_call_chain+0x65/0xa0
[<ffffffff80267626>] raw_notifier_call_chain+0x16/0x20
[<ffffffff805ba91b>] _cpu_down+0x1db/0x350
[<ffffffff8024a5b5>] disable_nonboot_cpus+0xe5/0x170
[<ffffffff80287ef5>] hibernation_snapshot+0x135/0x170
[<ffffffff8028b8a5>] snapshot_ioctl+0x425/0x620
[<ffffffff802f27a6>] vfs_ioctl+0x36/0xb0
[<ffffffff802f2b89>] do_vfs_ioctl+0x89/0x350
[<ffffffff802f2e9f>] sys_ioctl+0x4f/0x80
[<ffffffff8020bf0b>] system_call_fastpath+0x16/0x1b
[<ffffffffffffffff>] 0xffffffffffffffff
Here we have hibernation, which needs to call disable_nonboot_cpus. This
takes down all CPUs, and causes the workqueue code, now running off the
workqueue_cpu_callback, to call cleanup_workqueue_thread(), which
"acquires" the workqueue. I suspect this will also happen if you go into
sysfs and disable a CPU manually, which may help you reproduce this.
disable_nonboot_cpus calls cpu_map_update_begin() to avoid other things
interfering, and thus creates the dependency of the workqueue on that.
-> #4 (&usbhid->reset_work){--..}:
[<ffffffff802771e7>] check_prev_add+0x3b7/0x770
[<ffffffff80277b96>] validate_chain+0x5f6/0x6b0
[<ffffffff8027808f>] __lock_acquire+0x43f/0xa10
[<ffffffff802786f1>] lock_acquire+0x91/0xc0
[<ffffffff8025d287>] run_workqueue+0x107/0x250
[<ffffffff8025d47f>] worker_thread+0xaf/0x130
[<ffffffff80261dae>] kthread+0x4e/0x90
[<ffffffff8020cfba>] child_rip+0xa/0x20
[<ffffffffffffffff>] 0xffffffffffffffff
Now, of course, usbhid->reset_work runs off the schedule_work workqueue,
which was stopped during hibernation, so it depends on that workqueue.
Finally, we're back at the top, with input_disconnect_device() acquiring
&dev->mutex.
Now, how can a deadlock happen?
I think it cannot -- unless you have a polled USB device. The two
"&dev->mutex" instances here are from difference devices, but lockdep
cannot tell them apart, and if you have a polled USB device then the
same can happen.
Assume you had a polled USB driver using input_unregister_polled_device,
and thus input_unregister_device, in its usb_driver disconnect call. In
that case you could potentially trigger the deadlock when you manage to
get that usb device reset very very close before calling
disable_nonboot_cpus, so close that the usb reset_work is still
scheduled or something like that...
I don't really see a good way to solve it -- but I hope the analysis
helps some -- also adding lots of people to CC.
johannes
Download attachment "signature.asc" of type "application/pgp-signature" (837 bytes)
Powered by blists - more mailing lists