lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [<thread-prev] [day] [month] [year] [list]
Date:	Thu, 28 Feb 2008 10:44:37 +0530
From:	Gautham R Shenoy <ego@...ibm.com>
To:	Andrew Morton <akpm@...ux-foundation.org>
Cc:	Zdenek Kabelac <zdenek.kabelac@...il.com>,
	Kernel development list <linux-kernel@...r.kernel.org>,
	linux-usb@...r.kernel.org
Subject: Re: INFO: possible circular locking in the resume

On Tue, Feb 26, 2008 at 04:17:59PM -0800, Andrew Morton wrote:
> On Wed, 27 Feb 2008 00:29:21 +0100 "Zdenek Kabelac" <zdenek.kabelac@...il.com> wrote:
> 
> > Hi
> > 
> > I'm getting this INFO: from suspend/resume process.
> > (Kernel uses patches from this bugzilla to survive suspend
> > http://bugzilla.kernel.org/show_bug.cgi?id=10030)
> > (Config is available in the bugzilla)
> > 
> > 
> > [   42.097522] MMC: killing requests for dead queue
> > [   42.098738] PM: Finishing wakeup.
> > [   42.098740] Restarting tasks ... <6>usb 3-2: USB disconnect, address 2
> > [   42.129810] done.
> > [   42.258523]
> > [   42.258524] =======================================================
> > [   42.258626] [ INFO: possible circular locking dependency detected ]
> > [   42.258681] 2.6.25-rc3 #81
> > [   42.258736] -------------------------------------------------------
> > [   42.258787] sh/1973 is trying to acquire lock:
> > [   42.258839]  (&cpu_hotplug.lock){--..}, at: [<ffffffff8106d7a5>]
> > get_online_cpus+0x35/0x50
> 
> heh.

This have been pretty silent on this front for quite sometime now :)

This has nothing to do with bluetooth/srcu/cpufreq per se. 

If I am not mistaken, this can be reproduced by running cpu-hotplug 
in a tight loop, while trying to change the cpufreq governor to userspace.

The problem here being that on the read-path, we use cpu_hotplug.lock
to just increment the refcount. We don't hold the lock once we are done
with that. But on the write path, we hold the lock to block any readers.

And thus, in lockdep's dependency chain, we have this dependency between
sp->mutex --> per_cpu(cpu_policy_rwsem, cpu) --> cpu_hotplug.lock in the
reverse order in the write path.
and 

cpu_hotplug.lock --> sp->mutex on the read-path, i.e synchronize_srcu().

This can be solved by using a spin_lock to increment the refcount and
check for the writer, and a waitqueue for refcount
queueing the readers when a write is in progress, thus getting rid of
the mutex totally.

I have the patch, but it needs to be tested. Will post it later today.

Thanks for pointing this out :)

> 
> > [   42.259068]
> > [   42.259068] but task is already holding lock:
> > [   42.259166]  (&sp->mutex){--..}, at: [<ffffffff8105adb0>]
> > synchronize_srcu+0x20/0x90
> > [   42.259376]
> > [   42.259377] which lock already depends on the new lock.
> > [   42.259377]
> > [   42.259501]
> > [   42.259501] the existing dependency chain (in reverse order) is:
> > [   42.259501]
> > [   42.259501] -> #3 (&sp->mutex){--..}:
> > [   42.259501]        [<ffffffff81066d30>] __lock_acquire+0xfb0/0x1040
> > [   42.259501]        [<ffffffff81066e68>] lock_acquire+0xa8/0xf0
> > [   42.259501]        [<ffffffff812ef5ac>] mutex_lock_nested+0xcc/0x370
> > [   42.259501]        [<ffffffff8105adb0>] synchronize_srcu+0x20/0x90
> > [   42.259501]        [<ffffffff8105b2b4>]
> > srcu_notifier_chain_unregister+0x74/0xe0
> > [   42.259501]        [<ffffffff8125a337>] cpufreq_unregister_notifier+0x17/0x40
> > [   42.259501]        [<ffffffff8829e296>]
> > cpufreq_governor_userspace+0x1b6/0x1f4 [cpufreq_userspace]
> > [   42.259501]        [<ffffffff81259923>] __cpufreq_governor+0xd3/0x150
> > [   42.259501]        [<ffffffff8125a0b5>] __cpufreq_set_policy+0x115/0x170
> > [   42.259501]        [<ffffffff8125a819>] store_scaling_governor+0xc9/0x260
> > [   42.259501]        [<ffffffff8125b527>] store+0x87/0xa0
> > [   42.259501]        [<ffffffff811101fa>] sysfs_write_file+0xca/0x140
> > [   42.259501]        [<ffffffff810bb82b>] vfs_write+0xcb/0x170
> > [   42.259501]        [<ffffffff810bb9c0>] sys_write+0x50/0x90
> > [   42.259501]        [<ffffffff8100c70a>] system_call_after_swapgs+0x8a/0x8f
> > [   42.259501]        [<ffffffffffffffff>] 0xffffffffffffffff
> > [   42.259501]
> > [   42.259501] -> #2 (userspace_mutex){--..}:
> > [   42.259501]        [<ffffffff81066d30>] __lock_acquire+0xfb0/0x1040
> > [   42.259501]        [<ffffffff81066e68>] lock_acquire+0xa8/0xf0
> > [   42.259501]        [<ffffffff812ef5ac>] mutex_lock_nested+0xcc/0x370
> > [   42.259501]        [<ffffffff8829e13c>]
> > cpufreq_governor_userspace+0x5c/0x1f4 [cpufreq_userspace]
> > [   42.259501]        [<ffffffff81259923>] __cpufreq_governor+0xd3/0x150
> > [   42.259501]        [<ffffffff8125a0cb>] __cpufreq_set_policy+0x12b/0x170
> > [   42.259501]        [<ffffffff8125a819>] store_scaling_governor+0xc9/0x260
> > [   42.259501]        [<ffffffff8125b527>] store+0x87/0xa0
> > [   42.259501]        [<ffffffff811101fa>] sysfs_write_file+0xca/0x140
> > [   42.259501]        [<ffffffff810bb82b>] vfs_write+0xcb/0x170
> > [   42.259501]        [<ffffffff810bb9c0>] sys_write+0x50/0x90
> > [   42.259501]        [<ffffffff8100c70a>] system_call_after_swapgs+0x8a/0x8f
> > [   42.259501]        [<ffffffffffffffff>] 0xffffffffffffffff
> > [   42.259501]
> > [   42.259501] -> #1 (&per_cpu(cpu_policy_rwsem, cpu)){----}:
> > [   42.259501]        [<ffffffff81066d30>] __lock_acquire+0xfb0/0x1040
> > [   42.259501]        [<ffffffff81066e68>] lock_acquire+0xa8/0xf0
> > [   42.259501]        [<ffffffff812effc8>] down_write+0x38/0x70
> > [   42.259501]        [<ffffffff8125ae7c>] lock_policy_rwsem_write+0x4c/0x90
> > [   42.259501]        [<ffffffff812edbed>] cpufreq_cpu_callback+0x6d/0x90
> > [   42.259501]        [<ffffffff812f578f>] notifier_call_chain+0x3f/0x80
> > [   42.259501]        [<ffffffff8105b059>] __raw_notifier_call_chain+0x9/0x10
> > [   42.259501]        [<ffffffff8106d368>] _cpu_down+0xa8/0x290
> > [   42.259501]        [<ffffffff8106d621>] disable_nonboot_cpus+0x71/0x110
> > [   42.259501]        [<ffffffff810730d5>] suspend_devices_and_enter+0xe5/0x1a0
> > [   42.259501]        [<ffffffff81073306>] enter_state+0x146/0x1d0
> > [   42.259501]        [<ffffffff8107344a>] state_store+0xba/0x100
> > [   42.259501]        [<ffffffff811757f7>] kobj_attr_store+0x17/0x20
> > [   42.259501]        [<ffffffff811101fa>] sysfs_write_file+0xca/0x140
> > [   42.259501]        [<ffffffff810bb82b>] vfs_write+0xcb/0x170
> > [   42.259501]        [<ffffffff810bb9c0>] sys_write+0x50/0x90
> > [   42.259501]        [<ffffffff8100c70a>] system_call_after_swapgs+0x8a/0x8f
> > [   42.259501]        [<ffffffffffffffff>] 0xffffffffffffffff
> > [   42.259501]
> > [   42.259501] -> #0 (&cpu_hotplug.lock){--..}:
> > [   42.259501]        [<ffffffff81066c3d>] __lock_acquire+0xebd/0x1040
> > [   42.259501]        [<ffffffff81066e68>] lock_acquire+0xa8/0xf0
> > [   42.259501]        [<ffffffff812ef5ac>] mutex_lock_nested+0xcc/0x370
> > [   42.259501]        [<ffffffff8106d7a5>] get_online_cpus+0x35/0x50
> > [   42.259501]        [<ffffffff8103837e>] sched_getaffinity+0x1e/0xa0
> > [   42.259501]        [<ffffffff8108b386>] __synchronize_sched+0x16/0x90
> > [   42.259501]        [<ffffffff8105add5>] synchronize_srcu+0x45/0x90
> > [   42.259501]        [<ffffffff8105b2b4>]
> > srcu_notifier_chain_unregister+0x74/0xe0
> > [   42.259501]        [<ffffffff8125a337>] cpufreq_unregister_notifier+0x17/0x40
> > [   42.259501]        [<ffffffff8829e296>]
> > cpufreq_governor_userspace+0x1b6/0x1f4 [cpufreq_userspace]
> > [   42.259501]        [<ffffffff81259923>] __cpufreq_governor+0xd3/0x150
> > [   42.259501]        [<ffffffff8125a0b5>] __cpufreq_set_policy+0x115/0x170
> > [   42.259501]        [<ffffffff8125a819>] store_scaling_governor+0xc9/0x260
> > [   42.259501]        [<ffffffff8125b527>] store+0x87/0xa0
> > [   42.259501]        [<ffffffff811101fa>] sysfs_write_file+0xca/0x140
> > [   42.259501]        [<ffffffff810bb82b>] vfs_write+0xcb/0x170
> > [   42.259501]        [<ffffffff810bb9c0>] sys_write+0x50/0x90
> > [   42.259501]        [<ffffffff8100c70a>] system_call_after_swapgs+0x8a/0x8f
> > [   42.259501]        [<ffffffffffffffff>] 0xffffffffffffffff
> > [   42.259501]
> > [   42.259501] other info that might help us debug this:
> > [   42.259501]
> > [   42.259501] 4 locks held by sh/1973:
> > [   42.259501]  #0:  (&buffer->mutex){--..}, at: [<ffffffff81110173>]
> > sysfs_write_file+0x43/0x140
> > [   42.259501]  #1:  (&per_cpu(cpu_policy_rwsem, cpu)){----}, at:
> > [<ffffffff8125ae7c>] lock_policy_rwsem_write+0x4c/0x90
> > [   42.259501]  #2:  (userspace_mutex){--..}, at: [<ffffffff8829e22e>]
> > cpufreq_governor_userspace+0x14e/0x1f4 [cpufreq_userspace]
> > [   42.259501]  #3:  (&sp->mutex){--..}, at: [<ffffffff8105adb0>]
> > synchronize_srcu+0x20/0x90
> > [   42.259501]
> > [   42.259501] stack backtrace:
> > [   42.259501] Pid: 1973, comm: sh Not tainted 2.6.25-rc3 #81
> > [   42.259501]
> > [   42.259501] Call Trace:
> > [   42.259501]  [<ffffffff81065ac4>] print_circular_bug_tail+0x84/0x90
> > [   42.259501]  [<ffffffff81065872>] ? print_circular_bug_entry+0x52/0x60
> > [   42.259501]  [<ffffffff81066c3d>] __lock_acquire+0xebd/0x1040
> > [   42.259501]  [<ffffffff8106353b>] ? find_usage_backwards+0xcb/0x110
> > [   42.259502]  [<ffffffff8106353b>] ? find_usage_backwards+0xcb/0x110
> > [   42.259502]  [<ffffffff8106d7a5>] ? get_online_cpus+0x35/0x50
> > [   42.259502]  [<ffffffff81066e68>] lock_acquire+0xa8/0xf0
> > [   42.259502]  [<ffffffff8106d7a5>] ? get_online_cpus+0x35/0x50
> > [   42.259502]  [<ffffffff812ef5ac>] mutex_lock_nested+0xcc/0x370
> > [   42.259502]  [<ffffffff8106d7a5>] ? get_online_cpus+0x35/0x50
> > [   42.259502]  [<ffffffff81065649>] ? trace_hardirqs_on+0x139/0x1a0
> > [   42.259502]  [<ffffffff8106d7a5>] get_online_cpus+0x35/0x50
> > [   42.259502]  [<ffffffff8103837e>] sched_getaffinity+0x1e/0xa0
> > [   42.259502]  [<ffffffff8108b386>] __synchronize_sched+0x16/0x90
> > [   42.259502]  [<ffffffff8105add5>] synchronize_srcu+0x45/0x90
> > [   42.259502]  [<ffffffff8105b2b4>] srcu_notifier_chain_unregister+0x74/0xe0
> > [   42.259502]  [<ffffffff8125a337>] cpufreq_unregister_notifier+0x17/0x40
> > [   42.259502]  [<ffffffff8829e296>]
> > :cpufreq_userspace:cpufreq_governor_userspace+0x1b6/0x1f4
> > [   42.259502]  [<ffffffff81259923>] __cpufreq_governor+0xd3/0x150
> > [   42.259502]  [<ffffffff8125a0b5>] __cpufreq_set_policy+0x115/0x170
> > [   42.259502]  [<ffffffff8125a819>] store_scaling_governor+0xc9/0x260
> > [   42.259502]  [<ffffffff8125b650>] ? handle_update+0x0/0x10
> > [   42.259502]  [<ffffffff812efff0>] ? down_write+0x60/0x70
> > [   42.259502]  [<ffffffff8125b527>] store+0x87/0xa0
> > [   42.259502]  [<ffffffff811101fa>] sysfs_write_file+0xca/0x140
> > [   42.259502]  [<ffffffff810bb82b>] vfs_write+0xcb/0x170
> > [   42.259502]  [<ffffffff810bb9c0>] sys_write+0x50/0x90
> > [   42.259502]  [<ffffffff8100c70a>] system_call_after_swapgs+0x8a/0x8f
> > [   42.259502]
> > [   41.439840] mmc0: new SD card at address 5a61
> > [   41.440027] mmcblk0: mmc0:5a61 S016B 14560KiB
> > [   41.440123]  mmcblk0: p1
> > [   41.916006] usb 3-2: new full speed USB device using uhci_hcd and address 3
> > [   42.900716] usb 3-2: configuration #1 chosen from 1 choice
> > [   42.080814] usb 3-2: New USB device found, idVendor=0483, idProduct=2016
> > [   42.080939] usb 3-2: New USB device strings: Mfr=1, Product=2, SerialNumber=0
> > [   42.081094] usb 3-2: Product: Biometric Coprocessor
> > [   42.081212] usb 3-2: Manufacturer: STMicroelectronics
> > [   47.925651] PM: Syncing filesystems ... done.
> > [   47.925845] PM: Preparing system for mem sleep
> > [   47.925925] Freezing user space processes ... (elapsed 0.00 seconds) done.
> > [   47.926302] Freezing remaining freezable tasks ... (elapsed 0.00
> > seconds) done.
> > [   47.926596] PM: Entering mem sleep
> > [   47.929993] ACPI: Preparing to enter system sleep state S3
> > [   49.169572] sd 0:0:0:0: [sda] Synchronizing SCSI cache
> > [   49.169771] sd 0:0:0:0: [sda] Stopping disk
> > [   51.204365] mmc0: card 5a61 removed
> > [   51.204365] ACPI handle has no context!
> > [   51.204365] ACPI: PCI interrupt for device 0000:15:00.2 disabled
> > [   51.204365] ACPI handle has no context!
> 
> Something for Gautham and the USB people to scratch their heads over.

-- 
Thanks and Regards
gautham
--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ