[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <4CF95995.1070506@gmail.com>
Date: Fri, 03 Dec 2010 21:56:53 +0100
From: Jarek Poplawski <jarkao2@...il.com>
To: Yinghai Lu <yinghai@...nel.org>
CC: David Miller <davem@...emloft.net>,
NetDev <netdev@...r.kernel.org>, Ingo Molnar <mingo@...e.hu>,
Ron Mercer <ron.mercer@...gic.com>, linux-driver@...gic.com
Subject: Re: qlge warning
It looks like cancel_delayed_work_sync in ql_adapter_down is illegal.
We can't sync works with rtnl_lock while holding it in qlge_close.
Maintainers CC'ed.
Jarek P.
Yinghai Lu wrote:
> [ 290.233264] =======================================================
> [ 290.251780] [ INFO: possible circular locking dependency detected ]
> [ 290.271534] 2.6.37-rc4-tip-yh-05919-geb30094-dirty #308
> [ 290.271775] -------------------------------------------------------
> [ 290.291512] swapper/1 is trying to acquire lock:
> [ 290.291725] ((&(&qdev->mpi_port_cfg_work)->work)){+.+...}, at:
> [<ffffffff81096419>] wait_on_work+0x0/0xff
> [ 290.311643]
> [ 290.311644] but task is already holding lock:
> [ 290.311915] (rtnl_mutex){+.+.+.}, at: [<ffffffff81bb094d>]
> rtnl_lock+0x17/0x19
> [ 290.331681]
> [ 290.331682] which lock already depends on the new lock.
> [ 290.331684]
> [ 290.351491]
> [ 290.351492] the existing dependency chain (in reverse order) is:
> [ 290.351830]
> [ 290.351831] -> #1 (rtnl_mutex){+.+.+.}:
> [ 290.371562] [<ffffffff810ae6b6>] lock_acquire+0xca/0xf0
> [ 290.371824] [<ffffffff81cdbf5d>] mutex_lock_nested+0x60/0x2b8
> [ 290.391539] [<ffffffff81bb094d>] rtnl_lock+0x17/0x19
> [ 290.411250] [<ffffffff818501ad>] ql_mpi_port_cfg_work+0x1f/0x1ad
> [ 290.411606] [<ffffffff81095189>] process_one_work+0x234/0x3e8
> [ 290.431282] [<ffffffff81095663>] worker_thread+0x17f/0x261
> [ 290.431583] [<ffffffff8109a633>] kthread+0xa0/0xa8
> [ 290.451279] [<ffffffff8103a914>] kernel_thread_helper+0x4/0x10
> [ 290.451581]
> [ 290.451582] -> #0 ((&(&qdev->mpi_port_cfg_work)->work)){+.+...}:
> [ 290.471483] [<ffffffff810ada85>] __lock_acquire+0x113c/0x1813
> [ 290.491177] [<ffffffff810ae6b6>] lock_acquire+0xca/0xf0
> [ 290.491451] [<ffffffff8109646c>] wait_on_work+0x53/0xff
> [ 290.511128] [<ffffffff810965da>] __cancel_work_timer+0xc2/0x102
> [ 290.511434] [<ffffffff8109662c>] cancel_delayed_work_sync+0x12/0x14
> [ 290.531233] [<ffffffff81847646>] ql_cancel_all_work_sync+0x64/0x68
> [ 290.531563] [<ffffffff818499d5>] ql_adapter_down+0x23/0xf6
> [ 290.551298] [<ffffffff81849ca7>] qlge_close+0x67/0x76
> [ 290.571015] [<ffffffff81ba3853>] __dev_close+0x7b/0x89
> [ 290.571297] [<ffffffff81ba5535>] __dev_change_flags+0xad/0x131
> [ 290.590974] [<ffffffff81ba563a>] dev_change_flags+0x21/0x57
> [ 290.591280] [<ffffffff827de30e>] ic_close_devs+0x2e/0x48
> [ 290.610978] [<ffffffff827df332>] ip_auto_config+0xbc9/0xe84
> [ 290.611280] [<ffffffff810002da>] do_one_initcall+0x57/0x135
> [ 290.630977] [<ffffffff8278ef8a>] kernel_init+0x16c/0x1f6
> [ 290.631263] [<ffffffff8103a914>] kernel_thread_helper+0x4/0x10
> [ 290.651000]
> [ 290.651001] other info that might help us debug this:
> [ 290.651003]
> [ 290.670829] 1 lock held by swapper/1:
> [ 290.671013] #0: (rtnl_mutex){+.+.+.}, at: [<ffffffff81bb094d>]
> rtnl_lock+0x17/0x19
> [ 290.690819]
> [ 290.690820] stack backtrace:
> [ 290.691054] Pid: 1, comm: swapper Not tainted
> 2.6.37-rc4-tip-yh-05919-geb30094-dirty #308
> [ 290.710805] Call Trace:
> [ 290.710938] [<ffffffff810aa296>] ? print_circular_bug+0xaf/0xbe
> [ 290.730683] [<ffffffff810ada85>] ? __lock_acquire+0x113c/0x1813
> [ 290.730955] [<ffffffff81095d70>] ? wait_on_cpu_work+0xdb/0x114
> [ 290.750672] [<ffffffff81096419>] ? wait_on_work+0x0/0xff
> [ 290.750939] [<ffffffff810ae6b6>] ? lock_acquire+0xca/0xf0
> [ 290.770664] [<ffffffff81096419>] ? wait_on_work+0x0/0xff
> [ 290.770920] [<ffffffff8109646c>] ? wait_on_work+0x53/0xff
> [ 290.790575] [<ffffffff81096419>] ? wait_on_work+0x0/0xff
> [ 290.790821] [<ffffffff810965da>] ? __cancel_work_timer+0xc2/0x102
> [ 290.810559] [<ffffffff8109662c>] ? cancel_delayed_work_sync+0x12/0x14
> [ 290.810855] [<ffffffff81847646>] ? ql_cancel_all_work_sync+0x64/0x68
> [ 290.830594] [<ffffffff818499d5>] ? ql_adapter_down+0x23/0xf6
> [ 290.830867] [<ffffffff81849ca7>] ? qlge_close+0x67/0x76
> [ 290.850568] [<ffffffff81ba3853>] ? __dev_close+0x7b/0x89
> [ 290.850829] [<ffffffff81ba5535>] ? __dev_change_flags+0xad/0x131
> [ 290.870540] [<ffffffff81ba563a>] ? dev_change_flags+0x21/0x57
> [ 290.870815] [<ffffffff827de30e>] ? ic_close_devs+0x2e/0x48
> [ 290.890595] [<ffffffff827df332>] ? ip_auto_config+0xbc9/0xe84
> [ 290.910247] [<ffffffff81cda1e3>] ? printk+0x41/0x43
> [ 290.910488] [<ffffffff827de769>] ? ip_auto_config+0x0/0xe84
> [ 290.910747] [<ffffffff810002da>] ? do_one_initcall+0x57/0x135
> [ 290.930455] [<ffffffff8278ef8a>] ? kernel_init+0x16c/0x1f6
> [ 290.930743] [<ffffffff8103a914>] ? kernel_thread_helper+0x4/0x10
> [ 290.950419] [<ffffffff81cde23c>] ? restore_args+0x0/0x30
> [ 290.970152] [<ffffffff8278ee1e>] ? kernel_init+0x0/0x1f6
> [ 290.970398] [<ffffffff8103a910>] ? kernel_thread_helper+0x0/0x10
> --
> To unsubscribe from this list: send the line "unsubscribe netdev" in
> the body of a message to majordomo@...r.kernel.org
> More majordomo info at http://vger.kernel.org/majordomo-info.html
>
--
To unsubscribe from this list: send the line "unsubscribe netdev" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at http://vger.kernel.org/majordomo-info.html
Powered by blists - more mailing lists