lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [thread-next>] [day] [month] [year] [list]
Date:	Thu, 3 Oct 2013 15:08:30 -0400
From:	Dave Jones <davej@...hat.com>
To:	Linux Kernel <linux-kernel@...r.kernel.org>
Cc:	a.p.zijlstra@...llo.nl, gregkh@...uxfoundation.org,
	peter@...leysoftware.com
Subject: tty/perf lockdep trace.

 ======================================================
 [ INFO: possible circular locking dependency detected ]
 3.12.0-rc3+ #92 Not tainted
 -------------------------------------------------------
 trinity-child2/15191 is trying to acquire lock:
  (&rdp->nocb_wq){......}, at: [<ffffffff8108ff43>] __wake_up+0x23/0x50
 
but task is already holding lock:
  (&ctx->lock){-.-...}, at: [<ffffffff81154c19>] perf_event_exit_task+0x109/0x230

which lock already depends on the new lock.


the existing dependency chain (in reverse order) is:

-> #3 (&ctx->lock){-.-...}:
        [<ffffffff810cc243>] lock_acquire+0x93/0x200
        [<ffffffff81733f90>] _raw_spin_lock+0x40/0x80
        [<ffffffff811500ff>] __perf_event_task_sched_out+0x2df/0x5e0
        [<ffffffff81091b83>] perf_event_task_sched_out+0x93/0xa0
        [<ffffffff81732052>] __schedule+0x1d2/0xa20
        [<ffffffff81732f30>] preempt_schedule_irq+0x50/0xb0
        [<ffffffff817352b6>] retint_kernel+0x26/0x30
        [<ffffffff813eed04>] tty_flip_buffer_push+0x34/0x50
        [<ffffffff813f0504>] pty_write+0x54/0x60
        [<ffffffff813e900d>] n_tty_write+0x32d/0x4e0
        [<ffffffff813e5838>] tty_write+0x158/0x2d0
        [<ffffffff811c4850>] vfs_write+0xc0/0x1f0
        [<ffffffff811c52cc>] SyS_write+0x4c/0xa0
        [<ffffffff8173d4e4>] tracesys+0xdd/0xe2
 
-> #2 (&rq->lock){-.-.-.}:
        [<ffffffff810cc243>] lock_acquire+0x93/0x200
        [<ffffffff81733f90>] _raw_spin_lock+0x40/0x80
        [<ffffffff810980b2>] wake_up_new_task+0xc2/0x2e0
        [<ffffffff81054336>] do_fork+0x126/0x460
        [<ffffffff81054696>] kernel_thread+0x26/0x30
        [<ffffffff8171ff93>] rest_init+0x23/0x140
        [<ffffffff81ee1e4b>] start_kernel+0x3f6/0x403
        [<ffffffff81ee1571>] x86_64_start_reservations+0x2a/0x2c
        [<ffffffff81ee1664>] x86_64_start_kernel+0xf1/0xf4
 
-> #1 (&p->pi_lock){-.-.-.}:
        [<ffffffff810cc243>] lock_acquire+0x93/0x200
        [<ffffffff8173419b>] _raw_spin_lock_irqsave+0x4b/0x90
        [<ffffffff810979d1>] try_to_wake_up+0x31/0x350
        [<ffffffff81097d62>] default_wake_function+0x12/0x20
        [<ffffffff81084af8>] autoremove_wake_function+0x18/0x40
        [<ffffffff8108ea38>] __wake_up_common+0x58/0x90
        [<ffffffff8108ff59>] __wake_up+0x39/0x50
        [<ffffffff8110d4f8>] __call_rcu_nocb_enqueue+0xa8/0xc0
        [<ffffffff81111450>] __call_rcu+0x140/0x820
        [<ffffffff81111b8d>] call_rcu+0x1d/0x20
        [<ffffffff81093697>] cpu_attach_domain+0x287/0x360
        [<ffffffff81099d7e>] build_sched_domains+0xe5e/0x10a0
        [<ffffffff81efa7fc>] sched_init_smp+0x3b7/0x47a
        [<ffffffff81ee1f4e>] kernel_init_freeable+0xf6/0x202
        [<ffffffff817200be>] kernel_init+0xe/0x190
        [<ffffffff8173d22c>] ret_from_fork+0x7c/0xb0
 
-> #0 (&rdp->nocb_wq){......}:
        [<ffffffff810cb7ca>] __lock_acquire+0x191a/0x1be0
        [<ffffffff810cc243>] lock_acquire+0x93/0x200
        [<ffffffff8173419b>] _raw_spin_lock_irqsave+0x4b/0x90
        [<ffffffff8108ff43>] __wake_up+0x23/0x50
        [<ffffffff8110d4f8>] __call_rcu_nocb_enqueue+0xa8/0xc0
        [<ffffffff81111450>] __call_rcu+0x140/0x820
        [<ffffffff81111bb0>] kfree_call_rcu+0x20/0x30
        [<ffffffff81149abf>] put_ctx+0x4f/0x70
        [<ffffffff81154c3e>] perf_event_exit_task+0x12e/0x230
        [<ffffffff81056b8d>] do_exit+0x30d/0xcc0
        [<ffffffff8105893c>] do_group_exit+0x4c/0xc0
        [<ffffffff810589c4>] SyS_exit_group+0x14/0x20
        [<ffffffff8173d4e4>] tracesys+0xdd/0xe2
 
other info that might help us debug this:

Chain exists of:
  &rdp->nocb_wq --> &rq->lock --> &ctx->lock

  Possible unsafe locking scenario:

        CPU0                    CPU1
        ----                    ----
   lock(&ctx->lock);
                                lock(&rq->lock);
                                lock(&ctx->lock);
   lock(&rdp->nocb_wq);
 
 *** DEADLOCK ***

1 lock held by trinity-child2/15191:
 #0:  (&ctx->lock){-.-...}, at: [<ffffffff81154c19>] perf_event_exit_task+0x109/0x230

stack backtrace:
CPU: 2 PID: 15191 Comm: trinity-child2 Not tainted 3.12.0-rc3+ #92 
 ffffffff82565b70 ffff880070c2dbf8 ffffffff8172a363 ffffffff824edf40
 ffff880070c2dc38 ffffffff81726741 ffff880070c2dc90 ffff88022383b1c0
 ffff88022383aac0 0000000000000000 ffff88022383b188 ffff88022383b1c0
Call Trace:
 [<ffffffff8172a363>] dump_stack+0x4e/0x82
 [<ffffffff81726741>] print_circular_bug+0x200/0x20f
 [<ffffffff810cb7ca>] __lock_acquire+0x191a/0x1be0
 [<ffffffff810c6439>] ? get_lock_stats+0x19/0x60
 [<ffffffff8100b2f4>] ? native_sched_clock+0x24/0x80
 [<ffffffff810cc243>] lock_acquire+0x93/0x200
 [<ffffffff8108ff43>] ? __wake_up+0x23/0x50
 [<ffffffff8173419b>] _raw_spin_lock_irqsave+0x4b/0x90
 [<ffffffff8108ff43>] ? __wake_up+0x23/0x50
 [<ffffffff8108ff43>] __wake_up+0x23/0x50
 [<ffffffff8110d4f8>] __call_rcu_nocb_enqueue+0xa8/0xc0
 [<ffffffff81111450>] __call_rcu+0x140/0x820
 [<ffffffff8109bc8f>] ? local_clock+0x3f/0x50
 [<ffffffff81111bb0>] kfree_call_rcu+0x20/0x30
 [<ffffffff81149abf>] put_ctx+0x4f/0x70
 [<ffffffff81154c3e>] perf_event_exit_task+0x12e/0x230
 [<ffffffff81056b8d>] do_exit+0x30d/0xcc0
 [<ffffffff810c9af5>] ? trace_hardirqs_on_caller+0x115/0x1e0
 [<ffffffff810c9bcd>] ? trace_hardirqs_on+0xd/0x10
 [<ffffffff8105893c>] do_group_exit+0x4c/0xc0
 [<ffffffff810589c4>] SyS_exit_group+0x14/0x20
 [<ffffffff8173d4e4>] tracesys+0xdd/0xe2

--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ