lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Date:	Mon, 28 Jul 2014 19:08:21 -0400
From:	Sasha Levin <sasha.levin@...cle.com>
To:	Peter Zijlstra <peterz@...radead.org>
CC:	Ingo Molnar <mingo@...nel.org>,
	LKML <linux-kernel@...r.kernel.org>,
	Dave Jones <davej@...hat.com>
Subject: Re: sched: spinlock recursion in sched_rr_get_interval

On 07/07/2014 06:47 PM, Sasha Levin wrote:
> On 07/07/2014 04:05 PM, Peter Zijlstra wrote:
>> > On Mon, Jul 07, 2014 at 09:55:43AM -0400, Sasha Levin wrote:
>>> >> I've also had this one, which looks similar:
>>> >> 
>>> >> [10375.005884] BUG: spinlock recursion on CPU#0, modprobe/10965 [10375.006573]  lock: 0xffff8803a0fd7740, .magic: dead4ead, .owner: modprobe/10965, .owner_cpu: 15 [10375.007412] CPU: 0 PID: 10965 Comm: modprobe Tainted: G        W      3.16.0-rc3-next-20140704-sasha-00023-g26c0906-dirty #765
>> > 
>> > Something's fucked; so we have:
>> > 
>> > debug_spin_lock_before() SPIN_BUG_ON(lock->owner == current, "recursion");
>> > 
>> > Causing that, _HOWEVER_ look at .owner_cpu and the reporting cpu!! How can the lock owner, own the lock on cpu 15 and again contend with it on CPU 0. That's impossible.
>> > 
>> > About when-ish did you start seeing things like this? Lemme go stare hard at recent changes.
>> > 
> ~next-20140704 I guess, about when I reported the original issue.

Just wanted to add that this is still going on in -next:

[  860.050433] BUG: spinlock recursion on CPU#33, trinity-subchil/21438
[  860.051572]  lock: 0xffff8805fee10080, .magic: dead4ead, .owner: trinity-subchil/21438, .owner_cpu: -1
[  860.052943] CPU: 33 PID: 21438 Comm: trinity-subchil Not tainted 3.16.0-rc7-next-20140728-sasha-00029-ge067ff9 #976
[  860.053998]  ffff8805fee10080 ffff8805fe72bab0 ffffffffad464226 ffff8805ba163000
[  860.054820]  ffff8805fe72bad0 ffffffffaa1d7e76 ffff8805fee10080 ffffffffae88d599
[  860.055641]  ffff8805fe72baf0 ffffffffaa1d7ef6 ffff8805fee10080 ffff8805fee10080
[  860.056485] Call Trace:
[  860.056818]  [<ffffffffad464226>] dump_stack+0x4e/0x7a
[  860.057788]  [<ffffffffaa1d7e76>] spin_dump+0x86/0xe0
[  860.058620]  [<ffffffffaa1d7ef6>] spin_bug+0x26/0x30
[  860.059487]  [<ffffffffaa1d80bf>] do_raw_spin_lock+0x14f/0x1b0
[  860.060318]  [<ffffffffad4a9f01>] _raw_spin_lock+0x61/0x80
[  860.060318]  [<ffffffffaa1b4832>] ? load_balance+0x3a2/0xa50
[  860.060318]  [<ffffffffaa1b4832>] load_balance+0x3a2/0xa50
[  860.060318]  [<ffffffffaa1b541f>] pick_next_task_fair+0x53f/0xb00
[  860.060318]  [<ffffffffaa1b5300>] ? pick_next_task_fair+0x420/0xb00
[  860.060318]  [<ffffffffad4a89ab>] __schedule+0x16b/0x8c0
[  860.060318]  [<ffffffffaa2dbf18>] ? unlink_file_vma+0x38/0x60
[  860.060318]  [<ffffffffaa1a1903>] schedule_preempt_disabled+0x33/0x80
[  860.060318]  [<ffffffffaa1c633e>] mutex_lock_nested+0x1ae/0x620
[  860.060318]  [<ffffffffaa2dbf18>] ? unlink_file_vma+0x38/0x60
[  860.060318]  [<ffffffffaa2dbf18>] unlink_file_vma+0x38/0x60
[  860.060318]  [<ffffffffaa2d2b70>] free_pgtables+0xb0/0x130
[  860.060318]  [<ffffffffaa2df0d4>] exit_mmap+0xc4/0x180
[  860.060318]  [<ffffffffaa168aa3>] mmput+0x73/0x110
[  860.060318]  [<ffffffffaa16fa9a>] do_exit+0x2ca/0xc80
[  860.060318]  [<ffffffffaa1cecdb>] ? trace_hardirqs_on_caller+0xfb/0x280
[  860.060318]  [<ffffffffaa1cee6d>] ? trace_hardirqs_on+0xd/0x10
[  860.060318]  [<ffffffffaa1704de>] do_group_exit+0x4e/0xe0
[  860.060318]  [<ffffffffaa170584>] SyS_exit_group+0x14/0x20
[  860.060318]  [<ffffffffad4ab593>] tracesys+0xe1/0xe6


Thanks,
Sasha
--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ