lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <1394566262.2786.23.camel@buesod1.americas.hpqcorp.net>
Date:	Tue, 11 Mar 2014 12:31:02 -0700
From:	Davidlohr Bueso <davidlohr@...com>
To:	Mel Gorman <mgorman@...e.de>
Cc:	Sasha Levin <sasha.levin@...cle.com>,
	Rik van Riel <riel@...hat.com>,
	David Rientjes <rientjes@...gle.com>,
	Andrew Morton <akpm@...ux-foundation.org>,
	"linux-mm@...ck.org" <linux-mm@...ck.org>,
	LKML <linux-kernel@...r.kernel.org>, hhuang@...hat.com,
	knoel@...hat.com, aarcange@...hat.com
Subject: Re: [PATCH -mm] mm,numa,mprotect: always continue after finding a
 stable thp page

On Tue, 2014-03-11 at 18:06 +0000, Mel Gorman wrote:
> On Tue, Mar 11, 2014 at 01:33:00PM -0400, Sasha Levin wrote:
> > Okay. So just this patch on top of the latest -next shows the following issues:
> > 
> > 1. BUG in task_numa_work:
> > 
> > [  439.417171] BUG: unable to handle kernel paging request at ffff880e17530c00
> > [  439.418216] IP: [<ffffffff81299385>] vmacache_find+0x75/0xa0
> > [  439.419073] PGD 8904067 PUD 1028fcb067 PMD 1028f10067 PTE 8000000e17530060
> > [  439.420340] Oops: 0000 [#1] PREEMPT SMP DEBUG_PAGEALLOC
> > [  439.420340] Dumping ftrace buffer:
> > [  439.420340]    (ftrace buffer empty)
> > [  439.420340] Modules linked in:
> > [  439.420340] CPU: 12 PID: 9937 Comm: trinity-c212 Tainted: G        W    3.14.0-rc5-next-20140307-sasha-00009-g3b24300-dirty #137
> > [  439.420340] task: ffff880e1a45b000 ti: ffff880e1a490000 task.ti: ffff880e1a490000
> > [  439.420340] RIP: 0010:[<ffffffff81299385>]  [<ffffffff81299385>] vmacache_find+0x75/0xa0
> > [  439.420340] RSP: 0018:ffff880e1a491e68  EFLAGS: 00010286
> > [  439.420340] RAX: ffff880e17530c00 RBX: 0000000000000000 RCX: 0000000000000001
> > [  439.420340] RDX: 0000000000000001 RSI: 0000000000000000 RDI: ffff880e1a45b000
> > [  439.420340] RBP: ffff880e1a491e68 R08: 0000000000000000 R09: 0000000000000000
> > [  439.420340] R10: 0000000000000001 R11: 0000000000000000 R12: ffff880e1ab75000
> > [  439.420340] R13: ffff880e1ab75000 R14: 0000000000010000 R15: 0000000000000000
> > [  439.420340] FS:  00007f3458c05700(0000) GS:ffff880d2b800000(0000) knlGS:0000000000000000
> > [  439.420340] CS:  0010 DS: 0000 ES: 0000 CR0: 000000008005003b
> > [  439.420340] CR2: ffff880e17530c00 CR3: 0000000e1a472000 CR4: 00000000000006a0
> > [  439.420340] Stack:
> > [  439.420340]  ffff880e1a491e98 ffffffff812a7610 ffffffff8118de40 0000000000000117
> > [  439.420340]  00000001000036da ffff880e1a45b000 ffff880e1a491ef8 ffffffff8118de4b
> > [  439.420340]  ffff880e1a491ec8 ffffffff81269575 ffff880e1ab750a8 ffff880e1a45b000
> > [  439.420340] Call Trace:
> > [  439.420340]  [<ffffffff812a7610>] find_vma+0x20/0x90
> > [  439.420340]  [<ffffffff8118de40>] ? task_numa_work+0x130/0x360
> > [  439.420340]  [<ffffffff8118de4b>] task_numa_work+0x13b/0x360
> > [  439.420340]  [<ffffffff81269575>] ? context_tracking_user_exit+0x195/0x1d0
> > [  439.420340]  [<ffffffff8116c5be>] task_work_run+0xae/0xf0
> > [  439.420340]  [<ffffffff8106ffbe>] do_notify_resume+0x8e/0xe0
> > [  439.420340]  [<ffffffff844b17a2>] int_signal+0x12/0x17
> > [  439.420340] Code: 42 10 00 00 00 00 48 c7 42 18 00 00 00 00 eb 38 66 0f 1f 44 00 00 31 d2 48 89 c7 48 63 ca 48 8b 84 cf b8 07 00 00 48 85 c0 74 0b <48> 39 30 77 06 48 3b 70 08 72 12 ff c2 83 fa 04 75 de 66 0f 1f
> > [  439.420340] RIP  [<ffffffff81299385>] vmacache_find+0x75/0xa0
> > [  439.420340]  RSP <ffff880e1a491e68>
> > [  439.420340] CR2: ffff880e17530c00
> > 
> 
> Ok, this does not look related. It looks like damage from the VMA caching
> patches, possibly a use-after free. I'm skeptical that it's related to
> automatic NUMA balancing as such based on the second trace you posted.

Indeed, this issue is separate. It was reported a few days ago by
Fengguang: http://lkml.org/lkml/2014/3/9/201

Apparently this has only been seen in DEBUG_PAGEALLOC + trinity
scenarios, just like in Sasha's trace. I'm suspecting some invalidation
is missing (not sure what DEBUG_PAGEALLOC enables differently in this
aspect), as when we do the vmacache_find() the vma is bogus which
triggers the paging request error.

--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ