[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-Id: <8fd077773ca78811c68c7a537cf21761da5046fc.1461323133.git.jslaby@suse.cz>
Date: Fri, 22 Apr 2016 13:05:26 +0200
From: Jiri Slaby <jslaby@...e.cz>
To: stable@...r.kernel.org
Cc: linux-kernel@...r.kernel.org, Xishi Qiu <qiuxishi@...wei.com>,
Joonsoo Kim <js1304@...il.com>,
David Rientjes <rientjes@...gle.com>,
"Laura Abbott" <lauraa@...eaurora.org>,
Hui Zhu <zhuhui@...omi.com>,
Wang Xiaoqiang <wangxq10@....edu.cn>,
Andrew Morton <akpm@...ux-foundation.org>,
Linus Torvalds <torvalds@...ux-foundation.org>,
Jiri Slaby <jslaby@...e.cz>
Subject: [PATCH 3.12 43/78] mm: fix invalid node in alloc_migrate_target()
From: Xishi Qiu <qiuxishi@...wei.com>
3.12-stable review patch. If anyone has any objections, please let me know.
===============
commit 6f25a14a7053b69917e2ebea0d31dd444cd31fd5 upstream.
It is incorrect to use next_node to find a target node, it will return
MAX_NUMNODES or invalid node. This will lead to crash in buddy system
allocation.
Fixes: c8721bbbdd36 ("mm: memory-hotplug: enable memory hotplug to handle hugepage")
Signed-off-by: Xishi Qiu <qiuxishi@...wei.com>
Acked-by: Vlastimil Babka <vbabka@...e.cz>
Acked-by: Naoya Horiguchi <n-horiguchi@...jp.nec.com>
Cc: Joonsoo Kim <js1304@...il.com>
Cc: David Rientjes <rientjes@...gle.com>
Cc: "Laura Abbott" <lauraa@...eaurora.org>
Cc: Hui Zhu <zhuhui@...omi.com>
Cc: Wang Xiaoqiang <wangxq10@....edu.cn>
Signed-off-by: Andrew Morton <akpm@...ux-foundation.org>
Signed-off-by: Linus Torvalds <torvalds@...ux-foundation.org>
Signed-off-by: Jiri Slaby <jslaby@...e.cz>
---
mm/page_isolation.c | 8 ++++----
1 file changed, 4 insertions(+), 4 deletions(-)
diff --git a/mm/page_isolation.c b/mm/page_isolation.c
index d1473b2e9481..88bdbf48cf6c 100644
--- a/mm/page_isolation.c
+++ b/mm/page_isolation.c
@@ -259,11 +259,11 @@ struct page *alloc_migrate_target(struct page *page, unsigned long private,
* now as a simple work-around, we use the next node for destination.
*/
if (PageHuge(page)) {
- nodemask_t src = nodemask_of_node(page_to_nid(page));
- nodemask_t dst;
- nodes_complement(dst, src);
+ int node = next_online_node(page_to_nid(page));
+ if (node == MAX_NUMNODES)
+ node = first_online_node;
return alloc_huge_page_node(page_hstate(compound_head(page)),
- next_node(page_to_nid(page), dst));
+ node);
}
if (PageHighMem(page))
--
2.8.1
Powered by blists - more mailing lists