[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-Id: <20160418022616.415248358@linuxfoundation.org>
Date: Mon, 18 Apr 2016 11:28:05 +0900
From: Greg Kroah-Hartman <gregkh@...uxfoundation.org>
To: linux-kernel@...r.kernel.org
Cc: Greg Kroah-Hartman <gregkh@...uxfoundation.org>,
stable@...r.kernel.org, Xishi Qiu <qiuxishi@...wei.com>,
Vlastimil Babka <vbabka@...e.cz>,
Naoya Horiguchi <n-horiguchi@...jp.nec.com>,
Joonsoo Kim <js1304@...il.com>,
David Rientjes <rientjes@...gle.com>,
"Laura Abbott" <lauraa@...eaurora.org>,
Hui Zhu <zhuhui@...omi.com>,
Wang Xiaoqiang <wangxq10@....edu.cn>,
Andrew Morton <akpm@...ux-foundation.org>,
Linus Torvalds <torvalds@...ux-foundation.org>
Subject: [PATCH 4.5 013/124] mm: fix invalid node in alloc_migrate_target()
4.5-stable review patch. If anyone has any objections, please let me know.
------------------
From: Xishi Qiu <qiuxishi@...wei.com>
commit 6f25a14a7053b69917e2ebea0d31dd444cd31fd5 upstream.
It is incorrect to use next_node to find a target node, it will return
MAX_NUMNODES or invalid node. This will lead to crash in buddy system
allocation.
Fixes: c8721bbbdd36 ("mm: memory-hotplug: enable memory hotplug to handle hugepage")
Signed-off-by: Xishi Qiu <qiuxishi@...wei.com>
Acked-by: Vlastimil Babka <vbabka@...e.cz>
Acked-by: Naoya Horiguchi <n-horiguchi@...jp.nec.com>
Cc: Joonsoo Kim <js1304@...il.com>
Cc: David Rientjes <rientjes@...gle.com>
Cc: "Laura Abbott" <lauraa@...eaurora.org>
Cc: Hui Zhu <zhuhui@...omi.com>
Cc: Wang Xiaoqiang <wangxq10@....edu.cn>
Signed-off-by: Andrew Morton <akpm@...ux-foundation.org>
Signed-off-by: Linus Torvalds <torvalds@...ux-foundation.org>
Signed-off-by: Greg Kroah-Hartman <gregkh@...uxfoundation.org>
---
mm/page_isolation.c | 8 ++++----
1 file changed, 4 insertions(+), 4 deletions(-)
--- a/mm/page_isolation.c
+++ b/mm/page_isolation.c
@@ -289,11 +289,11 @@ struct page *alloc_migrate_target(struct
* now as a simple work-around, we use the next node for destination.
*/
if (PageHuge(page)) {
- nodemask_t src = nodemask_of_node(page_to_nid(page));
- nodemask_t dst;
- nodes_complement(dst, src);
+ int node = next_online_node(page_to_nid(page));
+ if (node == MAX_NUMNODES)
+ node = first_online_node;
return alloc_huge_page_node(page_hstate(compound_head(page)),
- next_node(page_to_nid(page), dst));
+ node);
}
if (PageHighMem(page))
Powered by blists - more mailing lists