lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite for Android: free password hash cracker in your pocket
[<prev] [next>] [<thread-prev] [day] [month] [year] [list]
Message-ID: <20240220183220.1451315-5-zi.yan@sent.com>
Date: Tue, 20 Feb 2024 13:32:20 -0500
From: Zi Yan <zi.yan@...t.com>
To: linux-mm@...ck.org,
	linux-kernel@...r.kernel.org
Cc: Zi Yan <ziy@...dia.com>,
	"Huang, Ying" <ying.huang@...el.com>,
	Ryan Roberts <ryan.roberts@....com>,
	Andrew Morton <akpm@...ux-foundation.org>,
	"Matthew Wilcox (Oracle)" <willy@...radead.org>,
	David Hildenbrand <david@...hat.com>,
	"Yin, Fengwei" <fengwei.yin@...el.com>,
	Yu Zhao <yuzhao@...gle.com>,
	Vlastimil Babka <vbabka@...e.cz>,
	"Kirill A . Shutemov" <kirill.shutemov@...ux.intel.com>,
	Johannes Weiner <hannes@...xchg.org>,
	Baolin Wang <baolin.wang@...ux.alibaba.com>,
	Kemeng Shi <shikemeng@...weicloud.com>,
	Mel Gorman <mgorman@...hsingularity.net>,
	Rohan Puri <rohan.puri15@...il.com>,
	Mcgrof Chamberlain <mcgrof@...nel.org>,
	Adam Manzanares <a.manzanares@...sung.com>,
	"Vishal Moola (Oracle)" <vishal.moola@...il.com>
Subject: [PATCH v7 4/4] mm/compaction: optimize >0 order folio compaction with free page split.

From: Zi Yan <ziy@...dia.com>

During migration in a memory compaction, free pages are placed in an array
of page lists based on their order.  But the desired free page order
(i.e., the order of a source page) might not be always present, thus
leading to migration failures and premature compaction termination.  Split
a high order free pages when source migration page has a lower order to
increase migration successful rate.

Note: merging free pages when a migration fails and a lower order free
page is returned via compaction_free() is possible, but there is too much
work.  Since the free pages are not buddy pages, it is hard to identify
these free pages using existing PFN-based page merging algorithm.

Signed-off-by: Zi Yan <ziy@...dia.com>
Reviewed-by: Baolin Wang <baolin.wang@...ux.alibaba.com>
Reviewed-by: Vlastimil Babka <vbabka@...e.cz>
Tested-by: Baolin Wang <baolin.wang@...ux.alibaba.com>
Tested-by: Yu Zhao <yuzhao@...gle.com>
Cc: Adam Manzanares <a.manzanares@...sung.com>
Cc: David Hildenbrand <david@...hat.com>
Cc: Huang Ying <ying.huang@...el.com>
Cc: Johannes Weiner <hannes@...xchg.org>
Cc: Kemeng Shi <shikemeng@...weicloud.com>
Cc: Kirill A. Shutemov <kirill.shutemov@...ux.intel.com>
Cc: Luis Chamberlain <mcgrof@...nel.org>
Cc: Matthew Wilcox (Oracle) <willy@...radead.org>
Cc: Mel Gorman <mgorman@...hsingularity.net>
Cc: Ryan Roberts <ryan.roberts@....com>
Cc: Vishal Moola (Oracle) <vishal.moola@...il.com>
Cc: Vlastimil Babka <vbabka@...e.cz>
Cc: Yin Fengwei <fengwei.yin@...el.com>
---
 mm/compaction.c | 35 ++++++++++++++++++++++++++++++-----
 1 file changed, 30 insertions(+), 5 deletions(-)

diff --git a/mm/compaction.c b/mm/compaction.c
index 112711752321..e70309674262 100644
--- a/mm/compaction.c
+++ b/mm/compaction.c
@@ -1856,15 +1856,40 @@ static struct folio *compaction_alloc(struct folio *src, unsigned long data)
 	struct compact_control *cc = (struct compact_control *)data;
 	struct folio *dst;
 	int order = folio_order(src);
+	bool has_isolated_pages = false;
+	int start_order;
+	struct page *freepage;
+	unsigned long size;
+
+again:
+	for (start_order = order; start_order < NR_PAGE_ORDERS; start_order++)
+		if (!list_empty(&cc->freepages[start_order]))
+			break;
 
-	if (list_empty(&cc->freepages[order])) {
-		isolate_freepages(cc);
-		if (list_empty(&cc->freepages[order]))
+	/* no free pages in the list */
+	if (start_order == NR_PAGE_ORDERS) {
+		if (has_isolated_pages)
 			return NULL;
+		isolate_freepages(cc);
+		has_isolated_pages = true;
+		goto again;
+	}
+
+	freepage = list_first_entry(&cc->freepages[start_order], struct page,
+				lru);
+	size = 1 << start_order;
+
+	list_del(&freepage->lru);
+
+	while (start_order > order) {
+		start_order--;
+		size >>= 1;
+
+		list_add(&freepage[size].lru, &cc->freepages[start_order]);
+		set_page_private(&freepage[size], start_order);
 	}
+	dst = (struct folio *)freepage;
 
-	dst = list_first_entry(&cc->freepages[order], struct folio, lru);
-	list_del(&dst->lru);
 	post_alloc_hook(&dst->page, order, __GFP_MOVABLE);
 	if (order)
 		prep_compound_page(&dst->page, order);
-- 
2.43.0


Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ