[<prev] [next>] [thread-next>] [day] [month] [year] [list]
Message-ID: <20250818185804.21044-1-joshua.hahnjy@gmail.com>
Date: Mon, 18 Aug 2025 11:58:03 -0700
From: Joshua Hahn <joshua.hahnjy@...il.com>
To: Johannes Weiner <hannes@...xchg.org>,
Chris Mason <clm@...com>
Cc: Andrew Morton <akpm@...ux-foundation.org>,
Vlastimil Babka <vbabka@...e.cz>,
Suren Baghdasaryan <surenb@...le.com>,
Michal Hocko <mhocko@...e.com>,
Brendan Jackman <jackmanb@...gle.com>,
Zi Yan <ziy@...dia.com>,
linux-mm@...ck.org,
linux-kernel@...r.kernel.org,
kernel-team@...a.com
Subject: [PATCH] mm/page_alloc: Occasionally relinquish zone lock in batch freeing
While testing workloads with high sustained memory pressure on large machines
(1TB memory, 316 CPUs), we saw an unexpectedly high number of softlockups.
Further investigation showed that the lock in free_pcppages_bulk was being held
for a long time, even being held while 2k+ pages were being freed.
Instead of holding the lock for the entirety of the freeing, check to see if
the zone lock is contended every pcp->batch pages. If there is contention,
relinquish the lock so that other processors have a change to grab the lock
and perform critical work.
In our fleet, we have seen that performing batched lock freeing has led to
significantly lower rates of softlockups, while incurring relatively small
regressions (relative to the workload and relative to the variation).
The following are a few synthetic benchmarks:
Test 1: Small machine (30G RAM, 36 CPUs)
stress-ng --vm 30 --vm-bytes 1G -M -t 100
+----------------------+---------------+-----------+
| Metric | Variation (%) | Delta (%) |
+----------------------+---------------+-----------+
| bogo ops | 0.0076 | -0.0183 |
| bogo ops/s (real) | 0.0064 | -0.0207 |
| bogo ops/s (usr+sys) | 0.3151 | +0.4141 |
+----------------------+---------------+-----------+
stress-ng --vm 20 --vm-bytes 3G -M -t 100
+----------------------+---------------+-----------+
| Metric | Variation (%) | Delta (%) |
+----------------------+---------------+-----------+
| bogo ops | 0.0295 | -0.0078 |
| bogo ops/s (real) | 0.0267 | -0.0177 |
| bogo ops/s (usr+sys) | 1.7079 | -0.0096 |
+----------------------+---------------+-----------+
Test 2: Big machine (250G RAM, 176 CPUs)
stress-ng --vm 50 --vm-bytes 5G -M -t 100
+----------------------+---------------+-----------+
| Metric | Variation (%) | Delta (%) |
+----------------------+---------------+-----------+
| bogo ops | 0.0362 | -0.0187 |
| bogo ops/s (real) | 0.0391 | -0.0220 |
| bogo ops/s (usr+sys) | 2.9603 | +1.3758 |
+----------------------+---------------+-----------+
stress-ng --vm 10 --vm-bytes 30G -M -t 100
+----------------------+---------------+-----------+
| Metric | Variation (%) | Delta (%) |
+----------------------+---------------+-----------+
| bogo ops | 2.3130 | -0.0754 |
| bogo ops/s (real) | 3.3069 | -0.8579 |
| bogo ops/s (usr+sys) | 4.0369 | -1.1985 |
+----------------------+---------------+-----------+
Suggested-by: Chris Mason <clm@...com>
Co-developed-by: Johannes Weiner <hannes@...xchg.org>
Signed-off-by: Joshua Hahn <joshua.hahnjy@...il.com>
---
mm/page_alloc.c | 15 ++++++++++++++-
1 file changed, 14 insertions(+), 1 deletion(-)
diff --git a/mm/page_alloc.c b/mm/page_alloc.c
index a8a84c3b5fe5..bd7a8da3e159 100644
--- a/mm/page_alloc.c
+++ b/mm/page_alloc.c
@@ -1238,6 +1238,8 @@ static void free_pcppages_bulk(struct zone *zone, int count,
* below while (list_empty(list)) loop.
*/
count = min(pcp->count, count);
+ if (!count)
+ return;
/* Ensure requested pindex is drained first. */
pindex = pindex - 1;
@@ -1247,6 +1249,7 @@ static void free_pcppages_bulk(struct zone *zone, int count,
while (count > 0) {
struct list_head *list;
int nr_pages;
+ int batch = min(count, pcp->batch);
/* Remove pages from lists in a round-robin fashion. */
do {
@@ -1267,12 +1270,22 @@ static void free_pcppages_bulk(struct zone *zone, int count,
/* must delete to avoid corrupting pcp list */
list_del(&page->pcp_list);
+ batch -= nr_pages;
count -= nr_pages;
pcp->count -= nr_pages;
__free_one_page(page, pfn, zone, order, mt, FPI_NONE);
trace_mm_page_pcpu_drain(page, order, mt);
- } while (count > 0 && !list_empty(list));
+ } while (batch > 0 && !list_empty(list));
+
+ /*
+ * Prevent starving the lock for other users; every pcp->batch
+ * pages freed, relinquish the zone lock if it is contended.
+ */
+ if (count && spin_is_contended(&zone->lock)) {
+ spin_unlock_irqrestore(&zone->lock, flags);
+ spin_lock_irqsave(&zone->lock, flags);
+ }
}
spin_unlock_irqrestore(&zone->lock, flags);
base-commit: 137a6423b60fe0785aada403679d3b086bb83062
--
2.47.3
Powered by blists - more mailing lists