[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <20081231110619.GA20534@csn.ul.ie>
Date: Wed, 31 Dec 2008 11:06:19 +0000
From: Mel Gorman <mel@....ul.ie>
To: Nick Piggin <npiggin@...e.de>
Cc: KOSAKI Motohiro <kosaki.motohiro@...fujitsu.com>,
LKML <linux-kernel@...r.kernel.org>,
linux-mm <linux-mm@...ck.org>,
Andrew Morton <akpm@...ux-foundation.org>,
wassim dagash <wassim.dagash@...il.com>
Subject: Re: [PATCH] mm: stop kswapd's infinite loop at high order
allocation
On Wed, Dec 31, 2008 at 02:32:33AM +0100, Nick Piggin wrote:
> On Tue, Dec 30, 2008 at 06:59:19PM +0000, Mel Gorman wrote:
> > On Tue, Dec 30, 2008 at 07:55:47PM +0900, KOSAKI Motohiro wrote:
> > >
> > > ok, wassim confirmed this patch works well.
> > >
> > >
> > > ==
> > > From: KOSAKI Motohiro <kosaki.motohiro@...fujitsu.com>
> > > Subject: [PATCH] mm: kswapd stop infinite loop at high order allocation
> > >
> > > Wassim Dagash reported following kswapd infinite loop problem.
> > >
> > > kswapd runs in some infinite loop trying to swap until order 10 of zone
> > > highmem is OK, While zone higmem (as I understand) has nothing to do
> > > with contiguous memory (cause there is no 1-1 mapping) which means
> > > kswapd will continue to try to balance order 10 of zone highmem
> > > forever (or until someone release a very large chunk of highmem).
> > >
> > > He proposed remove contenious checking on highmem at all.
> > > However hugepage on highmem need contenious highmem page.
> > >
> >
> > I'm lacking the original problem report, but contiguous order-10 pages are
> > indeed required for hugepages in highmem and reclaiming for them should not
> > be totally disabled at any point. While no 1-1 mapping exists for the kernel,
> > contiguity is still required.
>
> This doesn't totally disable them. It disables asynchronous reclaim for them
> until the next time kswapd kicks is kicked by a higher order allocator. The
> guy who kicked us off this time should go into direct reclaim.
>
I get that. The check to bail out is made after we've already done the
scanning. I wanted to be clear that contiguity in highmem is required for
hugepages.
>
> > kswapd gets a sc.order when it is known there is a process trying to get
> > high-order pages so it can reclaim at that order in an attempt to prevent
> > future direct reclaim at a high-order. Your patch does not appear to depend on
> > GFP_KERNEL at all so I found the comment misleading. Furthermore, asking it to
> > loop again at order-0 means it may scan and reclaim more memory unnecessarily
> > seeing as all_zones_ok was calculated based on a high-order value, not order-0.
>
> It shouldn't, because it should check all that.
>
Ok, with KOSAKI's patch we
1. Set order to 0 (and stop kswapd doing what it was asked to do)
2. goto loop_again
3. nr_reclaimed gets set to 0 (meaning we lose that value, but no biggie
as it doesn't get used by the caller anyway)
4. Reset all priorities
5. Do something like the following
for (priority = DEF_PRIORITY; priority >= 0; priority--) {
...
all_zones_ok = 1;
for (i = pgdat->nr_zones - 1; i >= 0; i--) {
...
if (inactive_anon_is_low(zone)) {
shrink_active_list(SWAP_CLUSTER_MAX, zone,
&sc, priority, 0);
}
if (!zone_watermark_ok(zone, order, zone->pages_high,
0, 0)) {
end_zone = i;
break;
}
}
}
So, by looping around, we could end up shrinking the active list again
before we recheck the zone watermarks depending on the size of the
inactive lists.
If the size of the lists is ok, I agree that we'll go through the lists,
do no reclaiming and exit out, albeit returning 0 when we have reclaimed pages.
>
> > While constantly looping trying to balance for high-orders is indeed bad,
> > I'm unconvinced this is the correct change. As we have already gone through
> > a priorities and scanned everything at the high-order, would it not make
> > more sense to do just give up with something like the following?
> >
> > /*
> > * If zones are still not balanced, loop again and continue attempting
> > * to rebalance the system. For high-order allocations, fragmentation
> > * can prevent the zones being rebalanced no matter how hard kswapd
> > * works, particularly on systems with little or no swap. For costly
> > * orders, just give up and assume interested processes will either
> > * direct reclaim or wake up kswapd as necessary.
> > */
> > if (!all_zones_ok && sc.order <= PAGE_ALLOC_COSTLY_ORDER) {
> > cond_resched();
> >
> > try_to_freeze();
> >
> > goto loop_again;
> > }
> >
> > I used PAGE_ALLOC_COSTLY_ORDER instead of sc.order == 0 because we are
> > expected to support allocations up to that order in a fairly reliable fashion.
>
> I actually think it's better to do it for all orders, because that
> constant is more or less arbitrary.
i.e.
if (!all_zones_ok && sc.order == 0) {
? or something else
What I did miss was that we have
if (nr_reclaimed >= SWAP_CLUSTER_MAX)
break;
so with my patch, kswapd is bailing out early without trying to reclaim for
high-orders that hard. That was not what I intended as it means we only ever
really rebalance the full system for order-0 pages and for everything else we
do relatively light scanning. The impact is that high-order users will direct
reclaim rather than depending on kswapd scanning very heavily. Arguably,
this is a good thing.
However, it also means that KOSAKI's and my patches only differs in that mine
bails early and KOSAKI rechecks everything at order-0, possibly reclaiming
more. If the comment was not so misleading, I'd have been a lot happier.
> It is possible a zone might become
> too fragmented to support this, but the allocating process has been OOMed or
> had their allocation satisfied from another zone. kswapd would have no way
> out of the loop even if the system no longer requires higher order allocations.
>
> IOW, I don't see a big downside, and there is a real upside.
>
> I think the patch is good.
>
Which one, KOSAKI's or my one?
Here is my one again which bails out for any high-order allocation after
just light scanning.
====
>From 0e09fe002d8e9956de227b880ef8458842b71ca9 Mon Sep 17 00:00:00 2001
From: Mel Gorman <mel@....ul.ie>
Date: Tue, 30 Dec 2008 18:53:23 +0000
Subject: [PATCH] mm: stop kswapd's infinite loop at high order allocation
Wassim Dagash reported the following (editted) kswapd infinite loop problem.
kswapd runs in some infinite loop trying to swap until order 10 of zone
highmem is OK.... kswapd will continue to try to balance order 10 of zone
highmem forever (or until someone release a very large chunk of highmem).
For costly high-order allocations, the system may never be balanced due to
fragmentation but kswapd should not infinitely loop as a result. The
following patch lets kswapd stop reclaiming in the event it cannot
balance zones and the order is high-order.
Reported-by: wassim dagash <wassim.dagash@...il.com>
Signed-off-by: Mel Gorman <mel@....ul.ie>
---
mm/vmscan.c | 11 ++++++++++-
1 files changed, 10 insertions(+), 1 deletions(-)
diff --git a/mm/vmscan.c b/mm/vmscan.c
index 62e7f62..7b0f412 100644
--- a/mm/vmscan.c
+++ b/mm/vmscan.c
@@ -1867,7 +1867,16 @@ out:
zone->prev_priority = temp_priority[i];
}
- if (!all_zones_ok) {
+
+ /*
+ * If zones are still not balanced, loop again and continue attempting
+ * to rebalance the system. For high-order allocations, fragmentation
+ * can prevent the zones being rebalanced no matter how hard kswapd
+ * works, particularly on systems with little or no swap. For
+ * high-orders, just give up and assume interested processes will
+ * either direct reclaim or wake up kswapd again as necessary.
+ */
+ if (!all_zones_ok && sc.order == 0) {
cond_resched();
try_to_freeze();
--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/
Powered by blists - more mailing lists