[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <20090610081132.GA27519@localhost>
Date: Wed, 10 Jun 2009 16:11:32 +0800
From: Wu Fengguang <fengguang.wu@...el.com>
To: Johannes Weiner <hannes@...xchg.org>
Cc: Andrew Morton <akpm@...ux-foundation.org>,
Rik van Riel <riel@...hat.com>,
Hugh Dickins <hugh.dickins@...cali.co.uk>,
Andi Kleen <andi@...stfloor.org>,
KAMEZAWA Hiroyuki <kamezawa.hiroyu@...fujitsu.com>,
Minchan Kim <minchan.kim@...il.com>,
"linux-mm@...ck.org" <linux-mm@...ck.org>,
"linux-kernel@...r.kernel.org" <linux-kernel@...r.kernel.org>
Subject: Re: [patch v3] swap: virtual swap readahead
On Wed, Jun 10, 2009 at 03:45:08PM +0800, Johannes Weiner wrote:
> Hi Fengguang,
>
> On Wed, Jun 10, 2009 at 01:03:42PM +0800, Wu Fengguang wrote:
> > On Wed, Jun 10, 2009 at 03:37:02AM +0800, Johannes Weiner wrote:
> > > On Tue, Jun 09, 2009 at 09:01:28PM +0200, Johannes Weiner wrote:
> > > > [resend with lists cc'd, sorry]
> > >
> > > [and fixed Hugh's email. crap]
> > >
> > > > Hi,
> > > >
> > > > here is a new iteration of the virtual swap readahead. Per Hugh's
> > > > suggestion, I moved the pte collecting to the callsite and thus out
> > > > ouf swap code. Unfortunately, I had to bound page_cluster due to an
> > > > array of that many swap entries on the stack, but I think it is better
> > > > to limit the cluster size to a sane maximum than using dynamic
> > > > allocation for this purpose.
> >
> > Hi Johannes,
> >
> > When stress testing your patch, I found it triggered many OOM kills.
> > Around the time of last OOMs, the memory usage is:
> >
> > total used free shared buffers cached
> > Mem: 474 468 5 0 0 239
> > -/+ buffers/cache: 229 244
> > Swap: 1023 221 802
>
> Wow, that really confused me for a second as we shouldn't read more
> pages ahead than without the patch, probably even less under stress.
Yup - swap readahead is much more challenging than sequential readahead,
in that it must be accurate enough given some really obscure patterns.
> So the problem has to be a runaway reading. And indeed, severe
> stupidity here:
>
> + window = cluster << PAGE_SHIFT;
> + min = addr & ~(window - 1);
> + max = min + cluster;
> + /*
> + * To keep the locking/highpte mapping simple, stay
> + * within the PTE range of one PMD entry.
> + */
> + limit = addr & PMD_MASK;
> + if (limit > min)
> + min = limit;
> + limit = pmd_addr_end(addr, max);
> + if (limit < max)
> + max = limit;
> + limit = max - min;
>
> The mistake is at the initial calculation of max. It should be
>
> max = min + window;
>
> The resulting problem is that min could get bigger than max when
> cluster is bigger than PMD_SHIFT. Did you use page_cluster == 5?
No I use the default 3.
btw, the mistake reflects bad named variables. How about rename
cluster => pages
window => bytes
?
> The initial min is aligned to a value below the PMD boundary and max
> based on it with a too small offset, staying below the PMD boundary as
> well. When min is rounded up, this becomes a bit large:
>
> limit = max - min;
>
> So if my brain is already functioning, fixing the initial max should
> be enough because either
>
> o window is smaller than PMD_SIZE, than we won't round down
> below a PMD boundary in the first place or
>
> o window is bigger than PMD_SIZE, than we can round down below
> a PMD boundary but adding window to that is garuanteed to
> cross the boundary again
>
> and thus max is always bigger than min.
>
> Fengguang, does this make sense? If so, the patch below should fix
> it.
Too bad, a quick test of the below patch freezes the box..
Thanks,
Fengguang
> --- a/mm/memory.c
> +++ b/mm/memory.c
> @@ -2467,7 +2467,7 @@ static int swap_readahead_ptes(struct mm
>
> window = cluster << PAGE_SHIFT;
> min = addr & ~(window - 1);
> - max = min + cluster;
> + max = min + window;
> /*
> * To keep the locking/highpte mapping simple, stay
> * within the PTE range of one PMD entry.
--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/
Powered by blists - more mailing lists