[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <ab3f9b940804301907y5a3e84e1l6cb41a339bc2241b@mail.gmail.com>
Date: Wed, 30 Apr 2008 19:07:58 -0700
From: "Tom May" <tom@...may.com>
To: "Daniel Spång" <daniel.spang@...il.com>
Cc: "KOSAKI Motohiro" <kosaki.motohiro@...fujitsu.com>,
linux-mm@...ck.org, linux-kernel@...r.kernel.org
Subject: Re: [PATCH 0/8][for -mm] mem_notify v6
On Wed, Apr 23, 2008 at 1:27 AM, Daniel Spång <daniel.spang@...il.com> wrote:
> Hi Tom
>
>
> On 4/17/08, Tom May <tom@...may.com> wrote:
> >
> > Here is the start and end of the output from the test program. At
> > each /dev/mem_notify notification Cached decreases, then eventually
> > Mapped decreases as well, which means the amount of time the program
> > has to free memory gets smaller and smaller. Finally the oom killer
> > is invoked because the program can't react quickly enough to free
> > memory, even though it can free at a faster rate than it can use
> > memory. My test is slow to free because it calls nanosleep, but this
> > is just a simulation of my actual program that has to perform garbage
> > collection before it can free memory.
>
> I have also seen this behaviour in my static tests with low mem
> notification on swapless systems. It is a problem with small programs
> (typically static test programs) where the text segment is only a few
> pages. I have not seen this behaviour in larger programs which use a
> larger working set. As long as the system working set is bigger than
> the amount of memory that needs to be allocated, between every
> notification reaction opportunity, it seems to be ok.
Hi Daniel,
You're saying the program's in-core text pages serve as a reserve that
the kernel can discard when it needs some memory, correct? And that
even if the kernel discards them, it will page them back in as a
matter of course as the program runs, to maintain the reserve? That
certainly makes sense.
In my case of a Java virtual machine, where I originally saw the
problem, most of the code is interpreted byte codes or jit-compiled
native code, all of which resides not in the text segment but in
anonymous pages that aren't backed by a file, and there is no swap
space. The actual text segment working set can be very small (memory
allocation, garbage collection, synchronization, other random native
code). And, as KOSAKI Motohiro pointed out, it may be wise to mlock
these areas. So the text working set doesn't make an adequate
reserve.
However, I can maintain a reserve of cached and/or mapped memory by
touching pages in the text segment (or any mapped file) as the final
step of low memory notification handling, if the cached page count is
getting low. For my purposes, this is nearly the same as having an
additional threshold-based notification, since it forces notifications
to occur while the kernel still has some memory to satisfy allocations
while userspace code works to free memory. And it's simple.
Unfortunately, this is more expensive than it could be since the pages
need to be read in from some device (mapping /dev/zero doesn't cause
pages to be allocated). What I'm looking for now is a cheap way to
populate the cache with pages that the kernel can throw away when it
needs to reclaim memory.
Thanks,
.tom
--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/
Powered by blists - more mailing lists