[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <AANLkTikT-svqverRLr7Mf6s-17VrOcP_BpyXFpDV=_7s@mail.gmail.com>
Date: Thu, 25 Nov 2010 16:33:01 +0100
From: Peter Schüller <scode@...tify.com>
To: Dave Hansen <dave@...ux.vnet.ibm.com>
Cc: Andrew Morton <akpm@...ux-foundation.org>,
linux-kernel@...r.kernel.org,
Mattias de Zalenski <zalenski@...tify.com>,
linux-mm@...ck.org
Subject: Re: Sudden and massive page cache eviction
> simple thing to do in any case. You can watch the entries in slabinfo
> and see if any of the ones with sizes over 4096 bytes are getting used
> often. You can also watch /proc/buddyinfo and see how often columns
> other than the first couple are moving around.
I collected some information from
/proc/{buddyinfo,meminfo,slabinfo,vmstat} and let it sit, polling
approximately once per minute. I have some results correlated with
another page eviction in graphs. The graph is here:
http://files.spotify.com/memcut/memgraph-20101124.png
The last sudden eviction there occurred somewhere between 22:30 and
22:45. Some URL:s that can be compared for those periods:
Before:
http://files.spotify.com/memcut/memstat-20101124/2010-11-24T22:39:30/vmstat
http://files.spotify.com/memcut/memstat-20101124/2010-11-24T22:39:30/buddyinfo
http://files.spotify.com/memcut/memstat-20101124/2010-11-24T22:39:30/meminfo
http://files.spotify.com/memcut/memstat-20101124/2010-11-24T22:39:30/slabinfo
After:
http://files.spotify.com/memcut/memstat-20101124/2010-11-24T22:45:31/vmstat
http://files.spotify.com/memcut/memstat-20101124/2010-11-24T22:45:31/buddyinfo
http://files.spotify.com/memcut/memstat-20101124/2010-11-24T22:45:31/meminfo
http://files.spotify.com/memcut/memstat-20101124/2010-11-24T22:45:31/slabinfo
A more complete set of files for several minutes before/after/during
is in this tarball:
http://files.spotify.com/memcut/memgraph-20101124.tar.gz
Diffing the two slabinfos above yields among other things the
following expected or at least plausible decreases since they
correlate with the symptomes:
-ext3_inode_cache 254638 693552 760 43 8 : tunables 0 0
0 : slabdata 16130 16130 0
+ext3_inode_cache 81698 468958 760 43 8 : tunables 0 0
0 : slabdata 10906 10906 0
-dentry 96504 344232 192 42 2 : tunables 0 0
0 : slabdata 8196 8196 0
+dentry 55628 243810 192 42 2 : tunables 0 0
0 : slabdata 5805 5805 0
And if my understanding is correct, buffer_head would be meta-data
associated with cached pages and thus be expected to drop in
correlation with less data cached:
-buffer_head 2109250 4979052 104 39 1 : tunables 0
0 0 : slabdata 127668 127668 0
+buffer_head 838859 4059822 104 39 1 : tunables 0
0 0 : slabdata 104098 104098 0
My knowledge of the implementation is lacking far too much to know
where best to look for the likely culprit in terms of the root cause
of the eviction. The one thing I thought looked suspicious was the
kmalloc increases:
-kmalloc-4096 301 328 4096 8 8 : tunables 0 0
0 : slabdata 41 41 0
+kmalloc-4096 637 680 4096 8 8 : tunables 0 0
0 : slabdata 85 85 0
-kmalloc-2048 18215 19696 2048 16 8 : tunables 0 0
0 : slabdata 1231 1231 0
+kmalloc-2048 41908 51792 2048 16 8 : tunables 0 0
0 : slabdata 3237 3237 0
-kmalloc-1024 85444 97280 1024 32 8 : tunables 0 0
0 : slabdata 3040 3040 0
+kmalloc-1024 267031 327104 1024 32 8 : tunables 0 0
0 : slabdata 10222 10222 0
-kmalloc-512 1988 2176 512 32 4 : tunables 0 0
0 : slabdata 68 68 0
+kmalloc-512 1692 2080 512 32 4 : tunables 0 0
0 : slabdata 65 65 0
-kmalloc-256 102588 119776 256 32 2 : tunables 0 0
0 : slabdata 3743 3743 0
+kmalloc-256 308470 370720 256 32 2 : tunables 0 0
0 : slabdata 11585 11585 0
-kmalloc-128 8435 9760 128 32 1 : tunables 0 0
0 : slabdata 305 305 0
+kmalloc-128 8524 9760 128 32 1 : tunables 0 0
0 : slabdata 305 305 0
-kmalloc-64 96176 405440 64 64 1 : tunables 0 0
0 : slabdata 6335 6335 0
+kmalloc-64 50001 352448 64 64 1 : tunables 0 0
0 : slabdata 5507 5507 0
If my interpretation and understanding is correct, this indicates that
for example, ~3000 to ~10000 3-order allocations resulting from 1 kb
kmalloc():s. Meaning about 0.2 gig ( 7000*4*8*1024/1024/1024). Add the
other ones and we get some more, but only a few hundred megs in total.
Going by the hypothesis that we are seeing the same thing as reported
by Simon Kirby (I'll respond to that E-Mail separately), the total
amount is (as far as I understand) not the important part, but the
fact that we saw a non-trivial increase in 3-order allocations would
perhaps be a consistent observation in that frequent 3-order
allocations might be more likely to trigger the behavior Simon
reports.
I can do additional post-processing on the data that was dumped (such
as graphing it), but I'm not sure which parts would be the most
interesting for figuring out what is going on. Is there something else
I should be collecting for that matter, than the
vmstat/slabinfo/buddyinfo/meminfo?
--
/ Peter Schuller aka scode
--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/
Powered by blists - more mailing lists