[<prev] [next>] [thread-next>] [day] [month] [year] [list]
Message-Id: <1249409546-6343-1-git-send-email-mel@csn.ul.ie>
Date: Tue, 4 Aug 2009 19:12:22 +0100
From: Mel Gorman <mel@....ul.ie>
To: Larry Woodman <lwoodman@...hat.com>,
Andrew Morton <akpm@...ux-foundation.org>
Cc: riel@...hat.com, Ingo Molnar <mingo@...e.hu>,
Peter Zijlstra <peterz@...radead.org>,
LKML <linux-kernel@...r.kernel.org>, linux-mm@...ck.org,
Mel Gorman <mel@....ul.ie>
Subject: [PATCH 0/4] Add some trace events for the page allocator v3
Changelog since V2
o Added Ack-ed By's from Rik
o Only call trace_mm_page_free_direct when page count reaches zero
o Rebase to 2.6.31-rc5
Changelog since V1
o Fix minor formatting error for the __rmqueue event
o Add event for __pagevec_free
o Bring naming more in line with Larry Woodman's tracing patch
o Add an example post-processing script for the trace events
This is V3 of a patchset to add some trace points for the page allocator. This
version adds some ACKs, drops the RFC from the headline and fixes one issue
where trace_mm_page_free_direct triggered more than it should. The following
four patches add some trace events for the page allocator under the heading
of kmem.
Patch 1 adds events for plain old allocate and freeing of pages
Patch 2 gives information useful for analysing fragmentation avoidance
Patch 3 tracks pages going to and from the buddy lists as an indirect
indication of zone lock hotness
Patch 4 adds a post-processing script that aggegates the events to
give a higher-level view
The first one could be used as an indicator as to whether the workload was
heavily dependant on the page allocator or not. You can make a guess based
on vmstat but you can't get a per-process breakdown. Depending on the call
path, the call_site for page allocation may be __get_free_pages() instead
of a useful callsite. Instead of passing down a return address similar to
slab debugging, the user should enable the stacktrace and seg-addr options
to get a proper stack trace.
The second patch would mainly be useful for users of hugepages and
particularly dynamic hugepage pool resizing as it could be used to tune
min_free_kbytes to a level that fragmentation was rarely a problem. My
main concern is that maybe I'm trying to jam too much into the TP_printk
that could be extrapolated after the fact if you were familiar with the
implementation. I couldn't determine if it was best to hold the hand of
the administrator even if it cost more to figure it out.
The third patch is trickier to draw conclusions from but high activity on
those events could explain why there were a large number of cache misses
on a page-allocator-intensive workload. The coalescing and splitting of
buddies involves a lot of writing of page metadata and cache line bounces
not to mention the acquisition of an interrupt-safe lock necessary to enter
this path.
The fourth patch parses the trace buffer to draw a higher-level picture of
what is going on broken down on a per-process basis.
.../postprocess/trace-pagealloc-postprocess.pl | 131 ++++++++++++++
include/trace/events/kmem.h | 184 ++++++++++++++++++++
mm/page_alloc.c | 14 ++-
3 files changed, 328 insertions(+), 1 deletions(-)
create mode 100755 Documentation/trace/postprocess/trace-pagealloc-postprocess.pl
--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/
Powered by blists - more mailing lists