[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-Id: <20090108093040.22d5f281.kamezawa.hiroyu@jp.fujitsu.com>
Date: Thu, 8 Jan 2009 09:30:40 +0900
From: KAMEZAWA Hiroyuki <kamezawa.hiroyu@...fujitsu.com>
To: Balbir Singh <balbir@...ux.vnet.ibm.com>
Cc: Andrew Morton <akpm@...ux-foundation.org>,
Sudhir Kumar <skumar@...ux.vnet.ibm.com>,
YAMAMOTO Takashi <yamamoto@...inux.co.jp>,
Paul Menage <menage@...gle.com>, lizf@...fujitsu.com,
linux-kernel@...r.kernel.org, linux-mm@...ck.org,
David Rientjes <rientjes@...gle.com>,
Pavel Emelianov <xemul@...nvz.org>, riel@...hat.com,
"kosaki.motohiro@...fujitsu.com" <kosaki.motohiro@...fujitsu.com>
Subject: Re: [RFC][PATCH 0/4] Memory controller soft limit patches
On Thu, 08 Jan 2009 00:11:10 +0530
Balbir Singh <balbir@...ux.vnet.ibm.com> wrote:
>
> Here is v1 of the new soft limit implementation. Soft limits is a new feature
> for the memory resource controller, something similar has existed in the
> group scheduler in the form of shares. We'll compare shares and soft limits
> below. I've had soft limit implementations earlier, but I've discarded those
> approaches in favour of this one.
>
> Soft limits are the most useful feature to have for environments where
> the administrator wants to overcommit the system, such that only on memory
> contention do the limits become active. The current soft limits implementation
> provides a soft_limit_in_bytes interface for the memory controller and not
> for memory+swap controller. The implementation maintains an RB-Tree of groups
> that exceed their soft limit and starts reclaiming from the group that
> exceeds this limit by the maximum amount.
>
> This is an RFC implementation and is not meant for inclusion
>
Core implemantation seems simple and the feature sounds good.
But, before reviewing into details, 3 points.
1. please fix current bugs on hierarchy management, before new feature.
AFAIK, OOM-Kill under hierarchy is broken. (I have patches but waits for
merge window close.)
I wonder there will be some others. Lockdep error which Nishimura reported
are all fixed now ?
2. You inserts reclaim-by-soft-limit into alloc_pages(). But, to do this,
you have to pass zonelist to try_to_free_mem_cgroup_pages() and have to modify
try_to_free_mem_cgroup_pages().
2-a) If not, when the memory request is for gfp_mask==GFP_DMA or allocation
is under a cpuset, memory reclaim will not work correctlly.
2-b) try_to_free_mem_cgroup_pages() cannot do good work for order > 1 allocation.
Please try fake-numa (or real NUMA machine) and cpuset.
3. If you want to insert hooks to "generic" page allocator, it's better to add CC to
Rik van Riel, Kosaki Motohiro, at leaset.
To be honest, I myself don't like to add a hook to alloc_pages() directly.
Can we implment call soft-limit like kswapd (or on kswapd()) ?
i.e. in moderate way ?
A happy new year,
-Kame
> TODOs
>
> 1. The shares interface is not yet implemented, the current soft limit
> implementation is not yet hierarchy aware. The end goal is to add
> a shares interface on top of soft limits and to maintain shares in
> a manner similar to the group scheduler
> 2. The current implementation maintains the delta from the soft limit
> and pushes back groups to their soft limits, a ratio of delta/soft_limit
> is more useful
> 3. It would be nice to have more targetted reclaim (in terms of pages to
> recalim) interface. So that groups are pushed back, close to their soft
> limits.
>
> Tests
> -----
>
> I've run two memory intensive workloads with differing soft limits and
> seen that they are pushed back to their soft limit on contention. Their usage
> was their soft limit plus additional memory that they were able to grab
> on the system.
>
> Please review, comment.
>
> Series
> ------
>
> memcg-soft-limit-documentation.patch
> memcg-add-soft-limit-interface.patch
> memcg-organize-over-soft-limit-groups.patch
> memcg-soft-limit-reclaim-on-contention.patch
>
> --
> Balbir
>
--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/
Powered by blists - more mailing lists