[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <20110920084531.GB27675@tiehlicka.suse.cz>
Date: Tue, 20 Sep 2011 10:45:32 +0200
From: Michal Hocko <mhocko@...e.cz>
To: Johannes Weiner <jweiner@...hat.com>
Cc: Andrew Morton <akpm@...ux-foundation.org>,
KAMEZAWA Hiroyuki <kamezawa.hiroyu@...fujitsu.com>,
Daisuke Nishimura <nishimura@....nes.nec.co.jp>,
Balbir Singh <bsingharora@...il.com>,
Ying Han <yinghan@...gle.com>,
Greg Thelen <gthelen@...gle.com>,
Michel Lespinasse <walken@...gle.com>,
Rik van Riel <riel@...hat.com>,
Minchan Kim <minchan.kim@...il.com>,
Christoph Hellwig <hch@...radead.org>, linux-mm@...ck.org,
linux-kernel@...r.kernel.org
Subject: Re: [patch 04/11] mm: memcg: per-priority per-zone hierarchy scan
generations
On Mon 12-09-11 12:57:21, Johannes Weiner wrote:
> Memory cgroup limit reclaim currently picks one memory cgroup out of
> the target hierarchy, remembers it as the last scanned child, and
> reclaims all zones in it with decreasing priority levels.
>
> The new hierarchy reclaim code will pick memory cgroups from the same
> hierarchy concurrently from different zones and priority levels, it
> becomes necessary that hierarchy roots not only remember the last
> scanned child, but do so for each zone and priority level.
>
> Furthermore, detecting full hierarchy round-trips reliably will become
> crucial, so instead of counting on one iterator site seeing a certain
> memory cgroup twice, use a generation counter that is increased every
> time the child with the highest ID has been visited.
In principle I think the patch is good. I have some concerns about
locking and I would really appreciate some more description (like you
provided in the other email in this thread).
>
> Signed-off-by: Johannes Weiner <jweiner@...hat.com>
> ---
> mm/memcontrol.c | 60 +++++++++++++++++++++++++++++++++++++++---------------
> 1 files changed, 43 insertions(+), 17 deletions(-)
>
> diff --git a/mm/memcontrol.c b/mm/memcontrol.c
> index 912c7c7..f4b404e 100644
> --- a/mm/memcontrol.c
> +++ b/mm/memcontrol.c
> @@ -121,6 +121,11 @@ struct mem_cgroup_stat_cpu {
> unsigned long targets[MEM_CGROUP_NTARGETS];
> };
>
> +struct mem_cgroup_iter_state {
> + int position;
> + unsigned int generation;
> +};
> +
> /*
> * per-zone information in memory controller.
> */
> @@ -131,6 +136,8 @@ struct mem_cgroup_per_zone {
> struct list_head lists[NR_LRU_LISTS];
> unsigned long count[NR_LRU_LISTS];
>
> + struct mem_cgroup_iter_state iter_state[DEF_PRIORITY + 1];
> +
> struct zone_reclaim_stat reclaim_stat;
> struct rb_node tree_node; /* RB tree node */
> unsigned long long usage_in_excess;/* Set to the value by which */
[...]
> @@ -781,9 +783,15 @@ struct mem_cgroup *try_get_mem_cgroup_from_mm(struct mm_struct *mm)
> return memcg;
> }
>
> +struct mem_cgroup_iter {
Wouldn't be mem_cgroup_zone_iter_state a better name. It is true it is
rather long but I find mem_cgroup_iter very confusing because the actual
position is stored in the zone's state. The other thing is that it looks
like we have two iterators in mem_cgroup_iter function now but in fact
the iter parameter is just a state when we start iteration.
> + struct zone *zone;
> + int priority;
> + unsigned int generation;
> +};
> +
> static struct mem_cgroup *mem_cgroup_iter(struct mem_cgroup *root,
> struct mem_cgroup *prev,
> - bool remember)
> + struct mem_cgroup_iter *iter)
I would rather see a different name for the last parameter
(iter_state?).
> {
> struct mem_cgroup *mem = NULL;
> int id = 0;
> @@ -791,7 +799,7 @@ static struct mem_cgroup *mem_cgroup_iter(struct mem_cgroup *root,
> if (!root)
> root = root_mem_cgroup;
>
> - if (prev && !remember)
> + if (prev && !iter)
> id = css_id(&prev->css);
>
> if (prev && prev != root)
> @@ -804,10 +812,20 @@ static struct mem_cgroup *mem_cgroup_iter(struct mem_cgroup *root,
> }
>
> while (!mem) {
> + struct mem_cgroup_iter_state *uninitialized_var(is);
> struct cgroup_subsys_state *css;
>
> - if (remember)
> - id = root->last_scanned_child;
> + if (iter) {
> + int nid = zone_to_nid(iter->zone);
> + int zid = zone_idx(iter->zone);
> + struct mem_cgroup_per_zone *mz;
> +
> + mz = mem_cgroup_zoneinfo(root, nid, zid);
> + is = &mz->iter_state[iter->priority];
> + if (prev && iter->generation != is->generation)
> + return NULL;
> + id = is->position;
Do we need any kind of locking here (spin_lock(&is->lock))?
If two parallel reclaimers start on the same zone and priority they will
see the same position and so bang on the same cgroup.
> + }
>
> rcu_read_lock();
> css = css_get_next(&mem_cgroup_subsys, id + 1, &root->css, &id);
> @@ -818,8 +836,13 @@ static struct mem_cgroup *mem_cgroup_iter(struct mem_cgroup *root,
> id = 0;
> rcu_read_unlock();
>
> - if (remember)
> - root->last_scanned_child = id;
> + if (iter) {
> + is->position = id;
> + if (!css)
> + is->generation++;
> + else if (!prev && mem)
> + iter->generation = is->generation;
unlock it here.
> + }
>
> if (prev && !css)
> return NULL;
[...]
--
Michal Hocko
SUSE Labs
SUSE LINUX s.r.o.
Lihovarska 1060/12
190 00 Praha 9
Czech Republic
--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/
Powered by blists - more mailing lists