[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <20110714110935.GK19408@tiehlicka.suse.cz>
Date: Thu, 14 Jul 2011 13:09:35 +0200
From: Michal Hocko <mhocko@...e.cz>
To: KAMEZAWA Hiroyuki <kamezawa.hiroyu@...fujitsu.com>
Cc: linux-mm@...ck.org, Balbir Singh <bsingharora@...il.com>,
Daisuke Nishimura <nishimura@....nes.nec.co.jp>,
linux-kernel@...r.kernel.org
Subject: Re: [PATCH 1/2] memcg: make oom_lock 0 and 1 based rather than
coutner
On Thu 14-07-11 19:17:28, KAMEZAWA Hiroyuki wrote:
> On Thu, 14 Jul 2011 11:51:52 +0200
> Michal Hocko <mhocko@...e.cz> wrote:
>
> > On Thu 14-07-11 18:30:14, KAMEZAWA Hiroyuki wrote:
> > > On Thu, 14 Jul 2011 11:00:17 +0200
> > > Michal Hocko <mhocko@...e.cz> wrote:
> > >
> > > > On Thu 14-07-11 11:59:13, KAMEZAWA Hiroyuki wrote:
> > > > > On Thu, 14 Jul 2011 10:02:59 +0900
> > > > > KAMEZAWA Hiroyuki <kamezawa.hiroyu@...fujitsu.com> wrote:
[...]
> > > ==
> > > for_each_mem_cgroup_tree(iter, mem) {
> > > - x = atomic_inc_return(&iter->oom_lock);
> > > - lock_count = max(x, lock_count);
> > > + x = !!atomic_add_unless(&iter->oom_lock, 1, 1);
> > > + if (lock_count == -1)
> > > + lock_count = x;
> > > +
> > > + /* New child can be created but we shouldn't race with
> > > + * somebody else trying to oom because we are under
> > > + * memcg_oom_mutex
> > > + */
> > > + BUG_ON(lock_count != x);
> > > }
> > > ==
> > >
> > > When, B,D,E is under OOM,
> > >
> > > A oom_lock = 0
> > > B oom_lock = 1
> > > C oom_lock = 0
> > > D oom_lock = 1
> > > E oom_lock = 1
> > >
> > > Here, assume A enters OOM.
> > >
> > > A oom_lock = 1 -- (*)
> > > B oom_lock = 1
> > > C oom_lock = 1
> > > D oom_lock = 1
> > > E oom_lock = 1
> > >
> > > because of (*), mem_cgroup_oom_lock() will return lock_count=1, true.
> > >
> > > Then, a new oom-killer will another oom-kiiler running in B-D-E.
> >
> > OK, does this mean that for_each_mem_cgroup_tree doesn't lock the whole
> > hierarchy at once?
>
> yes. this for_each_mem_cgroup_tree() just locks a subtree.
OK, then I really misunderstood the macro and now I see your points.
Thinking about it some more having a full hierarchy locked is not that
good idea after all. We would block also parallel branches which will
not bail out from OOM if we handle oom condition in another branch.
>
> > I have to confess that the behavior of mem_cgroup_start_loop is little
> > bit obscure to me. The comment says it searches for the cgroup with the
> > minimum ID - I assume this is the root of the hierarchy. Is this
> > correct?
> >
>
> No. Assume following sequence.
>
> 1. cgcreate -g memory:X css_id=5 assigned.
> ........far later.....
> 2. cgcreate -g memory:A css_id=30 assigned.
> 3. cgdelete -g memory:X css_id=5 freed.
> 4. cgcreate -g memory:A/B
> 5. cgcreate -g memory:A/C
> 6. cgcreate -g memory:A/B/D
> 7. cgcreate -g memory:A/B/E
>
> Then, css_id will be
> ==
> A css_id=30
> B css_id=5 # reuse X's id.
> C css_id=31
> D css_id=32
> E css_id=33
> ==
> Then, the search under "B" will find B->D->E
>
> The search under "A" will find B->A->C->D->E.
>
> > If yes then if we have oom in what-ever cgroup in the hierarchy then
> > the above code should lock the whole hierarchy and the above never
> > happens. Right?
>
> Yes and no. old code allows following happens at the same time.
>
> A
> B C
> D E F
>
> B-D-E goes into OOM because of B's limit.
> C-F goes into OOM because of C's limit
>
>
> When you stop OOM under A because of B's limit, C can't invoke OOM.
>
> After a little more consideration, my suggestion is,
>
> === lock ===
> bool success = true;
> ...
> for_each_mem_cgroup_tree(iter, mem) {
> success &= !!atomic_add_unless(&iter->oom_lock, 1, 1);
> /* "break" loop is not allowed because of css refcount....*/
> }
> return success.
>
> By this, when a sub-hierarchy is under OOM, don't invoke new OOM.
Hmm, I am afraid this will not work as well. The group tree traversing
depends on the creation order so we might end up seeing locked subtree
sooner than unlocked so we could grant the lock and see multiple OOMs.
We have to guarantee that we do not grant the lock if we encounter
already locked sub group (and then we have to clear oom_lock for all
groups that we have already visited).
> === unlock ===
> struct mem_cgroup *oom_root;
>
> oom_root = memcg;
> do {
> struct mem_cgroup *parent;
>
> parent = mem_cgroup_parent(oom_root);
> if (!parent || !parent->use_hierarchy)
> break;
>
> if (atomic_read(&parent->oom_lock))
> break;
> } while (1);
>
> for_each_mem_cgroup_tree(iter, oom_root)
> atomic_add_unless(&iter->oom_lock, -1, 0);
>
> By this, at unlock, unlock oom-lock of a hierarchy which was under oom_lock
> because of a sub-hierarchy was under OOM.
This would unlock also groups that might have a parallel oom lock.
A - B - C - D oom (from B)
- E - F oom (F)
unlock in what-ever branch will unlock also the parallel oom.
I will think about something else and return to your first patch if I
find it over complicated as well.
--
Michal Hocko
SUSE Labs
SUSE LINUX s.r.o.
Lihovarska 1060/12
190 00 Praha 9
Czech Republic
--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/
Powered by blists - more mailing lists