[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <20190329142033.GB2474@cmpxchg.org>
Date: Fri, 29 Mar 2019 10:20:33 -0400
From: Johannes Weiner <hannes@...xchg.org>
To: Shakeel Butt <shakeelb@...gle.com>
Cc: Matthew Wilcox <willy@...radead.org>,
Vladimir Davydov <vdavydov.dev@...il.com>,
Michal Hocko <mhocko@...e.com>,
Andrew Morton <akpm@...ux-foundation.org>,
Paolo Bonzini <pbonzini@...hat.com>,
Ben Gardon <bgardon@...gle.com>,
Radim Krčmář <rkrcmar@...hat.com>,
Linux MM <linux-mm@...ck.org>, kvm@...r.kernel.org,
LKML <linux-kernel@...r.kernel.org>
Subject: Re: [RFC PATCH] mm, kvm: account kvm_vcpu_mmap to kmemcg
On Thu, Mar 28, 2019 at 08:59:45PM -0700, Shakeel Butt wrote:
> On Thu, Mar 28, 2019 at 7:36 PM Matthew Wilcox <willy@...radead.org> wrote:
> > I don't understand why we need a PageKmemcg anyway. We already
> > have an entire pointer in struct page; can we not just check whether
> > page->mem_cgroup is NULL or not?
>
> PageKmemcg is for kmem while page->mem_cgroup is used for anon, file
> and kmem memory. So, page->mem_cgroup can not be used for NULL check
> unless we unify them. Not sure how complicated would that be.
A page flag warrants research into this.
The only reason we have PageKmemcg() is because of the way we do
memory type accounting at uncharge time:
if (!PageKmemcg(page)) {
unsigned int nr_pages = 1;
if (PageTransHuge(page)) {
nr_pages <<= compound_order(page);
ug->nr_huge += nr_pages;
}
if (PageAnon(page))
ug->nr_anon += nr_pages;
else {
ug->nr_file += nr_pages;
if (PageSwapBacked(page))
ug->nr_shmem += nr_pages;
}
ug->pgpgout++;
} else {
ug->nr_kmem += 1 << compound_order(page);
__ClearPageKmemcg(page);
}
[...]
if (!mem_cgroup_is_root(ug->memcg)) {
page_counter_uncharge(&ug->memcg->memory, nr_pages);
if (do_memsw_account())
page_counter_uncharge(&ug->memcg->memsw, nr_pages);
if (!cgroup_subsys_on_dfl(memory_cgrp_subsys) && ug->nr_kmem)
page_counter_uncharge(&ug->memcg->kmem, ug->nr_kmem);
memcg_oom_recover(ug->memcg);
}
local_irq_save(flags);
__mod_memcg_state(ug->memcg, MEMCG_RSS, -ug->nr_anon);
__mod_memcg_state(ug->memcg, MEMCG_CACHE, -ug->nr_file);
__mod_memcg_state(ug->memcg, MEMCG_RSS_HUGE, -ug->nr_huge);
__mod_memcg_state(ug->memcg, NR_SHMEM, -ug->nr_shmem);
__count_memcg_events(ug->memcg, PGPGOUT, ug->pgpgout);
But nothing says we have to have all these duplicate private counters,
or update them this late in the page's lifetime. The generic vmstat
counters in comparison are updated when 1) we know the page is going
away but 2) we still know the page type. We can do the same here.
We can either
a) Push up the MEMCG_RSS, MEMCG_CACHE etc. accounting sites to before
the pages are uncharged, when the page type is still known, but
page->mem_cgroup is exclusive, i.e. when they are deleted from page
cache or when their last pte is going away. This would be very
close to where the VM updates NR_ANON_MAPPED, NR_FILE_PAGES etc.
or
b) Tweak the existing NR_ANON_MAPPED, NR_FILE_PAGES, NR_ANON_THPS
accounting sites to use the lruvec_page_state infra and get rid of
the duplicate MEMCG_RSS, MEMCG_CACHE counters completely.
These sites would need slight adjustments, as they are sometimes
before commit_charge() set up page->mem_cgroup, but it doesn't look
too complicated to fix that ordering.
The latter would be a great cleanup, and frankly one that is long
overdue. There is no good reason for all this duplication. We'd not
only get rid of the private counters and the duplicate accounting
sites, it would drastically simplify charging and uncharging, and it
would even obviate the need for a separate kmem (un)charge path.
[ The cgroup1 memcg->kmem thing is the odd-one-out, but I think this
is complete legacy at this point and nobody is actively setting
limits on that counter anyway. We can break out an explicit v1-only
mem_cgroup_charge_legacy_kmem(), put it into the currently accounted
callsites for compatibility, and not add any new ones. ]
Powered by blists - more mailing lists