[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <486C7C82.4020600@linux.vnet.ibm.com>
Date: Thu, 03 Jul 2008 12:45:14 +0530
From: Balbir Singh <balbir@...ux.vnet.ibm.com>
To: KAMEZAWA Hiroyuki <kamezawa.hiroyu@...fujitsu.com>
CC: Hugh Dickins <hugh@...itas.com>,
Andrew Morton <akpm@...ux-foundation.org>,
Rik van Riel <riel@...hat.com>,
Lee Schermerhorn <lee.schermerhorn@...com>,
Balbir Singh <balbir@...ibm.com>, linux-kernel@...r.kernel.org
Subject: Re: [RFC][PATCH] memcg: change shmem handler.
KAMEZAWA Hiroyuki wrote:
> On Sun, 29 Jun 2008 01:22:22 +0100 (BST)
> Hugh Dickins <hugh@...itas.com> wrote:
>
>> add_to_page_cache_lru puts PageSwapBacked pages on the active_anon lru,
>> so shouldn't mem_cgroup_charge_common mirror that by setting FLAG_ACTIVE?
>>
>> Signed-off-by: Hugh Dickins <hugh@...itas.com>
>
> How about a patch like this ?
> ==
> a RFC patch: memcg-change-shmem-handler.
>
> Should be divided into 2-4 patches, and may have some problem.
> (I did many careless misses today.....)
>
> But please see idea/concepts. Maybe right way.
>
> shmem has following characteristics.
> In general.
> - seems like file cache
> - swap-backed
> - when it's swapped out, it's removed from radix-tree and added to swap.
> - when it's swapped in, it's removed from swap and added to radix-tree.
>
> With memcg.
> - shmem is treted just as a file-cache. So, started from inactive list.
> - shmem's page fault routine is sensitive to GFP_xxx in which used.
> (GFP_NOWAIT is used) and pre-charge is done before add_to_page_cache.
> - shmem's page is removed by mem_cgroup_uncharge_cache_page(), So,
> shmem's swapcache is not charged.
>
> This patch fixes some mess by
> - PAGE_CGROUP_FLAG_CACHE is deleted (and replaced by FLAG_FILE)
> - PAGE_CGROUP_FLAG_SHMEM is added.
> - add_to_page_cache_nocharge() is added.
> This avoids mem_cgroup_charge_cache_page(). This is useful when page is
> pre-charged.
> - uses add_to_page_cache_nocharge() also in hugemem.
> (I think hugemem controller should be independent from memcg.
> Balbir, how do you think ?)
I am not 100% sure of that right now. I definitely want different control
parameters (not included as a part of memory.limit_in_bytes). If there is
leverage from memory controller, we could consider adding that to it or a
separate controller if that makes more sense.
> - PageSwapBacked() is checked.
> (A imported patch from Hugh Dickins)
>
> As result.
> - shmem will be in SwapBacked/Active list at first.
Good
> - memcg has "shmem/tmpfs" counter.
>
I think this will be useful too (but I need to play with it)
> Signed-off-by: KAMEZAWA Hiroyuki <kamezawa.hiroyu@...fujitsu.com>
>
> --
> include/linux/pagemap.h | 16 +++++++++
> mm/filemap.c | 49 ++++++++++++++++++++++--------
> mm/hugetlb.c | 3 +
> mm/memcontrol.c | 78 +++++++++++++++++++++++++-----------------------
> mm/shmem.c | 17 ++++++----
> 5 files changed, 107 insertions(+), 56 deletions(-)
>
> Index: linux-2.6.26-rc5-mm3-kosaki/mm/memcontrol.c
> ===================================================================
> --- linux-2.6.26-rc5-mm3-kosaki.orig/mm/memcontrol.c 2008-06-30 15:02:52.000000000 +0900
> +++ linux-2.6.26-rc5-mm3-kosaki/mm/memcontrol.c 2008-06-30 16:26:34.000000000 +0900
> @@ -49,6 +49,7 @@
> */
> MEM_CGROUP_STAT_CACHE, /* # of pages charged as cache */
> MEM_CGROUP_STAT_RSS, /* # of pages charged as anon/swapcache */
> + MEM_CGROUP_STAT_SHMEM, /* # of pages charges as shmem/tmpfs */
> MEM_CGROUP_STAT_PGPGIN_COUNT, /* # of pages paged in */
> MEM_CGROUP_STAT_PGPGOUT_COUNT, /* # of pages paged out */
>
> @@ -160,10 +161,10 @@
> struct mem_cgroup *mem_cgroup;
> int flags;
> };
> -#define PAGE_CGROUP_FLAG_CACHE (0x1) /* charged as cache */
> -#define PAGE_CGROUP_FLAG_ACTIVE (0x2) /* page is active in this cgroup */
> -#define PAGE_CGROUP_FLAG_FILE (0x4) /* page is file system backed */
> -#define PAGE_CGROUP_FLAG_UNEVICTABLE (0x8) /* page is unevictableable */
> +#define PAGE_CGROUP_FLAG_ACTIVE (0x1) /* page is active in this cgroup */
> +#define PAGE_CGROUP_FLAG_FILE (0x2) /* page is file system backed */
> +#define PAGE_CGROUP_FLAG_UNEVICTABLE (0x4) /* page is unevictableable */
> +#define PAGE_CGROUP_FLAG_SHMEM (0x8) /* page is shmem/tmpfs */
>
> static int page_cgroup_nid(struct page_cgroup *pc)
> {
> @@ -178,6 +179,7 @@
> enum charge_type {
> MEM_CGROUP_CHARGE_TYPE_CACHE = 0,
> MEM_CGROUP_CHARGE_TYPE_MAPPED,
> + MEM_CGROUP_CHARGE_TYPE_SHMEM,
> MEM_CGROUP_CHARGE_TYPE_FORCE, /* used by force_empty */
> };
>
> @@ -191,8 +193,10 @@
> struct mem_cgroup_stat *stat = &mem->stat;
>
> VM_BUG_ON(!irqs_disabled());
> - if (flags & PAGE_CGROUP_FLAG_CACHE)
> + if (flags & PAGE_CGROUP_FLAG_FILE)
> __mem_cgroup_stat_add_safe(stat, MEM_CGROUP_STAT_CACHE, val);
> + else if (flags & PAGE_CGROUP_FLAG_SHMEM)
> + __mem_cgroup_stat_add_safe(stat, MEM_CGROUP_STAT_SHMEM, val);
> else
> __mem_cgroup_stat_add_safe(stat, MEM_CGROUP_STAT_RSS, val);
>
> @@ -573,12 +577,19 @@
> * If a page is accounted as a page cache, insert to inactive list.
> * If anon, insert to active list.
> */
> - if (ctype == MEM_CGROUP_CHARGE_TYPE_CACHE) {
> - pc->flags = PAGE_CGROUP_FLAG_CACHE;
> - if (page_is_file_cache(page))
> - pc->flags |= PAGE_CGROUP_FLAG_FILE;
> - } else
> + switch (ctype) {
> + case MEM_CGROUP_CHARGE_TYPE_CACHE:
> + pc->flags = PAGE_CGROUP_FLAG_FILE;
> + break;
> + case MEM_CGROUP_CHARGE_TYPE_SHMEM:
> + pc->flags = PAGE_CGROUP_FLAG_SHMEM | PAGE_CGROUP_FLAG_ACTIVE;
> + break;
> + case MEM_CGROUP_CHARGE_TYPE_MAPPED:
> pc->flags = PAGE_CGROUP_FLAG_ACTIVE;
> + break;
> + default:
> + BUG();
> + }
>
> lock_page_cgroup(page);
> if (unlikely(page_get_page_cgroup(page))) {
> @@ -625,28 +636,10 @@
> int mem_cgroup_cache_charge(struct page *page, struct mm_struct *mm,
> gfp_t gfp_mask)
> {
> - /*
> - * Corner case handling. This is called from add_to_page_cache()
> - * in usual. But some FS (shmem) precharges this page before calling it
> - * and call add_to_page_cache() with GFP_NOWAIT.
> - *
> - * For GFP_NOWAIT case, the page may be pre-charged before calling
> - * add_to_page_cache(). (See shmem.c) check it here and avoid to call
> - * charge twice. (It works but has to pay a bit larger cost.)
> - */
> - if (!(gfp_mask & __GFP_WAIT)) {
> - struct page_cgroup *pc;
> + enum charge_type ctype = MEM_CGROUP_CHARGE_TYPE_CACHE;
>
> - lock_page_cgroup(page);
> - pc = page_get_page_cgroup(page);
> - if (pc) {
> - VM_BUG_ON(pc->page != page);
> - VM_BUG_ON(!pc->mem_cgroup);
> - unlock_page_cgroup(page);
> - return 0;
> - }
> - unlock_page_cgroup(page);
> - }
> + if (PageSwapBacked(page))
> + ctype = MEM_CGROUP_CHARGE_TYPE_SHMEM;
>
> if (unlikely(!mm))
> mm = &init_mm;
> @@ -678,11 +671,19 @@
> goto unlock;
>
> VM_BUG_ON(pc->page != page);
> -
> + /*
> + * There are 2 cases.
> + * 1. anon pages are swapped out.
> + * 2. shmem pages are swapped out.
> + * In both case, PageSwapCache() returns 1 and we don't want to
> + * uncharge it.
> + */
> + if (PageSwapCache(page))
> + goto unlock;
> + /* When the page is unmapped, file-cache and shmem memory is alive */
> if ((ctype == MEM_CGROUP_CHARGE_TYPE_MAPPED)
> - && ((pc->flags & PAGE_CGROUP_FLAG_CACHE)
> - || page_mapped(page)
> - || PageSwapCache(page)))
> + && ((pc->flags & (PAGE_CGROUP_FLAG_FILE | PAGE_CGROUP_FLAG_SHMEM))
> + || page_mapped(page)))
> goto unlock;
>
> mz = page_cgroup_zoneinfo(pc);
> @@ -732,8 +733,10 @@
> if (pc) {
> mem = pc->mem_cgroup;
> css_get(&mem->css);
> - if (pc->flags & PAGE_CGROUP_FLAG_CACHE)
> + if (pc->flags & PAGE_CGROUP_FLAG_FILE)
> ctype = MEM_CGROUP_CHARGE_TYPE_CACHE;
> + else if (pc->flags & PAGE_CGROUP_FLAG_SHMEM)
> + ctype = MEM_CGROUP_CHARGE_TYPE_SHMEM;
> }
> unlock_page_cgroup(page);
> if (mem) {
> @@ -785,6 +788,8 @@
> progress = try_to_free_mem_cgroup_pages(mem, gfp_mask);
> } while (!progress && --retry);
>
> + css_put(&mem->css);
> +
Didn't Hugh fix this as a part of his patchset.
> if (!retry)
> return -ENOMEM;
> return 0;
> @@ -920,6 +925,7 @@
> } mem_cgroup_stat_desc[] = {
> [MEM_CGROUP_STAT_CACHE] = { "cache", PAGE_SIZE, },
> [MEM_CGROUP_STAT_RSS] = { "anon/swapcache", PAGE_SIZE, },
> + [MEM_CGROUP_STAT_SHMEM] = { "shmem/tmpfs", PAGE_SIZE, },
> [MEM_CGROUP_STAT_PGPGIN_COUNT] = {"pgpgin", 1, },
> [MEM_CGROUP_STAT_PGPGOUT_COUNT] = {"pgpgout", 1, },
> };
> Index: linux-2.6.26-rc5-mm3-kosaki/include/linux/pagemap.h
> ===================================================================
> --- linux-2.6.26-rc5-mm3-kosaki.orig/include/linux/pagemap.h 2008-06-30 15:00:04.000000000 +0900
> +++ linux-2.6.26-rc5-mm3-kosaki/include/linux/pagemap.h 2008-06-30 15:03:02.000000000 +0900
> @@ -258,6 +258,22 @@
> pgoff_t index, gfp_t gfp_mask);
> int add_to_page_cache_lru(struct page *page, struct address_space *mapping,
> pgoff_t index, gfp_t gfp_mask);
> +
> +#ifdef CONFIG_CGROUP_MEM_RES_CTLR
> +/*
> + * If the page is pre-charged before add_to_page_cache() this routine is
> + * light-weight.
> + */
> +int add_to_page_cache_nocharge(struct page *page, struct address_space *mapping,
> + pgoff_t index, gfp_t gfp_mask);
> +#else
> +static inline int add_to_page_cache_nocharge(struct page *page,
> + struct address_space *mapping, pgoff_t index,
> + gfp_t mask)
> +{
> + return add_to_page_cache(page, mapping, index, gfp_mask);
> +}
> +#endif
> extern void remove_from_page_cache(struct page *page);
> extern void __remove_from_page_cache(struct page *page);
>
> Index: linux-2.6.26-rc5-mm3-kosaki/mm/filemap.c
> ===================================================================
> --- linux-2.6.26-rc5-mm3-kosaki.orig/mm/filemap.c 2008-06-30 15:00:04.000000000 +0900
> +++ linux-2.6.26-rc5-mm3-kosaki/mm/filemap.c 2008-06-30 15:23:36.000000000 +0900
> @@ -442,8 +442,9 @@
> return err;
> }
>
> +
> /**
> - * add_to_page_cache - add newly allocated pagecache pages
> + * add_to_page_cache_nocharge() - add newly allocated pagecache pages
> * @page: page to add
> * @mapping: the page's address_space
> * @offset: page index
> @@ -454,22 +455,20 @@
> * The other page state flags were set by rmqueue().
> *
> * This function does not add the page to the LRU. The caller must do that.
> + * This doesn't call memory resource control routine.
> */
> -int add_to_page_cache(struct page *page, struct address_space *mapping,
> - pgoff_t offset, gfp_t gfp_mask)
> +
> +int add_to_page_cache_nocharge(struct page *page,
> + struct address_space *mapping,
> + pgoff_t offset, gfp_t gfp_mask)
> {
I wonder if we should modify a generic routine and add charge/no charge variants
to it. Charging is very memory controller specific operation. Can't we infer the
charging/no charging from gfp_mask or another flag?
> - int error = mem_cgroup_cache_charge(page, current->mm,
> - gfp_mask & ~__GFP_HIGHMEM);
> - if (error)
> - goto out;
>
> - error = radix_tree_preload(gfp_mask & ~__GFP_HIGHMEM);
> + int error = radix_tree_preload(gfp_mask & ~__GFP_HIGHMEM);
> if (error == 0) {
> page_cache_get(page);
> SetPageLocked(page);
> page->mapping = mapping;
> page->index = offset;
> -
> spin_lock_irq(&mapping->tree_lock);
> error = radix_tree_insert(&mapping->page_tree, offset, page);
> if (likely(!error)) {
> @@ -478,15 +477,39 @@
> } else {
> page->mapping = NULL;
> ClearPageLocked(page);
> - mem_cgroup_uncharge_cache_page(page);
> page_cache_release(page);
> }
>
> spin_unlock_irq(&mapping->tree_lock);
> radix_tree_preload_end();
> - } else
> - mem_cgroup_uncharge_cache_page(page);
> -out:
> + }
> + return error;
> +}
> +
> +/**
> + * add_to_page_cache - add newly allocated pagecache pages
> + * @page: page to add
> + * @mapping: the page's address_space
> + * @offset: page index
> + * @gfp_mask: page allocation mode
> + *
> + * This function is used to add newly allocated pagecache pages;
> + * the page is new, so we can just run SetPageLocked() against it.
> + * The other page state flags were set by rmqueue().
> + *
> + * This function does not add the page to the LRU. The caller must do that.
> + */
> +int add_to_page_cache(struct page *page, struct address_space *mapping,
> + pgoff_t offset, gfp_t gfp_mask)
> +{
> + int error = mem_cgroup_cache_charge(page, current->mm,
> + gfp_mask & ~__GFP_HIGHMEM);
> + if (!error) {
> + error = add_to_page_cache_nocharge(page, mapping, offset,
> + gfp_mask);
> + if (error)
> + mem_cgroup_uncharge_cache_page(page);
> + }
> return error;
> }
> EXPORT_SYMBOL(add_to_page_cache);
> Index: linux-2.6.26-rc5-mm3-kosaki/mm/shmem.c
> ===================================================================
> --- linux-2.6.26-rc5-mm3-kosaki.orig/mm/shmem.c 2008-06-30 15:00:04.000000000 +0900
> +++ linux-2.6.26-rc5-mm3-kosaki/mm/shmem.c 2008-06-30 15:37:11.000000000 +0900
> @@ -940,10 +940,8 @@
> spin_lock(&info->lock);
> ptr = shmem_swp_entry(info, idx, NULL);
> if (ptr && ptr->val == entry.val)
> - error = add_to_page_cache(page, inode->i_mapping, idx,
> + error = add_to_page_cache_nocharge(page, inode->i_mapping, idx,
> GFP_NOWAIT);
> - else /* we don't have to account this page. */
> - mem_cgroup_uncharge_cache_page(page);
>
> if (error == -EEXIST) {
> struct page *filepage = find_get_page(inode->i_mapping, idx);
> @@ -955,6 +953,7 @@
> */
> if (PageUptodate(filepage))
> error = 0;
> + mem_cgroup_uncharge_cache_page(page);
> page_cache_release(filepage);
> }
> }
> @@ -965,7 +964,9 @@
> shmem_swp_set(info, ptr, 0);
> swap_free(entry);
> error = 1; /* not an error, but entry was found */
> - }
> + } else
> + mem_cgroup_uncharge_cache_page(page);
> +
> if (ptr)
> shmem_swp_unmap(ptr);
> spin_unlock(&info->lock);
> @@ -1375,6 +1376,7 @@
> error = -ENOMEM;
> goto failed;
> }
> + SetPageSwapBacked(filepage);
>
> /* Precharge page while we can wait, compensate after */
> error = mem_cgroup_cache_charge(filepage, current->mm,
> @@ -1387,7 +1389,6 @@
> goto failed;
> }
>
> - SetPageSwapBacked(filepage);
I thought I saw a patch from Hugh for this one too.. but I might be missing the
chronological ordering of which came first.
> spin_lock(&info->lock);
> entry = shmem_swp_alloc(info, idx, sgp);
> if (IS_ERR(entry))
> @@ -1400,7 +1401,8 @@
> if (ret)
> mem_cgroup_uncharge_cache_page(filepage);
> else
> - ret = add_to_page_cache_lru(filepage, mapping,
> + ret = add_to_page_cache_nocharge(filepage,
> + mapping,
> idx, GFP_NOWAIT);
> /*
> * At add_to_page_cache_lru() failure, uncharge will
> @@ -1408,6 +1410,7 @@
> */
> if (ret) {
> spin_unlock(&info->lock);
> + mem_cgroup_uncharge_cache_page(filepage);
> page_cache_release(filepage);
> shmem_unacct_blocks(info->flags, 1);
> shmem_free_blocks(inode, 1);
> @@ -1415,6 +1418,8 @@
> if (error)
> goto failed;
> goto repeat;
> + } else {
> + lru_cache_add_active_anon(filepage);
> }
> info->flags |= SHMEM_PAGEIN;
> }
> Index: linux-2.6.26-rc5-mm3-kosaki/mm/hugetlb.c
> ===================================================================
> --- linux-2.6.26-rc5-mm3-kosaki.orig/mm/hugetlb.c 2008-06-25 18:07:09.000000000 +0900
> +++ linux-2.6.26-rc5-mm3-kosaki/mm/hugetlb.c 2008-06-30 15:08:16.000000000 +0900
> @@ -1813,7 +1813,8 @@
> int err;
> struct inode *inode = mapping->host;
>
> - err = add_to_page_cache(page, mapping, idx, GFP_KERNEL);
> + err = add_to_page_cache_nocharge(page, mapping,
> + idx, GFP_KERNEL);
> if (err) {
> put_page(page);
> if (err == -EEXIST)
>
--
Warm Regards,
Balbir Singh
Linux Technology Center
IBM, ISTL
--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/
Powered by blists - more mailing lists