lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <20160113165957.GJ17512@dhcp22.suse.cz>
Date:	Wed, 13 Jan 2016 17:59:58 +0100
From:	Michal Hocko <mhocko@...nel.org>
To:	Vladimir Davydov <vdavydov@...tuozzo.com>
Cc:	Andrew Morton <akpm@...ux-foundation.org>,
	Johannes Weiner <hannes@...xchg.org>, linux-mm@...ck.org,
	cgroups@...r.kernel.org, linux-kernel@...r.kernel.org
Subject: Re: [PATCH v2 6/7] mm: free swap cache aggressively if memcg swap is
 full

On Thu 17-12-15 15:29:59, Vladimir Davydov wrote:
> Swap cache pages are freed aggressively if swap is nearly full (>50%
> currently), because otherwise we are likely to stop scanning anonymous
> when we near the swap limit even if there is plenty of freeable swap
> cache pages. We should follow the same trend in case of memory cgroup,
> which has its own swap limit.
> 
> Signed-off-by: Vladimir Davydov <vdavydov@...tuozzo.com>
> Acked-by: Johannes Weiner <hannes@...xchg.org>

I cannot say I would be deeply familiar with swapcache internalls and
all the interaction with the reclaim but from what I understand the
patch looks sane
Acked-by: Michal Hocko <mhocko@...e.com>

> ---
> Changes in v2:
>  - Remove unnecessary PageSwapCache check from mem_cgroup_swap_full.
>  - Do not check swap limit on the legacy hierarchy.
> 
>  include/linux/swap.h |  6 ++++++
>  mm/memcontrol.c      | 22 ++++++++++++++++++++++
>  mm/memory.c          |  3 ++-
>  mm/swapfile.c        |  2 +-
>  mm/vmscan.c          |  2 +-
>  5 files changed, 32 insertions(+), 3 deletions(-)
> 
> diff --git a/include/linux/swap.h b/include/linux/swap.h
> index c544998dfbe7..5ebdbabc62f0 100644
> --- a/include/linux/swap.h
> +++ b/include/linux/swap.h
> @@ -552,6 +552,7 @@ extern void mem_cgroup_swapout(struct page *page, swp_entry_t entry);
>  extern int mem_cgroup_try_charge_swap(struct page *page, swp_entry_t entry);
>  extern void mem_cgroup_uncharge_swap(swp_entry_t entry);
>  extern long mem_cgroup_get_nr_swap_pages(struct mem_cgroup *memcg);
> +extern bool mem_cgroup_swap_full(struct page *page);
>  #else
>  static inline void mem_cgroup_swapout(struct page *page, swp_entry_t entry)
>  {
> @@ -571,6 +572,11 @@ static inline long mem_cgroup_get_nr_swap_pages(struct mem_cgroup *memcg)
>  {
>  	return get_nr_swap_pages();
>  }
> +
> +static inline bool mem_cgroup_swap_full(struct page *page)
> +{
> +	return vm_swap_full();
> +}
>  #endif
>  
>  #endif /* __KERNEL__*/
> diff --git a/mm/memcontrol.c b/mm/memcontrol.c
> index e0e498f5ca32..fc25dc211eaf 100644
> --- a/mm/memcontrol.c
> +++ b/mm/memcontrol.c
> @@ -5749,6 +5749,28 @@ long mem_cgroup_get_nr_swap_pages(struct mem_cgroup *memcg)
>  	return nr_swap_pages;
>  }
>  
> +bool mem_cgroup_swap_full(struct page *page)
> +{
> +	struct mem_cgroup *memcg;
> +
> +	VM_BUG_ON_PAGE(!PageLocked(page), page);
> +
> +	if (vm_swap_full())
> +		return true;
> +	if (!do_swap_account || !cgroup_subsys_on_dfl(memory_cgrp_subsys))
> +		return false;
> +
> +	memcg = page->mem_cgroup;
> +	if (!memcg)
> +		return false;
> +
> +	for (; memcg != root_mem_cgroup; memcg = parent_mem_cgroup(memcg))
> +		if (page_counter_read(&memcg->swap) * 2 >= memcg->swap.limit)
> +			return true;
> +
> +	return false;
> +}
> +
>  /* for remember boot option*/
>  #ifdef CONFIG_MEMCG_SWAP_ENABLED
>  static int really_do_swap_account __initdata = 1;
> diff --git a/mm/memory.c b/mm/memory.c
> index 3b115dcaa26e..2bd6a78c142b 100644
> --- a/mm/memory.c
> +++ b/mm/memory.c
> @@ -2563,7 +2563,8 @@ int do_swap_page(struct mm_struct *mm, struct vm_area_struct *vma,
>  	}
>  
>  	swap_free(entry);
> -	if (vm_swap_full() || (vma->vm_flags & VM_LOCKED) || PageMlocked(page))
> +	if (mem_cgroup_swap_full(page) ||
> +	    (vma->vm_flags & VM_LOCKED) || PageMlocked(page))
>  		try_to_free_swap(page);
>  	unlock_page(page);
>  	if (page != swapcache) {
> diff --git a/mm/swapfile.c b/mm/swapfile.c
> index efa279221302..ab1a8a619676 100644
> --- a/mm/swapfile.c
> +++ b/mm/swapfile.c
> @@ -1009,7 +1009,7 @@ int free_swap_and_cache(swp_entry_t entry)
>  		 * Also recheck PageSwapCache now page is locked (above).
>  		 */
>  		if (PageSwapCache(page) && !PageWriteback(page) &&
> -				(!page_mapped(page) || vm_swap_full())) {
> +		    (!page_mapped(page) || mem_cgroup_swap_full(page))) {
>  			delete_from_swap_cache(page);
>  			SetPageDirty(page);
>  		}
> diff --git a/mm/vmscan.c b/mm/vmscan.c
> index ab52d865d922..1cd88e9b0383 100644
> --- a/mm/vmscan.c
> +++ b/mm/vmscan.c
> @@ -1206,7 +1206,7 @@ cull_mlocked:
>  
>  activate_locked:
>  		/* Not a candidate for swapping, so reclaim swap space. */
> -		if (PageSwapCache(page) && vm_swap_full())
> +		if (PageSwapCache(page) && mem_cgroup_swap_full(page))
>  			try_to_free_swap(page);
>  		VM_BUG_ON_PAGE(PageActive(page), page);
>  		SetPageActive(page);
> -- 
> 2.1.4

-- 
Michal Hocko
SUSE Labs

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ