lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Date:	Fri, 29 Feb 2008 16:03:20 +0000
From:	Andy Whitcroft <apw@...dowen.org>
To:	Rik van Riel <riel@...hat.com>
Cc:	linux-kernel@...r.kernel.org,
	KOSAKI Motohiro <kosaki.motohiro@...fujitsu.com>,
	Lee Schermerhorn <Lee.Schermerhorn@...com>, linux-mm@...ck.org,
	Christoph Lameter <clameter@....com>
Subject: Re: [patch 02/21] Use an indexed array for LRU variables

On Thu, Feb 28, 2008 at 02:29:10PM -0500, Rik van Riel wrote:
> V3 [riel]: memcontrol LRU arrayification
> 
> V1 -> V2 [lts]:
> + Remove extraneous  __dec_zone_state(zone, NR_ACTIVE) pointed
>   out by Mel G.
> 
> >From clameter@....com Wed Aug 29 11:39:51 2007
> 
> Currently we are defining explicit variables for the inactive
> and active list. An indexed array can be more generic and avoid
> repeating similar code in several places in the reclaim code.
> 
> We are saving a few bytes in terms of code size:
> 
> Before:
> 
>    text    data     bss     dec     hex filename
> 4097753  573120 4092484 8763357  85b7dd vmlinux
> 
> After:
> 
>    text    data     bss     dec     hex filename
> 4097729  573120 4092484 8763333  85b7c5 vmlinux
> 
> Having an easy way to add new lru lists may ease future work on
> the reclaim code.
> 
> Signed-off-by: Rik van Riel <riel@...hat.com>
> Signed-off-by: Lee Schermerhorn <lee.schermerhorn@...com>
> Signed-off-by: Christoph Lameter <clameter@....com>
> 
>  include/linux/memcontrol.h |   17 +---
>  include/linux/mm_inline.h  |   33 +++++---
>  include/linux/mmzone.h     |   17 ++--
>  mm/memcontrol.c            |  116 ++++++++++------------------
>  mm/page_alloc.c            |    9 +-
>  mm/swap.c                  |    2 
>  mm/vmscan.c                |  141 +++++++++++++++++------------------
>  mm/vmstat.c                |    3 
>  8 files changed, 158 insertions(+), 180 deletions(-)
> 
> Index: linux-2.6.25-rc2-mm1/include/linux/mmzone.h
> ===================================================================
> --- linux-2.6.25-rc2-mm1.orig/include/linux/mmzone.h	2008-02-19 16:23:16.000000000 -0500
> +++ linux-2.6.25-rc2-mm1/include/linux/mmzone.h	2008-02-27 14:42:23.000000000 -0500
> @@ -80,8 +80,8 @@ struct zone_padding {
>  enum zone_stat_item {
>  	/* First 128 byte cacheline (assuming 64 bit words) */
>  	NR_FREE_PAGES,
> -	NR_INACTIVE,
> -	NR_ACTIVE,
> +	NR_INACTIVE,	/* must match order of LRU_[IN]ACTIVE */
> +	NR_ACTIVE,	/*  "     "     "   "       "         */

This little ordering constraint is a little nasty.  If we have enum_list
available at this point then we can make sure that these order correctly
automatically with something like this:

	#define NR_LRU(lru) NR_ ## lru = NR_LRU_BASE + LRU_ ## lru
	enum foo {
		NR_FOO,
		NR_BAR,

		NR_LRU_BASE,
		NR_LRU(ACTIVE),
		NR_LRU(INACTIVE),
		NR_LRU_END = NR_LRU_BASE + NR_LRU_LISTS - 1,

		NR_YOU,
		NR_ME,
	};

The NR_LRU_END must be last to ensure that counting continues for those
later.

>  	NR_ANON_PAGES,	/* Mapped anonymous pages */
>  	NR_FILE_MAPPED,	/* pagecache pages mapped into pagetables.
>  			   only modified from process context */
> @@ -105,6 +105,13 @@ enum zone_stat_item {
>  #endif
>  	NR_VM_ZONE_STAT_ITEMS };
>  
> +enum lru_list {
> +	LRU_INACTIVE,	/* must match order of NR_[IN]ACTIVE */
> +	LRU_ACTIVE,	/*  "     "     "   "       "        */
> +	NR_LRU_LISTS };
> +
> +#define for_each_lru(l) for (l = 0; l < NR_LRU_LISTS; l++)
> +
>  struct per_cpu_pages {
>  	int count;		/* number of pages in the list */
>  	int high;		/* high watermark, emptying needed */
> @@ -258,10 +265,8 @@ struct zone {
>  
>  	/* Fields commonly accessed by the page reclaim scanner */
>  	spinlock_t		lru_lock;	
> -	struct list_head	active_list;
> -	struct list_head	inactive_list;
> -	unsigned long		nr_scan_active;
> -	unsigned long		nr_scan_inactive;
> +	struct list_head	list[NR_LRU_LISTS];
> +	unsigned long		nr_scan[NR_LRU_LISTS];
>  	unsigned long		pages_scanned;	   /* since last reclaim */
>  	unsigned long		flags;		   /* zone flags, see below */
>  
> Index: linux-2.6.25-rc2-mm1/include/linux/mm_inline.h
> ===================================================================
> --- linux-2.6.25-rc2-mm1.orig/include/linux/mm_inline.h	2007-07-08 19:32:17.000000000 -0400
> +++ linux-2.6.25-rc2-mm1/include/linux/mm_inline.h	2008-02-27 14:42:23.000000000 -0500
> @@ -1,40 +1,51 @@
>  static inline void
> +add_page_to_lru_list(struct zone *zone, struct page *page, enum lru_list l)
> +{
> +	list_add(&page->lru, &zone->list[l]);
> +	__inc_zone_state(zone, NR_INACTIVE + l);

... and by adding a NR_LRU_BASE this implicit assumption that NR_INACTIVE 
is 0 would be removed as we would do this:

	__inc_zone_state(zone, NR_LRU_BASE + l);

> +}
> +
> +static inline void
> +del_page_from_lru_list(struct zone *zone, struct page *page, enum lru_list l)
> +{
> +	list_del(&page->lru);
> +	__dec_zone_state(zone, NR_INACTIVE + l);

and here.

> +}
> +
> +static inline void
>  add_page_to_active_list(struct zone *zone, struct page *page)
>  {
> -	list_add(&page->lru, &zone->active_list);
> -	__inc_zone_state(zone, NR_ACTIVE);
> +	add_page_to_lru_list(zone, page, LRU_ACTIVE);
>  }
>  
>  static inline void
>  add_page_to_inactive_list(struct zone *zone, struct page *page)
>  {
> -	list_add(&page->lru, &zone->inactive_list);
> -	__inc_zone_state(zone, NR_INACTIVE);
> +	add_page_to_lru_list(zone, page, LRU_INACTIVE);
>  }
>  
>  static inline void
>  del_page_from_active_list(struct zone *zone, struct page *page)
>  {
> -	list_del(&page->lru);
> -	__dec_zone_state(zone, NR_ACTIVE);
> +	del_page_from_lru_list(zone, page, LRU_ACTIVE);
>  }
>  
>  static inline void
>  del_page_from_inactive_list(struct zone *zone, struct page *page)
>  {
> -	list_del(&page->lru);
> -	__dec_zone_state(zone, NR_INACTIVE);
> +	del_page_from_lru_list(zone, page, LRU_INACTIVE);
>  }
>  
>  static inline void
>  del_page_from_lru(struct zone *zone, struct page *page)
>  {
> +	enum lru_list l = LRU_INACTIVE;
> +
>  	list_del(&page->lru);
>  	if (PageActive(page)) {
>  		__ClearPageActive(page);
> -		__dec_zone_state(zone, NR_ACTIVE);
> -	} else {
> -		__dec_zone_state(zone, NR_INACTIVE);
> +		l = LRU_ACTIVE;
>  	}
> +	__dec_zone_state(zone, NR_INACTIVE + l);

and here.

>  }
>  
> Index: linux-2.6.25-rc2-mm1/mm/page_alloc.c
> ===================================================================
> --- linux-2.6.25-rc2-mm1.orig/mm/page_alloc.c	2008-02-19 16:23:16.000000000 -0500
> +++ linux-2.6.25-rc2-mm1/mm/page_alloc.c	2008-02-27 14:42:23.000000000 -0500
> @@ -3411,6 +3411,7 @@ static void __meminit free_area_init_cor
>  	for (j = 0; j < MAX_NR_ZONES; j++) {
>  		struct zone *zone = pgdat->node_zones + j;
>  		unsigned long size, realsize, memmap_pages;
> +		enum lru_list l;
>  
>  		size = zone_spanned_pages_in_node(nid, j, zones_size);
>  		realsize = size - zone_absent_pages_in_node(nid, j,
> @@ -3460,10 +3461,10 @@ static void __meminit free_area_init_cor
>  		zone->prev_priority = DEF_PRIORITY;
>  
>  		zone_pcp_init(zone);
> -		INIT_LIST_HEAD(&zone->active_list);
> -		INIT_LIST_HEAD(&zone->inactive_list);
> -		zone->nr_scan_active = 0;
> -		zone->nr_scan_inactive = 0;
> +		for_each_lru(l) {
> +			INIT_LIST_HEAD(&zone->list[l]);
> +			zone->nr_scan[l] = 0;
> +		}
>  		zap_zone_vm_stats(zone);
>  		zone->flags = 0;
>  		if (!size)
> Index: linux-2.6.25-rc2-mm1/mm/swap.c
> ===================================================================
> --- linux-2.6.25-rc2-mm1.orig/mm/swap.c	2008-02-19 16:23:09.000000000 -0500
> +++ linux-2.6.25-rc2-mm1/mm/swap.c	2008-02-27 14:42:23.000000000 -0500
> @@ -118,7 +118,7 @@ static void pagevec_move_tail(struct pag
>  			spin_lock(&zone->lru_lock);
>  		}
>  		if (PageLRU(page) && !PageActive(page)) {
> -			list_move_tail(&page->lru, &zone->inactive_list);
> +			list_move_tail(&page->lru, &zone->list[LRU_INACTIVE]);
>  			pgmoved++;
>  		}
>  	}
> Index: linux-2.6.25-rc2-mm1/mm/vmscan.c
> ===================================================================
> --- linux-2.6.25-rc2-mm1.orig/mm/vmscan.c	2008-02-25 17:10:54.000000000 -0500
> +++ linux-2.6.25-rc2-mm1/mm/vmscan.c	2008-02-27 14:42:23.000000000 -0500
> @@ -804,10 +804,10 @@ static unsigned long isolate_pages_globa
>  					int active)
>  {
>  	if (active)
> -		return isolate_lru_pages(nr, &z->active_list, dst,
> +		return isolate_lru_pages(nr, &z->list[LRU_ACTIVE], dst,
>  						scanned, order, mode);
>  	else
> -		return isolate_lru_pages(nr, &z->inactive_list, dst,
> +		return isolate_lru_pages(nr, &z->list[LRU_INACTIVE], dst,
>  						scanned, order, mode);
>  }
>  
> @@ -954,10 +954,7 @@ static unsigned long shrink_inactive_lis
>  			VM_BUG_ON(PageLRU(page));
>  			SetPageLRU(page);
>  			list_del(&page->lru);
> -			if (PageActive(page))
> -				add_page_to_active_list(zone, page);
> -			else
> -				add_page_to_inactive_list(zone, page);
> +			add_page_to_lru_list(zone, page, PageActive(page));
>  			if (!pagevec_add(&pvec, page)) {
>  				spin_unlock_irq(&zone->lru_lock);
>  				__pagevec_release(&pvec);
> @@ -1125,11 +1122,14 @@ static void shrink_active_list(unsigned 
>  	int pgdeactivate = 0;
>  	unsigned long pgscanned;
>  	LIST_HEAD(l_hold);	/* The pages which were snipped off */
> -	LIST_HEAD(l_inactive);	/* Pages to go onto the inactive_list */
> -	LIST_HEAD(l_active);	/* Pages to go onto the active_list */
> +	struct list_head list[NR_LRU_LISTS];
>  	struct page *page;
>  	struct pagevec pvec;
>  	int reclaim_mapped = 0;
> +	enum lru_list l;
> +
> +	for_each_lru(l)
> +		INIT_LIST_HEAD(&list[l]);
>  
>  	if (sc->may_swap)
>  		reclaim_mapped = calc_reclaim_mapped(sc, zone, priority);
> @@ -1157,28 +1157,28 @@ static void shrink_active_list(unsigned 
>  			if (!reclaim_mapped ||
>  			    (total_swap_pages == 0 && PageAnon(page)) ||
>  			    page_referenced(page, 0, sc->mem_cgroup)) {
> -				list_add(&page->lru, &l_active);
> +				list_add(&page->lru, &list[LRU_ACTIVE]);
>  				continue;
>  			}
>  		} else if (TestClearPageReferenced(page)) {
> -			list_add(&page->lru, &l_active);
> +			list_add(&page->lru, &list[LRU_ACTIVE]);
>  			continue;
>  		}
> -		list_add(&page->lru, &l_inactive);
> +		list_add(&page->lru, &list[LRU_INACTIVE]);
>  	}
>  
>  	pagevec_init(&pvec, 1);
>  	pgmoved = 0;
>  	spin_lock_irq(&zone->lru_lock);
> -	while (!list_empty(&l_inactive)) {
> -		page = lru_to_page(&l_inactive);
> -		prefetchw_prev_lru_page(page, &l_inactive, flags);
> +	while (!list_empty(&list[LRU_INACTIVE])) {
> +		page = lru_to_page(&list[LRU_INACTIVE]);
> +		prefetchw_prev_lru_page(page, &list[LRU_INACTIVE], flags);
>  		VM_BUG_ON(PageLRU(page));
>  		SetPageLRU(page);
>  		VM_BUG_ON(!PageActive(page));
>  		ClearPageActive(page);
>  
> -		list_move(&page->lru, &zone->inactive_list);
> +		list_move(&page->lru, &zone->list[LRU_INACTIVE]);
>  		mem_cgroup_move_lists(page_get_page_cgroup(page), false);
>  		pgmoved++;
>  		if (!pagevec_add(&pvec, page)) {
> @@ -1201,13 +1201,13 @@ static void shrink_active_list(unsigned 
>  	}
>  
>  	pgmoved = 0;
> -	while (!list_empty(&l_active)) {
> -		page = lru_to_page(&l_active);
> -		prefetchw_prev_lru_page(page, &l_active, flags);
> +	while (!list_empty(&list[LRU_ACTIVE])) {
> +		page = lru_to_page(&list[LRU_ACTIVE]);
> +		prefetchw_prev_lru_page(page, &list[LRU_ACTIVE], flags);
>  		VM_BUG_ON(PageLRU(page));
>  		SetPageLRU(page);
>  		VM_BUG_ON(!PageActive(page));
> -		list_move(&page->lru, &zone->active_list);
> +		list_move(&page->lru, &zone->list[LRU_ACTIVE]);
>  		mem_cgroup_move_lists(page_get_page_cgroup(page), true);
>  		pgmoved++;
>  		if (!pagevec_add(&pvec, page)) {
> @@ -1227,65 +1227,64 @@ static void shrink_active_list(unsigned 
>  	pagevec_release(&pvec);
>  }
>  
> +static unsigned long shrink_list(enum lru_list l, unsigned long nr_to_scan,
> +	struct zone *zone, struct scan_control *sc, int priority)
> +{
> +	if (l == LRU_ACTIVE) {
> +		shrink_active_list(nr_to_scan, zone, sc, priority);
> +		return 0;
> +	}
> +	return shrink_inactive_list(nr_to_scan, zone, sc);
> +}
> +
>  /*
>   * This is a basic per-zone page freer.  Used by both kswapd and direct reclaim.
>   */
>  static unsigned long shrink_zone(int priority, struct zone *zone,
>  				struct scan_control *sc)
>  {
> -	unsigned long nr_active;
> -	unsigned long nr_inactive;
> +	unsigned long nr[NR_LRU_LISTS];
>  	unsigned long nr_to_scan;
>  	unsigned long nr_reclaimed = 0;
> +	enum lru_list l;
>  
>  	if (scan_global_lru(sc)) {
>  		/*
>  		 * Add one to nr_to_scan just to make sure that the kernel
>  		 * will slowly sift through the active list.
>  		 */
> -		zone->nr_scan_active +=
> -			(zone_page_state(zone, NR_ACTIVE) >> priority) + 1;
> -		nr_active = zone->nr_scan_active;
> -		zone->nr_scan_inactive +=
> -			(zone_page_state(zone, NR_INACTIVE) >> priority) + 1;
> -		nr_inactive = zone->nr_scan_inactive;
> -		if (nr_inactive >= sc->swap_cluster_max)
> -			zone->nr_scan_inactive = 0;
> -		else
> -			nr_inactive = 0;
> -
> -		if (nr_active >= sc->swap_cluster_max)
> -			zone->nr_scan_active = 0;
> -		else
> -			nr_active = 0;
> +		for_each_lru(l) {
> +			zone->nr_scan[l] += (zone_page_state(zone,
> +					NR_INACTIVE + l)  >> priority) + 1;

and here.

> +			nr[l] = zone->nr_scan[l];
> +			if (nr[l] >= sc->swap_cluster_max)
> +				zone->nr_scan[l] = 0;
> +			else
> +				nr[l] = 0;
> +		}
>  	} else {
>  		/*
>  		 * This reclaim occurs not because zone memory shortage but
>  		 * because memory controller hits its limit.
>  		 * Then, don't modify zone reclaim related data.
>  		 */
> -		nr_active = mem_cgroup_calc_reclaim_active(sc->mem_cgroup,
> -					zone, priority);
> +		nr[LRU_ACTIVE] = mem_cgroup_calc_reclaim(sc->mem_cgroup,
> +					zone, priority, LRU_ACTIVE);
>  
> -		nr_inactive = mem_cgroup_calc_reclaim_inactive(sc->mem_cgroup,
> -					zone, priority);
> +		nr[LRU_INACTIVE] = mem_cgroup_calc_reclaim(sc->mem_cgroup,
> +					zone, priority, LRU_INACTIVE);
>  	}
>  
> -
> -	while (nr_active || nr_inactive) {
> -		if (nr_active) {
> -			nr_to_scan = min(nr_active,
> +	while (nr[LRU_ACTIVE] || nr[LRU_INACTIVE]) {
> +		for_each_lru(l) {
> +			if (nr[l]) {
> +				nr_to_scan = min(nr[l],
>  					(unsigned long)sc->swap_cluster_max);
> -			nr_active -= nr_to_scan;
> -			shrink_active_list(nr_to_scan, zone, sc, priority);
> -		}
> +				nr[l] -= nr_to_scan;
>  
> -		if (nr_inactive) {
> -			nr_to_scan = min(nr_inactive,
> -					(unsigned long)sc->swap_cluster_max);
> -			nr_inactive -= nr_to_scan;
> -			nr_reclaimed += shrink_inactive_list(nr_to_scan, zone,
> -								sc);
> +				nr_reclaimed += shrink_list(l, nr_to_scan,
> +							zone, sc, priority);
> +			}
>  		}
>  	}
>  
> @@ -1802,6 +1801,7 @@ static unsigned long shrink_all_zones(un
>  {
>  	struct zone *zone;
>  	unsigned long nr_to_scan, ret = 0;
> +	enum lru_list l;
>  
>  	for_each_zone(zone) {
>  
> @@ -1811,28 +1811,25 @@ static unsigned long shrink_all_zones(un
>  		if (zone_is_all_unreclaimable(zone) && prio != DEF_PRIORITY)
>  			continue;
>  
> -		/* For pass = 0 we don't shrink the active list */
> -		if (pass > 0) {
> -			zone->nr_scan_active +=
> -				(zone_page_state(zone, NR_ACTIVE) >> prio) + 1;
> -			if (zone->nr_scan_active >= nr_pages || pass > 3) {
> -				zone->nr_scan_active = 0;
> +		for_each_lru(l) {
> +			/* For pass = 0 we don't shrink the active list */
> +			if (pass == 0 && l == LRU_ACTIVE)
> +				continue;
> +
> +			zone->nr_scan[l] +=
> +				(zone_page_state(zone, NR_INACTIVE + l)
> +								>> prio) + 1;
> +			if (zone->nr_scan[l] >= nr_pages || pass > 3) {
> +				zone->nr_scan[l] = 0;
>  				nr_to_scan = min(nr_pages,
> -					zone_page_state(zone, NR_ACTIVE));
> -				shrink_active_list(nr_to_scan, zone, sc, prio);
> +					zone_page_state(zone,
> +							NR_INACTIVE + l));
> +				ret += shrink_list(l, nr_to_scan, zone,
> +								sc, prio);
> +				if (ret >= nr_pages)
> +					return ret;
>  			}
>  		}
> -
> -		zone->nr_scan_inactive +=
> -			(zone_page_state(zone, NR_INACTIVE) >> prio) + 1;
> -		if (zone->nr_scan_inactive >= nr_pages || pass > 3) {
> -			zone->nr_scan_inactive = 0;
> -			nr_to_scan = min(nr_pages,
> -				zone_page_state(zone, NR_INACTIVE));
> -			ret += shrink_inactive_list(nr_to_scan, zone, sc);
> -			if (ret >= nr_pages)
> -				return ret;
> -		}
>  	}
>  
>  	return ret;
> Index: linux-2.6.25-rc2-mm1/mm/vmstat.c
> ===================================================================
> --- linux-2.6.25-rc2-mm1.orig/mm/vmstat.c	2008-02-19 16:23:16.000000000 -0500
> +++ linux-2.6.25-rc2-mm1/mm/vmstat.c	2008-02-26 21:24:23.000000000 -0500
> @@ -758,7 +758,8 @@ static void zoneinfo_show_print(struct s
>  		   zone->pages_low,
>  		   zone->pages_high,
>  		   zone->pages_scanned,
> -		   zone->nr_scan_active, zone->nr_scan_inactive,
> +		   zone->nr_scan[LRU_ACTIVE],
> +		   zone->nr_scan[LRU_INACTIVE],
>  		   zone->spanned_pages,
>  		   zone->present_pages);
>  
> Index: linux-2.6.25-rc2-mm1/include/linux/memcontrol.h
> ===================================================================
> --- linux-2.6.25-rc2-mm1.orig/include/linux/memcontrol.h	2008-02-19 16:23:16.000000000 -0500
> +++ linux-2.6.25-rc2-mm1/include/linux/memcontrol.h	2008-02-26 21:33:14.000000000 -0500
> @@ -67,10 +67,8 @@ extern void mem_cgroup_note_reclaim_prio
>  extern void mem_cgroup_record_reclaim_priority(struct mem_cgroup *mem,
>  							int priority);
>  
> -extern long mem_cgroup_calc_reclaim_active(struct mem_cgroup *mem,
> -				struct zone *zone, int priority);
> -extern long mem_cgroup_calc_reclaim_inactive(struct mem_cgroup *mem,
> -				struct zone *zone, int priority);
> +extern long mem_cgroup_calc_reclaim(struct mem_cgroup *mem, struct zone *zone,
> +					int priority, enum lru_list lru);
>  
>  #else /* CONFIG_CGROUP_MEM_CONT */
>  static inline void mm_init_cgroup(struct mm_struct *mm,
> @@ -168,14 +166,9 @@ static inline void mem_cgroup_record_rec
>  {
>  }
>  
> -static inline long mem_cgroup_calc_reclaim_active(struct mem_cgroup *mem,
> -					struct zone *zone, int priority)
> -{
> -	return 0;
> -}
> -
> -static inline long mem_cgroup_calc_reclaim_inactive(struct mem_cgroup *mem,
> -					struct zone *zone, int priority)
> +static inline long mem_cgroup_calc_reclaim(struct mem_cgroup *mem,
> +					struct zone *zone, int priority,
> +					enum lru_list lru)
>  {
>  	return 0;
>  }
> Index: linux-2.6.25-rc2-mm1/mm/memcontrol.c
> ===================================================================
> --- linux-2.6.25-rc2-mm1.orig/mm/memcontrol.c	2008-02-19 16:23:09.000000000 -0500
> +++ linux-2.6.25-rc2-mm1/mm/memcontrol.c	2008-02-27 14:43:02.000000000 -0500
> @@ -30,6 +30,7 @@
>  #include <linux/spinlock.h>
>  #include <linux/fs.h>
>  #include <linux/seq_file.h>
> +#include <linux/mm_inline.h>
>  
>  #include <asm/uaccess.h>
>  
> @@ -80,22 +81,13 @@ static s64 mem_cgroup_read_stat(struct m
>  /*
>   * per-zone information in memory controller.
>   */
> -
> -enum mem_cgroup_zstat_index {
> -	MEM_CGROUP_ZSTAT_ACTIVE,
> -	MEM_CGROUP_ZSTAT_INACTIVE,
> -
> -	NR_MEM_CGROUP_ZSTAT,
> -};
> -
>  struct mem_cgroup_per_zone {
>  	/*
>  	 * spin_lock to protect the per cgroup LRU
>  	 */
>  	spinlock_t		lru_lock;
> -	struct list_head	active_list;
> -	struct list_head	inactive_list;
> -	unsigned long count[NR_MEM_CGROUP_ZSTAT];
> +	struct list_head	lists[NR_LRU_LISTS];
> +	unsigned long		count[NR_LRU_LISTS];
>  };
>  /* Macro for accessing counter */
>  #define MEM_CGROUP_ZSTAT(mz, idx)	((mz)->count[(idx)])
> @@ -220,7 +212,7 @@ page_cgroup_zoneinfo(struct page_cgroup 
>  }
>  
>  static unsigned long mem_cgroup_get_all_zonestat(struct mem_cgroup *mem,
> -					enum mem_cgroup_zstat_index idx)
> +					enum lru_list idx)
>  {
>  	int nid, zid;
>  	struct mem_cgroup_per_zone *mz;
> @@ -346,13 +338,13 @@ static struct page_cgroup *clear_page_cg
>  
>  static void __mem_cgroup_remove_list(struct page_cgroup *pc)
>  {
> -	int from = pc->flags & PAGE_CGROUP_FLAG_ACTIVE;
>  	struct mem_cgroup_per_zone *mz = page_cgroup_zoneinfo(pc);
> +	int lru = LRU_INACTIVE;
>  
> -	if (from)
> -		MEM_CGROUP_ZSTAT(mz, MEM_CGROUP_ZSTAT_ACTIVE) -= 1;
> -	else
> -		MEM_CGROUP_ZSTAT(mz, MEM_CGROUP_ZSTAT_INACTIVE) -= 1;
> +	if (pc->flags & PAGE_CGROUP_FLAG_ACTIVE)
> +		lru += LRU_ACTIVE;

This is making the assumption that LRU_INACTIVE is 0, this should really be:

		lru += LRU_ACTIVE - LRU_INACTIVE;

> +
> +	MEM_CGROUP_ZSTAT(mz, lru) -= 1;
>  
>  	mem_cgroup_charge_statistics(pc->mem_cgroup, pc->flags, false);
>  	list_del_init(&pc->lru);
> @@ -360,16 +352,15 @@ static void __mem_cgroup_remove_list(str
>  
>  static void __mem_cgroup_add_list(struct page_cgroup *pc)
>  {
> -	int to = pc->flags & PAGE_CGROUP_FLAG_ACTIVE;
>  	struct mem_cgroup_per_zone *mz = page_cgroup_zoneinfo(pc);
> +	int lru = LRU_INACTIVE;
> +
> +	if (pc->flags & PAGE_CGROUP_FLAG_ACTIVE)
> +		lru += LRU_ACTIVE;

and here.

> +
> +	MEM_CGROUP_ZSTAT(mz, lru) += 1;
> +	list_add(&pc->lru, &mz->lists[lru]);
>  
> -	if (!to) {
> -		MEM_CGROUP_ZSTAT(mz, MEM_CGROUP_ZSTAT_INACTIVE) += 1;
> -		list_add(&pc->lru, &mz->inactive_list);
> -	} else {
> -		MEM_CGROUP_ZSTAT(mz, MEM_CGROUP_ZSTAT_ACTIVE) += 1;
> -		list_add(&pc->lru, &mz->active_list);
> -	}
>  	mem_cgroup_charge_statistics(pc->mem_cgroup, pc->flags, true);
>  }
>  
> @@ -377,21 +368,18 @@ static void __mem_cgroup_move_lists(stru
>  {
>  	int from = pc->flags & PAGE_CGROUP_FLAG_ACTIVE;
>  	struct mem_cgroup_per_zone *mz = page_cgroup_zoneinfo(pc);
> +	int lru = !!from;

This really is meant to mean:

	lru = (from)? LRU_ACTIVE : LRU_INACTIVE

and if those change it will break completely.  We should either spell it
out explicitly, or put some kind of build failure in place if they are
not those values.  If we want something without branches perhaps:

	int lru_list = { LRU_INACTIVE, LRU_ACTIVE };
	int lru = lru_list[!!from];

> -	if (from)
> -		MEM_CGROUP_ZSTAT(mz, MEM_CGROUP_ZSTAT_ACTIVE) -= 1;
> -	else
> -		MEM_CGROUP_ZSTAT(mz, MEM_CGROUP_ZSTAT_INACTIVE) -= 1;
> +	MEM_CGROUP_ZSTAT(mz, lru) -= 1;
>  
> -	if (active) {
> -		MEM_CGROUP_ZSTAT(mz, MEM_CGROUP_ZSTAT_ACTIVE) += 1;
> +	if (active)
>  		pc->flags |= PAGE_CGROUP_FLAG_ACTIVE;
> -		list_move(&pc->lru, &mz->active_list);
> -	} else {
> -		MEM_CGROUP_ZSTAT(mz, MEM_CGROUP_ZSTAT_INACTIVE) += 1;
> +	else
>  		pc->flags &= ~PAGE_CGROUP_FLAG_ACTIVE;
> -		list_move(&pc->lru, &mz->inactive_list);
> -	}
> +
> +	lru = !!active;
> +	MEM_CGROUP_ZSTAT(mz, lru) += 1;
> +	list_move(&pc->lru, &mz->lists[lru]);
>  }
>  
>  int task_in_mem_cgroup(struct task_struct *task, const struct mem_cgroup *mem)
> @@ -447,8 +435,8 @@ long mem_cgroup_reclaim_imbalance(struct
>  {
>  	unsigned long active, inactive;
>  	/* active and inactive are the number of pages. 'long' is ok.*/
> -	active = mem_cgroup_get_all_zonestat(mem, MEM_CGROUP_ZSTAT_ACTIVE);
> -	inactive = mem_cgroup_get_all_zonestat(mem, MEM_CGROUP_ZSTAT_INACTIVE);
> +	active = mem_cgroup_get_all_zonestat(mem, LRU_ACTIVE);
> +	inactive = mem_cgroup_get_all_zonestat(mem, LRU_INACTIVE);
>  	return (long) (active / (inactive + 1));
>  }
>  
> @@ -479,29 +467,17 @@ void mem_cgroup_record_reclaim_priority(
>   * (see include/linux/mmzone.h)
>   */
>  
> -long mem_cgroup_calc_reclaim_active(struct mem_cgroup *mem,
> -				   struct zone *zone, int priority)
> +long mem_cgroup_calc_reclaim(struct mem_cgroup *mem, struct zone *zone,
> +					int priority, enum lru_list lru)
>  {
> -	long nr_active;
> +	long nr_pages;
>  	int nid = zone->zone_pgdat->node_id;
>  	int zid = zone_idx(zone);
>  	struct mem_cgroup_per_zone *mz = mem_cgroup_zoneinfo(mem, nid, zid);
>  
> -	nr_active = MEM_CGROUP_ZSTAT(mz, MEM_CGROUP_ZSTAT_ACTIVE);
> -	return (nr_active >> priority);
> -}
> -
> -long mem_cgroup_calc_reclaim_inactive(struct mem_cgroup *mem,
> -					struct zone *zone, int priority)
> -{
> -	long nr_inactive;
> -	int nid = zone->zone_pgdat->node_id;
> -	int zid = zone_idx(zone);
> -	struct mem_cgroup_per_zone *mz = mem_cgroup_zoneinfo(mem, nid, zid);
> +	nr_pages = MEM_CGROUP_ZSTAT(mz, lru);
>  
> -	nr_inactive = MEM_CGROUP_ZSTAT(mz, MEM_CGROUP_ZSTAT_INACTIVE);
> -
> -	return (nr_inactive >> priority);
> +	return (nr_pages >> priority);
>  }
>  
>  unsigned long mem_cgroup_isolate_pages(unsigned long nr_to_scan,
> @@ -520,13 +496,10 @@ unsigned long mem_cgroup_isolate_pages(u
>  	int nid = z->zone_pgdat->node_id;
>  	int zid = zone_idx(z);
>  	struct mem_cgroup_per_zone *mz;
> +	int lru = !!active;

same issue here

>  
>  	mz = mem_cgroup_zoneinfo(mem_cont, nid, zid);
> -	if (active)
> -		src = &mz->active_list;
> -	else
> -		src = &mz->inactive_list;
> -
> +	src = &mz->lists[lru];
>  
>  	spin_lock(&mz->lru_lock);
>  	scan = 0;
> @@ -833,7 +806,7 @@ retry:
>  static void
>  mem_cgroup_force_empty_list(struct mem_cgroup *mem,
>  			    struct mem_cgroup_per_zone *mz,
> -			    int active)
> +			    enum lru_list lru)
>  {
>  	struct page_cgroup *pc;
>  	struct page *page;
> @@ -841,10 +814,7 @@ mem_cgroup_force_empty_list(struct mem_c
>  	unsigned long flags;
>  	struct list_head *list;
>  
> -	if (active)
> -		list = &mz->active_list;
> -	else
> -		list = &mz->inactive_list;
> +	list = &mz->lists[lru];
>  
>  	if (list_empty(list))
>  		return;
> @@ -894,11 +864,10 @@ int mem_cgroup_force_empty(struct mem_cg
>  		for_each_node_state(node, N_POSSIBLE)
>  			for (zid = 0; zid < MAX_NR_ZONES; zid++) {
>  				struct mem_cgroup_per_zone *mz;
> +				enum lru_list l;
>  				mz = mem_cgroup_zoneinfo(mem, node, zid);
> -				/* drop all page_cgroup in active_list */
> -				mem_cgroup_force_empty_list(mem, mz, 1);
> -				/* drop all page_cgroup in inactive_list */
> -				mem_cgroup_force_empty_list(mem, mz, 0);
> +				for_each_lru(l)
> +					mem_cgroup_force_empty_list(mem, mz, l);
>  			}
>  	}
>  	ret = 0;
> @@ -994,9 +963,9 @@ static int mem_control_stat_show(struct 
>  		unsigned long active, inactive;
>  
>  		inactive = mem_cgroup_get_all_zonestat(mem_cont,
> -						MEM_CGROUP_ZSTAT_INACTIVE);
> +						LRU_INACTIVE);
>  		active = mem_cgroup_get_all_zonestat(mem_cont,
> -						MEM_CGROUP_ZSTAT_ACTIVE);
> +						LRU_ACTIVE);
>  		seq_printf(m, "active %ld\n", (active) * PAGE_SIZE);
>  		seq_printf(m, "inactive %ld\n", (inactive) * PAGE_SIZE);
>  	}
> @@ -1052,6 +1021,7 @@ static int alloc_mem_cgroup_per_zone_inf
>  {
>  	struct mem_cgroup_per_node *pn;
>  	struct mem_cgroup_per_zone *mz;
> +	enum lru_list l;
>  	int zone;
>  	/*
>  	 * This routine is called against possible nodes.
> @@ -1073,9 +1043,9 @@ static int alloc_mem_cgroup_per_zone_inf
>  
>  	for (zone = 0; zone < MAX_NR_ZONES; zone++) {
>  		mz = &pn->zoneinfo[zone];
> -		INIT_LIST_HEAD(&mz->active_list);
> -		INIT_LIST_HEAD(&mz->inactive_list);
>  		spin_lock_init(&mz->lru_lock);
> +		for_each_lru(l)
> +			INIT_LIST_HEAD(&mz->lists[l]);
>  	}
>  	return 0;
>  }

-apw
--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ