lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Date:   Thu, 22 Jun 2017 08:44:54 +0200
From:   Michal Hocko <mhocko@...nel.org>
To:     Nikolay Borisov <nborisov@...e.com>
Cc:     linux-mm@...ck.org, mgorman@...hsingularity.net,
        cmetcalf@...lanox.com, minchan@...nel.org, vbabka@...e.cz,
        kirill.shutemov@...ux.intel.com, tj@...nel.org,
        cgroups@...r.kernel.org, linux-kernel@...r.kernel.org
Subject: Re: [PATCH] mm: Refactor conversion of pages to bytes macro
 definitions

On Tue 20-06-17 18:14:28, Nikolay Borisov wrote:
> Currently there are a multiple files with the following code:
>  #define K(x) ((x) << (PAGE_SHIFT - 10))
>  ... some code..
>  #undef K
> 
> This is mainly used to print out some memory-related statistics, where X is
> given in pages and the macro just converts it to kilobytes. In the future
> there is going to be more macros since there are intention to introduce
> byte-based memory counters [1]. This could lead to proliferation of
> multiple duplicated definition of various macros used to convert a quantity
> from one unit to another. Let's try and consolidate such definition in the
> mm.h header since currently it's being included in all files which exhibit
> this pattern. Also let's rename it to something a bit more verbose.
> 
> This patch doesn't introduce any functional changes
> 
> [1] https://patchwork.kernel.org/patch/9395205/
> 
> Signed-off-by: Nikolay Borisov <nborisov@...e.com>
> ---
>  arch/tile/mm/pgtable.c      |  2 --
>  drivers/base/node.c         | 66 ++++++++++++++++++-------------------
>  include/linux/mm.h          |  2 ++
>  kernel/debug/kdb/kdb_main.c |  3 +-
>  mm/backing-dev.c            | 22 +++++--------
>  mm/memcontrol.c             | 17 +++++-----
>  mm/oom_kill.c               | 19 +++++------
>  mm/page_alloc.c             | 80 ++++++++++++++++++++++-----------------------
>  8 files changed, 100 insertions(+), 111 deletions(-)

Those macros are quite trivial and we do not really save much code while
this touches a lot of code potentially causing some conflicts. So do we
really need this? I am usually very keen on removing duplication but
this doesn't seem to be worth all the troubles IMHO.

> 
> diff --git a/arch/tile/mm/pgtable.c b/arch/tile/mm/pgtable.c
> index 492a7361e58e..f04af570c1c2 100644
> --- a/arch/tile/mm/pgtable.c
> +++ b/arch/tile/mm/pgtable.c
> @@ -34,8 +34,6 @@
>  #include <asm/tlbflush.h>
>  #include <asm/homecache.h>
>  
> -#define K(x) ((x) << (PAGE_SHIFT-10))
> -
>  /**
>   * shatter_huge_page() - ensure a given address is mapped by a small page.
>   *
> diff --git a/drivers/base/node.c b/drivers/base/node.c
> index 5548f9686016..b6f563a3a3a9 100644
> --- a/drivers/base/node.c
> +++ b/drivers/base/node.c
> @@ -50,7 +50,6 @@ static inline ssize_t node_read_cpulist(struct device *dev,
>  static DEVICE_ATTR(cpumap,  S_IRUGO, node_read_cpumask, NULL);
>  static DEVICE_ATTR(cpulist, S_IRUGO, node_read_cpulist, NULL);
>  
> -#define K(x) ((x) << (PAGE_SHIFT - 10))
>  static ssize_t node_read_meminfo(struct device *dev,
>  			struct device_attribute *attr, char *buf)
>  {
> @@ -72,19 +71,19 @@ static ssize_t node_read_meminfo(struct device *dev,
>  		       "Node %d Inactive(file): %8lu kB\n"
>  		       "Node %d Unevictable:    %8lu kB\n"
>  		       "Node %d Mlocked:        %8lu kB\n",
> -		       nid, K(i.totalram),
> -		       nid, K(i.freeram),
> -		       nid, K(i.totalram - i.freeram),
> -		       nid, K(node_page_state(pgdat, NR_ACTIVE_ANON) +
> +		       nid, PtoK(i.totalram),
> +		       nid, PtoK(i.freeram),
> +		       nid, PtoK(i.totalram - i.freeram),
> +		       nid, PtoK(node_page_state(pgdat, NR_ACTIVE_ANON) +
>  				node_page_state(pgdat, NR_ACTIVE_FILE)),
> -		       nid, K(node_page_state(pgdat, NR_INACTIVE_ANON) +
> +		       nid, PtoK(node_page_state(pgdat, NR_INACTIVE_ANON) +
>  				node_page_state(pgdat, NR_INACTIVE_FILE)),
> -		       nid, K(node_page_state(pgdat, NR_ACTIVE_ANON)),
> -		       nid, K(node_page_state(pgdat, NR_INACTIVE_ANON)),
> -		       nid, K(node_page_state(pgdat, NR_ACTIVE_FILE)),
> -		       nid, K(node_page_state(pgdat, NR_INACTIVE_FILE)),
> -		       nid, K(node_page_state(pgdat, NR_UNEVICTABLE)),
> -		       nid, K(sum_zone_node_page_state(nid, NR_MLOCK)));
> +		       nid, PtoK(node_page_state(pgdat, NR_ACTIVE_ANON)),
> +		       nid, PtoK(node_page_state(pgdat, NR_INACTIVE_ANON)),
> +		       nid, PtoK(node_page_state(pgdat, NR_ACTIVE_FILE)),
> +		       nid, PtoK(node_page_state(pgdat, NR_INACTIVE_FILE)),
> +		       nid, PtoK(node_page_state(pgdat, NR_UNEVICTABLE)),
> +		       nid, PtoK(sum_zone_node_page_state(nid, NR_MLOCK)));
>  
>  #ifdef CONFIG_HIGHMEM
>  	n += sprintf(buf + n,
> @@ -92,10 +91,10 @@ static ssize_t node_read_meminfo(struct device *dev,
>  		       "Node %d HighFree:       %8lu kB\n"
>  		       "Node %d LowTotal:       %8lu kB\n"
>  		       "Node %d LowFree:        %8lu kB\n",
> -		       nid, K(i.totalhigh),
> -		       nid, K(i.freehigh),
> -		       nid, K(i.totalram - i.totalhigh),
> -		       nid, K(i.freeram - i.freehigh));
> +		       nid, PtoK(i.totalhigh),
> +		       nid, PtoK(i.freehigh),
> +		       nid, PtoK(i.totalram - i.totalhigh),
> +		       nid, PtoK(i.freeram - i.freehigh));
>  #endif
>  	n += sprintf(buf + n,
>  		       "Node %d Dirty:          %8lu kB\n"
> @@ -118,36 +117,35 @@ static ssize_t node_read_meminfo(struct device *dev,
>  		       "Node %d ShmemPmdMapped: %8lu kB\n"
>  #endif
>  			,
> -		       nid, K(node_page_state(pgdat, NR_FILE_DIRTY)),
> -		       nid, K(node_page_state(pgdat, NR_WRITEBACK)),
> -		       nid, K(node_page_state(pgdat, NR_FILE_PAGES)),
> -		       nid, K(node_page_state(pgdat, NR_FILE_MAPPED)),
> -		       nid, K(node_page_state(pgdat, NR_ANON_MAPPED)),
> -		       nid, K(i.sharedram),
> +		       nid, PtoK(node_page_state(pgdat, NR_FILE_DIRTY)),
> +		       nid, PtoK(node_page_state(pgdat, NR_WRITEBACK)),
> +		       nid, PtoK(node_page_state(pgdat, NR_FILE_PAGES)),
> +		       nid, PtoK(node_page_state(pgdat, NR_FILE_MAPPED)),
> +		       nid, PtoK(node_page_state(pgdat, NR_ANON_MAPPED)),
> +		       nid, PtoK(i.sharedram),
>  		       nid, sum_zone_node_page_state(nid, NR_KERNEL_STACK_KB),
> -		       nid, K(sum_zone_node_page_state(nid, NR_PAGETABLE)),
> -		       nid, K(node_page_state(pgdat, NR_UNSTABLE_NFS)),
> -		       nid, K(sum_zone_node_page_state(nid, NR_BOUNCE)),
> -		       nid, K(node_page_state(pgdat, NR_WRITEBACK_TEMP)),
> -		       nid, K(sum_zone_node_page_state(nid, NR_SLAB_RECLAIMABLE) +
> +		       nid, PtoK(sum_zone_node_page_state(nid, NR_PAGETABLE)),
> +		       nid, PtoK(node_page_state(pgdat, NR_UNSTABLE_NFS)),
> +		       nid, PtoK(sum_zone_node_page_state(nid, NR_BOUNCE)),
> +		       nid, PtoK(node_page_state(pgdat, NR_WRITEBACK_TEMP)),
> +		       nid, PtoK(sum_zone_node_page_state(nid, NR_SLAB_RECLAIMABLE) +
>  				sum_zone_node_page_state(nid, NR_SLAB_UNRECLAIMABLE)),
> -		       nid, K(sum_zone_node_page_state(nid, NR_SLAB_RECLAIMABLE)),
> +		       nid, PtoK(sum_zone_node_page_state(nid, NR_SLAB_RECLAIMABLE)),
>  #ifdef CONFIG_TRANSPARENT_HUGEPAGE
> -		       nid, K(sum_zone_node_page_state(nid, NR_SLAB_UNRECLAIMABLE)),
> -		       nid, K(node_page_state(pgdat, NR_ANON_THPS) *
> +		       nid, PtoK(sum_zone_node_page_state(nid, NR_SLAB_UNRECLAIMABLE)),
> +		       nid, PtoK(node_page_state(pgdat, NR_ANON_THPS) *
>  				       HPAGE_PMD_NR),
> -		       nid, K(node_page_state(pgdat, NR_SHMEM_THPS) *
> +		       nid, PtoK(node_page_state(pgdat, NR_SHMEM_THPS) *
>  				       HPAGE_PMD_NR),
> -		       nid, K(node_page_state(pgdat, NR_SHMEM_PMDMAPPED) *
> +		       nid, PtoK(node_page_state(pgdat, NR_SHMEM_PMDMAPPED) *
>  				       HPAGE_PMD_NR));
>  #else
> -		       nid, K(sum_zone_node_page_state(nid, NR_SLAB_UNRECLAIMABLE)));
> +		       nid, PtoK(sum_zone_node_page_state(nid, NR_SLAB_UNRECLAIMABLE)));
>  #endif
>  	n += hugetlb_report_node_meminfo(nid, buf + n);
>  	return n;
>  }
>  
> -#undef K
>  static DEVICE_ATTR(meminfo, S_IRUGO, node_read_meminfo, NULL);
>  
>  static ssize_t node_read_numastat(struct device *dev,
> diff --git a/include/linux/mm.h b/include/linux/mm.h
> index 6f543a47fc92..d8d80e2e9194 100644
> --- a/include/linux/mm.h
> +++ b/include/linux/mm.h
> @@ -93,6 +93,8 @@ extern int mmap_rnd_compat_bits __read_mostly;
>  #define mm_forbids_zeropage(X)	(0)
>  #endif
>  
> +#define PtoK(pages) ((pages) << (PAGE_SHIFT-10))
> +
>  /*
>   * Default maximum number of active map areas, this limits the number of vmas
>   * per mm struct. Users can overwrite this number by sysctl but there is a
> diff --git a/kernel/debug/kdb/kdb_main.c b/kernel/debug/kdb/kdb_main.c
> index c8146d53ca67..e833cb02d2c8 100644
> --- a/kernel/debug/kdb/kdb_main.c
> +++ b/kernel/debug/kdb/kdb_main.c
> @@ -2582,10 +2582,9 @@ static int kdb_summary(int argc, const char **argv)
>  #undef LOAD_INT
>  #undef LOAD_FRAC
>  	/* Display in kilobytes */
> -#define K(x) ((x) << (PAGE_SHIFT - 10))
>  	kdb_printf("\nMemTotal:       %8lu kB\nMemFree:        %8lu kB\n"
>  		   "Buffers:        %8lu kB\n",
> -		   K(val.totalram), K(val.freeram), K(val.bufferram));
> +		   PtoK(val.totalram), PtoK(val.freeram), PtoK(val.bufferram));
>  	return 0;
>  }
>  
> diff --git a/mm/backing-dev.c b/mm/backing-dev.c
> index f028a9a472fd..0c09dd103109 100644
> --- a/mm/backing-dev.c
> +++ b/mm/backing-dev.c
> @@ -67,7 +67,6 @@ static int bdi_debug_stats_show(struct seq_file *m, void *v)
>  	global_dirty_limits(&background_thresh, &dirty_thresh);
>  	wb_thresh = wb_calc_thresh(wb, dirty_thresh);
>  
> -#define K(x) ((x) << (PAGE_SHIFT - 10))
>  	seq_printf(m,
>  		   "BdiWriteback:       %10lu kB\n"
>  		   "BdiReclaimable:     %10lu kB\n"
> @@ -83,20 +82,19 @@ static int bdi_debug_stats_show(struct seq_file *m, void *v)
>  		   "b_dirty_time:       %10lu\n"
>  		   "bdi_list:           %10u\n"
>  		   "state:              %10lx\n",
> -		   (unsigned long) K(wb_stat(wb, WB_WRITEBACK)),
> -		   (unsigned long) K(wb_stat(wb, WB_RECLAIMABLE)),
> -		   K(wb_thresh),
> -		   K(dirty_thresh),
> -		   K(background_thresh),
> -		   (unsigned long) K(wb_stat(wb, WB_DIRTIED)),
> -		   (unsigned long) K(wb_stat(wb, WB_WRITTEN)),
> -		   (unsigned long) K(wb->write_bandwidth),
> +		   (unsigned long) PtoK(wb_stat(wb, WB_WRITEBACK)),
> +		   (unsigned long) PtoK(wb_stat(wb, WB_RECLAIMABLE)),
> +		   PtoK(wb_thresh),
> +		   PtoK(dirty_thresh),
> +		   PtoK(background_thresh),
> +		   (unsigned long) PtoK(wb_stat(wb, WB_DIRTIED)),
> +		   (unsigned long) PtoK(wb_stat(wb, WB_WRITTEN)),
> +		   (unsigned long) PtoK(wb->write_bandwidth),
>  		   nr_dirty,
>  		   nr_io,
>  		   nr_more_io,
>  		   nr_dirty_time,
>  		   !list_empty(&bdi->bdi_list), bdi->wb.state);
> -#undef K
>  
>  	return 0;
>  }
> @@ -155,8 +153,6 @@ static ssize_t read_ahead_kb_store(struct device *dev,
>  	return count;
>  }
>  
> -#define K(pages) ((pages) << (PAGE_SHIFT - 10))
> -
>  #define BDI_SHOW(name, expr)						\
>  static ssize_t name##_show(struct device *dev,				\
>  			   struct device_attribute *attr, char *page)	\
> @@ -167,7 +163,7 @@ static ssize_t name##_show(struct device *dev,				\
>  }									\
>  static DEVICE_ATTR_RW(name);
>  
> -BDI_SHOW(read_ahead_kb, K(bdi->ra_pages))
> +BDI_SHOW(read_ahead_kb, PtoK(bdi->ra_pages))
>  
>  static ssize_t min_ratio_store(struct device *dev,
>  		struct device_attribute *attr, const char *buf, size_t count)
> diff --git a/mm/memcontrol.c b/mm/memcontrol.c
> index 94172089f52f..f0f1f4dbe816 100644
> --- a/mm/memcontrol.c
> +++ b/mm/memcontrol.c
> @@ -1132,7 +1132,6 @@ static const char *const memcg1_stat_names[] = {
>  	"swap",
>  };
>  
> -#define K(x) ((x) << (PAGE_SHIFT-10))
>  /**
>   * mem_cgroup_print_oom_info: Print OOM information relevant to memory controller.
>   * @memcg: The memory cgroup that went over limit
> @@ -1162,14 +1161,14 @@ void mem_cgroup_print_oom_info(struct mem_cgroup *memcg, struct task_struct *p)
>  	rcu_read_unlock();
>  
>  	pr_info("memory: usage %llukB, limit %llukB, failcnt %lu\n",
> -		K((u64)page_counter_read(&memcg->memory)),
> -		K((u64)memcg->memory.limit), memcg->memory.failcnt);
> +		PtoK((u64)page_counter_read(&memcg->memory)),
> +		PtoK((u64)memcg->memory.limit), memcg->memory.failcnt);
>  	pr_info("memory+swap: usage %llukB, limit %llukB, failcnt %lu\n",
> -		K((u64)page_counter_read(&memcg->memsw)),
> -		K((u64)memcg->memsw.limit), memcg->memsw.failcnt);
> +		PtoK((u64)page_counter_read(&memcg->memsw)),
> +		PtoK((u64)memcg->memsw.limit), memcg->memsw.failcnt);
>  	pr_info("kmem: usage %llukB, limit %llukB, failcnt %lu\n",
> -		K((u64)page_counter_read(&memcg->kmem)),
> -		K((u64)memcg->kmem.limit), memcg->kmem.failcnt);
> +		PtoK((u64)page_counter_read(&memcg->kmem)),
> +		PtoK((u64)memcg->kmem.limit), memcg->kmem.failcnt);
>  
>  	for_each_mem_cgroup_tree(iter, memcg) {
>  		pr_info("Memory cgroup stats for ");
> @@ -1180,12 +1179,12 @@ void mem_cgroup_print_oom_info(struct mem_cgroup *memcg, struct task_struct *p)
>  			if (memcg1_stats[i] == MEMCG_SWAP && !do_swap_account)
>  				continue;
>  			pr_cont(" %s:%luKB", memcg1_stat_names[i],
> -				K(memcg_page_state(iter, memcg1_stats[i])));
> +				PtoK(memcg_page_state(iter, memcg1_stats[i])));
>  		}
>  
>  		for (i = 0; i < NR_LRU_LISTS; i++)
>  			pr_cont(" %s:%luKB", mem_cgroup_lru_names[i],
> -				K(mem_cgroup_nr_lru_pages(iter, BIT(i))));
> +				PtoK(mem_cgroup_nr_lru_pages(iter, BIT(i))));
>  
>  		pr_cont("\n");
>  	}
> diff --git a/mm/oom_kill.c b/mm/oom_kill.c
> index 04c9143a8625..d83801347c96 100644
> --- a/mm/oom_kill.c
> +++ b/mm/oom_kill.c
> @@ -1,6 +1,6 @@
>  /*
>   *  linux/mm/oom_kill.c
> - * 
> + *
>   *  Copyright (C)  1998,2000  Rik van Riel
>   *	Thanks go out to Claus Fischer for some serious inspiration and
>   *	for goading me into coding this file...
> @@ -435,8 +435,6 @@ static DECLARE_WAIT_QUEUE_HEAD(oom_victims_wait);
>  
>  static bool oom_killer_disabled __read_mostly;
>  
> -#define K(x) ((x) << (PAGE_SHIFT-10))
> -
>  /*
>   * task->mm can be NULL if the task is the exited group leader.  So to
>   * determine whether the task is using a particular mm, we examine all the
> @@ -533,9 +531,9 @@ static bool __oom_reap_task_mm(struct task_struct *tsk, struct mm_struct *mm)
>  	tlb_finish_mmu(&tlb, 0, -1);
>  	pr_info("oom_reaper: reaped process %d (%s), now anon-rss:%lukB, file-rss:%lukB, shmem-rss:%lukB\n",
>  			task_pid_nr(tsk), tsk->comm,
> -			K(get_mm_counter(mm, MM_ANONPAGES)),
> -			K(get_mm_counter(mm, MM_FILEPAGES)),
> -			K(get_mm_counter(mm, MM_SHMEMPAGES)));
> +			PtoK(get_mm_counter(mm, MM_ANONPAGES)),
> +			PtoK(get_mm_counter(mm, MM_FILEPAGES)),
> +			PtoK(get_mm_counter(mm, MM_SHMEMPAGES)));
>  	up_read(&mm->mmap_sem);
>  
>  	/*
> @@ -884,10 +882,10 @@ static void oom_kill_process(struct oom_control *oc, const char *message)
>  	do_send_sig_info(SIGKILL, SEND_SIG_FORCED, victim, true);
>  	mark_oom_victim(victim);
>  	pr_err("Killed process %d (%s) total-vm:%lukB, anon-rss:%lukB, file-rss:%lukB, shmem-rss:%lukB\n",
> -		task_pid_nr(victim), victim->comm, K(victim->mm->total_vm),
> -		K(get_mm_counter(victim->mm, MM_ANONPAGES)),
> -		K(get_mm_counter(victim->mm, MM_FILEPAGES)),
> -		K(get_mm_counter(victim->mm, MM_SHMEMPAGES)));
> +		task_pid_nr(victim), victim->comm, PtoK(victim->mm->total_vm),
> +		PtoK(get_mm_counter(victim->mm, MM_ANONPAGES)),
> +		PtoK(get_mm_counter(victim->mm, MM_FILEPAGES)),
> +		PtoK(get_mm_counter(victim->mm, MM_SHMEMPAGES)));
>  	task_unlock(victim);
>  
>  	/*
> @@ -929,7 +927,6 @@ static void oom_kill_process(struct oom_control *oc, const char *message)
>  	mmdrop(mm);
>  	put_task_struct(victim);
>  }
> -#undef K
>  
>  /*
>   * Determines whether the kernel must panic because of the panic_on_oom sysctl.
> diff --git a/mm/page_alloc.c b/mm/page_alloc.c
> index 2302f250d6b1..5785a2f8d7db 100644
> --- a/mm/page_alloc.c
> +++ b/mm/page_alloc.c
> @@ -4460,8 +4460,6 @@ static bool show_mem_node_skip(unsigned int flags, int nid, nodemask_t *nodemask
>  	return !node_isset(nid, *nodemask);
>  }
>  
> -#define K(x) ((x) << (PAGE_SHIFT-10))
> -
>  static void show_migration_types(unsigned char type)
>  {
>  	static const char types[MIGRATE_TYPES] = {
> @@ -4565,25 +4563,27 @@ void show_free_areas(unsigned int filter, nodemask_t *nodemask)
>  			" all_unreclaimable? %s"
>  			"\n",
>  			pgdat->node_id,
> -			K(node_page_state(pgdat, NR_ACTIVE_ANON)),
> -			K(node_page_state(pgdat, NR_INACTIVE_ANON)),
> -			K(node_page_state(pgdat, NR_ACTIVE_FILE)),
> -			K(node_page_state(pgdat, NR_INACTIVE_FILE)),
> -			K(node_page_state(pgdat, NR_UNEVICTABLE)),
> -			K(node_page_state(pgdat, NR_ISOLATED_ANON)),
> -			K(node_page_state(pgdat, NR_ISOLATED_FILE)),
> -			K(node_page_state(pgdat, NR_FILE_MAPPED)),
> -			K(node_page_state(pgdat, NR_FILE_DIRTY)),
> -			K(node_page_state(pgdat, NR_WRITEBACK)),
> -			K(node_page_state(pgdat, NR_SHMEM)),
> +			PtoK(node_page_state(pgdat, NR_ACTIVE_ANON)),
> +			PtoK(node_page_state(pgdat, NR_INACTIVE_ANON)),
> +			PtoK(node_page_state(pgdat, NR_ACTIVE_FILE)),
> +			PtoK(node_page_state(pgdat, NR_INACTIVE_FILE)),
> +			PtoK(node_page_state(pgdat, NR_UNEVICTABLE)),
> +			PtoK(node_page_state(pgdat, NR_ISOLATED_ANON)),
> +			PtoK(node_page_state(pgdat, NR_ISOLATED_FILE)),
> +			PtoK(node_page_state(pgdat, NR_FILE_MAPPED)),
> +			PtoK(node_page_state(pgdat, NR_FILE_DIRTY)),
> +			PtoK(node_page_state(pgdat, NR_WRITEBACK)),
> +			PtoK(node_page_state(pgdat, NR_SHMEM)),
>  #ifdef CONFIG_TRANSPARENT_HUGEPAGE
> -			K(node_page_state(pgdat, NR_SHMEM_THPS) * HPAGE_PMD_NR),
> -			K(node_page_state(pgdat, NR_SHMEM_PMDMAPPED)
> +			PtoK(node_page_state(pgdat, NR_SHMEM_THPS) *
> +			     HPAGE_PMD_NR),
> +			PtoK(node_page_state(pgdat, NR_SHMEM_PMDMAPPED)
>  					* HPAGE_PMD_NR),
> -			K(node_page_state(pgdat, NR_ANON_THPS) * HPAGE_PMD_NR),
> +			PtoK(node_page_state(pgdat, NR_ANON_THPS) *
> +			     HPAGE_PMD_NR),
>  #endif
> -			K(node_page_state(pgdat, NR_WRITEBACK_TEMP)),
> -			K(node_page_state(pgdat, NR_UNSTABLE_NFS)),
> +			PtoK(node_page_state(pgdat, NR_WRITEBACK_TEMP)),
> +			PtoK(node_page_state(pgdat, NR_UNSTABLE_NFS)),
>  			pgdat->kswapd_failures >= MAX_RECLAIM_RETRIES ?
>  				"yes" : "no");
>  	}
> @@ -4624,27 +4624,27 @@ void show_free_areas(unsigned int filter, nodemask_t *nodemask)
>  			" free_cma:%lukB"
>  			"\n",
>  			zone->name,
> -			K(zone_page_state(zone, NR_FREE_PAGES)),
> -			K(min_wmark_pages(zone)),
> -			K(low_wmark_pages(zone)),
> -			K(high_wmark_pages(zone)),
> -			K(zone_page_state(zone, NR_ZONE_ACTIVE_ANON)),
> -			K(zone_page_state(zone, NR_ZONE_INACTIVE_ANON)),
> -			K(zone_page_state(zone, NR_ZONE_ACTIVE_FILE)),
> -			K(zone_page_state(zone, NR_ZONE_INACTIVE_FILE)),
> -			K(zone_page_state(zone, NR_ZONE_UNEVICTABLE)),
> -			K(zone_page_state(zone, NR_ZONE_WRITE_PENDING)),
> -			K(zone->present_pages),
> -			K(zone->managed_pages),
> -			K(zone_page_state(zone, NR_MLOCK)),
> -			K(zone_page_state(zone, NR_SLAB_RECLAIMABLE)),
> -			K(zone_page_state(zone, NR_SLAB_UNRECLAIMABLE)),
> +			PtoK(zone_page_state(zone, NR_FREE_PAGES)),
> +			PtoK(min_wmark_pages(zone)),
> +			PtoK(low_wmark_pages(zone)),
> +			PtoK(high_wmark_pages(zone)),
> +			PtoK(zone_page_state(zone, NR_ZONE_ACTIVE_ANON)),
> +			PtoK(zone_page_state(zone, NR_ZONE_INACTIVE_ANON)),
> +			PtoK(zone_page_state(zone, NR_ZONE_ACTIVE_FILE)),
> +			PtoK(zone_page_state(zone, NR_ZONE_INACTIVE_FILE)),
> +			PtoK(zone_page_state(zone, NR_ZONE_UNEVICTABLE)),
> +			PtoK(zone_page_state(zone, NR_ZONE_WRITE_PENDING)),
> +			PtoK(zone->present_pages),
> +			PtoK(zone->managed_pages),
> +			PtoK(zone_page_state(zone, NR_MLOCK)),
> +			PtoK(zone_page_state(zone, NR_SLAB_RECLAIMABLE)),
> +			PtoK(zone_page_state(zone, NR_SLAB_UNRECLAIMABLE)),
>  			zone_page_state(zone, NR_KERNEL_STACK_KB),
> -			K(zone_page_state(zone, NR_PAGETABLE)),
> -			K(zone_page_state(zone, NR_BOUNCE)),
> -			K(free_pcp),
> -			K(this_cpu_read(zone->pageset->pcp.count)),
> -			K(zone_page_state(zone, NR_FREE_CMA_PAGES)));
> +			PtoK(zone_page_state(zone, NR_PAGETABLE)),
> +			PtoK(zone_page_state(zone, NR_BOUNCE)),
> +			PtoK(free_pcp),
> +			PtoK(this_cpu_read(zone->pageset->pcp.count)),
> +			PtoK(zone_page_state(zone, NR_FREE_CMA_PAGES)));
>  		printk("lowmem_reserve[]:");
>  		for (i = 0; i < MAX_NR_ZONES; i++)
>  			printk(KERN_CONT " %ld", zone->lowmem_reserve[i]);
> @@ -4678,11 +4678,11 @@ void show_free_areas(unsigned int filter, nodemask_t *nodemask)
>  		spin_unlock_irqrestore(&zone->lock, flags);
>  		for (order = 0; order < MAX_ORDER; order++) {
>  			printk(KERN_CONT "%lu*%lukB ",
> -			       nr[order], K(1UL) << order);
> +			       nr[order], PtoK(1UL) << order);
>  			if (nr[order])
>  				show_migration_types(types[order]);
>  		}
> -		printk(KERN_CONT "= %lukB\n", K(total));
> +		printk(KERN_CONT "= %lukB\n", PtoK(total));
>  	}
>  
>  	hugetlb_show_meminfo();
> -- 
> 2.7.4

-- 
Michal Hocko
SUSE Labs

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ