[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <1357299744.5273.4.camel@kernel.cn.ibm.com>
Date: Fri, 04 Jan 2013 05:42:24 -0600
From: Simon Jeons <simon.jeons@...il.com>
To: Mel Gorman <mgorman@...e.de>
Cc: Peter Zijlstra <a.p.zijlstra@...llo.nl>,
Andrea Arcangeli <aarcange@...hat.com>,
Ingo Molnar <mingo@...nel.org>, Rik van Riel <riel@...hat.com>,
Johannes Weiner <hannes@...xchg.org>,
Hugh Dickins <hughd@...gle.com>,
Thomas Gleixner <tglx@...utronix.de>,
Paul Turner <pjt@...gle.com>, Hillf Danton <dhillf@...il.com>,
David Rientjes <rientjes@...gle.com>,
Lee Schermerhorn <Lee.Schermerhorn@...com>,
Alex Shi <lkml.alex@...il.com>,
Srikar Dronamraju <srikar@...ux.vnet.ibm.com>,
Aneesh Kumar <aneesh.kumar@...ux.vnet.ibm.com>,
Linus Torvalds <torvalds@...ux-foundation.org>,
Andrew Morton <akpm@...ux-foundation.org>,
Linux-MM <linux-mm@...ck.org>,
LKML <linux-kernel@...r.kernel.org>
Subject: Re: [PATCH 29/49] mm: numa: Add pte updates, hinting and migration
stats
On Fri, 2012-12-07 at 10:23 +0000, Mel Gorman wrote:
> It is tricky to quantify the basic cost of automatic NUMA placement in a
> meaningful manner. This patch adds some vmstats that can be used as part
> of a basic costing model.
Hi Gorman,
>
> u = basic unit = sizeof(void *)
> Ca = cost of struct page access = sizeof(struct page) / u
> Cpte = Cost PTE access = Ca
> Cupdate = Cost PTE update = (2 * Cpte) + (2 * Wlock)
> where Cpte is incurred twice for a read and a write and Wlock
> is a constant representing the cost of taking or releasing a
> lock
> Cnumahint = Cost of a minor page fault = some high constant e.g. 1000
> Cpagerw = Cost to read or write a full page = Ca + PAGE_SIZE/u
Why cpagerw = Ca + PAGE_SIZE/u instead of Cpte + PAGE_SIZE/u ?
> Ci = Cost of page isolation = Ca + Wi
> where Wi is a constant that should reflect the approximate cost
> of the locking operation
> Cpagecopy = Cpagerw + (Cpagerw * Wnuma) + Ci + (Ci * Wnuma)
> where Wnuma is the approximate NUMA factor. 1 is local. 1.2
> would imply that remote accesses are 20% more expensive
>
> Balancing cost = Cpte * numa_pte_updates +
> Cnumahint * numa_hint_faults +
> Ci * numa_pages_migrated +
> Cpagecopy * numa_pages_migrated
>
Since Cpagecopy has already accumulated ci why count ci twice ?
> Note that numa_pages_migrated is used as a measure of how many pages
> were isolated even though it would miss pages that failed to migrate. A
> vmstat counter could have been added for it but the isolation cost is
> pretty marginal in comparison to the overall cost so it seemed overkill.
>
> The ideal way to measure automatic placement benefit would be to count
> the number of remote accesses versus local accesses and do something like
>
> benefit = (remote_accesses_before - remove_access_after) * Wnuma
>
> but the information is not readily available. As a workload converges, the
> expection would be that the number of remote numa hints would reduce to 0.
>
> convergence = numa_hint_faults_local / numa_hint_faults
> where this is measured for the last N number of
> numa hints recorded. When the workload is fully
> converged the value is 1.
>
convergence tend to 0 is better or 1 is better? If tend to 1, Cpte *
numa_pte_updates + Cnumahint * numa_hint_faults are just waste, where I
miss?
> This can measure if the placement policy is converging and how fast it is
> doing it.
>
> Signed-off-by: Mel Gorman <mgorman@...e.de>
> Acked-by: Rik van Riel <riel@...hat.com>
> ---
> include/linux/vm_event_item.h | 6 ++++++
> include/linux/vmstat.h | 8 ++++++++
> mm/huge_memory.c | 5 +++++
> mm/memory.c | 12 ++++++++++++
> mm/mempolicy.c | 2 ++
> mm/migrate.c | 3 ++-
> mm/vmstat.c | 6 ++++++
> 7 files changed, 41 insertions(+), 1 deletion(-)
>
> diff --git a/include/linux/vm_event_item.h b/include/linux/vm_event_item.h
> index a1f750b..dded0af 100644
> --- a/include/linux/vm_event_item.h
> +++ b/include/linux/vm_event_item.h
> @@ -38,6 +38,12 @@ enum vm_event_item { PGPGIN, PGPGOUT, PSWPIN, PSWPOUT,
> KSWAPD_LOW_WMARK_HIT_QUICKLY, KSWAPD_HIGH_WMARK_HIT_QUICKLY,
> KSWAPD_SKIP_CONGESTION_WAIT,
> PAGEOUTRUN, ALLOCSTALL, PGROTATED,
> +#ifdef CONFIG_BALANCE_NUMA
> + NUMA_PTE_UPDATES,
> + NUMA_HINT_FAULTS,
> + NUMA_HINT_FAULTS_LOCAL,
> + NUMA_PAGE_MIGRATE,
> +#endif
> #ifdef CONFIG_MIGRATION
> PGMIGRATE_SUCCESS, PGMIGRATE_FAIL,
> #endif
> diff --git a/include/linux/vmstat.h b/include/linux/vmstat.h
> index 92a86b2..dffccfa 100644
> --- a/include/linux/vmstat.h
> +++ b/include/linux/vmstat.h
> @@ -80,6 +80,14 @@ static inline void vm_events_fold_cpu(int cpu)
>
> #endif /* CONFIG_VM_EVENT_COUNTERS */
>
> +#ifdef CONFIG_BALANCE_NUMA
> +#define count_vm_numa_event(x) count_vm_event(x)
> +#define count_vm_numa_events(x, y) count_vm_events(x, y)
> +#else
> +#define count_vm_numa_event(x) do {} while (0)
> +#define count_vm_numa_events(x, y) do {} while (0)
> +#endif /* CONFIG_BALANCE_NUMA */
> +
> #define __count_zone_vm_events(item, zone, delta) \
> __count_vm_events(item##_NORMAL - ZONE_NORMAL + \
> zone_idx(zone), delta)
> diff --git a/mm/huge_memory.c b/mm/huge_memory.c
> index b3d4c4b..66e73cc 100644
> --- a/mm/huge_memory.c
> +++ b/mm/huge_memory.c
> @@ -1025,6 +1025,7 @@ int do_huge_pmd_numa_page(struct mm_struct *mm, struct vm_area_struct *vma,
> struct page *page = NULL;
> unsigned long haddr = addr & HPAGE_PMD_MASK;
> int target_nid;
> + int current_nid = -1;
>
> spin_lock(&mm->page_table_lock);
> if (unlikely(!pmd_same(pmd, *pmdp)))
> @@ -1033,6 +1034,10 @@ int do_huge_pmd_numa_page(struct mm_struct *mm, struct vm_area_struct *vma,
> page = pmd_page(pmd);
> get_page(page);
> spin_unlock(&mm->page_table_lock);
> + current_nid = page_to_nid(page);
> + count_vm_numa_event(NUMA_HINT_FAULTS);
> + if (current_nid == numa_node_id())
> + count_vm_numa_event(NUMA_HINT_FAULTS_LOCAL);
>
> target_nid = mpol_misplaced(page, vma, haddr);
> if (target_nid == -1)
> diff --git a/mm/memory.c b/mm/memory.c
> index 1d6f85a..47f5dd1 100644
> --- a/mm/memory.c
> +++ b/mm/memory.c
> @@ -3477,6 +3477,7 @@ int do_numa_page(struct mm_struct *mm, struct vm_area_struct *vma,
> set_pte_at(mm, addr, ptep, pte);
> update_mmu_cache(vma, addr, ptep);
>
> + count_vm_numa_event(NUMA_HINT_FAULTS);
> page = vm_normal_page(vma, addr, pte);
> if (!page) {
> pte_unmap_unlock(ptep, ptl);
> @@ -3485,6 +3486,8 @@ int do_numa_page(struct mm_struct *mm, struct vm_area_struct *vma,
>
> get_page(page);
> current_nid = page_to_nid(page);
> + if (current_nid == numa_node_id())
> + count_vm_numa_event(NUMA_HINT_FAULTS_LOCAL);
> target_nid = mpol_misplaced(page, vma, addr);
> pte_unmap_unlock(ptep, ptl);
> if (target_nid == -1) {
> @@ -3517,6 +3520,9 @@ static int do_pmd_numa_page(struct mm_struct *mm, struct vm_area_struct *vma,
> unsigned long offset;
> spinlock_t *ptl;
> bool numa = false;
> + int local_nid = numa_node_id();
> + unsigned long nr_faults = 0;
> + unsigned long nr_faults_local = 0;
>
> spin_lock(&mm->page_table_lock);
> pmd = *pmdp;
> @@ -3565,10 +3571,16 @@ static int do_pmd_numa_page(struct mm_struct *mm, struct vm_area_struct *vma,
> curr_nid = page_to_nid(page);
> task_numa_fault(curr_nid, 1);
>
> + nr_faults++;
> + if (curr_nid == local_nid)
> + nr_faults_local++;
> +
> pte = pte_offset_map_lock(mm, pmdp, addr, &ptl);
> }
> pte_unmap_unlock(orig_pte, ptl);
>
> + count_vm_numa_events(NUMA_HINT_FAULTS, nr_faults);
> + count_vm_numa_events(NUMA_HINT_FAULTS_LOCAL, nr_faults_local);
> return 0;
> }
> #else
> diff --git a/mm/mempolicy.c b/mm/mempolicy.c
> index a7a62fe..516491f 100644
> --- a/mm/mempolicy.c
> +++ b/mm/mempolicy.c
> @@ -583,6 +583,8 @@ unsigned long change_prot_numa(struct vm_area_struct *vma,
> BUILD_BUG_ON(_PAGE_NUMA != _PAGE_PROTNONE);
>
> nr_updated = change_protection(vma, addr, end, vma->vm_page_prot, 0, 1);
> + if (nr_updated)
> + count_vm_numa_events(NUMA_PTE_UPDATES, nr_updated);
>
> return nr_updated;
> }
> diff --git a/mm/migrate.c b/mm/migrate.c
> index 49878d7..4f55694 100644
> --- a/mm/migrate.c
> +++ b/mm/migrate.c
> @@ -1514,7 +1514,8 @@ int migrate_misplaced_page(struct page *page, int node)
> if (nr_remaining) {
> putback_lru_pages(&migratepages);
> isolated = 0;
> - }
> + } else
> + count_vm_numa_event(NUMA_PAGE_MIGRATE);
> }
> BUG_ON(!list_empty(&migratepages));
> out:
> diff --git a/mm/vmstat.c b/mm/vmstat.c
> index 3a067fa..cfa386da 100644
> --- a/mm/vmstat.c
> +++ b/mm/vmstat.c
> @@ -774,6 +774,12 @@ const char * const vmstat_text[] = {
>
> "pgrotated",
>
> +#ifdef CONFIG_BALANCE_NUMA
> + "numa_pte_updates",
> + "numa_hint_faults",
> + "numa_hint_faults_local",
> + "numa_pages_migrated",
> +#endif
> #ifdef CONFIG_MIGRATION
> "pgmigrate_success",
> "pgmigrate_fail",
--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/
Powered by blists - more mailing lists