[<prev] [next>] [day] [month] [year] [list]
Message-ID: <tip-fvywddcv5mj2lr0y76i7ai1j@git.kernel.org>
Date: Fri, 28 Sep 2012 00:55:43 -0700
From: tip-bot for Peter Zijlstra <a.p.zijlstra@...llo.nl>
To: linux-tip-commits@...r.kernel.org
Cc: linux-kernel@...r.kernel.org, hpa@...or.com, mingo@...nel.org,
torvalds@...ux-foundation.org, a.p.zijlstra@...llo.nl,
riel@...hat.com, akpm@...ux-foundation.org, tglx@...utronix.de
Subject: [tip:sched/numa] mm: Optimize do_prot_none()
Commit-ID: 1a201bacffd9f7cbc0f9a3b790dcb8108336784d
Gitweb: http://git.kernel.org/tip/1a201bacffd9f7cbc0f9a3b790dcb8108336784d
Author: Peter Zijlstra <a.p.zijlstra@...llo.nl>
AuthorDate: Wed, 26 Sep 2012 16:05:01 +0200
Committer: Ingo Molnar <mingo@...nel.org>
CommitDate: Thu, 27 Sep 2012 14:46:02 +0200
mm: Optimize do_prot_none()
Reduces do_prot_none() to a single pte_lock acquisition in the !migrate case.
Also flipping the protection bits back sooner avoids the chance of
other CPUs hitting it and causing a pile-up on the pte_lock.
This significantly reduces contention on the pte_lock in a NUMA page fault
benchmark, system time dropped from around 78% to 35% and __memset_sse2()
is the most expensive function instead of _raw_spin_lock()/handle_pte_fault().
Signed-off-by: Peter Zijlstra <a.p.zijlstra@...llo.nl>
Cc: Linus Torvalds <torvalds@...ux-foundation.org>
Cc: Andrew Morton <akpm@...ux-foundation.org>
Cc: Rik van Riel <riel@...hat.com>
Link: http://lkml.kernel.org/n/tip-fvywddcv5mj2lr0y76i7ai1j@git.kernel.org
Signed-off-by: Ingo Molnar <mingo@...nel.org>
---
mm/memory.c | 66 +++++++++++++++++++++++++----------------------------------
1 files changed, 28 insertions(+), 38 deletions(-)
diff --git a/mm/memory.c b/mm/memory.c
index d896a24..965eeef 100644
--- a/mm/memory.c
+++ b/mm/memory.c
@@ -3438,56 +3438,39 @@ static bool pte_prot_none(struct vm_area_struct *vma, pte_t pte)
return pte_same(pte, pte_modify(pte, vma_prot_none(vma)));
}
-static int do_prot_none(struct mm_struct *mm, struct vm_area_struct *vma,
- unsigned long address, pte_t *ptep, pmd_t *pmd,
- unsigned int flags, pte_t entry)
+#ifdef CONFIG_NUMA
+static void do_prot_none_numa(struct mm_struct *mm, struct vm_area_struct *vma,
+ unsigned long address, struct page *page)
{
- struct page *page = NULL;
- spinlock_t *ptl;
int node;
- ptl = pte_lockptr(mm, pmd);
- spin_lock(ptl);
- if (unlikely(!pte_same(*ptep, entry)))
- goto unlock;
-
-#ifdef CONFIG_NUMA
/*
* For NUMA systems we use the special PROT_NONE maps to drive
* lazy page migration, see MPOL_MF_LAZY and related.
*/
- page = vm_normal_page(vma, address, entry);
- if (!page)
- goto do_fixup_locked;
-
- get_page(page);
- pte_unmap_unlock(ptep, ptl);
-
node = mpol_misplaced(page, vma, address);
- if (node == -1)
- goto do_fixup;
+ if (node != -1)
+ migrate_misplaced_page(mm, page, node);
+}
+#else
+static void do_prot_none_numa(struct mm_struct *mm, struct vm_area_struct *vma,
+ unsigned long address, struct page *page)
+{
+}
+#endif /* CONFIG_NUMA */
- /*
- * Page migration will install a new pte with vma->vm_page_prot,
- * otherwise fall-through to the fixup. Next time,.. perhaps.
- */
- if (!migrate_misplaced_page(mm, page, node)) {
- put_page(page);
- return 0;
- }
+static int do_prot_none(struct mm_struct *mm, struct vm_area_struct *vma,
+ unsigned long address, pte_t *ptep, pmd_t *pmd,
+ unsigned int flags, pte_t entry)
+{
+ struct page *page = NULL;
+ spinlock_t *ptl;
-do_fixup:
- /*
- * OK, nothing to do,.. change the protection back to what it
- * ought to be.
- */
- ptep = pte_offset_map_lock(mm, pmd, address, &ptl);
+ ptl = pte_lockptr(mm, pmd);
+ spin_lock(ptl);
if (unlikely(!pte_same(*ptep, entry)))
goto unlock;
-do_fixup_locked:
-#endif /* CONFIG_NUMA */
-
flush_cache_page(vma, address, pte_pfn(entry));
ptep_modify_prot_start(mm, address, ptep);
@@ -3495,10 +3478,17 @@ do_fixup_locked:
ptep_modify_prot_commit(mm, address, ptep, entry);
update_mmu_cache(vma, address, ptep);
+
+ page = vm_normal_page(vma, address, entry);
+ if (page)
+ get_page(page);
+
unlock:
pte_unmap_unlock(ptep, ptl);
- if (page)
+ if (page) {
+ do_prot_none_numa(mm, vma, address, page);
put_page(page);
+ }
return 0;
}
--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/
Powered by blists - more mailing lists