[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <4C3C35B7.50101@cn.fujitsu.com>
Date: Tue, 13 Jul 2010 17:45:27 +0800
From: Xiao Guangrong <xiaoguangrong@...fujitsu.com>
To: Avi Kivity <avi@...hat.com>
CC: LKML <linux-kernel@...r.kernel.org>,
KVM list <kvm@...r.kernel.org>,
Marcelo Tosatti <mtosatti@...hat.com>
Subject: [PATCH 3/4] KVM: MMU: track dirty page in speculative path properly
In speculative path, the page is not real write-access, no need mark it
dirty, so clear dirty bit in this path and later examine this bit when
we release the page
Signed-off-by: Xiao Guangrong <xiaoguangrong@...fujitsu.com>
---
arch/x86/kvm/mmu.c | 24 +++++++++++-------------
1 files changed, 11 insertions(+), 13 deletions(-)
diff --git a/arch/x86/kvm/mmu.c b/arch/x86/kvm/mmu.c
index 67dbafa..5e9d4a0 100644
--- a/arch/x86/kvm/mmu.c
+++ b/arch/x86/kvm/mmu.c
@@ -315,21 +315,19 @@ static void set_spte_atomic(u64 *sptep, u64 new_spte)
pfn = spte_to_pfn(old_spte);
if (old_spte & shadow_accessed_mask)
kvm_set_pfn_accessed(pfn);
- if (is_writable_pte(old_spte))
+
+ if ((shadow_dirty_mask && (old_spte & shadow_dirty_mask)) ||
+ (!shadow_dirty_mask && is_writable_pte(old_spte)))
kvm_set_pfn_dirty(pfn);
}
static void update_spte(u64 *sptep, u64 new_spte)
{
- u64 old_spte;
-
- if (!shadow_accessed_mask || (new_spte & shadow_accessed_mask)) {
+ if ((!shadow_accessed_mask || (new_spte & shadow_accessed_mask)) &&
+ (!shadow_dirty_mask || (new_spte & shadow_dirty_mask)))
__set_spte(sptep, new_spte);
- } else {
- old_spte = __xchg_spte(sptep, new_spte);
- if (old_spte & shadow_accessed_mask)
- mark_page_accessed(pfn_to_page(spte_to_pfn(old_spte)));
- }
+ else
+ set_spte_atomic(sptep, new_spte);
}
static int mmu_topup_memory_cache(struct kvm_mmu_memory_cache *cache,
@@ -745,7 +743,7 @@ static int rmap_write_protect(struct kvm *kvm, u64 gfn)
}
spte = rmap_next(kvm, rmapp, spte);
}
- if (write_protected) {
+ if (!shadow_dirty_mask && write_protected) {
pfn_t pfn;
spte = rmap_next(kvm, rmapp, NULL);
@@ -1879,9 +1877,9 @@ static int set_spte(struct kvm_vcpu *vcpu, u64 *sptep,
* whether the guest actually used the pte (in order to detect
* demand paging).
*/
- spte = shadow_base_present_pte | shadow_dirty_mask;
+ spte = shadow_base_present_pte;
if (!speculative)
- spte |= shadow_accessed_mask;
+ spte |= shadow_accessed_mask | shadow_dirty_mask;
if (!dirty)
pte_access &= ~ACC_WRITE_MASK;
if (pte_access & ACC_EXEC_MASK)
@@ -2007,7 +2005,7 @@ static void mmu_set_spte(struct kvm_vcpu *vcpu, u64 *sptep,
if (rmap_count > RMAP_RECYCLE_THRESHOLD)
rmap_recycle(vcpu, sptep, gfn);
} else {
- if (was_writable)
+ if (!shadow_dirty_mask && was_writable)
kvm_release_pfn_dirty(pfn);
else
kvm_release_pfn_clean(pfn);
--
1.6.1.2
--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/
Powered by blists - more mailing lists