[<prev] [next>] [day] [month] [year] [list]
Message-Id: <1363768178-4596-1-git-send-email-xiaoguangrong@linux.vnet.ibm.com>
Date: Wed, 20 Mar 2013 16:29:38 +0800
From: Xiao Guangrong <xiaoguangrong@...ux.vnet.ibm.com>
To: mtosatti@...hat.com
Cc: gleb@...hat.com, linux-kernel@...r.kernel.org, kvm@...r.kernel.org,
Xiao Guangrong <xiaoguangrong@...ux.vnet.ibm.com>
Subject: [PATCH v2 0/7] KVM: MMU: fast zap all shadow pages
Changlog:
V2:
- do not reset n_requested_mmu_pages and n_max_mmu_pages
- batch free root shadow pages to reduce vcpu notification and mmu-lock
contention
- remove the first patch that introduce kvm->arch.mmu_cache since we only
'memset zero' on hashtable rather than all mmu cache members in this
version
- remove unnecessary kvm_reload_remote_mmus after kvm_mmu_zap_all
* Issue
The current kvm_mmu_zap_all is really slow - it is holding mmu-lock to
walk and zap all shadow pages one by one, also it need to zap all guest
page's rmap and all shadow page's parent spte list. Particularly, things
become worse if guest uses more memory or vcpus. It is not good for
scalability.
* Idea
Since all shadow page will be zapped, we can directly zap the mmu-cache
and rmap so that vcpu will fault on the new mmu-cache, after that, we can
directly free the memory used by old mmu-cache.
The root shadow page is little especial since they are currently used by
vcpus, we can not directly free them. So, we zap the root shadow pages and
re-add them into the new mmu-cache.
* TODO
(1): free root shadow pages by using generation-number
(2): drop unnecessary @npages from kvm_arch_create_memslot
* Performance
The testcase can be found at:
http://www.gossamer-threads.com/lists/engine?do=post_attachment;postatt_id=54896;list=linux
is used to measure the time of delete / add memslot. At that time, all vcpus
are waiting, that means, no mmu-lock contention. I believe the result be more
beautiful if other vcpus and mmu notification need to hold the mmu-lock.
Guest VCPU:6, Mem:2048M
before: Run 10 times, Avg time:46078825 ns.
after: Run 10 times, Avg time:21558774 ns. (+ 113%)
Xiao Guangrong (7):
KVM: MMU: introduce mmu_cache->pte_list_descs
KVM: x86: introduce memslot_set_lpage_disallowed
KVM: x86: introduce kvm_clear_all_gfn_page_info
KVM: MMU: delete shadow page from hash list in
kvm_mmu_prepare_zap_page
KVM: MMU: split kvm_mmu_prepare_zap_page
KVM: MMU: fast zap all shadow pages
KVM: MMU: drop unnecessary kvm_reload_remote_mmus after
kvm_mmu_zap_all
arch/x86/include/asm/kvm_host.h | 7 ++-
arch/x86/kvm/mmu.c | 105 ++++++++++++++++++++++++++++++++++-----
arch/x86/kvm/mmu.h | 1 +
arch/x86/kvm/x86.c | 87 +++++++++++++++++++++++++-------
include/linux/kvm_host.h | 1 +
5 files changed, 166 insertions(+), 35 deletions(-)
--
1.7.7.6
--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/
Powered by blists - more mailing lists