lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [thread-next>] [day] [month] [year] [list]
Message-Id: <20231009025510.342681-1-zhaotianrui@loongson.cn>
Date:   Mon,  9 Oct 2023 10:55:06 +0800
From:   Tianrui Zhao <zhaotianrui@...ngson.cn>
To:     Shuah Khan <shuah@...nel.org>, Paolo Bonzini <pbonzini@...hat.com>,
        linux-kernel@...r.kernel.org, kvm@...r.kernel.org,
        Sean Christopherson <seanjc@...gle.com>
Cc:     Vishal Annapurve <vannapurve@...gle.com>,
        Huacai Chen <chenhuacai@...nel.org>,
        WANG Xuerui <kernel@...0n.name>, loongarch@...ts.linux.dev,
        Peter Xu <peterx@...hat.com>,
        Vipin Sharma <vipinsh@...gle.com>, maobibo@...ngson.cn,
        zhaotianrui@...ngson.cn
Subject: [PATCH v3 0/4] KVM: selftests: Add LoongArch support

This patch series base on the Linux LoongArch KVM patch:
Based-on: <20230927030959.3629941-1-zhaotianrui@...ngson.cn> 

We add LoongArch support into KVM selftests and there are some KVM
test cases we have passed:
	demand_paging_test
	dirty_log_perf_test
	dirty_log_test
	guest_print_test
	kvm_binary_stats_test
	kvm_create_max_vcpus
	kvm_page_table_test
	memslot_modification_stress_test
	memslot_perf_test
	set_memory_region_test

Changes for v3:
1. Improve implementation of LoongArch VM page walk.
2. Add exception handler for LoongArch.
3. Add dirty_log_test, dirty_log_perf_test, guest_print_test
test cases for LoongArch.
4. Add __ASSEMBLER__ macro to distinguish asm file and c file.
5. Move ucall_arch_do_ucall to the header file and make it as
static inline to avoid function calls.
6. Change the DEFAULT_GUEST_TEST_MEM base addr for LoongArch.

Changes for v2:
1. We should use ".balign 4096" to align the assemble code with 4K in
exception.S instead of "align 12".
2. LoongArch only supports 3 or 4 levels page tables, so we remove the
hanlders for 2-levels page table.
3. Remove the DEFAULT_LOONGARCH_GUEST_STACK_VADDR_MIN and use the common
DEFAULT_GUEST_STACK_VADDR_MIN to allocate stack memory in guest.
4. Reorganize the test cases supported by LoongArch.
5. Fix some code comments.
6. Add kvm_binary_stats_test test case into LoongArch KVM selftests.

changes for v1:
1. Add kvm selftests header files for LoongArch.
2. Add processor tests for LoongArch KVM.
3. Add ucall tests for LoongArch KVM.
4. Add LoongArch tests into makefile.

All of the test cases results:
1..10
# timeout set to 120
# selftests: kvm: demand_paging_test
# Testing guest mode: PA-bits:36,  VA-bits:47, 16K pages
# guest physical test memory: [0xfbfffc000, 0xfffffc000)
# Finished creating vCPUs and starting uffd threads
# Started all vCPUs
# All vCPU threads joined
# Total guest execution time: 0.200804700s
# Overall demand paging rate: 326366.862927 pgs/sec
ok 1 selftests: kvm: demand_paging_test
# timeout set to 120
# selftests: kvm: dirty_log_perf_test
# Test iterations: 2
# Testing guest mode: PA-bits:36,  VA-bits:47, 16K pages
# guest physical test memory: [0xfbfffc000, 0xfffffc000)
# Random seed: 1
# Populate memory time: 0.201452560s
# Enabling dirty logging time: 0.000451670s
# 
# Iteration 1 dirty memory time: 0.051582140s
# Iteration 1 get dirty log time: 0.000010510s
# Iteration 1 clear dirty log time: 0.000421730s
# Iteration 2 dirty memory time: 0.046593760s
# Iteration 2 get dirty log time: 0.000002110s
# Iteration 2 clear dirty log time: 0.000418020s
# Disabling dirty logging time: 0.002948490s
# Get dirty log over 2 iterations took 0.000012620s. (Avg 0.000006310s/iteration)
# Clear dirty log over 2 iterations took 0.000839750s. (Avg 0.000419875s/iteration)
ok 2 selftests: kvm: dirty_log_perf_test
# timeout set to 120
# selftests: kvm: dirty_log_test
# Test iterations: 32, interval: 10 (ms)
# Testing Log Mode 'dirty-log'
# Testing guest mode: PA-bits:36,  VA-bits:47, 16K pages
# guest physical test memory offset: 0xfbfff0000
# Dirtied 453632 pages
# Total bits checked: dirty (436564), clear (1595145), track_next (70002)
# Testing Log Mode 'clear-log'
# Testing guest mode: PA-bits:36,  VA-bits:47, 16K pages
# guest physical test memory offset: 0xfbfff0000
# Dirtied 425984 pages
# Total bits checked: dirty (414397), clear (1617312), track_next (68152)
# Testing Log Mode 'dirty-ring'
# Testing guest mode: PA-bits:36,  VA-bits:47, 16K pages
# dirty ring count: 0x10000
# guest physical test memory offset: 0xfbfff0000
# vcpu stops because vcpu is kicked out...
# Notifying vcpu to continue
# vcpu continues now.
# Iteration 1 collected 3201 pages
# vcpu stops because dirty ring is full...
# vcpu continues now.
# vcpu stops because dirty ring is full...
# Notifying vcpu to continue
# Iteration 2 collected 65472 pages
# ......
# vcpu continues now.
# vcpu stops because vcpu is kicked out...
# vcpu continues now.
# vcpu stops because vcpu is kicked out...
# Notifying vcpu to continue
# vcpu continues now.
# Iteration 31 collected 12642 pages
# vcpu stops because dirty ring is full...
# vcpu continues now.
# Dirtied 7275520 pages
# Total bits checked: dirty (1165675), clear (866034), track_next (811358)
ok 3 selftests: kvm: dirty_log_test
# timeout set to 120
# selftests: kvm: guest_print_test
ok 4 selftests: kvm: guest_print_test
# timeout set to 120
# selftests: kvm: kvm_binary_stats_test
# TAP version 13
# 1..4
# ok 1 vm0
# ok 2 vm1
# ok 3 vm2
# ok 4 vm3
# # Totals: pass:4 fail:0 xfail:0 xpass:0 skip:0 error:0
ok 5 selftests: kvm: kvm_binary_stats_test
# timeout set to 120
# selftests: kvm: kvm_create_max_vcpus
# KVM_CAP_MAX_VCPU_ID: 256
# KVM_CAP_MAX_VCPUS: 256
# Testing creating 256 vCPUs, with IDs 0...255.
ok 6 selftests: kvm: kvm_create_max_vcpus
# timeout set to 120
# selftests: kvm: kvm_page_table_test
# Testing guest mode: PA-bits:36,  VA-bits:47, 16K pages
# Testing memory backing src type: anonymous
# Testing memory backing src granularity: 0x4000
# Testing memory size(aligned): 0x40000000
# Guest physical test memory offset: 0xfbfffc000
# Host  virtual  test memory offset: 0x7fffb0860000
# Number of testing vCPUs: 1
# Started all vCPUs successfully
# KVM_CREATE_MAPPINGS: total execution time: 0.200919330s
# 
# KVM_UPDATE_MAPPINGS: total execution time: 0.051182930s
# 
# KVM_ADJUST_MAPPINGS: total execution time: 0.010083590s
# 
ok 7 selftests: kvm: kvm_page_table_test
# timeout set to 120
# selftests: kvm: memslot_modification_stress_test
# Testing guest mode: PA-bits:36,  VA-bits:47, 16K pages
# guest physical test memory: [0xfbfffc000, 0xfffffc000)
# Finished creating vCPUs
# Started all vCPUs
# All vCPU threads joined
ok 8 selftests: kvm: memslot_modification_stress_test
# timeout set to 120
# selftests: kvm: memslot_perf_test
# Testing map performance with 1 runs, 5 seconds each
# Memslot count too high for this test, decrease the cap (max is 2053)
# 
# Testing unmap performance with 1 runs, 5 seconds each
# Memslot count too high for this test, decrease the cap (max is 8197)
# 
# Testing unmap chunked performance with 1 runs, 5 seconds each
# Memslot count too high for this test, decrease the cap (max is 8197)
# 
# Testing move active area performance with 1 runs, 5 seconds each
# Test took 0.761678900s for slot setup + 5.000014460s all iterations
# Done 120167 iterations, avg 0.000041608s each
# Best runtime result was 0.000041608s per iteration (with 120167 iterations)
# 
# Testing move inactive area performance with 1 runs, 5 seconds each
# Test took 0.771796550s for slot setup + 5.000018520s all iterations
# Done 136354 iterations, avg 0.000036669s each
# Best runtime result was 0.000036669s per iteration (with 136354 iterations)
# 
# Testing RW performance with 1 runs, 5 seconds each
# Test took 0.763568840s for slot setup + 5.002233800s all iterations
# Done 649 iterations, avg 0.007707602s each
# Best runtime result was 0.007707602s per iteration (with 649 iterations)
# Best slot setup time for the whole test area was 0.761678900s
ok 9 selftests: kvm: memslot_perf_test
# timeout set to 120
# selftests: kvm: set_memory_region_test
# Allowed number of memory slots: 32767
# Adding slots 0..32766, each memory region with 2048K size
ok 10 selftests: kvm: set_memory_region_test

Tianrui Zhao (4):
  KVM: selftests: Add KVM selftests header files for LoongArch
  KVM: selftests: Add core KVM selftests support for LoongArch
  KVM: selftests: Add ucall test support for LoongArch
  KVM: selftests: Add test cases for LoongArch

 tools/testing/selftests/kvm/Makefile          |  15 +
 .../selftests/kvm/include/kvm_util_base.h     |   5 +
 .../kvm/include/loongarch/processor.h         | 133 +++++++
 .../selftests/kvm/include/loongarch/ucall.h   |  20 ++
 .../testing/selftests/kvm/include/memstress.h |  10 +
 .../selftests/kvm/lib/loongarch/exception.S   |  59 ++++
 .../selftests/kvm/lib/loongarch/processor.c   | 333 ++++++++++++++++++
 .../selftests/kvm/lib/loongarch/ucall.c       |  38 ++
 8 files changed, 613 insertions(+)
 create mode 100644 tools/testing/selftests/kvm/include/loongarch/processor.h
 create mode 100644 tools/testing/selftests/kvm/include/loongarch/ucall.h
 create mode 100644 tools/testing/selftests/kvm/lib/loongarch/exception.S
 create mode 100644 tools/testing/selftests/kvm/lib/loongarch/processor.c
 create mode 100644 tools/testing/selftests/kvm/lib/loongarch/ucall.c

-- 
2.39.1

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ