[PATCH v5 0/4] KVM: selftests: Add LoongArch support

From: Tianrui Zhao
Date: Thu Nov 30 2023 - 06:31:28 EST


We add LoongArch support into KVM selftests and there are some KVM
test cases we have passed:
demand_paging_test
dirty_log_perf_test
dirty_log_test
guest_print_test
kvm_binary_stats_test
kvm_create_max_vcpus
kvm_page_table_test
memslot_modification_stress_test
memslot_perf_test
set_memory_region_test

Changes for v5:
1. In LoongArch kvm self tests, the DEFAULT_GUEST_TEST_MEM could be
0x130000000, it is different from the default value in memstress.h.
So we Move the definition of DEFAULT_GUEST_TEST_MEM into LoongArch
ucall.h, and add 'ifndef' condition for DEFAULT_GUEST_TEST_MEM
in memstress.h.

Changes for v4:
1. Remove the based-on flag, as the LoongArch KVM patch series
have been accepted by Linux kernel, so this can be applied directly
in kernel.

Changes for v3:
1. Improve implementation of LoongArch VM page walk.
2. Add exception handler for LoongArch.
3. Add dirty_log_test, dirty_log_perf_test, guest_print_test
test cases for LoongArch.
4. Add __ASSEMBLER__ macro to distinguish asm file and c file.
5. Move ucall_arch_do_ucall to the header file and make it as
static inline to avoid function calls.
6. Change the DEFAULT_GUEST_TEST_MEM base addr for LoongArch.

Changes for v2:
1. We should use ".balign 4096" to align the assemble code with 4K in
exception.S instead of "align 12".
2. LoongArch only supports 3 or 4 levels page tables, so we remove the
hanlders for 2-levels page table.
3. Remove the DEFAULT_LOONGARCH_GUEST_STACK_VADDR_MIN and use the common
DEFAULT_GUEST_STACK_VADDR_MIN to allocate stack memory in guest.
4. Reorganize the test cases supported by LoongArch.
5. Fix some code comments.
6. Add kvm_binary_stats_test test case into LoongArch KVM selftests.

changes for v1:
1. Add kvm selftests header files for LoongArch.
2. Add processor tests for LoongArch KVM.
3. Add ucall tests for LoongArch KVM.
4. Add LoongArch tests into makefile.

All of the test cases results:
1..10
timeout set to 120
selftests: kvm: demand_paging_test
Testing guest mode: PA-bits:36, VA-bits:47, 16K pages
guest physical test memory: [0xfbfffc000, 0xfffffc000)
Finished creating vCPUs and starting uffd threads
Started all vCPUs
All vCPU threads joined
Total guest execution time: 0.200804700s
Overall demand paging rate: 326366.862927 pgs/sec
ok 1 selftests: kvm: demand_paging_test
timeout set to 120
selftests: kvm: dirty_log_perf_test
Test iterations: 2
Testing guest mode: PA-bits:36, VA-bits:47, 16K pages
guest physical test memory: [0xfbfffc000, 0xfffffc000)
Random seed: 1
Populate memory time: 0.201452560s
Enabling dirty logging time: 0.000451670s

Iteration 1 dirty memory time: 0.051582140s
Iteration 1 get dirty log time: 0.000010510s
Iteration 1 clear dirty log time: 0.000421730s
Iteration 2 dirty memory time: 0.046593760s
Iteration 2 get dirty log time: 0.000002110s
Iteration 2 clear dirty log time: 0.000418020s
Disabling dirty logging time: 0.002948490s
Get dirty log over 2 iterations took 0.000012620s. (Avg 0.000006310s/iteration)
Clear dirty log over 2 iterations took 0.000839750s. (Avg 0.000419875s/iteration)
ok 2 selftests: kvm: dirty_log_perf_test
timeout set to 120
selftests: kvm: dirty_log_test
Test iterations: 32, interval: 10 (ms)
Testing Log Mode 'dirty-log'
Testing guest mode: PA-bits:36, VA-bits:47, 16K pages
guest physical test memory offset: 0xfbfff0000
Dirtied 453632 pages
Total bits checked: dirty (436564), clear (1595145), track_next (70002)
Testing Log Mode 'clear-log'
Testing guest mode: PA-bits:36, VA-bits:47, 16K pages
guest physical test memory offset: 0xfbfff0000
Dirtied 425984 pages
Total bits checked: dirty (414397), clear (1617312), track_next (68152)
Testing Log Mode 'dirty-ring'
Testing guest mode: PA-bits:36, VA-bits:47, 16K pages
dirty ring count: 0x10000
guest physical test memory offset: 0xfbfff0000
vcpu stops because vcpu is kicked out...
Notifying vcpu to continue
vcpu continues now.
Iteration 1 collected 3201 pages
vcpu stops because dirty ring is full...
vcpu continues now.
vcpu stops because dirty ring is full...
Notifying vcpu to continue
Iteration 2 collected 65472 pages
......
vcpu continues now.
vcpu stops because vcpu is kicked out...
vcpu continues now.
vcpu stops because vcpu is kicked out...
Notifying vcpu to continue
vcpu continues now.
Iteration 31 collected 12642 pages
vcpu stops because dirty ring is full...
vcpu continues now.
Dirtied 7275520 pages
Total bits checked: dirty (1165675), clear (866034), track_next (811358)
ok 3 selftests: kvm: dirty_log_test
timeout set to 120
selftests: kvm: guest_print_test
ok 4 selftests: kvm: guest_print_test
timeout set to 120
selftests: kvm: kvm_binary_stats_test
TAP version 13
1..4
ok 1 vm0
ok 2 vm1
ok 3 vm2
ok 4 vm3
# Totals: pass:4 fail:0 xfail:0 xpass:0 skip:0 error:0
ok 5 selftests: kvm: kvm_binary_stats_test
timeout set to 120
selftests: kvm: kvm_create_max_vcpus
KVM_CAP_MAX_VCPU_ID: 256
KVM_CAP_MAX_VCPUS: 256
Testing creating 256 vCPUs, with IDs 0...255.
ok 6 selftests: kvm: kvm_create_max_vcpus
timeout set to 120
selftests: kvm: kvm_page_table_test
Testing guest mode: PA-bits:36, VA-bits:47, 16K pages
Testing memory backing src type: anonymous
Testing memory backing src granularity: 0x4000
Testing memory size(aligned): 0x40000000
Guest physical test memory offset: 0xfbfffc000
Host virtual test memory offset: 0x7fffb0860000
Number of testing vCPUs: 1
Started all vCPUs successfully
KVM_CREATE_MAPPINGS: total execution time: 0.200919330s

KVM_UPDATE_MAPPINGS: total execution time: 0.051182930s

KVM_ADJUST_MAPPINGS: total execution time: 0.010083590s

ok 7 selftests: kvm: kvm_page_table_test
timeout set to 120
selftests: kvm: memslot_modification_stress_test
Testing guest mode: PA-bits:36, VA-bits:47, 16K pages
guest physical test memory: [0xfbfffc000, 0xfffffc000)
Finished creating vCPUs
Started all vCPUs
All vCPU threads joined
ok 8 selftests: kvm: memslot_modification_stress_test
timeout set to 120
selftests: kvm: memslot_perf_test
Testing map performance with 1 runs, 5 seconds each
Memslot count too high for this test, decrease the cap (max is 2053)

Testing unmap performance with 1 runs, 5 seconds each
Memslot count too high for this test, decrease the cap (max is 8197)

Testing unmap chunked performance with 1 runs, 5 seconds each
Memslot count too high for this test, decrease the cap (max is 8197)

Testing move active area performance with 1 runs, 5 seconds each
Test took 0.761678900s for slot setup + 5.000014460s all iterations
Done 120167 iterations, avg 0.000041608s each
Best runtime result was 0.000041608s per iteration (with 120167 iterations)

Testing move inactive area performance with 1 runs, 5 seconds each
Test took 0.771796550s for slot setup + 5.000018520s all iterations
Done 136354 iterations, avg 0.000036669s each
Best runtime result was 0.000036669s per iteration (with 136354 iterations)

Testing RW performance with 1 runs, 5 seconds each
Test took 0.763568840s for slot setup + 5.002233800s all iterations
Done 649 iterations, avg 0.007707602s each
Best runtime result was 0.007707602s per iteration (with 649 iterations)
Best slot setup time for the whole test area was 0.761678900s
ok 9 selftests: kvm: memslot_perf_test
timeout set to 120
selftests: kvm: set_memory_region_test
Allowed number of memory slots: 32767
Adding slots 0..32766, each memory region with 2048K size
ok 10 selftests: kvm: set_memory_region_test

Tianrui Zhao (4):
KVM: selftests: Add KVM selftests header files for LoongArch
KVM: selftests: Add core KVM selftests support for LoongArch
KVM: selftests: Add ucall test support for LoongArch
KVM: selftests: Add test cases for LoongArch

tools/testing/selftests/kvm/Makefile | 15 +
.../selftests/kvm/include/kvm_util_base.h | 5 +
.../kvm/include/loongarch/processor.h | 133 +++++++
.../selftests/kvm/include/loongarch/ucall.h | 28 ++
.../testing/selftests/kvm/include/memstress.h | 2 +
.../selftests/kvm/lib/loongarch/exception.S | 59 ++++
.../selftests/kvm/lib/loongarch/processor.c | 333 ++++++++++++++++++
.../selftests/kvm/lib/loongarch/ucall.c | 38 ++
8 files changed, 613 insertions(+)
create mode 100644 tools/testing/selftests/kvm/include/loongarch/processor.h
create mode 100644 tools/testing/selftests/kvm/include/loongarch/ucall.h
create mode 100644 tools/testing/selftests/kvm/lib/loongarch/exception.S
create mode 100644 tools/testing/selftests/kvm/lib/loongarch/processor.c
create mode 100644 tools/testing/selftests/kvm/lib/loongarch/ucall.c

--
2.39.1