[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-Id: <20231121115457.76269-10-cloudliang@tencent.com>
Date: Tue, 21 Nov 2023 19:54:57 +0800
From: Jinrong Liang <ljr.kernel@...il.com>
To: Sean Christopherson <seanjc@...gle.com>
Cc: Paolo Bonzini <pbonzini@...hat.com>, Like Xu <likexu@...cent.com>,
Jim Mattson <jmattson@...gle.com>,
Aaron Lewis <aaronlewis@...gle.com>,
Wanpeng Li <wanpengli@...cent.com>,
Jinrong Liang <cloudliang@...cent.com>,
Jinrong Liang <ljr.kernel@...il.com>, kvm@...r.kernel.org,
linux-kernel@...r.kernel.org
Subject: [PATCH 9/9] KVM: selftests: Test AMD Guest PerfMonV2
From: Jinrong Liang <cloudliang@...cent.com>
Add test case for AMD Guest PerfMonV2.
Signed-off-by: Jinrong Liang <cloudliang@...cent.com>
---
.../selftests/kvm/x86_64/pmu_counters_test.c | 35 +++++++++++++++++++
1 file changed, 35 insertions(+)
diff --git a/tools/testing/selftests/kvm/x86_64/pmu_counters_test.c b/tools/testing/selftests/kvm/x86_64/pmu_counters_test.c
index ca5b352ea6ae..aa44f2282996 100644
--- a/tools/testing/selftests/kvm/x86_64/pmu_counters_test.c
+++ b/tools/testing/selftests/kvm/x86_64/pmu_counters_test.c
@@ -722,6 +722,38 @@ static void set_amd_counters(uint8_t *nr_amd_ounters, uint64_t *ctrl_msr,
}
}
+static void guest_test_amd_perfmonv2(void)
+{
+ unsigned int i;
+
+ for (i = 0; i < AMD64_NR_COUNTERS_CORE; i++) {
+ wrmsr(MSR_F15H_PERF_CTL0 + i * 2, 0);
+ wrmsr(MSR_F15H_PERF_CTR0 + i * 2, ARCH_PERFMON_EVENTSEL_OS |
+ ARCH_PERFMON_EVENTSEL_ENABLE | AMD_ZEN_CORE_CYCLES);
+
+ wrmsr(MSR_AMD64_PERF_CNTR_GLOBAL_CTL, 0);
+ __asm__ __volatile__("loop ." : "+c"((int){NUM_BRANCHES}));
+ GUEST_ASSERT(!_rdpmc(i));
+
+ wrmsr(MSR_AMD64_PERF_CNTR_GLOBAL_CTL, BIT_ULL(i));
+ __asm__ __volatile__("loop ." : "+c"((int){NUM_BRANCHES}));
+ wrmsr(MSR_AMD64_PERF_CNTR_GLOBAL_CTL, 0);
+ GUEST_ASSERT(_rdpmc(i));
+
+ wrmsr(MSR_F15H_PERF_CTL0 + i * 2, (1ULL << 48) - 2);
+ wrmsr(MSR_AMD64_PERF_CNTR_GLOBAL_STATUS_CLR, 0xff);
+ wrmsr(MSR_AMD64_PERF_CNTR_GLOBAL_CTL, BIT_ULL(i));
+ __asm__ __volatile__("loop ." : "+c"((int){NUM_BRANCHES}));
+ wrmsr(MSR_AMD64_PERF_CNTR_GLOBAL_CTL, 0);
+ GUEST_ASSERT(rdmsr(MSR_AMD64_PERF_CNTR_GLOBAL_STATUS) &
+ BIT_ULL(i));
+
+ wrmsr(MSR_AMD64_PERF_CNTR_GLOBAL_STATUS_CLR, BIT_ULL(i));
+ GUEST_ASSERT(!(rdmsr(MSR_AMD64_PERF_CNTR_GLOBAL_STATUS) &
+ BIT_ULL(i)));
+ }
+}
+
static void guest_test_amd_counters(void)
{
bool guest_pmu_is_perfmonv2 = this_cpu_has(X86_FEATURE_PERFMON_V2);
@@ -747,6 +779,9 @@ static void guest_test_amd_counters(void)
}
}
+ if (guest_pmu_is_perfmonv2)
+ guest_test_amd_perfmonv2();
+
GUEST_DONE();
}
--
2.39.3
Powered by blists - more mailing lists