lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite for Android: free password hash cracker in your pocket
[<prev] [next>] [<thread-prev] [day] [month] [year] [list]
Message-ID: <83ee6240-b270-4057-bc8d-4dcc9ddc1b9b@linux.intel.com>
Date: Mon, 16 Jun 2025 10:55:26 -0400
From: "Liang, Kan" <kan.liang@...ux.intel.com>
To: Ian Rogers <irogers@...gle.com>, Peter Zijlstra <peterz@...radead.org>,
 Ingo Molnar <mingo@...hat.com>, Arnaldo Carvalho de Melo <acme@...nel.org>,
 Namhyung Kim <namhyung@...nel.org>, Mark Rutland <mark.rutland@....com>,
 Alexander Shishkin <alexander.shishkin@...ux.intel.com>,
 Jiri Olsa <jolsa@...nel.org>, Adrian Hunter <adrian.hunter@...el.com>,
 Paul Walmsley <paul.walmsley@...ive.com>, Palmer Dabbelt
 <palmer@...belt.com>, Albert Ou <aou@...s.berkeley.edu>,
 Alexandre Ghiti <alex@...ti.fr>, Ravi Bangoria <ravi.bangoria@....com>,
 Howard Chu <howardchu95@...il.com>, linux-perf-users@...r.kernel.org,
 linux-kernel@...r.kernel.org
Subject: Re: [PATCH v1] perf test: Expand user space event reading (rdpmc)
 tests



On 2025-06-13 8:45 p.m., Ian Rogers wrote:
> Test that disabling rdpmc support via /sys/bus/event_source/cpu*/rdpmc
> disables reading in the mmap (libperf read support will fallback to
> using a system call).
> Test all hybrid PMUs support rdpmc.
> Ensure hybrid PMUs use the correct CPU to rdpmc the correct
> event. Previously the test would open cycles or instructions with no
> extended type then rdpmc it on whatever CPU. This could fail/skip due
> to which CPU the test was scheduled upon.
> 
> Signed-off-by: Ian Rogers <irogers@...gle.com>

Thanks for the fix.

Reviewed-by: Kan Liang <kan.liang@...ux.intel.com>

Thanks,
Kan> ---
>  tools/perf/tests/mmap-basic.c | 289 +++++++++++++++++++++++++---------
>  tools/perf/tests/tests.h      |   9 ++
>  tools/perf/util/affinity.c    |  18 +++
>  tools/perf/util/affinity.h    |   2 +
>  4 files changed, 240 insertions(+), 78 deletions(-)
> 
> diff --git a/tools/perf/tests/mmap-basic.c b/tools/perf/tests/mmap-basic.c
> index 04b547c6bdbe..3c89d3001887 100644
> --- a/tools/perf/tests/mmap-basic.c
> +++ b/tools/perf/tests/mmap-basic.c
> @@ -1,15 +1,18 @@
>  // SPDX-License-Identifier: GPL-2.0
>  #include <errno.h>
> +#include <fcntl.h>
>  #include <inttypes.h>
>  #include <stdlib.h>
>  #include <perf/cpumap.h>
>  
> +#include "cpumap.h"
>  #include "debug.h"
>  #include "event.h"
>  #include "evlist.h"
>  #include "evsel.h"
>  #include "thread_map.h"
>  #include "tests.h"
> +#include "util/affinity.h"
>  #include "util/mmap.h"
>  #include "util/sample.h"
>  #include <linux/err.h>
> @@ -172,99 +175,199 @@ static int test__basic_mmap(struct test_suite *test __maybe_unused, int subtest
>  	return err;
>  }
>  
> -static int test_stat_user_read(int event)
> -{
> -	struct perf_counts_values counts = { .val = 0 };
> -	struct perf_thread_map *threads;
> -	struct perf_evsel *evsel;
> -	struct perf_event_mmap_page *pc;
> -	struct perf_event_attr attr = {
> -		.type	= PERF_TYPE_HARDWARE,
> -		.config	= event,
> -#ifdef __aarch64__
> -		.config1 = 0x2,		/* Request user access */
> -#endif
> -	};
> -	int err, i, ret = TEST_FAIL;
> -	bool opened = false, mapped = false;
> +enum user_read_state {
> +	USER_READ_ENABLED,
> +	USER_READ_DISABLED,
> +	USER_READ_UNKNOWN,
> +};
>  
> -	threads = perf_thread_map__new_dummy();
> -	TEST_ASSERT_VAL("failed to create threads", threads);
> +static enum user_read_state set_user_read(struct perf_pmu *pmu, enum user_read_state enabled)
> +{
> +	char buf[2] = {0, '\n'};
> +	ssize_t len;
> +	int events_fd, rdpmc_fd;
> +	enum user_read_state old_user_read = USER_READ_UNKNOWN;
> +
> +	if (enabled == USER_READ_UNKNOWN)
> +		return USER_READ_UNKNOWN;
> +
> +	events_fd = perf_pmu__event_source_devices_fd();
> +	if (events_fd < 0)
> +		return USER_READ_UNKNOWN;
> +
> +	rdpmc_fd = perf_pmu__pathname_fd(events_fd, pmu->name, "rdpmc", O_RDWR);
> +	if (rdpmc_fd < 0) {
> +		close(events_fd);
> +		return USER_READ_UNKNOWN;
> +	}
>  
> -	perf_thread_map__set_pid(threads, 0, 0);
> +	len = read(rdpmc_fd, buf, sizeof(buf));
> +	if (len != sizeof(buf))
> +		pr_debug("%s read failed\n", __func__);
>  
> -	evsel = perf_evsel__new(&attr);
> -	TEST_ASSERT_VAL("failed to create evsel", evsel);
> +	// Note, on Intel hybrid disabling on 1 PMU will implicitly disable on
> +	// all the core PMUs.
> +	old_user_read = (buf[0] == '1') ? USER_READ_ENABLED : USER_READ_DISABLED;
>  
> -	err = perf_evsel__open(evsel, NULL, threads);
> -	if (err) {
> -		pr_err("failed to open evsel: %s\n", strerror(-err));
> -		ret = TEST_SKIP;
> -		goto out;
> +	if (enabled != old_user_read) {
> +		buf[0] = (enabled == USER_READ_ENABLED) ? '1' : '0';
> +		len = write(rdpmc_fd, buf, sizeof(buf));
> +		if (len != sizeof(buf))
> +			pr_debug("%s write failed\n", __func__);
>  	}
> -	opened = true;
> +	close(rdpmc_fd);
> +	close(events_fd);
> +	return old_user_read;
> +}
>  
> -	err = perf_evsel__mmap(evsel, 0);
> -	if (err) {
> -		pr_err("failed to mmap evsel: %s\n", strerror(-err));
> -		goto out;
> +static int test_stat_user_read(u64 event, enum user_read_state enabled)
> +{
> +	struct perf_pmu *pmu = NULL;
> +	struct perf_thread_map *threads = perf_thread_map__new_dummy();
> +	int ret = TEST_OK;
> +
> +	pr_err("User space counter reading %" PRIu64 "\n", event);
> +	if (!threads) {
> +		pr_err("User space counter reading [Failed to create threads]\n");
> +		return TEST_FAIL;
>  	}
> -	mapped = true;
> +	perf_thread_map__set_pid(threads, 0, 0);
>  
> -	pc = perf_evsel__mmap_base(evsel, 0, 0);
> -	if (!pc) {
> -		pr_err("failed to get mmapped address\n");
> -		goto out;
> -	}
> +	while ((pmu = perf_pmus__scan_core(pmu)) != NULL) {
> +		enum user_read_state saved_user_read_state = set_user_read(pmu, enabled);
> +		struct perf_event_attr attr = {
> +			.type	= PERF_TYPE_HARDWARE,
> +			.config	= perf_pmus__supports_extended_type()
> +			? event | ((u64)pmu->type << PERF_PMU_TYPE_SHIFT)
> +				: event,
> +#ifdef __aarch64__
> +			.config1 = 0x2,		/* Request user access */
> +#endif
> +		};
> +		struct perf_evsel *evsel = NULL;
> +		int err;
> +		struct perf_event_mmap_page *pc;
> +		bool mapped = false, opened = false, rdpmc_supported;
> +		struct perf_counts_values counts = { .val = 0 };
> +
> +
> +		pr_debug("User space counter reading for PMU %s\n", pmu->name);
> +		/*
> +		 * Restrict scheduling to only use the rdpmc on the CPUs the
> +		 * event can be on. If the test doesn't run on the CPU of the
> +		 * event then the event will be disabled and the pc->index test
> +		 * will fail.
> +		 */
> +		if (pmu->cpus != NULL)
> +			cpu_map__set_affinity(pmu->cpus);
> +
> +		/* Make the evsel. */
> +		evsel = perf_evsel__new(&attr);
> +		if (!evsel) {
> +			pr_err("User space counter reading for PMU %s [Failed to allocate evsel]\n",
> +				pmu->name);
> +			ret = TEST_FAIL;
> +			goto cleanup;
> +		}
>  
> -	if (!pc->cap_user_rdpmc || !pc->index) {
> -		pr_err("userspace counter access not %s\n",
> -			!pc->cap_user_rdpmc ? "supported" : "enabled");
> -		ret = TEST_SKIP;
> -		goto out;
> -	}
> -	if (pc->pmc_width < 32) {
> -		pr_err("userspace counter width not set (%d)\n", pc->pmc_width);
> -		goto out;
> -	}
> +		err = perf_evsel__open(evsel, NULL, threads);
> +		if (err) {
> +			pr_err("User space counter reading for PMU %s [Failed to open evsel]\n",
> +				pmu->name);
> +			ret = TEST_SKIP;
> +			goto cleanup;
> +		}
> +		opened = true;
> +		err = perf_evsel__mmap(evsel, 0);
> +		if (err) {
> +			pr_err("User space counter reading for PMU %s [Failed to mmap evsel]\n",
> +				pmu->name);
> +			ret = TEST_FAIL;
> +			goto cleanup;
> +		}
> +		mapped = true;
> +
> +		pc = perf_evsel__mmap_base(evsel, 0, 0);
> +		if (!pc) {
> +			pr_err("User space counter reading for PMU %s [Failed to get mmaped address]\n",
> +				pmu->name);
> +			ret = TEST_FAIL;
> +			goto cleanup;
> +		}
>  
> -	perf_evsel__read(evsel, 0, 0, &counts);
> -	if (counts.val == 0) {
> -		pr_err("failed to read value for evsel\n");
> -		goto out;
> -	}
> +		if (saved_user_read_state == USER_READ_UNKNOWN)
> +			rdpmc_supported = pc->cap_user_rdpmc && pc->index;
> +		else
> +			rdpmc_supported = (enabled == USER_READ_ENABLED);
>  
> -	for (i = 0; i < 5; i++) {
> -		volatile int count = 0x10000 << i;
> -		__u64 start, end, last = 0;
> +		if (rdpmc_supported && (!pc->cap_user_rdpmc || !pc->index)) {
> +			pr_err("User space counter reading for PMU %s [Failed unexpected supported counter access %d %d]\n",
> +				pmu->name, pc->cap_user_rdpmc, pc->index);
> +			ret = TEST_FAIL;
> +			goto cleanup;
> +		}
>  
> -		pr_debug("\tloop = %u, ", count);
> +		if (!rdpmc_supported && pc->cap_user_rdpmc) {
> +			pr_err("User space counter reading for PMU %s [Failed unexpected unsupported counter access %d]\n",
> +				pmu->name, pc->cap_user_rdpmc);
> +			ret = TEST_FAIL;
> +			goto cleanup;
> +		}
> +
> +		if (rdpmc_supported && pc->pmc_width < 32) {
> +			pr_err("User space counter reading for PMU %s [Failed width not set %d]\n",
> +				pmu->name, pc->pmc_width);
> +			ret = TEST_FAIL;
> +			goto cleanup;
> +		}
>  
>  		perf_evsel__read(evsel, 0, 0, &counts);
> -		start = counts.val;
> +		if (counts.val == 0) {
> +			pr_err("User space counter reading for PMU %s [Failed read]\n", pmu->name);
> +			ret = TEST_FAIL;
> +			goto cleanup;
> +		}
>  
> -		while (count--) ;
> +		for (int i = 0; i < 5; i++) {
> +			volatile int count = 0x10000 << i;
> +			__u64 start, end, last = 0;
>  
> -		perf_evsel__read(evsel, 0, 0, &counts);
> -		end = counts.val;
> +			pr_debug("\tloop = %u, ", count);
>  
> -		if ((end - start) < last) {
> -			pr_err("invalid counter data: end=%llu start=%llu last= %llu\n",
> -				end, start, last);
> -			goto out;
> -		}
> -		last = end - start;
> -		pr_debug("count = %llu\n", end - start);
> -	}
> -	ret = TEST_OK;
> +			perf_evsel__read(evsel, 0, 0, &counts);
> +			start = counts.val;
> +
> +			while (count--) ;
>  
> -out:
> -	if (mapped)
> -		perf_evsel__munmap(evsel);
> -	if (opened)
> -		perf_evsel__close(evsel);
> -	perf_evsel__delete(evsel);
> +			perf_evsel__read(evsel, 0, 0, &counts);
> +			end = counts.val;
>  
> +			if ((end - start) < last) {
> +				pr_err("User space counter reading for PMU %s [Failed invalid counter data: end=%llu start=%llu last= %llu]\n",
> +					pmu->name, end, start, last);
> +				ret = TEST_FAIL;
> +				goto cleanup;
> +			}
> +			last = end - start;
> +			pr_debug("count = %llu\n", last);
> +		}
> +		pr_debug("User space counter reading for PMU %s [Success]\n", pmu->name);
> +cleanup:
> +		if (mapped)
> +			perf_evsel__munmap(evsel);
> +		if (opened)
> +			perf_evsel__close(evsel);
> +		perf_evsel__delete(evsel);
> +
> +		/* If the affinity was changed, then put it back to all CPUs. */
> +		if (pmu->cpus != NULL) {
> +			struct perf_cpu_map *cpus = cpu_map__online();
> +
> +			cpu_map__set_affinity(cpus);
> +			perf_cpu_map__put(cpus);
> +		}
> +		set_user_read(pmu, saved_user_read_state);
> +	}
>  	perf_thread_map__put(threads);
>  	return ret;
>  }
> @@ -272,20 +375,32 @@ static int test_stat_user_read(int event)
>  static int test__mmap_user_read_instr(struct test_suite *test __maybe_unused,
>  				      int subtest __maybe_unused)
>  {
> -	return test_stat_user_read(PERF_COUNT_HW_INSTRUCTIONS);
> +	return test_stat_user_read(PERF_COUNT_HW_INSTRUCTIONS, USER_READ_ENABLED);
>  }
>  
>  static int test__mmap_user_read_cycles(struct test_suite *test __maybe_unused,
>  				       int subtest __maybe_unused)
>  {
> -	return test_stat_user_read(PERF_COUNT_HW_CPU_CYCLES);
> +	return test_stat_user_read(PERF_COUNT_HW_CPU_CYCLES, USER_READ_ENABLED);
> +}
> +
> +static int test__mmap_user_read_instr_disabled(struct test_suite *test __maybe_unused,
> +					       int subtest __maybe_unused)
> +{
> +	return test_stat_user_read(PERF_COUNT_HW_INSTRUCTIONS, USER_READ_DISABLED);
> +}
> +
> +static int test__mmap_user_read_cycles_disabled(struct test_suite *test __maybe_unused,
> +						int subtest __maybe_unused)
> +{
> +	return test_stat_user_read(PERF_COUNT_HW_CPU_CYCLES, USER_READ_DISABLED);
>  }
>  
>  static struct test_case tests__basic_mmap[] = {
>  	TEST_CASE_REASON("Read samples using the mmap interface",
>  			 basic_mmap,
>  			 "permissions"),
> -	TEST_CASE_REASON("User space counter reading of instructions",
> +	TEST_CASE_REASON_EXCLUSIVE("User space counter reading of instructions",
>  			 mmap_user_read_instr,
>  #if defined(__i386__) || defined(__x86_64__) || defined(__aarch64__) || \
>  			 (defined(__riscv) && __riscv_xlen == 64)
> @@ -294,13 +409,31 @@ static struct test_case tests__basic_mmap[] = {
>  			 "unsupported"
>  #endif
>  		),
> -	TEST_CASE_REASON("User space counter reading of cycles",
> +	TEST_CASE_REASON_EXCLUSIVE("User space counter reading of cycles",
>  			 mmap_user_read_cycles,
>  #if defined(__i386__) || defined(__x86_64__) || defined(__aarch64__) || \
>  			 (defined(__riscv) && __riscv_xlen == 64)
>  			 "permissions"
>  #else
>  			 "unsupported"
> +#endif
> +		),
> +	TEST_CASE_REASON_EXCLUSIVE("User space counter disabling instructions",
> +			 mmap_user_read_instr_disabled,
> +#if defined(__i386__) || defined(__x86_64__) || defined(__aarch64__) || \
> +			 (defined(__riscv) && __riscv_xlen == 64)
> +			 "permissions"
> +#else
> +			 "unsupported"
> +#endif
> +		),
> +	TEST_CASE_REASON_EXCLUSIVE("User space counter disabling cycles",
> +			 mmap_user_read_cycles_disabled,
> +#if defined(__i386__) || defined(__x86_64__) || defined(__aarch64__) || \
> +			 (defined(__riscv) && __riscv_xlen == 64)
> +			 "permissions"
> +#else
> +			 "unsupported"
>  #endif
>  		),
>  	{	.name = NULL, }
> diff --git a/tools/perf/tests/tests.h b/tools/perf/tests/tests.h
> index bb7951c61971..4c128a959441 100644
> --- a/tools/perf/tests/tests.h
> +++ b/tools/perf/tests/tests.h
> @@ -71,6 +71,15 @@ struct test_suite {
>  		.exclusive = true,			\
>  	}
>  
> +#define TEST_CASE_REASON_EXCLUSIVE(description, _name, _reason)	\
> +	{						\
> +		.name = #_name,				\
> +		.desc = description,			\
> +		.run_case = test__##_name,		\
> +		.skip_reason = _reason,			\
> +		.exclusive = true,			\
> +	}
> +
>  #define DEFINE_SUITE(description, _name)		\
>  	struct test_case tests__##_name[] = {           \
>  		TEST_CASE(description, _name),		\
> diff --git a/tools/perf/util/affinity.c b/tools/perf/util/affinity.c
> index 38dc4524b7e8..4fe851334296 100644
> --- a/tools/perf/util/affinity.c
> +++ b/tools/perf/util/affinity.c
> @@ -5,6 +5,7 @@
>  #include <stdlib.h>
>  #include <linux/bitmap.h>
>  #include <linux/zalloc.h>
> +#include <perf/cpumap.h>
>  #include "perf.h"
>  #include "cpumap.h"
>  #include "affinity.h"
> @@ -83,3 +84,20 @@ void affinity__cleanup(struct affinity *a)
>  	if (a != NULL)
>  		__affinity__cleanup(a);
>  }
> +
> +void cpu_map__set_affinity(const struct perf_cpu_map *cpumap)
> +{
> +	int cpu_set_size = get_cpu_set_size();
> +	unsigned long *cpuset = bitmap_zalloc(cpu_set_size * 8);
> +	struct perf_cpu cpu;
> +	int idx;
> +
> +	if (!cpuset)
> +		return;
> +
> +	perf_cpu_map__for_each_cpu_skip_any(cpu, idx, cpumap)
> +		__set_bit(cpu.cpu, cpuset);
> +
> +	sched_setaffinity(0, cpu_set_size, (cpu_set_t *)cpuset);
> +	zfree(&cpuset);
> +}
> diff --git a/tools/perf/util/affinity.h b/tools/perf/util/affinity.h
> index 0ad6a18ef20c..7341194b2298 100644
> --- a/tools/perf/util/affinity.h
> +++ b/tools/perf/util/affinity.h
> @@ -4,6 +4,7 @@
>  
>  #include <stdbool.h>
>  
> +struct perf_cpu_map;
>  struct affinity {
>  	unsigned long *orig_cpus;
>  	unsigned long *sched_cpus;
> @@ -13,5 +14,6 @@ struct affinity {
>  void affinity__cleanup(struct affinity *a);
>  void affinity__set(struct affinity *a, int cpu);
>  int affinity__setup(struct affinity *a);
> +void cpu_map__set_affinity(const struct perf_cpu_map *cpumap);
>  
>  #endif // PERF_AFFINITY_H


Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ