lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <CAP-5=fXE02soq4z0ZjuPYQZTH9bkfrMXqe1nT9A7EcgG0cH6yw@mail.gmail.com>
Date: Mon, 12 Jan 2026 17:49:33 -0800
From: Ian Rogers <irogers@...gle.com>
To: Dapeng Mi <dapeng1.mi@...ux.intel.com>
Cc: Peter Zijlstra <peterz@...radead.org>, Ingo Molnar <mingo@...hat.com>, 
	Arnaldo Carvalho de Melo <acme@...nel.org>, Namhyung Kim <namhyung@...nel.org>, 
	Adrian Hunter <adrian.hunter@...el.com>, 
	Alexander Shishkin <alexander.shishkin@...ux.intel.com>, Andi Kleen <ak@...ux.intel.com>, 
	Eranian Stephane <eranian@...gle.com>, linux-kernel@...r.kernel.org, 
	linux-perf-users@...r.kernel.org, Dapeng Mi <dapeng1.mi@...el.com>, 
	Zide Chen <zide.chen@...el.com>, Falcon Thomas <thomas.falcon@...el.com>, 
	Xudong Hao <xudong.hao@...el.com>
Subject: Re: [Patch v2 7/7] perf/x86/intel: Add support for rdpmc user disable feature

On Sun, Jan 11, 2026 at 9:20 PM Dapeng Mi <dapeng1.mi@...ux.intel.com> wrote:
>
> Starting with Panther Cove, the rdpmc user disable feature is supported.
> This feature allows the perf system to disable user space rdpmc reads at
> the counter level.
>
> Currently, when a global counter is active, any user with rdpmc rights
> can read it, even if perf access permissions forbid it (e.g., disallow
> reading ring 0 counters). The rdpmc user disable feature mitigates this
> security concern.
>
> Details:
>
> - A new RDPMC_USR_DISABLE bit (bit 37) in each EVNTSELx MSR indicates
>   that the GP counter cannot be read by RDPMC in ring 3.
> - New RDPMC_USR_DISABLE bits in IA32_FIXED_CTR_CTRL MSR (bits 33, 37,
>   41, 45, etc.) for fixed counters 0, 1, 2, 3, etc.
> - When calling rdpmc instruction for counter x, the following pseudo
>   code demonstrates how the counter value is obtained:
>         If (!CPL0 && RDPMC_USR_DISABLE[x] == 1) ? 0 : counter_value;
> - RDPMC_USR_DISABLE is enumerated by CPUID.0x23.0.EBX[2].
>
> This patch extends the current global user space rdpmc control logic via
> the sysfs interface (/sys/devices/cpu/rdpmc) as follows:
>
> - rdpmc = 0:
>   Global user space rdpmc and counter-level user space rdpmc for all
>   counters are both disabled.
> - rdpmc = 1:
>   Global user space rdpmc is enabled during the mmap-enabled time window,
>   and counter-level user space rdpmc is enabled only for non-system-wide
>   events. This prevents counter data leaks as count data is cleared
>   during context switches.
> - rdpmc = 2:
>   Global user space rdpmc and counter-level user space rdpmc for all
>   counters are enabled unconditionally.
>
> The new rdpmc settings only affect newly activated perf events; currently
> active perf events remain unaffected. This simplifies and cleans up the
> code. The default value of rdpmc remains unchanged at 1.
>
> For more details about rdpmc user disable, please refer to chapter 15
> "RDPMC USER DISABLE" in ISE documentation.
>
> ISE: https://www.intel.com/content/www/us/en/content-details/869288/intel-architecture-instruction-set-extensions-programming-reference.html
>
> Signed-off-by: Dapeng Mi <dapeng1.mi@...ux.intel.com>
> ---
>  .../sysfs-bus-event_source-devices-rdpmc      | 40 +++++++++++++++++++
>  arch/x86/events/core.c                        | 21 ++++++++++
>  arch/x86/events/intel/core.c                  | 26 ++++++++++++
>  arch/x86/events/perf_event.h                  |  6 +++
>  arch/x86/include/asm/perf_event.h             |  8 +++-
>  5 files changed, 99 insertions(+), 2 deletions(-)
>  create mode 100644 Documentation/ABI/testing/sysfs-bus-event_source-devices-rdpmc
>
> diff --git a/Documentation/ABI/testing/sysfs-bus-event_source-devices-rdpmc b/Documentation/ABI/testing/sysfs-bus-event_source-devices-rdpmc
> new file mode 100644
> index 000000000000..d004527ab13e
> --- /dev/null
> +++ b/Documentation/ABI/testing/sysfs-bus-event_source-devices-rdpmc
> @@ -0,0 +1,40 @@
> +What:           /sys/bus/event_source/devices/cpu.../rdpmc
> +Date:           November 2011
> +KernelVersion:  3.10
> +Contact:        Linux kernel mailing list linux-kernel@...r.kernel.org
> +Description:    The /sys/bus/event_source/devices/cpu.../rdpmc attribute
> +                is used to show/manage if rdpmc instruction can be
> +                executed in user space. This attribute supports 3 numbers.
> +                - rdpmc = 0
> +                user space rdpmc is globally disabled for all PMU
> +                counters.
> +                - rdpmc = 1
> +                user space rdpmc is globally enabled only in event mmap
> +                ioctl called time window. If the mmap region is unmapped,
> +                user space rdpmc is disabled again.
> +                - rdpmc = 2
> +                user space rdpmc is globally enabled for all PMU
> +                counters.

Fwiw, I found it surprising in the test:
https://web.git.kernel.org/pub/scm/linux/kernel/git/perf/perf-tools-next.git/tree/tools/perf/tests/mmap-basic.c?h=perf-tools-next#n375
that the enable/disable of rdpmc on cpu_atom affected cpu_core and
vice-versa. Perhaps the docs could mention this.

Also fwiw, I remember Peter's proposal to improve rdpmc so that
restartable sequences or CPU affinities aren't necessary on hybrid
machines by handling faults in the kernel:
https://lore.kernel.org/linux-perf-users/20250618084522.GE1613376@noisy.programming.kicks-ass.net/
which imo would be a welcome addition. Perhaps without that fix we can
document the affinity/rseq needs.

Thanks,
Ian

> +
> +                In the Intel platforms supporting counter level's user
> +                space rdpmc disable feature (CPUID.23H.EBX[2] = 1), the
> +                meaning of 3 numbers is extended to
> +                - rdpmc = 0
> +                global user space rdpmc and counter level's user space
> +                rdpmc of all counters are both disabled.
> +                - rdpmc = 1
> +                No changes on behavior of global user space rdpmc.
> +                counter level's rdpmc of system-wide events is disabled
> +                but counter level's rdpmc of non-system-wide events is
> +                enabled.
> +                - rdpmc = 2
> +                global user space rdpmc and counter level's user space
> +                rdpmc of all counters are both enabled unconditionally.
> +
> +                The default value of rdpmc is 1.
> +
> +                Please notice global user space rdpmc's behavior would
> +                change immediately along with the rdpmc value's change,
> +                but the behavior of counter level's user space rdpmc
> +                won't take effect immediately until the event is
> +                reactivated or recreated.
> diff --git a/arch/x86/events/core.c b/arch/x86/events/core.c
> index c2717cb5034f..6df73e8398cd 100644
> --- a/arch/x86/events/core.c
> +++ b/arch/x86/events/core.c
> @@ -2616,6 +2616,27 @@ static ssize_t get_attr_rdpmc(struct device *cdev,
>         return snprintf(buf, 40, "%d\n", x86_pmu.attr_rdpmc);
>  }
>
> +/*
> + * Behaviors of rdpmc value:
> + * - rdpmc = 0
> + *    global user space rdpmc and counter level's user space rdpmc of all
> + *    counters are both disabled.
> + * - rdpmc = 1
> + *    global user space rdpmc is enabled in mmap enabled time window and
> + *    counter level's user space rdpmc is enabled for only non system-wide
> + *    events. Counter level's user space rdpmc of system-wide events is
> + *    still disabled by default. This won't introduce counter data leak for
> + *    non system-wide events since their count data would be cleared when
> + *    context switches.
> + * - rdpmc = 2
> + *    global user space rdpmc and counter level's user space rdpmc of all
> + *    counters are enabled unconditionally.
> + *
> + * Suppose the rdpmc value won't be changed frequently, don't dynamically
> + * reschedule events to make the new rpdmc value take effect on active perf
> + * events immediately, the new rdpmc value would only impact the new
> + * activated perf events. This makes code simpler and cleaner.
> + */
>  static ssize_t set_attr_rdpmc(struct device *cdev,
>                               struct device_attribute *attr,
>                               const char *buf, size_t count)
> diff --git a/arch/x86/events/intel/core.c b/arch/x86/events/intel/core.c
> index dd488a095f33..77cf849a1381 100644
> --- a/arch/x86/events/intel/core.c
> +++ b/arch/x86/events/intel/core.c
> @@ -3128,6 +3128,8 @@ static void intel_pmu_enable_fixed(struct perf_event *event)
>                 bits |= INTEL_FIXED_0_USER;
>         if (hwc->config & ARCH_PERFMON_EVENTSEL_OS)
>                 bits |= INTEL_FIXED_0_KERNEL;
> +       if (hwc->config & ARCH_PERFMON_EVENTSEL_RDPMC_USER_DISABLE)
> +               bits |= INTEL_FIXED_0_RDPMC_USER_DISABLE;
>
>         /*
>          * ANY bit is supported in v3 and up
> @@ -3263,6 +3265,26 @@ static void intel_pmu_enable_event_ext(struct perf_event *event)
>                 __intel_pmu_update_event_ext(hwc->idx, ext);
>  }
>
> +static void intel_pmu_update_rdpmc_user_disable(struct perf_event *event)
> +{
> +       /*
> +        * Counter scope's user-space rdpmc is disabled by default
> +        * except two cases.
> +        * a. rdpmc = 2 (user space rdpmc enabled unconditionally)
> +        * b. rdpmc = 1 and the event is not a system-wide event.
> +        *    The count of non-system-wide events would be cleared when
> +        *    context switches, so no count data is leaked.
> +        */
> +       if (x86_pmu_has_rdpmc_user_disable(event->pmu)) {
> +               if (x86_pmu.attr_rdpmc == X86_USER_RDPMC_ALWAYS_ENABLE ||
> +                   (x86_pmu.attr_rdpmc == X86_USER_RDPMC_CONDITIONAL_ENABLE &&
> +                    event->ctx->task))
> +                       event->hw.config &= ~ARCH_PERFMON_EVENTSEL_RDPMC_USER_DISABLE;
> +               else
> +                       event->hw.config |= ARCH_PERFMON_EVENTSEL_RDPMC_USER_DISABLE;
> +       }
> +}
> +
>  DEFINE_STATIC_CALL_NULL(intel_pmu_enable_event_ext, intel_pmu_enable_event_ext);
>
>  static void intel_pmu_enable_event(struct perf_event *event)
> @@ -3271,6 +3293,8 @@ static void intel_pmu_enable_event(struct perf_event *event)
>         struct hw_perf_event *hwc = &event->hw;
>         int idx = hwc->idx;
>
> +       intel_pmu_update_rdpmc_user_disable(event);
> +
>         if (unlikely(event->attr.precise_ip))
>                 static_call(x86_pmu_pebs_enable)(event);
>
> @@ -5863,6 +5887,8 @@ static void update_pmu_cap(struct pmu *pmu)
>                 hybrid(pmu, config_mask) |= ARCH_PERFMON_EVENTSEL_UMASK2;
>         if (ebx_0.split.eq)
>                 hybrid(pmu, config_mask) |= ARCH_PERFMON_EVENTSEL_EQ;
> +       if (ebx_0.split.rdpmc_user_disable)
> +               hybrid(pmu, config_mask) |= ARCH_PERFMON_EVENTSEL_RDPMC_USER_DISABLE;
>
>         if (eax_0.split.cntr_subleaf) {
>                 cpuid_count(ARCH_PERFMON_EXT_LEAF, ARCH_PERFMON_NUM_COUNTER_LEAF,
> diff --git a/arch/x86/events/perf_event.h b/arch/x86/events/perf_event.h
> index 24a81d2916e9..cd337f3ffd01 100644
> --- a/arch/x86/events/perf_event.h
> +++ b/arch/x86/events/perf_event.h
> @@ -1333,6 +1333,12 @@ static inline u64 x86_pmu_get_event_config(struct perf_event *event)
>         return event->attr.config & hybrid(event->pmu, config_mask);
>  }
>
> +static inline bool x86_pmu_has_rdpmc_user_disable(struct pmu *pmu)
> +{
> +       return !!(hybrid(pmu, config_mask) &
> +                ARCH_PERFMON_EVENTSEL_RDPMC_USER_DISABLE);
> +}
> +
>  extern struct event_constraint emptyconstraint;
>
>  extern struct event_constraint unconstrained;
> diff --git a/arch/x86/include/asm/perf_event.h b/arch/x86/include/asm/perf_event.h
> index 0d9af4135e0a..ff5acb8b199b 100644
> --- a/arch/x86/include/asm/perf_event.h
> +++ b/arch/x86/include/asm/perf_event.h
> @@ -33,6 +33,7 @@
>  #define ARCH_PERFMON_EVENTSEL_CMASK                    0xFF000000ULL
>  #define ARCH_PERFMON_EVENTSEL_BR_CNTR                  (1ULL << 35)
>  #define ARCH_PERFMON_EVENTSEL_EQ                       (1ULL << 36)
> +#define ARCH_PERFMON_EVENTSEL_RDPMC_USER_DISABLE       (1ULL << 37)
>  #define ARCH_PERFMON_EVENTSEL_UMASK2                   (0xFFULL << 40)
>
>  #define INTEL_FIXED_BITS_STRIDE                        4
> @@ -40,6 +41,7 @@
>  #define INTEL_FIXED_0_USER                             (1ULL << 1)
>  #define INTEL_FIXED_0_ANYTHREAD                        (1ULL << 2)
>  #define INTEL_FIXED_0_ENABLE_PMI                       (1ULL << 3)
> +#define INTEL_FIXED_0_RDPMC_USER_DISABLE               (1ULL << 33)
>  #define INTEL_FIXED_3_METRICS_CLEAR                    (1ULL << 2)
>
>  #define HSW_IN_TX                                      (1ULL << 32)
> @@ -50,7 +52,7 @@
>  #define INTEL_FIXED_BITS_MASK                                  \
>         (INTEL_FIXED_0_KERNEL | INTEL_FIXED_0_USER |            \
>          INTEL_FIXED_0_ANYTHREAD | INTEL_FIXED_0_ENABLE_PMI |   \
> -        ICL_FIXED_0_ADAPTIVE)
> +        ICL_FIXED_0_ADAPTIVE | INTEL_FIXED_0_RDPMC_USER_DISABLE)
>
>  #define intel_fixed_bits_by_idx(_idx, _bits)                   \
>         ((_bits) << ((_idx) * INTEL_FIXED_BITS_STRIDE))
> @@ -226,7 +228,9 @@ union cpuid35_ebx {
>                 unsigned int    umask2:1;
>                 /* EQ-bit Supported */
>                 unsigned int    eq:1;
> -               unsigned int    reserved:30;
> +               /* rdpmc user disable Supported */
> +               unsigned int    rdpmc_user_disable:1;
> +               unsigned int    reserved:29;
>         } split;
>         unsigned int            full;
>  };
> --
> 2.34.1
>

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ