[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <92465f174413553410aaa876223817b156808800.1689748843.git.sandipan.das@amd.com>
Date: Wed, 19 Jul 2023 12:25:39 +0530
From: Sandipan Das <sandipan.das@....com>
To: <linux-kernel@...r.kernel.org>, <linux-perf-users@...r.kernel.org>,
<x86@...nel.org>
CC: <peterz@...radead.org>, <mingo@...hat.com>, <acme@...nel.org>,
<mark.rutland@....com>, <alexander.shishkin@...ux.intel.com>,
<jolsa@...nel.org>, <namhyung@...nel.org>, <irogers@...gle.com>,
<adrian.hunter@...el.com>, <tglx@...utronix.de>, <bp@...en8.de>,
<dave.hansen@...ux.intel.com>, <hpa@...or.com>,
<eranian@...gle.com>, <ananth.narayan@....com>,
<ravi.bangoria@....com>, <santosh.shukla@....com>,
<sandipan.das@....com>
Subject: [PATCH 4/6] perf/x86/amd/uncore: Add group exclusivity
In some cases, it may be necessary to restrict opening PMU events to a
subset of CPUs. Uncore PMUs which require this restriction can use the
new group attribute in struct amd_uncore to set a valid uncore ID during
initialization. In the starting phase of hotplug, the per-CPU context
will be marked as unused if the ID in group does not match the uncore ID
for the CPU that is being onlined.
E.g. the Zen 4 memory controller (UMC) PMUs are specific to each active
memory channel and the MSR address space for the PERF_CTL and PERF_CTR
registers is reused on each socket. Thus, PMU events corresponding to a
memory controller should only be opened on CPUs belonging to the socket
associated with that memory controller.
Signed-off-by: Sandipan Das <sandipan.das@....com>
---
arch/x86/events/amd/uncore.c | 23 +++++++++++++++++++++--
1 file changed, 21 insertions(+), 2 deletions(-)
diff --git a/arch/x86/events/amd/uncore.c b/arch/x86/events/amd/uncore.c
index f17df6574ba5..6653e8e164bd 100644
--- a/arch/x86/events/amd/uncore.c
+++ b/arch/x86/events/amd/uncore.c
@@ -50,6 +50,7 @@ struct amd_uncore {
int num_counters;
int rdpmc_base;
u32 msr_base;
+ int group;
cpumask_t active_mask;
struct pmu pmu;
struct amd_uncore_ctx * __percpu *ctx;
@@ -423,6 +424,17 @@ static int amd_uncore_cpu_starting(unsigned int cpu)
uncore = &uncores[i];
ctx = *per_cpu_ptr(uncore->ctx, cpu);
ctx->id = uncore->id(cpu);
+
+ /*
+ * Reclaim the context if events can only be opened by CPUs
+ * within the same group
+ */
+ if (uncore->group >= 0 && ctx->id != uncore->group) {
+ hlist_add_head(&ctx->node, &uncore_unused_list);
+ *per_cpu_ptr(uncore->ctx, cpu) = NULL;
+ continue;
+ }
+
ctx = amd_uncore_find_online_sibling(ctx, uncore);
*per_cpu_ptr(uncore->ctx, cpu) = ctx;
}
@@ -453,7 +465,7 @@ static int amd_uncore_cpu_online(unsigned int cpu)
for (i = 0; i < num_uncores; i++) {
uncore = &uncores[i];
ctx = *per_cpu_ptr(uncore->ctx, cpu);
- if (cpu == ctx->cpu)
+ if (ctx && cpu == ctx->cpu)
cpumask_set_cpu(cpu, &uncore->active_mask);
}
@@ -469,12 +481,14 @@ static int amd_uncore_cpu_down_prepare(unsigned int cpu)
for (i = 0; i < num_uncores; i++) {
uncore = &uncores[i];
this = *per_cpu_ptr(uncore->ctx, cpu);
+ if (!this)
+ continue;
/* this cpu is going down, migrate to a shared sibling if possible */
for_each_online_cpu(j) {
that = *per_cpu_ptr(uncore->ctx, j);
- if (cpu == j)
+ if (!that || cpu == j)
continue;
if (this == that) {
@@ -499,6 +513,9 @@ static int amd_uncore_cpu_dead(unsigned int cpu)
for (i = 0; i < num_uncores; i++) {
uncore = &uncores[i];
ctx = *per_cpu_ptr(uncore->ctx, cpu);
+ if (!ctx)
+ continue;
+
if (cpu == ctx->cpu)
cpumask_clear_cpu(cpu, &uncore->active_mask);
@@ -584,6 +601,7 @@ static int amd_uncore_df_init(void)
uncore->msr_base = MSR_F15H_NB_PERF_CTL;
uncore->rdpmc_base = RDPMC_BASE_NB;
uncore->id = amd_uncore_df_id;
+ uncore->group = -1;
if (pmu_version >= 2) {
*df_attr++ = &format_attr_event14v2.attr;
@@ -693,6 +711,7 @@ static int amd_uncore_l3_init(void)
uncore->msr_base = MSR_F16H_L2I_PERF_CTL;
uncore->rdpmc_base = RDPMC_BASE_LLC;
uncore->id = amd_uncore_l3_id;
+ uncore->group = -1;
if (boot_cpu_data.x86 >= 0x17) {
*l3_attr++ = &format_attr_event8.attr;
--
2.34.1
Powered by blists - more mailing lists