lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [thread-next>] [day] [month] [year] [list]
Message-ID: <20110831124125.GA10443@quad>
Date:	Wed, 31 Aug 2011 14:41:25 +0200
From:	Stephane Eranian <eranian@...gle.com>
To:	linux-kernel@...r.kernel.org
Cc:	ming.m.lin@...el.com, peterz@...radead.org, mingo@...e.hu,
	andi@...stfloor.org, acme@...stprotocols.net
Subject: [PATCH] perf_event: fix Ming's Intel NHM/WSM/SNB uncore PMU patch


The following patch sits on top of Lin Ming's patch
which adds support for Intel uncore PMU on NHM/WSM/SNB.
The patch series was posted on LKML in July 2011.

The patches fixes:
  - nhm_uncore_pmu_enable_all() to also enabled the fixed uncore counters
  - uncore_pmu_add() to schedule for the actual max number of generic counters
  - fixed the NHM/WSM vs. SNB MSRs for the fixed counter, they are swapped 

The patch adds the following improvement:
  - add an event group validation function, to check if a group can ever be
    schedule. This way, the behavior is consistent with core PMU event groups.

Signed-off-by: Stephane Eranian <eranian@...gle.com>
---

diff --git a/arch/x86/include/asm/msr-index.h b/arch/x86/include/asm/msr-index.h
index e250977..283f292 100644
--- a/arch/x86/include/asm/msr-index.h
+++ b/arch/x86/include/asm/msr-index.h
@@ -426,8 +426,11 @@
 
 /* Intel Nehalem/Westmere/SandyBridge uncore performance counters */
 #define MSR_UNCORE_PERF_GLOBAL_CTRL	0x00000391
-#define MSR_UNCORE_FIXED_CTR_CTRL	0x00000394
-#define MSR_UNCORE_FIXED_CTR0		0x00000395
+#define MSR_NHM_UNCORE_FIXED_CTR0	0x00000394
+#define MSR_NHM_UNCORE_FIXED_CTR_CTRL	0x00000395
+
+#define MSR_SNB_UNCORE_FIXED_CTR0	0x00000395
+#define MSR_SNB_UNCORE_FIXED_CTR_CTRL	0x00000394
 
 #define MSR_NHM_UNCORE_PMC0		0x000003b0
 #define MSR_NHM_UNCORE_PERFEVTSEL0	0x000003c0
diff --git a/arch/x86/kernel/cpu/perf_event_intel_uncore.c b/arch/x86/kernel/cpu/perf_event_intel_uncore.c
index 1100589..70bd28c 100644
--- a/arch/x86/kernel/cpu/perf_event_intel_uncore.c
+++ b/arch/x86/kernel/cpu/perf_event_intel_uncore.c
@@ -23,8 +23,8 @@ static void uncore_fixed_hw_config(struct perf_event *event)
 {
 	struct hw_perf_event *hwc = &event->hw;
 
-	hwc->config_base = MSR_UNCORE_FIXED_CTR_CTRL;
-	hwc->event_base = MSR_UNCORE_FIXED_CTR0;
+	hwc->config_base = intel_uncore_pmu.fixed_config_base;
+	hwc->event_base = intel_uncore_pmu.fixed_event_base;
 }
 
 static void uncore_fixed_disable_event(struct perf_event *event)
@@ -63,7 +63,14 @@ static void uncore_pmu_disable_event(struct perf_event *event)
 
 static void nhm_uncore_pmu_enable_all(void)
 {
-	u64 ctrl = (1 << UNCORE_NUM_COUNTERS) - 1;
+	u64 ctrl, fmask;
+
+	/* generic counters */
+	ctrl = (1 << UNCORE_NUM_GENERIC_COUNTERS) - 1;
+
+	/* fixed counters */
+	fmask = (1 << UNCORE_NUM_FIXED_COUNTERS) - 1;
+	ctrl |= fmask << X86_PMC_IDX_FIXED;
 
 	wrmsrl(MSR_UNCORE_PERF_GLOBAL_CTRL, ctrl);
 }
@@ -96,6 +103,8 @@ static __initconst const struct intel_uncore_pmu nhm_uncore_pmu = {
 	.hw_config		= nhm_uncore_pmu_hw_config,
 	.cntval_bits		= 48,
 	.cntval_bits_fixed	= 48,
+	.fixed_config_base	= MSR_NHM_UNCORE_FIXED_CTR_CTRL,
+	.fixed_event_base	= MSR_NHM_UNCORE_FIXED_CTR0,
 };
 
 /* SandyBridge uncore pmu */
@@ -149,6 +158,8 @@ static __initconst const struct intel_uncore_pmu snb_uncore_pmu = {
 	.hw_config		= snb_uncore_pmu_hw_config,
 	.cntval_bits		= 44,
 	.cntval_bits_fixed	= 48,
+	.fixed_config_base	= MSR_SNB_UNCORE_FIXED_CTR_CTRL,
+	.fixed_event_base	= MSR_SNB_UNCORE_FIXED_CTR0,
 };
 
 static u64 uncore_perf_event_update(struct perf_event *event)
@@ -234,9 +245,43 @@ static void uncore_pmu_init_hrtimer(struct intel_uncore *uncore)
 
 static struct pmu uncore_pmu;
 
+static int uncore_validate_group(struct perf_event *event)
+{
+	struct perf_event *leader = event->group_leader;
+	struct perf_event *e = event->group_leader;
+	int gen = 0, fixed = 0;
+	int ret = 0;
+
+	if (leader->attr.config == UNCORE_FIXED_EVENT)
+		fixed++;
+	else
+		gen++;
+
+	if (event->attr.config == UNCORE_FIXED_EVENT)
+		fixed++;
+	else
+		gen++;
+
+	list_for_each_entry(e, &leader->sibling_list, group_entry) {
+		if (e->attr.config == UNCORE_FIXED_EVENT)
+			fixed++;
+		else
+			gen++;
+	}
+
+	if (fixed > UNCORE_NUM_FIXED_COUNTERS)
+		ret = -ENOSPC;
+
+	if (gen > UNCORE_NUM_GENERIC_COUNTERS)
+		ret = -ENOSPC;
+
+	return ret;
+}
+
 static int uncore_pmu_event_init(struct perf_event *event)
 {
 	struct hw_perf_event *hwc = &event->hw;
+	int ret = 0;
 
 	if (!uncore_pmu_initialized)
 		return -ENOENT;
@@ -256,7 +301,10 @@ static int uncore_pmu_event_init(struct perf_event *event)
 	if (hwc->sample_period)
 		return -EINVAL;
 
-	return 0;
+	if (event->group_leader != event)
+		ret = uncore_validate_group(event);
+
+	return ret;
 }
 
 static void uncore_pmu_start(struct perf_event *event, int flags)
@@ -290,7 +338,7 @@ static int uncore_pmu_add(struct perf_event *event, int flags)
 		goto fixed_event;
 	}
 
-	for (i = 0; i < X86_PMC_IDX_FIXED; i++) {
+	for (i = 0; i < UNCORE_NUM_GENERIC_COUNTERS; i++) {
 fixed_event:
 		if (!uncore->events[i]) {
 			uncore->events[i] = event;
diff --git a/arch/x86/kernel/cpu/perf_event_intel_uncore.h b/arch/x86/kernel/cpu/perf_event_intel_uncore.h
index c7392aa..79f3d19 100644
--- a/arch/x86/kernel/cpu/perf_event_intel_uncore.h
+++ b/arch/x86/kernel/cpu/perf_event_intel_uncore.h
@@ -61,4 +61,6 @@ struct intel_uncore_pmu {
 	void		(*hw_config)(struct perf_event *event);
 	int		cntval_bits;
 	int		cntval_bits_fixed;
+	int		fixed_config_base;
+	int		fixed_event_base;
 };
--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ