[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <ddeae082-c38d-a961-4d90-1fbc1c9f2726@linux.intel.com>
Date: Fri, 24 Jul 2020 15:10:52 -0400
From: "Liang, Kan" <kan.liang@...ux.intel.com>
To: peterz@...radead.org
Cc: acme@...hat.com, mingo@...nel.org, linux-kernel@...r.kernel.org,
jolsa@...nel.org, eranian@...gle.com,
alexander.shishkin@...ux.intel.com, ak@...ux.intel.com,
like.xu@...ux.intel.com
Subject: Re: [PATCH V7 08/14] perf/x86/intel: Generic support for hardware
TopDown metrics
On 7/24/2020 12:07 PM, Liang, Kan wrote:
>
>
> On 7/24/2020 11:27 AM, peterz@...radead.org wrote:
>> On Fri, Jul 24, 2020 at 03:19:06PM +0200, peterz@...radead.org wrote:
>>> On Thu, Jul 23, 2020 at 10:11:11AM -0700, kan.liang@...ux.intel.com
>>> wrote:
>>>> @@ -3375,6 +3428,72 @@ static int intel_pmu_hw_config(struct
>>>> perf_event *event)
>>>> if (event->attr.type != PERF_TYPE_RAW)
>>>> return 0;
>>>> + /*
>>>> + * Config Topdown slots and metric events
>>>> + *
>>>> + * The slots event on Fixed Counter 3 can support sampling,
>>>> + * which will be handled normally in x86_perf_event_update().
>>>> + *
>>>> + * The metric events don't support sampling.
>>>> + *
>>>> + * For counting, topdown slots and metric events will be
>>>> + * handled specially for event update.
>>>> + * A flag PERF_X86_EVENT_TOPDOWN is applied for the case.
>>>> + */
>>>> + if (x86_pmu.intel_cap.perf_metrics && is_topdown_event(event)) {
>>>> + if (is_metric_event(event)) {
>>>> + struct perf_event *leader = event->group_leader;
>>>> + struct perf_event *sibling;
>>>> +
>>>> + /* The metric events don't support sampling. */
>>>> + if (is_sampling_event(event))
>>>> + return -EINVAL;
>>>> +
>>>> + /* The metric events cannot be a group leader. */
>>>> + if (leader == event)
>>>> + return -EINVAL;
>>>> +
>>>> + /*
>>>> + * The slots event cannot be the leader of a topdown
>>>> + * sample-read group, e.g., {slots, topdown-retiring}:S
>>>> + */
>>>> + if (is_slots_event(leader) && is_sampling_event(leader))
>>>> + return -EINVAL;
>>>
>>> This has nothing to do with sample-read; SLOTS cannot be sampling when
>>> coupled with the METRIC stuff because hardware is daft.
>>>
>>> And you can have SAMPLE_READ on non-leader events just fine.
>>>
>>>> +
>>>> + /*
>>>> + * The slots event must be before the metric events,
>>>> + * because we only update the values of a topdown
>>>> + * group once with the slots event.
>>>> + */
>>>> + if (!is_slots_event(leader)) {
>>>> + for_each_sibling_event(sibling, leader) {
>>>> + if (is_slots_event(sibling))
>>>> + break;
>>>> + if (is_metric_event(sibling))
>>>> + return -EINVAL;
>>>> + }
>>>> + }
>>>
>>> Per the SIBLING patch this then wants to be:
>>>
>>> if (!is_slots_event(leader))
>>> return -EINVAL;
>>>
>>> event->event_caps |= PERF_EV_CAP_SIBLING.
>>> /*
>>> * Only once we have a METRICs sibling to we
>>> * need TopDown magic.
>>> */
>>> leader->hw.flags |= PERF_X86_EVENT_TOPDOWN;
Since we only set the flag for the SLOTS event now, the V7 patch will
treat the metric events as normal events, which trigger an error.
To fix the error, I think we may merge the changes as below with the
[PATCH 08/14] perf/x86/intel: Generic support for hardware TopDown metrics.
I think we don't need the PERF_X86_EVENT_TOPDOWN flag anymore.
If it's a non-sampling slots event, apply the special function.
If it's a metric event, do nothing.
diff --git a/arch/x86/events/core.c b/arch/x86/events/core.c
index 0f3d01562ded..02dfee0b6615 100644
--- a/arch/x86/events/core.c
+++ b/arch/x86/events/core.c
@@ -73,10 +73,10 @@ u64 x86_perf_event_update(struct perf_event *event)
u64 prev_raw_count, new_raw_count;
u64 delta;
- if (unlikely(!hwc->event_base))
+ if (unlikely(!hwc->event_base || is_metric_event(event)))
return 0;
- if (unlikely(is_topdown_count(event)) && x86_pmu.update_topdown_event)
+ if (unlikely(is_slots_count(event)) && x86_pmu.update_topdown_event)
return x86_pmu.update_topdown_event(event);
/*
@@ -1280,11 +1280,10 @@ int x86_perf_event_set_period(struct perf_event
*event)
s64 period = hwc->sample_period;
int ret = 0, idx = hwc->idx;
- if (unlikely(!hwc->event_base))
+ if (unlikely(!hwc->event_base || is_metric_event(event)))
return 0;
- if (unlikely(is_topdown_count(event)) &&
- x86_pmu.set_topdown_event_period)
+ if (unlikely(is_slots_count(event)) && x86_pmu.set_topdown_event_period)
return x86_pmu.set_topdown_event_period(event);
/*
diff --git a/arch/x86/events/intel/core.c b/arch/x86/events/intel/core.c
index 6cb079e0c9d9..010ac74afc09 100644
--- a/arch/x86/events/intel/core.c
+++ b/arch/x86/events/intel/core.c
@@ -2405,27 +2405,18 @@ static u64 icl_update_topdown_event(struct
perf_event *event)
return slots;
}
-static void intel_pmu_read_topdown_event(struct perf_event *event)
+static void intel_pmu_read_event(struct perf_event *event)
{
- struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events);
-
- /* Only need to call update_topdown_event() once for group read. */
- if ((cpuc->txn_flags & PERF_PMU_TXN_READ) &&
- !is_slots_event(event))
+ if (unlikely(is_metric_event(event)))
return;
- perf_pmu_disable(event->pmu);
- x86_pmu.update_topdown_event(event);
- perf_pmu_enable(event->pmu);
-}
-
-static void intel_pmu_read_event(struct perf_event *event)
-{
if (event->hw.flags & PERF_X86_EVENT_AUTO_RELOAD)
intel_pmu_auto_reload_read(event);
- else if (is_topdown_count(event) && x86_pmu.update_topdown_event)
- intel_pmu_read_topdown_event(event);
- else
+ else if (is_slots_count(event) && x86_pmu.update_topdown_event) {
+ perf_pmu_disable(event->pmu);
+ x86_pmu.update_topdown_event(event);
+ perf_pmu_enable(event->pmu);
+ } else
x86_perf_event_update(event);
}
@@ -3606,12 +3597,14 @@ static int intel_pmu_hw_config(struct perf_event
*event)
*
* The slots event on Fixed Counter 3 can support sampling,
* which will be handled normally in x86_perf_event_update().
- *
* The metric events don't support sampling.
*
- * For counting, topdown slots and metric events will be
- * handled specially for event update.
- * A flag PERF_X86_EVENT_TOPDOWN is applied for the case.
+ * For the counting of the slots and metric events, the rules
+ * as below have to be applied:
+ * - A standalone metric event or pure metric events group is
+ * not supported.
+ * - The metric events group must include the slots event.
+ * - The slots event must be the leader of the group.
*/
if (x86_pmu.intel_cap.perf_metrics && is_topdown_event(event)) {
if (event->attr.config1 || event->attr.config2)
@@ -3647,11 +3640,6 @@ static int intel_pmu_hw_config(struct perf_event
*event)
return -EINVAL;
event->event_caps |= PERF_EV_CAP_SIBLING;
- /*
- * Only once we have a METRICs sibling do we
- * need TopDown magic.
- */
- leader->hw.flags |= PERF_X86_EVENT_TOPDOWN;
}
}
diff --git a/arch/x86/events/perf_event.h b/arch/x86/events/perf_event.h
index 345442410a4d..153db209c105 100644
--- a/arch/x86/events/perf_event.h
+++ b/arch/x86/events/perf_event.h
@@ -79,12 +79,6 @@ static inline bool constraint_match(struct
event_constraint *c, u64 ecode)
#define PERF_X86_EVENT_PEBS_VIA_PT 0x0800 /* use PT buffer for PEBS */
#define PERF_X86_EVENT_PAIR 0x1000 /* Large Increment per Cycle */
#define PERF_X86_EVENT_LBR_SELECT 0x2000 /* Save/Restore MSR_LBR_SELECT */
-#define PERF_X86_EVENT_TOPDOWN 0x4000 /* Count Topdown slots/metrics
events */
-
-static inline bool is_topdown_count(struct perf_event *event)
-{
- return event->hw.flags & PERF_X86_EVENT_TOPDOWN;
-}
static inline bool is_metric_event(struct perf_event *event)
{
@@ -105,6 +99,14 @@ static inline bool is_topdown_event(struct
perf_event *event)
return is_metric_event(event) || is_slots_event(event);
}
+static inline bool is_slots_count(struct perf_event *event)
+{
+ if (is_slots_event(event) && !is_sampling_event(event))
+ return true;
+
+ return false;
+}
+
struct amd_nb {
int nb_id; /* NorthBridge id */
int refcnt; /* reference count */
Thanks,
Kan
>>>> + }
>>>> +
>>>> + if (!is_sampling_event(event)) {
>>>> + if (event->attr.config1 != 0)
>>>> + return -EINVAL;
>>>
>>> How does this depend on sampling?
>>>
>>>> + /*
>>>> + * The TopDown metrics events and slots event don't
>>>> + * support any filters.
>>>> + */
>>>> + if (event->attr.config & X86_ALL_EVENT_FLAGS)
>>>> + return -EINVAL;
>>>
>>> That seems independent of sampling too. Even a sampling SLOTS shouldn't
>>> be having any of those afaict.
>>>
>>>> +
>>>> + event->hw.flags |= PERF_X86_EVENT_TOPDOWN;
>>>
>>> This is confusing too, a !sampling SLOTS event without METRIC siblings
>>> shouldn't have this set, right? So arguably, this should be like above.
>>>
>>>> +
>>>> + event->event_caps |= PERF_EV_CAP_COEXIST;
>>>> +
>>>> + if (is_metric_event(event))
>>>> + event->hw.flags &= ~PERF_X86_EVENT_RDPMC_ALLOWED;
>>>
>>> This too seems like something that should be in the is_metric_event()
>>> branch above.
>>>
>>>> + }
>>>> + }
>>>> +
>>>> if (!(event->attr.config & ARCH_PERFMON_EVENTSEL_ANY))
>>>> return 0;
>>
>> FWIW, I pushed out a branch with all these changes in:
>>
>> git://git.kernel.org/pub/scm/linux/kernel/git/peterz/queue.git
>> perf/metric
>>
>> Just to get it some build love, if you want it differently, I'm happy to
>> throw it all out again.
>
> Thanks Peter.
>
> I will pull the branch and do more tests.
>
> Thanks,
> Kan
Powered by blists - more mailing lists