lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <AANLkTinP9zdHEoq96nf-eBektNc6g9fPRmjoQHTK1jKY@mail.gmail.com>
Date:	Tue, 1 Mar 2011 09:57:37 +0100
From:	Stephane Eranian <eranian@...gle.com>
To:	Lin Ming <ming.m.lin@...el.com>
Cc:	Peter Zijlstra <a.p.zijlstra@...llo.nl>,
	Ingo Molnar <mingo@...e.hu>, Andi Kleen <andi@...stfloor.org>,
	lkml <linux-kernel@...r.kernel.org>
Subject: Re: [PATCH v2 -tip] perf: x86, add SandyBridge support

On Tue, Mar 1, 2011 at 9:45 AM, Lin Ming <ming.m.lin@...el.com> wrote:
> On Tue, 2011-03-01 at 15:43 +0800, Stephane Eranian wrote:
>> On Mon, Feb 28, 2011 at 10:15 AM, Peter Zijlstra <a.p.zijlstra@...llo.nl> wrote:
>> > On Mon, 2011-02-28 at 15:22 +0800, Lin Ming wrote:
>> >> This patch adds basic SandyBridge support, including hardware cache
>> >> events and PEBS events support.
>> >>
>> >> LLC-* hareware cache events don't work for now, it depends on the
>> >> offcore patches.
>> >
>> > What's the status of those, Stephane reported some problems last I
>> > remember?
>> >
>> I tried the trick I mentioned and it seems to work.
>>
>> Something like below with hwc->extra_alloc.
>> Could probably find a better name for that field.
>
> Stephane,
>
> I'll integrate below changes to the offcore patches, OK?
>
Let me try one more test on this.
I want to show the case the caused the problem in the first place.

> diff --git a/arch/x86/kernel/cpu/perf_event_intel.c b/arch/x86/kernel/cpu/perf_event_intel.c
> index f152930..ac1d100 100644
> --- a/arch/x86/kernel/cpu/perf_event_intel.c
> +++ b/arch/x86/kernel/cpu/perf_event_intel.c
> @@ -906,7 +906,7 @@ intel_percore_constraints(struct cpu_hw_events *cpuc, struct perf_event *event)
>        int free_slot;
>        int found;
>
> -       if (!x86_pmu.percore_constraints)
> +       if (!x86_pmu.percore_constraints || hwc->extra_alloc)
>                return NULL;
>
>        for (c = x86_pmu.percore_constraints; c->cmask; c++) {
> @@ -931,6 +931,7 @@ intel_percore_constraints(struct cpu_hw_events *cpuc, struct perf_event *event)
>                                if (hwc->extra_config == era->extra_config) {
>                                        era->ref++;
>                                        cpuc->percore_used = 1;
> +                                       hwc->extra_alloc = 1;
>                                        c = NULL;
>                                }
>                                /* else conflict */
> @@ -945,6 +946,7 @@ intel_percore_constraints(struct cpu_hw_events *cpuc, struct perf_event *event)
>                        era->extra_reg = hwc->extra_reg;
>                        era->extra_config = hwc->extra_config;
>                        cpuc->percore_used = 1;
> +                       hwc->extra_alloc = 1;
>                        c = NULL;
>                }
>                raw_spin_unlock(&pc->lock);
> @@ -998,6 +1000,7 @@ static void intel_put_event_constraints(struct cpu_hw_events *cpuc,
>                            era->extra_config == hwc->extra_config &&
>                            era->extra_reg == er->msr) {
>                                era->ref--;
> +                               hwc->extra_alloc = 0;
>                                break;
>                        }
>                }
> diff --git a/include/linux/perf_event.h b/include/linux/perf_event.h
> index f531ce3..dbbf33a 100644
> --- a/include/linux/perf_event.h
> +++ b/include/linux/perf_event.h
> @@ -546,6 +546,7 @@ struct hw_perf_event {
>                        int             last_cpu;
>                        unsigned int    extra_reg;
>                        u64             extra_config;
> +                       int             extra_alloc;
>                };
>                struct { /* software */
>                        struct hrtimer  hrtimer;
>
>
>>
>> static struct event_constraint *
>> intel_percore_constraints(struct cpu_hw_events *cpuc, struct perf_event *event)
>> {
>>         struct hw_perf_event *hwc = &event->hw;
>>         unsigned int e = hwc->config & ARCH_PERFMON_EVENTSEL_EVENT;
>>         struct event_constraint *c;
>>         struct intel_percore *pc;
>>         struct er_account *era;
>>         int i;
>>         int free_slot;
>>         int found;
>>
>>         if (!x86_pmu.percore_constraints)
>>                 return NULL;
>>
>>         if (hwc->extra_alloc)
>>                 return NULL;
>>
>>         for (c = x86_pmu.percore_constraints; c->cmask; c++) {
>>                 if (e != c->code)
>>                         continue;
>>
>>                 /*
>>                  * Allocate resource per core.
>>                  */
>>                 c = NULL;
>>                 pc = cpuc->per_core;
>>                 if (!pc)
>>                         break;
>>                 c = &emptyconstraint;
>>                 raw_spin_lock(&pc->lock);
>>                 free_slot = -1;
>>                 found = 0;
>>                 for (i = 0; i < MAX_EXTRA_REGS; i++) {
>>                         era = &pc->regs[i];
>>                         if (era->ref > 0 && hwc->extra_reg == era->extra_reg) {
>>                                 /* Allow sharing same config */
>>                                 if (hwc->extra_config == era->extra_config) {
>>                                         era->ref++;
>>                                         cpuc->percore_used = 1;
>>                                         hwc->extra_alloc = 1;
>>                                         c = NULL;
>>                                 }
>>                                 /* else conflict */
>>                                 found = 1;
>>                                 break;
>>                         } else if (era->ref == 0 && free_slot == -1)
>>                                 free_slot = i;
>>                 }
>>                 if (!found && free_slot != -1) {
>>                         era = &pc->regs[free_slot];
>>                         era->ref = 1;
>>                         era->extra_reg = hwc->extra_reg;
>>                         era->extra_config = hwc->extra_config;
>>                         cpuc->percore_used = 1;
>>                         hwc->extra_alloc = 1;
>>                         c = NULL;
>>                 }
>>                 raw_spin_unlock(&pc->lock);
>>                 return c;
>>         }
>>
>>         return NULL;
>> }
>>
>> static void intel_put_event_constraints(struct cpu_hw_events *cpuc,
>>                                         struct perf_event *event)
>> {
>>         struct extra_reg *er;
>>         struct intel_percore *pc;
>>         struct er_account *era;
>>         struct hw_perf_event *hwc = &event->hw;
>>         int i, allref;
>>
>>         if (!cpuc->percore_used)
>>                 return;
>>
>>         for (er = x86_pmu.extra_regs; er->msr; er++) {
>>                 if (er->event != (hwc->config & er->config_mask))
>>                         continue;
>>
>>                 pc = cpuc->per_core;
>>                 raw_spin_lock(&pc->lock);
>>                 for (i = 0; i < MAX_EXTRA_REGS; i++) {
>>                         era = &pc->regs[i];
>>                         if (era->ref > 0 &&
>>                             era->extra_config == hwc->extra_config &&
>>                             era->extra_reg == er->msr) {
>>                                 era->ref--;
>>                                 hwc->extra_alloc = 0;
>>                                 break;
>>                         }
>>                 }
>>                 allref = 0;
>>                 for (i = 0; i < MAX_EXTRA_REGS; i++)
>>                         allref += pc->regs[i].ref;
>>                 if (allref == 0)
>>                         cpuc->percore_used = 0;
>>                 raw_spin_unlock(&pc->lock);
>>                 break;
>>         }
>> }
>>
>> >
>> >>  #define INTEL_EVENT_CONSTRAINT(c, n) \
>> >>       EVENT_CONSTRAINT(c, n, ARCH_PERFMON_EVENTSEL_EVENT)
>> >> +#define INTEL_EVENT_CONSTRAINT2(c, n)        \
>> >> +     EVENT_CONSTRAINT(c, n, INTEL_ARCH_EVENT_MASK)
>> >
>> > That's a particularly bad name, how about something like
>> >
>> > INTEL_UEVENT_CONSTRAINT or somesuch.
>> >
>> >> @@ -702,7 +738,13 @@ static void intel_ds_init(void)
>> >>                       printk(KERN_CONT "PEBS fmt1%c, ", pebs_type);
>> >>                       x86_pmu.pebs_record_size = sizeof(struct pebs_record_nhm);
>> >>                       x86_pmu.drain_pebs = intel_pmu_drain_pebs_nhm;
>> >> -                     x86_pmu.pebs_constraints = intel_nehalem_pebs_events;
>> >> +                     switch (boot_cpu_data.x86_model) {
>> >> +                     case 42: /* SandyBridge */
>> >> +                             x86_pmu.pebs_constraints = intel_snb_pebs_events;
>> >> +                             break;
>> >> +                     default:
>> >> +                             x86_pmu.pebs_constraints = intel_nehalem_pebs_events;
>> >> +                     }
>> >>                       break;
>> >>
>> >>               default:
>> >
>> > We already have this massive model switch right after this function,
>> > might as well move the pebs constraint assignment there.
>> >
>
>
>

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ