lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Date:   Thu, 9 Jun 2022 08:47:14 -0400
From:   "Liang, Kan" <kan.liang@...ux.intel.com>
To:     Namhyung Kim <namhyung@...il.com>,
        Xing Zhengjun <zhengjun.xing@...ux.intel.com>
Cc:     Arnaldo Carvalho de Melo <acme@...nel.org>,
        Peter Zijlstra <peterz@...radead.org>,
        Ingo Molnar <mingo@...hat.com>, alexander.shishkin@...el.com,
        Jiri Olsa <jolsa@...hat.com>,
        linux-kernel <linux-kernel@...r.kernel.org>,
        linux-perf-users <linux-perf-users@...r.kernel.org>,
        Ian Rogers <irogers@...gle.com>,
        Adrian Hunter <adrian.hunter@...el.com>,
        Andi Kleen <ak@...ux.intel.com>
Subject: Re: [PATCH 4/5] perf x86 evlist: Add default hybrid events for perf
 stat



On 6/8/2022 8:04 PM, Namhyung Kim wrote:
> Hello,
> 
> On Tue, Jun 7, 2022 at 12:31 AM <zhengjun.xing@...ux.intel.com> wrote:
>>
>> From: Kan Liang <kan.liang@...ux.intel.com>
>>
>> Provide a new solution to replace the reverted commit ac2dc29edd21
>> ("perf stat: Add default hybrid events").
>>
>> For the default software attrs, nothing is changed.
>> For the default hardware attrs, create a new evsel for each hybrid pmu.
>>
>> With the new solution, adding a new default attr will not require the
>> special support for the hybrid platform anymore.
>>
>> Also, the "--detailed" is supported on the hybrid platform
>>
>> With the patch,
>>
>> ./perf stat -a -ddd sleep 1
>>
>>   Performance counter stats for 'system wide':
>>
>>         32,231.06 msec cpu-clock                 #   32.056 CPUs utilized
>>               529      context-switches          #   16.413 /sec
>>                32      cpu-migrations            #    0.993 /sec
>>                69      page-faults               #    2.141 /sec
>>       176,754,151      cpu_core/cycles/          #    5.484 M/sec          (41.65%)
>>       161,695,280      cpu_atom/cycles/          #    5.017 M/sec          (49.92%)
>>        48,595,992      cpu_core/instructions/    #    1.508 M/sec          (49.98%)
>>        32,363,337      cpu_atom/instructions/    #    1.004 M/sec          (58.26%)
>>        10,088,639      cpu_core/branches/        #  313.010 K/sec          (58.31%)
>>         6,390,582      cpu_atom/branches/        #  198.274 K/sec          (58.26%)
>>           846,201      cpu_core/branch-misses/   #   26.254 K/sec          (66.65%)
>>           676,477      cpu_atom/branch-misses/   #   20.988 K/sec          (58.27%)
>>        14,290,070      cpu_core/L1-dcache-loads/ #  443.363 K/sec          (66.66%)
>>         9,983,532      cpu_atom/L1-dcache-loads/ #  309.749 K/sec          (58.27%)
>>           740,725      cpu_core/L1-dcache-load-misses/ #   22.982 K/sec    (66.66%)
>>   <not supported>      cpu_atom/L1-dcache-load-misses/
>>           480,441      cpu_core/LLC-loads/       #   14.906 K/sec          (66.67%)
>>           326,570      cpu_atom/LLC-loads/       #   10.132 K/sec          (58.27%)
>>               329      cpu_core/LLC-load-misses/ #   10.208 /sec           (66.68%)
>>                 0      cpu_atom/LLC-load-misses/ #    0.000 /sec           (58.32%)
>>   <not supported>      cpu_core/L1-icache-loads/
>>        21,982,491      cpu_atom/L1-icache-loads/ #  682.028 K/sec          (58.43%)
>>         4,493,189      cpu_core/L1-icache-load-misses/ #  139.406 K/sec    (33.34%)
>>         4,711,404      cpu_atom/L1-icache-load-misses/ #  146.176 K/sec    (50.08%)
>>        13,713,090      cpu_core/dTLB-loads/      #  425.462 K/sec          (33.34%)
>>         9,384,727      cpu_atom/dTLB-loads/      #  291.170 K/sec          (50.08%)
>>           157,387      cpu_core/dTLB-load-misses/ #    4.883 K/sec         (33.33%)
>>           108,328      cpu_atom/dTLB-load-misses/ #    3.361 K/sec         (50.08%)
>>   <not supported>      cpu_core/iTLB-loads/
>>   <not supported>      cpu_atom/iTLB-loads/
>>            37,655      cpu_core/iTLB-load-misses/ #    1.168 K/sec         (33.32%)
>>            61,661      cpu_atom/iTLB-load-misses/ #    1.913 K/sec         (50.03%)
>>   <not supported>      cpu_core/L1-dcache-prefetches/
>>   <not supported>      cpu_atom/L1-dcache-prefetches/
>>   <not supported>      cpu_core/L1-dcache-prefetch-misses/
>>   <not supported>      cpu_atom/L1-dcache-prefetch-misses/
>>
>>         1.005466919 seconds time elapsed
>>
>> Signed-off-by: Kan Liang <kan.liang@...ux.intel.com>
>> Signed-off-by: Zhengjun Xing <zhengjun.xing@...ux.intel.com>
>> ---
>>   tools/perf/arch/x86/util/evlist.c | 52 ++++++++++++++++++++++++++++++-
>>   tools/perf/util/evlist.c          |  2 +-
>>   tools/perf/util/evlist.h          |  2 ++
>>   3 files changed, 54 insertions(+), 2 deletions(-)
>>
>> diff --git a/tools/perf/arch/x86/util/evlist.c b/tools/perf/arch/x86/util/evlist.c
>> index 777bdf182a58..1b3f9e1a2287 100644
>> --- a/tools/perf/arch/x86/util/evlist.c
>> +++ b/tools/perf/arch/x86/util/evlist.c
>> @@ -4,16 +4,66 @@
>>   #include "util/evlist.h"
>>   #include "util/parse-events.h"
>>   #include "topdown.h"
>> +#include "util/event.h"
>> +#include "util/pmu-hybrid.h"
>>
>>   #define TOPDOWN_L1_EVENTS      "{slots,topdown-retiring,topdown-bad-spec,topdown-fe-bound,topdown-be-bound}"
>>   #define TOPDOWN_L2_EVENTS      "{slots,topdown-retiring,topdown-bad-spec,topdown-fe-bound,topdown-be-bound,topdown-heavy-ops,topdown-br-mispredict,topdown-fetch-lat,topdown-mem-bound}"
>>
>> +static int ___evlist__add_default_attrs(struct evlist *evlist, struct perf_event_attr *attrs, size_t nr_attrs)
>> +{
>> +       struct perf_cpu_map *cpus;
>> +       struct evsel *evsel, *n;
>> +       struct perf_pmu *pmu;
>> +       LIST_HEAD(head);
>> +       size_t i, j = 0;
>> +
>> +       for (i = 0; i < nr_attrs; i++)
>> +               event_attr_init(attrs + i);
>> +
>> +       if (!perf_pmu__has_hybrid())
>> +               return evlist__add_attrs(evlist, attrs, nr_attrs);
>> +
>> +       for (i = 0; i < nr_attrs; i++) {
>> +               if (attrs[i].type == PERF_TYPE_SOFTWARE) {
>> +                       evsel = evsel__new_idx(attrs + i, evlist->core.nr_entries + j);
> 
> Probably no need to calculate index (j) as it's updated
> later when it goes to the evlist...
> 
> 
>> +                       if (evsel == NULL)
>> +                               goto out_delete_partial_list;
>> +                       j++;
>> +                       list_add_tail(&evsel->core.node, &head);
>> +                       continue;
>> +               }
>> +
>> +               perf_pmu__for_each_hybrid_pmu(pmu) {
>> +                       evsel = evsel__new_idx(attrs + i, evlist->core.nr_entries + j);
>> +                       if (evsel == NULL)
>> +                               goto out_delete_partial_list;
>> +                       j++;
>> +                       evsel->core.attr.config |= (__u64)pmu->type << PERF_PMU_TYPE_SHIFT;
>> +                       cpus = perf_cpu_map__get(pmu->cpus);
>> +                       evsel->core.cpus = cpus;
>> +                       evsel->core.own_cpus = perf_cpu_map__get(cpus);
>> +                       evsel->pmu_name = strdup(pmu->name);
>> +                       list_add_tail(&evsel->core.node, &head);
>> +               }
>> +       }
>> +
>> +       evlist__splice_list_tail(evlist, &head);
> 
> ... like here.

Yes, the index of all new evsel will be updated when adding to the evlist.

Zhengjun, could you please handle the patch? Just set 0 idx for the new 
evsel should be good enough.


Thanks,
Kan

> 
> Thanks,
> Namhyung
> 
> 
>> +
>> +       return 0;
>> +
>> +out_delete_partial_list:
>> +       __evlist__for_each_entry_safe(&head, n, evsel)
>> +               evsel__delete(evsel);
>> +       return -1;
>> +}

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ