lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <669ac082-0652-6947-81e1-0f99de0521b5@amd.com>
Date:   Thu, 18 May 2023 07:43:00 +0530
From:   K Prateek Nayak <kprateek.nayak@....com>
To:     Ian Rogers <irogers@...gle.com>
Cc:     linux-perf-users@...r.kernel.org, linux-kernel@...r.kernel.org,
        acme@...nel.org, peterz@...radead.org, mingo@...hat.com,
        mark.rutland@....com, alexander.shishkin@...ux.intel.com,
        jolsa@...nel.org, namhyung@...nel.org, ravi.bangoria@....com,
        sandipan.das@....com, ananth.narayan@....com,
        gautham.shenoy@....com, eranian@...gle.com, puwen@...on.cn
Subject: Re: [PATCH v4 0/5] perf stat: Add option to aggregate data based on
 the cache topology

Hello Ian,

On 5/17/2023 11:28 PM, Ian Rogers wrote:
> On Wed, May 17, 2023 at 10:22 AM K Prateek Nayak <kprateek.nayak@....com> wrote:
>>
>> Motivation behind this feature is to aggregate the data at the LLC level
>> for chiplet based processors which currently do not expose the chiplet
>> details in sysfs cpu topology information.
>>
>> For the completeness of the feature, the series adds ability to
>> aggregate data at any cache level. Following is the example of the
>> output on a dual socket Zen3 processor with 2 x 64C/128T containing 8
>> chiplet per socket.
>>
>>   $ sudo perf stat --per-cache -a -e ls_dmnd_fills_from_sys.ext_cache_remote --\
>>     taskset -c 0-15,64-79,128-143,192-207\
>>     perf bench sched messaging -p -t -l 100000 -g 8
>>
>>     # Running 'sched/messaging' benchmark:
>>     # 20 sender and receiver threads per group
>>     # 8 groups == 320 threads run
>>
>>     Total time: 7.648 [sec]
>>
>>     Performance counter stats for 'system wide':
>>
>>     S0-D0-L3-ID0             16         17,145,912      ls_dmnd_fills_from_sys.ext_cache_remote
>>     S0-D0-L3-ID8             16         14,977,628      ls_dmnd_fills_from_sys.ext_cache_remote
>>     S0-D0-L3-ID16            16            262,539      ls_dmnd_fills_from_sys.ext_cache_remote
>>     S0-D0-L3-ID24            16              3,140      ls_dmnd_fills_from_sys.ext_cache_remote
>>     S0-D0-L3-ID32            16             27,403      ls_dmnd_fills_from_sys.ext_cache_remote
>>     S0-D0-L3-ID40            16             17,026      ls_dmnd_fills_from_sys.ext_cache_remote
>>     S0-D0-L3-ID48            16              7,292      ls_dmnd_fills_from_sys.ext_cache_remote
>>     S0-D0-L3-ID56            16              2,464      ls_dmnd_fills_from_sys.ext_cache_remote
>>     S1-D1-L3-ID64            16         22,489,306      ls_dmnd_fills_from_sys.ext_cache_remote
>>     S1-D1-L3-ID72            16         21,455,257      ls_dmnd_fills_from_sys.ext_cache_remote
>>     S1-D1-L3-ID80            16             11,619      ls_dmnd_fills_from_sys.ext_cache_remote
>>     S1-D1-L3-ID88            16             30,978      ls_dmnd_fills_from_sys.ext_cache_remote
>>     S1-D1-L3-ID96            16             37,628      ls_dmnd_fills_from_sys.ext_cache_remote
>>     S1-D1-L3-ID104           16             13,594      ls_dmnd_fills_from_sys.ext_cache_remote
>>     S1-D1-L3-ID112           16             10,164      ls_dmnd_fills_from_sys.ext_cache_remote
>>     S1-D1-L3-ID120           16             11,259      ls_dmnd_fills_from_sys.ext_cache_remote
>>
>>           7.779171484 seconds time elapsed
>>
>> [..snip..]
> 
> Acked-by: Ian Rogers <irogers@...gle.com>

Thank you for taking a look at the series and for the ack :)

> 
> Thanks,
> Ian
> 
>>  [..snip..]
>>
 
--
Thanks and Regards,
Prateek

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ