[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <669ac082-0652-6947-81e1-0f99de0521b5@amd.com>
Date: Thu, 18 May 2023 07:43:00 +0530
From: K Prateek Nayak <kprateek.nayak@....com>
To: Ian Rogers <irogers@...gle.com>
Cc: linux-perf-users@...r.kernel.org, linux-kernel@...r.kernel.org,
acme@...nel.org, peterz@...radead.org, mingo@...hat.com,
mark.rutland@....com, alexander.shishkin@...ux.intel.com,
jolsa@...nel.org, namhyung@...nel.org, ravi.bangoria@....com,
sandipan.das@....com, ananth.narayan@....com,
gautham.shenoy@....com, eranian@...gle.com, puwen@...on.cn
Subject: Re: [PATCH v4 0/5] perf stat: Add option to aggregate data based on
the cache topology
Hello Ian,
On 5/17/2023 11:28 PM, Ian Rogers wrote:
> On Wed, May 17, 2023 at 10:22 AM K Prateek Nayak <kprateek.nayak@....com> wrote:
>>
>> Motivation behind this feature is to aggregate the data at the LLC level
>> for chiplet based processors which currently do not expose the chiplet
>> details in sysfs cpu topology information.
>>
>> For the completeness of the feature, the series adds ability to
>> aggregate data at any cache level. Following is the example of the
>> output on a dual socket Zen3 processor with 2 x 64C/128T containing 8
>> chiplet per socket.
>>
>> $ sudo perf stat --per-cache -a -e ls_dmnd_fills_from_sys.ext_cache_remote --\
>> taskset -c 0-15,64-79,128-143,192-207\
>> perf bench sched messaging -p -t -l 100000 -g 8
>>
>> # Running 'sched/messaging' benchmark:
>> # 20 sender and receiver threads per group
>> # 8 groups == 320 threads run
>>
>> Total time: 7.648 [sec]
>>
>> Performance counter stats for 'system wide':
>>
>> S0-D0-L3-ID0 16 17,145,912 ls_dmnd_fills_from_sys.ext_cache_remote
>> S0-D0-L3-ID8 16 14,977,628 ls_dmnd_fills_from_sys.ext_cache_remote
>> S0-D0-L3-ID16 16 262,539 ls_dmnd_fills_from_sys.ext_cache_remote
>> S0-D0-L3-ID24 16 3,140 ls_dmnd_fills_from_sys.ext_cache_remote
>> S0-D0-L3-ID32 16 27,403 ls_dmnd_fills_from_sys.ext_cache_remote
>> S0-D0-L3-ID40 16 17,026 ls_dmnd_fills_from_sys.ext_cache_remote
>> S0-D0-L3-ID48 16 7,292 ls_dmnd_fills_from_sys.ext_cache_remote
>> S0-D0-L3-ID56 16 2,464 ls_dmnd_fills_from_sys.ext_cache_remote
>> S1-D1-L3-ID64 16 22,489,306 ls_dmnd_fills_from_sys.ext_cache_remote
>> S1-D1-L3-ID72 16 21,455,257 ls_dmnd_fills_from_sys.ext_cache_remote
>> S1-D1-L3-ID80 16 11,619 ls_dmnd_fills_from_sys.ext_cache_remote
>> S1-D1-L3-ID88 16 30,978 ls_dmnd_fills_from_sys.ext_cache_remote
>> S1-D1-L3-ID96 16 37,628 ls_dmnd_fills_from_sys.ext_cache_remote
>> S1-D1-L3-ID104 16 13,594 ls_dmnd_fills_from_sys.ext_cache_remote
>> S1-D1-L3-ID112 16 10,164 ls_dmnd_fills_from_sys.ext_cache_remote
>> S1-D1-L3-ID120 16 11,259 ls_dmnd_fills_from_sys.ext_cache_remote
>>
>> 7.779171484 seconds time elapsed
>>
>> [..snip..]
>
> Acked-by: Ian Rogers <irogers@...gle.com>
Thank you for taking a look at the series and for the ack :)
>
> Thanks,
> Ian
>
>> [..snip..]
>>
--
Thanks and Regards,
Prateek
Powered by blists - more mailing lists