[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <4f63c3d5-2a33-28ed-4e45-086045e9ab50@linux.intel.com>
Date: Fri, 21 Sep 2018 09:13:08 +0300
From: Alexey Budankov <alexey.budankov@...ux.intel.com>
To: Jiri Olsa <jolsa@...hat.com>
Cc: Jiri Olsa <jolsa@...nel.org>,
Arnaldo Carvalho de Melo <acme@...nel.org>,
lkml <linux-kernel@...r.kernel.org>,
Ingo Molnar <mingo@...nel.org>,
Namhyung Kim <namhyung@...nel.org>,
Alexander Shishkin <alexander.shishkin@...ux.intel.com>,
Peter Zijlstra <a.p.zijlstra@...llo.nl>,
Andi Kleen <andi@...stfloor.org>
Subject: Re: [RFCv2 00/48] perf tools: Add threads to record command
Hello Jiri,
On 14.09.2018 12:37, Alexey Budankov wrote:
> On 14.09.2018 11:28, Jiri Olsa wrote:
>> On Fri, Sep 14, 2018 at 10:26:53AM +0200, Jiri Olsa wrote:
>>
>> SNIP
>>
>>>>> The threaded monitoring currently can't monitor backward maps
>>>>> and there are probably more limitations which I haven't spotted
>>>>> yet.
>>>>>
>>>>> So far I tested on laptop:
>>>>> http://people.redhat.com/~jolsa/record_threads/test-4CPU.txt
>>>>>
>>>>> and a one bigger server:
>>>>> http://people.redhat.com/~jolsa/record_threads/test-208CPU.txt
>>>>>
>>>>> I can see decrease in recorded LOST events, but both the benchmark
>>>>> and the monitoring must be carefully configured wrt:
>>>>> - number of events (frequency)
>>>>> - size of the memory maps
>>>>> - size of events (callchains)
>>>>> - final perf.data size
>>>>>
>>>>> It's also available in:
>>>>> git://git.kernel.org/pub/scm/linux/kernel/git/jolsa/perf.git
>>>>> perf/record_threads
>>>>>
>>>>> thoughts? ;-) thanks
>>>>> jirka
>>>>
>>>> It is preferable to split into smaller pieces that bring
>>>> some improvement proved by metrics numbers and ready for
>>>> merging and upstream. Do we have more metrics than the
>>>> data loss from trace AIO patches?
>>>
>>> well the primary focus is to get more events in,
>>> so the LOST metric is the main one
>>
>> actualy I was hoping, could you please run it through the same
>> tests as you do for AIO code on some huge server?
>
> Yeah, I will, but it takes some time.
Here it is:
Hardware:
cat /proc/cpuinfo
processor : 271
vendor_id : GenuineIntel
cpu family : 6
model : 133
model name : Intel(R) Xeon Phi(TM) CPU 7285 @ 1.30GHz
stepping : 0
microcode : 0xe
cpu MHz : 1064.235
cache size : 1024 KB
physical id : 0
siblings : 272
core id : 73
cpu cores : 68
apicid : 295
initial apicid : 295
fpu : yes
fpu_exception : yes
cpuid level : 13
wp : yes
flags : fpu vme de pse tsc msr pae mce cx8 apic sep mtrr pge mca cmov pat pse36 clflush dts acpi mmx fxsr sse sse2 ss ht tm pbe syscall nx pdpe1gb rdtscp lm constant_tsc arch_perfmon pebs bts rep_good nopl xtopology nonstop_tsc cpuid aperfmperf pni pclmulqdq dtes64 monitor ds_cpl vmx est tm2 ssse3 fma cx16 xtpr pdcm sse4_1 sse4_2 x2apic movbe popcnt tsc_deadline_timer aes xsave avx f16c rdrand lahf_lm abm 3dnowprefetch ring3mwait cpuid_fault epb pti tpr_shadow vnmi flexpriority ept vpid fsgsbase tsc_adjust bmi1 avx2 smep bmi2 erms avx512f rdseed adx avx512pf avx512er avx512cd xsaveopt dtherm ida arat pln pts avx512_vpopcntdq avx512_4vnniw avx512_4fmaps
bugs : cpu_meltdown spectre_v1 spectre_v2
bogomips : 2594.07
clflush size : 64
cache_alignment : 64
address sizes : 46 bits physical, 48 bits virtual
power management:
uname -a
Linux nntpat98-196 4.18.0-rc7+ #2 SMP Thu Sep 6 13:24:37 MSK 2018 x86_64 x86_64 x86_64 GNU/Linux
cat /proc/sys/kernel/perf_event_paranoid
0
cat /proc/sys/kernel/perf_event_mlock_kb
516
cat /proc/sys/kernel/perf_event_max_sample_rate
3000
cat /etc/redhat-release
Red Hat Enterprise Linux Server release 7.5 (Maipo)
Metrics:
runtime overhead (%) : elapsed_time_under_profiling / elapsed_time
data loss (%) : paused_time / elapsed_time_under_profiling
LOST events : stat from perf report --stats
SAMPLE events : stat from perf report --stats
perf.data size (B) : size of trace file on disk
Events:
cpu/period=P,event=0x3c/Duk;CPU_CLK_UNHALTED.THREAD
cpu/period=P,umask=0x3/Duk;CPU_CLK_UNHALTED.REF_TSC
cpu/period=P,event=0xc0/Duk;INST_RETIRED.ANY
cpu/period=0xaae61,event=0xc2,umask=0x10/uk;UOPS_RETIRED.ALL
cpu/period=0x11171,event=0xc2,umask=0x20/uk;UOPS_RETIRED.SCALAR_SIMD
cpu/period=0x11171,event=0xc2,umask=0x40/uk;UOPS_RETIRED.PACKED_SIMD
=================================================
Command:
/usr/bin/time /tmp/vtune_amplifier_2019.574715/bin64/perf.thr record --threads=T \
-a -N -B -T -R --call-graph dwarf,1024 --user-regs=ip,bp,sp \
-e cpu/period=P,event=0x3c/Duk,\
cpu/period=P,umask=0x3/Duk,\
cpu/period=P,event=0xc0/Duk,\
cpu/period=0x30d40,event=0xc2,umask=0x10/uk,\
cpu/period=0x4e20,event=0xc2,umask=0x20/uk,\
cpu/period=0x4e20,event=0xc2,umask=0x40/uk \
--clockid=monotonic_raw -- ./matrix.(icc|gcc)
Workload: matrix multiplication in 256 threads
/usr/bin/time ./matrix.icc
Addr of buf1 = 0x7ff9faa73010
Offs of buf1 = 0x7ff9faa73180
Addr of buf2 = 0x7ff9f8a72010
Offs of buf2 = 0x7ff9f8a721c0
Addr of buf3 = 0x7ff9f6a71010
Offs of buf3 = 0x7ff9f6a71100
Addr of buf4 = 0x7ff9f4a70010
Offs of buf4 = 0x7ff9f4a70140
Threads #: 256 Pthreads
Matrix size: 2048
Using multiply kernel: multiply1
Freq = 0.997720 GHz
Execution time = 9.061 seconds
1639.55user 6.59system 0:07.12elapsed 23094%CPU (0avgtext+0avgdata 100448maxresident)k
96inputs+0outputs (1major+33839minor)pagefaults 0swaps
T : 272
P (period, ms) : 0.1
runtime overhead (%) : 45x ~ 323.54 / 7.12
data loss (%) : 96
LOST events : 323662
SAMPLE events : 31885479
perf.data size (GiB) : 42
P (period, ms) : 0.25
runtime overhead (%) : 25x ~ 180.76 / 7.12
data loss (%) : 69
LOST events : 10636
SAMPLE events : 18692998
perf.data size (GiB) : 23.5
P (period, ms) : 0.35
runtime overhead (%) : 16x ~ 119.49 / 7.12
data loss (%) : 1
LOST events : 6
SAMPLE events : 11178524
perf.data size (GiB) : 14
T : 128
P (period, ms) : 0.35
runtime overhead (%) : 15x ~ 111.98 / 7.12
data loss (%) : 62
LOST events : 2825
SAMPLE events : 11267247
perf.data size (GiB) : 15
T : 64
P (period, ms) : 0.35
runtime overhead (%) : 14x ~ 101.55 / 7.12
data loss (%) : 67
LOST events : 5155
SAMPLE events : 10966297
perf.data size (GiB) : 13.7
Workload: matrix multiplication in 128 threads
/usr/bin/time ./matrix.gcc
Addr of buf1 = 0x7f072e630010
Offs of buf1 = 0x7f072e630180
Addr of buf2 = 0x7f072c62f010
Offs of buf2 = 0x7f072c62f1c0
Addr of buf3 = 0x7f072a62e010
Offs of buf3 = 0x7f072a62e100
Addr of buf4 = 0x7f072862d010
Offs of buf4 = 0x7f072862d140
Threads #: 128 Pthreads
Matrix size: 2048
Using multiply kernel: multiply1
Execution time = 6.639 seconds
767.03user 11.17system 0:06.81elapsed 11424%CPU (0avgtext+0avgdata 100756maxresident)k
88inputs+0outputs (0major+139898minor)pagefaults 0swaps
T : 272
P (period, ms) : 0.1
runtime overhead (%) : 29x ~ 198.81 / 6.81
data loss (%) : 21
LOST events : 2502
SAMPLE events : 22481062
perf.data size (GiB) : 27.6
P (period, ms) : 0.25
runtime overhead (%) : 13x ~ 88.47 / 6.81
data loss (%) : 0
LOST events : 0
SAMPLE events : 9572787
perf.data size (GiB) : 11.3
P (period, ms) : 0.35
runtime overhead (%) : 10x ~ 67.11 / 6.81
data loss (%) : 1
LOST events : 137
SAMPLE events : 6985930
perf.data size (GiB) : 8
T : 128
P (period, ms) : 0.35
runtime overhead (%) : 9.5x ~ 64.33 / 6.81
data loss (%) : 1
LOST events : 3
SAMPLE events : 6666903
perf.data size (GiB) : 7.8
T : 64
P (period, ms) : 0.25
runtime overhead (%) : 17x ~ 114.27 / 6.81
data loss (%) : 2
LOST events : 52
SAMPLE events : 12643645
perf.data size (GiB) : 15.5
P (period, ms) : 0.35
runtime overhead (%) : 10x ~ 68.60 / 6.81
data loss (%) : 1
LOST events : 93
SAMPLE events : 7164368
perf.data size (GiB) : 8.5
Thanks,
Alexey
>
>>
>> thanks,
>> jirka
>>
>
Powered by blists - more mailing lists