lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <4f63c3d5-2a33-28ed-4e45-086045e9ab50@linux.intel.com>
Date:   Fri, 21 Sep 2018 09:13:08 +0300
From:   Alexey Budankov <alexey.budankov@...ux.intel.com>
To:     Jiri Olsa <jolsa@...hat.com>
Cc:     Jiri Olsa <jolsa@...nel.org>,
        Arnaldo Carvalho de Melo <acme@...nel.org>,
        lkml <linux-kernel@...r.kernel.org>,
        Ingo Molnar <mingo@...nel.org>,
        Namhyung Kim <namhyung@...nel.org>,
        Alexander Shishkin <alexander.shishkin@...ux.intel.com>,
        Peter Zijlstra <a.p.zijlstra@...llo.nl>,
        Andi Kleen <andi@...stfloor.org>
Subject: Re: [RFCv2 00/48] perf tools: Add threads to record command

Hello Jiri,

On 14.09.2018 12:37, Alexey Budankov wrote:
> On 14.09.2018 11:28, Jiri Olsa wrote:
>> On Fri, Sep 14, 2018 at 10:26:53AM +0200, Jiri Olsa wrote:
>>
>> SNIP
>>
>>>>> The threaded monitoring currently can't monitor backward maps
>>>>> and there are probably more limitations which I haven't spotted
>>>>> yet.
>>>>>
>>>>> So far I tested on laptop:
>>>>>   http://people.redhat.com/~jolsa/record_threads/test-4CPU.txt
>>>>>
>>>>> and a one bigger server:
>>>>>   http://people.redhat.com/~jolsa/record_threads/test-208CPU.txt
>>>>>
>>>>> I can see decrease in recorded LOST events, but both the benchmark
>>>>> and the monitoring must be carefully configured wrt:
>>>>>   - number of events (frequency)
>>>>>   - size of the memory maps
>>>>>   - size of events (callchains)
>>>>>   - final perf.data size
>>>>>
>>>>> It's also available in:
>>>>>   git://git.kernel.org/pub/scm/linux/kernel/git/jolsa/perf.git
>>>>>   perf/record_threads
>>>>>
>>>>> thoughts? ;-) thanks
>>>>> jirka
>>>>
>>>> It is preferable to split into smaller pieces that bring 
>>>> some improvement proved by metrics numbers and ready for 
>>>> merging and upstream. Do we have more metrics than the 
>>>> data loss from trace AIO patches?
>>>
>>> well the primary focus is to get more events in,
>>> so the LOST metric is the main one
>>
>> actualy I was hoping, could you please run it through the same
>> tests as you do for AIO code on some huge server? 
> 
> Yeah, I will, but it takes some time.

Here it is:

Hardware:
cat /proc/cpuinfo
processor	: 271
vendor_id	: GenuineIntel
cpu family	: 6
model		: 133
model name	: Intel(R) Xeon Phi(TM) CPU 7285 @ 1.30GHz
stepping	: 0
microcode	: 0xe
cpu MHz		: 1064.235
cache size	: 1024 KB
physical id	: 0
siblings	: 272
core id		: 73
cpu cores	: 68
apicid		: 295
initial apicid	: 295
fpu		: yes
fpu_exception	: yes
cpuid level	: 13
wp		: yes
flags		: fpu vme de pse tsc msr pae mce cx8 apic sep mtrr pge mca cmov pat pse36 clflush dts acpi mmx fxsr sse sse2 ss ht tm pbe syscall nx pdpe1gb rdtscp lm constant_tsc arch_perfmon pebs bts rep_good nopl xtopology nonstop_tsc cpuid aperfmperf pni pclmulqdq dtes64 monitor ds_cpl vmx est tm2 ssse3 fma cx16 xtpr pdcm sse4_1 sse4_2 x2apic movbe popcnt tsc_deadline_timer aes xsave avx f16c rdrand lahf_lm abm 3dnowprefetch ring3mwait cpuid_fault epb pti tpr_shadow vnmi flexpriority ept vpid fsgsbase tsc_adjust bmi1 avx2 smep bmi2 erms avx512f rdseed adx avx512pf avx512er avx512cd xsaveopt dtherm ida arat pln pts avx512_vpopcntdq avx512_4vnniw avx512_4fmaps
bugs		: cpu_meltdown spectre_v1 spectre_v2
bogomips	: 2594.07
clflush size	: 64
cache_alignment	: 64
address sizes	: 46 bits physical, 48 bits virtual
power management:

uname -a
Linux nntpat98-196 4.18.0-rc7+ #2 SMP Thu Sep 6 13:24:37 MSK 2018 x86_64 x86_64 x86_64 GNU/Linux

cat /proc/sys/kernel/perf_event_paranoid
0

cat /proc/sys/kernel/perf_event_mlock_kb 
516

cat /proc/sys/kernel/perf_event_max_sample_rate 
3000

cat /etc/redhat-release 
Red Hat Enterprise Linux Server release 7.5 (Maipo)

Metrics:
runtime overhead (%) : elapsed_time_under_profiling / elapsed_time
data loss (%)        : paused_time / elapsed_time_under_profiling
LOST events          : stat from perf report --stats
SAMPLE events        : stat from perf report --stats
perf.data size (B)   : size of trace file on disk

Events:
cpu/period=P,event=0x3c/Duk;CPU_CLK_UNHALTED.THREAD
cpu/period=P,umask=0x3/Duk;CPU_CLK_UNHALTED.REF_TSC
cpu/period=P,event=0xc0/Duk;INST_RETIRED.ANY
cpu/period=0xaae61,event=0xc2,umask=0x10/uk;UOPS_RETIRED.ALL
cpu/period=0x11171,event=0xc2,umask=0x20/uk;UOPS_RETIRED.SCALAR_SIMD
cpu/period=0x11171,event=0xc2,umask=0x40/uk;UOPS_RETIRED.PACKED_SIMD

=================================================

Command:
/usr/bin/time /tmp/vtune_amplifier_2019.574715/bin64/perf.thr record --threads=T \
	-a -N -B -T -R --call-graph dwarf,1024 --user-regs=ip,bp,sp \
        -e cpu/period=P,event=0x3c/Duk,\
           cpu/period=P,umask=0x3/Duk,\
           cpu/period=P,event=0xc0/Duk,\
           cpu/period=0x30d40,event=0xc2,umask=0x10/uk,\
           cpu/period=0x4e20,event=0xc2,umask=0x20/uk,\
           cpu/period=0x4e20,event=0xc2,umask=0x40/uk \
         --clockid=monotonic_raw -- ./matrix.(icc|gcc)

Workload: matrix multiplication in 256 threads

/usr/bin/time ./matrix.icc
Addr of buf1 = 0x7ff9faa73010
Offs of buf1 = 0x7ff9faa73180
Addr of buf2 = 0x7ff9f8a72010
Offs of buf2 = 0x7ff9f8a721c0
Addr of buf3 = 0x7ff9f6a71010
Offs of buf3 = 0x7ff9f6a71100
Addr of buf4 = 0x7ff9f4a70010
Offs of buf4 = 0x7ff9f4a70140
Threads #: 256 Pthreads
Matrix size: 2048
Using multiply kernel: multiply1
Freq = 0.997720 GHz
Execution time = 9.061 seconds
1639.55user 6.59system 0:07.12elapsed 23094%CPU (0avgtext+0avgdata 100448maxresident)k
96inputs+0outputs (1major+33839minor)pagefaults 0swaps

T : 272
        P (period, ms)       : 0.1
	runtime overhead (%) : 45x ~ 323.54 / 7.12
	data loss (%)        : 96
	LOST events          : 323662
	SAMPLE events        : 31885479
        perf.data size (GiB) : 42

	P (period, ms)       : 0.25
	runtime overhead (%) : 25x ~ 180.76 / 7.12
	data loss (%)        : 69 
	LOST events          : 10636
	SAMPLE events        : 18692998
        perf.data size (GiB) : 23.5

	P (period, ms)       : 0.35 
	runtime overhead (%) : 16x ~ 119.49 / 7.12
	data loss (%)        : 1
	LOST events          : 6
	SAMPLE events        : 11178524
        perf.data size (GiB) : 14

T : 128
	P (period, ms)       : 0.35 
	runtime overhead (%) : 15x ~ 111.98 / 7.12
	data loss (%)        : 62
	LOST events          : 2825
	SAMPLE events        : 11267247
        perf.data size (GiB) : 15

T : 64
	P (period, ms)       : 0.35 
	runtime overhead (%) : 14x ~ 101.55 / 7.12
	data loss (%)        : 67
	LOST events          : 5155
	SAMPLE events        : 10966297
        perf.data size (GiB) : 13.7

Workload: matrix multiplication in 128 threads

/usr/bin/time ./matrix.gcc
Addr of buf1 = 0x7f072e630010
Offs of buf1 = 0x7f072e630180
Addr of buf2 = 0x7f072c62f010
Offs of buf2 = 0x7f072c62f1c0
Addr of buf3 = 0x7f072a62e010
Offs of buf3 = 0x7f072a62e100
Addr of buf4 = 0x7f072862d010
Offs of buf4 = 0x7f072862d140
Threads #: 128 Pthreads
Matrix size: 2048
Using multiply kernel: multiply1
Execution time = 6.639 seconds
767.03user 11.17system 0:06.81elapsed 11424%CPU (0avgtext+0avgdata 100756maxresident)k
88inputs+0outputs (0major+139898minor)pagefaults 0swaps

T : 272
        P (period, ms)       : 0.1
	runtime overhead (%) : 29x ~ 198.81 / 6.81
	data loss (%)        : 21
	LOST events          : 2502
	SAMPLE events        : 22481062
        perf.data size (GiB) : 27.6

	P (period, ms)       : 0.25
	runtime overhead (%) : 13x ~ 88.47 / 6.81
	data loss (%)        : 0
	LOST events          : 0
	SAMPLE events        : 9572787
        perf.data size (GiB) : 11.3

	P (period, ms)       : 0.35 
	runtime overhead (%) : 10x ~ 67.11 / 6.81
	data loss (%)        : 1
	LOST events          : 137
	SAMPLE events        : 6985930
        perf.data size (GiB) : 8

T : 128
	P (period, ms)       : 0.35 
	runtime overhead (%) : 9.5x ~ 64.33 / 6.81
	data loss (%)        : 1
	LOST events          : 3
	SAMPLE events        : 6666903
        perf.data size (GiB) : 7.8

T : 64
	P (period, ms)       : 0.25
	runtime overhead (%) : 17x ~ 114.27 / 6.81
	data loss (%)        : 2
	LOST events          : 52
	SAMPLE events        : 12643645
        perf.data size (GiB) : 15.5

	P (period, ms)       : 0.35 
	runtime overhead (%) : 10x ~ 68.60 / 6.81
	data loss (%)        : 1
	LOST events          : 93
	SAMPLE events        : 7164368
        perf.data size (GiB) : 8.5

Thanks,
Alexey

> 
>>
>> thanks,
>> jirka
>>
> 

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ