lists.openwall.net | lists / announce owl-users owl-dev john-users john-dev passwdqc-users yescrypt popa3d-users / oss-security kernel-hardening musl sabotage tlsify passwords / crypt-dev xvendor / Bugtraq Full-Disclosure linux-kernel linux-netdev linux-ext4 linux-hardening linux-cve-announce PHC | |
Open Source and information security mailing list archives
| ||
|
Date: Sun, 10 Aug 2014 18:54:13 +0800 From: Fengguang Wu <fengguang.wu@...el.com> To: Peter Zijlstra <peterz@...radead.org> Cc: Vincent Guittot <vincent.guittot@...aro.org>, Dave Hansen <dave.hansen@...el.com>, LKML <linux-kernel@...r.kernel.org>, lkp@...org, Ingo Molnar <mingo@...nel.org>, Dietmar Eggemann <dietmar.eggemann@....com>, Preeti U Murthy <preeti@...ux.vnet.ibm.com> Subject: Re: [sched] 143e1e28cb4: +17.9% aim7.jobs-per-min, -9.7% hackbench.throughput On Sun, Aug 10, 2014 at 09:59:15AM +0200, Peter Zijlstra wrote: > On Sun, Aug 10, 2014 at 12:41:27PM +0800, Fengguang Wu wrote: > > Hi Vincent, > > > > FYI, we noticed some performance ups/downs on > > > > commit 143e1e28cb40bed836b0a06567208bd7347c9672 ("sched: Rework sched_domain topology definition") > > > > 128529 ± 1% +17.9% 151594 ± 0% brickland1/aim7/6000-page_test > > 76064 ± 3% -32.2% 51572 ± 6% brickland1/aim7/6000-page_test > > 59366697 ± 3% -46.1% 32017187 ± 7% brickland1/aim7/6000-page_test > > 2561 ± 7% -42.9% 1463 ± 9% brickland1/aim7/6000-page_test > > 9926 ± 2% -43.8% 5577 ± 4% brickland1/aim7/6000-page_test > > 19542 ± 9% -38.3% 12057 ± 4% brickland1/aim7/6000-page_test > > 993654 ± 2% -19.9% 795962 ± 3% brickland1/aim7/6000-page_test > > etc.. > > how does one read that? afaict its a random number generator.. The "brickland1/aim7/6000-page_test" is the test case part. The "TOTAL XXX" is the metric part. One test run may generate lots of metrics, reflecting different aspect of the system dynamics. This view may be easier to read, by grouping the metrics by test case. test case: brickland1/aim7/6000-page_test 128529 ± 1% +17.9% 151594 ± 0% TOTAL aim7.jobs-per-min 582269 ±14% -55.6% 258617 ±16% TOTAL softirqs.SCHED 59366697 ± 3% -46.1% 32017187 ± 7% TOTAL cpuidle.C1-IVT.time 54543 ±11% -37.2% 34252 ±16% TOTAL cpuidle.C1-IVT.usage 2561 ± 7% -42.9% 1463 ± 9% TOTAL numa-numastat.node2.other_node 9926 ± 2% -43.8% 5577 ± 4% TOTAL proc-vmstat.numa_other 2627 ±12% -49.1% 1337 ±12% TOTAL numa-numastat.node1.other_node 19542 ± 9% -38.3% 12057 ± 4% TOTAL cpuidle.C1E-IVT.usage 2455 ±10% -41.0% 1448 ± 9% TOTAL numa-numastat.node0.other_node 471304 ±11% -31.4% 323251 ± 8% TOTAL numa-vmstat.node1.nr_anon_pages 2281 ±12% -41.8% 1327 ±16% TOTAL numa-numastat.node3.other_node 1903446 ±11% -30.7% 1318156 ± 7% TOTAL numa-meminfo.node1.AnonPages 518274 ±11% -30.4% 360742 ± 8% TOTAL numa-vmstat.node1.nr_active_anon 2097138 ±10% -30.0% 1469003 ± 8% TOTAL numa-meminfo.node1.Active(anon) 49527464 ± 6% -32.4% 33488833 ± 4% TOTAL cpuidle.C1E-IVT.time 2118206 ±10% -29.7% 1488874 ± 7% TOTAL numa-meminfo.node1.Active 76064 ± 3% -32.2% 51572 ± 6% TOTAL cpuidle.C6-IVT.usage 188938 ±33% -41.3% 110966 ±16% TOTAL numa-meminfo.node2.PageTables 47262 ±35% -42.3% 27273 ±16% TOTAL numa-vmstat.node2.nr_page_table_pages 1944687 ±10% -25.8% 1443923 ±16% TOTAL numa-meminfo.node3.Active(anon) 1754763 ±11% -26.6% 1288713 ±16% TOTAL numa-meminfo.node3.AnonPages 1964722 ±10% -25.5% 1464696 ±16% TOTAL numa-meminfo.node3.Active 432109 ± 9% -26.2% 318886 ±14% TOTAL numa-vmstat.node3.nr_anon_pages 479527 ± 9% -25.3% 358029 ±14% TOTAL numa-vmstat.node3.nr_active_anon 463719 ± 8% -24.7% 349388 ± 7% TOTAL numa-vmstat.node0.nr_anon_pages 3157742 ±16% -26.5% 2320253 ±10% TOTAL numa-meminfo.node1.MemUsed 7303589 ± 2% -24.8% 5495829 ± 3% TOTAL meminfo.AnonPages 8064024 ± 2% -24.0% 6132677 ± 3% TOTAL meminfo.Active(anon) 511455 ± 8% -23.9% 389447 ± 7% TOTAL numa-vmstat.node0.nr_active_anon 1818612 ± 2% -24.9% 1365670 ± 3% TOTAL proc-vmstat.nr_anon_pages 2007155 ± 2% -24.3% 1518688 ± 3% TOTAL proc-vmstat.nr_active_anon 8145316 ± 2% -23.7% 6213832 ± 3% TOTAL meminfo.Active 1850230 ± 8% -24.1% 1405061 ± 8% TOTAL numa-meminfo.node0.AnonPages 6.567e+11 ± 3% -21.4% 5.16e+11 ± 4% TOTAL meminfo.Committed_AS 2044097 ± 7% -23.5% 1562809 ± 8% TOTAL numa-meminfo.node0.Active(anon) 2064106 ± 7% -23.3% 1582792 ± 8% TOTAL numa-meminfo.node0.Active 235358 ± 5% -19.8% 188793 ± 3% TOTAL proc-vmstat.pgmigrate_success 235358 ± 5% -19.8% 188793 ± 3% TOTAL proc-vmstat.numa_pages_migrated 433235 ± 4% -18.1% 354845 ± 5% TOTAL numa-vmstat.node2.nr_anon_pages 198747 ±23% -28.0% 143034 ± 3% TOTAL proc-vmstat.nr_page_table_pages 3187 ± 5% -18.5% 2599 ± 6% TOTAL numa-vmstat.node0.numa_other 796281 ±23% -27.7% 575352 ± 3% TOTAL meminfo.PageTables 1395062 ± 6% -19.0% 1130108 ± 3% TOTAL proc-vmstat.numa_hint_faults 477037 ± 4% -17.2% 394983 ± 5% TOTAL numa-vmstat.node2.nr_active_anon 2829 ±10% +18.7% 3357 ± 3% TOTAL numa-vmstat.node2.nr_alloc_batch 993654 ± 2% -19.9% 795962 ± 3% TOTAL softirqs.RCU 2706 ± 4% +26.1% 3411 ± 5% TOTAL numa-vmstat.node1.nr_alloc_batch 2725835 ± 4% -17.5% 2247537 ± 4% TOTAL numa-meminfo.node2.MemUsed 393637 ± 6% -15.3% 333296 ± 2% TOTAL proc-vmstat.numa_hint_faults_local 2.82 ± 3% +21.9% 3.43 ± 4% TOTAL turbostat.%pc2 4.40 ± 2% +22.0% 5.37 ± 4% TOTAL turbostat.%c6 1742111 ± 4% -16.9% 1447181 ± 5% TOTAL numa-meminfo.node2.AnonPages 15865125 ± 1% -15.0% 13485882 ± 1% TOTAL softirqs.TIMER 1923000 ± 4% -16.4% 1608509 ± 5% TOTAL numa-meminfo.node2.Active(anon) 1943185 ± 4% -16.2% 1629057 ± 5% TOTAL numa-meminfo.node2.Active 3077 ± 1% +14.5% 3523 ± 0% TOTAL proc-vmstat.pgactivate 329 ± 1% -13.3% 285 ± 0% TOTAL uptime.boot 13158 ±13% -14.4% 11261 ± 4% TOTAL numa-meminfo.node3.SReclaimable 3289 ±13% -14.4% 2815 ± 4% TOTAL numa-vmstat.node3.nr_slab_reclaimable 3150464 ± 2% -24.2% 2387551 ± 3% TOTAL time.voluntary_context_switches 281 ± 1% -15.1% 238 ± 0% TOTAL time.elapsed_time 29294 ± 1% -14.3% 25093 ± 0% TOTAL time.system_time 4529818 ± 1% -8.8% 4129398 ± 1% TOTAL time.involuntary_context_switches 15.75 ± 1% -3.4% 15.21 ± 0% TOTAL turbostat.RAM_W 10655 ± 0% +1.4% 10802 ± 0% TOTAL time.percent_of_cpu_this_job_got Thanks, Fengguang -- To unsubscribe from this list: send the line "unsubscribe linux-kernel" in the body of a message to majordomo@...r.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html Please read the FAQ at http://www.tux.org/lkml/
Powered by blists - more mailing lists