lists.openwall.net | lists / announce owl-users owl-dev john-users john-dev passwdqc-users yescrypt popa3d-users / oss-security kernel-hardening musl sabotage tlsify passwords / crypt-dev xvendor / Bugtraq Full-Disclosure linux-kernel linux-netdev linux-ext4 linux-hardening linux-cve-announce PHC | |
Open Source and information security mailing list archives
| ||
|
Date: Mon, 27 Oct 2014 10:14:25 +0800 From: kernel test robot <fengguang.wu@...el.com> To: Jason Low <jason.low2@...com> Cc: Ingo Molnar <mingo@...nel.org>, LKML <linux-kernel@...r.kernel.org>, lkp@...org, Yuanhan Liu <yuanhan.liu@...ux.intel.com> Subject: [LKP] [sched] 8236d907ab3: 2% vm-scalability.throughput FYI, we noticed the below changes on commit 8236d907ab3411ad452280faa8b26c1347327380 ("sched: Reduce contention in update_cfs_rq_blocked_load()") 5cd038f53ed9ec7a 8236d907ab3411ad452280faa8 testbox/testcase/testparams ---------------- -------------------------- --------------------------- %stddev %change %stddev \ | \ 8957484 ± 0% +1.9% 9123335 ± 0% lkp-nex06/vm-scalability/performance-300s-anon-r-rand-mt 8977343 ± 0% +2.0% 9152823 ± 0% lkp-nex06/vm-scalability/performance-300s-anon-r-rand 8967408 +1.9% 9138067 GEO-MEAN vm-scalability.throughput 5cd038f53ed9ec7a 8236d907ab3411ad452280faa8 ---------------- -------------------------- 4.05 ± 5% -99.0% 0.04 ± 0% lkp-nex06/vm-scalability/performance-300s-anon-r-rand-mt 3.95 ± 4% -98.9% 0.04 ± 9% lkp-nex06/vm-scalability/performance-300s-anon-r-rand 4.00 -99.0% 0.04 GEO-MEAN perf-profile.cpu-cycles.task_tick_fair.scheduler_tick.update_process_times.tick_sched_handle.tick_sched_timer 5cd038f53ed9ec7a 8236d907ab3411ad452280faa8 ---------------- -------------------------- 3.88 ± 4% -99.5% 0.02 ± 0% lkp-nex06/vm-scalability/performance-300s-anon-r-rand-mt 3.83 ± 2% -99.5% 0.02 ± 0% lkp-nex06/vm-scalability/performance-300s-anon-r-rand 3.85 -99.5% 0.02 GEO-MEAN perf-profile.cpu-cycles.update_cfs_rq_blocked_load.task_tick_fair.scheduler_tick.update_process_times.tick_sched_handle 5cd038f53ed9ec7a 8236d907ab3411ad452280faa8 ---------------- -------------------------- 61 ± 43% -70.8% 18 ± 24% lkp-nex06/vm-scalability/performance-300s-anon-r-rand 61 -70.8% 17 GEO-MEAN sched_debug.cfs_rq[50]:/.tg_load_contrib 5cd038f53ed9ec7a 8236d907ab3411ad452280faa8 ---------------- -------------------------- 7257 ± 37% +106.2% 14962 ± 49% lkp-nex06/vm-scalability/performance-300s-anon-r-rand 7257 +106.2% 14962 GEO-MEAN sched_debug.cpu#1.sched_count 5cd038f53ed9ec7a 8236d907ab3411ad452280faa8 ---------------- -------------------------- 0.06 ± 15% +96.8% 0.12 ± 18% lkp-nex06/vm-scalability/performance-300s-anon-r-rand-mt 0.06 +96.8% 0.12 GEO-MEAN turbostat.%c1 5cd038f53ed9ec7a 8236d907ab3411ad452280faa8 ---------------- -------------------------- 415 ± 14% -24.7% 313 ± 33% lkp-nex06/vm-scalability/performance-300s-anon-r-rand-mt 415 -24.7% 312 GEO-MEAN slabinfo.blkdev_queue.num_objs 5cd038f53ed9ec7a 8236d907ab3411ad452280faa8 ---------------- -------------------------- 2080 ± 21% +78.4% 3711 ± 44% lkp-nex06/vm-scalability/performance-300s-anon-r-rand 2080 +78.4% 3711 GEO-MEAN sched_debug.cpu#33.ttwu_count 5cd038f53ed9ec7a 8236d907ab3411ad452280faa8 ---------------- -------------------------- 58 ± 39% +93.1% 112 ± 34% lkp-nex06/vm-scalability/performance-300s-anon-r-rand-mt 58 +93.1% 112 GEO-MEAN sched_debug.cfs_rq[11]:/.tg_load_contrib 5cd038f53ed9ec7a 8236d907ab3411ad452280faa8 ---------------- -------------------------- 5708 ± 13% -14.8% 4862 ± 19% lkp-nex06/vm-scalability/performance-300s-anon-r-rand 5708 -14.8% 4862 GEO-MEAN meminfo.AnonHugePages 5cd038f53ed9ec7a 8236d907ab3411ad452280faa8 ---------------- -------------------------- 487 ± 9% +42.3% 693 ± 14% lkp-nex06/vm-scalability/performance-300s-anon-r-rand-mt 487 +42.3% 693 GEO-MEAN sched_debug.cpu#20.sched_goidle 5cd038f53ed9ec7a 8236d907ab3411ad452280faa8 ---------------- -------------------------- 15 ± 10% +37.3% 20 ± 12% lkp-nex06/vm-scalability/performance-300s-anon-r-rand 15 +37.3% 20 GEO-MEAN sched_debug.cpu#26.load 5cd038f53ed9ec7a 8236d907ab3411ad452280faa8 ---------------- -------------------------- 18 ± 7% -19.4% 15 ± 5% lkp-nex06/vm-scalability/performance-300s-anon-r-rand 18 -19.4% 15 GEO-MEAN sched_debug.cpu#49.cpu_load[3] 5cd038f53ed9ec7a 8236d907ab3411ad452280faa8 ---------------- -------------------------- 231 ± 14% -25.0% 173 ± 20% lkp-nex06/vm-scalability/performance-300s-anon-r-rand 231 -25.0% 173 GEO-MEAN sched_debug.cpu#58.sched_goidle 5cd038f53ed9ec7a 8236d907ab3411ad452280faa8 ---------------- -------------------------- 1441314 ± 3% +26.9% 1829712 ± 2% lkp-nex06/vm-scalability/performance-300s-anon-r-rand-mt 1414086 ± 8% +23.6% 1747307 ± 8% lkp-nex06/vm-scalability/performance-300s-anon-r-rand 1427635 +25.2% 1788034 GEO-MEAN softirqs.RCU 5cd038f53ed9ec7a 8236d907ab3411ad452280faa8 ---------------- -------------------------- 17 ± 6% -16.9% 14 ± 5% lkp-nex06/vm-scalability/performance-300s-anon-r-rand 17 -16.9% 14 GEO-MEAN sched_debug.cpu#49.cpu_load[4] 5cd038f53ed9ec7a 8236d907ab3411ad452280faa8 ---------------- -------------------------- 19 ± 9% -16.8% 15 ± 7% lkp-nex06/vm-scalability/performance-300s-anon-r-rand 18 -16.8% 15 GEO-MEAN sched_debug.cpu#49.cpu_load[2] 5cd038f53ed9ec7a 8236d907ab3411ad452280faa8 ---------------- -------------------------- 1196 ± 8% +17.3% 1403 ± 5% lkp-nex06/vm-scalability/performance-300s-anon-r-rand-mt 1196 +17.3% 1403 GEO-MEAN sched_debug.cpu#49.sched_goidle 5cd038f53ed9ec7a 8236d907ab3411ad452280faa8 ---------------- -------------------------- 98963 ± 0% +52.9% 151281 ± 38% lkp-nex06/vm-scalability/performance-300s-anon-r-rand 98963 +52.9% 151281 GEO-MEAN numa-meminfo.node3.FilePages 5cd038f53ed9ec7a 8236d907ab3411ad452280faa8 ---------------- -------------------------- 24740 ± 0% +52.9% 37819 ± 38% lkp-nex06/vm-scalability/performance-300s-anon-r-rand 24740 +52.9% 37819 GEO-MEAN numa-vmstat.node3.nr_file_pages 5cd038f53ed9ec7a 8236d907ab3411ad452280faa8 ---------------- -------------------------- 38000 ± 1% +15.9% 44042 ± 0% lkp-nex06/vm-scalability/performance-300s-anon-r-rand-mt 38312 ± 2% +13.9% 43643 ± 1% lkp-nex06/vm-scalability/performance-300s-anon-r-rand 38156 +14.9% 43842 GEO-MEAN proc-vmstat.pgactivate 5cd038f53ed9ec7a 8236d907ab3411ad452280faa8 ---------------- -------------------------- 3.29 ± 6% +15.3% 3.80 ± 0% lkp-nex06/vm-scalability/performance-300s-anon-r-rand-mt 3.33 ± 2% +14.0% 3.79 ± 0% lkp-nex06/vm-scalability/performance-300s-anon-r-rand 3.31 +14.7% 3.79 GEO-MEAN perf-profile.cpu-cycles.__drand48_iterate 5cd038f53ed9ec7a 8236d907ab3411ad452280faa8 ---------------- -------------------------- 3.34 ± 6% +14.9% 3.83 ± 0% lkp-nex06/vm-scalability/performance-300s-anon-r-rand-mt 3.37 ± 2% +13.8% 3.84 ± 0% lkp-nex06/vm-scalability/performance-300s-anon-r-rand 3.35 +14.3% 3.83 GEO-MEAN perf-profile.cpu-cycles.__nrand48_r 5cd038f53ed9ec7a 8236d907ab3411ad452280faa8 ---------------- -------------------------- 81.57 ± 0% +10.8% 90.40 ± 0% lkp-nex06/vm-scalability/performance-300s-anon-r-rand-mt 81.61 ± 0% +10.8% 90.40 ± 0% lkp-nex06/vm-scalability/performance-300s-anon-r-rand 81.59 +10.8% 90.40 GEO-MEAN perf-profile.cpu-cycles.do_unit 5cd038f53ed9ec7a 8236d907ab3411ad452280faa8 ---------------- -------------------------- 39878 ± 0% -9.5% 36107 ± 0% lkp-nex06/vm-scalability/performance-300s-anon-r-rand-mt 39774 ± 0% -9.3% 36061 ± 0% lkp-nex06/vm-scalability/performance-300s-anon-r-rand 39826 -9.4% 36084 GEO-MEAN proc-vmstat.nr_shmem 5cd038f53ed9ec7a 8236d907ab3411ad452280faa8 ---------------- -------------------------- 159494 ± 0% -9.5% 144409 ± 0% lkp-nex06/vm-scalability/performance-300s-anon-r-rand-mt 159156 ± 0% -9.3% 144313 ± 0% lkp-nex06/vm-scalability/performance-300s-anon-r-rand 159325 -9.4% 144361 GEO-MEAN meminfo.Shmem 5cd038f53ed9ec7a 8236d907ab3411ad452280faa8 ---------------- -------------------------- 144353 ± 2% -3.7% 138984 ± 2% lkp-nex06/vm-scalability/performance-300s-anon-r-rand-mt 144353 -3.7% 138984 GEO-MEAN time.involuntary_context_switches lkp-nex06: Nehalem-EX Memory: 64G vm-scalability.throughput 9.15e+06 ++---------------------------------------------------------------+ O O O O O O O O O O O O O O O O O O O O O | | O O O O O O 9.1e+06 ++ | | | | | 9.05e+06 ++ | | | 9e+06 ++ *.. | | : .*.*..*. .*.*.. | | .* .*.. : *.*. *. .* *. .* | 8.95e+06 *+ + .* * *.*. : + *. | | *. : + | | * | 8.9e+06 ++---------------------------------------------------------------+ perf-profile.cpu-cycles.do_unit 91 ++---------------------------------------------------------------------+ 90 O+ O O O O O O O O O O O O O O O O O O O O O O O O O O | | 89 ++ | 88 ++ | | | 87 ++ | 86 ++ | 85 ++ | | | 84 ++ | 83 ++ .*.. .*.. | | .* *. *.*..*..*..*. | 82 *+.*.*.. .*. *.. .*. .*.. .*.* | 81 ++------*--------------------------------*----*-----*------------------+ 4.5 ++--------------------------------------------------------------------+ | *.. .*..*..*.*..*.. | 4 *+. + *..*. * *.. | 3.5 ++ * *.. *..*. .. * | | .. *..*..*.*..* | 3 ++ * | 2.5 ++ | | | 2 ++ | 1.5 ++ | | | 1 ++ | 0.5 ++ | | | 0 O+-O-O--O--O-O--O--O--O-O--O--O-O--O--O-O--O--O-O--O--O--O-O--O--O-O--O 4.5 ++--------------------------------------------------------------------+ | .*. .*.. *.. .*.. | 4 *+ *. .*.. + *..* *..*.. | 3.5 ++ *.*.. .*.. *..*..*.*. * * | | *. + | 3 ++ * | 2.5 ++ | | | 2 ++ | 1.5 ++ | | | 1 ++ | 0.5 ++ | | | 0 O+-O-O--O--O-O--O--O--O-O--O--O-O--O--O-O--O--O-O--O--O--O-O--O--O-O--O [*] bisect-good sample [O] bisect-bad sample To reproduce: apt-get install ruby ruby-oj git clone git://git.kernel.org/pub/scm/linux/kernel/git/wfg/lkp-tests.git cd lkp-tests bin/setup-local job.yaml # the job file attached in this email bin/run-local job.yaml Disclaimer: Results have been estimated based on internal Intel analysis and are provided for informational purposes only. Any difference in system hardware or software design or configuration may affect actual performance. Thanks, Fengguang View attachment "job.yaml" of type "text/plain" (1663 bytes) View attachment "reproduce" of type "text/plain" (5107 bytes) _______________________________________________ LKP mailing list LKP@...ux.intel.com
Powered by blists - more mailing lists