lists.openwall.net | lists / announce owl-users owl-dev john-users john-dev passwdqc-users yescrypt popa3d-users / oss-security kernel-hardening musl sabotage tlsify passwords / crypt-dev xvendor / Bugtraq Full-Disclosure linux-kernel linux-netdev linux-ext4 linux-hardening linux-cve-announce PHC | |
Open Source and information security mailing list archives
| ||
|
Date: Mon, 7 Nov 2016 10:48:10 +0800 From: kernel test robot <xiaolong.ye@...el.com> To: Huang Ying <ying.huang@...el.com> Cc: Linus Torvalds <torvalds@...ux-foundation.org>, Rik van Riel <riel@...hat.com>, Hugh Dickins <hughd@...gle.com>, Shaohua Li <shli@...nel.org>, Minchan Kim <minchan@...nel.org>, Mel Gorman <mgorman@...hsingularity.net>, Tejun Heo <tj@...nel.org>, Wu Fengguang <fengguang.wu@...el.com>, Dave Hansen <dave.hansen@...el.com>, Andrew Morton <akpm@...ux-foundation.org>, LKML <linux-kernel@...r.kernel.org>, lkp@...org Subject: [lkp] [mm] 371a096edf: vm-scalability.throughput 6.1% improvement Greeting, FYI, we noticed a 6.1% improvement of vm-scalability.throughput due to commit: commit 371a096edf43a8c71844cf71c20765c8b21d07d9 ("mm: don't use radix tree writeback tags for pages in swap cache") https://git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux.git master in testcase: vm-scalability on test machine: 72 threads Intel(R) Xeon(R) CPU E5-2699 v3 @ 2.30GHz with 128G memory with following parameters: runtime: 300 thp_enabled: never thp_defrag: never nr_task: 16 nr_ssd: 1 test: swap-w-seq cpufreq_governor: performance The motivation behind this suite is to exercise functions and regions of the mm/ of the Linux kernel which are of interest to us. In addition to that, the commit also has significant impact on the following tests: +------------------+-----------------------------------------------------------------------+ | testcase: change | vm-scalability: vm-scalability.throughput 14.1% improvement | | test machine | 72 threads Intel(R) Xeon(R) CPU E5-2699 v3 @ 2.30GHz with 128G memory | | test parameters | cpufreq_governor=performance | | | nr_ssd=1 | | | nr_task=64 | | | runtime=300 | | | test=swap-w-seq | | | thp_defrag=never | | | thp_enabled=always | +------------------+-----------------------------------------------------------------------+ Disclaimer: Results have been estimated based on internal Intel analysis and are provided for informational purposes only. Any difference in system hardware or software design or configuration may affect actual performance. Details are as below: --------------------------------------------------------------------------------------------------> To reproduce: git clone git://git.kernel.org/pub/scm/linux/kernel/git/wfg/lkp-tests.git cd lkp-tests bin/lkp install job.yaml # job file is attached in this email bin/lkp run job.yaml ========================================================================================= compiler/cpufreq_governor/kconfig/nr_ssd/nr_task/rootfs/runtime/tbox_group/test/testcase/thp_defrag/thp_enabled: gcc-6/performance/x86_64-rhel-7.2/1/16/debian-x86_64-2016-08-31.cgz/300/lkp-hsw-ep4/swap-w-seq/vm-scalability/never/never commit: 1d8bf926f8 ("mm/bootmem.c: replace kzalloc() by kzalloc_node()") 371a096edf ("mm: don't use radix tree writeback tags for pages in swap cache") 1d8bf926f8739bd3 371a096edf43a8c71844cf71c2 ---------------- -------------------------- fail:runs %reproduction fail:runs | | | %stddev %change %stddev \ | \ 2213672 ± 0% +6.1% 2348422 ± 1% vm-scalability.throughput 534544 ± 1% +6.8% 570794 ± 4% vm-scalability.time.involuntary_context_switches 1.458e+08 ± 0% +7.0% 1.56e+08 ± 1% vm-scalability.time.minor_page_faults 4991 ± 50% +3615.9% 185462 ± 31% vm-scalability.time.voluntary_context_switches 0.08 ± 27% +3440.8% 2.69 ± 50% turbostat.CPU%c3 229747 ± 5% +47.7% 339402 ± 11% softirqs.RCU 270638 ± 1% +13.0% 305899 ± 2% softirqs.SCHED 1277270 ± 0% +7.8% 1377440 ± 1% vmstat.io.bo 1277264 ± 0% +7.8% 1377434 ± 1% vmstat.swap.so 5148 ± 0% +50.8% 7765 ± 9% vmstat.system.cs 110151 ± 0% +30.6% 143842 ± 11% meminfo.SUnreclaim 9703 ± 20% -32.8% 6516 ± 28% meminfo.Shmem 783039 ± 5% -26.3% 577083 ± 17% meminfo.SwapCached 1167 ± 27% +27840.3% 326175 ± 61% meminfo.Writeback 637164 ± 5% -25.4% 475615 ± 9% numa-meminfo.node0.FilePages 243.40 ± 36% +19536.4% 47794 ± 58% numa-meminfo.node0.Writeback 50438 ± 4% +31.7% 66437 ± 12% numa-meminfo.node1.SUnreclaim 83150 ± 4% +17.5% 97667 ± 8% numa-meminfo.node1.Slab 1045 ± 24% +25899.2% 271899 ± 60% numa-meminfo.node1.Writeback 159271 ± 5% -25.4% 118746 ± 9% numa-vmstat.node0.nr_file_pages 59.60 ± 37% +19680.4% 11789 ± 61% numa-vmstat.node0.nr_writeback 60.00 ± 37% +19544.0% 11786 ± 61% numa-vmstat.node0.nr_zone_write_pending 12609 ± 4% +31.5% 16579 ± 12% numa-vmstat.node1.nr_slab_unreclaimable 265.30 ± 19% +25322.4% 67445 ± 61% numa-vmstat.node1.nr_writeback 266.80 ± 19% +25179.8% 67446 ± 61% numa-vmstat.node1.nr_zone_write_pending 7424364 ± 50% +715.3% 60527951 ± 47% cpuidle.C1-HSW.time 85779 ± 8% +394.3% 423973 ± 36% cpuidle.C1-HSW.usage 11362590 ± 16% +1931.5% 2.308e+08 ± 58% cpuidle.C1E-HSW.time 58721 ± 12% +1054.1% 677675 ± 52% cpuidle.C1E-HSW.usage 60495231 ± 13% +1096.4% 7.237e+08 ± 46% cpuidle.C3-HSW.time 181445 ± 13% +599.7% 1269496 ± 43% cpuidle.C3-HSW.usage 2126 ± 9% +798.9% 19117 ± 46% cpuidle.POLL.usage 40653369 ± 70% -88.3% 4756647 ± 58% proc-vmstat.compact_migrate_scanned 370154 ± 3% -13.7% 319306 ± 8% proc-vmstat.nr_file_pages 27536 ± 0% +30.4% 35899 ± 11% proc-vmstat.nr_slab_unreclaimable 71520720 ± 5% +12.4% 80357490 ± 3% proc-vmstat.nr_vmscan_write 292.40 ± 31% +27493.5% 80683 ± 62% proc-vmstat.nr_writeback 294.00 ± 31% +27342.9% 80682 ± 62% proc-vmstat.nr_zone_write_pending 1.262e+08 ± 3% +15.9% 1.463e+08 ± 1% proc-vmstat.numa_pte_updates 87331985 ± 1% +33.7% 1.168e+08 ± 3% proc-vmstat.pgrotated 28607602 ± 2% +22.0% 34898577 ± 2% proc-vmstat.pgscan_kswapd 14766402 ± 4% +20.4% 17782412 ± 4% proc-vmstat.pgsteal_kswapd 1828938 ± 1% +52.5% 2790009 ± 9% perf-stat.context-switches 17676 ± 6% +309.2% 72326 ± 29% perf-stat.cpu-migrations 0.06 ± 11% +112.1% 0.13 ± 23% perf-stat.dTLB-load-miss-rate% 9.969e+08 ± 13% +93.7% 1.931e+09 ± 18% perf-stat.dTLB-load-misses 0.22 ± 2% +10.0% 0.24 ± 5% perf-stat.dTLB-store-miss-rate% 8.69e+08 ± 4% +18.2% 1.027e+09 ± 9% perf-stat.dTLB-store-misses 36.04 ± 2% +8.4% 39.05 ± 2% perf-stat.iTLB-load-miss-rate% 1.331e+08 ± 3% +19.8% 1.594e+08 ± 6% perf-stat.iTLB-load-misses 1.463e+08 ± 0% +7.0% 1.565e+08 ± 1% perf-stat.minor-faults 60.76 ± 1% +2.8% 62.46 ± 0% perf-stat.node-store-miss-rate% 1.463e+08 ± 0% +7.0% 1.565e+08 ± 1% perf-stat.page-faults 178.50 ± 14% +201.1% 537.40 ± 14% slabinfo.bdev_cache.active_objs 178.50 ± 14% +201.1% 537.40 ± 14% slabinfo.bdev_cache.num_objs 444.80 ± 12% +66.6% 741.10 ± 13% slabinfo.file_lock_cache.active_objs 444.80 ± 12% +66.6% 741.10 ± 13% slabinfo.file_lock_cache.num_objs 4019 ± 1% +100.4% 8054 ± 13% slabinfo.kmalloc-1024.active_objs 128.10 ± 0% +100.5% 256.90 ± 14% slabinfo.kmalloc-1024.active_slabs 4100 ± 0% +100.8% 8232 ± 14% slabinfo.kmalloc-1024.num_objs 128.10 ± 0% +100.5% 256.90 ± 14% slabinfo.kmalloc-1024.num_slabs 7578 ± 0% +21.8% 9232 ± 7% slabinfo.kmalloc-192.active_objs 7609 ± 0% +21.4% 9238 ± 7% slabinfo.kmalloc-192.num_objs 4829 ± 1% +47.7% 7134 ± 11% slabinfo.kmalloc-2048.active_objs 312.10 ± 2% +45.2% 453.10 ± 11% slabinfo.kmalloc-2048.active_slabs 4898 ± 1% +47.9% 7246 ± 11% slabinfo.kmalloc-2048.num_objs 312.10 ± 2% +45.2% 453.10 ± 11% slabinfo.kmalloc-2048.num_slabs 16028 ± 6% +479.7% 92924 ± 51% slabinfo.kmalloc-256.active_objs 351.20 ± 4% +1071.3% 4113 ± 56% slabinfo.kmalloc-256.active_slabs 16315 ± 5% +487.8% 95897 ± 51% slabinfo.kmalloc-256.num_objs 351.20 ± 4% +1071.3% 4113 ± 56% slabinfo.kmalloc-256.num_slabs 751.10 ± 6% +74.9% 1313 ± 14% slabinfo.nsproxy.active_objs 751.10 ± 6% +74.9% 1313 ± 14% slabinfo.nsproxy.num_objs 37639 ± 4% -16.0% 31603 ± 6% slabinfo.radix_tree_node.active_objs 705.70 ± 4% -17.4% 583.10 ± 7% slabinfo.radix_tree_node.active_slabs 39498 ± 3% -17.4% 32627 ± 7% slabinfo.radix_tree_node.num_objs 705.70 ± 4% -17.4% 583.10 ± 7% slabinfo.radix_tree_node.num_slabs 146695 ± 1% -28.0% 105561 ± 17% sched_debug.cfs_rq:/.exec_clock.max 57.20 ± 73% +574.1% 385.58 ± 26% sched_debug.cfs_rq:/.exec_clock.min 50781 ± 2% -26.8% 37173 ± 18% sched_debug.cfs_rq:/.exec_clock.stddev 150.31 ± 5% -25.0% 112.67 ± 17% sched_debug.cfs_rq:/.load_avg.stddev 2792970 ± 4% -40.8% 1653864 ± 29% sched_debug.cfs_rq:/.min_vruntime.max 918436 ± 5% -38.0% 569297 ± 31% sched_debug.cfs_rq:/.min_vruntime.stddev 842.53 ± 5% -31.2% 579.83 ± 18% sched_debug.cfs_rq:/.runnable_load_avg.max 134.83 ± 5% -34.0% 89.01 ± 24% sched_debug.cfs_rq:/.runnable_load_avg.stddev 918795 ± 5% -37.9% 570175 ± 31% sched_debug.cfs_rq:/.spread0.stddev 380.58 ± 2% -14.8% 324.26 ± 10% sched_debug.cfs_rq:/.util_avg.stddev 42.86 ± 17% +1021.2% 480.60 ± 88% sched_debug.cpu.clock.stddev 42.86 ± 17% +1021.2% 480.60 ± 88% sched_debug.cpu.clock_task.stddev 843.67 ± 5% -31.3% 579.83 ± 18% sched_debug.cpu.cpu_load[0].max 134.98 ± 5% -34.0% 89.02 ± 24% sched_debug.cpu.cpu_load[0].stddev 136.36 ± 5% -30.6% 94.67 ± 26% sched_debug.cpu.cpu_load[3].stddev 136.89 ± 5% -31.3% 94.00 ± 26% sched_debug.cpu.cpu_load[4].stddev 441.52 ± 5% -16.4% 369.26 ± 13% sched_debug.cpu.curr->pid.avg 0.00 ± 10% +846.3% 0.00 ± 86% sched_debug.cpu.next_balance.stddev 155148 ± 0% -16.9% 128991 ± 11% sched_debug.cpu.nr_load_updates.max 38237 ± 2% -27.9% 27553 ± 19% sched_debug.cpu.nr_load_updates.stddev 12989 ± 1% +66.7% 21655 ± 11% sched_debug.cpu.nr_switches.avg 38611 ± 9% +50.9% 58253 ± 13% sched_debug.cpu.nr_switches.max 1120 ± 32% +350.3% 5042 ± 29% sched_debug.cpu.nr_switches.min 10487 ± 3% +27.6% 13383 ± 9% sched_debug.cpu.nr_switches.stddev 0.01 ± 14% +1005.7% 0.09 ± 38% sched_debug.cpu.nr_uninterruptible.avg 22.03 ± 19% +124.5% 49.47 ± 28% sched_debug.cpu.nr_uninterruptible.max -25.55 ±-32% +173.8% -69.95 ±-27% sched_debug.cpu.nr_uninterruptible.min 7.95 ± 16% +144.6% 19.46 ± 18% sched_debug.cpu.nr_uninterruptible.stddev 13007 ± 2% +66.2% 21615 ± 11% sched_debug.cpu.sched_count.avg 664.22 ± 53% +590.3% 4584 ± 32% sched_debug.cpu.sched_count.min 1797 ± 3% +189.0% 5196 ± 21% sched_debug.cpu.sched_goidle.avg 112.57 ± 48% +1648.7% 1968 ± 35% sched_debug.cpu.sched_goidle.min 6018 ± 1% +76.9% 10646 ± 12% sched_debug.cpu.ttwu_count.avg 17865 ± 3% +107.9% 37145 ± 21% sched_debug.cpu.ttwu_count.max 93.22 ± 36% +600.3% 652.75 ± 27% sched_debug.cpu.ttwu_count.min 5368 ± 2% +69.2% 9082 ± 14% sched_debug.cpu.ttwu_count.stddev 4852 ± 1% +36.6% 6626 ± 8% sched_debug.cpu.ttwu_local.avg 55.75 ± 24% +310.6% 228.88 ± 16% sched_debug.cpu.ttwu_local.min *************************************************************************************************** lkp-hsw-ep4: 72 threads Intel(R) Xeon(R) CPU E5-2699 v3 @ 2.30GHz with 128G memory ========================================================================================= compiler/cpufreq_governor/kconfig/nr_ssd/nr_task/rootfs/runtime/tbox_group/test/testcase/thp_defrag/thp_enabled: gcc-6/performance/x86_64-rhel-7.2/1/64/debian-x86_64-2016-08-31.cgz/300/lkp-hsw-ep4/swap-w-seq/vm-scalability/never/always commit: 1d8bf926f8 ("mm/bootmem.c: replace kzalloc() by kzalloc_node()") 371a096edf ("mm: don't use radix tree writeback tags for pages in swap cache") 1d8bf926f8739bd3 371a096edf43a8c71844cf71c2 ---------------- -------------------------- fail:runs %reproduction fail:runs | | | %stddev %change %stddev \ | \ 2125843 ± 0% +14.1% 2426196 ± 1% vm-scalability.throughput 741119 ± 3% +6.0% 785919 ± 2% vm-scalability.time.involuntary_context_switches 1.074e+08 ± 1% +16.5% 1.251e+08 ± 1% vm-scalability.time.minor_page_faults 6243 ± 0% -40.2% 3733 ± 10% vm-scalability.time.percent_of_cpu_this_job_got 21748 ± 0% -41.0% 12831 ± 10% vm-scalability.time.system_time 315.07 ± 1% -10.9% 280.79 ± 3% vm-scalability.time.user_time 333774 ± 2% +12.3% 374973 ± 3% softirqs.SCHED 11885858 ± 0% -39.4% 7205748 ± 9% softirqs.TIMER 40607746 ± 1% +25.4% 50919220 ± 12% numa-numastat.node0.numa_foreign 40597356 ± 1% +25.4% 50910950 ± 12% numa-numastat.node1.numa_miss 0.25 ±173% +1200.0% 3.25 ± 70% numa-numastat.node1.other_node 1211829 ± 1% +16.0% 1405484 ± 1% vmstat.io.bo 4939033 ± 3% +16.1% 5733178 ± 6% vmstat.memory.free 2.1e+08 ± 1% +14.9% 2.412e+08 ± 2% vmstat.memory.swpd 0.00 ± 0% +Inf% 31.50 ± 17% vmstat.procs.b 67.00 ± 0% -42.5% 38.50 ± 10% vmstat.procs.r 1211824 ± 1% +16.0% 1405478 ± 1% vmstat.swap.so 92.51 ± 0% -39.5% 56.01 ± 9% turbostat.%Busy 2583 ± 0% -39.4% 1566 ± 9% turbostat.Avg_MHz 6.96 ± 2% +406.3% 35.26 ± 12% turbostat.CPU%c1 0.01 ± 0% +24025.0% 2.41 ± 19% turbostat.CPU%c3 0.52 ± 21% +1125.7% 6.31 ± 14% turbostat.CPU%c6 258.28 ± 0% -12.0% 227.24 ± 1% turbostat.PkgWatt 55.36 ± 0% +1.1% 55.98 ± 0% turbostat.RAMWatt 28505239 ± 10% -31.5% 19524763 ± 15% meminfo.AnonHugePages 5537897 ± 9% -14.4% 4740114 ± 9% meminfo.DirectMap2M 5822789 ± 2% +13.2% 6594234 ± 5% meminfo.MemAvailable 6222814 ± 2% +12.5% 7001717 ± 4% meminfo.MemFree 709649 ± 0% +12.6% 798807 ± 5% meminfo.PageTables 111455 ± 0% +55.1% 172917 ± 6% meminfo.SUnreclaim 169503 ± 0% +37.7% 233321 ± 4% meminfo.Slab 1504686 ± 3% -24.8% 1131932 ± 15% meminfo.SwapCached 3028 ± 10% +22884.4% 695969 ± 20% meminfo.Writeback 43194655 ± 11% +619.4% 3.107e+08 ± 16% cpuidle.C1-HSW.time 622561 ± 17% +104.2% 1271319 ± 17% cpuidle.C1-HSW.usage 14115248 ± 15% +9297.9% 1.327e+09 ± 16% cpuidle.C1E-HSW.time 84418 ± 8% +2439.5% 2143769 ± 16% cpuidle.C1E-HSW.usage 53083465 ± 15% +2556.0% 1.41e+09 ± 21% cpuidle.C3-HSW.time 154076 ± 13% +1115.8% 1873320 ± 20% cpuidle.C3-HSW.usage 1.803e+09 ± 3% +350.2% 8.117e+09 ± 10% cpuidle.C6-HSW.time 1936956 ± 3% +347.5% 8667200 ± 11% cpuidle.C6-HSW.usage 14995046 ± 10% +2237.0% 3.504e+08 ± 14% cpuidle.POLL.time 13351 ± 17% +202.3% 40357 ± 15% cpuidle.POLL.usage 22648511 ± 14% -36.8% 14307991 ± 19% numa-meminfo.node0.AnonHugePages 1045808 ± 5% -47.9% 544659 ± 15% numa-meminfo.node0.FilePages 58493 ± 4% +57.0% 91837 ± 11% numa-meminfo.node0.SUnreclaim 85451 ± 4% +40.6% 120107 ± 9% numa-meminfo.node0.Slab 1041 ± 12% +15041.9% 157627 ± 13% numa-meminfo.node0.Writeback 5767065 ± 4% -32.4% 3896120 ± 36% numa-meminfo.node1.AnonHugePages 293938 ± 5% +50.9% 443468 ± 27% numa-meminfo.node1.PageTables 52953 ± 4% +53.2% 81112 ± 14% numa-meminfo.node1.SUnreclaim 17882 ± 4% -55.2% 8018 ± 59% numa-meminfo.node1.Shmem 84052 ± 5% +34.8% 113281 ± 11% numa-meminfo.node1.Slab 2236 ± 8% +23862.8% 535987 ± 24% numa-meminfo.node1.Writeback 11242 ± 13% -38.0% 6974 ± 21% numa-vmstat.node0.nr_anon_transparent_hugepages 262464 ± 4% -48.5% 135174 ± 17% numa-vmstat.node0.nr_file_pages 1215804 ± 3% +13.8% 1383278 ± 4% numa-vmstat.node0.nr_free_pages 1287 ± 17% +102.0% 2599 ± 24% numa-vmstat.node0.nr_isolated_anon 14625 ± 4% +54.0% 22521 ± 12% numa-vmstat.node0.nr_slab_unreclaimable 10793807 ± 9% +36.2% 14705195 ± 5% numa-vmstat.node0.nr_vmscan_write 233.00 ± 7% +15561.6% 36491 ± 15% numa-vmstat.node0.nr_writeback 10793782 ± 9% +35.9% 14668927 ± 5% numa-vmstat.node0.nr_written 233.00 ± 8% +15561.5% 36491 ± 15% numa-vmstat.node0.nr_zone_write_pending 22566322 ± 3% +31.3% 29625222 ± 12% numa-vmstat.node0.numa_foreign 2872 ± 4% -27.1% 2093 ± 23% numa-vmstat.node1.nr_anon_transparent_hugepages 404362 ± 9% +34.0% 541983 ± 19% numa-vmstat.node1.nr_free_pages 156.00 ± 11% +95.4% 304.75 ± 15% numa-vmstat.node1.nr_pages_scanned 13235 ± 4% +50.2% 19874 ± 14% numa-vmstat.node1.nr_slab_unreclaimable 49196401 ± 1% +24.1% 61044573 ± 7% numa-vmstat.node1.nr_vmscan_write 572.50 ± 14% +21710.0% 124862 ± 28% numa-vmstat.node1.nr_writeback 49196179 ± 1% +23.8% 60919929 ± 7% numa-vmstat.node1.nr_written 574.25 ± 14% +21643.7% 124863 ± 28% numa-vmstat.node1.nr_zone_write_pending 22550830 ± 3% +31.3% 29618804 ± 12% numa-vmstat.node1.numa_miss 4.086e+12 ± 0% -28.4% 2.928e+12 ± 5% perf-stat.branch-instructions 0.10 ± 0% +52.0% 0.15 ± 6% perf-stat.branch-miss-rate% 4.119e+09 ± 0% +8.5% 4.471e+09 ± 1% perf-stat.branch-misses 30.75 ± 0% -13.1% 26.72 ± 3% perf-stat.cache-miss-rate% 1.046e+10 ± 1% +9.9% 1.149e+10 ± 1% perf-stat.cache-misses 3.4e+10 ± 0% +26.5% 4.303e+10 ± 2% perf-stat.cache-references 6.545e+13 ± 0% -37.9% 4.065e+13 ± 7% perf-stat.cpu-cycles 195546 ± 4% +32.7% 259393 ± 6% perf-stat.cpu-migrations 0.04 ± 12% +84.7% 0.08 ± 10% perf-stat.dTLB-load-miss-rate% 4.077e+12 ± 0% -33.5% 2.713e+12 ± 8% perf-stat.dTLB-loads 0.18 ± 4% +16.4% 0.21 ± 2% perf-stat.dTLB-store-miss-rate% 6.92e+08 ± 3% +17.8% 8.149e+08 ± 9% perf-stat.dTLB-store-misses 37.76 ± 2% -41.7% 22.01 ± 5% perf-stat.iTLB-load-miss-rate% 42983189 ± 3% +36.3% 58591746 ± 8% perf-stat.iTLB-load-misses 70857240 ± 2% +194.4% 2.086e+08 ± 11% perf-stat.iTLB-loads 1.656e+13 ± 0% -27.9% 1.194e+13 ± 5% perf-stat.instructions 385661 ± 3% -46.6% 205896 ± 12% perf-stat.instructions-per-iTLB-miss 0.25 ± 0% +16.3% 0.29 ± 2% perf-stat.ipc 1.079e+08 ± 1% +16.4% 1.256e+08 ± 1% perf-stat.minor-faults 85.52 ± 0% -4.7% 81.49 ± 0% perf-stat.node-load-miss-rate% 9.377e+08 ± 2% +28.9% 1.208e+09 ± 4% perf-stat.node-loads 55.12 ± 0% +2.7% 56.60 ± 1% perf-stat.node-store-miss-rate% 2.191e+09 ± 0% +4.7% 2.294e+09 ± 1% perf-stat.node-store-misses 1.079e+08 ± 1% +16.4% 1.256e+08 ± 1% perf-stat.page-faults 174.75 ± 11% +206.4% 535.50 ± 38% slabinfo.bdev_cache.active_objs 174.75 ± 11% +206.4% 535.50 ± 38% slabinfo.bdev_cache.num_objs 6222 ± 1% +18.0% 7341 ± 5% slabinfo.cred_jar.active_objs 6222 ± 1% +18.0% 7341 ± 5% slabinfo.cred_jar.num_objs 453.00 ± 11% +65.0% 747.25 ± 16% slabinfo.file_lock_cache.active_objs 453.00 ± 11% +65.0% 747.25 ± 16% slabinfo.file_lock_cache.num_objs 4046 ± 1% +101.9% 8169 ± 3% slabinfo.kmalloc-1024.active_objs 126.50 ± 0% +106.5% 261.25 ± 3% slabinfo.kmalloc-1024.active_slabs 4071 ± 0% +105.6% 8372 ± 3% slabinfo.kmalloc-1024.num_objs 126.50 ± 0% +106.5% 261.25 ± 3% slabinfo.kmalloc-1024.num_slabs 7564 ± 0% +17.1% 8853 ± 1% slabinfo.kmalloc-192.active_objs 7615 ± 0% +16.3% 8860 ± 1% slabinfo.kmalloc-192.num_objs 4766 ± 1% +69.8% 8094 ± 8% slabinfo.kmalloc-2048.active_objs 307.00 ± 2% +67.9% 515.50 ± 8% slabinfo.kmalloc-2048.active_slabs 4826 ± 1% +70.5% 8227 ± 8% slabinfo.kmalloc-2048.num_objs 307.00 ± 2% +67.9% 515.50 ± 8% slabinfo.kmalloc-2048.num_slabs 14592 ± 1% +934.5% 150965 ± 34% slabinfo.kmalloc-256.active_objs 342.25 ± 1% +1924.0% 6927 ± 34% slabinfo.kmalloc-256.active_slabs 14906 ± 1% +934.8% 154254 ± 34% slabinfo.kmalloc-256.num_objs 342.25 ± 1% +1924.0% 6927 ± 34% slabinfo.kmalloc-256.num_slabs 12576 ± 3% +51.5% 19055 ± 7% slabinfo.kmalloc-512.active_objs 218.50 ± 15% +56.3% 341.50 ± 10% slabinfo.kmalloc-512.active_slabs 12727 ± 3% +54.0% 19595 ± 9% slabinfo.kmalloc-512.num_objs 218.50 ± 15% +56.3% 341.50 ± 10% slabinfo.kmalloc-512.num_slabs 765.50 ± 4% +69.1% 1294 ± 21% slabinfo.nsproxy.active_objs 765.50 ± 4% +69.1% 1294 ± 21% slabinfo.nsproxy.num_objs 1612 ± 13% -69.5% 492.25 ± 96% proc-vmstat.compact_fail 1619 ± 12% -69.1% 501.25 ± 95% proc-vmstat.compact_stall 13831 ± 9% -31.2% 9522 ± 16% proc-vmstat.nr_anon_transparent_hugepages 140989 ± 6% +25.8% 177310 ± 4% proc-vmstat.nr_dirty_background_threshold 282323 ± 6% +25.8% 355056 ± 4% proc-vmstat.nr_dirty_threshold 471239 ± 3% -20.9% 372789 ± 11% proc-vmstat.nr_file_pages 1496965 ± 6% +24.3% 1860851 ± 3% proc-vmstat.nr_free_pages 5139 ± 4% +21.5% 6246 ± 2% proc-vmstat.nr_isolated_anon 178332 ± 1% +13.1% 201722 ± 5% proc-vmstat.nr_page_table_pages 147.75 ± 8% +100.8% 296.75 ± 18% proc-vmstat.nr_pages_scanned 27862 ± 0% +52.7% 42558 ± 6% proc-vmstat.nr_slab_unreclaimable 60269851 ± 1% +25.1% 75424523 ± 5% proc-vmstat.nr_vmscan_write 774.25 ± 2% +21072.7% 163929 ± 22% proc-vmstat.nr_writeback 1.073e+08 ± 1% +15.5% 1.239e+08 ± 2% proc-vmstat.nr_written 775.50 ± 2% +21038.5% 163928 ± 22% proc-vmstat.nr_zone_write_pending 46256859 ± 1% +26.7% 58617918 ± 8% proc-vmstat.numa_foreign 13246 ± 6% -46.1% 7141 ± 17% proc-vmstat.numa_hint_faults 8187 ± 2% -42.4% 4713 ± 20% proc-vmstat.numa_hint_faults_local 46256859 ± 1% +26.7% 58617918 ± 8% proc-vmstat.numa_miss 1.394e+08 ± 0% +13.0% 1.575e+08 ± 1% proc-vmstat.pgalloc_normal 1.084e+08 ± 1% +15.4% 1.251e+08 ± 2% proc-vmstat.pgdeactivate 1.079e+08 ± 1% +16.4% 1.257e+08 ± 1% proc-vmstat.pgfault 1.377e+08 ± 1% +13.3% 1.559e+08 ± 1% proc-vmstat.pgfree 3711 ± 22% -67.7% 1198 ± 32% proc-vmstat.pgmigrate_fail 4.293e+08 ± 1% +15.5% 4.958e+08 ± 2% proc-vmstat.pgpgout 76556208 ± 1% +21.6% 93113321 ± 2% proc-vmstat.pgrefill 61504865 ± 2% +85.7% 1.142e+08 ± 5% proc-vmstat.pgrotated 1.754e+08 ± 1% +26.7% 2.223e+08 ± 4% proc-vmstat.pgscan_direct 8178772 ± 3% +61.6% 13217350 ± 18% proc-vmstat.pgscan_kswapd 1.007e+08 ± 1% +12.8% 1.136e+08 ± 2% proc-vmstat.pgsteal_direct 6418690 ± 7% +57.7% 10125267 ± 23% proc-vmstat.pgsteal_kswapd 1.073e+08 ± 1% +15.5% 1.239e+08 ± 2% proc-vmstat.pswpout 209541 ± 1% +16.5% 244206 ± 1% proc-vmstat.thp_fault_fallback 142799 ± 0% -43.9% 80147 ± 5% sched_debug.cfs_rq:/.exec_clock.avg 131668 ± 1% -69.0% 40852 ± 14% sched_debug.cfs_rq:/.exec_clock.min 5099 ± 13% +520.8% 31657 ± 20% sched_debug.cfs_rq:/.exec_clock.stddev 227.88 ± 4% +35.0% 307.70 ± 15% sched_debug.cfs_rq:/.load_avg.min 246.74 ± 26% -53.7% 114.30 ± 23% sched_debug.cfs_rq:/.load_avg.stddev 9367426 ± 0% -58.8% 3856380 ± 10% sched_debug.cfs_rq:/.min_vruntime.avg 10027888 ± 0% -46.2% 5391954 ± 6% sched_debug.cfs_rq:/.min_vruntime.max 8536208 ± 2% -70.5% 2518516 ± 18% sched_debug.cfs_rq:/.min_vruntime.min 308749 ± 8% +188.8% 891811 ± 21% sched_debug.cfs_rq:/.min_vruntime.stddev 0.82 ± 2% -31.3% 0.56 ± 15% sched_debug.cfs_rq:/.nr_running.avg 0.23 ± 13% +43.2% 0.33 ± 8% sched_debug.cfs_rq:/.nr_running.stddev 36.70 ± 13% -67.1% 12.06 ± 20% sched_debug.cfs_rq:/.nr_spread_over.avg 123.42 ± 19% -68.2% 39.20 ± 35% sched_debug.cfs_rq:/.nr_spread_over.max 8.92 ± 14% -64.9% 3.13 ± 52% sched_debug.cfs_rq:/.nr_spread_over.min 23.11 ± 20% -74.8% 5.83 ± 27% sched_debug.cfs_rq:/.nr_spread_over.stddev 32.16 ± 6% -37.6% 20.08 ± 33% sched_debug.cfs_rq:/.runnable_load_avg.avg 653879 ± 47% +135.8% 1541932 ± 15% sched_debug.cfs_rq:/.spread0.max 315599 ± 9% +184.2% 896957 ± 21% sched_debug.cfs_rq:/.spread0.stddev 898.96 ± 1% -27.4% 652.99 ± 15% sched_debug.cfs_rq:/.util_avg.avg 673946 ± 3% +24.3% 837742 ± 4% sched_debug.cpu.avg_idle.avg 49038 ± 8% +259.6% 176360 ± 60% sched_debug.cpu.avg_idle.min 301898 ± 6% -25.5% 225046 ± 21% sched_debug.cpu.avg_idle.stddev 190551 ± 0% +9.0% 207609 ± 2% sched_debug.cpu.clock.max 233.98 ± 3% +2164.8% 5299 ± 53% sched_debug.cpu.clock.stddev 190551 ± 0% +9.0% 207609 ± 2% sched_debug.cpu.clock_task.max 233.98 ± 3% +2164.8% 5299 ± 53% sched_debug.cpu.clock_task.stddev 32.11 ± 6% -37.8% 19.98 ± 33% sched_debug.cpu.cpu_load[0].avg 35.42 ± 9% -37.8% 22.02 ± 26% sched_debug.cpu.cpu_load[1].avg 133.36 ± 15% -39.1% 81.19 ± 37% sched_debug.cpu.cpu_load[1].stddev 34.65 ± 7% -38.1% 21.45 ± 27% sched_debug.cpu.cpu_load[2].avg 126.41 ± 9% -38.0% 78.32 ± 38% sched_debug.cpu.cpu_load[2].stddev 34.35 ± 6% -38.4% 21.15 ± 28% sched_debug.cpu.cpu_load[3].avg 123.58 ± 6% -38.4% 76.07 ± 38% sched_debug.cpu.cpu_load[3].stddev 34.45 ± 5% -38.7% 21.11 ± 28% sched_debug.cpu.cpu_load[4].avg 123.29 ± 4% -38.7% 75.54 ± 38% sched_debug.cpu.cpu_load[4].stddev 1367 ± 2% -31.7% 934.09 ± 15% sched_debug.cpu.curr->pid.avg 0.00 ± 2% +2051.8% 0.01 ± 53% sched_debug.cpu.next_balance.stddev 148590 ± 1% -25.6% 110559 ± 3% sched_debug.cpu.nr_load_updates.avg 139700 ± 1% -43.5% 78865 ± 5% sched_debug.cpu.nr_load_updates.min 3733 ± 13% +519.4% 23123 ± 21% sched_debug.cpu.nr_load_updates.stddev 0.84 ± 2% -31.9% 0.57 ± 17% sched_debug.cpu.nr_running.avg 0.28 ± 10% +28.5% 0.36 ± 10% sched_debug.cpu.nr_running.stddev 14395 ± 9% +34.0% 19289 ± 23% sched_debug.cpu.nr_switches.min 0.11 ± 43% +470.4% 0.62 ± 52% sched_debug.cpu.nr_uninterruptible.avg 7400 ± 21% +31.3% 9718 ± 15% sched_debug.cpu.sched_goidle.avg 1050 ± 7% +245.9% 3632 ± 38% sched_debug.cpu.sched_goidle.min 40897 ± 23% +43.3% 58610 ± 7% sched_debug.cpu.ttwu_count.max 10300 ± 11% -51.3% 5018 ± 29% sched_debug.cpu.ttwu_count.min 5838 ± 26% +99.6% 11652 ± 12% sched_debug.cpu.ttwu_count.stddev 7617 ± 7% +18.5% 9024 ± 5% sched_debug.cpu.ttwu_local.avg 4191 ± 25% +36.8% 5733 ± 9% sched_debug.cpu.ttwu_local.stddev Thanks, Xiaolong View attachment "config-4.8.0-10066-g371a096" of type "text/plain" (153641 bytes) View attachment "job-script" of type "text/plain" (7082 bytes) View attachment "job.yaml" of type "text/plain" (4740 bytes) View attachment "reproduce" of type "text/plain" (575 bytes)
Powered by blists - more mailing lists