[<prev] [next>] [<thread-prev] [day] [month] [year] [list]
Message-Id: <20190303183419.GA4365@linux.ibm.com>
Date: Sun, 3 Mar 2019 10:34:19 -0800
From: "Paul E. McKenney" <paulmck@...ux.ibm.com>
To: kernel test robot <rong.a.chen@...el.com>
Cc: LKML <linux-kernel@...r.kernel.org>,
Linus Torvalds <torvalds@...ux-foundation.org>, lkp@...org
Subject: Re: [LKP] [rcu] 7e28c5af4e: will-it-scale.per_process_ops 84.7%
improvement
On Thu, Feb 28, 2019 at 12:50:15PM +0800, kernel test robot wrote:
> Greeting,
>
> FYI, we noticed a 84.7% improvement of will-it-scale.per_process_ops due to commit:
>
>
> commit: 7e28c5af4ef6b539334aa5de40feca0c041c94df ("rcu: Eliminate ->rcu_qs_ctr from the rcu_dynticks structure")
> https://git.kernel.org/cgit/linux/kernel/git/torvalds/linux.git master
>
> in testcase: will-it-scale
> on test machine: 288 threads Intel(R) Xeon Phi(TM) CPU 7295 @ 1.50GHz with 80G memory
> with following parameters:
>
> nr_task: 100%
> mode: process
> test: brk1
> cpufreq_governor: performance
>
> test-description: Will It Scale takes a testcase and runs it from 1 through to n parallel copies to see if the testcase will scale. It builds both a process and threads based test in order to see any differences between the two.
> test-url: https://github.com/antonblanchard/will-it-scale
It looks like getting rid of the rcu_dynticks structure improved cache
locality. Or am I missing something in the wealth of statistic herein?
Thanx, Paul
> Details are as below:
> -------------------------------------------------------------------------------------------------->
>
>
> To reproduce:
>
> git clone https://github.com/intel/lkp-tests.git
> cd lkp-tests
> bin/lkp install job.yaml # job file is attached in this email
> bin/lkp run job.yaml
>
> =========================================================================================
> compiler/cpufreq_governor/kconfig/mode/nr_task/rootfs/tbox_group/test/testcase:
> gcc-7/performance/x86_64-rhel-7.2/process/100%/debian-x86_64-2018-04-03.cgz/lkp-knm01/brk1/will-it-scale
>
> commit:
> c5bacd9417 ("rcu: Motivate Tiny RCU forward progress")
> 7e28c5af4e ("rcu: Eliminate ->rcu_qs_ctr from the rcu_dynticks structure")
>
> c5bacd94173ec49d 7e28c5af4ef6b539334aa5de40
> ---------------- --------------------------
> %stddev %change %stddev
> \ | \
> 48605 +84.7% 89756 ± 4% will-it-scale.per_process_ops
> 13998509 +84.7% 25850040 ± 4% will-it-scale.workload
> 0.01 ± 63% -0.0 0.00 ±131% mpstat.cpu.soft%
> 11.86 ± 3% -1.2 10.65 ± 4% mpstat.cpu.usr%
> 662910 +15.8% 767829 ± 3% numa-numastat.node0.local_node
> 662889 +15.8% 767801 ± 3% numa-numastat.node0.numa_hit
> 2149 ± 4% -16.9% 1785 ± 2% vmstat.system.cs
> 727237 ± 3% -14.9% 618668 vmstat.system.in
> 284156 ± 4% +32.5% 376625 numa-meminfo.node0.Active
> 283908 ± 4% +32.6% 376374 numa-meminfo.node0.Active(anon)
> 110629 ± 15% -77.7% 24625 ± 18% numa-meminfo.node0.Inactive
> 110393 ± 15% -77.9% 24400 ± 19% numa-meminfo.node0.Inactive(anon)
> 284255 ± 4% +32.5% 376666 meminfo.Active
> 284007 ± 4% +32.5% 376415 meminfo.Active(anon)
> 110855 ± 15% -77.8% 24620 ± 18% meminfo.Inactive
> 110619 ± 15% -77.9% 24395 ± 19% meminfo.Inactive(anon)
> 8.00 ±173% +11303.1% 912.25 ± 87% meminfo.Mlocked
> 11072 -15.7% 9338 ± 3% meminfo.max_used_kB
> 71004 ± 4% +32.6% 94124 numa-vmstat.node0.nr_active_anon
> 27812 ± 15% -78.1% 6100 ± 19% numa-vmstat.node0.nr_inactive_anon
> 1.00 ±173% +13100.0% 132.00 ± 87% numa-vmstat.node0.nr_mlock
> 71002 ± 4% +32.6% 94122 numa-vmstat.node0.nr_zone_active_anon
> 27811 ± 15% -78.1% 6100 ± 19% numa-vmstat.node0.nr_zone_inactive_anon
> 1.00 ±173% +9450.0% 95.50 ± 87% numa-vmstat.node1.nr_mlock
> 70939 ± 4% +32.8% 94186 proc-vmstat.nr_active_anon
> 27523 ± 15% -77.8% 6101 ± 19% proc-vmstat.nr_inactive_anon
> 2.25 ±173% +10044.4% 228.25 ± 86% proc-vmstat.nr_mlock
> 70939 ± 4% +32.8% 94186 proc-vmstat.nr_zone_active_anon
> 27523 ± 15% -77.8% 6101 ± 19% proc-vmstat.nr_zone_inactive_anon
> 680060 +15.8% 787707 ± 2% proc-vmstat.numa_hit
> 680060 +15.8% 787707 ± 2% proc-vmstat.numa_local
> 140509 ± 10% -45.3% 76820 ± 24% proc-vmstat.numa_pte_updates
> 895.50 ± 30% +480.0% 5193 ± 36% proc-vmstat.pgactivate
> 774283 +13.9% 882060 ± 3% proc-vmstat.pgalloc_normal
> 653457 +21.9% 796258 ± 3% proc-vmstat.pgfault
> 610954 ± 2% +32.4% 808662 ± 3% proc-vmstat.pgfree
> 5299 ± 6% +106.4% 10939 ± 36% softirqs.CPU10.RCU
> 106293 ± 5% +28.2% 136274 ± 15% softirqs.CPU126.TIMER
> 108234 ± 6% +26.3% 136696 ± 15% softirqs.CPU127.TIMER
> 109747 ± 7% +26.4% 138696 ± 11% softirqs.CPU135.TIMER
> 111732 ± 8% +24.4% 138967 ± 11% softirqs.CPU138.TIMER
> 107031 ± 8% +26.7% 135558 ± 14% softirqs.CPU148.TIMER
> 100841 +23.1% 124155 ± 15% softirqs.CPU15.TIMER
> 104640 ± 4% +27.8% 133719 ± 16% softirqs.CPU172.TIMER
> 103142 ± 3% +30.9% 135023 ± 13% softirqs.CPU190.TIMER
> 106698 ± 10% +27.5% 136023 ± 13% softirqs.CPU197.TIMER
> 110268 ± 2% +22.5% 135052 ± 12% softirqs.CPU205.TIMER
> 106111 ± 8% +23.9% 131436 ± 14% softirqs.CPU206.TIMER
> 106894 ± 6% +28.7% 137574 ± 14% softirqs.CPU213.TIMER
> 108418 ± 8% +28.3% 139100 ± 12% softirqs.CPU214.TIMER
> 100170 ± 10% +32.8% 132978 ± 15% softirqs.CPU215.TIMER
> 100836 ± 2% +30.3% 131406 ± 17% softirqs.CPU217.TIMER
> 112497 ± 8% +22.9% 138291 ± 13% softirqs.CPU219.TIMER
> 108669 ± 10% +22.1% 132720 ± 15% softirqs.CPU220.TIMER
> 108584 ± 8% +25.5% 136230 ± 13% softirqs.CPU227.TIMER
> 106251 ± 6% +28.6% 136686 ± 11% softirqs.CPU236.TIMER
> 105296 ± 7% +31.3% 138251 ± 12% softirqs.CPU238.TIMER
> 3915 ± 8% +77.4% 6945 ± 30% softirqs.CPU249.RCU
> 101465 ± 2% +26.5% 128388 ± 19% softirqs.CPU271.TIMER
> 104242 ± 8% +23.7% 128907 ± 17% softirqs.CPU70.TIMER
> 107020 ± 7% +26.9% 135845 ± 13% softirqs.CPU92.TIMER
> 9.659e+09 ± 10% +16.3% 1.124e+10 ± 3% perf-stat.i.branch-instructions
> 6.69 +1.0 7.69 ± 2% perf-stat.i.branch-miss-rate%
> 6.443e+08 ± 10% +33.3% 8.59e+08 ± 2% perf-stat.i.branch-misses
> 4.10 ± 6% +0.9 5.04 ± 10% perf-stat.i.cache-miss-rate%
> 67464898 ± 10% +25.2% 84491315 ± 4% perf-stat.i.cache-misses
> 2820 ± 13% -32.4% 1906 ± 3% perf-stat.i.context-switches
> 15.34 -37.0% 9.67 ± 3% perf-stat.i.cpi
> 7.067e+11 ± 11% -30.2% 4.932e+11 ± 4% perf-stat.i.cpu-cycles
> 71.93 ± 10% -41.9% 41.79 ± 9% perf-stat.i.cpu-migrations
> 2.36 -0.0 2.33 perf-stat.i.iTLB-load-miss-rate%
> 1.106e+09 ± 10% +15.1% 1.273e+09 ± 2% perf-stat.i.iTLB-load-misses
> 4.569e+10 ± 10% +17.0% 5.345e+10 ± 3% perf-stat.i.iTLB-loads
> 4.621e+10 ± 10% +16.1% 5.364e+10 ± 3% perf-stat.i.instructions
> 0.07 +64.1% 0.11 ± 3% perf-stat.i.ipc
> 3524 ± 11% -18.3% 2878 perf-stat.i.minor-faults
> 993359 ± 11% -32.0% 675215 ± 2% perf-stat.i.msec
> 3529 ± 11% -18.5% 2878 perf-stat.i.page-faults
> 6.67 +1.0 7.65 perf-stat.overall.branch-miss-rate%
> 3.91 +0.6 4.54 ± 3% perf-stat.overall.cache-miss-rate%
> 15.28 -39.8% 9.20 ± 3% perf-stat.overall.cpi
> 2.36 -0.0 2.33 perf-stat.overall.iTLB-load-miss-rate%
> 41.78 +0.9% 42.13 perf-stat.overall.instructions-per-iTLB-miss
> 0.07 +66.4% 0.11 ± 3% perf-stat.overall.ipc
> 580565 -5.5% 548645 perf-stat.overall.path-length
> 1.699e+12 +74.9% 2.971e+12 ± 4% perf-stat.total.branch-instructions
> 1.133e+11 +100.6% 2.272e+11 ± 5% perf-stat.total.branch-misses
> 1.186e+10 +88.0% 2.23e+10 perf-stat.total.cache-misses
> 3.032e+11 +62.2% 4.918e+11 ± 2% perf-stat.total.cache-references
> 1.242e+14 +4.8% 1.302e+14 perf-stat.total.cpu-cycles
> 12654 -13.0% 11014 ± 5% perf-stat.total.cpu-migrations
> 1.945e+11 +73.0% 3.365e+11 ± 3% perf-stat.total.iTLB-load-misses
> 8.038e+12 +75.8% 1.413e+13 ± 4% perf-stat.total.iTLB-loads
> 8.127e+12 +74.5% 1.418e+13 ± 4% perf-stat.total.instructions
> 619183 +22.9% 760791 ± 3% perf-stat.total.minor-faults
> 620031 +22.7% 760750 ± 3% perf-stat.total.page-faults
> 23472574 ± 28% -50.0% 11736977 ± 60% sched_debug.cfs_rq:/.MIN_vruntime.max
> 1380731 ± 28% -46.9% 733243 ± 66% sched_debug.cfs_rq:/.MIN_vruntime.stddev
> 109109 +17.1% 127734 sched_debug.cfs_rq:/.exec_clock.avg
> 115740 +12.4% 130080 sched_debug.cfs_rq:/.exec_clock.max
> 4252 ± 4% -50.4% 2110 ± 36% sched_debug.cfs_rq:/.exec_clock.stddev
> 6243 ± 12% -14.0% 5369 ± 15% sched_debug.cfs_rq:/.load.avg
> 887713 ± 24% -30.6% 616115 ± 43% sched_debug.cfs_rq:/.load.max
> 53788 ± 23% -30.4% 37427 ± 40% sched_debug.cfs_rq:/.load.stddev
> 23472574 ± 28% -50.0% 11736977 ± 60% sched_debug.cfs_rq:/.max_vruntime.max
> 1380731 ± 28% -46.9% 733243 ± 66% sched_debug.cfs_rq:/.max_vruntime.stddev
> 36862110 +17.8% 43411425 sched_debug.cfs_rq:/.min_vruntime.avg
> 39448691 +12.5% 44387703 sched_debug.cfs_rq:/.min_vruntime.max
> 1668979 ± 3% -45.0% 917652 ± 30% sched_debug.cfs_rq:/.min_vruntime.stddev
> 1.38 ± 9% -9.1% 1.25 ± 6% sched_debug.cfs_rq:/.nr_running.max
> 0.56 ± 19% +42.2% 0.80 sched_debug.cfs_rq:/.nr_running.min
> 0.07 ± 13% -32.7% 0.05 ± 15% sched_debug.cfs_rq:/.nr_running.stddev
> 2.25 ± 3% -51.5% 1.09 ± 2% sched_debug.cfs_rq:/.nr_spread_over.avg
> 6235 ± 12% -14.0% 5361 ± 16% sched_debug.cfs_rq:/.runnable_weight.avg
> 887713 ± 24% -30.6% 616088 ± 43% sched_debug.cfs_rq:/.runnable_weight.max
> 53791 ± 23% -30.4% 37423 ± 40% sched_debug.cfs_rq:/.runnable_weight.stddev
> 513480 ± 46% -68.1% 163689 ± 16% sched_debug.cfs_rq:/.spread0.avg
> 992070 ± 25% -55.6% 440313 ± 8% sched_debug.cfs_rq:/.spread0.max
> 271.31 ± 25% +54.1% 418.15 ± 23% sched_debug.cfs_rq:/.util_avg.min
> 71.69 ± 6% -14.9% 61.03 ± 7% sched_debug.cfs_rq:/.util_avg.stddev
> 638.86 +18.5% 756.85 ± 3% sched_debug.cfs_rq:/.util_est_enqueued.avg
> 131.07 ± 2% -17.0% 108.78 ± 5% sched_debug.cfs_rq:/.util_est_enqueued.stddev
> 1617727 ± 4% -24.5% 1222141 ± 11% sched_debug.cpu.avg_idle.avg
> 167800 ± 2% +13.3% 190124 ± 4% sched_debug.cpu.clock.avg
> 174380 +10.3% 192313 ± 4% sched_debug.cpu.clock.max
> 161230 ± 2% +16.5% 187888 ± 4% sched_debug.cpu.clock.min
> 3797 ± 4% -66.4% 1276 ± 8% sched_debug.cpu.clock.stddev
> 167800 ± 2% +13.3% 190124 ± 4% sched_debug.cpu.clock_task.avg
> 174380 +10.3% 192313 ± 4% sched_debug.cpu.clock_task.max
> 161230 ± 2% +16.5% 187888 ± 4% sched_debug.cpu.clock_task.min
> 3797 ± 4% -66.4% 1276 ± 8% sched_debug.cpu.clock_task.stddev
> 5416 ± 2% +15.6% 6262 ± 5% sched_debug.cpu.curr->pid.max
> 2045 ± 18% +20.6% 2466 ± 3% sched_debug.cpu.curr->pid.min
> 892728 ± 7% -37.5% 557989 ± 5% sched_debug.cpu.max_idle_balance_cost.avg
> 4113479 ± 18% -51.6% 1989769 ± 36% sched_debug.cpu.max_idle_balance_cost.max
> 461057 ± 17% -67.6% 149491 ± 36% sched_debug.cpu.max_idle_balance_cost.stddev
> 0.00 ± 4% -65.8% 0.00 ± 23% sched_debug.cpu.next_balance.stddev
> 117031 +16.4% 136279 sched_debug.cpu.nr_load_updates.avg
> 139472 ± 3% +16.0% 161763 ± 4% sched_debug.cpu.nr_load_updates.max
> 109218 +22.1% 133390 sched_debug.cpu.nr_load_updates.min
> 3910 ± 3% -46.0% 2111 ± 19% sched_debug.cpu.nr_load_updates.stddev
> 0.16 ± 4% -19.4% 0.13 ± 3% sched_debug.cpu.nr_running.stddev
> 1108 +12.5% 1247 ± 2% sched_debug.cpu.nr_switches.avg
> 26767 ± 16% +60.1% 42860 ± 15% sched_debug.cpu.nr_switches.max
> 2342 ± 4% +34.2% 3144 ± 7% sched_debug.cpu.nr_switches.stddev
> 29066 ± 11% +83.5% 53341 ± 36% sched_debug.cpu.sched_count.max
> 2473 ± 2% +61.3% 3988 ± 33% sched_debug.cpu.sched_count.stddev
> 275.25 ± 3% +33.5% 367.54 sched_debug.cpu.ttwu_count.avg
> 10910 ± 17% +67.3% 18254 ± 13% sched_debug.cpu.ttwu_count.max
> 50.31 ± 2% +19.9% 60.30 ± 2% sched_debug.cpu.ttwu_count.min
> 832.06 ± 10% +46.1% 1215 ± 7% sched_debug.cpu.ttwu_count.stddev
> 176.33 +42.8% 251.81 ± 2% sched_debug.cpu.ttwu_local.avg
> 10535 ± 18% +69.7% 17877 ± 14% sched_debug.cpu.ttwu_local.max
> 31.56 +15.2% 36.35 sched_debug.cpu.ttwu_local.min
> 729.04 ± 13% +57.2% 1146 ± 9% sched_debug.cpu.ttwu_local.stddev
> 161164 ± 2% +16.6% 187864 ± 4% sched_debug.cpu_clk
> 160513 ± 2% +16.6% 187213 ± 4% sched_debug.ktime
> 1.31 ± 24% +39.1% 1.82 sched_debug.rt_rq:/.rt_runtime.stddev
> 161891 ± 2% +16.3% 188213 ± 4% sched_debug.sched_clk
> 745.75 ± 61% -87.9% 90.25 ±173% interrupts.36:IR-PCI-MSI.2621442-edge.eth1-TxRx-1
> 961.75 ± 48% -75.5% 235.25 ±173% interrupts.37:IR-PCI-MSI.2621443-edge.eth1-TxRx-2
> 4649 ± 16% -25.2% 3475 ± 3% interrupts.CPU100.NMI:Non-maskable_interrupts
> 4649 ± 16% -25.2% 3475 ± 3% interrupts.CPU100.PMI:Performance_monitoring_interrupts
> 30.50 ± 17% +277.9% 115.25 ± 49% interrupts.CPU109.RES:Rescheduling_interrupts
> 552027 ± 5% +12.2% 619466 ± 2% interrupts.CPU110.LOC:Local_timer_interrupts
> 274.50 ± 59% -44.9% 151.25 ± 92% interrupts.CPU116.RES:Rescheduling_interrupts
> 37.50 ± 47% +323.3% 158.75 ± 43% interrupts.CPU121.RES:Rescheduling_interrupts
> 7275 ± 12% -45.9% 3937 ± 25% interrupts.CPU125.NMI:Non-maskable_interrupts
> 7275 ± 12% -45.9% 3937 ± 25% interrupts.CPU125.PMI:Performance_monitoring_interrupts
> 5151 ± 34% -32.6% 3470 ± 6% interrupts.CPU140.NMI:Non-maskable_interrupts
> 5151 ± 34% -32.6% 3470 ± 6% interrupts.CPU140.PMI:Performance_monitoring_interrupts
> 5589 ± 28% -26.0% 4134 ± 32% interrupts.CPU153.NMI:Non-maskable_interrupts
> 5589 ± 28% -26.0% 4134 ± 32% interrupts.CPU153.PMI:Performance_monitoring_interrupts
> 745.75 ± 61% -87.9% 90.25 ±173% interrupts.CPU16.36:IR-PCI-MSI.2621442-edge.eth1-TxRx-1
> 961.75 ± 48% -75.5% 235.25 ±173% interrupts.CPU17.37:IR-PCI-MSI.2621443-edge.eth1-TxRx-2
> 6606 ± 25% -38.6% 4057 ± 27% interrupts.CPU171.NMI:Non-maskable_interrupts
> 6606 ± 25% -38.6% 4057 ± 27% interrupts.CPU171.PMI:Performance_monitoring_interrupts
> 5916 ± 21% -28.9% 4205 ± 28% interrupts.CPU172.NMI:Non-maskable_interrupts
> 5916 ± 21% -28.9% 4205 ± 28% interrupts.CPU172.PMI:Performance_monitoring_interrupts
> 85.50 ± 36% -79.2% 17.75 ± 72% interrupts.CPU176.RES:Rescheduling_interrupts
> 74.00 ± 79% -79.1% 15.50 ± 27% interrupts.CPU177.RES:Rescheduling_interrupts
> 6513 ± 21% -36.4% 4139 ± 24% interrupts.CPU180.NMI:Non-maskable_interrupts
> 6513 ± 21% -36.4% 4139 ± 24% interrupts.CPU180.PMI:Performance_monitoring_interrupts
> 17.50 ± 32% +371.4% 82.50 ±101% interrupts.CPU181.RES:Rescheduling_interrupts
> 41.00 ±101% +232.9% 136.50 ± 79% interrupts.CPU183.RES:Rescheduling_interrupts
> 5998 ± 31% -29.3% 4238 ± 18% interrupts.CPU192.NMI:Non-maskable_interrupts
> 5998 ± 31% -29.3% 4238 ± 18% interrupts.CPU192.PMI:Performance_monitoring_interrupts
> 5812 ± 21% -40.6% 3450 ± 4% interrupts.CPU198.NMI:Non-maskable_interrupts
> 5812 ± 21% -40.6% 3450 ± 4% interrupts.CPU198.PMI:Performance_monitoring_interrupts
> 5605 ± 22% -27.0% 4093 ± 26% interrupts.CPU200.NMI:Non-maskable_interrupts
> 5605 ± 22% -27.0% 4093 ± 26% interrupts.CPU200.PMI:Performance_monitoring_interrupts
> 5528 ± 27% -38.1% 3423 ± 2% interrupts.CPU212.NMI:Non-maskable_interrupts
> 5528 ± 27% -38.1% 3423 ± 2% interrupts.CPU212.PMI:Performance_monitoring_interrupts
> 6447 ± 27% -31.2% 4433 ± 39% interrupts.CPU216.NMI:Non-maskable_interrupts
> 6447 ± 27% -31.2% 4433 ± 39% interrupts.CPU216.PMI:Performance_monitoring_interrupts
> 561218 ± 5% +10.4% 619618 ± 2% interrupts.CPU220.LOC:Local_timer_interrupts
> 201.00 ±134% -85.0% 30.25 ± 53% interrupts.CPU228.RES:Rescheduling_interrupts
> 5757 ± 22% -35.4% 3719 ± 4% interrupts.CPU229.NMI:Non-maskable_interrupts
> 5757 ± 22% -35.4% 3719 ± 4% interrupts.CPU229.PMI:Performance_monitoring_interrupts
> 165.75 ± 46% -70.6% 48.75 ±103% interrupts.CPU230.RES:Rescheduling_interrupts
> 23.75 ± 38% +346.3% 106.00 ± 87% interrupts.CPU236.RES:Rescheduling_interrupts
> 4576 ± 15% -26.0% 3385 ± 3% interrupts.CPU248.NMI:Non-maskable_interrupts
> 4576 ± 15% -26.0% 3385 ± 3% interrupts.CPU248.PMI:Performance_monitoring_interrupts
> 606.75 ±140% -92.7% 44.00 ± 69% interrupts.CPU252.RES:Rescheduling_interrupts
> 80.00 ± 62% -67.5% 26.00 ± 30% interrupts.CPU256.RES:Rescheduling_interrupts
> 5696 ± 30% -26.9% 4164 ± 26% interrupts.CPU259.NMI:Non-maskable_interrupts
> 5696 ± 30% -26.9% 4164 ± 26% interrupts.CPU259.PMI:Performance_monitoring_interrupts
> 1796 ± 99% -76.2% 428.00 ± 45% interrupts.CPU26.RES:Rescheduling_interrupts
> 109.50 ± 70% -66.0% 37.25 ± 55% interrupts.CPU266.RES:Rescheduling_interrupts
> 143.00 ± 38% -69.1% 44.25 ± 90% interrupts.CPU271.RES:Rescheduling_interrupts
> 5526 ± 29% -34.5% 3619 interrupts.CPU278.NMI:Non-maskable_interrupts
> 5526 ± 29% -34.5% 3619 interrupts.CPU278.PMI:Performance_monitoring_interrupts
> 94.75 ± 42% -55.1% 42.50 ± 64% interrupts.CPU278.RES:Rescheduling_interrupts
> 6671 ± 19% -31.4% 4579 ± 31% interrupts.CPU281.NMI:Non-maskable_interrupts
> 6671 ± 19% -31.4% 4579 ± 31% interrupts.CPU281.PMI:Performance_monitoring_interrupts
> 7140 ± 25% -45.1% 3917 ± 2% interrupts.CPU37.NMI:Non-maskable_interrupts
> 7140 ± 25% -45.1% 3917 ± 2% interrupts.CPU37.PMI:Performance_monitoring_interrupts
> 6176 ± 29% -37.9% 3834 ± 3% interrupts.CPU38.NMI:Non-maskable_interrupts
> 6176 ± 29% -37.9% 3834 ± 3% interrupts.CPU38.PMI:Performance_monitoring_interrupts
> 968.50 ±109% -75.0% 241.75 ± 42% interrupts.CPU38.RES:Rescheduling_interrupts
> 6057 ± 24% -28.2% 4352 ± 23% interrupts.CPU41.NMI:Non-maskable_interrupts
> 6057 ± 24% -28.2% 4352 ± 23% interrupts.CPU41.PMI:Performance_monitoring_interrupts
> 120.00 ± 99% +376.2% 571.50 ± 85% interrupts.CPU47.RES:Rescheduling_interrupts
> 177.75 ± 31% +247.4% 617.50 ± 88% interrupts.CPU48.RES:Rescheduling_interrupts
> 110.00 ± 83% +116.6% 238.25 ± 26% interrupts.CPU49.RES:Rescheduling_interrupts
> 6925 ± 22% -45.4% 3781 ± 3% interrupts.CPU50.NMI:Non-maskable_interrupts
> 6925 ± 22% -45.4% 3781 ± 3% interrupts.CPU50.PMI:Performance_monitoring_interrupts
> 180.25 ± 97% +685.2% 1415 ±115% interrupts.CPU51.RES:Rescheduling_interrupts
> 182.50 ± 53% +1719.2% 3320 ± 62% interrupts.CPU52.RES:Rescheduling_interrupts
> 5043 ± 18% -26.0% 3730 ± 5% interrupts.CPU57.NMI:Non-maskable_interrupts
> 5043 ± 18% -26.0% 3730 ± 5% interrupts.CPU57.PMI:Performance_monitoring_interrupts
> 3655 ± 9% +33.0% 4863 ± 21% interrupts.CPU60.NMI:Non-maskable_interrupts
> 3655 ± 9% +33.0% 4863 ± 21% interrupts.CPU60.PMI:Performance_monitoring_interrupts
> 6130 ± 24% -28.1% 4410 ± 21% interrupts.CPU64.NMI:Non-maskable_interrupts
> 6130 ± 24% -28.1% 4410 ± 21% interrupts.CPU64.PMI:Performance_monitoring_interrupts
> 5811 ± 25% -36.1% 3713 interrupts.CPU76.NMI:Non-maskable_interrupts
> 5811 ± 25% -36.1% 3713 interrupts.CPU76.PMI:Performance_monitoring_interrupts
> 6036 ± 28% -39.4% 3660 ± 5% interrupts.CPU80.NMI:Non-maskable_interrupts
> 6036 ± 28% -39.4% 3660 ± 5% interrupts.CPU80.PMI:Performance_monitoring_interrupts
> 6494 ± 25% -43.9% 3640 ± 5% interrupts.CPU82.NMI:Non-maskable_interrupts
> 6494 ± 25% -43.9% 3640 ± 5% interrupts.CPU82.PMI:Performance_monitoring_interrupts
> 6881 ± 13% -48.8% 3522 ± 3% interrupts.CPU84.NMI:Non-maskable_interrupts
> 6881 ± 13% -48.8% 3522 ± 3% interrupts.CPU84.PMI:Performance_monitoring_interrupts
> 110.75 ± 19% +119.4% 243.00 ± 19% interrupts.CPU9.RES:Rescheduling_interrupts
>
>
>
> will-it-scale.per_process_ops
>
> 95000 +-+-----------------------------------O-----------------------------+
> 90000 +-+ O O O O O |
> | O O OO O O |
> 85000 +-+ OOO O O O |
> 80000 +-+ OO O O O |
> | O O O |
> 75000 +-+O O |
> 70000 OO+ O O O |
> 65000 +-+ |
> | |
> 60000 +-+ |
> 55000 +-+ + + .+ .+ ++. |
> | .+++.+ ++.+ :.+ + ++ + +++.+++.++ + .+++.+ +. + |
> 50000 +-+ + + :+ +++ + .++ + ++.+++.+|
> 45000 +-+-----------------------------------------------------------------+
>
>
> [*] bisect-good sample
> [O] bisect-bad sample
>
>
>
> Disclaimer:
> Results have been estimated based on internal Intel analysis and are provided
> for informational purposes only. Any difference in system hardware or software
> design or configuration may affect actual performance.
>
>
> Thanks,
> Rong Chen
> ---
>
> #! jobs/will-it-scale-100.yaml
> suite: will-it-scale
> testcase: will-it-scale
> category: benchmark
> nr_task: 100%
> will-it-scale:
> mode: process
> test: brk1
> job_origin: "/lkp/lkp/.src-20181229-091801/allot/cyclic:p1:linux-devel:devel-hourly/lkp-knm01/will-it-scale-100.yaml"
>
> #! queue options
> queue: bisect
> testbox: lkp-knm01
> tbox_group: lkp-knm01
> submit_id: 5c2b509f0b9a935849d3d1a6
> job_file: "/lkp/jobs/scheduled/lkp-knm01/will-it-scale-performance-process-100%-brk1-debian-x86_64-2018-04-03.cgz-7e28c5af4ef6b539334aa5de40feca0c041c94df-20190101-22601-1ncxbj9-0.yaml"
> id: 3b65780fc5b8920f2e29ad867cd42a75eee7bd74
> queuer_version: "/lkp/lkp/.src-20181229-164014"
>
> #! hosts/lkp-knm01
>
> #! include/category/benchmark
> kmsg:
> boot-time:
> iostat:
> heartbeat:
> vmstat:
> numa-numastat:
> numa-vmstat:
> numa-meminfo:
> proc-vmstat:
> proc-stat:
> meminfo:
> slabinfo:
> interrupts:
> lock_stat:
> latency_stats:
> softirqs:
> bdi_dev_mapping:
> diskstats:
> nfsstat:
> cpuidle:
> cpufreq-stats:
> turbostat:
> sched_debug:
> perf-stat:
> mpstat:
> perf-profile:
>
> #! include/category/ALL
> cpufreq_governor: performance
>
> #! include/queue/cyclic
> commit: 7e28c5af4ef6b539334aa5de40feca0c041c94df
>
> #! default params
> kconfig: x86_64-rhel-7.2
> compiler: gcc-7
> rootfs: debian-x86_64-2018-04-03.cgz
> enqueue_time: 2019-01-01 19:35:59.961441013 +08:00
> _id: 5c2b509f0b9a935849d3d1a6
> _rt: "/result/will-it-scale/performance-process-100%-brk1/lkp-knm01/debian-x86_64-2018-04-03.cgz/x86_64-rhel-7.2/gcc-7/7e28c5af4ef6b539334aa5de40feca0c041c94df"
>
> #! schedule options
> user: lkp
> head_commit: ab4756ff8ee9144cb7045eff9043ed2c6522e3b1
> base_commit: 8fe28cb58bcb235034b64cbbb7550a8a43fd88be
> branch: linux-devel/devel-hourly-2018123112
> result_root: "/result/will-it-scale/performance-process-100%-brk1/lkp-knm01/debian-x86_64-2018-04-03.cgz/x86_64-rhel-7.2/gcc-7/7e28c5af4ef6b539334aa5de40feca0c041c94df/0"
> scheduler_version: "/lkp/lkp/.src-20181229-164014"
> LKP_SERVER: inn
> max_uptime: 1500
> initrd: "/osimage/debian/debian-x86_64-2018-04-03.cgz"
> bootloader_append:
> - root=/dev/ram0
> - user=lkp
> - job=/lkp/jobs/scheduled/lkp-knm01/will-it-scale-performance-process-100%-brk1-debian-x86_64-2018-04-03.cgz-7e28c5af4ef6b539334aa5de40feca0c041c94df-20190101-22601-1ncxbj9-0.yaml
> - ARCH=x86_64
> - kconfig=x86_64-rhel-7.2
> - branch=linux-devel/devel-hourly-2018123112
> - commit=7e28c5af4ef6b539334aa5de40feca0c041c94df
> - BOOT_IMAGE=/pkg/linux/x86_64-rhel-7.2/gcc-7/7e28c5af4ef6b539334aa5de40feca0c041c94df/vmlinuz-4.19.0-rc1-00100-g7e28c5a
> - max_uptime=1500
> - RESULT_ROOT=/result/will-it-scale/performance-process-100%-brk1/lkp-knm01/debian-x86_64-2018-04-03.cgz/x86_64-rhel-7.2/gcc-7/7e28c5af4ef6b539334aa5de40feca0c041c94df/0
> - LKP_SERVER=inn
> - debug
> - apic=debug
> - sysrq_always_enabled
> - rcupdate.rcu_cpu_stall_timeout=100
> - net.ifnames=0
> - printk.devkmsg=on
> - panic=-1
> - softlockup_panic=1
> - nmi_watchdog=panic
> - oops=panic
> - load_ramdisk=2
> - prompt_ramdisk=0
> - drbd.minor_count=8
> - systemd.log_level=err
> - ignore_loglevel
> - console=tty0
> - earlyprintk=ttyS0,115200
> - console=ttyS0,115200
> - vga=normal
> - rw
> modules_initrd: "/pkg/linux/x86_64-rhel-7.2/gcc-7/7e28c5af4ef6b539334aa5de40feca0c041c94df/modules.cgz"
> bm_initrd: "/osimage/deps/debian-x86_64-2018-04-03.cgz/run-ipconfig_2018-04-03.cgz,/osimage/deps/debian-x86_64-2018-04-03.cgz/lkp_2018-04-03.cgz,/osimage/deps/debian-x86_64-2018-04-03.cgz/rsync-rootfs_2018-04-03.cgz,/osimage/deps/debian-x86_64-2018-04-03.cgz/will-it-scale_2018-05-17.cgz,/osimage/pkg/debian-x86_64-2018-04-03.cgz/will-it-scale-x86_64-decad85_2018-06-07.cgz,/osimage/deps/debian-x86_64-2018-04-03.cgz/mpstat_2018-06-19.cgz,/osimage/deps/debian-x86_64-2018-04-03.cgz/turbostat_2018-05-17.cgz,/osimage/pkg/debian-x86_64-2018-04-03.cgz/turbostat-x86_64-d5256b2_2018-05-18.cgz,/osimage/deps/debian-x86_64-2018-04-03.cgz/perf_2019-01-01.cgz,/osimage/pkg/debian-x86_64-2018-04-03.cgz/perf-x86_64-e1ef035d272e_2019-01-01.cgz,/osimage/deps/debian-x86_64-2018-04-03.cgz/hw_2016-11-15.cgz"
> lkp_initrd: "/lkp/lkp/lkp-x86_64.cgz"
> site: inn
>
> #! /lkp/lkp/.src-20181229-164014/include/site/inn
> LKP_CGI_PORT: 80
> LKP_CIFS_PORT: 139
> oom-killer:
> watchdog:
>
> #! runtime status
> repeat_to: 2
> model: Knights Mill
> nr_node: 1
> nr_cpu: 288
> memory: 80G
> hdd_partitions:
> swap_partitions: LABEL=SWAP
> rootfs_partition: LABEL=LKP-ROOTFS
>
> #! user overrides
> kernel: "/pkg/linux/x86_64-rhel-7.2/gcc-7/7e28c5af4ef6b539334aa5de40feca0c041c94df/vmlinuz-4.19.0-rc1-00100-g7e28c5a"
> dequeue_time: 2019-01-01 20:16:40.604583551 +08:00
> job_state: finished
> loadavg: 60.12 134.31 72.59 1/1845 11197
> start_time: '1546345124'
> end_time: '1546345447'
> version: "/lkp/lkp/.src-20181229-164014"
>
> for cpu_dir in /sys/devices/system/cpu/cpu[0-9]*
> do
> online_file="$cpu_dir"/online
> [ -f "$online_file" ] && [ "$(cat "$online_file")" -eq 0 ] && continue
>
> file="$cpu_dir"/cpufreq/scaling_governor
> [ -f "$file" ] && echo "performance" > "$file"
> done
>
> "python2" "./runtest.py" "brk1" "295" "process" "288"
Powered by blists - more mailing lists