lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  PHC 
Open Source and information security mailing list archives
 
Hash Suite for Android: free password hash cracker in your pocket
[<prev] [next>] [<thread-prev] [day] [month] [year] [list]
Date:   Sun, 3 Mar 2019 10:34:19 -0800
From:   "Paul E. McKenney" <paulmck@...ux.ibm.com>
To:     kernel test robot <rong.a.chen@...el.com>
Cc:     LKML <linux-kernel@...r.kernel.org>,
        Linus Torvalds <torvalds@...ux-foundation.org>, lkp@...org
Subject: Re: [LKP] [rcu]  7e28c5af4e:  will-it-scale.per_process_ops 84.7%
 improvement

On Thu, Feb 28, 2019 at 12:50:15PM +0800, kernel test robot wrote:
> Greeting,
> 
> FYI, we noticed a 84.7% improvement of will-it-scale.per_process_ops due to commit:
> 
> 
> commit: 7e28c5af4ef6b539334aa5de40feca0c041c94df ("rcu: Eliminate ->rcu_qs_ctr from the rcu_dynticks structure")
> https://git.kernel.org/cgit/linux/kernel/git/torvalds/linux.git master
> 
> in testcase: will-it-scale
> on test machine: 288 threads Intel(R) Xeon Phi(TM) CPU 7295 @ 1.50GHz with 80G memory
> with following parameters:
> 
> 	nr_task: 100%
> 	mode: process
> 	test: brk1
> 	cpufreq_governor: performance
> 
> test-description: Will It Scale takes a testcase and runs it from 1 through to n parallel copies to see if the testcase will scale. It builds both a process and threads based test in order to see any differences between the two.
> test-url: https://github.com/antonblanchard/will-it-scale

It looks like getting rid of the rcu_dynticks structure improved cache
locality.  Or am I missing something in the wealth of statistic herein?

							Thanx, Paul

> Details are as below:
> -------------------------------------------------------------------------------------------------->
> 
> 
> To reproduce:
> 
>         git clone https://github.com/intel/lkp-tests.git
>         cd lkp-tests
>         bin/lkp install job.yaml  # job file is attached in this email
>         bin/lkp run     job.yaml
> 
> =========================================================================================
> compiler/cpufreq_governor/kconfig/mode/nr_task/rootfs/tbox_group/test/testcase:
>   gcc-7/performance/x86_64-rhel-7.2/process/100%/debian-x86_64-2018-04-03.cgz/lkp-knm01/brk1/will-it-scale
> 
> commit: 
>   c5bacd9417 ("rcu: Motivate Tiny RCU forward progress")
>   7e28c5af4e ("rcu: Eliminate ->rcu_qs_ctr from the rcu_dynticks structure")
> 
> c5bacd94173ec49d 7e28c5af4ef6b539334aa5de40 
> ---------------- -------------------------- 
>          %stddev     %change         %stddev
>              \          |                \  
>      48605           +84.7%      89756 ±  4%  will-it-scale.per_process_ops
>   13998509           +84.7%   25850040 ±  4%  will-it-scale.workload
>       0.01 ± 63%      -0.0        0.00 ±131%  mpstat.cpu.soft%
>      11.86 ±  3%      -1.2       10.65 ±  4%  mpstat.cpu.usr%
>     662910           +15.8%     767829 ±  3%  numa-numastat.node0.local_node
>     662889           +15.8%     767801 ±  3%  numa-numastat.node0.numa_hit
>       2149 ±  4%     -16.9%       1785 ±  2%  vmstat.system.cs
>     727237 ±  3%     -14.9%     618668        vmstat.system.in
>     284156 ±  4%     +32.5%     376625        numa-meminfo.node0.Active
>     283908 ±  4%     +32.6%     376374        numa-meminfo.node0.Active(anon)
>     110629 ± 15%     -77.7%      24625 ± 18%  numa-meminfo.node0.Inactive
>     110393 ± 15%     -77.9%      24400 ± 19%  numa-meminfo.node0.Inactive(anon)
>     284255 ±  4%     +32.5%     376666        meminfo.Active
>     284007 ±  4%     +32.5%     376415        meminfo.Active(anon)
>     110855 ± 15%     -77.8%      24620 ± 18%  meminfo.Inactive
>     110619 ± 15%     -77.9%      24395 ± 19%  meminfo.Inactive(anon)
>       8.00 ±173%  +11303.1%     912.25 ± 87%  meminfo.Mlocked
>      11072           -15.7%       9338 ±  3%  meminfo.max_used_kB
>      71004 ±  4%     +32.6%      94124        numa-vmstat.node0.nr_active_anon
>      27812 ± 15%     -78.1%       6100 ± 19%  numa-vmstat.node0.nr_inactive_anon
>       1.00 ±173%  +13100.0%     132.00 ± 87%  numa-vmstat.node0.nr_mlock
>      71002 ±  4%     +32.6%      94122        numa-vmstat.node0.nr_zone_active_anon
>      27811 ± 15%     -78.1%       6100 ± 19%  numa-vmstat.node0.nr_zone_inactive_anon
>       1.00 ±173%   +9450.0%      95.50 ± 87%  numa-vmstat.node1.nr_mlock
>      70939 ±  4%     +32.8%      94186        proc-vmstat.nr_active_anon
>      27523 ± 15%     -77.8%       6101 ± 19%  proc-vmstat.nr_inactive_anon
>       2.25 ±173%  +10044.4%     228.25 ± 86%  proc-vmstat.nr_mlock
>      70939 ±  4%     +32.8%      94186        proc-vmstat.nr_zone_active_anon
>      27523 ± 15%     -77.8%       6101 ± 19%  proc-vmstat.nr_zone_inactive_anon
>     680060           +15.8%     787707 ±  2%  proc-vmstat.numa_hit
>     680060           +15.8%     787707 ±  2%  proc-vmstat.numa_local
>     140509 ± 10%     -45.3%      76820 ± 24%  proc-vmstat.numa_pte_updates
>     895.50 ± 30%    +480.0%       5193 ± 36%  proc-vmstat.pgactivate
>     774283           +13.9%     882060 ±  3%  proc-vmstat.pgalloc_normal
>     653457           +21.9%     796258 ±  3%  proc-vmstat.pgfault
>     610954 ±  2%     +32.4%     808662 ±  3%  proc-vmstat.pgfree
>       5299 ±  6%    +106.4%      10939 ± 36%  softirqs.CPU10.RCU
>     106293 ±  5%     +28.2%     136274 ± 15%  softirqs.CPU126.TIMER
>     108234 ±  6%     +26.3%     136696 ± 15%  softirqs.CPU127.TIMER
>     109747 ±  7%     +26.4%     138696 ± 11%  softirqs.CPU135.TIMER
>     111732 ±  8%     +24.4%     138967 ± 11%  softirqs.CPU138.TIMER
>     107031 ±  8%     +26.7%     135558 ± 14%  softirqs.CPU148.TIMER
>     100841           +23.1%     124155 ± 15%  softirqs.CPU15.TIMER
>     104640 ±  4%     +27.8%     133719 ± 16%  softirqs.CPU172.TIMER
>     103142 ±  3%     +30.9%     135023 ± 13%  softirqs.CPU190.TIMER
>     106698 ± 10%     +27.5%     136023 ± 13%  softirqs.CPU197.TIMER
>     110268 ±  2%     +22.5%     135052 ± 12%  softirqs.CPU205.TIMER
>     106111 ±  8%     +23.9%     131436 ± 14%  softirqs.CPU206.TIMER
>     106894 ±  6%     +28.7%     137574 ± 14%  softirqs.CPU213.TIMER
>     108418 ±  8%     +28.3%     139100 ± 12%  softirqs.CPU214.TIMER
>     100170 ± 10%     +32.8%     132978 ± 15%  softirqs.CPU215.TIMER
>     100836 ±  2%     +30.3%     131406 ± 17%  softirqs.CPU217.TIMER
>     112497 ±  8%     +22.9%     138291 ± 13%  softirqs.CPU219.TIMER
>     108669 ± 10%     +22.1%     132720 ± 15%  softirqs.CPU220.TIMER
>     108584 ±  8%     +25.5%     136230 ± 13%  softirqs.CPU227.TIMER
>     106251 ±  6%     +28.6%     136686 ± 11%  softirqs.CPU236.TIMER
>     105296 ±  7%     +31.3%     138251 ± 12%  softirqs.CPU238.TIMER
>       3915 ±  8%     +77.4%       6945 ± 30%  softirqs.CPU249.RCU
>     101465 ±  2%     +26.5%     128388 ± 19%  softirqs.CPU271.TIMER
>     104242 ±  8%     +23.7%     128907 ± 17%  softirqs.CPU70.TIMER
>     107020 ±  7%     +26.9%     135845 ± 13%  softirqs.CPU92.TIMER
>  9.659e+09 ± 10%     +16.3%  1.124e+10 ±  3%  perf-stat.i.branch-instructions
>       6.69            +1.0        7.69 ±  2%  perf-stat.i.branch-miss-rate%
>  6.443e+08 ± 10%     +33.3%   8.59e+08 ±  2%  perf-stat.i.branch-misses
>       4.10 ±  6%      +0.9        5.04 ± 10%  perf-stat.i.cache-miss-rate%
>   67464898 ± 10%     +25.2%   84491315 ±  4%  perf-stat.i.cache-misses
>       2820 ± 13%     -32.4%       1906 ±  3%  perf-stat.i.context-switches
>      15.34           -37.0%       9.67 ±  3%  perf-stat.i.cpi
>  7.067e+11 ± 11%     -30.2%  4.932e+11 ±  4%  perf-stat.i.cpu-cycles
>      71.93 ± 10%     -41.9%      41.79 ±  9%  perf-stat.i.cpu-migrations
>       2.36            -0.0        2.33        perf-stat.i.iTLB-load-miss-rate%
>  1.106e+09 ± 10%     +15.1%  1.273e+09 ±  2%  perf-stat.i.iTLB-load-misses
>  4.569e+10 ± 10%     +17.0%  5.345e+10 ±  3%  perf-stat.i.iTLB-loads
>  4.621e+10 ± 10%     +16.1%  5.364e+10 ±  3%  perf-stat.i.instructions
>       0.07           +64.1%       0.11 ±  3%  perf-stat.i.ipc
>       3524 ± 11%     -18.3%       2878        perf-stat.i.minor-faults
>     993359 ± 11%     -32.0%     675215 ±  2%  perf-stat.i.msec
>       3529 ± 11%     -18.5%       2878        perf-stat.i.page-faults
>       6.67            +1.0        7.65        perf-stat.overall.branch-miss-rate%
>       3.91            +0.6        4.54 ±  3%  perf-stat.overall.cache-miss-rate%
>      15.28           -39.8%       9.20 ±  3%  perf-stat.overall.cpi
>       2.36            -0.0        2.33        perf-stat.overall.iTLB-load-miss-rate%
>      41.78            +0.9%      42.13        perf-stat.overall.instructions-per-iTLB-miss
>       0.07           +66.4%       0.11 ±  3%  perf-stat.overall.ipc
>     580565            -5.5%     548645        perf-stat.overall.path-length
>  1.699e+12           +74.9%  2.971e+12 ±  4%  perf-stat.total.branch-instructions
>  1.133e+11          +100.6%  2.272e+11 ±  5%  perf-stat.total.branch-misses
>  1.186e+10           +88.0%   2.23e+10        perf-stat.total.cache-misses
>  3.032e+11           +62.2%  4.918e+11 ±  2%  perf-stat.total.cache-references
>  1.242e+14            +4.8%  1.302e+14        perf-stat.total.cpu-cycles
>      12654           -13.0%      11014 ±  5%  perf-stat.total.cpu-migrations
>  1.945e+11           +73.0%  3.365e+11 ±  3%  perf-stat.total.iTLB-load-misses
>  8.038e+12           +75.8%  1.413e+13 ±  4%  perf-stat.total.iTLB-loads
>  8.127e+12           +74.5%  1.418e+13 ±  4%  perf-stat.total.instructions
>     619183           +22.9%     760791 ±  3%  perf-stat.total.minor-faults
>     620031           +22.7%     760750 ±  3%  perf-stat.total.page-faults
>   23472574 ± 28%     -50.0%   11736977 ± 60%  sched_debug.cfs_rq:/.MIN_vruntime.max
>    1380731 ± 28%     -46.9%     733243 ± 66%  sched_debug.cfs_rq:/.MIN_vruntime.stddev
>     109109           +17.1%     127734        sched_debug.cfs_rq:/.exec_clock.avg
>     115740           +12.4%     130080        sched_debug.cfs_rq:/.exec_clock.max
>       4252 ±  4%     -50.4%       2110 ± 36%  sched_debug.cfs_rq:/.exec_clock.stddev
>       6243 ± 12%     -14.0%       5369 ± 15%  sched_debug.cfs_rq:/.load.avg
>     887713 ± 24%     -30.6%     616115 ± 43%  sched_debug.cfs_rq:/.load.max
>      53788 ± 23%     -30.4%      37427 ± 40%  sched_debug.cfs_rq:/.load.stddev
>   23472574 ± 28%     -50.0%   11736977 ± 60%  sched_debug.cfs_rq:/.max_vruntime.max
>    1380731 ± 28%     -46.9%     733243 ± 66%  sched_debug.cfs_rq:/.max_vruntime.stddev
>   36862110           +17.8%   43411425        sched_debug.cfs_rq:/.min_vruntime.avg
>   39448691           +12.5%   44387703        sched_debug.cfs_rq:/.min_vruntime.max
>    1668979 ±  3%     -45.0%     917652 ± 30%  sched_debug.cfs_rq:/.min_vruntime.stddev
>       1.38 ±  9%      -9.1%       1.25 ±  6%  sched_debug.cfs_rq:/.nr_running.max
>       0.56 ± 19%     +42.2%       0.80        sched_debug.cfs_rq:/.nr_running.min
>       0.07 ± 13%     -32.7%       0.05 ± 15%  sched_debug.cfs_rq:/.nr_running.stddev
>       2.25 ±  3%     -51.5%       1.09 ±  2%  sched_debug.cfs_rq:/.nr_spread_over.avg
>       6235 ± 12%     -14.0%       5361 ± 16%  sched_debug.cfs_rq:/.runnable_weight.avg
>     887713 ± 24%     -30.6%     616088 ± 43%  sched_debug.cfs_rq:/.runnable_weight.max
>      53791 ± 23%     -30.4%      37423 ± 40%  sched_debug.cfs_rq:/.runnable_weight.stddev
>     513480 ± 46%     -68.1%     163689 ± 16%  sched_debug.cfs_rq:/.spread0.avg
>     992070 ± 25%     -55.6%     440313 ±  8%  sched_debug.cfs_rq:/.spread0.max
>     271.31 ± 25%     +54.1%     418.15 ± 23%  sched_debug.cfs_rq:/.util_avg.min
>      71.69 ±  6%     -14.9%      61.03 ±  7%  sched_debug.cfs_rq:/.util_avg.stddev
>     638.86           +18.5%     756.85 ±  3%  sched_debug.cfs_rq:/.util_est_enqueued.avg
>     131.07 ±  2%     -17.0%     108.78 ±  5%  sched_debug.cfs_rq:/.util_est_enqueued.stddev
>    1617727 ±  4%     -24.5%    1222141 ± 11%  sched_debug.cpu.avg_idle.avg
>     167800 ±  2%     +13.3%     190124 ±  4%  sched_debug.cpu.clock.avg
>     174380           +10.3%     192313 ±  4%  sched_debug.cpu.clock.max
>     161230 ±  2%     +16.5%     187888 ±  4%  sched_debug.cpu.clock.min
>       3797 ±  4%     -66.4%       1276 ±  8%  sched_debug.cpu.clock.stddev
>     167800 ±  2%     +13.3%     190124 ±  4%  sched_debug.cpu.clock_task.avg
>     174380           +10.3%     192313 ±  4%  sched_debug.cpu.clock_task.max
>     161230 ±  2%     +16.5%     187888 ±  4%  sched_debug.cpu.clock_task.min
>       3797 ±  4%     -66.4%       1276 ±  8%  sched_debug.cpu.clock_task.stddev
>       5416 ±  2%     +15.6%       6262 ±  5%  sched_debug.cpu.curr->pid.max
>       2045 ± 18%     +20.6%       2466 ±  3%  sched_debug.cpu.curr->pid.min
>     892728 ±  7%     -37.5%     557989 ±  5%  sched_debug.cpu.max_idle_balance_cost.avg
>    4113479 ± 18%     -51.6%    1989769 ± 36%  sched_debug.cpu.max_idle_balance_cost.max
>     461057 ± 17%     -67.6%     149491 ± 36%  sched_debug.cpu.max_idle_balance_cost.stddev
>       0.00 ±  4%     -65.8%       0.00 ± 23%  sched_debug.cpu.next_balance.stddev
>     117031           +16.4%     136279        sched_debug.cpu.nr_load_updates.avg
>     139472 ±  3%     +16.0%     161763 ±  4%  sched_debug.cpu.nr_load_updates.max
>     109218           +22.1%     133390        sched_debug.cpu.nr_load_updates.min
>       3910 ±  3%     -46.0%       2111 ± 19%  sched_debug.cpu.nr_load_updates.stddev
>       0.16 ±  4%     -19.4%       0.13 ±  3%  sched_debug.cpu.nr_running.stddev
>       1108           +12.5%       1247 ±  2%  sched_debug.cpu.nr_switches.avg
>      26767 ± 16%     +60.1%      42860 ± 15%  sched_debug.cpu.nr_switches.max
>       2342 ±  4%     +34.2%       3144 ±  7%  sched_debug.cpu.nr_switches.stddev
>      29066 ± 11%     +83.5%      53341 ± 36%  sched_debug.cpu.sched_count.max
>       2473 ±  2%     +61.3%       3988 ± 33%  sched_debug.cpu.sched_count.stddev
>     275.25 ±  3%     +33.5%     367.54        sched_debug.cpu.ttwu_count.avg
>      10910 ± 17%     +67.3%      18254 ± 13%  sched_debug.cpu.ttwu_count.max
>      50.31 ±  2%     +19.9%      60.30 ±  2%  sched_debug.cpu.ttwu_count.min
>     832.06 ± 10%     +46.1%       1215 ±  7%  sched_debug.cpu.ttwu_count.stddev
>     176.33           +42.8%     251.81 ±  2%  sched_debug.cpu.ttwu_local.avg
>      10535 ± 18%     +69.7%      17877 ± 14%  sched_debug.cpu.ttwu_local.max
>      31.56           +15.2%      36.35        sched_debug.cpu.ttwu_local.min
>     729.04 ± 13%     +57.2%       1146 ±  9%  sched_debug.cpu.ttwu_local.stddev
>     161164 ±  2%     +16.6%     187864 ±  4%  sched_debug.cpu_clk
>     160513 ±  2%     +16.6%     187213 ±  4%  sched_debug.ktime
>       1.31 ± 24%     +39.1%       1.82        sched_debug.rt_rq:/.rt_runtime.stddev
>     161891 ±  2%     +16.3%     188213 ±  4%  sched_debug.sched_clk
>     745.75 ± 61%     -87.9%      90.25 ±173%  interrupts.36:IR-PCI-MSI.2621442-edge.eth1-TxRx-1
>     961.75 ± 48%     -75.5%     235.25 ±173%  interrupts.37:IR-PCI-MSI.2621443-edge.eth1-TxRx-2
>       4649 ± 16%     -25.2%       3475 ±  3%  interrupts.CPU100.NMI:Non-maskable_interrupts
>       4649 ± 16%     -25.2%       3475 ±  3%  interrupts.CPU100.PMI:Performance_monitoring_interrupts
>      30.50 ± 17%    +277.9%     115.25 ± 49%  interrupts.CPU109.RES:Rescheduling_interrupts
>     552027 ±  5%     +12.2%     619466 ±  2%  interrupts.CPU110.LOC:Local_timer_interrupts
>     274.50 ± 59%     -44.9%     151.25 ± 92%  interrupts.CPU116.RES:Rescheduling_interrupts
>      37.50 ± 47%    +323.3%     158.75 ± 43%  interrupts.CPU121.RES:Rescheduling_interrupts
>       7275 ± 12%     -45.9%       3937 ± 25%  interrupts.CPU125.NMI:Non-maskable_interrupts
>       7275 ± 12%     -45.9%       3937 ± 25%  interrupts.CPU125.PMI:Performance_monitoring_interrupts
>       5151 ± 34%     -32.6%       3470 ±  6%  interrupts.CPU140.NMI:Non-maskable_interrupts
>       5151 ± 34%     -32.6%       3470 ±  6%  interrupts.CPU140.PMI:Performance_monitoring_interrupts
>       5589 ± 28%     -26.0%       4134 ± 32%  interrupts.CPU153.NMI:Non-maskable_interrupts
>       5589 ± 28%     -26.0%       4134 ± 32%  interrupts.CPU153.PMI:Performance_monitoring_interrupts
>     745.75 ± 61%     -87.9%      90.25 ±173%  interrupts.CPU16.36:IR-PCI-MSI.2621442-edge.eth1-TxRx-1
>     961.75 ± 48%     -75.5%     235.25 ±173%  interrupts.CPU17.37:IR-PCI-MSI.2621443-edge.eth1-TxRx-2
>       6606 ± 25%     -38.6%       4057 ± 27%  interrupts.CPU171.NMI:Non-maskable_interrupts
>       6606 ± 25%     -38.6%       4057 ± 27%  interrupts.CPU171.PMI:Performance_monitoring_interrupts
>       5916 ± 21%     -28.9%       4205 ± 28%  interrupts.CPU172.NMI:Non-maskable_interrupts
>       5916 ± 21%     -28.9%       4205 ± 28%  interrupts.CPU172.PMI:Performance_monitoring_interrupts
>      85.50 ± 36%     -79.2%      17.75 ± 72%  interrupts.CPU176.RES:Rescheduling_interrupts
>      74.00 ± 79%     -79.1%      15.50 ± 27%  interrupts.CPU177.RES:Rescheduling_interrupts
>       6513 ± 21%     -36.4%       4139 ± 24%  interrupts.CPU180.NMI:Non-maskable_interrupts
>       6513 ± 21%     -36.4%       4139 ± 24%  interrupts.CPU180.PMI:Performance_monitoring_interrupts
>      17.50 ± 32%    +371.4%      82.50 ±101%  interrupts.CPU181.RES:Rescheduling_interrupts
>      41.00 ±101%    +232.9%     136.50 ± 79%  interrupts.CPU183.RES:Rescheduling_interrupts
>       5998 ± 31%     -29.3%       4238 ± 18%  interrupts.CPU192.NMI:Non-maskable_interrupts
>       5998 ± 31%     -29.3%       4238 ± 18%  interrupts.CPU192.PMI:Performance_monitoring_interrupts
>       5812 ± 21%     -40.6%       3450 ±  4%  interrupts.CPU198.NMI:Non-maskable_interrupts
>       5812 ± 21%     -40.6%       3450 ±  4%  interrupts.CPU198.PMI:Performance_monitoring_interrupts
>       5605 ± 22%     -27.0%       4093 ± 26%  interrupts.CPU200.NMI:Non-maskable_interrupts
>       5605 ± 22%     -27.0%       4093 ± 26%  interrupts.CPU200.PMI:Performance_monitoring_interrupts
>       5528 ± 27%     -38.1%       3423 ±  2%  interrupts.CPU212.NMI:Non-maskable_interrupts
>       5528 ± 27%     -38.1%       3423 ±  2%  interrupts.CPU212.PMI:Performance_monitoring_interrupts
>       6447 ± 27%     -31.2%       4433 ± 39%  interrupts.CPU216.NMI:Non-maskable_interrupts
>       6447 ± 27%     -31.2%       4433 ± 39%  interrupts.CPU216.PMI:Performance_monitoring_interrupts
>     561218 ±  5%     +10.4%     619618 ±  2%  interrupts.CPU220.LOC:Local_timer_interrupts
>     201.00 ±134%     -85.0%      30.25 ± 53%  interrupts.CPU228.RES:Rescheduling_interrupts
>       5757 ± 22%     -35.4%       3719 ±  4%  interrupts.CPU229.NMI:Non-maskable_interrupts
>       5757 ± 22%     -35.4%       3719 ±  4%  interrupts.CPU229.PMI:Performance_monitoring_interrupts
>     165.75 ± 46%     -70.6%      48.75 ±103%  interrupts.CPU230.RES:Rescheduling_interrupts
>      23.75 ± 38%    +346.3%     106.00 ± 87%  interrupts.CPU236.RES:Rescheduling_interrupts
>       4576 ± 15%     -26.0%       3385 ±  3%  interrupts.CPU248.NMI:Non-maskable_interrupts
>       4576 ± 15%     -26.0%       3385 ±  3%  interrupts.CPU248.PMI:Performance_monitoring_interrupts
>     606.75 ±140%     -92.7%      44.00 ± 69%  interrupts.CPU252.RES:Rescheduling_interrupts
>      80.00 ± 62%     -67.5%      26.00 ± 30%  interrupts.CPU256.RES:Rescheduling_interrupts
>       5696 ± 30%     -26.9%       4164 ± 26%  interrupts.CPU259.NMI:Non-maskable_interrupts
>       5696 ± 30%     -26.9%       4164 ± 26%  interrupts.CPU259.PMI:Performance_monitoring_interrupts
>       1796 ± 99%     -76.2%     428.00 ± 45%  interrupts.CPU26.RES:Rescheduling_interrupts
>     109.50 ± 70%     -66.0%      37.25 ± 55%  interrupts.CPU266.RES:Rescheduling_interrupts
>     143.00 ± 38%     -69.1%      44.25 ± 90%  interrupts.CPU271.RES:Rescheduling_interrupts
>       5526 ± 29%     -34.5%       3619        interrupts.CPU278.NMI:Non-maskable_interrupts
>       5526 ± 29%     -34.5%       3619        interrupts.CPU278.PMI:Performance_monitoring_interrupts
>      94.75 ± 42%     -55.1%      42.50 ± 64%  interrupts.CPU278.RES:Rescheduling_interrupts
>       6671 ± 19%     -31.4%       4579 ± 31%  interrupts.CPU281.NMI:Non-maskable_interrupts
>       6671 ± 19%     -31.4%       4579 ± 31%  interrupts.CPU281.PMI:Performance_monitoring_interrupts
>       7140 ± 25%     -45.1%       3917 ±  2%  interrupts.CPU37.NMI:Non-maskable_interrupts
>       7140 ± 25%     -45.1%       3917 ±  2%  interrupts.CPU37.PMI:Performance_monitoring_interrupts
>       6176 ± 29%     -37.9%       3834 ±  3%  interrupts.CPU38.NMI:Non-maskable_interrupts
>       6176 ± 29%     -37.9%       3834 ±  3%  interrupts.CPU38.PMI:Performance_monitoring_interrupts
>     968.50 ±109%     -75.0%     241.75 ± 42%  interrupts.CPU38.RES:Rescheduling_interrupts
>       6057 ± 24%     -28.2%       4352 ± 23%  interrupts.CPU41.NMI:Non-maskable_interrupts
>       6057 ± 24%     -28.2%       4352 ± 23%  interrupts.CPU41.PMI:Performance_monitoring_interrupts
>     120.00 ± 99%    +376.2%     571.50 ± 85%  interrupts.CPU47.RES:Rescheduling_interrupts
>     177.75 ± 31%    +247.4%     617.50 ± 88%  interrupts.CPU48.RES:Rescheduling_interrupts
>     110.00 ± 83%    +116.6%     238.25 ± 26%  interrupts.CPU49.RES:Rescheduling_interrupts
>       6925 ± 22%     -45.4%       3781 ±  3%  interrupts.CPU50.NMI:Non-maskable_interrupts
>       6925 ± 22%     -45.4%       3781 ±  3%  interrupts.CPU50.PMI:Performance_monitoring_interrupts
>     180.25 ± 97%    +685.2%       1415 ±115%  interrupts.CPU51.RES:Rescheduling_interrupts
>     182.50 ± 53%   +1719.2%       3320 ± 62%  interrupts.CPU52.RES:Rescheduling_interrupts
>       5043 ± 18%     -26.0%       3730 ±  5%  interrupts.CPU57.NMI:Non-maskable_interrupts
>       5043 ± 18%     -26.0%       3730 ±  5%  interrupts.CPU57.PMI:Performance_monitoring_interrupts
>       3655 ±  9%     +33.0%       4863 ± 21%  interrupts.CPU60.NMI:Non-maskable_interrupts
>       3655 ±  9%     +33.0%       4863 ± 21%  interrupts.CPU60.PMI:Performance_monitoring_interrupts
>       6130 ± 24%     -28.1%       4410 ± 21%  interrupts.CPU64.NMI:Non-maskable_interrupts
>       6130 ± 24%     -28.1%       4410 ± 21%  interrupts.CPU64.PMI:Performance_monitoring_interrupts
>       5811 ± 25%     -36.1%       3713        interrupts.CPU76.NMI:Non-maskable_interrupts
>       5811 ± 25%     -36.1%       3713        interrupts.CPU76.PMI:Performance_monitoring_interrupts
>       6036 ± 28%     -39.4%       3660 ±  5%  interrupts.CPU80.NMI:Non-maskable_interrupts
>       6036 ± 28%     -39.4%       3660 ±  5%  interrupts.CPU80.PMI:Performance_monitoring_interrupts
>       6494 ± 25%     -43.9%       3640 ±  5%  interrupts.CPU82.NMI:Non-maskable_interrupts
>       6494 ± 25%     -43.9%       3640 ±  5%  interrupts.CPU82.PMI:Performance_monitoring_interrupts
>       6881 ± 13%     -48.8%       3522 ±  3%  interrupts.CPU84.NMI:Non-maskable_interrupts
>       6881 ± 13%     -48.8%       3522 ±  3%  interrupts.CPU84.PMI:Performance_monitoring_interrupts
>     110.75 ± 19%    +119.4%     243.00 ± 19%  interrupts.CPU9.RES:Rescheduling_interrupts
> 
> 
>                                                                                 
>                             will-it-scale.per_process_ops                       
>                                                                                 
>   95000 +-+-----------------------------------O-----------------------------+   
>   90000 +-+              O  O   O O               O                         |   
>         |               O O  OO  O               O                          |   
>   85000 +-+         OOO                  O O   O                            |   
>   80000 +-+    OO O                       O  O                              |   
>         |    O                       O O                                    |   
>   75000 +-+O     O                                                          |   
>   70000 OO+ O                      O  O                                     |   
>   65000 +-+                                                                 |   
>         |                                                                   |   
>   60000 +-+                                                                 |   
>   55000 +-+     +    +     .+  .+ ++.                                       |   
>         | .+++.+ ++.+ :.+ +  ++  +   +++.+++.++  +   .+++.+     +. +        |   
>   50000 +-+           +  +                     :+ +++      + .++  + ++.+++.+|   
>   45000 +-+-----------------------------------------------------------------+   
>                                                                                 
>                                                                                 
> [*] bisect-good sample
> [O] bisect-bad  sample
> 
> 
> 
> Disclaimer:
> Results have been estimated based on internal Intel analysis and are provided
> for informational purposes only. Any difference in system hardware or software
> design or configuration may affect actual performance.
> 
> 
> Thanks,
> Rong Chen

> ---
> 
> #! jobs/will-it-scale-100.yaml
> suite: will-it-scale
> testcase: will-it-scale
> category: benchmark
> nr_task: 100%
> will-it-scale:
>   mode: process
>   test: brk1
> job_origin: "/lkp/lkp/.src-20181229-091801/allot/cyclic:p1:linux-devel:devel-hourly/lkp-knm01/will-it-scale-100.yaml"
> 
> #! queue options
> queue: bisect
> testbox: lkp-knm01
> tbox_group: lkp-knm01
> submit_id: 5c2b509f0b9a935849d3d1a6
> job_file: "/lkp/jobs/scheduled/lkp-knm01/will-it-scale-performance-process-100%-brk1-debian-x86_64-2018-04-03.cgz-7e28c5af4ef6b539334aa5de40feca0c041c94df-20190101-22601-1ncxbj9-0.yaml"
> id: 3b65780fc5b8920f2e29ad867cd42a75eee7bd74
> queuer_version: "/lkp/lkp/.src-20181229-164014"
> 
> #! hosts/lkp-knm01
> 
> #! include/category/benchmark
> kmsg: 
> boot-time: 
> iostat: 
> heartbeat: 
> vmstat: 
> numa-numastat: 
> numa-vmstat: 
> numa-meminfo: 
> proc-vmstat: 
> proc-stat: 
> meminfo: 
> slabinfo: 
> interrupts: 
> lock_stat: 
> latency_stats: 
> softirqs: 
> bdi_dev_mapping: 
> diskstats: 
> nfsstat: 
> cpuidle: 
> cpufreq-stats: 
> turbostat: 
> sched_debug: 
> perf-stat: 
> mpstat: 
> perf-profile: 
> 
> #! include/category/ALL
> cpufreq_governor: performance
> 
> #! include/queue/cyclic
> commit: 7e28c5af4ef6b539334aa5de40feca0c041c94df
> 
> #! default params
> kconfig: x86_64-rhel-7.2
> compiler: gcc-7
> rootfs: debian-x86_64-2018-04-03.cgz
> enqueue_time: 2019-01-01 19:35:59.961441013 +08:00
> _id: 5c2b509f0b9a935849d3d1a6
> _rt: "/result/will-it-scale/performance-process-100%-brk1/lkp-knm01/debian-x86_64-2018-04-03.cgz/x86_64-rhel-7.2/gcc-7/7e28c5af4ef6b539334aa5de40feca0c041c94df"
> 
> #! schedule options
> user: lkp
> head_commit: ab4756ff8ee9144cb7045eff9043ed2c6522e3b1
> base_commit: 8fe28cb58bcb235034b64cbbb7550a8a43fd88be
> branch: linux-devel/devel-hourly-2018123112
> result_root: "/result/will-it-scale/performance-process-100%-brk1/lkp-knm01/debian-x86_64-2018-04-03.cgz/x86_64-rhel-7.2/gcc-7/7e28c5af4ef6b539334aa5de40feca0c041c94df/0"
> scheduler_version: "/lkp/lkp/.src-20181229-164014"
> LKP_SERVER: inn
> max_uptime: 1500
> initrd: "/osimage/debian/debian-x86_64-2018-04-03.cgz"
> bootloader_append:
> - root=/dev/ram0
> - user=lkp
> - job=/lkp/jobs/scheduled/lkp-knm01/will-it-scale-performance-process-100%-brk1-debian-x86_64-2018-04-03.cgz-7e28c5af4ef6b539334aa5de40feca0c041c94df-20190101-22601-1ncxbj9-0.yaml
> - ARCH=x86_64
> - kconfig=x86_64-rhel-7.2
> - branch=linux-devel/devel-hourly-2018123112
> - commit=7e28c5af4ef6b539334aa5de40feca0c041c94df
> - BOOT_IMAGE=/pkg/linux/x86_64-rhel-7.2/gcc-7/7e28c5af4ef6b539334aa5de40feca0c041c94df/vmlinuz-4.19.0-rc1-00100-g7e28c5a
> - max_uptime=1500
> - RESULT_ROOT=/result/will-it-scale/performance-process-100%-brk1/lkp-knm01/debian-x86_64-2018-04-03.cgz/x86_64-rhel-7.2/gcc-7/7e28c5af4ef6b539334aa5de40feca0c041c94df/0
> - LKP_SERVER=inn
> - debug
> - apic=debug
> - sysrq_always_enabled
> - rcupdate.rcu_cpu_stall_timeout=100
> - net.ifnames=0
> - printk.devkmsg=on
> - panic=-1
> - softlockup_panic=1
> - nmi_watchdog=panic
> - oops=panic
> - load_ramdisk=2
> - prompt_ramdisk=0
> - drbd.minor_count=8
> - systemd.log_level=err
> - ignore_loglevel
> - console=tty0
> - earlyprintk=ttyS0,115200
> - console=ttyS0,115200
> - vga=normal
> - rw
> modules_initrd: "/pkg/linux/x86_64-rhel-7.2/gcc-7/7e28c5af4ef6b539334aa5de40feca0c041c94df/modules.cgz"
> bm_initrd: "/osimage/deps/debian-x86_64-2018-04-03.cgz/run-ipconfig_2018-04-03.cgz,/osimage/deps/debian-x86_64-2018-04-03.cgz/lkp_2018-04-03.cgz,/osimage/deps/debian-x86_64-2018-04-03.cgz/rsync-rootfs_2018-04-03.cgz,/osimage/deps/debian-x86_64-2018-04-03.cgz/will-it-scale_2018-05-17.cgz,/osimage/pkg/debian-x86_64-2018-04-03.cgz/will-it-scale-x86_64-decad85_2018-06-07.cgz,/osimage/deps/debian-x86_64-2018-04-03.cgz/mpstat_2018-06-19.cgz,/osimage/deps/debian-x86_64-2018-04-03.cgz/turbostat_2018-05-17.cgz,/osimage/pkg/debian-x86_64-2018-04-03.cgz/turbostat-x86_64-d5256b2_2018-05-18.cgz,/osimage/deps/debian-x86_64-2018-04-03.cgz/perf_2019-01-01.cgz,/osimage/pkg/debian-x86_64-2018-04-03.cgz/perf-x86_64-e1ef035d272e_2019-01-01.cgz,/osimage/deps/debian-x86_64-2018-04-03.cgz/hw_2016-11-15.cgz"
> lkp_initrd: "/lkp/lkp/lkp-x86_64.cgz"
> site: inn
> 
> #! /lkp/lkp/.src-20181229-164014/include/site/inn
> LKP_CGI_PORT: 80
> LKP_CIFS_PORT: 139
> oom-killer: 
> watchdog: 
> 
> #! runtime status
> repeat_to: 2
> model: Knights Mill
> nr_node: 1
> nr_cpu: 288
> memory: 80G
> hdd_partitions: 
> swap_partitions: LABEL=SWAP
> rootfs_partition: LABEL=LKP-ROOTFS
> 
> #! user overrides
> kernel: "/pkg/linux/x86_64-rhel-7.2/gcc-7/7e28c5af4ef6b539334aa5de40feca0c041c94df/vmlinuz-4.19.0-rc1-00100-g7e28c5a"
> dequeue_time: 2019-01-01 20:16:40.604583551 +08:00
> job_state: finished
> loadavg: 60.12 134.31 72.59 1/1845 11197
> start_time: '1546345124'
> end_time: '1546345447'
> version: "/lkp/lkp/.src-20181229-164014"

> 
> for cpu_dir in /sys/devices/system/cpu/cpu[0-9]*
> do
> 	online_file="$cpu_dir"/online
> 	[ -f "$online_file" ] && [ "$(cat "$online_file")" -eq 0 ] && continue
> 
> 	file="$cpu_dir"/cpufreq/scaling_governor
> 	[ -f "$file" ] && echo "performance" > "$file"
> done
> 
>  "python2" "./runtest.py" "brk1" "295" "process" "288"

Powered by blists - more mailing lists