lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [thread-next>] [day] [month] [year] [list]
Message-ID: <20190228045015.GB10770@shao2-debian>
Date:   Thu, 28 Feb 2019 12:50:15 +0800
From:   kernel test robot <rong.a.chen@...el.com>
To:     "Paul E. McKenney" <paulmck@...ux.vnet.ibm.com>
Cc:     LKML <linux-kernel@...r.kernel.org>,
        Linus Torvalds <torvalds@...ux-foundation.org>, lkp@...org
Subject: [LKP] [rcu]  7e28c5af4e:  will-it-scale.per_process_ops 84.7%
 improvement

Greeting,

FYI, we noticed a 84.7% improvement of will-it-scale.per_process_ops due to commit:


commit: 7e28c5af4ef6b539334aa5de40feca0c041c94df ("rcu: Eliminate ->rcu_qs_ctr from the rcu_dynticks structure")
https://git.kernel.org/cgit/linux/kernel/git/torvalds/linux.git master

in testcase: will-it-scale
on test machine: 288 threads Intel(R) Xeon Phi(TM) CPU 7295 @ 1.50GHz with 80G memory
with following parameters:

	nr_task: 100%
	mode: process
	test: brk1
	cpufreq_governor: performance

test-description: Will It Scale takes a testcase and runs it from 1 through to n parallel copies to see if the testcase will scale. It builds both a process and threads based test in order to see any differences between the two.
test-url: https://github.com/antonblanchard/will-it-scale



Details are as below:
-------------------------------------------------------------------------------------------------->


To reproduce:

        git clone https://github.com/intel/lkp-tests.git
        cd lkp-tests
        bin/lkp install job.yaml  # job file is attached in this email
        bin/lkp run     job.yaml

=========================================================================================
compiler/cpufreq_governor/kconfig/mode/nr_task/rootfs/tbox_group/test/testcase:
  gcc-7/performance/x86_64-rhel-7.2/process/100%/debian-x86_64-2018-04-03.cgz/lkp-knm01/brk1/will-it-scale

commit: 
  c5bacd9417 ("rcu: Motivate Tiny RCU forward progress")
  7e28c5af4e ("rcu: Eliminate ->rcu_qs_ctr from the rcu_dynticks structure")

c5bacd94173ec49d 7e28c5af4ef6b539334aa5de40 
---------------- -------------------------- 
         %stddev     %change         %stddev
             \          |                \  
     48605           +84.7%      89756 ±  4%  will-it-scale.per_process_ops
  13998509           +84.7%   25850040 ±  4%  will-it-scale.workload
      0.01 ± 63%      -0.0        0.00 ±131%  mpstat.cpu.soft%
     11.86 ±  3%      -1.2       10.65 ±  4%  mpstat.cpu.usr%
    662910           +15.8%     767829 ±  3%  numa-numastat.node0.local_node
    662889           +15.8%     767801 ±  3%  numa-numastat.node0.numa_hit
      2149 ±  4%     -16.9%       1785 ±  2%  vmstat.system.cs
    727237 ±  3%     -14.9%     618668        vmstat.system.in
    284156 ±  4%     +32.5%     376625        numa-meminfo.node0.Active
    283908 ±  4%     +32.6%     376374        numa-meminfo.node0.Active(anon)
    110629 ± 15%     -77.7%      24625 ± 18%  numa-meminfo.node0.Inactive
    110393 ± 15%     -77.9%      24400 ± 19%  numa-meminfo.node0.Inactive(anon)
    284255 ±  4%     +32.5%     376666        meminfo.Active
    284007 ±  4%     +32.5%     376415        meminfo.Active(anon)
    110855 ± 15%     -77.8%      24620 ± 18%  meminfo.Inactive
    110619 ± 15%     -77.9%      24395 ± 19%  meminfo.Inactive(anon)
      8.00 ±173%  +11303.1%     912.25 ± 87%  meminfo.Mlocked
     11072           -15.7%       9338 ±  3%  meminfo.max_used_kB
     71004 ±  4%     +32.6%      94124        numa-vmstat.node0.nr_active_anon
     27812 ± 15%     -78.1%       6100 ± 19%  numa-vmstat.node0.nr_inactive_anon
      1.00 ±173%  +13100.0%     132.00 ± 87%  numa-vmstat.node0.nr_mlock
     71002 ±  4%     +32.6%      94122        numa-vmstat.node0.nr_zone_active_anon
     27811 ± 15%     -78.1%       6100 ± 19%  numa-vmstat.node0.nr_zone_inactive_anon
      1.00 ±173%   +9450.0%      95.50 ± 87%  numa-vmstat.node1.nr_mlock
     70939 ±  4%     +32.8%      94186        proc-vmstat.nr_active_anon
     27523 ± 15%     -77.8%       6101 ± 19%  proc-vmstat.nr_inactive_anon
      2.25 ±173%  +10044.4%     228.25 ± 86%  proc-vmstat.nr_mlock
     70939 ±  4%     +32.8%      94186        proc-vmstat.nr_zone_active_anon
     27523 ± 15%     -77.8%       6101 ± 19%  proc-vmstat.nr_zone_inactive_anon
    680060           +15.8%     787707 ±  2%  proc-vmstat.numa_hit
    680060           +15.8%     787707 ±  2%  proc-vmstat.numa_local
    140509 ± 10%     -45.3%      76820 ± 24%  proc-vmstat.numa_pte_updates
    895.50 ± 30%    +480.0%       5193 ± 36%  proc-vmstat.pgactivate
    774283           +13.9%     882060 ±  3%  proc-vmstat.pgalloc_normal
    653457           +21.9%     796258 ±  3%  proc-vmstat.pgfault
    610954 ±  2%     +32.4%     808662 ±  3%  proc-vmstat.pgfree
      5299 ±  6%    +106.4%      10939 ± 36%  softirqs.CPU10.RCU
    106293 ±  5%     +28.2%     136274 ± 15%  softirqs.CPU126.TIMER
    108234 ±  6%     +26.3%     136696 ± 15%  softirqs.CPU127.TIMER
    109747 ±  7%     +26.4%     138696 ± 11%  softirqs.CPU135.TIMER
    111732 ±  8%     +24.4%     138967 ± 11%  softirqs.CPU138.TIMER
    107031 ±  8%     +26.7%     135558 ± 14%  softirqs.CPU148.TIMER
    100841           +23.1%     124155 ± 15%  softirqs.CPU15.TIMER
    104640 ±  4%     +27.8%     133719 ± 16%  softirqs.CPU172.TIMER
    103142 ±  3%     +30.9%     135023 ± 13%  softirqs.CPU190.TIMER
    106698 ± 10%     +27.5%     136023 ± 13%  softirqs.CPU197.TIMER
    110268 ±  2%     +22.5%     135052 ± 12%  softirqs.CPU205.TIMER
    106111 ±  8%     +23.9%     131436 ± 14%  softirqs.CPU206.TIMER
    106894 ±  6%     +28.7%     137574 ± 14%  softirqs.CPU213.TIMER
    108418 ±  8%     +28.3%     139100 ± 12%  softirqs.CPU214.TIMER
    100170 ± 10%     +32.8%     132978 ± 15%  softirqs.CPU215.TIMER
    100836 ±  2%     +30.3%     131406 ± 17%  softirqs.CPU217.TIMER
    112497 ±  8%     +22.9%     138291 ± 13%  softirqs.CPU219.TIMER
    108669 ± 10%     +22.1%     132720 ± 15%  softirqs.CPU220.TIMER
    108584 ±  8%     +25.5%     136230 ± 13%  softirqs.CPU227.TIMER
    106251 ±  6%     +28.6%     136686 ± 11%  softirqs.CPU236.TIMER
    105296 ±  7%     +31.3%     138251 ± 12%  softirqs.CPU238.TIMER
      3915 ±  8%     +77.4%       6945 ± 30%  softirqs.CPU249.RCU
    101465 ±  2%     +26.5%     128388 ± 19%  softirqs.CPU271.TIMER
    104242 ±  8%     +23.7%     128907 ± 17%  softirqs.CPU70.TIMER
    107020 ±  7%     +26.9%     135845 ± 13%  softirqs.CPU92.TIMER
 9.659e+09 ± 10%     +16.3%  1.124e+10 ±  3%  perf-stat.i.branch-instructions
      6.69            +1.0        7.69 ±  2%  perf-stat.i.branch-miss-rate%
 6.443e+08 ± 10%     +33.3%   8.59e+08 ±  2%  perf-stat.i.branch-misses
      4.10 ±  6%      +0.9        5.04 ± 10%  perf-stat.i.cache-miss-rate%
  67464898 ± 10%     +25.2%   84491315 ±  4%  perf-stat.i.cache-misses
      2820 ± 13%     -32.4%       1906 ±  3%  perf-stat.i.context-switches
     15.34           -37.0%       9.67 ±  3%  perf-stat.i.cpi
 7.067e+11 ± 11%     -30.2%  4.932e+11 ±  4%  perf-stat.i.cpu-cycles
     71.93 ± 10%     -41.9%      41.79 ±  9%  perf-stat.i.cpu-migrations
      2.36            -0.0        2.33        perf-stat.i.iTLB-load-miss-rate%
 1.106e+09 ± 10%     +15.1%  1.273e+09 ±  2%  perf-stat.i.iTLB-load-misses
 4.569e+10 ± 10%     +17.0%  5.345e+10 ±  3%  perf-stat.i.iTLB-loads
 4.621e+10 ± 10%     +16.1%  5.364e+10 ±  3%  perf-stat.i.instructions
      0.07           +64.1%       0.11 ±  3%  perf-stat.i.ipc
      3524 ± 11%     -18.3%       2878        perf-stat.i.minor-faults
    993359 ± 11%     -32.0%     675215 ±  2%  perf-stat.i.msec
      3529 ± 11%     -18.5%       2878        perf-stat.i.page-faults
      6.67            +1.0        7.65        perf-stat.overall.branch-miss-rate%
      3.91            +0.6        4.54 ±  3%  perf-stat.overall.cache-miss-rate%
     15.28           -39.8%       9.20 ±  3%  perf-stat.overall.cpi
      2.36            -0.0        2.33        perf-stat.overall.iTLB-load-miss-rate%
     41.78            +0.9%      42.13        perf-stat.overall.instructions-per-iTLB-miss
      0.07           +66.4%       0.11 ±  3%  perf-stat.overall.ipc
    580565            -5.5%     548645        perf-stat.overall.path-length
 1.699e+12           +74.9%  2.971e+12 ±  4%  perf-stat.total.branch-instructions
 1.133e+11          +100.6%  2.272e+11 ±  5%  perf-stat.total.branch-misses
 1.186e+10           +88.0%   2.23e+10        perf-stat.total.cache-misses
 3.032e+11           +62.2%  4.918e+11 ±  2%  perf-stat.total.cache-references
 1.242e+14            +4.8%  1.302e+14        perf-stat.total.cpu-cycles
     12654           -13.0%      11014 ±  5%  perf-stat.total.cpu-migrations
 1.945e+11           +73.0%  3.365e+11 ±  3%  perf-stat.total.iTLB-load-misses
 8.038e+12           +75.8%  1.413e+13 ±  4%  perf-stat.total.iTLB-loads
 8.127e+12           +74.5%  1.418e+13 ±  4%  perf-stat.total.instructions
    619183           +22.9%     760791 ±  3%  perf-stat.total.minor-faults
    620031           +22.7%     760750 ±  3%  perf-stat.total.page-faults
  23472574 ± 28%     -50.0%   11736977 ± 60%  sched_debug.cfs_rq:/.MIN_vruntime.max
   1380731 ± 28%     -46.9%     733243 ± 66%  sched_debug.cfs_rq:/.MIN_vruntime.stddev
    109109           +17.1%     127734        sched_debug.cfs_rq:/.exec_clock.avg
    115740           +12.4%     130080        sched_debug.cfs_rq:/.exec_clock.max
      4252 ±  4%     -50.4%       2110 ± 36%  sched_debug.cfs_rq:/.exec_clock.stddev
      6243 ± 12%     -14.0%       5369 ± 15%  sched_debug.cfs_rq:/.load.avg
    887713 ± 24%     -30.6%     616115 ± 43%  sched_debug.cfs_rq:/.load.max
     53788 ± 23%     -30.4%      37427 ± 40%  sched_debug.cfs_rq:/.load.stddev
  23472574 ± 28%     -50.0%   11736977 ± 60%  sched_debug.cfs_rq:/.max_vruntime.max
   1380731 ± 28%     -46.9%     733243 ± 66%  sched_debug.cfs_rq:/.max_vruntime.stddev
  36862110           +17.8%   43411425        sched_debug.cfs_rq:/.min_vruntime.avg
  39448691           +12.5%   44387703        sched_debug.cfs_rq:/.min_vruntime.max
   1668979 ±  3%     -45.0%     917652 ± 30%  sched_debug.cfs_rq:/.min_vruntime.stddev
      1.38 ±  9%      -9.1%       1.25 ±  6%  sched_debug.cfs_rq:/.nr_running.max
      0.56 ± 19%     +42.2%       0.80        sched_debug.cfs_rq:/.nr_running.min
      0.07 ± 13%     -32.7%       0.05 ± 15%  sched_debug.cfs_rq:/.nr_running.stddev
      2.25 ±  3%     -51.5%       1.09 ±  2%  sched_debug.cfs_rq:/.nr_spread_over.avg
      6235 ± 12%     -14.0%       5361 ± 16%  sched_debug.cfs_rq:/.runnable_weight.avg
    887713 ± 24%     -30.6%     616088 ± 43%  sched_debug.cfs_rq:/.runnable_weight.max
     53791 ± 23%     -30.4%      37423 ± 40%  sched_debug.cfs_rq:/.runnable_weight.stddev
    513480 ± 46%     -68.1%     163689 ± 16%  sched_debug.cfs_rq:/.spread0.avg
    992070 ± 25%     -55.6%     440313 ±  8%  sched_debug.cfs_rq:/.spread0.max
    271.31 ± 25%     +54.1%     418.15 ± 23%  sched_debug.cfs_rq:/.util_avg.min
     71.69 ±  6%     -14.9%      61.03 ±  7%  sched_debug.cfs_rq:/.util_avg.stddev
    638.86           +18.5%     756.85 ±  3%  sched_debug.cfs_rq:/.util_est_enqueued.avg
    131.07 ±  2%     -17.0%     108.78 ±  5%  sched_debug.cfs_rq:/.util_est_enqueued.stddev
   1617727 ±  4%     -24.5%    1222141 ± 11%  sched_debug.cpu.avg_idle.avg
    167800 ±  2%     +13.3%     190124 ±  4%  sched_debug.cpu.clock.avg
    174380           +10.3%     192313 ±  4%  sched_debug.cpu.clock.max
    161230 ±  2%     +16.5%     187888 ±  4%  sched_debug.cpu.clock.min
      3797 ±  4%     -66.4%       1276 ±  8%  sched_debug.cpu.clock.stddev
    167800 ±  2%     +13.3%     190124 ±  4%  sched_debug.cpu.clock_task.avg
    174380           +10.3%     192313 ±  4%  sched_debug.cpu.clock_task.max
    161230 ±  2%     +16.5%     187888 ±  4%  sched_debug.cpu.clock_task.min
      3797 ±  4%     -66.4%       1276 ±  8%  sched_debug.cpu.clock_task.stddev
      5416 ±  2%     +15.6%       6262 ±  5%  sched_debug.cpu.curr->pid.max
      2045 ± 18%     +20.6%       2466 ±  3%  sched_debug.cpu.curr->pid.min
    892728 ±  7%     -37.5%     557989 ±  5%  sched_debug.cpu.max_idle_balance_cost.avg
   4113479 ± 18%     -51.6%    1989769 ± 36%  sched_debug.cpu.max_idle_balance_cost.max
    461057 ± 17%     -67.6%     149491 ± 36%  sched_debug.cpu.max_idle_balance_cost.stddev
      0.00 ±  4%     -65.8%       0.00 ± 23%  sched_debug.cpu.next_balance.stddev
    117031           +16.4%     136279        sched_debug.cpu.nr_load_updates.avg
    139472 ±  3%     +16.0%     161763 ±  4%  sched_debug.cpu.nr_load_updates.max
    109218           +22.1%     133390        sched_debug.cpu.nr_load_updates.min
      3910 ±  3%     -46.0%       2111 ± 19%  sched_debug.cpu.nr_load_updates.stddev
      0.16 ±  4%     -19.4%       0.13 ±  3%  sched_debug.cpu.nr_running.stddev
      1108           +12.5%       1247 ±  2%  sched_debug.cpu.nr_switches.avg
     26767 ± 16%     +60.1%      42860 ± 15%  sched_debug.cpu.nr_switches.max
      2342 ±  4%     +34.2%       3144 ±  7%  sched_debug.cpu.nr_switches.stddev
     29066 ± 11%     +83.5%      53341 ± 36%  sched_debug.cpu.sched_count.max
      2473 ±  2%     +61.3%       3988 ± 33%  sched_debug.cpu.sched_count.stddev
    275.25 ±  3%     +33.5%     367.54        sched_debug.cpu.ttwu_count.avg
     10910 ± 17%     +67.3%      18254 ± 13%  sched_debug.cpu.ttwu_count.max
     50.31 ±  2%     +19.9%      60.30 ±  2%  sched_debug.cpu.ttwu_count.min
    832.06 ± 10%     +46.1%       1215 ±  7%  sched_debug.cpu.ttwu_count.stddev
    176.33           +42.8%     251.81 ±  2%  sched_debug.cpu.ttwu_local.avg
     10535 ± 18%     +69.7%      17877 ± 14%  sched_debug.cpu.ttwu_local.max
     31.56           +15.2%      36.35        sched_debug.cpu.ttwu_local.min
    729.04 ± 13%     +57.2%       1146 ±  9%  sched_debug.cpu.ttwu_local.stddev
    161164 ±  2%     +16.6%     187864 ±  4%  sched_debug.cpu_clk
    160513 ±  2%     +16.6%     187213 ±  4%  sched_debug.ktime
      1.31 ± 24%     +39.1%       1.82        sched_debug.rt_rq:/.rt_runtime.stddev
    161891 ±  2%     +16.3%     188213 ±  4%  sched_debug.sched_clk
    745.75 ± 61%     -87.9%      90.25 ±173%  interrupts.36:IR-PCI-MSI.2621442-edge.eth1-TxRx-1
    961.75 ± 48%     -75.5%     235.25 ±173%  interrupts.37:IR-PCI-MSI.2621443-edge.eth1-TxRx-2
      4649 ± 16%     -25.2%       3475 ±  3%  interrupts.CPU100.NMI:Non-maskable_interrupts
      4649 ± 16%     -25.2%       3475 ±  3%  interrupts.CPU100.PMI:Performance_monitoring_interrupts
     30.50 ± 17%    +277.9%     115.25 ± 49%  interrupts.CPU109.RES:Rescheduling_interrupts
    552027 ±  5%     +12.2%     619466 ±  2%  interrupts.CPU110.LOC:Local_timer_interrupts
    274.50 ± 59%     -44.9%     151.25 ± 92%  interrupts.CPU116.RES:Rescheduling_interrupts
     37.50 ± 47%    +323.3%     158.75 ± 43%  interrupts.CPU121.RES:Rescheduling_interrupts
      7275 ± 12%     -45.9%       3937 ± 25%  interrupts.CPU125.NMI:Non-maskable_interrupts
      7275 ± 12%     -45.9%       3937 ± 25%  interrupts.CPU125.PMI:Performance_monitoring_interrupts
      5151 ± 34%     -32.6%       3470 ±  6%  interrupts.CPU140.NMI:Non-maskable_interrupts
      5151 ± 34%     -32.6%       3470 ±  6%  interrupts.CPU140.PMI:Performance_monitoring_interrupts
      5589 ± 28%     -26.0%       4134 ± 32%  interrupts.CPU153.NMI:Non-maskable_interrupts
      5589 ± 28%     -26.0%       4134 ± 32%  interrupts.CPU153.PMI:Performance_monitoring_interrupts
    745.75 ± 61%     -87.9%      90.25 ±173%  interrupts.CPU16.36:IR-PCI-MSI.2621442-edge.eth1-TxRx-1
    961.75 ± 48%     -75.5%     235.25 ±173%  interrupts.CPU17.37:IR-PCI-MSI.2621443-edge.eth1-TxRx-2
      6606 ± 25%     -38.6%       4057 ± 27%  interrupts.CPU171.NMI:Non-maskable_interrupts
      6606 ± 25%     -38.6%       4057 ± 27%  interrupts.CPU171.PMI:Performance_monitoring_interrupts
      5916 ± 21%     -28.9%       4205 ± 28%  interrupts.CPU172.NMI:Non-maskable_interrupts
      5916 ± 21%     -28.9%       4205 ± 28%  interrupts.CPU172.PMI:Performance_monitoring_interrupts
     85.50 ± 36%     -79.2%      17.75 ± 72%  interrupts.CPU176.RES:Rescheduling_interrupts
     74.00 ± 79%     -79.1%      15.50 ± 27%  interrupts.CPU177.RES:Rescheduling_interrupts
      6513 ± 21%     -36.4%       4139 ± 24%  interrupts.CPU180.NMI:Non-maskable_interrupts
      6513 ± 21%     -36.4%       4139 ± 24%  interrupts.CPU180.PMI:Performance_monitoring_interrupts
     17.50 ± 32%    +371.4%      82.50 ±101%  interrupts.CPU181.RES:Rescheduling_interrupts
     41.00 ±101%    +232.9%     136.50 ± 79%  interrupts.CPU183.RES:Rescheduling_interrupts
      5998 ± 31%     -29.3%       4238 ± 18%  interrupts.CPU192.NMI:Non-maskable_interrupts
      5998 ± 31%     -29.3%       4238 ± 18%  interrupts.CPU192.PMI:Performance_monitoring_interrupts
      5812 ± 21%     -40.6%       3450 ±  4%  interrupts.CPU198.NMI:Non-maskable_interrupts
      5812 ± 21%     -40.6%       3450 ±  4%  interrupts.CPU198.PMI:Performance_monitoring_interrupts
      5605 ± 22%     -27.0%       4093 ± 26%  interrupts.CPU200.NMI:Non-maskable_interrupts
      5605 ± 22%     -27.0%       4093 ± 26%  interrupts.CPU200.PMI:Performance_monitoring_interrupts
      5528 ± 27%     -38.1%       3423 ±  2%  interrupts.CPU212.NMI:Non-maskable_interrupts
      5528 ± 27%     -38.1%       3423 ±  2%  interrupts.CPU212.PMI:Performance_monitoring_interrupts
      6447 ± 27%     -31.2%       4433 ± 39%  interrupts.CPU216.NMI:Non-maskable_interrupts
      6447 ± 27%     -31.2%       4433 ± 39%  interrupts.CPU216.PMI:Performance_monitoring_interrupts
    561218 ±  5%     +10.4%     619618 ±  2%  interrupts.CPU220.LOC:Local_timer_interrupts
    201.00 ±134%     -85.0%      30.25 ± 53%  interrupts.CPU228.RES:Rescheduling_interrupts
      5757 ± 22%     -35.4%       3719 ±  4%  interrupts.CPU229.NMI:Non-maskable_interrupts
      5757 ± 22%     -35.4%       3719 ±  4%  interrupts.CPU229.PMI:Performance_monitoring_interrupts
    165.75 ± 46%     -70.6%      48.75 ±103%  interrupts.CPU230.RES:Rescheduling_interrupts
     23.75 ± 38%    +346.3%     106.00 ± 87%  interrupts.CPU236.RES:Rescheduling_interrupts
      4576 ± 15%     -26.0%       3385 ±  3%  interrupts.CPU248.NMI:Non-maskable_interrupts
      4576 ± 15%     -26.0%       3385 ±  3%  interrupts.CPU248.PMI:Performance_monitoring_interrupts
    606.75 ±140%     -92.7%      44.00 ± 69%  interrupts.CPU252.RES:Rescheduling_interrupts
     80.00 ± 62%     -67.5%      26.00 ± 30%  interrupts.CPU256.RES:Rescheduling_interrupts
      5696 ± 30%     -26.9%       4164 ± 26%  interrupts.CPU259.NMI:Non-maskable_interrupts
      5696 ± 30%     -26.9%       4164 ± 26%  interrupts.CPU259.PMI:Performance_monitoring_interrupts
      1796 ± 99%     -76.2%     428.00 ± 45%  interrupts.CPU26.RES:Rescheduling_interrupts
    109.50 ± 70%     -66.0%      37.25 ± 55%  interrupts.CPU266.RES:Rescheduling_interrupts
    143.00 ± 38%     -69.1%      44.25 ± 90%  interrupts.CPU271.RES:Rescheduling_interrupts
      5526 ± 29%     -34.5%       3619        interrupts.CPU278.NMI:Non-maskable_interrupts
      5526 ± 29%     -34.5%       3619        interrupts.CPU278.PMI:Performance_monitoring_interrupts
     94.75 ± 42%     -55.1%      42.50 ± 64%  interrupts.CPU278.RES:Rescheduling_interrupts
      6671 ± 19%     -31.4%       4579 ± 31%  interrupts.CPU281.NMI:Non-maskable_interrupts
      6671 ± 19%     -31.4%       4579 ± 31%  interrupts.CPU281.PMI:Performance_monitoring_interrupts
      7140 ± 25%     -45.1%       3917 ±  2%  interrupts.CPU37.NMI:Non-maskable_interrupts
      7140 ± 25%     -45.1%       3917 ±  2%  interrupts.CPU37.PMI:Performance_monitoring_interrupts
      6176 ± 29%     -37.9%       3834 ±  3%  interrupts.CPU38.NMI:Non-maskable_interrupts
      6176 ± 29%     -37.9%       3834 ±  3%  interrupts.CPU38.PMI:Performance_monitoring_interrupts
    968.50 ±109%     -75.0%     241.75 ± 42%  interrupts.CPU38.RES:Rescheduling_interrupts
      6057 ± 24%     -28.2%       4352 ± 23%  interrupts.CPU41.NMI:Non-maskable_interrupts
      6057 ± 24%     -28.2%       4352 ± 23%  interrupts.CPU41.PMI:Performance_monitoring_interrupts
    120.00 ± 99%    +376.2%     571.50 ± 85%  interrupts.CPU47.RES:Rescheduling_interrupts
    177.75 ± 31%    +247.4%     617.50 ± 88%  interrupts.CPU48.RES:Rescheduling_interrupts
    110.00 ± 83%    +116.6%     238.25 ± 26%  interrupts.CPU49.RES:Rescheduling_interrupts
      6925 ± 22%     -45.4%       3781 ±  3%  interrupts.CPU50.NMI:Non-maskable_interrupts
      6925 ± 22%     -45.4%       3781 ±  3%  interrupts.CPU50.PMI:Performance_monitoring_interrupts
    180.25 ± 97%    +685.2%       1415 ±115%  interrupts.CPU51.RES:Rescheduling_interrupts
    182.50 ± 53%   +1719.2%       3320 ± 62%  interrupts.CPU52.RES:Rescheduling_interrupts
      5043 ± 18%     -26.0%       3730 ±  5%  interrupts.CPU57.NMI:Non-maskable_interrupts
      5043 ± 18%     -26.0%       3730 ±  5%  interrupts.CPU57.PMI:Performance_monitoring_interrupts
      3655 ±  9%     +33.0%       4863 ± 21%  interrupts.CPU60.NMI:Non-maskable_interrupts
      3655 ±  9%     +33.0%       4863 ± 21%  interrupts.CPU60.PMI:Performance_monitoring_interrupts
      6130 ± 24%     -28.1%       4410 ± 21%  interrupts.CPU64.NMI:Non-maskable_interrupts
      6130 ± 24%     -28.1%       4410 ± 21%  interrupts.CPU64.PMI:Performance_monitoring_interrupts
      5811 ± 25%     -36.1%       3713        interrupts.CPU76.NMI:Non-maskable_interrupts
      5811 ± 25%     -36.1%       3713        interrupts.CPU76.PMI:Performance_monitoring_interrupts
      6036 ± 28%     -39.4%       3660 ±  5%  interrupts.CPU80.NMI:Non-maskable_interrupts
      6036 ± 28%     -39.4%       3660 ±  5%  interrupts.CPU80.PMI:Performance_monitoring_interrupts
      6494 ± 25%     -43.9%       3640 ±  5%  interrupts.CPU82.NMI:Non-maskable_interrupts
      6494 ± 25%     -43.9%       3640 ±  5%  interrupts.CPU82.PMI:Performance_monitoring_interrupts
      6881 ± 13%     -48.8%       3522 ±  3%  interrupts.CPU84.NMI:Non-maskable_interrupts
      6881 ± 13%     -48.8%       3522 ±  3%  interrupts.CPU84.PMI:Performance_monitoring_interrupts
    110.75 ± 19%    +119.4%     243.00 ± 19%  interrupts.CPU9.RES:Rescheduling_interrupts


                                                                                
                            will-it-scale.per_process_ops                       
                                                                                
  95000 +-+-----------------------------------O-----------------------------+   
  90000 +-+              O  O   O O               O                         |   
        |               O O  OO  O               O                          |   
  85000 +-+         OOO                  O O   O                            |   
  80000 +-+    OO O                       O  O                              |   
        |    O                       O O                                    |   
  75000 +-+O     O                                                          |   
  70000 OO+ O                      O  O                                     |   
  65000 +-+                                                                 |   
        |                                                                   |   
  60000 +-+                                                                 |   
  55000 +-+     +    +     .+  .+ ++.                                       |   
        | .+++.+ ++.+ :.+ +  ++  +   +++.+++.++  +   .+++.+     +. +        |   
  50000 +-+           +  +                     :+ +++      + .++  + ++.+++.+|   
  45000 +-+-----------------------------------------------------------------+   
                                                                                
                                                                                
[*] bisect-good sample
[O] bisect-bad  sample



Disclaimer:
Results have been estimated based on internal Intel analysis and are provided
for informational purposes only. Any difference in system hardware or software
design or configuration may affect actual performance.


Thanks,
Rong Chen

View attachment "job.yaml" of type "text/plain" (4615 bytes)

View attachment "reproduce" of type "text/plain" (309 bytes)

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ