[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <CAKfTPtB3jQWd52FTyKNk5w8mmoDnu+jwYgkFBOiOKjb_BjxqTw@mail.gmail.com>
Date: Thu, 14 May 2020 19:09:35 +0200
From: Vincent Guittot <vincent.guittot@...aro.org>
To: kernel test robot <oliver.sang@...el.com>
Cc: Ingo Molnar <mingo@...nel.org>, Ben Segall <bsegall@...gle.com>,
Dietmar Eggemann <dietmar.eggemann@....com>,
Juri Lelli <juri.lelli@...hat.com>,
Linus Torvalds <torvalds@...ux-foundation.org>,
Mel Gorman <mgorman@...e.de>, Mike Galbraith <efault@....de>,
Peter Zijlstra <peterz@...radead.org>,
Steven Rostedt <rostedt@...dmis.org>,
Thomas Gleixner <tglx@...utronix.de>,
LKML <linux-kernel@...r.kernel.org>, lkp@...ts.01.org,
OTC LSE PnP <otc.lse.pnp@...el.com>
Subject: Re: [sched/fair] 0b0695f2b3: phoronix-test-suite.compress-gzip.0.seconds
19.8% regression
Hi Oliver,
On Thu, 14 May 2020 at 16:05, kernel test robot <oliver.sang@...el.com> wrote:
>
> Hi Vincent Guittot,
>
> Below report FYI.
> Last year, we actually reported an improvement "[sched/fair] 0b0695f2b3:
> vm-scalability.median 3.1% improvement" on link [1].
> but now we found the regression on pts.compress-gzip.
> This seems align with what showed in "[v4,00/10] sched/fair: rework the CFS
> load balance" (link [2]), where showed the reworked load balance could have
> both positive and negative effect for different test suites.
We have tried to run all possible use cases but it's impossible to
covers all so there were a possibility that one that is not covered,
would regressed.
> And also from link [3], the patch set risks regressions.
>
> We also confirmed this regression on another platform
> (Intel(R) Core(TM) i7-8700 CPU @ 3.20GHz with 8G memory),
> below is the data (lower is better).
> v5.4 4.1
> fcf0553db6f4c79387864f6e4ab4a891601f395e 4.01
> 0b0695f2b34a4afa3f6e9aa1ff0e5336d8dad912 4.89
> v5.5 5.18
> v5.6 4.62
> v5.7-rc2 4.53
> v5.7-rc3 4.59
>
> It seems there are some recovery on latest kernels, but not fully back.
> We were just wondering whether you could share some lights the further works
> on the load balance after patch set [2] which could cause the performance
> change?
> And whether you have plan to refine the load balance algorithm further?
I'm going to have a look at your regression to understand what is
going wrong and how it can be fixed
Thanks
Vincent
> thanks
>
> [1] https://lists.01.org/hyperkitty/list/lkp@lists.01.org/thread/SANC7QLYZKUNMM6O7UNR3OAQAKS5BESE/
> [2] https://lore.kernel.org/patchwork/cover/1141687/
> [3] https://www.phoronix.com/scan.php?page=news_item&px=Linux-5.5-Scheduler
>
>
>
> Below is the detail regression report FYI.
>
> Greeting,
>
> FYI, we noticed a 19.8% regression of phoronix-test-suite.compress-gzip.0.seconds due to commit:
>
>
> commit: 0b0695f2b34a4afa3f6e9aa1ff0e5336d8dad912 ("sched/fair: Rework load_balance()")
> https://git.kernel.org/cgit/linux/kernel/git/torvalds/linux.git master
>
> in testcase: phoronix-test-suite
> on test machine: 16 threads Intel(R) Xeon(R) E-2278G CPU @ 3.40GHz with 32G memory
> with following parameters:
>
> test: compress-gzip-1.2.0
> cpufreq_governor: performance
> ucode: 0xca
>
> test-description: The Phoronix Test Suite is the most comprehensive testing and benchmarking platform available that provides an extensible framework for which new tests can be easily added.
> test-url: http://www.phoronix-test-suite.com/
>
> In addition to that, the commit also has significant impact on the following tests:
>
> +------------------+-----------------------------------------------------------------------+
> | testcase: change | phoronix-test-suite: |
> | test machine | 12 threads Intel(R) Core(TM) i7-8700 CPU @ 3.20GHz with 8G memory |
> | test parameters | cpufreq_governor=performance |
> | | test=compress-gzip-1.2.0 |
> +------------------+-----------------------------------------------------------------------+
> | testcase: change | vm-scalability: vm-scalability.median 3.1% improvement |
> | test machine | 104 threads Skylake with 192G memory |
> | test parameters | cpufreq_governor=performance |
> | | runtime=300s |
> | | size=8T |
> | | test=anon-cow-seq |
> | | ucode=0x2000064 |
> +------------------+-----------------------------------------------------------------------+
> | testcase: change | stress-ng: stress-ng.fault.ops_per_sec -23.1% regression |
> | test machine | 88 threads Intel(R) Xeon(R) CPU E5-2699 v4 @ 2.20GHz with 128G memory |
> | test parameters | class=scheduler |
> | | cpufreq_governor=performance |
> | | disk=1HDD |
> | | nr_threads=100% |
> | | sc_pid_max=4194304 |
> | | testtime=1s |
> | | ucode=0xb000038 |
> +------------------+-----------------------------------------------------------------------+
> | testcase: change | stress-ng: stress-ng.schedpolicy.ops_per_sec -33.3% regression |
> | test machine | 96 threads Intel(R) Xeon(R) Gold 6252 CPU @ 2.10GHz with 192G memory |
> | test parameters | class=interrupt |
> | | cpufreq_governor=performance |
> | | disk=1HDD |
> | | nr_threads=100% |
> | | testtime=1s |
> | | ucode=0x500002c |
> +------------------+-----------------------------------------------------------------------+
> | testcase: change | stress-ng: stress-ng.schedpolicy.ops_per_sec 42.3% improvement |
> | test machine | 88 threads Intel(R) Xeon(R) CPU E5-2699 v4 @ 2.20GHz with 128G memory |
> | test parameters | class=interrupt |
> | | cpufreq_governor=performance |
> | | disk=1HDD |
> | | nr_threads=100% |
> | | testtime=30s |
> | | ucode=0xb000038 |
> +------------------+-----------------------------------------------------------------------+
> | testcase: change | stress-ng: stress-ng.schedpolicy.ops_per_sec 35.1% improvement |
> | test machine | 88 threads Intel(R) Xeon(R) CPU E5-2699 v4 @ 2.20GHz with 128G memory |
> | test parameters | class=interrupt |
> | | cpufreq_governor=performance |
> | | disk=1HDD |
> | | nr_threads=100% |
> | | testtime=1s |
> | | ucode=0xb000038 |
> +------------------+-----------------------------------------------------------------------+
> | testcase: change | stress-ng: stress-ng.ioprio.ops_per_sec -20.7% regression |
> | test machine | 96 threads Intel(R) Xeon(R) Gold 6252 CPU @ 2.10GHz with 192G memory |
> | test parameters | class=os |
> | | cpufreq_governor=performance |
> | | disk=1HDD |
> | | fs=ext4 |
> | | nr_threads=100% |
> | | testtime=1s |
> | | ucode=0x500002b |
> +------------------+-----------------------------------------------------------------------+
> | testcase: change | stress-ng: stress-ng.schedpolicy.ops_per_sec 43.0% improvement |
> | test machine | 88 threads Intel(R) Xeon(R) CPU E5-2699 v4 @ 2.20GHz with 128G memory |
> | test parameters | class=interrupt |
> | | cpufreq_governor=performance |
> | | disk=1HDD |
> | | nr_threads=100% |
> | | testtime=30s |
> | | ucode=0xb000038 |
> +------------------+-----------------------------------------------------------------------+
>
>
> If you fix the issue, kindly add following tag
> Reported-by: kernel test robot <oliver.sang@...el.com>
>
>
> Details are as below:
> -------------------------------------------------------------------------------------------------->
>
>
> To reproduce:
>
> git clone https://github.com/intel/lkp-tests.git
> cd lkp-tests
> bin/lkp install job.yaml # job file is attached in this email
> bin/lkp run job.yaml
>
> =========================================================================================
> compiler/cpufreq_governor/kconfig/rootfs/tbox_group/test/testcase/ucode:
> gcc-7/performance/x86_64-lck-7983/clear-x86_64-phoronix-30140/lkp-cfl-e1/compress-gzip-1.2.0/phoronix-test-suite/0xca
>
> commit:
> fcf0553db6 ("sched/fair: Remove meaningless imbalance calculation")
> 0b0695f2b3 ("sched/fair: Rework load_balance()")
>
> fcf0553db6f4c793 0b0695f2b34a4afa3f6e9aa1ff0
> ---------------- ---------------------------
> fail:runs %reproduction fail:runs
> | | |
> :4 4% 0:7 perf-profile.children.cycles-pp.error_entry
> %stddev %change %stddev
> \ | \
> 6.01 +19.8% 7.20 phoronix-test-suite.compress-gzip.0.seconds
> 147.57 ą 8% +25.1% 184.54 phoronix-test-suite.time.elapsed_time
> 147.57 ą 8% +25.1% 184.54 phoronix-test-suite.time.elapsed_time.max
> 52926 ą 8% -23.8% 40312 meminfo.max_used_kB
> 0.11 ą 7% -0.0 0.09 ą 3% mpstat.cpu.all.soft%
> 242384 -1.4% 238931 proc-vmstat.nr_inactive_anon
> 242384 -1.4% 238931 proc-vmstat.nr_zone_inactive_anon
> 1.052e+08 ą 27% +56.5% 1.647e+08 ą 10% cpuidle.C1E.time
> 1041078 ą 22% +54.7% 1610786 ą 7% cpuidle.C1E.usage
> 3.414e+08 ą 6% +57.6% 5.381e+08 ą 28% cpuidle.C6.time
> 817897 ą 3% +50.1% 1227607 ą 11% cpuidle.C6.usage
> 2884 -4.2% 2762 turbostat.Avg_MHz
> 1041024 ą 22% +54.7% 1610657 ą 7% turbostat.C1E
> 817802 ą 3% +50.1% 1227380 ą 11% turbostat.C6
> 66.75 -2.0% 65.42 turbostat.CorWatt
> 67.28 -2.0% 65.94 turbostat.PkgWatt
> 32.50 +6.2% 34.50 vmstat.cpu.id
> 62.50 -2.4% 61.00 vmstat.cpu.us
> 2443 ą 2% -28.9% 1738 ą 2% vmstat.io.bi
> 23765 ą 4% +16.5% 27685 vmstat.system.cs
> 37860 -7.1% 35180 ą 2% vmstat.system.in
> 3.474e+09 ą 3% -12.7% 3.032e+09 perf-stat.i.branch-instructions
> 1.344e+08 ą 2% -11.6% 1.188e+08 perf-stat.i.branch-misses
> 13033225 ą 4% -19.0% 10561032 perf-stat.i.cache-misses
> 5.105e+08 ą 3% -15.3% 4.322e+08 perf-stat.i.cache-references
> 24205 ą 4% +16.3% 28161 perf-stat.i.context-switches
> 30.25 ą 2% +39.7% 42.27 ą 2% perf-stat.i.cpi
> 4.63e+10 -4.7% 4.412e+10 perf-stat.i.cpu-cycles
> 3147 ą 4% -8.4% 2882 ą 2% perf-stat.i.cpu-migrations
> 16724 ą 2% +45.9% 24406 ą 5% perf-stat.i.cycles-between-cache-misses
> 0.18 ą 13% -0.1 0.12 ą 4% perf-stat.i.dTLB-load-miss-rate%
> 4.822e+09 ą 3% -11.9% 4.248e+09 perf-stat.i.dTLB-loads
> 0.07 ą 8% -0.0 0.05 ą 16% perf-stat.i.dTLB-store-miss-rate%
> 1.623e+09 ą 2% -11.5% 1.436e+09 perf-stat.i.dTLB-stores
> 1007120 ą 3% -8.9% 917854 ą 2% perf-stat.i.iTLB-load-misses
> 1.816e+10 ą 3% -12.2% 1.594e+10 perf-stat.i.instructions
> 2.06 ą 54% -66.0% 0.70 perf-stat.i.major-faults
> 29896 ą 13% -35.2% 19362 ą 8% perf-stat.i.minor-faults
> 0.00 ą 9% -0.0 0.00 ą 6% perf-stat.i.node-load-miss-rate%
> 1295134 ą 3% -14.2% 1111173 perf-stat.i.node-loads
> 3064949 ą 4% -18.7% 2491063 ą 2% perf-stat.i.node-stores
> 29898 ą 13% -35.2% 19363 ą 8% perf-stat.i.page-faults
> 28.10 -3.5% 27.12 perf-stat.overall.MPKI
> 2.55 -0.1 2.44 ą 2% perf-stat.overall.cache-miss-rate%
> 2.56 ą 3% +8.5% 2.77 perf-stat.overall.cpi
> 3567 ą 5% +17.3% 4186 perf-stat.overall.cycles-between-cache-misses
> 0.02 ą 3% +0.0 0.02 ą 3% perf-stat.overall.dTLB-load-miss-rate%
> 18031 -3.6% 17375 ą 2% perf-stat.overall.instructions-per-iTLB-miss
> 0.39 ą 3% -7.9% 0.36 perf-stat.overall.ipc
> 3.446e+09 ą 3% -12.6% 3.011e+09 perf-stat.ps.branch-instructions
> 1.333e+08 ą 2% -11.5% 1.18e+08 perf-stat.ps.branch-misses
> 12927998 ą 4% -18.8% 10491818 perf-stat.ps.cache-misses
> 5.064e+08 ą 3% -15.2% 4.293e+08 perf-stat.ps.cache-references
> 24011 ą 4% +16.5% 27973 perf-stat.ps.context-switches
> 4.601e+10 -4.6% 4.391e+10 perf-stat.ps.cpu-cycles
> 3121 ą 4% -8.3% 2863 ą 2% perf-stat.ps.cpu-migrations
> 4.783e+09 ą 3% -11.8% 4.219e+09 perf-stat.ps.dTLB-loads
> 1.61e+09 ą 2% -11.4% 1.426e+09 perf-stat.ps.dTLB-stores
> 999100 ą 3% -8.7% 911974 ą 2% perf-stat.ps.iTLB-load-misses
> 1.802e+10 ą 3% -12.1% 1.584e+10 perf-stat.ps.instructions
> 2.04 ą 54% -65.9% 0.70 perf-stat.ps.major-faults
> 29656 ą 13% -35.1% 19237 ą 8% perf-stat.ps.minor-faults
> 1284601 ą 3% -14.1% 1103823 perf-stat.ps.node-loads
> 3039931 ą 4% -18.6% 2474451 ą 2% perf-stat.ps.node-stores
> 29658 ą 13% -35.1% 19238 ą 8% perf-stat.ps.page-faults
> 50384 ą 2% +16.5% 58713 ą 4% softirqs.CPU0.RCU
> 33143 ą 2% +19.9% 39731 ą 2% softirqs.CPU0.SCHED
> 72672 +24.0% 90109 softirqs.CPU0.TIMER
> 22182 ą 4% +26.3% 28008 ą 4% softirqs.CPU1.SCHED
> 74465 ą 4% +26.3% 94027 ą 3% softirqs.CPU1.TIMER
> 18680 ą 7% +29.2% 24135 ą 3% softirqs.CPU10.SCHED
> 75941 ą 2% +21.8% 92486 ą 7% softirqs.CPU10.TIMER
> 48991 ą 4% +22.7% 60105 ą 5% softirqs.CPU11.RCU
> 18666 ą 6% +28.4% 23976 ą 4% softirqs.CPU11.SCHED
> 74896 ą 6% +24.4% 93173 ą 3% softirqs.CPU11.TIMER
> 49490 +20.5% 59659 ą 2% softirqs.CPU12.RCU
> 18973 ą 7% +26.0% 23909 ą 3% softirqs.CPU12.SCHED
> 50620 +19.9% 60677 ą 6% softirqs.CPU13.RCU
> 19136 ą 6% +23.2% 23577 ą 4% softirqs.CPU13.SCHED
> 74812 +33.3% 99756 ą 7% softirqs.CPU13.TIMER
> 50824 +15.9% 58881 ą 3% softirqs.CPU14.RCU
> 19550 ą 5% +24.1% 24270 ą 4% softirqs.CPU14.SCHED
> 76801 +22.8% 94309 ą 4% softirqs.CPU14.TIMER
> 51844 +11.5% 57795 ą 3% softirqs.CPU15.RCU
> 19204 ą 8% +28.4% 24662 ą 2% softirqs.CPU15.SCHED
> 74751 +29.9% 97127 ą 3% softirqs.CPU15.TIMER
> 50307 +17.4% 59062 ą 4% softirqs.CPU2.RCU
> 22150 +12.2% 24848 softirqs.CPU2.SCHED
> 79653 ą 2% +21.6% 96829 ą 10% softirqs.CPU2.TIMER
> 50833 +21.1% 61534 ą 4% softirqs.CPU3.RCU
> 18935 ą 2% +32.0% 25002 ą 3% softirqs.CPU3.SCHED
> 50569 +15.8% 58570 ą 4% softirqs.CPU4.RCU
> 20509 ą 5% +18.3% 24271 softirqs.CPU4.SCHED
> 80942 ą 2% +15.3% 93304 ą 3% softirqs.CPU4.TIMER
> 50692 +16.5% 59067 ą 4% softirqs.CPU5.RCU
> 20237 ą 3% +18.2% 23914 ą 3% softirqs.CPU5.SCHED
> 78963 +21.8% 96151 ą 2% softirqs.CPU5.TIMER
> 19709 ą 7% +20.1% 23663 softirqs.CPU6.SCHED
> 81250 +15.9% 94185 softirqs.CPU6.TIMER
> 51379 +15.0% 59108 softirqs.CPU7.RCU
> 19642 ą 5% +28.4% 25227 ą 3% softirqs.CPU7.SCHED
> 78299 ą 2% +30.3% 102021 ą 4% softirqs.CPU7.TIMER
> 49723 +19.0% 59169 ą 4% softirqs.CPU8.RCU
> 20138 ą 6% +21.7% 24501 ą 2% softirqs.CPU8.SCHED
> 75256 ą 3% +22.8% 92419 ą 2% softirqs.CPU8.TIMER
> 50406 ą 2% +17.4% 59178 ą 4% softirqs.CPU9.RCU
> 19182 ą 9% +24.2% 23831 ą 6% softirqs.CPU9.SCHED
> 73572 ą 5% +30.4% 95951 ą 8% softirqs.CPU9.TIMER
> 812363 +16.6% 946858 ą 3% softirqs.RCU
> 330042 ą 4% +23.5% 407533 softirqs.SCHED
> 1240046 +22.5% 1519539 softirqs.TIMER
> 251015 ą 21% -84.2% 39587 ą106% sched_debug.cfs_rq:/.MIN_vruntime.avg
> 537847 ą 4% -44.8% 297100 ą 66% sched_debug.cfs_rq:/.MIN_vruntime.max
> 257990 ą 5% -63.4% 94515 ą 82% sched_debug.cfs_rq:/.MIN_vruntime.stddev
> 38935 +47.9% 57601 sched_debug.cfs_rq:/.exec_clock.avg
> 44119 +40.6% 62013 sched_debug.cfs_rq:/.exec_clock.max
> 37622 +49.9% 56404 sched_debug.cfs_rq:/.exec_clock.min
> 47287 ą 7% -70.3% 14036 ą 88% sched_debug.cfs_rq:/.load.min
> 67.17 -52.9% 31.62 ą 31% sched_debug.cfs_rq:/.load_avg.min
> 251015 ą 21% -84.2% 39588 ą106% sched_debug.cfs_rq:/.max_vruntime.avg
> 537847 ą 4% -44.8% 297103 ą 66% sched_debug.cfs_rq:/.max_vruntime.max
> 257991 ą 5% -63.4% 94516 ą 82% sched_debug.cfs_rq:/.max_vruntime.stddev
> 529078 ą 3% +45.2% 768398 sched_debug.cfs_rq:/.min_vruntime.avg
> 547175 ą 2% +44.1% 788582 sched_debug.cfs_rq:/.min_vruntime.max
> 496420 +48.3% 736148 ą 2% sched_debug.cfs_rq:/.min_vruntime.min
> 1.33 ą 15% -44.0% 0.75 ą 32% sched_debug.cfs_rq:/.nr_running.avg
> 0.83 ą 20% -70.0% 0.25 ą 70% sched_debug.cfs_rq:/.nr_running.min
> 0.54 ą 8% -15.9% 0.45 ą 7% sched_debug.cfs_rq:/.nr_running.stddev
> 0.33 +62.9% 0.54 ą 8% sched_debug.cfs_rq:/.nr_spread_over.avg
> 1.33 +54.7% 2.06 ą 17% sched_debug.cfs_rq:/.nr_spread_over.max
> 0.44 ą 7% +56.4% 0.69 ą 6% sched_debug.cfs_rq:/.nr_spread_over.stddev
> 130.83 ą 14% -25.6% 97.37 ą 15% sched_debug.cfs_rq:/.runnable_load_avg.avg
> 45.33 ą 6% -79.3% 9.38 ą 70% sched_debug.cfs_rq:/.runnable_load_avg.min
> 47283 ą 7% -70.9% 13741 ą 89% sched_debug.cfs_rq:/.runnable_weight.min
> 1098 ą 8% -27.6% 795.02 ą 24% sched_debug.cfs_rq:/.util_avg.avg
> 757.50 ą 9% -51.3% 369.25 ą 10% sched_debug.cfs_rq:/.util_avg.min
> 762.39 ą 11% -44.4% 424.04 ą 46% sched_debug.cfs_rq:/.util_est_enqueued.avg
> 314.00 ą 18% -78.5% 67.38 ą100% sched_debug.cfs_rq:/.util_est_enqueued.min
> 142951 ą 5% +22.8% 175502 ą 3% sched_debug.cpu.avg_idle.avg
> 72112 -18.3% 58937 ą 13% sched_debug.cpu.avg_idle.stddev
> 127638 ą 7% +39.3% 177858 ą 5% sched_debug.cpu.clock.avg
> 127643 ą 7% +39.3% 177862 ą 5% sched_debug.cpu.clock.max
> 127633 ą 7% +39.3% 177855 ą 5% sched_debug.cpu.clock.min
> 126720 ą 7% +39.4% 176681 ą 5% sched_debug.cpu.clock_task.avg
> 127168 ą 7% +39.3% 177179 ą 5% sched_debug.cpu.clock_task.max
> 125240 ą 7% +39.5% 174767 ą 5% sched_debug.cpu.clock_task.min
> 563.60 ą 2% +25.9% 709.78 ą 9% sched_debug.cpu.clock_task.stddev
> 1.66 ą 18% -37.5% 1.04 ą 32% sched_debug.cpu.nr_running.avg
> 0.83 ą 20% -62.5% 0.31 ą 87% sched_debug.cpu.nr_running.min
> 127617 ą 3% +52.9% 195080 sched_debug.cpu.nr_switches.avg
> 149901 ą 6% +45.2% 217652 sched_debug.cpu.nr_switches.max
> 108182 ą 5% +61.6% 174808 sched_debug.cpu.nr_switches.min
> 0.20 ą 5% -62.5% 0.07 ą 67% sched_debug.cpu.nr_uninterruptible.avg
> -29.33 -13.5% -25.38 sched_debug.cpu.nr_uninterruptible.min
> 92666 ą 8% +66.8% 154559 sched_debug.cpu.sched_count.avg
> 104565 ą 11% +57.2% 164374 sched_debug.cpu.sched_count.max
> 80272 ą 10% +77.2% 142238 sched_debug.cpu.sched_count.min
> 38029 ą 10% +80.4% 68608 sched_debug.cpu.sched_goidle.avg
> 43413 ą 11% +68.5% 73149 sched_debug.cpu.sched_goidle.max
> 32420 ą 11% +94.5% 63069 sched_debug.cpu.sched_goidle.min
> 51567 ą 8% +60.7% 82878 sched_debug.cpu.ttwu_count.avg
> 79015 ą 9% +45.2% 114717 ą 4% sched_debug.cpu.ttwu_count.max
> 42919 ą 9% +63.3% 70086 sched_debug.cpu.ttwu_count.min
> 127632 ą 7% +39.3% 177854 ą 5% sched_debug.cpu_clk
> 125087 ą 7% +40.1% 175285 ą 5% sched_debug.ktime
> 127882 ą 6% +39.3% 178163 ą 5% sched_debug.sched_clk
> 146.00 ą 13% +902.9% 1464 ą143% interrupts.133:IR-PCI-MSI.2097153-edge.eth1-TxRx-0
> 3375 ą 93% -94.8% 174.75 ą 26% interrupts.134:IR-PCI-MSI.2097154-edge.eth1-TxRx-1
> 297595 ą 8% +22.8% 365351 ą 2% interrupts.CPU0.LOC:Local_timer_interrupts
> 8402 -21.7% 6577 ą 25% interrupts.CPU0.NMI:Non-maskable_interrupts
> 8402 -21.7% 6577 ą 25% interrupts.CPU0.PMI:Performance_monitoring_interrupts
> 937.00 ą 2% +18.1% 1106 ą 11% interrupts.CPU0.RES:Rescheduling_interrupts
> 146.00 ą 13% +902.9% 1464 ą143% interrupts.CPU1.133:IR-PCI-MSI.2097153-edge.eth1-TxRx-0
> 297695 ą 8% +22.7% 365189 ą 2% interrupts.CPU1.LOC:Local_timer_interrupts
> 8412 -20.9% 6655 ą 25% interrupts.CPU1.NMI:Non-maskable_interrupts
> 8412 -20.9% 6655 ą 25% interrupts.CPU1.PMI:Performance_monitoring_interrupts
> 297641 ą 8% +22.7% 365268 ą 2% interrupts.CPU10.LOC:Local_timer_interrupts
> 8365 -10.9% 7455 ą 3% interrupts.CPU10.NMI:Non-maskable_interrupts
> 8365 -10.9% 7455 ą 3% interrupts.CPU10.PMI:Performance_monitoring_interrupts
> 297729 ą 8% +22.7% 365238 ą 2% interrupts.CPU11.LOC:Local_timer_interrupts
> 8376 -21.8% 6554 ą 26% interrupts.CPU11.NMI:Non-maskable_interrupts
> 8376 -21.8% 6554 ą 26% interrupts.CPU11.PMI:Performance_monitoring_interrupts
> 297394 ą 8% +22.8% 365274 ą 2% interrupts.CPU12.LOC:Local_timer_interrupts
> 8393 -10.5% 7512 ą 3% interrupts.CPU12.NMI:Non-maskable_interrupts
> 8393 -10.5% 7512 ą 3% interrupts.CPU12.PMI:Performance_monitoring_interrupts
> 297744 ą 8% +22.7% 365243 ą 2% interrupts.CPU13.LOC:Local_timer_interrupts
> 8353 -10.6% 7469 ą 3% interrupts.CPU13.NMI:Non-maskable_interrupts
> 8353 -10.6% 7469 ą 3% interrupts.CPU13.PMI:Performance_monitoring_interrupts
> 148.50 ą 17% -24.2% 112.50 ą 8% interrupts.CPU13.TLB:TLB_shootdowns
> 297692 ą 8% +22.7% 365311 ą 2% interrupts.CPU14.LOC:Local_timer_interrupts
> 8374 -10.4% 7501 ą 4% interrupts.CPU14.NMI:Non-maskable_interrupts
> 8374 -10.4% 7501 ą 4% interrupts.CPU14.PMI:Performance_monitoring_interrupts
> 297453 ą 8% +22.8% 365311 ą 2% interrupts.CPU15.LOC:Local_timer_interrupts
> 8336 -22.8% 6433 ą 26% interrupts.CPU15.NMI:Non-maskable_interrupts
> 8336 -22.8% 6433 ą 26% interrupts.CPU15.PMI:Performance_monitoring_interrupts
> 699.50 ą 21% +51.3% 1058 ą 10% interrupts.CPU15.RES:Rescheduling_interrupts
> 3375 ą 93% -94.8% 174.75 ą 26% interrupts.CPU2.134:IR-PCI-MSI.2097154-edge.eth1-TxRx-1
> 297685 ą 8% +22.7% 365273 ą 2% interrupts.CPU2.LOC:Local_timer_interrupts
> 8357 -21.2% 6584 ą 25% interrupts.CPU2.NMI:Non-maskable_interrupts
> 8357 -21.2% 6584 ą 25% interrupts.CPU2.PMI:Performance_monitoring_interrupts
> 164.00 ą 30% -23.0% 126.25 ą 32% interrupts.CPU2.TLB:TLB_shootdowns
> 297352 ą 8% +22.9% 365371 ą 2% interrupts.CPU3.LOC:Local_timer_interrupts
> 8383 -10.6% 7493 ą 4% interrupts.CPU3.NMI:Non-maskable_interrupts
> 8383 -10.6% 7493 ą 4% interrupts.CPU3.PMI:Performance_monitoring_interrupts
> 780.50 ą 3% +32.7% 1035 ą 6% interrupts.CPU3.RES:Rescheduling_interrupts
> 297595 ą 8% +22.8% 365415 ą 2% interrupts.CPU4.LOC:Local_timer_interrupts
> 8382 -21.4% 6584 ą 25% interrupts.CPU4.NMI:Non-maskable_interrupts
> 8382 -21.4% 6584 ą 25% interrupts.CPU4.PMI:Performance_monitoring_interrupts
> 297720 ą 8% +22.7% 365347 ą 2% interrupts.CPU5.LOC:Local_timer_interrupts
> 8353 -32.0% 5679 ą 34% interrupts.CPU5.NMI:Non-maskable_interrupts
> 8353 -32.0% 5679 ą 34% interrupts.CPU5.PMI:Performance_monitoring_interrupts
> 727.00 ą 16% +98.3% 1442 ą 47% interrupts.CPU5.RES:Rescheduling_interrupts
> 297620 ą 8% +22.8% 365343 ą 2% interrupts.CPU6.LOC:Local_timer_interrupts
> 8388 -10.3% 7526 ą 4% interrupts.CPU6.NMI:Non-maskable_interrupts
> 8388 -10.3% 7526 ą 4% interrupts.CPU6.PMI:Performance_monitoring_interrupts
> 156.50 ą 3% -27.6% 113.25 ą 16% interrupts.CPU6.TLB:TLB_shootdowns
> 297690 ą 8% +22.7% 365363 ą 2% interrupts.CPU7.LOC:Local_timer_interrupts
> 8390 -23.1% 6449 ą 25% interrupts.CPU7.NMI:Non-maskable_interrupts
> 8390 -23.1% 6449 ą 25% interrupts.CPU7.PMI:Performance_monitoring_interrupts
> 918.00 ą 16% +49.4% 1371 ą 7% interrupts.CPU7.RES:Rescheduling_interrupts
> 120.00 ą 35% +70.8% 205.00 ą 17% interrupts.CPU7.TLB:TLB_shootdowns
> 297731 ą 8% +22.7% 365368 ą 2% interrupts.CPU8.LOC:Local_timer_interrupts
> 8393 -32.5% 5668 ą 35% interrupts.CPU8.NMI:Non-maskable_interrupts
> 8393 -32.5% 5668 ą 35% interrupts.CPU8.PMI:Performance_monitoring_interrupts
> 297779 ą 8% +22.7% 365399 ą 2% interrupts.CPU9.LOC:Local_timer_interrupts
> 8430 -10.8% 7517 ą 2% interrupts.CPU9.NMI:Non-maskable_interrupts
> 8430 -10.8% 7517 ą 2% interrupts.CPU9.PMI:Performance_monitoring_interrupts
> 956.50 +13.5% 1085 ą 4% interrupts.CPU9.RES:Rescheduling_interrupts
> 4762118 ą 8% +22.7% 5845069 ą 2% interrupts.LOC:Local_timer_interrupts
> 134093 -18.2% 109662 ą 11% interrupts.NMI:Non-maskable_interrupts
> 134093 -18.2% 109662 ą 11% interrupts.PMI:Performance_monitoring_interrupts
> 66.97 ą 9% -29.9 37.12 ą 49% perf-profile.calltrace.cycles-pp.deflate
> 66.67 ą 9% -29.7 36.97 ą 50% perf-profile.calltrace.cycles-pp.deflate_medium.deflate
> 43.24 ą 9% -18.6 24.61 ą 49% perf-profile.calltrace.cycles-pp.longest_match.deflate_medium.deflate
> 2.29 ą 14% -1.2 1.05 ą 58% perf-profile.calltrace.cycles-pp.deflateSetDictionary
> 0.74 ą 6% -0.5 0.27 ą100% perf-profile.calltrace.cycles-pp.read.__libc_start_main
> 0.74 ą 7% -0.5 0.27 ą100% perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.read.__libc_start_main
> 0.73 ą 7% -0.5 0.27 ą100% perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.read.__libc_start_main
> 0.73 ą 7% -0.5 0.27 ą100% perf-profile.calltrace.cycles-pp.__x64_sys_read.do_syscall_64.entry_SYSCALL_64_after_hwframe.read.__libc_start_main
> 0.73 ą 7% -0.5 0.27 ą100% perf-profile.calltrace.cycles-pp.ksys_read.__x64_sys_read.do_syscall_64.entry_SYSCALL_64_after_hwframe.read
> 0.26 ą100% +0.6 0.88 ą 45% perf-profile.calltrace.cycles-pp.vfs_statx.__do_sys_newfstatat.__x64_sys_newfstatat.do_syscall_64.entry_SYSCALL_64_after_hwframe
> 0.34 ą100% +0.7 1.02 ą 42% perf-profile.calltrace.cycles-pp.do_sys_open.__x64_sys_openat.do_syscall_64.entry_SYSCALL_64_after_hwframe
> 0.28 ą100% +0.7 0.96 ą 44% perf-profile.calltrace.cycles-pp.__do_sys_newfstatat.__x64_sys_newfstatat.do_syscall_64.entry_SYSCALL_64_after_hwframe
> 0.28 ą100% +0.7 0.96 ą 44% perf-profile.calltrace.cycles-pp.__x64_sys_newfstatat.do_syscall_64.entry_SYSCALL_64_after_hwframe
> 0.34 ą100% +0.7 1.03 ą 42% perf-profile.calltrace.cycles-pp.__x64_sys_openat.do_syscall_64.entry_SYSCALL_64_after_hwframe
> 0.00 +0.8 0.77 ą 35% perf-profile.calltrace.cycles-pp.hrtimer_interrupt.smp_apic_timer_interrupt.apic_timer_interrupt.cpuidle_enter_state.cpuidle_enter
> 0.56 ą 9% +0.8 1.40 ą 45% perf-profile.calltrace.cycles-pp.__schedule.schedule.futex_wait_queue_me.futex_wait.do_futex
> 0.58 ą 10% +0.9 1.43 ą 45% perf-profile.calltrace.cycles-pp.schedule.futex_wait_queue_me.futex_wait.do_futex.__x64_sys_futex
> 0.33 ą100% +0.9 1.21 ą 28% perf-profile.calltrace.cycles-pp.menu_select.cpuidle_select.do_idle.cpu_startup_entry.start_secondary
> 0.34 ą 99% +0.9 1.27 ą 30% perf-profile.calltrace.cycles-pp.cpuidle_select.do_idle.cpu_startup_entry.start_secondary.secondary_startup_64
> 0.00 +1.0 0.96 ą 62% perf-profile.calltrace.cycles-pp.poll_idle.cpuidle_enter_state.cpuidle_enter.do_idle.cpu_startup_entry
> 0.62 ą 9% +1.0 1.60 ą 52% perf-profile.calltrace.cycles-pp.futex_wait_queue_me.futex_wait.do_futex.__x64_sys_futex.do_syscall_64
> 0.68 ą 10% +1.0 1.73 ą 51% perf-profile.calltrace.cycles-pp.futex_wait.do_futex.__x64_sys_futex.do_syscall_64.entry_SYSCALL_64_after_hwframe
> 0.46 ą100% +1.1 1.60 ą 43% perf-profile.calltrace.cycles-pp.smp_apic_timer_interrupt.apic_timer_interrupt.cpuidle_enter_state.cpuidle_enter.do_idle
> 0.47 ą100% +1.2 1.62 ą 43% perf-profile.calltrace.cycles-pp.apic_timer_interrupt.cpuidle_enter_state.cpuidle_enter.do_idle.cpu_startup_entry
> 17.73 ą 21% +19.1 36.84 ą 25% perf-profile.calltrace.cycles-pp.intel_idle.cpuidle_enter_state.cpuidle_enter.do_idle.cpu_startup_entry
> 17.75 ą 20% +19.9 37.63 ą 26% perf-profile.calltrace.cycles-pp.cpuidle_enter_state.cpuidle_enter.do_idle.cpu_startup_entry.start_secondary
> 17.84 ą 20% +20.0 37.82 ą 26% perf-profile.calltrace.cycles-pp.cpuidle_enter.do_idle.cpu_startup_entry.start_secondary.secondary_startup_64
> 18.83 ą 20% +21.2 40.05 ą 27% perf-profile.calltrace.cycles-pp.do_idle.cpu_startup_entry.start_secondary.secondary_startup_64
> 18.89 ą 20% +21.2 40.11 ą 27% perf-profile.calltrace.cycles-pp.cpu_startup_entry.start_secondary.secondary_startup_64
> 18.89 ą 20% +21.2 40.12 ą 27% perf-profile.calltrace.cycles-pp.start_secondary.secondary_startup_64
> 20.14 ą 20% +22.5 42.66 ą 27% perf-profile.calltrace.cycles-pp.secondary_startup_64
> 66.97 ą 9% -29.9 37.12 ą 49% perf-profile.children.cycles-pp.deflate
> 66.83 ą 9% -29.8 37.06 ą 49% perf-profile.children.cycles-pp.deflate_medium
> 43.58 ą 9% -18.8 24.80 ą 49% perf-profile.children.cycles-pp.longest_match
> 2.29 ą 14% -1.2 1.12 ą 43% perf-profile.children.cycles-pp.deflateSetDictionary
> 0.84 -0.3 0.58 ą 19% perf-profile.children.cycles-pp.read
> 0.52 ą 13% -0.2 0.27 ą 43% perf-profile.children.cycles-pp.fill_window
> 0.06 +0.0 0.08 ą 13% perf-profile.children.cycles-pp.update_stack_state
> 0.07 ą 14% +0.0 0.11 ą 24% perf-profile.children.cycles-pp._raw_spin_unlock_irqrestore
> 0.03 ą100% +0.1 0.09 ą 19% perf-profile.children.cycles-pp.find_next_and_bit
> 0.00 +0.1 0.06 ą 15% perf-profile.children.cycles-pp.refcount_inc_not_zero_checked
> 0.03 ą100% +0.1 0.08 ą 33% perf-profile.children.cycles-pp.free_pcppages_bulk
> 0.07 ą 7% +0.1 0.12 ą 36% perf-profile.children.cycles-pp.syscall_return_via_sysret
> 0.00 +0.1 0.06 ą 28% perf-profile.children.cycles-pp.rb_erase
> 0.03 ą100% +0.1 0.09 ą 24% perf-profile.children.cycles-pp.shmem_undo_range
> 0.03 ą100% +0.1 0.09 ą 28% perf-profile.children.cycles-pp.unlinkat
> 0.03 ą100% +0.1 0.09 ą 28% perf-profile.children.cycles-pp.__x64_sys_unlinkat
> 0.03 ą100% +0.1 0.09 ą 28% perf-profile.children.cycles-pp.do_unlinkat
> 0.03 ą100% +0.1 0.09 ą 28% perf-profile.children.cycles-pp.ovl_destroy_inode
> 0.03 ą100% +0.1 0.09 ą 28% perf-profile.children.cycles-pp.shmem_evict_inode
> 0.03 ą100% +0.1 0.09 ą 28% perf-profile.children.cycles-pp.shmem_truncate_range
> 0.05 +0.1 0.12 ą 38% perf-profile.children.cycles-pp.unmap_vmas
> 0.00 +0.1 0.07 ą 30% perf-profile.children.cycles-pp.timerqueue_del
> 0.00 +0.1 0.07 ą 26% perf-profile.children.cycles-pp.idle_cpu
> 0.09 ą 17% +0.1 0.15 ą 19% perf-profile.children.cycles-pp.__update_load_avg_se
> 0.00 +0.1 0.07 ą 33% perf-profile.children.cycles-pp.unmap_region
> 0.00 +0.1 0.07 ą 33% perf-profile.children.cycles-pp.__alloc_fd
> 0.03 ą100% +0.1 0.10 ą 31% perf-profile.children.cycles-pp.destroy_inode
> 0.03 ą100% +0.1 0.10 ą 30% perf-profile.children.cycles-pp.evict
> 0.06 ą 16% +0.1 0.13 ą 35% perf-profile.children.cycles-pp.ovl_override_creds
> 0.00 +0.1 0.07 ą 26% perf-profile.children.cycles-pp.kernel_text_address
> 0.00 +0.1 0.07 ą 41% perf-profile.children.cycles-pp.file_remove_privs
> 0.07 ą 23% +0.1 0.14 ą 47% perf-profile.children.cycles-pp.hrtimer_next_event_without
> 0.03 ą100% +0.1 0.10 ą 24% perf-profile.children.cycles-pp.__dentry_kill
> 0.03 ą100% +0.1 0.10 ą 29% perf-profile.children.cycles-pp.dentry_unlink_inode
> 0.03 ą100% +0.1 0.10 ą 29% perf-profile.children.cycles-pp.iput
> 0.03 ą100% +0.1 0.10 ą 54% perf-profile.children.cycles-pp.__close_fd
> 0.08 ą 25% +0.1 0.15 ą 35% perf-profile.children.cycles-pp.__switch_to
> 0.00 +0.1 0.07 ą 29% perf-profile.children.cycles-pp.__switch_to_asm
> 0.00 +0.1 0.08 ą 24% perf-profile.children.cycles-pp.__kernel_text_address
> 0.03 ą100% +0.1 0.11 ą 51% perf-profile.children.cycles-pp.enqueue_hrtimer
> 0.03 ą100% +0.1 0.11 ą 33% perf-profile.children.cycles-pp.rcu_gp_kthread_wake
> 0.03 ą100% +0.1 0.11 ą 33% perf-profile.children.cycles-pp.swake_up_one
> 0.03 ą100% +0.1 0.11 ą 33% perf-profile.children.cycles-pp.swake_up_locked
> 0.10 ą 30% +0.1 0.18 ą 17% perf-profile.children.cycles-pp.irqtime_account_irq
> 0.03 ą100% +0.1 0.11 ą 38% perf-profile.children.cycles-pp.unmap_page_range
> 0.00 +0.1 0.09 ą 37% perf-profile.children.cycles-pp.putname
> 0.03 ą100% +0.1 0.11 ą 28% perf-profile.children.cycles-pp.filemap_map_pages
> 0.07 ą 28% +0.1 0.16 ą 35% perf-profile.children.cycles-pp.getname
> 0.03 ą100% +0.1 0.11 ą 40% perf-profile.children.cycles-pp.unmap_single_vma
> 0.08 ą 29% +0.1 0.17 ą 38% perf-profile.children.cycles-pp.queued_spin_lock_slowpath
> 0.03 ą100% +0.1 0.12 ą 54% perf-profile.children.cycles-pp.setlocale
> 0.03 ą100% +0.1 0.12 ą 60% perf-profile.children.cycles-pp.__open64_nocancel
> 0.00 +0.1 0.09 ą 31% perf-profile.children.cycles-pp.__hrtimer_get_next_event
> 0.00 +0.1 0.10 ą 28% perf-profile.children.cycles-pp.get_unused_fd_flags
> 0.00 +0.1 0.10 ą 65% perf-profile.children.cycles-pp.timerqueue_add
> 0.07 ą 28% +0.1 0.17 ą 42% perf-profile.children.cycles-pp.__hrtimer_next_event_base
> 0.03 ą100% +0.1 0.12 ą 51% perf-profile.children.cycles-pp.__x64_sys_close
> 0.00 +0.1 0.10 ą 38% perf-profile.children.cycles-pp.do_lookup_x
> 0.03 ą100% +0.1 0.12 ą 23% perf-profile.children.cycles-pp.kmem_cache_free
> 0.04 ą100% +0.1 0.14 ą 53% perf-profile.children.cycles-pp.__do_munmap
> 0.00 +0.1 0.10 ą 35% perf-profile.children.cycles-pp.unwind_get_return_address
> 0.00 +0.1 0.10 ą 49% perf-profile.children.cycles-pp.shmem_add_to_page_cache
> 0.07 ą 20% +0.1 0.18 ą 25% perf-profile.children.cycles-pp.find_next_bit
> 0.08 ą 25% +0.1 0.18 ą 34% perf-profile.children.cycles-pp.dput
> 0.11 ą 33% +0.1 0.21 ą 37% perf-profile.children.cycles-pp.perf_mux_hrtimer_handler
> 0.08 ą 5% +0.1 0.19 ą 27% perf-profile.children.cycles-pp._raw_spin_lock_irqsave
> 0.00 +0.1 0.11 ą 52% perf-profile.children.cycles-pp.rcu_idle_exit
> 0.03 ą100% +0.1 0.14 ą 18% perf-profile.children.cycles-pp.entry_SYSCALL_64
> 0.08 +0.1 0.19 ą 43% perf-profile.children.cycles-pp.exit_mmap
> 0.09 ą 22% +0.1 0.20 ą 57% perf-profile.children.cycles-pp.set_next_entity
> 0.07 ą 7% +0.1 0.18 ą 60% perf-profile.children.cycles-pp.switch_mm_irqs_off
> 0.10 ą 26% +0.1 0.21 ą 32% perf-profile.children.cycles-pp.sched_clock
> 0.12 ą 25% +0.1 0.23 ą 39% perf-profile.children.cycles-pp.update_cfs_group
> 0.07 ą 14% +0.1 0.18 ą 45% perf-profile.children.cycles-pp.lapic_next_deadline
> 0.08 ą 5% +0.1 0.20 ą 44% perf-profile.children.cycles-pp.mmput
> 0.11 ą 18% +0.1 0.23 ą 41% perf-profile.children.cycles-pp.clockevents_program_event
> 0.07 ą 7% +0.1 0.18 ą 40% perf-profile.children.cycles-pp.strncpy_from_user
> 0.00 +0.1 0.12 ą 48% perf-profile.children.cycles-pp.flush_old_exec
> 0.11 ą 18% +0.1 0.23 ą 37% perf-profile.children.cycles-pp.native_sched_clock
> 0.09 ą 17% +0.1 0.21 ą 46% perf-profile.children.cycles-pp._dl_sysdep_start
> 0.12 ą 19% +0.1 0.26 ą 37% perf-profile.children.cycles-pp.tick_program_event
> 0.09 ą 33% +0.1 0.23 ą 61% perf-profile.children.cycles-pp.mmap_region
> 0.14 ą 21% +0.1 0.28 ą 39% perf-profile.children.cycles-pp.sched_clock_cpu
> 0.11 ą 27% +0.1 0.25 ą 56% perf-profile.children.cycles-pp.do_mmap
> 0.11 ą 36% +0.1 0.26 ą 57% perf-profile.children.cycles-pp.ksys_mmap_pgoff
> 0.09 ą 17% +0.1 0.23 ą 48% perf-profile.children.cycles-pp.lookup_fast
> 0.04 ą100% +0.2 0.19 ą 48% perf-profile.children.cycles-pp.open_path
> 0.11 ą 30% +0.2 0.27 ą 58% perf-profile.children.cycles-pp.vm_mmap_pgoff
> 0.11 ą 27% +0.2 0.28 ą 37% perf-profile.children.cycles-pp.update_blocked_averages
> 0.11 +0.2 0.29 ą 38% perf-profile.children.cycles-pp.search_binary_handler
> 0.11 ą 4% +0.2 0.29 ą 39% perf-profile.children.cycles-pp.load_elf_binary
> 0.11 ą 30% +0.2 0.30 ą 50% perf-profile.children.cycles-pp.inode_permission
> 0.04 ą100% +0.2 0.24 ą 55% perf-profile.children.cycles-pp.__GI___open64_nocancel
> 0.15 ą 29% +0.2 0.35 ą 34% perf-profile.children.cycles-pp.getname_flags
> 0.16 ą 25% +0.2 0.38 ą 26% perf-profile.children.cycles-pp.get_next_timer_interrupt
> 0.18 ą 11% +0.2 0.41 ą 39% perf-profile.children.cycles-pp.execve
> 0.19 ą 5% +0.2 0.42 ą 37% perf-profile.children.cycles-pp.__x64_sys_execve
> 0.18 ą 2% +0.2 0.42 ą 37% perf-profile.children.cycles-pp.__do_execve_file
> 0.32 ą 18% +0.3 0.57 ą 33% perf-profile.children.cycles-pp.__account_scheduler_latency
> 0.21 ą 17% +0.3 0.48 ą 47% perf-profile.children.cycles-pp.schedule_idle
> 0.20 ą 19% +0.3 0.49 ą 33% perf-profile.children.cycles-pp.tick_nohz_next_event
> 0.21 ą 26% +0.3 0.55 ą 41% perf-profile.children.cycles-pp.find_busiest_group
> 0.32 ą 26% +0.3 0.67 ą 52% perf-profile.children.cycles-pp.__handle_mm_fault
> 0.22 ą 25% +0.4 0.57 ą 49% perf-profile.children.cycles-pp.filename_lookup
> 0.34 ą 27% +0.4 0.72 ą 50% perf-profile.children.cycles-pp.handle_mm_fault
> 0.42 ą 32% +0.4 0.80 ą 43% perf-profile.children.cycles-pp.shmem_getpage_gfp
> 0.36 ą 23% +0.4 0.77 ą 41% perf-profile.children.cycles-pp.load_balance
> 0.41 ą 30% +0.4 0.82 ą 50% perf-profile.children.cycles-pp.do_page_fault
> 0.39 ą 30% +0.4 0.80 ą 50% perf-profile.children.cycles-pp.__do_page_fault
> 0.28 ą 22% +0.4 0.70 ą 37% perf-profile.children.cycles-pp.tick_nohz_get_sleep_length
> 0.43 ą 31% +0.4 0.86 ą 49% perf-profile.children.cycles-pp.page_fault
> 0.31 ą 25% +0.5 0.77 ą 46% perf-profile.children.cycles-pp.user_path_at_empty
> 0.36 ą 20% +0.5 0.84 ą 34% perf-profile.children.cycles-pp.newidle_balance
> 0.45 ą 21% +0.5 0.95 ą 44% perf-profile.children.cycles-pp.vfs_statx
> 0.46 ą 20% +0.5 0.97 ą 43% perf-profile.children.cycles-pp.__do_sys_newfstatat
> 0.47 ą 20% +0.5 0.98 ą 44% perf-profile.children.cycles-pp.__x64_sys_newfstatat
> 0.29 ą 37% +0.5 0.81 ą 32% perf-profile.children.cycles-pp.io_serial_in
> 0.53 ą 25% +0.5 1.06 ą 49% perf-profile.children.cycles-pp.path_openat
> 0.55 ą 24% +0.5 1.09 ą 50% perf-profile.children.cycles-pp.do_filp_open
> 0.35 ą 2% +0.5 0.90 ą 60% perf-profile.children.cycles-pp.uart_console_write
> 0.35 ą 4% +0.6 0.91 ą 60% perf-profile.children.cycles-pp.console_unlock
> 0.35 ą 4% +0.6 0.91 ą 60% perf-profile.children.cycles-pp.univ8250_console_write
> 0.35 ą 4% +0.6 0.91 ą 60% perf-profile.children.cycles-pp.serial8250_console_write
> 0.82 ą 23% +0.6 1.42 ą 31% perf-profile.children.cycles-pp.__hrtimer_run_queues
> 0.47 ą 28% +0.6 1.10 ą 39% perf-profile.children.cycles-pp.irq_work_interrupt
> 0.47 ą 28% +0.6 1.10 ą 39% perf-profile.children.cycles-pp.smp_irq_work_interrupt
> 0.47 ą 28% +0.6 1.10 ą 39% perf-profile.children.cycles-pp.irq_work_run
> 0.47 ą 28% +0.6 1.10 ą 39% perf-profile.children.cycles-pp.perf_duration_warn
> 0.47 ą 28% +0.6 1.10 ą 39% perf-profile.children.cycles-pp.printk
> 0.47 ą 28% +0.6 1.10 ą 39% perf-profile.children.cycles-pp.vprintk_func
> 0.47 ą 28% +0.6 1.10 ą 39% perf-profile.children.cycles-pp.vprintk_default
> 0.47 ą 28% +0.6 1.11 ą 39% perf-profile.children.cycles-pp.irq_work_run_list
> 0.49 ą 31% +0.6 1.13 ą 39% perf-profile.children.cycles-pp.vprintk_emit
> 0.54 ą 19% +0.6 1.17 ą 38% perf-profile.children.cycles-pp.pick_next_task_fair
> 0.32 ą 7% +0.7 1.02 ą 56% perf-profile.children.cycles-pp.poll_idle
> 0.60 ą 15% +0.7 1.31 ą 29% perf-profile.children.cycles-pp.menu_select
> 0.65 ą 27% +0.7 1.36 ą 45% perf-profile.children.cycles-pp.do_sys_open
> 0.62 ą 15% +0.7 1.36 ą 31% perf-profile.children.cycles-pp.cpuidle_select
> 0.66 ą 26% +0.7 1.39 ą 44% perf-profile.children.cycles-pp.__x64_sys_openat
> 1.11 ą 22% +0.9 2.03 ą 31% perf-profile.children.cycles-pp.hrtimer_interrupt
> 0.89 ą 24% +0.9 1.81 ą 42% perf-profile.children.cycles-pp.futex_wait_queue_me
> 1.16 ą 27% +1.0 2.13 ą 36% perf-profile.children.cycles-pp.schedule
> 0.97 ą 23% +1.0 1.97 ą 42% perf-profile.children.cycles-pp.futex_wait
> 1.33 ą 25% +1.2 2.55 ą 39% perf-profile.children.cycles-pp.__schedule
> 1.84 ą 26% +1.6 3.42 ą 36% perf-profile.children.cycles-pp.smp_apic_timer_interrupt
> 1.76 ą 22% +1.6 3.41 ą 40% perf-profile.children.cycles-pp.do_futex
> 1.79 ą 22% +1.7 3.49 ą 41% perf-profile.children.cycles-pp.__x64_sys_futex
> 2.23 ą 20% +1.7 3.98 ą 37% perf-profile.children.cycles-pp.apic_timer_interrupt
> 17.73 ą 21% +19.1 36.86 ą 25% perf-profile.children.cycles-pp.intel_idle
> 19.00 ą 21% +21.1 40.13 ą 26% perf-profile.children.cycles-pp.cpuidle_enter_state
> 19.02 ą 21% +21.2 40.19 ą 26% perf-profile.children.cycles-pp.cpuidle_enter
> 18.89 ą 20% +21.2 40.12 ą 27% perf-profile.children.cycles-pp.start_secondary
> 20.14 ą 20% +22.5 42.65 ą 27% perf-profile.children.cycles-pp.cpu_startup_entry
> 20.08 ą 20% +22.5 42.59 ą 27% perf-profile.children.cycles-pp.do_idle
> 20.14 ą 20% +22.5 42.66 ą 27% perf-profile.children.cycles-pp.secondary_startup_64
> 43.25 ą 9% -18.6 24.63 ą 49% perf-profile.self.cycles-pp.longest_match
> 22.74 ą 11% -10.8 11.97 ą 50% perf-profile.self.cycles-pp.deflate_medium
> 2.26 ą 14% -1.2 1.11 ą 44% perf-profile.self.cycles-pp.deflateSetDictionary
> 0.51 ą 12% -0.3 0.24 ą 57% perf-profile.self.cycles-pp.fill_window
> 0.07 ą 7% +0.0 0.10 ą 24% perf-profile.self.cycles-pp._raw_spin_unlock_irqrestore
> 0.07 ą 7% +0.1 0.12 ą 36% perf-profile.self.cycles-pp.syscall_return_via_sysret
> 0.08 ą 12% +0.1 0.14 ą 15% perf-profile.self.cycles-pp.__update_load_avg_se
> 0.06 +0.1 0.13 ą 27% perf-profile.self.cycles-pp._raw_spin_lock_irqsave
> 0.08 ą 25% +0.1 0.15 ą 37% perf-profile.self.cycles-pp.__switch_to
> 0.06 ą 16% +0.1 0.13 ą 29% perf-profile.self.cycles-pp.get_page_from_freelist
> 0.00 +0.1 0.07 ą 29% perf-profile.self.cycles-pp.__switch_to_asm
> 0.05 +0.1 0.13 ą 57% perf-profile.self.cycles-pp.switch_mm_irqs_off
> 0.00 +0.1 0.08 ą 41% perf-profile.self.cycles-pp.interrupt_entry
> 0.00 +0.1 0.08 ą 61% perf-profile.self.cycles-pp.run_timer_softirq
> 0.07 ą 23% +0.1 0.15 ą 43% perf-profile.self.cycles-pp.__hrtimer_next_event_base
> 0.03 ą100% +0.1 0.12 ą 43% perf-profile.self.cycles-pp.update_cfs_group
> 0.08 ą 29% +0.1 0.17 ą 38% perf-profile.self.cycles-pp.queued_spin_lock_slowpath
> 0.00 +0.1 0.09 ą 29% perf-profile.self.cycles-pp.strncpy_from_user
> 0.06 ą 16% +0.1 0.15 ą 24% perf-profile.self.cycles-pp.find_next_bit
> 0.00 +0.1 0.10 ą 35% perf-profile.self.cycles-pp.do_lookup_x
> 0.00 +0.1 0.10 ą 13% perf-profile.self.cycles-pp.kmem_cache_free
> 0.06 ą 16% +0.1 0.16 ą 30% perf-profile.self.cycles-pp.do_idle
> 0.03 ą100% +0.1 0.13 ą 18% perf-profile.self.cycles-pp.entry_SYSCALL_64
> 0.03 ą100% +0.1 0.14 ą 41% perf-profile.self.cycles-pp.update_blocked_averages
> 0.11 ą 18% +0.1 0.22 ą 37% perf-profile.self.cycles-pp.native_sched_clock
> 0.07 ą 14% +0.1 0.18 ą 46% perf-profile.self.cycles-pp.lapic_next_deadline
> 0.00 +0.1 0.12 ą 65% perf-profile.self.cycles-pp.set_next_entity
> 0.12 ą 28% +0.1 0.27 ą 32% perf-profile.self.cycles-pp.cpuidle_enter_state
> 0.15 ą 3% +0.2 0.32 ą 39% perf-profile.self.cycles-pp.io_serial_out
> 0.25 ą 4% +0.2 0.48 ą 19% perf-profile.self.cycles-pp.menu_select
> 0.15 ą 22% +0.3 0.42 ą 46% perf-profile.self.cycles-pp.find_busiest_group
> 0.29 ą 37% +0.4 0.71 ą 42% perf-profile.self.cycles-pp.io_serial_in
> 0.32 ą 7% +0.7 1.02 ą 56% perf-profile.self.cycles-pp.poll_idle
> 17.73 ą 21% +19.1 36.79 ą 25% perf-profile.self.cycles-pp.intel_idle
>
>
>
> phoronix-test-suite.compress-gzip.0.seconds
>
> 8 +-----------------------------------------------------------------------+
> | O O O O O O O O |
> 7 |-+ O O O O O O O O O |
> 6 |-+ + + + |
> | + : + + : + + + : |
> 5 |-+ : : : : :: : : : : |
> | :: : : : :: : : : :: : : |
> 4 |:+: : : : : : : : : : : : : : : : : |
> |: : : : : : : : : + + : : + : : : : : : : |
> 3 |:+: : : : : : : : : : : : : : : : : : : : |
> 2 |:+: : : : : : : : : : : : : : : : : : : : : : : |
> |: : : : : : : : : : : : : : : : : : : : : : : : |
> 1 |-: :: : : : : : : : : :: :: :: : : |
> | : : : : : : : : : : : : |
> 0 +-----------------------------------------------------------------------+
>
>
> [*] bisect-good sample
> [O] bisect-bad sample
>
> ***************************************************************************************************
> lkp-cfl-d1: 12 threads Intel(R) Core(TM) i7-8700 CPU @ 3.20GHz with 8G memory
>
>
> ***************************************************************************************************
> lkp-skl-fpga01: 104 threads Skylake with 192G memory
> =========================================================================================
> compiler/cpufreq_governor/kconfig/rootfs/runtime/size/tbox_group/test/testcase/ucode:
> gcc-7/performance/x86_64-rhel-7.6/debian-x86_64-2019-09-23.cgz/300s/8T/lkp-skl-fpga01/anon-cow-seq/vm-scalability/0x2000064
>
> commit:
> fcf0553db6 ("sched/fair: Remove meaningless imbalance calculation")
> 0b0695f2b3 ("sched/fair: Rework load_balance()")
>
> fcf0553db6f4c793 0b0695f2b34a4afa3f6e9aa1ff0
> ---------------- ---------------------------
> %stddev %change %stddev
> \ | \
> 413301 +3.1% 426103 vm-scalability.median
> 0.04 ą 2% -34.0% 0.03 ą 12% vm-scalability.median_stddev
> 43837589 +2.4% 44902458 vm-scalability.throughput
> 181085 -18.7% 147221 vm-scalability.time.involuntary_context_switches
> 12762365 ą 2% +3.9% 13262025 vm-scalability.time.minor_page_faults
> 7773 +2.9% 7997 vm-scalability.time.percent_of_cpu_this_job_got
> 11449 +1.2% 11589 vm-scalability.time.system_time
> 12024 +4.7% 12584 vm-scalability.time.user_time
> 439194 ą 2% +46.0% 641402 ą 2% vm-scalability.time.voluntary_context_switches
> 1.148e+10 +5.0% 1.206e+10 vm-scalability.workload
> 0.00 ą 54% +0.0 0.00 ą 17% mpstat.cpu.all.iowait%
> 4767597 +52.5% 7268430 ą 41% numa-numastat.node1.local_node
> 4781030 +52.3% 7280347 ą 41% numa-numastat.node1.numa_hit
> 24.75 -9.1% 22.50 ą 2% vmstat.cpu.id
> 37.50 +4.7% 39.25 vmstat.cpu.us
> 6643 ą 3% +15.1% 7647 vmstat.system.cs
> 12220504 +33.4% 16298593 ą 4% cpuidle.C1.time
> 260215 ą 6% +55.3% 404158 ą 3% cpuidle.C1.usage
> 4986034 ą 3% +56.2% 7786811 ą 2% cpuidle.POLL.time
> 145941 ą 3% +61.2% 235218 ą 2% cpuidle.POLL.usage
> 1990 +3.0% 2049 turbostat.Avg_MHz
> 254633 ą 6% +56.7% 398892 ą 4% turbostat.C1
> 0.04 +0.0 0.05 turbostat.C1%
> 309.99 +1.5% 314.75 turbostat.RAMWatt
> 1688 ą 11% +17.4% 1983 ą 5% slabinfo.UNIX.active_objs
> 1688 ą 11% +17.4% 1983 ą 5% slabinfo.UNIX.num_objs
> 2460 ą 3% -15.8% 2072 ą 11% slabinfo.dmaengine-unmap-16.active_objs
> 2460 ą 3% -15.8% 2072 ą 11% slabinfo.dmaengine-unmap-16.num_objs
> 2814 ą 9% +14.6% 3225 ą 4% slabinfo.sock_inode_cache.active_objs
> 2814 ą 9% +14.6% 3225 ą 4% slabinfo.sock_inode_cache.num_objs
> 0.67 ą 5% +0.1 0.73 ą 3% perf-profile.calltrace.cycles-pp.__alloc_pages_nodemask.alloc_pages_vma.do_huge_pmd_wp_page.__handle_mm_fault.handle_mm_fault
> 0.68 ą 6% +0.1 0.74 ą 2% perf-profile.calltrace.cycles-pp.alloc_pages_vma.do_huge_pmd_wp_page.__handle_mm_fault.handle_mm_fault.__do_page_fault
> 0.05 +0.0 0.07 ą 7% perf-profile.children.cycles-pp.schedule
> 0.06 +0.0 0.08 ą 6% perf-profile.children.cycles-pp.__wake_up_common
> 0.06 ą 7% +0.0 0.08 ą 6% perf-profile.children.cycles-pp.wake_up_page_bit
> 0.23 ą 7% +0.0 0.28 ą 5% perf-profile.children.cycles-pp._raw_spin_lock_irqsave
> 0.00 +0.1 0.05 perf-profile.children.cycles-pp.drm_fb_helper_sys_imageblit
> 0.00 +0.1 0.05 perf-profile.children.cycles-pp.sys_imageblit
> 29026 ą 3% -26.7% 21283 ą 44% numa-vmstat.node0.nr_inactive_anon
> 30069 ą 3% -20.5% 23905 ą 26% numa-vmstat.node0.nr_shmem
> 12120 ą 2% -15.5% 10241 ą 12% numa-vmstat.node0.nr_slab_reclaimable
> 29026 ą 3% -26.7% 21283 ą 44% numa-vmstat.node0.nr_zone_inactive_anon
> 4010893 +16.1% 4655889 ą 9% numa-vmstat.node1.nr_active_anon
> 3982581 +16.3% 4632344 ą 9% numa-vmstat.node1.nr_anon_pages
> 6861 +16.1% 7964 ą 8% numa-vmstat.node1.nr_anon_transparent_hugepages
> 2317 ą 42% +336.9% 10125 ą 93% numa-vmstat.node1.nr_inactive_anon
> 6596 ą 4% +18.2% 7799 ą 14% numa-vmstat.node1.nr_kernel_stack
> 9629 ą 8% +66.4% 16020 ą 41% numa-vmstat.node1.nr_shmem
> 7558 ą 3% +26.5% 9561 ą 14% numa-vmstat.node1.nr_slab_reclaimable
> 4010227 +16.1% 4655056 ą 9% numa-vmstat.node1.nr_zone_active_anon
> 2317 ą 42% +336.9% 10125 ą 93% numa-vmstat.node1.nr_zone_inactive_anon
> 2859663 ą 2% +46.2% 4179500 ą 36% numa-vmstat.node1.numa_hit
> 2680260 ą 2% +49.3% 4002218 ą 37% numa-vmstat.node1.numa_local
> 116661 ą 3% -26.3% 86010 ą 44% numa-meminfo.node0.Inactive
> 116192 ą 3% -26.7% 85146 ą 44% numa-meminfo.node0.Inactive(anon)
> 48486 ą 2% -15.5% 40966 ą 12% numa-meminfo.node0.KReclaimable
> 48486 ą 2% -15.5% 40966 ą 12% numa-meminfo.node0.SReclaimable
> 120367 ą 3% -20.5% 95642 ą 26% numa-meminfo.node0.Shmem
> 16210528 +15.2% 18673368 ą 6% numa-meminfo.node1.Active
> 16210394 +15.2% 18673287 ą 6% numa-meminfo.node1.Active(anon)
> 14170064 +15.6% 16379835 ą 7% numa-meminfo.node1.AnonHugePages
> 16113351 +15.3% 18577254 ą 7% numa-meminfo.node1.AnonPages
> 10534 ą 33% +293.8% 41480 ą 92% numa-meminfo.node1.Inactive
> 9262 ą 42% +338.2% 40589 ą 93% numa-meminfo.node1.Inactive(anon)
> 30235 ą 3% +26.5% 38242 ą 14% numa-meminfo.node1.KReclaimable
> 6594 ą 4% +18.3% 7802 ą 14% numa-meminfo.node1.KernelStack
> 17083646 +15.1% 19656922 ą 7% numa-meminfo.node1.MemUsed
> 30235 ą 3% +26.5% 38242 ą 14% numa-meminfo.node1.SReclaimable
> 38540 ą 8% +66.4% 64117 ą 42% numa-meminfo.node1.Shmem
> 106342 +19.8% 127451 ą 11% numa-meminfo.node1.Slab
> 9479688 +4.5% 9905902 proc-vmstat.nr_active_anon
> 9434298 +4.5% 9856978 proc-vmstat.nr_anon_pages
> 16194 +4.3% 16895 proc-vmstat.nr_anon_transparent_hugepages
> 276.75 +3.6% 286.75 proc-vmstat.nr_dirtied
> 3888633 -1.1% 3845882 proc-vmstat.nr_dirty_background_threshold
> 7786774 -1.1% 7701168 proc-vmstat.nr_dirty_threshold
> 39168820 -1.1% 38741444 proc-vmstat.nr_free_pages
> 50391 +1.0% 50904 proc-vmstat.nr_slab_unreclaimable
> 257.50 +3.6% 266.75 proc-vmstat.nr_written
> 9479678 +4.5% 9905895 proc-vmstat.nr_zone_active_anon
> 1501517 -5.9% 1412958 proc-vmstat.numa_hint_faults
> 1075936 -13.1% 934706 proc-vmstat.numa_hint_faults_local
> 17306395 +4.8% 18141722 proc-vmstat.numa_hit
> 5211079 +4.2% 5427541 proc-vmstat.numa_huge_pte_updates
> 17272620 +4.8% 18107691 proc-vmstat.numa_local
> 33774 +0.8% 34031 proc-vmstat.numa_other
> 690793 ą 3% -13.7% 596166 ą 2% proc-vmstat.numa_pages_migrated
> 2.669e+09 +4.2% 2.78e+09 proc-vmstat.numa_pte_updates
> 2.755e+09 +5.6% 2.909e+09 proc-vmstat.pgalloc_normal
> 13573227 ą 2% +3.6% 14060842 proc-vmstat.pgfault
> 2.752e+09 +5.6% 2.906e+09 proc-vmstat.pgfree
> 1.723e+08 ą 2% +14.3% 1.97e+08 ą 8% proc-vmstat.pgmigrate_fail
> 690793 ą 3% -13.7% 596166 ą 2% proc-vmstat.pgmigrate_success
> 5015265 +5.0% 5266730 proc-vmstat.thp_deferred_split_page
> 5019661 +5.0% 5271482 proc-vmstat.thp_fault_alloc
> 18284 ą 62% -79.9% 3681 ą172% sched_debug.cfs_rq:/.MIN_vruntime.avg
> 1901618 ą 62% -89.9% 192494 ą172% sched_debug.cfs_rq:/.MIN_vruntime.max
> 185571 ą 62% -85.8% 26313 ą172% sched_debug.cfs_rq:/.MIN_vruntime.stddev
> 15241 ą 6% -36.6% 9655 ą 6% sched_debug.cfs_rq:/.exec_clock.stddev
> 18284 ą 62% -79.9% 3681 ą172% sched_debug.cfs_rq:/.max_vruntime.avg
> 1901618 ą 62% -89.9% 192494 ą172% sched_debug.cfs_rq:/.max_vruntime.max
> 185571 ą 62% -85.8% 26313 ą172% sched_debug.cfs_rq:/.max_vruntime.stddev
> 898812 ą 7% -31.2% 618552 ą 5% sched_debug.cfs_rq:/.min_vruntime.stddev
> 10.30 ą 12% +34.5% 13.86 ą 6% sched_debug.cfs_rq:/.nr_spread_over.avg
> 34.75 ą 8% +95.9% 68.08 ą 4% sched_debug.cfs_rq:/.nr_spread_over.max
> 9.12 ą 11% +82.3% 16.62 ą 9% sched_debug.cfs_rq:/.nr_spread_over.stddev
> -1470498 -31.9% -1000709 sched_debug.cfs_rq:/.spread0.min
> 899820 ą 7% -31.2% 618970 ą 5% sched_debug.cfs_rq:/.spread0.stddev
> 1589 ą 9% -19.2% 1284 ą 9% sched_debug.cfs_rq:/.util_avg.max
> 0.54 ą 39% +7484.6% 41.08 ą 92% sched_debug.cfs_rq:/.util_est_enqueued.min
> 238.84 ą 8% -33.2% 159.61 ą 26% sched_debug.cfs_rq:/.util_est_enqueued.stddev
> 10787 ą 2% +13.8% 12274 sched_debug.cpu.nr_switches.avg
> 35242 ą 9% +32.3% 46641 ą 25% sched_debug.cpu.nr_switches.max
> 9139 ą 3% +16.4% 10636 sched_debug.cpu.sched_count.avg
> 32025 ą 10% +34.6% 43091 ą 27% sched_debug.cpu.sched_count.max
> 4016 ą 2% +14.7% 4606 ą 5% sched_debug.cpu.sched_count.min
> 2960 +38.3% 4093 sched_debug.cpu.sched_goidle.avg
> 11201 ą 24% +75.8% 19691 ą 26% sched_debug.cpu.sched_goidle.max
> 1099 ą 6% +56.9% 1725 ą 6% sched_debug.cpu.sched_goidle.min
> 1877 ą 10% +32.5% 2487 ą 17% sched_debug.cpu.sched_goidle.stddev
> 4348 ą 3% +19.3% 5188 sched_debug.cpu.ttwu_count.avg
> 17832 ą 11% +78.6% 31852 ą 29% sched_debug.cpu.ttwu_count.max
> 1699 ą 6% +28.2% 2178 ą 7% sched_debug.cpu.ttwu_count.min
> 1357 ą 10% -22.6% 1050 ą 4% sched_debug.cpu.ttwu_local.avg
> 11483 ą 5% -25.0% 8614 ą 15% sched_debug.cpu.ttwu_local.max
> 1979 ą 12% -36.8% 1251 ą 10% sched_debug.cpu.ttwu_local.stddev
> 3.941e+10 +5.0% 4.137e+10 perf-stat.i.branch-instructions
> 0.02 ą 50% -0.0 0.02 ą 5% perf-stat.i.branch-miss-rate%
> 67.94 -3.9 63.99 perf-stat.i.cache-miss-rate%
> 8.329e+08 -1.9% 8.17e+08 perf-stat.i.cache-misses
> 1.224e+09 +4.5% 1.28e+09 perf-stat.i.cache-references
> 6650 ą 3% +15.5% 7678 perf-stat.i.context-switches
> 1.64 -1.8% 1.61 perf-stat.i.cpi
> 2.037e+11 +2.8% 2.095e+11 perf-stat.i.cpu-cycles
> 257.56 -4.0% 247.13 perf-stat.i.cpu-migrations
> 244.94 +4.5% 255.91 perf-stat.i.cycles-between-cache-misses
> 1189446 ą 2% +3.2% 1227527 perf-stat.i.dTLB-load-misses
> 2.669e+10 +4.7% 2.794e+10 perf-stat.i.dTLB-loads
> 0.00 ą 7% -0.0 0.00 perf-stat.i.dTLB-store-miss-rate%
> 337782 +4.5% 353044 perf-stat.i.dTLB-store-misses
> 9.096e+09 +4.7% 9.526e+09 perf-stat.i.dTLB-stores
> 39.50 +2.1 41.64 perf-stat.i.iTLB-load-miss-rate%
> 296305 ą 2% +9.0% 323020 perf-stat.i.iTLB-load-misses
> 1.238e+11 +4.9% 1.299e+11 perf-stat.i.instructions
> 428249 ą 2% -4.4% 409553 perf-stat.i.instructions-per-iTLB-miss
> 0.61 +1.6% 0.62 perf-stat.i.ipc
> 44430 +3.8% 46121 perf-stat.i.minor-faults
> 54.82 +3.9 58.73 perf-stat.i.node-load-miss-rate%
> 68519419 ą 4% -11.7% 60479057 ą 6% perf-stat.i.node-load-misses
> 49879161 ą 3% -20.7% 39554915 ą 4% perf-stat.i.node-loads
> 44428 +3.8% 46119 perf-stat.i.page-faults
> 0.02 -0.0 0.01 ą 5% perf-stat.overall.branch-miss-rate%
> 68.03 -4.2 63.83 perf-stat.overall.cache-miss-rate%
> 1.65 -2.0% 1.61 perf-stat.overall.cpi
> 244.61 +4.8% 256.41 perf-stat.overall.cycles-between-cache-misses
> 30.21 +2.2 32.38 perf-stat.overall.iTLB-load-miss-rate%
> 417920 ą 2% -3.7% 402452 perf-stat.overall.instructions-per-iTLB-miss
> 0.61 +2.1% 0.62 perf-stat.overall.ipc
> 57.84 +2.6 60.44 perf-stat.overall.node-load-miss-rate%
> 3.925e+10 +5.1% 4.124e+10 perf-stat.ps.branch-instructions
> 8.295e+08 -1.8% 8.144e+08 perf-stat.ps.cache-misses
> 1.219e+09 +4.6% 1.276e+09 perf-stat.ps.cache-references
> 6625 ą 3% +15.4% 7648 perf-stat.ps.context-switches
> 2.029e+11 +2.9% 2.088e+11 perf-stat.ps.cpu-cycles
> 256.82 -4.2% 246.09 perf-stat.ps.cpu-migrations
> 1184763 ą 2% +3.3% 1223366 perf-stat.ps.dTLB-load-misses
> 2.658e+10 +4.8% 2.786e+10 perf-stat.ps.dTLB-loads
> 336658 +4.5% 351710 perf-stat.ps.dTLB-store-misses
> 9.059e+09 +4.8% 9.497e+09 perf-stat.ps.dTLB-stores
> 295140 ą 2% +9.0% 321824 perf-stat.ps.iTLB-load-misses
> 1.233e+11 +5.0% 1.295e+11 perf-stat.ps.instructions
> 44309 +3.7% 45933 perf-stat.ps.minor-faults
> 68208972 ą 4% -11.6% 60272675 ą 6% perf-stat.ps.node-load-misses
> 49689740 ą 3% -20.7% 39401789 ą 4% perf-stat.ps.node-loads
> 44308 +3.7% 45932 perf-stat.ps.page-faults
> 3.732e+13 +5.1% 3.922e+13 perf-stat.total.instructions
> 14949 ą 2% +14.5% 17124 ą 11% softirqs.CPU0.SCHED
> 9940 +37.8% 13700 ą 24% softirqs.CPU1.SCHED
> 9370 ą 2% +28.2% 12014 ą 16% softirqs.CPU10.SCHED
> 17637 ą 2% -16.5% 14733 ą 16% softirqs.CPU101.SCHED
> 17846 ą 3% -17.4% 14745 ą 16% softirqs.CPU103.SCHED
> 9552 +24.7% 11916 ą 17% softirqs.CPU11.SCHED
> 9210 ą 5% +27.9% 11784 ą 16% softirqs.CPU12.SCHED
> 9378 ą 3% +27.7% 11974 ą 16% softirqs.CPU13.SCHED
> 9164 ą 2% +29.4% 11856 ą 18% softirqs.CPU14.SCHED
> 9215 +21.2% 11170 ą 19% softirqs.CPU15.SCHED
> 9118 ą 2% +29.1% 11772 ą 16% softirqs.CPU16.SCHED
> 9413 +29.2% 12165 ą 18% softirqs.CPU17.SCHED
> 9309 ą 2% +29.9% 12097 ą 17% softirqs.CPU18.SCHED
> 9423 +26.1% 11880 ą 15% softirqs.CPU19.SCHED
> 9010 ą 7% +37.8% 12420 ą 18% softirqs.CPU2.SCHED
> 9382 ą 3% +27.0% 11916 ą 15% softirqs.CPU20.SCHED
> 9102 ą 4% +30.0% 11830 ą 16% softirqs.CPU21.SCHED
> 9543 ą 3% +23.4% 11780 ą 18% softirqs.CPU22.SCHED
> 8998 ą 5% +29.2% 11630 ą 18% softirqs.CPU24.SCHED
> 9254 ą 2% +23.9% 11462 ą 19% softirqs.CPU25.SCHED
> 18450 ą 4% -16.9% 15341 ą 16% softirqs.CPU26.SCHED
> 17551 ą 4% -14.8% 14956 ą 13% softirqs.CPU27.SCHED
> 17575 ą 4% -14.6% 15010 ą 14% softirqs.CPU28.SCHED
> 17515 ą 5% -14.2% 15021 ą 13% softirqs.CPU29.SCHED
> 17715 ą 2% -16.1% 14856 ą 13% softirqs.CPU30.SCHED
> 17754 ą 4% -16.1% 14904 ą 13% softirqs.CPU31.SCHED
> 17675 ą 2% -17.0% 14679 ą 21% softirqs.CPU32.SCHED
> 17625 ą 2% -16.0% 14813 ą 13% softirqs.CPU34.SCHED
> 17619 ą 2% -14.7% 15024 ą 14% softirqs.CPU35.SCHED
> 17887 ą 3% -17.0% 14841 ą 14% softirqs.CPU36.SCHED
> 17658 ą 3% -16.3% 14771 ą 12% softirqs.CPU38.SCHED
> 17501 ą 2% -15.3% 14816 ą 14% softirqs.CPU39.SCHED
> 9360 ą 2% +25.4% 11740 ą 14% softirqs.CPU4.SCHED
> 17699 ą 4% -16.2% 14827 ą 14% softirqs.CPU42.SCHED
> 17580 ą 3% -16.5% 14679 ą 15% softirqs.CPU43.SCHED
> 17658 ą 3% -17.1% 14644 ą 14% softirqs.CPU44.SCHED
> 17452 ą 4% -14.0% 15001 ą 15% softirqs.CPU46.SCHED
> 17599 ą 4% -17.4% 14544 ą 14% softirqs.CPU47.SCHED
> 17792 ą 3% -16.5% 14864 ą 14% softirqs.CPU48.SCHED
> 17333 ą 2% -16.7% 14445 ą 14% softirqs.CPU49.SCHED
> 9483 +32.3% 12547 ą 24% softirqs.CPU5.SCHED
> 17842 ą 3% -15.9% 14997 ą 16% softirqs.CPU51.SCHED
> 9051 ą 2% +23.3% 11160 ą 13% softirqs.CPU52.SCHED
> 9385 ą 3% +25.2% 11752 ą 16% softirqs.CPU53.SCHED
> 9446 ą 6% +24.9% 11798 ą 14% softirqs.CPU54.SCHED
> 10006 ą 6% +22.4% 12249 ą 14% softirqs.CPU55.SCHED
> 9657 +22.0% 11780 ą 16% softirqs.CPU57.SCHED
> 9399 +27.5% 11980 ą 15% softirqs.CPU58.SCHED
> 9234 ą 3% +27.7% 11795 ą 14% softirqs.CPU59.SCHED
> 9726 ą 6% +24.0% 12062 ą 16% softirqs.CPU6.SCHED
> 9165 ą 2% +23.7% 11342 ą 14% softirqs.CPU60.SCHED
> 9357 ą 2% +25.8% 11774 ą 15% softirqs.CPU61.SCHED
> 9406 ą 3% +25.2% 11780 ą 16% softirqs.CPU62.SCHED
> 9489 +23.2% 11688 ą 15% softirqs.CPU63.SCHED
> 9399 ą 2% +23.5% 11604 ą 16% softirqs.CPU65.SCHED
> 8950 ą 2% +31.6% 11774 ą 16% softirqs.CPU66.SCHED
> 9260 +21.7% 11267 ą 19% softirqs.CPU67.SCHED
> 9187 +27.1% 11672 ą 17% softirqs.CPU68.SCHED
> 9443 ą 2% +25.5% 11847 ą 17% softirqs.CPU69.SCHED
> 9144 ą 3% +28.0% 11706 ą 16% softirqs.CPU7.SCHED
> 9276 ą 2% +28.0% 11871 ą 17% softirqs.CPU70.SCHED
> 9494 +21.4% 11526 ą 14% softirqs.CPU71.SCHED
> 9124 ą 3% +27.8% 11657 ą 17% softirqs.CPU72.SCHED
> 9189 ą 3% +25.9% 11568 ą 16% softirqs.CPU73.SCHED
> 9392 ą 2% +23.7% 11619 ą 16% softirqs.CPU74.SCHED
> 17821 ą 3% -14.7% 15197 ą 17% softirqs.CPU78.SCHED
> 17581 ą 2% -15.7% 14827 ą 15% softirqs.CPU79.SCHED
> 9123 +28.2% 11695 ą 15% softirqs.CPU8.SCHED
> 17524 ą 2% -16.7% 14601 ą 14% softirqs.CPU80.SCHED
> 17644 ą 3% -16.2% 14782 ą 14% softirqs.CPU81.SCHED
> 17705 ą 3% -18.6% 14414 ą 22% softirqs.CPU84.SCHED
> 17679 ą 2% -14.1% 15185 ą 11% softirqs.CPU85.SCHED
> 17434 ą 3% -15.5% 14724 ą 14% softirqs.CPU86.SCHED
> 17409 ą 2% -15.0% 14794 ą 13% softirqs.CPU87.SCHED
> 17470 ą 3% -15.7% 14730 ą 13% softirqs.CPU88.SCHED
> 17748 ą 4% -17.1% 14721 ą 12% softirqs.CPU89.SCHED
> 9323 +28.0% 11929 ą 17% softirqs.CPU9.SCHED
> 17471 ą 2% -16.9% 14525 ą 13% softirqs.CPU90.SCHED
> 17900 ą 3% -17.0% 14850 ą 14% softirqs.CPU94.SCHED
> 17599 ą 4% -17.4% 14544 ą 15% softirqs.CPU95.SCHED
> 17697 ą 4% -17.7% 14569 ą 13% softirqs.CPU96.SCHED
> 17561 ą 3% -15.1% 14901 ą 13% softirqs.CPU97.SCHED
> 17404 ą 3% -16.1% 14601 ą 13% softirqs.CPU98.SCHED
> 17802 ą 3% -19.4% 14344 ą 15% softirqs.CPU99.SCHED
> 1310 ą 10% -17.0% 1088 ą 5% interrupts.CPU1.RES:Rescheduling_interrupts
> 3427 +13.3% 3883 ą 9% interrupts.CPU10.CAL:Function_call_interrupts
> 736.50 ą 20% +34.4% 989.75 ą 17% interrupts.CPU100.RES:Rescheduling_interrupts
> 3421 ą 3% +14.6% 3921 ą 9% interrupts.CPU101.CAL:Function_call_interrupts
> 4873 ą 8% +16.2% 5662 ą 7% interrupts.CPU101.NMI:Non-maskable_interrupts
> 4873 ą 8% +16.2% 5662 ą 7% interrupts.CPU101.PMI:Performance_monitoring_interrupts
> 629.50 ą 19% +83.2% 1153 ą 46% interrupts.CPU101.RES:Rescheduling_interrupts
> 661.75 ą 14% +25.7% 832.00 ą 13% interrupts.CPU102.RES:Rescheduling_interrupts
> 4695 ą 5% +15.5% 5420 ą 9% interrupts.CPU103.NMI:Non-maskable_interrupts
> 4695 ą 5% +15.5% 5420 ą 9% interrupts.CPU103.PMI:Performance_monitoring_interrupts
> 3460 +12.1% 3877 ą 9% interrupts.CPU11.CAL:Function_call_interrupts
> 691.50 ą 7% +41.0% 975.00 ą 32% interrupts.CPU19.RES:Rescheduling_interrupts
> 3413 ą 2% +13.4% 3870 ą 10% interrupts.CPU20.CAL:Function_call_interrupts
> 3413 ą 2% +13.4% 3871 ą 10% interrupts.CPU22.CAL:Function_call_interrupts
> 863.00 ą 36% +45.3% 1254 ą 24% interrupts.CPU23.RES:Rescheduling_interrupts
> 659.75 ą 12% +83.4% 1209 ą 20% interrupts.CPU26.RES:Rescheduling_interrupts
> 615.00 ą 10% +87.8% 1155 ą 14% interrupts.CPU27.RES:Rescheduling_interrupts
> 663.75 ą 5% +67.9% 1114 ą 7% interrupts.CPU28.RES:Rescheduling_interrupts
> 3421 ą 4% +13.4% 3879 ą 9% interrupts.CPU29.CAL:Function_call_interrupts
> 805.25 ą 16% +33.0% 1071 ą 15% interrupts.CPU29.RES:Rescheduling_interrupts
> 3482 ą 3% +11.0% 3864 ą 8% interrupts.CPU3.CAL:Function_call_interrupts
> 819.75 ą 19% +48.4% 1216 ą 12% interrupts.CPU30.RES:Rescheduling_interrupts
> 777.25 ą 8% +31.6% 1023 ą 6% interrupts.CPU31.RES:Rescheduling_interrupts
> 844.50 ą 25% +41.7% 1196 ą 20% interrupts.CPU32.RES:Rescheduling_interrupts
> 722.75 ą 14% +94.2% 1403 ą 26% interrupts.CPU33.RES:Rescheduling_interrupts
> 3944 ą 25% +36.8% 5394 ą 9% interrupts.CPU34.NMI:Non-maskable_interrupts
> 3944 ą 25% +36.8% 5394 ą 9% interrupts.CPU34.PMI:Performance_monitoring_interrupts
> 781.75 ą 9% +45.3% 1136 ą 27% interrupts.CPU34.RES:Rescheduling_interrupts
> 735.50 ą 9% +33.3% 980.75 ą 4% interrupts.CPU35.RES:Rescheduling_interrupts
> 691.75 ą 10% +41.6% 979.50 ą 13% interrupts.CPU36.RES:Rescheduling_interrupts
> 727.00 ą 16% +47.7% 1074 ą 15% interrupts.CPU37.RES:Rescheduling_interrupts
> 4413 ą 7% +24.9% 5511 ą 9% interrupts.CPU38.NMI:Non-maskable_interrupts
> 4413 ą 7% +24.9% 5511 ą 9% interrupts.CPU38.PMI:Performance_monitoring_interrupts
> 708.75 ą 25% +62.6% 1152 ą 22% interrupts.CPU38.RES:Rescheduling_interrupts
> 666.50 ą 7% +57.8% 1052 ą 13% interrupts.CPU39.RES:Rescheduling_interrupts
> 765.75 ą 11% +25.2% 958.75 ą 14% interrupts.CPU4.RES:Rescheduling_interrupts
> 3395 ą 2% +15.1% 3908 ą 10% interrupts.CPU40.CAL:Function_call_interrupts
> 770.00 ą 16% +45.3% 1119 ą 18% interrupts.CPU40.RES:Rescheduling_interrupts
> 740.50 ą 26% +61.9% 1198 ą 19% interrupts.CPU41.RES:Rescheduling_interrupts
> 3459 ą 2% +12.9% 3905 ą 11% interrupts.CPU42.CAL:Function_call_interrupts
> 4530 ą 5% +22.8% 5564 ą 9% interrupts.CPU42.NMI:Non-maskable_interrupts
> 4530 ą 5% +22.8% 5564 ą 9% interrupts.CPU42.PMI:Performance_monitoring_interrupts
> 3330 ą 25% +60.0% 5328 ą 10% interrupts.CPU44.NMI:Non-maskable_interrupts
> 3330 ą 25% +60.0% 5328 ą 10% interrupts.CPU44.PMI:Performance_monitoring_interrupts
> 686.25 ą 9% +48.4% 1018 ą 10% interrupts.CPU44.RES:Rescheduling_interrupts
> 702.00 ą 15% +38.6% 973.25 ą 5% interrupts.CPU45.RES:Rescheduling_interrupts
> 4742 ą 7% +19.3% 5657 ą 8% interrupts.CPU46.NMI:Non-maskable_interrupts
> 4742 ą 7% +19.3% 5657 ą 8% interrupts.CPU46.PMI:Performance_monitoring_interrupts
> 732.75 ą 6% +51.9% 1113 ą 7% interrupts.CPU46.RES:Rescheduling_interrupts
> 775.50 ą 17% +41.3% 1095 ą 6% interrupts.CPU47.RES:Rescheduling_interrupts
> 670.75 ą 5% +60.7% 1078 ą 6% interrupts.CPU48.RES:Rescheduling_interrupts
> 4870 ą 8% +16.5% 5676 ą 7% interrupts.CPU49.NMI:Non-maskable_interrupts
> 4870 ą 8% +16.5% 5676 ą 7% interrupts.CPU49.PMI:Performance_monitoring_interrupts
> 694.75 ą 12% +25.8% 874.00 ą 11% interrupts.CPU49.RES:Rescheduling_interrupts
> 686.00 ą 9% +52.0% 1042 ą 20% interrupts.CPU50.RES:Rescheduling_interrupts
> 3361 +17.2% 3938 ą 9% interrupts.CPU51.CAL:Function_call_interrupts
> 4707 ą 6% +16.0% 5463 ą 8% interrupts.CPU51.NMI:Non-maskable_interrupts
> 4707 ą 6% +16.0% 5463 ą 8% interrupts.CPU51.PMI:Performance_monitoring_interrupts
> 638.75 ą 12% +28.6% 821.25 ą 15% interrupts.CPU54.RES:Rescheduling_interrupts
> 677.50 ą 8% +51.8% 1028 ą 29% interrupts.CPU58.RES:Rescheduling_interrupts
> 3465 ą 2% +12.0% 3880 ą 9% interrupts.CPU6.CAL:Function_call_interrupts
> 641.25 ą 2% +26.1% 808.75 ą 10% interrupts.CPU60.RES:Rescheduling_interrupts
> 599.75 ą 2% +45.6% 873.50 ą 8% interrupts.CPU62.RES:Rescheduling_interrupts
> 661.50 ą 9% +52.4% 1008 ą 27% interrupts.CPU63.RES:Rescheduling_interrupts
> 611.00 ą 12% +31.1% 801.00 ą 13% interrupts.CPU69.RES:Rescheduling_interrupts
> 3507 ą 2% +10.8% 3888 ą 9% interrupts.CPU7.CAL:Function_call_interrupts
> 664.00 ą 5% +32.3% 878.50 ą 23% interrupts.CPU70.RES:Rescheduling_interrupts
> 5780 ą 9% -38.8% 3540 ą 37% interrupts.CPU73.NMI:Non-maskable_interrupts
> 5780 ą 9% -38.8% 3540 ą 37% interrupts.CPU73.PMI:Performance_monitoring_interrupts
> 5787 ą 9% -26.7% 4243 ą 28% interrupts.CPU76.NMI:Non-maskable_interrupts
> 5787 ą 9% -26.7% 4243 ą 28% interrupts.CPU76.PMI:Performance_monitoring_interrupts
> 751.50 ą 15% +88.0% 1413 ą 37% interrupts.CPU78.RES:Rescheduling_interrupts
> 725.50 ą 12% +82.9% 1327 ą 36% interrupts.CPU79.RES:Rescheduling_interrupts
> 714.00 ą 18% +33.2% 951.00 ą 15% interrupts.CPU80.RES:Rescheduling_interrupts
> 706.25 ą 19% +55.6% 1098 ą 27% interrupts.CPU82.RES:Rescheduling_interrupts
> 4524 ą 6% +19.6% 5409 ą 8% interrupts.CPU83.NMI:Non-maskable_interrupts
> 4524 ą 6% +19.6% 5409 ą 8% interrupts.CPU83.PMI:Performance_monitoring_interrupts
> 666.75 ą 15% +37.3% 915.50 ą 4% interrupts.CPU83.RES:Rescheduling_interrupts
> 782.50 ą 26% +57.6% 1233 ą 21% interrupts.CPU84.RES:Rescheduling_interrupts
> 622.75 ą 12% +77.8% 1107 ą 17% interrupts.CPU85.RES:Rescheduling_interrupts
> 3465 ą 3% +13.5% 3933 ą 9% interrupts.CPU86.CAL:Function_call_interrupts
> 714.75 ą 14% +47.0% 1050 ą 10% interrupts.CPU86.RES:Rescheduling_interrupts
> 3519 ą 2% +11.7% 3929 ą 9% interrupts.CPU87.CAL:Function_call_interrupts
> 582.75 ą 10% +54.2% 898.75 ą 11% interrupts.CPU87.RES:Rescheduling_interrupts
> 713.00 ą 10% +36.6% 974.25 ą 11% interrupts.CPU88.RES:Rescheduling_interrupts
> 690.50 ą 13% +53.0% 1056 ą 13% interrupts.CPU89.RES:Rescheduling_interrupts
> 3477 +11.0% 3860 ą 8% interrupts.CPU9.CAL:Function_call_interrupts
> 684.50 ą 14% +39.7% 956.25 ą 11% interrupts.CPU90.RES:Rescheduling_interrupts
> 3946 ą 21% +39.8% 5516 ą 10% interrupts.CPU91.NMI:Non-maskable_interrupts
> 3946 ą 21% +39.8% 5516 ą 10% interrupts.CPU91.PMI:Performance_monitoring_interrupts
> 649.00 ą 13% +54.3% 1001 ą 6% interrupts.CPU91.RES:Rescheduling_interrupts
> 674.25 ą 21% +39.5% 940.25 ą 11% interrupts.CPU92.RES:Rescheduling_interrupts
> 3971 ą 26% +41.2% 5606 ą 8% interrupts.CPU94.NMI:Non-maskable_interrupts
> 3971 ą 26% +41.2% 5606 ą 8% interrupts.CPU94.PMI:Performance_monitoring_interrupts
> 4129 ą 22% +33.2% 5499 ą 9% interrupts.CPU95.NMI:Non-maskable_interrupts
> 4129 ą 22% +33.2% 5499 ą 9% interrupts.CPU95.PMI:Performance_monitoring_interrupts
> 685.75 ą 14% +38.0% 946.50 ą 9% interrupts.CPU96.RES:Rescheduling_interrupts
> 4630 ą 11% +18.3% 5477 ą 8% interrupts.CPU97.NMI:Non-maskable_interrupts
> 4630 ą 11% +18.3% 5477 ą 8% interrupts.CPU97.PMI:Performance_monitoring_interrupts
> 4835 ą 9% +16.3% 5622 ą 9% interrupts.CPU98.NMI:Non-maskable_interrupts
> 4835 ą 9% +16.3% 5622 ą 9% interrupts.CPU98.PMI:Performance_monitoring_interrupts
> 596.25 ą 11% +81.8% 1083 ą 9% interrupts.CPU98.RES:Rescheduling_interrupts
> 674.75 ą 17% +43.7% 969.50 ą 5% interrupts.CPU99.RES:Rescheduling_interrupts
> 78.25 ą 13% +21.4% 95.00 ą 10% interrupts.IWI:IRQ_work_interrupts
> 85705 ą 6% +26.0% 107990 ą 6% interrupts.RES:Rescheduling_interrupts
>
>
>
> ***************************************************************************************************
> lkp-bdw-ep6: 88 threads Intel(R) Xeon(R) CPU E5-2699 v4 @ 2.20GHz with 128G memory
> =========================================================================================
> class/compiler/cpufreq_governor/disk/kconfig/nr_threads/rootfs/sc_pid_max/tbox_group/testcase/testtime/ucode:
> scheduler/gcc-7/performance/1HDD/x86_64-rhel-7.6/100%/debian-x86_64-2019-11-14.cgz/4194304/lkp-bdw-ep6/stress-ng/1s/0xb000038
>
> commit:
> fcf0553db6 ("sched/fair: Remove meaningless imbalance calculation")
> 0b0695f2b3 ("sched/fair: Rework load_balance()")
>
> fcf0553db6f4c793 0b0695f2b34a4afa3f6e9aa1ff0
> ---------------- ---------------------------
> %stddev %change %stddev
> \ | \
> 887157 ą 4% -23.1% 682080 ą 3% stress-ng.fault.ops
> 887743 ą 4% -23.1% 682337 ą 3% stress-ng.fault.ops_per_sec
> 9537184 ą 10% -21.2% 7518352 ą 14% stress-ng.hrtimers.ops_per_sec
> 360922 ą 13% -21.1% 284734 ą 6% stress-ng.kill.ops
> 361115 ą 13% -21.1% 284810 ą 6% stress-ng.kill.ops_per_sec
> 23260649 -26.9% 17006477 ą 24% stress-ng.mq.ops
> 23255884 -26.9% 17004540 ą 24% stress-ng.mq.ops_per_sec
> 3291588 ą 3% +42.5% 4690316 ą 2% stress-ng.schedpolicy.ops
> 3327913 ą 3% +41.5% 4709770 ą 2% stress-ng.schedpolicy.ops_per_sec
> 48.14 -2.2% 47.09 stress-ng.time.elapsed_time
> 48.14 -2.2% 47.09 stress-ng.time.elapsed_time.max
> 5480 +3.7% 5681 stress-ng.time.percent_of_cpu_this_job_got
> 2249 +1.3% 2278 stress-ng.time.system_time
> 902759 ą 4% -22.6% 698616 ą 3% proc-vmstat.unevictable_pgs_culled
> 98767954 ą 7% +16.4% 1.15e+08 ą 7% cpuidle.C1.time
> 1181676 ą 12% -43.2% 671022 ą 37% cpuidle.C6.usage
> 2.21 ą 7% +0.4 2.62 ą 10% turbostat.C1%
> 1176838 ą 12% -43.2% 668921 ą 37% turbostat.C6
> 3961223 ą 4% +12.8% 4469620 ą 5% vmstat.memory.cache
> 439.50 ą 3% +14.7% 504.00 ą 9% vmstat.procs.r
> 0.42 ą 7% -15.6% 0.35 ą 13% sched_debug.cfs_rq:/.nr_running.stddev
> 0.00 ą 4% -18.1% 0.00 ą 16% sched_debug.cpu.next_balance.stddev
> 0.41 ą 7% -15.1% 0.35 ą 13% sched_debug.cpu.nr_running.stddev
> 9367 ą 9% -12.8% 8166 ą 2% softirqs.CPU1.SCHED
> 35143 ą 6% -12.0% 30930 ą 2% softirqs.CPU22.TIMER
> 31997 ą 4% -7.5% 29595 ą 2% softirqs.CPU27.TIMER
> 3.64 ą173% -100.0% 0.00 iostat.sda.await.max
> 3.64 ą173% -100.0% 0.00 iostat.sda.r_await.max
> 3.90 ą173% -100.0% 0.00 iostat.sdc.await.max
> 3.90 ą173% -100.0% 0.00 iostat.sdc.r_await.max
> 12991737 ą 10% +61.5% 20979642 ą 8% numa-numastat.node0.local_node
> 13073590 ą 10% +61.1% 21059448 ą 8% numa-numastat.node0.numa_hit
> 20903562 ą 3% -32.2% 14164789 ą 3% numa-numastat.node1.local_node
> 20993788 ą 3% -32.1% 14245636 ą 3% numa-numastat.node1.numa_hit
> 90229 ą 4% -10.4% 80843 ą 9% numa-numastat.node1.other_node
> 50.75 ą 90% +1732.0% 929.75 ą147% interrupts.CPU23.IWI:IRQ_work_interrupts
> 40391 ą 59% -57.0% 17359 ą 11% interrupts.CPU24.RES:Rescheduling_interrupts
> 65670 ą 11% -48.7% 33716 ą 54% interrupts.CPU42.RES:Rescheduling_interrupts
> 42201 ą 46% -57.1% 18121 ą 35% interrupts.CPU49.RES:Rescheduling_interrupts
> 293869 ą 44% +103.5% 598082 ą 23% interrupts.CPU52.LOC:Local_timer_interrupts
> 17367 ą 8% +120.5% 38299 ą 44% interrupts.CPU55.RES:Rescheduling_interrupts
> 1.127e+08 +3.8% 1.17e+08 ą 2% perf-stat.i.branch-misses
> 11.10 +1.2 12.26 ą 6% perf-stat.i.cache-miss-rate%
> 4.833e+10 ą 3% +4.7% 5.06e+10 perf-stat.i.instructions
> 15009442 ą 4% +14.3% 17150138 ą 3% perf-stat.i.node-load-misses
> 47.12 ą 5% +3.2 50.37 ą 5% perf-stat.i.node-store-miss-rate%
> 6016833 ą 7% +17.0% 7036803 ą 3% perf-stat.i.node-store-misses
> 1.044e+10 ą 2% +4.0% 1.086e+10 perf-stat.ps.branch-instructions
> 1.364e+10 ą 3% +4.0% 1.418e+10 perf-stat.ps.dTLB-loads
> 4.804e+10 ą 2% +4.1% 5.003e+10 perf-stat.ps.instructions
> 14785608 ą 5% +11.3% 16451530 ą 3% perf-stat.ps.node-load-misses
> 5968712 ą 7% +13.4% 6769847 ą 3% perf-stat.ps.node-store-misses
> 13588 ą 4% +29.4% 17585 ą 9% slabinfo.Acpi-State.active_objs
> 13588 ą 4% +29.4% 17585 ą 9% slabinfo.Acpi-State.num_objs
> 20859 ą 3% -8.6% 19060 ą 4% slabinfo.kmalloc-192.num_objs
> 488.00 ą 25% +41.0% 688.00 ą 5% slabinfo.kmalloc-rcl-128.active_objs
> 488.00 ą 25% +41.0% 688.00 ą 5% slabinfo.kmalloc-rcl-128.num_objs
> 39660 ą 3% +11.8% 44348 ą 2% slabinfo.radix_tree_node.active_objs
> 44284 ą 3% +12.3% 49720 slabinfo.radix_tree_node.num_objs
> 5811 ą 15% +16.1% 6746 ą 14% slabinfo.sighand_cache.active_objs
> 402.00 ą 15% +17.5% 472.50 ą 14% slabinfo.sighand_cache.active_slabs
> 6035 ą 15% +17.5% 7091 ą 14% slabinfo.sighand_cache.num_objs
> 402.00 ą 15% +17.5% 472.50 ą 14% slabinfo.sighand_cache.num_slabs
> 10282 ą 10% +12.9% 11604 ą 9% slabinfo.signal_cache.active_objs
> 11350 ą 10% +12.8% 12808 ą 9% slabinfo.signal_cache.num_objs
> 732920 ą 9% +162.0% 1919987 ą 11% numa-meminfo.node0.Active
> 732868 ą 9% +162.0% 1919814 ą 11% numa-meminfo.node0.Active(anon)
> 545019 ą 6% +61.0% 877443 ą 17% numa-meminfo.node0.AnonHugePages
> 695015 ą 10% +46.8% 1020150 ą 14% numa-meminfo.node0.AnonPages
> 638322 ą 4% +448.2% 3499399 ą 5% numa-meminfo.node0.FilePages
> 81008 ą 14% +2443.4% 2060329 ą 3% numa-meminfo.node0.Inactive
> 80866 ą 14% +2447.4% 2060022 ą 3% numa-meminfo.node0.Inactive(anon)
> 86504 ą 10% +2287.3% 2065084 ą 3% numa-meminfo.node0.Mapped
> 2010104 +160.8% 5242366 ą 5% numa-meminfo.node0.MemUsed
> 16453 ą 15% +159.2% 42640 numa-meminfo.node0.PageTables
> 112769 ą 13% +2521.1% 2955821 ą 7% numa-meminfo.node0.Shmem
> 1839527 ą 4% -60.2% 732645 ą 23% numa-meminfo.node1.Active
> 1839399 ą 4% -60.2% 732637 ą 23% numa-meminfo.node1.Active(anon)
> 982237 ą 7% -45.9% 531445 ą 27% numa-meminfo.node1.AnonHugePages
> 1149348 ą 8% -41.2% 676067 ą 25% numa-meminfo.node1.AnonPages
> 3170649 ą 4% -77.2% 723230 ą 7% numa-meminfo.node1.FilePages
> 1960718 ą 4% -91.8% 160773 ą 31% numa-meminfo.node1.Inactive
> 1960515 ą 4% -91.8% 160722 ą 31% numa-meminfo.node1.Inactive(anon)
> 118489 ą 11% -20.2% 94603 ą 3% numa-meminfo.node1.KReclaimable
> 1966065 ą 4% -91.5% 166789 ą 29% numa-meminfo.node1.Mapped
> 5034310 ą 3% -60.2% 2003121 ą 9% numa-meminfo.node1.MemUsed
> 42684 ą 10% -64.2% 15283 ą 21% numa-meminfo.node1.PageTables
> 118489 ą 11% -20.2% 94603 ą 3% numa-meminfo.node1.SReclaimable
> 2644708 ą 5% -91.9% 214268 ą 24% numa-meminfo.node1.Shmem
> 147513 ą 20% +244.2% 507737 ą 7% numa-vmstat.node0.nr_active_anon
> 137512 ą 21% +105.8% 282999 ą 3% numa-vmstat.node0.nr_anon_pages
> 210.25 ą 33% +124.7% 472.50 ą 11% numa-vmstat.node0.nr_anon_transparent_hugepages
> 158008 ą 4% +454.7% 876519 ą 6% numa-vmstat.node0.nr_file_pages
> 18416 ą 27% +2711.4% 517747 ą 3% numa-vmstat.node0.nr_inactive_anon
> 26255 ą 22% +34.3% 35251 ą 10% numa-vmstat.node0.nr_kernel_stack
> 19893 ą 23% +2509.5% 519129 ą 3% numa-vmstat.node0.nr_mapped
> 3928 ą 22% +179.4% 10976 ą 4% numa-vmstat.node0.nr_page_table_pages
> 26623 ą 18% +2681.9% 740635 ą 7% numa-vmstat.node0.nr_shmem
> 147520 ą 20% +244.3% 507885 ą 7% numa-vmstat.node0.nr_zone_active_anon
> 18415 ą 27% +2711.5% 517739 ą 3% numa-vmstat.node0.nr_zone_inactive_anon
> 6937137 ą 8% +55.9% 10814957 ą 7% numa-vmstat.node0.numa_hit
> 6860210 ą 8% +56.6% 10739902 ą 7% numa-vmstat.node0.numa_local
> 425559 ą 13% -52.9% 200300 ą 17% numa-vmstat.node1.nr_active_anon
> 786341 ą 4% -76.6% 183664 ą 7% numa-vmstat.node1.nr_file_pages
> 483646 ą 4% -90.8% 44606 ą 29% numa-vmstat.node1.nr_inactive_anon
> 485120 ą 4% -90.5% 46130 ą 27% numa-vmstat.node1.nr_mapped
> 10471 ą 6% -61.3% 4048 ą 18% numa-vmstat.node1.nr_page_table_pages
> 654852 ą 5% -91.4% 56439 ą 25% numa-vmstat.node1.nr_shmem
> 29681 ą 11% -20.3% 23669 ą 3% numa-vmstat.node1.nr_slab_reclaimable
> 425556 ą 13% -52.9% 200359 ą 17% numa-vmstat.node1.nr_zone_active_anon
> 483649 ą 4% -90.8% 44600 ą 29% numa-vmstat.node1.nr_zone_inactive_anon
> 10527487 ą 5% -31.3% 7233899 ą 6% numa-vmstat.node1.numa_hit
> 10290625 ą 5% -31.9% 7006050 ą 7% numa-vmstat.node1.numa_local
>
>
>
> ***************************************************************************************************
> lkp-csl-2sp5: 96 threads Intel(R) Xeon(R) Gold 6252 CPU @ 2.10GHz with 192G memory
> =========================================================================================
> class/compiler/cpufreq_governor/disk/kconfig/nr_threads/rootfs/tbox_group/testcase/testtime/ucode:
> interrupt/gcc-7/performance/1HDD/x86_64-fedora-25/100%/debian-x86_64-2019-11-14.cgz/lkp-csl-2sp5/stress-ng/1s/0x500002c
>
> commit:
> fcf0553db6 ("sched/fair: Remove meaningless imbalance calculation")
> 0b0695f2b3 ("sched/fair: Rework load_balance()")
>
> fcf0553db6f4c793 0b0695f2b34a4afa3f6e9aa1ff0
> ---------------- ---------------------------
> %stddev %change %stddev
> \ | \
> 6684836 -33.3% 4457559 ą 4% stress-ng.schedpolicy.ops
> 6684766 -33.3% 4457633 ą 4% stress-ng.schedpolicy.ops_per_sec
> 19978129 -28.8% 14231813 ą 16% stress-ng.time.involuntary_context_switches
> 82.49 ą 2% -5.2% 78.23 stress-ng.time.user_time
> 106716 ą 29% +40.3% 149697 ą 2% meminfo.max_used_kB
> 4.07 ą 22% +1.2 5.23 ą 5% mpstat.cpu.all.irq%
> 2721317 ą 10% +66.5% 4531100 ą 22% cpuidle.POLL.time
> 71470 ą 18% +41.1% 100822 ą 11% cpuidle.POLL.usage
> 841.00 ą 41% -50.4% 417.25 ą 17% numa-meminfo.node0.Dirty
> 7096 ą 7% +25.8% 8930 ą 9% numa-meminfo.node1.KernelStack
> 68752 ą 90% -45.9% 37169 ą143% sched_debug.cfs_rq:/.runnable_weight.stddev
> 654.93 ą 11% +19.3% 781.09 ą 2% sched_debug.cpu.clock_task.stddev
> 183.06 ą 83% -76.9% 42.20 ą 17% iostat.sda.await.max
> 627.47 ą102% -96.7% 20.52 ą 38% iostat.sda.r_await.max
> 183.08 ą 83% -76.9% 42.24 ą 17% iostat.sda.w_await.max
> 209.00 ą 41% -50.2% 104.00 ą 17% numa-vmstat.node0.nr_dirty
> 209.50 ą 41% -50.4% 104.00 ą 17% numa-vmstat.node0.nr_zone_write_pending
> 6792 ą 8% +34.4% 9131 ą 7% numa-vmstat.node1.nr_kernel_stack
> 3.57 ą173% +9.8 13.38 ą 25% perf-profile.calltrace.cycles-pp.proc_reg_read.vfs_read.ksys_read.do_syscall_64.entry_SYSCALL_64_after_hwframe
> 3.57 ą173% +9.8 13.38 ą 25% perf-profile.calltrace.cycles-pp.seq_read.proc_reg_read.vfs_read.ksys_read.do_syscall_64
> 3.57 ą173% +9.8 13.39 ą 25% perf-profile.children.cycles-pp.proc_reg_read
> 3.57 ą173% +12.6 16.16 ą 28% perf-profile.children.cycles-pp.seq_read
> 7948 ą 56% -53.1% 3730 ą 5% softirqs.CPU25.RCU
> 6701 ą 33% -46.7% 3570 ą 5% softirqs.CPU34.RCU
> 8232 ą 89% -60.5% 3247 softirqs.CPU50.RCU
> 326269 ą 16% -27.4% 236940 softirqs.RCU
> 68066 +7.9% 73438 proc-vmstat.nr_active_anon
> 67504 +7.8% 72783 proc-vmstat.nr_anon_pages
> 7198 ą 19% +34.2% 9658 ą 2% proc-vmstat.nr_page_table_pages
> 40664 ą 8% +10.1% 44766 proc-vmstat.nr_slab_unreclaimable
> 68066 +7.9% 73438 proc-vmstat.nr_zone_active_anon
> 1980169 ą 4% -5.3% 1875307 proc-vmstat.numa_hit
> 1960247 ą 4% -5.4% 1855033 proc-vmstat.numa_local
> 956008 ą 16% -17.8% 786247 proc-vmstat.pgfault
> 26598 ą 76% +301.2% 106716 ą 45% interrupts.CPU1.RES:Rescheduling_interrupts
> 151212 ą 39% -67.3% 49451 ą 57% interrupts.CPU26.RES:Rescheduling_interrupts
> 1013586 ą 2% -10.9% 903528 ą 7% interrupts.CPU27.LOC:Local_timer_interrupts
> 1000980 ą 2% -11.4% 886740 ą 8% interrupts.CPU31.LOC:Local_timer_interrupts
> 1021043 ą 3% -9.9% 919686 ą 6% interrupts.CPU32.LOC:Local_timer_interrupts
> 125222 ą 51% -86.0% 17483 ą106% interrupts.CPU33.RES:Rescheduling_interrupts
> 1003735 ą 2% -11.1% 891833 ą 8% interrupts.CPU34.LOC:Local_timer_interrupts
> 1021799 ą 2% -13.2% 886665 ą 8% interrupts.CPU38.LOC:Local_timer_interrupts
> 997788 ą 2% -13.2% 866427 ą 10% interrupts.CPU42.LOC:Local_timer_interrupts
> 1001618 -11.6% 885490 ą 9% interrupts.CPU45.LOC:Local_timer_interrupts
> 22321 ą 58% +550.3% 145153 ą 22% interrupts.CPU9.RES:Rescheduling_interrupts
> 3151 ą 53% +67.3% 5273 ą 8% slabinfo.avc_xperms_data.active_objs
> 3151 ą 53% +67.3% 5273 ą 8% slabinfo.avc_xperms_data.num_objs
> 348.75 ą 13% +39.8% 487.50 ą 5% slabinfo.biovec-128.active_objs
> 348.75 ą 13% +39.8% 487.50 ą 5% slabinfo.biovec-128.num_objs
> 13422 ą 97% +121.1% 29678 ą 2% slabinfo.btrfs_extent_map.active_objs
> 14638 ą 98% +117.8% 31888 ą 2% slabinfo.btrfs_extent_map.num_objs
> 3835 ą 18% +40.9% 5404 ą 7% slabinfo.dmaengine-unmap-16.active_objs
> 3924 ą 18% +39.9% 5490 ą 8% slabinfo.dmaengine-unmap-16.num_objs
> 3482 ą 96% +119.1% 7631 ą 10% slabinfo.khugepaged_mm_slot.active_objs
> 3573 ą 96% +119.4% 7839 ą 10% slabinfo.khugepaged_mm_slot.num_objs
> 8629 ą 52% -49.2% 4384 slabinfo.kmalloc-rcl-64.active_objs
> 8629 ą 52% -49.2% 4384 slabinfo.kmalloc-rcl-64.num_objs
> 2309 ą 57% +82.1% 4206 ą 5% slabinfo.mnt_cache.active_objs
> 2336 ą 57% +80.8% 4224 ą 5% slabinfo.mnt_cache.num_objs
> 5320 ą 48% +69.1% 8999 ą 23% slabinfo.pool_workqueue.active_objs
> 165.75 ą 48% +69.4% 280.75 ą 23% slabinfo.pool_workqueue.active_slabs
> 5320 ą 48% +69.2% 8999 ą 23% slabinfo.pool_workqueue.num_objs
> 165.75 ą 48% +69.4% 280.75 ą 23% slabinfo.pool_workqueue.num_slabs
> 3306 ą 15% +27.0% 4199 ą 3% slabinfo.task_group.active_objs
> 3333 ą 16% +30.1% 4336 ą 3% slabinfo.task_group.num_objs
> 14.74 ą 2% +1.8 16.53 ą 2% perf-stat.i.cache-miss-rate%
> 22459727 ą 20% +46.7% 32955572 ą 4% perf-stat.i.cache-misses
> 33575 ą 19% +68.8% 56658 ą 13% perf-stat.i.cpu-migrations
> 0.03 ą 20% +0.0 0.05 ą 8% perf-stat.i.dTLB-load-miss-rate%
> 6351703 ą 33% +47.2% 9352532 ą 9% perf-stat.i.dTLB-load-misses
> 0.45 ą 3% -3.0% 0.44 perf-stat.i.ipc
> 4711345 ą 18% +43.9% 6780944 ą 7% perf-stat.i.node-load-misses
> 82.51 +4.5 86.97 perf-stat.i.node-store-miss-rate%
> 2861142 ą 31% +60.8% 4601146 ą 5% perf-stat.i.node-store-misses
> 0.92 ą 6% -0.1 0.85 ą 2% perf-stat.overall.branch-miss-rate%
> 0.02 ą 3% +0.0 0.02 ą 4% perf-stat.overall.dTLB-store-miss-rate%
> 715.05 ą 5% +9.9% 785.50 ą 4% perf-stat.overall.instructions-per-iTLB-miss
> 0.44 ą 2% -5.4% 0.42 ą 2% perf-stat.overall.ipc
> 79.67 +2.1 81.80 ą 2% perf-stat.overall.node-store-miss-rate%
> 22237897 ą 19% +46.4% 32560557 ą 5% perf-stat.ps.cache-misses
> 32491 ą 18% +70.5% 55390 ą 13% perf-stat.ps.cpu-migrations
> 6071108 ą 31% +45.0% 8804767 ą 9% perf-stat.ps.dTLB-load-misses
> 1866 ą 98% -91.9% 150.48 ą 2% perf-stat.ps.major-faults
> 4593546 ą 16% +42.4% 6541402 ą 7% perf-stat.ps.node-load-misses
> 2757176 ą 29% +58.4% 4368169 ą 5% perf-stat.ps.node-store-misses
> 1.303e+12 ą 3% -9.8% 1.175e+12 ą 3% perf-stat.total.instructions
>
>
>
> ***************************************************************************************************
> lkp-bdw-ep6: 88 threads Intel(R) Xeon(R) CPU E5-2699 v4 @ 2.20GHz with 128G memory
> =========================================================================================
> class/compiler/cpufreq_governor/disk/kconfig/nr_threads/rootfs/tbox_group/testcase/testtime/ucode:
> interrupt/gcc-7/performance/1HDD/x86_64-rhel-7.6/100%/debian-x86_64-2019-11-14.cgz/lkp-bdw-ep6/stress-ng/30s/0xb000038
>
> commit:
> fcf0553db6 ("sched/fair: Remove meaningless imbalance calculation")
> 0b0695f2b3 ("sched/fair: Rework load_balance()")
>
> fcf0553db6f4c793 0b0695f2b34a4afa3f6e9aa1ff0
> ---------------- ---------------------------
> fail:runs %reproduction fail:runs
> | | |
> 1:4 -25% :4 dmesg.WARNING:at#for_ip_interrupt_entry/0x
> %stddev %change %stddev
> \ | \
> 98245522 +42.3% 1.398e+08 stress-ng.schedpolicy.ops
> 3274860 +42.3% 4661027 stress-ng.schedpolicy.ops_per_sec
> 3.473e+08 -9.7% 3.137e+08 stress-ng.sigq.ops
> 11576537 -9.7% 10454846 stress-ng.sigq.ops_per_sec
> 38097605 ą 6% +10.3% 42011440 ą 4% stress-ng.sigrt.ops
> 1269646 ą 6% +10.3% 1400024 ą 4% stress-ng.sigrt.ops_per_sec
> 3.628e+08 ą 4% -21.5% 2.848e+08 ą 10% stress-ng.time.involuntary_context_switches
> 7040 +2.9% 7245 stress-ng.time.percent_of_cpu_this_job_got
> 15.09 ą 3% -13.4% 13.07 ą 5% iostat.cpu.idle
> 14.82 ą 3% -2.0 12.80 ą 5% mpstat.cpu.all.idle%
> 3.333e+08 ą 17% +59.9% 5.331e+08 ą 22% cpuidle.C1.time
> 5985148 ą 23% +112.5% 12719679 ą 20% cpuidle.C1E.usage
> 14.50 ą 3% -12.1% 12.75 ą 6% vmstat.cpu.id
> 1113131 ą 2% -10.5% 996285 ą 3% vmstat.system.cs
> 2269 +2.4% 2324 turbostat.Avg_MHz
> 0.64 ą 17% +0.4 1.02 ą 23% turbostat.C1%
> 5984799 ą 23% +112.5% 12719086 ą 20% turbostat.C1E
> 4.17 ą 32% -46.0% 2.25 ą 38% turbostat.Pkg%pc2
> 216.57 +2.1% 221.12 turbostat.PkgWatt
> 13.33 ą 3% +3.9% 13.84 turbostat.RAMWatt
> 99920 +13.6% 113486 ą 15% proc-vmstat.nr_active_anon
> 5738 +1.2% 5806 proc-vmstat.nr_inactive_anon
> 46788 +2.1% 47749 proc-vmstat.nr_slab_unreclaimable
> 99920 +13.6% 113486 ą 15% proc-vmstat.nr_zone_active_anon
> 5738 +1.2% 5806 proc-vmstat.nr_zone_inactive_anon
> 3150 ą 2% +35.4% 4265 ą 33% proc-vmstat.numa_huge_pte_updates
> 1641223 +34.3% 2203844 ą 32% proc-vmstat.numa_pte_updates
> 13575 ą 18% +62.1% 21999 ą 4% slabinfo.ext4_extent_status.active_objs
> 13954 ą 17% +57.7% 21999 ą 4% slabinfo.ext4_extent_status.num_objs
> 2527 ą 4% +9.8% 2774 ą 2% slabinfo.khugepaged_mm_slot.active_objs
> 2527 ą 4% +9.8% 2774 ą 2% slabinfo.khugepaged_mm_slot.num_objs
> 57547 ą 8% -15.3% 48743 ą 9% slabinfo.kmalloc-rcl-64.active_objs
> 898.75 ą 8% -15.3% 761.00 ą 9% slabinfo.kmalloc-rcl-64.active_slabs
> 57547 ą 8% -15.3% 48743 ą 9% slabinfo.kmalloc-rcl-64.num_objs
> 898.75 ą 8% -15.3% 761.00 ą 9% slabinfo.kmalloc-rcl-64.num_slabs
> 1.014e+10 +1.7% 1.031e+10 perf-stat.i.branch-instructions
> 13.37 ą 4% +2.0 15.33 ą 3% perf-stat.i.cache-miss-rate%
> 1.965e+11 +2.6% 2.015e+11 perf-stat.i.cpu-cycles
> 20057708 ą 4% +13.9% 22841468 ą 4% perf-stat.i.iTLB-loads
> 4.973e+10 +1.4% 5.042e+10 perf-stat.i.instructions
> 3272 ą 2% +2.9% 3366 perf-stat.i.minor-faults
> 4500892 ą 3% +18.9% 5351518 ą 6% perf-stat.i.node-store-misses
> 3.91 +1.3% 3.96 perf-stat.overall.cpi
> 69.62 -1.5 68.11 perf-stat.overall.iTLB-load-miss-rate%
> 1.047e+10 +1.3% 1.061e+10 perf-stat.ps.branch-instructions
> 1117454 ą 2% -10.6% 999467 ą 3% perf-stat.ps.context-switches
> 1.986e+11 +2.4% 2.033e+11 perf-stat.ps.cpu-cycles
> 19614413 ą 4% +13.6% 22288555 ą 4% perf-stat.ps.iTLB-loads
> 3493 -1.1% 3453 perf-stat.ps.minor-faults
> 4546636 ą 3% +17.0% 5321658 ą 5% perf-stat.ps.node-store-misses
> 0.64 ą 3% -0.2 0.44 ą 57% perf-profile.calltrace.cycles-pp.common_timer_get.do_timer_gettime.__x64_sys_timer_gettime.do_syscall_64.entry_SYSCALL_64_after_hwframe
> 0.66 ą 3% -0.1 0.58 ą 7% perf-profile.children.cycles-pp.common_timer_get
> 0.44 ą 4% -0.1 0.39 ą 5% perf-profile.children.cycles-pp.posix_ktime_get_ts
> 0.39 ą 5% -0.0 0.34 ą 6% perf-profile.children.cycles-pp.ktime_get_ts64
> 0.07 ą 17% +0.0 0.10 ą 8% perf-profile.children.cycles-pp.task_tick_fair
> 0.08 ą 15% +0.0 0.11 ą 7% perf-profile.children.cycles-pp.scheduler_tick
> 0.46 ą 5% +0.1 0.54 ą 6% perf-profile.children.cycles-pp.__might_sleep
> 0.69 ą 8% +0.2 0.85 ą 12% perf-profile.children.cycles-pp.___might_sleep
> 0.90 ą 5% -0.2 0.73 ą 9% perf-profile.self.cycles-pp.__might_fault
> 0.40 ą 6% -0.1 0.33 ą 9% perf-profile.self.cycles-pp.do_timer_gettime
> 0.50 ą 4% -0.1 0.45 ą 7% perf-profile.self.cycles-pp.put_itimerspec64
> 0.32 ą 2% -0.0 0.27 ą 9% perf-profile.self.cycles-pp.update_curr_fair
> 0.20 ą 6% -0.0 0.18 ą 2% perf-profile.self.cycles-pp.ktime_get_ts64
> 0.08 ą 23% +0.0 0.12 ą 8% perf-profile.self.cycles-pp._raw_spin_trylock
> 0.42 ą 5% +0.1 0.50 ą 6% perf-profile.self.cycles-pp.__might_sleep
> 0.66 ą 9% +0.2 0.82 ą 12% perf-profile.self.cycles-pp.___might_sleep
> 47297 ą 13% +19.7% 56608 ą 5% softirqs.CPU13.SCHED
> 47070 ą 3% +20.5% 56735 ą 7% softirqs.CPU2.SCHED
> 55443 ą 9% -20.2% 44250 ą 2% softirqs.CPU28.SCHED
> 56633 ą 3% -12.6% 49520 ą 7% softirqs.CPU34.SCHED
> 56599 ą 11% -18.0% 46384 ą 2% softirqs.CPU36.SCHED
> 56909 ą 9% -18.4% 46438 ą 6% softirqs.CPU40.SCHED
> 45062 ą 9% +28.1% 57709 ą 9% softirqs.CPU45.SCHED
> 43959 +28.7% 56593 ą 9% softirqs.CPU49.SCHED
> 46235 ą 10% +22.2% 56506 ą 11% softirqs.CPU5.SCHED
> 44779 ą 12% +22.5% 54859 ą 11% softirqs.CPU57.SCHED
> 46739 ą 10% +21.1% 56579 ą 8% softirqs.CPU6.SCHED
> 53129 ą 4% -13.1% 46149 ą 8% softirqs.CPU70.SCHED
> 55822 ą 7% -20.5% 44389 ą 8% softirqs.CPU73.SCHED
> 56011 ą 5% -11.4% 49610 ą 7% softirqs.CPU77.SCHED
> 55263 ą 9% -13.2% 47942 ą 12% softirqs.CPU78.SCHED
> 58792 ą 14% -21.3% 46291 ą 9% softirqs.CPU81.SCHED
> 53341 ą 7% -13.7% 46041 ą 10% softirqs.CPU83.SCHED
> 59096 ą 15% -23.9% 44998 ą 6% softirqs.CPU85.SCHED
> 36647 -98.5% 543.00 ą 61% numa-meminfo.node0.Active(file)
> 620922 ą 4% -10.4% 556566 ą 5% numa-meminfo.node0.FilePages
> 21243 ą 3% -36.2% 13543 ą 41% numa-meminfo.node0.Inactive
> 20802 ą 3% -35.3% 13455 ą 42% numa-meminfo.node0.Inactive(anon)
> 15374 ą 9% -27.2% 11193 ą 8% numa-meminfo.node0.KernelStack
> 21573 -34.7% 14084 ą 14% numa-meminfo.node0.Mapped
> 1136795 ą 5% -12.4% 995965 ą 6% numa-meminfo.node0.MemUsed
> 16420 ą 6% -66.0% 5580 ą 18% numa-meminfo.node0.PageTables
> 108182 ą 2% -18.5% 88150 ą 3% numa-meminfo.node0.SUnreclaim
> 166467 ą 2% -15.8% 140184 ą 4% numa-meminfo.node0.Slab
> 181705 ą 36% +63.8% 297623 ą 10% numa-meminfo.node1.Active
> 320.75 ą 27% +11187.0% 36203 numa-meminfo.node1.Active(file)
> 2208 ą 38% +362.1% 10207 ą 54% numa-meminfo.node1.Inactive
> 2150 ą 39% +356.0% 9804 ą 58% numa-meminfo.node1.Inactive(anon)
> 41819 ą 10% +17.3% 49068 ą 6% numa-meminfo.node1.KReclaimable
> 11711 ą 5% +47.2% 17238 ą 22% numa-meminfo.node1.KernelStack
> 10642 +68.3% 17911 ą 11% numa-meminfo.node1.Mapped
> 952520 ą 6% +20.3% 1146337 ą 3% numa-meminfo.node1.MemUsed
> 12342 ą 15% +92.4% 23741 ą 9% numa-meminfo.node1.PageTables
> 41819 ą 10% +17.3% 49068 ą 6% numa-meminfo.node1.SReclaimable
> 80394 ą 3% +27.1% 102206 ą 3% numa-meminfo.node1.SUnreclaim
> 122214 ą 3% +23.8% 151275 ą 3% numa-meminfo.node1.Slab
> 9160 -98.5% 135.25 ą 61% numa-vmstat.node0.nr_active_file
> 155223 ą 4% -10.4% 139122 ą 5% numa-vmstat.node0.nr_file_pages
> 5202 ą 3% -35.4% 3362 ą 42% numa-vmstat.node0.nr_inactive_anon
> 109.50 ą 14% -80.1% 21.75 ą160% numa-vmstat.node0.nr_inactive_file
> 14757 ą 3% -34.4% 9676 ą 12% numa-vmstat.node0.nr_kernel_stack
> 5455 -34.9% 3549 ą 12% numa-vmstat.node0.nr_mapped
> 4069 ą 6% -68.3% 1289 ą 24% numa-vmstat.node0.nr_page_table_pages
> 26943 ą 2% -19.2% 21761 ą 3% numa-vmstat.node0.nr_slab_unreclaimable
> 2240 ą 6% -97.8% 49.00 ą 69% numa-vmstat.node0.nr_written
> 9160 -98.5% 135.25 ą 61% numa-vmstat.node0.nr_zone_active_file
> 5202 ą 3% -35.4% 3362 ą 42% numa-vmstat.node0.nr_zone_inactive_anon
> 109.50 ą 14% -80.1% 21.75 ą160% numa-vmstat.node0.nr_zone_inactive_file
> 79.75 ą 28% +11247.0% 9049 numa-vmstat.node1.nr_active_file
> 542.25 ą 41% +352.1% 2451 ą 58% numa-vmstat.node1.nr_inactive_anon
> 14.00 ą140% +617.9% 100.50 ą 35% numa-vmstat.node1.nr_inactive_file
> 11182 ą 4% +28.9% 14415 ą 4% numa-vmstat.node1.nr_kernel_stack
> 2728 ą 3% +67.7% 4576 ą 9% numa-vmstat.node1.nr_mapped
> 3056 ą 15% +88.2% 5754 ą 8% numa-vmstat.node1.nr_page_table_pages
> 10454 ą 10% +17.3% 12262 ą 7% numa-vmstat.node1.nr_slab_reclaimable
> 20006 ą 3% +25.0% 25016 ą 3% numa-vmstat.node1.nr_slab_unreclaimable
> 19.00 ą 52% +11859.2% 2272 ą 2% numa-vmstat.node1.nr_written
> 79.75 ą 28% +11247.0% 9049 numa-vmstat.node1.nr_zone_active_file
> 542.25 ą 41% +352.1% 2451 ą 58% numa-vmstat.node1.nr_zone_inactive_anon
> 14.00 ą140% +617.9% 100.50 ą 35% numa-vmstat.node1.nr_zone_inactive_file
> 173580 ą 21% +349.5% 780280 ą 7% sched_debug.cfs_rq:/.MIN_vruntime.avg
> 6891819 ą 37% +109.1% 14412817 ą 9% sched_debug.cfs_rq:/.MIN_vruntime.max
> 1031500 ą 25% +189.1% 2982452 ą 8% sched_debug.cfs_rq:/.MIN_vruntime.stddev
> 149079 +13.6% 169354 ą 2% sched_debug.cfs_rq:/.exec_clock.min
> 8550 ą 3% -59.7% 3442 ą 32% sched_debug.cfs_rq:/.exec_clock.stddev
> 4.95 ą 6% -15.2% 4.20 ą 10% sched_debug.cfs_rq:/.load_avg.min
> 173580 ą 21% +349.5% 780280 ą 7% sched_debug.cfs_rq:/.max_vruntime.avg
> 6891819 ą 37% +109.1% 14412817 ą 9% sched_debug.cfs_rq:/.max_vruntime.max
> 1031500 ą 25% +189.1% 2982452 ą 8% sched_debug.cfs_rq:/.max_vruntime.stddev
> 16144141 +27.9% 20645199 ą 6% sched_debug.cfs_rq:/.min_vruntime.avg
> 17660392 +27.7% 22546402 ą 4% sched_debug.cfs_rq:/.min_vruntime.max
> 13747718 +36.8% 18802595 ą 5% sched_debug.cfs_rq:/.min_vruntime.min
> 0.17 ą 11% +35.0% 0.22 ą 15% sched_debug.cfs_rq:/.nr_running.stddev
> 10.64 ą 14% -26.4% 7.83 ą 12% sched_debug.cpu.clock.stddev
> 10.64 ą 14% -26.4% 7.83 ą 12% sched_debug.cpu.clock_task.stddev
> 7093 ą 42% -65.9% 2420 ą120% sched_debug.cpu.curr->pid.min
> 2434979 ą 2% -18.6% 1981697 ą 3% sched_debug.cpu.nr_switches.avg
> 3993189 ą 6% -22.2% 3104832 ą 5% sched_debug.cpu.nr_switches.max
> -145.03 -42.8% -82.90 sched_debug.cpu.nr_uninterruptible.min
> 2097122 ą 6% +38.7% 2908923 ą 6% sched_debug.cpu.sched_count.min
> 809684 ą 13% -30.5% 562929 ą 17% sched_debug.cpu.sched_count.stddev
> 307565 ą 4% -15.1% 261231 ą 3% sched_debug.cpu.ttwu_count.min
> 207286 ą 6% -16.4% 173387 ą 3% sched_debug.cpu.ttwu_local.min
> 125963 ą 23% +53.1% 192849 ą 2% sched_debug.cpu.ttwu_local.stddev
> 2527246 +10.8% 2800959 ą 3% sched_debug.cpu.yld_count.avg
> 1294266 ą 4% +53.7% 1989264 ą 2% sched_debug.cpu.yld_count.min
> 621332 ą 9% -38.4% 382813 ą 22% sched_debug.cpu.yld_count.stddev
> 899.50 ą 28% -48.2% 465.75 ą 42% interrupts.37:IR-PCI-MSI.1572868-edge.eth0-TxRx-3
> 372.50 ą 7% +169.5% 1004 ą 40% interrupts.40:IR-PCI-MSI.1572871-edge.eth0-TxRx-6
> 6201 ą 8% +17.9% 7309 ą 3% interrupts.CPU0.CAL:Function_call_interrupts
> 653368 ą 47% +159.4% 1695029 ą 17% interrupts.CPU0.RES:Rescheduling_interrupts
> 7104 ą 7% +13.6% 8067 interrupts.CPU1.CAL:Function_call_interrupts
> 2094 ą 59% +89.1% 3962 ą 10% interrupts.CPU10.TLB:TLB_shootdowns
> 7309 ą 8% +11.2% 8125 interrupts.CPU11.CAL:Function_call_interrupts
> 2089 ą 62% +86.2% 3890 ą 11% interrupts.CPU13.TLB:TLB_shootdowns
> 7068 ą 8% +15.2% 8144 ą 2% interrupts.CPU14.CAL:Function_call_interrupts
> 7112 ą 7% +13.6% 8079 ą 3% interrupts.CPU15.CAL:Function_call_interrupts
> 1950 ą 61% +103.5% 3968 ą 11% interrupts.CPU15.TLB:TLB_shootdowns
> 899.50 ą 28% -48.2% 465.75 ą 42% interrupts.CPU16.37:IR-PCI-MSI.1572868-edge.eth0-TxRx-3
> 2252 ą 47% +62.6% 3664 ą 15% interrupts.CPU16.TLB:TLB_shootdowns
> 7111 ą 8% +14.8% 8167 ą 3% interrupts.CPU18.CAL:Function_call_interrupts
> 1972 ą 60% +96.3% 3872 ą 9% interrupts.CPU18.TLB:TLB_shootdowns
> 372.50 ą 7% +169.5% 1004 ą 40% interrupts.CPU19.40:IR-PCI-MSI.1572871-edge.eth0-TxRx-6
> 2942 ą 12% -57.5% 1251 ą 22% interrupts.CPU22.TLB:TLB_shootdowns
> 7819 -12.2% 6861 ą 3% interrupts.CPU23.CAL:Function_call_interrupts
> 3327 ą 12% -62.7% 1241 ą 29% interrupts.CPU23.TLB:TLB_shootdowns
> 7767 ą 3% -14.0% 6683 ą 5% interrupts.CPU24.CAL:Function_call_interrupts
> 3185 ą 21% -63.8% 1154 ą 14% interrupts.CPU24.TLB:TLB_shootdowns
> 7679 ą 4% -11.3% 6812 ą 2% interrupts.CPU25.CAL:Function_call_interrupts
> 3004 ą 28% -63.4% 1100 ą 7% interrupts.CPU25.TLB:TLB_shootdowns
> 3187 ą 17% -61.3% 1232 ą 35% interrupts.CPU26.TLB:TLB_shootdowns
> 3193 ą 16% -59.3% 1299 ą 34% interrupts.CPU27.TLB:TLB_shootdowns
> 3059 ą 21% -58.0% 1285 ą 32% interrupts.CPU28.TLB:TLB_shootdowns
> 7798 ą 4% -13.8% 6719 ą 7% interrupts.CPU29.CAL:Function_call_interrupts
> 3122 ą 20% -62.3% 1178 ą 37% interrupts.CPU29.TLB:TLB_shootdowns
> 7727 ą 2% -11.6% 6827 ą 5% interrupts.CPU30.CAL:Function_call_interrupts
> 3102 ą 18% -59.4% 1259 ą 33% interrupts.CPU30.TLB:TLB_shootdowns
> 3269 ą 24% -58.1% 1371 ą 48% interrupts.CPU31.TLB:TLB_shootdowns
> 7918 ą 3% -14.5% 6771 interrupts.CPU32.CAL:Function_call_interrupts
> 3324 ą 18% -70.7% 973.50 ą 18% interrupts.CPU32.TLB:TLB_shootdowns
> 2817 ą 27% -60.2% 1121 ą 26% interrupts.CPU33.TLB:TLB_shootdowns
> 7956 ą 3% -11.8% 7018 ą 4% interrupts.CPU34.CAL:Function_call_interrupts
> 3426 ą 21% -70.3% 1018 ą 29% interrupts.CPU34.TLB:TLB_shootdowns
> 3121 ą 17% -70.3% 926.75 ą 22% interrupts.CPU35.TLB:TLB_shootdowns
> 7596 ą 4% -10.6% 6793 ą 3% interrupts.CPU36.CAL:Function_call_interrupts
> 2900 ą 30% -62.3% 1094 ą 34% interrupts.CPU36.TLB:TLB_shootdowns
> 7863 -13.1% 6833 ą 2% interrupts.CPU37.CAL:Function_call_interrupts
> 3259 ą 15% -65.9% 1111 ą 20% interrupts.CPU37.TLB:TLB_shootdowns
> 3230 ą 26% -64.0% 1163 ą 39% interrupts.CPU38.TLB:TLB_shootdowns
> 7728 ą 5% -13.8% 6662 ą 7% interrupts.CPU39.CAL:Function_call_interrupts
> 2950 ą 29% -61.6% 1133 ą 26% interrupts.CPU39.TLB:TLB_shootdowns
> 6864 ą 3% +18.7% 8147 interrupts.CPU4.CAL:Function_call_interrupts
> 1847 ą 59% +118.7% 4039 ą 7% interrupts.CPU4.TLB:TLB_shootdowns
> 7951 ą 6% -15.0% 6760 ą 2% interrupts.CPU40.CAL:Function_call_interrupts
> 3200 ą 30% -72.3% 886.50 ą 39% interrupts.CPU40.TLB:TLB_shootdowns
> 7819 ą 6% -11.3% 6933 ą 2% interrupts.CPU41.CAL:Function_call_interrupts
> 3149 ą 28% -62.9% 1169 ą 24% interrupts.CPU41.TLB:TLB_shootdowns
> 7884 ą 4% -11.0% 7019 ą 2% interrupts.CPU42.CAL:Function_call_interrupts
> 3248 ą 16% -63.4% 1190 ą 23% interrupts.CPU42.TLB:TLB_shootdowns
> 7659 ą 5% -12.7% 6690 ą 3% interrupts.CPU43.CAL:Function_call_interrupts
> 490732 ą 20% +114.5% 1052606 ą 47% interrupts.CPU43.RES:Rescheduling_interrupts
> 1432688 ą 34% -67.4% 467217 ą 43% interrupts.CPU47.RES:Rescheduling_interrupts
> 7122 ą 8% +16.0% 8259 ą 3% interrupts.CPU48.CAL:Function_call_interrupts
> 1868 ą 65% +118.4% 4079 ą 8% interrupts.CPU48.TLB:TLB_shootdowns
> 7165 ą 8% +11.3% 7977 ą 5% interrupts.CPU49.CAL:Function_call_interrupts
> 1961 ą 59% +98.4% 3891 ą 4% interrupts.CPU49.TLB:TLB_shootdowns
> 461807 ą 47% +190.8% 1342990 ą 48% interrupts.CPU5.RES:Rescheduling_interrupts
> 7167 ą 7% +15.4% 8273 interrupts.CPU50.CAL:Function_call_interrupts
> 2027 ą 51% +103.9% 4134 ą 8% interrupts.CPU50.TLB:TLB_shootdowns
> 7163 ą 9% +16.3% 8328 interrupts.CPU51.CAL:Function_call_interrupts
> 660073 ą 33% +74.0% 1148640 ą 25% interrupts.CPU51.RES:Rescheduling_interrupts
> 2043 ą 64% +95.8% 4000 ą 5% interrupts.CPU51.TLB:TLB_shootdowns
> 7428 ą 9% +13.5% 8434 ą 2% interrupts.CPU52.CAL:Function_call_interrupts
> 2280 ą 61% +85.8% 4236 ą 9% interrupts.CPU52.TLB:TLB_shootdowns
> 7144 ą 11% +17.8% 8413 interrupts.CPU53.CAL:Function_call_interrupts
> 1967 ą 67% +104.7% 4026 ą 5% interrupts.CPU53.TLB:TLB_shootdowns
> 7264 ą 10% +15.6% 8394 ą 4% interrupts.CPU54.CAL:Function_call_interrupts
> 7045 ą 11% +18.7% 8365 ą 2% interrupts.CPU56.CAL:Function_call_interrupts
> 2109 ą 59% +91.6% 4041 ą 10% interrupts.CPU56.TLB:TLB_shootdowns
> 7307 ą 9% +15.3% 8428 ą 2% interrupts.CPU57.CAL:Function_call_interrupts
> 2078 ą 64% +96.5% 4085 ą 6% interrupts.CPU57.TLB:TLB_shootdowns
> 6834 ą 12% +19.8% 8190 ą 3% interrupts.CPU58.CAL:Function_call_interrupts
> 612496 ą 85% +122.5% 1362815 ą 27% interrupts.CPU58.RES:Rescheduling_interrupts
> 1884 ą 69% +112.0% 3995 ą 8% interrupts.CPU58.TLB:TLB_shootdowns
> 7185 ą 8% +15.9% 8329 interrupts.CPU59.CAL:Function_call_interrupts
> 1982 ą 58% +101.1% 3986 ą 5% interrupts.CPU59.TLB:TLB_shootdowns
> 7051 ą 6% +13.1% 7975 interrupts.CPU6.CAL:Function_call_interrupts
> 1831 ą 49% +102.1% 3701 ą 8% interrupts.CPU6.TLB:TLB_shootdowns
> 7356 ą 8% +16.2% 8548 interrupts.CPU60.CAL:Function_call_interrupts
> 2124 ą 57% +92.8% 4096 ą 5% interrupts.CPU60.TLB:TLB_shootdowns
> 7243 ą 9% +15.1% 8334 interrupts.CPU61.CAL:Function_call_interrupts
> 572423 ą 71% +110.0% 1201919 ą 40% interrupts.CPU61.RES:Rescheduling_interrupts
> 7295 ą 9% +14.7% 8369 interrupts.CPU63.CAL:Function_call_interrupts
> 2139 ą 57% +85.7% 3971 ą 3% interrupts.CPU63.TLB:TLB_shootdowns
> 7964 ą 2% -15.6% 6726 ą 5% interrupts.CPU66.CAL:Function_call_interrupts
> 3198 ą 21% -65.0% 1119 ą 24% interrupts.CPU66.TLB:TLB_shootdowns
> 8103 ą 2% -17.5% 6687 ą 9% interrupts.CPU67.CAL:Function_call_interrupts
> 3357 ą 18% -62.9% 1244 ą 32% interrupts.CPU67.TLB:TLB_shootdowns
> 7772 ą 2% -14.0% 6687 ą 8% interrupts.CPU68.CAL:Function_call_interrupts
> 2983 ą 17% -59.2% 1217 ą 15% interrupts.CPU68.TLB:TLB_shootdowns
> 7986 ą 4% -13.8% 6887 ą 4% interrupts.CPU69.CAL:Function_call_interrupts
> 3192 ą 24% -65.0% 1117 ą 30% interrupts.CPU69.TLB:TLB_shootdowns
> 7070 ą 6% +14.6% 8100 ą 2% interrupts.CPU7.CAL:Function_call_interrupts
> 697891 ą 32% +54.4% 1077890 ą 18% interrupts.CPU7.RES:Rescheduling_interrupts
> 1998 ą 55% +97.1% 3938 ą 10% interrupts.CPU7.TLB:TLB_shootdowns
> 8085 -13.4% 7002 ą 3% interrupts.CPU70.CAL:Function_call_interrupts
> 1064985 ą 35% -62.5% 398986 ą 29% interrupts.CPU70.RES:Rescheduling_interrupts
> 3347 ą 12% -61.7% 1280 ą 24% interrupts.CPU70.TLB:TLB_shootdowns
> 2916 ą 16% -58.8% 1201 ą 39% interrupts.CPU71.TLB:TLB_shootdowns
> 3314 ą 19% -61.3% 1281 ą 26% interrupts.CPU72.TLB:TLB_shootdowns
> 3119 ą 18% -61.5% 1200 ą 39% interrupts.CPU73.TLB:TLB_shootdowns
> 7992 ą 4% -12.6% 6984 ą 3% interrupts.CPU74.CAL:Function_call_interrupts
> 3187 ą 21% -56.8% 1378 ą 40% interrupts.CPU74.TLB:TLB_shootdowns
> 7953 ą 4% -12.0% 6999 ą 4% interrupts.CPU75.CAL:Function_call_interrupts
> 3072 ą 26% -56.8% 1327 ą 34% interrupts.CPU75.TLB:TLB_shootdowns
> 8119 ą 5% -12.4% 7109 ą 7% interrupts.CPU76.CAL:Function_call_interrupts
> 3418 ą 20% -67.5% 1111 ą 31% interrupts.CPU76.TLB:TLB_shootdowns
> 7804 ą 5% -11.4% 6916 ą 4% interrupts.CPU77.CAL:Function_call_interrupts
> 7976 ą 5% -14.4% 6826 ą 3% interrupts.CPU78.CAL:Function_call_interrupts
> 3209 ą 27% -71.8% 904.75 ą 28% interrupts.CPU78.TLB:TLB_shootdowns
> 8187 ą 4% -14.6% 6991 ą 3% interrupts.CPU79.CAL:Function_call_interrupts
> 3458 ą 20% -67.5% 1125 ą 36% interrupts.CPU79.TLB:TLB_shootdowns
> 7122 ą 7% +14.2% 8136 ą 2% interrupts.CPU8.CAL:Function_call_interrupts
> 2096 ą 63% +87.4% 3928 ą 8% interrupts.CPU8.TLB:TLB_shootdowns
> 8130 ą 5% -17.2% 6728 ą 5% interrupts.CPU81.CAL:Function_call_interrupts
> 3253 ą 24% -70.6% 955.00 ą 38% interrupts.CPU81.TLB:TLB_shootdowns
> 7940 ą 5% -13.9% 6839 ą 5% interrupts.CPU82.CAL:Function_call_interrupts
> 2952 ą 26% -66.3% 996.00 ą 51% interrupts.CPU82.TLB:TLB_shootdowns
> 7900 ą 6% -13.4% 6844 ą 3% interrupts.CPU83.CAL:Function_call_interrupts
> 3012 ą 34% -68.3% 956.00 ą 17% interrupts.CPU83.TLB:TLB_shootdowns
> 7952 ą 6% -15.8% 6695 ą 2% interrupts.CPU84.CAL:Function_call_interrupts
> 3049 ą 31% -75.5% 746.50 ą 27% interrupts.CPU84.TLB:TLB_shootdowns
> 8065 ą 6% -15.7% 6798 interrupts.CPU85.CAL:Function_call_interrupts
> 3222 ą 23% -69.7% 976.00 ą 13% interrupts.CPU85.TLB:TLB_shootdowns
> 8049 ą 5% -13.2% 6983 ą 4% interrupts.CPU86.CAL:Function_call_interrupts
> 3159 ą 19% -61.9% 1202 ą 27% interrupts.CPU86.TLB:TLB_shootdowns
> 8154 ą 8% -16.9% 6773 ą 3% interrupts.CPU87.CAL:Function_call_interrupts
> 1432962 ą 21% -48.5% 737989 ą 30% interrupts.CPU87.RES:Rescheduling_interrupts
> 3186 ą 33% -72.3% 881.75 ą 21% interrupts.CPU87.TLB:TLB_shootdowns
>
>
>
> ***************************************************************************************************
> lkp-bdw-ep6: 88 threads Intel(R) Xeon(R) CPU E5-2699 v4 @ 2.20GHz with 128G memory
> =========================================================================================
> class/compiler/cpufreq_governor/disk/kconfig/nr_threads/rootfs/tbox_group/testcase/testtime/ucode:
> interrupt/gcc-7/performance/1HDD/x86_64-rhel-7.6/100%/debian-x86_64-2019-11-14.cgz/lkp-bdw-ep6/stress-ng/1s/0xb000038
>
> commit:
> fcf0553db6 ("sched/fair: Remove meaningless imbalance calculation")
> 0b0695f2b3 ("sched/fair: Rework load_balance()")
>
> fcf0553db6f4c793 0b0695f2b34a4afa3f6e9aa1ff0
> ---------------- ---------------------------
> %stddev %change %stddev
> \ | \
> 3345449 +35.1% 4518187 ą 5% stress-ng.schedpolicy.ops
> 3347036 +35.1% 4520740 ą 5% stress-ng.schedpolicy.ops_per_sec
> 11464910 ą 6% -23.3% 8796455 ą 11% stress-ng.sigq.ops
> 11452565 ą 6% -23.3% 8786844 ą 11% stress-ng.sigq.ops_per_sec
> 228736 +20.7% 276087 ą 20% stress-ng.sleep.ops
> 157479 +23.0% 193722 ą 21% stress-ng.sleep.ops_per_sec
> 14584704 -5.8% 13744640 ą 4% stress-ng.timerfd.ops
> 14546032 -5.7% 13718862 ą 4% stress-ng.timerfd.ops_per_sec
> 27.24 ą105% +283.9% 104.58 ą109% iostat.sdb.r_await.max
> 122324 ą 35% +63.9% 200505 ą 21% meminfo.AnonHugePages
> 47267 ą 26% +155.2% 120638 ą 45% numa-meminfo.node1.AnonHugePages
> 22880 ą 6% -9.9% 20605 ą 3% softirqs.CPU57.TIMER
> 636196 ą 24% +38.5% 880847 ą 7% cpuidle.C1.usage
> 55936214 ą 20% +63.9% 91684673 ą 18% cpuidle.C1E.time
> 1.175e+08 ą 22% +101.8% 2.372e+08 ą 29% cpuidle.C3.time
> 4.242e+08 ą 6% -39.1% 2.584e+08 ą 39% cpuidle.C6.time
> 59.50 ą 34% +66.0% 98.75 ą 22% proc-vmstat.nr_anon_transparent_hugepages
> 25612 ą 10% +13.8% 29146 ą 4% proc-vmstat.nr_kernel_stack
> 2783465 ą 9% +14.5% 3187157 ą 9% proc-vmstat.pgalloc_normal
> 1743 ą 28% +43.8% 2507 ą 23% proc-vmstat.thp_deferred_split_page
> 1765 ą 30% +43.2% 2529 ą 22% proc-vmstat.thp_fault_alloc
> 811.00 ą 3% -13.8% 699.00 ą 7% slabinfo.kmem_cache_node.active_objs
> 864.00 ą 3% -13.0% 752.00 ą 7% slabinfo.kmem_cache_node.num_objs
> 8686 ą 7% +13.6% 9869 ą 3% slabinfo.pid.active_objs
> 8690 ą 7% +13.8% 9890 ą 3% slabinfo.pid.num_objs
> 9813 ą 6% +15.7% 11352 ą 3% slabinfo.task_delay_info.active_objs
> 9813 ą 6% +15.7% 11352 ą 3% slabinfo.task_delay_info.num_objs
> 79.22 ą 10% -41.1% 46.68 ą 22% sched_debug.cfs_rq:/.load_avg.avg
> 242.49 ą 6% -29.6% 170.70 ą 17% sched_debug.cfs_rq:/.load_avg.stddev
> 43.14 ą 29% -67.1% 14.18 ą 66% sched_debug.cfs_rq:/.removed.load_avg.avg
> 201.73 ą 15% -50.1% 100.68 ą 60% sched_debug.cfs_rq:/.removed.load_avg.stddev
> 1987 ą 28% -67.3% 650.09 ą 66% sched_debug.cfs_rq:/.removed.runnable_sum.avg
> 9298 ą 15% -50.3% 4616 ą 60% sched_debug.cfs_rq:/.removed.runnable_sum.stddev
> 18.17 ą 27% -68.6% 5.70 ą 63% sched_debug.cfs_rq:/.removed.util_avg.avg
> 87.61 ą 13% -52.6% 41.48 ą 59% sched_debug.cfs_rq:/.removed.util_avg.stddev
> 633327 ą 24% +38.4% 876596 ą 7% turbostat.C1
> 2.75 ą 22% +1.8 4.52 ą 17% turbostat.C1E%
> 5.76 ą 22% +6.1 11.82 ą 30% turbostat.C3%
> 20.69 ą 5% -8.1 12.63 ą 38% turbostat.C6%
> 15.62 ą 6% +18.4% 18.50 ą 8% turbostat.CPU%c1
> 1.56 ą 16% +208.5% 4.82 ą 38% turbostat.CPU%c3
> 12.81 ą 4% -48.1% 6.65 ą 43% turbostat.CPU%c6
> 5.02 ą 8% -34.6% 3.28 ą 14% turbostat.Pkg%pc2
> 0.85 ą 57% -84.7% 0.13 ą173% turbostat.Pkg%pc6
> 88.25 ą 13% +262.6% 320.00 ą 71% interrupts.CPU10.TLB:TLB_shootdowns
> 116.25 ą 36% +151.6% 292.50 ą 68% interrupts.CPU19.TLB:TLB_shootdowns
> 109.25 ą 8% +217.4% 346.75 ą106% interrupts.CPU2.TLB:TLB_shootdowns
> 15180 ą111% +303.9% 61314 ą 32% interrupts.CPU23.RES:Rescheduling_interrupts
> 111.50 ą 26% +210.3% 346.00 ą 79% interrupts.CPU3.TLB:TLB_shootdowns
> 86.50 ą 35% +413.0% 443.75 ą 66% interrupts.CPU33.TLB:TLB_shootdowns
> 728.00 ą 8% +29.6% 943.50 ą 16% interrupts.CPU38.CAL:Function_call_interrupts
> 1070 ą 72% +84.9% 1979 ą 9% interrupts.CPU54.76:IR-PCI-MSI.512000-edge.ahci[0000:00:1f.2]
> 41429 ą 64% -73.7% 10882 ą 73% interrupts.CPU59.RES:Rescheduling_interrupts
> 26330 ą 85% -73.3% 7022 ą 86% interrupts.CPU62.RES:Rescheduling_interrupts
> 103.00 ą 22% +181.3% 289.75 ą 92% interrupts.CPU65.TLB:TLB_shootdowns
> 100.00 ą 40% +365.0% 465.00 ą 71% interrupts.CPU70.TLB:TLB_shootdowns
> 110.25 ą 18% +308.4% 450.25 ą 71% interrupts.CPU80.TLB:TLB_shootdowns
> 93.50 ą 42% +355.1% 425.50 ą 82% interrupts.CPU84.TLB:TLB_shootdowns
> 104.50 ą 18% +289.7% 407.25 ą 68% interrupts.CPU87.TLB:TLB_shootdowns
> 1.76 ą 3% -0.1 1.66 ą 4% perf-stat.i.branch-miss-rate%
> 8.08 ą 6% +2.0 10.04 perf-stat.i.cache-miss-rate%
> 18031213 ą 4% +27.2% 22939937 ą 3% perf-stat.i.cache-misses
> 4.041e+08 -1.9% 3.965e+08 perf-stat.i.cache-references
> 31764 ą 26% -40.6% 18859 ą 10% perf-stat.i.cycles-between-cache-misses
> 66.18 -1.5 64.71 perf-stat.i.iTLB-load-miss-rate%
> 4503482 ą 8% +19.5% 5382698 ą 5% perf-stat.i.node-load-misses
> 3892859 ą 2% +16.6% 4538750 ą 4% perf-stat.i.node-store-misses
> 1526815 ą 13% +25.8% 1921178 ą 9% perf-stat.i.node-stores
> 4.72 ą 4% +1.3 6.00 ą 3% perf-stat.overall.cache-miss-rate%
> 9120 ą 6% -18.9% 7394 ą 2% perf-stat.overall.cycles-between-cache-misses
> 18237318 ą 4% +25.4% 22866104 ą 3% perf-stat.ps.cache-misses
> 4392089 ą 8% +18.1% 5189251 ą 5% perf-stat.ps.node-load-misses
> 1629766 ą 2% +17.9% 1920947 ą 13% perf-stat.ps.node-loads
> 3694566 ą 2% +16.1% 4288126 ą 4% perf-stat.ps.node-store-misses
> 1536866 ą 12% +23.7% 1901141 ą 7% perf-stat.ps.node-stores
> 38.20 ą 18% -13.2 24.96 ą 10% perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe
> 38.20 ą 18% -13.2 24.96 ą 10% perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe
> 7.98 ą 67% -7.2 0.73 ą173% perf-profile.calltrace.cycles-pp.smp_call_function_single.event_function_call.perf_remove_from_context.perf_event_release_kernel.perf_release
> 7.98 ą 67% -7.2 0.73 ą173% perf-profile.calltrace.cycles-pp.event_function_call.perf_remove_from_context.perf_event_release_kernel.perf_release.__fput
> 7.98 ą 67% -7.2 0.73 ą173% perf-profile.calltrace.cycles-pp.perf_remove_from_context.perf_event_release_kernel.perf_release.__fput.task_work_run
> 11.86 ą 41% -6.8 5.07 ą 62% perf-profile.calltrace.cycles-pp.exit_to_usermode_loop.do_syscall_64.entry_SYSCALL_64_after_hwframe
> 11.86 ą 41% -6.8 5.07 ą 62% perf-profile.calltrace.cycles-pp.do_signal.exit_to_usermode_loop.do_syscall_64.entry_SYSCALL_64_after_hwframe
> 11.86 ą 41% -6.8 5.07 ą 62% perf-profile.calltrace.cycles-pp.get_signal.do_signal.exit_to_usermode_loop.do_syscall_64.entry_SYSCALL_64_after_hwframe
> 11.86 ą 41% -6.8 5.07 ą 62% perf-profile.calltrace.cycles-pp.do_group_exit.get_signal.do_signal.exit_to_usermode_loop.do_syscall_64
> 11.86 ą 41% -6.8 5.07 ą 62% perf-profile.calltrace.cycles-pp.do_exit.do_group_exit.get_signal.do_signal.exit_to_usermode_loop
> 11.86 ą 41% -6.8 5.07 ą 62% perf-profile.calltrace.cycles-pp.task_work_run.do_exit.do_group_exit.get_signal.do_signal
> 4.27 ą 66% -3.5 0.73 ą173% perf-profile.calltrace.cycles-pp.read
> 4.05 ą 71% -3.3 0.73 ą173% perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.read
> 4.05 ą 71% -3.3 0.73 ą173% perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.read
> 13.30 ą 38% -8.2 5.07 ą 62% perf-profile.children.cycles-pp.task_work_run
> 12.47 ą 46% -7.4 5.07 ą 62% perf-profile.children.cycles-pp.exit_to_usermode_loop
> 12.47 ą 46% -7.4 5.07 ą 62% perf-profile.children.cycles-pp.__fput
> 7.98 ą 67% -7.2 0.73 ą173% perf-profile.children.cycles-pp.perf_remove_from_context
> 11.86 ą 41% -6.8 5.07 ą 62% perf-profile.children.cycles-pp.do_signal
> 11.86 ą 41% -6.8 5.07 ą 62% perf-profile.children.cycles-pp.get_signal
> 9.43 ą 21% -4.7 4.72 ą 67% perf-profile.children.cycles-pp.ksys_read
> 9.43 ą 21% -4.7 4.72 ą 67% perf-profile.children.cycles-pp.vfs_read
> 4.27 ą 66% -3.5 0.73 ą173% perf-profile.children.cycles-pp.read
> 3.86 ą101% -3.1 0.71 ą173% perf-profile.children.cycles-pp._raw_spin_lock
> 3.86 ą101% -3.1 0.71 ą173% perf-profile.children.cycles-pp.native_queued_spin_lock_slowpath
> 3.86 ą101% -3.1 0.71 ą173% perf-profile.self.cycles-pp.native_queued_spin_lock_slowpath
>
>
>
> ***************************************************************************************************
> lkp-csl-2sp5: 96 threads Intel(R) Xeon(R) Gold 6252 CPU @ 2.10GHz with 192G memory
> =========================================================================================
> class/compiler/cpufreq_governor/disk/fs/kconfig/nr_threads/rootfs/tbox_group/testcase/testtime/ucode:
> os/gcc-7/performance/1HDD/ext4/x86_64-rhel-7.6/100%/debian-x86_64-2019-09-23.cgz/lkp-csl-2sp5/stress-ng/1s/0x500002b
>
> commit:
> fcf0553db6 ("sched/fair: Remove meaningless imbalance calculation")
> 0b0695f2b3 ("sched/fair: Rework load_balance()")
>
> fcf0553db6f4c793 0b0695f2b34a4afa3f6e9aa1ff0
> ---------------- ---------------------------
> fail:runs %reproduction fail:runs
> | | |
> :2 50% 1:8 dmesg.WARNING:at_ip_selinux_file_ioctl/0x
> %stddev %change %stddev
> \ | \
> 122451 ą 11% -19.9% 98072 ą 15% stress-ng.ioprio.ops
> 116979 ą 11% -20.7% 92815 ą 16% stress-ng.ioprio.ops_per_sec
> 274187 ą 21% -26.7% 201013 ą 11% stress-ng.kill.ops
> 274219 ą 21% -26.7% 201040 ą 11% stress-ng.kill.ops_per_sec
> 3973765 -10.1% 3570462 ą 5% stress-ng.lockf.ops
> 3972581 -10.2% 3568935 ą 5% stress-ng.lockf.ops_per_sec
> 10719 ą 8% -39.9% 6442 ą 22% stress-ng.procfs.ops
> 9683 ą 3% -39.3% 5878 ą 22% stress-ng.procfs.ops_per_sec
> 6562721 -35.1% 4260609 ą 8% stress-ng.schedpolicy.ops
> 6564233 -35.1% 4261479 ą 8% stress-ng.schedpolicy.ops_per_sec
> 1070988 +21.4% 1299977 ą 7% stress-ng.sigrt.ops
> 1061773 +21.2% 1286618 ą 7% stress-ng.sigrt.ops_per_sec
> 1155684 ą 5% -14.8% 984531 ą 16% stress-ng.symlink.ops
> 991624 ą 4% -23.8% 755147 ą 41% stress-ng.symlink.ops_per_sec
> 6925 -12.1% 6086 ą 27% stress-ng.time.percent_of_cpu_this_job_got
> 24.68 +9.3 33.96 ą 52% mpstat.cpu.all.idle%
> 171.00 ą 2% -55.3% 76.50 ą 60% numa-vmstat.node1.nr_inactive_file
> 171.00 ą 2% -55.3% 76.50 ą 60% numa-vmstat.node1.nr_zone_inactive_file
> 2.032e+11 -12.5% 1.777e+11 ą 27% perf-stat.i.cpu-cycles
> 2.025e+11 -12.0% 1.782e+11 ą 27% perf-stat.ps.cpu-cycles
> 25.00 +37.5% 34.38 ą 51% vmstat.cpu.id
> 68.00 -13.2% 59.00 ą 27% vmstat.cpu.sy
> 25.24 +37.0% 34.57 ą 51% iostat.cpu.idle
> 68.21 -12.7% 59.53 ą 27% iostat.cpu.system
> 4.31 ą100% +200.6% 12.96 ą 63% iostat.sda.r_await.max
> 1014 ą 2% -17.1% 841.00 ą 10% meminfo.Inactive(file)
> 30692 ą 12% -20.9% 24280 ą 30% meminfo.Mlocked
> 103627 ą 27% -32.7% 69720 meminfo.Percpu
> 255.50 ą 2% -18.1% 209.25 ą 10% proc-vmstat.nr_inactive_file
> 255.50 ą 2% -18.1% 209.25 ą 10% proc-vmstat.nr_zone_inactive_file
> 185035 ą 22% -22.2% 143917 ą 25% proc-vmstat.pgmigrate_success
> 2107 -12.3% 1848 ą 27% turbostat.Avg_MHz
> 69.00 -7.1% 64.12 ą 8% turbostat.PkgTmp
> 94.63 -2.2% 92.58 ą 4% turbostat.RAMWatt
> 96048 +26.8% 121800 ą 8% softirqs.CPU10.NET_RX
> 96671 ą 4% +34.2% 129776 ą 6% softirqs.CPU15.NET_RX
> 171243 ą 3% -12.9% 149135 ą 8% softirqs.CPU25.NET_RX
> 165317 ą 4% -11.4% 146494 ą 9% softirqs.CPU27.NET_RX
> 139558 -24.5% 105430 ą 14% softirqs.CPU58.NET_RX
> 147836 -15.8% 124408 ą 6% softirqs.CPU63.NET_RX
> 129568 -13.8% 111624 ą 10% softirqs.CPU66.NET_RX
> 1050 ą 2% +14.2% 1198 ą 9% slabinfo.biovec-128.active_objs
> 1050 ą 2% +14.2% 1198 ą 9% slabinfo.biovec-128.num_objs
> 23129 +19.6% 27668 ą 6% slabinfo.kmalloc-512.active_objs
> 766.50 +17.4% 899.75 ą 6% slabinfo.kmalloc-512.active_slabs
> 24535 +17.4% 28806 ą 6% slabinfo.kmalloc-512.num_objs
> 766.50 +17.4% 899.75 ą 6% slabinfo.kmalloc-512.num_slabs
> 1039 ą 4% -4.3% 994.12 ą 6% slabinfo.sock_inode_cache.active_slabs
> 40527 ą 4% -4.3% 38785 ą 6% slabinfo.sock_inode_cache.num_objs
> 1039 ą 4% -4.3% 994.12 ą 6% slabinfo.sock_inode_cache.num_slabs
> 1549456 -43.6% 873443 ą 24% sched_debug.cfs_rq:/.min_vruntime.stddev
> 73.25 ą 5% +74.8% 128.03 ą 31% sched_debug.cfs_rq:/.nr_spread_over.stddev
> 18.60 ą 57% -63.8% 6.73 ą 64% sched_debug.cfs_rq:/.removed.load_avg.avg
> 79.57 ą 44% -44.1% 44.52 ą 55% sched_debug.cfs_rq:/.removed.load_avg.stddev
> 857.10 ą 57% -63.8% 310.09 ą 64% sched_debug.cfs_rq:/.removed.runnable_sum.avg
> 3664 ą 44% -44.1% 2049 ą 55% sched_debug.cfs_rq:/.removed.runnable_sum.stddev
> 4.91 ą 42% -45.3% 2.69 ą 61% sched_debug.cfs_rq:/.removed.util_avg.avg
> 1549544 -43.6% 874006 ą 24% sched_debug.cfs_rq:/.spread0.stddev
> 786.14 ą 6% -20.1% 628.46 ą 23% sched_debug.cfs_rq:/.util_avg.avg
> 1415 ą 8% -16.7% 1178 ą 18% sched_debug.cfs_rq:/.util_avg.max
> 467435 ą 15% +46.7% 685829 ą 15% sched_debug.cpu.avg_idle.avg
> 17972 ą 8% +631.2% 131410 ą 34% sched_debug.cpu.avg_idle.min
> 7.66 ą 26% +209.7% 23.72 ą 54% sched_debug.cpu.clock.stddev
> 7.66 ą 26% +209.7% 23.72 ą 54% sched_debug.cpu.clock_task.stddev
> 618063 ą 5% -17.0% 513085 ą 5% sched_debug.cpu.max_idle_balance_cost.max
> 12083 ą 28% -85.4% 1768 ą231% sched_debug.cpu.max_idle_balance_cost.stddev
> 12857 ą 16% +2117.7% 285128 ą106% sched_debug.cpu.yld_count.min
> 0.55 ą 6% -0.2 0.37 ą 51% perf-profile.children.cycles-pp.fpregs_assert_state_consistent
> 0.30 ą 21% -0.2 0.14 ą105% perf-profile.children.cycles-pp.yield_task_fair
> 0.32 ą 6% -0.2 0.16 ą 86% perf-profile.children.cycles-pp.rmap_walk_anon
> 0.19 -0.1 0.10 ą 86% perf-profile.children.cycles-pp.page_mapcount_is_zero
> 0.19 -0.1 0.10 ą 86% perf-profile.children.cycles-pp.total_mapcount
> 0.14 -0.1 0.09 ą 29% perf-profile.children.cycles-pp.start_kernel
> 0.11 ą 9% -0.0 0.07 ą 47% perf-profile.children.cycles-pp.__switch_to
> 0.10 ą 14% -0.0 0.06 ą 45% perf-profile.children.cycles-pp.switch_fpu_return
> 0.08 ą 6% -0.0 0.04 ą 79% perf-profile.children.cycles-pp.__update_load_avg_se
> 0.12 ą 13% -0.0 0.09 ą 23% perf-profile.children.cycles-pp.native_write_msr
> 0.31 ą 6% -0.2 0.15 ą 81% perf-profile.self.cycles-pp.poll_idle
> 0.50 ą 6% -0.2 0.35 ą 50% perf-profile.self.cycles-pp.fpregs_assert_state_consistent
> 0.18 ą 2% -0.1 0.10 ą 86% perf-profile.self.cycles-pp.total_mapcount
> 0.10 ą 14% -0.0 0.06 ą 45% perf-profile.self.cycles-pp.switch_fpu_return
> 0.10 ą 10% -0.0 0.06 ą 47% perf-profile.self.cycles-pp.__switch_to
> 0.07 ą 7% -0.0 0.03 ą100% perf-profile.self.cycles-pp.prep_new_page
> 0.07 ą 7% -0.0 0.03 ą100% perf-profile.self.cycles-pp.llist_add_batch
> 0.07 ą 14% -0.0 0.04 ą 79% perf-profile.self.cycles-pp.__update_load_avg_se
> 0.12 ą 13% -0.0 0.09 ą 23% perf-profile.self.cycles-pp.native_write_msr
> 66096 ą 99% -99.8% 148.50 ą 92% interrupts.109:PCI-MSI.31981642-edge.i40e-eth0-TxRx-73
> 543.50 ą 39% -73.3% 145.38 ą 81% interrupts.112:PCI-MSI.31981645-edge.i40e-eth0-TxRx-76
> 169.00 ą 28% -55.3% 75.50 ą 83% interrupts.116:PCI-MSI.31981649-edge.i40e-eth0-TxRx-80
> 224.00 ą 14% -57.6% 95.00 ą 87% interrupts.121:PCI-MSI.31981654-edge.i40e-eth0-TxRx-85
> 680.00 ą 28% -80.5% 132.75 ą 82% interrupts.43:PCI-MSI.31981576-edge.i40e-eth0-TxRx-7
> 327.50 ą 31% -39.0% 199.62 ą 60% interrupts.60:PCI-MSI.31981593-edge.i40e-eth0-TxRx-24
> 217.50 ą 19% -51.7% 105.12 ą 79% interrupts.63:PCI-MSI.31981596-edge.i40e-eth0-TxRx-27
> 375.00 ą 46% -78.5% 80.50 ą 82% interrupts.70:PCI-MSI.31981603-edge.i40e-eth0-TxRx-34
> 196.50 ą 3% -51.6% 95.12 ą 74% interrupts.72:PCI-MSI.31981605-edge.i40e-eth0-TxRx-36
> 442.50 ą 45% -73.1% 118.88 ą 90% interrupts.79:PCI-MSI.31981612-edge.i40e-eth0-TxRx-43
> 271.00 ą 8% -53.2% 126.88 ą 75% interrupts.82:PCI-MSI.31981615-edge.i40e-eth0-TxRx-46
> 145448 ą 4% -41.6% 84975 ą 42% interrupts.CPU1.RES:Rescheduling_interrupts
> 11773 ą 19% -38.1% 7290 ą 52% interrupts.CPU13.TLB:TLB_shootdowns
> 24177 ą 15% +356.5% 110368 ą 58% interrupts.CPU16.RES:Rescheduling_interrupts
> 3395 ą 3% +78.3% 6055 ą 18% interrupts.CPU17.NMI:Non-maskable_interrupts
> 3395 ą 3% +78.3% 6055 ą 18% interrupts.CPU17.PMI:Performance_monitoring_interrupts
> 106701 ą 41% -55.6% 47425 ą 56% interrupts.CPU18.RES:Rescheduling_interrupts
> 327.50 ą 31% -39.3% 198.88 ą 60% interrupts.CPU24.60:PCI-MSI.31981593-edge.i40e-eth0-TxRx-24
> 411618 +53.6% 632283 ą 77% interrupts.CPU25.LOC:Local_timer_interrupts
> 16189 ą 26% -53.0% 7611 ą 66% interrupts.CPU25.TLB:TLB_shootdowns
> 407253 +54.4% 628596 ą 78% interrupts.CPU26.LOC:Local_timer_interrupts
> 216.50 ą 19% -51.8% 104.25 ą 80% interrupts.CPU27.63:PCI-MSI.31981596-edge.i40e-eth0-TxRx-27
> 7180 -20.9% 5682 ą 25% interrupts.CPU29.NMI:Non-maskable_interrupts
> 7180 -20.9% 5682 ą 25% interrupts.CPU29.PMI:Performance_monitoring_interrupts
> 15186 ą 12% -45.5% 8276 ą 49% interrupts.CPU3.TLB:TLB_shootdowns
> 13092 ą 19% -29.5% 9231 ą 35% interrupts.CPU30.TLB:TLB_shootdowns
> 13204 ą 26% -29.3% 9336 ą 19% interrupts.CPU31.TLB:TLB_shootdowns
> 374.50 ą 46% -78.7% 79.62 ą 83% interrupts.CPU34.70:PCI-MSI.31981603-edge.i40e-eth0-TxRx-34
> 7188 -25.6% 5345 ą 26% interrupts.CPU35.NMI:Non-maskable_interrupts
> 7188 -25.6% 5345 ą 26% interrupts.CPU35.PMI:Performance_monitoring_interrupts
> 196.00 ą 4% -52.0% 94.12 ą 75% interrupts.CPU36.72:PCI-MSI.31981605-edge.i40e-eth0-TxRx-36
> 12170 ą 20% -34.3% 7998 ą 32% interrupts.CPU39.TLB:TLB_shootdowns
> 442.00 ą 45% -73.3% 118.12 ą 91% interrupts.CPU43.79:PCI-MSI.31981612-edge.i40e-eth0-TxRx-43
> 12070 ą 15% -37.2% 7581 ą 49% interrupts.CPU43.TLB:TLB_shootdowns
> 7177 -27.6% 5195 ą 26% interrupts.CPU45.NMI:Non-maskable_interrupts
> 7177 -27.6% 5195 ą 26% interrupts.CPU45.PMI:Performance_monitoring_interrupts
> 271.00 ą 8% -53.4% 126.38 ą 75% interrupts.CPU46.82:PCI-MSI.31981615-edge.i40e-eth0-TxRx-46
> 3591 +84.0% 6607 ą 12% interrupts.CPU46.NMI:Non-maskable_interrupts
> 3591 +84.0% 6607 ą 12% interrupts.CPU46.PMI:Performance_monitoring_interrupts
> 57614 ą 30% -34.0% 38015 ą 28% interrupts.CPU46.RES:Rescheduling_interrupts
> 149154 ą 41% -47.2% 78808 ą 51% interrupts.CPU51.RES:Rescheduling_interrupts
> 30366 ą 28% +279.5% 115229 ą 42% interrupts.CPU52.RES:Rescheduling_interrupts
> 29690 +355.5% 135237 ą 57% interrupts.CPU54.RES:Rescheduling_interrupts
> 213106 ą 2% -66.9% 70545 ą 43% interrupts.CPU59.RES:Rescheduling_interrupts
> 225753 ą 7% -72.9% 61212 ą 72% interrupts.CPU60.RES:Rescheduling_interrupts
> 12430 ą 14% -41.5% 7276 ą 52% interrupts.CPU61.TLB:TLB_shootdowns
> 44552 ą 22% +229.6% 146864 ą 36% interrupts.CPU65.RES:Rescheduling_interrupts
> 126088 ą 56% -35.3% 81516 ą 73% interrupts.CPU66.RES:Rescheduling_interrupts
> 170880 ą 15% -62.9% 63320 ą 52% interrupts.CPU68.RES:Rescheduling_interrupts
> 186033 ą 10% -39.8% 112012 ą 41% interrupts.CPU69.RES:Rescheduling_interrupts
> 679.50 ą 29% -80.5% 132.25 ą 82% interrupts.CPU7.43:PCI-MSI.31981576-edge.i40e-eth0-TxRx-7
> 124750 ą 18% -39.4% 75553 ą 43% interrupts.CPU7.RES:Rescheduling_interrupts
> 158500 ą 47% -52.1% 75915 ą 67% interrupts.CPU71.RES:Rescheduling_interrupts
> 11846 ą 11% -32.5% 8001 ą 47% interrupts.CPU72.TLB:TLB_shootdowns
> 66095 ą 99% -99.8% 147.62 ą 93% interrupts.CPU73.109:PCI-MSI.31981642-edge.i40e-eth0-TxRx-73
> 7221 ą 2% -31.0% 4982 ą 35% interrupts.CPU73.NMI:Non-maskable_interrupts
> 7221 ą 2% -31.0% 4982 ą 35% interrupts.CPU73.PMI:Performance_monitoring_interrupts
> 15304 ą 14% -47.9% 7972 ą 31% interrupts.CPU73.TLB:TLB_shootdowns
> 10918 ą 3% -31.9% 7436 ą 36% interrupts.CPU74.TLB:TLB_shootdowns
> 543.00 ą 39% -73.3% 144.75 ą 81% interrupts.CPU76.112:PCI-MSI.31981645-edge.i40e-eth0-TxRx-76
> 12214 ą 14% -40.9% 7220 ą 38% interrupts.CPU79.TLB:TLB_shootdowns
> 168.00 ą 29% -55.7% 74.50 ą 85% interrupts.CPU80.116:PCI-MSI.31981649-edge.i40e-eth0-TxRx-80
> 28619 ą 3% +158.4% 73939 ą 44% interrupts.CPU80.RES:Rescheduling_interrupts
> 12258 -34.3% 8056 ą 29% interrupts.CPU80.TLB:TLB_shootdowns
> 7214 -19.5% 5809 ą 24% interrupts.CPU82.NMI:Non-maskable_interrupts
> 7214 -19.5% 5809 ą 24% interrupts.CPU82.PMI:Performance_monitoring_interrupts
> 13522 ą 11% -41.2% 7949 ą 29% interrupts.CPU84.TLB:TLB_shootdowns
> 223.50 ą 14% -57.8% 94.25 ą 88% interrupts.CPU85.121:PCI-MSI.31981654-edge.i40e-eth0-TxRx-85
> 11989 ą 2% -31.7% 8194 ą 22% interrupts.CPU85.TLB:TLB_shootdowns
> 121153 ą 29% -41.4% 70964 ą 58% interrupts.CPU86.RES:Rescheduling_interrupts
> 11731 ą 8% -40.7% 6957 ą 36% interrupts.CPU86.TLB:TLB_shootdowns
> 12192 ą 22% -35.8% 7824 ą 43% interrupts.CPU87.TLB:TLB_shootdowns
> 11603 ą 19% -31.8% 7915 ą 41% interrupts.CPU89.TLB:TLB_shootdowns
> 10471 ą 5% -27.0% 7641 ą 31% interrupts.CPU91.TLB:TLB_shootdowns
> 7156 -20.9% 5658 ą 23% interrupts.CPU92.NMI:Non-maskable_interrupts
> 7156 -20.9% 5658 ą 23% interrupts.CPU92.PMI:Performance_monitoring_interrupts
> 99802 ą 20% -43.6% 56270 ą 47% interrupts.CPU92.RES:Rescheduling_interrupts
> 109162 ą 18% -28.7% 77839 ą 26% interrupts.CPU93.RES:Rescheduling_interrupts
> 15044 ą 29% -44.4% 8359 ą 30% interrupts.CPU93.TLB:TLB_shootdowns
> 110749 ą 19% -47.3% 58345 ą 48% interrupts.CPU94.RES:Rescheduling_interrupts
> 7245 -21.4% 5697 ą 25% interrupts.CPU95.NMI:Non-maskable_interrupts
> 7245 -21.4% 5697 ą 25% interrupts.CPU95.PMI:Performance_monitoring_interrupts
> 1969 ą 5% +491.7% 11653 ą 81% interrupts.IWI:IRQ_work_interrupts
>
>
>
> ***************************************************************************************************
> lkp-bdw-ep6: 88 threads Intel(R) Xeon(R) CPU E5-2699 v4 @ 2.20GHz with 128G memory
> =========================================================================================
> class/compiler/cpufreq_governor/disk/kconfig/nr_threads/rootfs/tbox_group/testcase/testtime/ucode:
> interrupt/gcc-7/performance/1HDD/x86_64-rhel-7.6/100%/debian-x86_64-2019-09-23.cgz/lkp-bdw-ep6/stress-ng/30s/0xb000038
>
> commit:
> fcf0553db6 ("sched/fair: Remove meaningless imbalance calculation")
> 0b0695f2b3 ("sched/fair: Rework load_balance()")
>
> fcf0553db6f4c793 0b0695f2b34a4afa3f6e9aa1ff0
> ---------------- ---------------------------
> %stddev %change %stddev
> \ | \
> 98318389 +43.0% 1.406e+08 stress-ng.schedpolicy.ops
> 3277346 +43.0% 4685146 stress-ng.schedpolicy.ops_per_sec
> 3.506e+08 ą 4% -10.3% 3.146e+08 ą 3% stress-ng.sigq.ops
> 11684738 ą 4% -10.3% 10485353 ą 3% stress-ng.sigq.ops_per_sec
> 3.628e+08 ą 6% -19.4% 2.925e+08 ą 6% stress-ng.time.involuntary_context_switches
> 29456 +2.8% 30285 stress-ng.time.system_time
> 7636655 ą 9% +46.6% 11197377 ą 27% cpuidle.C1E.usage
> 1111483 ą 3% -9.5% 1005829 vmstat.system.cs
> 22638222 ą 4% +16.5% 26370816 ą 11% meminfo.Committed_AS
> 28908 ą 6% +24.6% 36020 ą 16% meminfo.KernelStack
> 7636543 ą 9% +46.6% 11196090 ą 27% turbostat.C1E
> 3.46 ą 16% -61.2% 1.35 ą 7% turbostat.Pkg%pc2
> 217.54 +1.7% 221.33 turbostat.PkgWatt
> 13.34 ą 2% +5.8% 14.11 turbostat.RAMWatt
> 525.50 ą 8% -15.7% 443.00 ą 12% slabinfo.biovec-128.active_objs
> 525.50 ą 8% -15.7% 443.00 ą 12% slabinfo.biovec-128.num_objs
> 28089 ą 12% -33.0% 18833 ą 22% slabinfo.pool_workqueue.active_objs
> 877.25 ą 12% -32.6% 591.00 ą 21% slabinfo.pool_workqueue.active_slabs
> 28089 ą 12% -32.6% 18925 ą 21% slabinfo.pool_workqueue.num_objs
> 877.25 ą 12% -32.6% 591.00 ą 21% slabinfo.pool_workqueue.num_slabs
> 846.75 ą 6% -18.0% 694.75 ą 9% slabinfo.skbuff_fclone_cache.active_objs
> 846.75 ą 6% -18.0% 694.75 ą 9% slabinfo.skbuff_fclone_cache.num_objs
> 63348 ą 6% -20.7% 50261 ą 4% softirqs.CPU14.SCHED
> 44394 ą 4% +21.4% 53880 ą 8% softirqs.CPU42.SCHED
> 52246 ą 7% -15.1% 44352 softirqs.CPU47.SCHED
> 58350 ą 4% -11.0% 51914 ą 7% softirqs.CPU6.SCHED
> 58009 ą 7% -23.8% 44206 ą 4% softirqs.CPU63.SCHED
> 49166 ą 6% +23.4% 60683 ą 9% softirqs.CPU68.SCHED
> 44594 ą 7% +14.3% 50951 ą 8% softirqs.CPU78.SCHED
> 46407 ą 9% +19.6% 55515 ą 8% softirqs.CPU84.SCHED
> 55555 ą 8% -15.5% 46933 ą 4% softirqs.CPU9.SCHED
> 198757 ą 18% +44.1% 286316 ą 9% numa-meminfo.node0.Active
> 189280 ą 19% +37.1% 259422 ą 7% numa-meminfo.node0.Active(anon)
> 110438 ą 33% +68.3% 185869 ą 16% numa-meminfo.node0.AnonHugePages
> 143458 ą 28% +67.7% 240547 ą 13% numa-meminfo.node0.AnonPages
> 12438 ą 16% +61.9% 20134 ą 37% numa-meminfo.node0.KernelStack
> 1004379 ą 7% +16.4% 1168764 ą 4% numa-meminfo.node0.MemUsed
> 357111 ą 24% -41.6% 208655 ą 29% numa-meminfo.node1.Active
> 330094 ą 22% -39.6% 199339 ą 32% numa-meminfo.node1.Active(anon)
> 265924 ą 25% -52.2% 127138 ą 46% numa-meminfo.node1.AnonHugePages
> 314059 ą 22% -49.6% 158305 ą 36% numa-meminfo.node1.AnonPages
> 15386 ą 16% -25.1% 11525 ą 15% numa-meminfo.node1.KernelStack
> 1200805 ą 11% -18.6% 977595 ą 7% numa-meminfo.node1.MemUsed
> 965.50 ą 15% -29.3% 682.25 ą 43% numa-meminfo.node1.Mlocked
> 46762 ą 18% +37.8% 64452 ą 8% numa-vmstat.node0.nr_active_anon
> 35393 ą 27% +68.9% 59793 ą 12% numa-vmstat.node0.nr_anon_pages
> 52.75 ą 33% +71.1% 90.25 ą 15% numa-vmstat.node0.nr_anon_transparent_hugepages
> 15.00 ą 96% +598.3% 104.75 ą 15% numa-vmstat.node0.nr_inactive_file
> 11555 ą 22% +68.9% 19513 ą 41% numa-vmstat.node0.nr_kernel_stack
> 550.25 ą162% +207.5% 1691 ą 48% numa-vmstat.node0.nr_written
> 46762 ą 18% +37.8% 64452 ą 8% numa-vmstat.node0.nr_zone_active_anon
> 15.00 ą 96% +598.3% 104.75 ą 15% numa-vmstat.node0.nr_zone_inactive_file
> 82094 ą 22% -39.5% 49641 ą 32% numa-vmstat.node1.nr_active_anon
> 78146 ą 23% -49.5% 39455 ą 37% numa-vmstat.node1.nr_anon_pages
> 129.00 ą 25% -52.3% 61.50 ą 47% numa-vmstat.node1.nr_anon_transparent_hugepages
> 107.75 ą 12% -85.4% 15.75 ą103% numa-vmstat.node1.nr_inactive_file
> 14322 ą 11% -21.1% 11304 ą 11% numa-vmstat.node1.nr_kernel_stack
> 241.00 ą 15% -29.5% 170.00 ą 43% numa-vmstat.node1.nr_mlock
> 82094 ą 22% -39.5% 49641 ą 32% numa-vmstat.node1.nr_zone_active_anon
> 107.75 ą 12% -85.4% 15.75 ą103% numa-vmstat.node1.nr_zone_inactive_file
> 0.81 ą 5% +0.2 0.99 ą 10% perf-profile.calltrace.cycles-pp.task_rq_lock.task_sched_runtime.cpu_clock_sample.posix_cpu_timer_get.do_timer_gettime
> 0.60 ą 11% +0.2 0.83 ą 9% perf-profile.calltrace.cycles-pp.___might_sleep.__might_fault._copy_to_user.put_itimerspec64.__x64_sys_timer_gettime
> 1.73 ą 9% +0.3 2.05 ą 8% perf-profile.calltrace.cycles-pp.__might_fault._copy_to_user.put_itimerspec64.__x64_sys_timer_gettime.do_syscall_64
> 3.92 ą 5% +0.6 4.49 ą 7% perf-profile.calltrace.cycles-pp.task_sched_runtime.cpu_clock_sample.posix_cpu_timer_get.do_timer_gettime.__x64_sys_timer_gettime
> 4.17 ą 4% +0.6 4.78 ą 7% perf-profile.calltrace.cycles-pp.cpu_clock_sample.posix_cpu_timer_get.do_timer_gettime.__x64_sys_timer_gettime.do_syscall_64
> 5.72 ą 3% +0.7 6.43 ą 7% perf-profile.calltrace.cycles-pp.posix_cpu_timer_get.do_timer_gettime.__x64_sys_timer_gettime.do_syscall_64.entry_SYSCALL_64_after_hwframe
> 0.24 ą 54% -0.2 0.07 ą131% perf-profile.children.cycles-pp.ext4_inode_csum_set
> 0.45 ą 3% +0.1 0.56 ą 4% perf-profile.children.cycles-pp.__might_sleep
> 0.84 ą 5% +0.2 1.03 ą 9% perf-profile.children.cycles-pp.task_rq_lock
> 0.66 ą 8% +0.2 0.88 ą 7% perf-profile.children.cycles-pp.___might_sleep
> 1.83 ą 9% +0.3 2.16 ą 8% perf-profile.children.cycles-pp.__might_fault
> 4.04 ą 5% +0.6 4.62 ą 7% perf-profile.children.cycles-pp.task_sched_runtime
> 4.24 ą 4% +0.6 4.87 ą 7% perf-profile.children.cycles-pp.cpu_clock_sample
> 5.77 ą 3% +0.7 6.48 ą 7% perf-profile.children.cycles-pp.posix_cpu_timer_get
> 0.22 ą 11% +0.1 0.28 ą 15% perf-profile.self.cycles-pp.cpu_clock_sample
> 0.47 ą 7% +0.1 0.55 ą 5% perf-profile.self.cycles-pp.update_curr
> 0.28 ą 5% +0.1 0.38 ą 14% perf-profile.self.cycles-pp.task_rq_lock
> 0.42 ą 3% +0.1 0.53 ą 4% perf-profile.self.cycles-pp.__might_sleep
> 0.50 ą 5% +0.1 0.61 ą 11% perf-profile.self.cycles-pp.task_sched_runtime
> 0.63 ą 9% +0.2 0.85 ą 7% perf-profile.self.cycles-pp.___might_sleep
> 9180611 ą 5% +40.1% 12859327 ą 14% sched_debug.cfs_rq:/.MIN_vruntime.max
> 1479571 ą 6% +57.6% 2331469 ą 14% sched_debug.cfs_rq:/.MIN_vruntime.stddev
> 7951 ą 6% -52.5% 3773 ą 17% sched_debug.cfs_rq:/.exec_clock.stddev
> 321306 ą 39% -44.2% 179273 sched_debug.cfs_rq:/.load.max
> 9180613 ą 5% +40.1% 12859327 ą 14% sched_debug.cfs_rq:/.max_vruntime.max
> 1479571 ą 6% +57.6% 2331469 ą 14% sched_debug.cfs_rq:/.max_vruntime.stddev
> 16622378 +20.0% 19940069 ą 7% sched_debug.cfs_rq:/.min_vruntime.avg
> 18123901 +19.7% 21686545 ą 6% sched_debug.cfs_rq:/.min_vruntime.max
> 14338218 ą 3% +27.4% 18267927 ą 7% sched_debug.cfs_rq:/.min_vruntime.min
> 0.17 ą 16% +23.4% 0.21 ą 11% sched_debug.cfs_rq:/.nr_running.stddev
> 319990 ą 39% -44.6% 177347 sched_debug.cfs_rq:/.runnable_weight.max
> -2067420 -33.5% -1375445 sched_debug.cfs_rq:/.spread0.min
> 1033 ą 8% -13.7% 891.85 ą 3% sched_debug.cfs_rq:/.util_est_enqueued.max
> 93676 ą 16% -29.0% 66471 ą 17% sched_debug.cpu.avg_idle.min
> 10391 ą 52% +118.9% 22750 ą 15% sched_debug.cpu.curr->pid.avg
> 14393 ą 35% +113.2% 30689 ą 17% sched_debug.cpu.curr->pid.max
> 3041 ą 38% +161.8% 7963 ą 11% sched_debug.cpu.curr->pid.stddev
> 3.38 ą 6% -16.3% 2.83 ą 5% sched_debug.cpu.nr_running.max
> 2412687 ą 4% -16.0% 2027251 ą 3% sched_debug.cpu.nr_switches.avg
> 4038819 ą 3% -20.2% 3223112 ą 5% sched_debug.cpu.nr_switches.max
> 834203 ą 17% -37.8% 518798 ą 27% sched_debug.cpu.nr_switches.stddev
> 45.85 ą 13% +41.2% 64.75 ą 18% sched_debug.cpu.nr_uninterruptible.max
> 1937209 ą 2% +58.5% 3070891 ą 3% sched_debug.cpu.sched_count.min
> 1074023 ą 13% -57.9% 451958 ą 12% sched_debug.cpu.sched_count.stddev
> 1283769 ą 7% +65.1% 2118907 ą 7% sched_debug.cpu.yld_count.min
> 714244 ą 5% -51.9% 343373 ą 22% sched_debug.cpu.yld_count.stddev
> 12.54 ą 9% -18.8% 10.18 ą 15% perf-stat.i.MPKI
> 1.011e+10 +2.6% 1.038e+10 perf-stat.i.branch-instructions
> 13.22 ą 5% +2.5 15.75 ą 3% perf-stat.i.cache-miss-rate%
> 21084021 ą 6% +33.9% 28231058 ą 6% perf-stat.i.cache-misses
> 1143861 ą 5% -12.1% 1005721 ą 6% perf-stat.i.context-switches
> 1.984e+11 +1.8% 2.02e+11 perf-stat.i.cpu-cycles
> 1.525e+10 +1.3% 1.544e+10 perf-stat.i.dTLB-loads
> 65.46 -2.7 62.76 ą 3% perf-stat.i.iTLB-load-miss-rate%
> 20360883 ą 4% +10.5% 22500874 ą 4% perf-stat.i.iTLB-loads
> 4.963e+10 +2.0% 5.062e+10 perf-stat.i.instructions
> 181557 -2.4% 177113 perf-stat.i.msec
> 5350122 ą 8% +26.5% 6765332 ą 7% perf-stat.i.node-load-misses
> 4264320 ą 3% +24.8% 5321600 ą 4% perf-stat.i.node-store-misses
> 6.12 ą 5% +1.5 7.60 ą 2% perf-stat.overall.cache-miss-rate%
> 7646 ą 6% -17.7% 6295 ą 3% perf-stat.overall.cycles-between-cache-misses
> 69.29 -1.1 68.22 perf-stat.overall.iTLB-load-miss-rate%
> 61.11 ą 2% +6.6 67.71 ą 5% perf-stat.overall.node-load-miss-rate%
> 74.82 +1.8 76.58 perf-stat.overall.node-store-miss-rate%
> 1.044e+10 +1.8% 1.063e+10 perf-stat.ps.branch-instructions
> 26325951 ą 6% +22.9% 32366684 ą 2% perf-stat.ps.cache-misses
> 1115530 ą 3% -9.5% 1009780 perf-stat.ps.context-switches
> 1.536e+10 +1.0% 1.552e+10 perf-stat.ps.dTLB-loads
> 44718416 ą 2% +5.8% 47308605 ą 3% perf-stat.ps.iTLB-load-misses
> 19831973 ą 4% +11.1% 22040029 ą 4% perf-stat.ps.iTLB-loads
> 5.064e+10 +1.4% 5.137e+10 perf-stat.ps.instructions
> 5454694 ą 9% +26.4% 6892365 ą 6% perf-stat.ps.node-load-misses
> 4263688 ą 4% +24.9% 5325279 ą 4% perf-stat.ps.node-store-misses
> 3.001e+13 +1.7% 3.052e+13 perf-stat.total.instructions
> 18550 -74.9% 4650 ą173% interrupts.76:IR-PCI-MSI.512000-edge.ahci[0000:00:1f.2]
> 7642 ą 9% -20.4% 6086 ą 2% interrupts.CPU0.CAL:Function_call_interrupts
> 4376 ą 22% -75.4% 1077 ą 41% interrupts.CPU0.TLB:TLB_shootdowns
> 8402 ą 5% -19.0% 6806 interrupts.CPU1.CAL:Function_call_interrupts
> 4559 ą 20% -73.7% 1199 ą 15% interrupts.CPU1.TLB:TLB_shootdowns
> 8423 ą 4% -20.2% 6725 ą 2% interrupts.CPU10.CAL:Function_call_interrupts
> 4536 ą 14% -75.0% 1135 ą 20% interrupts.CPU10.TLB:TLB_shootdowns
> 8303 ą 3% -18.2% 6795 ą 2% interrupts.CPU11.CAL:Function_call_interrupts
> 4404 ą 11% -71.6% 1250 ą 35% interrupts.CPU11.TLB:TLB_shootdowns
> 8491 ą 6% -21.3% 6683 interrupts.CPU12.CAL:Function_call_interrupts
> 4723 ą 20% -77.2% 1077 ą 17% interrupts.CPU12.TLB:TLB_shootdowns
> 8403 ą 5% -20.3% 6700 ą 2% interrupts.CPU13.CAL:Function_call_interrupts
> 4557 ą 19% -74.2% 1175 ą 22% interrupts.CPU13.TLB:TLB_shootdowns
> 8459 ą 4% -18.6% 6884 interrupts.CPU14.CAL:Function_call_interrupts
> 4559 ą 18% -69.8% 1376 ą 13% interrupts.CPU14.TLB:TLB_shootdowns
> 8305 ą 7% -17.7% 6833 ą 2% interrupts.CPU15.CAL:Function_call_interrupts
> 4261 ą 25% -67.6% 1382 ą 24% interrupts.CPU15.TLB:TLB_shootdowns
> 8277 ą 5% -19.1% 6696 ą 3% interrupts.CPU16.CAL:Function_call_interrupts
> 4214 ą 22% -69.6% 1282 ą 8% interrupts.CPU16.TLB:TLB_shootdowns
> 8258 ą 5% -18.9% 6694 ą 3% interrupts.CPU17.CAL:Function_call_interrupts
> 4461 ą 19% -74.1% 1155 ą 21% interrupts.CPU17.TLB:TLB_shootdowns
> 8457 ą 6% -20.6% 6717 interrupts.CPU18.CAL:Function_call_interrupts
> 4889 ą 34% +60.0% 7822 interrupts.CPU18.NMI:Non-maskable_interrupts
> 4889 ą 34% +60.0% 7822 interrupts.CPU18.PMI:Performance_monitoring_interrupts
> 4731 ą 22% -77.2% 1078 ą 10% interrupts.CPU18.TLB:TLB_shootdowns
> 8160 ą 5% -18.1% 6684 interrupts.CPU19.CAL:Function_call_interrupts
> 4311 ą 20% -74.2% 1114 ą 13% interrupts.CPU19.TLB:TLB_shootdowns
> 8464 ą 2% -18.2% 6927 ą 3% interrupts.CPU2.CAL:Function_call_interrupts
> 4938 ą 14% -70.5% 1457 ą 18% interrupts.CPU2.TLB:TLB_shootdowns
> 8358 ą 6% -19.7% 6715 ą 3% interrupts.CPU20.CAL:Function_call_interrupts
> 4567 ą 24% -74.6% 1160 ą 35% interrupts.CPU20.TLB:TLB_shootdowns
> 8460 ą 4% -22.3% 6577 ą 2% interrupts.CPU21.CAL:Function_call_interrupts
> 4514 ą 18% -76.0% 1084 ą 22% interrupts.CPU21.TLB:TLB_shootdowns
> 6677 ą 6% +19.6% 7988 ą 9% interrupts.CPU22.CAL:Function_call_interrupts
> 1288 ą 14% +209.1% 3983 ą 35% interrupts.CPU22.TLB:TLB_shootdowns
> 6751 ą 2% +24.0% 8370 ą 9% interrupts.CPU23.CAL:Function_call_interrupts
> 1037 ą 29% +323.0% 4388 ą 36% interrupts.CPU23.TLB:TLB_shootdowns
> 6844 +20.6% 8251 ą 9% interrupts.CPU24.CAL:Function_call_interrupts
> 1205 ą 17% +229.2% 3967 ą 40% interrupts.CPU24.TLB:TLB_shootdowns
> 6880 +21.9% 8389 ą 7% interrupts.CPU25.CAL:Function_call_interrupts
> 1228 ą 19% +245.2% 4240 ą 35% interrupts.CPU25.TLB:TLB_shootdowns
> 6494 ą 8% +25.1% 8123 ą 9% interrupts.CPU26.CAL:Function_call_interrupts
> 1141 ą 13% +262.5% 4139 ą 32% interrupts.CPU26.TLB:TLB_shootdowns
> 6852 +19.2% 8166 ą 7% interrupts.CPU27.CAL:Function_call_interrupts
> 1298 ą 8% +197.1% 3857 ą 31% interrupts.CPU27.TLB:TLB_shootdowns
> 6563 ą 6% +25.2% 8214 ą 8% interrupts.CPU28.CAL:Function_call_interrupts
> 1176 ą 8% +237.1% 3964 ą 33% interrupts.CPU28.TLB:TLB_shootdowns
> 6842 ą 2% +21.4% 8308 ą 8% interrupts.CPU29.CAL:Function_call_interrupts
> 1271 ą 11% +223.8% 4118 ą 33% interrupts.CPU29.TLB:TLB_shootdowns
> 8418 ą 3% -21.1% 6643 ą 2% interrupts.CPU3.CAL:Function_call_interrupts
> 4677 ą 11% -75.1% 1164 ą 16% interrupts.CPU3.TLB:TLB_shootdowns
> 6798 ą 3% +21.8% 8284 ą 7% interrupts.CPU30.CAL:Function_call_interrupts
> 1219 ą 12% +236.3% 4102 ą 30% interrupts.CPU30.TLB:TLB_shootdowns
> 6503 ą 4% +25.9% 8186 ą 6% interrupts.CPU31.CAL:Function_call_interrupts
> 1046 ą 15% +289.1% 4072 ą 32% interrupts.CPU31.TLB:TLB_shootdowns
> 6949 ą 3% +17.2% 8141 ą 8% interrupts.CPU32.CAL:Function_call_interrupts
> 1241 ą 23% +210.6% 3854 ą 34% interrupts.CPU32.TLB:TLB_shootdowns
> 1487 ą 26% +161.6% 3889 ą 46% interrupts.CPU33.TLB:TLB_shootdowns
> 1710 ą 44% +140.1% 4105 ą 36% interrupts.CPU34.TLB:TLB_shootdowns
> 6957 ą 2% +15.2% 8012 ą 9% interrupts.CPU35.CAL:Function_call_interrupts
> 1165 ą 8% +223.1% 3765 ą 38% interrupts.CPU35.TLB:TLB_shootdowns
> 1423 ą 24% +173.4% 3892 ą 33% interrupts.CPU36.TLB:TLB_shootdowns
> 1279 ą 29% +224.2% 4148 ą 39% interrupts.CPU37.TLB:TLB_shootdowns
> 1301 ą 20% +226.1% 4244 ą 35% interrupts.CPU38.TLB:TLB_shootdowns
> 6906 ą 2% +18.5% 8181 ą 8% interrupts.CPU39.CAL:Function_call_interrupts
> 368828 ą 20% +96.2% 723710 ą 7% interrupts.CPU39.RES:Rescheduling_interrupts
> 1438 ą 12% +174.8% 3951 ą 33% interrupts.CPU39.TLB:TLB_shootdowns
> 8399 ą 5% -19.2% 6788 ą 2% interrupts.CPU4.CAL:Function_call_interrupts
> 4567 ą 18% -72.7% 1245 ą 28% interrupts.CPU4.TLB:TLB_shootdowns
> 6895 +22.4% 8439 ą 9% interrupts.CPU40.CAL:Function_call_interrupts
> 1233 ą 11% +247.1% 4280 ą 36% interrupts.CPU40.TLB:TLB_shootdowns
> 6819 ą 2% +21.3% 8274 ą 9% interrupts.CPU41.CAL:Function_call_interrupts
> 1260 ą 14% +207.1% 3871 ą 38% interrupts.CPU41.TLB:TLB_shootdowns
> 1301 ą 9% +204.7% 3963 ą 36% interrupts.CPU42.TLB:TLB_shootdowns
> 6721 ą 3% +22.3% 8221 ą 7% interrupts.CPU43.CAL:Function_call_interrupts
> 1237 ą 19% +224.8% 4017 ą 35% interrupts.CPU43.TLB:TLB_shootdowns
> 8422 ą 8% -22.7% 6506 ą 5% interrupts.CPU44.CAL:Function_call_interrupts
> 15261375 ą 7% -7.8% 14064176 interrupts.CPU44.LOC:Local_timer_interrupts
> 4376 ą 25% -75.7% 1063 ą 26% interrupts.CPU44.TLB:TLB_shootdowns
> 8451 ą 5% -23.7% 6448 ą 6% interrupts.CPU45.CAL:Function_call_interrupts
> 4351 ą 18% -74.9% 1094 ą 12% interrupts.CPU45.TLB:TLB_shootdowns
> 8705 ą 6% -21.2% 6860 ą 2% interrupts.CPU46.CAL:Function_call_interrupts
> 4787 ą 20% -69.5% 1462 ą 16% interrupts.CPU46.TLB:TLB_shootdowns
> 8334 ą 3% -18.9% 6763 interrupts.CPU47.CAL:Function_call_interrupts
> 4126 ą 10% -71.3% 1186 ą 18% interrupts.CPU47.TLB:TLB_shootdowns
> 8578 ą 4% -21.7% 6713 interrupts.CPU48.CAL:Function_call_interrupts
> 4520 ą 15% -74.5% 1154 ą 23% interrupts.CPU48.TLB:TLB_shootdowns
> 8450 ą 8% -18.8% 6863 ą 3% interrupts.CPU49.CAL:Function_call_interrupts
> 4494 ą 24% -66.5% 1505 ą 22% interrupts.CPU49.TLB:TLB_shootdowns
> 8307 ą 4% -18.0% 6816 ą 2% interrupts.CPU5.CAL:Function_call_interrupts
> 7845 -37.4% 4908 ą 34% interrupts.CPU5.NMI:Non-maskable_interrupts
> 7845 -37.4% 4908 ą 34% interrupts.CPU5.PMI:Performance_monitoring_interrupts
> 4429 ą 17% -69.8% 1339 ą 20% interrupts.CPU5.TLB:TLB_shootdowns
> 8444 ą 4% -21.7% 6613 interrupts.CPU50.CAL:Function_call_interrupts
> 4282 ą 16% -76.0% 1029 ą 17% interrupts.CPU50.TLB:TLB_shootdowns
> 8750 ą 6% -22.2% 6803 interrupts.CPU51.CAL:Function_call_interrupts
> 4755 ą 20% -73.1% 1277 ą 15% interrupts.CPU51.TLB:TLB_shootdowns
> 8478 ą 6% -20.2% 6766 ą 2% interrupts.CPU52.CAL:Function_call_interrupts
> 4337 ą 20% -72.6% 1190 ą 22% interrupts.CPU52.TLB:TLB_shootdowns
> 8604 ą 7% -21.5% 6750 ą 4% interrupts.CPU53.CAL:Function_call_interrupts
> 4649 ą 17% -74.3% 1193 ą 23% interrupts.CPU53.TLB:TLB_shootdowns
> 8317 ą 9% -19.4% 6706 ą 3% interrupts.CPU54.CAL:Function_call_interrupts
> 4372 ą 12% -75.4% 1076 ą 29% interrupts.CPU54.TLB:TLB_shootdowns
> 8439 ą 3% -18.5% 6876 interrupts.CPU55.CAL:Function_call_interrupts
> 4415 ą 11% -71.6% 1254 ą 17% interrupts.CPU55.TLB:TLB_shootdowns
> 8869 ą 6% -22.6% 6864 ą 2% interrupts.CPU56.CAL:Function_call_interrupts
> 517594 ą 13% +123.3% 1155539 ą 25% interrupts.CPU56.RES:Rescheduling_interrupts
> 5085 ą 22% -74.9% 1278 ą 17% interrupts.CPU56.TLB:TLB_shootdowns
> 8682 ą 4% -21.7% 6796 ą 2% interrupts.CPU57.CAL:Function_call_interrupts
> 4808 ą 19% -74.1% 1243 ą 13% interrupts.CPU57.TLB:TLB_shootdowns
> 8626 ą 7% -21.8% 6746 ą 2% interrupts.CPU58.CAL:Function_call_interrupts
> 4816 ą 20% -79.1% 1007 ą 28% interrupts.CPU58.TLB:TLB_shootdowns
> 8759 ą 8% -20.3% 6984 interrupts.CPU59.CAL:Function_call_interrupts
> 4840 ą 22% -70.6% 1423 ą 14% interrupts.CPU59.TLB:TLB_shootdowns
> 8167 ą 6% -19.0% 6615 ą 2% interrupts.CPU6.CAL:Function_call_interrupts
> 4129 ą 21% -75.4% 1017 ą 24% interrupts.CPU6.TLB:TLB_shootdowns
> 8910 ą 4% -23.7% 6794 ą 3% interrupts.CPU60.CAL:Function_call_interrupts
> 5017 ą 12% -77.8% 1113 ą 15% interrupts.CPU60.TLB:TLB_shootdowns
> 8689 ą 5% -21.6% 6808 interrupts.CPU61.CAL:Function_call_interrupts
> 4715 ą 20% -77.6% 1055 ą 19% interrupts.CPU61.TLB:TLB_shootdowns
> 8574 ą 4% -18.9% 6953 ą 2% interrupts.CPU62.CAL:Function_call_interrupts
> 4494 ą 17% -72.3% 1244 ą 7% interrupts.CPU62.TLB:TLB_shootdowns
> 8865 ą 3% -25.4% 6614 ą 7% interrupts.CPU63.CAL:Function_call_interrupts
> 4870 ą 12% -76.8% 1130 ą 12% interrupts.CPU63.TLB:TLB_shootdowns
> 8724 ą 7% -20.2% 6958 ą 3% interrupts.CPU64.CAL:Function_call_interrupts
> 4736 ą 16% -72.6% 1295 ą 7% interrupts.CPU64.TLB:TLB_shootdowns
> 8717 ą 6% -23.7% 6653 ą 4% interrupts.CPU65.CAL:Function_call_interrupts
> 4626 ą 19% -76.5% 1087 ą 21% interrupts.CPU65.TLB:TLB_shootdowns
> 6671 +24.7% 8318 ą 9% interrupts.CPU66.CAL:Function_call_interrupts
> 1091 ą 8% +249.8% 3819 ą 32% interrupts.CPU66.TLB:TLB_shootdowns
> 6795 ą 2% +26.9% 8624 ą 9% interrupts.CPU67.CAL:Function_call_interrupts
> 1098 ą 24% +299.5% 4388 ą 39% interrupts.CPU67.TLB:TLB_shootdowns
> 6704 ą 5% +25.8% 8431 ą 8% interrupts.CPU68.CAL:Function_call_interrupts
> 1214 ą 15% +236.1% 4083 ą 36% interrupts.CPU68.TLB:TLB_shootdowns
> 1049 ą 15% +326.2% 4473 ą 33% interrupts.CPU69.TLB:TLB_shootdowns
> 8554 ą 6% -19.6% 6874 ą 2% interrupts.CPU7.CAL:Function_call_interrupts
> 4753 ą 19% -71.7% 1344 ą 16% interrupts.CPU7.TLB:TLB_shootdowns
> 1298 ą 13% +227.4% 4249 ą 38% interrupts.CPU70.TLB:TLB_shootdowns
> 6976 +19.9% 8362 ą 7% interrupts.CPU71.CAL:Function_call_interrupts
> 1232748 ą 18% -57.3% 525824 ą 33% interrupts.CPU71.RES:Rescheduling_interrupts
> 1253 ą 9% +211.8% 3909 ą 31% interrupts.CPU71.TLB:TLB_shootdowns
> 1316 ą 22% +188.7% 3800 ą 33% interrupts.CPU72.TLB:TLB_shootdowns
> 6665 ą 5% +26.5% 8429 ą 8% interrupts.CPU73.CAL:Function_call_interrupts
> 1202 ą 13% +234.1% 4017 ą 37% interrupts.CPU73.TLB:TLB_shootdowns
> 6639 ą 5% +27.0% 8434 ą 8% interrupts.CPU74.CAL:Function_call_interrupts
> 1079 ą 16% +269.4% 3986 ą 36% interrupts.CPU74.TLB:TLB_shootdowns
> 1055 ą 12% +301.2% 4235 ą 34% interrupts.CPU75.TLB:TLB_shootdowns
> 7011 ą 3% +21.6% 8522 ą 8% interrupts.CPU76.CAL:Function_call_interrupts
> 1223 ą 13% +230.7% 4047 ą 35% interrupts.CPU76.TLB:TLB_shootdowns
> 6886 ą 7% +25.6% 8652 ą 10% interrupts.CPU77.CAL:Function_call_interrupts
> 1316 ą 16% +229.8% 4339 ą 36% interrupts.CPU77.TLB:TLB_shootdowns
> 7343 ą 5% +19.1% 8743 ą 9% interrupts.CPU78.CAL:Function_call_interrupts
> 1699 ą 37% +144.4% 4152 ą 31% interrupts.CPU78.TLB:TLB_shootdowns
> 7136 ą 4% +21.4% 8666 ą 9% interrupts.CPU79.CAL:Function_call_interrupts
> 1094 ą 13% +276.2% 4118 ą 34% interrupts.CPU79.TLB:TLB_shootdowns
> 8531 ą 5% -19.5% 6869 ą 2% interrupts.CPU8.CAL:Function_call_interrupts
> 4764 ą 16% -71.0% 1382 ą 14% interrupts.CPU8.TLB:TLB_shootdowns
> 1387 ą 29% +181.8% 3910 ą 38% interrupts.CPU80.TLB:TLB_shootdowns
> 1114 ą 30% +259.7% 4007 ą 36% interrupts.CPU81.TLB:TLB_shootdowns
> 7012 +23.9% 8685 ą 8% interrupts.CPU82.CAL:Function_call_interrupts
> 1274 ą 12% +255.4% 4530 ą 27% interrupts.CPU82.TLB:TLB_shootdowns
> 6971 ą 3% +23.8% 8628 ą 9% interrupts.CPU83.CAL:Function_call_interrupts
> 1156 ą 18% +260.1% 4162 ą 34% interrupts.CPU83.TLB:TLB_shootdowns
> 7030 ą 4% +21.0% 8504 ą 8% interrupts.CPU84.CAL:Function_call_interrupts
> 1286 ą 23% +224.0% 4166 ą 31% interrupts.CPU84.TLB:TLB_shootdowns
> 7059 +22.4% 8644 ą 11% interrupts.CPU85.CAL:Function_call_interrupts
> 1421 ą 22% +208.8% 4388 ą 33% interrupts.CPU85.TLB:TLB_shootdowns
> 7018 ą 2% +22.8% 8615 ą 9% interrupts.CPU86.CAL:Function_call_interrupts
> 1258 ą 8% +231.1% 4167 ą 34% interrupts.CPU86.TLB:TLB_shootdowns
> 1338 ą 3% +217.9% 4255 ą 31% interrupts.CPU87.TLB:TLB_shootdowns
> 8376 ą 4% -19.0% 6787 ą 2% interrupts.CPU9.CAL:Function_call_interrupts
> 4466 ą 17% -71.2% 1286 ą 18% interrupts.CPU9.TLB:TLB_shootdowns
>
>
>
>
>
> Disclaimer:
> Results have been estimated based on internal Intel analysis and are provided
> for informational purposes only. Any difference in system hardware or software
> design or configuration may affect actual performance.
>
>
> Thanks,
> Oliver Sang
>
Powered by blists - more mailing lists