lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Date:   Mon, 16 Jul 2018 04:01:44 -0400 (EDT)
From:   Xiao Ni <xni@...hat.com>
To:     Aaron Lu <aaron.lu@...el.com>
Cc:     kernel test robot <xiaolong.ye@...el.com>,
        Stephen Rothwell <sfr@...b.auug.org.au>, lkp@...org,
        LKML <linux-kernel@...r.kernel.org>, Shaohua Li <shli@...com>,
        Ming Lei <ming.lei@...hat.com>
Subject: Re: [LKP] [lkp-robot] [MD] 5a409b4f56: aim7.jobs-per-min -27.5%
 regression

Hi Aaron

I have no update for this yet. I'll have a look this week and give response later.

Regards
Xiao

----- Original Message -----
> From: "Aaron Lu" <aaron.lu@...el.com>
> To: "Xiao Ni" <xni@...hat.com>
> Cc: "kernel test robot" <xiaolong.ye@...el.com>, "Stephen Rothwell" <sfr@...b.auug.org.au>, lkp@...org, "LKML"
> <linux-kernel@...r.kernel.org>, "Shaohua Li" <shli@...com>, "Ming Lei" <ming.lei@...hat.com>
> Sent: Monday, July 16, 2018 3:54:30 PM
> Subject: Re: [LKP] [lkp-robot] [MD] 5a409b4f56: aim7.jobs-per-min -27.5% regression
> 
> Ping...
> Any update on this?
> Feel free to ask me for any additional data if you need.
> 
> Thanks,
> Aaron
> 
> On Mon, Jun 04, 2018 at 02:42:03PM +0800, kernel test robot wrote:
> > 
> > Greeting,
> > 
> > FYI, we noticed a -27.5% regression of aim7.jobs-per-min due to commit:
> > 
> > 
> > commit: 5a409b4f56d50b212334f338cb8465d65550cd85 ("MD: fix lock contention
> > for flush bios")
> > https://git.kernel.org/cgit/linux/kernel/git/next/linux-next.git master
> > 
> > in testcase: aim7
> > on test machine: 40 threads Intel(R) Xeon(R) CPU E5-2690 v2 @ 3.00GHz with
> > 384G memory
> > with following parameters:
> > 
> > 	disk: 4BRD_12G
> > 	md: RAID1
> > 	fs: xfs
> > 	test: sync_disk_rw
> > 	load: 600
> > 	cpufreq_governor: performance
> > 
> > test-description: AIM7 is a traditional UNIX system level benchmark suite
> > which is used to test and measure the performance of multiuser system.
> > test-url: https://sourceforge.net/projects/aimbench/files/aim-suite7/
> > 
> > 
> > Details are as below:
> > -------------------------------------------------------------------------------------------------->
> > 
> > =========================================================================================
> > compiler/cpufreq_governor/disk/fs/kconfig/load/md/rootfs/tbox_group/test/testcase:
> >   gcc-7/performance/4BRD_12G/xfs/x86_64-rhel-7.2/600/RAID1/debian-x86_64-2016-08-31.cgz/lkp-ivb-ep01/sync_disk_rw/aim7
> > 
> > commit:
> >   448ec638c6 ("md/raid5: Assigning NULL to sh->batch_head before testing
> >   bit R5_Overlap of a stripe")
> >   5a409b4f56 ("MD: fix lock contention for flush bios")
> > 
> > 448ec638c6bcf369 5a409b4f56d50b212334f338cb
> > ---------------- --------------------------
> >          %stddev     %change         %stddev
> >              \          |                \
> >       1640           -27.5%       1189        aim7.jobs-per-min
> >       2194           +37.9%       3026        aim7.time.elapsed_time
> >       2194           +37.9%       3026        aim7.time.elapsed_time.max
> >   50990311           -95.8%    2148266
> >   aim7.time.involuntary_context_switches
> >     107965 ±  4%     -26.4%      79516 ±  2%  aim7.time.minor_page_faults
> >      49.14           +82.5%      89.66 ±  2%  aim7.time.user_time
> >  7.123e+08           -35.7%  4.582e+08
> >  aim7.time.voluntary_context_switches
> >     672282           +36.8%     919615
> >     interrupts.CAL:Function_call_interrupts
> >   16631387 ±  2%     -39.9%    9993075 ±  7%  softirqs.RCU
> >    9708009          +186.1%   27778773        softirqs.SCHED
> >   33436649           +45.5%   48644912        softirqs.TIMER
> >       4.16            -2.1        2.01        mpstat.cpu.idle%
> >       0.24 ±  2%     +27.7       27.91        mpstat.cpu.iowait%
> >      95.51           -25.6       69.94        mpstat.cpu.sys%
> >       0.09            +0.0        0.13        mpstat.cpu.usr%
> >    6051756 ±  3%     +59.0%    9623085
> >    numa-numastat.node0.local_node
> >    6055311 ±  3%     +59.0%    9626996        numa-numastat.node0.numa_hit
> >    6481209 ±  3%     +48.4%    9616310
> >    numa-numastat.node1.local_node
> >    6485866 ±  3%     +48.3%    9620756        numa-numastat.node1.numa_hit
> >      61404           -27.7%      44424        vmstat.io.bo
> >       2.60 ± 18%  +11519.2%     302.10        vmstat.procs.b
> >     304.10           -84.9%      45.80 ±  2%  vmstat.procs.r
> >     400477           -43.5%     226094        vmstat.system.cs
> >     166461           -49.9%      83332        vmstat.system.in
> >      78397           +27.0%      99567        meminfo.Dirty
> >      14427           +18.4%      17082        meminfo.Inactive(anon)
> >       1963 ±  5%      +5.4%       2068 ±  4%  meminfo.Mlocked
> >     101143          +991.0%    1103488        meminfo.SUnreclaim
> >      53684 ±  4%     -18.1%      43946 ±  3%  meminfo.Shmem
> >     175580          +571.4%    1178829        meminfo.Slab
> >      39406           +26.2%      49717        numa-meminfo.node0.Dirty
> >    1767204 ± 10%     +37.2%    2425487 ±  2%  numa-meminfo.node0.MemUsed
> >      51634 ± 18%    +979.3%     557316        numa-meminfo.node0.SUnreclaim
> >      92259 ± 13%    +551.7%     601288        numa-meminfo.node0.Slab
> >      38969           +28.0%      49863        numa-meminfo.node1.Dirty
> >    1895204 ± 10%     +24.7%    2363037 ±  3%  numa-meminfo.node1.MemUsed
> >      49512 ± 19%   +1003.1%     546165        numa-meminfo.node1.SUnreclaim
> >      83323 ± 14%    +593.1%     577534        numa-meminfo.node1.Slab
> >  2.524e+09          +894.5%   2.51e+10        cpuidle.C1.time
> >   50620790          +316.5%  2.109e+08        cpuidle.C1.usage
> >  3.965e+08         +1871.1%  7.815e+09        cpuidle.C1E.time
> >    5987788          +186.1%   17129412        cpuidle.C1E.usage
> >  2.506e+08           +97.5%  4.948e+08 ±  2%  cpuidle.C3.time
> >    2923498           -55.7%    1295033        cpuidle.C3.usage
> >  5.327e+08          +179.9%  1.491e+09        cpuidle.C6.time
> >     779874 ±  2%    +229.3%    2567769        cpuidle.C6.usage
> >    6191357         +3333.6%  2.126e+08        cpuidle.POLL.time
> >     204095         +1982.1%    4249504        cpuidle.POLL.usage
> >       9850           +26.3%      12444        numa-vmstat.node0.nr_dirty
> >      12908 ± 18%    +979.3%     139321
> >      numa-vmstat.node0.nr_slab_unreclaimable
> >       8876           +29.6%      11505
> >       numa-vmstat.node0.nr_zone_write_pending
> >    3486319 ±  4%     +55.1%    5407021        numa-vmstat.node0.numa_hit
> >    3482713 ±  4%     +55.1%    5403066        numa-vmstat.node0.numa_local
> >       9743           +28.1%      12479        numa-vmstat.node1.nr_dirty
> >      12377 ± 19%   +1003.1%     136532
> >      numa-vmstat.node1.nr_slab_unreclaimable
> >       9287           +30.0%      12074
> >       numa-vmstat.node1.nr_zone_write_pending
> >    3678995 ±  4%     +44.8%    5326772        numa-vmstat.node1.numa_hit
> >    3497785 ±  4%     +47.1%    5145705        numa-vmstat.node1.numa_local
> >     252.70          +100.2%     505.90
> >     slabinfo.biovec-max.active_objs
> >     282.70           +99.1%     562.90        slabinfo.biovec-max.num_objs
> >       2978 ± 17%     +52.5%       4543 ± 14%
> >       slabinfo.dmaengine-unmap-16.active_objs
> >       2978 ± 17%     +52.5%       4543 ± 14%
> >       slabinfo.dmaengine-unmap-16.num_objs
> >       2078          +147.9%       5153 ± 11%
> >       slabinfo.ip6_dst_cache.active_objs
> >       2078          +148.1%       5157 ± 11%
> >       slabinfo.ip6_dst_cache.num_objs
> >       5538 ±  2%     +26.2%       6990 ±  3%
> >       slabinfo.kmalloc-1024.active_objs
> >       5586 ±  3%     +27.1%       7097 ±  3%
> >       slabinfo.kmalloc-1024.num_objs
> >       6878           +47.6%      10151 ±  5%
> >       slabinfo.kmalloc-192.active_objs
> >       6889           +47.5%      10160 ±  5%  slabinfo.kmalloc-192.num_objs
> >       9843 ±  5%  +1.6e+05%   16002876
> >       slabinfo.kmalloc-64.active_objs
> >     161.90 ±  4%  +1.5e+05%     250044
> >     slabinfo.kmalloc-64.active_slabs
> >      10386 ±  4%  +1.5e+05%   16002877        slabinfo.kmalloc-64.num_objs
> >     161.90 ±  4%  +1.5e+05%     250044        slabinfo.kmalloc-64.num_slabs
> >     432.80 ± 12%     +45.2%     628.50 ±  6%
> >     slabinfo.nfs_read_data.active_objs
> >     432.80 ± 12%     +45.2%     628.50 ±  6%
> >     slabinfo.nfs_read_data.num_objs
> >       3956           -23.1%       3041
> >       slabinfo.pool_workqueue.active_objs
> >       4098           -19.8%       3286
> >       slabinfo.pool_workqueue.num_objs
> >     360.50 ± 15%     +56.6%     564.70 ± 11%
> >     slabinfo.secpath_cache.active_objs
> >     360.50 ± 15%     +56.6%     564.70 ± 11%
> >     slabinfo.secpath_cache.num_objs
> >      35373 ±  2%      -8.3%      32432        proc-vmstat.nr_active_anon
> >      19595           +27.1%      24914        proc-vmstat.nr_dirty
> >       3607           +18.4%       4270        proc-vmstat.nr_inactive_anon
> >     490.30 ±  5%      +5.4%     516.90 ±  4%  proc-vmstat.nr_mlock
> >      13421 ±  4%     -18.1%      10986 ±  3%  proc-vmstat.nr_shmem
> >      18608            +1.2%      18834
> >      proc-vmstat.nr_slab_reclaimable
> >      25286          +991.0%     275882
> >      proc-vmstat.nr_slab_unreclaimable
> >      35405 ±  2%      -8.3%      32465
> >      proc-vmstat.nr_zone_active_anon
> >       3607           +18.4%       4270
> >       proc-vmstat.nr_zone_inactive_anon
> >      18161           +29.8%      23572
> >      proc-vmstat.nr_zone_write_pending
> >      76941 ±  5%     -36.8%      48622 ±  4%  proc-vmstat.numa_hint_faults
> >      33878 ±  7%     -35.5%      21836 ±  5%
> >      proc-vmstat.numa_hint_faults_local
> >   12568956           +53.3%   19272377        proc-vmstat.numa_hit
> >   12560739           +53.4%   19264015        proc-vmstat.numa_local
> >      17938 ±  3%     -33.5%      11935 ±  2%
> >      proc-vmstat.numa_pages_migrated
> >      78296 ±  5%     -36.0%      50085 ±  4%  proc-vmstat.numa_pte_updates
> >       8848 ±  6%     -38.2%       5466 ±  6%  proc-vmstat.pgactivate
> >    8874568 ±  8%    +368.7%   41590920        proc-vmstat.pgalloc_normal
> >    5435965           +39.2%    7564148        proc-vmstat.pgfault
> >   12863707          +255.1%   45683570        proc-vmstat.pgfree
> >      17938 ±  3%     -33.5%      11935 ±  2%  proc-vmstat.pgmigrate_success
> >  1.379e+13           -40.8%   8.17e+12        perf-stat.branch-instructions
> >       0.30            +0.1        0.42        perf-stat.branch-miss-rate%
> >    4.2e+10           -17.6%  3.462e+10        perf-stat.branch-misses
> >      15.99            +3.8       19.74        perf-stat.cache-miss-rate%
> >  3.779e+10           -21.6%  2.963e+10        perf-stat.cache-misses
> >  2.364e+11           -36.5%  1.501e+11        perf-stat.cache-references
> >  8.795e+08           -22.2%   6.84e+08        perf-stat.context-switches
> >       4.44            -7.2%       4.12        perf-stat.cpi
> >  2.508e+14           -44.5%  1.393e+14        perf-stat.cpu-cycles
> >   36915392           +60.4%   59211221        perf-stat.cpu-migrations
> >       0.29 ±  2%      +0.0        0.34 ±  4%
> >       perf-stat.dTLB-load-miss-rate%
> >   4.14e+10           -30.2%   2.89e+10 ±  4%  perf-stat.dTLB-load-misses
> >  1.417e+13           -40.1%  8.491e+12        perf-stat.dTLB-loads
> >       0.20 ±  4%      -0.0        0.18 ±  5%
> >       perf-stat.dTLB-store-miss-rate%
> >  3.072e+09 ±  4%     -28.0%   2.21e+09 ±  4%  perf-stat.dTLB-store-misses
> >  1.535e+12           -20.2%  1.225e+12        perf-stat.dTLB-stores
> >      90.73           -11.7       79.07
> >      perf-stat.iTLB-load-miss-rate%
> >  8.291e+09            -6.6%  7.743e+09        perf-stat.iTLB-load-misses
> >  8.473e+08          +141.8%  2.049e+09 ±  3%  perf-stat.iTLB-loads
> >  5.646e+13           -40.2%  3.378e+13        perf-stat.instructions
> >       6810           -35.9%       4362
> >       perf-stat.instructions-per-iTLB-miss
> >       0.23            +7.8%       0.24        perf-stat.ipc
> >    5326672           +39.2%    7413706        perf-stat.minor-faults
> >  1.873e+10           -29.9%  1.312e+10        perf-stat.node-load-misses
> >  2.093e+10           -29.2%  1.481e+10        perf-stat.node-loads
> >      39.38            -0.7       38.72
> >      perf-stat.node-store-miss-rate%
> >  1.087e+10           -16.6%  9.069e+09        perf-stat.node-store-misses
> >  1.673e+10           -14.2%  1.435e+10        perf-stat.node-stores
> >    5326695           +39.2%    7413708        perf-stat.page-faults
> >    1875095 ±  7%     -54.8%     846645 ± 16%
> >    sched_debug.cfs_rq:/.MIN_vruntime.avg
> >   32868920 ±  6%     -35.7%   21150379 ± 14%
> >   sched_debug.cfs_rq:/.MIN_vruntime.max
> >    7267340 ±  5%     -44.7%    4015798 ± 14%
> >    sched_debug.cfs_rq:/.MIN_vruntime.stddev
> >       4278 ±  7%     -54.7%       1939 ± 11%
> >       sched_debug.cfs_rq:/.exec_clock.stddev
> >     245.48 ±  2%     +65.3%     405.75 ±  7%
> >     sched_debug.cfs_rq:/.load_avg.avg
> >       2692 ±  6%    +126.0%       6087 ±  7%
> >       sched_debug.cfs_rq:/.load_avg.max
> >      33.09           -73.0%       8.94 ±  7%
> >      sched_debug.cfs_rq:/.load_avg.min
> >     507.40 ±  4%    +128.0%       1156 ±  7%
> >     sched_debug.cfs_rq:/.load_avg.stddev
> >    1875095 ±  7%     -54.8%     846645 ± 16%
> >    sched_debug.cfs_rq:/.max_vruntime.avg
> >   32868921 ±  6%     -35.7%   21150379 ± 14%
> >   sched_debug.cfs_rq:/.max_vruntime.max
> >    7267341 ±  5%     -44.7%    4015798 ± 14%
> >    sched_debug.cfs_rq:/.max_vruntime.stddev
> >   35887197           -13.2%   31149130
> >   sched_debug.cfs_rq:/.min_vruntime.avg
> >   37385506           -14.3%   32043914
> >   sched_debug.cfs_rq:/.min_vruntime.max
> >   34416296           -12.3%   30183927
> >   sched_debug.cfs_rq:/.min_vruntime.min
> >    1228844 ±  8%     -52.6%     582759 ±  4%
> >    sched_debug.cfs_rq:/.min_vruntime.stddev
> >       0.83           -28.1%       0.60 ±  6%
> >       sched_debug.cfs_rq:/.nr_running.avg
> >       2.07 ±  3%     -24.6%       1.56 ±  8%
> >       sched_debug.cfs_rq:/.nr_running.max
> >      20.52 ±  4%     -48.8%      10.52 ±  3%
> >      sched_debug.cfs_rq:/.nr_spread_over.avg
> >      35.96 ±  5%     -42.2%      20.77 ±  9%
> >      sched_debug.cfs_rq:/.nr_spread_over.max
> >       8.97 ± 11%     -44.5%       4.98 ±  8%
> >       sched_debug.cfs_rq:/.nr_spread_over.min
> >       6.40 ± 12%     -45.5%       3.49 ±  7%
> >       sched_debug.cfs_rq:/.nr_spread_over.stddev
> >      21.78 ±  7%    +143.3%      53.00 ±  9%
> >      sched_debug.cfs_rq:/.runnable_load_avg.avg
> >     328.86 ± 18%    +303.4%       1326 ± 14%
> >     sched_debug.cfs_rq:/.runnable_load_avg.max
> >      55.97 ± 17%    +286.0%     216.07 ± 13%
> >      sched_debug.cfs_rq:/.runnable_load_avg.stddev
> >       0.10 ± 29%     -82.4%       0.02 ± 50%
> >       sched_debug.cfs_rq:/.spread.avg
> >       3.43 ± 25%     -79.9%       0.69 ± 50%
> >       sched_debug.cfs_rq:/.spread.max
> >       0.56 ± 26%     -80.7%       0.11 ± 50%
> >       sched_debug.cfs_rq:/.spread.stddev
> >    1228822 ±  8%     -52.6%     582732 ±  4%
> >    sched_debug.cfs_rq:/.spread0.stddev
> >     992.30           -24.9%     745.56 ±  2%
> >     sched_debug.cfs_rq:/.util_avg.avg
> >       1485           -18.1%       1217 ±  2%
> >       sched_debug.cfs_rq:/.util_avg.max
> >     515.45 ±  2%     -25.2%     385.73 ±  6%
> >     sched_debug.cfs_rq:/.util_avg.min
> >     201.54           -14.9%     171.52 ±  3%
> >     sched_debug.cfs_rq:/.util_avg.stddev
> >     248.73 ±  6%     -38.1%     154.02 ±  8%
> >     sched_debug.cfs_rq:/.util_est_enqueued.avg
> >     222.78 ±  3%     -15.8%     187.58 ±  2%
> >     sched_debug.cfs_rq:/.util_est_enqueued.stddev
> >      77097 ±  4%    +278.4%     291767 ± 11%  sched_debug.cpu.avg_idle.avg
> >     181319 ±  6%    +298.7%     722862 ±  3%  sched_debug.cpu.avg_idle.max
> >      19338          +392.3%      95203 ± 17%  sched_debug.cpu.avg_idle.min
> >      34877 ±  6%    +303.5%     140732 ±  6%
> >      sched_debug.cpu.avg_idle.stddev
> >    1107408           +37.6%    1523823        sched_debug.cpu.clock.avg
> >    1107427           +37.6%    1523834        sched_debug.cpu.clock.max
> >    1107385           +37.6%    1523811        sched_debug.cpu.clock.min
> >      13.10 ±  9%     -48.1%       6.80 ±  8%  sched_debug.cpu.clock.stddev
> >    1107408           +37.6%    1523823
> >    sched_debug.cpu.clock_task.avg
> >    1107427           +37.6%    1523834
> >    sched_debug.cpu.clock_task.max
> >    1107385           +37.6%    1523811
> >    sched_debug.cpu.clock_task.min
> >      13.10 ±  9%     -48.1%       6.80 ±  8%
> >      sched_debug.cpu.clock_task.stddev
> >      30.36 ±  7%    +107.7%      63.06 ± 12%
> >      sched_debug.cpu.cpu_load[0].avg
> >     381.48 ± 18%    +269.8%       1410 ± 18%
> >     sched_debug.cpu.cpu_load[0].max
> >      63.92 ± 18%    +262.2%     231.50 ± 17%
> >      sched_debug.cpu.cpu_load[0].stddev
> >      31.34 ±  5%    +118.4%      68.44 ±  9%
> >      sched_debug.cpu.cpu_load[1].avg
> >     323.62 ± 17%    +349.5%       1454 ± 14%
> >     sched_debug.cpu.cpu_load[1].max
> >      53.23 ± 16%    +350.3%     239.71 ± 13%
> >      sched_debug.cpu.cpu_load[1].stddev
> >      32.15 ±  3%    +129.4%      73.74 ±  6%
> >      sched_debug.cpu.cpu_load[2].avg
> >     285.20 ± 14%    +420.8%       1485 ±  9%
> >     sched_debug.cpu.cpu_load[2].max
> >      46.66 ± 12%    +430.0%     247.32 ±  8%
> >      sched_debug.cpu.cpu_load[2].stddev
> >      33.02 ±  2%    +133.2%      77.00 ±  3%
> >      sched_debug.cpu.cpu_load[3].avg
> >     252.16 ± 10%    +481.2%       1465 ±  7%
> >     sched_debug.cpu.cpu_load[3].max
> >      40.74 ±  8%    +503.2%     245.72 ±  6%
> >      sched_debug.cpu.cpu_load[3].stddev
> >      33.86          +131.5%      78.38 ±  2%
> >      sched_debug.cpu.cpu_load[4].avg
> >     219.81 ±  8%    +522.6%       1368 ±  5%
> >     sched_debug.cpu.cpu_load[4].max
> >      35.45 ±  7%    +554.2%     231.90 ±  4%
> >      sched_debug.cpu.cpu_load[4].stddev
> >       2600 ±  4%     -30.5%       1807 ±  4%  sched_debug.cpu.curr->pid.avg
> >      25309 ±  4%     -19.5%      20367 ±  4%  sched_debug.cpu.curr->pid.max
> >       4534 ±  7%     -21.2%       3573 ±  5%
> >       sched_debug.cpu.curr->pid.stddev
> >       0.00 ±  2%     -27.6%       0.00 ±  6%
> >       sched_debug.cpu.next_balance.stddev
> >    1083917           +38.6%    1502777
> >    sched_debug.cpu.nr_load_updates.avg
> >    1088142           +38.6%    1508302
> >    sched_debug.cpu.nr_load_updates.max
> >    1082048           +38.7%    1501073
> >    sched_debug.cpu.nr_load_updates.min
> >       3.53 ±  6%     -73.0%       0.95 ±  6%
> >       sched_debug.cpu.nr_running.avg
> >      11.54 ±  3%     -62.1%       4.37 ± 10%
> >      sched_debug.cpu.nr_running.max
> >       3.10 ±  3%     -66.8%       1.03 ±  9%
> >       sched_debug.cpu.nr_running.stddev
> >   10764176           -22.4%    8355047
> >   sched_debug.cpu.nr_switches.avg
> >   10976436           -22.2%    8545010
> >   sched_debug.cpu.nr_switches.max
> >   10547712           -22.8%    8143037
> >   sched_debug.cpu.nr_switches.min
> >     148628 ±  3%     -22.7%     114880 ±  7%
> >     sched_debug.cpu.nr_switches.stddev
> >      11.13 ±  2%     +24.5%      13.85
> >      sched_debug.cpu.nr_uninterruptible.avg
> >       6420 ±  8%     -48.7%       3296 ± 11%
> >       sched_debug.cpu.nr_uninterruptible.max
> >      -5500           -37.2%      -3455
> >      sched_debug.cpu.nr_uninterruptible.min
> >       3784 ±  6%     -47.2%       1997 ±  4%
> >       sched_debug.cpu.nr_uninterruptible.stddev
> >   10812670           -22.7%    8356821
> >   sched_debug.cpu.sched_count.avg
> >   11020646           -22.5%    8546277
> >   sched_debug.cpu.sched_count.max
> >   10601390           -23.2%    8144743
> >   sched_debug.cpu.sched_count.min
> >     144529 ±  3%     -20.9%     114359 ±  7%
> >     sched_debug.cpu.sched_count.stddev
> >     706116          +259.0%    2534721
> >     sched_debug.cpu.sched_goidle.avg
> >     771307          +232.4%    2564059
> >     sched_debug.cpu.sched_goidle.max
> >     644658          +286.9%    2494236
> >     sched_debug.cpu.sched_goidle.min
> >      49847 ±  6%     -67.9%      15979 ±  7%
> >      sched_debug.cpu.sched_goidle.stddev
> >    9618827           -39.9%    5780369
> >    sched_debug.cpu.ttwu_count.avg
> >    8990451           -61.7%    3441265 ±  4%
> >    sched_debug.cpu.ttwu_count.min
> >     418563 ± 25%    +244.2%    1440565 ±  7%
> >     sched_debug.cpu.ttwu_count.stddev
> >     640964           -93.7%      40366 ±  2%
> >     sched_debug.cpu.ttwu_local.avg
> >     679527           -92.1%      53476 ±  4%
> >     sched_debug.cpu.ttwu_local.max
> >     601661           -94.9%      30636 ±  3%
> >     sched_debug.cpu.ttwu_local.min
> >      24242 ± 21%     -77.7%       5405 ±  9%
> >      sched_debug.cpu.ttwu_local.stddev
> >    1107383           +37.6%    1523810        sched_debug.cpu_clk
> >    1107383           +37.6%    1523810        sched_debug.ktime
> >       0.00           -49.4%       0.00 ± 65%
> >       sched_debug.rt_rq:/.rt_nr_migratory.avg
> >       0.03           -49.4%       0.01 ± 65%
> >       sched_debug.rt_rq:/.rt_nr_migratory.max
> >       0.00           -49.4%       0.00 ± 65%
> >       sched_debug.rt_rq:/.rt_nr_migratory.stddev
> >       0.00           -49.4%       0.00 ± 65%
> >       sched_debug.rt_rq:/.rt_nr_running.avg
> >       0.03           -49.4%       0.01 ± 65%
> >       sched_debug.rt_rq:/.rt_nr_running.max
> >       0.00           -49.4%       0.00 ± 65%
> >       sched_debug.rt_rq:/.rt_nr_running.stddev
> >       0.01 ±  8%     +79.9%       0.01 ± 23%
> >       sched_debug.rt_rq:/.rt_time.avg
> >    1107805           +37.6%    1524235        sched_debug.sched_clk
> >      87.59           -87.6        0.00
> >      perf-profile.calltrace.cycles-pp.md_flush_request.raid1_make_request.md_handle_request.md_make_request.generic_make_request
> >      87.57           -87.6        0.00
> >      perf-profile.calltrace.cycles-pp.submit_bio_wait.blkdev_issue_flush.xfs_file_fsync.xfs_file_write_iter.__vfs_write
> >      87.59           -87.5        0.05 ±299%
> >      perf-profile.calltrace.cycles-pp.blkdev_issue_flush.xfs_file_fsync.xfs_file_write_iter.__vfs_write.vfs_write
> >      87.51           -87.5        0.00
> >      perf-profile.calltrace.cycles-pp.generic_make_request.submit_bio.submit_bio_wait.blkdev_issue_flush.xfs_file_fsync
> >      87.51           -87.5        0.00
> >      perf-profile.calltrace.cycles-pp.submit_bio.submit_bio_wait.blkdev_issue_flush.xfs_file_fsync.xfs_file_write_iter
> >      87.50           -87.5        0.00
> >      perf-profile.calltrace.cycles-pp.md_make_request.generic_make_request.submit_bio.submit_bio_wait.blkdev_issue_flush
> >      87.50           -87.5        0.00
> >      perf-profile.calltrace.cycles-pp.md_handle_request.md_make_request.generic_make_request.submit_bio.submit_bio_wait
> >      82.37           -82.4        0.00
> >      perf-profile.calltrace.cycles-pp._raw_spin_lock_irq.md_flush_request.raid1_make_request.md_handle_request.md_make_request
> >      82.23           -82.2        0.00
> >      perf-profile.calltrace.cycles-pp.native_queued_spin_lock_slowpath._raw_spin_lock_irq.md_flush_request.raid1_make_request.md_handle_request
> >      87.79           -25.0       62.75 ±  8%
> >      perf-profile.calltrace.cycles-pp.raid1_make_request.md_handle_request.md_make_request.generic_make_request.submit_bio
> >      92.78           -13.0       79.76
> >      perf-profile.calltrace.cycles-pp.xfs_file_fsync.xfs_file_write_iter.__vfs_write.vfs_write.ksys_write
> >      93.08           -12.6       80.49
> >      perf-profile.calltrace.cycles-pp.xfs_file_write_iter.__vfs_write.vfs_write.ksys_write.do_syscall_64
> >      93.08           -12.6       80.50
> >      perf-profile.calltrace.cycles-pp.__vfs_write.vfs_write.ksys_write.do_syscall_64.entry_SYSCALL_64_after_hwframe
> >      93.11           -12.6       80.56
> >      perf-profile.calltrace.cycles-pp.vfs_write.ksys_write.do_syscall_64.entry_SYSCALL_64_after_hwframe
> >      93.11           -12.6       80.56
> >      perf-profile.calltrace.cycles-pp.ksys_write.do_syscall_64.entry_SYSCALL_64_after_hwframe
> >      93.14           -12.5       80.64
> >      perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe
> >      93.15           -12.5       80.65
> >      perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe
> >       3.40 ±  2%      -1.4        1.97 ±  8%
> >       perf-profile.calltrace.cycles-pp.worker_thread.kthread.ret_from_fork
> >       3.33 ±  2%      -1.4        1.96 ±  9%
> >       perf-profile.calltrace.cycles-pp.process_one_work.worker_thread.kthread.ret_from_fork
> >       1.12 ±  2%      -0.7        0.42 ± 68%
> >       perf-profile.calltrace.cycles-pp.__save_stack_trace.save_stack_trace_tsk.__account_scheduler_latency.enqueue_entity.enqueue_task_fair
> >       1.16 ±  2%      -0.6        0.60 ± 17%
> >       perf-profile.calltrace.cycles-pp.save_stack_trace_tsk.__account_scheduler_latency.enqueue_entity.enqueue_task_fair.ttwu_do_activate
> >       0.00            +0.6        0.59 ± 15%
> >       perf-profile.calltrace.cycles-pp._raw_spin_lock_irqsave.__wake_up_common_lock.raid1_write_request.raid1_make_request.md_handle_request
> >       0.00            +0.6        0.64 ± 15%
> >       perf-profile.calltrace.cycles-pp.__wake_up_common_lock.raid1_write_request.raid1_make_request.md_handle_request.md_make_request
> >       0.00            +0.7        0.65 ± 10%
> >       perf-profile.calltrace.cycles-pp.enqueue_entity.enqueue_task_fair.ttwu_do_activate.sched_ttwu_pending.do_idle
> >       0.00            +0.7        0.68 ± 10%
> >       perf-profile.calltrace.cycles-pp.enqueue_task_fair.ttwu_do_activate.sched_ttwu_pending.do_idle.cpu_startup_entry
> >       0.00            +0.7        0.69 ± 10%
> >       perf-profile.calltrace.cycles-pp.ttwu_do_activate.sched_ttwu_pending.do_idle.cpu_startup_entry.start_secondary
> >       0.00            +0.8        0.79 ± 11%
> >       perf-profile.calltrace.cycles-pp.sched_ttwu_pending.do_idle.cpu_startup_entry.start_secondary.secondary_startup_64
> >       0.00            +0.8        0.83 ±  7%
> >       perf-profile.calltrace.cycles-pp.__schedule.schedule.raid1_write_request.raid1_make_request.md_handle_request
> >       0.62 ±  3%      +0.8        1.45 ± 22%
> >       perf-profile.calltrace.cycles-pp.native_queued_spin_lock_slowpath._raw_spin_lock_irqsave.remove_wait_queue.xlog_wait.__xfs_log_force_lsn
> >       0.00            +0.8        0.83 ±  7%
> >       perf-profile.calltrace.cycles-pp.schedule.raid1_write_request.raid1_make_request.md_handle_request.md_make_request
> >       0.63 ±  2%      +0.8        1.46 ± 22%
> >       perf-profile.calltrace.cycles-pp.remove_wait_queue.xlog_wait.__xfs_log_force_lsn.xfs_log_force_lsn.xfs_file_fsync
> >       0.62 ±  2%      +0.8        1.46 ± 22%
> >       perf-profile.calltrace.cycles-pp._raw_spin_lock_irqsave.remove_wait_queue.xlog_wait.__xfs_log_force_lsn.xfs_log_force_lsn
> >       3.92 ±  2%      +0.9        4.79 ±  6%
> >       perf-profile.calltrace.cycles-pp.ret_from_fork
> >       3.92 ±  2%      +0.9        4.79 ±  6%
> >       perf-profile.calltrace.cycles-pp.kthread.ret_from_fork
> >       0.69 ±  2%      +0.9        1.64 ± 23%
> >       perf-profile.calltrace.cycles-pp.xlog_wait.__xfs_log_force_lsn.xfs_log_force_lsn.xfs_file_fsync.xfs_file_write_iter
> >       0.00            +1.2        1.17 ±  8%
> >       perf-profile.calltrace.cycles-pp._raw_spin_unlock_irqrestore.prepare_to_wait_event.raid1_write_request.raid1_make_request.md_handle_request
> >       0.00            +1.2        1.23 ± 18%
> >       perf-profile.calltrace.cycles-pp.prepare_to_wait_event.raid1_write_request.raid1_make_request.md_handle_request.submit_flushes
> >       0.00            +1.3        1.27 ± 17%
> >       perf-profile.calltrace.cycles-pp.raid1_write_request.raid1_make_request.md_handle_request.submit_flushes.process_one_work
> >       0.00            +1.3        1.27 ± 17%
> >       perf-profile.calltrace.cycles-pp.md_handle_request.submit_flushes.process_one_work.worker_thread.kthread
> >       0.00            +1.3        1.27 ± 17%
> >       perf-profile.calltrace.cycles-pp.raid1_make_request.md_handle_request.submit_flushes.process_one_work.worker_thread
> >       0.00            +1.3        1.27 ± 17%
> >       perf-profile.calltrace.cycles-pp.submit_flushes.process_one_work.worker_thread.kthread.ret_from_fork
> >       0.00            +1.6        1.65 ± 14%
> >       perf-profile.calltrace.cycles-pp.try_to_wake_up.autoremove_wake_function.__wake_up_common.__wake_up_common_lock.raid_end_bio_io
> >       0.00            +1.7        1.71 ± 14%
> >       perf-profile.calltrace.cycles-pp.autoremove_wake_function.__wake_up_common.__wake_up_common_lock.raid_end_bio_io.raid1_end_write_request
> >       0.00            +1.7        1.71 ± 14%
> >       perf-profile.calltrace.cycles-pp.__wake_up_common.__wake_up_common_lock.raid_end_bio_io.raid1_end_write_request.brd_make_request
> >       0.00            +1.9        1.86 ± 13%
> >       perf-profile.calltrace.cycles-pp.__wake_up_common_lock.raid_end_bio_io.raid1_end_write_request.brd_make_request.generic_make_request
> >       0.00            +2.1        2.10 ± 10%
> >       perf-profile.calltrace.cycles-pp.native_queued_spin_lock_slowpath._raw_spin_lock_irqsave.remove_wait_queue.__xfs_log_force_lsn.xfs_log_force_lsn
> >       0.00            +2.1        2.10 ± 10%
> >       perf-profile.calltrace.cycles-pp._raw_spin_lock_irqsave.remove_wait_queue.__xfs_log_force_lsn.xfs_log_force_lsn.xfs_file_fsync
> >       0.00            +2.1        2.11 ± 10%
> >       perf-profile.calltrace.cycles-pp.remove_wait_queue.__xfs_log_force_lsn.xfs_log_force_lsn.xfs_file_fsync.xfs_file_write_iter
> >       0.00            +2.2        2.16 ± 10%
> >       perf-profile.calltrace.cycles-pp.raid_end_bio_io.raid1_end_write_request.brd_make_request.generic_make_request.flush_bio_list
> >       2.24 ±  4%      +2.2        4.44 ± 15%
> >       perf-profile.calltrace.cycles-pp.xfs_log_force_lsn.xfs_file_fsync.xfs_file_write_iter.__vfs_write.vfs_write
> >       0.00            +2.3        2.25 ± 10%
> >       perf-profile.calltrace.cycles-pp.raid1_end_write_request.brd_make_request.generic_make_request.flush_bio_list.flush_pending_writes
> >       0.00            +2.3        2.30 ± 20%
> >       perf-profile.calltrace.cycles-pp.native_queued_spin_lock_slowpath._raw_spin_lock_irqsave.raid1_write_request.raid1_make_request.md_handle_request
> >       0.00            +2.4        2.35 ± 20%
> >       perf-profile.calltrace.cycles-pp._raw_spin_lock_irqsave.raid1_write_request.raid1_make_request.md_handle_request.md_make_request
> >       0.37 ± 65%      +2.4        2.81 ±  7%
> >       perf-profile.calltrace.cycles-pp.md_thread.kthread.ret_from_fork
> >       0.26 ±100%      +2.5        2.81 ±  7%
> >       perf-profile.calltrace.cycles-pp.raid1d.md_thread.kthread.ret_from_fork
> >       0.26 ±100%      +2.5        2.81 ±  7%
> >       perf-profile.calltrace.cycles-pp.flush_pending_writes.raid1d.md_thread.kthread.ret_from_fork
> >       0.26 ±100%      +2.6        2.81 ±  7%
> >       perf-profile.calltrace.cycles-pp.flush_bio_list.flush_pending_writes.raid1d.md_thread.kthread
> >       0.10 ±200%      +2.7        2.76 ±  7%
> >       perf-profile.calltrace.cycles-pp.generic_make_request.flush_bio_list.flush_pending_writes.raid1d.md_thread
> >       0.00            +2.7        2.73 ±  7%
> >       perf-profile.calltrace.cycles-pp.brd_make_request.generic_make_request.flush_bio_list.flush_pending_writes.raid1d
> >       1.20 ±  3%      +3.1        4.35 ± 15%
> >       perf-profile.calltrace.cycles-pp.__xfs_log_force_lsn.xfs_log_force_lsn.xfs_file_fsync.xfs_file_write_iter.__vfs_write
> >       0.63 ±  6%      +3.8        4.38 ± 27%
> >       perf-profile.calltrace.cycles-pp.native_queued_spin_lock_slowpath._raw_spin_lock_irqsave.remove_wait_queue.__xfs_log_force_lsn.xfs_file_fsync
> >       0.63 ±  5%      +3.8        4.39 ± 27%
> >       perf-profile.calltrace.cycles-pp._raw_spin_lock_irqsave.remove_wait_queue.__xfs_log_force_lsn.xfs_file_fsync.xfs_file_write_iter
> >       0.63 ±  5%      +3.8        4.40 ± 27%
> >       perf-profile.calltrace.cycles-pp.remove_wait_queue.__xfs_log_force_lsn.xfs_file_fsync.xfs_file_write_iter.__vfs_write
> >       1.26 ±  5%      +5.3        6.55 ± 27%
> >       perf-profile.calltrace.cycles-pp.native_queued_spin_lock_slowpath._raw_spin_lock.__xfs_log_force_lsn.xfs_file_fsync.xfs_file_write_iter
> >       1.27 ±  5%      +5.3        6.55 ± 27%
> >       perf-profile.calltrace.cycles-pp._raw_spin_lock.__xfs_log_force_lsn.xfs_file_fsync.xfs_file_write_iter.__vfs_write
> >       1.30 ±  4%      +8.4        9.72 ±  9%
> >       perf-profile.calltrace.cycles-pp.intel_idle.cpuidle_enter_state.do_idle.cpu_startup_entry.start_secondary
> >       1.33 ±  4%      +8.9       10.26 ±  9%
> >       perf-profile.calltrace.cycles-pp.cpuidle_enter_state.do_idle.cpu_startup_entry.start_secondary.secondary_startup_64
> >       2.28 ±  2%      +9.1       11.36 ± 27%
> >       perf-profile.calltrace.cycles-pp.__xfs_log_force_lsn.xfs_file_fsync.xfs_file_write_iter.__vfs_write.vfs_write
> >       1.59 ±  4%     +10.4       11.97 ±  9%
> >       perf-profile.calltrace.cycles-pp.do_idle.cpu_startup_entry.start_secondary.secondary_startup_64
> >       1.59 ±  4%     +10.4       11.98 ±  9%
> >       perf-profile.calltrace.cycles-pp.cpu_startup_entry.start_secondary.secondary_startup_64
> >       1.59 ±  4%     +10.4       11.98 ±  9%
> >       perf-profile.calltrace.cycles-pp.start_secondary.secondary_startup_64
> >       1.63 ±  4%     +10.8       12.47 ±  8%
> >       perf-profile.calltrace.cycles-pp.secondary_startup_64
> >       0.00           +57.7       57.66 ± 10%
> >       perf-profile.calltrace.cycles-pp.native_queued_spin_lock_slowpath._raw_spin_lock_irqsave.prepare_to_wait_event.raid1_write_request.raid1_make_request
> >       0.00           +57.7       57.73 ± 10%
> >       perf-profile.calltrace.cycles-pp._raw_spin_lock_irqsave.prepare_to_wait_event.raid1_write_request.raid1_make_request.md_handle_request
> >       0.05 ±299%     +57.8       57.85 ±  9%
> >       perf-profile.calltrace.cycles-pp.prepare_to_wait_event.raid1_write_request.raid1_make_request.md_handle_request.md_make_request
> >       0.19 ±154%     +62.5       62.73 ±  8%
> >       perf-profile.calltrace.cycles-pp.raid1_write_request.raid1_make_request.md_handle_request.md_make_request.generic_make_request
> >       0.19 ±154%     +62.6       62.76 ±  8%
> >       perf-profile.calltrace.cycles-pp.md_handle_request.md_make_request.generic_make_request.submit_bio.xfs_submit_ioend
> >       0.19 ±154%     +62.6       62.79 ±  8%
> >       perf-profile.calltrace.cycles-pp.md_make_request.generic_make_request.submit_bio.xfs_submit_ioend.xfs_vm_writepages
> >       0.20 ±154%     +62.6       62.81 ±  8%
> >       perf-profile.calltrace.cycles-pp.generic_make_request.submit_bio.xfs_submit_ioend.xfs_vm_writepages.do_writepages
> >       0.20 ±154%     +62.6       62.81 ±  8%
> >       perf-profile.calltrace.cycles-pp.submit_bio.xfs_submit_ioend.xfs_vm_writepages.do_writepages.__filemap_fdatawrite_range
> >       0.20 ±154%     +62.6       62.82 ±  8%
> >       perf-profile.calltrace.cycles-pp.xfs_submit_ioend.xfs_vm_writepages.do_writepages.__filemap_fdatawrite_range.file_write_and_wait_range
> >       0.29 ±125%     +62.8       63.09 ±  8%
> >       perf-profile.calltrace.cycles-pp.xfs_vm_writepages.do_writepages.__filemap_fdatawrite_range.file_write_and_wait_range.xfs_file_fsync
> >       0.29 ±126%     +62.8       63.10 ±  8%
> >       perf-profile.calltrace.cycles-pp.do_writepages.__filemap_fdatawrite_range.file_write_and_wait_range.xfs_file_fsync.xfs_file_write_iter
> >       0.29 ±125%     +62.8       63.11 ±  8%
> >       perf-profile.calltrace.cycles-pp.__filemap_fdatawrite_range.file_write_and_wait_range.xfs_file_fsync.xfs_file_write_iter.__vfs_write
> >       0.62 ± 41%     +62.9       63.52 ±  7%
> >       perf-profile.calltrace.cycles-pp.file_write_and_wait_range.xfs_file_fsync.xfs_file_write_iter.__vfs_write.vfs_write
> >      88.51           -88.2        0.26 ± 19%
> >      perf-profile.children.cycles-pp.md_flush_request
> >      87.57           -87.2        0.35 ± 19%
> >      perf-profile.children.cycles-pp.submit_bio_wait
> >      87.59           -87.2        0.39 ± 19%
> >      perf-profile.children.cycles-pp.blkdev_issue_flush
> >      83.26           -83.2        0.02 ±123%
> >      perf-profile.children.cycles-pp._raw_spin_lock_irq
> >      88.85           -25.7       63.11 ±  8%
> >      perf-profile.children.cycles-pp.md_make_request
> >      88.90           -25.7       63.17 ±  8%
> >      perf-profile.children.cycles-pp.submit_bio
> >      88.83           -24.5       64.31 ±  8%
> >      perf-profile.children.cycles-pp.raid1_make_request
> >      88.84           -24.5       64.33 ±  8%
> >      perf-profile.children.cycles-pp.md_handle_request
> >      89.38           -23.5       65.92 ±  7%
> >      perf-profile.children.cycles-pp.generic_make_request
> >      89.90           -13.4       76.51 ±  2%
> >      perf-profile.children.cycles-pp.native_queued_spin_lock_slowpath
> >      92.79           -13.0       79.76
> >      perf-profile.children.cycles-pp.xfs_file_fsync
> >      93.08           -12.6       80.49
> >      perf-profile.children.cycles-pp.xfs_file_write_iter
> >      93.09           -12.6       80.54
> >      perf-profile.children.cycles-pp.__vfs_write
> >      93.13           -12.5       80.60
> >      perf-profile.children.cycles-pp.vfs_write
> >      93.13           -12.5       80.61
> >      perf-profile.children.cycles-pp.ksys_write
> >      93.22           -12.4       80.83
> >      perf-profile.children.cycles-pp.do_syscall_64
> >      93.22           -12.4       80.83
> >      perf-profile.children.cycles-pp.entry_SYSCALL_64_after_hwframe
> >       3.40 ±  2%      -1.4        1.97 ±  8%
> >       perf-profile.children.cycles-pp.worker_thread
> >       3.33 ±  2%      -1.4        1.96 ±  9%
> >       perf-profile.children.cycles-pp.process_one_work
> >       1.03 ±  7%      -1.0        0.07 ± 37%
> >       perf-profile.children.cycles-pp.xlog_cil_force_lsn
> >       1.69 ±  2%      -0.7        0.96 ±  4%
> >       perf-profile.children.cycles-pp.reschedule_interrupt
> >       1.66 ±  2%      -0.7        0.94 ±  4%
> >       perf-profile.children.cycles-pp.scheduler_ipi
> >       1.13 ±  2%      -0.7        0.47 ± 11%
> >       perf-profile.children.cycles-pp.finish_wait
> >       0.54 ±  8%      -0.4        0.10 ± 38%
> >       perf-profile.children.cycles-pp.xlog_cil_push
> >       0.49 ±  9%      -0.4        0.09 ± 35%
> >       perf-profile.children.cycles-pp.xlog_write
> >       0.10 ±  8%      -0.1        0.04 ± 67%
> >       perf-profile.children.cycles-pp.flush_work
> >       0.20 ±  5%      -0.0        0.16 ± 11%
> >       perf-profile.children.cycles-pp.reweight_entity
> >       0.06 ± 10%      +0.0        0.10 ± 23%
> >       perf-profile.children.cycles-pp.brd_lookup_page
> >       0.18 ±  5%      +0.0        0.23 ± 13%
> >       perf-profile.children.cycles-pp.__update_load_avg_se
> >       0.02 ±153%      +0.1        0.07 ± 16%
> >       perf-profile.children.cycles-pp.delay_tsc
> >       0.03 ±100%      +0.1        0.08 ± 15%
> >       perf-profile.children.cycles-pp.find_next_bit
> >       0.08 ±  5%      +0.1        0.14 ± 14%
> >       perf-profile.children.cycles-pp.native_write_msr
> >       0.29 ±  4%      +0.1        0.36 ±  8%
> >       perf-profile.children.cycles-pp.__orc_find
> >       0.40 ±  4%      +0.1        0.46 ±  7%
> >       perf-profile.children.cycles-pp.dequeue_task_fair
> >       0.11 ± 11%      +0.1        0.18 ± 14%
> >       perf-profile.children.cycles-pp.__module_text_address
> >       0.12 ±  8%      +0.1        0.19 ± 13%
> >       perf-profile.children.cycles-pp.is_module_text_address
> >       0.04 ± 50%      +0.1        0.12 ± 19%
> >       perf-profile.children.cycles-pp.kmem_cache_alloc
> >       0.00            +0.1        0.08 ± 11%
> >       perf-profile.children.cycles-pp.clear_page_erms
> >       0.00            +0.1        0.08 ± 28%
> >       perf-profile.children.cycles-pp.__indirect_thunk_start
> >       0.01 ±200%      +0.1        0.10 ± 25%
> >       perf-profile.children.cycles-pp.xfs_trans_alloc
> >       0.00            +0.1        0.09 ± 18%
> >       perf-profile.children.cycles-pp.md_wakeup_thread
> >       0.00            +0.1        0.09 ± 26%
> >       perf-profile.children.cycles-pp.rebalance_domains
> >       0.00            +0.1        0.09 ± 26%
> >       perf-profile.children.cycles-pp.get_next_timer_interrupt
> >       0.00            +0.1        0.09 ± 20%
> >       perf-profile.children.cycles-pp.ktime_get
> >       0.18 ±  4%      +0.1        0.27 ± 12%
> >       perf-profile.children.cycles-pp.idle_cpu
> >       0.20 ±  6%      +0.1        0.30 ±  9%
> >       perf-profile.children.cycles-pp.unwind_get_return_address
> >       0.16 ± 10%      +0.1        0.25 ± 13%
> >       perf-profile.children.cycles-pp.__module_address
> >       0.03 ±100%      +0.1        0.13 ±  8%
> >       perf-profile.children.cycles-pp.brd_insert_page
> >       0.06 ±  9%      +0.1        0.16 ± 14%
> >       perf-profile.children.cycles-pp.task_tick_fair
> >       0.08 ± 12%      +0.1        0.18 ± 24%
> >       perf-profile.children.cycles-pp.bio_alloc_bioset
> >       0.03 ± 81%      +0.1        0.14 ± 27%
> >       perf-profile.children.cycles-pp.generic_make_request_checks
> >       0.17 ±  7%      +0.1        0.28 ± 11%
> >       perf-profile.children.cycles-pp.__kernel_text_address
> >       0.11 ±  9%      +0.1        0.22 ± 15%
> >       perf-profile.children.cycles-pp.wake_up_page_bit
> >       0.16 ±  6%      +0.1        0.27 ± 10%
> >       perf-profile.children.cycles-pp.kernel_text_address
> >       0.00            +0.1        0.11 ± 11%
> >       perf-profile.children.cycles-pp.get_page_from_freelist
> >       0.00            +0.1        0.11 ± 19%
> >       perf-profile.children.cycles-pp.perf_mux_hrtimer_handler
> >       0.00            +0.1        0.11 ±  7%
> >       perf-profile.children.cycles-pp.__alloc_pages_nodemask
> >       0.08 ± 10%      +0.1        0.19 ± 22%
> >       perf-profile.children.cycles-pp.xfs_do_writepage
> >       0.25 ±  4%      +0.1        0.37 ± 10%
> >       perf-profile.children.cycles-pp.switch_mm_irqs_off
> >       0.00            +0.1        0.12 ± 13%
> >       perf-profile.children.cycles-pp.switch_mm
> >       0.08 ± 38%      +0.1        0.20 ± 19%
> >       perf-profile.children.cycles-pp.io_serial_in
> >       0.18 ±  5%      +0.1        0.31 ±  7%
> >       perf-profile.children.cycles-pp.dequeue_entity
> >       0.00            +0.1        0.13 ± 26%
> >       perf-profile.children.cycles-pp.tick_nohz_next_event
> >       0.06 ± 11%      +0.1        0.19 ± 19%
> >       perf-profile.children.cycles-pp.mempool_alloc
> >       0.32 ±  5%      +0.1        0.45 ±  6%
> >       perf-profile.children.cycles-pp.orc_find
> >       0.15 ± 10%      +0.1        0.29 ± 19%
> >       perf-profile.children.cycles-pp.xfs_destroy_ioend
> >       0.15 ± 11%      +0.1        0.30 ± 18%
> >       perf-profile.children.cycles-pp.call_bio_endio
> >       0.08 ± 17%      +0.2        0.23 ± 25%
> >       perf-profile.children.cycles-pp.xlog_state_done_syncing
> >       0.00            +0.2        0.15 ± 22%
> >       perf-profile.children.cycles-pp.tick_nohz_get_sleep_length
> >       0.12 ±  8%      +0.2        0.27 ± 23%
> >       perf-profile.children.cycles-pp.write_cache_pages
> >       0.10 ± 16%      +0.2        0.26 ± 16%
> >       perf-profile.children.cycles-pp.wait_for_xmitr
> >       0.10 ± 19%      +0.2        0.25 ± 14%
> >       perf-profile.children.cycles-pp.serial8250_console_putchar
> >       0.10 ± 17%      +0.2        0.26 ± 13%
> >       perf-profile.children.cycles-pp.uart_console_write
> >       0.10 ± 16%      +0.2        0.26 ± 15%
> >       perf-profile.children.cycles-pp.serial8250_console_write
> >       0.11 ± 15%      +0.2        0.27 ± 15%
> >       perf-profile.children.cycles-pp.console_unlock
> >       0.09 ±  9%      +0.2        0.26 ± 12%
> >       perf-profile.children.cycles-pp.scheduler_tick
> >       0.10 ± 18%      +0.2        0.28 ± 15%
> >       perf-profile.children.cycles-pp.irq_work_run_list
> >       0.10 ± 15%      +0.2        0.28 ± 14%
> >       perf-profile.children.cycles-pp.xlog_state_do_callback
> >       0.09 ± 12%      +0.2        0.27 ± 16%
> >       perf-profile.children.cycles-pp.irq_work_run
> >       0.09 ± 12%      +0.2        0.27 ± 16%
> >       perf-profile.children.cycles-pp.printk
> >       0.09 ± 12%      +0.2        0.27 ± 16%
> >       perf-profile.children.cycles-pp.vprintk_emit
> >       0.09 ± 12%      +0.2        0.27 ± 17%
> >       perf-profile.children.cycles-pp.irq_work_interrupt
> >       0.09 ± 12%      +0.2        0.27 ± 17%
> >       perf-profile.children.cycles-pp.smp_irq_work_interrupt
> >       0.00            +0.2        0.18 ± 16%
> >       perf-profile.children.cycles-pp.poll_idle
> >       0.30 ±  4%      +0.2        0.49 ± 11%
> >       perf-profile.children.cycles-pp.update_load_avg
> >       1.39 ±  2%      +0.2        1.59 ±  6%
> >       perf-profile.children.cycles-pp.__save_stack_trace
> >       1.43            +0.2        1.65 ±  6%
> >       perf-profile.children.cycles-pp.save_stack_trace_tsk
> >       0.14 ± 13%      +0.2        0.36 ± 13%
> >       perf-profile.children.cycles-pp.update_process_times
> >       0.00            +0.2        0.23 ± 22%
> >       perf-profile.children.cycles-pp.find_busiest_group
> >       0.22 ±  6%      +0.2        0.45 ± 18%
> >       perf-profile.children.cycles-pp.brd_do_bvec
> >       0.14 ± 13%      +0.2        0.38 ± 14%
> >       perf-profile.children.cycles-pp.tick_sched_handle
> >       0.10 ±  8%      +0.2        0.34 ± 26%
> >       perf-profile.children.cycles-pp.xfs_log_commit_cil
> >       0.07 ± 10%      +0.3        0.33 ± 23%
> >       perf-profile.children.cycles-pp.io_schedule
> >       0.03 ± 83%      +0.3        0.29 ± 27%
> >       perf-profile.children.cycles-pp.__softirqentry_text_start
> >       0.11 ±  5%      +0.3        0.36 ± 25%
> >       perf-profile.children.cycles-pp.__xfs_trans_commit
> >       0.06 ± 36%      +0.3        0.31 ± 26%
> >       perf-profile.children.cycles-pp.irq_exit
> >       0.08 ±  9%      +0.3        0.35 ± 23%
> >       perf-profile.children.cycles-pp.wait_on_page_bit_common
> >       0.15 ± 12%      +0.3        0.42 ± 14%
> >       perf-profile.children.cycles-pp.tick_sched_timer
> >       0.10 ± 11%      +0.3        0.39 ± 22%
> >       perf-profile.children.cycles-pp.__filemap_fdatawait_range
> >       0.06 ± 12%      +0.3        0.37 ±  9%
> >       perf-profile.children.cycles-pp.schedule_idle
> >       0.02 ±153%      +0.3        0.34 ± 17%
> >       perf-profile.children.cycles-pp.menu_select
> >       0.17 ±  5%      +0.3        0.49 ± 22%
> >       perf-profile.children.cycles-pp.xfs_vn_update_time
> >       0.19 ± 12%      +0.3        0.51 ± 18%
> >       perf-profile.children.cycles-pp.xlog_iodone
> >       0.18 ±  5%      +0.3        0.51 ± 22%
> >       perf-profile.children.cycles-pp.file_update_time
> >       0.18 ±  5%      +0.3        0.51 ± 21%
> >       perf-profile.children.cycles-pp.xfs_file_aio_write_checks
> >       0.21 ± 11%      +0.4        0.60 ± 15%
> >       perf-profile.children.cycles-pp.__hrtimer_run_queues
> >       0.26 ±  6%      +0.4        0.69 ± 16%
> >       perf-profile.children.cycles-pp.pick_next_task_fair
> >       1.20 ±  2%      +0.4        1.64 ± 10%
> >       perf-profile.children.cycles-pp.schedule
> >       0.28 ±  5%      +0.4        0.72 ± 21%
> >       perf-profile.children.cycles-pp.xfs_file_buffered_aio_write
> >       0.00            +0.4        0.44 ± 22%
> >       perf-profile.children.cycles-pp.load_balance
> >       0.25 ±  8%      +0.5        0.74 ± 15%
> >       perf-profile.children.cycles-pp.hrtimer_interrupt
> >       1.30 ±  2%      +0.7        2.00 ±  9%
> >       perf-profile.children.cycles-pp.__schedule
> >       0.31 ±  8%      +0.8        1.09 ± 16%
> >       perf-profile.children.cycles-pp.smp_apic_timer_interrupt
> >       0.31 ±  8%      +0.8        1.09 ± 16%
> >       perf-profile.children.cycles-pp.apic_timer_interrupt
> >       3.92 ±  2%      +0.9        4.79 ±  6%
> >       perf-profile.children.cycles-pp.ret_from_fork
> >       3.92 ±  2%      +0.9        4.79 ±  6%
> >       perf-profile.children.cycles-pp.kthread
> >       0.69 ±  2%      +0.9        1.64 ± 23%
> >       perf-profile.children.cycles-pp.xlog_wait
> >       0.08 ± 13%      +1.2        1.27 ± 17%
> >       perf-profile.children.cycles-pp.submit_flushes
> >       0.16 ±  9%      +1.6        1.74 ±  4%
> >       perf-profile.children.cycles-pp._raw_spin_unlock_irqrestore
> >       0.17 ±  9%      +2.0        2.16 ± 10%
> >       perf-profile.children.cycles-pp.raid_end_bio_io
> >       0.21 ±  6%      +2.0        2.25 ± 10%
> >       perf-profile.children.cycles-pp.raid1_end_write_request
> >       2.24 ±  4%      +2.2        4.44 ± 15%
> >       perf-profile.children.cycles-pp.xfs_log_force_lsn
> >       0.46 ±  6%      +2.3        2.73 ±  7%
> >       perf-profile.children.cycles-pp.brd_make_request
> >       0.51 ±  6%      +2.3        2.81 ±  7%
> >       perf-profile.children.cycles-pp.md_thread
> >       0.49 ±  6%      +2.3        2.81 ±  7%
> >       perf-profile.children.cycles-pp.raid1d
> >       0.49 ±  6%      +2.3        2.81 ±  7%
> >       perf-profile.children.cycles-pp.flush_pending_writes
> >       0.49 ±  6%      +2.3        2.81 ±  7%
> >       perf-profile.children.cycles-pp.flush_bio_list
> >       1.80 ±  3%      +5.6        7.44 ± 27%
> >       perf-profile.children.cycles-pp._raw_spin_lock
> >       2.12 ±  4%      +5.8        7.97 ± 20%
> >       perf-profile.children.cycles-pp.remove_wait_queue
> >       1.33 ±  4%      +8.8       10.12 ±  8%
> >       perf-profile.children.cycles-pp.intel_idle
> >       1.37 ±  4%      +9.3       10.71 ±  8%
> >       perf-profile.children.cycles-pp.cpuidle_enter_state
> >       1.59 ±  4%     +10.4       11.98 ±  9%
> >       perf-profile.children.cycles-pp.start_secondary
> >       1.63 ±  4%     +10.8       12.47 ±  8%
> >       perf-profile.children.cycles-pp.secondary_startup_64
> >       1.63 ±  4%     +10.8       12.47 ±  8%
> >       perf-profile.children.cycles-pp.cpu_startup_entry
> >       1.63 ±  4%     +10.9       12.49 ±  8%
> >       perf-profile.children.cycles-pp.do_idle
> >       3.48           +12.2       15.72 ± 23%
> >       perf-profile.children.cycles-pp.__xfs_log_force_lsn
> >       1.36 ± 12%     +57.8       59.12 ± 10%
> >       perf-profile.children.cycles-pp.prepare_to_wait_event
> >       0.43 ± 38%     +62.4       62.82 ±  8%
> >       perf-profile.children.cycles-pp.xfs_submit_ioend
> >       0.55 ± 29%     +62.5       63.10 ±  8%
> >       perf-profile.children.cycles-pp.xfs_vm_writepages
> >       0.55 ± 30%     +62.5       63.10 ±  8%
> >       perf-profile.children.cycles-pp.do_writepages
> >       0.55 ± 29%     +62.6       63.11 ±  8%
> >       perf-profile.children.cycles-pp.__filemap_fdatawrite_range
> >       0.66 ± 25%     +62.9       63.52 ±  7%
> >       perf-profile.children.cycles-pp.file_write_and_wait_range
> >       0.39 ± 43%     +63.6       64.02 ±  8%
> >       perf-profile.children.cycles-pp.raid1_write_request
> >       5.43 ±  3%     +64.2       69.64 ±  5%
> >       perf-profile.children.cycles-pp._raw_spin_lock_irqsave
> >      89.86           -13.5       76.31 ±  2%
> >      perf-profile.self.cycles-pp.native_queued_spin_lock_slowpath
> >       0.14 ±  8%      -0.0        0.09 ± 19%
> >       perf-profile.self.cycles-pp.md_flush_request
> >       0.10 ± 12%      -0.0        0.07 ± 21%
> >       perf-profile.self.cycles-pp.account_entity_enqueue
> >       0.06 ±  7%      +0.0        0.08 ± 12%
> >       perf-profile.self.cycles-pp.pick_next_task_fair
> >       0.05 ± 12%      +0.0        0.08 ± 18%
> >       perf-profile.self.cycles-pp.___perf_sw_event
> >       0.15 ±  6%      +0.0        0.18 ±  9%
> >       perf-profile.self.cycles-pp.__update_load_avg_se
> >       0.17 ±  4%      +0.0        0.22 ± 10%
> >       perf-profile.self.cycles-pp.__schedule
> >       0.10 ± 11%      +0.1        0.15 ± 11%
> >       perf-profile.self.cycles-pp._raw_spin_lock
> >       0.02 ±153%      +0.1        0.07 ± 16%
> >       perf-profile.self.cycles-pp.delay_tsc
> >       0.02 ±152%      +0.1        0.07 ± 23%
> >       perf-profile.self.cycles-pp.set_next_entity
> >       0.03 ±100%      +0.1        0.08 ± 15%
> >       perf-profile.self.cycles-pp.find_next_bit
> >       0.08 ±  5%      +0.1        0.14 ± 14%
> >       perf-profile.self.cycles-pp.native_write_msr
> >       0.01 ±200%      +0.1        0.07 ± 23%
> >       perf-profile.self.cycles-pp.kmem_cache_alloc
> >       0.29 ±  4%      +0.1        0.36 ±  8%
> >       perf-profile.self.cycles-pp.__orc_find
> >       0.14 ±  7%      +0.1        0.21 ± 12%
> >       perf-profile.self.cycles-pp.switch_mm_irqs_off
> >       0.00            +0.1        0.08 ± 11%
> >       perf-profile.self.cycles-pp.clear_page_erms
> >       0.00            +0.1        0.08 ± 28%
> >       perf-profile.self.cycles-pp.__indirect_thunk_start
> >       0.00            +0.1        0.08 ± 20%
> >       perf-profile.self.cycles-pp.md_wakeup_thread
> >       0.34 ±  6%      +0.1        0.43 ± 12%
> >       perf-profile.self.cycles-pp._raw_spin_lock_irqsave
> >       0.18 ±  4%      +0.1        0.27 ± 12%
> >       perf-profile.self.cycles-pp.idle_cpu
> >       0.16 ± 10%      +0.1        0.25 ± 13%
> >       perf-profile.self.cycles-pp.__module_address
> >       0.06 ± 11%      +0.1        0.17 ± 14%
> >       perf-profile.self.cycles-pp._raw_spin_unlock_irqrestore
> >       0.08 ± 38%      +0.1        0.20 ± 19%
> >       perf-profile.self.cycles-pp.io_serial_in
> >       0.18 ±  5%      +0.1        0.32 ± 15%
> >       perf-profile.self.cycles-pp.update_load_avg
> >       0.00            +0.1        0.15 ± 17%
> >       perf-profile.self.cycles-pp.poll_idle
> >       0.00            +0.2        0.15 ± 16%
> >       perf-profile.self.cycles-pp.menu_select
> >       0.00            +0.2        0.18 ± 24%
> >       perf-profile.self.cycles-pp.find_busiest_group
> >       0.02 ±152%      +0.3        0.35 ± 21%
> >       perf-profile.self.cycles-pp.raid1_write_request
> >       1.33 ±  4%      +8.8       10.12 ±  8%
> >       perf-profile.self.cycles-pp.intel_idle
> > 
> > 
> >                                                                                 
> >                                  aim7.jobs-per-min
> >                                                                                 
> >   1700
> >   +-+------------------------------------------------------------------+
> >        |+ ++++++ :+   ++++ ++++     +++     ++++++ +  +  ++++++++++++ ++
> >        |++|
> >   1600 +-+       + +++    +    +++++   ++.++      + ++ ++            +  ++
> >   |
> >        |                                                                    |   
> >        |                                                                    |   
> >   1500 +-+
> >   |
> >        |                                                                    |   
> >   1400 +-+
> >   |
> >        |                                                                    |   
> >   1300 +-+
> >   |
> >        |                                                                    |   
> >        O  OO         OO  O              O          O
> >        |
> >   1200 +OO  OOOOOOOOO  OO OOOOOOOOOOOOOO  OOOOOOOOO O
> >   |
> >        |                                                                    |   
> >   1100
> >   +-+------------------------------------------------------------------+
> >                                                                                 
> >                                                                                                                                                                 
> >                                                                                 
> > [*] bisect-good sample
> > [O] bisect-bad  sample
> > 
> > 
> > Disclaimer:
> > Results have been estimated based on internal Intel analysis and are
> > provided
> > for informational purposes only. Any difference in system hardware or
> > software
> > design or configuration may affect actual performance.
> > 
> > 
> > Thanks,
> > Xiaolong
> 

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ