lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Date:   Mon, 16 Jul 2018 15:54:30 +0800
From:   Aaron Lu <aaron.lu@...el.com>
To:     Xiao Ni <xni@...hat.com>
Cc:     kernel test robot <xiaolong.ye@...el.com>,
        Stephen Rothwell <sfr@...b.auug.org.au>, lkp@...org,
        LKML <linux-kernel@...r.kernel.org>, Shaohua Li <shli@...com>,
        Ming Lei <ming.lei@...hat.com>
Subject: Re: [LKP] [lkp-robot] [MD] 5a409b4f56: aim7.jobs-per-min -27.5%
 regression

Ping...
Any update on this?
Feel free to ask me for any additional data if you need.

Thanks,
Aaron

On Mon, Jun 04, 2018 at 02:42:03PM +0800, kernel test robot wrote:
> 
> Greeting,
> 
> FYI, we noticed a -27.5% regression of aim7.jobs-per-min due to commit:
> 
> 
> commit: 5a409b4f56d50b212334f338cb8465d65550cd85 ("MD: fix lock contention for flush bios")
> https://git.kernel.org/cgit/linux/kernel/git/next/linux-next.git master
> 
> in testcase: aim7
> on test machine: 40 threads Intel(R) Xeon(R) CPU E5-2690 v2 @ 3.00GHz with 384G memory
> with following parameters:
> 
> 	disk: 4BRD_12G
> 	md: RAID1
> 	fs: xfs
> 	test: sync_disk_rw
> 	load: 600
> 	cpufreq_governor: performance
> 
> test-description: AIM7 is a traditional UNIX system level benchmark suite which is used to test and measure the performance of multiuser system.
> test-url: https://sourceforge.net/projects/aimbench/files/aim-suite7/
> 
> 
> Details are as below:
> -------------------------------------------------------------------------------------------------->
> 
> =========================================================================================
> compiler/cpufreq_governor/disk/fs/kconfig/load/md/rootfs/tbox_group/test/testcase:
>   gcc-7/performance/4BRD_12G/xfs/x86_64-rhel-7.2/600/RAID1/debian-x86_64-2016-08-31.cgz/lkp-ivb-ep01/sync_disk_rw/aim7
> 
> commit: 
>   448ec638c6 ("md/raid5: Assigning NULL to sh->batch_head before testing bit R5_Overlap of a stripe")
>   5a409b4f56 ("MD: fix lock contention for flush bios")
> 
> 448ec638c6bcf369 5a409b4f56d50b212334f338cb 
> ---------------- -------------------------- 
>          %stddev     %change         %stddev
>              \          |                \  
>       1640           -27.5%       1189        aim7.jobs-per-min
>       2194           +37.9%       3026        aim7.time.elapsed_time
>       2194           +37.9%       3026        aim7.time.elapsed_time.max
>   50990311           -95.8%    2148266        aim7.time.involuntary_context_switches
>     107965 ±  4%     -26.4%      79516 ±  2%  aim7.time.minor_page_faults
>      49.14           +82.5%      89.66 ±  2%  aim7.time.user_time
>  7.123e+08           -35.7%  4.582e+08        aim7.time.voluntary_context_switches
>     672282           +36.8%     919615        interrupts.CAL:Function_call_interrupts
>   16631387 ±  2%     -39.9%    9993075 ±  7%  softirqs.RCU
>    9708009          +186.1%   27778773        softirqs.SCHED
>   33436649           +45.5%   48644912        softirqs.TIMER
>       4.16            -2.1        2.01        mpstat.cpu.idle%
>       0.24 ±  2%     +27.7       27.91        mpstat.cpu.iowait%
>      95.51           -25.6       69.94        mpstat.cpu.sys%
>       0.09            +0.0        0.13        mpstat.cpu.usr%
>    6051756 ±  3%     +59.0%    9623085        numa-numastat.node0.local_node
>    6055311 ±  3%     +59.0%    9626996        numa-numastat.node0.numa_hit
>    6481209 ±  3%     +48.4%    9616310        numa-numastat.node1.local_node
>    6485866 ±  3%     +48.3%    9620756        numa-numastat.node1.numa_hit
>      61404           -27.7%      44424        vmstat.io.bo
>       2.60 ± 18%  +11519.2%     302.10        vmstat.procs.b
>     304.10           -84.9%      45.80 ±  2%  vmstat.procs.r
>     400477           -43.5%     226094        vmstat.system.cs
>     166461           -49.9%      83332        vmstat.system.in
>      78397           +27.0%      99567        meminfo.Dirty
>      14427           +18.4%      17082        meminfo.Inactive(anon)
>       1963 ±  5%      +5.4%       2068 ±  4%  meminfo.Mlocked
>     101143          +991.0%    1103488        meminfo.SUnreclaim
>      53684 ±  4%     -18.1%      43946 ±  3%  meminfo.Shmem
>     175580          +571.4%    1178829        meminfo.Slab
>      39406           +26.2%      49717        numa-meminfo.node0.Dirty
>    1767204 ± 10%     +37.2%    2425487 ±  2%  numa-meminfo.node0.MemUsed
>      51634 ± 18%    +979.3%     557316        numa-meminfo.node0.SUnreclaim
>      92259 ± 13%    +551.7%     601288        numa-meminfo.node0.Slab
>      38969           +28.0%      49863        numa-meminfo.node1.Dirty
>    1895204 ± 10%     +24.7%    2363037 ±  3%  numa-meminfo.node1.MemUsed
>      49512 ± 19%   +1003.1%     546165        numa-meminfo.node1.SUnreclaim
>      83323 ± 14%    +593.1%     577534        numa-meminfo.node1.Slab
>  2.524e+09          +894.5%   2.51e+10        cpuidle.C1.time
>   50620790          +316.5%  2.109e+08        cpuidle.C1.usage
>  3.965e+08         +1871.1%  7.815e+09        cpuidle.C1E.time
>    5987788          +186.1%   17129412        cpuidle.C1E.usage
>  2.506e+08           +97.5%  4.948e+08 ±  2%  cpuidle.C3.time
>    2923498           -55.7%    1295033        cpuidle.C3.usage
>  5.327e+08          +179.9%  1.491e+09        cpuidle.C6.time
>     779874 ±  2%    +229.3%    2567769        cpuidle.C6.usage
>    6191357         +3333.6%  2.126e+08        cpuidle.POLL.time
>     204095         +1982.1%    4249504        cpuidle.POLL.usage
>       9850           +26.3%      12444        numa-vmstat.node0.nr_dirty
>      12908 ± 18%    +979.3%     139321        numa-vmstat.node0.nr_slab_unreclaimable
>       8876           +29.6%      11505        numa-vmstat.node0.nr_zone_write_pending
>    3486319 ±  4%     +55.1%    5407021        numa-vmstat.node0.numa_hit
>    3482713 ±  4%     +55.1%    5403066        numa-vmstat.node0.numa_local
>       9743           +28.1%      12479        numa-vmstat.node1.nr_dirty
>      12377 ± 19%   +1003.1%     136532        numa-vmstat.node1.nr_slab_unreclaimable
>       9287           +30.0%      12074        numa-vmstat.node1.nr_zone_write_pending
>    3678995 ±  4%     +44.8%    5326772        numa-vmstat.node1.numa_hit
>    3497785 ±  4%     +47.1%    5145705        numa-vmstat.node1.numa_local
>     252.70          +100.2%     505.90        slabinfo.biovec-max.active_objs
>     282.70           +99.1%     562.90        slabinfo.biovec-max.num_objs
>       2978 ± 17%     +52.5%       4543 ± 14%  slabinfo.dmaengine-unmap-16.active_objs
>       2978 ± 17%     +52.5%       4543 ± 14%  slabinfo.dmaengine-unmap-16.num_objs
>       2078          +147.9%       5153 ± 11%  slabinfo.ip6_dst_cache.active_objs
>       2078          +148.1%       5157 ± 11%  slabinfo.ip6_dst_cache.num_objs
>       5538 ±  2%     +26.2%       6990 ±  3%  slabinfo.kmalloc-1024.active_objs
>       5586 ±  3%     +27.1%       7097 ±  3%  slabinfo.kmalloc-1024.num_objs
>       6878           +47.6%      10151 ±  5%  slabinfo.kmalloc-192.active_objs
>       6889           +47.5%      10160 ±  5%  slabinfo.kmalloc-192.num_objs
>       9843 ±  5%  +1.6e+05%   16002876        slabinfo.kmalloc-64.active_objs
>     161.90 ±  4%  +1.5e+05%     250044        slabinfo.kmalloc-64.active_slabs
>      10386 ±  4%  +1.5e+05%   16002877        slabinfo.kmalloc-64.num_objs
>     161.90 ±  4%  +1.5e+05%     250044        slabinfo.kmalloc-64.num_slabs
>     432.80 ± 12%     +45.2%     628.50 ±  6%  slabinfo.nfs_read_data.active_objs
>     432.80 ± 12%     +45.2%     628.50 ±  6%  slabinfo.nfs_read_data.num_objs
>       3956           -23.1%       3041        slabinfo.pool_workqueue.active_objs
>       4098           -19.8%       3286        slabinfo.pool_workqueue.num_objs
>     360.50 ± 15%     +56.6%     564.70 ± 11%  slabinfo.secpath_cache.active_objs
>     360.50 ± 15%     +56.6%     564.70 ± 11%  slabinfo.secpath_cache.num_objs
>      35373 ±  2%      -8.3%      32432        proc-vmstat.nr_active_anon
>      19595           +27.1%      24914        proc-vmstat.nr_dirty
>       3607           +18.4%       4270        proc-vmstat.nr_inactive_anon
>     490.30 ±  5%      +5.4%     516.90 ±  4%  proc-vmstat.nr_mlock
>      13421 ±  4%     -18.1%      10986 ±  3%  proc-vmstat.nr_shmem
>      18608            +1.2%      18834        proc-vmstat.nr_slab_reclaimable
>      25286          +991.0%     275882        proc-vmstat.nr_slab_unreclaimable
>      35405 ±  2%      -8.3%      32465        proc-vmstat.nr_zone_active_anon
>       3607           +18.4%       4270        proc-vmstat.nr_zone_inactive_anon
>      18161           +29.8%      23572        proc-vmstat.nr_zone_write_pending
>      76941 ±  5%     -36.8%      48622 ±  4%  proc-vmstat.numa_hint_faults
>      33878 ±  7%     -35.5%      21836 ±  5%  proc-vmstat.numa_hint_faults_local
>   12568956           +53.3%   19272377        proc-vmstat.numa_hit
>   12560739           +53.4%   19264015        proc-vmstat.numa_local
>      17938 ±  3%     -33.5%      11935 ±  2%  proc-vmstat.numa_pages_migrated
>      78296 ±  5%     -36.0%      50085 ±  4%  proc-vmstat.numa_pte_updates
>       8848 ±  6%     -38.2%       5466 ±  6%  proc-vmstat.pgactivate
>    8874568 ±  8%    +368.7%   41590920        proc-vmstat.pgalloc_normal
>    5435965           +39.2%    7564148        proc-vmstat.pgfault
>   12863707          +255.1%   45683570        proc-vmstat.pgfree
>      17938 ±  3%     -33.5%      11935 ±  2%  proc-vmstat.pgmigrate_success
>  1.379e+13           -40.8%   8.17e+12        perf-stat.branch-instructions
>       0.30            +0.1        0.42        perf-stat.branch-miss-rate%
>    4.2e+10           -17.6%  3.462e+10        perf-stat.branch-misses
>      15.99            +3.8       19.74        perf-stat.cache-miss-rate%
>  3.779e+10           -21.6%  2.963e+10        perf-stat.cache-misses
>  2.364e+11           -36.5%  1.501e+11        perf-stat.cache-references
>  8.795e+08           -22.2%   6.84e+08        perf-stat.context-switches
>       4.44            -7.2%       4.12        perf-stat.cpi
>  2.508e+14           -44.5%  1.393e+14        perf-stat.cpu-cycles
>   36915392           +60.4%   59211221        perf-stat.cpu-migrations
>       0.29 ±  2%      +0.0        0.34 ±  4%  perf-stat.dTLB-load-miss-rate%
>   4.14e+10           -30.2%   2.89e+10 ±  4%  perf-stat.dTLB-load-misses
>  1.417e+13           -40.1%  8.491e+12        perf-stat.dTLB-loads
>       0.20 ±  4%      -0.0        0.18 ±  5%  perf-stat.dTLB-store-miss-rate%
>  3.072e+09 ±  4%     -28.0%   2.21e+09 ±  4%  perf-stat.dTLB-store-misses
>  1.535e+12           -20.2%  1.225e+12        perf-stat.dTLB-stores
>      90.73           -11.7       79.07        perf-stat.iTLB-load-miss-rate%
>  8.291e+09            -6.6%  7.743e+09        perf-stat.iTLB-load-misses
>  8.473e+08          +141.8%  2.049e+09 ±  3%  perf-stat.iTLB-loads
>  5.646e+13           -40.2%  3.378e+13        perf-stat.instructions
>       6810           -35.9%       4362        perf-stat.instructions-per-iTLB-miss
>       0.23            +7.8%       0.24        perf-stat.ipc
>    5326672           +39.2%    7413706        perf-stat.minor-faults
>  1.873e+10           -29.9%  1.312e+10        perf-stat.node-load-misses
>  2.093e+10           -29.2%  1.481e+10        perf-stat.node-loads
>      39.38            -0.7       38.72        perf-stat.node-store-miss-rate%
>  1.087e+10           -16.6%  9.069e+09        perf-stat.node-store-misses
>  1.673e+10           -14.2%  1.435e+10        perf-stat.node-stores
>    5326695           +39.2%    7413708        perf-stat.page-faults
>    1875095 ±  7%     -54.8%     846645 ± 16%  sched_debug.cfs_rq:/.MIN_vruntime.avg
>   32868920 ±  6%     -35.7%   21150379 ± 14%  sched_debug.cfs_rq:/.MIN_vruntime.max
>    7267340 ±  5%     -44.7%    4015798 ± 14%  sched_debug.cfs_rq:/.MIN_vruntime.stddev
>       4278 ±  7%     -54.7%       1939 ± 11%  sched_debug.cfs_rq:/.exec_clock.stddev
>     245.48 ±  2%     +65.3%     405.75 ±  7%  sched_debug.cfs_rq:/.load_avg.avg
>       2692 ±  6%    +126.0%       6087 ±  7%  sched_debug.cfs_rq:/.load_avg.max
>      33.09           -73.0%       8.94 ±  7%  sched_debug.cfs_rq:/.load_avg.min
>     507.40 ±  4%    +128.0%       1156 ±  7%  sched_debug.cfs_rq:/.load_avg.stddev
>    1875095 ±  7%     -54.8%     846645 ± 16%  sched_debug.cfs_rq:/.max_vruntime.avg
>   32868921 ±  6%     -35.7%   21150379 ± 14%  sched_debug.cfs_rq:/.max_vruntime.max
>    7267341 ±  5%     -44.7%    4015798 ± 14%  sched_debug.cfs_rq:/.max_vruntime.stddev
>   35887197           -13.2%   31149130        sched_debug.cfs_rq:/.min_vruntime.avg
>   37385506           -14.3%   32043914        sched_debug.cfs_rq:/.min_vruntime.max
>   34416296           -12.3%   30183927        sched_debug.cfs_rq:/.min_vruntime.min
>    1228844 ±  8%     -52.6%     582759 ±  4%  sched_debug.cfs_rq:/.min_vruntime.stddev
>       0.83           -28.1%       0.60 ±  6%  sched_debug.cfs_rq:/.nr_running.avg
>       2.07 ±  3%     -24.6%       1.56 ±  8%  sched_debug.cfs_rq:/.nr_running.max
>      20.52 ±  4%     -48.8%      10.52 ±  3%  sched_debug.cfs_rq:/.nr_spread_over.avg
>      35.96 ±  5%     -42.2%      20.77 ±  9%  sched_debug.cfs_rq:/.nr_spread_over.max
>       8.97 ± 11%     -44.5%       4.98 ±  8%  sched_debug.cfs_rq:/.nr_spread_over.min
>       6.40 ± 12%     -45.5%       3.49 ±  7%  sched_debug.cfs_rq:/.nr_spread_over.stddev
>      21.78 ±  7%    +143.3%      53.00 ±  9%  sched_debug.cfs_rq:/.runnable_load_avg.avg
>     328.86 ± 18%    +303.4%       1326 ± 14%  sched_debug.cfs_rq:/.runnable_load_avg.max
>      55.97 ± 17%    +286.0%     216.07 ± 13%  sched_debug.cfs_rq:/.runnable_load_avg.stddev
>       0.10 ± 29%     -82.4%       0.02 ± 50%  sched_debug.cfs_rq:/.spread.avg
>       3.43 ± 25%     -79.9%       0.69 ± 50%  sched_debug.cfs_rq:/.spread.max
>       0.56 ± 26%     -80.7%       0.11 ± 50%  sched_debug.cfs_rq:/.spread.stddev
>    1228822 ±  8%     -52.6%     582732 ±  4%  sched_debug.cfs_rq:/.spread0.stddev
>     992.30           -24.9%     745.56 ±  2%  sched_debug.cfs_rq:/.util_avg.avg
>       1485           -18.1%       1217 ±  2%  sched_debug.cfs_rq:/.util_avg.max
>     515.45 ±  2%     -25.2%     385.73 ±  6%  sched_debug.cfs_rq:/.util_avg.min
>     201.54           -14.9%     171.52 ±  3%  sched_debug.cfs_rq:/.util_avg.stddev
>     248.73 ±  6%     -38.1%     154.02 ±  8%  sched_debug.cfs_rq:/.util_est_enqueued.avg
>     222.78 ±  3%     -15.8%     187.58 ±  2%  sched_debug.cfs_rq:/.util_est_enqueued.stddev
>      77097 ±  4%    +278.4%     291767 ± 11%  sched_debug.cpu.avg_idle.avg
>     181319 ±  6%    +298.7%     722862 ±  3%  sched_debug.cpu.avg_idle.max
>      19338          +392.3%      95203 ± 17%  sched_debug.cpu.avg_idle.min
>      34877 ±  6%    +303.5%     140732 ±  6%  sched_debug.cpu.avg_idle.stddev
>    1107408           +37.6%    1523823        sched_debug.cpu.clock.avg
>    1107427           +37.6%    1523834        sched_debug.cpu.clock.max
>    1107385           +37.6%    1523811        sched_debug.cpu.clock.min
>      13.10 ±  9%     -48.1%       6.80 ±  8%  sched_debug.cpu.clock.stddev
>    1107408           +37.6%    1523823        sched_debug.cpu.clock_task.avg
>    1107427           +37.6%    1523834        sched_debug.cpu.clock_task.max
>    1107385           +37.6%    1523811        sched_debug.cpu.clock_task.min
>      13.10 ±  9%     -48.1%       6.80 ±  8%  sched_debug.cpu.clock_task.stddev
>      30.36 ±  7%    +107.7%      63.06 ± 12%  sched_debug.cpu.cpu_load[0].avg
>     381.48 ± 18%    +269.8%       1410 ± 18%  sched_debug.cpu.cpu_load[0].max
>      63.92 ± 18%    +262.2%     231.50 ± 17%  sched_debug.cpu.cpu_load[0].stddev
>      31.34 ±  5%    +118.4%      68.44 ±  9%  sched_debug.cpu.cpu_load[1].avg
>     323.62 ± 17%    +349.5%       1454 ± 14%  sched_debug.cpu.cpu_load[1].max
>      53.23 ± 16%    +350.3%     239.71 ± 13%  sched_debug.cpu.cpu_load[1].stddev
>      32.15 ±  3%    +129.4%      73.74 ±  6%  sched_debug.cpu.cpu_load[2].avg
>     285.20 ± 14%    +420.8%       1485 ±  9%  sched_debug.cpu.cpu_load[2].max
>      46.66 ± 12%    +430.0%     247.32 ±  8%  sched_debug.cpu.cpu_load[2].stddev
>      33.02 ±  2%    +133.2%      77.00 ±  3%  sched_debug.cpu.cpu_load[3].avg
>     252.16 ± 10%    +481.2%       1465 ±  7%  sched_debug.cpu.cpu_load[3].max
>      40.74 ±  8%    +503.2%     245.72 ±  6%  sched_debug.cpu.cpu_load[3].stddev
>      33.86          +131.5%      78.38 ±  2%  sched_debug.cpu.cpu_load[4].avg
>     219.81 ±  8%    +522.6%       1368 ±  5%  sched_debug.cpu.cpu_load[4].max
>      35.45 ±  7%    +554.2%     231.90 ±  4%  sched_debug.cpu.cpu_load[4].stddev
>       2600 ±  4%     -30.5%       1807 ±  4%  sched_debug.cpu.curr->pid.avg
>      25309 ±  4%     -19.5%      20367 ±  4%  sched_debug.cpu.curr->pid.max
>       4534 ±  7%     -21.2%       3573 ±  5%  sched_debug.cpu.curr->pid.stddev
>       0.00 ±  2%     -27.6%       0.00 ±  6%  sched_debug.cpu.next_balance.stddev
>    1083917           +38.6%    1502777        sched_debug.cpu.nr_load_updates.avg
>    1088142           +38.6%    1508302        sched_debug.cpu.nr_load_updates.max
>    1082048           +38.7%    1501073        sched_debug.cpu.nr_load_updates.min
>       3.53 ±  6%     -73.0%       0.95 ±  6%  sched_debug.cpu.nr_running.avg
>      11.54 ±  3%     -62.1%       4.37 ± 10%  sched_debug.cpu.nr_running.max
>       3.10 ±  3%     -66.8%       1.03 ±  9%  sched_debug.cpu.nr_running.stddev
>   10764176           -22.4%    8355047        sched_debug.cpu.nr_switches.avg
>   10976436           -22.2%    8545010        sched_debug.cpu.nr_switches.max
>   10547712           -22.8%    8143037        sched_debug.cpu.nr_switches.min
>     148628 ±  3%     -22.7%     114880 ±  7%  sched_debug.cpu.nr_switches.stddev
>      11.13 ±  2%     +24.5%      13.85        sched_debug.cpu.nr_uninterruptible.avg
>       6420 ±  8%     -48.7%       3296 ± 11%  sched_debug.cpu.nr_uninterruptible.max
>      -5500           -37.2%      -3455        sched_debug.cpu.nr_uninterruptible.min
>       3784 ±  6%     -47.2%       1997 ±  4%  sched_debug.cpu.nr_uninterruptible.stddev
>   10812670           -22.7%    8356821        sched_debug.cpu.sched_count.avg
>   11020646           -22.5%    8546277        sched_debug.cpu.sched_count.max
>   10601390           -23.2%    8144743        sched_debug.cpu.sched_count.min
>     144529 ±  3%     -20.9%     114359 ±  7%  sched_debug.cpu.sched_count.stddev
>     706116          +259.0%    2534721        sched_debug.cpu.sched_goidle.avg
>     771307          +232.4%    2564059        sched_debug.cpu.sched_goidle.max
>     644658          +286.9%    2494236        sched_debug.cpu.sched_goidle.min
>      49847 ±  6%     -67.9%      15979 ±  7%  sched_debug.cpu.sched_goidle.stddev
>    9618827           -39.9%    5780369        sched_debug.cpu.ttwu_count.avg
>    8990451           -61.7%    3441265 ±  4%  sched_debug.cpu.ttwu_count.min
>     418563 ± 25%    +244.2%    1440565 ±  7%  sched_debug.cpu.ttwu_count.stddev
>     640964           -93.7%      40366 ±  2%  sched_debug.cpu.ttwu_local.avg
>     679527           -92.1%      53476 ±  4%  sched_debug.cpu.ttwu_local.max
>     601661           -94.9%      30636 ±  3%  sched_debug.cpu.ttwu_local.min
>      24242 ± 21%     -77.7%       5405 ±  9%  sched_debug.cpu.ttwu_local.stddev
>    1107383           +37.6%    1523810        sched_debug.cpu_clk
>    1107383           +37.6%    1523810        sched_debug.ktime
>       0.00           -49.4%       0.00 ± 65%  sched_debug.rt_rq:/.rt_nr_migratory.avg
>       0.03           -49.4%       0.01 ± 65%  sched_debug.rt_rq:/.rt_nr_migratory.max
>       0.00           -49.4%       0.00 ± 65%  sched_debug.rt_rq:/.rt_nr_migratory.stddev
>       0.00           -49.4%       0.00 ± 65%  sched_debug.rt_rq:/.rt_nr_running.avg
>       0.03           -49.4%       0.01 ± 65%  sched_debug.rt_rq:/.rt_nr_running.max
>       0.00           -49.4%       0.00 ± 65%  sched_debug.rt_rq:/.rt_nr_running.stddev
>       0.01 ±  8%     +79.9%       0.01 ± 23%  sched_debug.rt_rq:/.rt_time.avg
>    1107805           +37.6%    1524235        sched_debug.sched_clk
>      87.59           -87.6        0.00        perf-profile.calltrace.cycles-pp.md_flush_request.raid1_make_request.md_handle_request.md_make_request.generic_make_request
>      87.57           -87.6        0.00        perf-profile.calltrace.cycles-pp.submit_bio_wait.blkdev_issue_flush.xfs_file_fsync.xfs_file_write_iter.__vfs_write
>      87.59           -87.5        0.05 ±299%  perf-profile.calltrace.cycles-pp.blkdev_issue_flush.xfs_file_fsync.xfs_file_write_iter.__vfs_write.vfs_write
>      87.51           -87.5        0.00        perf-profile.calltrace.cycles-pp.generic_make_request.submit_bio.submit_bio_wait.blkdev_issue_flush.xfs_file_fsync
>      87.51           -87.5        0.00        perf-profile.calltrace.cycles-pp.submit_bio.submit_bio_wait.blkdev_issue_flush.xfs_file_fsync.xfs_file_write_iter
>      87.50           -87.5        0.00        perf-profile.calltrace.cycles-pp.md_make_request.generic_make_request.submit_bio.submit_bio_wait.blkdev_issue_flush
>      87.50           -87.5        0.00        perf-profile.calltrace.cycles-pp.md_handle_request.md_make_request.generic_make_request.submit_bio.submit_bio_wait
>      82.37           -82.4        0.00        perf-profile.calltrace.cycles-pp._raw_spin_lock_irq.md_flush_request.raid1_make_request.md_handle_request.md_make_request
>      82.23           -82.2        0.00        perf-profile.calltrace.cycles-pp.native_queued_spin_lock_slowpath._raw_spin_lock_irq.md_flush_request.raid1_make_request.md_handle_request
>      87.79           -25.0       62.75 ±  8%  perf-profile.calltrace.cycles-pp.raid1_make_request.md_handle_request.md_make_request.generic_make_request.submit_bio
>      92.78           -13.0       79.76        perf-profile.calltrace.cycles-pp.xfs_file_fsync.xfs_file_write_iter.__vfs_write.vfs_write.ksys_write
>      93.08           -12.6       80.49        perf-profile.calltrace.cycles-pp.xfs_file_write_iter.__vfs_write.vfs_write.ksys_write.do_syscall_64
>      93.08           -12.6       80.50        perf-profile.calltrace.cycles-pp.__vfs_write.vfs_write.ksys_write.do_syscall_64.entry_SYSCALL_64_after_hwframe
>      93.11           -12.6       80.56        perf-profile.calltrace.cycles-pp.vfs_write.ksys_write.do_syscall_64.entry_SYSCALL_64_after_hwframe
>      93.11           -12.6       80.56        perf-profile.calltrace.cycles-pp.ksys_write.do_syscall_64.entry_SYSCALL_64_after_hwframe
>      93.14           -12.5       80.64        perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe
>      93.15           -12.5       80.65        perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe
>       3.40 ±  2%      -1.4        1.97 ±  8%  perf-profile.calltrace.cycles-pp.worker_thread.kthread.ret_from_fork
>       3.33 ±  2%      -1.4        1.96 ±  9%  perf-profile.calltrace.cycles-pp.process_one_work.worker_thread.kthread.ret_from_fork
>       1.12 ±  2%      -0.7        0.42 ± 68%  perf-profile.calltrace.cycles-pp.__save_stack_trace.save_stack_trace_tsk.__account_scheduler_latency.enqueue_entity.enqueue_task_fair
>       1.16 ±  2%      -0.6        0.60 ± 17%  perf-profile.calltrace.cycles-pp.save_stack_trace_tsk.__account_scheduler_latency.enqueue_entity.enqueue_task_fair.ttwu_do_activate
>       0.00            +0.6        0.59 ± 15%  perf-profile.calltrace.cycles-pp._raw_spin_lock_irqsave.__wake_up_common_lock.raid1_write_request.raid1_make_request.md_handle_request
>       0.00            +0.6        0.64 ± 15%  perf-profile.calltrace.cycles-pp.__wake_up_common_lock.raid1_write_request.raid1_make_request.md_handle_request.md_make_request
>       0.00            +0.7        0.65 ± 10%  perf-profile.calltrace.cycles-pp.enqueue_entity.enqueue_task_fair.ttwu_do_activate.sched_ttwu_pending.do_idle
>       0.00            +0.7        0.68 ± 10%  perf-profile.calltrace.cycles-pp.enqueue_task_fair.ttwu_do_activate.sched_ttwu_pending.do_idle.cpu_startup_entry
>       0.00            +0.7        0.69 ± 10%  perf-profile.calltrace.cycles-pp.ttwu_do_activate.sched_ttwu_pending.do_idle.cpu_startup_entry.start_secondary
>       0.00            +0.8        0.79 ± 11%  perf-profile.calltrace.cycles-pp.sched_ttwu_pending.do_idle.cpu_startup_entry.start_secondary.secondary_startup_64
>       0.00            +0.8        0.83 ±  7%  perf-profile.calltrace.cycles-pp.__schedule.schedule.raid1_write_request.raid1_make_request.md_handle_request
>       0.62 ±  3%      +0.8        1.45 ± 22%  perf-profile.calltrace.cycles-pp.native_queued_spin_lock_slowpath._raw_spin_lock_irqsave.remove_wait_queue.xlog_wait.__xfs_log_force_lsn
>       0.00            +0.8        0.83 ±  7%  perf-profile.calltrace.cycles-pp.schedule.raid1_write_request.raid1_make_request.md_handle_request.md_make_request
>       0.63 ±  2%      +0.8        1.46 ± 22%  perf-profile.calltrace.cycles-pp.remove_wait_queue.xlog_wait.__xfs_log_force_lsn.xfs_log_force_lsn.xfs_file_fsync
>       0.62 ±  2%      +0.8        1.46 ± 22%  perf-profile.calltrace.cycles-pp._raw_spin_lock_irqsave.remove_wait_queue.xlog_wait.__xfs_log_force_lsn.xfs_log_force_lsn
>       3.92 ±  2%      +0.9        4.79 ±  6%  perf-profile.calltrace.cycles-pp.ret_from_fork
>       3.92 ±  2%      +0.9        4.79 ±  6%  perf-profile.calltrace.cycles-pp.kthread.ret_from_fork
>       0.69 ±  2%      +0.9        1.64 ± 23%  perf-profile.calltrace.cycles-pp.xlog_wait.__xfs_log_force_lsn.xfs_log_force_lsn.xfs_file_fsync.xfs_file_write_iter
>       0.00            +1.2        1.17 ±  8%  perf-profile.calltrace.cycles-pp._raw_spin_unlock_irqrestore.prepare_to_wait_event.raid1_write_request.raid1_make_request.md_handle_request
>       0.00            +1.2        1.23 ± 18%  perf-profile.calltrace.cycles-pp.prepare_to_wait_event.raid1_write_request.raid1_make_request.md_handle_request.submit_flushes
>       0.00            +1.3        1.27 ± 17%  perf-profile.calltrace.cycles-pp.raid1_write_request.raid1_make_request.md_handle_request.submit_flushes.process_one_work
>       0.00            +1.3        1.27 ± 17%  perf-profile.calltrace.cycles-pp.md_handle_request.submit_flushes.process_one_work.worker_thread.kthread
>       0.00            +1.3        1.27 ± 17%  perf-profile.calltrace.cycles-pp.raid1_make_request.md_handle_request.submit_flushes.process_one_work.worker_thread
>       0.00            +1.3        1.27 ± 17%  perf-profile.calltrace.cycles-pp.submit_flushes.process_one_work.worker_thread.kthread.ret_from_fork
>       0.00            +1.6        1.65 ± 14%  perf-profile.calltrace.cycles-pp.try_to_wake_up.autoremove_wake_function.__wake_up_common.__wake_up_common_lock.raid_end_bio_io
>       0.00            +1.7        1.71 ± 14%  perf-profile.calltrace.cycles-pp.autoremove_wake_function.__wake_up_common.__wake_up_common_lock.raid_end_bio_io.raid1_end_write_request
>       0.00            +1.7        1.71 ± 14%  perf-profile.calltrace.cycles-pp.__wake_up_common.__wake_up_common_lock.raid_end_bio_io.raid1_end_write_request.brd_make_request
>       0.00            +1.9        1.86 ± 13%  perf-profile.calltrace.cycles-pp.__wake_up_common_lock.raid_end_bio_io.raid1_end_write_request.brd_make_request.generic_make_request
>       0.00            +2.1        2.10 ± 10%  perf-profile.calltrace.cycles-pp.native_queued_spin_lock_slowpath._raw_spin_lock_irqsave.remove_wait_queue.__xfs_log_force_lsn.xfs_log_force_lsn
>       0.00            +2.1        2.10 ± 10%  perf-profile.calltrace.cycles-pp._raw_spin_lock_irqsave.remove_wait_queue.__xfs_log_force_lsn.xfs_log_force_lsn.xfs_file_fsync
>       0.00            +2.1        2.11 ± 10%  perf-profile.calltrace.cycles-pp.remove_wait_queue.__xfs_log_force_lsn.xfs_log_force_lsn.xfs_file_fsync.xfs_file_write_iter
>       0.00            +2.2        2.16 ± 10%  perf-profile.calltrace.cycles-pp.raid_end_bio_io.raid1_end_write_request.brd_make_request.generic_make_request.flush_bio_list
>       2.24 ±  4%      +2.2        4.44 ± 15%  perf-profile.calltrace.cycles-pp.xfs_log_force_lsn.xfs_file_fsync.xfs_file_write_iter.__vfs_write.vfs_write
>       0.00            +2.3        2.25 ± 10%  perf-profile.calltrace.cycles-pp.raid1_end_write_request.brd_make_request.generic_make_request.flush_bio_list.flush_pending_writes
>       0.00            +2.3        2.30 ± 20%  perf-profile.calltrace.cycles-pp.native_queued_spin_lock_slowpath._raw_spin_lock_irqsave.raid1_write_request.raid1_make_request.md_handle_request
>       0.00            +2.4        2.35 ± 20%  perf-profile.calltrace.cycles-pp._raw_spin_lock_irqsave.raid1_write_request.raid1_make_request.md_handle_request.md_make_request
>       0.37 ± 65%      +2.4        2.81 ±  7%  perf-profile.calltrace.cycles-pp.md_thread.kthread.ret_from_fork
>       0.26 ±100%      +2.5        2.81 ±  7%  perf-profile.calltrace.cycles-pp.raid1d.md_thread.kthread.ret_from_fork
>       0.26 ±100%      +2.5        2.81 ±  7%  perf-profile.calltrace.cycles-pp.flush_pending_writes.raid1d.md_thread.kthread.ret_from_fork
>       0.26 ±100%      +2.6        2.81 ±  7%  perf-profile.calltrace.cycles-pp.flush_bio_list.flush_pending_writes.raid1d.md_thread.kthread
>       0.10 ±200%      +2.7        2.76 ±  7%  perf-profile.calltrace.cycles-pp.generic_make_request.flush_bio_list.flush_pending_writes.raid1d.md_thread
>       0.00            +2.7        2.73 ±  7%  perf-profile.calltrace.cycles-pp.brd_make_request.generic_make_request.flush_bio_list.flush_pending_writes.raid1d
>       1.20 ±  3%      +3.1        4.35 ± 15%  perf-profile.calltrace.cycles-pp.__xfs_log_force_lsn.xfs_log_force_lsn.xfs_file_fsync.xfs_file_write_iter.__vfs_write
>       0.63 ±  6%      +3.8        4.38 ± 27%  perf-profile.calltrace.cycles-pp.native_queued_spin_lock_slowpath._raw_spin_lock_irqsave.remove_wait_queue.__xfs_log_force_lsn.xfs_file_fsync
>       0.63 ±  5%      +3.8        4.39 ± 27%  perf-profile.calltrace.cycles-pp._raw_spin_lock_irqsave.remove_wait_queue.__xfs_log_force_lsn.xfs_file_fsync.xfs_file_write_iter
>       0.63 ±  5%      +3.8        4.40 ± 27%  perf-profile.calltrace.cycles-pp.remove_wait_queue.__xfs_log_force_lsn.xfs_file_fsync.xfs_file_write_iter.__vfs_write
>       1.26 ±  5%      +5.3        6.55 ± 27%  perf-profile.calltrace.cycles-pp.native_queued_spin_lock_slowpath._raw_spin_lock.__xfs_log_force_lsn.xfs_file_fsync.xfs_file_write_iter
>       1.27 ±  5%      +5.3        6.55 ± 27%  perf-profile.calltrace.cycles-pp._raw_spin_lock.__xfs_log_force_lsn.xfs_file_fsync.xfs_file_write_iter.__vfs_write
>       1.30 ±  4%      +8.4        9.72 ±  9%  perf-profile.calltrace.cycles-pp.intel_idle.cpuidle_enter_state.do_idle.cpu_startup_entry.start_secondary
>       1.33 ±  4%      +8.9       10.26 ±  9%  perf-profile.calltrace.cycles-pp.cpuidle_enter_state.do_idle.cpu_startup_entry.start_secondary.secondary_startup_64
>       2.28 ±  2%      +9.1       11.36 ± 27%  perf-profile.calltrace.cycles-pp.__xfs_log_force_lsn.xfs_file_fsync.xfs_file_write_iter.__vfs_write.vfs_write
>       1.59 ±  4%     +10.4       11.97 ±  9%  perf-profile.calltrace.cycles-pp.do_idle.cpu_startup_entry.start_secondary.secondary_startup_64
>       1.59 ±  4%     +10.4       11.98 ±  9%  perf-profile.calltrace.cycles-pp.cpu_startup_entry.start_secondary.secondary_startup_64
>       1.59 ±  4%     +10.4       11.98 ±  9%  perf-profile.calltrace.cycles-pp.start_secondary.secondary_startup_64
>       1.63 ±  4%     +10.8       12.47 ±  8%  perf-profile.calltrace.cycles-pp.secondary_startup_64
>       0.00           +57.7       57.66 ± 10%  perf-profile.calltrace.cycles-pp.native_queued_spin_lock_slowpath._raw_spin_lock_irqsave.prepare_to_wait_event.raid1_write_request.raid1_make_request
>       0.00           +57.7       57.73 ± 10%  perf-profile.calltrace.cycles-pp._raw_spin_lock_irqsave.prepare_to_wait_event.raid1_write_request.raid1_make_request.md_handle_request
>       0.05 ±299%     +57.8       57.85 ±  9%  perf-profile.calltrace.cycles-pp.prepare_to_wait_event.raid1_write_request.raid1_make_request.md_handle_request.md_make_request
>       0.19 ±154%     +62.5       62.73 ±  8%  perf-profile.calltrace.cycles-pp.raid1_write_request.raid1_make_request.md_handle_request.md_make_request.generic_make_request
>       0.19 ±154%     +62.6       62.76 ±  8%  perf-profile.calltrace.cycles-pp.md_handle_request.md_make_request.generic_make_request.submit_bio.xfs_submit_ioend
>       0.19 ±154%     +62.6       62.79 ±  8%  perf-profile.calltrace.cycles-pp.md_make_request.generic_make_request.submit_bio.xfs_submit_ioend.xfs_vm_writepages
>       0.20 ±154%     +62.6       62.81 ±  8%  perf-profile.calltrace.cycles-pp.generic_make_request.submit_bio.xfs_submit_ioend.xfs_vm_writepages.do_writepages
>       0.20 ±154%     +62.6       62.81 ±  8%  perf-profile.calltrace.cycles-pp.submit_bio.xfs_submit_ioend.xfs_vm_writepages.do_writepages.__filemap_fdatawrite_range
>       0.20 ±154%     +62.6       62.82 ±  8%  perf-profile.calltrace.cycles-pp.xfs_submit_ioend.xfs_vm_writepages.do_writepages.__filemap_fdatawrite_range.file_write_and_wait_range
>       0.29 ±125%     +62.8       63.09 ±  8%  perf-profile.calltrace.cycles-pp.xfs_vm_writepages.do_writepages.__filemap_fdatawrite_range.file_write_and_wait_range.xfs_file_fsync
>       0.29 ±126%     +62.8       63.10 ±  8%  perf-profile.calltrace.cycles-pp.do_writepages.__filemap_fdatawrite_range.file_write_and_wait_range.xfs_file_fsync.xfs_file_write_iter
>       0.29 ±125%     +62.8       63.11 ±  8%  perf-profile.calltrace.cycles-pp.__filemap_fdatawrite_range.file_write_and_wait_range.xfs_file_fsync.xfs_file_write_iter.__vfs_write
>       0.62 ± 41%     +62.9       63.52 ±  7%  perf-profile.calltrace.cycles-pp.file_write_and_wait_range.xfs_file_fsync.xfs_file_write_iter.__vfs_write.vfs_write
>      88.51           -88.2        0.26 ± 19%  perf-profile.children.cycles-pp.md_flush_request
>      87.57           -87.2        0.35 ± 19%  perf-profile.children.cycles-pp.submit_bio_wait
>      87.59           -87.2        0.39 ± 19%  perf-profile.children.cycles-pp.blkdev_issue_flush
>      83.26           -83.2        0.02 ±123%  perf-profile.children.cycles-pp._raw_spin_lock_irq
>      88.85           -25.7       63.11 ±  8%  perf-profile.children.cycles-pp.md_make_request
>      88.90           -25.7       63.17 ±  8%  perf-profile.children.cycles-pp.submit_bio
>      88.83           -24.5       64.31 ±  8%  perf-profile.children.cycles-pp.raid1_make_request
>      88.84           -24.5       64.33 ±  8%  perf-profile.children.cycles-pp.md_handle_request
>      89.38           -23.5       65.92 ±  7%  perf-profile.children.cycles-pp.generic_make_request
>      89.90           -13.4       76.51 ±  2%  perf-profile.children.cycles-pp.native_queued_spin_lock_slowpath
>      92.79           -13.0       79.76        perf-profile.children.cycles-pp.xfs_file_fsync
>      93.08           -12.6       80.49        perf-profile.children.cycles-pp.xfs_file_write_iter
>      93.09           -12.6       80.54        perf-profile.children.cycles-pp.__vfs_write
>      93.13           -12.5       80.60        perf-profile.children.cycles-pp.vfs_write
>      93.13           -12.5       80.61        perf-profile.children.cycles-pp.ksys_write
>      93.22           -12.4       80.83        perf-profile.children.cycles-pp.do_syscall_64
>      93.22           -12.4       80.83        perf-profile.children.cycles-pp.entry_SYSCALL_64_after_hwframe
>       3.40 ±  2%      -1.4        1.97 ±  8%  perf-profile.children.cycles-pp.worker_thread
>       3.33 ±  2%      -1.4        1.96 ±  9%  perf-profile.children.cycles-pp.process_one_work
>       1.03 ±  7%      -1.0        0.07 ± 37%  perf-profile.children.cycles-pp.xlog_cil_force_lsn
>       1.69 ±  2%      -0.7        0.96 ±  4%  perf-profile.children.cycles-pp.reschedule_interrupt
>       1.66 ±  2%      -0.7        0.94 ±  4%  perf-profile.children.cycles-pp.scheduler_ipi
>       1.13 ±  2%      -0.7        0.47 ± 11%  perf-profile.children.cycles-pp.finish_wait
>       0.54 ±  8%      -0.4        0.10 ± 38%  perf-profile.children.cycles-pp.xlog_cil_push
>       0.49 ±  9%      -0.4        0.09 ± 35%  perf-profile.children.cycles-pp.xlog_write
>       0.10 ±  8%      -0.1        0.04 ± 67%  perf-profile.children.cycles-pp.flush_work
>       0.20 ±  5%      -0.0        0.16 ± 11%  perf-profile.children.cycles-pp.reweight_entity
>       0.06 ± 10%      +0.0        0.10 ± 23%  perf-profile.children.cycles-pp.brd_lookup_page
>       0.18 ±  5%      +0.0        0.23 ± 13%  perf-profile.children.cycles-pp.__update_load_avg_se
>       0.02 ±153%      +0.1        0.07 ± 16%  perf-profile.children.cycles-pp.delay_tsc
>       0.03 ±100%      +0.1        0.08 ± 15%  perf-profile.children.cycles-pp.find_next_bit
>       0.08 ±  5%      +0.1        0.14 ± 14%  perf-profile.children.cycles-pp.native_write_msr
>       0.29 ±  4%      +0.1        0.36 ±  8%  perf-profile.children.cycles-pp.__orc_find
>       0.40 ±  4%      +0.1        0.46 ±  7%  perf-profile.children.cycles-pp.dequeue_task_fair
>       0.11 ± 11%      +0.1        0.18 ± 14%  perf-profile.children.cycles-pp.__module_text_address
>       0.12 ±  8%      +0.1        0.19 ± 13%  perf-profile.children.cycles-pp.is_module_text_address
>       0.04 ± 50%      +0.1        0.12 ± 19%  perf-profile.children.cycles-pp.kmem_cache_alloc
>       0.00            +0.1        0.08 ± 11%  perf-profile.children.cycles-pp.clear_page_erms
>       0.00            +0.1        0.08 ± 28%  perf-profile.children.cycles-pp.__indirect_thunk_start
>       0.01 ±200%      +0.1        0.10 ± 25%  perf-profile.children.cycles-pp.xfs_trans_alloc
>       0.00            +0.1        0.09 ± 18%  perf-profile.children.cycles-pp.md_wakeup_thread
>       0.00            +0.1        0.09 ± 26%  perf-profile.children.cycles-pp.rebalance_domains
>       0.00            +0.1        0.09 ± 26%  perf-profile.children.cycles-pp.get_next_timer_interrupt
>       0.00            +0.1        0.09 ± 20%  perf-profile.children.cycles-pp.ktime_get
>       0.18 ±  4%      +0.1        0.27 ± 12%  perf-profile.children.cycles-pp.idle_cpu
>       0.20 ±  6%      +0.1        0.30 ±  9%  perf-profile.children.cycles-pp.unwind_get_return_address
>       0.16 ± 10%      +0.1        0.25 ± 13%  perf-profile.children.cycles-pp.__module_address
>       0.03 ±100%      +0.1        0.13 ±  8%  perf-profile.children.cycles-pp.brd_insert_page
>       0.06 ±  9%      +0.1        0.16 ± 14%  perf-profile.children.cycles-pp.task_tick_fair
>       0.08 ± 12%      +0.1        0.18 ± 24%  perf-profile.children.cycles-pp.bio_alloc_bioset
>       0.03 ± 81%      +0.1        0.14 ± 27%  perf-profile.children.cycles-pp.generic_make_request_checks
>       0.17 ±  7%      +0.1        0.28 ± 11%  perf-profile.children.cycles-pp.__kernel_text_address
>       0.11 ±  9%      +0.1        0.22 ± 15%  perf-profile.children.cycles-pp.wake_up_page_bit
>       0.16 ±  6%      +0.1        0.27 ± 10%  perf-profile.children.cycles-pp.kernel_text_address
>       0.00            +0.1        0.11 ± 11%  perf-profile.children.cycles-pp.get_page_from_freelist
>       0.00            +0.1        0.11 ± 19%  perf-profile.children.cycles-pp.perf_mux_hrtimer_handler
>       0.00            +0.1        0.11 ±  7%  perf-profile.children.cycles-pp.__alloc_pages_nodemask
>       0.08 ± 10%      +0.1        0.19 ± 22%  perf-profile.children.cycles-pp.xfs_do_writepage
>       0.25 ±  4%      +0.1        0.37 ± 10%  perf-profile.children.cycles-pp.switch_mm_irqs_off
>       0.00            +0.1        0.12 ± 13%  perf-profile.children.cycles-pp.switch_mm
>       0.08 ± 38%      +0.1        0.20 ± 19%  perf-profile.children.cycles-pp.io_serial_in
>       0.18 ±  5%      +0.1        0.31 ±  7%  perf-profile.children.cycles-pp.dequeue_entity
>       0.00            +0.1        0.13 ± 26%  perf-profile.children.cycles-pp.tick_nohz_next_event
>       0.06 ± 11%      +0.1        0.19 ± 19%  perf-profile.children.cycles-pp.mempool_alloc
>       0.32 ±  5%      +0.1        0.45 ±  6%  perf-profile.children.cycles-pp.orc_find
>       0.15 ± 10%      +0.1        0.29 ± 19%  perf-profile.children.cycles-pp.xfs_destroy_ioend
>       0.15 ± 11%      +0.1        0.30 ± 18%  perf-profile.children.cycles-pp.call_bio_endio
>       0.08 ± 17%      +0.2        0.23 ± 25%  perf-profile.children.cycles-pp.xlog_state_done_syncing
>       0.00            +0.2        0.15 ± 22%  perf-profile.children.cycles-pp.tick_nohz_get_sleep_length
>       0.12 ±  8%      +0.2        0.27 ± 23%  perf-profile.children.cycles-pp.write_cache_pages
>       0.10 ± 16%      +0.2        0.26 ± 16%  perf-profile.children.cycles-pp.wait_for_xmitr
>       0.10 ± 19%      +0.2        0.25 ± 14%  perf-profile.children.cycles-pp.serial8250_console_putchar
>       0.10 ± 17%      +0.2        0.26 ± 13%  perf-profile.children.cycles-pp.uart_console_write
>       0.10 ± 16%      +0.2        0.26 ± 15%  perf-profile.children.cycles-pp.serial8250_console_write
>       0.11 ± 15%      +0.2        0.27 ± 15%  perf-profile.children.cycles-pp.console_unlock
>       0.09 ±  9%      +0.2        0.26 ± 12%  perf-profile.children.cycles-pp.scheduler_tick
>       0.10 ± 18%      +0.2        0.28 ± 15%  perf-profile.children.cycles-pp.irq_work_run_list
>       0.10 ± 15%      +0.2        0.28 ± 14%  perf-profile.children.cycles-pp.xlog_state_do_callback
>       0.09 ± 12%      +0.2        0.27 ± 16%  perf-profile.children.cycles-pp.irq_work_run
>       0.09 ± 12%      +0.2        0.27 ± 16%  perf-profile.children.cycles-pp.printk
>       0.09 ± 12%      +0.2        0.27 ± 16%  perf-profile.children.cycles-pp.vprintk_emit
>       0.09 ± 12%      +0.2        0.27 ± 17%  perf-profile.children.cycles-pp.irq_work_interrupt
>       0.09 ± 12%      +0.2        0.27 ± 17%  perf-profile.children.cycles-pp.smp_irq_work_interrupt
>       0.00            +0.2        0.18 ± 16%  perf-profile.children.cycles-pp.poll_idle
>       0.30 ±  4%      +0.2        0.49 ± 11%  perf-profile.children.cycles-pp.update_load_avg
>       1.39 ±  2%      +0.2        1.59 ±  6%  perf-profile.children.cycles-pp.__save_stack_trace
>       1.43            +0.2        1.65 ±  6%  perf-profile.children.cycles-pp.save_stack_trace_tsk
>       0.14 ± 13%      +0.2        0.36 ± 13%  perf-profile.children.cycles-pp.update_process_times
>       0.00            +0.2        0.23 ± 22%  perf-profile.children.cycles-pp.find_busiest_group
>       0.22 ±  6%      +0.2        0.45 ± 18%  perf-profile.children.cycles-pp.brd_do_bvec
>       0.14 ± 13%      +0.2        0.38 ± 14%  perf-profile.children.cycles-pp.tick_sched_handle
>       0.10 ±  8%      +0.2        0.34 ± 26%  perf-profile.children.cycles-pp.xfs_log_commit_cil
>       0.07 ± 10%      +0.3        0.33 ± 23%  perf-profile.children.cycles-pp.io_schedule
>       0.03 ± 83%      +0.3        0.29 ± 27%  perf-profile.children.cycles-pp.__softirqentry_text_start
>       0.11 ±  5%      +0.3        0.36 ± 25%  perf-profile.children.cycles-pp.__xfs_trans_commit
>       0.06 ± 36%      +0.3        0.31 ± 26%  perf-profile.children.cycles-pp.irq_exit
>       0.08 ±  9%      +0.3        0.35 ± 23%  perf-profile.children.cycles-pp.wait_on_page_bit_common
>       0.15 ± 12%      +0.3        0.42 ± 14%  perf-profile.children.cycles-pp.tick_sched_timer
>       0.10 ± 11%      +0.3        0.39 ± 22%  perf-profile.children.cycles-pp.__filemap_fdatawait_range
>       0.06 ± 12%      +0.3        0.37 ±  9%  perf-profile.children.cycles-pp.schedule_idle
>       0.02 ±153%      +0.3        0.34 ± 17%  perf-profile.children.cycles-pp.menu_select
>       0.17 ±  5%      +0.3        0.49 ± 22%  perf-profile.children.cycles-pp.xfs_vn_update_time
>       0.19 ± 12%      +0.3        0.51 ± 18%  perf-profile.children.cycles-pp.xlog_iodone
>       0.18 ±  5%      +0.3        0.51 ± 22%  perf-profile.children.cycles-pp.file_update_time
>       0.18 ±  5%      +0.3        0.51 ± 21%  perf-profile.children.cycles-pp.xfs_file_aio_write_checks
>       0.21 ± 11%      +0.4        0.60 ± 15%  perf-profile.children.cycles-pp.__hrtimer_run_queues
>       0.26 ±  6%      +0.4        0.69 ± 16%  perf-profile.children.cycles-pp.pick_next_task_fair
>       1.20 ±  2%      +0.4        1.64 ± 10%  perf-profile.children.cycles-pp.schedule
>       0.28 ±  5%      +0.4        0.72 ± 21%  perf-profile.children.cycles-pp.xfs_file_buffered_aio_write
>       0.00            +0.4        0.44 ± 22%  perf-profile.children.cycles-pp.load_balance
>       0.25 ±  8%      +0.5        0.74 ± 15%  perf-profile.children.cycles-pp.hrtimer_interrupt
>       1.30 ±  2%      +0.7        2.00 ±  9%  perf-profile.children.cycles-pp.__schedule
>       0.31 ±  8%      +0.8        1.09 ± 16%  perf-profile.children.cycles-pp.smp_apic_timer_interrupt
>       0.31 ±  8%      +0.8        1.09 ± 16%  perf-profile.children.cycles-pp.apic_timer_interrupt
>       3.92 ±  2%      +0.9        4.79 ±  6%  perf-profile.children.cycles-pp.ret_from_fork
>       3.92 ±  2%      +0.9        4.79 ±  6%  perf-profile.children.cycles-pp.kthread
>       0.69 ±  2%      +0.9        1.64 ± 23%  perf-profile.children.cycles-pp.xlog_wait
>       0.08 ± 13%      +1.2        1.27 ± 17%  perf-profile.children.cycles-pp.submit_flushes
>       0.16 ±  9%      +1.6        1.74 ±  4%  perf-profile.children.cycles-pp._raw_spin_unlock_irqrestore
>       0.17 ±  9%      +2.0        2.16 ± 10%  perf-profile.children.cycles-pp.raid_end_bio_io
>       0.21 ±  6%      +2.0        2.25 ± 10%  perf-profile.children.cycles-pp.raid1_end_write_request
>       2.24 ±  4%      +2.2        4.44 ± 15%  perf-profile.children.cycles-pp.xfs_log_force_lsn
>       0.46 ±  6%      +2.3        2.73 ±  7%  perf-profile.children.cycles-pp.brd_make_request
>       0.51 ±  6%      +2.3        2.81 ±  7%  perf-profile.children.cycles-pp.md_thread
>       0.49 ±  6%      +2.3        2.81 ±  7%  perf-profile.children.cycles-pp.raid1d
>       0.49 ±  6%      +2.3        2.81 ±  7%  perf-profile.children.cycles-pp.flush_pending_writes
>       0.49 ±  6%      +2.3        2.81 ±  7%  perf-profile.children.cycles-pp.flush_bio_list
>       1.80 ±  3%      +5.6        7.44 ± 27%  perf-profile.children.cycles-pp._raw_spin_lock
>       2.12 ±  4%      +5.8        7.97 ± 20%  perf-profile.children.cycles-pp.remove_wait_queue
>       1.33 ±  4%      +8.8       10.12 ±  8%  perf-profile.children.cycles-pp.intel_idle
>       1.37 ±  4%      +9.3       10.71 ±  8%  perf-profile.children.cycles-pp.cpuidle_enter_state
>       1.59 ±  4%     +10.4       11.98 ±  9%  perf-profile.children.cycles-pp.start_secondary
>       1.63 ±  4%     +10.8       12.47 ±  8%  perf-profile.children.cycles-pp.secondary_startup_64
>       1.63 ±  4%     +10.8       12.47 ±  8%  perf-profile.children.cycles-pp.cpu_startup_entry
>       1.63 ±  4%     +10.9       12.49 ±  8%  perf-profile.children.cycles-pp.do_idle
>       3.48           +12.2       15.72 ± 23%  perf-profile.children.cycles-pp.__xfs_log_force_lsn
>       1.36 ± 12%     +57.8       59.12 ± 10%  perf-profile.children.cycles-pp.prepare_to_wait_event
>       0.43 ± 38%     +62.4       62.82 ±  8%  perf-profile.children.cycles-pp.xfs_submit_ioend
>       0.55 ± 29%     +62.5       63.10 ±  8%  perf-profile.children.cycles-pp.xfs_vm_writepages
>       0.55 ± 30%     +62.5       63.10 ±  8%  perf-profile.children.cycles-pp.do_writepages
>       0.55 ± 29%     +62.6       63.11 ±  8%  perf-profile.children.cycles-pp.__filemap_fdatawrite_range
>       0.66 ± 25%     +62.9       63.52 ±  7%  perf-profile.children.cycles-pp.file_write_and_wait_range
>       0.39 ± 43%     +63.6       64.02 ±  8%  perf-profile.children.cycles-pp.raid1_write_request
>       5.43 ±  3%     +64.2       69.64 ±  5%  perf-profile.children.cycles-pp._raw_spin_lock_irqsave
>      89.86           -13.5       76.31 ±  2%  perf-profile.self.cycles-pp.native_queued_spin_lock_slowpath
>       0.14 ±  8%      -0.0        0.09 ± 19%  perf-profile.self.cycles-pp.md_flush_request
>       0.10 ± 12%      -0.0        0.07 ± 21%  perf-profile.self.cycles-pp.account_entity_enqueue
>       0.06 ±  7%      +0.0        0.08 ± 12%  perf-profile.self.cycles-pp.pick_next_task_fair
>       0.05 ± 12%      +0.0        0.08 ± 18%  perf-profile.self.cycles-pp.___perf_sw_event
>       0.15 ±  6%      +0.0        0.18 ±  9%  perf-profile.self.cycles-pp.__update_load_avg_se
>       0.17 ±  4%      +0.0        0.22 ± 10%  perf-profile.self.cycles-pp.__schedule
>       0.10 ± 11%      +0.1        0.15 ± 11%  perf-profile.self.cycles-pp._raw_spin_lock
>       0.02 ±153%      +0.1        0.07 ± 16%  perf-profile.self.cycles-pp.delay_tsc
>       0.02 ±152%      +0.1        0.07 ± 23%  perf-profile.self.cycles-pp.set_next_entity
>       0.03 ±100%      +0.1        0.08 ± 15%  perf-profile.self.cycles-pp.find_next_bit
>       0.08 ±  5%      +0.1        0.14 ± 14%  perf-profile.self.cycles-pp.native_write_msr
>       0.01 ±200%      +0.1        0.07 ± 23%  perf-profile.self.cycles-pp.kmem_cache_alloc
>       0.29 ±  4%      +0.1        0.36 ±  8%  perf-profile.self.cycles-pp.__orc_find
>       0.14 ±  7%      +0.1        0.21 ± 12%  perf-profile.self.cycles-pp.switch_mm_irqs_off
>       0.00            +0.1        0.08 ± 11%  perf-profile.self.cycles-pp.clear_page_erms
>       0.00            +0.1        0.08 ± 28%  perf-profile.self.cycles-pp.__indirect_thunk_start
>       0.00            +0.1        0.08 ± 20%  perf-profile.self.cycles-pp.md_wakeup_thread
>       0.34 ±  6%      +0.1        0.43 ± 12%  perf-profile.self.cycles-pp._raw_spin_lock_irqsave
>       0.18 ±  4%      +0.1        0.27 ± 12%  perf-profile.self.cycles-pp.idle_cpu
>       0.16 ± 10%      +0.1        0.25 ± 13%  perf-profile.self.cycles-pp.__module_address
>       0.06 ± 11%      +0.1        0.17 ± 14%  perf-profile.self.cycles-pp._raw_spin_unlock_irqrestore
>       0.08 ± 38%      +0.1        0.20 ± 19%  perf-profile.self.cycles-pp.io_serial_in
>       0.18 ±  5%      +0.1        0.32 ± 15%  perf-profile.self.cycles-pp.update_load_avg
>       0.00            +0.1        0.15 ± 17%  perf-profile.self.cycles-pp.poll_idle
>       0.00            +0.2        0.15 ± 16%  perf-profile.self.cycles-pp.menu_select
>       0.00            +0.2        0.18 ± 24%  perf-profile.self.cycles-pp.find_busiest_group
>       0.02 ±152%      +0.3        0.35 ± 21%  perf-profile.self.cycles-pp.raid1_write_request
>       1.33 ±  4%      +8.8       10.12 ±  8%  perf-profile.self.cycles-pp.intel_idle
> 
> 
>                                                                                 
>                                  aim7.jobs-per-min                              
>                                                                                 
>   1700 +-+------------------------------------------------------------------+   
>        |+ ++++++ :+   ++++ ++++     +++     ++++++ +  +  ++++++++++++ ++  ++|   
>   1600 +-+       + +++    +    +++++   ++.++      + ++ ++            +  ++  |   
>        |                                                                    |   
>        |                                                                    |   
>   1500 +-+                                                                  |   
>        |                                                                    |   
>   1400 +-+                                                                  |   
>        |                                                                    |   
>   1300 +-+                                                                  |   
>        |                                                                    |   
>        O  OO         OO  O              O          O                        |   
>   1200 +OO  OOOOOOOOO  OO OOOOOOOOOOOOOO  OOOOOOOOO O                       |   
>        |                                                                    |   
>   1100 +-+------------------------------------------------------------------+   
>                                                                                 
>                                                                                                                                                                 
>                                                                                 
> [*] bisect-good sample
> [O] bisect-bad  sample
> 
> 
> Disclaimer:
> Results have been estimated based on internal Intel analysis and are provided
> for informational purposes only. Any difference in system hardware or software
> design or configuration may affect actual performance.
> 
> 
> Thanks,
> Xiaolong

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ