lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [thread-next>] [day] [month] [year] [list]
Message-ID: <20180604064203.GG16472@yexl-desktop>
Date:   Mon, 4 Jun 2018 14:42:03 +0800
From:   kernel test robot <xiaolong.ye@...el.com>
To:     Xiao Ni <xni@...hat.com>
Cc:     Shaohua Li <shli@...com>, Ming Lei <ming.lei@...hat.com>,
        LKML <linux-kernel@...r.kernel.org>,
        Stephen Rothwell <sfr@...b.auug.org.au>, lkp@...org
Subject: [lkp-robot] [MD]  5a409b4f56:  aim7.jobs-per-min -27.5% regression


Greeting,

FYI, we noticed a -27.5% regression of aim7.jobs-per-min due to commit:


commit: 5a409b4f56d50b212334f338cb8465d65550cd85 ("MD: fix lock contention for flush bios")
https://git.kernel.org/cgit/linux/kernel/git/next/linux-next.git master

in testcase: aim7
on test machine: 40 threads Intel(R) Xeon(R) CPU E5-2690 v2 @ 3.00GHz with 384G memory
with following parameters:

	disk: 4BRD_12G
	md: RAID1
	fs: xfs
	test: sync_disk_rw
	load: 600
	cpufreq_governor: performance

test-description: AIM7 is a traditional UNIX system level benchmark suite which is used to test and measure the performance of multiuser system.
test-url: https://sourceforge.net/projects/aimbench/files/aim-suite7/


Details are as below:
-------------------------------------------------------------------------------------------------->

=========================================================================================
compiler/cpufreq_governor/disk/fs/kconfig/load/md/rootfs/tbox_group/test/testcase:
  gcc-7/performance/4BRD_12G/xfs/x86_64-rhel-7.2/600/RAID1/debian-x86_64-2016-08-31.cgz/lkp-ivb-ep01/sync_disk_rw/aim7

commit: 
  448ec638c6 ("md/raid5: Assigning NULL to sh->batch_head before testing bit R5_Overlap of a stripe")
  5a409b4f56 ("MD: fix lock contention for flush bios")

448ec638c6bcf369 5a409b4f56d50b212334f338cb 
---------------- -------------------------- 
         %stddev     %change         %stddev
             \          |                \  
      1640           -27.5%       1189        aim7.jobs-per-min
      2194           +37.9%       3026        aim7.time.elapsed_time
      2194           +37.9%       3026        aim7.time.elapsed_time.max
  50990311           -95.8%    2148266        aim7.time.involuntary_context_switches
    107965 ±  4%     -26.4%      79516 ±  2%  aim7.time.minor_page_faults
     49.14           +82.5%      89.66 ±  2%  aim7.time.user_time
 7.123e+08           -35.7%  4.582e+08        aim7.time.voluntary_context_switches
    672282           +36.8%     919615        interrupts.CAL:Function_call_interrupts
  16631387 ±  2%     -39.9%    9993075 ±  7%  softirqs.RCU
   9708009          +186.1%   27778773        softirqs.SCHED
  33436649           +45.5%   48644912        softirqs.TIMER
      4.16            -2.1        2.01        mpstat.cpu.idle%
      0.24 ±  2%     +27.7       27.91        mpstat.cpu.iowait%
     95.51           -25.6       69.94        mpstat.cpu.sys%
      0.09            +0.0        0.13        mpstat.cpu.usr%
   6051756 ±  3%     +59.0%    9623085        numa-numastat.node0.local_node
   6055311 ±  3%     +59.0%    9626996        numa-numastat.node0.numa_hit
   6481209 ±  3%     +48.4%    9616310        numa-numastat.node1.local_node
   6485866 ±  3%     +48.3%    9620756        numa-numastat.node1.numa_hit
     61404           -27.7%      44424        vmstat.io.bo
      2.60 ± 18%  +11519.2%     302.10        vmstat.procs.b
    304.10           -84.9%      45.80 ±  2%  vmstat.procs.r
    400477           -43.5%     226094        vmstat.system.cs
    166461           -49.9%      83332        vmstat.system.in
     78397           +27.0%      99567        meminfo.Dirty
     14427           +18.4%      17082        meminfo.Inactive(anon)
      1963 ±  5%      +5.4%       2068 ±  4%  meminfo.Mlocked
    101143          +991.0%    1103488        meminfo.SUnreclaim
     53684 ±  4%     -18.1%      43946 ±  3%  meminfo.Shmem
    175580          +571.4%    1178829        meminfo.Slab
     39406           +26.2%      49717        numa-meminfo.node0.Dirty
   1767204 ± 10%     +37.2%    2425487 ±  2%  numa-meminfo.node0.MemUsed
     51634 ± 18%    +979.3%     557316        numa-meminfo.node0.SUnreclaim
     92259 ± 13%    +551.7%     601288        numa-meminfo.node0.Slab
     38969           +28.0%      49863        numa-meminfo.node1.Dirty
   1895204 ± 10%     +24.7%    2363037 ±  3%  numa-meminfo.node1.MemUsed
     49512 ± 19%   +1003.1%     546165        numa-meminfo.node1.SUnreclaim
     83323 ± 14%    +593.1%     577534        numa-meminfo.node1.Slab
 2.524e+09          +894.5%   2.51e+10        cpuidle.C1.time
  50620790          +316.5%  2.109e+08        cpuidle.C1.usage
 3.965e+08         +1871.1%  7.815e+09        cpuidle.C1E.time
   5987788          +186.1%   17129412        cpuidle.C1E.usage
 2.506e+08           +97.5%  4.948e+08 ±  2%  cpuidle.C3.time
   2923498           -55.7%    1295033        cpuidle.C3.usage
 5.327e+08          +179.9%  1.491e+09        cpuidle.C6.time
    779874 ±  2%    +229.3%    2567769        cpuidle.C6.usage
   6191357         +3333.6%  2.126e+08        cpuidle.POLL.time
    204095         +1982.1%    4249504        cpuidle.POLL.usage
      9850           +26.3%      12444        numa-vmstat.node0.nr_dirty
     12908 ± 18%    +979.3%     139321        numa-vmstat.node0.nr_slab_unreclaimable
      8876           +29.6%      11505        numa-vmstat.node0.nr_zone_write_pending
   3486319 ±  4%     +55.1%    5407021        numa-vmstat.node0.numa_hit
   3482713 ±  4%     +55.1%    5403066        numa-vmstat.node0.numa_local
      9743           +28.1%      12479        numa-vmstat.node1.nr_dirty
     12377 ± 19%   +1003.1%     136532        numa-vmstat.node1.nr_slab_unreclaimable
      9287           +30.0%      12074        numa-vmstat.node1.nr_zone_write_pending
   3678995 ±  4%     +44.8%    5326772        numa-vmstat.node1.numa_hit
   3497785 ±  4%     +47.1%    5145705        numa-vmstat.node1.numa_local
    252.70          +100.2%     505.90        slabinfo.biovec-max.active_objs
    282.70           +99.1%     562.90        slabinfo.biovec-max.num_objs
      2978 ± 17%     +52.5%       4543 ± 14%  slabinfo.dmaengine-unmap-16.active_objs
      2978 ± 17%     +52.5%       4543 ± 14%  slabinfo.dmaengine-unmap-16.num_objs
      2078          +147.9%       5153 ± 11%  slabinfo.ip6_dst_cache.active_objs
      2078          +148.1%       5157 ± 11%  slabinfo.ip6_dst_cache.num_objs
      5538 ±  2%     +26.2%       6990 ±  3%  slabinfo.kmalloc-1024.active_objs
      5586 ±  3%     +27.1%       7097 ±  3%  slabinfo.kmalloc-1024.num_objs
      6878           +47.6%      10151 ±  5%  slabinfo.kmalloc-192.active_objs
      6889           +47.5%      10160 ±  5%  slabinfo.kmalloc-192.num_objs
      9843 ±  5%  +1.6e+05%   16002876        slabinfo.kmalloc-64.active_objs
    161.90 ±  4%  +1.5e+05%     250044        slabinfo.kmalloc-64.active_slabs
     10386 ±  4%  +1.5e+05%   16002877        slabinfo.kmalloc-64.num_objs
    161.90 ±  4%  +1.5e+05%     250044        slabinfo.kmalloc-64.num_slabs
    432.80 ± 12%     +45.2%     628.50 ±  6%  slabinfo.nfs_read_data.active_objs
    432.80 ± 12%     +45.2%     628.50 ±  6%  slabinfo.nfs_read_data.num_objs
      3956           -23.1%       3041        slabinfo.pool_workqueue.active_objs
      4098           -19.8%       3286        slabinfo.pool_workqueue.num_objs
    360.50 ± 15%     +56.6%     564.70 ± 11%  slabinfo.secpath_cache.active_objs
    360.50 ± 15%     +56.6%     564.70 ± 11%  slabinfo.secpath_cache.num_objs
     35373 ±  2%      -8.3%      32432        proc-vmstat.nr_active_anon
     19595           +27.1%      24914        proc-vmstat.nr_dirty
      3607           +18.4%       4270        proc-vmstat.nr_inactive_anon
    490.30 ±  5%      +5.4%     516.90 ±  4%  proc-vmstat.nr_mlock
     13421 ±  4%     -18.1%      10986 ±  3%  proc-vmstat.nr_shmem
     18608            +1.2%      18834        proc-vmstat.nr_slab_reclaimable
     25286          +991.0%     275882        proc-vmstat.nr_slab_unreclaimable
     35405 ±  2%      -8.3%      32465        proc-vmstat.nr_zone_active_anon
      3607           +18.4%       4270        proc-vmstat.nr_zone_inactive_anon
     18161           +29.8%      23572        proc-vmstat.nr_zone_write_pending
     76941 ±  5%     -36.8%      48622 ±  4%  proc-vmstat.numa_hint_faults
     33878 ±  7%     -35.5%      21836 ±  5%  proc-vmstat.numa_hint_faults_local
  12568956           +53.3%   19272377        proc-vmstat.numa_hit
  12560739           +53.4%   19264015        proc-vmstat.numa_local
     17938 ±  3%     -33.5%      11935 ±  2%  proc-vmstat.numa_pages_migrated
     78296 ±  5%     -36.0%      50085 ±  4%  proc-vmstat.numa_pte_updates
      8848 ±  6%     -38.2%       5466 ±  6%  proc-vmstat.pgactivate
   8874568 ±  8%    +368.7%   41590920        proc-vmstat.pgalloc_normal
   5435965           +39.2%    7564148        proc-vmstat.pgfault
  12863707          +255.1%   45683570        proc-vmstat.pgfree
     17938 ±  3%     -33.5%      11935 ±  2%  proc-vmstat.pgmigrate_success
 1.379e+13           -40.8%   8.17e+12        perf-stat.branch-instructions
      0.30            +0.1        0.42        perf-stat.branch-miss-rate%
   4.2e+10           -17.6%  3.462e+10        perf-stat.branch-misses
     15.99            +3.8       19.74        perf-stat.cache-miss-rate%
 3.779e+10           -21.6%  2.963e+10        perf-stat.cache-misses
 2.364e+11           -36.5%  1.501e+11        perf-stat.cache-references
 8.795e+08           -22.2%   6.84e+08        perf-stat.context-switches
      4.44            -7.2%       4.12        perf-stat.cpi
 2.508e+14           -44.5%  1.393e+14        perf-stat.cpu-cycles
  36915392           +60.4%   59211221        perf-stat.cpu-migrations
      0.29 ±  2%      +0.0        0.34 ±  4%  perf-stat.dTLB-load-miss-rate%
  4.14e+10           -30.2%   2.89e+10 ±  4%  perf-stat.dTLB-load-misses
 1.417e+13           -40.1%  8.491e+12        perf-stat.dTLB-loads
      0.20 ±  4%      -0.0        0.18 ±  5%  perf-stat.dTLB-store-miss-rate%
 3.072e+09 ±  4%     -28.0%   2.21e+09 ±  4%  perf-stat.dTLB-store-misses
 1.535e+12           -20.2%  1.225e+12        perf-stat.dTLB-stores
     90.73           -11.7       79.07        perf-stat.iTLB-load-miss-rate%
 8.291e+09            -6.6%  7.743e+09        perf-stat.iTLB-load-misses
 8.473e+08          +141.8%  2.049e+09 ±  3%  perf-stat.iTLB-loads
 5.646e+13           -40.2%  3.378e+13        perf-stat.instructions
      6810           -35.9%       4362        perf-stat.instructions-per-iTLB-miss
      0.23            +7.8%       0.24        perf-stat.ipc
   5326672           +39.2%    7413706        perf-stat.minor-faults
 1.873e+10           -29.9%  1.312e+10        perf-stat.node-load-misses
 2.093e+10           -29.2%  1.481e+10        perf-stat.node-loads
     39.38            -0.7       38.72        perf-stat.node-store-miss-rate%
 1.087e+10           -16.6%  9.069e+09        perf-stat.node-store-misses
 1.673e+10           -14.2%  1.435e+10        perf-stat.node-stores
   5326695           +39.2%    7413708        perf-stat.page-faults
   1875095 ±  7%     -54.8%     846645 ± 16%  sched_debug.cfs_rq:/.MIN_vruntime.avg
  32868920 ±  6%     -35.7%   21150379 ± 14%  sched_debug.cfs_rq:/.MIN_vruntime.max
   7267340 ±  5%     -44.7%    4015798 ± 14%  sched_debug.cfs_rq:/.MIN_vruntime.stddev
      4278 ±  7%     -54.7%       1939 ± 11%  sched_debug.cfs_rq:/.exec_clock.stddev
    245.48 ±  2%     +65.3%     405.75 ±  7%  sched_debug.cfs_rq:/.load_avg.avg
      2692 ±  6%    +126.0%       6087 ±  7%  sched_debug.cfs_rq:/.load_avg.max
     33.09           -73.0%       8.94 ±  7%  sched_debug.cfs_rq:/.load_avg.min
    507.40 ±  4%    +128.0%       1156 ±  7%  sched_debug.cfs_rq:/.load_avg.stddev
   1875095 ±  7%     -54.8%     846645 ± 16%  sched_debug.cfs_rq:/.max_vruntime.avg
  32868921 ±  6%     -35.7%   21150379 ± 14%  sched_debug.cfs_rq:/.max_vruntime.max
   7267341 ±  5%     -44.7%    4015798 ± 14%  sched_debug.cfs_rq:/.max_vruntime.stddev
  35887197           -13.2%   31149130        sched_debug.cfs_rq:/.min_vruntime.avg
  37385506           -14.3%   32043914        sched_debug.cfs_rq:/.min_vruntime.max
  34416296           -12.3%   30183927        sched_debug.cfs_rq:/.min_vruntime.min
   1228844 ±  8%     -52.6%     582759 ±  4%  sched_debug.cfs_rq:/.min_vruntime.stddev
      0.83           -28.1%       0.60 ±  6%  sched_debug.cfs_rq:/.nr_running.avg
      2.07 ±  3%     -24.6%       1.56 ±  8%  sched_debug.cfs_rq:/.nr_running.max
     20.52 ±  4%     -48.8%      10.52 ±  3%  sched_debug.cfs_rq:/.nr_spread_over.avg
     35.96 ±  5%     -42.2%      20.77 ±  9%  sched_debug.cfs_rq:/.nr_spread_over.max
      8.97 ± 11%     -44.5%       4.98 ±  8%  sched_debug.cfs_rq:/.nr_spread_over.min
      6.40 ± 12%     -45.5%       3.49 ±  7%  sched_debug.cfs_rq:/.nr_spread_over.stddev
     21.78 ±  7%    +143.3%      53.00 ±  9%  sched_debug.cfs_rq:/.runnable_load_avg.avg
    328.86 ± 18%    +303.4%       1326 ± 14%  sched_debug.cfs_rq:/.runnable_load_avg.max
     55.97 ± 17%    +286.0%     216.07 ± 13%  sched_debug.cfs_rq:/.runnable_load_avg.stddev
      0.10 ± 29%     -82.4%       0.02 ± 50%  sched_debug.cfs_rq:/.spread.avg
      3.43 ± 25%     -79.9%       0.69 ± 50%  sched_debug.cfs_rq:/.spread.max
      0.56 ± 26%     -80.7%       0.11 ± 50%  sched_debug.cfs_rq:/.spread.stddev
   1228822 ±  8%     -52.6%     582732 ±  4%  sched_debug.cfs_rq:/.spread0.stddev
    992.30           -24.9%     745.56 ±  2%  sched_debug.cfs_rq:/.util_avg.avg
      1485           -18.1%       1217 ±  2%  sched_debug.cfs_rq:/.util_avg.max
    515.45 ±  2%     -25.2%     385.73 ±  6%  sched_debug.cfs_rq:/.util_avg.min
    201.54           -14.9%     171.52 ±  3%  sched_debug.cfs_rq:/.util_avg.stddev
    248.73 ±  6%     -38.1%     154.02 ±  8%  sched_debug.cfs_rq:/.util_est_enqueued.avg
    222.78 ±  3%     -15.8%     187.58 ±  2%  sched_debug.cfs_rq:/.util_est_enqueued.stddev
     77097 ±  4%    +278.4%     291767 ± 11%  sched_debug.cpu.avg_idle.avg
    181319 ±  6%    +298.7%     722862 ±  3%  sched_debug.cpu.avg_idle.max
     19338          +392.3%      95203 ± 17%  sched_debug.cpu.avg_idle.min
     34877 ±  6%    +303.5%     140732 ±  6%  sched_debug.cpu.avg_idle.stddev
   1107408           +37.6%    1523823        sched_debug.cpu.clock.avg
   1107427           +37.6%    1523834        sched_debug.cpu.clock.max
   1107385           +37.6%    1523811        sched_debug.cpu.clock.min
     13.10 ±  9%     -48.1%       6.80 ±  8%  sched_debug.cpu.clock.stddev
   1107408           +37.6%    1523823        sched_debug.cpu.clock_task.avg
   1107427           +37.6%    1523834        sched_debug.cpu.clock_task.max
   1107385           +37.6%    1523811        sched_debug.cpu.clock_task.min
     13.10 ±  9%     -48.1%       6.80 ±  8%  sched_debug.cpu.clock_task.stddev
     30.36 ±  7%    +107.7%      63.06 ± 12%  sched_debug.cpu.cpu_load[0].avg
    381.48 ± 18%    +269.8%       1410 ± 18%  sched_debug.cpu.cpu_load[0].max
     63.92 ± 18%    +262.2%     231.50 ± 17%  sched_debug.cpu.cpu_load[0].stddev
     31.34 ±  5%    +118.4%      68.44 ±  9%  sched_debug.cpu.cpu_load[1].avg
    323.62 ± 17%    +349.5%       1454 ± 14%  sched_debug.cpu.cpu_load[1].max
     53.23 ± 16%    +350.3%     239.71 ± 13%  sched_debug.cpu.cpu_load[1].stddev
     32.15 ±  3%    +129.4%      73.74 ±  6%  sched_debug.cpu.cpu_load[2].avg
    285.20 ± 14%    +420.8%       1485 ±  9%  sched_debug.cpu.cpu_load[2].max
     46.66 ± 12%    +430.0%     247.32 ±  8%  sched_debug.cpu.cpu_load[2].stddev
     33.02 ±  2%    +133.2%      77.00 ±  3%  sched_debug.cpu.cpu_load[3].avg
    252.16 ± 10%    +481.2%       1465 ±  7%  sched_debug.cpu.cpu_load[3].max
     40.74 ±  8%    +503.2%     245.72 ±  6%  sched_debug.cpu.cpu_load[3].stddev
     33.86          +131.5%      78.38 ±  2%  sched_debug.cpu.cpu_load[4].avg
    219.81 ±  8%    +522.6%       1368 ±  5%  sched_debug.cpu.cpu_load[4].max
     35.45 ±  7%    +554.2%     231.90 ±  4%  sched_debug.cpu.cpu_load[4].stddev
      2600 ±  4%     -30.5%       1807 ±  4%  sched_debug.cpu.curr->pid.avg
     25309 ±  4%     -19.5%      20367 ±  4%  sched_debug.cpu.curr->pid.max
      4534 ±  7%     -21.2%       3573 ±  5%  sched_debug.cpu.curr->pid.stddev
      0.00 ±  2%     -27.6%       0.00 ±  6%  sched_debug.cpu.next_balance.stddev
   1083917           +38.6%    1502777        sched_debug.cpu.nr_load_updates.avg
   1088142           +38.6%    1508302        sched_debug.cpu.nr_load_updates.max
   1082048           +38.7%    1501073        sched_debug.cpu.nr_load_updates.min
      3.53 ±  6%     -73.0%       0.95 ±  6%  sched_debug.cpu.nr_running.avg
     11.54 ±  3%     -62.1%       4.37 ± 10%  sched_debug.cpu.nr_running.max
      3.10 ±  3%     -66.8%       1.03 ±  9%  sched_debug.cpu.nr_running.stddev
  10764176           -22.4%    8355047        sched_debug.cpu.nr_switches.avg
  10976436           -22.2%    8545010        sched_debug.cpu.nr_switches.max
  10547712           -22.8%    8143037        sched_debug.cpu.nr_switches.min
    148628 ±  3%     -22.7%     114880 ±  7%  sched_debug.cpu.nr_switches.stddev
     11.13 ±  2%     +24.5%      13.85        sched_debug.cpu.nr_uninterruptible.avg
      6420 ±  8%     -48.7%       3296 ± 11%  sched_debug.cpu.nr_uninterruptible.max
     -5500           -37.2%      -3455        sched_debug.cpu.nr_uninterruptible.min
      3784 ±  6%     -47.2%       1997 ±  4%  sched_debug.cpu.nr_uninterruptible.stddev
  10812670           -22.7%    8356821        sched_debug.cpu.sched_count.avg
  11020646           -22.5%    8546277        sched_debug.cpu.sched_count.max
  10601390           -23.2%    8144743        sched_debug.cpu.sched_count.min
    144529 ±  3%     -20.9%     114359 ±  7%  sched_debug.cpu.sched_count.stddev
    706116          +259.0%    2534721        sched_debug.cpu.sched_goidle.avg
    771307          +232.4%    2564059        sched_debug.cpu.sched_goidle.max
    644658          +286.9%    2494236        sched_debug.cpu.sched_goidle.min
     49847 ±  6%     -67.9%      15979 ±  7%  sched_debug.cpu.sched_goidle.stddev
   9618827           -39.9%    5780369        sched_debug.cpu.ttwu_count.avg
   8990451           -61.7%    3441265 ±  4%  sched_debug.cpu.ttwu_count.min
    418563 ± 25%    +244.2%    1440565 ±  7%  sched_debug.cpu.ttwu_count.stddev
    640964           -93.7%      40366 ±  2%  sched_debug.cpu.ttwu_local.avg
    679527           -92.1%      53476 ±  4%  sched_debug.cpu.ttwu_local.max
    601661           -94.9%      30636 ±  3%  sched_debug.cpu.ttwu_local.min
     24242 ± 21%     -77.7%       5405 ±  9%  sched_debug.cpu.ttwu_local.stddev
   1107383           +37.6%    1523810        sched_debug.cpu_clk
   1107383           +37.6%    1523810        sched_debug.ktime
      0.00           -49.4%       0.00 ± 65%  sched_debug.rt_rq:/.rt_nr_migratory.avg
      0.03           -49.4%       0.01 ± 65%  sched_debug.rt_rq:/.rt_nr_migratory.max
      0.00           -49.4%       0.00 ± 65%  sched_debug.rt_rq:/.rt_nr_migratory.stddev
      0.00           -49.4%       0.00 ± 65%  sched_debug.rt_rq:/.rt_nr_running.avg
      0.03           -49.4%       0.01 ± 65%  sched_debug.rt_rq:/.rt_nr_running.max
      0.00           -49.4%       0.00 ± 65%  sched_debug.rt_rq:/.rt_nr_running.stddev
      0.01 ±  8%     +79.9%       0.01 ± 23%  sched_debug.rt_rq:/.rt_time.avg
   1107805           +37.6%    1524235        sched_debug.sched_clk
     87.59           -87.6        0.00        perf-profile.calltrace.cycles-pp.md_flush_request.raid1_make_request.md_handle_request.md_make_request.generic_make_request
     87.57           -87.6        0.00        perf-profile.calltrace.cycles-pp.submit_bio_wait.blkdev_issue_flush.xfs_file_fsync.xfs_file_write_iter.__vfs_write
     87.59           -87.5        0.05 ±299%  perf-profile.calltrace.cycles-pp.blkdev_issue_flush.xfs_file_fsync.xfs_file_write_iter.__vfs_write.vfs_write
     87.51           -87.5        0.00        perf-profile.calltrace.cycles-pp.generic_make_request.submit_bio.submit_bio_wait.blkdev_issue_flush.xfs_file_fsync
     87.51           -87.5        0.00        perf-profile.calltrace.cycles-pp.submit_bio.submit_bio_wait.blkdev_issue_flush.xfs_file_fsync.xfs_file_write_iter
     87.50           -87.5        0.00        perf-profile.calltrace.cycles-pp.md_make_request.generic_make_request.submit_bio.submit_bio_wait.blkdev_issue_flush
     87.50           -87.5        0.00        perf-profile.calltrace.cycles-pp.md_handle_request.md_make_request.generic_make_request.submit_bio.submit_bio_wait
     82.37           -82.4        0.00        perf-profile.calltrace.cycles-pp._raw_spin_lock_irq.md_flush_request.raid1_make_request.md_handle_request.md_make_request
     82.23           -82.2        0.00        perf-profile.calltrace.cycles-pp.native_queued_spin_lock_slowpath._raw_spin_lock_irq.md_flush_request.raid1_make_request.md_handle_request
     87.79           -25.0       62.75 ±  8%  perf-profile.calltrace.cycles-pp.raid1_make_request.md_handle_request.md_make_request.generic_make_request.submit_bio
     92.78           -13.0       79.76        perf-profile.calltrace.cycles-pp.xfs_file_fsync.xfs_file_write_iter.__vfs_write.vfs_write.ksys_write
     93.08           -12.6       80.49        perf-profile.calltrace.cycles-pp.xfs_file_write_iter.__vfs_write.vfs_write.ksys_write.do_syscall_64
     93.08           -12.6       80.50        perf-profile.calltrace.cycles-pp.__vfs_write.vfs_write.ksys_write.do_syscall_64.entry_SYSCALL_64_after_hwframe
     93.11           -12.6       80.56        perf-profile.calltrace.cycles-pp.vfs_write.ksys_write.do_syscall_64.entry_SYSCALL_64_after_hwframe
     93.11           -12.6       80.56        perf-profile.calltrace.cycles-pp.ksys_write.do_syscall_64.entry_SYSCALL_64_after_hwframe
     93.14           -12.5       80.64        perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe
     93.15           -12.5       80.65        perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe
      3.40 ±  2%      -1.4        1.97 ±  8%  perf-profile.calltrace.cycles-pp.worker_thread.kthread.ret_from_fork
      3.33 ±  2%      -1.4        1.96 ±  9%  perf-profile.calltrace.cycles-pp.process_one_work.worker_thread.kthread.ret_from_fork
      1.12 ±  2%      -0.7        0.42 ± 68%  perf-profile.calltrace.cycles-pp.__save_stack_trace.save_stack_trace_tsk.__account_scheduler_latency.enqueue_entity.enqueue_task_fair
      1.16 ±  2%      -0.6        0.60 ± 17%  perf-profile.calltrace.cycles-pp.save_stack_trace_tsk.__account_scheduler_latency.enqueue_entity.enqueue_task_fair.ttwu_do_activate
      0.00            +0.6        0.59 ± 15%  perf-profile.calltrace.cycles-pp._raw_spin_lock_irqsave.__wake_up_common_lock.raid1_write_request.raid1_make_request.md_handle_request
      0.00            +0.6        0.64 ± 15%  perf-profile.calltrace.cycles-pp.__wake_up_common_lock.raid1_write_request.raid1_make_request.md_handle_request.md_make_request
      0.00            +0.7        0.65 ± 10%  perf-profile.calltrace.cycles-pp.enqueue_entity.enqueue_task_fair.ttwu_do_activate.sched_ttwu_pending.do_idle
      0.00            +0.7        0.68 ± 10%  perf-profile.calltrace.cycles-pp.enqueue_task_fair.ttwu_do_activate.sched_ttwu_pending.do_idle.cpu_startup_entry
      0.00            +0.7        0.69 ± 10%  perf-profile.calltrace.cycles-pp.ttwu_do_activate.sched_ttwu_pending.do_idle.cpu_startup_entry.start_secondary
      0.00            +0.8        0.79 ± 11%  perf-profile.calltrace.cycles-pp.sched_ttwu_pending.do_idle.cpu_startup_entry.start_secondary.secondary_startup_64
      0.00            +0.8        0.83 ±  7%  perf-profile.calltrace.cycles-pp.__schedule.schedule.raid1_write_request.raid1_make_request.md_handle_request
      0.62 ±  3%      +0.8        1.45 ± 22%  perf-profile.calltrace.cycles-pp.native_queued_spin_lock_slowpath._raw_spin_lock_irqsave.remove_wait_queue.xlog_wait.__xfs_log_force_lsn
      0.00            +0.8        0.83 ±  7%  perf-profile.calltrace.cycles-pp.schedule.raid1_write_request.raid1_make_request.md_handle_request.md_make_request
      0.63 ±  2%      +0.8        1.46 ± 22%  perf-profile.calltrace.cycles-pp.remove_wait_queue.xlog_wait.__xfs_log_force_lsn.xfs_log_force_lsn.xfs_file_fsync
      0.62 ±  2%      +0.8        1.46 ± 22%  perf-profile.calltrace.cycles-pp._raw_spin_lock_irqsave.remove_wait_queue.xlog_wait.__xfs_log_force_lsn.xfs_log_force_lsn
      3.92 ±  2%      +0.9        4.79 ±  6%  perf-profile.calltrace.cycles-pp.ret_from_fork
      3.92 ±  2%      +0.9        4.79 ±  6%  perf-profile.calltrace.cycles-pp.kthread.ret_from_fork
      0.69 ±  2%      +0.9        1.64 ± 23%  perf-profile.calltrace.cycles-pp.xlog_wait.__xfs_log_force_lsn.xfs_log_force_lsn.xfs_file_fsync.xfs_file_write_iter
      0.00            +1.2        1.17 ±  8%  perf-profile.calltrace.cycles-pp._raw_spin_unlock_irqrestore.prepare_to_wait_event.raid1_write_request.raid1_make_request.md_handle_request
      0.00            +1.2        1.23 ± 18%  perf-profile.calltrace.cycles-pp.prepare_to_wait_event.raid1_write_request.raid1_make_request.md_handle_request.submit_flushes
      0.00            +1.3        1.27 ± 17%  perf-profile.calltrace.cycles-pp.raid1_write_request.raid1_make_request.md_handle_request.submit_flushes.process_one_work
      0.00            +1.3        1.27 ± 17%  perf-profile.calltrace.cycles-pp.md_handle_request.submit_flushes.process_one_work.worker_thread.kthread
      0.00            +1.3        1.27 ± 17%  perf-profile.calltrace.cycles-pp.raid1_make_request.md_handle_request.submit_flushes.process_one_work.worker_thread
      0.00            +1.3        1.27 ± 17%  perf-profile.calltrace.cycles-pp.submit_flushes.process_one_work.worker_thread.kthread.ret_from_fork
      0.00            +1.6        1.65 ± 14%  perf-profile.calltrace.cycles-pp.try_to_wake_up.autoremove_wake_function.__wake_up_common.__wake_up_common_lock.raid_end_bio_io
      0.00            +1.7        1.71 ± 14%  perf-profile.calltrace.cycles-pp.autoremove_wake_function.__wake_up_common.__wake_up_common_lock.raid_end_bio_io.raid1_end_write_request
      0.00            +1.7        1.71 ± 14%  perf-profile.calltrace.cycles-pp.__wake_up_common.__wake_up_common_lock.raid_end_bio_io.raid1_end_write_request.brd_make_request
      0.00            +1.9        1.86 ± 13%  perf-profile.calltrace.cycles-pp.__wake_up_common_lock.raid_end_bio_io.raid1_end_write_request.brd_make_request.generic_make_request
      0.00            +2.1        2.10 ± 10%  perf-profile.calltrace.cycles-pp.native_queued_spin_lock_slowpath._raw_spin_lock_irqsave.remove_wait_queue.__xfs_log_force_lsn.xfs_log_force_lsn
      0.00            +2.1        2.10 ± 10%  perf-profile.calltrace.cycles-pp._raw_spin_lock_irqsave.remove_wait_queue.__xfs_log_force_lsn.xfs_log_force_lsn.xfs_file_fsync
      0.00            +2.1        2.11 ± 10%  perf-profile.calltrace.cycles-pp.remove_wait_queue.__xfs_log_force_lsn.xfs_log_force_lsn.xfs_file_fsync.xfs_file_write_iter
      0.00            +2.2        2.16 ± 10%  perf-profile.calltrace.cycles-pp.raid_end_bio_io.raid1_end_write_request.brd_make_request.generic_make_request.flush_bio_list
      2.24 ±  4%      +2.2        4.44 ± 15%  perf-profile.calltrace.cycles-pp.xfs_log_force_lsn.xfs_file_fsync.xfs_file_write_iter.__vfs_write.vfs_write
      0.00            +2.3        2.25 ± 10%  perf-profile.calltrace.cycles-pp.raid1_end_write_request.brd_make_request.generic_make_request.flush_bio_list.flush_pending_writes
      0.00            +2.3        2.30 ± 20%  perf-profile.calltrace.cycles-pp.native_queued_spin_lock_slowpath._raw_spin_lock_irqsave.raid1_write_request.raid1_make_request.md_handle_request
      0.00            +2.4        2.35 ± 20%  perf-profile.calltrace.cycles-pp._raw_spin_lock_irqsave.raid1_write_request.raid1_make_request.md_handle_request.md_make_request
      0.37 ± 65%      +2.4        2.81 ±  7%  perf-profile.calltrace.cycles-pp.md_thread.kthread.ret_from_fork
      0.26 ±100%      +2.5        2.81 ±  7%  perf-profile.calltrace.cycles-pp.raid1d.md_thread.kthread.ret_from_fork
      0.26 ±100%      +2.5        2.81 ±  7%  perf-profile.calltrace.cycles-pp.flush_pending_writes.raid1d.md_thread.kthread.ret_from_fork
      0.26 ±100%      +2.6        2.81 ±  7%  perf-profile.calltrace.cycles-pp.flush_bio_list.flush_pending_writes.raid1d.md_thread.kthread
      0.10 ±200%      +2.7        2.76 ±  7%  perf-profile.calltrace.cycles-pp.generic_make_request.flush_bio_list.flush_pending_writes.raid1d.md_thread
      0.00            +2.7        2.73 ±  7%  perf-profile.calltrace.cycles-pp.brd_make_request.generic_make_request.flush_bio_list.flush_pending_writes.raid1d
      1.20 ±  3%      +3.1        4.35 ± 15%  perf-profile.calltrace.cycles-pp.__xfs_log_force_lsn.xfs_log_force_lsn.xfs_file_fsync.xfs_file_write_iter.__vfs_write
      0.63 ±  6%      +3.8        4.38 ± 27%  perf-profile.calltrace.cycles-pp.native_queued_spin_lock_slowpath._raw_spin_lock_irqsave.remove_wait_queue.__xfs_log_force_lsn.xfs_file_fsync
      0.63 ±  5%      +3.8        4.39 ± 27%  perf-profile.calltrace.cycles-pp._raw_spin_lock_irqsave.remove_wait_queue.__xfs_log_force_lsn.xfs_file_fsync.xfs_file_write_iter
      0.63 ±  5%      +3.8        4.40 ± 27%  perf-profile.calltrace.cycles-pp.remove_wait_queue.__xfs_log_force_lsn.xfs_file_fsync.xfs_file_write_iter.__vfs_write
      1.26 ±  5%      +5.3        6.55 ± 27%  perf-profile.calltrace.cycles-pp.native_queued_spin_lock_slowpath._raw_spin_lock.__xfs_log_force_lsn.xfs_file_fsync.xfs_file_write_iter
      1.27 ±  5%      +5.3        6.55 ± 27%  perf-profile.calltrace.cycles-pp._raw_spin_lock.__xfs_log_force_lsn.xfs_file_fsync.xfs_file_write_iter.__vfs_write
      1.30 ±  4%      +8.4        9.72 ±  9%  perf-profile.calltrace.cycles-pp.intel_idle.cpuidle_enter_state.do_idle.cpu_startup_entry.start_secondary
      1.33 ±  4%      +8.9       10.26 ±  9%  perf-profile.calltrace.cycles-pp.cpuidle_enter_state.do_idle.cpu_startup_entry.start_secondary.secondary_startup_64
      2.28 ±  2%      +9.1       11.36 ± 27%  perf-profile.calltrace.cycles-pp.__xfs_log_force_lsn.xfs_file_fsync.xfs_file_write_iter.__vfs_write.vfs_write
      1.59 ±  4%     +10.4       11.97 ±  9%  perf-profile.calltrace.cycles-pp.do_idle.cpu_startup_entry.start_secondary.secondary_startup_64
      1.59 ±  4%     +10.4       11.98 ±  9%  perf-profile.calltrace.cycles-pp.cpu_startup_entry.start_secondary.secondary_startup_64
      1.59 ±  4%     +10.4       11.98 ±  9%  perf-profile.calltrace.cycles-pp.start_secondary.secondary_startup_64
      1.63 ±  4%     +10.8       12.47 ±  8%  perf-profile.calltrace.cycles-pp.secondary_startup_64
      0.00           +57.7       57.66 ± 10%  perf-profile.calltrace.cycles-pp.native_queued_spin_lock_slowpath._raw_spin_lock_irqsave.prepare_to_wait_event.raid1_write_request.raid1_make_request
      0.00           +57.7       57.73 ± 10%  perf-profile.calltrace.cycles-pp._raw_spin_lock_irqsave.prepare_to_wait_event.raid1_write_request.raid1_make_request.md_handle_request
      0.05 ±299%     +57.8       57.85 ±  9%  perf-profile.calltrace.cycles-pp.prepare_to_wait_event.raid1_write_request.raid1_make_request.md_handle_request.md_make_request
      0.19 ±154%     +62.5       62.73 ±  8%  perf-profile.calltrace.cycles-pp.raid1_write_request.raid1_make_request.md_handle_request.md_make_request.generic_make_request
      0.19 ±154%     +62.6       62.76 ±  8%  perf-profile.calltrace.cycles-pp.md_handle_request.md_make_request.generic_make_request.submit_bio.xfs_submit_ioend
      0.19 ±154%     +62.6       62.79 ±  8%  perf-profile.calltrace.cycles-pp.md_make_request.generic_make_request.submit_bio.xfs_submit_ioend.xfs_vm_writepages
      0.20 ±154%     +62.6       62.81 ±  8%  perf-profile.calltrace.cycles-pp.generic_make_request.submit_bio.xfs_submit_ioend.xfs_vm_writepages.do_writepages
      0.20 ±154%     +62.6       62.81 ±  8%  perf-profile.calltrace.cycles-pp.submit_bio.xfs_submit_ioend.xfs_vm_writepages.do_writepages.__filemap_fdatawrite_range
      0.20 ±154%     +62.6       62.82 ±  8%  perf-profile.calltrace.cycles-pp.xfs_submit_ioend.xfs_vm_writepages.do_writepages.__filemap_fdatawrite_range.file_write_and_wait_range
      0.29 ±125%     +62.8       63.09 ±  8%  perf-profile.calltrace.cycles-pp.xfs_vm_writepages.do_writepages.__filemap_fdatawrite_range.file_write_and_wait_range.xfs_file_fsync
      0.29 ±126%     +62.8       63.10 ±  8%  perf-profile.calltrace.cycles-pp.do_writepages.__filemap_fdatawrite_range.file_write_and_wait_range.xfs_file_fsync.xfs_file_write_iter
      0.29 ±125%     +62.8       63.11 ±  8%  perf-profile.calltrace.cycles-pp.__filemap_fdatawrite_range.file_write_and_wait_range.xfs_file_fsync.xfs_file_write_iter.__vfs_write
      0.62 ± 41%     +62.9       63.52 ±  7%  perf-profile.calltrace.cycles-pp.file_write_and_wait_range.xfs_file_fsync.xfs_file_write_iter.__vfs_write.vfs_write
     88.51           -88.2        0.26 ± 19%  perf-profile.children.cycles-pp.md_flush_request
     87.57           -87.2        0.35 ± 19%  perf-profile.children.cycles-pp.submit_bio_wait
     87.59           -87.2        0.39 ± 19%  perf-profile.children.cycles-pp.blkdev_issue_flush
     83.26           -83.2        0.02 ±123%  perf-profile.children.cycles-pp._raw_spin_lock_irq
     88.85           -25.7       63.11 ±  8%  perf-profile.children.cycles-pp.md_make_request
     88.90           -25.7       63.17 ±  8%  perf-profile.children.cycles-pp.submit_bio
     88.83           -24.5       64.31 ±  8%  perf-profile.children.cycles-pp.raid1_make_request
     88.84           -24.5       64.33 ±  8%  perf-profile.children.cycles-pp.md_handle_request
     89.38           -23.5       65.92 ±  7%  perf-profile.children.cycles-pp.generic_make_request
     89.90           -13.4       76.51 ±  2%  perf-profile.children.cycles-pp.native_queued_spin_lock_slowpath
     92.79           -13.0       79.76        perf-profile.children.cycles-pp.xfs_file_fsync
     93.08           -12.6       80.49        perf-profile.children.cycles-pp.xfs_file_write_iter
     93.09           -12.6       80.54        perf-profile.children.cycles-pp.__vfs_write
     93.13           -12.5       80.60        perf-profile.children.cycles-pp.vfs_write
     93.13           -12.5       80.61        perf-profile.children.cycles-pp.ksys_write
     93.22           -12.4       80.83        perf-profile.children.cycles-pp.do_syscall_64
     93.22           -12.4       80.83        perf-profile.children.cycles-pp.entry_SYSCALL_64_after_hwframe
      3.40 ±  2%      -1.4        1.97 ±  8%  perf-profile.children.cycles-pp.worker_thread
      3.33 ±  2%      -1.4        1.96 ±  9%  perf-profile.children.cycles-pp.process_one_work
      1.03 ±  7%      -1.0        0.07 ± 37%  perf-profile.children.cycles-pp.xlog_cil_force_lsn
      1.69 ±  2%      -0.7        0.96 ±  4%  perf-profile.children.cycles-pp.reschedule_interrupt
      1.66 ±  2%      -0.7        0.94 ±  4%  perf-profile.children.cycles-pp.scheduler_ipi
      1.13 ±  2%      -0.7        0.47 ± 11%  perf-profile.children.cycles-pp.finish_wait
      0.54 ±  8%      -0.4        0.10 ± 38%  perf-profile.children.cycles-pp.xlog_cil_push
      0.49 ±  9%      -0.4        0.09 ± 35%  perf-profile.children.cycles-pp.xlog_write
      0.10 ±  8%      -0.1        0.04 ± 67%  perf-profile.children.cycles-pp.flush_work
      0.20 ±  5%      -0.0        0.16 ± 11%  perf-profile.children.cycles-pp.reweight_entity
      0.06 ± 10%      +0.0        0.10 ± 23%  perf-profile.children.cycles-pp.brd_lookup_page
      0.18 ±  5%      +0.0        0.23 ± 13%  perf-profile.children.cycles-pp.__update_load_avg_se
      0.02 ±153%      +0.1        0.07 ± 16%  perf-profile.children.cycles-pp.delay_tsc
      0.03 ±100%      +0.1        0.08 ± 15%  perf-profile.children.cycles-pp.find_next_bit
      0.08 ±  5%      +0.1        0.14 ± 14%  perf-profile.children.cycles-pp.native_write_msr
      0.29 ±  4%      +0.1        0.36 ±  8%  perf-profile.children.cycles-pp.__orc_find
      0.40 ±  4%      +0.1        0.46 ±  7%  perf-profile.children.cycles-pp.dequeue_task_fair
      0.11 ± 11%      +0.1        0.18 ± 14%  perf-profile.children.cycles-pp.__module_text_address
      0.12 ±  8%      +0.1        0.19 ± 13%  perf-profile.children.cycles-pp.is_module_text_address
      0.04 ± 50%      +0.1        0.12 ± 19%  perf-profile.children.cycles-pp.kmem_cache_alloc
      0.00            +0.1        0.08 ± 11%  perf-profile.children.cycles-pp.clear_page_erms
      0.00            +0.1        0.08 ± 28%  perf-profile.children.cycles-pp.__indirect_thunk_start
      0.01 ±200%      +0.1        0.10 ± 25%  perf-profile.children.cycles-pp.xfs_trans_alloc
      0.00            +0.1        0.09 ± 18%  perf-profile.children.cycles-pp.md_wakeup_thread
      0.00            +0.1        0.09 ± 26%  perf-profile.children.cycles-pp.rebalance_domains
      0.00            +0.1        0.09 ± 26%  perf-profile.children.cycles-pp.get_next_timer_interrupt
      0.00            +0.1        0.09 ± 20%  perf-profile.children.cycles-pp.ktime_get
      0.18 ±  4%      +0.1        0.27 ± 12%  perf-profile.children.cycles-pp.idle_cpu
      0.20 ±  6%      +0.1        0.30 ±  9%  perf-profile.children.cycles-pp.unwind_get_return_address
      0.16 ± 10%      +0.1        0.25 ± 13%  perf-profile.children.cycles-pp.__module_address
      0.03 ±100%      +0.1        0.13 ±  8%  perf-profile.children.cycles-pp.brd_insert_page
      0.06 ±  9%      +0.1        0.16 ± 14%  perf-profile.children.cycles-pp.task_tick_fair
      0.08 ± 12%      +0.1        0.18 ± 24%  perf-profile.children.cycles-pp.bio_alloc_bioset
      0.03 ± 81%      +0.1        0.14 ± 27%  perf-profile.children.cycles-pp.generic_make_request_checks
      0.17 ±  7%      +0.1        0.28 ± 11%  perf-profile.children.cycles-pp.__kernel_text_address
      0.11 ±  9%      +0.1        0.22 ± 15%  perf-profile.children.cycles-pp.wake_up_page_bit
      0.16 ±  6%      +0.1        0.27 ± 10%  perf-profile.children.cycles-pp.kernel_text_address
      0.00            +0.1        0.11 ± 11%  perf-profile.children.cycles-pp.get_page_from_freelist
      0.00            +0.1        0.11 ± 19%  perf-profile.children.cycles-pp.perf_mux_hrtimer_handler
      0.00            +0.1        0.11 ±  7%  perf-profile.children.cycles-pp.__alloc_pages_nodemask
      0.08 ± 10%      +0.1        0.19 ± 22%  perf-profile.children.cycles-pp.xfs_do_writepage
      0.25 ±  4%      +0.1        0.37 ± 10%  perf-profile.children.cycles-pp.switch_mm_irqs_off
      0.00            +0.1        0.12 ± 13%  perf-profile.children.cycles-pp.switch_mm
      0.08 ± 38%      +0.1        0.20 ± 19%  perf-profile.children.cycles-pp.io_serial_in
      0.18 ±  5%      +0.1        0.31 ±  7%  perf-profile.children.cycles-pp.dequeue_entity
      0.00            +0.1        0.13 ± 26%  perf-profile.children.cycles-pp.tick_nohz_next_event
      0.06 ± 11%      +0.1        0.19 ± 19%  perf-profile.children.cycles-pp.mempool_alloc
      0.32 ±  5%      +0.1        0.45 ±  6%  perf-profile.children.cycles-pp.orc_find
      0.15 ± 10%      +0.1        0.29 ± 19%  perf-profile.children.cycles-pp.xfs_destroy_ioend
      0.15 ± 11%      +0.1        0.30 ± 18%  perf-profile.children.cycles-pp.call_bio_endio
      0.08 ± 17%      +0.2        0.23 ± 25%  perf-profile.children.cycles-pp.xlog_state_done_syncing
      0.00            +0.2        0.15 ± 22%  perf-profile.children.cycles-pp.tick_nohz_get_sleep_length
      0.12 ±  8%      +0.2        0.27 ± 23%  perf-profile.children.cycles-pp.write_cache_pages
      0.10 ± 16%      +0.2        0.26 ± 16%  perf-profile.children.cycles-pp.wait_for_xmitr
      0.10 ± 19%      +0.2        0.25 ± 14%  perf-profile.children.cycles-pp.serial8250_console_putchar
      0.10 ± 17%      +0.2        0.26 ± 13%  perf-profile.children.cycles-pp.uart_console_write
      0.10 ± 16%      +0.2        0.26 ± 15%  perf-profile.children.cycles-pp.serial8250_console_write
      0.11 ± 15%      +0.2        0.27 ± 15%  perf-profile.children.cycles-pp.console_unlock
      0.09 ±  9%      +0.2        0.26 ± 12%  perf-profile.children.cycles-pp.scheduler_tick
      0.10 ± 18%      +0.2        0.28 ± 15%  perf-profile.children.cycles-pp.irq_work_run_list
      0.10 ± 15%      +0.2        0.28 ± 14%  perf-profile.children.cycles-pp.xlog_state_do_callback
      0.09 ± 12%      +0.2        0.27 ± 16%  perf-profile.children.cycles-pp.irq_work_run
      0.09 ± 12%      +0.2        0.27 ± 16%  perf-profile.children.cycles-pp.printk
      0.09 ± 12%      +0.2        0.27 ± 16%  perf-profile.children.cycles-pp.vprintk_emit
      0.09 ± 12%      +0.2        0.27 ± 17%  perf-profile.children.cycles-pp.irq_work_interrupt
      0.09 ± 12%      +0.2        0.27 ± 17%  perf-profile.children.cycles-pp.smp_irq_work_interrupt
      0.00            +0.2        0.18 ± 16%  perf-profile.children.cycles-pp.poll_idle
      0.30 ±  4%      +0.2        0.49 ± 11%  perf-profile.children.cycles-pp.update_load_avg
      1.39 ±  2%      +0.2        1.59 ±  6%  perf-profile.children.cycles-pp.__save_stack_trace
      1.43            +0.2        1.65 ±  6%  perf-profile.children.cycles-pp.save_stack_trace_tsk
      0.14 ± 13%      +0.2        0.36 ± 13%  perf-profile.children.cycles-pp.update_process_times
      0.00            +0.2        0.23 ± 22%  perf-profile.children.cycles-pp.find_busiest_group
      0.22 ±  6%      +0.2        0.45 ± 18%  perf-profile.children.cycles-pp.brd_do_bvec
      0.14 ± 13%      +0.2        0.38 ± 14%  perf-profile.children.cycles-pp.tick_sched_handle
      0.10 ±  8%      +0.2        0.34 ± 26%  perf-profile.children.cycles-pp.xfs_log_commit_cil
      0.07 ± 10%      +0.3        0.33 ± 23%  perf-profile.children.cycles-pp.io_schedule
      0.03 ± 83%      +0.3        0.29 ± 27%  perf-profile.children.cycles-pp.__softirqentry_text_start
      0.11 ±  5%      +0.3        0.36 ± 25%  perf-profile.children.cycles-pp.__xfs_trans_commit
      0.06 ± 36%      +0.3        0.31 ± 26%  perf-profile.children.cycles-pp.irq_exit
      0.08 ±  9%      +0.3        0.35 ± 23%  perf-profile.children.cycles-pp.wait_on_page_bit_common
      0.15 ± 12%      +0.3        0.42 ± 14%  perf-profile.children.cycles-pp.tick_sched_timer
      0.10 ± 11%      +0.3        0.39 ± 22%  perf-profile.children.cycles-pp.__filemap_fdatawait_range
      0.06 ± 12%      +0.3        0.37 ±  9%  perf-profile.children.cycles-pp.schedule_idle
      0.02 ±153%      +0.3        0.34 ± 17%  perf-profile.children.cycles-pp.menu_select
      0.17 ±  5%      +0.3        0.49 ± 22%  perf-profile.children.cycles-pp.xfs_vn_update_time
      0.19 ± 12%      +0.3        0.51 ± 18%  perf-profile.children.cycles-pp.xlog_iodone
      0.18 ±  5%      +0.3        0.51 ± 22%  perf-profile.children.cycles-pp.file_update_time
      0.18 ±  5%      +0.3        0.51 ± 21%  perf-profile.children.cycles-pp.xfs_file_aio_write_checks
      0.21 ± 11%      +0.4        0.60 ± 15%  perf-profile.children.cycles-pp.__hrtimer_run_queues
      0.26 ±  6%      +0.4        0.69 ± 16%  perf-profile.children.cycles-pp.pick_next_task_fair
      1.20 ±  2%      +0.4        1.64 ± 10%  perf-profile.children.cycles-pp.schedule
      0.28 ±  5%      +0.4        0.72 ± 21%  perf-profile.children.cycles-pp.xfs_file_buffered_aio_write
      0.00            +0.4        0.44 ± 22%  perf-profile.children.cycles-pp.load_balance
      0.25 ±  8%      +0.5        0.74 ± 15%  perf-profile.children.cycles-pp.hrtimer_interrupt
      1.30 ±  2%      +0.7        2.00 ±  9%  perf-profile.children.cycles-pp.__schedule
      0.31 ±  8%      +0.8        1.09 ± 16%  perf-profile.children.cycles-pp.smp_apic_timer_interrupt
      0.31 ±  8%      +0.8        1.09 ± 16%  perf-profile.children.cycles-pp.apic_timer_interrupt
      3.92 ±  2%      +0.9        4.79 ±  6%  perf-profile.children.cycles-pp.ret_from_fork
      3.92 ±  2%      +0.9        4.79 ±  6%  perf-profile.children.cycles-pp.kthread
      0.69 ±  2%      +0.9        1.64 ± 23%  perf-profile.children.cycles-pp.xlog_wait
      0.08 ± 13%      +1.2        1.27 ± 17%  perf-profile.children.cycles-pp.submit_flushes
      0.16 ±  9%      +1.6        1.74 ±  4%  perf-profile.children.cycles-pp._raw_spin_unlock_irqrestore
      0.17 ±  9%      +2.0        2.16 ± 10%  perf-profile.children.cycles-pp.raid_end_bio_io
      0.21 ±  6%      +2.0        2.25 ± 10%  perf-profile.children.cycles-pp.raid1_end_write_request
      2.24 ±  4%      +2.2        4.44 ± 15%  perf-profile.children.cycles-pp.xfs_log_force_lsn
      0.46 ±  6%      +2.3        2.73 ±  7%  perf-profile.children.cycles-pp.brd_make_request
      0.51 ±  6%      +2.3        2.81 ±  7%  perf-profile.children.cycles-pp.md_thread
      0.49 ±  6%      +2.3        2.81 ±  7%  perf-profile.children.cycles-pp.raid1d
      0.49 ±  6%      +2.3        2.81 ±  7%  perf-profile.children.cycles-pp.flush_pending_writes
      0.49 ±  6%      +2.3        2.81 ±  7%  perf-profile.children.cycles-pp.flush_bio_list
      1.80 ±  3%      +5.6        7.44 ± 27%  perf-profile.children.cycles-pp._raw_spin_lock
      2.12 ±  4%      +5.8        7.97 ± 20%  perf-profile.children.cycles-pp.remove_wait_queue
      1.33 ±  4%      +8.8       10.12 ±  8%  perf-profile.children.cycles-pp.intel_idle
      1.37 ±  4%      +9.3       10.71 ±  8%  perf-profile.children.cycles-pp.cpuidle_enter_state
      1.59 ±  4%     +10.4       11.98 ±  9%  perf-profile.children.cycles-pp.start_secondary
      1.63 ±  4%     +10.8       12.47 ±  8%  perf-profile.children.cycles-pp.secondary_startup_64
      1.63 ±  4%     +10.8       12.47 ±  8%  perf-profile.children.cycles-pp.cpu_startup_entry
      1.63 ±  4%     +10.9       12.49 ±  8%  perf-profile.children.cycles-pp.do_idle
      3.48           +12.2       15.72 ± 23%  perf-profile.children.cycles-pp.__xfs_log_force_lsn
      1.36 ± 12%     +57.8       59.12 ± 10%  perf-profile.children.cycles-pp.prepare_to_wait_event
      0.43 ± 38%     +62.4       62.82 ±  8%  perf-profile.children.cycles-pp.xfs_submit_ioend
      0.55 ± 29%     +62.5       63.10 ±  8%  perf-profile.children.cycles-pp.xfs_vm_writepages
      0.55 ± 30%     +62.5       63.10 ±  8%  perf-profile.children.cycles-pp.do_writepages
      0.55 ± 29%     +62.6       63.11 ±  8%  perf-profile.children.cycles-pp.__filemap_fdatawrite_range
      0.66 ± 25%     +62.9       63.52 ±  7%  perf-profile.children.cycles-pp.file_write_and_wait_range
      0.39 ± 43%     +63.6       64.02 ±  8%  perf-profile.children.cycles-pp.raid1_write_request
      5.43 ±  3%     +64.2       69.64 ±  5%  perf-profile.children.cycles-pp._raw_spin_lock_irqsave
     89.86           -13.5       76.31 ±  2%  perf-profile.self.cycles-pp.native_queued_spin_lock_slowpath
      0.14 ±  8%      -0.0        0.09 ± 19%  perf-profile.self.cycles-pp.md_flush_request
      0.10 ± 12%      -0.0        0.07 ± 21%  perf-profile.self.cycles-pp.account_entity_enqueue
      0.06 ±  7%      +0.0        0.08 ± 12%  perf-profile.self.cycles-pp.pick_next_task_fair
      0.05 ± 12%      +0.0        0.08 ± 18%  perf-profile.self.cycles-pp.___perf_sw_event
      0.15 ±  6%      +0.0        0.18 ±  9%  perf-profile.self.cycles-pp.__update_load_avg_se
      0.17 ±  4%      +0.0        0.22 ± 10%  perf-profile.self.cycles-pp.__schedule
      0.10 ± 11%      +0.1        0.15 ± 11%  perf-profile.self.cycles-pp._raw_spin_lock
      0.02 ±153%      +0.1        0.07 ± 16%  perf-profile.self.cycles-pp.delay_tsc
      0.02 ±152%      +0.1        0.07 ± 23%  perf-profile.self.cycles-pp.set_next_entity
      0.03 ±100%      +0.1        0.08 ± 15%  perf-profile.self.cycles-pp.find_next_bit
      0.08 ±  5%      +0.1        0.14 ± 14%  perf-profile.self.cycles-pp.native_write_msr
      0.01 ±200%      +0.1        0.07 ± 23%  perf-profile.self.cycles-pp.kmem_cache_alloc
      0.29 ±  4%      +0.1        0.36 ±  8%  perf-profile.self.cycles-pp.__orc_find
      0.14 ±  7%      +0.1        0.21 ± 12%  perf-profile.self.cycles-pp.switch_mm_irqs_off
      0.00            +0.1        0.08 ± 11%  perf-profile.self.cycles-pp.clear_page_erms
      0.00            +0.1        0.08 ± 28%  perf-profile.self.cycles-pp.__indirect_thunk_start
      0.00            +0.1        0.08 ± 20%  perf-profile.self.cycles-pp.md_wakeup_thread
      0.34 ±  6%      +0.1        0.43 ± 12%  perf-profile.self.cycles-pp._raw_spin_lock_irqsave
      0.18 ±  4%      +0.1        0.27 ± 12%  perf-profile.self.cycles-pp.idle_cpu
      0.16 ± 10%      +0.1        0.25 ± 13%  perf-profile.self.cycles-pp.__module_address
      0.06 ± 11%      +0.1        0.17 ± 14%  perf-profile.self.cycles-pp._raw_spin_unlock_irqrestore
      0.08 ± 38%      +0.1        0.20 ± 19%  perf-profile.self.cycles-pp.io_serial_in
      0.18 ±  5%      +0.1        0.32 ± 15%  perf-profile.self.cycles-pp.update_load_avg
      0.00            +0.1        0.15 ± 17%  perf-profile.self.cycles-pp.poll_idle
      0.00            +0.2        0.15 ± 16%  perf-profile.self.cycles-pp.menu_select
      0.00            +0.2        0.18 ± 24%  perf-profile.self.cycles-pp.find_busiest_group
      0.02 ±152%      +0.3        0.35 ± 21%  perf-profile.self.cycles-pp.raid1_write_request
      1.33 ±  4%      +8.8       10.12 ±  8%  perf-profile.self.cycles-pp.intel_idle


                                                                                
                                 aim7.jobs-per-min                              
                                                                                
  1700 +-+------------------------------------------------------------------+   
       |+ ++++++ :+   ++++ ++++     +++     ++++++ +  +  ++++++++++++ ++  ++|   
  1600 +-+       + +++    +    +++++   ++.++      + ++ ++            +  ++  |   
       |                                                                    |   
       |                                                                    |   
  1500 +-+                                                                  |   
       |                                                                    |   
  1400 +-+                                                                  |   
       |                                                                    |   
  1300 +-+                                                                  |   
       |                                                                    |   
       O  OO         OO  O              O          O                        |   
  1200 +OO  OOOOOOOOO  OO OOOOOOOOOOOOOO  OOOOOOOOO O                       |   
       |                                                                    |   
  1100 +-+------------------------------------------------------------------+   
                                                                                
                                                                                                                                                                
                                                                                
[*] bisect-good sample
[O] bisect-bad  sample


Disclaimer:
Results have been estimated based on internal Intel analysis and are provided
for informational purposes only. Any difference in system hardware or software
design or configuration may affect actual performance.


Thanks,
Xiaolong

View attachment "config-4.17.0-rc1-00189-g5a409b4" of type "text/plain" (164401 bytes)

View attachment "job-script" of type "text/plain" (7703 bytes)

View attachment "job.yaml" of type "text/plain" (5344 bytes)

View attachment "reproduce" of type "text/plain" (1026 bytes)

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ