[<prev] [next>] [thread-next>] [day] [month] [year] [list]
Message-ID: <20180604064203.GG16472@yexl-desktop>
Date: Mon, 4 Jun 2018 14:42:03 +0800
From: kernel test robot <xiaolong.ye@...el.com>
To: Xiao Ni <xni@...hat.com>
Cc: Shaohua Li <shli@...com>, Ming Lei <ming.lei@...hat.com>,
LKML <linux-kernel@...r.kernel.org>,
Stephen Rothwell <sfr@...b.auug.org.au>, lkp@...org
Subject: [lkp-robot] [MD] 5a409b4f56: aim7.jobs-per-min -27.5% regression
Greeting,
FYI, we noticed a -27.5% regression of aim7.jobs-per-min due to commit:
commit: 5a409b4f56d50b212334f338cb8465d65550cd85 ("MD: fix lock contention for flush bios")
https://git.kernel.org/cgit/linux/kernel/git/next/linux-next.git master
in testcase: aim7
on test machine: 40 threads Intel(R) Xeon(R) CPU E5-2690 v2 @ 3.00GHz with 384G memory
with following parameters:
disk: 4BRD_12G
md: RAID1
fs: xfs
test: sync_disk_rw
load: 600
cpufreq_governor: performance
test-description: AIM7 is a traditional UNIX system level benchmark suite which is used to test and measure the performance of multiuser system.
test-url: https://sourceforge.net/projects/aimbench/files/aim-suite7/
Details are as below:
-------------------------------------------------------------------------------------------------->
=========================================================================================
compiler/cpufreq_governor/disk/fs/kconfig/load/md/rootfs/tbox_group/test/testcase:
gcc-7/performance/4BRD_12G/xfs/x86_64-rhel-7.2/600/RAID1/debian-x86_64-2016-08-31.cgz/lkp-ivb-ep01/sync_disk_rw/aim7
commit:
448ec638c6 ("md/raid5: Assigning NULL to sh->batch_head before testing bit R5_Overlap of a stripe")
5a409b4f56 ("MD: fix lock contention for flush bios")
448ec638c6bcf369 5a409b4f56d50b212334f338cb
---------------- --------------------------
%stddev %change %stddev
\ | \
1640 -27.5% 1189 aim7.jobs-per-min
2194 +37.9% 3026 aim7.time.elapsed_time
2194 +37.9% 3026 aim7.time.elapsed_time.max
50990311 -95.8% 2148266 aim7.time.involuntary_context_switches
107965 ± 4% -26.4% 79516 ± 2% aim7.time.minor_page_faults
49.14 +82.5% 89.66 ± 2% aim7.time.user_time
7.123e+08 -35.7% 4.582e+08 aim7.time.voluntary_context_switches
672282 +36.8% 919615 interrupts.CAL:Function_call_interrupts
16631387 ± 2% -39.9% 9993075 ± 7% softirqs.RCU
9708009 +186.1% 27778773 softirqs.SCHED
33436649 +45.5% 48644912 softirqs.TIMER
4.16 -2.1 2.01 mpstat.cpu.idle%
0.24 ± 2% +27.7 27.91 mpstat.cpu.iowait%
95.51 -25.6 69.94 mpstat.cpu.sys%
0.09 +0.0 0.13 mpstat.cpu.usr%
6051756 ± 3% +59.0% 9623085 numa-numastat.node0.local_node
6055311 ± 3% +59.0% 9626996 numa-numastat.node0.numa_hit
6481209 ± 3% +48.4% 9616310 numa-numastat.node1.local_node
6485866 ± 3% +48.3% 9620756 numa-numastat.node1.numa_hit
61404 -27.7% 44424 vmstat.io.bo
2.60 ± 18% +11519.2% 302.10 vmstat.procs.b
304.10 -84.9% 45.80 ± 2% vmstat.procs.r
400477 -43.5% 226094 vmstat.system.cs
166461 -49.9% 83332 vmstat.system.in
78397 +27.0% 99567 meminfo.Dirty
14427 +18.4% 17082 meminfo.Inactive(anon)
1963 ± 5% +5.4% 2068 ± 4% meminfo.Mlocked
101143 +991.0% 1103488 meminfo.SUnreclaim
53684 ± 4% -18.1% 43946 ± 3% meminfo.Shmem
175580 +571.4% 1178829 meminfo.Slab
39406 +26.2% 49717 numa-meminfo.node0.Dirty
1767204 ± 10% +37.2% 2425487 ± 2% numa-meminfo.node0.MemUsed
51634 ± 18% +979.3% 557316 numa-meminfo.node0.SUnreclaim
92259 ± 13% +551.7% 601288 numa-meminfo.node0.Slab
38969 +28.0% 49863 numa-meminfo.node1.Dirty
1895204 ± 10% +24.7% 2363037 ± 3% numa-meminfo.node1.MemUsed
49512 ± 19% +1003.1% 546165 numa-meminfo.node1.SUnreclaim
83323 ± 14% +593.1% 577534 numa-meminfo.node1.Slab
2.524e+09 +894.5% 2.51e+10 cpuidle.C1.time
50620790 +316.5% 2.109e+08 cpuidle.C1.usage
3.965e+08 +1871.1% 7.815e+09 cpuidle.C1E.time
5987788 +186.1% 17129412 cpuidle.C1E.usage
2.506e+08 +97.5% 4.948e+08 ± 2% cpuidle.C3.time
2923498 -55.7% 1295033 cpuidle.C3.usage
5.327e+08 +179.9% 1.491e+09 cpuidle.C6.time
779874 ± 2% +229.3% 2567769 cpuidle.C6.usage
6191357 +3333.6% 2.126e+08 cpuidle.POLL.time
204095 +1982.1% 4249504 cpuidle.POLL.usage
9850 +26.3% 12444 numa-vmstat.node0.nr_dirty
12908 ± 18% +979.3% 139321 numa-vmstat.node0.nr_slab_unreclaimable
8876 +29.6% 11505 numa-vmstat.node0.nr_zone_write_pending
3486319 ± 4% +55.1% 5407021 numa-vmstat.node0.numa_hit
3482713 ± 4% +55.1% 5403066 numa-vmstat.node0.numa_local
9743 +28.1% 12479 numa-vmstat.node1.nr_dirty
12377 ± 19% +1003.1% 136532 numa-vmstat.node1.nr_slab_unreclaimable
9287 +30.0% 12074 numa-vmstat.node1.nr_zone_write_pending
3678995 ± 4% +44.8% 5326772 numa-vmstat.node1.numa_hit
3497785 ± 4% +47.1% 5145705 numa-vmstat.node1.numa_local
252.70 +100.2% 505.90 slabinfo.biovec-max.active_objs
282.70 +99.1% 562.90 slabinfo.biovec-max.num_objs
2978 ± 17% +52.5% 4543 ± 14% slabinfo.dmaengine-unmap-16.active_objs
2978 ± 17% +52.5% 4543 ± 14% slabinfo.dmaengine-unmap-16.num_objs
2078 +147.9% 5153 ± 11% slabinfo.ip6_dst_cache.active_objs
2078 +148.1% 5157 ± 11% slabinfo.ip6_dst_cache.num_objs
5538 ± 2% +26.2% 6990 ± 3% slabinfo.kmalloc-1024.active_objs
5586 ± 3% +27.1% 7097 ± 3% slabinfo.kmalloc-1024.num_objs
6878 +47.6% 10151 ± 5% slabinfo.kmalloc-192.active_objs
6889 +47.5% 10160 ± 5% slabinfo.kmalloc-192.num_objs
9843 ± 5% +1.6e+05% 16002876 slabinfo.kmalloc-64.active_objs
161.90 ± 4% +1.5e+05% 250044 slabinfo.kmalloc-64.active_slabs
10386 ± 4% +1.5e+05% 16002877 slabinfo.kmalloc-64.num_objs
161.90 ± 4% +1.5e+05% 250044 slabinfo.kmalloc-64.num_slabs
432.80 ± 12% +45.2% 628.50 ± 6% slabinfo.nfs_read_data.active_objs
432.80 ± 12% +45.2% 628.50 ± 6% slabinfo.nfs_read_data.num_objs
3956 -23.1% 3041 slabinfo.pool_workqueue.active_objs
4098 -19.8% 3286 slabinfo.pool_workqueue.num_objs
360.50 ± 15% +56.6% 564.70 ± 11% slabinfo.secpath_cache.active_objs
360.50 ± 15% +56.6% 564.70 ± 11% slabinfo.secpath_cache.num_objs
35373 ± 2% -8.3% 32432 proc-vmstat.nr_active_anon
19595 +27.1% 24914 proc-vmstat.nr_dirty
3607 +18.4% 4270 proc-vmstat.nr_inactive_anon
490.30 ± 5% +5.4% 516.90 ± 4% proc-vmstat.nr_mlock
13421 ± 4% -18.1% 10986 ± 3% proc-vmstat.nr_shmem
18608 +1.2% 18834 proc-vmstat.nr_slab_reclaimable
25286 +991.0% 275882 proc-vmstat.nr_slab_unreclaimable
35405 ± 2% -8.3% 32465 proc-vmstat.nr_zone_active_anon
3607 +18.4% 4270 proc-vmstat.nr_zone_inactive_anon
18161 +29.8% 23572 proc-vmstat.nr_zone_write_pending
76941 ± 5% -36.8% 48622 ± 4% proc-vmstat.numa_hint_faults
33878 ± 7% -35.5% 21836 ± 5% proc-vmstat.numa_hint_faults_local
12568956 +53.3% 19272377 proc-vmstat.numa_hit
12560739 +53.4% 19264015 proc-vmstat.numa_local
17938 ± 3% -33.5% 11935 ± 2% proc-vmstat.numa_pages_migrated
78296 ± 5% -36.0% 50085 ± 4% proc-vmstat.numa_pte_updates
8848 ± 6% -38.2% 5466 ± 6% proc-vmstat.pgactivate
8874568 ± 8% +368.7% 41590920 proc-vmstat.pgalloc_normal
5435965 +39.2% 7564148 proc-vmstat.pgfault
12863707 +255.1% 45683570 proc-vmstat.pgfree
17938 ± 3% -33.5% 11935 ± 2% proc-vmstat.pgmigrate_success
1.379e+13 -40.8% 8.17e+12 perf-stat.branch-instructions
0.30 +0.1 0.42 perf-stat.branch-miss-rate%
4.2e+10 -17.6% 3.462e+10 perf-stat.branch-misses
15.99 +3.8 19.74 perf-stat.cache-miss-rate%
3.779e+10 -21.6% 2.963e+10 perf-stat.cache-misses
2.364e+11 -36.5% 1.501e+11 perf-stat.cache-references
8.795e+08 -22.2% 6.84e+08 perf-stat.context-switches
4.44 -7.2% 4.12 perf-stat.cpi
2.508e+14 -44.5% 1.393e+14 perf-stat.cpu-cycles
36915392 +60.4% 59211221 perf-stat.cpu-migrations
0.29 ± 2% +0.0 0.34 ± 4% perf-stat.dTLB-load-miss-rate%
4.14e+10 -30.2% 2.89e+10 ± 4% perf-stat.dTLB-load-misses
1.417e+13 -40.1% 8.491e+12 perf-stat.dTLB-loads
0.20 ± 4% -0.0 0.18 ± 5% perf-stat.dTLB-store-miss-rate%
3.072e+09 ± 4% -28.0% 2.21e+09 ± 4% perf-stat.dTLB-store-misses
1.535e+12 -20.2% 1.225e+12 perf-stat.dTLB-stores
90.73 -11.7 79.07 perf-stat.iTLB-load-miss-rate%
8.291e+09 -6.6% 7.743e+09 perf-stat.iTLB-load-misses
8.473e+08 +141.8% 2.049e+09 ± 3% perf-stat.iTLB-loads
5.646e+13 -40.2% 3.378e+13 perf-stat.instructions
6810 -35.9% 4362 perf-stat.instructions-per-iTLB-miss
0.23 +7.8% 0.24 perf-stat.ipc
5326672 +39.2% 7413706 perf-stat.minor-faults
1.873e+10 -29.9% 1.312e+10 perf-stat.node-load-misses
2.093e+10 -29.2% 1.481e+10 perf-stat.node-loads
39.38 -0.7 38.72 perf-stat.node-store-miss-rate%
1.087e+10 -16.6% 9.069e+09 perf-stat.node-store-misses
1.673e+10 -14.2% 1.435e+10 perf-stat.node-stores
5326695 +39.2% 7413708 perf-stat.page-faults
1875095 ± 7% -54.8% 846645 ± 16% sched_debug.cfs_rq:/.MIN_vruntime.avg
32868920 ± 6% -35.7% 21150379 ± 14% sched_debug.cfs_rq:/.MIN_vruntime.max
7267340 ± 5% -44.7% 4015798 ± 14% sched_debug.cfs_rq:/.MIN_vruntime.stddev
4278 ± 7% -54.7% 1939 ± 11% sched_debug.cfs_rq:/.exec_clock.stddev
245.48 ± 2% +65.3% 405.75 ± 7% sched_debug.cfs_rq:/.load_avg.avg
2692 ± 6% +126.0% 6087 ± 7% sched_debug.cfs_rq:/.load_avg.max
33.09 -73.0% 8.94 ± 7% sched_debug.cfs_rq:/.load_avg.min
507.40 ± 4% +128.0% 1156 ± 7% sched_debug.cfs_rq:/.load_avg.stddev
1875095 ± 7% -54.8% 846645 ± 16% sched_debug.cfs_rq:/.max_vruntime.avg
32868921 ± 6% -35.7% 21150379 ± 14% sched_debug.cfs_rq:/.max_vruntime.max
7267341 ± 5% -44.7% 4015798 ± 14% sched_debug.cfs_rq:/.max_vruntime.stddev
35887197 -13.2% 31149130 sched_debug.cfs_rq:/.min_vruntime.avg
37385506 -14.3% 32043914 sched_debug.cfs_rq:/.min_vruntime.max
34416296 -12.3% 30183927 sched_debug.cfs_rq:/.min_vruntime.min
1228844 ± 8% -52.6% 582759 ± 4% sched_debug.cfs_rq:/.min_vruntime.stddev
0.83 -28.1% 0.60 ± 6% sched_debug.cfs_rq:/.nr_running.avg
2.07 ± 3% -24.6% 1.56 ± 8% sched_debug.cfs_rq:/.nr_running.max
20.52 ± 4% -48.8% 10.52 ± 3% sched_debug.cfs_rq:/.nr_spread_over.avg
35.96 ± 5% -42.2% 20.77 ± 9% sched_debug.cfs_rq:/.nr_spread_over.max
8.97 ± 11% -44.5% 4.98 ± 8% sched_debug.cfs_rq:/.nr_spread_over.min
6.40 ± 12% -45.5% 3.49 ± 7% sched_debug.cfs_rq:/.nr_spread_over.stddev
21.78 ± 7% +143.3% 53.00 ± 9% sched_debug.cfs_rq:/.runnable_load_avg.avg
328.86 ± 18% +303.4% 1326 ± 14% sched_debug.cfs_rq:/.runnable_load_avg.max
55.97 ± 17% +286.0% 216.07 ± 13% sched_debug.cfs_rq:/.runnable_load_avg.stddev
0.10 ± 29% -82.4% 0.02 ± 50% sched_debug.cfs_rq:/.spread.avg
3.43 ± 25% -79.9% 0.69 ± 50% sched_debug.cfs_rq:/.spread.max
0.56 ± 26% -80.7% 0.11 ± 50% sched_debug.cfs_rq:/.spread.stddev
1228822 ± 8% -52.6% 582732 ± 4% sched_debug.cfs_rq:/.spread0.stddev
992.30 -24.9% 745.56 ± 2% sched_debug.cfs_rq:/.util_avg.avg
1485 -18.1% 1217 ± 2% sched_debug.cfs_rq:/.util_avg.max
515.45 ± 2% -25.2% 385.73 ± 6% sched_debug.cfs_rq:/.util_avg.min
201.54 -14.9% 171.52 ± 3% sched_debug.cfs_rq:/.util_avg.stddev
248.73 ± 6% -38.1% 154.02 ± 8% sched_debug.cfs_rq:/.util_est_enqueued.avg
222.78 ± 3% -15.8% 187.58 ± 2% sched_debug.cfs_rq:/.util_est_enqueued.stddev
77097 ± 4% +278.4% 291767 ± 11% sched_debug.cpu.avg_idle.avg
181319 ± 6% +298.7% 722862 ± 3% sched_debug.cpu.avg_idle.max
19338 +392.3% 95203 ± 17% sched_debug.cpu.avg_idle.min
34877 ± 6% +303.5% 140732 ± 6% sched_debug.cpu.avg_idle.stddev
1107408 +37.6% 1523823 sched_debug.cpu.clock.avg
1107427 +37.6% 1523834 sched_debug.cpu.clock.max
1107385 +37.6% 1523811 sched_debug.cpu.clock.min
13.10 ± 9% -48.1% 6.80 ± 8% sched_debug.cpu.clock.stddev
1107408 +37.6% 1523823 sched_debug.cpu.clock_task.avg
1107427 +37.6% 1523834 sched_debug.cpu.clock_task.max
1107385 +37.6% 1523811 sched_debug.cpu.clock_task.min
13.10 ± 9% -48.1% 6.80 ± 8% sched_debug.cpu.clock_task.stddev
30.36 ± 7% +107.7% 63.06 ± 12% sched_debug.cpu.cpu_load[0].avg
381.48 ± 18% +269.8% 1410 ± 18% sched_debug.cpu.cpu_load[0].max
63.92 ± 18% +262.2% 231.50 ± 17% sched_debug.cpu.cpu_load[0].stddev
31.34 ± 5% +118.4% 68.44 ± 9% sched_debug.cpu.cpu_load[1].avg
323.62 ± 17% +349.5% 1454 ± 14% sched_debug.cpu.cpu_load[1].max
53.23 ± 16% +350.3% 239.71 ± 13% sched_debug.cpu.cpu_load[1].stddev
32.15 ± 3% +129.4% 73.74 ± 6% sched_debug.cpu.cpu_load[2].avg
285.20 ± 14% +420.8% 1485 ± 9% sched_debug.cpu.cpu_load[2].max
46.66 ± 12% +430.0% 247.32 ± 8% sched_debug.cpu.cpu_load[2].stddev
33.02 ± 2% +133.2% 77.00 ± 3% sched_debug.cpu.cpu_load[3].avg
252.16 ± 10% +481.2% 1465 ± 7% sched_debug.cpu.cpu_load[3].max
40.74 ± 8% +503.2% 245.72 ± 6% sched_debug.cpu.cpu_load[3].stddev
33.86 +131.5% 78.38 ± 2% sched_debug.cpu.cpu_load[4].avg
219.81 ± 8% +522.6% 1368 ± 5% sched_debug.cpu.cpu_load[4].max
35.45 ± 7% +554.2% 231.90 ± 4% sched_debug.cpu.cpu_load[4].stddev
2600 ± 4% -30.5% 1807 ± 4% sched_debug.cpu.curr->pid.avg
25309 ± 4% -19.5% 20367 ± 4% sched_debug.cpu.curr->pid.max
4534 ± 7% -21.2% 3573 ± 5% sched_debug.cpu.curr->pid.stddev
0.00 ± 2% -27.6% 0.00 ± 6% sched_debug.cpu.next_balance.stddev
1083917 +38.6% 1502777 sched_debug.cpu.nr_load_updates.avg
1088142 +38.6% 1508302 sched_debug.cpu.nr_load_updates.max
1082048 +38.7% 1501073 sched_debug.cpu.nr_load_updates.min
3.53 ± 6% -73.0% 0.95 ± 6% sched_debug.cpu.nr_running.avg
11.54 ± 3% -62.1% 4.37 ± 10% sched_debug.cpu.nr_running.max
3.10 ± 3% -66.8% 1.03 ± 9% sched_debug.cpu.nr_running.stddev
10764176 -22.4% 8355047 sched_debug.cpu.nr_switches.avg
10976436 -22.2% 8545010 sched_debug.cpu.nr_switches.max
10547712 -22.8% 8143037 sched_debug.cpu.nr_switches.min
148628 ± 3% -22.7% 114880 ± 7% sched_debug.cpu.nr_switches.stddev
11.13 ± 2% +24.5% 13.85 sched_debug.cpu.nr_uninterruptible.avg
6420 ± 8% -48.7% 3296 ± 11% sched_debug.cpu.nr_uninterruptible.max
-5500 -37.2% -3455 sched_debug.cpu.nr_uninterruptible.min
3784 ± 6% -47.2% 1997 ± 4% sched_debug.cpu.nr_uninterruptible.stddev
10812670 -22.7% 8356821 sched_debug.cpu.sched_count.avg
11020646 -22.5% 8546277 sched_debug.cpu.sched_count.max
10601390 -23.2% 8144743 sched_debug.cpu.sched_count.min
144529 ± 3% -20.9% 114359 ± 7% sched_debug.cpu.sched_count.stddev
706116 +259.0% 2534721 sched_debug.cpu.sched_goidle.avg
771307 +232.4% 2564059 sched_debug.cpu.sched_goidle.max
644658 +286.9% 2494236 sched_debug.cpu.sched_goidle.min
49847 ± 6% -67.9% 15979 ± 7% sched_debug.cpu.sched_goidle.stddev
9618827 -39.9% 5780369 sched_debug.cpu.ttwu_count.avg
8990451 -61.7% 3441265 ± 4% sched_debug.cpu.ttwu_count.min
418563 ± 25% +244.2% 1440565 ± 7% sched_debug.cpu.ttwu_count.stddev
640964 -93.7% 40366 ± 2% sched_debug.cpu.ttwu_local.avg
679527 -92.1% 53476 ± 4% sched_debug.cpu.ttwu_local.max
601661 -94.9% 30636 ± 3% sched_debug.cpu.ttwu_local.min
24242 ± 21% -77.7% 5405 ± 9% sched_debug.cpu.ttwu_local.stddev
1107383 +37.6% 1523810 sched_debug.cpu_clk
1107383 +37.6% 1523810 sched_debug.ktime
0.00 -49.4% 0.00 ± 65% sched_debug.rt_rq:/.rt_nr_migratory.avg
0.03 -49.4% 0.01 ± 65% sched_debug.rt_rq:/.rt_nr_migratory.max
0.00 -49.4% 0.00 ± 65% sched_debug.rt_rq:/.rt_nr_migratory.stddev
0.00 -49.4% 0.00 ± 65% sched_debug.rt_rq:/.rt_nr_running.avg
0.03 -49.4% 0.01 ± 65% sched_debug.rt_rq:/.rt_nr_running.max
0.00 -49.4% 0.00 ± 65% sched_debug.rt_rq:/.rt_nr_running.stddev
0.01 ± 8% +79.9% 0.01 ± 23% sched_debug.rt_rq:/.rt_time.avg
1107805 +37.6% 1524235 sched_debug.sched_clk
87.59 -87.6 0.00 perf-profile.calltrace.cycles-pp.md_flush_request.raid1_make_request.md_handle_request.md_make_request.generic_make_request
87.57 -87.6 0.00 perf-profile.calltrace.cycles-pp.submit_bio_wait.blkdev_issue_flush.xfs_file_fsync.xfs_file_write_iter.__vfs_write
87.59 -87.5 0.05 ±299% perf-profile.calltrace.cycles-pp.blkdev_issue_flush.xfs_file_fsync.xfs_file_write_iter.__vfs_write.vfs_write
87.51 -87.5 0.00 perf-profile.calltrace.cycles-pp.generic_make_request.submit_bio.submit_bio_wait.blkdev_issue_flush.xfs_file_fsync
87.51 -87.5 0.00 perf-profile.calltrace.cycles-pp.submit_bio.submit_bio_wait.blkdev_issue_flush.xfs_file_fsync.xfs_file_write_iter
87.50 -87.5 0.00 perf-profile.calltrace.cycles-pp.md_make_request.generic_make_request.submit_bio.submit_bio_wait.blkdev_issue_flush
87.50 -87.5 0.00 perf-profile.calltrace.cycles-pp.md_handle_request.md_make_request.generic_make_request.submit_bio.submit_bio_wait
82.37 -82.4 0.00 perf-profile.calltrace.cycles-pp._raw_spin_lock_irq.md_flush_request.raid1_make_request.md_handle_request.md_make_request
82.23 -82.2 0.00 perf-profile.calltrace.cycles-pp.native_queued_spin_lock_slowpath._raw_spin_lock_irq.md_flush_request.raid1_make_request.md_handle_request
87.79 -25.0 62.75 ± 8% perf-profile.calltrace.cycles-pp.raid1_make_request.md_handle_request.md_make_request.generic_make_request.submit_bio
92.78 -13.0 79.76 perf-profile.calltrace.cycles-pp.xfs_file_fsync.xfs_file_write_iter.__vfs_write.vfs_write.ksys_write
93.08 -12.6 80.49 perf-profile.calltrace.cycles-pp.xfs_file_write_iter.__vfs_write.vfs_write.ksys_write.do_syscall_64
93.08 -12.6 80.50 perf-profile.calltrace.cycles-pp.__vfs_write.vfs_write.ksys_write.do_syscall_64.entry_SYSCALL_64_after_hwframe
93.11 -12.6 80.56 perf-profile.calltrace.cycles-pp.vfs_write.ksys_write.do_syscall_64.entry_SYSCALL_64_after_hwframe
93.11 -12.6 80.56 perf-profile.calltrace.cycles-pp.ksys_write.do_syscall_64.entry_SYSCALL_64_after_hwframe
93.14 -12.5 80.64 perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe
93.15 -12.5 80.65 perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe
3.40 ± 2% -1.4 1.97 ± 8% perf-profile.calltrace.cycles-pp.worker_thread.kthread.ret_from_fork
3.33 ± 2% -1.4 1.96 ± 9% perf-profile.calltrace.cycles-pp.process_one_work.worker_thread.kthread.ret_from_fork
1.12 ± 2% -0.7 0.42 ± 68% perf-profile.calltrace.cycles-pp.__save_stack_trace.save_stack_trace_tsk.__account_scheduler_latency.enqueue_entity.enqueue_task_fair
1.16 ± 2% -0.6 0.60 ± 17% perf-profile.calltrace.cycles-pp.save_stack_trace_tsk.__account_scheduler_latency.enqueue_entity.enqueue_task_fair.ttwu_do_activate
0.00 +0.6 0.59 ± 15% perf-profile.calltrace.cycles-pp._raw_spin_lock_irqsave.__wake_up_common_lock.raid1_write_request.raid1_make_request.md_handle_request
0.00 +0.6 0.64 ± 15% perf-profile.calltrace.cycles-pp.__wake_up_common_lock.raid1_write_request.raid1_make_request.md_handle_request.md_make_request
0.00 +0.7 0.65 ± 10% perf-profile.calltrace.cycles-pp.enqueue_entity.enqueue_task_fair.ttwu_do_activate.sched_ttwu_pending.do_idle
0.00 +0.7 0.68 ± 10% perf-profile.calltrace.cycles-pp.enqueue_task_fair.ttwu_do_activate.sched_ttwu_pending.do_idle.cpu_startup_entry
0.00 +0.7 0.69 ± 10% perf-profile.calltrace.cycles-pp.ttwu_do_activate.sched_ttwu_pending.do_idle.cpu_startup_entry.start_secondary
0.00 +0.8 0.79 ± 11% perf-profile.calltrace.cycles-pp.sched_ttwu_pending.do_idle.cpu_startup_entry.start_secondary.secondary_startup_64
0.00 +0.8 0.83 ± 7% perf-profile.calltrace.cycles-pp.__schedule.schedule.raid1_write_request.raid1_make_request.md_handle_request
0.62 ± 3% +0.8 1.45 ± 22% perf-profile.calltrace.cycles-pp.native_queued_spin_lock_slowpath._raw_spin_lock_irqsave.remove_wait_queue.xlog_wait.__xfs_log_force_lsn
0.00 +0.8 0.83 ± 7% perf-profile.calltrace.cycles-pp.schedule.raid1_write_request.raid1_make_request.md_handle_request.md_make_request
0.63 ± 2% +0.8 1.46 ± 22% perf-profile.calltrace.cycles-pp.remove_wait_queue.xlog_wait.__xfs_log_force_lsn.xfs_log_force_lsn.xfs_file_fsync
0.62 ± 2% +0.8 1.46 ± 22% perf-profile.calltrace.cycles-pp._raw_spin_lock_irqsave.remove_wait_queue.xlog_wait.__xfs_log_force_lsn.xfs_log_force_lsn
3.92 ± 2% +0.9 4.79 ± 6% perf-profile.calltrace.cycles-pp.ret_from_fork
3.92 ± 2% +0.9 4.79 ± 6% perf-profile.calltrace.cycles-pp.kthread.ret_from_fork
0.69 ± 2% +0.9 1.64 ± 23% perf-profile.calltrace.cycles-pp.xlog_wait.__xfs_log_force_lsn.xfs_log_force_lsn.xfs_file_fsync.xfs_file_write_iter
0.00 +1.2 1.17 ± 8% perf-profile.calltrace.cycles-pp._raw_spin_unlock_irqrestore.prepare_to_wait_event.raid1_write_request.raid1_make_request.md_handle_request
0.00 +1.2 1.23 ± 18% perf-profile.calltrace.cycles-pp.prepare_to_wait_event.raid1_write_request.raid1_make_request.md_handle_request.submit_flushes
0.00 +1.3 1.27 ± 17% perf-profile.calltrace.cycles-pp.raid1_write_request.raid1_make_request.md_handle_request.submit_flushes.process_one_work
0.00 +1.3 1.27 ± 17% perf-profile.calltrace.cycles-pp.md_handle_request.submit_flushes.process_one_work.worker_thread.kthread
0.00 +1.3 1.27 ± 17% perf-profile.calltrace.cycles-pp.raid1_make_request.md_handle_request.submit_flushes.process_one_work.worker_thread
0.00 +1.3 1.27 ± 17% perf-profile.calltrace.cycles-pp.submit_flushes.process_one_work.worker_thread.kthread.ret_from_fork
0.00 +1.6 1.65 ± 14% perf-profile.calltrace.cycles-pp.try_to_wake_up.autoremove_wake_function.__wake_up_common.__wake_up_common_lock.raid_end_bio_io
0.00 +1.7 1.71 ± 14% perf-profile.calltrace.cycles-pp.autoremove_wake_function.__wake_up_common.__wake_up_common_lock.raid_end_bio_io.raid1_end_write_request
0.00 +1.7 1.71 ± 14% perf-profile.calltrace.cycles-pp.__wake_up_common.__wake_up_common_lock.raid_end_bio_io.raid1_end_write_request.brd_make_request
0.00 +1.9 1.86 ± 13% perf-profile.calltrace.cycles-pp.__wake_up_common_lock.raid_end_bio_io.raid1_end_write_request.brd_make_request.generic_make_request
0.00 +2.1 2.10 ± 10% perf-profile.calltrace.cycles-pp.native_queued_spin_lock_slowpath._raw_spin_lock_irqsave.remove_wait_queue.__xfs_log_force_lsn.xfs_log_force_lsn
0.00 +2.1 2.10 ± 10% perf-profile.calltrace.cycles-pp._raw_spin_lock_irqsave.remove_wait_queue.__xfs_log_force_lsn.xfs_log_force_lsn.xfs_file_fsync
0.00 +2.1 2.11 ± 10% perf-profile.calltrace.cycles-pp.remove_wait_queue.__xfs_log_force_lsn.xfs_log_force_lsn.xfs_file_fsync.xfs_file_write_iter
0.00 +2.2 2.16 ± 10% perf-profile.calltrace.cycles-pp.raid_end_bio_io.raid1_end_write_request.brd_make_request.generic_make_request.flush_bio_list
2.24 ± 4% +2.2 4.44 ± 15% perf-profile.calltrace.cycles-pp.xfs_log_force_lsn.xfs_file_fsync.xfs_file_write_iter.__vfs_write.vfs_write
0.00 +2.3 2.25 ± 10% perf-profile.calltrace.cycles-pp.raid1_end_write_request.brd_make_request.generic_make_request.flush_bio_list.flush_pending_writes
0.00 +2.3 2.30 ± 20% perf-profile.calltrace.cycles-pp.native_queued_spin_lock_slowpath._raw_spin_lock_irqsave.raid1_write_request.raid1_make_request.md_handle_request
0.00 +2.4 2.35 ± 20% perf-profile.calltrace.cycles-pp._raw_spin_lock_irqsave.raid1_write_request.raid1_make_request.md_handle_request.md_make_request
0.37 ± 65% +2.4 2.81 ± 7% perf-profile.calltrace.cycles-pp.md_thread.kthread.ret_from_fork
0.26 ±100% +2.5 2.81 ± 7% perf-profile.calltrace.cycles-pp.raid1d.md_thread.kthread.ret_from_fork
0.26 ±100% +2.5 2.81 ± 7% perf-profile.calltrace.cycles-pp.flush_pending_writes.raid1d.md_thread.kthread.ret_from_fork
0.26 ±100% +2.6 2.81 ± 7% perf-profile.calltrace.cycles-pp.flush_bio_list.flush_pending_writes.raid1d.md_thread.kthread
0.10 ±200% +2.7 2.76 ± 7% perf-profile.calltrace.cycles-pp.generic_make_request.flush_bio_list.flush_pending_writes.raid1d.md_thread
0.00 +2.7 2.73 ± 7% perf-profile.calltrace.cycles-pp.brd_make_request.generic_make_request.flush_bio_list.flush_pending_writes.raid1d
1.20 ± 3% +3.1 4.35 ± 15% perf-profile.calltrace.cycles-pp.__xfs_log_force_lsn.xfs_log_force_lsn.xfs_file_fsync.xfs_file_write_iter.__vfs_write
0.63 ± 6% +3.8 4.38 ± 27% perf-profile.calltrace.cycles-pp.native_queued_spin_lock_slowpath._raw_spin_lock_irqsave.remove_wait_queue.__xfs_log_force_lsn.xfs_file_fsync
0.63 ± 5% +3.8 4.39 ± 27% perf-profile.calltrace.cycles-pp._raw_spin_lock_irqsave.remove_wait_queue.__xfs_log_force_lsn.xfs_file_fsync.xfs_file_write_iter
0.63 ± 5% +3.8 4.40 ± 27% perf-profile.calltrace.cycles-pp.remove_wait_queue.__xfs_log_force_lsn.xfs_file_fsync.xfs_file_write_iter.__vfs_write
1.26 ± 5% +5.3 6.55 ± 27% perf-profile.calltrace.cycles-pp.native_queued_spin_lock_slowpath._raw_spin_lock.__xfs_log_force_lsn.xfs_file_fsync.xfs_file_write_iter
1.27 ± 5% +5.3 6.55 ± 27% perf-profile.calltrace.cycles-pp._raw_spin_lock.__xfs_log_force_lsn.xfs_file_fsync.xfs_file_write_iter.__vfs_write
1.30 ± 4% +8.4 9.72 ± 9% perf-profile.calltrace.cycles-pp.intel_idle.cpuidle_enter_state.do_idle.cpu_startup_entry.start_secondary
1.33 ± 4% +8.9 10.26 ± 9% perf-profile.calltrace.cycles-pp.cpuidle_enter_state.do_idle.cpu_startup_entry.start_secondary.secondary_startup_64
2.28 ± 2% +9.1 11.36 ± 27% perf-profile.calltrace.cycles-pp.__xfs_log_force_lsn.xfs_file_fsync.xfs_file_write_iter.__vfs_write.vfs_write
1.59 ± 4% +10.4 11.97 ± 9% perf-profile.calltrace.cycles-pp.do_idle.cpu_startup_entry.start_secondary.secondary_startup_64
1.59 ± 4% +10.4 11.98 ± 9% perf-profile.calltrace.cycles-pp.cpu_startup_entry.start_secondary.secondary_startup_64
1.59 ± 4% +10.4 11.98 ± 9% perf-profile.calltrace.cycles-pp.start_secondary.secondary_startup_64
1.63 ± 4% +10.8 12.47 ± 8% perf-profile.calltrace.cycles-pp.secondary_startup_64
0.00 +57.7 57.66 ± 10% perf-profile.calltrace.cycles-pp.native_queued_spin_lock_slowpath._raw_spin_lock_irqsave.prepare_to_wait_event.raid1_write_request.raid1_make_request
0.00 +57.7 57.73 ± 10% perf-profile.calltrace.cycles-pp._raw_spin_lock_irqsave.prepare_to_wait_event.raid1_write_request.raid1_make_request.md_handle_request
0.05 ±299% +57.8 57.85 ± 9% perf-profile.calltrace.cycles-pp.prepare_to_wait_event.raid1_write_request.raid1_make_request.md_handle_request.md_make_request
0.19 ±154% +62.5 62.73 ± 8% perf-profile.calltrace.cycles-pp.raid1_write_request.raid1_make_request.md_handle_request.md_make_request.generic_make_request
0.19 ±154% +62.6 62.76 ± 8% perf-profile.calltrace.cycles-pp.md_handle_request.md_make_request.generic_make_request.submit_bio.xfs_submit_ioend
0.19 ±154% +62.6 62.79 ± 8% perf-profile.calltrace.cycles-pp.md_make_request.generic_make_request.submit_bio.xfs_submit_ioend.xfs_vm_writepages
0.20 ±154% +62.6 62.81 ± 8% perf-profile.calltrace.cycles-pp.generic_make_request.submit_bio.xfs_submit_ioend.xfs_vm_writepages.do_writepages
0.20 ±154% +62.6 62.81 ± 8% perf-profile.calltrace.cycles-pp.submit_bio.xfs_submit_ioend.xfs_vm_writepages.do_writepages.__filemap_fdatawrite_range
0.20 ±154% +62.6 62.82 ± 8% perf-profile.calltrace.cycles-pp.xfs_submit_ioend.xfs_vm_writepages.do_writepages.__filemap_fdatawrite_range.file_write_and_wait_range
0.29 ±125% +62.8 63.09 ± 8% perf-profile.calltrace.cycles-pp.xfs_vm_writepages.do_writepages.__filemap_fdatawrite_range.file_write_and_wait_range.xfs_file_fsync
0.29 ±126% +62.8 63.10 ± 8% perf-profile.calltrace.cycles-pp.do_writepages.__filemap_fdatawrite_range.file_write_and_wait_range.xfs_file_fsync.xfs_file_write_iter
0.29 ±125% +62.8 63.11 ± 8% perf-profile.calltrace.cycles-pp.__filemap_fdatawrite_range.file_write_and_wait_range.xfs_file_fsync.xfs_file_write_iter.__vfs_write
0.62 ± 41% +62.9 63.52 ± 7% perf-profile.calltrace.cycles-pp.file_write_and_wait_range.xfs_file_fsync.xfs_file_write_iter.__vfs_write.vfs_write
88.51 -88.2 0.26 ± 19% perf-profile.children.cycles-pp.md_flush_request
87.57 -87.2 0.35 ± 19% perf-profile.children.cycles-pp.submit_bio_wait
87.59 -87.2 0.39 ± 19% perf-profile.children.cycles-pp.blkdev_issue_flush
83.26 -83.2 0.02 ±123% perf-profile.children.cycles-pp._raw_spin_lock_irq
88.85 -25.7 63.11 ± 8% perf-profile.children.cycles-pp.md_make_request
88.90 -25.7 63.17 ± 8% perf-profile.children.cycles-pp.submit_bio
88.83 -24.5 64.31 ± 8% perf-profile.children.cycles-pp.raid1_make_request
88.84 -24.5 64.33 ± 8% perf-profile.children.cycles-pp.md_handle_request
89.38 -23.5 65.92 ± 7% perf-profile.children.cycles-pp.generic_make_request
89.90 -13.4 76.51 ± 2% perf-profile.children.cycles-pp.native_queued_spin_lock_slowpath
92.79 -13.0 79.76 perf-profile.children.cycles-pp.xfs_file_fsync
93.08 -12.6 80.49 perf-profile.children.cycles-pp.xfs_file_write_iter
93.09 -12.6 80.54 perf-profile.children.cycles-pp.__vfs_write
93.13 -12.5 80.60 perf-profile.children.cycles-pp.vfs_write
93.13 -12.5 80.61 perf-profile.children.cycles-pp.ksys_write
93.22 -12.4 80.83 perf-profile.children.cycles-pp.do_syscall_64
93.22 -12.4 80.83 perf-profile.children.cycles-pp.entry_SYSCALL_64_after_hwframe
3.40 ± 2% -1.4 1.97 ± 8% perf-profile.children.cycles-pp.worker_thread
3.33 ± 2% -1.4 1.96 ± 9% perf-profile.children.cycles-pp.process_one_work
1.03 ± 7% -1.0 0.07 ± 37% perf-profile.children.cycles-pp.xlog_cil_force_lsn
1.69 ± 2% -0.7 0.96 ± 4% perf-profile.children.cycles-pp.reschedule_interrupt
1.66 ± 2% -0.7 0.94 ± 4% perf-profile.children.cycles-pp.scheduler_ipi
1.13 ± 2% -0.7 0.47 ± 11% perf-profile.children.cycles-pp.finish_wait
0.54 ± 8% -0.4 0.10 ± 38% perf-profile.children.cycles-pp.xlog_cil_push
0.49 ± 9% -0.4 0.09 ± 35% perf-profile.children.cycles-pp.xlog_write
0.10 ± 8% -0.1 0.04 ± 67% perf-profile.children.cycles-pp.flush_work
0.20 ± 5% -0.0 0.16 ± 11% perf-profile.children.cycles-pp.reweight_entity
0.06 ± 10% +0.0 0.10 ± 23% perf-profile.children.cycles-pp.brd_lookup_page
0.18 ± 5% +0.0 0.23 ± 13% perf-profile.children.cycles-pp.__update_load_avg_se
0.02 ±153% +0.1 0.07 ± 16% perf-profile.children.cycles-pp.delay_tsc
0.03 ±100% +0.1 0.08 ± 15% perf-profile.children.cycles-pp.find_next_bit
0.08 ± 5% +0.1 0.14 ± 14% perf-profile.children.cycles-pp.native_write_msr
0.29 ± 4% +0.1 0.36 ± 8% perf-profile.children.cycles-pp.__orc_find
0.40 ± 4% +0.1 0.46 ± 7% perf-profile.children.cycles-pp.dequeue_task_fair
0.11 ± 11% +0.1 0.18 ± 14% perf-profile.children.cycles-pp.__module_text_address
0.12 ± 8% +0.1 0.19 ± 13% perf-profile.children.cycles-pp.is_module_text_address
0.04 ± 50% +0.1 0.12 ± 19% perf-profile.children.cycles-pp.kmem_cache_alloc
0.00 +0.1 0.08 ± 11% perf-profile.children.cycles-pp.clear_page_erms
0.00 +0.1 0.08 ± 28% perf-profile.children.cycles-pp.__indirect_thunk_start
0.01 ±200% +0.1 0.10 ± 25% perf-profile.children.cycles-pp.xfs_trans_alloc
0.00 +0.1 0.09 ± 18% perf-profile.children.cycles-pp.md_wakeup_thread
0.00 +0.1 0.09 ± 26% perf-profile.children.cycles-pp.rebalance_domains
0.00 +0.1 0.09 ± 26% perf-profile.children.cycles-pp.get_next_timer_interrupt
0.00 +0.1 0.09 ± 20% perf-profile.children.cycles-pp.ktime_get
0.18 ± 4% +0.1 0.27 ± 12% perf-profile.children.cycles-pp.idle_cpu
0.20 ± 6% +0.1 0.30 ± 9% perf-profile.children.cycles-pp.unwind_get_return_address
0.16 ± 10% +0.1 0.25 ± 13% perf-profile.children.cycles-pp.__module_address
0.03 ±100% +0.1 0.13 ± 8% perf-profile.children.cycles-pp.brd_insert_page
0.06 ± 9% +0.1 0.16 ± 14% perf-profile.children.cycles-pp.task_tick_fair
0.08 ± 12% +0.1 0.18 ± 24% perf-profile.children.cycles-pp.bio_alloc_bioset
0.03 ± 81% +0.1 0.14 ± 27% perf-profile.children.cycles-pp.generic_make_request_checks
0.17 ± 7% +0.1 0.28 ± 11% perf-profile.children.cycles-pp.__kernel_text_address
0.11 ± 9% +0.1 0.22 ± 15% perf-profile.children.cycles-pp.wake_up_page_bit
0.16 ± 6% +0.1 0.27 ± 10% perf-profile.children.cycles-pp.kernel_text_address
0.00 +0.1 0.11 ± 11% perf-profile.children.cycles-pp.get_page_from_freelist
0.00 +0.1 0.11 ± 19% perf-profile.children.cycles-pp.perf_mux_hrtimer_handler
0.00 +0.1 0.11 ± 7% perf-profile.children.cycles-pp.__alloc_pages_nodemask
0.08 ± 10% +0.1 0.19 ± 22% perf-profile.children.cycles-pp.xfs_do_writepage
0.25 ± 4% +0.1 0.37 ± 10% perf-profile.children.cycles-pp.switch_mm_irqs_off
0.00 +0.1 0.12 ± 13% perf-profile.children.cycles-pp.switch_mm
0.08 ± 38% +0.1 0.20 ± 19% perf-profile.children.cycles-pp.io_serial_in
0.18 ± 5% +0.1 0.31 ± 7% perf-profile.children.cycles-pp.dequeue_entity
0.00 +0.1 0.13 ± 26% perf-profile.children.cycles-pp.tick_nohz_next_event
0.06 ± 11% +0.1 0.19 ± 19% perf-profile.children.cycles-pp.mempool_alloc
0.32 ± 5% +0.1 0.45 ± 6% perf-profile.children.cycles-pp.orc_find
0.15 ± 10% +0.1 0.29 ± 19% perf-profile.children.cycles-pp.xfs_destroy_ioend
0.15 ± 11% +0.1 0.30 ± 18% perf-profile.children.cycles-pp.call_bio_endio
0.08 ± 17% +0.2 0.23 ± 25% perf-profile.children.cycles-pp.xlog_state_done_syncing
0.00 +0.2 0.15 ± 22% perf-profile.children.cycles-pp.tick_nohz_get_sleep_length
0.12 ± 8% +0.2 0.27 ± 23% perf-profile.children.cycles-pp.write_cache_pages
0.10 ± 16% +0.2 0.26 ± 16% perf-profile.children.cycles-pp.wait_for_xmitr
0.10 ± 19% +0.2 0.25 ± 14% perf-profile.children.cycles-pp.serial8250_console_putchar
0.10 ± 17% +0.2 0.26 ± 13% perf-profile.children.cycles-pp.uart_console_write
0.10 ± 16% +0.2 0.26 ± 15% perf-profile.children.cycles-pp.serial8250_console_write
0.11 ± 15% +0.2 0.27 ± 15% perf-profile.children.cycles-pp.console_unlock
0.09 ± 9% +0.2 0.26 ± 12% perf-profile.children.cycles-pp.scheduler_tick
0.10 ± 18% +0.2 0.28 ± 15% perf-profile.children.cycles-pp.irq_work_run_list
0.10 ± 15% +0.2 0.28 ± 14% perf-profile.children.cycles-pp.xlog_state_do_callback
0.09 ± 12% +0.2 0.27 ± 16% perf-profile.children.cycles-pp.irq_work_run
0.09 ± 12% +0.2 0.27 ± 16% perf-profile.children.cycles-pp.printk
0.09 ± 12% +0.2 0.27 ± 16% perf-profile.children.cycles-pp.vprintk_emit
0.09 ± 12% +0.2 0.27 ± 17% perf-profile.children.cycles-pp.irq_work_interrupt
0.09 ± 12% +0.2 0.27 ± 17% perf-profile.children.cycles-pp.smp_irq_work_interrupt
0.00 +0.2 0.18 ± 16% perf-profile.children.cycles-pp.poll_idle
0.30 ± 4% +0.2 0.49 ± 11% perf-profile.children.cycles-pp.update_load_avg
1.39 ± 2% +0.2 1.59 ± 6% perf-profile.children.cycles-pp.__save_stack_trace
1.43 +0.2 1.65 ± 6% perf-profile.children.cycles-pp.save_stack_trace_tsk
0.14 ± 13% +0.2 0.36 ± 13% perf-profile.children.cycles-pp.update_process_times
0.00 +0.2 0.23 ± 22% perf-profile.children.cycles-pp.find_busiest_group
0.22 ± 6% +0.2 0.45 ± 18% perf-profile.children.cycles-pp.brd_do_bvec
0.14 ± 13% +0.2 0.38 ± 14% perf-profile.children.cycles-pp.tick_sched_handle
0.10 ± 8% +0.2 0.34 ± 26% perf-profile.children.cycles-pp.xfs_log_commit_cil
0.07 ± 10% +0.3 0.33 ± 23% perf-profile.children.cycles-pp.io_schedule
0.03 ± 83% +0.3 0.29 ± 27% perf-profile.children.cycles-pp.__softirqentry_text_start
0.11 ± 5% +0.3 0.36 ± 25% perf-profile.children.cycles-pp.__xfs_trans_commit
0.06 ± 36% +0.3 0.31 ± 26% perf-profile.children.cycles-pp.irq_exit
0.08 ± 9% +0.3 0.35 ± 23% perf-profile.children.cycles-pp.wait_on_page_bit_common
0.15 ± 12% +0.3 0.42 ± 14% perf-profile.children.cycles-pp.tick_sched_timer
0.10 ± 11% +0.3 0.39 ± 22% perf-profile.children.cycles-pp.__filemap_fdatawait_range
0.06 ± 12% +0.3 0.37 ± 9% perf-profile.children.cycles-pp.schedule_idle
0.02 ±153% +0.3 0.34 ± 17% perf-profile.children.cycles-pp.menu_select
0.17 ± 5% +0.3 0.49 ± 22% perf-profile.children.cycles-pp.xfs_vn_update_time
0.19 ± 12% +0.3 0.51 ± 18% perf-profile.children.cycles-pp.xlog_iodone
0.18 ± 5% +0.3 0.51 ± 22% perf-profile.children.cycles-pp.file_update_time
0.18 ± 5% +0.3 0.51 ± 21% perf-profile.children.cycles-pp.xfs_file_aio_write_checks
0.21 ± 11% +0.4 0.60 ± 15% perf-profile.children.cycles-pp.__hrtimer_run_queues
0.26 ± 6% +0.4 0.69 ± 16% perf-profile.children.cycles-pp.pick_next_task_fair
1.20 ± 2% +0.4 1.64 ± 10% perf-profile.children.cycles-pp.schedule
0.28 ± 5% +0.4 0.72 ± 21% perf-profile.children.cycles-pp.xfs_file_buffered_aio_write
0.00 +0.4 0.44 ± 22% perf-profile.children.cycles-pp.load_balance
0.25 ± 8% +0.5 0.74 ± 15% perf-profile.children.cycles-pp.hrtimer_interrupt
1.30 ± 2% +0.7 2.00 ± 9% perf-profile.children.cycles-pp.__schedule
0.31 ± 8% +0.8 1.09 ± 16% perf-profile.children.cycles-pp.smp_apic_timer_interrupt
0.31 ± 8% +0.8 1.09 ± 16% perf-profile.children.cycles-pp.apic_timer_interrupt
3.92 ± 2% +0.9 4.79 ± 6% perf-profile.children.cycles-pp.ret_from_fork
3.92 ± 2% +0.9 4.79 ± 6% perf-profile.children.cycles-pp.kthread
0.69 ± 2% +0.9 1.64 ± 23% perf-profile.children.cycles-pp.xlog_wait
0.08 ± 13% +1.2 1.27 ± 17% perf-profile.children.cycles-pp.submit_flushes
0.16 ± 9% +1.6 1.74 ± 4% perf-profile.children.cycles-pp._raw_spin_unlock_irqrestore
0.17 ± 9% +2.0 2.16 ± 10% perf-profile.children.cycles-pp.raid_end_bio_io
0.21 ± 6% +2.0 2.25 ± 10% perf-profile.children.cycles-pp.raid1_end_write_request
2.24 ± 4% +2.2 4.44 ± 15% perf-profile.children.cycles-pp.xfs_log_force_lsn
0.46 ± 6% +2.3 2.73 ± 7% perf-profile.children.cycles-pp.brd_make_request
0.51 ± 6% +2.3 2.81 ± 7% perf-profile.children.cycles-pp.md_thread
0.49 ± 6% +2.3 2.81 ± 7% perf-profile.children.cycles-pp.raid1d
0.49 ± 6% +2.3 2.81 ± 7% perf-profile.children.cycles-pp.flush_pending_writes
0.49 ± 6% +2.3 2.81 ± 7% perf-profile.children.cycles-pp.flush_bio_list
1.80 ± 3% +5.6 7.44 ± 27% perf-profile.children.cycles-pp._raw_spin_lock
2.12 ± 4% +5.8 7.97 ± 20% perf-profile.children.cycles-pp.remove_wait_queue
1.33 ± 4% +8.8 10.12 ± 8% perf-profile.children.cycles-pp.intel_idle
1.37 ± 4% +9.3 10.71 ± 8% perf-profile.children.cycles-pp.cpuidle_enter_state
1.59 ± 4% +10.4 11.98 ± 9% perf-profile.children.cycles-pp.start_secondary
1.63 ± 4% +10.8 12.47 ± 8% perf-profile.children.cycles-pp.secondary_startup_64
1.63 ± 4% +10.8 12.47 ± 8% perf-profile.children.cycles-pp.cpu_startup_entry
1.63 ± 4% +10.9 12.49 ± 8% perf-profile.children.cycles-pp.do_idle
3.48 +12.2 15.72 ± 23% perf-profile.children.cycles-pp.__xfs_log_force_lsn
1.36 ± 12% +57.8 59.12 ± 10% perf-profile.children.cycles-pp.prepare_to_wait_event
0.43 ± 38% +62.4 62.82 ± 8% perf-profile.children.cycles-pp.xfs_submit_ioend
0.55 ± 29% +62.5 63.10 ± 8% perf-profile.children.cycles-pp.xfs_vm_writepages
0.55 ± 30% +62.5 63.10 ± 8% perf-profile.children.cycles-pp.do_writepages
0.55 ± 29% +62.6 63.11 ± 8% perf-profile.children.cycles-pp.__filemap_fdatawrite_range
0.66 ± 25% +62.9 63.52 ± 7% perf-profile.children.cycles-pp.file_write_and_wait_range
0.39 ± 43% +63.6 64.02 ± 8% perf-profile.children.cycles-pp.raid1_write_request
5.43 ± 3% +64.2 69.64 ± 5% perf-profile.children.cycles-pp._raw_spin_lock_irqsave
89.86 -13.5 76.31 ± 2% perf-profile.self.cycles-pp.native_queued_spin_lock_slowpath
0.14 ± 8% -0.0 0.09 ± 19% perf-profile.self.cycles-pp.md_flush_request
0.10 ± 12% -0.0 0.07 ± 21% perf-profile.self.cycles-pp.account_entity_enqueue
0.06 ± 7% +0.0 0.08 ± 12% perf-profile.self.cycles-pp.pick_next_task_fair
0.05 ± 12% +0.0 0.08 ± 18% perf-profile.self.cycles-pp.___perf_sw_event
0.15 ± 6% +0.0 0.18 ± 9% perf-profile.self.cycles-pp.__update_load_avg_se
0.17 ± 4% +0.0 0.22 ± 10% perf-profile.self.cycles-pp.__schedule
0.10 ± 11% +0.1 0.15 ± 11% perf-profile.self.cycles-pp._raw_spin_lock
0.02 ±153% +0.1 0.07 ± 16% perf-profile.self.cycles-pp.delay_tsc
0.02 ±152% +0.1 0.07 ± 23% perf-profile.self.cycles-pp.set_next_entity
0.03 ±100% +0.1 0.08 ± 15% perf-profile.self.cycles-pp.find_next_bit
0.08 ± 5% +0.1 0.14 ± 14% perf-profile.self.cycles-pp.native_write_msr
0.01 ±200% +0.1 0.07 ± 23% perf-profile.self.cycles-pp.kmem_cache_alloc
0.29 ± 4% +0.1 0.36 ± 8% perf-profile.self.cycles-pp.__orc_find
0.14 ± 7% +0.1 0.21 ± 12% perf-profile.self.cycles-pp.switch_mm_irqs_off
0.00 +0.1 0.08 ± 11% perf-profile.self.cycles-pp.clear_page_erms
0.00 +0.1 0.08 ± 28% perf-profile.self.cycles-pp.__indirect_thunk_start
0.00 +0.1 0.08 ± 20% perf-profile.self.cycles-pp.md_wakeup_thread
0.34 ± 6% +0.1 0.43 ± 12% perf-profile.self.cycles-pp._raw_spin_lock_irqsave
0.18 ± 4% +0.1 0.27 ± 12% perf-profile.self.cycles-pp.idle_cpu
0.16 ± 10% +0.1 0.25 ± 13% perf-profile.self.cycles-pp.__module_address
0.06 ± 11% +0.1 0.17 ± 14% perf-profile.self.cycles-pp._raw_spin_unlock_irqrestore
0.08 ± 38% +0.1 0.20 ± 19% perf-profile.self.cycles-pp.io_serial_in
0.18 ± 5% +0.1 0.32 ± 15% perf-profile.self.cycles-pp.update_load_avg
0.00 +0.1 0.15 ± 17% perf-profile.self.cycles-pp.poll_idle
0.00 +0.2 0.15 ± 16% perf-profile.self.cycles-pp.menu_select
0.00 +0.2 0.18 ± 24% perf-profile.self.cycles-pp.find_busiest_group
0.02 ±152% +0.3 0.35 ± 21% perf-profile.self.cycles-pp.raid1_write_request
1.33 ± 4% +8.8 10.12 ± 8% perf-profile.self.cycles-pp.intel_idle
aim7.jobs-per-min
1700 +-+------------------------------------------------------------------+
|+ ++++++ :+ ++++ ++++ +++ ++++++ + + ++++++++++++ ++ ++|
1600 +-+ + +++ + +++++ ++.++ + ++ ++ + ++ |
| |
| |
1500 +-+ |
| |
1400 +-+ |
| |
1300 +-+ |
| |
O OO OO O O O |
1200 +OO OOOOOOOOO OO OOOOOOOOOOOOOO OOOOOOOOO O |
| |
1100 +-+------------------------------------------------------------------+
[*] bisect-good sample
[O] bisect-bad sample
Disclaimer:
Results have been estimated based on internal Intel analysis and are provided
for informational purposes only. Any difference in system hardware or software
design or configuration may affect actual performance.
Thanks,
Xiaolong
View attachment "config-4.17.0-rc1-00189-g5a409b4" of type "text/plain" (164401 bytes)
View attachment "job-script" of type "text/plain" (7703 bytes)
View attachment "job.yaml" of type "text/plain" (5344 bytes)
View attachment "reproduce" of type "text/plain" (1026 bytes)
Powered by blists - more mailing lists