[<prev] [next>] [day] [month] [year] [list]
Message-ID: <20210712145638.GC13620@xsang-OptiPlex-9020>
Date: Mon, 12 Jul 2021 22:56:38 +0800
From: kernel test robot <oliver.sang@...el.com>
To: Dave Chinner <dchinner@...hat.com>
Cc: "Darrick J. Wong" <djwong@...nel.org>,
Chandan Babu R <chandanrlinux@...il.com>,
Brian Foster <bfoster@...hat.com>,
Allison Henderson <allison.henderson@...cle.com>,
LKML <linux-kernel@...r.kernel.org>, lkp@...ts.01.org,
lkp@...el.com, ying.huang@...el.com, feng.tang@...el.com,
zhengjun.xing@...ux.intel.com
Subject: [xfs] a79b28c284: fsmark.files_per_sec -4.6% regression
Greeting,
FYI, we noticed a -4.6% regression of fsmark.files_per_sec due to commit:
commit: a79b28c284fd910bb291dbf307a26f4d432e88f3 ("xfs: separate CIL commit record IO")
https://git.kernel.org/cgit/linux/kernel/git/torvalds/linux.git master
in testcase: fsmark
on test machine: 96 threads 2 sockets Intel(R) Xeon(R) Gold 6252 CPU @ 2.10GHz with 512G memory
with following parameters:
iterations: 1x
nr_threads: 32t
disk: 1SSD
fs: xfs
filesize: 8K
test_size: 400M
sync_method: fsyncBeforeClose
nr_directories: 16d
nr_files_per_directory: 256fpd
cpufreq_governor: performance
ucode: 0x5003006
test-description: The fsmark is a file system benchmark to test synchronous write workloads, for example, mail servers workload.
test-url: https://sourceforge.net/projects/fsmark/
In addition to that, the commit also has significant impact on the following tests:
+------------------+---------------------------------------------------------------------------------+
| testcase: change | aim7: aim7.jobs-per-min 22.0% improvement |
| test machine | 88 threads 2 sockets Intel(R) Xeon(R) Gold 6238M CPU @ 2.10GHz with 128G memory |
| test parameters | cpufreq_governor=performance |
| | disk=4BRD_12G |
| | fs=xfs |
| | load=300 |
| | md=RAID0 |
| | test=sync_disk_rw |
| | ucode=0x5003006 |
+------------------+---------------------------------------------------------------------------------+
If you fix the issue, kindly add following tag
Reported-by: kernel test robot <oliver.sang@...el.com>
Details are as below:
-------------------------------------------------------------------------------------------------->
To reproduce:
git clone https://github.com/intel/lkp-tests.git
cd lkp-tests
bin/lkp install job.yaml # job file is attached in this email
bin/lkp split-job --compatible job.yaml # generate the yaml file for lkp run
bin/lkp run generated-yaml-file
=========================================================================================
compiler/cpufreq_governor/disk/filesize/fs/iterations/kconfig/nr_directories/nr_files_per_directory/nr_threads/rootfs/sync_method/tbox_group/test_size/testcase/ucode:
gcc-9/performance/1SSD/8K/xfs/1x/x86_64-rhel-8.3/16d/256fpd/32t/debian-10.4-x86_64-20200603.cgz/fsyncBeforeClose/lkp-csl-2sp7/400M/fsmark/0x5003006
commit:
18842e0a4f ("xfs: Fix 64-bit division on 32-bit in xlog_state_switch_iclogs()")
a79b28c284 ("xfs: separate CIL commit record IO")
18842e0a4f48564b a79b28c284fd910bb291dbf307a
---------------- ---------------------------
%stddev %change %stddev
\ | \
16388 -4.6% 15631 ± 2% fsmark.files_per_sec
19379 ± 6% -31.7% 13238 ± 3% fsmark.time.involuntary_context_switches
294578 +11.5% 328546 fsmark.time.voluntary_context_switches
11335 ± 11% +67.3% 18968 ± 56% cpuidle.POLL.usage
2860 ±199% -98.4% 45.67 ±127% softirqs.CPU72.TIMER
114218 ± 7% -11.2% 101430 vmstat.io.bo
23503 ± 12% +19.7% 28140 ± 9% numa-vmstat.node0.nr_slab_unreclaimable
588.67 ± 39% -44.8% 325.17 ± 50% numa-vmstat.node1.nr_page_table_pages
94014 ± 12% +19.7% 112564 ± 9% numa-meminfo.node0.SUnreclaim
164603 ± 67% -70.4% 48754 ± 74% numa-meminfo.node1.Inactive
2357 ± 39% -44.8% 1301 ± 50% numa-meminfo.node1.PageTables
70708 +0.7% 71212 proc-vmstat.nr_inactive_anon
18040 -3.2% 17455 proc-vmstat.nr_kernel_stack
70708 +0.7% 71212 proc-vmstat.nr_zone_inactive_anon
370332 +2.0% 377771 proc-vmstat.pgalloc_normal
157090 ± 41% +34.6% 211411 proc-vmstat.pgfree
3271411 ± 3% -8.3% 3001095 ± 3% perf-stat.i.iTLB-load-misses
2245484 ± 53% -64.3% 802425 ± 66% perf-stat.i.node-load-misses
56.94 ± 24% -22.2 34.72 ± 29% perf-stat.i.node-store-miss-rate%
1090824 ± 57% -65.7% 374199 ± 70% perf-stat.i.node-store-misses
59.80 ± 26% -23.7 36.06 ± 32% perf-stat.overall.node-store-miss-rate%
0.99 ± 45% -48.8% 0.51 ± 31% perf-stat.ps.major-faults
1716083 ± 54% -64.0% 618497 ± 70% perf-stat.ps.node-load-misses
834091 ± 58% -65.4% 288523 ± 74% perf-stat.ps.node-store-misses
487.67 ± 17% -35.0% 317.17 ± 12% slabinfo.biovec-max.active_objs
487.67 ± 17% -35.0% 317.17 ± 12% slabinfo.biovec-max.num_objs
8026 ± 5% +60.8% 12901 ± 3% slabinfo.kmalloc-1k.active_objs
252.33 ± 5% +61.1% 406.50 ± 3% slabinfo.kmalloc-1k.active_slabs
8086 ± 5% +61.0% 13017 ± 3% slabinfo.kmalloc-1k.num_objs
252.33 ± 5% +61.1% 406.50 ± 3% slabinfo.kmalloc-1k.num_slabs
2465 ± 6% -21.0% 1946 ± 14% slabinfo.pool_workqueue.active_objs
2475 ± 6% -20.9% 1958 ± 14% slabinfo.pool_workqueue.num_objs
18532 ± 7% -12.6% 16189 slabinfo.xfs_ili.active_objs
18570 ± 7% -12.6% 16222 slabinfo.xfs_ili.num_objs
57483 ± 5% -10.4% 51530 ± 3% interrupts.CAL:Function_call_interrupts
818.17 ± 45% -38.5% 503.00 interrupts.CPU11.CAL:Function_call_interrupts
572.83 ± 10% -12.6% 500.67 interrupts.CPU15.CAL:Function_call_interrupts
667.17 ± 20% -29.4% 470.83 ± 14% interrupts.CPU17.CAL:Function_call_interrupts
623.17 ± 12% -18.1% 510.50 ± 3% interrupts.CPU18.CAL:Function_call_interrupts
588.83 ± 2% -14.2% 505.00 interrupts.CPU19.CAL:Function_call_interrupts
606.33 ± 7% -17.3% 501.17 interrupts.CPU21.CAL:Function_call_interrupts
907.00 ± 30% -32.8% 609.67 ± 17% interrupts.CPU25.CAL:Function_call_interrupts
588.67 ± 5% -12.2% 516.67 ± 3% interrupts.CPU3.CAL:Function_call_interrupts
604.00 ± 13% -16.0% 507.50 ± 5% interrupts.CPU31.CAL:Function_call_interrupts
573.50 ± 3% -16.0% 481.67 ± 15% interrupts.CPU4.CAL:Function_call_interrupts
617.17 ± 15% -17.7% 507.83 ± 4% interrupts.CPU44.CAL:Function_call_interrupts
595.00 ± 4% -13.4% 515.33 ± 4% interrupts.CPU49.CAL:Function_call_interrupts
572.17 ± 4% -8.5% 523.67 ± 5% interrupts.CPU52.CAL:Function_call_interrupts
581.83 ± 5% -11.6% 514.17 ± 4% interrupts.CPU53.CAL:Function_call_interrupts
578.67 -10.2% 519.50 ± 4% interrupts.CPU54.CAL:Function_call_interrupts
581.00 ± 3% -12.4% 508.67 interrupts.CPU56.CAL:Function_call_interrupts
582.17 ± 3% -11.8% 513.67 interrupts.CPU57.CAL:Function_call_interrupts
581.67 ± 4% -13.2% 504.83 interrupts.CPU59.CAL:Function_call_interrupts
630.83 ± 18% -20.5% 501.50 interrupts.CPU61.CAL:Function_call_interrupts
633.00 ± 26% -19.5% 509.67 ± 3% interrupts.CPU64.CAL:Function_call_interrupts
604.00 ± 14% -16.6% 503.83 ± 4% interrupts.CPU75.CAL:Function_call_interrupts
603.67 ± 12% -16.0% 507.33 ± 4% interrupts.CPU78.CAL:Function_call_interrupts
602.17 ± 13% -16.0% 506.00 ± 4% interrupts.CPU80.CAL:Function_call_interrupts
618.33 ± 13% -14.3% 530.17 ± 7% interrupts.CPU90.CAL:Function_call_interrupts
616.00 ± 13% -15.2% 522.67 ± 5% interrupts.CPU91.CAL:Function_call_interrupts
57.10 ± 15% -16.4 40.71 ± 26% perf-profile.calltrace.cycles-pp.secondary_startup_64_no_verify
57.10 ± 15% -16.4 40.71 ± 26% perf-profile.calltrace.cycles-pp.start_secondary.secondary_startup_64_no_verify
57.10 ± 15% -16.4 40.71 ± 26% perf-profile.calltrace.cycles-pp.cpu_startup_entry.start_secondary.secondary_startup_64_no_verify
57.10 ± 15% -16.4 40.71 ± 26% perf-profile.calltrace.cycles-pp.do_idle.cpu_startup_entry.start_secondary.secondary_startup_64_no_verify
14.67 ± 60% -10.1 4.57 ±148% perf-profile.calltrace.cycles-pp.seq_read_iter.proc_reg_read_iter.new_sync_read.vfs_read.ksys_read
14.67 ± 60% -10.1 4.57 ±148% perf-profile.calltrace.cycles-pp.proc_reg_read_iter.new_sync_read.vfs_read.ksys_read.do_syscall_64
14.67 ± 60% -8.7 6.02 ±161% perf-profile.calltrace.cycles-pp.new_sync_read.vfs_read.ksys_read.do_syscall_64.entry_SYSCALL_64_after_hwframe
12.00 ± 86% -7.4 4.56 ±148% perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.read
12.00 ± 86% -7.4 4.56 ±148% perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.read
12.00 ± 86% -7.4 4.56 ±148% perf-profile.calltrace.cycles-pp.ksys_read.do_syscall_64.entry_SYSCALL_64_after_hwframe.read
12.00 ± 86% -7.4 4.56 ±148% perf-profile.calltrace.cycles-pp.vfs_read.ksys_read.do_syscall_64.entry_SYSCALL_64_after_hwframe.read
12.00 ± 86% -7.4 4.56 ±148% perf-profile.calltrace.cycles-pp.read
5.45 ±104% -4.7 0.72 ±223% perf-profile.calltrace.cycles-pp.arch_show_interrupts.seq_read_iter.proc_reg_read_iter.new_sync_read.vfs_read
6.14 ±108% -4.7 1.45 ±223% perf-profile.calltrace.cycles-pp.ksys_read.do_syscall_64.entry_SYSCALL_64_after_hwframe
6.14 ±108% -4.7 1.45 ±223% perf-profile.calltrace.cycles-pp.vfs_read.ksys_read.do_syscall_64.entry_SYSCALL_64_after_hwframe
4.75 ±104% -3.6 1.19 ±223% perf-profile.calltrace.cycles-pp.__x64_sys_getdents64.do_syscall_64.entry_SYSCALL_64_after_hwframe
4.75 ±104% -3.6 1.19 ±223% perf-profile.calltrace.cycles-pp.iterate_dir.__x64_sys_getdents64.do_syscall_64.entry_SYSCALL_64_after_hwframe
6.55 ±100% -2.7 3.84 ±143% perf-profile.calltrace.cycles-pp.show_interrupts.seq_read_iter.proc_reg_read_iter.new_sync_read.vfs_read
57.10 ± 15% -16.4 40.71 ± 26% perf-profile.children.cycles-pp.secondary_startup_64_no_verify
57.10 ± 15% -16.4 40.71 ± 26% perf-profile.children.cycles-pp.start_secondary
57.10 ± 15% -16.4 40.71 ± 26% perf-profile.children.cycles-pp.cpu_startup_entry
57.10 ± 15% -16.4 40.71 ± 26% perf-profile.children.cycles-pp.do_idle
18.14 ± 62% -12.8 5.29 ±155% perf-profile.children.cycles-pp.seq_read_iter
18.14 ± 62% -12.1 6.02 ±161% perf-profile.children.cycles-pp.ksys_read
18.14 ± 62% -12.1 6.02 ±161% perf-profile.children.cycles-pp.vfs_read
14.67 ± 60% -10.1 4.56 ±148% perf-profile.children.cycles-pp.proc_reg_read_iter
14.67 ± 60% -8.7 6.02 ±161% perf-profile.children.cycles-pp.new_sync_read
12.00 ± 86% -6.7 5.29 ±155% perf-profile.children.cycles-pp.read
5.45 ±104% -4.7 0.72 ±223% perf-profile.children.cycles-pp.arch_show_interrupts
5.75 ±105% -4.3 1.45 ±223% perf-profile.children.cycles-pp.vsnprintf
5.75 ±105% -3.6 2.17 ±223% perf-profile.children.cycles-pp.seq_vprintf
5.75 ±105% -3.6 2.17 ±223% perf-profile.children.cycles-pp.seq_printf
4.75 ±104% -3.6 1.19 ±223% perf-profile.children.cycles-pp.__x64_sys_getdents64
4.75 ±104% -3.6 1.19 ±223% perf-profile.children.cycles-pp.iterate_dir
6.55 ±100% -2.7 3.84 ±143% perf-profile.children.cycles-pp.show_interrupts
5.45 ±104% -5.4 0.00 perf-profile.self.cycles-pp.arch_show_interrupts
fsmark.files_per_sec
17000 +-------------------------------------------------------------------+
|+.+++ + :++.++ + +.++++.++++.+ +.+ + .++++. + .++ +.+ + .+ + |
16500 |-+ + + ++.+ :: :: + ++ + + + ::+. |
16000 |-+ O O OO O OO+O OOO OO O O+O O O + +|
| O O O O O O OOO O O OO OO O O|
15500 |O+ O O O O |
15000 |-+ O O O O O O O |
| O |
14500 |-+ |
14000 |-+ O |
| O |
13500 |-+ O |
13000 |-+ |
| O |
12500 +-------------------------------------------------------------------+
fsmark.time.voluntary_context_switches
350000 +------------------------------------------------------------------+
| O O |
340000 |-+ |
| |
| O O O O O |
330000 |OO O OO OOO O O OOOO OOOO OOOO OOOOO OOOO OOO OO OO OO OO|
| O O |
320000 |-+ O |
| |
310000 |-+ |
| |
| |
300000 |-+ |
|++.++++.++++.++++.++++.++++.++++.+++++.++++.++++.++++.++++.++++.++|
290000 +------------------------------------------------------------------+
[*] bisect-good sample
[O] bisect-bad sample
***************************************************************************************************
lkp-csl-2sp9: 88 threads 2 sockets Intel(R) Xeon(R) Gold 6238M CPU @ 2.10GHz with 128G memory
=========================================================================================
compiler/cpufreq_governor/disk/fs/kconfig/load/md/rootfs/tbox_group/test/testcase/ucode:
gcc-9/performance/4BRD_12G/xfs/x86_64-rhel-8.3/300/RAID0/debian-10.4-x86_64-20200603.cgz/lkp-csl-2sp9/sync_disk_rw/aim7/0x5003006
commit:
18842e0a4f ("xfs: Fix 64-bit division on 32-bit in xlog_state_switch_iclogs()")
a79b28c284 ("xfs: separate CIL commit record IO")
18842e0a4f48564b a79b28c284fd910bb291dbf307a
---------------- ---------------------------
%stddev %change %stddev
\ | \
13879 +22.0% 16929 aim7.jobs-per-min
129.73 -18.0% 106.37 aim7.time.elapsed_time
129.73 -18.0% 106.37 aim7.time.elapsed_time.max
1647556 +16.5% 1919576 aim7.time.involuntary_context_switches
41390 ± 3% -11.2% 36759 aim7.time.minor_page_faults
5461 -17.9% 4483 aim7.time.system_time
72997986 -10.5% 65359678 aim7.time.voluntary_context_switches
0.01 +0.0 0.02 ± 9% mpstat.cpu.all.iowait%
0.54 -0.1 0.48 ± 2% mpstat.cpu.all.usr%
2982166 -9.3% 2704857 numa-numastat.node1.local_node
3008831 -9.0% 2737175 numa-numastat.node1.numa_hit
58057 +10.9% 64387 ± 5% slabinfo.anon_vma_chain.active_objs
58142 +10.7% 64387 ± 5% slabinfo.anon_vma_chain.num_objs
168.11 -13.9% 144.77 uptime.boot
8419 -11.9% 7419 uptime.idle
558231 +21.8% 679782 vmstat.io.bo
3562019 -20.1% 2846094 vmstat.memory.cache
1098857 +12.1% 1231460 vmstat.system.cs
9.824e+08 +11.9% 1.099e+09 ± 2% cpuidle.C1.time
21821055 +11.0% 24216149 ± 2% cpuidle.C1.usage
3.872e+09 ± 10% -29.1% 2.745e+09 ± 15% cpuidle.C1E.time
46682234 -20.9% 36905264 ± 3% cpuidle.C1E.usage
1122822 +12.3% 1260895 cpuidle.POLL.usage
8.51 +3.1 11.59 turbostat.C1%
46681814 -20.9% 36904474 ± 3% turbostat.C1E
29377272 -16.7% 24460328 ± 3% turbostat.IRQ
64.33 -5.7% 60.67 turbostat.PkgTmp
55.96 +1.3% 56.68 turbostat.RAMWatt
335419 ± 2% -69.3% 102817 ± 8% meminfo.Active
335163 ± 2% -69.4% 102561 ± 8% meminfo.Active(anon)
131608 -13.7% 113590 meminfo.AnonHugePages
3403302 -20.7% 2700472 meminfo.Cached
1483499 -47.5% 778943 ± 2% meminfo.Committed_AS
977192 -48.9% 499429 ± 2% meminfo.Inactive
831379 -57.0% 357444 ± 3% meminfo.Inactive(anon)
364174 -83.5% 60025 ± 3% meminfo.Mapped
5288714 -13.5% 4573105 meminfo.Memused
888801 -78.6% 189797 ± 10% meminfo.Shmem
5365908 -13.0% 4670485 meminfo.max_used_kB
91803 ± 6% -78.3% 19879 ± 32% numa-meminfo.node0.Active
91632 ± 6% -78.5% 19709 ± 32% numa-meminfo.node0.Active(anon)
144688 ± 8% -70.1% 43249 ± 9% numa-meminfo.node0.Mapped
278561 ± 5% -79.3% 57643 ± 37% numa-meminfo.node0.Shmem
244509 -65.6% 84084 ± 2% numa-meminfo.node1.Active
244424 -65.6% 83999 ± 2% numa-meminfo.node1.Active(anon)
1475164 ± 64% -75.5% 361168 ± 29% numa-meminfo.node1.FilePages
555525 ± 16% -65.2% 193567 ± 32% numa-meminfo.node1.Inactive
482733 ± 19% -74.4% 123553 ± 50% numa-meminfo.node1.Inactive(anon)
221734 ± 5% -92.4% 16893 ± 11% numa-meminfo.node1.Mapped
2341429 ± 42% -51.0% 1147446 ± 25% numa-meminfo.node1.MemUsed
612619 ± 2% -78.0% 134759 ± 4% numa-meminfo.node1.Shmem
22915 ± 6% -78.5% 4927 ± 32% numa-vmstat.node0.nr_active_anon
36199 ± 8% -70.1% 10813 ± 9% numa-vmstat.node0.nr_mapped
69661 ± 5% -79.3% 14415 ± 36% numa-vmstat.node0.nr_shmem
22915 ± 6% -78.5% 4927 ± 32% numa-vmstat.node0.nr_zone_active_anon
7370 ± 5% -6.5% 6890 ± 3% numa-vmstat.node0.nr_zone_write_pending
61128 -65.6% 21003 ± 2% numa-vmstat.node1.nr_active_anon
368867 ± 64% -75.5% 90307 ± 29% numa-vmstat.node1.nr_file_pages
120723 ± 19% -74.4% 30895 ± 50% numa-vmstat.node1.nr_inactive_anon
55470 ± 6% -92.4% 4222 ± 11% numa-vmstat.node1.nr_mapped
153215 ± 2% -78.0% 33695 ± 4% numa-vmstat.node1.nr_shmem
61128 -65.6% 21002 ± 2% numa-vmstat.node1.nr_zone_active_anon
120723 ± 19% -74.4% 30894 ± 50% numa-vmstat.node1.nr_zone_inactive_anon
83793 ± 2% -69.4% 25645 ± 8% proc-vmstat.nr_active_anon
69383 -2.5% 67666 proc-vmstat.nr_anon_pages
850838 -20.7% 675125 proc-vmstat.nr_file_pages
207856 -57.0% 89363 ± 3% proc-vmstat.nr_inactive_anon
36450 -2.6% 35495 proc-vmstat.nr_inactive_file
91055 -83.5% 15006 ± 3% proc-vmstat.nr_mapped
222214 -78.6% 47456 ± 11% proc-vmstat.nr_shmem
83793 ± 2% -69.4% 25645 ± 8% proc-vmstat.nr_zone_active_anon
207856 -57.0% 89363 ± 3% proc-vmstat.nr_zone_inactive_anon
36450 -2.6% 35495 proc-vmstat.nr_zone_inactive_file
15021 ± 2% -5.8% 14149 proc-vmstat.nr_zone_write_pending
266167 -35.6% 171458 ± 11% proc-vmstat.numa_hint_faults
140448 ± 3% -37.3% 88003 ± 10% proc-vmstat.numa_hint_faults_local
5650544 -6.4% 5290494 ± 2% proc-vmstat.numa_hit
5570801 -6.5% 5210756 ± 2% proc-vmstat.numa_local
380199 -50.1% 189829 ± 8% proc-vmstat.numa_pte_updates
5717786 -5.7% 5391070 proc-vmstat.pgalloc_normal
808432 -25.4% 602788 ± 3% proc-vmstat.pgfault
29460 -13.0% 25626 ± 2% proc-vmstat.pgreuse
4.307e+09 +3.6% 4.464e+09 perf-stat.i.branch-instructions
39861920 +11.3% 44384422 perf-stat.i.branch-misses
23.30 +0.8 24.14 perf-stat.i.cache-miss-rate%
52855610 +17.7% 62197817 perf-stat.i.cache-misses
2.166e+08 +12.0% 2.426e+08 perf-stat.i.cache-references
1118434 +12.8% 1261487 perf-stat.i.context-switches
6.75 -3.9% 6.48 perf-stat.i.cpi
1.308e+11 +1.1% 1.322e+11 perf-stat.i.cpu-cycles
178676 +17.9% 210712 perf-stat.i.cpu-migrations
2593 ± 4% -12.7% 2265 ± 7% perf-stat.i.cycles-between-cache-misses
5.075e+09 +5.3% 5.342e+09 perf-stat.i.dTLB-loads
512902 ± 7% +17.5% 602445 ± 5% perf-stat.i.dTLB-store-misses
1.715e+09 +10.7% 1.899e+09 perf-stat.i.dTLB-stores
8968888 +12.4% 10078328 perf-stat.i.iTLB-load-misses
16507011 +15.0% 18977719 ± 3% perf-stat.i.iTLB-loads
1.925e+10 +4.6% 2.014e+10 perf-stat.i.instructions
2247 -4.8% 2139 perf-stat.i.instructions-per-iTLB-miss
1.49 +1.1% 1.50 perf-stat.i.metric.GHz
610.38 ± 2% +19.5% 729.31 ± 5% perf-stat.i.metric.K/sec
128.52 +5.6% 135.69 perf-stat.i.metric.M/sec
5939 -9.8% 5356 ± 2% perf-stat.i.minor-faults
19010898 +18.2% 22479012 perf-stat.i.node-load-misses
2079877 +11.3% 2314010 perf-stat.i.node-loads
8988994 +17.4% 10554292 perf-stat.i.node-store-misses
1813880 +13.9% 2065595 perf-stat.i.node-stores
5941 -9.8% 5359 ± 2% perf-stat.i.page-faults
11.25 +7.1% 12.05 perf-stat.overall.MPKI
0.93 +0.1 0.99 perf-stat.overall.branch-miss-rate%
24.41 +1.2 25.64 perf-stat.overall.cache-miss-rate%
6.79 -3.3% 6.57 perf-stat.overall.cpi
2474 -14.1% 2125 perf-stat.overall.cycles-between-cache-misses
2146 -6.9% 1998 perf-stat.overall.instructions-per-iTLB-miss
0.15 +3.4% 0.15 perf-stat.overall.ipc
4.274e+09 +3.5% 4.423e+09 perf-stat.ps.branch-instructions
39550311 +11.2% 43964111 perf-stat.ps.branch-misses
52450879 +17.5% 61628414 perf-stat.ps.cache-misses
2.149e+08 +11.9% 2.404e+08 perf-stat.ps.cache-references
1109849 +12.6% 1249872 perf-stat.ps.context-switches
177305 +17.8% 208787 perf-stat.ps.cpu-migrations
5.037e+09 +5.1% 5.293e+09 perf-stat.ps.dTLB-loads
508884 ± 7% +17.3% 597002 ± 5% perf-stat.ps.dTLB-store-misses
1.702e+09 +10.5% 1.881e+09 perf-stat.ps.dTLB-stores
8901107 +12.2% 9985139 perf-stat.ps.iTLB-load-misses
16380357 +14.8% 18802343 ± 3% perf-stat.ps.iTLB-loads
1.911e+10 +4.4% 1.995e+10 perf-stat.ps.instructions
5893 -10.0% 5301 ± 2% perf-stat.ps.minor-faults
18864869 +18.1% 22272948 perf-stat.ps.node-load-misses
2063954 +11.1% 2293240 perf-stat.ps.node-loads
8920023 +17.2% 10457333 perf-stat.ps.node-store-misses
1800221 +13.7% 2046553 perf-stat.ps.node-stores
5895 -10.0% 5304 ± 2% perf-stat.ps.page-faults
2.486e+12 -14.2% 2.134e+12 perf-stat.total.instructions
5157270 -6.3% 4834857 interrupts.CAL:Function_call_interrupts
258813 -19.8% 207639 ± 4% interrupts.CPU0.LOC:Local_timer_interrupts
258857 -20.0% 207130 ± 4% interrupts.CPU1.LOC:Local_timer_interrupts
258867 -19.9% 207436 ± 4% interrupts.CPU10.LOC:Local_timer_interrupts
258829 -19.8% 207511 ± 4% interrupts.CPU11.LOC:Local_timer_interrupts
258768 -19.8% 207624 ± 4% interrupts.CPU12.LOC:Local_timer_interrupts
258820 -19.8% 207594 ± 4% interrupts.CPU13.LOC:Local_timer_interrupts
258806 -19.8% 207625 ± 4% interrupts.CPU14.LOC:Local_timer_interrupts
258664 -19.6% 207905 ± 4% interrupts.CPU15.LOC:Local_timer_interrupts
6059 ± 2% -8.9% 5518 ± 2% interrupts.CPU15.RES:Rescheduling_interrupts
258859 -19.8% 207619 ± 4% interrupts.CPU16.LOC:Local_timer_interrupts
6175 ± 4% -10.3% 5540 ± 3% interrupts.CPU16.RES:Rescheduling_interrupts
258722 -19.7% 207647 ± 4% interrupts.CPU17.LOC:Local_timer_interrupts
258646 -19.7% 207580 ± 4% interrupts.CPU18.LOC:Local_timer_interrupts
258799 -19.8% 207564 ± 4% interrupts.CPU19.LOC:Local_timer_interrupts
258801 -19.8% 207563 ± 4% interrupts.CPU2.LOC:Local_timer_interrupts
258723 -19.8% 207565 ± 4% interrupts.CPU20.LOC:Local_timer_interrupts
258804 -19.8% 207615 ± 4% interrupts.CPU21.LOC:Local_timer_interrupts
57725 ± 2% -7.5% 53408 ± 4% interrupts.CPU22.CAL:Function_call_interrupts
258842 -20.3% 206199 ± 5% interrupts.CPU22.LOC:Local_timer_interrupts
57984 ± 2% -7.6% 53551 ± 3% interrupts.CPU23.CAL:Function_call_interrupts
258760 -20.3% 206156 ± 5% interrupts.CPU23.LOC:Local_timer_interrupts
58472 -8.1% 53738 ± 3% interrupts.CPU24.CAL:Function_call_interrupts
258969 -20.4% 206187 ± 5% interrupts.CPU24.LOC:Local_timer_interrupts
59121 -8.9% 53835 ± 3% interrupts.CPU25.CAL:Function_call_interrupts
258770 -20.3% 206220 ± 5% interrupts.CPU25.LOC:Local_timer_interrupts
59203 -8.8% 53976 ± 4% interrupts.CPU26.CAL:Function_call_interrupts
258771 -20.3% 206166 ± 5% interrupts.CPU26.LOC:Local_timer_interrupts
57731 ± 2% -7.9% 53145 ± 3% interrupts.CPU27.CAL:Function_call_interrupts
258765 -20.3% 206110 ± 5% interrupts.CPU27.LOC:Local_timer_interrupts
59022 -9.1% 53645 ± 4% interrupts.CPU28.CAL:Function_call_interrupts
259049 -20.5% 206059 ± 5% interrupts.CPU28.LOC:Local_timer_interrupts
58573 -8.2% 53783 ± 4% interrupts.CPU29.CAL:Function_call_interrupts
258649 -20.3% 206190 ± 5% interrupts.CPU29.LOC:Local_timer_interrupts
258909 -19.9% 207403 ± 4% interrupts.CPU3.LOC:Local_timer_interrupts
58830 -8.8% 53648 ± 3% interrupts.CPU30.CAL:Function_call_interrupts
258775 -20.3% 206235 ± 5% interrupts.CPU30.LOC:Local_timer_interrupts
58760 -8.4% 53810 ± 3% interrupts.CPU31.CAL:Function_call_interrupts
258780 -20.4% 206118 ± 5% interrupts.CPU31.LOC:Local_timer_interrupts
58551 -8.7% 53477 ± 3% interrupts.CPU32.CAL:Function_call_interrupts
258680 -20.3% 206191 ± 5% interrupts.CPU32.LOC:Local_timer_interrupts
58657 ± 2% -8.8% 53517 ± 4% interrupts.CPU33.CAL:Function_call_interrupts
258752 -20.4% 206003 ± 5% interrupts.CPU33.LOC:Local_timer_interrupts
58161 -7.4% 53868 ± 4% interrupts.CPU34.CAL:Function_call_interrupts
258778 -20.4% 206106 ± 5% interrupts.CPU34.LOC:Local_timer_interrupts
58619 -8.4% 53676 ± 4% interrupts.CPU35.CAL:Function_call_interrupts
258721 -20.3% 206175 ± 5% interrupts.CPU35.LOC:Local_timer_interrupts
58349 -7.6% 53907 ± 4% interrupts.CPU36.CAL:Function_call_interrupts
258753 -20.3% 206154 ± 5% interrupts.CPU36.LOC:Local_timer_interrupts
58714 -9.2% 53323 ± 4% interrupts.CPU37.CAL:Function_call_interrupts
258694 -20.2% 206499 ± 5% interrupts.CPU37.LOC:Local_timer_interrupts
58291 -7.7% 53799 ± 4% interrupts.CPU38.CAL:Function_call_interrupts
259023 -20.4% 206168 ± 5% interrupts.CPU38.LOC:Local_timer_interrupts
58924 -7.8% 54338 ± 4% interrupts.CPU39.CAL:Function_call_interrupts
258920 -20.4% 206187 ± 5% interrupts.CPU39.LOC:Local_timer_interrupts
258817 -19.8% 207532 ± 4% interrupts.CPU4.LOC:Local_timer_interrupts
5957 ± 3% -8.7% 5441 ± 3% interrupts.CPU4.RES:Rescheduling_interrupts
258718 -20.3% 206124 ± 5% interrupts.CPU40.LOC:Local_timer_interrupts
57580 ± 2% -7.3% 53377 ± 4% interrupts.CPU41.CAL:Function_call_interrupts
258739 -20.3% 206212 ± 5% interrupts.CPU41.LOC:Local_timer_interrupts
57848 ± 2% -7.7% 53401 ± 3% interrupts.CPU42.CAL:Function_call_interrupts
258729 -20.3% 206208 ± 5% interrupts.CPU42.LOC:Local_timer_interrupts
58375 -7.0% 54263 ± 4% interrupts.CPU43.CAL:Function_call_interrupts
258771 -20.3% 206256 ± 5% interrupts.CPU43.LOC:Local_timer_interrupts
258781 -19.8% 207611 ± 4% interrupts.CPU44.LOC:Local_timer_interrupts
258798 -19.8% 207429 ± 4% interrupts.CPU45.LOC:Local_timer_interrupts
258765 -19.8% 207525 ± 4% interrupts.CPU46.LOC:Local_timer_interrupts
5658 ± 2% -8.3% 5188 ± 2% interrupts.CPU46.RES:Rescheduling_interrupts
258787 -19.8% 207521 ± 4% interrupts.CPU47.LOC:Local_timer_interrupts
5631 -9.1% 5116 ± 2% interrupts.CPU47.RES:Rescheduling_interrupts
258813 -19.8% 207610 ± 4% interrupts.CPU48.LOC:Local_timer_interrupts
258805 -19.8% 207627 ± 4% interrupts.CPU49.LOC:Local_timer_interrupts
258833 -19.6% 207990 ± 4% interrupts.CPU5.LOC:Local_timer_interrupts
258740 -19.8% 207549 ± 4% interrupts.CPU50.LOC:Local_timer_interrupts
258775 -19.8% 207483 ± 4% interrupts.CPU51.LOC:Local_timer_interrupts
258790 -19.8% 207538 ± 4% interrupts.CPU52.LOC:Local_timer_interrupts
258802 -19.8% 207564 ± 4% interrupts.CPU53.LOC:Local_timer_interrupts
258827 -19.8% 207583 ± 4% interrupts.CPU54.LOC:Local_timer_interrupts
258601 -19.8% 207522 ± 4% interrupts.CPU55.LOC:Local_timer_interrupts
5690 ± 2% -7.7% 5251 ± 2% interrupts.CPU55.RES:Rescheduling_interrupts
258790 -19.8% 207540 ± 4% interrupts.CPU56.LOC:Local_timer_interrupts
258822 -19.8% 207524 ± 4% interrupts.CPU57.LOC:Local_timer_interrupts
258775 -19.8% 207552 ± 4% interrupts.CPU58.LOC:Local_timer_interrupts
258818 -19.8% 207591 ± 4% interrupts.CPU59.LOC:Local_timer_interrupts
258798 -19.8% 207591 ± 4% interrupts.CPU6.LOC:Local_timer_interrupts
258785 -19.8% 207634 ± 4% interrupts.CPU60.LOC:Local_timer_interrupts
258835 -19.8% 207540 ± 4% interrupts.CPU61.LOC:Local_timer_interrupts
5474 -8.1% 5032 ± 3% interrupts.CPU61.RES:Rescheduling_interrupts
258841 -19.8% 207593 ± 4% interrupts.CPU62.LOC:Local_timer_interrupts
258828 -19.8% 207511 ± 4% interrupts.CPU63.LOC:Local_timer_interrupts
258799 -19.8% 207562 ± 4% interrupts.CPU64.LOC:Local_timer_interrupts
258810 -19.8% 207574 ± 4% interrupts.CPU65.LOC:Local_timer_interrupts
58711 -8.7% 53620 ± 4% interrupts.CPU66.CAL:Function_call_interrupts
258759 -20.3% 206149 ± 5% interrupts.CPU66.LOC:Local_timer_interrupts
58760 -8.7% 53648 ± 4% interrupts.CPU67.CAL:Function_call_interrupts
258798 -20.3% 206243 ± 5% interrupts.CPU67.LOC:Local_timer_interrupts
5834 -8.3% 5348 ± 4% interrupts.CPU67.RES:Rescheduling_interrupts
58749 -8.8% 53569 ± 4% interrupts.CPU68.CAL:Function_call_interrupts
258762 -20.3% 206156 ± 5% interrupts.CPU68.LOC:Local_timer_interrupts
58503 -8.5% 53511 ± 4% interrupts.CPU69.CAL:Function_call_interrupts
258766 -20.3% 206179 ± 5% interrupts.CPU69.LOC:Local_timer_interrupts
258815 -19.8% 207607 ± 4% interrupts.CPU7.LOC:Local_timer_interrupts
58243 -9.0% 53008 ± 4% interrupts.CPU70.CAL:Function_call_interrupts
258760 -20.3% 206133 ± 5% interrupts.CPU70.LOC:Local_timer_interrupts
58155 -8.6% 53157 ± 4% interrupts.CPU71.CAL:Function_call_interrupts
258816 -20.3% 206150 ± 5% interrupts.CPU71.LOC:Local_timer_interrupts
58617 -8.4% 53666 ± 4% interrupts.CPU72.CAL:Function_call_interrupts
258800 -20.3% 206195 ± 5% interrupts.CPU72.LOC:Local_timer_interrupts
56202 ± 3% -5.3% 53235 ± 4% interrupts.CPU73.CAL:Function_call_interrupts
258813 -20.4% 206117 ± 5% interrupts.CPU73.LOC:Local_timer_interrupts
58430 -8.8% 53264 ± 4% interrupts.CPU74.CAL:Function_call_interrupts
258763 -20.3% 206118 ± 5% interrupts.CPU74.LOC:Local_timer_interrupts
58240 -8.7% 53145 ± 4% interrupts.CPU75.CAL:Function_call_interrupts
258774 -20.3% 206145 ± 5% interrupts.CPU75.LOC:Local_timer_interrupts
58193 -7.9% 53585 ± 4% interrupts.CPU76.CAL:Function_call_interrupts
258747 -20.3% 206179 ± 5% interrupts.CPU76.LOC:Local_timer_interrupts
6221 -14.0% 5352 ± 2% interrupts.CPU76.RES:Rescheduling_interrupts
59053 -9.8% 53280 ± 4% interrupts.CPU77.CAL:Function_call_interrupts
258776 -20.3% 206177 ± 5% interrupts.CPU77.LOC:Local_timer_interrupts
6068 -15.4% 5132 interrupts.CPU77.RES:Rescheduling_interrupts
57716 -8.0% 53124 ± 4% interrupts.CPU78.CAL:Function_call_interrupts
258754 -20.3% 206213 ± 5% interrupts.CPU78.LOC:Local_timer_interrupts
5830 -13.3% 5056 interrupts.CPU78.RES:Rescheduling_interrupts
58692 -8.6% 53667 ± 4% interrupts.CPU79.CAL:Function_call_interrupts
258794 -20.3% 206167 ± 5% interrupts.CPU79.LOC:Local_timer_interrupts
6035 -14.1% 5186 interrupts.CPU79.RES:Rescheduling_interrupts
258767 -19.8% 207598 ± 4% interrupts.CPU8.LOC:Local_timer_interrupts
58102 -8.1% 53421 ± 4% interrupts.CPU80.CAL:Function_call_interrupts
258638 -20.3% 206237 ± 5% interrupts.CPU80.LOC:Local_timer_interrupts
5924 ± 2% -11.9% 5217 ± 2% interrupts.CPU80.RES:Rescheduling_interrupts
58216 -9.3% 52787 ± 3% interrupts.CPU81.CAL:Function_call_interrupts
258674 -20.3% 206164 ± 5% interrupts.CPU81.LOC:Local_timer_interrupts
6053 -11.9% 5335 interrupts.CPU81.RES:Rescheduling_interrupts
58210 -8.7% 53122 ± 4% interrupts.CPU82.CAL:Function_call_interrupts
258680 -20.3% 206108 ± 5% interrupts.CPU82.LOC:Local_timer_interrupts
6617 ± 5% -18.7% 5379 interrupts.CPU82.RES:Rescheduling_interrupts
58532 -9.1% 53179 ± 4% interrupts.CPU83.CAL:Function_call_interrupts
258721 -20.3% 206164 ± 5% interrupts.CPU83.LOC:Local_timer_interrupts
5855 ± 2% -14.3% 5015 ± 2% interrupts.CPU83.RES:Rescheduling_interrupts
58049 -8.3% 53240 ± 4% interrupts.CPU84.CAL:Function_call_interrupts
258762 -20.4% 206103 ± 5% interrupts.CPU84.LOC:Local_timer_interrupts
58529 -8.7% 53457 ± 4% interrupts.CPU85.CAL:Function_call_interrupts
258746 -20.3% 206226 ± 5% interrupts.CPU85.LOC:Local_timer_interrupts
58118 -8.2% 53328 ± 4% interrupts.CPU86.CAL:Function_call_interrupts
258769 -20.4% 206092 ± 5% interrupts.CPU86.LOC:Local_timer_interrupts
258862 -20.3% 206240 ± 5% interrupts.CPU87.LOC:Local_timer_interrupts
258817 -19.8% 207486 ± 4% interrupts.CPU9.LOC:Local_timer_interrupts
5855 -9.1% 5321 interrupts.CPU9.RES:Rescheduling_interrupts
2099 -23.1% 1613 ± 8% interrupts.IWI:IRQ_work_interrupts
22773085 -20.1% 18204614 ± 4% interrupts.LOC:Local_timer_interrupts
18980 -11.0% 16886 ± 3% softirqs.CPU0.RCU
33504 -15.1% 28441 softirqs.CPU0.SCHED
17793 ± 7% -13.5% 15389 ± 5% softirqs.CPU1.RCU
31237 ± 4% -13.8% 26937 ± 6% softirqs.CPU1.SCHED
16570 ± 2% -12.4% 14507 ± 4% softirqs.CPU10.RCU
29665 -14.1% 25496 ± 2% softirqs.CPU10.SCHED
30185 -17.4% 24946 ± 2% softirqs.CPU11.SCHED
30314 ± 2% -15.8% 25525 ± 2% softirqs.CPU12.SCHED
16626 ± 2% -12.0% 14628 ± 4% softirqs.CPU13.RCU
29804 -15.7% 25129 softirqs.CPU13.SCHED
29843 -15.9% 25104 softirqs.CPU14.SCHED
29723 -14.0% 25574 ± 2% softirqs.CPU15.SCHED
16869 ± 2% -12.7% 14725 ± 4% softirqs.CPU16.RCU
30037 -15.9% 25247 ± 2% softirqs.CPU16.SCHED
16731 ± 3% -10.3% 15002 ± 5% softirqs.CPU17.RCU
29838 -16.1% 25032 softirqs.CPU17.SCHED
17111 ± 4% -13.0% 14891 ± 2% softirqs.CPU18.RCU
30271 -15.8% 25493 softirqs.CPU18.SCHED
29760 -14.1% 25557 ± 4% softirqs.CPU19.SCHED
31874 ± 2% -16.9% 26476 ± 2% softirqs.CPU2.SCHED
18355 ± 8% -18.4% 14976 ± 6% softirqs.CPU20.RCU
29939 -15.5% 25295 ± 2% softirqs.CPU20.SCHED
29891 -15.7% 25204 ± 2% softirqs.CPU21.SCHED
16564 -11.1% 14727 softirqs.CPU22.RCU
29167 ± 2% -13.1% 25349 softirqs.CPU22.SCHED
16200 ± 2% -10.0% 14581 ± 2% softirqs.CPU23.RCU
29737 ± 2% -15.4% 25156 softirqs.CPU23.SCHED
29658 -15.3% 25110 softirqs.CPU24.SCHED
16316 -11.0% 14527 ± 2% softirqs.CPU25.RCU
29893 -16.3% 25012 softirqs.CPU25.SCHED
16615 -11.4% 14715 softirqs.CPU26.RCU
29791 -16.5% 24880 softirqs.CPU26.SCHED
16479 -9.0% 14999 ± 5% softirqs.CPU27.RCU
29569 -14.1% 25401 softirqs.CPU27.SCHED
16763 ± 3% -11.5% 14836 softirqs.CPU28.RCU
30093 -16.1% 25253 softirqs.CPU28.SCHED
16407 -11.7% 14482 ± 2% softirqs.CPU29.RCU
29681 -15.3% 25135 softirqs.CPU29.SCHED
17138 ± 4% -13.0% 14913 ± 5% softirqs.CPU3.RCU
29927 -13.4% 25914 ± 2% softirqs.CPU3.SCHED
16520 -11.6% 14611 ± 2% softirqs.CPU30.RCU
29770 -16.3% 24908 softirqs.CPU30.SCHED
16676 ± 2% -12.4% 14606 ± 4% softirqs.CPU31.RCU
29802 -15.8% 25086 softirqs.CPU31.SCHED
16998 ± 3% -13.3% 14735 ± 2% softirqs.CPU32.RCU
30551 ± 4% -17.1% 25319 softirqs.CPU32.SCHED
16477 -10.7% 14720 softirqs.CPU33.RCU
29801 ± 2% -16.0% 25025 softirqs.CPU33.SCHED
16949 ± 3% -10.6% 15156 ± 4% softirqs.CPU34.RCU
29803 -15.6% 25147 softirqs.CPU34.SCHED
16807 ± 3% -11.8% 14821 softirqs.CPU35.RCU
30197 -16.3% 25273 softirqs.CPU35.SCHED
29782 -15.8% 25072 softirqs.CPU36.SCHED
16477 -10.3% 14772 softirqs.CPU37.RCU
29601 -13.9% 25498 softirqs.CPU37.SCHED
16904 ± 2% -13.0% 14701 ± 2% softirqs.CPU38.RCU
29795 ± 2% -16.6% 24842 softirqs.CPU38.SCHED
16908 ± 3% -14.5% 14453 ± 2% softirqs.CPU39.RCU
29966 -16.5% 25032 softirqs.CPU39.SCHED
29868 -14.2% 25620 softirqs.CPU4.SCHED
17225 ± 6% -16.1% 14454 ± 4% softirqs.CPU40.RCU
29566 -15.3% 25057 softirqs.CPU40.SCHED
16558 -13.1% 14393 ± 4% softirqs.CPU41.RCU
29721 -14.5% 25411 softirqs.CPU41.SCHED
16780 -13.3% 14543 ± 3% softirqs.CPU42.RCU
29813 -15.9% 25065 softirqs.CPU42.SCHED
16691 -11.4% 14793 ± 2% softirqs.CPU43.RCU
29058 ± 2% -15.9% 24424 softirqs.CPU43.SCHED
29016 -13.3% 25153 ± 2% softirqs.CPU44.SCHED
16236 -11.5% 14361 ± 4% softirqs.CPU45.RCU
29485 -13.4% 25525 softirqs.CPU45.SCHED
29643 -16.0% 24890 ± 2% softirqs.CPU46.SCHED
29747 -14.7% 25386 softirqs.CPU47.SCHED
29573 -14.2% 25365 softirqs.CPU48.SCHED
16929 -14.7% 14446 ± 4% softirqs.CPU49.RCU
29866 -15.5% 25238 ± 2% softirqs.CPU49.SCHED
30110 -15.5% 25444 ± 2% softirqs.CPU5.SCHED
16731 ± 4% -14.9% 14246 ± 4% softirqs.CPU50.RCU
29804 -14.9% 25359 ± 2% softirqs.CPU50.SCHED
16865 ± 5% -14.2% 14467 ± 6% softirqs.CPU51.RCU
29674 -15.9% 24950 ± 2% softirqs.CPU51.SCHED
16599 -11.1% 14751 ± 5% softirqs.CPU52.RCU
29968 -15.0% 25467 softirqs.CPU52.SCHED
29848 -15.2% 25326 ± 2% softirqs.CPU53.SCHED
16584 ± 3% -10.5% 14841 ± 7% softirqs.CPU54.RCU
29831 -14.9% 25379 ± 2% softirqs.CPU54.SCHED
16716 ± 4% -12.2% 14669 ± 5% softirqs.CPU55.RCU
29856 -15.6% 25207 ± 2% softirqs.CPU55.SCHED
16415 ± 2% -10.8% 14642 ± 4% softirqs.CPU56.RCU
29554 -14.9% 25148 softirqs.CPU56.SCHED
16554 ± 3% -12.6% 14467 ± 5% softirqs.CPU57.RCU
30048 -16.5% 25096 ± 2% softirqs.CPU57.SCHED
16407 ± 2% -9.8% 14803 ± 8% softirqs.CPU58.RCU
29805 -15.5% 25191 softirqs.CPU58.SCHED
16966 ± 6% -14.2% 14563 ± 4% softirqs.CPU59.RCU
29788 -14.6% 25446 ± 2% softirqs.CPU59.SCHED
17270 ± 5% -14.1% 14842 ± 5% softirqs.CPU6.RCU
29950 -14.1% 25724 ± 2% softirqs.CPU6.SCHED
16423 ± 4% -10.4% 14707 ± 6% softirqs.CPU60.RCU
29734 -13.7% 25658 ± 2% softirqs.CPU60.SCHED
16520 ± 2% -11.5% 14614 ± 5% softirqs.CPU61.RCU
29831 -15.3% 25259 ± 2% softirqs.CPU61.SCHED
29301 -13.6% 25312 softirqs.CPU62.SCHED
16495 ± 2% -12.0% 14520 ± 4% softirqs.CPU63.RCU
29708 -14.3% 25458 softirqs.CPU63.SCHED
16599 ± 2% -13.4% 14369 ± 3% softirqs.CPU64.RCU
29984 -16.0% 25189 softirqs.CPU64.SCHED
16851 ± 3% -10.8% 15031 ± 8% softirqs.CPU65.RCU
29863 -15.0% 25384 ± 2% softirqs.CPU65.SCHED
18595 ± 15% -20.2% 14841 ± 2% softirqs.CPU66.RCU
29553 -14.8% 25187 softirqs.CPU66.SCHED
30196 -15.7% 25465 ± 2% softirqs.CPU67.SCHED
16409 -9.6% 14827 softirqs.CPU68.RCU
30172 -16.8% 25097 softirqs.CPU68.SCHED
30371 ± 3% -15.4% 25695 ± 3% softirqs.CPU69.SCHED
16520 ± 2% -11.0% 14710 ± 5% softirqs.CPU7.RCU
29770 -15.4% 25176 ± 2% softirqs.CPU7.SCHED
16658 ± 2% -10.7% 14868 ± 4% softirqs.CPU70.RCU
29774 -15.6% 25123 softirqs.CPU70.SCHED
16370 -11.6% 14475 ± 4% softirqs.CPU71.RCU
29799 -16.1% 24998 softirqs.CPU71.SCHED
30299 -17.7% 24938 softirqs.CPU72.SCHED
29062 ± 3% -14.0% 24986 softirqs.CPU73.SCHED
16933 ± 5% -11.7% 14945 softirqs.CPU74.RCU
30213 ± 2% -16.7% 25171 softirqs.CPU74.SCHED
15997 -11.2% 14199 ± 4% softirqs.CPU75.RCU
29767 -15.3% 25220 softirqs.CPU75.SCHED
16015 -11.2% 14225 ± 3% softirqs.CPU76.RCU
29681 -15.3% 25132 softirqs.CPU76.SCHED
29659 -15.6% 25036 softirqs.CPU77.SCHED
15990 -12.1% 14048 ± 3% softirqs.CPU78.RCU
29814 -16.1% 25013 softirqs.CPU78.SCHED
16210 ± 2% -13.0% 14101 ± 3% softirqs.CPU79.RCU
29738 -15.8% 25042 softirqs.CPU79.SCHED
16905 ± 5% -14.1% 14527 ± 4% softirqs.CPU8.RCU
29852 -15.5% 25239 ± 2% softirqs.CPU8.SCHED
16792 ± 6% -14.9% 14293 ± 3% softirqs.CPU80.RCU
30105 -17.0% 24988 softirqs.CPU80.SCHED
16428 ± 2% -13.3% 14236 ± 3% softirqs.CPU81.RCU
29787 -16.0% 25019 softirqs.CPU81.SCHED
29496 -14.4% 25253 softirqs.CPU82.SCHED
29706 -14.9% 25293 softirqs.CPU83.SCHED
15801 -10.1% 14210 softirqs.CPU84.RCU
29653 -15.1% 25181 softirqs.CPU84.SCHED
16095 -12.9% 14020 softirqs.CPU85.RCU
29799 -15.0% 25323 softirqs.CPU85.SCHED
16123 -10.7% 14391 ± 2% softirqs.CPU86.RCU
29644 -14.0% 25502 ± 2% softirqs.CPU86.SCHED
16822 ± 2% -13.1% 14626 softirqs.CPU87.RCU
28356 -13.3% 24594 softirqs.CPU87.SCHED
17085 ± 3% -13.2% 14838 ± 6% softirqs.CPU9.RCU
29917 -14.3% 25653 ± 3% softirqs.CPU9.SCHED
1467534 -11.5% 1298751 ± 3% softirqs.RCU
2628721 -15.3% 2226374 softirqs.SCHED
20535 -9.4% 18603 softirqs.TIMER
11.00 -10.2 0.77 ± 4% perf-profile.calltrace.cycles-pp.__xfs_log_force_lsn.xfs_file_fsync.xfs_file_buffered_write.new_sync_write.vfs_write
7.10 -7.1 0.00 perf-profile.calltrace.cycles-pp._raw_spin_lock.__xfs_log_force_lsn.xfs_file_fsync.xfs_file_buffered_write.new_sync_write
7.08 -7.1 0.00 perf-profile.calltrace.cycles-pp.native_queued_spin_lock_slowpath._raw_spin_lock.__xfs_log_force_lsn.xfs_file_fsync.xfs_file_buffered_write
68.11 -1.6 66.52 perf-profile.calltrace.cycles-pp.xfs_file_fsync.xfs_file_buffered_write.new_sync_write.vfs_write.ksys_write
70.73 -1.5 69.20 perf-profile.calltrace.cycles-pp.xfs_file_buffered_write.new_sync_write.vfs_write.ksys_write.do_syscall_64
70.75 -1.5 69.22 perf-profile.calltrace.cycles-pp.new_sync_write.vfs_write.ksys_write.do_syscall_64.entry_SYSCALL_64_after_hwframe
70.83 -1.5 69.32 perf-profile.calltrace.cycles-pp.vfs_write.ksys_write.do_syscall_64.entry_SYSCALL_64_after_hwframe.write
70.84 -1.5 69.33 perf-profile.calltrace.cycles-pp.ksys_write.do_syscall_64.entry_SYSCALL_64_after_hwframe.write
70.93 -1.5 69.42 perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.write
70.92 -1.5 69.41 perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.write
71.09 -1.5 69.61 perf-profile.calltrace.cycles-pp.write
0.82 -0.1 0.71 perf-profile.calltrace.cycles-pp.xlog_ioend_work.process_one_work.worker_thread.kthread.ret_from_fork
1.01 -0.1 0.94 ± 2% perf-profile.calltrace.cycles-pp.xlog_cil_push_work.process_one_work.worker_thread.kthread.ret_from_fork
0.54 +0.0 0.57 ± 2% perf-profile.calltrace.cycles-pp.schedule_idle.do_idle.cpu_startup_entry.start_secondary.secondary_startup_64_no_verify
0.52 +0.0 0.55 ± 2% perf-profile.calltrace.cycles-pp.__schedule.schedule_idle.do_idle.cpu_startup_entry.start_secondary
0.62 +0.0 0.66 ± 2% perf-profile.calltrace.cycles-pp.xlog_state_do_callback.xlog_ioend_work.process_one_work.worker_thread.kthread
2.79 +0.1 2.87 ± 2% perf-profile.calltrace.cycles-pp.__flush_work.xlog_cil_force_lsn.xfs_log_force_lsn.xfs_file_fsync.xfs_file_buffered_write
0.52 ± 2% +0.1 0.61 perf-profile.calltrace.cycles-pp.wait_for_completion.__flush_work.xlog_cil_force_lsn.xfs_log_force_lsn.xfs_file_fsync
0.56 +0.1 0.66 perf-profile.calltrace.cycles-pp.iomap_apply.iomap_file_buffered_write.xfs_file_buffered_write.new_sync_write.vfs_write
0.56 +0.1 0.66 perf-profile.calltrace.cycles-pp.iomap_file_buffered_write.xfs_file_buffered_write.new_sync_write.vfs_write.ksys_write
0.95 ± 4% +0.1 1.07 ± 5% perf-profile.calltrace.cycles-pp.brd_submit_bio.submit_bio_noacct.submit_bio.iomap_submit_ioend.xfs_vm_writepages
0.54 +0.1 0.66 perf-profile.calltrace.cycles-pp.complete.process_one_work.worker_thread.kthread.ret_from_fork
0.70 +0.2 0.85 perf-profile.calltrace.cycles-pp.md_submit_flush_data.process_one_work.worker_thread.kthread.ret_from_fork
1.16 ± 4% +0.2 1.32 ± 4% perf-profile.calltrace.cycles-pp.submit_bio.iomap_submit_ioend.xfs_vm_writepages.do_writepages.__filemap_fdatawrite_range
1.13 ± 4% +0.2 1.28 ± 4% perf-profile.calltrace.cycles-pp.submit_bio_noacct.submit_bio.iomap_submit_ioend.xfs_vm_writepages.do_writepages
1.17 ± 3% +0.2 1.33 ± 4% perf-profile.calltrace.cycles-pp.iomap_submit_ioend.xfs_vm_writepages.do_writepages.__filemap_fdatawrite_range.file_write_and_wait_range
4.28 +0.2 4.45 perf-profile.calltrace.cycles-pp.process_one_work.worker_thread.kthread.ret_from_fork
4.56 +0.2 4.75 perf-profile.calltrace.cycles-pp.worker_thread.kthread.ret_from_fork
4.56 +0.2 4.76 perf-profile.calltrace.cycles-pp.ret_from_fork
4.56 +0.2 4.76 perf-profile.calltrace.cycles-pp.kthread.ret_from_fork
1.96 ± 5% +0.3 2.22 ± 5% perf-profile.calltrace.cycles-pp.__filemap_fdatawrite_range.file_write_and_wait_range.xfs_file_fsync.xfs_file_buffered_write.new_sync_write
1.95 ± 5% +0.3 2.21 ± 5% perf-profile.calltrace.cycles-pp.do_writepages.__filemap_fdatawrite_range.file_write_and_wait_range.xfs_file_fsync.xfs_file_buffered_write
1.94 ± 5% +0.3 2.21 ± 5% perf-profile.calltrace.cycles-pp.xfs_vm_writepages.do_writepages.__filemap_fdatawrite_range.file_write_and_wait_range.xfs_file_fsync
2.09 ± 5% +0.3 2.36 ± 5% perf-profile.calltrace.cycles-pp.file_write_and_wait_range.xfs_file_fsync.xfs_file_buffered_write.new_sync_write.vfs_write
8.93 +0.4 9.33 perf-profile.calltrace.cycles-pp.remove_wait_queue.xlog_cil_force_lsn.xfs_log_force_lsn.xfs_file_fsync.xfs_file_buffered_write
8.90 +0.4 9.30 perf-profile.calltrace.cycles-pp._raw_spin_lock_irqsave.remove_wait_queue.xlog_cil_force_lsn.xfs_log_force_lsn.xfs_file_fsync
8.86 +0.4 9.27 perf-profile.calltrace.cycles-pp.native_queued_spin_lock_slowpath._raw_spin_lock_irqsave.remove_wait_queue.xlog_cil_force_lsn.xfs_log_force_lsn
0.17 ±141% +0.4 0.58 perf-profile.calltrace.cycles-pp.prepare_to_wait_event.md_flush_request.raid0_make_request.md_handle_request.md_submit_bio
0.00 +0.5 0.52 perf-profile.calltrace.cycles-pp.iomap_write_actor.iomap_apply.iomap_file_buffered_write.xfs_file_buffered_write.new_sync_write
0.00 +0.6 0.58 perf-profile.calltrace.cycles-pp.try_to_wake_up.swake_up_locked.complete.process_one_work.worker_thread
0.00 +0.6 0.61 perf-profile.calltrace.cycles-pp.swake_up_locked.complete.process_one_work.worker_thread.kthread
9.75 +0.8 10.59 perf-profile.calltrace.cycles-pp.native_queued_spin_lock_slowpath._raw_spin_lock_irqsave.remove_wait_queue.xlog_wait_on_iclog.__xfs_log_force_lsn
21.50 +1.1 22.59 perf-profile.calltrace.cycles-pp.intel_idle.cpuidle_enter_state.cpuidle_enter.do_idle.cpu_startup_entry
22.08 +1.1 23.19 perf-profile.calltrace.cycles-pp.cpuidle_enter.do_idle.cpu_startup_entry.start_secondary.secondary_startup_64_no_verify
22.05 +1.1 23.16 perf-profile.calltrace.cycles-pp.cpuidle_enter_state.cpuidle_enter.do_idle.cpu_startup_entry.start_secondary
23.93 +1.2 25.15 perf-profile.calltrace.cycles-pp.secondary_startup_64_no_verify
23.63 +1.3 24.88 perf-profile.calltrace.cycles-pp.do_idle.cpu_startup_entry.start_secondary.secondary_startup_64_no_verify
23.64 +1.3 24.90 perf-profile.calltrace.cycles-pp.cpu_startup_entry.start_secondary.secondary_startup_64_no_verify
23.64 +1.3 24.90 perf-profile.calltrace.cycles-pp.start_secondary.secondary_startup_64_no_verify
21.00 +1.9 22.92 perf-profile.calltrace.cycles-pp.native_queued_spin_lock_slowpath._raw_spin_lock_irq.md_flush_request.raid0_make_request.md_handle_request
21.09 +1.9 23.03 perf-profile.calltrace.cycles-pp._raw_spin_lock_irq.md_flush_request.raid0_make_request.md_handle_request.md_submit_bio
13.72 +2.0 15.74 ± 2% perf-profile.calltrace.cycles-pp.__xfs_log_force_lsn.xfs_log_force_lsn.xfs_file_fsync.xfs_file_buffered_write.new_sync_write
22.13 +2.1 24.24 perf-profile.calltrace.cycles-pp.md_flush_request.raid0_make_request.md_handle_request.md_submit_bio.submit_bio_noacct
22.16 +2.1 24.27 perf-profile.calltrace.cycles-pp.raid0_make_request.md_handle_request.md_submit_bio.submit_bio_noacct.submit_bio
22.22 +2.1 24.34 perf-profile.calltrace.cycles-pp.md_handle_request.md_submit_bio.submit_bio_noacct.submit_bio.submit_bio_wait
22.28 +2.1 24.41 perf-profile.calltrace.cycles-pp.md_submit_bio.submit_bio_noacct.submit_bio.submit_bio_wait.blkdev_issue_flush
22.34 +2.1 24.48 perf-profile.calltrace.cycles-pp.submit_bio.submit_bio_wait.blkdev_issue_flush.xfs_file_fsync.xfs_file_buffered_write
22.37 +2.1 24.51 perf-profile.calltrace.cycles-pp.submit_bio_wait.blkdev_issue_flush.xfs_file_fsync.xfs_file_buffered_write.new_sync_write
22.38 +2.1 24.53 perf-profile.calltrace.cycles-pp.blkdev_issue_flush.xfs_file_fsync.xfs_file_buffered_write.new_sync_write.vfs_write
22.33 +2.1 24.48 perf-profile.calltrace.cycles-pp.submit_bio_noacct.submit_bio.submit_bio_wait.blkdev_issue_flush.xfs_file_fsync
6.70 ± 2% +3.7 10.35 ± 2% perf-profile.calltrace.cycles-pp.native_queued_spin_lock_slowpath._raw_spin_lock.xlog_cil_force_lsn.xfs_log_force_lsn.xfs_file_fsync
6.80 ± 2% +3.7 10.46 ± 2% perf-profile.calltrace.cycles-pp._raw_spin_lock.xlog_cil_force_lsn.xfs_log_force_lsn.xfs_file_fsync.xfs_file_buffered_write
18.82 +4.2 23.01 perf-profile.calltrace.cycles-pp.xlog_cil_force_lsn.xfs_log_force_lsn.xfs_file_fsync.xfs_file_buffered_write.new_sync_write
6.05 +4.6 10.62 perf-profile.calltrace.cycles-pp._raw_spin_lock_irqsave.remove_wait_queue.xlog_wait_on_iclog.__xfs_log_force_lsn.xfs_log_force_lsn
6.06 +4.6 10.65 perf-profile.calltrace.cycles-pp.remove_wait_queue.xlog_wait_on_iclog.__xfs_log_force_lsn.xfs_log_force_lsn.xfs_file_fsync
6.25 +4.7 10.97 perf-profile.calltrace.cycles-pp.xlog_wait_on_iclog.__xfs_log_force_lsn.xfs_log_force_lsn.xfs_file_fsync.xfs_file_buffered_write
32.55 +6.2 38.76 perf-profile.calltrace.cycles-pp.xfs_log_force_lsn.xfs_file_fsync.xfs_file_buffered_write.new_sync_write.vfs_write
24.73 -8.2 16.50 ± 2% perf-profile.children.cycles-pp.__xfs_log_force_lsn
20.13 -3.4 16.69 perf-profile.children.cycles-pp._raw_spin_lock
66.06 -2.2 63.83 perf-profile.children.cycles-pp.native_queued_spin_lock_slowpath
68.11 -1.6 66.52 perf-profile.children.cycles-pp.xfs_file_fsync
70.73 -1.5 69.20 perf-profile.children.cycles-pp.xfs_file_buffered_write
70.76 -1.5 69.23 perf-profile.children.cycles-pp.new_sync_write
70.84 -1.5 69.32 perf-profile.children.cycles-pp.vfs_write
70.85 -1.5 69.33 perf-profile.children.cycles-pp.ksys_write
71.12 -1.5 69.62 perf-profile.children.cycles-pp.entry_SYSCALL_64_after_hwframe
71.10 -1.5 69.61 perf-profile.children.cycles-pp.do_syscall_64
71.12 -1.5 69.64 perf-profile.children.cycles-pp.write
21.61 -0.9 20.73 perf-profile.children.cycles-pp.remove_wait_queue
23.44 -0.7 22.73 perf-profile.children.cycles-pp._raw_spin_lock_irqsave
0.62 -0.2 0.43 ± 2% perf-profile.children.cycles-pp.xlog_write
0.20 ± 6% -0.1 0.05 perf-profile.children.cycles-pp.xlog_state_done_syncing
0.82 -0.1 0.71 perf-profile.children.cycles-pp.xlog_ioend_work
0.39 -0.1 0.29 ± 2% perf-profile.children.cycles-pp.xlog_state_release_iclog
1.01 -0.1 0.94 ± 2% perf-profile.children.cycles-pp.xlog_cil_push_work
0.23 ± 11% -0.1 0.16 ± 24% perf-profile.children.cycles-pp.xlog_grant_add_space
0.24 ± 17% -0.1 0.17 ± 19% perf-profile.children.cycles-pp.xfs_log_ticket_ungrant
0.13 -0.1 0.07 perf-profile.children.cycles-pp.xlog_state_get_iclog_space
0.68 -0.0 0.64 ± 2% perf-profile.children.cycles-pp.asm_sysvec_apic_timer_interrupt
0.43 -0.0 0.40 ± 3% perf-profile.children.cycles-pp.__sysvec_apic_timer_interrupt
0.60 -0.0 0.57 ± 2% perf-profile.children.cycles-pp.sysvec_apic_timer_interrupt
0.42 -0.0 0.39 ± 2% perf-profile.children.cycles-pp.hrtimer_interrupt
0.05 +0.0 0.06 perf-profile.children.cycles-pp.__radix_tree_lookup
0.12 +0.0 0.13 perf-profile.children.cycles-pp.__update_load_avg_cfs_rq
0.08 +0.0 0.09 perf-profile.children.cycles-pp.iomap_set_page_dirty
0.08 +0.0 0.09 perf-profile.children.cycles-pp.__list_add_valid
0.06 +0.0 0.07 perf-profile.children.cycles-pp.ttwu_do_wakeup
0.09 +0.0 0.10 ± 4% perf-profile.children.cycles-pp.iomap_set_range_uptodate
0.07 ± 6% +0.0 0.09 ± 5% perf-profile.children.cycles-pp.copyin
0.21 ± 2% +0.0 0.22 perf-profile.children.cycles-pp.ttwu_queue_wakelist
0.09 ± 5% +0.0 0.11 ± 4% perf-profile.children.cycles-pp.xfs_btree_lookup
0.08 +0.0 0.09 ± 5% perf-profile.children.cycles-pp.iov_iter_copy_from_user_atomic
0.09 ± 5% +0.0 0.10 perf-profile.children.cycles-pp.llseek
0.10 ± 4% +0.0 0.11 ± 4% perf-profile.children.cycles-pp.queue_work_on
0.16 ± 5% +0.0 0.18 ± 2% perf-profile.children.cycles-pp.update_rq_clock
0.13 ± 3% +0.0 0.14 ± 3% perf-profile.children.cycles-pp.syscall_exit_to_user_mode
0.09 ± 5% +0.0 0.10 ± 4% perf-profile.children.cycles-pp.__queue_work
0.05 ± 8% +0.0 0.07 ± 11% perf-profile.children.cycles-pp.xfs_map_blocks
0.12 ± 3% +0.0 0.14 ± 5% perf-profile.children.cycles-pp.exit_to_user_mode_prepare
0.11 ± 4% +0.0 0.13 perf-profile.children.cycles-pp.pagecache_get_page
0.11 +0.0 0.13 ± 3% perf-profile.children.cycles-pp.set_task_cpu
0.19 ± 4% +0.0 0.21 perf-profile.children.cycles-pp.__list_del_entry_valid
0.07 ± 6% +0.0 0.09 ± 5% perf-profile.children.cycles-pp.insert_work
0.08 ± 6% +0.0 0.10 ± 4% perf-profile.children.cycles-pp.migrate_task_rq_fair
0.17 ± 2% +0.0 0.19 ± 2% perf-profile.children.cycles-pp.iomap_write_begin
0.16 ± 3% +0.0 0.18 ± 2% perf-profile.children.cycles-pp.perf_trace_sched_wakeup_template
0.12 +0.0 0.14 perf-profile.children.cycles-pp.grab_cache_page_write_begin
0.28 +0.0 0.30 ± 2% perf-profile.children.cycles-pp.pick_next_task_fair
0.11 ± 4% +0.0 0.13 perf-profile.children.cycles-pp.xfs_trans_committed_bulk
0.09 ± 5% +0.0 0.11 perf-profile.children.cycles-pp.xfs_buffered_write_iomap_begin
0.16 +0.0 0.19 ± 6% perf-profile.children.cycles-pp.update_cfs_group
0.14 ± 3% +0.0 0.17 ± 4% perf-profile.children.cycles-pp.__switch_to_asm
0.17 ± 4% +0.0 0.20 ± 7% perf-profile.children.cycles-pp.xfs_inode_item_format
0.15 +0.0 0.18 ± 4% perf-profile.children.cycles-pp.iomap_write_end
0.30 +0.0 0.33 perf-profile.children.cycles-pp.select_idle_cpu
0.19 ± 4% +0.0 0.22 ± 2% perf-profile.children.cycles-pp.xfs_bmap_add_extent_unwritten_real
0.17 ± 4% +0.0 0.20 ± 2% perf-profile.children.cycles-pp.xlog_cil_process_committed
0.17 ± 4% +0.0 0.20 ± 2% perf-profile.children.cycles-pp.xlog_cil_committed
1.25 +0.0 1.29 perf-profile.children.cycles-pp.__wake_up_common_lock
0.26 +0.0 0.30 perf-profile.children.cycles-pp.available_idle_cpu
0.19 ± 4% +0.0 0.23 ± 3% perf-profile.children.cycles-pp.xfs_bmapi_convert_unwritten
0.48 +0.0 0.51 ± 2% perf-profile.children.cycles-pp.dequeue_entity
0.02 ±141% +0.0 0.06 ± 8% perf-profile.children.cycles-pp.xfs_iext_lookup_extent
0.15 ± 6% +0.0 0.19 ± 4% perf-profile.children.cycles-pp.poll_idle
0.11 ± 4% +0.0 0.15 ± 12% perf-profile.children.cycles-pp.submit_bio_checks
0.62 +0.0 0.66 ± 2% perf-profile.children.cycles-pp.xlog_state_do_callback
0.54 +0.0 0.58 ± 2% perf-profile.children.cycles-pp.schedule_idle
0.23 ± 3% +0.0 0.27 perf-profile.children.cycles-pp.xfs_bmapi_write
0.38 +0.0 0.42 perf-profile.children.cycles-pp.xlog_state_clean_iclog
0.37 ± 3% +0.0 0.42 perf-profile.children.cycles-pp.sched_ttwu_pending
0.40 +0.0 0.45 perf-profile.children.cycles-pp.select_idle_sibling
0.59 +0.0 0.64 perf-profile.children.cycles-pp.dequeue_task_fair
0.53 +0.0 0.58 ± 2% perf-profile.children.cycles-pp.select_task_rq_fair
0.00 +0.1 0.05 perf-profile.children.cycles-pp.__pagevec_release
0.56 +0.1 0.61 ± 3% perf-profile.children.cycles-pp.update_load_avg
0.00 +0.1 0.05 ± 8% perf-profile.children.cycles-pp.kfree
0.00 +0.1 0.05 ± 8% perf-profile.children.cycles-pp.submit_flushes
0.00 +0.1 0.05 ± 8% perf-profile.children.cycles-pp.bio_alloc_bioset
0.41 +0.1 0.46 perf-profile.children.cycles-pp.brd_do_bvec
1.19 +0.1 1.24 perf-profile.children.cycles-pp.__wake_up_common
0.47 ± 2% +0.1 0.52 perf-profile.children.cycles-pp.enqueue_entity
0.59 +0.1 0.64 perf-profile.children.cycles-pp.enqueue_task_fair
0.61 ± 2% +0.1 0.68 perf-profile.children.cycles-pp.ttwu_do_activate
0.00 +0.1 0.07 ± 35% perf-profile.children.cycles-pp.blk_throtl_bio
0.44 +0.1 0.51 perf-profile.children.cycles-pp.flush_smp_call_function_from_idle
0.45 +0.1 0.52 perf-profile.children.cycles-pp.iomap_write_actor
0.44 +0.1 0.51 perf-profile.children.cycles-pp.schedule_timeout
2.79 +0.1 2.87 ± 2% perf-profile.children.cycles-pp.__flush_work
0.52 ± 2% +0.1 0.62 perf-profile.children.cycles-pp.wait_for_completion
0.39 ± 2% +0.1 0.49 perf-profile.children.cycles-pp.autoremove_wake_function
0.49 ± 3% +0.1 0.58 perf-profile.children.cycles-pp.prepare_to_wait_event
0.56 +0.1 0.66 perf-profile.children.cycles-pp.iomap_apply
0.56 +0.1 0.66 perf-profile.children.cycles-pp.iomap_file_buffered_write
1.32 +0.1 1.43 perf-profile.children.cycles-pp.schedule
0.49 +0.1 0.61 perf-profile.children.cycles-pp.swake_up_locked
0.54 +0.1 0.67 perf-profile.children.cycles-pp.complete
1.00 ± 4% +0.1 1.13 ± 5% perf-profile.children.cycles-pp.brd_submit_bio
0.71 +0.1 0.85 perf-profile.children.cycles-pp.md_submit_flush_data
1.83 +0.2 1.98 perf-profile.children.cycles-pp.__schedule
4.29 +0.2 4.45 perf-profile.children.cycles-pp.process_one_work
1.17 ± 3% +0.2 1.33 ± 4% perf-profile.children.cycles-pp.iomap_submit_ioend
1.80 +0.2 1.99 perf-profile.children.cycles-pp.try_to_wake_up
4.56 +0.2 4.76 perf-profile.children.cycles-pp.ret_from_fork
4.56 +0.2 4.76 perf-profile.children.cycles-pp.kthread
4.56 +0.2 4.76 perf-profile.children.cycles-pp.worker_thread
1.96 ± 5% +0.3 2.22 ± 5% perf-profile.children.cycles-pp.__filemap_fdatawrite_range
1.95 ± 5% +0.3 2.21 ± 5% perf-profile.children.cycles-pp.do_writepages
1.95 ± 5% +0.3 2.21 ± 5% perf-profile.children.cycles-pp.xfs_vm_writepages
2.09 ± 5% +0.3 2.36 ± 5% perf-profile.children.cycles-pp.file_write_and_wait_range
21.76 +1.1 22.82 perf-profile.children.cycles-pp.intel_idle
22.34 +1.1 23.43 perf-profile.children.cycles-pp.cpuidle_enter
22.34 +1.1 23.43 perf-profile.children.cycles-pp.cpuidle_enter_state
23.93 +1.2 25.15 perf-profile.children.cycles-pp.secondary_startup_64_no_verify
23.93 +1.2 25.15 perf-profile.children.cycles-pp.cpu_startup_entry
23.92 +1.2 25.15 perf-profile.children.cycles-pp.do_idle
23.64 +1.3 24.90 perf-profile.children.cycles-pp.start_secondary
10.12 +1.3 11.45 perf-profile.children.cycles-pp.xlog_wait_on_iclog
23.85 +2.0 25.87 perf-profile.children.cycles-pp._raw_spin_lock_irq
22.33 +2.1 24.47 perf-profile.children.cycles-pp.md_flush_request
22.37 +2.1 24.51 perf-profile.children.cycles-pp.submit_bio_wait
22.38 +2.1 24.53 perf-profile.children.cycles-pp.blkdev_issue_flush
22.43 +2.2 24.59 perf-profile.children.cycles-pp.raid0_make_request
22.53 +2.2 24.70 perf-profile.children.cycles-pp.md_handle_request
22.62 +2.2 24.80 perf-profile.children.cycles-pp.md_submit_bio
23.73 +2.3 26.07 perf-profile.children.cycles-pp.submit_bio
23.75 +2.3 26.09 perf-profile.children.cycles-pp.submit_bio_noacct
18.82 +4.2 23.01 perf-profile.children.cycles-pp.xlog_cil_force_lsn
32.56 +6.2 38.77 perf-profile.children.cycles-pp.xfs_log_force_lsn
65.90 -2.2 63.69 perf-profile.self.cycles-pp.native_queued_spin_lock_slowpath
0.22 ± 9% -0.1 0.16 ± 24% perf-profile.self.cycles-pp.xlog_grant_add_space
0.21 ± 13% -0.1 0.15 ± 16% perf-profile.self.cycles-pp.xfs_log_ticket_ungrant
0.08 +0.0 0.09 perf-profile.self.cycles-pp.__list_add_valid
0.06 +0.0 0.07 perf-profile.self.cycles-pp.write
0.20 ± 2% +0.0 0.21 perf-profile.self.cycles-pp.menu_select
0.11 ± 4% +0.0 0.12 perf-profile.self.cycles-pp.xfs_log_commit_cil
0.09 +0.0 0.10 ± 4% perf-profile.self.cycles-pp.iomap_set_range_uptodate
0.07 ± 6% +0.0 0.09 ± 5% perf-profile.self.cycles-pp.insert_work
0.07 +0.0 0.08 ± 5% perf-profile.self.cycles-pp.memcpy_erms
0.07 ± 7% +0.0 0.08 perf-profile.self.cycles-pp.perf_trace_sched_wakeup_template
0.06 +0.0 0.07 ± 6% perf-profile.self.cycles-pp.flush_smp_call_function_from_idle
0.11 ± 4% +0.0 0.12 ± 3% perf-profile.self.cycles-pp.try_to_wake_up
0.09 ± 5% +0.0 0.10 ± 4% perf-profile.self.cycles-pp.__switch_to
0.08 ± 12% +0.0 0.09 ± 10% perf-profile.self.cycles-pp.xfs_inode_item_format
0.07 ± 6% +0.0 0.09 perf-profile.self.cycles-pp.xlog_cil_force_lsn
0.06 ± 8% +0.0 0.07 ± 6% perf-profile.self.cycles-pp.prepare_to_wait_event
0.16 +0.0 0.18 ± 5% perf-profile.self.cycles-pp.update_cfs_group
0.18 ± 2% +0.0 0.21 perf-profile.self.cycles-pp.__list_del_entry_valid
0.14 ± 3% +0.0 0.17 ± 4% perf-profile.self.cycles-pp.__switch_to_asm
0.36 ± 2% +0.0 0.40 perf-profile.self.cycles-pp.__schedule
0.33 +0.0 0.37 perf-profile.self.cycles-pp.brd_do_bvec
0.30 +0.0 0.33 ± 5% perf-profile.self.cycles-pp.update_load_avg
0.26 ± 3% +0.0 0.30 perf-profile.self.cycles-pp.available_idle_cpu
0.14 ± 6% +0.0 0.18 ± 5% perf-profile.self.cycles-pp.poll_idle
0.57 ± 3% +0.0 0.61 ± 2% perf-profile.self.cycles-pp._raw_spin_lock_irqsave
0.29 +0.0 0.34 ± 3% perf-profile.self.cycles-pp._raw_spin_lock_irq
0.00 +0.1 0.05 perf-profile.self.cycles-pp.kmem_cache_alloc
0.00 +0.1 0.05 ± 8% perf-profile.self.cycles-pp.kfree
0.00 +0.1 0.06 ± 8% perf-profile.self.cycles-pp.migrate_task_rq_fair
0.00 +0.1 0.06 ± 16% perf-profile.self.cycles-pp.percpu_counter_add_batch
21.76 +1.1 22.82 perf-profile.self.cycles-pp.intel_idle
0.04 ± 3% -13.2% 0.04 ± 4% perf-sched.sch_delay.avg.ms.__xfs_log_force_lsn.xfs_log_force_lsn.xfs_file_fsync.xfs_file_buffered_write
0.02 ± 17% -41.4% 0.01 ± 14% perf-sched.sch_delay.avg.ms.do_wait.kernel_wait4.__do_sys_wait4.do_syscall_64
0.24 ± 27% -62.7% 0.09 ± 64% perf-sched.sch_delay.avg.ms.exit_to_user_mode_prepare.irqentry_exit_to_user_mode.asm_sysvec_reschedule_ipi.[unknown]
0.11 ± 12% -59.5% 0.04 ± 16% perf-sched.sch_delay.avg.ms.exit_to_user_mode_prepare.syscall_exit_to_user_mode.do_syscall_64.entry_SYSCALL_64_after_hwframe
0.09 -40.1% 0.05 ± 3% perf-sched.sch_delay.avg.ms.io_schedule.wait_on_page_bit.wait_on_page_writeback.__filemap_fdatawait_range
0.08 ± 2% -19.0% 0.07 perf-sched.sch_delay.avg.ms.md_flush_request.raid0_make_request.md_handle_request.md_submit_bio
0.01 ± 35% -100.0% 0.00 perf-sched.sch_delay.avg.ms.preempt_schedule_common.__cond_resched.__alloc_pages.pagecache_get_page.grab_cache_page_write_begin
0.01 ± 5% -23.8% 0.01 ± 4% perf-sched.sch_delay.avg.ms.preempt_schedule_common.__cond_resched.__flush_work.xlog_cil_force_lsn.xfs_log_force_lsn
0.09 ± 26% -69.2% 0.03 ±100% perf-sched.sch_delay.avg.ms.preempt_schedule_common.__cond_resched.__kmalloc.kmem_alloc.kmem_alloc_large
0.13 ± 68% -74.1% 0.03 ±102% perf-sched.sch_delay.avg.ms.preempt_schedule_common.__cond_resched.do_user_addr_fault.exc_page_fault.asm_exc_page_fault
0.07 ± 22% -41.1% 0.04 ± 36% perf-sched.sch_delay.avg.ms.preempt_schedule_common.__cond_resched.down_read.xfs_file_fsync.xfs_file_buffered_write
0.04 ± 15% -41.0% 0.03 ± 15% perf-sched.sch_delay.avg.ms.preempt_schedule_common.__cond_resched.down_write.xfs_ilock.xfs_buffered_write_iomap_begin
0.11 ± 58% -76.4% 0.03 ±107% perf-sched.sch_delay.avg.ms.preempt_schedule_common.__cond_resched.down_write.xfs_ilock.xfs_file_buffered_write
0.04 ± 28% -63.3% 0.01 ± 17% perf-sched.sch_delay.avg.ms.preempt_schedule_common.__cond_resched.down_write.xfs_ilock.xfs_vn_update_time
0.00 ± 28% -100.0% 0.00 perf-sched.sch_delay.avg.ms.preempt_schedule_common.__cond_resched.down_write_killable.vm_mmap_pgoff.ksys_mmap_pgoff
0.26 -40.4% 0.15 perf-sched.sch_delay.avg.ms.preempt_schedule_common.__cond_resched.process_one_work.worker_thread.kthread
0.08 ± 26% -30.9% 0.06 ± 22% perf-sched.sch_delay.avg.ms.preempt_schedule_common.__cond_resched.stop_one_cpu.migrate_task_to.task_numa_migrate
0.14 ± 16% -83.5% 0.02 ±101% perf-sched.sch_delay.avg.ms.preempt_schedule_common.__cond_resched.submit_bio_checks.submit_bio_noacct.raid0_make_request
0.03 ± 5% -19.0% 0.03 perf-sched.sch_delay.avg.ms.preempt_schedule_common.__cond_resched.submit_bio_checks.submit_bio_noacct.submit_bio
0.11 ± 3% -63.9% 0.04 ± 4% perf-sched.sch_delay.avg.ms.preempt_schedule_common.__cond_resched.wait_for_completion.__flush_work.xlog_cil_force_lsn
0.07 ± 6% -64.7% 0.02 ± 39% perf-sched.sch_delay.avg.ms.preempt_schedule_common.__cond_resched.wait_for_completion.stop_two_cpus.migrate_swap
0.37 ± 2% -19.9% 0.29 ± 2% perf-sched.sch_delay.avg.ms.preempt_schedule_common.__cond_resched.wait_for_completion_io_timeout.submit_bio_wait.blkdev_issue_flush
0.09 ± 29% -58.2% 0.04 ± 36% perf-sched.sch_delay.avg.ms.preempt_schedule_common.__cond_resched.write_cache_pages.iomap_writepages.xfs_vm_writepages
0.03 ± 21% -41.4% 0.02 ± 17% perf-sched.sch_delay.avg.ms.rcu_gp_kthread.kthread.ret_from_fork
0.03 ± 28% +80.9% 0.05 ± 17% perf-sched.sch_delay.avg.ms.rwsem_down_read_slowpath.xfs_log_commit_cil.__xfs_trans_commit.xfs_free_eofblocks
0.19 ± 20% -84.8% 0.03 ±110% perf-sched.sch_delay.avg.ms.rwsem_down_read_slowpath.xfs_log_commit_cil.__xfs_trans_commit.xfs_inactive_ifree
0.13 ± 13% -46.4% 0.07 ± 44% perf-sched.sch_delay.avg.ms.rwsem_down_read_slowpath.xfs_log_commit_cil.__xfs_trans_commit.xfs_remove
0.07 ± 13% -61.3% 0.03 ± 37% perf-sched.sch_delay.avg.ms.rwsem_down_read_slowpath.xfs_log_commit_cil.__xfs_trans_commit.xfs_trans_roll
0.19 ± 3% -36.5% 0.12 ± 7% perf-sched.sch_delay.avg.ms.rwsem_down_read_slowpath.xfs_log_commit_cil.__xfs_trans_commit.xfs_vn_update_time
0.12 ± 15% -74.7% 0.03 ±117% perf-sched.sch_delay.avg.ms.rwsem_down_write_slowpath.do_unlinkat.do_syscall_64.entry_SYSCALL_64_after_hwframe
0.05 ± 11% -46.9% 0.03 ± 4% perf-sched.sch_delay.avg.ms.rwsem_down_write_slowpath.xlog_cil_push_work.process_one_work.worker_thread
0.00 ± 10% +28.6% 0.01 perf-sched.sch_delay.avg.ms.schedule_hrtimeout_range_clock.poll_schedule_timeout.constprop.0.do_select
0.05 ± 5% -52.9% 0.02 ± 4% perf-sched.sch_delay.avg.ms.schedule_timeout.__down.down.xfs_buf_lock
0.04 ± 57% -100.0% 0.00 perf-sched.sch_delay.avg.ms.schedule_timeout.__down.down.xlog_write_iclog
0.03 ± 31% -57.3% 0.01 ± 36% perf-sched.sch_delay.avg.ms.schedule_timeout.kcompactd.kthread.ret_from_fork
0.03 ± 8% -25.5% 0.02 ± 10% perf-sched.sch_delay.avg.ms.schedule_timeout.rcu_gp_kthread.kthread.ret_from_fork
0.06 -25.7% 0.04 perf-sched.sch_delay.avg.ms.schedule_timeout.wait_for_completion.__flush_work.xlog_cil_force_lsn
0.03 ± 7% -15.6% 0.02 ± 5% perf-sched.sch_delay.avg.ms.smpboot_thread_fn.kthread.ret_from_fork
0.02 -16.7% 0.01 ± 3% perf-sched.sch_delay.avg.ms.worker_thread.kthread.ret_from_fork
0.02 ± 4% -27.5% 0.01 ± 3% perf-sched.sch_delay.avg.ms.xlog_cil_force_lsn.xfs_log_force_lsn.xfs_file_fsync.xfs_file_buffered_write
0.00 -100.0% 0.00 perf-sched.sch_delay.avg.ms.xlog_state_get_iclog_space.xlog_write.xlog_cil_push_work.process_one_work
0.11 ± 2% -60.7% 0.04 ± 2% perf-sched.sch_delay.avg.ms.xlog_wait_on_iclog.__xfs_log_force_lsn.xfs_file_fsync.xfs_file_buffered_write
0.06 -11.5% 0.05 perf-sched.sch_delay.avg.ms.xlog_wait_on_iclog.__xfs_log_force_lsn.xfs_log_force_lsn.xfs_file_fsync
5.52 ± 28% -51.9% 2.66 ± 47% perf-sched.sch_delay.max.ms.__xfs_log_force_lsn.xfs_log_force_lsn.xfs_file_fsync.xfs_file_buffered_write
2.27 ± 6% -47.7% 1.19 ± 56% perf-sched.sch_delay.max.ms.exit_to_user_mode_prepare.irqentry_exit_to_user_mode.asm_sysvec_reschedule_ipi.[unknown]
2.36 ± 6% -42.5% 1.36 ± 9% perf-sched.sch_delay.max.ms.exit_to_user_mode_prepare.syscall_exit_to_user_mode.do_syscall_64.entry_SYSCALL_64_after_hwframe
0.01 ± 56% -100.0% 0.00 perf-sched.sch_delay.max.ms.preempt_schedule_common.__cond_resched.__alloc_pages.pagecache_get_page.grab_cache_page_write_begin
0.22 ± 67% -88.5% 0.03 ± 12% perf-sched.sch_delay.max.ms.preempt_schedule_common.__cond_resched.__filemap_fdatawait_range.file_write_and_wait_range.xfs_file_fsync
2.00 ± 14% -35.0% 1.30 ± 30% perf-sched.sch_delay.max.ms.preempt_schedule_common.__cond_resched.__flush_work.xlog_cil_force_lsn.xfs_log_force_lsn
2.24 ± 20% -75.6% 0.55 ± 73% perf-sched.sch_delay.max.ms.preempt_schedule_common.__cond_resched.__kmalloc.kmem_alloc.kmem_alloc_large
0.74 ± 75% -85.2% 0.11 ± 82% perf-sched.sch_delay.max.ms.preempt_schedule_common.__cond_resched.do_user_addr_fault.exc_page_fault.asm_exc_page_fault
2.18 ± 8% -43.3% 1.24 ± 35% perf-sched.sch_delay.max.ms.preempt_schedule_common.__cond_resched.down_read.xfs_file_fsync.xfs_file_buffered_write
1.72 ± 39% -71.1% 0.50 ± 72% perf-sched.sch_delay.max.ms.preempt_schedule_common.__cond_resched.down_read.xfs_log_commit_cil.__xfs_trans_commit
2.17 ± 3% -42.6% 1.25 ± 9% perf-sched.sch_delay.max.ms.preempt_schedule_common.__cond_resched.down_write.xfs_ilock.xfs_buffered_write_iomap_begin
1.60 ± 62% -86.3% 0.22 ±106% perf-sched.sch_delay.max.ms.preempt_schedule_common.__cond_resched.down_write.xfs_ilock.xfs_file_buffered_write
1.38 ± 37% -47.5% 0.72 ± 14% perf-sched.sch_delay.max.ms.preempt_schedule_common.__cond_resched.down_write.xfs_ilock.xfs_vn_update_time
0.00 ± 28% -100.0% 0.00 perf-sched.sch_delay.max.ms.preempt_schedule_common.__cond_resched.down_write_killable.vm_mmap_pgoff.ksys_mmap_pgoff
1.97 ± 10% -30.2% 1.38 ± 5% perf-sched.sch_delay.max.ms.preempt_schedule_common.__cond_resched.mempool_alloc.md_submit_bio.submit_bio_noacct
1.85 ± 7% -31.2% 1.27 ± 2% perf-sched.sch_delay.max.ms.preempt_schedule_common.__cond_resched.process_one_work.worker_thread.kthread
1.59 ± 25% -89.9% 0.16 ± 93% perf-sched.sch_delay.max.ms.preempt_schedule_common.__cond_resched.submit_bio_checks.submit_bio_noacct.raid0_make_request
2.56 ± 9% -28.7% 1.82 ± 8% perf-sched.sch_delay.max.ms.preempt_schedule_common.__cond_resched.wait_for_completion.__flush_work.xlog_cil_force_lsn
2.31 ± 11% -62.5% 0.87 ± 15% perf-sched.sch_delay.max.ms.preempt_schedule_common.__cond_resched.write_cache_pages.iomap_writepages.xfs_vm_writepages
1.06 ± 20% -59.5% 0.43 ± 48% perf-sched.sch_delay.max.ms.rwsem_down_read_slowpath.xfs_log_commit_cil.__xfs_trans_commit.xfs_create
0.94 ± 40% -77.1% 0.22 ±129% perf-sched.sch_delay.max.ms.rwsem_down_read_slowpath.xfs_log_commit_cil.__xfs_trans_commit.xfs_inactive_ifree
1.61 ± 62% -51.7% 0.78 ± 15% perf-sched.sch_delay.max.ms.rwsem_down_read_slowpath.xfs_log_commit_cil.__xfs_trans_commit.xfs_trans_roll
0.62 ± 24% -75.1% 0.15 ±133% perf-sched.sch_delay.max.ms.rwsem_down_write_slowpath.do_unlinkat.do_syscall_64.entry_SYSCALL_64_after_hwframe
1.11 ± 9% -61.9% 0.42 ± 8% perf-sched.sch_delay.max.ms.rwsem_down_write_slowpath.xlog_cil_push_work.process_one_work.worker_thread
0.01 ± 8% +135.3% 0.01 ± 7% perf-sched.sch_delay.max.ms.schedule_hrtimeout_range_clock.poll_schedule_timeout.constprop.0.do_select
0.18 ± 28% -89.7% 0.02 ± 13% perf-sched.sch_delay.max.ms.schedule_hrtimeout_range_clock.poll_schedule_timeout.constprop.0.do_sys_poll
2.54 ± 42% -46.1% 1.37 ± 26% perf-sched.sch_delay.max.ms.schedule_timeout.__down.down.xfs_buf_lock
0.33 ± 78% -100.0% 0.00 perf-sched.sch_delay.max.ms.schedule_timeout.__down.down.xlog_write_iclog
0.70 ± 51% -78.2% 0.15 ± 65% perf-sched.sch_delay.max.ms.schedule_timeout.kcompactd.kthread.ret_from_fork
0.01 ± 14% -100.0% 0.00 perf-sched.sch_delay.max.ms.xlog_state_get_iclog_space.xlog_write.xlog_cil_push_work.process_one_work
4.73 ± 12% -46.7% 2.52 ± 61% perf-sched.sch_delay.max.ms.xlog_wait_on_iclog.__xfs_log_force_lsn.xfs_file_fsync.xfs_file_buffered_write
5.49 ± 20% +24.0% 6.80 ± 3% perf-sched.sch_delay.max.ms.xlog_wait_on_iclog.__xfs_log_force_lsn.xfs_log_force_lsn.xfs_file_fsync
0.06 -23.4% 0.05 perf-sched.total_sch_delay.average.ms
1.45 -11.8% 1.28 perf-sched.total_wait_and_delay.average.ms
4218600 -22.8% 3256636 ± 9% perf-sched.total_wait_and_delay.count.ms
8577 ± 4% -20.4% 6826 ± 10% perf-sched.total_wait_and_delay.max.ms
1.39 -11.3% 1.23 perf-sched.total_wait_time.average.ms
8577 ± 4% -20.4% 6826 ± 10% perf-sched.total_wait_time.max.ms
0.78 +36.9% 1.07 ± 2% perf-sched.wait_and_delay.avg.ms.__xfs_log_force_lsn.xfs_log_force_lsn.xfs_file_fsync.xfs_file_buffered_write
0.49 ±141% +1.3e+05% 618.51 ± 16% perf-sched.wait_and_delay.avg.ms.devkmsg_read.vfs_read.ksys_read.do_syscall_64
788.02 ± 5% -32.2% 533.91 ± 4% perf-sched.wait_and_delay.avg.ms.do_nanosleep.hrtimer_nanosleep.__x64_sys_nanosleep.do_syscall_64
0.50 ±141% +1.2e+05% 618.53 ± 16% perf-sched.wait_and_delay.avg.ms.do_syslog.part.0.kmsg_read.vfs_read
272.43 -61.4% 105.13 ± 6% perf-sched.wait_and_delay.avg.ms.do_task_dead.do_exit.do_group_exit.__x64_sys_exit_group.do_syscall_64
0.34 -44.4% 0.19 ± 2% perf-sched.wait_and_delay.avg.ms.io_schedule.wait_on_page_bit.wait_on_page_writeback.__filemap_fdatawait_range
0.83 +13.0% 0.94 perf-sched.wait_and_delay.avg.ms.md_flush_request.raid0_make_request.md_handle_request.md_submit_bio
127.64 ± 22% +59.7% 203.80 ± 15% perf-sched.wait_and_delay.avg.ms.pipe_read.new_sync_read.vfs_read.ksys_read
20.10 ±101% -87.9% 2.43 ± 95% perf-sched.wait_and_delay.avg.ms.preempt_schedule_common.__cond_resched.__alloc_pages.alloc_pages_vma.shmem_alloc_page
4.68 ± 14% -79.3% 0.97 ±141% perf-sched.wait_and_delay.avg.ms.preempt_schedule_common.__cond_resched.pagecache_get_page.shmem_getpage_gfp.shmem_write_begin
5.34 ± 29% -42.3% 3.08 ± 10% perf-sched.wait_and_delay.avg.ms.rwsem_down_read_slowpath.xfs_log_commit_cil.__xfs_trans_commit.xfs_iomap_write_unwritten
3.68 -47.5% 1.93 ± 2% perf-sched.wait_and_delay.avg.ms.rwsem_down_read_slowpath.xfs_log_commit_cil.__xfs_trans_commit.xfs_trans_roll
213.02 ± 3% +33.6% 284.56 ± 21% perf-sched.wait_and_delay.avg.ms.schedule_hrtimeout_range_clock.ep_poll.do_epoll_wait.__x64_sys_epoll_wait
7.00 ± 2% -12.6% 6.12 ± 7% perf-sched.wait_and_delay.avg.ms.schedule_timeout.rcu_gp_kthread.kthread.ret_from_fork
0.54 -42.8% 0.31 perf-sched.wait_and_delay.avg.ms.schedule_timeout.wait_for_completion.__flush_work.xlog_cil_force_lsn
487.35 ± 3% -17.6% 401.38 ± 4% perf-sched.wait_and_delay.avg.ms.smpboot_thread_fn.kthread.ret_from_fork
4.28 ± 2% -23.3% 3.28 perf-sched.wait_and_delay.avg.ms.worker_thread.kthread.ret_from_fork
0.92 -12.7% 0.80 perf-sched.wait_and_delay.avg.ms.xlog_cil_force_lsn.xfs_log_force_lsn.xfs_file_fsync.xfs_file_buffered_write
0.97 -71.8% 0.27 perf-sched.wait_and_delay.avg.ms.xlog_wait_on_iclog.__xfs_log_force_lsn.xfs_file_fsync.xfs_file_buffered_write
20.00 -26.7% 14.67 ± 12% perf-sched.wait_and_delay.count.__x64_sys_pause.do_syscall_64.entry_SYSCALL_64_after_hwframe.[unknown]
245148 -86.0% 34284 ± 10% perf-sched.wait_and_delay.count.__xfs_log_force_lsn.xfs_log_force_lsn.xfs_file_fsync.xfs_file_buffered_write
0.67 ±141% +1600.0% 11.33 ± 8% perf-sched.wait_and_delay.count.devkmsg_read.vfs_read.ksys_read.do_syscall_64
0.67 ±141% +1600.0% 11.33 ± 8% perf-sched.wait_and_delay.count.do_syslog.part.0.kmsg_read.vfs_read
247.00 +119.8% 543.00 ± 3% perf-sched.wait_and_delay.count.do_task_dead.do_exit.do_group_exit.__x64_sys_exit_group.do_syscall_64
168.00 ± 70% +178.4% 467.67 ± 4% perf-sched.wait_and_delay.count.do_wait.kernel_wait4.__do_sys_wait4.do_syscall_64
251.67 ± 12% -73.8% 66.00 ±141% perf-sched.wait_and_delay.count.exit_to_user_mode_prepare.syscall_exit_to_user_mode.do_syscall_64.entry_SYSCALL_64_after_hwframe
201465 -21.1% 158976 ± 9% perf-sched.wait_and_delay.count.io_schedule.wait_on_page_bit.wait_on_page_writeback.__filemap_fdatawait_range
808550 -16.0% 679238 ± 9% perf-sched.wait_and_delay.count.md_flush_request.raid0_make_request.md_handle_request.md_submit_bio
1320 ± 21% -56.4% 576.33 ± 20% perf-sched.wait_and_delay.count.pipe_read.new_sync_read.vfs_read.ksys_read
27.67 ± 14% -43.4% 15.67 ± 23% perf-sched.wait_and_delay.count.preempt_schedule_common.__cond_resched.generic_perform_write.__generic_file_write_iter.generic_file_write_iter
8.33 ± 11% -88.0% 1.00 ±141% perf-sched.wait_and_delay.count.preempt_schedule_common.__cond_resched.pagecache_get_page.shmem_getpage_gfp.shmem_write_begin
7471 -53.7% 3455 ± 10% perf-sched.wait_and_delay.count.preempt_schedule_common.__cond_resched.process_one_work.worker_thread.kthread
969.67 -20.1% 774.33 ± 12% perf-sched.wait_and_delay.count.preempt_schedule_common.__cond_resched.wait_for_completion.affine_move_task.__set_cpus_allowed_ptr
246.33 ± 7% -21.7% 193.00 ± 15% perf-sched.wait_and_delay.count.rcu_gp_kthread.kthread.ret_from_fork
118.33 ± 8% +18.9% 140.67 ± 4% perf-sched.wait_and_delay.count.rwsem_down_read_slowpath.xfs_log_commit_cil.__xfs_trans_commit.xfs_trans_roll
1035 -37.9% 643.67 ± 10% perf-sched.wait_and_delay.count.schedule_hrtimeout_range_clock.poll_schedule_timeout.constprop.0.do_sys_poll
39.33 ± 2% -27.1% 28.67 ± 11% perf-sched.wait_and_delay.count.schedule_timeout.kcompactd.kthread.ret_from_fork
1217352 -15.4% 1030088 ± 9% perf-sched.wait_and_delay.count.schedule_timeout.wait_for_completion.__flush_work.xlog_cil_force_lsn
206.67 -25.2% 154.67 ± 9% perf-sched.wait_and_delay.count.schedule_timeout.xfsaild.kthread.ret_from_fork
1994 ± 3% -13.6% 1722 ± 13% perf-sched.wait_and_delay.count.smpboot_thread_fn.kthread.ret_from_fork
470869 -13.6% 406938 ± 9% perf-sched.wait_and_delay.count.worker_thread.kthread.ret_from_fork
517416 -27.7% 374206 ± 9% perf-sched.wait_and_delay.count.xlog_cil_force_lsn.xfs_log_force_lsn.xfs_file_fsync.xfs_file_buffered_write
241356 -86.0% 33853 ± 10% perf-sched.wait_and_delay.count.xlog_wait_on_iclog.__xfs_log_force_lsn.xfs_file_fsync.xfs_file_buffered_write
8.75 ± 23% -44.0% 4.90 ± 16% perf-sched.wait_and_delay.max.ms.__xfs_log_force_lsn.xfs_log_force_lsn.xfs_file_fsync.xfs_file_buffered_write
0.98 ±141% +7e+05% 6824 ± 10% perf-sched.wait_and_delay.max.ms.devkmsg_read.vfs_read.ksys_read.do_syscall_64
0.99 ±141% +6.9e+05% 6824 ± 10% perf-sched.wait_and_delay.max.ms.do_syslog.part.0.kmsg_read.vfs_read
999.86 -66.7% 333.06 ±141% perf-sched.wait_and_delay.max.ms.exit_to_user_mode_prepare.syscall_exit_to_user_mode.do_syscall_64.entry_SYSCALL_64_after_hwframe
13.31 ± 16% -39.9% 8.00 ± 16% perf-sched.wait_and_delay.max.ms.io_schedule.wait_on_page_bit.wait_on_page_writeback.__filemap_fdatawait_range
90.59 ± 72% -72.0% 25.40 ± 12% perf-sched.wait_and_delay.max.ms.md_flush_request.raid0_make_request.md_handle_request.md_submit_bio
1013 +573.9% 6826 ± 10% perf-sched.wait_and_delay.max.ms.pipe_read.new_sync_read.vfs_read.ksys_read
85.23 ±123% -94.7% 4.51 ± 88% perf-sched.wait_and_delay.max.ms.preempt_schedule_common.__cond_resched.__alloc_pages.alloc_pages_vma.shmem_alloc_page
174.15 ± 55% -78.9% 36.75 ± 95% perf-sched.wait_and_delay.max.ms.preempt_schedule_common.__cond_resched.generic_perform_write.__generic_file_write_iter.generic_file_write_iter
13.38 ± 19% -84.0% 2.14 ±141% perf-sched.wait_and_delay.max.ms.preempt_schedule_common.__cond_resched.pagecache_get_page.shmem_getpage_gfp.shmem_write_begin
63.82 ± 16% +3174.1% 2089 ±132% perf-sched.wait_and_delay.max.ms.preempt_schedule_common.__cond_resched.process_one_work.worker_thread.kthread
6.62 ± 11% -20.9% 5.24 ± 2% perf-sched.wait_and_delay.max.ms.rcu_gp_kthread.kthread.ret_from_fork
500.62 +1181.3% 6414 ± 11% perf-sched.wait_and_delay.max.ms.schedule_hrtimeout_range_clock.ep_poll.do_epoll_wait.__x64_sys_epoll_wait
8033 ± 10% -40.4% 4787 ± 22% perf-sched.wait_and_delay.max.ms.smpboot_thread_fn.kthread.ret_from_fork
8134 ± 4% -34.5% 5325 ± 7% perf-sched.wait_and_delay.max.ms.worker_thread.kthread.ret_from_fork
0.74 +39.6% 1.04 ± 2% perf-sched.wait_time.avg.ms.__xfs_log_force_lsn.xfs_log_force_lsn.xfs_file_fsync.xfs_file_buffered_write
1.32 ± 8% +46932.4% 618.48 ± 16% perf-sched.wait_time.avg.ms.devkmsg_read.vfs_read.ksys_read.do_syscall_64
787.97 ± 5% -32.2% 533.88 ± 4% perf-sched.wait_time.avg.ms.do_nanosleep.hrtimer_nanosleep.__x64_sys_nanosleep.do_syscall_64
1.33 ± 8% +46567.6% 618.50 ± 16% perf-sched.wait_time.avg.ms.do_syslog.part.0.kmsg_read.vfs_read
272.41 -61.4% 105.12 ± 6% perf-sched.wait_time.avg.ms.do_task_dead.do_exit.do_group_exit.__x64_sys_exit_group.do_syscall_64
0.13 ± 24% -41.9% 0.08 ± 45% perf-sched.wait_time.avg.ms.exit_to_user_mode_prepare.irqentry_exit_to_user_mode.asm_sysvec_call_function_single.[unknown]
0.25 -45.6% 0.14 perf-sched.wait_time.avg.ms.io_schedule.wait_on_page_bit.wait_on_page_writeback.__filemap_fdatawait_range
0.75 +16.4% 0.88 perf-sched.wait_time.avg.ms.md_flush_request.raid0_make_request.md_handle_request.md_submit_bio
127.63 ± 22% +59.7% 203.79 ± 15% perf-sched.wait_time.avg.ms.pipe_read.new_sync_read.vfs_read.ksys_read
20.10 ±101% -87.5% 2.51 ± 89% perf-sched.wait_time.avg.ms.preempt_schedule_common.__cond_resched.__alloc_pages.alloc_pages_vma.shmem_alloc_page
0.05 ±109% -100.0% 0.00 perf-sched.wait_time.avg.ms.preempt_schedule_common.__cond_resched.__alloc_pages.pagecache_get_page.grab_cache_page_write_begin
0.04 ± 3% -33.6% 0.03 ± 8% perf-sched.wait_time.avg.ms.preempt_schedule_common.__cond_resched.__filemap_fdatawait_range.file_write_and_wait_range.xfs_file_fsync
0.70 -38.4% 0.43 perf-sched.wait_time.avg.ms.preempt_schedule_common.__cond_resched.__flush_work.xlog_cil_force_lsn.xfs_log_force_lsn
1.40 ± 41% -66.0% 0.48 ± 86% perf-sched.wait_time.avg.ms.preempt_schedule_common.__cond_resched.do_user_addr_fault.exc_page_fault.asm_exc_page_fault
0.16 ± 12% -35.7% 0.10 ± 18% perf-sched.wait_time.avg.ms.preempt_schedule_common.__cond_resched.down_read.xfs_file_fsync.xfs_file_buffered_write
0.20 ± 37% -66.6% 0.07 ± 37% perf-sched.wait_time.avg.ms.preempt_schedule_common.__cond_resched.down_read.xfs_log_commit_cil.__xfs_trans_commit
0.15 ± 15% -39.6% 0.09 ± 6% perf-sched.wait_time.avg.ms.preempt_schedule_common.__cond_resched.down_write.xfs_ilock.xfs_buffered_write_iomap_begin
0.27 ± 45% -67.2% 0.09 ± 39% perf-sched.wait_time.avg.ms.preempt_schedule_common.__cond_resched.down_write.xfs_ilock.xfs_file_buffered_write
0.19 ± 16% -58.3% 0.08 ± 29% perf-sched.wait_time.avg.ms.preempt_schedule_common.__cond_resched.down_write.xfs_ilock.xfs_vn_update_time
0.22 ± 24% -35.1% 0.14 ± 33% perf-sched.wait_time.avg.ms.preempt_schedule_common.__cond_resched.iomap_write_actor.iomap_apply.iomap_file_buffered_write
0.23 ± 16% -21.8% 0.18 ± 6% perf-sched.wait_time.avg.ms.preempt_schedule_common.__cond_resched.kmem_cache_alloc.xfs_trans_alloc.xfs_vn_update_time
0.08 ± 96% +345.4% 0.34 ± 66% perf-sched.wait_time.avg.ms.preempt_schedule_common.__cond_resched.mempool_alloc.bio_alloc_bioset.submit_flushes
0.52 ± 2% +42.1% 0.73 ± 3% perf-sched.wait_time.avg.ms.preempt_schedule_common.__cond_resched.mempool_alloc.md_submit_bio.submit_bio_noacct
0.00 ±141% +14412.5% 0.39 ±113% perf-sched.wait_time.avg.ms.preempt_schedule_common.__cond_resched.mnt_want_write.do_unlinkat.do_syscall_64
4.68 ± 14% -76.7% 1.09 ±118% perf-sched.wait_time.avg.ms.preempt_schedule_common.__cond_resched.pagecache_get_page.shmem_getpage_gfp.shmem_write_begin
0.81 ± 15% -70.7% 0.24 ± 15% perf-sched.wait_time.avg.ms.preempt_schedule_common.__cond_resched.stop_one_cpu.migrate_task_to.task_numa_migrate
0.25 ± 10% -62.7% 0.09 ± 61% perf-sched.wait_time.avg.ms.preempt_schedule_common.__cond_resched.submit_bio_checks.submit_bio_noacct.raid0_make_request
0.33 ± 6% -39.7% 0.20 ± 4% perf-sched.wait_time.avg.ms.preempt_schedule_common.__cond_resched.wait_for_completion.__flush_work.xlog_cil_force_lsn
0.75 ± 17% -61.3% 0.29 ± 22% perf-sched.wait_time.avg.ms.preempt_schedule_common.__cond_resched.wait_for_completion.stop_two_cpus.migrate_swap
1.07 +11.9% 1.19 perf-sched.wait_time.avg.ms.preempt_schedule_common.__cond_resched.wait_for_completion_io_timeout.submit_bio_wait.blkdev_issue_flush
0.18 ± 25% -49.3% 0.09 ± 32% perf-sched.wait_time.avg.ms.preempt_schedule_common.__cond_resched.write_cache_pages.iomap_writepages.xfs_vm_writepages
0.22 ± 8% -31.9% 0.15 ± 25% perf-sched.wait_time.avg.ms.rwsem_down_read_slowpath.xfs_log_commit_cil.__xfs_trans_commit.xfs_bmapi_convert_delalloc
0.22 ± 15% -46.3% 0.12 ± 21% perf-sched.wait_time.avg.ms.rwsem_down_read_slowpath.xfs_log_commit_cil.__xfs_trans_commit.xfs_create
0.14 ± 8% -21.6% 0.11 ± 3% perf-sched.wait_time.avg.ms.rwsem_down_read_slowpath.xfs_log_commit_cil.__xfs_trans_commit.xfs_free_eofblocks
0.31 ± 15% -76.0% 0.08 ± 42% perf-sched.wait_time.avg.ms.rwsem_down_read_slowpath.xfs_log_commit_cil.__xfs_trans_commit.xfs_inactive_ifree
5.33 ± 29% -42.4% 3.07 ± 10% perf-sched.wait_time.avg.ms.rwsem_down_read_slowpath.xfs_log_commit_cil.__xfs_trans_commit.xfs_iomap_write_unwritten
0.34 ± 14% -33.2% 0.23 ± 23% perf-sched.wait_time.avg.ms.rwsem_down_read_slowpath.xfs_log_commit_cil.__xfs_trans_commit.xfs_remove
3.61 -47.3% 1.90 ± 2% perf-sched.wait_time.avg.ms.rwsem_down_read_slowpath.xfs_log_commit_cil.__xfs_trans_commit.xfs_trans_roll
0.50 ± 4% -42.0% 0.29 ± 7% perf-sched.wait_time.avg.ms.rwsem_down_read_slowpath.xfs_log_commit_cil.__xfs_trans_commit.xfs_vn_update_time
213.00 ± 3% +33.6% 284.53 ± 21% perf-sched.wait_time.avg.ms.schedule_hrtimeout_range_clock.ep_poll.do_epoll_wait.__x64_sys_epoll_wait
0.65 ± 9% -28.3% 0.46 ± 8% perf-sched.wait_time.avg.ms.schedule_timeout.__down.down.xfs_buf_lock
0.33 ± 40% -100.0% 0.00 perf-sched.wait_time.avg.ms.schedule_timeout.__down.down.xlog_write_iclog
6.97 ± 2% -12.6% 6.09 ± 7% perf-sched.wait_time.avg.ms.schedule_timeout.rcu_gp_kthread.kthread.ret_from_fork
0.48 -44.9% 0.26 perf-sched.wait_time.avg.ms.schedule_timeout.wait_for_completion.__flush_work.xlog_cil_force_lsn
0.10 ± 20% -64.3% 0.04 ± 75% perf-sched.wait_time.avg.ms.schedule_timeout.wait_for_completion.stop_two_cpus.migrate_swap
487.33 ± 3% -17.6% 401.36 ± 4% perf-sched.wait_time.avg.ms.smpboot_thread_fn.kthread.ret_from_fork
4.26 ± 2% -23.4% 3.27 perf-sched.wait_time.avg.ms.worker_thread.kthread.ret_from_fork
0.90 -12.5% 0.79 perf-sched.wait_time.avg.ms.xlog_cil_force_lsn.xfs_log_force_lsn.xfs_file_fsync.xfs_file_buffered_write
0.39 ±128% -100.0% 0.00 perf-sched.wait_time.avg.ms.xlog_state_get_iclog_space.xlog_write.xlog_cil_push_work.process_one_work
0.87 -73.2% 0.23 perf-sched.wait_time.avg.ms.xlog_wait_on_iclog.__xfs_log_force_lsn.xfs_file_fsync.xfs_file_buffered_write
5.18 ± 8% -35.3% 3.35 ± 2% perf-sched.wait_time.max.ms.__xfs_log_force_lsn.xfs_log_force_lsn.xfs_file_fsync.xfs_file_buffered_write
2.63 ± 8% +2.6e+05% 6824 ± 10% perf-sched.wait_time.max.ms.devkmsg_read.vfs_read.ksys_read.do_syscall_64
2.65 ± 8% +2.6e+05% 6824 ± 10% perf-sched.wait_time.max.ms.do_syslog.part.0.kmsg_read.vfs_read
3.16 ± 38% -49.3% 1.60 ± 72% perf-sched.wait_time.max.ms.exit_to_user_mode_prepare.irqentry_exit_to_user_mode.asm_sysvec_call_function_single.[unknown]
999.83 -66.2% 337.99 ±138% perf-sched.wait_time.max.ms.exit_to_user_mode_prepare.syscall_exit_to_user_mode.do_syscall_64.entry_SYSCALL_64_after_hwframe
7.37 ± 16% -41.0% 4.35 ± 5% perf-sched.wait_time.max.ms.io_schedule.wait_on_page_bit.wait_on_page_writeback.__filemap_fdatawait_range
90.51 ± 72% -71.9% 25.40 ± 12% perf-sched.wait_time.max.ms.md_flush_request.raid0_make_request.md_handle_request.md_submit_bio
1013 +573.9% 6826 ± 10% perf-sched.wait_time.max.ms.pipe_read.new_sync_read.vfs_read.ksys_read
85.23 ±123% -94.6% 4.59 ± 85% perf-sched.wait_time.max.ms.preempt_schedule_common.__cond_resched.__alloc_pages.alloc_pages_vma.shmem_alloc_page
0.08 ±121% -100.0% 0.00 perf-sched.wait_time.max.ms.preempt_schedule_common.__cond_resched.__alloc_pages.pagecache_get_page.grab_cache_page_write_begin
0.34 ± 24% -84.0% 0.05 ± 15% perf-sched.wait_time.max.ms.preempt_schedule_common.__cond_resched.__filemap_fdatawait_range.file_write_and_wait_range.xfs_file_fsync
4.02 ± 9% -44.5% 2.23 ± 36% perf-sched.wait_time.max.ms.preempt_schedule_common.__cond_resched.__flush_work.xlog_cil_force_lsn.xfs_log_force_lsn
6.69 ± 65% -66.8% 2.22 ± 86% perf-sched.wait_time.max.ms.preempt_schedule_common.__cond_resched.do_user_addr_fault.exc_page_fault.asm_exc_page_fault
2.84 -50.4% 1.41 ± 29% perf-sched.wait_time.max.ms.preempt_schedule_common.__cond_resched.down_read.xfs_file_fsync.xfs_file_buffered_write
2.16 ± 34% -74.2% 0.56 ± 59% perf-sched.wait_time.max.ms.preempt_schedule_common.__cond_resched.down_read.xfs_log_commit_cil.__xfs_trans_commit
3.30 ± 13% -42.0% 1.91 ± 23% perf-sched.wait_time.max.ms.preempt_schedule_common.__cond_resched.down_write.xfs_ilock.xfs_buffered_write_iomap_begin
2.18 ± 19% -46.4% 1.17 ± 11% perf-sched.wait_time.max.ms.preempt_schedule_common.__cond_resched.down_write.xfs_ilock.xfs_vn_update_time
174.15 ± 55% -78.9% 36.75 ± 95% perf-sched.wait_time.max.ms.preempt_schedule_common.__cond_resched.generic_perform_write.__generic_file_write_iter.generic_file_write_iter
2.15 ± 22% -36.2% 1.37 ± 10% perf-sched.wait_time.max.ms.preempt_schedule_common.__cond_resched.kmem_cache_alloc.xfs_trans_alloc.xfs_vn_update_time
1.62 ± 31% -55.2% 0.72 ± 57% perf-sched.wait_time.max.ms.preempt_schedule_common.__cond_resched.mempool_alloc.bio_alloc_bioset.iomap_writepage_map
0.08 ± 96% +345.4% 0.34 ± 66% perf-sched.wait_time.max.ms.preempt_schedule_common.__cond_resched.mempool_alloc.bio_alloc_bioset.submit_flushes
0.00 ±141% +14412.5% 0.39 ±113% perf-sched.wait_time.max.ms.preempt_schedule_common.__cond_resched.mnt_want_write.do_unlinkat.do_syscall_64
13.38 ± 19% -81.9% 2.43 ±116% perf-sched.wait_time.max.ms.preempt_schedule_common.__cond_resched.pagecache_get_page.shmem_getpage_gfp.shmem_write_begin
63.48 ± 17% +3191.3% 2089 ±132% perf-sched.wait_time.max.ms.preempt_schedule_common.__cond_resched.process_one_work.worker_thread.kthread
21.36 ± 34% -63.4% 7.81 ± 26% perf-sched.wait_time.max.ms.preempt_schedule_common.__cond_resched.stop_one_cpu.migrate_task_to.task_numa_migrate
0.85 ± 47% -49.3% 0.43 ± 69% perf-sched.wait_time.max.ms.preempt_schedule_common.__cond_resched.stop_one_cpu.sched_exec.bprm_execve
1.98 ± 29% -53.5% 0.92 ± 66% perf-sched.wait_time.max.ms.preempt_schedule_common.__cond_resched.submit_bio_checks.submit_bio_noacct.raid0_make_request
3.17 ± 7% -25.2% 2.37 ± 14% perf-sched.wait_time.max.ms.preempt_schedule_common.__cond_resched.wait_for_completion.__flush_work.xlog_cil_force_lsn
25.72 ± 7% -45.8% 13.95 ± 8% perf-sched.wait_time.max.ms.preempt_schedule_common.__cond_resched.wait_for_completion.stop_two_cpus.migrate_swap
3.07 ± 8% -56.4% 1.34 ± 21% perf-sched.wait_time.max.ms.preempt_schedule_common.__cond_resched.write_cache_pages.iomap_writepages.xfs_vm_writepages
1.62 ± 49% -68.9% 0.50 ± 69% perf-sched.wait_time.max.ms.preempt_schedule_common.__cond_resched.xfs_trans_alloc.xfs_vn_update_time.file_update_time
1.23 ± 17% -25.0% 0.92 ± 26% perf-sched.wait_time.max.ms.rwsem_down_read_slowpath.xfs_log_commit_cil.__xfs_trans_commit.xfs_bmapi_convert_delalloc
1.40 ± 28% -54.1% 0.64 ± 14% perf-sched.wait_time.max.ms.rwsem_down_read_slowpath.xfs_log_commit_cil.__xfs_trans_commit.xfs_create
1.14 ± 34% -67.1% 0.37 ± 50% perf-sched.wait_time.max.ms.rwsem_down_read_slowpath.xfs_log_commit_cil.__xfs_trans_commit.xfs_inactive_ifree
0.40 ± 71% -70.6% 0.12 ± 95% perf-sched.wait_time.max.ms.rwsem_down_write_slowpath.path_openat.do_filp_open.do_sys_openat2
500.24 +1182.3% 6414 ± 11% perf-sched.wait_time.max.ms.schedule_hrtimeout_range_clock.ep_poll.do_epoll_wait.__x64_sys_epoll_wait
48.99 ± 43% -54.0% 22.53 ± 17% perf-sched.wait_time.max.ms.schedule_timeout.__down.down.xfs_buf_lock
1.29 ± 19% -100.0% 0.00 perf-sched.wait_time.max.ms.schedule_timeout.__down.down.xlog_write_iclog
0.10 ± 20% -64.3% 0.04 ± 75% perf-sched.wait_time.max.ms.schedule_timeout.wait_for_completion.stop_two_cpus.migrate_swap
8033 ± 10% -40.4% 4787 ± 22% perf-sched.wait_time.max.ms.smpboot_thread_fn.kthread.ret_from_fork
8134 ± 4% -34.5% 5325 ± 7% perf-sched.wait_time.max.ms.worker_thread.kthread.ret_from_fork
1.88 ±135% -100.0% 0.00 perf-sched.wait_time.max.ms.xlog_state_get_iclog_space.xlog_write.xlog_cil_push_work.process_one_work
Disclaimer:
Results have been estimated based on internal Intel analysis and are provided
for informational purposes only. Any difference in system hardware or software
design or configuration may affect actual performance.
---
0DAY/LKP+ Test Infrastructure Open Source Technology Center
https://lists.01.org/hyperkitty/list/lkp@lists.01.org Intel Corporation
Thanks,
Oliver Sang
View attachment "config-5.13.0-rc4-00087-ga79b28c284fd" of type "text/plain" (174272 bytes)
View attachment "job-script" of type "text/plain" (8619 bytes)
View attachment "job.yaml" of type "text/plain" (5823 bytes)
View attachment "reproduce" of type "text/plain" (918 bytes)
Powered by blists - more mailing lists