[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <Zw1IHVLclhiBjDkP@bfoster>
Date: Mon, 14 Oct 2024 12:34:37 -0400
From: Brian Foster <bfoster@...hat.com>
To: kernel test robot <oliver.sang@...el.com>
Cc: oe-lkp@...ts.linux.dev, lkp@...el.com, linux-kernel@...r.kernel.org,
Christian Brauner <brauner@...nel.org>,
"Darrick J. Wong" <djwong@...nel.org>,
Josef Bacik <josef@...icpanda.com>, linux-xfs@...r.kernel.org,
linux-fsdevel@...r.kernel.org, ying.huang@...el.com,
feng.tang@...el.com, fengwei.yin@...el.com
Subject: Re: [linus:master] [iomap] c5c810b94c:
stress-ng.metamix.ops_per_sec -98.4% regression
On Mon, Oct 14, 2024 at 03:55:24PM +0800, kernel test robot wrote:
>
>
> Hello,
>
> kernel test robot noticed a -98.4% regression of stress-ng.metamix.ops_per_sec on:
>
>
> commit: c5c810b94cfd818fc2f58c96feee58a9e5ead96d ("iomap: fix handling of dirty folios over unwritten extents")
> https://git.kernel.org/cgit/linux/kernel/git/torvalds/linux.git master
>
> testcase: stress-ng
> config: x86_64-rhel-8.3
> compiler: gcc-12
> test machine: 64 threads 2 sockets Intel(R) Xeon(R) Gold 6346 CPU @ 3.10GHz (Ice Lake) with 256G memory
> parameters:
>
> nr_threads: 100%
> disk: 1HDD
> testtime: 60s
> fs: xfs
> test: metamix
> cpufreq_governor: performance
>
>
>
>
> If you fix the issue in a separate patch/commit (i.e. not just a new version of
> the same patch/commit), kindly add following tags
> | Reported-by: kernel test robot <oliver.sang@...el.com>
> | Closes: https://lore.kernel.org/oe-lkp/202410141536.1167190b-oliver.sang@intel.com
>
>
> Details are as below:
> -------------------------------------------------------------------------------------------------->
>
>
> The kernel config and materials to reproduce are available at:
> https://download.01.org/0day-ci/archive/20241014/202410141536.1167190b-oliver.sang@intel.com
>
So I basically just run this on a >64xcpu guest and reproduce the delta:
stress-ng --timeout 60 --times --verify --metrics --no-rand-seed --metamix 64
The short of it is that with tracing enabled, I see a very large number
of extending writes across unwritten mappings, which basically means XFS
eof zeroing is calling zero range and hitting the newly introduced
flush. This is all pretty much expected given the patch.
I ran a quick experiment to skip the flush on sub-4k ranges in favor of
doing explicit folio zeroing. The idea with that is that the range is
likely restricted to single folio and since it's dirty, we can assume
unwritten conversion is imminent and just explicitly zero the range. I
still see a decent number of flushes from larger ranges in that
experiment, but that still seems to get things pretty close to my
baseline test (on a 6.10 distro kernel).
That said, the next thing on my todo list in this area is to try out the
revalidation based iteration scheme that was discussed on earlier review
passes of the flush fix, with the idea that would elide the flush
entirely on filesystems that support revalidation (i.e. XFS). Given that
the current approach was intended to be a simple and correct fallback
for fs' without revalidation support, I'm a little reluctant to
introduce a heuristic just to appease this benchmark on XFS.
So... my current thinking is that is I'm going to prioritize hacking on
the reval scheme over optimizing the fallback solution and see how that
goes. If that doesn't pan out for whatever reason and/or if some other
more practical workload comes up where this is a practical problem, then
I suspect it will be easy enough to work around with a heuristic to
balance flushes vs. explicit folio zeroing.
For example, I still have another prototype lying around that let
iomap_truncate_page() default to zeroing rather than flushing since the
operation is by design restricted to a partial range of the EOF folio.
With the caveat that I haven't fully dug into this stress-ng test to
know exactly what it's doing, I suspect something like that could help
elide flushes entirely from this workload with some fairly small tweaks
in the fs.
Brian
> =========================================================================================
> compiler/cpufreq_governor/disk/fs/kconfig/nr_threads/rootfs/tbox_group/test/testcase/testtime:
> gcc-12/performance/1HDD/xfs/x86_64-rhel-8.3/100%/debian-12-x86_64-20240206.cgz/lkp-icl-2sp8/metamix/stress-ng/60s
>
> commit:
> 6f634eb080 ("filemap: fix htmldoc warning for mapping_align_index()")
> c5c810b94c ("iomap: fix handling of dirty folios over unwritten extents")
>
> 6f634eb080161baa c5c810b94cfd818fc2f58c96fee
> ---------------- ---------------------------
> %stddev %change %stddev
> \ | \
> 5.762e+09 ± 6% +1784.6% 1.086e+11 ± 3% cpuidle..time
> 537216 ± 8% +846.9% 5087100 ± 3% cpuidle..usage
> 5.83 ± 13% -72.0% 1.63 ± 3% iostat.cpu.idle
> 93.48 +5.1% 98.28 iostat.cpu.iowait
> 134.32 ± 4% +1197.6% 1742 ± 3% uptime.boot
> 2789 ± 3% +54.0% 4295 ± 4% uptime.idle
> 47.17 ±144% -81.6% 8.67 ± 24% perf-c2c.DRAM.local
> 89.00 ± 77% -58.4% 37.00 ± 9% perf-c2c.HITM.local
> 42.33 ± 86% -63.0% 15.67 ± 26% perf-c2c.HITM.remote
> 609419 ± 10% +144.5% 1489941 ± 18% numa-numastat.node0.local_node
> 628123 ± 10% +142.2% 1521154 ± 17% numa-numastat.node0.numa_hit
> 537937 ± 4% +288.1% 2087519 ± 10% numa-numastat.node1.local_node
> 585710 ± 4% +262.4% 2122854 ± 10% numa-numastat.node1.numa_hit
> 33057 ± 5% -94.2% 1926 ± 2% vmstat.io.bo
> 1.72 ± 6% -37.5% 1.08 vmstat.procs.r
> 5454 ± 6% -45.2% 2990 vmstat.system.cs
> 5999 ± 4% -53.5% 2790 vmstat.system.in
> 3.79 ± 19% -2.3 1.52 ± 3% mpstat.cpu.all.idle%
> 0.02 ± 23% -0.0 0.01 ± 3% mpstat.cpu.all.irq%
> 0.01 ± 4% -0.0 0.00 mpstat.cpu.all.soft%
> 0.25 ± 6% -0.2 0.05 ± 4% mpstat.cpu.all.sys%
> 0.40 ± 5% -0.4 0.03 ± 7% mpstat.cpu.all.usr%
> 6.44 ± 5% -13.8% 5.55 ± 3% mpstat.max_utilization_pct
> 1991 ± 14% -68.8% 621.17 ± 9% stress-ng.metamix.ops
> 23.12 ± 10% -98.4% 0.37 ± 7% stress-ng.metamix.ops_per_sec
> 87.77 ± 6% +1831.7% 1695 ± 3% stress-ng.time.elapsed_time
> 87.77 ± 6% +1831.7% 1695 ± 3% stress-ng.time.elapsed_time.max
> 120134 ± 2% -5.1% 114001 stress-ng.time.minor_page_faults
> 5.67 ± 8% -82.4% 1.00 stress-ng.time.percent_of_cpu_this_job_got
> 4.90 ± 7% +331.2% 21.13 ± 4% stress-ng.time.system_time
> 63630 ± 9% +1332.9% 911761 ± 3% stress-ng.time.voluntary_context_switches
> 25272 ± 6% +100.3% 50631 meminfo.Active
> 20787 ± 10% +96.7% 40898 ± 2% meminfo.Active(anon)
> 4485 ± 15% +117.0% 9733 meminfo.Active(file)
> 207516 ± 6% +216.6% 656984 ± 10% meminfo.AnonHugePages
> 377749 ± 4% +31.5% 496804 meminfo.Dirty
> 1808866 -20.6% 1436964 meminfo.Inactive
> 902066 -43.7% 508085 meminfo.Inactive(file)
> 6425133 -10.6% 5746563 meminfo.Memused
> 11.47 +17676.6% 2038 meminfo.Mlocked
> 102534 +22.7% 125834 meminfo.Shmem
> 119495 -95.0% 5925 ± 3% meminfo.Writeback
> 7239789 -14.4% 6193951 meminfo.max_used_kB
> 2230 ± 16% +72.8% 3853 ± 5% numa-meminfo.node0.Active(file)
> 84262 ± 27% +532.6% 533075 ± 25% numa-meminfo.node0.AnonHugePages
> 401080 ± 65% +81.0% 726026 ± 10% numa-meminfo.node0.AnonPages.max
> 450689 ± 3% -59.7% 181817 ± 4% numa-meminfo.node0.Inactive(file)
> 15857 ± 4% -12.5% 13880 ± 4% numa-meminfo.node0.KernelStack
> 5.73 ±100% +14883.5% 858.83 ± 96% numa-meminfo.node0.Mlocked
> 59744 ± 3% -96.6% 2047 ± 4% numa-meminfo.node0.Writeback
> 16267 ± 11% +121.2% 35987 ± 16% numa-meminfo.node1.Active
> 14010 ± 13% +114.9% 30108 ± 20% numa-meminfo.node1.Active(anon)
> 2257 ± 15% +160.4% 5879 ± 4% numa-meminfo.node1.Active(file)
> 188486 ± 7% +69.0% 318533 ± 3% numa-meminfo.node1.Dirty
> 956395 ± 30% -43.4% 541330 ± 20% numa-meminfo.node1.Inactive
> 452002 ± 5% -27.8% 326319 ± 3% numa-meminfo.node1.Inactive(file)
> 150109 ± 13% +16.7% 175200 ± 5% numa-meminfo.node1.Slab
> 59749 ± 4% -93.5% 3891 ± 5% numa-meminfo.node1.Writeback
> 556.82 ± 16% +73.0% 963.40 ± 5% numa-vmstat.node0.nr_active_file
> 41.21 ± 27% +531.7% 260.29 ± 25% numa-vmstat.node0.nr_anon_transparent_hugepages
> 379768 ± 11% -23.8% 289236 ± 3% numa-vmstat.node0.nr_dirtied
> 112900 ± 2% -59.7% 45451 ± 4% numa-vmstat.node0.nr_inactive_file
> 15877 ± 4% -12.6% 13883 ± 3% numa-vmstat.node0.nr_kernel_stack
> 1.44 ±100% +14824.7% 214.67 ± 96% numa-vmstat.node0.nr_mlock
> 14977 ± 3% -96.6% 512.83 ± 4% numa-vmstat.node0.nr_writeback
> 379768 ± 11% -25.6% 282645 ± 3% numa-vmstat.node0.nr_written
> 556.84 ± 16% +73.0% 963.40 ± 5% numa-vmstat.node0.nr_zone_active_file
> 112900 ± 2% -59.7% 45451 ± 4% numa-vmstat.node0.nr_zone_inactive_file
> 62482 ± 3% -27.8% 45088 ± 4% numa-vmstat.node0.nr_zone_write_pending
> 625814 ± 10% +143.0% 1520756 ± 17% numa-vmstat.node0.numa_hit
> 607109 ± 10% +145.4% 1489543 ± 18% numa-vmstat.node0.numa_local
> 3496 ± 13% +115.3% 7527 ± 20% numa-vmstat.node1.nr_active_anon
> 563.82 ± 16% +160.7% 1469 ± 4% numa-vmstat.node1.nr_active_file
> 380179 ± 9% +38.2% 525240 ± 5% numa-vmstat.node1.nr_dirtied
> 47231 ± 7% +68.6% 79622 ± 3% numa-vmstat.node1.nr_dirty
> 113239 ± 5% -28.0% 81571 ± 3% numa-vmstat.node1.nr_inactive_file
> 14977 ± 4% -93.5% 974.17 ± 5% numa-vmstat.node1.nr_writeback
> 380179 ± 9% +35.0% 513207 ± 5% numa-vmstat.node1.nr_written
> 3496 ± 13% +115.3% 7527 ± 20% numa-vmstat.node1.nr_zone_active_anon
> 563.82 ± 16% +160.7% 1469 ± 4% numa-vmstat.node1.nr_zone_active_file
> 113239 ± 5% -28.0% 81571 ± 3% numa-vmstat.node1.nr_zone_inactive_file
> 62209 ± 6% +29.6% 80597 ± 3% numa-vmstat.node1.nr_zone_write_pending
> 583795 ± 4% +263.5% 2121826 ± 10% numa-vmstat.node1.numa_hit
> 535988 ± 4% +289.3% 2086491 ± 10% numa-vmstat.node1.numa_local
> 5190 ± 10% +97.0% 10224 ± 2% proc-vmstat.nr_active_anon
> 1122 ± 14% +116.7% 2433 proc-vmstat.nr_active_file
> 208668 +2.2% 213362 proc-vmstat.nr_anon_pages
> 101.32 ± 6% +216.6% 320.79 ± 10% proc-vmstat.nr_anon_transparent_hugepages
> 94628 ± 4% +31.3% 124208 proc-vmstat.nr_dirty
> 1051112 -8.7% 959304 proc-vmstat.nr_file_pages
> 226820 +2.4% 232205 proc-vmstat.nr_inactive_anon
> 225925 -43.8% 127025 proc-vmstat.nr_inactive_file
> 29400 -6.6% 27458 proc-vmstat.nr_kernel_stack
> 2.88 +17612.4% 509.76 proc-vmstat.nr_mlock
> 22780 -6.0% 21412 ± 3% proc-vmstat.nr_page_table_pages
> 25696 ± 2% +22.4% 31461 proc-vmstat.nr_shmem
> 26966 +2.2% 27573 proc-vmstat.nr_slab_reclaimable
> 63926 +2.0% 65209 proc-vmstat.nr_slab_unreclaimable
> 29903 -95.0% 1484 ± 3% proc-vmstat.nr_writeback
> 5190 ± 10% +97.0% 10224 ± 2% proc-vmstat.nr_zone_active_anon
> 1122 ± 14% +116.7% 2433 proc-vmstat.nr_zone_active_file
> 226820 +2.4% 232205 proc-vmstat.nr_zone_inactive_anon
> 225925 -43.8% 127025 proc-vmstat.nr_zone_inactive_file
> 1215783 ± 6% +199.9% 3646420 ± 3% proc-vmstat.numa_hit
> 1149305 ± 7% +211.5% 3579877 ± 3% proc-vmstat.numa_local
> 89633 ± 7% +349.7% 403114 ± 3% proc-vmstat.pgactivate
> 1312602 ± 6% +200.5% 3944776 ± 3% proc-vmstat.pgalloc_normal
> 415149 ± 3% +897.4% 4140633 ± 3% proc-vmstat.pgfault
> 1306213 ± 6% +200.3% 3923139 ± 3% proc-vmstat.pgfree
> 16256 ± 3% +1057.0% 188090 ± 3% proc-vmstat.pgreuse
> 1.82 ± 4% -9.6% 1.65 perf-stat.i.MPKI
> 3.76e+08 ± 6% -89.1% 40978627 ± 4% perf-stat.i.branch-instructions
> 2.30 ± 3% -0.6 1.66 perf-stat.i.branch-miss-rate%
> 17129191 ± 7% -93.0% 1196180 ± 7% perf-stat.i.branch-misses
> 16.84 ± 3% -6.5 10.36 perf-stat.i.cache-miss-rate%
> 1341312 ± 4% -81.2% 251609 perf-stat.i.cache-misses
> 7592299 ± 3% -73.3% 2030865 perf-stat.i.cache-references
> 5488 ± 6% -45.6% 2986 perf-stat.i.context-switches
> 1.35 ± 4% +8.7% 1.47 perf-stat.i.cpi
> 1.867e+09 ± 5% -86.0% 2.607e+08 ± 6% perf-stat.i.cpu-cycles
> 199.44 ± 2% -59.4% 80.94 perf-stat.i.cpu-migrations
> 1655 ± 13% -41.4% 969.67 perf-stat.i.cycles-between-cache-misses
> 1.841e+09 ± 6% -89.2% 1.994e+08 ± 4% perf-stat.i.instructions
> 0.87 -18.5% 0.71 perf-stat.i.ipc
> 0.57 ± 40% -97.6% 0.01 ± 43% perf-stat.i.major-faults
> 0.01 ±141% +30993.6% 2.48 ± 5% perf-stat.i.metric.K/sec
> 3665 ± 3% -36.0% 2345 perf-stat.i.minor-faults
> 3666 ± 3% -36.0% 2345 perf-stat.i.page-faults
> 0.73 ± 5% +72.7% 1.26 ± 4% perf-stat.overall.MPKI
> 4.55 -1.6 2.92 ± 3% perf-stat.overall.branch-miss-rate%
> 17.66 ± 2% -5.3 12.39 perf-stat.overall.cache-miss-rate%
> 1.02 ± 5% +28.1% 1.31 ± 3% perf-stat.overall.cpi
> 1397 ± 5% -25.8% 1036 ± 5% perf-stat.overall.cycles-between-cache-misses
> 0.98 ± 5% -22.1% 0.77 ± 3% perf-stat.overall.ipc
> 3.728e+08 ± 6% -89.0% 41014226 ± 4% perf-stat.ps.branch-instructions
> 16957001 ± 7% -92.9% 1198002 ± 7% perf-stat.ps.branch-misses
> 1332233 ± 4% -81.1% 251650 perf-stat.ps.cache-misses
> 7543069 ± 3% -73.1% 2030809 perf-stat.ps.cache-references
> 5443 ± 6% -45.2% 2985 perf-stat.ps.context-switches
> 63285 +1.1% 63962 perf-stat.ps.cpu-clock
> 1.859e+09 ± 5% -86.0% 2.611e+08 ± 6% perf-stat.ps.cpu-cycles
> 198.47 ± 3% -59.2% 80.91 perf-stat.ps.cpu-migrations
> 1.826e+09 ± 6% -89.1% 1.996e+08 ± 4% perf-stat.ps.instructions
> 0.58 ± 41% -97.6% 0.01 ± 43% perf-stat.ps.major-faults
> 3640 ± 3% -35.6% 2344 perf-stat.ps.minor-faults
> 3640 ± 3% -35.6% 2344 perf-stat.ps.page-faults
> 63285 +1.1% 63962 perf-stat.ps.task-clock
> 1.637e+11 ± 5% +106.8% 3.387e+11 ± 4% perf-stat.total.instructions
> 0.01 ± 31% -100.0% 0.00 perf-sched.sch_delay.avg.ms.__cond_resched.mempool_alloc_noprof.bio_alloc_bioset.iomap_writepage_map_blocks.iomap_writepage_map
> 0.01 ± 11% -100.0% 0.00 perf-sched.sch_delay.avg.ms.schedule_timeout.__wait_for_common.__flush_workqueue.xlog_cil_push_now.isra
> 0.00 ± 16% -100.0% 0.00 perf-sched.sch_delay.avg.ms.schedule_timeout.xfsaild.kthread.ret_from_fork
> 0.00 ± 17% -100.0% 0.00 perf-sched.sch_delay.avg.ms.xlog_force_lsn.xfs_log_force_seq.xfs_file_fsync.__x64_sys_fdatasync
> 0.01 ± 11% -100.0% 0.00 perf-sched.sch_delay.avg.ms.xlog_wait_on_iclog.xfs_file_fsync.__x64_sys_fdatasync.do_syscall_64
> 0.00 ± 12% -100.0% 0.00 perf-sched.sch_delay.avg.ms.xlog_wait_on_iclog.xfs_log_force_seq.xfs_file_fsync.__x64_sys_fdatasync
> 0.01 ± 8% -100.0% 0.00 perf-sched.sch_delay.avg.ms.xlog_wait_on_iclog.xlog_cil_push_work.process_one_work.worker_thread
> 0.02 ±135% -100.0% 0.00 perf-sched.sch_delay.max.ms.__cond_resched.mempool_alloc_noprof.bio_alloc_bioset.iomap_writepage_map_blocks.iomap_writepage_map
> 0.14 ± 79% -88.5% 0.02 ±103% perf-sched.sch_delay.max.ms.io_schedule.rq_qos_wait.wbt_wait.__rq_qos_throttle
> 0.02 ± 77% -67.9% 0.01 ± 5% perf-sched.sch_delay.max.ms.schedule_hrtimeout_range_clock.do_poll.constprop.0.do_sys_poll
> 0.02 ± 38% -100.0% 0.00 perf-sched.sch_delay.max.ms.schedule_timeout.__wait_for_common.__flush_workqueue.xlog_cil_push_now.isra
> 0.01 ± 44% -100.0% 0.00 perf-sched.sch_delay.max.ms.schedule_timeout.xfsaild.kthread.ret_from_fork
> 0.00 ± 26% -100.0% 0.00 perf-sched.sch_delay.max.ms.xlog_force_lsn.xfs_log_force_seq.xfs_file_fsync.__x64_sys_fdatasync
> 0.01 ±103% -100.0% 0.00 perf-sched.sch_delay.max.ms.xlog_wait_on_iclog.xfs_file_fsync.__x64_sys_fdatasync.do_syscall_64
> 0.00 ± 11% -100.0% 0.00 perf-sched.sch_delay.max.ms.xlog_wait_on_iclog.xfs_log_force_seq.xfs_file_fsync.__x64_sys_fdatasync
> 0.01 ± 11% -100.0% 0.00 perf-sched.sch_delay.max.ms.xlog_wait_on_iclog.xlog_cil_push_work.process_one_work.worker_thread
> 8119 ± 9% -50.9% 3990 ± 15% perf-sched.total_wait_and_delay.count.ms
> 765.01 ± 48% -100.0% 0.01 perf-sched.wait_and_delay.avg.ms.io_schedule.folio_wait_bit_common.folio_wait_writeback.__filemap_fdatawait_range
> 8.46 ± 6% +27.6% 10.79 ± 16% perf-sched.wait_and_delay.avg.ms.schedule_timeout.rcu_gp_fqs_loop.rcu_gp_kthread.kthread
> 50.66 -100.0% 0.00 perf-sched.wait_and_delay.avg.ms.schedule_timeout.xfsaild.kthread.ret_from_fork
> 134.86 ± 13% +158.8% 348.98 ± 18% perf-sched.wait_and_delay.avg.ms.worker_thread.kthread.ret_from_fork.ret_from_fork_asm
> 65.33 ± 70% +353.1% 296.00 ± 64% perf-sched.wait_and_delay.count.io_schedule.folio_wait_bit_common.folio_wait_writeback.__filemap_fdatawait_range
> 1366 ± 18% -78.6% 293.00 ± 64% perf-sched.wait_and_delay.count.io_schedule.rq_qos_wait.wbt_wait.__rq_qos_throttle
> 97.17 -100.0% 0.00 perf-sched.wait_and_delay.count.schedule_timeout.xfsaild.kthread.ret_from_fork
> 3007 ± 18% -77.6% 674.17 ± 29% perf-sched.wait_and_delay.count.worker_thread.kthread.ret_from_fork.ret_from_fork_asm
> 2858 ± 50% -100.0% 0.04 ± 72% perf-sched.wait_and_delay.max.ms.io_schedule.folio_wait_bit_common.folio_wait_writeback.__filemap_fdatawait_range
> 103.17 ± 10% +73.5% 179.00 ± 24% perf-sched.wait_and_delay.max.ms.schedule_timeout.rcu_gp_fqs_loop.rcu_gp_kthread.kthread
> 59.66 ± 10% -100.0% 0.00 perf-sched.wait_and_delay.max.ms.schedule_timeout.xfsaild.kthread.ret_from_fork
> 765.00 ± 48% -100.0% 0.00 perf-sched.wait_time.avg.ms.io_schedule.folio_wait_bit_common.folio_wait_writeback.__filemap_fdatawait_range
> 93.17 ± 32% -100.0% 0.00 perf-sched.wait_time.avg.ms.schedule_timeout.__wait_for_common.__flush_workqueue.xlog_cil_push_now.isra
> 8.45 ± 6% +27.6% 10.78 ± 16% perf-sched.wait_time.avg.ms.schedule_timeout.rcu_gp_fqs_loop.rcu_gp_kthread.kthread
> 50.65 -100.0% 0.00 perf-sched.wait_time.avg.ms.schedule_timeout.xfsaild.kthread.ret_from_fork
> 134.64 ± 13% +158.9% 348.54 ± 18% perf-sched.wait_time.avg.ms.worker_thread.kthread.ret_from_fork.ret_from_fork_asm
> 0.01 ± 22% -100.0% 0.00 perf-sched.wait_time.avg.ms.xlog_force_lsn.xfs_log_force_seq.xfs_file_fsync.__x64_sys_fdatasync
> 202.39 ± 62% -100.0% 0.00 perf-sched.wait_time.avg.ms.xlog_wait_on_iclog.xfs_file_fsync.__x64_sys_fdatasync.do_syscall_64
> 248.16 ± 50% -100.0% 0.00 perf-sched.wait_time.avg.ms.xlog_wait_on_iclog.xfs_log_force_seq.xfs_file_fsync.__x64_sys_fdatasync
> 219.73 ± 46% -100.0% 0.00 perf-sched.wait_time.avg.ms.xlog_wait_on_iclog.xlog_cil_push_work.process_one_work.worker_thread
> 2858 ± 50% -100.0% 0.00 perf-sched.wait_time.max.ms.io_schedule.folio_wait_bit_common.folio_wait_writeback.__filemap_fdatawait_range
> 468.55 ± 29% -100.0% 0.00 perf-sched.wait_time.max.ms.schedule_timeout.__wait_for_common.__flush_workqueue.xlog_cil_push_now.isra
> 103.16 ± 10% +73.5% 179.00 ± 24% perf-sched.wait_time.max.ms.schedule_timeout.rcu_gp_fqs_loop.rcu_gp_kthread.kthread
> 59.65 ± 10% -100.0% 0.00 perf-sched.wait_time.max.ms.schedule_timeout.xfsaild.kthread.ret_from_fork
> 0.02 ± 25% -100.0% 0.00 perf-sched.wait_time.max.ms.xlog_force_lsn.xfs_log_force_seq.xfs_file_fsync.__x64_sys_fdatasync
> 348.84 ± 58% -100.0% 0.00 perf-sched.wait_time.max.ms.xlog_wait_on_iclog.xfs_file_fsync.__x64_sys_fdatasync.do_syscall_64
> 486.25 ± 23% -100.0% 0.00 perf-sched.wait_time.max.ms.xlog_wait_on_iclog.xfs_log_force_seq.xfs_file_fsync.__x64_sys_fdatasync
> 473.19 ± 38% -100.0% 0.00 perf-sched.wait_time.max.ms.xlog_wait_on_iclog.xlog_cil_push_work.process_one_work.worker_thread
> 7346 ± 4% +23.0% 9033 ± 7% sched_debug.cfs_rq:/.avg_vruntime.avg
> 848.55 ± 8% +123.3% 1894 ± 11% sched_debug.cfs_rq:/.avg_vruntime.min
> 0.14 ± 14% -71.2% 0.04 ± 2% sched_debug.cfs_rq:/.h_nr_running.avg
> 0.32 ± 9% -41.3% 0.19 ± 2% sched_debug.cfs_rq:/.h_nr_running.stddev
> 139317 ±191% -88.9% 15437 ± 4% sched_debug.cfs_rq:/.load.avg
> 8499863 ±201% -92.8% 613636 ± 4% sched_debug.cfs_rq:/.load.max
> 247.33 ± 32% -85.6% 35.54 ± 46% sched_debug.cfs_rq:/.load_avg.avg
> 632.63 ± 71% -74.1% 163.67 ± 68% sched_debug.cfs_rq:/.load_avg.stddev
> 7346 ± 4% +23.0% 9033 ± 7% sched_debug.cfs_rq:/.min_vruntime.avg
> 848.55 ± 8% +123.3% 1894 ± 11% sched_debug.cfs_rq:/.min_vruntime.min
> 0.14 ± 14% -71.2% 0.04 ± 2% sched_debug.cfs_rq:/.nr_running.avg
> 0.32 ± 9% -41.3% 0.19 ± 2% sched_debug.cfs_rq:/.nr_running.stddev
> 62.04 ± 22% -89.5% 6.51 ± 42% sched_debug.cfs_rq:/.removed.load_avg.avg
> 597.33 ± 31% -63.0% 220.77 ± 53% sched_debug.cfs_rq:/.removed.load_avg.max
> 171.33 ± 9% -80.1% 34.06 ± 47% sched_debug.cfs_rq:/.removed.load_avg.stddev
> 18.96 ± 25% -85.8% 2.70 ± 45% sched_debug.cfs_rq:/.removed.runnable_avg.avg
> 308.25 ± 31% -64.0% 111.09 ± 54% sched_debug.cfs_rq:/.removed.runnable_avg.max
> 60.15 ± 17% -74.0% 15.67 ± 51% sched_debug.cfs_rq:/.removed.runnable_avg.stddev
> 18.96 ± 25% -85.8% 2.70 ± 45% sched_debug.cfs_rq:/.removed.util_avg.avg
> 308.25 ± 31% -64.0% 111.09 ± 54% sched_debug.cfs_rq:/.removed.util_avg.max
> 60.15 ± 17% -74.0% 15.67 ± 51% sched_debug.cfs_rq:/.removed.util_avg.stddev
> 237.66 ± 5% -85.8% 33.72 ± 3% sched_debug.cfs_rq:/.runnable_avg.avg
> 1104 ± 12% -45.6% 600.94 ± 2% sched_debug.cfs_rq:/.runnable_avg.max
> 226.39 ± 5% -57.8% 95.63 ± 2% sched_debug.cfs_rq:/.runnable_avg.stddev
> 236.68 ± 5% -85.9% 33.42 ± 4% sched_debug.cfs_rq:/.util_avg.avg
> 1103 ± 12% -46.3% 591.99 sched_debug.cfs_rq:/.util_avg.max
> 226.37 ± 5% -58.2% 94.62 ± 2% sched_debug.cfs_rq:/.util_avg.stddev
> 18.18 ± 24% -83.7% 2.96 ± 17% sched_debug.cfs_rq:/.util_est.avg
> 338.67 ± 13% -67.8% 109.03 ± 15% sched_debug.cfs_rq:/.util_est.max
> 67.93 ± 17% -76.8% 15.79 ± 14% sched_debug.cfs_rq:/.util_est.stddev
> 803668 ± 2% +18.9% 955672 sched_debug.cpu.avg_idle.avg
> 198446 ± 5% -41.6% 115938 ± 11% sched_debug.cpu.avg_idle.stddev
> 74790 +1064.4% 870887 ± 3% sched_debug.cpu.clock.avg
> 74795 +1064.4% 870893 ± 3% sched_debug.cpu.clock.max
> 74786 +1064.5% 870879 ± 3% sched_debug.cpu.clock.min
> 2.48 ± 4% +26.3% 3.14 ± 8% sched_debug.cpu.clock.stddev
> 74525 +1068.1% 870538 ± 3% sched_debug.cpu.clock_task.avg
> 74776 +1064.6% 870850 ± 3% sched_debug.cpu.clock_task.max
> 66679 +1193.0% 862189 ± 3% sched_debug.cpu.clock_task.min
> 339.14 ± 15% +27.9% 433.69 ± 4% sched_debug.cpu.curr->pid.avg
> 4069 +498.1% 24338 ± 3% sched_debug.cpu.curr->pid.max
> 895.15 ± 7% +243.7% 3076 ± 3% sched_debug.cpu.curr->pid.stddev
> 0.14 ± 15% -74.6% 0.04 ± 8% sched_debug.cpu.nr_running.avg
> 0.31 ± 11% -44.7% 0.17 ± 6% sched_debug.cpu.nr_running.stddev
> 4679 ± 4% +794.9% 41879 ± 4% sched_debug.cpu.nr_switches.avg
> 27295 ± 11% +571.5% 183299 ± 4% sched_debug.cpu.nr_switches.max
> 1238 ± 20% +985.0% 13436 ± 12% sched_debug.cpu.nr_switches.min
> 4550 ± 12% +581.7% 31022 ± 7% sched_debug.cpu.nr_switches.stddev
> 8.07 +60.9% 12.99 ± 2% sched_debug.cpu.nr_uninterruptible.avg
> 35.50 ± 10% +49.5% 53.06 ± 14% sched_debug.cpu.nr_uninterruptible.max
> 10.63 ± 7% +33.5% 14.19 ± 12% sched_debug.cpu.nr_uninterruptible.stddev
> 74787 +1064.5% 870884 ± 3% sched_debug.cpu_clk
> 73624 +1081.3% 869721 ± 3% sched_debug.ktime
> 75383 +1056.2% 871587 ± 3% sched_debug.sched_clk
> 12.22 ± 39% -12.2 0.00 perf-profile.calltrace.cycles-pp.fdatasync.stress_metamix
> 12.15 ± 39% -12.2 0.00 perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.fdatasync.stress_metamix
> 12.15 ± 39% -12.2 0.00 perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.fdatasync.stress_metamix
> 12.02 ± 40% -12.0 0.00 perf-profile.calltrace.cycles-pp.__x64_sys_fdatasync.do_syscall_64.entry_SYSCALL_64_after_hwframe.fdatasync.stress_metamix
> 12.02 ± 40% -12.0 0.00 perf-profile.calltrace.cycles-pp.xfs_file_fsync.__x64_sys_fdatasync.do_syscall_64.entry_SYSCALL_64_after_hwframe.fdatasync
> 9.67 ± 50% -9.2 0.44 ±112% perf-profile.calltrace.cycles-pp.iomap_file_buffered_write.xfs_file_buffered_write.vfs_write.ksys_write.do_syscall_64
> 8.34 ± 50% -7.9 0.41 ±108% perf-profile.calltrace.cycles-pp.iomap_write_iter.iomap_file_buffered_write.xfs_file_buffered_write.vfs_write.ksys_write
> 7.12 ± 30% -7.1 0.00 perf-profile.calltrace.cycles-pp.file_write_and_wait_range.xfs_file_fsync.__x64_sys_fdatasync.do_syscall_64.entry_SYSCALL_64_after_hwframe
> 6.81 ± 32% -6.8 0.00 perf-profile.calltrace.cycles-pp.__filemap_fdatawrite_range.file_write_and_wait_range.xfs_file_fsync.__x64_sys_fdatasync.do_syscall_64
> 6.81 ± 32% -6.8 0.00 perf-profile.calltrace.cycles-pp.filemap_fdatawrite_wbc.__filemap_fdatawrite_range.file_write_and_wait_range.xfs_file_fsync.__x64_sys_fdatasync
> 6.80 ± 32% -6.8 0.00 perf-profile.calltrace.cycles-pp.do_writepages.filemap_fdatawrite_wbc.__filemap_fdatawrite_range.file_write_and_wait_range.xfs_file_fsync
> 6.80 ± 32% -6.8 0.00 perf-profile.calltrace.cycles-pp.xfs_vm_writepages.do_writepages.filemap_fdatawrite_wbc.__filemap_fdatawrite_range.file_write_and_wait_range
> 6.44 ± 49% -6.3 0.12 ±223% perf-profile.calltrace.cycles-pp.iomap_write_begin.iomap_write_iter.iomap_file_buffered_write.xfs_file_buffered_write.vfs_write
> 5.25 ± 45% -5.2 0.00 perf-profile.calltrace.cycles-pp.iomap_writepage_map.iomap_writepages.xfs_vm_writepages.do_writepages.filemap_fdatawrite_wbc
> 4.70 ± 47% -4.7 0.00 perf-profile.calltrace.cycles-pp.read.stress_metamix
> 4.69 ± 46% -4.7 0.00 perf-profile.calltrace.cycles-pp.iomap_writepage_map_blocks.iomap_writepage_map.iomap_writepages.xfs_vm_writepages.do_writepages
> 4.68 ± 63% -4.7 0.00 perf-profile.calltrace.cycles-pp.unlink.stress_metamix
> 4.66 ± 63% -4.7 0.00 perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.unlink.stress_metamix
> 4.66 ± 63% -4.7 0.00 perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.unlink.stress_metamix
> 4.64 ± 63% -4.6 0.00 perf-profile.calltrace.cycles-pp.__x64_sys_unlink.do_syscall_64.entry_SYSCALL_64_after_hwframe.unlink.stress_metamix
> 4.64 ± 63% -4.6 0.00 perf-profile.calltrace.cycles-pp.do_unlinkat.__x64_sys_unlink.do_syscall_64.entry_SYSCALL_64_after_hwframe.unlink
> 4.12 ± 48% -4.1 0.00 perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.read.stress_metamix
> 4.08 ± 48% -4.1 0.00 perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.read.stress_metamix
> 4.01 ± 64% -4.0 0.00 perf-profile.calltrace.cycles-pp.evict.do_unlinkat.__x64_sys_unlink.do_syscall_64.entry_SYSCALL_64_after_hwframe
> 4.00 ± 64% -4.0 0.00 perf-profile.calltrace.cycles-pp.truncate_inode_pages_range.evict.do_unlinkat.__x64_sys_unlink.do_syscall_64
> 3.96 ± 48% -4.0 0.00 perf-profile.calltrace.cycles-pp.ksys_read.do_syscall_64.entry_SYSCALL_64_after_hwframe.read.stress_metamix
> 3.88 ± 52% -3.9 0.00 perf-profile.calltrace.cycles-pp.iomap_submit_ioend.iomap_writepage_map_blocks.iomap_writepage_map.iomap_writepages.xfs_vm_writepages
> 3.87 ± 53% -3.9 0.00 perf-profile.calltrace.cycles-pp.submit_bio_noacct_nocheck.iomap_submit_ioend.iomap_writepage_map_blocks.iomap_writepage_map.iomap_writepages
> 3.86 ± 53% -3.9 0.00 perf-profile.calltrace.cycles-pp.__submit_bio.submit_bio_noacct_nocheck.iomap_submit_ioend.iomap_writepage_map_blocks.iomap_writepage_map
> 6.96 ± 11% -3.3 3.62 ± 29% perf-profile.calltrace.cycles-pp.worker_thread.kthread.ret_from_fork.ret_from_fork_asm
> 6.29 ± 14% -3.1 3.16 ± 28% perf-profile.calltrace.cycles-pp.process_one_work.worker_thread.kthread.ret_from_fork.ret_from_fork_asm
> 0.65 ± 64% +0.5 1.12 ± 25% perf-profile.calltrace.cycles-pp._nohz_idle_balance.handle_softirqs.__irq_exit_rcu.sysvec_call_function_single.asm_sysvec_call_function_single
> 0.22 ±141% +0.6 0.85 ± 28% perf-profile.calltrace.cycles-pp.ksys_write.do_syscall_64.entry_SYSCALL_64_after_hwframe.write
> 0.11 ±223% +0.6 0.74 ± 28% perf-profile.calltrace.cycles-pp.pipe_read.vfs_read.ksys_read.do_syscall_64.entry_SYSCALL_64_after_hwframe
> 0.22 ±141% +0.7 0.88 ± 27% perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.write
> 0.22 ±141% +0.7 0.88 ± 27% perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.write
> 0.50 ± 76% +0.7 1.15 ± 25% perf-profile.calltrace.cycles-pp.__sysvec_posted_msi_notification.sysvec_posted_msi_notification.asm_sysvec_posted_msi_notification.acpi_safe_halt.acpi_idle_enter
> 0.22 ±142% +0.7 0.94 ± 25% perf-profile.calltrace.cycles-pp.__open64_nocancel.setlocale
> 0.27 ±141% +0.8 1.07 ± 28% perf-profile.calltrace.cycles-pp.write
> 0.33 ±102% +0.8 1.14 ± 22% perf-profile.calltrace.cycles-pp.do_read_fault.do_fault.__handle_mm_fault.handle_mm_fault.do_user_addr_fault
> 0.22 ±145% +0.8 1.05 ± 24% perf-profile.calltrace.cycles-pp.filemap_map_pages.do_read_fault.do_fault.__handle_mm_fault.handle_mm_fault
> 0.43 ±108% +0.9 1.35 ± 24% perf-profile.calltrace.cycles-pp.do_fault.__handle_mm_fault.handle_mm_fault.do_user_addr_fault.exc_page_fault
> 0.16 ±223% +1.0 1.17 ± 36% perf-profile.calltrace.cycles-pp.blk_mq_submit_bio.__submit_bio.submit_bio_noacct_nocheck.iomap_submit_ioend.iomap_writepages
> 0.15 ±223% +1.1 1.20 ± 38% perf-profile.calltrace.cycles-pp.blk_mq_dispatch_rq_list.__blk_mq_do_dispatch_sched.__blk_mq_sched_dispatch_requests.blk_mq_sched_dispatch_requests.blk_mq_run_hw_queue
> 0.66 ± 98% +1.1 1.76 ± 33% perf-profile.calltrace.cycles-pp.cmd_stat.run_builtin.handle_internal_command.main
> 0.66 ± 98% +1.1 1.76 ± 33% perf-profile.calltrace.cycles-pp.dispatch_events.cmd_stat.run_builtin.handle_internal_command.main
> 0.66 ± 98% +1.1 1.76 ± 33% perf-profile.calltrace.cycles-pp.process_interval.dispatch_events.cmd_stat.run_builtin.handle_internal_command
> 0.54 ±119% +1.1 1.68 ± 34% perf-profile.calltrace.cycles-pp.exit_mmap.mmput.exit_mm.do_exit.do_group_exit
> 0.54 ±118% +1.2 1.70 ± 33% perf-profile.calltrace.cycles-pp.exit_mm.do_exit.do_group_exit.__x64_sys_exit_group.x64_sys_call
> 0.54 ±118% +1.2 1.70 ± 33% perf-profile.calltrace.cycles-pp.mmput.exit_mm.do_exit.do_group_exit.__x64_sys_exit_group
> 1.56 ± 19% +1.2 2.73 ± 22% perf-profile.calltrace.cycles-pp.handle_mm_fault.do_user_addr_fault.exc_page_fault.asm_exc_page_fault
> 0.00 +1.2 1.18 ± 40% perf-profile.calltrace.cycles-pp.scsi_queue_rq.blk_mq_dispatch_rq_list.__blk_mq_do_dispatch_sched.__blk_mq_sched_dispatch_requests.blk_mq_sched_dispatch_requests
> 0.45 ±148% +1.2 1.66 ± 33% perf-profile.calltrace.cycles-pp.read_counters.process_interval.dispatch_events.cmd_stat.run_builtin
> 0.72 ± 98% +1.3 1.99 ± 32% perf-profile.calltrace.cycles-pp.handle_internal_command.main
> 0.72 ± 98% +1.3 1.99 ± 32% perf-profile.calltrace.cycles-pp.main
> 0.72 ± 98% +1.3 1.99 ± 32% perf-profile.calltrace.cycles-pp.run_builtin.handle_internal_command.main
> 0.48 ±151% +1.3 1.76 ± 32% perf-profile.calltrace.cycles-pp.vm_mmap_pgoff.ksys_mmap_pgoff.do_syscall_64.entry_SYSCALL_64_after_hwframe
> 1.80 ± 20% +1.3 3.10 ± 23% perf-profile.calltrace.cycles-pp.do_user_addr_fault.exc_page_fault.asm_exc_page_fault
> 1.82 ± 20% +1.3 3.12 ± 23% perf-profile.calltrace.cycles-pp.exc_page_fault.asm_exc_page_fault
> 0.64 ±114% +1.4 2.00 ± 32% perf-profile.calltrace.cycles-pp.__x64_sys_exit_group.x64_sys_call.do_syscall_64.entry_SYSCALL_64_after_hwframe
> 0.64 ±114% +1.4 2.00 ± 32% perf-profile.calltrace.cycles-pp.do_exit.do_group_exit.__x64_sys_exit_group.x64_sys_call.do_syscall_64
> 0.64 ±114% +1.4 2.00 ± 32% perf-profile.calltrace.cycles-pp.do_group_exit.__x64_sys_exit_group.x64_sys_call.do_syscall_64.entry_SYSCALL_64_after_hwframe
> 0.52 ±109% +1.4 1.88 ± 27% perf-profile.calltrace.cycles-pp.do_sys_openat2.__x64_sys_openat.do_syscall_64.entry_SYSCALL_64_after_hwframe
> 0.53 ±109% +1.4 1.92 ± 28% perf-profile.calltrace.cycles-pp.__x64_sys_openat.do_syscall_64.entry_SYSCALL_64_after_hwframe
> 0.93 ± 83% +1.4 2.32 ± 32% perf-profile.calltrace.cycles-pp.x64_sys_call.do_syscall_64.entry_SYSCALL_64_after_hwframe
> 0.91 ± 89% +1.5 2.41 ± 42% perf-profile.calltrace.cycles-pp.do_mmap.vm_mmap_pgoff.ksys_mmap_pgoff.do_syscall_64.entry_SYSCALL_64_after_hwframe
> 0.71 ±110% +1.6 2.32 ± 42% perf-profile.calltrace.cycles-pp.mmap_region.do_mmap.vm_mmap_pgoff.ksys_mmap_pgoff.do_syscall_64
> 0.66 ±112% +1.7 2.35 ± 30% perf-profile.calltrace.cycles-pp.setlocale
> 2.08 ± 26% +1.7 3.82 ± 29% perf-profile.calltrace.cycles-pp.asm_exc_page_fault
> 1.05 ± 78% +1.9 2.94 ± 28% perf-profile.calltrace.cycles-pp.ksys_read.do_syscall_64.entry_SYSCALL_64_after_hwframe.read
> 1.46 ± 49% +1.9 3.35 ± 19% perf-profile.calltrace.cycles-pp.asm_sysvec_posted_msi_notification.acpi_safe_halt.acpi_idle_enter.cpuidle_enter_state.cpuidle_enter
> 1.10 ± 78% +1.9 3.02 ± 28% perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.read
> 1.10 ± 78% +1.9 3.02 ± 28% perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.read
> 1.30 ± 69% +2.0 3.28 ± 32% perf-profile.calltrace.cycles-pp.load_elf_binary.search_binary_handler.exec_binprm.bprm_execve.do_execveat_common
> 1.32 ± 68% +2.0 3.30 ± 31% perf-profile.calltrace.cycles-pp.exec_binprm.bprm_execve.do_execveat_common.__x64_sys_execve.do_syscall_64
> 1.32 ± 68% +2.0 3.30 ± 31% perf-profile.calltrace.cycles-pp.search_binary_handler.exec_binprm.bprm_execve.do_execveat_common.__x64_sys_execve
> 1.43 ± 70% +2.0 3.45 ± 30% perf-profile.calltrace.cycles-pp.bprm_execve.do_execveat_common.__x64_sys_execve.do_syscall_64.entry_SYSCALL_64_after_hwframe
> 1.11 ± 78% +2.1 3.24 ± 27% perf-profile.calltrace.cycles-pp.read
> 2.01 ± 71% +3.0 4.99 ± 22% perf-profile.calltrace.cycles-pp.__x64_sys_execve.do_syscall_64.entry_SYSCALL_64_after_hwframe.execve
> 2.01 ± 71% +3.0 5.00 ± 22% perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.execve
> 2.01 ± 71% +3.0 5.00 ± 22% perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.execve
> 2.01 ± 71% +3.0 5.00 ± 22% perf-profile.calltrace.cycles-pp.execve
> 2.00 ± 70% +3.0 4.99 ± 22% perf-profile.calltrace.cycles-pp.do_execveat_common.__x64_sys_execve.do_syscall_64.entry_SYSCALL_64_after_hwframe.execve
> 1.01 ± 54% +3.3 4.27 ± 31% perf-profile.calltrace.cycles-pp.native_queued_spin_lock_slowpath._raw_spin_lock.dd_dispatch_request.__blk_mq_do_dispatch_sched.__blk_mq_sched_dispatch_requests
> 1.01 ± 54% +3.4 4.36 ± 30% perf-profile.calltrace.cycles-pp._raw_spin_lock.dd_dispatch_request.__blk_mq_do_dispatch_sched.__blk_mq_sched_dispatch_requests.blk_mq_sched_dispatch_requests
> 1.05 ± 55% +3.4 4.47 ± 31% perf-profile.calltrace.cycles-pp.dd_dispatch_request.__blk_mq_do_dispatch_sched.__blk_mq_sched_dispatch_requests.blk_mq_sched_dispatch_requests.blk_mq_run_hw_queue
> 2.33 ± 39% +4.0 6.29 ± 28% perf-profile.calltrace.cycles-pp.__blk_mq_sched_dispatch_requests.blk_mq_sched_dispatch_requests.blk_mq_run_hw_queue.blk_mq_dispatch_plug_list.blk_mq_flush_plug_list
> 2.33 ± 39% +4.0 6.31 ± 28% perf-profile.calltrace.cycles-pp.blk_mq_sched_dispatch_requests.blk_mq_run_hw_queue.blk_mq_dispatch_plug_list.blk_mq_flush_plug_list.__blk_flush_plug
> 2.35 ± 39% +4.0 6.35 ± 27% perf-profile.calltrace.cycles-pp.blk_mq_run_hw_queue.blk_mq_dispatch_plug_list.blk_mq_flush_plug_list.__blk_flush_plug.__submit_bio
> 1.71 ± 50% +4.5 6.23 ± 29% perf-profile.calltrace.cycles-pp.__blk_mq_do_dispatch_sched.__blk_mq_sched_dispatch_requests.blk_mq_sched_dispatch_requests.blk_mq_run_hw_queue.blk_mq_dispatch_plug_list
> 3.24 ± 66% +4.8 8.01 ± 27% perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe
> 3.24 ± 66% +4.8 8.04 ± 27% perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe
> 3.55 ± 40% +5.2 8.77 ± 29% perf-profile.calltrace.cycles-pp.blk_mq_dispatch_plug_list.blk_mq_flush_plug_list.__blk_flush_plug.__submit_bio.submit_bio_noacct_nocheck
> 3.55 ± 40% +5.2 8.77 ± 29% perf-profile.calltrace.cycles-pp.blk_mq_flush_plug_list.__blk_flush_plug.__submit_bio.submit_bio_noacct_nocheck.iomap_submit_ioend
> 3.98 ± 55% +8.1 12.11 ± 30% perf-profile.calltrace.cycles-pp.xfs_file_write_checks.xfs_file_buffered_write.vfs_write.ksys_write.do_syscall_64
> 0.56 ±164% +8.2 8.77 ± 29% perf-profile.calltrace.cycles-pp.__blk_flush_plug.__submit_bio.submit_bio_noacct_nocheck.iomap_submit_ioend.iomap_writepages
> 3.33 ± 56% +8.3 11.60 ± 31% perf-profile.calltrace.cycles-pp.iomap_zero_range.xfs_file_write_checks.xfs_file_buffered_write.vfs_write.ksys_write
> 0.91 ±108% +9.0 9.94 ± 29% perf-profile.calltrace.cycles-pp.__submit_bio.submit_bio_noacct_nocheck.iomap_submit_ioend.iomap_writepages.xfs_vm_writepages
> 0.91 ±108% +9.0 9.94 ± 29% perf-profile.calltrace.cycles-pp.submit_bio_noacct_nocheck.iomap_submit_ioend.iomap_writepages.xfs_vm_writepages.do_writepages
> 0.92 ±107% +9.0 9.95 ± 29% perf-profile.calltrace.cycles-pp.iomap_submit_ioend.iomap_writepages.xfs_vm_writepages.do_writepages.filemap_fdatawrite_wbc
> 0.00 +10.3 10.29 ± 30% perf-profile.calltrace.cycles-pp.do_writepages.filemap_fdatawrite_wbc.__filemap_fdatawrite_range.filemap_write_and_wait_range.iomap_zero_range
> 0.00 +10.3 10.29 ± 30% perf-profile.calltrace.cycles-pp.xfs_vm_writepages.do_writepages.filemap_fdatawrite_wbc.__filemap_fdatawrite_range.filemap_write_and_wait_range
> 0.00 +10.4 10.40 ± 30% perf-profile.calltrace.cycles-pp.__filemap_fdatawrite_range.filemap_write_and_wait_range.iomap_zero_range.xfs_file_write_checks.xfs_file_buffered_write
> 0.00 +10.4 10.40 ± 30% perf-profile.calltrace.cycles-pp.filemap_fdatawrite_wbc.__filemap_fdatawrite_range.filemap_write_and_wait_range.iomap_zero_range.xfs_file_write_checks
> 0.00 +11.2 11.22 ± 30% perf-profile.calltrace.cycles-pp.filemap_write_and_wait_range.iomap_zero_range.xfs_file_write_checks.xfs_file_buffered_write.vfs_write
> 12.24 ± 39% -12.2 0.00 perf-profile.children.cycles-pp.fdatasync
> 12.02 ± 40% -12.0 0.00 perf-profile.children.cycles-pp.__x64_sys_fdatasync
> 12.02 ± 40% -12.0 0.00 perf-profile.children.cycles-pp.xfs_file_fsync
> 9.68 ± 50% -9.1 0.57 ± 69% perf-profile.children.cycles-pp.iomap_file_buffered_write
> 8.37 ± 50% -7.8 0.53 ± 66% perf-profile.children.cycles-pp.iomap_write_iter
> 7.12 ± 30% -7.1 0.00 perf-profile.children.cycles-pp.file_write_and_wait_range
> 6.48 ± 49% -6.1 0.35 ± 65% perf-profile.children.cycles-pp.iomap_write_begin
> 5.25 ± 45% -5.0 0.22 ± 74% perf-profile.children.cycles-pp.iomap_writepage_map
> 4.74 ± 60% -4.7 0.00 perf-profile.children.cycles-pp.unlink
> 4.69 ± 60% -4.7 0.00 perf-profile.children.cycles-pp.__x64_sys_unlink
> 4.69 ± 60% -4.7 0.00 perf-profile.children.cycles-pp.do_unlinkat
> 4.70 ± 46% -4.6 0.14 ± 97% perf-profile.children.cycles-pp.iomap_writepage_map_blocks
> 4.31 ± 55% -4.2 0.16 ±108% perf-profile.children.cycles-pp.iomap_iter
> 4.13 ± 61% -4.1 0.00 perf-profile.children.cycles-pp.truncate_inode_pages_range
> 4.06 ± 61% -4.1 0.01 ±223% perf-profile.children.cycles-pp.evict
> 3.90 ± 48% -3.8 0.11 ± 82% perf-profile.children.cycles-pp.__iomap_write_begin
> 3.82 ± 56% -3.7 0.08 ± 80% perf-profile.children.cycles-pp.xfs_buffered_write_iomap_begin
> 3.42 ± 47% -3.4 0.07 ±111% perf-profile.children.cycles-pp.zero_user_segments
> 6.97 ± 11% -3.4 3.62 ± 29% perf-profile.children.cycles-pp.worker_thread
> 3.45 ± 46% -3.2 0.24 ± 35% perf-profile.children.cycles-pp.memset_orig
> 3.25 ± 45% -3.1 0.11 ± 88% perf-profile.children.cycles-pp.filemap_read
> 6.29 ± 14% -3.1 3.16 ± 28% perf-profile.children.cycles-pp.process_one_work
> 3.18 ± 54% -2.6 0.59 ± 50% perf-profile.children.cycles-pp.folios_put_refs
> 2.46 ± 67% -2.4 0.10 ± 65% perf-profile.children.cycles-pp.__page_cache_release
> 2.51 ± 50% -2.2 0.28 ± 52% perf-profile.children.cycles-pp.__filemap_get_folio
> 1.58 ± 51% -1.5 0.10 ± 84% perf-profile.children.cycles-pp.filemap_add_folio
> 1.52 ± 45% -1.4 0.10 ±119% perf-profile.children.cycles-pp.copy_page_to_iter
> 1.48 ± 44% -1.4 0.12 ± 75% perf-profile.children.cycles-pp._copy_to_iter
> 2.55 ± 16% -1.1 1.41 ± 36% perf-profile.children.cycles-pp.pick_next_task_fair
> 1.22 ± 24% -1.1 0.10 ± 71% perf-profile.children.cycles-pp.mod_delayed_work_on
> 1.16 ± 50% -1.0 0.13 ± 78% perf-profile.children.cycles-pp.open64
> 0.95 ± 31% -0.9 0.06 ± 84% perf-profile.children.cycles-pp.try_to_grab_pending
> 0.99 ± 45% -0.9 0.10 ± 71% perf-profile.children.cycles-pp.kblockd_mod_delayed_work_on
> 0.88 ± 50% -0.8 0.06 ±111% perf-profile.children.cycles-pp.filemap_get_pages
> 0.81 ± 48% -0.7 0.06 ±113% perf-profile.children.cycles-pp.filemap_get_read_batch
> 1.08 ± 28% -0.6 0.45 ± 39% perf-profile.children.cycles-pp.clear_bhb_loop
> 0.70 ± 51% -0.6 0.08 ±125% perf-profile.children.cycles-pp.folio_alloc_noprof
> 0.66 ± 53% -0.6 0.06 ± 84% perf-profile.children.cycles-pp.__filemap_add_folio
> 0.59 ± 32% -0.6 0.04 ±100% perf-profile.children.cycles-pp.xfs_map_blocks
> 0.88 ± 14% -0.5 0.38 ± 51% perf-profile.children.cycles-pp.__lruvec_stat_mod_folio
> 0.58 ± 38% -0.5 0.13 ± 81% perf-profile.children.cycles-pp.writeback_iter
> 0.71 ± 29% -0.4 0.26 ± 51% perf-profile.children.cycles-pp.entry_SYSCALL_64
> 0.52 ± 43% -0.4 0.12 ± 91% perf-profile.children.cycles-pp.writeback_get_folio
> 0.44 ± 53% -0.4 0.06 ±130% perf-profile.children.cycles-pp.__folio_start_writeback
> 0.49 ± 20% -0.4 0.12 ± 62% perf-profile.children.cycles-pp.percpu_counter_add_batch
> 0.35 ± 38% -0.3 0.08 ± 54% perf-profile.children.cycles-pp.touch_atime
> 0.63 ± 23% -0.3 0.37 ± 66% perf-profile.children.cycles-pp.get_page_from_freelist
> 0.47 ± 33% -0.3 0.21 ± 79% perf-profile.children.cycles-pp.xas_load
> 0.30 ± 35% -0.2 0.09 ± 86% perf-profile.children.cycles-pp.rmqueue
> 0.24 ± 29% -0.2 0.05 ± 71% perf-profile.children.cycles-pp.__mod_node_page_state
> 0.13 ± 22% -0.1 0.04 ±101% perf-profile.children.cycles-pp.lock_timer_base
> 0.16 ± 31% -0.1 0.08 ± 85% perf-profile.children.cycles-pp.xas_find
> 0.16 ± 22% -0.1 0.08 ± 78% perf-profile.children.cycles-pp.xfs_trans_reserve
> 0.05 ± 78% +0.1 0.12 ± 37% perf-profile.children.cycles-pp.hrtimer_try_to_cancel
> 0.00 +0.1 0.08 ± 10% perf-profile.children.cycles-pp.change_protection_range
> 0.13 ± 33% +0.1 0.21 ± 38% perf-profile.children.cycles-pp.wakeup_preempt
> 0.05 ±108% +0.1 0.14 ± 24% perf-profile.children.cycles-pp.perf_event_read
> 0.05 ±108% +0.1 0.14 ± 24% perf-profile.children.cycles-pp.smp_call_function_single
> 0.04 ±118% +0.1 0.14 ± 37% perf-profile.children.cycles-pp.run_ksoftirqd
> 0.02 ±143% +0.1 0.12 ± 60% perf-profile.children.cycles-pp.__poll
> 0.02 ±143% +0.1 0.12 ± 60% perf-profile.children.cycles-pp.__x64_sys_poll
> 0.02 ±143% +0.1 0.12 ± 60% perf-profile.children.cycles-pp.do_sys_poll
> 0.01 ±223% +0.1 0.11 ± 48% perf-profile.children.cycles-pp.lockref_put_return
> 0.02 ±142% +0.1 0.12 ± 60% perf-profile.children.cycles-pp.do_poll
> 0.07 ± 75% +0.1 0.18 ± 34% perf-profile.children.cycles-pp.switch_fpu_return
> 0.04 ±112% +0.1 0.15 ± 66% perf-profile.children.cycles-pp.getenv
> 0.01 ±223% +0.1 0.12 ± 39% perf-profile.children.cycles-pp.folio_putback_lru
> 0.04 ±114% +0.1 0.16 ± 37% perf-profile.children.cycles-pp.generic_exec_single
> 0.02 ±223% +0.1 0.14 ± 38% perf-profile.children.cycles-pp.expand_downwards
> 0.12 ± 44% +0.1 0.24 ± 15% perf-profile.children.cycles-pp.__check_object_size
> 0.14 ± 29% +0.1 0.27 ± 19% perf-profile.children.cycles-pp.vma_alloc_folio_noprof
> 0.01 ±223% +0.1 0.14 ± 37% perf-profile.children.cycles-pp.xfsaild
> 0.01 ±223% +0.1 0.14 ± 37% perf-profile.children.cycles-pp.xfsaild_push
> 0.01 ±223% +0.1 0.14 ± 39% perf-profile.children.cycles-pp.get_cpu_sleep_time_us
> 0.06 ± 83% +0.1 0.19 ± 62% perf-profile.children.cycles-pp.__d_add
> 0.10 ± 92% +0.1 0.24 ± 15% perf-profile.children.cycles-pp.shift_arg_pages
> 0.00 +0.1 0.14 ± 37% perf-profile.children.cycles-pp.get_idle_time
> 0.13 ± 37% +0.1 0.27 ± 19% perf-profile.children.cycles-pp.folio_alloc_mpol_noprof
> 0.11 ± 80% +0.1 0.26 ± 39% perf-profile.children.cycles-pp.mm_init
> 0.07 ± 49% +0.1 0.22 ± 35% perf-profile.children.cycles-pp.pte_alloc_one
> 0.22 ± 47% +0.2 0.37 ± 33% perf-profile.children.cycles-pp.scsi_mq_get_budget
> 0.06 ±103% +0.2 0.21 ± 53% perf-profile.children.cycles-pp.unlink_anon_vmas
> 0.14 ± 79% +0.2 0.30 ± 30% perf-profile.children.cycles-pp.mas_wr_node_store
> 0.04 ±107% +0.2 0.20 ± 70% perf-profile.children.cycles-pp.dyntick_save_progress_counter
> 0.05 ±132% +0.2 0.22 ± 66% perf-profile.children.cycles-pp.sysfs_kf_seq_show
> 0.05 ±136% +0.2 0.22 ± 66% perf-profile.children.cycles-pp.dev_attr_show
> 0.07 ± 89% +0.2 0.24 ± 28% perf-profile.children.cycles-pp.__cmd_record
> 0.07 ± 89% +0.2 0.24 ± 28% perf-profile.children.cycles-pp.cmd_record
> 0.05 ±125% +0.2 0.22 ± 43% perf-profile.children.cycles-pp.move_queued_task
> 0.08 ± 14% +0.2 0.26 ± 51% perf-profile.children.cycles-pp.flush_smp_call_function_queue
> 0.07 ±127% +0.2 0.25 ± 31% perf-profile.children.cycles-pp.__dentry_kill
> 0.04 ± 77% +0.2 0.22 ± 62% perf-profile.children.cycles-pp.mas_split
> 0.08 ± 27% +0.2 0.27 ± 47% perf-profile.children.cycles-pp.mas_alloc_nodes
> 0.14 ± 62% +0.2 0.34 ± 29% perf-profile.children.cycles-pp.create_elf_tables
> 0.01 ±223% +0.2 0.21 ± 60% perf-profile.children.cycles-pp.__put_user_8
> 0.12 ± 64% +0.2 0.32 ± 48% perf-profile.children.cycles-pp.seq_printf
> 0.21 ± 36% +0.2 0.43 ± 37% perf-profile.children.cycles-pp.vfs_statx
> 0.23 ± 73% +0.2 0.45 ± 38% perf-profile.children.cycles-pp.sbitmap_get
> 0.15 ±104% +0.2 0.38 ± 52% perf-profile.children.cycles-pp.get_arg_page
> 0.24 ± 70% +0.2 0.47 ± 42% perf-profile.children.cycles-pp.load_elf_interp
> 0.08 ± 52% +0.2 0.31 ± 44% perf-profile.children.cycles-pp.__get_user_8
> 0.13 ± 48% +0.3 0.38 ± 51% perf-profile.children.cycles-pp.rcu_gp_fqs_loop
> 0.21 ± 66% +0.3 0.46 ± 46% perf-profile.children.cycles-pp.vsnprintf
> 0.12 ± 61% +0.3 0.38 ± 30% perf-profile.children.cycles-pp.slab_show
> 0.15 ± 65% +0.3 0.40 ± 50% perf-profile.children.cycles-pp.rep_stos_alternative
> 0.23 ± 32% +0.3 0.49 ± 32% perf-profile.children.cycles-pp.path_lookupat
> 0.13 ± 82% +0.3 0.40 ± 50% perf-profile.children.cycles-pp.perf_evsel__read
> 0.13 ± 62% +0.3 0.40 ± 24% perf-profile.children.cycles-pp.step_into
> 0.14 ± 70% +0.3 0.42 ± 35% perf-profile.children.cycles-pp.alloc_anon_folio
> 0.23 ± 32% +0.3 0.51 ± 28% perf-profile.children.cycles-pp.filename_lookup
> 0.06 ± 88% +0.3 0.34 ± 48% perf-profile.children.cycles-pp.rseq_ip_fixup
> 0.18 ± 39% +0.3 0.47 ± 38% perf-profile.children.cycles-pp.dput
> 0.16 ± 43% +0.3 0.45 ± 37% perf-profile.children.cycles-pp.rcu_gp_kthread
> 0.28 ± 69% +0.3 0.57 ± 32% perf-profile.children.cycles-pp.__vfork
> 0.24 ± 68% +0.3 0.54 ± 29% perf-profile.children.cycles-pp.__x64_sys_sched_setaffinity
> 0.09 ± 65% +0.3 0.42 ± 38% perf-profile.children.cycles-pp.__rseq_handle_notify_resume
> 0.36 ± 40% +0.3 0.68 ± 27% perf-profile.children.cycles-pp.__do_sys_newfstatat
> 0.23 ± 66% +0.3 0.58 ± 56% perf-profile.children.cycles-pp.free_pgtables
> 0.32 ± 42% +0.4 0.67 ± 30% perf-profile.children.cycles-pp.tick_irq_enter
> 0.43 ± 39% +0.4 0.78 ± 19% perf-profile.children.cycles-pp.irq_enter_rcu
> 0.21 ± 55% +0.4 0.57 ± 41% perf-profile.children.cycles-pp.readn
> 0.16 ± 33% +0.4 0.52 ± 50% perf-profile.children.cycles-pp.dup_mmap
> 0.15 ± 77% +0.4 0.51 ± 39% perf-profile.children.cycles-pp.show_stat
> 0.34 ± 69% +0.4 0.71 ± 42% perf-profile.children.cycles-pp.zap_present_ptes
> 0.21 ± 67% +0.4 0.60 ± 23% perf-profile.children.cycles-pp.i2c_outb
> 0.23 ± 79% +0.4 0.62 ± 36% perf-profile.children.cycles-pp.seq_read
> 0.18 ± 40% +0.4 0.59 ± 46% perf-profile.children.cycles-pp.dup_mm
> 0.30 ± 67% +0.4 0.72 ± 36% perf-profile.children.cycles-pp.tmigr_handle_remote_up
> 0.20 ± 65% +0.4 0.63 ± 58% perf-profile.children.cycles-pp.sync_regs
> 0.25 ± 66% +0.4 0.68 ± 23% perf-profile.children.cycles-pp.try_address
> 0.27 ± 74% +0.5 0.72 ± 24% perf-profile.children.cycles-pp.output_poll_execute
> 0.28 ± 70% +0.5 0.74 ± 28% perf-profile.children.cycles-pp.pipe_read
> 0.26 ± 70% +0.5 0.72 ± 24% perf-profile.children.cycles-pp.__i2c_transfer
> 0.26 ± 70% +0.5 0.72 ± 24% perf-profile.children.cycles-pp.bit_xfer
> 0.26 ± 70% +0.5 0.72 ± 24% perf-profile.children.cycles-pp.drm_connector_helper_detect_from_ddc
> 0.26 ± 70% +0.5 0.72 ± 24% perf-profile.children.cycles-pp.drm_do_probe_ddc_edid
> 0.26 ± 70% +0.5 0.72 ± 24% perf-profile.children.cycles-pp.drm_helper_probe_detect_ctx
> 0.26 ± 70% +0.5 0.72 ± 24% perf-profile.children.cycles-pp.drm_probe_ddc
> 0.26 ± 70% +0.5 0.72 ± 24% perf-profile.children.cycles-pp.i2c_transfer
> 0.14 ± 61% +0.5 0.61 ± 35% perf-profile.children.cycles-pp.balance_fair
> 0.23 ± 68% +0.5 0.69 ± 34% perf-profile.children.cycles-pp.folio_wait_bit_common
> 0.28 ± 81% +0.5 0.77 ± 29% perf-profile.children.cycles-pp.pipe_write
> 0.35 ± 68% +0.5 0.84 ± 41% perf-profile.children.cycles-pp.tmigr_handle_remote
> 0.24 ± 81% +0.5 0.74 ± 34% perf-profile.children.cycles-pp.copy_strings
> 0.27 ± 72% +0.5 0.78 ± 50% perf-profile.children.cycles-pp.exec_mmap
> 0.19 ± 83% +0.5 0.70 ± 34% perf-profile.children.cycles-pp.folio_wait_writeback
> 0.30 ± 67% +0.5 0.83 ± 31% perf-profile.children.cycles-pp.do_anonymous_page
> 0.44 ± 59% +0.6 1.00 ± 46% perf-profile.children.cycles-pp.zap_pmd_range
> 0.35 ± 86% +0.6 0.93 ± 31% perf-profile.children.cycles-pp.collapse_huge_page
> 0.36 ± 84% +0.6 0.93 ± 30% perf-profile.children.cycles-pp.khugepaged
> 0.35 ± 85% +0.6 0.93 ± 30% perf-profile.children.cycles-pp.hpage_collapse_scan_pmd
> 0.35 ± 85% +0.6 0.93 ± 30% perf-profile.children.cycles-pp.khugepaged_scan_mm_slot
> 0.34 ± 63% +0.6 0.92 ± 37% perf-profile.children.cycles-pp.evlist_cpu_iterator__next
> 0.24 ± 42% +0.6 0.89 ± 47% perf-profile.children.cycles-pp.scsi_dispatch_cmd
> 0.32 ± 62% +0.7 0.99 ± 45% perf-profile.children.cycles-pp.begin_new_exec
> 0.41 ± 58% +0.7 1.10 ± 21% perf-profile.children.cycles-pp.__open64_nocancel
> 0.37 ± 48% +0.7 1.06 ± 34% perf-profile.children.cycles-pp._Fork
> 0.54 ± 66% +0.7 1.27 ± 29% perf-profile.children.cycles-pp.sched_setaffinity
> 0.78 ± 51% +0.8 1.56 ± 28% perf-profile.children.cycles-pp.link_path_walk
> 0.40 ± 35% +0.8 1.19 ± 40% perf-profile.children.cycles-pp.scsi_queue_rq
> 0.51 ± 74% +0.8 1.30 ± 39% perf-profile.children.cycles-pp.elf_load
> 0.34 ± 54% +0.8 1.15 ± 34% perf-profile.children.cycles-pp.smpboot_thread_fn
> 0.72 ± 69% +0.9 1.66 ± 33% perf-profile.children.cycles-pp.read_counters
> 0.77 ± 70% +1.0 1.76 ± 33% perf-profile.children.cycles-pp.cmd_stat
> 0.77 ± 70% +1.0 1.76 ± 33% perf-profile.children.cycles-pp.dispatch_events
> 0.77 ± 70% +1.0 1.76 ± 33% perf-profile.children.cycles-pp.process_interval
> 1.00 ± 19% +1.0 2.01 ± 26% perf-profile.children.cycles-pp.filemap_map_pages
> 1.39 ± 34% +1.1 2.48 ± 22% perf-profile.children.cycles-pp.asm_sysvec_posted_msi_notification
> 0.85 ± 68% +1.1 1.99 ± 32% perf-profile.children.cycles-pp.handle_internal_command
> 0.85 ± 68% +1.1 1.99 ± 32% perf-profile.children.cycles-pp.main
> 0.85 ± 68% +1.1 1.99 ± 32% perf-profile.children.cycles-pp.run_builtin
> 0.88 ± 64% +1.2 2.04 ± 32% perf-profile.children.cycles-pp.do_group_exit
> 0.88 ± 64% +1.2 2.05 ± 32% perf-profile.children.cycles-pp.__x64_sys_exit_group
> 0.64 ± 66% +1.2 1.87 ± 37% perf-profile.children.cycles-pp.seq_read_iter
> 1.24 ± 54% +1.2 2.48 ± 31% perf-profile.children.cycles-pp.x64_sys_call
> 1.03 ± 66% +1.3 2.30 ± 30% perf-profile.children.cycles-pp.do_exit
> 0.88 ± 64% +1.5 2.35 ± 30% perf-profile.children.cycles-pp.setlocale
> 1.08 ± 71% +1.5 2.56 ± 26% perf-profile.children.cycles-pp.mmput
> 1.05 ± 70% +1.5 2.55 ± 27% perf-profile.children.cycles-pp.exit_mmap
> 3.54 ± 34% +1.8 5.33 ± 20% perf-profile.children.cycles-pp.handle_softirqs
> 1.61 ± 55% +1.8 3.40 ± 35% perf-profile.children.cycles-pp.mmap_region
> 1.30 ± 69% +2.0 3.28 ± 32% perf-profile.children.cycles-pp.load_elf_binary
> 1.32 ± 68% +2.0 3.30 ± 31% perf-profile.children.cycles-pp.search_binary_handler
> 1.32 ± 68% +2.0 3.32 ± 31% perf-profile.children.cycles-pp.exec_binprm
> 1.45 ± 70% +2.1 3.52 ± 30% perf-profile.children.cycles-pp.bprm_execve
> 2.01 ± 71% +3.0 5.01 ± 22% perf-profile.children.cycles-pp.__x64_sys_execve
> 2.01 ± 71% +3.0 5.01 ± 23% perf-profile.children.cycles-pp.execve
> 2.00 ± 70% +3.0 5.01 ± 22% perf-profile.children.cycles-pp.do_execveat_common
> 1.18 ± 39% +3.3 4.48 ± 31% perf-profile.children.cycles-pp.dd_dispatch_request
> 2.77 ± 34% +3.6 6.33 ± 28% perf-profile.children.cycles-pp.__blk_mq_sched_dispatch_requests
> 2.77 ± 34% +3.6 6.34 ± 28% perf-profile.children.cycles-pp.blk_mq_sched_dispatch_requests
> 2.59 ± 34% +3.8 6.35 ± 27% perf-profile.children.cycles-pp.blk_mq_run_hw_queue
> 1.94 ± 37% +4.3 6.24 ± 29% perf-profile.children.cycles-pp.__blk_mq_do_dispatch_sched
> 3.76 ± 36% +5.0 8.79 ± 30% perf-profile.children.cycles-pp.__blk_flush_plug
> 3.76 ± 36% +5.0 8.79 ± 30% perf-profile.children.cycles-pp.blk_mq_dispatch_plug_list
> 3.76 ± 36% +5.0 8.79 ± 30% perf-profile.children.cycles-pp.blk_mq_flush_plug_list
> 4.84 ± 35% +5.1 9.96 ± 29% perf-profile.children.cycles-pp.iomap_submit_ioend
> 4.01 ± 55% +8.1 12.11 ± 30% perf-profile.children.cycles-pp.xfs_file_write_checks
> 3.34 ± 56% +8.3 11.60 ± 31% perf-profile.children.cycles-pp.iomap_zero_range
> 0.00 +11.2 11.22 ± 30% perf-profile.children.cycles-pp.filemap_write_and_wait_range
> 3.43 ± 46% -3.2 0.24 ± 35% perf-profile.self.cycles-pp.memset_orig
> 1.42 ± 47% -1.3 0.08 ±121% perf-profile.self.cycles-pp._copy_to_iter
> 1.07 ± 27% -0.6 0.45 ± 39% perf-profile.self.cycles-pp.clear_bhb_loop
> 0.38 ± 23% -0.3 0.12 ± 62% perf-profile.self.cycles-pp.percpu_counter_add_batch
> 0.30 ± 33% -0.2 0.09 ± 88% perf-profile.self.cycles-pp.do_syscall_64
> 0.22 ± 30% -0.2 0.05 ± 71% perf-profile.self.cycles-pp.__mod_node_page_state
> 0.13 ± 20% -0.1 0.04 ±103% perf-profile.self.cycles-pp.__page_cache_release
> 0.01 ±223% +0.1 0.11 ± 48% perf-profile.self.cycles-pp.lockref_put_return
> 0.02 ±142% +0.1 0.13 ± 65% perf-profile.self.cycles-pp.xfs_ag_block_count
> 0.01 ±223% +0.1 0.13 ± 45% perf-profile.self.cycles-pp.get_cpu_sleep_time_us
> 0.04 ±112% +0.2 0.19 ± 68% perf-profile.self.cycles-pp.dyntick_save_progress_counter
> 0.14 ± 35% +0.2 0.31 ± 35% perf-profile.self.cycles-pp.sched_balance_domains
> 0.08 ± 52% +0.2 0.31 ± 44% perf-profile.self.cycles-pp.__get_user_8
> 0.06 ± 92% +0.3 0.34 ± 60% perf-profile.self.cycles-pp.fold_vm_numa_events
> 0.20 ± 65% +0.4 0.63 ± 58% perf-profile.self.cycles-pp.sync_regs
> 0.26 ± 34% +0.5 0.75 ± 37% perf-profile.self.cycles-pp.filemap_map_pages
>
>
>
>
> Disclaimer:
> Results have been estimated based on internal Intel analysis and are provided
> for informational purposes only. Any difference in system hardware or software
> design or configuration may affect actual performance.
>
>
> --
> 0-DAY CI Kernel Test Service
> https://github.com/intel/lkp-tests/wiki
>
Powered by blists - more mailing lists