[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <816f906d-eafe-0271-9edd-7f25fd757fa0@oracle.com>
Date: Thu, 22 Apr 2021 22:26:44 -0700
From: Junxiao Bi <junxiao.bi@...cle.com>
To: kernel test robot <oliver.sang@...el.com>
Cc: 0day robot <lkp@...el.com>, LKML <linux-kernel@...r.kernel.org>,
lkp@...ts.01.org, ying.huang@...el.com, feng.tang@...el.com,
zhengjun.xing@...el.com, linux-block@...r.kernel.org,
axboe@...nel.dk
Subject: Re: [block] 658f2fb7d2:
fxmark.hdd_f2fs_dbench_client_72_directio.works/sec -21.4% regression
On 4/22/21 7:55 PM, kernel test robot wrote:
>
> Greeting,
>
> FYI, we noticed a -21.4% regression of fxmark.hdd_f2fs_dbench_client_72_directio.works/sec due to commit:
>
>
> commit: 658f2fb7d2139d7a218083b7e8688dd288c6512d ("[PATCH] block: fix io hung by block throttle")
> url: https://github.com/0day-ci/linux/commits/Junxiao-Bi/block-fix-io-hung-by-block-throttle/20210415-052219
> base: https://git.kernel.org/cgit/linux/kernel/git/axboe/linux-block.git for-next
>
> in testcase: fxmark
> on test machine: 288 threads Intel(R) Xeon Phi(TM) CPU 7295 @ 1.50GHz with 80G memory
> with following parameters:
>
> disk: 1HDD
> media: hdd
> test: dbench_client
> fstype: f2fs
> directio: directio
This test is doing directio? The throttle didn't take care direct io,
see wbt_should_throttle(), in that case, the changed code will not even
be run.
Thanks,
Junxiao.
> cpufreq_governor: performance
> ucode: 0x11
>
>
>
>
> If you fix the issue, kindly add following tag
> Reported-by: kernel test robot <oliver.sang@...el.com>
>
>
> Details are as below:
> -------------------------------------------------------------------------------------------------->
>
>
> To reproduce:
>
> git clone https://github.com/intel/lkp-tests.git
> cd lkp-tests
> bin/lkp install job.yaml # job file is attached in this email
> bin/lkp split-job --compatible job.yaml
> bin/lkp run compatible-job.yaml
>
> =========================================================================================
> compiler/cpufreq_governor/directio/disk/fstype/kconfig/media/rootfs/tbox_group/test/testcase/ucode:
> gcc-9/performance/directio/1HDD/f2fs/x86_64-rhel-8.3/hdd/debian-10.4-x86_64-20200603.cgz/lkp-knm01/dbench_client/fxmark/0x11
>
> commit:
> 901b8e44f6 ("Merge branch 'for-5.13/io_uring' into for-next")
> 658f2fb7d2 ("block: fix io hung by block throttle")
>
> 901b8e44f6b08436 658f2fb7d2139d7a218083b7e86
> ---------------- ---------------------------
> %stddev %change %stddev
> \ | \
> 66.02 ± 2% -14.2% 56.65 ± 3% fxmark.hdd_f2fs_dbench_client_36_directio.sys_sec
> 3.04 ± 2% -14.3% 2.61 ± 3% fxmark.hdd_f2fs_dbench_client_36_directio.sys_util
> 19.47 ± 3% -13.4% 16.85 ± 3% fxmark.hdd_f2fs_dbench_client_36_directio.user_sec
> 0.90 ± 3% -13.5% 0.78 ± 3% fxmark.hdd_f2fs_dbench_client_36_directio.user_util
> 128.18 ± 2% -18.2% 104.90 ± 5% fxmark.hdd_f2fs_dbench_client_36_directio.works/sec
> 8.65 -9.0% 7.87 ± 3% fxmark.hdd_f2fs_dbench_client_54_directio.softirq_sec
> 0.27 -9.1% 0.24 ± 3% fxmark.hdd_f2fs_dbench_client_54_directio.softirq_util
> 103.82 ± 3% -20.8% 82.22 ± 3% fxmark.hdd_f2fs_dbench_client_54_directio.sys_sec
> 3.19 ± 3% -20.9% 2.52 ± 3% fxmark.hdd_f2fs_dbench_client_54_directio.sys_util
> 29.38 ± 2% -18.4% 23.98 fxmark.hdd_f2fs_dbench_client_54_directio.user_sec
> 0.90 ± 2% -18.5% 0.74 fxmark.hdd_f2fs_dbench_client_54_directio.user_util
> 162.84 ± 4% -24.8% 122.39 ± 3% fxmark.hdd_f2fs_dbench_client_54_directio.works/sec
> 10.29 -8.7% 9.40 ± 3% fxmark.hdd_f2fs_dbench_client_72_directio.softirq_sec
> 0.24 -8.8% 0.22 ± 3% fxmark.hdd_f2fs_dbench_client_72_directio.softirq_util
> 132.66 ± 2% -19.3% 107.01 ± 2% fxmark.hdd_f2fs_dbench_client_72_directio.sys_sec
> 3.05 ± 2% -19.4% 2.46 ± 2% fxmark.hdd_f2fs_dbench_client_72_directio.sys_util
> 29.72 ± 3% -18.2% 24.31 ± 2% fxmark.hdd_f2fs_dbench_client_72_directio.user_sec
> 0.68 ± 3% -18.3% 0.56 ± 2% fxmark.hdd_f2fs_dbench_client_72_directio.user_util
> 183.55 ± 3% -21.4% 144.32 ± 3% fxmark.hdd_f2fs_dbench_client_72_directio.works/sec
> 26566804 -9.9% 23936362 fxmark.time.file_system_outputs
> 72184 -2.6% 70322 fxmark.time.involuntary_context_switches
> 88.33 -15.9% 74.29 fxmark.time.percent_of_cpu_this_job_got
> 357.13 -15.9% 300.26 fxmark.time.system_time
> 91.39 -14.5% 78.18 fxmark.time.user_time
> 930788 ± 3% -30.6% 645881 ± 2% fxmark.time.voluntary_context_switches
> 9.40 -2.7% 9.14 iostat.cpu.system
> 2.70 -2.0% 2.64 iostat.cpu.user
> 6281 ± 6% -21.4% 4939 ± 8% meminfo.Active(anon)
> 71273 ± 4% -20.3% 56818 ± 7% meminfo.Shmem
> 6267 ± 5% -21.1% 4946 ± 8% numa-meminfo.node0.Active(anon)
> 71236 ± 4% -20.2% 56833 ± 7% numa-meminfo.node0.Shmem
> 6729 -12.4% 5897 vmstat.io.bo
> 6385 ± 2% -17.6% 5262 vmstat.system.cs
> 6088 ± 2% -18.6% 4958 perf-stat.i.context-switches
> 176.01 -15.1% 149.46 perf-stat.i.cpu-migrations
> 6074 ± 2% -18.6% 4945 perf-stat.ps.context-switches
> 175.89 -15.1% 149.36 perf-stat.ps.cpu-migrations
> 1570 ± 6% -21.5% 1233 ± 8% numa-vmstat.node0.nr_active_anon
> 2302914 -11.0% 2050036 numa-vmstat.node0.nr_dirtied
> 17815 ± 4% -20.3% 14202 ± 7% numa-vmstat.node0.nr_shmem
> 540895 -14.6% 461814 numa-vmstat.node0.nr_written
> 1570 ± 6% -21.5% 1233 ± 8% numa-vmstat.node0.nr_zone_active_anon
> 2679 ± 4% -15.2% 2272 ± 4% slabinfo.dmaengine-unmap-16.active_objs
> 3266 ± 7% -13.0% 2841 ± 4% slabinfo.dmaengine-unmap-16.num_objs
> 5826 ± 4% -12.6% 5093 ± 3% slabinfo.ext4_io_end.active_objs
> 5994 ± 4% -12.3% 5257 ± 3% slabinfo.ext4_io_end.num_objs
> 18698 ± 2% -9.5% 16920 slabinfo.ext4_pending_reservation.active_objs
> 18783 ± 2% -9.5% 17007 slabinfo.ext4_pending_reservation.num_objs
> 32425 ± 2% -11.6% 28671 ± 2% slabinfo.f2fs_free_nid.active_objs
> 32622 ± 2% -11.6% 28825 ± 2% slabinfo.f2fs_free_nid.num_objs
> 1584 ± 5% -22.1% 1233 ± 8% proc-vmstat.nr_active_anon
> 20434 -2.0% 20027 proc-vmstat.nr_active_file
> 3316154 -9.9% 2986902 proc-vmstat.nr_dirtied
> 405121 -0.9% 401332 proc-vmstat.nr_file_pages
> 82265 -4.4% 78653 proc-vmstat.nr_inactive_anon
> 17866 ± 4% -20.2% 14252 ± 7% proc-vmstat.nr_shmem
> 850118 -12.4% 744638 proc-vmstat.nr_written
> 1584 ± 5% -22.1% 1233 ± 8% proc-vmstat.nr_zone_active_anon
> 20434 -2.0% 20027 proc-vmstat.nr_zone_active_file
> 82265 -4.4% 78653 proc-vmstat.nr_zone_inactive_anon
> 4971269 -6.1% 4667688 proc-vmstat.numa_hit
> 4971268 -6.1% 4667687 proc-vmstat.numa_local
> 293150 -4.6% 279668 proc-vmstat.pgactivate
> 5110253 -5.9% 4808331 proc-vmstat.pgalloc_normal
> 144088 -17.3% 119141 proc-vmstat.pgdeactivate
> 5087969 -5.8% 4791351 proc-vmstat.pgfree
> 3400472 -12.4% 2978555 proc-vmstat.pgpgout
> 126836 -16.8% 105510 proc-vmstat.pgrotated
> 616731 -5.1% 585318 interrupts.CAL:Function_call_interrupts
> 6785 ± 4% -11.8% 5983 ± 4% interrupts.CPU11.CAL:Function_call_interrupts
> 6543 ± 3% -12.3% 5735 ± 3% interrupts.CPU17.CAL:Function_call_interrupts
> 34.17 ± 19% +185.6% 97.57 ±108% interrupts.CPU18.TLB:TLB_shootdowns
> 402.33 ± 33% -42.1% 232.86 ± 21% interrupts.CPU2.NMI:Non-maskable_interrupts
> 402.33 ± 33% -42.1% 232.86 ± 21% interrupts.CPU2.PMI:Performance_monitoring_interrupts
> 5327 ± 6% -15.6% 4496 ± 6% interrupts.CPU20.CAL:Function_call_interrupts
> 5017 ± 4% -12.6% 4384 ± 6% interrupts.CPU21.CAL:Function_call_interrupts
> 5106 ± 4% -12.6% 4461 ± 6% interrupts.CPU22.CAL:Function_call_interrupts
> 5046 ± 3% -13.0% 4389 ± 5% interrupts.CPU25.CAL:Function_call_interrupts
> 5126 ± 2% -13.7% 4422 ± 5% interrupts.CPU26.CAL:Function_call_interrupts
> 5025 ± 5% -14.1% 4318 ± 3% interrupts.CPU27.CAL:Function_call_interrupts
> 5074 ± 5% -12.3% 4451 ± 4% interrupts.CPU28.CAL:Function_call_interrupts
> 4985 ± 3% -12.0% 4386 ± 3% interrupts.CPU29.CAL:Function_call_interrupts
> 5168 ± 7% -14.2% 4436 ± 6% interrupts.CPU31.CAL:Function_call_interrupts
> 5390 ± 10% -16.7% 4492 ± 8% interrupts.CPU33.CAL:Function_call_interrupts
> 4929 ± 2% -11.1% 4379 ± 5% interrupts.CPU34.CAL:Function_call_interrupts
> 4942 ± 3% -11.7% 4366 ± 5% interrupts.CPU35.CAL:Function_call_interrupts
> 3625 ± 3% -8.2% 3329 ± 5% interrupts.CPU39.CAL:Function_call_interrupts
> 3658 ± 4% -10.0% 3290 ± 3% interrupts.CPU40.CAL:Function_call_interrupts
> 3825 ± 6% -16.3% 3203 ± 6% interrupts.CPU42.CAL:Function_call_interrupts
> 3665 ± 4% -15.7% 3088 ± 10% interrupts.CPU43.CAL:Function_call_interrupts
> 3664 ± 6% -10.5% 3279 ± 6% interrupts.CPU44.CAL:Function_call_interrupts
> 3788 ± 7% -13.4% 3280 ± 4% interrupts.CPU46.CAL:Function_call_interrupts
> 3895 ± 15% -16.3% 3261 ± 4% interrupts.CPU47.CAL:Function_call_interrupts
> 3945 ± 7% -15.0% 3352 ± 5% interrupts.CPU48.CAL:Function_call_interrupts
> 3838 ± 10% -15.7% 3237 ± 4% interrupts.CPU49.CAL:Function_call_interrupts
> 6927 ± 5% -13.5% 5991 ± 8% interrupts.CPU5.CAL:Function_call_interrupts
> 3459 ± 6% -13.1% 3005 ± 6% interrupts.CPU53.CAL:Function_call_interrupts
> 2564 ± 3% -10.2% 2302 ± 7% interrupts.CPU54.CAL:Function_call_interrupts
> 2517 ± 2% -8.9% 2294 ± 6% interrupts.CPU56.CAL:Function_call_interrupts
> 2517 ± 3% -10.7% 2246 ± 7% interrupts.CPU70.CAL:Function_call_interrupts
> 20310 ± 4% -13.1% 17648 ± 8% interrupts.NMI:Non-maskable_interrupts
> 20310 ± 4% -13.1% 17648 ± 8% interrupts.PMI:Performance_monitoring_interrupts
> 0.06 ± 91% -86.0% 0.01 ±130% perf-sched.sch_delay.avg.ms.preempt_schedule_common.__cond_resched.pagecache_get_page.__get_node_page.f2fs_update_inode_page
> 0.04 ± 6% +16.1% 0.05 ± 8% perf-sched.sch_delay.avg.ms.rwsem_down_read_slowpath.__submit_merged_write_cond.f2fs_fsync_node_pages.f2fs_do_sync_file
> 0.03 ± 9% +110.0% 0.07 ± 97% perf-sched.sch_delay.avg.ms.rwsem_down_write_slowpath.set_node_addr.__write_node_page.f2fs_fsync_node_pages
> 0.67 ± 66% -66.9% 0.22 ± 43% perf-sched.sch_delay.max.ms.issue_flush_thread.kthread.ret_from_fork
> 0.19 ± 89% -93.9% 0.01 ±127% perf-sched.sch_delay.max.ms.preempt_schedule_common.__cond_resched.pagecache_get_page.__get_node_page.f2fs_update_inode_page
> 0.45 ± 43% -56.5% 0.20 ± 68% perf-sched.sch_delay.max.ms.rwsem_down_read_slowpath.f2fs_submit_merged_ipu_write.f2fs_wait_on_page_writeback.f2fs_wait_on_node_pages_writeback
> 8.98 ±198% -93.0% 0.63 ± 28% perf-sched.sch_delay.max.ms.rwsem_down_write_slowpath.set_node_addr.f2fs_new_node_page.f2fs_new_inode_page
> 0.08 ± 34% -41.0% 0.05 ± 38% perf-sched.sch_delay.max.ms.schedule_hrtimeout_range_clock.poll_schedule_timeout.constprop.0.do_sys_poll
> 45.26 ± 6% +28.5% 58.17 ± 7% perf-sched.total_wait_and_delay.average.ms
> 61502 ± 6% -27.7% 44475 ± 11% perf-sched.total_wait_and_delay.count.ms
> 45.23 ± 6% +28.5% 58.14 ± 7% perf-sched.total_wait_time.average.ms
> 6.59 ± 10% +39.7% 9.20 ± 10% perf-sched.wait_and_delay.avg.ms.io_schedule.wait_on_page_bit.wait_on_page_writeback.f2fs_wait_on_page_writeback
> 4.48 ± 9% -53.8% 2.07 ± 71% perf-sched.wait_and_delay.avg.ms.rwsem_down_read_slowpath.f2fs_need_inode_block_update.f2fs_do_sync_file.do_fsync
> 9.17 ± 14% -31.8% 6.25 ± 11% perf-sched.wait_and_delay.avg.ms.rwsem_down_write_slowpath.set_node_addr.__write_node_page.f2fs_fsync_node_pages
> 7540 ± 11% -26.6% 5536 ± 17% perf-sched.wait_and_delay.count.io_schedule.wait_on_page_bit.wait_on_page_writeback.__filemap_fdatawait_range
> 19838 ± 8% -35.6% 12768 ± 15% perf-sched.wait_and_delay.count.io_schedule.wait_on_page_bit.wait_on_page_writeback.f2fs_wait_on_page_writeback
> 911.33 ± 6% -29.8% 639.43 ± 10% perf-sched.wait_and_delay.count.rwsem_down_read_slowpath.__submit_merged_write_cond.f2fs_fsync_node_pages.f2fs_do_sync_file
> 2795 ± 6% -43.1% 1589 ± 20% perf-sched.wait_and_delay.count.rwsem_down_read_slowpath.__submit_merged_write_cond.f2fs_wait_on_page_writeback.f2fs_wait_on_node_pages_writeback
> 491.67 ± 14% -46.0% 265.29 ± 49% perf-sched.wait_and_delay.count.rwsem_down_read_slowpath.f2fs_get_node_info.__write_node_page.f2fs_fsync_node_pages
> 1726 ± 28% -48.5% 889.29 ± 42% perf-sched.wait_and_delay.count.rwsem_down_read_slowpath.f2fs_get_node_info.f2fs_do_write_data_page.f2fs_write_single_data_page
> 839.00 ± 31% -59.6% 338.57 ± 56% perf-sched.wait_and_delay.count.rwsem_down_read_slowpath.f2fs_get_node_info.truncate_node.f2fs_remove_inode_page
> 509.67 ± 12% -46.6% 272.00 ± 48% perf-sched.wait_and_delay.count.rwsem_down_read_slowpath.f2fs_need_dentry_mark.f2fs_fsync_node_pages.f2fs_do_sync_file
> 497.83 ± 12% -58.8% 205.00 ± 67% perf-sched.wait_and_delay.count.rwsem_down_read_slowpath.f2fs_need_inode_block_update.f2fs_do_sync_file.do_fsync
> 267.17 ± 23% -75.7% 64.86 ±128% perf-sched.wait_and_delay.count.rwsem_down_write_slowpath.f2fs_merge_page_bio.f2fs_inplace_write_data.f2fs_do_write_data_page
> 1539 ± 5% -49.4% 779.14 ± 15% perf-sched.wait_and_delay.count.rwsem_down_write_slowpath.f2fs_submit_page_write.do_write_page.f2fs_do_write_node_page
> 791.00 ± 13% -45.2% 433.57 ± 27% perf-sched.wait_and_delay.count.rwsem_down_write_slowpath.set_node_addr.__write_node_page.f2fs_fsync_node_pages
> 910.67 ± 20% -54.1% 417.86 ± 38% perf-sched.wait_and_delay.count.rwsem_down_write_slowpath.set_node_addr.f2fs_new_node_page.f2fs_new_inode_page
> 856.83 ± 29% -60.2% 341.14 ± 57% perf-sched.wait_and_delay.count.rwsem_down_write_slowpath.set_node_addr.truncate_node.f2fs_remove_inode_page
> 3465 ± 2% -21.4% 2725 ± 8% perf-sched.wait_and_delay.count.schedule_timeout.wait_for_completion.f2fs_issue_flush.f2fs_do_sync_file
> 142.28 ± 8% +1126.8% 1745 ± 97% perf-sched.wait_and_delay.max.ms.io_schedule.wait_on_page_bit.wait_on_page_writeback.__filemap_fdatawait_range
> 5186 ± 2% +22.8% 6370 ± 11% perf-sched.wait_and_delay.max.ms.worker_thread.kthread.ret_from_fork
> 6.56 ± 10% +39.7% 9.17 ± 10% perf-sched.wait_time.avg.ms.io_schedule.wait_on_page_bit.wait_on_page_writeback.f2fs_wait_on_page_writeback
> 15.00 ± 88% -75.6% 3.65 ±165% perf-sched.wait_time.avg.ms.preempt_schedule_common.__cond_resched.pagecache_get_page.__get_node_page.f2fs_find_in_inline_dir
> 2.74 ± 17% +90.3% 5.22 ± 24% perf-sched.wait_time.avg.ms.rwsem_down_read_slowpath.f2fs_get_node_info.__write_node_page.f2fs_fsync_node_pages
> 4.46 ± 9% -43.4% 2.52 ± 35% perf-sched.wait_time.avg.ms.rwsem_down_read_slowpath.f2fs_need_inode_block_update.f2fs_do_sync_file.do_fsync
> 9.14 ± 14% -32.3% 6.18 ± 12% perf-sched.wait_time.avg.ms.rwsem_down_write_slowpath.set_node_addr.__write_node_page.f2fs_fsync_node_pages
> 142.23 ± 8% +1127.2% 1745 ± 97% perf-sched.wait_time.max.ms.io_schedule.wait_on_page_bit.wait_on_page_writeback.__filemap_fdatawait_range
> 33.23 ± 80% -62.6% 12.44 ±158% perf-sched.wait_time.max.ms.preempt_schedule_common.__cond_resched.mnt_want_write.do_unlinkat.do_syscall_64
> 52.12 ± 53% -73.9% 13.59 ±161% perf-sched.wait_time.max.ms.preempt_schedule_common.__cond_resched.pagecache_get_page.__get_node_page.f2fs_update_inode_page
> 5186 ± 2% +22.8% 6370 ± 11% perf-sched.wait_time.max.ms.worker_thread.kthread.ret_from_fork
> 43.08 ± 8% -8.5 34.60 ± 11% perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe
> 41.46 ± 9% -8.2 33.29 ± 11% perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe
> 4.91 ± 10% -1.2 3.67 ± 12% perf-profile.calltrace.cycles-pp.vfs_write.ksys_pwrite64.do_syscall_64.entry_SYSCALL_64_after_hwframe
> 4.94 ± 10% -1.2 3.70 ± 12% perf-profile.calltrace.cycles-pp.ksys_pwrite64.do_syscall_64.entry_SYSCALL_64_after_hwframe
> 4.77 ± 11% -1.2 3.57 ± 12% perf-profile.calltrace.cycles-pp.new_sync_write.vfs_write.ksys_pwrite64.do_syscall_64.entry_SYSCALL_64_after_hwframe
> 4.72 ± 11% -1.2 3.53 ± 12% perf-profile.calltrace.cycles-pp.f2fs_file_write_iter.new_sync_write.vfs_write.ksys_pwrite64.do_syscall_64
> 5.97 ± 9% -1.2 4.80 ± 11% perf-profile.calltrace.cycles-pp.path_openat.do_filp_open.do_sys_openat2.do_sys_open.do_syscall_64
> 6.06 ± 9% -1.2 4.89 ± 10% perf-profile.calltrace.cycles-pp.do_filp_open.do_sys_openat2.do_sys_open.do_syscall_64.entry_SYSCALL_64_after_hwframe
> 3.52 ± 13% -0.9 2.60 ± 14% perf-profile.calltrace.cycles-pp.do_unlinkat.do_syscall_64.entry_SYSCALL_64_after_hwframe
> 3.06 ± 12% -0.8 2.30 ± 14% perf-profile.calltrace.cycles-pp.__generic_file_write_iter.f2fs_file_write_iter.new_sync_write.vfs_write.ksys_pwrite64
> 2.62 ± 15% -0.7 1.91 ± 16% perf-profile.calltrace.cycles-pp.evict.do_unlinkat.do_syscall_64.entry_SYSCALL_64_after_hwframe
> 2.52 ± 15% -0.7 1.82 ± 16% perf-profile.calltrace.cycles-pp.f2fs_evict_inode.evict.do_unlinkat.do_syscall_64.entry_SYSCALL_64_after_hwframe
> 3.45 ± 5% -0.7 2.79 ± 11% perf-profile.calltrace.cycles-pp.__do_sys_newstat.do_syscall_64.entry_SYSCALL_64_after_hwframe
> 2.16 ± 15% -0.6 1.52 ± 15% perf-profile.calltrace.cycles-pp.f2fs_create.path_openat.do_filp_open.do_sys_openat2.do_sys_open
> 3.30 ± 6% -0.6 2.67 ± 11% perf-profile.calltrace.cycles-pp.vfs_statx.__do_sys_newstat.do_syscall_64.entry_SYSCALL_64_after_hwframe
> 2.43 ± 9% -0.6 1.84 ± 10% perf-profile.calltrace.cycles-pp.f2fs_wait_on_node_pages_writeback.f2fs_do_sync_file.do_fsync.__x64_sys_fsync.do_syscall_64
> 2.37 ± 9% -0.6 1.80 ± 11% perf-profile.calltrace.cycles-pp.f2fs_wait_on_page_writeback.f2fs_wait_on_node_pages_writeback.f2fs_do_sync_file.do_fsync.__x64_sys_fsync
> 2.19 ± 6% -0.4 1.75 ± 10% perf-profile.calltrace.cycles-pp.path_lookupat.filename_lookup.vfs_statx.__do_sys_newstat.do_syscall_64
> 1.39 ± 16% -0.4 0.96 ± 17% perf-profile.calltrace.cycles-pp.f2fs_do_add_link.f2fs_create.path_openat.do_filp_open.do_sys_openat2
> 1.34 ± 16% -0.4 0.91 ± 18% perf-profile.calltrace.cycles-pp.f2fs_add_dentry.f2fs_do_add_link.f2fs_create.path_openat.do_filp_open
> 1.33 ± 16% -0.4 0.90 ± 18% perf-profile.calltrace.cycles-pp.f2fs_add_inline_entry.f2fs_add_dentry.f2fs_do_add_link.f2fs_create.path_openat
> 1.71 ± 15% -0.4 1.29 ± 12% perf-profile.calltrace.cycles-pp.schedule_idle.do_idle.cpu_startup_entry.start_secondary.secondary_startup_64_no_verify
> 1.02 ± 14% -0.4 0.60 ± 43% perf-profile.calltrace.cycles-pp.f2fs_remove_inode_page.f2fs_evict_inode.evict.do_unlinkat.do_syscall_64
> 1.69 ± 15% -0.4 1.27 ± 12% perf-profile.calltrace.cycles-pp.__schedule.schedule_idle.do_idle.cpu_startup_entry.start_secondary
> 1.59 ± 7% -0.4 1.18 ± 11% perf-profile.calltrace.cycles-pp.link_path_walk.path_lookupat.filename_lookup.vfs_statx.__do_sys_newstat
> 1.47 ± 9% -0.4 1.06 ± 8% perf-profile.calltrace.cycles-pp.f2fs_preallocate_blocks.f2fs_file_write_iter.new_sync_write.vfs_write.ksys_pwrite64
> 0.94 ± 14% -0.4 0.54 ± 44% perf-profile.calltrace.cycles-pp.truncate_node.f2fs_remove_inode_page.f2fs_evict_inode.evict.do_unlinkat
> 1.78 ± 10% -0.4 1.42 ± 16% perf-profile.calltrace.cycles-pp.ksys_pread64.do_syscall_64.entry_SYSCALL_64_after_hwframe
> 1.74 ± 11% -0.3 1.39 ± 17% perf-profile.calltrace.cycles-pp.vfs_read.ksys_pread64.do_syscall_64.entry_SYSCALL_64_after_hwframe
> 0.77 ± 15% -0.3 0.45 ± 65% perf-profile.calltrace.cycles-pp.perf_event_output_forward.__perf_event_overflow.perf_swevent_overflow.perf_tp_event.perf_trace_sched_switch
> 1.07 ± 7% -0.3 0.75 ± 9% perf-profile.calltrace.cycles-pp.f2fs_convert_inline_inode.f2fs_preallocate_blocks.f2fs_file_write_iter.new_sync_write.vfs_write
> 0.61 ± 9% -0.3 0.31 ± 86% perf-profile.calltrace.cycles-pp.user_path_at_empty.vfs_statx.__do_sys_newstat.do_syscall_64.entry_SYSCALL_64_after_hwframe
> 0.62 ± 12% -0.3 0.32 ± 87% perf-profile.calltrace.cycles-pp.truncate_inode_pages_range.f2fs_evict_inode.evict.do_unlinkat.do_syscall_64
> 1.46 ± 10% -0.3 1.18 ± 20% perf-profile.calltrace.cycles-pp.f2fs_file_read_iter.new_sync_read.vfs_read.ksys_pread64.do_syscall_64
> 1.39 ± 10% -0.3 1.11 ± 19% perf-profile.calltrace.cycles-pp.filemap_read.f2fs_file_read_iter.new_sync_read.vfs_read.ksys_pread64
> 0.91 ± 8% -0.3 0.64 ± 13% perf-profile.calltrace.cycles-pp.f2fs_convert_inline_page.f2fs_convert_inline_inode.f2fs_preallocate_blocks.f2fs_file_write_iter.new_sync_write
> 1.40 ± 9% -0.3 1.13 ± 15% perf-profile.calltrace.cycles-pp.syscall_exit_to_user_mode.entry_SYSCALL_64_after_hwframe
> 1.00 ± 15% -0.2 0.77 ± 14% perf-profile.calltrace.cycles-pp.__filemap_fdatawait_range.file_write_and_wait_range.f2fs_do_sync_file.do_fsync.__x64_sys_fsync
> 0.95 ± 15% -0.2 0.72 ± 15% perf-profile.calltrace.cycles-pp.wait_on_page_writeback.__filemap_fdatawait_range.file_write_and_wait_range.f2fs_do_sync_file.do_fsync
> 0.94 ± 15% -0.2 0.71 ± 15% perf-profile.calltrace.cycles-pp.wait_on_page_bit.wait_on_page_writeback.__filemap_fdatawait_range.file_write_and_wait_range.f2fs_do_sync_file
> 0.91 ± 15% -0.2 0.70 ± 16% perf-profile.calltrace.cycles-pp.io_schedule.wait_on_page_bit.wait_on_page_writeback.__filemap_fdatawait_range.file_write_and_wait_range
> 0.90 ± 15% -0.2 0.69 ± 15% perf-profile.calltrace.cycles-pp.schedule.io_schedule.wait_on_page_bit.wait_on_page_writeback.__filemap_fdatawait_range
> 1.16 ± 10% -0.2 0.96 ± 15% perf-profile.calltrace.cycles-pp.exit_to_user_mode_prepare.syscall_exit_to_user_mode.entry_SYSCALL_64_after_hwframe
> 1.12 ± 9% -0.2 0.93 ± 13% perf-profile.calltrace.cycles-pp.link_path_walk.path_openat.do_filp_open.do_sys_openat2.do_sys_open
> 0.83 ± 9% -0.2 0.66 ± 10% perf-profile.calltrace.cycles-pp.walk_component.link_path_walk.path_lookupat.filename_lookup.vfs_statx
> 0.74 ± 9% +0.1 0.88 ± 8% perf-profile.calltrace.cycles-pp.update_sd_lb_stats.find_busiest_group.load_balance.rebalance_domains.__softirqentry_text_start
> 0.82 ± 8% +0.1 0.96 ± 7% perf-profile.calltrace.cycles-pp.find_busiest_group.load_balance.rebalance_domains.__softirqentry_text_start.irq_exit_rcu
> 1.16 ± 4% +0.2 1.38 ± 9% perf-profile.calltrace.cycles-pp.load_balance.rebalance_domains.__softirqentry_text_start.irq_exit_rcu.sysvec_apic_timer_interrupt
> 1.36 ± 4% +0.2 1.61 ± 7% perf-profile.calltrace.cycles-pp.rebalance_domains.__softirqentry_text_start.irq_exit_rcu.sysvec_apic_timer_interrupt.asm_sysvec_apic_timer_interrupt
> 1.53 ± 4% +0.3 1.84 ± 11% perf-profile.calltrace.cycles-pp.scheduler_tick.update_process_times.tick_sched_handle.tick_sched_timer.__hrtimer_run_queues
> 1.05 ± 5% +0.3 1.36 ± 7% perf-profile.calltrace.cycles-pp.ktime_get.clockevents_program_event.hrtimer_interrupt.__sysvec_apic_timer_interrupt.sysvec_apic_timer_interrupt
> 3.10 ± 2% +0.7 3.80 ± 9% perf-profile.calltrace.cycles-pp.update_process_times.tick_sched_handle.tick_sched_timer.__hrtimer_run_queues.hrtimer_interrupt
> 3.54 ± 2% +0.7 4.26 ± 8% perf-profile.calltrace.cycles-pp.__softirqentry_text_start.irq_exit_rcu.sysvec_apic_timer_interrupt.asm_sysvec_apic_timer_interrupt.cpuidle_enter_state
> 3.35 ± 2% +0.8 4.12 ± 10% perf-profile.calltrace.cycles-pp.tick_sched_handle.tick_sched_timer.__hrtimer_run_queues.hrtimer_interrupt.__sysvec_apic_timer_interrupt
> 4.14 ± 3% +0.8 4.97 ± 9% perf-profile.calltrace.cycles-pp.irq_exit_rcu.sysvec_apic_timer_interrupt.asm_sysvec_apic_timer_interrupt.cpuidle_enter_state.cpuidle_enter
> 43.33 ± 7% +8.1 51.42 ± 9% perf-profile.calltrace.cycles-pp.secondary_startup_64_no_verify
> 42.38 ± 7% +8.2 50.54 ± 9% perf-profile.calltrace.cycles-pp.do_idle.cpu_startup_entry.start_secondary.secondary_startup_64_no_verify
> 42.46 ± 7% +8.2 50.64 ± 9% perf-profile.calltrace.cycles-pp.start_secondary.secondary_startup_64_no_verify
> 42.46 ± 7% +8.2 50.64 ± 9% perf-profile.calltrace.cycles-pp.cpu_startup_entry.start_secondary.secondary_startup_64_no_verify
> 48.00 ± 6% -7.9 40.13 ± 9% perf-profile.children.cycles-pp.entry_SYSCALL_64_after_hwframe
> 46.03 ± 7% -7.6 38.47 ± 9% perf-profile.children.cycles-pp.do_syscall_64
> 8.17 ± 10% -1.9 6.28 ± 10% perf-profile.children.cycles-pp.__schedule
> 6.38 ± 9% -1.5 4.89 ± 11% perf-profile.children.cycles-pp.schedule
> 6.11 ± 10% -1.3 4.77 ± 12% perf-profile.children.cycles-pp.perf_tp_event
> 5.66 ± 9% -1.3 4.33 ± 10% perf-profile.children.cycles-pp.vfs_write
> 5.49 ± 9% -1.3 4.20 ± 10% perf-profile.children.cycles-pp.new_sync_write
> 5.86 ± 11% -1.3 4.58 ± 12% perf-profile.children.cycles-pp.perf_swevent_overflow
> 5.83 ± 11% -1.3 4.55 ± 12% perf-profile.children.cycles-pp.__perf_event_overflow
> 5.77 ± 11% -1.3 4.50 ± 12% perf-profile.children.cycles-pp.perf_event_output_forward
> 4.94 ± 10% -1.2 3.70 ± 12% perf-profile.children.cycles-pp.ksys_pwrite64
> 4.72 ± 11% -1.2 3.53 ± 12% perf-profile.children.cycles-pp.f2fs_file_write_iter
> 4.99 ± 11% -1.2 3.81 ± 13% perf-profile.children.cycles-pp.perf_prepare_sample
> 4.50 ± 11% -1.1 3.40 ± 13% perf-profile.children.cycles-pp.get_perf_callchain
> 4.52 ± 12% -1.1 3.43 ± 13% perf-profile.children.cycles-pp.perf_callchain
> 3.69 ± 12% -1.0 2.73 ± 12% perf-profile.children.cycles-pp.perf_callchain_kernel
> 3.52 ± 13% -0.9 2.60 ± 14% perf-profile.children.cycles-pp.do_unlinkat
> 3.54 ± 11% -0.8 2.72 ± 12% perf-profile.children.cycles-pp.__generic_file_write_iter
> 3.31 ± 13% -0.8 2.55 ± 12% perf-profile.children.cycles-pp.generic_perform_write
> 3.08 ± 10% -0.7 2.35 ± 10% perf-profile.children.cycles-pp.io_schedule
> 2.64 ± 15% -0.7 1.93 ± 16% perf-profile.children.cycles-pp.evict
> 2.53 ± 15% -0.7 1.83 ± 16% perf-profile.children.cycles-pp.f2fs_evict_inode
> 3.14 ± 9% -0.7 2.44 ± 11% perf-profile.children.cycles-pp.wait_on_page_writeback
> 2.93 ± 8% -0.7 2.24 ± 11% perf-profile.children.cycles-pp.f2fs_wait_on_page_writeback
> 3.11 ± 9% -0.7 2.41 ± 11% perf-profile.children.cycles-pp.wait_on_page_bit
> 1.45 ± 20% -0.7 0.79 ± 22% perf-profile.children.cycles-pp.rwsem_down_write_slowpath
> 2.16 ± 15% -0.6 1.52 ± 15% perf-profile.children.cycles-pp.f2fs_create
> 3.93 ± 4% -0.6 3.33 ± 10% perf-profile.children.cycles-pp.__do_sys_newstat
> 3.79 ± 5% -0.6 3.20 ± 10% perf-profile.children.cycles-pp.vfs_statx
> 2.43 ± 9% -0.6 1.84 ± 10% perf-profile.children.cycles-pp.f2fs_wait_on_node_pages_writeback
> 3.30 ± 6% -0.6 2.71 ± 10% perf-profile.children.cycles-pp.link_path_walk
> 2.20 ± 13% -0.6 1.65 ± 13% perf-profile.children.cycles-pp.perf_trace_sched_switch
> 1.26 ± 11% -0.5 0.74 ± 21% perf-profile.children.cycles-pp.rwsem_down_read_slowpath
> 1.54 ± 15% -0.5 1.03 ± 17% perf-profile.children.cycles-pp.do_write_page
> 0.98 ± 22% -0.5 0.47 ± 18% perf-profile.children.cycles-pp.native_queued_spin_lock_slowpath
> 1.96 ± 13% -0.5 1.47 ± 16% perf-profile.children.cycles-pp.pagecache_get_page
> 2.11 ± 11% -0.5 1.63 ± 9% perf-profile.children.cycles-pp.dequeue_task_fair
> 2.33 ± 11% -0.5 1.85 ± 15% perf-profile.children.cycles-pp.newidle_balance
> 1.45 ± 16% -0.5 0.99 ± 16% perf-profile.children.cycles-pp.f2fs_do_add_link
> 2.02 ± 11% -0.4 1.57 ± 10% perf-profile.children.cycles-pp.dequeue_entity
> 1.39 ± 16% -0.4 0.94 ± 17% perf-profile.children.cycles-pp.f2fs_add_dentry
> 1.38 ± 16% -0.4 0.94 ± 17% perf-profile.children.cycles-pp.f2fs_add_inline_entry
> 1.74 ± 15% -0.4 1.30 ± 12% perf-profile.children.cycles-pp.schedule_idle
> 3.00 ± 5% -0.4 2.58 ± 9% perf-profile.children.cycles-pp.filename_lookup
> 1.48 ± 9% -0.4 1.06 ± 8% perf-profile.children.cycles-pp.f2fs_preallocate_blocks
> 1.78 ± 10% -0.4 1.42 ± 16% perf-profile.children.cycles-pp.ksys_pread64
> 1.02 ± 14% -0.4 0.67 ± 18% perf-profile.children.cycles-pp.f2fs_remove_inode_page
> 0.95 ± 14% -0.3 0.60 ± 21% perf-profile.children.cycles-pp.truncate_node
> 1.10 ± 16% -0.3 0.75 ± 21% perf-profile.children.cycles-pp.f2fs_outplace_write_data
> 0.78 ± 19% -0.3 0.44 ± 20% perf-profile.children.cycles-pp.rwsem_wake
> 1.09 ± 7% -0.3 0.76 ± 9% perf-profile.children.cycles-pp.f2fs_convert_inline_inode
> 1.38 ± 11% -0.3 1.06 ± 11% perf-profile.children.cycles-pp.perf_trace_sched_wakeup_template
> 0.76 ± 20% -0.3 0.44 ± 20% perf-profile.children.cycles-pp.wake_up_q
> 1.50 ± 9% -0.3 1.20 ± 17% perf-profile.children.cycles-pp.filemap_read
> 0.81 ± 15% -0.3 0.52 ± 16% perf-profile.children.cycles-pp.f2fs_submit_page_write
> 0.63 ± 17% -0.3 0.34 ± 42% perf-profile.children.cycles-pp.f2fs_get_node_info
> 1.39 ± 12% -0.3 1.10 ± 15% perf-profile.children.cycles-pp.__unwind_start
> 0.52 ± 29% -0.3 0.24 ± 50% perf-profile.children.cycles-pp.osq_lock
> 1.46 ± 10% -0.3 1.18 ± 20% perf-profile.children.cycles-pp.f2fs_file_read_iter
> 1.19 ± 14% -0.3 0.91 ± 16% perf-profile.children.cycles-pp._raw_spin_lock
> 0.91 ± 8% -0.3 0.64 ± 13% perf-profile.children.cycles-pp.f2fs_convert_inline_page
> 0.95 ± 11% -0.2 0.70 ± 15% perf-profile.children.cycles-pp.__orc_find
> 1.18 ± 5% -0.2 0.93 ± 7% perf-profile.children.cycles-pp.__submit_merged_write_cond
> 1.00 ± 15% -0.2 0.77 ± 14% perf-profile.children.cycles-pp.__filemap_fdatawait_range
> 2.31 ± 7% -0.2 2.08 ± 10% perf-profile.children.cycles-pp.walk_component
> 0.84 ± 12% -0.2 0.63 ± 13% perf-profile.children.cycles-pp.inode_permission
> 0.78 ± 11% -0.2 0.57 ± 20% perf-profile.children.cycles-pp.add_to_page_cache_lru
> 0.65 ± 12% -0.2 0.46 ± 17% perf-profile.children.cycles-pp.__add_to_page_cache_locked
> 0.64 ± 11% -0.2 0.45 ± 12% perf-profile.children.cycles-pp.orc_find
> 0.36 ± 24% -0.2 0.19 ± 25% perf-profile.children.cycles-pp.f2fs_invalidate_blocks
> 0.53 ± 24% -0.2 0.37 ± 16% perf-profile.children.cycles-pp.generic_permission
> 0.58 ± 21% -0.2 0.41 ± 13% perf-profile.children.cycles-pp.finish_task_switch
> 0.76 ± 25% -0.2 0.61 ± 12% perf-profile.children.cycles-pp.kmem_cache_free
> 0.57 ± 10% -0.1 0.43 ± 14% perf-profile.children.cycles-pp.sched_ttwu_pending
> 0.52 ± 14% -0.1 0.37 ± 18% perf-profile.children.cycles-pp.asm_sysvec_call_function_single
> 0.48 ± 24% -0.1 0.34 ± 14% perf-profile.children.cycles-pp.make_kuid
> 0.48 ± 14% -0.1 0.35 ± 17% perf-profile.children.cycles-pp.sysvec_call_function_single
> 0.60 ± 10% -0.1 0.48 ± 15% perf-profile.children.cycles-pp.__entry_text_start
> 0.32 ± 21% -0.1 0.20 ± 21% perf-profile.children.cycles-pp.filemap_get_pages
> 0.67 ± 14% -0.1 0.55 ± 18% perf-profile.children.cycles-pp.step_into
> 0.45 ± 24% -0.1 0.34 ± 13% perf-profile.children.cycles-pp.map_id_range_down
> 0.60 ± 10% -0.1 0.49 ± 10% perf-profile.children.cycles-pp.vfs_unlink
> 0.33 ± 13% -0.1 0.24 ± 16% perf-profile.children.cycles-pp.do_renameat2
> 0.24 ± 16% -0.1 0.17 ± 24% perf-profile.children.cycles-pp.aa_get_task_label
> 0.17 ± 26% -0.1 0.11 ± 10% perf-profile.children.cycles-pp.filename_parentat
> 0.20 ± 15% -0.1 0.14 ± 24% perf-profile.children.cycles-pp.__sysvec_call_function_single
> 0.09 ± 24% -0.1 0.04 ± 63% perf-profile.children.cycles-pp.xas_create
> 0.14 ± 10% -0.1 0.09 ± 27% perf-profile.children.cycles-pp.set_page_dirty
> 0.15 ± 18% -0.1 0.10 ± 12% perf-profile.children.cycles-pp.path_parentat
> 0.12 ± 35% -0.0 0.07 ± 25% perf-profile.children.cycles-pp.truncate_cleanup_page
> 0.17 ± 9% -0.0 0.12 ± 17% perf-profile.children.cycles-pp.__f2fs_find_entry
> 0.14 ± 11% -0.0 0.10 ± 25% perf-profile.children.cycles-pp.mem_cgroup_uncharge_list
> 0.11 ± 19% -0.0 0.07 ± 49% perf-profile.children.cycles-pp.set_task_cpu
> 0.26 ± 7% -0.0 0.22 ± 17% perf-profile.children.cycles-pp.entry_SYSCALL_64_safe_stack
> 0.19 ± 14% -0.0 0.14 ± 16% perf-profile.children.cycles-pp.lru_add_drain_cpu
> 0.15 ± 11% -0.0 0.11 ± 15% perf-profile.children.cycles-pp.f2fs_find_in_inline_dir
> 0.08 ± 18% -0.0 0.04 ± 65% perf-profile.children.cycles-pp.perf_trace_buf_update
> 0.12 ± 16% -0.0 0.09 ± 26% perf-profile.children.cycles-pp.__test_set_page_writeback
> 0.14 ± 13% -0.0 0.11 ± 7% perf-profile.children.cycles-pp.__mark_inode_dirty
> 0.15 ± 10% -0.0 0.12 ± 11% perf-profile.children.cycles-pp.__list_add_valid
> 0.15 ± 14% +0.0 0.20 ± 8% perf-profile.children.cycles-pp.nr_iowait_cpu
> 0.30 ± 9% +0.1 0.37 ± 11% perf-profile.children.cycles-pp.mmput
> 0.29 ± 8% +0.1 0.37 ± 10% perf-profile.children.cycles-pp.exit_mmap
> 0.16 ± 11% +0.1 0.23 ± 16% perf-profile.children.cycles-pp.rb_next
> 0.03 ±100% +0.1 0.11 ± 32% perf-profile.children.cycles-pp.sys_imageblit
> 0.03 ±100% +0.1 0.11 ± 32% perf-profile.children.cycles-pp.drm_fb_helper_sys_imageblit
> 0.18 ± 10% +0.1 0.27 ± 19% perf-profile.children.cycles-pp.trigger_load_balance
> 0.50 ± 5% +0.1 0.58 ± 11% perf-profile.children.cycles-pp.__hrtimer_next_event_base
> 0.05 ± 45% +0.1 0.14 ± 42% perf-profile.children.cycles-pp.vt_console_print
> 1.06 ± 7% +0.2 1.26 ± 12% perf-profile.children.cycles-pp.update_blocked_averages
> 1.43 ± 5% +0.3 1.69 ± 7% perf-profile.children.cycles-pp.rebalance_domains
> 1.79 ± 5% +0.4 2.20 ± 11% perf-profile.children.cycles-pp.clockevents_program_event
> 43.33 ± 7% +8.1 51.42 ± 9% perf-profile.children.cycles-pp.secondary_startup_64_no_verify
> 43.33 ± 7% +8.1 51.42 ± 9% perf-profile.children.cycles-pp.cpu_startup_entry
> 43.32 ± 7% +8.1 51.42 ± 9% perf-profile.children.cycles-pp.do_idle
> 42.46 ± 7% +8.2 50.64 ± 9% perf-profile.children.cycles-pp.start_secondary
> 0.49 ± 28% -0.3 0.23 ± 48% perf-profile.self.cycles-pp.osq_lock
> 0.94 ± 11% -0.3 0.69 ± 16% perf-profile.self.cycles-pp.__orc_find
> 1.07 ± 7% -0.2 0.85 ± 13% perf-profile.self.cycles-pp.copy_user_enhanced_fast_string
> 0.60 ± 10% -0.1 0.46 ± 16% perf-profile.self.cycles-pp.__entry_text_start
> 0.29 ± 12% -0.1 0.23 ± 15% perf-profile.self.cycles-pp.__schedule
> 0.17 ± 15% -0.1 0.12 ± 14% perf-profile.self.cycles-pp.update_curr
> 0.14 ± 16% -0.0 0.10 ± 20% perf-profile.self.cycles-pp.path_init
> 0.21 ± 6% -0.0 0.17 ± 14% perf-profile.self.cycles-pp.entry_SYSCALL_64_safe_stack
> 0.09 ± 18% -0.0 0.05 ± 42% perf-profile.self.cycles-pp.current_time
> 0.17 ± 8% -0.0 0.13 ± 17% perf-profile.self.cycles-pp.f2fs_fill_dentries
> 0.08 ± 11% -0.0 0.05 ± 44% perf-profile.self.cycles-pp.path_lookupat
> 0.10 ± 9% -0.0 0.08 ± 21% perf-profile.self.cycles-pp.__lookup_mnt
> 0.09 ± 20% +0.1 0.16 ± 24% perf-profile.self.cycles-pp.rcu_core
> 0.15 ± 14% +0.1 0.23 ± 19% perf-profile.self.cycles-pp.trigger_load_balance
>
>
>
> fxmark.hdd_f2fs_dbench_client_72_directio.works_sec
>
> 200 +---------------------------------------------------------------------+
> | +. |
> 190 |.+.+ +. .+ +. .+.+.+.+.+.+.+. .+.+.+. .+.+. : +.|
> 180 |-+ + + +.+ + + + + + +. .+. : |
> | + +.+.+.+ + +.+ |
> 170 |-+ |
> | |
> 160 |-+ O |
> | O |
> 150 |-O O O O O O O O O O O O |
> 140 |-+ O O O O O O O O |
> | O O O O |
> 130 |-+ O O O O |
> | O |
> 120 +---------------------------------------------------------------------+
>
>
> fxmark.hdd_f2fs_dbench_client_72_directio.sys_sec
>
> 145 +---------------------------------------------------------------------+
> 140 |-+ + + + .+ |
> |.+. : + + : : + .+. .+.+. .+.+. +. .+ .+ : |
> 135 |-+ +. : + : : + + +.+ + +.+.+. + +.+ + .+ : |
> 130 |-+ + + + + + +.+.+ +.|
> 125 |-+ + |
> 120 |-+ |
> | |
> 115 |-O O O |
> 110 |-+ O O O O O O O O O O O O O O |
> 105 |-+ O O O O O O O O O O |
> 100 |-+ O O O |
> | |
> 95 |-+ O |
> 90 +---------------------------------------------------------------------+
>
>
> fxmark.hdd_f2fs_dbench_client_72_directio.sys_util
>
> 3.4 +---------------------------------------------------------------------+
> | + |
> 3.2 |.+ +. + : + .+ .+. .+. .+ |
> | +.+. + + : + + + + .+ +.+ +.+. .+.+. .+. .+ + |
> 3 |-+ + + +. + + +.+.+ + +. .+.+ +.|
> | + + |
> 2.8 |-+ |
> | |
> 2.6 |-O O O O |
> | O O O O O O O O O O O O O O O O |
> 2.4 |-+ O O O O O O O |
> | O O O |
> 2.2 |-+ O |
> | |
> 2 +---------------------------------------------------------------------+
>
>
> fxmark.hdd_f2fs_dbench_client_54_directio.sys_sec
>
> 110 +---------------------------------------------------------------------+
> |.+ .+.+ +.+ .+ .+.+ |
> 105 |-++ .+ +. +. .+ + + + .+.+ + +.+ : + + :|
> 100 |-+ + + + +.+ + +.+ + + + + + + : + +. :|
> | + + .+.+ + + + + |
> 95 |-+ + |
> | |
> 90 |-+ |
> | O |
> 85 |-+ O O O O O O O O O O O |
> 80 |-O O O O O O O |
> | O O O O O O O O O |
> 75 |-+ O O O |
> | |
> 70 +---------------------------------------------------------------------+
>
>
> fxmark.hdd_f2fs_dbench_client_54_directio.sys_util
>
> 3.3 +---------------------------------------------------------------------+
> 3.2 |-++ .+ .+.+ + .+ + .+ + +. .+ +.+.+. |
> | + + .+.+.+ .+.+.+ + +.+ + + + + + + + + :|
> 3.1 |-+ + : +.+ + + + + :|
> 3 |-+ : + + |
> 2.9 |-+ + |
> 2.8 |-+ |
> | |
> 2.7 |-+ O |
> 2.6 |-+ O O O O O O O O O |
> 2.5 |-O O O O O |
> 2.4 |-+ O O O O O O O O O O O O |
> | O O O |
> 2.3 |-+ O |
> 2.2 +---------------------------------------------------------------------+
>
>
> fxmark.time.user_time
>
> 96 +----------------------------------------------------------------------+
> 94 |-+ .+ |
> |.+. .+.+ + .+ +.+.+. .+.+. .+.+ .+.+ .+ |
> 92 |-+ + +.+.+. .+.+ + + +. + + .+.+.+ + .+.+.+ : |
> 90 |-+ + + + + : +|
> | + |
> 88 |-+ |
> 86 |-+ |
> 84 |-+ |
> | |
> 82 |-+ O |
> 80 |-O O O O O O O |
> | O O O O O O O O O |
> 78 |-+ O O O O O O O O O O O O O O |
> 76 +----------------------------------------------------------------------+
>
>
> fxmark.time.system_time
>
> 370 +---------------------------------------------------------------------+
> |.+.+. .+.+.+. .+. .+.+.+.+.+.+.+.+.+ +. .+. .+ |
> 360 |-+ + +.+. .+ + + + +.+ +. .+.+.+ : |
> 350 |-+ +.+ + + : +|
> | + |
> 340 |-+ |
> 330 |-+ |
> | |
> 320 |-+ |
> 310 |-O O |
> | O O O O O O O O O O |
> 300 |-+ O O O O O O O O O O O O O O O O O |
> 290 |-+ O |
> | O |
> 280 +---------------------------------------------------------------------+
>
>
> fxmark.time.voluntary_context_switches
>
> 1.05e+06 +----------------------------------------------------------------+
> 1e+06 |-+ +. |
> |.+.+.++.+.+.+.+. .+. + +.+.++.+.+.+.+ ++.+.+.+. .+ |
> 950000 |-+ +.++ + + + +. .++ : |
> 900000 |-+ + + : +|
> | + |
> 850000 |-+ |
> 800000 |-+ |
> 750000 |-+ |
> | O |
> 700000 |-O O O |
> 650000 |-+ O O O O O O O O OO O O O O O O O |
> | O O O O O O O O O |
> 600000 |-+ O |
> 550000 +----------------------------------------------------------------+
>
>
> fxmark.time.file_system_outputs
>
> 2.75e+07 +----------------------------------------------------------------+
> | + |
> 2.7e+07 |.+. .++.+.+. + + .+.+.+. +.+. .+.+ + .+.+ .+ |
> 2.65e+07 |-+ + +. .+. + + + + + + :.+ + .+.++ : :|
> | + + + + + : :|
> 2.6e+07 |-+ + |
> | |
> 2.55e+07 |-+ |
> | |
> 2.5e+07 |-+ |
> 2.45e+07 |-+ O O |
> | O O O O O O O |
> 2.4e+07 |-+ O O O O O OO O O O |
> | O O O O O O O O O O O |
> 2.35e+07 +----------------------------------------------------------------+
>
>
> [*] bisect-good sample
> [O] bisect-bad sample
>
>
>
> Disclaimer:
> Results have been estimated based on internal Intel analysis and are provided
> for informational purposes only. Any difference in system hardware or software
> design or configuration may affect actual performance.
>
>
> ---
> 0DAY/LKP+ Test Infrastructure Open Source Technology Center
> https://lists.01.org/hyperkitty/list/lkp@lists.01.org Intel Corporation
>
> Thanks,
> Oliver Sang
>
Powered by blists - more mailing lists