lists.openwall.net | lists / announce owl-users owl-dev john-users john-dev passwdqc-users yescrypt popa3d-users / oss-security kernel-hardening musl sabotage tlsify passwords / crypt-dev xvendor / Bugtraq Full-Disclosure linux-kernel linux-netdev linux-ext4 linux-hardening linux-cve-announce PHC | |
Open Source and information security mailing list archives
| ||
|
Message-ID: <202506032249.cd0b4bc-lkp@intel.com> Date: Tue, 3 Jun 2025 22:49:47 +0800 From: kernel test robot <oliver.sang@...el.com> To: Zhang Yi <yi.zhang@...wei.com> CC: <oe-lkp@...ts.linux.dev>, <lkp@...el.com>, <linux-kernel@...r.kernel.org>, Theodore Ts'o <tytso@....edu>, <linux-ext4@...r.kernel.org>, <oliver.sang@...el.com> Subject: [linus:master] [ext4] 7ac67301e8: aim7.jobs-per-min 22.9% regression Hello, kernel test robot noticed a 22.9% regression of aim7.jobs-per-min on: commit: 7ac67301e82f02b77a5c8e7377a1f414ef108b84 ("ext4: enable large folio for regular file") https://git.kernel.org/cgit/linux/kernel/git/torvalds/linux.git master [test failed on linux-next/master 3a83b350b5be4b4f6bd895eecf9a92080200ee5d] testcase: aim7 config: x86_64-rhel-9.4 compiler: gcc-12 test machine: 96 threads 2 sockets Intel(R) Xeon(R) Platinum 8260L CPU @ 2.40GHz (Cascade Lake) with 128G memory parameters: disk: 4BRD_12G md: RAID0 fs: ext4 test: creat-clo load: 1000 cpufreq_governor: performance In addition to that, the commit also has significant impact on the following tests: +------------------+----------------------------------------------------------------------------------------------------+ | testcase: change | filebench: filebench.sum_operations/s 13.9% improvement | | test machine | 192 threads 4 sockets Intel(R) Xeon(R) Platinum 9242 CPU @ 2.30GHz (Cascade Lake) with 176G memory | | test parameters | cpufreq_governor=performance | | | disk=1SSD | | | fs=ext4 | | | test=randomread.f | +------------------+----------------------------------------------------------------------------------------------------+ If you fix the issue in a separate patch/commit (i.e. not just a new version of the same patch/commit), kindly add following tags | Reported-by: kernel test robot <oliver.sang@...el.com> | Closes: https://lore.kernel.org/oe-lkp/202506032249.cd0b4bc-lkp@intel.com Details are as below: --------------------------------------------------------------------------------------------------> The kernel config and materials to reproduce are available at: https://download.01.org/0day-ci/archive/20250603/202506032249.cd0b4bc-lkp@intel.com ========================================================================================= compiler/cpufreq_governor/disk/fs/kconfig/load/md/rootfs/tbox_group/test/testcase: gcc-12/performance/4BRD_12G/ext4/x86_64-rhel-9.4/1000/RAID0/debian-12-x86_64-20240206.cgz/lkp-csl-2sp3/creat-clo/aim7 commit: 01e807e18f ("ext4: make online defragmentation support large folios") 7ac67301e8 ("ext4: enable large folio for regular file") 01e807e18fd87937 7ac67301e82f02b77a5c8e7377a ---------------- --------------------------- %stddev %change %stddev \ | \ 1.039e+10 ± 33% +141.0% 2.504e+10 cpuidle..time 20464297 ± 49% +227.8% 67090730 ± 2% cpuidle..usage 21.42 ± 22% +88.3% 40.33 iostat.cpu.idle 78.35 ± 6% -24.1% 59.45 iostat.cpu.system 552.32 ± 6% +26.6% 699.14 uptime.boot 14635 ± 23% +101.0% 29415 ± 2% uptime.idle 68206 ± 39% +145.2% 167235 meminfo.Buffers 71959 ± 37% +137.4% 170796 meminfo.Inactive 71959 ± 37% +137.4% 170796 meminfo.Inactive(file) 894.92 ± 11% +40.9% 1260 meminfo.Mlocked 1125010 ± 22% +155.5% 2874061 numa-numastat.node0.local_node 1183030 ± 22% +146.6% 2917772 numa-numastat.node0.numa_hit 1270609 ± 29% +133.6% 2968203 ± 2% numa-numastat.node1.local_node 1312306 ± 27% +130.5% 3024283 ± 2% numa-numastat.node1.numa_hit 1182349 ± 22% +146.8% 2917852 numa-vmstat.node0.numa_hit 1124328 ± 22% +155.6% 2874142 numa-vmstat.node0.numa_local 1311405 ± 27% +130.6% 3024180 ± 2% numa-vmstat.node1.numa_hit 1269707 ± 29% +133.8% 2968101 ± 2% numa-vmstat.node1.numa_local 21.46 ± 22% +87.9% 40.32 vmstat.cpu.id 9467 ± 39% +214.4% 29766 vmstat.io.bo 94.65 -27.6% 68.51 ± 6% vmstat.procs.r 77160 ± 39% +175.9% 212860 vmstat.system.cs 110041 ± 3% +18.3% 130180 vmstat.system.in 21.11 ± 23% +19.0 40.15 mpstat.cpu.all.idle% 0.00 ±116% +0.0 0.05 ± 22% mpstat.cpu.all.iowait% 0.51 ± 5% +0.1 0.63 mpstat.cpu.all.irq% 0.04 ± 27% +0.1 0.09 mpstat.cpu.all.soft% 78.11 ± 6% -19.2 58.90 mpstat.cpu.all.sys% 0.22 ± 7% -0.1 0.17 ± 2% mpstat.cpu.all.usr% 28.67 ± 15% +2158.7% 647.50 mpstat.max_utilization.seconds 12003 ± 6% -22.9% 9255 aim7.jobs-per-min 502.51 ± 7% +29.0% 648.48 aim7.time.elapsed_time 502.51 ± 7% +29.0% 648.48 aim7.time.elapsed_time.max 3006810 ± 34% +197.8% 8952958 aim7.time.involuntary_context_switches 2361 +1.5% 2396 aim7.time.maximum_resident_set_size 120010 ± 2% -7.7% 110766 ± 2% aim7.time.minor_page_faults 7505 ± 6% -25.3% 5607 aim7.time.percent_of_cpu_this_job_got 37471 -3.1% 36310 aim7.time.system_time 23471247 ± 53% +256.2% 83610596 aim7.time.voluntary_context_switches 1052963 +1.8% 1071691 proc-vmstat.nr_file_pages 17995 ± 37% +137.3% 42699 proc-vmstat.nr_inactive_file 18010 -2.0% 17656 proc-vmstat.nr_mapped 223.82 ± 11% +40.8% 315.11 proc-vmstat.nr_mlock 134427 -4.5% 128396 proc-vmstat.nr_shmem 33925 +1.9% 34558 proc-vmstat.nr_slab_reclaimable 17995 ± 37% +137.3% 42699 proc-vmstat.nr_zone_inactive_file 2495758 ± 24% +138.1% 5942873 proc-vmstat.numa_hit 2396041 ± 25% +143.9% 5843083 proc-vmstat.numa_local 2727788 ± 22% +125.1% 6139263 proc-vmstat.pgalloc_normal 1536291 ± 6% +22.1% 1875129 proc-vmstat.pgfault 2558575 ± 24% +132.6% 5950566 proc-vmstat.pgfree 4923796 ± 50% +293.3% 19362948 proc-vmstat.pgpgout 84986 ± 4% +12.4% 95485 proc-vmstat.pgreuse 0.61 ± 3% +23.7% 0.75 ± 2% perf-stat.i.MPKI 1.175e+10 ± 7% -33.7% 7.789e+09 ± 2% perf-stat.i.branch-instructions 0.25 ± 8% +0.1 0.32 ± 4% perf-stat.i.branch-miss-rate% 15144423 ± 4% +25.7% 19035266 perf-stat.i.branch-misses 38.23 ± 5% -8.6 29.64 perf-stat.i.cache-miss-rate% 34494766 ± 6% -17.6% 28420268 perf-stat.i.cache-misses 89945462 +6.7% 95927456 perf-stat.i.cache-references 77676 ± 39% +174.9% 213543 perf-stat.i.context-switches 4.00 ± 2% +12.1% 4.48 ± 2% perf-stat.i.cpi 2.328e+11 ± 6% -26.3% 1.715e+11 perf-stat.i.cpu-cycles 6559 ± 27% +135.2% 15428 ± 3% perf-stat.i.cpu-migrations 6657 -10.0% 5989 perf-stat.i.cycles-between-cache-misses 5.832e+10 ± 7% -34.8% 3.8e+10 ± 2% perf-stat.i.instructions 0.26 -11.5% 0.23 perf-stat.i.ipc 0.61 ± 58% +263.1% 2.22 perf-stat.i.metric.K/sec 2826 -4.4% 2701 perf-stat.i.minor-faults 2826 -4.4% 2701 perf-stat.i.page-faults 0.59 ± 2% +26.3% 0.75 ± 2% perf-stat.overall.MPKI 0.13 ± 13% +0.1 0.24 ± 3% perf-stat.overall.branch-miss-rate% 38.34 ± 5% -8.7 29.62 perf-stat.overall.cache-miss-rate% 4.00 ± 2% +13.0% 4.52 ± 2% perf-stat.overall.cpi 6747 -10.6% 6034 perf-stat.overall.cycles-between-cache-misses 0.25 ± 2% -11.5% 0.22 ± 2% perf-stat.overall.ipc 1.173e+10 ± 7% -33.7% 7.776e+09 ± 2% perf-stat.ps.branch-instructions 15120179 ± 4% +25.8% 19015313 perf-stat.ps.branch-misses 34435926 ± 6% -17.6% 28376133 perf-stat.ps.cache-misses 89790928 +6.7% 95792014 perf-stat.ps.cache-references 77356 ± 39% +175.6% 213185 perf-stat.ps.context-switches 2.324e+11 ± 6% -26.3% 1.712e+11 perf-stat.ps.cpu-cycles 6536 ± 27% +135.7% 15403 ± 3% perf-stat.ps.cpu-migrations 5.823e+10 ± 7% -34.8% 3.794e+10 ± 2% perf-stat.ps.instructions 2820 -4.4% 2696 perf-stat.ps.minor-faults 2820 -4.4% 2696 perf-stat.ps.page-faults 2.915e+13 ± 2% -15.5% 2.464e+13 ± 2% perf-stat.total.instructions 13893356 ± 10% -28.1% 9985283 ± 4% sched_debug.cfs_rq:/.avg_vruntime.avg 14866508 ± 9% -25.0% 11157233 ± 2% sched_debug.cfs_rq:/.avg_vruntime.max 13368653 ± 10% -29.6% 9413191 ± 4% sched_debug.cfs_rq:/.avg_vruntime.min 217979 ± 4% +132.7% 507327 ± 18% sched_debug.cfs_rq:/.avg_vruntime.stddev 2.40 ± 18% +40.2% 3.36 ± 8% sched_debug.cfs_rq:/.load_avg.min 13893356 ± 10% -28.1% 9985283 ± 4% sched_debug.cfs_rq:/.min_vruntime.avg 14866508 ± 9% -25.0% 11157233 ± 2% sched_debug.cfs_rq:/.min_vruntime.max 13368653 ± 10% -29.6% 9413191 ± 4% sched_debug.cfs_rq:/.min_vruntime.min 217979 ± 4% +132.7% 507327 ± 18% sched_debug.cfs_rq:/.min_vruntime.stddev 0.68 ± 9% -27.0% 0.50 ± 2% sched_debug.cfs_rq:/.nr_queued.avg 0.38 ± 11% +26.8% 0.48 sched_debug.cfs_rq:/.nr_queued.stddev 136.33 ± 39% -33.6% 90.51 ± 4% sched_debug.cfs_rq:/.removed.load_avg.max 30.24 ± 18% -36.3% 19.26 ± 20% sched_debug.cfs_rq:/.removed.load_avg.stddev 1397 ± 26% +53.5% 2145 ± 4% sched_debug.cfs_rq:/.runnable_avg.avg 2938 ± 18% +44.0% 4230 ± 16% sched_debug.cfs_rq:/.runnable_avg.max 651.69 ± 9% -26.3% 480.36 sched_debug.cfs_rq:/.util_avg.avg 1626 ± 16% -33.3% 1084 ± 16% sched_debug.cfs_rq:/.util_avg.max 274.82 ± 14% -38.0% 170.26 ± 8% sched_debug.cfs_rq:/.util_avg.stddev 825.51 ± 11% -46.6% 440.94 ± 4% sched_debug.cfs_rq:/.util_est.avg 3315 ± 19% -48.0% 1725 ± 4% sched_debug.cfs_rq:/.util_est.max 813.04 ± 22% -37.6% 507.58 ± 4% sched_debug.cfs_rq:/.util_est.stddev 695475 ± 14% -27.5% 504320 ± 3% sched_debug.cpu.avg_idle.avg 947694 ± 6% -18.0% 776966 ± 4% sched_debug.cpu.avg_idle.max 123841 ± 12% -24.0% 94117 ± 3% sched_debug.cpu.avg_idle.stddev 288821 ± 5% +24.5% 359697 ± 4% sched_debug.cpu.clock.avg 288831 ± 5% +24.5% 359707 ± 4% sched_debug.cpu.clock.max 288810 ± 5% +24.5% 359686 ± 4% sched_debug.cpu.clock.min 287274 ± 5% +24.4% 357262 ± 4% sched_debug.cpu.clock_task.avg 287497 ± 5% +24.4% 357608 ± 4% sched_debug.cpu.clock_task.max 279648 ± 5% +24.9% 349351 ± 4% sched_debug.cpu.clock_task.min 3160 ± 9% -25.4% 2357 sched_debug.cpu.curr->pid.avg 1873 ± 11% +29.4% 2424 ± 3% sched_debug.cpu.curr->pid.stddev 225517 ± 52% +202.2% 681409 ± 4% sched_debug.cpu.nr_switches.avg 238933 ± 50% +199.7% 716085 ± 4% sched_debug.cpu.nr_switches.max 215091 ± 53% +200.4% 646099 ± 5% sched_debug.cpu.nr_switches.min 6339 ± 35% +322.0% 26749 ± 15% sched_debug.cpu.nr_switches.stddev 470.00 ± 17% +89.1% 888.96 ± 10% sched_debug.cpu.nr_uninterruptible.max -436.21 +166.4% -1162 sched_debug.cpu.nr_uninterruptible.min 175.08 ± 20% +126.7% 396.98 ± 6% sched_debug.cpu.nr_uninterruptible.stddev 288810 ± 5% +24.5% 359686 ± 4% sched_debug.cpu_clk 288250 ± 5% +24.6% 359126 ± 4% sched_debug.ktime 289382 ± 5% +24.5% 360266 ± 4% sched_debug.sched_clk 0.43 ± 27% +62.1% 0.70 ± 14% perf-sched.sch_delay.avg.ms.__cond_resched.__ext4_handle_dirty_metadata.ext4_orphan_add.ext4_truncate.ext4_setattr 0.14 ±116% -78.4% 0.03 ±136% perf-sched.sch_delay.avg.ms.__cond_resched.__wait_for_common.wait_for_completion_state.kernel_clone.__x64_sys_vfork 0.04 ± 40% +541.3% 0.28 ±160% perf-sched.sch_delay.avg.ms.__cond_resched.bdev_getblk.jbd2_journal_get_descriptor_buffer.jbd2_journal_commit_transaction.kjournald2 0.43 ± 51% +126.8% 0.97 ± 31% perf-sched.sch_delay.avg.ms.__cond_resched.dput.__fput.__x64_sys_close.do_syscall_64 0.76 ± 76% +93.3% 1.47 ± 19% perf-sched.sch_delay.avg.ms.__cond_resched.ext4_journal_check_start.__ext4_journal_start_sb.ext4_truncate.ext4_setattr 0.89 ±146% -100.0% 0.00 perf-sched.sch_delay.avg.ms.__cond_resched.stop_one_cpu.migrate_task_to.task_numa_migrate.isra 0.11 ± 60% -78.1% 0.02 ±117% perf-sched.sch_delay.avg.ms.__cond_resched.stop_one_cpu.sched_exec.bprm_execve.part 0.01 ± 59% -75.4% 0.00 ± 31% perf-sched.sch_delay.avg.ms.anon_pipe_read.vfs_read.ksys_read.do_syscall_64 0.14 ± 86% -91.3% 0.01 ± 70% perf-sched.sch_delay.avg.ms.do_wait.kernel_wait4.do_syscall_64.entry_SYSCALL_64_after_hwframe 0.82 ± 8% -26.4% 0.60 ± 9% perf-sched.sch_delay.avg.ms.kjournald2.kthread.ret_from_fork.ret_from_fork_asm 0.07 ± 46% +649.3% 0.50 ±136% perf-sched.sch_delay.avg.ms.schedule_hrtimeout_range_clock.ep_poll.do_epoll_wait.__x64_sys_epoll_wait 0.37 ± 37% -74.9% 0.09 ± 55% perf-sched.sch_delay.avg.ms.schedule_timeout.__wait_for_common.wait_for_completion_state.kernel_clone 0.05 ± 62% +196.8% 0.15 ± 71% perf-sched.sch_delay.avg.ms.schedule_timeout.rcu_gp_fqs_loop.rcu_gp_kthread.kthread 0.04 ±104% +177.2% 0.10 ± 11% perf-sched.sch_delay.max.ms.__cond_resched.__find_get_block_slow.find_get_block_common.bdev_getblk.jbd2_journal_get_descriptor_buffer 0.31 ±127% -90.0% 0.03 ±136% perf-sched.sch_delay.max.ms.__cond_resched.__wait_for_common.wait_for_completion_state.kernel_clone.__x64_sys_vfork 0.09 ± 44% +23615.3% 20.95 ±207% perf-sched.sch_delay.max.ms.__cond_resched.bdev_getblk.jbd2_journal_get_descriptor_buffer.jbd2_journal_commit_transaction.kjournald2 0.01 ± 42% -81.0% 0.00 ±142% perf-sched.sch_delay.max.ms.__cond_resched.down_write_killable.exec_mmap.begin_new_exec.load_elf_binary 2.11 ± 76% +132.6% 4.91 ± 19% perf-sched.sch_delay.max.ms.__cond_resched.ext4_journal_check_start.__ext4_journal_start_sb.ext4_truncate.ext4_setattr 2.29 ± 69% -100.0% 0.00 perf-sched.sch_delay.max.ms.__cond_resched.stop_one_cpu.migrate_task_to.task_numa_migrate.isra 0.68 ± 79% -79.6% 0.14 ±149% perf-sched.sch_delay.max.ms.__cond_resched.stop_one_cpu.sched_exec.bprm_execve.part 9.93 ±114% -88.5% 1.14 ± 97% perf-sched.sch_delay.max.ms.anon_pipe_read.vfs_read.ksys_read.do_syscall_64 0.95 ± 93% +1396.1% 14.16 ±192% perf-sched.sch_delay.max.ms.schedule_hrtimeout_range_clock.ep_poll.do_epoll_wait.__x64_sys_epoll_wait 1232 ±131% -97.0% 36.96 ±124% perf-sched.total_sch_delay.max.ms 63.11 ± 65% -66.8% 20.95 ± 2% perf-sched.total_wait_and_delay.average.ms 100984 ± 48% +85.7% 187485 perf-sched.total_wait_and_delay.count.ms 61.91 ± 66% -68.1% 19.75 ± 2% perf-sched.total_wait_time.average.ms 88.55 ±133% -72.8% 24.08 ± 4% perf-sched.wait_and_delay.avg.ms.__cond_resched.__ext4_mark_inode_dirty.ext4_dirty_inode.__mark_inode_dirty.ext4_setattr 59.33 ± 81% -62.6% 22.21 ± 2% perf-sched.wait_and_delay.avg.ms.schedule_preempt_disabled.__mutex_lock.constprop.0.ext4_orphan_add 71.70 ± 76% -76.6% 16.81 ± 3% perf-sched.wait_and_delay.avg.ms.schedule_preempt_disabled.__mutex_lock.constprop.0.ext4_orphan_del 74.50 ± 98% +1048.1% 855.33 ± 9% perf-sched.wait_and_delay.count.__cond_resched.__ext4_journal_get_write_access.ext4_orphan_add.ext4_truncate.ext4_setattr 0.67 ±141% +925.0% 6.83 ± 24% perf-sched.wait_and_delay.count.__cond_resched.__find_get_block_slow.find_get_block_common.bdev_getblk.jbd2_journal_get_descriptor_buffer 31.50 ±141% +210.1% 97.67 perf-sched.wait_and_delay.count.__cond_resched.bdev_getblk.jbd2_journal_get_descriptor_buffer.jbd2_journal_commit_transaction.kjournald2 1.00 ±115% +266.7% 3.67 ± 40% perf-sched.wait_and_delay.count.__cond_resched.down_write.ext4_truncate.ext4_setattr.notify_change 107.67 ±148% +1574.6% 1803 ± 27% perf-sched.wait_and_delay.count.io_schedule.bit_wait_io.__wait_on_bit_lock.out_of_line_wait_on_bit_lock 39510 ± 53% +78.9% 70701 perf-sched.wait_and_delay.count.schedule_preempt_disabled.__mutex_lock.constprop.0.ext4_orphan_add 1392 ± 18% -38.3% 859.50 ± 2% perf-sched.wait_and_delay.count.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm 947.20 ±110% -94.3% 53.91 ± 50% perf-sched.wait_and_delay.max.ms.__cond_resched.__ext4_mark_inode_dirty.ext4_dirty_inode.__mark_inode_dirty.ext4_setattr 2500 ± 63% -96.6% 86.00 ± 88% perf-sched.wait_and_delay.max.ms.schedule_preempt_disabled.__mutex_lock.constprop.0.ext4_orphan_add 2480 ± 64% -97.2% 70.57 ± 74% perf-sched.wait_and_delay.max.ms.schedule_preempt_disabled.__mutex_lock.constprop.0.ext4_orphan_del 71.88 ± 65% -72.8% 19.58 ± 5% perf-sched.wait_time.avg.ms.__cond_resched.__ext4_handle_dirty_metadata.ext4_orphan_add.ext4_truncate.ext4_setattr 113.18 ±116% -77.1% 25.93 ± 2% perf-sched.wait_time.avg.ms.__cond_resched.__ext4_journal_get_write_access.ext4_orphan_add.ext4_truncate.ext4_setattr 88.09 ±134% -72.9% 23.91 ± 4% perf-sched.wait_time.avg.ms.__cond_resched.__ext4_mark_inode_dirty.ext4_dirty_inode.__mark_inode_dirty.ext4_setattr 81.81 ± 78% -69.7% 24.80 ± 6% perf-sched.wait_time.avg.ms.__cond_resched.__find_get_block_slow.find_get_block_common.bdev_getblk.__ext4_get_inode_loc 48.73 ± 76% -88.2% 5.73 ±223% perf-sched.wait_time.avg.ms.__cond_resched.__wait_for_common.stop_two_cpus.migrate_swap.task_numa_migrate 65.97 ± 87% -58.7% 27.24 ± 4% perf-sched.wait_time.avg.ms.__cond_resched.dput.terminate_walk.path_openat.do_filp_open 104.27 ± 92% -76.1% 24.89 ± 6% perf-sched.wait_time.avg.ms.__cond_resched.ext4_journal_check_start.__ext4_journal_start_sb.ext4_dirty_inode.__mark_inode_dirty 58.53 ± 48% -55.1% 26.29 ± 10% perf-sched.wait_time.avg.ms.__cond_resched.kmem_cache_alloc_noprof.alloc_empty_file.path_openat.do_filp_open 64.58 ± 74% -59.6% 26.11 ± 8% perf-sched.wait_time.avg.ms.__cond_resched.kmem_cache_alloc_noprof.getname_flags.part.0 56.99 ± 47% -58.9% 23.45 ± 19% perf-sched.wait_time.avg.ms.__cond_resched.kmem_cache_alloc_noprof.jbd2__journal_start.ext4_truncate.ext4_setattr 112.58 ±105% -76.3% 26.65 ± 6% perf-sched.wait_time.avg.ms.__cond_resched.mnt_want_write.do_open.path_openat.do_filp_open 242.61 ±111% -100.0% 0.00 perf-sched.wait_time.avg.ms.__cond_resched.stop_one_cpu.migrate_task_to.task_numa_migrate.isra 59.19 ± 91% -57.5% 25.15 ± 7% perf-sched.wait_time.avg.ms.__cond_resched.task_work_run.syscall_exit_to_user_mode.do_syscall_64.entry_SYSCALL_64_after_hwframe 2.67 ± 52% +470.8% 15.24 ± 4% perf-sched.wait_time.avg.ms.kjournald2.kthread.ret_from_fork.ret_from_fork_asm 57.13 ± 85% -64.9% 20.04 ± 2% perf-sched.wait_time.avg.ms.schedule_preempt_disabled.__mutex_lock.constprop.0.ext4_orphan_add 70.26 ± 78% -77.8% 15.58 ± 3% perf-sched.wait_time.avg.ms.schedule_preempt_disabled.__mutex_lock.constprop.0.ext4_orphan_del 113.86 ±110% -92.6% 8.40 ± 13% perf-sched.wait_time.avg.ms.schedule_preempt_disabled.rwsem_down_write_slowpath.down_write.open_last_lookups 1.22 ± 25% -37.3% 0.76 ± 5% perf-sched.wait_time.avg.ms.schedule_timeout.__wait_for_common.wait_for_completion_state.kernel_clone 6.62 ± 28% +48.7% 9.85 ± 6% perf-sched.wait_time.avg.ms.schedule_timeout.rcu_gp_fqs_loop.rcu_gp_kthread.kthread 1507 ± 87% -97.0% 45.93 ± 29% perf-sched.wait_time.max.ms.__cond_resched.__ext4_handle_dirty_metadata.ext4_do_update_inode.isra.0 1646 ± 93% -97.5% 40.37 ± 17% perf-sched.wait_time.max.ms.__cond_resched.__ext4_handle_dirty_metadata.ext4_orphan_add.ext4_truncate.ext4_setattr 1142 ±116% -96.0% 46.24 ± 28% perf-sched.wait_time.max.ms.__cond_resched.__ext4_journal_get_write_access.ext4_orphan_add.ext4_truncate.ext4_setattr 913.53 ±116% -95.1% 44.79 ± 27% perf-sched.wait_time.max.ms.__cond_resched.__ext4_mark_inode_dirty.ext4_dirty_inode.__mark_inode_dirty.ext4_setattr 107.27 ± 85% -72.9% 29.08 ± 8% perf-sched.wait_time.max.ms.__cond_resched.__ext4_mark_inode_dirty.ext4_ext_truncate.ext4_truncate.ext4_setattr 582.25 ±159% -94.2% 34.03 ± 5% perf-sched.wait_time.max.ms.__cond_resched.__find_get_block_slow.find_get_block_common.bdev_getblk.__ext4_get_inode_loc 111.60 ± 91% -70.0% 33.52 ± 12% perf-sched.wait_time.max.ms.__cond_resched.__fput.__x64_sys_close.do_syscall_64.entry_SYSCALL_64_after_hwframe 718.87 ±195% -96.2% 27.67 ± 46% perf-sched.wait_time.max.ms.__cond_resched.__kmalloc_noprof.ext4_ext_remove_space.ext4_ext_truncate.ext4_truncate 155.77 ± 75% -78.3% 33.76 ± 7% perf-sched.wait_time.max.ms.__cond_resched.__mutex_lock.constprop.0.ext4_orphan_add 787.80 ±193% -95.9% 31.91 ± 6% perf-sched.wait_time.max.ms.__cond_resched.__mutex_lock.constprop.0.ext4_orphan_del 71.24 ±112% -92.0% 5.73 ±223% perf-sched.wait_time.max.ms.__cond_resched.__wait_for_common.stop_two_cpus.migrate_swap.task_numa_migrate 1596 ± 97% -97.5% 40.35 ± 17% perf-sched.wait_time.max.ms.__cond_resched.bdev_getblk.__ext4_get_inode_loc.ext4_get_inode_loc.ext4_reserve_inode_write 567.42 ±149% -93.9% 34.41 ± 2% perf-sched.wait_time.max.ms.__cond_resched.down_read.ext4_xattr_get.__vfs_getxattr.cap_inode_need_killpriv 111.57 ± 66% -69.7% 33.76 ± 3% perf-sched.wait_time.max.ms.__cond_resched.dput.__fput.__x64_sys_close.do_syscall_64 179.12 ± 55% -81.2% 33.67 ± 4% perf-sched.wait_time.max.ms.__cond_resched.dput.terminate_walk.path_openat.do_filp_open 494.41 ±155% -93.1% 34.32 ± 5% perf-sched.wait_time.max.ms.__cond_resched.ext4_journal_check_start.__ext4_journal_start_sb.ext4_dirty_inode.__mark_inode_dirty 1024 ±118% -95.5% 45.74 ± 29% perf-sched.wait_time.max.ms.__cond_resched.ext4_orphan_add.ext4_truncate.ext4_setattr.notify_change 165.88 ± 60% -80.1% 33.03 ± 4% perf-sched.wait_time.max.ms.__cond_resched.kmem_cache_alloc_noprof.alloc_empty_file.path_openat.do_filp_open 113.61 ± 87% -73.2% 30.42 ± 8% perf-sched.wait_time.max.ms.__cond_resched.kmem_cache_alloc_noprof.getname_flags.part.0 127.53 ± 72% -74.2% 32.88 ± 11% perf-sched.wait_time.max.ms.__cond_resched.kmem_cache_alloc_noprof.jbd2__journal_start.ext4_dirty_inode.__mark_inode_dirty 102.22 ± 54% -70.4% 30.29 ± 12% perf-sched.wait_time.max.ms.__cond_resched.kmem_cache_alloc_noprof.jbd2__journal_start.ext4_truncate.ext4_setattr 895.85 ±122% -96.0% 35.84 ± 11% perf-sched.wait_time.max.ms.__cond_resched.mnt_want_write.do_open.path_openat.do_filp_open 1523 ±100% -100.0% 0.00 perf-sched.wait_time.max.ms.__cond_resched.stop_one_cpu.migrate_task_to.task_numa_migrate.isra 72.88 ± 95% -56.3% 31.88 ± 5% perf-sched.wait_time.max.ms.__cond_resched.task_work_run.syscall_exit_to_user_mode.do_syscall_64.entry_SYSCALL_64_after_hwframe 558.24 ±151% -92.9% 39.83 ± 32% perf-sched.wait_time.max.ms.irqentry_exit_to_user_mode.asm_sysvec_apic_timer_interrupt.[unknown] 2439 ± 67% -97.5% 60.23 ± 55% perf-sched.wait_time.max.ms.schedule_preempt_disabled.__mutex_lock.constprop.0.ext4_orphan_add 2431 ± 68% -97.7% 54.78 ± 46% perf-sched.wait_time.max.ms.schedule_preempt_disabled.__mutex_lock.constprop.0.ext4_orphan_del 1533 ± 91% -97.5% 37.66 ± 4% perf-sched.wait_time.max.ms.schedule_preempt_disabled.rwsem_down_write_slowpath.down_write.do_unlinkat 1127 ±133% -96.9% 35.52 ± 8% perf-sched.wait_time.max.ms.schedule_preempt_disabled.rwsem_down_write_slowpath.down_write.open_last_lookups 716.22 ± 56% -93.7% 45.07 ± 31% perf-sched.wait_time.max.ms.syscall_exit_to_user_mode.do_syscall_64.entry_SYSCALL_64_after_hwframe.[unknown] 43.31 ± 3% -10.6 32.68 ± 4% perf-profile.calltrace.cycles-pp.osq_lock.__mutex_lock.ext4_orphan_del.ext4_truncate.ext4_setattr 43.97 ± 3% -10.5 33.46 ± 4% perf-profile.calltrace.cycles-pp.osq_lock.__mutex_lock.ext4_orphan_add.ext4_truncate.ext4_setattr 44.73 ± 3% -10.4 34.28 ± 4% perf-profile.calltrace.cycles-pp.ext4_orphan_del.ext4_truncate.ext4_setattr.notify_change.do_truncate 43.90 ± 3% -10.3 33.57 ± 4% perf-profile.calltrace.cycles-pp.__mutex_lock.ext4_orphan_del.ext4_truncate.ext4_setattr.notify_change 92.63 ± 2% -8.5 84.12 perf-profile.calltrace.cycles-pp.ext4_truncate.ext4_setattr.notify_change.do_truncate.do_open 45.41 -6.9 38.48 ± 2% perf-profile.calltrace.cycles-pp.__mutex_lock.ext4_orphan_add.ext4_truncate.ext4_setattr.notify_change 46.25 -6.7 39.54 ± 2% perf-profile.calltrace.cycles-pp.ext4_orphan_add.ext4_truncate.ext4_setattr.notify_change.do_truncate 96.15 -1.4 94.76 perf-profile.calltrace.cycles-pp.ext4_setattr.notify_change.do_truncate.do_open.path_openat 96.18 -1.4 94.80 perf-profile.calltrace.cycles-pp.notify_change.do_truncate.do_open.path_openat.do_filp_open 96.29 -1.4 94.92 perf-profile.calltrace.cycles-pp.do_truncate.do_open.path_openat.do_filp_open.do_sys_openat2 96.44 -1.4 95.08 perf-profile.calltrace.cycles-pp.do_open.path_openat.do_filp_open.do_sys_openat2.__x64_sys_creat 96.58 -1.3 95.29 perf-profile.calltrace.cycles-pp.path_openat.do_filp_open.do_sys_openat2.__x64_sys_creat.do_syscall_64 96.59 -1.3 95.30 perf-profile.calltrace.cycles-pp.do_filp_open.do_sys_openat2.__x64_sys_creat.do_syscall_64.entry_SYSCALL_64_after_hwframe 96.67 -1.3 95.40 perf-profile.calltrace.cycles-pp.do_sys_openat2.__x64_sys_creat.do_syscall_64.entry_SYSCALL_64_after_hwframe.creat64 96.68 -1.3 95.40 perf-profile.calltrace.cycles-pp.__x64_sys_creat.do_syscall_64.entry_SYSCALL_64_after_hwframe.creat64 96.76 -1.2 95.55 perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.creat64 96.76 -1.2 95.55 perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.creat64 96.85 -1.2 95.67 perf-profile.calltrace.cycles-pp.creat64 2.45 ± 6% +0.4 2.81 ± 3% perf-profile.calltrace.cycles-pp.intel_idle.cpuidle_enter_state.cpuidle_enter.cpuidle_idle_call.do_idle 2.50 ± 5% +0.5 3.03 ± 3% perf-profile.calltrace.cycles-pp.cpuidle_enter_state.cpuidle_enter.cpuidle_idle_call.do_idle.cpu_startup_entry 2.50 ± 5% +0.5 3.04 ± 3% perf-profile.calltrace.cycles-pp.cpuidle_enter.cpuidle_idle_call.do_idle.cpu_startup_entry.start_secondary 2.51 ± 5% +0.6 3.12 ± 4% perf-profile.calltrace.cycles-pp.cpuidle_idle_call.do_idle.cpu_startup_entry.start_secondary.common_startup_64 2.56 ± 4% +0.8 3.36 ± 4% perf-profile.calltrace.cycles-pp.do_idle.cpu_startup_entry.start_secondary.common_startup_64 2.56 ± 4% +0.8 3.36 ± 4% perf-profile.calltrace.cycles-pp.cpu_startup_entry.start_secondary.common_startup_64 2.56 ± 4% +0.8 3.36 ± 4% perf-profile.calltrace.cycles-pp.start_secondary.common_startup_64 2.58 ± 4% +0.8 3.39 ± 4% perf-profile.calltrace.cycles-pp.common_startup_64 0.76 ± 80% +3.4 4.18 ± 9% perf-profile.calltrace.cycles-pp.native_queued_spin_lock_slowpath._raw_spin_lock_irqsave.__mutex_lock.ext4_orphan_add.ext4_truncate 0.76 ± 80% +3.5 4.22 ± 9% perf-profile.calltrace.cycles-pp._raw_spin_lock_irqsave.__mutex_lock.ext4_orphan_add.ext4_truncate.ext4_setattr 2.89 ± 62% +6.3 9.16 ± 6% perf-profile.calltrace.cycles-pp.native_queued_spin_lock_slowpath.queued_read_lock_slowpath.start_this_handle.jbd2__journal_start.ext4_dirty_inode 2.90 ± 62% +6.3 9.21 ± 6% perf-profile.calltrace.cycles-pp.queued_read_lock_slowpath.start_this_handle.jbd2__journal_start.ext4_dirty_inode.__mark_inode_dirty 3.13 ± 60% +6.9 10.03 ± 6% perf-profile.calltrace.cycles-pp.jbd2__journal_start.ext4_dirty_inode.__mark_inode_dirty.ext4_setattr.notify_change 3.12 ± 61% +6.9 10.02 ± 6% perf-profile.calltrace.cycles-pp.start_this_handle.jbd2__journal_start.ext4_dirty_inode.__mark_inode_dirty.ext4_setattr 3.38 ± 57% +7.0 10.42 ± 6% perf-profile.calltrace.cycles-pp.ext4_dirty_inode.__mark_inode_dirty.ext4_setattr.notify_change.do_truncate 3.42 ± 56% +7.1 10.53 ± 6% perf-profile.calltrace.cycles-pp.__mark_inode_dirty.ext4_setattr.notify_change.do_truncate.do_open 0.61 ± 58% +8.1 8.73 ± 19% perf-profile.calltrace.cycles-pp.queued_read_lock_slowpath.start_this_handle.jbd2__journal_start.ext4_truncate.ext4_setattr 0.32 ±141% +8.2 8.52 ± 19% perf-profile.calltrace.cycles-pp.native_queued_spin_lock_slowpath.queued_read_lock_slowpath.start_this_handle.jbd2__journal_start.ext4_truncate 0.78 ± 56% +8.7 9.47 ± 18% perf-profile.calltrace.cycles-pp.start_this_handle.jbd2__journal_start.ext4_truncate.ext4_setattr.notify_change 0.79 ± 56% +8.7 9.48 ± 18% perf-profile.calltrace.cycles-pp.jbd2__journal_start.ext4_truncate.ext4_setattr.notify_change.do_truncate 87.39 ± 3% -21.1 66.25 ± 4% perf-profile.children.cycles-pp.osq_lock 89.41 ± 2% -17.3 72.15 ± 3% perf-profile.children.cycles-pp.__mutex_lock 44.78 ± 3% -10.5 34.33 ± 4% perf-profile.children.cycles-pp.ext4_orphan_del 92.63 ± 2% -8.5 84.12 perf-profile.children.cycles-pp.ext4_truncate 46.30 -6.7 39.58 ± 2% perf-profile.children.cycles-pp.ext4_orphan_add 96.15 -1.4 94.76 perf-profile.children.cycles-pp.ext4_setattr 96.18 -1.4 94.80 perf-profile.children.cycles-pp.notify_change 96.29 -1.4 94.92 perf-profile.children.cycles-pp.do_truncate 96.44 -1.4 95.08 perf-profile.children.cycles-pp.do_open 96.59 -1.3 95.29 perf-profile.children.cycles-pp.path_openat 96.60 -1.3 95.31 perf-profile.children.cycles-pp.do_filp_open 96.68 -1.3 95.40 perf-profile.children.cycles-pp.__x64_sys_creat 96.67 -1.3 95.40 perf-profile.children.cycles-pp.do_sys_openat2 96.86 -1.2 95.68 perf-profile.children.cycles-pp.creat64 97.09 -1.1 95.94 perf-profile.children.cycles-pp.do_syscall_64 97.10 -1.1 95.95 perf-profile.children.cycles-pp.entry_SYSCALL_64_after_hwframe 0.62 ± 2% -0.1 0.51 ± 3% perf-profile.children.cycles-pp.mutex_spin_on_owner 0.10 ± 8% -0.0 0.07 perf-profile.children.cycles-pp.osq_unlock 0.09 +0.0 0.10 perf-profile.children.cycles-pp.security_inode_need_killpriv 0.09 +0.0 0.10 ± 4% perf-profile.children.cycles-pp.vfs_open 0.06 ± 8% +0.0 0.07 perf-profile.children.cycles-pp.evict 0.06 ± 8% +0.0 0.07 perf-profile.children.cycles-pp.ext4_evict_inode 0.10 ± 3% +0.0 0.12 ± 6% perf-profile.children.cycles-pp.__ext4_journal_start_sb 0.11 ± 3% +0.0 0.13 ± 2% perf-profile.children.cycles-pp.__x64_sys_close 0.10 ± 4% +0.0 0.12 ± 5% perf-profile.children.cycles-pp.mutex_lock 0.08 ± 4% +0.0 0.10 ± 6% perf-profile.children.cycles-pp.update_cfs_group 0.06 ± 9% +0.0 0.08 perf-profile.children.cycles-pp.__ext4_unlink 0.06 ± 9% +0.0 0.08 perf-profile.children.cycles-pp.ext4_unlink 0.06 ± 9% +0.0 0.08 perf-profile.children.cycles-pp.vfs_unlink 0.08 ± 5% +0.0 0.11 ± 3% perf-profile.children.cycles-pp.__ext4_handle_dirty_metadata 0.22 ± 2% +0.0 0.25 ± 2% perf-profile.children.cycles-pp.__close 0.12 ± 6% +0.0 0.15 ± 3% perf-profile.children.cycles-pp.entry_SYSRETQ_unsafe_stack 0.58 +0.0 0.61 perf-profile.children.cycles-pp.ext4_do_update_inode 0.17 ± 4% +0.0 0.20 ± 6% perf-profile.children.cycles-pp.task_tick_fair 0.13 ± 6% +0.0 0.16 ± 4% perf-profile.children.cycles-pp.__x64_sys_unlink 0.13 ± 6% +0.0 0.16 ± 4% perf-profile.children.cycles-pp.do_unlinkat 0.13 ± 6% +0.0 0.16 ± 4% perf-profile.children.cycles-pp.unlink 0.15 ± 4% +0.0 0.19 ± 5% perf-profile.children.cycles-pp.find_get_block_common 0.16 ± 4% +0.0 0.20 ± 5% perf-profile.children.cycles-pp.bdev_getblk 0.22 ± 4% +0.0 0.27 ± 4% perf-profile.children.cycles-pp.sched_tick 0.00 +0.1 0.05 perf-profile.children.cycles-pp.__cond_resched 0.00 +0.1 0.05 perf-profile.children.cycles-pp.jbd2_write_access_granted 0.00 +0.1 0.05 perf-profile.children.cycles-pp.task_mm_cid_work 0.01 ±223% +0.1 0.07 ± 11% perf-profile.children.cycles-pp.task_work_run 0.01 ±223% +0.1 0.07 ± 8% perf-profile.children.cycles-pp.stop_this_handle 0.00 +0.1 0.06 ± 11% perf-profile.children.cycles-pp.menu_select 0.00 +0.1 0.06 ± 11% perf-profile.children.cycles-pp.select_idle_cpu 0.08 ± 18% +0.1 0.15 ± 9% perf-profile.children.cycles-pp.syscall_exit_to_user_mode 0.00 +0.1 0.06 ± 7% perf-profile.children.cycles-pp.schedule_idle 0.34 ± 8% +0.1 0.41 ± 3% perf-profile.children.cycles-pp.__mutex_unlock_slowpath 0.01 ±223% +0.1 0.08 ± 6% perf-profile.children.cycles-pp.open_last_lookups 0.08 ± 11% +0.1 0.15 ± 6% perf-profile.children.cycles-pp.jbd2_journal_stop 0.10 ± 10% +0.1 0.16 ± 4% perf-profile.children.cycles-pp.__ext4_journal_stop 0.32 ± 3% +0.1 0.38 ± 3% perf-profile.children.cycles-pp.update_process_times 0.00 +0.1 0.07 ± 10% perf-profile.children.cycles-pp.select_idle_sibling 0.34 ± 4% +0.1 0.42 ± 5% perf-profile.children.cycles-pp.tick_nohz_handler 0.01 ±223% +0.1 0.09 ± 6% perf-profile.children.cycles-pp.update_curr 0.00 +0.1 0.08 ± 8% perf-profile.children.cycles-pp.dequeue_entity 0.06 ± 14% +0.1 0.15 ± 5% perf-profile.children.cycles-pp._raw_spin_lock 0.00 +0.1 0.09 ± 12% perf-profile.children.cycles-pp.select_task_rq_fair 0.00 +0.1 0.09 ± 9% perf-profile.children.cycles-pp.raw_spin_rq_lock_nested 0.00 +0.1 0.10 ± 10% perf-profile.children.cycles-pp.select_task_rq 0.00 +0.1 0.10 ± 4% perf-profile.children.cycles-pp.handle_softirqs 0.62 ± 2% +0.1 0.72 ± 2% perf-profile.children.cycles-pp.ext4_reserve_inode_write 0.00 +0.1 0.10 ± 10% perf-profile.children.cycles-pp.__sysvec_call_function_single 0.01 ±223% +0.1 0.11 ± 4% perf-profile.children.cycles-pp.__irq_exit_rcu 0.00 +0.1 0.10 ± 23% perf-profile.children.cycles-pp.prepare_to_wait_exclusive 0.00 +0.1 0.10 ± 11% perf-profile.children.cycles-pp.sysvec_call_function_single 0.38 ± 4% +0.1 0.49 ± 2% perf-profile.children.cycles-pp.__hrtimer_run_queues 0.00 +0.1 0.11 ± 11% perf-profile.children.cycles-pp.enqueue_entity 0.08 ± 22% +0.1 0.18 ± 8% perf-profile.children.cycles-pp.update_load_avg 0.00 +0.1 0.11 ± 10% perf-profile.children.cycles-pp.asm_sysvec_call_function_single 0.73 ± 2% +0.1 0.85 ± 3% perf-profile.children.cycles-pp.__ext4_mark_inode_dirty 0.12 ± 14% +0.1 0.25 ± 7% perf-profile.children.cycles-pp.__ext4_journal_get_write_access 0.44 ± 4% +0.1 0.57 perf-profile.children.cycles-pp.__sysvec_apic_timer_interrupt 0.44 ± 4% +0.1 0.57 perf-profile.children.cycles-pp.hrtimer_interrupt 0.11 ± 12% +0.1 0.24 ± 8% perf-profile.children.cycles-pp.jbd2_journal_get_write_access 0.01 ±223% +0.1 0.14 ± 10% perf-profile.children.cycles-pp.try_to_block_task 0.01 ±223% +0.1 0.15 ± 14% perf-profile.children.cycles-pp.queued_write_lock_slowpath 0.01 ±223% +0.1 0.15 ± 10% perf-profile.children.cycles-pp.flush_smp_call_function_queue 0.09 ± 22% +0.1 0.24 ± 7% perf-profile.children.cycles-pp._raw_read_lock 0.01 ±223% +0.1 0.16 ± 9% perf-profile.children.cycles-pp.dequeue_entities 0.01 ±223% +0.1 0.16 ± 7% perf-profile.children.cycles-pp.dequeue_task_fair 0.12 ± 19% +0.2 0.28 ± 9% perf-profile.children.cycles-pp.try_to_wake_up 0.03 ±101% +0.2 0.19 ± 10% perf-profile.children.cycles-pp.enqueue_task_fair 0.02 ±141% +0.2 0.18 ± 12% perf-profile.children.cycles-pp.autoremove_wake_function 0.03 ±103% +0.2 0.20 ± 9% perf-profile.children.cycles-pp.enqueue_task 0.02 ±141% +0.2 0.19 ± 13% perf-profile.children.cycles-pp.__wake_up 0.03 ±101% +0.2 0.20 ± 9% perf-profile.children.cycles-pp.ttwu_do_activate 0.02 ±141% +0.2 0.19 ± 13% perf-profile.children.cycles-pp.__wake_up_common 0.03 ±103% +0.2 0.20 ± 11% perf-profile.children.cycles-pp.sched_ttwu_pending 0.04 ±101% +0.2 0.22 ± 9% perf-profile.children.cycles-pp.update_sg_lb_stats 0.04 ±102% +0.2 0.24 ± 9% perf-profile.children.cycles-pp.update_sd_lb_stats 0.04 ±101% +0.2 0.24 ± 8% perf-profile.children.cycles-pp.sched_balance_find_src_group 0.04 ±102% +0.2 0.24 ± 9% perf-profile.children.cycles-pp.__flush_smp_call_function_queue 0.48 ± 6% +0.2 0.70 ± 2% perf-profile.children.cycles-pp.sysvec_apic_timer_interrupt 0.50 ± 6% +0.2 0.74 ± 2% perf-profile.children.cycles-pp.asm_sysvec_apic_timer_interrupt 0.04 ±102% +0.2 0.28 ± 10% perf-profile.children.cycles-pp.jbd2_journal_commit_transaction 0.04 ±102% +0.2 0.29 ± 11% perf-profile.children.cycles-pp.kjournald2 0.10 ± 30% +0.2 0.36 ± 14% perf-profile.children.cycles-pp.kthread 0.10 ± 30% +0.2 0.36 ± 14% perf-profile.children.cycles-pp.ret_from_fork 0.10 ± 30% +0.2 0.36 ± 14% perf-profile.children.cycles-pp.ret_from_fork_asm 0.07 ± 58% +0.3 0.35 ± 8% perf-profile.children.cycles-pp.sched_balance_rq 0.06 ± 77% +0.3 0.36 ± 8% perf-profile.children.cycles-pp.sched_balance_newidle 0.10 ± 59% +0.3 0.42 ± 7% perf-profile.children.cycles-pp.schedule_preempt_disabled 0.09 ± 61% +0.3 0.43 ± 8% perf-profile.children.cycles-pp.pick_next_task_fair 0.09 ± 61% +0.4 0.44 ± 8% perf-profile.children.cycles-pp.__pick_next_task 2.47 ± 6% +0.4 2.83 ± 3% perf-profile.children.cycles-pp.intel_idle 0.14 ± 48% +0.6 0.70 ± 8% perf-profile.children.cycles-pp.schedule 2.52 ± 5% +0.6 3.07 ± 3% perf-profile.children.cycles-pp.cpuidle_enter 2.52 ± 5% +0.6 3.07 ± 3% perf-profile.children.cycles-pp.cpuidle_enter_state 0.16 ± 49% +0.6 0.76 ± 8% perf-profile.children.cycles-pp.__schedule 0.12 ± 56% +0.6 0.72 ± 11% perf-profile.children.cycles-pp.add_transaction_credits 2.53 ± 4% +0.6 3.15 ± 4% perf-profile.children.cycles-pp.cpuidle_idle_call 2.56 ± 4% +0.8 3.36 ± 4% perf-profile.children.cycles-pp.start_secondary 2.58 ± 4% +0.8 3.39 ± 4% perf-profile.children.cycles-pp.common_startup_64 2.58 ± 4% +0.8 3.39 ± 4% perf-profile.children.cycles-pp.cpu_startup_entry 2.58 ± 4% +0.8 3.39 ± 4% perf-profile.children.cycles-pp.do_idle 1.15 ± 52% +3.9 5.01 ± 9% perf-profile.children.cycles-pp._raw_spin_lock_irqsave 3.38 ± 57% +7.0 10.42 ± 6% perf-profile.children.cycles-pp.ext4_dirty_inode 3.42 ± 56% +7.1 10.53 ± 6% perf-profile.children.cycles-pp.__mark_inode_dirty 3.57 ± 58% +14.5 18.04 ± 11% perf-profile.children.cycles-pp.queued_read_lock_slowpath 3.98 ± 56% +15.6 19.60 ± 11% perf-profile.children.cycles-pp.start_this_handle 4.01 ± 56% +15.6 19.62 ± 11% perf-profile.children.cycles-pp.jbd2__journal_start 4.54 ± 59% +18.3 22.85 ± 10% perf-profile.children.cycles-pp.native_queued_spin_lock_slowpath 86.96 ± 3% -21.1 65.83 ± 4% perf-profile.self.cycles-pp.osq_lock 0.62 ± 2% -0.1 0.50 ± 3% perf-profile.self.cycles-pp.mutex_spin_on_owner 0.17 ± 2% -0.0 0.13 ± 4% perf-profile.self.cycles-pp._raw_spin_lock_irqsave 0.28 -0.0 0.25 ± 3% perf-profile.self.cycles-pp.__mutex_lock 0.09 ± 5% -0.0 0.07 ± 7% perf-profile.self.cycles-pp.osq_unlock 0.22 -0.0 0.20 ± 3% perf-profile.self.cycles-pp.ext4_mark_iloc_dirty 0.05 ± 8% +0.0 0.06 ± 7% perf-profile.self.cycles-pp.__ext4_handle_dirty_metadata 0.13 ± 3% +0.0 0.15 ± 5% perf-profile.self.cycles-pp.ext4_orphan_add 0.08 ± 4% +0.0 0.10 ± 6% perf-profile.self.cycles-pp.update_cfs_group 0.09 +0.0 0.12 ± 5% perf-profile.self.cycles-pp.find_get_block_common 0.12 ± 6% +0.0 0.15 ± 3% perf-profile.self.cycles-pp.entry_SYSRETQ_unsafe_stack 0.06 ± 15% +0.0 0.10 ± 7% perf-profile.self.cycles-pp.jbd2_journal_get_write_access 0.00 +0.1 0.05 perf-profile.self.cycles-pp.jbd2_write_access_granted 0.00 +0.1 0.05 perf-profile.self.cycles-pp.task_mm_cid_work 0.04 ± 45% +0.1 0.10 ± 4% perf-profile.self.cycles-pp._raw_spin_lock 0.01 ±223% +0.1 0.07 ± 8% perf-profile.self.cycles-pp.stop_this_handle 0.01 ±223% +0.1 0.07 ± 6% perf-profile.self.cycles-pp.jbd2_journal_stop 0.02 ±141% +0.1 0.09 ± 9% perf-profile.self.cycles-pp.update_load_avg 0.03 ±100% +0.1 0.16 ± 9% perf-profile.self.cycles-pp.update_sg_lb_stats 0.09 ± 22% +0.1 0.23 ± 8% perf-profile.self.cycles-pp._raw_read_lock 0.06 ± 75% +0.2 0.25 ± 8% perf-profile.self.cycles-pp.add_transaction_credits 0.04 ±102% +0.2 0.24 ± 11% perf-profile.self.cycles-pp.queued_read_lock_slowpath 2.47 ± 6% +0.4 2.83 ± 3% perf-profile.self.cycles-pp.intel_idle 0.18 ± 30% +0.4 0.56 ± 9% perf-profile.self.cycles-pp.start_this_handle 4.50 ± 59% +18.2 22.66 ± 10% perf-profile.self.cycles-pp.native_queued_spin_lock_slowpath *************************************************************************************************** lkp-csl-2sp10: 192 threads 4 sockets Intel(R) Xeon(R) Platinum 9242 CPU @ 2.30GHz (Cascade Lake) with 176G memory ========================================================================================= compiler/cpufreq_governor/disk/fs/kconfig/rootfs/tbox_group/test/testcase: gcc-12/performance/1SSD/ext4/x86_64-rhel-9.4/debian-12-x86_64-20240206.cgz/lkp-csl-2sp10/randomread.f/filebench commit: 01e807e18f ("ext4: make online defragmentation support large folios") 7ac67301e8 ("ext4: enable large folio for regular file") 01e807e18fd87937 7ac67301e82f02b77a5c8e7377a ---------------- --------------------------- %stddev %change %stddev \ | \ 1339 ± 4% -10.2% 1202 ± 4% perf-c2c.DRAM.remote 0.08 ± 13% -20.2% 0.06 ± 12% perf-sched.sch_delay.max.ms.devkmsg_read.vfs_read.ksys_read.do_syscall_64 31210 +1.1% 31555 vmstat.io.bo 3177 ±218% +81142.5% 2581268 ± 99% numa-meminfo.node2.Inactive 3177 ±218% +81142.5% 2581268 ± 99% numa-meminfo.node2.Inactive(file) 794.16 ±218% +81233.1% 645917 ± 99% numa-vmstat.node2.nr_inactive_file 794.16 ±218% +81233.0% 645916 ± 99% numa-vmstat.node2.nr_zone_inactive_file 66252 -4.0% 63604 proc-vmstat.nr_slab_reclaimable 2328335 -56.1% 1022608 proc-vmstat.numa_hit 2030258 -64.3% 724500 proc-vmstat.numa_local 2695 ± 2% +13.9% 3071 filebench.sum_bytes_mb/s 20707377 ± 2% +13.9% 23594148 filebench.sum_operations 345075 ± 2% +13.9% 393184 filebench.sum_operations/s 345075 ± 2% +13.9% 393184 filebench.sum_reads/s 0.00 -33.3% 0.00 filebench.sum_time_ms/op 11.85 +6.0% 12.56 perf-stat.i.MPKI 14126211 +1.5% 14333361 perf-stat.i.branch-misses 19038125 +11.3% 21192935 perf-stat.i.cache-misses 47685799 +7.9% 51436964 perf-stat.i.cache-references 2.69 -1.4% 2.65 perf-stat.i.cpi 12.56 +11.8% 14.04 perf-stat.overall.MPKI 4.48 +0.1 4.55 perf-stat.overall.branch-miss-rate% 39.99 +1.3 41.27 perf-stat.overall.cache-miss-rate% 168.91 -10.3% 151.57 perf-stat.overall.cycles-between-cache-misses 13998510 +1.5% 14206269 perf-stat.ps.branch-misses 18912177 +11.4% 21059829 perf-stat.ps.cache-misses 47298361 +7.9% 51025248 perf-stat.ps.cache-references 2.498e+11 -1.4% 2.463e+11 perf-stat.total.instructions Disclaimer: Results have been estimated based on internal Intel analysis and are provided for informational purposes only. Any difference in system hardware or software design or configuration may affect actual performance. -- 0-DAY CI Kernel Test Service https://github.com/intel/lkp-tests/wiki
Powered by blists - more mailing lists