[<prev] [next>] [day] [month] [year] [list]
Message-ID: <202601121044.6d8b5bac-lkp@intel.com>
Date: Mon, 12 Jan 2026 13:54:15 +0800
From: kernel test robot <oliver.sang@...el.com>
To: "Paul E. McKenney" <paulmck@...nel.org>
CC: <oe-lkp@...ts.linux.dev>, <lkp@...el.com>, Boqun Feng
<boqun.feng@...il.com>, kernel test robot <oliver.sang@...el.com>, "Andrii
Nakryiko" <andrii@...nel.org>, Alexei Starovoitov <ast@...nel.org>, "Peter
Zijlstra" <peterz@...radead.org>, Joel Fernandes <joelagnelf@...dia.com>,
<rcu@...r.kernel.org>, <linux-kernel@...r.kernel.org>
Subject: [linux-next:master] [rcu] c27cea4416:
stress-ng.fd-abuse.ops_per_sec 7.1% regression
Hello,
for this change, we reported
"[paulmckrcu:dev.2025.12.16a] [rcu] 1ac50ec628: stress-ng.memfd.ops_per_sec 3.4% improvement"
in
https://lore.kernel.org/all/202601071316.992a1d32-lkp@intel.com/
now we notice this commit is in linux-next/master.
we still catpure similar improvement as below. at the same time, we notice a
regression in another stress-ng test.
we don't have enough capability to analyze if this change really has performance
impact or just the code footprint change and the corresponding cache line change
impact these stress-ng tests.
since data is stable, we just report out FYI. please educate us if this report
is less meaningful. thanks
kernel test robot noticed a 7.1% regression of stress-ng.fd-abuse.ops_per_sec on:
commit: c27cea4416a396a1c5b6b3529dd925f92a69e7d3 ("rcu: Re-implement RCU Tasks Trace in terms of SRCU-fast")
https://git.kernel.org/cgit/linux/kernel/git/next/linux-next.git master
[still regression on linux-next/master f417b7ffcbef7d76b0d8860518f50dae0e7e5eda]
testcase: stress-ng
config: x86_64-rhel-9.4
compiler: gcc-14
test machine: 256 threads 2 sockets Intel(R) Xeon(R) 6768P CPU @ 2.4GHz (Granite Rapids) with 64G memory
parameters:
nr_threads: 100%
testtime: 60s
test: fd-abuse
cpufreq_governor: performance
In addition to that, the commit also has significant impact on the following tests:
+------------------+---------------------------------------------------------------------------------------------+
| testcase: change | stress-ng: stress-ng.memfd.ops_per_sec 3.5% improvement |
| test machine | 192 threads 2 sockets Intel(R) Xeon(R) 6740E CPU @ 2.4GHz (Sierra Forest) with 256G memory |
| test parameters | cpufreq_governor=performance |
| | nr_threads=100% |
| | test=memfd |
| | testtime=60s |
+------------------+---------------------------------------------------------------------------------------------+
If you fix the issue in a separate patch/commit (i.e. not just a new version of
the same patch/commit), kindly add following tags
| Reported-by: kernel test robot <oliver.sang@...el.com>
| Closes: https://lore.kernel.org/oe-lkp/202601121044.6d8b5bac-lkp@intel.com
Details are as below:
-------------------------------------------------------------------------------------------------->
The kernel config and materials to reproduce are available at:
https://download.01.org/0day-ci/archive/20260112/202601121044.6d8b5bac-lkp@intel.com
=========================================================================================
compiler/cpufreq_governor/kconfig/nr_threads/rootfs/tbox_group/test/testcase/testtime:
gcc-14/performance/x86_64-rhel-9.4/100%/debian-13-x86_64-20250902.cgz/lkp-gnr-2sp4/fd-abuse/stress-ng/60s
commit:
v6.19-rc1
c27cea4416 ("rcu: Re-implement RCU Tasks Trace in terms of SRCU-fast")
v6.19-rc1 c27cea4416a396a1c5b6b3529dd
---------------- ---------------------------
%stddev %change %stddev
\ | \
35053 ± 14% +133.6% 81881 ±126% numa-meminfo.node0.Shmem
8763 ± 14% +133.8% 20487 ±126% numa-vmstat.node0.nr_shmem
2.404e+09 -7.1% 2.234e+09 stress-ng.fd-abuse.ops
40118365 -7.1% 37281947 stress-ng.fd-abuse.ops_per_sec
147.78 -6.9% 137.58 stress-ng.time.user_time
94705 ± 2% -4.7% 90252 proc-vmstat.nr_slab_reclaimable
565933 -3.3% 547436 proc-vmstat.nr_slab_unreclaimable
6030483 ± 2% -6.0% 5669811 proc-vmstat.numa_hit
5766412 ± 2% -6.3% 5405494 proc-vmstat.numa_local
8618500 ± 2% -7.1% 8009296 proc-vmstat.pgalloc_normal
6877608 ± 4% -8.6% 6283782 ± 2% proc-vmstat.pgfree
0.28 -7.8% 0.26 perf-stat.i.MPKI
2.52e+10 -2.2% 2.464e+10 perf-stat.i.branch-instructions
27782554 -10.8% 24789456 perf-stat.i.cache-misses
1.509e+08 ± 5% -7.8% 1.391e+08 perf-stat.i.cache-references
8.10 +2.7% 8.31 perf-stat.i.cpi
34826 +13.6% 39567 perf-stat.i.cycles-between-cache-misses
1.099e+11 -2.6% 1.071e+11 perf-stat.i.instructions
0.13 -2.5% 0.12 perf-stat.i.ipc
0.25 -8.6% 0.23 perf-stat.overall.MPKI
8.18 +2.7% 8.39 perf-stat.overall.cpi
32484 +12.3% 36472 perf-stat.overall.cycles-between-cache-misses
0.12 -2.6% 0.12 perf-stat.overall.ipc
2.48e+10 -2.2% 2.425e+10 perf-stat.ps.branch-instructions
27233991 -10.9% 24262443 perf-stat.ps.cache-misses
1.479e+08 ± 5% -7.8% 1.363e+08 perf-stat.ps.cache-references
1.082e+11 -2.6% 1.054e+11 perf-stat.ps.instructions
6.605e+12 -2.5% 6.443e+12 perf-stat.total.instructions
1.75 -0.1 1.63 perf-profile.calltrace.cycles-pp.stress_fd_lseek
0.56 ± 2% -0.1 0.45 ± 37% perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.stress_fd_lseek
0.58 ± 2% -0.0 0.54 ± 2% perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.stress_fd_lseek
2.20 +0.0 2.21 perf-profile.calltrace.cycles-pp.inode_sb_list_add.new_inode.__shmem_get_inode.__shmem_file_setup.shmem_zero_setup_desc
49.74 +0.1 49.83 perf-profile.calltrace.cycles-pp.__mmap
49.71 +0.1 49.80 perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.__mmap
49.70 +0.1 49.80 perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.__mmap
47.24 +0.1 47.34 perf-profile.calltrace.cycles-pp.do_mmap.vm_mmap_pgoff.do_syscall_64.entry_SYSCALL_64_after_hwframe.__mmap
47.24 +0.1 47.35 perf-profile.calltrace.cycles-pp.vm_mmap_pgoff.do_syscall_64.entry_SYSCALL_64_after_hwframe.__mmap
47.19 +0.1 47.30 perf-profile.calltrace.cycles-pp.__mmap_region.do_mmap.vm_mmap_pgoff.do_syscall_64.entry_SYSCALL_64_after_hwframe
47.09 +0.1 47.21 perf-profile.calltrace.cycles-pp.__mmap_new_vma.__mmap_region.do_mmap.vm_mmap_pgoff.do_syscall_64
46.81 +0.1 46.93 perf-profile.calltrace.cycles-pp.new_inode.__shmem_get_inode.__shmem_file_setup.shmem_zero_setup.__mmap_new_vma
47.01 +0.1 47.13 perf-profile.calltrace.cycles-pp.shmem_zero_setup.__mmap_new_vma.__mmap_region.do_mmap.vm_mmap_pgoff
47.01 +0.1 47.13 perf-profile.calltrace.cycles-pp.__shmem_file_setup.shmem_zero_setup.__mmap_new_vma.__mmap_region.do_mmap
46.89 +0.1 47.01 perf-profile.calltrace.cycles-pp.__shmem_get_inode.__shmem_file_setup.shmem_zero_setup.__mmap_new_vma.__mmap_region
48.80 +0.1 48.93 perf-profile.calltrace.cycles-pp.native_queued_spin_lock_slowpath._raw_spin_lock.inode_sb_list_add.new_inode.__shmem_get_inode
47.17 +0.1 47.30 perf-profile.calltrace.cycles-pp.__munmap
47.16 +0.1 47.28 perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.__munmap
47.16 +0.1 47.28 perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.__munmap
46.74 +0.1 46.87 perf-profile.calltrace.cycles-pp.inode_sb_list_add.new_inode.__shmem_get_inode.__shmem_file_setup.shmem_zero_setup
47.00 +0.1 47.14 perf-profile.calltrace.cycles-pp.__fput.task_work_run.exit_to_user_mode_loop.do_syscall_64.entry_SYSCALL_64_after_hwframe
46.98 +0.1 47.12 perf-profile.calltrace.cycles-pp.__dentry_kill.finish_dput.__fput.task_work_run.exit_to_user_mode_loop
48.89 +0.1 49.03 perf-profile.calltrace.cycles-pp._raw_spin_lock.inode_sb_list_add.new_inode.__shmem_get_inode.__shmem_file_setup
47.01 +0.1 47.15 perf-profile.calltrace.cycles-pp.exit_to_user_mode_loop.do_syscall_64.entry_SYSCALL_64_after_hwframe.__munmap
46.62 +0.1 46.76 perf-profile.calltrace.cycles-pp.native_queued_spin_lock_slowpath._raw_spin_lock.evict.__dentry_kill.finish_dput
46.98 +0.1 47.13 perf-profile.calltrace.cycles-pp.finish_dput.__fput.task_work_run.exit_to_user_mode_loop.do_syscall_64
47.01 +0.1 47.15 perf-profile.calltrace.cycles-pp.task_work_run.exit_to_user_mode_loop.do_syscall_64.entry_SYSCALL_64_after_hwframe.__munmap
46.77 +0.1 46.92 perf-profile.calltrace.cycles-pp._raw_spin_lock.evict.__dentry_kill.finish_dput.__fput
46.88 +0.1 47.02 perf-profile.calltrace.cycles-pp.evict.__dentry_kill.finish_dput.__fput.task_work_run
1.76 -0.1 1.63 perf-profile.children.cycles-pp.stress_fd_lseek
0.44 -0.0 0.41 perf-profile.children.cycles-pp.entry_SYSCALL_64
0.47 -0.0 0.44 perf-profile.children.cycles-pp.llseek
0.24 -0.0 0.22 ± 2% perf-profile.children.cycles-pp.entry_SYSRETQ_unsafe_stack
0.11 ± 4% -0.0 0.09 ± 5% perf-profile.children.cycles-pp.__inet_listen_sk
0.20 -0.0 0.19 perf-profile.children.cycles-pp.__vm_munmap
0.20 -0.0 0.19 perf-profile.children.cycles-pp.__x64_sys_munmap
0.11 -0.0 0.10 ± 3% perf-profile.children.cycles-pp.__sys_listen
0.11 -0.0 0.10 ± 3% perf-profile.children.cycles-pp.__x64_sys_listen
0.20 -0.0 0.19 perf-profile.children.cycles-pp.arch_exit_to_user_mode_prepare
0.10 -0.0 0.09 perf-profile.children.cycles-pp.inet_csk_get_port
0.09 -0.0 0.08 perf-profile.children.cycles-pp.__get_unmapped_area
0.09 -0.0 0.08 perf-profile.children.cycles-pp.__x64_sys_pselect6
0.08 -0.0 0.07 perf-profile.children.cycles-pp.alloc_inode
0.12 -0.0 0.11 perf-profile.children.cycles-pp.kmem_cache_alloc_noprof
0.11 -0.0 0.10 perf-profile.children.cycles-pp.rcu_core
0.08 -0.0 0.07 perf-profile.children.cycles-pp.run_ksoftirqd
0.08 -0.0 0.07 perf-profile.children.cycles-pp.smpboot_thread_fn
98.70 +0.1 98.79 perf-profile.children.cycles-pp.entry_SYSCALL_64_after_hwframe
49.76 +0.1 49.85 perf-profile.children.cycles-pp.__mmap
98.63 +0.1 98.73 perf-profile.children.cycles-pp.do_syscall_64
49.68 +0.1 49.78 perf-profile.children.cycles-pp.vm_mmap_pgoff
49.66 +0.1 49.76 perf-profile.children.cycles-pp.do_mmap
49.55 +0.1 49.66 perf-profile.children.cycles-pp.__mmap_region
47.17 +0.1 47.28 perf-profile.children.cycles-pp.__mmap_new_vma
47.01 +0.1 47.13 perf-profile.children.cycles-pp.shmem_zero_setup
47.25 +0.1 47.37 perf-profile.children.cycles-pp.__munmap
49.22 +0.1 49.35 perf-profile.children.cycles-pp.__shmem_file_setup
47.07 +0.1 47.20 perf-profile.children.cycles-pp.__fput
49.10 +0.1 49.23 perf-profile.children.cycles-pp.__shmem_get_inode
49.02 +0.1 49.15 perf-profile.children.cycles-pp.new_inode
48.94 +0.1 49.08 perf-profile.children.cycles-pp.inode_sb_list_add
47.03 +0.1 47.17 perf-profile.children.cycles-pp.exit_to_user_mode_loop
46.98 +0.1 47.13 perf-profile.children.cycles-pp.__dentry_kill
46.98 +0.1 47.13 perf-profile.children.cycles-pp.finish_dput
47.01 +0.1 47.15 perf-profile.children.cycles-pp.task_work_run
46.88 +0.1 47.02 perf-profile.children.cycles-pp.evict
95.49 +0.3 95.75 perf-profile.children.cycles-pp.native_queued_spin_lock_slowpath
95.75 +0.3 96.03 perf-profile.children.cycles-pp._raw_spin_lock
0.18 ± 2% -0.0 0.17 ± 2% perf-profile.self.cycles-pp.arch_exit_to_user_mode_prepare
0.16 -0.0 0.15 perf-profile.self.cycles-pp.do_syscall_64
0.23 -0.0 0.22 perf-profile.self.cycles-pp.entry_SYSRETQ_unsafe_stack
0.11 -0.0 0.10 perf-profile.self.cycles-pp.entry_SYSCALL_64
0.08 -0.0 0.07 perf-profile.self.cycles-pp.entry_SYSCALL_64_after_hwframe
0.15 -0.0 0.14 perf-profile.self.cycles-pp.llseek
95.08 +0.3 95.35 perf-profile.self.cycles-pp.native_queued_spin_lock_slowpath
***************************************************************************************************
lkp-srf-2sp2: 192 threads 2 sockets Intel(R) Xeon(R) 6740E CPU @ 2.4GHz (Sierra Forest) with 256G memory
=========================================================================================
compiler/cpufreq_governor/kconfig/nr_threads/rootfs/tbox_group/test/testcase/testtime:
gcc-14/performance/x86_64-rhel-9.4/100%/debian-13-x86_64-20250902.cgz/lkp-srf-2sp2/memfd/stress-ng/60s
commit:
v6.19-rc1
c27cea4416 ("rcu: Re-implement RCU Tasks Trace in terms of SRCU-fast")
v6.19-rc1 c27cea4416a396a1c5b6b3529dd
---------------- ---------------------------
%stddev %change %stddev
\ | \
257709 +47.4% 379897 meminfo.AnonHugePages
185.79 +48.4% 275.65 boot-time.boot
35229 +49.0% 52482 boot-time.idle
279.39 +32.1% 369.06 uptime.boot
41662 +41.3% 58880 uptime.idle
107491 -3.7% 103488 stress-ng.memfd.nanosecs_per_memfd_create_call
193402 +3.5% 200220 stress-ng.memfd.ops
3224 +3.5% 3338 stress-ng.memfd.ops_per_sec
99066540 +3.5% 1.026e+08 stress-ng.time.minor_page_faults
190038 +1.7% 193305 proc-vmstat.nr_anon_pages
125.88 +47.4% 185.50 proc-vmstat.nr_anon_transparent_hugepages
39831 -3.0% 38648 proc-vmstat.nr_kernel_stack
1.525e+08 +3.5% 1.579e+08 proc-vmstat.numa_hit
1.523e+08 +3.5% 1.577e+08 proc-vmstat.numa_local
1.629e+08 +3.6% 1.688e+08 proc-vmstat.pgalloc_normal
99551866 +3.5% 1.03e+08 proc-vmstat.pgfault
1.622e+08 +3.6% 1.681e+08 proc-vmstat.pgfree
247638 +36.3% 337493 sched_debug.cpu.clock.avg
247674 +36.3% 337526 sched_debug.cpu.clock.max
247599 +36.3% 337457 sched_debug.cpu.clock.min
247436 +36.3% 337285 sched_debug.cpu.clock_task.avg
247527 +36.3% 337377 sched_debug.cpu.clock_task.max
241459 +36.8% 330354 sched_debug.cpu.clock_task.min
247600 +36.3% 337458 sched_debug.cpu_clk
247342 +36.3% 337200 sched_debug.ktime
248152 +36.1% 337703 sched_debug.sched_clk
2.25 +1.5% 2.29 perf-stat.i.MPKI
1.644e+10 +1.9% 1.675e+10 perf-stat.i.branch-instructions
0.24 ± 4% +0.0 0.25 perf-stat.i.branch-miss-rate%
38026731 ± 5% +9.0% 41438665 perf-stat.i.branch-misses
1.712e+08 +3.6% 1.774e+08 perf-stat.i.cache-misses
2.891e+08 +3.3% 2.987e+08 perf-stat.i.cache-references
7.88 -2.0% 7.72 perf-stat.i.cpi
294.71 +2.6% 302.24 perf-stat.i.cpu-migrations
3506 -3.5% 3385 perf-stat.i.cycles-between-cache-misses
7.629e+10 +2.1% 7.785e+10 perf-stat.i.instructions
0.13 +2.1% 0.13 perf-stat.i.ipc
17.00 +3.5% 17.60 perf-stat.i.metric.K/sec
1632141 +3.5% 1689961 perf-stat.i.minor-faults
1632141 +3.5% 1689961 perf-stat.i.page-faults
2.25 +1.5% 2.28 perf-stat.overall.MPKI
0.23 ± 4% +0.0 0.25 perf-stat.overall.branch-miss-rate%
7.91 -2.0% 7.75 perf-stat.overall.cpi
3521 -3.4% 3401 perf-stat.overall.cycles-between-cache-misses
0.13 +2.0% 0.13 perf-stat.overall.ipc
1.618e+10 +1.8% 1.648e+10 perf-stat.ps.branch-instructions
37315073 ± 5% +9.1% 40712563 perf-stat.ps.branch-misses
1.686e+08 +3.5% 1.745e+08 perf-stat.ps.cache-misses
2.849e+08 +3.2% 2.941e+08 perf-stat.ps.cache-references
6628 +1.1% 6704 perf-stat.ps.context-switches
288.02 +2.7% 295.69 perf-stat.ps.cpu-migrations
7.506e+10 +2.0% 7.657e+10 perf-stat.ps.instructions
1606028 +3.5% 1662399 perf-stat.ps.minor-faults
1606028 +3.5% 1662399 perf-stat.ps.page-faults
45.98 -0.3 45.73 perf-profile.calltrace.cycles-pp._raw_spin_lock.inode_sb_list_add.new_inode.__shmem_get_inode.__shmem_file_setup
46.05 -0.3 45.80 perf-profile.calltrace.cycles-pp.inode_sb_list_add.new_inode.__shmem_get_inode.__shmem_file_setup.__x64_sys_memfd_create
45.86 -0.2 45.61 perf-profile.calltrace.cycles-pp.native_queued_spin_lock_slowpath._raw_spin_lock.inode_sb_list_add.new_inode.__shmem_get_inode
46.28 -0.2 46.03 perf-profile.calltrace.cycles-pp.__shmem_get_inode.__shmem_file_setup.__x64_sys_memfd_create.do_syscall_64.entry_SYSCALL_64_after_hwframe
46.19 -0.2 45.94 perf-profile.calltrace.cycles-pp.new_inode.__shmem_get_inode.__shmem_file_setup.__x64_sys_memfd_create.do_syscall_64
46.54 -0.2 46.30 perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.memfd_create
46.49 -0.2 46.25 perf-profile.calltrace.cycles-pp.__shmem_file_setup.__x64_sys_memfd_create.do_syscall_64.entry_SYSCALL_64_after_hwframe.memfd_create
46.54 -0.2 46.30 perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.memfd_create
46.56 -0.2 46.33 perf-profile.calltrace.cycles-pp.memfd_create
46.53 -0.2 46.30 perf-profile.calltrace.cycles-pp.__x64_sys_memfd_create.do_syscall_64.entry_SYSCALL_64_after_hwframe.memfd_create
45.51 -0.2 45.29 perf-profile.calltrace.cycles-pp._raw_spin_lock.evict.__dentry_kill.finish_dput.__fput
45.34 -0.2 45.13 perf-profile.calltrace.cycles-pp.native_queued_spin_lock_slowpath._raw_spin_lock.evict.__dentry_kill.finish_dput
46.33 -0.2 46.15 perf-profile.calltrace.cycles-pp.evict.__dentry_kill.finish_dput.__fput.task_work_run
46.54 -0.2 46.36 perf-profile.calltrace.cycles-pp.finish_dput.__fput.task_work_run.exit_to_user_mode_loop.do_syscall_64
46.67 -0.2 46.48 perf-profile.calltrace.cycles-pp.close_range
46.54 -0.2 46.36 perf-profile.calltrace.cycles-pp.__dentry_kill.finish_dput.__fput.task_work_run.exit_to_user_mode_loop
46.62 -0.2 46.44 perf-profile.calltrace.cycles-pp.__fput.task_work_run.exit_to_user_mode_loop.do_syscall_64.entry_SYSCALL_64_after_hwframe
46.67 -0.2 46.48 perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.close_range
46.67 -0.2 46.48 perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.close_range
46.64 -0.2 46.46 perf-profile.calltrace.cycles-pp.exit_to_user_mode_loop.do_syscall_64.entry_SYSCALL_64_after_hwframe.close_range
46.64 -0.2 46.46 perf-profile.calltrace.cycles-pp.task_work_run.exit_to_user_mode_loop.do_syscall_64.entry_SYSCALL_64_after_hwframe.close_range
1.03 +0.0 1.05 perf-profile.calltrace.cycles-pp.handle_mm_fault.do_user_addr_fault.exc_page_fault.asm_exc_page_fault.stress_memfd_child
1.28 +0.0 1.31 perf-profile.calltrace.cycles-pp.asm_exc_page_fault.stress_memfd_child
1.00 +0.0 1.02 perf-profile.calltrace.cycles-pp.__handle_mm_fault.handle_mm_fault.do_user_addr_fault.exc_page_fault.asm_exc_page_fault
1.12 +0.0 1.15 perf-profile.calltrace.cycles-pp.exc_page_fault.asm_exc_page_fault.stress_memfd_child
0.78 ± 2% +0.0 0.82 perf-profile.calltrace.cycles-pp.shmem_evict_inode.evict.__dentry_kill.finish_dput.__fput
0.75 ± 2% +0.0 0.79 perf-profile.calltrace.cycles-pp.shmem_undo_range.shmem_evict_inode.evict.__dentry_kill.finish_dput
0.87 +0.0 0.91 ± 3% perf-profile.calltrace.cycles-pp.__mmap_region.mmap_region.do_mmap.vm_mmap_pgoff.ksys_mmap_pgoff
0.88 +0.0 0.92 ± 3% perf-profile.calltrace.cycles-pp.mmap_region.do_mmap.vm_mmap_pgoff.ksys_mmap_pgoff.do_syscall_64
1.09 +0.1 1.15 ± 2% perf-profile.calltrace.cycles-pp.vm_mmap_pgoff.ksys_mmap_pgoff.do_syscall_64.entry_SYSCALL_64_after_hwframe.__mmap
1.10 +0.1 1.17 ± 2% perf-profile.calltrace.cycles-pp.ksys_mmap_pgoff.do_syscall_64.entry_SYSCALL_64_after_hwframe.__mmap
1.12 +0.1 1.18 ± 2% perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.__mmap
1.12 +0.1 1.19 ± 3% perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.__mmap
1.07 +0.1 1.14 ± 3% perf-profile.calltrace.cycles-pp.do_mmap.vm_mmap_pgoff.ksys_mmap_pgoff.do_syscall_64.entry_SYSCALL_64_after_hwframe
1.30 +0.1 1.37 ± 2% perf-profile.calltrace.cycles-pp.__mmap
0.60 ± 4% +0.1 0.70 ± 5% perf-profile.calltrace.cycles-pp.folios_put_refs.shmem_undo_range.shmem_fallocate.vfs_fallocate.__x64_sys_fallocate
1.16 ± 3% +0.2 1.35 ± 7% perf-profile.calltrace.cycles-pp.shmem_undo_range.shmem_fallocate.vfs_fallocate.__x64_sys_fallocate.do_syscall_64
1.71 ± 2% +0.2 1.93 ± 5% perf-profile.calltrace.cycles-pp.shmem_fallocate.vfs_fallocate.__x64_sys_fallocate.do_syscall_64.entry_SYSCALL_64_after_hwframe
1.74 ± 2% +0.2 1.97 ± 5% perf-profile.calltrace.cycles-pp.__x64_sys_fallocate.do_syscall_64.entry_SYSCALL_64_after_hwframe
2.26 +0.2 2.48 ± 4% perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe
2.25 +0.2 2.48 ± 4% perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe
1.72 ± 2% +0.2 1.95 ± 5% perf-profile.calltrace.cycles-pp.vfs_fallocate.__x64_sys_fallocate.do_syscall_64.entry_SYSCALL_64_after_hwframe
92.24 -0.5 91.78 perf-profile.children.cycles-pp._raw_spin_lock
46.05 -0.3 45.80 perf-profile.children.cycles-pp.inode_sb_list_add
46.28 -0.2 46.03 perf-profile.children.cycles-pp.__shmem_get_inode
46.19 -0.2 45.94 perf-profile.children.cycles-pp.new_inode
46.49 -0.2 46.25 perf-profile.children.cycles-pp.__shmem_file_setup
46.54 -0.2 46.30 perf-profile.children.cycles-pp.__x64_sys_memfd_create
46.57 -0.2 46.34 perf-profile.children.cycles-pp.memfd_create
92.33 -0.2 92.11 perf-profile.children.cycles-pp.native_queued_spin_lock_slowpath
46.82 -0.2 46.63 perf-profile.children.cycles-pp.evict
47.03 -0.2 46.84 perf-profile.children.cycles-pp.finish_dput
47.03 -0.2 46.84 perf-profile.children.cycles-pp.__dentry_kill
47.11 -0.2 46.92 perf-profile.children.cycles-pp.__fput
46.65 -0.2 46.46 perf-profile.children.cycles-pp.exit_to_user_mode_loop
46.65 -0.2 46.46 perf-profile.children.cycles-pp.task_work_run
46.67 -0.2 46.49 perf-profile.children.cycles-pp.close_range
97.54 -0.1 97.44 perf-profile.children.cycles-pp.do_syscall_64
97.56 -0.1 97.46 perf-profile.children.cycles-pp.entry_SYSCALL_64_after_hwframe
0.26 -0.0 0.24 perf-profile.children.cycles-pp.shmem_add_to_page_cache
0.24 ± 2% -0.0 0.23 perf-profile.children.cycles-pp.xas_store
0.19 ± 2% -0.0 0.17 ± 2% perf-profile.children.cycles-pp.truncate_inode_folio
0.54 -0.0 0.53 perf-profile.children.cycles-pp.shmem_alloc_and_add_folio
0.13 -0.0 0.12 perf-profile.children.cycles-pp.__filemap_remove_folio
0.07 -0.0 0.06 perf-profile.children.cycles-pp.mas_walk
0.09 -0.0 0.08 perf-profile.children.cycles-pp.vms_gather_munmap_vmas
0.17 +0.0 0.18 perf-profile.children.cycles-pp.do_ftruncate
0.09 +0.0 0.10 perf-profile.children.cycles-pp.notify_change
0.31 +0.0 0.32 perf-profile.children.cycles-pp.llseek
0.23 +0.0 0.24 ± 2% perf-profile.children.cycles-pp.ftruncate64
0.15 ± 3% +0.0 0.17 ± 2% perf-profile.children.cycles-pp.flush_tlb_mm_range
1.03 +0.0 1.05 perf-profile.children.cycles-pp.handle_mm_fault
0.13 ± 3% +0.0 0.15 ± 2% perf-profile.children.cycles-pp.unmapped_area_topdown
0.46 +0.0 0.48 perf-profile.children.cycles-pp.run_ksoftirqd
0.16 ± 2% +0.0 0.18 ± 2% perf-profile.children.cycles-pp.__get_unmapped_area
0.15 ± 2% +0.0 0.17 ± 2% perf-profile.children.cycles-pp.arch_get_unmapped_area_topdown
1.00 +0.0 1.02 perf-profile.children.cycles-pp.__handle_mm_fault
0.52 +0.0 0.54 perf-profile.children.cycles-pp.rcu_core
0.16 ± 3% +0.0 0.18 ± 3% perf-profile.children.cycles-pp.shmem_get_unmapped_area
0.32 +0.0 0.34 perf-profile.children.cycles-pp.unmap_page_range
0.30 +0.0 0.32 perf-profile.children.cycles-pp.zap_pmd_range
0.20 ± 2% +0.0 0.22 perf-profile.children.cycles-pp.zap_page_range_single_batched
0.28 +0.0 0.31 perf-profile.children.cycles-pp.zap_pte_range
0.23 +0.0 0.26 perf-profile.children.cycles-pp.unmap_mapping_range
0.13 ± 3% +0.0 0.16 ± 3% perf-profile.children.cycles-pp.vm_unmapped_area
0.15 +0.0 0.18 perf-profile.children.cycles-pp.perf_event_mmap
0.14 +0.0 0.17 perf-profile.children.cycles-pp.perf_event_mmap_event
0.21 +0.0 0.24 perf-profile.children.cycles-pp.zap_page_range_single
1.44 +0.0 1.47 perf-profile.children.cycles-pp.asm_exc_page_fault
0.78 ± 2% +0.0 0.82 perf-profile.children.cycles-pp.shmem_evict_inode
0.87 +0.0 0.91 ± 3% perf-profile.children.cycles-pp.__mmap_region
0.88 +0.0 0.92 ± 3% perf-profile.children.cycles-pp.mmap_region
0.00 +0.1 0.05 perf-profile.children.cycles-pp.fault_dirty_shared_page
0.00 +0.1 0.05 perf-profile.children.cycles-pp.shmem_recalc_inode
1.07 +0.1 1.14 ± 3% perf-profile.children.cycles-pp.do_mmap
1.10 +0.1 1.17 ± 2% perf-profile.children.cycles-pp.ksys_mmap_pgoff
1.09 +0.1 1.16 ± 3% perf-profile.children.cycles-pp.vm_mmap_pgoff
1.30 +0.1 1.38 ± 2% perf-profile.children.cycles-pp.__mmap
0.41 ± 3% +0.1 0.51 ± 11% perf-profile.children.cycles-pp.__folio_batch_release
0.41 ± 3% +0.1 0.51 ± 11% perf-profile.children.cycles-pp.folio_batch_move_lru
0.41 ± 3% +0.1 0.51 ± 11% perf-profile.children.cycles-pp.lru_add_drain_cpu
0.51 ± 4% +0.1 0.65 ± 11% perf-profile.children.cycles-pp.__page_cache_release
1.11 ± 3% +0.1 1.24 ± 4% perf-profile.children.cycles-pp.folios_put_refs
1.71 ± 2% +0.2 1.93 ± 5% perf-profile.children.cycles-pp.shmem_fallocate
1.74 ± 2% +0.2 1.97 ± 5% perf-profile.children.cycles-pp.__x64_sys_fallocate
1.72 ± 2% +0.2 1.95 ± 5% perf-profile.children.cycles-pp.vfs_fallocate
0.66 ± 4% +0.2 0.89 ± 14% perf-profile.children.cycles-pp.folio_lruvec_lock_irqsave
0.70 ± 4% +0.2 0.93 ± 14% perf-profile.children.cycles-pp._raw_spin_lock_irqsave
1.92 ± 2% +0.2 2.15 ± 5% perf-profile.children.cycles-pp.shmem_undo_range
91.97 -0.2 91.76 perf-profile.self.cycles-pp.native_queued_spin_lock_slowpath
0.07 -0.0 0.06 perf-profile.self.cycles-pp.mas_walk
0.08 +0.0 0.09 perf-profile.self.cycles-pp.entry_SYSCALL_64
Disclaimer:
Results have been estimated based on internal Intel analysis and are provided
for informational purposes only. Any difference in system hardware or software
design or configuration may affect actual performance.
--
0-DAY CI Kernel Test Service
https://github.com/intel/lkp-tests/wiki
Powered by blists - more mailing lists