[<prev] [next>] [day] [month] [year] [list]
Message-ID: <202411072132.a8d2cf0f-oliver.sang@intel.com>
Date: Thu, 7 Nov 2024 22:10:26 +0800
From: kernel test robot <oliver.sang@...el.com>
To: Vlastimil Babka <vbabka@...e.cz>
CC: <oe-lkp@...ts.linux.dev>, <lkp@...el.com>, <linux-kernel@...r.kernel.org>,
Andrew Morton <akpm@...ux-foundation.org>, Michael Matz <matz@...e.de>,
Gabriel Krisman Bertazi <gabriel@...sman.be>, Matthias Bodenbinder
<matthias@...enbinder.de>, Lorenzo Stoakes <lorenzo.stoakes@...cle.com>,
"Yang Shi" <yang@...amperecomputing.com>, Rik van Riel <riel@...riel.com>,
Jann Horn <jannh@...gle.com>, "Liam R. Howlett" <Liam.Howlett@...cle.com>,
Petr Tesarik <ptesarik@...e.com>, Thorsten Leemhuis
<regressions@...mhuis.info>, <linux-mm@...ck.org>, <ying.huang@...el.com>,
<feng.tang@...el.com>, <fengwei.yin@...el.com>, <oliver.sang@...el.com>
Subject: [linus:master] [mm, mmap] d4148aeab4: will-it-scale.per_process_ops
3888.9% improvement
Hello,
kernel test robot noticed a 3888.9% improvement of will-it-scale.per_process_ops on:
commit: d4148aeab412432bf928f311eca8a2ba52bb05df ("mm, mmap: limit THP alignment of anonymous mappings to PMD-aligned sizes")
https://git.kernel.org/cgit/linux/kernel/git/torvalds/linux.git master
[test failed on linux-next/master 5b913f5d7d7fe0f567dea8605f21da6eaa1735fb]
testcase: will-it-scale
config: x86_64-rhel-8.3
compiler: gcc-12
test machine: 224 threads 4 sockets Intel(R) Xeon(R) Platinum 8380H CPU @ 2.90GHz (Cooper Lake) with 192G memory
parameters:
nr_task: 100%
mode: process
test: malloc1
cpufreq_governor: performance
In addition to that, the commit also has significant impact on the following tests:
+------------------+---------------------------------------------------------------------------------------------+
| testcase: change | stress-ng: stress-ng.radixsort.ops_per_sec 9.2% regression |
| test machine | 224 threads 2 sockets Intel(R) Xeon(R) Platinum 8480CTDX (Sapphire Rapids) with 256G memory |
| test parameters | cpufreq_governor=performance |
| | nr_threads=100% |
| | test=radixsort |
| | testtime=60s |
+------------------+---------------------------------------------------------------------------------------------+
Details are as below:
-------------------------------------------------------------------------------------------------->
The kernel config and materials to reproduce are available at:
https://download.01.org/0day-ci/archive/20241107/202411072132.a8d2cf0f-oliver.sang@intel.com
=========================================================================================
compiler/cpufreq_governor/kconfig/mode/nr_task/rootfs/tbox_group/test/testcase:
gcc-12/performance/x86_64-rhel-8.3/process/100%/debian-12-x86_64-20240206.cgz/lkp-cpl-4sp2/malloc1/will-it-scale
commit:
15e8156713 ("mm: shrinker: avoid memleak in alloc_shrinker_info")
d4148aeab4 ("mm, mmap: limit THP alignment of anonymous mappings to PMD-aligned sizes")
15e8156713cc3803 d4148aeab412432bf928f311eca
---------------- ---------------------------
%stddev %change %stddev
\ | \
255351 +4.5% 266817 vmstat.system.in
1.00 -0.8 0.20 ± 7% mpstat.cpu.all.irq%
0.13 +1.1 1.28 ± 32% mpstat.cpu.all.soft%
0.49 ± 15% +1.3 1.77 ± 27% mpstat.cpu.all.usr%
4143 ± 5% +181.3% 11654 ± 7% perf-c2c.DRAM.remote
2500 ± 3% +2209.9% 57748 ± 15% perf-c2c.HITM.local
3810 ± 2% +1461.0% 59473 ± 15% perf-c2c.HITM.total
78888 +3883.3% 3142322 ± 22% will-it-scale.224.processes
351.67 +3888.9% 14027 ± 22% will-it-scale.per_process_ops
78888 +3883.3% 3142322 ± 22% will-it-scale.workload
702940 ± 15% +123.4% 1570646 ± 11% meminfo.Active
702881 ± 15% +123.5% 1570598 ± 11% meminfo.Active(anon)
3953778 +28.4% 5075703 ± 4% meminfo.Cached
382618 +117.7% 832966 ± 10% meminfo.SUnreclaim
913559 ± 5% +122.8% 2035523 ± 11% meminfo.Shmem
528864 +85.5% 980938 ± 8% meminfo.Slab
12139918 +3685.1% 4.595e+08 ± 22% numa-numastat.node0.local_node
12236476 +3655.6% 4.595e+08 ± 22% numa-numastat.node0.numa_hit
12182775 +3990.8% 4.984e+08 ± 23% numa-numastat.node1.local_node
12264802 +3963.8% 4.984e+08 ± 23% numa-numastat.node1.numa_hit
12222359 +3921.0% 4.915e+08 ± 23% numa-numastat.node2.local_node
12320084 +3889.7% 4.915e+08 ± 23% numa-numastat.node2.numa_hit
12317018 +3904.7% 4.933e+08 ± 22% numa-numastat.node3.local_node
12388608 +3881.9% 4.933e+08 ± 22% numa-numastat.node3.numa_hit
27453504 ± 11% +20.0% 32957345 ± 2% sched_debug.cfs_rq:/.avg_vruntime.min
1.78 ± 2% +12.5% 2.00 ± 4% sched_debug.cfs_rq:/.h_nr_running.max
27453504 ± 11% +20.0% 32957377 ± 2% sched_debug.cfs_rq:/.min_vruntime.min
384.34 ± 6% +28.2% 492.57 ± 4% sched_debug.cfs_rq:/.util_est.avg
113.34 ± 7% +87.5% 212.55 ± 11% sched_debug.cfs_rq:/.util_est.stddev
395.67 ± 6% -91.8% 32.63 ± 7% sched_debug.cpu.clock.stddev
0.00 ± 5% -85.8% 0.00 ± 26% sched_debug.cpu.next_balance.stddev
1.78 ± 2% +12.5% 2.00 ± 8% sched_debug.cpu.nr_running.max
5.99 ± 9% -16.9% 4.98 ± 8% sched_debug.cpu.nr_uninterruptible.stddev
369561 ± 5% -48.5% 190196 ± 65% numa-meminfo.node0.AnonPages.max
108019 ± 4% +102.2% 218399 ± 12% numa-meminfo.node0.SUnreclaim
148129 ± 16% +86.2% 275784 ± 14% numa-meminfo.node0.Slab
73527 ± 61% -53.5% 34174 ± 80% numa-meminfo.node1.Mapped
94870 ± 4% +122.7% 211285 ± 9% numa-meminfo.node1.SUnreclaim
136858 ± 15% +83.3% 250864 ± 11% numa-meminfo.node1.Slab
89356 ± 2% +124.1% 200203 ± 10% numa-meminfo.node2.SUnreclaim
118047 ± 16% +89.0% 223166 ± 15% numa-meminfo.node2.Slab
698747 ± 15% +117.2% 1517832 ± 12% numa-meminfo.node3.Active
698731 ± 15% +117.2% 1517816 ± 12% numa-meminfo.node3.Active(anon)
90353 ± 7% +122.1% 200636 ± 10% numa-meminfo.node3.SUnreclaim
902916 ± 5% +118.4% 1972406 ± 12% numa-meminfo.node3.Shmem
125802 ± 17% +81.8% 228694 ± 8% numa-meminfo.node3.Slab
175727 ± 15% +123.5% 392704 ± 11% proc-vmstat.nr_active_anon
988408 +28.4% 1268872 ± 4% proc-vmstat.nr_file_pages
228353 ± 5% +122.8% 508826 ± 11% proc-vmstat.nr_shmem
36558 +1.2% 36996 proc-vmstat.nr_slab_reclaimable
95650 +117.2% 207775 ± 10% proc-vmstat.nr_slab_unreclaimable
175727 ± 15% +123.5% 392704 ± 11% proc-vmstat.nr_zone_active_anon
61863 ± 20% +50.4% 93011 ± 17% proc-vmstat.numa_hint_faults
27849 ± 47% +108.0% 57912 ± 22% proc-vmstat.numa_hint_faults_local
49211744 +3847.8% 1.943e+09 ± 22% proc-vmstat.numa_hit
48863846 +3875.5% 1.943e+09 ± 22% proc-vmstat.numa_local
102439 ± 38% +399.0% 511155 ± 5% proc-vmstat.pgactivate
1.218e+10 -83.3% 2.036e+09 ± 23% proc-vmstat.pgalloc_normal
25163379 +3665.3% 9.475e+08 ± 22% proc-vmstat.pgfault
1.218e+10 -83.3% 2.035e+09 ± 23% proc-vmstat.pgfree
23741008 -100.0% 0.00 proc-vmstat.thp_fault_alloc
27004 ± 4% +102.3% 54625 ± 12% numa-vmstat.node0.nr_slab_unreclaimable
12236166 +3655.7% 4.595e+08 ± 22% numa-vmstat.node0.numa_hit
12139608 +3685.2% 4.595e+08 ± 22% numa-vmstat.node0.numa_local
18419 ± 61% -53.5% 8573 ± 80% numa-vmstat.node1.nr_mapped
23716 ± 4% +122.9% 52871 ± 9% numa-vmstat.node1.nr_slab_unreclaimable
12263982 +3964.1% 4.984e+08 ± 23% numa-vmstat.node1.numa_hit
12181955 +3991.1% 4.984e+08 ± 23% numa-vmstat.node1.numa_local
22339 ± 2% +124.8% 50226 ± 10% numa-vmstat.node2.nr_slab_unreclaimable
12319433 +3889.9% 4.915e+08 ± 23% numa-vmstat.node2.numa_hit
12221708 +3921.2% 4.915e+08 ± 23% numa-vmstat.node2.numa_local
174568 ± 15% +117.3% 379260 ± 12% numa-vmstat.node3.nr_active_anon
225581 ± 5% +118.6% 493053 ± 12% numa-vmstat.node3.nr_shmem
22588 ± 7% +122.8% 50325 ± 11% numa-vmstat.node3.nr_slab_unreclaimable
174566 ± 15% +117.3% 379259 ± 12% numa-vmstat.node3.nr_zone_active_anon
12386775 +3882.4% 4.933e+08 ± 22% numa-vmstat.node3.numa_hit
12315185 +3905.2% 4.933e+08 ± 22% numa-vmstat.node3.numa_local
20.80 -88.0% 2.51 ± 4% perf-stat.i.MPKI
1.314e+09 +1555.3% 2.175e+10 ± 18% perf-stat.i.branch-instructions
0.79 ± 3% -0.4 0.35 perf-stat.i.branch-miss-rate%
12157447 ± 2% +510.1% 74172493 ± 18% perf-stat.i.branch-misses
61.19 -29.2 31.97 ± 2% perf-stat.i.cache-miss-rate%
1.255e+08 +104.5% 2.567e+08 ± 19% perf-stat.i.cache-misses
2.043e+08 +292.8% 8.024e+08 ± 18% perf-stat.i.cache-references
139.74 -93.8% 8.69 ± 27% perf-stat.i.cpi
264.51 +8.5% 286.93 perf-stat.i.cpu-migrations
6664 -47.9% 3471 ± 28% perf-stat.i.cycles-between-cache-misses
6.183e+09 +1555.5% 1.024e+11 ± 18% perf-stat.i.instructions
0.01 ± 23% +808.8% 0.12 ± 18% perf-stat.i.ipc
82374 +3701.0% 3131041 ± 22% perf-stat.i.minor-faults
82375 +3701.0% 3131042 ± 22% perf-stat.i.page-faults
20.32 -91.8% 1.67 ± 70% perf-stat.overall.MPKI
0.92 ± 2% -0.7 0.23 ± 70% perf-stat.overall.branch-miss-rate%
61.44 -39.9 21.51 ± 70% perf-stat.overall.cache-miss-rate%
136.21 -95.6% 6.05 ± 80% perf-stat.overall.cpi
6703 -63.8% 2428 ± 80% perf-stat.overall.cycles-between-cache-misses
23815471 -71.7% 6746184 ± 71% perf-stat.overall.path-length
2.41 ± 6% -93.2% 0.16 ± 78% perf-sched.sch_delay.avg.ms.__cond_resched.__alloc_pages_noprof.alloc_pages_mpol_noprof.folio_alloc_mpol_noprof.vma_alloc_folio_noprof
0.60 ± 21% -100.0% 0.00 perf-sched.sch_delay.avg.ms.__cond_resched.__alloc_pages_noprof.alloc_pages_mpol_noprof.pte_alloc_one.__do_huge_pmd_anonymous_page
0.74 ± 36% -100.0% 0.00 perf-sched.sch_delay.avg.ms.__cond_resched.__anon_vma_prepare.__vmf_anon_prepare.do_huge_pmd_anonymous_page.__handle_mm_fault
2.00 ± 27% -76.6% 0.47 ±149% perf-sched.sch_delay.avg.ms.__cond_resched.__kmalloc_cache_noprof.perf_event_mmap_event.perf_event_mmap.mmap_region
0.35 ± 4% -78.5% 0.07 ± 10% perf-sched.sch_delay.avg.ms.__cond_resched.__wait_for_common.affine_move_task.__set_cpus_allowed_ptr.__sched_setaffinity
1.99 ± 20% -77.4% 0.45 ± 93% perf-sched.sch_delay.avg.ms.__cond_resched.dput.step_into.link_path_walk.part
2.15 ± 14% -98.2% 0.04 ±196% perf-sched.sch_delay.avg.ms.__cond_resched.dput.terminate_walk.path_openat.do_filp_open
0.57 ± 7% -100.0% 0.00 perf-sched.sch_delay.avg.ms.__cond_resched.folio_zero_user.__do_huge_pmd_anonymous_page.__handle_mm_fault.handle_mm_fault
0.84 ± 96% -100.0% 0.00 perf-sched.sch_delay.avg.ms.__cond_resched.kmem_cache_alloc_noprof.__anon_vma_prepare.__vmf_anon_prepare.do_huge_pmd_anonymous_page
1.41 ± 39% -93.0% 0.10 ±223% perf-sched.sch_delay.avg.ms.__cond_resched.kmem_cache_alloc_noprof.getname_flags.part.0
1.52 ± 31% -100.0% 0.00 perf-sched.sch_delay.avg.ms.__cond_resched.kmem_cache_alloc_noprof.security_file_alloc.init_file.alloc_empty_file
1.31 ± 17% -100.0% 0.00 perf-sched.sch_delay.avg.ms.__cond_resched.mutex_lock.fifo_open.do_dentry_open.vfs_open
0.05 ± 4% -76.9% 0.01 ± 27% perf-sched.sch_delay.avg.ms.__cond_resched.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
1.13 ± 33% -99.6% 0.00 ±223% perf-sched.sch_delay.avg.ms.__cond_resched.task_work_run.irqentry_exit_to_user_mode.asm_exc_page_fault.[unknown]
2.20 ± 19% -93.4% 0.14 ±223% perf-sched.sch_delay.avg.ms.__cond_resched.wp_page_copy.__handle_mm_fault.handle_mm_fault.do_user_addr_fault
1.65 ± 30% -70.6% 0.48 ± 65% perf-sched.sch_delay.avg.ms.do_nanosleep.hrtimer_nanosleep.common_nsleep.__x64_sys_clock_nanosleep
0.50 ± 10% -89.7% 0.05 ± 50% perf-sched.sch_delay.avg.ms.do_wait.kernel_wait4.__do_sys_wait4.do_syscall_64
0.03 ± 21% -63.7% 0.01 ± 53% perf-sched.sch_delay.avg.ms.ipmi_thread.kthread.ret_from_fork.ret_from_fork_asm
0.93 ± 11% -54.4% 0.42 ± 35% perf-sched.sch_delay.avg.ms.irqentry_exit_to_user_mode.asm_exc_page_fault.[unknown].[unknown]
2.62 ± 33% -51.4% 1.27 ± 32% perf-sched.sch_delay.avg.ms.irqentry_exit_to_user_mode.asm_sysvec_apic_timer_interrupt.[unknown].[unknown]
0.81 ±133% -94.4% 0.05 ± 13% perf-sched.sch_delay.avg.ms.pipe_read.vfs_read.ksys_read.do_syscall_64
1.11 ± 12% -51.4% 0.54 ± 20% perf-sched.sch_delay.avg.ms.schedule_hrtimeout_range_clock.do_poll.constprop.0.do_sys_poll
0.49 ± 37% -77.7% 0.11 ± 52% perf-sched.sch_delay.avg.ms.schedule_timeout.__wait_for_common.wait_for_completion_state.kernel_clone
0.08 ± 3% -71.3% 0.02 ±107% perf-sched.sch_delay.avg.ms.schedule_timeout.kcompactd.kthread.ret_from_fork
0.05 ± 33% -80.9% 0.01 ± 11% perf-sched.sch_delay.avg.ms.schedule_timeout.rcu_gp_fqs_loop.rcu_gp_kthread.kthread
0.05 ± 8% -81.4% 0.01 ± 41% perf-sched.sch_delay.avg.ms.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
2.15 ± 13% -77.3% 0.49 ± 28% perf-sched.sch_delay.avg.ms.wait_for_partner.fifo_open.do_dentry_open.vfs_open
4.04 -57.7% 1.71 ± 97% perf-sched.sch_delay.max.ms.__cond_resched.__alloc_pages_noprof.alloc_pages_mpol_noprof.folio_alloc_mpol_noprof.vma_alloc_folio_noprof
4.09 ± 2% -100.0% 0.00 perf-sched.sch_delay.max.ms.__cond_resched.__alloc_pages_noprof.alloc_pages_mpol_noprof.pte_alloc_one.__do_huge_pmd_anonymous_page
1.19 ± 74% +593.4% 8.27 ± 62% perf-sched.sch_delay.max.ms.__cond_resched.__alloc_pages_noprof.alloc_pages_mpol_noprof.pte_alloc_one.__pte_alloc
0.84 ±173% +695.6% 6.69 ± 78% perf-sched.sch_delay.max.ms.__cond_resched.__anon_vma_prepare.__vmf_anon_prepare.do_anonymous_page.__handle_mm_fault
4.00 -100.0% 0.00 perf-sched.sch_delay.max.ms.__cond_resched.__anon_vma_prepare.__vmf_anon_prepare.do_huge_pmd_anonymous_page.__handle_mm_fault
2.71 ± 21% -82.7% 0.47 ±149% perf-sched.sch_delay.max.ms.__cond_resched.__kmalloc_cache_noprof.perf_event_mmap_event.perf_event_mmap.mmap_region
7.66 ± 93% -91.2% 0.67 ±223% perf-sched.sch_delay.max.ms.__cond_resched.__kmalloc_cache_noprof.vmstat_start.seq_read_iter.proc_reg_read_iter
0.01 ±223% +9242.3% 0.81 ±179% perf-sched.sch_delay.max.ms.__cond_resched.__kmalloc_node_noprof.alloc_slab_obj_exts.allocate_slab.___slab_alloc
8.12 ±103% +186.4% 23.26 ± 73% perf-sched.sch_delay.max.ms.__cond_resched.__tlb_batch_free_encoded_pages.tlb_finish_mmu.vms_clear_ptes.vms_complete_munmap_vmas
2.34 ± 11% -75.7% 0.57 ± 75% perf-sched.sch_delay.max.ms.__cond_resched.dput.step_into.link_path_walk.part
2.77 ± 21% -96.7% 0.09 ±212% perf-sched.sch_delay.max.ms.__cond_resched.dput.terminate_walk.path_openat.do_filp_open
36.48 ± 69% -100.0% 0.00 perf-sched.sch_delay.max.ms.__cond_resched.folio_zero_user.__do_huge_pmd_anonymous_page.__handle_mm_fault.handle_mm_fault
0.37 ±223% +998.1% 4.04 ± 35% perf-sched.sch_delay.max.ms.__cond_resched.kmem_cache_alloc_noprof.__anon_vma_prepare.__vmf_anon_prepare.do_anonymous_page
2.26 ± 78% -100.0% 0.00 perf-sched.sch_delay.max.ms.__cond_resched.kmem_cache_alloc_noprof.__anon_vma_prepare.__vmf_anon_prepare.do_huge_pmd_anonymous_page
1.83 ± 40% -94.6% 0.10 ±223% perf-sched.sch_delay.max.ms.__cond_resched.kmem_cache_alloc_noprof.getname_flags.part.0
1.75 ± 25% -100.0% 0.00 perf-sched.sch_delay.max.ms.__cond_resched.kmem_cache_alloc_noprof.security_file_alloc.init_file.alloc_empty_file
1.51 ± 30% -100.0% 0.00 perf-sched.sch_delay.max.ms.__cond_resched.mutex_lock.fifo_open.do_dentry_open.vfs_open
0.13 ± 50% -86.1% 0.02 ± 46% perf-sched.sch_delay.max.ms.__cond_resched.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
4.17 ± 5% -99.8% 0.01 ±223% perf-sched.sch_delay.max.ms.__cond_resched.task_work_run.irqentry_exit_to_user_mode.asm_exc_page_fault.[unknown]
4.02 +510.8% 24.57 ± 38% perf-sched.sch_delay.max.ms.__cond_resched.unmap_vmas.vms_clear_ptes.vms_complete_munmap_vmas.do_vmi_align_munmap
3.99 -92.2% 0.31 ±223% perf-sched.sch_delay.max.ms.__cond_resched.wp_page_copy.__handle_mm_fault.handle_mm_fault.do_user_addr_fault
4.11 ± 5% +33.0% 5.47 ± 27% perf-sched.sch_delay.max.ms.__cond_resched.zap_pmd_range.isra.0.unmap_page_range
4.54 ± 12% -54.1% 2.08 ± 65% perf-sched.sch_delay.max.ms.do_nanosleep.hrtimer_nanosleep.common_nsleep.__x64_sys_clock_nanosleep
4.00 ± 25% -70.1% 1.20 ± 30% perf-sched.sch_delay.max.ms.do_wait.kernel_wait4.__do_sys_wait4.do_syscall_64
0.03 ± 17% -60.8% 0.01 ± 64% perf-sched.sch_delay.max.ms.ipmi_thread.kthread.ret_from_fork.ret_from_fork_asm
4.08 ± 2% -61.1% 1.59 ±108% perf-sched.sch_delay.max.ms.irqentry_exit_to_user_mode.asm_exc_page_fault.[unknown]
389.98 ±121% -96.7% 13.00 ± 33% perf-sched.sch_delay.max.ms.pipe_read.vfs_read.ksys_read.do_syscall_64
9.71 ±119% -78.8% 2.06 ± 58% perf-sched.sch_delay.max.ms.schedule_timeout.__wait_for_common.wait_for_completion_state.kernel_clone
8.60 ±210% -99.1% 0.08 ± 36% perf-sched.sch_delay.max.ms.schedule_timeout.rcu_gp_fqs_loop.rcu_gp_kthread.kthread
4.05 ± 2% -34.0% 2.67 ± 42% perf-sched.sch_delay.max.ms.wait_for_partner.fifo_open.do_dentry_open.vfs_open
8.02 ± 5% -55.7% 3.55 ± 11% perf-sched.sch_delay.max.ms.worker_thread.kthread.ret_from_fork.ret_from_fork_asm
0.47 ± 5% -13.9% 0.40 ± 13% perf-sched.total_sch_delay.average.ms
396.80 ± 9% -100.0% 0.00 perf-sched.wait_and_delay.avg.ms.__cond_resched.__kmalloc_cache_noprof.vmstat_start.seq_read_iter.proc_reg_read_iter
0.17 ±223% +596.6% 1.17 ± 14% perf-sched.wait_and_delay.avg.ms.__cond_resched.__tlb_batch_free_encoded_pages.tlb_finish_mmu.vms_clear_ptes.vms_complete_munmap_vmas
1.15 ± 7% -100.0% 0.00 perf-sched.wait_and_delay.avg.ms.__cond_resched.folio_zero_user.__do_huge_pmd_anonymous_page.__handle_mm_fault.handle_mm_fault
417.15 ± 5% +75.4% 731.48 ± 12% perf-sched.wait_and_delay.avg.ms.do_nanosleep.hrtimer_nanosleep.common_nsleep.__x64_sys_clock_nanosleep
11.50 ± 25% -100.0% 0.00 perf-sched.wait_and_delay.count.__cond_resched.__kmalloc_cache_noprof.vmstat_start.seq_read_iter.proc_reg_read_iter
26.17 ±223% +15367.5% 4047 ± 49% perf-sched.wait_and_delay.count.__cond_resched.__tlb_batch_free_encoded_pages.tlb_finish_mmu.vms_clear_ptes.vms_complete_munmap_vmas
15379 ± 15% -100.0% 0.00 perf-sched.wait_and_delay.count.__cond_resched.folio_zero_user.__do_huge_pmd_anonymous_page.__handle_mm_fault.handle_mm_fault
30.33 ±223% +836.8% 284.17 ± 53% perf-sched.wait_and_delay.count.irqentry_exit_to_user_mode.asm_exc_page_fault.[unknown].[unknown]
1000 -100.0% 0.00 perf-sched.wait_and_delay.max.ms.__cond_resched.__kmalloc_cache_noprof.vmstat_start.seq_read_iter.proc_reg_read_iter
1.34 ±223% +3359.9% 46.51 ± 73% perf-sched.wait_and_delay.max.ms.__cond_resched.__tlb_batch_free_encoded_pages.tlb_finish_mmu.vms_clear_ptes.vms_complete_munmap_vmas
72.95 ± 69% -100.0% 0.00 perf-sched.wait_and_delay.max.ms.__cond_resched.folio_zero_user.__do_huge_pmd_anonymous_page.__handle_mm_fault.handle_mm_fault
13.11 ±122% -61.3% 5.08 perf-sched.wait_and_delay.max.ms.schedule_timeout.rcu_gp_fqs_loop.rcu_gp_kthread.kthread
2.41 ± 6% -93.5% 0.16 ± 76% perf-sched.wait_time.avg.ms.__cond_resched.__alloc_pages_noprof.alloc_pages_mpol_noprof.folio_alloc_mpol_noprof.vma_alloc_folio_noprof
0.60 ± 21% -100.0% 0.00 perf-sched.wait_time.avg.ms.__cond_resched.__alloc_pages_noprof.alloc_pages_mpol_noprof.pte_alloc_one.__do_huge_pmd_anonymous_page
0.74 ± 36% -100.0% 0.00 perf-sched.wait_time.avg.ms.__cond_resched.__anon_vma_prepare.__vmf_anon_prepare.do_huge_pmd_anonymous_page.__handle_mm_fault
2.00 ± 27% -76.6% 0.47 ±149% perf-sched.wait_time.avg.ms.__cond_resched.__kmalloc_cache_noprof.perf_event_mmap_event.perf_event_mmap.mmap_region
394.43 ± 10% -99.8% 0.67 ±223% perf-sched.wait_time.avg.ms.__cond_resched.__kmalloc_cache_noprof.vmstat_start.seq_read_iter.proc_reg_read_iter
1.99 ± 20% -77.8% 0.44 ± 95% perf-sched.wait_time.avg.ms.__cond_resched.dput.step_into.link_path_walk.part
2.15 ± 14% -98.4% 0.03 ±193% perf-sched.wait_time.avg.ms.__cond_resched.dput.terminate_walk.path_openat.do_filp_open
0.57 ± 7% -100.0% 0.00 perf-sched.wait_time.avg.ms.__cond_resched.folio_zero_user.__do_huge_pmd_anonymous_page.__handle_mm_fault.handle_mm_fault
0.84 ± 96% -100.0% 0.00 perf-sched.wait_time.avg.ms.__cond_resched.kmem_cache_alloc_noprof.__anon_vma_prepare.__vmf_anon_prepare.do_huge_pmd_anonymous_page
1.41 ± 40% -93.0% 0.10 ±223% perf-sched.wait_time.avg.ms.__cond_resched.kmem_cache_alloc_noprof.getname_flags.part.0
84.95 ±219% -100.0% 0.00 perf-sched.wait_time.avg.ms.__cond_resched.kmem_cache_alloc_noprof.security_file_alloc.init_file.alloc_empty_file
1.31 ± 17% -100.0% 0.00 perf-sched.wait_time.avg.ms.__cond_resched.mutex_lock.fifo_open.do_dentry_open.vfs_open
1.13 ± 33% -99.6% 0.00 ±223% perf-sched.wait_time.avg.ms.__cond_resched.task_work_run.irqentry_exit_to_user_mode.asm_exc_page_fault.[unknown]
2.11 ± 23% -93.1% 0.14 ±223% perf-sched.wait_time.avg.ms.__cond_resched.wp_page_copy.__handle_mm_fault.handle_mm_fault.do_user_addr_fault
415.50 ± 5% +75.9% 731.00 ± 13% perf-sched.wait_time.avg.ms.do_nanosleep.hrtimer_nanosleep.common_nsleep.__x64_sys_clock_nanosleep
5.81 ± 41% -74.2% 1.50 ± 9% perf-sched.wait_time.avg.ms.do_wait.kernel_wait4.__do_sys_wait4.do_syscall_64
0.93 ± 12% -54.5% 0.42 ± 35% perf-sched.wait_time.avg.ms.irqentry_exit_to_user_mode.asm_exc_page_fault.[unknown].[unknown]
9.89 ±101% -87.1% 1.27 ± 32% perf-sched.wait_time.avg.ms.irqentry_exit_to_user_mode.asm_sysvec_apic_timer_interrupt.[unknown].[unknown]
13.22 ± 5% -77.3% 3.01 ± 14% perf-sched.wait_time.avg.ms.schedule_timeout.__wait_for_common.wait_for_completion_state.kernel_clone
2.13 ± 15% -77.6% 0.48 ± 28% perf-sched.wait_time.avg.ms.wait_for_partner.fifo_open.do_dentry_open.vfs_open
4.04 -57.7% 1.71 ± 97% perf-sched.wait_time.max.ms.__cond_resched.__alloc_pages_noprof.alloc_pages_mpol_noprof.folio_alloc_mpol_noprof.vma_alloc_folio_noprof
4.09 ± 2% -100.0% 0.00 perf-sched.wait_time.max.ms.__cond_resched.__alloc_pages_noprof.alloc_pages_mpol_noprof.pte_alloc_one.__do_huge_pmd_anonymous_page
1.19 ± 74% +593.4% 8.27 ± 62% perf-sched.wait_time.max.ms.__cond_resched.__alloc_pages_noprof.alloc_pages_mpol_noprof.pte_alloc_one.__pte_alloc
0.84 ±173% +695.6% 6.69 ± 78% perf-sched.wait_time.max.ms.__cond_resched.__anon_vma_prepare.__vmf_anon_prepare.do_anonymous_page.__handle_mm_fault
4.00 -100.0% 0.00 perf-sched.wait_time.max.ms.__cond_resched.__anon_vma_prepare.__vmf_anon_prepare.do_huge_pmd_anonymous_page.__handle_mm_fault
2.71 ± 21% -82.7% 0.47 ±149% perf-sched.wait_time.max.ms.__cond_resched.__kmalloc_cache_noprof.perf_event_mmap_event.perf_event_mmap.mmap_region
1000 -99.9% 0.67 ±223% perf-sched.wait_time.max.ms.__cond_resched.__kmalloc_cache_noprof.vmstat_start.seq_read_iter.proc_reg_read_iter
8.12 ±103% +186.4% 23.26 ± 73% perf-sched.wait_time.max.ms.__cond_resched.__tlb_batch_free_encoded_pages.tlb_finish_mmu.vms_clear_ptes.vms_complete_munmap_vmas
2.34 ± 11% -75.7% 0.57 ± 75% perf-sched.wait_time.max.ms.__cond_resched.dput.step_into.link_path_walk.part
2.77 ± 21% -96.7% 0.09 ±212% perf-sched.wait_time.max.ms.__cond_resched.dput.terminate_walk.path_openat.do_filp_open
36.48 ± 69% -100.0% 0.00 perf-sched.wait_time.max.ms.__cond_resched.folio_zero_user.__do_huge_pmd_anonymous_page.__handle_mm_fault.handle_mm_fault
0.37 ±223% +998.1% 4.04 ± 35% perf-sched.wait_time.max.ms.__cond_resched.kmem_cache_alloc_noprof.__anon_vma_prepare.__vmf_anon_prepare.do_anonymous_page
2.26 ± 78% -100.0% 0.00 perf-sched.wait_time.max.ms.__cond_resched.kmem_cache_alloc_noprof.__anon_vma_prepare.__vmf_anon_prepare.do_huge_pmd_anonymous_page
1.83 ± 40% -94.6% 0.10 ±223% perf-sched.wait_time.max.ms.__cond_resched.kmem_cache_alloc_noprof.getname_flags.part.0
168.46 ±221% -100.0% 0.00 perf-sched.wait_time.max.ms.__cond_resched.kmem_cache_alloc_noprof.security_file_alloc.init_file.alloc_empty_file
1.51 ± 30% -100.0% 0.00 perf-sched.wait_time.max.ms.__cond_resched.mutex_lock.fifo_open.do_dentry_open.vfs_open
4.17 ± 5% -99.8% 0.01 ±223% perf-sched.wait_time.max.ms.__cond_resched.task_work_run.irqentry_exit_to_user_mode.asm_exc_page_fault.[unknown]
4.02 +510.8% 24.57 ± 38% perf-sched.wait_time.max.ms.__cond_resched.unmap_vmas.vms_clear_ptes.vms_complete_munmap_vmas.do_vmi_align_munmap
3.99 -92.2% 0.31 ±223% perf-sched.wait_time.max.ms.__cond_resched.wp_page_copy.__handle_mm_fault.handle_mm_fault.do_user_addr_fault
4.11 ± 5% +33.0% 5.47 ± 27% perf-sched.wait_time.max.ms.__cond_resched.zap_pmd_range.isra.0.unmap_page_range
250.47 ±134% -89.0% 27.62 ± 59% perf-sched.wait_time.max.ms.do_wait.kernel_wait4.__do_sys_wait4.do_syscall_64
4.08 ± 2% -61.1% 1.59 ±108% perf-sched.wait_time.max.ms.irqentry_exit_to_user_mode.asm_exc_page_fault.[unknown]
499.41 +278.7% 1891 ± 64% perf-sched.wait_time.max.ms.schedule_hrtimeout_range_clock.ep_poll.do_epoll_wait.__x64_sys_epoll_wait
54.24 ± 46% -59.6% 21.93 ± 68% perf-sched.wait_time.max.ms.schedule_timeout.__wait_for_common.wait_for_completion_state.kernel_clone
4.05 ± 2% -34.0% 2.67 ± 42% perf-sched.wait_time.max.ms.wait_for_partner.fifo_open.do_dentry_open.vfs_open
96.58 -96.6 0.00 perf-profile.calltrace.cycles-pp.folio_zero_user.__do_huge_pmd_anonymous_page.__handle_mm_fault.handle_mm_fault.do_user_addr_fault
96.37 -96.4 0.00 perf-profile.calltrace.cycles-pp.__do_huge_pmd_anonymous_page.__handle_mm_fault.handle_mm_fault.do_user_addr_fault.exc_page_fault
96.69 -90.4 6.27 ± 11% perf-profile.calltrace.cycles-pp.__handle_mm_fault.handle_mm_fault.do_user_addr_fault.exc_page_fault.asm_exc_page_fault
96.73 -90.3 6.45 ± 11% perf-profile.calltrace.cycles-pp.handle_mm_fault.do_user_addr_fault.exc_page_fault.asm_exc_page_fault
97.00 -90.2 6.79 ± 10% perf-profile.calltrace.cycles-pp.asm_exc_page_fault
96.77 -90.1 6.68 ± 10% perf-profile.calltrace.cycles-pp.do_user_addr_fault.exc_page_fault.asm_exc_page_fault
96.77 -90.1 6.69 ± 10% perf-profile.calltrace.cycles-pp.exc_page_fault.asm_exc_page_fault
87.45 -87.4 0.00 perf-profile.calltrace.cycles-pp.clear_page_erms.folio_zero_user.__do_huge_pmd_anonymous_page.__handle_mm_fault.handle_mm_fault
0.00 +0.6 0.61 ± 11% perf-profile.calltrace.cycles-pp.lru_add.folio_batch_move_lru.lru_add_drain_cpu.lru_add_drain.vms_clear_ptes
0.00 +0.6 0.62 ± 12% perf-profile.calltrace.cycles-pp.__memcg_slab_post_alloc_hook.kmem_cache_alloc_noprof.__anon_vma_prepare.__vmf_anon_prepare.do_anonymous_page
0.00 +0.7 0.67 ± 10% perf-profile.calltrace.cycles-pp.kmem_cache_alloc_noprof.__anon_vma_prepare.__vmf_anon_prepare.do_anonymous_page.__handle_mm_fault
0.00 +0.7 0.69 ± 27% perf-profile.calltrace.cycles-pp.get_mem_cgroup_from_mm.__mem_cgroup_charge.alloc_anon_folio.do_anonymous_page.__handle_mm_fault
0.00 +0.8 0.82 ± 10% perf-profile.calltrace.cycles-pp.__lruvec_stat_mod_folio.folio_remove_rmap_ptes.zap_present_ptes.zap_pte_range.zap_pmd_range
0.00 +0.8 0.84 ± 13% perf-profile.calltrace.cycles-pp.__lruvec_stat_mod_folio.pte_alloc_one.__pte_alloc.do_anonymous_page.__handle_mm_fault
0.00 +0.8 0.84 ± 10% perf-profile.calltrace.cycles-pp.folio_remove_rmap_ptes.zap_present_ptes.zap_pte_range.zap_pmd_range.unmap_page_range
0.00 +0.8 0.85 ± 8% perf-profile.calltrace.cycles-pp.__lruvec_stat_mod_folio.___pte_free_tlb.free_pud_range.free_p4d_range.free_pgd_range
0.00 +0.9 0.86 ± 9% perf-profile.calltrace.cycles-pp.__anon_vma_prepare.__vmf_anon_prepare.do_anonymous_page.__handle_mm_fault.handle_mm_fault
0.00 +0.9 0.87 ± 8% perf-profile.calltrace.cycles-pp.___pte_free_tlb.free_pud_range.free_p4d_range.free_pgd_range.free_pgtables
0.00 +0.9 0.88 ± 9% perf-profile.calltrace.cycles-pp.__lruvec_stat_mod_folio.__folio_mod_stat.folio_add_new_anon_rmap.do_anonymous_page.__handle_mm_fault
0.00 +0.9 0.88 ± 10% perf-profile.calltrace.cycles-pp.__folio_mod_stat.folio_add_new_anon_rmap.do_anonymous_page.__handle_mm_fault.handle_mm_fault
0.00 +0.9 0.90 ± 9% perf-profile.calltrace.cycles-pp.zap_present_ptes.zap_pte_range.zap_pmd_range.unmap_page_range.unmap_vmas
0.00 +0.9 0.90 ± 9% perf-profile.calltrace.cycles-pp.__vmf_anon_prepare.do_anonymous_page.__handle_mm_fault.handle_mm_fault.do_user_addr_fault
0.00 +0.9 0.91 ± 10% perf-profile.calltrace.cycles-pp.folio_add_new_anon_rmap.do_anonymous_page.__handle_mm_fault.handle_mm_fault.do_user_addr_fault
0.00 +1.0 1.03 ± 9% perf-profile.calltrace.cycles-pp.free_pud_range.free_p4d_range.free_pgd_range.free_pgtables.vms_clear_ptes
0.00 +1.0 1.05 ± 9% perf-profile.calltrace.cycles-pp.free_p4d_range.free_pgd_range.free_pgtables.vms_clear_ptes.vms_complete_munmap_vmas
0.00 +1.1 1.07 ± 9% perf-profile.calltrace.cycles-pp.free_pgd_range.free_pgtables.vms_clear_ptes.vms_complete_munmap_vmas.do_vmi_align_munmap
0.00 +1.1 1.11 ± 75% perf-profile.calltrace.cycles-pp.uncharge_folio.__mem_cgroup_uncharge_folios.folios_put_refs.free_pages_and_swap_cache.__tlb_batch_free_encoded_pages
0.00 +1.3 1.29 ± 46% perf-profile.calltrace.cycles-pp.__mem_cgroup_charge.alloc_anon_folio.do_anonymous_page.__handle_mm_fault.handle_mm_fault
0.00 +1.4 1.38 ± 9% perf-profile.calltrace.cycles-pp.free_pgtables.vms_clear_ptes.vms_complete_munmap_vmas.do_vmi_align_munmap.do_vmi_munmap
0.00 +1.5 1.52 ± 28% perf-profile.calltrace.cycles-pp.__memcg_kmem_charge_page.__alloc_pages_noprof.alloc_pages_mpol_noprof.pte_alloc_one.__pte_alloc
0.00 +1.5 1.54 ± 34% perf-profile.calltrace.cycles-pp.alloc_anon_folio.do_anonymous_page.__handle_mm_fault.handle_mm_fault.do_user_addr_fault
0.00 +1.7 1.73 ± 22% perf-profile.calltrace.cycles-pp.__alloc_pages_noprof.alloc_pages_mpol_noprof.pte_alloc_one.__pte_alloc.do_anonymous_page
0.00 +1.7 1.74 ± 18% perf-profile.calltrace.cycles-pp.mmap_region.do_mmap.vm_mmap_pgoff.do_syscall_64.entry_SYSCALL_64_after_hwframe
0.00 +1.8 1.75 ± 21% perf-profile.calltrace.cycles-pp.alloc_pages_mpol_noprof.pte_alloc_one.__pte_alloc.do_anonymous_page.__handle_mm_fault
0.00 +2.0 2.02 ± 14% perf-profile.calltrace.cycles-pp.zap_pte_range.zap_pmd_range.unmap_page_range.unmap_vmas.vms_clear_ptes
0.00 +2.1 2.10 ± 19% perf-profile.calltrace.cycles-pp.do_mmap.vm_mmap_pgoff.do_syscall_64.entry_SYSCALL_64_after_hwframe.__mmap
0.00 +2.2 2.22 ± 19% perf-profile.calltrace.cycles-pp.vm_mmap_pgoff.do_syscall_64.entry_SYSCALL_64_after_hwframe.__mmap
0.00 +2.3 2.28 ± 19% perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.__mmap
0.00 +2.3 2.29 ± 19% perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.__mmap
0.00 +2.5 2.46 ± 19% perf-profile.calltrace.cycles-pp.__mmap
0.00 +2.5 2.47 ± 16% perf-profile.calltrace.cycles-pp.zap_pmd_range.unmap_page_range.unmap_vmas.vms_clear_ptes.vms_complete_munmap_vmas
0.00 +2.5 2.54 ± 16% perf-profile.calltrace.cycles-pp.unmap_page_range.unmap_vmas.vms_clear_ptes.vms_complete_munmap_vmas.do_vmi_align_munmap
0.00 +2.6 2.58 ± 16% perf-profile.calltrace.cycles-pp.unmap_vmas.vms_clear_ptes.vms_complete_munmap_vmas.do_vmi_align_munmap.do_vmi_munmap
0.00 +2.6 2.63 ± 12% perf-profile.calltrace.cycles-pp.pte_alloc_one.__pte_alloc.do_anonymous_page.__handle_mm_fault.handle_mm_fault
0.00 +2.7 2.69 ± 12% perf-profile.calltrace.cycles-pp.__pte_alloc.do_anonymous_page.__handle_mm_fault.handle_mm_fault.do_user_addr_fault
0.00 +3.2 3.21 ± 38% perf-profile.calltrace.cycles-pp.page_counter_cancel.page_counter_uncharge.uncharge_batch.__mem_cgroup_uncharge_folios.folios_put_refs
0.00 +4.0 4.02 ± 37% perf-profile.calltrace.cycles-pp.page_counter_uncharge.uncharge_batch.__mem_cgroup_uncharge_folios.folios_put_refs.free_pages_and_swap_cache
0.00 +5.0 5.04 ± 23% perf-profile.calltrace.cycles-pp.uncharge_batch.__mem_cgroup_uncharge_folios.folios_put_refs.free_pages_and_swap_cache.__tlb_batch_free_encoded_pages
0.00 +6.2 6.16 ± 7% perf-profile.calltrace.cycles-pp.__mem_cgroup_uncharge_folios.folios_put_refs.free_pages_and_swap_cache.__tlb_batch_free_encoded_pages.tlb_finish_mmu
0.00 +6.2 6.16 ± 12% perf-profile.calltrace.cycles-pp.do_anonymous_page.__handle_mm_fault.handle_mm_fault.do_user_addr_fault.exc_page_fault
0.00 +36.4 36.44 ± 2% perf-profile.calltrace.cycles-pp.native_queued_spin_lock_slowpath._raw_spin_lock_irqsave.folio_lruvec_lock_irqsave.folio_batch_move_lru.lru_add_drain_cpu
0.00 +36.6 36.57 ± 2% perf-profile.calltrace.cycles-pp._raw_spin_lock_irqsave.folio_lruvec_lock_irqsave.folio_batch_move_lru.lru_add_drain_cpu.lru_add_drain
0.00 +36.8 36.80 ± 3% perf-profile.calltrace.cycles-pp.native_queued_spin_lock_slowpath._raw_spin_lock_irqsave.folio_lruvec_lock_irqsave.__page_cache_release.folios_put_refs
0.00 +36.9 36.93 ± 3% perf-profile.calltrace.cycles-pp._raw_spin_lock_irqsave.folio_lruvec_lock_irqsave.__page_cache_release.folios_put_refs.free_pages_and_swap_cache
0.00 +37.0 37.00 perf-profile.calltrace.cycles-pp.folio_lruvec_lock_irqsave.folio_batch_move_lru.lru_add_drain_cpu.lru_add_drain.vms_clear_ptes
0.00 +37.3 37.33 ± 3% perf-profile.calltrace.cycles-pp.folio_lruvec_lock_irqsave.__page_cache_release.folios_put_refs.free_pages_and_swap_cache.__tlb_batch_free_encoded_pages
0.00 +37.9 37.87 ± 3% perf-profile.calltrace.cycles-pp.__page_cache_release.folios_put_refs.free_pages_and_swap_cache.__tlb_batch_free_encoded_pages.tlb_finish_mmu
0.00 +38.4 38.42 perf-profile.calltrace.cycles-pp.folio_batch_move_lru.lru_add_drain_cpu.lru_add_drain.vms_clear_ptes.vms_complete_munmap_vmas
0.00 +38.4 38.45 perf-profile.calltrace.cycles-pp.lru_add_drain_cpu.lru_add_drain.vms_clear_ptes.vms_complete_munmap_vmas.do_vmi_align_munmap
0.00 +38.5 38.45 perf-profile.calltrace.cycles-pp.lru_add_drain.vms_clear_ptes.vms_complete_munmap_vmas.do_vmi_align_munmap.do_vmi_munmap
0.78 ± 2% +44.1 44.87 perf-profile.calltrace.cycles-pp.folios_put_refs.free_pages_and_swap_cache.__tlb_batch_free_encoded_pages.tlb_finish_mmu.vms_clear_ptes
0.79 ± 2% +44.2 44.99 perf-profile.calltrace.cycles-pp.free_pages_and_swap_cache.__tlb_batch_free_encoded_pages.tlb_finish_mmu.vms_clear_ptes.vms_complete_munmap_vmas
0.79 ± 2% +44.2 45.00 perf-profile.calltrace.cycles-pp.__tlb_batch_free_encoded_pages.tlb_finish_mmu.vms_clear_ptes.vms_complete_munmap_vmas.do_vmi_align_munmap
0.86 ± 2% +44.4 45.24 perf-profile.calltrace.cycles-pp.tlb_finish_mmu.vms_clear_ptes.vms_complete_munmap_vmas.do_vmi_align_munmap.do_vmi_munmap
1.23 ± 2% +86.5 87.70 perf-profile.calltrace.cycles-pp.vms_clear_ptes.vms_complete_munmap_vmas.do_vmi_align_munmap.do_vmi_munmap.__vm_munmap
1.25 +86.6 87.86 perf-profile.calltrace.cycles-pp.vms_complete_munmap_vmas.do_vmi_align_munmap.do_vmi_munmap.__vm_munmap.__x64_sys_munmap
1.44 +87.1 88.55 perf-profile.calltrace.cycles-pp.do_vmi_align_munmap.do_vmi_munmap.__vm_munmap.__x64_sys_munmap.do_syscall_64
1.48 +87.1 88.60 perf-profile.calltrace.cycles-pp.do_vmi_munmap.__vm_munmap.__x64_sys_munmap.do_syscall_64.entry_SYSCALL_64_after_hwframe
1.49 +87.2 88.66 perf-profile.calltrace.cycles-pp.__vm_munmap.__x64_sys_munmap.do_syscall_64.entry_SYSCALL_64_after_hwframe.__munmap
1.50 +87.2 88.66 perf-profile.calltrace.cycles-pp.__x64_sys_munmap.do_syscall_64.entry_SYSCALL_64_after_hwframe.__munmap
1.52 +87.2 88.72 perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.__munmap
1.52 +87.2 88.73 perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.__munmap
1.53 +87.4 88.90 perf-profile.calltrace.cycles-pp.__munmap
96.37 -96.4 0.00 perf-profile.children.cycles-pp.__do_huge_pmd_anonymous_page
95.62 -95.6 0.00 perf-profile.children.cycles-pp.folio_zero_user
96.71 -90.4 6.28 ± 11% perf-profile.children.cycles-pp.__handle_mm_fault
96.75 -90.3 6.46 ± 11% perf-profile.children.cycles-pp.handle_mm_fault
97.02 -90.2 6.83 ± 10% perf-profile.children.cycles-pp.asm_exc_page_fault
96.79 -90.1 6.68 ± 10% perf-profile.children.cycles-pp.do_user_addr_fault
96.80 -90.1 6.70 ± 10% perf-profile.children.cycles-pp.exc_page_fault
89.46 -89.3 0.20 ± 22% perf-profile.children.cycles-pp.clear_page_erms
4.15 -3.8 0.35 ± 23% perf-profile.children.cycles-pp.__cond_resched
0.81 ± 3% -0.6 0.23 ± 7% perf-profile.children.cycles-pp.__sysvec_apic_timer_interrupt
0.80 ± 3% -0.6 0.22 ± 7% perf-profile.children.cycles-pp.hrtimer_interrupt
0.73 ± 3% -0.5 0.18 ± 8% perf-profile.children.cycles-pp.__hrtimer_run_queues
0.60 ± 4% -0.5 0.14 ± 5% perf-profile.children.cycles-pp.tick_nohz_handler
0.83 ± 16% -0.4 0.39 ± 3% perf-profile.children.cycles-pp.__cmd_record
0.53 ± 3% -0.4 0.12 ± 6% perf-profile.children.cycles-pp.update_process_times
0.52 -0.4 0.12 ± 28% perf-profile.children.cycles-pp.free_unref_folios
0.44 ± 3% -0.3 0.12 ± 6% perf-profile.children.cycles-pp.cmd_record
0.44 ± 3% -0.3 0.12 ± 6% perf-profile.children.cycles-pp.record__mmap_read_evlist
0.44 ± 3% -0.3 0.12 ± 6% perf-profile.children.cycles-pp.handle_internal_command
0.44 ± 3% -0.3 0.12 ± 6% perf-profile.children.cycles-pp.main
0.44 ± 3% -0.3 0.12 ± 6% perf-profile.children.cycles-pp.run_builtin
0.43 ± 3% -0.3 0.11 ± 6% perf-profile.children.cycles-pp.perf_mmap__push
0.34 ± 3% -0.3 0.07 ± 17% perf-profile.children.cycles-pp.record__pushfn
0.34 ± 4% -0.3 0.07 ± 17% perf-profile.children.cycles-pp.write
0.34 ± 4% -0.3 0.07 ± 17% perf-profile.children.cycles-pp.writen
0.32 ± 3% -0.3 0.07 ± 15% perf-profile.children.cycles-pp.ksys_write
0.32 ± 3% -0.2 0.07 ± 16% perf-profile.children.cycles-pp.vfs_write
0.30 ± 3% -0.2 0.07 ± 16% perf-profile.children.cycles-pp.shmem_file_write_iter
0.30 ± 3% -0.2 0.06 ± 19% perf-profile.children.cycles-pp.generic_perform_write
0.29 ± 4% -0.2 0.09 ± 4% perf-profile.children.cycles-pp.sched_tick
0.38 ± 33% -0.1 0.24 ± 8% perf-profile.children.cycles-pp.process_simple
0.20 ± 5% -0.1 0.06 ± 8% perf-profile.children.cycles-pp.task_tick_fair
0.38 ± 33% -0.1 0.24 ± 8% perf-profile.children.cycles-pp.ordered_events__queue
0.38 ± 33% -0.1 0.24 ± 8% perf-profile.children.cycles-pp.queue_event
0.00 +0.1 0.10 ± 22% perf-profile.children.cycles-pp.sync_regs
0.00 +0.1 0.10 ± 59% perf-profile.children.cycles-pp.__count_memcg_events
0.00 +0.1 0.10 ± 24% perf-profile.children.cycles-pp.___perf_sw_event
0.00 +0.1 0.10 ± 26% perf-profile.children.cycles-pp.perf_event_mmap_output
0.00 +0.1 0.11 ± 24% perf-profile.children.cycles-pp.native_flush_tlb_local
0.00 +0.1 0.11 ± 8% perf-profile.children.cycles-pp.__put_anon_vma
0.00 +0.1 0.11 ± 17% perf-profile.children.cycles-pp.rmqueue
0.00 +0.1 0.12 ± 26% perf-profile.children.cycles-pp.find_mergeable_anon_vma
0.00 +0.1 0.12 ± 23% perf-profile.children.cycles-pp.flush_tlb_func
0.00 +0.1 0.13 ± 28% perf-profile.children.cycles-pp.__call_rcu_common
0.00 +0.1 0.13 ± 24% perf-profile.children.cycles-pp.__perf_sw_event
0.00 +0.1 0.13 ± 38% perf-profile.children.cycles-pp.obj_cgroup_uncharge_pages
0.05 +0.1 0.18 ± 23% perf-profile.children.cycles-pp.flush_tlb_mm_range
0.00 +0.1 0.14 ± 27% perf-profile.children.cycles-pp.___slab_alloc
0.00 +0.1 0.14 ± 23% perf-profile.children.cycles-pp.mas_preallocate
0.00 +0.1 0.15 ± 18% perf-profile.children.cycles-pp.mas_empty_area_rev
0.00 +0.2 0.16 ± 22% perf-profile.children.cycles-pp.clear_bhb_loop
0.02 ±141% +0.2 0.17 ± 25% perf-profile.children.cycles-pp.perf_iterate_sb
0.00 +0.2 0.16 ± 26% perf-profile.children.cycles-pp.mas_alloc_nodes
0.07 ± 5% +0.2 0.24 ± 25% perf-profile.children.cycles-pp.vms_gather_munmap_vmas
0.00 +0.2 0.17 ± 20% perf-profile.children.cycles-pp.obj_cgroup_charge
0.00 +0.2 0.18 ± 25% perf-profile.children.cycles-pp.mas_walk
0.05 +0.2 0.24 ± 24% perf-profile.children.cycles-pp.mas_find
0.06 ± 6% +0.2 0.27 ± 11% perf-profile.children.cycles-pp.unlink_anon_vmas
0.05 ± 7% +0.2 0.26 ± 7% perf-profile.children.cycles-pp.free_unref_page
0.07 ± 11% +0.2 0.28 ± 21% perf-profile.children.cycles-pp.perf_event_mmap_event
0.07 ± 9% +0.2 0.31 ± 22% perf-profile.children.cycles-pp.perf_event_mmap
0.00 +0.2 0.24 ± 20% perf-profile.children.cycles-pp.vm_unmapped_area
0.02 ± 99% +0.2 0.27 ± 20% perf-profile.children.cycles-pp.arch_get_unmapped_area_topdown
0.06 ± 6% +0.2 0.31 ± 20% perf-profile.children.cycles-pp.__get_unmapped_area
0.00 +0.3 0.26 ± 29% perf-profile.children.cycles-pp.rcu_cblist_dequeue
0.00 +0.3 0.26 ± 6% perf-profile.children.cycles-pp.free_pcppages_bulk
0.00 +0.3 0.26 ± 23% perf-profile.children.cycles-pp.mas_store_prealloc
0.00 +0.3 0.27 ± 7% perf-profile.children.cycles-pp.__put_partials
0.00 +0.3 0.28 ± 7% perf-profile.children.cycles-pp.free_unref_page_commit
0.00 +0.3 0.28 ± 19% perf-profile.children.cycles-pp.vm_area_free_rcu_cb
0.14 ± 2% +0.3 0.48 ± 25% perf-profile.children.cycles-pp.mas_store_gfp
0.05 ± 7% +0.3 0.40 ± 4% perf-profile.children.cycles-pp.__memcg_slab_free_hook
0.10 +0.4 0.47 ± 24% perf-profile.children.cycles-pp.mas_wr_node_store
0.00 +0.5 0.46 ± 27% perf-profile.children.cycles-pp.__slab_free
0.07 +0.5 0.55 ± 10% perf-profile.children.cycles-pp.vm_area_alloc
0.06 +0.5 0.61 ± 12% perf-profile.children.cycles-pp.lru_add
0.00 +0.7 0.70 ± 26% perf-profile.children.cycles-pp.get_mem_cgroup_from_mm
0.00 +0.7 0.72 ± 24% perf-profile.children.cycles-pp.__mod_memcg_state
0.12 ± 6% +0.8 0.88 ± 10% perf-profile.children.cycles-pp.__folio_mod_stat
0.13 ± 5% +0.8 0.91 ± 10% perf-profile.children.cycles-pp.folio_add_new_anon_rmap
0.04 ± 71% +0.8 0.83 ± 13% perf-profile.children.cycles-pp.mod_objcg_state
0.00 +0.8 0.83 ± 35% perf-profile.children.cycles-pp.propagate_protected_usage
0.00 +0.8 0.84 ± 10% perf-profile.children.cycles-pp.folio_remove_rmap_ptes
0.00 +0.9 0.86 ± 9% perf-profile.children.cycles-pp.__anon_vma_prepare
0.00 +0.9 0.88 ± 8% perf-profile.children.cycles-pp.___pte_free_tlb
0.01 ±223% +0.9 0.90 ± 9% perf-profile.children.cycles-pp.__vmf_anon_prepare
0.00 +0.9 0.90 ± 9% perf-profile.children.cycles-pp.zap_present_ptes
0.06 +1.0 1.02 ± 10% perf-profile.children.cycles-pp.__memcg_slab_post_alloc_hook
0.05 ± 7% +1.0 1.02 ± 33% perf-profile.children.cycles-pp.native_irq_return_iret
0.00 +1.0 1.04 ± 9% perf-profile.children.cycles-pp.free_pud_range
0.00 +1.1 1.05 ± 9% perf-profile.children.cycles-pp.free_p4d_range
0.00 +1.1 1.07 ± 9% perf-profile.children.cycles-pp.free_pgd_range
0.00 +1.1 1.11 ± 75% perf-profile.children.cycles-pp.uncharge_folio
0.12 ± 4% +1.1 1.24 ± 11% perf-profile.children.cycles-pp.kmem_cache_free
0.10 ± 4% +1.2 1.30 ± 46% perf-profile.children.cycles-pp.__mem_cgroup_charge
0.11 ± 4% +1.2 1.34 ± 9% perf-profile.children.cycles-pp.kmem_cache_alloc_noprof
0.16 ± 3% +1.3 1.41 ± 16% perf-profile.children.cycles-pp.handle_softirqs
0.13 +1.3 1.41 ± 16% perf-profile.children.cycles-pp.rcu_core
0.12 ± 4% +1.3 1.41 ± 16% perf-profile.children.cycles-pp.rcu_do_batch
0.10 ± 3% +1.3 1.39 ± 9% perf-profile.children.cycles-pp.free_pgtables
0.06 +1.5 1.53 ± 28% perf-profile.children.cycles-pp.__memcg_kmem_charge_page
0.28 ± 4% +1.5 1.75 ± 18% perf-profile.children.cycles-pp.mmap_region
0.00 +1.5 1.54 ± 34% perf-profile.children.cycles-pp.alloc_anon_folio
0.38 ± 2% +1.6 1.96 ± 17% perf-profile.children.cycles-pp.__alloc_pages_noprof
0.39 ± 2% +1.6 2.00 ± 16% perf-profile.children.cycles-pp.alloc_pages_mpol_noprof
0.34 ± 2% +1.8 2.11 ± 19% perf-profile.children.cycles-pp.do_mmap
0.36 ± 2% +1.9 2.24 ± 19% perf-profile.children.cycles-pp.vm_mmap_pgoff
0.00 +2.0 2.02 ± 14% perf-profile.children.cycles-pp.zap_pte_range
0.37 ± 3% +2.1 2.48 ± 19% perf-profile.children.cycles-pp.__mmap
0.22 ± 2% +2.3 2.50 ± 16% perf-profile.children.cycles-pp.zap_pmd_range
0.23 ± 3% +2.3 2.55 ± 16% perf-profile.children.cycles-pp.unmap_page_range
0.24 ± 3% +2.3 2.58 ± 16% perf-profile.children.cycles-pp.unmap_vmas
0.16 +2.5 2.63 ± 12% perf-profile.children.cycles-pp.pte_alloc_one
0.08 ± 12% +2.6 2.71 ± 32% perf-profile.children.cycles-pp.__mod_memcg_lruvec_state
0.00 +2.7 2.69 ± 12% perf-profile.children.cycles-pp.__pte_alloc
0.05 +3.2 3.22 ± 38% perf-profile.children.cycles-pp.page_counter_cancel
0.22 ± 5% +3.2 3.40 ± 10% perf-profile.children.cycles-pp.__lruvec_stat_mod_folio
0.05 +4.0 4.03 ± 37% perf-profile.children.cycles-pp.page_counter_uncharge
0.06 +5.0 5.04 ± 23% perf-profile.children.cycles-pp.uncharge_batch
0.06 ± 7% +6.1 6.16 ± 7% perf-profile.children.cycles-pp.__mem_cgroup_uncharge_folios
0.00 +6.2 6.16 ± 12% perf-profile.children.cycles-pp.do_anonymous_page
0.10 ± 5% +37.8 37.88 ± 3% perf-profile.children.cycles-pp.__page_cache_release
0.12 ± 3% +38.3 38.46 perf-profile.children.cycles-pp.folio_batch_move_lru
0.00 +38.5 38.46 perf-profile.children.cycles-pp.lru_add_drain_cpu
0.00 +38.5 38.47 perf-profile.children.cycles-pp.lru_add_drain
0.79 ± 2% +44.2 44.99 perf-profile.children.cycles-pp.free_pages_and_swap_cache
0.79 ± 2% +44.2 45.01 perf-profile.children.cycles-pp.__tlb_batch_free_encoded_pages
0.78 ± 2% +44.2 45.01 perf-profile.children.cycles-pp.folios_put_refs
0.86 ± 2% +44.4 45.24 perf-profile.children.cycles-pp.tlb_finish_mmu
0.12 ± 6% +73.4 73.50 ± 2% perf-profile.children.cycles-pp.native_queued_spin_lock_slowpath
0.21 ± 4% +73.6 73.77 ± 2% perf-profile.children.cycles-pp._raw_spin_lock_irqsave
0.08 +74.3 74.35 ± 2% perf-profile.children.cycles-pp.folio_lruvec_lock_irqsave
1.23 ± 2% +86.5 87.71 perf-profile.children.cycles-pp.vms_clear_ptes
1.26 +86.6 87.87 perf-profile.children.cycles-pp.vms_complete_munmap_vmas
1.44 +87.1 88.55 perf-profile.children.cycles-pp.do_vmi_align_munmap
1.48 +87.1 88.60 perf-profile.children.cycles-pp.do_vmi_munmap
1.50 +87.2 88.66 perf-profile.children.cycles-pp.__x64_sys_munmap
1.49 +87.2 88.66 perf-profile.children.cycles-pp.__vm_munmap
1.53 +87.4 88.93 perf-profile.children.cycles-pp.__munmap
2.30 +88.8 91.10 perf-profile.children.cycles-pp.do_syscall_64
2.30 +88.8 91.12 perf-profile.children.cycles-pp.entry_SYSCALL_64_after_hwframe
88.52 -88.3 0.20 ± 22% perf-profile.self.cycles-pp.clear_page_erms
3.36 -3.2 0.20 ± 23% perf-profile.self.cycles-pp.__cond_resched
0.49 -0.5 0.03 ± 70% perf-profile.self.cycles-pp.free_unref_folios
0.37 ± 33% -0.1 0.24 ± 8% perf-profile.self.cycles-pp.queue_event
0.00 +0.1 0.09 ± 22% perf-profile.self.cycles-pp.___perf_sw_event
0.00 +0.1 0.10 ± 64% perf-profile.self.cycles-pp.__count_memcg_events
0.00 +0.1 0.10 ± 22% perf-profile.self.cycles-pp.sync_regs
0.00 +0.1 0.10 ± 23% perf-profile.self.cycles-pp.perf_event_mmap_output
0.00 +0.1 0.10 ± 47% perf-profile.self.cycles-pp.obj_cgroup_uncharge_pages
0.00 +0.1 0.11 ± 24% perf-profile.self.cycles-pp.native_flush_tlb_local
0.00 +0.1 0.11 ± 21% perf-profile.self.cycles-pp.kmem_cache_free
0.00 +0.1 0.11 ± 25% perf-profile.self.cycles-pp.__memcg_slab_free_hook
0.00 +0.1 0.12 ± 22% perf-profile.self.cycles-pp.__page_cache_release
0.00 +0.1 0.13 ± 24% perf-profile.self.cycles-pp.mmap_region
0.00 +0.1 0.13 ± 26% perf-profile.self.cycles-pp.obj_cgroup_charge
0.00 +0.2 0.15 ± 22% perf-profile.self.cycles-pp.__memcg_slab_post_alloc_hook
0.00 +0.2 0.16 ± 22% perf-profile.self.cycles-pp.clear_bhb_loop
0.00 +0.2 0.16 ± 25% perf-profile.self.cycles-pp.free_pud_range
0.00 +0.2 0.17 ± 24% perf-profile.self.cycles-pp.kmem_cache_alloc_noprof
0.00 +0.2 0.17 ± 23% perf-profile.self.cycles-pp.lru_add
0.00 +0.2 0.17 ± 24% perf-profile.self.cycles-pp.mas_walk
0.00 +0.2 0.18 ± 22% perf-profile.self.cycles-pp.folio_batch_move_lru
0.09 +0.2 0.28 ± 29% perf-profile.self.cycles-pp._raw_spin_lock_irqsave
0.06 ± 9% +0.2 0.26 ± 25% perf-profile.self.cycles-pp.mas_wr_node_store
0.00 +0.2 0.23 ± 21% perf-profile.self.cycles-pp.zap_pmd_range
0.00 +0.3 0.25 ± 28% perf-profile.self.cycles-pp.rcu_cblist_dequeue
0.00 +0.3 0.28 ± 21% perf-profile.self.cycles-pp.folios_put_refs
0.00 +0.4 0.44 ± 26% perf-profile.self.cycles-pp.__slab_free
0.00 +0.5 0.46 ± 33% perf-profile.self.cycles-pp.mod_objcg_state
0.00 +0.5 0.48 ± 96% perf-profile.self.cycles-pp.__mem_cgroup_charge
0.00 +0.6 0.59 ± 32% perf-profile.self.cycles-pp.uncharge_batch
0.00 +0.7 0.69 ± 24% perf-profile.self.cycles-pp.__mod_memcg_state
0.00 +0.7 0.69 ± 27% perf-profile.self.cycles-pp.get_mem_cgroup_from_mm
0.00 +0.8 0.81 ± 35% perf-profile.self.cycles-pp.propagate_protected_usage
0.00 +0.8 0.82 ± 36% perf-profile.self.cycles-pp.folio_lruvec_lock_irqsave
0.05 ± 7% +1.0 1.02 ± 33% perf-profile.self.cycles-pp.native_irq_return_iret
0.00 +1.0 1.02 ± 22% perf-profile.self.cycles-pp.zap_pte_range
0.00 +1.1 1.10 ± 76% perf-profile.self.cycles-pp.uncharge_folio
0.00 +1.1 1.14 ± 37% perf-profile.self.cycles-pp.__memcg_kmem_charge_page
0.08 ± 4% +1.7 1.78 ± 35% perf-profile.self.cycles-pp.__lruvec_stat_mod_folio
0.07 ± 16% +2.6 2.62 ± 33% perf-profile.self.cycles-pp.__mod_memcg_lruvec_state
0.05 +3.1 3.16 ± 38% perf-profile.self.cycles-pp.page_counter_cancel
0.12 ± 7% +73.4 73.50 ± 2% perf-profile.self.cycles-pp.native_queued_spin_lock_slowpath
***************************************************************************************************
lkp-spr-r02: 224 threads 2 sockets Intel(R) Xeon(R) Platinum 8480CTDX (Sapphire Rapids) with 256G memory
=========================================================================================
compiler/cpufreq_governor/kconfig/nr_threads/rootfs/tbox_group/test/testcase/testtime:
gcc-12/performance/x86_64-rhel-8.3/100%/debian-12-x86_64-20240206.cgz/lkp-spr-r02/radixsort/stress-ng/60s
commit:
15e8156713 ("mm: shrinker: avoid memleak in alloc_shrinker_info")
d4148aeab4 ("mm, mmap: limit THP alignment of anonymous mappings to PMD-aligned sizes")
15e8156713cc3803 d4148aeab412432bf928f311eca
---------------- ---------------------------
%stddev %change %stddev
\ | \
18.98 ± 5% +30.5% 24.78 ± 14% sched_debug.cpu.clock.stddev
0.05 ± 7% +886.3% 0.50 ± 86% perf-sched.sch_delay.avg.ms.__cond_resched.__wait_for_common.affine_move_task.__set_cpus_allowed_ptr.__sched_setaffinity
1.14 ± 74% -55.5% 0.51 ± 37% perf-sched.sch_delay.avg.ms.__x64_sys_pause.do_syscall_64.entry_SYSCALL_64_after_hwframe.[unknown]
265008 ± 13% -37.2% 166403 ± 22% meminfo.Active
264976 ± 13% -37.2% 166371 ± 22% meminfo.Active(anon)
1315990 -63.8% 475807 ± 3% meminfo.AnonHugePages
164713 -9.2% 149574 stress-ng.radixsort.ops
2745 -9.2% 2492 stress-ng.radixsort.ops_per_sec
78925 +289.9% 307739 stress-ng.time.minor_page_faults
327.88 ± 3% -67.1% 107.89 ± 3% numa-vmstat.node0.nr_anon_transparent_hugepages
63711 ± 11% -40.7% 37791 ± 25% numa-vmstat.node1.nr_active_anon
317.83 ± 3% -60.5% 125.70 ± 8% numa-vmstat.node1.nr_anon_transparent_hugepages
293623 ± 18% +35.1% 396637 ± 14% numa-vmstat.node1.nr_inactive_anon
63710 ± 11% -40.7% 37791 ± 25% numa-vmstat.node1.nr_zone_active_anon
293622 ± 18% +35.1% 396635 ± 14% numa-vmstat.node1.nr_zone_inactive_anon
668445 ± 3% -67.1% 220222 ± 3% numa-meminfo.node0.AnonHugePages
1317034 ± 13% -23.2% 1010970 ± 16% numa-meminfo.node0.AnonPages.max
254241 ± 11% -40.5% 151233 ± 26% numa-meminfo.node1.Active
254219 ± 11% -40.5% 151227 ± 26% numa-meminfo.node1.Active(anon)
647491 ± 3% -60.4% 256719 ± 8% numa-meminfo.node1.AnonHugePages
1170941 ± 18% +35.3% 1583801 ± 14% numa-meminfo.node1.Inactive
1170783 ± 18% +35.3% 1583760 ± 14% numa-meminfo.node1.Inactive(anon)
66923 ± 13% -36.6% 42442 ± 22% proc-vmstat.nr_active_anon
642.60 -63.8% 232.40 ± 3% proc-vmstat.nr_anon_transparent_hugepages
597472 ± 2% +5.6% 630776 proc-vmstat.nr_inactive_anon
66923 ± 13% -36.6% 42442 ± 22% proc-vmstat.nr_zone_active_anon
597472 ± 2% +5.6% 630776 proc-vmstat.nr_zone_inactive_anon
1188533 +19.2% 1416356 proc-vmstat.numa_hit
2502 -63.5% 913.67 proc-vmstat.numa_huge_pte_updates
956729 +23.2% 1178316 proc-vmstat.numa_local
1495630 ± 2% -54.0% 687360 ± 6% proc-vmstat.numa_pte_updates
820025 +26.8% 1040066 proc-vmstat.pgfault
6.20 +9.6% 6.79 perf-stat.i.MPKI
4.203e+10 -6.8% 3.916e+10 perf-stat.i.branch-instructions
6.15 +0.2 6.30 perf-stat.i.branch-miss-rate%
2.636e+09 -6.2% 2.474e+09 perf-stat.i.branch-misses
2.56 ± 2% +8.5% 2.77 perf-stat.i.cpi
2.444e+11 -6.9% 2.276e+11 perf-stat.i.instructions
0.40 -9.1% 0.36 perf-stat.i.ipc
2.20 ± 2% +90.0% 4.18 ± 2% perf-stat.i.metric.K/sec
10808 +45.4% 15713 ± 2% perf-stat.i.minor-faults
10809 +45.4% 15713 ± 2% perf-stat.i.page-faults
6.34 +7.9% 6.84 perf-stat.overall.MPKI
37.76 -0.6 37.12 perf-stat.overall.cache-miss-rate%
2.54 +9.9% 2.80 perf-stat.overall.cpi
401.20 +1.8% 408.57 perf-stat.overall.cycles-between-cache-misses
0.39 -9.0% 0.36 perf-stat.overall.ipc
4.134e+10 ± 2% -7.4% 3.829e+10 perf-stat.ps.branch-instructions
2.592e+09 -6.7% 2.419e+09 perf-stat.ps.branch-misses
2.404e+11 ± 2% -7.4% 2.226e+11 perf-stat.ps.instructions
10358 +36.2% 14109 ± 2% perf-stat.ps.minor-faults
10358 +36.2% 14109 ± 2% perf-stat.ps.page-faults
1.525e+13 -9.9% 1.374e+13 perf-stat.total.instructions
0.60 ± 10% -0.3 0.28 ±100% perf-profile.calltrace.cycles-pp.__perf_mmap__read_init.perf_mmap__read_init.perf_mmap__push.record__mmap_read_evlist.__cmd_record
0.71 ± 4% +0.1 0.77 ± 3% perf-profile.calltrace.cycles-pp.update_load_avg.task_tick_fair.sched_tick.update_process_times.tick_nohz_handler
1.25 ± 9% +0.3 1.54 ± 4% perf-profile.calltrace.cycles-pp.update_curr.task_tick_fair.sched_tick.update_process_times.tick_nohz_handler
0.26 ±100% +0.3 0.57 ± 3% perf-profile.calltrace.cycles-pp.__irq_exit_rcu.sysvec_apic_timer_interrupt.asm_sysvec_apic_timer_interrupt
2.64 ± 6% +0.4 3.09 ± 3% perf-profile.calltrace.cycles-pp.task_tick_fair.sched_tick.update_process_times.tick_nohz_handler.__hrtimer_run_queues
0.08 ±223% +0.5 0.55 ± 4% perf-profile.calltrace.cycles-pp.account_user_time.update_process_times.tick_nohz_handler.__hrtimer_run_queues.hrtimer_interrupt
5.00 ± 6% +0.7 5.75 ± 3% perf-profile.calltrace.cycles-pp.sched_tick.update_process_times.tick_nohz_handler.__hrtimer_run_queues.hrtimer_interrupt
7.66 ± 5% +1.1 8.72 ± 3% perf-profile.calltrace.cycles-pp.update_process_times.tick_nohz_handler.__hrtimer_run_queues.hrtimer_interrupt.__sysvec_apic_timer_interrupt
8.10 ± 5% +1.1 9.18 ± 3% perf-profile.calltrace.cycles-pp.tick_nohz_handler.__hrtimer_run_queues.hrtimer_interrupt.__sysvec_apic_timer_interrupt.sysvec_apic_timer_interrupt
8.38 ± 5% +1.1 9.51 ± 3% perf-profile.calltrace.cycles-pp.__hrtimer_run_queues.hrtimer_interrupt.__sysvec_apic_timer_interrupt.sysvec_apic_timer_interrupt.asm_sysvec_apic_timer_interrupt
10.78 ± 5% +1.3 12.08 ± 3% perf-profile.calltrace.cycles-pp.hrtimer_interrupt.__sysvec_apic_timer_interrupt.sysvec_apic_timer_interrupt.asm_sysvec_apic_timer_interrupt
11.04 ± 5% +1.3 12.36 ± 3% perf-profile.calltrace.cycles-pp.__sysvec_apic_timer_interrupt.sysvec_apic_timer_interrupt.asm_sysvec_apic_timer_interrupt
11.77 ± 5% +1.4 13.19 ± 2% perf-profile.calltrace.cycles-pp.sysvec_apic_timer_interrupt.asm_sysvec_apic_timer_interrupt
12.68 ± 5% +1.6 14.23 ± 2% perf-profile.calltrace.cycles-pp.asm_sysvec_apic_timer_interrupt
12.84 -0.4 12.47 ± 2% perf-profile.children.cycles-pp.strcmp@plt
0.65 ± 10% -0.2 0.50 ± 30% perf-profile.children.cycles-pp.__perf_mmap__read_init
0.12 ± 3% +0.0 0.14 ± 4% perf-profile.children.cycles-pp.sched_clock
0.05 ± 8% +0.0 0.08 ± 8% perf-profile.children.cycles-pp.rb_next
0.09 ± 7% +0.0 0.13 ± 10% perf-profile.children.cycles-pp.timerqueue_del
0.20 ± 4% +0.0 0.23 ± 4% perf-profile.children.cycles-pp.lapic_next_deadline
0.40 ± 7% +0.0 0.44 ± 5% perf-profile.children.cycles-pp.handle_softirqs
0.42 ± 4% +0.0 0.47 ± 5% perf-profile.children.cycles-pp.native_irq_return_iret
0.20 ± 9% +0.1 0.26 ± 6% perf-profile.children.cycles-pp.__cgroup_account_cputime_field
0.12 ± 4% +0.1 0.19 ± 10% perf-profile.children.cycles-pp._raw_spin_lock
0.75 ± 4% +0.1 0.82 ± 3% perf-profile.children.cycles-pp.update_load_avg
0.40 ± 4% +0.1 0.47 ± 2% perf-profile.children.cycles-pp.hrtimer_active
0.34 ± 9% +0.1 0.41 ± 5% perf-profile.children.cycles-pp.irqtime_account_irq
0.52 ± 5% +0.1 0.60 ± 3% perf-profile.children.cycles-pp.__irq_exit_rcu
0.12 ± 9% +0.1 0.20 ± 6% perf-profile.children.cycles-pp.perf_trace_sched_stat_runtime
0.36 ± 2% +0.1 0.44 ± 3% perf-profile.children.cycles-pp.task_mm_cid_work
0.50 ± 6% +0.1 0.58 ± 4% perf-profile.children.cycles-pp.account_user_time
0.38 ± 2% +0.1 0.48 ± 3% perf-profile.children.cycles-pp.task_work_run
0.46 ± 2% +0.1 0.56 ± 3% perf-profile.children.cycles-pp.irqentry_exit_to_user_mode
1.32 ± 9% +0.3 1.64 ± 4% perf-profile.children.cycles-pp.update_curr
2.76 ± 6% +0.5 3.25 ± 3% perf-profile.children.cycles-pp.task_tick_fair
5.24 ± 6% +0.8 6.04 ± 3% perf-profile.children.cycles-pp.sched_tick
7.99 ± 5% +1.2 9.14 ± 3% perf-profile.children.cycles-pp.update_process_times
8.45 ± 5% +1.2 9.63 ± 3% perf-profile.children.cycles-pp.tick_nohz_handler
8.75 ± 5% +1.2 9.99 ± 3% perf-profile.children.cycles-pp.__hrtimer_run_queues
11.23 ± 5% +1.4 12.66 ± 3% perf-profile.children.cycles-pp.hrtimer_interrupt
11.50 ± 5% +1.4 12.94 ± 3% perf-profile.children.cycles-pp.__sysvec_apic_timer_interrupt
12.26 ± 5% +1.6 13.82 ± 3% perf-profile.children.cycles-pp.sysvec_apic_timer_interrupt
13.22 ± 5% +1.7 14.92 ± 2% perf-profile.children.cycles-pp.asm_sysvec_apic_timer_interrupt
6.12 -0.2 5.92 ± 2% perf-profile.self.cycles-pp.strcmp@plt
0.09 ± 11% +0.0 0.11 ± 6% perf-profile.self.cycles-pp.__hrtimer_run_queues
0.10 ± 9% +0.0 0.12 ± 5% perf-profile.self.cycles-pp.task_tick_fair
0.05 ± 8% +0.0 0.08 ± 10% perf-profile.self.cycles-pp.rb_next
0.20 ± 4% +0.0 0.23 ± 3% perf-profile.self.cycles-pp.lapic_next_deadline
0.42 ± 4% +0.0 0.47 ± 5% perf-profile.self.cycles-pp.native_irq_return_iret
0.16 ± 14% +0.1 0.21 ± 4% perf-profile.self.cycles-pp.__cgroup_account_cputime_field
0.12 ± 4% +0.1 0.18 ± 10% perf-profile.self.cycles-pp._raw_spin_lock
0.24 ± 9% +0.1 0.30 ± 7% perf-profile.self.cycles-pp.irqtime_account_irq
0.40 ± 4% +0.1 0.46 ± 3% perf-profile.self.cycles-pp.hrtimer_active
0.10 ± 3% +0.1 0.18 ± 8% perf-profile.self.cycles-pp.hrtimer_interrupt
0.11 ± 11% +0.1 0.20 ± 6% perf-profile.self.cycles-pp.perf_trace_sched_stat_runtime
0.33 ± 3% +0.1 0.42 ± 4% perf-profile.self.cycles-pp.task_mm_cid_work
0.88 ± 11% +0.2 1.05 ± 5% perf-profile.self.cycles-pp.update_curr
Disclaimer:
Results have been estimated based on internal Intel analysis and are provided
for informational purposes only. Any difference in system hardware or software
design or configuration may affect actual performance.
--
0-DAY CI Kernel Test Service
https://github.com/intel/lkp-tests/wiki
Powered by blists - more mailing lists