[<prev] [next>] [thread-next>] [day] [month] [year] [list]
Message-ID: <202401221624.cb53a8ca-oliver.sang@intel.com>
Date: Mon, 22 Jan 2024 16:39:05 +0800
From: kernel test robot <oliver.sang@...el.com>
To: Yosry Ahmed <yosryahmed@...gle.com>
CC: <oe-lkp@...ts.linux.dev>, <lkp@...el.com>, <linux-kernel@...r.kernel.org>,
Andrew Morton <akpm@...ux-foundation.org>, Johannes Weiner
<hannes@...xchg.org>, Domenico Cerasuolo <cerasuolodomenico@...il.com>,
Shakeel Butt <shakeelb@...gle.com>, Chris Li <chrisl@...nel.org>, Greg Thelen
<gthelen@...gle.com>, Ivan Babrou <ivan@...udflare.com>, Michal Hocko
<mhocko@...nel.org>, Michal Koutny <mkoutny@...e.com>, Muchun Song
<muchun.song@...ux.dev>, Roman Gushchin <roman.gushchin@...ux.dev>, Tejun Heo
<tj@...nel.org>, Waiman Long <longman@...hat.com>, Wei Xu
<weixugc@...gle.com>, <cgroups@...r.kernel.org>, <linux-mm@...ck.org>,
<ying.huang@...el.com>, <feng.tang@...el.com>, <fengwei.yin@...el.com>,
<oliver.sang@...el.com>
Subject: [linus:master] [mm] 8d59d2214c: vm-scalability.throughput -36.6%
regression
hi, Yosry Ahmed,
per your suggestion in
https://lore.kernel.org/all/CAJD7tkameJBrJQxRj+ibKL6-yd-i0wyoyv2cgZdh3ZepA1p7wA@mail.gmail.com/
"I think it would be useful to know if there are
regressions/improvements in other microbenchmarks, at least to
investigate whether they represent real regressions."
we still report below two regressions to you just FYI what we observed in our
microbenchmark tests.
(we still captured will-it-scale::fallocate regression but ignore here per
your commit message)
Hello,
kernel test robot noticed a -36.6% regression of vm-scalability.throughput on:
commit: 8d59d2214c2362e7a9d185d80b613e632581af7b ("mm: memcg: make stats flushing threshold per-memcg")
https://git.kernel.org/cgit/linux/kernel/git/torvalds/linux.git master
testcase: vm-scalability
test machine: 224 threads 4 sockets Intel(R) Xeon(R) Platinum 8380H CPU @ 2.90GHz (Cooper Lake) with 192G memory
parameters:
runtime: 300s
size: 1T
test: lru-shm
cpufreq_governor: performance
test-description: The motivation behind this suite is to exercise functions and regions of the mm/ of the Linux kernel which are of interest to us.
test-url: https://git.kernel.org/cgit/linux/kernel/git/wfg/vm-scalability.git/
In addition to that, the commit also has significant impact on the following tests:
+------------------+----------------------------------------------------------------------------------------------------+
| testcase: change | will-it-scale: will-it-scale.per_process_ops -32.3% regression |
| test machine | 104 threads 2 sockets (Skylake) with 192G memory |
| test parameters | cpufreq_governor=performance |
| | mode=process |
| | nr_task=50% |
| | test=tlb_flush2 |
+------------------+----------------------------------------------------------------------------------------------------+
If you fix the issue in a separate patch/commit (i.e. not just a new version of
the same patch/commit), kindly add following tags
| Reported-by: kernel test robot <oliver.sang@...el.com>
| Closes: https://lore.kernel.org/oe-lkp/202401221624.cb53a8ca-oliver.sang@intel.com
Details are as below:
-------------------------------------------------------------------------------------------------->
The kernel config and materials to reproduce are available at:
https://download.01.org/0day-ci/archive/20240122/202401221624.cb53a8ca-oliver.sang@intel.com
=========================================================================================
compiler/cpufreq_governor/kconfig/rootfs/runtime/size/tbox_group/test/testcase:
gcc-12/performance/x86_64-rhel-8.3/debian-11.1-x86_64-20220510.cgz/300s/1T/lkp-cpl-4sp2/lru-shm/vm-scalability
commit:
e0bf1dc859 ("mm: memcg: move vmstats structs definition above flushing code")
8d59d2214c ("mm: memcg: make stats flushing threshold per-memcg")
e0bf1dc859fdd08e 8d59d2214c2362e7a9d185d80b6
---------------- ---------------------------
%stddev %change %stddev
\ | \
0.01 +86.7% 0.02 vm-scalability.free_time
946447 -37.8% 588327 vm-scalability.median
2.131e+08 -36.6% 1.351e+08 vm-scalability.throughput
284.74 +6.3% 302.62 vm-scalability.time.elapsed_time
284.74 +6.3% 302.62 vm-scalability.time.elapsed_time.max
30485 +14.8% 34987 vm-scalability.time.involuntary_context_switches
1893 +43.6% 2718 vm-scalability.time.percent_of_cpu_this_job_got
3855 +67.7% 6467 vm-scalability.time.system_time
1537 +14.5% 1760 vm-scalability.time.user_time
120009 -5.6% 113290 vm-scalability.time.voluntary_context_switches
6.46 +3.5 9.95 mpstat.cpu.all.sys%
21.22 +38.8% 29.46 vmstat.procs.r
0.01 ± 20% +1887.0% 0.18 ±203% perf-sched.sch_delay.max.ms.__cond_resched.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
0.01 ± 28% +63.3% 0.01 ± 29% perf-sched.sch_delay.max.ms.wait_for_partner.fifo_open.do_dentry_open.do_open
113624 ± 5% +14.0% 129566 ± 3% meminfo.Active
113476 ± 5% +14.0% 129417 ± 3% meminfo.Active(anon)
3987746 +46.0% 5821636 meminfo.Mapped
16345 +14.6% 18729 meminfo.PageTables
474.17 ± 3% -88.9% 52.50 ±125% perf-c2c.DRAM.local
483.17 ± 5% -79.3% 99.83 ± 70% perf-c2c.DRAM.remote
1045 ± 5% -71.9% 294.00 ± 63% perf-c2c.HITM.local
119.50 ± 10% -78.8% 25.33 ± 20% perf-c2c.HITM.remote
392.33 +35.4% 531.17 turbostat.Avg_MHz
10.35 +3.7 14.00 turbostat.Busy%
90.56 -3.7 86.86 turbostat.C1%
0.28 ± 5% -31.5% 0.19 turbostat.IPC
481.33 +2.5% 493.38 turbostat.PkgWatt
999019 ± 3% +44.4% 1442651 ± 2% numa-meminfo.node0.Mapped
1005687 ± 4% +44.1% 1449402 ± 3% numa-meminfo.node1.Mapped
3689 ± 3% +21.7% 4490 ± 7% numa-meminfo.node1.PageTables
980589 ± 2% +42.3% 1395777 ± 2% numa-meminfo.node2.Mapped
96484 ± 5% +22.0% 117715 ± 4% numa-meminfo.node3.Active
96430 ± 5% +22.1% 117694 ± 4% numa-meminfo.node3.Active(anon)
991367 ± 3% +42.7% 1414337 ± 4% numa-meminfo.node3.Mapped
251219 ± 3% +44.8% 363745 ± 2% numa-vmstat.node0.nr_mapped
253252 ± 2% +44.6% 366087 ± 3% numa-vmstat.node1.nr_mapped
927.67 ± 3% +21.9% 1130 ± 7% numa-vmstat.node1.nr_page_table_pages
248171 ± 2% +42.5% 353541 ± 4% numa-vmstat.node2.nr_mapped
24188 ± 5% +21.6% 29410 ± 4% numa-vmstat.node3.nr_active_anon
245825 ± 2% +45.5% 357622 ± 3% numa-vmstat.node3.nr_mapped
1038 ± 11% +17.8% 1224 ± 6% numa-vmstat.node3.nr_page_table_pages
24188 ± 5% +21.6% 29410 ± 4% numa-vmstat.node3.nr_zone_active_anon
28376 ± 5% +14.0% 32338 ± 3% proc-vmstat.nr_active_anon
993504 +46.6% 1456136 proc-vmstat.nr_mapped
4060 +15.5% 4691 proc-vmstat.nr_page_table_pages
28376 ± 5% +14.0% 32338 ± 3% proc-vmstat.nr_zone_active_anon
1.066e+09 -2.0% 1.045e+09 proc-vmstat.numa_hit
1.065e+09 -2.0% 1.044e+09 proc-vmstat.numa_local
5659 +5.6% 5978 proc-vmstat.unevictable_pgs_culled
34604288 +3.7% 35898496 proc-vmstat.unevictable_pgs_scanned
1223376 ± 14% +119.1% 2680582 ± 9% sched_debug.cfs_rq:/.avg_vruntime.avg
1673909 ± 14% +97.6% 3308254 ± 8% sched_debug.cfs_rq:/.avg_vruntime.max
810795 ± 15% +145.8% 1993289 ± 9% sched_debug.cfs_rq:/.avg_vruntime.min
156233 ± 8% +55.1% 242331 ± 6% sched_debug.cfs_rq:/.avg_vruntime.stddev
1223376 ± 14% +119.1% 2680582 ± 9% sched_debug.cfs_rq:/.min_vruntime.avg
1673909 ± 14% +97.6% 3308254 ± 8% sched_debug.cfs_rq:/.min_vruntime.max
810795 ± 15% +145.8% 1993289 ± 9% sched_debug.cfs_rq:/.min_vruntime.min
156233 ± 8% +55.1% 242331 ± 6% sched_debug.cfs_rq:/.min_vruntime.stddev
126445 ± 3% -11.0% 112493 ± 4% sched_debug.cpu.avg_idle.stddev
1447 ± 15% +32.0% 1910 ± 9% sched_debug.cpu.nr_switches.min
0.71 +13.4% 0.80 perf-stat.i.MPKI
2.343e+10 -7.9% 2.157e+10 perf-stat.i.branch-instructions
0.36 -0.0 0.35 perf-stat.i.branch-miss-rate%
30833194 -7.3% 28584190 perf-stat.i.branch-misses
26.04 -1.4 24.66 perf-stat.i.cache-miss-rate%
51345490 ± 3% +40.7% 72258633 ± 3% perf-stat.i.cache-misses
1.616e+08 ± 6% +58.6% 2.562e+08 ± 6% perf-stat.i.cache-references
1.29 +9.4% 1.42 perf-stat.i.cpi
8.394e+10 +33.7% 1.122e+11 perf-stat.i.cpu-cycles
505.77 -2.6% 492.52 perf-stat.i.cpu-migrations
0.03 +0.0 0.03 ± 2% perf-stat.i.dTLB-load-miss-rate%
2.335e+10 -7.4% 2.162e+10 perf-stat.i.dTLB-loads
0.03 +0.0 0.03 perf-stat.i.dTLB-store-miss-rate%
3948344 -8.0% 3633633 perf-stat.i.dTLB-store-misses
6.549e+09 -7.0% 6.09e+09 perf-stat.i.dTLB-stores
17546602 -22.8% 13551001 perf-stat.i.iTLB-load-misses
2552560 -2.6% 2485876 perf-stat.i.iTLB-loads
8.367e+10 -7.5% 7.737e+10 perf-stat.i.instructions
4706 +7.7% 5070 perf-stat.i.instructions-per-iTLB-miss
0.81 -12.0% 0.72 perf-stat.i.ipc
1.59 ± 3% -22.3% 1.23 ± 4% perf-stat.i.major-faults
0.37 +34.2% 0.49 perf-stat.i.metric.GHz
233.98 -6.9% 217.90 perf-stat.i.metric.M/sec
3619177 -9.5% 3276556 perf-stat.i.minor-faults
74.28 +4.8 79.04 perf-stat.i.node-load-miss-rate%
2898733 ± 4% +49.0% 4320557 perf-stat.i.node-load-misses
1928237 ± 4% -11.9% 1698426 perf-stat.i.node-loads
13383344 ± 2% +4.7% 14013398 ± 3% perf-stat.i.node-stores
3619179 -9.5% 3276558 perf-stat.i.page-faults
0.61 ± 3% +52.5% 0.94 ± 3% perf-stat.overall.MPKI
31.95 ± 2% -3.6 28.34 ± 3% perf-stat.overall.cache-miss-rate%
1.00 +45.0% 1.45 perf-stat.overall.cpi
0.07 +0.0 0.08 ± 4% perf-stat.overall.dTLB-load-miss-rate%
87.62 -2.6 85.05 perf-stat.overall.iTLB-load-miss-rate%
4778 +20.2% 5745 perf-stat.overall.instructions-per-iTLB-miss
1.00 -31.0% 0.69 perf-stat.overall.ipc
59.75 ± 3% +11.8 71.59 perf-stat.overall.node-load-miss-rate%
5145 +1.8% 5239 perf-stat.overall.path-length
2.405e+10 -6.3% 2.252e+10 perf-stat.ps.branch-instructions
31203502 -6.4% 29219514 perf-stat.ps.branch-misses
52696784 ± 3% +43.4% 75547948 ± 3% perf-stat.ps.cache-misses
1.652e+08 ± 6% +61.7% 2.672e+08 ± 7% perf-stat.ps.cache-references
8.584e+10 +36.3% 1.17e+11 perf-stat.ps.cpu-cycles
506.29 -2.0% 496.05 perf-stat.ps.cpu-migrations
2.395e+10 -5.9% 2.254e+10 perf-stat.ps.dTLB-loads
4059043 -6.2% 3806002 perf-stat.ps.dTLB-store-misses
6.688e+09 -5.7% 6.308e+09 perf-stat.ps.dTLB-stores
17944396 -21.8% 14028927 perf-stat.ps.iTLB-load-misses
2534093 -2.7% 2465233 perf-stat.ps.iTLB-loads
8.575e+10 -6.0% 8.059e+10 perf-stat.ps.instructions
1.60 ± 3% -23.2% 1.23 ± 4% perf-stat.ps.major-faults
3726053 -7.7% 3439511 perf-stat.ps.minor-faults
2942507 ± 4% +52.0% 4472428 perf-stat.ps.node-load-misses
1980077 ± 4% -10.4% 1774633 perf-stat.ps.node-loads
13780660 ± 2% +6.8% 14716100 ± 3% perf-stat.ps.node-stores
3726055 -7.7% 3439513 perf-stat.ps.page-faults
37.11 -6.7 30.40 ± 6% perf-profile.calltrace.cycles-pp.asm_sysvec_apic_timer_interrupt.acpi_safe_halt.acpi_idle_enter.cpuidle_enter_state.cpuidle_enter
21.14 -3.8 17.36 ± 7% perf-profile.calltrace.cycles-pp.secondary_startup_64_no_verify
21.05 -3.8 17.29 ± 7% perf-profile.calltrace.cycles-pp.do_idle.cpu_startup_entry.start_secondary.secondary_startup_64_no_verify
21.05 -3.8 17.29 ± 7% perf-profile.calltrace.cycles-pp.cpu_startup_entry.start_secondary.secondary_startup_64_no_verify
21.05 -3.8 17.29 ± 7% perf-profile.calltrace.cycles-pp.start_secondary.secondary_startup_64_no_verify
21.00 -3.8 17.25 ± 7% perf-profile.calltrace.cycles-pp.cpuidle_idle_call.do_idle.cpu_startup_entry.start_secondary.secondary_startup_64_no_verify
20.70 -3.7 17.00 ± 7% perf-profile.calltrace.cycles-pp.cpuidle_enter.cpuidle_idle_call.do_idle.cpu_startup_entry.start_secondary
20.69 -3.7 16.99 ± 7% perf-profile.calltrace.cycles-pp.cpuidle_enter_state.cpuidle_enter.cpuidle_idle_call.do_idle.cpu_startup_entry
20.64 -3.7 16.95 ± 7% perf-profile.calltrace.cycles-pp.acpi_idle_enter.cpuidle_enter_state.cpuidle_enter.cpuidle_idle_call.do_idle
9.51 ± 3% -1.9 7.57 ± 2% perf-profile.calltrace.cycles-pp.do_rw_once
4.54 -1.4 3.19 perf-profile.calltrace.cycles-pp.filemap_map_pages.do_read_fault.do_fault.__handle_mm_fault.handle_mm_fault
2.83 -0.9 1.96 perf-profile.calltrace.cycles-pp.next_uptodate_folio.filemap_map_pages.do_read_fault.do_fault.__handle_mm_fault
3.90 -0.6 3.34 ± 5% perf-profile.calltrace.cycles-pp.clear_page_erms.shmem_get_folio_gfp.shmem_fault.__do_fault.do_read_fault
4.44 ± 6% -0.5 3.98 ± 3% perf-profile.calltrace.cycles-pp.get_mem_cgroup_from_mm.__mem_cgroup_charge.shmem_alloc_and_add_folio.shmem_get_folio_gfp.shmem_fault
1.17 ± 3% -0.4 0.73 ± 6% perf-profile.calltrace.cycles-pp.shmem_inode_acct_blocks.shmem_alloc_and_add_folio.shmem_get_folio_gfp.shmem_fault.__do_fault
1.42 ± 2% -0.4 0.99 ± 2% perf-profile.calltrace.cycles-pp.shmem_alloc_folio.shmem_alloc_and_add_folio.shmem_get_folio_gfp.shmem_fault.__do_fault
1.32 ± 2% -0.4 0.91 perf-profile.calltrace.cycles-pp.alloc_pages_mpol.shmem_alloc_folio.shmem_alloc_and_add_folio.shmem_get_folio_gfp.shmem_fault
1.19 ± 2% -0.4 0.82 perf-profile.calltrace.cycles-pp.__alloc_pages.alloc_pages_mpol.shmem_alloc_folio.shmem_alloc_and_add_folio.shmem_get_folio_gfp
0.96 ± 2% -0.3 0.65 ± 2% perf-profile.calltrace.cycles-pp.get_page_from_freelist.__alloc_pages.alloc_pages_mpol.shmem_alloc_folio.shmem_alloc_and_add_folio
0.98 ± 2% -0.3 0.68 ± 4% perf-profile.calltrace.cycles-pp.sync_regs.asm_exc_page_fault.do_access
1.21 +0.5 1.69 ± 5% perf-profile.calltrace.cycles-pp.__munmap
1.21 +0.5 1.69 ± 5% perf-profile.calltrace.cycles-pp.__vm_munmap.__x64_sys_munmap.do_syscall_64.entry_SYSCALL_64_after_hwframe.__munmap
1.21 +0.5 1.69 ± 5% perf-profile.calltrace.cycles-pp.__x64_sys_munmap.do_syscall_64.entry_SYSCALL_64_after_hwframe.__munmap
1.21 +0.5 1.69 ± 5% perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.__munmap
1.21 +0.5 1.69 ± 5% perf-profile.calltrace.cycles-pp.do_vmi_align_munmap.do_vmi_munmap.__vm_munmap.__x64_sys_munmap.do_syscall_64
1.21 +0.5 1.69 ± 5% perf-profile.calltrace.cycles-pp.do_vmi_munmap.__vm_munmap.__x64_sys_munmap.do_syscall_64.entry_SYSCALL_64_after_hwframe
1.21 +0.5 1.69 ± 5% perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.__munmap
1.20 +0.5 1.68 ± 5% perf-profile.calltrace.cycles-pp.unmap_page_range.unmap_vmas.unmap_region.do_vmi_align_munmap.do_vmi_munmap
1.20 +0.5 1.68 ± 5% perf-profile.calltrace.cycles-pp.unmap_vmas.unmap_region.do_vmi_align_munmap.do_vmi_munmap.__vm_munmap
1.20 +0.5 1.68 ± 5% perf-profile.calltrace.cycles-pp.zap_pmd_range.unmap_page_range.unmap_vmas.unmap_region.do_vmi_align_munmap
1.20 +0.5 1.69 ± 5% perf-profile.calltrace.cycles-pp.unmap_region.do_vmi_align_munmap.do_vmi_munmap.__vm_munmap.__x64_sys_munmap
1.18 +0.5 1.67 ± 6% perf-profile.calltrace.cycles-pp.zap_pte_range.zap_pmd_range.unmap_page_range.unmap_vmas.unmap_region
0.84 ± 2% +0.6 1.43 ± 5% perf-profile.calltrace.cycles-pp.lru_add_fn.folio_batch_move_lru.folio_add_lru.shmem_alloc_and_add_folio.shmem_get_folio_gfp
0.58 ± 3% +0.6 1.18 ± 5% perf-profile.calltrace.cycles-pp.page_remove_rmap.zap_pte_range.zap_pmd_range.unmap_page_range.unmap_vmas
0.00 +0.8 0.79 ± 4% perf-profile.calltrace.cycles-pp.__mod_memcg_lruvec_state.__mod_lruvec_page_state.page_remove_rmap.zap_pte_range.zap_pmd_range
0.00 +1.0 1.02 ± 5% perf-profile.calltrace.cycles-pp.__mod_memcg_lruvec_state.lru_add_fn.folio_batch_move_lru.folio_add_lru.shmem_alloc_and_add_folio
0.00 +1.1 1.08 ± 4% perf-profile.calltrace.cycles-pp.__mod_lruvec_page_state.page_remove_rmap.zap_pte_range.zap_pmd_range.unmap_page_range
0.00 +1.5 1.46 ± 5% perf-profile.calltrace.cycles-pp.__count_memcg_events.handle_mm_fault.do_user_addr_fault.exc_page_fault.asm_exc_page_fault
3.29 ± 3% +1.9 5.19 perf-profile.calltrace.cycles-pp.finish_fault.do_read_fault.do_fault.__handle_mm_fault.handle_mm_fault
3.02 ± 4% +2.0 5.00 perf-profile.calltrace.cycles-pp.set_pte_range.finish_fault.do_read_fault.do_fault.__handle_mm_fault
2.84 ± 4% +2.0 4.86 perf-profile.calltrace.cycles-pp.folio_add_file_rmap_range.set_pte_range.finish_fault.do_read_fault.do_fault
2.73 ± 4% +2.0 4.77 perf-profile.calltrace.cycles-pp.__mod_lruvec_page_state.folio_add_file_rmap_range.set_pte_range.finish_fault.do_read_fault
1.48 ± 4% +2.1 3.56 ± 2% perf-profile.calltrace.cycles-pp.__mod_memcg_lruvec_state.__mod_lruvec_page_state.folio_add_file_rmap_range.set_pte_range.finish_fault
0.57 ± 4% +2.8 3.35 ± 2% perf-profile.calltrace.cycles-pp.__count_memcg_events.mem_cgroup_commit_charge.__mem_cgroup_charge.shmem_alloc_and_add_folio.shmem_get_folio_gfp
1.96 ± 5% +2.9 4.86 ± 2% perf-profile.calltrace.cycles-pp.mem_cgroup_commit_charge.__mem_cgroup_charge.shmem_alloc_and_add_folio.shmem_get_folio_gfp.shmem_fault
3.65 ± 2% +3.1 6.77 ± 2% perf-profile.calltrace.cycles-pp.shmem_add_to_page_cache.shmem_alloc_and_add_folio.shmem_get_folio_gfp.shmem_fault.__do_fault
0.80 ± 4% +3.1 3.92 ± 3% perf-profile.calltrace.cycles-pp.__mod_memcg_lruvec_state.__mod_lruvec_page_state.shmem_add_to_page_cache.shmem_alloc_and_add_folio.shmem_get_folio_gfp
2.68 ± 3% +3.4 6.08 ± 2% perf-profile.calltrace.cycles-pp.__mod_lruvec_page_state.shmem_add_to_page_cache.shmem_alloc_and_add_folio.shmem_get_folio_gfp.shmem_fault
7.71 ± 6% +3.9 11.66 ± 2% perf-profile.calltrace.cycles-pp.__mem_cgroup_charge.shmem_alloc_and_add_folio.shmem_get_folio_gfp.shmem_fault.__do_fault
67.18 +6.3 73.46 ± 3% perf-profile.calltrace.cycles-pp.do_access
1.46 ± 9% +7.1 8.57 ± 16% perf-profile.calltrace.cycles-pp._raw_spin_lock_irqsave.folio_lruvec_lock_irqsave.folio_batch_move_lru.folio_add_lru.shmem_alloc_and_add_folio
1.50 ± 9% +7.1 8.61 ± 16% perf-profile.calltrace.cycles-pp.folio_lruvec_lock_irqsave.folio_batch_move_lru.folio_add_lru.shmem_alloc_and_add_folio.shmem_get_folio_gfp
1.38 ± 10% +7.1 8.51 ± 16% perf-profile.calltrace.cycles-pp.native_queued_spin_lock_slowpath._raw_spin_lock_irqsave.folio_lruvec_lock_irqsave.folio_batch_move_lru.folio_add_lru
51.46 +7.6 59.08 ± 3% perf-profile.calltrace.cycles-pp.asm_exc_page_fault.do_access
2.98 ± 5% +7.7 10.66 ± 14% perf-profile.calltrace.cycles-pp.folio_add_lru.shmem_alloc_and_add_folio.shmem_get_folio_gfp.shmem_fault.__do_fault
2.84 ± 6% +7.7 10.56 ± 14% perf-profile.calltrace.cycles-pp.folio_batch_move_lru.folio_add_lru.shmem_alloc_and_add_folio.shmem_get_folio_gfp.shmem_fault
34.18 +8.5 42.68 ± 4% perf-profile.calltrace.cycles-pp.__do_fault.do_read_fault.do_fault.__handle_mm_fault.handle_mm_fault
34.14 +8.5 42.64 ± 4% perf-profile.calltrace.cycles-pp.shmem_fault.__do_fault.do_read_fault.do_fault.__handle_mm_fault
33.95 +8.6 42.51 ± 4% perf-profile.calltrace.cycles-pp.shmem_get_folio_gfp.shmem_fault.__do_fault.do_read_fault.do_fault
42.88 +8.8 51.70 ± 4% perf-profile.calltrace.cycles-pp.__handle_mm_fault.handle_mm_fault.do_user_addr_fault.exc_page_fault.asm_exc_page_fault
42.34 +9.0 51.30 ± 4% perf-profile.calltrace.cycles-pp.do_fault.__handle_mm_fault.handle_mm_fault.do_user_addr_fault.exc_page_fault
42.29 +9.0 51.28 ± 4% perf-profile.calltrace.cycles-pp.do_read_fault.do_fault.__handle_mm_fault.handle_mm_fault.do_user_addr_fault
45.07 +9.6 54.62 ± 4% perf-profile.calltrace.cycles-pp.exc_page_fault.asm_exc_page_fault.do_access
44.95 +9.6 54.53 ± 4% perf-profile.calltrace.cycles-pp.do_user_addr_fault.exc_page_fault.asm_exc_page_fault.do_access
43.72 +9.9 53.64 ± 4% perf-profile.calltrace.cycles-pp.handle_mm_fault.do_user_addr_fault.exc_page_fault.asm_exc_page_fault.do_access
17.28 ± 2% +13.8 31.05 ± 6% perf-profile.calltrace.cycles-pp.shmem_alloc_and_add_folio.shmem_get_folio_gfp.shmem_fault.__do_fault.do_read_fault
21.14 -3.8 17.36 ± 7% perf-profile.children.cycles-pp.cpu_startup_entry
21.14 -3.8 17.36 ± 7% perf-profile.children.cycles-pp.do_idle
21.14 -3.8 17.36 ± 7% perf-profile.children.cycles-pp.secondary_startup_64_no_verify
21.09 -3.8 17.33 ± 7% perf-profile.children.cycles-pp.cpuidle_idle_call
21.05 -3.8 17.29 ± 7% perf-profile.children.cycles-pp.start_secondary
20.79 -3.7 17.07 ± 7% perf-profile.children.cycles-pp.cpuidle_enter
20.78 -3.7 17.07 ± 7% perf-profile.children.cycles-pp.cpuidle_enter_state
20.72 -3.7 17.02 ± 7% perf-profile.children.cycles-pp.acpi_idle_enter
20.71 -3.7 17.01 ± 7% perf-profile.children.cycles-pp.acpi_safe_halt
20.79 -3.6 17.19 ± 6% perf-profile.children.cycles-pp.asm_sysvec_apic_timer_interrupt
11.52 -3.1 8.42 perf-profile.children.cycles-pp.do_rw_once
4.62 -1.4 3.24 perf-profile.children.cycles-pp.filemap_map_pages
2.89 -0.9 2.00 perf-profile.children.cycles-pp.next_uptodate_folio
3.98 -0.6 3.39 ± 5% perf-profile.children.cycles-pp.clear_page_erms
4.46 ± 6% -0.5 3.99 ± 3% perf-profile.children.cycles-pp.get_mem_cgroup_from_mm
1.18 ± 4% -0.4 0.74 ± 6% perf-profile.children.cycles-pp.shmem_inode_acct_blocks
1.44 ± 2% -0.4 1.00 ± 2% perf-profile.children.cycles-pp.shmem_alloc_folio
1.40 -0.4 0.99 perf-profile.children.cycles-pp.alloc_pages_mpol
1.27 -0.4 0.90 perf-profile.children.cycles-pp.__alloc_pages
1.01 ± 2% -0.3 0.68 perf-profile.children.cycles-pp.get_page_from_freelist
1.02 ± 2% -0.3 0.70 ± 4% perf-profile.children.cycles-pp.sync_regs
0.77 ± 2% -0.3 0.51 perf-profile.children.cycles-pp.rmqueue
0.81 ± 2% -0.2 0.60 perf-profile.children.cycles-pp.__perf_sw_event
0.53 ± 3% -0.2 0.34 ± 2% perf-profile.children.cycles-pp.__rmqueue_pcplist
0.68 ± 2% -0.2 0.50 ± 5% perf-profile.children.cycles-pp.__mod_lruvec_state
0.65 ± 6% -0.2 0.47 ± 2% perf-profile.children.cycles-pp._raw_spin_lock
0.47 ± 3% -0.2 0.29 ± 2% perf-profile.children.cycles-pp.rmqueue_bulk
0.65 ± 2% -0.2 0.49 perf-profile.children.cycles-pp.___perf_sw_event
0.64 ± 4% -0.1 0.49 ± 5% perf-profile.children.cycles-pp.xas_load
0.54 -0.1 0.39 ± 4% perf-profile.children.cycles-pp.__mod_node_page_state
0.49 ± 2% -0.1 0.35 ± 3% perf-profile.children.cycles-pp.lock_vma_under_rcu
0.54 ± 5% -0.1 0.40 ± 2% perf-profile.children.cycles-pp.xas_find
0.39 ± 4% -0.1 0.28 ± 3% perf-profile.children.cycles-pp.__pte_offset_map_lock
0.39 ± 3% -0.1 0.29 ± 3% perf-profile.children.cycles-pp.xas_descend
0.32 ± 4% -0.1 0.22 ± 8% perf-profile.children.cycles-pp.__dquot_alloc_space
0.30 ± 3% -0.1 0.22 ± 3% perf-profile.children.cycles-pp.mas_walk
0.20 ± 13% -0.1 0.13 ± 5% perf-profile.children.cycles-pp.shmem_recalc_inode
0.26 ± 2% -0.1 0.19 ± 3% perf-profile.children.cycles-pp.filemap_get_entry
0.18 ± 5% -0.1 0.12 ± 5% perf-profile.children.cycles-pp.xas_find_conflict
0.28 ± 4% -0.1 0.22 ± 8% perf-profile.children.cycles-pp.__x64_sys_execve
0.28 ± 4% -0.1 0.22 ± 8% perf-profile.children.cycles-pp.do_execveat_common
0.28 ± 4% -0.1 0.22 ± 8% perf-profile.children.cycles-pp.execve
0.29 ± 3% -0.1 0.24 ± 8% perf-profile.children.cycles-pp.asm_sysvec_call_function_single
0.16 ± 5% -0.1 0.11 ± 8% perf-profile.children.cycles-pp.error_entry
0.14 ± 5% -0.0 0.09 ± 8% perf-profile.children.cycles-pp.__percpu_counter_limited_add
0.15 ± 5% -0.0 0.10 ± 10% perf-profile.children.cycles-pp.inode_add_bytes
0.07 ± 6% -0.0 0.02 ± 99% perf-profile.children.cycles-pp.__folio_throttle_swaprate
0.10 -0.0 0.06 ± 13% perf-profile.children.cycles-pp.security_vm_enough_memory_mm
0.18 ± 7% -0.0 0.14 ± 13% perf-profile.children.cycles-pp._raw_spin_lock_irq
0.16 ± 5% -0.0 0.12 perf-profile.children.cycles-pp.handle_pte_fault
0.17 ± 7% -0.0 0.12 ± 4% perf-profile.children.cycles-pp.xas_start
0.14 ± 6% -0.0 0.10 ± 3% perf-profile.children.cycles-pp.__pte_offset_map
0.07 ± 5% -0.0 0.03 ± 70% perf-profile.children.cycles-pp.policy_nodemask
0.16 ± 4% -0.0 0.13 ± 12% perf-profile.children.cycles-pp.folio_mark_accessed
0.19 ± 4% -0.0 0.16 ± 8% perf-profile.children.cycles-pp.bprm_execve
0.11 ± 9% -0.0 0.08 ± 6% perf-profile.children.cycles-pp.down_read_trylock
0.16 ± 6% -0.0 0.13 ± 5% perf-profile.children.cycles-pp.percpu_counter_add_batch
0.11 ± 6% -0.0 0.08 ± 6% perf-profile.children.cycles-pp.up_read
0.15 ± 7% -0.0 0.12 ± 13% perf-profile.children.cycles-pp.folio_unlock
0.10 ± 4% -0.0 0.07 ± 6% perf-profile.children.cycles-pp.__libc_fork
0.07 ± 6% -0.0 0.04 ± 45% perf-profile.children.cycles-pp.ksys_read
0.10 ± 3% -0.0 0.07 ± 11% perf-profile.children.cycles-pp.kernel_clone
0.09 ± 5% -0.0 0.06 ± 7% perf-profile.children.cycles-pp.mem_cgroup_update_lru_size
0.09 ± 5% -0.0 0.06 ± 11% perf-profile.children.cycles-pp.__x64_sys_openat
0.08 ± 8% -0.0 0.06 ± 11% perf-profile.children.cycles-pp.do_filp_open
0.08 ± 8% -0.0 0.06 ± 11% perf-profile.children.cycles-pp.path_openat
0.07 -0.0 0.04 ± 45% perf-profile.children.cycles-pp.vfs_read
0.09 ± 4% -0.0 0.06 ± 7% perf-profile.children.cycles-pp.__do_sys_clone
0.10 ± 6% -0.0 0.08 ± 6% perf-profile.children.cycles-pp.pte_offset_map_nolock
0.08 ± 8% -0.0 0.06 ± 11% perf-profile.children.cycles-pp.do_sys_openat2
0.07 ± 5% -0.0 0.04 ± 45% perf-profile.children.cycles-pp.copy_process
0.16 ± 5% -0.0 0.14 ± 6% perf-profile.children.cycles-pp.exec_binprm
0.10 ± 6% -0.0 0.08 ± 7% perf-profile.children.cycles-pp.__vm_enough_memory
0.16 ± 4% -0.0 0.14 ± 6% perf-profile.children.cycles-pp.search_binary_handler
0.08 -0.0 0.06 ± 9% perf-profile.children.cycles-pp.__irqentry_text_end
0.09 ± 5% -0.0 0.07 ± 7% perf-profile.children.cycles-pp._compound_head
0.15 ± 5% -0.0 0.13 ± 7% perf-profile.children.cycles-pp.xas_create
0.15 ± 4% -0.0 0.14 ± 8% perf-profile.children.cycles-pp.load_elf_binary
0.12 ± 4% -0.0 0.10 ± 3% perf-profile.children.cycles-pp.kmem_cache_alloc_lru
0.05 ± 8% +0.0 0.08 ± 8% perf-profile.children.cycles-pp.propagate_protected_usage
0.25 ± 2% +0.0 0.30 ± 4% perf-profile.children.cycles-pp.page_counter_try_charge
0.02 ±141% +0.0 0.06 ± 7% perf-profile.children.cycles-pp.mod_objcg_state
0.00 +0.1 0.07 ± 14% perf-profile.children.cycles-pp.tlb_finish_mmu
1.25 +0.5 1.72 ± 5% perf-profile.children.cycles-pp.unmap_vmas
1.24 +0.5 1.71 ± 5% perf-profile.children.cycles-pp.zap_pte_range
1.24 +0.5 1.71 ± 5% perf-profile.children.cycles-pp.unmap_page_range
1.24 +0.5 1.71 ± 5% perf-profile.children.cycles-pp.zap_pmd_range
1.21 +0.5 1.69 ± 5% perf-profile.children.cycles-pp.__munmap
1.22 +0.5 1.71 ± 5% perf-profile.children.cycles-pp.__vm_munmap
1.21 +0.5 1.70 ± 5% perf-profile.children.cycles-pp.__x64_sys_munmap
1.25 +0.5 1.74 ± 5% perf-profile.children.cycles-pp.do_vmi_align_munmap
1.25 +0.5 1.74 ± 5% perf-profile.children.cycles-pp.do_vmi_munmap
1.22 +0.5 1.72 ± 5% perf-profile.children.cycles-pp.unmap_region
0.85 ± 2% +0.6 1.44 ± 5% perf-profile.children.cycles-pp.lru_add_fn
0.60 ± 3% +0.6 1.20 ± 4% perf-profile.children.cycles-pp.page_remove_rmap
3.30 ± 3% +1.9 5.20 perf-profile.children.cycles-pp.finish_fault
3.04 ± 4% +2.0 5.01 perf-profile.children.cycles-pp.set_pte_range
2.85 ± 4% +2.0 4.87 perf-profile.children.cycles-pp.folio_add_file_rmap_range
1.97 ± 5% +2.9 4.88 ± 2% perf-profile.children.cycles-pp.mem_cgroup_commit_charge
3.69 ± 2% +3.1 6.80 ± 2% perf-profile.children.cycles-pp.shmem_add_to_page_cache
7.74 ± 6% +3.9 11.69 ± 2% perf-profile.children.cycles-pp.__mem_cgroup_charge
0.80 ± 4% +4.0 4.85 ± 3% perf-profile.children.cycles-pp.__count_memcg_events
6.12 ± 3% +6.1 12.18 perf-profile.children.cycles-pp.__mod_lruvec_page_state
2.99 ± 3% +6.6 9.56 ± 2% perf-profile.children.cycles-pp.__mod_memcg_lruvec_state
61.44 +6.7 68.11 ± 3% perf-profile.children.cycles-pp.do_access
1.58 ± 9% +7.1 8.72 ± 16% perf-profile.children.cycles-pp._raw_spin_lock_irqsave
1.45 ± 9% +7.2 8.63 ± 16% perf-profile.children.cycles-pp.native_queued_spin_lock_slowpath
1.53 ± 9% +7.2 8.72 ± 16% perf-profile.children.cycles-pp.folio_lruvec_lock_irqsave
2.98 ± 5% +7.7 10.67 ± 14% perf-profile.children.cycles-pp.folio_add_lru
2.86 ± 6% +7.8 10.63 ± 14% perf-profile.children.cycles-pp.folio_batch_move_lru
49.12 +8.3 57.47 ± 3% perf-profile.children.cycles-pp.asm_exc_page_fault
34.19 +8.5 42.68 ± 4% perf-profile.children.cycles-pp.__do_fault
34.15 +8.5 42.65 ± 4% perf-profile.children.cycles-pp.shmem_fault
33.99 +8.6 42.54 ± 4% perf-profile.children.cycles-pp.shmem_get_folio_gfp
43.06 +8.8 51.84 ± 4% perf-profile.children.cycles-pp.__handle_mm_fault
42.43 +8.9 51.37 ± 4% perf-profile.children.cycles-pp.do_fault
42.38 +9.0 51.34 ± 4% perf-profile.children.cycles-pp.do_read_fault
45.26 +9.5 54.78 ± 4% perf-profile.children.cycles-pp.exc_page_fault
45.15 +9.5 54.69 ± 4% perf-profile.children.cycles-pp.do_user_addr_fault
43.91 +9.9 53.80 ± 4% perf-profile.children.cycles-pp.handle_mm_fault
17.31 ± 2% +13.8 31.07 ± 5% perf-profile.children.cycles-pp.shmem_alloc_and_add_folio
12.24 -4.5 7.76 ± 3% perf-profile.self.cycles-pp.shmem_get_folio_gfp
17.96 -3.3 14.66 ± 4% perf-profile.self.cycles-pp.acpi_safe_halt
10.95 -3.2 7.74 perf-profile.self.cycles-pp.do_rw_once
5.96 -1.4 4.58 ± 2% perf-profile.self.cycles-pp.do_access
2.40 -0.8 1.64 perf-profile.self.cycles-pp.next_uptodate_folio
3.92 -0.6 3.36 ± 5% perf-profile.self.cycles-pp.clear_page_erms
4.40 ± 6% -0.5 3.95 ± 3% perf-profile.self.cycles-pp.get_mem_cgroup_from_mm
1.52 ± 2% -0.4 1.10 ± 2% perf-profile.self.cycles-pp.filemap_map_pages
1.02 ± 2% -0.3 0.70 ± 4% perf-profile.self.cycles-pp.sync_regs
0.50 ± 7% -0.2 0.27 ± 5% perf-profile.self.cycles-pp.shmem_inode_acct_blocks
0.63 ± 5% -0.2 0.46 ± 3% perf-profile.self.cycles-pp._raw_spin_lock
0.42 ± 2% -0.1 0.27 ± 2% perf-profile.self.cycles-pp.rmqueue_bulk
0.52 -0.1 0.38 ± 4% perf-profile.self.cycles-pp.__mod_node_page_state
0.56 ± 2% -0.1 0.42 perf-profile.self.cycles-pp.___perf_sw_event
0.31 ± 3% -0.1 0.20 ± 2% perf-profile.self.cycles-pp.shmem_add_to_page_cache
0.38 ± 4% -0.1 0.28 perf-profile.self.cycles-pp.__handle_mm_fault
0.36 ± 4% -0.1 0.26 ± 2% perf-profile.self.cycles-pp.xas_descend
0.30 ± 2% -0.1 0.22 ± 2% perf-profile.self.cycles-pp.mas_walk
0.33 ± 3% -0.1 0.26 ± 10% perf-profile.self.cycles-pp.lru_add_fn
0.20 ± 3% -0.1 0.14 ± 5% perf-profile.self.cycles-pp.asm_exc_page_fault
0.21 ± 5% -0.1 0.15 ± 6% perf-profile.self.cycles-pp.get_page_from_freelist
0.26 ± 9% -0.1 0.20 ± 15% perf-profile.self.cycles-pp.xas_store
0.16 ± 7% -0.1 0.11 ± 6% perf-profile.self.cycles-pp.__perf_sw_event
0.18 ± 2% -0.1 0.13 ± 5% perf-profile.self.cycles-pp.__alloc_pages
0.22 ± 4% -0.1 0.17 ± 4% perf-profile.self.cycles-pp.handle_mm_fault
0.20 ± 8% -0.1 0.14 ± 5% perf-profile.self.cycles-pp.xas_find
0.15 ± 6% -0.0 0.10 ± 7% perf-profile.self.cycles-pp.error_entry
0.17 ± 2% -0.0 0.12 ± 6% perf-profile.self.cycles-pp.__dquot_alloc_space
0.17 ± 6% -0.0 0.13 ± 10% perf-profile.self.cycles-pp._raw_spin_lock_irq
0.22 ± 4% -0.0 0.18 ± 9% perf-profile.self.cycles-pp.xas_load
0.23 ± 4% -0.0 0.19 ± 10% perf-profile.self.cycles-pp.zap_pte_range
0.12 ± 7% -0.0 0.08 ± 10% perf-profile.self.cycles-pp.__percpu_counter_limited_add
0.14 ± 3% -0.0 0.09 ± 7% perf-profile.self.cycles-pp.rmqueue
0.15 ± 6% -0.0 0.10 ± 9% perf-profile.self.cycles-pp.__mod_lruvec_state
0.15 ± 2% -0.0 0.11 ± 6% perf-profile.self.cycles-pp.do_user_addr_fault
0.12 ± 7% -0.0 0.08 ± 5% perf-profile.self.cycles-pp.folio_add_lru
0.16 ± 7% -0.0 0.12 ± 4% perf-profile.self.cycles-pp.xas_start
0.06 ± 7% -0.0 0.02 ± 99% perf-profile.self.cycles-pp.finish_fault
0.16 ± 4% -0.0 0.12 ± 12% perf-profile.self.cycles-pp.folio_mark_accessed
0.11 ± 8% -0.0 0.08 ± 6% perf-profile.self.cycles-pp.__pte_offset_map_lock
0.13 ± 6% -0.0 0.09 ± 4% perf-profile.self.cycles-pp.__pte_offset_map
0.11 ± 9% -0.0 0.08 ± 6% perf-profile.self.cycles-pp.down_read_trylock
0.12 ± 3% -0.0 0.09 ± 5% perf-profile.self.cycles-pp.do_read_fault
0.09 ± 5% -0.0 0.06 ± 7% perf-profile.self.cycles-pp.mem_cgroup_update_lru_size
0.16 ± 4% -0.0 0.12 ± 4% perf-profile.self.cycles-pp.percpu_counter_add_batch
0.11 ± 6% -0.0 0.08 ± 5% perf-profile.self.cycles-pp.shmem_alloc_and_add_folio
0.08 ± 8% -0.0 0.05 perf-profile.self.cycles-pp.xas_find_conflict
0.12 ± 4% -0.0 0.09 ± 7% perf-profile.self.cycles-pp.folio_add_file_rmap_range
0.10 ± 6% -0.0 0.07 ± 6% perf-profile.self.cycles-pp.up_read
0.12 ± 4% -0.0 0.10 ± 6% perf-profile.self.cycles-pp._raw_spin_lock_irqsave
0.09 ± 4% -0.0 0.07 ± 7% perf-profile.self.cycles-pp.exc_page_fault
0.13 ± 6% -0.0 0.10 ± 9% perf-profile.self.cycles-pp.page_remove_rmap
0.08 -0.0 0.06 ± 8% perf-profile.self.cycles-pp.__irqentry_text_end
0.19 ± 5% -0.0 0.17 ± 5% perf-profile.self.cycles-pp.cgroup_rstat_updated
0.09 ± 6% -0.0 0.07 ± 5% perf-profile.self.cycles-pp.set_pte_range
0.07 ± 5% -0.0 0.05 ± 7% perf-profile.self.cycles-pp._compound_head
0.08 -0.0 0.06 ± 9% perf-profile.self.cycles-pp.lock_vma_under_rcu
0.05 ± 8% +0.0 0.08 ± 8% perf-profile.self.cycles-pp.propagate_protected_usage
2.93 ± 4% +0.4 3.35 ± 3% perf-profile.self.cycles-pp.__mod_lruvec_page_state
0.77 ± 7% +1.5 2.23 ± 3% perf-profile.self.cycles-pp.__mem_cgroup_charge
0.75 ± 4% +4.0 4.80 ± 3% perf-profile.self.cycles-pp.__count_memcg_events
2.83 ± 3% +6.6 9.40 ± 2% perf-profile.self.cycles-pp.__mod_memcg_lruvec_state
1.45 ± 9% +7.2 8.63 ± 16% perf-profile.self.cycles-pp.native_queued_spin_lock_slowpath
***************************************************************************************************
lkp-skl-fpga01: 104 threads 2 sockets (Skylake) with 192G memory
=========================================================================================
compiler/cpufreq_governor/kconfig/mode/nr_task/rootfs/tbox_group/test/testcase:
gcc-12/performance/x86_64-rhel-8.3/process/50%/debian-11.1-x86_64-20220510.cgz/lkp-skl-fpga01/tlb_flush2/will-it-scale
commit:
e0bf1dc859 ("mm: memcg: move vmstats structs definition above flushing code")
8d59d2214c ("mm: memcg: make stats flushing threshold per-memcg")
e0bf1dc859fdd08e 8d59d2214c2362e7a9d185d80b6
---------------- ---------------------------
%stddev %change %stddev
\ | \
4.05 -1.2 2.81 mpstat.cpu.all.usr%
193.83 ± 6% +69.3% 328.17 ± 8% perf-c2c.DRAM.local
1216 ± 8% +27.1% 1546 ± 6% perf-c2c.DRAM.remote
150.33 ± 13% -40.0% 90.17 ± 13% perf-c2c.HITM.remote
0.04 -25.0% 0.03 turbostat.IPC
316.16 -1.5% 311.47 turbostat.PkgWatt
30.54 +4.9% 32.04 turbostat.RAMWatt
2132437 -32.3% 1444430 will-it-scale.52.processes
41008 -32.3% 27776 will-it-scale.per_process_ops
2132437 -32.3% 1444430 will-it-scale.workload
3.113e+08 ± 3% -31.7% 2.125e+08 ± 4% numa-numastat.node0.local_node
3.114e+08 ± 3% -31.7% 2.126e+08 ± 4% numa-numastat.node0.numa_hit
3.322e+08 ± 2% -32.5% 2.243e+08 ± 3% numa-numastat.node1.local_node
3.323e+08 ± 2% -32.5% 2.243e+08 ± 3% numa-numastat.node1.numa_hit
3.114e+08 ± 3% -31.7% 2.126e+08 ± 4% numa-vmstat.node0.numa_hit
3.113e+08 ± 3% -31.7% 2.125e+08 ± 4% numa-vmstat.node0.numa_local
3.323e+08 ± 2% -32.5% 2.243e+08 ± 3% numa-vmstat.node1.numa_hit
3.322e+08 ± 2% -32.5% 2.243e+08 ± 3% numa-vmstat.node1.numa_local
0.00 ± 19% -61.1% 0.00 ± 31% perf-sched.sch_delay.avg.ms.do_task_dead.do_exit.do_group_exit.__x64_sys_exit_group.do_syscall_64
217.07 ± 11% -46.4% 116.39 ± 23% perf-sched.wait_and_delay.avg.ms.do_task_dead.do_exit.do_group_exit.__x64_sys_exit_group.do_syscall_64
218.50 ± 6% +19.1% 260.33 ± 4% perf-sched.wait_and_delay.count.__cond_resched.zap_page_range_single.madvise_vma_behavior.do_madvise.__x64_sys_madvise
217.06 ± 11% -46.4% 116.38 ± 23% perf-sched.wait_time.avg.ms.do_task_dead.do_exit.do_group_exit.__x64_sys_exit_group.do_syscall_64
6.436e+08 -32.1% 4.369e+08 proc-vmstat.numa_hit
6.435e+08 -32.1% 4.368e+08 proc-vmstat.numa_local
6.432e+08 -32.1% 4.368e+08 proc-vmstat.pgalloc_normal
1.286e+09 -32.1% 8.726e+08 proc-vmstat.pgfault
6.432e+08 -32.1% 4.367e+08 proc-vmstat.pgfree
170696 ± 8% +3.4% 176515 ± 8% sched_debug.cpu.clock.avg
170703 ± 8% +3.4% 176522 ± 8% sched_debug.cpu.clock.max
170689 ± 8% +3.4% 176508 ± 8% sched_debug.cpu.clock.min
169431 ± 8% +3.4% 175248 ± 8% sched_debug.cpu.clock_task.avg
169630 ± 8% +3.4% 175429 ± 8% sched_debug.cpu.clock_task.max
162542 ± 8% +3.5% 168260 ± 8% sched_debug.cpu.clock_task.min
170690 ± 8% +3.4% 176508 ± 8% sched_debug.cpu_clk
170117 ± 8% +3.4% 175938 ± 8% sched_debug.ktime
171259 ± 8% +3.4% 177078 ± 8% sched_debug.sched_clk
4.06 +80.8% 7.34 perf-stat.i.MPKI
4.066e+09 -23.3% 3.12e+09 perf-stat.i.branch-instructions
0.57 -0.0 0.55 perf-stat.i.branch-miss-rate%
23478297 -25.0% 17605102 perf-stat.i.branch-misses
17.25 +7.0 24.27 perf-stat.i.cache-miss-rate%
82715093 ± 2% +35.9% 1.124e+08 perf-stat.i.cache-misses
4.795e+08 ± 2% -3.4% 4.63e+08 perf-stat.i.cache-references
7.14 +32.9% 9.49 perf-stat.i.cpi
134.85 -1.2% 133.29 perf-stat.i.cpu-migrations
1760 ± 2% -26.5% 1294 perf-stat.i.cycles-between-cache-misses
0.26 -0.0 0.24 perf-stat.i.dTLB-load-miss-rate%
13461491 -31.7% 9190211 perf-stat.i.dTLB-load-misses
5.141e+09 -24.1% 3.902e+09 perf-stat.i.dTLB-loads
0.45 -0.0 0.44 perf-stat.i.dTLB-store-miss-rate%
12934403 -32.2% 8773143 perf-stat.i.dTLB-store-misses
2.841e+09 -29.9% 1.992e+09 perf-stat.i.dTLB-stores
14.76 +1.4 16.18 ± 4% perf-stat.i.iTLB-load-miss-rate%
7454399 ± 2% -22.7% 5760387 ± 4% perf-stat.i.iTLB-load-misses
43026423 -30.6% 29840650 perf-stat.i.iTLB-loads
2.042e+10 -24.7% 1.538e+10 perf-stat.i.instructions
0.14 -24.6% 0.11 perf-stat.i.ipc
815.65 -20.2% 651.03 perf-stat.i.metric.K/sec
120.43 -24.3% 91.11 perf-stat.i.metric.M/sec
4264808 -32.2% 2892980 perf-stat.i.minor-faults
11007315 ± 2% +39.7% 15375516 perf-stat.i.node-load-misses
1459152 ± 6% +45.1% 2116827 ± 5% perf-stat.i.node-loads
7872989 ± 2% -26.2% 5812458 perf-stat.i.node-store-misses
4264808 -32.2% 2892980 perf-stat.i.page-faults
4.05 +80.4% 7.31 perf-stat.overall.MPKI
0.58 -0.0 0.57 perf-stat.overall.branch-miss-rate%
17.25 +7.0 24.27 perf-stat.overall.cache-miss-rate%
7.13 +32.7% 9.46 perf-stat.overall.cpi
1759 ± 2% -26.5% 1294 perf-stat.overall.cycles-between-cache-misses
0.26 -0.0 0.23 perf-stat.overall.dTLB-load-miss-rate%
0.45 -0.0 0.44 perf-stat.overall.dTLB-store-miss-rate%
14.77 +1.4 16.18 ± 4% perf-stat.overall.iTLB-load-miss-rate%
0.14 -24.7% 0.11 perf-stat.overall.ipc
2882666 +11.2% 3206246 perf-stat.overall.path-length
4.052e+09 -23.3% 3.11e+09 perf-stat.ps.branch-instructions
23421504 -25.0% 17574476 perf-stat.ps.branch-misses
82419384 ± 2% +35.9% 1.12e+08 perf-stat.ps.cache-misses
4.778e+08 ± 2% -3.4% 4.614e+08 perf-stat.ps.cache-references
134.44 -1.1% 132.98 perf-stat.ps.cpu-migrations
13415064 -31.7% 9160067 perf-stat.ps.dTLB-load-misses
5.124e+09 -24.1% 3.89e+09 perf-stat.ps.dTLB-loads
12889609 -32.2% 8744145 perf-stat.ps.dTLB-store-misses
2.831e+09 -29.9% 1.986e+09 perf-stat.ps.dTLB-stores
7428050 ± 2% -22.7% 5741276 ± 4% perf-stat.ps.iTLB-load-misses
42877049 -30.6% 29741122 perf-stat.ps.iTLB-loads
2.035e+10 -24.7% 1.533e+10 perf-stat.ps.instructions
4250034 -32.2% 2883410 perf-stat.ps.minor-faults
10968228 ± 2% +39.7% 15322266 perf-stat.ps.node-load-misses
1454274 ± 6% +45.1% 2109746 ± 5% perf-stat.ps.node-loads
7845298 ± 2% -26.2% 5792864 perf-stat.ps.node-store-misses
4250034 -32.2% 2883410 perf-stat.ps.page-faults
6.147e+12 -24.7% 4.631e+12 perf-stat.total.instructions
26.77 -1.8 24.93 ± 3% perf-profile.calltrace.cycles-pp.intel_idle_ibrs.cpuidle_enter_state.cpuidle_enter.cpuidle_idle_call.do_idle
26.75 -1.8 24.92 ± 2% perf-profile.calltrace.cycles-pp.cpuidle_enter.cpuidle_idle_call.do_idle.cpu_startup_entry.start_secondary
26.75 -1.8 24.92 ± 2% perf-profile.calltrace.cycles-pp.cpuidle_idle_call.do_idle.cpu_startup_entry.start_secondary.secondary_startup_64_no_verify
26.84 -1.8 25.00 ± 3% perf-profile.calltrace.cycles-pp.cpuidle_enter_state.cpuidle_enter.cpuidle_idle_call.do_idle.cpu_startup_entry
26.75 -1.8 24.92 ± 2% perf-profile.calltrace.cycles-pp.cpu_startup_entry.start_secondary.secondary_startup_64_no_verify
26.75 -1.8 24.92 ± 2% perf-profile.calltrace.cycles-pp.do_idle.cpu_startup_entry.start_secondary.secondary_startup_64_no_verify
26.75 -1.8 24.92 ± 2% perf-profile.calltrace.cycles-pp.start_secondary.secondary_startup_64_no_verify
27.05 -1.8 25.29 ± 3% perf-profile.calltrace.cycles-pp.secondary_startup_64_no_verify
13.02 ± 2% -1.4 11.60 ± 4% perf-profile.calltrace.cycles-pp.testcase
5.54 ± 5% -1.0 4.52 ± 3% perf-profile.calltrace.cycles-pp.__mem_cgroup_uncharge_list.release_pages.tlb_batch_pages_flush.tlb_finish_mmu.zap_page_range_single
1.37 ± 2% -0.9 0.51 ± 58% perf-profile.calltrace.cycles-pp.asm_exc_page_fault.__madvise
10.38 ± 3% -0.8 9.54 ± 2% perf-profile.calltrace.cycles-pp.asm_exc_page_fault.testcase
2.38 ± 2% -0.8 1.63 ± 3% perf-profile.calltrace.cycles-pp.page_counter_uncharge.uncharge_batch.__mem_cgroup_uncharge_list.release_pages.tlb_batch_pages_flush
4.02 ± 3% -0.7 3.32 ± 3% perf-profile.calltrace.cycles-pp.uncharge_batch.__mem_cgroup_uncharge_list.release_pages.tlb_batch_pages_flush.tlb_finish_mmu
1.92 ± 4% -0.4 1.49 ± 2% perf-profile.calltrace.cycles-pp.irqentry_exit_to_user_mode.asm_exc_page_fault.testcase
1.36 ± 2% -0.4 0.99 perf-profile.calltrace.cycles-pp.__irqentry_text_end.testcase
1.30 ± 10% -0.4 0.94 ± 6% perf-profile.calltrace.cycles-pp.get_mem_cgroup_from_mm.__mem_cgroup_charge.do_anonymous_page.__handle_mm_fault.handle_mm_fault
1.50 ± 11% -0.3 1.19 ± 5% perf-profile.calltrace.cycles-pp.uncharge_folio.__mem_cgroup_uncharge_list.release_pages.tlb_batch_pages_flush.tlb_finish_mmu
1.13 ± 3% -0.3 0.83 perf-profile.calltrace.cycles-pp.flush_tlb_mm_range.tlb_finish_mmu.zap_page_range_single.madvise_vma_behavior.do_madvise
0.71 ± 3% -0.3 0.43 ± 44% perf-profile.calltrace.cycles-pp.syscall_return_via_sysret.__madvise
1.02 ± 3% -0.3 0.75 perf-profile.calltrace.cycles-pp.flush_tlb_func.flush_tlb_mm_range.tlb_finish_mmu.zap_page_range_single.madvise_vma_behavior
0.97 ± 3% -0.3 0.72 perf-profile.calltrace.cycles-pp.native_flush_tlb_one_user.flush_tlb_func.flush_tlb_mm_range.tlb_finish_mmu.zap_page_range_single
0.77 ± 2% -0.2 0.58 ± 2% perf-profile.calltrace.cycles-pp.syscall_exit_to_user_mode.do_syscall_64.entry_SYSCALL_64_after_hwframe.__madvise
0.71 ± 2% -0.1 0.60 ± 3% perf-profile.calltrace.cycles-pp.propagate_protected_usage.page_counter_uncharge.uncharge_batch.__mem_cgroup_uncharge_list.release_pages
1.20 +0.1 1.34 perf-profile.calltrace.cycles-pp.unmap_page_range.zap_page_range_single.madvise_vma_behavior.do_madvise.__x64_sys_madvise
1.10 ± 2% +0.2 1.28 perf-profile.calltrace.cycles-pp.zap_pmd_range.unmap_page_range.zap_page_range_single.madvise_vma_behavior.do_madvise
1.04 ± 2% +0.2 1.24 perf-profile.calltrace.cycles-pp.zap_pte_range.zap_pmd_range.unmap_page_range.zap_page_range_single.madvise_vma_behavior
0.83 +0.2 1.07 ± 2% perf-profile.calltrace.cycles-pp.lru_add_fn.folio_batch_move_lru.lru_add_drain_cpu.lru_add_drain.zap_page_range_single
0.81 ± 2% +0.3 1.08 perf-profile.calltrace.cycles-pp.page_remove_rmap.zap_pte_range.zap_pmd_range.unmap_page_range.zap_page_range_single
0.88 ± 10% +0.3 1.16 ± 4% perf-profile.calltrace.cycles-pp.mem_cgroup_commit_charge.__mem_cgroup_charge.do_anonymous_page.__handle_mm_fault.handle_mm_fault
0.71 ± 2% +0.3 1.00 perf-profile.calltrace.cycles-pp.__mod_lruvec_page_state.page_remove_rmap.zap_pte_range.zap_pmd_range.unmap_page_range
0.76 ± 3% +0.3 1.09 ± 2% perf-profile.calltrace.cycles-pp.folio_add_new_anon_rmap.do_anonymous_page.__handle_mm_fault.handle_mm_fault.do_user_addr_fault
0.73 ± 3% +0.3 1.07 ± 2% perf-profile.calltrace.cycles-pp.__mod_lruvec_page_state.folio_add_new_anon_rmap.do_anonymous_page.__handle_mm_fault.handle_mm_fault
0.00 +0.6 0.55 ± 2% perf-profile.calltrace.cycles-pp.__count_memcg_events.mem_cgroup_commit_charge.__mem_cgroup_charge.do_anonymous_page.__handle_mm_fault
6.60 ± 4% +0.6 7.18 ± 3% perf-profile.calltrace.cycles-pp.exc_page_fault.asm_exc_page_fault.testcase
6.54 ± 4% +0.6 7.13 ± 3% perf-profile.calltrace.cycles-pp.do_user_addr_fault.exc_page_fault.asm_exc_page_fault.testcase
0.00 +0.7 0.74 ± 3% perf-profile.calltrace.cycles-pp.__mod_memcg_lruvec_state.release_pages.tlb_batch_pages_flush.tlb_finish_mmu.zap_page_range_single
0.00 +0.8 0.79 ± 2% perf-profile.calltrace.cycles-pp.__mod_memcg_lruvec_state.__mod_lruvec_page_state.page_remove_rmap.zap_pte_range.zap_pmd_range
0.00 +0.8 0.79 ± 3% perf-profile.calltrace.cycles-pp.__mod_memcg_lruvec_state.lru_add_fn.folio_batch_move_lru.lru_add_drain_cpu.lru_add_drain
0.00 +0.8 0.80 ± 3% perf-profile.calltrace.cycles-pp.__mod_memcg_lruvec_state.__mod_lruvec_page_state.folio_add_new_anon_rmap.do_anonymous_page.__handle_mm_fault
5.80 ± 5% +0.8 6.60 ± 3% perf-profile.calltrace.cycles-pp.handle_mm_fault.do_user_addr_fault.exc_page_fault.asm_exc_page_fault.testcase
0.00 +0.8 0.82 perf-profile.calltrace.cycles-pp.__count_memcg_events.uncharge_batch.__mem_cgroup_uncharge_list.release_pages.tlb_batch_pages_flush
0.69 ± 4% +0.9 1.59 ± 2% perf-profile.calltrace.cycles-pp.__count_memcg_events.handle_mm_fault.do_user_addr_fault.exc_page_fault.asm_exc_page_fault
30.43 +1.1 31.57 perf-profile.calltrace.cycles-pp.tlb_finish_mmu.zap_page_range_single.madvise_vma_behavior.do_madvise.__x64_sys_madvise
29.22 +1.5 30.69 perf-profile.calltrace.cycles-pp.tlb_batch_pages_flush.tlb_finish_mmu.zap_page_range_single.madvise_vma_behavior.do_madvise
29.05 +1.5 30.56 perf-profile.calltrace.cycles-pp.release_pages.tlb_batch_pages_flush.tlb_finish_mmu.zap_page_range_single.madvise_vma_behavior
22.56 ± 2% +2.3 24.87 perf-profile.calltrace.cycles-pp.folio_lruvec_lock_irqsave.release_pages.tlb_batch_pages_flush.tlb_finish_mmu.zap_page_range_single
22.36 ± 2% +2.3 24.70 perf-profile.calltrace.cycles-pp._raw_spin_lock_irqsave.folio_lruvec_lock_irqsave.release_pages.tlb_batch_pages_flush.tlb_finish_mmu
22.11 ± 2% +2.4 24.55 perf-profile.calltrace.cycles-pp.native_queued_spin_lock_slowpath._raw_spin_lock_irqsave.folio_lruvec_lock_irqsave.release_pages.tlb_batch_pages_flush
22.70 +2.6 25.35 perf-profile.calltrace.cycles-pp.folio_lruvec_lock_irqsave.folio_batch_move_lru.lru_add_drain_cpu.lru_add_drain.zap_page_range_single
22.38 +2.7 25.08 perf-profile.calltrace.cycles-pp._raw_spin_lock_irqsave.folio_lruvec_lock_irqsave.folio_batch_move_lru.lru_add_drain_cpu.lru_add_drain
24.10 +2.7 26.82 perf-profile.calltrace.cycles-pp.lru_add_drain.zap_page_range_single.madvise_vma_behavior.do_madvise.__x64_sys_madvise
24.09 +2.7 26.82 perf-profile.calltrace.cycles-pp.lru_add_drain_cpu.lru_add_drain.zap_page_range_single.madvise_vma_behavior.do_madvise
24.07 +2.7 26.79 perf-profile.calltrace.cycles-pp.folio_batch_move_lru.lru_add_drain_cpu.lru_add_drain.zap_page_range_single.madvise_vma_behavior
22.14 +2.8 24.93 perf-profile.calltrace.cycles-pp.native_queued_spin_lock_slowpath._raw_spin_lock_irqsave.folio_lruvec_lock_irqsave.folio_batch_move_lru.lru_add_drain_cpu
59.76 +2.9 62.64 perf-profile.calltrace.cycles-pp.__madvise
57.63 +3.5 61.10 perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.__madvise
57.27 +3.6 60.85 perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.__madvise
56.41 +3.8 60.20 perf-profile.calltrace.cycles-pp.__x64_sys_madvise.do_syscall_64.entry_SYSCALL_64_after_hwframe.__madvise
56.37 +3.8 60.17 perf-profile.calltrace.cycles-pp.do_madvise.__x64_sys_madvise.do_syscall_64.entry_SYSCALL_64_after_hwframe.__madvise
55.94 +3.9 59.88 perf-profile.calltrace.cycles-pp.madvise_vma_behavior.do_madvise.__x64_sys_madvise.do_syscall_64.entry_SYSCALL_64_after_hwframe
55.85 +4.0 59.82 perf-profile.calltrace.cycles-pp.zap_page_range_single.madvise_vma_behavior.do_madvise.__x64_sys_madvise.do_syscall_64
26.75 -1.8 24.92 ± 2% perf-profile.children.cycles-pp.start_secondary
26.98 -1.8 25.22 ± 3% perf-profile.children.cycles-pp.intel_idle_ibrs
27.05 -1.8 25.29 ± 3% perf-profile.children.cycles-pp.cpu_startup_entry
27.05 -1.8 25.29 ± 3% perf-profile.children.cycles-pp.do_idle
27.05 -1.8 25.29 ± 3% perf-profile.children.cycles-pp.secondary_startup_64_no_verify
27.05 -1.8 25.29 ± 3% perf-profile.children.cycles-pp.cpuidle_enter
27.05 -1.8 25.29 ± 3% perf-profile.children.cycles-pp.cpuidle_enter_state
27.05 -1.8 25.29 ± 3% perf-profile.children.cycles-pp.cpuidle_idle_call
13.66 ± 2% -1.3 12.38 perf-profile.children.cycles-pp.testcase
5.55 ± 5% -1.0 4.52 ± 3% perf-profile.children.cycles-pp.__mem_cgroup_uncharge_list
2.39 ± 2% -0.8 1.63 ± 3% perf-profile.children.cycles-pp.page_counter_uncharge
4.03 ± 3% -0.7 3.32 ± 3% perf-profile.children.cycles-pp.uncharge_batch
1.96 ± 4% -0.4 1.52 ± 2% perf-profile.children.cycles-pp.irqentry_exit_to_user_mode
1.30 -0.4 0.94 ± 2% perf-profile.children.cycles-pp.error_entry
1.36 ± 2% -0.4 0.99 perf-profile.children.cycles-pp.__irqentry_text_end
1.30 ± 10% -0.4 0.94 ± 6% perf-profile.children.cycles-pp.get_mem_cgroup_from_mm
1.51 ± 11% -0.3 1.19 ± 5% perf-profile.children.cycles-pp.uncharge_folio
1.14 ± 3% -0.3 0.84 perf-profile.children.cycles-pp.flush_tlb_mm_range
1.02 ± 3% -0.3 0.75 perf-profile.children.cycles-pp.flush_tlb_func
0.98 ± 3% -0.3 0.72 perf-profile.children.cycles-pp.native_flush_tlb_one_user
0.73 ± 2% -0.2 0.52 ± 2% perf-profile.children.cycles-pp.syscall_return_via_sysret
0.69 ± 2% -0.2 0.50 ± 2% perf-profile.children.cycles-pp.native_irq_return_iret
0.79 ± 2% -0.2 0.60 ± 2% perf-profile.children.cycles-pp.syscall_exit_to_user_mode
0.51 ± 2% -0.1 0.38 ± 2% perf-profile.children.cycles-pp.sync_regs
0.41 ± 3% -0.1 0.29 ± 3% perf-profile.children.cycles-pp.__perf_sw_event
0.44 ± 2% -0.1 0.32 ± 2% perf-profile.children.cycles-pp.vma_alloc_folio
0.72 ± 2% -0.1 0.61 ± 3% perf-profile.children.cycles-pp.propagate_protected_usage
0.39 -0.1 0.28 ± 2% perf-profile.children.cycles-pp.alloc_pages_mpol
0.35 ± 3% -0.1 0.25 ± 3% perf-profile.children.cycles-pp.__alloc_pages
0.34 ± 2% -0.1 0.24 ± 4% perf-profile.children.cycles-pp.___perf_sw_event
0.30 ± 3% -0.1 0.21 ± 5% perf-profile.children.cycles-pp.lock_vma_under_rcu
0.32 ± 2% -0.1 0.24 perf-profile.children.cycles-pp.entry_SYSCALL_64
0.12 ± 4% -0.1 0.03 ± 70% perf-profile.children.cycles-pp.down_read
0.25 ± 3% -0.1 0.18 ± 4% perf-profile.children.cycles-pp.mas_walk
0.25 ± 3% -0.1 0.18 ± 2% perf-profile.children.cycles-pp.get_page_from_freelist
0.17 ± 4% -0.1 0.11 ± 3% perf-profile.children.cycles-pp.__pte_offset_map_lock
0.14 ± 3% -0.0 0.10 ± 3% perf-profile.children.cycles-pp.clear_page_erms
0.17 ± 2% -0.0 0.12 ± 3% perf-profile.children.cycles-pp.find_vma_prev
0.13 ± 2% -0.0 0.09 perf-profile.children.cycles-pp.percpu_counter_add_batch
0.11 ± 4% -0.0 0.07 ± 10% perf-profile.children.cycles-pp.__cond_resched
0.13 ± 2% -0.0 0.10 ± 7% perf-profile.children.cycles-pp.free_pages_and_swap_cache
0.06 ± 7% -0.0 0.03 ± 70% perf-profile.children.cycles-pp.unmap_vmas
0.11 ± 3% -0.0 0.08 ± 6% perf-profile.children.cycles-pp.free_unref_page_list
0.06 -0.0 0.03 ± 70% perf-profile.children.cycles-pp.exit_to_user_mode_prepare
0.09 ± 7% -0.0 0.06 ± 6% perf-profile.children.cycles-pp.free_swap_cache
0.09 ± 7% -0.0 0.07 ± 7% perf-profile.children.cycles-pp.__munmap
0.09 ± 8% -0.0 0.06 ± 6% perf-profile.children.cycles-pp._raw_spin_lock
0.09 ± 5% -0.0 0.06 ± 6% perf-profile.children.cycles-pp.handle_pte_fault
0.08 ± 8% -0.0 0.06 ± 6% perf-profile.children.cycles-pp.do_vmi_munmap
0.08 ± 4% -0.0 0.06 ± 6% perf-profile.children.cycles-pp.__mod_lruvec_state
0.07 ± 6% -0.0 0.05 ± 8% perf-profile.children.cycles-pp.rmqueue
0.07 ± 9% -0.0 0.05 ± 7% perf-profile.children.cycles-pp.unmap_region
0.08 ± 8% -0.0 0.06 ± 6% perf-profile.children.cycles-pp.__vm_munmap
0.08 ± 8% -0.0 0.06 ± 6% perf-profile.children.cycles-pp.__x64_sys_munmap
0.08 ± 8% -0.0 0.06 ± 6% perf-profile.children.cycles-pp.do_vmi_align_munmap
0.08 ± 5% -0.0 0.07 ± 7% perf-profile.children.cycles-pp.try_charge_memcg
1.27 +0.1 1.40 perf-profile.children.cycles-pp.unmap_page_range
1.17 +0.2 1.32 perf-profile.children.cycles-pp.zap_pmd_range
1.12 +0.2 1.29 perf-profile.children.cycles-pp.zap_pte_range
0.84 +0.2 1.07 ± 2% perf-profile.children.cycles-pp.lru_add_fn
0.81 ± 2% +0.3 1.08 perf-profile.children.cycles-pp.page_remove_rmap
0.89 ± 10% +0.3 1.16 ± 4% perf-profile.children.cycles-pp.mem_cgroup_commit_charge
0.77 ± 3% +0.3 1.09 ± 2% perf-profile.children.cycles-pp.folio_add_new_anon_rmap
6.62 ± 4% +0.6 7.19 ± 3% perf-profile.children.cycles-pp.exc_page_fault
6.56 ± 4% +0.6 7.14 ± 3% perf-profile.children.cycles-pp.do_user_addr_fault
1.44 ± 2% +0.6 2.08 ± 2% perf-profile.children.cycles-pp.__mod_lruvec_page_state
5.80 ± 5% +0.8 6.61 ± 3% perf-profile.children.cycles-pp.handle_mm_fault
30.44 +1.1 31.58 perf-profile.children.cycles-pp.tlb_finish_mmu
29.23 +1.5 30.69 perf-profile.children.cycles-pp.tlb_batch_pages_flush
29.19 +1.5 30.66 perf-profile.children.cycles-pp.release_pages
1.63 ± 5% +1.5 3.13 ± 2% perf-profile.children.cycles-pp.__mod_memcg_lruvec_state
1.32 ± 4% +1.6 2.97 ± 2% perf-profile.children.cycles-pp.__count_memcg_events
24.12 +2.7 26.84 perf-profile.children.cycles-pp.lru_add_drain
24.12 +2.7 26.84 perf-profile.children.cycles-pp.lru_add_drain_cpu
24.09 +2.7 26.81 perf-profile.children.cycles-pp.folio_batch_move_lru
59.80 +2.9 62.68 perf-profile.children.cycles-pp.__madvise
57.82 +3.4 61.26 perf-profile.children.cycles-pp.entry_SYSCALL_64_after_hwframe
57.44 +3.5 60.99 perf-profile.children.cycles-pp.do_syscall_64
56.41 +3.8 60.20 perf-profile.children.cycles-pp.__x64_sys_madvise
56.37 +3.8 60.17 perf-profile.children.cycles-pp.do_madvise
55.94 +3.9 59.88 perf-profile.children.cycles-pp.madvise_vma_behavior
55.85 +4.0 59.82 perf-profile.children.cycles-pp.zap_page_range_single
45.26 +5.0 50.23 perf-profile.children.cycles-pp.folio_lruvec_lock_irqsave
44.75 +5.0 49.80 perf-profile.children.cycles-pp._raw_spin_lock_irqsave
44.26 +5.2 49.50 perf-profile.children.cycles-pp.native_queued_spin_lock_slowpath
26.98 -1.8 25.22 ± 3% perf-profile.self.cycles-pp.intel_idle_ibrs
1.67 ± 3% -0.6 1.02 ± 3% perf-profile.self.cycles-pp.page_counter_uncharge
1.92 ± 5% -0.4 1.49 ± 2% perf-profile.self.cycles-pp.irqentry_exit_to_user_mode
1.47 ± 2% -0.4 1.06 ± 2% perf-profile.self.cycles-pp.testcase
1.36 ± 2% -0.4 0.99 perf-profile.self.cycles-pp.__irqentry_text_end
1.30 -0.4 0.94 perf-profile.self.cycles-pp.error_entry
1.30 ± 10% -0.4 0.94 ± 6% perf-profile.self.cycles-pp.get_mem_cgroup_from_mm
1.18 ± 8% -0.3 0.86 ± 6% perf-profile.self.cycles-pp.uncharge_batch
1.50 ± 11% -0.3 1.19 ± 5% perf-profile.self.cycles-pp.uncharge_folio
0.98 ± 3% -0.3 0.72 perf-profile.self.cycles-pp.native_flush_tlb_one_user
0.71 ± 2% -0.2 0.51 ± 2% perf-profile.self.cycles-pp.syscall_return_via_sysret
0.69 ± 2% -0.2 0.50 ± 2% perf-profile.self.cycles-pp.native_irq_return_iret
0.50 ± 4% -0.2 0.30 ± 5% perf-profile.self.cycles-pp._raw_spin_lock_irqsave
0.75 ± 2% -0.2 0.56 ± 2% perf-profile.self.cycles-pp.syscall_exit_to_user_mode
0.51 ± 2% -0.1 0.38 ± 2% perf-profile.self.cycles-pp.sync_regs
0.35 ± 3% -0.1 0.23 ± 2% perf-profile.self.cycles-pp.folio_batch_move_lru
0.36 ± 5% -0.1 0.24 ± 2% perf-profile.self.cycles-pp.lru_add_fn
0.39 ± 2% -0.1 0.27 ± 2% perf-profile.self.cycles-pp.entry_SYSCALL_64_after_hwframe
0.72 ± 2% -0.1 0.61 ± 3% perf-profile.self.cycles-pp.propagate_protected_usage
0.45 -0.1 0.34 ± 2% perf-profile.self.cycles-pp.release_pages
0.54 ± 4% -0.1 0.45 ± 4% perf-profile.self.cycles-pp.__mod_lruvec_page_state
0.30 ± 2% -0.1 0.21 ± 3% perf-profile.self.cycles-pp.___perf_sw_event
0.52 ± 5% -0.1 0.43 ± 5% perf-profile.self.cycles-pp.folio_lruvec_lock_irqsave
0.28 ± 3% -0.1 0.21 perf-profile.self.cycles-pp.entry_SYSCALL_64
0.25 ± 3% -0.1 0.18 ± 4% perf-profile.self.cycles-pp.mas_walk
0.24 ± 2% -0.1 0.17 ± 4% perf-profile.self.cycles-pp.__handle_mm_fault
0.16 ± 4% -0.1 0.10 ± 9% perf-profile.self.cycles-pp.zap_pte_range
0.14 ± 4% -0.0 0.10 ± 4% perf-profile.self.cycles-pp.clear_page_erms
0.08 ± 6% -0.0 0.03 ± 70% perf-profile.self.cycles-pp.__cond_resched
0.13 -0.0 0.09 perf-profile.self.cycles-pp.percpu_counter_add_batch
0.14 ± 5% -0.0 0.11 ± 3% perf-profile.self.cycles-pp.handle_mm_fault
0.11 ± 3% -0.0 0.08 ± 6% perf-profile.self.cycles-pp.do_user_addr_fault
0.08 ± 6% -0.0 0.04 ± 44% perf-profile.self.cycles-pp.__perf_sw_event
0.07 ± 10% -0.0 0.04 ± 44% perf-profile.self.cycles-pp.tlb_finish_mmu
0.09 ± 7% -0.0 0.06 ± 6% perf-profile.self.cycles-pp.free_swap_cache
0.08 ± 7% -0.0 0.05 ± 8% perf-profile.self.cycles-pp.lock_vma_under_rcu
0.09 ± 8% -0.0 0.06 ± 6% perf-profile.self.cycles-pp._raw_spin_lock
0.07 ± 7% -0.0 0.04 ± 44% perf-profile.self.cycles-pp.asm_exc_page_fault
0.10 ± 3% -0.0 0.08 ± 6% perf-profile.self.cycles-pp.page_remove_rmap
0.08 ± 6% -0.0 0.05 ± 8% perf-profile.self.cycles-pp.flush_tlb_mm_range
0.08 ± 6% -0.0 0.06 ± 9% perf-profile.self.cycles-pp.do_anonymous_page
0.08 ± 7% -0.0 0.06 ± 6% perf-profile.self.cycles-pp.unmap_page_range
0.08 ± 5% -0.0 0.06 ± 6% perf-profile.self.cycles-pp.__alloc_pages
0.08 ± 6% -0.0 0.06 ± 8% perf-profile.self.cycles-pp.do_madvise
0.07 ± 10% -0.0 0.05 ± 8% perf-profile.self.cycles-pp.up_read
1.58 ± 6% +1.5 3.09 ± 2% perf-profile.self.cycles-pp.__mod_memcg_lruvec_state
1.27 ± 5% +1.7 2.93 ± 2% perf-profile.self.cycles-pp.__count_memcg_events
44.25 +5.2 49.50 perf-profile.self.cycles-pp.native_queued_spin_lock_slowpath
Disclaimer:
Results have been estimated based on internal Intel analysis and are provided
for informational purposes only. Any difference in system hardware or software
design or configuration may affect actual performance.
--
0-DAY CI Kernel Test Service
https://github.com/intel/lkp-tests/wiki
Powered by blists - more mailing lists