[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <202403220943.96dde419-oliver.sang@intel.com>
Date: Fri, 22 Mar 2024 10:12:04 +0800
From: kernel test robot <oliver.sang@...el.com>
To: Robert Richter <rrichter@....com>
CC: <oe-lkp@...ts.linux.dev>, <lkp@...el.com>, <linux-acpi@...r.kernel.org>,
<ying.huang@...el.com>, <feng.tang@...el.com>, <fengwei.yin@...el.com>,
"Rafael J. Wysocki" <rafael@...nel.org>, Dave Hansen
<dave.hansen@...ux.intel.com>, Dan Williams <dan.j.williams@...el.com>,
Alison Schofield <alison.schofield@...el.com>,
<linux-kernel@...r.kernel.org>, <linux-cxl@...r.kernel.org>, Robert Richter
<rrichter@....com>, Len Brown <lenb@...nel.org>, <oliver.sang@...el.com>
Subject: Re: [PATCH v2 3/3] ACPI/NUMA: Remove architecture dependent
remainings
Hello,
kernel test robot noticed a -43.6% regression of vm-scalability.throughput on:
commit: 09f84589a8ed1162c0274a1b5b3dadb98ee3155e ("[PATCH v2 3/3] ACPI/NUMA: Remove architecture dependent remainings")
url: https://github.com/intel-lab-lkp/linux/commits/Robert-Richter/x86-numa-Fix-SRAT-lookup-of-CFMWS-ranges-with-numa_fill_memblks/20240319-200345
base: https://git.kernel.org/cgit/linux/kernel/git/rafael/linux-pm.git linux-next
patch link: https://lore.kernel.org/all/20240319120026.2246389-4-rrichter@amd.com/
patch subject: [PATCH v2 3/3] ACPI/NUMA: Remove architecture dependent remainings
testcase: vm-scalability
test machine: 128 threads 2 sockets Intel(R) Xeon(R) Gold 6338 CPU @ 2.00GHz (Ice Lake) with 256G memory
parameters:
runtime: 300s
size: 512G
test: anon-cow-rand-hugetlb
cpufreq_governor: performance
In addition to that, the commit also has significant impact on the following tests:
+------------------+-----------------------------------------------------------------------------------------------+
| testcase: change | phoronix-test-suite: phoronix-test-suite.npb.SP.B.total_mop_s -58.1% regression |
| test machine | 96 threads 2 sockets Intel(R) Xeon(R) Gold 6252 CPU @ 2.10GHz (Cascade Lake) with 512G memory |
| test parameters | cpufreq_governor=performance |
| | option_a=SP.B [Scalar Penta-diagonal solver] |
| | test=npb-1.4.5 |
+------------------+-----------------------------------------------------------------------------------------------+
| testcase: change | will-it-scale: will-it-scale.per_process_ops -38.1% regression |
| test machine | 104 threads 2 sockets (Skylake) with 192G memory |
| test parameters | cpufreq_governor=performance |
| | mode=process |
| | nr_task=100% |
| | test=mmap1 |
+------------------+-----------------------------------------------------------------------------------------------+
| testcase: change | phoronix-test-suite: phoronix-test-suite.asmfish.0.nodes_second -23.9% regression |
| test machine | 96 threads 2 sockets Intel(R) Xeon(R) Gold 6252 CPU @ 2.10GHz (Cascade Lake) with 512G memory |
| test parameters | cpufreq_governor=performance |
| | test=asmfish-1.1.2 |
+------------------+-----------------------------------------------------------------------------------------------+
If you fix the issue in a separate patch/commit (i.e. not just a new version of
the same patch/commit), kindly add following tags
| Reported-by: kernel test robot <oliver.sang@...el.com>
| Closes: https://lore.kernel.org/oe-lkp/202403220943.96dde419-oliver.sang@intel.com
Details are as below:
-------------------------------------------------------------------------------------------------->
The kernel config and materials to reproduce are available at:
https://download.01.org/0day-ci/archive/20240322/202403220943.96dde419-oliver.sang@intel.com
=========================================================================================
compiler/cpufreq_governor/kconfig/rootfs/runtime/size/tbox_group/test/testcase:
gcc-12/performance/x86_64-rhel-8.3/debian-12-x86_64-20240206.cgz/300s/512G/lkp-icl-2sp2/anon-cow-rand-hugetlb/vm-scalability
commit:
9ef3ad0f03 ("ACPI/NUMA: Print CXL Early Discovery Table (CEDT)")
09f84589a8 ("ACPI/NUMA: Remove architecture dependent remainings")
9ef3ad0f032818c5 09f84589a8ed1162c0274a1b5b3
---------------- ---------------------------
%stddev %change %stddev
\ | \
4.277e+09 ± 6% -70.5% 1.261e+09 ± 6% cpuidle..time
735569 ± 2% -34.8% 479905 ± 3% cpuidle..usage
243.64 ± 2% -24.9% 182.92 uptime.boot
9961 ± 4% -29.0% 7073 ± 4% uptime.idle
16.94 ± 3% -9.9 7.06 ± 8% mpstat.cpu.all.idle%
0.35 -0.1 0.29 mpstat.cpu.all.irq%
0.03 ± 3% -0.0 0.02 mpstat.cpu.all.soft%
12.05 -2.1 10.00 mpstat.cpu.all.sys%
70.63 +12.0 82.63 mpstat.cpu.all.usr%
17.78 ± 3% -52.2% 8.49 ± 6% vmstat.cpu.id
70.01 +16.5% 81.54 vmstat.cpu.us
107.54 +10.6% 118.93 vmstat.procs.r
2652 ± 2% -8.9% 2417 ± 5% vmstat.system.cs
141617 ± 3% -13.4% 122650 vmstat.system.in
573464 ± 8% -100.0% 0.00 numa-numastat.node0.local_node
632802 ± 4% -100.0% 0.00 numa-numastat.node0.numa_hit
59338 ± 51% -100.0% 0.00 numa-numastat.node0.other_node
677948 ± 7% -100.0% 0.00 numa-numastat.node1.local_node
751205 ± 4% -100.0% 0.00 numa-numastat.node1.numa_hit
73258 ± 41% -100.0% 0.00 numa-numastat.node1.other_node
95544 ± 27% -59.6% 38591 ± 5% meminfo.Active
95399 ± 27% -59.7% 38433 ± 5% meminfo.Active(anon)
10487673 ± 7% -28.6% 7487857 ± 6% meminfo.DirectMap2M
43001 +14.2% 49088 meminfo.HugePages_Surp
43001 +14.2% 49088 meminfo.HugePages_Total
88066074 +14.2% 1.005e+08 meminfo.Hugetlb
1.356e+08 -9.5% 1.228e+08 meminfo.MemAvailable
1.366e+08 -9.1% 1.241e+08 meminfo.MemFree
94133967 +13.0% 1.064e+08 meminfo.Memused
241580 ± 8% -26.3% 178118 meminfo.Shmem
267728 -81.4% 49921 meminfo.VmallocUsed
51764 ± 2% -37.7% 32258 vm-scalability.median
6.46 ± 19% -5.3 1.21 ± 20% vm-scalability.median_stddev%
2.86 ± 18% -2.7 0.11 ± 31% vm-scalability.stddev%
7315147 -43.6% 4122657 vm-scalability.throughput
193.69 ± 2% -32.1% 131.50 vm-scalability.time.elapsed_time
193.69 ± 2% -32.1% 131.50 vm-scalability.time.elapsed_time.max
138604 ± 4% -21.6% 108601 ± 7% vm-scalability.time.involuntary_context_switches
578101 -55.9% 254781 vm-scalability.time.minor_page_faults
10610 +12.8% 11965 vm-scalability.time.percent_of_cpu_this_job_got
2981 ± 3% -43.3% 1691 vm-scalability.time.system_time
17565 -20.0% 14044 vm-scalability.time.user_time
15370 ± 4% -51.6% 7439 ± 5% vm-scalability.time.voluntary_context_switches
1.159e+09 -55.6% 5.146e+08 vm-scalability.workload
495078 ± 11% -71.4% 141719 ± 6% sched_debug.cfs_rq:/.avg_vruntime.stddev
0.28 ± 21% -33.3% 0.19 ± 21% sched_debug.cfs_rq:/.h_nr_running.stddev
2570 ± 80% +109.3% 5381 sched_debug.cfs_rq:/.load.min
2.76 ± 52% +68.8% 4.67 sched_debug.cfs_rq:/.load_avg.min
495078 ± 11% -71.4% 141719 ± 6% sched_debug.cfs_rq:/.min_vruntime.stddev
0.23 ± 31% -53.6% 0.11 ± 15% sched_debug.cfs_rq:/.nr_running.stddev
242.06 ± 32% -50.0% 121.10 ± 9% sched_debug.cfs_rq:/.runnable_avg.stddev
235.93 ± 81% +137.9% 561.33 ± 19% sched_debug.cfs_rq:/.util_avg.min
226.75 ± 31% -52.3% 108.21 ± 13% sched_debug.cfs_rq:/.util_avg.stddev
192.74 ± 41% -73.6% 50.87 ± 50% sched_debug.cfs_rq:/.util_est.stddev
370662 ± 45% +53.6% 569380 ± 2% sched_debug.cpu.avg_idle.min
4569 ± 11% -23.5% 3494 sched_debug.cpu.curr->pid.avg
6312 ± 6% -26.8% 4619 sched_debug.cpu.curr->pid.max
37956 ± 30% -50.9% 18650 ± 70% sched_debug.cpu.max_idle_balance_cost.stddev
2949 ± 7% -22.4% 2287 ± 3% sched_debug.cpu.nr_switches.avg
821.53 ± 12% -31.9% 559.11 ± 6% sched_debug.cpu.nr_switches.min
0.25 ±101% -99.8% 0.00 ± 92% sched_debug.rt_rq:.rt_time.avg
32.43 ±101% -99.8% 0.06 ± 92% sched_debug.rt_rq:.rt_time.max
2.85 ±101% -99.8% 0.01 ± 92% sched_debug.rt_rq:.rt_time.stddev
507357 -55.6% 225492 proc-vmstat.htlb_buddy_alloc_success
23772 ± 27% -59.2% 9703 ± 5% proc-vmstat.nr_active_anon
3376161 -9.1% 3070574 proc-vmstat.nr_dirty_background_threshold
6760578 -9.1% 6148657 proc-vmstat.nr_dirty_threshold
823112 -1.9% 807289 proc-vmstat.nr_file_pages
34080722 -8.7% 31109176 proc-vmstat.nr_free_pages
25608 +3.2% 26432 proc-vmstat.nr_kernel_stack
2710 +3.5% 2804 proc-vmstat.nr_page_table_pages
60299 ± 8% -26.2% 44490 proc-vmstat.nr_shmem
28254 -1.1% 27945 proc-vmstat.nr_slab_reclaimable
66139 -1.8% 64958 proc-vmstat.nr_slab_unreclaimable
23772 ± 27% -59.2% 9703 ± 5% proc-vmstat.nr_zone_active_anon
33857 ± 16% -100.0% 0.00 proc-vmstat.numa_hint_faults
12661 ± 72% -100.0% 0.00 proc-vmstat.numa_hint_faults_local
1386858 -43.2% 787481 proc-vmstat.numa_hit
1254263 -37.2% 787477 proc-vmstat.numa_local
132597 -100.0% 0.00 proc-vmstat.numa_other
41301 ± 42% -100.0% 0.00 proc-vmstat.numa_pages_migrated
183630 ± 19% -100.0% 0.00 proc-vmstat.numa_pte_updates
2.608e+08 -55.5% 1.161e+08 proc-vmstat.pgalloc_normal
1247964 -55.1% 559863 proc-vmstat.pgfault
2.597e+08 -55.4% 1.159e+08 proc-vmstat.pgfree
41301 ± 42% -100.0% 0.00 proc-vmstat.pgmigrate_success
35726 ± 6% -55.9% 15762 ± 3% proc-vmstat.pgreuse
15.99 -14.2% 13.72 perf-stat.i.MPKI
1.556e+10 ± 3% -34.0% 1.027e+10 perf-stat.i.branch-instructions
0.10 ± 4% +0.1 0.18 ± 4% perf-stat.i.branch-miss-rate%
11581986 ± 3% +29.1% 14946939 perf-stat.i.branch-misses
8.579e+08 ± 2% -36.0% 5.495e+08 perf-stat.i.cache-misses
8.808e+08 ± 2% -35.7% 5.662e+08 perf-stat.i.cache-references
2591 ± 2% -12.0% 2281 ± 4% perf-stat.i.context-switches
4.97 +47.0% 7.31 perf-stat.i.cpi
2.814e+11 +11.0% 3.122e+11 perf-stat.i.cpu-cycles
218.48 -11.0% 194.53 perf-stat.i.cpu-migrations
337.74 +73.1% 584.61 perf-stat.i.cycles-between-cache-misses
6.664e+10 ± 3% -33.9% 4.408e+10 perf-stat.i.instructions
0.25 -36.7% 0.16 perf-stat.i.ipc
2.00 ± 12% -47.4% 1.05 ± 16% perf-stat.i.major-faults
6085 -37.6% 3798 perf-stat.i.minor-faults
6087 -37.6% 3799 perf-stat.i.page-faults
12.96 -3.2% 12.55 perf-stat.overall.MPKI
0.07 +0.1 0.15 perf-stat.overall.branch-miss-rate%
4.25 +67.1% 7.10 perf-stat.overall.cpi
327.67 +72.6% 565.42 perf-stat.overall.cycles-between-cache-misses
0.24 -40.2% 0.14 perf-stat.overall.ipc
10931 +3.1% 11273 perf-stat.overall.path-length
1.518e+10 ± 2% -33.1% 1.015e+10 perf-stat.ps.branch-instructions
10985194 ± 2% +35.0% 14830464 perf-stat.ps.branch-misses
8.425e+08 ± 2% -35.1% 5.466e+08 perf-stat.ps.cache-misses
8.655e+08 ± 2% -34.9% 5.635e+08 perf-stat.ps.cache-references
2574 -11.3% 2282 ± 4% perf-stat.ps.context-switches
2.759e+11 +12.0% 3.09e+11 perf-stat.ps.cpu-cycles
215.11 ± 2% -9.9% 193.81 perf-stat.ps.cpu-migrations
6.501e+10 ± 2% -33.0% 4.355e+10 perf-stat.ps.instructions
2.04 ± 11% -47.9% 1.06 ± 16% perf-stat.ps.major-faults
6066 -36.8% 3832 perf-stat.ps.minor-faults
6068 -36.8% 3833 perf-stat.ps.page-faults
1.267e+13 -54.2% 5.802e+12 perf-stat.total.instructions
792.04 ± 23% -100.0% 0.00 numa-vmstat.node0.nr_active_anon
48328 ± 63% -100.0% 0.00 numa-vmstat.node0.nr_anon_pages
229536 ± 90% -100.0% 0.00 numa-vmstat.node0.nr_file_pages
23714921 ± 2% -100.0% 0.00 numa-vmstat.node0.nr_free_pages
53088 ± 73% -100.0% 0.00 numa-vmstat.node0.nr_inactive_anon
12744 ± 7% -100.0% 0.00 numa-vmstat.node0.nr_kernel_stack
5388 ± 45% -100.0% 0.00 numa-vmstat.node0.nr_mapped
890.81 ± 13% -100.0% 0.00 numa-vmstat.node0.nr_page_table_pages
6425 ±158% -100.0% 0.00 numa-vmstat.node0.nr_shmem
10347 ± 33% -100.0% 0.00 numa-vmstat.node0.nr_slab_reclaimable
32990 ± 10% -100.0% 0.00 numa-vmstat.node0.nr_slab_unreclaimable
223791 ± 90% -100.0% 0.00 numa-vmstat.node0.nr_unevictable
792.05 ± 23% -100.0% 0.00 numa-vmstat.node0.nr_zone_active_anon
53088 ± 73% -100.0% 0.00 numa-vmstat.node0.nr_zone_inactive_anon
223791 ± 90% -100.0% 0.00 numa-vmstat.node0.nr_zone_unevictable
633548 ± 4% -100.0% 0.00 numa-vmstat.node0.numa_hit
574210 ± 8% -100.0% 0.00 numa-vmstat.node0.numa_local
59338 ± 51% -100.0% 0.00 numa-vmstat.node0.numa_other
23125 ± 27% -100.0% 0.00 numa-vmstat.node1.nr_active_anon
115486 ± 26% -100.0% 0.00 numa-vmstat.node1.nr_anon_pages
593673 ± 34% -100.0% 0.00 numa-vmstat.node1.nr_file_pages
10376888 ± 3% -100.0% 0.00 numa-vmstat.node1.nr_free_pages
146240 ± 26% -100.0% 0.00 numa-vmstat.node1.nr_inactive_anon
12863 ± 6% -100.0% 0.00 numa-vmstat.node1.nr_kernel_stack
13956 ± 5% -100.0% 0.00 numa-vmstat.node1.nr_mapped
1818 ± 6% -100.0% 0.00 numa-vmstat.node1.nr_page_table_pages
53973 ± 15% -100.0% 0.00 numa-vmstat.node1.nr_shmem
17907 ± 19% -100.0% 0.00 numa-vmstat.node1.nr_slab_reclaimable
33150 ± 10% -100.0% 0.00 numa-vmstat.node1.nr_slab_unreclaimable
539462 ± 37% -100.0% 0.00 numa-vmstat.node1.nr_unevictable
23125 ± 27% -100.0% 0.00 numa-vmstat.node1.nr_zone_active_anon
146241 ± 26% -100.0% 0.00 numa-vmstat.node1.nr_zone_inactive_anon
539462 ± 37% -100.0% 0.00 numa-vmstat.node1.nr_zone_unevictable
750435 ± 4% -100.0% 0.00 numa-vmstat.node1.numa_hit
677178 ± 7% -100.0% 0.00 numa-vmstat.node1.numa_local
73258 ± 41% -100.0% 0.00 numa-vmstat.node1.numa_other
3216 ± 24% -100.0% 0.00 numa-meminfo.node0.Active
3165 ± 23% -100.0% 0.00 numa-meminfo.node0.Active(anon)
193141 ± 63% -100.0% 0.00 numa-meminfo.node0.AnonPages
245306 ± 55% -100.0% 0.00 numa-meminfo.node0.AnonPages.max
918145 ± 90% -100.0% 0.00 numa-meminfo.node0.FilePages
16988 ± 6% -100.0% 0.00 numa-meminfo.node0.HugePages_Surp
16988 ± 6% -100.0% 0.00 numa-meminfo.node0.HugePages_Total
212488 ± 73% -100.0% 0.00 numa-meminfo.node0.Inactive
212188 ± 73% -100.0% 0.00 numa-meminfo.node0.Inactive(anon)
41389 ± 33% -100.0% 0.00 numa-meminfo.node0.KReclaimable
12743 ± 7% -100.0% 0.00 numa-meminfo.node0.KernelStack
21242 ± 45% -100.0% 0.00 numa-meminfo.node0.Mapped
94746960 ± 2% -100.0% 0.00 numa-meminfo.node0.MemFree
1.317e+08 -100.0% 0.00 numa-meminfo.node0.MemTotal
36938247 ± 5% -100.0% 0.00 numa-meminfo.node0.MemUsed
3559 ± 13% -100.0% 0.00 numa-meminfo.node0.PageTables
41389 ± 33% -100.0% 0.00 numa-meminfo.node0.SReclaimable
131966 ± 10% -100.0% 0.00 numa-meminfo.node0.SUnreclaim
25702 ±158% -100.0% 0.00 numa-meminfo.node0.Shmem
173356 ± 15% -100.0% 0.00 numa-meminfo.node0.Slab
895165 ± 90% -100.0% 0.00 numa-meminfo.node0.Unevictable
92044 ± 27% -100.0% 0.00 numa-meminfo.node1.Active
91950 ± 27% -100.0% 0.00 numa-meminfo.node1.Active(anon)
94963 ± 57% -100.0% 0.00 numa-meminfo.node1.AnonHugePages
462070 ± 26% -100.0% 0.00 numa-meminfo.node1.AnonPages
521211 ± 21% -100.0% 0.00 numa-meminfo.node1.AnonPages.max
2374512 ± 34% -100.0% 0.00 numa-meminfo.node1.FilePages
26160 ± 3% -100.0% 0.00 numa-meminfo.node1.HugePages_Surp
26160 ± 3% -100.0% 0.00 numa-meminfo.node1.HugePages_Total
586305 ± 26% -100.0% 0.00 numa-meminfo.node1.Inactive
585444 ± 26% -100.0% 0.00 numa-meminfo.node1.Inactive(anon)
71636 ± 19% -100.0% 0.00 numa-meminfo.node1.KReclaimable
12870 ± 6% -100.0% 0.00 numa-meminfo.node1.KernelStack
55637 ± 5% -100.0% 0.00 numa-meminfo.node1.Mapped
41532114 ± 3% -100.0% 0.00 numa-meminfo.node1.MemFree
99030534 -100.0% 0.00 numa-meminfo.node1.MemTotal
57498419 ± 2% -100.0% 0.00 numa-meminfo.node1.MemUsed
7297 ± 6% -100.0% 0.00 numa-meminfo.node1.PageTables
71636 ± 19% -100.0% 0.00 numa-meminfo.node1.SReclaimable
132602 ± 10% -100.0% 0.00 numa-meminfo.node1.SUnreclaim
215707 ± 15% -100.0% 0.00 numa-meminfo.node1.Shmem
204238 ± 12% -100.0% 0.00 numa-meminfo.node1.Slab
2157849 ± 37% -100.0% 0.00 numa-meminfo.node1.Unevictable
56.43 ± 18% -56.4 0.00 perf-profile.calltrace.cycles-pp.do_rw_once
40.90 ± 15% -40.9 0.00 perf-profile.calltrace.cycles-pp.lrand48_r@plt
19.35 ± 72% -19.4 0.00 perf-profile.calltrace.cycles-pp.do_access
17.15 ± 84% -17.2 0.00 perf-profile.calltrace.cycles-pp.asm_exc_page_fault.do_access
17.12 ± 84% -17.1 0.00 perf-profile.calltrace.cycles-pp.exc_page_fault.asm_exc_page_fault.do_access
17.12 ± 84% -17.1 0.00 perf-profile.calltrace.cycles-pp.do_user_addr_fault.exc_page_fault.asm_exc_page_fault.do_access
17.10 ± 84% -17.1 0.00 perf-profile.calltrace.cycles-pp.handle_mm_fault.do_user_addr_fault.exc_page_fault.asm_exc_page_fault.do_access
17.09 ± 84% -17.1 0.00 perf-profile.calltrace.cycles-pp.hugetlb_fault.handle_mm_fault.do_user_addr_fault.exc_page_fault.asm_exc_page_fault
13.57 ± 84% -13.6 0.00 perf-profile.calltrace.cycles-pp.hugetlb_wp.hugetlb_fault.handle_mm_fault.do_user_addr_fault.exc_page_fault
13.30 ± 84% -13.3 0.00 perf-profile.calltrace.cycles-pp.copy_user_large_folio.hugetlb_wp.hugetlb_fault.handle_mm_fault.do_user_addr_fault
13.28 ± 84% -13.3 0.00 perf-profile.calltrace.cycles-pp.copy_subpage.copy_user_large_folio.hugetlb_wp.hugetlb_fault.handle_mm_fault
13.16 ± 84% -13.2 0.00 perf-profile.calltrace.cycles-pp.copy_mc_enhanced_fast_string.copy_subpage.copy_user_large_folio.hugetlb_wp.hugetlb_fault
0.00 +2.1 2.06 ± 9% perf-profile.calltrace.cycles-pp.__mmput.exit_mm.do_exit.do_group_exit.get_signal
0.00 +2.1 2.06 ± 9% perf-profile.calltrace.cycles-pp.exit_mm.do_exit.do_group_exit.get_signal.arch_do_signal_or_restart
0.00 +2.1 2.06 ± 9% perf-profile.calltrace.cycles-pp.put_files_struct.do_exit.do_group_exit.get_signal.arch_do_signal_or_restart
0.00 +3.3 3.35 ± 39% perf-profile.calltrace.cycles-pp.common_startup_64
0.00 +3.8 3.83 ± 43% perf-profile.calltrace.cycles-pp.__x64_sys_openat.do_syscall_64.entry_SYSCALL_64_after_hwframe
0.00 +3.8 3.83 ± 43% perf-profile.calltrace.cycles-pp.do_sys_openat2.__x64_sys_openat.do_syscall_64.entry_SYSCALL_64_after_hwframe
0.00 +4.3 4.32 ± 43% perf-profile.calltrace.cycles-pp.sw_perf_event_destroy._free_event.perf_event_release_kernel.perf_release.__fput
0.00 +4.4 4.37 ± 49% perf-profile.calltrace.cycles-pp.__x64_sys_exit_group.do_syscall_64.entry_SYSCALL_64_after_hwframe
0.00 +4.4 4.37 ± 49% perf-profile.calltrace.cycles-pp.do_exit.do_group_exit.__x64_sys_exit_group.do_syscall_64.entry_SYSCALL_64_after_hwframe
0.00 +4.4 4.37 ± 49% perf-profile.calltrace.cycles-pp.do_group_exit.__x64_sys_exit_group.do_syscall_64.entry_SYSCALL_64_after_hwframe
0.00 +4.6 4.63 ± 41% perf-profile.calltrace.cycles-pp._free_event.perf_event_release_kernel.perf_release.__fput.task_work_run
0.00 +5.2 5.19 ± 25% perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.write.writen.record__pushfn
0.00 +5.2 5.19 ± 25% perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.write.writen.record__pushfn.perf_mmap__push
0.00 +5.2 5.19 ± 25% perf-profile.calltrace.cycles-pp.generic_perform_write.shmem_file_write_iter.vfs_write.ksys_write.do_syscall_64
0.00 +5.2 5.19 ± 25% perf-profile.calltrace.cycles-pp.ksys_write.do_syscall_64.entry_SYSCALL_64_after_hwframe.write.writen
0.00 +5.2 5.19 ± 25% perf-profile.calltrace.cycles-pp.record__pushfn.perf_mmap__push.record__mmap_read_evlist.__cmd_record.cmd_record
0.00 +5.2 5.19 ± 25% perf-profile.calltrace.cycles-pp.shmem_file_write_iter.vfs_write.ksys_write.do_syscall_64.entry_SYSCALL_64_after_hwframe
0.00 +5.2 5.19 ± 25% perf-profile.calltrace.cycles-pp.write.writen.record__pushfn.perf_mmap__push.record__mmap_read_evlist
0.00 +5.2 5.19 ± 25% perf-profile.calltrace.cycles-pp.writen.record__pushfn.perf_mmap__push.record__mmap_read_evlist.__cmd_record
0.00 +5.4 5.38 ± 34% perf-profile.calltrace.cycles-pp.exit_mmap.__mmput.exit_mm.do_exit.do_group_exit
0.00 +5.5 5.50 ± 28% perf-profile.calltrace.cycles-pp.vfs_write.ksys_write.do_syscall_64.entry_SYSCALL_64_after_hwframe.write
0.00 +6.2 6.18 ± 9% perf-profile.calltrace.cycles-pp.perf_mmap__push.record__mmap_read_evlist.__cmd_record.cmd_record.run_builtin
0.00 +6.2 6.18 ± 9% perf-profile.calltrace.cycles-pp.record__mmap_read_evlist.__cmd_record.cmd_record.run_builtin.main
0.00 +6.3 6.34 ± 50% perf-profile.calltrace.cycles-pp.do_filp_open.do_sys_openat2.__x64_sys_openat.do_syscall_64.entry_SYSCALL_64_after_hwframe
0.00 +6.3 6.34 ± 50% perf-profile.calltrace.cycles-pp.path_openat.do_filp_open.do_sys_openat2.__x64_sys_openat.do_syscall_64
0.00 +6.8 6.82 ± 41% perf-profile.calltrace.cycles-pp.__handle_mm_fault.handle_mm_fault.do_user_addr_fault.exc_page_fault.asm_exc_page_fault
0.00 +6.9 6.94 ± 23% perf-profile.calltrace.cycles-pp.__cmd_record.cmd_record.run_builtin.main
0.00 +6.9 6.94 ± 23% perf-profile.calltrace.cycles-pp.cmd_record.run_builtin.main
0.00 +8.6 8.58 ± 19% perf-profile.calltrace.cycles-pp.main
0.00 +8.6 8.58 ± 19% perf-profile.calltrace.cycles-pp.run_builtin.main
0.00 +11.4 11.42 ± 30% perf-profile.calltrace.cycles-pp.vsnprintf.seq_printf.show_interrupts.seq_read_iter.proc_reg_read_iter
0.00 +12.2 12.18 ± 26% perf-profile.calltrace.cycles-pp.smp_call_function_single.event_function_call.perf_event_release_kernel.perf_release.__fput
0.00 +12.5 12.53 ± 26% perf-profile.calltrace.cycles-pp.event_function_call.perf_event_release_kernel.perf_release.__fput.task_work_run
0.00 +12.7 12.71 ± 17% perf-profile.calltrace.cycles-pp.seq_printf.show_interrupts.seq_read_iter.proc_reg_read_iter.vfs_read
0.00 +17.2 17.20 ± 17% perf-profile.calltrace.cycles-pp.proc_reg_read_iter.vfs_read.ksys_read.do_syscall_64.entry_SYSCALL_64_after_hwframe
0.00 +17.2 17.20 ± 17% perf-profile.calltrace.cycles-pp.seq_read_iter.proc_reg_read_iter.vfs_read.ksys_read.do_syscall_64
0.00 +17.2 17.20 ± 17% perf-profile.calltrace.cycles-pp.show_interrupts.seq_read_iter.proc_reg_read_iter.vfs_read.ksys_read
0.00 +17.8 17.82 ± 9% perf-profile.calltrace.cycles-pp.perf_event_release_kernel.perf_release.__fput.task_work_run.do_exit
0.00 +17.8 17.82 ± 9% perf-profile.calltrace.cycles-pp.perf_release.__fput.task_work_run.do_exit.do_group_exit
0.00 +21.6 21.56 ± 7% perf-profile.calltrace.cycles-pp.__fput.task_work_run.do_exit.do_group_exit.get_signal
0.00 +22.6 22.64 ± 9% perf-profile.calltrace.cycles-pp.task_work_run.do_exit.do_group_exit.get_signal.arch_do_signal_or_restart
0.00 +24.2 24.16 ± 13% perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.read
0.00 +24.2 24.16 ± 13% perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.read
0.00 +24.2 24.16 ± 13% perf-profile.calltrace.cycles-pp.ksys_read.do_syscall_64.entry_SYSCALL_64_after_hwframe.read
0.00 +24.2 24.16 ± 13% perf-profile.calltrace.cycles-pp.vfs_read.ksys_read.do_syscall_64.entry_SYSCALL_64_after_hwframe.read
0.00 +25.2 25.18 ± 17% perf-profile.calltrace.cycles-pp.read
0.00 +26.8 26.76 ± 9% perf-profile.calltrace.cycles-pp.arch_do_signal_or_restart.syscall_exit_to_user_mode.do_syscall_64.entry_SYSCALL_64_after_hwframe
0.00 +26.8 26.76 ± 9% perf-profile.calltrace.cycles-pp.do_exit.do_group_exit.get_signal.arch_do_signal_or_restart.syscall_exit_to_user_mode
0.00 +26.8 26.76 ± 9% perf-profile.calltrace.cycles-pp.do_group_exit.get_signal.arch_do_signal_or_restart.syscall_exit_to_user_mode.do_syscall_64
0.00 +26.8 26.76 ± 9% perf-profile.calltrace.cycles-pp.get_signal.arch_do_signal_or_restart.syscall_exit_to_user_mode.do_syscall_64.entry_SYSCALL_64_after_hwframe
0.00 +27.4 27.39 ± 8% perf-profile.calltrace.cycles-pp.syscall_exit_to_user_mode.do_syscall_64.entry_SYSCALL_64_after_hwframe
0.00 +41.1 41.08 ± 10% perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe
0.00 +41.1 41.08 ± 10% perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe
77.16 ± 17% -77.2 0.00 perf-profile.children.cycles-pp.do_rw_once
21.07 ± 14% -21.1 0.00 perf-profile.children.cycles-pp.lrand48_r@plt
19.67 ± 71% -19.7 0.00 perf-profile.children.cycles-pp.do_access
17.09 ± 84% -17.1 0.00 perf-profile.children.cycles-pp.hugetlb_fault
13.57 ± 84% -13.6 0.00 perf-profile.children.cycles-pp.hugetlb_wp
13.30 ± 84% -13.3 0.00 perf-profile.children.cycles-pp.copy_user_large_folio
13.28 ± 84% -13.3 0.00 perf-profile.children.cycles-pp.copy_subpage
13.27 ± 84% -12.9 0.38 ±223% perf-profile.children.cycles-pp.copy_mc_enhanced_fast_string
0.00 +2.1 2.06 ± 9% perf-profile.children.cycles-pp.put_files_struct
0.02 ±142% +3.3 3.35 ± 39% perf-profile.children.cycles-pp.common_startup_64
0.02 ±142% +3.3 3.35 ± 39% perf-profile.children.cycles-pp.cpu_startup_entry
0.02 ±142% +3.3 3.35 ± 39% perf-profile.children.cycles-pp.do_idle
0.00 +3.8 3.77 ± 51% perf-profile.children.cycles-pp.kmem_cache_free
0.00 +4.3 4.32 ± 43% perf-profile.children.cycles-pp.sw_perf_event_destroy
0.01 ±223% +4.4 4.37 ± 49% perf-profile.children.cycles-pp.__x64_sys_exit_group
0.00 +4.4 4.41 ± 39% perf-profile.children.cycles-pp.link_path_walk
0.00 +4.6 4.63 ± 41% perf-profile.children.cycles-pp._free_event
0.06 ± 85% +5.1 5.19 ± 25% perf-profile.children.cycles-pp.record__pushfn
0.06 ± 85% +5.1 5.19 ± 25% perf-profile.children.cycles-pp.writen
0.04 ±108% +5.2 5.19 ± 25% perf-profile.children.cycles-pp.generic_perform_write
0.04 ±108% +5.2 5.19 ± 25% perf-profile.children.cycles-pp.shmem_file_write_iter
0.01 ±223% +5.4 5.38 ± 34% perf-profile.children.cycles-pp.__mmput
0.01 ±223% +5.4 5.38 ± 34% perf-profile.children.cycles-pp.exit_mm
0.01 ±223% +5.4 5.38 ± 34% perf-profile.children.cycles-pp.exit_mmap
0.08 ± 98% +5.4 5.50 ± 28% perf-profile.children.cycles-pp.write
0.08 ±100% +5.4 5.50 ± 28% perf-profile.children.cycles-pp.ksys_write
0.08 ± 99% +5.4 5.50 ± 28% perf-profile.children.cycles-pp.vfs_write
0.07 ± 86% +6.1 6.18 ± 9% perf-profile.children.cycles-pp.record__mmap_read_evlist
0.06 ± 85% +6.1 6.18 ± 9% perf-profile.children.cycles-pp.perf_mmap__push
0.00 +6.3 6.34 ± 50% perf-profile.children.cycles-pp.do_filp_open
0.00 +6.3 6.34 ± 50% perf-profile.children.cycles-pp.path_openat
0.00 +6.5 6.50 ± 40% perf-profile.children.cycles-pp.__handle_mm_fault
0.00 +7.0 7.04 ± 54% perf-profile.children.cycles-pp.__x64_sys_openat
0.00 +7.0 7.04 ± 54% perf-profile.children.cycles-pp.do_sys_openat2
0.08 ± 82% +8.5 8.58 ± 19% perf-profile.children.cycles-pp.main
0.08 ± 82% +8.5 8.58 ± 19% perf-profile.children.cycles-pp.run_builtin
0.08 ± 85% +8.5 8.58 ± 19% perf-profile.children.cycles-pp.__cmd_record
0.08 ± 85% +8.5 8.58 ± 19% perf-profile.children.cycles-pp.cmd_record
0.00 +11.7 11.74 ± 30% perf-profile.children.cycles-pp.vsnprintf
0.00 +12.7 12.71 ± 17% perf-profile.children.cycles-pp.seq_printf
0.00 +13.8 13.81 ± 19% perf-profile.children.cycles-pp.smp_call_function_single
0.00 +14.2 14.17 ± 18% perf-profile.children.cycles-pp.event_function_call
0.00 +17.2 17.20 ± 17% perf-profile.children.cycles-pp.proc_reg_read_iter
0.00 +17.2 17.20 ± 17% perf-profile.children.cycles-pp.show_interrupts
0.00 +17.8 17.82 ± 9% perf-profile.children.cycles-pp.perf_event_release_kernel
0.00 +17.8 17.82 ± 9% perf-profile.children.cycles-pp.perf_release
0.00 +22.3 22.30 ± 11% perf-profile.children.cycles-pp.__fput
0.02 ±142% +22.9 22.96 ± 8% perf-profile.children.cycles-pp.task_work_run
0.00 +23.2 23.22 ± 10% perf-profile.children.cycles-pp.seq_read_iter
0.00 +24.5 24.47 ± 13% perf-profile.children.cycles-pp.ksys_read
0.00 +24.5 24.47 ± 13% perf-profile.children.cycles-pp.vfs_read
0.00 +25.2 25.18 ± 13% perf-profile.children.cycles-pp.read
0.00 +26.8 26.76 ± 9% perf-profile.children.cycles-pp.arch_do_signal_or_restart
0.00 +26.8 26.76 ± 9% perf-profile.children.cycles-pp.get_signal
0.00 +27.1 27.08 ± 8% perf-profile.children.cycles-pp.syscall_exit_to_user_mode
0.01 ±223% +31.1 31.13 ± 8% perf-profile.children.cycles-pp.do_exit
0.01 ±223% +31.1 31.13 ± 8% perf-profile.children.cycles-pp.do_group_exit
0.18 ± 66% +85.5 85.67 perf-profile.children.cycles-pp.do_syscall_64
0.18 ± 66% +85.8 86.02 perf-profile.children.cycles-pp.entry_SYSCALL_64_after_hwframe
76.28 ± 17% -76.3 0.00 perf-profile.self.cycles-pp.do_rw_once
13.17 ± 84% -12.8 0.38 ±223% perf-profile.self.cycles-pp.copy_mc_enhanced_fast_string
0.00 +3.7 3.73 ± 53% perf-profile.self.cycles-pp.show_interrupts
0.00 +10.8 10.84 ± 40% perf-profile.self.cycles-pp.smp_call_function_single
***************************************************************************************************
lkp-csl-2sp7: 96 threads 2 sockets Intel(R) Xeon(R) Gold 6252 CPU @ 2.10GHz (Cascade Lake) with 512G memory
=========================================================================================
compiler/cpufreq_governor/kconfig/option_a/rootfs/tbox_group/test/testcase:
gcc-12/performance/x86_64-rhel-8.3/SP.B [Scalar Penta-diagonal solver]/debian-12-x86_64-phoronix/lkp-csl-2sp7/npb-1.4.5/phoronix-test-suite
commit:
9ef3ad0f03 ("ACPI/NUMA: Print CXL Early Discovery Table (CEDT)")
09f84589a8 ("ACPI/NUMA: Remove architecture dependent remainings")
9ef3ad0f032818c5 09f84589a8ed1162c0274a1b5b3
---------------- ---------------------------
%stddev %change %stddev
\ | \
14.98 +1.9% 15.26 boot-time.dhcp
6.328e+09 +62.7% 1.03e+10 cpuidle..time
6773424 +61.9% 10963809 cpuidle..usage
136.23 +50.7% 205.27 uptime.boot
10617 +38.0% 14647 uptime.idle
231376 +16.8% 270293 meminfo.Active
74278 +37.5% 102161 meminfo.Active(anon)
276857 ± 6% +31.8% 364759 ± 3% meminfo.AnonHugePages
11829930 ± 3% -38.6% 7260501 ± 11% meminfo.DirectMap2M
261359 -83.6% 42734 meminfo.VmallocUsed
0.15 -0.1 0.08 ± 2% mpstat.cpu.all.iowait%
1.82 +0.8 2.62 mpstat.cpu.all.irq%
0.09 +0.1 0.14 mpstat.cpu.all.soft%
0.89 +0.3 1.14 mpstat.cpu.all.sys%
20.41 +5.9 26.33 mpstat.cpu.all.usr%
523084 ± 23% -100.0% 0.00 numa-numastat.node0.local_node
577073 ± 19% -100.0% 0.00 numa-numastat.node0.numa_hit
53991 ± 50% -100.0% 0.00 numa-numastat.node0.other_node
560471 ± 22% -100.0% 0.00 numa-numastat.node1.local_node
606158 ± 18% -100.0% 0.00 numa-numastat.node1.numa_hit
45685 ± 60% -100.0% 0.00 numa-numastat.node1.other_node
76.91 -8.8% 70.16 vmstat.cpu.id
19.78 +31.9% 26.08 vmstat.cpu.us
3610 -43.7% 2032 vmstat.io.bi
0.29 ± 48% -65.8% 0.10 ± 6% vmstat.procs.b
21.91 +29.1% 28.29 vmstat.procs.r
6078 ± 4% -27.6% 4398 ± 3% vmstat.system.cs
22219 -58.1% 9301 phoronix-test-suite.npb.SP.B.total_mop_s
84.27 +79.7% 151.42 phoronix-test-suite.time.elapsed_time
84.27 +79.7% 151.42 phoronix-test-suite.time.elapsed_time.max
9092 +61.6% 14688 phoronix-test-suite.time.involuntary_context_switches
1556010 -61.5% 598472 phoronix-test-suite.time.minor_page_faults
2019 +30.6% 2637 phoronix-test-suite.time.percent_of_cpu_this_job_got
64.18 +140.5% 154.34 phoronix-test-suite.time.system_time
1637 +134.4% 3839 phoronix-test-suite.time.user_time
18650 +37.3% 25601 proc-vmstat.nr_active_anon
39298 +6.9% 42026 proc-vmstat.nr_active_file
296059 +7.8% 319032 proc-vmstat.nr_anon_pages
644447 +1.9% 657001 proc-vmstat.nr_file_pages
7113787 +17.7% 8370599 proc-vmstat.nr_foll_pin_acquired
7113780 +17.7% 8370587 proc-vmstat.nr_foll_pin_released
468080 +4.9% 490832 proc-vmstat.nr_inactive_anon
21467 +1.3% 21743 proc-vmstat.nr_kernel_stack
40569 +3.0% 41805 proc-vmstat.nr_mapped
3552 +6.1% 3767 proc-vmstat.nr_page_table_pages
190633 +3.4% 197135 proc-vmstat.nr_shmem
21713 +2.3% 22201 proc-vmstat.nr_slab_reclaimable
18650 +37.3% 25601 proc-vmstat.nr_zone_active_anon
39298 +6.9% 42026 proc-vmstat.nr_zone_active_file
468080 +4.9% 490832 proc-vmstat.nr_zone_inactive_anon
978315 -100.0% 0.00 proc-vmstat.numa_hint_faults
976417 -100.0% 0.00 proc-vmstat.numa_hint_faults_local
1184528 -1.1% 1171481 proc-vmstat.numa_hit
1319 -100.0% 0.00 proc-vmstat.numa_huge_pte_updates
1084853 +8.0% 1171481 proc-vmstat.numa_local
99677 -100.0% 0.00 proc-vmstat.numa_other
3465 ± 63% -100.0% 0.00 proc-vmstat.numa_pages_migrated
1700329 -100.0% 0.00 proc-vmstat.numa_pte_updates
71755 +5.5% 75670 proc-vmstat.pgactivate
1444406 +0.9% 1458081 proc-vmstat.pgalloc_normal
2020069 -49.4% 1022408 proc-vmstat.pgfault
1268206 +3.8% 1316886 proc-vmstat.pgfree
3465 ± 63% -100.0% 0.00 proc-vmstat.pgmigrate_success
100000 ± 47% -100.0% 0.00 numa-meminfo.node0.Active
13486 ± 48% -100.0% 0.00 numa-meminfo.node0.Active(anon)
86514 ± 49% -100.0% 0.00 numa-meminfo.node0.Active(file)
143140 ± 29% -100.0% 0.00 numa-meminfo.node0.AnonHugePages
554668 ± 43% -100.0% 0.00 numa-meminfo.node0.AnonPages
720916 ± 33% -100.0% 0.00 numa-meminfo.node0.AnonPages.max
1320009 ± 59% -100.0% 0.00 numa-meminfo.node0.FilePages
1773835 ± 32% -100.0% 0.00 numa-meminfo.node0.Inactive
678653 ± 49% -100.0% 0.00 numa-meminfo.node0.Inactive(anon)
1095181 ± 64% -100.0% 0.00 numa-meminfo.node0.Inactive(file)
46789 ± 28% -100.0% 0.00 numa-meminfo.node0.KReclaimable
11095 ± 4% -100.0% 0.00 numa-meminfo.node0.KernelStack
101067 ± 28% -100.0% 0.00 numa-meminfo.node0.Mapped
2.611e+08 -100.0% 0.00 numa-meminfo.node0.MemFree
2.638e+08 -100.0% 0.00 numa-meminfo.node0.MemTotal
2732014 ± 23% -100.0% 0.00 numa-meminfo.node0.MemUsed
7763 ± 6% -100.0% 0.00 numa-meminfo.node0.PageTables
46789 ± 28% -100.0% 0.00 numa-meminfo.node0.SReclaimable
125988 ± 9% -100.0% 0.00 numa-meminfo.node0.SUnreclaim
138154 ±125% -100.0% 0.00 numa-meminfo.node0.Shmem
172777 ± 12% -100.0% 0.00 numa-meminfo.node0.Slab
3231 ± 5% -100.0% 0.00 numa-meminfo.node0.Unevictable
131809 ± 35% -100.0% 0.00 numa-meminfo.node1.Active
61136 ± 10% -100.0% 0.00 numa-meminfo.node1.Active(anon)
70672 ± 59% -100.0% 0.00 numa-meminfo.node1.Active(file)
133009 ± 33% -100.0% 0.00 numa-meminfo.node1.AnonHugePages
629614 ± 38% -100.0% 0.00 numa-meminfo.node1.AnonPages
819699 ± 29% -100.0% 0.00 numa-meminfo.node1.AnonPages.max
1257793 ± 62% -100.0% 0.00 numa-meminfo.node1.FilePages
1756533 ± 32% -100.0% 0.00 numa-meminfo.node1.Inactive
1193976 ± 28% -100.0% 0.00 numa-meminfo.node1.Inactive(anon)
562556 ±126% -100.0% 0.00 numa-meminfo.node1.Inactive(file)
40060 ± 33% -100.0% 0.00 numa-meminfo.node1.KReclaimable
10374 ± 4% -100.0% 0.00 numa-meminfo.node1.KernelStack
60225 ± 48% -100.0% 0.00 numa-meminfo.node1.Mapped
2.614e+08 -100.0% 0.00 numa-meminfo.node1.MemFree
2.642e+08 -100.0% 0.00 numa-meminfo.node1.MemTotal
2770706 ± 22% -100.0% 0.00 numa-meminfo.node1.MemUsed
6462 ± 8% -100.0% 0.00 numa-meminfo.node1.PageTables
40060 ± 33% -100.0% 0.00 numa-meminfo.node1.SReclaimable
116360 ± 10% -100.0% 0.00 numa-meminfo.node1.SUnreclaim
624395 ± 27% -100.0% 0.00 numa-meminfo.node1.Shmem
156420 ± 13% -100.0% 0.00 numa-meminfo.node1.Slab
0.01 ± 17% +56.1% 0.02 ± 16% perf-sched.sch_delay.avg.ms.__cond_resched.stop_one_cpu.sched_exec.bprm_execve.part
0.03 ± 51% +147.3% 0.08 ± 39% perf-sched.sch_delay.avg.ms.__x64_sys_pause.do_syscall_64.entry_SYSCALL_64_after_hwframe.[unknown]
0.02 ± 24% +132.8% 0.05 ± 52% perf-sched.sch_delay.avg.ms.do_nanosleep.hrtimer_nanosleep.common_nsleep.__x64_sys_clock_nanosleep
0.01 ± 5% +110.7% 0.02 ± 13% perf-sched.sch_delay.avg.ms.do_task_dead.do_exit.do_group_exit.__x64_sys_exit_group.do_syscall_64
0.02 ± 7% +207.9% 0.05 ± 87% perf-sched.sch_delay.avg.ms.do_wait.kernel_wait4.__do_sys_wait4.do_syscall_64
0.02 ± 71% +159.7% 0.06 ± 37% perf-sched.sch_delay.avg.ms.futex_wait_queue.__futex_wait.futex_wait.do_futex
0.00 +4144.4% 0.13 ±213% perf-sched.sch_delay.avg.ms.pipe_read.vfs_read.ksys_read.do_syscall_64
0.01 ± 16% +135.3% 0.03 ± 33% perf-sched.sch_delay.avg.ms.schedule_hrtimeout_range_clock.do_select.core_sys_select.do_pselect.constprop
0.03 ± 37% +81.9% 0.05 ± 14% perf-sched.sch_delay.avg.ms.schedule_hrtimeout_range_clock.ep_poll.do_epoll_wait.__x64_sys_epoll_wait
0.01 ± 5% +117.5% 0.02 ± 40% perf-sched.sch_delay.avg.ms.schedule_timeout.__wait_for_common.wait_for_completion_state.kernel_clone
0.02 ± 2% +74.2% 0.03 ± 5% perf-sched.sch_delay.avg.ms.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
0.01 ± 18% +70.5% 0.01 ± 12% perf-sched.sch_delay.avg.ms.wait_for_partner.fifo_open.do_dentry_open.do_open
0.05 ± 56% +186.6% 0.14 ± 43% perf-sched.sch_delay.max.ms.__x64_sys_pause.do_syscall_64.entry_SYSCALL_64_after_hwframe.[unknown]
0.09 ± 52% +1385.3% 1.38 ±198% perf-sched.sch_delay.max.ms.do_wait.kernel_wait4.__do_sys_wait4.do_syscall_64
0.06 ± 27% +2.6e+05% 167.11 ±223% perf-sched.sch_delay.max.ms.pipe_read.vfs_read.ksys_read.do_syscall_64
0.04 ± 37% +277.6% 0.13 ± 44% perf-sched.sch_delay.max.ms.schedule_hrtimeout_range_clock.do_select.core_sys_select.do_pselect.constprop
0.05 ± 22% +190.8% 0.14 ± 24% perf-sched.sch_delay.max.ms.schedule_hrtimeout_range_clock.ep_poll.do_epoll_wait.__x64_sys_epoll_wait
0.05 ± 10% +3160.1% 1.61 ±197% perf-sched.sch_delay.max.ms.schedule_hrtimeout_range_clock.ep_poll.do_epoll_wait.do_epoll_pwait.part
0.04 ± 21% +222.5% 0.14 ± 78% perf-sched.sch_delay.max.ms.schedule_timeout.__wait_for_common.wait_for_completion_state.kernel_clone
0.04 ± 33% +193.0% 0.12 ± 68% perf-sched.sch_delay.max.ms.schedule_timeout.rcu_gp_fqs_loop.rcu_gp_kthread.kthread
8369 ± 7% -10.6% 7478 ± 4% perf-sched.total_wait_and_delay.count.ms
253.51 ± 3% +22.4% 310.30 ± 2% perf-sched.wait_and_delay.avg.ms.do_task_dead.do_exit.do_group_exit.__x64_sys_exit_group.do_syscall_64
1.03 ± 5% +106.4% 2.13 ± 5% perf-sched.wait_and_delay.avg.ms.do_wait.kernel_wait4.__do_sys_wait4.do_syscall_64
0.10 ± 5% +73.2% 0.18 ± 8% perf-sched.wait_and_delay.avg.ms.irqentry_exit_to_user_mode.asm_sysvec_apic_timer_interrupt.[unknown].[unknown]
1.51 ± 4% +62.7% 2.46 ± 8% perf-sched.wait_and_delay.avg.ms.schedule_timeout.__wait_for_common.wait_for_completion_state.kernel_clone
124.00 -48.8% 63.50 perf-sched.wait_and_delay.count.do_task_dead.do_exit.do_group_exit.__x64_sys_exit_group.do_syscall_64
123.00 -49.3% 62.33 perf-sched.wait_and_delay.count.do_wait.kernel_wait4.__do_sys_wait4.do_syscall_64
2125 -34.1% 1400 ± 27% perf-sched.wait_and_delay.count.pipe_read.vfs_read.ksys_read.do_syscall_64
86.00 -52.3% 41.00 perf-sched.wait_and_delay.count.schedule_timeout.__wait_for_common.wait_for_completion_state.kernel_clone
19.83 -49.6% 10.00 perf-sched.wait_and_delay.count.schedule_timeout.kcompactd.kthread.ret_from_fork
550.00 ± 2% -12.2% 482.67 ± 2% perf-sched.wait_and_delay.count.worker_thread.kthread.ret_from_fork.ret_from_fork_asm
21.99 ± 3% +26.6% 27.84 ± 7% perf-sched.wait_and_delay.max.ms.do_wait.kernel_wait4.__do_sys_wait4.do_syscall_64
5.86 ± 6% +78.7% 10.47 ± 8% perf-sched.wait_and_delay.max.ms.schedule_timeout.__wait_for_common.wait_for_completion_state.kernel_clone
253.50 ± 3% +22.4% 310.28 ± 2% perf-sched.wait_time.avg.ms.do_task_dead.do_exit.do_group_exit.__x64_sys_exit_group.do_syscall_64
1.02 ± 5% +104.8% 2.08 ± 7% perf-sched.wait_time.avg.ms.do_wait.kernel_wait4.__do_sys_wait4.do_syscall_64
0.10 ± 17% +66.2% 0.17 ± 11% perf-sched.wait_time.avg.ms.irqentry_exit_to_user_mode.asm_sysvec_apic_timer_interrupt.[unknown]
0.10 ± 5% +73.2% 0.18 ± 8% perf-sched.wait_time.avg.ms.irqentry_exit_to_user_mode.asm_sysvec_apic_timer_interrupt.[unknown].[unknown]
0.04 ±110% +332.7% 0.16 ± 57% perf-sched.wait_time.avg.ms.irqentry_exit_to_user_mode.asm_sysvec_call_function_single.[unknown].[unknown]
1.50 ± 4% +62.3% 2.43 ± 8% perf-sched.wait_time.avg.ms.schedule_timeout.__wait_for_common.wait_for_completion_state.kernel_clone
0.00 ± 40% +750.0% 0.02 ±122% perf-sched.wait_time.avg.ms.wait_for_partner.fifo_open.do_dentry_open.do_open
21.97 ± 3% +26.6% 27.82 ± 7% perf-sched.wait_time.max.ms.do_wait.kernel_wait4.__do_sys_wait4.do_syscall_64
0.35 ± 16% +48.7% 0.52 ± 7% perf-sched.wait_time.max.ms.irqentry_exit_to_user_mode.asm_sysvec_apic_timer_interrupt.[unknown]
0.06 ±108% +379.9% 0.29 ± 50% perf-sched.wait_time.max.ms.irqentry_exit_to_user_mode.asm_sysvec_call_function_single.[unknown].[unknown]
5.84 ± 6% +78.8% 10.44 ± 8% perf-sched.wait_time.max.ms.schedule_timeout.__wait_for_common.wait_for_completion_state.kernel_clone
0.03 ± 16% +851.5% 0.31 ±157% perf-sched.wait_time.max.ms.wait_for_partner.fifo_open.do_dentry_open.do_open
3375 ± 48% -100.0% 0.00 numa-vmstat.node0.nr_active_anon
21639 ± 49% -100.0% 0.00 numa-vmstat.node0.nr_active_file
138732 ± 43% -100.0% 0.00 numa-vmstat.node0.nr_anon_pages
69.78 ± 29% -100.0% 0.00 numa-vmstat.node0.nr_anon_transparent_hugepages
330019 ± 59% -100.0% 0.00 numa-vmstat.node0.nr_file_pages
3551956 -100.0% 0.00 numa-vmstat.node0.nr_foll_pin_acquired
3551953 -100.0% 0.00 numa-vmstat.node0.nr_foll_pin_released
65277218 -100.0% 0.00 numa-vmstat.node0.nr_free_pages
169733 ± 49% -100.0% 0.00 numa-vmstat.node0.nr_inactive_anon
273801 ± 64% -100.0% 0.00 numa-vmstat.node0.nr_inactive_file
11096 ± 4% -100.0% 0.00 numa-vmstat.node0.nr_kernel_stack
25457 ± 28% -100.0% 0.00 numa-vmstat.node0.nr_mapped
1937 ± 6% -100.0% 0.00 numa-vmstat.node0.nr_page_table_pages
34539 ±125% -100.0% 0.00 numa-vmstat.node0.nr_shmem
11695 ± 28% -100.0% 0.00 numa-vmstat.node0.nr_slab_reclaimable
31493 ± 9% -100.0% 0.00 numa-vmstat.node0.nr_slab_unreclaimable
807.98 ± 5% -100.0% 0.00 numa-vmstat.node0.nr_unevictable
3375 ± 48% -100.0% 0.00 numa-vmstat.node0.nr_zone_active_anon
21639 ± 49% -100.0% 0.00 numa-vmstat.node0.nr_zone_active_file
169735 ± 49% -100.0% 0.00 numa-vmstat.node0.nr_zone_inactive_anon
273801 ± 64% -100.0% 0.00 numa-vmstat.node0.nr_zone_inactive_file
807.98 ± 5% -100.0% 0.00 numa-vmstat.node0.nr_zone_unevictable
576852 ± 19% -100.0% 0.00 numa-vmstat.node0.numa_hit
522864 ± 24% -100.0% 0.00 numa-vmstat.node0.numa_local
53991 ± 50% -100.0% 0.00 numa-vmstat.node0.numa_other
15300 ± 10% -100.0% 0.00 numa-vmstat.node1.nr_active_anon
17669 ± 59% -100.0% 0.00 numa-vmstat.node1.nr_active_file
157621 ± 38% -100.0% 0.00 numa-vmstat.node1.nr_anon_pages
314384 ± 62% -100.0% 0.00 numa-vmstat.node1.nr_file_pages
3552396 -100.0% 0.00 numa-vmstat.node1.nr_foll_pin_acquired
3552390 -100.0% 0.00 numa-vmstat.node1.nr_foll_pin_released
65353661 -100.0% 0.00 numa-vmstat.node1.nr_free_pages
298619 ± 28% -100.0% 0.00 numa-vmstat.node1.nr_inactive_anon
140594 ±126% -100.0% 0.00 numa-vmstat.node1.nr_inactive_file
10373 ± 4% -100.0% 0.00 numa-vmstat.node1.nr_kernel_stack
15147 ± 48% -100.0% 0.00 numa-vmstat.node1.nr_mapped
1615 ± 8% -100.0% 0.00 numa-vmstat.node1.nr_page_table_pages
156079 ± 27% -100.0% 0.00 numa-vmstat.node1.nr_shmem
10015 ± 33% -100.0% 0.00 numa-vmstat.node1.nr_slab_reclaimable
29089 ± 10% -100.0% 0.00 numa-vmstat.node1.nr_slab_unreclaimable
15300 ± 10% -100.0% 0.00 numa-vmstat.node1.nr_zone_active_anon
17669 ± 59% -100.0% 0.00 numa-vmstat.node1.nr_zone_active_file
298620 ± 28% -100.0% 0.00 numa-vmstat.node1.nr_zone_inactive_anon
140594 ±126% -100.0% 0.00 numa-vmstat.node1.nr_zone_inactive_file
605836 ± 18% -100.0% 0.00 numa-vmstat.node1.numa_hit
560149 ± 22% -100.0% 0.00 numa-vmstat.node1.numa_local
45685 ± 60% -100.0% 0.00 numa-vmstat.node1.numa_other
9.24 -26.9% 6.75 perf-stat.i.MPKI
3.006e+09 ± 3% +30.6% 3.924e+09 perf-stat.i.branch-instructions
2.88 -0.7 2.19 ± 2% perf-stat.i.branch-miss-rate%
34648844 -23.3% 26559921 perf-stat.i.branch-misses
44.09 +8.1 52.19 perf-stat.i.cache-miss-rate%
4.032e+08 -44.5% 2.239e+08 perf-stat.i.cache-misses
6.104e+08 -42.7% 3.496e+08 perf-stat.i.cache-references
6150 ± 4% -27.2% 4479 ± 3% perf-stat.i.context-switches
1.98 ± 2% +40.0% 2.78 perf-stat.i.cpi
5.387e+10 +33.5% 7.189e+10 perf-stat.i.cpu-cycles
156.13 -9.6% 141.17 perf-stat.i.cpu-migrations
666.32 ± 3% -13.9% 574.01 perf-stat.i.cycles-between-cache-misses
1379778 -36.7% 873874 ± 2% perf-stat.i.dTLB-load-misses
1.241e+10 ± 2% -5.1% 1.178e+10 perf-stat.i.dTLB-loads
0.04 ± 4% -0.0 0.03 ± 3% perf-stat.i.dTLB-store-miss-rate%
976671 -45.5% 532272 perf-stat.i.dTLB-store-misses
4.839e+09 -12.5% 4.236e+09 perf-stat.i.dTLB-stores
41.13 ± 4% -3.8 37.28 perf-stat.i.iTLB-load-miss-rate%
1105884 -24.3% 837557 perf-stat.i.iTLB-load-misses
1971195 ± 11% -18.1% 1613463 perf-stat.i.iTLB-loads
26637 ± 2% +45.0% 38621 perf-stat.i.instructions-per-iTLB-miss
0.61 -26.0% 0.45 perf-stat.i.ipc
12.95 ± 3% -33.3% 8.64 ± 3% perf-stat.i.major-faults
0.56 +33.4% 0.75 perf-stat.i.metric.GHz
1256 -37.8% 781.77 perf-stat.i.metric.K/sec
217.05 ± 2% -2.8% 210.92 perf-stat.i.metric.M/sec
22076 -73.0% 5958 perf-stat.i.minor-faults
30.69 ± 2% +26.8 57.44 perf-stat.i.node-load-miss-rate%
5638356 +188.6% 16272394 perf-stat.i.node-load-misses
51095437 -68.4% 16149472 perf-stat.i.node-loads
24.17 ± 5% +33.5 57.66 perf-stat.i.node-store-miss-rate%
3634105 +207.6% 11177771 perf-stat.i.node-store-misses
33521254 -68.1% 10709700 perf-stat.i.node-stores
22089 -73.0% 5966 perf-stat.i.page-faults
13.73 -44.7% 7.60 perf-stat.overall.MPKI
1.12 ± 2% -0.5 0.65 perf-stat.overall.branch-miss-rate%
66.19 -1.9 64.29 perf-stat.overall.cache-miss-rate%
1.83 +33.2% 2.44 perf-stat.overall.cpi
133.26 +140.7% 320.76 perf-stat.overall.cycles-between-cache-misses
0.01 -0.0 0.01 ± 3% perf-stat.overall.dTLB-load-miss-rate%
0.02 -0.0 0.01 perf-stat.overall.dTLB-store-miss-rate%
27181 ± 2% +33.5% 36284 perf-stat.overall.instructions-per-iTLB-miss
0.55 -25.0% 0.41 perf-stat.overall.ipc
9.93 +40.3 50.21 perf-stat.overall.node-load-miss-rate%
9.78 +41.3 51.05 perf-stat.overall.node-store-miss-rate%
3.017e+09 ± 3% +31.3% 3.961e+09 perf-stat.ps.branch-instructions
33863841 -23.7% 25852545 perf-stat.ps.branch-misses
4.08e+08 -44.2% 2.279e+08 perf-stat.ps.cache-misses
6.164e+08 -42.5% 3.544e+08 perf-stat.ps.cache-references
5959 ± 4% -27.9% 4295 ± 4% perf-stat.ps.context-switches
5.437e+10 +34.4% 7.309e+10 perf-stat.ps.cpu-cycles
153.77 -10.2% 138.05 perf-stat.ps.cpu-migrations
1362804 -37.6% 851037 ± 2% perf-stat.ps.dTLB-load-misses
1.253e+10 ± 2% -4.7% 1.194e+10 perf-stat.ps.dTLB-loads
982670 -45.5% 535635 perf-stat.ps.dTLB-store-misses
4.886e+09 -12.1% 4.294e+09 perf-stat.ps.dTLB-stores
1093709 -24.4% 826521 perf-stat.ps.iTLB-load-misses
1967627 ± 10% -17.5% 1623220 perf-stat.ps.iTLB-loads
12.44 ± 3% -35.3% 8.05 ± 3% perf-stat.ps.major-faults
21849 -74.4% 5599 perf-stat.ps.minor-faults
5700455 +191.0% 16587479 perf-stat.ps.node-load-misses
51713905 -68.2% 16452569 perf-stat.ps.node-loads
3675605 +209.5% 11374853 perf-stat.ps.node-store-misses
33923436 -67.8% 10906955 perf-stat.ps.node-stores
21862 -74.3% 5607 perf-stat.ps.page-faults
2.516e+12 +82.7% 4.596e+12 perf-stat.total.instructions
7647 ± 15% +7433.9% 576144 ± 2% sched_debug.cfs_rq:/.avg_vruntime.avg
30894 ± 29% +3982.0% 1261110 ± 6% sched_debug.cfs_rq:/.avg_vruntime.max
2728 ± 33% +1541.3% 44777 ± 91% sched_debug.cfs_rq:/.avg_vruntime.min
4017 ± 12% +7295.9% 297155 ± 5% sched_debug.cfs_rq:/.avg_vruntime.stddev
0.07 ± 33% +400.4% 0.37 ± 4% sched_debug.cfs_rq:/.h_nr_running.avg
0.26 ± 14% +62.7% 0.42 ± 2% sched_debug.cfs_rq:/.h_nr_running.stddev
1477 ± 93% +658.6% 11208 ± 24% sched_debug.cfs_rq:/.load.avg
7647 ± 15% +7433.9% 576145 ± 2% sched_debug.cfs_rq:/.min_vruntime.avg
30894 ± 29% +3982.0% 1261110 ± 6% sched_debug.cfs_rq:/.min_vruntime.max
2728 ± 33% +1541.3% 44777 ± 91% sched_debug.cfs_rq:/.min_vruntime.min
4017 ± 12% +7295.9% 297155 ± 5% sched_debug.cfs_rq:/.min_vruntime.stddev
0.07 ± 33% +400.4% 0.37 ± 4% sched_debug.cfs_rq:/.nr_running.avg
0.26 ± 14% +62.7% 0.42 ± 2% sched_debug.cfs_rq:/.nr_running.stddev
250.19 ± 8% +65.6% 414.25 ± 4% sched_debug.cfs_rq:/.runnable_avg.avg
276.31 ± 5% +48.5% 410.18 sched_debug.cfs_rq:/.runnable_avg.stddev
249.62 ± 8% +65.5% 413.18 ± 4% sched_debug.cfs_rq:/.util_avg.avg
276.00 ± 5% +48.6% 410.18 sched_debug.cfs_rq:/.util_avg.stddev
23.11 ± 19% +497.9% 138.20 ± 5% sched_debug.cfs_rq:/.util_est.avg
845.17 ± 8% -16.7% 704.06 ± 9% sched_debug.cfs_rq:/.util_est.max
109.08 ± 8% +65.1% 180.12 ± 4% sched_debug.cfs_rq:/.util_est.stddev
6.61 ±128% +214.3% 20.79 ± 70% sched_debug.cfs_rq:/init.scope.avg_vruntime.min
6.61 ±128% +214.3% 20.79 ± 70% sched_debug.cfs_rq:/init.scope.min_vruntime.min
50610 +61.8% 81883 ± 34% sched_debug.cfs_rq:/init.scope.se->exec_start.avg
50661 +61.9% 82013 ± 34% sched_debug.cfs_rq:/init.scope.se->exec_start.max
50549 +61.5% 81627 ± 35% sched_debug.cfs_rq:/init.scope.se->exec_start.min
153.50 ±153% +494.9% 913.22 ± 62% sched_debug.cfs_rq:/init.scope.se->load.weight.min
6.80 ±133% +216.4% 21.51 ± 69% sched_debug.cfs_rq:/init.scope.se->sum_exec_runtime.min
6475 ± 32% +1515.3% 104591 ± 91% sched_debug.cfs_rq:/init.scope.se->vruntime.min
13999 ± 20% -66.0% 4760 ± 33% sched_debug.cfs_rq:/init.scope.tg_load_avg.avg
13999 ± 20% -66.0% 4760 ± 33% sched_debug.cfs_rq:/init.scope.tg_load_avg.max
13999 ± 20% -66.0% 4760 ± 33% sched_debug.cfs_rq:/init.scope.tg_load_avg.min
3518 ± 15% +18809.0% 665385 ± 2% sched_debug.cfs_rq:/system.slice.avg_vruntime.avg
23752 ± 49% +5906.4% 1426667 ± 7% sched_debug.cfs_rq:/system.slice.avg_vruntime.max
1472 ± 31% +3291.4% 49946 ± 90% sched_debug.cfs_rq:/system.slice.avg_vruntime.min
2920 ± 39% +11804.1% 347701 ± 7% sched_debug.cfs_rq:/system.slice.avg_vruntime.stddev
0.09 ± 34% +424.6% 0.46 ± 4% sched_debug.cfs_rq:/system.slice.h_nr_running.avg
0.28 ± 14% +46.3% 0.41 ± 2% sched_debug.cfs_rq:/system.slice.h_nr_running.stddev
2067 ± 52% +412.4% 10591 ± 6% sched_debug.cfs_rq:/system.slice.load.avg
10.36 ± 6% +16.9% 12.12 ± 4% sched_debug.cfs_rq:/system.slice.load_avg.avg
68.67 ± 21% -47.2% 36.28 ± 17% sched_debug.cfs_rq:/system.slice.load_avg.max
13.24 ± 10% -20.1% 10.58 ± 6% sched_debug.cfs_rq:/system.slice.load_avg.stddev
3518 ± 15% +18809.0% 665385 ± 2% sched_debug.cfs_rq:/system.slice.min_vruntime.avg
23752 ± 49% +5906.4% 1426667 ± 7% sched_debug.cfs_rq:/system.slice.min_vruntime.max
1472 ± 31% +3291.4% 49946 ± 90% sched_debug.cfs_rq:/system.slice.min_vruntime.min
2920 ± 39% +11804.1% 347701 ± 7% sched_debug.cfs_rq:/system.slice.min_vruntime.stddev
0.09 ± 34% +424.6% 0.46 ± 4% sched_debug.cfs_rq:/system.slice.nr_running.avg
0.28 ± 14% +46.3% 0.41 ± 2% sched_debug.cfs_rq:/system.slice.nr_running.stddev
287.37 ± 7% +75.3% 503.83 ± 5% sched_debug.cfs_rq:/system.slice.runnable_avg.avg
277.25 ± 5% +43.0% 396.44 sched_debug.cfs_rq:/system.slice.runnable_avg.stddev
6.50 ± 55% +91.9% 12.48 ± 8% sched_debug.cfs_rq:/system.slice.se->avg.load_avg.avg
287.37 ± 7% +75.3% 503.88 ± 5% sched_debug.cfs_rq:/system.slice.se->avg.runnable_avg.avg
277.25 ± 5% +43.0% 396.37 sched_debug.cfs_rq:/system.slice.se->avg.runnable_avg.stddev
287.35 ± 7% +75.3% 503.74 ± 5% sched_debug.cfs_rq:/system.slice.se->avg.util_avg.avg
277.24 ± 5% +43.0% 396.49 sched_debug.cfs_rq:/system.slice.se->avg.util_avg.stddev
50573 +116.7% 109592 sched_debug.cfs_rq:/system.slice.se->exec_start.avg
50741 +117.9% 110587 sched_debug.cfs_rq:/system.slice.se->exec_start.max
43876 +137.1% 104039 sched_debug.cfs_rq:/system.slice.se->exec_start.min
745.42 +42.9% 1065 ± 9% sched_debug.cfs_rq:/system.slice.se->exec_start.stddev
327582 ± 23% -63.1% 120920 ± 31% sched_debug.cfs_rq:/system.slice.se->load.weight.max
48399 ± 22% -58.8% 19928 ± 25% sched_debug.cfs_rq:/system.slice.se->load.weight.stddev
330.49 ± 22% +4461.4% 15074 sched_debug.cfs_rq:/system.slice.se->sum_exec_runtime.avg
10758 ± 44% +226.9% 35171 ± 7% sched_debug.cfs_rq:/system.slice.se->sum_exec_runtime.max
29.77 ± 30% +4193.1% 1277 ± 84% sched_debug.cfs_rq:/system.slice.se->sum_exec_runtime.min
1382 ± 35% +484.7% 8085 ± 7% sched_debug.cfs_rq:/system.slice.se->sum_exec_runtime.stddev
7861 ± 13% +7258.1% 578439 sched_debug.cfs_rq:/system.slice.se->vruntime.avg
29255 ± 32% +4127.3% 1236710 ± 6% sched_debug.cfs_rq:/system.slice.se->vruntime.max
2957 ± 30% +1554.1% 48917 ± 83% sched_debug.cfs_rq:/system.slice.se->vruntime.min
4044 ± 16% +7281.5% 298577 ± 6% sched_debug.cfs_rq:/system.slice.se->vruntime.stddev
302.45 ± 62% -77.0% 69.42 ± 88% sched_debug.cfs_rq:/system.slice.tg_load_avg.stddev
287.35 ± 7% +75.3% 503.74 ± 5% sched_debug.cfs_rq:/system.slice.util_avg.avg
277.24 ± 5% +43.0% 396.48 sched_debug.cfs_rq:/system.slice.util_avg.stddev
1.57 ±193% +1094.7% 18.73 ± 51% sched_debug.cfs_rq:/system.slice/containerd.service.avg_vruntime.min
0.86 ± 26% +364.2% 3.97 ± 16% sched_debug.cfs_rq:/system.slice/containerd.service.load_avg.avg
3.83 ± 55% +140.6% 9.22 ± 19% sched_debug.cfs_rq:/system.slice/containerd.service.load_avg.max
1.41 ± 48% +148.1% 3.51 ± 22% sched_debug.cfs_rq:/system.slice/containerd.service.load_avg.stddev
1.57 ±193% +1094.7% 18.73 ± 51% sched_debug.cfs_rq:/system.slice/containerd.service.min_vruntime.min
0.88 ± 28% +355.8% 4.01 ± 16% sched_debug.cfs_rq:/system.slice/containerd.service.runnable_avg.avg
4.00 ± 55% +130.6% 9.22 ± 19% sched_debug.cfs_rq:/system.slice/containerd.service.runnable_avg.max
1.47 ± 49% +138.4% 3.50 ± 22% sched_debug.cfs_rq:/system.slice/containerd.service.runnable_avg.stddev
0.26 ± 92% +390.2% 1.29 ± 20% sched_debug.cfs_rq:/system.slice/containerd.service.se->avg.load_avg.avg
0.85 ± 32% +368.2% 3.99 ± 16% sched_debug.cfs_rq:/system.slice/containerd.service.se->avg.runnable_avg.avg
3.83 ± 61% +137.7% 9.11 ± 18% sched_debug.cfs_rq:/system.slice/containerd.service.se->avg.runnable_avg.max
1.41 ± 53% +145.3% 3.46 ± 21% sched_debug.cfs_rq:/system.slice/containerd.service.se->avg.runnable_avg.stddev
0.85 ± 32% +365.3% 3.96 ± 17% sched_debug.cfs_rq:/system.slice/containerd.service.se->avg.util_avg.avg
3.83 ± 61% +136.2% 9.06 ± 19% sched_debug.cfs_rq:/system.slice/containerd.service.se->avg.util_avg.max
1.41 ± 53% +143.9% 3.44 ± 22% sched_debug.cfs_rq:/system.slice/containerd.service.se->avg.util_avg.stddev
50173 +118.7% 109726 sched_debug.cfs_rq:/system.slice/containerd.service.se->exec_start.avg
50702 +117.3% 110151 sched_debug.cfs_rq:/system.slice/containerd.service.se->exec_start.max
48149 ± 6% +125.9% 108746 sched_debug.cfs_rq:/system.slice/containerd.service.se->exec_start.min
738718 ± 15% -20.1% 590310 ± 6% sched_debug.cfs_rq:/system.slice/containerd.service.se->load.weight.max
2.11 ±154% +825.1% 19.48 ± 48% sched_debug.cfs_rq:/system.slice/containerd.service.se->sum_exec_runtime.min
3846 ± 24% +12555.9% 486815 ± 20% sched_debug.cfs_rq:/system.slice/containerd.service.se->vruntime.avg
8152 ± 52% +8857.9% 730249 ± 21% sched_debug.cfs_rq:/system.slice/containerd.service.se->vruntime.max
2008 ± 22% +14293.3% 289128 ± 44% sched_debug.cfs_rq:/system.slice/containerd.service.se->vruntime.min
2041 ± 71% +8227.2% 170022 ± 38% sched_debug.cfs_rq:/system.slice/containerd.service.se->vruntime.stddev
5.20 ± 39% +266.8% 19.08 ± 22% sched_debug.cfs_rq:/system.slice/containerd.service.tg_load_avg.avg
5.33 ± 38% +293.8% 21.00 ± 25% sched_debug.cfs_rq:/system.slice/containerd.service.tg_load_avg.max
5.00 ± 38% +255.6% 17.78 ± 22% sched_debug.cfs_rq:/system.slice/containerd.service.tg_load_avg.min
0.16 ±141% +651.8% 1.19 ± 70% sched_debug.cfs_rq:/system.slice/containerd.service.tg_load_avg.stddev
0.86 ± 26% +361.1% 3.95 ± 17% sched_debug.cfs_rq:/system.slice/containerd.service.tg_load_avg_contrib.avg
3.83 ± 55% +140.6% 9.22 ± 19% sched_debug.cfs_rq:/system.slice/containerd.service.tg_load_avg_contrib.max
1.41 ± 48% +146.1% 3.48 ± 22% sched_debug.cfs_rq:/system.slice/containerd.service.tg_load_avg_contrib.stddev
0.88 ± 28% +353.0% 3.99 ± 17% sched_debug.cfs_rq:/system.slice/containerd.service.util_avg.avg
4.00 ± 55% +129.2% 9.17 ± 20% sched_debug.cfs_rq:/system.slice/containerd.service.util_avg.max
1.47 ± 49% +137.0% 3.48 ± 22% sched_debug.cfs_rq:/system.slice/containerd.service.util_avg.stddev
50765 +119.4% 111370 sched_debug.cpu.clock.avg
50771 +119.4% 111382 sched_debug.cpu.clock.max
50759 +119.4% 111356 sched_debug.cpu.clock.min
3.27 ± 10% +126.1% 7.40 ± 21% sched_debug.cpu.clock.stddev
50601 +116.6% 109610 sched_debug.cpu.clock_task.avg
50745 +118.0% 110614 sched_debug.cpu.clock_task.max
43880 +134.7% 102986 sched_debug.cpu.clock_task.min
696.32 +51.9% 1057 ± 6% sched_debug.cpu.clock_task.stddev
165.30 ± 33% +792.2% 1474 ± 2% sched_debug.cpu.curr->pid.avg
2493 +85.2% 4618 sched_debug.cpu.curr->pid.max
587.01 ± 16% +230.2% 1938 sched_debug.cpu.curr->pid.stddev
0.00 ± 22% +76.0% 0.00 ± 7% sched_debug.cpu.next_balance.stddev
0.07 ± 28% +293.0% 0.29 ± 3% sched_debug.cpu.nr_running.avg
0.26 ± 12% +60.0% 0.42 ± 2% sched_debug.cpu.nr_running.stddev
1606 ± 3% +188.0% 4627 ± 3% sched_debug.cpu.nr_switches.avg
11893 ± 21% +400.8% 59557 ± 2% sched_debug.cpu.nr_switches.max
254.67 ± 13% +299.7% 1017 ± 17% sched_debug.cpu.nr_switches.min
2062 ± 11% +220.6% 6611 ± 3% sched_debug.cpu.nr_switches.stddev
0.14 ± 17% -76.4% 0.03 ± 14% sched_debug.cpu.nr_uninterruptible.avg
50761 +119.4% 111355 sched_debug.cpu_clk
50082 +121.0% 110688 sched_debug.ktime
51315 +118.1% 111904 sched_debug.sched_clk
19.88 ± 35% -19.9 0.00 perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe
19.88 ± 35% -19.9 0.00 perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe
19.20 ± 16% -19.2 0.00 perf-profile.calltrace.cycles-pp.write
17.84 ± 19% -17.8 0.00 perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.write
17.84 ± 19% -17.8 0.00 perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.write
17.84 ± 19% -17.8 0.00 perf-profile.calltrace.cycles-pp.generic_perform_write.shmem_file_write_iter.vfs_write.ksys_write.do_syscall_64
17.84 ± 19% -17.8 0.00 perf-profile.calltrace.cycles-pp.ksys_write.do_syscall_64.entry_SYSCALL_64_after_hwframe.write
17.84 ± 19% -17.8 0.00 perf-profile.calltrace.cycles-pp.shmem_file_write_iter.vfs_write.ksys_write.do_syscall_64.entry_SYSCALL_64_after_hwframe
17.84 ± 19% -17.8 0.00 perf-profile.calltrace.cycles-pp.vfs_write.ksys_write.do_syscall_64.entry_SYSCALL_64_after_hwframe.write
12.83 ± 44% -12.8 0.00 perf-profile.calltrace.cycles-pp.arch_do_signal_or_restart.syscall_exit_to_user_mode.do_syscall_64.entry_SYSCALL_64_after_hwframe
12.83 ± 44% -12.8 0.00 perf-profile.calltrace.cycles-pp.do_exit.do_group_exit.get_signal.arch_do_signal_or_restart.syscall_exit_to_user_mode
12.83 ± 44% -12.8 0.00 perf-profile.calltrace.cycles-pp.do_group_exit.get_signal.arch_do_signal_or_restart.syscall_exit_to_user_mode.do_syscall_64
12.83 ± 44% -12.8 0.00 perf-profile.calltrace.cycles-pp.get_signal.arch_do_signal_or_restart.syscall_exit_to_user_mode.do_syscall_64.entry_SYSCALL_64_after_hwframe
12.83 ± 44% -12.8 0.00 perf-profile.calltrace.cycles-pp.syscall_exit_to_user_mode.do_syscall_64.entry_SYSCALL_64_after_hwframe
11.79 ± 52% -11.8 0.00 perf-profile.calltrace.cycles-pp.task_work_run.do_exit.do_group_exit.get_signal.arch_do_signal_or_restart
10.96 ± 48% -11.0 0.00 perf-profile.calltrace.cycles-pp.__fput.task_work_run.do_exit.do_group_exit.get_signal
9.79 ± 55% -9.8 0.00 perf-profile.calltrace.cycles-pp.perf_event_release_kernel.perf_release.__fput.task_work_run.do_exit
9.79 ± 55% -9.8 0.00 perf-profile.calltrace.cycles-pp.perf_release.__fput.task_work_run.do_exit.do_group_exit
9.42 ± 39% -9.4 0.00 perf-profile.calltrace.cycles-pp.asm_exc_page_fault.fault_in_readable.fault_in_iov_iter_readable.generic_perform_write.shmem_file_write_iter
6.64 ± 37% -6.6 0.00 perf-profile.calltrace.cycles-pp.event_function_call.perf_event_release_kernel.perf_release.__fput.task_work_run
6.55 ± 65% -6.5 0.00 perf-profile.calltrace.cycles-pp.asm_exc_page_fault
6.41 ± 20% -6.4 0.00 perf-profile.calltrace.cycles-pp.fault_in_iov_iter_readable.generic_perform_write.shmem_file_write_iter.vfs_write.ksys_write
6.35 ± 42% -6.4 0.00 perf-profile.calltrace.cycles-pp.smp_call_function_single.event_function_call.perf_event_release_kernel.perf_release.__fput
6.07 ± 62% -6.1 0.00 perf-profile.calltrace.cycles-pp.do_user_addr_fault.exc_page_fault.asm_exc_page_fault
6.07 ± 62% -6.1 0.00 perf-profile.calltrace.cycles-pp.exc_page_fault.asm_exc_page_fault
5.99 ± 73% -6.0 0.00 perf-profile.calltrace.cycles-pp.__handle_mm_fault.handle_mm_fault.do_user_addr_fault.exc_page_fault.asm_exc_page_fault
5.54 ± 55% -5.5 0.00 perf-profile.calltrace.cycles-pp.copy_page_from_iter_atomic.generic_perform_write.shmem_file_write_iter.vfs_write.ksys_write
5.54 ± 55% -5.5 0.00 perf-profile.calltrace.cycles-pp.rep_movs_alternative.copy_page_from_iter_atomic.generic_perform_write.shmem_file_write_iter.vfs_write
5.03 ± 78% -5.0 0.00 perf-profile.calltrace.cycles-pp.handle_mm_fault.do_user_addr_fault.exc_page_fault.asm_exc_page_fault
0.00 +1.0 0.96 ± 13% perf-profile.calltrace.cycles-pp.opal_progress
0.00 +1.8 1.83 perf-profile.calltrace.cycles-pp.tzetar_.z_solve_.adi_
0.00 +2.2 2.16 ± 4% perf-profile.calltrace.cycles-pp.rep_movs_alternative._copy_to_iter.copy_page_to_iter.process_vm_rw_single_vec.process_vm_rw_core
0.00 +2.2 2.20 ± 4% perf-profile.calltrace.cycles-pp._copy_to_iter.copy_page_to_iter.process_vm_rw_single_vec.process_vm_rw_core.process_vm_rw
0.00 +2.2 2.20 ± 4% perf-profile.calltrace.cycles-pp.copy_page_to_iter.process_vm_rw_single_vec.process_vm_rw_core.process_vm_rw.__x64_sys_process_vm_readv
0.00 +2.5 2.50 ± 4% perf-profile.calltrace.cycles-pp.process_vm_rw_single_vec.process_vm_rw_core.process_vm_rw.__x64_sys_process_vm_readv.do_syscall_64
0.00 +2.7 2.70 ± 4% perf-profile.calltrace.cycles-pp.process_vm_rw_core.process_vm_rw.__x64_sys_process_vm_readv.do_syscall_64.entry_SYSCALL_64_after_hwframe
0.00 +2.7 2.73 ± 4% perf-profile.calltrace.cycles-pp.__x64_sys_process_vm_readv.do_syscall_64.entry_SYSCALL_64_after_hwframe.process_vm_readv
0.00 +2.7 2.73 ± 4% perf-profile.calltrace.cycles-pp.process_vm_rw.__x64_sys_process_vm_readv.do_syscall_64.entry_SYSCALL_64_after_hwframe.process_vm_readv
0.00 +2.7 2.74 ± 4% perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.process_vm_readv
0.00 +2.7 2.75 ± 4% perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.process_vm_readv
0.00 +2.8 2.82 ± 4% perf-profile.calltrace.cycles-pp.process_vm_readv
0.00 +3.1 3.09 ± 2% perf-profile.calltrace.cycles-pp.txinvr_.adi_
0.00 +3.2 3.16 perf-profile.calltrace.cycles-pp.copy_faces_
0.00 +3.4 3.40 ± 3% perf-profile.calltrace.cycles-pp.add_.MAIN__
0.00 +3.9 3.91 ± 3% perf-profile.calltrace.cycles-pp.MAIN__
0.00 +5.5 5.52 perf-profile.calltrace.cycles-pp.lhsx_.x_solve_.adi_
0.00 +6.2 6.25 ± 3% perf-profile.calltrace.cycles-pp.lhsy_.y_solve_.adi_
0.00 +7.4 7.38 perf-profile.calltrace.cycles-pp.lhsz_.z_solve_.adi_
0.00 +9.1 9.14 perf-profile.calltrace.cycles-pp.compute_rhs_.adi_
0.00 +12.1 12.09 ± 2% perf-profile.calltrace.cycles-pp.y_solve_.adi_
0.00 +12.4 12.42 perf-profile.calltrace.cycles-pp.x_solve_.adi_
0.00 +14.5 14.55 perf-profile.calltrace.cycles-pp.z_solve_.adi_
0.00 +51.6 51.58 perf-profile.calltrace.cycles-pp.adi_
51.56 ± 15% -48.2 3.36 ± 4% perf-profile.children.cycles-pp.do_syscall_64
51.56 ± 15% -48.2 3.36 ± 4% perf-profile.children.cycles-pp.entry_SYSCALL_64_after_hwframe
19.20 ± 16% -19.2 0.00 perf-profile.children.cycles-pp.write
17.84 ± 19% -17.8 0.00 perf-profile.children.cycles-pp.generic_perform_write
17.84 ± 19% -17.8 0.00 perf-profile.children.cycles-pp.ksys_write
17.84 ± 19% -17.8 0.00 perf-profile.children.cycles-pp.shmem_file_write_iter
17.84 ± 19% -17.8 0.00 perf-profile.children.cycles-pp.vfs_write
16.00 ± 34% -15.9 0.12 ± 15% perf-profile.children.cycles-pp.do_exit
16.00 ± 34% -15.9 0.12 ± 15% perf-profile.children.cycles-pp.do_group_exit
13.39 ± 40% -13.4 0.00 perf-profile.children.cycles-pp.syscall_exit_to_user_mode
12.83 ± 44% -12.8 0.00 perf-profile.children.cycles-pp.arch_do_signal_or_restart
12.83 ± 44% -12.8 0.00 perf-profile.children.cycles-pp.get_signal
12.81 ± 36% -12.8 0.00 perf-profile.children.cycles-pp.__fput
12.54 ± 46% -12.5 0.02 ± 99% perf-profile.children.cycles-pp.task_work_run
12.82 ± 42% -12.4 0.38 ± 35% perf-profile.children.cycles-pp.asm_exc_page_fault
10.45 ± 34% -10.4 0.00 perf-profile.children.cycles-pp.event_function_call
10.16 ± 37% -10.2 0.00 perf-profile.children.cycles-pp.smp_call_function_single
9.79 ± 55% -9.8 0.00 perf-profile.children.cycles-pp.perf_event_release_kernel
9.79 ± 55% -9.8 0.00 perf-profile.children.cycles-pp.perf_release
6.41 ± 20% -6.4 0.00 perf-profile.children.cycles-pp.fault_in_iov_iter_readable
6.41 ± 20% -6.4 0.00 perf-profile.children.cycles-pp.fault_in_readable
5.54 ± 55% -5.5 0.00 perf-profile.children.cycles-pp.copy_page_from_iter_atomic
4.88 ± 38% -4.8 0.13 ± 14% perf-profile.children.cycles-pp.__mmput
4.61 ± 31% -4.5 0.13 ± 14% perf-profile.children.cycles-pp.exit_mmap
4.38 ± 37% -4.3 0.13 ± 5% perf-profile.children.cycles-pp.__intel_pmu_enable_all
5.54 ± 55% -3.3 2.19 ± 4% perf-profile.children.cycles-pp.rep_movs_alternative
3.18 ± 26% -3.1 0.11 ± 15% perf-profile.children.cycles-pp.exit_mm
3.18 ± 26% -3.1 0.12 ± 15% perf-profile.children.cycles-pp.__x64_sys_exit_group
0.00 +0.1 0.05 perf-profile.children.cycles-pp.irqentry_exit_to_user_mode
0.00 +0.1 0.05 perf-profile.children.cycles-pp.irqtime_account_irq
0.00 +0.1 0.05 ± 7% perf-profile.children.cycles-pp.enqueue_hrtimer
0.00 +0.1 0.05 ± 7% perf-profile.children.cycles-pp.malloc
0.00 +0.1 0.05 ± 8% perf-profile.children.cycles-pp.__update_load_avg_se
0.00 +0.1 0.05 ± 8% perf-profile.children.cycles-pp.ktime_get_update_offsets_now
0.00 +0.1 0.06 ± 9% perf-profile.children.cycles-pp.run_rebalance_domains
0.00 +0.1 0.06 ± 9% perf-profile.children.cycles-pp.timerqueue_del
0.00 +0.1 0.06 ± 11% perf-profile.children.cycles-pp.__update_load_avg_cfs_rq
0.00 +0.1 0.06 ± 6% perf-profile.children.cycles-pp.clockevents_program_event
0.00 +0.1 0.06 ± 11% perf-profile.children.cycles-pp.find_get_task_by_vpid
0.00 +0.1 0.06 ± 11% perf-profile.children.cycles-pp.runtime.selectgo
0.00 +0.1 0.06 ± 13% perf-profile.children.cycles-pp.runtime.mcall
0.00 +0.1 0.06 ± 13% perf-profile.children.cycles-pp.runtime.park_m
0.00 +0.1 0.06 ± 13% perf-profile.children.cycles-pp.runtime.schedule
0.00 +0.1 0.06 ± 11% perf-profile.children.cycles-pp.try_grab_page
0.00 +0.1 0.06 ± 14% perf-profile.children.cycles-pp.github.com/containerd/containerd/gc/scheduler.(*gcScheduler).run
0.00 +0.1 0.06 ± 14% perf-profile.children.cycles-pp.github.com/containerd/containerd/gc/scheduler.init.0.func1.1
0.00 +0.1 0.07 ± 7% perf-profile.children.cycles-pp.rcu_pending
0.00 +0.1 0.07 ± 7% perf-profile.children.cycles-pp.tick_irq_enter
0.00 +0.1 0.07 ± 7% perf-profile.children.cycles-pp.yama_ptrace_access_check
0.00 +0.1 0.07 ± 5% perf-profile.children.cycles-pp.do_poll
0.00 +0.1 0.07 ± 13% perf-profile.children.cycles-pp.update_cfs_group
0.00 +0.1 0.07 ± 9% perf-profile.children.cycles-pp.account_user_time
0.00 +0.1 0.07 ± 9% perf-profile.children.cycles-pp.pinvr_
0.00 +0.1 0.07 ± 6% perf-profile.children.cycles-pp.do_sys_poll
0.00 +0.1 0.08 ± 6% perf-profile.children.cycles-pp.__x64_sys_poll
0.00 +0.1 0.08 ± 6% perf-profile.children.cycles-pp.irq_enter_rcu
0.00 +0.1 0.08 ± 11% perf-profile.children.cycles-pp.ninvr_
0.00 +0.1 0.09 ± 8% perf-profile.children.cycles-pp.__poll
0.00 +0.1 0.09 ± 8% perf-profile.children.cycles-pp.rcu_sched_clock_irq
0.00 +0.1 0.09 ± 10% perf-profile.children.cycles-pp.update_rq_clock_task
0.00 +0.1 0.09 ± 7% perf-profile.children.cycles-pp.rebalance_domains
0.00 +0.1 0.10 ± 18% perf-profile.children.cycles-pp.folio_lruvec_lock_irqsave
0.00 +0.1 0.10 ± 8% perf-profile.children.cycles-pp.security_ptrace_access_check
0.00 +0.1 0.10 ± 19% perf-profile.children.cycles-pp.runtime.goexit.abi0
0.00 +0.1 0.10 ± 10% perf-profile.children.cycles-pp.tick_nohz_next_event
0.00 +0.1 0.11 ± 7% perf-profile.children.cycles-pp.ptrace_may_access
0.00 +0.1 0.12 ± 7% perf-profile.children.cycles-pp.mca_btl_vader_poll_handle_frag
0.00 +0.1 0.12 ± 11% perf-profile.children.cycles-pp.follow_page_pte
0.00 +0.1 0.12 ± 15% perf-profile.children.cycles-pp.native_queued_spin_lock_slowpath
0.00 +0.1 0.13 ± 5% perf-profile.children.cycles-pp.mm_access
0.00 +0.1 0.13 ± 16% perf-profile.children.cycles-pp._raw_spin_lock_irqsave
0.00 +0.1 0.13 ± 8% perf-profile.children.cycles-pp.tick_nohz_get_sleep_length
0.00 +0.1 0.15 ± 6% perf-profile.children.cycles-pp.memcpy@plt
0.00 +0.2 0.16 ± 11% perf-profile.children.cycles-pp.opal_progress@plt
0.00 +0.2 0.16 ± 7% perf-profile.children.cycles-pp.ompi_request_default_wait_all
0.00 +0.2 0.19 ± 7% perf-profile.children.cycles-pp.__do_softirq
0.00 +0.2 0.20 ± 12% perf-profile.children.cycles-pp.__irqentry_text_end
0.00 +0.2 0.20 ± 2% perf-profile.children.cycles-pp.update_load_avg
0.00 +0.2 0.21 ± 7% perf-profile.children.cycles-pp.verify_
0.00 +0.2 0.22 ± 5% perf-profile.children.cycles-pp.menu_select
0.00 +0.2 0.22 ± 3% perf-profile.children.cycles-pp.irq_exit_rcu
0.00 +0.2 0.22 ± 7% perf-profile.children.cycles-pp.__get_user_pages
0.00 +0.2 0.23 ± 7% perf-profile.children.cycles-pp.__gup_longterm_locked
0.00 +0.2 0.24 ± 5% perf-profile.children.cycles-pp.pin_user_pages_remote
0.00 +0.5 0.46 ± 80% perf-profile.children.cycles-pp.arch_call_rest_init
0.00 +0.5 0.46 ± 80% perf-profile.children.cycles-pp.rest_init
0.00 +0.5 0.46 ± 80% perf-profile.children.cycles-pp.start_kernel
0.00 +0.5 0.46 ± 80% perf-profile.children.cycles-pp.x86_64_start_kernel
0.00 +0.5 0.46 ± 80% perf-profile.children.cycles-pp.x86_64_start_reservations
0.00 +0.5 0.46 ± 11% perf-profile.children.cycles-pp.ompi_coll_libnbc_progress
0.00 +1.1 1.06 ± 13% perf-profile.children.cycles-pp.opal_progress
0.00 +1.8 1.83 ± 2% perf-profile.children.cycles-pp.tzetar_
0.00 +2.2 2.20 ± 4% perf-profile.children.cycles-pp._copy_to_iter
0.00 +2.2 2.20 ± 4% perf-profile.children.cycles-pp.copy_page_to_iter
0.00 +2.5 2.50 ± 4% perf-profile.children.cycles-pp.process_vm_rw_single_vec
0.00 +2.7 2.70 ± 4% perf-profile.children.cycles-pp.process_vm_rw_core
0.00 +2.7 2.73 ± 4% perf-profile.children.cycles-pp.__x64_sys_process_vm_readv
0.00 +2.7 2.73 ± 4% perf-profile.children.cycles-pp.process_vm_rw
0.00 +2.8 2.82 ± 4% perf-profile.children.cycles-pp.process_vm_readv
0.00 +3.1 3.10 ± 2% perf-profile.children.cycles-pp.txinvr_
0.00 +3.2 3.20 perf-profile.children.cycles-pp.copy_faces_
0.00 +3.4 3.41 ± 3% perf-profile.children.cycles-pp.add_
0.00 +3.9 3.91 ± 3% perf-profile.children.cycles-pp.MAIN__
0.00 +5.5 5.54 perf-profile.children.cycles-pp.lhsx_
0.00 +6.3 6.27 ± 3% perf-profile.children.cycles-pp.lhsy_
0.00 +7.4 7.40 perf-profile.children.cycles-pp.lhsz_
0.00 +9.4 9.38 perf-profile.children.cycles-pp.compute_rhs_
0.00 +12.1 12.10 ± 2% perf-profile.children.cycles-pp.y_solve_
0.00 +12.4 12.44 perf-profile.children.cycles-pp.x_solve_
0.00 +14.6 14.56 perf-profile.children.cycles-pp.z_solve_
0.00 +51.6 51.58 perf-profile.children.cycles-pp.adi_
6.35 ± 42% -6.4 0.00 perf-profile.self.cycles-pp.smp_call_function_single
4.38 ± 37% -4.3 0.13 ± 5% perf-profile.self.cycles-pp.__intel_pmu_enable_all
5.14 ± 44% -3.0 2.14 ± 4% perf-profile.self.cycles-pp.rep_movs_alternative
0.00 +0.1 0.06 ± 11% perf-profile.self.cycles-pp.__update_load_avg_cfs_rq
0.00 +0.1 0.06 ± 6% perf-profile.self.cycles-pp.try_grab_page
0.00 +0.1 0.06 perf-profile.self.cycles-pp.menu_select
0.00 +0.1 0.06 ± 19% perf-profile.self.cycles-pp.scheduler_tick
0.00 +0.1 0.06 ± 11% perf-profile.self.cycles-pp.update_cfs_group
0.00 +0.1 0.07 ± 14% perf-profile.self.cycles-pp.task_tick_fair
0.00 +0.1 0.07 ± 9% perf-profile.self.cycles-pp.pinvr_
0.00 +0.1 0.07 ± 10% perf-profile.self.cycles-pp.ninvr_
0.00 +0.1 0.08 ± 9% perf-profile.self.cycles-pp.memcpy@plt
0.00 +0.1 0.08 ± 18% perf-profile.self.cycles-pp.opal_progress@plt
0.00 +0.1 0.10 ± 6% perf-profile.self.cycles-pp.update_load_avg
0.00 +0.1 0.10 ± 6% perf-profile.self.cycles-pp.cpuidle_enter_state
0.00 +0.1 0.12 ± 8% perf-profile.self.cycles-pp.mca_btl_vader_poll_handle_frag
0.00 +0.1 0.12 ± 15% perf-profile.self.cycles-pp.native_queued_spin_lock_slowpath
0.00 +0.2 0.16 ± 7% perf-profile.self.cycles-pp.ompi_request_default_wait_all
0.00 +0.2 0.20 ± 12% perf-profile.self.cycles-pp.__irqentry_text_end
0.00 +0.4 0.44 ± 11% perf-profile.self.cycles-pp.ompi_coll_libnbc_progress
0.00 +0.9 0.92 ± 12% perf-profile.self.cycles-pp.opal_progress
0.00 +1.8 1.79 ± 2% perf-profile.self.cycles-pp.tzetar_
0.00 +3.0 3.04 ± 2% perf-profile.self.cycles-pp.txinvr_
0.00 +3.1 3.15 perf-profile.self.cycles-pp.copy_faces_
0.00 +3.4 3.36 ± 3% perf-profile.self.cycles-pp.add_
0.00 +5.2 5.21 perf-profile.self.cycles-pp.z_solve_
0.00 +5.4 5.39 perf-profile.self.cycles-pp.lhsx_
0.00 +5.6 5.62 perf-profile.self.cycles-pp.y_solve_
0.00 +6.1 6.09 ± 3% perf-profile.self.cycles-pp.lhsy_
0.00 +6.7 6.66 perf-profile.self.cycles-pp.x_solve_
0.00 +7.2 7.22 perf-profile.self.cycles-pp.lhsz_
0.00 +9.1 9.11 perf-profile.self.cycles-pp.compute_rhs_
***************************************************************************************************
lkp-skl-fpga01: 104 threads 2 sockets (Skylake) with 192G memory
=========================================================================================
compiler/cpufreq_governor/kconfig/mode/nr_task/rootfs/tbox_group/test/testcase:
gcc-12/performance/x86_64-rhel-8.3/process/100%/debian-12-x86_64-20240206.cgz/lkp-skl-fpga01/mmap1/will-it-scale
commit:
9ef3ad0f03 ("ACPI/NUMA: Print CXL Early Discovery Table (CEDT)")
09f84589a8 ("ACPI/NUMA: Remove architecture dependent remainings")
9ef3ad0f032818c5 09f84589a8ed1162c0274a1b5b3
---------------- ---------------------------
%stddev %change %stddev
\ | \
1.967e+08 ± 12% +104.1% 4.015e+08 ± 36% cpuidle..time
3475 +13.2% 3933 ± 9% uptime.idle
117.40 +15.4% 135.50 ± 5% vmstat.procs.r
7130 -31.4% 4890 vmstat.system.cs
7005637 -38.1% 4334264 will-it-scale.104.processes
67361 -38.1% 41675 will-it-scale.per_process_ops
7005637 -38.1% 4334264 will-it-scale.workload
0.45 -0.1 0.34 mpstat.cpu.all.irq%
21.80 +24.1 45.86 mpstat.cpu.all.soft%
67.95 -21.2 46.74 mpstat.cpu.all.sys%
9.30 -3.4 5.94 mpstat.cpu.all.usr%
10720 ± 6% -50.5% 5308 ± 7% perf-c2c.DRAM.local
669.83 ± 4% +478.1% 3872 ± 6% perf-c2c.DRAM.remote
6822 ± 5% -40.8% 4037 ± 7% perf-c2c.HITM.local
421.00 ± 5% +577.2% 2850 ± 6% perf-c2c.HITM.remote
1.49e+08 -100.0% 0.00 numa-numastat.node0.local_node
1.491e+08 -100.0% 0.00 numa-numastat.node0.numa_hit
59368 ± 59% -100.0% 0.00 numa-numastat.node0.other_node
1.567e+08 -100.0% 0.00 numa-numastat.node1.local_node
1.567e+08 -100.0% 0.00 numa-numastat.node1.numa_hit
48388 ± 72% -100.0% 0.00 numa-numastat.node1.other_node
10352298 ± 7% -55.6% 4591957 ± 15% meminfo.DirectMap2M
2206149 ± 3% -21.4% 1733730 ± 5% meminfo.Inactive
2203855 ± 3% -21.4% 1731430 ± 5% meminfo.Inactive(anon)
6859527 -10.6% 6129805 meminfo.Memused
49043 -12.2% 43044 meminfo.Percpu
495530 -22.7% 382930 ± 2% meminfo.SUnreclaim
1816899 ± 4% -22.8% 1403126 ± 5% meminfo.Shmem
599136 -18.9% 485960 meminfo.Slab
257232 -84.9% 38717 meminfo.VmallocUsed
1219270 -8.5% 1115918 proc-vmstat.nr_file_pages
550889 ± 3% -21.4% 432828 ± 5% proc-vmstat.nr_inactive_anon
21408 +1.5% 21731 proc-vmstat.nr_kernel_stack
3324 -4.1% 3187 proc-vmstat.nr_page_table_pages
454144 ± 4% -22.8% 350800 ± 5% proc-vmstat.nr_shmem
125759 -21.9% 98264 proc-vmstat.nr_slab_unreclaimable
550889 ± 3% -21.4% 432828 ± 5% proc-vmstat.nr_zone_inactive_anon
55488 ± 18% -100.0% 0.00 proc-vmstat.numa_hint_faults
40437 ± 29% -100.0% 0.00 proc-vmstat.numa_hint_faults_local
3.058e+08 -28.2% 2.195e+08 proc-vmstat.numa_hit
3.057e+08 -28.2% 2.195e+08 proc-vmstat.numa_local
107756 -100.0% 0.00 proc-vmstat.numa_other
18231 ± 96% -100.0% 0.00 proc-vmstat.numa_pages_migrated
182874 ± 22% -100.0% 0.00 proc-vmstat.numa_pte_updates
90886 +122.4% 202099 ± 73% proc-vmstat.pgactivate
5.952e+08 -28.2% 4.271e+08 proc-vmstat.pgalloc_normal
1042723 -39.2% 634241 proc-vmstat.pgfault
5.943e+08 -28.2% 4.265e+08 proc-vmstat.pgfree
18231 ± 96% -100.0% 0.00 proc-vmstat.pgmigrate_success
44763 ± 6% -27.1% 32626 proc-vmstat.pgreuse
2.02 +12.4% 2.27 perf-stat.i.MPKI
2.933e+10 -36.2% 1.871e+10 perf-stat.i.branch-instructions
0.82 -0.0 0.80 perf-stat.i.branch-miss-rate%
2.382e+08 -37.7% 1.485e+08 perf-stat.i.branch-misses
62.01 +12.4 74.38 perf-stat.i.cache-miss-rate%
3.092e+08 -28.9% 2.199e+08 perf-stat.i.cache-misses
4.978e+08 -40.9% 2.943e+08 perf-stat.i.cache-references
7111 -32.2% 4822 perf-stat.i.context-switches
1.83 +59.0% 2.91 perf-stat.i.cpi
217.02 +39.3% 302.23 perf-stat.i.cpu-migrations
905.82 +41.5% 1281 perf-stat.i.cycles-between-cache-misses
1.528e+11 -36.7% 9.67e+10 perf-stat.i.instructions
0.55 -36.9% 0.35 perf-stat.i.ipc
0.11 ± 29% -51.5% 0.05 ± 37% perf-stat.i.major-faults
3266 -40.8% 1932 perf-stat.i.minor-faults
3266 -40.8% 1932 perf-stat.i.page-faults
2.02 +12.4% 2.27 perf-stat.overall.MPKI
0.81 -0.0 0.79 perf-stat.overall.branch-miss-rate%
62.11 +12.6 74.69 perf-stat.overall.cache-miss-rate%
1.83 +58.9% 2.91 perf-stat.overall.cpi
905.85 +41.4% 1280 perf-stat.overall.cycles-between-cache-misses
0.55 -37.1% 0.34 perf-stat.overall.ipc
6581748 +3.8% 6830787 perf-stat.overall.path-length
2.923e+10 -36.2% 1.865e+10 perf-stat.ps.branch-instructions
2.373e+08 -37.7% 1.479e+08 perf-stat.ps.branch-misses
3.082e+08 -28.9% 2.192e+08 perf-stat.ps.cache-misses
4.962e+08 -40.9% 2.934e+08 perf-stat.ps.cache-references
7083 -32.2% 4802 perf-stat.ps.context-switches
215.95 +39.3% 300.81 perf-stat.ps.cpu-migrations
1.523e+11 -36.7% 9.636e+10 perf-stat.ps.instructions
0.11 ± 29% -51.1% 0.05 ± 37% perf-stat.ps.major-faults
3244 -41.2% 1908 perf-stat.ps.minor-faults
3245 -41.2% 1908 perf-stat.ps.page-faults
4.611e+13 -35.8% 2.961e+13 perf-stat.total.instructions
13971 ± 89% -100.0% 0.00 numa-vmstat.node0.nr_active_anon
97064 ± 52% -100.0% 0.00 numa-vmstat.node0.nr_anon_pages
551160 ± 47% -100.0% 0.00 numa-vmstat.node0.nr_file_pages
23600699 -100.0% 0.00 numa-vmstat.node0.nr_free_pages
97975 ± 52% -100.0% 0.00 numa-vmstat.node0.nr_inactive_anon
11699 ± 4% -100.0% 0.00 numa-vmstat.node0.nr_kernel_stack
14296 ± 11% -100.0% 0.00 numa-vmstat.node0.nr_mapped
1544 ± 6% -100.0% 0.00 numa-vmstat.node0.nr_page_table_pages
15651 ± 83% -100.0% 0.00 numa-vmstat.node0.nr_shmem
15308 ± 24% -100.0% 0.00 numa-vmstat.node0.nr_slab_reclaimable
67651 ± 3% -100.0% 0.00 numa-vmstat.node0.nr_slab_unreclaimable
536155 ± 47% -100.0% 0.00 numa-vmstat.node0.nr_unevictable
13971 ± 89% -100.0% 0.00 numa-vmstat.node0.nr_zone_active_anon
97974 ± 52% -100.0% 0.00 numa-vmstat.node0.nr_zone_inactive_anon
536155 ± 47% -100.0% 0.00 numa-vmstat.node0.nr_zone_unevictable
1.491e+08 -100.0% 0.00 numa-vmstat.node0.numa_hit
1.49e+08 -100.0% 0.00 numa-vmstat.node0.numa_local
59368 ± 59% -100.0% 0.00 numa-vmstat.node0.numa_other
67762 ± 21% -100.0% 0.00 numa-vmstat.node1.nr_active_anon
82222 ± 62% -100.0% 0.00 numa-vmstat.node1.nr_anon_pages
668044 ± 37% -100.0% 0.00 numa-vmstat.node1.nr_file_pages
23857326 -100.0% 0.00 numa-vmstat.node1.nr_free_pages
452882 ± 9% -100.0% 0.00 numa-vmstat.node1.nr_inactive_anon
9709 ± 5% -100.0% 0.00 numa-vmstat.node1.nr_kernel_stack
148558 ± 4% -100.0% 0.00 numa-vmstat.node1.nr_mapped
1779 ± 5% -100.0% 0.00 numa-vmstat.node1.nr_page_table_pages
438426 ± 5% -100.0% 0.00 numa-vmstat.node1.nr_shmem
10592 ± 35% -100.0% 0.00 numa-vmstat.node1.nr_slab_reclaimable
61563 ± 4% -100.0% 0.00 numa-vmstat.node1.nr_slab_unreclaimable
229017 ±111% -100.0% 0.00 numa-vmstat.node1.nr_unevictable
67762 ± 21% -100.0% 0.00 numa-vmstat.node1.nr_zone_active_anon
452882 ± 9% -100.0% 0.00 numa-vmstat.node1.nr_zone_inactive_anon
229017 ±111% -100.0% 0.00 numa-vmstat.node1.nr_zone_unevictable
1.567e+08 -100.0% 0.00 numa-vmstat.node1.numa_hit
1.566e+08 -100.0% 0.00 numa-vmstat.node1.numa_local
48388 ± 72% -100.0% 0.00 numa-vmstat.node1.numa_other
55900 ± 89% -100.0% 0.00 numa-meminfo.node0.Active
55801 ± 89% -100.0% 0.00 numa-meminfo.node0.Active(anon)
388254 ± 52% -100.0% 0.00 numa-meminfo.node0.AnonPages
440142 ± 48% -100.0% 0.00 numa-meminfo.node0.AnonPages.max
2204609 ± 47% -100.0% 0.00 numa-meminfo.node0.FilePages
392333 ± 52% -100.0% 0.00 numa-meminfo.node0.Inactive
391951 ± 52% -100.0% 0.00 numa-meminfo.node0.Inactive(anon)
61236 ± 24% -100.0% 0.00 numa-meminfo.node0.KReclaimable
11699 ± 4% -100.0% 0.00 numa-meminfo.node0.KernelStack
56750 ± 12% -100.0% 0.00 numa-meminfo.node0.Mapped
94405300 -100.0% 0.00 numa-meminfo.node0.MemFree
97679448 -100.0% 0.00 numa-meminfo.node0.MemTotal
3274147 ± 37% -100.0% 0.00 numa-meminfo.node0.MemUsed
6174 ± 6% -100.0% 0.00 numa-meminfo.node0.PageTables
61236 ± 24% -100.0% 0.00 numa-meminfo.node0.SReclaimable
268262 ± 4% -100.0% 0.00 numa-meminfo.node0.SUnreclaim
62576 ± 83% -100.0% 0.00 numa-meminfo.node0.Shmem
329498 ± 7% -100.0% 0.00 numa-meminfo.node0.Slab
2144623 ± 47% -100.0% 0.00 numa-meminfo.node0.Unevictable
271268 ± 21% -100.0% 0.00 numa-meminfo.node1.Active
270777 ± 21% -100.0% 0.00 numa-meminfo.node1.Active(anon)
328899 ± 62% -100.0% 0.00 numa-meminfo.node1.AnonPages
368362 ± 54% -100.0% 0.00 numa-meminfo.node1.AnonPages.max
2672717 ± 37% -100.0% 0.00 numa-meminfo.node1.FilePages
1814269 ± 9% -100.0% 0.00 numa-meminfo.node1.Inactive
1812357 ± 9% -100.0% 0.00 numa-meminfo.node1.Inactive(anon)
42376 ± 35% -100.0% 0.00 numa-meminfo.node1.KReclaimable
9709 ± 5% -100.0% 0.00 numa-meminfo.node1.KernelStack
594500 ± 4% -100.0% 0.00 numa-meminfo.node1.Mapped
95440412 -100.0% 0.00 numa-meminfo.node1.MemFree
99034620 -100.0% 0.00 numa-meminfo.node1.MemTotal
3594207 ± 32% -100.0% 0.00 numa-meminfo.node1.MemUsed
7121 ± 5% -100.0% 0.00 numa-meminfo.node1.PageTables
42376 ± 35% -100.0% 0.00 numa-meminfo.node1.SReclaimable
234056 ± 3% -100.0% 0.00 numa-meminfo.node1.SUnreclaim
1754246 ± 5% -100.0% 0.00 numa-meminfo.node1.Shmem
276432 ± 7% -100.0% 0.00 numa-meminfo.node1.Slab
916068 ±111% -100.0% 0.00 numa-meminfo.node1.Unevictable
12846621 -31.1% 8851378 sched_debug.cfs_rq:/.avg_vruntime.avg
13026668 -29.8% 9146405 sched_debug.cfs_rq:/.avg_vruntime.max
12010831 -34.2% 7905615 sched_debug.cfs_rq:/.avg_vruntime.min
172319 ± 4% +46.4% 252250 ± 4% sched_debug.cfs_rq:/.avg_vruntime.stddev
0.98 +15.9% 1.14 ± 4% sched_debug.cfs_rq:/.h_nr_running.avg
0.38 ± 4% +25.9% 0.48 ± 5% sched_debug.cfs_rq:/.h_nr_running.stddev
1482184 ± 18% +81.3% 2687007 ± 18% sched_debug.cfs_rq:/.left_deadline.avg
12984350 -29.9% 9098475 sched_debug.cfs_rq:/.left_deadline.max
1482165 ± 18% +81.3% 2686955 ± 18% sched_debug.cfs_rq:/.left_vruntime.avg
12984176 -29.9% 9098312 sched_debug.cfs_rq:/.left_vruntime.max
110018 ± 15% +151.7% 276928 ± 20% sched_debug.cfs_rq:/.load.avg
279353 ± 7% +39.9% 390900 ± 5% sched_debug.cfs_rq:/.load.stddev
153.45 ± 7% +165.5% 407.41 ± 5% sched_debug.cfs_rq:/.load_avg.avg
465.14 ± 19% +59.0% 739.61 ± 3% sched_debug.cfs_rq:/.load_avg.max
67.81 ± 5% +162.1% 177.72 ± 15% sched_debug.cfs_rq:/.load_avg.min
75.29 ± 10% +48.6% 111.88 ± 7% sched_debug.cfs_rq:/.load_avg.stddev
12846638 -31.1% 8851412 sched_debug.cfs_rq:/.min_vruntime.avg
13026687 -29.8% 9146424 sched_debug.cfs_rq:/.min_vruntime.max
12010851 -34.2% 7905665 sched_debug.cfs_rq:/.min_vruntime.min
172320 ± 4% +46.4% 252248 ± 4% sched_debug.cfs_rq:/.min_vruntime.stddev
0.95 +15.5% 1.10 ± 4% sched_debug.cfs_rq:/.nr_running.avg
0.32 ± 5% +31.7% 0.42 ± 4% sched_debug.cfs_rq:/.nr_running.stddev
1482165 ± 18% +81.3% 2686956 ± 18% sched_debug.cfs_rq:/.right_vruntime.avg
12984176 -29.9% 9098312 sched_debug.cfs_rq:/.right_vruntime.max
1047 +24.4% 1303 sched_debug.cfs_rq:/.runnable_avg.avg
2000 +20.3% 2405 ± 2% sched_debug.cfs_rq:/.runnable_avg.max
159.06 ± 4% +39.3% 221.64 ± 6% sched_debug.cfs_rq:/.runnable_avg.stddev
1151 ± 3% +15.3% 1328 ± 3% sched_debug.cfs_rq:/.util_avg.max
78.12 ± 15% +34.9% 105.36 ± 8% sched_debug.cfs_rq:/.util_avg.stddev
734.30 ± 4% -10.1% 659.87 ± 3% sched_debug.cfs_rq:/.util_est.avg
159.60 ± 9% +79.3% 286.17 ± 5% sched_debug.cfs_rq:/.util_est.stddev
18.37 ± 4% +22.8% 22.56 ± 4% sched_debug.cpu.clock.stddev
3171 ± 2% -18.0% 2599 ± 7% sched_debug.cpu.curr->pid.avg
7419 ± 2% -24.5% 5599 ± 3% sched_debug.cpu.curr->pid.max
1202 ± 6% +24.3% 1494 ± 4% sched_debug.cpu.curr->pid.stddev
0.00 ± 2% +12.6% 0.00 sched_debug.cpu.next_balance.stddev
0.98 +15.9% 1.14 ± 4% sched_debug.cpu.nr_running.avg
0.38 ± 4% +25.4% 0.47 ± 5% sched_debug.cpu.nr_running.stddev
11316 -28.4% 8107 sched_debug.cpu.nr_switches.avg
8294 -35.9% 5317 sched_debug.cpu.nr_switches.min
2590 ± 6% +20.3% 3115 ± 7% sched_debug.cpu.nr_switches.stddev
48.09 -14.2 33.91 perf-profile.calltrace.cycles-pp.__munmap
36.16 -13.8 22.35 perf-profile.calltrace.cycles-pp.__mmap
42.68 -12.0 30.66 perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.__munmap
30.76 -11.6 19.12 perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.__mmap
41.53 -11.6 29.96 perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.__munmap
40.88 -11.3 29.57 perf-profile.calltrace.cycles-pp.__x64_sys_munmap.do_syscall_64.entry_SYSCALL_64_after_hwframe.__munmap
40.76 -11.3 29.50 perf-profile.calltrace.cycles-pp.__vm_munmap.__x64_sys_munmap.do_syscall_64.entry_SYSCALL_64_after_hwframe.__munmap
29.62 -11.2 18.42 perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.__mmap
40.13 -11.0 29.13 perf-profile.calltrace.cycles-pp.do_vmi_munmap.__vm_munmap.__x64_sys_munmap.do_syscall_64.entry_SYSCALL_64_after_hwframe
28.85 -10.9 17.98 perf-profile.calltrace.cycles-pp.vm_mmap_pgoff.do_syscall_64.entry_SYSCALL_64_after_hwframe.__mmap
39.44 -10.7 28.72 perf-profile.calltrace.cycles-pp.do_vmi_align_munmap.do_vmi_munmap.__vm_munmap.__x64_sys_munmap.do_syscall_64
27.54 -10.3 17.21 perf-profile.calltrace.cycles-pp.do_mmap.vm_mmap_pgoff.do_syscall_64.entry_SYSCALL_64_after_hwframe.__mmap
23.80 -8.8 14.98 perf-profile.calltrace.cycles-pp.mmap_region.do_mmap.vm_mmap_pgoff.do_syscall_64.entry_SYSCALL_64_after_hwframe
12.41 -5.1 7.34 perf-profile.calltrace.cycles-pp.unmap_region.do_vmi_align_munmap.do_vmi_munmap.__vm_munmap.__x64_sys_munmap
11.64 -4.5 7.10 perf-profile.calltrace.cycles-pp.mas_store_prealloc.mmap_region.do_mmap.vm_mmap_pgoff.do_syscall_64
23.99 -4.4 19.62 perf-profile.calltrace.cycles-pp.mas_store_gfp.do_vmi_align_munmap.do_vmi_munmap.__vm_munmap.__x64_sys_munmap
10.31 -4.1 6.25 perf-profile.calltrace.cycles-pp.mas_wr_bnode.mas_store_prealloc.mmap_region.do_mmap.vm_mmap_pgoff
9.12 -3.7 5.39 perf-profile.calltrace.cycles-pp.unmap_vmas.unmap_region.do_vmi_align_munmap.do_vmi_munmap.__vm_munmap
8.76 -3.6 5.17 perf-profile.calltrace.cycles-pp.unmap_page_range.unmap_vmas.unmap_region.do_vmi_align_munmap.do_vmi_munmap
8.65 -3.4 5.29 perf-profile.calltrace.cycles-pp.mas_split.mas_wr_bnode.mas_store_prealloc.mmap_region.do_mmap
18.03 -3.3 14.74 perf-profile.calltrace.cycles-pp.mas_wr_spanning_store.mas_store_gfp.do_vmi_align_munmap.do_vmi_munmap.__vm_munmap
7.98 -3.3 4.71 perf-profile.calltrace.cycles-pp.zap_pmd_range.unmap_page_range.unmap_vmas.unmap_region.do_vmi_align_munmap
6.44 -2.4 4.00 perf-profile.calltrace.cycles-pp.mas_spanning_rebalance.mas_wr_spanning_store.mas_store_gfp.do_vmi_align_munmap.do_vmi_munmap
3.32 -1.4 1.95 perf-profile.calltrace.cycles-pp.__cond_resched.zap_pmd_range.unmap_page_range.unmap_vmas.unmap_region
3.35 -1.4 2.00 perf-profile.calltrace.cycles-pp.get_unmapped_area.do_mmap.vm_mmap_pgoff.do_syscall_64.entry_SYSCALL_64_after_hwframe
2.98 -1.2 1.78 perf-profile.calltrace.cycles-pp.thp_get_unmapped_area.get_unmapped_area.do_mmap.vm_mmap_pgoff.do_syscall_64
2.75 -1.1 1.64 perf-profile.calltrace.cycles-pp.arch_get_unmapped_area_topdown.thp_get_unmapped_area.get_unmapped_area.do_mmap.vm_mmap_pgoff
2.60 -1.0 1.55 perf-profile.calltrace.cycles-pp.perf_event_mmap.mmap_region.do_mmap.vm_mmap_pgoff.do_syscall_64
3.00 -1.0 1.96 perf-profile.calltrace.cycles-pp.vm_area_alloc.mmap_region.do_mmap.vm_mmap_pgoff.do_syscall_64
2.35 -1.0 1.37 perf-profile.calltrace.cycles-pp.mas_topiary_replace.mas_spanning_rebalance.mas_wr_spanning_store.mas_store_gfp.do_vmi_align_munmap
2.31 -0.9 1.37 perf-profile.calltrace.cycles-pp.vm_unmapped_area.arch_get_unmapped_area_topdown.thp_get_unmapped_area.get_unmapped_area.do_mmap
2.29 -0.9 1.37 perf-profile.calltrace.cycles-pp.perf_event_mmap_event.perf_event_mmap.mmap_region.do_mmap.vm_mmap_pgoff
2.16 -0.9 1.24 perf-profile.calltrace.cycles-pp.mas_topiary_replace.mas_split.mas_wr_bnode.mas_store_prealloc.mmap_region
2.54 -0.9 1.66 perf-profile.calltrace.cycles-pp.kmem_cache_alloc.vm_area_alloc.mmap_region.do_mmap.vm_mmap_pgoff
2.01 -0.8 1.19 perf-profile.calltrace.cycles-pp.free_pgd_range.unmap_region.do_vmi_align_munmap.do_vmi_munmap.__vm_munmap
2.04 -0.8 1.25 perf-profile.calltrace.cycles-pp.syscall_return_via_sysret.__mmap
2.05 -0.8 1.25 perf-profile.calltrace.cycles-pp.syscall_return_via_sysret.__munmap
5.22 -0.8 4.43 ± 3% perf-profile.calltrace.cycles-pp.mas_nomem.mas_store_gfp.do_vmi_align_munmap.do_vmi_munmap.__vm_munmap
1.83 -0.7 1.09 perf-profile.calltrace.cycles-pp.free_p4d_range.free_pgd_range.unmap_region.do_vmi_align_munmap.do_vmi_munmap
2.99 -0.7 2.25 perf-profile.calltrace.cycles-pp.mas_preallocate.mmap_region.do_mmap.vm_mmap_pgoff.do_syscall_64
5.12 -0.7 4.38 ± 3% perf-profile.calltrace.cycles-pp.mas_destroy.mas_nomem.mas_store_gfp.do_vmi_align_munmap.do_vmi_munmap
0.99 ± 19% -0.7 0.26 ±100% perf-profile.calltrace.cycles-pp.__do_softirq.irq_exit_rcu.sysvec_apic_timer_interrupt.asm_sysvec_apic_timer_interrupt.__put_partials
1.60 -0.7 0.94 perf-profile.calltrace.cycles-pp.mast_fill_bnode.mas_split.mas_wr_bnode.mas_store_prealloc.mmap_region
1.58 -0.6 0.94 perf-profile.calltrace.cycles-pp.free_pud_range.free_p4d_range.free_pgd_range.unmap_region.do_vmi_align_munmap
1.65 -0.6 1.02 perf-profile.calltrace.cycles-pp.entry_SYSRETQ_unsafe_stack.__mmap
1.65 -0.6 1.03 perf-profile.calltrace.cycles-pp.entry_SYSRETQ_unsafe_stack.__munmap
1.53 -0.6 0.92 perf-profile.calltrace.cycles-pp.perf_iterate_sb.perf_event_mmap_event.perf_event_mmap.mmap_region.do_mmap
1.38 -0.6 0.82 perf-profile.calltrace.cycles-pp.mas_empty_area_rev.vm_unmapped_area.arch_get_unmapped_area_topdown.thp_get_unmapped_area.get_unmapped_area
2.42 -0.5 1.91 ± 2% perf-profile.calltrace.cycles-pp.mas_alloc_nodes.mas_preallocate.mmap_region.do_mmap.vm_mmap_pgoff
1.11 -0.5 0.64 perf-profile.calltrace.cycles-pp.mast_split_data.mas_split.mas_wr_bnode.mas_store_prealloc.mmap_region
0.88 -0.5 0.42 ± 44% perf-profile.calltrace.cycles-pp.mab_mas_cp.mast_split_data.mas_split.mas_wr_bnode.mas_store_prealloc
2.64 ± 2% -0.4 2.21 ± 3% perf-profile.calltrace.cycles-pp.kmem_cache_free_bulk.mas_destroy.mas_nomem.mas_store_gfp.do_vmi_align_munmap
1.04 -0.4 0.61 perf-profile.calltrace.cycles-pp.mas_rev_awalk.mas_empty_area_rev.vm_unmapped_area.arch_get_unmapped_area_topdown.thp_get_unmapped_area
1.07 -0.4 0.66 perf-profile.calltrace.cycles-pp.perf_event_mmap_output.perf_iterate_sb.perf_event_mmap_event.perf_event_mmap.mmap_region
0.94 -0.4 0.54 perf-profile.calltrace.cycles-pp.mas_store_b_node.mas_wr_bnode.mas_store_prealloc.mmap_region.do_mmap
0.86 -0.4 0.51 perf-profile.calltrace.cycles-pp.do_vmi_munmap.mmap_region.do_mmap.vm_mmap_pgoff.do_syscall_64
1.29 -0.3 0.97 perf-profile.calltrace.cycles-pp.mab_mas_cp.mas_spanning_rebalance.mas_wr_spanning_store.mas_store_gfp.do_vmi_align_munmap
0.97 -0.3 0.66 ± 2% perf-profile.calltrace.cycles-pp.mas_split_final_node.mas_split.mas_wr_bnode.mas_store_prealloc.mmap_region
0.55 -0.3 0.26 ±100% perf-profile.calltrace.cycles-pp.asm_sysvec_apic_timer_interrupt.__put_partials.kmem_cache_free.mas_destroy.mas_nomem
0.55 -0.3 0.26 ±100% perf-profile.calltrace.cycles-pp.sysvec_apic_timer_interrupt.asm_sysvec_apic_timer_interrupt.__put_partials.kmem_cache_free.mas_destroy
0.55 -0.3 0.26 ±100% perf-profile.calltrace.cycles-pp.irq_exit_rcu.sysvec_apic_timer_interrupt.asm_sysvec_apic_timer_interrupt.__put_partials.kmem_cache_free
0.85 -0.3 0.56 perf-profile.calltrace.cycles-pp.mas_push_data.mas_split.mas_wr_bnode.mas_store_prealloc.mmap_region
0.82 -0.3 0.56 ± 2% perf-profile.calltrace.cycles-pp.mas_destroy.mas_store_prealloc.mmap_region.do_mmap.vm_mmap_pgoff
0.69 -0.3 0.43 ± 44% perf-profile.calltrace.cycles-pp.mas_pop_node.mas_split.mas_wr_bnode.mas_store_prealloc.mmap_region
1.82 -0.3 1.56 ± 2% perf-profile.calltrace.cycles-pp.kmem_cache_alloc_bulk.mas_alloc_nodes.mas_preallocate.mmap_region.do_mmap
1.77 ± 2% -0.2 1.58 ± 4% perf-profile.calltrace.cycles-pp.kmem_cache_free.mas_destroy.mas_nomem.mas_store_gfp.do_vmi_align_munmap
1.52 ± 2% -0.1 1.39 ± 3% perf-profile.calltrace.cycles-pp.__kmem_cache_alloc_bulk.kmem_cache_alloc_bulk.mas_alloc_nodes.mas_preallocate.mmap_region
0.62 -0.0 0.59 ± 2% perf-profile.calltrace.cycles-pp.kmem_cache_free.vm_area_free_rcu_cb.rcu_do_batch.rcu_core.__do_softirq
0.92 +0.0 0.96 ± 2% perf-profile.calltrace.cycles-pp.allocate_slab.___slab_alloc.__kmem_cache_alloc_bulk.kmem_cache_alloc_bulk.mas_alloc_nodes
0.85 ± 4% +0.2 1.01 ± 2% perf-profile.calltrace.cycles-pp.rcu_cblist_dequeue.rcu_do_batch.rcu_core.__do_softirq.run_ksoftirqd
1.44 ± 2% +0.2 1.68 ± 2% perf-profile.calltrace.cycles-pp.__do_softirq.irq_exit_rcu.sysvec_apic_timer_interrupt.asm_sysvec_apic_timer_interrupt.get_partial_node
1.44 ± 2% +0.2 1.69 ± 3% perf-profile.calltrace.cycles-pp.irq_exit_rcu.sysvec_apic_timer_interrupt.asm_sysvec_apic_timer_interrupt.get_partial_node.___slab_alloc
1.44 ± 2% +0.3 1.70 ± 2% perf-profile.calltrace.cycles-pp.asm_sysvec_apic_timer_interrupt.get_partial_node.___slab_alloc.__kmem_cache_alloc_bulk.kmem_cache_alloc_bulk
1.44 ± 2% +0.3 1.70 ± 2% perf-profile.calltrace.cycles-pp.sysvec_apic_timer_interrupt.asm_sysvec_apic_timer_interrupt.get_partial_node.___slab_alloc.__kmem_cache_alloc_bulk
1.24 ± 2% +0.3 1.59 ± 2% perf-profile.calltrace.cycles-pp.kmem_cache_free.rcu_do_batch.rcu_core.__do_softirq.irq_exit_rcu
8.54 +0.4 8.89 ± 2% perf-profile.calltrace.cycles-pp.mas_alloc_nodes.mas_wr_spanning_store.mas_store_gfp.do_vmi_align_munmap.do_vmi_munmap
8.10 +0.5 8.64 ± 2% perf-profile.calltrace.cycles-pp.kmem_cache_alloc_bulk.mas_alloc_nodes.mas_wr_spanning_store.mas_store_gfp.do_vmi_align_munmap
7.92 +0.6 8.53 ± 2% perf-profile.calltrace.cycles-pp.__kmem_cache_alloc_bulk.kmem_cache_alloc_bulk.mas_alloc_nodes.mas_wr_spanning_store.mas_store_gfp
7.30 +0.9 8.16 ± 2% perf-profile.calltrace.cycles-pp.___slab_alloc.__kmem_cache_alloc_bulk.kmem_cache_alloc_bulk.mas_alloc_nodes.mas_wr_spanning_store
1.85 +1.1 2.93 ± 4% perf-profile.calltrace.cycles-pp.native_queued_spin_lock_slowpath._raw_spin_lock_irqsave.__put_partials.kmem_cache_free.rcu_do_batch
1.88 +1.1 2.96 ± 4% perf-profile.calltrace.cycles-pp._raw_spin_lock_irqsave.__put_partials.kmem_cache_free.rcu_do_batch.rcu_core
3.44 ± 2% +2.1 5.51 ± 4% perf-profile.calltrace.cycles-pp._raw_spin_lock_irqsave.get_partial_node.___slab_alloc.__kmem_cache_alloc_bulk.kmem_cache_alloc_bulk
3.40 ± 2% +2.1 5.49 ± 4% perf-profile.calltrace.cycles-pp.native_queued_spin_lock_slowpath._raw_spin_lock_irqsave.get_partial_node.___slab_alloc.__kmem_cache_alloc_bulk
5.40 ± 2% +2.3 7.66 ± 3% perf-profile.calltrace.cycles-pp.get_partial_node.___slab_alloc.__kmem_cache_alloc_bulk.kmem_cache_alloc_bulk.mas_alloc_nodes
6.49 +3.1 9.58 ± 5% perf-profile.calltrace.cycles-pp.native_queued_spin_lock_slowpath._raw_spin_lock_irqsave.__slab_free.kmem_cache_free.rcu_do_batch
9.25 +3.1 12.33 ± 3% perf-profile.calltrace.cycles-pp.__slab_free.kmem_cache_free.rcu_do_batch.rcu_core.__do_softirq
6.61 +3.1 9.72 ± 5% perf-profile.calltrace.cycles-pp._raw_spin_lock_irqsave.__slab_free.kmem_cache_free.rcu_do_batch.rcu_core
0.65 ± 8% +10.4 11.04 ± 3% perf-profile.calltrace.cycles-pp._raw_spin_lock_irqsave.free_pcppages_bulk.free_unref_page_commit.free_unref_page.kmem_cache_free
0.70 ± 8% +10.4 11.12 ± 3% perf-profile.calltrace.cycles-pp.free_pcppages_bulk.free_unref_page_commit.free_unref_page.kmem_cache_free.rcu_do_batch
0.73 ± 7% +10.5 11.18 ± 3% perf-profile.calltrace.cycles-pp.free_unref_page.kmem_cache_free.rcu_do_batch.rcu_core.__do_softirq
0.71 ± 8% +10.5 11.16 ± 3% perf-profile.calltrace.cycles-pp.free_unref_page_commit.free_unref_page.kmem_cache_free.rcu_do_batch.rcu_core
0.84 ± 7% +13.3 14.15 ± 3% perf-profile.calltrace.cycles-pp._raw_spin_lock_irqsave.free_pcppages_bulk.free_unref_page_commit.free_unref_page.__put_partials
0.91 ± 6% +13.4 14.26 ± 3% perf-profile.calltrace.cycles-pp.free_pcppages_bulk.free_unref_page_commit.free_unref_page.__put_partials.kmem_cache_free
0.92 ± 6% +13.4 14.32 ± 3% perf-profile.calltrace.cycles-pp.free_unref_page_commit.free_unref_page.__put_partials.kmem_cache_free.rcu_do_batch
0.94 ± 6% +13.4 14.34 ± 3% perf-profile.calltrace.cycles-pp.free_unref_page.__put_partials.kmem_cache_free.rcu_do_batch.rcu_core
2.95 +15.1 18.08 perf-profile.calltrace.cycles-pp.__put_partials.kmem_cache_free.rcu_do_batch.rcu_core.__do_softirq
1.45 ± 7% +23.6 25.09 ± 3% perf-profile.calltrace.cycles-pp.native_queued_spin_lock_slowpath._raw_spin_lock_irqsave.free_pcppages_bulk.free_unref_page_commit.free_unref_page
13.04 +28.0 40.99 perf-profile.calltrace.cycles-pp.kmem_cache_free.rcu_do_batch.rcu_core.__do_softirq.run_ksoftirqd
15.02 +28.1 43.09 perf-profile.calltrace.cycles-pp.kthread.ret_from_fork.ret_from_fork_asm
15.02 +28.1 43.09 perf-profile.calltrace.cycles-pp.ret_from_fork.ret_from_fork_asm
15.02 +28.1 43.09 perf-profile.calltrace.cycles-pp.ret_from_fork_asm
14.98 +28.1 43.06 perf-profile.calltrace.cycles-pp.rcu_core.__do_softirq.run_ksoftirqd.smpboot_thread_fn.kthread
14.96 +28.1 43.04 perf-profile.calltrace.cycles-pp.rcu_do_batch.rcu_core.__do_softirq.run_ksoftirqd.smpboot_thread_fn
14.98 +28.1 43.06 perf-profile.calltrace.cycles-pp.__do_softirq.run_ksoftirqd.smpboot_thread_fn.kthread.ret_from_fork
14.98 +28.1 43.06 perf-profile.calltrace.cycles-pp.run_ksoftirqd.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
15.00 +28.1 43.08 perf-profile.calltrace.cycles-pp.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
73.58 -23.7 49.89 perf-profile.children.cycles-pp.entry_SYSCALL_64_after_hwframe
71.25 -22.8 48.46 perf-profile.children.cycles-pp.do_syscall_64
48.23 -14.2 34.00 perf-profile.children.cycles-pp.__munmap
36.32 -13.9 22.44 perf-profile.children.cycles-pp.__mmap
41.03 -11.4 29.66 perf-profile.children.cycles-pp.do_vmi_munmap
40.90 -11.3 29.58 perf-profile.children.cycles-pp.__x64_sys_munmap
40.79 -11.3 29.52 perf-profile.children.cycles-pp.__vm_munmap
28.87 -10.9 18.00 perf-profile.children.cycles-pp.vm_mmap_pgoff
39.53 -10.8 28.78 perf-profile.children.cycles-pp.do_vmi_align_munmap
27.55 -10.3 17.22 perf-profile.children.cycles-pp.do_mmap
23.89 -8.9 15.04 perf-profile.children.cycles-pp.mmap_region
12.45 -5.1 7.37 perf-profile.children.cycles-pp.unmap_region
11.65 -4.5 7.11 perf-profile.children.cycles-pp.mas_store_prealloc
24.04 -4.4 19.64 perf-profile.children.cycles-pp.mas_store_gfp
10.32 -4.1 6.26 perf-profile.children.cycles-pp.mas_wr_bnode
10.22 -4.0 6.24 ± 2% perf-profile.children.cycles-pp.sysvec_apic_timer_interrupt
10.25 -4.0 6.28 perf-profile.children.cycles-pp.asm_sysvec_apic_timer_interrupt
9.84 -3.9 5.93 ± 2% perf-profile.children.cycles-pp.irq_exit_rcu
9.14 -3.7 5.39 perf-profile.children.cycles-pp.unmap_vmas
8.78 -3.6 5.18 perf-profile.children.cycles-pp.unmap_page_range
8.73 -3.4 5.34 perf-profile.children.cycles-pp.mas_split
8.20 -3.4 4.84 perf-profile.children.cycles-pp.zap_pmd_range
18.05 -3.3 14.75 perf-profile.children.cycles-pp.mas_wr_spanning_store
6.49 -2.4 4.04 perf-profile.children.cycles-pp.mas_spanning_rebalance
4.54 -1.9 2.62 perf-profile.children.cycles-pp.mas_topiary_replace
4.66 -1.9 2.75 perf-profile.children.cycles-pp.__cond_resched
4.13 -1.6 2.52 perf-profile.children.cycles-pp.syscall_return_via_sysret
3.37 -1.4 2.01 perf-profile.children.cycles-pp.get_unmapped_area
3.54 -1.4 2.19 perf-profile.children.cycles-pp.entry_SYSRETQ_unsafe_stack
1.34 ± 3% -1.3 0.07 ± 5% perf-profile.children.cycles-pp.get_any_partial
2.98 -1.2 1.78 perf-profile.children.cycles-pp.thp_get_unmapped_area
3.35 -1.2 2.16 perf-profile.children.cycles-pp.kmem_cache_alloc
2.77 -1.1 1.65 perf-profile.children.cycles-pp.arch_get_unmapped_area_topdown
2.60 -1.0 1.56 perf-profile.children.cycles-pp.perf_event_mmap
3.01 -1.0 1.96 perf-profile.children.cycles-pp.vm_area_alloc
5.97 -1.0 4.94 ± 3% perf-profile.children.cycles-pp.mas_destroy
2.32 -0.9 1.38 perf-profile.children.cycles-pp.vm_unmapped_area
2.33 -0.9 1.39 perf-profile.children.cycles-pp.perf_event_mmap_event
2.13 -0.9 1.26 perf-profile.children.cycles-pp.mas_find
2.71 -0.8 1.88 perf-profile.children.cycles-pp.mab_mas_cp
2.02 -0.8 1.20 perf-profile.children.cycles-pp.free_pgd_range
1.94 -0.8 1.14 perf-profile.children.cycles-pp.mas_mab_cp
5.24 -0.8 4.44 ± 3% perf-profile.children.cycles-pp.mas_nomem
1.66 -0.8 0.90 perf-profile.children.cycles-pp.__call_rcu_common
3.02 -0.8 2.26 perf-profile.children.cycles-pp.mas_preallocate
1.84 -0.7 1.10 perf-profile.children.cycles-pp.free_p4d_range
1.87 -0.7 1.13 perf-profile.children.cycles-pp.entry_SYSCALL_64
1.61 -0.7 0.95 perf-profile.children.cycles-pp.mast_fill_bnode
1.59 -0.6 0.94 perf-profile.children.cycles-pp.free_pud_range
1.54 -0.6 0.92 perf-profile.children.cycles-pp.perf_iterate_sb
1.41 -0.6 0.82 perf-profile.children.cycles-pp.mas_store_b_node
1.42 -0.6 0.84 perf-profile.children.cycles-pp.mas_empty_area_rev
1.24 -0.5 0.74 perf-profile.children.cycles-pp.rcu_all_qs
1.46 -0.5 0.98 perf-profile.children.cycles-pp.mas_pop_node
1.12 -0.5 0.65 perf-profile.children.cycles-pp.mast_split_data
1.46 -0.5 1.00 perf-profile.children.cycles-pp.vm_area_free_rcu_cb
2.66 ± 2% -0.4 2.22 ± 3% perf-profile.children.cycles-pp.kmem_cache_free_bulk
1.08 -0.4 0.64 perf-profile.children.cycles-pp.mas_walk
1.04 -0.4 0.62 perf-profile.children.cycles-pp.mas_rev_awalk
1.08 -0.4 0.66 perf-profile.children.cycles-pp.perf_event_mmap_output
0.95 -0.4 0.58 ± 2% perf-profile.children.cycles-pp.memcpy_orig
0.75 -0.3 0.43 perf-profile.children.cycles-pp.mtree_range_walk
0.98 -0.3 0.66 perf-profile.children.cycles-pp.mas_split_final_node
0.75 -0.3 0.44 perf-profile.children.cycles-pp.__memcg_slab_post_alloc_hook
0.86 -0.3 0.57 perf-profile.children.cycles-pp.mas_push_data
0.70 -0.3 0.41 perf-profile.children.cycles-pp.security_mmap_file
1.48 ± 3% -0.3 1.18 ± 2% perf-profile.children.cycles-pp.rcu_cblist_dequeue
0.66 -0.3 0.39 perf-profile.children.cycles-pp.mas_next_slot
0.70 -0.3 0.42 perf-profile.children.cycles-pp.__memcg_slab_free_hook
0.66 -0.3 0.38 perf-profile.children.cycles-pp.mas_prev_slot
0.64 -0.3 0.37 perf-profile.children.cycles-pp.mas_wr_store_entry
0.59 -0.2 0.34 perf-profile.children.cycles-pp.mas_leaf_max_gap
0.57 -0.2 0.34 ± 2% perf-profile.children.cycles-pp.mas_wr_walk_index
0.53 -0.2 0.30 perf-profile.children.cycles-pp.build_detached_freelist
0.50 -0.2 0.31 perf-profile.children.cycles-pp.entry_SYSCALL_64_safe_stack
0.44 -0.2 0.26 perf-profile.children.cycles-pp.mas_wr_walk
0.39 -0.2 0.22 ± 2% perf-profile.children.cycles-pp.mod_objcg_state
0.43 -0.2 0.26 perf-profile.children.cycles-pp.free_pgtables
0.44 -0.2 0.27 perf-profile.children.cycles-pp.rcu_segcblist_enqueue
0.36 -0.2 0.22 ± 2% perf-profile.children.cycles-pp.userfaultfd_unmap_complete
0.36 ± 2% -0.1 0.21 ± 2% perf-profile.children.cycles-pp.down_write_killable
0.36 -0.1 0.22 ± 2% perf-profile.children.cycles-pp.mas_next_sibling
0.31 -0.1 0.18 ± 2% perf-profile.children.cycles-pp.mas_ascend
0.30 -0.1 0.18 ± 2% perf-profile.children.cycles-pp.tlb_finish_mmu
0.32 ± 2% -0.1 0.20 ± 3% perf-profile.children.cycles-pp.down_write
0.29 ± 3% -0.1 0.18 ± 2% perf-profile.children.cycles-pp.up_write
0.31 ± 2% -0.1 0.20 ± 3% perf-profile.children.cycles-pp.obj_cgroup_charge
0.36 -0.1 0.25 perf-profile.children.cycles-pp.syscall_exit_to_user_mode
0.26 ± 2% -0.1 0.16 ± 2% perf-profile.children.cycles-pp.security_mmap_addr
0.24 ± 3% -0.1 0.15 ± 2% perf-profile.children.cycles-pp.security_vm_enough_memory_mm
0.23 ± 2% -0.1 0.14 ± 3% perf-profile.children.cycles-pp.vma_set_page_prot
0.22 ± 2% -0.1 0.12 ± 3% perf-profile.children.cycles-pp.mt_free_rcu
0.21 ± 2% -0.1 0.12 ± 3% perf-profile.children.cycles-pp.vm_get_page_prot
0.22 ± 2% -0.1 0.13 ± 3% perf-profile.children.cycles-pp.mas_prev_range
0.85 -0.1 0.78 ± 2% perf-profile.children.cycles-pp.shuffle_freelist
0.19 ± 3% -0.1 0.12 ± 3% perf-profile.children.cycles-pp.cap_vm_enough_memory
0.17 ± 2% -0.1 0.10 ± 4% perf-profile.children.cycles-pp.mas_prev
0.17 ± 2% -0.1 0.10 ± 4% perf-profile.children.cycles-pp.remove_vma
0.39 ± 2% -0.1 0.32 perf-profile.children.cycles-pp.__sysvec_apic_timer_interrupt
0.39 ± 2% -0.1 0.32 perf-profile.children.cycles-pp.hrtimer_interrupt
0.16 ± 3% -0.1 0.09 ± 5% perf-profile.children.cycles-pp.testcase
0.16 ± 3% -0.1 0.09 perf-profile.children.cycles-pp.percpu_counter_add_batch
0.16 ± 2% -0.1 0.10 ± 5% perf-profile.children.cycles-pp.sized_strscpy
0.16 ± 3% -0.1 0.09 perf-profile.children.cycles-pp.tlb_gather_mmu
0.16 ± 3% -0.1 0.09 ± 4% perf-profile.children.cycles-pp.strnlen
0.34 ± 3% -0.1 0.28 perf-profile.children.cycles-pp.__hrtimer_run_queues
0.14 ± 2% -0.1 0.08 ± 4% perf-profile.children.cycles-pp.lru_add_drain
0.28 ± 3% -0.1 0.22 ± 3% perf-profile.children.cycles-pp.tick_nohz_handler
0.13 ± 3% -0.1 0.08 ± 6% perf-profile.children.cycles-pp.ima_file_mmap
0.16 ± 3% -0.1 0.10 ± 4% perf-profile.children.cycles-pp.should_failslab
0.13 ± 3% -0.1 0.07 perf-profile.children.cycles-pp.syscall_exit_to_user_mode_prepare
0.17 ± 2% -0.1 0.11 ± 6% perf-profile.children.cycles-pp.current_obj_cgroup
0.13 ± 3% -0.1 0.07 ± 6% perf-profile.children.cycles-pp.mab_calc_split
0.11 ± 5% -0.1 0.06 perf-profile.children.cycles-pp.put_cpu_partial
0.12 -0.0 0.07 perf-profile.children.cycles-pp.mas_wr_store_setup
0.07 ± 6% -0.0 0.02 ± 99% perf-profile.children.cycles-pp.unmap_single_vma
0.25 ± 3% -0.0 0.20 ± 4% perf-profile.children.cycles-pp.update_process_times
0.11 -0.0 0.06 ± 7% perf-profile.children.cycles-pp.downgrade_write
0.12 ± 3% -0.0 0.07 ± 5% perf-profile.children.cycles-pp.mas_data_end
0.12 ± 4% -0.0 0.07 perf-profile.children.cycles-pp.mas_next
0.11 ± 4% -0.0 0.06 ± 6% perf-profile.children.cycles-pp.apparmor_mmap_file
0.10 ± 4% -0.0 0.06 ± 6% perf-profile.children.cycles-pp.cap_mmap_addr
0.11 ± 4% -0.0 0.07 perf-profile.children.cycles-pp.lru_add_drain_cpu
0.10 -0.0 0.06 perf-profile.children.cycles-pp.tlb_batch_pages_flush
0.10 ± 4% -0.0 0.06 ± 8% perf-profile.children.cycles-pp.mas_wr_end_piv
0.10 ± 5% -0.0 0.06 perf-profile.children.cycles-pp.may_expand_vm
0.08 -0.0 0.04 ± 44% perf-profile.children.cycles-pp.mas_node_count_gfp
0.09 ± 4% -0.0 0.05 perf-profile.children.cycles-pp.mas_next_range
0.10 ± 4% -0.0 0.06 perf-profile.children.cycles-pp.cap_mmap_file
0.09 ± 5% -0.0 0.06 ± 8% perf-profile.children.cycles-pp.__memcpy
0.09 ± 5% -0.0 0.06 ± 8% perf-profile.children.cycles-pp.__x64_sys_mmap
0.08 ± 5% -0.0 0.05 perf-profile.children.cycles-pp.mas_prev_setup
0.08 ± 5% -0.0 0.05 perf-profile.children.cycles-pp.refill_obj_stock
0.08 ± 5% -0.0 0.05 perf-profile.children.cycles-pp.__vm_enough_memory
0.10 ± 3% -0.0 0.07 ± 7% perf-profile.children.cycles-pp.kfree
0.08 ± 4% -0.0 0.05 perf-profile.children.cycles-pp.up_read
0.12 -0.0 0.10 ± 4% perf-profile.children.cycles-pp.discard_slab
0.16 ± 3% +0.0 0.20 ± 2% perf-profile.children.cycles-pp.__free_one_page
1.26 +0.1 1.32 ± 2% perf-profile.children.cycles-pp.allocate_slab
0.17 ± 2% +0.1 0.29 ± 6% perf-profile.children.cycles-pp.rmqueue
0.13 ± 2% +0.1 0.26 ± 7% perf-profile.children.cycles-pp.__rmqueue_pcplist
0.10 ± 4% +0.1 0.25 ± 6% perf-profile.children.cycles-pp.rmqueue_bulk
0.31 ± 2% +0.2 0.46 ± 4% perf-profile.children.cycles-pp.__alloc_pages
0.27 ± 2% +0.2 0.42 ± 3% perf-profile.children.cycles-pp.get_page_from_freelist
0.00 +0.5 0.46 ± 3% perf-profile.children.cycles-pp.perf_c2c__record
0.04 ± 44% +0.5 0.50 ± 2% perf-profile.children.cycles-pp.main
0.04 ± 44% +0.5 0.50 ± 2% perf-profile.children.cycles-pp.run_builtin
9.45 +0.5 9.92 ± 2% perf-profile.children.cycles-pp.__kmem_cache_alloc_bulk
0.02 ±141% +0.5 0.50 ± 2% perf-profile.children.cycles-pp.cmd_record
9.06 +0.8 9.81 ± 2% perf-profile.children.cycles-pp.___slab_alloc
6.60 ± 2% +1.1 7.69 ± 3% perf-profile.children.cycles-pp.get_partial_node
8.58 +14.3 22.89 perf-profile.children.cycles-pp.__put_partials
24.83 +24.2 49.03 perf-profile.children.cycles-pp.__do_softirq
24.81 +24.2 49.02 perf-profile.children.cycles-pp.rcu_core
24.80 +24.2 49.01 perf-profile.children.cycles-pp.rcu_do_batch
24.97 +24.3 49.31 perf-profile.children.cycles-pp.kmem_cache_free
2.44 ± 5% +24.8 27.19 ± 3% perf-profile.children.cycles-pp.free_pcppages_bulk
2.55 ± 5% +24.8 27.36 ± 3% perf-profile.children.cycles-pp.free_unref_page
2.49 ± 5% +24.8 27.32 ± 3% perf-profile.children.cycles-pp.free_unref_page_commit
15.02 +28.1 43.09 perf-profile.children.cycles-pp.kthread
15.02 +28.1 43.09 perf-profile.children.cycles-pp.ret_from_fork
15.02 +28.1 43.09 perf-profile.children.cycles-pp.ret_from_fork_asm
14.98 +28.1 43.06 perf-profile.children.cycles-pp.run_ksoftirqd
15.00 +28.1 43.08 perf-profile.children.cycles-pp.smpboot_thread_fn
22.12 +28.3 50.41 perf-profile.children.cycles-pp._raw_spin_lock_irqsave
21.74 +28.3 50.06 perf-profile.children.cycles-pp.native_queued_spin_lock_slowpath
4.12 -1.6 2.52 perf-profile.self.cycles-pp.syscall_return_via_sysret
4.04 -1.6 2.47 perf-profile.self.cycles-pp.zap_pmd_range
4.82 -1.4 3.41 perf-profile.self.cycles-pp.__slab_free
3.50 -1.3 2.17 perf-profile.self.cycles-pp.entry_SYSRETQ_unsafe_stack
3.13 -1.2 1.88 perf-profile.self.cycles-pp.__cond_resched
2.93 -1.2 1.77 perf-profile.self.cycles-pp.mas_topiary_replace
2.38 -0.9 1.45 perf-profile.self.cycles-pp.entry_SYSCALL_64_after_hwframe
1.94 -0.8 1.16 perf-profile.self.cycles-pp.kmem_cache_free
1.64 -0.6 0.99 perf-profile.self.cycles-pp.entry_SYSCALL_64
1.51 -0.6 0.93 perf-profile.self.cycles-pp.mas_wr_spanning_store
2.01 -0.6 1.45 perf-profile.self.cycles-pp.mab_mas_cp
1.41 -0.6 0.85 perf-profile.self.cycles-pp.mas_mab_cp
1.45 -0.6 0.89 perf-profile.self.cycles-pp.free_pud_range
1.25 -0.5 0.79 ± 2% perf-profile.self.cycles-pp.kmem_cache_alloc
1.01 -0.4 0.56 perf-profile.self.cycles-pp.__call_rcu_common
1.14 -0.4 0.71 perf-profile.self.cycles-pp.mas_spanning_rebalance
1.30 -0.4 0.90 perf-profile.self.cycles-pp.mas_pop_node
0.94 -0.4 0.57 perf-profile.self.cycles-pp.mas_walk
0.97 -0.4 0.61 perf-profile.self.cycles-pp.perf_event_mmap_output
0.86 -0.3 0.52 perf-profile.self.cycles-pp.mas_rev_awalk
0.87 -0.3 0.54 perf-profile.self.cycles-pp.rcu_all_qs
0.85 -0.3 0.52 perf-profile.self.cycles-pp.do_vmi_align_munmap
0.84 -0.3 0.51 perf-profile.self.cycles-pp.mast_fill_bnode
0.83 -0.3 0.51 perf-profile.self.cycles-pp.mmap_region
0.78 -0.3 0.47 perf-profile.self.cycles-pp.__kmem_cache_alloc_bulk
0.84 -0.3 0.53 ± 2% perf-profile.self.cycles-pp.memcpy_orig
0.73 -0.3 0.44 perf-profile.self.cycles-pp.mas_split
1.45 ± 3% -0.3 1.17 ± 2% perf-profile.self.cycles-pp.rcu_cblist_dequeue
0.66 -0.3 0.40 perf-profile.self.cycles-pp.mtree_range_walk
0.60 -0.2 0.36 perf-profile.self.cycles-pp.mas_find
0.92 -0.2 0.69 perf-profile.self.cycles-pp.___slab_alloc
0.60 -0.2 0.36 perf-profile.self.cycles-pp.mas_prev_slot
0.59 -0.2 0.35 perf-profile.self.cycles-pp.mas_wr_bnode
0.59 -0.2 0.36 ± 2% perf-profile.self.cycles-pp.mas_next_slot
0.53 -0.2 0.32 perf-profile.self.cycles-pp.mas_leaf_max_gap
0.52 -0.2 0.32 perf-profile.self.cycles-pp.mas_wr_walk_index
0.48 -0.2 0.28 ± 2% perf-profile.self.cycles-pp.build_detached_freelist
0.50 -0.2 0.31 ± 2% perf-profile.self.cycles-pp.__memcg_slab_post_alloc_hook
0.46 -0.2 0.28 ± 2% perf-profile.self.cycles-pp.unmap_page_range
0.42 ± 2% -0.2 0.25 ± 2% perf-profile.self.cycles-pp.perf_iterate_sb
0.43 -0.2 0.27 ± 2% perf-profile.self.cycles-pp.rcu_segcblist_enqueue
0.40 -0.2 0.24 perf-profile.self.cycles-pp.mas_wr_walk
0.40 -0.2 0.25 ± 2% perf-profile.self.cycles-pp.do_syscall_64
0.38 -0.2 0.23 perf-profile.self.cycles-pp.kmem_cache_alloc_bulk
0.34 -0.1 0.20 ± 2% perf-profile.self.cycles-pp.mod_objcg_state
0.36 -0.1 0.22 ± 2% perf-profile.self.cycles-pp.vm_unmapped_area
0.38 -0.1 0.24 ± 2% perf-profile.self.cycles-pp.__memcg_slab_free_hook
0.36 -0.1 0.23 ± 2% perf-profile.self.cycles-pp.mas_store_gfp
0.36 -0.1 0.22 perf-profile.self.cycles-pp.perf_event_mmap_event
0.36 -0.1 0.23 perf-profile.self.cycles-pp.arch_get_unmapped_area_topdown
0.33 ± 2% -0.1 0.20 perf-profile.self.cycles-pp.mas_alloc_nodes
0.33 -0.1 0.20 perf-profile.self.cycles-pp.security_mmap_file
0.33 -0.1 0.20 perf-profile.self.cycles-pp.kmem_cache_free_bulk
0.44 -0.1 0.32 ± 2% perf-profile.self.cycles-pp.mas_push_data
0.34 -0.1 0.21 perf-profile.self.cycles-pp.mas_empty_area_rev
0.32 -0.1 0.19 perf-profile.self.cycles-pp.userfaultfd_unmap_complete
0.30 -0.1 0.18 ± 2% perf-profile.self.cycles-pp.mas_store_b_node
0.27 -0.1 0.15 perf-profile.self.cycles-pp.__mmap
0.30 ± 2% -0.1 0.19 ± 2% perf-profile.self.cycles-pp.mas_wr_store_entry
0.27 ± 2% -0.1 0.15 ± 3% perf-profile.self.cycles-pp.__munmap
0.22 ± 3% -0.1 0.11 ± 4% perf-profile.self.cycles-pp.get_partial_node
0.27 -0.1 0.16 ± 3% perf-profile.self.cycles-pp.mas_ascend
0.27 ± 2% -0.1 0.16 ± 3% perf-profile.self.cycles-pp.mas_preallocate
0.28 ± 2% -0.1 0.17 ± 2% perf-profile.self.cycles-pp.do_vmi_munmap
0.24 ± 2% -0.1 0.14 ± 3% perf-profile.self.cycles-pp.__vm_munmap
0.26 ± 2% -0.1 0.16 perf-profile.self.cycles-pp.do_mmap
0.26 ± 2% -0.1 0.16 perf-profile.self.cycles-pp.up_write
0.25 -0.1 0.15 perf-profile.self.cycles-pp.vm_area_alloc
0.24 -0.1 0.15 ± 2% perf-profile.self.cycles-pp.unmap_region
0.14 -0.1 0.05 perf-profile.self.cycles-pp.get_any_partial
0.23 -0.1 0.14 ± 2% perf-profile.self.cycles-pp.perf_event_mmap
0.21 ± 2% -0.1 0.12 ± 3% perf-profile.self.cycles-pp.mt_free_rcu
0.22 -0.1 0.13 ± 2% perf-profile.self.cycles-pp.mas_destroy
0.22 ± 2% -0.1 0.14 perf-profile.self.cycles-pp.free_p4d_range
0.22 ± 2% -0.1 0.13 ± 3% perf-profile.self.cycles-pp.mas_next_sibling
0.19 ± 3% -0.1 0.11 ± 3% perf-profile.self.cycles-pp.vm_get_page_prot
0.20 ± 2% -0.1 0.12 ± 3% perf-profile.self.cycles-pp.down_write_killable
0.21 ± 2% -0.1 0.14 ± 3% perf-profile.self.cycles-pp.obj_cgroup_charge
0.20 ± 2% -0.1 0.12 perf-profile.self.cycles-pp.mast_split_data
0.38 ± 2% -0.1 0.30 ± 2% perf-profile.self.cycles-pp.vm_area_free_rcu_cb
0.18 ± 3% -0.1 0.11 perf-profile.self.cycles-pp.free_pgtables
0.19 -0.1 0.12 ± 3% perf-profile.self.cycles-pp.thp_get_unmapped_area
0.19 -0.1 0.12 ± 3% perf-profile.self.cycles-pp.vm_mmap_pgoff
0.19 ± 2% -0.1 0.12 ± 3% perf-profile.self.cycles-pp.tlb_finish_mmu
0.17 ± 2% -0.1 0.10 ± 4% perf-profile.self.cycles-pp.cap_vm_enough_memory
0.17 ± 2% -0.1 0.11 ± 6% perf-profile.self.cycles-pp.mas_split_final_node
0.17 ± 2% -0.1 0.10 ± 4% perf-profile.self.cycles-pp.syscall_exit_to_user_mode
0.16 ± 2% -0.1 0.10 ± 3% perf-profile.self.cycles-pp.mas_store_prealloc
0.17 ± 2% -0.1 0.11 ± 4% perf-profile.self.cycles-pp.unmap_vmas
0.16 -0.1 0.10 ± 4% perf-profile.self.cycles-pp.free_pgd_range
0.20 ± 2% -0.1 0.13 ± 5% perf-profile.self.cycles-pp.down_write
0.17 ± 4% -0.1 0.11 ± 3% perf-profile.self.cycles-pp.__put_partials
0.15 ± 3% -0.1 0.09 ± 4% perf-profile.self.cycles-pp.rcu_do_batch
0.14 -0.1 0.08 ± 5% perf-profile.self.cycles-pp.security_mmap_addr
0.14 ± 3% -0.1 0.08 perf-profile.self.cycles-pp.tlb_gather_mmu
0.14 ± 3% -0.1 0.08 ± 5% perf-profile.self.cycles-pp.strnlen
0.13 ± 3% -0.1 0.08 perf-profile.self.cycles-pp.mas_prev_range
0.13 ± 5% -0.1 0.08 perf-profile.self.cycles-pp.percpu_counter_add_batch
0.11 ± 4% -0.1 0.06 ± 9% perf-profile.self.cycles-pp.put_cpu_partial
0.13 ± 2% -0.1 0.08 ± 4% perf-profile.self.cycles-pp.current_obj_cgroup
0.13 -0.1 0.08 perf-profile.self.cycles-pp.sized_strscpy
0.13 ± 2% -0.0 0.08 ± 4% perf-profile.self.cycles-pp.get_unmapped_area
0.12 ± 3% -0.0 0.07 perf-profile.self.cycles-pp.ima_file_mmap
0.12 -0.0 0.07 ± 5% perf-profile.self.cycles-pp.mas_prev
0.11 -0.0 0.06 ± 7% perf-profile.self.cycles-pp.syscall_exit_to_user_mode_prepare
0.12 ± 3% -0.0 0.08 ± 4% perf-profile.self.cycles-pp.entry_SYSCALL_64_safe_stack
0.10 ± 3% -0.0 0.06 ± 8% perf-profile.self.cycles-pp.downgrade_write
0.10 -0.0 0.06 perf-profile.self.cycles-pp.__x64_sys_munmap
0.10 -0.0 0.06 perf-profile.self.cycles-pp.lru_add_drain_cpu
0.10 ± 4% -0.0 0.06 ± 6% perf-profile.self.cycles-pp.apparmor_mmap_file
0.11 ± 3% -0.0 0.07 perf-profile.self.cycles-pp.mab_calc_split
0.10 ± 3% -0.0 0.06 ± 6% perf-profile.self.cycles-pp.vma_set_page_prot
0.07 -0.0 0.03 ± 70% perf-profile.self.cycles-pp.tlb_batch_pages_flush
0.09 ± 4% -0.0 0.05 ± 7% perf-profile.self.cycles-pp.may_expand_vm
0.08 ± 6% -0.0 0.04 ± 44% perf-profile.self.cycles-pp.kfree
0.09 ± 5% -0.0 0.05 ± 7% perf-profile.self.cycles-pp.cap_mmap_file
0.10 ± 4% -0.0 0.07 ± 5% perf-profile.self.cycles-pp.mas_data_end
0.38 -0.0 0.35 perf-profile.self.cycles-pp._raw_spin_lock_irqsave
0.09 -0.0 0.06 ± 8% perf-profile.self.cycles-pp.cap_mmap_addr
0.09 -0.0 0.06 ± 8% perf-profile.self.cycles-pp.mas_wr_store_setup
0.09 ± 4% -0.0 0.06 ± 6% perf-profile.self.cycles-pp.remove_vma
0.08 ± 5% -0.0 0.05 perf-profile.self.cycles-pp.refill_obj_stock
0.08 ± 4% -0.0 0.05 perf-profile.self.cycles-pp.mas_nomem
0.08 -0.0 0.05 perf-profile.self.cycles-pp.__x64_sys_mmap
0.08 -0.0 0.05 perf-profile.self.cycles-pp.mas_next
0.13 -0.0 0.10 ± 3% perf-profile.self.cycles-pp.setup_object
0.63 -0.0 0.61 perf-profile.self.cycles-pp.shuffle_freelist
0.12 -0.0 0.10 ± 4% perf-profile.self.cycles-pp.discard_slab
0.08 ± 4% +0.0 0.10 ± 4% perf-profile.self.cycles-pp.rmqueue_bulk
0.16 ± 3% +0.0 0.19 ± 2% perf-profile.self.cycles-pp.__free_one_page
0.00 +0.1 0.06 ± 9% perf-profile.self.cycles-pp.get_page_from_freelist
0.00 +0.1 0.06 ± 7% perf-profile.self.cycles-pp.free_unref_page_commit
21.74 +28.3 50.06 perf-profile.self.cycles-pp.native_queued_spin_lock_slowpath
***************************************************************************************************
lkp-csl-2sp7: 96 threads 2 sockets Intel(R) Xeon(R) Gold 6252 CPU @ 2.10GHz (Cascade Lake) with 512G memory
=========================================================================================
compiler/cpufreq_governor/kconfig/rootfs/tbox_group/test/testcase:
gcc-12/performance/x86_64-rhel-8.3/debian-12-x86_64-phoronix/lkp-csl-2sp7/asmfish-1.1.2/phoronix-test-suite
commit:
9ef3ad0f03 ("ACPI/NUMA: Print CXL Early Discovery Table (CEDT)")
09f84589a8 ("ACPI/NUMA: Remove architecture dependent remainings")
9ef3ad0f032818c5 09f84589a8ed1162c0274a1b5b3
---------------- ---------------------------
%stddev %change %stddev
\ | \
14.98 +1.9% 15.27 boot-time.dhcp
93848406 -23.9% 71464814 phoronix-test-suite.asmfish.0.nodes_second
0.06 ± 61% -64.0% 0.02 ± 8% vmstat.procs.b
1183 ± 26% +689.3% 9337 ± 9% perf-c2c.DRAM.remote
112.50 ± 36% +2692.2% 3141 ± 10% perf-c2c.HITM.remote
11131904 ± 13% -33.9% 7360921 ± 7% meminfo.DirectMap2M
1251 ± 31% +56.6% 1959 ± 19% meminfo.Mlocked
4391 ± 8% +16.1% 5099 ± 7% meminfo.Unevictable
260856 -84.1% 41446 meminfo.VmallocUsed
0.05 ± 11% -0.0 0.01 ± 4% mpstat.cpu.all.iowait%
0.02 ± 3% -0.0 0.02 ± 6% mpstat.cpu.all.soft%
0.06 ± 6% -0.0 0.04 mpstat.cpu.all.sys%
31.83 ± 2% -17.1% 26.40 mpstat.max_utilization.seconds
986818 ± 45% -100.0% 0.00 numa-numastat.node0.local_node
1027333 ± 45% -100.0% 0.00 numa-numastat.node0.numa_hit
40474 ± 74% -100.0% 0.00 numa-numastat.node0.other_node
968499 ± 58% -100.0% 0.00 numa-numastat.node1.local_node
1032398 ± 53% -100.0% 0.00 numa-numastat.node1.numa_hit
63895 ± 44% -100.0% 0.00 numa-numastat.node1.other_node
313.19 ± 31% +56.4% 489.72 ± 19% proc-vmstat.nr_mlock
2722 -3.0% 2641 proc-vmstat.nr_page_table_pages
58851 -0.7% 58431 proc-vmstat.nr_slab_unreclaimable
1098 ± 8% +16.1% 1274 ± 7% proc-vmstat.nr_unevictable
1098 ± 8% +16.1% 1274 ± 7% proc-vmstat.nr_zone_unevictable
49674 ± 59% -100.0% 0.00 proc-vmstat.numa_hint_faults
36892 ± 57% -100.0% 0.00 proc-vmstat.numa_hint_faults_local
2061577 ± 48% -34.9% 1342268 proc-vmstat.numa_hit
14970 ± 76% -100.0% 0.00 proc-vmstat.numa_huge_pte_updates
1957163 ± 50% -31.4% 1342877 proc-vmstat.numa_local
104370 ± 3% -100.0% 0.00 proc-vmstat.numa_other
1784768 ± 75% -100.0% 0.00 proc-vmstat.numa_pages_migrated
7772796 ± 75% -100.0% 0.00 proc-vmstat.numa_pte_updates
5403170 ± 62% -54.3% 2470734 proc-vmstat.pgalloc_normal
2641844 ± 60% -45.6% 1438131 proc-vmstat.pgfault
5223988 ± 64% -56.2% 2290502 proc-vmstat.pgfree
1784768 ± 75% -100.0% 0.00 proc-vmstat.pgmigrate_success
98843 ± 61% -39.3% 60039 ± 3% proc-vmstat.pgreuse
3478 ± 75% -100.0% 0.00 proc-vmstat.thp_migration_success
137922 ± 38% -100.0% 0.00 numa-meminfo.node0.Active
39195 ± 31% -100.0% 0.00 numa-meminfo.node0.Active(anon)
98727 ± 43% -100.0% 0.00 numa-meminfo.node0.Active(file)
867525 ± 22% -100.0% 0.00 numa-meminfo.node0.AnonHugePages
1116586 ± 19% -100.0% 0.00 numa-meminfo.node0.AnonPages
1448118 ± 15% -100.0% 0.00 numa-meminfo.node0.AnonPages.max
1414495 ± 62% -100.0% 0.00 numa-meminfo.node0.FilePages
2393152 ± 28% -100.0% 0.00 numa-meminfo.node0.Inactive
1467103 ± 16% -100.0% 0.00 numa-meminfo.node0.Inactive(anon)
926048 ± 78% -100.0% 0.00 numa-meminfo.node0.Inactive(file)
42206 ± 33% -100.0% 0.00 numa-meminfo.node0.KReclaimable
10393 ± 8% -100.0% 0.00 numa-meminfo.node0.KernelStack
84153 ± 30% -100.0% 0.00 numa-meminfo.node0.Mapped
2.606e+08 -100.0% 0.00 numa-meminfo.node0.MemFree
2.638e+08 -100.0% 0.00 numa-meminfo.node0.MemTotal
3202155 ± 24% -100.0% 0.00 numa-meminfo.node0.MemUsed
5609 ± 19% -100.0% 0.00 numa-meminfo.node0.PageTables
42206 ± 33% -100.0% 0.00 numa-meminfo.node0.SReclaimable
110987 ± 21% -100.0% 0.00 numa-meminfo.node0.SUnreclaim
389372 ± 45% -100.0% 0.00 numa-meminfo.node0.Shmem
153193 ± 23% -100.0% 0.00 numa-meminfo.node0.Slab
3500 ± 14% -100.0% 0.00 numa-meminfo.node0.Unevictable
278067 ± 19% -100.0% 0.00 numa-meminfo.node1.Active
214635 ± 8% -100.0% 0.00 numa-meminfo.node1.Active(anon)
63432 ± 67% -100.0% 0.00 numa-meminfo.node1.Active(file)
680824 ± 23% -100.0% 0.00 numa-meminfo.node1.AnonHugePages
829328 ± 25% -100.0% 0.00 numa-meminfo.node1.AnonPages
1414486 ± 14% -100.0% 0.00 numa-meminfo.node1.AnonPages.max
1390641 ± 64% -100.0% 0.00 numa-meminfo.node1.FilePages
1942799 ± 35% -100.0% 0.00 numa-meminfo.node1.Inactive
1173590 ± 19% -100.0% 0.00 numa-meminfo.node1.Inactive(anon)
769209 ± 94% -100.0% 0.00 numa-meminfo.node1.Inactive(file)
42052 ± 34% -100.0% 0.00 numa-meminfo.node1.KReclaimable
10353 ± 7% -100.0% 0.00 numa-meminfo.node1.KernelStack
56786 ± 44% -100.0% 0.00 numa-meminfo.node1.Mapped
2.612e+08 -100.0% 0.00 numa-meminfo.node1.MemFree
2.642e+08 -100.0% 0.00 numa-meminfo.node1.MemTotal
3009654 ± 26% -100.0% 0.00 numa-meminfo.node1.MemUsed
5280 ± 19% -100.0% 0.00 numa-meminfo.node1.PageTables
42052 ± 34% -100.0% 0.00 numa-meminfo.node1.SReclaimable
124421 ± 19% -100.0% 0.00 numa-meminfo.node1.SUnreclaim
557220 ± 32% -100.0% 0.00 numa-meminfo.node1.Shmem
166473 ± 22% -100.0% 0.00 numa-meminfo.node1.Slab
9797 ± 31% -100.0% 0.00 numa-vmstat.node0.nr_active_anon
24680 ± 43% -100.0% 0.00 numa-vmstat.node0.nr_active_file
279210 ± 19% -100.0% 0.00 numa-vmstat.node0.nr_anon_pages
423.69 ± 22% -100.0% 0.00 numa-vmstat.node0.nr_anon_transparent_hugepages
353618 ± 62% -100.0% 0.00 numa-vmstat.node0.nr_file_pages
65159849 -100.0% 0.00 numa-vmstat.node0.nr_free_pages
366837 ± 16% -100.0% 0.00 numa-vmstat.node0.nr_inactive_anon
231510 ± 78% -100.0% 0.00 numa-vmstat.node0.nr_inactive_file
10394 ± 8% -100.0% 0.00 numa-vmstat.node0.nr_kernel_stack
21187 ± 29% -100.0% 0.00 numa-vmstat.node0.nr_mapped
1402 ± 19% -100.0% 0.00 numa-vmstat.node0.nr_page_table_pages
97341 ± 45% -100.0% 0.00 numa-vmstat.node0.nr_shmem
10551 ± 33% -100.0% 0.00 numa-vmstat.node0.nr_slab_reclaimable
27747 ± 21% -100.0% 0.00 numa-vmstat.node0.nr_slab_unreclaimable
875.15 ± 14% -100.0% 0.00 numa-vmstat.node0.nr_unevictable
9797 ± 31% -100.0% 0.00 numa-vmstat.node0.nr_zone_active_anon
24680 ± 43% -100.0% 0.00 numa-vmstat.node0.nr_zone_active_file
366837 ± 16% -100.0% 0.00 numa-vmstat.node0.nr_zone_inactive_anon
231510 ± 78% -100.0% 0.00 numa-vmstat.node0.nr_zone_inactive_file
875.15 ± 14% -100.0% 0.00 numa-vmstat.node0.nr_zone_unevictable
1027333 ± 45% -100.0% 0.00 numa-vmstat.node0.numa_hit
986816 ± 45% -100.0% 0.00 numa-vmstat.node0.numa_local
40474 ± 74% -100.0% 0.00 numa-vmstat.node0.numa_other
53643 ± 8% -100.0% 0.00 numa-vmstat.node1.nr_active_anon
15857 ± 67% -100.0% 0.00 numa-vmstat.node1.nr_active_file
207230 ± 25% -100.0% 0.00 numa-vmstat.node1.nr_anon_pages
332.22 ± 23% -100.0% 0.00 numa-vmstat.node1.nr_anon_transparent_hugepages
347657 ± 64% -100.0% 0.00 numa-vmstat.node1.nr_file_pages
65294238 -100.0% 0.00 numa-vmstat.node1.nr_free_pages
293307 ± 19% -100.0% 0.00 numa-vmstat.node1.nr_inactive_anon
192302 ± 94% -100.0% 0.00 numa-vmstat.node1.nr_inactive_file
10354 ± 7% -100.0% 0.00 numa-vmstat.node1.nr_kernel_stack
14315 ± 43% -100.0% 0.00 numa-vmstat.node1.nr_mapped
1320 ± 19% -100.0% 0.00 numa-vmstat.node1.nr_page_table_pages
139303 ± 32% -100.0% 0.00 numa-vmstat.node1.nr_shmem
10512 ± 34% -100.0% 0.00 numa-vmstat.node1.nr_slab_reclaimable
31105 ± 19% -100.0% 0.00 numa-vmstat.node1.nr_slab_unreclaimable
53643 ± 8% -100.0% 0.00 numa-vmstat.node1.nr_zone_active_anon
15857 ± 67% -100.0% 0.00 numa-vmstat.node1.nr_zone_active_file
293306 ± 19% -100.0% 0.00 numa-vmstat.node1.nr_zone_inactive_anon
192302 ± 94% -100.0% 0.00 numa-vmstat.node1.nr_zone_inactive_file
1032447 ± 53% -100.0% 0.00 numa-vmstat.node1.numa_hit
968548 ± 57% -100.0% 0.00 numa-vmstat.node1.numa_local
63895 ± 44% -100.0% 0.00 numa-vmstat.node1.numa_other
2.01 ± 5% -10.8% 1.80 ± 6% perf-sched.sch_delay.avg.ms.__cond_resched.stop_one_cpu.sched_exec.bprm_execve.part
0.02 ± 25% -31.8% 0.01 ± 9% perf-sched.sch_delay.avg.ms.do_nanosleep.hrtimer_nanosleep.__x64_sys_nanosleep.do_syscall_64
0.24 ± 20% -28.4% 0.17 ± 20% perf-sched.sch_delay.avg.ms.do_task_dead.do_exit.do_group_exit.__x64_sys_exit_group.do_syscall_64
0.01 ± 8% -100.0% 0.00 perf-sched.sch_delay.avg.ms.io_schedule.migration_entry_wait_on_locked.__handle_mm_fault.handle_mm_fault
0.00 ± 33% -100.0% 0.00 perf-sched.sch_delay.avg.ms.io_schedule.migration_entry_wait_on_locked.migration_entry_wait.do_swap_page
0.02 ± 4% +12.1% 0.02 ± 3% perf-sched.sch_delay.avg.ms.irq_thread.kthread.ret_from_fork.ret_from_fork_asm
0.02 ± 2% +21.2% 0.02 ± 6% perf-sched.sch_delay.avg.ms.rcu_gp_kthread.kthread.ret_from_fork.ret_from_fork_asm
0.02 ± 50% -35.5% 0.01 ± 7% perf-sched.sch_delay.avg.ms.schedule_hrtimeout_range_clock.ep_poll.do_epoll_wait.do_epoll_pwait.part
0.02 ± 4% +21.3% 0.02 ± 4% perf-sched.sch_delay.avg.ms.schedule_timeout.kcompactd.kthread.ret_from_fork
0.02 ± 19% -28.6% 0.02 ± 17% perf-sched.sch_delay.avg.ms.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
3.91 -21.9% 3.05 ± 23% perf-sched.sch_delay.max.ms.do_wait.kernel_wait4.__do_sys_wait4.do_syscall_64
0.04 ± 13% -100.0% 0.00 perf-sched.sch_delay.max.ms.io_schedule.migration_entry_wait_on_locked.__handle_mm_fault.handle_mm_fault
0.02 ± 33% -100.0% 0.00 perf-sched.sch_delay.max.ms.io_schedule.migration_entry_wait_on_locked.migration_entry_wait.do_swap_page
0.32 ±208% -100.0% 0.00 perf-sched.sch_delay.max.ms.irqentry_exit_to_user_mode.asm_sysvec_apic_timer_interrupt.[unknown].[unknown]
0.01 ± 35% -100.0% 0.00 perf-sched.sch_delay.max.ms.irqentry_exit_to_user_mode.asm_sysvec_call_function_single.[unknown].[unknown]
0.03 ± 45% -100.0% 0.00 perf-sched.sch_delay.max.ms.irqentry_exit_to_user_mode.asm_sysvec_reschedule_ipi.[unknown].[unknown]
2.31 ± 31% -55.0% 1.04 ± 98% perf-sched.sch_delay.max.ms.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
65.46 ± 6% -15.5% 55.34 ± 2% perf-sched.total_wait_and_delay.average.ms
65.43 ± 6% -15.5% 55.31 ± 2% perf-sched.total_wait_time.average.ms
2.01 ± 5% -100.0% 0.00 perf-sched.wait_and_delay.avg.ms.__cond_resched.stop_one_cpu.sched_exec.bprm_execve.part
3.39 ± 9% -100.0% 0.00 perf-sched.wait_and_delay.avg.ms.do_wait.kernel_wait4.__do_sys_wait4.do_syscall_64
0.62 ± 7% -100.0% 0.00 perf-sched.wait_and_delay.avg.ms.io_schedule.migration_entry_wait_on_locked.__handle_mm_fault.handle_mm_fault
0.15 ± 11% -18.2% 0.13 ± 3% perf-sched.wait_and_delay.avg.ms.irqentry_exit_to_user_mode.asm_sysvec_apic_timer_interrupt.[unknown].[unknown]
51.61 ± 7% -12.3% 45.26 ± 6% perf-sched.wait_and_delay.avg.ms.pipe_read.vfs_read.ksys_read.do_syscall_64
3.59 ± 16% -28.0% 2.58 ± 21% perf-sched.wait_and_delay.avg.ms.schedule_hrtimeout_range_clock.ep_poll.do_epoll_wait.do_epoll_pwait.part
2.44 ± 20% -100.0% 0.00 perf-sched.wait_and_delay.avg.ms.schedule_timeout.__wait_for_common.wait_for_completion_state.kernel_clone
18.95 ± 4% +50.7% 28.56 ± 4% perf-sched.wait_and_delay.avg.ms.schedule_timeout.rcu_gp_fqs_loop.rcu_gp_kthread.kthread
494.08 +11.4% 550.61 perf-sched.wait_and_delay.avg.ms.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
37.67 ± 18% -50.1% 18.80 ± 6% perf-sched.wait_and_delay.count.__cond_resched.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
104.50 ± 5% -100.0% 0.00 perf-sched.wait_and_delay.count.__cond_resched.stop_one_cpu.sched_exec.bprm_execve.part
124.83 -48.7% 64.00 perf-sched.wait_and_delay.count.do_task_dead.do_exit.do_group_exit.__x64_sys_exit_group.do_syscall_64
114.67 ± 2% -100.0% 0.00 perf-sched.wait_and_delay.count.do_wait.kernel_wait4.__do_sys_wait4.do_syscall_64
749.83 ± 5% -100.0% 0.00 perf-sched.wait_and_delay.count.io_schedule.migration_entry_wait_on_locked.__handle_mm_fault.handle_mm_fault
1323 ± 6% -26.3% 975.20 ± 8% perf-sched.wait_and_delay.count.pipe_read.vfs_read.ksys_read.do_syscall_64
1427 ± 14% +50.1% 2143 ± 10% perf-sched.wait_and_delay.count.schedule_hrtimeout_range_clock.ep_poll.do_epoll_wait.do_epoll_pwait.part
86.17 -100.0% 0.00 perf-sched.wait_and_delay.count.schedule_timeout.__wait_for_common.wait_for_completion_state.kernel_clone
19.83 -51.6% 9.60 ± 5% perf-sched.wait_and_delay.count.schedule_timeout.kcompactd.kthread.ret_from_fork
254.33 ± 5% -33.6% 168.80 ± 5% perf-sched.wait_and_delay.count.schedule_timeout.rcu_gp_fqs_loop.rcu_gp_kthread.kthread
854.17 -10.1% 768.00 perf-sched.wait_and_delay.count.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
574.00 ± 7% -41.3% 336.80 ± 9% perf-sched.wait_and_delay.count.worker_thread.kthread.ret_from_fork.ret_from_fork_asm
3.93 -100.0% 0.00 perf-sched.wait_and_delay.max.ms.__cond_resched.stop_one_cpu.sched_exec.bprm_execve.part
47.29 ± 55% -100.0% 0.00 perf-sched.wait_and_delay.max.ms.do_wait.kernel_wait4.__do_sys_wait4.do_syscall_64
2.88 ± 19% -100.0% 0.00 perf-sched.wait_and_delay.max.ms.io_schedule.migration_entry_wait_on_locked.__handle_mm_fault.handle_mm_fault
31.00 ± 94% -100.0% 0.00 perf-sched.wait_and_delay.max.ms.schedule_timeout.__wait_for_common.wait_for_completion_state.kernel_clone
288.79 ±116% -98.7% 3.87 ± 6% perf-sched.wait_time.avg.ms.__cond_resched.__alloc_pages.alloc_pages_mpol.shmem_alloc_folio.shmem_alloc_and_add_folio
0.61 ± 7% -100.0% 0.00 perf-sched.wait_time.avg.ms.io_schedule.migration_entry_wait_on_locked.__handle_mm_fault.handle_mm_fault
0.07 ± 41% -100.0% 0.00 perf-sched.wait_time.avg.ms.io_schedule.migration_entry_wait_on_locked.migration_entry_wait.do_swap_page
0.15 ± 11% -18.1% 0.13 ± 3% perf-sched.wait_time.avg.ms.irqentry_exit_to_user_mode.asm_sysvec_apic_timer_interrupt.[unknown].[unknown]
51.58 ± 7% -12.7% 45.03 ± 7% perf-sched.wait_time.avg.ms.pipe_read.vfs_read.ksys_read.do_syscall_64
3.57 ± 16% -28.0% 2.57 ± 21% perf-sched.wait_time.avg.ms.schedule_hrtimeout_range_clock.ep_poll.do_epoll_wait.do_epoll_pwait.part
2.17 ± 18% +49.0% 3.23 ± 18% perf-sched.wait_time.avg.ms.schedule_timeout.__wait_for_common.wait_for_completion_state.kernel_clone
18.94 ± 4% +50.7% 28.54 ± 4% perf-sched.wait_time.avg.ms.schedule_timeout.rcu_gp_fqs_loop.rcu_gp_kthread.kthread
494.06 +11.4% 550.60 perf-sched.wait_time.avg.ms.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
2.87 ± 19% -100.0% 0.00 perf-sched.wait_time.max.ms.io_schedule.migration_entry_wait_on_locked.__handle_mm_fault.handle_mm_fault
0.74 ± 86% -100.0% 0.00 perf-sched.wait_time.max.ms.io_schedule.migration_entry_wait_on_locked.migration_entry_wait.do_swap_page
4.83 ± 14% -33.7% 3.21 ± 49% perf-sched.wait_time.max.ms.irqentry_exit_to_user_mode.asm_sysvec_apic_timer_interrupt.[unknown]
0.86 +113.9% 1.84 perf-stat.i.MPKI
2.197e+10 -22.8% 1.697e+10 perf-stat.i.branch-instructions
8.68 -0.2 8.52 perf-stat.i.branch-miss-rate%
1.9e+09 -23.3% 1.458e+09 perf-stat.i.branch-misses
27.65 +23.9 51.52 perf-stat.i.cache-miss-rate%
1.621e+08 +77.1% 2.871e+08 perf-stat.i.cache-misses
5.87e+08 -7.6% 5.422e+08 perf-stat.i.cache-references
1.16 +38.1% 1.60 perf-stat.i.cpi
2.225e+11 +7.5% 2.393e+11 perf-stat.i.cpu-cycles
154.02 -16.1% 129.22 perf-stat.i.cpu-migrations
1391 -38.3% 857.88 perf-stat.i.cycles-between-cache-misses
0.06 ± 4% -0.0 0.05 ± 8% perf-stat.i.dTLB-load-miss-rate%
16519218 ± 3% -27.0% 12056970 ± 11% perf-stat.i.dTLB-load-misses
6.726e+10 -22.7% 5.201e+10 perf-stat.i.dTLB-loads
325201 -12.1% 285915 ± 3% perf-stat.i.dTLB-store-misses
2.069e+10 -22.8% 1.597e+10 perf-stat.i.dTLB-stores
1471789 -7.6% 1360644 ± 2% perf-stat.i.iTLB-load-misses
2.235e+11 -22.7% 1.728e+11 perf-stat.i.instructions
161257 -16.2% 135135 perf-stat.i.instructions-per-iTLB-miss
0.98 -26.7% 0.72 perf-stat.i.ipc
2.32 +7.5% 2.49 perf-stat.i.metric.GHz
316.36 +194.6% 931.87 perf-stat.i.metric.K/sec
1151 -22.6% 890.51 perf-stat.i.metric.M/sec
3133 ± 4% -46.7% 1669 perf-stat.i.minor-faults
27.70 +52.6 80.31 perf-stat.i.node-load-miss-rate%
5227534 ± 2% +993.2% 57147855 perf-stat.i.node-load-misses
15752155 -13.8% 13581362 perf-stat.i.node-loads
51.43 +41.3 92.70 perf-stat.i.node-store-miss-rate%
2278574 +614.2% 16272457 ± 2% perf-stat.i.node-store-misses
2080120 -48.2% 1078354 perf-stat.i.node-stores
3133 ± 4% -46.7% 1669 perf-stat.i.page-faults
0.73 +129.0% 1.66 perf-stat.overall.MPKI
27.63 +25.3 52.96 perf-stat.overall.cache-miss-rate%
1.00 +39.1% 1.38 perf-stat.overall.cpi
1372 -39.3% 833.36 perf-stat.overall.cycles-between-cache-misses
0.00 +0.0 0.00 ± 3% perf-stat.overall.dTLB-store-miss-rate%
151892 -16.2% 127248 ± 3% perf-stat.overall.instructions-per-iTLB-miss
1.00 -28.1% 0.72 perf-stat.overall.ipc
24.91 +55.9 80.79 perf-stat.overall.node-load-miss-rate%
52.27 +41.5 93.78 perf-stat.overall.node-store-miss-rate%
2.194e+10 -22.7% 1.696e+10 perf-stat.ps.branch-instructions
1.898e+09 -23.2% 1.457e+09 perf-stat.ps.branch-misses
1.619e+08 +77.2% 2.869e+08 perf-stat.ps.cache-misses
5.86e+08 -7.6% 5.417e+08 perf-stat.ps.cache-references
2.222e+11 +7.6% 2.391e+11 perf-stat.ps.cpu-cycles
153.79 -16.1% 128.97 perf-stat.ps.cpu-migrations
16497488 ± 3% -26.9% 12054157 ± 11% perf-stat.ps.dTLB-load-misses
6.717e+10 -22.6% 5.198e+10 perf-stat.ps.dTLB-loads
324755 -12.1% 285616 ± 3% perf-stat.ps.dTLB-store-misses
2.066e+10 -22.8% 1.596e+10 perf-stat.ps.dTLB-stores
1469530 -7.6% 1358463 ± 2% perf-stat.ps.iTLB-load-misses
2.232e+11 -22.6% 1.727e+11 perf-stat.ps.instructions
3129 ± 4% -46.8% 1664 perf-stat.ps.minor-faults
5220451 ± 2% +993.6% 57090593 perf-stat.ps.node-load-misses
15736071 -13.8% 13570457 perf-stat.ps.node-loads
2275264 +614.3% 16251080 ± 2% perf-stat.ps.node-store-misses
2077508 -48.1% 1077622 perf-stat.ps.node-stores
3129 ± 4% -46.8% 1664 perf-stat.ps.page-faults
2.84 ± 72% -2.5 0.34 ± 82% perf-profile.calltrace.cycles-pp.asm_exc_page_fault
2.78 ± 74% -2.4 0.33 ± 81% perf-profile.calltrace.cycles-pp.exc_page_fault.asm_exc_page_fault
9.33 ± 6% -1.9 7.45 ± 8% perf-profile.calltrace.cycles-pp.ktime_get_update_offsets_now.hrtimer_interrupt.__sysvec_apic_timer_interrupt.sysvec_apic_timer_interrupt.asm_sysvec_apic_timer_interrupt
0.96 ± 12% -0.4 0.55 ± 54% perf-profile.calltrace.cycles-pp.rcu_core.__do_softirq.irq_exit_rcu.sysvec_apic_timer_interrupt.asm_sysvec_apic_timer_interrupt
1.38 ± 11% -0.3 1.12 ± 8% perf-profile.calltrace.cycles-pp.rcu_sched_clock_irq.update_process_times.tick_nohz_handler.__hrtimer_run_queues.hrtimer_interrupt
0.50 ± 46% +0.4 0.87 ± 8% perf-profile.calltrace.cycles-pp.kthread.ret_from_fork.ret_from_fork_asm
0.50 ± 46% +0.4 0.87 ± 8% perf-profile.calltrace.cycles-pp.ret_from_fork.ret_from_fork_asm
0.50 ± 46% +0.4 0.87 ± 8% perf-profile.calltrace.cycles-pp.ret_from_fork_asm
2.71 ± 5% +0.4 3.09 ± 3% perf-profile.calltrace.cycles-pp.__update_load_avg_se.update_load_avg.task_tick_fair.scheduler_tick.update_process_times
0.52 ± 46% +0.5 1.03 ± 28% perf-profile.calltrace.cycles-pp.hrtimer_forward.tick_nohz_handler.__hrtimer_run_queues.hrtimer_interrupt.__sysvec_apic_timer_interrupt
7.24 ± 6% +0.7 7.92 ± 5% perf-profile.calltrace.cycles-pp.update_load_avg.task_tick_fair.scheduler_tick.update_process_times.tick_nohz_handler
3.41 ± 12% +1.5 4.91 ± 13% perf-profile.calltrace.cycles-pp.ktime_get.tick_nohz_handler.__hrtimer_run_queues.hrtimer_interrupt.__sysvec_apic_timer_interrupt
18.04 ± 4% +1.8 19.80 ± 5% perf-profile.calltrace.cycles-pp.task_tick_fair.scheduler_tick.update_process_times.tick_nohz_handler.__hrtimer_run_queues
3.34 ± 61% -2.5 0.87 ± 9% perf-profile.children.cycles-pp.asm_exc_page_fault
3.19 ± 65% -2.4 0.75 ± 7% perf-profile.children.cycles-pp.exc_page_fault
3.18 ± 65% -2.4 0.74 ± 6% perf-profile.children.cycles-pp.do_user_addr_fault
3.09 ± 68% -2.4 0.68 ± 7% perf-profile.children.cycles-pp.handle_mm_fault
3.02 ± 69% -2.4 0.63 ± 8% perf-profile.children.cycles-pp.__handle_mm_fault
9.34 ± 6% -1.9 7.45 ± 8% perf-profile.children.cycles-pp.ktime_get_update_offsets_now
0.99 ± 12% -0.3 0.69 ± 24% perf-profile.children.cycles-pp.rcu_core
1.40 ± 13% -0.3 1.15 ± 7% perf-profile.children.cycles-pp.rcu_sched_clock_irq
0.46 ± 29% -0.2 0.24 ± 23% perf-profile.children.cycles-pp.check_cpu_stall
0.49 ± 16% -0.2 0.32 ± 28% perf-profile.children.cycles-pp.vm_mmap_pgoff
0.26 ± 33% -0.2 0.09 ± 57% perf-profile.children.cycles-pp._raw_spin_trylock
0.48 ± 16% -0.2 0.32 ± 27% perf-profile.children.cycles-pp.do_mmap
0.44 ± 20% -0.2 0.28 ± 20% perf-profile.children.cycles-pp.setlocale
0.36 ± 21% -0.1 0.21 ± 33% perf-profile.children.cycles-pp.ksys_mmap_pgoff
0.36 ± 19% -0.1 0.22 ± 22% perf-profile.children.cycles-pp.do_read_fault
0.19 ± 11% -0.1 0.09 ± 64% perf-profile.children.cycles-pp.kmem_cache_alloc
0.42 ± 17% -0.1 0.32 ± 24% perf-profile.children.cycles-pp.do_fault
0.28 ± 26% -0.1 0.18 ± 18% perf-profile.children.cycles-pp.avg_vruntime
0.36 ± 16% -0.1 0.27 ± 7% perf-profile.children.cycles-pp.exit_mm
0.28 ± 16% -0.1 0.19 ± 36% perf-profile.children.cycles-pp.__open64_nocancel
0.14 ± 18% -0.1 0.06 ± 30% perf-profile.children.cycles-pp.__mmap
0.12 ± 25% -0.1 0.04 ± 86% perf-profile.children.cycles-pp.__mem_cgroup_charge
0.15 ± 48% -0.1 0.08 ± 25% perf-profile.children.cycles-pp.__cond_resched
0.03 ±102% +0.1 0.12 ± 33% perf-profile.children.cycles-pp.task_mm_cid_work
0.26 ± 33% +0.1 0.36 ± 9% perf-profile.children.cycles-pp.worker_thread
0.05 ± 71% +0.1 0.15 ± 20% perf-profile.children.cycles-pp.apparmor_file_permission
0.05 ± 75% +0.1 0.16 ± 20% perf-profile.children.cycles-pp.security_file_permission
0.06 ± 74% +0.1 0.16 ± 20% perf-profile.children.cycles-pp.rw_verify_area
0.08 ± 14% +0.2 0.24 ± 30% perf-profile.children.cycles-pp.copy_mc_fragile
0.06 ± 87% +0.2 0.22 ± 32% perf-profile.children.cycles-pp.hpage_collapse_scan_pmd
0.06 ± 87% +0.2 0.22 ± 32% perf-profile.children.cycles-pp.khugepaged_scan_mm_slot
0.06 ± 92% +0.2 0.22 ± 32% perf-profile.children.cycles-pp.khugepaged
0.05 ± 91% +0.2 0.22 ± 32% perf-profile.children.cycles-pp.collapse_huge_page
0.25 ± 26% +0.2 0.45 ± 26% perf-profile.children.cycles-pp.update_rq_clock
0.28 ± 37% +0.2 0.50 ± 24% perf-profile.children.cycles-pp.runtime.epollwait.abi0
0.58 ± 15% +0.3 0.87 ± 8% perf-profile.children.cycles-pp.kthread
0.61 ± 15% +0.3 0.91 ± 6% perf-profile.children.cycles-pp.ret_from_fork_asm
0.60 ± 15% +0.3 0.90 ± 7% perf-profile.children.cycles-pp.ret_from_fork
1.22 ± 14% +0.3 1.52 ± 5% perf-profile.children.cycles-pp.sched_clock_cpu
0.20 ± 50% +0.4 0.55 ± 43% perf-profile.children.cycles-pp.copy_page_from_iter_atomic
0.71 ± 9% +0.4 1.11 ± 28% perf-profile.children.cycles-pp.hrtimer_forward
2.81 ± 6% +0.4 3.24 ± 3% perf-profile.children.cycles-pp.__update_load_avg_se
18.15 ± 4% +1.7 19.88 ± 5% perf-profile.children.cycles-pp.task_tick_fair
4.84 ± 9% +4.5 9.33 ± 56% perf-profile.children.cycles-pp.ktime_get
8.97 ± 7% -1.9 7.10 ± 8% perf-profile.self.cycles-pp.ktime_get_update_offsets_now
2.47 ± 12% -0.8 1.67 ± 41% perf-profile.self.cycles-pp.update_process_times
0.46 ± 29% -0.2 0.24 ± 23% perf-profile.self.cycles-pp.check_cpu_stall
0.26 ± 33% -0.2 0.09 ± 57% perf-profile.self.cycles-pp._raw_spin_trylock
0.02 ±142% +0.1 0.08 ± 22% perf-profile.self.cycles-pp.perf_poll
0.13 ± 28% +0.1 0.22 ± 19% perf-profile.self.cycles-pp.rebalance_domains
0.02 ±141% +0.1 0.12 ± 33% perf-profile.self.cycles-pp.task_mm_cid_work
0.08 ± 14% +0.2 0.24 ± 30% perf-profile.self.cycles-pp.copy_mc_fragile
0.70 ± 10% +0.4 1.09 ± 28% perf-profile.self.cycles-pp.hrtimer_forward
2.52 ± 9% +0.5 2.98 ± 5% perf-profile.self.cycles-pp.__update_load_avg_se
3.83 ± 12% +4.5 8.29 ± 62% perf-profile.self.cycles-pp.ktime_get
0.81 ± 9% +12.2% 0.91 ± 2% sched_debug.cfs_rq:/.h_nr_running.min
8252 ± 8% +15.1% 9494 ± 3% sched_debug.cfs_rq:/.load.min
7.77 ± 7% +14.8% 8.92 ± 3% sched_debug.cfs_rq:/.load_avg.min
0.81 ± 9% +12.2% 0.91 ± 2% sched_debug.cfs_rq:/.nr_running.min
825.73 ± 8% +12.7% 930.23 ± 3% sched_debug.cfs_rq:/.runnable_avg.min
798.93 ± 8% +12.3% 897.30 ± 4% sched_debug.cfs_rq:/.util_avg.min
0.81 ± 9% +12.2% 0.91 ± 2% sched_debug.cfs_rq:/system.slice.h_nr_running.min
8609 ± 9% +13.0% 9727 ± 3% sched_debug.cfs_rq:/system.slice.load.min
11.30 ± 7% -9.6% 10.22 sched_debug.cfs_rq:/system.slice.load_avg.avg
65.78 ± 30% -59.6% 26.56 ± 8% sched_debug.cfs_rq:/system.slice.load_avg.max
8.05 ± 8% +12.5% 9.05 ± 3% sched_debug.cfs_rq:/system.slice.load_avg.min
9.36 ± 37% -70.7% 2.75 ± 9% sched_debug.cfs_rq:/system.slice.load_avg.stddev
0.81 ± 9% +12.2% 0.91 ± 2% sched_debug.cfs_rq:/system.slice.nr_running.min
825.67 ± 8% +12.7% 930.23 ± 3% sched_debug.cfs_rq:/system.slice.runnable_avg.min
10.94 ± 42% -58.2% 4.58 ± 61% sched_debug.cfs_rq:/system.slice.se->avg.load_avg.stddev
825.67 ± 8% +12.7% 930.23 ± 3% sched_debug.cfs_rq:/system.slice.se->avg.runnable_avg.min
797.77 ± 8% +12.4% 896.95 ± 4% sched_debug.cfs_rq:/system.slice.se->avg.util_avg.min
122819 ± 40% -51.2% 59903 ± 44% sched_debug.cfs_rq:/system.slice.se->load.weight.max
25488 ± 50% -75.6% 6211 ± 42% sched_debug.cfs_rq:/system.slice.se->load.weight.stddev
8.06 ± 9% +12.4% 9.05 ± 3% sched_debug.cfs_rq:/system.slice.tg_load_avg_contrib.min
797.77 ± 8% +12.4% 896.95 ± 4% sched_debug.cfs_rq:/system.slice.util_avg.min
128.18 ± 22% +91.7% 245.75 ± 11% sched_debug.cfs_rq:/system.slice/containerd.service.avg_vruntime.avg
234.17 ± 25% +79.1% 419.30 ± 12% sched_debug.cfs_rq:/system.slice/containerd.service.avg_vruntime.max
43.50 ± 22% +120.5% 95.92 ± 28% sched_debug.cfs_rq:/system.slice/containerd.service.avg_vruntime.min
81.66 ± 29% +74.6% 142.59 ± 24% sched_debug.cfs_rq:/system.slice/containerd.service.avg_vruntime.stddev
2.80 ± 33% +81.1% 5.06 ± 33% sched_debug.cfs_rq:/system.slice/containerd.service.load_avg.avg
5.64 ± 29% +82.3% 10.28 ± 31% sched_debug.cfs_rq:/system.slice/containerd.service.load_avg.max
2.27 ± 34% +93.6% 4.40 ± 37% sched_debug.cfs_rq:/system.slice/containerd.service.load_avg.stddev
128.18 ± 22% +91.7% 245.75 ± 11% sched_debug.cfs_rq:/system.slice/containerd.service.min_vruntime.avg
234.17 ± 25% +79.1% 419.30 ± 12% sched_debug.cfs_rq:/system.slice/containerd.service.min_vruntime.max
43.50 ± 22% +120.5% 95.92 ± 28% sched_debug.cfs_rq:/system.slice/containerd.service.min_vruntime.min
81.66 ± 29% +74.6% 142.59 ± 24% sched_debug.cfs_rq:/system.slice/containerd.service.min_vruntime.stddev
2.79 ± 33% +81.1% 5.06 ± 33% sched_debug.cfs_rq:/system.slice/containerd.service.runnable_avg.avg
5.62 ± 29% +82.6% 10.27 ± 31% sched_debug.cfs_rq:/system.slice/containerd.service.runnable_avg.max
2.26 ± 33% +94.0% 4.39 ± 38% sched_debug.cfs_rq:/system.slice/containerd.service.runnable_avg.stddev
2.78 ± 33% +81.5% 5.05 ± 33% sched_debug.cfs_rq:/system.slice/containerd.service.se->avg.runnable_avg.avg
5.61 ± 29% +83.1% 10.27 ± 31% sched_debug.cfs_rq:/system.slice/containerd.service.se->avg.runnable_avg.max
2.26 ± 33% +94.6% 4.40 ± 37% sched_debug.cfs_rq:/system.slice/containerd.service.se->avg.runnable_avg.stddev
136.98 ± 21% +94.7% 266.75 ± 14% sched_debug.cfs_rq:/system.slice/containerd.service.se->sum_exec_runtime.avg
252.42 ± 24% +82.7% 461.23 ± 16% sched_debug.cfs_rq:/system.slice/containerd.service.se->sum_exec_runtime.max
45.53 ± 21% +119.8% 100.06 ± 28% sched_debug.cfs_rq:/system.slice/containerd.service.se->sum_exec_runtime.min
89.12 ± 27% +79.0% 159.55 ± 28% sched_debug.cfs_rq:/system.slice/containerd.service.se->sum_exec_runtime.stddev
8.81 ± 37% +4295.1% 387.34 ±118% sched_debug.cfs_rq:/system.slice/containerd.service.tg_load_avg.avg
9.12 ± 36% +4150.1% 387.69 ±118% sched_debug.cfs_rq:/system.slice/containerd.service.tg_load_avg.max
8.51 ± 37% +4444.4% 386.92 ±118% sched_debug.cfs_rq:/system.slice/containerd.service.tg_load_avg.min
2.80 ± 33% +81.1% 5.06 ± 33% sched_debug.cfs_rq:/system.slice/containerd.service.tg_load_avg_contrib.avg
5.64 ± 29% +82.3% 10.28 ± 31% sched_debug.cfs_rq:/system.slice/containerd.service.tg_load_avg_contrib.max
2.27 ± 34% +93.6% 4.40 ± 37% sched_debug.cfs_rq:/system.slice/containerd.service.tg_load_avg_contrib.stddev
0.17 ± 8% -60.1% 0.07 ±122% sched_debug.cfs_rq:/system.slice/lkp-bootstrap.service.h_nr_running.stddev
180165 ± 8% -60.0% 72004 ±122% sched_debug.cfs_rq:/system.slice/lkp-bootstrap.service.load.stddev
187.40 ± 15% -67.8% 60.33 ±122% sched_debug.cfs_rq:/system.slice/lkp-bootstrap.service.load_avg.stddev
0.17 ± 8% -60.1% 0.07 ±122% sched_debug.cfs_rq:/system.slice/lkp-bootstrap.service.nr_running.stddev
6.76 ± 69% -76.3% 1.60 ±129% sched_debug.cfs_rq:/system.slice/lkp-bootstrap.service.removed.load_avg.avg
147.10 ± 62% -79.3% 30.38 ±122% sched_debug.cfs_rq:/system.slice/lkp-bootstrap.service.removed.load_avg.max
28.32 ± 61% -76.8% 6.58 ±124% sched_debug.cfs_rq:/system.slice/lkp-bootstrap.service.removed.load_avg.stddev
2.89 ± 77% -75.8% 0.70 ±124% sched_debug.cfs_rq:/system.slice/lkp-bootstrap.service.removed.runnable_avg.avg
57.35 ± 50% -73.5% 15.17 ±122% sched_debug.cfs_rq:/system.slice/lkp-bootstrap.service.removed.runnable_avg.max
11.43 ± 64% -73.8% 2.99 ±122% sched_debug.cfs_rq:/system.slice/lkp-bootstrap.service.removed.runnable_avg.stddev
2.87 ± 77% -75.7% 0.70 ±124% sched_debug.cfs_rq:/system.slice/lkp-bootstrap.service.removed.util_avg.avg
55.94 ± 49% -72.9% 15.17 ±122% sched_debug.cfs_rq:/system.slice/lkp-bootstrap.service.removed.util_avg.max
11.28 ± 64% -73.5% 2.99 ±122% sched_debug.cfs_rq:/system.slice/lkp-bootstrap.service.removed.util_avg.stddev
69.45 ± 22% -63.6% 25.26 ±122% sched_debug.cfs_rq:/system.slice/lkp-bootstrap.service.runnable_avg.stddev
11.31 ± 7% -64.9% 3.97 ±122% sched_debug.cfs_rq:/system.slice/lkp-bootstrap.service.se->avg.load_avg.avg
60.48 ± 40% -83.8% 9.82 ±122% sched_debug.cfs_rq:/system.slice/lkp-bootstrap.service.se->avg.load_avg.max
9.81 ± 46% -89.7% 1.01 ±122% sched_debug.cfs_rq:/system.slice/lkp-bootstrap.service.se->avg.load_avg.stddev
69.41 ± 22% -63.6% 25.24 ±122% sched_debug.cfs_rq:/system.slice/lkp-bootstrap.service.se->avg.runnable_avg.stddev
43.67 ± 37% -73.2% 11.72 ±126% sched_debug.cfs_rq:/system.slice/lkp-bootstrap.service.se->avg.util_avg.stddev
25759 ± 39% -80.9% 4908 ±122% sched_debug.cfs_rq:/system.slice/lkp-bootstrap.service.se->load.weight.avg
102176 ± 47% -84.0% 16350 ±125% sched_debug.cfs_rq:/system.slice/lkp-bootstrap.service.se->load.weight.max
23115 ± 57% -92.6% 1718 ±126% sched_debug.cfs_rq:/system.slice/lkp-bootstrap.service.se->load.weight.stddev
850.39 ± 47% -74.6% 216.04 ±123% sched_debug.cfs_rq:/system.slice/lkp-bootstrap.service.tg_load_avg.stddev
188.15 ± 15% -68.1% 60.09 ±122% sched_debug.cfs_rq:/system.slice/lkp-bootstrap.service.tg_load_avg_contrib.stddev
43.66 ± 37% -73.2% 11.72 ±126% sched_debug.cfs_rq:/system.slice/lkp-bootstrap.service.util_avg.stddev
835849 +17.1% 979195 sched_debug.cpu.avg_idle.avg
390705 ± 8% +27.5% 498342 ± 8% sched_debug.cpu.avg_idle.min
112505 ± 5% -37.1% 70769 ± 6% sched_debug.cpu.avg_idle.stddev
13678 ± 52% -34.9% 8908 sched_debug.cpu.curr->pid.max
839.26 ± 43% -46.1% 452.62 ± 21% sched_debug.cpu.curr->pid.stddev
0.81 ± 9% +12.2% 0.91 ± 2% sched_debug.cpu.nr_running.min
15.22 ± 26% +51.7% 23.09 ± 11% sched_debug.cpu.nr_uninterruptible.max
-17.17 +58.4% -27.19 sched_debug.cpu.nr_uninterruptible.min
5.71 ± 16% +31.9% 7.54 ± 10% sched_debug.cpu.nr_uninterruptible.stddev
Disclaimer:
Results have been estimated based on internal Intel analysis and are provided
for informational purposes only. Any difference in system hardware or software
design or configuration may affect actual performance.
--
0-DAY CI Kernel Test Service
https://github.com/intel/lkp-tests/wiki
Powered by blists - more mailing lists