lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [<thread-prev] [day] [month] [year] [list]
Message-ID: <72e017b9-deb6-44fa-91d6-716ee2c39cbc@intel.com>
Date:   Wed, 13 Sep 2023 21:09:56 +0800
From:   "Yin, Fengwei" <fengwei.yin@...el.com>
To:     kernel test robot <oliver.sang@...el.com>,
        Matthew Wilcox <willy@...radead.org>
CC:     <oe-lkp@...ts.linux.dev>, <lkp@...el.com>,
        <linux-kernel@...r.kernel.org>,
        Andrew Morton <akpm@...ux-foundation.org>,
        Matthew Wilcox <willy@...radead.org>,
        <linux-fsdevel@...r.kernel.org>, <ying.huang@...el.com>,
        <feng.tang@...el.com>
Subject: Re: [linus:master] [filemap] de74976eb6:
 stress-ng.bad-altstack.ops_per_sec -25.2% regression

Add Matthew and I'd like to hear his comments here.

On 9/11/2023 4:27 PM, kernel test robot wrote:
> 
> 
> Hello,
> 
> kernel test robot noticed a -25.2% regression of stress-ng.bad-altstack.ops_per_sec on:
> 
> 
> commit: de74976eb65151a2f568e477fc2e0032df5b22b4 ("filemap: add filemap_map_folio_range()")
> https://git.kernel.org/cgit/linux/kernel/git/torvalds/linux.git master
> 
> testcase: stress-ng
> test machine: 64 threads 2 sockets Intel(R) Xeon(R) Gold 6346 CPU @ 3.10GHz (Ice Lake) with 256G memory
> parameters:
> 
> 	nr_threads: 100%
> 	testtime: 60s
> 	class: memory
> 	test: bad-altstack
> 	cpufreq_governor: performance
> 
> 
> In addition to that, the commit also has significant impact on the following tests:
> 
> +------------------+---------------------------------------------------------------------------------------------+
> | testcase: change | stress-ng: stress-ng.fork.ops_per_sec -13.3% regression                                     |
> | test machine     | 64 threads 2 sockets Intel(R) Xeon(R) Gold 6346 CPU @ 3.10GHz (Ice Lake) with 256G memory   |
> | test parameters  | class=pts                                                                                   |
> |                  | cpufreq_governor=performance                                                                |
> |                  | nr_threads=100%                                                                             |
> |                  | test=fork                                                                                   |
> |                  | testtime=60s                                                                                |
> +------------------+---------------------------------------------------------------------------------------------+
> | testcase: change | vm-scalability: vm-scalability.throughput -11.1% regression                                 |
> | test machine     | 224 threads 2 sockets Intel(R) Xeon(R) Platinum 8480L (Sapphire Rapids) with 512G memory    |
> | test parameters  | cpufreq_governor=performance                                                                |
> |                  | runtime=300s                                                                                |
> |                  | test=mmap-pread-rand                                                                        |
> +------------------+---------------------------------------------------------------------------------------------+
> | testcase: change | stress-ng: stress-ng.zombie.ops_per_sec -74.5% regression                                   |
> | test machine     | 224 threads 2 sockets Intel(R) Xeon(R) Platinum 8480CTDX (Sapphire Rapids) with 256G memory |
> | test parameters  | class=scheduler                                                                             |
> |                  | cpufreq_governor=performance                                                                |
> |                  | nr_threads=100%                                                                             |
> |                  | sc_pid_max=4194304                                                                          |
> |                  | test=zombie                                                                                 |
> |                  | testtime=60s                                                                                |
> +------------------+---------------------------------------------------------------------------------------------+
I saw the stress-ng.zombie has very unstable result. I expect around 20% regression.

After several round of profiling and testing, my understanding is fixed to the filemap_map_folio_range()
is not suitable for order-0 folio.

So I made following change to split the order-0 folio handling out instread of leave to
filemap_map_folio_range(). And all the regressions here are gone (commit 2334949153 is with the patch below):


9f1f5b60e76d44fa de74976eb65151a2f568e477fc2 2334949153c5e312b6f6c1f4852
---------------- --------------------------- ---------------------------
   3685088 ±  2%     -46.7%    1964990 ± 49%      -0.9%    3651955        stress-ng.zombie.ops
     56732 ±  2%     -49.7%      28544 ± 61%      -0.9%      56216        stress-ng.zombie.ops_per_sec

   3843810           -25.5%    2862662            +1.0%    3884164        stress-ng.bad-altstack.ops
     64061           -25.5%      47709            +1.1%      64734        stress-ng.bad-altstack.ops_per_sec

   1709026           -14.3%    1464872 ±  2%      +0.3%    1713950        stress-ng.fork.ops
     28483           -14.3%      24414 ±  2%      +0.3%      28565        stress-ng.fork.ops_per_sec

     61874           -13.2%      53731            +0.7%      62318        vm-scalability.median
      1210           -26.1        1184 ± 13%     +70.9        1281 ±  3%  vm-scalability.stddev%
  13527663           -12.8%   11796658            +0.5%   13595987        vm-scalability.throughput


Hi Matthew,
I am not sure whether it's worthy to add filemap_map_one_folio() to handle order-0 folio here.
But it does fix all regressions LKP reported.

I also compared the code generated for these two functions:
  - The code size of filemap_map_folio_range() is 126 bytes larger than filemap_map_one_folio()
  - There is loop in filemap_map_folio_range()


The changes of 2334949153c5e312b6f6c1f4852 is as following:

diff --git a/mm/filemap.c b/mm/filemap.c
index bdc1e0b811bf..ae30319eb005 100644
--- a/mm/filemap.c
+++ b/mm/filemap.c
@@ -3473,21 +3473,19 @@ static struct folio *next_uptodate_folio(struct xa_state *xas,
  */
 static vm_fault_t filemap_map_folio_range(struct vm_fault *vmf,
 			struct folio *folio, unsigned long start,
-			unsigned long addr, unsigned int nr_pages)
+			unsigned long addr, unsigned int nr_pages,
+			unsigned int *mmap_miss)
 {
 	vm_fault_t ret = 0;
 	struct vm_area_struct *vma = vmf->vma;
-	struct file *file = vma->vm_file;
 	struct page *page = folio_page(folio, start);
-	unsigned int mmap_miss = READ_ONCE(file->f_ra.mmap_miss);
 	unsigned int ref_count = 0, count = 0;
 
 	do {
 		if (PageHWPoison(page))
 			continue;
 
-		if (mmap_miss > 0)
-			mmap_miss--;
+		(*mmap_miss)++;
 
 		/*
 		 * NOTE: If there're PTE markers, we'll leave them to be
@@ -3509,7 +3507,36 @@ static vm_fault_t filemap_map_folio_range(struct vm_fault *vmf,
 	vmf->pte -= nr_pages;
 
 	folio_ref_add(folio, ref_count);
-	WRITE_ONCE(file->f_ra.mmap_miss, mmap_miss);
+
+	return ret;
+}
+
+static vm_fault_t filemap_map_one_folio(struct vm_fault *vmf,
+			struct folio *folio, unsigned long addr,
+			unsigned int *mmap_miss)
+{
+	vm_fault_t ret = 0;
+	struct vm_area_struct *vma = vmf->vma;
+
+	if (PageHWPoison(&folio->page))
+		return ret;
+
+	(*mmap_miss)++;
+
+	/*
+	 * NOTE: If there're PTE markers, we'll leave them to be
+	 * handled in the specific fault path, and it'll prohibit
+	 * the fault-around logic.
+	 */
+	if (!pte_none(ptep_get(vmf->pte)))
+		return ret;
+
+	if (vmf->address == addr)
+		ret = VM_FAULT_NOPAGE;
+
+	do_set_pte(vmf, &folio->page, addr);
+	update_mmu_cache(vma, addr, vmf->pte);
+	folio_ref_inc(folio);
 
 	return ret;
 }
@@ -3525,7 +3552,7 @@ vm_fault_t filemap_map_pages(struct vm_fault *vmf,
 	XA_STATE(xas, &mapping->i_pages, start_pgoff);
 	struct folio *folio;
 	vm_fault_t ret = 0;
-	int nr_pages = 0;
+	unsigned int mmap_miss = 0, mmap_miss_saved, nr_pages = 0;
 
 	rcu_read_lock();
 	folio = next_uptodate_folio(&xas, mapping, end_pgoff);
@@ -3553,25 +3580,28 @@ vm_fault_t filemap_map_pages(struct vm_fault *vmf,
 		end = folio->index + folio_nr_pages(folio) - 1;
 		nr_pages = min(end, end_pgoff) - xas.xa_index + 1;
 
-		/*
-		 * NOTE: If there're PTE markers, we'll leave them to be
-		 * handled in the specific fault path, and it'll prohibit the
-		 * fault-around logic.
-		 */
-		if (!pte_none(ptep_get(vmf->pte)))
-			goto unlock;
-
-		ret |= filemap_map_folio_range(vmf, folio,
-				xas.xa_index - folio->index, addr, nr_pages);
+		if (!folio_test_large(folio)) {
+			ret |= filemap_map_one_folio(vmf, folio,
+					addr, &mmap_miss);
+		} else {
+			ret |= filemap_map_folio_range(vmf, folio,
+					xas.xa_index - folio->index, addr,
+					nr_pages, &mmap_miss);
+		}
 
-unlock:
 		folio_unlock(folio);
 		folio_put(folio);
-		folio = next_uptodate_folio(&xas, mapping, end_pgoff);
-	} while (folio);
+	} while ((folio = next_uptodate_folio(&xas, mapping, end_pgoff)) != NULL);
 	pte_unmap_unlock(vmf->pte, vmf->ptl);
 out:
 	rcu_read_unlock();
+
+	mmap_miss_saved = READ_ONCE(file->f_ra.mmap_miss);
+	if (mmap_miss >= mmap_miss_saved)
+		WRITE_ONCE(file->f_ra.mmap_miss, 0);
+	else
+		WRITE_ONCE(file->f_ra.mmap_miss, mmap_miss_saved - mmap_miss);
+
 	return ret;
 }
 EXPORT_SYMBOL(filemap_map_pages);



Regards
Yin, Fengwei


> 
> 
> If you fix the issue in a separate patch/commit (i.e. not just a new version of
> the same patch/commit), kindly add following tags
> | Reported-by: kernel test robot <oliver.sang@...el.com>
> | Closes: https://lore.kernel.org/oe-lkp/202309111556.b2aa3d7a-oliver.sang@intel.com
> 
> 
> Details are as below:
> -------------------------------------------------------------------------------------------------->
> 
> 
> The kernel config and materials to reproduce are available at:
> https://download.01.org/0day-ci/archive/20230911/202309111556.b2aa3d7a-oliver.sang@intel.com
> 
> =========================================================================================
> class/compiler/cpufreq_governor/kconfig/nr_threads/rootfs/tbox_group/test/testcase/testtime:
>   memory/gcc-12/performance/x86_64-rhel-8.3/100%/debian-11.1-x86_64-20220510.cgz/lkp-icl-2sp8/bad-altstack/stress-ng/60s
> 
> commit: 
>   9f1f5b60e7 ("mm: use flush_icache_pages() in do_set_pmd()")
>   de74976eb6 ("filemap: add filemap_map_folio_range()")
> 
> 9f1f5b60e76d44fa de74976eb65151a2f568e477fc2 
> ---------------- --------------------------- 
>          %stddev     %change         %stddev
>              \          |                \  
>    5005804           -25.4%    3732280 ±  4%  cpuidle..usage
>     226369           -22.8%     174812 ±  5%  vmstat.system.cs
>     196009           -10.2%     176042 ±  3%  vmstat.system.in
>       1.35            -0.1        1.22 ±  2%  mpstat.cpu.all.irq%
>       0.67            -0.1        0.53 ±  5%  mpstat.cpu.all.soft%
>       3.69 ±  3%      -0.5        3.21 ±  5%  mpstat.cpu.all.usr%
>    4245150 ±  6%     -25.8%    3149301 ± 29%  numa-meminfo.node1.AnonPages
>    4945434 ±  4%     -26.2%    3647859 ± 26%  numa-meminfo.node1.Inactive
>    4945301 ±  4%     -26.2%    3647723 ± 26%  numa-meminfo.node1.Inactive(anon)
>   73394698 ±  2%     -28.1%   52759723 ±  9%  numa-numastat.node0.local_node
>   73440754 ±  2%     -28.1%   52810446 ±  9%  numa-numastat.node0.numa_hit
>   72370904 ±  2%     -29.3%   51147333 ±  9%  numa-numastat.node1.local_node
>   72404527 ±  2%     -29.3%   51170297 ±  9%  numa-numastat.node1.numa_hit
>    4842492 ±  2%     -14.5%    4141547 ±  4%  meminfo.AnonPages
>    7390687           -11.8%    6519765 ±  3%  meminfo.Committed_AS
>    5853448 ±  2%     -15.7%    4935195 ±  4%  meminfo.Inactive
>    5853245 ±  2%     -15.7%    4934992 ±  4%  meminfo.Inactive(anon)
>    1150839           -19.3%     928885 ±  3%  meminfo.Shmem
>    4978080           -25.8%    3694247 ±  4%  turbostat.C1
>       0.15           -24.4%       0.11 ±  4%  turbostat.IPC
>      25.82 ± 23%     -24.6        1.24 ± 94%  turbostat.PKG_%
>       9925          +106.7%      20517 ± 26%  turbostat.POLL
>     398.49            -4.0%     382.58        turbostat.PkgWatt
>     101.71           -12.3%      89.24        turbostat.RAMWatt
>   73443008 ±  2%     -28.1%   52812536 ±  9%  numa-vmstat.node0.numa_hit
>   73396953 ±  2%     -28.1%   52761813 ±  9%  numa-vmstat.node0.numa_local
>    1061960 ±  6%     -25.8%     787891 ± 29%  numa-vmstat.node1.nr_anon_pages
>    1237001 ±  4%     -26.2%     912544 ± 26%  numa-vmstat.node1.nr_inactive_anon
>    1236999 ±  4%     -26.2%     912542 ± 26%  numa-vmstat.node1.nr_zone_inactive_anon
>   72405504 ±  2%     -29.3%   51172088 ±  9%  numa-vmstat.node1.numa_hit
>   72371881 ±  2%     -29.3%   51149123 ±  9%  numa-vmstat.node1.numa_local
>    3878334           -25.2%    2899973 ±  3%  stress-ng.bad-altstack.ops
>      64637           -25.2%      48331 ±  3%  stress-ng.bad-altstack.ops_per_sec
>    1020183            -9.3%     924917 ±  3%  stress-ng.time.involuntary_context_switches
>      31629 ± 11%    +686.1%     248632 ± 19%  stress-ng.time.major_page_faults
>  1.775e+08 ±  2%     -23.0%  1.366e+08 ±  3%  stress-ng.time.minor_page_faults
>       5351            +1.0%       5407        stress-ng.time.percent_of_cpu_this_job_got
>       2175           +14.1%       2481 ±  2%  stress-ng.time.system_time
>       1152           -23.9%     877.35 ±  7%  stress-ng.time.user_time
>    6559992           -20.8%    5193194 ±  2%  stress-ng.time.voluntary_context_switches
>       4.50 ± 16%     -68.5%       1.42 ± 59%  sched_debug.cfs_rq:/.load_avg.min
>     610.20 ±  3%     -11.2%     541.92 ±  4%  sched_debug.cfs_rq:/.runnable_avg.avg
>     590.68 ±  3%     -11.0%     525.71 ±  4%  sched_debug.cfs_rq:/.util_avg.avg
>       1432 ±  7%     -15.5%       1210 ± 11%  sched_debug.cfs_rq:/.util_avg.max
>     118.53 ± 23%     -49.5%      59.82 ± 18%  sched_debug.cfs_rq:/.util_est_enqueued.avg
>     175.92 ± 13%     -30.6%     122.01 ± 10%  sched_debug.cfs_rq:/.util_est_enqueued.stddev
>     402951 ±  5%     +20.2%     484486 ±  8%  sched_debug.cpu.avg_idle.avg
>    1848476           -25.3%    1381292 ±  3%  sched_debug.cpu.curr->pid.max
>     902123 ±  4%     -25.6%     671441 ±  4%  sched_debug.cpu.curr->pid.stddev
>     113163           -22.0%      88310 ±  4%  sched_debug.cpu.nr_switches.avg
>    1211031 ±  2%     -14.5%    1036001 ±  4%  proc-vmstat.nr_anon_pages
>     973846            -5.7%     918330        proc-vmstat.nr_file_pages
>    1463754 ±  2%     -15.7%    1234145 ±  4%  proc-vmstat.nr_inactive_anon
>       6642 ±  2%      -6.4%       6218        proc-vmstat.nr_page_table_pages
>     287748           -19.3%     232230 ±  3%  proc-vmstat.nr_shmem
>    1463754 ±  2%     -15.7%    1234145 ±  4%  proc-vmstat.nr_zone_inactive_anon
>  1.458e+08 ±  2%     -28.7%   1.04e+08 ±  4%  proc-vmstat.numa_hit
>  1.458e+08 ±  2%     -28.7%  1.039e+08 ±  4%  proc-vmstat.numa_local
>  1.538e+08 ±  2%     -28.6%  1.097e+08 ±  4%  proc-vmstat.pgalloc_normal
>  1.848e+08           -22.3%  1.436e+08 ±  3%  proc-vmstat.pgfault
>  1.488e+08 ±  2%     -28.3%  1.067e+08 ±  4%  proc-vmstat.pgfree
>    7485052           -24.2%    5673016 ±  3%  proc-vmstat.pgreuse
>      17.90            -6.9%      16.67        perf-stat.i.MPKI
>  1.884e+10           -27.2%  1.372e+10 ±  4%  perf-stat.i.branch-instructions
>       0.91            -0.0        0.87        perf-stat.i.branch-miss-rate%
>  1.627e+08           -30.5%   1.13e+08 ±  5%  perf-stat.i.branch-misses
>      27.37            -1.1       26.26        perf-stat.i.cache-miss-rate%
>  4.766e+08 ±  2%     -34.5%  3.123e+08 ±  6%  perf-stat.i.cache-misses
>  1.692e+09           -31.4%  1.161e+09 ±  5%  perf-stat.i.cache-references
>     241702           -23.6%     184568 ±  5%  perf-stat.i.context-switches
>       2.19           +35.9%       2.97 ±  2%  perf-stat.i.cpi
>      65598           -32.1%      44526 ±  7%  perf-stat.i.cpu-migrations
>     582.31 ±  5%     +33.0%     774.35 ±  3%  perf-stat.i.cycles-between-cache-misses
>   34494957 ±  6%     -21.2%   27167191 ±  9%  perf-stat.i.dTLB-load-misses
>  2.329e+10           -25.8%  1.729e+10 ±  4%  perf-stat.i.dTLB-loads
>   21905610 ±  4%     -22.9%   16890155 ±  8%  perf-stat.i.dTLB-store-misses
>  1.189e+10           -23.6%  9.089e+09 ±  4%  perf-stat.i.dTLB-stores
>  9.182e+10           -26.3%  6.764e+10 ±  4%  perf-stat.i.instructions
>       0.47 ±  2%     -23.1%       0.36 ±  3%  perf-stat.i.ipc
>     514.31 ± 12%    +717.4%       4204 ± 20%  perf-stat.i.major-faults
>       1254 ± 12%     +63.7%       2052 ±  3%  perf-stat.i.metric.K/sec
>     872.91           -26.1%     645.30 ±  4%  perf-stat.i.metric.M/sec
>    2910945 ±  2%     -23.2%    2235593 ±  4%  perf-stat.i.minor-faults
>      92.03            -0.8       91.22        perf-stat.i.node-load-miss-rate%
>    1.2e+08 ±  2%     -29.7%   84436459 ±  6%  perf-stat.i.node-load-misses
>    8939523 ±  5%     -19.6%    7184335 ±  7%  perf-stat.i.node-loads
>      55.38            -7.1       48.31        perf-stat.i.node-store-miss-rate%
>   76680149           -31.5%   52522811 ±  5%  perf-stat.i.node-store-misses
>   59789508            -9.3%   54216942 ±  6%  perf-stat.i.node-stores
>    3063984 ±  2%     -23.2%    2352460 ±  4%  perf-stat.i.page-faults
>      18.25            -6.4%      17.08        perf-stat.overall.MPKI
>       0.85            -0.0        0.81        perf-stat.overall.branch-miss-rate%
>      28.13            -1.2       26.92        perf-stat.overall.cache-miss-rate%
>       2.21           +34.6%       2.97 ±  2%  perf-stat.overall.cpi
>     430.75 ±  2%     +50.4%     647.65 ±  4%  perf-stat.overall.cycles-between-cache-misses
>       0.45           -25.6%       0.34 ±  2%  perf-stat.overall.ipc
>      56.77            -6.8       49.96        perf-stat.overall.node-store-miss-rate%
>  1.835e+10           -26.0%  1.357e+10 ±  3%  perf-stat.ps.branch-instructions
>  1.562e+08           -29.2%  1.105e+08 ±  5%  perf-stat.ps.branch-misses
>  4.587e+08 ±  2%     -32.9%  3.076e+08 ±  5%  perf-stat.ps.cache-misses
>  1.631e+09           -29.9%  1.142e+09 ±  4%  perf-stat.ps.cache-references
>     230601           -22.4%     179055 ±  4%  perf-stat.ps.context-switches
>      62819           -32.0%      42714 ±  7%  perf-stat.ps.cpu-migrations
>   34685174 ±  5%     -20.9%   27444426 ±  9%  perf-stat.ps.dTLB-load-misses
>  2.269e+10           -24.6%  1.711e+10 ±  3%  perf-stat.ps.dTLB-loads
>   21412527 ±  4%     -21.8%   16743593 ±  8%  perf-stat.ps.dTLB-store-misses
>   1.16e+10           -22.4%  8.999e+09 ±  3%  perf-stat.ps.dTLB-stores
>  8.935e+10           -25.2%  6.687e+10 ±  3%  perf-stat.ps.instructions
>     494.42 ± 12%    +691.7%       3914 ± 20%  perf-stat.ps.major-faults
>    2841237 ±  2%     -22.1%    2214148 ±  3%  perf-stat.ps.minor-faults
>  1.158e+08 ±  2%     -28.5%   82760390 ±  5%  perf-stat.ps.node-load-misses
>    9347688 ±  6%     -19.6%    7516467 ±  9%  perf-stat.ps.node-loads
>   75112388           -30.5%   52227314 ±  5%  perf-stat.ps.node-store-misses
>   57203372            -8.5%   52340890 ±  5%  perf-stat.ps.node-stores
>    2990035 ±  2%     -22.1%    2329223 ±  3%  perf-stat.ps.page-faults
>  5.556e+12 ±  2%     -24.1%  4.214e+12 ±  2%  perf-stat.total.instructions
>       0.03 ± 10%     -34.1%       0.02 ± 15%  perf-sched.sch_delay.avg.ms.__cond_resched.__alloc_pages.pte_alloc_one.do_read_fault.do_fault
>       0.03 ± 11%     -30.4%       0.02 ± 31%  perf-sched.sch_delay.avg.ms.__cond_resched.__kmem_cache_alloc_node.kmalloc_trace.single_open.do_dentry_open
>       0.02 ±  2%     -26.6%       0.02 ±  8%  perf-sched.sch_delay.avg.ms.__cond_resched.__put_anon_vma.unlink_anon_vmas.free_pgtables.exit_mmap
>       0.02 ±  3%     -26.9%       0.02 ± 11%  perf-sched.sch_delay.avg.ms.__cond_resched.down_write.free_pgtables.exit_mmap.__mmput
>       0.03 ±  5%     -29.2%       0.02 ± 16%  perf-sched.sch_delay.avg.ms.__cond_resched.down_write.unlink_anon_vmas.free_pgtables.exit_mmap
>       0.02 ±  5%     -22.9%       0.02 ± 12%  perf-sched.sch_delay.avg.ms.__cond_resched.down_write.unlink_file_vma.free_pgtables.exit_mmap
>       0.03 ± 21%     -35.3%       0.02 ± 23%  perf-sched.sch_delay.avg.ms.__cond_resched.dput.path_put.exit_fs.do_exit
>       0.03 ±  7%     -35.0%       0.02 ± 23%  perf-sched.sch_delay.avg.ms.__cond_resched.dput.step_into.open_last_lookups.path_openat
>       0.03 ± 15%     -26.3%       0.02 ± 15%  perf-sched.sch_delay.avg.ms.__cond_resched.dput.terminate_walk.path_openat.do_filp_open
>       0.03 ±  2%     -32.4%       0.02 ± 13%  perf-sched.sch_delay.avg.ms.__cond_resched.exit_mmap.__mmput.exit_mm.do_exit
>       0.03 ± 11%     -28.8%       0.02 ± 10%  perf-sched.sch_delay.avg.ms.__cond_resched.kmem_cache_alloc.alloc_empty_file.path_openat.do_filp_open
>       0.02 ±  5%     -13.3%       0.02 ±  2%  perf-sched.sch_delay.avg.ms.__cond_resched.kmem_cache_alloc.anon_vma_fork.dup_mmap.dup_mm
>       0.02 ± 17%     -52.2%       0.01 ± 13%  perf-sched.sch_delay.avg.ms.__cond_resched.kmem_cache_alloc.dup_mm.constprop.0
>       0.04 ± 30%     -49.8%       0.02 ± 17%  perf-sched.sch_delay.avg.ms.__cond_resched.kmem_cache_alloc.security_file_alloc.init_file.alloc_empty_file
>       0.02 ±105%     -89.8%       0.00 ±142%  perf-sched.sch_delay.avg.ms.__cond_resched.kmem_cache_alloc_node.alloc_vmap_area.__get_vm_area_node.__vmalloc_node_range
>       0.04 ±  4%     -26.8%       0.03 ±  7%  perf-sched.sch_delay.avg.ms.__cond_resched.lock_mm_and_find_vma.do_user_addr_fault.exc_page_fault.asm_exc_page_fault
>       0.03 ± 16%     -38.5%       0.02 ± 33%  perf-sched.sch_delay.avg.ms.__cond_resched.mnt_want_write.open_last_lookups.path_openat.do_filp_open
>       0.05 ± 86%     -79.8%       0.01 ± 95%  perf-sched.sch_delay.avg.ms.__cond_resched.mutex_lock.__fdget_pos.ksys_write.do_syscall_64
>       0.03 ± 14%     -31.2%       0.02 ± 25%  perf-sched.sch_delay.avg.ms.__cond_resched.mutex_lock.futex_exit_release.exit_mm_release.exit_mm
>       0.03 ±  7%     -24.5%       0.02 ± 13%  perf-sched.sch_delay.avg.ms.__cond_resched.put_files_struct.do_exit.do_group_exit.get_signal
>       0.03 ±  7%     -40.7%       0.02 ± 26%  perf-sched.sch_delay.avg.ms.__cond_resched.remove_vma.exit_mmap.__mmput.exit_mm
>       0.04 ± 24%     -46.2%       0.02 ± 27%  perf-sched.sch_delay.avg.ms.__cond_resched.task_work_run.exit_to_user_mode_loop.exit_to_user_mode_prepare.syscall_exit_to_user_mode
>       0.03 ±  2%     -32.6%       0.02 ± 12%  perf-sched.sch_delay.avg.ms.__cond_resched.tlb_batch_pages_flush.tlb_finish_mmu.exit_mmap.__mmput
>       0.03 ±  2%     -33.1%       0.02 ± 11%  perf-sched.sch_delay.avg.ms.__cond_resched.zap_pmd_range.isra.0.unmap_page_range
>       0.03           -46.2%       0.01 ± 11%  perf-sched.sch_delay.avg.ms.do_task_dead.do_exit.do_group_exit.get_signal.arch_do_signal_or_restart
>       0.02           -11.1%       0.02 ±  3%  perf-sched.sch_delay.avg.ms.do_wait.kernel_wait4.__do_sys_wait4.do_syscall_64
>       0.03           -24.7%       0.02 ±  8%  perf-sched.sch_delay.avg.ms.exit_to_user_mode_loop.exit_to_user_mode_prepare.irqentry_exit_to_user_mode.asm_exc_page_fault
>       0.04 ±  8%     -45.0%       0.02 ± 15%  perf-sched.sch_delay.avg.ms.exit_to_user_mode_loop.exit_to_user_mode_prepare.irqentry_exit_to_user_mode.asm_sysvec_apic_timer_interrupt
>       0.02 ±  4%     -27.0%       0.02 ± 10%  perf-sched.sch_delay.avg.ms.exit_to_user_mode_loop.exit_to_user_mode_prepare.syscall_exit_to_user_mode.do_syscall_64
>       0.03 ±  5%     -41.6%       0.02 ± 17%  perf-sched.sch_delay.avg.ms.io_schedule.folio_wait_bit_common.__do_fault.do_read_fault
>       0.04 ±  9%     -47.1%       0.02 ± 12%  perf-sched.sch_delay.avg.ms.io_schedule.folio_wait_bit_common.filemap_fault.__do_fault
>       0.69 ± 10%     +39.0%       0.96 ± 10%  perf-sched.sch_delay.max.ms.__cond_resched.__put_anon_vma.unlink_anon_vmas.free_pgtables.exit_mmap
>       1.30 ±108%     -85.6%       0.19 ± 19%  perf-sched.sch_delay.max.ms.__cond_resched.down_write.anon_vma_clone.anon_vma_fork.dup_mmap
>       0.32 ± 18%     -92.3%       0.02 ± 19%  perf-sched.sch_delay.max.ms.__cond_resched.kmem_cache_alloc.dup_mm.constprop.0
>       1.66 ±101%     -82.8%       0.29 ± 24%  perf-sched.sch_delay.max.ms.__cond_resched.kmem_cache_alloc.vm_area_dup.dup_mmap.dup_mm
>       0.02 ± 94%     -90.6%       0.00 ±142%  perf-sched.sch_delay.max.ms.__cond_resched.kmem_cache_alloc_node.alloc_vmap_area.__get_vm_area_node.__vmalloc_node_range
>       0.55 ± 12%     -16.8%       0.46 ± 12%  perf-sched.sch_delay.max.ms.__cond_resched.mutex_lock.futex_exit_release.exit_mm_release.exit_mm
>       0.14 ± 51%     -58.5%       0.06 ± 71%  perf-sched.sch_delay.max.ms.__cond_resched.mutex_lock.ldt_dup_context.dup_mmap.dup_mm
>       0.41 ± 19%     -42.8%       0.24 ± 59%  perf-sched.sch_delay.max.ms.__cond_resched.mutex_lock.perf_event_exit_task.do_exit.do_group_exit
>       0.56 ± 14%     -41.6%       0.32 ± 29%  perf-sched.sch_delay.max.ms.__cond_resched.task_work_run.exit_to_user_mode_loop.exit_to_user_mode_prepare.syscall_exit_to_user_mode
>       0.25 ± 24%     -69.9%       0.07 ± 66%  perf-sched.sch_delay.max.ms.__cond_resched.uprobe_start_dup_mmap.dup_mmap.dup_mm.constprop
>       3.72 ± 21%     -41.8%       2.16 ± 28%  perf-sched.sch_delay.max.ms.do_task_dead.do_exit.do_group_exit.get_signal.arch_do_signal_or_restart
>       0.62 ±  8%     +77.7%       1.11 ±  9%  perf-sched.sch_delay.max.ms.io_schedule.folio_wait_bit_common.filemap_fault.__do_fault
>       0.15 ± 64%     -70.6%       0.04 ± 56%  perf-sched.sch_delay.max.ms.schedule_preempt_disabled.rwsem_down_write_slowpath.down_write.dup_mmap
>       0.45 ± 13%     -30.1%       0.32 ± 21%  perf-sched.sch_delay.max.ms.schedule_preempt_disabled.rwsem_down_write_slowpath.down_write.unlink_file_vma
>       0.02           -27.5%       0.02 ±  5%  perf-sched.total_sch_delay.average.ms
>       1.56           +31.8%       2.06 ±  3%  perf-sched.total_wait_and_delay.average.ms
>     827124           -21.1%     652241 ±  3%  perf-sched.total_wait_and_delay.count.ms
>       1.54           +32.6%       2.04 ±  3%  perf-sched.total_wait_time.average.ms
>       0.59           +66.9%       0.98 ±  5%  perf-sched.wait_and_delay.avg.ms.do_wait.kernel_wait4.__do_sys_wait4.do_syscall_64
>     299.52 ±  5%      -8.5%     273.93 ±  3%  perf-sched.wait_and_delay.avg.ms.pipe_read.vfs_read.ksys_read.do_syscall_64
>      41.36 ±  3%     +18.5%      49.03 ±  6%  perf-sched.wait_and_delay.avg.ms.schedule_hrtimeout_range_clock.do_poll.constprop.0.do_sys_poll
>      16.06 ±  2%     +30.9%      21.02 ±  3%  perf-sched.wait_and_delay.avg.ms.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
>       5.75           +28.4%       7.39 ±  3%  perf-sched.wait_and_delay.avg.ms.worker_thread.kthread.ret_from_fork.ret_from_fork_asm
>      12.67 ± 17%     -53.9%       5.83 ± 58%  perf-sched.wait_and_delay.count.__cond_resched.__alloc_pages.__folio_alloc.vma_alloc_folio.shmem_alloc_folio
>       7866           -30.9%       5433 ±  9%  perf-sched.wait_and_delay.count.__cond_resched.__put_anon_vma.unlink_anon_vmas.free_pgtables.exit_mmap
>       9011           -45.3%       4932 ± 10%  perf-sched.wait_and_delay.count.__cond_resched.exit_mmap.__mmput.exit_mm.do_exit
>      48.00 ± 15%     -24.7%      36.17 ± 17%  perf-sched.wait_and_delay.count.__cond_resched.generic_perform_write.generic_file_write_iter.vfs_write.ksys_write
>       8.33 ± 22%     -66.0%       2.83 ± 47%  perf-sched.wait_and_delay.count.__cond_resched.shmem_get_folio_gfp.shmem_write_begin.generic_perform_write.generic_file_write_iter
>       9855           -57.4%       4200 ± 12%  perf-sched.wait_and_delay.count.__cond_resched.tlb_batch_pages_flush.tlb_finish_mmu.exit_mmap.__mmput
>       7774           -11.5%       6877 ±  4%  perf-sched.wait_and_delay.count.__cond_resched.zap_pmd_range.isra.0.unmap_page_range
>     317531           -25.2%     237490 ±  4%  perf-sched.wait_and_delay.count.do_task_dead.do_exit.do_group_exit.get_signal.arch_do_signal_or_restart
>     318099           -25.2%     237911 ±  4%  perf-sched.wait_and_delay.count.do_wait.kernel_wait4.__do_sys_wait4.do_syscall_64
>      15702           +94.1%      30484 ±  2%  perf-sched.wait_and_delay.count.exit_to_user_mode_loop.exit_to_user_mode_prepare.irqentry_exit_to_user_mode.asm_exc_page_fault
>     233.17 ±  5%      +8.6%     253.33 ±  3%  perf-sched.wait_and_delay.count.pipe_read.vfs_read.ksys_read.do_syscall_64
>     320.00 ±  3%     -11.0%     284.83 ±  9%  perf-sched.wait_and_delay.count.schedule_hrtimeout_range_clock.do_poll.constprop.0.do_sys_poll
>      37291           -22.1%      29068 ±  3%  perf-sched.wait_and_delay.count.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
>      56531           -21.9%      44128 ±  3%  perf-sched.wait_and_delay.count.worker_thread.kthread.ret_from_fork.ret_from_fork_asm
>       0.02 ±  9%     +75.2%       0.04 ± 16%  perf-sched.wait_time.avg.ms.__cond_resched.__alloc_pages.__folio_alloc.vma_alloc_folio.wp_page_copy
>       0.57 ±  5%     +65.6%       0.95 ±  9%  perf-sched.wait_time.avg.ms.__cond_resched.__alloc_pages.__get_free_pages.pgd_alloc.mm_init
>       0.57 ±  3%     +64.9%       0.94 ±  4%  perf-sched.wait_time.avg.ms.__cond_resched.__alloc_pages.__pmd_alloc.copy_p4d_range.copy_page_range
>       0.57 ±  7%     +68.7%       0.97 ±  5%  perf-sched.wait_time.avg.ms.__cond_resched.__alloc_pages.__pud_alloc.copy_p4d_range.copy_page_range
>       0.44 ± 45%     +79.5%       0.78 ± 13%  perf-sched.wait_time.avg.ms.__cond_resched.__alloc_pages.allocate_slab.___slab_alloc.kmem_cache_alloc_node
>       0.58 ±  4%     +67.3%       0.98 ±  3%  perf-sched.wait_time.avg.ms.__cond_resched.__alloc_pages.pte_alloc_one.__pte_alloc.copy_pte_range
>       0.01 ± 11%     +79.5%       0.02 ± 19%  perf-sched.wait_time.avg.ms.__cond_resched.__alloc_pages.pte_alloc_one.do_read_fault.do_fault
>       0.57           +57.0%       0.90 ± 16%  perf-sched.wait_time.avg.ms.__cond_resched.__kmem_cache_alloc_node.__kmalloc.security_task_alloc.copy_process
>       0.58 ±  5%     +65.1%       0.96 ±  5%  perf-sched.wait_time.avg.ms.__cond_resched.__kmem_cache_alloc_node.__kmalloc_node.__vmalloc_area_node.__vmalloc_node_range
>       0.57 ±  7%     +71.2%       0.97 ± 16%  perf-sched.wait_time.avg.ms.__cond_resched.__kmem_cache_alloc_node.__kmalloc_node.memcg_alloc_slab_cgroups.allocate_slab
>       0.54 ± 10%     +86.7%       1.01 ± 11%  perf-sched.wait_time.avg.ms.__cond_resched.__mutex_lock.constprop.0.pcpu_alloc
>       0.02 ±  5%     +46.6%       0.03 ± 14%  perf-sched.wait_time.avg.ms.__cond_resched.__put_anon_vma.unlink_anon_vmas.free_pgtables.exit_mmap
>       0.57 ±  4%     +71.6%       0.98 ± 11%  perf-sched.wait_time.avg.ms.__cond_resched.__vmalloc_area_node.__vmalloc_node_range.alloc_thread_stack_node.dup_task_struct
>       0.60 ±  4%     +62.1%       0.97 ±  6%  perf-sched.wait_time.avg.ms.__cond_resched.cgroup_css_set_fork.cgroup_can_fork.copy_process.kernel_clone
>       0.58 ±  2%     +67.5%       0.97 ±  6%  perf-sched.wait_time.avg.ms.__cond_resched.copy_page_range.dup_mmap.dup_mm.constprop
>       0.58           +68.4%       0.97 ±  5%  perf-sched.wait_time.avg.ms.__cond_resched.copy_pte_range.copy_p4d_range.copy_page_range.dup_mmap
>       0.59 ±  3%     +66.2%       0.98 ±  6%  perf-sched.wait_time.avg.ms.__cond_resched.dentry_kill.dput.proc_invalidate_siblings_dcache.release_task
>       0.03 ±  8%     +65.2%       0.05 ± 21%  perf-sched.wait_time.avg.ms.__cond_resched.down_read.acct_collect.do_exit.do_group_exit
>       0.58 ±  2%     +67.4%       0.97 ±  5%  perf-sched.wait_time.avg.ms.__cond_resched.down_write.anon_vma_clone.anon_vma_fork.dup_mmap
>       0.58           +69.3%       0.99 ±  6%  perf-sched.wait_time.avg.ms.__cond_resched.down_write.anon_vma_fork.dup_mmap.dup_mm
>       0.58           +69.2%       0.98 ±  5%  perf-sched.wait_time.avg.ms.__cond_resched.down_write.dup_mmap.dup_mm.constprop
>       0.58          +195.7%       1.71 ± 97%  perf-sched.wait_time.avg.ms.__cond_resched.down_write.dup_userfaultfd.dup_mmap.dup_mm
>       0.02 ±  3%     +40.0%       0.03 ± 13%  perf-sched.wait_time.avg.ms.__cond_resched.down_write.free_pgtables.exit_mmap.__mmput
>       0.02 ± 65%    +137.5%       0.05 ± 23%  perf-sched.wait_time.avg.ms.__cond_resched.down_write.mprotect_fixup.do_mprotect_pkey.__x64_sys_mprotect
>       0.02 ±  4%     +42.3%       0.03 ± 19%  perf-sched.wait_time.avg.ms.__cond_resched.down_write.unlink_file_vma.free_pgtables.exit_mmap
>       0.02 ±  5%     +25.8%       0.03 ± 10%  perf-sched.wait_time.avg.ms.__cond_resched.exit_mmap.__mmput.exit_mm.do_exit
>       0.61 ±  7%     +64.8%       1.00 ±  4%  perf-sched.wait_time.avg.ms.__cond_resched.kmem_cache_alloc.alloc_pid.copy_process.kernel_clone
>       0.58           +69.6%       0.98 ±  6%  perf-sched.wait_time.avg.ms.__cond_resched.kmem_cache_alloc.anon_vma_fork.dup_mmap.dup_mm
>       0.57 ±  4%     +73.5%       1.00 ±  7%  perf-sched.wait_time.avg.ms.__cond_resched.kmem_cache_alloc.copy_fs_struct.copy_process.kernel_clone
>       0.56 ±  3%     +71.9%       0.96 ± 12%  perf-sched.wait_time.avg.ms.__cond_resched.kmem_cache_alloc.copy_sighand.copy_process.kernel_clone
>       0.57 ±  5%     +63.7%       0.93 ± 10%  perf-sched.wait_time.avg.ms.__cond_resched.kmem_cache_alloc.copy_signal.copy_process.kernel_clone
>       0.58 ±  4%     +64.6%       0.96 ±  9%  perf-sched.wait_time.avg.ms.__cond_resched.kmem_cache_alloc.dup_mm.constprop.0
>       0.57          +348.5%       2.57 ±140%  perf-sched.wait_time.avg.ms.__cond_resched.kmem_cache_alloc.prepare_creds.copy_creds.copy_process
>       0.58           +69.4%       0.98 ±  5%  perf-sched.wait_time.avg.ms.__cond_resched.kmem_cache_alloc.vm_area_dup.dup_mmap.dup_mm
>       0.58 ±  6%     +82.1%       1.06 ±  4%  perf-sched.wait_time.avg.ms.__cond_resched.kmem_cache_alloc_node.dup_task_struct.copy_process.kernel_clone
>       0.03 ±  6%     +42.6%       0.04 ± 19%  perf-sched.wait_time.avg.ms.__cond_resched.mmput.getrusage.__do_sys_getrusage.do_syscall_64
>       0.03 ± 20%     +55.3%       0.04 ± 25%  perf-sched.wait_time.avg.ms.__cond_resched.mnt_want_write.open_last_lookups.path_openat.do_filp_open
>       0.02 ± 13%     +43.4%       0.03 ± 10%  perf-sched.wait_time.avg.ms.__cond_resched.mutex_lock.futex_exit_release.exit_mm_release.exit_mm
>       0.57 ±  4%     +69.7%       0.97 ±  4%  perf-sched.wait_time.avg.ms.__cond_resched.mutex_lock.ldt_dup_context.dup_mmap.dup_mm
>       0.58          +208.4%       1.78 ±100%  perf-sched.wait_time.avg.ms.__cond_resched.mutex_lock_killable.pcpu_alloc.__percpu_counter_init.mm_init
>       0.57 ±  2%     +71.2%       0.97 ±  9%  perf-sched.wait_time.avg.ms.__cond_resched.mutex_lock_killable.pcpu_alloc.mm_init.dup_mm
>       0.03 ± 27%     +58.2%       0.04 ± 15%  perf-sched.wait_time.avg.ms.__cond_resched.slab_pre_alloc_hook.constprop.0.kmem_cache_alloc_lru
>       0.58           +69.7%       0.98 ±  9%  perf-sched.wait_time.avg.ms.__cond_resched.uprobe_start_dup_mmap.dup_mmap.dup_mm.constprop
>       0.03 ±  9%     +38.6%       0.04 ±  7%  perf-sched.wait_time.avg.ms.__cond_resched.wp_page_copy.__handle_mm_fault.handle_mm_fault.do_user_addr_fault
>       0.03 ±  3%    +134.6%       0.06 ± 92%  perf-sched.wait_time.avg.ms.__cond_resched.zap_pmd_range.isra.0.unmap_page_range
>       0.04           +21.2%       0.04 ±  3%  perf-sched.wait_time.avg.ms.do_task_dead.do_exit.do_group_exit.get_signal.arch_do_signal_or_restart
>       0.57           +69.4%       0.96 ±  5%  perf-sched.wait_time.avg.ms.do_wait.kernel_wait4.__do_sys_wait4.do_syscall_64
>       0.02 ±  4%     +18.1%       0.02 ±  6%  perf-sched.wait_time.avg.ms.exit_to_user_mode_loop.exit_to_user_mode_prepare.irqentry_exit_to_user_mode.asm_exc_page_fault
>       0.01 ± 11%     +44.3%       0.01 ± 10%  perf-sched.wait_time.avg.ms.io_schedule.folio_wait_bit_common.__do_fault.do_read_fault
>       0.01 ± 27%     +83.6%       0.02 ±  5%  perf-sched.wait_time.avg.ms.io_schedule.folio_wait_bit_common.filemap_fault.__do_fault
>     299.51 ±  5%      -8.5%     273.91 ±  3%  perf-sched.wait_time.avg.ms.pipe_read.vfs_read.ksys_read.do_syscall_64
>      41.36 ±  3%     +18.5%      49.03 ±  6%  perf-sched.wait_time.avg.ms.schedule_hrtimeout_range_clock.do_poll.constprop.0.do_sys_poll
>       0.41 ±  2%     +74.7%       0.72 ± 11%  perf-sched.wait_time.avg.ms.schedule_preempt_disabled.__mutex_lock.constprop.0.pcpu_alloc
>       0.27 ±  5%    +118.9%       0.60 ± 10%  perf-sched.wait_time.avg.ms.schedule_preempt_disabled.rwsem_down_write_slowpath.down_write.anon_vma_clone
>       0.27 ±  3%    +130.0%       0.61 ±  9%  perf-sched.wait_time.avg.ms.schedule_preempt_disabled.rwsem_down_write_slowpath.down_write.anon_vma_fork
>       0.36 ± 14%    +128.3%       0.83 ± 28%  perf-sched.wait_time.avg.ms.schedule_preempt_disabled.rwsem_down_write_slowpath.down_write.dup_mmap
>      16.04 ±  2%     +30.9%      20.99 ±  3%  perf-sched.wait_time.avg.ms.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
>       5.73           +28.6%       7.37 ±  3%  perf-sched.wait_time.avg.ms.worker_thread.kthread.ret_from_fork.ret_from_fork_asm
>       1.06 ± 11%     +38.4%       1.46 ± 10%  perf-sched.wait_time.max.ms.__cond_resched.__alloc_pages.__folio_alloc.vma_alloc_folio.wp_page_copy
>       1.05 ± 15%     +28.0%       1.35 ±  9%  perf-sched.wait_time.max.ms.__cond_resched.__alloc_pages.__get_free_pages.pgd_alloc.mm_init
>       0.92 ± 16%     +54.5%       1.42 ±  7%  perf-sched.wait_time.max.ms.__cond_resched.__alloc_pages.__pud_alloc.copy_p4d_range.copy_page_range
>       0.96 ±  8%     +46.0%       1.41 ± 12%  perf-sched.wait_time.max.ms.__cond_resched.__kmem_cache_alloc_node.__kmalloc_node.memcg_alloc_slab_cgroups.allocate_slab
>       0.75 ± 30%     +85.9%       1.39 ± 12%  perf-sched.wait_time.max.ms.__cond_resched.__mutex_lock.constprop.0.pcpu_alloc
>       1.07 ± 18%     +46.9%       1.57 ± 16%  perf-sched.wait_time.max.ms.__cond_resched.__vmalloc_area_node.__vmalloc_node_range.alloc_thread_stack_node.dup_task_struct
>       0.41 ±  7%     -39.3%       0.25 ± 46%  perf-sched.wait_time.max.ms.__cond_resched.down_read.exit_mmap.__mmput.exit_mm
>       1.28 ±  6%     +61.6%       2.07 ± 14%  perf-sched.wait_time.max.ms.__cond_resched.down_write.anon_vma_fork.dup_mmap.dup_mm
>       4.00 ± 25%     -44.6%       2.22 ± 10%  perf-sched.wait_time.max.ms.__cond_resched.kmem_cache_alloc.anon_vma_fork.dup_mmap.dup_mm
>       0.85 ± 10%     +51.1%       1.29 ± 20%  perf-sched.wait_time.max.ms.__cond_resched.kmem_cache_alloc.copy_sighand.copy_process.kernel_clone
>       0.98 ± 10%     +54.2%       1.52 ± 16%  perf-sched.wait_time.max.ms.__cond_resched.kmem_cache_alloc.dup_mm.constprop.0
>       1.33 ±  7%  +12540.5%     168.29 ±221%  perf-sched.wait_time.max.ms.__cond_resched.kmem_cache_alloc.prepare_creds.copy_creds.copy_process
>       0.96 ±  8%     +65.7%       1.59 ± 12%  perf-sched.wait_time.max.ms.__cond_resched.kmem_cache_alloc_node.dup_task_struct.copy_process.kernel_clone
>       1.10 ± 12%     +40.3%       1.55 ± 10%  perf-sched.wait_time.max.ms.__cond_resched.mutex_lock.ldt_dup_context.dup_mmap.dup_mm
>       1.44 ± 11%  +11607.0%     168.31 ±221%  perf-sched.wait_time.max.ms.__cond_resched.mutex_lock_killable.pcpu_alloc.__percpu_counter_init.mm_init
>       1.13 ± 17%     +48.9%       1.68 ± 16%  perf-sched.wait_time.max.ms.__cond_resched.mutex_lock_killable.pcpu_alloc.mm_init.dup_mm
>       1.18 ±  7%     +45.0%       1.72 ± 23%  perf-sched.wait_time.max.ms.__cond_resched.uprobe_start_dup_mmap.dup_mmap.dup_mm.constprop
>      39.25           -14.6       24.65 ±  5%  perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.__libc_fork
>      39.25           -14.6       24.65 ±  5%  perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.__libc_fork
>      39.23           -14.6       24.64 ±  5%  perf-profile.calltrace.cycles-pp.__do_sys_clone.do_syscall_64.entry_SYSCALL_64_after_hwframe.__libc_fork
>      39.23           -14.6       24.64 ±  5%  perf-profile.calltrace.cycles-pp.kernel_clone.__do_sys_clone.do_syscall_64.entry_SYSCALL_64_after_hwframe.__libc_fork
>      38.39           -14.3       24.05 ±  5%  perf-profile.calltrace.cycles-pp.copy_process.kernel_clone.__do_sys_clone.do_syscall_64.entry_SYSCALL_64_after_hwframe
>      35.84           -13.6       22.28 ±  5%  perf-profile.calltrace.cycles-pp.dup_mm.copy_process.kernel_clone.__do_sys_clone.do_syscall_64
>      41.72           -13.1       28.60 ±  4%  perf-profile.calltrace.cycles-pp.__libc_fork
>      35.34           -12.9       22.40 ±  4%  perf-profile.calltrace.cycles-pp.do_group_exit.get_signal.arch_do_signal_or_restart.exit_to_user_mode_loop.exit_to_user_mode_prepare
>      35.33           -12.9       22.40 ±  4%  perf-profile.calltrace.cycles-pp.do_exit.do_group_exit.get_signal.arch_do_signal_or_restart.exit_to_user_mode_loop
>      35.78           -12.6       23.23 ±  4%  perf-profile.calltrace.cycles-pp.arch_do_signal_or_restart.exit_to_user_mode_loop.exit_to_user_mode_prepare.irqentry_exit_to_user_mode.asm_exc_page_fault
>      35.54           -12.5       23.02 ±  4%  perf-profile.calltrace.cycles-pp.get_signal.arch_do_signal_or_restart.exit_to_user_mode_loop.exit_to_user_mode_prepare.irqentry_exit_to_user_mode
>      34.76           -12.3       22.44 ±  4%  perf-profile.calltrace.cycles-pp.exit_to_user_mode_prepare.irqentry_exit_to_user_mode.asm_exc_page_fault
>      34.76           -12.3       22.44 ±  4%  perf-profile.calltrace.cycles-pp.irqentry_exit_to_user_mode.asm_exc_page_fault
>      34.75           -12.3       22.43 ±  4%  perf-profile.calltrace.cycles-pp.exit_to_user_mode_loop.exit_to_user_mode_prepare.irqentry_exit_to_user_mode.asm_exc_page_fault
>      33.06           -12.3       20.80 ±  5%  perf-profile.calltrace.cycles-pp.dup_mmap.dup_mm.copy_process.kernel_clone.__do_sys_clone
>      32.78           -11.7       21.04 ±  4%  perf-profile.calltrace.cycles-pp.exit_mm.do_exit.do_group_exit.get_signal.arch_do_signal_or_restart
>      32.59           -11.6       20.94 ±  4%  perf-profile.calltrace.cycles-pp.__mmput.exit_mm.do_exit.do_group_exit.get_signal
>      32.51           -11.6       20.87 ±  4%  perf-profile.calltrace.cycles-pp.exit_mmap.__mmput.exit_mm.do_exit.do_group_exit
>      17.35            -8.1        9.28 ±  6%  perf-profile.calltrace.cycles-pp.anon_vma_fork.dup_mmap.dup_mm.copy_process.kernel_clone
>      14.50            -6.0        8.51 ±  5%  perf-profile.calltrace.cycles-pp.free_pgtables.exit_mmap.__mmput.exit_mm.do_exit
>      10.14            -4.6        5.59 ±  6%  perf-profile.calltrace.cycles-pp.unlink_anon_vmas.free_pgtables.exit_mmap.__mmput.exit_mm
>      10.56            -4.5        6.04 ±  6%  perf-profile.calltrace.cycles-pp.anon_vma_clone.anon_vma_fork.dup_mmap.dup_mm.copy_process
>       4.98            -2.9        2.09 ± 10%  perf-profile.calltrace.cycles-pp.down_write.anon_vma_fork.dup_mmap.dup_mm.copy_process
>       4.78            -2.8        1.94 ± 10%  perf-profile.calltrace.cycles-pp.rwsem_down_write_slowpath.down_write.anon_vma_fork.dup_mmap.dup_mm
>       4.65            -2.8        1.87 ± 10%  perf-profile.calltrace.cycles-pp.rwsem_optimistic_spin.rwsem_down_write_slowpath.down_write.anon_vma_fork.dup_mmap
>       9.00            -2.5        6.49 ±  4%  perf-profile.calltrace.cycles-pp.unmap_vmas.exit_mmap.__mmput.exit_mm.do_exit
>       8.45            -2.4        6.04 ±  4%  perf-profile.calltrace.cycles-pp.unmap_page_range.unmap_vmas.exit_mmap.__mmput.exit_mm
>       8.21            -2.3        5.86 ±  4%  perf-profile.calltrace.cycles-pp.zap_pmd_range.unmap_page_range.unmap_vmas.exit_mmap.__mmput
>       3.93            -2.3        1.64 ± 10%  perf-profile.calltrace.cycles-pp.down_write.anon_vma_clone.anon_vma_fork.dup_mmap.dup_mm
>       3.70            -2.3        1.45 ±  9%  perf-profile.calltrace.cycles-pp.down_write.unlink_anon_vmas.free_pgtables.exit_mmap.__mmput
>       7.83            -2.2        5.60 ±  4%  perf-profile.calltrace.cycles-pp.zap_pte_range.zap_pmd_range.unmap_page_range.unmap_vmas.exit_mmap
>       3.30            -2.1        1.20 ± 10%  perf-profile.calltrace.cycles-pp.rwsem_down_write_slowpath.down_write.unlink_anon_vmas.free_pgtables.exit_mmap
>       3.50            -2.1        1.41 ± 10%  perf-profile.calltrace.cycles-pp.rwsem_down_write_slowpath.down_write.anon_vma_clone.anon_vma_fork.dup_mmap
>       5.09 ±  2%      -2.0        3.04 ±  6%  perf-profile.calltrace.cycles-pp.tlb_finish_mmu.exit_mmap.__mmput.exit_mm.do_exit
>       5.04 ±  2%      -2.0        3.00 ±  6%  perf-profile.calltrace.cycles-pp.tlb_batch_pages_flush.tlb_finish_mmu.exit_mmap.__mmput.exit_mm
>       3.38            -2.0        1.35 ± 10%  perf-profile.calltrace.cycles-pp.rwsem_optimistic_spin.rwsem_down_write_slowpath.down_write.anon_vma_clone.anon_vma_fork
>       3.16            -2.0        1.14 ± 10%  perf-profile.calltrace.cycles-pp.rwsem_optimistic_spin.rwsem_down_write_slowpath.down_write.unlink_anon_vmas.free_pgtables
>       2.79            -1.8        0.97 ± 12%  perf-profile.calltrace.cycles-pp.osq_lock.rwsem_optimistic_spin.rwsem_down_write_slowpath.down_write.anon_vma_fork
>       4.10 ±  3%      -1.7        2.42 ±  6%  perf-profile.calltrace.cycles-pp.release_pages.tlb_batch_pages_flush.tlb_finish_mmu.exit_mmap.__mmput
>       1.88            -1.5        0.39 ± 71%  perf-profile.calltrace.cycles-pp.osq_lock.rwsem_optimistic_spin.rwsem_down_write_slowpath.down_write.unlink_anon_vmas
>       1.85            -1.3        0.60 ± 12%  perf-profile.calltrace.cycles-pp.osq_lock.rwsem_optimistic_spin.rwsem_down_write_slowpath.down_write.anon_vma_clone
>       2.64            -1.2        1.39 ±  6%  perf-profile.calltrace.cycles-pp.mm_init.dup_mm.copy_process.kernel_clone.__do_sys_clone
>       3.64            -1.1        2.52 ±  4%  perf-profile.calltrace.cycles-pp.anon_vma_interval_tree_insert.anon_vma_clone.anon_vma_fork.dup_mmap.dup_mm
>       3.31            -1.1        2.25 ±  4%  perf-profile.calltrace.cycles-pp.vm_area_dup.dup_mmap.dup_mm.copy_process.kernel_clone
>       2.98            -1.0        1.94 ±  4%  perf-profile.calltrace.cycles-pp.wait4
>       2.92            -1.0        1.90 ±  4%  perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.wait4
>       2.91            -1.0        1.89 ±  4%  perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.wait4
>       2.88            -1.0        1.86 ±  4%  perf-profile.calltrace.cycles-pp.__do_sys_wait4.do_syscall_64.entry_SYSCALL_64_after_hwframe.wait4
>       2.88            -1.0        1.86 ±  4%  perf-profile.calltrace.cycles-pp.kernel_wait4.__do_sys_wait4.do_syscall_64.entry_SYSCALL_64_after_hwframe.wait4
>       3.92 ±  2%      -1.0        2.91 ±  3%  perf-profile.calltrace.cycles-pp.copy_page_range.dup_mmap.dup_mm.copy_process.kernel_clone
>       2.82            -1.0        1.82 ±  4%  perf-profile.calltrace.cycles-pp.do_wait.kernel_wait4.__do_sys_wait4.do_syscall_64.entry_SYSCALL_64_after_hwframe
>       1.90            -1.0        0.94 ±  7%  perf-profile.calltrace.cycles-pp.__percpu_counter_init.mm_init.dup_mm.copy_process.kernel_clone
>       3.73 ±  2%      -0.9        2.79 ±  3%  perf-profile.calltrace.cycles-pp.copy_p4d_range.copy_page_range.dup_mmap.dup_mm.copy_process
>       2.36 ±  5%      -0.9        1.45 ±  5%  perf-profile.calltrace.cycles-pp.unlink_file_vma.free_pgtables.exit_mmap.__mmput.exit_mm
>       1.67            -0.9        0.81 ±  8%  perf-profile.calltrace.cycles-pp.pcpu_alloc.__percpu_counter_init.mm_init.dup_mm.copy_process
>       1.54            -0.9        0.69 ± 10%  perf-profile.calltrace.cycles-pp.rwsem_spin_on_owner.rwsem_optimistic_spin.rwsem_down_write_slowpath.down_write.anon_vma_fork
>       3.13 ±  3%      -0.8        2.33 ±  3%  perf-profile.calltrace.cycles-pp.copy_pte_range.copy_p4d_range.copy_page_range.dup_mmap.dup_mm
>       1.83 ±  6%      -0.7        1.09 ±  8%  perf-profile.calltrace.cycles-pp._compound_head.zap_pte_range.zap_pmd_range.unmap_page_range.unmap_vmas
>       1.60 ±  2%      -0.7        0.87 ±  6%  perf-profile.calltrace.cycles-pp.__put_anon_vma.unlink_anon_vmas.free_pgtables.exit_mmap.__mmput
>       2.32            -0.7        1.62 ±  4%  perf-profile.calltrace.cycles-pp.kmem_cache_alloc.vm_area_dup.dup_mmap.dup_mm.copy_process
>       1.26            -0.7        0.58 ± 10%  perf-profile.calltrace.cycles-pp.rwsem_spin_on_owner.rwsem_optimistic_spin.rwsem_down_write_slowpath.down_write.anon_vma_clone
>       1.46 ±  6%      -0.7        0.80 ±  7%  perf-profile.calltrace.cycles-pp.down_write.unlink_file_vma.free_pgtables.exit_mmap.__mmput
>       1.92 ±  5%      -0.7        1.25 ±  6%  perf-profile.calltrace.cycles-pp.down_write.dup_mmap.dup_mm.copy_process.kernel_clone
>       1.48            -0.5        0.97 ±  5%  perf-profile.calltrace.cycles-pp.schedule.do_wait.kernel_wait4.__do_sys_wait4.do_syscall_64
>       2.05            -0.5        1.55 ±  3%  perf-profile.calltrace.cycles-pp.mas_store.dup_mmap.dup_mm.copy_process.kernel_clone
>       1.47            -0.5        0.97 ±  5%  perf-profile.calltrace.cycles-pp.__schedule.schedule.do_wait.kernel_wait4.__do_sys_wait4
>       0.75            -0.5        0.26 ±100%  perf-profile.calltrace.cycles-pp.kmem_cache_free.unlink_anon_vmas.free_pgtables.exit_mmap.__mmput
>       0.85            -0.5        0.36 ± 70%  perf-profile.calltrace.cycles-pp.free_swap_cache.free_pages_and_swap_cache.tlb_batch_pages_flush.tlb_finish_mmu.exit_mmap
>       1.38            -0.5        0.91 ±  3%  perf-profile.calltrace.cycles-pp.dup_task_struct.copy_process.kernel_clone.__do_sys_clone.do_syscall_64
>       1.86 ±  2%      -0.5        1.39 ±  6%  perf-profile.calltrace.cycles-pp.copy_present_pte.copy_pte_range.copy_p4d_range.copy_page_range.dup_mmap
>       1.12            -0.4        0.70 ±  4%  perf-profile.calltrace.cycles-pp.alloc_thread_stack_node.dup_task_struct.copy_process.kernel_clone.__do_sys_clone
>       1.18            -0.4        0.75 ±  5%  perf-profile.calltrace.cycles-pp.ret_from_fork_asm.__libc_fork
>       1.13            -0.4        0.72 ±  5%  perf-profile.calltrace.cycles-pp.ret_from_fork.ret_from_fork_asm.__libc_fork
>       0.67            -0.4        0.26 ±100%  perf-profile.calltrace.cycles-pp.cpuidle_enter.cpuidle_idle_call.do_idle.cpu_startup_entry.start_secondary
>       0.66            -0.4        0.26 ±100%  perf-profile.calltrace.cycles-pp.cpuidle_enter_state.cpuidle_enter.cpuidle_idle_call.do_idle.cpu_startup_entry
>       1.11            -0.4        0.70 ±  5%  perf-profile.calltrace.cycles-pp.schedule_tail.ret_from_fork.ret_from_fork_asm.__libc_fork
>       2.78 ±  3%      -0.4        2.37 ±  3%  perf-profile.calltrace.cycles-pp.page_remove_rmap.zap_pte_range.zap_pmd_range.unmap_page_range.unmap_vmas
>       1.23 ±  4%      -0.4        0.84 ±  4%  perf-profile.calltrace.cycles-pp.up_write.free_pgtables.exit_mmap.__mmput.exit_mm
>       1.21            -0.4        0.83 ±  7%  perf-profile.calltrace.cycles-pp.__anon_vma_interval_tree_remove.unlink_anon_vmas.free_pgtables.exit_mmap.__mmput
>       1.22            -0.4        0.85 ±  4%  perf-profile.calltrace.cycles-pp.__vm_area_free.exit_mmap.__mmput.exit_mm.do_exit
>       1.14            -0.3        0.81 ±  4%  perf-profile.calltrace.cycles-pp.memcg_slab_post_alloc_hook.kmem_cache_alloc.vm_area_dup.dup_mmap.dup_mm
>       0.87            -0.3        0.54 ±  5%  perf-profile.calltrace.cycles-pp.free_pages_and_swap_cache.tlb_batch_pages_flush.tlb_finish_mmu.exit_mmap.__mmput
>       1.39            -0.3        1.07 ±  3%  perf-profile.calltrace.cycles-pp.up_write.dup_mmap.dup_mm.copy_process.kernel_clone
>       0.94            -0.3        0.66 ±  4%  perf-profile.calltrace.cycles-pp.wait_task_zombie.do_wait.kernel_wait4.__do_sys_wait4.do_syscall_64
>       1.23            -0.3        0.96 ±  2%  perf-profile.calltrace.cycles-pp.mas_wr_bnode.mas_store.dup_mmap.dup_mm.copy_process
>       0.88            -0.3        0.61 ±  4%  perf-profile.calltrace.cycles-pp.release_task.wait_task_zombie.do_wait.kernel_wait4.__do_sys_wait4
>       1.09            -0.3        0.83 ±  3%  perf-profile.calltrace.cycles-pp.secondary_startup_64_no_verify
>       1.07            -0.2        0.82 ±  3%  perf-profile.calltrace.cycles-pp.do_idle.cpu_startup_entry.start_secondary.secondary_startup_64_no_verify
>       1.07            -0.2        0.82 ±  3%  perf-profile.calltrace.cycles-pp.cpu_startup_entry.start_secondary.secondary_startup_64_no_verify
>       1.07            -0.2        0.82 ±  3%  perf-profile.calltrace.cycles-pp.start_secondary.secondary_startup_64_no_verify
>       0.80            -0.2        0.56 ±  4%  perf-profile.calltrace.cycles-pp.__schedule.do_task_dead.do_exit.do_group_exit.get_signal
>       0.81            -0.2        0.56 ±  4%  perf-profile.calltrace.cycles-pp.do_task_dead.do_exit.do_group_exit.get_signal.arch_do_signal_or_restart
>       0.79            -0.2        0.54 ±  4%  perf-profile.calltrace.cycles-pp.kmem_cache_alloc.anon_vma_fork.dup_mmap.dup_mm.copy_process
>       0.82            -0.2        0.57 ±  6%  perf-profile.calltrace.cycles-pp.remove_vma.exit_mmap.__mmput.exit_mm.do_exit
>       0.89            -0.2        0.64 ±  5%  perf-profile.calltrace.cycles-pp.kmem_cache_alloc.anon_vma_clone.anon_vma_fork.dup_mmap.dup_mm
>       0.79            -0.2        0.55 ±  4%  perf-profile.calltrace.cycles-pp.wake_up_new_task.kernel_clone.__do_sys_clone.do_syscall_64.entry_SYSCALL_64_after_hwframe
>       1.08            -0.2        0.84 ±  3%  perf-profile.calltrace.cycles-pp.mas_split.mas_wr_bnode.mas_store.dup_mmap.dup_mm
>       1.06            -0.2        0.83 ±  2%  perf-profile.calltrace.cycles-pp.irqentry_exit_to_user_mode.asm_exc_page_fault.stress_bad_altstack
>       1.06            -0.2        0.83 ±  2%  perf-profile.calltrace.cycles-pp.exit_to_user_mode_loop.exit_to_user_mode_prepare.irqentry_exit_to_user_mode.asm_exc_page_fault.stress_bad_altstack
>       1.06            -0.2        0.83 ±  2%  perf-profile.calltrace.cycles-pp.exit_to_user_mode_prepare.irqentry_exit_to_user_mode.asm_exc_page_fault.stress_bad_altstack
>       1.00            -0.2        0.78 ±  3%  perf-profile.calltrace.cycles-pp.__slab_free.unlink_anon_vmas.free_pgtables.exit_mmap.__mmput
>       1.05            -0.2        0.84        perf-profile.calltrace.cycles-pp.do_filp_open.do_sys_openat2.__x64_sys_openat.do_syscall_64.entry_SYSCALL_64_after_hwframe
>       1.05            -0.2        0.84        perf-profile.calltrace.cycles-pp.path_openat.do_filp_open.do_sys_openat2.__x64_sys_openat.do_syscall_64
>       1.11            -0.2        0.91        perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.open64
>       1.10            -0.2        0.90        perf-profile.calltrace.cycles-pp.__x64_sys_openat.do_syscall_64.entry_SYSCALL_64_after_hwframe.open64
>       1.11            -0.2        0.91        perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.open64
>       1.10            -0.2        0.90        perf-profile.calltrace.cycles-pp.do_sys_openat2.__x64_sys_openat.do_syscall_64.entry_SYSCALL_64_after_hwframe.open64
>       0.77            -0.2        0.58 ±  4%  perf-profile.calltrace.cycles-pp.cpuidle_idle_call.do_idle.cpu_startup_entry.start_secondary.secondary_startup_64_no_verify
>       0.72 ±  3%      -0.2        0.54 ±  2%  perf-profile.calltrace.cycles-pp.vma_interval_tree_remove.unlink_file_vma.free_pgtables.exit_mmap.__mmput
>       1.14            -0.1        0.99        perf-profile.calltrace.cycles-pp.open64
>       0.73 ±  2%      -0.1        0.65 ±  6%  perf-profile.calltrace.cycles-pp.do_set_pte.filemap_map_pages.do_read_fault.do_fault.__handle_mm_fault
>       0.58            -0.1        0.53 ±  2%  perf-profile.calltrace.cycles-pp.__slab_free.exit_mmap.__mmput.exit_mm.do_exit
>       0.59 ±  2%      +0.1        0.67 ±  2%  perf-profile.calltrace.cycles-pp.dup_userfaultfd.dup_mmap.dup_mm.copy_process.kernel_clone
>       0.74 ± 19%      +0.3        1.00 ± 14%  perf-profile.calltrace.cycles-pp.exc_page_fault.asm_exc_page_fault.stress_mwc_reseed
>       0.74 ± 20%      +0.3        1.00 ± 14%  perf-profile.calltrace.cycles-pp.do_user_addr_fault.exc_page_fault.asm_exc_page_fault.stress_mwc_reseed
>       0.67 ± 20%      +0.3        0.95 ± 14%  perf-profile.calltrace.cycles-pp.handle_mm_fault.do_user_addr_fault.exc_page_fault.asm_exc_page_fault.stress_mwc_reseed
>       1.06 ±  3%      +0.3        1.38 ± 30%  perf-profile.calltrace.cycles-pp.exc_page_fault.asm_exc_page_fault.stress_bad_altstack
>       0.98 ±  3%      +0.3        1.31 ± 31%  perf-profile.calltrace.cycles-pp.do_user_addr_fault.exc_page_fault.asm_exc_page_fault.stress_bad_altstack
>       0.84 ±  3%      +0.4        1.20 ± 34%  perf-profile.calltrace.cycles-pp.handle_mm_fault.do_user_addr_fault.exc_page_fault.asm_exc_page_fault.stress_bad_altstack
>       0.09 ±223%      +0.6        0.74 ± 15%  perf-profile.calltrace.cycles-pp.shim_nanosleep_uint64
>       0.00            +0.7        0.66 ± 16%  perf-profile.calltrace.cycles-pp.__sigsetjmp@plt
>       1.22            +1.9        3.13 ±  4%  perf-profile.calltrace.cycles-pp.asm_exc_page_fault.__libc_fork
>       0.99            +2.0        2.97 ±  5%  perf-profile.calltrace.cycles-pp.do_user_addr_fault.exc_page_fault.asm_exc_page_fault.__libc_fork
>       1.00            +2.0        2.98 ±  5%  perf-profile.calltrace.cycles-pp.exc_page_fault.asm_exc_page_fault.__libc_fork
>       0.00            +2.0        1.98 ±  5%  perf-profile.calltrace.cycles-pp.next_uptodate_folio.filemap_map_pages.do_read_fault.do_fault.__handle_mm_fault
>       0.70            +2.0        2.70 ±  6%  perf-profile.calltrace.cycles-pp.do_sys_openat2.__x64_sys_openat.do_syscall_64.entry_SYSCALL_64_after_hwframe.__open64_nocancel
>       0.71 ±  2%      +2.0        2.71 ±  6%  perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.__open64_nocancel
>       0.71 ±  2%      +2.0        2.71 ±  6%  perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.__open64_nocancel
>       0.72 ±  2%      +2.0        2.72 ±  6%  perf-profile.calltrace.cycles-pp.__open64_nocancel
>       0.71            +2.0        2.71 ±  6%  perf-profile.calltrace.cycles-pp.__x64_sys_openat.do_syscall_64.entry_SYSCALL_64_after_hwframe.__open64_nocancel
>       0.00            +2.0        2.02 ± 16%  perf-profile.calltrace.cycles-pp.handle_mm_fault.do_user_addr_fault.exc_page_fault.asm_exc_page_fault.strlen@plt
>       0.80 ±  2%      +2.0        2.83 ±  5%  perf-profile.calltrace.cycles-pp.handle_mm_fault.do_user_addr_fault.exc_page_fault.asm_exc_page_fault.__libc_fork
>       0.00            +2.0        2.04 ± 16%  perf-profile.calltrace.cycles-pp.exc_page_fault.asm_exc_page_fault.strlen@plt
>       0.00            +2.0        2.04 ± 16%  perf-profile.calltrace.cycles-pp.do_user_addr_fault.exc_page_fault.asm_exc_page_fault.strlen@plt
>       0.00            +2.1        2.05 ± 16%  perf-profile.calltrace.cycles-pp.asm_exc_page_fault.strlen@plt
>       0.52 ±  3%      +2.1        2.58 ±  6%  perf-profile.calltrace.cycles-pp.asm_exc_page_fault.__sigsetjmp
>       0.52 ±  2%      +2.1        2.58 ±  6%  perf-profile.calltrace.cycles-pp.__sigsetjmp
>       0.00            +2.1        2.06 ± 15%  perf-profile.calltrace.cycles-pp.strlen@plt
>       0.44 ± 44%      +2.1        2.56 ±  6%  perf-profile.calltrace.cycles-pp.setrlimit64
>       0.00            +2.4        2.37 ±  7%  perf-profile.calltrace.cycles-pp.handle_mm_fault.do_user_addr_fault.exc_page_fault.asm_exc_page_fault.strncpy_from_user
>       0.00            +2.4        2.39 ±  7%  perf-profile.calltrace.cycles-pp.exc_page_fault.asm_exc_page_fault.strncpy_from_user.getname_flags.do_sys_openat2
>       0.00            +2.4        2.39 ±  7%  perf-profile.calltrace.cycles-pp.do_user_addr_fault.exc_page_fault.asm_exc_page_fault.strncpy_from_user.getname_flags
>       0.00            +2.4        2.40 ±  7%  perf-profile.calltrace.cycles-pp.handle_mm_fault.do_user_addr_fault.exc_page_fault.asm_exc_page_fault.setrlimit64
>       0.00            +2.4        2.41 ±  7%  perf-profile.calltrace.cycles-pp.asm_exc_page_fault.strncpy_from_user.getname_flags.do_sys_openat2.__x64_sys_openat
>       0.00            +2.4        2.42 ±  7%  perf-profile.calltrace.cycles-pp.exc_page_fault.asm_exc_page_fault.setrlimit64
>       0.00            +2.4        2.42 ±  7%  perf-profile.calltrace.cycles-pp.do_user_addr_fault.exc_page_fault.asm_exc_page_fault.setrlimit64
>       0.00            +2.4        2.42 ±  7%  perf-profile.calltrace.cycles-pp.strncpy_from_user.getname_flags.do_sys_openat2.__x64_sys_openat.do_syscall_64
>       0.00            +2.4        2.43 ±  6%  perf-profile.calltrace.cycles-pp.asm_exc_page_fault.setrlimit64
>       0.00            +2.5        2.45 ±  7%  perf-profile.calltrace.cycles-pp.getname_flags.do_sys_openat2.__x64_sys_openat.do_syscall_64.entry_SYSCALL_64_after_hwframe
>       0.08 ±223%      +2.5        2.54 ±  6%  perf-profile.calltrace.cycles-pp.exc_page_fault.asm_exc_page_fault.__sigsetjmp
>       0.00            +2.5        2.50 ±  6%  perf-profile.calltrace.cycles-pp.handle_mm_fault.do_user_addr_fault.exc_page_fault.asm_exc_page_fault.__sigsetjmp
>       0.00            +2.5        2.54 ±  6%  perf-profile.calltrace.cycles-pp.do_user_addr_fault.exc_page_fault.asm_exc_page_fault.__sigsetjmp
>      40.24            +2.6       42.88 ±  4%  perf-profile.calltrace.cycles-pp.asm_exc_page_fault
>       5.12 ±  4%     +15.0       20.17 ± 12%  perf-profile.calltrace.cycles-pp.exc_page_fault.asm_exc_page_fault
>       5.10 ±  4%     +15.1       20.16 ± 12%  perf-profile.calltrace.cycles-pp.do_user_addr_fault.exc_page_fault.asm_exc_page_fault
>       4.57 ±  4%     +15.2       19.75 ± 12%  perf-profile.calltrace.cycles-pp.handle_mm_fault.do_user_addr_fault.exc_page_fault.asm_exc_page_fault
>       6.51 ±  3%     +32.7       39.16 ±  6%  perf-profile.calltrace.cycles-pp.__handle_mm_fault.handle_mm_fault.do_user_addr_fault.exc_page_fault.asm_exc_page_fault
>       4.52 ±  4%     +33.2       37.68 ±  6%  perf-profile.calltrace.cycles-pp.do_read_fault.do_fault.__handle_mm_fault.handle_mm_fault.do_user_addr_fault
>       4.54 ±  4%     +33.2       37.70 ±  6%  perf-profile.calltrace.cycles-pp.do_fault.__handle_mm_fault.handle_mm_fault.do_user_addr_fault.exc_page_fault
>       3.95 ±  8%     +33.3       37.25 ±  6%  perf-profile.calltrace.cycles-pp.filemap_map_pages.do_read_fault.do_fault.__handle_mm_fault.handle_mm_fault
>      39.23           -14.6       24.64 ±  5%  perf-profile.children.cycles-pp.__do_sys_clone
>      39.23           -14.6       24.64 ±  5%  perf-profile.children.cycles-pp.kernel_clone
>      38.39           -14.3       24.05 ±  5%  perf-profile.children.cycles-pp.copy_process
>      45.65           -14.2       31.50 ±  3%  perf-profile.children.cycles-pp.entry_SYSCALL_64_after_hwframe
>      45.61           -14.1       31.47 ±  3%  perf-profile.children.cycles-pp.do_syscall_64
>      35.84           -13.6       22.28 ±  5%  perf-profile.children.cycles-pp.dup_mm
>      41.88           -13.2       28.71 ±  4%  perf-profile.children.cycles-pp.__libc_fork
>      36.14           -12.6       23.54 ±  4%  perf-profile.children.cycles-pp.exit_to_user_mode_prepare
>      36.04           -12.6       23.46 ±  4%  perf-profile.children.cycles-pp.exit_to_user_mode_loop
>      35.79           -12.6       23.23 ±  4%  perf-profile.children.cycles-pp.arch_do_signal_or_restart
>      35.86           -12.6       23.31 ±  4%  perf-profile.children.cycles-pp.irqentry_exit_to_user_mode
>      35.54           -12.5       23.02 ±  4%  perf-profile.children.cycles-pp.get_signal
>      35.34           -12.5       22.85 ±  4%  perf-profile.children.cycles-pp.do_group_exit
>      35.34           -12.5       22.85 ±  4%  perf-profile.children.cycles-pp.do_exit
>      33.16           -12.3       20.87 ±  5%  perf-profile.children.cycles-pp.dup_mmap
>      32.83           -11.8       21.08 ±  4%  perf-profile.children.cycles-pp.exit_mm
>      32.60           -11.7       20.95 ±  4%  perf-profile.children.cycles-pp.__mmput
>      32.54           -11.6       20.89 ±  4%  perf-profile.children.cycles-pp.exit_mmap
>      17.56            -9.3        8.30 ±  7%  perf-profile.children.cycles-pp.down_write
>      13.98            -8.4        5.54 ±  9%  perf-profile.children.cycles-pp.rwsem_down_write_slowpath
>      13.39            -8.1        5.28 ±  9%  perf-profile.children.cycles-pp.rwsem_optimistic_spin
>      17.36            -8.1        9.29 ±  7%  perf-profile.children.cycles-pp.anon_vma_fork
>      14.52            -6.0        8.52 ±  5%  perf-profile.children.cycles-pp.free_pgtables
>       8.49            -5.8        2.71 ± 12%  perf-profile.children.cycles-pp.osq_lock
>      10.17            -4.6        5.61 ±  6%  perf-profile.children.cycles-pp.unlink_anon_vmas
>      10.58            -4.5        6.05 ±  6%  perf-profile.children.cycles-pp.anon_vma_clone
>       9.02            -2.5        6.50 ±  4%  perf-profile.children.cycles-pp.unmap_vmas
>       4.42            -2.4        1.98 ±  8%  perf-profile.children.cycles-pp.rwsem_spin_on_owner
>       8.47            -2.4        6.05 ±  4%  perf-profile.children.cycles-pp.unmap_page_range
>       8.23            -2.3        5.88 ±  4%  perf-profile.children.cycles-pp.zap_pmd_range
>       8.04            -2.3        5.75 ±  4%  perf-profile.children.cycles-pp.zap_pte_range
>       5.09 ±  2%      -2.0        3.05 ±  6%  perf-profile.children.cycles-pp.tlb_finish_mmu
>       5.04 ±  2%      -2.0        3.01 ±  6%  perf-profile.children.cycles-pp.tlb_batch_pages_flush
>       4.14 ±  3%      -1.7        2.45 ±  6%  perf-profile.children.cycles-pp.release_pages
>       2.31            -1.4        0.90 ± 10%  perf-profile.children.cycles-pp.native_queued_spin_lock_slowpath
>       4.43            -1.3        3.14 ±  4%  perf-profile.children.cycles-pp.kmem_cache_alloc
>       4.25            -1.3        2.99 ±  3%  perf-profile.children.cycles-pp.up_write
>       2.64            -1.2        1.40 ±  6%  perf-profile.children.cycles-pp.mm_init
>       2.26            -1.1        1.12 ±  6%  perf-profile.children.cycles-pp.pcpu_alloc
>       3.67            -1.1        2.54 ±  4%  perf-profile.children.cycles-pp.anon_vma_interval_tree_insert
>       3.33            -1.1        2.26 ±  5%  perf-profile.children.cycles-pp.vm_area_dup
>       2.98            -1.0        1.94 ±  4%  perf-profile.children.cycles-pp.wait4
>       2.66 ±  6%      -1.0        1.64 ±  8%  perf-profile.children.cycles-pp._compound_head
>       2.88            -1.0        1.87 ±  4%  perf-profile.children.cycles-pp.__do_sys_wait4
>       2.88            -1.0        1.86 ±  4%  perf-profile.children.cycles-pp.kernel_wait4
>       3.93 ±  2%      -1.0        2.92 ±  3%  perf-profile.children.cycles-pp.copy_page_range
>       2.82            -1.0        1.82 ±  4%  perf-profile.children.cycles-pp.do_wait
>       1.90            -1.0        0.95 ±  7%  perf-profile.children.cycles-pp.__percpu_counter_init
>       3.74 ±  2%      -0.9        2.79 ±  3%  perf-profile.children.cycles-pp.copy_p4d_range
>       2.38 ±  5%      -0.9        1.46 ±  5%  perf-profile.children.cycles-pp.unlink_file_vma
>       2.56            -0.8        1.75 ±  3%  perf-profile.children.cycles-pp.kmem_cache_free
>       3.14 ±  2%      -0.8        2.35 ±  3%  perf-profile.children.cycles-pp.copy_pte_range
>       2.69            -0.8        1.90 ±  4%  perf-profile.children.cycles-pp.__schedule
>       1.93            -0.8        1.16 ±  5%  perf-profile.children.cycles-pp.ret_from_fork_asm
>       1.88            -0.8        1.13 ±  6%  perf-profile.children.cycles-pp.ret_from_fork
>       1.61            -0.7        0.87 ±  6%  perf-profile.children.cycles-pp.__put_anon_vma
>       1.40 ±  2%      -0.7        0.72 ±  7%  perf-profile.children.cycles-pp._raw_spin_lock_irqsave
>       2.31            -0.7        1.64 ±  5%  perf-profile.children.cycles-pp.memcg_slab_post_alloc_hook
>       0.89 ±  3%      -0.7        0.24 ± 13%  perf-profile.children.cycles-pp.__mutex_lock
>       1.55            -0.6        0.93 ±  6%  perf-profile.children.cycles-pp.__mmdrop
>       1.66            -0.6        1.11 ±  4%  perf-profile.children.cycles-pp.mod_objcg_state
>       1.49            -0.5        0.94 ±  5%  perf-profile.children.cycles-pp.finish_task_switch
>       2.06            -0.5        1.56 ±  3%  perf-profile.children.cycles-pp.mas_store
>       2.44            -0.5        1.96 ±  3%  perf-profile.children.cycles-pp.__slab_free
>       1.60            -0.5        1.12 ±  4%  perf-profile.children.cycles-pp.schedule
>       2.10            -0.5        1.62 ±  2%  perf-profile.children.cycles-pp._raw_spin_lock
>       1.89 ±  2%      -0.5        1.41 ±  6%  perf-profile.children.cycles-pp.copy_present_pte
>       1.38            -0.5        0.92 ±  3%  perf-profile.children.cycles-pp.dup_task_struct
>       1.44            -0.5        0.98 ±  4%  perf-profile.children.cycles-pp.pick_next_task_fair
>       1.34            -0.4        0.90 ±  5%  perf-profile.children.cycles-pp.newidle_balance
>       1.12            -0.4        0.70 ±  4%  perf-profile.children.cycles-pp.alloc_thread_stack_node
>       0.91 ±  2%      -0.4        0.48 ±  5%  perf-profile.children.cycles-pp.__vmalloc_node_range
>       1.24            -0.4        0.83 ±  5%  perf-profile.children.cycles-pp.load_balance
>       1.11            -0.4        0.70 ±  5%  perf-profile.children.cycles-pp.schedule_tail
>       2.82 ±  3%      -0.4        2.42 ±  2%  perf-profile.children.cycles-pp.page_remove_rmap
>       0.80 ±  2%      -0.4        0.41 ±  5%  perf-profile.children.cycles-pp.__get_vm_area_node
>       0.78 ±  2%      -0.4        0.39 ±  4%  perf-profile.children.cycles-pp.alloc_vmap_area
>       1.22 ±  2%      -0.4        0.84 ±  8%  perf-profile.children.cycles-pp.__anon_vma_interval_tree_remove
>       1.24            -0.4        0.86 ±  4%  perf-profile.children.cycles-pp.__vm_area_free
>       1.46 ±  3%      -0.4        1.10 ±  4%  perf-profile.children.cycles-pp.__alloc_pages
>       0.91            -0.3        0.57 ±  5%  perf-profile.children.cycles-pp.free_pages_and_swap_cache
>       0.75 ±  2%      -0.3        0.41 ±  7%  perf-profile.children.cycles-pp.kthread
>       0.88            -0.3        0.54 ±  5%  perf-profile.children.cycles-pp.free_swap_cache
>       0.62            -0.3        0.30 ±  7%  perf-profile.children.cycles-pp.free_percpu
>       1.11            -0.3        0.80 ±  4%  perf-profile.children.cycles-pp.__do_softirq
>       0.56 ±  4%      -0.3        0.26 ±  3%  perf-profile.children.cycles-pp.__perf_sw_event
>       0.91            -0.3        0.60 ±  6%  perf-profile.children.cycles-pp.find_busiest_group
>       0.48 ±  2%      -0.3        0.18 ± 11%  perf-profile.children.cycles-pp.queued_write_lock_slowpath
>       0.90            -0.3        0.60 ±  5%  perf-profile.children.cycles-pp.update_sd_lb_stats
>       0.53 ±  4%      -0.3        0.23 ±  3%  perf-profile.children.cycles-pp.___perf_sw_event
>       0.99            -0.3        0.70 ±  5%  perf-profile.children.cycles-pp.rcu_core
>       0.96            -0.3        0.68 ±  5%  perf-profile.children.cycles-pp.rcu_do_batch
>       0.94            -0.3        0.66 ±  4%  perf-profile.children.cycles-pp.wait_task_zombie
>       0.57            -0.3        0.28 ±  6%  perf-profile.children.cycles-pp.percpu_counter_destroy
>       0.64 ± 12%      -0.3        0.36 ± 17%  perf-profile.children.cycles-pp.machine__process_fork_event
>       0.82            -0.3        0.54 ±  5%  perf-profile.children.cycles-pp.update_sg_lb_stats
>       0.88            -0.3        0.61 ±  4%  perf-profile.children.cycles-pp.release_task
>       1.23            -0.3        0.96 ±  2%  perf-profile.children.cycles-pp.mas_wr_bnode
>       1.34            -0.3        1.07        perf-profile.children.cycles-pp.do_filp_open
>       1.33            -0.3        1.07        perf-profile.children.cycles-pp.path_openat
>       0.92 ±  4%      -0.3        0.66 ±  4%  perf-profile.children.cycles-pp.get_page_from_freelist
>       1.09            -0.3        0.83 ±  3%  perf-profile.children.cycles-pp.secondary_startup_64_no_verify
>       1.09            -0.3        0.83 ±  3%  perf-profile.children.cycles-pp.cpu_startup_entry
>       0.79 ±  9%      -0.3        0.54 ± 14%  perf-profile.children.cycles-pp.perf_session__deliver_event
>       1.08            -0.3        0.83 ±  3%  perf-profile.children.cycles-pp.do_idle
>       0.81 ±  9%      -0.2        0.56 ± 13%  perf-profile.children.cycles-pp.perf_session__process_user_event
>       0.81 ±  9%      -0.2        0.56 ± 13%  perf-profile.children.cycles-pp.__ordered_events__flush
>       0.82            -0.2        0.58 ±  6%  perf-profile.children.cycles-pp.remove_vma
>       1.07            -0.2        0.82 ±  3%  perf-profile.children.cycles-pp.start_secondary
>       0.81            -0.2        0.56 ±  5%  perf-profile.children.cycles-pp.do_task_dead
>       0.79            -0.2        0.55 ±  4%  perf-profile.children.cycles-pp.wake_up_new_task
>       1.09            -0.2        0.85 ±  2%  perf-profile.children.cycles-pp.mas_split
>       0.41 ±  7%      -0.2        0.18 ± 14%  perf-profile.children.cycles-pp.folio_lruvec_lock_irqsave
>       0.69 ±  5%      -0.2        0.46 ±  6%  perf-profile.children.cycles-pp.osq_unlock
>       1.41            -0.2        1.18 ±  2%  perf-profile.children.cycles-pp.asm_sysvec_apic_timer_interrupt
>       1.37            -0.2        1.14 ±  2%  perf-profile.children.cycles-pp.sysvec_apic_timer_interrupt
>       0.74            -0.2        0.52 ±  4%  perf-profile.children.cycles-pp.__cond_resched
>       0.72 ±  2%      -0.2        0.50 ±  5%  perf-profile.children.cycles-pp.mas_next_slot
>       0.81 ±  4%      -0.2        0.60 ±  7%  perf-profile.children.cycles-pp.pte_alloc_one
>       0.51 ±  4%      -0.2        0.30 ±  6%  perf-profile.children.cycles-pp.__rb_erase_color
>       0.81            -0.2        0.60 ±  3%  perf-profile.children.cycles-pp.lock_vma_under_rcu
>       0.73            -0.2        0.52 ±  6%  perf-profile.children.cycles-pp.fput
>       0.79            -0.2        0.59 ±  3%  perf-profile.children.cycles-pp.cpuidle_idle_call
>       0.68            -0.2        0.48 ±  5%  perf-profile.children.cycles-pp.select_task_rq_fair
>       0.68 ±  2%      -0.2        0.48 ±  4%  perf-profile.children.cycles-pp.mas_wr_store_entry
>       0.66            -0.2        0.46 ±  6%  perf-profile.children.cycles-pp.__percpu_counter_sum
>       0.32 ±  2%      -0.2        0.12 ±  8%  perf-profile.children.cycles-pp.queued_read_lock_slowpath
>       0.39 ±  5%      -0.2        0.20 ±  8%  perf-profile.children.cycles-pp.worker_thread
>       1.56            -0.2        1.37        perf-profile.children.cycles-pp.do_set_pte
>       0.50            -0.2        0.32 ±  6%  perf-profile.children.cycles-pp.exit_notify
>       0.68 ±  8%      -0.2        0.50 ± 10%  perf-profile.children.cycles-pp.__pte_alloc
>       0.35 ±  5%      -0.2        0.17 ±  9%  perf-profile.children.cycles-pp.process_one_work
>       0.57 ±  2%      -0.2        0.39 ±  4%  perf-profile.children.cycles-pp.percpu_counter_add_batch
>       0.54 ±  5%      -0.2        0.36 ±  5%  perf-profile.children.cycles-pp.clear_page_erms
>       0.53            -0.2        0.35 ±  5%  perf-profile.children.cycles-pp.__list_del_entry_valid
>       0.40 ± 13%      -0.2        0.22 ± 17%  perf-profile.children.cycles-pp.____machine__findnew_thread
>       0.56            -0.2        0.38 ±  5%  perf-profile.children.cycles-pp.find_idlest_cpu
>       0.80            -0.2        0.62 ±  3%  perf-profile.children.cycles-pp.__irq_exit_rcu
>       0.72 ±  3%      -0.2        0.55 ±  2%  perf-profile.children.cycles-pp.vma_interval_tree_remove
>       0.64            -0.2        0.46 ±  4%  perf-profile.children.cycles-pp.acpi_safe_halt
>       0.68            -0.2        0.51 ±  3%  perf-profile.children.cycles-pp.cpuidle_enter_state
>       0.68            -0.2        0.51 ±  4%  perf-profile.children.cycles-pp.cpuidle_enter
>       0.31 ±  5%      -0.2        0.14 ± 11%  perf-profile.children.cycles-pp.vfree
>       0.32 ±  5%      -0.2        0.15 ± 10%  perf-profile.children.cycles-pp.delayed_vfree_work
>       0.64            -0.2        0.47 ±  3%  perf-profile.children.cycles-pp.acpi_idle_enter
>       0.36 ±  3%      -0.2        0.20 ±  9%  perf-profile.children.cycles-pp.smpboot_thread_fn
>       1.15            -0.2        1.00        perf-profile.children.cycles-pp.open64
>       0.49            -0.2        0.34 ±  5%  perf-profile.children.cycles-pp.find_idlest_group
>       0.50            -0.1        0.35 ±  4%  perf-profile.children.cycles-pp.acct_collect
>       0.34 ±  2%      -0.1        0.19 ±  7%  perf-profile.children.cycles-pp.vma_interval_tree_insert_after
>       0.26 ±  7%      -0.1        0.12 ± 12%  perf-profile.children.cycles-pp.remove_vm_area
>       0.36 ±  5%      -0.1        0.21 ±  6%  perf-profile.children.cycles-pp.__rb_insert_augmented
>       0.33 ±  3%      -0.1        0.18 ± 10%  perf-profile.children.cycles-pp.run_ksoftirqd
>       0.49 ±  2%      -0.1        0.35 ±  4%  perf-profile.children.cycles-pp.___slab_alloc
>       0.62 ±  3%      -0.1        0.48 ±  5%  perf-profile.children.cycles-pp.link_path_walk
>       0.45            -0.1        0.32 ±  4%  perf-profile.children.cycles-pp.update_sg_wakeup_stats
>       0.22 ±  3%      -0.1        0.08 ± 11%  perf-profile.children.cycles-pp.mutex_spin_on_owner
>       0.54 ±  2%      -0.1        0.40 ±  3%  perf-profile.children.cycles-pp.sync_regs
>       0.44 ±  2%      -0.1        0.31 ±  4%  perf-profile.children.cycles-pp.mas_wr_append
>       0.42 ±  3%      -0.1        0.29 ±  5%  perf-profile.children.cycles-pp.get_obj_cgroup_from_current
>       0.47 ±  4%      -0.1        0.34 ±  6%  perf-profile.children.cycles-pp.walk_component
>       0.44 ±  2%      -0.1        0.32 ±  3%  perf-profile.children.cycles-pp.obj_cgroup_charge
>       0.19 ±  9%      -0.1        0.08 ± 12%  perf-profile.children.cycles-pp.find_unlink_vmap_area
>       0.28 ± 15%      -0.1        0.17 ± 20%  perf-profile.children.cycles-pp.nsinfo__new
>       0.28 ± 13%      -0.1        0.17 ± 17%  perf-profile.children.cycles-pp.thread__new
>       0.28 ± 14%      -0.1        0.16 ± 19%  perf-profile.children.cycles-pp.__xstat64
>       0.46            -0.1        0.35 ±  3%  perf-profile.children.cycles-pp.open_last_lookups
>       0.43 ±  2%      -0.1        0.32 ±  4%  perf-profile.children.cycles-pp.mas_push_data
>       0.34            -0.1        0.23 ±  3%  perf-profile.children.cycles-pp.dequeue_task_fair
>       0.36            -0.1        0.25 ±  5%  perf-profile.children.cycles-pp.mas_update_gap
>       0.26 ± 14%      -0.1        0.16 ± 18%  perf-profile.children.cycles-pp.__do_sys_newstat
>       0.34 ±  2%      -0.1        0.24 ±  5%  perf-profile.children.cycles-pp.asm_sysvec_call_function_single
>       0.48 ±  3%      -0.1        0.38 ±  6%  perf-profile.children.cycles-pp.__memcg_kmem_charge_page
>       0.26 ± 14%      -0.1        0.16 ± 17%  perf-profile.children.cycles-pp.vfs_fstatat
>       0.51            -0.1        0.41 ±  3%  perf-profile.children.cycles-pp.native_irq_return_iret
>       0.31 ±  3%      -0.1        0.21 ±  6%  perf-profile.children.cycles-pp.__mod_memcg_lruvec_state
>       0.34 ±  3%      -0.1        0.24 ±  5%  perf-profile.children.cycles-pp.cgroup_rstat_updated
>       0.23 ± 13%      -0.1        0.13 ± 14%  perf-profile.children.cycles-pp.maps__clone
>       0.34 ±  2%      -0.1        0.24 ±  3%  perf-profile.children.cycles-pp._find_next_bit
>       0.35            -0.1        0.25 ±  4%  perf-profile.children.cycles-pp.update_load_avg
>       0.24 ± 14%      -0.1        0.14 ± 18%  perf-profile.children.cycles-pp.vfs_statx
>       0.30 ±  3%      -0.1        0.21 ±  5%  perf-profile.children.cycles-pp.memcg_account_kmem
>       0.43 ±  2%      -0.1        0.34 ±  3%  perf-profile.children.cycles-pp.proc_invalidate_siblings_dcache
>       0.35 ±  2%      -0.1        0.26 ±  3%  perf-profile.children.cycles-pp.mtree_range_walk
>       0.22 ± 15%      -0.1        0.13 ± 18%  perf-profile.children.cycles-pp.filename_lookup
>       0.32 ±  3%      -0.1        0.23 ±  6%  perf-profile.children.cycles-pp.__lookup_slow
>       0.28            -0.1        0.19 ±  5%  perf-profile.children.cycles-pp.__anon_vma_interval_tree_augment_rotate
>       0.22 ± 15%      -0.1        0.13 ± 18%  perf-profile.children.cycles-pp.path_lookupat
>       0.29 ±  2%      -0.1        0.21 ±  5%  perf-profile.children.cycles-pp.activate_task
>       0.29 ±  2%      -0.1        0.21 ±  7%  perf-profile.children.cycles-pp.refill_obj_stock
>       0.26            -0.1        0.18 ±  3%  perf-profile.children.cycles-pp.dequeue_entity
>       0.21 ±  3%      -0.1        0.13 ±  6%  perf-profile.children.cycles-pp.free_unref_page
>       0.37 ±  2%      -0.1        0.28 ±  3%  perf-profile.children.cycles-pp.rmqueue
>       0.28 ±  2%      -0.1        0.20 ±  5%  perf-profile.children.cycles-pp.enqueue_task_fair
>       0.32            -0.1        0.24 ±  2%  perf-profile.children.cycles-pp.lookup_open
>       0.36 ±  2%      -0.1        0.28 ±  3%  perf-profile.children.cycles-pp.__put_user_4
>       0.24            -0.1        0.16 ±  7%  perf-profile.children.cycles-pp.new_inode
>       0.26            -0.1        0.18 ±  6%  perf-profile.children.cycles-pp.proc_pid_make_inode
>       0.31 ±  2%      -0.1        0.24 ±  3%  perf-profile.children.cycles-pp.mas_walk
>       0.30 ±  5%      -0.1        0.22 ±  9%  perf-profile.children.cycles-pp.__wp_page_copy_user
>       0.16 ±  3%      -0.1        0.08 ±  8%  perf-profile.children.cycles-pp.free_pcppages_bulk
>       0.44            -0.1        0.36 ±  3%  perf-profile.children.cycles-pp.__nptl_set_robust
>       0.29 ±  5%      -0.1        0.22 ±  8%  perf-profile.children.cycles-pp.copy_mc_enhanced_fast_string
>       0.26 ±  3%      -0.1        0.19 ± 10%  perf-profile.children.cycles-pp.dput
>       0.24 ±  6%      -0.1        0.17 ±  5%  perf-profile.children.cycles-pp.__mod_lruvec_page_state
>       0.30            -0.1        0.23 ±  3%  perf-profile.children.cycles-pp.__read_nocancel
>       0.18 ± 13%      -0.1        0.11 ± 21%  perf-profile.children.cycles-pp.__pud_alloc
>       0.25            -0.1        0.18 ±  3%  perf-profile.children.cycles-pp.sched_move_task
>       0.23 ±  2%      -0.1        0.16 ±  5%  perf-profile.children.cycles-pp.rcu_all_qs
>       0.18 ±  2%      -0.1        0.12 ±  8%  perf-profile.children.cycles-pp.__list_add_valid
>       0.30            -0.1        0.23 ±  4%  perf-profile.children.cycles-pp.put_cred_rcu
>       0.15 ±  6%      -0.1        0.08 ±  8%  perf-profile.children.cycles-pp.mark_page_accessed
>       0.30            -0.1        0.23 ±  2%  perf-profile.children.cycles-pp.ksys_read
>       0.29            -0.1        0.22 ±  3%  perf-profile.children.cycles-pp.vfs_read
>       0.23 ±  5%      -0.1        0.16 ±  5%  perf-profile.children.cycles-pp.__get_obj_cgroup_from_memcg
>       0.30            -0.1        0.23 ±  4%  perf-profile.children.cycles-pp.try_to_wake_up
>       0.29 ±  3%      -0.1        0.23 ±  4%  perf-profile.children.cycles-pp.__reclaim_stacks
>       0.30            -0.1        0.23 ±  4%  perf-profile.children.cycles-pp.down_read_trylock
>       0.22 ±  2%      -0.1        0.15 ±  7%  perf-profile.children.cycles-pp.cpu_util
>       0.26            -0.1        0.19 ±  3%  perf-profile.children.cycles-pp.seq_read_iter
>       0.22            -0.1        0.16 ±  4%  perf-profile.children.cycles-pp.d_alloc_parallel
>       0.21            -0.1        0.15 ±  6%  perf-profile.children.cycles-pp.enqueue_entity
>       0.21 ±  3%      -0.1        0.14 ±  8%  perf-profile.children.cycles-pp.folio_batch_move_lru
>       0.20            -0.1        0.14 ±  5%  perf-profile.children.cycles-pp.mas_leaf_max_gap
>       0.46 ±  2%      -0.1        0.40 ±  4%  perf-profile.children.cycles-pp.memset_orig
>       0.28            -0.1        0.22 ±  4%  perf-profile.children.cycles-pp.do_notify_parent
>       0.16 ±  3%      -0.1        0.11 ±  8%  perf-profile.children.cycles-pp.alloc_inode
>       0.16 ±  2%      -0.1        0.10 ±  4%  perf-profile.children.cycles-pp.kmem_cache_alloc_lru
>       0.20            -0.1        0.14 ±  6%  perf-profile.children.cycles-pp.vm_normal_page
>       0.19            -0.1        0.13 ±  4%  perf-profile.children.cycles-pp.sysvec_call_function_single
>       0.20 ±  4%      -0.1        0.14 ±  6%  perf-profile.children.cycles-pp.lru_add_drain
>       0.20 ±  3%      -0.1        0.14 ±  8%  perf-profile.children.cycles-pp.lru_add_drain_cpu
>       0.17            -0.1        0.12 ±  6%  perf-profile.children.cycles-pp.d_alloc
>       0.26            -0.1        0.21 ±  5%  perf-profile.children.cycles-pp.__wake_up_common_lock
>       0.22 ±  2%      -0.1        0.16 ±  4%  perf-profile.children.cycles-pp.proc_pident_lookup
>       0.12 ±  4%      -0.1        0.07 ±  5%  perf-profile.children.cycles-pp.rmqueue_bulk
>       0.09 ± 12%      -0.1        0.04 ± 71%  perf-profile.children.cycles-pp.maps__insert
>       0.20 ±  2%      -0.1        0.14 ±  6%  perf-profile.children.cycles-pp.__exit_signal
>       0.56            -0.1        0.51        perf-profile.children.cycles-pp.__sysvec_apic_timer_interrupt
>       0.56            -0.1        0.50        perf-profile.children.cycles-pp.hrtimer_interrupt
>       0.17            -0.1        0.12 ±  5%  perf-profile.children.cycles-pp.proc_pident_instantiate
>       0.29 ±  2%      -0.1        0.24 ±  3%  perf-profile.children.cycles-pp.d_invalidate
>       0.22 ±  2%      -0.1        0.17 ±  5%  perf-profile.children.cycles-pp.mab_mas_cp
>       0.18 ±  2%      -0.1        0.13 ±  7%  perf-profile.children.cycles-pp.mast_fill_bnode
>       0.24            -0.0        0.19 ±  4%  perf-profile.children.cycles-pp.schedule_idle
>       0.20 ±  2%      -0.0        0.14 ±  5%  perf-profile.children.cycles-pp.pcpu_alloc_area
>       0.14 ±  3%      -0.0        0.10 ±  5%  perf-profile.children.cycles-pp.mas_alloc_nodes
>       0.18 ±  2%      -0.0        0.14 ±  3%  perf-profile.children.cycles-pp.loadavg_proc_show
>       0.15 ±  4%      -0.0        0.10 ±  8%  perf-profile.children.cycles-pp.__unfreeze_partials
>       0.26            -0.0        0.22 ±  5%  perf-profile.children.cycles-pp.__memcpy
>       0.22 ±  9%      -0.0        0.17 ±  4%  perf-profile.children.cycles-pp.__pmd_alloc
>       0.21 ±  3%      -0.0        0.16 ±  5%  perf-profile.children.cycles-pp.lookup_fast
>       0.11 ±  3%      -0.0        0.06 ±  9%  perf-profile.children.cycles-pp.__free_one_page
>       0.14 ±  3%      -0.0        0.10 ±  7%  perf-profile.children.cycles-pp.mas_expected_entries
>       0.18 ±  2%      -0.0        0.14 ±  5%  perf-profile.children.cycles-pp.mas_split_final_node
>       0.51            -0.0        0.47        perf-profile.children.cycles-pp.__hrtimer_run_queues
>       0.13 ±  3%      -0.0        0.09 ±  8%  perf-profile.children.cycles-pp.__kmem_cache_alloc_bulk
>       0.26            -0.0        0.22 ±  4%  perf-profile.children.cycles-pp.shrink_dcache_parent
>       0.15 ±  2%      -0.0        0.10 ±  7%  perf-profile.children.cycles-pp.__tlb_remove_page_size
>       0.16 ±  3%      -0.0        0.11 ±  5%  perf-profile.children.cycles-pp.__flush_smp_call_function_queue
>       0.15 ±  4%      -0.0        0.11 ±  6%  perf-profile.children.cycles-pp.mas_find
>       0.14            -0.0        0.10 ±  5%  perf-profile.children.cycles-pp.kmem_cache_alloc_bulk
>       0.19 ±  2%      -0.0        0.14 ±  4%  perf-profile.children.cycles-pp.update_curr
>       0.17 ±  2%      -0.0        0.13 ±  3%  perf-profile.children.cycles-pp.prepare_creds
>       0.18 ±  6%      -0.0        0.14 ±  5%  perf-profile.children.cycles-pp.step_into
>       0.25            -0.0        0.20 ±  3%  perf-profile.children.cycles-pp.__wake_up_common
>       0.16 ±  2%      -0.0        0.12 ±  4%  perf-profile.children.cycles-pp.__sysvec_call_function_single
>       0.11            -0.0        0.07 ±  7%  perf-profile.children.cycles-pp.pcpu_free_area
>       0.07 ±  5%      -0.0        0.02 ± 99%  perf-profile.children.cycles-pp.uncharge_batch
>       0.07 ±  5%      -0.0        0.02 ± 99%  perf-profile.children.cycles-pp.__switch_to_asm
>       0.16 ±  4%      -0.0        0.11 ±  3%  perf-profile.children.cycles-pp.free_pgd_range
>       0.08 ± 13%      -0.0        0.04 ± 71%  perf-profile.children.cycles-pp.malloc
>       0.46            -0.0        0.42 ±  2%  perf-profile.children.cycles-pp.tick_sched_timer
>       0.08 ±  6%      -0.0        0.04 ± 71%  perf-profile.children.cycles-pp.__d_lookup_rcu
>       0.21 ±  2%      -0.0        0.17 ±  2%  perf-profile.children.cycles-pp.asm_sysvec_reschedule_ipi
>       0.07 ±  7%      -0.0        0.02 ± 99%  perf-profile.children.cycles-pp.switch_fpu_return
>       0.44            -0.0        0.40 ±  2%  perf-profile.children.cycles-pp.tick_sched_handle
>       0.44            -0.0        0.40        perf-profile.children.cycles-pp.update_process_times
>       0.20            -0.0        0.16 ±  4%  perf-profile.children.cycles-pp.__dentry_kill
>       0.14 ±  5%      -0.0        0.10 ±  4%  perf-profile.children.cycles-pp.free_p4d_range
>       0.14 ±  5%      -0.0        0.10 ±  4%  perf-profile.children.cycles-pp.free_pud_range
>       0.09 ±  5%      -0.0        0.05 ±  8%  perf-profile.children.cycles-pp.rcu_cblist_dequeue
>       0.14 ±  3%      -0.0        0.10 ±  5%  perf-profile.children.cycles-pp.sched_ttwu_pending
>       0.13 ±  3%      -0.0        0.09 ±  7%  perf-profile.children.cycles-pp.allocate_slab
>       0.17 ±  2%      -0.0        0.14 ±  5%  perf-profile.children.cycles-pp.proc_root_lookup
>       0.17 ±  2%      -0.0        0.14 ±  5%  perf-profile.children.cycles-pp.proc_pid_lookup
>       0.16 ±  2%      -0.0        0.12 ±  4%  perf-profile.children.cycles-pp.perf_iterate_sb
>       0.11 ±  4%      -0.0        0.08 ±  6%  perf-profile.children.cycles-pp.folio_mark_accessed
>       0.16 ±  2%      -0.0        0.12 ±  3%  perf-profile.children.cycles-pp.ttwu_do_activate
>       0.13            -0.0        0.09 ±  5%  perf-profile.children.cycles-pp.free_unref_page_list
>       0.11 ±  6%      -0.0        0.07 ±  5%  perf-profile.children.cycles-pp.__vmalloc_area_node
>       0.06 ±  6%      -0.0        0.02 ± 99%  perf-profile.children.cycles-pp.slab_pre_alloc_hook
>       0.06 ±  6%      -0.0        0.02 ± 99%  perf-profile.children.cycles-pp.entry_SYSRETQ_unsafe_stack
>       0.11 ±  4%      -0.0        0.08 ±  6%  perf-profile.children.cycles-pp.detach_tasks
>       0.36 ±  3%      -0.0        0.33 ±  3%  perf-profile.children.cycles-pp.unmap_single_vma
>       0.16 ±  3%      -0.0        0.13 ±  6%  perf-profile.children.cycles-pp.__get_free_pages
>       0.12 ±  4%      -0.0        0.08 ±  8%  perf-profile.children.cycles-pp.__kmem_cache_alloc_node
>       0.06            -0.0        0.02 ± 99%  perf-profile.children.cycles-pp.update_rq_clock
>       0.20 ±  4%      -0.0        0.16 ±  3%  perf-profile.children.cycles-pp.handle_signal
>       0.19 ±  3%      -0.0        0.16 ±  3%  perf-profile.children.cycles-pp.get_sigframe
>       0.15            -0.0        0.12 ±  4%  perf-profile.children.cycles-pp.alloc_empty_file
>       0.15 ±  8%      -0.0        0.12 ±  9%  perf-profile.children.cycles-pp.do_anonymous_page
>       0.10 ±  3%      -0.0        0.06 ±  7%  perf-profile.children.cycles-pp.__mem_cgroup_uncharge_list
>       0.16 ±  4%      -0.0        0.13 ±  6%  perf-profile.children.cycles-pp.pgd_alloc
>       0.12 ±  4%      -0.0        0.08 ±  8%  perf-profile.children.cycles-pp.__d_alloc
>       0.11            -0.0        0.08 ±  6%  perf-profile.children.cycles-pp.mas_mab_cp
>       0.13 ±  2%      -0.0        0.10 ±  5%  perf-profile.children.cycles-pp.__pte_offset_map
>       0.07 ±  6%      -0.0        0.04 ± 44%  perf-profile.children.cycles-pp.shuffle_freelist
>       0.07 ±  6%      -0.0        0.04 ± 44%  perf-profile.children.cycles-pp._find_next_and_bit
>       0.07 ±  6%      -0.0        0.04 ± 44%  perf-profile.children.cycles-pp._find_next_zero_bit
>       0.37 ±  2%      -0.0        0.34 ±  2%  perf-profile.children.cycles-pp.scheduler_tick
>       0.12 ±  5%      -0.0        0.09 ±  5%  perf-profile.children.cycles-pp.mas_wr_walk
>       0.12            -0.0        0.09 ±  4%  perf-profile.children.cycles-pp.alloc_pid
>       0.20 ±  4%      -0.0        0.16 ±  3%  perf-profile.children.cycles-pp.x64_setup_rt_frame
>       0.19            -0.0        0.16 ±  3%  perf-profile.children.cycles-pp.shrink_dentry_list
>       0.13 ±  6%      -0.0        0.10 ±  8%  perf-profile.children.cycles-pp.flush_tlb_func
>       0.14 ±  9%      -0.0        0.11 ±  6%  perf-profile.children.cycles-pp.pick_link
>       0.06 ±  7%      -0.0        0.03 ± 70%  perf-profile.children.cycles-pp.__mt_destroy
>       0.25            -0.0        0.22 ±  2%  perf-profile.children.cycles-pp.mas_topiary_replace
>       0.16 ±  3%      -0.0        0.12 ±  4%  perf-profile.children.cycles-pp.do_coredump
>       0.11 ±  4%      -0.0        0.08 ±  6%  perf-profile.children.cycles-pp.native_flush_tlb_one_user
>       0.10 ±  4%      -0.0        0.08 ±  6%  perf-profile.children.cycles-pp.__put_task_struct
>       0.10 ±  5%      -0.0        0.07 ±  7%  perf-profile.children.cycles-pp.mas_pop_node
>       0.09            -0.0        0.06 ±  6%  perf-profile.children.cycles-pp.raw_spin_rq_lock_nested
>       0.07            -0.0        0.04 ± 44%  perf-profile.children.cycles-pp.sched_cgroup_fork
>       0.07            -0.0        0.04 ± 44%  perf-profile.children.cycles-pp.__init_rwsem
>       0.07            -0.0        0.04 ± 44%  perf-profile.children.cycles-pp.security_cred_free
>       0.07 ±  6%      -0.0        0.04 ± 45%  perf-profile.children.cycles-pp.lockref_get_not_dead
>       0.06 ±  6%      -0.0        0.03 ± 70%  perf-profile.children.cycles-pp.cgroup_can_fork
>       0.06 ±  6%      -0.0        0.03 ± 70%  perf-profile.children.cycles-pp.cpu_clock_sample_group
>       0.21 ±  2%      -0.0        0.18 ±  3%  perf-profile.children.cycles-pp.flush_tlb_mm_range
>       0.14 ±  3%      -0.0        0.11 ±  6%  perf-profile.children.cycles-pp.rep_stos_alternative
>       0.10            -0.0        0.07 ±  6%  perf-profile.children.cycles-pp.copy_creds
>       0.06            -0.0        0.03 ± 70%  perf-profile.children.cycles-pp.update_rq_clock_task
>       0.14 ±  4%      -0.0        0.11 ±  6%  perf-profile.children.cycles-pp.copy_fpstate_to_sigframe
>       0.16 ±  2%      -0.0        0.13 ±  2%  perf-profile.children.cycles-pp.do_open
>       0.18 ±  3%      -0.0        0.16 ±  3%  perf-profile.children.cycles-pp._raw_spin_lock_irq
>       0.10 ±  4%      -0.0        0.07 ±  5%  perf-profile.children.cycles-pp.seq_printf
>       0.10 ±  4%      -0.0        0.07 ±  9%  perf-profile.children.cycles-pp.proc_pid_make_base_inode
>       0.08 ±  6%      -0.0        0.05 ±  7%  perf-profile.children.cycles-pp.free_unref_page_prepare
>       0.08 ±  5%      -0.0        0.06 ±  9%  perf-profile.children.cycles-pp.inode_init_always
>       0.08 ±  6%      -0.0        0.05 ±  7%  perf-profile.children.cycles-pp.account_kernel_stack
>       0.08 ±  4%      -0.0        0.06 ±  8%  perf-profile.children.cycles-pp.__alloc_pages_bulk
>       0.08 ±  5%      -0.0        0.06 ±  6%  perf-profile.children.cycles-pp.nr_running
>       0.10 ±  3%      -0.0        0.08 ±  6%  perf-profile.children.cycles-pp.proc_pid_instantiate
>       0.12            -0.0        0.10 ±  5%  perf-profile.children.cycles-pp.select_task_rq
>       0.10 ±  6%      -0.0        0.07 ±  6%  perf-profile.children.cycles-pp.obj_cgroup_uncharge_pages
>       0.22 ±  3%      -0.0        0.19 ±  3%  perf-profile.children.cycles-pp.vma_alloc_folio
>       0.13 ±  3%      -0.0        0.10 ±  4%  perf-profile.children.cycles-pp.vsnprintf
>       0.08 ±  4%      -0.0        0.06 ±  6%  perf-profile.children.cycles-pp.__mod_lruvec_state
>       0.08 ±  4%      -0.0        0.06 ±  9%  perf-profile.children.cycles-pp.get_partial_node
>       0.09 ±  5%      -0.0        0.07        perf-profile.children.cycles-pp.update_blocked_averages
>       0.14 ±  2%      -0.0        0.12 ±  4%  perf-profile.children.cycles-pp.__do_sys_prlimit64
>       0.07 ±  6%      -0.0        0.05 ±  7%  perf-profile.children.cycles-pp.mas_next_node
>       0.08 ±  6%      -0.0        0.06 ±  9%  perf-profile.children.cycles-pp.exit_task_stack_account
>       0.06 ±  7%      -0.0        0.04 ± 44%  perf-profile.children.cycles-pp.terminate_walk
>       0.11 ±  4%      -0.0        0.08 ±  5%  perf-profile.children.cycles-pp.thread_group_cputime
>       0.11 ±  3%      -0.0        0.09 ±  5%  perf-profile.children.cycles-pp.perf_event_task_output
>       0.09            -0.0        0.07 ± 10%  perf-profile.children.cycles-pp.stress_get_setting
>       0.08 ±  6%      -0.0        0.06 ±  8%  perf-profile.children.cycles-pp.lockref_put_return
>       0.11 ±  4%      -0.0        0.09 ±  5%  perf-profile.children.cycles-pp.do_prlimit
>       0.08 ±  5%      -0.0        0.06 ±  7%  perf-profile.children.cycles-pp.pte_offset_map_nolock
>       0.06 ±  6%      -0.0        0.04 ± 44%  perf-profile.children.cycles-pp.mprotect_fixup
>       0.07            -0.0        0.05        perf-profile.children.cycles-pp.__do_sys_sysinfo
>       0.07            -0.0        0.05        perf-profile.children.cycles-pp.add_device_randomness
>       0.11            -0.0        0.09        perf-profile.children.cycles-pp.dup_fd
>       0.08 ±  5%      -0.0        0.06 ±  7%  perf-profile.children.cycles-pp.__entry_text_start
>       0.18 ±  3%      -0.0        0.16 ±  4%  perf-profile.children.cycles-pp.__folio_alloc
>       0.13 ±  5%      -0.0        0.11 ±  3%  perf-profile.children.cycles-pp.lock_mm_and_find_vma
>       0.11 ±  6%      -0.0        0.09 ±  4%  perf-profile.children.cycles-pp.__count_memcg_events
>       0.08 ±  4%      -0.0        0.06 ± 13%  perf-profile.children.cycles-pp.lru_add_fn
>       0.07            -0.0        0.05 ±  7%  perf-profile.children.cycles-pp._find_next_or_bit
>       0.07            -0.0        0.05 ±  7%  perf-profile.children.cycles-pp.get_zeroed_page
>       0.08            -0.0        0.06 ±  6%  perf-profile.children.cycles-pp.init_file
>       0.08            -0.0        0.06 ± 11%  perf-profile.children.cycles-pp.put_files_struct
>       0.07 ±  5%      -0.0        0.05 ±  8%  perf-profile.children.cycles-pp.kmem_cache_free_bulk
>       0.07 ±  5%      -0.0        0.05        perf-profile.children.cycles-pp.menu_select
>       0.07 ±  5%      -0.0        0.05        perf-profile.children.cycles-pp.do_sysinfo
>       0.06            -0.0        0.04 ± 44%  perf-profile.children.cycles-pp.set_next_entity
>       0.06            -0.0        0.04 ± 44%  perf-profile.children.cycles-pp.mutex_unlock
>       0.16 ±  3%      -0.0        0.14 ±  3%  perf-profile.children.cycles-pp.__mem_cgroup_charge
>       0.09 ±  4%      -0.0        0.08 ±  6%  perf-profile.children.cycles-pp.copy_signal
>       0.07 ±  5%      -0.0        0.05 ±  7%  perf-profile.children.cycles-pp.perf_event_task
>       0.07 ±  5%      -0.0        0.06 ±  9%  perf-profile.children.cycles-pp.mt_find
>       0.07            -0.0        0.05 ±  8%  perf-profile.children.cycles-pp.perf_event_fork
>       0.18 ±  2%      -0.0        0.16 ±  3%  perf-profile.children.cycles-pp.try_charge_memcg
>       0.09            -0.0        0.07 ±  6%  perf-profile.children.cycles-pp.put_ucounts
>       0.09            -0.0        0.07 ±  6%  perf-profile.children.cycles-pp._atomic_dec_and_lock_irqsave
>       0.09            -0.0        0.07 ±  6%  perf-profile.children.cycles-pp.__x64_sys_mprotect
>       0.09            -0.0        0.07 ±  6%  perf-profile.children.cycles-pp.do_mprotect_pkey
>       0.08 ±  5%      -0.0        0.07 ±  7%  perf-profile.children.cycles-pp.__update_load_avg_cfs_rq
>       0.07 ±  7%      -0.0        0.05        perf-profile.children.cycles-pp.error_entry
>       0.07 ±  7%      -0.0        0.05        perf-profile.children.cycles-pp.__update_load_avg_se
>       0.07 ±  6%      -0.0        0.06 ±  6%  perf-profile.children.cycles-pp.kmem_cache_alloc_node
>       0.07 ±  5%      -0.0        0.05 ±  8%  perf-profile.children.cycles-pp.select_idle_sibling
>       0.07            -0.0        0.06 ±  9%  perf-profile.children.cycles-pp.__p4d_alloc
>       0.12 ±  3%      -0.0        0.11 ±  3%  perf-profile.children.cycles-pp.do_dentry_open
>       0.11 ±  3%      -0.0        0.10 ±  4%  perf-profile.children.cycles-pp.evict
>       0.08 ±  6%      -0.0        0.06        perf-profile.children.cycles-pp.mas_store_b_node
>       0.10 ±  5%      -0.0        0.08 ±  4%  perf-profile.children.cycles-pp.filp_close
>       0.15 ±  3%      -0.0        0.13 ±  3%  perf-profile.children.cycles-pp.__close_nocancel
>       0.11 ±  4%      -0.0        0.09 ±  5%  perf-profile.children.cycles-pp.up_read
>       0.15 ±  2%      -0.0        0.14 ±  2%  perf-profile.children.cycles-pp.arch_dup_task_struct
>       0.07 ±  5%      -0.0        0.06        perf-profile.children.cycles-pp.find_vma
>       0.06 ±  7%      -0.0        0.05 ±  8%  perf-profile.children.cycles-pp.update_rlimit_cpu
>       0.08 ±  6%      -0.0        0.06 ±  7%  perf-profile.children.cycles-pp.switch_mm_irqs_off
>       0.06 ±  6%      -0.0        0.05        perf-profile.children.cycles-pp.task_rq_lock
>       0.05            +0.0        0.07 ±  8%  perf-profile.children.cycles-pp.pmd_install
>       0.09 ±  9%      +0.0        0.12 ±  6%  perf-profile.children.cycles-pp.xas_find
>       0.05 ± 46%      +0.0        0.09 ±  9%  perf-profile.children.cycles-pp.xas_load
>       0.53 ±  3%      +0.0        0.58        perf-profile.children.cycles-pp.__pte_offset_map_lock
>       0.00            +0.1        0.07 ± 10%  perf-profile.children.cycles-pp.__errno_location
>       0.05            +0.1        0.12 ± 12%  perf-profile.children.cycles-pp.__do_fault
>       0.61            +0.1        0.68        perf-profile.children.cycles-pp.dup_userfaultfd
>       0.00            +0.1        0.08 ± 17%  perf-profile.children.cycles-pp.getloadavg
>       0.00            +0.1        0.09 ± 15%  perf-profile.children.cycles-pp.filemap_fault
>       0.79 ±  2%      +0.1        0.89        perf-profile.children.cycles-pp.page_add_file_rmap
>       0.00            +0.1        0.12 ±  9%  perf-profile.children.cycles-pp.__getpid
>       0.42 ±  4%      +0.3        0.68 ± 16%  perf-profile.children.cycles-pp.__sigsetjmp@plt
>       0.36 ±  5%      +0.3        0.64 ± 17%  perf-profile.children.cycles-pp.stress_align_address
>       0.42 ± 14%      +0.3        0.74 ± 15%  perf-profile.children.cycles-pp.shim_nanosleep_uint64
>       0.41 ±  4%      +1.7        2.12 ± 15%  perf-profile.children.cycles-pp.strlen@plt
>       1.81            +1.8        3.61 ±  4%  perf-profile.children.cycles-pp.do_sys_openat2
>       1.81            +1.8        3.62 ±  4%  perf-profile.children.cycles-pp.__x64_sys_openat
>       0.72 ±  2%      +2.0        2.72 ±  6%  perf-profile.children.cycles-pp.__open64_nocancel
>       0.56 ±  4%      +2.0        2.59 ±  6%  perf-profile.children.cycles-pp.setrlimit64
>       0.58 ±  3%      +2.0        2.63 ±  6%  perf-profile.children.cycles-pp.__sigsetjmp
>       0.44 ±  2%      +2.1        2.50 ±  7%  perf-profile.children.cycles-pp.getname_flags
>       0.41            +2.1        2.47 ±  7%  perf-profile.children.cycles-pp.strncpy_from_user
>       0.00            +4.0        3.99        perf-profile.children.cycles-pp.next_uptodate_folio
>      49.10           +17.6       66.70 ±  2%  perf-profile.children.cycles-pp.asm_exc_page_fault
>      11.96 ±  2%     +30.5       42.47 ±  6%  perf-profile.children.cycles-pp.exc_page_fault
>      11.81 ±  2%     +30.5       42.35 ±  6%  perf-profile.children.cycles-pp.do_user_addr_fault
>      10.45 ±  2%     +30.9       41.30 ±  6%  perf-profile.children.cycles-pp.handle_mm_fault
>       9.88 ±  2%     +31.2       41.04 ±  6%  perf-profile.children.cycles-pp.__handle_mm_fault
>       7.67 ±  2%     +31.2       38.92 ±  6%  perf-profile.children.cycles-pp.filemap_map_pages
>       8.00 ±  2%     +31.3       39.28 ±  6%  perf-profile.children.cycles-pp.do_fault
>       7.94 ±  2%     +31.3       39.24 ±  6%  perf-profile.children.cycles-pp.do_read_fault
>       8.37            -5.7        2.68 ± 11%  perf-profile.self.cycles-pp.osq_lock
>       4.34            -2.4        1.94 ±  9%  perf-profile.self.cycles-pp.rwsem_spin_on_owner
>       2.30            -1.4        0.90 ± 10%  perf-profile.self.cycles-pp.native_queued_spin_lock_slowpath
>       3.45 ±  3%      -1.4        2.08 ±  6%  perf-profile.self.cycles-pp.release_pages
>       4.10            -1.2        2.92 ±  3%  perf-profile.self.cycles-pp.up_write
>       3.61            -1.1        2.50 ±  4%  perf-profile.self.cycles-pp.anon_vma_interval_tree_insert
>       2.55 ±  7%      -1.0        1.56 ±  8%  perf-profile.self.cycles-pp._compound_head
>       2.53            -0.9        1.58 ±  7%  perf-profile.self.cycles-pp.zap_pte_range
>       3.25            -0.7        2.53 ±  3%  perf-profile.self.cycles-pp.down_write
>       1.13 ±  2%      -0.5        0.64 ±  6%  perf-profile.self.cycles-pp.anon_vma_clone
>       2.39            -0.5        1.92 ±  3%  perf-profile.self.cycles-pp.__slab_free
>       1.36            -0.5        0.91 ±  3%  perf-profile.self.cycles-pp.mod_objcg_state
>       1.56            -0.4        1.14 ±  5%  perf-profile.self.cycles-pp.memcg_slab_post_alloc_hook
>       1.26            -0.4        0.87 ±  3%  perf-profile.self.cycles-pp.kmem_cache_free
>       1.18            -0.4        0.81 ±  7%  perf-profile.self.cycles-pp.__anon_vma_interval_tree_remove
>       2.71 ±  3%      -0.4        2.35 ±  2%  perf-profile.self.cycles-pp.page_remove_rmap
>       1.52            -0.3        1.18 ±  3%  perf-profile.self.cycles-pp.dup_mmap
>       0.79 ±  2%      -0.3        0.44 ±  6%  perf-profile.self.cycles-pp.unlink_anon_vmas
>       0.90            -0.3        0.58 ±  7%  perf-profile.self.cycles-pp.vm_area_dup
>       0.82            -0.3        0.51 ±  5%  perf-profile.self.cycles-pp.free_swap_cache
>       0.56 ±  4%      -0.3        0.25 ±  9%  perf-profile.self.cycles-pp.rwsem_down_write_slowpath
>       0.48 ±  5%      -0.3        0.20 ±  4%  perf-profile.self.cycles-pp.___perf_sw_event
>       0.90            -0.3        0.63 ±  4%  perf-profile.self.cycles-pp.kmem_cache_alloc
>       0.68 ±  4%      -0.2        0.45 ±  6%  perf-profile.self.cycles-pp.osq_unlock
>       1.14 ±  6%      -0.2        0.92 ±  6%  perf-profile.self.cycles-pp.copy_present_pte
>       0.46            -0.2        0.25 ±  2%  perf-profile.self.cycles-pp.do_set_pte
>       0.70            -0.2        0.50 ±  6%  perf-profile.self.cycles-pp.fput
>       0.58            -0.2        0.38 ±  4%  perf-profile.self.cycles-pp.update_sg_lb_stats
>       0.62            -0.2        0.43 ±  5%  perf-profile.self.cycles-pp.mas_next_slot
>       0.54            -0.2        0.35 ±  2%  perf-profile.self.cycles-pp.anon_vma_fork
>       0.44 ±  4%      -0.2        0.26 ±  6%  perf-profile.self.cycles-pp.__rb_erase_color
>       0.53 ±  5%      -0.2        0.36 ±  6%  perf-profile.self.cycles-pp.clear_page_erms
>       0.53            -0.2        0.36 ±  4%  perf-profile.self.cycles-pp.percpu_counter_add_batch
>       0.71 ±  3%      -0.2        0.54 ±  2%  perf-profile.self.cycles-pp.vma_interval_tree_remove
>       0.51            -0.2        0.34 ±  5%  perf-profile.self.cycles-pp.__list_del_entry_valid
>       0.56            -0.2        0.40 ±  6%  perf-profile.self.cycles-pp.__percpu_counter_sum
>       0.39            -0.2        0.23 ±  4%  perf-profile.self.cycles-pp.__put_anon_vma
>       0.43            -0.2        0.27 ±  5%  perf-profile.self.cycles-pp.pcpu_alloc
>       0.44 ±  8%      -0.2        0.29 ±  7%  perf-profile.self.cycles-pp.rwsem_optimistic_spin
>       0.32 ±  2%      -0.1        0.18 ±  7%  perf-profile.self.cycles-pp.vma_interval_tree_insert_after
>       0.34 ±  5%      -0.1        0.20 ±  5%  perf-profile.self.cycles-pp.__rb_insert_augmented
>       0.53            -0.1        0.39 ±  4%  perf-profile.self.cycles-pp._raw_spin_lock_irqsave
>       0.54 ±  2%      -0.1        0.40 ±  4%  perf-profile.self.cycles-pp.sync_regs
>       0.22 ±  3%      -0.1        0.08 ±  8%  perf-profile.self.cycles-pp.mutex_spin_on_owner
>       0.43            -0.1        0.30 ±  4%  perf-profile.self.cycles-pp.__cond_resched
>       0.36            -0.1        0.24 ±  3%  perf-profile.self.cycles-pp.acct_collect
>       0.38            -0.1        0.26 ±  4%  perf-profile.self.cycles-pp.update_sg_wakeup_stats
>       0.30 ± 14%      -0.1        0.18 ± 22%  perf-profile.self.cycles-pp.copy_pte_range
>       0.51            -0.1        0.41 ±  3%  perf-profile.self.cycles-pp.native_irq_return_iret
>       0.21            -0.1        0.11 ±  9%  perf-profile.self.cycles-pp.queued_write_lock_slowpath
>       0.33 ±  3%      -0.1        0.23 ±  3%  perf-profile.self.cycles-pp.obj_cgroup_charge
>       0.38            -0.1        0.30 ±  2%  perf-profile.self.cycles-pp.acpi_safe_halt
>       0.34 ±  2%      -0.1        0.26 ±  3%  perf-profile.self.cycles-pp.mtree_range_walk
>       0.27            -0.1        0.19 ±  4%  perf-profile.self.cycles-pp.__anon_vma_interval_tree_augment_rotate
>       0.27            -0.1        0.19 ±  6%  perf-profile.self.cycles-pp.refill_obj_stock
>       0.23 ±  4%      -0.1        0.16 ±  4%  perf-profile.self.cycles-pp.__mod_memcg_lruvec_state
>       0.14 ±  3%      -0.1        0.07 ± 11%  perf-profile.self.cycles-pp.__vm_area_free
>       0.26 ±  2%      -0.1        0.18 ±  2%  perf-profile.self.cycles-pp.___slab_alloc
>       0.24 ±  2%      -0.1        0.17 ±  5%  perf-profile.self.cycles-pp.lock_vma_under_rcu
>       0.29 ±  4%      -0.1        0.22 ±  8%  perf-profile.self.cycles-pp.copy_mc_enhanced_fast_string
>       0.22 ±  4%      -0.1        0.15 ±  7%  perf-profile.self.cycles-pp.get_obj_cgroup_from_current
>       0.17 ±  2%      -0.1        0.11 ±  8%  perf-profile.self.cycles-pp.__list_add_valid
>       0.29            -0.1        0.22 ±  3%  perf-profile.self.cycles-pp.down_read_trylock
>       0.26 ±  2%      -0.1        0.19 ±  4%  perf-profile.self.cycles-pp.__libc_fork
>       0.20 ±  2%      -0.1        0.14 ±  8%  perf-profile.self.cycles-pp.cpu_util
>       0.19 ±  3%      -0.1        0.13 ±  5%  perf-profile.self.cycles-pp.mas_update_gap
>       0.12 ±  7%      -0.1        0.06 ± 11%  perf-profile.self.cycles-pp.mark_page_accessed
>       0.21 ±  3%      -0.1        0.15 ±  4%  perf-profile.self.cycles-pp.unmap_page_range
>       0.21 ±  5%      -0.1        0.15 ±  7%  perf-profile.self.cycles-pp.__get_obj_cgroup_from_memcg
>       0.22 ±  3%      -0.1        0.16 ±  5%  perf-profile.self.cycles-pp.memcg_account_kmem
>       0.18 ±  3%      -0.1        0.12 ±  6%  perf-profile.self.cycles-pp.update_load_avg
>       0.19            -0.1        0.14 ±  5%  perf-profile.self.cycles-pp.mas_leaf_max_gap
>       0.16 ±  2%      -0.1        0.11 ±  7%  perf-profile.self.cycles-pp.copy_page_range
>       0.12 ±  4%      -0.1        0.07 ±  9%  perf-profile.self.cycles-pp.queued_read_lock_slowpath
>       0.20 ±  6%      -0.1        0.15 ±  6%  perf-profile.self.cycles-pp.cgroup_rstat_updated
>       0.20 ±  2%      -0.1        0.14 ±  4%  perf-profile.self.cycles-pp._find_next_bit
>       0.10 ±  5%      -0.1        0.04 ± 45%  perf-profile.self.cycles-pp.__free_one_page
>       0.44 ±  2%      -0.1        0.39 ±  4%  perf-profile.self.cycles-pp.memset_orig
>       0.19 ±  2%      -0.0        0.14 ± 11%  perf-profile.self.cycles-pp.stress_bad_altstack
>       0.13 ±  8%      -0.0        0.08 ±  5%  perf-profile.self.cycles-pp.__mod_lruvec_page_state
>       0.16 ±  2%      -0.0        0.12 ±  6%  perf-profile.self.cycles-pp.rcu_all_qs
>       0.13 ±  4%      -0.0        0.09 ±  8%  perf-profile.self.cycles-pp.zap_pmd_range
>       0.07 ±  5%      -0.0        0.02 ± 99%  perf-profile.self.cycles-pp.__switch_to_asm
>       0.14 ±  3%      -0.0        0.10 ±  6%  perf-profile.self.cycles-pp.__pte_offset_map_lock
>       0.25 ±  2%      -0.0        0.21 ±  5%  perf-profile.self.cycles-pp.__memcpy
>       0.12 ±  4%      -0.0        0.07 ±  6%  perf-profile.self.cycles-pp.free_pgtables
>       0.08 ±  6%      -0.0        0.03 ± 70%  perf-profile.self.cycles-pp.__d_lookup_rcu
>       0.07 ±  7%      -0.0        0.02 ± 99%  perf-profile.self.cycles-pp._find_next_and_bit
>       0.14 ±  3%      -0.0        0.10 ±  4%  perf-profile.self.cycles-pp.mas_store
>       0.14 ±  3%      -0.0        0.10 ±  7%  perf-profile.self.cycles-pp.vm_normal_page
>       0.12 ±  4%      -0.0        0.08 ±  6%  perf-profile.self.cycles-pp.__kmem_cache_alloc_bulk
>       0.17 ±  2%      -0.0        0.14 ±  3%  perf-profile.self.cycles-pp.mab_mas_cp
>       0.07 ± 12%      -0.0        0.04 ± 71%  perf-profile.self.cycles-pp.malloc
>       0.09 ±  6%      -0.0        0.05 ±  8%  perf-profile.self.cycles-pp.rcu_cblist_dequeue
>       0.06 ±  6%      -0.0        0.02 ± 99%  perf-profile.self.cycles-pp.error_entry
>       0.07 ±  5%      -0.0        0.03 ± 70%  perf-profile.self.cycles-pp.free_pages_and_swap_cache
>       0.09            -0.0        0.06 ±  9%  perf-profile.self.cycles-pp.free_percpu
>       0.12 ±  4%      -0.0        0.08 ±  4%  perf-profile.self.cycles-pp.handle_mm_fault
>       0.09 ±  4%      -0.0        0.06 ±  6%  perf-profile.self.cycles-pp.remove_vma
>       0.10 ±  3%      -0.0        0.07 ±  7%  perf-profile.self.cycles-pp.mas_find
>       0.10 ±  5%      -0.0        0.06 ±  7%  perf-profile.self.cycles-pp.folio_mark_accessed
>       0.09 ±  5%      -0.0        0.06 ±  9%  perf-profile.self.cycles-pp.exit_mmap
>       0.25 ±  2%      -0.0        0.22 ±  3%  perf-profile.self.cycles-pp.mas_topiary_replace
>       0.09 ±  5%      -0.0        0.06 ±  8%  perf-profile.self.cycles-pp.__tlb_remove_page_size
>       0.11 ±  4%      -0.0        0.08 ±  5%  perf-profile.self.cycles-pp.mas_wr_walk
>       0.11 ±  4%      -0.0        0.08 ±  6%  perf-profile.self.cycles-pp.native_flush_tlb_one_user
>       0.09 ±  5%      -0.0        0.06 ±  7%  perf-profile.self.cycles-pp.mas_pop_node
>       0.08 ±  5%      -0.0        0.06 ±  8%  perf-profile.self.cycles-pp.mas_push_data
>       0.07            -0.0        0.04 ± 44%  perf-profile.self.cycles-pp._find_next_zero_bit
>       0.06 ±  6%      -0.0        0.03 ± 70%  perf-profile.self.cycles-pp.__sigsetjmp
>       0.08 ±  5%      -0.0        0.06 ±  9%  perf-profile.self.cycles-pp.__put_user_4
>       0.34 ±  4%      -0.0        0.32 ±  3%  perf-profile.self.cycles-pp.unmap_single_vma
>       0.10 ±  3%      -0.0        0.07 ±  6%  perf-profile.self.cycles-pp.mas_wr_store_entry
>       0.06            -0.0        0.03 ± 70%  perf-profile.self.cycles-pp.mas_wr_bnode
>       0.06            -0.0        0.03 ± 70%  perf-profile.self.cycles-pp.__reclaim_stacks
>       0.11 ±  3%      -0.0        0.09 ±  5%  perf-profile.self.cycles-pp.__pte_offset_map
>       0.09 ±  5%      -0.0        0.06 ±  6%  perf-profile.self.cycles-pp.mas_wr_append
>       0.11 ±  4%      -0.0        0.09 ±  4%  perf-profile.self.cycles-pp.update_curr
>       0.11 ±  3%      -0.0        0.09 ±  4%  perf-profile.self.cycles-pp.do_user_addr_fault
>       0.08            -0.0        0.06 ±  8%  perf-profile.self.cycles-pp.copy_process
>       0.07 ±  7%      -0.0        0.04 ± 45%  perf-profile.self.cycles-pp.prepare_creds
>       0.09 ±  4%      -0.0        0.06 ±  7%  perf-profile.self.cycles-pp.mas_mab_cp
>       0.12 ±  4%      -0.0        0.09 ±  4%  perf-profile.self.cycles-pp.__schedule
>       0.13 ±  3%      -0.0        0.10 ±  4%  perf-profile.self.cycles-pp.copy_p4d_range
>       0.08 ±  4%      -0.0        0.06 ±  8%  perf-profile.self.cycles-pp.__perf_sw_event
>       0.08 ±  4%      -0.0        0.06        perf-profile.self.cycles-pp.mm_init
>       0.08 ±  5%      -0.0        0.06 ±  7%  perf-profile.self.cycles-pp.mast_fill_bnode
>       0.09 ±  5%      -0.0        0.06 ±  7%  perf-profile.self.cycles-pp.stress_get_setting
>       0.08 ±  6%      -0.0        0.06 ±  8%  perf-profile.self.cycles-pp.lockref_put_return
>       0.09 ±  5%      -0.0        0.07 ±  7%  perf-profile.self.cycles-pp.asm_exc_page_fault
>       0.07            -0.0        0.05        perf-profile.self.cycles-pp.nr_running
>       0.06            -0.0        0.04 ± 44%  perf-profile.self.cycles-pp.strlen@plt
>       0.06            -0.0        0.04 ± 44%  perf-profile.self.cycles-pp.free_pud_range
>       0.06            -0.0        0.04 ± 44%  perf-profile.self.cycles-pp.mutex_unlock
>       0.14 ±  3%      -0.0        0.12 ±  5%  perf-profile.self.cycles-pp.perf_adjust_freq_unthr_context
>       0.08 ± 10%      -0.0        0.06 ±  6%  perf-profile.self.cycles-pp.rmqueue
>       0.09            -0.0        0.07 ±  6%  perf-profile.self.cycles-pp._atomic_dec_and_lock_irqsave
>       0.07 ±  7%      -0.0        0.05 ±  7%  perf-profile.self.cycles-pp.dup_fd
>       0.08 ±  5%      -0.0        0.07 ±  5%  perf-profile.self.cycles-pp.pcpu_alloc_area
>       0.08 ±  6%      -0.0        0.06 ±  7%  perf-profile.self.cycles-pp.__update_load_avg_cfs_rq
>       0.08 ±  6%      -0.0        0.06 ±  6%  perf-profile.self.cycles-pp.switch_mm_irqs_off
>       0.10 ±  4%      -0.0        0.09 ±  7%  perf-profile.self.cycles-pp.up_read
>       0.06 ±  6%      -0.0        0.05        perf-profile.self.cycles-pp.get_page_from_freelist
>       0.00            +0.1        0.05 ±  8%  perf-profile.self.cycles-pp.ptep_clear_flush
>       0.76 ±  2%      +0.1        0.87        perf-profile.self.cycles-pp.page_add_file_rmap
>       0.00            +3.8        3.80        perf-profile.self.cycles-pp.next_uptodate_folio
>       1.41 ±  2%     +31.7       33.08 ±  7%  perf-profile.self.cycles-pp.filemap_map_pages
> 
> 
> ***************************************************************************************************
> lkp-icl-2sp7: 64 threads 2 sockets Intel(R) Xeon(R) Gold 6346 CPU @ 3.10GHz (Ice Lake) with 256G memory
> =========================================================================================
> class/compiler/cpufreq_governor/kconfig/nr_threads/rootfs/tbox_group/test/testcase/testtime:
>   pts/gcc-12/performance/x86_64-rhel-8.3/100%/debian-11.1-x86_64-20220510.cgz/lkp-icl-2sp7/fork/stress-ng/60s
> 
> commit: 
>   9f1f5b60e7 ("mm: use flush_icache_pages() in do_set_pmd()")
>   de74976eb6 ("filemap: add filemap_map_folio_range()")
> 
> 9f1f5b60e76d44fa de74976eb65151a2f568e477fc2 
> ---------------- --------------------------- 
>          %stddev     %change         %stddev
>              \          |                \  
>    1393423 ±  3%     -20.9%    1102032 ±  2%  cpuidle..usage
>      76864           -12.8%      66991 ±  2%  vmstat.system.cs
>       0.00 ± 29%      +0.1        0.11 ±  7%  mpstat.cpu.all.iowait%
>       4.35            -0.4        3.93        mpstat.cpu.all.soft%
>     133293 ±  4%      -9.2%     120977 ±  6%  numa-meminfo.node1.SUnreclaim
>     471554 ± 13%     -25.7%     350336 ± 12%  numa-meminfo.node1.Shmem
>    2863944            -8.0%    2634003        meminfo.Inactive
>    2863779            -8.0%    2633836        meminfo.Inactive(anon)
>     267633 ±  3%      -7.5%     247668 ±  4%  meminfo.SUnreclaim
>     626825           -10.2%     562986        meminfo.Shmem
>   30299397           -16.5%   25290170 ±  4%  numa-numastat.node0.local_node
>   30574273           -16.5%   25525439 ±  4%  numa-numastat.node0.numa_hit
>     270840 ±  6%     -14.2%     232249 ± 14%  numa-numastat.node0.other_node
>   29636034 ±  2%     -13.3%   25688715 ±  6%  numa-numastat.node1.local_node
>   29942858 ±  2%     -13.3%   25954938 ±  5%  numa-numastat.node1.numa_hit
>   30580946           -16.5%   25534090 ±  4%  numa-vmstat.node0.numa_hit
>   30306070           -16.5%   25298821 ±  4%  numa-vmstat.node0.numa_local
>     270840 ±  6%     -14.2%     232249 ± 14%  numa-vmstat.node0.numa_other
>     118001 ± 13%     -25.7%      87690 ± 12%  numa-vmstat.node1.nr_shmem
>      33409 ±  4%      -9.5%      30222 ±  6%  numa-vmstat.node1.nr_slab_unreclaimable
>   29948612 ±  2%     -13.3%   25956024 ±  5%  numa-vmstat.node1.numa_hit
>   29641789 ±  2%     -13.3%   25689801 ±  6%  numa-vmstat.node1.numa_local
>    1701805           -13.3%    1474809        stress-ng.fork.ops
>      28363           -13.3%      24579        stress-ng.fork.ops_per_sec
>     196070 ±  2%     -20.8%     155373 ±  4%  stress-ng.time.involuntary_context_switches
>      14911 ±  6%    +964.6%     158745 ±  7%  stress-ng.time.major_page_faults
>  1.284e+08           -18.7%  1.045e+08 ± 11%  stress-ng.time.minor_page_faults
>       3119            +2.0%       3181        stress-ng.time.system_time
>     113.69 ±  2%     -32.8%      76.41 ±  2%  stress-ng.time.user_time
>    3159974            -7.1%    2934998        stress-ng.time.voluntary_context_switches
>       9213 ±  9%    +180.8%      25876 ±  4%  turbostat.C1
>       0.01 ± 31%      +0.0        0.05        turbostat.C1%
>    1042465 ±  3%     -32.7%     701201 ±  5%  turbostat.C1E
>       3.75 ±  2%      -0.5        3.28 ±  3%  turbostat.C1E%
>       0.10           -20.0%       0.08        turbostat.IPC
>       1650 ±  2%     +56.6%       2584 ±  9%  turbostat.POLL
>     230.66            -2.8%     224.15        turbostat.PkgWatt
>      75.92            -5.2%      71.96        turbostat.RAMWatt
>     637764 ±  2%     -12.3%     559588 ±  6%  sched_debug.cfs_rq:/.MIN_vruntime.max
>     263.91 ± 10%    +173.2%     721.05 ± 66%  sched_debug.cfs_rq:/.load_avg.avg
>     637764 ±  2%     -12.3%     559589 ±  6%  sched_debug.cfs_rq:/.max_vruntime.max
>     617832           -14.3%     529781 ±  3%  sched_debug.cfs_rq:/.min_vruntime.avg
>     439310 ±  3%     -15.9%     369572 ±  6%  sched_debug.cfs_rq:/.min_vruntime.min
>      93.37 ± 15%     -19.4%      75.29 ± 17%  sched_debug.cfs_rq:/.util_est_enqueued.avg
>       4.17 ± 17%     +34.3%       5.60 ±  9%  sched_debug.cpu.clock.stddev
>     814830           -13.3%     706518        sched_debug.cpu.curr->pid.max
>     403830           -14.7%     344526        sched_debug.cpu.curr->pid.stddev
>      40522           -12.6%      35397 ±  2%  sched_debug.cpu.nr_switches.avg
>      29933 ±  5%     -12.0%      26346 ±  7%  sched_debug.cpu.nr_switches.min
>      30.67 ± 22%     +52.2%      46.67 ± 16%  sched_debug.cpu.nr_uninterruptible.max
>     -28.92           +74.4%     -50.42        sched_debug.cpu.nr_uninterruptible.min
>      11.77 ±  5%     +59.2%      18.73 ± 12%  sched_debug.cpu.nr_uninterruptible.stddev
>     575349            -7.1%     534305        proc-vmstat.nr_anon_pages
>     842911            -1.9%     827005        proc-vmstat.nr_file_pages
>     716938            -8.0%     659721        proc-vmstat.nr_inactive_anon
>     130143            -4.3%     124519 ±  2%  proc-vmstat.nr_mapped
>       4924            -6.0%       4627 ±  2%  proc-vmstat.nr_page_table_pages
>     156822           -10.1%     140917        proc-vmstat.nr_shmem
>      66832 ±  3%      -6.9%      62207 ±  4%  proc-vmstat.nr_slab_unreclaimable
>     716938            -8.0%     659721        proc-vmstat.nr_zone_inactive_anon
>   60509078 ±  2%     -14.9%   51474061 ±  3%  proc-vmstat.numa_hit
>   59927380 ±  2%     -14.9%   50972568 ±  3%  proc-vmstat.numa_local
>     577989 ±  2%     -14.0%     496858 ±  3%  proc-vmstat.numa_other
>   66640119           -15.0%   56627849 ±  3%  proc-vmstat.pgalloc_normal
>  1.312e+08           -18.2%  1.073e+08 ± 10%  proc-vmstat.pgfault
>   64445803           -14.9%   54819989 ±  3%  proc-vmstat.pgfree
>   60119718           -22.7%   46481479 ± 25%  proc-vmstat.pgreuse
>  1.068e+10           -16.6%  8.899e+09 ±  2%  perf-stat.i.branch-instructions
>   75840999           -16.2%   63544754 ±  2%  perf-stat.i.branch-misses
>      30.59            -1.1       29.46        perf-stat.i.cache-miss-rate%
>   1.92e+08           -20.0%  1.535e+08 ±  2%  perf-stat.i.cache-misses
>  6.283e+08           -16.5%  5.247e+08 ±  2%  perf-stat.i.cache-references
>      79953           -15.1%      67856 ±  2%  perf-stat.i.context-switches
>       3.43           +19.7%       4.11 ±  2%  perf-stat.i.cpi
>      22388 ±  2%     -38.2%      13831 ±  6%  perf-stat.i.cpu-migrations
>     984.38           +23.2%       1212 ±  2%  perf-stat.i.cycles-between-cache-misses
>   19740484 ±  6%     -16.3%   16521409 ±  8%  perf-stat.i.dTLB-load-misses
>    1.3e+10           -15.6%  1.097e+10 ±  2%  perf-stat.i.dTLB-loads
>       0.24 ±  3%      -0.0        0.22 ±  6%  perf-stat.i.dTLB-store-miss-rate%
>   15956338 ±  2%     -19.3%   12872776 ±  7%  perf-stat.i.dTLB-store-misses
>  6.422e+09           -12.9%  5.594e+09        perf-stat.i.dTLB-stores
>  5.162e+10           -15.9%   4.34e+10 ±  2%  perf-stat.i.instructions
>       0.32           -14.6%       0.27 ±  2%  perf-stat.i.ipc
>     250.36 ±  7%   +1059.2%       2902 ±  7%  perf-stat.i.major-faults
>       2036           -15.2%       1727 ±  2%  perf-stat.i.metric.K/sec
>     480.18           -15.5%     405.54 ±  2%  perf-stat.i.metric.M/sec
>    2110044           -19.1%    1707218 ± 10%  perf-stat.i.minor-faults
>   80646029           -18.5%   65745733 ±  2%  perf-stat.i.node-load-misses
>   12640859 ±  3%     -12.5%   11055255 ±  2%  perf-stat.i.node-loads
>   50235371           -22.8%   38806291 ±  3%  perf-stat.i.node-store-misses
>   46512626 ±  2%     -20.9%   36805650 ±  3%  perf-stat.i.node-stores
>    2110294           -19.0%    1710120 ± 10%  perf-stat.i.page-faults
>      30.78            -1.0       29.76        perf-stat.overall.cache-miss-rate%
>       3.50           +17.8%       4.12 ±  2%  perf-stat.overall.cpi
>     928.07           +22.5%       1137 ±  2%  perf-stat.overall.cycles-between-cache-misses
>       0.25 ±  3%      -0.0        0.23 ±  6%  perf-stat.overall.dTLB-store-miss-rate%
>       0.29           -15.1%       0.24 ±  2%  perf-stat.overall.ipc
>      85.78            -0.9       84.87        perf-stat.overall.node-load-miss-rate%
>  1.017e+10           -15.5%  8.597e+09        perf-stat.ps.branch-instructions
>   71096858           -14.6%   60750696 ±  2%  perf-stat.ps.branch-misses
>  1.854e+08           -18.1%  1.519e+08 ±  2%  perf-stat.ps.cache-misses
>  6.023e+08           -15.3%  5.103e+08 ±  2%  perf-stat.ps.cache-references
>      77490           -13.6%      66934 ±  2%  perf-stat.ps.context-switches
>      21219 ±  2%     -33.9%      14030 ±  5%  perf-stat.ps.cpu-migrations
>   19795627 ±  6%     -16.1%   16616085 ±  7%  perf-stat.ps.dTLB-load-misses
>   1.24e+10           -14.4%  1.061e+10        perf-stat.ps.dTLB-loads
>   15378428 ±  2%     -18.5%   12530798 ±  6%  perf-stat.ps.dTLB-store-misses
>  6.158e+09           -11.8%  5.429e+09        perf-stat.ps.dTLB-stores
>  4.918e+10           -14.8%  4.191e+10        perf-stat.ps.instructions
>     232.74 ±  6%    +991.1%       2539 ±  7%  perf-stat.ps.major-faults
>    2027968           -18.2%    1657895 ± 10%  perf-stat.ps.minor-faults
>   77365251           -16.9%   64327762 ±  2%  perf-stat.ps.node-load-misses
>   12821005 ±  3%     -10.6%   11467882 ±  2%  perf-stat.ps.node-loads
>   47651364           -20.9%   37701874 ±  2%  perf-stat.ps.node-store-misses
>   45671351 ±  2%     -18.5%   37216427 ±  3%  perf-stat.ps.node-stores
>    2028200           -18.1%    1660434 ± 10%  perf-stat.ps.page-faults
>  3.018e+12           -15.4%  2.552e+12 ±  3%  perf-stat.total.instructions
>       0.08 ±  4%     -10.0%       0.07 ±  2%  perf-sched.sch_delay.avg.ms.__cond_resched.__kmem_cache_alloc_node.kmalloc_node_trace.alloc_fair_sched_group.sched_create_group
>       0.09 ±  5%     -22.9%       0.07 ±  8%  perf-sched.sch_delay.avg.ms.__cond_resched.__kmem_cache_alloc_node.kmalloc_trace.sched_autogroup_create_attach.ksys_setsid
>       0.05 ± 84%     -84.4%       0.01 ±102%  perf-sched.sch_delay.avg.ms.__cond_resched.__vmalloc_area_node.__vmalloc_node_range.alloc_thread_stack_node.dup_task_struct
>       0.06 ± 21%     -46.9%       0.03 ± 26%  perf-sched.sch_delay.avg.ms.__cond_resched.copy_pte_range.copy_p4d_range.copy_page_range.dup_mmap
>       0.03 ± 18%     -39.7%       0.02 ± 37%  perf-sched.sch_delay.avg.ms.__cond_resched.down_write.anon_vma_clone.anon_vma_fork.dup_mmap
>       0.04 ± 30%     -47.5%       0.02 ± 37%  perf-sched.sch_delay.avg.ms.__cond_resched.down_write.anon_vma_fork.dup_mmap.dup_mm
>       0.06 ± 57%     -61.6%       0.02 ± 34%  perf-sched.sch_delay.avg.ms.__cond_resched.down_write.dup_userfaultfd.dup_mmap.dup_mm
>       0.09 ±  6%    +109.2%       0.18 ±  8%  perf-sched.sch_delay.avg.ms.__cond_resched.lock_mm_and_find_vma.do_user_addr_fault.exc_page_fault.asm_exc_page_fault
>       0.07           -28.5%       0.05 ±  2%  perf-sched.sch_delay.avg.ms.do_task_dead.do_exit.do_group_exit.__x64_sys_exit_group.do_syscall_64
>       0.05           -28.3%       0.03 ±  6%  perf-sched.sch_delay.avg.ms.do_wait.kernel_wait4.__do_sys_wait4.do_syscall_64
>       0.05 ± 18%     -37.7%       0.03 ± 21%  perf-sched.sch_delay.avg.ms.exit_to_user_mode_loop.exit_to_user_mode_prepare.irqentry_exit_to_user_mode.asm_sysvec_apic_timer_interrupt
>       0.07 ± 23%     +73.4%       0.12 ± 26%  perf-sched.sch_delay.avg.ms.exit_to_user_mode_loop.exit_to_user_mode_prepare.irqentry_exit_to_user_mode.asm_sysvec_reschedule_ipi
>       0.05 ±  7%     -28.9%       0.03 ± 20%  perf-sched.sch_delay.avg.ms.exit_to_user_mode_loop.exit_to_user_mode_prepare.syscall_exit_to_user_mode.do_syscall_64
>       0.07 ± 21%    +234.5%       0.25 ± 12%  perf-sched.sch_delay.avg.ms.exit_to_user_mode_loop.exit_to_user_mode_prepare.syscall_exit_to_user_mode.ret_from_fork_asm
>       0.09 ± 15%     -46.7%       0.05 ±  2%  perf-sched.sch_delay.avg.ms.io_schedule.folio_wait_bit_common.filemap_fault.__do_fault
>       0.02 ± 19%     -42.7%       0.01 ± 42%  perf-sched.sch_delay.avg.ms.schedule_preempt_disabled.rwsem_down_write_slowpath.down_write.anon_vma_fork
>       0.07 ±  8%     +69.0%       0.11 ± 12%  perf-sched.sch_delay.avg.ms.schedule_timeout.rcu_gp_fqs_loop.rcu_gp_kthread.kthread
>       0.12 ±143%     -92.2%       0.01 ±110%  perf-sched.sch_delay.max.ms.__cond_resched.__vmalloc_area_node.__vmalloc_node_range.alloc_thread_stack_node.dup_task_struct
>       0.46 ± 32%     -62.3%       0.18 ± 79%  perf-sched.sch_delay.max.ms.__cond_resched.down_write.anon_vma_fork.dup_mmap.dup_mm
>       1.83 ± 35%     +46.7%       2.69 ±  8%  perf-sched.sch_delay.max.ms.__cond_resched.down_write_killable.exec_mmap.begin_new_exec.load_elf_binary
>       1.07 ± 29%     -58.9%       0.44 ± 58%  perf-sched.sch_delay.max.ms.schedule_preempt_disabled.rwsem_down_write_slowpath.down_write.anon_vma_fork
>       0.07 ±  4%     -26.5%       0.05 ±  4%  perf-sched.total_sch_delay.average.ms
>       3.98           +15.2%       4.59        perf-sched.total_wait_and_delay.average.ms
>     325106           -11.8%     286809 ±  2%  perf-sched.total_wait_and_delay.count.ms
>       3.91           +16.0%       4.53 ±  2%  perf-sched.total_wait_time.average.ms
>       1.37           +30.7%       1.79 ±  3%  perf-sched.wait_and_delay.avg.ms.do_wait.kernel_wait4.__do_sys_wait4.do_syscall_64
>      82.89          +113.4%     176.92 ± 10%  perf-sched.wait_and_delay.avg.ms.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
>      22.41 ±  3%    +105.6%      46.08 ± 10%  perf-sched.wait_and_delay.avg.ms.worker_thread.kthread.ret_from_fork.ret_from_fork_asm
>       3309 ±  3%     -75.0%     827.67 ±100%  perf-sched.wait_and_delay.count.__cond_resched.__kmem_cache_alloc_node.kmalloc_node_trace.alloc_fair_sched_group.sched_create_group
>      27.67 ± 17%     -38.0%      17.17 ± 26%  perf-sched.wait_and_delay.count.__cond_resched.generic_perform_write.generic_file_write_iter.vfs_write.ksys_write
>     140277           -12.3%     123044        perf-sched.wait_and_delay.count.do_task_dead.do_exit.do_group_exit.__x64_sys_exit_group.do_syscall_64
>     140531           -12.3%     123246        perf-sched.wait_and_delay.count.do_wait.kernel_wait4.__do_sys_wait4.do_syscall_64
>       5685 ±  3%     -17.0%       4721 ±  8%  perf-sched.wait_and_delay.count.exit_to_user_mode_loop.exit_to_user_mode_prepare.irqentry_exit_to_user_mode.asm_exc_page_fault
>       2160 ±  9%     -35.7%       1389 ± 45%  perf-sched.wait_and_delay.count.exit_to_user_mode_loop.exit_to_user_mode_prepare.syscall_exit_to_user_mode.do_syscall_64
>       7268           -52.9%       3422 ± 10%  perf-sched.wait_and_delay.count.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
>      14506 ±  3%     -51.5%       7037 ± 11%  perf-sched.wait_and_delay.count.worker_thread.kthread.ret_from_fork.ret_from_fork_asm
>       0.05 ± 33%    +178.6%       0.15 ± 23%  perf-sched.wait_time.avg.ms.__cond_resched.__alloc_pages.__folio_alloc.vma_alloc_folio.wp_page_copy
>       1.39 ±  7%     +40.7%       1.96 ± 13%  perf-sched.wait_time.avg.ms.__cond_resched.__alloc_pages.pte_alloc_one.__pte_alloc.copy_pte_range
>       1.26 ±  9%     +33.2%       1.68 ± 13%  perf-sched.wait_time.avg.ms.__cond_resched.__kmem_cache_alloc_node.__kmalloc.security_prepare_creds.prepare_creds
>       1.33 ± 10%     +46.2%       1.95 ± 11%  perf-sched.wait_time.avg.ms.__cond_resched.__kmem_cache_alloc_node.__kmalloc.security_task_alloc.copy_process
>       0.04 ±  4%     +29.5%       0.06 ±  6%  perf-sched.wait_time.avg.ms.__cond_resched.__kmem_cache_alloc_node.kmalloc_node_trace.alloc_fair_sched_group.sched_create_group
>       0.03 ± 10%     +48.4%       0.04 ± 19%  perf-sched.wait_time.avg.ms.__cond_resched.__kmem_cache_alloc_node.kmalloc_trace.sched_autogroup_create_attach.ksys_setsid
>       1.27 ±  6%     +37.9%       1.75 ±  7%  perf-sched.wait_time.avg.ms.__cond_resched.cgroup_css_set_fork.cgroup_can_fork.copy_process.kernel_clone
>       1.32 ±  3%     +32.3%       1.75 ±  3%  perf-sched.wait_time.avg.ms.__cond_resched.copy_page_range.dup_mmap.dup_mm.constprop
>       1.34 ±  2%     +30.2%       1.75 ±  4%  perf-sched.wait_time.avg.ms.__cond_resched.copy_pte_range.copy_p4d_range.copy_page_range.dup_mmap
>       1.28 ±  3%     +36.0%       1.75 ±  6%  perf-sched.wait_time.avg.ms.__cond_resched.down_write.anon_vma_clone.anon_vma_fork.dup_mmap
>       1.30 ±  2%     +37.5%       1.78 ±  6%  perf-sched.wait_time.avg.ms.__cond_resched.down_write.anon_vma_fork.dup_mmap.dup_mm
>       1.31           +34.1%       1.75 ±  4%  perf-sched.wait_time.avg.ms.__cond_resched.down_write.dup_mmap.dup_mm.constprop
>       1.32 ±  4%     +35.1%       1.78 ±  3%  perf-sched.wait_time.avg.ms.__cond_resched.down_write.dup_userfaultfd.dup_mmap.dup_mm
>       0.03 ± 48%    +343.4%       0.12 ± 63%  perf-sched.wait_time.avg.ms.__cond_resched.dput.__fput.task_work_run.exit_to_user_mode_loop
>       1.25 ± 13%     +41.9%       1.77 ±  8%  perf-sched.wait_time.avg.ms.__cond_resched.kmem_cache_alloc.alloc_pid.copy_process.kernel_clone
>       1.31           +33.4%       1.75 ±  4%  perf-sched.wait_time.avg.ms.__cond_resched.kmem_cache_alloc.anon_vma_fork.dup_mmap.dup_mm
>       1.30 ± 13%     +40.1%       1.82 ±  6%  perf-sched.wait_time.avg.ms.__cond_resched.kmem_cache_alloc.copy_fs_struct.copy_process.kernel_clone
>       1.24 ± 18%     +58.8%       1.97 ± 14%  perf-sched.wait_time.avg.ms.__cond_resched.kmem_cache_alloc.copy_signal.copy_process.kernel_clone
>       1.29 ±  4%     +32.8%       1.72 ± 10%  perf-sched.wait_time.avg.ms.__cond_resched.kmem_cache_alloc.dup_mm.constprop.0
>       1.32           +34.9%       1.78 ±  3%  perf-sched.wait_time.avg.ms.__cond_resched.kmem_cache_alloc.vm_area_dup.dup_mmap.dup_mm
>       1.30 ±  2%     +29.8%       1.69 ±  3%  perf-sched.wait_time.avg.ms.__cond_resched.mutex_lock_killable.pcpu_alloc.__percpu_counter_init.mm_init
>       1.41 ±  5%     +24.3%       1.76 ± 10%  perf-sched.wait_time.avg.ms.__cond_resched.mutex_lock_killable.pcpu_alloc.mm_init.dup_mm
>       1.27 ±  3%     +38.8%       1.76 ± 10%  perf-sched.wait_time.avg.ms.__cond_resched.uprobe_start_dup_mmap.dup_mmap.dup_mm.constprop
>       0.23 ±  5%     +16.9%       0.26 ±  4%  perf-sched.wait_time.avg.ms.do_task_dead.do_exit.do_group_exit.__x64_sys_exit_group.do_syscall_64
>       1.33           +32.7%       1.76 ±  3%  perf-sched.wait_time.avg.ms.do_wait.kernel_wait4.__do_sys_wait4.do_syscall_64
>       0.04 ± 24%    +556.4%       0.25 ± 21%  perf-sched.wait_time.avg.ms.exit_to_user_mode_loop.exit_to_user_mode_prepare.irqentry_exit_to_user_mode.asm_sysvec_call_function_single
>       0.08 ± 57%    +174.5%       0.21 ± 22%  perf-sched.wait_time.avg.ms.exit_to_user_mode_loop.exit_to_user_mode_prepare.irqentry_exit_to_user_mode.asm_sysvec_reschedule_ipi
>       0.01 ± 84%    +683.3%       0.04 ± 28%  perf-sched.wait_time.avg.ms.exit_to_user_mode_loop.exit_to_user_mode_prepare.syscall_exit_to_user_mode.ret_from_fork_asm
>       0.02 ± 20%     +38.8%       0.03 ±  4%  perf-sched.wait_time.avg.ms.io_schedule.folio_wait_bit_common.filemap_fault.__do_fault
>       0.89 ±  8%     +39.8%       1.24 ± 15%  perf-sched.wait_time.avg.ms.schedule_preempt_disabled.__mutex_lock.constprop.0.pcpu_alloc
>       0.55 ±  5%     +38.9%       0.76 ± 13%  perf-sched.wait_time.avg.ms.schedule_preempt_disabled.rwsem_down_write_slowpath.down_write.anon_vma_fork
>      82.82          +113.5%     176.85 ± 10%  perf-sched.wait_time.avg.ms.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
>      22.11 ±  3%    +107.0%      45.76 ± 10%  perf-sched.wait_time.avg.ms.worker_thread.kthread.ret_from_fork.ret_from_fork_asm
>       1.35 ± 18%     +43.6%       1.94 ± 26%  perf-sched.wait_time.max.ms.__cond_resched.__mutex_lock.constprop.0.pcpu_alloc
>       0.53 ± 72%    +221.3%       1.72 ± 34%  perf-sched.wait_time.max.ms.__cond_resched.down_write_killable.exec_mmap.begin_new_exec.load_elf_binary
>       0.26 ±141%   +1052.5%       2.96 ± 83%  perf-sched.wait_time.max.ms.__cond_resched.dput.__fput.task_work_run.exit_to_user_mode_loop
>       2.12 ±  9%     +49.7%       3.18 ± 19%  perf-sched.wait_time.max.ms.__cond_resched.uprobe_start_dup_mmap.dup_mmap.dup_mm.constprop
>       1.28 ± 21%    +403.9%       6.45 ±101%  perf-sched.wait_time.max.ms.exit_to_user_mode_loop.exit_to_user_mode_prepare.irqentry_exit_to_user_mode.asm_sysvec_call_function_single
>       0.58 ± 73%    +168.7%       1.56 ± 21%  perf-sched.wait_time.max.ms.exit_to_user_mode_loop.exit_to_user_mode_prepare.syscall_exit_to_user_mode.ret_from_fork_asm
>       1.30 ± 32%    +151.4%       3.27 ± 19%  perf-sched.wait_time.max.ms.io_schedule.folio_wait_bit_common.filemap_fault.__do_fault
>       4.35 ± 11%     +68.9%       7.35 ± 22%  perf-sched.wait_time.max.ms.schedule_timeout.__wait_for_common.wait_for_completion_state.kernel_clone
>      35.78            -8.5       27.26 ±  2%  perf-profile.calltrace.cycles-pp.do_exit.do_group_exit.__x64_sys_exit_group.do_syscall_64.entry_SYSCALL_64_after_hwframe
>      35.19            -8.4       26.83 ±  3%  perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.__libc_fork
>      35.18            -8.4       26.82 ±  3%  perf-profile.calltrace.cycles-pp.__do_sys_clone.do_syscall_64.entry_SYSCALL_64_after_hwframe.__libc_fork
>      35.19            -8.4       26.83 ±  3%  perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.__libc_fork
>      35.18            -8.4       26.82 ±  3%  perf-profile.calltrace.cycles-pp.kernel_clone.__do_sys_clone.do_syscall_64.entry_SYSCALL_64_after_hwframe.__libc_fork
>      34.32            -8.2       26.14 ±  3%  perf-profile.calltrace.cycles-pp.copy_process.kernel_clone.__do_sys_clone.do_syscall_64.entry_SYSCALL_64_after_hwframe
>      35.26            -8.0       27.27 ±  2%  perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe
>      35.26            -8.0       27.27 ±  2%  perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe
>      35.24            -8.0       27.26 ±  2%  perf-profile.calltrace.cycles-pp.__x64_sys_exit_group.do_syscall_64.entry_SYSCALL_64_after_hwframe
>      35.24            -8.0       27.26 ±  2%  perf-profile.calltrace.cycles-pp.do_group_exit.__x64_sys_exit_group.do_syscall_64.entry_SYSCALL_64_after_hwframe
>      32.41            -7.6       24.78 ±  3%  perf-profile.calltrace.cycles-pp.dup_mm.copy_process.kernel_clone.__do_sys_clone.do_syscall_64
>      33.53            -7.5       26.02 ±  2%  perf-profile.calltrace.cycles-pp.exit_mmap.__mmput.exit_mm.do_exit.do_group_exit
>      33.64            -7.5       26.13 ±  2%  perf-profile.calltrace.cycles-pp.exit_mm.do_exit.do_group_exit.__x64_sys_exit_group.do_syscall_64
>      33.63            -7.5       26.12 ±  2%  perf-profile.calltrace.cycles-pp.__mmput.exit_mm.do_exit.do_group_exit.__x64_sys_exit_group
>      36.99            -7.4       29.61 ±  2%  perf-profile.calltrace.cycles-pp.__libc_fork
>      29.90            -7.0       22.86 ±  3%  perf-profile.calltrace.cycles-pp.dup_mmap.dup_mm.copy_process.kernel_clone.__do_sys_clone
>      15.68            -4.1       11.56 ±  4%  perf-profile.calltrace.cycles-pp.anon_vma_fork.dup_mmap.dup_mm.copy_process.kernel_clone
>      12.53            -2.9        9.59 ±  3%  perf-profile.calltrace.cycles-pp.free_pgtables.exit_mmap.__mmput.exit_mm.do_exit
>      13.12            -2.9       10.26 ±  2%  perf-profile.calltrace.cycles-pp.unmap_vmas.exit_mmap.__mmput.exit_mm.do_exit
>      12.54            -2.7        9.84 ±  2%  perf-profile.calltrace.cycles-pp.unmap_page_range.unmap_vmas.exit_mmap.__mmput.exit_mm
>      12.40            -2.7        9.72 ±  2%  perf-profile.calltrace.cycles-pp.zap_pmd_range.unmap_page_range.unmap_vmas.exit_mmap.__mmput
>      12.14            -2.6        9.50        perf-profile.calltrace.cycles-pp.zap_pte_range.zap_pmd_range.unmap_page_range.unmap_vmas.exit_mmap
>      10.06            -2.5        7.60 ±  3%  perf-profile.calltrace.cycles-pp.anon_vma_clone.anon_vma_fork.dup_mmap.dup_mm.copy_process
>       8.72            -2.3        6.41 ±  4%  perf-profile.calltrace.cycles-pp.unlink_anon_vmas.free_pgtables.exit_mmap.__mmput.exit_mm
>       5.30            -1.5        3.84 ±  3%  perf-profile.calltrace.cycles-pp._compound_head.zap_pte_range.zap_pmd_range.unmap_page_range.unmap_vmas
>       4.55 ±  2%      -1.5        3.10 ±  5%  perf-profile.calltrace.cycles-pp.down_write.anon_vma_fork.dup_mmap.dup_mm.copy_process
>       4.40 ±  2%      -1.5        2.94 ±  6%  perf-profile.calltrace.cycles-pp.down_write.anon_vma_clone.anon_vma_fork.dup_mmap.dup_mm
>       4.12 ±  2%      -1.4        2.67 ±  7%  perf-profile.calltrace.cycles-pp.down_write.unlink_anon_vmas.free_pgtables.exit_mmap.__mmput
>       4.39 ±  2%      -1.4        2.95 ±  5%  perf-profile.calltrace.cycles-pp.rwsem_down_write_slowpath.down_write.anon_vma_fork.dup_mmap.dup_mm
>       4.33 ±  2%      -1.4        2.91 ±  5%  perf-profile.calltrace.cycles-pp.rwsem_optimistic_spin.rwsem_down_write_slowpath.down_write.anon_vma_fork.dup_mmap
>       4.14 ±  2%      -1.4        2.73 ±  6%  perf-profile.calltrace.cycles-pp.rwsem_down_write_slowpath.down_write.anon_vma_clone.anon_vma_fork.dup_mmap
>       4.10 ±  2%      -1.4        2.70 ±  6%  perf-profile.calltrace.cycles-pp.rwsem_optimistic_spin.rwsem_down_write_slowpath.down_write.anon_vma_clone.anon_vma_fork
>       3.78 ±  2%      -1.4        2.39 ±  7%  perf-profile.calltrace.cycles-pp.rwsem_down_write_slowpath.down_write.unlink_anon_vmas.free_pgtables.exit_mmap
>       3.74 ±  2%      -1.4        2.36 ±  8%  perf-profile.calltrace.cycles-pp.rwsem_optimistic_spin.rwsem_down_write_slowpath.down_write.unlink_anon_vmas.free_pgtables
>       5.39            -1.3        4.07 ±  3%  perf-profile.calltrace.cycles-pp.copy_page_range.dup_mmap.dup_mm.copy_process.kernel_clone
>       5.30            -1.3        3.99 ±  3%  perf-profile.calltrace.cycles-pp.copy_p4d_range.copy_page_range.dup_mmap.dup_mm.copy_process
>       5.26            -1.3        4.01 ±  3%  perf-profile.calltrace.cycles-pp.tlb_finish_mmu.exit_mmap.__mmput.exit_mm.do_exit
>       5.22            -1.2        3.97 ±  3%  perf-profile.calltrace.cycles-pp.tlb_batch_pages_flush.tlb_finish_mmu.exit_mmap.__mmput.exit_mm
>       4.83 ±  2%      -1.2        3.62 ±  3%  perf-profile.calltrace.cycles-pp.copy_pte_range.copy_p4d_range.copy_page_range.dup_mmap.dup_mm
>       2.58 ±  3%      -1.1        1.53 ± 10%  perf-profile.calltrace.cycles-pp.osq_lock.rwsem_optimistic_spin.rwsem_down_write_slowpath.down_write.unlink_anon_vmas
>       2.72 ±  3%      -1.0        1.67 ±  7%  perf-profile.calltrace.cycles-pp.osq_lock.rwsem_optimistic_spin.rwsem_down_write_slowpath.down_write.anon_vma_clone
>       4.17            -1.0        3.15 ±  3%  perf-profile.calltrace.cycles-pp.release_pages.tlb_batch_pages_flush.tlb_finish_mmu.exit_mmap.__mmput
>       2.78 ±  2%      -1.0        1.80 ±  6%  perf-profile.calltrace.cycles-pp.osq_lock.rwsem_optimistic_spin.rwsem_down_write_slowpath.down_write.anon_vma_fork
>       3.54            -0.9        2.64 ±  3%  perf-profile.calltrace.cycles-pp.copy_present_pte.copy_pte_range.copy_p4d_range.copy_page_range.dup_mmap
>       2.91            -0.8        2.07 ±  4%  perf-profile.calltrace.cycles-pp.wait4
>       2.88            -0.8        2.05 ±  4%  perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.wait4
>       2.87            -0.8        2.04 ±  4%  perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.wait4
>       2.83            -0.8        2.01 ±  4%  perf-profile.calltrace.cycles-pp.__do_sys_wait4.do_syscall_64.entry_SYSCALL_64_after_hwframe.wait4
>       2.82            -0.8        2.00 ±  4%  perf-profile.calltrace.cycles-pp.kernel_wait4.__do_sys_wait4.do_syscall_64.entry_SYSCALL_64_after_hwframe.wait4
>       2.77            -0.8        1.96 ±  4%  perf-profile.calltrace.cycles-pp.do_wait.kernel_wait4.__do_sys_wait4.do_syscall_64.entry_SYSCALL_64_after_hwframe
>       8.13            -0.8        7.37 ±  2%  perf-profile.calltrace.cycles-pp.setsid
>       8.11            -0.8        7.35 ±  2%  perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.setsid
>       8.11            -0.8        7.35 ±  2%  perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.setsid
>       8.07            -0.8        7.32 ±  2%  perf-profile.calltrace.cycles-pp.__x64_sys_setsid.do_syscall_64.entry_SYSCALL_64_after_hwframe.setsid
>       8.07            -0.8        7.32 ±  2%  perf-profile.calltrace.cycles-pp.ksys_setsid.__x64_sys_setsid.do_syscall_64.entry_SYSCALL_64_after_hwframe.setsid
>       1.74            -0.7        1.08 ±  5%  perf-profile.calltrace.cycles-pp._compound_head.copy_present_pte.copy_pte_range.copy_p4d_range.copy_page_range
>       7.80            -0.6        7.18        perf-profile.calltrace.cycles-pp.sched_autogroup_create_attach.ksys_setsid.__x64_sys_setsid.do_syscall_64.entry_SYSCALL_64_after_hwframe
>       2.50            -0.6        1.90 ±  2%  perf-profile.calltrace.cycles-pp.vm_area_dup.dup_mmap.dup_mm.copy_process.kernel_clone
>       3.58            -0.6        3.01 ±  2%  perf-profile.calltrace.cycles-pp.anon_vma_interval_tree_insert.anon_vma_clone.anon_vma_fork.dup_mmap.dup_mm
>       2.42 ±  2%      -0.6        1.86 ±  4%  perf-profile.calltrace.cycles-pp.mm_init.dup_mm.copy_process.kernel_clone.__do_sys_clone
>       0.80            -0.5        0.26 ±100%  perf-profile.calltrace.cycles-pp.dup_task_struct.copy_process.kernel_clone.__do_sys_clone.do_syscall_64
>       2.81            -0.5        2.30 ±  3%  perf-profile.calltrace.cycles-pp.online_fair_sched_group.sched_autogroup_create_attach.ksys_setsid.__x64_sys_setsid.do_syscall_64
>       1.92 ±  9%      -0.4        1.50 ±  6%  perf-profile.calltrace.cycles-pp.down_write.dup_mmap.dup_mm.copy_process.kernel_clone
>       1.31            -0.4        0.90 ±  4%  perf-profile.calltrace.cycles-pp.rwsem_spin_on_owner.rwsem_optimistic_spin.rwsem_down_write_slowpath.down_write.anon_vma_fork
>       2.30 ±  4%      -0.4        1.90 ±  5%  perf-profile.calltrace.cycles-pp.unlink_file_vma.free_pgtables.exit_mmap.__mmput.exit_mm
>       1.67            -0.4        1.29 ±  4%  perf-profile.calltrace.cycles-pp.schedule.do_wait.kernel_wait4.__do_sys_wait4.do_syscall_64
>       1.66            -0.4        1.28 ±  4%  perf-profile.calltrace.cycles-pp.__schedule.schedule.do_wait.kernel_wait4.__do_sys_wait4
>       1.68 ±  2%      -0.4        1.30 ±  4%  perf-profile.calltrace.cycles-pp.__percpu_counter_init.mm_init.dup_mm.copy_process.kernel_clone
>       1.52            -0.4        1.17 ±  2%  perf-profile.calltrace.cycles-pp.kmem_cache_alloc.vm_area_dup.dup_mmap.dup_mm.copy_process
>       3.53            -0.3        3.20        perf-profile.calltrace.cycles-pp.page_remove_rmap.zap_pte_range.zap_pmd_range.unmap_page_range.unmap_vmas
>       1.46 ±  3%      -0.3        1.13 ±  5%  perf-profile.calltrace.cycles-pp.pcpu_alloc.__percpu_counter_init.mm_init.dup_mm.copy_process
>       1.42 ±  7%      -0.3        1.10 ±  8%  perf-profile.calltrace.cycles-pp.down_write.unlink_file_vma.free_pgtables.exit_mmap.__mmput
>       1.18            -0.3        0.87 ±  5%  perf-profile.calltrace.cycles-pp.rwsem_spin_on_owner.rwsem_optimistic_spin.rwsem_down_write_slowpath.down_write.anon_vma_clone
>       1.00 ± 15%      -0.3        0.69 ± 13%  perf-profile.calltrace.cycles-pp.rwsem_down_write_slowpath.down_write.dup_mmap.dup_mm.copy_process
>       0.96 ± 15%      -0.3        0.66 ± 13%  perf-profile.calltrace.cycles-pp.rwsem_optimistic_spin.rwsem_down_write_slowpath.down_write.dup_mmap.dup_mm
>       0.97 ±  2%      -0.3        0.68 ±  4%  perf-profile.calltrace.cycles-pp.rwsem_spin_on_owner.rwsem_optimistic_spin.rwsem_down_write_slowpath.down_write.unlink_anon_vmas
>       4.07            -0.3        3.80        perf-profile.calltrace.cycles-pp.kmalloc_node_trace.alloc_fair_sched_group.sched_create_group.sched_autogroup_create_attach.ksys_setsid
>       4.04            -0.3        3.77        perf-profile.calltrace.cycles-pp.__kmem_cache_alloc_node.kmalloc_node_trace.alloc_fair_sched_group.sched_create_group.sched_autogroup_create_attach
>       0.90 ± 11%      -0.3        0.64 ± 13%  perf-profile.calltrace.cycles-pp.rwsem_down_write_slowpath.down_write.unlink_file_vma.free_pgtables.exit_mmap
>       0.86 ± 11%      -0.2        0.62 ± 13%  perf-profile.calltrace.cycles-pp.rwsem_optimistic_spin.rwsem_down_write_slowpath.down_write.unlink_file_vma.free_pgtables
>       1.87            -0.2        1.63        perf-profile.calltrace.cycles-pp.memcg_slab_post_alloc_hook.__kmem_cache_alloc_node.kmalloc_node_trace.alloc_fair_sched_group.sched_create_group
>       1.19            -0.2        0.96 ±  2%  perf-profile.calltrace.cycles-pp.__anon_vma_interval_tree_remove.unlink_anon_vmas.free_pgtables.exit_mmap.__mmput
>       1.00 ±  2%      -0.2        0.78 ±  3%  perf-profile.calltrace.cycles-pp.free_pages_and_swap_cache.tlb_batch_pages_flush.tlb_finish_mmu.exit_mmap.__mmput
>       0.98 ±  2%      -0.2        0.77 ±  2%  perf-profile.calltrace.cycles-pp.free_swap_cache.free_pages_and_swap_cache.tlb_batch_pages_flush.tlb_finish_mmu.exit_mmap
>       1.01            -0.2        0.81 ±  2%  perf-profile.calltrace.cycles-pp.raw_spin_rq_lock_nested.online_fair_sched_group.sched_autogroup_create_attach.ksys_setsid.__x64_sys_setsid
>       1.00            -0.2        0.80 ±  2%  perf-profile.calltrace.cycles-pp._raw_spin_lock.raw_spin_rq_lock_nested.online_fair_sched_group.sched_autogroup_create_attach.ksys_setsid
>       1.58            -0.2        1.39 ±  3%  perf-profile.calltrace.cycles-pp.___slab_alloc.__kmem_cache_alloc_node.kmalloc_node_trace.alloc_fair_sched_group.sched_create_group
>       0.86            -0.2        0.67 ±  2%  perf-profile.calltrace.cycles-pp.memcg_slab_post_alloc_hook.kmem_cache_alloc.vm_area_dup.dup_mmap.dup_mm
>       0.62 ±  2%      -0.2        0.43 ± 44%  perf-profile.calltrace.cycles-pp.__put_anon_vma.unlink_anon_vmas.free_pgtables.exit_mmap.__mmput
>       0.82            -0.2        0.64 ±  3%  perf-profile.calltrace.cycles-pp.wake_up_new_task.kernel_clone.__do_sys_clone.do_syscall_64.entry_SYSCALL_64_after_hwframe
>       0.69            -0.2        0.53 ±  2%  perf-profile.calltrace.cycles-pp.__vm_area_free.exit_mmap.__mmput.exit_mm.do_exit
>       0.72            -0.1        0.57 ±  3%  perf-profile.calltrace.cycles-pp.attach_entity_cfs_rq.online_fair_sched_group.sched_autogroup_create_attach.ksys_setsid.__x64_sys_setsid
>       1.00 ±  2%      -0.1        0.86 ±  3%  perf-profile.calltrace.cycles-pp.up_write.dup_mmap.dup_mm.copy_process.kernel_clone
>       0.73            -0.1        0.59 ±  2%  perf-profile.calltrace.cycles-pp.__slab_free.unlink_anon_vmas.free_pgtables.exit_mmap.__mmput
>       0.74 ±  2%      -0.1        0.61 ±  2%  perf-profile.calltrace.cycles-pp.remove_vma.exit_mmap.__mmput.exit_mm.do_exit
>       0.66 ±  2%      -0.1        0.55 ±  3%  perf-profile.calltrace.cycles-pp.fput.remove_vma.exit_mmap.__mmput.exit_mm
>       1.03            -0.1        0.92 ±  2%  perf-profile.calltrace.cycles-pp.mas_store.dup_mmap.dup_mm.copy_process.kernel_clone
>       0.88 ±  2%      -0.1        0.78 ±  2%  perf-profile.calltrace.cycles-pp.up_write.free_pgtables.exit_mmap.__mmput.exit_mm
>       0.69            -0.1        0.60 ±  3%  perf-profile.calltrace.cycles-pp.deactivate_slab.___slab_alloc.__kmem_cache_alloc_node.kmalloc_node_trace.alloc_fair_sched_group
>       0.61            -0.1        0.54 ±  3%  perf-profile.calltrace.cycles-pp.kmem_cache_alloc.anon_vma_clone.anon_vma_fork.dup_mmap.dup_mm
>       0.67            -0.1        0.60 ±  2%  perf-profile.calltrace.cycles-pp.update_rq_clock_task.online_fair_sched_group.sched_autogroup_create_attach.ksys_setsid.__x64_sys_setsid
>       0.62            -0.1        0.56        perf-profile.calltrace.cycles-pp.mas_wr_bnode.mas_store.dup_mmap.dup_mm.copy_process
>       0.79            -0.1        0.73 ±  2%  perf-profile.calltrace.cycles-pp.vma_interval_tree_remove.unlink_file_vma.free_pgtables.exit_mmap.__mmput
>       0.69 ±  9%      +0.3        0.94 ±  8%  perf-profile.calltrace.cycles-pp.stress_fork_fn
>       0.63 ± 10%      +0.3        0.89 ±  8%  perf-profile.calltrace.cycles-pp.asm_exc_page_fault.stress_fork_fn
>       0.58 ±  9%      +0.3        0.85 ±  8%  perf-profile.calltrace.cycles-pp.exc_page_fault.asm_exc_page_fault.stress_fork_fn
>       0.58 ±  9%      +0.3        0.85 ±  8%  perf-profile.calltrace.cycles-pp.do_user_addr_fault.exc_page_fault.asm_exc_page_fault.stress_fork_fn
>       0.45 ± 45%      +0.4        0.82 ±  9%  perf-profile.calltrace.cycles-pp.handle_mm_fault.do_user_addr_fault.exc_page_fault.asm_exc_page_fault.stress_fork_fn
>       0.00            +0.7        0.69 ± 10%  perf-profile.calltrace.cycles-pp.rcu_do_batch.rcu_core.__do_softirq.__irq_exit_rcu.sysvec_apic_timer_interrupt
>       0.00            +0.7        0.70 ± 10%  perf-profile.calltrace.cycles-pp.rcu_core.__do_softirq.__irq_exit_rcu.sysvec_apic_timer_interrupt.asm_sysvec_apic_timer_interrupt
>       0.00            +0.7        0.70 ±  8%  perf-profile.calltrace.cycles-pp.handle_mm_fault.do_user_addr_fault.exc_page_fault.asm_exc_page_fault.getpid@plt
>       0.00            +0.7        0.70 ± 10%  perf-profile.calltrace.cycles-pp.__do_softirq.__irq_exit_rcu.sysvec_apic_timer_interrupt.asm_sysvec_apic_timer_interrupt.filemap_map_pages
>       0.00            +0.7        0.70 ± 10%  perf-profile.calltrace.cycles-pp.__irq_exit_rcu.sysvec_apic_timer_interrupt.asm_sysvec_apic_timer_interrupt.filemap_map_pages.do_read_fault
>       0.00            +0.7        0.71 ±  8%  perf-profile.calltrace.cycles-pp.do_user_addr_fault.exc_page_fault.asm_exc_page_fault.getpid@plt
>       0.00            +0.7        0.71 ±  9%  perf-profile.calltrace.cycles-pp.exc_page_fault.asm_exc_page_fault.getpid@plt
>       0.00            +0.7        0.72 ±  8%  perf-profile.calltrace.cycles-pp.asm_exc_page_fault.getpid@plt
>       0.00            +0.7        0.73 ±  9%  perf-profile.calltrace.cycles-pp.handle_mm_fault.do_user_addr_fault.exc_page_fault.asm_exc_page_fault.shim_vhangup
>       0.00            +0.7        0.73 ±  9%  perf-profile.calltrace.cycles-pp.getpid@plt
>       0.00            +0.8        0.75 ±  9%  perf-profile.calltrace.cycles-pp.exc_page_fault.asm_exc_page_fault.shim_vhangup
>       0.00            +0.8        0.75 ±  9%  perf-profile.calltrace.cycles-pp.do_user_addr_fault.exc_page_fault.asm_exc_page_fault.shim_vhangup
>       0.00            +0.8        0.77 ±  9%  perf-profile.calltrace.cycles-pp.asm_exc_page_fault.shim_vhangup
>       0.00            +0.8        0.78 ±  8%  perf-profile.calltrace.cycles-pp.shim_vhangup
>       0.00            +0.8        0.80 ± 10%  perf-profile.calltrace.cycles-pp.sysvec_apic_timer_interrupt.asm_sysvec_apic_timer_interrupt.filemap_map_pages.do_read_fault.do_fault
>       0.00            +0.8        0.81 ± 10%  perf-profile.calltrace.cycles-pp.asm_sysvec_apic_timer_interrupt.filemap_map_pages.do_read_fault.do_fault.__handle_mm_fault
>       1.03            +1.3        2.36 ±  5%  perf-profile.calltrace.cycles-pp.asm_exc_page_fault.__libc_fork
>       0.90            +1.4        2.25 ±  5%  perf-profile.calltrace.cycles-pp.do_user_addr_fault.exc_page_fault.asm_exc_page_fault.__libc_fork
>       0.90            +1.4        2.25 ±  5%  perf-profile.calltrace.cycles-pp.exc_page_fault.asm_exc_page_fault.__libc_fork
>       0.74 ±  2%      +1.4        2.15 ±  6%  perf-profile.calltrace.cycles-pp.handle_mm_fault.do_user_addr_fault.exc_page_fault.asm_exc_page_fault.__libc_fork
>       0.00            +1.8        1.78 ±  7%  perf-profile.calltrace.cycles-pp.handle_mm_fault.do_user_addr_fault.exc_page_fault.asm_exc_page_fault.strchrnul@plt
>       0.00            +1.8        1.78 ±  7%  perf-profile.calltrace.cycles-pp.handle_mm_fault.do_user_addr_fault.exc_page_fault.asm_exc_page_fault.vhangup
>       0.00            +1.8        1.79 ±  7%  perf-profile.calltrace.cycles-pp.exc_page_fault.asm_exc_page_fault.strchrnul@plt
>       0.00            +1.8        1.79 ±  7%  perf-profile.calltrace.cycles-pp.do_user_addr_fault.exc_page_fault.asm_exc_page_fault.strchrnul@plt
>       0.00            +1.8        1.80 ±  7%  perf-profile.calltrace.cycles-pp.do_user_addr_fault.exc_page_fault.asm_exc_page_fault.vhangup
>       0.00            +1.8        1.80 ±  7%  perf-profile.calltrace.cycles-pp.exc_page_fault.asm_exc_page_fault.vhangup
>       0.00            +1.8        1.80 ±  7%  perf-profile.calltrace.cycles-pp.handle_mm_fault.do_user_addr_fault.exc_page_fault.asm_exc_page_fault.__sched_yield
>       0.00            +1.8        1.80 ±  7%  perf-profile.calltrace.cycles-pp.asm_exc_page_fault.strchrnul@plt
>       0.00            +1.8        1.80 ±  7%  perf-profile.calltrace.cycles-pp.handle_mm_fault.do_user_addr_fault.exc_page_fault.asm_exc_page_fault.__snprintf_chk
>       0.00            +1.8        1.80 ±  7%  perf-profile.calltrace.cycles-pp.asm_exc_page_fault.vhangup
>       0.00            +1.8        1.81 ±  7%  perf-profile.calltrace.cycles-pp.do_user_addr_fault.exc_page_fault.asm_exc_page_fault.__sched_yield
>       0.00            +1.8        1.81 ±  7%  perf-profile.calltrace.cycles-pp.exc_page_fault.asm_exc_page_fault.__sched_yield
>       0.00            +1.8        1.81 ±  7%  perf-profile.calltrace.cycles-pp.exc_page_fault.asm_exc_page_fault.__snprintf_chk
>       0.00            +1.8        1.81 ±  7%  perf-profile.calltrace.cycles-pp.do_user_addr_fault.exc_page_fault.asm_exc_page_fault.__snprintf_chk
>       0.00            +1.8        1.82 ±  7%  perf-profile.calltrace.cycles-pp.strchrnul@plt
>       0.00            +1.8        1.82 ±  7%  perf-profile.calltrace.cycles-pp.asm_exc_page_fault.__snprintf_chk
>       0.00            +1.8        1.83 ±  7%  perf-profile.calltrace.cycles-pp.asm_exc_page_fault.__sched_yield
>       0.00            +1.8        1.84 ±  7%  perf-profile.calltrace.cycles-pp.vhangup
>       0.00            +1.8        1.84 ±  7%  perf-profile.calltrace.cycles-pp.__snprintf_chk
>       0.00            +1.9        1.92 ±  7%  perf-profile.calltrace.cycles-pp.__sched_yield
>       0.00            +2.9        2.87 ±  2%  perf-profile.calltrace.cycles-pp.next_uptodate_folio.filemap_map_pages.do_read_fault.do_fault.__handle_mm_fault
>       7.24 ±  2%     +10.9       18.18 ±  9%  perf-profile.calltrace.cycles-pp.asm_exc_page_fault
>       6.80 ±  2%     +11.0       17.80 ±  9%  perf-profile.calltrace.cycles-pp.exc_page_fault.asm_exc_page_fault
>       6.78 ±  2%     +11.0       17.78 ± 10%  perf-profile.calltrace.cycles-pp.do_user_addr_fault.exc_page_fault.asm_exc_page_fault
>       6.08 ±  2%     +11.2       17.26 ± 10%  perf-profile.calltrace.cycles-pp.handle_mm_fault.do_user_addr_fault.exc_page_fault.asm_exc_page_fault
>       6.64 ±  4%     +22.7       29.37 ±  6%  perf-profile.calltrace.cycles-pp.__handle_mm_fault.handle_mm_fault.do_user_addr_fault.exc_page_fault.asm_exc_page_fault
>       4.79 ±  2%     +23.1       27.93 ±  7%  perf-profile.calltrace.cycles-pp.filemap_map_pages.do_read_fault.do_fault.__handle_mm_fault.handle_mm_fault
>       4.91 ±  2%     +23.3       28.25 ±  7%  perf-profile.calltrace.cycles-pp.do_read_fault.do_fault.__handle_mm_fault.handle_mm_fault.do_user_addr_fault
>       4.92 ±  2%     +23.3       28.26 ±  7%  perf-profile.calltrace.cycles-pp.do_fault.__handle_mm_fault.handle_mm_fault.do_user_addr_fault.exc_page_fault
>      82.82           -18.3       64.50 ±  2%  perf-profile.children.cycles-pp.do_syscall_64
>      82.84           -18.3       64.51 ±  2%  perf-profile.children.cycles-pp.entry_SYSCALL_64_after_hwframe
>      35.18            -8.4       26.82 ±  3%  perf-profile.children.cycles-pp.__do_sys_clone
>      35.18            -8.4       26.82 ±  3%  perf-profile.children.cycles-pp.kernel_clone
>      34.32            -8.2       26.15 ±  3%  perf-profile.children.cycles-pp.copy_process
>      35.78            -8.1       27.66 ±  2%  perf-profile.children.cycles-pp.do_exit
>      35.78            -8.1       27.66 ±  2%  perf-profile.children.cycles-pp.__x64_sys_exit_group
>      35.78            -8.1       27.66 ±  2%  perf-profile.children.cycles-pp.do_group_exit
>      32.41            -7.6       24.78 ±  3%  perf-profile.children.cycles-pp.dup_mm
>      33.68            -7.5       26.16 ±  2%  perf-profile.children.cycles-pp.exit_mm
>      33.54            -7.5       26.03 ±  2%  perf-profile.children.cycles-pp.exit_mmap
>      33.63            -7.5       26.13 ±  2%  perf-profile.children.cycles-pp.__mmput
>      37.08            -7.4       29.68 ±  2%  perf-profile.children.cycles-pp.__libc_fork
>      29.95            -7.1       22.89 ±  3%  perf-profile.children.cycles-pp.dup_mmap
>       6.50            -6.5        0.00        perf-profile.children.cycles-pp.next_uptodate_page
>      17.25            -5.3       12.00 ±  5%  perf-profile.children.cycles-pp.down_write
>      14.39            -4.8        9.56 ±  7%  perf-profile.children.cycles-pp.rwsem_down_write_slowpath
>      14.18            -4.8        9.40 ±  7%  perf-profile.children.cycles-pp.rwsem_optimistic_spin
>      15.69            -4.1       11.56 ±  4%  perf-profile.children.cycles-pp.anon_vma_fork
>      10.16            -3.8        6.40 ±  8%  perf-profile.children.cycles-pp.osq_lock
>      12.54            -2.9        9.60 ±  3%  perf-profile.children.cycles-pp.free_pgtables
>      13.13            -2.9       10.27 ±  2%  perf-profile.children.cycles-pp.unmap_vmas
>      12.54            -2.7        9.84 ±  2%  perf-profile.children.cycles-pp.unmap_page_range
>      12.41            -2.7        9.73 ±  2%  perf-profile.children.cycles-pp.zap_pmd_range
>      12.31            -2.7        9.64        perf-profile.children.cycles-pp.zap_pte_range
>      10.06            -2.5        7.60 ±  3%  perf-profile.children.cycles-pp.anon_vma_clone
>       8.73            -2.3        6.42 ±  4%  perf-profile.children.cycles-pp.unlink_anon_vmas
>       7.28            -2.2        5.04 ±  3%  perf-profile.children.cycles-pp._compound_head
>       5.34            -1.3        4.00 ±  3%  perf-profile.children.cycles-pp.native_queued_spin_lock_slowpath
>       5.40            -1.3        4.07 ±  3%  perf-profile.children.cycles-pp.copy_page_range
>       5.30            -1.3        3.99 ±  3%  perf-profile.children.cycles-pp.copy_p4d_range
>       5.26            -1.3        4.01 ±  3%  perf-profile.children.cycles-pp.tlb_finish_mmu
>       5.22            -1.2        3.97 ±  3%  perf-profile.children.cycles-pp.tlb_batch_pages_flush
>       4.84 ±  2%      -1.2        3.64 ±  3%  perf-profile.children.cycles-pp.copy_pte_range
>       3.86            -1.1        2.77 ±  4%  perf-profile.children.cycles-pp.rwsem_spin_on_owner
>       4.19            -1.0        3.16 ±  3%  perf-profile.children.cycles-pp.release_pages
>       3.59            -0.9        2.68 ±  3%  perf-profile.children.cycles-pp.copy_present_pte
>       2.92            -0.8        2.08 ±  4%  perf-profile.children.cycles-pp.wait4
>       2.83            -0.8        2.01 ±  4%  perf-profile.children.cycles-pp.__do_sys_wait4
>       2.82            -0.8        2.00 ±  4%  perf-profile.children.cycles-pp.kernel_wait4
>       2.77            -0.8        1.96 ±  4%  perf-profile.children.cycles-pp.do_wait
>       8.14            -0.8        7.37 ±  2%  perf-profile.children.cycles-pp.setsid
>       8.08            -0.8        7.32 ±  2%  perf-profile.children.cycles-pp.__x64_sys_setsid
>       8.08            -0.8        7.32 ±  2%  perf-profile.children.cycles-pp.ksys_setsid
>       5.74            -0.7        5.03 ±  2%  perf-profile.children.cycles-pp.__do_softirq
>       1.20 ±  2%      -0.7        0.53 ±  8%  perf-profile.children.cycles-pp.ret_from_fork_asm
>       2.56            -0.7        1.90 ±  4%  perf-profile.children.cycles-pp.__schedule
>       1.15 ±  2%      -0.7        0.50 ±  8%  perf-profile.children.cycles-pp.ret_from_fork
>       5.59            -0.6        4.95 ±  2%  perf-profile.children.cycles-pp.rcu_core
>       5.56            -0.6        4.91 ±  2%  perf-profile.children.cycles-pp.rcu_do_batch
>       7.80            -0.6        7.18        perf-profile.children.cycles-pp.sched_autogroup_create_attach
>       2.86            -0.6        2.25 ±  2%  perf-profile.children.cycles-pp.kmem_cache_alloc
>       1.23 ±  2%      -0.6        0.62 ±  7%  perf-profile.children.cycles-pp.queued_write_lock_slowpath
>       2.51            -0.6        1.90 ±  2%  perf-profile.children.cycles-pp.vm_area_dup
>       3.60            -0.6        3.03 ±  2%  perf-profile.children.cycles-pp.anon_vma_interval_tree_insert
>       6.22            -0.6        5.65 ±  2%  perf-profile.children.cycles-pp.sysvec_apic_timer_interrupt
>       6.26            -0.6        5.70 ±  2%  perf-profile.children.cycles-pp.asm_sysvec_apic_timer_interrupt
>       2.43 ±  2%      -0.6        1.87 ±  4%  perf-profile.children.cycles-pp.mm_init
>       3.52            -0.6        2.96        perf-profile.children.cycles-pp.memcg_slab_post_alloc_hook
>       2.82            -0.5        2.31 ±  2%  perf-profile.children.cycles-pp.online_fair_sched_group
>       1.10            -0.5        0.60 ±  7%  perf-profile.children.cycles-pp.pick_next_task_fair
>       1.42 ±  4%      -0.5        0.92 ±  6%  perf-profile.children.cycles-pp.__alloc_pages
>       5.43            -0.5        4.94 ±  2%  perf-profile.children.cycles-pp.__irq_exit_rcu
>       1.02 ±  2%      -0.5        0.54 ±  7%  perf-profile.children.cycles-pp.newidle_balance
>       4.50            -0.5        4.03 ±  2%  perf-profile.children.cycles-pp._raw_spin_lock_irqsave
>       2.84            -0.5        2.38 ±  2%  perf-profile.children.cycles-pp.__slab_free
>       2.06 ±  2%      -0.5        1.61 ±  4%  perf-profile.children.cycles-pp.pcpu_alloc
>       0.92 ±  2%      -0.4        0.47 ±  7%  perf-profile.children.cycles-pp.load_balance
>       2.31 ±  4%      -0.4        1.91 ±  5%  perf-profile.children.cycles-pp.unlink_file_vma
>       1.68 ±  2%      -0.4        1.31 ±  4%  perf-profile.children.cycles-pp.__percpu_counter_init
>       2.84            -0.4        2.49 ±  2%  perf-profile.children.cycles-pp.up_write
>       3.52            -0.4        3.17 ±  2%  perf-profile.children.cycles-pp.sched_free_group_rcu
>       1.95            -0.4        1.60        perf-profile.children.cycles-pp._raw_spin_lock
>       1.49 ±  2%      -0.3        1.15 ±  3%  perf-profile.children.cycles-pp.finish_task_switch
>       0.65 ±  3%      -0.3        0.31 ±  6%  perf-profile.children.cycles-pp.schedule_tail
>       3.48            -0.3        3.14 ±  2%  perf-profile.children.cycles-pp.free_fair_sched_group
>       1.86            -0.3        1.52 ±  3%  perf-profile.children.cycles-pp.schedule
>       1.51            -0.3        1.18 ±  2%  perf-profile.children.cycles-pp.kmem_cache_free
>       1.91            -0.3        1.59 ±  3%  perf-profile.children.cycles-pp.___slab_alloc
>       3.56            -0.3        3.24        perf-profile.children.cycles-pp.page_remove_rmap
>       0.67 ±  2%      -0.3        0.35 ±  7%  perf-profile.children.cycles-pp.update_sd_lb_stats
>       0.68 ±  2%      -0.3        0.36 ±  7%  perf-profile.children.cycles-pp.find_busiest_group
>       0.48 ±  3%      -0.3        0.18 ± 12%  perf-profile.children.cycles-pp.kthread
>       0.63 ±  2%      -0.3        0.33 ±  8%  perf-profile.children.cycles-pp.update_sg_lb_stats
>       4.19            -0.3        3.89        perf-profile.children.cycles-pp.__kmem_cache_alloc_node
>       0.84 ±  4%      -0.3        0.54 ±  6%  perf-profile.children.cycles-pp.get_page_from_freelist
>       0.50 ±  2%      -0.3        0.20 ± 10%  perf-profile.children.cycles-pp.__perf_sw_event
>       0.80            -0.3        0.51 ±  4%  perf-profile.children.cycles-pp.dup_task_struct
>       0.48 ±  2%      -0.3        0.19 ±  9%  perf-profile.children.cycles-pp.___perf_sw_event
>       0.62 ±  2%      -0.3        0.34 ±  8%  perf-profile.children.cycles-pp.secondary_startup_64_no_verify
>       0.62 ±  2%      -0.3        0.34 ±  8%  perf-profile.children.cycles-pp.cpu_startup_entry
>       0.62 ±  2%      -0.3        0.34 ±  8%  perf-profile.children.cycles-pp.do_idle
>       0.79 ±  8%      -0.3        0.52 ± 10%  perf-profile.children.cycles-pp.pte_alloc_one
>       0.61 ±  2%      -0.3        0.34 ±  7%  perf-profile.children.cycles-pp.start_secondary
>       1.29 ±  2%      -0.3        1.01 ±  3%  perf-profile.children.cycles-pp.__mmdrop
>       4.08            -0.3        3.81        perf-profile.children.cycles-pp.kmalloc_node_trace
>       0.89 ±  5%      -0.2        0.65 ±  7%  perf-profile.children.cycles-pp.__mutex_lock
>       0.53 ±  2%      -0.2        0.29 ±  4%  perf-profile.children.cycles-pp.alloc_thread_stack_node
>       0.35 ±  5%      -0.2        0.11 ± 12%  perf-profile.children.cycles-pp.smpboot_thread_fn
>       1.09            -0.2        0.86 ±  2%  perf-profile.children.cycles-pp.raw_spin_rq_lock_nested
>       1.20            -0.2        0.96 ±  2%  perf-profile.children.cycles-pp.__anon_vma_interval_tree_remove
>       0.63 ± 11%      -0.2        0.40 ± 11%  perf-profile.children.cycles-pp.__pte_alloc
>       0.49 ±  3%      -0.2        0.26 ±  6%  perf-profile.children.cycles-pp.queued_read_lock_slowpath
>       1.04 ±  2%      -0.2        0.81 ±  3%  perf-profile.children.cycles-pp.free_pages_and_swap_cache
>       0.32 ±  5%      -0.2        0.10 ± 14%  perf-profile.children.cycles-pp.run_ksoftirqd
>       0.50 ±  2%      -0.2        0.28 ±  6%  perf-profile.children.cycles-pp.do_task_dead
>       1.01 ±  2%      -0.2        0.79 ±  2%  perf-profile.children.cycles-pp.free_swap_cache
>       0.67            -0.2        0.46 ±  5%  perf-profile.children.cycles-pp.lock_vma_under_rcu
>       0.26 ±  5%      -0.2        0.06 ± 11%  perf-profile.children.cycles-pp.__vmalloc_node_range
>       0.50 ±  3%      -0.2        0.30 ±  8%  perf-profile.children.cycles-pp.clear_page_erms
>       0.60 ±  2%      -0.2        0.41 ±  3%  perf-profile.children.cycles-pp.exit_notify
>       0.55 ±  5%      -0.2        0.36 ±  8%  perf-profile.children.cycles-pp.__memcg_kmem_charge_page
>       1.29            -0.2        1.10 ±  2%  perf-profile.children.cycles-pp.sched_unregister_group_rcu
>       0.54 ±  2%      -0.2        0.36 ±  5%  perf-profile.children.cycles-pp.wait_task_zombie
>       1.28            -0.2        1.10 ±  2%  perf-profile.children.cycles-pp.unregister_fair_sched_group
>       0.82            -0.2        0.64 ±  3%  perf-profile.children.cycles-pp.wake_up_new_task
>       0.38 ±  2%      -0.2        0.21 ±  6%  perf-profile.children.cycles-pp.__setpgid
>       0.37 ±  3%      -0.2        0.20 ±  8%  perf-profile.children.cycles-pp.cpuidle_idle_call
>       0.50 ±  2%      -0.2        0.33 ±  4%  perf-profile.children.cycles-pp.release_task
>       0.52 ± 14%      -0.2        0.35 ± 11%  perf-profile.children.cycles-pp.machine__process_fork_event
>       0.34 ±  2%      -0.2        0.18 ±  5%  perf-profile.children.cycles-pp.__do_sys_setpgid
>       0.69            -0.2        0.53 ±  2%  perf-profile.children.cycles-pp.__vm_area_free
>       0.83 ±  2%      -0.2        0.68 ±  4%  perf-profile.children.cycles-pp.wp_page_copy
>       0.74            -0.1        0.59 ±  3%  perf-profile.children.cycles-pp.attach_entity_cfs_rq
>       0.32 ±  4%      -0.1        0.17 ±  8%  perf-profile.children.cycles-pp.cpuidle_enter
>       0.31 ±  4%      -0.1        0.17 ±  9%  perf-profile.children.cycles-pp.cpuidle_enter_state
>       0.69            -0.1        0.55 ±  3%  perf-profile.children.cycles-pp.update_load_avg
>       0.53            -0.1        0.39 ±  4%  perf-profile.children.cycles-pp.__list_del_entry_valid
>       0.50            -0.1        0.36 ±  4%  perf-profile.children.cycles-pp.unmap_single_vma
>       0.75 ±  2%      -0.1        0.61 ±  2%  perf-profile.children.cycles-pp.remove_vma
>       0.54            -0.1        0.41 ±  2%  perf-profile.children.cycles-pp._exit
>       1.04            -0.1        0.91 ±  2%  perf-profile.children.cycles-pp.remove_entity_load_avg
>       0.84            -0.1        0.72 ±  3%  perf-profile.children.cycles-pp.mod_objcg_state
>       0.68            -0.1        0.55 ±  2%  perf-profile.children.cycles-pp.select_task_rq_fair
>       0.44 ±  2%      -0.1        0.32 ±  2%  perf-profile.children.cycles-pp.acct_collect
>       0.68 ±  2%      -0.1        0.57 ±  2%  perf-profile.children.cycles-pp.fput
>       1.03            -0.1        0.92 ±  2%  perf-profile.children.cycles-pp.mas_store
>       0.62            -0.1        0.51 ±  4%  perf-profile.children.cycles-pp.get_partial_node
>       0.39 ±  3%      -0.1        0.28 ±  7%  perf-profile.children.cycles-pp.copy_mc_enhanced_fast_string
>       0.62 ±  2%      -0.1        0.51 ±  2%  perf-profile.children.cycles-pp.__put_anon_vma
>       0.39 ±  3%      -0.1        0.28 ±  6%  perf-profile.children.cycles-pp.__wp_page_copy_user
>       0.31 ±  4%      -0.1        0.20 ±  7%  perf-profile.children.cycles-pp.rmqueue
>       0.57            -0.1        0.46 ±  2%  perf-profile.children.cycles-pp.find_idlest_cpu
>       0.32 ±  4%      -0.1        0.21 ±  7%  perf-profile.children.cycles-pp.folio_lruvec_lock_irqsave
>       0.48 ±  3%      -0.1        0.37 ±  3%  perf-profile.children.cycles-pp.free_percpu
>       0.71            -0.1        0.61 ±  4%  perf-profile.children.cycles-pp.deactivate_slab
>       0.44 ±  4%      -0.1        0.34 ±  4%  perf-profile.children.cycles-pp.percpu_counter_destroy
>       0.14 ±  4%      -0.1        0.05 ± 45%  perf-profile.children.cycles-pp.rmqueue_bulk
>       0.30 ±  5%      -0.1        0.21 ± 10%  perf-profile.children.cycles-pp.cgroup_rstat_updated
>       0.76            -0.1        0.68 ±  3%  perf-profile.children.cycles-pp.update_rq_clock_task
>       0.58 ±  2%      -0.1        0.49 ±  3%  perf-profile.children.cycles-pp.__percpu_counter_sum
>       0.53            -0.1        0.44 ±  2%  perf-profile.children.cycles-pp.find_idlest_group
>       0.28 ± 14%      -0.1        0.19 ± 11%  perf-profile.children.cycles-pp.____machine__findnew_thread
>       1.97 ±  2%      -0.1        1.88        perf-profile.children.cycles-pp.__unfreeze_partials
>       0.51            -0.1        0.42 ±  2%  perf-profile.children.cycles-pp.update_sg_wakeup_stats
>       0.26 ±  5%      -0.1        0.18 ±  8%  perf-profile.children.cycles-pp.memcg_account_kmem
>       0.24 ± 15%      -0.1        0.15 ± 12%  perf-profile.children.cycles-pp.maps__clone
>       0.78            -0.1        0.70 ±  2%  perf-profile.children.cycles-pp.__sysvec_apic_timer_interrupt
>       0.18 ±  5%      -0.1        0.10 ± 12%  perf-profile.children.cycles-pp.free_unref_page
>       0.78            -0.1        0.70 ±  2%  perf-profile.children.cycles-pp.hrtimer_interrupt
>       0.23 ±  2%      -0.1        0.15 ±  6%  perf-profile.children.cycles-pp.down_read_trylock
>       0.18 ±  2%      -0.1        0.10 ±  8%  perf-profile.children.cycles-pp.schedule_idle
>       0.69            -0.1        0.61 ±  3%  perf-profile.children.cycles-pp.__hrtimer_run_queues
>       0.38            -0.1        0.30 ± 22%  perf-profile.children.cycles-pp.do_wp_page
>       0.40 ±  2%      -0.1        0.32 ±  2%  perf-profile.children.cycles-pp.attach_entity_load_avg
>       0.31 ±  2%      -0.1        0.24 ±  3%  perf-profile.children.cycles-pp.dequeue_task_fair
>       0.43 ±  2%      -0.1        0.35 ± 10%  perf-profile.children.cycles-pp.sync_regs
>       0.40 ±  2%      -0.1        0.32 ±  2%  perf-profile.children.cycles-pp.__rb_erase_color
>       0.60            -0.1        0.52 ±  3%  perf-profile.children.cycles-pp.tick_sched_handle
>       0.59            -0.1        0.52 ±  2%  perf-profile.children.cycles-pp.update_process_times
>       0.26 ±  2%      -0.1        0.18 ±  6%  perf-profile.children.cycles-pp.mas_walk
>       0.62 ±  2%      -0.1        0.55 ±  3%  perf-profile.children.cycles-pp.tick_sched_timer
>       0.20 ±  2%      -0.1        0.13 ±  5%  perf-profile.children.cycles-pp.__put_user_4
>       0.13 ±  6%      -0.1        0.06 ± 15%  perf-profile.children.cycles-pp.free_pcppages_bulk
>       0.79            -0.1        0.73 ±  2%  perf-profile.children.cycles-pp.vma_interval_tree_remove
>       0.25            -0.1        0.18 ±  2%  perf-profile.children.cycles-pp.mtree_range_walk
>       0.23            -0.1        0.16 ±  3%  perf-profile.children.cycles-pp.activate_task
>       0.40 ±  2%      -0.1        0.34 ±  2%  perf-profile.children.cycles-pp.mas_next_slot
>       0.33 ±  4%      -0.1        0.26 ±  5%  perf-profile.children.cycles-pp.percpu_counter_add_batch
>       0.27 ±  2%      -0.1        0.20 ±  2%  perf-profile.children.cycles-pp.kfree
>       0.12 ±  3%      -0.1        0.06 ± 13%  perf-profile.children.cycles-pp.worker_thread
>       0.24 ±  2%      -0.1        0.18 ±  4%  perf-profile.children.cycles-pp.dequeue_entity
>       0.62            -0.1        0.56        perf-profile.children.cycles-pp.mas_wr_bnode
>       0.16 ±  4%      -0.1        0.10 ±  6%  perf-profile.children.cycles-pp.update_blocked_averages
>       0.18 ±  4%      -0.1        0.12 ±  5%  perf-profile.children.cycles-pp.pgd_alloc
>       0.14 ±  3%      -0.1        0.07 ± 10%  perf-profile.children.cycles-pp.intel_idle
>       0.24            -0.1        0.18 ±  4%  perf-profile.children.cycles-pp.enqueue_task_fair
>       0.18 ±  4%      -0.1        0.12 ±  4%  perf-profile.children.cycles-pp.__get_free_pages
>       0.15            -0.1        0.09 ±  6%  perf-profile.children.cycles-pp.allocate_slab
>       0.56            -0.1        0.50 ±  2%  perf-profile.children.cycles-pp.mas_split
>       0.29            -0.1        0.24 ±  3%  perf-profile.children.cycles-pp.__update_load_avg_cfs_rq
>       0.50 ±  3%      -0.1        0.44 ±  4%  perf-profile.children.cycles-pp.osq_unlock
>       0.49 ±  2%      -0.1        0.43 ±  2%  perf-profile.children.cycles-pp.scheduler_tick
>       0.38            -0.1        0.32 ±  2%  perf-profile.children.cycles-pp.sched_move_task
>       0.13 ± 11%      -0.1        0.08 ± 12%  perf-profile.children.cycles-pp.do_anonymous_page
>       0.20 ±  8%      -0.1        0.15 ±  8%  perf-profile.children.cycles-pp.try_charge_memcg
>       0.27 ±  2%      -0.1        0.22 ±  5%  perf-profile.children.cycles-pp.__memcpy
>       0.18 ± 13%      -0.1        0.13 ± 12%  perf-profile.children.cycles-pp.nsinfo__new
>       0.30 ±  3%      -0.0        0.25 ±  2%  perf-profile.children.cycles-pp._find_next_bit
>       0.42 ±  2%      -0.0        0.38 ±  2%  perf-profile.children.cycles-pp.__cond_resched
>       0.18 ±  6%      -0.0        0.13 ±  8%  perf-profile.children.cycles-pp.mutex_spin_on_owner
>       0.18 ± 13%      -0.0        0.13 ± 12%  perf-profile.children.cycles-pp.thread__new
>       0.18 ±  2%      -0.0        0.13 ±  2%  perf-profile.children.cycles-pp.vma_alloc_folio
>       0.24 ±  5%      -0.0        0.19 ±  2%  perf-profile.children.cycles-pp.obj_cgroup_charge
>       0.22 ±  4%      -0.0        0.17 ±  4%  perf-profile.children.cycles-pp.__reclaim_stacks
>       0.18 ±  2%      -0.0        0.13 ±  4%  perf-profile.children.cycles-pp.enqueue_entity
>       0.16 ±  4%      -0.0        0.11 ±  6%  perf-profile.children.cycles-pp.__mem_cgroup_charge
>       0.51 ±  5%      -0.0        0.46 ±  3%  perf-profile.children.cycles-pp.__put_task_struct
>       0.16 ±  3%      -0.0        0.11 ±  3%  perf-profile.children.cycles-pp.__folio_alloc
>       0.24 ±  4%      -0.0        0.19 ±  3%  perf-profile.children.cycles-pp.vma_interval_tree_insert_after
>       0.19 ±  3%      -0.0        0.15 ±  5%  perf-profile.children.cycles-pp.__list_add_valid
>       0.16 ± 12%      -0.0        0.11 ± 12%  perf-profile.children.cycles-pp.vfs_statx
>       0.17 ± 13%      -0.0        0.12 ± 13%  perf-profile.children.cycles-pp.__do_sys_newstat
>       0.12 ±  4%      -0.0        0.08 ±  4%  perf-profile.children.cycles-pp.cpu_util
>       0.33 ±  2%      -0.0        0.29 ±  4%  perf-profile.children.cycles-pp.mas_wr_store_entry
>       0.25 ±  2%      -0.0        0.21 ±  2%  perf-profile.children.cycles-pp.do_notify_parent
>       0.24 ±  2%      -0.0        0.20 ±  2%  perf-profile.children.cycles-pp.__rb_insert_augmented
>       0.21 ±  2%      -0.0        0.17 ±  6%  perf-profile.children.cycles-pp.arch_dup_task_struct
>       0.24 ±  3%      -0.0        0.20 ±  2%  perf-profile.children.cycles-pp.__anon_vma_interval_tree_augment_rotate
>       0.22 ±  2%      -0.0        0.18 ±  3%  perf-profile.children.cycles-pp.mas_push_data
>       0.23 ±  2%      -0.0        0.19 ±  2%  perf-profile.children.cycles-pp.perf_event_task_tick
>       0.10 ±  4%      -0.0        0.06 ±  6%  perf-profile.children.cycles-pp.mas_expected_entries
>       0.10 ±  4%      -0.0        0.06 ±  6%  perf-profile.children.cycles-pp.mas_alloc_nodes
>       0.09 ±  4%      -0.0        0.05 ±  8%  perf-profile.children.cycles-pp.__kmem_cache_alloc_bulk
>       0.23 ±  3%      -0.0        0.19 ±  3%  perf-profile.children.cycles-pp.__wake_up_common_lock
>       0.23 ±  2%      -0.0        0.19 ±  2%  perf-profile.children.cycles-pp.perf_adjust_freq_unthr_context
>       0.09 ±  4%      -0.0        0.06 ±  9%  perf-profile.children.cycles-pp.kmem_cache_alloc_bulk
>       0.24 ±  2%      -0.0        0.20        perf-profile.children.cycles-pp.autogroup_move_group
>       0.13 ±  8%      -0.0        0.09 ±  9%  perf-profile.children.cycles-pp.charge_memcg
>       0.10 ±  4%      -0.0        0.07 ±  5%  perf-profile.children.cycles-pp.__update_blocked_fair
>       0.07 ± 11%      -0.0        0.04 ± 71%  perf-profile.children.cycles-pp.step_into
>       0.24 ±  2%      -0.0        0.21        perf-profile.children.cycles-pp.vm_normal_page
>       0.15 ±  3%      -0.0        0.12 ±  3%  perf-profile.children.cycles-pp.shim_waitpid
>       0.12 ±  7%      -0.0        0.09 ±  5%  perf-profile.children.cycles-pp.__count_memcg_events
>       0.11 ±  3%      -0.0        0.08 ±  6%  perf-profile.children.cycles-pp.mas_split_final_node
>       0.08 ±  4%      -0.0        0.05        perf-profile.children.cycles-pp.shuffle_freelist
>       0.13 ±  2%      -0.0        0.10 ±  4%  perf-profile.children.cycles-pp.syscall_exit_to_user_mode
>       0.21 ±  2%      -0.0        0.18 ±  5%  perf-profile.children.cycles-pp.exit_to_user_mode_prepare
>       0.16 ±  5%      -0.0        0.14 ±  5%  perf-profile.children.cycles-pp.__mod_memcg_lruvec_state
>       0.12 ±  3%      -0.0        0.09 ±  5%  perf-profile.children.cycles-pp.mab_mas_cp
>       0.22 ±  2%      -0.0        0.19 ±  3%  perf-profile.children.cycles-pp.mas_wr_append
>       0.06            -0.0        0.03 ± 70%  perf-profile.children.cycles-pp.free_unref_page_prepare
>       0.14 ±  4%      -0.0        0.11 ±  6%  perf-profile.children.cycles-pp.__tlb_remove_page_size
>       0.15 ±  3%      -0.0        0.12 ±  5%  perf-profile.children.cycles-pp.__exit_signal
>       0.08 ±  7%      -0.0        0.06 ±  9%  perf-profile.children.cycles-pp.switch_fpu_return
>       0.12 ±  5%      -0.0        0.10 ±  4%  perf-profile.children.cycles-pp.__mod_lruvec_page_state
>       0.12 ±  4%      -0.0        0.09 ±  4%  perf-profile.children.cycles-pp.folio_batch_move_lru
>       0.16 ±  3%      -0.0        0.13 ±  5%  perf-profile.children.cycles-pp.update_curr
>       0.23 ±  2%      -0.0        0.21 ±  3%  perf-profile.children.cycles-pp.try_to_wake_up
>       0.17 ±  4%      -0.0        0.15 ±  4%  perf-profile.children.cycles-pp.mas_update_gap
>       0.15 ±  4%      -0.0        0.13 ±  2%  perf-profile.children.cycles-pp.pcpu_alloc_area
>       0.11 ±  3%      -0.0        0.09 ±  5%  perf-profile.children.cycles-pp.ttwu_do_activate
>       0.07 ±  5%      -0.0        0.05        perf-profile.children.cycles-pp.get_zeroed_page
>       0.07 ± 12%      -0.0        0.05        perf-profile.children.cycles-pp.rb_next
>       0.12 ±  4%      -0.0        0.09 ±  5%  perf-profile.children.cycles-pp.alloc_pid
>       0.08 ±  4%      -0.0        0.06 ±  6%  perf-profile.children.cycles-pp.mark_page_accessed
>       0.15 ±  3%      -0.0        0.13 ±  3%  perf-profile.children.cycles-pp.refill_obj_stock
>       0.11 ±  4%      -0.0        0.09 ±  5%  perf-profile.children.cycles-pp._raw_spin_trylock
>       0.08 ±  5%      -0.0        0.06 ±  7%  perf-profile.children.cycles-pp.copy_signal
>       0.07 ±  5%      -0.0        0.05        perf-profile.children.cycles-pp.__radix_tree_lookup
>       0.07 ±  5%      -0.0        0.05 ±  8%  perf-profile.children.cycles-pp.__p4d_alloc
>       0.11 ±  3%      -0.0        0.09        perf-profile.children.cycles-pp.lru_add_drain_cpu
>       0.09            -0.0        0.07 ± 10%  perf-profile.children.cycles-pp.pte_offset_map_nolock
>       0.08 ±  6%      -0.0        0.06 ±  6%  perf-profile.children.cycles-pp.exit_to_user_mode_loop
>       0.14 ±  3%      -0.0        0.13 ±  3%  perf-profile.children.cycles-pp._raw_spin_lock_irq
>       0.11 ±  3%      -0.0        0.09 ±  4%  perf-profile.children.cycles-pp.lru_add_drain
>       0.10 ±  4%      -0.0        0.08 ±  4%  perf-profile.children.cycles-pp.mas_leaf_max_gap
>       0.09            -0.0        0.08 ±  6%  perf-profile.children.cycles-pp.copy_creds
>       0.08 ±  6%      -0.0        0.06 ±  6%  perf-profile.children.cycles-pp._raw_write_lock_irq
>       0.12 ±  3%      -0.0        0.11 ±  3%  perf-profile.children.cycles-pp.perf_iterate_sb
>       0.08 ±  5%      -0.0        0.07        perf-profile.children.cycles-pp.pcpu_free_area
>       0.08 ±  5%      -0.0        0.07        perf-profile.children.cycles-pp.dup_fd
>       0.08 ±  8%      -0.0        0.06 ±  7%  perf-profile.children.cycles-pp.__mem_cgroup_uncharge_list
>       0.07 ±  7%      -0.0        0.05 ±  7%  perf-profile.children.cycles-pp._find_next_zero_bit
>       0.09 ±  4%      -0.0        0.07 ±  6%  perf-profile.children.cycles-pp.free_unref_page_list
>       0.08 ±  5%      -0.0        0.07 ±  5%  perf-profile.children.cycles-pp.prepare_creds
>       0.09            -0.0        0.08 ±  6%  perf-profile.children.cycles-pp.__pte_offset_map
>       0.09 ±  4%      -0.0        0.08 ±  4%  perf-profile.children.cycles-pp.update_rq_clock
>       0.10 ±  4%      -0.0        0.09        perf-profile.children.cycles-pp.select_task_rq
>       0.10 ±  4%      -0.0        0.09 ±  5%  perf-profile.children.cycles-pp.__update_load_avg_blocked_se
>       0.07            -0.0        0.06        perf-profile.children.cycles-pp.perf_event_task_output
>       0.07            -0.0        0.06        perf-profile.children.cycles-pp.propagate_entity_cfs_rq
>       0.06            -0.0        0.05        perf-profile.children.cycles-pp.sched_clock_cpu
>       0.06            -0.0        0.05        perf-profile.children.cycles-pp.__kmem_cache_free
>       0.38            +0.0        0.40        perf-profile.children.cycles-pp.dup_userfaultfd
>       0.06 ± 11%      +0.1        0.12 ±  6%  perf-profile.children.cycles-pp.__getpid
>       0.29            +0.1        0.35 ±  4%  perf-profile.children.cycles-pp.__pte_offset_map_lock
>       0.04 ± 45%      +0.1        0.10 ± 10%  perf-profile.children.cycles-pp.xas_find
>       0.21 ±  3%      +0.1        0.28 ±  3%  perf-profile.children.cycles-pp.init_entity_runnable_average
>       0.83 ±  3%      +0.1        0.91 ±  2%  perf-profile.children.cycles-pp.do_set_pte
>       0.00            +0.1        0.09 ± 10%  perf-profile.children.cycles-pp.xas_load
>       0.00            +0.1        0.10 ±  8%  perf-profile.children.cycles-pp.io_schedule
>       0.00            +0.1        0.10 ±  5%  perf-profile.children.cycles-pp.folio_wait_bit_common
>       0.06 ±  6%      +0.1        0.18 ±  7%  perf-profile.children.cycles-pp.init_tg_cfs_entry
>       0.00            +0.1        0.14 ±  8%  perf-profile.children.cycles-pp.filemap_fault
>       0.00            +0.1        0.14 ±  8%  perf-profile.children.cycles-pp.__do_fault
>       0.47 ±  3%      +0.2        0.63 ±  3%  perf-profile.children.cycles-pp.page_add_file_rmap
>       0.73 ±  9%      +0.2        0.97 ±  7%  perf-profile.children.cycles-pp.stress_fork_fn
>       0.43 ±  8%      +0.3        0.77 ±  8%  perf-profile.children.cycles-pp.getpid@plt
>       0.42            +0.4        0.80 ±  8%  perf-profile.children.cycles-pp.shim_vhangup
>       0.44            +1.4        1.85 ±  7%  perf-profile.children.cycles-pp.vhangup
>       0.40            +1.5        1.85 ±  7%  perf-profile.children.cycles-pp.__snprintf_chk
>       0.48            +1.5        1.93 ±  7%  perf-profile.children.cycles-pp.__sched_yield
>       0.39 ±  2%      +1.5        1.85 ±  7%  perf-profile.children.cycles-pp.strchrnul@plt
>       0.00            +4.9        4.90        perf-profile.children.cycles-pp.next_uptodate_folio
>      13.05           +19.4       32.42 ±  5%  perf-profile.children.cycles-pp.asm_exc_page_fault
>      12.10           +19.5       31.63 ±  6%  perf-profile.children.cycles-pp.exc_page_fault
>      12.07           +19.5       31.61 ±  6%  perf-profile.children.cycles-pp.do_user_addr_fault
>      10.88           +19.8       30.73 ±  6%  perf-profile.children.cycles-pp.handle_mm_fault
>      10.35           +20.1       30.50 ±  6%  perf-profile.children.cycles-pp.__handle_mm_fault
>       8.37           +20.4       28.77 ±  6%  perf-profile.children.cycles-pp.filemap_map_pages
>       8.61           +20.5       29.08 ±  6%  perf-profile.children.cycles-pp.do_fault
>       8.60           +20.5       29.07 ±  6%  perf-profile.children.cycles-pp.do_read_fault
>       6.06            -6.1        0.00        perf-profile.self.cycles-pp.next_uptodate_page
>       9.78            -3.6        6.19 ±  8%  perf-profile.self.cycles-pp.osq_lock
>       6.76            -2.1        4.70 ±  3%  perf-profile.self.cycles-pp._compound_head
>       5.31            -1.3        4.00 ±  3%  perf-profile.self.cycles-pp.native_queued_spin_lock_slowpath
>       3.62            -1.0        2.62 ±  4%  perf-profile.self.cycles-pp.rwsem_spin_on_owner
>       3.42            -0.8        2.65 ±  3%  perf-profile.self.cycles-pp.release_pages
>       2.74 ±  6%      -0.8        1.98 ±  4%  perf-profile.self.cycles-pp.zap_pte_range
>       3.38            -0.5        2.85 ±  2%  perf-profile.self.cycles-pp.anon_vma_interval_tree_insert
>       2.98            -0.5        2.51        perf-profile.self.cycles-pp.memcg_slab_post_alloc_hook
>       2.72            -0.4        2.28 ±  2%  perf-profile.self.cycles-pp.__slab_free
>       2.53            -0.3        2.20        perf-profile.self.cycles-pp.down_write
>       2.64            -0.3        2.34 ±  2%  perf-profile.self.cycles-pp.up_write
>       3.31            -0.3        3.03        perf-profile.self.cycles-pp.page_remove_rmap
>       0.42 ±  2%      -0.3        0.15 ± 10%  perf-profile.self.cycles-pp.___perf_sw_event
>       0.51 ±  2%      -0.2        0.26 ±  8%  perf-profile.self.cycles-pp.update_sg_lb_stats
>       0.79 ±  2%      -0.2        0.56 ±  3%  perf-profile.self.cycles-pp.anon_vma_clone
>       0.88            -0.2        0.66 ±  2%  perf-profile.self.cycles-pp.vm_area_dup
>       1.66            -0.2        1.44 ±  2%  perf-profile.self.cycles-pp.copy_present_pte
>       1.12            -0.2        0.91 ±  2%  perf-profile.self.cycles-pp.__anon_vma_interval_tree_remove
>       0.77 ±  2%      -0.2        0.57 ±  3%  perf-profile.self.cycles-pp.kmem_cache_free
>       1.24            -0.2        1.05 ±  2%  perf-profile.self.cycles-pp.dup_mmap
>       0.47 ±  4%      -0.2        0.28 ±  6%  perf-profile.self.cycles-pp.clear_page_erms
>       0.90 ±  2%      -0.2        0.72 ±  2%  perf-profile.self.cycles-pp.free_swap_cache
>       1.17            -0.1        1.03        perf-profile.self.cycles-pp._raw_spin_lock_irqsave
>       0.51            -0.1        0.38 ±  4%  perf-profile.self.cycles-pp.__list_del_entry_valid
>       0.63            -0.1        0.50 ±  2%  perf-profile.self.cycles-pp.unlink_anon_vmas
>       0.46            -0.1        0.33 ±  3%  perf-profile.self.cycles-pp.unmap_single_vma
>       0.51            -0.1        0.40        perf-profile.self.cycles-pp.kmem_cache_alloc
>       0.36 ±  3%      -0.1        0.26 ±  6%  perf-profile.self.cycles-pp.copy_mc_enhanced_fast_string
>       0.34 ±  2%      -0.1        0.24 ±  3%  perf-profile.self.cycles-pp.acct_collect
>       0.64 ±  2%      -0.1        0.54 ±  3%  perf-profile.self.cycles-pp.fput
>       0.28            -0.1        0.19 ±  4%  perf-profile.self.cycles-pp.queued_write_lock_slowpath
>       0.64            -0.1        0.55 ±  2%  perf-profile.self.cycles-pp.mod_objcg_state
>       0.46            -0.1        0.38 ±  3%  perf-profile.self.cycles-pp.update_sg_wakeup_stats
>       0.72            -0.1        0.64 ±  3%  perf-profile.self.cycles-pp.update_rq_clock_task
>       0.40 ±  2%      -0.1        0.32 ±  2%  perf-profile.self.cycles-pp.attach_entity_load_avg
>       0.51 ±  2%      -0.1        0.43 ±  3%  perf-profile.self.cycles-pp.__percpu_counter_sum
>       0.42 ±  2%      -0.1        0.35 ± 10%  perf-profile.self.cycles-pp.sync_regs
>       0.38            -0.1        0.31 ±  4%  perf-profile.self.cycles-pp.remove_entity_load_avg
>       0.40            -0.1        0.33 ±  3%  perf-profile.self.cycles-pp.pcpu_alloc
>       0.21 ±  2%      -0.1        0.14 ±  6%  perf-profile.self.cycles-pp.down_read_trylock
>       0.31            -0.1        0.24 ±  3%  perf-profile.self.cycles-pp.update_load_avg
>       1.26            -0.1        1.19        perf-profile.self.cycles-pp._raw_spin_lock
>       0.34 ± 12%      -0.1        0.28 ±  9%  perf-profile.self.cycles-pp.rwsem_optimistic_spin
>       0.27            -0.1        0.20 ±  2%  perf-profile.self.cycles-pp.kfree
>       0.14 ±  3%      -0.1        0.07 ± 10%  perf-profile.self.cycles-pp.intel_idle
>       0.23 ±  2%      -0.1        0.17 ±  5%  perf-profile.self.cycles-pp.mtree_range_walk
>       0.28            -0.1        0.22 ±  3%  perf-profile.self.cycles-pp.__update_load_avg_cfs_rq
>       0.20            -0.1        0.14 ±  4%  perf-profile.self.cycles-pp.lock_vma_under_rcu
>       0.29 ±  3%      -0.1        0.23 ±  5%  perf-profile.self.cycles-pp.percpu_counter_add_batch
>       0.19 ±  6%      -0.1        0.14 ±  7%  perf-profile.self.cycles-pp.memcg_account_kmem
>       0.18 ± 11%      -0.1        0.12 ± 12%  perf-profile.self.cycles-pp.cgroup_rstat_updated
>       0.34 ±  2%      -0.1        0.28 ±  3%  perf-profile.self.cycles-pp.mas_next_slot
>       0.24 ±  3%      -0.1        0.18 ±  4%  perf-profile.self.cycles-pp.unregister_fair_sched_group
>       0.20 ± 10%      -0.1        0.14 ± 13%  perf-profile.self.cycles-pp.rwsem_down_write_slowpath
>       0.39            -0.1        0.34 ±  2%  perf-profile.self.cycles-pp.___slab_alloc
>       0.13 ±  9%      -0.1        0.08 ±  6%  perf-profile.self.cycles-pp.__memcg_kmem_charge_page
>       0.33 ±  2%      -0.0        0.28 ±  2%  perf-profile.self.cycles-pp.__rb_erase_color
>       0.08 ±  6%      -0.0        0.02 ± 99%  perf-profile.self.cycles-pp.__kmem_cache_alloc_bulk
>       0.74            -0.0        0.69 ±  2%  perf-profile.self.cycles-pp.vma_interval_tree_remove
>       0.16 ±  7%      -0.0        0.12 ±  8%  perf-profile.self.cycles-pp.try_charge_memcg
>       0.07 ± 12%      -0.0        0.02 ± 99%  perf-profile.self.cycles-pp.rb_next
>       0.08 ± 17%      -0.0        0.04 ± 71%  perf-profile.self.cycles-pp.____machine__findnew_thread
>       0.17 ±  6%      -0.0        0.12 ±  7%  perf-profile.self.cycles-pp.mutex_spin_on_owner
>       0.47 ±  3%      -0.0        0.42 ±  4%  perf-profile.self.cycles-pp.osq_unlock
>       0.24 ±  2%      -0.0        0.20 ±  5%  perf-profile.self.cycles-pp.__memcpy
>       0.22 ±  4%      -0.0        0.18 ±  4%  perf-profile.self.cycles-pp.vma_interval_tree_insert_after
>       0.12            -0.0        0.08 ±  7%  perf-profile.self.cycles-pp.cpu_util
>       0.18 ±  5%      -0.0        0.14 ±  2%  perf-profile.self.cycles-pp.do_set_pte
>       0.06 ±  7%      -0.0        0.02 ± 99%  perf-profile.self.cycles-pp.remove_vma
>       0.17 ±  6%      -0.0        0.13 ±  3%  perf-profile.self.cycles-pp.obj_cgroup_charge
>       0.22 ±  2%      -0.0        0.18 ±  3%  perf-profile.self.cycles-pp.__rb_insert_augmented
>       0.18 ±  2%      -0.0        0.14 ±  4%  perf-profile.self.cycles-pp.__list_add_valid
>       0.22 ±  4%      -0.0        0.18 ±  2%  perf-profile.self.cycles-pp.__anon_vma_interval_tree_augment_rotate
>       0.07 ±  5%      -0.0        0.03 ± 70%  perf-profile.self.cycles-pp.mark_page_accessed
>       0.16 ±  3%      -0.0        0.12 ±  6%  perf-profile.self.cycles-pp.perf_adjust_freq_unthr_context
>       0.06            -0.0        0.02 ± 99%  perf-profile.self.cycles-pp.mm_init
>       0.15 ±  5%      -0.0        0.12 ±  4%  perf-profile.self.cycles-pp.__libc_fork
>       0.07 ±  5%      -0.0        0.04 ± 44%  perf-profile.self.cycles-pp.__update_blocked_fair
>       0.10 ±  3%      -0.0        0.07        perf-profile.self.cycles-pp.queued_read_lock_slowpath
>       0.24 ±  2%      -0.0        0.21 ±  2%  perf-profile.self.cycles-pp.deactivate_slab
>       0.17 ±  4%      -0.0        0.14        perf-profile.self.cycles-pp.vm_normal_page
>       0.18 ±  2%      -0.0        0.16 ±  3%  perf-profile.self.cycles-pp.get_partial_node
>       0.09 ±  5%      -0.0        0.06        perf-profile.self.cycles-pp.__vm_area_free
>       0.06            -0.0        0.03 ± 70%  perf-profile.self.cycles-pp.pte_offset_map_nolock
>       0.13 ±  2%      -0.0        0.10 ±  3%  perf-profile.self.cycles-pp.online_fair_sched_group
>       0.10 ±  3%      -0.0        0.08 ±  6%  perf-profile.self.cycles-pp.free_pgtables
>       0.08 ± 11%      -0.0        0.06 ±  9%  perf-profile.self.cycles-pp.stress_fork_fn
>       0.14 ±  5%      -0.0        0.12 ±  3%  perf-profile.self.cycles-pp._find_next_bit
>       0.09 ±  5%      -0.0        0.07 ±  5%  perf-profile.self.cycles-pp.mab_mas_cp
>       0.13 ±  2%      -0.0        0.11 ±  4%  perf-profile.self.cycles-pp.__unfreeze_partials
>       0.12 ±  4%      -0.0        0.10 ±  3%  perf-profile.self.cycles-pp.__mod_memcg_lruvec_state
>       0.10 ±  4%      -0.0        0.09 ±  5%  perf-profile.self.cycles-pp._raw_spin_trylock
>       0.11 ±  6%      -0.0        0.09        perf-profile.self.cycles-pp.unmap_page_range
>       0.09 ±  6%      -0.0        0.07 ±  6%  perf-profile.self.cycles-pp.__schedule
>       0.08 ±  5%      -0.0        0.07 ±  7%  perf-profile.self.cycles-pp.update_curr
>       0.08 ±  6%      -0.0        0.06 ±  6%  perf-profile.self.cycles-pp._raw_write_lock_irq
>       0.10 ±  4%      -0.0        0.08 ±  5%  perf-profile.self.cycles-pp.get_obj_cgroup_from_current
>       0.07 ±  6%      -0.0        0.06        perf-profile.self.cycles-pp.free_percpu
>       0.06 ±  7%      -0.0        0.05 ±  7%  perf-profile.self.cycles-pp._find_next_zero_bit
>       0.07            -0.0        0.06 ±  6%  perf-profile.self.cycles-pp.mas_store
>       0.10 ±  4%      -0.0        0.09 ±  5%  perf-profile.self.cycles-pp.__update_load_avg_blocked_se
>       0.06 ±  6%      -0.0        0.05        perf-profile.self.cycles-pp.__radix_tree_lookup
>       0.07            -0.0        0.06        perf-profile.self.cycles-pp.copy_process
>       0.05 ±  8%      +0.0        0.07        perf-profile.self.cycles-pp.exit_to_user_mode_prepare
>       0.22 ±  3%      +0.0        0.24        perf-profile.self.cycles-pp.alloc_fair_sched_group
>       0.20 ±  3%      +0.1        0.25 ±  3%  perf-profile.self.cycles-pp.init_entity_runnable_average
>       0.28 ±  3%      +0.1        0.35 ±  3%  perf-profile.self.cycles-pp.__kmem_cache_alloc_node
>       0.05            +0.1        0.17 ±  7%  perf-profile.self.cycles-pp.init_tg_cfs_entry
>       0.43 ±  3%      +0.2        0.58 ±  3%  perf-profile.self.cycles-pp.page_add_file_rmap
>       0.00            +4.5        4.50        perf-profile.self.cycles-pp.next_uptodate_folio
>       0.90           +20.6       21.46 ±  8%  perf-profile.self.cycles-pp.filemap_map_pages
> 
> 
> 
> ***************************************************************************************************
> lkp-spr-2sp3: 224 threads 2 sockets Intel(R) Xeon(R) Platinum 8480L (Sapphire Rapids) with 512G memory
> =========================================================================================
> compiler/cpufreq_governor/kconfig/rootfs/runtime/tbox_group/test/testcase:
>   gcc-12/performance/x86_64-rhel-8.3/debian-11.1-x86_64-20220510.cgz/300s/lkp-spr-2sp3/mmap-pread-rand/vm-scalability
> 
> commit: 
>   9f1f5b60e7 ("mm: use flush_icache_pages() in do_set_pmd()")
>   de74976eb6 ("filemap: add filemap_map_folio_range()")
> 
> 9f1f5b60e76d44fa de74976eb65151a2f568e477fc2 
> ---------------- --------------------------- 
>          %stddev     %change         %stddev
>              \          |                \  
>      61637           -11.8%      54369        vm-scalability.median
>       1671 ± 51%    -499.9        1171 ±  4%  vm-scalability.stddev%
>   13458014           -11.1%   11962548        vm-scalability.throughput
>       2750 ± 12%    +293.5%      10822 ±  5%  vm-scalability.time.system_time
>      64216           -11.7%      56706        vm-scalability.time.user_time
>   4.03e+09           -10.8%  3.596e+09        vm-scalability.workload
>     159605            +1.8%     162466        proc-vmstat.nr_shmem
>      92.33           -12.1%      81.17        vmstat.cpu.us
>       0.01 ±  3%      +0.0        0.02 ±  7%  mpstat.cpu.all.soft%
>       4.02 ± 12%     +11.7       15.68 ±  5%  mpstat.cpu.all.sys%
>      93.19           -11.2       82.03        mpstat.cpu.all.usr%
>       0.01 ±  5%     -13.0%       0.01 ±  6%  perf-sched.sch_delay.avg.ms.schedule_timeout.kcompactd.kthread.ret_from_fork
>       2.02 ± 45%     +48.0%       2.99 ± 15%  perf-sched.wait_and_delay.avg.ms.rcu_gp_kthread.kthread.ret_from_fork.ret_from_fork_asm
>       2.42 ±  8%     +23.1%       2.98 ± 15%  perf-sched.wait_time.avg.ms.rcu_gp_kthread.kthread.ret_from_fork.ret_from_fork_asm
>     177.45           -19.7      157.76        turbostat.PKG_%
>     694.20            -2.1%     679.50        turbostat.PkgWatt
>      62.93            -7.5%      58.23        turbostat.RAMWatt
>       0.49 ±  3%      -0.0        0.46        perf-profile.children.cycles-pp.hrtimer_interrupt
>       0.43 ±  5%      -0.0        0.40        perf-profile.children.cycles-pp.__hrtimer_run_queues
>       0.50 ±  3%      -0.0        0.47        perf-profile.children.cycles-pp.__sysvec_apic_timer_interrupt
>       0.40 ±  5%      -0.0        0.37 ±  2%  perf-profile.children.cycles-pp.tick_sched_timer
>       0.38 ±  5%      -0.0        0.35        perf-profile.children.cycles-pp.update_process_times
>     943206           +41.7%    1336270 ±  7%  sched_debug.cpu.avg_idle.avg
>    1823893 ± 13%    +100.4%    3655497 ± 16%  sched_debug.cpu.avg_idle.max
>     172988 ± 16%    +284.9%     665748 ± 21%  sched_debug.cpu.avg_idle.stddev
>     587283 ±  7%     +31.5%     772030 ±  8%  sched_debug.cpu.max_idle_balance_cost.max
>       7394 ± 32%    +592.1%      51171 ± 31%  sched_debug.cpu.max_idle_balance_cost.stddev
>      22.38            -4.9%      21.27        perf-stat.i.MPKI
>  3.153e+10           -10.2%  2.833e+10        perf-stat.i.branch-instructions
>       0.03 ±  5%      +0.0        0.07 ±  3%  perf-stat.i.branch-miss-rate%
>      95.88            -4.4       91.45        perf-stat.i.cache-miss-rate%
>  2.907e+09           -10.3%  2.607e+09        perf-stat.i.cache-misses
>  2.991e+09           -10.1%   2.69e+09        perf-stat.i.cache-references
>       5.07 ±  2%    +116.1%      10.97 ±  8%  perf-stat.i.cpi
>     295.41 ±  4%    +405.0%       1491 ±  9%  perf-stat.i.cycles-between-cache-misses
>       3.53            -0.3        3.22        perf-stat.i.dTLB-load-miss-rate%
>  1.468e+09           -10.4%  1.316e+09        perf-stat.i.dTLB-load-misses
>  3.905e+10           -10.1%  3.509e+10        perf-stat.i.dTLB-loads
>       0.02 ±  2%      +0.0        0.05 ±  8%  perf-stat.i.dTLB-store-miss-rate%
>    2215806 ±  2%      -9.0%    2016922 ±  3%  perf-stat.i.dTLB-store-misses
>  1.214e+10           -10.0%  1.094e+10        perf-stat.i.dTLB-stores
>  1.325e+11           -10.1%  1.191e+11        perf-stat.i.instructions
>       0.20 ±  2%     -10.4%       0.18        perf-stat.i.ipc
>      29.74 ±  9%    +198.9%      88.90 ± 11%  perf-stat.i.metric.K/sec
>     395.53           -10.2%     355.27        perf-stat.i.metric.M/sec
>     203404 ±  3%      -7.0%     189209        perf-stat.i.minor-faults
>  1.062e+09 ±  2%     -15.6%  8.962e+08 ±  2%  perf-stat.i.node-load-misses
>  1.831e+09 ±  2%      -7.4%  1.696e+09 ±  3%  perf-stat.i.node-loads
>     204552 ±  3%      -6.9%     190351        perf-stat.i.page-faults
>       0.02 ±  2%      +0.0        0.02 ±  2%  perf-stat.overall.branch-miss-rate%
>       4.98           +14.6%       5.70        perf-stat.overall.cpi
>     226.86           +15.1%     261.02        perf-stat.overall.cycles-between-cache-misses
>       0.20           -12.8%       0.18        perf-stat.overall.ipc
>      10081            -1.4%       9943        perf-stat.overall.path-length
>  3.143e+10           -12.4%  2.753e+10        perf-stat.ps.branch-instructions
>  2.896e+09           -12.6%   2.53e+09        perf-stat.ps.cache-misses
>   2.98e+09           -12.3%  2.613e+09        perf-stat.ps.cache-references
>  1.463e+09           -12.7%  1.277e+09        perf-stat.ps.dTLB-load-misses
>  3.892e+10           -12.4%  3.411e+10        perf-stat.ps.dTLB-loads
>    2296781 ±  2%      -7.7%    2120487 ±  3%  perf-stat.ps.dTLB-store-misses
>  1.211e+10           -12.1%  1.065e+10        perf-stat.ps.dTLB-stores
>  1.321e+11           -12.3%  1.158e+11        perf-stat.ps.instructions
>  1.058e+09 ±  2%     -17.8%  8.694e+08 ±  2%  perf-stat.ps.node-load-misses
>  1.824e+09 ±  2%      -9.8%  1.646e+09 ±  3%  perf-stat.ps.node-loads
>  4.063e+13           -12.0%  3.576e+13        perf-stat.total.instructions
> 
> 
> 
> ***************************************************************************************************
> lkp-spr-r02: 224 threads 2 sockets Intel(R) Xeon(R) Platinum 8480CTDX (Sapphire Rapids) with 256G memory
> =========================================================================================
> class/compiler/cpufreq_governor/kconfig/nr_threads/rootfs/sc_pid_max/tbox_group/test/testcase/testtime:
>   scheduler/gcc-12/performance/x86_64-rhel-8.3/100%/debian-11.1-x86_64-20220510.cgz/4194304/lkp-spr-r02/zombie/stress-ng/60s
> 
> commit: 
>   9f1f5b60e7 ("mm: use flush_icache_pages() in do_set_pmd()")
>   de74976eb6 ("filemap: add filemap_map_folio_range()")
> 
> 9f1f5b60e76d44fa de74976eb65151a2f568e477fc2 
> ---------------- --------------------------- 
>          %stddev     %change         %stddev
>              \          |                \  
>     175389 ±  8%     +48.3%     260092 ± 14%  sched_debug.cpu.avg_idle.stddev
>  1.052e+09 ±  4%    +177.7%  2.921e+09 ± 59%  cpuidle..time
>    1486111 ±  8%   +1935.1%   30243535 ± 87%  cpuidle..usage
>       0.01 ± 40%      +5.1        5.09 ± 69%  mpstat.cpu.all.iowait%
>       1.07 ±  3%      -0.8        0.30 ± 98%  mpstat.cpu.all.soft%
>       1.06            -0.5        0.52 ± 61%  mpstat.cpu.all.usr%
>       0.00       +1.2e+107%     122169 ± 65%  vmstat.procs.b
>     291.67 ±  4%   +2001.1%       6128 ± 88%  vmstat.procs.r
>     356111           +60.6%     571836 ± 54%  vmstat.system.in
>      57294         +1320.7%     813995 ±119%  meminfo.KernelStack
>      74462 ±  3%   +3257.4%    2499959 ±130%  meminfo.PageTables
>   25447288           -39.9%   15287038 ± 41%  meminfo.SUnreclaim
>   25634838           -39.4%   15544405 ± 40%  meminfo.Slab
>   66497461 ±  3%     -51.9%   31974064 ± 48%  numa-numastat.node0.local_node
>   66655914 ±  3%     -51.6%   32294668 ± 47%  numa-numastat.node0.numa_hit
>   70406387 ±  3%     -47.7%   36843120 ± 52%  numa-numastat.node1.local_node
>   70506925 ±  3%     -47.4%   37117073 ± 52%  numa-numastat.node1.numa_hit
>      78192 ± 50%    +174.0%     214273 ± 38%  numa-numastat.node1.other_node
>       3390 ± 18%   +4084.9%     141867 ± 53%  turbostat.C1
>       0.15 ±  2%     -56.0%       0.07 ± 44%  turbostat.IPC
>       2056 ± 25%  +1.4e+06%   28675451 ± 92%  turbostat.POLL
>     609.02           -12.0%     535.83 ±  6%  turbostat.PkgWatt
>      35.37           -35.4%      22.85 ± 19%  turbostat.RAMWatt
>     271305 ±  9%    +179.1%     757275 ± 38%  stress-ng.time.involuntary_context_switches
>      66094 ± 15%   +3591.7%    2440028 ± 71%  stress-ng.time.major_page_faults
>  1.498e+08 ±  4%     -54.2%   68628961 ± 56%  stress-ng.time.minor_page_faults
>      12962          +276.6%      48813 ± 81%  stress-ng.time.system_time
>     121.53 ±  3%     -56.3%      53.13 ± 55%  stress-ng.time.user_time
>    3643735           -53.8%    1681822 ± 56%  stress-ng.zombie.ops
>      56119           -74.5%      14319 ±105%  stress-ng.zombie.ops_per_sec
>     385088 ± 37%    +817.9%    3534727 ± 80%  numa-meminfo.node0.AnonPages.max
>      28420 ±  7%   +1062.4%     330365 ±140%  numa-meminfo.node0.KernelStack
>      35679 ± 14%   +2710.3%    1002703 ±153%  numa-meminfo.node0.PageTables
>   12105886 ±  3%     -40.7%    7174505 ± 47%  numa-meminfo.node0.SUnreclaim
>   12220251 ±  3%     -40.4%    7283187 ± 47%  numa-meminfo.node0.Slab
>    1290286 ± 12%     -46.2%     694602 ± 49%  numa-meminfo.node1.FilePages
>     610714 ±  2%     -41.0%     360265 ± 46%  numa-meminfo.node1.Mapped
>   13247897 ±  3%     -45.3%    7243195 ± 51%  numa-meminfo.node1.SUnreclaim
>    1217022 ± 10%     -45.6%     662370 ± 46%  numa-meminfo.node1.Shmem
>   13320721 ±  3%     -44.2%    7434743 ± 48%  numa-meminfo.node1.Slab
>       0.00       +1.7e+104%     172.83 ±138%  proc-vmstat.nr_isolated_anon
>      57369         +1059.9%     665452 ±120%  proc-vmstat.nr_kernel_stack
>      18723 ±  3%   +2604.6%     506387 ±132%  proc-vmstat.nr_page_table_pages
>    6342780           -42.4%    3650813 ± 46%  proc-vmstat.nr_slab_unreclaimable
>  1.371e+08 ±  3%     -49.4%   69417360 ± 50%  proc-vmstat.numa_hit
>  1.369e+08 ±  3%     -49.7%   68822802 ± 50%  proc-vmstat.numa_local
>  1.547e+08 ±  2%     -48.5%   79708856 ± 49%  proc-vmstat.pgalloc_normal
>   1.52e+08 ±  4%     -50.7%   74893245 ± 54%  proc-vmstat.pgfault
>  1.524e+08 ±  2%     -49.2%   77369894 ± 49%  proc-vmstat.pgfree
>    9681032 ±  7%     -57.7%    4096941 ± 60%  proc-vmstat.pgreuse
>      28414 ±  7%    +949.9%     298310 ±166%  numa-vmstat.node0.nr_kernel_stack
>       8996 ± 14%   +2424.4%     227103 ±180%  numa-vmstat.node0.nr_page_table_pages
>    3025579 ±  3%     -42.3%    1745312 ± 52%  numa-vmstat.node0.nr_slab_unreclaimable
>   66655838 ±  3%     -51.6%   32293907 ± 47%  numa-vmstat.node0.numa_hit
>   66497386 ±  3%     -51.9%   31973303 ± 48%  numa-vmstat.node0.numa_local
>     322703 ± 12%     -50.1%     161011 ± 58%  numa-vmstat.node1.nr_file_pages
>     152754 ±  2%     -44.9%      84140 ± 57%  numa-vmstat.node1.nr_mapped
>     304388 ± 10%     -49.8%     152952 ± 56%  numa-vmstat.node1.nr_shmem
>    3310966 ±  3%     -46.8%    1761156 ± 53%  numa-vmstat.node1.nr_slab_unreclaimable
>   70504864 ±  3%     -47.4%   37117466 ± 52%  numa-vmstat.node1.numa_hit
>   70404328 ±  3%     -47.7%   36843513 ± 52%  numa-vmstat.node1.numa_local
>      78191 ± 50%    +174.0%     214273 ± 38%  numa-vmstat.node1.numa_other
>       9.37 ±  2%     -20.6%       7.44 ± 15%  perf-stat.i.MPKI
>   1.58e+08           -15.9%  1.329e+08 ± 11%  perf-stat.i.branch-misses
>  5.873e+08 ±  3%     -38.8%  3.592e+08 ± 14%  perf-stat.i.cache-misses
>  1.574e+09 ±  2%     -36.4%      1e+09 ± 15%  perf-stat.i.cache-references
>      78615           +80.3%     141727 ± 10%  perf-stat.i.context-switches
>       3.69           +48.7%       5.48 ± 23%  perf-stat.i.cpi
>      28686 ±  5%     +68.0%      48183 ± 24%  perf-stat.i.cpu-migrations
>       1094 ±  4%    +318.6%       4582 ± 87%  perf-stat.i.cycles-between-cache-misses
>   93120973 ±  2%     -24.2%   70578623 ± 21%  perf-stat.i.dTLB-load-misses
>       0.56            -0.1        0.41 ± 23%  perf-stat.i.dTLB-store-miss-rate%
>   59029929 ±  2%     -39.3%   35849349 ± 26%  perf-stat.i.dTLB-store-misses
>  9.925e+09 ±  2%     -26.1%  7.339e+09 ± 15%  perf-stat.i.dTLB-stores
>       1009 ± 15%   +3078.5%      32082 ± 24%  perf-stat.i.major-faults
>       2.67           -10.3%       2.39 ± 10%  perf-stat.i.metric.GHz
>       1003 ±  2%     -29.4%     708.55 ± 16%  perf-stat.i.metric.K/sec
>     387.65           -37.1%     243.89 ± 22%  perf-stat.i.metric.M/sec
>    2300158 ±  4%     -32.9%    1542813 ± 19%  perf-stat.i.minor-faults
>  1.368e+08 ±  2%     -26.3%  1.008e+08 ± 14%  perf-stat.i.node-load-misses
>    2301168 ±  4%     -31.4%    1578147 ± 18%  perf-stat.i.page-faults
>       9.65 ±  2%     -44.6%       5.35 ± 51%  perf-stat.overall.MPKI
>       3.70          +125.9%       8.36 ± 37%  perf-stat.overall.cpi
>       1021 ±  3%    +631.6%       7471 ± 88%  perf-stat.overall.cycles-between-cache-misses
>       0.59            -0.2        0.38 ± 39%  perf-stat.overall.dTLB-store-miss-rate%
>       0.27           -47.4%       0.14 ± 43%  perf-stat.overall.ipc
>  3.248e+10           -45.9%  1.758e+10 ± 27%  perf-stat.ps.branch-instructions
>   1.53e+08           -59.0%   62792893 ± 80%  perf-stat.ps.branch-misses
>  5.787e+08 ±  3%     -69.5%  1.767e+08 ± 76%  perf-stat.ps.cache-misses
>  1.541e+09 ±  2%     -68.0%   4.93e+08 ± 77%  perf-stat.ps.cache-references
>   91095243 ±  2%     -59.5%   36919468 ± 61%  perf-stat.ps.dTLB-load-misses
>  4.134e+10           -51.2%  2.016e+10 ± 37%  perf-stat.ps.dTLB-loads
>   57601991 ±  3%     -68.8%   17958458 ± 97%  perf-stat.ps.dTLB-store-misses
>  9.696e+09 ±  2%     -62.5%  3.641e+09 ± 78%  perf-stat.ps.dTLB-stores
>  1.596e+11           -50.2%  7.949e+10 ± 35%  perf-stat.ps.instructions
>     983.10 ± 16%   +1528.2%      16006 ± 47%  perf-stat.ps.major-faults
>    2249530 ±  5%     -65.8%     769259 ± 89%  perf-stat.ps.minor-faults
>  1.351e+08 ±  2%     -62.6%   50573350 ± 80%  perf-stat.ps.node-load-misses
>   21223230 ±  5%     -51.4%   10324363 ± 54%  perf-stat.ps.node-loads
>    2250513 ±  5%     -65.1%     785451 ± 87%  perf-stat.ps.page-faults
> 
> 
> 
> 
> 
> Disclaimer:
> Results have been estimated based on internal Intel analysis and are provided
> for informational purposes only. Any difference in system hardware or software
> design or configuration may affect actual performance.
> 
> 

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ