lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [thread-next>] [day] [month] [year] [list]
Message-ID: <202401221624.cb53a8ca-oliver.sang@intel.com>
Date: Mon, 22 Jan 2024 16:39:05 +0800
From: kernel test robot <oliver.sang@...el.com>
To: Yosry Ahmed <yosryahmed@...gle.com>
CC: <oe-lkp@...ts.linux.dev>, <lkp@...el.com>, <linux-kernel@...r.kernel.org>,
	Andrew Morton <akpm@...ux-foundation.org>, Johannes Weiner
	<hannes@...xchg.org>, Domenico Cerasuolo <cerasuolodomenico@...il.com>,
	Shakeel Butt <shakeelb@...gle.com>, Chris Li <chrisl@...nel.org>, Greg Thelen
	<gthelen@...gle.com>, Ivan Babrou <ivan@...udflare.com>, Michal Hocko
	<mhocko@...nel.org>, Michal Koutny <mkoutny@...e.com>, Muchun Song
	<muchun.song@...ux.dev>, Roman Gushchin <roman.gushchin@...ux.dev>, Tejun Heo
	<tj@...nel.org>, Waiman Long <longman@...hat.com>, Wei Xu
	<weixugc@...gle.com>, <cgroups@...r.kernel.org>, <linux-mm@...ck.org>,
	<ying.huang@...el.com>, <feng.tang@...el.com>, <fengwei.yin@...el.com>,
	<oliver.sang@...el.com>
Subject: [linus:master] [mm]  8d59d2214c:  vm-scalability.throughput -36.6%
 regression



hi, Yosry Ahmed,

per your suggestion in
https://lore.kernel.org/all/CAJD7tkameJBrJQxRj+ibKL6-yd-i0wyoyv2cgZdh3ZepA1p7wA@mail.gmail.com/
"I think it would be useful to know if there are
regressions/improvements in other microbenchmarks, at least to
investigate whether they represent real regressions."

we still report below two regressions to you just FYI what we observed in our
microbenchmark tests.
(we still captured will-it-scale::fallocate regression but ignore here per
your commit message)


Hello,

kernel test robot noticed a -36.6% regression of vm-scalability.throughput on:


commit: 8d59d2214c2362e7a9d185d80b613e632581af7b ("mm: memcg: make stats flushing threshold per-memcg")
https://git.kernel.org/cgit/linux/kernel/git/torvalds/linux.git master

testcase: vm-scalability
test machine: 224 threads 4 sockets Intel(R) Xeon(R) Platinum 8380H CPU @ 2.90GHz (Cooper Lake) with 192G memory
parameters:

	runtime: 300s
	size: 1T
	test: lru-shm
	cpufreq_governor: performance

test-description: The motivation behind this suite is to exercise functions and regions of the mm/ of the Linux kernel which are of interest to us.
test-url: https://git.kernel.org/cgit/linux/kernel/git/wfg/vm-scalability.git/

In addition to that, the commit also has significant impact on the following tests:

+------------------+----------------------------------------------------------------------------------------------------+
| testcase: change | will-it-scale: will-it-scale.per_process_ops -32.3% regression                                     |
| test machine     | 104 threads 2 sockets (Skylake) with 192G memory                                                   |
| test parameters  | cpufreq_governor=performance                                                                       |
|                  | mode=process                                                                                       |
|                  | nr_task=50%                                                                                        |
|                  | test=tlb_flush2                                                                                    |
+------------------+----------------------------------------------------------------------------------------------------+


If you fix the issue in a separate patch/commit (i.e. not just a new version of
the same patch/commit), kindly add following tags
| Reported-by: kernel test robot <oliver.sang@...el.com>
| Closes: https://lore.kernel.org/oe-lkp/202401221624.cb53a8ca-oliver.sang@intel.com


Details are as below:
-------------------------------------------------------------------------------------------------->


The kernel config and materials to reproduce are available at:
https://download.01.org/0day-ci/archive/20240122/202401221624.cb53a8ca-oliver.sang@intel.com

=========================================================================================
compiler/cpufreq_governor/kconfig/rootfs/runtime/size/tbox_group/test/testcase:
  gcc-12/performance/x86_64-rhel-8.3/debian-11.1-x86_64-20220510.cgz/300s/1T/lkp-cpl-4sp2/lru-shm/vm-scalability

commit: 
  e0bf1dc859 ("mm: memcg: move vmstats structs definition above flushing code")
  8d59d2214c ("mm: memcg: make stats flushing threshold per-memcg")

e0bf1dc859fdd08e 8d59d2214c2362e7a9d185d80b6 
---------------- --------------------------- 
         %stddev     %change         %stddev
             \          |                \  
      0.01           +86.7%       0.02        vm-scalability.free_time
    946447           -37.8%     588327        vm-scalability.median
 2.131e+08           -36.6%  1.351e+08        vm-scalability.throughput
    284.74            +6.3%     302.62        vm-scalability.time.elapsed_time
    284.74            +6.3%     302.62        vm-scalability.time.elapsed_time.max
     30485           +14.8%      34987        vm-scalability.time.involuntary_context_switches
      1893           +43.6%       2718        vm-scalability.time.percent_of_cpu_this_job_got
      3855           +67.7%       6467        vm-scalability.time.system_time
      1537           +14.5%       1760        vm-scalability.time.user_time
    120009            -5.6%     113290        vm-scalability.time.voluntary_context_switches
      6.46            +3.5        9.95        mpstat.cpu.all.sys%
     21.22           +38.8%      29.46        vmstat.procs.r
      0.01 ± 20%   +1887.0%       0.18 ±203%  perf-sched.sch_delay.max.ms.__cond_resched.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
      0.01 ± 28%     +63.3%       0.01 ± 29%  perf-sched.sch_delay.max.ms.wait_for_partner.fifo_open.do_dentry_open.do_open
    113624 ±  5%     +14.0%     129566 ±  3%  meminfo.Active
    113476 ±  5%     +14.0%     129417 ±  3%  meminfo.Active(anon)
   3987746           +46.0%    5821636        meminfo.Mapped
     16345           +14.6%      18729        meminfo.PageTables
    474.17 ±  3%     -88.9%      52.50 ±125%  perf-c2c.DRAM.local
    483.17 ±  5%     -79.3%      99.83 ± 70%  perf-c2c.DRAM.remote
      1045 ±  5%     -71.9%     294.00 ± 63%  perf-c2c.HITM.local
    119.50 ± 10%     -78.8%      25.33 ± 20%  perf-c2c.HITM.remote
    392.33           +35.4%     531.17        turbostat.Avg_MHz
     10.35            +3.7       14.00        turbostat.Busy%
     90.56            -3.7       86.86        turbostat.C1%
      0.28 ±  5%     -31.5%       0.19        turbostat.IPC
    481.33            +2.5%     493.38        turbostat.PkgWatt
    999019 ±  3%     +44.4%    1442651 ±  2%  numa-meminfo.node0.Mapped
   1005687 ±  4%     +44.1%    1449402 ±  3%  numa-meminfo.node1.Mapped
      3689 ±  3%     +21.7%       4490 ±  7%  numa-meminfo.node1.PageTables
    980589 ±  2%     +42.3%    1395777 ±  2%  numa-meminfo.node2.Mapped
     96484 ±  5%     +22.0%     117715 ±  4%  numa-meminfo.node3.Active
     96430 ±  5%     +22.1%     117694 ±  4%  numa-meminfo.node3.Active(anon)
    991367 ±  3%     +42.7%    1414337 ±  4%  numa-meminfo.node3.Mapped
    251219 ±  3%     +44.8%     363745 ±  2%  numa-vmstat.node0.nr_mapped
    253252 ±  2%     +44.6%     366087 ±  3%  numa-vmstat.node1.nr_mapped
    927.67 ±  3%     +21.9%       1130 ±  7%  numa-vmstat.node1.nr_page_table_pages
    248171 ±  2%     +42.5%     353541 ±  4%  numa-vmstat.node2.nr_mapped
     24188 ±  5%     +21.6%      29410 ±  4%  numa-vmstat.node3.nr_active_anon
    245825 ±  2%     +45.5%     357622 ±  3%  numa-vmstat.node3.nr_mapped
      1038 ± 11%     +17.8%       1224 ±  6%  numa-vmstat.node3.nr_page_table_pages
     24188 ±  5%     +21.6%      29410 ±  4%  numa-vmstat.node3.nr_zone_active_anon
     28376 ±  5%     +14.0%      32338 ±  3%  proc-vmstat.nr_active_anon
    993504           +46.6%    1456136        proc-vmstat.nr_mapped
      4060           +15.5%       4691        proc-vmstat.nr_page_table_pages
     28376 ±  5%     +14.0%      32338 ±  3%  proc-vmstat.nr_zone_active_anon
 1.066e+09            -2.0%  1.045e+09        proc-vmstat.numa_hit
 1.065e+09            -2.0%  1.044e+09        proc-vmstat.numa_local
      5659            +5.6%       5978        proc-vmstat.unevictable_pgs_culled
  34604288            +3.7%   35898496        proc-vmstat.unevictable_pgs_scanned
   1223376 ± 14%    +119.1%    2680582 ±  9%  sched_debug.cfs_rq:/.avg_vruntime.avg
   1673909 ± 14%     +97.6%    3308254 ±  8%  sched_debug.cfs_rq:/.avg_vruntime.max
    810795 ± 15%    +145.8%    1993289 ±  9%  sched_debug.cfs_rq:/.avg_vruntime.min
    156233 ±  8%     +55.1%     242331 ±  6%  sched_debug.cfs_rq:/.avg_vruntime.stddev
   1223376 ± 14%    +119.1%    2680582 ±  9%  sched_debug.cfs_rq:/.min_vruntime.avg
   1673909 ± 14%     +97.6%    3308254 ±  8%  sched_debug.cfs_rq:/.min_vruntime.max
    810795 ± 15%    +145.8%    1993289 ±  9%  sched_debug.cfs_rq:/.min_vruntime.min
    156233 ±  8%     +55.1%     242331 ±  6%  sched_debug.cfs_rq:/.min_vruntime.stddev
    126445 ±  3%     -11.0%     112493 ±  4%  sched_debug.cpu.avg_idle.stddev
      1447 ± 15%     +32.0%       1910 ±  9%  sched_debug.cpu.nr_switches.min
      0.71           +13.4%       0.80        perf-stat.i.MPKI
 2.343e+10            -7.9%  2.157e+10        perf-stat.i.branch-instructions
      0.36            -0.0        0.35        perf-stat.i.branch-miss-rate%
  30833194            -7.3%   28584190        perf-stat.i.branch-misses
     26.04            -1.4       24.66        perf-stat.i.cache-miss-rate%
  51345490 ±  3%     +40.7%   72258633 ±  3%  perf-stat.i.cache-misses
 1.616e+08 ±  6%     +58.6%  2.562e+08 ±  6%  perf-stat.i.cache-references
      1.29            +9.4%       1.42        perf-stat.i.cpi
 8.394e+10           +33.7%  1.122e+11        perf-stat.i.cpu-cycles
    505.77            -2.6%     492.52        perf-stat.i.cpu-migrations
      0.03            +0.0        0.03 ±  2%  perf-stat.i.dTLB-load-miss-rate%
 2.335e+10            -7.4%  2.162e+10        perf-stat.i.dTLB-loads
      0.03            +0.0        0.03        perf-stat.i.dTLB-store-miss-rate%
   3948344            -8.0%    3633633        perf-stat.i.dTLB-store-misses
 6.549e+09            -7.0%   6.09e+09        perf-stat.i.dTLB-stores
  17546602           -22.8%   13551001        perf-stat.i.iTLB-load-misses
   2552560            -2.6%    2485876        perf-stat.i.iTLB-loads
 8.367e+10            -7.5%  7.737e+10        perf-stat.i.instructions
      4706            +7.7%       5070        perf-stat.i.instructions-per-iTLB-miss
      0.81           -12.0%       0.72        perf-stat.i.ipc
      1.59 ±  3%     -22.3%       1.23 ±  4%  perf-stat.i.major-faults
      0.37           +34.2%       0.49        perf-stat.i.metric.GHz
    233.98            -6.9%     217.90        perf-stat.i.metric.M/sec
   3619177            -9.5%    3276556        perf-stat.i.minor-faults
     74.28            +4.8       79.04        perf-stat.i.node-load-miss-rate%
   2898733 ±  4%     +49.0%    4320557        perf-stat.i.node-load-misses
   1928237 ±  4%     -11.9%    1698426        perf-stat.i.node-loads
  13383344 ±  2%      +4.7%   14013398 ±  3%  perf-stat.i.node-stores
   3619179            -9.5%    3276558        perf-stat.i.page-faults
      0.61 ±  3%     +52.5%       0.94 ±  3%  perf-stat.overall.MPKI
     31.95 ±  2%      -3.6       28.34 ±  3%  perf-stat.overall.cache-miss-rate%
      1.00           +45.0%       1.45        perf-stat.overall.cpi
      0.07            +0.0        0.08 ±  4%  perf-stat.overall.dTLB-load-miss-rate%
     87.62            -2.6       85.05        perf-stat.overall.iTLB-load-miss-rate%
      4778           +20.2%       5745        perf-stat.overall.instructions-per-iTLB-miss
      1.00           -31.0%       0.69        perf-stat.overall.ipc
     59.75 ±  3%     +11.8       71.59        perf-stat.overall.node-load-miss-rate%
      5145            +1.8%       5239        perf-stat.overall.path-length
 2.405e+10            -6.3%  2.252e+10        perf-stat.ps.branch-instructions
  31203502            -6.4%   29219514        perf-stat.ps.branch-misses
  52696784 ±  3%     +43.4%   75547948 ±  3%  perf-stat.ps.cache-misses
 1.652e+08 ±  6%     +61.7%  2.672e+08 ±  7%  perf-stat.ps.cache-references
 8.584e+10           +36.3%   1.17e+11        perf-stat.ps.cpu-cycles
    506.29            -2.0%     496.05        perf-stat.ps.cpu-migrations
 2.395e+10            -5.9%  2.254e+10        perf-stat.ps.dTLB-loads
   4059043            -6.2%    3806002        perf-stat.ps.dTLB-store-misses
 6.688e+09            -5.7%  6.308e+09        perf-stat.ps.dTLB-stores
  17944396           -21.8%   14028927        perf-stat.ps.iTLB-load-misses
   2534093            -2.7%    2465233        perf-stat.ps.iTLB-loads
 8.575e+10            -6.0%  8.059e+10        perf-stat.ps.instructions
      1.60 ±  3%     -23.2%       1.23 ±  4%  perf-stat.ps.major-faults
   3726053            -7.7%    3439511        perf-stat.ps.minor-faults
   2942507 ±  4%     +52.0%    4472428        perf-stat.ps.node-load-misses
   1980077 ±  4%     -10.4%    1774633        perf-stat.ps.node-loads
  13780660 ±  2%      +6.8%   14716100 ±  3%  perf-stat.ps.node-stores
   3726055            -7.7%    3439513        perf-stat.ps.page-faults
     37.11            -6.7       30.40 ±  6%  perf-profile.calltrace.cycles-pp.asm_sysvec_apic_timer_interrupt.acpi_safe_halt.acpi_idle_enter.cpuidle_enter_state.cpuidle_enter
     21.14            -3.8       17.36 ±  7%  perf-profile.calltrace.cycles-pp.secondary_startup_64_no_verify
     21.05            -3.8       17.29 ±  7%  perf-profile.calltrace.cycles-pp.do_idle.cpu_startup_entry.start_secondary.secondary_startup_64_no_verify
     21.05            -3.8       17.29 ±  7%  perf-profile.calltrace.cycles-pp.cpu_startup_entry.start_secondary.secondary_startup_64_no_verify
     21.05            -3.8       17.29 ±  7%  perf-profile.calltrace.cycles-pp.start_secondary.secondary_startup_64_no_verify
     21.00            -3.8       17.25 ±  7%  perf-profile.calltrace.cycles-pp.cpuidle_idle_call.do_idle.cpu_startup_entry.start_secondary.secondary_startup_64_no_verify
     20.70            -3.7       17.00 ±  7%  perf-profile.calltrace.cycles-pp.cpuidle_enter.cpuidle_idle_call.do_idle.cpu_startup_entry.start_secondary
     20.69            -3.7       16.99 ±  7%  perf-profile.calltrace.cycles-pp.cpuidle_enter_state.cpuidle_enter.cpuidle_idle_call.do_idle.cpu_startup_entry
     20.64            -3.7       16.95 ±  7%  perf-profile.calltrace.cycles-pp.acpi_idle_enter.cpuidle_enter_state.cpuidle_enter.cpuidle_idle_call.do_idle
      9.51 ±  3%      -1.9        7.57 ±  2%  perf-profile.calltrace.cycles-pp.do_rw_once
      4.54            -1.4        3.19        perf-profile.calltrace.cycles-pp.filemap_map_pages.do_read_fault.do_fault.__handle_mm_fault.handle_mm_fault
      2.83            -0.9        1.96        perf-profile.calltrace.cycles-pp.next_uptodate_folio.filemap_map_pages.do_read_fault.do_fault.__handle_mm_fault
      3.90            -0.6        3.34 ±  5%  perf-profile.calltrace.cycles-pp.clear_page_erms.shmem_get_folio_gfp.shmem_fault.__do_fault.do_read_fault
      4.44 ±  6%      -0.5        3.98 ±  3%  perf-profile.calltrace.cycles-pp.get_mem_cgroup_from_mm.__mem_cgroup_charge.shmem_alloc_and_add_folio.shmem_get_folio_gfp.shmem_fault
      1.17 ±  3%      -0.4        0.73 ±  6%  perf-profile.calltrace.cycles-pp.shmem_inode_acct_blocks.shmem_alloc_and_add_folio.shmem_get_folio_gfp.shmem_fault.__do_fault
      1.42 ±  2%      -0.4        0.99 ±  2%  perf-profile.calltrace.cycles-pp.shmem_alloc_folio.shmem_alloc_and_add_folio.shmem_get_folio_gfp.shmem_fault.__do_fault
      1.32 ±  2%      -0.4        0.91        perf-profile.calltrace.cycles-pp.alloc_pages_mpol.shmem_alloc_folio.shmem_alloc_and_add_folio.shmem_get_folio_gfp.shmem_fault
      1.19 ±  2%      -0.4        0.82        perf-profile.calltrace.cycles-pp.__alloc_pages.alloc_pages_mpol.shmem_alloc_folio.shmem_alloc_and_add_folio.shmem_get_folio_gfp
      0.96 ±  2%      -0.3        0.65 ±  2%  perf-profile.calltrace.cycles-pp.get_page_from_freelist.__alloc_pages.alloc_pages_mpol.shmem_alloc_folio.shmem_alloc_and_add_folio
      0.98 ±  2%      -0.3        0.68 ±  4%  perf-profile.calltrace.cycles-pp.sync_regs.asm_exc_page_fault.do_access
      1.21            +0.5        1.69 ±  5%  perf-profile.calltrace.cycles-pp.__munmap
      1.21            +0.5        1.69 ±  5%  perf-profile.calltrace.cycles-pp.__vm_munmap.__x64_sys_munmap.do_syscall_64.entry_SYSCALL_64_after_hwframe.__munmap
      1.21            +0.5        1.69 ±  5%  perf-profile.calltrace.cycles-pp.__x64_sys_munmap.do_syscall_64.entry_SYSCALL_64_after_hwframe.__munmap
      1.21            +0.5        1.69 ±  5%  perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.__munmap
      1.21            +0.5        1.69 ±  5%  perf-profile.calltrace.cycles-pp.do_vmi_align_munmap.do_vmi_munmap.__vm_munmap.__x64_sys_munmap.do_syscall_64
      1.21            +0.5        1.69 ±  5%  perf-profile.calltrace.cycles-pp.do_vmi_munmap.__vm_munmap.__x64_sys_munmap.do_syscall_64.entry_SYSCALL_64_after_hwframe
      1.21            +0.5        1.69 ±  5%  perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.__munmap
      1.20            +0.5        1.68 ±  5%  perf-profile.calltrace.cycles-pp.unmap_page_range.unmap_vmas.unmap_region.do_vmi_align_munmap.do_vmi_munmap
      1.20            +0.5        1.68 ±  5%  perf-profile.calltrace.cycles-pp.unmap_vmas.unmap_region.do_vmi_align_munmap.do_vmi_munmap.__vm_munmap
      1.20            +0.5        1.68 ±  5%  perf-profile.calltrace.cycles-pp.zap_pmd_range.unmap_page_range.unmap_vmas.unmap_region.do_vmi_align_munmap
      1.20            +0.5        1.69 ±  5%  perf-profile.calltrace.cycles-pp.unmap_region.do_vmi_align_munmap.do_vmi_munmap.__vm_munmap.__x64_sys_munmap
      1.18            +0.5        1.67 ±  6%  perf-profile.calltrace.cycles-pp.zap_pte_range.zap_pmd_range.unmap_page_range.unmap_vmas.unmap_region
      0.84 ±  2%      +0.6        1.43 ±  5%  perf-profile.calltrace.cycles-pp.lru_add_fn.folio_batch_move_lru.folio_add_lru.shmem_alloc_and_add_folio.shmem_get_folio_gfp
      0.58 ±  3%      +0.6        1.18 ±  5%  perf-profile.calltrace.cycles-pp.page_remove_rmap.zap_pte_range.zap_pmd_range.unmap_page_range.unmap_vmas
      0.00            +0.8        0.79 ±  4%  perf-profile.calltrace.cycles-pp.__mod_memcg_lruvec_state.__mod_lruvec_page_state.page_remove_rmap.zap_pte_range.zap_pmd_range
      0.00            +1.0        1.02 ±  5%  perf-profile.calltrace.cycles-pp.__mod_memcg_lruvec_state.lru_add_fn.folio_batch_move_lru.folio_add_lru.shmem_alloc_and_add_folio
      0.00            +1.1        1.08 ±  4%  perf-profile.calltrace.cycles-pp.__mod_lruvec_page_state.page_remove_rmap.zap_pte_range.zap_pmd_range.unmap_page_range
      0.00            +1.5        1.46 ±  5%  perf-profile.calltrace.cycles-pp.__count_memcg_events.handle_mm_fault.do_user_addr_fault.exc_page_fault.asm_exc_page_fault
      3.29 ±  3%      +1.9        5.19        perf-profile.calltrace.cycles-pp.finish_fault.do_read_fault.do_fault.__handle_mm_fault.handle_mm_fault
      3.02 ±  4%      +2.0        5.00        perf-profile.calltrace.cycles-pp.set_pte_range.finish_fault.do_read_fault.do_fault.__handle_mm_fault
      2.84 ±  4%      +2.0        4.86        perf-profile.calltrace.cycles-pp.folio_add_file_rmap_range.set_pte_range.finish_fault.do_read_fault.do_fault
      2.73 ±  4%      +2.0        4.77        perf-profile.calltrace.cycles-pp.__mod_lruvec_page_state.folio_add_file_rmap_range.set_pte_range.finish_fault.do_read_fault
      1.48 ±  4%      +2.1        3.56 ±  2%  perf-profile.calltrace.cycles-pp.__mod_memcg_lruvec_state.__mod_lruvec_page_state.folio_add_file_rmap_range.set_pte_range.finish_fault
      0.57 ±  4%      +2.8        3.35 ±  2%  perf-profile.calltrace.cycles-pp.__count_memcg_events.mem_cgroup_commit_charge.__mem_cgroup_charge.shmem_alloc_and_add_folio.shmem_get_folio_gfp
      1.96 ±  5%      +2.9        4.86 ±  2%  perf-profile.calltrace.cycles-pp.mem_cgroup_commit_charge.__mem_cgroup_charge.shmem_alloc_and_add_folio.shmem_get_folio_gfp.shmem_fault
      3.65 ±  2%      +3.1        6.77 ±  2%  perf-profile.calltrace.cycles-pp.shmem_add_to_page_cache.shmem_alloc_and_add_folio.shmem_get_folio_gfp.shmem_fault.__do_fault
      0.80 ±  4%      +3.1        3.92 ±  3%  perf-profile.calltrace.cycles-pp.__mod_memcg_lruvec_state.__mod_lruvec_page_state.shmem_add_to_page_cache.shmem_alloc_and_add_folio.shmem_get_folio_gfp
      2.68 ±  3%      +3.4        6.08 ±  2%  perf-profile.calltrace.cycles-pp.__mod_lruvec_page_state.shmem_add_to_page_cache.shmem_alloc_and_add_folio.shmem_get_folio_gfp.shmem_fault
      7.71 ±  6%      +3.9       11.66 ±  2%  perf-profile.calltrace.cycles-pp.__mem_cgroup_charge.shmem_alloc_and_add_folio.shmem_get_folio_gfp.shmem_fault.__do_fault
     67.18            +6.3       73.46 ±  3%  perf-profile.calltrace.cycles-pp.do_access
      1.46 ±  9%      +7.1        8.57 ± 16%  perf-profile.calltrace.cycles-pp._raw_spin_lock_irqsave.folio_lruvec_lock_irqsave.folio_batch_move_lru.folio_add_lru.shmem_alloc_and_add_folio
      1.50 ±  9%      +7.1        8.61 ± 16%  perf-profile.calltrace.cycles-pp.folio_lruvec_lock_irqsave.folio_batch_move_lru.folio_add_lru.shmem_alloc_and_add_folio.shmem_get_folio_gfp
      1.38 ± 10%      +7.1        8.51 ± 16%  perf-profile.calltrace.cycles-pp.native_queued_spin_lock_slowpath._raw_spin_lock_irqsave.folio_lruvec_lock_irqsave.folio_batch_move_lru.folio_add_lru
     51.46            +7.6       59.08 ±  3%  perf-profile.calltrace.cycles-pp.asm_exc_page_fault.do_access
      2.98 ±  5%      +7.7       10.66 ± 14%  perf-profile.calltrace.cycles-pp.folio_add_lru.shmem_alloc_and_add_folio.shmem_get_folio_gfp.shmem_fault.__do_fault
      2.84 ±  6%      +7.7       10.56 ± 14%  perf-profile.calltrace.cycles-pp.folio_batch_move_lru.folio_add_lru.shmem_alloc_and_add_folio.shmem_get_folio_gfp.shmem_fault
     34.18            +8.5       42.68 ±  4%  perf-profile.calltrace.cycles-pp.__do_fault.do_read_fault.do_fault.__handle_mm_fault.handle_mm_fault
     34.14            +8.5       42.64 ±  4%  perf-profile.calltrace.cycles-pp.shmem_fault.__do_fault.do_read_fault.do_fault.__handle_mm_fault
     33.95            +8.6       42.51 ±  4%  perf-profile.calltrace.cycles-pp.shmem_get_folio_gfp.shmem_fault.__do_fault.do_read_fault.do_fault
     42.88            +8.8       51.70 ±  4%  perf-profile.calltrace.cycles-pp.__handle_mm_fault.handle_mm_fault.do_user_addr_fault.exc_page_fault.asm_exc_page_fault
     42.34            +9.0       51.30 ±  4%  perf-profile.calltrace.cycles-pp.do_fault.__handle_mm_fault.handle_mm_fault.do_user_addr_fault.exc_page_fault
     42.29            +9.0       51.28 ±  4%  perf-profile.calltrace.cycles-pp.do_read_fault.do_fault.__handle_mm_fault.handle_mm_fault.do_user_addr_fault
     45.07            +9.6       54.62 ±  4%  perf-profile.calltrace.cycles-pp.exc_page_fault.asm_exc_page_fault.do_access
     44.95            +9.6       54.53 ±  4%  perf-profile.calltrace.cycles-pp.do_user_addr_fault.exc_page_fault.asm_exc_page_fault.do_access
     43.72            +9.9       53.64 ±  4%  perf-profile.calltrace.cycles-pp.handle_mm_fault.do_user_addr_fault.exc_page_fault.asm_exc_page_fault.do_access
     17.28 ±  2%     +13.8       31.05 ±  6%  perf-profile.calltrace.cycles-pp.shmem_alloc_and_add_folio.shmem_get_folio_gfp.shmem_fault.__do_fault.do_read_fault
     21.14            -3.8       17.36 ±  7%  perf-profile.children.cycles-pp.cpu_startup_entry
     21.14            -3.8       17.36 ±  7%  perf-profile.children.cycles-pp.do_idle
     21.14            -3.8       17.36 ±  7%  perf-profile.children.cycles-pp.secondary_startup_64_no_verify
     21.09            -3.8       17.33 ±  7%  perf-profile.children.cycles-pp.cpuidle_idle_call
     21.05            -3.8       17.29 ±  7%  perf-profile.children.cycles-pp.start_secondary
     20.79            -3.7       17.07 ±  7%  perf-profile.children.cycles-pp.cpuidle_enter
     20.78            -3.7       17.07 ±  7%  perf-profile.children.cycles-pp.cpuidle_enter_state
     20.72            -3.7       17.02 ±  7%  perf-profile.children.cycles-pp.acpi_idle_enter
     20.71            -3.7       17.01 ±  7%  perf-profile.children.cycles-pp.acpi_safe_halt
     20.79            -3.6       17.19 ±  6%  perf-profile.children.cycles-pp.asm_sysvec_apic_timer_interrupt
     11.52            -3.1        8.42        perf-profile.children.cycles-pp.do_rw_once
      4.62            -1.4        3.24        perf-profile.children.cycles-pp.filemap_map_pages
      2.89            -0.9        2.00        perf-profile.children.cycles-pp.next_uptodate_folio
      3.98            -0.6        3.39 ±  5%  perf-profile.children.cycles-pp.clear_page_erms
      4.46 ±  6%      -0.5        3.99 ±  3%  perf-profile.children.cycles-pp.get_mem_cgroup_from_mm
      1.18 ±  4%      -0.4        0.74 ±  6%  perf-profile.children.cycles-pp.shmem_inode_acct_blocks
      1.44 ±  2%      -0.4        1.00 ±  2%  perf-profile.children.cycles-pp.shmem_alloc_folio
      1.40            -0.4        0.99        perf-profile.children.cycles-pp.alloc_pages_mpol
      1.27            -0.4        0.90        perf-profile.children.cycles-pp.__alloc_pages
      1.01 ±  2%      -0.3        0.68        perf-profile.children.cycles-pp.get_page_from_freelist
      1.02 ±  2%      -0.3        0.70 ±  4%  perf-profile.children.cycles-pp.sync_regs
      0.77 ±  2%      -0.3        0.51        perf-profile.children.cycles-pp.rmqueue
      0.81 ±  2%      -0.2        0.60        perf-profile.children.cycles-pp.__perf_sw_event
      0.53 ±  3%      -0.2        0.34 ±  2%  perf-profile.children.cycles-pp.__rmqueue_pcplist
      0.68 ±  2%      -0.2        0.50 ±  5%  perf-profile.children.cycles-pp.__mod_lruvec_state
      0.65 ±  6%      -0.2        0.47 ±  2%  perf-profile.children.cycles-pp._raw_spin_lock
      0.47 ±  3%      -0.2        0.29 ±  2%  perf-profile.children.cycles-pp.rmqueue_bulk
      0.65 ±  2%      -0.2        0.49        perf-profile.children.cycles-pp.___perf_sw_event
      0.64 ±  4%      -0.1        0.49 ±  5%  perf-profile.children.cycles-pp.xas_load
      0.54            -0.1        0.39 ±  4%  perf-profile.children.cycles-pp.__mod_node_page_state
      0.49 ±  2%      -0.1        0.35 ±  3%  perf-profile.children.cycles-pp.lock_vma_under_rcu
      0.54 ±  5%      -0.1        0.40 ±  2%  perf-profile.children.cycles-pp.xas_find
      0.39 ±  4%      -0.1        0.28 ±  3%  perf-profile.children.cycles-pp.__pte_offset_map_lock
      0.39 ±  3%      -0.1        0.29 ±  3%  perf-profile.children.cycles-pp.xas_descend
      0.32 ±  4%      -0.1        0.22 ±  8%  perf-profile.children.cycles-pp.__dquot_alloc_space
      0.30 ±  3%      -0.1        0.22 ±  3%  perf-profile.children.cycles-pp.mas_walk
      0.20 ± 13%      -0.1        0.13 ±  5%  perf-profile.children.cycles-pp.shmem_recalc_inode
      0.26 ±  2%      -0.1        0.19 ±  3%  perf-profile.children.cycles-pp.filemap_get_entry
      0.18 ±  5%      -0.1        0.12 ±  5%  perf-profile.children.cycles-pp.xas_find_conflict
      0.28 ±  4%      -0.1        0.22 ±  8%  perf-profile.children.cycles-pp.__x64_sys_execve
      0.28 ±  4%      -0.1        0.22 ±  8%  perf-profile.children.cycles-pp.do_execveat_common
      0.28 ±  4%      -0.1        0.22 ±  8%  perf-profile.children.cycles-pp.execve
      0.29 ±  3%      -0.1        0.24 ±  8%  perf-profile.children.cycles-pp.asm_sysvec_call_function_single
      0.16 ±  5%      -0.1        0.11 ±  8%  perf-profile.children.cycles-pp.error_entry
      0.14 ±  5%      -0.0        0.09 ±  8%  perf-profile.children.cycles-pp.__percpu_counter_limited_add
      0.15 ±  5%      -0.0        0.10 ± 10%  perf-profile.children.cycles-pp.inode_add_bytes
      0.07 ±  6%      -0.0        0.02 ± 99%  perf-profile.children.cycles-pp.__folio_throttle_swaprate
      0.10            -0.0        0.06 ± 13%  perf-profile.children.cycles-pp.security_vm_enough_memory_mm
      0.18 ±  7%      -0.0        0.14 ± 13%  perf-profile.children.cycles-pp._raw_spin_lock_irq
      0.16 ±  5%      -0.0        0.12        perf-profile.children.cycles-pp.handle_pte_fault
      0.17 ±  7%      -0.0        0.12 ±  4%  perf-profile.children.cycles-pp.xas_start
      0.14 ±  6%      -0.0        0.10 ±  3%  perf-profile.children.cycles-pp.__pte_offset_map
      0.07 ±  5%      -0.0        0.03 ± 70%  perf-profile.children.cycles-pp.policy_nodemask
      0.16 ±  4%      -0.0        0.13 ± 12%  perf-profile.children.cycles-pp.folio_mark_accessed
      0.19 ±  4%      -0.0        0.16 ±  8%  perf-profile.children.cycles-pp.bprm_execve
      0.11 ±  9%      -0.0        0.08 ±  6%  perf-profile.children.cycles-pp.down_read_trylock
      0.16 ±  6%      -0.0        0.13 ±  5%  perf-profile.children.cycles-pp.percpu_counter_add_batch
      0.11 ±  6%      -0.0        0.08 ±  6%  perf-profile.children.cycles-pp.up_read
      0.15 ±  7%      -0.0        0.12 ± 13%  perf-profile.children.cycles-pp.folio_unlock
      0.10 ±  4%      -0.0        0.07 ±  6%  perf-profile.children.cycles-pp.__libc_fork
      0.07 ±  6%      -0.0        0.04 ± 45%  perf-profile.children.cycles-pp.ksys_read
      0.10 ±  3%      -0.0        0.07 ± 11%  perf-profile.children.cycles-pp.kernel_clone
      0.09 ±  5%      -0.0        0.06 ±  7%  perf-profile.children.cycles-pp.mem_cgroup_update_lru_size
      0.09 ±  5%      -0.0        0.06 ± 11%  perf-profile.children.cycles-pp.__x64_sys_openat
      0.08 ±  8%      -0.0        0.06 ± 11%  perf-profile.children.cycles-pp.do_filp_open
      0.08 ±  8%      -0.0        0.06 ± 11%  perf-profile.children.cycles-pp.path_openat
      0.07            -0.0        0.04 ± 45%  perf-profile.children.cycles-pp.vfs_read
      0.09 ±  4%      -0.0        0.06 ±  7%  perf-profile.children.cycles-pp.__do_sys_clone
      0.10 ±  6%      -0.0        0.08 ±  6%  perf-profile.children.cycles-pp.pte_offset_map_nolock
      0.08 ±  8%      -0.0        0.06 ± 11%  perf-profile.children.cycles-pp.do_sys_openat2
      0.07 ±  5%      -0.0        0.04 ± 45%  perf-profile.children.cycles-pp.copy_process
      0.16 ±  5%      -0.0        0.14 ±  6%  perf-profile.children.cycles-pp.exec_binprm
      0.10 ±  6%      -0.0        0.08 ±  7%  perf-profile.children.cycles-pp.__vm_enough_memory
      0.16 ±  4%      -0.0        0.14 ±  6%  perf-profile.children.cycles-pp.search_binary_handler
      0.08            -0.0        0.06 ±  9%  perf-profile.children.cycles-pp.__irqentry_text_end
      0.09 ±  5%      -0.0        0.07 ±  7%  perf-profile.children.cycles-pp._compound_head
      0.15 ±  5%      -0.0        0.13 ±  7%  perf-profile.children.cycles-pp.xas_create
      0.15 ±  4%      -0.0        0.14 ±  8%  perf-profile.children.cycles-pp.load_elf_binary
      0.12 ±  4%      -0.0        0.10 ±  3%  perf-profile.children.cycles-pp.kmem_cache_alloc_lru
      0.05 ±  8%      +0.0        0.08 ±  8%  perf-profile.children.cycles-pp.propagate_protected_usage
      0.25 ±  2%      +0.0        0.30 ±  4%  perf-profile.children.cycles-pp.page_counter_try_charge
      0.02 ±141%      +0.0        0.06 ±  7%  perf-profile.children.cycles-pp.mod_objcg_state
      0.00            +0.1        0.07 ± 14%  perf-profile.children.cycles-pp.tlb_finish_mmu
      1.25            +0.5        1.72 ±  5%  perf-profile.children.cycles-pp.unmap_vmas
      1.24            +0.5        1.71 ±  5%  perf-profile.children.cycles-pp.zap_pte_range
      1.24            +0.5        1.71 ±  5%  perf-profile.children.cycles-pp.unmap_page_range
      1.24            +0.5        1.71 ±  5%  perf-profile.children.cycles-pp.zap_pmd_range
      1.21            +0.5        1.69 ±  5%  perf-profile.children.cycles-pp.__munmap
      1.22            +0.5        1.71 ±  5%  perf-profile.children.cycles-pp.__vm_munmap
      1.21            +0.5        1.70 ±  5%  perf-profile.children.cycles-pp.__x64_sys_munmap
      1.25            +0.5        1.74 ±  5%  perf-profile.children.cycles-pp.do_vmi_align_munmap
      1.25            +0.5        1.74 ±  5%  perf-profile.children.cycles-pp.do_vmi_munmap
      1.22            +0.5        1.72 ±  5%  perf-profile.children.cycles-pp.unmap_region
      0.85 ±  2%      +0.6        1.44 ±  5%  perf-profile.children.cycles-pp.lru_add_fn
      0.60 ±  3%      +0.6        1.20 ±  4%  perf-profile.children.cycles-pp.page_remove_rmap
      3.30 ±  3%      +1.9        5.20        perf-profile.children.cycles-pp.finish_fault
      3.04 ±  4%      +2.0        5.01        perf-profile.children.cycles-pp.set_pte_range
      2.85 ±  4%      +2.0        4.87        perf-profile.children.cycles-pp.folio_add_file_rmap_range
      1.97 ±  5%      +2.9        4.88 ±  2%  perf-profile.children.cycles-pp.mem_cgroup_commit_charge
      3.69 ±  2%      +3.1        6.80 ±  2%  perf-profile.children.cycles-pp.shmem_add_to_page_cache
      7.74 ±  6%      +3.9       11.69 ±  2%  perf-profile.children.cycles-pp.__mem_cgroup_charge
      0.80 ±  4%      +4.0        4.85 ±  3%  perf-profile.children.cycles-pp.__count_memcg_events
      6.12 ±  3%      +6.1       12.18        perf-profile.children.cycles-pp.__mod_lruvec_page_state
      2.99 ±  3%      +6.6        9.56 ±  2%  perf-profile.children.cycles-pp.__mod_memcg_lruvec_state
     61.44            +6.7       68.11 ±  3%  perf-profile.children.cycles-pp.do_access
      1.58 ±  9%      +7.1        8.72 ± 16%  perf-profile.children.cycles-pp._raw_spin_lock_irqsave
      1.45 ±  9%      +7.2        8.63 ± 16%  perf-profile.children.cycles-pp.native_queued_spin_lock_slowpath
      1.53 ±  9%      +7.2        8.72 ± 16%  perf-profile.children.cycles-pp.folio_lruvec_lock_irqsave
      2.98 ±  5%      +7.7       10.67 ± 14%  perf-profile.children.cycles-pp.folio_add_lru
      2.86 ±  6%      +7.8       10.63 ± 14%  perf-profile.children.cycles-pp.folio_batch_move_lru
     49.12            +8.3       57.47 ±  3%  perf-profile.children.cycles-pp.asm_exc_page_fault
     34.19            +8.5       42.68 ±  4%  perf-profile.children.cycles-pp.__do_fault
     34.15            +8.5       42.65 ±  4%  perf-profile.children.cycles-pp.shmem_fault
     33.99            +8.6       42.54 ±  4%  perf-profile.children.cycles-pp.shmem_get_folio_gfp
     43.06            +8.8       51.84 ±  4%  perf-profile.children.cycles-pp.__handle_mm_fault
     42.43            +8.9       51.37 ±  4%  perf-profile.children.cycles-pp.do_fault
     42.38            +9.0       51.34 ±  4%  perf-profile.children.cycles-pp.do_read_fault
     45.26            +9.5       54.78 ±  4%  perf-profile.children.cycles-pp.exc_page_fault
     45.15            +9.5       54.69 ±  4%  perf-profile.children.cycles-pp.do_user_addr_fault
     43.91            +9.9       53.80 ±  4%  perf-profile.children.cycles-pp.handle_mm_fault
     17.31 ±  2%     +13.8       31.07 ±  5%  perf-profile.children.cycles-pp.shmem_alloc_and_add_folio
     12.24            -4.5        7.76 ±  3%  perf-profile.self.cycles-pp.shmem_get_folio_gfp
     17.96            -3.3       14.66 ±  4%  perf-profile.self.cycles-pp.acpi_safe_halt
     10.95            -3.2        7.74        perf-profile.self.cycles-pp.do_rw_once
      5.96            -1.4        4.58 ±  2%  perf-profile.self.cycles-pp.do_access
      2.40            -0.8        1.64        perf-profile.self.cycles-pp.next_uptodate_folio
      3.92            -0.6        3.36 ±  5%  perf-profile.self.cycles-pp.clear_page_erms
      4.40 ±  6%      -0.5        3.95 ±  3%  perf-profile.self.cycles-pp.get_mem_cgroup_from_mm
      1.52 ±  2%      -0.4        1.10 ±  2%  perf-profile.self.cycles-pp.filemap_map_pages
      1.02 ±  2%      -0.3        0.70 ±  4%  perf-profile.self.cycles-pp.sync_regs
      0.50 ±  7%      -0.2        0.27 ±  5%  perf-profile.self.cycles-pp.shmem_inode_acct_blocks
      0.63 ±  5%      -0.2        0.46 ±  3%  perf-profile.self.cycles-pp._raw_spin_lock
      0.42 ±  2%      -0.1        0.27 ±  2%  perf-profile.self.cycles-pp.rmqueue_bulk
      0.52            -0.1        0.38 ±  4%  perf-profile.self.cycles-pp.__mod_node_page_state
      0.56 ±  2%      -0.1        0.42        perf-profile.self.cycles-pp.___perf_sw_event
      0.31 ±  3%      -0.1        0.20 ±  2%  perf-profile.self.cycles-pp.shmem_add_to_page_cache
      0.38 ±  4%      -0.1        0.28        perf-profile.self.cycles-pp.__handle_mm_fault
      0.36 ±  4%      -0.1        0.26 ±  2%  perf-profile.self.cycles-pp.xas_descend
      0.30 ±  2%      -0.1        0.22 ±  2%  perf-profile.self.cycles-pp.mas_walk
      0.33 ±  3%      -0.1        0.26 ± 10%  perf-profile.self.cycles-pp.lru_add_fn
      0.20 ±  3%      -0.1        0.14 ±  5%  perf-profile.self.cycles-pp.asm_exc_page_fault
      0.21 ±  5%      -0.1        0.15 ±  6%  perf-profile.self.cycles-pp.get_page_from_freelist
      0.26 ±  9%      -0.1        0.20 ± 15%  perf-profile.self.cycles-pp.xas_store
      0.16 ±  7%      -0.1        0.11 ±  6%  perf-profile.self.cycles-pp.__perf_sw_event
      0.18 ±  2%      -0.1        0.13 ±  5%  perf-profile.self.cycles-pp.__alloc_pages
      0.22 ±  4%      -0.1        0.17 ±  4%  perf-profile.self.cycles-pp.handle_mm_fault
      0.20 ±  8%      -0.1        0.14 ±  5%  perf-profile.self.cycles-pp.xas_find
      0.15 ±  6%      -0.0        0.10 ±  7%  perf-profile.self.cycles-pp.error_entry
      0.17 ±  2%      -0.0        0.12 ±  6%  perf-profile.self.cycles-pp.__dquot_alloc_space
      0.17 ±  6%      -0.0        0.13 ± 10%  perf-profile.self.cycles-pp._raw_spin_lock_irq
      0.22 ±  4%      -0.0        0.18 ±  9%  perf-profile.self.cycles-pp.xas_load
      0.23 ±  4%      -0.0        0.19 ± 10%  perf-profile.self.cycles-pp.zap_pte_range
      0.12 ±  7%      -0.0        0.08 ± 10%  perf-profile.self.cycles-pp.__percpu_counter_limited_add
      0.14 ±  3%      -0.0        0.09 ±  7%  perf-profile.self.cycles-pp.rmqueue
      0.15 ±  6%      -0.0        0.10 ±  9%  perf-profile.self.cycles-pp.__mod_lruvec_state
      0.15 ±  2%      -0.0        0.11 ±  6%  perf-profile.self.cycles-pp.do_user_addr_fault
      0.12 ±  7%      -0.0        0.08 ±  5%  perf-profile.self.cycles-pp.folio_add_lru
      0.16 ±  7%      -0.0        0.12 ±  4%  perf-profile.self.cycles-pp.xas_start
      0.06 ±  7%      -0.0        0.02 ± 99%  perf-profile.self.cycles-pp.finish_fault
      0.16 ±  4%      -0.0        0.12 ± 12%  perf-profile.self.cycles-pp.folio_mark_accessed
      0.11 ±  8%      -0.0        0.08 ±  6%  perf-profile.self.cycles-pp.__pte_offset_map_lock
      0.13 ±  6%      -0.0        0.09 ±  4%  perf-profile.self.cycles-pp.__pte_offset_map
      0.11 ±  9%      -0.0        0.08 ±  6%  perf-profile.self.cycles-pp.down_read_trylock
      0.12 ±  3%      -0.0        0.09 ±  5%  perf-profile.self.cycles-pp.do_read_fault
      0.09 ±  5%      -0.0        0.06 ±  7%  perf-profile.self.cycles-pp.mem_cgroup_update_lru_size
      0.16 ±  4%      -0.0        0.12 ±  4%  perf-profile.self.cycles-pp.percpu_counter_add_batch
      0.11 ±  6%      -0.0        0.08 ±  5%  perf-profile.self.cycles-pp.shmem_alloc_and_add_folio
      0.08 ±  8%      -0.0        0.05        perf-profile.self.cycles-pp.xas_find_conflict
      0.12 ±  4%      -0.0        0.09 ±  7%  perf-profile.self.cycles-pp.folio_add_file_rmap_range
      0.10 ±  6%      -0.0        0.07 ±  6%  perf-profile.self.cycles-pp.up_read
      0.12 ±  4%      -0.0        0.10 ±  6%  perf-profile.self.cycles-pp._raw_spin_lock_irqsave
      0.09 ±  4%      -0.0        0.07 ±  7%  perf-profile.self.cycles-pp.exc_page_fault
      0.13 ±  6%      -0.0        0.10 ±  9%  perf-profile.self.cycles-pp.page_remove_rmap
      0.08            -0.0        0.06 ±  8%  perf-profile.self.cycles-pp.__irqentry_text_end
      0.19 ±  5%      -0.0        0.17 ±  5%  perf-profile.self.cycles-pp.cgroup_rstat_updated
      0.09 ±  6%      -0.0        0.07 ±  5%  perf-profile.self.cycles-pp.set_pte_range
      0.07 ±  5%      -0.0        0.05 ±  7%  perf-profile.self.cycles-pp._compound_head
      0.08            -0.0        0.06 ±  9%  perf-profile.self.cycles-pp.lock_vma_under_rcu
      0.05 ±  8%      +0.0        0.08 ±  8%  perf-profile.self.cycles-pp.propagate_protected_usage
      2.93 ±  4%      +0.4        3.35 ±  3%  perf-profile.self.cycles-pp.__mod_lruvec_page_state
      0.77 ±  7%      +1.5        2.23 ±  3%  perf-profile.self.cycles-pp.__mem_cgroup_charge
      0.75 ±  4%      +4.0        4.80 ±  3%  perf-profile.self.cycles-pp.__count_memcg_events
      2.83 ±  3%      +6.6        9.40 ±  2%  perf-profile.self.cycles-pp.__mod_memcg_lruvec_state
      1.45 ±  9%      +7.2        8.63 ± 16%  perf-profile.self.cycles-pp.native_queued_spin_lock_slowpath


***************************************************************************************************
lkp-skl-fpga01: 104 threads 2 sockets (Skylake) with 192G memory
=========================================================================================
compiler/cpufreq_governor/kconfig/mode/nr_task/rootfs/tbox_group/test/testcase:
  gcc-12/performance/x86_64-rhel-8.3/process/50%/debian-11.1-x86_64-20220510.cgz/lkp-skl-fpga01/tlb_flush2/will-it-scale

commit: 
  e0bf1dc859 ("mm: memcg: move vmstats structs definition above flushing code")
  8d59d2214c ("mm: memcg: make stats flushing threshold per-memcg")

e0bf1dc859fdd08e 8d59d2214c2362e7a9d185d80b6 
---------------- --------------------------- 
         %stddev     %change         %stddev
             \          |                \  
      4.05            -1.2        2.81        mpstat.cpu.all.usr%
    193.83 ±  6%     +69.3%     328.17 ±  8%  perf-c2c.DRAM.local
      1216 ±  8%     +27.1%       1546 ±  6%  perf-c2c.DRAM.remote
    150.33 ± 13%     -40.0%      90.17 ± 13%  perf-c2c.HITM.remote
      0.04           -25.0%       0.03        turbostat.IPC
    316.16            -1.5%     311.47        turbostat.PkgWatt
     30.54            +4.9%      32.04        turbostat.RAMWatt
   2132437           -32.3%    1444430        will-it-scale.52.processes
     41008           -32.3%      27776        will-it-scale.per_process_ops
   2132437           -32.3%    1444430        will-it-scale.workload
 3.113e+08 ±  3%     -31.7%  2.125e+08 ±  4%  numa-numastat.node0.local_node
 3.114e+08 ±  3%     -31.7%  2.126e+08 ±  4%  numa-numastat.node0.numa_hit
 3.322e+08 ±  2%     -32.5%  2.243e+08 ±  3%  numa-numastat.node1.local_node
 3.323e+08 ±  2%     -32.5%  2.243e+08 ±  3%  numa-numastat.node1.numa_hit
 3.114e+08 ±  3%     -31.7%  2.126e+08 ±  4%  numa-vmstat.node0.numa_hit
 3.113e+08 ±  3%     -31.7%  2.125e+08 ±  4%  numa-vmstat.node0.numa_local
 3.323e+08 ±  2%     -32.5%  2.243e+08 ±  3%  numa-vmstat.node1.numa_hit
 3.322e+08 ±  2%     -32.5%  2.243e+08 ±  3%  numa-vmstat.node1.numa_local
      0.00 ± 19%     -61.1%       0.00 ± 31%  perf-sched.sch_delay.avg.ms.do_task_dead.do_exit.do_group_exit.__x64_sys_exit_group.do_syscall_64
    217.07 ± 11%     -46.4%     116.39 ± 23%  perf-sched.wait_and_delay.avg.ms.do_task_dead.do_exit.do_group_exit.__x64_sys_exit_group.do_syscall_64
    218.50 ±  6%     +19.1%     260.33 ±  4%  perf-sched.wait_and_delay.count.__cond_resched.zap_page_range_single.madvise_vma_behavior.do_madvise.__x64_sys_madvise
    217.06 ± 11%     -46.4%     116.38 ± 23%  perf-sched.wait_time.avg.ms.do_task_dead.do_exit.do_group_exit.__x64_sys_exit_group.do_syscall_64
 6.436e+08           -32.1%  4.369e+08        proc-vmstat.numa_hit
 6.435e+08           -32.1%  4.368e+08        proc-vmstat.numa_local
 6.432e+08           -32.1%  4.368e+08        proc-vmstat.pgalloc_normal
 1.286e+09           -32.1%  8.726e+08        proc-vmstat.pgfault
 6.432e+08           -32.1%  4.367e+08        proc-vmstat.pgfree
    170696 ±  8%      +3.4%     176515 ±  8%  sched_debug.cpu.clock.avg
    170703 ±  8%      +3.4%     176522 ±  8%  sched_debug.cpu.clock.max
    170689 ±  8%      +3.4%     176508 ±  8%  sched_debug.cpu.clock.min
    169431 ±  8%      +3.4%     175248 ±  8%  sched_debug.cpu.clock_task.avg
    169630 ±  8%      +3.4%     175429 ±  8%  sched_debug.cpu.clock_task.max
    162542 ±  8%      +3.5%     168260 ±  8%  sched_debug.cpu.clock_task.min
    170690 ±  8%      +3.4%     176508 ±  8%  sched_debug.cpu_clk
    170117 ±  8%      +3.4%     175938 ±  8%  sched_debug.ktime
    171259 ±  8%      +3.4%     177078 ±  8%  sched_debug.sched_clk
      4.06           +80.8%       7.34        perf-stat.i.MPKI
 4.066e+09           -23.3%   3.12e+09        perf-stat.i.branch-instructions
      0.57            -0.0        0.55        perf-stat.i.branch-miss-rate%
  23478297           -25.0%   17605102        perf-stat.i.branch-misses
     17.25            +7.0       24.27        perf-stat.i.cache-miss-rate%
  82715093 ±  2%     +35.9%  1.124e+08        perf-stat.i.cache-misses
 4.795e+08 ±  2%      -3.4%   4.63e+08        perf-stat.i.cache-references
      7.14           +32.9%       9.49        perf-stat.i.cpi
    134.85            -1.2%     133.29        perf-stat.i.cpu-migrations
      1760 ±  2%     -26.5%       1294        perf-stat.i.cycles-between-cache-misses
      0.26            -0.0        0.24        perf-stat.i.dTLB-load-miss-rate%
  13461491           -31.7%    9190211        perf-stat.i.dTLB-load-misses
 5.141e+09           -24.1%  3.902e+09        perf-stat.i.dTLB-loads
      0.45            -0.0        0.44        perf-stat.i.dTLB-store-miss-rate%
  12934403           -32.2%    8773143        perf-stat.i.dTLB-store-misses
 2.841e+09           -29.9%  1.992e+09        perf-stat.i.dTLB-stores
     14.76            +1.4       16.18 ±  4%  perf-stat.i.iTLB-load-miss-rate%
   7454399 ±  2%     -22.7%    5760387 ±  4%  perf-stat.i.iTLB-load-misses
  43026423           -30.6%   29840650        perf-stat.i.iTLB-loads
 2.042e+10           -24.7%  1.538e+10        perf-stat.i.instructions
      0.14           -24.6%       0.11        perf-stat.i.ipc
    815.65           -20.2%     651.03        perf-stat.i.metric.K/sec
    120.43           -24.3%      91.11        perf-stat.i.metric.M/sec
   4264808           -32.2%    2892980        perf-stat.i.minor-faults
  11007315 ±  2%     +39.7%   15375516        perf-stat.i.node-load-misses
   1459152 ±  6%     +45.1%    2116827 ±  5%  perf-stat.i.node-loads
   7872989 ±  2%     -26.2%    5812458        perf-stat.i.node-store-misses
   4264808           -32.2%    2892980        perf-stat.i.page-faults
      4.05           +80.4%       7.31        perf-stat.overall.MPKI
      0.58            -0.0        0.57        perf-stat.overall.branch-miss-rate%
     17.25            +7.0       24.27        perf-stat.overall.cache-miss-rate%
      7.13           +32.7%       9.46        perf-stat.overall.cpi
      1759 ±  2%     -26.5%       1294        perf-stat.overall.cycles-between-cache-misses
      0.26            -0.0        0.23        perf-stat.overall.dTLB-load-miss-rate%
      0.45            -0.0        0.44        perf-stat.overall.dTLB-store-miss-rate%
     14.77            +1.4       16.18 ±  4%  perf-stat.overall.iTLB-load-miss-rate%
      0.14           -24.7%       0.11        perf-stat.overall.ipc
   2882666           +11.2%    3206246        perf-stat.overall.path-length
 4.052e+09           -23.3%   3.11e+09        perf-stat.ps.branch-instructions
  23421504           -25.0%   17574476        perf-stat.ps.branch-misses
  82419384 ±  2%     +35.9%   1.12e+08        perf-stat.ps.cache-misses
 4.778e+08 ±  2%      -3.4%  4.614e+08        perf-stat.ps.cache-references
    134.44            -1.1%     132.98        perf-stat.ps.cpu-migrations
  13415064           -31.7%    9160067        perf-stat.ps.dTLB-load-misses
 5.124e+09           -24.1%   3.89e+09        perf-stat.ps.dTLB-loads
  12889609           -32.2%    8744145        perf-stat.ps.dTLB-store-misses
 2.831e+09           -29.9%  1.986e+09        perf-stat.ps.dTLB-stores
   7428050 ±  2%     -22.7%    5741276 ±  4%  perf-stat.ps.iTLB-load-misses
  42877049           -30.6%   29741122        perf-stat.ps.iTLB-loads
 2.035e+10           -24.7%  1.533e+10        perf-stat.ps.instructions
   4250034           -32.2%    2883410        perf-stat.ps.minor-faults
  10968228 ±  2%     +39.7%   15322266        perf-stat.ps.node-load-misses
   1454274 ±  6%     +45.1%    2109746 ±  5%  perf-stat.ps.node-loads
   7845298 ±  2%     -26.2%    5792864        perf-stat.ps.node-store-misses
   4250034           -32.2%    2883410        perf-stat.ps.page-faults
 6.147e+12           -24.7%  4.631e+12        perf-stat.total.instructions
     26.77            -1.8       24.93 ±  3%  perf-profile.calltrace.cycles-pp.intel_idle_ibrs.cpuidle_enter_state.cpuidle_enter.cpuidle_idle_call.do_idle
     26.75            -1.8       24.92 ±  2%  perf-profile.calltrace.cycles-pp.cpuidle_enter.cpuidle_idle_call.do_idle.cpu_startup_entry.start_secondary
     26.75            -1.8       24.92 ±  2%  perf-profile.calltrace.cycles-pp.cpuidle_idle_call.do_idle.cpu_startup_entry.start_secondary.secondary_startup_64_no_verify
     26.84            -1.8       25.00 ±  3%  perf-profile.calltrace.cycles-pp.cpuidle_enter_state.cpuidle_enter.cpuidle_idle_call.do_idle.cpu_startup_entry
     26.75            -1.8       24.92 ±  2%  perf-profile.calltrace.cycles-pp.cpu_startup_entry.start_secondary.secondary_startup_64_no_verify
     26.75            -1.8       24.92 ±  2%  perf-profile.calltrace.cycles-pp.do_idle.cpu_startup_entry.start_secondary.secondary_startup_64_no_verify
     26.75            -1.8       24.92 ±  2%  perf-profile.calltrace.cycles-pp.start_secondary.secondary_startup_64_no_verify
     27.05            -1.8       25.29 ±  3%  perf-profile.calltrace.cycles-pp.secondary_startup_64_no_verify
     13.02 ±  2%      -1.4       11.60 ±  4%  perf-profile.calltrace.cycles-pp.testcase
      5.54 ±  5%      -1.0        4.52 ±  3%  perf-profile.calltrace.cycles-pp.__mem_cgroup_uncharge_list.release_pages.tlb_batch_pages_flush.tlb_finish_mmu.zap_page_range_single
      1.37 ±  2%      -0.9        0.51 ± 58%  perf-profile.calltrace.cycles-pp.asm_exc_page_fault.__madvise
     10.38 ±  3%      -0.8        9.54 ±  2%  perf-profile.calltrace.cycles-pp.asm_exc_page_fault.testcase
      2.38 ±  2%      -0.8        1.63 ±  3%  perf-profile.calltrace.cycles-pp.page_counter_uncharge.uncharge_batch.__mem_cgroup_uncharge_list.release_pages.tlb_batch_pages_flush
      4.02 ±  3%      -0.7        3.32 ±  3%  perf-profile.calltrace.cycles-pp.uncharge_batch.__mem_cgroup_uncharge_list.release_pages.tlb_batch_pages_flush.tlb_finish_mmu
      1.92 ±  4%      -0.4        1.49 ±  2%  perf-profile.calltrace.cycles-pp.irqentry_exit_to_user_mode.asm_exc_page_fault.testcase
      1.36 ±  2%      -0.4        0.99        perf-profile.calltrace.cycles-pp.__irqentry_text_end.testcase
      1.30 ± 10%      -0.4        0.94 ±  6%  perf-profile.calltrace.cycles-pp.get_mem_cgroup_from_mm.__mem_cgroup_charge.do_anonymous_page.__handle_mm_fault.handle_mm_fault
      1.50 ± 11%      -0.3        1.19 ±  5%  perf-profile.calltrace.cycles-pp.uncharge_folio.__mem_cgroup_uncharge_list.release_pages.tlb_batch_pages_flush.tlb_finish_mmu
      1.13 ±  3%      -0.3        0.83        perf-profile.calltrace.cycles-pp.flush_tlb_mm_range.tlb_finish_mmu.zap_page_range_single.madvise_vma_behavior.do_madvise
      0.71 ±  3%      -0.3        0.43 ± 44%  perf-profile.calltrace.cycles-pp.syscall_return_via_sysret.__madvise
      1.02 ±  3%      -0.3        0.75        perf-profile.calltrace.cycles-pp.flush_tlb_func.flush_tlb_mm_range.tlb_finish_mmu.zap_page_range_single.madvise_vma_behavior
      0.97 ±  3%      -0.3        0.72        perf-profile.calltrace.cycles-pp.native_flush_tlb_one_user.flush_tlb_func.flush_tlb_mm_range.tlb_finish_mmu.zap_page_range_single
      0.77 ±  2%      -0.2        0.58 ±  2%  perf-profile.calltrace.cycles-pp.syscall_exit_to_user_mode.do_syscall_64.entry_SYSCALL_64_after_hwframe.__madvise
      0.71 ±  2%      -0.1        0.60 ±  3%  perf-profile.calltrace.cycles-pp.propagate_protected_usage.page_counter_uncharge.uncharge_batch.__mem_cgroup_uncharge_list.release_pages
      1.20            +0.1        1.34        perf-profile.calltrace.cycles-pp.unmap_page_range.zap_page_range_single.madvise_vma_behavior.do_madvise.__x64_sys_madvise
      1.10 ±  2%      +0.2        1.28        perf-profile.calltrace.cycles-pp.zap_pmd_range.unmap_page_range.zap_page_range_single.madvise_vma_behavior.do_madvise
      1.04 ±  2%      +0.2        1.24        perf-profile.calltrace.cycles-pp.zap_pte_range.zap_pmd_range.unmap_page_range.zap_page_range_single.madvise_vma_behavior
      0.83            +0.2        1.07 ±  2%  perf-profile.calltrace.cycles-pp.lru_add_fn.folio_batch_move_lru.lru_add_drain_cpu.lru_add_drain.zap_page_range_single
      0.81 ±  2%      +0.3        1.08        perf-profile.calltrace.cycles-pp.page_remove_rmap.zap_pte_range.zap_pmd_range.unmap_page_range.zap_page_range_single
      0.88 ± 10%      +0.3        1.16 ±  4%  perf-profile.calltrace.cycles-pp.mem_cgroup_commit_charge.__mem_cgroup_charge.do_anonymous_page.__handle_mm_fault.handle_mm_fault
      0.71 ±  2%      +0.3        1.00        perf-profile.calltrace.cycles-pp.__mod_lruvec_page_state.page_remove_rmap.zap_pte_range.zap_pmd_range.unmap_page_range
      0.76 ±  3%      +0.3        1.09 ±  2%  perf-profile.calltrace.cycles-pp.folio_add_new_anon_rmap.do_anonymous_page.__handle_mm_fault.handle_mm_fault.do_user_addr_fault
      0.73 ±  3%      +0.3        1.07 ±  2%  perf-profile.calltrace.cycles-pp.__mod_lruvec_page_state.folio_add_new_anon_rmap.do_anonymous_page.__handle_mm_fault.handle_mm_fault
      0.00            +0.6        0.55 ±  2%  perf-profile.calltrace.cycles-pp.__count_memcg_events.mem_cgroup_commit_charge.__mem_cgroup_charge.do_anonymous_page.__handle_mm_fault
      6.60 ±  4%      +0.6        7.18 ±  3%  perf-profile.calltrace.cycles-pp.exc_page_fault.asm_exc_page_fault.testcase
      6.54 ±  4%      +0.6        7.13 ±  3%  perf-profile.calltrace.cycles-pp.do_user_addr_fault.exc_page_fault.asm_exc_page_fault.testcase
      0.00            +0.7        0.74 ±  3%  perf-profile.calltrace.cycles-pp.__mod_memcg_lruvec_state.release_pages.tlb_batch_pages_flush.tlb_finish_mmu.zap_page_range_single
      0.00            +0.8        0.79 ±  2%  perf-profile.calltrace.cycles-pp.__mod_memcg_lruvec_state.__mod_lruvec_page_state.page_remove_rmap.zap_pte_range.zap_pmd_range
      0.00            +0.8        0.79 ±  3%  perf-profile.calltrace.cycles-pp.__mod_memcg_lruvec_state.lru_add_fn.folio_batch_move_lru.lru_add_drain_cpu.lru_add_drain
      0.00            +0.8        0.80 ±  3%  perf-profile.calltrace.cycles-pp.__mod_memcg_lruvec_state.__mod_lruvec_page_state.folio_add_new_anon_rmap.do_anonymous_page.__handle_mm_fault
      5.80 ±  5%      +0.8        6.60 ±  3%  perf-profile.calltrace.cycles-pp.handle_mm_fault.do_user_addr_fault.exc_page_fault.asm_exc_page_fault.testcase
      0.00            +0.8        0.82        perf-profile.calltrace.cycles-pp.__count_memcg_events.uncharge_batch.__mem_cgroup_uncharge_list.release_pages.tlb_batch_pages_flush
      0.69 ±  4%      +0.9        1.59 ±  2%  perf-profile.calltrace.cycles-pp.__count_memcg_events.handle_mm_fault.do_user_addr_fault.exc_page_fault.asm_exc_page_fault
     30.43            +1.1       31.57        perf-profile.calltrace.cycles-pp.tlb_finish_mmu.zap_page_range_single.madvise_vma_behavior.do_madvise.__x64_sys_madvise
     29.22            +1.5       30.69        perf-profile.calltrace.cycles-pp.tlb_batch_pages_flush.tlb_finish_mmu.zap_page_range_single.madvise_vma_behavior.do_madvise
     29.05            +1.5       30.56        perf-profile.calltrace.cycles-pp.release_pages.tlb_batch_pages_flush.tlb_finish_mmu.zap_page_range_single.madvise_vma_behavior
     22.56 ±  2%      +2.3       24.87        perf-profile.calltrace.cycles-pp.folio_lruvec_lock_irqsave.release_pages.tlb_batch_pages_flush.tlb_finish_mmu.zap_page_range_single
     22.36 ±  2%      +2.3       24.70        perf-profile.calltrace.cycles-pp._raw_spin_lock_irqsave.folio_lruvec_lock_irqsave.release_pages.tlb_batch_pages_flush.tlb_finish_mmu
     22.11 ±  2%      +2.4       24.55        perf-profile.calltrace.cycles-pp.native_queued_spin_lock_slowpath._raw_spin_lock_irqsave.folio_lruvec_lock_irqsave.release_pages.tlb_batch_pages_flush
     22.70            +2.6       25.35        perf-profile.calltrace.cycles-pp.folio_lruvec_lock_irqsave.folio_batch_move_lru.lru_add_drain_cpu.lru_add_drain.zap_page_range_single
     22.38            +2.7       25.08        perf-profile.calltrace.cycles-pp._raw_spin_lock_irqsave.folio_lruvec_lock_irqsave.folio_batch_move_lru.lru_add_drain_cpu.lru_add_drain
     24.10            +2.7       26.82        perf-profile.calltrace.cycles-pp.lru_add_drain.zap_page_range_single.madvise_vma_behavior.do_madvise.__x64_sys_madvise
     24.09            +2.7       26.82        perf-profile.calltrace.cycles-pp.lru_add_drain_cpu.lru_add_drain.zap_page_range_single.madvise_vma_behavior.do_madvise
     24.07            +2.7       26.79        perf-profile.calltrace.cycles-pp.folio_batch_move_lru.lru_add_drain_cpu.lru_add_drain.zap_page_range_single.madvise_vma_behavior
     22.14            +2.8       24.93        perf-profile.calltrace.cycles-pp.native_queued_spin_lock_slowpath._raw_spin_lock_irqsave.folio_lruvec_lock_irqsave.folio_batch_move_lru.lru_add_drain_cpu
     59.76            +2.9       62.64        perf-profile.calltrace.cycles-pp.__madvise
     57.63            +3.5       61.10        perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.__madvise
     57.27            +3.6       60.85        perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.__madvise
     56.41            +3.8       60.20        perf-profile.calltrace.cycles-pp.__x64_sys_madvise.do_syscall_64.entry_SYSCALL_64_after_hwframe.__madvise
     56.37            +3.8       60.17        perf-profile.calltrace.cycles-pp.do_madvise.__x64_sys_madvise.do_syscall_64.entry_SYSCALL_64_after_hwframe.__madvise
     55.94            +3.9       59.88        perf-profile.calltrace.cycles-pp.madvise_vma_behavior.do_madvise.__x64_sys_madvise.do_syscall_64.entry_SYSCALL_64_after_hwframe
     55.85            +4.0       59.82        perf-profile.calltrace.cycles-pp.zap_page_range_single.madvise_vma_behavior.do_madvise.__x64_sys_madvise.do_syscall_64
     26.75            -1.8       24.92 ±  2%  perf-profile.children.cycles-pp.start_secondary
     26.98            -1.8       25.22 ±  3%  perf-profile.children.cycles-pp.intel_idle_ibrs
     27.05            -1.8       25.29 ±  3%  perf-profile.children.cycles-pp.cpu_startup_entry
     27.05            -1.8       25.29 ±  3%  perf-profile.children.cycles-pp.do_idle
     27.05            -1.8       25.29 ±  3%  perf-profile.children.cycles-pp.secondary_startup_64_no_verify
     27.05            -1.8       25.29 ±  3%  perf-profile.children.cycles-pp.cpuidle_enter
     27.05            -1.8       25.29 ±  3%  perf-profile.children.cycles-pp.cpuidle_enter_state
     27.05            -1.8       25.29 ±  3%  perf-profile.children.cycles-pp.cpuidle_idle_call
     13.66 ±  2%      -1.3       12.38        perf-profile.children.cycles-pp.testcase
      5.55 ±  5%      -1.0        4.52 ±  3%  perf-profile.children.cycles-pp.__mem_cgroup_uncharge_list
      2.39 ±  2%      -0.8        1.63 ±  3%  perf-profile.children.cycles-pp.page_counter_uncharge
      4.03 ±  3%      -0.7        3.32 ±  3%  perf-profile.children.cycles-pp.uncharge_batch
      1.96 ±  4%      -0.4        1.52 ±  2%  perf-profile.children.cycles-pp.irqentry_exit_to_user_mode
      1.30            -0.4        0.94 ±  2%  perf-profile.children.cycles-pp.error_entry
      1.36 ±  2%      -0.4        0.99        perf-profile.children.cycles-pp.__irqentry_text_end
      1.30 ± 10%      -0.4        0.94 ±  6%  perf-profile.children.cycles-pp.get_mem_cgroup_from_mm
      1.51 ± 11%      -0.3        1.19 ±  5%  perf-profile.children.cycles-pp.uncharge_folio
      1.14 ±  3%      -0.3        0.84        perf-profile.children.cycles-pp.flush_tlb_mm_range
      1.02 ±  3%      -0.3        0.75        perf-profile.children.cycles-pp.flush_tlb_func
      0.98 ±  3%      -0.3        0.72        perf-profile.children.cycles-pp.native_flush_tlb_one_user
      0.73 ±  2%      -0.2        0.52 ±  2%  perf-profile.children.cycles-pp.syscall_return_via_sysret
      0.69 ±  2%      -0.2        0.50 ±  2%  perf-profile.children.cycles-pp.native_irq_return_iret
      0.79 ±  2%      -0.2        0.60 ±  2%  perf-profile.children.cycles-pp.syscall_exit_to_user_mode
      0.51 ±  2%      -0.1        0.38 ±  2%  perf-profile.children.cycles-pp.sync_regs
      0.41 ±  3%      -0.1        0.29 ±  3%  perf-profile.children.cycles-pp.__perf_sw_event
      0.44 ±  2%      -0.1        0.32 ±  2%  perf-profile.children.cycles-pp.vma_alloc_folio
      0.72 ±  2%      -0.1        0.61 ±  3%  perf-profile.children.cycles-pp.propagate_protected_usage
      0.39            -0.1        0.28 ±  2%  perf-profile.children.cycles-pp.alloc_pages_mpol
      0.35 ±  3%      -0.1        0.25 ±  3%  perf-profile.children.cycles-pp.__alloc_pages
      0.34 ±  2%      -0.1        0.24 ±  4%  perf-profile.children.cycles-pp.___perf_sw_event
      0.30 ±  3%      -0.1        0.21 ±  5%  perf-profile.children.cycles-pp.lock_vma_under_rcu
      0.32 ±  2%      -0.1        0.24        perf-profile.children.cycles-pp.entry_SYSCALL_64
      0.12 ±  4%      -0.1        0.03 ± 70%  perf-profile.children.cycles-pp.down_read
      0.25 ±  3%      -0.1        0.18 ±  4%  perf-profile.children.cycles-pp.mas_walk
      0.25 ±  3%      -0.1        0.18 ±  2%  perf-profile.children.cycles-pp.get_page_from_freelist
      0.17 ±  4%      -0.1        0.11 ±  3%  perf-profile.children.cycles-pp.__pte_offset_map_lock
      0.14 ±  3%      -0.0        0.10 ±  3%  perf-profile.children.cycles-pp.clear_page_erms
      0.17 ±  2%      -0.0        0.12 ±  3%  perf-profile.children.cycles-pp.find_vma_prev
      0.13 ±  2%      -0.0        0.09        perf-profile.children.cycles-pp.percpu_counter_add_batch
      0.11 ±  4%      -0.0        0.07 ± 10%  perf-profile.children.cycles-pp.__cond_resched
      0.13 ±  2%      -0.0        0.10 ±  7%  perf-profile.children.cycles-pp.free_pages_and_swap_cache
      0.06 ±  7%      -0.0        0.03 ± 70%  perf-profile.children.cycles-pp.unmap_vmas
      0.11 ±  3%      -0.0        0.08 ±  6%  perf-profile.children.cycles-pp.free_unref_page_list
      0.06            -0.0        0.03 ± 70%  perf-profile.children.cycles-pp.exit_to_user_mode_prepare
      0.09 ±  7%      -0.0        0.06 ±  6%  perf-profile.children.cycles-pp.free_swap_cache
      0.09 ±  7%      -0.0        0.07 ±  7%  perf-profile.children.cycles-pp.__munmap
      0.09 ±  8%      -0.0        0.06 ±  6%  perf-profile.children.cycles-pp._raw_spin_lock
      0.09 ±  5%      -0.0        0.06 ±  6%  perf-profile.children.cycles-pp.handle_pte_fault
      0.08 ±  8%      -0.0        0.06 ±  6%  perf-profile.children.cycles-pp.do_vmi_munmap
      0.08 ±  4%      -0.0        0.06 ±  6%  perf-profile.children.cycles-pp.__mod_lruvec_state
      0.07 ±  6%      -0.0        0.05 ±  8%  perf-profile.children.cycles-pp.rmqueue
      0.07 ±  9%      -0.0        0.05 ±  7%  perf-profile.children.cycles-pp.unmap_region
      0.08 ±  8%      -0.0        0.06 ±  6%  perf-profile.children.cycles-pp.__vm_munmap
      0.08 ±  8%      -0.0        0.06 ±  6%  perf-profile.children.cycles-pp.__x64_sys_munmap
      0.08 ±  8%      -0.0        0.06 ±  6%  perf-profile.children.cycles-pp.do_vmi_align_munmap
      0.08 ±  5%      -0.0        0.07 ±  7%  perf-profile.children.cycles-pp.try_charge_memcg
      1.27            +0.1        1.40        perf-profile.children.cycles-pp.unmap_page_range
      1.17            +0.2        1.32        perf-profile.children.cycles-pp.zap_pmd_range
      1.12            +0.2        1.29        perf-profile.children.cycles-pp.zap_pte_range
      0.84            +0.2        1.07 ±  2%  perf-profile.children.cycles-pp.lru_add_fn
      0.81 ±  2%      +0.3        1.08        perf-profile.children.cycles-pp.page_remove_rmap
      0.89 ± 10%      +0.3        1.16 ±  4%  perf-profile.children.cycles-pp.mem_cgroup_commit_charge
      0.77 ±  3%      +0.3        1.09 ±  2%  perf-profile.children.cycles-pp.folio_add_new_anon_rmap
      6.62 ±  4%      +0.6        7.19 ±  3%  perf-profile.children.cycles-pp.exc_page_fault
      6.56 ±  4%      +0.6        7.14 ±  3%  perf-profile.children.cycles-pp.do_user_addr_fault
      1.44 ±  2%      +0.6        2.08 ±  2%  perf-profile.children.cycles-pp.__mod_lruvec_page_state
      5.80 ±  5%      +0.8        6.61 ±  3%  perf-profile.children.cycles-pp.handle_mm_fault
     30.44            +1.1       31.58        perf-profile.children.cycles-pp.tlb_finish_mmu
     29.23            +1.5       30.69        perf-profile.children.cycles-pp.tlb_batch_pages_flush
     29.19            +1.5       30.66        perf-profile.children.cycles-pp.release_pages
      1.63 ±  5%      +1.5        3.13 ±  2%  perf-profile.children.cycles-pp.__mod_memcg_lruvec_state
      1.32 ±  4%      +1.6        2.97 ±  2%  perf-profile.children.cycles-pp.__count_memcg_events
     24.12            +2.7       26.84        perf-profile.children.cycles-pp.lru_add_drain
     24.12            +2.7       26.84        perf-profile.children.cycles-pp.lru_add_drain_cpu
     24.09            +2.7       26.81        perf-profile.children.cycles-pp.folio_batch_move_lru
     59.80            +2.9       62.68        perf-profile.children.cycles-pp.__madvise
     57.82            +3.4       61.26        perf-profile.children.cycles-pp.entry_SYSCALL_64_after_hwframe
     57.44            +3.5       60.99        perf-profile.children.cycles-pp.do_syscall_64
     56.41            +3.8       60.20        perf-profile.children.cycles-pp.__x64_sys_madvise
     56.37            +3.8       60.17        perf-profile.children.cycles-pp.do_madvise
     55.94            +3.9       59.88        perf-profile.children.cycles-pp.madvise_vma_behavior
     55.85            +4.0       59.82        perf-profile.children.cycles-pp.zap_page_range_single
     45.26            +5.0       50.23        perf-profile.children.cycles-pp.folio_lruvec_lock_irqsave
     44.75            +5.0       49.80        perf-profile.children.cycles-pp._raw_spin_lock_irqsave
     44.26            +5.2       49.50        perf-profile.children.cycles-pp.native_queued_spin_lock_slowpath
     26.98            -1.8       25.22 ±  3%  perf-profile.self.cycles-pp.intel_idle_ibrs
      1.67 ±  3%      -0.6        1.02 ±  3%  perf-profile.self.cycles-pp.page_counter_uncharge
      1.92 ±  5%      -0.4        1.49 ±  2%  perf-profile.self.cycles-pp.irqentry_exit_to_user_mode
      1.47 ±  2%      -0.4        1.06 ±  2%  perf-profile.self.cycles-pp.testcase
      1.36 ±  2%      -0.4        0.99        perf-profile.self.cycles-pp.__irqentry_text_end
      1.30            -0.4        0.94        perf-profile.self.cycles-pp.error_entry
      1.30 ± 10%      -0.4        0.94 ±  6%  perf-profile.self.cycles-pp.get_mem_cgroup_from_mm
      1.18 ±  8%      -0.3        0.86 ±  6%  perf-profile.self.cycles-pp.uncharge_batch
      1.50 ± 11%      -0.3        1.19 ±  5%  perf-profile.self.cycles-pp.uncharge_folio
      0.98 ±  3%      -0.3        0.72        perf-profile.self.cycles-pp.native_flush_tlb_one_user
      0.71 ±  2%      -0.2        0.51 ±  2%  perf-profile.self.cycles-pp.syscall_return_via_sysret
      0.69 ±  2%      -0.2        0.50 ±  2%  perf-profile.self.cycles-pp.native_irq_return_iret
      0.50 ±  4%      -0.2        0.30 ±  5%  perf-profile.self.cycles-pp._raw_spin_lock_irqsave
      0.75 ±  2%      -0.2        0.56 ±  2%  perf-profile.self.cycles-pp.syscall_exit_to_user_mode
      0.51 ±  2%      -0.1        0.38 ±  2%  perf-profile.self.cycles-pp.sync_regs
      0.35 ±  3%      -0.1        0.23 ±  2%  perf-profile.self.cycles-pp.folio_batch_move_lru
      0.36 ±  5%      -0.1        0.24 ±  2%  perf-profile.self.cycles-pp.lru_add_fn
      0.39 ±  2%      -0.1        0.27 ±  2%  perf-profile.self.cycles-pp.entry_SYSCALL_64_after_hwframe
      0.72 ±  2%      -0.1        0.61 ±  3%  perf-profile.self.cycles-pp.propagate_protected_usage
      0.45            -0.1        0.34 ±  2%  perf-profile.self.cycles-pp.release_pages
      0.54 ±  4%      -0.1        0.45 ±  4%  perf-profile.self.cycles-pp.__mod_lruvec_page_state
      0.30 ±  2%      -0.1        0.21 ±  3%  perf-profile.self.cycles-pp.___perf_sw_event
      0.52 ±  5%      -0.1        0.43 ±  5%  perf-profile.self.cycles-pp.folio_lruvec_lock_irqsave
      0.28 ±  3%      -0.1        0.21        perf-profile.self.cycles-pp.entry_SYSCALL_64
      0.25 ±  3%      -0.1        0.18 ±  4%  perf-profile.self.cycles-pp.mas_walk
      0.24 ±  2%      -0.1        0.17 ±  4%  perf-profile.self.cycles-pp.__handle_mm_fault
      0.16 ±  4%      -0.1        0.10 ±  9%  perf-profile.self.cycles-pp.zap_pte_range
      0.14 ±  4%      -0.0        0.10 ±  4%  perf-profile.self.cycles-pp.clear_page_erms
      0.08 ±  6%      -0.0        0.03 ± 70%  perf-profile.self.cycles-pp.__cond_resched
      0.13            -0.0        0.09        perf-profile.self.cycles-pp.percpu_counter_add_batch
      0.14 ±  5%      -0.0        0.11 ±  3%  perf-profile.self.cycles-pp.handle_mm_fault
      0.11 ±  3%      -0.0        0.08 ±  6%  perf-profile.self.cycles-pp.do_user_addr_fault
      0.08 ±  6%      -0.0        0.04 ± 44%  perf-profile.self.cycles-pp.__perf_sw_event
      0.07 ± 10%      -0.0        0.04 ± 44%  perf-profile.self.cycles-pp.tlb_finish_mmu
      0.09 ±  7%      -0.0        0.06 ±  6%  perf-profile.self.cycles-pp.free_swap_cache
      0.08 ±  7%      -0.0        0.05 ±  8%  perf-profile.self.cycles-pp.lock_vma_under_rcu
      0.09 ±  8%      -0.0        0.06 ±  6%  perf-profile.self.cycles-pp._raw_spin_lock
      0.07 ±  7%      -0.0        0.04 ± 44%  perf-profile.self.cycles-pp.asm_exc_page_fault
      0.10 ±  3%      -0.0        0.08 ±  6%  perf-profile.self.cycles-pp.page_remove_rmap
      0.08 ±  6%      -0.0        0.05 ±  8%  perf-profile.self.cycles-pp.flush_tlb_mm_range
      0.08 ±  6%      -0.0        0.06 ±  9%  perf-profile.self.cycles-pp.do_anonymous_page
      0.08 ±  7%      -0.0        0.06 ±  6%  perf-profile.self.cycles-pp.unmap_page_range
      0.08 ±  5%      -0.0        0.06 ±  6%  perf-profile.self.cycles-pp.__alloc_pages
      0.08 ±  6%      -0.0        0.06 ±  8%  perf-profile.self.cycles-pp.do_madvise
      0.07 ± 10%      -0.0        0.05 ±  8%  perf-profile.self.cycles-pp.up_read
      1.58 ±  6%      +1.5        3.09 ±  2%  perf-profile.self.cycles-pp.__mod_memcg_lruvec_state
      1.27 ±  5%      +1.7        2.93 ±  2%  perf-profile.self.cycles-pp.__count_memcg_events
     44.25            +5.2       49.50        perf-profile.self.cycles-pp.native_queued_spin_lock_slowpath



Disclaimer:
Results have been estimated based on internal Intel analysis and are provided
for informational purposes only. Any difference in system hardware or software
design or configuration may affect actual performance.


-- 
0-DAY CI Kernel Test Service
https://github.com/intel/lkp-tests/wiki


Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ