lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [thread-next>] [day] [month] [year] [list]
Message-ID: <202504152235.188dcce9-lkp@intel.com>
Date: Wed, 16 Apr 2025 15:01:21 +0800
From: kernel test robot <oliver.sang@...el.com>
To: David Hildenbrand <david@...hat.com>
CC: <oe-lkp@...ts.linux.dev>, <lkp@...el.com>, <linux-kernel@...r.kernel.org>,
	Andrew Morton <akpm@...ux-foundation.org>, Andy Lutomirks^H^Hski
	<luto@...nel.org>, Borislav Betkov <bp@...en8.de>, Dave Hansen
	<dave.hansen@...ux.intel.com>, Ingo Molnar <mingo@...hat.com>, Jann Horn
	<jannh@...gle.com>, Johannes Weiner <hannes@...xchg.org>, Jonathan Corbet
	<corbet@....net>, "Kirill A. Shutemov" <kirill.shutemov@...ux.intel.com>,
	Lance Yang <ioworker0@...il.com>, Liam Howlett <liam.howlett@...cle.com>,
	Lorenzo Stoakes <lorenzo.stoakes@...cle.com>, Matthew Wilcow
	<willy@...radead.org>, Michal Koutn <mkoutny@...e.com>, Muchun Song
	<muchun.song@...ux.dev>, tejun heo <tj@...nel.org>, Thomas Gleixner
	<tglx@...utronix.de>, Vlastimil Babka <vbabka@...e.cz>, Zefan Li
	<lizefan.x@...edance.com>, <linux-mm@...ck.org>, <oliver.sang@...el.com>
Subject: [linus:master] [mm/rmap]  6af8cb80d3:  vm-scalability.throughput
 7.8% regression



Hello,

kernel test robot noticed a 7.8% regression of vm-scalability.throughput on:


commit: 6af8cb80d3a9a6bbd521d8a7c949b4eafb7dba5d ("mm/rmap: basic MM owner tracking for large folios (!hugetlb)")
https://git.kernel.org/cgit/linux/kernel/git/torvalds/linux.git master


testcase: vm-scalability
config: x86_64-rhel-9.4
compiler: gcc-12
test machine: 256 threads 2 sockets GENUINE INTEL(R) XEON(R) (Sierra Forest) with 128G memory
parameters:

	runtime: 300s
	size: 8T
	test: anon-cow-seq
	cpufreq_governor: performance




If you fix the issue in a separate patch/commit (i.e. not just a new version of
the same patch/commit), kindly add following tags
| Reported-by: kernel test robot <oliver.sang@...el.com>
| Closes: https://lore.kernel.org/oe-lkp/202504152235.188dcce9-lkp@intel.com


Details are as below:
-------------------------------------------------------------------------------------------------->


The kernel config and materials to reproduce are available at:
https://download.01.org/0day-ci/archive/20250415/202504152235.188dcce9-lkp@intel.com

=========================================================================================
compiler/cpufreq_governor/kconfig/rootfs/runtime/size/tbox_group/test/testcase:
  gcc-12/performance/x86_64-rhel-9.4/debian-12-x86_64-20240206.cgz/300s/8T/lkp-srf-2sp1/anon-cow-seq/vm-scalability

commit: 
  448854478a ("mm/rmap: use folio_large_nr_pages() in add/remove functions")
  6af8cb80d3 ("mm/rmap: basic MM owner tracking for large folios (!hugetlb)")

448854478ab2f477 6af8cb80d3a9a6bbd521d8a7c94 
---------------- --------------------------- 
         %stddev     %change         %stddev
             \          |                \  
     53400           +12.6%      60113        uptime.idle
 1.457e+10 ±  3%     +45.8%  2.125e+10 ±  3%  cpuidle..time
   9999188 ±  2%     +41.4%   14142800 ±  2%  cpuidle..usage
     19401 ±  7%     -21.8%      15176 ± 15%  perf-c2c.DRAM.local
     42430 ±  4%     +51.6%      64345 ± 12%  perf-c2c.HITM.local
     52676 ±  6%     +44.2%      75955 ± 11%  perf-c2c.HITM.total
 8.894e+08           -15.0%  7.557e+08 ±  7%  numa-numastat.node0.local_node
 8.895e+08           -15.0%  7.559e+08 ±  7%  numa-numastat.node0.numa_hit
 8.879e+08           -18.0%  7.284e+08 ±  8%  numa-numastat.node1.local_node
  8.88e+08           -18.0%  7.285e+08 ±  8%  numa-numastat.node1.numa_hit
     15.74 ±  4%     +54.4%      24.31 ±  4%  vmstat.cpu.id
    214.36           -12.0%     188.68        vmstat.procs.r
      9162            -6.0%       8609 ±  2%  vmstat.system.cs
    359131            -3.1%     347966        vmstat.system.in
     15.07 ±  4%      +8.8       23.85 ±  4%  mpstat.cpu.all.idle%
      0.67            -0.1        0.57        mpstat.cpu.all.irq%
      0.17 ±  7%      -0.0        0.14 ±  5%  mpstat.cpu.all.soft%
     75.25            -7.8       67.47        mpstat.cpu.all.sys%
      4.42            -0.9        3.52        mpstat.cpu.all.usr%
  30105953           -40.3%   17974636 ±  2%  meminfo.Active
  30105953           -40.3%   17974636 ±  2%  meminfo.Active(anon)
  28734383           -40.0%   17245061 ±  2%  meminfo.AnonPages
  88649090           +12.1%   99375037        meminfo.MemAvailable
  89179467           +12.0%   99904712        meminfo.MemFree
  42488096           -25.2%   31762847        meminfo.Memused
  73422718           -16.1%   61622071 ±  5%  meminfo.max_used_kB
  14976533 ±  2%     -40.8%    8859130 ±  2%  numa-meminfo.node0.Active
  14976533 ±  2%     -40.8%    8859130 ±  2%  numa-meminfo.node0.Active(anon)
  14404444 ±  2%     -40.1%    8625820 ±  2%  numa-meminfo.node0.AnonPages
   2961490 ± 40%     -58.4%    1231139 ±129%  numa-meminfo.node0.FilePages
  43362532 ±  2%     +16.4%   50494564 ±  3%  numa-meminfo.node0.MemFree
  22336223 ±  5%     -31.9%   15204191 ± 11%  numa-meminfo.node0.MemUsed
  15156929           -39.9%    9115542 ±  3%  numa-meminfo.node1.Active
  15156929           -39.9%    9115542 ±  3%  numa-meminfo.node1.Active(anon)
  14334430           -39.9%    8615995 ±  3%  numa-meminfo.node1.AnonPages
  31426803 ±  4%      -6.1%   29506154 ±  4%  numa-meminfo.node1.AnonPages.max
  20191977 ±  6%     -18.0%   16564588 ±  8%  numa-meminfo.node1.MemUsed
    138520           -22.3%     107667        vm-scalability.median
      3.14 ±  3%      +6.4        9.57 ±  9%  vm-scalability.median_stddev%
      2.35 ±  8%     +12.7       15.05 ±  7%  vm-scalability.stddev%
  36255698            -7.8%   33421568        vm-scalability.throughput
    589253 ±  2%     -36.1%     376589 ±  5%  vm-scalability.time.involuntary_context_switches
 1.772e+09           -16.5%   1.48e+09        vm-scalability.time.minor_page_faults
     20491           -10.6%      18325        vm-scalability.time.percent_of_cpu_this_job_got
     58690            -9.8%      52967        vm-scalability.time.system_time
      3314           -19.5%       2668        vm-scalability.time.user_time
    423126           -25.5%     315143 ±  2%  vm-scalability.time.voluntary_context_switches
 7.968e+09           -16.5%  6.652e+09        vm-scalability.workload
   3752122 ±  2%     -40.9%    2218075 ±  3%  numa-vmstat.node0.nr_active_anon
   3604567 ±  2%     -40.1%    2157950 ±  2%  numa-vmstat.node0.nr_anon_pages
    740400 ± 40%     -58.4%     307780 ±129%  numa-vmstat.node0.nr_file_pages
  10838075 ±  2%     +16.4%   12619573 ±  3%  numa-vmstat.node0.nr_free_pages
   3751375 ±  2%     -40.9%    2217819 ±  3%  numa-vmstat.node0.nr_zone_active_anon
 8.895e+08           -15.0%  7.559e+08 ±  7%  numa-vmstat.node0.numa_hit
 8.894e+08           -15.0%  7.557e+08 ±  7%  numa-vmstat.node0.numa_local
   3814298           -40.7%    2260493 ±  2%  numa-vmstat.node1.nr_active_anon
   3602611           -40.7%    2135607 ±  2%  numa-vmstat.node1.nr_anon_pages
   3813553           -40.7%    2260182 ±  2%  numa-vmstat.node1.nr_zone_active_anon
  8.88e+08           -18.0%  7.285e+08 ±  8%  numa-vmstat.node1.numa_hit
 8.879e+08           -18.0%  7.284e+08 ±  8%  numa-vmstat.node1.numa_local
  29828859 ± 10%     -25.9%   22114899 ± 12%  sched_debug.cfs_rq:/.avg_vruntime.avg
  31729867 ± 10%     -23.1%   24394442 ± 12%  sched_debug.cfs_rq:/.avg_vruntime.max
  27439158 ± 11%     -30.2%   19146267 ± 13%  sched_debug.cfs_rq:/.avg_vruntime.min
   1805955 ± 23%     -31.8%    1231121 ± 22%  sched_debug.cfs_rq:/.left_deadline.stddev
   1805948 ± 23%     -31.8%    1231114 ± 22%  sched_debug.cfs_rq:/.left_vruntime.stddev
  29828862 ± 10%     -25.9%   22114901 ± 12%  sched_debug.cfs_rq:/.min_vruntime.avg
  31729867 ± 10%     -23.1%   24394442 ± 12%  sched_debug.cfs_rq:/.min_vruntime.max
  27439158 ± 11%     -30.2%   19146267 ± 13%  sched_debug.cfs_rq:/.min_vruntime.min
   1805948 ± 23%     -31.8%    1231116 ± 22%  sched_debug.cfs_rq:/.right_vruntime.stddev
    468.50 ±  5%     -25.1%     351.00 ± 24%  sched_debug.cfs_rq:/.util_est.avg
      1538 ±  7%     -27.1%       1121 ±  7%  sched_debug.cfs_rq:/.util_est.max
    291424 ±  4%      -4.1%     279460 ±  4%  sched_debug.cpu.clock_task.min
      1210 ±  2%     +12.2%       1357 ±  3%  sched_debug.cpu.clock_task.stddev
     24329 ±  7%     -13.3%      21085 ±  6%  sched_debug.cpu.curr->pid.max
      5671 ±  7%     -11.3%       5028 ± 10%  sched_debug.cpu.nr_switches.avg
     31944 ±  6%     +25.3%      40025 ± 17%  sched_debug.cpu.nr_switches.max
      3279 ±  8%     -18.3%       2678 ±  8%  sched_debug.cpu.nr_switches.min
      3073 ±  9%     +29.1%       3967 ± 15%  sched_debug.cpu.nr_switches.stddev
   7540874           -40.6%    4480163 ±  2%  proc-vmstat.nr_active_anon
   7192890           -40.2%    4298978 ±  2%  proc-vmstat.nr_anon_pages
   2209141           +12.2%    2479543        proc-vmstat.nr_dirty_background_threshold
   4423684           +12.2%    4965150        proc-vmstat.nr_dirty_threshold
  22285247           +12.1%   24987745        proc-vmstat.nr_free_pages
     45347            -1.1%      44832        proc-vmstat.nr_kernel_stack
     34897            -8.5%      31928        proc-vmstat.nr_page_table_pages
   7540867           -40.6%    4480162 ±  2%  proc-vmstat.nr_zone_active_anon
    144536 ±  6%     +40.8%     203521 ±  9%  proc-vmstat.numa_hint_faults_local
 1.778e+09           -16.5%  1.484e+09        proc-vmstat.numa_hit
   1475274           +57.2%    2319414 ±  2%  proc-vmstat.numa_huge_pte_updates
 1.777e+09           -16.5%  1.484e+09        proc-vmstat.numa_local
    267417            -0.9%     264919        proc-vmstat.numa_other
   1726476 ±  4%    +182.5%    4877430 ±  3%  proc-vmstat.numa_pages_migrated
 7.943e+08           +51.1%    1.2e+09 ±  2%  proc-vmstat.numa_pte_updates
 1.793e+09           -16.2%  1.503e+09        proc-vmstat.pgalloc_normal
 1.774e+09           -16.5%  1.481e+09        proc-vmstat.pgfault
 1.793e+09           -16.2%  1.502e+09        proc-vmstat.pgfree
   7187712           +18.2%    8494165 ±  2%  proc-vmstat.pgmigrate_fail
   1726476 ±  4%    +182.5%    4877430 ±  3%  proc-vmstat.pgmigrate_success
    180904           -10.4%     162040        proc-vmstat.pgreuse
     13401           -16.5%      11190        proc-vmstat.thp_fault_alloc
     14046           +18.2%      16598 ±  2%  proc-vmstat.thp_migration_fail
      3199 ±  4%    +193.0%       9375 ±  4%  proc-vmstat.thp_migration_success
   3430623           -16.5%    2864565        proc-vmstat.thp_split_pmd
      8447           +16.5%       9841        proc-vmstat.unevictable_pgs_culled
      3.45           +12.3%       3.87        perf-stat.i.MPKI
 3.587e+10           -20.3%  2.859e+10        perf-stat.i.branch-instructions
      0.06            +0.0        0.10 ±  6%  perf-stat.i.branch-miss-rate%
  16449735           +19.0%   19582727        perf-stat.i.branch-misses
     54.86            +2.2       57.09        perf-stat.i.cache-miss-rate%
 4.407e+08           -11.2%  3.915e+08        perf-stat.i.cache-misses
 8.022e+08           -14.3%  6.875e+08        perf-stat.i.cache-references
      9138            -4.7%       8709 ±  2%  perf-stat.i.context-switches
      4.31           +45.0%       6.25 ±  6%  perf-stat.i.cpi
  5.52e+11            -9.4%      5e+11        perf-stat.i.cpu-cycles
    803.00            -9.9%     723.58        perf-stat.i.cpu-migrations
      1255           +28.5%       1612 ±  6%  perf-stat.i.cycles-between-cache-misses
 1.278e+11           -20.4%  1.017e+11        perf-stat.i.instructions
      0.24            -8.5%       0.22        perf-stat.i.ipc
     45.69           -16.2%      38.30        perf-stat.i.metric.K/sec
   5871870           -18.2%    4801162        perf-stat.i.minor-faults
   5871868           -18.2%    4801161        perf-stat.i.page-faults
      3.45           +11.8%       3.86        perf-stat.overall.MPKI
      0.04            +0.0        0.07        perf-stat.overall.branch-miss-rate%
     54.96            +2.0       56.96        perf-stat.overall.cache-miss-rate%
      4.32           +12.5%       4.86        perf-stat.overall.cpi
      0.23           -11.1%       0.21        perf-stat.overall.ipc
      4825            -3.0%       4679        perf-stat.overall.path-length
 3.562e+10           -19.1%  2.882e+10        perf-stat.ps.branch-instructions
  15804911           +21.0%   19125930        perf-stat.ps.branch-misses
 4.379e+08            -9.8%   3.95e+08        perf-stat.ps.cache-misses
 7.968e+08           -13.0%  6.935e+08        perf-stat.ps.cache-references
      9012            -5.3%       8538 ±  2%  perf-stat.ps.context-switches
 5.481e+11            -9.2%  4.975e+11        perf-stat.ps.cpu-cycles
    793.13           -10.7%     708.06        perf-stat.ps.cpu-migrations
 1.269e+11           -19.3%  1.024e+11        perf-stat.ps.instructions
   5836046           -16.8%    4855858        perf-stat.ps.minor-faults
   5836046           -16.8%    4855859        perf-stat.ps.page-faults
 3.845e+13           -19.0%  3.113e+13        perf-stat.total.instructions
     42.65 ±218%     -99.7%       0.11 ±191%  perf-sched.sch_delay.avg.ms.__cond_resched.__alloc_frozen_pages_noprof.alloc_pages_mpol.alloc_pages_noprof.__pmd_alloc
      0.48 ±146%     -99.8%       0.00 ±223%  perf-sched.sch_delay.avg.ms.__cond_resched.__kmalloc_cache_noprof.single_open.single_open_size.do_dentry_open
      2.29 ± 24%     -60.6%       0.90 ± 36%  perf-sched.sch_delay.avg.ms.__cond_resched.__split_huge_pmd.do_huge_pmd_wp_page.__handle_mm_fault.handle_mm_fault
      0.25 ± 89%     -93.4%       0.02 ± 59%  perf-sched.sch_delay.avg.ms.__cond_resched.down_write_killable.exec_mmap.begin_new_exec.load_elf_binary
      4.09 ± 79%     -92.4%       0.31 ±177%  perf-sched.sch_delay.avg.ms.__cond_resched.dput.__fput.__x64_sys_close.do_syscall_64
      3.78 ± 97%    -100.0%       0.00        perf-sched.sch_delay.avg.ms.__cond_resched.kmem_cache_alloc_noprof.alloc_pid.copy_process.kernel_clone
      2.93 ± 26%     -64.2%       1.05 ± 24%  perf-sched.sch_delay.avg.ms.__cond_resched.wp_page_copy.__handle_mm_fault.handle_mm_fault.do_user_addr_fault
      0.82 ± 26%     -49.9%       0.41 ± 52%  perf-sched.sch_delay.avg.ms.devkmsg_read.vfs_read.ksys_read.do_syscall_64
      3.27 ± 23%     +49.0%       4.87 ± 14%  perf-sched.sch_delay.avg.ms.irqentry_exit_to_user_mode.asm_exc_page_fault.[unknown]
      3.66 ± 22%     -62.4%       1.38 ± 34%  perf-sched.sch_delay.avg.ms.irqentry_exit_to_user_mode.asm_sysvec_apic_timer_interrupt.[unknown]
      0.07 ± 23%     +89.3%       0.14 ± 62%  perf-sched.sch_delay.avg.ms.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
      0.98 ± 10%     +30.9%       1.28 ± 23%  perf-sched.sch_delay.avg.ms.worker_thread.kthread.ret_from_fork.ret_from_fork_asm
    170.74 ±217%     -99.9%       0.20 ±200%  perf-sched.sch_delay.max.ms.__cond_resched.__alloc_frozen_pages_noprof.alloc_pages_mpol.alloc_pages_noprof.__pmd_alloc
      0.87 ±161%     -99.9%       0.00 ±223%  perf-sched.sch_delay.max.ms.__cond_resched.__kmalloc_cache_noprof.single_open.single_open_size.do_dentry_open
     10.33 ± 20%     -51.8%       4.98 ± 36%  perf-sched.sch_delay.max.ms.__cond_resched.__split_huge_pmd.do_huge_pmd_wp_page.__handle_mm_fault.handle_mm_fault
     11.45 ± 47%     -83.9%       1.84 ±145%  perf-sched.sch_delay.max.ms.__cond_resched.change_pmd_range.isra.0.change_pud_range
      1.13 ± 96%     -98.2%       0.02 ± 54%  perf-sched.sch_delay.max.ms.__cond_resched.down_write_killable.exec_mmap.begin_new_exec.load_elf_binary
     11.78 ± 66%     -95.1%       0.58 ±190%  perf-sched.sch_delay.max.ms.__cond_resched.dput.__fput.__x64_sys_close.do_syscall_64
     12.61 ±121%    -100.0%       0.00        perf-sched.sch_delay.max.ms.__cond_resched.kmem_cache_alloc_noprof.alloc_pid.copy_process.kernel_clone
     73.69 ±137%     -83.0%      12.50 ± 35%  perf-sched.sch_delay.max.ms.pipe_read.vfs_read.ksys_read.do_syscall_64
      1.00 ±137%   +1001.5%      11.02 ±162%  perf-sched.sch_delay.max.ms.schedule_preempt_disabled.rwsem_down_read_slowpath.down_read.folio_lock_anon_vma_read
      0.09 ± 56%    +319.7%       0.37 ± 65%  perf-sched.sch_delay.max.ms.schedule_preempt_disabled.rwsem_down_write_slowpath.down_write.__put_anon_vma
      9.93 ± 55%     -77.0%       2.29 ±104%  perf-sched.sch_delay.max.ms.wait_for_partner.fifo_open.do_dentry_open.vfs_open
     13.75 ± 31%   +1908.6%     276.09 ±203%  perf-sched.sch_delay.max.ms.worker_thread.kthread.ret_from_fork.ret_from_fork_asm
      9.29 ± 13%     +78.0%      16.54 ± 22%  perf-sched.wait_and_delay.avg.ms.__cond_resched.__alloc_frozen_pages_noprof.alloc_pages_mpol.folio_alloc_mpol_noprof.vma_alloc_folio_noprof
    818.50 ± 23%    -100.0%       0.00        perf-sched.wait_and_delay.avg.ms.__cond_resched.kmem_cache_alloc_noprof.alloc_pid.copy_process.kernel_clone
    164.97 ±107%    +241.3%     563.02 ± 31%  perf-sched.wait_and_delay.avg.ms.__x64_sys_pause.do_syscall_64.entry_SYSCALL_64_after_hwframe.[unknown]
      7.60 ±  8%     -16.3%       6.36 ±  5%  perf-sched.wait_and_delay.avg.ms.devkmsg_read.vfs_read.ksys_read.do_syscall_64
      9.12 ±  9%     +41.1%      12.86 ±  7%  perf-sched.wait_and_delay.avg.ms.irqentry_exit_to_user_mode.asm_exc_page_fault.[unknown]
    830.40 ± 18%    -100.0%       0.00        perf-sched.wait_and_delay.avg.ms.pipe_write.vfs_write.ksys_write.do_syscall_64
     15.56 ±  4%     -17.7%      12.81 ±  8%  perf-sched.wait_and_delay.avg.ms.schedule_hrtimeout_range.ep_poll.do_epoll_wait.__x64_sys_epoll_wait
      1608 ± 19%     -81.7%     293.67 ± 15%  perf-sched.wait_and_delay.count.__cond_resched.__alloc_frozen_pages_noprof.alloc_pages_mpol.folio_alloc_mpol_noprof.vma_alloc_folio_noprof
      2.50 ± 68%    -100.0%       0.00        perf-sched.wait_and_delay.count.__cond_resched.kmem_cache_alloc_noprof.alloc_pid.copy_process.kernel_clone
     27.67 ± 70%     -84.3%       4.33 ±134%  perf-sched.wait_and_delay.count.__cond_resched.refresh_cpu_vm_stats.vmstat_update.process_one_work.worker_thread
      3154 ± 20%     -73.3%     842.50 ± 17%  perf-sched.wait_and_delay.count.__cond_resched.wp_page_copy.__handle_mm_fault.handle_mm_fault.do_user_addr_fault
    586.50 ±  5%     +26.4%     741.50 ±  8%  perf-sched.wait_and_delay.count.devkmsg_read.vfs_read.ksys_read.do_syscall_64
    426.17 ± 25%     -88.3%      49.67 ±141%  perf-sched.wait_and_delay.count.irqentry_exit_to_user_mode.asm_sysvec_apic_timer_interrupt.[unknown]
      3.67 ± 20%    -100.0%       0.00        perf-sched.wait_and_delay.count.pipe_write.vfs_write.ksys_write.do_syscall_64
    946.33 ± 12%     +48.7%       1407 ± 17%  perf-sched.wait_and_delay.count.schedule_hrtimeout_range.do_poll.constprop.0.do_sys_poll
    611.17 ±  6%     +23.9%     757.17 ±  7%  perf-sched.wait_and_delay.count.schedule_hrtimeout_range.ep_poll.do_epoll_wait.__x64_sys_epoll_wait
      1027 ±  3%    -100.0%       0.00        perf-sched.wait_and_delay.max.ms.__cond_resched.kmem_cache_alloc_noprof.alloc_pid.copy_process.kernel_clone
      1553 ±  5%    -100.0%       0.00        perf-sched.wait_and_delay.max.ms.pipe_write.vfs_write.ksys_write.do_syscall_64
      5.87 ± 11%    +126.7%      13.30 ± 25%  perf-sched.wait_time.avg.ms.__cond_resched.__alloc_frozen_pages_noprof.alloc_pages_mpol.folio_alloc_mpol_noprof.vma_alloc_folio_noprof
    498.88 ± 57%    -100.0%       0.00 ±223%  perf-sched.wait_time.avg.ms.__cond_resched.__kmalloc_cache_noprof.single_open.single_open_size.do_dentry_open
      3.55 ±169%    +514.3%      21.79 ± 30%  perf-sched.wait_time.avg.ms.__cond_resched.__kmalloc_noprof.inotify_handle_inode_event.send_to_group.fsnotify
      4.09 ± 79%     -92.4%       0.31 ±179%  perf-sched.wait_time.avg.ms.__cond_resched.dput.__fput.__x64_sys_close.do_syscall_64
    814.72 ± 24%    -100.0%       0.00        perf-sched.wait_time.avg.ms.__cond_resched.kmem_cache_alloc_noprof.alloc_pid.copy_process.kernel_clone
      0.93 ± 73%     -86.9%       0.12 ±122%  perf-sched.wait_time.avg.ms.__cond_resched.stop_one_cpu.sched_exec.bprm_execve.part
    165.07 ±106%    +240.8%     562.58 ± 31%  perf-sched.wait_time.avg.ms.__x64_sys_pause.do_syscall_64.entry_SYSCALL_64_after_hwframe.[unknown]
      6.78 ±  6%     -12.2%       5.95 ±  3%  perf-sched.wait_time.avg.ms.devkmsg_read.vfs_read.ksys_read.do_syscall_64
      5.85 ±  5%     +36.7%       8.00 ±  8%  perf-sched.wait_time.avg.ms.irqentry_exit_to_user_mode.asm_exc_page_fault.[unknown]
    829.13 ± 19%     -99.9%       0.91 ±223%  perf-sched.wait_time.avg.ms.pipe_write.vfs_write.ksys_write.do_syscall_64
     15.07 ±  4%     -17.5%      12.43 ±  7%  perf-sched.wait_time.avg.ms.schedule_hrtimeout_range.ep_poll.do_epoll_wait.__x64_sys_epoll_wait
      6.88 ± 33%     -51.4%       3.34 ± 38%  perf-sched.wait_time.avg.ms.schedule_timeout.__wait_for_common.wait_for_completion_state.kernel_clone
     80.29 ± 10%     +31.1%     105.26 ± 20%  perf-sched.wait_time.max.ms.__cond_resched.__alloc_frozen_pages_noprof.alloc_pages_mpol.folio_alloc_mpol_noprof.vma_alloc_folio_noprof
    830.56 ± 44%    -100.0%       0.00 ±223%  perf-sched.wait_time.max.ms.__cond_resched.__kmalloc_cache_noprof.single_open.single_open_size.do_dentry_open
      6.25 ±188%   +2184.2%     142.83 ± 55%  perf-sched.wait_time.max.ms.__cond_resched.__kmalloc_noprof.inotify_handle_inode_event.send_to_group.fsnotify
     11.78 ± 66%     -95.1%       0.58 ±190%  perf-sched.wait_time.max.ms.__cond_resched.dput.__fput.__x64_sys_close.do_syscall_64
      1020 ±  2%    -100.0%       0.00        perf-sched.wait_time.max.ms.__cond_resched.kmem_cache_alloc_noprof.alloc_pid.copy_process.kernel_clone
     14.39 ± 57%     -83.7%       2.34 ±137%  perf-sched.wait_time.max.ms.__cond_resched.stop_one_cpu.sched_exec.bprm_execve.part
     58.23 ± 39%     +83.4%     106.83 ± 31%  perf-sched.wait_time.max.ms.io_schedule.migration_entry_wait_on_locked.migration_entry_wait.do_swap_page
      1553 ±  5%     -99.9%       1.13 ±223%  perf-sched.wait_time.max.ms.pipe_write.vfs_write.ksys_write.do_syscall_64
     34.14 ± 80%    +138.5%      81.42 ± 38%  perf-sched.wait_time.max.ms.schedule_preempt_disabled.rwsem_down_write_slowpath.down_write.unlink_anon_vmas
    175.44 ±210%     -98.7%       2.29 ±104%  perf-sched.wait_time.max.ms.wait_for_partner.fifo_open.do_dentry_open.vfs_open




Disclaimer:
Results have been estimated based on internal Intel analysis and are provided
for informational purposes only. Any difference in system hardware or software
design or configuration may affect actual performance.


-- 
0-DAY CI Kernel Test Service
https://github.com/intel/lkp-tests/wiki


Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ