lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [day] [month] [year] [list]
Message-ID: <202409241539.4b10c59b-oliver.sang@intel.com>
Date: Tue, 24 Sep 2024 15:31:18 +0800
From: kernel test robot <oliver.sang@...el.com>
To: Aristeu Rozanski <aris@...hat.com>
CC: <oe-lkp@...ts.linux.dev>, <lkp@...el.com>, <linux-kernel@...r.kernel.org>,
	Andrew Morton <akpm@...ux-foundation.org>, Muchun Song
	<muchun.song@...ux.dev>, Aristeu Rozanski <aris@...vo.org>, David Hildenbrand
	<david@...hat.com>, Vishal Moola <vishal.moola@...il.com>,
	<linux-mm@...ck.org>, <ying.huang@...el.com>, <feng.tang@...el.com>,
	<fengwei.yin@...el.com>, <oliver.sang@...el.com>
Subject: [linus:master] [hugetlb]  003af997c8:  vm-scalability.throughput
 -44.1% regression



Hello,

kernel test robot noticed a -44.1% regression of vm-scalability.throughput on:


commit: 003af997c8a945493859dd1a2d015cc9387ff27a ("hugetlb: force allocating surplus hugepages on mempolicy allowed nodes")
https://git.kernel.org/cgit/linux/kernel/git/torvalds/linux.git master

testcase: vm-scalability
test machine: 128 threads 2 sockets Intel(R) Xeon(R) Gold 6338 CPU @ 2.00GHz (Ice Lake) with 256G memory
parameters:

	runtime: 300s
	size: 512G
	test: anon-w-rand-hugetlb
	cpufreq_governor: performance




If you fix the issue in a separate patch/commit (i.e. not just a new version of
the same patch/commit), kindly add following tags
| Reported-by: kernel test robot <oliver.sang@...el.com>
| Closes: https://lore.kernel.org/oe-lkp/202409241539.4b10c59b-oliver.sang@intel.com


Details are as below:
-------------------------------------------------------------------------------------------------->


The kernel config and materials to reproduce are available at:
https://download.01.org/0day-ci/archive/20240924/202409241539.4b10c59b-oliver.sang@intel.com

=========================================================================================
compiler/cpufreq_governor/kconfig/rootfs/runtime/size/tbox_group/test/testcase:
  gcc-12/performance/x86_64-rhel-8.3/debian-12-x86_64-20240206.cgz/300s/512G/lkp-icl-2sp2/anon-w-rand-hugetlb/vm-scalability

commit: 
  64548bc534 ("mm/damon/paddr: initialize nr_succeeded in __damon_pa_migrate_folio_list()")
  003af997c8 ("hugetlb: force allocating surplus hugepages on mempolicy allowed nodes")

64548bc5347bd051 003af997c8a945493859dd1a2d0 
---------------- --------------------------- 
         %stddev     %change         %stddev
             \          |                \  
 3.539e+09 ± 10%     -48.1%  1.836e+09 ±  9%  cpuidle..time
    244.26 ±  6%     +39.2%     340.02        uptime.boot
      9914 ±  6%     -21.2%       7808        uptime.idle
     14.40 ±  8%      -9.6        4.81 ±  9%  mpstat.cpu.all.idle%
      0.02 ±  2%      +0.0        0.03        mpstat.cpu.all.soft%
      6.61 ±  5%      +3.3        9.87        mpstat.cpu.all.sys%
      4.00          +129.2%       9.17 ± 30%  mpstat.max_utilization.seconds
      3.80 ± 80%   +2470.2%      97.67 ±  7%  perf-c2c.DRAM.local
     27.60 ± 31%   +4957.4%       1395        perf-c2c.DRAM.remote
     22.00 ± 18%   +3278.8%     743.33 ±  4%  perf-c2c.HITM.local
      9.20 ± 40%   +6387.3%     596.83 ±  3%  perf-c2c.HITM.remote
     64.40 ±200%  +63684.7%      41077 ± 20%  numa-numastat.node0.numa_foreign
    662185 ±  6%     +50.0%     993375 ±  5%  numa-numastat.node0.numa_hit
     38632 ± 70%    +825.8%     357657 ±  9%  numa-numastat.node0.other_node
    723573 ±  7%     -19.4%     583311 ±  8%  numa-numastat.node1.numa_hit
     64.40 ±200%  +64112.2%      41352 ± 20%  numa-numastat.node1.numa_miss
     15.17 ±  7%     -63.9%       5.48 ±  8%  vmstat.cpu.id
     78.08            +8.3%      84.58        vmstat.cpu.us
    110.68           +11.2%     123.04        vmstat.procs.r
      2778 ±  2%     -11.2%       2465 ±  3%  vmstat.system.cs
    142558 ±  2%     +11.8%     159336        vmstat.system.in
      0.00 ±  6%     -28.7%       0.00 ±  3%  vm-scalability.free_time
     57226 ±  5%     -41.4%      33548        vm-scalability.median
     15.32 ± 36%     -14.7        0.58 ± 20%  vm-scalability.median_stddev%
     15.16 ± 43%     -14.7        0.42 ± 48%  vm-scalability.stddev%
   7692177 ±  5%     -44.1%    4299393        vm-scalability.throughput
    189.12 ±  7%     +52.1%     287.59        vm-scalability.time.elapsed_time
    189.12 ±  7%     +52.1%     287.59        vm-scalability.time.elapsed_time.max
    142450 ±  7%     +63.3%     232670 ±  5%  vm-scalability.time.involuntary_context_switches
     10933           +11.2%      12161        vm-scalability.time.percent_of_cpu_this_job_got
      1574 ± 12%    +128.3%       3594        vm-scalability.time.system_time
     19105 ±  7%     +64.3%      31384        vm-scalability.time.user_time
      4986 ±  5%     +41.0%       7031        vm-scalability.time.voluntary_context_switches
    107334 ± 18%    +140.2%     257850 ±  2%  meminfo.Active
    107199 ± 18%    +140.4%     257707 ±  2%  meminfo.Active(anon)
   1060745 ±  2%     +13.3%    1202096        meminfo.Committed_AS
      1593 ± 17%     +88.9%       3009 ±  4%  meminfo.HugePages_Free
      1593 ± 17%     +88.9%       3010 ±  3%  meminfo.HugePages_Rsvd
     54941           +11.5%      61250        meminfo.HugePages_Surp
     54941           +11.5%      61250        meminfo.HugePages_Total
 1.125e+08           +11.5%  1.254e+08        meminfo.Hugetlb
      1203           -50.3%     597.80 ± 82%  meminfo.Inactive(file)
  1.44e+08            -8.9%  1.313e+08        meminfo.MemAvailable
 1.451e+08            -8.8%  1.324e+08        meminfo.MemFree
 1.186e+08           +10.7%  1.314e+08        meminfo.Memused
    255862 ± 10%     +58.6%     405684        meminfo.Shmem
  18068249 ± 12%     -90.6%    1706023 ±  6%  numa-vmstat.node0.nr_free_pages
    179.13 ± 56%     -95.6%       7.82 ±105%  numa-vmstat.node0.nr_inactive_file
      1479 ± 10%     -37.5%     924.24 ± 20%  numa-vmstat.node0.nr_page_table_pages
    179.13 ± 56%     -95.6%       7.82 ±105%  numa-vmstat.node0.nr_zone_inactive_file
     64.40 ±200%  +63684.7%      41077 ± 20%  numa-vmstat.node0.numa_foreign
    662831 ±  6%     +49.8%     993007 ±  5%  numa-vmstat.node0.numa_hit
     38632 ± 70%    +825.8%     357657 ±  9%  numa-vmstat.node0.numa_other
     25754 ± 19%    +104.8%      52745 ± 20%  numa-vmstat.node1.nr_active_anon
  18214205 ± 13%     +72.3%   31383668        numa-vmstat.node1.nr_free_pages
      1266 ± 13%     +48.7%       1883 ± 10%  numa-vmstat.node1.nr_page_table_pages
     43011 ± 44%     +80.8%      77760 ± 25%  numa-vmstat.node1.nr_shmem
     25754 ± 19%    +104.8%      52745 ± 20%  numa-vmstat.node1.nr_zone_active_anon
    722032 ±  7%     -19.3%     582363 ±  8%  numa-vmstat.node1.numa_hit
     64.40 ±200%  +64112.2%      41352 ± 20%  numa-vmstat.node1.numa_miss
    763.05 ± 34%    +287.1%       2954 ±  6%  numa-meminfo.node0.HugePages_Free
     27885 ± 15%    +113.2%      59456        numa-meminfo.node0.HugePages_Surp
     27885 ± 15%    +113.2%      59456        numa-meminfo.node0.HugePages_Total
    716.02 ± 56%     -95.6%      31.28 ±105%  numa-meminfo.node0.Inactive(file)
  72184008 ± 12%     -90.4%    6907510 ± 11%  numa-meminfo.node0.MemFree
  59500944 ± 15%    +109.7%  1.248e+08        numa-meminfo.node0.MemUsed
      5935 ± 10%     -37.7%       3696 ± 19%  numa-meminfo.node0.PageTables
    102909 ± 19%    +104.9%     210848 ± 21%  numa-meminfo.node1.Active
    102855 ± 19%    +104.9%     210789 ± 21%  numa-meminfo.node1.Active(anon)
    756.97 ± 40%     -99.9%       0.99 ±223%  numa-meminfo.node1.HugePages_Free
     27143 ± 16%     -93.5%       1757 ± 33%  numa-meminfo.node1.HugePages_Surp
     27143 ± 16%     -93.5%       1757 ± 33%  numa-meminfo.node1.HugePages_Total
  72746397 ± 13%     +72.6%  1.255e+08        numa-meminfo.node1.MemFree
  59311794 ± 16%     -89.0%    6524390        numa-meminfo.node1.MemUsed
      5070 ± 13%     +48.3%       7518 ± 10%  numa-meminfo.node1.PageTables
    172326 ± 44%     +80.4%     310895 ± 25%  numa-meminfo.node1.Shmem
      0.01 ±122%    +311.8%       0.04 ± 79%  perf-sched.sch_delay.avg.ms.__cond_resched.down_write_killable.exec_mmap.begin_new_exec.load_elf_binary
      0.01 ± 84%    +998.4%       0.16 ±117%  perf-sched.sch_delay.avg.ms.__cond_resched.dput.__fput.__x64_sys_close.do_syscall_64
      0.01 ± 98%    +669.5%       0.07 ± 51%  perf-sched.sch_delay.avg.ms.__cond_resched.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
      0.13 ± 86%    +329.8%       0.54 ± 54%  perf-sched.sch_delay.avg.ms.do_wait.kernel_wait4.__do_sys_wait4.do_syscall_64
      0.01 ± 50%    +248.7%       0.05 ±  7%  perf-sched.sch_delay.avg.ms.irq_thread.kthread.ret_from_fork.ret_from_fork_asm
      0.00 ±200%    +1e+05%       1.04 ± 35%  perf-sched.sch_delay.avg.ms.irqentry_exit_to_user_mode.asm_exc_page_fault.[unknown]
      0.03 ± 18%    +531.6%       0.18 ±151%  perf-sched.sch_delay.avg.ms.pipe_read.vfs_read.ksys_read.do_syscall_64
      0.41 ± 76%    +103.8%       0.84 ± 18%  perf-sched.sch_delay.avg.ms.schedule_hrtimeout_range_clock.do_poll.constprop.0.do_sys_poll
      0.02 ± 19%    +639.7%       0.14 ± 61%  perf-sched.sch_delay.avg.ms.schedule_timeout.kcompactd.kthread.ret_from_fork
      0.01 ± 36%    +216.9%       0.04 ± 19%  perf-sched.sch_delay.avg.ms.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
      0.02 ± 81%   +2492.3%       0.44 ±133%  perf-sched.sch_delay.max.ms.__cond_resched.dput.__fput.__x64_sys_close.do_syscall_64
      0.02 ± 87%   +2603.1%       0.52 ±106%  perf-sched.sch_delay.max.ms.__cond_resched.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
      0.03 ± 50%    +509.5%       0.17 ± 14%  perf-sched.sch_delay.max.ms.irq_thread.kthread.ret_from_fork.ret_from_fork_asm
      0.00 ±200%  +3.7e+05%       3.66 ± 14%  perf-sched.sch_delay.max.ms.irqentry_exit_to_user_mode.asm_exc_page_fault.[unknown]
      0.04 ± 45%   +4905.2%       1.94 ± 72%  perf-sched.sch_delay.max.ms.schedule_timeout.kcompactd.kthread.ret_from_fork
      0.20 ± 68%    +215.9%       0.63 ± 47%  perf-sched.total_sch_delay.average.ms
     60.34 ±136%   +1104.4%     726.77 ± 77%  perf-sched.total_sch_delay.max.ms
    306.32 ± 61%     +57.1%     481.25 ±  2%  perf-sched.wait_and_delay.avg.ms.schedule_hrtimeout_range_clock.do_poll.constprop.0.do_sys_poll
      2.24 ± 49%     +79.0%       4.01 ± 20%  perf-sched.wait_and_delay.avg.ms.schedule_timeout.__wait_for_common.wait_for_completion_state.kernel_clone
     11.87 ± 29%     -37.8%       7.38 ±  8%  perf-sched.wait_and_delay.avg.ms.schedule_timeout.rcu_gp_fqs_loop.rcu_gp_kthread.kthread
    434.40 ± 29%     +51.8%     659.33 ±  8%  perf-sched.wait_and_delay.count.schedule_timeout.rcu_gp_fqs_loop.rcu_gp_kthread.kthread
      6.37 ± 37%     +62.1%      10.33 ± 22%  perf-sched.wait_and_delay.max.ms.irqentry_exit_to_user_mode.asm_sysvec_reschedule_ipi.[unknown]
      0.01 ± 84%    +998.4%       0.16 ±117%  perf-sched.wait_time.avg.ms.__cond_resched.dput.__fput.__x64_sys_close.do_syscall_64
      1.97 ± 47%     +84.7%       3.64 ± 19%  perf-sched.wait_time.avg.ms.schedule_timeout.__wait_for_common.wait_for_completion_state.kernel_clone
     11.80 ± 29%     -38.2%       7.29 ±  8%  perf-sched.wait_time.avg.ms.schedule_timeout.rcu_gp_fqs_loop.rcu_gp_kthread.kthread
      0.02 ± 81%   +2492.3%       0.44 ±133%  perf-sched.wait_time.max.ms.__cond_resched.dput.__fput.__x64_sys_close.do_syscall_64
      3.28 ± 30%    +147.1%       8.10 ± 45%  perf-sched.wait_time.max.ms.irqentry_exit_to_user_mode.asm_sysvec_reschedule_ipi.[unknown]
     17.20 ±  2%      -6.0%      16.17        perf-stat.i.MPKI
 1.569e+10 ±  7%     -33.7%  1.041e+10        perf-stat.i.branch-instructions
  10996521 ±  6%     -32.9%    7383042 ±  2%  perf-stat.i.branch-misses
 8.129e+08 ±  7%     -34.7%  5.307e+08        perf-stat.i.cache-misses
 8.307e+08 ±  7%     -34.6%  5.436e+08        perf-stat.i.cache-references
      2701           -11.3%       2396 ±  3%  perf-stat.i.context-switches
      5.62 ± 10%     +70.2%       9.57        perf-stat.i.cpi
 2.884e+11           +10.8%  3.195e+11        perf-stat.i.cpu-cycles
    374.61 ± 10%     +60.3%     600.58        perf-stat.i.cycles-between-cache-misses
 6.744e+10 ±  7%     -33.7%  4.472e+10        perf-stat.i.instructions
      0.25 ±  7%     -40.2%       0.15 ±  2%  perf-stat.i.ipc
      0.20 ± 57%     -72.5%       0.05 ± 97%  perf-stat.i.major-faults
      6025 ±  4%     -20.3%       4801        perf-stat.i.minor-faults
      6025 ±  4%     -20.3%       4801        perf-stat.i.page-faults
     12.13            -1.2%      11.98        perf-stat.overall.MPKI
      4.32 ±  7%     +66.7%       7.19        perf-stat.overall.cpi
    355.77 ±  8%     +68.7%     600.33        perf-stat.overall.cycles-between-cache-misses
      0.23 ±  7%     -40.3%       0.14        perf-stat.overall.ipc
     10566            +1.3%      10705        perf-stat.overall.path-length
 1.543e+10 ±  7%     -33.6%  1.024e+10        perf-stat.ps.branch-instructions
  10668497 ±  6%     -32.6%    7189782 ±  2%  perf-stat.ps.branch-misses
 8.043e+08 ±  7%     -34.5%  5.269e+08        perf-stat.ps.cache-misses
 8.224e+08 ±  7%     -34.3%  5.402e+08        perf-stat.ps.cache-references
      2686           -10.4%       2406 ±  3%  perf-stat.ps.context-switches
 2.845e+11           +11.2%  3.163e+11        perf-stat.ps.cpu-cycles
 6.629e+10 ±  7%     -33.7%  4.397e+10        perf-stat.ps.instructions
      0.20 ± 56%     -71.7%       0.06 ± 96%  perf-stat.ps.major-faults
      6139 ±  4%     -19.7%       4928        perf-stat.ps.minor-faults
      6139 ±  4%     -19.7%       4928        perf-stat.ps.page-faults
 1.254e+13            +1.3%  1.271e+13        perf-stat.total.instructions
  12803673 ±122%   +2705.8%  3.592e+08 ± 20%  proc-vmstat.compact_daemon_free_scanned
    188006 ±147%   +2684.0%    5234004 ± 55%  proc-vmstat.compact_daemon_migrate_scanned
  12803673 ±122%   +2705.8%  3.592e+08 ± 20%  proc-vmstat.compact_free_scanned
    188006 ±147%   +2684.0%    5234004 ± 55%  proc-vmstat.compact_migrate_scanned
     26765 ± 18%    +140.9%      64490 ±  2%  proc-vmstat.nr_active_anon
   3601411            -9.1%    3272731        proc-vmstat.nr_dirty_background_threshold
   7211629            -9.1%    6553464        proc-vmstat.nr_dirty_threshold
    837964            +4.4%     875032        proc-vmstat.nr_file_pages
  36351167            -9.1%   33059841        proc-vmstat.nr_free_pages
    300.91           -50.4%     149.15 ± 82%  proc-vmstat.nr_inactive_file
      2743            +2.3%       2805        proc-vmstat.nr_page_table_pages
     63967 ±  9%     +58.7%     101505        proc-vmstat.nr_shmem
     28763            -6.7%      26832 ±  2%  proc-vmstat.nr_slab_reclaimable
     26765 ± 18%    +140.9%      64490 ±  2%  proc-vmstat.nr_zone_active_anon
    300.91           -50.4%     149.15 ± 82%  proc-vmstat.nr_zone_inactive_file
    506.80 ±170%   +8005.2%      41077 ± 20%  proc-vmstat.numa_foreign
     19032 ± 45%     +77.0%      33682 ± 16%  proc-vmstat.numa_hint_faults
   1393599 ±  2%     +13.3%    1579111        proc-vmstat.numa_hit
    506.80 ±170%   +8059.6%      41352 ± 20%  proc-vmstat.numa_miss
    133133          +232.2%     442280 ±  8%  proc-vmstat.numa_other
      8765 ± 63%    +271.3%      32545 ± 17%  proc-vmstat.numa_pages_migrated
    124049 ± 35%     +69.0%     209637 ± 14%  proc-vmstat.numa_pte_updates
     69397 ± 14%     +41.1%      97902        proc-vmstat.pgactivate
    512.00 ±200%   +3700.7%      19459 ±  3%  proc-vmstat.pgalloc_dma
     80179 ±200%   +3713.5%    3057605        proc-vmstat.pgalloc_dma32
 2.651e+08            -1.0%  2.624e+08        proc-vmstat.pgalloc_normal
   1229462 ±  2%     +21.5%    1494275        proc-vmstat.pgfault
    154.00 ±199%   +2344.7%       3764 ± 76%  proc-vmstat.pgmigrate_fail
     20244 ± 71%    +221.2%      65033 ± 19%  proc-vmstat.pgmigrate_success
     33783 ±  5%     +29.3%      43678 ±  2%  proc-vmstat.pgreuse
   8518493 ± 21%     +89.5%   16143327        sched_debug.cfs_rq:/.avg_vruntime.avg
   9012023 ± 20%     +83.6%   16543009        sched_debug.cfs_rq:/.avg_vruntime.max
   7203984 ± 20%     +98.9%   14327528 ±  2%  sched_debug.cfs_rq:/.avg_vruntime.min
      0.65 ± 12%     +27.0%       0.83        sched_debug.cfs_rq:/.h_nr_running.avg
      0.30 ± 13%    +133.3%       0.70 ± 14%  sched_debug.cfs_rq:/.h_nr_running.min
      0.26 ± 17%     -35.6%       0.17 ±  6%  sched_debug.cfs_rq:/.h_nr_running.stddev
      2410 ± 14%    +133.2%       5621 ± 14%  sched_debug.cfs_rq:/.load.min
      2.27 ± 11%    +122.1%       5.03 ± 11%  sched_debug.cfs_rq:/.load_avg.min
   8518493 ± 21%     +89.5%   16143327        sched_debug.cfs_rq:/.min_vruntime.avg
   9012023 ± 20%     +83.6%   16543009        sched_debug.cfs_rq:/.min_vruntime.max
   7203984 ± 20%     +98.9%   14327528 ±  2%  sched_debug.cfs_rq:/.min_vruntime.min
      0.65 ± 12%     +26.5%       0.82        sched_debug.cfs_rq:/.nr_running.avg
      0.30 ± 13%    +133.3%       0.70 ± 14%  sched_debug.cfs_rq:/.nr_running.min
      0.21 ± 27%     -61.7%       0.08 ± 21%  sched_debug.cfs_rq:/.nr_running.stddev
    702.48 ± 11%     +24.0%     871.29        sched_debug.cfs_rq:/.runnable_avg.avg
    282.30 ± 22%    +136.3%     667.03 ± 20%  sched_debug.cfs_rq:/.runnable_avg.min
    240.81 ± 22%     -47.7%     126.02 ± 15%  sched_debug.cfs_rq:/.runnable_avg.stddev
    696.46 ± 10%     +23.9%     863.11        sched_debug.cfs_rq:/.util_avg.avg
    227.72 ± 38%    +140.3%     547.27 ± 28%  sched_debug.cfs_rq:/.util_avg.min
    219.06 ± 27%     -53.9%     101.05 ± 19%  sched_debug.cfs_rq:/.util_avg.stddev
    503.97 ± 18%     +53.7%     774.51 ±  3%  sched_debug.cfs_rq:/.util_est.avg
      1047 ± 14%     +28.5%       1346 ± 14%  sched_debug.cfs_rq:/.util_est.max
    279.14 ± 11%     -37.2%     175.37 ± 16%  sched_debug.cfs_rq:/.util_est.stddev
   1111522 ±  7%     +15.5%    1283701 ±  5%  sched_debug.cpu.avg_idle.max
    351610 ± 23%     +48.8%     523030 ±  6%  sched_debug.cpu.avg_idle.min
    126692 ± 11%     +35.4%     171555        sched_debug.cpu.clock.avg
    126710 ± 11%     +35.4%     171591        sched_debug.cpu.clock.max
    126672 ± 11%     +35.4%     171515        sched_debug.cpu.clock.min
    126285 ± 11%     +35.4%     171003        sched_debug.cpu.clock_task.avg
    126456 ± 11%     +35.4%     171228        sched_debug.cpu.clock_task.max
    117086 ± 12%     +38.3%     161970        sched_debug.cpu.clock_task.min
      4361 ± 16%     +44.0%       6281        sched_debug.cpu.curr->pid.avg
      6032 ±  7%     +20.8%       7284        sched_debug.cpu.curr->pid.max
      2157 ± 26%    +118.4%       4711 ± 24%  sched_debug.cpu.curr->pid.min
    992.57 ± 24%     -58.1%     415.92 ± 40%  sched_debug.cpu.curr->pid.stddev
      0.64 ± 12%     +29.0%       0.83        sched_debug.cpu.nr_running.avg
      0.30 ± 13%    +133.3%       0.70 ± 14%  sched_debug.cpu.nr_running.min
      0.27 ± 15%     -37.9%       0.17 ±  8%  sched_debug.cpu.nr_running.stddev
      2936 ±  9%     +22.6%       3599 ±  3%  sched_debug.cpu.nr_switches.avg
    701.50 ± 18%     +43.1%       1003 ±  3%  sched_debug.cpu.nr_switches.min
    126673 ± 11%     +35.4%     171514        sched_debug.cpu_clk
    125408 ± 12%     +35.8%     170282        sched_debug.ktime
    127719 ± 11%     +35.0%     172457        sched_debug.sched_clk
     16.02 ± 41%     -14.3        1.75        perf-profile.calltrace.cycles-pp.do_access
     13.69 ± 51%     -13.7        0.00        perf-profile.calltrace.cycles-pp.asm_exc_page_fault.do_access
     13.66 ± 51%     -13.7        0.00        perf-profile.calltrace.cycles-pp.do_user_addr_fault.exc_page_fault.asm_exc_page_fault.do_access
     13.66 ± 51%     -13.7        0.00        perf-profile.calltrace.cycles-pp.exc_page_fault.asm_exc_page_fault.do_access
     13.65 ± 51%     -13.6        0.00        perf-profile.calltrace.cycles-pp.handle_mm_fault.do_user_addr_fault.exc_page_fault.asm_exc_page_fault.do_access
     13.63 ± 51%     -13.6        0.00        perf-profile.calltrace.cycles-pp.hugetlb_fault.handle_mm_fault.do_user_addr_fault.exc_page_fault.asm_exc_page_fault
     12.86 ± 51%     -12.9        0.00        perf-profile.calltrace.cycles-pp.hugetlb_no_page.hugetlb_fault.handle_mm_fault.do_user_addr_fault.exc_page_fault
     12.60 ± 51%     -12.6        0.00        perf-profile.calltrace.cycles-pp.folio_zero_user.hugetlb_no_page.hugetlb_fault.handle_mm_fault.do_user_addr_fault
     11.75 ± 51%     -11.7        0.00        perf-profile.calltrace.cycles-pp.clear_page_erms.folio_zero_user.hugetlb_no_page.hugetlb_fault.handle_mm_fault
      2.29 ±  7%      -0.7        1.55        perf-profile.calltrace.cycles-pp.nrand48_r
      1.15 ± 29%      +1.9        3.05 ±  7%  perf-profile.calltrace.cycles-pp.lrand48_r
     44.16 ±  7%      +7.7       51.86 ±  7%  perf-profile.calltrace.cycles-pp.lrand48_r@plt
     57.17 ±  9%     +11.6       68.72 ±  3%  perf-profile.calltrace.cycles-pp.do_rw_once
     16.34 ± 41%     -14.4        1.95        perf-profile.children.cycles-pp.do_access
     13.70 ± 51%     -13.7        0.00        perf-profile.children.cycles-pp.asm_exc_page_fault
     13.68 ± 51%     -13.7        0.00        perf-profile.children.cycles-pp.exc_page_fault
     13.68 ± 51%     -13.7        0.00        perf-profile.children.cycles-pp.do_user_addr_fault
     13.66 ± 51%     -13.7        0.00        perf-profile.children.cycles-pp.handle_mm_fault
     13.63 ± 51%     -13.6        0.00        perf-profile.children.cycles-pp.hugetlb_fault
     12.86 ± 51%     -12.9        0.00        perf-profile.children.cycles-pp.hugetlb_no_page
     12.68 ± 51%     -12.7        0.00        perf-profile.children.cycles-pp.folio_zero_user
     12.00 ± 51%     -12.0        0.00        perf-profile.children.cycles-pp.clear_page_erms
      3.04 ±  7%      -1.0        2.07        perf-profile.children.cycles-pp.nrand48_r
      1.03 ± 49%      -1.0        0.06 ±  6%  perf-profile.children.cycles-pp.entry_SYSCALL_64_after_hwframe
      1.02 ± 49%      -1.0        0.06 ±  6%  perf-profile.children.cycles-pp.do_syscall_64
      0.64 ± 13%      -0.2        0.39 ±  5%  perf-profile.children.cycles-pp.asm_sysvec_apic_timer_interrupt
      0.45 ±  9%      -0.2        0.28 ±  8%  perf-profile.children.cycles-pp.sysvec_apic_timer_interrupt
      0.42 ±  9%      -0.2        0.26 ±  8%  perf-profile.children.cycles-pp.__sysvec_apic_timer_interrupt
      0.41 ± 10%      -0.2        0.25 ±  8%  perf-profile.children.cycles-pp.hrtimer_interrupt
      0.30 ± 10%      -0.1        0.18 ±  7%  perf-profile.children.cycles-pp.__hrtimer_run_queues
      0.29 ± 10%      -0.1        0.17 ±  7%  perf-profile.children.cycles-pp.tick_nohz_handler
      0.26 ± 10%      -0.1        0.15 ±  6%  perf-profile.children.cycles-pp.update_process_times
      0.14 ±  9%      -0.0        0.10 ±  6%  perf-profile.children.cycles-pp.sched_tick
      0.08 ± 17%      -0.0        0.04 ± 71%  perf-profile.children.cycles-pp.ktime_get
      0.08 ± 15%      -0.0        0.06 ± 16%  perf-profile.children.cycles-pp.clockevents_program_event
      0.94 ±  9%      +0.8        1.74 ±  6%  perf-profile.children.cycles-pp.lrand48_r
     22.76 ±  7%      +4.7       27.50 ±  7%  perf-profile.children.cycles-pp.lrand48_r@plt
     79.53 ±  8%     +15.3       94.84        perf-profile.children.cycles-pp.do_rw_once
     11.89 ± 51%     -11.9        0.00        perf-profile.self.cycles-pp.clear_page_erms
      2.73 ±  7%      -0.9        1.87        perf-profile.self.cycles-pp.nrand48_r
      2.33 ± 12%      -0.6        1.75        perf-profile.self.cycles-pp.do_access
      0.33 ±  8%      -0.1        0.23        perf-profile.self.cycles-pp.lrand48_r
      0.62 ± 12%      +0.9        1.52 ±  7%  perf-profile.self.cycles-pp.lrand48_r@plt
     78.68 ±  8%     +15.5       94.20        perf-profile.self.cycles-pp.do_rw_once




Disclaimer:
Results have been estimated based on internal Intel analysis and are provided
for informational purposes only. Any difference in system hardware or software
design or configuration may affect actual performance.


-- 
0-DAY CI Kernel Test Service
https://github.com/intel/lkp-tests/wiki


Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ