[<prev] [next>] [thread-next>] [day] [month] [year] [list]
Message-ID: <202402081435.c3faa4b4-oliver.sang@intel.com>
Date: Thu, 8 Feb 2024 15:36:46 +0800
From: kernel test robot <oliver.sang@...el.com>
To: Ryan Roberts <ryan.roberts@....com>
CC: <oe-lkp@...ts.linux.dev>, <lkp@...el.com>, <linux-kernel@...r.kernel.org>,
Andrew Morton <akpm@...ux-foundation.org>, Yang Shi <shy828301@...il.com>,
Matthew Wilcox <willy@...radead.org>, Rik van Riel <riel@...riel.com>,
<linux-mm@...ck.org>, <ying.huang@...el.com>, <feng.tang@...el.com>,
<fengwei.yin@...el.com>, <oliver.sang@...el.com>
Subject: [linus:master] [mm] 96204e1531: will-it-scale.per_thread_ops -3.9%
regression
Hello,
kernel test robot noticed a -3.9% regression of will-it-scale.per_thread_ops on:
commit: 96204e15310c218fd9355bdcacd02fed1d18070e ("mm: thp_get_unmapped_area must honour topdown preference")
https://git.kernel.org/cgit/linux/kernel/git/torvalds/linux.git master
testcase: will-it-scale
test machine: 224 threads 2 sockets Intel(R) Xeon(R) Platinum 8480CTDX (Sapphire Rapids) with 512G memory
parameters:
test: page_fault1
cpufreq_governor: performance
If you fix the issue in a separate patch/commit (i.e. not just a new version of
the same patch/commit), kindly add following tags
| Reported-by: kernel test robot <oliver.sang@...el.com>
| Closes: https://lore.kernel.org/oe-lkp/202402081435.c3faa4b4-oliver.sang@intel.com
Details are as below:
-------------------------------------------------------------------------------------------------->
The kernel config and materials to reproduce are available at:
https://download.01.org/0day-ci/archive/20240208/202402081435.c3faa4b4-oliver.sang@intel.com
=========================================================================================
compiler/cpufreq_governor/kconfig/rootfs/tbox_group/test/testcase:
gcc-12/performance/x86_64-rhel-8.3/debian-11.1-x86_64-20220510.cgz/lkp-spr-2sp4/page_fault1/will-it-scale
commit:
4ef9ad19e1 ("mm: huge_memory: don't force huge page alignment on 32 bit")
96204e1531 ("mm: thp_get_unmapped_area must honour topdown preference")
4ef9ad19e17676b9 96204e15310c218fd9355bdcacd
---------------- ---------------------------
%stddev %change %stddev
\ | \
32.93 -3.5 29.43 mpstat.cpu.all.sys%
77.89 ± 2% -10.2% 69.98 vmstat.procs.r
4678 +14.0% 5332 vmstat.system.cs
251960 -3.4% 243306 vmstat.system.in
1049 -10.2% 943.00 turbostat.Avg_MHz
36.22 -3.7 32.54 turbostat.Busy%
29.10 +1.8 30.87 turbostat.C1E%
0.35 +15.7% 0.40 turbostat.CPU%c6
565.62 -2.3% 552.53 turbostat.PkgWatt
123847 -18.4% 101096 meminfo.Active
123657 -18.3% 101000 meminfo.Active(anon)
8127530 +10.3% 8966990 meminfo.AnonHugePages
8298705 +10.3% 9156780 meminfo.AnonPages
8417195 +10.1% 9267954 meminfo.Inactive
8416461 +10.1% 9267742 meminfo.Inactive(anon)
138121 -16.9% 114720 meminfo.Shmem
4852175 ± 2% +9.2% 5296708 numa-meminfo.node0.AnonHugePages
4941270 ± 2% +9.0% 5387476 numa-meminfo.node0.AnonPages
121571 -18.9% 98652 numa-meminfo.node1.Active
121528 -18.9% 98620 numa-meminfo.node1.Active(anon)
3270590 +12.4% 3674514 numa-meminfo.node1.AnonHugePages
3352686 +12.6% 3773824 numa-meminfo.node1.AnonPages
3395079 +12.4% 3815833 numa-meminfo.node1.Inactive
3394971 +12.4% 3815762 numa-meminfo.node1.Inactive(anon)
125041 -18.4% 102036 numa-meminfo.node1.Shmem
1219136 ± 2% +10.3% 1344427 numa-vmstat.node0.nr_anon_pages
2337 +10.4% 2581 numa-vmstat.node0.nr_anon_transparent_hugepages
1237208 ± 2% +10.1% 1362084 numa-vmstat.node0.nr_inactive_anon
1236994 ± 2% +10.1% 1362125 numa-vmstat.node0.nr_zone_inactive_anon
30361 -18.8% 24656 numa-vmstat.node1.nr_active_anon
828422 +13.5% 940280 numa-vmstat.node1.nr_anon_pages
1577 +13.3% 1788 numa-vmstat.node1.nr_anon_transparent_hugepages
838873 +13.3% 950629 ± 2% numa-vmstat.node1.nr_inactive_anon
31244 -18.4% 25510 numa-vmstat.node1.nr_shmem
30361 -18.8% 24656 numa-vmstat.node1.nr_zone_active_anon
838696 +13.4% 950674 ± 2% numa-vmstat.node1.nr_zone_inactive_anon
30907 -18.4% 25233 proc-vmstat.nr_active_anon
2068862 +11.0% 2295458 proc-vmstat.nr_anon_pages
3957 +11.0% 4390 proc-vmstat.nr_anon_transparent_hugepages
2099316 +10.6% 2322543 proc-vmstat.nr_inactive_anon
7379 +4.4% 7701 proc-vmstat.nr_page_table_pages
34526 -17.0% 28665 proc-vmstat.nr_shmem
30907 -18.4% 25233 proc-vmstat.nr_zone_active_anon
2099307 +10.6% 2322537 proc-vmstat.nr_zone_inactive_anon
26039 ± 10% -16.4% 21774 ± 6% proc-vmstat.numa_hint_faults_local
51127959 -3.7% 49228917 proc-vmstat.numa_hit
50891640 -3.7% 48994322 proc-vmstat.numa_local
32315 -21.8% 25269 proc-vmstat.pgactivate
1.265e+10 -3.8% 1.217e+10 proc-vmstat.pgalloc_normal
25947640 -3.6% 25012103 proc-vmstat.pgfault
1.265e+10 -3.8% 1.217e+10 proc-vmstat.pgfree
24650837 -3.8% 23717158 proc-vmstat.thp_fault_alloc
44754589 -15.5% 37839290 will-it-scale.112.threads
74.20 +11.2% 82.49 will-it-scale.112.threads_idle
36349325 -6.0% 34182802 will-it-scale.168.threads
72.79 +11.7% 81.28 will-it-scale.168.threads_idle
40356275 -5.6% 38083547 will-it-scale.224.threads
66.97 +19.5% 80.06 will-it-scale.224.threads_idle
28146956 -15.5% 23782499 will-it-scale.56.threads
78.16 +8.4% 84.70 will-it-scale.56.threads_idle
795871 -3.9% 764972 will-it-scale.per_thread_ops
67103 -4.9% 63834 will-it-scale.time.involuntary_context_switches
17508243 ± 4% +6.9% 18724480 will-it-scale.time.maximum_resident_set_size
22382078 -3.6% 21577203 will-it-scale.time.minor_page_faults
7050 -7.3% 6534 will-it-scale.time.percent_of_cpu_this_job_got
20998 -7.4% 19448 will-it-scale.time.system_time
1075 -3.3% 1040 will-it-scale.time.user_time
236326 +40.3% 331522 will-it-scale.time.voluntary_context_switches
4.056e+08 -3.9% 3.899e+08 will-it-scale.workload
5880710 -9.8% 5305863 sched_debug.cfs_rq:/.avg_vruntime.avg
10300738 ± 2% -11.0% 9163982 ± 2% sched_debug.cfs_rq:/.avg_vruntime.max
1772532 ± 6% -15.5% 1498526 ± 4% sched_debug.cfs_rq:/.avg_vruntime.min
2066050 -12.1% 1817081 sched_debug.cfs_rq:/.avg_vruntime.stddev
321.14 ± 36% -63.9% 115.93 ± 57% sched_debug.cfs_rq:/.load_avg.avg
1945 ± 20% -47.8% 1015 ± 50% sched_debug.cfs_rq:/.load_avg.stddev
5880710 -9.8% 5305863 sched_debug.cfs_rq:/.min_vruntime.avg
10300738 ± 2% -11.0% 9163982 ± 2% sched_debug.cfs_rq:/.min_vruntime.max
1772532 ± 6% -15.5% 1498526 ± 4% sched_debug.cfs_rq:/.min_vruntime.min
2066050 -12.1% 1817081 sched_debug.cfs_rq:/.min_vruntime.stddev
233.57 ± 8% -24.3% 176.83 ± 3% sched_debug.cfs_rq:/.runnable_avg.avg
958.36 ± 4% -9.2% 870.11 ± 3% sched_debug.cfs_rq:/.runnable_avg.max
240.06 ± 5% -20.4% 190.99 ± 5% sched_debug.cfs_rq:/.runnable_avg.stddev
232.74 ± 9% -24.3% 176.20 ± 3% sched_debug.cfs_rq:/.util_avg.avg
238.62 ± 5% -20.4% 189.97 ± 5% sched_debug.cfs_rq:/.util_avg.stddev
134.97 ± 20% -46.6% 72.12 ± 9% sched_debug.cfs_rq:/.util_est.avg
211.62 ± 7% -30.9% 146.18 ± 3% sched_debug.cfs_rq:/.util_est.stddev
85860 ± 4% +20.2% 103165 ± 4% sched_debug.cpu.avg_idle.stddev
2360 ± 15% -32.2% 1600 ± 6% sched_debug.cpu.curr->pid.avg
4127 ± 5% -11.4% 3655 ± 2% sched_debug.cpu.curr->pid.stddev
0.18 ± 13% -30.8% 0.13 ± 5% sched_debug.cpu.nr_running.avg
0.35 ± 3% -11.4% 0.31 sched_debug.cpu.nr_running.stddev
4008 +10.9% 4445 sched_debug.cpu.nr_switches.avg
0.25 ± 9% +22.4% 0.31 ± 2% sched_debug.cpu.nr_uninterruptible.avg
234.42 +10.7% 259.59 perf-stat.i.MPKI
1.038e+09 -2.6% 1.012e+09 perf-stat.i.branch-instructions
0.45 +0.0 0.47 perf-stat.i.branch-miss-rate%
1.192e+09 +4.1% 1.241e+09 perf-stat.i.cache-misses
1.307e+09 +3.2% 1.349e+09 perf-stat.i.cache-references
4550 +15.1% 5238 perf-stat.i.context-switches
37.73 -11.3% 33.45 perf-stat.i.cpi
2.282e+11 -10.6% 2.041e+11 perf-stat.i.cpu-cycles
395.06 -4.0% 379.37 perf-stat.i.cycles-between-cache-misses
0.08 -0.0 0.07 perf-stat.i.dTLB-load-miss-rate%
957972 -11.5% 848218 perf-stat.i.dTLB-load-misses
1.142e+09 -2.6% 1.112e+09 perf-stat.i.dTLB-loads
4327055 -3.0% 4195482 perf-stat.i.dTLB-store-misses
9.982e+08 -3.2% 9.66e+08 perf-stat.i.dTLB-stores
5.057e+09 -2.7% 4.92e+09 perf-stat.i.instructions
0.08 +4.6% 0.09 perf-stat.i.ipc
1.02 -10.5% 0.91 perf-stat.i.metric.GHz
245.40 -1.7% 241.21 perf-stat.i.metric.K/sec
19.83 -1.0% 19.63 perf-stat.i.metric.M/sec
81125 -3.9% 77994 perf-stat.i.minor-faults
9961580 -4.6% 9501596 perf-stat.i.node-loads
81127 -3.9% 77996 perf-stat.i.page-faults
239.06 +6.3% 254.04 perf-stat.overall.MPKI
0.41 +0.0 0.42 perf-stat.overall.branch-miss-rate%
45.33 -7.8% 41.81 perf-stat.overall.cpi
189.64 -13.2% 164.59 perf-stat.overall.cycles-between-cache-misses
0.09 -0.0 0.08 perf-stat.overall.dTLB-load-miss-rate%
0.02 +8.4% 0.02 perf-stat.overall.ipc
3930 +1.4% 3987 perf-stat.overall.path-length
1.044e+09 -2.5% 1.017e+09 perf-stat.ps.branch-instructions
1.215e+09 +3.5% 1.257e+09 perf-stat.ps.cache-misses
1.329e+09 +2.7% 1.365e+09 perf-stat.ps.cache-references
4559 +14.8% 5231 perf-stat.ps.context-switches
2.304e+11 -10.2% 2.069e+11 perf-stat.ps.cpu-cycles
977167 -11.5% 864759 perf-stat.ps.dTLB-load-misses
1.146e+09 -2.5% 1.117e+09 perf-stat.ps.dTLB-loads
4393673 -3.1% 4258515 perf-stat.ps.dTLB-store-misses
1.006e+09 -3.2% 9.745e+08 perf-stat.ps.dTLB-stores
5.082e+09 -2.6% 4.948e+09 perf-stat.ps.instructions
82026 -3.8% 78940 perf-stat.ps.minor-faults
10075965 -4.4% 9629094 perf-stat.ps.node-loads
82028 -3.8% 78942 perf-stat.ps.page-faults
1.594e+12 -2.5% 1.555e+12 perf-stat.total.instructions
82.91 -2.8 80.13 perf-profile.calltrace.cycles-pp.clear_page_erms.clear_huge_page.__do_huge_pmd_anonymous_page.__handle_mm_fault.handle_mm_fault
94.27 -1.2 93.10 perf-profile.calltrace.cycles-pp.testcase
92.74 -1.2 91.59 perf-profile.calltrace.cycles-pp.__handle_mm_fault.handle_mm_fault.do_user_addr_fault.exc_page_fault.asm_exc_page_fault
92.76 -1.1 91.62 perf-profile.calltrace.cycles-pp.handle_mm_fault.do_user_addr_fault.exc_page_fault.asm_exc_page_fault.testcase
92.82 -1.1 91.69 perf-profile.calltrace.cycles-pp.exc_page_fault.asm_exc_page_fault.testcase
92.82 -1.1 91.69 perf-profile.calltrace.cycles-pp.do_user_addr_fault.exc_page_fault.asm_exc_page_fault.testcase
94.24 -1.0 93.19 perf-profile.calltrace.cycles-pp.asm_exc_page_fault.testcase
90.44 -1.0 89.46 perf-profile.calltrace.cycles-pp.clear_huge_page.__do_huge_pmd_anonymous_page.__handle_mm_fault.handle_mm_fault.do_user_addr_fault
90.76 -0.9 89.81 perf-profile.calltrace.cycles-pp.__do_huge_pmd_anonymous_page.__handle_mm_fault.handle_mm_fault.do_user_addr_fault.exc_page_fault
0.59 ± 2% -0.2 0.34 ± 70% perf-profile.calltrace.cycles-pp.sysvec_apic_timer_interrupt.asm_sysvec_apic_timer_interrupt.clear_page_erms.clear_huge_page.__do_huge_pmd_anonymous_page
1.98 -0.2 1.77 perf-profile.calltrace.cycles-pp.do_huge_pmd_anonymous_page.__handle_mm_fault.handle_mm_fault.do_user_addr_fault.exc_page_fault
1.98 -0.2 1.76 perf-profile.calltrace.cycles-pp.vma_alloc_folio.do_huge_pmd_anonymous_page.__handle_mm_fault.handle_mm_fault.do_user_addr_fault
1.80 -0.2 1.60 perf-profile.calltrace.cycles-pp.prep_compound_page.get_page_from_freelist.__alloc_pages.alloc_pages_mpol.vma_alloc_folio
1.85 -0.2 1.65 perf-profile.calltrace.cycles-pp.__alloc_pages.alloc_pages_mpol.vma_alloc_folio.do_huge_pmd_anonymous_page.__handle_mm_fault
1.84 -0.2 1.64 perf-profile.calltrace.cycles-pp.get_page_from_freelist.__alloc_pages.alloc_pages_mpol.vma_alloc_folio.do_huge_pmd_anonymous_page
1.85 -0.2 1.66 perf-profile.calltrace.cycles-pp.alloc_pages_mpol.vma_alloc_folio.do_huge_pmd_anonymous_page.__handle_mm_fault.handle_mm_fault
0.98 -0.1 0.84 perf-profile.calltrace.cycles-pp.asm_sysvec_apic_timer_interrupt.clear_page_erms.clear_huge_page.__do_huge_pmd_anonymous_page.__handle_mm_fault
1.33 -0.1 1.28 perf-profile.calltrace.cycles-pp.free_unref_page_prepare.free_unref_page.release_pages.tlb_batch_pages_flush.tlb_finish_mmu
1.51 -0.0 1.46 perf-profile.calltrace.cycles-pp.tlb_finish_mmu.unmap_region.do_vmi_align_munmap.do_vmi_munmap.__vm_munmap
1.47 -0.0 1.43 perf-profile.calltrace.cycles-pp.tlb_batch_pages_flush.tlb_finish_mmu.unmap_region.do_vmi_align_munmap.do_vmi_munmap
1.38 -0.0 1.33 perf-profile.calltrace.cycles-pp.free_unref_page.release_pages.tlb_batch_pages_flush.tlb_finish_mmu.unmap_region
1.46 -0.0 1.42 perf-profile.calltrace.cycles-pp.release_pages.tlb_batch_pages_flush.tlb_finish_mmu.unmap_region.do_vmi_align_munmap
1.64 -0.0 1.60 perf-profile.calltrace.cycles-pp.unmap_region.do_vmi_align_munmap.do_vmi_munmap.__vm_munmap.__x64_sys_munmap
0.78 ± 3% +0.2 0.96 perf-profile.calltrace.cycles-pp.intel_idle_xstate.cpuidle_enter_state.cpuidle_enter.cpuidle_idle_call.do_idle
0.64 +0.2 0.82 ± 3% perf-profile.calltrace.cycles-pp.__hrtimer_run_queues.hrtimer_interrupt.__sysvec_apic_timer_interrupt.sysvec_apic_timer_interrupt.asm_sysvec_apic_timer_interrupt
0.82 ± 3% +0.3 1.12 ± 4% perf-profile.calltrace.cycles-pp.hrtimer_interrupt.__sysvec_apic_timer_interrupt.sysvec_apic_timer_interrupt.asm_sysvec_apic_timer_interrupt.cpuidle_enter_state
0.84 ± 3% +0.3 1.16 ± 4% perf-profile.calltrace.cycles-pp.__sysvec_apic_timer_interrupt.sysvec_apic_timer_interrupt.asm_sysvec_apic_timer_interrupt.cpuidle_enter_state.cpuidle_enter
0.26 ±100% +0.4 0.70 perf-profile.calltrace.cycles-pp.intel_idle.cpuidle_enter_state.cpuidle_enter.cpuidle_idle_call.do_idle
1.26 ± 5% +0.5 1.74 ± 5% perf-profile.calltrace.cycles-pp.sysvec_apic_timer_interrupt.asm_sysvec_apic_timer_interrupt.cpuidle_enter_state.cpuidle_enter.cpuidle_idle_call
1.40 ± 4% +0.5 1.92 ± 4% perf-profile.calltrace.cycles-pp.asm_sysvec_apic_timer_interrupt.cpuidle_enter_state.cpuidle_enter.cpuidle_idle_call.do_idle
3.45 +0.9 4.35 perf-profile.calltrace.cycles-pp.__cond_resched.clear_huge_page.__do_huge_pmd_anonymous_page.__handle_mm_fault.handle_mm_fault
2.75 +1.0 3.70 ± 2% perf-profile.calltrace.cycles-pp.cpuidle_enter_state.cpuidle_enter.cpuidle_idle_call.do_idle.cpu_startup_entry
2.88 +1.0 3.87 ± 2% perf-profile.calltrace.cycles-pp.cpuidle_enter.cpuidle_idle_call.do_idle.cpu_startup_entry.start_secondary
3.08 +1.0 4.12 ± 2% perf-profile.calltrace.cycles-pp.cpuidle_idle_call.do_idle.cpu_startup_entry.start_secondary.secondary_startup_64_no_verify
3.12 +1.1 4.18 ± 2% perf-profile.calltrace.cycles-pp.do_idle.cpu_startup_entry.start_secondary.secondary_startup_64_no_verify
3.12 +1.1 4.18 ± 2% perf-profile.calltrace.cycles-pp.cpu_startup_entry.start_secondary.secondary_startup_64_no_verify
3.12 +1.1 4.18 ± 2% perf-profile.calltrace.cycles-pp.start_secondary.secondary_startup_64_no_verify
3.15 +1.1 4.22 ± 2% perf-profile.calltrace.cycles-pp.secondary_startup_64_no_verify
84.16 -2.9 81.22 perf-profile.children.cycles-pp.clear_page_erms
92.75 -1.2 91.60 perf-profile.children.cycles-pp.__handle_mm_fault
92.77 -1.1 91.63 perf-profile.children.cycles-pp.handle_mm_fault
94.98 -1.1 93.84 perf-profile.children.cycles-pp.testcase
92.83 -1.1 91.70 perf-profile.children.cycles-pp.do_user_addr_fault
92.83 -1.1 91.70 perf-profile.children.cycles-pp.exc_page_fault
93.55 -1.1 92.47 perf-profile.children.cycles-pp.asm_exc_page_fault
90.13 -1.0 89.13 perf-profile.children.cycles-pp.clear_huge_page
90.76 -0.9 89.81 perf-profile.children.cycles-pp.__do_huge_pmd_anonymous_page
1.98 -0.2 1.77 perf-profile.children.cycles-pp.do_huge_pmd_anonymous_page
1.98 -0.2 1.76 perf-profile.children.cycles-pp.vma_alloc_folio
1.80 -0.2 1.60 perf-profile.children.cycles-pp.prep_compound_page
1.88 -0.2 1.69 perf-profile.children.cycles-pp.get_page_from_freelist
1.94 -0.2 1.76 perf-profile.children.cycles-pp.__alloc_pages
1.94 -0.2 1.76 perf-profile.children.cycles-pp.alloc_pages_mpol
0.38 -0.1 0.25 ± 2% perf-profile.children.cycles-pp.asm_sysvec_call_function
1.38 -0.0 1.33 perf-profile.children.cycles-pp.free_unref_page_prepare
1.51 -0.0 1.46 perf-profile.children.cycles-pp.tlb_finish_mmu
1.39 -0.0 1.34 perf-profile.children.cycles-pp.free_unref_page
1.47 -0.0 1.42 perf-profile.children.cycles-pp.release_pages
1.47 -0.0 1.43 perf-profile.children.cycles-pp.tlb_batch_pages_flush
1.64 -0.0 1.60 perf-profile.children.cycles-pp.unmap_region
0.21 ± 3% -0.0 0.18 ± 3% perf-profile.children.cycles-pp.task_tick_fair
0.12 ± 3% -0.0 0.10 ± 4% perf-profile.children.cycles-pp.folio_prep_large_rmappable
0.09 ± 4% +0.0 0.10 ± 4% perf-profile.children.cycles-pp.free_tail_page_prepare
0.17 ± 2% +0.0 0.19 ± 3% perf-profile.children.cycles-pp.pte_alloc_one
0.17 ± 4% +0.0 0.19 perf-profile.children.cycles-pp.perf_rotate_context
0.06 ± 11% +0.0 0.08 ± 8% perf-profile.children.cycles-pp._raw_spin_lock_irq
0.06 ± 7% +0.0 0.08 ± 5% perf-profile.children.cycles-pp.update_sg_lb_stats
0.06 ± 8% +0.0 0.08 ± 6% perf-profile.children.cycles-pp.clockevents_program_event
0.12 ± 4% +0.0 0.14 ± 3% perf-profile.children.cycles-pp.__intel_pmu_enable_all
0.16 ± 4% +0.0 0.18 ± 3% perf-profile.children.cycles-pp.native_irq_return_iret
0.08 ± 4% +0.0 0.10 ± 4% perf-profile.children.cycles-pp.find_busiest_group
0.08 ± 6% +0.0 0.10 ± 4% perf-profile.children.cycles-pp.update_sd_lb_stats
0.05 +0.0 0.08 ± 8% perf-profile.children.cycles-pp.__schedule
0.11 ± 3% +0.0 0.14 ± 3% perf-profile.children.cycles-pp.load_balance
0.17 ± 8% +0.0 0.21 ± 3% perf-profile.children.cycles-pp.menu_select
0.01 ±223% +0.0 0.06 ± 6% perf-profile.children.cycles-pp.native_sched_clock
0.00 +0.1 0.05 perf-profile.children.cycles-pp.irqtime_account_irq
0.00 +0.1 0.05 perf-profile.children.cycles-pp.run_rebalance_domains
0.00 +0.1 0.05 perf-profile.children.cycles-pp.rwsem_optimistic_spin
0.00 +0.1 0.05 perf-profile.children.cycles-pp.update_blocked_averages
0.00 +0.1 0.05 ± 7% perf-profile.children.cycles-pp.__mmap
0.00 +0.1 0.05 ± 7% perf-profile.children.cycles-pp.vm_mmap_pgoff
0.00 +0.1 0.06 ± 8% perf-profile.children.cycles-pp.lapic_next_deadline
0.11 ± 4% +0.1 0.17 ± 4% perf-profile.children.cycles-pp.rebalance_domains
0.00 +0.1 0.06 ± 11% perf-profile.children.cycles-pp._raw_spin_trylock
0.00 +0.1 0.06 ± 6% perf-profile.children.cycles-pp.down_write_killable
0.00 +0.1 0.06 ± 6% perf-profile.children.cycles-pp.read_tsc
0.00 +0.1 0.06 ± 6% perf-profile.children.cycles-pp.schedule
0.79 ± 4% +0.1 0.87 ± 3% perf-profile.children.cycles-pp.tick_nohz_highres_handler
0.19 ± 3% +0.1 0.27 ± 3% perf-profile.children.cycles-pp.__do_softirq
0.21 ± 3% +0.1 0.31 ± 3% perf-profile.children.cycles-pp.irq_exit_rcu
0.00 +0.1 0.10 ± 3% perf-profile.children.cycles-pp.rwsem_down_write_slowpath
1.17 +0.1 1.29 perf-profile.children.cycles-pp.__hrtimer_run_queues
0.79 ± 3% +0.2 0.97 perf-profile.children.cycles-pp.intel_idle_xstate
0.50 ± 3% +0.2 0.70 perf-profile.children.cycles-pp.intel_idle
1.45 ± 2% +0.2 1.67 ± 3% perf-profile.children.cycles-pp.hrtimer_interrupt
1.48 ± 2% +0.2 1.73 ± 3% perf-profile.children.cycles-pp.__sysvec_apic_timer_interrupt
2.50 ± 2% +0.4 2.89 ± 3% perf-profile.children.cycles-pp.asm_sysvec_apic_timer_interrupt
1.94 ± 4% +0.4 2.35 ± 4% perf-profile.children.cycles-pp.sysvec_apic_timer_interrupt
3.60 +0.9 4.50 perf-profile.children.cycles-pp.__cond_resched
2.91 +1.0 3.90 ± 2% perf-profile.children.cycles-pp.cpuidle_enter
2.90 +1.0 3.90 ± 2% perf-profile.children.cycles-pp.cpuidle_enter_state
3.11 +1.0 4.15 ± 2% perf-profile.children.cycles-pp.cpuidle_idle_call
3.12 +1.1 4.18 ± 2% perf-profile.children.cycles-pp.start_secondary
3.15 +1.1 4.22 ± 2% perf-profile.children.cycles-pp.cpu_startup_entry
3.15 +1.1 4.22 ± 2% perf-profile.children.cycles-pp.do_idle
3.15 +1.1 4.22 ± 2% perf-profile.children.cycles-pp.secondary_startup_64_no_verify
83.46 -2.9 80.61 perf-profile.self.cycles-pp.clear_page_erms
1.79 -0.2 1.59 perf-profile.self.cycles-pp.prep_compound_page
1.30 -0.1 1.25 perf-profile.self.cycles-pp.free_unref_page_prepare
0.12 ± 3% -0.0 0.10 ± 4% perf-profile.self.cycles-pp.folio_prep_large_rmappable
0.05 +0.0 0.06 ± 7% perf-profile.self.cycles-pp.update_sg_lb_stats
0.06 ± 11% +0.0 0.08 ± 8% perf-profile.self.cycles-pp._raw_spin_lock_irq
0.12 ± 4% +0.0 0.14 ± 3% perf-profile.self.cycles-pp.__intel_pmu_enable_all
0.16 ± 4% +0.0 0.18 ± 3% perf-profile.self.cycles-pp.native_irq_return_iret
0.07 ± 6% +0.0 0.10 ± 5% perf-profile.self.cycles-pp.menu_select
0.20 ± 3% +0.0 0.25 ± 2% perf-profile.self.cycles-pp.cpuidle_enter_state
0.00 +0.1 0.06 ± 8% perf-profile.self.cycles-pp.lapic_next_deadline
0.00 +0.1 0.06 ± 6% perf-profile.self.cycles-pp.native_sched_clock
0.00 +0.1 0.06 ± 13% perf-profile.self.cycles-pp._raw_spin_trylock
0.00 +0.1 0.06 ± 9% perf-profile.self.cycles-pp.read_tsc
0.78 ± 3% +0.2 0.97 perf-profile.self.cycles-pp.intel_idle_xstate
0.50 ± 3% +0.2 0.70 perf-profile.self.cycles-pp.intel_idle
2.93 +0.9 3.83 perf-profile.self.cycles-pp.__cond_resched
2.59 +1.0 3.61 perf-profile.self.cycles-pp.clear_huge_page
Disclaimer:
Results have been estimated based on internal Intel analysis and are provided
for informational purposes only. Any difference in system hardware or software
design or configuration may affect actual performance.
--
0-DAY CI Kernel Test Service
https://github.com/intel/lkp-tests/wiki
Powered by blists - more mailing lists