[<prev] [next>] [<thread-prev] [day] [month] [year] [list]
Message-ID: <fd1adcc1-ed9b-44f8-b884-e7d0aa5e40b2@arm.com>
Date: Thu, 8 Feb 2024 15:40:33 +0000
From: Ryan Roberts <ryan.roberts@....com>
To: kernel test robot <oliver.sang@...el.com>
Cc: oe-lkp@...ts.linux.dev, lkp@...el.com, linux-kernel@...r.kernel.org,
Andrew Morton <akpm@...ux-foundation.org>, Yang Shi <shy828301@...il.com>,
Matthew Wilcox <willy@...radead.org>, Rik van Riel <riel@...riel.com>,
linux-mm@...ck.org, ying.huang@...el.com, feng.tang@...el.com,
fengwei.yin@...el.com
Subject: Re: [linus:master] [mm] 96204e1531: will-it-scale.per_thread_ops
-3.9% regression
On 08/02/2024 07:36, kernel test robot wrote:
>
>
> Hello,
>
> kernel test robot noticed a -3.9% regression of will-it-scale.per_thread_ops on:
Hi,
I've tried reproducing this on x86_64, but I see fluctuations of +/- 5% between
runs without changing the kernel, when running the page_fault1 test. So I'm not
sure a -3.9% regression is real?
The fingered change is actually a fix for a functional bug, so I'm not sure we
should be preferring performance over correctness anyway :)
I'm not planning to do anything further here, unless someone shouts that there
is a real problem.
>
>
> commit: 96204e15310c218fd9355bdcacd02fed1d18070e ("mm: thp_get_unmapped_area must honour topdown preference")
> https://git.kernel.org/cgit/linux/kernel/git/torvalds/linux.git master
>
> testcase: will-it-scale
> test machine: 224 threads 2 sockets Intel(R) Xeon(R) Platinum 8480CTDX (Sapphire Rapids) with 512G memory
> parameters:
>
> test: page_fault1
> cpufreq_governor: performance
>
>
>
>
> If you fix the issue in a separate patch/commit (i.e. not just a new version of
> the same patch/commit), kindly add following tags
> | Reported-by: kernel test robot <oliver.sang@...el.com>
> | Closes: https://lore.kernel.org/oe-lkp/202402081435.c3faa4b4-oliver.sang@intel.com
>
>
> Details are as below:
> -------------------------------------------------------------------------------------------------->
>
>
> The kernel config and materials to reproduce are available at:
> https://download.01.org/0day-ci/archive/20240208/202402081435.c3faa4b4-oliver.sang@intel.com
>
> =========================================================================================
> compiler/cpufreq_governor/kconfig/rootfs/tbox_group/test/testcase:
> gcc-12/performance/x86_64-rhel-8.3/debian-11.1-x86_64-20220510.cgz/lkp-spr-2sp4/page_fault1/will-it-scale
>
> commit:
> 4ef9ad19e1 ("mm: huge_memory: don't force huge page alignment on 32 bit")
> 96204e1531 ("mm: thp_get_unmapped_area must honour topdown preference")
>
I don't really understand the below table. Is a positive or negative change
better? There appear to be tests with much more fluctuation than the reported
page_fault1 test. Are these real?
Thanks,
Ryan
> 4ef9ad19e17676b9 96204e15310c218fd9355bdcacd
> ---------------- ---------------------------
> %stddev %change %stddev
> \ | \
> 32.93 -3.5 29.43 mpstat.cpu.all.sys%
> 77.89 � 2% -10.2% 69.98 vmstat.procs.r
> 4678 +14.0% 5332 vmstat.system.cs
> 251960 -3.4% 243306 vmstat.system.in
> 1049 -10.2% 943.00 turbostat.Avg_MHz
> 36.22 -3.7 32.54 turbostat.Busy%
> 29.10 +1.8 30.87 turbostat.C1E%
> 0.35 +15.7% 0.40 turbostat.CPU%c6
> 565.62 -2.3% 552.53 turbostat.PkgWatt
> 123847 -18.4% 101096 meminfo.Active
> 123657 -18.3% 101000 meminfo.Active(anon)
> 8127530 +10.3% 8966990 meminfo.AnonHugePages
> 8298705 +10.3% 9156780 meminfo.AnonPages
> 8417195 +10.1% 9267954 meminfo.Inactive
> 8416461 +10.1% 9267742 meminfo.Inactive(anon)
> 138121 -16.9% 114720 meminfo.Shmem
> 4852175 � 2% +9.2% 5296708 numa-meminfo.node0.AnonHugePages
> 4941270 � 2% +9.0% 5387476 numa-meminfo.node0.AnonPages
> 121571 -18.9% 98652 numa-meminfo.node1.Active
> 121528 -18.9% 98620 numa-meminfo.node1.Active(anon)
> 3270590 +12.4% 3674514 numa-meminfo.node1.AnonHugePages
> 3352686 +12.6% 3773824 numa-meminfo.node1.AnonPages
> 3395079 +12.4% 3815833 numa-meminfo.node1.Inactive
> 3394971 +12.4% 3815762 numa-meminfo.node1.Inactive(anon)
> 125041 -18.4% 102036 numa-meminfo.node1.Shmem
> 1219136 � 2% +10.3% 1344427 numa-vmstat.node0.nr_anon_pages
> 2337 +10.4% 2581 numa-vmstat.node0.nr_anon_transparent_hugepages
> 1237208 � 2% +10.1% 1362084 numa-vmstat.node0.nr_inactive_anon
> 1236994 � 2% +10.1% 1362125 numa-vmstat.node0.nr_zone_inactive_anon
> 30361 -18.8% 24656 numa-vmstat.node1.nr_active_anon
> 828422 +13.5% 940280 numa-vmstat.node1.nr_anon_pages
> 1577 +13.3% 1788 numa-vmstat.node1.nr_anon_transparent_hugepages
> 838873 +13.3% 950629 � 2% numa-vmstat.node1.nr_inactive_anon
> 31244 -18.4% 25510 numa-vmstat.node1.nr_shmem
> 30361 -18.8% 24656 numa-vmstat.node1.nr_zone_active_anon
> 838696 +13.4% 950674 � 2% numa-vmstat.node1.nr_zone_inactive_anon
> 30907 -18.4% 25233 proc-vmstat.nr_active_anon
> 2068862 +11.0% 2295458 proc-vmstat.nr_anon_pages
> 3957 +11.0% 4390 proc-vmstat.nr_anon_transparent_hugepages
> 2099316 +10.6% 2322543 proc-vmstat.nr_inactive_anon
> 7379 +4.4% 7701 proc-vmstat.nr_page_table_pages
> 34526 -17.0% 28665 proc-vmstat.nr_shmem
> 30907 -18.4% 25233 proc-vmstat.nr_zone_active_anon
> 2099307 +10.6% 2322537 proc-vmstat.nr_zone_inactive_anon
> 26039 � 10% -16.4% 21774 � 6% proc-vmstat.numa_hint_faults_local
> 51127959 -3.7% 49228917 proc-vmstat.numa_hit
> 50891640 -3.7% 48994322 proc-vmstat.numa_local
> 32315 -21.8% 25269 proc-vmstat.pgactivate
> 1.265e+10 -3.8% 1.217e+10 proc-vmstat.pgalloc_normal
> 25947640 -3.6% 25012103 proc-vmstat.pgfault
> 1.265e+10 -3.8% 1.217e+10 proc-vmstat.pgfree
> 24650837 -3.8% 23717158 proc-vmstat.thp_fault_alloc
> 44754589 -15.5% 37839290 will-it-scale.112.threads
> 74.20 +11.2% 82.49 will-it-scale.112.threads_idle
> 36349325 -6.0% 34182802 will-it-scale.168.threads
> 72.79 +11.7% 81.28 will-it-scale.168.threads_idle
> 40356275 -5.6% 38083547 will-it-scale.224.threads
> 66.97 +19.5% 80.06 will-it-scale.224.threads_idle
> 28146956 -15.5% 23782499 will-it-scale.56.threads
> 78.16 +8.4% 84.70 will-it-scale.56.threads_idle
> 795871 -3.9% 764972 will-it-scale.per_thread_ops
> 67103 -4.9% 63834 will-it-scale.time.involuntary_context_switches
> 17508243 � 4% +6.9% 18724480 will-it-scale.time.maximum_resident_set_size
> 22382078 -3.6% 21577203 will-it-scale.time.minor_page_faults
> 7050 -7.3% 6534 will-it-scale.time.percent_of_cpu_this_job_got
> 20998 -7.4% 19448 will-it-scale.time.system_time
> 1075 -3.3% 1040 will-it-scale.time.user_time
> 236326 +40.3% 331522 will-it-scale.time.voluntary_context_switches
> 4.056e+08 -3.9% 3.899e+08 will-it-scale.workload
> 5880710 -9.8% 5305863 sched_debug.cfs_rq:/.avg_vruntime.avg
> 10300738 � 2% -11.0% 9163982 � 2% sched_debug.cfs_rq:/.avg_vruntime.max
> 1772532 � 6% -15.5% 1498526 � 4% sched_debug.cfs_rq:/.avg_vruntime.min
> 2066050 -12.1% 1817081 sched_debug.cfs_rq:/.avg_vruntime.stddev
> 321.14 � 36% -63.9% 115.93 � 57% sched_debug.cfs_rq:/.load_avg.avg
> 1945 � 20% -47.8% 1015 � 50% sched_debug.cfs_rq:/.load_avg.stddev
> 5880710 -9.8% 5305863 sched_debug.cfs_rq:/.min_vruntime.avg
> 10300738 � 2% -11.0% 9163982 � 2% sched_debug.cfs_rq:/.min_vruntime.max
> 1772532 � 6% -15.5% 1498526 � 4% sched_debug.cfs_rq:/.min_vruntime.min
> 2066050 -12.1% 1817081 sched_debug.cfs_rq:/.min_vruntime.stddev
> 233.57 � 8% -24.3% 176.83 � 3% sched_debug.cfs_rq:/.runnable_avg.avg
> 958.36 � 4% -9.2% 870.11 � 3% sched_debug.cfs_rq:/.runnable_avg.max
> 240.06 � 5% -20.4% 190.99 � 5% sched_debug.cfs_rq:/.runnable_avg.stddev
> 232.74 � 9% -24.3% 176.20 � 3% sched_debug.cfs_rq:/.util_avg.avg
> 238.62 � 5% -20.4% 189.97 � 5% sched_debug.cfs_rq:/.util_avg.stddev
> 134.97 � 20% -46.6% 72.12 � 9% sched_debug.cfs_rq:/.util_est.avg
> 211.62 � 7% -30.9% 146.18 � 3% sched_debug.cfs_rq:/.util_est.stddev
> 85860 � 4% +20.2% 103165 � 4% sched_debug.cpu.avg_idle.stddev
> 2360 � 15% -32.2% 1600 � 6% sched_debug.cpu.curr->pid.avg
> 4127 � 5% -11.4% 3655 � 2% sched_debug.cpu.curr->pid.stddev
> 0.18 � 13% -30.8% 0.13 � 5% sched_debug.cpu.nr_running.avg
> 0.35 � 3% -11.4% 0.31 sched_debug.cpu.nr_running.stddev
> 4008 +10.9% 4445 sched_debug.cpu.nr_switches.avg
> 0.25 � 9% +22.4% 0.31 � 2% sched_debug.cpu.nr_uninterruptible.avg
> 234.42 +10.7% 259.59 perf-stat.i.MPKI
> 1.038e+09 -2.6% 1.012e+09 perf-stat.i.branch-instructions
> 0.45 +0.0 0.47 perf-stat.i.branch-miss-rate%
> 1.192e+09 +4.1% 1.241e+09 perf-stat.i.cache-misses
> 1.307e+09 +3.2% 1.349e+09 perf-stat.i.cache-references
> 4550 +15.1% 5238 perf-stat.i.context-switches
> 37.73 -11.3% 33.45 perf-stat.i.cpi
> 2.282e+11 -10.6% 2.041e+11 perf-stat.i.cpu-cycles
> 395.06 -4.0% 379.37 perf-stat.i.cycles-between-cache-misses
> 0.08 -0.0 0.07 perf-stat.i.dTLB-load-miss-rate%
> 957972 -11.5% 848218 perf-stat.i.dTLB-load-misses
> 1.142e+09 -2.6% 1.112e+09 perf-stat.i.dTLB-loads
> 4327055 -3.0% 4195482 perf-stat.i.dTLB-store-misses
> 9.982e+08 -3.2% 9.66e+08 perf-stat.i.dTLB-stores
> 5.057e+09 -2.7% 4.92e+09 perf-stat.i.instructions
> 0.08 +4.6% 0.09 perf-stat.i.ipc
> 1.02 -10.5% 0.91 perf-stat.i.metric.GHz
> 245.40 -1.7% 241.21 perf-stat.i.metric.K/sec
> 19.83 -1.0% 19.63 perf-stat.i.metric.M/sec
> 81125 -3.9% 77994 perf-stat.i.minor-faults
> 9961580 -4.6% 9501596 perf-stat.i.node-loads
> 81127 -3.9% 77996 perf-stat.i.page-faults
> 239.06 +6.3% 254.04 perf-stat.overall.MPKI
> 0.41 +0.0 0.42 perf-stat.overall.branch-miss-rate%
> 45.33 -7.8% 41.81 perf-stat.overall.cpi
> 189.64 -13.2% 164.59 perf-stat.overall.cycles-between-cache-misses
> 0.09 -0.0 0.08 perf-stat.overall.dTLB-load-miss-rate%
> 0.02 +8.4% 0.02 perf-stat.overall.ipc
> 3930 +1.4% 3987 perf-stat.overall.path-length
> 1.044e+09 -2.5% 1.017e+09 perf-stat.ps.branch-instructions
> 1.215e+09 +3.5% 1.257e+09 perf-stat.ps.cache-misses
> 1.329e+09 +2.7% 1.365e+09 perf-stat.ps.cache-references
> 4559 +14.8% 5231 perf-stat.ps.context-switches
> 2.304e+11 -10.2% 2.069e+11 perf-stat.ps.cpu-cycles
> 977167 -11.5% 864759 perf-stat.ps.dTLB-load-misses
> 1.146e+09 -2.5% 1.117e+09 perf-stat.ps.dTLB-loads
> 4393673 -3.1% 4258515 perf-stat.ps.dTLB-store-misses
> 1.006e+09 -3.2% 9.745e+08 perf-stat.ps.dTLB-stores
> 5.082e+09 -2.6% 4.948e+09 perf-stat.ps.instructions
> 82026 -3.8% 78940 perf-stat.ps.minor-faults
> 10075965 -4.4% 9629094 perf-stat.ps.node-loads
> 82028 -3.8% 78942 perf-stat.ps.page-faults
> 1.594e+12 -2.5% 1.555e+12 perf-stat.total.instructions
> 82.91 -2.8 80.13 perf-profile.calltrace.cycles-pp.clear_page_erms.clear_huge_page.__do_huge_pmd_anonymous_page.__handle_mm_fault.handle_mm_fault
> 94.27 -1.2 93.10 perf-profile.calltrace.cycles-pp.testcase
> 92.74 -1.2 91.59 perf-profile.calltrace.cycles-pp.__handle_mm_fault.handle_mm_fault.do_user_addr_fault.exc_page_fault.asm_exc_page_fault
> 92.76 -1.1 91.62 perf-profile.calltrace.cycles-pp.handle_mm_fault.do_user_addr_fault.exc_page_fault.asm_exc_page_fault.testcase
> 92.82 -1.1 91.69 perf-profile.calltrace.cycles-pp.exc_page_fault.asm_exc_page_fault.testcase
> 92.82 -1.1 91.69 perf-profile.calltrace.cycles-pp.do_user_addr_fault.exc_page_fault.asm_exc_page_fault.testcase
> 94.24 -1.0 93.19 perf-profile.calltrace.cycles-pp.asm_exc_page_fault.testcase
> 90.44 -1.0 89.46 perf-profile.calltrace.cycles-pp.clear_huge_page.__do_huge_pmd_anonymous_page.__handle_mm_fault.handle_mm_fault.do_user_addr_fault
> 90.76 -0.9 89.81 perf-profile.calltrace.cycles-pp.__do_huge_pmd_anonymous_page.__handle_mm_fault.handle_mm_fault.do_user_addr_fault.exc_page_fault
> 0.59 � 2% -0.2 0.34 � 70% perf-profile.calltrace.cycles-pp.sysvec_apic_timer_interrupt.asm_sysvec_apic_timer_interrupt.clear_page_erms.clear_huge_page.__do_huge_pmd_anonymous_page
> 1.98 -0.2 1.77 perf-profile.calltrace.cycles-pp.do_huge_pmd_anonymous_page.__handle_mm_fault.handle_mm_fault.do_user_addr_fault.exc_page_fault
> 1.98 -0.2 1.76 perf-profile.calltrace.cycles-pp.vma_alloc_folio.do_huge_pmd_anonymous_page.__handle_mm_fault.handle_mm_fault.do_user_addr_fault
> 1.80 -0.2 1.60 perf-profile.calltrace.cycles-pp.prep_compound_page.get_page_from_freelist.__alloc_pages.alloc_pages_mpol.vma_alloc_folio
> 1.85 -0.2 1.65 perf-profile.calltrace.cycles-pp.__alloc_pages.alloc_pages_mpol.vma_alloc_folio.do_huge_pmd_anonymous_page.__handle_mm_fault
> 1.84 -0.2 1.64 perf-profile.calltrace.cycles-pp.get_page_from_freelist.__alloc_pages.alloc_pages_mpol.vma_alloc_folio.do_huge_pmd_anonymous_page
> 1.85 -0.2 1.66 perf-profile.calltrace.cycles-pp.alloc_pages_mpol.vma_alloc_folio.do_huge_pmd_anonymous_page.__handle_mm_fault.handle_mm_fault
> 0.98 -0.1 0.84 perf-profile.calltrace.cycles-pp.asm_sysvec_apic_timer_interrupt.clear_page_erms.clear_huge_page.__do_huge_pmd_anonymous_page.__handle_mm_fault
> 1.33 -0.1 1.28 perf-profile.calltrace.cycles-pp.free_unref_page_prepare.free_unref_page.release_pages.tlb_batch_pages_flush.tlb_finish_mmu
> 1.51 -0.0 1.46 perf-profile.calltrace.cycles-pp.tlb_finish_mmu.unmap_region.do_vmi_align_munmap.do_vmi_munmap.__vm_munmap
> 1.47 -0.0 1.43 perf-profile.calltrace.cycles-pp.tlb_batch_pages_flush.tlb_finish_mmu.unmap_region.do_vmi_align_munmap.do_vmi_munmap
> 1.38 -0.0 1.33 perf-profile.calltrace.cycles-pp.free_unref_page.release_pages.tlb_batch_pages_flush.tlb_finish_mmu.unmap_region
> 1.46 -0.0 1.42 perf-profile.calltrace.cycles-pp.release_pages.tlb_batch_pages_flush.tlb_finish_mmu.unmap_region.do_vmi_align_munmap
> 1.64 -0.0 1.60 perf-profile.calltrace.cycles-pp.unmap_region.do_vmi_align_munmap.do_vmi_munmap.__vm_munmap.__x64_sys_munmap
> 0.78 � 3% +0.2 0.96 perf-profile.calltrace.cycles-pp.intel_idle_xstate.cpuidle_enter_state.cpuidle_enter.cpuidle_idle_call.do_idle
> 0.64 +0.2 0.82 � 3% perf-profile.calltrace.cycles-pp.__hrtimer_run_queues.hrtimer_interrupt.__sysvec_apic_timer_interrupt.sysvec_apic_timer_interrupt.asm_sysvec_apic_timer_interrupt
> 0.82 � 3% +0.3 1.12 � 4% perf-profile.calltrace.cycles-pp.hrtimer_interrupt.__sysvec_apic_timer_interrupt.sysvec_apic_timer_interrupt.asm_sysvec_apic_timer_interrupt.cpuidle_enter_state
> 0.84 � 3% +0.3 1.16 � 4% perf-profile.calltrace.cycles-pp.__sysvec_apic_timer_interrupt.sysvec_apic_timer_interrupt.asm_sysvec_apic_timer_interrupt.cpuidle_enter_state.cpuidle_enter
> 0.26 �100% +0.4 0.70 perf-profile.calltrace.cycles-pp.intel_idle.cpuidle_enter_state.cpuidle_enter.cpuidle_idle_call.do_idle
> 1.26 � 5% +0.5 1.74 � 5% perf-profile.calltrace.cycles-pp.sysvec_apic_timer_interrupt.asm_sysvec_apic_timer_interrupt.cpuidle_enter_state.cpuidle_enter.cpuidle_idle_call
> 1.40 � 4% +0.5 1.92 � 4% perf-profile.calltrace.cycles-pp.asm_sysvec_apic_timer_interrupt.cpuidle_enter_state.cpuidle_enter.cpuidle_idle_call.do_idle
> 3.45 +0.9 4.35 perf-profile.calltrace.cycles-pp.__cond_resched.clear_huge_page.__do_huge_pmd_anonymous_page.__handle_mm_fault.handle_mm_fault
> 2.75 +1.0 3.70 � 2% perf-profile.calltrace.cycles-pp.cpuidle_enter_state.cpuidle_enter.cpuidle_idle_call.do_idle.cpu_startup_entry
> 2.88 +1.0 3.87 � 2% perf-profile.calltrace.cycles-pp.cpuidle_enter.cpuidle_idle_call.do_idle.cpu_startup_entry.start_secondary
> 3.08 +1.0 4.12 � 2% perf-profile.calltrace.cycles-pp.cpuidle_idle_call.do_idle.cpu_startup_entry.start_secondary.secondary_startup_64_no_verify
> 3.12 +1.1 4.18 � 2% perf-profile.calltrace.cycles-pp.do_idle.cpu_startup_entry.start_secondary.secondary_startup_64_no_verify
> 3.12 +1.1 4.18 � 2% perf-profile.calltrace.cycles-pp.cpu_startup_entry.start_secondary.secondary_startup_64_no_verify
> 3.12 +1.1 4.18 � 2% perf-profile.calltrace.cycles-pp.start_secondary.secondary_startup_64_no_verify
> 3.15 +1.1 4.22 � 2% perf-profile.calltrace.cycles-pp.secondary_startup_64_no_verify
> 84.16 -2.9 81.22 perf-profile.children.cycles-pp.clear_page_erms
> 92.75 -1.2 91.60 perf-profile.children.cycles-pp.__handle_mm_fault
> 92.77 -1.1 91.63 perf-profile.children.cycles-pp.handle_mm_fault
> 94.98 -1.1 93.84 perf-profile.children.cycles-pp.testcase
> 92.83 -1.1 91.70 perf-profile.children.cycles-pp.do_user_addr_fault
> 92.83 -1.1 91.70 perf-profile.children.cycles-pp.exc_page_fault
> 93.55 -1.1 92.47 perf-profile.children.cycles-pp.asm_exc_page_fault
> 90.13 -1.0 89.13 perf-profile.children.cycles-pp.clear_huge_page
> 90.76 -0.9 89.81 perf-profile.children.cycles-pp.__do_huge_pmd_anonymous_page
> 1.98 -0.2 1.77 perf-profile.children.cycles-pp.do_huge_pmd_anonymous_page
> 1.98 -0.2 1.76 perf-profile.children.cycles-pp.vma_alloc_folio
> 1.80 -0.2 1.60 perf-profile.children.cycles-pp.prep_compound_page
> 1.88 -0.2 1.69 perf-profile.children.cycles-pp.get_page_from_freelist
> 1.94 -0.2 1.76 perf-profile.children.cycles-pp.__alloc_pages
> 1.94 -0.2 1.76 perf-profile.children.cycles-pp.alloc_pages_mpol
> 0.38 -0.1 0.25 � 2% perf-profile.children.cycles-pp.asm_sysvec_call_function
> 1.38 -0.0 1.33 perf-profile.children.cycles-pp.free_unref_page_prepare
> 1.51 -0.0 1.46 perf-profile.children.cycles-pp.tlb_finish_mmu
> 1.39 -0.0 1.34 perf-profile.children.cycles-pp.free_unref_page
> 1.47 -0.0 1.42 perf-profile.children.cycles-pp.release_pages
> 1.47 -0.0 1.43 perf-profile.children.cycles-pp.tlb_batch_pages_flush
> 1.64 -0.0 1.60 perf-profile.children.cycles-pp.unmap_region
> 0.21 � 3% -0.0 0.18 � 3% perf-profile.children.cycles-pp.task_tick_fair
> 0.12 � 3% -0.0 0.10 � 4% perf-profile.children.cycles-pp.folio_prep_large_rmappable
> 0.09 � 4% +0.0 0.10 � 4% perf-profile.children.cycles-pp.free_tail_page_prepare
> 0.17 � 2% +0.0 0.19 � 3% perf-profile.children.cycles-pp.pte_alloc_one
> 0.17 � 4% +0.0 0.19 perf-profile.children.cycles-pp.perf_rotate_context
> 0.06 � 11% +0.0 0.08 � 8% perf-profile.children.cycles-pp._raw_spin_lock_irq
> 0.06 � 7% +0.0 0.08 � 5% perf-profile.children.cycles-pp.update_sg_lb_stats
> 0.06 � 8% +0.0 0.08 � 6% perf-profile.children.cycles-pp.clockevents_program_event
> 0.12 � 4% +0.0 0.14 � 3% perf-profile.children.cycles-pp.__intel_pmu_enable_all
> 0.16 � 4% +0.0 0.18 � 3% perf-profile.children.cycles-pp.native_irq_return_iret
> 0.08 � 4% +0.0 0.10 � 4% perf-profile.children.cycles-pp.find_busiest_group
> 0.08 � 6% +0.0 0.10 � 4% perf-profile.children.cycles-pp.update_sd_lb_stats
> 0.05 +0.0 0.08 � 8% perf-profile.children.cycles-pp.__schedule
> 0.11 � 3% +0.0 0.14 � 3% perf-profile.children.cycles-pp.load_balance
> 0.17 � 8% +0.0 0.21 � 3% perf-profile.children.cycles-pp.menu_select
> 0.01 �223% +0.0 0.06 � 6% perf-profile.children.cycles-pp.native_sched_clock
> 0.00 +0.1 0.05 perf-profile.children.cycles-pp.irqtime_account_irq
> 0.00 +0.1 0.05 perf-profile.children.cycles-pp.run_rebalance_domains
> 0.00 +0.1 0.05 perf-profile.children.cycles-pp.rwsem_optimistic_spin
> 0.00 +0.1 0.05 perf-profile.children.cycles-pp.update_blocked_averages
> 0.00 +0.1 0.05 � 7% perf-profile.children.cycles-pp.__mmap
> 0.00 +0.1 0.05 � 7% perf-profile.children.cycles-pp.vm_mmap_pgoff
> 0.00 +0.1 0.06 � 8% perf-profile.children.cycles-pp.lapic_next_deadline
> 0.11 � 4% +0.1 0.17 � 4% perf-profile.children.cycles-pp.rebalance_domains
> 0.00 +0.1 0.06 � 11% perf-profile.children.cycles-pp._raw_spin_trylock
> 0.00 +0.1 0.06 � 6% perf-profile.children.cycles-pp.down_write_killable
> 0.00 +0.1 0.06 � 6% perf-profile.children.cycles-pp.read_tsc
> 0.00 +0.1 0.06 � 6% perf-profile.children.cycles-pp.schedule
> 0.79 � 4% +0.1 0.87 � 3% perf-profile.children.cycles-pp.tick_nohz_highres_handler
> 0.19 � 3% +0.1 0.27 � 3% perf-profile.children.cycles-pp.__do_softirq
> 0.21 � 3% +0.1 0.31 � 3% perf-profile.children.cycles-pp.irq_exit_rcu
> 0.00 +0.1 0.10 � 3% perf-profile.children.cycles-pp.rwsem_down_write_slowpath
> 1.17 +0.1 1.29 perf-profile.children.cycles-pp.__hrtimer_run_queues
> 0.79 � 3% +0.2 0.97 perf-profile.children.cycles-pp.intel_idle_xstate
> 0.50 � 3% +0.2 0.70 perf-profile.children.cycles-pp.intel_idle
> 1.45 � 2% +0.2 1.67 � 3% perf-profile.children.cycles-pp.hrtimer_interrupt
> 1.48 � 2% +0.2 1.73 � 3% perf-profile.children.cycles-pp.__sysvec_apic_timer_interrupt
> 2.50 � 2% +0.4 2.89 � 3% perf-profile.children.cycles-pp.asm_sysvec_apic_timer_interrupt
> 1.94 � 4% +0.4 2.35 � 4% perf-profile.children.cycles-pp.sysvec_apic_timer_interrupt
> 3.60 +0.9 4.50 perf-profile.children.cycles-pp.__cond_resched
> 2.91 +1.0 3.90 � 2% perf-profile.children.cycles-pp.cpuidle_enter
> 2.90 +1.0 3.90 � 2% perf-profile.children.cycles-pp.cpuidle_enter_state
> 3.11 +1.0 4.15 � 2% perf-profile.children.cycles-pp.cpuidle_idle_call
> 3.12 +1.1 4.18 � 2% perf-profile.children.cycles-pp.start_secondary
> 3.15 +1.1 4.22 � 2% perf-profile.children.cycles-pp.cpu_startup_entry
> 3.15 +1.1 4.22 � 2% perf-profile.children.cycles-pp.do_idle
> 3.15 +1.1 4.22 � 2% perf-profile.children.cycles-pp.secondary_startup_64_no_verify
> 83.46 -2.9 80.61 perf-profile.self.cycles-pp.clear_page_erms
> 1.79 -0.2 1.59 perf-profile.self.cycles-pp.prep_compound_page
> 1.30 -0.1 1.25 perf-profile.self.cycles-pp.free_unref_page_prepare
> 0.12 � 3% -0.0 0.10 � 4% perf-profile.self.cycles-pp.folio_prep_large_rmappable
> 0.05 +0.0 0.06 � 7% perf-profile.self.cycles-pp.update_sg_lb_stats
> 0.06 � 11% +0.0 0.08 � 8% perf-profile.self.cycles-pp._raw_spin_lock_irq
> 0.12 � 4% +0.0 0.14 � 3% perf-profile.self.cycles-pp.__intel_pmu_enable_all
> 0.16 � 4% +0.0 0.18 � 3% perf-profile.self.cycles-pp.native_irq_return_iret
> 0.07 � 6% +0.0 0.10 � 5% perf-profile.self.cycles-pp.menu_select
> 0.20 � 3% +0.0 0.25 � 2% perf-profile.self.cycles-pp.cpuidle_enter_state
> 0.00 +0.1 0.06 � 8% perf-profile.self.cycles-pp.lapic_next_deadline
> 0.00 +0.1 0.06 � 6% perf-profile.self.cycles-pp.native_sched_clock
> 0.00 +0.1 0.06 � 13% perf-profile.self.cycles-pp._raw_spin_trylock
> 0.00 +0.1 0.06 � 9% perf-profile.self.cycles-pp.read_tsc
> 0.78 � 3% +0.2 0.97 perf-profile.self.cycles-pp.intel_idle_xstate
> 0.50 � 3% +0.2 0.70 perf-profile.self.cycles-pp.intel_idle
> 2.93 +0.9 3.83 perf-profile.self.cycles-pp.__cond_resched
> 2.59 +1.0 3.61 perf-profile.self.cycles-pp.clear_huge_page
>
>
>
>
> Disclaimer:
> Results have been estimated based on internal Intel analysis and are provided
> for informational purposes only. Any difference in system hardware or software
> design or configuration may affect actual performance.
>
>
Powered by blists - more mailing lists