[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <CAOUHufb1qteTQ+rLHBgu5yVQNYPBVrv7wCfEYC85y_pjQQ3EfQ@mail.gmail.com>
Date: Fri, 19 Jul 2024 10:06:46 -0600
From: Yu Zhao <yuzhao@...gle.com>
To: Oliver Sang <oliver.sang@...el.com>
Cc: Janosch Frank <frankja@...ux.ibm.com>, oe-lkp@...ts.linux.dev, lkp@...el.com,
Linux Memory Management List <linux-mm@...ck.org>, Andrew Morton <akpm@...ux-foundation.org>,
Muchun Song <muchun.song@...ux.dev>, David Hildenbrand <david@...hat.com>,
Frank van der Linden <fvdl@...gle.com>, Matthew Wilcox <willy@...radead.org>, Peter Xu <peterx@...hat.com>,
Yang Shi <yang@...amperecomputing.com>, linux-kernel@...r.kernel.org,
ying.huang@...el.com, feng.tang@...el.com, fengwei.yin@...el.com,
Christian Borntraeger <borntraeger@...ux.ibm.com>, Claudio Imbrenda <imbrenda@...ux.ibm.com>,
Marc Hartmayer <mhartmay@...ux.ibm.com>, Heiko Carstens <hca@...ux.ibm.com>,
Yosry Ahmed <yosryahmed@...gle.com>
Subject: Re: [linux-next:master] [mm/hugetlb_vmemmap] 875fa64577:
vm-scalability.throughput -34.3% regression
On Fri, Jul 19, 2024 at 2:44 AM Oliver Sang <oliver.sang@...el.com> wrote:
>
> hi, Yu Zhao,
>
> On Wed, Jul 17, 2024 at 09:44:33AM -0600, Yu Zhao wrote:
> > On Wed, Jul 17, 2024 at 2:36 AM Yu Zhao <yuzhao@...gle.com> wrote:
> > >
> > > Hi Janosch and Oliver,
> > >
> > > On Wed, Jul 17, 2024 at 1:57 AM Janosch Frank <frankja@...ux.ibm.com> wrote:
> > > >
> > > > On 7/9/24 07:11, kernel test robot wrote:
> > > > > Hello,
> > > > >
> > > > > kernel test robot noticed a -34.3% regression of vm-scalability.throughput on:
> > > > >
> > > > >
> > > > > commit: 875fa64577da9bc8e9963ee14fef8433f20653e7 ("mm/hugetlb_vmemmap: fix race with speculative PFN walkers")
> > > > > https://git.kernel.org/cgit/linux/kernel/git/next/linux-next.git master
> > > > >
> > > > > [still regression on linux-next/master 0b58e108042b0ed28a71cd7edf5175999955b233]
> > > > >
> > > > This has hit s390 huge page backed KVM guests as well.
> > > > Our simple start/stop test case went from ~5 to over 50 seconds of runtime.
> > >
> > > Could you try the attached patch please? Thank you.
> >
> > Thanks, Yosry, for spotting the following typo:
> > flags &= VMEMMAP_SYNCHRONIZE_RCU;
> > It's supposed to be:
> > flags &= ~VMEMMAP_SYNCHRONIZE_RCU;
> >
> > Reattaching v2 with the above typo fixed. Please let me know, Janosch & Oliver.
>
> since the commit is in mainline now, I directly apply your v2 patch upon
> bd225530a4c71 ("mm/hugetlb_vmemmap: fix race with speculative PFN walkers")
>
> in our tests, your v2 patch not only recovers the performance regression,
Thanks for verifying the fix!
> it even has +13.7% performance improvement than 5a4d8944d6b1e (parent of
> bd225530a4c71)
Glad to hear!
(The original patch improved and regressed the performance at the same
time, but the regression is bigger. The fix removed the regression and
surfaced the improvement.)
> detail is as below
>
> =========================================================================================
> compiler/cpufreq_governor/kconfig/rootfs/runtime/size/tbox_group/test/testcase:
> gcc-13/performance/x86_64-rhel-8.3/debian-12-x86_64-20240206.cgz/300s/512G/lkp-icl-2sp2/anon-cow-rand-hugetlb/vm-scalability
>
> commit:
> 5a4d8944d6b1e ("cachestat: do not flush stats in recency check")
> bd225530a4c71 ("mm/hugetlb_vmemmap: fix race with speculative PFN walkers")
> 9a5b87b521401 <---- your v2 patch
>
> 5a4d8944d6b1e1aa bd225530a4c717714722c373144 9a5b87b5214018a2be217dc4648
> ---------------- --------------------------- ---------------------------
> %stddev %change %stddev %change %stddev
> \ | \ | \
> 4.271e+09 ± 10% +348.4% 1.915e+10 ± 6% -39.9% 2.567e+09 ± 20% cpuidle..time
> 774593 ± 4% +1060.9% 8992186 ± 6% -17.2% 641254 cpuidle..usage
> 555365 ± 8% +28.0% 710795 ± 2% -4.5% 530157 ± 5% numa-numastat.node0.local_node
> 629633 ± 4% +23.0% 774346 ± 5% +0.6% 633264 ± 4% numa-numastat.node0.numa_hit
> 255.76 ± 2% +31.1% 335.40 ± 3% -13.8% 220.53 ± 2% uptime.boot
> 10305 ± 6% +144.3% 25171 ± 5% -17.1% 8543 ± 8% uptime.idle
> 1.83 ± 58% +96200.0% 1765 ±155% +736.4% 15.33 ± 24% perf-c2c.DRAM.local
> 33.00 ± 16% +39068.2% 12925 ±122% +95.5% 64.50 ± 49% perf-c2c.DRAM.remote
> 21.33 ± 8% +2361.7% 525.17 ± 31% +271.1% 79.17 ± 52% perf-c2c.HITM.local
> 9.17 ± 21% +3438.2% 324.33 ± 57% +270.9% 34.00 ± 60% perf-c2c.HITM.remote
> 16.11 ± 7% +37.1 53.16 ± 2% -4.6 11.50 ± 19% mpstat.cpu.all.idle%
> 0.34 ± 2% -0.1 0.22 +0.0 0.35 ± 3% mpstat.cpu.all.irq%
> 0.03 ± 5% +0.0 0.04 ± 8% -0.0 0.02 mpstat.cpu.all.soft%
> 10.58 ± 4% -9.5 1.03 ± 36% +0.1 10.71 ± 2% mpstat.cpu.all.sys%
> 72.94 ± 2% -27.4 45.55 ± 3% +4.5 77.41 ± 2% mpstat.cpu.all.usr%
> 6.00 ± 16% +230.6% 19.83 ± 5% +8.3% 6.50 ± 17% mpstat.max_utilization.seconds
> 16.95 ± 7% +215.5% 53.48 ± 2% -26.2% 12.51 ± 16% vmstat.cpu.id
> 72.33 ± 2% -37.4% 45.31 ± 3% +6.0% 76.65 ± 2% vmstat.cpu.us
> 2.254e+08 -0.0% 2.254e+08 +14.7% 2.584e+08 vmstat.memory.free
> 108.30 -43.3% 61.43 ± 2% +5.4% 114.12 ± 2% vmstat.procs.r
> 2659 +162.6% 6982 ± 3% +3.6% 2753 ± 4% vmstat.system.cs
> 136384 ± 4% -21.9% 106579 ± 7% +13.3% 154581 ± 3% vmstat.system.in
> 203.41 ± 2% +39.2% 283.06 ± 4% -17.1% 168.71 ± 2% time.elapsed_time
> 203.41 ± 2% +39.2% 283.06 ± 4% -17.1% 168.71 ± 2% time.elapsed_time.max
> 148901 ± 6% -45.6% 81059 ± 4% -8.8% 135748 ± 8% time.involuntary_context_switches
> 169.83 ± 23% +85.3% 314.67 ± 8% +7.9% 183.33 ± 7% time.major_page_faults
> 10697 -43.4% 6050 ± 2% +5.6% 11294 ± 2% time.percent_of_cpu_this_job_got
> 2740 ± 6% -86.7% 365.06 ± 43% -16.1% 2298 time.system_time
> 19012 -11.9% 16746 -11.9% 16747 time.user_time
> 14412 ± 5% +4432.0% 653187 -16.6% 12025 ± 3% time.voluntary_context_switches
> 50095 ± 2% -31.5% 34325 ± 2% +18.6% 59408 vm-scalability.median
> 8.25 ± 16% -3.4 4.84 ± 22% -6.6 1.65 ± 15% vm-scalability.median_stddev%
> 6863720 -34.0% 4532485 +13.7% 7805408 vm-scalability.throughput
> 203.41 ± 2% +39.2% 283.06 ± 4% -17.1% 168.71 ± 2% vm-scalability.time.elapsed_time
> 203.41 ± 2% +39.2% 283.06 ± 4% -17.1% 168.71 ± 2% vm-scalability.time.elapsed_time.max
> 148901 ± 6% -45.6% 81059 ± 4% -8.8% 135748 ± 8% vm-scalability.time.involuntary_context_switches
> 10697 -43.4% 6050 ± 2% +5.6% 11294 ± 2% vm-scalability.time.percent_of_cpu_this_job_got
> 2740 ± 6% -86.7% 365.06 ± 43% -16.1% 2298 vm-scalability.time.system_time
> 19012 -11.9% 16746 -11.9% 16747 vm-scalability.time.user_time
> 14412 ± 5% +4432.0% 653187 -16.6% 12025 ± 3% vm-scalability.time.voluntary_context_switches
> 1.159e+09 +0.0% 1.159e+09 +1.6% 1.178e+09 vm-scalability.workload
> 22900043 ± 4% +1.2% 23166356 ± 6% -16.7% 19076170 ± 5% numa-vmstat.node0.nr_free_pages
> 42856 ± 43% +998.5% 470779 ± 51% +318.6% 179409 ±154% numa-vmstat.node0.nr_unevictable
> 42856 ± 43% +998.5% 470779 ± 51% +318.6% 179409 ±154% numa-vmstat.node0.nr_zone_unevictable
> 629160 ± 4% +22.9% 773391 ± 5% +0.5% 632570 ± 4% numa-vmstat.node0.numa_hit
> 554892 ± 8% +27.9% 709841 ± 2% -4.6% 529463 ± 5% numa-vmstat.node0.numa_local
> 27469 ± 14% +0.0% 27475 ± 41% -31.7% 18763 ± 13% numa-vmstat.node1.nr_active_anon
> 767179 ± 2% -55.8% 339212 ± 72% -19.7% 616417 ± 43% numa-vmstat.node1.nr_file_pages
> 10693349 ± 5% +46.3% 15639681 ± 7% +69.4% 18112002 ± 3% numa-vmstat.node1.nr_free_pages
> 14210 ± 27% -65.0% 4973 ± 49% -34.7% 9280 ± 39% numa-vmstat.node1.nr_mapped
> 724050 ± 2% -59.1% 296265 ± 82% -18.9% 587498 ± 47% numa-vmstat.node1.nr_unevictable
> 27469 ± 14% +0.0% 27475 ± 41% -31.7% 18763 ± 13% numa-vmstat.node1.nr_zone_active_anon
> 724050 ± 2% -59.1% 296265 ± 82% -18.9% 587498 ± 47% numa-vmstat.node1.nr_zone_unevictable
> 120619 ± 11% +13.6% 137042 ± 27% -31.2% 82976 ± 7% meminfo.Active
> 120472 ± 11% +13.6% 136895 ± 27% -31.2% 82826 ± 7% meminfo.Active(anon)
> 70234807 +14.6% 80512468 +10.2% 77431344 meminfo.CommitLimit
> 2.235e+08 +0.1% 2.237e+08 +15.1% 2.573e+08 meminfo.DirectMap1G
> 44064 -22.8% 34027 ± 2% +20.7% 53164 ± 2% meminfo.HugePages_Surp
> 44064 -22.8% 34027 ± 2% +20.7% 53164 ± 2% meminfo.HugePages_Total
> 90243440 -22.8% 69688103 ± 2% +20.7% 1.089e+08 ± 2% meminfo.Hugetlb
> 70163 ± 29% -42.6% 40293 ± 11% -21.9% 54789 ± 15% meminfo.Mapped
> 1.334e+08 +15.5% 1.541e+08 +10.7% 1.477e+08 meminfo.MemAvailable
> 1.344e+08 +15.4% 1.551e+08 +10.7% 1.488e+08 meminfo.MemFree
> 2.307e+08 +0.0% 2.307e+08 +14.3% 2.637e+08 meminfo.MemTotal
> 96309843 -21.5% 75639108 ± 2% +19.4% 1.15e+08 ± 2% meminfo.Memused
> 259553 ± 2% -0.9% 257226 ± 15% -10.5% 232211 ± 4% meminfo.Shmem
> 1.2e+08 -2.4% 1.172e+08 +13.3% 1.36e+08 meminfo.max_used_kB
> 18884 ± 10% -7.2% 17519 ± 15% +37.6% 25983 ± 6% numa-meminfo.node0.HugePages_Surp
> 18884 ± 10% -7.2% 17519 ± 15% +37.6% 25983 ± 6% numa-meminfo.node0.HugePages_Total
> 91526744 ± 4% +1.2% 92620825 ± 6% -16.7% 76248423 ± 5% numa-meminfo.node0.MemFree
> 40158207 ± 9% -2.7% 39064126 ± 15% +38.0% 55436528 ± 7% numa-meminfo.node0.MemUsed
> 171426 ± 43% +998.5% 1883116 ± 51% +318.6% 717638 ±154% numa-meminfo.node0.Unevictable
> 110091 ± 14% -0.1% 109981 ± 41% -31.7% 75226 ± 13% numa-meminfo.node1.Active
> 110025 ± 14% -0.1% 109915 ± 41% -31.7% 75176 ± 13% numa-meminfo.node1.Active(anon)
> 3068496 ± 2% -55.8% 1356754 ± 72% -19.6% 2466084 ± 43% numa-meminfo.node1.FilePages
> 25218 ± 4% -34.7% 16475 ± 12% +7.9% 27213 ± 3% numa-meminfo.node1.HugePages_Surp
> 25218 ± 4% -34.7% 16475 ± 12% +7.9% 27213 ± 3% numa-meminfo.node1.HugePages_Total
> 55867 ± 27% -65.5% 19266 ± 50% -34.4% 36671 ± 38% numa-meminfo.node1.Mapped
> 42795888 ± 5% +46.1% 62520130 ± 7% +69.3% 72441496 ± 3% numa-meminfo.node1.MemFree
> 99028084 +0.0% 99028084 +33.4% 1.321e+08 numa-meminfo.node1.MemTotal
> 56232195 ± 3% -35.1% 36507953 ± 12% +6.0% 59616707 ± 4% numa-meminfo.node1.MemUsed
> 2896199 ± 2% -59.1% 1185064 ± 82% -18.9% 2349991 ± 47% numa-meminfo.node1.Unevictable
> 507357 +0.0% 507357 +1.7% 516000 proc-vmstat.htlb_buddy_alloc_success
> 29942 ± 10% +14.3% 34235 ± 27% -30.7% 20740 ± 7% proc-vmstat.nr_active_anon
> 3324095 +15.7% 3847387 +10.9% 3686860 proc-vmstat.nr_dirty_background_threshold
> 6656318 +15.7% 7704181 +10.9% 7382735 proc-vmstat.nr_dirty_threshold
> 33559092 +15.6% 38798108 +10.9% 37209133 proc-vmstat.nr_free_pages
> 197697 ± 2% -2.5% 192661 +1.0% 199623 proc-vmstat.nr_inactive_anon
> 17939 ± 28% -42.5% 10307 ± 11% -22.4% 13927 ± 14% proc-vmstat.nr_mapped
> 2691 -7.1% 2501 +2.9% 2769 proc-vmstat.nr_page_table_pages
> 64848 ± 2% -0.7% 64386 ± 15% -10.6% 57987 ± 4% proc-vmstat.nr_shmem
> 29942 ± 10% +14.3% 34235 ± 27% -30.7% 20740 ± 7% proc-vmstat.nr_zone_active_anon
> 197697 ± 2% -2.5% 192661 +1.0% 199623 proc-vmstat.nr_zone_inactive_anon
> 1403095 +9.3% 1534152 ± 2% -3.2% 1358244 proc-vmstat.numa_hit
> 1267544 +10.6% 1401482 ± 2% -3.4% 1224210 proc-vmstat.numa_local
> 2.608e+08 +0.1% 2.609e+08 +1.7% 2.651e+08 proc-vmstat.pgalloc_normal
> 1259957 +13.4% 1428284 ± 2% -6.5% 1178198 proc-vmstat.pgfault
> 2.591e+08 +0.3% 2.6e+08 +2.3% 2.649e+08 proc-vmstat.pgfree
> 36883 ± 3% +18.5% 43709 ± 5% -12.2% 32371 ± 3% proc-vmstat.pgreuse
> 1.88 ± 16% -0.6 1.33 ±100% +0.9 2.80 ± 11% perf-profile.calltrace.cycles-pp.nrand48_r
> 16.19 ± 85% +28.6 44.75 ± 95% -11.4 4.78 ±218% perf-profile.calltrace.cycles-pp.hugetlb_fault.handle_mm_fault.do_user_addr_fault.exc_page_fault.asm_exc_page_fault
> 16.20 ± 85% +28.6 44.78 ± 95% -11.4 4.78 ±218% perf-profile.calltrace.cycles-pp.handle_mm_fault.do_user_addr_fault.exc_page_fault.asm_exc_page_fault.do_access
> 16.22 ± 85% +28.6 44.82 ± 95% -11.4 4.79 ±218% perf-profile.calltrace.cycles-pp.exc_page_fault.asm_exc_page_fault.do_access
> 16.22 ± 85% +28.6 44.82 ± 95% -11.4 4.79 ±218% perf-profile.calltrace.cycles-pp.do_user_addr_fault.exc_page_fault.asm_exc_page_fault.do_access
> 16.24 ± 85% +28.8 45.01 ± 95% -11.4 4.80 ±218% perf-profile.calltrace.cycles-pp.asm_exc_page_fault.do_access
> 12.42 ± 84% +29.5 41.89 ± 95% -8.8 3.65 ±223% perf-profile.calltrace.cycles-pp.copy_mc_enhanced_fast_string.copy_subpage.copy_user_large_folio.hugetlb_wp.hugetlb_fault
> 12.52 ± 84% +29.6 42.08 ± 95% -8.8 3.68 ±223% perf-profile.calltrace.cycles-pp.copy_subpage.copy_user_large_folio.hugetlb_wp.hugetlb_fault.handle_mm_fault
> 12.53 ± 84% +29.7 42.23 ± 95% -8.9 3.68 ±223% perf-profile.calltrace.cycles-pp.copy_user_large_folio.hugetlb_wp.hugetlb_fault.handle_mm_fault.do_user_addr_fault
> 12.80 ± 84% +30.9 43.65 ± 95% -9.0 3.76 ±223% perf-profile.calltrace.cycles-pp.hugetlb_wp.hugetlb_fault.handle_mm_fault.do_user_addr_fault.exc_page_fault
> 2.50 ± 17% -0.7 1.78 ±100% +1.2 3.73 ± 11% perf-profile.children.cycles-pp.nrand48_r
> 16.24 ± 85% +28.6 44.87 ± 95% -11.4 4.79 ±218% perf-profile.children.cycles-pp.do_user_addr_fault
> 16.24 ± 85% +28.6 44.87 ± 95% -11.4 4.79 ±218% perf-profile.children.cycles-pp.exc_page_fault
> 16.20 ± 85% +28.7 44.86 ± 95% -11.4 4.78 ±218% perf-profile.children.cycles-pp.hugetlb_fault
> 16.22 ± 85% +28.7 44.94 ± 95% -11.4 4.79 ±218% perf-profile.children.cycles-pp.handle_mm_fault
> 16.26 ± 85% +28.8 45.06 ± 95% -11.5 4.80 ±218% perf-profile.children.cycles-pp.asm_exc_page_fault
> 12.51 ± 84% +29.5 42.01 ± 95% -8.8 3.75 ±218% perf-profile.children.cycles-pp.copy_mc_enhanced_fast_string
> 12.52 ± 84% +29.6 42.11 ± 95% -8.8 3.75 ±218% perf-profile.children.cycles-pp.copy_subpage
> 12.53 ± 84% +29.7 42.25 ± 95% -8.8 3.76 ±218% perf-profile.children.cycles-pp.copy_user_large_folio
> 12.80 ± 84% +30.9 43.65 ± 95% -9.0 3.83 ±218% perf-profile.children.cycles-pp.hugetlb_wp
> 2.25 ± 17% -0.7 1.59 ±100% +1.1 3.36 ± 11% perf-profile.self.cycles-pp.nrand48_r
> 1.74 ± 21% -0.5 1.25 ± 92% +1.2 2.94 ± 13% perf-profile.self.cycles-pp.do_access
> 0.27 ± 17% -0.1 0.19 ±100% +0.1 0.40 ± 11% perf-profile.self.cycles-pp.lrand48_r
> 12.41 ± 84% +29.4 41.80 ± 95% -8.7 3.72 ±218% perf-profile.self.cycles-pp.copy_mc_enhanced_fast_string
> 350208 ± 16% -2.7% 340891 ± 36% -47.2% 184918 ± 9% sched_debug.cfs_rq:/.avg_vruntime.stddev
> 16833 ±149% -100.0% 3.19 ±100% -100.0% 0.58 ±179% sched_debug.cfs_rq:/.left_deadline.avg
> 2154658 ±149% -100.0% 317.15 ± 93% -100.0% 74.40 ±179% sched_debug.cfs_rq:/.left_deadline.max
> 189702 ±149% -100.0% 29.47 ± 94% -100.0% 6.55 ±179% sched_debug.cfs_rq:/.left_deadline.stddev
> 16833 ±149% -100.0% 3.05 ±102% -100.0% 0.58 ±179% sched_debug.cfs_rq:/.left_vruntime.avg
> 2154613 ±149% -100.0% 298.70 ± 95% -100.0% 74.06 ±179% sched_debug.cfs_rq:/.left_vruntime.max
> 189698 ±149% -100.0% 27.96 ± 96% -100.0% 6.52 ±179% sched_debug.cfs_rq:/.left_vruntime.stddev
> 350208 ± 16% -2.7% 340891 ± 36% -47.2% 184918 ± 9% sched_debug.cfs_rq:/.min_vruntime.stddev
> 52.88 ± 14% -19.5% 42.56 ± 39% +22.8% 64.94 ± 9% sched_debug.cfs_rq:/.removed.load_avg.stddev
> 16833 ±149% -100.0% 3.05 ±102% -100.0% 0.58 ±179% sched_debug.cfs_rq:/.right_vruntime.avg
> 2154613 ±149% -100.0% 298.70 ± 95% -100.0% 74.11 ±179% sched_debug.cfs_rq:/.right_vruntime.max
> 189698 ±149% -100.0% 27.96 ± 96% -100.0% 6.53 ±179% sched_debug.cfs_rq:/.right_vruntime.stddev
> 1588 ± 9% -31.2% 1093 ± 18% -20.0% 1270 ± 16% sched_debug.cfs_rq:/.runnable_avg.max
> 676.36 ± 7% -94.8% 35.08 ± 42% -2.7% 657.82 ± 3% sched_debug.cfs_rq:/.util_est.avg
> 1339 ± 8% -74.5% 341.42 ± 24% -22.6% 1037 ± 23% sched_debug.cfs_rq:/.util_est.max
> 152.67 ± 35% -72.3% 42.35 ± 21% -14.9% 129.89 ± 33% sched_debug.cfs_rq:/.util_est.stddev
> 1116839 ± 7% -7.1% 1037321 ± 4% +22.9% 1372316 ± 11% sched_debug.cpu.avg_idle.max
> 126915 ± 10% +31.6% 166966 ± 6% -12.2% 111446 ± 2% sched_debug.cpu.clock.avg
> 126930 ± 10% +31.6% 166977 ± 6% -12.2% 111459 ± 2% sched_debug.cpu.clock.max
> 126899 ± 10% +31.6% 166949 ± 6% -12.2% 111428 ± 2% sched_debug.cpu.clock.min
> 126491 ± 10% +31.7% 166537 ± 6% -12.2% 111078 ± 2% sched_debug.cpu.clock_task.avg
> 126683 ± 10% +31.6% 166730 ± 6% -12.2% 111237 ± 2% sched_debug.cpu.clock_task.max
> 117365 ± 11% +33.6% 156775 ± 6% -13.0% 102099 ± 2% sched_debug.cpu.clock_task.min
> 2826 ± 10% +178.1% 7858 ± 8% -10.3% 2534 ± 6% sched_debug.cpu.nr_switches.avg
> 755.38 ± 15% +423.8% 3956 ± 14% -15.2% 640.33 ± 3% sched_debug.cpu.nr_switches.min
> 126900 ± 10% +31.6% 166954 ± 6% -12.2% 111432 ± 2% sched_debug.cpu_clk
> 125667 ± 10% +31.9% 165721 ± 6% -12.3% 110200 ± 2% sched_debug.ktime
> 0.54 ±141% -99.9% 0.00 ±132% -99.9% 0.00 ±114% sched_debug.rt_rq:.rt_time.avg
> 69.73 ±141% -99.9% 0.06 ±132% -99.9% 0.07 ±114% sched_debug.rt_rq:.rt_time.max
> 6.14 ±141% -99.9% 0.01 ±132% -99.9% 0.01 ±114% sched_debug.rt_rq:.rt_time.stddev
> 127860 ± 10% +31.3% 167917 ± 6% -12.1% 112402 ± 2% sched_debug.sched_clk
> 15.99 +363.6% 74.14 ± 6% +10.1% 17.61 perf-stat.i.MPKI
> 1.467e+10 ± 2% -32.0% 9.975e+09 ± 3% +21.3% 1.779e+10 ± 2% perf-stat.i.branch-instructions
> 0.10 ± 5% +0.6 0.68 ± 5% +0.0 0.11 ± 4% perf-stat.i.branch-miss-rate%
> 10870114 ± 3% -26.4% 8001551 ± 3% +15.7% 12580898 ± 2% perf-stat.i.branch-misses
> 97.11 -20.0 77.11 -0.0 97.10 perf-stat.i.cache-miss-rate%
> 8.118e+08 ± 2% -32.5% 5.482e+08 ± 3% +23.1% 9.992e+08 ± 2% perf-stat.i.cache-misses
> 8.328e+08 ± 2% -28.4% 5.963e+08 ± 3% +22.8% 1.023e+09 ± 2% perf-stat.i.cache-references
> 2601 ± 2% +172.3% 7083 ± 3% +2.5% 2665 ± 5% perf-stat.i.context-switches
> 5.10 +39.5% 7.11 ± 9% -9.2% 4.62 perf-stat.i.cpi
> 2.826e+11 -44.1% 1.58e+11 ± 2% +5.7% 2.987e+11 ± 2% perf-stat.i.cpu-cycles
> 216.56 +42.4% 308.33 ± 6% +2.2% 221.23 perf-stat.i.cpu-migrations
> 358.79 -0.3% 357.70 ± 21% -14.1% 308.23 perf-stat.i.cycles-between-cache-misses
> 6.286e+10 ± 2% -31.7% 4.293e+10 ± 3% +21.3% 7.626e+10 ± 2% perf-stat.i.instructions
> 0.24 +39.9% 0.33 ± 4% +13.6% 0.27 perf-stat.i.ipc
> 5844 -16.9% 4856 ± 2% +12.5% 6577 perf-stat.i.minor-faults
> 5846 -16.9% 4857 ± 2% +12.5% 6578 perf-stat.i.page-faults
> 13.00 -2.2% 12.72 +1.2% 13.15 perf-stat.overall.MPKI
> 0.07 +0.0 0.08 -0.0 0.07 perf-stat.overall.branch-miss-rate%
> 97.44 -5.3 92.09 +0.2 97.66 perf-stat.overall.cache-miss-rate%
> 4.51 -18.4% 3.68 -13.0% 3.92 perf-stat.overall.cpi
> 346.76 -16.6% 289.11 -14.0% 298.06 perf-stat.overall.cycles-between-cache-misses
> 0.22 +22.6% 0.27 +15.0% 0.26 perf-stat.overall.ipc
> 10906 -3.4% 10541 -1.1% 10784 perf-stat.overall.path-length
> 1.445e+10 ± 2% -30.7% 1.001e+10 ± 3% +21.2% 1.752e+10 ± 2% perf-stat.ps.branch-instructions
> 10469697 ± 3% -23.5% 8005730 ± 3% +18.3% 12387061 ± 2% perf-stat.ps.branch-misses
> 8.045e+08 ± 2% -31.9% 5.478e+08 ± 3% +22.7% 9.874e+08 ± 2% perf-stat.ps.cache-misses
> 8.257e+08 ± 2% -27.9% 5.95e+08 ± 3% +22.5% 1.011e+09 ± 2% perf-stat.ps.cache-references
> 2584 ± 2% +169.3% 6958 ± 3% +2.7% 2654 ± 4% perf-stat.ps.context-switches
> 2.789e+11 -43.2% 1.583e+11 ± 2% +5.5% 2.943e+11 ± 2% perf-stat.ps.cpu-cycles
> 214.69 +41.8% 304.37 ± 6% +2.2% 219.46 perf-stat.ps.cpu-migrations
> 6.19e+10 ± 2% -30.4% 4.309e+10 ± 3% +21.3% 7.507e+10 ± 2% perf-stat.ps.instructions
> 5849 -18.0% 4799 ± 2% +12.3% 6568 ± 2% perf-stat.ps.minor-faults
> 5851 -18.0% 4800 ± 2% +12.3% 6570 ± 2% perf-stat.ps.page-faults
> 1.264e+13 -3.4% 1.222e+13 +0.5% 1.27e+13 perf-stat.total.instructions
Powered by blists - more mailing lists