[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <f06f7ebf-ec9c-3126-f00a-a3578ecee3f6@linux.intel.com>
Date: Mon, 13 Apr 2020 16:37:58 +0800
From: Xing Zhengjun <zhengjun.xing@...ux.intel.com>
To: kernel test robot <rong.a.chen@...el.com>,
Ritesh Harjani <riteshh@...ux.ibm.com>
Cc: Theodore Ts'o <tytso@....edu>, kbuild test robot <lkp@...el.com>,
Jan Kara <jack@...e.cz>,
"Darrick J. Wong" <darrick.wong@...cle.com>,
LKML <linux-kernel@...r.kernel.org>, lkp@...ts.01.org
Subject: Re: [LKP] [ext4] d3b6f23f71: stress-ng.fiemap.ops_per_sec -60.5%
regression
Hi Harjani,
Do you have time to take a look at this? Thanks.
On 4/7/2020 4:00 PM, kernel test robot wrote:
> Greeting,
>
> FYI, we noticed a -60.5% regression of stress-ng.fiemap.ops_per_sec due to commit:
>
>
> commit: d3b6f23f71670007817a5d59f3fbafab2b794e8c ("ext4: move ext4_fiemap to use iomap framework")
> https://git.kernel.org/cgit/linux/kernel/git/torvalds/linux.git master
>
> in testcase: stress-ng
> on test machine: 96 threads Intel(R) Xeon(R) Gold 6252 CPU @ 2.10GHz with 192G memory
> with following parameters:
>
> nr_threads: 10%
> disk: 1HDD
> testtime: 1s
> class: os
> cpufreq_governor: performance
> ucode: 0x500002c
> fs: ext4
>
>
>
>
>
>
> Details are as below:
> -------------------------------------------------------------------------------------------------->
>
>
> To reproduce:
>
> git clone https://github.com/intel/lkp-tests.git
> cd lkp-tests
> bin/lkp install job.yaml # job file is attached in this email
> bin/lkp run job.yaml
>
> =========================================================================================
> class/compiler/cpufreq_governor/disk/fs/kconfig/nr_threads/rootfs/tbox_group/testcase/testtime/ucode:
> os/gcc-7/performance/1HDD/ext4/x86_64-rhel-7.6/10%/debian-x86_64-20191114.cgz/lkp-csl-2sp5/stress-ng/1s/0x500002c
>
> commit:
> b2c5764262 ("ext4: make ext4_ind_map_blocks work with fiemap")
> d3b6f23f71 ("ext4: move ext4_fiemap to use iomap framework")
>
> b2c5764262edded1 d3b6f23f71670007817a5d59f3f
> ---------------- ---------------------------
> fail:runs %reproduction fail:runs
> | | |
> :4 25% 1:4 dmesg.WARNING:at#for_ip_interrupt_entry/0x
> 2:4 5% 2:4 perf-profile.calltrace.cycles-pp.sync_regs.error_entry
> 2:4 6% 3:4 perf-profile.calltrace.cycles-pp.error_entry
> 3:4 9% 3:4 perf-profile.children.cycles-pp.error_entry
> 0:4 1% 0:4 perf-profile.self.cycles-pp.error_entry
> %stddev %change %stddev
> \ | \
> 28623 +28.2% 36703 ± 12% stress-ng.daemon.ops
> 28632 +28.2% 36704 ± 12% stress-ng.daemon.ops_per_sec
> 566.00 ± 22% -53.2% 265.00 ± 53% stress-ng.dev.ops
> 278.81 ± 22% -53.0% 131.00 ± 54% stress-ng.dev.ops_per_sec
> 73160 -60.6% 28849 ± 3% stress-ng.fiemap.ops
> 72471 -60.5% 28612 ± 3% stress-ng.fiemap.ops_per_sec
> 23421 ± 12% +21.2% 28388 ± 6% stress-ng.filename.ops
> 22638 ± 12% +20.3% 27241 ± 6% stress-ng.filename.ops_per_sec
> 21.25 ± 7% -10.6% 19.00 ± 3% stress-ng.iomix.ops
> 38.75 ± 49% -47.7% 20.25 ± 96% stress-ng.memhotplug.ops
> 34.45 ± 52% -51.8% 16.62 ±106% stress-ng.memhotplug.ops_per_sec
> 1734 ± 10% +31.4% 2278 ± 10% stress-ng.resources.ops
> 807.56 ± 5% +35.2% 1091 ± 8% stress-ng.resources.ops_per_sec
> 1007356 ± 3% -16.5% 840642 ± 9% stress-ng.revio.ops
> 1007692 ± 3% -16.6% 840711 ± 9% stress-ng.revio.ops_per_sec
> 21812 ± 3% +16.0% 25294 ± 5% stress-ng.sysbadaddr.ops
> 21821 ± 3% +15.9% 25294 ± 5% stress-ng.sysbadaddr.ops_per_sec
> 440.75 ± 4% +21.9% 537.25 ± 9% stress-ng.sysfs.ops
> 440.53 ± 4% +21.9% 536.86 ± 9% stress-ng.sysfs.ops_per_sec
> 13286582 -11.1% 11805520 ± 6% stress-ng.time.file_system_outputs
> 68253896 +2.4% 69860122 stress-ng.time.minor_page_faults
> 197.00 ± 4% -15.9% 165.75 ± 12% stress-ng.xattr.ops
> 192.45 ± 5% -16.1% 161.46 ± 11% stress-ng.xattr.ops_per_sec
> 15310 +62.5% 24875 ± 22% stress-ng.zombie.ops
> 15310 +62.5% 24874 ± 22% stress-ng.zombie.ops_per_sec
> 203.50 ± 12% -47.3% 107.25 ± 49% vmstat.io.bi
> 861318 ± 18% -29.7% 605884 ± 5% meminfo.AnonHugePages
> 1062742 ± 14% -20.2% 847853 ± 3% meminfo.AnonPages
> 31093 ± 6% +9.6% 34090 ± 3% meminfo.KernelStack
> 7151 ± 34% +55.8% 11145 ± 9% meminfo.Mlocked
> 1.082e+08 ± 5% -40.2% 64705429 ± 31% numa-numastat.node0.local_node
> 1.082e+08 ± 5% -40.2% 64739883 ± 31% numa-numastat.node0.numa_hit
> 46032662 ± 21% +104.3% 94042918 ± 20% numa-numastat.node1.local_node
> 46074205 ± 21% +104.2% 94072810 ± 20% numa-numastat.node1.numa_hit
> 3942 ± 3% +14.2% 4501 ± 4% slabinfo.pool_workqueue.active_objs
> 4098 ± 3% +14.3% 4683 ± 4% slabinfo.pool_workqueue.num_objs
> 4817 ± 7% +13.3% 5456 ± 8% slabinfo.proc_dir_entry.active_objs
> 5153 ± 6% +12.5% 5797 ± 8% slabinfo.proc_dir_entry.num_objs
> 18598 ± 13% -33.1% 12437 ± 20% sched_debug.cfs_rq:/.load.avg
> 452595 ± 56% -71.4% 129637 ± 76% sched_debug.cfs_rq:/.load.max
> 67675 ± 35% -55.1% 30377 ± 42% sched_debug.cfs_rq:/.load.stddev
> 18114 ± 12% -33.7% 12011 ± 20% sched_debug.cfs_rq:/.runnable_weight.avg
> 448215 ± 58% -72.8% 121789 ± 82% sched_debug.cfs_rq:/.runnable_weight.max
> 67083 ± 37% -56.3% 29305 ± 43% sched_debug.cfs_rq:/.runnable_weight.stddev
> -38032 +434.3% -203212 sched_debug.cfs_rq:/.spread0.avg
> -204466 +95.8% -400301 sched_debug.cfs_rq:/.spread0.min
> 90.02 ± 25% -58.1% 37.69 ± 52% sched_debug.cfs_rq:/.util_est_enqueued.avg
> 677.54 ± 6% -39.3% 411.50 ± 22% sched_debug.cfs_rq:/.util_est_enqueued.max
> 196.57 ± 8% -47.6% 103.05 ± 36% sched_debug.cfs_rq:/.util_est_enqueued.stddev
> 3.34 ± 23% +34.1% 4.48 ± 4% sched_debug.cpu.clock.stddev
> 3.34 ± 23% +34.1% 4.48 ± 4% sched_debug.cpu.clock_task.stddev
> 402872 ± 7% -11.9% 354819 ± 2% proc-vmstat.nr_active_anon
> 1730331 -9.5% 1566418 ± 5% proc-vmstat.nr_dirtied
> 31042 ± 6% +9.3% 33915 ± 3% proc-vmstat.nr_kernel_stack
> 229047 -2.4% 223615 proc-vmstat.nr_mapped
> 74008 ± 7% +20.5% 89163 ± 8% proc-vmstat.nr_written
> 402872 ± 7% -11.9% 354819 ± 2% proc-vmstat.nr_zone_active_anon
> 50587 ± 11% -25.2% 37829 ± 14% proc-vmstat.numa_pages_migrated
> 457500 -23.1% 351918 ± 31% proc-vmstat.numa_pte_updates
> 81382485 +1.9% 82907822 proc-vmstat.pgfault
> 2.885e+08 ± 5% -13.3% 2.502e+08 ± 6% proc-vmstat.pgfree
> 42206 ± 12% -46.9% 22399 ± 49% proc-vmstat.pgpgin
> 431233 ± 13% -64.8% 151736 ±109% proc-vmstat.pgrotated
> 176754 ± 7% -40.2% 105637 ± 31% proc-vmstat.thp_fault_alloc
> 314.50 ± 82% +341.5% 1388 ± 44% proc-vmstat.unevictable_pgs_stranded
> 1075269 ± 14% -41.3% 631388 ± 17% numa-meminfo.node0.Active
> 976056 ± 12% -39.7% 588727 ± 19% numa-meminfo.node0.Active(anon)
> 426857 ± 22% -36.4% 271375 ± 13% numa-meminfo.node0.AnonHugePages
> 558590 ± 19% -36.4% 355402 ± 14% numa-meminfo.node0.AnonPages
> 1794824 ± 9% -28.8% 1277157 ± 20% numa-meminfo.node0.FilePages
> 8517 ± 92% -82.7% 1473 ± 89% numa-meminfo.node0.Inactive(file)
> 633118 ± 2% -41.7% 368920 ± 36% numa-meminfo.node0.Mapped
> 2958038 ± 12% -27.7% 2139271 ± 12% numa-meminfo.node0.MemUsed
> 181401 ± 5% -13.7% 156561 ± 4% numa-meminfo.node0.SUnreclaim
> 258124 ± 6% -13.0% 224535 ± 5% numa-meminfo.node0.Slab
> 702083 ± 16% +31.0% 919406 ± 11% numa-meminfo.node1.Active
> 38663 ±107% +137.8% 91951 ± 31% numa-meminfo.node1.Active(file)
> 1154975 ± 7% +41.6% 1635593 ± 12% numa-meminfo.node1.FilePages
> 395813 ± 25% +62.8% 644533 ± 16% numa-meminfo.node1.Inactive
> 394313 ± 25% +62.5% 640686 ± 16% numa-meminfo.node1.Inactive(anon)
> 273317 +88.8% 515976 ± 25% numa-meminfo.node1.Mapped
> 2279237 ± 6% +25.7% 2865582 ± 7% numa-meminfo.node1.MemUsed
> 10830 ± 18% +29.6% 14033 ± 9% numa-meminfo.node1.PageTables
> 149390 ± 3% +23.2% 184085 ± 3% numa-meminfo.node1.SUnreclaim
> 569542 ± 16% +74.8% 995336 ± 21% numa-meminfo.node1.Shmem
> 220774 ± 5% +20.3% 265656 ± 3% numa-meminfo.node1.Slab
> 35623587 ± 5% -11.7% 31444514 ± 3% perf-stat.i.cache-misses
> 2.576e+08 ± 5% -6.8% 2.4e+08 ± 2% perf-stat.i.cache-references
> 3585 -7.3% 3323 ± 5% perf-stat.i.cpu-migrations
> 180139 ± 2% +4.2% 187668 perf-stat.i.minor-faults
> 69.13 +2.6 71.75 perf-stat.i.node-load-miss-rate%
> 4313695 ± 2% -7.4% 3994957 ± 2% perf-stat.i.node-load-misses
> 5466253 ± 11% -17.3% 4521173 ± 6% perf-stat.i.node-loads
> 2818674 ± 6% -15.8% 2372542 ± 5% perf-stat.i.node-stores
> 227810 +4.6% 238290 perf-stat.i.page-faults
> 12.67 ± 4% -7.2% 11.76 ± 2% perf-stat.overall.MPKI
> 1.01 ± 4% -0.0 0.97 ± 3% perf-stat.overall.branch-miss-rate%
> 1044 +13.1% 1181 ± 4% perf-stat.overall.cycles-between-cache-misses
> 40.37 ± 4% +3.6 44.00 ± 2% perf-stat.overall.node-store-miss-rate%
> 36139526 ± 5% -12.5% 31625519 ± 3% perf-stat.ps.cache-misses
> 2.566e+08 ± 5% -6.9% 2.389e+08 ± 2% perf-stat.ps.cache-references
> 3562 -7.2% 3306 ± 5% perf-stat.ps.cpu-migrations
> 179088 +4.2% 186579 perf-stat.ps.minor-faults
> 4323383 ± 2% -7.5% 3999214 perf-stat.ps.node-load-misses
> 5607721 ± 10% -18.5% 4568664 ± 6% perf-stat.ps.node-loads
> 2855134 ± 7% -16.4% 2387345 ± 5% perf-stat.ps.node-stores
> 226270 +4.6% 236709 perf-stat.ps.page-faults
> 242305 ± 10% -42.4% 139551 ± 18% numa-vmstat.node0.nr_active_anon
> 135983 ± 17% -37.4% 85189 ± 10% numa-vmstat.node0.nr_anon_pages
> 209.25 ± 16% -38.1% 129.50 ± 10% numa-vmstat.node0.nr_anon_transparent_hugepages
> 449367 ± 9% -29.7% 315804 ± 20% numa-vmstat.node0.nr_file_pages
> 2167 ± 90% -80.6% 419.75 ± 98% numa-vmstat.node0.nr_inactive_file
> 157405 ± 3% -41.4% 92206 ± 35% numa-vmstat.node0.nr_mapped
> 2022 ± 30% -73.3% 539.25 ± 91% numa-vmstat.node0.nr_mlock
> 3336 ± 10% -24.3% 2524 ± 25% numa-vmstat.node0.nr_page_table_pages
> 286158 ± 10% -41.2% 168337 ± 37% numa-vmstat.node0.nr_shmem
> 45493 ± 5% -14.1% 39094 ± 4% numa-vmstat.node0.nr_slab_unreclaimable
> 242294 ± 10% -42.4% 139547 ± 18% numa-vmstat.node0.nr_zone_active_anon
> 2167 ± 90% -80.6% 419.75 ± 98% numa-vmstat.node0.nr_zone_inactive_file
> 54053924 ± 8% -39.3% 32786242 ± 34% numa-vmstat.node0.numa_hit
> 53929628 ± 8% -39.5% 32619715 ± 34% numa-vmstat.node0.numa_local
> 9701 ±107% +136.9% 22985 ± 31% numa-vmstat.node1.nr_active_file
> 202.50 ± 16% -25.1% 151.75 ± 23% numa-vmstat.node1.nr_anon_transparent_hugepages
> 284922 ± 7% +43.3% 408195 ± 13% numa-vmstat.node1.nr_file_pages
> 96002 ± 26% +67.5% 160850 ± 17% numa-vmstat.node1.nr_inactive_anon
> 68077 ± 2% +90.3% 129533 ± 25% numa-vmstat.node1.nr_mapped
> 138482 ± 15% +79.2% 248100 ± 22% numa-vmstat.node1.nr_shmem
> 37396 ± 3% +23.3% 46094 ± 3% numa-vmstat.node1.nr_slab_unreclaimable
> 9701 ±107% +136.9% 22985 ± 31% numa-vmstat.node1.nr_zone_active_file
> 96005 ± 26% +67.5% 160846 ± 17% numa-vmstat.node1.nr_zone_inactive_anon
> 23343661 ± 17% +99.9% 46664267 ± 23% numa-vmstat.node1.numa_hit
> 23248487 ± 17% +100.5% 46610447 ± 23% numa-vmstat.node1.numa_local
> 105745 ± 23% +112.6% 224805 ± 24% softirqs.CPU0.NET_RX
> 133310 ± 36% -45.3% 72987 ± 52% softirqs.CPU1.NET_RX
> 170110 ± 55% -66.8% 56407 ±147% softirqs.CPU11.NET_RX
> 91465 ± 36% -65.2% 31858 ±112% softirqs.CPU13.NET_RX
> 164491 ± 57% -77.7% 36641 ±121% softirqs.CPU15.NET_RX
> 121069 ± 55% -99.3% 816.75 ± 96% softirqs.CPU17.NET_RX
> 81019 ± 4% -8.7% 73967 ± 4% softirqs.CPU20.RCU
> 72143 ± 63% -89.8% 7360 ±172% softirqs.CPU22.NET_RX
> 270663 ± 17% -57.9% 113915 ± 45% softirqs.CPU24.NET_RX
> 20149 ± 76% +474.1% 115680 ± 62% softirqs.CPU26.NET_RX
> 14033 ± 70% +977.5% 151211 ± 75% softirqs.CPU27.NET_RX
> 27834 ± 94% +476.1% 160357 ± 28% softirqs.CPU28.NET_RX
> 35346 ± 68% +212.0% 110290 ± 30% softirqs.CPU29.NET_RX
> 34347 ±103% +336.5% 149941 ± 32% softirqs.CPU32.NET_RX
> 70077 ± 3% +10.8% 77624 ± 3% softirqs.CPU34.RCU
> 36453 ± 84% +339.6% 160253 ± 42% softirqs.CPU36.NET_RX
> 72367 ± 2% +10.6% 80043 softirqs.CPU37.RCU
> 25239 ±118% +267.7% 92799 ± 45% softirqs.CPU38.NET_RX
> 4995 ±170% +1155.8% 62734 ± 62% softirqs.CPU39.NET_RX
> 4641 ±145% +1611.3% 79432 ± 90% softirqs.CPU42.NET_RX
> 7192 ± 65% +918.0% 73225 ± 66% softirqs.CPU45.NET_RX
> 1772 ±166% +1837.4% 34344 ± 63% softirqs.CPU46.NET_RX
> 13149 ± 81% +874.7% 128170 ± 58% softirqs.CPU47.NET_RX
> 86484 ± 94% -92.6% 6357 ±172% softirqs.CPU48.NET_RX
> 129128 ± 27% -95.8% 5434 ±172% softirqs.CPU55.NET_RX
> 82772 ± 59% -91.7% 6891 ±164% softirqs.CPU56.NET_RX
> 145313 ± 57% -87.8% 17796 ± 88% softirqs.CPU57.NET_RX
> 118160 ± 33% -86.3% 16226 ±109% softirqs.CPU58.NET_RX
> 94576 ± 56% -94.1% 5557 ±173% softirqs.CPU6.NET_RX
> 82900 ± 77% -66.8% 27508 ±171% softirqs.CPU62.NET_RX
> 157291 ± 30% -81.1% 29656 ±111% softirqs.CPU64.NET_RX
> 135101 ± 28% -80.2% 26748 ± 90% softirqs.CPU67.NET_RX
> 146574 ± 56% -100.0% 69.75 ± 98% softirqs.CPU68.NET_RX
> 81347 ± 2% -9.0% 74024 ± 2% softirqs.CPU68.RCU
> 201729 ± 37% -99.6% 887.50 ±107% softirqs.CPU69.NET_RX
> 108454 ± 78% -97.9% 2254 ±169% softirqs.CPU70.NET_RX
> 55289 ±104% -89.3% 5942 ±172% softirqs.CPU71.NET_RX
> 10112 ±172% +964.6% 107651 ± 89% softirqs.CPU72.NET_RX
> 3136 ±171% +1522.2% 50879 ± 66% softirqs.CPU73.NET_RX
> 13353 ± 79% +809.2% 121407 ±101% softirqs.CPU74.NET_RX
> 75194 ± 3% +10.3% 82957 ± 5% softirqs.CPU75.RCU
> 11002 ±173% +1040.8% 125512 ± 61% softirqs.CPU76.NET_RX
> 2463 ±173% +2567.3% 65708 ± 77% softirqs.CPU78.NET_RX
> 25956 ± 3% -7.8% 23932 ± 3% softirqs.CPU78.SCHED
> 16366 ±150% +340.7% 72125 ± 91% softirqs.CPU82.NET_RX
> 14553 ±130% +1513.4% 234809 ± 27% softirqs.CPU93.NET_RX
> 26314 -9.2% 23884 ± 3% softirqs.CPU93.SCHED
> 4582 ± 88% +4903.4% 229268 ± 23% softirqs.CPU94.NET_RX
> 11214 ±111% +1762.5% 208867 ± 18% softirqs.CPU95.NET_RX
> 1.53 ± 27% -0.5 0.99 ± 17% perf-profile.calltrace.cycles-pp.exit_to_usermode_loop.do_syscall_64.entry_SYSCALL_64_after_hwframe
> 1.52 ± 27% -0.5 0.99 ± 17% perf-profile.calltrace.cycles-pp.do_signal.exit_to_usermode_loop.do_syscall_64.entry_SYSCALL_64_after_hwframe
> 1.39 ± 29% -0.5 0.88 ± 21% perf-profile.calltrace.cycles-pp.do_group_exit.get_signal.do_signal.exit_to_usermode_loop.do_syscall_64
> 1.39 ± 29% -0.5 0.88 ± 21% perf-profile.calltrace.cycles-pp.get_signal.do_signal.exit_to_usermode_loop.do_syscall_64.entry_SYSCALL_64_after_hwframe
> 0.50 ± 59% +0.3 0.81 ± 13% perf-profile.calltrace.cycles-pp.filemap_map_pages.handle_pte_fault.__handle_mm_fault.handle_mm_fault.do_page_fault
> 5.70 ± 9% +0.8 6.47 ± 7% perf-profile.calltrace.cycles-pp.do_exit.do_group_exit.get_signal.do_signal.exit_to_usermode_loop
> 5.48 ± 9% +0.8 6.27 ± 7% perf-profile.calltrace.cycles-pp.exit_mmap.mmput.do_exit.do_group_exit.get_signal
> 5.49 ± 9% +0.8 6.28 ± 7% perf-profile.calltrace.cycles-pp.mmput.do_exit.do_group_exit.get_signal.do_signal
> 4.30 ± 4% +1.3 5.60 ± 7% perf-profile.calltrace.cycles-pp.do_group_exit.get_signal.do_signal.exit_to_usermode_loop.prepare_exit_to_usermode
> 4.40 ± 4% +1.3 5.69 ± 7% perf-profile.calltrace.cycles-pp.prepare_exit_to_usermode.swapgs_restore_regs_and_return_to_usermode
> 4.37 ± 4% +1.3 5.66 ± 7% perf-profile.calltrace.cycles-pp.exit_to_usermode_loop.prepare_exit_to_usermode.swapgs_restore_regs_and_return_to_usermode
> 4.36 ± 4% +1.3 5.66 ± 7% perf-profile.calltrace.cycles-pp.do_signal.exit_to_usermode_loop.prepare_exit_to_usermode.swapgs_restore_regs_and_return_to_usermode
> 4.33 ± 4% +1.3 5.62 ± 7% perf-profile.calltrace.cycles-pp.get_signal.do_signal.exit_to_usermode_loop.prepare_exit_to_usermode.swapgs_restore_regs_and_return_to_usermode
> 4.44 ± 4% +1.3 5.74 ± 7% perf-profile.calltrace.cycles-pp.swapgs_restore_regs_and_return_to_usermode
> 3.20 ± 10% -2.4 0.78 ±156% perf-profile.children.cycles-pp.copy_page
> 0.16 ± 9% -0.1 0.08 ± 64% perf-profile.children.cycles-pp.irq_work_interrupt
> 0.16 ± 9% -0.1 0.08 ± 64% perf-profile.children.cycles-pp.smp_irq_work_interrupt
> 0.24 ± 5% -0.1 0.17 ± 18% perf-profile.children.cycles-pp.irq_work_run_list
> 0.16 ± 9% -0.1 0.10 ± 24% perf-profile.children.cycles-pp.irq_work_run
> 0.16 ± 9% -0.1 0.10 ± 24% perf-profile.children.cycles-pp.printk
> 0.23 ± 6% -0.1 0.17 ± 9% perf-profile.children.cycles-pp.__do_execve_file
> 0.08 ± 14% -0.1 0.03 ±100% perf-profile.children.cycles-pp.delay_tsc
> 0.16 ± 6% -0.1 0.11 ± 9% perf-profile.children.cycles-pp.load_elf_binary
> 0.16 ± 7% -0.0 0.12 ± 13% perf-profile.children.cycles-pp.search_binary_handler
> 0.20 ± 7% -0.0 0.15 ± 10% perf-profile.children.cycles-pp.call_usermodehelper_exec_async
> 0.19 ± 6% -0.0 0.15 ± 11% perf-profile.children.cycles-pp.do_execve
> 0.08 ± 10% -0.0 0.04 ± 59% perf-profile.children.cycles-pp.__vunmap
> 0.15 ± 3% -0.0 0.11 ± 7% perf-profile.children.cycles-pp.rcu_idle_exit
> 0.12 ± 10% -0.0 0.09 ± 14% perf-profile.children.cycles-pp.__switch_to_asm
> 0.09 ± 13% -0.0 0.07 ± 5% perf-profile.children.cycles-pp.des3_ede_encrypt
> 0.06 ± 11% +0.0 0.09 ± 13% perf-profile.children.cycles-pp.mark_page_accessed
> 0.15 ± 5% +0.0 0.19 ± 12% perf-profile.children.cycles-pp.apparmor_cred_prepare
> 0.22 ± 8% +0.0 0.27 ± 11% perf-profile.children.cycles-pp.mem_cgroup_throttle_swaprate
> 0.17 ± 2% +0.0 0.22 ± 12% perf-profile.children.cycles-pp.security_prepare_creds
> 0.95 ± 17% +0.3 1.22 ± 14% perf-profile.children.cycles-pp.filemap_map_pages
> 5.92 ± 8% +0.7 6.65 ± 7% perf-profile.children.cycles-pp.get_signal
> 5.66 ± 9% +0.8 6.44 ± 7% perf-profile.children.cycles-pp.mmput
> 5.65 ± 9% +0.8 6.43 ± 7% perf-profile.children.cycles-pp.exit_mmap
> 4.40 ± 4% +1.3 5.70 ± 7% perf-profile.children.cycles-pp.prepare_exit_to_usermode
> 4.45 ± 4% +1.3 5.75 ± 7% perf-profile.children.cycles-pp.swapgs_restore_regs_and_return_to_usermode
> 3.16 ± 10% -2.4 0.77 ±155% perf-profile.self.cycles-pp.copy_page
> 0.08 ± 14% -0.1 0.03 ±100% perf-profile.self.cycles-pp.delay_tsc
> 0.12 ± 10% -0.0 0.09 ± 14% perf-profile.self.cycles-pp.__switch_to_asm
> 0.08 ± 12% -0.0 0.06 ± 17% perf-profile.self.cycles-pp.enqueue_task_fair
> 0.09 ± 13% -0.0 0.07 ± 5% perf-profile.self.cycles-pp.des3_ede_encrypt
> 0.07 ± 13% +0.0 0.08 ± 19% perf-profile.self.cycles-pp.__lru_cache_add
> 0.19 ± 9% +0.0 0.22 ± 10% perf-profile.self.cycles-pp.mem_cgroup_throttle_swaprate
> 0.15 ± 5% +0.0 0.19 ± 11% perf-profile.self.cycles-pp.apparmor_cred_prepare
> 0.05 ± 58% +0.0 0.09 ± 13% perf-profile.self.cycles-pp.mark_page_accessed
> 0.58 ± 10% +0.2 0.80 ± 20% perf-profile.self.cycles-pp.release_pages
> 0.75 ±173% +1.3e+05% 1005 ±100% interrupts.127:PCI-MSI.31981660-edge.i40e-eth0-TxRx-91
> 820.75 ±111% -99.9% 0.50 ±173% interrupts.47:PCI-MSI.31981580-edge.i40e-eth0-TxRx-11
> 449.25 ± 86% -100.0% 0.00 interrupts.53:PCI-MSI.31981586-edge.i40e-eth0-TxRx-17
> 33.25 ±157% -100.0% 0.00 interrupts.57:PCI-MSI.31981590-edge.i40e-eth0-TxRx-21
> 0.75 ±110% +63533.3% 477.25 ±162% interrupts.61:PCI-MSI.31981594-edge.i40e-eth0-TxRx-25
> 561.50 ±160% -100.0% 0.00 interrupts.65:PCI-MSI.31981598-edge.i40e-eth0-TxRx-29
> 82921 ± 8% -11.1% 73748 ± 6% interrupts.CPU11.CAL:Function_call_interrupts
> 66509 ± 30% -32.6% 44828 ± 8% interrupts.CPU14.TLB:TLB_shootdowns
> 43105 ± 98% -90.3% 4183 ± 21% interrupts.CPU17.RES:Rescheduling_interrupts
> 148719 ± 70% -69.4% 45471 ± 16% interrupts.CPU17.TLB:TLB_shootdowns
> 85589 ± 42% -52.2% 40884 ± 5% interrupts.CPU20.TLB:TLB_shootdowns
> 222472 ± 41% -98.0% 4360 ± 45% interrupts.CPU22.RES:Rescheduling_interrupts
> 0.50 ±173% +95350.0% 477.25 ±162% interrupts.CPU25.61:PCI-MSI.31981594-edge.i40e-eth0-TxRx-25
> 76029 ± 10% +14.9% 87389 ± 5% interrupts.CPU25.CAL:Function_call_interrupts
> 399042 ± 6% +13.4% 452479 ± 8% interrupts.CPU27.LOC:Local_timer_interrupts
> 561.00 ±161% -100.0% 0.00 interrupts.CPU29.65:PCI-MSI.31981598-edge.i40e-eth0-TxRx-29
> 7034 ± 46% +1083.8% 83279 ±138% interrupts.CPU29.RES:Rescheduling_interrupts
> 17829 ± 99% -71.0% 5172 ± 16% interrupts.CPU30.RES:Rescheduling_interrupts
> 5569 ± 15% +2414.7% 140059 ± 94% interrupts.CPU31.RES:Rescheduling_interrupts
> 37674 ± 16% +36.6% 51473 ± 25% interrupts.CPU31.TLB:TLB_shootdowns
> 47905 ± 39% +76.6% 84583 ± 38% interrupts.CPU34.TLB:TLB_shootdowns
> 568.75 ±140% +224.8% 1847 ± 90% interrupts.CPU36.NMI:Non-maskable_interrupts
> 568.75 ±140% +224.8% 1847 ± 90% interrupts.CPU36.PMI:Performance_monitoring_interrupts
> 4236 ± 25% +2168.5% 96092 ± 90% interrupts.CPU36.RES:Rescheduling_interrupts
> 52717 ± 27% +43.3% 75565 ± 28% interrupts.CPU37.TLB:TLB_shootdowns
> 41418 ± 9% +136.6% 98010 ± 50% interrupts.CPU39.TLB:TLB_shootdowns
> 5551 ± 8% +847.8% 52615 ± 66% interrupts.CPU40.RES:Rescheduling_interrupts
> 4746 ± 25% +865.9% 45841 ± 91% interrupts.CPU42.RES:Rescheduling_interrupts
> 37556 ± 11% +24.6% 46808 ± 6% interrupts.CPU42.TLB:TLB_shootdowns
> 21846 ±124% -84.4% 3415 ± 46% interrupts.CPU48.RES:Rescheduling_interrupts
> 891.50 ± 22% -35.2% 577.25 ± 40% interrupts.CPU49.NMI:Non-maskable_interrupts
> 891.50 ± 22% -35.2% 577.25 ± 40% interrupts.CPU49.PMI:Performance_monitoring_interrupts
> 20459 ±120% -79.2% 4263 ± 14% interrupts.CPU49.RES:Rescheduling_interrupts
> 59840 ± 21% -23.1% 46042 ± 16% interrupts.CPU5.TLB:TLB_shootdowns
> 65200 ± 19% -34.5% 42678 ± 9% interrupts.CPU51.TLB:TLB_shootdowns
> 70923 ±153% -94.0% 4270 ± 29% interrupts.CPU53.RES:Rescheduling_interrupts
> 65312 ± 22% -28.7% 46578 ± 14% interrupts.CPU56.TLB:TLB_shootdowns
> 65828 ± 24% -33.4% 43846 ± 4% interrupts.CPU59.TLB:TLB_shootdowns
> 72558 ±156% -93.2% 4906 ± 9% interrupts.CPU6.RES:Rescheduling_interrupts
> 68698 ± 34% -32.6% 46327 ± 18% interrupts.CPU61.TLB:TLB_shootdowns
> 109745 ± 44% -57.4% 46711 ± 16% interrupts.CPU62.TLB:TLB_shootdowns
> 89714 ± 44% -48.5% 46198 ± 7% interrupts.CPU63.TLB:TLB_shootdowns
> 59380 ±136% -91.5% 5066 ± 13% interrupts.CPU69.RES:Rescheduling_interrupts
> 40094 ± 18% +133.9% 93798 ± 44% interrupts.CPU78.TLB:TLB_shootdowns
> 129884 ± 72% -55.3% 58034 ±157% interrupts.CPU8.RES:Rescheduling_interrupts
> 69984 ± 11% +51.4% 105957 ± 20% interrupts.CPU80.CAL:Function_call_interrupts
> 32857 ± 10% +128.7% 75131 ± 36% interrupts.CPU80.TLB:TLB_shootdowns
> 35726 ± 16% +34.6% 48081 ± 12% interrupts.CPU82.TLB:TLB_shootdowns
> 73820 ± 17% +28.2% 94643 ± 8% interrupts.CPU84.CAL:Function_call_interrupts
> 38829 ± 28% +190.3% 112736 ± 42% interrupts.CPU84.TLB:TLB_shootdowns
> 36129 ± 4% +47.6% 53329 ± 13% interrupts.CPU85.TLB:TLB_shootdowns
> 4693 ± 7% +1323.0% 66793 ±145% interrupts.CPU86.RES:Rescheduling_interrupts
> 38003 ± 11% +94.8% 74031 ± 43% interrupts.CPU86.TLB:TLB_shootdowns
> 78022 ± 3% +7.9% 84210 ± 3% interrupts.CPU87.CAL:Function_call_interrupts
> 36359 ± 6% +54.9% 56304 ± 48% interrupts.CPU88.TLB:TLB_shootdowns
> 89031 ±105% -95.0% 4475 ± 40% interrupts.CPU9.RES:Rescheduling_interrupts
> 40085 ± 11% +60.6% 64368 ± 27% interrupts.CPU91.TLB:TLB_shootdowns
> 42244 ± 10% +44.8% 61162 ± 35% interrupts.CPU94.TLB:TLB_shootdowns
> 40959 ± 15% +109.4% 85780 ± 41% interrupts.CPU95.TLB:TLB_shootdowns
>
>
>
> stress-ng.fiemap.ops
>
> 80000 +-------------------------------------------------------------------+
> 75000 |..+. .+.. .+..+.. .+. .+.. |
> | +..+..+..+.+. .+..+.. .+ +. +. +.+..+..+..+.+..|
> 70000 |-+ + +. |
> 65000 |-+ |
> 60000 |-+ |
> 55000 |-+ |
> | |
> 50000 |-+ |
> 45000 |-+ |
> 40000 |-+ |
> 35000 |-+ O |
> | O O O O O |
> 30000 |-+ O O O O O O O O O O O O O O O O O |
> 25000 +-------------------------------------------------------------------+
>
>
> stress-ng.fiemap.ops_per_sec
>
> 80000 +-------------------------------------------------------------------+
> 75000 |.. .+.. .+.. |
> | +. .+..+..+.+. .+..+.. .+.+. +..+.+..+..+.+..+..+..+.+..|
> 70000 |-+ +. + +. |
> 65000 |-+ |
> 60000 |-+ |
> 55000 |-+ |
> | |
> 50000 |-+ |
> 45000 |-+ |
> 40000 |-+ |
> 35000 |-+ O |
> | O O O O |
> 30000 |-+ O O O O O O O O O O O O O O O O |
> 25000 +-------------------------------------------------------------------+
>
>
> [*] bisect-good sample
> [O] bisect-bad sample
>
>
>
> Disclaimer:
> Results have been estimated based on internal Intel analysis and are provided
> for informational purposes only. Any difference in system hardware or software
> design or configuration may affect actual performance.
>
>
> Thanks,
> Rong Chen
>
>
> _______________________________________________
> LKP mailing list -- lkp@...ts.01.org
> To unsubscribe send an email to lkp-leave@...ts.01.org
>
--
Zhengjun Xing
Powered by blists - more mailing lists