[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <7bba55fd-955a-355e-4df1-7b4493b125a6@linux.intel.com>
Date: Tue, 16 Jun 2020 15:13:47 +0800
From: Xing Zhengjun <zhengjun.xing@...ux.intel.com>
To: Ritesh Harjani <riteshh@...ux.ibm.com>,
kernel test robot <rong.a.chen@...el.com>
Cc: Theodore Ts'o <tytso@....edu>, kbuild test robot <lkp@...el.com>,
Jan Kara <jack@...e.cz>,
"Darrick J. Wong" <darrick.wong@...cle.com>,
LKML <linux-kernel@...r.kernel.org>, lkp@...ts.01.org
Subject: Re: [LKP] [ext4] d3b6f23f71: stress-ng.fiemap.ops_per_sec -60.5%
regression
Hi Ritesh,
I test, the regression still existed in v5.8-rc1. Do you have time
to take a look at it? Thanks.
On 4/14/2020 1:49 PM, Xing Zhengjun wrote:
> Thanks for your quick response, if you need any more test information
> about the regression, please let me known.
>
> On 4/13/2020 6:56 PM, Ritesh Harjani wrote:
>>
>>
>> On 4/13/20 2:07 PM, Xing Zhengjun wrote:
>>> Hi Harjani,
>>>
>>> Do you have time to take a look at this? Thanks.
>>
>> Hello Xing,
>>
>> I do want to look into this. But as of now I am stuck with another
>> mballoc failure issue. I will get back at this once I have some handle
>> over that one.
>>
>> BTW, are you planning to take look at this?
>>
>> -ritesh
>>
>>
>>>
>>> On 4/7/2020 4:00 PM, kernel test robot wrote:
>>>> Greeting,
>>>>
>>>> FYI, we noticed a -60.5% regression of stress-ng.fiemap.ops_per_sec
>>>> due to commit:
>>>>
>>>>
>>>> commit: d3b6f23f71670007817a5d59f3fbafab2b794e8c ("ext4: move
>>>> ext4_fiemap to use iomap framework")
>>>> https://git.kernel.org/cgit/linux/kernel/git/torvalds/linux.git master
>>>>
>>>> in testcase: stress-ng
>>>> on test machine: 96 threads Intel(R) Xeon(R) Gold 6252 CPU @ 2.10GHz
>>>> with 192G memory
>>>> with following parameters:
>>>>
>>>> nr_threads: 10%
>>>> disk: 1HDD
>>>> testtime: 1s
>>>> class: os
>>>> cpufreq_governor: performance
>>>> ucode: 0x500002c
>>>> fs: ext4
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>> Details are as below:
>>>> -------------------------------------------------------------------------------------------------->
>>>>
>>>>
>>>>
>>>> To reproduce:
>>>>
>>>> git clone https://github.com/intel/lkp-tests.git
>>>> cd lkp-tests
>>>> bin/lkp install job.yaml # job file is attached in this email
>>>> bin/lkp run job.yaml
>>>>
>>>> =========================================================================================
>>>>
>>>> class/compiler/cpufreq_governor/disk/fs/kconfig/nr_threads/rootfs/tbox_group/testcase/testtime/ucode:
>>>>
>>>> os/gcc-7/performance/1HDD/ext4/x86_64-rhel-7.6/10%/debian-x86_64-20191114.cgz/lkp-csl-2sp5/stress-ng/1s/0x500002c
>>>>
>>>>
>>>> commit:
>>>> b2c5764262 ("ext4: make ext4_ind_map_blocks work with fiemap")
>>>> d3b6f23f71 ("ext4: move ext4_fiemap to use iomap framework")
>>>>
>>>> b2c5764262edded1 d3b6f23f71670007817a5d59f3f
>>>> ---------------- ---------------------------
>>>> fail:runs %reproduction fail:runs
>>>> | | |
>>>> :4 25% 1:4
>>>> dmesg.WARNING:at#for_ip_interrupt_entry/0x
>>>> 2:4 5% 2:4
>>>> perf-profile.calltrace.cycles-pp.sync_regs.error_entry
>>>> 2:4 6% 3:4
>>>> perf-profile.calltrace.cycles-pp.error_entry
>>>> 3:4 9% 3:4
>>>> perf-profile.children.cycles-pp.error_entry
>>>> 0:4 1% 0:4
>>>> perf-profile.self.cycles-pp.error_entry
>>>> %stddev %change %stddev
>>>> \ | \
>>>> 28623 +28.2% 36703 ± 12% stress-ng.daemon.ops
>>>> 28632 +28.2% 36704 ± 12%
>>>> stress-ng.daemon.ops_per_sec
>>>> 566.00 ± 22% -53.2% 265.00 ± 53% stress-ng.dev.ops
>>>> 278.81 ± 22% -53.0% 131.00 ± 54%
>>>> stress-ng.dev.ops_per_sec
>>>> 73160 -60.6% 28849 ± 3% stress-ng.fiemap.ops
>>>> 72471 -60.5% 28612 ± 3%
>>>> stress-ng.fiemap.ops_per_sec
>>>> 23421 ± 12% +21.2% 28388 ± 6% stress-ng.filename.ops
>>>> 22638 ± 12% +20.3% 27241 ± 6%
>>>> stress-ng.filename.ops_per_sec
>>>> 21.25 ± 7% -10.6% 19.00 ± 3% stress-ng.iomix.ops
>>>> 38.75 ± 49% -47.7% 20.25 ± 96% stress-ng.memhotplug.ops
>>>> 34.45 ± 52% -51.8% 16.62 ±106%
>>>> stress-ng.memhotplug.ops_per_sec
>>>> 1734 ± 10% +31.4% 2278 ± 10% stress-ng.resources.ops
>>>> 807.56 ± 5% +35.2% 1091 ± 8%
>>>> stress-ng.resources.ops_per_sec
>>>> 1007356 ± 3% -16.5% 840642 ± 9% stress-ng.revio.ops
>>>> 1007692 ± 3% -16.6% 840711 ± 9%
>>>> stress-ng.revio.ops_per_sec
>>>> 21812 ± 3% +16.0% 25294 ± 5% stress-ng.sysbadaddr.ops
>>>> 21821 ± 3% +15.9% 25294 ± 5%
>>>> stress-ng.sysbadaddr.ops_per_sec
>>>> 440.75 ± 4% +21.9% 537.25 ± 9% stress-ng.sysfs.ops
>>>> 440.53 ± 4% +21.9% 536.86 ± 9%
>>>> stress-ng.sysfs.ops_per_sec
>>>> 13286582 -11.1% 11805520 ± 6%
>>>> stress-ng.time.file_system_outputs
>>>> 68253896 +2.4% 69860122
>>>> stress-ng.time.minor_page_faults
>>>> 197.00 ± 4% -15.9% 165.75 ± 12% stress-ng.xattr.ops
>>>> 192.45 ± 5% -16.1% 161.46 ± 11%
>>>> stress-ng.xattr.ops_per_sec
>>>> 15310 +62.5% 24875 ± 22% stress-ng.zombie.ops
>>>> 15310 +62.5% 24874 ± 22%
>>>> stress-ng.zombie.ops_per_sec
>>>> 203.50 ± 12% -47.3% 107.25 ± 49% vmstat.io.bi
>>>> 861318 ± 18% -29.7% 605884 ± 5% meminfo.AnonHugePages
>>>> 1062742 ± 14% -20.2% 847853 ± 3% meminfo.AnonPages
>>>> 31093 ± 6% +9.6% 34090 ± 3% meminfo.KernelStack
>>>> 7151 ± 34% +55.8% 11145 ± 9% meminfo.Mlocked
>>>> 1.082e+08 ± 5% -40.2% 64705429 ± 31%
>>>> numa-numastat.node0.local_node
>>>> 1.082e+08 ± 5% -40.2% 64739883 ± 31%
>>>> numa-numastat.node0.numa_hit
>>>> 46032662 ± 21% +104.3% 94042918 ± 20%
>>>> numa-numastat.node1.local_node
>>>> 46074205 ± 21% +104.2% 94072810 ± 20%
>>>> numa-numastat.node1.numa_hit
>>>> 3942 ± 3% +14.2% 4501 ± 4%
>>>> slabinfo.pool_workqueue.active_objs
>>>> 4098 ± 3% +14.3% 4683 ± 4%
>>>> slabinfo.pool_workqueue.num_objs
>>>> 4817 ± 7% +13.3% 5456 ± 8%
>>>> slabinfo.proc_dir_entry.active_objs
>>>> 5153 ± 6% +12.5% 5797 ± 8%
>>>> slabinfo.proc_dir_entry.num_objs
>>>> 18598 ± 13% -33.1% 12437 ± 20%
>>>> sched_debug.cfs_rq:/.load.avg
>>>> 452595 ± 56% -71.4% 129637 ± 76%
>>>> sched_debug.cfs_rq:/.load.max
>>>> 67675 ± 35% -55.1% 30377 ± 42%
>>>> sched_debug.cfs_rq:/.load.stddev
>>>> 18114 ± 12% -33.7% 12011 ± 20%
>>>> sched_debug.cfs_rq:/.runnable_weight.avg
>>>> 448215 ± 58% -72.8% 121789 ± 82%
>>>> sched_debug.cfs_rq:/.runnable_weight.max
>>>> 67083 ± 37% -56.3% 29305 ± 43%
>>>> sched_debug.cfs_rq:/.runnable_weight.stddev
>>>> -38032 +434.3% -203212
>>>> sched_debug.cfs_rq:/.spread0.avg
>>>> -204466 +95.8% -400301
>>>> sched_debug.cfs_rq:/.spread0.min
>>>> 90.02 ± 25% -58.1% 37.69 ± 52%
>>>> sched_debug.cfs_rq:/.util_est_enqueued.avg
>>>> 677.54 ± 6% -39.3% 411.50 ± 22%
>>>> sched_debug.cfs_rq:/.util_est_enqueued.max
>>>> 196.57 ± 8% -47.6% 103.05 ± 36%
>>>> sched_debug.cfs_rq:/.util_est_enqueued.stddev
>>>> 3.34 ± 23% +34.1% 4.48 ± 4%
>>>> sched_debug.cpu.clock.stddev
>>>> 3.34 ± 23% +34.1% 4.48 ± 4%
>>>> sched_debug.cpu.clock_task.stddev
>>>> 402872 ± 7% -11.9% 354819 ± 2%
>>>> proc-vmstat.nr_active_anon
>>>> 1730331 -9.5% 1566418 ± 5% proc-vmstat.nr_dirtied
>>>> 31042 ± 6% +9.3% 33915 ± 3%
>>>> proc-vmstat.nr_kernel_stack
>>>> 229047 -2.4% 223615 proc-vmstat.nr_mapped
>>>> 74008 ± 7% +20.5% 89163 ± 8% proc-vmstat.nr_written
>>>> 402872 ± 7% -11.9% 354819 ± 2%
>>>> proc-vmstat.nr_zone_active_anon
>>>> 50587 ± 11% -25.2% 37829 ± 14%
>>>> proc-vmstat.numa_pages_migrated
>>>> 457500 -23.1% 351918 ± 31%
>>>> proc-vmstat.numa_pte_updates
>>>> 81382485 +1.9% 82907822 proc-vmstat.pgfault
>>>> 2.885e+08 ± 5% -13.3% 2.502e+08 ± 6% proc-vmstat.pgfree
>>>> 42206 ± 12% -46.9% 22399 ± 49% proc-vmstat.pgpgin
>>>> 431233 ± 13% -64.8% 151736 ±109% proc-vmstat.pgrotated
>>>> 176754 ± 7% -40.2% 105637 ± 31%
>>>> proc-vmstat.thp_fault_alloc
>>>> 314.50 ± 82% +341.5% 1388 ± 44%
>>>> proc-vmstat.unevictable_pgs_stranded
>>>> 1075269 ± 14% -41.3% 631388 ± 17%
>>>> numa-meminfo.node0.Active
>>>> 976056 ± 12% -39.7% 588727 ± 19%
>>>> numa-meminfo.node0.Active(anon)
>>>> 426857 ± 22% -36.4% 271375 ± 13%
>>>> numa-meminfo.node0.AnonHugePages
>>>> 558590 ± 19% -36.4% 355402 ± 14%
>>>> numa-meminfo.node0.AnonPages
>>>> 1794824 ± 9% -28.8% 1277157 ± 20%
>>>> numa-meminfo.node0.FilePages
>>>> 8517 ± 92% -82.7% 1473 ± 89%
>>>> numa-meminfo.node0.Inactive(file)
>>>> 633118 ± 2% -41.7% 368920 ± 36%
>>>> numa-meminfo.node0.Mapped
>>>> 2958038 ± 12% -27.7% 2139271 ± 12%
>>>> numa-meminfo.node0.MemUsed
>>>> 181401 ± 5% -13.7% 156561 ± 4%
>>>> numa-meminfo.node0.SUnreclaim
>>>> 258124 ± 6% -13.0% 224535 ± 5% numa-meminfo.node0.Slab
>>>> 702083 ± 16% +31.0% 919406 ± 11%
>>>> numa-meminfo.node1.Active
>>>> 38663 ±107% +137.8% 91951 ± 31%
>>>> numa-meminfo.node1.Active(file)
>>>> 1154975 ± 7% +41.6% 1635593 ± 12%
>>>> numa-meminfo.node1.FilePages
>>>> 395813 ± 25% +62.8% 644533 ± 16%
>>>> numa-meminfo.node1.Inactive
>>>> 394313 ± 25% +62.5% 640686 ± 16%
>>>> numa-meminfo.node1.Inactive(anon)
>>>> 273317 +88.8% 515976 ± 25%
>>>> numa-meminfo.node1.Mapped
>>>> 2279237 ± 6% +25.7% 2865582 ± 7%
>>>> numa-meminfo.node1.MemUsed
>>>> 10830 ± 18% +29.6% 14033 ± 9%
>>>> numa-meminfo.node1.PageTables
>>>> 149390 ± 3% +23.2% 184085 ± 3%
>>>> numa-meminfo.node1.SUnreclaim
>>>> 569542 ± 16% +74.8% 995336 ± 21% numa-meminfo.node1.Shmem
>>>> 220774 ± 5% +20.3% 265656 ± 3% numa-meminfo.node1.Slab
>>>> 35623587 ± 5% -11.7% 31444514 ± 3% perf-stat.i.cache-misses
>>>> 2.576e+08 ± 5% -6.8% 2.4e+08 ± 2%
>>>> perf-stat.i.cache-references
>>>> 3585 -7.3% 3323 ± 5%
>>>> perf-stat.i.cpu-migrations
>>>> 180139 ± 2% +4.2% 187668 perf-stat.i.minor-faults
>>>> 69.13 +2.6 71.75
>>>> perf-stat.i.node-load-miss-rate%
>>>> 4313695 ± 2% -7.4% 3994957 ± 2%
>>>> perf-stat.i.node-load-misses
>>>> 5466253 ± 11% -17.3% 4521173 ± 6% perf-stat.i.node-loads
>>>> 2818674 ± 6% -15.8% 2372542 ± 5% perf-stat.i.node-stores
>>>> 227810 +4.6% 238290 perf-stat.i.page-faults
>>>> 12.67 ± 4% -7.2% 11.76 ± 2% perf-stat.overall.MPKI
>>>> 1.01 ± 4% -0.0 0.97 ± 3%
>>>> perf-stat.overall.branch-miss-rate%
>>>> 1044 +13.1% 1181 ± 4%
>>>> perf-stat.overall.cycles-between-cache-misses
>>>> 40.37 ± 4% +3.6 44.00 ± 2%
>>>> perf-stat.overall.node-store-miss-rate%
>>>> 36139526 ± 5% -12.5% 31625519 ± 3%
>>>> perf-stat.ps.cache-misses
>>>> 2.566e+08 ± 5% -6.9% 2.389e+08 ± 2%
>>>> perf-stat.ps.cache-references
>>>> 3562 -7.2% 3306 ± 5%
>>>> perf-stat.ps.cpu-migrations
>>>> 179088 +4.2% 186579
>>>> perf-stat.ps.minor-faults
>>>> 4323383 ± 2% -7.5% 3999214 perf-stat.ps.node-load-misses
>>>> 5607721 ± 10% -18.5% 4568664 ± 6% perf-stat.ps.node-loads
>>>> 2855134 ± 7% -16.4% 2387345 ± 5% perf-stat.ps.node-stores
>>>> 226270 +4.6% 236709 perf-stat.ps.page-faults
>>>> 242305 ± 10% -42.4% 139551 ± 18%
>>>> numa-vmstat.node0.nr_active_anon
>>>> 135983 ± 17% -37.4% 85189 ± 10%
>>>> numa-vmstat.node0.nr_anon_pages
>>>> 209.25 ± 16% -38.1% 129.50 ± 10%
>>>> numa-vmstat.node0.nr_anon_transparent_hugepages
>>>> 449367 ± 9% -29.7% 315804 ± 20%
>>>> numa-vmstat.node0.nr_file_pages
>>>> 2167 ± 90% -80.6% 419.75 ± 98%
>>>> numa-vmstat.node0.nr_inactive_file
>>>> 157405 ± 3% -41.4% 92206 ± 35%
>>>> numa-vmstat.node0.nr_mapped
>>>> 2022 ± 30% -73.3% 539.25 ± 91%
>>>> numa-vmstat.node0.nr_mlock
>>>> 3336 ± 10% -24.3% 2524 ± 25%
>>>> numa-vmstat.node0.nr_page_table_pages
>>>> 286158 ± 10% -41.2% 168337 ± 37%
>>>> numa-vmstat.node0.nr_shmem
>>>> 45493 ± 5% -14.1% 39094 ± 4%
>>>> numa-vmstat.node0.nr_slab_unreclaimable
>>>> 242294 ± 10% -42.4% 139547 ± 18%
>>>> numa-vmstat.node0.nr_zone_active_anon
>>>> 2167 ± 90% -80.6% 419.75 ± 98%
>>>> numa-vmstat.node0.nr_zone_inactive_file
>>>> 54053924 ± 8% -39.3% 32786242 ± 34%
>>>> numa-vmstat.node0.numa_hit
>>>> 53929628 ± 8% -39.5% 32619715 ± 34%
>>>> numa-vmstat.node0.numa_local
>>>> 9701 ±107% +136.9% 22985 ± 31%
>>>> numa-vmstat.node1.nr_active_file
>>>> 202.50 ± 16% -25.1% 151.75 ± 23%
>>>> numa-vmstat.node1.nr_anon_transparent_hugepages
>>>> 284922 ± 7% +43.3% 408195 ± 13%
>>>> numa-vmstat.node1.nr_file_pages
>>>> 96002 ± 26% +67.5% 160850 ± 17%
>>>> numa-vmstat.node1.nr_inactive_anon
>>>> 68077 ± 2% +90.3% 129533 ± 25%
>>>> numa-vmstat.node1.nr_mapped
>>>> 138482 ± 15% +79.2% 248100 ± 22%
>>>> numa-vmstat.node1.nr_shmem
>>>> 37396 ± 3% +23.3% 46094 ± 3%
>>>> numa-vmstat.node1.nr_slab_unreclaimable
>>>> 9701 ±107% +136.9% 22985 ± 31%
>>>> numa-vmstat.node1.nr_zone_active_file
>>>> 96005 ± 26% +67.5% 160846 ± 17%
>>>> numa-vmstat.node1.nr_zone_inactive_anon
>>>> 23343661 ± 17% +99.9% 46664267 ± 23%
>>>> numa-vmstat.node1.numa_hit
>>>> 23248487 ± 17% +100.5% 46610447 ± 23%
>>>> numa-vmstat.node1.numa_local
>>>> 105745 ± 23% +112.6% 224805 ± 24% softirqs.CPU0.NET_RX
>>>> 133310 ± 36% -45.3% 72987 ± 52% softirqs.CPU1.NET_RX
>>>> 170110 ± 55% -66.8% 56407 ±147% softirqs.CPU11.NET_RX
>>>> 91465 ± 36% -65.2% 31858 ±112% softirqs.CPU13.NET_RX
>>>> 164491 ± 57% -77.7% 36641 ±121% softirqs.CPU15.NET_RX
>>>> 121069 ± 55% -99.3% 816.75 ± 96% softirqs.CPU17.NET_RX
>>>> 81019 ± 4% -8.7% 73967 ± 4% softirqs.CPU20.RCU
>>>> 72143 ± 63% -89.8% 7360 ±172% softirqs.CPU22.NET_RX
>>>> 270663 ± 17% -57.9% 113915 ± 45% softirqs.CPU24.NET_RX
>>>> 20149 ± 76% +474.1% 115680 ± 62% softirqs.CPU26.NET_RX
>>>> 14033 ± 70% +977.5% 151211 ± 75% softirqs.CPU27.NET_RX
>>>> 27834 ± 94% +476.1% 160357 ± 28% softirqs.CPU28.NET_RX
>>>> 35346 ± 68% +212.0% 110290 ± 30% softirqs.CPU29.NET_RX
>>>> 34347 ±103% +336.5% 149941 ± 32% softirqs.CPU32.NET_RX
>>>> 70077 ± 3% +10.8% 77624 ± 3% softirqs.CPU34.RCU
>>>> 36453 ± 84% +339.6% 160253 ± 42% softirqs.CPU36.NET_RX
>>>> 72367 ± 2% +10.6% 80043 softirqs.CPU37.RCU
>>>> 25239 ±118% +267.7% 92799 ± 45% softirqs.CPU38.NET_RX
>>>> 4995 ±170% +1155.8% 62734 ± 62% softirqs.CPU39.NET_RX
>>>> 4641 ±145% +1611.3% 79432 ± 90% softirqs.CPU42.NET_RX
>>>> 7192 ± 65% +918.0% 73225 ± 66% softirqs.CPU45.NET_RX
>>>> 1772 ±166% +1837.4% 34344 ± 63% softirqs.CPU46.NET_RX
>>>> 13149 ± 81% +874.7% 128170 ± 58% softirqs.CPU47.NET_RX
>>>> 86484 ± 94% -92.6% 6357 ±172% softirqs.CPU48.NET_RX
>>>> 129128 ± 27% -95.8% 5434 ±172% softirqs.CPU55.NET_RX
>>>> 82772 ± 59% -91.7% 6891 ±164% softirqs.CPU56.NET_RX
>>>> 145313 ± 57% -87.8% 17796 ± 88% softirqs.CPU57.NET_RX
>>>> 118160 ± 33% -86.3% 16226 ±109% softirqs.CPU58.NET_RX
>>>> 94576 ± 56% -94.1% 5557 ±173% softirqs.CPU6.NET_RX
>>>> 82900 ± 77% -66.8% 27508 ±171% softirqs.CPU62.NET_RX
>>>> 157291 ± 30% -81.1% 29656 ±111% softirqs.CPU64.NET_RX
>>>> 135101 ± 28% -80.2% 26748 ± 90% softirqs.CPU67.NET_RX
>>>> 146574 ± 56% -100.0% 69.75 ± 98% softirqs.CPU68.NET_RX
>>>> 81347 ± 2% -9.0% 74024 ± 2% softirqs.CPU68.RCU
>>>> 201729 ± 37% -99.6% 887.50 ±107% softirqs.CPU69.NET_RX
>>>> 108454 ± 78% -97.9% 2254 ±169% softirqs.CPU70.NET_RX
>>>> 55289 ±104% -89.3% 5942 ±172% softirqs.CPU71.NET_RX
>>>> 10112 ±172% +964.6% 107651 ± 89% softirqs.CPU72.NET_RX
>>>> 3136 ±171% +1522.2% 50879 ± 66% softirqs.CPU73.NET_RX
>>>> 13353 ± 79% +809.2% 121407 ±101% softirqs.CPU74.NET_RX
>>>> 75194 ± 3% +10.3% 82957 ± 5% softirqs.CPU75.RCU
>>>> 11002 ±173% +1040.8% 125512 ± 61% softirqs.CPU76.NET_RX
>>>> 2463 ±173% +2567.3% 65708 ± 77% softirqs.CPU78.NET_RX
>>>> 25956 ± 3% -7.8% 23932 ± 3% softirqs.CPU78.SCHED
>>>> 16366 ±150% +340.7% 72125 ± 91% softirqs.CPU82.NET_RX
>>>> 14553 ±130% +1513.4% 234809 ± 27% softirqs.CPU93.NET_RX
>>>> 26314 -9.2% 23884 ± 3% softirqs.CPU93.SCHED
>>>> 4582 ± 88% +4903.4% 229268 ± 23% softirqs.CPU94.NET_RX
>>>> 11214 ±111% +1762.5% 208867 ± 18% softirqs.CPU95.NET_RX
>>>> 1.53 ± 27% -0.5 0.99 ± 17%
>>>> perf-profile.calltrace.cycles-pp.exit_to_usermode_loop.do_syscall_64.entry_SYSCALL_64_after_hwframe
>>>>
>>>> 1.52 ± 27% -0.5 0.99 ± 17%
>>>> perf-profile.calltrace.cycles-pp.do_signal.exit_to_usermode_loop.do_syscall_64.entry_SYSCALL_64_after_hwframe
>>>>
>>>> 1.39 ± 29% -0.5 0.88 ± 21%
>>>> perf-profile.calltrace.cycles-pp.do_group_exit.get_signal.do_signal.exit_to_usermode_loop.do_syscall_64
>>>>
>>>> 1.39 ± 29% -0.5 0.88 ± 21%
>>>> perf-profile.calltrace.cycles-pp.get_signal.do_signal.exit_to_usermode_loop.do_syscall_64.entry_SYSCALL_64_after_hwframe
>>>>
>>>> 0.50 ± 59% +0.3 0.81 ± 13%
>>>> perf-profile.calltrace.cycles-pp.filemap_map_pages.handle_pte_fault.__handle_mm_fault.handle_mm_fault.do_page_fault
>>>>
>>>> 5.70 ± 9% +0.8 6.47 ± 7%
>>>> perf-profile.calltrace.cycles-pp.do_exit.do_group_exit.get_signal.do_signal.exit_to_usermode_loop
>>>>
>>>> 5.48 ± 9% +0.8 6.27 ± 7%
>>>> perf-profile.calltrace.cycles-pp.exit_mmap.mmput.do_exit.do_group_exit.get_signal
>>>>
>>>> 5.49 ± 9% +0.8 6.28 ± 7%
>>>> perf-profile.calltrace.cycles-pp.mmput.do_exit.do_group_exit.get_signal.do_signal
>>>>
>>>> 4.30 ± 4% +1.3 5.60 ± 7%
>>>> perf-profile.calltrace.cycles-pp.do_group_exit.get_signal.do_signal.exit_to_usermode_loop.prepare_exit_to_usermode
>>>>
>>>> 4.40 ± 4% +1.3 5.69 ± 7%
>>>> perf-profile.calltrace.cycles-pp.prepare_exit_to_usermode.swapgs_restore_regs_and_return_to_usermode
>>>>
>>>> 4.37 ± 4% +1.3 5.66 ± 7%
>>>> perf-profile.calltrace.cycles-pp.exit_to_usermode_loop.prepare_exit_to_usermode.swapgs_restore_regs_and_return_to_usermode
>>>>
>>>> 4.36 ± 4% +1.3 5.66 ± 7%
>>>> perf-profile.calltrace.cycles-pp.do_signal.exit_to_usermode_loop.prepare_exit_to_usermode.swapgs_restore_regs_and_return_to_usermode
>>>>
>>>> 4.33 ± 4% +1.3 5.62 ± 7%
>>>> perf-profile.calltrace.cycles-pp.get_signal.do_signal.exit_to_usermode_loop.prepare_exit_to_usermode.swapgs_restore_regs_and_return_to_usermode
>>>>
>>>> 4.44 ± 4% +1.3 5.74 ± 7%
>>>> perf-profile.calltrace.cycles-pp.swapgs_restore_regs_and_return_to_usermode
>>>>
>>>> 3.20 ± 10% -2.4 0.78 ±156%
>>>> perf-profile.children.cycles-pp.copy_page
>>>> 0.16 ± 9% -0.1 0.08 ± 64%
>>>> perf-profile.children.cycles-pp.irq_work_interrupt
>>>> 0.16 ± 9% -0.1 0.08 ± 64%
>>>> perf-profile.children.cycles-pp.smp_irq_work_interrupt
>>>> 0.24 ± 5% -0.1 0.17 ± 18%
>>>> perf-profile.children.cycles-pp.irq_work_run_list
>>>> 0.16 ± 9% -0.1 0.10 ± 24%
>>>> perf-profile.children.cycles-pp.irq_work_run
>>>> 0.16 ± 9% -0.1 0.10 ± 24%
>>>> perf-profile.children.cycles-pp.printk
>>>> 0.23 ± 6% -0.1 0.17 ± 9%
>>>> perf-profile.children.cycles-pp.__do_execve_file
>>>> 0.08 ± 14% -0.1 0.03 ±100%
>>>> perf-profile.children.cycles-pp.delay_tsc
>>>> 0.16 ± 6% -0.1 0.11 ± 9%
>>>> perf-profile.children.cycles-pp.load_elf_binary
>>>> 0.16 ± 7% -0.0 0.12 ± 13%
>>>> perf-profile.children.cycles-pp.search_binary_handler
>>>> 0.20 ± 7% -0.0 0.15 ± 10%
>>>> perf-profile.children.cycles-pp.call_usermodehelper_exec_async
>>>> 0.19 ± 6% -0.0 0.15 ± 11%
>>>> perf-profile.children.cycles-pp.do_execve
>>>> 0.08 ± 10% -0.0 0.04 ± 59%
>>>> perf-profile.children.cycles-pp.__vunmap
>>>> 0.15 ± 3% -0.0 0.11 ± 7%
>>>> perf-profile.children.cycles-pp.rcu_idle_exit
>>>> 0.12 ± 10% -0.0 0.09 ± 14%
>>>> perf-profile.children.cycles-pp.__switch_to_asm
>>>> 0.09 ± 13% -0.0 0.07 ± 5%
>>>> perf-profile.children.cycles-pp.des3_ede_encrypt
>>>> 0.06 ± 11% +0.0 0.09 ± 13%
>>>> perf-profile.children.cycles-pp.mark_page_accessed
>>>> 0.15 ± 5% +0.0 0.19 ± 12%
>>>> perf-profile.children.cycles-pp.apparmor_cred_prepare
>>>> 0.22 ± 8% +0.0 0.27 ± 11%
>>>> perf-profile.children.cycles-pp.mem_cgroup_throttle_swaprate
>>>> 0.17 ± 2% +0.0 0.22 ± 12%
>>>> perf-profile.children.cycles-pp.security_prepare_creds
>>>> 0.95 ± 17% +0.3 1.22 ± 14%
>>>> perf-profile.children.cycles-pp.filemap_map_pages
>>>> 5.92 ± 8% +0.7 6.65 ± 7%
>>>> perf-profile.children.cycles-pp.get_signal
>>>> 5.66 ± 9% +0.8 6.44 ± 7%
>>>> perf-profile.children.cycles-pp.mmput
>>>> 5.65 ± 9% +0.8 6.43 ± 7%
>>>> perf-profile.children.cycles-pp.exit_mmap
>>>> 4.40 ± 4% +1.3 5.70 ± 7%
>>>> perf-profile.children.cycles-pp.prepare_exit_to_usermode
>>>> 4.45 ± 4% +1.3 5.75 ± 7%
>>>> perf-profile.children.cycles-pp.swapgs_restore_regs_and_return_to_usermode
>>>>
>>>> 3.16 ± 10% -2.4 0.77 ±155%
>>>> perf-profile.self.cycles-pp.copy_page
>>>> 0.08 ± 14% -0.1 0.03 ±100%
>>>> perf-profile.self.cycles-pp.delay_tsc
>>>> 0.12 ± 10% -0.0 0.09 ± 14%
>>>> perf-profile.self.cycles-pp.__switch_to_asm
>>>> 0.08 ± 12% -0.0 0.06 ± 17%
>>>> perf-profile.self.cycles-pp.enqueue_task_fair
>>>> 0.09 ± 13% -0.0 0.07 ± 5%
>>>> perf-profile.self.cycles-pp.des3_ede_encrypt
>>>> 0.07 ± 13% +0.0 0.08 ± 19%
>>>> perf-profile.self.cycles-pp.__lru_cache_add
>>>> 0.19 ± 9% +0.0 0.22 ± 10%
>>>> perf-profile.self.cycles-pp.mem_cgroup_throttle_swaprate
>>>> 0.15 ± 5% +0.0 0.19 ± 11%
>>>> perf-profile.self.cycles-pp.apparmor_cred_prepare
>>>> 0.05 ± 58% +0.0 0.09 ± 13%
>>>> perf-profile.self.cycles-pp.mark_page_accessed
>>>> 0.58 ± 10% +0.2 0.80 ± 20%
>>>> perf-profile.self.cycles-pp.release_pages
>>>> 0.75 ±173% +1.3e+05% 1005 ±100%
>>>> interrupts.127:PCI-MSI.31981660-edge.i40e-eth0-TxRx-91
>>>> 820.75 ±111% -99.9% 0.50 ±173%
>>>> interrupts.47:PCI-MSI.31981580-edge.i40e-eth0-TxRx-11
>>>> 449.25 ± 86% -100.0% 0.00
>>>> interrupts.53:PCI-MSI.31981586-edge.i40e-eth0-TxRx-17
>>>> 33.25 ±157% -100.0% 0.00
>>>> interrupts.57:PCI-MSI.31981590-edge.i40e-eth0-TxRx-21
>>>> 0.75 ±110% +63533.3% 477.25 ±162%
>>>> interrupts.61:PCI-MSI.31981594-edge.i40e-eth0-TxRx-25
>>>> 561.50 ±160% -100.0% 0.00
>>>> interrupts.65:PCI-MSI.31981598-edge.i40e-eth0-TxRx-29
>>>> 82921 ± 8% -11.1% 73748 ± 6%
>>>> interrupts.CPU11.CAL:Function_call_interrupts
>>>> 66509 ± 30% -32.6% 44828 ± 8%
>>>> interrupts.CPU14.TLB:TLB_shootdowns
>>>> 43105 ± 98% -90.3% 4183 ± 21%
>>>> interrupts.CPU17.RES:Rescheduling_interrupts
>>>> 148719 ± 70% -69.4% 45471 ± 16%
>>>> interrupts.CPU17.TLB:TLB_shootdowns
>>>> 85589 ± 42% -52.2% 40884 ± 5%
>>>> interrupts.CPU20.TLB:TLB_shootdowns
>>>> 222472 ± 41% -98.0% 4360 ± 45%
>>>> interrupts.CPU22.RES:Rescheduling_interrupts
>>>> 0.50 ±173% +95350.0% 477.25 ±162%
>>>> interrupts.CPU25.61:PCI-MSI.31981594-edge.i40e-eth0-TxRx-25
>>>> 76029 ± 10% +14.9% 87389 ± 5%
>>>> interrupts.CPU25.CAL:Function_call_interrupts
>>>> 399042 ± 6% +13.4% 452479 ± 8%
>>>> interrupts.CPU27.LOC:Local_timer_interrupts
>>>> 561.00 ±161% -100.0% 0.00
>>>> interrupts.CPU29.65:PCI-MSI.31981598-edge.i40e-eth0-TxRx-29
>>>> 7034 ± 46% +1083.8% 83279 ±138%
>>>> interrupts.CPU29.RES:Rescheduling_interrupts
>>>> 17829 ± 99% -71.0% 5172 ± 16%
>>>> interrupts.CPU30.RES:Rescheduling_interrupts
>>>> 5569 ± 15% +2414.7% 140059 ± 94%
>>>> interrupts.CPU31.RES:Rescheduling_interrupts
>>>> 37674 ± 16% +36.6% 51473 ± 25%
>>>> interrupts.CPU31.TLB:TLB_shootdowns
>>>> 47905 ± 39% +76.6% 84583 ± 38%
>>>> interrupts.CPU34.TLB:TLB_shootdowns
>>>> 568.75 ±140% +224.8% 1847 ± 90%
>>>> interrupts.CPU36.NMI:Non-maskable_interrupts
>>>> 568.75 ±140% +224.8% 1847 ± 90%
>>>> interrupts.CPU36.PMI:Performance_monitoring_interrupts
>>>> 4236 ± 25% +2168.5% 96092 ± 90%
>>>> interrupts.CPU36.RES:Rescheduling_interrupts
>>>> 52717 ± 27% +43.3% 75565 ± 28%
>>>> interrupts.CPU37.TLB:TLB_shootdowns
>>>> 41418 ± 9% +136.6% 98010 ± 50%
>>>> interrupts.CPU39.TLB:TLB_shootdowns
>>>> 5551 ± 8% +847.8% 52615 ± 66%
>>>> interrupts.CPU40.RES:Rescheduling_interrupts
>>>> 4746 ± 25% +865.9% 45841 ± 91%
>>>> interrupts.CPU42.RES:Rescheduling_interrupts
>>>> 37556 ± 11% +24.6% 46808 ± 6%
>>>> interrupts.CPU42.TLB:TLB_shootdowns
>>>> 21846 ±124% -84.4% 3415 ± 46%
>>>> interrupts.CPU48.RES:Rescheduling_interrupts
>>>> 891.50 ± 22% -35.2% 577.25 ± 40%
>>>> interrupts.CPU49.NMI:Non-maskable_interrupts
>>>> 891.50 ± 22% -35.2% 577.25 ± 40%
>>>> interrupts.CPU49.PMI:Performance_monitoring_interrupts
>>>> 20459 ±120% -79.2% 4263 ± 14%
>>>> interrupts.CPU49.RES:Rescheduling_interrupts
>>>> 59840 ± 21% -23.1% 46042 ± 16%
>>>> interrupts.CPU5.TLB:TLB_shootdowns
>>>> 65200 ± 19% -34.5% 42678 ± 9%
>>>> interrupts.CPU51.TLB:TLB_shootdowns
>>>> 70923 ±153% -94.0% 4270 ± 29%
>>>> interrupts.CPU53.RES:Rescheduling_interrupts
>>>> 65312 ± 22% -28.7% 46578 ± 14%
>>>> interrupts.CPU56.TLB:TLB_shootdowns
>>>> 65828 ± 24% �� -33.4% 43846 ± 4%
>>>> interrupts.CPU59.TLB:TLB_shootdowns
>>>> 72558 ±156% -93.2% 4906 ± 9%
>>>> interrupts.CPU6.RES:Rescheduling_interrupts
>>>> 68698 ± 34% -32.6% 46327 ± 18%
>>>> interrupts.CPU61.TLB:TLB_shootdowns
>>>> 109745 ± 44% -57.4% 46711 ± 16%
>>>> interrupts.CPU62.TLB:TLB_shootdowns
>>>> 89714 ± 44% -48.5% 46198 ± 7%
>>>> interrupts.CPU63.TLB:TLB_shootdowns
>>>> 59380 ±136% -91.5% 5066 ± 13%
>>>> interrupts.CPU69.RES:Rescheduling_interrupts
>>>> 40094 ± 18% +133.9% 93798 ± 44%
>>>> interrupts.CPU78.TLB:TLB_shootdowns
>>>> 129884 ± 72% -55.3% 58034 ±157%
>>>> interrupts.CPU8.RES:Rescheduling_interrupts
>>>> 69984 ± 11% +51.4% 105957 ± 20%
>>>> interrupts.CPU80.CAL:Function_call_interrupts
>>>> 32857 ± 10% +128.7% 75131 ± 36%
>>>> interrupts.CPU80.TLB:TLB_shootdowns
>>>> 35726 ± 16% +34.6% 48081 ± 12%
>>>> interrupts.CPU82.TLB:TLB_shootdowns
>>>> 73820 ± 17% +28.2% 94643 ± 8%
>>>> interrupts.CPU84.CAL:Function_call_interrupts
>>>> 38829 ± 28% +190.3% 112736 ± 42%
>>>> interrupts.CPU84.TLB:TLB_shootdowns
>>>> 36129 ± 4% +47.6% 53329 ± 13%
>>>> interrupts.CPU85.TLB:TLB_shootdowns
>>>> 4693 ± 7% +1323.0% 66793 ±145%
>>>> interrupts.CPU86.RES:Rescheduling_interrupts
>>>> 38003 ± 11% +94.8% 74031 ± 43%
>>>> interrupts.CPU86.TLB:TLB_shootdowns
>>>> 78022 ± 3% +7.9% 84210 ± 3%
>>>> interrupts.CPU87.CAL:Function_call_interrupts
>>>> 36359 ± 6% +54.9% 56304 ± 48%
>>>> interrupts.CPU88.TLB:TLB_shootdowns
>>>> 89031 ±105% -95.0% 4475 ± 40%
>>>> interrupts.CPU9.RES:Rescheduling_interrupts
>>>> 40085 ± 11% +60.6% 64368 ± 27%
>>>> interrupts.CPU91.TLB:TLB_shootdowns
>>>> 42244 ± 10% +44.8% 61162 ± 35%
>>>> interrupts.CPU94.TLB:TLB_shootdowns
>>>> 40959 ± 15% +109.4% 85780 ± 41%
>>>> interrupts.CPU95.TLB:TLB_shootdowns
>>>>
>>>>
>>>> stress-ng.fiemap.ops
>>>> 80000
>>>> +-------------------------------------------------------------------+
>>>> 75000 |..+. .+.. .+..+.. .+.
>>>> .+.. |
>>>> | +..+..+..+.+. .+..+.. .+ +. +.
>>>> +.+..+..+..+.+..|
>>>> 70000 |-+ +
>>>> +. |
>>>> 65000
>>>> |-+ |
>>>> 60000
>>>> |-+ |
>>>> 55000
>>>> |-+ |
>>>> | |
>>>> 50000
>>>> |-+ |
>>>> 45000
>>>> |-+ |
>>>> 40000
>>>> |-+ |
>>>> 35000 |-+ O |
>>>> | O O O O
>>>> O |
>>>> 30000 |-+ O O O O O O O O O O O O O O
>>>> O O O |
>>>> 25000
>>>> +-------------------------------------------------------------------+
>>>> stress-ng.fiemap.ops_per_sec
>>>> 80000
>>>> +-------------------------------------------------------------------+
>>>> 75000 |.. .+.. .+.. |
>>>> | +. .+..+..+.+. .+..+.. .+.+.
>>>> +..+.+..+..+.+..+..+..+.+..|
>>>> 70000 |-+ +. +
>>>> +. |
>>>> 65000
>>>> |-+ |
>>>> 60000
>>>> |-+ |
>>>> 55000
>>>> |-+ |
>>>> | |
>>>> 50000
>>>> |-+ |
>>>> 45000
>>>> |-+ |
>>>> 40000
>>>> |-+ |
>>>> 35000 |-+ O |
>>>> | O O O
>>>> O |
>>>> 30000 |-+ O O O O O O O O O O O O O
>>>> O O O |
>>>> 25000
>>>> +-------------------------------------------------------------------+
>>>> [*] bisect-good sample
>>>> [O] bisect-bad sample
>>>>
>>>>
>>>>
>>>> Disclaimer:
>>>> Results have been estimated based on internal Intel analysis and are
>>>> provided
>>>> for informational purposes only. Any difference in system hardware
>>>> or software
>>>> design or configuration may affect actual performance.
>>>>
>>>>
>>>> Thanks,
>>>> Rong Chen
>>>>
>>>>
>>>> _______________________________________________
>>>> LKP mailing list -- lkp@...ts.01.org
>>>> To unsubscribe send an email to lkp-leave@...ts.01.org
>>>>
>>>
>>
>
--
Zhengjun Xing
Powered by blists - more mailing lists