[<prev] [next>] [day] [month] [year] [list]
Message-ID: <20210531141909.GA19669@xsang-OptiPlex-9020>
Date: Mon, 31 May 2021 22:19:09 +0800
From: kernel test robot <oliver.sang@...el.com>
To: chenxiang <chenxiang66@...ilicon.com>
Cc: 0day robot <lkp@...el.com>, LKML <linux-kernel@...r.kernel.org>,
lkp@...ts.01.org, ying.huang@...el.com, feng.tang@...el.com,
zhengjun.xing@...ux.intel.com, axboe@...nel.dk, tj@...nel.org,
martin.petersen@...cle.com, linux-ide@...r.kernel.org,
linux-scsi@...r.kernel.org, linuxarm@...neuler.org,
linuxarm@...wei.com, Xiang Chen <chenxiang66@...ilicon.com>
Subject: [libata] 2c76f9f255: stress-ng.copy-file.ops_per_sec -25.9%
regression
Greeting,
FYI, we noticed a -25.9% regression of stress-ng.copy-file.ops_per_sec due to commit:
commit: 2c76f9f255f01743b65a16b667355452a1f69b99 ("[PATCH] libata: configure max sectors properly")
url: https://github.com/0day-ci/linux/commits/chenxiang/libata-configure-max-sectors-properly/20210526-094051
base: https://git.kernel.org/cgit/linux/kernel/git/axboe/linux-block.git for-next
in testcase: stress-ng
on test machine: 96 threads 2 sockets Intel(R) Xeon(R) Gold 6252 CPU @ 2.10GHz with 512G memory
with following parameters:
nr_threads: 10%
disk: 1HDD
testtime: 60s
fs: f2fs
class: filesystem
test: copy-file
cpufreq_governor: performance
ucode: 0x5003006
In addition to that, the commit also has significant impact on the following tests:
+------------------+--------------------------------------------------------------------------------+
| testcase: change | stress-ng: stress-ng.readahead.ops_per_sec 71.5% improvement |
| test machine | 96 threads 2 sockets Intel(R) Xeon(R) Gold 6252 CPU @ 2.10GHz with 192G memory |
| test parameters | class=os |
| | cpufreq_governor=performance |
| | disk=1HDD |
| | fs=ext4 |
| | nr_threads=10% |
| | test=readahead |
| | testtime=60s |
| | ucode=0x5003006 |
+------------------+--------------------------------------------------------------------------------+
If you fix the issue, kindly add following tag
Reported-by: kernel test robot <oliver.sang@...el.com>
Details are as below:
-------------------------------------------------------------------------------------------------->
To reproduce:
git clone https://github.com/intel/lkp-tests.git
cd lkp-tests
bin/lkp install job.yaml # job file is attached in this email
bin/lkp split-job --compatible job.yaml # generate the yaml file for lkp run
bin/lkp run generated-yaml-file
=========================================================================================
class/compiler/cpufreq_governor/disk/fs/kconfig/nr_threads/rootfs/tbox_group/test/testcase/testtime/ucode:
filesystem/gcc-9/performance/1HDD/f2fs/x86_64-rhel-8.3/10%/debian-10.4-x86_64-20200603.cgz/lkp-csl-2sp7/copy-file/stress-ng/60s/0x5003006
commit:
6bcec6cee5 ("Merge branch 'for-5.14/io_uring' into for-next")
2c76f9f255 ("libata: configure max sectors properly")
6bcec6cee54edf7e 2c76f9f255f01743b65a16b6673
---------------- ---------------------------
%stddev %change %stddev
\ | \
2882 -25.9% 2136 stress-ng.copy-file.ops
47.84 -25.9% 35.43 stress-ng.copy-file.ops_per_sec
11911328 -25.7% 8848424 stress-ng.time.file_system_outputs
16.67 ± 4% -40.0% 10.00 stress-ng.time.percent_of_cpu_this_job_got
10.63 ± 3% -38.4% 6.55 ± 3% stress-ng.time.system_time
115088 ± 7% -38.6% 70635 ± 47% stress-ng.time.voluntary_context_switches
7482 ± 15% +28.7% 9626 ± 5% softirqs.CPU50.SCHED
94.62 +1.4% 95.90 iostat.cpu.idle
3.29 ± 8% -36.5% 2.09 ± 8% iostat.cpu.iowait
3.12 ± 8% -1.1 1.99 ± 8% mpstat.cpu.all.iowait%
0.22 ± 5% -0.1 0.15 ± 7% mpstat.cpu.all.sys%
1765 ± 8% +70.9% 3017 ± 9% slabinfo.dmaengine-unmap-16.active_objs
1790 ± 8% +68.6% 3019 ± 9% slabinfo.dmaengine-unmap-16.num_objs
661211 ± 10% -46.4% 354112 ± 56% numa-meminfo.node0.Active
659853 ± 10% -46.5% 352767 ± 56% numa-meminfo.node0.Active(file)
4128 ± 13% -78.1% 903.00 ± 58% numa-meminfo.node0.Writeback
94.17 +1.5% 95.57 vmstat.cpu.id
88096 ± 8% -30.0% 61677 ± 8% vmstat.io.bo
2916956 ± 4% -8.7% 2662410 ± 4% vmstat.memory.cache
5184 ± 9% -29.0% 3679 ± 24% vmstat.system.cs
669025 ± 8% -25.4% 499071 ± 8% meminfo.Active
666140 ± 8% -25.5% 496122 ± 8% meminfo.Active(file)
2853772 ± 5% -9.1% 2593698 ± 5% meminfo.Cached
4908 ± 19% +61.2% 7909 ± 9% meminfo.Dirty
4300 ± 15% -64.2% 1540 ± 44% meminfo.Writeback
162504 ± 10% -46.5% 86958 ± 56% numa-vmstat.node0.nr_active_file
782687 ± 5% -40.6% 464832 ± 58% numa-vmstat.node0.nr_dirtied
1016 ± 17% -83.9% 163.29 ± 60% numa-vmstat.node0.nr_writeback
781002 ± 5% -40.6% 463755 ± 58% numa-vmstat.node0.nr_written
162504 ± 10% -46.5% 86958 ± 56% numa-vmstat.node0.nr_zone_active_file
1682 ± 9% -36.2% 1072 ± 55% numa-vmstat.node0.nr_zone_write_pending
12554 +41.9% 17815 interrupts.315:PCI-MSI.376832-edge.ahci[0000:00:17.0]
114120 ± 4% -22.4% 88521 ± 22% interrupts.CAL:Function_call_interrupts
3736 ± 80% -83.7% 610.57 ± 22% interrupts.CPU1.CAL:Function_call_interrupts
146.67 ± 38% -27.8% 105.86 ± 40% interrupts.CPU15.NMI:Non-maskable_interrupts
146.67 ± 38% -27.8% 105.86 ± 40% interrupts.CPU15.PMI:Performance_monitoring_interrupts
159.00 ± 39% -41.8% 92.57 ± 42% interrupts.CPU55.NMI:Non-maskable_interrupts
159.00 ± 39% -41.8% 92.57 ± 42% interrupts.CPU55.PMI:Performance_monitoring_interrupts
101.33 ± 21% +47.6% 149.57 ± 17% interrupts.CPU66.NMI:Non-maskable_interrupts
101.33 ± 21% +47.6% 149.57 ± 17% interrupts.CPU66.PMI:Performance_monitoring_interrupts
166540 ± 8% -25.4% 124177 ± 8% proc-vmstat.nr_active_file
1495731 -25.7% 1111429 proc-vmstat.nr_dirtied
1159 ± 12% +69.6% 1967 ± 10% proc-vmstat.nr_dirty
712459 ± 5% -9.0% 648586 ± 5% proc-vmstat.nr_file_pages
1128 ± 12% -69.6% 342.71 ± 21% proc-vmstat.nr_writeback
1495710 -25.7% 1111195 proc-vmstat.nr_written
166540 ± 8% -25.4% 124177 ± 8% proc-vmstat.nr_zone_active_file
279115 -9.3% 253162 proc-vmstat.pgactivate
5983119 -25.7% 4445141 proc-vmstat.pgpgout
3.115e+08 ± 10% -15.5% 2.631e+08 ± 14% perf-stat.i.branch-instructions
5107 ± 9% -30.0% 3573 ± 26% perf-stat.i.context-switches
3.829e+08 ± 7% -13.9% 3.299e+08 ± 10% perf-stat.i.dTLB-loads
1.881e+08 ± 5% -11.1% 1.672e+08 ± 7% perf-stat.i.dTLB-stores
1.529e+09 ± 10% -15.6% 1.291e+09 ± 14% perf-stat.i.instructions
1801 ± 8% -13.7% 1554 ± 12% perf-stat.i.instructions-per-iTLB-miss
9.19 ± 8% -13.9% 7.91 ± 11% perf-stat.i.metric.M/sec
483599 ± 10% -35.8% 310368 ± 9% perf-stat.i.node-loads
389216 ± 11% -33.0% 260605 ± 9% perf-stat.i.node-stores
1834 ± 9% -13.6% 1585 ± 14% perf-stat.overall.instructions-per-iTLB-miss
3.069e+08 ± 10% -15.5% 2.593e+08 ± 14% perf-stat.ps.branch-instructions
5029 ± 9% -30.0% 3522 ± 26% perf-stat.ps.context-switches
3.772e+08 ± 7% -13.8% 3.251e+08 ± 10% perf-stat.ps.dTLB-loads
1.852e+08 ± 5% -11.0% 1.648e+08 ± 7% perf-stat.ps.dTLB-stores
1.506e+09 ± 10% -15.6% 1.272e+09 ± 14% perf-stat.ps.instructions
476080 ± 10% -35.8% 305808 ± 9% perf-stat.ps.node-loads
383190 ± 11% -33.0% 256784 ± 9% perf-stat.ps.node-stores
6.03 ± 13% -2.8 3.21 ± 21% perf-profile.calltrace.cycles-pp.__generic_file_write_iter.f2fs_file_write_iter.do_iter_readv_writev.do_iter_write.iter_file_splice_write
6.17 ± 14% -2.8 3.37 ± 22% perf-profile.calltrace.cycles-pp.do_iter_write.iter_file_splice_write.direct_splice_actor.splice_direct_to_actor.do_splice_direct
6.17 ± 14% -2.8 3.37 ± 22% perf-profile.calltrace.cycles-pp.do_iter_readv_writev.do_iter_write.iter_file_splice_write.direct_splice_actor.splice_direct_to_actor
6.01 ± 13% -2.8 3.21 ± 21% perf-profile.calltrace.cycles-pp.generic_perform_write.__generic_file_write_iter.f2fs_file_write_iter.do_iter_readv_writev.do_iter_write
6.17 ± 14% -2.8 3.37 ± 22% perf-profile.calltrace.cycles-pp.f2fs_file_write_iter.do_iter_readv_writev.do_iter_write.iter_file_splice_write.direct_splice_actor
9.91 ± 12% -2.8 7.14 ± 17% perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe
9.91 ± 12% -2.8 7.14 ± 17% perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe
6.23 ± 14% -2.8 3.46 ± 23% perf-profile.calltrace.cycles-pp.direct_splice_actor.splice_direct_to_actor.do_splice_direct.vfs_copy_file_range.__x64_sys_copy_file_range
6.23 ± 14% -2.8 3.46 ± 23% perf-profile.calltrace.cycles-pp.iter_file_splice_write.direct_splice_actor.splice_direct_to_actor.do_splice_direct.vfs_copy_file_range
6.67 ± 13% -2.7 4.00 ± 19% perf-profile.calltrace.cycles-pp.splice_direct_to_actor.do_splice_direct.vfs_copy_file_range.__x64_sys_copy_file_range.do_syscall_64
6.67 ± 13% -2.7 4.00 ± 19% perf-profile.calltrace.cycles-pp.do_splice_direct.vfs_copy_file_range.__x64_sys_copy_file_range.do_syscall_64.entry_SYSCALL_64_after_hwframe
6.67 ± 13% -2.7 4.00 ± 20% perf-profile.calltrace.cycles-pp.vfs_copy_file_range.__x64_sys_copy_file_range.do_syscall_64.entry_SYSCALL_64_after_hwframe
6.67 ± 13% -2.7 4.00 ± 20% perf-profile.calltrace.cycles-pp.__x64_sys_copy_file_range.do_syscall_64.entry_SYSCALL_64_after_hwframe
2.86 ± 20% -1.9 0.94 ± 23% perf-profile.calltrace.cycles-pp.f2fs_write_end.generic_perform_write.__generic_file_write_iter.f2fs_file_write_iter.do_iter_readv_writev
2.71 ± 20% -1.9 0.83 ± 24% perf-profile.calltrace.cycles-pp.f2fs_set_data_page_dirty.f2fs_write_end.generic_perform_write.__generic_file_write_iter.f2fs_file_write_iter
2.65 ± 13% -0.9 1.75 ± 23% perf-profile.calltrace.cycles-pp.iov_iter_copy_from_user_atomic.generic_perform_write.__generic_file_write_iter.f2fs_file_write_iter.do_iter_readv_writev
2.12 ± 16% -0.5 1.63 ± 13% perf-profile.calltrace.cycles-pp.f2fs_write_single_data_page.f2fs_write_cache_pages.f2fs_write_data_pages.do_writepages.__filemap_fdatawrite_range
1.98 ± 15% -0.4 1.53 ± 14% perf-profile.calltrace.cycles-pp.f2fs_do_write_data_page.f2fs_write_single_data_page.f2fs_write_cache_pages.f2fs_write_data_pages.do_writepages
0.98 ± 22% -0.3 0.72 ± 9% perf-profile.calltrace.cycles-pp.f2fs_outplace_write_data.f2fs_do_write_data_page.f2fs_write_single_data_page.f2fs_write_cache_pages.f2fs_write_data_pages
0.75 ± 11% -0.2 0.58 ± 8% perf-profile.calltrace.cycles-pp.do_write_page.f2fs_outplace_write_data.f2fs_do_write_data_page.f2fs_write_single_data_page.f2fs_write_cache_pages
86.95 +2.6 89.54 perf-profile.calltrace.cycles-pp.do_idle.cpu_startup_entry.start_secondary.secondary_startup_64_no_verify
6.02 ± 13% -2.8 3.21 ± 21% perf-profile.children.cycles-pp.generic_perform_write
6.03 ± 13% -2.8 3.21 ± 21% perf-profile.children.cycles-pp.__generic_file_write_iter
6.17 ± 14% -2.8 3.37 ± 22% perf-profile.children.cycles-pp.do_iter_write
6.17 ± 14% -2.8 3.37 ± 22% perf-profile.children.cycles-pp.do_iter_readv_writev
6.17 ± 14% -2.8 3.37 ± 22% perf-profile.children.cycles-pp.f2fs_file_write_iter
6.23 ± 14% -2.8 3.46 ± 23% perf-profile.children.cycles-pp.direct_splice_actor
6.23 ± 14% -2.8 3.46 ± 23% perf-profile.children.cycles-pp.iter_file_splice_write
10.70 ± 11% -2.7 7.98 ± 15% perf-profile.children.cycles-pp.entry_SYSCALL_64_after_hwframe
10.69 ± 11% -2.7 7.97 ± 16% perf-profile.children.cycles-pp.do_syscall_64
6.67 ± 13% -2.7 4.00 ± 19% perf-profile.children.cycles-pp.splice_direct_to_actor
6.67 ± 13% -2.7 4.00 ± 19% perf-profile.children.cycles-pp.do_splice_direct
6.67 ± 13% -2.7 4.00 ± 20% perf-profile.children.cycles-pp.vfs_copy_file_range
6.67 ± 13% -2.7 4.00 ± 20% perf-profile.children.cycles-pp.__x64_sys_copy_file_range
2.87 ± 20% -1.9 0.94 ± 23% perf-profile.children.cycles-pp.f2fs_write_end
2.71 ± 20% -1.9 0.83 ± 24% perf-profile.children.cycles-pp.f2fs_set_data_page_dirty
2.13 ± 24% -1.8 0.34 ± 18% perf-profile.children.cycles-pp.f2fs_update_dirty_page
2.56 ± 19% -1.6 1.00 ± 15% perf-profile.children.cycles-pp._raw_spin_lock
1.59 ± 30% -1.5 0.10 ± 29% perf-profile.children.cycles-pp.native_queued_spin_lock_slowpath
2.66 ± 13% -0.9 1.75 ± 23% perf-profile.children.cycles-pp.iov_iter_copy_from_user_atomic
2.13 ± 16% -0.5 1.63 ± 13% perf-profile.children.cycles-pp.f2fs_write_single_data_page
1.98 ± 15% -0.4 1.53 ± 14% perf-profile.children.cycles-pp.f2fs_do_write_data_page
0.98 ± 23% -0.3 0.72 ± 9% perf-profile.children.cycles-pp.f2fs_outplace_write_data
0.76 ± 12% -0.2 0.58 ± 8% perf-profile.children.cycles-pp.do_write_page
0.51 ± 7% -0.1 0.37 ± 7% perf-profile.children.cycles-pp.f2fs_allocate_data_block
0.14 ± 16% -0.0 0.10 ± 25% perf-profile.children.cycles-pp.___might_sleep
0.04 ± 71% +0.1 0.09 ± 19% perf-profile.children.cycles-pp.ksys_mmap_pgoff
0.05 ± 75% +0.1 0.10 ± 19% perf-profile.children.cycles-pp.add_to_page_cache_lru
0.05 ± 74% +0.1 0.11 ± 17% perf-profile.children.cycles-pp.mmap_region
0.01 ±223% +0.1 0.07 ± 21% perf-profile.children.cycles-pp.__add_to_page_cache_locked
0.06 ± 50% +0.1 0.12 ± 16% perf-profile.children.cycles-pp.do_mmap
0.02 ±149% +0.3 0.29 ± 38% perf-profile.children.cycles-pp.mutex_spin_on_owner
0.13 ± 91% +0.5 0.63 ± 39% perf-profile.children.cycles-pp.__mutex_lock
1.47 ± 27% -1.4 0.10 ± 29% perf-profile.self.cycles-pp.native_queued_spin_lock_slowpath
2.58 ± 14% -0.9 1.66 ± 21% perf-profile.self.cycles-pp.iov_iter_copy_from_user_atomic
0.32 ± 16% -0.2 0.07 ± 14% perf-profile.self.cycles-pp.f2fs_update_dirty_page
0.13 ± 16% -0.1 0.07 ± 71% perf-profile.self.cycles-pp.f2fs_do_write_data_page
0.14 ± 17% -0.0 0.10 ± 28% perf-profile.self.cycles-pp.___might_sleep
0.02 ±149% +0.3 0.29 ± 38% perf-profile.self.cycles-pp.mutex_spin_on_owner
stress-ng.time.percent_of_cpu_this_job_got
18 +----------------------------------------------------------------------+
| + : :: + : +: |
17 |-+ + +. .+ +.+ +.++.++. +. .+ +.+ + + .++ +. + :.+|
16 |+.+ +++ + ++.++.+ + ++ ++ +.+ + : + + + |
| + : : : |
15 |-+ + + |
| |
14 |-+ |
| |
13 |-+ |
12 |-+ |
| |
11 |-+O O |
| |
10 +----------------------------------------------------------------------+
stress-ng.time.file_system_outputs
1.25e+07 +----------------------------------------------------------------+
| .+ |
1.2e+07 |+.+++.+ ++ ++.+++.++++.+++.+++.++++.+++.+++.++++.+++.+++.++++.+|
1.15e+07 |-+ + |
| |
1.1e+07 |-+ |
| |
1.05e+07 |-+ |
| |
1e+07 |-+ |
9.5e+06 |-+ |
| |
9e+06 |-+OO OO O O O O O O |
|O O OO O O OO O O OO O |
8.5e+06 +----------------------------------------------------------------+
stress-ng.copy-file.ops
3000 +--------------------------------------------------------------------+
2900 |+. +.+ .+ +. +. + +.+ ++. +.+ +.+ +.+ .++ .++ .+ .+ .+|
| + ++ + + + +.++ +.+ + + + + + +.++.+++ ++ + |
2800 |-+ |
2700 |-+ |
| |
2600 |-+ |
2500 |-+ |
2400 |-+ |
| |
2300 |-+ |
2200 |-+O |
|O O OOO OOO O O OOO O OO OO |
2100 |-+ OO O O O |
2000 +--------------------------------------------------------------------+
stress-ng.copy-file.ops_per_sec
50 +----------------------------------------------------------------------+
|+. .+ .+ .+ +. +.+ .+ .+ + +. .+ |
48 |-+++ +.+++ + + ++.+ ++.++.++ ++ +.+ +.+ ++ ++.++.+++.++.++.+|
46 |-+ |
| |
44 |-+ |
| |
42 |-+ |
| |
40 |-+ |
38 |-+ |
| |
36 |-+O OO OO O OO O |
|O O OO O O OOO O OOO O |
34 +----------------------------------------------------------------------+
[*] bisect-good sample
[O] bisect-bad sample
***************************************************************************************************
lkp-csl-2sp5: 96 threads 2 sockets Intel(R) Xeon(R) Gold 6252 CPU @ 2.10GHz with 192G memory
=========================================================================================
class/compiler/cpufreq_governor/disk/fs/kconfig/nr_threads/rootfs/tbox_group/test/testcase/testtime/ucode:
os/gcc-9/performance/1HDD/ext4/x86_64-rhel-8.3/10%/debian-10.4-x86_64-20200603.cgz/lkp-csl-2sp5/readahead/stress-ng/60s/0x5003006
commit:
6bcec6cee5 ("Merge branch 'for-5.14/io_uring' into for-next")
2c76f9f255 ("libata: configure max sectors properly")
6bcec6cee54edf7e 2c76f9f255f01743b65a16b6673
---------------- ---------------------------
%stddev %change %stddev
\ | \
1.07e+08 +71.5% 1.836e+08 ± 6% stress-ng.readahead.ops
1783834 +71.5% 3059261 ± 6% stress-ng.readahead.ops_per_sec
479.07 -11.1% 426.10 ± 6% stress-ng.time.system_time
55.77 ± 2% +93.4% 107.84 ± 25% stress-ng.time.user_time
6632490 ±221% -99.4% 42440 ± 11% cpuidle.POLL.time
0.04 ± 8% +20.0% 0.05 ± 13% perf-sched.wait_and_delay.avg.ms.pipe_read.new_sync_read.vfs_read.ksys_read
19597 -1.0% 19407 proc-vmstat.pgactivate
1726 ± 3% -6.0% 1623 vmstat.system.cs
1910 ± 97% -74.5% 487.57 ±198% interrupts.CPU59.NMI:Non-maskable_interrupts
1910 ± 97% -74.5% 487.57 ±198% interrupts.CPU59.PMI:Performance_monitoring_interrupts
8.39 ± 2% -10.6% 7.50 ± 5% iostat.cpu.system
0.92 ± 3% +87.6% 1.72 ± 27% iostat.cpu.user
57046 ± 3% -30.2% 39790 ± 51% numa-vmstat.node0.nr_anon_pages
6475 ± 28% +267.8% 23819 ± 85% numa-vmstat.node1.nr_anon_pages
0.02 ± 4% +0.0 0.03 ± 6% mpstat.cpu.all.iowait%
7.06 ± 2% -0.8 6.22 ± 5% mpstat.cpu.all.sys%
0.87 ± 3% +0.8 1.62 ± 27% mpstat.cpu.all.usr%
97178 ± 9% -38.7% 59552 ± 54% numa-meminfo.node0.AnonHugePages
228163 ± 3% -30.2% 159145 ± 51% numa-meminfo.node0.AnonPages
164183 ± 9% -13.0% 142812 ± 14% numa-meminfo.node0.Slab
5744 ± 18% +515.7% 35368 ± 98% numa-meminfo.node1.AnonHugePages
25887 ± 28% +267.9% 95228 ± 85% numa-meminfo.node1.AnonPages
44270 ± 13% +156.2% 113422 ± 70% numa-meminfo.node1.AnonPages.max
9369 ± 6% -12.5% 8197 ± 20% softirqs.CPU59.SCHED
9695 ± 3% -12.7% 8463 ± 8% softirqs.CPU61.SCHED
9770 ± 4% -9.0% 8887 ± 10% softirqs.CPU66.SCHED
9473 ± 6% -14.5% 8096 ± 11% softirqs.CPU84.SCHED
9441 ± 6% -18.0% 7745 ± 20% softirqs.CPU85.SCHED
12271 ± 17% -15.3% 10392 ± 7% softirqs.TIMER
2.36 ± 21% +76.4% 4.16 ± 29% perf-stat.i.MPKI
1.545e+10 ± 2% -10.5% 1.382e+10 ± 6% perf-stat.i.branch-instructions
13484423 ± 3% +42.6% 19227872 ± 3% perf-stat.i.branch-misses
21671037 ± 12% +56.0% 33811435 ± 18% perf-stat.i.cache-misses
1.164e+08 ± 2% +66.2% 1.934e+08 ± 6% perf-stat.i.cache-references
1499 ± 4% -5.2% 1421 ± 2% perf-stat.i.context-switches
1323 ± 13% -26.6% 971.56 ± 23% perf-stat.i.cycles-between-cache-misses
1.749e+10 -14.4% 1.498e+10 ± 7% perf-stat.i.dTLB-loads
1.073e+10 ± 2% -19.9% 8.6e+09 ± 6% perf-stat.i.dTLB-stores
4054770 ± 6% +61.9% 6566022 ± 10% perf-stat.i.iTLB-load-misses
6.471e+10 ± 2% -13.4% 5.604e+10 ± 6% perf-stat.i.instructions
15495 ± 7% -46.3% 8315 ± 12% perf-stat.i.instructions-per-iTLB-miss
2.37 -13.1% 2.06 ± 6% perf-stat.i.ipc
201.39 ± 13% +43.3% 288.53 ± 13% perf-stat.i.metric.K/sec
456.15 -14.2% 391.60 ± 6% perf-stat.i.metric.M/sec
14406997 ± 14% +58.0% 22762612 ± 19% perf-stat.i.node-loads
111506 ± 28% +93.5% 215774 ± 25% perf-stat.i.node-store-misses
1.80 +91.9% 3.45 perf-stat.overall.MPKI
0.09 ± 2% +0.1 0.14 ± 4% perf-stat.overall.branch-miss-rate%
0.41 +14.7% 0.47 ± 7% perf-stat.overall.cpi
1235 ± 12% -35.5% 796.96 ± 17% perf-stat.overall.cycles-between-cache-misses
69.64 ± 4% +6.5 76.12 ± 3% perf-stat.overall.iTLB-load-miss-rate%
16027 ± 7% -46.1% 8642 ± 12% perf-stat.overall.instructions-per-iTLB-miss
2.46 -12.5% 2.15 ± 6% perf-stat.overall.ipc
1.521e+10 ± 2% -10.5% 1.362e+10 ± 6% perf-stat.ps.branch-instructions
13272497 ± 3% +42.7% 18933267 ± 3% perf-stat.ps.branch-misses
21338470 ± 12% +56.0% 33294227 ± 18% perf-stat.ps.cache-misses
1.146e+08 ± 2% +66.2% 1.905e+08 ± 6% perf-stat.ps.cache-references
1476 ± 4% -5.3% 1399 ± 2% perf-stat.ps.context-switches
1.723e+10 -14.3% 1.476e+10 ± 7% perf-stat.ps.dTLB-loads
1.057e+10 ± 2% -19.9% 8.471e+09 ± 6% perf-stat.ps.dTLB-stores
3992776 ± 6% +62.0% 6467207 ± 10% perf-stat.ps.iTLB-load-misses
6.372e+10 -13.4% 5.52e+10 ± 6% perf-stat.ps.instructions
14185966 ± 14% +58.0% 22414685 ± 19% perf-stat.ps.node-loads
109739 ± 28% +93.7% 212520 ± 25% perf-stat.ps.node-store-misses
4.112e+12 -12.5% 3.599e+12 ± 6% perf-stat.total.instructions
55.43 -11.8 43.67 ± 7% perf-profile.calltrace.cycles-pp.page_cache_ra_unbounded.generic_fadvise.ksys_readahead.do_syscall_64.entry_SYSCALL_64_after_hwframe
55.92 -11.4 44.53 ± 7% perf-profile.calltrace.cycles-pp.generic_fadvise.ksys_readahead.do_syscall_64.entry_SYSCALL_64_after_hwframe
56.20 -11.2 45.04 ± 7% perf-profile.calltrace.cycles-pp.ksys_readahead.do_syscall_64.entry_SYSCALL_64_after_hwframe
39.67 -8.6 31.12 ± 6% perf-profile.calltrace.cycles-pp.xa_load.page_cache_ra_unbounded.generic_fadvise.ksys_readahead.do_syscall_64
30.84 -6.6 24.21 ± 6% perf-profile.calltrace.cycles-pp.xas_load.xa_load.page_cache_ra_unbounded.generic_fadvise.ksys_readahead
64.08 -2.8 61.29 ± 2% perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe
10.02 ± 2% -2.2 7.85 ± 7% perf-profile.calltrace.cycles-pp.read_pages.page_cache_ra_unbounded.generic_fadvise.ksys_readahead.do_syscall_64
7.17 ± 2% -1.5 5.65 ± 7% perf-profile.calltrace.cycles-pp.xas_start.xas_load.xa_load.page_cache_ra_unbounded.generic_fadvise
2.10 ± 3% -0.4 1.69 ± 8% perf-profile.calltrace.cycles-pp.rcu_read_unlock_strict.xa_load.page_cache_ra_unbounded.generic_fadvise.ksys_readahead
0.45 ± 45% +0.4 0.90 ± 7% perf-profile.calltrace.cycles-pp.__entry_text_start
0.00 +0.6 0.62 ± 8% perf-profile.calltrace.cycles-pp.touch_atime.filemap_read.new_sync_read.vfs_read.ksys_pread64
1.16 ± 9% +1.1 2.26 ± 12% perf-profile.calltrace.cycles-pp.filemap_get_read_batch.filemap_get_pages.filemap_read.new_sync_read.vfs_read
1.26 ± 8% +1.2 2.44 ± 11% perf-profile.calltrace.cycles-pp.filemap_get_pages.filemap_read.new_sync_read.vfs_read.ksys_pread64
0.00 +2.4 2.41 ±186% perf-profile.calltrace.cycles-pp.syscall_exit_to_user_mode.do_syscall_64.entry_SYSCALL_64_after_hwframe
3.23 ± 3% +2.5 5.76 ± 9% perf-profile.calltrace.cycles-pp.copy_user_enhanced_fast_string.copyout.copy_page_to_iter.filemap_read.new_sync_read
3.28 ± 3% +2.6 5.86 ± 9% perf-profile.calltrace.cycles-pp.copyout.copy_page_to_iter.filemap_read.new_sync_read.vfs_read
3.61 ± 3% +2.8 6.46 ± 9% perf-profile.calltrace.cycles-pp.copy_page_to_iter.filemap_read.new_sync_read.vfs_read.ksys_pread64
5.93 ± 2% +4.9 10.84 ± 8% perf-profile.calltrace.cycles-pp.filemap_read.new_sync_read.vfs_read.ksys_pread64.do_syscall_64
6.25 ± 2% +5.2 11.43 ± 8% perf-profile.calltrace.cycles-pp.new_sync_read.vfs_read.ksys_pread64.do_syscall_64.entry_SYSCALL_64_after_hwframe
6.92 ± 2% +5.7 12.65 ± 8% perf-profile.calltrace.cycles-pp.vfs_read.ksys_pread64.do_syscall_64.entry_SYSCALL_64_after_hwframe
7.24 ± 2% +6.0 13.25 ± 8% perf-profile.calltrace.cycles-pp.ksys_pread64.do_syscall_64.entry_SYSCALL_64_after_hwframe
55.50 -11.7 43.75 ± 7% perf-profile.children.cycles-pp.page_cache_ra_unbounded
55.94 -11.4 44.56 ± 7% perf-profile.children.cycles-pp.generic_fadvise
56.22 -11.2 45.07 ± 7% perf-profile.children.cycles-pp.ksys_readahead
39.69 -8.6 31.12 ± 7% perf-profile.children.cycles-pp.xa_load
31.04 -6.5 24.55 ± 7% perf-profile.children.cycles-pp.xas_load
64.19 -2.8 61.43 ± 2% perf-profile.children.cycles-pp.do_syscall_64
10.06 ± 2% -2.2 7.90 ± 7% perf-profile.children.cycles-pp.read_pages
7.17 ± 2% -1.4 5.74 ± 7% perf-profile.children.cycles-pp.xas_start
2.14 ± 3% -0.4 1.73 ± 7% perf-profile.children.cycles-pp.rcu_read_unlock_strict
0.06 ± 13% +0.0 0.10 ± 10% perf-profile.children.cycles-pp.syscall_exit_to_user_mode_prepare
0.02 ±141% +0.0 0.06 ± 13% perf-profile.children.cycles-pp.get_next_timer_interrupt
0.07 ± 10% +0.1 0.12 ± 11% perf-profile.children.cycles-pp.__might_sleep
0.08 ± 12% +0.1 0.13 ± 8% perf-profile.children.cycles-pp.___might_sleep
0.07 ± 7% +0.1 0.12 ± 9% perf-profile.children.cycles-pp.__might_fault
0.06 ± 14% +0.1 0.12 ± 5% perf-profile.children.cycles-pp.make_kgid
0.06 ± 8% +0.1 0.11 ± 12% perf-profile.children.cycles-pp.aa_file_perm
0.03 ±100% +0.1 0.08 ± 18% perf-profile.children.cycles-pp.timestamp_truncate
0.00 +0.1 0.06 ± 10% perf-profile.children.cycles-pp.rcu_nocb_flush_deferred_wakeup
0.00 +0.1 0.06 ± 14% perf-profile.children.cycles-pp.generic_file_read_iter
0.00 +0.1 0.07 ± 13% perf-profile.children.cycles-pp.do_page_cache_ra
0.00 +0.1 0.07 ± 10% perf-profile.children.cycles-pp.rcu_all_qs
0.00 +0.1 0.07 ± 9% perf-profile.children.cycles-pp.rw_verify_area
0.08 ± 7% +0.1 0.15 ± 10% perf-profile.children.cycles-pp.make_kuid
0.06 ± 15% +0.1 0.13 ± 11% perf-profile.children.cycles-pp.__cond_resched
0.10 ± 4% +0.1 0.18 ± 13% perf-profile.children.cycles-pp.current_time
0.12 ± 8% +0.1 0.21 ± 11% perf-profile.children.cycles-pp.syscall_enter_from_user_mode
0.14 ± 11% +0.1 0.23 ± 10% perf-profile.children.cycles-pp.mark_page_accessed
0.12 ± 14% +0.1 0.22 ± 8% perf-profile.children.cycles-pp.ext4_file_read_iter
0.11 ± 7% +0.1 0.22 ± 8% perf-profile.children.cycles-pp.map_id_range_down
0.17 ± 8% +0.1 0.30 ± 10% perf-profile.children.cycles-pp.entry_SYSCALL_64_safe_stack
0.34 ± 15% +0.2 0.50 ± 7% perf-profile.children.cycles-pp.scheduler_tick
0.19 ± 9% +0.2 0.34 ± 12% perf-profile.children.cycles-pp.common_file_perm
0.21 ± 3% +0.2 0.37 ± 11% perf-profile.children.cycles-pp.__fsnotify_parent
0.21 ± 7% +0.2 0.39 ± 8% perf-profile.children.cycles-pp.force_page_cache_ra
0.21 ± 6% +0.2 0.39 ± 7% perf-profile.children.cycles-pp.exit_to_user_mode_prepare
0.26 ± 8% +0.2 0.48 ± 11% perf-profile.children.cycles-pp.security_file_permission
0.32 ± 5% +0.2 0.57 ± 9% perf-profile.children.cycles-pp.atime_needs_update
0.35 ± 5% +0.3 0.63 ± 8% perf-profile.children.cycles-pp.touch_atime
0.34 ± 7% +0.3 0.65 ± 9% perf-profile.children.cycles-pp.__fget_light
0.53 ± 7% +0.4 0.90 ± 7% perf-profile.children.cycles-pp.__entry_text_start
0.83 ± 4% +0.7 1.51 ± 5% perf-profile.children.cycles-pp.syscall_return_via_sysret
1.17 ± 8% +1.1 2.28 ± 12% perf-profile.children.cycles-pp.filemap_get_read_batch
1.27 ± 8% +1.2 2.46 ± 11% perf-profile.children.cycles-pp.filemap_get_pages
0.35 ± 4% +2.1 2.48 ±180% perf-profile.children.cycles-pp.syscall_exit_to_user_mode
3.26 ± 3% +2.6 5.82 ± 9% perf-profile.children.cycles-pp.copy_user_enhanced_fast_string
3.28 ± 3% +2.6 5.86 ± 9% perf-profile.children.cycles-pp.copyout
3.62 ± 3% +2.9 6.49 ± 9% perf-profile.children.cycles-pp.copy_page_to_iter
5.96 ± 2% +4.9 10.90 ± 8% perf-profile.children.cycles-pp.filemap_read
6.26 ± 2% +5.2 11.45 ± 8% perf-profile.children.cycles-pp.new_sync_read
6.95 ± 2% +5.7 12.69 ± 8% perf-profile.children.cycles-pp.vfs_read
7.25 ± 2% +6.0 13.25 ± 8% perf-profile.children.cycles-pp.ksys_pread64
23.77 -5.0 18.73 ± 7% perf-profile.self.cycles-pp.xas_load
9.96 ± 2% -2.2 7.78 ± 7% perf-profile.self.cycles-pp.read_pages
8.74 ± 2% -1.9 6.82 ± 7% perf-profile.self.cycles-pp.xa_load
6.03 ± 2% -1.2 4.84 ± 7% perf-profile.self.cycles-pp.xas_start
5.77 ± 2% -1.0 4.78 ± 9% perf-profile.self.cycles-pp.page_cache_ra_unbounded
1.09 ± 2% -0.2 0.88 ± 6% perf-profile.self.cycles-pp.rcu_read_unlock_strict
0.06 ± 11% +0.0 0.10 ± 11% perf-profile.self.cycles-pp.__might_sleep
0.06 ± 14% +0.0 0.11 ± 9% perf-profile.self.cycles-pp.security_file_permission
0.07 ± 8% +0.1 0.12 ± 11% perf-profile.self.cycles-pp.___might_sleep
0.09 ± 16% +0.1 0.14 ± 14% perf-profile.self.cycles-pp.atime_needs_update
0.03 ±100% +0.1 0.08 ± 10% perf-profile.self.cycles-pp.syscall_exit_to_user_mode_prepare
0.08 ± 8% +0.1 0.14 ± 14% perf-profile.self.cycles-pp.ksys_pread64
0.03 ± 70% +0.1 0.09 ± 9% perf-profile.self.cycles-pp.aa_file_perm
0.00 +0.1 0.06 ± 13% perf-profile.self.cycles-pp.generic_file_read_iter
0.00 +0.1 0.06 ± 11% perf-profile.self.cycles-pp.touch_atime
0.00 +0.1 0.06 ± 11% perf-profile.self.cycles-pp.rw_verify_area
0.00 +0.1 0.07 ± 11% perf-profile.self.cycles-pp.__cond_resched
0.01 ±223% +0.1 0.08 ± 16% perf-profile.self.cycles-pp.timestamp_truncate
0.09 ± 7% +0.1 0.17 ± 11% perf-profile.self.cycles-pp.syscall_enter_from_user_mode
0.10 ± 8% +0.1 0.18 ± 7% perf-profile.self.cycles-pp.filemap_get_pages
0.12 ± 11% +0.1 0.21 ± 11% perf-profile.self.cycles-pp.mark_page_accessed
0.11 ± 8% +0.1 0.20 ± 8% perf-profile.self.cycles-pp.map_id_range_down
0.12 ± 16% +0.1 0.22 ± 7% perf-profile.self.cycles-pp.ext4_file_read_iter
0.15 ± 6% +0.1 0.26 ± 7% perf-profile.self.cycles-pp.ksys_readahead
0.12 ± 10% +0.1 0.23 ± 13% perf-profile.self.cycles-pp.new_sync_read
0.12 ± 5% +0.1 0.24 ± 37% perf-profile.self.cycles-pp.do_syscall_64
0.17 ± 9% +0.1 0.30 ± 10% perf-profile.self.cycles-pp.entry_SYSCALL_64_safe_stack
0.18 ± 5% +0.1 0.32 ± 8% perf-profile.self.cycles-pp.exit_to_user_mode_prepare
0.19 ± 9% +0.1 0.32 ± 8% perf-profile.self.cycles-pp.vfs_read
0.19 ± 7% +0.2 0.35 ± 7% perf-profile.self.cycles-pp.copy_page_to_iter
0.20 ± 9% +0.2 0.37 ± 8% perf-profile.self.cycles-pp.force_page_cache_ra
0.20 ± 5% +0.2 0.37 ± 12% perf-profile.self.cycles-pp.__fsnotify_parent
0.20 ± 11% +0.2 0.38 ± 10% perf-profile.self.cycles-pp.generic_fadvise
0.31 ± 3% +0.3 0.58 ± 4% perf-profile.self.cycles-pp.entry_SYSCALL_64_after_hwframe
0.33 ± 6% +0.3 0.62 ± 9% perf-profile.self.cycles-pp.__fget_light
0.53 ± 7% +0.4 0.90 ± 7% perf-profile.self.cycles-pp.__entry_text_start
0.56 ± 2% +0.5 1.03 ± 7% perf-profile.self.cycles-pp.filemap_read
0.81 ± 4% +0.7 1.48 ± 5% perf-profile.self.cycles-pp.syscall_return_via_sysret
0.97 ± 11% +0.9 1.91 ± 13% perf-profile.self.cycles-pp.filemap_get_read_batch
0.09 ± 4% +1.9 2.02 ±223% perf-profile.self.cycles-pp.syscall_exit_to_user_mode
3.24 ± 3% +2.5 5.78 ± 9% perf-profile.self.cycles-pp.copy_user_enhanced_fast_string
Disclaimer:
Results have been estimated based on internal Intel analysis and are provided
for informational purposes only. Any difference in system hardware or software
design or configuration may affect actual performance.
---
0DAY/LKP+ Test Infrastructure Open Source Technology Center
https://lists.01.org/hyperkitty/list/lkp@lists.01.org Intel Corporation
Thanks,
Oliver Sang
View attachment "config-5.13.0-rc3-00037-g2c76f9f255f0" of type "text/plain" (174118 bytes)
View attachment "job-script" of type "text/plain" (8744 bytes)
View attachment "job.yaml" of type "text/plain" (5963 bytes)
View attachment "reproduce" of type "text/plain" (531 bytes)
Powered by blists - more mailing lists