[<prev] [next>] [thread-next>] [day] [month] [year] [list]
Message-ID: <20160809143359.GA11220@yexl-desktop>
Date: Tue, 9 Aug 2016 22:33:59 +0800
From: kernel test robot <xiaolong.ye@...el.com>
To: Christoph Hellwig <hch@....de>
Cc: Dave Chinner <david@...morbit.com>,
Bob Peterson <rpeterso@...hat.com>,
LKML <linux-kernel@...r.kernel.org>,
Linus Torvalds <torvalds@...ux-foundation.org>, lkp@...org
Subject: [lkp] [xfs] 68a9f5e700: aim7.jobs-per-min -13.6% regression
FYI, we noticed a -13.6% regression of aim7.jobs-per-min due to commit:
commit 68a9f5e7007c1afa2cf6830b690a90d0187c0684 ("xfs: implement iomap based buffered write path")
https://git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux.git master
in testcase: aim7
on test machine: 48 threads Ivytown Ivy Bridge-EP with 64G memory
with following parameters:
disk: 1BRD_48G
fs: xfs
test: disk_wrt
load: 3000
cpufreq_governor: performance
Disclaimer:
Results have been estimated based on internal Intel analysis and are provided
for informational purposes only. Any difference in system hardware or software
design or configuration may affect actual performance.
Details are as below:
-------------------------------------------------------------------------------------------------->
To reproduce:
git clone git://git.kernel.org/pub/scm/linux/kernel/git/wfg/lkp-tests.git
cd lkp-tests
bin/lkp install job.yaml # job file is attached in this email
bin/lkp run job.yaml
=========================================================================================
compiler/cpufreq_governor/disk/fs/kconfig/load/rootfs/tbox_group/test/testcase:
gcc-6/performance/1BRD_48G/xfs/x86_64-rhel/3000/debian-x86_64-2015-02-07.cgz/ivb44/disk_wrt/aim7
commit:
f0c6bcba74 ("xfs: reorder zeroing and flushing sequence in truncate")
68a9f5e700 ("xfs: implement iomap based buffered write path")
f0c6bcba74ac51cb 68a9f5e7007c1afa2cf6830b69
---------------- --------------------------
%stddev %change %stddev
\ | \
486586 ± 0% -13.6% 420342 ± 0% aim7.jobs-per-min
37.23 ± 0% +15.6% 43.04 ± 0% aim7.time.elapsed_time
37.23 ± 0% +15.6% 43.04 ± 0% aim7.time.elapsed_time.max
6424 ± 1% +31.3% 8432 ± 1% aim7.time.involuntary_context_switches
151288 ± 0% +2.8% 155579 ± 0% aim7.time.minor_page_faults
376.31 ± 0% +28.5% 483.48 ± 0% aim7.time.system_time
429058 ± 0% -20.0% 343371 ± 0% aim7.time.voluntary_context_switches
16014 ± 0% +28.8% 20628 ± 1% meminfo.Active(file)
127154 ± 9% -14.4% 108893 ± 11% softirqs.SCHED
14084 ± 18% -33.1% 9421 ± 17% numa-numastat.node1.numa_foreign
15461 ± 17% -31.4% 10598 ± 13% numa-numastat.node1.numa_miss
24561 ± 0% -27.2% 17873 ± 1% vmstat.system.cs
47289 ± 0% +1.2% 47866 ± 0% vmstat.system.in
7868 ± 1% +27.3% 10013 ± 6% numa-meminfo.node0.Active(file)
8148 ± 1% +29.5% 10554 ± 7% numa-meminfo.node1.Active(file)
81041 ± 3% +30.0% 105374 ± 24% numa-meminfo.node1.Slab
1966 ± 1% +30.1% 2558 ± 4% numa-vmstat.node0.nr_active_file
4204 ± 3% +17.1% 4921 ± 8% numa-vmstat.node0.nr_alloc_batch
2037 ± 1% +26.6% 2579 ± 5% numa-vmstat.node1.nr_active_file
4003 ± 0% +28.1% 5129 ± 1% proc-vmstat.nr_active_file
979.25 ± 0% +63.7% 1602 ± 1% proc-vmstat.pgactivate
4699 ± 3% +162.6% 12340 ± 73% proc-vmstat.pgpgout
50.23 ± 19% -27.3% 36.50 ± 17% sched_debug.cpu.cpu_load[1].avg
466.50 ± 29% -51.8% 225.00 ± 73% sched_debug.cpu.cpu_load[1].max
77.78 ± 33% -50.6% 38.40 ± 57% sched_debug.cpu.cpu_load[1].stddev
300.50 ± 33% -52.9% 141.50 ± 48% sched_debug.cpu.cpu_load[2].max
1836 ± 10% +65.5% 3039 ± 8% slabinfo.scsi_data_buffer.active_objs
1836 ± 10% +65.5% 3039 ± 8% slabinfo.scsi_data_buffer.num_objs
431.75 ± 10% +65.6% 715.00 ± 8% slabinfo.xfs_efd_item.active_objs
431.75 ± 10% +65.6% 715.00 ± 8% slabinfo.xfs_efd_item.num_objs
24.26 ± 0% +8.7% 26.36 ± 0% turbostat.%Busy
686.75 ± 0% +9.1% 749.25 ± 0% turbostat.Avg_MHz
0.29 ± 1% -24.3% 0.22 ± 1% turbostat.CPU%c3
91.39 ± 2% +3.6% 94.71 ± 0% turbostat.CorWatt
121.88 ± 1% +2.8% 125.23 ± 0% turbostat.PkgWatt
53643508 ± 0% -19.6% 43119128 ± 2% cpuidle.C1-IVT.time
318952 ± 0% -25.7% 237018 ± 0% cpuidle.C1-IVT.usage
3471235 ± 2% -16.9% 2886121 ± 2% cpuidle.C1E-IVT.time
46642 ± 1% -22.4% 36214 ± 0% cpuidle.C1E-IVT.usage
12601665 ± 1% -21.8% 9854467 ± 1% cpuidle.C3-IVT.time
79872 ± 1% -19.6% 64244 ± 1% cpuidle.C3-IVT.usage
1.292e+09 ± 0% +13.7% 1.47e+09 ± 0% cpuidle.C6-IVT.time
5131 ±121% -100.0% 0.00 ± -1% latency_stats.avg.nfs_wait_on_request.nfs_updatepage.nfs_write_end.generic_perform_write.__generic_file_write_iter.generic_file_write_iter.nfs_file_write.__vfs_write.vfs_write.SyS_write.entry_SYSCALL_64_fastpath
5131 ±121% -100.0% 0.00 ± -1% latency_stats.max.nfs_wait_on_request.nfs_updatepage.nfs_write_end.generic_perform_write.__generic_file_write_iter.generic_file_write_iter.nfs_file_write.__vfs_write.vfs_write.SyS_write.entry_SYSCALL_64_fastpath
9739 ± 99% -99.0% 95.50 ± 10% latency_stats.max.submit_bio_wait.blkdev_issue_flush.ext4_sync_fs.sync_fs_one_sb.iterate_supers.sys_sync.entry_SYSCALL_64_fastpath
7739 ± 81% -72.1% 2162 ± 52% latency_stats.max.wait_on_page_bit.__filemap_fdatawait_range.filemap_fdatawait_keep_errors.sync_inodes_sb.sync_inodes_one_sb.iterate_supers.sys_sync.entry_SYSCALL_64_fastpath
5131 ±121% -100.0% 0.00 ± -1% latency_stats.sum.nfs_wait_on_request.nfs_updatepage.nfs_write_end.generic_perform_write.__generic_file_write_iter.generic_file_write_iter.nfs_file_write.__vfs_write.vfs_write.SyS_write.entry_SYSCALL_64_fastpath
10459 ± 97% -97.5% 262.75 ± 5% latency_stats.sum.submit_bio_wait.blkdev_issue_flush.ext4_sync_fs.sync_fs_one_sb.iterate_supers.sys_sync.entry_SYSCALL_64_fastpath
9097 ± 81% -72.5% 2505 ± 45% latency_stats.sum.wait_on_page_bit.__filemap_fdatawait_range.filemap_fdatawait_keep_errors.sync_inodes_sb.sync_inodes_one_sb.iterate_supers.sys_sync.entry_SYSCALL_64_fastpath
2.59e+11 ± 6% +24.1% 3.213e+11 ± 4% perf-stat.branch-instructions
0.41 ± 2% -9.5% 0.38 ± 1% perf-stat.branch-miss-rate
1.072e+09 ± 4% +12.5% 1.206e+09 ± 3% perf-stat.branch-misses
972882 ± 0% -17.4% 803990 ± 0% perf-stat.context-switches
1.472e+12 ± 6% +22.4% 1.801e+12 ± 5% perf-stat.cpu-cycles
100350 ± 1% -5.1% 95219 ± 1% perf-stat.cpu-migrations
7.315e+08 ± 24% +60.4% 1.174e+09 ± 37% perf-stat.dTLB-load-misses
3.225e+11 ± 5% +36.4% 4.398e+11 ± 2% perf-stat.dTLB-loads
2.176e+11 ± 9% +44.6% 3.147e+11 ± 6% perf-stat.dTLB-stores
1.452e+12 ± 6% +29.5% 1.879e+12 ± 4% perf-stat.instructions
42168 ± 16% +27.5% 53751 ± 6% perf-stat.instructions-per-iTLB-miss
0.99 ± 0% +5.7% 1.04 ± 0% perf-stat.ipc
252401 ± 0% +6.6% 269148 ± 0% perf-stat.minor-faults
10.16 ± 3% +13.0% 11.48 ± 3% perf-stat.node-store-miss-rate
24842185 ± 2% +11.9% 27804764 ± 1% perf-stat.node-store-misses
252321 ± 0% +6.6% 268999 ± 0% perf-stat.page-faults
aim7.jobs-per-min
540000 ++-----------------------------------------------------------------+
520000 **.* *.**. .**.* |
| *.**.**.* ** *.**.**.**.**.* |
500000 ++ : |
480000 ++ *.**.**.**.**.**.**.**.**.*|
| |
460000 ++ |
440000 ++ |
420000 ++ O OO OO OO OO OO OO
|O O O OO O O O O O |
400000 O+ OO O OO O O O OO OO OO O O OO |
380000 ++ |
| |
360000 ++ O OO O |
340000 ++-----------------------------------------------------------------+
[*] bisect-good sample
[O] bisect-bad sample
Thanks,
Xiaolong
View attachment "config-4.7.0-rc1-00007-g68a9f5e" of type "text/plain" (151225 bytes)
View attachment "job.yaml" of type "text/plain" (3944 bytes)
View attachment "reproduce" of type "text/plain" (305 bytes)
Powered by blists - more mailing lists