[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <20200508085848.GS5770@shao2-debian>
Date: Fri, 8 May 2020 16:58:48 +0800
From: kernel test robot <rong.a.chen@...el.com>
To: Sayali Lokhande <sayalil@...eaurora.org>
Cc: jaegeuk@...nel.org, yuchao0@...wei.com,
linux-f2fs-devel@...ts.sourceforge.net, stummala@...eaurora.org,
linux-kernel@...r.kernel.org,
Sayali Lokhande <sayalil@...eaurora.org>, lkp@...ts.01.org
Subject: [f2fs] da55a3ad36: filebench.sum_bytes_mb/s -44.4% regression
Greeting,
FYI, we noticed a -44.4% regression of filebench.sum_bytes_mb/s due to commit:
commit: da55a3ad3669f17a05e00ec819901293c9cc7bf3 ("[PATCH V4] f2fs: Avoid double lock for cp_rwsem during checkpoint")
url: https://github.com/0day-ci/linux/commits/Sayali-Lokhande/f2fs-Avoid-double-lock-for-cp_rwsem-during-checkpoint/20200501-024326
base: https://git.kernel.org/cgit/linux/kernel/git/jaegeuk/f2fs.git dev-test
in testcase: filebench
on test machine: 8 threads Intel(R) Core(TM) i7-4770 CPU @ 3.40GHz with 8G memory
with following parameters:
disk: 1HDD
fs: f2fs
test: createfiles.f
cpufreq_governor: performance
ucode: 0x27
If you fix the issue, kindly add following tag
Reported-by: kernel test robot <rong.a.chen@...el.com>
Details are as below:
-------------------------------------------------------------------------------------------------->
To reproduce:
git clone https://github.com/intel/lkp-tests.git
cd lkp-tests
bin/lkp install job.yaml # job file is attached in this email
bin/lkp run job.yaml
=========================================================================================
compiler/cpufreq_governor/disk/fs/kconfig/rootfs/tbox_group/test/testcase/ucode:
gcc-7/performance/1HDD/f2fs/x86_64-rhel-7.6/debian-x86_64-20191114.cgz/lkp-hsw-d01/createfiles.f/filebench/0x27
commit:
dcd94277cf ("f2fs: add compressed/gc data read IO stat")
da55a3ad36 ("f2fs: Avoid double lock for cp_rwsem during checkpoint")
dcd94277cf7b4778 da55a3ad3669f17a05e00ec8199
---------------- ---------------------------
%stddev %change %stddev
\ | \
390.50 -44.4% 216.93 ± 28% filebench.sum_bytes_mb/s
74982 -44.5% 41652 ± 28% filebench.sum_operations/s
0.19 ± 3% +55.6% 0.29 ± 29% filebench.sum_time_ms/op
24992 -44.4% 13884 ± 28% filebench.sum_writes/s
11.21 ± 2% +24.5% 13.96 ± 3% filebench.time.elapsed_time
11.21 ± 2% +24.5% 13.96 ± 3% filebench.time.elapsed_time.max
2599052 +8.1% 2808701 filebench.time.file_system_outputs
34.50 -16.9% 28.67 ± 3% filebench.time.percent_of_cpu_this_job_got
19623 ± 3% +14.0% 22364 filebench.time.voluntary_context_switches
12.67 +2.8% 13.02 boot-time.dhcp
8.25 +3.3 11.54 mpstat.cpu.all.iowait%
4.25 ± 4% -0.5 3.73 ± 4% mpstat.cpu.all.sys%
3.84 ± 2% -0.6 3.25 ± 3% mpstat.cpu.all.usr%
10853 ± 3% +21.6% 13198 ± 6% softirqs.CPU0.TIMER
9855 +21.3% 11958 ± 8% softirqs.CPU4.TIMER
9749 ± 3% +36.2% 13275 ± 21% softirqs.CPU6.TIMER
30472 +11.1% 33865 softirqs.SCHED
957129 ± 18% -33.2% 639243 ± 28% cpuidle.C1.time
39341259 ± 22% -66.3% 13274567 ± 93% cpuidle.C6.time
45789 ± 18% -60.4% 18150 ± 87% cpuidle.C6.usage
32265 ± 19% -40.2% 19301 ± 12% cpuidle.POLL.time
8071 ± 12% -15.2% 6845 ± 11% cpuidle.POLL.usage
84.00 -2.4% 82.00 vmstat.cpu.id
7.00 +42.9% 10.00 vmstat.cpu.wa
14528 -15.9% 12222 ± 2% vmstat.io.bi
20968 +73.5% 36378 ± 6% vmstat.io.bo
1387456 +17.2% 1626013 vmstat.memory.cache
4842 +15.5% 5594 ± 2% interrupts.19:IR-IO-APIC.19-fasteoi.ata_piix
166.00 ± 74% +183.1% 470.00 ± 42% interrupts.28:IR-PCI-MSI.409600-edge.eth0
27387 ± 5% +178.5% 76272 ± 17% interrupts.CPU2.LOC:Local_timer_interrupts
30881 ± 20% +63.4% 50460 ± 25% interrupts.CPU3.LOC:Local_timer_interrupts
166.00 ± 74% +183.1% 470.00 ± 42% interrupts.CPU4.28:IR-PCI-MSI.409600-edge.eth0
75.50 ± 17% +32.0% 99.67 ± 3% interrupts.CPU4.CAL:Function_call_interrupts
175.50 -19.5% 141.33 ± 17% interrupts.CPU4.RES:Rescheduling_interrupts
116.00 ± 17% +53.7% 178.33 ± 4% interrupts.CPU6.RES:Rescheduling_interrupts
82.00 ± 6% +33.7% 109.67 ± 4% interrupts.CPU7.RES:Rescheduling_interrupts
373139 ± 2% +20.3% 448922 ± 4% interrupts.LOC:Local_timer_interrupts
84.66 -2.7% 82.38 iostat.cpu.idle
7.19 +43.5% 10.33 ± 3% iostat.cpu.iowait
4.33 -8.0% 3.98 ± 5% iostat.cpu.system
3.81 -13.2% 3.31 ± 3% iostat.cpu.user
1.25 +14.0% 1.42 ± 3% iostat.sda.avgqu-sz
17.43 +1308.1% 245.35 ±131% iostat.sda.await.max
313.22 -14.1% 269.02 ± 4% iostat.sda.r/s
14565 -15.9% 12254 ± 2% iostat.sda.rkB/s
3322 -16.0% 2789 ± 2% iostat.sda.rrqm/s
3.09 ± 3% +467.6% 17.52 ±109% iostat.sda.svctm
8.71 +2685.7% 242.77 ±133% iostat.sda.svctm.max
69.97 ± 2% +7.7% 75.33 ± 3% iostat.sda.util
35.44 ± 2% +94.3% 68.86 ± 7% iostat.sda.w/s
17.43 +1308.1% 245.35 ±131% iostat.sda.w_await.max
22983 +66.0% 38143 ± 6% iostat.sda.wkB/s
319440 +25.3% 400307 ± 2% meminfo.Active
280118 ± 2% +12.6% 315548 meminfo.Active(anon)
39321 ± 4% +115.6% 84758 ± 6% meminfo.Active(file)
66008 +64.7% 108726 ± 3% meminfo.AnonHugePages
279917 ± 2% +12.7% 315339 meminfo.AnonPages
1339189 +16.9% 1565959 meminfo.Cached
161579 ± 2% +98.1% 320046 ± 8% meminfo.Dirty
201499 ± 3% +90.1% 383044 ± 6% meminfo.Inactive
155133 ± 4% +115.6% 334509 ± 6% meminfo.Inactive(file)
63886 +36.5% 87191 ± 2% meminfo.KReclaimable
1819283 +15.7% 2104828 meminfo.Memused
63886 +36.5% 87191 ± 2% meminfo.SReclaimable
98391 +23.8% 121806 ± 2% meminfo.Slab
502.00 ± 13% +84.8% 927.67 ± 10% meminfo.Writeback
219588 -16.8% 182704 ± 3% meminfo.max_used_kB
70801 +11.4% 78906 proc-vmstat.nr_active_anon
9773 ± 3% +116.4% 21146 ± 7% proc-vmstat.nr_active_file
70754 +11.4% 78854 proc-vmstat.nr_anon_pages
324216 ± 3% +14.1% 369827 ± 4% proc-vmstat.nr_dirtied
40081 +99.1% 79787 ± 9% proc-vmstat.nr_dirty
334411 +17.0% 391272 proc-vmstat.nr_file_pages
1558377 -4.5% 1487609 proc-vmstat.nr_free_pages
11461 +5.7% 12118 ± 2% proc-vmstat.nr_inactive_anon
38579 ± 3% +116.3% 83463 ± 7% proc-vmstat.nr_inactive_file
3993 +1.6% 4055 proc-vmstat.nr_kernel_stack
14932 +4.4% 15587 ± 2% proc-vmstat.nr_mapped
1043 +4.0% 1085 proc-vmstat.nr_page_table_pages
11558 +5.8% 12228 ± 2% proc-vmstat.nr_shmem
15979 +36.3% 21783 ± 3% proc-vmstat.nr_slab_reclaimable
134.50 ± 5% +78.7% 240.33 ± 3% proc-vmstat.nr_writeback
71015 ± 2% +106.0% 146295 ± 3% proc-vmstat.nr_written
70801 +11.5% 78912 proc-vmstat.nr_zone_active_anon
9771 ± 3% +116.4% 21145 ± 7% proc-vmstat.nr_zone_active_file
11460 +5.7% 12117 ± 2% proc-vmstat.nr_zone_inactive_anon
38572 ± 3% +116.4% 83466 ± 7% proc-vmstat.nr_zone_inactive_file
40223 +99.0% 80023 ± 9% proc-vmstat.nr_zone_write_pending
285084 ± 2% +105.4% 585492 ± 3% proc-vmstat.pgpgout
6835 ± 2% -11.2% 6072 ± 9% slabinfo.anon_vma_chain.num_objs
51857 +21.8% 63172 ± 2% slabinfo.dentry.active_objs
2509 +20.9% 3035 ± 2% slabinfo.dentry.active_slabs
52713 +20.9% 63747 ± 2% slabinfo.dentry.num_objs
2509 +20.9% 3035 ± 2% slabinfo.dentry.num_slabs
1385 ± 9% +123.0% 3090 ± 12% slabinfo.f2fs_extent_node.active_objs
1562 ± 6% +102.7% 3167 ± 12% slabinfo.f2fs_extent_node.num_objs
11274 ± 8% +91.6% 21599 ± 6% slabinfo.f2fs_extent_tree.active_objs
224.00 ± 8% +90.2% 426.00 ± 6% slabinfo.f2fs_extent_tree.active_slabs
11471 ± 8% +89.6% 21747 ± 6% slabinfo.f2fs_extent_tree.num_objs
224.00 ± 8% +90.2% 426.00 ± 6% slabinfo.f2fs_extent_tree.num_slabs
10359 ± 3% +110.1% 21760 ± 6% slabinfo.f2fs_inode_cache.active_objs
372.50 ± 3% +109.0% 778.67 ± 6% slabinfo.f2fs_inode_cache.active_slabs
10437 ± 3% +109.0% 21817 ± 6% slabinfo.f2fs_inode_cache.num_objs
372.50 ± 3% +109.0% 778.67 ± 6% slabinfo.f2fs_inode_cache.num_slabs
43654 ± 4% +40.1% 61180 slabinfo.f2fs_nat_entry.active_objs
341.50 ± 4% +40.5% 479.67 slabinfo.f2fs_nat_entry.active_slabs
43751 ± 4% +40.5% 61453 slabinfo.f2fs_nat_entry.num_objs
341.50 ± 4% +40.5% 479.67 slabinfo.f2fs_nat_entry.num_slabs
7680 ± 6% -8.9% 6997 ± 6% slabinfo.kmalloc-8.active_objs
7680 ± 6% -8.9% 6997 ± 6% slabinfo.kmalloc-8.num_objs
22811 +54.6% 35261 ± 3% slabinfo.radix_tree_node.active_objs
855.00 +49.5% 1278 ± 3% slabinfo.radix_tree_node.active_slabs
23949 +49.5% 35802 ± 3% slabinfo.radix_tree_node.num_objs
855.00 +49.5% 1278 ± 3% slabinfo.radix_tree_node.num_slabs
2.53 ±100% +137.9% 6.03 ± 11% sched_debug.cfs_rq:/.exec_clock.avg
131763 ± 2% -30.7% 91313 ± 11% sched_debug.cfs_rq:/.load.avg
43314 ± 20% -100.0% 0.00 sched_debug.cfs_rq:/.load.min
58283 ± 18% +45.6% 84867 ± 7% sched_debug.cfs_rq:/.load.stddev
533.31 -20.8% 422.50 ± 7% sched_debug.cfs_rq:/.load_avg.avg
6147 -20.9% 4861 ± 11% sched_debug.cfs_rq:/.min_vruntime.min
1.00 -29.2% 0.71 ± 8% sched_debug.cfs_rq:/.nr_running.avg
1.00 -100.0% 0.00 sched_debug.cfs_rq:/.nr_running.min
128.00 -100.0% 0.00 sched_debug.cfs_rq:/.removed.load_avg.avg
1024 -100.0% 0.00 sched_debug.cfs_rq:/.removed.load_avg.max
338.66 -100.0% 0.00 sched_debug.cfs_rq:/.removed.load_avg.stddev
128.00 -100.0% 0.00 sched_debug.cfs_rq:/.removed.runnable_avg.avg
1024 -100.0% 0.00 sched_debug.cfs_rq:/.removed.runnable_avg.max
338.66 -100.0% 0.00 sched_debug.cfs_rq:/.removed.runnable_avg.stddev
20.81 ± 92% -100.0% 0.00 sched_debug.cfs_rq:/.removed.util_avg.avg
166.50 ± 92% -100.0% 0.00 sched_debug.cfs_rq:/.removed.util_avg.max
55.06 ± 92% -100.0% 0.00 sched_debug.cfs_rq:/.removed.util_avg.stddev
113.25 ± 35% -61.3% 43.88 ± 70% sched_debug.cfs_rq:/.util_est_enqueued.avg
470.50 ± 8% -42.3% 271.67 ± 68% sched_debug.cfs_rq:/.util_est_enqueued.max
1.50 ± 33% -100.0% 0.00 sched_debug.cfs_rq:/.util_est_enqueued.min
170.87 ± 16% -46.2% 91.94 ± 68% sched_debug.cfs_rq:/.util_est_enqueued.stddev
821833 ± 21% -18.7% 668072 ± 11% sched_debug.cpu.avg_idle.max
271414 ± 14% -12.8% 236616 ± 10% sched_debug.cpu.avg_idle.stddev
0.96 ± 42% -62.0% 0.37 ± 55% sched_debug.cpu.clock.stddev
0.96 ± 42% -62.0% 0.37 ± 55% sched_debug.cpu.clock_task.stddev
1106 ± 13% -22.8% 853.96 ± 7% sched_debug.cpu.curr->pid.avg
240.88 ± 56% +127.4% 547.88 ± 5% sched_debug.cpu.curr->pid.stddev
0.00 ± 2% -28.8% 0.00 ± 15% sched_debug.cpu.next_balance.stddev
1.56 ± 28% -54.7% 0.71 ± 8% sched_debug.cpu.nr_running.avg
3.00 ± 33% -66.7% 1.00 sched_debug.cpu.nr_running.max
3552 ± 13% +41.3% 5019 ± 18% sched_debug.cpu.nr_switches.min
filebench.sum_operations_s
80000 +-------------------------------------------------------------------+
|+.++++.+++.+++ +++ ++.++ ++ +++ ++ +.++ ++++.+++.+ + +|
70000 |-+ : : : : : :: : : :: : : : : : :|
60000 |-+ : : : : : :: : : :: : : : : : :|
| : : : : : :: : : :: : : : : : :|
50000 |O+ OO OO O :O: OO: O: :O: : : : : : : : : : : : |
| : : : : : : : : : : : : : : : : : |
40000 |-+ : : : : : : : : : : : : : : : : : |
| :: : : :: : : :: : : :: : :: |
30000 |-+ :: : : :: : : :: : : :: : :: |
20000 |-+ :: : : O :: O: : :: : : :: : :: |
| O : :: : : : : : : : : : |
10000 |-+ : :: : : : : : : : : : |
| : :: : : : : : : : : : |
0 +-------------------------------------------------------------------+
filebench.sum_writes_s
25000 +-------------------------------------------------------------------+
| : : : : : :: : : :: : : : : : :|
| : : : : : :: : : :: : : : : : :|
20000 |-+ : : : : : :: : : :: : : : : : :|
| : : : : : : : : : : : : : : : : : |
|O OO OO O :O: OO: O: :O: : : : : : : : : : : : |
15000 |-+ : : : : : : : : : : : : : : : : : |
| :: : : :: : : : : : : :: :::: |
10000 |-+ :: : : :: : : :: : : :: : :: |
| :: : : O :: O: : :: : : :: : :: |
| :: : : :: : : :: : : :: : :: |
5000 |-+ O : :: : : : : : : : : : |
| : :: : : : : : : : : : |
| : :: : : : : : : : : : |
0 +-------------------------------------------------------------------+
filebench.sum_bytes_mb_s
400 +---------------------------------------------------------------------+
| : : : : : : : : : : : : : : : : :|
350 |-+ : : : : : : : : : : : : : : : : :|
300 |-+ : : : : : : : : : : : : : : : : :|
| : : : : : : : : : : : : : : : : : |
250 |O+ OO O OO :O:O : O:O :O: : : : : : : : : : : : |
| : : : : : : : : : : : : : : : : : |
200 |-+ :: : : :: : : :: : : :: : ::: |
| : : : : : : : : : : :: :: |
150 |-+ : : : O : O: : : : : : :: :: |
100 |-+ : : : : : : : : : : :: :: |
| O : :: : : : : : : : : : |
50 |-+ : :: : : : : : : : : : |
| : :: : : : : : : : : : |
0 +---------------------------------------------------------------------+
[*] bisect-good sample
[O] bisect-bad sample
Disclaimer:
Results have been estimated based on internal Intel analysis and are provided
for informational purposes only. Any difference in system hardware or software
design or configuration may affect actual performance.
Thanks,
Rong Chen
View attachment "config-5.6.0-11881-gda55a3ad3669f1" of type "text/plain" (206513 bytes)
View attachment "job-script" of type "text/plain" (7790 bytes)
View attachment "job.yaml" of type "text/plain" (5411 bytes)
View attachment "reproduce" of type "text/plain" (572 bytes)
Powered by blists - more mailing lists