lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Date:   Fri, 8 May 2020 16:58:48 +0800
From:   kernel test robot <rong.a.chen@...el.com>
To:     Sayali Lokhande <sayalil@...eaurora.org>
Cc:     jaegeuk@...nel.org, yuchao0@...wei.com,
        linux-f2fs-devel@...ts.sourceforge.net, stummala@...eaurora.org,
        linux-kernel@...r.kernel.org,
        Sayali Lokhande <sayalil@...eaurora.org>, lkp@...ts.01.org
Subject: [f2fs] da55a3ad36: filebench.sum_bytes_mb/s -44.4% regression

Greeting,

FYI, we noticed a -44.4% regression of filebench.sum_bytes_mb/s due to commit:


commit: da55a3ad3669f17a05e00ec819901293c9cc7bf3 ("[PATCH V4] f2fs: Avoid double lock for cp_rwsem during checkpoint")
url: https://github.com/0day-ci/linux/commits/Sayali-Lokhande/f2fs-Avoid-double-lock-for-cp_rwsem-during-checkpoint/20200501-024326
base: https://git.kernel.org/cgit/linux/kernel/git/jaegeuk/f2fs.git dev-test

in testcase: filebench
on test machine: 8 threads Intel(R) Core(TM) i7-4770 CPU @ 3.40GHz with 8G memory
with following parameters:

	disk: 1HDD
	fs: f2fs
	test: createfiles.f
	cpufreq_governor: performance
	ucode: 0x27




If you fix the issue, kindly add following tag
Reported-by: kernel test robot <rong.a.chen@...el.com>


Details are as below:
-------------------------------------------------------------------------------------------------->


To reproduce:

        git clone https://github.com/intel/lkp-tests.git
        cd lkp-tests
        bin/lkp install job.yaml  # job file is attached in this email
        bin/lkp run     job.yaml

=========================================================================================
compiler/cpufreq_governor/disk/fs/kconfig/rootfs/tbox_group/test/testcase/ucode:
  gcc-7/performance/1HDD/f2fs/x86_64-rhel-7.6/debian-x86_64-20191114.cgz/lkp-hsw-d01/createfiles.f/filebench/0x27

commit: 
  dcd94277cf ("f2fs: add compressed/gc data read IO stat")
  da55a3ad36 ("f2fs: Avoid double lock for cp_rwsem during checkpoint")

dcd94277cf7b4778 da55a3ad3669f17a05e00ec8199 
---------------- --------------------------- 
         %stddev     %change         %stddev
             \          |                \  
    390.50           -44.4%     216.93 ± 28%  filebench.sum_bytes_mb/s
     74982           -44.5%      41652 ± 28%  filebench.sum_operations/s
      0.19 ±  3%     +55.6%       0.29 ± 29%  filebench.sum_time_ms/op
     24992           -44.4%      13884 ± 28%  filebench.sum_writes/s
     11.21 ±  2%     +24.5%      13.96 ±  3%  filebench.time.elapsed_time
     11.21 ±  2%     +24.5%      13.96 ±  3%  filebench.time.elapsed_time.max
   2599052            +8.1%    2808701        filebench.time.file_system_outputs
     34.50           -16.9%      28.67 ±  3%  filebench.time.percent_of_cpu_this_job_got
     19623 ±  3%     +14.0%      22364        filebench.time.voluntary_context_switches
     12.67            +2.8%      13.02        boot-time.dhcp
      8.25            +3.3       11.54        mpstat.cpu.all.iowait%
      4.25 ±  4%      -0.5        3.73 ±  4%  mpstat.cpu.all.sys%
      3.84 ±  2%      -0.6        3.25 ±  3%  mpstat.cpu.all.usr%
     10853 ±  3%     +21.6%      13198 ±  6%  softirqs.CPU0.TIMER
      9855           +21.3%      11958 ±  8%  softirqs.CPU4.TIMER
      9749 ±  3%     +36.2%      13275 ± 21%  softirqs.CPU6.TIMER
     30472           +11.1%      33865        softirqs.SCHED
    957129 ± 18%     -33.2%     639243 ± 28%  cpuidle.C1.time
  39341259 ± 22%     -66.3%   13274567 ± 93%  cpuidle.C6.time
     45789 ± 18%     -60.4%      18150 ± 87%  cpuidle.C6.usage
     32265 ± 19%     -40.2%      19301 ± 12%  cpuidle.POLL.time
      8071 ± 12%     -15.2%       6845 ± 11%  cpuidle.POLL.usage
     84.00            -2.4%      82.00        vmstat.cpu.id
      7.00           +42.9%      10.00        vmstat.cpu.wa
     14528           -15.9%      12222 ±  2%  vmstat.io.bi
     20968           +73.5%      36378 ±  6%  vmstat.io.bo
   1387456           +17.2%    1626013        vmstat.memory.cache
      4842           +15.5%       5594 ±  2%  interrupts.19:IR-IO-APIC.19-fasteoi.ata_piix
    166.00 ± 74%    +183.1%     470.00 ± 42%  interrupts.28:IR-PCI-MSI.409600-edge.eth0
     27387 ±  5%    +178.5%      76272 ± 17%  interrupts.CPU2.LOC:Local_timer_interrupts
     30881 ± 20%     +63.4%      50460 ± 25%  interrupts.CPU3.LOC:Local_timer_interrupts
    166.00 ± 74%    +183.1%     470.00 ± 42%  interrupts.CPU4.28:IR-PCI-MSI.409600-edge.eth0
     75.50 ± 17%     +32.0%      99.67 ±  3%  interrupts.CPU4.CAL:Function_call_interrupts
    175.50           -19.5%     141.33 ± 17%  interrupts.CPU4.RES:Rescheduling_interrupts
    116.00 ± 17%     +53.7%     178.33 ±  4%  interrupts.CPU6.RES:Rescheduling_interrupts
     82.00 ±  6%     +33.7%     109.67 ±  4%  interrupts.CPU7.RES:Rescheduling_interrupts
    373139 ±  2%     +20.3%     448922 ±  4%  interrupts.LOC:Local_timer_interrupts
     84.66            -2.7%      82.38        iostat.cpu.idle
      7.19           +43.5%      10.33 ±  3%  iostat.cpu.iowait
      4.33            -8.0%       3.98 ±  5%  iostat.cpu.system
      3.81           -13.2%       3.31 ±  3%  iostat.cpu.user
      1.25           +14.0%       1.42 ±  3%  iostat.sda.avgqu-sz
     17.43         +1308.1%     245.35 ±131%  iostat.sda.await.max
    313.22           -14.1%     269.02 ±  4%  iostat.sda.r/s
     14565           -15.9%      12254 ±  2%  iostat.sda.rkB/s
      3322           -16.0%       2789 ±  2%  iostat.sda.rrqm/s
      3.09 ±  3%    +467.6%      17.52 ±109%  iostat.sda.svctm
      8.71         +2685.7%     242.77 ±133%  iostat.sda.svctm.max
     69.97 ±  2%      +7.7%      75.33 ±  3%  iostat.sda.util
     35.44 ±  2%     +94.3%      68.86 ±  7%  iostat.sda.w/s
     17.43         +1308.1%     245.35 ±131%  iostat.sda.w_await.max
     22983           +66.0%      38143 ±  6%  iostat.sda.wkB/s
    319440           +25.3%     400307 ±  2%  meminfo.Active
    280118 ±  2%     +12.6%     315548        meminfo.Active(anon)
     39321 ±  4%    +115.6%      84758 ±  6%  meminfo.Active(file)
     66008           +64.7%     108726 ±  3%  meminfo.AnonHugePages
    279917 ±  2%     +12.7%     315339        meminfo.AnonPages
   1339189           +16.9%    1565959        meminfo.Cached
    161579 ±  2%     +98.1%     320046 ±  8%  meminfo.Dirty
    201499 ±  3%     +90.1%     383044 ±  6%  meminfo.Inactive
    155133 ±  4%    +115.6%     334509 ±  6%  meminfo.Inactive(file)
     63886           +36.5%      87191 ±  2%  meminfo.KReclaimable
   1819283           +15.7%    2104828        meminfo.Memused
     63886           +36.5%      87191 ±  2%  meminfo.SReclaimable
     98391           +23.8%     121806 ±  2%  meminfo.Slab
    502.00 ± 13%     +84.8%     927.67 ± 10%  meminfo.Writeback
    219588           -16.8%     182704 ±  3%  meminfo.max_used_kB
     70801           +11.4%      78906        proc-vmstat.nr_active_anon
      9773 ±  3%    +116.4%      21146 ±  7%  proc-vmstat.nr_active_file
     70754           +11.4%      78854        proc-vmstat.nr_anon_pages
    324216 ±  3%     +14.1%     369827 ±  4%  proc-vmstat.nr_dirtied
     40081           +99.1%      79787 ±  9%  proc-vmstat.nr_dirty
    334411           +17.0%     391272        proc-vmstat.nr_file_pages
   1558377            -4.5%    1487609        proc-vmstat.nr_free_pages
     11461            +5.7%      12118 ±  2%  proc-vmstat.nr_inactive_anon
     38579 ±  3%    +116.3%      83463 ±  7%  proc-vmstat.nr_inactive_file
      3993            +1.6%       4055        proc-vmstat.nr_kernel_stack
     14932            +4.4%      15587 ±  2%  proc-vmstat.nr_mapped
      1043            +4.0%       1085        proc-vmstat.nr_page_table_pages
     11558            +5.8%      12228 ±  2%  proc-vmstat.nr_shmem
     15979           +36.3%      21783 ±  3%  proc-vmstat.nr_slab_reclaimable
    134.50 ±  5%     +78.7%     240.33 ±  3%  proc-vmstat.nr_writeback
     71015 ±  2%    +106.0%     146295 ±  3%  proc-vmstat.nr_written
     70801           +11.5%      78912        proc-vmstat.nr_zone_active_anon
      9771 ±  3%    +116.4%      21145 ±  7%  proc-vmstat.nr_zone_active_file
     11460            +5.7%      12117 ±  2%  proc-vmstat.nr_zone_inactive_anon
     38572 ±  3%    +116.4%      83466 ±  7%  proc-vmstat.nr_zone_inactive_file
     40223           +99.0%      80023 ±  9%  proc-vmstat.nr_zone_write_pending
    285084 ±  2%    +105.4%     585492 ±  3%  proc-vmstat.pgpgout
      6835 ±  2%     -11.2%       6072 ±  9%  slabinfo.anon_vma_chain.num_objs
     51857           +21.8%      63172 ±  2%  slabinfo.dentry.active_objs
      2509           +20.9%       3035 ±  2%  slabinfo.dentry.active_slabs
     52713           +20.9%      63747 ±  2%  slabinfo.dentry.num_objs
      2509           +20.9%       3035 ±  2%  slabinfo.dentry.num_slabs
      1385 ±  9%    +123.0%       3090 ± 12%  slabinfo.f2fs_extent_node.active_objs
      1562 ±  6%    +102.7%       3167 ± 12%  slabinfo.f2fs_extent_node.num_objs
     11274 ±  8%     +91.6%      21599 ±  6%  slabinfo.f2fs_extent_tree.active_objs
    224.00 ±  8%     +90.2%     426.00 ±  6%  slabinfo.f2fs_extent_tree.active_slabs
     11471 ±  8%     +89.6%      21747 ±  6%  slabinfo.f2fs_extent_tree.num_objs
    224.00 ±  8%     +90.2%     426.00 ±  6%  slabinfo.f2fs_extent_tree.num_slabs
     10359 ±  3%    +110.1%      21760 ±  6%  slabinfo.f2fs_inode_cache.active_objs
    372.50 ±  3%    +109.0%     778.67 ±  6%  slabinfo.f2fs_inode_cache.active_slabs
     10437 ±  3%    +109.0%      21817 ±  6%  slabinfo.f2fs_inode_cache.num_objs
    372.50 ±  3%    +109.0%     778.67 ±  6%  slabinfo.f2fs_inode_cache.num_slabs
     43654 ±  4%     +40.1%      61180        slabinfo.f2fs_nat_entry.active_objs
    341.50 ±  4%     +40.5%     479.67        slabinfo.f2fs_nat_entry.active_slabs
     43751 ±  4%     +40.5%      61453        slabinfo.f2fs_nat_entry.num_objs
    341.50 ±  4%     +40.5%     479.67        slabinfo.f2fs_nat_entry.num_slabs
      7680 ±  6%      -8.9%       6997 ±  6%  slabinfo.kmalloc-8.active_objs
      7680 ±  6%      -8.9%       6997 ±  6%  slabinfo.kmalloc-8.num_objs
     22811           +54.6%      35261 ±  3%  slabinfo.radix_tree_node.active_objs
    855.00           +49.5%       1278 ±  3%  slabinfo.radix_tree_node.active_slabs
     23949           +49.5%      35802 ±  3%  slabinfo.radix_tree_node.num_objs
    855.00           +49.5%       1278 ±  3%  slabinfo.radix_tree_node.num_slabs
      2.53 ±100%    +137.9%       6.03 ± 11%  sched_debug.cfs_rq:/.exec_clock.avg
    131763 ±  2%     -30.7%      91313 ± 11%  sched_debug.cfs_rq:/.load.avg
     43314 ± 20%    -100.0%       0.00        sched_debug.cfs_rq:/.load.min
     58283 ± 18%     +45.6%      84867 ±  7%  sched_debug.cfs_rq:/.load.stddev
    533.31           -20.8%     422.50 ±  7%  sched_debug.cfs_rq:/.load_avg.avg
      6147           -20.9%       4861 ± 11%  sched_debug.cfs_rq:/.min_vruntime.min
      1.00           -29.2%       0.71 ±  8%  sched_debug.cfs_rq:/.nr_running.avg
      1.00          -100.0%       0.00        sched_debug.cfs_rq:/.nr_running.min
    128.00          -100.0%       0.00        sched_debug.cfs_rq:/.removed.load_avg.avg
      1024          -100.0%       0.00        sched_debug.cfs_rq:/.removed.load_avg.max
    338.66          -100.0%       0.00        sched_debug.cfs_rq:/.removed.load_avg.stddev
    128.00          -100.0%       0.00        sched_debug.cfs_rq:/.removed.runnable_avg.avg
      1024          -100.0%       0.00        sched_debug.cfs_rq:/.removed.runnable_avg.max
    338.66          -100.0%       0.00        sched_debug.cfs_rq:/.removed.runnable_avg.stddev
     20.81 ± 92%    -100.0%       0.00        sched_debug.cfs_rq:/.removed.util_avg.avg
    166.50 ± 92%    -100.0%       0.00        sched_debug.cfs_rq:/.removed.util_avg.max
     55.06 ± 92%    -100.0%       0.00        sched_debug.cfs_rq:/.removed.util_avg.stddev
    113.25 ± 35%     -61.3%      43.88 ± 70%  sched_debug.cfs_rq:/.util_est_enqueued.avg
    470.50 ±  8%     -42.3%     271.67 ± 68%  sched_debug.cfs_rq:/.util_est_enqueued.max
      1.50 ± 33%    -100.0%       0.00        sched_debug.cfs_rq:/.util_est_enqueued.min
    170.87 ± 16%     -46.2%      91.94 ± 68%  sched_debug.cfs_rq:/.util_est_enqueued.stddev
    821833 ± 21%     -18.7%     668072 ± 11%  sched_debug.cpu.avg_idle.max
    271414 ± 14%     -12.8%     236616 ± 10%  sched_debug.cpu.avg_idle.stddev
      0.96 ± 42%     -62.0%       0.37 ± 55%  sched_debug.cpu.clock.stddev
      0.96 ± 42%     -62.0%       0.37 ± 55%  sched_debug.cpu.clock_task.stddev
      1106 ± 13%     -22.8%     853.96 ±  7%  sched_debug.cpu.curr->pid.avg
    240.88 ± 56%    +127.4%     547.88 ±  5%  sched_debug.cpu.curr->pid.stddev
      0.00 ±  2%     -28.8%       0.00 ± 15%  sched_debug.cpu.next_balance.stddev
      1.56 ± 28%     -54.7%       0.71 ±  8%  sched_debug.cpu.nr_running.avg
      3.00 ± 33%     -66.7%       1.00        sched_debug.cpu.nr_running.max
      3552 ± 13%     +41.3%       5019 ± 18%  sched_debug.cpu.nr_switches.min


                                                                                
                             filebench.sum_operations_s                         
                                                                                
  80000 +-------------------------------------------------------------------+   
        |+.++++.+++.+++  +++   ++.++  ++    +++  ++    +.++  ++++.+++.+ +  +|   
  70000 |-+           :  : :   :   :  ::    : :  ::    :  :  :        : :  :|   
  60000 |-+           :  : :   :   :  ::    : :  ::    :  :  :        : :  :|   
        |             :  : :   :   :  ::    : :  ::    :  :  :        : :  :|   
  50000 |O+ OO   OO O :O: OO: O:   :O: :    :  : : :   :  : :         : : : |   
        |             : :   :  :   : : :    :  : : :   :  : :         : : : |   
  40000 |-+           : :   :  :   : : :    :  : : :   :  : :         : : : |   
        |              ::   : :     ::  :  :   ::  :  :    ::          : :: |   
  30000 |-+            ::   : :     ::  :  :   ::  :  :    ::          : :: |   
  20000 |-+            ::   : :   O :: O:  :   ::  :  :    ::          : :: |   
        |            O :     ::     :   :  :    :   : :    :           : :  |   
  10000 |-+            :     ::     :   :  :    :   : :    :           : :  |   
        |              :     ::     :   :  :    :   : :    :           : :  |   
      0 +-------------------------------------------------------------------+   
                                                                                
                                                                                                                                                                
                               filebench.sum_writes_s                           
                                                                                
  25000 +-------------------------------------------------------------------+   
        |             :  : :   :   :  ::    : :  ::    :  :  :        : :  :|   
        |             :  : :   :   :  ::    : :  ::    :  :  :        : :  :|   
  20000 |-+           :  : :   :   :  ::    : :  ::    :  :  :        : :  :|   
        |             : :   :  :   : : :    :  : : :   :  : :         : : : |   
        |O  OO   OO O :O: OO: O:   :O: :    :  : : :   :  : :         : : : |   
  15000 |-+           : :   :  :   : : :    :  : : :   :  : :         : : : |   
        |              ::   :  :    ::  :   :  : : :   :   ::          :::: |   
  10000 |-+            ::   : :     ::  :  :   ::  :  :    ::          : :: |   
        |              ::   : :   O :: O:  :   ::  :  :    ::          : :: |   
        |              ::   : :     ::  :  :   ::  :  :    ::          : :: |   
   5000 |-+          O :     ::     :   :  :    :   : :    :           : :  |   
        |              :     ::     :   :  :    :   : :    :           : :  |   
        |              :     ::     :   :  :    :   : :    :           : :  |   
      0 +-------------------------------------------------------------------+   
                                                                                
                                                                                                                                                                
                             filebench.sum_bytes_mb_s                           
                                                                                
  400 +---------------------------------------------------------------------+   
      |              : :  :   :   : : :    :  : : :    :  : :        :  :  :|   
  350 |-+            : :  :   :   : : :    :  : : :    :  : :        :  :  :|   
  300 |-+            : :  :   :   : : :    :  : : :    :  : :        :  :  :|   
      |              : :  :  :    : : :    :  : : :    :  : :         : : : |   
  250 |O+ OO   O OO  :O:O : O:O   :O: :    :  : : :    :  : :         : : : |   
      |              : :  :  :    : : :    :  : : :    :  : :         : : : |   
  200 |-+             ::   : :     ::  :   :   ::  :   :   ::         : ::: |   
      |               :    : :     :   :  :    :   :  :    :          :: :: |   
  150 |-+             :    : :  O  :  O:  :    :   :  :    :          :: :: |   
  100 |-+             :    : :     :   :  :    :   :  :    :          :: :: |   
      |            O  :    ::      :   :  :    :   :  :    :           : :  |   
   50 |-+             :    ::      :   :  :    :   :  :    :           : :  |   
      |               :    ::      :   :  :    :   :  :    :           : :  |   
    0 +---------------------------------------------------------------------+   
                                                                                
                                                                                
[*] bisect-good sample
[O] bisect-bad  sample



Disclaimer:
Results have been estimated based on internal Intel analysis and are provided
for informational purposes only. Any difference in system hardware or software
design or configuration may affect actual performance.


Thanks,
Rong Chen


View attachment "config-5.6.0-11881-gda55a3ad3669f1" of type "text/plain" (206513 bytes)

View attachment "job-script" of type "text/plain" (7790 bytes)

View attachment "job.yaml" of type "text/plain" (5411 bytes)

View attachment "reproduce" of type "text/plain" (572 bytes)

Powered by blists - more mailing lists