lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [day] [month] [year] [list]
Message-ID: <874mdti4xm.fsf@yhuang-dev.intel.com>
Date:	Mon, 01 Feb 2016 09:40:21 +0800
From:	kernel test robot <ying.huang@...ux.intel.com>
TO:	Jaegeuk Kim <jaegeuk@...nel.org>
CC:	0day robot <fengguang.wu@...el.com>
Subject: [lkp] [f2fs] ea32c36edc: -41.5% fsmark.time.file_system_outputs

FYI, we noticed the below changes on

https://github.com/0day-ci/linux Jaegeuk-Kim/f2fs-avoid-multiple-node-page-writes-due-to-inline_data/20160124-041711
commit ea32c36edcf58f8c27653f3e5bc41f8c0b41e235 ("f2fs: avoid multiple node page writes due to inline_data")


=========================================================================================
compiler/cpufreq_governor/disk/filesize/fs/iterations/kconfig/nr_directories/nr_files_per_directory/nr_threads/rootfs/sync_method/tbox_group/test_size/testcase:
  gcc-4.9/performance/1HDD/9B/f2fs/1x/x86_64-rhel/16d/256fpd/32t/debian-x86_64-2015-02-07.cgz/fsyncBeforeClose/lkp-ne04/400M/fsmark

commit: 
  7fdec82af6a9e190e53d07a1463d2a9ac49a8750
  ea32c36edcf58f8c27653f3e5bc41f8c0b41e235

7fdec82af6a9e190 ea32c36edcf58f8c27653f3e5b 
---------------- -------------------------- 
         %stddev     %change         %stddev
             \          |                \  
      1278 ±  0%      -3.2%       1236 ±  0%  fsmark.files_per_sec
   1874824 ±  0%     -41.5%    1097506 ±  0%  fsmark.time.file_system_outputs
     14748 ±  2%      +9.2%      16099 ±  1%  fsmark.time.involuntary_context_switches
     44.50 ±  4%     +19.1%      53.00 ±  0%  fsmark.time.percent_of_cpu_this_job_got
     36.30 ±  1%     +18.6%      43.03 ±  0%  fsmark.time.system_time
    870163 ±  0%     +19.6%    1041002 ±  0%  fsmark.time.voluntary_context_switches
     22186 ±  1%     +10.4%      24490 ±  1%  vmstat.system.cs
    326553 ±  4%     -10.6%     292026 ±  0%  meminfo.Active
    292915 ±  4%     -11.9%     258035 ±  0%  meminfo.Active(file)
     30038 ± 10%     +17.1%      35170 ± 13%  softirqs.RCU
     31822 ±  9%     +16.4%      37037 ± 13%  softirqs.SCHED
    143471 ±  4%     -15.6%     121104 ±  2%  numa-meminfo.node1.Active
     17081 ± 18%     -21.7%      13370 ± 13%  numa-meminfo.node1.Active(anon)
    126389 ±  2%     -14.8%     107733 ±  2%  numa-meminfo.node1.Active(file)
     16928 ± 18%     -21.9%      13222 ± 13%  numa-meminfo.node1.AnonPages
     75337 ± 20%     -42.1%      43626 ±  1%  numa-vmstat.node0.nr_dirtied
      4266 ± 18%     -21.8%       3338 ± 13%  numa-vmstat.node1.nr_active_anon
     31597 ±  2%     -14.8%      26931 ±  2%  numa-vmstat.node1.nr_active_file
      4226 ± 18%     -22.0%       3297 ± 13%  numa-vmstat.node1.nr_anon_pages
     73228 ±  4%     -11.9%      64502 ±  0%  proc-vmstat.nr_active_file
    233617 ±  0%     -41.5%     136640 ±  0%  proc-vmstat.nr_dirtied
      6178 ±  0%     -11.1%       5489 ±  0%  proc-vmstat.numa_pte_updates
     50891 ±  0%     -39.8%      30627 ±  0%  proc-vmstat.pgactivate
   1874824 ±  0%     -41.5%    1097506 ±  0%  time.file_system_outputs
     44.50 ±  4%     +19.1%      53.00 ±  0%  time.percent_of_cpu_this_job_got
     36.30 ±  1%     +18.6%      43.03 ±  0%  time.system_time
    870163 ±  0%     +19.6%    1041002 ±  0%  time.voluntary_context_switches
  63126565 ±  2%     +16.8%   73717213 ±  1%  cpuidle.C1-NHM.time
  49681942 ±  1%     +34.6%   66867204 ±  3%  cpuidle.C1E-NHM.time
     56832 ±  1%     +20.6%      68527 ±  5%  cpuidle.C1E-NHM.usage
 3.428e+08 ±  1%     +14.9%  3.938e+08 ±  3%  cpuidle.C3-NHM.time
    314858 ±  3%     +32.6%     417584 ± 11%  cpuidle.C3-NHM.usage
     10.96 ±  6%      +6.8%      11.70 ±  1%  turbostat.%Busy
     20.82 ±  5%     +15.7%      24.10 ±  1%  turbostat.CPU%c1
     28.15 ±  4%     +13.0%      31.82 ±  3%  turbostat.CPU%c3
     40.07 ±  7%     -19.2%      32.38 ±  4%  turbostat.CPU%c6
      9.22 ±  6%     -23.4%       7.06 ±  3%  turbostat.Pkg%pc3
     36043 ±  6%    -100.0%       0.00 ± -1%  latency_stats.avg.call_rwsem_down_read_failed.f2fs_write_data_page.[f2fs].__f2fs_writepage.[f2fs].f2fs_write_cache_pages.[f2fs].f2fs_write_data_pages.[f2fs].do_writepages.__filemap_fdatawrite_range.filemap_write_and_wait_range.f2fs_sync_file.[f2fs].vfs_fsync_range.do_fsync.SyS_fsync
      7336 ±  0%    +698.9%      58613 ±  1%  latency_stats.hits.call_rwsem_down_read_failed.get_node_info.[f2fs].f2fs_write_node_page.[f2fs].sync_node_pages.[f2fs].f2fs_sync_file.[f2fs].vfs_fsync_range.do_fsync.SyS_fsync.entry_SYSCALL_64_fastpath
     22886 ±  0%    -100.0%       0.00 ± -1%  latency_stats.hits.call_rwsem_down_read_failed.get_node_info.[f2fs].read_node_page.[f2fs].__get_node_page.[f2fs].get_node_page.[f2fs].get_dnode_of_data.[f2fs].f2fs_write_inline_data.[f2fs].f2fs_write_data_page.[f2fs].__f2fs_writepage.[f2fs].f2fs_write_cache_pages.[f2fs].f2fs_write_data_pages.[f2fs].do_writepages
      0.00 ± -1%      +Inf%      79634 ±  0%  latency_stats.hits.call_rwsem_down_read_failed.get_node_info.[f2fs].read_node_page.[f2fs].__get_node_page.[f2fs].get_node_page.[f2fs].get_dnode_of_data.[f2fs].f2fs_write_inline_data.[f2fs].f2fs_write_end.[f2fs].generic_perform_write.__generic_file_write_iter.generic_file_write_iter.f2fs_file_write_iter.[f2fs]
     75051 ±  0%    -100.0%       0.00 ± -1%  latency_stats.hits.call_rwsem_down_read_failed.get_node_info.[f2fs].read_node_page.[f2fs].__get_node_page.[f2fs].get_node_page.[f2fs].get_node_page.[f2fs].update_inode_page.[f2fs].f2fs_write_end.[f2fs].generic_perform_write.__generic_file_write_iter.generic_file_write_iter.f2fs_file_write_iter.[f2fs]
     17214 ±  0%    +321.2%      72498 ±  0%  latency_stats.hits.call_rwsem_down_read_failed.is_checkpointed_node.[f2fs].f2fs_sync_file.[f2fs].vfs_fsync_range.do_fsync.SyS_fsync.entry_SYSCALL_64_fastpath
     11958 ±  0%    +454.6%      66321 ±  0%  latency_stats.hits.call_rwsem_down_read_failed.need_dentry_mark.[f2fs].sync_node_pages.[f2fs].f2fs_sync_file.[f2fs].vfs_fsync_range.do_fsync.SyS_fsync.entry_SYSCALL_64_fastpath
     74681 ±  0%    -100.0%       8.25 ± 61%  latency_stats.hits.f2fs_write_data_pages.[f2fs].do_writepages.__filemap_fdatawrite_range.filemap_write_and_wait_range.f2fs_sync_file.[f2fs].vfs_fsync_range.do_fsync.SyS_fsync.entry_SYSCALL_64_fastpath
    130560 ± 11%    -100.0%       0.00 ± -1%  latency_stats.max.call_rwsem_down_read_failed.f2fs_write_data_page.[f2fs].__f2fs_writepage.[f2fs].f2fs_write_cache_pages.[f2fs].f2fs_write_data_pages.[f2fs].do_writepages.__filemap_fdatawrite_range.filemap_write_and_wait_range.f2fs_sync_file.[f2fs].vfs_fsync_range.do_fsync.SyS_fsync
    270069 ± 14%    -100.0%     125.25 ± 81%  latency_stats.max.f2fs_write_data_pages.[f2fs].do_writepages.__filemap_fdatawrite_range.filemap_write_and_wait_range.f2fs_sync_file.[f2fs].vfs_fsync_range.do_fsync.SyS_fsync.entry_SYSCALL_64_fastpath
  12155740 ±  1%    -100.0%       0.00 ± -1%  latency_stats.sum.call_rwsem_down_read_failed.f2fs_write_data_page.[f2fs].__f2fs_writepage.[f2fs].f2fs_write_cache_pages.[f2fs].f2fs_write_data_pages.[f2fs].do_writepages.__filemap_fdatawrite_range.filemap_write_and_wait_range.f2fs_sync_file.[f2fs].vfs_fsync_range.do_fsync.SyS_fsync
   1498927 ±  1%    +858.3%   14364772 ±  0%  latency_stats.sum.call_rwsem_down_read_failed.get_node_info.[f2fs].f2fs_write_node_page.[f2fs].sync_node_pages.[f2fs].f2fs_sync_file.[f2fs].vfs_fsync_range.do_fsync.SyS_fsync.entry_SYSCALL_64_fastpath
     22102 ±  9%    +187.5%      63548 ± 19%  latency_stats.sum.call_rwsem_down_read_failed.get_node_info.[f2fs].read_node_page.[f2fs].__get_node_page.[f2fs].get_node_page.[f2fs].get_dnode_of_data.[f2fs].do_write_data_page.[f2fs].f2fs_write_data_page.[f2fs].__f2fs_writepage.[f2fs].f2fs_write_cache_pages.[f2fs].f2fs_write_data_pages.[f2fs].do_writepages
   5128495 ±  1%    -100.0%       0.00 ± -1%  latency_stats.sum.call_rwsem_down_read_failed.get_node_info.[f2fs].read_node_page.[f2fs].__get_node_page.[f2fs].get_node_page.[f2fs].get_dnode_of_data.[f2fs].f2fs_write_inline_data.[f2fs].f2fs_write_data_page.[f2fs].__f2fs_writepage.[f2fs].f2fs_write_cache_pages.[f2fs].f2fs_write_data_pages.[f2fs].do_writepages
      0.00 ± -1%      +Inf%   29676287 ±  0%  latency_stats.sum.call_rwsem_down_read_failed.get_node_info.[f2fs].read_node_page.[f2fs].__get_node_page.[f2fs].get_node_page.[f2fs].get_dnode_of_data.[f2fs].f2fs_write_inline_data.[f2fs].f2fs_write_end.[f2fs].generic_perform_write.__generic_file_write_iter.generic_file_write_iter.f2fs_file_write_iter.[f2fs]
  24113264 ±  1%    -100.0%       0.00 ± -1%  latency_stats.sum.call_rwsem_down_read_failed.get_node_info.[f2fs].read_node_page.[f2fs].__get_node_page.[f2fs].get_node_page.[f2fs].get_node_page.[f2fs].update_inode_page.[f2fs].f2fs_write_end.[f2fs].generic_perform_write.__generic_file_write_iter.generic_file_write_iter.f2fs_file_write_iter.[f2fs]
     11707 ±  6%    +229.8%      38617 ± 24%  latency_stats.sum.call_rwsem_down_read_failed.get_node_info.[f2fs].write_data_page.[f2fs].do_write_data_page.[f2fs].f2fs_write_data_page.[f2fs].__f2fs_writepage.[f2fs].f2fs_write_cache_pages.[f2fs].f2fs_write_data_pages.[f2fs].do_writepages.__filemap_fdatawrite_range.filemap_fdatawrite.sync_dirty_inodes.[f2fs]
   3670765 ±  0%    +549.9%   23857878 ±  0%  latency_stats.sum.call_rwsem_down_read_failed.is_checkpointed_node.[f2fs].f2fs_sync_file.[f2fs].vfs_fsync_range.do_fsync.SyS_fsync.entry_SYSCALL_64_fastpath
   2546436 ±  0%    +680.8%   19883041 ±  0%  latency_stats.sum.call_rwsem_down_read_failed.need_dentry_mark.[f2fs].sync_node_pages.[f2fs].f2fs_sync_file.[f2fs].vfs_fsync_range.do_fsync.SyS_fsync.entry_SYSCALL_64_fastpath
   5243465 ±  2%    +258.7%   18809634 ±  1%  latency_stats.sum.call_rwsem_down_read_failed.need_inode_block_update.[f2fs].f2fs_sync_file.[f2fs].vfs_fsync_range.do_fsync.SyS_fsync.entry_SYSCALL_64_fastpath
     13216 ± 86%     -58.0%       5547 ± 26%  latency_stats.sum.call_rwsem_down_write_failed.f2fs_init_extent_tree.[f2fs].f2fs_new_inode.[f2fs].f2fs_create.[f2fs].vfs_create.path_openat.do_filp_open.do_sys_open.SyS_open.entry_SYSCALL_64_fastpath
   3020759 ±  2%    +333.5%   13096379 ±  0%  latency_stats.sum.call_rwsem_down_write_failed.set_node_addr.[f2fs].f2fs_write_node_page.[f2fs].sync_node_pages.[f2fs].f2fs_sync_file.[f2fs].vfs_fsync_range.do_fsync.SyS_fsync.entry_SYSCALL_64_fastpath
 1.722e+08 ±  2%    -100.0%     630.50 ± 92%  latency_stats.sum.f2fs_write_data_pages.[f2fs].do_writepages.__filemap_fdatawrite_range.filemap_write_and_wait_range.f2fs_sync_file.[f2fs].vfs_fsync_range.do_fsync.SyS_fsync.entry_SYSCALL_64_fastpath
    896.92 ± 10%     +23.0%       1103 ±  6%  sched_debug.cfs_rq:/.exec_clock.12
      1663 ±  9%     +40.9%       2343 ± 24%  sched_debug.cfs_rq:/.exec_clock.2
    767.96 ±  8%     +16.5%     894.94 ±  3%  sched_debug.cfs_rq:/.exec_clock.min
      3.12 ± 17%     -88.0%       0.38 ±110%  sched_debug.cfs_rq:/.load_avg.min
      3666 ± 13%     +35.9%       4980 ±  6%  sched_debug.cfs_rq:/.min_vruntime.12
      4101 ± 10%     +60.7%       6589 ± 26%  sched_debug.cfs_rq:/.min_vruntime.14
      4702 ±  7%     +50.3%       7067 ± 12%  sched_debug.cfs_rq:/.min_vruntime.15
      7043 ± 14%     +25.9%       8870 ± 11%  sched_debug.cfs_rq:/.min_vruntime.2
      4100 ±  4%     +48.3%       6081 ± 15%  sched_debug.cfs_rq:/.min_vruntime.9
      6575 ±  3%     +17.8%       7744 ±  3%  sched_debug.cfs_rq:/.min_vruntime.avg
      3239 ±  3%     +31.0%       4242 ±  9%  sched_debug.cfs_rq:/.min_vruntime.min
      0.33 ±141%   +1175.0%       4.25 ± 69%  sched_debug.cfs_rq:/.nr_spread_over.2
     -4695 ±-17%     -40.2%      -2807 ±-39%  sched_debug.cfs_rq:/.spread0.15
    136.75 ± 26%     -25.8%     101.50 ± 15%  sched_debug.cfs_rq:/.util_avg.3
     48.62 ± 21%     -80.7%       9.38 ± 59%  sched_debug.cfs_rq:/.util_avg.min
     43.46 ± 16%     +60.2%      69.63 ± 21%  sched_debug.cfs_rq:/.util_avg.stddev
      9236 ±  4%      +9.8%      10143 ±  6%  sched_debug.cpu.nr_load_updates.10
      9179 ±  4%     +11.8%      10258 ±  5%  sched_debug.cpu.nr_load_updates.14
      9296 ±  4%      +9.8%      10211 ±  4%  sched_debug.cpu.nr_load_updates.8
     64213 ±  7%     +12.9%      72481 ±  7%  sched_debug.cpu.nr_switches.0
     49843 ±  2%     +19.5%      59574 ±  8%  sched_debug.cpu.nr_switches.1
     35136 ± 10%     +23.1%      43259 ±  5%  sched_debug.cpu.nr_switches.10
     27937 ±  6%     +23.4%      34467 ±  6%  sched_debug.cpu.nr_switches.11
     38343 ± 10%     +21.4%      46565 ± 10%  sched_debug.cpu.nr_switches.12
     35454 ± 10%     +21.8%      43192 ±  3%  sched_debug.cpu.nr_switches.14
     29114 ±  7%     +29.6%      37725 ± 20%  sched_debug.cpu.nr_switches.15
     58979 ±  5%     +17.3%      69170 ±  7%  sched_debug.cpu.nr_switches.2
     56391 ±  5%     +23.7%      69761 ±  8%  sched_debug.cpu.nr_switches.6
     35766 ±  9%     +25.9%      45013 ±  2%  sched_debug.cpu.nr_switches.8
     43509 ±  5%     +16.8%      50818 ±  3%  sched_debug.cpu.nr_switches.avg
     65222 ±  7%     +12.5%      73372 ±  7%  sched_debug.cpu.nr_switches.max
     27549 ±  6%     +18.2%      32567 ±  3%  sched_debug.cpu.nr_switches.min
     -6534 ± -5%     +70.0%     -11111 ± -2%  sched_debug.cpu.nr_uninterruptible.0
   -958.25 ±-13%     +42.1%      -1361 ±-10%  sched_debug.cpu.nr_uninterruptible.1
      1716 ± 20%     +46.1%       2507 ±  6%  sched_debug.cpu.nr_uninterruptible.10
      1373 ±  8%     +56.0%       2143 ± 10%  sched_debug.cpu.nr_uninterruptible.11
      1599 ±  7%     +71.7%       2744 ±  6%  sched_debug.cpu.nr_uninterruptible.12
      1351 ± 20%     +52.4%       2060 ±  3%  sched_debug.cpu.nr_uninterruptible.13
      1824 ±  8%     +40.8%       2568 ±  4%  sched_debug.cpu.nr_uninterruptible.14
      1334 ± 10%     +68.0%       2241 ±  4%  sched_debug.cpu.nr_uninterruptible.15
     -1140 ±-14%     +44.6%      -1649 ± -4%  sched_debug.cpu.nr_uninterruptible.5
     -1139 ±-15%     +41.1%      -1607 ±-17%  sched_debug.cpu.nr_uninterruptible.7
      1167 ± 13%     +70.0%       1983 ±  4%  sched_debug.cpu.nr_uninterruptible.8
      1319 ±  7%     +46.4%       1930 ±  6%  sched_debug.cpu.nr_uninterruptible.9
      1956 ±  8%     +41.9%       2775 ±  5%  sched_debug.cpu.nr_uninterruptible.max
     -6537 ± -5%     +70.0%     -11114 ± -2%  sched_debug.cpu.nr_uninterruptible.min
      2035 ±  1%     +62.9%       3315 ±  1%  sched_debug.cpu.nr_uninterruptible.stddev
     51578 ±  2%     +17.3%      60512 ±  7%  sched_debug.cpu.sched_count.1
     35156 ± 10%     +24.2%      43650 ±  4%  sched_debug.cpu.sched_count.10
     27959 ±  6%     +25.4%      35049 ±  4%  sched_debug.cpu.sched_count.11
     38363 ± 10%     +25.0%      47965 ±  7%  sched_debug.cpu.sched_count.12
     35618 ±  9%     +21.8%      43369 ±  2%  sched_debug.cpu.sched_count.14
     29134 ±  7%     +34.3%      39119 ± 18%  sched_debug.cpu.sched_count.15
     50741 ±  8%     +30.2%      66049 ± 23%  sched_debug.cpu.sched_count.3
     62138 ±  8%     +14.3%      71037 ±  6%  sched_debug.cpu.sched_count.6
     36028 ±  9%     +25.0%      45039 ±  2%  sched_debug.cpu.sched_count.8
     27623 ±  6%     +18.4%      32700 ±  3%  sched_debug.cpu.sched_count.min
     28067 ±  8%     +10.2%      30921 ±  8%  sched_debug.cpu.sched_goidle.0
     21211 ±  3%     +20.1%      25465 ±  8%  sched_debug.cpu.sched_goidle.1
     13529 ± 10%     +20.1%      16248 ±  6%  sched_debug.cpu.sched_goidle.10
     11159 ±  5%     +21.5%      13554 ±  6%  sched_debug.cpu.sched_goidle.11
     13666 ± 11%     +18.9%      16252 ±  3%  sched_debug.cpu.sched_goidle.14
     11607 ±  8%     +29.2%      15001 ± 25%  sched_debug.cpu.sched_goidle.15
     25465 ±  6%     +15.8%      29484 ±  8%  sched_debug.cpu.sched_goidle.2
     24349 ±  5%     +22.5%      29827 ± 10%  sched_debug.cpu.sched_goidle.6
     13756 ±  9%     +23.4%      16972 ±  3%  sched_debug.cpu.sched_goidle.8
     18266 ±  4%     +14.8%      20973 ±  4%  sched_debug.cpu.sched_goidle.avg
     28513 ±  7%     +11.4%      31770 ±  7%  sched_debug.cpu.sched_goidle.max
     11005 ±  5%     +15.1%      12672 ±  3%  sched_debug.cpu.sched_goidle.min
     15905 ±  9%     +22.8%      19533 ±  8%  sched_debug.cpu.ttwu_count.12
     13379 ± 10%     +22.5%      16391 ±  9%  sched_debug.cpu.ttwu_count.13
     15038 ± 11%     +29.0%      19395 ±  8%  sched_debug.cpu.ttwu_count.14
     26020 ±  4%     +22.2%      31789 ±  3%  sched_debug.cpu.ttwu_count.2
     25547 ±  7%     +16.5%      29762 ±  5%  sched_debug.cpu.ttwu_count.3
     26667 ±  5%     +31.7%      35117 ±  9%  sched_debug.cpu.ttwu_count.4
     26022 ±  9%     +15.1%      29953 ±  3%  sched_debug.cpu.ttwu_count.5
     27546 ±  6%     +31.4%      36193 ±  8%  sched_debug.cpu.ttwu_count.6
     25556 ±  5%     +24.9%      31908 ± 13%  sched_debug.cpu.ttwu_count.7
     24624 ±  5%     +18.5%      29191 ±  3%  sched_debug.cpu.ttwu_count.avg
     11891 ±  7%     +23.4%      14671 ±  5%  sched_debug.cpu.ttwu_count.min
     21466 ±  2%     +33.3%      28616 ±  2%  sched_debug.cpu.ttwu_local.0
      6323 ±  9%     +29.1%       8161 ±  3%  sched_debug.cpu.ttwu_local.1
      4485 ±  9%     +31.7%       5907 ±  6%  sched_debug.cpu.ttwu_local.10
      3238 ±  3%     +25.5%       4063 ±  6%  sched_debug.cpu.ttwu_local.11
      5061 ± 10%     +22.7%       6211 ±  9%  sched_debug.cpu.ttwu_local.12
      3264 ± 10%     +28.0%       4178 ±  5%  sched_debug.cpu.ttwu_local.13
      4339 ±  9%     +35.0%       5859 ±  4%  sched_debug.cpu.ttwu_local.14
      3638 ± 10%     +21.4%       4415 ±  5%  sched_debug.cpu.ttwu_local.15
      6344 ±  6%     +37.4%       8716 ±  2%  sched_debug.cpu.ttwu_local.2
      5839 ±  9%     +28.8%       7523 ±  6%  sched_debug.cpu.ttwu_local.3
      6459 ±  6%     +48.7%       9602 ± 13%  sched_debug.cpu.ttwu_local.4
      5643 ±  9%     +29.8%       7322 ±  5%  sched_debug.cpu.ttwu_local.5
      6341 ±  4%     +58.9%      10079 ± 14%  sched_debug.cpu.ttwu_local.6
      5599 ±  4%     +34.9%       7552 ±  5%  sched_debug.cpu.ttwu_local.7
      4901 ±  6%     +28.0%       6275 ±  5%  sched_debug.cpu.ttwu_local.8
      6031 ±  3%     +33.5%       8053 ±  3%  sched_debug.cpu.ttwu_local.avg
     21563 ±  2%     +34.9%      29080 ±  3%  sched_debug.cpu.ttwu_local.max
      3061 ±  6%     +29.6%       3966 ±  4%  sched_debug.cpu.ttwu_local.min
      4253 ±  3%     +36.5%       5807 ±  4%  sched_debug.cpu.ttwu_local.stddev


lkp-ne04: Nehalem-EP
Memory: 12G


                            fsmark.time.file_system_outputs

  1.9e+06 **-**-*-**--***-****-***-****-***-***-****-***-****-***-***-****-**
          |    *    *                                                       |
  1.8e+06 ++                                                                |
  1.7e+06 ++                                                                |
          |                                                                 |
  1.6e+06 ++                                                                |
  1.5e+06 ++                                                                |
          |                                                                 |
  1.4e+06 ++                                                                |
  1.3e+06 ++                                                                |
          |                                                                 |
  1.2e+06 ++                                                                |
  1.1e+06 ++                        O    OO OO                              |
          OO OOOO OOO OOO OOOO OOO O OO O     O                             |
    1e+06 ++----------------------------------------------------------------+


                               proc-vmstat.nr_dirtied

  240000 ++----------------*--*---------------------------------------------+
         **.***.***.****.**  * *.***.***.****.***.***.***.***.****.***.***.**
  220000 ++                                                                 |
         |                                                                  |
         |                                                                  |
  200000 ++                                                                 |
         |                                                                  |
  180000 ++                                                                 |
         |                                                                  |
  160000 ++                                                                 |
         |                                                                  |
         |                                                                  |
  140000 OO OOO OOO OOOO OOO OOO OOO OOO OOOO O                             |
         |                                                                  |
  120000 ++-----------------------------------------------------------------+


                               proc-vmstat.pgactivate

  55000 ++------------------------------------------------------------------+
        |                                                                   |
        **.***.***.* *.*                                              .** .**
  50000 ++          *  :                                             *   *  |
        |               :                                            :      |
        |               **.***.***.***.***.***.***.***.***.* *.***.**       |
  45000 ++                                                  *               |
        |                                                                   |
  40000 ++                                                                  |
        OO OOO OOO OOO OOO OOO                                              |
        |                                                                   |
  35000 ++                                                                  |
        |                                                                   |
        |                                                                   |
  30000 ++---------------------OOO-OOO-OOO-OOO------------------------------+


	[*] bisect-good sample
	[O] bisect-bad  sample

To reproduce:

        git clone git://git.kernel.org/pub/scm/linux/kernel/git/wfg/lkp-tests.git
        cd lkp-tests
        bin/lkp install job.yaml  # job file is attached in this email
        bin/lkp run     job.yaml


Disclaimer:
Results have been estimated based on internal Intel analysis and are provided
for informational purposes only. Any difference in system hardware or software
design or configuration may affect actual performance.


Thanks,
Ying Huang

View attachment "job.yaml" of type "text/plain" (3869 bytes)

Download attachment "reproduce.sh" of type "application/x-sh" (2127 bytes)

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ