lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [thread-next>] [day] [month] [year] [list]
Message-ID: <20160517045805.GA16472@yexl-desktop>
Date:	Tue, 17 May 2016 12:58:05 +0800
From:	kernel test robot <xiaolong.ye@...el.com>
To:	Johannes Weiner <hannes@...xchg.org>
Cc:	Sergey Senozhatsky <sergey.senozhatsky@...il.com>,
	Vladimir Davydov <vdavydov@...tuozzo.com>,
	Michal Hocko <mhocko@...e.cz>,
	David Rientjes <rientjes@...gle.com>,
	LKML <linux-kernel@...r.kernel.org>, lkp@...org
Subject: [mm] 23047a96d7: vm-scalability.throughput -23.8% regression

FYI, we noticed vm-scalability.throughput -23.8% regression due to commit:

commit 23047a96d7cfcfca1a6d026ecaec526ea4803e9e ("mm: workingset: per-cgroup cache thrash detection")
https://git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux.git master

in testcase: vm-scalability
on test machine: lkp-hsw01: 56 threads Grantley Haswell-EP with 64G memory
with following conditions: cpufreq_governor=performance/runtime=300s/test=lru-file-readtwice


Details are as below:
-------------------------------------------------------------------------------------------------->


=========================================================================================
compiler/cpufreq_governor/kconfig/rootfs/runtime/tbox_group/test/testcase:
  gcc-4.9/performance/x86_64-rhel/debian-x86_64-2015-02-07.cgz/300s/lkp-hsw01/lru-file-readtwice/vm-scalability

commit: 
  612e44939c3c77245ac80843c0c7876c8cf97282
  23047a96d7cfcfca1a6d026ecaec526ea4803e9e

612e44939c3c7724 23047a96d7cfcfca1a6d026eca 
---------------- -------------------------- 
         %stddev     %change         %stddev
             \          |                \  
  28384711 ±  0%     -23.8%   21621405 ±  0%  vm-scalability.throughput
   1854112 ±  0%      -7.7%    1711141 ±  0%  vm-scalability.time.involuntary_context_switches
    176.03 ±  0%     -22.2%     136.95 ±  1%  vm-scalability.time.user_time
    302905 ±  2%     -31.2%     208386 ±  0%  vm-scalability.time.voluntary_context_switches
      0.92 ±  2%     +51.0%       1.38 ±  2%  perf-profile.cycles-pp.kswapd
    754212 ±  1%     -29.2%     533832 ±  2%  softirqs.RCU
     20518 ±  2%      -8.1%      18866 ±  2%  vmstat.system.cs
     10574 ± 19%     +29.9%      13737 ±  8%  numa-meminfo.node0.Mapped
     13490 ± 13%     -36.6%       8549 ± 17%  numa-meminfo.node1.Mapped
    583.00 ±  8%     +18.8%     692.50 ±  5%  slabinfo.avc_xperms_node.active_objs
    583.00 ±  8%     +18.8%     692.50 ±  5%  slabinfo.avc_xperms_node.num_objs
    176.03 ±  0%     -22.2%     136.95 ±  1%  time.user_time
    302905 ±  2%     -31.2%     208386 ±  0%  time.voluntary_context_switches
    263.42 ±  0%      -3.0%     255.52 ±  0%  turbostat.PkgWatt
     61.05 ±  0%     -12.7%      53.26 ±  0%  turbostat.RAMWatt
      1868 ± 16%     -43.7%       1052 ± 13%  cpuidle.C1-HSW.usage
      1499 ±  9%     -30.3%       1045 ± 12%  cpuidle.C3-HSW.usage
     16071 ±  4%     -15.0%      13664 ±  3%  cpuidle.C6-HSW.usage
     17572 ± 27%     -59.1%       7179 ±  5%  cpuidle.POLL.usage
 4.896e+08 ±  0%     -20.7%  3.884e+08 ±  0%  numa-numastat.node0.local_node
  71305376 ±  2%     -19.7%   57223573 ±  4%  numa-numastat.node0.numa_foreign
 4.896e+08 ±  0%     -20.7%  3.884e+08 ±  0%  numa-numastat.node0.numa_hit
  43760475 ±  3%     -22.1%   34074417 ±  5%  numa-numastat.node0.numa_miss
  43765010 ±  3%     -22.1%   34078937 ±  5%  numa-numastat.node0.other_node
 4.586e+08 ±  0%     -25.7%  3.408e+08 ±  1%  numa-numastat.node1.local_node
  43760472 ±  3%     -22.1%   34074417 ±  5%  numa-numastat.node1.numa_foreign
 4.586e+08 ±  0%     -25.7%  3.408e+08 ±  1%  numa-numastat.node1.numa_hit
  71305376 ±  2%     -19.7%   57223573 ±  4%  numa-numastat.node1.numa_miss
  71311721 ±  2%     -19.7%   57229904 ±  4%  numa-numastat.node1.other_node
    543.25 ±  3%     -15.0%     461.50 ±  3%  numa-vmstat.node0.nr_isolated_file
      2651 ± 19%     +30.2%       3451 ±  8%  numa-vmstat.node0.nr_mapped
      1226 ±  6%     -31.7%     837.25 ±  9%  numa-vmstat.node0.nr_pages_scanned
  37111278 ±  1%     -20.6%   29474561 ±  3%  numa-vmstat.node0.numa_foreign
 2.568e+08 ±  0%     -21.0%  2.028e+08 ±  0%  numa-vmstat.node0.numa_hit
 2.567e+08 ±  0%     -21.0%  2.027e+08 ±  0%  numa-vmstat.node0.numa_local
  22595209 ±  2%     -22.9%   17420980 ±  4%  numa-vmstat.node0.numa_miss
  22665391 ±  2%     -22.8%   17490378 ±  4%  numa-vmstat.node0.numa_other
     88.25 ±173%   +1029.7%     997.00 ± 63%  numa-vmstat.node0.workingset_activate
   3965715 ±  0%     -24.9%    2977998 ±  0%  numa-vmstat.node0.workingset_nodereclaim
     90.25 ±170%   +1006.4%     998.50 ± 63%  numa-vmstat.node0.workingset_refault
    612.50 ±  3%      -9.4%     554.75 ±  4%  numa-vmstat.node1.nr_alloc_batch
      3279 ± 14%     -34.1%       2161 ± 17%  numa-vmstat.node1.nr_mapped
  22597658 ±  2%     -22.9%   17423271 ±  4%  numa-vmstat.node1.numa_foreign
 2.403e+08 ±  0%     -25.9%  1.781e+08 ±  1%  numa-vmstat.node1.numa_hit
 2.403e+08 ±  0%     -25.9%  1.781e+08 ±  1%  numa-vmstat.node1.numa_local
  37115261 ±  1%     -20.6%   29478460 ±  3%  numa-vmstat.node1.numa_miss
  37136533 ±  1%     -20.6%   29500409 ±  3%  numa-vmstat.node1.numa_other
      6137 ±173%    +257.3%      21927 ± 60%  numa-vmstat.node1.workingset_activate
   3237162 ±  0%     -30.6%    2246385 ±  1%  numa-vmstat.node1.workingset_nodereclaim
      6139 ±173%    +257.2%      21930 ± 60%  numa-vmstat.node1.workingset_refault
    501243 ±  0%     -26.9%     366510 ±  1%  proc-vmstat.allocstall
     28483 ±  0%     -50.7%      14047 ±  3%  proc-vmstat.kswapd_low_wmark_hit_quickly
 1.151e+08 ±  0%     -20.7%   91297990 ±  0%  proc-vmstat.numa_foreign
 9.482e+08 ±  0%     -23.1%  7.293e+08 ±  0%  proc-vmstat.numa_hit
 9.482e+08 ±  0%     -23.1%  7.293e+08 ±  0%  proc-vmstat.numa_local
 1.151e+08 ±  0%     -20.7%   91297990 ±  0%  proc-vmstat.numa_miss
 1.151e+08 ±  0%     -20.7%   91308842 ±  0%  proc-vmstat.numa_other
     31562 ±  0%     -47.1%      16687 ±  2%  proc-vmstat.pageoutrun
 1.048e+09 ±  0%     -22.8%  8.088e+08 ±  0%  proc-vmstat.pgactivate
  28481000 ±  0%     -21.3%   22422907 ±  0%  proc-vmstat.pgalloc_dma32
 1.035e+09 ±  0%     -22.9%  7.984e+08 ±  0%  proc-vmstat.pgalloc_normal
 1.041e+09 ±  0%     -23.0%  8.024e+08 ±  0%  proc-vmstat.pgdeactivate
 1.063e+09 ±  0%     -22.8%    8.2e+08 ±  0%  proc-vmstat.pgfree
      2458 ± 91%     -93.5%     160.75 ± 29%  proc-vmstat.pgmigrate_success
  27571690 ±  0%     -20.6%   21889554 ±  0%  proc-vmstat.pgrefill_dma32
 1.014e+09 ±  0%     -23.0%  7.805e+08 ±  0%  proc-vmstat.pgrefill_normal
  25263166 ±  0%     -27.4%   18337251 ±  1%  proc-vmstat.pgscan_direct_dma32
 9.377e+08 ±  0%     -26.9%  6.852e+08 ±  1%  proc-vmstat.pgscan_direct_normal
   2134103 ±  1%     +57.6%    3363418 ±  6%  proc-vmstat.pgscan_kswapd_dma32
  69594167 ±  0%     +26.7%   88192786 ±  2%  proc-vmstat.pgscan_kswapd_normal
  25260851 ±  0%     -27.4%   18335464 ±  1%  proc-vmstat.pgsteal_direct_dma32
 9.376e+08 ±  0%     -26.9%  6.852e+08 ±  1%  proc-vmstat.pgsteal_direct_normal
   2133563 ±  1%     +57.6%    3362346 ±  6%  proc-vmstat.pgsteal_kswapd_dma32
  69585316 ±  0%     +26.7%   88176045 ±  2%  proc-vmstat.pgsteal_kswapd_normal
  17530080 ±  0%     -23.3%   13440416 ±  0%  proc-vmstat.slabs_scanned
      6226 ±173%    +268.2%      22924 ± 58%  proc-vmstat.workingset_activate
   7202139 ±  0%     -27.5%    5223203 ±  0%  proc-vmstat.workingset_nodereclaim
      6230 ±173%    +268.0%      22929 ± 58%  proc-vmstat.workingset_refault
    123.70 ± 12%     +26.7%     156.79 ± 11%  sched_debug.cfs_rq:/.load.stddev
     42.08 ±  1%     +23.3%      51.90 ±  8%  sched_debug.cfs_rq:/.load_avg.avg
    779.50 ±  2%     +20.7%     940.83 ±  5%  sched_debug.cfs_rq:/.load_avg.max
      9.46 ±  8%     -13.7%       8.17 ±  1%  sched_debug.cfs_rq:/.load_avg.min
    123.38 ±  2%     +31.4%     162.10 ±  6%  sched_debug.cfs_rq:/.load_avg.stddev
    304497 ± 22%     +65.6%     504169 ±  7%  sched_debug.cfs_rq:/.min_vruntime.stddev
     25.74 ±  8%     +33.9%      34.46 ±  8%  sched_debug.cfs_rq:/.runnable_load_avg.avg
    481.33 ± 11%     +50.5%     724.54 ± 11%  sched_debug.cfs_rq:/.runnable_load_avg.max
     69.65 ± 15%     +62.2%     112.95 ± 12%  sched_debug.cfs_rq:/.runnable_load_avg.stddev
  -1363122 ±-14%     +52.6%   -2080627 ±-10%  sched_debug.cfs_rq:/.spread0.min
    304448 ± 22%     +65.6%     504111 ±  7%  sched_debug.cfs_rq:/.spread0.stddev
    733220 ±  5%     +13.0%     828548 ±  1%  sched_debug.cpu.avg_idle.avg
    123344 ± 11%     +73.4%     213827 ± 27%  sched_debug.cpu.avg_idle.min
    233732 ±  5%     -13.5%     202264 ±  6%  sched_debug.cpu.avg_idle.stddev
     26.93 ±  9%     +27.8%      34.42 ±  8%  sched_debug.cpu.cpu_load[0].avg
     78.79 ± 19%     +43.7%     113.20 ± 12%  sched_debug.cpu.cpu_load[0].stddev
     26.23 ±  8%     +30.5%      34.23 ±  7%  sched_debug.cpu.cpu_load[1].avg
    513.17 ± 12%     +38.6%     711.12 ± 11%  sched_debug.cpu.cpu_load[1].max
     73.34 ± 15%     +50.7%     110.55 ± 11%  sched_debug.cpu.cpu_load[1].stddev
     25.93 ±  6%     +32.6%      34.40 ±  6%  sched_debug.cpu.cpu_load[2].avg
    488.38 ±  8%     +44.8%     706.96 ± 10%  sched_debug.cpu.cpu_load[2].max
     69.79 ± 10%     +56.9%     109.52 ± 10%  sched_debug.cpu.cpu_load[2].stddev
     25.89 ±  4%     +35.1%      34.97 ±  4%  sched_debug.cpu.cpu_load[3].avg
    467.83 ±  7%     +50.2%     702.71 ±  9%  sched_debug.cpu.cpu_load[3].max
     67.27 ±  9%     +63.6%     110.03 ±  8%  sched_debug.cpu.cpu_load[3].stddev
     25.83 ±  4%     +37.2%      35.44 ±  3%  sched_debug.cpu.cpu_load[4].avg
    445.29 ±  9%     +56.7%     697.88 ±  8%  sched_debug.cpu.cpu_load[4].max
     64.41 ±  9%     +72.4%     111.02 ±  6%  sched_debug.cpu.cpu_load[4].stddev
    123.66 ± 12%     +28.2%     158.54 ± 11%  sched_debug.cpu.load.stddev
      1.56 ±  1%      +9.8%       1.71 ±  0%  sched_debug.cpu.nr_running.avg
      0.46 ± 12%     +28.4%       0.59 ±  6%  sched_debug.cpu.nr_running.stddev
     57967 ±  3%      -9.8%      52290 ±  2%  sched_debug.cpu.nr_switches.avg
    270099 ±  9%     -16.4%     225748 ±  7%  sched_debug.cpu.nr_switches.max
     27370 ±  1%     -13.3%      23723 ±  0%  sched_debug.cpu.nr_switches.min
     55749 ±  7%     -14.3%      47767 ±  5%  sched_debug.cpu.nr_switches.stddev
    -55.33 ±-19%     -40.4%     -32.96 ± -2%  sched_debug.cpu.nr_uninterruptible.min

=========================================================================================
compiler/kconfig/rootfs/sleep/tbox_group/testcase:
  gcc-5/x86_64-randconfig-a0-04240012/yocto-minimal-i386.cgz/1/vm-kbuild-yocto-ia32/boot

commit: 
  612e44939c3c77245ac80843c0c7876c8cf97282
  23047a96d7cfcfca1a6d026ecaec526ea4803e9e

612e44939c3c7724 23047a96d7cfcfca1a6d026eca 
---------------- -------------------------- 
       fail:runs  %reproduction    fail:runs
           |             |             |    
           :50           2%           1:180   kmsg.augmented_rbtree_testing
           :50         216%         108:180   last_state.is_incomplete_run



                           vm-scalability.time.user_time

  180 *+**-*-----*-*-*--*-*---*----*-*-**------**-*-**-**------*-**---------+
  175 ++     **.*     *    *.*  *.*       **.*            *.**     +  .*.**.*
      |                                                             **      |
  170 ++                                                                    |
  165 ++                                                                    |
      |                                                                     |
  160 ++                                                                    |
  155 ++                                                                    |
  150 ++                                                                    |
      | O                                                                   |
  145 O+ O O OO    O                                                        |
  140 ++        OO                                     O                    |
      |                   OO O  O  O O     O   OO O O                       |
  135 ++              O O     O   O    OO O  O       O                      |
  130 ++-------------O------------------------------------------------------+


                               vm-scalability.throughput

  2.9e+07 ++------*--------------------------*---------------*--------------+
          |.*   *  *.* .**.**.* .*.**. *.  .*  **. *. .*   *  *.* .* .**.**.*
  2.8e+07 *+ *.*      *        *      *  **       *  *  *.*      *  *       |
  2.7e+07 ++                                                                |
          |                                                                 |
  2.6e+07 ++                                                                |
          |                                                                 |
  2.5e+07 ++                                                                |
          |                                                                 |
  2.4e+07 ++                                                                |
  2.3e+07 ++O  OO     O                                                     |
          O  O    OO O                                                      |
  2.2e+07 ++                                                                |
          |             OO OO OO O OO OO OO OO OO OO O OO                   |
  2.1e+07 ++----------------------------------------------------------------+



	[*] bisect-good sample
	[O] bisect-bad  sample

To reproduce:

        git clone git://git.kernel.org/pub/scm/linux/kernel/git/wfg/lkp-tests.git
        cd lkp-tests
        bin/lkp install job.yaml  # job file is attached in this email
        bin/lkp run     job.yaml


Disclaimer:
Results have been estimated based on internal Intel analysis and are provided
for informational purposes only. Any difference in system hardware or software
design or configuration may affect actual performance.


Thanks,
Xiaolong

View attachment "job.yaml" of type "text/plain" (3531 bytes)

View attachment "reproduce" of type "text/plain" (12116 bytes)

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ