lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [day] [month] [year] [list]
Date:	Mon, 16 Feb 2015 15:25:14 +0800
From:	Huang Ying <ying.huang@...el.com>
To:	Trond Myklebust <trond.myklebust@...marydata.com>
Cc:	LKML <linux-kernel@...r.kernel.org>, LKP ML <lkp@...org>
Subject: [LKP] [SUNRPC] 127b21b89f9: nfs loopback setup: +129.7%
 fsmark.app_overhead, +14.2% fsmark.time.involuntary_context_switches

FYI, we noticed the below changes on

git://git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux.git master
commit 127b21b89f9d8ba0dc23e47b8c35d8a0bac9d6fc ("SUNRPC: Adjust rpciod workqueue parameters")


testbox/testcase/testparams: nhm4/fsmark/performance-1x-32t-1HDD-xfs-nfsv4-5K-400M-fsyncBeforeClose-16d-256fpd

40dd4b7aee1a8c3b  127b21b89f9d8ba0dc23e47b8c  
----------------  --------------------------  
         %stddev     %change         %stddev
             \          |                \  
   8133446 ±  0%      -1.3%    8023711 ±  0%  fsmark.app_overhead
     52620 ±  0%     +60.6%      84523 ±  0%  fsmark.time.involuntary_context_switches
         5 ± 37%    +708.7%         46 ± 16%  sched_debug.cpu#1.cpu_load[4]
         7 ± 47%    +906.9%         73 ± 46%  sched_debug.cpu#3.cpu_load[3]
         7 ± 33%    +528.6%         44 ± 21%  sched_debug.cpu#1.cpu_load[2]
         6 ± 32%    +577.8%         45 ± 19%  sched_debug.cpu#1.cpu_load[3]
         7 ± 39%    +753.6%         59 ± 27%  sched_debug.cpu#3.cpu_load[4]
         8 ± 39%    +434.4%         42 ± 41%  sched_debug.cpu#2.cpu_load[4]
         7 ± 29%     -67.7%          2 ± 34%  sched_debug.cfs_rq[1]:/.nr_spread_over
       475 ± 42%    +264.9%       1733 ± 17%  sched_debug.cfs_rq[2]:/.blocked_load_avg
         9 ± 46%    +254.1%         32 ± 42%  sched_debug.cpu#5.cpu_load[4]
       495 ± 41%    +260.8%       1787 ± 17%  sched_debug.cfs_rq[2]:/.tg_load_contrib
       450 ± 26%    +193.5%       1321 ± 14%  sched_debug.cfs_rq[3]:/.tg_load_contrib
       440 ± 29%    +194.7%       1296 ± 14%  sched_debug.cfs_rq[3]:/.blocked_load_avg
       586 ± 30%    +129.7%       1345 ± 17%  sched_debug.cfs_rq[1]:/.blocked_load_avg
       617 ± 27%    +144.2%       1506 ± 15%  sched_debug.cfs_rq[1]:/.tg_load_contrib
       936 ± 28%     +91.6%       1793 ± 16%  sched_debug.cfs_rq[5]:/.tg_load_contrib
       922 ± 29%     +92.0%       1771 ± 16%  sched_debug.cfs_rq[5]:/.blocked_load_avg
    144242 ± 26%     +70.3%     245615 ± 14%  sched_debug.cpu#3.ttwu_local
    159375 ± 23%     +61.9%     257969 ± 13%  sched_debug.cpu#3.sched_goidle
      1754 ± 29%     +61.0%       2824 ± 24%  sched_debug.cfs_rq[0]:/.blocked_load_avg
    171393 ± 22%     +59.6%     273569 ± 13%  sched_debug.cpu#3.ttwu_count
    346764 ± 22%     +58.3%     548977 ± 13%  sched_debug.cpu#3.nr_switches
    346840 ± 22%     +58.3%     549053 ± 13%  sched_debug.cpu#3.sched_count
      1975 ± 19%     +50.7%       2976 ± 19%  sched_debug.cfs_rq[0]:/.tg_load_contrib
     17240 ± 12%     +37.6%      23728 ±  8%  sched_debug.cpu#3.nr_load_updates
     52620 ±  0%     +60.6%      84523 ±  0%  time.involuntary_context_switches
      9440 ±  6%     +61.6%      15259 ±  5%  sched_debug.cfs_rq[0]:/.tg_load_avg
      5200 ± 16%     +49.2%       7759 ± 18%  sched_debug.cfs_rq[3]:/.min_vruntime
      9394 ±  6%     +60.8%      15110 ±  6%  sched_debug.cfs_rq[2]:/.tg_load_avg
      9408 ±  6%     +62.2%      15260 ±  5%  sched_debug.cfs_rq[1]:/.tg_load_avg
      9387 ±  6%     +59.8%      14999 ±  7%  sched_debug.cfs_rq[3]:/.tg_load_avg
      9370 ±  6%     +59.7%      14961 ±  6%  sched_debug.cfs_rq[7]:/.tg_load_avg
      9371 ±  6%     +58.3%      14832 ±  7%  sched_debug.cfs_rq[6]:/.tg_load_avg
      9389 ±  6%     +59.0%      14924 ±  6%  sched_debug.cfs_rq[5]:/.tg_load_avg
      9391 ±  6%     +59.1%      14937 ±  6%  sched_debug.cfs_rq[4]:/.tg_load_avg
      2492 ± 10%     +22.8%       3059 ±  8%  sched_debug.cfs_rq[3]:/.exec_clock
    654180 ± 10%     -17.6%     539086 ± 10%  sched_debug.cpu#3.avg_idle
        67 ± 22%     +29.9%         88 ±  4%  sched_debug.cfs_rq[3]:/.tg_runnable_contrib
    610001 ±  7%     -18.3%     498400 ±  8%  sched_debug.cpu#0.avg_idle
    132930 ± 23%     +39.8%     185843 ±  4%  sched_debug.cpu#2.ttwu_local
      4543 ± 14%     +34.5%       6112 ±  4%  sched_debug.cfs_rq[2]:/.min_vruntime
    161157 ± 19%     +33.0%     214350 ±  3%  sched_debug.cpu#2.ttwu_count
    148740 ± 20%     +33.5%     198618 ±  3%  sched_debug.cpu#2.sched_goidle
    325967 ± 19%     +32.1%     430669 ±  3%  sched_debug.cpu#2.nr_switches
    326041 ± 19%     +32.1%     430744 ±  3%  sched_debug.cpu#2.sched_count
     17437 ± 14%     +21.7%      21219 ±  8%  sched_debug.cpu#2.nr_load_updates
      3598 ±  7%     +14.9%       4134 ±  5%  sched_debug.cfs_rq[0]:/.avg->runnable_avg_sum
      3338 ±  8%     +18.6%       3959 ±  9%  sched_debug.cfs_rq[1]:/.avg->runnable_avg_sum
      1018 ± 31%     -23.9%        775 ±  2%  sched_debug.cfs_rq[4]:/.exec_clock
        72 ±  8%     +18.3%         85 ±  9%  sched_debug.cfs_rq[1]:/.tg_runnable_contrib
      8293 ± 45%     -30.4%       5770 ±  4%  sched_debug.cpu#4.nr_load_updates
      2420 ±  7%     +13.7%       2752 ±  0%  sched_debug.cfs_rq[2]:/.exec_clock
      9386 ±  1%      +2.3%       9598 ±  1%  vmstat.io.bo
     40903 ±  0%      +2.6%      41951 ±  0%  vmstat.system.cs

testbox/testcase/testparams: nhm4/fsmark/performance-1x-32t-1HDD-xfs-nfsv4-16MB-60G-NoSync-16d-256fpd

40dd4b7aee1a8c3b  127b21b89f9d8ba0dc23e47b8c  
----------------  --------------------------  
  50440392 ±  6%    +129.7%  1.158e+08 ±  8%  fsmark.app_overhead
    226074 ±  0%     +14.2%     258118 ±  0%  fsmark.time.involuntary_context_switches
         8 ±  5%     +21.2%         10 ±  0%  fsmark.time.percent_of_cpu_this_job_got
     60.69 ±  0%     +13.5%      68.86 ±  0%  fsmark.time.system_time
   2523773 ±  0%      +1.4%    2558878 ±  0%  fsmark.time.voluntary_context_switches
         8 ± 39%    +188.2%         24 ± 24%  sched_debug.cpu#2.cpu_load[3]
         8 ± 31%    +197.1%         25 ± 13%  sched_debug.cpu#2.cpu_load[4]
    283664 ± 13%    +237.6%     957681 ±  7%  proc-vmstat.pgsteal_direct_dma32
     30933 ± 17%    +205.1%      94381 ± 30%  sched_debug.cpu#5.ttwu_local
      2119 ± 14%    +231.0%       7014 ±  7%  proc-vmstat.allocstall
        16 ± 36%     +59.1%         26 ±  3%  sched_debug.cpu#0.cpu_load[4]
        18 ± 30%     +85.3%         34 ± 26%  sched_debug.cpu#5.cpu_load[4]
    213724 ±  4%     +22.0%     260658 ± 17%  sched_debug.cpu#6.ttwu_count
    330125 ± 12%    +195.4%     975107 ±  8%  proc-vmstat.pgscan_direct_dma32
       471 ± 49%    +156.7%       1210 ± 34%  sched_debug.cpu#1.curr->pid
      1717 ± 11%    +131.8%       3981 ±  2%  proc-vmstat.nr_dirty
       217 ±  9%    +136.3%        514 ± 13%  sched_debug.cfs_rq[1]:/.blocked_load_avg
       229 ± 10%    +138.3%        546 ± 18%  sched_debug.cfs_rq[1]:/.tg_load_contrib
      7162 ± 13%    +121.3%      15849 ±  1%  meminfo.Dirty
       521 ± 11%     +81.3%        945 ± 19%  sched_debug.cfs_rq[5]:/.tg_load_contrib
       506 ± 12%     +84.3%        932 ± 19%  sched_debug.cfs_rq[5]:/.blocked_load_avg
         3 ±  0%     +66.7%          5 ±  0%  vmstat.procs.b
       537 ± 20%     +37.6%        739 ± 10%  sched_debug.cfs_rq[0]:/.blocked_load_avg
       574 ± 19%     +41.1%        810 ± 12%  sched_debug.cfs_rq[0]:/.tg_load_contrib
       195 ±  6%     -33.7%        129 ± 11%  sched_debug.cpu#0.nr_uninterruptible
    226074 ±  0%     +14.2%     258118 ±  0%  time.involuntary_context_switches
  29991075 ± 14%     -22.0%   23379253 ± 19%  cpuidle.POLL.time
     54075 ±  2%     +42.5%      77063 ±  1%  meminfo.Writeback
      3770 ±  6%     +51.9%       5725 ±  8%  sched_debug.cfs_rq[0]:/.tg_load_avg
      3762 ±  6%     +50.0%       5645 ±  7%  sched_debug.cfs_rq[2]:/.tg_load_avg
      3768 ±  6%     +50.9%       5687 ±  7%  sched_debug.cfs_rq[1]:/.tg_load_avg
      3767 ±  6%     +49.3%       5624 ±  7%  sched_debug.cfs_rq[3]:/.tg_load_avg
      3756 ±  6%     +48.7%       5585 ±  8%  sched_debug.cfs_rq[7]:/.tg_load_avg
      3762 ±  6%     +48.9%       5602 ±  8%  sched_debug.cfs_rq[6]:/.tg_load_avg
      3762 ±  6%     +48.6%       5591 ±  8%  sched_debug.cfs_rq[5]:/.tg_load_avg
      3761 ±  6%     +50.1%       5647 ±  7%  sched_debug.cfs_rq[4]:/.tg_load_avg
     13632 ±  3%     +40.9%      19210 ±  1%  proc-vmstat.nr_writeback
     10367 ±  3%     +29.3%      13406 ±  9%  sched_debug.cfs_rq[6]:/.min_vruntime
     89961 ±  1%     +35.0%     121410 ±  1%  meminfo.NFS_Unstable
     22527 ±  1%     +35.6%      30545 ±  1%  proc-vmstat.nr_unstable
    207975 ±  0%     +34.6%     280001 ±  8%  sched_debug.cpu#5.ttwu_count
     10227 ±  4%     +30.6%      13360 ±  2%  sched_debug.cfs_rq[5]:/.min_vruntime
    143095 ±  4%     -25.8%     106195 ±  4%  proc-vmstat.pgactivate
    759903 ± 15%     +19.2%     905820 ± 13%  sched_debug.cpu#0.nr_switches
    760233 ± 15%     +19.2%     906200 ± 13%  sched_debug.cpu#0.sched_count
       589 ±  7%     +27.7%        753 ±  8%  sched_debug.cfs_rq[4]:/.blocked_load_avg
         8 ±  5%     +21.2%         10 ±  0%  time.percent_of_cpu_this_job_got
      3703 ±  5%     +41.5%       5241 ± 31%  proc-vmstat.pgsteal_kswapd_dma
       613 ±  6%     +27.1%        779 ±  9%  sched_debug.cfs_rq[4]:/.tg_load_contrib
     18002 ±  9%     +24.5%      22412 ± 10%  sched_debug.cfs_rq[0]:/.min_vruntime
      4051 ±  4%     +38.7%       5619 ± 29%  proc-vmstat.pgscan_kswapd_dma
       570 ± 16%     +30.2%        742 ± 19%  sched_debug.cfs_rq[7]:/.tg_load_contrib
     67791 ±  3%     -20.1%      54186 ±  9%  proc-vmstat.kswapd_inodesteal
      2168 ±  1%     +22.0%       2645 ±  0%  slabinfo.kmalloc-128.active_slabs
      2168 ±  1%     +22.0%       2645 ±  0%  slabinfo.kmalloc-128.num_slabs
     69411 ±  1%     +22.0%      84660 ±  0%  slabinfo.kmalloc-128.num_objs
        13 ± 23%     -44.4%          7 ± 34%  sched_debug.cfs_rq[0]:/.nr_spread_over
     66478 ±  1%     +18.6%      78816 ±  0%  slabinfo.kmalloc-128.active_objs
      0.50 ±  3%     +14.0%       0.57 ±  5%  time.user_time
      4983 ±  4%     +31.3%       6541 ± 26%  proc-vmstat.pgalloc_dma
     60407 ±  3%     -15.5%      51037 ±  3%  proc-vmstat.nr_active_file
    241585 ±  3%     -15.5%     204154 ±  3%  meminfo.Active(file)
       536 ±  3%     -13.1%        466 ±  2%  proc-vmstat.kswapd_high_wmark_hit_quickly
      3892 ± 18%     -22.4%       3018 ±  0%  uptime.idle
    264741 ±  3%     -13.8%     228127 ±  3%  meminfo.Active
     60.69 ±  0%     +13.5%      68.86 ±  0%  time.system_time
     11151 ±  6%     +19.0%      13267 ±  1%  sched_debug.cfs_rq[4]:/.min_vruntime
    500911 ±  1%     -10.7%     447478 ±  0%  cpuidle.C6-NHM.usage
 4.834e+08 ±  3%     +11.3%   5.38e+08 ±  2%  cpuidle.C3-NHM.time
     19446 ±  0%      +0.9%      19618 ±  0%  vmstat.system.cs

testbox/testcase/testparams: nhm4/fsmark/performance-1x-32t-1HDD-xfs-nfsv4-8K-400M-fsyncBeforeClose-16d-256fpd

40dd4b7aee1a8c3b  127b21b89f9d8ba0dc23e47b8c  
----------------  --------------------------  
     33823 ±  0%     +56.8%      53052 ±  0%  fsmark.time.involuntary_context_switches
     49306 ± 44%     -75.7%      11963 ± 24%  sched_debug.cpu#6.sched_goidle
         1 ± 34%    +440.0%          6 ± 46%  sched_debug.cfs_rq[4]:/.nr_spread_over
    105610 ± 41%     -69.9%      31780 ± 17%  sched_debug.cpu#6.nr_switches
    105651 ± 41%     -69.9%      31817 ± 17%  sched_debug.cpu#6.sched_count
        11 ± 20%     -47.7%          5 ± 33%  sched_debug.cfs_rq[1]:/.nr_spread_over
     44125 ± 49%     -82.9%       7541 ± 41%  sched_debug.cpu#6.ttwu_local
     51122 ± 42%     -71.1%      14769 ± 19%  sched_debug.cpu#6.ttwu_count
      5455 ± 26%     -45.5%       2974 ±  5%  sched_debug.cpu#6.nr_load_updates
       800 ± 34%    +148.9%       1991 ± 37%  sched_debug.cfs_rq[0]:/.blocked_load_avg
      1031 ± 26%    +114.1%       2209 ± 34%  sched_debug.cfs_rq[0]:/.tg_load_contrib
         8 ± 32%     +45.7%         12 ± 18%  sched_debug.cfs_rq[2]:/.nr_spread_over
     33823 ±  0%     +56.8%      53052 ±  0%  time.involuntary_context_switches
     10789 ±  9%     +25.4%      13532 ±  5%  sched_debug.cfs_rq[0]:/.tg_load_avg
     10783 ±  9%     +25.2%      13497 ±  5%  sched_debug.cfs_rq[2]:/.tg_load_avg
     10777 ±  9%     +25.2%      13494 ±  5%  sched_debug.cfs_rq[1]:/.tg_load_avg
     10799 ±  9%     +24.9%      13490 ±  5%  sched_debug.cfs_rq[3]:/.tg_load_avg
     10734 ±  9%     +25.4%      13456 ±  5%  sched_debug.cfs_rq[7]:/.tg_load_avg
     10743 ±  9%     +25.6%      13495 ±  5%  sched_debug.cfs_rq[6]:/.tg_load_avg
     10743 ±  9%     +25.7%      13502 ±  5%  sched_debug.cfs_rq[5]:/.tg_load_avg
     10798 ±  9%     +24.8%      13481 ±  5%  sched_debug.cfs_rq[4]:/.tg_load_avg
       796 ± 14%     -21.0%        629 ±  5%  sched_debug.cfs_rq[6]:/.exec_clock
      2161 ± 17%     -25.4%       1612 ±  5%  sched_debug.cfs_rq[6]:/.min_vruntime
    297542 ±  9%     -22.0%     232177 ± 17%  sched_debug.cpu#0.nr_switches
    297589 ±  9%     -22.0%     232227 ± 17%  sched_debug.cpu#0.sched_count
      1901 ±  8%     +23.2%       2342 ± 10%  sched_debug.cfs_rq[6]:/.tg_load_contrib
      7060 ± 10%     -13.3%       6122 ±  7%  sched_debug.cfs_rq[0]:/.min_vruntime
    748854 ±  5%     -20.0%     599063 ±  6%  sched_debug.cpu#0.avg_idle
      1851 ±  6%     +24.1%       2298 ± 10%  sched_debug.cfs_rq[6]:/.blocked_load_avg
    140093 ±  9%     -23.4%     107299 ± 19%  sched_debug.cpu#0.ttwu_local
    141664 ±  9%     -23.0%     109136 ± 18%  sched_debug.cpu#0.sched_goidle
        52 ± 13%     +29.7%         67 ±  2%  sched_debug.cfs_rq[2]:/.tg_runnable_contrib
      2440 ± 13%     +28.1%       3126 ±  2%  sched_debug.cfs_rq[2]:/.avg->runnable_avg_sum
    165266 ±  8%     -19.5%     133017 ± 15%  sched_debug.cpu#0.ttwu_count
     17128 ±  5%     -14.3%      14673 ±  8%  sched_debug.cpu#0.nr_load_updates
     34290 ±  0%      +2.0%      34989 ±  0%  vmstat.system.cs

nhm4: Nehalem
Memory: 4G

                          time.involuntary_context_switches

  60000 ++------------------------------------------------------------------+
        |                                                                   |
  55000 O+OO O O OO O OO O OO O O OO O OO O OO O O OO O OO O OO O O OO      |
        |                                                              O OO O
        |                                                                   |
  50000 ++                                                                  |
        |                                                                   |
  45000 ++                                                                  |
        |                                                                   |
  40000 ++                                                                  |
        |                                                                   |
        |                                                                   |
  35000 ++*   .*.**.*.* .*.**.   .**.*.  .*.* .*.*.      *. .**.*.*.*       |
        *  *.*         *      *.*      **    *     **.*.*  *                |
  30000 ++------------------------------------------------------------------+

	[*] bisect-good sample
	[O] bisect-bad  sample

To reproduce:

	apt-get install ruby
	git clone git://git.kernel.org/pub/scm/linux/kernel/git/wfg/lkp-tests.git
	cd lkp-tests
	bin/setup-local job.yaml # the job file attached in this email
	bin/run-local   job.yaml


Disclaimer:
Results have been estimated based on internal Intel analysis and are provided
for informational purposes only. Any difference in system hardware or software
design or configuration may affect actual performance.


Thanks,
Ying Huang


View attachment "job.yaml" of type "text/plain" (2010 bytes)

View attachment "reproduce" of type "text/plain" (770 bytes)

_______________________________________________
LKP mailing list
LKP@...ux.intel.com

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ