lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite for Android: free password hash cracker in your pocket
[<prev] [next>] [day] [month] [year] [list]
Date:	Wed, 1 Oct 2014 15:50:20 +0800
From:	Fengguang Wu <fengguang.wu@...el.com>
To:	Dirk Brandewie <dirk.j.brandewie@...el.com>
Cc:	Dave Hansen <dave.hansen@...el.com>,
	"Rafael J. Wysocki" <rafael.j.wysocki@...el.com>,
	LKML <linux-kernel@...r.kernel.org>, lkp@...org
Subject: [intel_pstate] adacdf3f2b8: +119.9% aim9.shell_rtns_3.ops_per_sec,
 +51.6% turbostat.Pkg_W

Hi Dirk,

FYI, we noticed the below changes on commit
adacdf3f2b8e65aa441613cf61c4f598e9042690 ("intel_pstate: Remove C0 tracking")

test case: brickland3/aim9/300s-shell_rtns_3

      v3.15-rc8  adacdf3f2b8e65aa441613cf6 
---------------  ------------------------- 
       125 ± 5%    +119.9%        275 ± 1%  TOTAL aim9.shell_rtns_3.ops_per_sec
     96.81 ± 3%     +51.6%     146.77 ± 5%  TOTAL turbostat.Pkg_W
     36.74 ± 9%    +121.4%      81.34 ±10%  TOTAL turbostat.Cor_W
     38.36 ± 1%     +63.6%      62.76 ± 0%  TOTAL turbostat.RAM_W
    -13794 ±-5%     -13.4%     -11946 ±-6%  TOTAL sched_debug.cfs_rq[1]:/.spread0
    -10828 ±-8%     -34.7%      -7069 ±-20%  TOTAL sched_debug.cfs_rq[34]:/.spread0
    -14141 ±-8%     -19.1%     -11441 ±-19%  TOTAL sched_debug.cfs_rq[24]:/.spread0
    -13819 ±-7%     -13.6%     -11944 ±-6%  TOTAL sched_debug.cfs_rq[7]:/.spread0
      6006 ± 8%     -71.6%       1703 ±21%  TOTAL sched_debug.cpu#33.ttwu_local
      6281 ±36%     -68.3%       1988 ±49%  TOTAL sched_debug.cpu#7.ttwu_count
      3177 ±47%    +235.5%      10660 ± 6%  TOTAL cpuidle.C1-IVT-4S.usage
       268 ±45%     -68.5%         84 ±31%  TOTAL sched_debug.cpu#45.ttwu_local
      6658 ±34%     -66.1%       2260 ±30%  TOTAL sched_debug.cpu#21.ttwu_count
       292 ±44%     -71.7%         82 ±11%  TOTAL sched_debug.cpu#23.ttwu_local
      5351 ± 6%     -61.8%       2045 ±29%  TOTAL sched_debug.cpu#48.ttwu_local
      2395 ±29%     -62.9%        888 ±14%  TOTAL sched_debug.cpu#37.ttwu_count
      2269 ±11%    +144.0%       5537 ±26%  TOTAL sched_debug.cfs_rq[91]:/.blocked_load_avg
      2040 ±17%    +154.5%       5192 ±14%  TOTAL sched_debug.cfs_rq[106]:/.blocked_load_avg
      1.24 ± 6%    +154.8%       3.15 ± 0%  TOTAL turbostat.GHz
      2417 ±10%    +135.2%       5685 ±25%  TOTAL sched_debug.cfs_rq[91]:/.tg_load_contrib
        69 ±29%     -52.4%         33 ±36%  TOTAL sched_debug.cfs_rq[41]:/.avg->runnable_avg_sum
      4422 ±29%     -57.6%       1875 ± 9%  TOTAL sched_debug.cpu#3.ttwu_local
      2210 ±14%    +140.9%       5324 ±14%  TOTAL sched_debug.cfs_rq[106]:/.tg_load_contrib
      1445 ±16%    +126.6%       3276 ±14%  TOTAL sched_debug.cfs_rq[3]:/.blocked_load_avg
      1448 ±16%    +127.4%       3293 ±14%  TOTAL sched_debug.cfs_rq[3]:/.tg_load_contrib
    561248 ± 4%    +145.3%    1376953 ± 0%  TOTAL cpuidle.C6-IVT-4S.usage
      4975 ±28%     -63.7%       1805 ±13%  TOTAL sched_debug.cpu#4.ttwu_local
      1348 ±19%    +137.8%       3206 ±12%  TOTAL sched_debug.cfs_rq[48]:/.blocked_load_avg
      1696 ±13%    +106.7%       3507 ±15%  TOTAL sched_debug.cfs_rq[32]:/.tg_load_contrib
      1684 ±13%    +106.6%       3478 ±15%  TOTAL sched_debug.cfs_rq[32]:/.blocked_load_avg
      1619 ±22%    +118.1%       3532 ±13%  TOTAL sched_debug.cfs_rq[17]:/.blocked_load_avg
      1626 ±22%    +117.4%       3537 ±13%  TOTAL sched_debug.cfs_rq[17]:/.tg_load_contrib
      1354 ±19%    +137.0%       3209 ±12%  TOTAL sched_debug.cfs_rq[48]:/.tg_load_contrib
     21314 ± 5%    +125.6%      48083 ± 2%  TOTAL sched_debug.cfs_rq[85]:/.tg_load_avg
     21409 ± 5%    +125.1%      48199 ± 1%  TOTAL sched_debug.cfs_rq[83]:/.tg_load_avg
     21340 ± 5%    +125.8%      48193 ± 1%  TOTAL sched_debug.cfs_rq[84]:/.tg_load_avg
     21291 ± 5%    +125.7%      48060 ± 2%  TOTAL sched_debug.cfs_rq[86]:/.tg_load_avg
     21191 ± 6%    +126.2%      47929 ± 1%  TOTAL sched_debug.cfs_rq[102]:/.tg_load_avg
     21266 ± 5%    +126.0%      48058 ± 2%  TOTAL sched_debug.cfs_rq[90]:/.tg_load_avg
     21289 ± 5%    +125.7%      48054 ± 2%  TOTAL sched_debug.cfs_rq[89]:/.tg_load_avg
     21186 ± 6%    +126.2%      47929 ± 1%  TOTAL sched_debug.cfs_rq[101]:/.tg_load_avg
     21314 ± 6%    +125.8%      48131 ± 1%  TOTAL sched_debug.cfs_rq[113]:/.tg_load_avg
     21266 ± 5%    +126.1%      48083 ± 2%  TOTAL sched_debug.cfs_rq[91]:/.tg_load_avg
     21298 ± 6%    +125.3%      47981 ± 1%  TOTAL sched_debug.cfs_rq[106]:/.tg_load_avg
     21309 ± 6%    +125.0%      47953 ± 1%  TOTAL sched_debug.cfs_rq[105]:/.tg_load_avg
     21178 ± 6%    +126.4%      47953 ± 1%  TOTAL sched_debug.cfs_rq[100]:/.tg_load_avg
     21236 ± 6%    +126.5%      48095 ± 1%  TOTAL sched_debug.cfs_rq[93]:/.tg_load_avg
     21271 ± 5%    +126.0%      48077 ± 1%  TOTAL sched_debug.cfs_rq[88]:/.tg_load_avg
     21286 ± 5%    +126.0%      48096 ± 2%  TOTAL sched_debug.cfs_rq[87]:/.tg_load_avg
     21269 ± 6%    +126.1%      48093 ± 1%  TOTAL sched_debug.cfs_rq[92]:/.tg_load_avg
     21291 ± 6%    +125.2%      47956 ± 1%  TOTAL sched_debug.cfs_rq[104]:/.tg_load_avg
     21303 ± 6%    +125.3%      48005 ± 1%  TOTAL sched_debug.cfs_rq[107]:/.tg_load_avg
     21247 ± 6%    +125.7%      47957 ± 1%  TOTAL sched_debug.cfs_rq[94]:/.tg_load_avg
     21350 ± 6%    +125.7%      48185 ± 1%  TOTAL sched_debug.cfs_rq[119]:/.tg_load_avg
     21357 ± 6%    +125.3%      48108 ± 1%  TOTAL sched_debug.cfs_rq[114]:/.tg_load_avg
     21263 ± 6%    +125.6%      47968 ± 1%  TOTAL sched_debug.cfs_rq[103]:/.tg_load_avg
     21362 ± 6%    +125.4%      48154 ± 1%  TOTAL sched_debug.cfs_rq[118]:/.tg_load_avg
     21470 ± 5%    +124.6%      48223 ± 1%  TOTAL sched_debug.cfs_rq[82]:/.tg_load_avg
      1513 ±24%    +122.0%       3358 ±16%  TOTAL sched_debug.cfs_rq[47]:/.blocked_load_avg
     21170 ± 6%    +126.4%      47936 ± 1%  TOTAL sched_debug.cfs_rq[98]:/.tg_load_avg
     21216 ± 6%    +126.0%      47943 ± 1%  TOTAL sched_debug.cfs_rq[95]:/.tg_load_avg
     21183 ± 6%    +126.3%      47936 ± 1%  TOTAL sched_debug.cfs_rq[97]:/.tg_load_avg
     21351 ± 6%    +125.3%      48106 ± 1%  TOTAL sched_debug.cfs_rq[115]:/.tg_load_avg
     21194 ± 6%    +126.2%      47952 ± 1%  TOTAL sched_debug.cfs_rq[96]:/.tg_load_avg
     21181 ± 6%    +126.3%      47929 ± 1%  TOTAL sched_debug.cfs_rq[99]:/.tg_load_avg
     21366 ± 6%    +125.1%      48101 ± 1%  TOTAL sched_debug.cfs_rq[116]:/.tg_load_avg
     21352 ± 6%    +125.5%      48145 ± 1%  TOTAL sched_debug.cfs_rq[112]:/.tg_load_avg
     21381 ± 6%    +125.1%      48131 ± 1%  TOTAL sched_debug.cfs_rq[117]:/.tg_load_avg
     21507 ± 5%    +124.3%      48244 ± 1%  TOTAL sched_debug.cfs_rq[81]:/.tg_load_avg
     21346 ± 6%    +125.5%      48126 ± 1%  TOTAL sched_debug.cfs_rq[111]:/.tg_load_avg
     22339 ± 4%    +124.5%      50156 ± 1%  TOTAL sched_debug.cfs_rq[5]:/.tg_load_avg
     21569 ± 5%    +123.7%      48256 ± 1%  TOTAL sched_debug.cfs_rq[80]:/.tg_load_avg
     21343 ± 6%    +125.0%      48018 ± 1%  TOTAL sched_debug.cfs_rq[108]:/.tg_load_avg
      1528 ±23%    +120.8%       3373 ±16%  TOTAL sched_debug.cfs_rq[47]:/.tg_load_contrib
     21616 ± 5%    +123.2%      48245 ± 2%  TOTAL sched_debug.cfs_rq[78]:/.tg_load_avg
     21595 ± 4%    +124.7%      48525 ± 1%  TOTAL sched_debug.cfs_rq[41]:/.tg_load_avg
     21622 ± 5%    +123.4%      48294 ± 2%  TOTAL sched_debug.cfs_rq[77]:/.tg_load_avg
     21571 ± 4%    +123.7%      48245 ± 2%  TOTAL sched_debug.cfs_rq[79]:/.tg_load_avg
     22460 ± 4%    +124.3%      50377 ± 1%  TOTAL sched_debug.cfs_rq[4]:/.tg_load_avg
     22257 ± 5%    +124.2%      49910 ± 1%  TOTAL sched_debug.cfs_rq[8]:/.tg_load_avg
     22291 ± 5%    +124.0%      49922 ± 1%  TOTAL sched_debug.cfs_rq[7]:/.tg_load_avg
     22586 ± 4%    +123.3%      50430 ± 1%  TOTAL sched_debug.cfs_rq[3]:/.tg_load_avg
     22236 ± 5%    +124.1%      49831 ± 1%  TOTAL sched_debug.cfs_rq[9]:/.tg_load_avg
     21599 ± 4%    +124.4%      48473 ± 1%  TOTAL sched_debug.cfs_rq[42]:/.tg_load_avg
     22118 ± 6%    +123.8%      49501 ± 1%  TOTAL sched_debug.cfs_rq[12]:/.tg_load_avg
     21591 ± 4%    +124.8%      48544 ± 1%  TOTAL sched_debug.cfs_rq[40]:/.tg_load_avg
     21348 ± 6%    +125.3%      48090 ± 1%  TOTAL sched_debug.cfs_rq[110]:/.tg_load_avg
     21636 ± 4%    +123.4%      48331 ± 1%  TOTAL sched_debug.cfs_rq[43]:/.tg_load_avg
     22170 ± 5%    +123.5%      49543 ± 1%  TOTAL sched_debug.cfs_rq[11]:/.tg_load_avg
     22117 ± 6%    +123.8%      49505 ± 1%  TOTAL sched_debug.cfs_rq[13]:/.tg_load_avg
     21656 ± 5%    +122.8%      48260 ± 1%  TOTAL sched_debug.cfs_rq[44]:/.tg_load_avg
     22206 ± 5%    +123.4%      49613 ± 1%  TOTAL sched_debug.cfs_rq[10]:/.tg_load_avg
     22307 ± 4%    +124.3%      50042 ± 1%  TOTAL sched_debug.cfs_rq[6]:/.tg_load_avg
      1438 ±22%    +123.8%       3218 ±13%  TOTAL sched_debug.cfs_rq[18]:/.blocked_load_avg
     21389 ± 6%    +124.7%      48064 ± 1%  TOTAL sched_debug.cfs_rq[109]:/.tg_load_avg
      1438 ±22%    +124.0%       3222 ±13%  TOTAL sched_debug.cfs_rq[18]:/.tg_load_contrib
     21612 ± 4%    +124.6%      48541 ± 1%  TOTAL sched_debug.cfs_rq[39]:/.tg_load_avg
       103 ±15%     -56.4%         45 ±32%  TOTAL sched_debug.cfs_rq[30]:/.avg->runnable_avg_sum
     22746 ± 4%    +122.6%      50626 ± 0%  TOTAL sched_debug.cfs_rq[2]:/.tg_load_avg
     22689 ± 4%    +122.6%      50512 ± 0%  TOTAL sched_debug.cfs_rq[1]:/.tg_load_avg
      1498 ±12%    +118.2%       3271 ±16%  TOTAL sched_debug.cfs_rq[33]:/.tg_load_contrib
     21710 ± 5%    +122.1%      48223 ± 1%  TOTAL sched_debug.cfs_rq[45]:/.tg_load_avg
     21675 ± 5%    +123.9%      48522 ± 1%  TOTAL sched_debug.cfs_rq[68]:/.tg_load_avg
     22702 ± 4%    +123.0%      50638 ± 1%  TOTAL sched_debug.cfs_rq[0]:/.tg_load_avg
     21791 ± 5%    +120.8%      48114 ± 1%  TOTAL sched_debug.cfs_rq[47]:/.tg_load_avg
     21768 ± 5%    +122.1%      48352 ± 1%  TOTAL sched_debug.cfs_rq[57]:/.tg_load_avg
     21611 ± 4%    +123.9%      48393 ± 1%  TOTAL sched_debug.cfs_rq[72]:/.tg_load_avg
     21668 ± 4%    +124.2%      48578 ± 1%  TOTAL sched_debug.cfs_rq[38]:/.tg_load_avg
     21661 ± 5%    +123.5%      48416 ± 1%  TOTAL sched_debug.cfs_rq[71]:/.tg_load_avg
     21653 ± 5%    +123.0%      48291 ± 2%  TOTAL sched_debug.cfs_rq[76]:/.tg_load_avg
     21748 ± 5%    +122.7%      48426 ± 1%  TOTAL sched_debug.cfs_rq[62]:/.tg_load_avg
     21770 ± 5%    +121.2%      48162 ± 1%  TOTAL sched_debug.cfs_rq[46]:/.tg_load_avg
     21688 ± 5%    +122.7%      48295 ± 2%  TOTAL sched_debug.cfs_rq[75]:/.tg_load_avg
     21651 ± 5%    +123.5%      48392 ± 1%  TOTAL sched_debug.cfs_rq[74]:/.tg_load_avg
    101524 ± 6%    +122.3%     225684 ± 4%  TOTAL proc-vmstat.pgalloc_dma32
     21758 ± 5%    +122.5%      48417 ± 1%  TOTAL sched_debug.cfs_rq[63]:/.tg_load_avg
     21721 ± 5%    +123.2%      48488 ± 1%  TOTAL sched_debug.cfs_rq[67]:/.tg_load_avg
      2057 ±10%    +137.4%       4885 ±18%  TOTAL sched_debug.cfs_rq[61]:/.blocked_load_avg
     21704 ± 5%    +123.2%      48439 ± 1%  TOTAL sched_debug.cfs_rq[64]:/.tg_load_avg
     21695 ± 5%    +123.2%      48422 ± 1%  TOTAL sched_debug.cfs_rq[70]:/.tg_load_avg
     21706 ± 5%    +123.1%      48428 ± 1%  TOTAL sched_debug.cfs_rq[69]:/.tg_load_avg
     21837 ± 5%    +121.2%      48294 ± 1%  TOTAL sched_debug.cfs_rq[56]:/.tg_load_avg
     21761 ± 4%    +122.3%      48370 ± 1%  TOTAL sched_debug.cfs_rq[61]:/.tg_load_avg
     21769 ± 4%    +123.2%      48581 ± 1%  TOTAL sched_debug.cfs_rq[37]:/.tg_load_avg
     21704 ± 5%    +123.4%      48479 ± 1%  TOTAL sched_debug.cfs_rq[66]:/.tg_load_avg
     21643 ± 5%    +123.5%      48365 ± 1%  TOTAL sched_debug.cfs_rq[73]:/.tg_load_avg
     21693 ± 5%    +123.5%      48480 ± 1%  TOTAL sched_debug.cfs_rq[65]:/.tg_load_avg
      1498 ±12%    +117.5%       3260 ±16%  TOTAL sched_debug.cfs_rq[33]:/.blocked_load_avg
     21762 ± 4%    +122.0%      48313 ± 1%  TOTAL sched_debug.cfs_rq[58]:/.tg_load_avg
     21873 ± 5%    +120.1%      48132 ± 1%  TOTAL sched_debug.cfs_rq[48]:/.tg_load_avg
     21778 ± 4%    +123.3%      48623 ± 1%  TOTAL sched_debug.cfs_rq[36]:/.tg_load_avg
     21887 ± 3%    +122.3%      48647 ± 1%  TOTAL sched_debug.cfs_rq[35]:/.tg_load_avg
     22106 ± 6%    +123.1%      49330 ± 1%  TOTAL sched_debug.cfs_rq[14]:/.tg_load_avg
     21743 ± 5%    +122.5%      48382 ± 1%  TOTAL sched_debug.cfs_rq[60]:/.tg_load_avg
     21761 ± 5%    +122.2%      48362 ± 1%  TOTAL sched_debug.cfs_rq[59]:/.tg_load_avg
     21933 ± 3%    +121.8%      48653 ± 1%  TOTAL sched_debug.cfs_rq[34]:/.tg_load_avg
     22003 ± 3%    +121.9%      48833 ± 1%  TOTAL sched_debug.cfs_rq[29]:/.tg_load_avg
     21927 ± 5%    +119.2%      48069 ± 1%  TOTAL sched_debug.cfs_rq[49]:/.tg_load_avg
     21970 ± 4%    +123.1%      49009 ± 2%  TOTAL sched_debug.cfs_rq[22]:/.tg_load_avg
     21979 ± 5%    +123.1%      49045 ± 1%  TOTAL sched_debug.cfs_rq[21]:/.tg_load_avg
     22033 ± 5%    +123.1%      49153 ± 1%  TOTAL sched_debug.cfs_rq[19]:/.tg_load_avg
     21979 ± 5%    +118.8%      48084 ± 1%  TOTAL sched_debug.cfs_rq[52]:/.tg_load_avg
     22059 ± 3%    +121.2%      48794 ± 1%  TOTAL sched_debug.cfs_rq[28]:/.tg_load_avg
     21984 ± 4%    +122.9%      48996 ± 2%  TOTAL sched_debug.cfs_rq[23]:/.tg_load_avg
     21965 ± 4%    +119.1%      48128 ± 1%  TOTAL sched_debug.cfs_rq[53]:/.tg_load_avg
     22037 ± 3%    +121.4%      48784 ± 1%  TOTAL sched_debug.cfs_rq[30]:/.tg_load_avg
     22069 ± 3%    +121.1%      48793 ± 1%  TOTAL sched_debug.cfs_rq[31]:/.tg_load_avg
     21957 ± 4%    +119.6%      48208 ± 1%  TOTAL sched_debug.cfs_rq[54]:/.tg_load_avg
     21954 ± 5%    +118.8%      48030 ± 1%  TOTAL sched_debug.cfs_rq[50]:/.tg_load_avg
     22117 ± 6%    +122.8%      49277 ± 1%  TOTAL sched_debug.cfs_rq[15]:/.tg_load_avg
     21978 ± 5%    +118.6%      48046 ± 1%  TOTAL sched_debug.cfs_rq[51]:/.tg_load_avg
     21963 ± 4%    +119.7%      48246 ± 1%  TOTAL sched_debug.cfs_rq[55]:/.tg_load_avg
     22012 ± 3%    +121.3%      48712 ± 1%  TOTAL sched_debug.cfs_rq[32]:/.tg_load_avg
     22037 ± 3%    +121.6%      48838 ± 1%  TOTAL sched_debug.cfs_rq[27]:/.tg_load_avg
     22029 ± 3%    +122.0%      48908 ± 1%  TOTAL sched_debug.cfs_rq[26]:/.tg_load_avg
     21990 ± 4%    +122.3%      48891 ± 1%  TOTAL sched_debug.cfs_rq[25]:/.tg_load_avg
     22010 ± 4%    +122.2%      48914 ± 1%  TOTAL sched_debug.cfs_rq[24]:/.tg_load_avg
     21985 ± 3%    +121.4%      48671 ± 1%  TOTAL sched_debug.cfs_rq[33]:/.tg_load_avg
     22023 ± 5%    +122.8%      49058 ± 2%  TOTAL sched_debug.cfs_rq[20]:/.tg_load_avg
     22054 ± 5%    +123.3%      49240 ± 1%  TOTAL sched_debug.cfs_rq[17]:/.tg_load_avg
     22112 ± 6%    +122.7%      49239 ± 1%  TOTAL sched_debug.cfs_rq[16]:/.tg_load_avg
     22058 ± 5%    +122.8%      49137 ± 1%  TOTAL sched_debug.cfs_rq[18]:/.tg_load_avg
      2237 ± 9%    +124.1%       5013 ±18%  TOTAL sched_debug.cfs_rq[61]:/.tg_load_contrib
     15018 ± 5%    +114.2%      32164 ± 0%  TOTAL sched_debug.cpu#61.ttwu_local
     15138 ± 5%    +113.2%      32273 ± 1%  TOTAL sched_debug.cpu#91.ttwu_local
     15141 ± 6%    +116.6%      32798 ± 4%  TOTAL sched_debug.cpu#106.ttwu_local
  13386019 ± 4%    +113.7%   28610387 ± 2%  TOTAL proc-vmstat.pgalloc_normal
  13484197 ± 4%    +113.8%   28830514 ± 2%  TOTAL proc-vmstat.pgfree
     76806 ± 5%    +111.6%     162514 ± 1%  TOTAL sched_debug.cpu#91.nr_switches
     76808 ± 5%    +111.8%     162667 ± 1%  TOTAL sched_debug.cpu#91.sched_count
  13084681 ± 4%    +113.2%   27900527 ± 2%  TOTAL proc-vmstat.numa_local
  13084697 ± 4%    +113.2%   27900563 ± 2%  TOTAL proc-vmstat.numa_hit
     76596 ± 6%    +112.9%     163099 ± 2%  TOTAL sched_debug.cpu#106.nr_switches
     30771 ± 6%    +111.9%      65201 ± 2%  TOTAL sched_debug.cpu#106.sched_goidle
     30881 ± 5%    +110.9%      65122 ± 2%  TOTAL sched_debug.cpu#91.sched_goidle
  18389572 ± 5%    +111.8%   38945748 ± 1%  TOTAL proc-vmstat.pgfault
   3250107 ± 5%    +111.8%    6885051 ± 1%  TOTAL numa-numastat.node1.local_node
   3250110 ± 5%    +111.8%    6885062 ± 1%  TOTAL numa-numastat.node1.numa_hit
   3262764 ± 5%    +115.3%    7024497 ± 4%  TOTAL numa-numastat.node2.numa_hit
   3262758 ± 5%    +115.3%    7024485 ± 4%  TOTAL numa-numastat.node2.local_node
   3279215 ± 4%    +113.9%    7015147 ± 4%  TOTAL numa-numastat.node0.numa_hit
   3279211 ± 4%    +113.9%    7015136 ± 4%  TOTAL numa-numastat.node0.local_node
     76121 ± 4%    +112.9%     162034 ± 1%  TOTAL sched_debug.cpu#61.nr_switches
     77527 ± 6%    +109.0%     162036 ± 1%  TOTAL sched_debug.cpu#61.sched_count
    243.30 ±33%     -51.6%     117.69 ±29%  TOTAL sched_debug.cfs_rq[92]:/.exec_clock
     30594 ± 4%    +112.1%      64904 ± 2%  TOTAL sched_debug.cpu#61.sched_goidle
   3281833 ± 4%    +109.8%    6886537 ± 1%  TOTAL numa-numastat.node3.local_node
   3281836 ± 4%    +109.8%    6886541 ± 1%  TOTAL numa-numastat.node3.numa_hit
     78218 ± 6%    +109.1%     163583 ± 3%  TOTAL sched_debug.cpu#106.sched_count
   1727502 ± 6%    +107.5%    3583823 ± 4%  TOTAL numa-vmstat.node2.numa_local
   1742994 ± 5%    +103.7%    3550217 ± 1%  TOTAL numa-vmstat.node3.numa_local
   1794367 ± 5%    +101.6%    3617858 ± 1%  TOTAL numa-vmstat.node1.numa_local
   1810000 ± 5%    +102.6%    3666376 ± 3%  TOTAL numa-vmstat.node2.numa_hit
   1825404 ± 5%     +99.0%    3632638 ± 1%  TOTAL numa-vmstat.node3.numa_hit
   1816414 ± 5%    +101.8%    3666109 ± 3%  TOTAL numa-vmstat.node0.numa_local
   1843627 ± 6%    +100.6%    3698703 ± 1%  TOTAL numa-vmstat.node1.numa_hit
      3135 ±12%     -46.7%       1672 ±15%  TOTAL sched_debug.cpu#34.ttwu_local
   1849929 ± 4%     +98.3%    3668167 ± 3%  TOTAL numa-vmstat.node0.numa_hit
      8992 ±13%     -50.9%       4418 ±41%  TOTAL sched_debug.cpu#30.sched_count
    241.28 ±24%     -40.6%     143.43 ±25%  TOTAL sched_debug.cfs_rq[11]:/.exec_clock
     18020 ±39%     -55.2%       8066 ±16%  TOTAL sched_debug.cpu#4.ttwu_count
       319 ±22%     +61.6%        516 ±13%  TOTAL cpuidle.C1E-IVT-4S.usage
      4156 ±15%     -47.6%       2176 ±41%  TOTAL sched_debug.cpu#30.sched_goidle
      8343 ±15%     -47.6%       4375 ±41%  TOTAL sched_debug.cpu#30.nr_switches
     29165 ± 1%     +76.4%      51461 ± 3%  TOTAL sched_debug.cpu#106.ttwu_count
     28980 ± 2%     +73.4%      50247 ± 1%  TOTAL sched_debug.cpu#61.ttwu_count
     29138 ± 1%     +74.5%      50853 ± 1%  TOTAL sched_debug.cpu#91.ttwu_count
     22537 ± 8%     +70.7%      38465 ±19%  TOTAL sched_debug.cpu#47.ttwu_count
      1641 ± 3%     +67.9%       2757 ± 1%  TOTAL proc-vmstat.pgactivate
       131 ±19%     -37.4%         82 ± 5%  TOTAL sched_debug.cpu#106.cpu_load[4]
     13130 ± 2%     +62.4%      21321 ± 8%  TOTAL sched_debug.cpu#47.sched_goidle
      7089 ±13%     +54.9%      10979 ±11%  TOTAL sched_debug.cpu#20.sched_goidle
     26562 ± 2%     +61.5%      42903 ± 7%  TOTAL sched_debug.cpu#47.nr_switches
     14233 ±13%     +54.5%      21991 ±11%  TOTAL sched_debug.cpu#20.nr_switches
        88 ±17%     +54.3%        135 ±25%  TOTAL sched_debug.cpu#107.ttwu_local
      4777 ±12%     +54.7%       7389 ±14%  TOTAL sched_debug.cfs_rq[34]:/.min_vruntime
       119 ±12%     -32.7%         80 ± 8%  TOTAL sched_debug.cpu#61.cpu_load[4]
       149 ±17%     -33.5%         99 ± 9%  TOTAL sched_debug.cpu#106.cpu_load[3]
     11071 ±17%     -26.7%       8120 ±22%  TOTAL sched_debug.cpu#34.ttwu_count
      3831 ± 6%     +42.6%       5463 ± 7%  TOTAL numa-meminfo.node2.KernelStack
      1712 ±11%     -43.0%        975 ±22%  TOTAL sched_debug.cpu#1.ttwu_local
       239 ± 6%     +41.7%        339 ± 7%  TOTAL numa-vmstat.node2.nr_kernel_stack
      3638 ±24%     -32.8%       2443 ±32%  TOTAL sched_debug.cpu#1.ttwu_count
       135 ± 7%     -37.0%         85 ±11%  TOTAL sched_debug.cpu#91.cpu_load[4]
      5131 ±18%     -21.3%       4038 ± 5%  TOTAL meminfo.AnonHugePages
       227 ±12%     -28.6%        162 ±18%  TOTAL sched_debug.cpu#91.cpu_load[0]
     66199 ± 6%     +49.7%      99076 ± 1%  TOTAL sched_debug.cpu#106.nr_load_updates
     31880 ± 6%     +41.2%      45012 ±10%  TOTAL sched_debug.cpu#47.sched_count
     13581 ± 3%     +47.7%      20066 ± 5%  TOTAL sched_debug.cpu#32.sched_goidle
     29309 ±12%     +41.2%      41372 ± 9%  TOTAL sched_debug.cpu#32.sched_count
     69667 ± 4%     +42.5%      99307 ± 1%  TOTAL sched_debug.cpu#91.nr_load_updates
       160 ±15%     -26.9%        117 ±18%  TOTAL sched_debug.cfs_rq[61]:/.load
     27436 ± 3%     +47.3%      40401 ± 5%  TOTAL sched_debug.cpu#32.nr_switches
     70549 ± 4%     +41.8%     100061 ± 1%  TOTAL sched_debug.cpu#61.nr_load_updates
     13693 ± 5%     +48.4%      20325 ± 6%  TOTAL sched_debug.cpu#17.sched_goidle
      3973 ± 6%     +41.4%       5619 ± 6%  TOTAL numa-meminfo.node3.KernelStack
     27719 ± 5%     +47.9%      40984 ± 6%  TOTAL sched_debug.cpu#17.nr_switches
       248 ± 6%     +40.7%        349 ± 6%  TOTAL numa-vmstat.node3.nr_kernel_stack
      6508 ± 1%     +49.1%       9705 ±22%  TOTAL sched_debug.cpu#35.sched_goidle
       138 ±14%     -27.4%        100 ±10%  TOTAL sched_debug.cpu#61.cpu_load[3]
     13073 ± 2%     +48.7%      19438 ±22%  TOTAL sched_debug.cpu#35.nr_switches
       666 ±14%     +50.7%       1004 ±17%  TOTAL cpuidle.C3-IVT-4S.usage
        80 ±33%     -45.9%         43 ±34%  TOTAL sched_debug.cfs_rq[39]:/.avg->runnable_avg_sum
     19457 ± 7%     +31.6%      25610 ± 7%  TOTAL sched_debug.cpu#47.nr_load_updates
     21711 ±13%     +45.4%      31570 ±19%  TOTAL sched_debug.cpu#17.ttwu_count
     13418 ± 3%     +45.3%      19492 ±22%  TOTAL sched_debug.cpu#35.sched_count
     22622 ± 5%     +37.5%      31103 ±15%  TOTAL sched_debug.cpu#32.ttwu_count
        21 ± 9%     -19.6%         17 ±12%  TOTAL sched_debug.cpu#99.ttwu_local
       191 ±12%     -25.1%        143 ±10%  TOTAL sched_debug.cpu#91.cpu_load[1]
       154 ± 9%     -31.6%        105 ±11%  TOTAL sched_debug.cpu#91.cpu_load[3]
       160 ±15%     -27.2%        116 ±14%  TOTAL sched_debug.cpu#106.cpu_load[2]
     15464 ±15%     +43.3%      22163 ±11%  TOTAL sched_debug.cpu#20.sched_count
       176 ±14%     -24.4%        133 ±23%  TOTAL sched_debug.cpu#106.cpu_load[1]
       169 ±10%     -25.2%        126 ±10%  TOTAL sched_debug.cpu#91.cpu_load[2]
        20 ±10%     -15.5%         17 ± 8%  TOTAL sched_debug.cpu#74.ttwu_local
     87151 ±17%     +34.6%     117307 ± 3%  TOTAL sched_debug.cfs_rq[106]:/.spread0
     23131 ± 9%     -20.8%      18314 ±12%  TOTAL sched_debug.cpu#33.ttwu_count
      9097 ±11%     -23.5%       6955 ± 8%  TOTAL sched_debug.cfs_rq[0]:/.exec_clock
      1485 ± 2%     +25.6%       1866 ± 7%  TOTAL proc-vmstat.nr_kernel_stack
      1788 ± 2%     -20.6%       1419 ± 0%  TOTAL sched_debug.cfs_rq[106]:/.tg->runnable_avg
      1810 ± 2%     -20.6%       1436 ± 0%  TOTAL sched_debug.cfs_rq[119]:/.tg->runnable_avg
      1809 ± 2%     -20.7%       1435 ± 0%  TOTAL sched_debug.cfs_rq[118]:/.tg->runnable_avg
      1784 ± 2%     -20.6%       1417 ± 0%  TOTAL sched_debug.cfs_rq[105]:/.tg->runnable_avg
      1798 ± 2%     -20.7%       1426 ± 0%  TOTAL sched_debug.cfs_rq[111]:/.tg->runnable_avg
      1803 ± 2%     -20.5%       1433 ± 0%  TOTAL sched_debug.cfs_rq[115]:/.tg->runnable_avg
      1801 ± 2%     -20.5%       1431 ± 0%  TOTAL sched_debug.cfs_rq[114]:/.tg->runnable_avg
      1790 ± 2%     -20.6%       1421 ± 0%  TOTAL sched_debug.cfs_rq[107]:/.tg->runnable_avg
      1799 ± 2%     -20.6%       1428 ± 0%  TOTAL sched_debug.cfs_rq[112]:/.tg->runnable_avg
      1792 ± 2%     -20.6%       1423 ± 0%  TOTAL sched_debug.cfs_rq[108]:/.tg->runnable_avg
      1782 ± 2%     -20.5%       1415 ± 0%  TOTAL sched_debug.cfs_rq[104]:/.tg->runnable_avg
      1800 ± 2%     -20.6%       1430 ± 0%  TOTAL sched_debug.cfs_rq[113]:/.tg->runnable_avg
      1805 ± 2%     -20.6%       1434 ± 0%  TOTAL sched_debug.cfs_rq[116]:/.tg->runnable_avg
      1806 ± 2%     -20.6%       1435 ± 0%  TOTAL sched_debug.cfs_rq[117]:/.tg->runnable_avg
      1795 ± 2%     -20.7%       1424 ± 0%  TOTAL sched_debug.cfs_rq[109]:/.tg->runnable_avg
     95310 ± 4%     +23.7%     117875 ± 2%  TOTAL sched_debug.cfs_rq[91]:/.spread0
      1796 ± 2%     -20.7%       1425 ± 0%  TOTAL sched_debug.cfs_rq[110]:/.tg->runnable_avg
      1778 ± 2%     -20.5%       1414 ± 0%  TOTAL sched_debug.cfs_rq[103]:/.tg->runnable_avg
      1771 ± 2%     -20.3%       1411 ± 0%  TOTAL sched_debug.cfs_rq[100]:/.tg->runnable_avg
      1772 ± 2%     -20.3%       1413 ± 0%  TOTAL sched_debug.cfs_rq[101]:/.tg->runnable_avg
      1768 ± 2%     -20.3%       1410 ± 0%  TOTAL sched_debug.cfs_rq[99]:/.tg->runnable_avg
      1774 ± 2%     -20.3%       1413 ± 0%  TOTAL sched_debug.cfs_rq[102]:/.tg->runnable_avg
     97534 ± 4%     +21.8%     118768 ± 5%  TOTAL sched_debug.cfs_rq[61]:/.spread0
      1766 ± 2%     -20.2%       1408 ± 0%  TOTAL sched_debug.cfs_rq[98]:/.tg->runnable_avg
      1762 ± 2%     -20.1%       1407 ± 0%  TOTAL sched_debug.cfs_rq[97]:/.tg->runnable_avg
      1760 ± 2%     -20.1%       1405 ± 0%  TOTAL sched_debug.cfs_rq[96]:/.tg->runnable_avg
      1756 ± 2%     -20.0%       1405 ± 0%  TOTAL sched_debug.cfs_rq[95]:/.tg->runnable_avg
      1747 ± 2%     -19.8%       1400 ± 0%  TOTAL sched_debug.cfs_rq[92]:/.tg->runnable_avg
      1753 ± 2%     -19.9%       1404 ± 0%  TOTAL sched_debug.cfs_rq[94]:/.tg->runnable_avg
      1751 ± 2%     -19.9%       1402 ± 0%  TOTAL sched_debug.cfs_rq[93]:/.tg->runnable_avg
      1743 ± 2%     -19.8%       1398 ± 0%  TOTAL sched_debug.cfs_rq[91]:/.tg->runnable_avg
     23871 ± 2%     +24.3%      29667 ± 8%  TOTAL meminfo.KernelStack
      1739 ± 2%     -19.7%       1397 ± 0%  TOTAL sched_debug.cfs_rq[90]:/.tg->runnable_avg
      1734 ± 2%     -19.6%       1395 ± 0%  TOTAL sched_debug.cfs_rq[89]:/.tg->runnable_avg
      1729 ± 2%     -19.4%       1394 ± 0%  TOTAL sched_debug.cfs_rq[88]:/.tg->runnable_avg
      1725 ± 2%     -19.3%       1392 ± 0%  TOTAL sched_debug.cfs_rq[87]:/.tg->runnable_avg
      1724 ± 2%     -19.3%       1390 ± 0%  TOTAL sched_debug.cfs_rq[86]:/.tg->runnable_avg
      1721 ± 2%     -19.3%       1389 ± 0%  TOTAL sched_debug.cfs_rq[85]:/.tg->runnable_avg
      1718 ± 2%     -19.2%       1388 ± 0%  TOTAL sched_debug.cfs_rq[84]:/.tg->runnable_avg
    102757 ±13%     +28.2%     131768 ± 3%  TOTAL sched_debug.cfs_rq[106]:/.min_vruntime
      1699 ± 2%     -19.0%       1376 ± 0%  TOTAL sched_debug.cfs_rq[77]:/.tg->runnable_avg
      1701 ± 2%     -19.0%       1378 ± 0%  TOTAL sched_debug.cfs_rq[78]:/.tg->runnable_avg
      1692 ± 2%     -18.9%       1373 ± 0%  TOTAL sched_debug.cfs_rq[75]:/.tg->runnable_avg
      1695 ± 2%     -19.0%       1373 ± 0%  TOTAL sched_debug.cfs_rq[76]:/.tg->runnable_avg
      1715 ± 2%     -19.1%       1387 ± 0%  TOTAL sched_debug.cfs_rq[83]:/.tg->runnable_avg
     21038 ± 5%     +20.1%      25263 ± 5%  TOTAL sched_debug.cpu#17.nr_load_updates
      1683 ± 2%     -18.8%       1367 ± 0%  TOTAL sched_debug.cfs_rq[71]:/.tg->runnable_avg
      1709 ± 2%     -18.9%       1385 ± 0%  TOTAL sched_debug.cfs_rq[82]:/.tg->runnable_avg
      1701 ± 2%     -18.9%       1379 ± 0%  TOTAL sched_debug.cfs_rq[79]:/.tg->runnable_avg
      1686 ± 2%     -18.8%       1369 ± 0%  TOTAL sched_debug.cfs_rq[73]:/.tg->runnable_avg
      1681 ± 2%     -18.8%       1365 ± 0%  TOTAL sched_debug.cfs_rq[70]:/.tg->runnable_avg
      1705 ± 2%     -18.9%       1382 ± 0%  TOTAL sched_debug.cfs_rq[80]:/.tg->runnable_avg
      1683 ± 2%     -18.7%       1368 ± 0%  TOTAL sched_debug.cfs_rq[72]:/.tg->runnable_avg
      1672 ± 2%     -18.5%       1362 ± 0%  TOTAL sched_debug.cfs_rq[67]:/.tg->runnable_avg
      1688 ± 2%     -18.8%       1371 ± 0%  TOTAL sched_debug.cfs_rq[74]:/.tg->runnable_avg
      1663 ± 2%     -18.5%       1356 ± 0%  TOTAL sched_debug.cfs_rq[63]:/.tg->runnable_avg
      1679 ± 2%     -18.7%       1364 ± 0%  TOTAL sched_debug.cfs_rq[69]:/.tg->runnable_avg
      1670 ± 2%     -18.5%       1362 ± 0%  TOTAL sched_debug.cfs_rq[66]:/.tg->runnable_avg
      1675 ± 2%     -18.6%       1363 ± 0%  TOTAL sched_debug.cfs_rq[68]:/.tg->runnable_avg
      1665 ± 2%     -18.5%       1357 ± 0%  TOTAL sched_debug.cfs_rq[64]:/.tg->runnable_avg
      1707 ± 2%     -18.9%       1384 ± 0%  TOTAL sched_debug.cfs_rq[81]:/.tg->runnable_avg
      1667 ± 2%     -18.5%       1359 ± 0%  TOTAL sched_debug.cfs_rq[65]:/.tg->runnable_avg
       152 ±14%     -18.9%        123 ±12%  TOTAL sched_debug.cpu#61.cpu_load[2]
      1658 ± 2%     -18.3%       1355 ± 0%  TOTAL sched_debug.cfs_rq[62]:/.tg->runnable_avg
      1652 ± 2%     -18.1%       1353 ± 0%  TOTAL sched_debug.cfs_rq[61]:/.tg->runnable_avg
      1650 ± 2%     -18.0%       1352 ± 0%  TOTAL sched_debug.cfs_rq[60]:/.tg->runnable_avg
      1643 ± 2%     -17.9%       1348 ± 0%  TOTAL sched_debug.cfs_rq[57]:/.tg->runnable_avg
    113140 ± 4%     +17.8%     133227 ± 5%  TOTAL sched_debug.cfs_rq[61]:/.min_vruntime
      1648 ± 2%     -18.0%       1351 ± 0%  TOTAL sched_debug.cfs_rq[59]:/.tg->runnable_avg
    110916 ± 4%     +19.3%     132335 ± 2%  TOTAL sched_debug.cfs_rq[91]:/.min_vruntime
      1625 ± 1%     -17.2%       1346 ± 0%  TOTAL sched_debug.cfs_rq[55]:/.tg->runnable_avg
      1638 ± 2%     -17.8%       1347 ± 0%  TOTAL sched_debug.cfs_rq[56]:/.tg->runnable_avg
      1646 ± 2%     -17.9%       1350 ± 0%  TOTAL sched_debug.cfs_rq[58]:/.tg->runnable_avg
      1615 ± 1%     -17.0%       1340 ± 0%  TOTAL sched_debug.cfs_rq[51]:/.tg->runnable_avg
      1616 ± 1%     -17.0%       1341 ± 0%  TOTAL sched_debug.cfs_rq[52]:/.tg->runnable_avg
      1620 ± 1%     -17.1%       1343 ± 0%  TOTAL sched_debug.cfs_rq[53]:/.tg->runnable_avg
      1611 ± 1%     -17.0%       1337 ± 0%  TOTAL sched_debug.cfs_rq[50]:/.tg->runnable_avg
      1622 ± 1%     -17.1%       1345 ± 0%  TOTAL sched_debug.cfs_rq[54]:/.tg->runnable_avg
      1605 ± 1%     -16.8%       1335 ± 0%  TOTAL sched_debug.cfs_rq[49]:/.tg->runnable_avg
      1601 ± 1%     -16.9%       1331 ± 0%  TOTAL sched_debug.cfs_rq[48]:/.tg->runnable_avg
     14782 ± 5%     -18.7%      12017 ± 7%  TOTAL sched_debug.cfs_rq[91]:/.avg->runnable_avg_sum
      1595 ± 1%     -16.8%       1327 ± 0%  TOTAL sched_debug.cfs_rq[47]:/.tg->runnable_avg
       321 ± 5%     -18.9%        260 ± 7%  TOTAL sched_debug.cfs_rq[91]:/.tg_runnable_contrib
     17810 ±34%     +39.1%      24766 ± 4%  TOTAL sched_debug.cpu#2.nr_load_updates
      1590 ± 1%     -16.6%       1326 ± 0%  TOTAL sched_debug.cfs_rq[46]:/.tg->runnable_avg
      1587 ± 1%     -16.5%       1324 ± 0%  TOTAL sched_debug.cfs_rq[45]:/.tg->runnable_avg
      1581 ± 1%     -16.3%       1323 ± 0%  TOTAL sched_debug.cfs_rq[44]:/.tg->runnable_avg
     20364 ± 6%     +21.0%      24646 ± 5%  TOTAL sched_debug.cpu#32.nr_load_updates
     23451 ±11%     +16.0%      27201 ±10%  TOTAL sched_debug.cpu#18.nr_load_updates
      1576 ± 1%     -16.1%       1322 ± 0%  TOTAL sched_debug.cfs_rq[43]:/.tg->runnable_avg
      4393 ± 3%     +17.0%       5138 ± 2%  TOTAL slabinfo.signal_cache.num_objs
      1573 ± 1%     -16.0%       1321 ± 0%  TOTAL sched_debug.cfs_rq[42]:/.tg->runnable_avg
      1568 ± 1%     -15.9%       1319 ± 0%  TOTAL sched_debug.cfs_rq[41]:/.tg->runnable_avg
      1564 ± 1%     -15.7%       1318 ± 0%  TOTAL sched_debug.cfs_rq[40]:/.tg->runnable_avg
       296 ± 4%     -14.6%        253 ± 4%  TOTAL sched_debug.cfs_rq[61]:/.tg_runnable_contrib
      1560 ± 1%     -15.5%       1318 ± 0%  TOTAL sched_debug.cfs_rq[39]:/.tg->runnable_avg
      1554 ± 1%     -15.3%       1317 ± 0%  TOTAL sched_debug.cfs_rq[38]:/.tg->runnable_avg
     13680 ± 3%     -14.7%      11667 ± 4%  TOTAL sched_debug.cfs_rq[61]:/.avg->runnable_avg_sum
      1544 ± 1%     -14.8%       1315 ± 0%  TOTAL sched_debug.cfs_rq[37]:/.tg->runnable_avg
      3309 ± 5%     -17.4%       2734 ± 2%  TOTAL sched_debug.cfs_rq[32]:/.exec_clock
      1534 ± 1%     -14.6%       1310 ± 0%  TOTAL sched_debug.cfs_rq[34]:/.tg->runnable_avg
      1537 ± 1%     -14.6%       1312 ± 0%  TOTAL sched_debug.cfs_rq[35]:/.tg->runnable_avg
      1540 ± 1%     -14.7%       1314 ± 0%  TOTAL sched_debug.cfs_rq[36]:/.tg->runnable_avg
       553 ±43%     +59.0%        879 ± 3%  TOTAL numa-vmstat.node0.nr_kernel_stack
      1530 ± 1%     -14.5%       1308 ± 0%  TOTAL sched_debug.cfs_rq[33]:/.tg->runnable_avg
      1523 ± 1%     -14.3%       1306 ± 0%  TOTAL sched_debug.cfs_rq[32]:/.tg->runnable_avg
      8851 ±43%     +59.3%      14097 ± 3%  TOTAL numa-meminfo.node0.KernelStack
      1519 ± 1%     -14.1%       1306 ± 0%  TOTAL sched_debug.cfs_rq[31]:/.tg->runnable_avg
      1516 ± 1%     -14.0%       1304 ± 0%  TOTAL sched_debug.cfs_rq[30]:/.tg->runnable_avg
      1510 ± 1%     -14.0%       1300 ± 0%  TOTAL sched_debug.cfs_rq[28]:/.tg->runnable_avg
      1513 ± 1%     -14.0%       1302 ± 0%  TOTAL sched_debug.cfs_rq[29]:/.tg->runnable_avg
      1507 ± 1%     -13.9%       1297 ± 0%  TOTAL sched_debug.cfs_rq[27]:/.tg->runnable_avg
      1504 ± 1%     -13.8%       1296 ± 0%  TOTAL sched_debug.cfs_rq[26]:/.tg->runnable_avg
      1496 ± 1%     -13.6%       1293 ± 0%  TOTAL sched_debug.cfs_rq[24]:/.tg->runnable_avg
      1492 ± 1%     -13.5%       1292 ± 0%  TOTAL sched_debug.cfs_rq[23]:/.tg->runnable_avg
      1499 ± 1%     -13.6%       1295 ± 0%  TOTAL sched_debug.cfs_rq[25]:/.tg->runnable_avg
    795010 ± 3%     -10.6%     710653 ± 6%  TOTAL sched_debug.cpu#32.avg_idle
      1489 ± 1%     -13.3%       1291 ± 0%  TOTAL sched_debug.cfs_rq[22]:/.tg->runnable_avg
      1467 ± 1%     -12.9%       1278 ± 0%  TOTAL sched_debug.cfs_rq[17]:/.tg->runnable_avg
      1485 ± 1%     -13.2%       1290 ± 0%  TOTAL sched_debug.cfs_rq[21]:/.tg->runnable_avg
      1463 ± 1%     -12.8%       1276 ± 0%  TOTAL sched_debug.cfs_rq[16]:/.tg->runnable_avg
      1027 ± 6%     +16.2%       1194 ± 4%  TOTAL slabinfo.kmalloc-192.active_slabs
      1027 ± 6%     +16.2%       1194 ± 4%  TOTAL slabinfo.kmalloc-192.num_slabs
     43031 ± 6%     +16.3%      50041 ± 4%  TOTAL slabinfo.kmalloc-192.active_objs
     43170 ± 6%     +16.2%      50161 ± 4%  TOTAL slabinfo.kmalloc-192.num_objs
      1472 ± 1%     -12.9%       1282 ± 0%  TOTAL sched_debug.cfs_rq[18]:/.tg->runnable_avg
      1479 ± 1%     -13.0%       1287 ± 0%  TOTAL sched_debug.cfs_rq[20]:/.tg->runnable_avg
      1456 ± 1%     -12.5%       1273 ± 0%  TOTAL sched_debug.cfs_rq[15]:/.tg->runnable_avg
      1452 ± 1%     -12.3%       1273 ± 0%  TOTAL sched_debug.cfs_rq[14]:/.tg->runnable_avg
       862 ± 8%     -12.9%        750 ± 5%  TOTAL slabinfo.RAW.num_objs
       862 ± 8%     -12.9%        750 ± 5%  TOTAL slabinfo.RAW.active_objs
      1475 ± 1%     -12.9%       1284 ± 0%  TOTAL sched_debug.cfs_rq[19]:/.tg->runnable_avg
      4393 ± 3%     +14.5%       5028 ± 2%  TOTAL slabinfo.signal_cache.active_objs
      1446 ± 1%     -12.1%       1272 ± 0%  TOTAL sched_debug.cfs_rq[12]:/.tg->runnable_avg
      1448 ± 1%     -12.1%       1273 ± 0%  TOTAL sched_debug.cfs_rq[13]:/.tg->runnable_avg
      1442 ± 1%     -11.9%       1271 ± 0%  TOTAL sched_debug.cfs_rq[11]:/.tg->runnable_avg
      1439 ± 1%     -11.7%       1271 ± 0%  TOTAL sched_debug.cfs_rq[10]:/.tg->runnable_avg
      1437 ± 1%     -11.5%       1271 ± 0%  TOTAL sched_debug.cfs_rq[9]:/.tg->runnable_avg
      1431 ± 1%     -11.2%       1270 ± 0%  TOTAL sched_debug.cfs_rq[8]:/.tg->runnable_avg
      1428 ± 1%     -11.1%       1269 ± 0%  TOTAL sched_debug.cfs_rq[7]:/.tg->runnable_avg
      1423 ± 1%     -10.8%       1270 ± 0%  TOTAL sched_debug.cfs_rq[6]:/.tg->runnable_avg
      1421 ± 1%     -10.6%       1270 ± 0%  TOTAL sched_debug.cfs_rq[5]:/.tg->runnable_avg
      1418 ± 1%     -10.5%       1269 ± 0%  TOTAL sched_debug.cfs_rq[4]:/.tg->runnable_avg
      1417 ± 1%     -10.5%       1268 ± 0%  TOTAL sched_debug.cfs_rq[3]:/.tg->runnable_avg
      5041 ± 4%     +12.8%       5687 ± 1%  TOTAL slabinfo.task_xstate.active_objs
      5041 ± 4%     +12.8%       5687 ± 1%  TOTAL slabinfo.task_xstate.num_objs
        20 ±18%     -18.6%         16 ± 2%  TOTAL sched_debug.cpu#104.ttwu_local
     83828 ± 1%      +9.4%      91675 ± 3%  TOTAL slabinfo.kmalloc-64.active_objs
      1406 ± 1%     -10.1%       1264 ± 0%  TOTAL sched_debug.cfs_rq[2]:/.tg->runnable_avg
      1404 ± 1%     -10.1%       1262 ± 0%  TOTAL sched_debug.cfs_rq[1]:/.tg->runnable_avg
    109592 ± 4%      +6.3%     116546 ± 2%  TOTAL numa-meminfo.node1.FilePages
     27397 ± 4%      +6.3%      29136 ± 2%  TOTAL numa-vmstat.node1.nr_file_pages
        36 ± 2%      +8.3%         39 ± 2%  TOTAL turbostat.CTMP
      1382 ± 1%      -9.2%       1255 ± 0%  TOTAL sched_debug.cfs_rq[0]:/.tg->runnable_avg
     52240 ± 5%      +8.9%      56888 ± 4%  TOTAL numa-meminfo.node0.Slab
     31564 ± 7%     +14.9%      36254 ± 5%  TOTAL numa-meminfo.node1.Active
      1331 ± 0%      +8.1%       1439 ± 3%  TOTAL slabinfo.kmalloc-64.active_slabs
      1331 ± 0%      +8.1%       1439 ± 3%  TOTAL slabinfo.kmalloc-64.num_slabs
     85255 ± 0%      +8.1%      92172 ± 3%  TOTAL slabinfo.kmalloc-64.num_objs
    217201 ± 5%    +125.5%     489860 ± 0%  TOTAL time.voluntary_context_switches
  17206167 ± 5%    +118.8%   37639010 ± 1%  TOTAL time.minor_page_faults
    115930 ± 5%    +116.5%     251005 ± 1%  TOTAL time.involuntary_context_switches
      0.00 ± 9%    +121.4%       0.00 ±10%  TOTAL energy.energy-cores
      0.00 ± 1%     +63.6%       0.00 ± 0%  TOTAL energy.energy-ram
      0.00 ± 3%     +51.6%       0.00 ± 5%  TOTAL energy.energy-pkg
      7352 ± 1%     +39.9%      10285 ± 0%  TOTAL vmstat.system.cs
     89.70 ± 0%     -14.0%      77.11 ± 1%  TOTAL time.user_time
      1.06 ± 0%     -12.9%       0.92 ± 0%  TOTAL turbostat.%c0
       214 ± 0%      +5.9%        227 ± 0%  TOTAL time.system_time


                                 time.user_time

  95 ++---------------------------------------------------------------------+
     |                *..                                                   |
     |               :   *..  .*.                                           |
  90 *+.*..*..*.*..  :      *.   *..*..*..*..*                              |
     |              :                                                       |
     |             *                                                        |
  85 ++                                                                     |
     |                                                                      |
  80 ++                                                                     |
     |          O                                                     O     |
     |  O  O  O    O                                             O          O
  75 O+               O     O    O     O                   O  O    O     O  |
     |                   O     O    O     O  O  O O  O  O                   |
     |                                                                      |
  70 ++---------------------------------------------------------------------+


                                 time.system_time

  235 ++--------------------------------------------------------------------+
      |                                                                     |
      |                        O                                            |
  230 ++                          O  O  O O  O  O  O  O O                   |
      O                   O O                              O  O     O    O  |
      |  O  O O     O  O                                         O    O     O
  225 ++         O                                                          |
      |                                                                     |
  220 ++                                                                    |
      |                                                                     |
      |             *                                                       |
  215 ++           + :           .*..                                       |
      |           +   :   *.*..*.    *..*.*..*                              |
      *..*..*.*..*    : ..                                                  |
  210 ++---------------*----------------------------------------------------+


                           time.voluntary_context_switches

  500000 ++------------------------------------------------------O--O--O----O
         |                                                               O  |
  450000 ++                                                    O            |
         |                    O                             O               |
  400000 ++                 O    O  O O  O  O  O O  O  O O                  |
         |  O       O                                                       |
  350000 O+   O  O    O  O                                                  |
         |                                                                  |
  300000 ++                                                                 |
         |                                                                  |
  250000 ++                                                                 |
         *..*.*..*..*.          .*..*.     .*..                             |
  200000 ++           *..*..*.*.      *..*.    *                            |
         |                                                                  |
  150000 ++-----------------------------------------------------------------+


                                  energy.energy-cores

  2.5e-08 ++----------------------------------------------------------------+
          |                                                            O    |
          |                                                                 |
    2e-08 ++                                                                |
          |                                                      O  O    O  O
          |                                                    O            |
  1.5e-08 ++                   O                                            |
          |  O       O                    O                 O               |
    1e-08 O+   O  O    O  O O     O O  O    O  O  O O  O  O                 |
          |            *..*.*..*..*.*..*..*.*..*                            |
          |            :                                                    |
    5e-09 ++          :                                                     |
          |           :                                                     |
          |          :                                                      |
        0 *+-*-*--*--*------------------------------------------------------+


                                   energy.energy-pkg

    4e-08 ++----------------------------------------------------------------+
          |                                                            O    |
  3.5e-08 ++                                                                O
          |                                                    O O  O    O  |
    3e-08 ++                   O                                            |
  2.5e-08 ++ O       O      O     O O  O  O O  O  O O  O  O O               |
          O    O  O    O..O.*..*..*.*..*..*.*..*                            |
    2e-08 ++           :                                                    |
          |            :                                                    |
  1.5e-08 ++          :                                                     |
    1e-08 ++          :                                                     |
          |           :                                                     |
    5e-09 ++          :                                                     |
          |          :                                                      |
        0 *+-*-*--*--*------------------------------------------------------+


                           aim9.shell_rtns_3.ops_per_sec

  300 ++--------------------------------------------------------------------+
  280 ++                                                              O     |
      |                                                          O  O    O  O
  260 ++                                                      O             |
  240 ++                    O                              O                |
      |                   O    O  O  O  O O  O  O  O  O O                   |
  220 ++ O       O                                                          |
  200 O+    O O     O  O                                                    |
  180 ++                                                                    |
      |                                                                     |
  160 ++                                                                    |
  140 ++                                                                    |
      |                  .*.*..*..*..*..*.*..*                              |
  120 *+.*..*.*..*..*..*.                                                   |
  100 ++--------------------------------------------------------------------+


                                   turbostat.%c0

  1.12 ++-------------------------------------------------------------------+
   1.1 ++                 *..*                                              |
       |                 +    +                                             |
  1.08 ++   .*.        .*      +                                            |
  1.06 *+.*.   *..*..*.         *..*.*..*..*..*                             |
  1.04 ++                                                                   |
  1.02 ++                                                                   |
       |                                                                    |
     1 ++                                                                   |
  0.98 O+    O       O  O O          O                                      |
  0.96 ++ O    O  O          O  O  O    O  O  O O  O  O  O                  |
  0.94 ++                                                  O  O             |
       |                                                              O     |
  0.92 ++                                                        O  O    O  O
   0.9 ++-------------------------------------------------------------------+


                                  turbostat.Pkg_W

  170 ++--------------------------------------------------------------------+
      |                                                               O     |
  160 ++                                                                    |
  150 ++                                                                    |
      |                                                                     O
  140 ++                                                         O  O    O  |
      |                                                                     |
  130 ++                                                      O             |
      |                     O                                               |
  120 ++                                                                    |
  110 ++ O       O                      O                  O                |
      O     O O     O  O  O    O  O  O    O  O  O  O  O O                   |
  100 ++                                                                    |
      *..*..*.*..*..*..*..*.*..*..*..*..*.*..*                              |
   90 ++--------------------------------------------------------------------+


                                  turbostat.Cor_W

  100 ++--------------------------------------------------------------O-----+
      |                                                                     |
   90 ++                                                                    |
      |                                                                     |
   80 ++                                                                    O
      |                                                          O  O    O  |
   70 ++                                                      O             |
      |                     O                                               |
   60 ++                                                                    |
      |  O                                                                  |
   50 ++         O        O    O  O  O  O O  O  O  O  O O  O                |
      O     O O     O  O                                                    |
   40 ++                                                                    |
      *..*..*.*..*..*..*..*.*..*..*..*..*.*..*                              |
   30 ++--------------------------------------------------------------------+

	[*] bisect-good sample
	[O] bisect-bad  sample

To reproduce:

	apt-get install ruby ruby-oj
	git clone git://git.kernel.org/pub/scm/linux/kernel/git/wfg/lkp-tests.git
	cd lkp-tests
	bin/setup-local job.yaml # the job file attached in this email
	bin/run-local   job.yaml

Disclaimer:
Results have been estimated based on internal Intel analysis and are provided
for informational purposes only. Any difference in system hardware or software
design or configuration may affect actual performance.

Thanks,
Fengguang

View attachment "job.yaml" of type "text/plain" (1286 bytes)

_______________________________________________
LKP mailing list
LKP@...ux.intel.com

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ