lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [day] [month] [year] [list]
Message-ID: <1428628056.6726.62.camel@intel.com>
Date:	Fri, 10 Apr 2015 09:07:36 +0800
From:	Huang Ying <ying.huang@...el.com>
To:	Mel Gorman <mgorman@...e.de>
Cc:	Linus Torvalds <torvalds@...ux-foundation.org>,
	LKML <linux-kernel@...r.kernel.org>, LKP ML <lkp@...org>
Subject: [LKP] [mm] bea66fbd11a: -13.5% aim7.time.minor_page_faults, -41.1%
 time.system_time

FYI, we noticed the below changes on

git://git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux.git master
commit bea66fbd11af1ca98ae26855eea41eda8582923e ("mm: numa: group related processes based on VMA flags instead of page table flags")


testbox/testcase/testparams: brickland3/aim7/performance-3000-sieve

98cf21c61a7f5419  bea66fbd11af1ca98ae26855ee  
----------------  --------------------------  
         %stddev     %change         %stddev
             \          |                \  
 1.155e+08 ±  1%     -13.5%   99915100 ±  1%  aim7.time.minor_page_faults
    904971 ±  1%     -64.8%     318137 ± 10%  aim7.time.voluntary_context_switches
    120.22 ±  1%     -41.1%      70.87 ±  3%  aim7.time.system_time
   6122485 ±  0%      -1.6%    6025423 ±  0%  aim7.time.involuntary_context_switches
      1464 ± 14%     -66.0%        498 ± 37%  cpuidle.C1-IVT-4S.usage
       470 ±  1%     -65.0%        164 ± 26%  vmstat.procs.b
    125221 ±  0%      -2.2%     122500 ±  0%  vmstat.system.in
     32780 ±  0%     -10.4%      29374 ±  1%  vmstat.system.cs
 1.155e+08 ±  1%     -13.5%   99915100 ±  1%  time.minor_page_faults
    120.22 ±  1%     -41.1%      70.87 ±  3%  time.system_time
    904971 ±  1%     -64.8%     318137 ± 10%  time.voluntary_context_switches
        84 ± 43%     -67.0%         28 ± 23%  numa-numastat.node0.other_node
  56733780 ±  1%     -12.9%   49397913 ±  1%  proc-vmstat.numa_hint_faults
  55660319 ±  1%     -13.2%   48314771 ±  1%  proc-vmstat.numa_hint_faults_local
   3054195 ±  0%     -13.5%    2642998 ±  0%  proc-vmstat.numa_hit
   3053893 ±  0%     -13.5%    2642870 ±  0%  proc-vmstat.numa_local
   1010298 ±  1%     -41.6%     590406 ±  1%  proc-vmstat.numa_pages_migrated
  57648883 ±  1%     -13.5%   49892114 ±  1%  proc-vmstat.numa_pte_updates
   3076975 ±  0%     -13.6%    2659661 ±  0%  proc-vmstat.pgalloc_normal
  1.16e+08 ±  1%     -13.4%  1.005e+08 ±  1%  proc-vmstat.pgfault
   3003186 ±  0%     -12.3%    2633789 ±  0%  proc-vmstat.pgfree
   1010298 ±  1%     -41.6%     590406 ±  1%  proc-vmstat.pgmigrate_success
         6 ± 12%     +33.3%          9 ±  7%  sched_debug.cfs_rq[104]:/.load
         5 ± 28%     +66.7%          8 ± 21%  sched_debug.cfs_rq[106]:/.load
         5 ± 28%     +61.9%          8 ± 17%  sched_debug.cfs_rq[106]:/.runnable_load_avg
      5069 ±  4%      +6.5%       5399 ±  5%  sched_debug.cfs_rq[109]:/.tg_load_avg
         6 ±  6%     +20.0%          7 ±  6%  sched_debug.cfs_rq[110]:/.load
      5060 ±  6%     +10.2%       5574 ±  5%  sched_debug.cfs_rq[115]:/.tg_load_avg
         6 ±  0%     +20.8%          7 ±  5%  sched_debug.cfs_rq[115]:/.load
      4974 ±  4%     +11.8%       5562 ±  6%  sched_debug.cfs_rq[116]:/.tg_load_avg
      4906 ±  2%     +11.3%       5459 ±  7%  sched_debug.cfs_rq[117]:/.tg_load_avg
         5 ± 14%     +50.0%          7 ± 11%  sched_debug.cfs_rq[118]:/.load
         5 ±  8%     +38.1%          7 ± 11%  sched_debug.cfs_rq[118]:/.runnable_load_avg
         6 ± 13%    +500.0%         37 ± 42%  sched_debug.cfs_rq[26]:/.tg_load_contrib
  20927468 ±  4%      +8.4%   22685964 ±  4%  sched_debug.cfs_rq[27]:/.min_vruntime
  21203485 ±  3%     +13.4%   24049232 ±  7%  sched_debug.cfs_rq[28]:/.min_vruntime
  18134631 ±  6%     +13.3%   20548482 ± 12%  sched_debug.cfs_rq[29]:/.min_vruntime
         7 ± 14%     -30.0%          5 ± 20%  sched_debug.cfs_rq[2]:/.runnable_load_avg
         5 ± 15%     +36.4%          7 ± 14%  sched_debug.cfs_rq[45]:/.runnable_load_avg
  22214373 ±  6%      +9.3%   24278287 ±  4%  sched_debug.cfs_rq[64]:/.min_vruntime
         7 ± 20%     -40.0%          4 ± 24%  sched_debug.cfs_rq[67]:/.load
         8 ± 15%     -40.6%          4 ± 31%  sched_debug.cfs_rq[67]:/.runnable_load_avg
         9 ± 37%     -35.9%          6 ±  6%  sched_debug.cfs_rq[6]:/.runnable_load_avg
         9 ± 30%     -35.1%          6 ±  0%  sched_debug.cfs_rq[6]:/.load
        66 ± 41%     -68.9%         20 ± 38%  sched_debug.cfs_rq[72]:/.blocked_load_avg
        74 ± 38%     -63.4%         27 ± 29%  sched_debug.cfs_rq[72]:/.tg_load_contrib
         7 ± 14%     -33.3%          5 ± 24%  sched_debug.cfs_rq[73]:/.runnable_load_avg
         7 ± 17%     -24.1%          5 ± 15%  sched_debug.cfs_rq[73]:/.load
         8 ± 13%     -29.4%          6 ± 11%  sched_debug.cfs_rq[7]:/.load
         8 ± 13%     -27.3%          6 ± 11%  sched_debug.cfs_rq[7]:/.runnable_load_avg
         7 ±  6%     -26.7%          5 ± 15%  sched_debug.cfs_rq[82]:/.load
         8 ± 21%     -38.2%          5 ± 34%  sched_debug.cfs_rq[84]:/.runnable_load_avg
  20672959 ±  3%     +18.6%   24508843 ±  8%  sched_debug.cfs_rq[84]:/.min_vruntime
         7 ±  6%     -26.7%          5 ± 20%  sched_debug.cfs_rq[9]:/.load
         8 ± 17%     -35.3%          5 ± 20%  sched_debug.cfs_rq[9]:/.runnable_load_avg
     12917 ± 12%     -27.9%       9309 ±  4%  sched_debug.cpu#0.ttwu_count
      4804 ±  3%      -9.3%       4356 ±  7%  sched_debug.cpu#0.ttwu_local
      1907 ± 36%     -61.0%        744 ± 25%  sched_debug.cpu#100.ttwu_local
      5295 ± 24%     -56.4%       2309 ± 25%  sched_debug.cpu#100.ttwu_count
     32486 ±  4%     -17.8%      26713 ±  7%  sched_debug.cpu#104.nr_switches
     33183 ±  4%     -14.9%      28232 ±  7%  sched_debug.cpu#104.sched_count
         6 ± 12%     +33.3%          9 ±  7%  sched_debug.cpu#104.load
        15 ± 11%     +52.5%         23 ±  4%  sched_debug.cpu#104.nr_running
         5 ± 28%     +57.1%          8 ± 13%  sched_debug.cpu#106.cpu_load[0]
        12 ± 22%     +70.8%         20 ±  2%  sched_debug.cpu#106.nr_running
       242 ± 17%     -29.9%        169 ± 29%  sched_debug.cpu#106.sched_goidle
         5 ± 28%     +66.7%          8 ± 21%  sched_debug.cpu#106.load
        15 ±  4%     +20.0%         18 ±  5%  sched_debug.cpu#11.nr_running
      1608 ± 25%     -49.3%        815 ± 32%  sched_debug.cpu#11.ttwu_local
      2025 ± 28%     -30.6%       1405 ± 19%  sched_debug.cpu#110.ttwu_count
         6 ±  6%     +20.0%          7 ±  6%  sched_debug.cpu#110.load
       673 ± 15%     -32.9%        452 ± 25%  sched_debug.cpu#110.ttwu_local
        14 ±  9%     +28.1%         18 ±  4%  sched_debug.cpu#110.nr_running
        13 ± 18%     +39.6%         18 ±  8%  sched_debug.cpu#112.nr_running
      3580 ± 36%     -52.6%       1698 ± 36%  sched_debug.cpu#114.ttwu_count
        14 ± 10%     +25.9%         18 ±  8%  sched_debug.cpu#115.nr_running
       240 ±  9%     -31.8%        164 ± 31%  sched_debug.cpu#118.sched_goidle
         5 ±  9%     +31.8%          7 ±  5%  sched_debug.cpu#118.cpu_load[3]
         5 ±  9%     +31.8%          7 ±  5%  sched_debug.cpu#118.cpu_load[4]
         5 ±  8%     +38.1%          7 ± 11%  sched_debug.cpu#118.cpu_load[1]
         5 ±  8%     +38.1%          7 ± 11%  sched_debug.cpu#118.cpu_load[2]
         5 ± 14%     +50.0%          7 ± 11%  sched_debug.cpu#118.load
         5 ±  8%     +38.1%          7 ± 11%  sched_debug.cpu#118.cpu_load[0]
        12 ±  5%     +54.2%         18 ± 16%  sched_debug.cpu#118.nr_running
      5276 ± 14%     -39.7%       3182 ± 42%  sched_debug.cpu#16.ttwu_count
      6741 ± 46%     -66.3%       2270 ± 28%  sched_debug.cpu#17.ttwu_count
         7 ± 14%     -30.0%          5 ± 20%  sched_debug.cpu#2.cpu_load[0]
       376 ± 18%    +132.8%        877 ± 29%  sched_debug.cpu#21.sched_goidle
      6779 ± 20%     -60.5%       2681 ± 40%  sched_debug.cpu#23.ttwu_count
      5932 ± 36%     -57.0%       2552 ± 38%  sched_debug.cpu#26.ttwu_count
      1871 ± 43%     -50.9%        919 ± 38%  sched_debug.cpu#3.ttwu_local
      3782 ± 14%     -34.2%       2488 ± 11%  sched_debug.cpu#30.ttwu_count
        13 ±  3%     +36.4%         18 ± 12%  sched_debug.cpu#31.nr_running
      1664 ± 43%     -48.2%        861 ± 22%  sched_debug.cpu#32.ttwu_local
        14 ± 10%     +29.3%         18 ±  7%  sched_debug.cpu#32.nr_running
     31176 ± 12%     -16.1%      26153 ±  4%  sched_debug.cpu#32.sched_count
      5754 ± 22%     -54.7%       2606 ± 25%  sched_debug.cpu#36.ttwu_count
      1067 ± 16%     -31.1%        735 ± 16%  sched_debug.cpu#38.ttwu_local
     28900 ±  5%     -11.0%      25726 ±  7%  sched_debug.cpu#39.sched_count
      1257 ± 37%     -44.8%        693 ± 49%  sched_debug.cpu#39.ttwu_local
      4893 ± 21%     -54.0%       2249 ± 47%  sched_debug.cpu#40.ttwu_count
     28781 ±  4%      -9.0%      26197 ±  7%  sched_debug.cpu#40.sched_count
      4800 ± 29%     -51.7%       2320 ± 13%  sched_debug.cpu#42.ttwu_count
      1628 ± 35%     -53.3%        760 ± 26%  sched_debug.cpu#42.ttwu_local
     28887 ±  5%      -9.1%      26257 ±  7%  sched_debug.cpu#42.sched_count
         5 ± 15%     +36.4%          7 ± 14%  sched_debug.cpu#45.cpu_load[1]
         5 ± 15%     +36.4%          7 ± 14%  sched_debug.cpu#45.cpu_load[0]
     29036 ±  3%     -11.5%      25689 ±  4%  sched_debug.cpu#45.nr_switches
        13 ± 16%     +42.6%         19 ± 14%  sched_debug.cpu#45.nr_running
      6709 ± 35%     -69.9%       2020 ± 36%  sched_debug.cpu#45.ttwu_count
     30362 ±  6%     -13.8%      26186 ±  4%  sched_debug.cpu#45.sched_count
      1784 ± 24%     -53.0%        838 ± 23%  sched_debug.cpu#45.ttwu_local
     29093 ±  5%      -9.0%      26479 ±  5%  sched_debug.cpu#46.sched_count
        11 ± 13%     +57.4%         18 ± 22%  sched_debug.cpu#48.nr_running
        12 ± 23%     +40.8%         17 ±  8%  sched_debug.cpu#49.nr_running
      1020 ±  9%     -21.7%        799 ± 12%  sched_debug.cpu#50.ttwu_local
       514 ± 11%     -28.1%        370 ± 15%  sched_debug.cpu#52.sched_goidle
      5126 ± 36%     -63.0%       1898 ± 11%  sched_debug.cpu#53.ttwu_count
     29319 ±  5%     -11.6%      25920 ±  5%  sched_debug.cpu#54.sched_count
        13 ± 17%     +39.6%         18 ± 14%  sched_debug.cpu#54.nr_running
     28346 ±  3%      -9.9%      25537 ±  5%  sched_debug.cpu#54.nr_switches
       558 ±  5%     -35.9%        358 ± 12%  sched_debug.cpu#54.sched_goidle
      3768 ± 19%     -54.5%       1714 ± 43%  sched_debug.cpu#55.ttwu_count
     30565 ± 10%     -13.6%      26415 ±  2%  sched_debug.cpu#55.sched_count
      1582 ± 20%     -50.3%        787 ± 45%  sched_debug.cpu#56.ttwu_local
      6921 ± 37%     -73.0%       1871 ± 41%  sched_debug.cpu#56.ttwu_count
      1232 ± 18%     -47.0%        653 ± 20%  sched_debug.cpu#57.ttwu_local
       626 ± 41%     -38.0%        388 ± 11%  sched_debug.cpu#57.sched_goidle
     32348 ± 16%     -17.6%      26669 ±  2%  sched_debug.cpu#57.sched_count
      1441 ± 42%     -51.7%        695 ± 23%  sched_debug.cpu#58.ttwu_local
     41619 ± 45%     -35.7%      26769 ±  5%  sched_debug.cpu#59.sched_count
         9 ± 30%     -35.1%          6 ±  0%  sched_debug.cpu#6.load
         9 ± 37%     -35.9%          6 ±  6%  sched_debug.cpu#6.cpu_load[0]
         9 ± 37%     -35.9%          6 ±  6%  sched_debug.cpu#6.cpu_load[1]
         9 ± 39%     -34.2%          6 ±  6%  sched_debug.cpu#6.cpu_load[2]
     28059 ±  4%      -8.6%      25653 ±  4%  sched_debug.cpu#60.sched_count
      4532 ± 36%     -68.8%       1413 ± 40%  sched_debug.cpu#60.ttwu_count
      4914 ± 38%     -60.7%       1929 ± 36%  sched_debug.cpu#62.ttwu_count
      1128 ±  7%     -42.9%        644 ± 20%  sched_debug.cpu#62.ttwu_local
         7 ± 20%     -40.0%          4 ± 24%  sched_debug.cpu#67.load
     28008 ±  2%      -9.4%      25363 ±  5%  sched_debug.cpu#67.sched_count
     27655 ±  3%      -9.5%      25024 ±  5%  sched_debug.cpu#67.nr_switches
      1403 ± 21%     -56.1%        616 ±  8%  sched_debug.cpu#67.ttwu_local
      4889 ± 30%     -54.4%       2228 ± 36%  sched_debug.cpu#67.ttwu_count
      1174 ± 36%     -59.3%        477 ± 27%  sched_debug.cpu#69.ttwu_local
        15 ±  8%     +28.3%         19 ± 14%  sched_debug.cpu#69.nr_running
         8 ± 13%     -29.4%          6 ± 11%  sched_debug.cpu#7.load
      4411 ± 33%     -41.1%       2596 ± 36%  sched_debug.cpu#7.ttwu_count
      1528 ± 23%     -65.7%        524 ± 21%  sched_debug.cpu#71.ttwu_local
     28435 ±  3%     -11.3%      25224 ±  4%  sched_debug.cpu#71.sched_count
         7 ± 14%     -35.5%          5 ± 24%  sched_debug.cpu#73.cpu_load[2]
         8 ± 15%     -37.5%          5 ± 24%  sched_debug.cpu#73.cpu_load[3]
         7 ± 14%     -33.3%          5 ± 24%  sched_debug.cpu#73.cpu_load[1]
         8 ± 15%     -34.4%          5 ± 20%  sched_debug.cpu#73.cpu_load[4]
         7 ± 14%     -33.3%          5 ± 24%  sched_debug.cpu#73.cpu_load[0]
         7 ± 17%     -24.1%          5 ± 15%  sched_debug.cpu#73.load
     32392 ±  4%     -17.2%      26828 ±  5%  sched_debug.cpu#74.nr_switches
     34243 ±  8%     -17.3%      28320 ±  4%  sched_debug.cpu#74.sched_count
      1728 ± 29%     -50.9%        849 ± 39%  sched_debug.cpu#75.ttwu_local
      5389 ± 21%     -61.6%       2072 ± 39%  sched_debug.cpu#75.ttwu_count
     28023 ±  4%      -8.3%      25700 ±  6%  sched_debug.cpu#75.nr_switches
      3398 ± 11%     -33.9%       2248 ± 38%  sched_debug.cpu#76.ttwu_count
         7 ±  6%     -26.7%          5 ± 15%  sched_debug.cpu#82.load
         8 ± 16%     -37.1%          5 ± 32%  sched_debug.cpu#84.cpu_load[3]
         8 ± 21%     -40.0%          5 ± 34%  sched_debug.cpu#84.cpu_load[1]
         8 ± 16%     -40.0%          5 ± 34%  sched_debug.cpu#84.cpu_load[2]
         8 ± 16%     -37.1%          5 ± 32%  sched_debug.cpu#84.cpu_load[4]
         8 ± 21%     -38.2%          5 ± 34%  sched_debug.cpu#84.cpu_load[0]
      1555 ± 43%     -56.6%        675 ± 38%  sched_debug.cpu#87.ttwu_local
     33842 ±  1%     -15.1%      28727 ±  6%  sched_debug.cpu#89.sched_count
     33213 ±  1%     -18.8%      26968 ±  7%  sched_debug.cpu#89.nr_switches
         7 ±  6%     -26.7%          5 ± 20%  sched_debug.cpu#9.load
        12 ± 10%     +37.3%         17 ± 16%  sched_debug.cpu#90.nr_running
         5 ±  8%     +33.3%          7 ± 17%  sched_debug.cpu#90.cpu_load[4]
         5 ±  8%     +33.3%          7 ± 17%  sched_debug.cpu#90.cpu_load[3]
      2539 ± 15%     -66.6%        847 ± 39%  sched_debug.cpu#96.ttwu_count
      1124 ± 47%     -49.8%        564 ± 23%  sched_debug.cpu#97.ttwu_local
         4 ± 10%     +58.8%          6 ± 21%  sched_debug.cpu#99.cpu_load[3]
         4 ± 10%     +47.1%          6 ± 13%  sched_debug.cpu#99.cpu_load[4]
         4 ± 10%     +58.8%          6 ± 21%  sched_debug.cpu#99.cpu_load[2]
      1124 ± 34%     -53.8%        519 ± 12%  sched_debug.cpu#99.ttwu_local

testbox/testcase/testparams: lkp-ne02/netperf/performance-300s-200%-TCP_SENDFILE

98cf21c61a7f5419  bea66fbd11af1ca98ae26855ee  
----------------  --------------------------  
      3618 ±  2%      -8.1%       3326 ±  2%  slabinfo.proc_inode_cache.num_objs
      0.26 ±  7%      -8.7%       0.24 ±  5%  turbostat.CPU%c1
      2937 ±  0%     +12.9%       3314 ±  9%  numa-meminfo.node1.PageTables
    387540 ± 29%     -49.3%     196421 ± 39%  cpuidle.C1-NHM.time
     85799 ±  6%    +214.0%     269448 ± 37%  numa-vmstat.node0.numa_local
       733 ±  0%     +12.9%        827 ±  9%  numa-vmstat.node1.nr_page_table_pages
     69008 ±  4%    +555.5%     452353 ± 47%  numa-numastat.node0.local_node
     68999 ±  4%    +555.6%     452349 ± 47%  numa-numastat.node0.numa_hit
      3720 ±  9%     +67.5%       6232 ± 21%  proc-vmstat.pgalloc_dma
    550092 ±  8%     +32.5%     728886 ± 12%  proc-vmstat.pgalloc_dma32
       124 ±  6%      -5.4%        117 ±  7%  sched_debug.cfs_rq[0]:/.load
      5872 ± 11%     -17.9%       4821 ± 10%  sched_debug.cfs_rq[10]:/.tg_load_avg
       129 ±  4%      -7.6%        119 ±  5%  sched_debug.cfs_rq[10]:/.runnable_load_avg
       129 ±  4%     -10.4%        116 ±  4%  sched_debug.cfs_rq[10]:/.load
      5817 ± 13%     -17.2%       4816 ±  9%  sched_debug.cfs_rq[11]:/.tg_load_avg
   1511306 ±  3%      -7.4%    1399297 ±  5%  sched_debug.cfs_rq[12]:/.min_vruntime
      5799 ± 13%     -17.2%       4800 ± 10%  sched_debug.cfs_rq[12]:/.tg_load_avg
       629 ± 33%     -51.0%        308 ± 35%  sched_debug.cfs_rq[12]:/.tg_load_contrib
   1511248 ±  3%     -11.9%    1331134 ±  5%  sched_debug.cfs_rq[12]:/.MIN_vruntime
   1511249 ±  3%     -11.9%    1331134 ±  5%  sched_debug.cfs_rq[12]:/.max_vruntime
       522 ±  1%     -55.4%        233 ± 15%  sched_debug.cfs_rq[13]:/.tg_load_contrib
       406 ±  0%     -71.9%        114 ± 33%  sched_debug.cfs_rq[13]:/.blocked_load_avg
      5789 ± 13%     -17.3%       4788 ± 10%  sched_debug.cfs_rq[13]:/.tg_load_avg
      5784 ± 13%     -17.7%       4760 ±  9%  sched_debug.cfs_rq[14]:/.tg_load_avg
       643 ± 16%     -73.9%        168 ± 34%  sched_debug.cfs_rq[15]:/.tg_load_contrib
      5760 ± 12%     -17.7%       4742 ±  9%  sched_debug.cfs_rq[15]:/.tg_load_avg
       153 ±  9%    +142.6%        371 ± 33%  sched_debug.cfs_rq[2]:/.tg_load_contrib
       178 ± 20%    +191.3%        518 ± 37%  sched_debug.cfs_rq[3]:/.tg_load_contrib
   1484495 ±  1%     -10.4%    1330254 ±  0%  sched_debug.cfs_rq[4]:/.max_vruntime
   1484495 ±  1%     -10.4%    1330254 ±  0%  sched_debug.cfs_rq[4]:/.MIN_vruntime
   1525410 ±  4%      -9.1%    1387280 ±  5%  sched_debug.cfs_rq[4]:/.min_vruntime
       363 ±  0%     -38.9%        222 ± 31%  sched_debug.cfs_rq[5]:/.tg_load_contrib
      5881 ± 12%     -14.0%       5059 ±  5%  sched_debug.cfs_rq[6]:/.tg_load_avg
       654 ±  4%     -70.7%        191 ± 28%  sched_debug.cfs_rq[7]:/.tg_load_contrib
      5915 ± 11%     -14.7%       5043 ±  5%  sched_debug.cfs_rq[7]:/.tg_load_avg
      5895 ± 11%     -14.8%       5024 ±  5%  sched_debug.cfs_rq[8]:/.tg_load_avg
       144 ±  9%     -18.2%        117 ±  5%  sched_debug.cfs_rq[8]:/.load
        72 ± 18%    +305.9%        294 ± 38%  sched_debug.cfs_rq[8]:/.blocked_load_avg
       202 ±  6%    +104.7%        413 ± 26%  sched_debug.cfs_rq[8]:/.tg_load_contrib
      5887 ± 11%     -15.1%       4998 ±  6%  sched_debug.cfs_rq[9]:/.tg_load_avg
       128 ±  5%      -7.4%        119 ±  6%  sched_debug.cpu#0.cpu_load[4]
       128 ±  6%      -6.6%        119 ±  6%  sched_debug.cpu#0.cpu_load[3]
       127 ±  6%      -5.5%        120 ±  5%  sched_debug.cpu#0.cpu_load[2]
     12677 ± 35%     -49.6%       6383 ± 48%  sched_debug.cpu#1.sched_goidle
       127 ±  3%      -7.5%        117 ±  4%  sched_debug.cpu#10.cpu_load[3]
       127 ±  5%     -10.6%        114 ±  5%  sched_debug.cpu#10.load
       127 ±  3%      -7.3%        117 ±  4%  sched_debug.cpu#10.cpu_load[4]
    285821 ± 31%     -31.8%     194934 ±  0%  sched_debug.cpu#12.ttwu_local
    289858 ± 31%     -31.6%     198184 ±  0%  sched_debug.cpu#12.ttwu_count
    382013 ± 45%     -47.0%     202396 ±  0%  sched_debug.cpu#12.nr_switches
    382110 ± 45%     -47.0%     202488 ±  0%  sched_debug.cpu#12.sched_count
      1504 ± 24%     -22.1%       1171 ± 29%  sched_debug.cpu#12.curr->pid
      1670 ± 24%     -72.9%        453 ± 17%  sched_debug.cpu#13.sched_goidle
      1840 ± 18%     -41.9%       1068 ± 15%  sched_debug.cpu#14.curr->pid
       114 ±  0%      +6.4%        121 ±  4%  sched_debug.cpu#15.cpu_load[2]
      1494 ± 18%     -29.7%       1050 ± 13%  sched_debug.cpu#15.curr->pid
    211892 ±  5%      -5.7%     199867 ±  0%  sched_debug.cpu#2.ttwu_local
    330774 ± 35%     -37.9%     205319 ±  2%  sched_debug.cpu#4.ttwu_count
    324839 ± 36%     -38.4%     200110 ±  2%  sched_debug.cpu#4.ttwu_local
      1397 ±  0%     -23.7%       1066 ± 13%  sched_debug.cpu#6.curr->pid
       128 ±  2%      -7.2%        118 ±  4%  sched_debug.cpu#8.cpu_load[4]
       144 ±  7%     -18.2%        118 ±  5%  sched_debug.cpu#8.load
       141 ± 11%     -18.1%        115 ±  5%  sched_debug.cpu#9.load

brickland3: Brickland Ivy Bridge-EX
Memory: 512G

lkp-ne02: Nehalem-EP
Memory: 5G




                               aim7.time.system_time

  140 ++--------------------------------------------------------------------+
  130 ++       .*..   .*..                                                  |
      |    ..*.     ..    *..      .*..                          .*         |
  120 *+.*.        *         *...*.      ..*..*..*...  .*..*...*.           |
  110 ++                               *.            *.                     |
      |                                                                     |
  100 ++                                                                    |
   90 ++                                                                    |
   80 ++                                                                    |
      |                                                                  O  |
   70 ++                                                          O  O      O
   60 ++                                                                    |
      |  O                                              O  O   O            |
   50 ++     O     O      O      O         O     O                          |
   40 O+--------O------O-----O------O--O------O------O----------------------+


                         aim7.time.voluntary_context_switches

  1.1e+06 ++----------------------------------------------------------------+
          |               *..                                               |
    1e+06 ++   .*..     ..           .*..                                   |
   900000 *+.*.    *...*     *..*..*.      ..*..*..*..  .*..*..*...*        |
          |                              *.           *.                    |
   800000 ++                                                                |
   700000 ++                                                                |
          |                                                                 |
   600000 ++                                                                |
   500000 ++                                                                |
          |                                                                 |
   400000 ++                       O         O  O  O  O     O  O         O  |
   300000 ++ O  O  O   O  O     O     O  O               O            O     O
          O                  O                                     O        |
   200000 ++----------------------------------------------------------------+


                                 time.system_time

  140 ++--------------------------------------------------------------------+
  130 ++       .*..   .*..                                                  |
      |    ..*.     ..    *..      .*..                          .*         |
  120 *+.*.        *         *...*.      ..*..*..*...  .*..*...*.           |
  110 ++                               *.            *.                     |
      |                                                                     |
  100 ++                                                                    |
   90 ++                                                                    |
   80 ++                                                                    |
      |                                                                  O  |
   70 ++                                                          O  O      O
   60 ++                                                                    |
      |  O                                              O  O   O            |
   50 ++     O     O      O      O         O     O                          |
   40 O+--------O------O-----O------O--O------O------O----------------------+


                            time.voluntary_context_switches

  1.1e+06 ++----------------------------------------------------------------+
          |               *..                                               |
    1e+06 ++   .*..     ..           .*..                                   |
   900000 *+.*.    *...*     *..*..*.      ..*..*..*..  .*..*..*...*        |
          |                              *.           *.                    |
   800000 ++                                                                |
   700000 ++                                                                |
          |                                                                 |
   600000 ++                                                                |
   500000 ++                                                                |
          |                                                                 |
   400000 ++                       O         O  O  O  O     O  O         O  |
   300000 ++ O  O  O   O  O     O     O  O               O            O     O
          O                  O                                     O        |
   200000 ++----------------------------------------------------------------+


                                  vmstat.procs.b

  500 ++-----*--*--*--------------------------------------------------------+
      *.. ..           *..*..*...*..*..*...*..     ..*..*..*...*..*         |
  450 ++ *                                    *..*.                         |
  400 ++                                                                    |
      |                                                                     |
  350 ++                                                                    |
      |                                                                     |
  300 ++                                                                    |
      |                                                                     |
  250 ++                                                                    |
  200 ++               O     O                             O             O  |
      |  O         O             O  O  O   O  O  O   O  O      O            O
  150 ++     O  O         O                                                 |
      O                                                              O      |
  100 ++----------------------------------------------------------O---------+


                                  vmstat.system.in

  126000 ++-----------------------------------------------------------------+
         | .*..                                                             |
  125500 *+      ..*..  .*..         .*..      *..      .*..   *..*         |
  125000 ++    *.     *.    *..  ..*.        ..   *...*.     ..             |
         |                     *.        *..*               *               |
  124500 ++                                                                 |
  124000 ++                                                                 |
         |                                                                  |
  123500 ++                                                                 |
  123000 ++ O                                                               |
         |            O                     O     O         O  O            |
  122500 ++    O         O         O     O                        O   O  O  O
  122000 ++        O        O  O      O        O      O  O                  |
         |                                                                  |
  121500 O+-----------------------------------------------------------------+


                                  vmstat.system.cs

  33500 ++------------------------------------------------------------------+
  33000 ++   .*.       .*..      .*..*..                                    |
        *..*.   ..   *.    *...*.         ..*..*..*..  ..*..*..*..*         |
  32500 ++         ..                   *.           *.                     |
  32000 ++        *                                                         |
        |                                                                   |
  31500 ++                                                                  |
  31000 ++                                                                  |
  30500 ++                                                                  |
        |                                                   O               |
  30000 ++        O  O  O      O  O     O   O  O  O  O         O         O  |
  29500 O+ O  O            O         O                   O                  |
        |                                                             O     O
  29000 ++                                                        O         |
  28500 ++------------------------------------------------------------------+


                                 proc-vmstat.numa_hit

  3.3e+06 ++--------------*-------------------------------------------------+
          |        *.    : :                                                |
  3.2e+06 ++      +  ..  :  :                                               |
          *..*.. +      :   :                  .*..                         |
  3.1e+06 ++    *      *     *..*..        ..*.    *..  .*..  .*...*        |
    3e+06 ++                       *..*..*.           *.    *.              |
          |                                                                 |
  2.9e+06 ++                                                                |
          |                                                                 |
  2.8e+06 ++                                                                |
  2.7e+06 ++                                                                |
          |  O                                                        O     O
  2.6e+06 O+    O      O  O  O     O  O  O   O  O  O  O        O   O     O  |
          |        O            O                        O  O               |
  2.5e+06 ++----------------------------------------------------------------+


                                proc-vmstat.numa_local

  3.3e+06 ++--------------*-------------------------------------------------+
          |        *.    : :                                                |
  3.2e+06 ++      +  ..  :  :                                               |
          *..*.. +      :   :                  .*..                         |
  3.1e+06 ++    *      *     *..*..        ..*.    *..  .*..  .*...*        |
    3e+06 ++                       *..*..*.           *.    *.              |
          |                                                                 |
  2.9e+06 ++                                                                |
          |                                                                 |
  2.8e+06 ++                                                                |
  2.7e+06 ++                                                                |
          |  O                                                              O
  2.6e+06 O+    O      O  O  O     O  O  O   O  O  O  O        O   O  O  O  |
          |        O            O                        O  O               |
  2.5e+06 ++----------------------------------------------------------------+


                              proc-vmstat.pgalloc_normal

  3.4e+06 ++----------------------------------------------------------------+
          |               *                                                 |
  3.3e+06 ++             : :                                                |
  3.2e+06 ++       *..   :  :                                               |
          | .*.. ..   . :   :                  .*..                         |
  3.1e+06 *+    *      *     *..*..     .*...*.    *..  .*..  .*...*        |
    3e+06 ++                       *..*.              *.    *.              |
          |                                                                 |
  2.9e+06 ++                                                                |
  2.8e+06 ++                                                                |
          |                                                                 |
  2.7e+06 ++ O                                                     O        |
  2.6e+06 O+    O      O  O  O     O  O  O   O  O  O  O     O  O      O  O  O
          |        O            O                        O                  |
  2.5e+06 ++----------------------------------------------------------------+


                            proc-vmstat.numa_pages_migrated

  1.3e+06 ++----------------------------------------------------------------+
          |               *                                                 |
  1.2e+06 ++       *.    : :                                                |
          |       +  ..  :  :                                               |
  1.1e+06 *+.*.. +      :   :                  .*..                         |
    1e+06 ++    *      *     *..*..        ..*.    *..  .*..  .*...*        |
          |                        *..*..*.           *.    *.              |
   900000 ++                                                                |
          |                                                                 |
   800000 ++                                                                |
   700000 ++                                                                |
          |                                                                 |
   600000 ++ O                                                        O  O  O
          O            O  O        O  O  O   O  O  O  O     O  O   O        |
   500000 ++----O--O---------O--O------------------------O------------------+


                             proc-vmstat.pgmigrate_success

  1.3e+06 ++----------------------------------------------------------------+
          |               *                                                 |
  1.2e+06 ++       *.    : :                                                |
          |       +  ..  :  :                                               |
  1.1e+06 *+.*.. +      :   :                  .*..                         |
    1e+06 ++    *      *     *..*..        ..*.    *..  .*..  .*...*        |
          |                        *..*..*.           *.    *.              |
   900000 ++                                                                |
          |                                                                 |
   800000 ++                                                                |
   700000 ++                                                                |
          |                                                                 |
   600000 ++ O                                                        O  O  O
          O            O  O        O  O  O   O  O  O  O     O  O   O        |
   500000 ++----O--O---------O--O------------------------O------------------+


	[*] bisect-good sample
	[O] bisect-bad  sample

To reproduce:

	apt-get install ruby
	git clone git://git.kernel.org/pub/scm/linux/kernel/git/wfg/lkp-tests.git
	cd lkp-tests
	bin/setup-local job.yaml # the job file attached in this email
	bin/run-local   job.yaml


Disclaimer:
Results have been estimated based on internal Intel analysis and are provided
for informational purposes only. Any difference in system hardware or software
design or configuration may affect actual performance.


Thanks,
Ying Huang


View attachment "job.yaml" of type "text/plain" (1926 bytes)

View attachment "reproduce" of type "text/plain" (8890 bytes)

_______________________________________________
LKP mailing list
LKP@...ux.intel.com

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ