lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite for Android: free password hash cracker in your pocket
[<prev] [next>] [thread-next>] [day] [month] [year] [list]
Date:	Wed, 27 Apr 2016 11:15:56 +0800
From:	kernel test robot <xiaolong.ye@...el.com>
To:	Michal Hocko <mhocko@...e.com>
Cc:	Stephen Rothwell <sfr@...b.auug.org.au>,
	Hillf Danton <hillf.zj@...baba-inc.com>,
	Johannes Weiner <hannes@...xchg.org>,
	Mel Gorman <mgorman@...e.de>,
	David Rientjes <rientjes@...gle.com>,
	Tetsuo Handa <penguin-kernel@...ove.SAKURA.ne.jp>,
	KAMEZAWA Hiroyuki <kamezawa.hiroyu@...fujitsu.com>,
	Andrew Morton <akpm@...ux-foundation.org>,
	LKML <linux-kernel@...r.kernel.org>, lkp@...org
Subject: [lkp] [mm, oom] faad2185f4: vm-scalability.throughput -11.8%
 regression

FYI, we noticed vm-scalability.throughput -11.8% regression with the following commit:

https://git.kernel.org/pub/scm/linux/kernel/git/next/linux-next.git master
commit faad2185f482578d50d363746006a1b95dde9d0a ("mm, oom: rework oom detection")

on test machine: lkp-hsw-ep2: 72 threads Brickland Haswell-EP with 128G memory


=========================================================================================
compiler/cpufreq_governor/kconfig/nr_pmem/nr_task/rootfs/tbox_group/test/testcase/thp_defrag/thp_enabled:
  gcc-4.9/performance/x86_64-rhel-pmem/1/16/debian-x86_64-2015-02-07.cgz/lkp-hsw-ep2/swap-w-rand/vm-scalability/never/never

commit: 
  0da9597ac9c0adb8a521b9935fbe43d8b0e8cc64
  faad2185f482578d50d363746006a1b95dde9d0a

0da9597ac9c0adb8 faad2185f482578d50d3637460 
---------------- -------------------------- 
       fail:runs  %reproduction    fail:runs
         %stddev     %change         %stddev
             \          |                \  
     43802 ±  0%     -11.8%      38653 ±  0%  vm-scalability.throughput
    310.35 ±  0%    -100.0%       0.00 ± -1%  vm-scalability.time.elapsed_time
    310.35 ±  0%    -100.0%       0.00 ± -1%  vm-scalability.time.elapsed_time.max
    234551 ±  6%    -100.0%       0.00 ± -1%  vm-scalability.time.involuntary_context_switches
  44654748 ±  9%    -100.0%       0.00 ± -1%  vm-scalability.time.major_page_faults
   2442686 ± 11%    -100.0%       0.00 ± -1%  vm-scalability.time.maximum_resident_set_size
  34477365 ±  0%    -100.0%       0.00 ± -1%  vm-scalability.time.minor_page_faults
      4096 ±  0%    -100.0%       0.00 ± -1%  vm-scalability.time.page_size
      1595 ±  0%    -100.0%       0.00 ± -1%  vm-scalability.time.percent_of_cpu_this_job_got
      4935 ±  0%    -100.0%       0.00 ± -1%  vm-scalability.time.system_time
     19.08 ±  6%    -100.0%       0.00 ± -1%  vm-scalability.time.user_time
    342.89 ±  0%     -71.7%      96.99 ± -1%  uptime.boot
     18719 ±  1%     -70.3%       5555 ±  0%  uptime.idle
    227271 ±  3%     -68.0%      72623 ±  0%  softirqs.RCU
    208173 ±  7%     -69.7%      63118 ±  0%  softirqs.SCHED
   3204631 ±  1%     -73.0%     866292 ±  0%  softirqs.TIMER
    739.50 ±  0%      -1.6%     728.00 ±  0%  turbostat.Avg_MHz
     61.50 ±  3%     +20.3%      74.00 ± -1%  turbostat.CoreTmp
      0.07 ± 57%   +1092.7%       0.82 ±-121%  turbostat.Pkg%pc2
     64.75 ±  2%     +14.3%      74.00 ± -1%  turbostat.PkgTmp
     51.45 ±  0%      +1.8%      52.39 ± -1%  turbostat.RAMWatt
    789322 ±  4%     +49.2%    1177649 ±  0%  vmstat.memory.free
  53141272 ±  1%     -45.8%   28781900 ±  0%  vmstat.memory.swpd
      0.00 ±  0%      +Inf%       1.00 ±-100%  vmstat.procs.b
    780938 ±  7%     +66.2%    1297589 ±  0%  vmstat.swap.so
      4217 ±  6%    +103.4%       8576 ±  0%  vmstat.system.cs
    204460 ±  6%     +62.0%     331270 ±  0%  vmstat.system.in
   9128034 ± 43%     -85.7%    1306182 ±  0%  cpuidle.C1E-HSW.time
      5009 ± 52%     -88.9%     557.00 ±  0%  cpuidle.C1E-HSW.usage
      9110 ±130%     -93.3%     611.00 ±  0%  cpuidle.C3-HSW.usage
 1.655e+10 ±  0%     -79.5%  3.397e+09 ±  0%  cpuidle.C6-HSW.time
    621881 ±  2%     -71.5%     177398 ±  0%  cpuidle.C6-HSW.usage
  53981965 ± 58%     -80.4%   10553789 ±  0%  cpuidle.POLL.time
     85773 ±  9%     -18.4%      69982 ±  0%  cpuidle.POLL.usage
   2925199 ± 94%     -75.8%     706866 ±  0%  numa-numastat.node0.local_node
   2931002 ± 93%     -75.6%     716120 ±  0%  numa-numastat.node0.numa_hit
  12041792 ± 24%     -67.4%    3919657 ±  0%  numa-numastat.node0.numa_miss
  12047595 ± 24%     -67.4%    3928911 ±  0%  numa-numastat.node0.other_node
  64592910 ± 10%     -66.5%   21635175 ±  0%  numa-numastat.node1.local_node
  12041716 ± 24%     -67.5%    3919210 ±  0%  numa-numastat.node1.numa_foreign
  64601023 ± 10%     -66.5%   21639833 ±  0%  numa-numastat.node1.numa_hit
      4730 ± 13%    +290.9%      18491 ±  0%  meminfo.Inactive(file)
     12978 ±  8%     +46.3%      18985 ±  0%  meminfo.Mapped
    703327 ±  9%     +72.4%    1212584 ±  0%  meminfo.MemAvailable
    732344 ±  8%     +65.0%    1208500 ±  0%  meminfo.MemFree
     99286 ±  4%     +30.3%     129348 ±  0%  meminfo.SReclaimable
      3920 ± 21%    +332.5%      16955 ±  0%  meminfo.Shmem
    206164 ±  2%     +14.7%     236528 ±  0%  meminfo.Slab
      1113 ± 10%     +23.6%       1377 ±  0%  meminfo.SwapCached
  47130509 ±  3%     +53.1%   72150055 ±  0%  meminfo.SwapFree
      1012 ± 12%     +60.9%       1628 ±  0%  slabinfo.blkdev_requests.active_objs
      1012 ± 12%     +60.9%       1628 ±  0%  slabinfo.blkdev_requests.num_objs
      1531 ±  5%     +12.5%       1722 ±  0%  slabinfo.mnt_cache.active_objs
      1531 ±  5%     +12.5%       1722 ±  0%  slabinfo.mnt_cache.num_objs
      9719 ±  9%     -16.8%       8087 ±  0%  slabinfo.proc_inode_cache.num_objs
     92960 ±  6%     +69.6%     157683 ±  0%  slabinfo.radix_tree_node.active_objs
      9336 ±  9%     +35.2%      12624 ±  0%  slabinfo.radix_tree_node.active_slabs
     95203 ±  6%     +66.0%     158075 ±  0%  slabinfo.radix_tree_node.num_objs
      9336 ±  9%     +35.2%      12624 ±  0%  slabinfo.radix_tree_node.num_slabs
    310.35 ±  0%    -100.0%       0.00 ± -1%  time.elapsed_time
    310.35 ±  0%    -100.0%       0.00 ± -1%  time.elapsed_time.max
    600.00 ± 27%    -100.0%       0.00 ± -1%  time.file_system_inputs
    234551 ±  6%    -100.0%       0.00 ± -1%  time.involuntary_context_switches
  44654748 ±  9%    -100.0%       0.00 ± -1%  time.major_page_faults
   2442686 ± 11%    -100.0%       0.00 ± -1%  time.maximum_resident_set_size
  34477365 ±  0%    -100.0%       0.00 ± -1%  time.minor_page_faults
      4096 ±  0%    -100.0%       0.00 ± -1%  time.page_size
      1595 ±  0%    -100.0%       0.00 ± -1%  time.percent_of_cpu_this_job_got
      4935 ±  0%    -100.0%       0.00 ± -1%  time.system_time
     19.08 ±  6%    -100.0%       0.00 ± -1%  time.user_time
    390.50 ± 34%    -100.0%       0.00 ± -1%  time.voluntary_context_switches
    914507 ±  7%     -13.3%     792912 ±  0%  numa-meminfo.node0.Active
    913915 ±  7%     -13.5%     790259 ±  0%  numa-meminfo.node0.Active(anon)
    592.00 ± 31%    +348.1%       2653 ±  0%  numa-meminfo.node0.Active(file)
   1217059 ±  7%     -13.7%    1049893 ±  0%  numa-meminfo.node0.AnonPages
    306384 ±  7%     -12.0%     269631 ±  0%  numa-meminfo.node0.Inactive
    304389 ±  7%     -14.4%     260426 ±  0%  numa-meminfo.node0.Inactive(anon)
      1995 ±  8%    +361.3%       9204 ±  0%  numa-meminfo.node0.Inactive(file)
      5801 ±  4%     +16.7%       6772 ±  0%  numa-meminfo.node0.Mapped
     32196 ±  7%     +36.4%      43932 ±  0%  numa-meminfo.node0.MemFree
     55651 ±  5%     +10.6%      61563 ±  0%  numa-meminfo.node0.SUnreclaim
      2966 ± 15%    +232.9%       9875 ±  0%  numa-meminfo.node1.Inactive(file)
      7446 ± 13%     +67.7%      12486 ±  0%  numa-meminfo.node1.Mapped
    679948 ±  6%     +76.7%    1201231 ±  0%  numa-meminfo.node1.MemFree
     66811 ±  7%     +48.5%      99246 ±  0%  numa-meminfo.node1.SReclaimable
     51227 ±  5%     -11.0%      45616 ±  0%  numa-meminfo.node1.SUnreclaim
      3090 ± 39%    +415.6%      15932 ±  0%  numa-meminfo.node1.Shmem
    118039 ±  3%     +22.7%     144863 ±  0%  numa-meminfo.node1.Slab
      0.00 ± -1%      +Inf%       1.58 ±-63%  perf-profile.cycles-pp.__alloc_pages_slowpath.constprop.93.__alloc_pages_nodemask.alloc_kmem_pages_node.copy_process._do_fork
      0.00 ± -1%      +Inf%      26.40 ± -3%  perf-profile.cycles-pp.__alloc_pages_slowpath.constprop.93.__alloc_pages_nodemask.alloc_pages_vma.__read_swap_cache_async.read_swap_cache_async
      0.00 ± -1%      +Inf%      39.64 ± -2%  perf-profile.cycles-pp.__alloc_pages_slowpath.constprop.93.__alloc_pages_nodemask.alloc_pages_vma.handle_mm_fault.__do_page_fault
      5.20 ±140%    -100.0%       0.00 ± -1%  perf-profile.cycles-pp.do_try_to_free_pages.try_to_free_pages.__alloc_pages_nodemask.alloc_kmem_pages_node.copy_process
     25.02 ± 10%    -100.0%       0.00 ± -1%  perf-profile.cycles-pp.do_try_to_free_pages.try_to_free_pages.__alloc_pages_nodemask.alloc_pages_vma.__read_swap_cache_async
     38.03 ±  9%    -100.0%       0.00 ± -1%  perf-profile.cycles-pp.do_try_to_free_pages.try_to_free_pages.__alloc_pages_nodemask.alloc_pages_vma.handle_mm_fault
      0.00 ± -1%      +Inf%       1.59 ±-62%  perf-profile.cycles-pp.do_try_to_free_pages.try_to_free_pages.__alloc_pages_slowpath.__alloc_pages_nodemask.alloc_kmem_pages_node
      0.00 ± -1%      +Inf%      65.24 ± -1%  perf-profile.cycles-pp.do_try_to_free_pages.try_to_free_pages.__alloc_pages_slowpath.__alloc_pages_nodemask.alloc_pages_vma
      5.20 ±140%    -100.0%       0.00 ± -1%  perf-profile.cycles-pp.shrink_zone.do_try_to_free_pages.try_to_free_pages.__alloc_pages_nodemask.alloc_kmem_pages_node
     63.09 ±  8%    -100.0%       0.00 ± -1%  perf-profile.cycles-pp.shrink_zone.do_try_to_free_pages.try_to_free_pages.__alloc_pages_nodemask.alloc_pages_vma
      0.00 ± -1%      +Inf%      67.08 ± -1%  perf-profile.cycles-pp.shrink_zone.do_try_to_free_pages.try_to_free_pages.__alloc_pages_slowpath.__alloc_pages_nodemask
     69.00 ±  2%    -100.0%       0.00 ± -1%  perf-profile.cycles-pp.shrink_zone_memcg.shrink_zone.do_try_to_free_pages.try_to_free_pages.__alloc_pages_nodemask
      0.00 ± -1%      +Inf%      66.87 ± -1%  perf-profile.cycles-pp.shrink_zone_memcg.shrink_zone.do_try_to_free_pages.try_to_free_pages.__alloc_pages_slowpath
      5.20 ±140%    -100.0%       0.00 ± -1%  perf-profile.cycles-pp.try_to_free_pages.__alloc_pages_nodemask.alloc_kmem_pages_node.copy_process._do_fork
     25.05 ± 11%    -100.0%       0.00 ± -1%  perf-profile.cycles-pp.try_to_free_pages.__alloc_pages_nodemask.alloc_pages_vma.__read_swap_cache_async.read_swap_cache_async
     38.06 ±  9%    -100.0%       0.00 ± -1%  perf-profile.cycles-pp.try_to_free_pages.__alloc_pages_nodemask.alloc_pages_vma.handle_mm_fault.__do_page_fault
      0.00 ± -1%      +Inf%       1.59 ±-62%  perf-profile.cycles-pp.try_to_free_pages.__alloc_pages_slowpath.__alloc_pages_nodemask.alloc_kmem_pages_node.copy_process
      0.00 ± -1%      +Inf%      26.22 ± -3%  perf-profile.cycles-pp.try_to_free_pages.__alloc_pages_slowpath.__alloc_pages_nodemask.alloc_pages_vma.__read_swap_cache_async
      0.00 ± -1%      +Inf%      39.01 ± -2%  perf-profile.cycles-pp.try_to_free_pages.__alloc_pages_slowpath.__alloc_pages_nodemask.alloc_pages_vma.handle_mm_fault
    228466 ±  7%     -13.4%     197776 ±  0%  numa-vmstat.node0.nr_active_anon
    147.50 ± 31%    +334.6%     641.00 ±  0%  numa-vmstat.node0.nr_active_file
    304255 ±  7%     -13.6%     262822 ±  0%  numa-vmstat.node0.nr_anon_pages
      8062 ±  8%     +34.3%      10829 ±  0%  numa-vmstat.node0.nr_free_pages
     76095 ±  7%     -14.3%      65250 ±  0%  numa-vmstat.node0.nr_inactive_anon
    498.00 ±  8%    +347.8%       2230 ±  0%  numa-vmstat.node0.nr_inactive_file
      1466 ±  5%     +15.0%       1686 ±  0%  numa-vmstat.node0.nr_mapped
     13912 ±  5%     +10.6%      15390 ±  0%  numa-vmstat.node0.nr_slab_unreclaimable
   7585474 ±  5%     -73.6%    2005989 ±  0%  numa-vmstat.node0.nr_vmscan_write
   7585495 ±  5%     -73.6%    2006038 ±  0%  numa-vmstat.node0.nr_written
   2042806 ± 73%     -60.3%     810553 ±  0%  numa-vmstat.node0.numa_hit
   1973969 ± 76%     -62.6%     737625 ±  0%  numa-vmstat.node0.numa_local
   6640606 ± 22%     -72.4%    1834872 ±  0%  numa-vmstat.node0.numa_miss
   6709443 ± 22%     -71.6%    1907800 ±  0%  numa-vmstat.node0.numa_other
    169806 ±  5%     +71.9%     291868 ±  0%  numa-vmstat.node1.nr_free_pages
    740.75 ± 15%    +223.1%       2393 ±  0%  numa-vmstat.node1.nr_inactive_file
      1860 ± 13%     +69.4%       3153 ±  0%  numa-vmstat.node1.nr_mapped
    767.88 ± 39%    +415.4%       3958 ±  0%  numa-vmstat.node1.nr_shmem
     16698 ±  7%     +46.9%      24534 ±  0%  numa-vmstat.node1.nr_slab_reclaimable
     12806 ±  5%     -10.9%      11405 ±  0%  numa-vmstat.node1.nr_slab_unreclaimable
  27889038 ±  6%     -68.4%    8818477 ±  0%  numa-vmstat.node1.nr_vmscan_write
  27889106 ±  6%     -68.4%    8818479 ±  0%  numa-vmstat.node1.nr_written
   6640458 ± 22%     -72.4%    1834609 ±  0%  numa-vmstat.node1.numa_foreign
  38283180 ±  9%     -71.4%   10962630 ±  0%  numa-vmstat.node1.numa_hit
  38265243 ±  9%     -71.4%   10948754 ±  0%  numa-vmstat.node1.numa_local
    539498 ±  6%     -66.2%     182224 ±  0%  proc-vmstat.allocstall
    144.38 ± 22%     -96.5%       5.00 ±-20%  proc-vmstat.compact_fail
  15889726 ± 25%     -87.2%    2027142 ±  0%  proc-vmstat.compact_free_scanned
      7424 ± 21%     -95.5%     337.00 ±  0%  proc-vmstat.compact_isolated
     18421 ±120%     -98.3%     310.00 ±  0%  proc-vmstat.compact_migrate_scanned
    192.00 ± 21%     -96.4%       7.00 ±-14%  proc-vmstat.compact_stall
     49525 ± 43%    +154.6%     126090 ±  0%  proc-vmstat.kswapd_low_wmark_hit_quickly
     17344 ±  4%     +73.0%      30013 ±  0%  proc-vmstat.nr_dirty_background_threshold
     34690 ±  4%     +73.0%      60026 ±  0%  proc-vmstat.nr_dirty_threshold
    180484 ±  4%     +67.9%     303083 ±  0%  proc-vmstat.nr_free_pages
      1227 ± 10%    +276.8%       4623 ±  0%  proc-vmstat.nr_inactive_file
      3303 ±  6%     +43.0%       4722 ±  0%  proc-vmstat.nr_mapped
      1012 ± 17%    +321.7%       4270 ±  0%  proc-vmstat.nr_shmem
     24900 ±  4%     +29.6%      32265 ±  0%  proc-vmstat.nr_slab_reclaimable
  35587470 ±  5%     -69.7%   10775004 ±  0%  proc-vmstat.nr_vmscan_write
  61007414 ±  6%     -65.6%   21016129 ±  0%  proc-vmstat.nr_written
  16970144 ± 12%     -37.0%   10686065 ±  0%  proc-vmstat.numa_foreign
  10074000 ±  1%     -45.2%    5519749 ±  0%  proc-vmstat.numa_hint_faults
   9673661 ±  5%     -44.4%    5377833 ±  0%  proc-vmstat.numa_hint_faults_local
  67528367 ±  6%     -67.0%   22278204 ±  0%  proc-vmstat.numa_hit
  67514451 ±  6%     -67.0%   22264292 ±  0%  proc-vmstat.numa_local
  16969897 ± 12%     -37.0%   10686272 ±  0%  proc-vmstat.numa_miss
  16983813 ± 12%     -37.0%   10700184 ±  0%  proc-vmstat.numa_other
  41943046 ±  1%     -43.9%   23535513 ±  0%  proc-vmstat.numa_pte_updates
     49539 ± 43%    +154.5%     126102 ±  0%  proc-vmstat.pageoutrun
  45300466 ±  9%     -79.2%    9418945 ±  0%  proc-vmstat.pgactivate
    558557 ± 14%     -34.1%     367818 ±  0%  proc-vmstat.pgalloc_dma
  14967174 ±  3%     -69.1%    4626484 ±  0%  proc-vmstat.pgalloc_dma32
  71037855 ±  7%     -57.6%   30119030 ±  0%  proc-vmstat.pgalloc_normal
  62292933 ±  6%     -65.2%   21706559 ±  0%  proc-vmstat.pgdeactivate
  79824509 ±  5%     -56.2%   34976920 ±  0%  proc-vmstat.pgfault
  86163698 ±  6%     -68.5%   27162999 ±  0%  proc-vmstat.pgfree
  44685673 ±  9%     -79.4%    9192073 ±  0%  proc-vmstat.pgmajfault
  13765509 ±  7%     -69.7%    4168976 ±  0%  proc-vmstat.pgrefill_dma32
  48547731 ±  6%     -63.8%   17561899 ±  0%  proc-vmstat.pgrefill_normal
  12122138 ±  7%     -69.7%    3675632 ±  0%  proc-vmstat.pgscan_direct_dma32
  67953310 ±  7%     -66.4%   22842830 ±  0%  proc-vmstat.pgscan_direct_normal
  11915527 ± 10%     -79.3%    2460668 ±  0%  proc-vmstat.pgscan_kswapd_dma32
  15559179 ±  9%     -80.7%    2995996 ±  0%  proc-vmstat.pgscan_kswapd_normal
   8844259 ±  8%     -70.8%    2582588 ±  0%  proc-vmstat.pgsteal_direct_dma32
  43061102 ±  7%     -64.0%   15515081 ±  0%  proc-vmstat.pgsteal_direct_normal
   4732303 ±  6%     -69.0%    1469200 ±  0%  proc-vmstat.pgsteal_kswapd_dma32
   4380170 ±  7%     -66.6%    1462100 ±  0%  proc-vmstat.pgsteal_kswapd_normal
  44709819 ±  9%     -79.4%    9217280 ±  0%  proc-vmstat.pswpin
  61007674 ±  6%     -65.6%   21016726 ±  0%  proc-vmstat.pswpout
     37.61 ±  8%     -39.3%      22.83 ± -4%  sched_debug.cfs_rq:/.load.avg
    884.52 ±  5%     -36.7%     559.50 ±  0%  sched_debug.cfs_rq:/.load.max
    146.88 ±  5%     -38.3%      90.64 ± -1%  sched_debug.cfs_rq:/.load.stddev
     47.93 ±  5%     +28.5%      61.60 ± -1%  sched_debug.cfs_rq:/.load_avg.avg
      1095 ± 10%     +52.2%       1667 ±  0%  sched_debug.cfs_rq:/.load_avg.max
    170.96 ±  7%     +39.6%     238.66 ±  0%  sched_debug.cfs_rq:/.load_avg.stddev
    578829 ±  2%     -80.5%     112739 ±  0%  sched_debug.cfs_rq:/.min_vruntime.avg
   2507544 ±  0%     -80.8%     482665 ±  0%  sched_debug.cfs_rq:/.min_vruntime.max
    998179 ±  1%     -82.2%     177613 ±  0%  sched_debug.cfs_rq:/.min_vruntime.stddev
      0.24 ±  2%     -37.1%       0.15 ±-654%  sched_debug.cfs_rq:/.nr_running.avg
      0.41 ±  1%     -22.1%       0.32 ±-312%  sched_debug.cfs_rq:/.nr_running.stddev
     34.69 ±  0%     -38.3%      21.40 ± -4%  sched_debug.cfs_rq:/.runnable_load_avg.avg
    849.33 ±  0%     -37.9%     527.50 ±  0%  sched_debug.cfs_rq:/.runnable_load_avg.max
    138.61 ±  0%     -38.4%      85.43 ± -1%  sched_debug.cfs_rq:/.runnable_load_avg.stddev
    444145 ± 30%     -87.1%      57376 ±  0%  sched_debug.cfs_rq:/.spread0.avg
   2372869 ±  5%     -82.0%     427303 ±  0%  sched_debug.cfs_rq:/.spread0.max
    998183 ±  1%     -82.2%     177613 ±  0%  sched_debug.cfs_rq:/.spread0.stddev
    242.15 ±  1%     -28.8%     172.45 ±  0%  sched_debug.cfs_rq:/.util_avg.avg
    392.49 ±  0%     -21.4%     308.56 ±  0%  sched_debug.cfs_rq:/.util_avg.stddev
    184988 ±  1%     -64.1%      66460 ±  0%  sched_debug.cpu.clock.avg
    184996 ±  1%     -64.1%      66466 ±  0%  sched_debug.cpu.clock.max
    184978 ±  1%     -64.1%      66453 ±  0%  sched_debug.cpu.clock.min
      5.60 ± 18%     -30.8%       3.88 ±-25%  sched_debug.cpu.clock.stddev
    184988 ±  1%     -64.1%      66460 ±  0%  sched_debug.cpu.clock_task.avg
    184996 ±  1%     -64.1%      66466 ±  0%  sched_debug.cpu.clock_task.max
    184978 ±  1%     -64.1%      66453 ±  0%  sched_debug.cpu.clock_task.min
      5.60 ± 18%     -30.8%       3.88 ±-25%  sched_debug.cpu.clock_task.stddev
     36.54 ±  4%     -42.2%      21.11 ± -4%  sched_debug.cpu.cpu_load[0].avg
    950.98 ±  7%     -44.5%     527.50 ±  0%  sched_debug.cpu.cpu_load[0].max
    151.40 ±  6%     -43.7%      85.22 ± -1%  sched_debug.cpu.cpu_load[0].stddev
     35.91 ±  2%     -41.2%      21.10 ± -4%  sched_debug.cpu.cpu_load[1].avg
    899.77 ±  3%     -41.4%     527.50 ±  0%  sched_debug.cpu.cpu_load[1].max
    145.18 ±  3%     -41.3%      85.22 ± -1%  sched_debug.cpu.cpu_load[1].stddev
     35.61 ±  2%     -40.7%      21.12 ± -4%  sched_debug.cpu.cpu_load[2].avg
    877.87 ±  2%     -39.9%     527.50 ±  0%  sched_debug.cpu.cpu_load[2].max
    142.54 ±  2%     -40.2%      85.23 ± -1%  sched_debug.cpu.cpu_load[2].stddev
     35.37 ±  2%     -40.1%      21.20 ± -4%  sched_debug.cpu.cpu_load[3].avg
    867.60 ±  2%     -39.2%     527.50 ±  0%  sched_debug.cpu.cpu_load[3].max
    141.21 ±  2%     -39.6%      85.33 ± -1%  sched_debug.cpu.cpu_load[3].stddev
     35.16 ±  1%     -39.6%      21.24 ± -4%  sched_debug.cpu.cpu_load[4].avg
    858.88 ±  2%     -38.6%     527.50 ±  0%  sched_debug.cpu.cpu_load[4].max
    140.16 ±  2%     -39.0%      85.43 ± -1%  sched_debug.cpu.cpu_load[4].stddev
    456.75 ±  2%     -41.9%     265.40 ±  0%  sched_debug.cpu.curr->pid.avg
      5331 ±  1%     -53.3%       2491 ±  0%  sched_debug.cpu.curr->pid.max
    912.17 ±  2%     -38.3%     562.54 ±  0%  sched_debug.cpu.curr->pid.stddev
     37.90 ±  7%     -39.8%      22.83 ± -4%  sched_debug.cpu.load.avg
    904.00 ±  7%     -38.1%     559.50 ±  0%  sched_debug.cpu.load.max
    149.42 ±  6%     -39.3%      90.64 ± -1%  sched_debug.cpu.load.stddev
      0.00 ±  5%     -31.3%       0.00 ±-4394145%  sched_debug.cpu.next_balance.stddev
     72445 ±  0%     -75.8%      17509 ±  0%  sched_debug.cpu.nr_load_updates.avg
    155443 ±  0%     -77.5%      34902 ±  0%  sched_debug.cpu.nr_load_updates.max
     19501 ± 33%     -61.4%       7530 ±  0%  sched_debug.cpu.nr_load_updates.min
     46290 ±  1%     -80.6%       8995 ±  0%  sched_debug.cpu.nr_load_updates.stddev
      0.25 ±  3%     -34.8%       0.16 ±-626%  sched_debug.cpu.nr_running.avg
      1.12 ±  6%     +33.3%       1.50 ±-66%  sched_debug.cpu.nr_running.max
      0.42 ±  2%     -14.1%       0.36 ±-276%  sched_debug.cpu.nr_running.stddev
      9841 ±  4%     -54.7%       4459 ±  0%  sched_debug.cpu.nr_switches.avg
    323.71 ± 11%     -26.8%     237.00 ±  0%  sched_debug.cpu.nr_switches.min
     10606 ±  6%     -26.9%       7748 ±  0%  sched_debug.cpu.nr_switches.stddev
      0.00 ± 68%   +1100.0%       0.05 ±-2057%  sched_debug.cpu.nr_uninterruptible.avg
    184975 ±  1%     -64.1%      66455 ±  0%  sched_debug.cpu_clk
    181780 ±  1%     -65.3%      63027 ±  0%  sched_debug.ktime
      0.12 ±  5%    +205.5%       0.36 ±-275%  sched_debug.rt_rq:/.rt_time.avg
      4.33 ±  7%    +211.4%      13.48 ± -7%  sched_debug.rt_rq:/.rt_time.max
      0.60 ±  6%    +208.6%       1.85 ±-54%  sched_debug.rt_rq:/.rt_time.stddev
    184975 ±  1%     -64.1%      66455 ±  0%  sched_debug.sched_clk


lkp-hsw-ep2: 72 threads Brickland Haswell-EP with 128G memory


	[*] bisect-good sample
	[O] bisect-bad  sample

To reproduce:

        git clone git://git.kernel.org/pub/scm/linux/kernel/git/wfg/lkp-tests.git
        cd lkp-tests
        bin/lkp install job.yaml  # job file is attached in this email
        bin/lkp run     job.yaml


Disclaimer:
Results have been estimated based on internal Intel analysis and are provided
for informational purposes only. Any difference in system hardware or software
design or configuration may affect actual performance.


Thanks,
Xiaolong

View attachment "job.yaml" of type "text/plain" (3787 bytes)

View attachment "reproduce" of type "text/plain" (7236 bytes)

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ