lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Date:	Mon, 15 Aug 2016 22:14:55 +0800
From:	Fengguang Wu <fengguang.wu@...el.com>
To:	Christoph Hellwig <hch@....de>
Cc:	Dave Chinner <david@...morbit.com>,
	Ye Xiaolong <xiaolong.ye@...el.com>,
	Linus Torvalds <torvalds@...ux-foundation.org>,
	LKML <linux-kernel@...r.kernel.org>,
	Bob Peterson <rpeterso@...hat.com>, LKP <lkp@...org>
Subject: Re: [LKP] [lkp] [xfs] 68a9f5e700: aim7.jobs-per-min -13.6% regression

Hi Christoph,

On Sun, Aug 14, 2016 at 06:17:24PM +0200, Christoph Hellwig wrote:
>Snipping the long contest:
>
>I think there are three observations here:
>
> (1) removing the mark_page_accessed (which is the only significant
>     change in the parent commit)  hurts the
>     aim7/1BRD_48G-xfs-disk_rr-3000-performance/ivb44 test.
>     I'd still rather stick to the filemap version and let the
>     VM people sort it out.  How do the numbers for this test
>     look for XFS vs say ext4 and btrfs?
> (2) lots of additional spinlock contention in the new case.  A quick
>     check shows that I fat-fingered my rewrite so that we do
>     the xfs_inode_set_eofblocks_tag call now for the pure lookup
>     case, and pretty much all new cycles come from that.
> (3) Boy, are those xfs_inode_set_eofblocks_tag calls expensive, and
>     we're already doing way to many even without my little bug above.
>
>So I've force pushed a new version of the iomap-fixes branch with
>(2) fixed, and also a little patch to xfs_inode_set_eofblocks_tag a
>lot less expensive slotted in before that.  Would be good to see
>the numbers with that.

The aim7 1BRD tests finished and there are ups and downs, with overall
performance remain flat.

99091700659f4df9  74a242ad94d13436a1644c0b45  bf4dc6e4ecc2a3d042029319bc  testcase/testparams/testbox
----------------  --------------------------  --------------------------  ---------------------------
         %stddev     %change         %stddev     %change         %stddev
             \          |                \          |                \  
    159926                      157324                      158574        GEO-MEAN aim7.jobs-per-min
     70897               5%      74137               4%      73775        aim7/1BRD_48G-xfs-creat-clo-1500-performance/ivb44
    485217 ±  3%                492431                      477533        aim7/1BRD_48G-xfs-disk_rd-9000-performance/ivb44
    360451             -19%     292980             -17%     299377        aim7/1BRD_48G-xfs-disk_rr-3000-performance/ivb44
    338114                      338410               5%     354078        aim7/1BRD_48G-xfs-disk_rw-3000-performance/ivb44
     60130 ±  5%         4%      62438               5%      62923        aim7/1BRD_48G-xfs-disk_src-3000-performance/ivb44
    403144                      397790                      410648        aim7/1BRD_48G-xfs-disk_wrt-3000-performance/ivb44
     26327                       26534                       26128        aim7/1BRD_48G-xfs-sync_disk_rw-600-performance/ivb44

The new commit bf4dc6e ("xfs: rewrite and optimize the delalloc write
path") improves the aim7/1BRD_48G-xfs-disk_rw-3000-performance/ivb44
case by 5%. Here are the detailed numbers:

aim7/1BRD_48G-xfs-disk_rw-3000-performance/ivb44

74a242ad94d13436  bf4dc6e4ecc2a3d042029319bc
----------------  --------------------------
         %stddev     %change         %stddev
             \          |                \
    338410               5%     354078        aim7.jobs-per-min
    404390               8%     435117        aim7.time.voluntary_context_switches
      2502              -4%       2396        aim7.time.maximum_resident_set_size
     15018              -9%      13701        aim7.time.involuntary_context_switches
       900             -11%        801        aim7.time.system_time
     17432              11%      19365        vmstat.system.cs
     47736 ± 19%       -24%      36087        interrupts.CAL:Function_call_interrupts
   2129646              31%    2790638        proc-vmstat.pgalloc_dma32
    379503              13%     429384        numa-meminfo.node0.Dirty
     15018              -9%      13701        time.involuntary_context_switches
       900             -11%        801        time.system_time
      1560              10%       1716        slabinfo.mnt_cache.active_objs
      1560              10%       1716        slabinfo.mnt_cache.num_objs
     61.53               -4      57.45 ±  4%  perf-profile.cycles-pp.intel_idle.cpuidle_enter_state.cpuidle_enter.call_cpuidle.cpu_startup_entry
     61.63               -4      57.55 ±  4%  perf-profile.func.cycles-pp.intel_idle
   1007188 ± 16%       156%    2577911 ±  6%  numa-numastat.node0.numa_miss
   9662857 ±  4%       -13%    8420159 ±  3%  numa-numastat.node0.numa_foreign
   1008220 ± 16%       155%    2570630 ±  6%  numa-numastat.node1.numa_foreign
   9664033 ±  4%       -13%    8413184 ±  3%  numa-numastat.node1.numa_miss
  26519887 ±  3%        18%   31322674        cpuidle.C1-IVT.time
    122238              16%     142383        cpuidle.C1-IVT.usage
     46548              11%      51645        cpuidle.C1E-IVT.usage
  17253419              13%   19567582        cpuidle.C3-IVT.time
     86847              13%      98333        cpuidle.C3-IVT.usage
    482033 ± 12%       108%    1000665 ±  8%  numa-vmstat.node0.numa_miss
     94689              14%     107744        numa-vmstat.node0.nr_zone_write_pending
     94677              14%     107718        numa-vmstat.node0.nr_dirty
   3156643 ±  3%       -20%    2527460 ±  3%  numa-vmstat.node0.numa_foreign
    429288 ± 12%       129%     983053 ±  8%  numa-vmstat.node1.numa_foreign
   3104193 ±  3%       -19%    2510128        numa-vmstat.node1.numa_miss
      6.43 ±  5%        51%       9.70 ± 11%  turbostat.Pkg%pc2
      0.30              28%       0.38        turbostat.CPU%c3
      9.71                        9.92        turbostat.RAMWatt
       158                         154        turbostat.PkgWatt
       125              -3%        121        turbostat.CorWatt
      1141              -6%       1078        turbostat.Avg_MHz
     38.70              -6%      36.48        turbostat.%Busy
      5.03 ± 11%       -51%       2.46 ± 40%  turbostat.Pkg%pc6
      8.33 ± 48%        88%      15.67 ± 36%  sched_debug.cfs_rq:/.runnable_load_avg.max
      1947 ±  3%       -12%       1710 ±  7%  sched_debug.cfs_rq:/.spread0.stddev
      1936 ±  3%       -12%       1698 ±  8%  sched_debug.cfs_rq:/.min_vruntime.stddev
      2170 ± 10%       -14%       1863 ±  6%  sched_debug.cfs_rq:/.load_avg.max
    220926 ± 18%        37%     303192 ±  5%  sched_debug.cpu.avg_idle.stddev
      0.06 ± 13%       357%       0.28 ± 23%  sched_debug.rt_rq:/.rt_time.avg
      0.37 ± 10%       240%       1.25 ± 15%  sched_debug.rt_rq:/.rt_time.stddev
      2.54 ± 10%       160%       6.59 ± 10%  sched_debug.rt_rq:/.rt_time.max
      0.32 ± 19%        29%       0.42 ± 10%  perf-stat.dTLB-load-miss-rate
    964727               7%    1028830        perf-stat.context-switches
    176406               4%     184289        perf-stat.cpu-migrations
      0.29               4%       0.30        perf-stat.branch-miss-rate
 1.634e+09                   1.673e+09        perf-stat.node-store-misses
     23.60                       23.99        perf-stat.node-store-miss-rate
     40.01                       40.57        perf-stat.cache-miss-rate
      0.95              -8%       0.87        perf-stat.ipc
 3.203e+12              -9%  2.928e+12        perf-stat.cpu-cycles
 1.506e+09             -11%  1.345e+09        perf-stat.branch-misses
     50.64 ± 13%       -14%      43.45 ±  4%  perf-stat.iTLB-load-miss-rate
 5.285e+11             -14%  4.523e+11        perf-stat.branch-instructions
 3.042e+12             -16%  2.551e+12        perf-stat.instructions
 7.996e+11             -18%  6.584e+11        perf-stat.dTLB-loads
 5.569e+11 ±  4%       -18%  4.578e+11        perf-stat.dTLB-stores


Here are the detailed numbers for the slowed down case:

aim7/1BRD_48G-xfs-disk_rr-3000-performance/ivb44

99091700659f4df9  bf4dc6e4ecc2a3d042029319bc
----------------  --------------------------
         %stddev      change         %stddev
             \          |                \
    360451             -17%     299377        aim7.jobs-per-min
     12806             481%      74447        aim7.time.involuntary_context_switches
       755              44%       1086        aim7.time.system_time
     50.17              20%      60.36        aim7.time.elapsed_time
     50.17              20%      60.36        aim7.time.elapsed_time.max
    438148                      446012        aim7.time.voluntary_context_switches
     37798 ± 16%       780%     332583 ±  8%  interrupts.CAL:Function_call_interrupts
     78.82 ±  5%        18%      93.35 ±  5%  uptime.boot
      2847 ±  7%        11%       3160 ±  7%  uptime.idle
    147490 ±  8%        34%     197261 ±  3%  softirqs.RCU
    648159              29%     839283        softirqs.TIMER
    160830              10%     177144        softirqs.SCHED
   3845352 ±  4%        91%    7349133        numa-numastat.node0.numa_miss
   4686838 ±  5%        67%    7835640        numa-numastat.node0.numa_foreign
   3848455 ±  4%        91%    7352436        numa-numastat.node1.numa_foreign
   4689920 ±  5%        67%    7838734        numa-numastat.node1.numa_miss
     50.17              20%      60.36        time.elapsed_time.max
     12806             481%      74447        time.involuntary_context_switches
       755              44%       1086        time.system_time
     50.17              20%      60.36        time.elapsed_time
      1563              18%       1846        time.percent_of_cpu_this_job_got
     11699 ± 19%      3738%     449048        vmstat.io.bo
  18836969             -16%   15789996        vmstat.memory.free
        16              19%         19        vmstat.procs.r
     19377             459%     108364        vmstat.system.cs
     48255              11%      53537        vmstat.system.in
   2357299              25%    2951384        meminfo.Inactive(file)
   2366381              25%    2960468        meminfo.Inactive
   1575292              -9%    1429971        meminfo.Cached
  19342499             -17%   16100340        meminfo.MemFree
   1057904             -20%     842987        meminfo.Dirty
      1057              21%       1284        turbostat.Avg_MHz
     35.78              21%      43.24        turbostat.%Busy
      9.95              15%      11.47        turbostat.RAMWatt
        74 ±  5%        10%         81        turbostat.CoreTmp
        74 ±  4%        10%         81        turbostat.PkgTmp
       118               8%        128        turbostat.CorWatt
       151               7%        162        turbostat.PkgWatt
     29.06             -23%      22.39        turbostat.CPU%c6
       487 ± 89%      3e+04      26448 ± 57%  latency_stats.max.down.xfs_buf_lock._xfs_buf_find.xfs_buf_get_map.xfs_buf_read_map.xfs_trans_read_buf_map.xfs_read_agf.xfs_alloc_read_agf.xfs_alloc_fix_freelist.xfs_free_extent_fix_freelist.xfs_free_extent.xfs_trans_free_extent
      1823 ± 82%      2e+06    1913796 ± 38%  latency_stats.sum.down.xfs_buf_lock._xfs_buf_find.xfs_buf_get_map.xfs_buf_read_map.xfs_trans_read_buf_map.xfs_read_agf.xfs_alloc_read_agf.xfs_alloc_fix_freelist.xfs_free_extent_fix_freelist.xfs_free_extent.xfs_trans_free_extent
    208475 ± 43%      1e+06    1409494 ±  5%  latency_stats.sum.wait_on_page_bit.truncate_inode_pages_range.truncate_inode_pages_final.evict.iput.dentry_unlink_inode.__dentry_kill.dput.__fput.____fput.task_work_run.exit_to_usermode_loop
      6884 ± 73%      8e+04      90790 ±  9%  latency_stats.sum.call_rwsem_down_read_failed.xfs_log_commit_cil.__xfs_trans_commit.xfs_trans_commit.xfs_vn_update_time.file_update_time.xfs_file_aio_write_checks.xfs_file_buffered_aio_write.xfs_file_write_iter.__vfs_write.vfs_write.SyS_write
      1598 ± 20%      3e+04      35015 ± 27%  latency_stats.sum.call_rwsem_down_read_failed.xfs_log_commit_cil.__xfs_trans_commit.__xfs_trans_roll.xfs_trans_roll.xfs_itruncate_extents.xfs_free_eofblocks.xfs_release.xfs_file_release.__fput.____fput.task_work_run
      2006 ± 25%      3e+04      31143 ± 35%  latency_stats.sum.call_rwsem_down_read_failed.xfs_log_commit_cil.__xfs_trans_commit.__xfs_trans_roll.xfs_trans_roll.xfs_itruncate_extents.xfs_inactive_truncate.xfs_inactive.xfs_fs_destroy_inode.destroy_inode.evict.iput
        29 ±101%      1e+04      10214 ± 29%  latency_stats.sum.call_rwsem_down_read_failed.xfs_log_commit_cil.__xfs_trans_commit.__xfs_trans_roll.xfs_trans_roll.xfs_defer_trans_roll.xfs_defer_finish.xfs_itruncate_extents.xfs_inactive_truncate.xfs_inactive.xfs_fs_destroy_inode.destroy_inode
      1206 ± 51%      9e+03       9919 ± 25%  latency_stats.sum.call_rwsem_down_read_failed.xfs_log_commit_cil.__xfs_trans_commit.xfs_trans_commit.xfs_vn_update_time.touch_atime.generic_file_read_iter.xfs_file_buffered_aio_read.xfs_file_read_iter.__vfs_read.vfs_read.SyS_read
  29869205 ±  4%       -10%   26804569        cpuidle.C1-IVT.time
   5737726              39%    7952214        cpuidle.C1E-IVT.time
     51141              17%      59958        cpuidle.C1E-IVT.usage
  18377551              37%   25176426        cpuidle.C3-IVT.time
     96067              17%     112045        cpuidle.C3-IVT.usage
   1806811              12%    2024041        cpuidle.C6-IVT.usage
   1104420 ± 36%       204%    3361085 ± 27%  cpuidle.POLL.time
       281 ± 10%        20%        338        cpuidle.POLL.usage
      5.61 ± 11%       -0.5       5.12 ± 18%  perf-profile.cycles-pp.irq_exit.smp_apic_timer_interrupt.apic_timer_interrupt.cpuidle_enter.call_cpuidle
      5.85 ±  6%       -0.8       5.06 ± 15%  perf-profile.cycles-pp.hrtimer_interrupt.local_apic_timer_interrupt.smp_apic_timer_interrupt.apic_timer_interrupt.cpuidle_enter
      6.32 ±  6%       -0.9       5.42 ± 15%  perf-profile.cycles-pp.local_apic_timer_interrupt.smp_apic_timer_interrupt.apic_timer_interrupt.cpuidle_enter.call_cpuidle
     15.77 ±  8%         -2      13.83 ± 17%  perf-profile.cycles-pp.smp_apic_timer_interrupt.apic_timer_interrupt.cpuidle_enter.call_cpuidle.cpu_startup_entry
     16.04 ±  8%         -2      14.01 ± 15%  perf-profile.cycles-pp.apic_timer_interrupt.cpuidle_enter.call_cpuidle.cpu_startup_entry.start_secondary
     60.25 ±  4%         -7      53.03 ±  7%  perf-profile.cycles-pp.intel_idle.cpuidle_enter_state.cpuidle_enter.call_cpuidle.cpu_startup_entry
     60.41 ±  4%         -7      53.12 ±  7%  perf-profile.func.cycles-pp.intel_idle
   1174104              22%    1436859        numa-meminfo.node0.Inactive
   1167471              22%    1428271        numa-meminfo.node0.Inactive(file)
    770811              -9%     698147        numa-meminfo.node0.FilePages
  20707294             -12%   18281509 ±  6%  numa-meminfo.node0.Active
  20613745             -12%   18180987 ±  6%  numa-meminfo.node0.Active(file)
   9676639             -17%    8003627        numa-meminfo.node0.MemFree
    509906             -22%     396192        numa-meminfo.node0.Dirty
   1189539              28%    1524697        numa-meminfo.node1.Inactive(file)
   1191989              28%    1525194        numa-meminfo.node1.Inactive
    804508             -10%     727067        numa-meminfo.node1.FilePages
   9654540             -16%    8077810        numa-meminfo.node1.MemFree
    547956             -19%     441933        numa-meminfo.node1.Dirty
       396 ± 12%       485%       2320 ± 37%  slabinfo.bio-1.num_objs
       396 ± 12%       481%       2303 ± 37%  slabinfo.bio-1.active_objs
        73             140%        176 ± 14%  slabinfo.kmalloc-128.active_slabs
        73             140%        176 ± 14%  slabinfo.kmalloc-128.num_slabs
      4734              94%       9171 ± 11%  slabinfo.kmalloc-128.num_objs
      4734              88%       8917 ± 13%  slabinfo.kmalloc-128.active_objs
     16238             -10%      14552 ±  3%  slabinfo.kmalloc-256.active_objs
     17189             -13%      15033 ±  3%  slabinfo.kmalloc-256.num_objs
     20651              96%      40387 ± 17%  slabinfo.radix_tree_node.active_objs
       398              91%        761 ± 17%  slabinfo.radix_tree_node.active_slabs
       398              91%        761 ± 17%  slabinfo.radix_tree_node.num_slabs
     22313              91%      42650 ± 17%  slabinfo.radix_tree_node.num_objs
        32             638%        236 ± 28%  slabinfo.xfs_efd_item.active_slabs
        32             638%        236 ± 28%  slabinfo.xfs_efd_item.num_slabs
      1295             281%       4934 ± 23%  slabinfo.xfs_efd_item.num_objs
      1295             280%       4923 ± 23%  slabinfo.xfs_efd_item.active_objs
      1661              81%       3000 ± 42%  slabinfo.xfs_log_ticket.num_objs
      1661              78%       2952 ± 42%  slabinfo.xfs_log_ticket.active_objs
      2617              49%       3905 ± 30%  slabinfo.xfs_trans.num_objs
      2617              48%       3870 ± 31%  slabinfo.xfs_trans.active_objs
   1015933             567%    6779099        perf-stat.context-switches
 4.864e+08             126%  1.101e+09        perf-stat.node-load-misses
 1.179e+09             103%  2.399e+09        perf-stat.node-loads
      0.06 ± 34%        92%       0.12 ± 11%  perf-stat.dTLB-store-miss-rate
 2.985e+08 ± 32%        86%  5.542e+08 ± 11%  perf-stat.dTLB-store-misses
 2.551e+09 ± 15%        81%  4.625e+09 ± 13%  perf-stat.dTLB-load-misses
      0.39 ± 14%        66%       0.65 ± 13%  perf-stat.dTLB-load-miss-rate
  1.26e+09              60%  2.019e+09        perf-stat.node-store-misses
  46072661 ± 27%        49%   68472915        perf-stat.iTLB-loads
 2.738e+12 ±  4%        43%  3.916e+12        perf-stat.cpu-cycles
     21.48              32%      28.35        perf-stat.node-store-miss-rate
 1.612e+10 ±  3%        28%  2.066e+10        perf-stat.cache-references
 1.669e+09 ±  3%        24%  2.063e+09        perf-stat.branch-misses
 6.816e+09 ±  3%        20%  8.179e+09        perf-stat.cache-misses
    177699              18%     209145        perf-stat.cpu-migrations
      0.39              13%       0.44        perf-stat.branch-miss-rate
 4.606e+09              11%  5.102e+09        perf-stat.node-stores
 4.329e+11 ±  4%         9%  4.727e+11        perf-stat.branch-instructions
 6.458e+11               9%  7.046e+11        perf-stat.dTLB-loads
     29.19               8%      31.45        perf-stat.node-load-miss-rate
    286173               8%     308115        perf-stat.page-faults
    286191               8%     308109        perf-stat.minor-faults
  45084934               4%   47073719        perf-stat.iTLB-load-misses
     42.28              -6%      39.58        perf-stat.cache-miss-rate
     50.62 ± 16%       -19%      40.75        perf-stat.iTLB-load-miss-rate
      0.89             -28%       0.64        perf-stat.ipc
         2 ± 36%     4e+07%     970191        proc-vmstat.pgrotated
       150 ± 21%     1e+07%   15356485 ±  3%  proc-vmstat.nr_vmscan_immediate_reclaim
     76823 ± 35%     56899%   43788651        proc-vmstat.pgscan_direct
    153407 ± 19%      4483%    7031431        proc-vmstat.nr_written
    619699 ± 19%      4441%   28139689        proc-vmstat.pgpgout
   5342421            1061%   62050709        proc-vmstat.pgactivate
        47 ± 25%       354%        217        proc-vmstat.nr_pages_scanned
   8542963 ±  3%        78%   15182914        proc-vmstat.numa_miss
   8542963 ±  3%        78%   15182715        proc-vmstat.numa_foreign
   2820568              31%    3699073        proc-vmstat.pgalloc_dma32
    589234              25%     738160        proc-vmstat.nr_zone_inactive_file
    589240              25%     738155        proc-vmstat.nr_inactive_file
  61347830              13%   69522958        proc-vmstat.pgfree
    393711              -9%     356981        proc-vmstat.nr_file_pages
   4831749             -17%    4020131        proc-vmstat.nr_free_pages
  61252784             -18%   50183773        proc-vmstat.pgrefill
  61245420             -18%   50176301        proc-vmstat.pgdeactivate
    264397             -20%     210222        proc-vmstat.nr_zone_write_pending
    264367             -20%     210188        proc-vmstat.nr_dirty
  60420248             -39%   36646178        proc-vmstat.pgscan_kswapd
  60373976             -44%   33735064        proc-vmstat.pgsteal_kswapd
      1753             -98%         43 ± 18%  proc-vmstat.pageoutrun
      1095             -98%         25 ± 17%  proc-vmstat.kswapd_low_wmark_hit_quickly
       656 ±  3%       -98%         15 ± 24%  proc-vmstat.kswapd_high_wmark_hit_quickly
         0                     1136221        numa-vmstat.node0.workingset_refault
         0                     1136221        numa-vmstat.node0.workingset_activate
        23 ± 45%     1e+07%    2756907        numa-vmstat.node0.nr_vmscan_immediate_reclaim
     37618 ± 24%      3234%    1254165        numa-vmstat.node0.nr_written
   1346538 ±  4%       104%    2748439        numa-vmstat.node0.numa_miss
   1577620 ±  5%        80%    2842882        numa-vmstat.node0.numa_foreign
    291242              23%     357407        numa-vmstat.node0.nr_inactive_file
    291237              23%     357390        numa-vmstat.node0.nr_zone_inactive_file
  13961935              12%   15577331        numa-vmstat.node0.numa_local
  13961938              12%   15577332        numa-vmstat.node0.numa_hit
     39831              10%      43768        numa-vmstat.node0.nr_unevictable
     39831              10%      43768        numa-vmstat.node0.nr_zone_unevictable
    193467             -10%     174639        numa-vmstat.node0.nr_file_pages
   5147212             -12%    4542321 ±  6%  numa-vmstat.node0.nr_active_file
   5147237             -12%    4542325 ±  6%  numa-vmstat.node0.nr_zone_active_file
   2426129             -17%    2008637        numa-vmstat.node0.nr_free_pages
    128285             -23%      99206        numa-vmstat.node0.nr_zone_write_pending
    128259             -23%      99183        numa-vmstat.node0.nr_dirty
         0                     1190594        numa-vmstat.node1.workingset_refault
         0                     1190594        numa-vmstat.node1.workingset_activate
        21 ± 36%     1e+07%    3120425 ±  4%  numa-vmstat.node1.nr_vmscan_immediate_reclaim
     38541 ± 26%      3336%    1324185        numa-vmstat.node1.nr_written
   1316819 ±  4%       105%    2699075        numa-vmstat.node1.numa_foreign
   1547929 ±  4%        80%    2793491        numa-vmstat.node1.numa_miss
    296714              28%     381124        numa-vmstat.node1.nr_zone_inactive_file
    296714              28%     381123        numa-vmstat.node1.nr_inactive_file
  14311131              10%   15750908        numa-vmstat.node1.numa_hit
  14311130              10%   15750905        numa-vmstat.node1.numa_local
    201164             -10%     181742        numa-vmstat.node1.nr_file_pages
   2422825             -16%    2027750        numa-vmstat.node1.nr_free_pages
    137069             -19%     110501        numa-vmstat.node1.nr_zone_write_pending
    137069             -19%     110497        numa-vmstat.node1.nr_dirty
       737 ± 29%     27349%     202387        sched_debug.cfs_rq:/.min_vruntime.min
      3637 ± 20%      7919%     291675        sched_debug.cfs_rq:/.min_vruntime.avg
     11.00 ± 44%      4892%     549.17 ±  9%  sched_debug.cfs_rq:/.runnable_load_avg.max
      2.12 ± 36%      4853%     105.12 ±  5%  sched_debug.cfs_rq:/.runnable_load_avg.stddev
      1885 ±  6%      4189%      80870        sched_debug.cfs_rq:/.min_vruntime.stddev
      1896 ±  6%      4166%      80895        sched_debug.cfs_rq:/.spread0.stddev
     10774 ± 13%      4113%     453925        sched_debug.cfs_rq:/.min_vruntime.max
      1.02 ± 19%      2630%      27.72 ±  7%  sched_debug.cfs_rq:/.runnable_load_avg.avg
     63060 ± 45%       776%     552157        sched_debug.cfs_rq:/.load.max
     14442 ± 21%       590%      99615 ± 14%  sched_debug.cfs_rq:/.load.stddev
      8397 ±  9%       309%      34370 ± 12%  sched_debug.cfs_rq:/.load.avg
     46.02 ± 24%       176%     126.96 ±  6%  sched_debug.cfs_rq:/.util_avg.stddev
       817              19%        974 ±  3%  sched_debug.cfs_rq:/.util_avg.max
       721             -17%        600 ±  3%  sched_debug.cfs_rq:/.util_avg.avg
       595 ± 11%       -38%        371 ±  7%  sched_debug.cfs_rq:/.util_avg.min
      1484 ± 20%       -47%        792 ±  5%  sched_debug.cfs_rq:/.load_avg.min
      1798 ±  4%       -50%        903 ±  5%  sched_debug.cfs_rq:/.load_avg.avg
       322 ±  8%      7726%      25239 ±  8%  sched_debug.cpu.nr_switches.min
       969            7238%      71158        sched_debug.cpu.nr_switches.avg
      2.23 ± 40%      4650%     106.14 ±  4%  sched_debug.cpu.cpu_load[0].stddev
       943 ±  4%      3475%      33730 ±  3%  sched_debug.cpu.nr_switches.stddev
      0.87 ± 25%      3057%      27.46 ±  7%  sched_debug.cpu.cpu_load[0].avg
      5.43 ± 13%      2232%     126.61        sched_debug.cpu.nr_uninterruptible.stddev
      6131 ±  3%      2028%     130453        sched_debug.cpu.nr_switches.max
      1.58 ± 29%      1852%      30.90 ±  4%  sched_debug.cpu.cpu_load[4].avg
      2.00 ± 49%      1422%      30.44 ±  5%  sched_debug.cpu.cpu_load[3].avg
     63060 ± 45%      1053%     726920 ± 32%  sched_debug.cpu.load.max
     21.25 ± 44%       777%     186.33 ±  7%  sched_debug.cpu.nr_uninterruptible.max
     14419 ± 21%       731%     119865 ± 31%  sched_debug.cpu.load.stddev
      3586             381%      17262        sched_debug.cpu.nr_load_updates.min
      8286 ±  8%       364%      38414 ± 17%  sched_debug.cpu.load.avg
      5444             303%      21956        sched_debug.cpu.nr_load_updates.avg
      1156             231%       3827        sched_debug.cpu.nr_load_updates.stddev
      8603 ±  4%       222%      27662        sched_debug.cpu.nr_load_updates.max
      1410             165%       3735        sched_debug.cpu.curr->pid.max
     28742 ± 15%       120%      63101 ±  7%  sched_debug.cpu.clock.min
     28742 ± 15%       120%      63101 ±  7%  sched_debug.cpu.clock_task.min
     28748 ± 15%       120%      63107 ±  7%  sched_debug.cpu.clock.avg
     28748 ± 15%       120%      63107 ±  7%  sched_debug.cpu.clock_task.avg
     28751 ± 15%       120%      63113 ±  7%  sched_debug.cpu.clock.max
     28751 ± 15%       120%      63113 ±  7%  sched_debug.cpu.clock_task.max
       442 ± 11%        93%        854 ± 15%  sched_debug.cpu.curr->pid.avg
       618 ±  3%        72%       1065 ±  4%  sched_debug.cpu.curr->pid.stddev
      1.88 ± 11%        50%       2.83 ±  8%  sched_debug.cpu.clock.stddev
      1.88 ± 11%        50%       2.83 ±  8%  sched_debug.cpu.clock_task.stddev
      5.22 ±  9%       -55%       2.34 ± 23%  sched_debug.rt_rq:/.rt_time.max
      0.85             -55%       0.38 ± 28%  sched_debug.rt_rq:/.rt_time.stddev
      0.17             -56%       0.07 ± 33%  sched_debug.rt_rq:/.rt_time.avg
     27633 ± 16%       124%      61980 ±  8%  sched_debug.ktime
     28745 ± 15%       120%      63102 ±  7%  sched_debug.sched_clk
     28745 ± 15%       120%      63102 ±  7%  sched_debug.cpu_clk

Thanks,
Fengguang

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ