lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Date:   Mon, 13 Apr 2020 18:51:24 +0800
From:   Feng Tang <feng.tang@...el.com>
To:     Chao Yu <yuchao0@...wei.com>,
        kernel test robot <rong.a.chen@...el.com>
Cc:     Jaegeuk Kim <jaegeuk@...nel.org>,
        LKML <linux-kernel@...r.kernel.org>,
        Linus Torvalds <torvalds@...ux-foundation.org>,
        lkp@...ts.01.org
Subject: Re: [LKP] [f2fs] fe1897eaa6:  aim7.jobs-per-min -60.9% regression

On Sun, Dec 08, 2019 at 11:41:46PM +0800, kernel test robot wrote:
> Greeting,
> 
> FYI, we noticed a -60.9% regression of aim7.jobs-per-min due to commit:
> 
> 
> commit: fe1897eaa6646f5a64a4cee0e6473ed9887d324b ("f2fs: fix to update time in lazytime mode")
> https://git.kernel.org/cgit/linux/kernel/git/torvalds/linux.git master
> 
> in testcase: aim7
> on test machine: 48 threads Intel(R) Xeon(R) CPU E5-2697 v2 @ 2.70GHz with 64G memory
> with following parameters:
> 
> 	disk: 4BRD_12G
> 	md: RAID1
> 	fs: f2fs
> 	test: sync_disk_rw
> 	load: 200
> 	cpufreq_governor: performance
> 	ucode: 0x42e
> 
> test-description: AIM7 is a traditional UNIX system level benchmark suite which is used to test and measure the performance of multiuser system.
> test-url: https://sourceforge.net/projects/aimbench/files/aim-suite7/
> 
> 
> 
> If you fix the issue, kindly add following tag
> Reported-by: kernel test robot <rong.a.chen@...el.com>
> 
> 
> Details are as below:
> -------------------------------------------------------------------------------------------------->
> 
> 
> To reproduce:
> 
>         git clone https://github.com/intel/lkp-tests.git
>         cd lkp-tests
>         bin/lkp install job.yaml  # job file is attached in this email
>         bin/lkp run     job.yaml
> 
> =========================================================================================
> compiler/cpufreq_governor/disk/fs/kconfig/load/md/rootfs/tbox_group/test/testcase/ucode:
>   gcc-7/performance/4BRD_12G/f2fs/x86_64-rhel-7.6/200/RAID1/debian-x86_64-2019-11-14.cgz/lkp-ivb-2ep1/sync_disk_rw/aim7/0x42e
> 
> commit: 
>   b145b0eb20 (" ARM and x86 bugfixes of all kinds.  The most visible one is that migrating")
>   fe1897eaa6 ("f2fs: fix to update time in lazytime mode")
> 
> b145b0eb2031a620 fe1897eaa6646f5a64a4cee0e64 
> ---------------- --------------------------- 
>        fail:runs  %reproduction    fail:runs
>            |             |             |    
>           1:4          -25%            :4     dmesg.WARNING:at#for_ip_interrupt_entry/0x
>          %stddev     %change         %stddev
>              \          |                \  
>       1208           -60.9%     473.12        aim7.jobs-per-min

Hi Chao Yu,

Any thoughts on this? 

IIUC, this commit is a fix, and if you think it is necessary, we will
stop tracking this regresion.

Thanks,
Feng


>     992.70          +155.6%       2537        aim7.time.elapsed_time
>     992.70          +155.6%       2537        aim7.time.elapsed_time.max
>  1.078e+08           +60.0%  1.724e+08        aim7.time.file_system_outputs
>    9646301            -4.2%    9238717        aim7.time.involuntary_context_switches
>      24107 ±  3%     +58.2%      38130 ± 17%  aim7.time.minor_page_faults
>       3295          +202.3%       9963        aim7.time.system_time
>      26.75           +51.4%      40.49        aim7.time.user_time
>   5.02e+08          +210.5%  1.559e+09        aim7.time.voluntary_context_switches
>    9294452 ±  4%     -15.7%    7835497 ± 16%  numa-numastat.node0.local_node
>    9298417 ±  4%     -15.7%    7840991 ± 16%  numa-numastat.node0.numa_hit
>       3.66 ±  3%      +4.2        7.82 ± 18%  mpstat.cpu.all.idle%
>       0.01 ± 32%      -0.0        0.00 ± 31%  mpstat.cpu.all.soft%
>       0.06            -0.0        0.04        mpstat.cpu.all.usr%
>       3.85 ±  3%    +105.0%       7.89 ± 17%  iostat.cpu.idle
>      89.80            -4.9%      85.36        iostat.cpu.iowait
>       6.29            +6.6%       6.70        iostat.cpu.system
>      23968           -48.4%      12356        iostat.md0.w/s
>      53780           -37.1%      33824        iostat.md0.wkB/s
>      89.00            -4.8%      84.75        vmstat.cpu.wa
>      53779           -37.1%      33833        vmstat.io.bo
>   29311046           -28.6%   20935254        vmstat.memory.free
>    1026388           +19.4%    1225080        vmstat.system.cs
>     103289            +8.5%     112038        vmstat.system.in
>      24013 ± 29%     -58.5%       9976 ±120%  numa-meminfo.node0.Dirty
>     119019 ± 10%     +44.6%     172139 ± 13%  numa-meminfo.node1.Active
>       3503 ± 56%    +464.8%      19786 ± 38%  numa-meminfo.node1.Active(file)
>      40477 ± 16%     -38.5%      24896 ± 47%  numa-meminfo.node1.Dirty
>   19966008 ±  2%     -29.5%   14084594 ± 44%  numa-meminfo.node1.MemFree
>   13014643 ±  4%     +45.2%   18896052 ± 32%  numa-meminfo.node1.MemUsed
>      12018 ± 41%    +184.8%      34228 ± 35%  numa-meminfo.node1.Shmem
>   2.09e+10          +127.4%  4.752e+10 ±  4%  cpuidle.C1.time
>  2.558e+08          +140.1%  6.143e+08 ±  4%  cpuidle.C1.usage
>  3.017e+09 ± 10%    +301.2%  1.211e+10 ±  9%  cpuidle.C1E.time
>   18225961 ±  9%    +430.1%   96624314 ±  6%  cpuidle.C1E.usage
>  4.627e+08 ± 14%    +147.5%  1.145e+09 ± 20%  cpuidle.C6.time
>     802166 ± 17%    +186.9%    2301023 ± 26%  cpuidle.C6.usage
>  1.799e+10          +148.9%  4.476e+10        cpuidle.POLL.time
>  2.616e+08 ±  2%    +246.4%  9.063e+08 ±  2%  cpuidle.POLL.usage
>       6017 ± 29%     -58.5%       2496 ±120%  numa-vmstat.node0.nr_dirty
>       5742 ± 29%     -65.0%       2011 ±120%  numa-vmstat.node0.nr_zone_write_pending
>     874.50 ± 56%    +465.6%       4946 ± 38%  numa-vmstat.node1.nr_active_file
>      10139 ± 16%     -38.6%       6224 ± 47%  numa-vmstat.node1.nr_dirty
>    4992144 ±  2%     -29.5%    3521200 ± 44%  numa-vmstat.node1.nr_free_pages
>       3003 ± 41%    +184.9%       8556 ± 35%  numa-vmstat.node1.nr_shmem
>     874.50 ± 56%    +465.6%       4946 ± 38%  numa-vmstat.node1.nr_zone_active_file
>       9775 ± 16%     -47.9%       5089 ± 47%  numa-vmstat.node1.nr_zone_write_pending
>  2.558e+08          +140.1%  6.143e+08 ±  4%  turbostat.C1
>      43.77            -4.8       38.98 ±  3%  turbostat.C1%
>   18225642 ±  9%    +430.2%   96624037 ±  6%  turbostat.C1E
>       6.31 ± 10%      +3.6        9.94 ±  9%  turbostat.C1E%
>     795117 ± 17%    +188.4%    2292823 ± 26%  turbostat.C6
>       0.38 ± 42%     -63.6%       0.14 ± 13%  turbostat.CPU%c6
>  1.028e+08          +176.7%  2.844e+08        turbostat.IRQ
>       0.23 ± 33%     -69.1%       0.07 ± 22%  turbostat.Pkg%pc2
>     275064           +12.3%     308846 ±  4%  meminfo.Active
>       8036 ± 21%    +261.7%      29072 ± 52%  meminfo.Active(file)
>      64400           -45.9%      34850        meminfo.Dirty
>      16481           +10.8%      18264        meminfo.Inactive(anon)
>     147067           +12.8%     165858 ±  2%  meminfo.KReclaimable
>   29120417           -28.8%   20740075        meminfo.MemAvailable
>   29335165           -28.6%   20936804        meminfo.MemFree
>   36511133           +23.0%   44909491        meminfo.Memused
>     147067           +12.8%     165858 ±  2%  meminfo.SReclaimable
>      36567 ±  6%     +37.2%      50173        meminfo.Shmem
>      51580           -60.8%      20220        meminfo.max_used_kB
>       4.27 ±  7%     -13.9%       3.67 ±  4%  perf-stat.i.MPKI
>       0.71 ±  3%      -0.0        0.67        perf-stat.i.branch-miss-rate%
>      17.93            +3.5       21.41 ±  2%  perf-stat.i.cache-miss-rate%
>    1029337           +19.1%    1226025        perf-stat.i.context-switches
>      11421 ±  4%     +41.1%      16120 ±  6%  perf-stat.i.cpu-migrations
>       0.42 ±  2%      +0.0        0.45 ±  6%  perf-stat.i.dTLB-load-miss-rate%
>       0.11 ± 27%      -0.0        0.08 ±  3%  perf-stat.i.dTLB-store-miss-rate%
>      84.04 ±  2%      -2.3       81.75 ±  2%  perf-stat.i.iTLB-load-miss-rate%
>    7375571            +7.3%    7917282        perf-stat.i.iTLB-load-misses
>       3.91 ±  7%      -9.2%       3.56 ±  4%  perf-stat.overall.MPKI
>      17.90            +4.1       21.96 ±  3%  perf-stat.overall.cache-miss-rate%
>       0.41 ±  2%      +0.0        0.45 ±  6%  perf-stat.overall.dTLB-load-miss-rate%
>       0.10 ± 27%      -0.0        0.08 ±  3%  perf-stat.overall.dTLB-store-miss-rate%
>      83.99 ±  2%      -2.3       81.72 ±  2%  perf-stat.overall.iTLB-load-miss-rate%
>    1028292           +19.2%    1225532        perf-stat.ps.context-switches
>      11410 ±  4%     +41.2%      16114 ±  6%  perf-stat.ps.cpu-migrations
>    7368087            +7.4%    7914119        perf-stat.ps.iTLB-load-misses
>  2.558e+13 ±  2%    +157.0%  6.575e+13 ±  3%  perf-stat.total.instructions
>       9128 ±  2%     -25.3%       6821 ±  4%  slabinfo.dmaengine-unmap-16.active_objs
>       9175 ±  2%     -24.4%       6938 ±  4%  slabinfo.dmaengine-unmap-16.num_objs
>      28309           +12.5%      31840        slabinfo.f2fs_extent_node.active_objs
>      28704           +13.0%      32429        slabinfo.f2fs_extent_node.num_objs
>     964.00 ±  5%     -13.3%     836.00 ±  6%  slabinfo.kmem_cache_node.active_objs
>       1008 ±  5%     -12.7%     880.00 ±  6%  slabinfo.kmem_cache_node.num_objs
>      13017           -33.1%       8707        slabinfo.nat_entry.active_objs
>      13036           -33.2%       8707        slabinfo.nat_entry.num_objs
>     433.75 ±  9%     +32.4%     574.50 ±  2%  slabinfo.nfs_read_data.active_objs
>     433.75 ±  9%     +32.4%     574.50 ±  2%  slabinfo.nfs_read_data.num_objs
>     697.50 ±  3%     +48.6%       1036 ± 33%  slabinfo.numa_policy.active_objs
>     697.50 ±  3%     +48.6%       1036 ± 33%  slabinfo.numa_policy.num_objs
>     153098           +22.3%     187296        slabinfo.radix_tree_node.active_objs
>       2733           +22.3%       3344        slabinfo.radix_tree_node.active_slabs
>     153098           +22.3%     187296        slabinfo.radix_tree_node.num_objs
>       2733           +22.3%       3344        slabinfo.radix_tree_node.num_slabs
>     542.00 ± 10%     +29.5%     702.00 ±  6%  slabinfo.xfrm_state.active_objs
>     542.00 ± 10%     +29.5%     702.00 ±  6%  slabinfo.xfrm_state.num_objs
>      66757            +4.8%      69942        proc-vmstat.nr_active_anon
>       2009 ± 21%    +261.7%       7267 ± 52%  proc-vmstat.nr_active_file
>      96.00            +3.9%      99.75        proc-vmstat.nr_anon_transparent_hugepages
>   13466177           +60.0%   21542226        proc-vmstat.nr_dirtied
>      16154           -46.0%       8719        proc-vmstat.nr_dirty
>     726643           -28.8%     517541        proc-vmstat.nr_dirty_background_threshold
>    1455064           -28.8%    1036348        proc-vmstat.nr_dirty_threshold
>     311402            +2.4%     318957        proc-vmstat.nr_file_pages
>    7332853           -28.6%    5234441        proc-vmstat.nr_free_pages
>       4120           +10.8%       4565        proc-vmstat.nr_inactive_anon
>      13939            -1.2%      13772        proc-vmstat.nr_kernel_stack
>       2845            -1.5%       2803        proc-vmstat.nr_page_table_pages
>       9142 ±  6%     +37.2%      12542        proc-vmstat.nr_shmem
>      36767           +12.8%      41463 ±  2%  proc-vmstat.nr_slab_reclaimable
>   13387539           +60.3%   21463781        proc-vmstat.nr_written
>      66757            +4.8%      69942        proc-vmstat.nr_zone_active_anon
>       2009 ± 21%    +261.7%       7267 ± 52%  proc-vmstat.nr_zone_active_file
>       4120           +10.8%       4565        proc-vmstat.nr_zone_inactive_anon
>      15494           -54.2%       7089        proc-vmstat.nr_zone_write_pending
>     582146 ±102%    +628.1%    4238850 ±  6%  proc-vmstat.numa_foreign
>      13506 ±  3%    +114.5%      28974 ± 31%  proc-vmstat.numa_hint_faults
>       8913 ±  6%    +148.6%      22154 ± 35%  proc-vmstat.numa_hint_faults_local
>     582146 ±102%    +628.1%    4238850 ±  6%  proc-vmstat.numa_miss
>     597885 ± 99%    +611.6%    4254585 ±  6%  proc-vmstat.numa_other
>    1014174 ± 20%     -46.1%     547135 ± 57%  proc-vmstat.pgalloc_dma32
>   16157005           +22.4%   19769138        proc-vmstat.pgalloc_normal
>    2516271          +152.8%    6360808        proc-vmstat.pgfault
>    4866589           +64.4%    8002519        proc-vmstat.pgfree
>   53553588           +60.3%   85856803        proc-vmstat.pgpgout
>       1.55 ±  4%      +0.1        1.65        perf-profile.calltrace.cycles-pp.__schedule.schedule.io_schedule.wait_on_page_bit.f2fs_wait_on_page_writeback
>       1.58 ±  4%      +0.1        1.69        perf-profile.calltrace.cycles-pp.schedule.io_schedule.wait_on_page_bit.f2fs_wait_on_page_writeback.f2fs_wait_on_node_pages_writeback
>       1.60 ±  4%      +0.1        1.71        perf-profile.calltrace.cycles-pp.io_schedule.wait_on_page_bit.f2fs_wait_on_page_writeback.f2fs_wait_on_node_pages_writeback.f2fs_do_sync_file
>       0.75 ±  2%      +0.2        0.91 ±  2%  perf-profile.calltrace.cycles-pp.menu_select.do_idle.cpu_startup_entry.start_secondary.secondary_startup_64
>       0.70 ±  4%      -0.3        0.36 ±  7%  perf-profile.children.cycles-pp.__generic_file_write_iter
>       0.69 ±  4%      -0.3        0.35 ±  7%  perf-profile.children.cycles-pp.generic_perform_write
>       0.57 ±  5%      -0.3        0.25 ±  6%  perf-profile.children.cycles-pp.__filemap_fdatawrite_range
>       0.58 ±  5%      -0.3        0.26 ±  7%  perf-profile.children.cycles-pp.file_write_and_wait_range
>       0.57 ±  6%      -0.3        0.25 ±  6%  perf-profile.children.cycles-pp.do_writepages
>       0.57 ±  5%      -0.3        0.25 ±  6%  perf-profile.children.cycles-pp.f2fs_write_data_pages
>       0.64 ±  5%      -0.3        0.34 ±  6%  perf-profile.children.cycles-pp.f2fs_write_begin
>       0.57 ±  5%      -0.2        0.32 ±  7%  perf-profile.children.cycles-pp.f2fs_put_page
>       0.31 ± 10%      -0.2        0.11 ±  7%  perf-profile.children.cycles-pp.blk_finish_plug
>       0.31 ± 10%      -0.2        0.11 ±  7%  perf-profile.children.cycles-pp.blk_flush_plug_list
>       0.30 ± 11%      -0.2        0.11 ±  6%  perf-profile.children.cycles-pp.raid1_unplug
>       0.26            -0.1        0.14 ± 12%  perf-profile.children.cycles-pp.f2fs_write_cache_pages
>       0.22 ± 20%      -0.1        0.11 ± 41%  perf-profile.children.cycles-pp.brd_do_bvec
>       0.14 ± 12%      -0.1        0.03 ±100%  perf-profile.children.cycles-pp.submit_bio_wait
>       0.21 ± 10%      -0.1        0.11 ± 10%  perf-profile.children.cycles-pp.submit_bio
>       0.23 ± 17%      -0.1        0.13 ± 31%  perf-profile.children.cycles-pp.worker_thread
>       0.20 ±  3%      -0.1        0.11 ± 14%  perf-profile.children.cycles-pp.__write_data_page
>       0.14 ± 15%      -0.1        0.05 ±  9%  perf-profile.children.cycles-pp.__submit_flush_wait
>       0.17 ± 12%      -0.1        0.09 ± 12%  perf-profile.children.cycles-pp.md_make_request
>       0.16 ± 11%      -0.1        0.08 ± 14%  perf-profile.children.cycles-pp.md_handle_request
>       0.15 ±  9%      -0.1        0.07 ± 11%  perf-profile.children.cycles-pp.raid1_make_request
>       0.14 ± 17%      -0.1        0.06 ± 11%  perf-profile.children.cycles-pp.f2fs_issue_flush
>       0.14 ±  6%      -0.1        0.08 ± 19%  perf-profile.children.cycles-pp.f2fs_do_write_data_page
>       0.10 ± 33%      -0.0        0.06 ± 11%  perf-profile.children.cycles-pp.reschedule_interrupt
>       0.09 ± 37%      -0.0        0.05 ±  8%  perf-profile.children.cycles-pp.scheduler_ipi
>       0.06 ±  6%      +0.0        0.08 ± 10%  perf-profile.children.cycles-pp.up_read
>       0.10 ±  4%      +0.0        0.12 ± 13%  perf-profile.children.cycles-pp.pick_next_task_idle
>       0.20 ±  2%      +0.0        0.23 ±  5%  perf-profile.children.cycles-pp.get_next_timer_interrupt
>       0.16 ±  6%      +0.0        0.19 ±  4%  perf-profile.children.cycles-pp.irq_exit
>       0.01 ±173%      +0.1        0.07 ±  7%  perf-profile.children.cycles-pp.put_prev_task_fair
>       0.27 ±  3%      +0.1        0.33 ±  4%  perf-profile.children.cycles-pp.tick_nohz_next_event
>       0.35 ±  3%      +0.1        0.42 ±  3%  perf-profile.children.cycles-pp.tick_nohz_get_sleep_length
>       1.60 ±  4%      +0.1        1.72        perf-profile.children.cycles-pp.io_schedule
>       0.78 ±  2%      +0.1        0.92 ±  2%  perf-profile.children.cycles-pp.menu_select
>       0.06 ±  6%      +0.0        0.08 ± 10%  perf-profile.self.cycles-pp.up_read
>       0.08 ±  5%      +0.0        0.11 ± 11%  perf-profile.self.cycles-pp.cpuidle_enter_state
>       0.35 ±  5%      +0.0        0.39 ±  5%  perf-profile.self.cycles-pp.f2fs_wait_on_node_pages_writeback
>       0.35 ±  4%      +0.1        0.41 ±  3%  perf-profile.self.cycles-pp.menu_select
>       8096 ± 28%    +278.9%      30679 ± 20%  sched_debug.cfs_rq:/.MIN_vruntime.avg
>     158062 ± 17%    +299.5%     631442 ±  8%  sched_debug.cfs_rq:/.MIN_vruntime.max
>      32629 ± 22%    +289.7%     127148 ± 13%  sched_debug.cfs_rq:/.MIN_vruntime.stddev
>      44053          +192.3%     128774        sched_debug.cfs_rq:/.exec_clock.avg
>     140530 ±  9%    +377.8%     671471 ± 29%  sched_debug.cfs_rq:/.exec_clock.max
>      33108 ±  2%    +109.9%      69484 ±  2%  sched_debug.cfs_rq:/.exec_clock.min
>      20291 ±  7%    +386.1%      98642 ± 21%  sched_debug.cfs_rq:/.exec_clock.stddev
>       0.25 ± 82%     -88.4%       0.03 ±131%  sched_debug.cfs_rq:/.load_avg.min
>       8096 ± 28%    +278.9%      30679 ± 20%  sched_debug.cfs_rq:/.max_vruntime.avg
>     158062 ± 17%    +299.5%     631442 ±  8%  sched_debug.cfs_rq:/.max_vruntime.max
>      32629 ± 22%    +289.7%     127148 ± 13%  sched_debug.cfs_rq:/.max_vruntime.stddev
>     184170          +218.6%     586859 ±  2%  sched_debug.cfs_rq:/.min_vruntime.avg
>     251516 ±  4%    +254.5%     891702 ± 11%  sched_debug.cfs_rq:/.min_vruntime.max
>     165614          +131.3%     383025 ±  2%  sched_debug.cfs_rq:/.min_vruntime.min
>      15286 ±  9%   +1057.3%     176903 ± 10%  sched_debug.cfs_rq:/.min_vruntime.stddev
>       0.67 ±  6%     +45.5%       0.98 ±  9%  sched_debug.cfs_rq:/.nr_spread_over.avg
>       0.74 ± 20%     +32.4%       0.98 ± 13%  sched_debug.cfs_rq:/.nr_spread_over.stddev
>       3.43 ± 29%     -60.7%       1.35 ±100%  sched_debug.cfs_rq:/.removed.load_avg.avg
>      75.10 ± 34%     -84.1%      11.91 ±100%  sched_debug.cfs_rq:/.removed.load_avg.max
>      15.24 ± 23%     -75.5%       3.74 ±100%  sched_debug.cfs_rq:/.removed.load_avg.stddev
>     157.58 ± 29%     -60.5%      62.19 ±100%  sched_debug.cfs_rq:/.removed.runnable_sum.avg
>       3459 ± 34%     -84.0%     552.75 ± 99%  sched_debug.cfs_rq:/.removed.runnable_sum.max
>     700.79 ± 23%     -75.4%     172.66 ±100%  sched_debug.cfs_rq:/.removed.runnable_sum.stddev
>      34.81 ± 42%     -87.0%       4.53 ±100%  sched_debug.cfs_rq:/.removed.util_avg.max
>       5.89 ± 31%     -79.0%       1.23 ±100%  sched_debug.cfs_rq:/.removed.util_avg.stddev
>      59642 ± 58%    +506.4%     361657 ± 61%  sched_debug.cfs_rq:/.spread0.max
>      15286 ±  9%   +1057.2%     176904 ± 10%  sched_debug.cfs_rq:/.spread0.stddev
>      15.81 ± 15%     +28.7%      20.34 ±  2%  sched_debug.cfs_rq:/.util_est_enqueued.avg
>     553.04 ± 15%     +42.8%     789.60 ±  2%  sched_debug.cfs_rq:/.util_est_enqueued.max
>      83.37 ± 15%     +37.7%     114.78 ±  2%  sched_debug.cfs_rq:/.util_est_enqueued.stddev
>     130721           -17.9%     107295        sched_debug.cpu.avg_idle.avg
>      32650 ±  7%     -16.9%      27145 ±  6%  sched_debug.cpu.avg_idle.min
>     509212          +151.7%    1281639        sched_debug.cpu.clock.avg
>     509215          +151.7%    1281642        sched_debug.cpu.clock.max
>     509209          +151.7%    1281636        sched_debug.cpu.clock.min
>     509212          +151.7%    1281639        sched_debug.cpu.clock_task.avg
>     509215          +151.7%    1281642        sched_debug.cpu.clock_task.max
>     509209          +151.7%    1281636        sched_debug.cpu.clock_task.min
>     480.69 ±  2%     +89.2%     909.42 ±  3%  sched_debug.cpu.curr->pid.avg
>      14037          +135.4%      33040        sched_debug.cpu.curr->pid.max
>       2146 ±  2%    +125.6%       4843        sched_debug.cpu.curr->pid.stddev
>       0.12 ±  5%     +10.2%       0.13 ±  3%  sched_debug.cpu.nr_running.avg
>   10174129          +208.2%   31360192        sched_debug.cpu.nr_switches.avg
>   10849045          +290.1%   42321541 ±  3%  sched_debug.cpu.nr_switches.max
>     564008 ± 15%   +1719.4%   10261545 ±  8%  sched_debug.cpu.nr_switches.stddev
>     -87.16          +262.4%    -315.83        sched_debug.cpu.nr_uninterruptible.min
>      66.84 ± 14%    +143.1%     162.48 ± 21%  sched_debug.cpu.nr_uninterruptible.stddev
>   10172460          +208.3%   31358619        sched_debug.cpu.sched_count.avg
>   10847113          +290.1%   42319816 ±  3%  sched_debug.cpu.sched_count.max
>     563945 ± 15%   +1719.6%   10261603 ±  8%  sched_debug.cpu.sched_count.stddev
>    4957584          +211.1%   15425344        sched_debug.cpu.sched_goidle.avg
>    5286410          +291.7%   20709361 ±  3%  sched_debug.cpu.sched_goidle.max
>     272603 ± 14%   +1717.5%    4954528 ±  7%  sched_debug.cpu.sched_goidle.stddev
>    5120024          +209.4%   15841794        sched_debug.cpu.ttwu_count.avg
>   58965381 ± 14%    +567.2%  3.934e+08 ± 33%  sched_debug.cpu.ttwu_count.max
>     132694 ±  2%     -41.0%      78274 ±  5%  sched_debug.cpu.ttwu_count.min
>   11266172 ±  8%    +482.6%   65635267 ± 21%  sched_debug.cpu.ttwu_count.stddev
>     114955           +26.1%     144930 ±  3%  sched_debug.cpu.ttwu_local.max
>      83595 ±  4%     -53.0%      39284 ± 30%  sched_debug.cpu.ttwu_local.min
>       5967 ± 16%    +460.9%      33468 ±  9%  sched_debug.cpu.ttwu_local.stddev
>     509209          +151.7%    1281636        sched_debug.cpu_clk
>     506629          +152.5%    1279056        sched_debug.ktime
>     509627          +151.6%    1282053        sched_debug.sched_clk

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ