lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [<thread-prev] [day] [month] [year] [list]
Date:	Thu, 21 Aug 2014 23:00:50 +0800
From:	Fengguang Wu <fengguang.wu@...el.com>
To:	Rik van Riel <riel@...hat.com>
Cc:	Dave Hansen <dave.hansen@...el.com>,
	LKML <linux-kernel@...r.kernel.org>, lkp@...org,
	Ingo Molnar <mingo@...e.hu>
Subject: Re: [sched/fair] caeb178c60f: +252.0% cpuidle.C1-SNB.time, +3.1%
 turbostat.Pkg_W

On Thu, Aug 21, 2014 at 10:16:13AM -0400, Rik van Riel wrote:
> On 08/21/2014 10:01 AM, Fengguang Wu wrote:
> > Hi Rik,
> > 
> > FYI, we noticed the below changes on
> > 
> > git://git.kernel.org/pub/scm/linux/kernel/git/tip/tip.git sched/core
> > commit caeb178c60f4f93f1b45c0bc056b5cf6d217b67f ("sched/fair: Make update_sd_pick_busiest() return 'true' on a busier sd")
> > 
> > testbox/testcase/testparams: lkp-sb03/nepim/300s-100%-tcp6
> 
> Is this good or bad?

It seems mixed results. The throughput is 2.4% better in sequential
write test, while the power consumption (turbostat.Pkg_W) increases
by 3.1% in the nepim/300s-100%-tcp test.

> The numbers suggest the xfs + raid5 workload is doing around 2.4%
> more IO to disk per second with this change in, and there is more

Right.

> CPU idle time in the system...

Sorry "cpuidle" is the monitor name. You can find its code here:

https://git.kernel.org/cgit/linux/kernel/git/wfg/lkp-tests.git/tree/monitors/cpuidle

"cpuidle.C1-SNB.time" means the time spend in C1 state.

> For the tcp test, I see no throughput numbers, but I see more
> idle time as well as more time in turbo mode, and more softirqs,
> which could mean that more packets were handled.

Again, "turbostat" is a monitor name. "turbostat.Pkg_W" means the
CPU package watts reported by the turbostat tool.

> Does the patch introduce any performance issues, or did it
> simply trip up something in the statistics that your script
> noticed?

In normal LKP reports, only changed stats are listed. Here is the
performance/power index comparison, which lists all performance/power
related stats. The index is geometric average of all results. Baseline
is 100 for 743cb1ff191f00f.

   100      perf-index (the larger, the better)
    98     power-index (the larger, the better)

743cb1ff191f00f  caeb178c60f4f93f1b45c0bc0  testbox/testcase/testparams
---------------  -------------------------  ---------------------------
     %stddev        %change               %stddev
            \          |                 /
    691053 ± 4%      -5.1%     656100 ± 4%  lkp-sb03/nepim/300s-100%-tcp
    570185 ± 7%      +5.4%     600774 ± 4%  lkp-sb03/nepim/300s-100%-tcp6
   1261238 ± 5%      -0.3%    1256875 ± 4%  TOTAL nepim.tcp.avg.kbps_in

743cb1ff191f00f  caeb178c60f4f93f1b45c0bc0  
---------------  -------------------------  
    691216 ± 4%      -5.1%     656264 ± 4%  lkp-sb03/nepim/300s-100%-tcp
    570347 ± 7%      +5.4%     600902 ± 4%  lkp-sb03/nepim/300s-100%-tcp6
   1261564 ± 5%      -0.3%    1257167 ± 4%  TOTAL nepim.tcp.avg.kbps_out

743cb1ff191f00f  caeb178c60f4f93f1b45c0bc0  
---------------  -------------------------  
     77.48 ± 1%      +3.1%      79.91 ± 1%  lkp-sb03/nepim/300s-100%-tcp
     79.69 ± 2%      -0.6%      79.21 ± 1%  lkp-sb03/nepim/300s-100%-tcp6
    157.17 ± 2%      +1.2%     159.13 ± 1%  TOTAL turbostat.Pkg_W

743cb1ff191f00f  caeb178c60f4f93f1b45c0bc0  
---------------  -------------------------  
      6.05 ± 1%      +1.2%       6.12 ± 1%  lkp-sb03/nepim/300s-100%-tcp
      6.06 ± 0%      +1.0%       6.12 ± 1%  lkp-sb03/nepim/300s-100%-tcp6
     12.11 ± 1%      +1.1%      12.24 ± 1%  TOTAL turbostat.%c0

743cb1ff191f00f  caeb178c60f4f93f1b45c0bc0                   
---------------  -------------------------                   
    325759 ± 0%      +2.4%     333577 ± 0%  lkp-st02/dd-write/5m-11HDD-RAID5-cfq-xfs-1dd
    325759 ± 0%      +2.4%     333577 ± 0%  TOTAL iostat.md0.wkB/s

The nepim throughput numbers are not stable enough comparing to the
change, so are not regarded as real changes in the original email.
I will need to increase its test time to make it more stable..

Thanks,
Fengguang

> > 743cb1ff191f00f  caeb178c60f4f93f1b45c0bc0
> > ---------------  -------------------------
> >   29718911 ±45%    +329.5%  1.277e+08 ±10%  cpuidle.C1E-SNB.time
> >        861 ±34%   +1590.4%      14564 ±31%  cpuidle.C3-SNB.usage
> >   1.65e+08 ±20%    +175.4%  4.544e+08 ±15%  cpuidle.C1-SNB.time
> >         24 ±41%    +247.6%         86 ±23%  numa-numastat.node1.other_node
> >      27717 ±11%     +98.7%      55085 ± 6%  softirqs.RCU
> >     180767 ±11%     +86.7%     337416 ±10%  cpuidle.C7-SNB.usage
> >     104591 ±14%     +77.4%     185581 ±10%  cpuidle.C1E-SNB.usage
> >        384 ±10%     +33.3%        512 ±11%  slabinfo.kmem_cache.num_objs
> >        384 ±10%     +33.3%        512 ±11%  slabinfo.kmem_cache.active_objs
> >        494 ± 8%     +25.9%        622 ± 9%  slabinfo.kmem_cache_node.active_objs
> >        512 ± 7%     +25.0%        640 ± 8%  slabinfo.kmem_cache_node.num_objs
> >      83427 ± 6%     +10.3%      92028 ± 5%  meminfo.DirectMap4k
> >       9508 ± 1%     +21.3%      11534 ± 7%  slabinfo.kmalloc-512.active_objs
> >       9838 ± 1%     +20.5%      11852 ± 6%  slabinfo.kmalloc-512.num_objs
> >      53997 ± 6%     +11.1%      59981 ± 4%  numa-meminfo.node1.Slab
> >       2662 ± 3%      -9.0%       2424 ± 3%  slabinfo.kmalloc-96.active_objs
> >       2710 ± 3%      -8.6%       2478 ± 3%  slabinfo.kmalloc-96.num_objs
> >        921 ±41%   +3577.7%      33901 ±14%  time.involuntary_context_switches
> >       2371 ± 2%     +15.5%       2739 ± 2%  vmstat.system.in
> > 
> > testbox/testcase/testparams: lkp-sb03/nepim/300s-100%-tcp
> > 
> > 743cb1ff191f00f  caeb178c60f4f93f1b45c0bc0
> > ---------------  -------------------------
> >   20657207 ±31%    +358.2%   94650352 ±18%  cpuidle.C1E-SNB.time
> >   29718911 ±45%    +329.5%  1.277e+08 ±10%  cpuidle.C1E-SNB.time
> >        861 ±34%   +1590.4%      14564 ±31%  cpuidle.C3-SNB.usage
> >       0.05 ±46%    +812.5%       0.44 ±34%  turbostat.%c3
> >   1.12e+08 ±25%    +364.8%  5.207e+08 ±15%  cpuidle.C1-SNB.time
> >   1.65e+08 ±20%    +175.4%  4.544e+08 ±15%  cpuidle.C1-SNB.time
> >         35 ±19%    +105.6%         72 ±28%  numa-numastat.node1.other_node
> >         24 ±41%    +247.6%         86 ±23%  numa-numastat.node1.other_node
> >         43 ±22%     +86.2%         80 ±26%  numa-vmstat.node0.nr_dirtied
> >      24576 ± 6%    +113.9%      52574 ± 1%  softirqs.RCU
> >      27717 ±11%     +98.7%      55085 ± 6%  softirqs.RCU
> >     211533 ± 6%     +58.4%     334990 ± 8%  cpuidle.C7-SNB.usage
> >     180767 ±11%     +86.7%     337416 ±10%  cpuidle.C7-SNB.usage
> >      77739 ±13%     +52.9%     118876 ±18%  cpuidle.C1E-SNB.usage
> >     104591 ±14%     +77.4%     185581 ±10%  cpuidle.C1E-SNB.usage
> >      32.09 ±14%     -24.8%      24.12 ±18%  turbostat.%pc2
> >       9.04 ± 6%     +41.6%      12.80 ± 6%  turbostat.%c1
> >        384 ±10%     +33.3%        512 ±11%  slabinfo.kmem_cache.num_objs
> >        384 ±10%     +33.3%        512 ±11%  slabinfo.kmem_cache.active_objs
> >        494 ± 8%     +25.9%        622 ± 9%  slabinfo.kmem_cache_node.active_objs
> >        512 ± 7%     +25.0%        640 ± 8%  slabinfo.kmem_cache_node.num_objs
> >        379 ± 9%     +16.7%        443 ± 7%  numa-vmstat.node0.nr_page_table_pages
> >      83427 ± 6%     +10.3%      92028 ± 5%  meminfo.DirectMap4k
> >       1579 ± 6%     -15.3%       1338 ± 7%  numa-meminfo.node1.PageTables
> >        394 ± 6%     -15.1%        334 ± 7%  numa-vmstat.node1.nr_page_table_pages
> >       1509 ± 7%     +16.6%       1760 ± 7%  numa-meminfo.node0.PageTables
> >      12681 ± 1%     -17.3%      10482 ±14%  numa-meminfo.node1.AnonPages
> >       3169 ± 1%     -17.3%       2620 ±14%  numa-vmstat.node1.nr_anon_pages
> >      10171 ± 3%     +10.9%      11283 ± 3%  slabinfo.kmalloc-512.active_objs
> >       9508 ± 1%     +21.3%      11534 ± 7%  slabinfo.kmalloc-512.active_objs
> >      10481 ± 3%     +10.9%      11620 ± 3%  slabinfo.kmalloc-512.num_objs
> >       9838 ± 1%     +20.5%      11852 ± 6%  slabinfo.kmalloc-512.num_objs
> >      53997 ± 6%     +11.1%      59981 ± 4%  numa-meminfo.node1.Slab
> >       5072 ± 1%     +11.6%       5662 ± 3%  slabinfo.kmalloc-2048.num_objs
> >       4974 ± 1%     +11.6%       5551 ± 3%  slabinfo.kmalloc-2048.active_objs
> >      12824 ± 2%     -16.1%      10754 ±14%  numa-meminfo.node1.Active(anon)
> >       3205 ± 2%     -16.2%       2687 ±14%  numa-vmstat.node1.nr_active_anon
> >       2662 ± 3%      -9.0%       2424 ± 3%  slabinfo.kmalloc-96.active_objs
> >       2710 ± 3%      -8.6%       2478 ± 3%  slabinfo.kmalloc-96.num_objs
> >      15791 ± 1%     +15.2%      18192 ± 9%  numa-meminfo.node0.AnonPages
> >       3949 ± 1%     +15.2%       4549 ± 9%  numa-vmstat.node0.nr_anon_pages
> >      13669 ± 1%      -7.5%      12645 ± 2%  slabinfo.kmalloc-16.num_objs
> >        662 ±23%   +4718.6%      31918 ±12%  time.involuntary_context_switches
> >        921 ±41%   +3577.7%      33901 ±14%  time.involuntary_context_switches
> >       2463 ± 1%     +13.1%       2786 ± 3%  vmstat.system.in
> >       2371 ± 2%     +15.5%       2739 ± 2%  vmstat.system.in
> >      49.40 ± 2%      +4.8%      51.79 ± 2%  turbostat.Cor_W
> >      77.48 ± 1%      +3.1%      79.91 ± 1%  turbostat.Pkg_W
> > 
> > testbox/testcase/testparams: lkp-st02/dd-write/5m-11HDD-RAID5-cfq-xfs-1dd
> > 
> > 743cb1ff191f00f  caeb178c60f4f93f1b45c0bc0
> > ---------------  -------------------------
> >      18571 ± 7%     +31.4%      24396 ± 4%  proc-vmstat.pgscan_direct_normal
> >      39983 ± 2%     +38.3%      55286 ± 0%  perf-stat.cpu-migrations
> >    4193962 ± 2%     +20.9%    5072009 ± 3%  perf-stat.iTLB-load-misses
> >  4.568e+09 ± 2%     -17.2%  3.781e+09 ± 1%  perf-stat.L1-icache-load-misses
> >  1.762e+10 ± 0%      -7.8%  1.625e+10 ± 1%  perf-stat.cache-references
> >  1.408e+09 ± 1%      -6.6%  1.315e+09 ± 1%  perf-stat.branch-load-misses
> >  1.407e+09 ± 1%      -6.5%  1.316e+09 ± 1%  perf-stat.branch-misses
> >  6.839e+09 ± 1%      +5.0%  7.185e+09 ± 2%  perf-stat.LLC-loads
> >  1.558e+10 ± 0%      +3.5%  1.612e+10 ± 1%  perf-stat.L1-dcache-load-misses
> >  1.318e+12 ± 0%      +3.4%  1.363e+12 ± 0%  perf-stat.L1-icache-loads
> >  2.979e+10 ± 1%      +2.4%  3.051e+10 ± 0%  perf-stat.L1-dcache-store-misses
> >  1.893e+11 ± 0%      +2.5%   1.94e+11 ± 0%  perf-stat.branch-instructions
> >  2.298e+11 ± 0%      +2.7%  2.361e+11 ± 0%  perf-stat.L1-dcache-stores
> >  1.016e+12 ± 0%      +2.6%  1.042e+12 ± 0%  perf-stat.instructions
> >  1.892e+11 ± 0%      +2.5%   1.94e+11 ± 0%  perf-stat.branch-loads
> >   3.71e+11 ± 0%      +2.4%  3.799e+11 ± 0%  perf-stat.dTLB-loads
> >  3.711e+11 ± 0%      +2.3%  3.798e+11 ± 0%  perf-stat.L1-dcache-loads
> >     325768 ± 0%      +2.7%     334461 ± 0%  vmstat.io.bo
> >       8083 ± 0%      +2.4%       8278 ± 0%  iostat.sdf.wrqm/s
> >       8083 ± 0%      +2.4%       8278 ± 0%  iostat.sdk.wrqm/s
> >       8082 ± 0%      +2.4%       8276 ± 0%  iostat.sdg.wrqm/s
> >      32615 ± 0%      +2.4%      33398 ± 0%  iostat.sdf.wkB/s
> >      32617 ± 0%      +2.4%      33401 ± 0%  iostat.sdk.wkB/s
> >      32612 ± 0%      +2.4%      33393 ± 0%  iostat.sdg.wkB/s
> >       8083 ± 0%      +2.4%       8277 ± 0%  iostat.sdl.wrqm/s
> >       8083 ± 0%      +2.4%       8276 ± 0%  iostat.sdi.wrqm/s
> >       8082 ± 0%      +2.4%       8277 ± 0%  iostat.sdc.wrqm/s
> >      32614 ± 0%      +2.4%      33396 ± 0%  iostat.sdl.wkB/s
> >       8083 ± 0%      +2.4%       8278 ± 0%  iostat.sde.wrqm/s
> >       8082 ± 0%      +2.4%       8277 ± 0%  iostat.sdh.wrqm/s
> >       8083 ± 0%      +2.4%       8277 ± 0%  iostat.sdd.wrqm/s
> >      32614 ± 0%      +2.4%      33393 ± 0%  iostat.sdi.wkB/s
> >      32611 ± 0%      +2.4%      33395 ± 0%  iostat.sdc.wkB/s
> >     325759 ± 0%      +2.4%     333577 ± 0%  iostat.md0.wkB/s
> >       1274 ± 0%      +2.4%       1305 ± 0%  iostat.md0.w/s
> >       8082 ± 0%      +2.4%       8277 ± 0%  iostat.sdb.wrqm/s
> >      32618 ± 0%      +2.4%      33398 ± 0%  iostat.sde.wkB/s
> >      32612 ± 0%      +2.4%      33395 ± 0%  iostat.sdh.wkB/s
> >      32618 ± 0%      +2.4%      33397 ± 0%  iostat.sdd.wkB/s
> >       8084 ± 0%      +2.4%       8278 ± 0%  iostat.sdj.wrqm/s
> >      32611 ± 0%      +2.4%      33396 ± 0%  iostat.sdb.wkB/s
> >      32618 ± 0%      +2.4%      33400 ± 0%  iostat.sdj.wkB/s
> >    2.3e+11 ± 0%      +2.5%  2.357e+11 ± 0%  perf-stat.dTLB-stores
> >       4898 ± 0%      +2.1%       5003 ± 0%  vmstat.system.cs
> >  1.017e+12 ± 0%      +2.4%  1.042e+12 ± 0%  perf-stat.iTLB-loads
> >    1518279 ± 0%      +2.1%    1549457 ± 0%  perf-stat.context-switches
> >  1.456e+12 ± 0%      +1.4%  1.476e+12 ± 0%  perf-stat.cpu-cycles
> >  1.456e+12 ± 0%      +1.3%  1.475e+12 ± 0%  perf-stat.ref-cycles
> >  1.819e+11 ± 0%      +1.3%  1.843e+11 ± 0%  perf-stat.bus-cycles
> > 
> > lkp-sb03 is a Sandy Bridge-EP server.
> > Memory: 64G
> > Architecture:          x86_64
> > CPU op-mode(s):        32-bit, 64-bit
> > Byte Order:            Little Endian
> > CPU(s):                32
> > On-line CPU(s) list:   0-31
> > Thread(s) per core:    2
> > Core(s) per socket:    8
> > Socket(s):             2
> > NUMA node(s):          2
> > Vendor ID:             GenuineIntel
> > CPU family:            6
> > Model:                 45
> > Stepping:              6
> > CPU MHz:               3500.613
> > BogoMIPS:              5391.16
> > Virtualization:        VT-x
> > L1d cache:             32K
> > L1i cache:             32K
> > L2 cache:              256K
> > L3 cache:              20480K
> > NUMA node0 CPU(s):     0-7,16-23
> > NUMA node1 CPU(s):     8-15,24-31
> > 
> > lkp-st02 is Core2
> > Memory: 8G
> > 
> > 
> > 
> > 
> >                           time.involuntary_context_switches
> > 
> >   40000 O+------------------------------------------------------------------+
> >         |        O               O          O                               |
> >   35000 ++O  O O        O   O             O                                 |
> >   30000 ++         O                 O         O                            |
> >         |                 O    O   O                                        |
> >   25000 ++            O                 O                                   |
> >         |                                                                   |
> >   20000 ++                                                                  |
> >         |                                                                   |
> >   15000 ++                                                                  |
> >   10000 ++                                                                  |
> >         |                                                                   |
> >    5000 ++                                                                  |
> >         |                                                .*.                |
> >       0 *+*--*-*-*-*--*-*-*-*--*-*-*-*--*-*-*--*-*-*-*--*---*-*--*-*-*-*--*-*
> > 
> > 
> > 	[*] bisect-good sample
> > 	[O] bisect-bad  sample
> > 
> > 
> > Disclaimer:
> > Results have been estimated based on internal Intel analysis and are provided
> > for informational purposes only. Any difference in system hardware or software
> > design or configuration may affect actual performance.
> > 
> > Thanks,
> > Fengguang
> > 
--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ