lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [thread-next>] [day] [month] [year] [list]
Date:	Tue, 5 Jan 2016 19:30:52 -0800
From:	"Paul E. McKenney" <paulmck@...ux.vnet.ibm.com>
To:	kernel test robot <ying.huang@...ux.intel.com>
Cc:	lkp@...org, LKML <linux-kernel@...r.kernel.org>,
	Mike Galbraith <bitbucket@...ine.de>,
	Thomas Gleixner <tglx@...utronix.de>,
	Sebastian Andrzej Siewior <bigeasy@...utronix.de>
Subject: Re: [lkp] [rcu] b8084f70e0: No primary result change, 1414.7%
 unixbench.time.involuntary_context_switches

On Wed, Jan 06, 2016 at 10:47:43AM +0800, kernel test robot wrote:
> FYI, we noticed the below changes on
> 
> https://git.kernel.org/pub/scm/linux/kernel/git/rt/linux-rt-devel.git for-kbuild-bot/current-stable
> commit b8084f70e028697b342d6388b244d6171717fc54 ("rcu: Eliminate softirq processing from rcutree")

This is a performance degradation, correct?  If so, looks like this commit
is not ready for mainline unless set up to be configured on only for
-rt workloads.

Or am I missing the plot somewhere here?

								Thanx, Paul

> =========================================================================================
> compiler/cpufreq_governor/kconfig/nr_task/rootfs/tbox_group/test/testcase:
>   gcc-4.9/performance/x86_64-rhel/1/debian-x86_64-2015-02-07.cgz/lkp-ivb-d02/fsbuffer/unixbench
> 
> commit: 
>   ee644236a506642e20637f954d286f49a76ed83a
>   b8084f70e028697b342d6388b244d6171717fc54
> 
> ee644236a506642e b8084f70e028697b342d6388b2 
> ---------------- -------------------------- 
>          %stddev     %change         %stddev
>              \          |                \  
>     257.25 ± 18%   +1414.7%       3896 ± 13%  unixbench.time.involuntary_context_switches
>    4425148 ± 22%     -61.9%    1684712 ± 29%  cpuidle.C1E-IVB.time
>      83.75 ± 48%     +91.6%     160.50 ± 31%  cpuidle.POLL.usage
>       1902 ±  7%     -12.0%       1674 ±  1%  slabinfo.anon_vma.active_objs
>       1902 ±  7%     -12.0%       1674 ±  1%  slabinfo.anon_vma.num_objs
>     257.25 ± 18%   +1414.7%       3896 ± 13%  time.involuntary_context_switches
>       8.15 ±  6%     -14.1%       7.00 ±  1%  time.user_time
>      34345 ± 14%     -33.3%      22908 ± 33%  sched_debug.cfs_rq:/.exec_clock.2
>      34987 ± 13%     -19.7%      28093 ± 15%  sched_debug.cfs_rq:/.exec_clock.max
>     642.75 ±  5%     +59.6%       1025 ± 55%  sched_debug.cfs_rq:/.exec_clock.min
>      14639 ± 13%     -21.8%      11443 ± 16%  sched_debug.cfs_rq:/.exec_clock.stddev
>     127.75 ± 16%     -26.2%      94.25 ± 25%  sched_debug.cfs_rq:/.load_avg.3
>      38865 ± 12%     -36.0%      24885 ± 34%  sched_debug.cfs_rq:/.min_vruntime.2
>      39385 ± 12%     -22.8%      30395 ± 16%  sched_debug.cfs_rq:/.min_vruntime.max
>       1090 ± 11%     +45.3%       1584 ± 32%  sched_debug.cfs_rq:/.min_vruntime.min
>      16286 ± 13%     -25.1%      12200 ± 18%  sched_debug.cfs_rq:/.min_vruntime.stddev
>       2.00 ±  0%     -50.0%       1.00 ±  0%  sched_debug.cfs_rq:/.nr_spread_over.0
>       2.00 ±  0%     -50.0%       1.00 ±  0%  sched_debug.cfs_rq:/.nr_spread_over.max
>       0.87 ±  0%     -42.3%       0.50 ±  0%  sched_debug.cfs_rq:/.nr_spread_over.stddev
>      28.50 ± 62%    +338.6%     125.00 ±114%  sched_debug.cfs_rq:/.runnable_load_avg.1
>      16286 ± 13%     -25.1%      12200 ± 18%  sched_debug.cfs_rq:/.spread0.stddev
>     127.75 ± 16%     -26.0%      94.50 ± 25%  sched_debug.cfs_rq:/.tg_load_avg_contrib.3
>     626662 ±  5%     -13.2%     544121 ±  9%  sched_debug.cpu.avg_idle.0
>      20.25 ± 28%    +419.8%     105.25 ±132%  sched_debug.cpu.cpu_load[3].1
>     692.40 ±  7%     -15.2%     586.96 ±  9%  sched_debug.cpu.curr->pid.avg
>      36187 ± 12%     -31.5%      24793 ± 29%  sched_debug.cpu.nr_load_updates.2
>      15625 ± 11%     +61.6%      25245 ± 23%  sched_debug.cpu.nr_switches.0
>      20008 ±  4%     +14.1%      22828 ±  4%  sched_debug.cpu.nr_switches.avg
>      23333 ±  8%     +40.9%      32872 ± 18%  sched_debug.cpu.sched_count.0
>      21974 ±  4%     +12.7%      24771 ±  4%  sched_debug.cpu.sched_count.avg
>      31771 ±  9%     +22.6%      38942 ±  5%  sched_debug.cpu.sched_count.max
>       9514 ± 28%     +75.8%      16728 ±  7%  sched_debug.cpu.ttwu_count.0
>       4323 ± 10%     +40.6%       6080 ± 30%  sched_debug.cpu.ttwu_count.2
>      10068 ±  4%     +15.6%      11639 ±  4%  sched_debug.cpu.ttwu_count.avg
>       4101 ± 10%     +21.7%       4990 ±  8%  sched_debug.cpu.ttwu_count.min
>       3676 ± 21%    +102.5%       7443 ±  8%  sched_debug.cpu.ttwu_local.0
>       2014 ± 15%     +43.2%       2884 ± 10%  sched_debug.cpu.ttwu_local.2
>       3508 ±  1%     +24.8%       4377 ±  4%  sched_debug.cpu.ttwu_local.avg
>       5554 ± 19%     +34.0%       7444 ±  8%  sched_debug.cpu.ttwu_local.max
>       1846 ± 16%     +40.2%       2588 ±  6%  sched_debug.cpu.ttwu_local.min
> 
> =========================================================================================
> compiler/cpufreq_governor/kconfig/nr_task/rootfs/tbox_group/test/testcase:
>   gcc-4.9/performance/x86_64-rhel/1/debian-x86_64-2015-02-07.cgz/lkp-ivb-d02/fsdisk/unixbench
> 
> commit: 
>   ee644236a506642e20637f954d286f49a76ed83a
>   b8084f70e028697b342d6388b244d6171717fc54
> 
> ee644236a506642e b8084f70e028697b342d6388b2 
> ---------------- -------------------------- 
>        fail:runs  %reproduction    fail:runs
>            |             |             |    
>            :4           25%           1:4     kmsg.DHCP/BOOTP:Reply_not_for_us,op[#]xid[#]
>          %stddev     %change         %stddev
>              \          |                \  
>     265.75 ± 18%   +1313.4%       3756 ±  9%  unixbench.time.involuntary_context_switches
>     659.50 ±  6%     +16.5%     768.00 ±  7%  slabinfo.kmalloc-512.num_objs
>     265.75 ± 18%   +1313.4%       3756 ±  9%  time.involuntary_context_switches
>    5911312 ± 13%     -55.9%    2606719 ± 10%  cpuidle.C1E-IVB.time
>      95195 ±172%    +815.7%     871684 ±140%  cpuidle.POLL.time
>      51.75 ± 10%     -30.9%      35.75 ± 24%  cpuidle.POLL.usage
>      13361 ± 22%     -27.0%       9754 ±  4%  sched_debug.cfs_rq:/.exec_clock.stddev
>      14632 ± 24%     -28.1%      10524 ±  5%  sched_debug.cfs_rq:/.min_vruntime.stddev
>      14632 ± 24%     -28.1%      10524 ±  5%  sched_debug.cfs_rq:/.spread0.stddev
>     197.25 ± 14%     +52.9%     301.50 ± 41%  sched_debug.cfs_rq:/.util_avg.0
>     552666 ±  8%      -7.7%     510200 ±  7%  sched_debug.cpu.avg_idle.1
>      13610 ± 30%     +59.7%      21734 ±  7%  sched_debug.cpu.nr_switches.0
>      21145 ±  1%     +12.3%      23752 ±  7%  sched_debug.cpu.nr_switches.avg
>      10105 ± 21%     +62.1%      16380 ±  7%  sched_debug.cpu.nr_switches.min
>      11293 ± 28%     -44.4%       6276 ± 16%  sched_debug.cpu.nr_switches.stddev
>      21314 ± 19%     +38.1%      29437 ±  5%  sched_debug.cpu.sched_count.0
>      11261 ± 22%     +57.2%      17703 ±  9%  sched_debug.cpu.sched_count.min
>      11158 ± 27%     -36.7%       7068 ± 12%  sched_debug.cpu.sched_count.stddev
>       5854 ± 35%     +57.0%       9194 ±  8%  sched_debug.cpu.sched_goidle.0
>       3713 ± 27%     +68.4%       6253 ± 11%  sched_debug.cpu.sched_goidle.min
>       5638 ± 29%     -44.5%       3131 ± 18%  sched_debug.cpu.sched_goidle.stddev
>       8286 ± 32%     +85.0%      15333 ± 15%  sched_debug.cpu.ttwu_count.0
>      10697 ±  1%     +13.2%      12107 ±  7%  sched_debug.cpu.ttwu_count.avg
>       4845 ±  5%     +17.8%       5707 ±  8%  sched_debug.cpu.ttwu_count.min
>       2886 ± 47%    +105.5%       5931 ± 16%  sched_debug.cpu.ttwu_local.0
>       3667 ±  3%     +17.0%       4291 ±  1%  sched_debug.cpu.ttwu_local.avg
>       2127 ± 17%     +29.2%       2748 ±  7%  sched_debug.cpu.ttwu_local.min
> 
> 
> lkp-ivb-d02: Ivy Bridge
> Memory: 8G
> 
>                     unixbench.time.involuntary_context_switches
> 
>   5000 ++-------------------------------------------------------------------+
>   4500 O+                                                                   |
>        |                O              O      O            O                O
>   4000 ++                            O                            O  O      |
>   3500 ++O  O O  O    O           O         O                O  O      O  O |
>        |                   O    O         O                                 |
>   3000 ++                                          O    O                   |
>   2500 ++                                                                   |
>   2000 ++                                                                   |
>        |                                                                    |
>   1500 ++                                                                   |
>   1000 ++                                                                   |
>        |.*.. .*..         .*.  .*.*..             .*..                      |
>    500 *+   *    *.*..*.*.   *.       .*.. .*.*..*    *.*..*.*              |
>      0 ++----------O---------O-------*----*------O----O---------------------+
> 
> 	[*] bisect-good sample
> 	[O] bisect-bad  sample
> 
> To reproduce:
> 
>         git clone git://git.kernel.org/pub/scm/linux/kernel/git/wfg/lkp-tests.git
>         cd lkp-tests
>         bin/lkp install job.yaml  # job file is attached in this email
>         bin/lkp run     job.yaml
> 
> 
> Disclaimer:
> Results have been estimated based on internal Intel analysis and are provided
> for informational purposes only. Any difference in system hardware or software
> design or configuration may affect actual performance.
> 
> 
> Thanks,
> Ying Huang

> ---
> LKP_SERVER: inn
> LKP_CGI_PORT: 80
> LKP_CIFS_PORT: 139
> testcase: unixbench
> default-monitors:
>   wait: activate-monitor
>   kmsg: 
>   uptime: 
>   iostat: 
>   vmstat: 
>   numa-numastat: 
>   numa-vmstat: 
>   numa-meminfo: 
>   proc-vmstat: 
>   proc-stat:
>     interval: 10
>   meminfo: 
>   slabinfo: 
>   interrupts: 
>   lock_stat: 
>   latency_stats: 
>   softirqs: 
>   bdi_dev_mapping: 
>   diskstats: 
>   nfsstat: 
>   cpuidle: 
>   cpufreq-stats: 
>   turbostat: 
>   pmeter: 
>   sched_debug:
>     interval: 60
> cpufreq_governor: performance
> default-watchdogs:
>   oom-killer: 
>   watchdog: 
> commit: b8084f70e028697b342d6388b244d6171717fc54
> model: Ivy Bridge
> nr_cpu: 4
> memory: 8G
> nr_hdd_partitions: 1
> hdd_partitions: "/dev/disk/by-id/ata-ST1000DM003-1CH162_Z1DBQSB0-part1"
> swap_partitions: "/dev/disk/by-id/ata-ST1000DM003-1CH162_Z1DBQSB0-part3"
> rootfs_partition: "/dev/disk/by-id/ata-ST1000DM003-1CH162_Z1DBQSB0-part4"
> netconsole_port: 66723
> category: benchmark
> nr_task: 1
> unixbench:
>   test: fsdisk
> queue: bisect
> testbox: lkp-ivb-d02
> tbox_group: lkp-ivb-d02
> kconfig: x86_64-rhel
> enqueue_time: 2016-01-05 16:24:11.468660360 +08:00
> id: 62c43434378ef50b95fcd4973759daf66a148734
> user: lkp
> compiler: gcc-4.9
> head_commit: 5bae791e5bb8895ac3d65634231c99d79a0e0002
> base_commit: 168309855a7d1e16db751e9c647119fe2d2dc878
> branch: linux-devel/devel-hourly-2016010502
> rootfs: debian-x86_64-2015-02-07.cgz
> result_root: "/result/unixbench/performance-1-fsdisk/lkp-ivb-d02/debian-x86_64-2015-02-07.cgz/x86_64-rhel/gcc-4.9/b8084f70e028697b342d6388b244d6171717fc54/0"
> job_file: "/lkp/scheduled/lkp-ivb-d02/bisect_unixbench-performance-1-fsdisk-debian-x86_64-2015-02-07.cgz-x86_64-rhel-b8084f70e028697b342d6388b244d6171717fc54-20160105-89027-193oala-0.yaml"
> max_uptime: 922.0999999999999
> initrd: "/osimage/debian/debian-x86_64-2015-02-07.cgz"
> bootloader_append:
> - root=/dev/ram0
> - user=lkp
> - job=/lkp/scheduled/lkp-ivb-d02/bisect_unixbench-performance-1-fsdisk-debian-x86_64-2015-02-07.cgz-x86_64-rhel-b8084f70e028697b342d6388b244d6171717fc54-20160105-89027-193oala-0.yaml
> - ARCH=x86_64
> - kconfig=x86_64-rhel
> - branch=linux-devel/devel-hourly-2016010502
> - commit=b8084f70e028697b342d6388b244d6171717fc54
> - BOOT_IMAGE=/pkg/linux/x86_64-rhel/gcc-4.9/b8084f70e028697b342d6388b244d6171717fc54/vmlinuz-4.4.0-rc6-00236-gb8084f7
> - max_uptime=922
> - RESULT_ROOT=/result/unixbench/performance-1-fsdisk/lkp-ivb-d02/debian-x86_64-2015-02-07.cgz/x86_64-rhel/gcc-4.9/b8084f70e028697b342d6388b244d6171717fc54/0
> - LKP_SERVER=inn
> - |2-
> 
> 
>   earlyprintk=ttyS0,115200 systemd.log_level=err
>   debug apic=debug sysrq_always_enabled rcupdate.rcu_cpu_stall_timeout=100
>   panic=-1 softlockup_panic=1 nmi_watchdog=panic oops=panic load_ramdisk=2 prompt_ramdisk=0
>   console=ttyS0,115200 console=tty0 vga=normal
> 
>   rw
> lkp_initrd: "/lkp/lkp/lkp-x86_64.cgz"
> modules_initrd: "/pkg/linux/x86_64-rhel/gcc-4.9/b8084f70e028697b342d6388b244d6171717fc54/modules.cgz"
> bm_initrd: "/osimage/deps/debian-x86_64-2015-02-07.cgz/lkp.cgz,/osimage/deps/debian-x86_64-2015-02-07.cgz/run-ipconfig.cgz,/osimage/deps/debian-x86_64-2015-02-07.cgz/turbostat.cgz,/lkp/benchmarks/turbostat.cgz,/lkp/benchmarks/unixbench.cgz"
> linux_headers_initrd: "/pkg/linux/x86_64-rhel/gcc-4.9/b8084f70e028697b342d6388b244d6171717fc54/linux-headers.cgz"
> repeat_to: 2
> kernel: "/pkg/linux/x86_64-rhel/gcc-4.9/b8084f70e028697b342d6388b244d6171717fc54/vmlinuz-4.4.0-rc6-00236-gb8084f7"
> dequeue_time: 2016-01-05 16:39:47.706440461 +08:00
> job_state: finished
> loadavg: 0.58 0.25 0.10 1/123 2689
> start_time: '1451983209'
> end_time: '1451983353'
> version: "/lkp/lkp/.src-20160105-162125"


--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ