lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <20150121075254.GC9719@linux.vnet.ibm.com>
Date:	Tue, 20 Jan 2015 23:52:54 -0800
From:	"Paul E. McKenney" <paulmck@...ux.vnet.ibm.com>
To:	Huang Ying <ying.huang@...el.com>
Cc:	LKML <linux-kernel@...r.kernel.org>, LKP ML <lkp@...org>
Subject: Re: [LKP] [rcu] 00f5159aeb6: +55.3%
 will-it-scale.time.involuntary_context_switches

On Wed, Jan 21, 2015 at 03:32:44PM +0800, Huang Ying wrote:
> FYI, we noticed the below changes on
> 
> git://git.kernel.org/pub/scm/linux/kernel/git/paulmck/linux-rcu.git stall.2015.01.06a
> commit 00f5159aeb68f08ce29597be86be87d6db0c1ba1 ("rcu: Run grace-period kthreads at real-time priority")

Glad this helps!

However, the version that I am pushing during the upcoming merge window
requires a boot parameter rcutree.kthread_prio= set to a value in the
range 1..99 to have effect.  The reason is that we need to focus any
risks from this change on the people needing it.

							Thanx, Paul

> testbox/testcase/testparams: lkp-a03/will-it-scale/performance-lock1
> 
> e3663b1024d1f946  00f5159aeb68f08ce29597be86  
> ----------------  --------------------------  
>          %stddev     %change         %stddev
>              \          |                \  
>      11321 ±  5%     +55.3%      17582 ± 14%  will-it-scale.time.involuntary_context_switches
>      28999 ± 44%    +109.3%      60700 ± 25%  sched_debug.cpu#1.ttwu_count
>      58384 ± 44%    +103.6%     118851 ± 24%  sched_debug.cpu#1.nr_switches
>      26144 ± 48%     +90.3%      49762 ± 26%  sched_debug.cpu#1.sched_goidle
>        827 ± 34%     -57.3%        353 ± 21%  cpuidle.C2-ATM.usage
>      58976 ± 43%    +103.0%     119712 ± 24%  sched_debug.cpu#1.sched_count
>        503 ± 29%     -56.5%        219 ± 27%  cpuidle.C4-ATM.usage
>      11321 ±  5%     +55.3%      17582 ± 14%  time.involuntary_context_switches
>        156 ± 14%     -29.6%        110 ± 19%  cpuidle.POLL.usage
>      50583 ±  2%     -21.7%      39630 ±  4%  cpuidle.C6-ATM.usage
>   51597156 ±  2%     +23.7%   63848564 ±  3%  cpuidle.C1E-ATM.time
>      39349 ±  2%     -12.0%      34622 ±  2%  softirqs.SCHED
>     758790 ±  8%     -15.3%     642336 ± 12%  sched_debug.cpu#2.sched_count
>     378981 ±  8%     -15.6%     319776 ± 12%  sched_debug.cpu#2.ttwu_count
>     374440 ±  8%     -15.5%     316513 ± 12%  sched_debug.cpu#2.ttwu_local
>       5307 ±  0%      -1.0%       5255 ±  0%  vmstat.system.in
> 
> testbox/testcase/testparams: wsm/will-it-scale/performance-open2
> 
> e3663b1024d1f946  00f5159aeb68f08ce29597be86  
> ----------------  --------------------------  
>      22128 ±  3%    +120.2%      48719 ±  8%  will-it-scale.time.involuntary_context_switches
>       5577 ± 45%   +1492.2%      88802 ± 33%  sched_debug.cpu#6.sched_count
>       5457 ± 46%   +1526.2%      88751 ± 33%  sched_debug.cpu#6.nr_switches
>       0.45 ± 43%     -73.3%       0.12 ± 15%  turbostat.%c3
>         10 ± 24%     -55.8%          4 ± 40%  sched_debug.cpu#9.nr_uninterruptible
>        147 ± 11%     -58.5%         61 ± 32%  sched_debug.cpu#11.load
>        147 ± 11%     -58.5%         61 ± 32%  sched_debug.cfs_rq[11]:/.load
>   10821106 ± 39%     -59.8%    4347430 ± 20%  cpuidle.C3-NHM.time
>        356 ± 41%     +76.6%        629 ±  3%  sched_debug.cfs_rq[6]:/.blocked_load_avg
>      22128 ±  3%    +120.2%      48719 ±  8%  time.involuntary_context_switches
>        145 ± 36%     -35.9%         93 ± 30%  sched_debug.cpu#5.load
>        469 ± 29%     +55.9%        732 ±  4%  sched_debug.cfs_rq[6]:/.tg_load_contrib
>        145 ± 36%     -34.0%         95 ± 29%  sched_debug.cfs_rq[5]:/.load
>         99 ±  8%     -27.8%         72 ± 24%  sched_debug.cpu#9.load
>         99 ±  8%     -27.8%         72 ± 24%  sched_debug.cfs_rq[9]:/.load
>     219871 ±  6%     -33.3%     146545 ± 19%  sched_debug.cpu#3.ttwu_local
>     224610 ±  6%     -31.3%     154342 ± 18%  sched_debug.cpu#3.sched_goidle
>     451877 ±  6%     -31.1%     311378 ± 18%  sched_debug.cpu#3.sched_count
>     451737 ±  6%     -31.1%     311279 ± 18%  sched_debug.cpu#3.nr_switches
>     225208 ±  6%     -31.3%     154688 ± 18%  sched_debug.cpu#3.ttwu_count
>       4122 ±  7%     -23.6%       3148 ± 17%  cpuidle.C3-NHM.usage
>         84 ± 11%     -27.4%         61 ±  5%  sched_debug.cpu#7.cpu_load[0]
>     154565 ±  2%     -29.6%     108768 ±  9%  cpuidle.C6-NHM.usage
>     124949 ± 25%     +41.4%     176670 ± 15%  sched_debug.cpu#2.sched_goidle
>      74177 ±  2%     -23.6%      56653 ±  4%  sched_debug.cpu#3.nr_load_updates
>         79 ±  6%     -23.1%         60 ±  5%  sched_debug.cpu#7.cpu_load[1]
>      50107 ±  6%     -20.9%      39621 ±  7%  sched_debug.cfs_rq[3]:/.exec_clock
>         76 ±  4%     -20.6%         60 ±  6%  sched_debug.cpu#7.cpu_load[2]
>     385983 ±  9%     -20.4%     307347 ± 12%  sched_debug.cfs_rq[3]:/.min_vruntime
>         74 ±  2%     -18.5%         60 ±  6%  sched_debug.cpu#7.cpu_load[3]
>       1.50 ±  6%     -10.0%       1.35 ±  9%  perf-profile.cpu-cycles.selinux_file_free_security.security_file_free.__fput.____fput.task_work_run
>      46086 ±  1%     -25.7%      34260 ±  5%  softirqs.SCHED
>         72 ±  1%     -15.9%         60 ±  7%  sched_debug.cpu#7.cpu_load[4]
>         74 ± 13%     -17.7%         61 ±  2%  sched_debug.cfs_rq[7]:/.runnable_load_avg
>         65 ±  2%     +16.9%         76 ±  4%  sched_debug.cpu#1.cpu_load[3]
>       1.94 ±  5%      -9.9%       1.74 ±  6%  perf-profile.cpu-cycles.security_file_free.__fput.____fput.task_work_run.do_notify_resume
>      58851 ±  1%     +20.5%      70942 ±  9%  sched_debug.cfs_rq[1]:/.exec_clock
>      69880 ±  1%     -16.9%      58047 ± 11%  sched_debug.cpu#7.nr_load_updates
>         60 ±  5%      +9.5%         66 ±  3%  sched_debug.cpu#9.cpu_load[4]
>      12064 ±  7%     +12.4%      13561 ±  5%  slabinfo.kmalloc-64.num_objs
>         67 ±  4%     +14.9%         77 ±  4%  sched_debug.cpu#1.cpu_load[2]
>      53070 ±  7%     +18.3%      62803 ±  5%  sched_debug.cfs_rq[9]:/.exec_clock
>         64 ±  1%     +16.3%         74 ±  4%  sched_debug.cpu#1.cpu_load[4]
>      11895 ±  7%     +12.9%      13429 ±  5%  slabinfo.kmalloc-64.active_objs
>      67415 ±  1%     -16.4%      56345 ± 11%  sched_debug.cfs_rq[7]:/.exec_clock
>      58403 ±  4%     +12.0%      65434 ±  6%  sched_debug.cpu#9.nr_load_updates
>     428535 ±  8%     +18.3%     507136 ±  6%  sched_debug.cfs_rq[9]:/.min_vruntime
>     505551 ±  1%     +16.3%     588207 ±  8%  sched_debug.cfs_rq[1]:/.min_vruntime
>       1818 ±  5%      -8.4%       1665 ±  3%  slabinfo.proc_inode_cache.active_objs
>       1818 ±  5%      -8.4%       1665 ±  3%  slabinfo.proc_inode_cache.num_objs
>        616 ±  0%     +11.4%        686 ±  4%  sched_debug.cfs_rq[1]:/.tg_runnable_contrib
>       2259 ± 14%     +23.5%       2790 ±  5%  sched_debug.cpu#3.curr->pid
>      28336 ±  0%     +11.2%      31517 ±  4%  sched_debug.cfs_rq[1]:/.avg->runnable_avg_sum
>     576795 ±  1%     -14.6%     492717 ± 10%  sched_debug.cfs_rq[7]:/.min_vruntime
>       7176 ±  0%      +2.1%       7323 ±  0%  vmstat.system.cs
>      10798 ±  0%      -1.1%      10675 ±  0%  vmstat.system.in
> 
> lkp-a03: Atom
> Memory: 8G
> 
> wsm: Westmere
> Memory: 6G
> 
> 
> 
> 
>                    will-it-scale.time.involuntary_context_switches
> 
>   70000 ++------------------------------------------------------------------+
>         |                                                                   |
>   60000 ++                                                         O        |
>         O        O    O                  O    O   O    O    O   O      O    |
>   50000 ++   O O    O        O                           O    O           O O
>         | O                    O  O        O    O                    O      |
>   40000 ++              O  O        O O              O                      |
>         |                                                                   |
>   30000 ++                                                                  |
>         |     .*.                       .*.*..                      .*      |
>   20000 *+*..*   *..*.*    *.*.*..*.*.*.      *.*.*..*.*.*..*.*.*..*        |
>         |              :  :                                                 |
>   10000 ++             :  :                                                 |
>         |               ::                                                  |
>       0 ++--------------*---------------------------------------------------+
> 
> 
> 	[*] bisect-good sample
> 	[O] bisect-bad  sample
> 
> To reproduce:
> 
> 	apt-get install ruby ruby-oj
> 	git clone git://git.kernel.org/pub/scm/linux/kernel/git/wfg/lkp-tests.git
> 	cd lkp-tests
> 	bin/setup-local job.yaml # the job file attached in this email
> 	bin/run-local   job.yaml
> 
> 
> Disclaimer:
> Results have been estimated based on internal Intel analysis and are provided
> for informational purposes only. Any difference in system hardware or software
> design or configuration may affect actual performance.
> 
> 
> Thanks,
> Huang, Ying
> 
> 

> ---
> testcase: will-it-scale
> default-monitors:
>   wait: pre-test
>   uptime: 
>   iostat: 
>   vmstat: 
>   numa-numastat: 
>   numa-vmstat: 
>   numa-meminfo: 
>   proc-vmstat: 
>   proc-stat: 
>   meminfo: 
>   slabinfo: 
>   interrupts: 
>   lock_stat: 
>   latency_stats: 
>   softirqs: 
>   bdi_dev_mapping: 
>   diskstats: 
>   cpuidle: 
>   cpufreq: 
>   turbostat: 
>   sched_debug:
>     interval: 10
>   pmeter: 
> default_watchdogs:
>   watch-oom: 
>   watchdog: 
> cpufreq_governor:
> - performance
> commit: 634b0bd490b7ebd7a054cea4f7e0d25748bde678
> model: Westmere
> memory: 6G
> nr_hdd_partitions: 1
> hdd_partitions: 
> swap_partitions: 
> rootfs_partition: 
> netconsole_port: 6667
> perf-profile:
>   freq: 800
> will-it-scale:
>   test:
>   - open2
> testbox: wsm
> tbox_group: wsm
> kconfig: x86_64-rhel
> enqueue_time: 2015-01-16 17:43:47.976990621 +08:00
> head_commit: 634b0bd490b7ebd7a054cea4f7e0d25748bde678
> base_commit: eaa27f34e91a14cdceed26ed6c6793ec1d186115
> branch: next/master
> kernel: "/kernel/x86_64-rhel/634b0bd490b7ebd7a054cea4f7e0d25748bde678/vmlinuz-3.19.0-rc4-next-20150116-g634b0bd"
> user: lkp
> queue: cyclic
> rootfs: debian-x86_64-2015-01-13.cgz
> result_root: "/result/wsm/will-it-scale/performance-open2/debian-x86_64-2015-01-13.cgz/x86_64-rhel/634b0bd490b7ebd7a054cea4f7e0d25748bde678/0"
> job_file: "/lkp/scheduled/wsm/cyclic_will-it-scale-performance-open2-x86_64-rhel-HEAD-634b0bd490b7ebd7a054cea4f7e0d25748bde678-0.yaml"
> dequeue_time: 2015-01-16 23:50:07.247065623 +08:00
> nr_cpu: "$(nproc)"
> job_state: finished
> loadavg: 8.12 4.97 2.06 1/146 6693
> start_time: '1421423439'
> end_time: '1421423743'
> version: "/lkp/lkp/.src-20150116-113525"

> ./runtest.py open2 32 1 6 9 12

> _______________________________________________
> LKP mailing list
> LKP@...ux.intel.com
> 

--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ