[<prev] [next>] [thread-next>] [day] [month] [year] [list]
Message-ID: <1421825564.6126.29.camel@intel.com>
Date: Wed, 21 Jan 2015 15:32:44 +0800
From: Huang Ying <ying.huang@...el.com>
To: "Paul E. McKenney" <paulmck@...ux.vnet.ibm.com>
Cc: LKML <linux-kernel@...r.kernel.org>, LKP ML <lkp@...org>
Subject: [LKP] [rcu] 00f5159aeb6: +55.3%
will-it-scale.time.involuntary_context_switches
FYI, we noticed the below changes on
git://git.kernel.org/pub/scm/linux/kernel/git/paulmck/linux-rcu.git stall.2015.01.06a
commit 00f5159aeb68f08ce29597be86be87d6db0c1ba1 ("rcu: Run grace-period kthreads at real-time priority")
testbox/testcase/testparams: lkp-a03/will-it-scale/performance-lock1
e3663b1024d1f946 00f5159aeb68f08ce29597be86
---------------- --------------------------
%stddev %change %stddev
\ | \
11321 ± 5% +55.3% 17582 ± 14% will-it-scale.time.involuntary_context_switches
28999 ± 44% +109.3% 60700 ± 25% sched_debug.cpu#1.ttwu_count
58384 ± 44% +103.6% 118851 ± 24% sched_debug.cpu#1.nr_switches
26144 ± 48% +90.3% 49762 ± 26% sched_debug.cpu#1.sched_goidle
827 ± 34% -57.3% 353 ± 21% cpuidle.C2-ATM.usage
58976 ± 43% +103.0% 119712 ± 24% sched_debug.cpu#1.sched_count
503 ± 29% -56.5% 219 ± 27% cpuidle.C4-ATM.usage
11321 ± 5% +55.3% 17582 ± 14% time.involuntary_context_switches
156 ± 14% -29.6% 110 ± 19% cpuidle.POLL.usage
50583 ± 2% -21.7% 39630 ± 4% cpuidle.C6-ATM.usage
51597156 ± 2% +23.7% 63848564 ± 3% cpuidle.C1E-ATM.time
39349 ± 2% -12.0% 34622 ± 2% softirqs.SCHED
758790 ± 8% -15.3% 642336 ± 12% sched_debug.cpu#2.sched_count
378981 ± 8% -15.6% 319776 ± 12% sched_debug.cpu#2.ttwu_count
374440 ± 8% -15.5% 316513 ± 12% sched_debug.cpu#2.ttwu_local
5307 ± 0% -1.0% 5255 ± 0% vmstat.system.in
testbox/testcase/testparams: wsm/will-it-scale/performance-open2
e3663b1024d1f946 00f5159aeb68f08ce29597be86
---------------- --------------------------
22128 ± 3% +120.2% 48719 ± 8% will-it-scale.time.involuntary_context_switches
5577 ± 45% +1492.2% 88802 ± 33% sched_debug.cpu#6.sched_count
5457 ± 46% +1526.2% 88751 ± 33% sched_debug.cpu#6.nr_switches
0.45 ± 43% -73.3% 0.12 ± 15% turbostat.%c3
10 ± 24% -55.8% 4 ± 40% sched_debug.cpu#9.nr_uninterruptible
147 ± 11% -58.5% 61 ± 32% sched_debug.cpu#11.load
147 ± 11% -58.5% 61 ± 32% sched_debug.cfs_rq[11]:/.load
10821106 ± 39% -59.8% 4347430 ± 20% cpuidle.C3-NHM.time
356 ± 41% +76.6% 629 ± 3% sched_debug.cfs_rq[6]:/.blocked_load_avg
22128 ± 3% +120.2% 48719 ± 8% time.involuntary_context_switches
145 ± 36% -35.9% 93 ± 30% sched_debug.cpu#5.load
469 ± 29% +55.9% 732 ± 4% sched_debug.cfs_rq[6]:/.tg_load_contrib
145 ± 36% -34.0% 95 ± 29% sched_debug.cfs_rq[5]:/.load
99 ± 8% -27.8% 72 ± 24% sched_debug.cpu#9.load
99 ± 8% -27.8% 72 ± 24% sched_debug.cfs_rq[9]:/.load
219871 ± 6% -33.3% 146545 ± 19% sched_debug.cpu#3.ttwu_local
224610 ± 6% -31.3% 154342 ± 18% sched_debug.cpu#3.sched_goidle
451877 ± 6% -31.1% 311378 ± 18% sched_debug.cpu#3.sched_count
451737 ± 6% -31.1% 311279 ± 18% sched_debug.cpu#3.nr_switches
225208 ± 6% -31.3% 154688 ± 18% sched_debug.cpu#3.ttwu_count
4122 ± 7% -23.6% 3148 ± 17% cpuidle.C3-NHM.usage
84 ± 11% -27.4% 61 ± 5% sched_debug.cpu#7.cpu_load[0]
154565 ± 2% -29.6% 108768 ± 9% cpuidle.C6-NHM.usage
124949 ± 25% +41.4% 176670 ± 15% sched_debug.cpu#2.sched_goidle
74177 ± 2% -23.6% 56653 ± 4% sched_debug.cpu#3.nr_load_updates
79 ± 6% -23.1% 60 ± 5% sched_debug.cpu#7.cpu_load[1]
50107 ± 6% -20.9% 39621 ± 7% sched_debug.cfs_rq[3]:/.exec_clock
76 ± 4% -20.6% 60 ± 6% sched_debug.cpu#7.cpu_load[2]
385983 ± 9% -20.4% 307347 ± 12% sched_debug.cfs_rq[3]:/.min_vruntime
74 ± 2% -18.5% 60 ± 6% sched_debug.cpu#7.cpu_load[3]
1.50 ± 6% -10.0% 1.35 ± 9% perf-profile.cpu-cycles.selinux_file_free_security.security_file_free.__fput.____fput.task_work_run
46086 ± 1% -25.7% 34260 ± 5% softirqs.SCHED
72 ± 1% -15.9% 60 ± 7% sched_debug.cpu#7.cpu_load[4]
74 ± 13% -17.7% 61 ± 2% sched_debug.cfs_rq[7]:/.runnable_load_avg
65 ± 2% +16.9% 76 ± 4% sched_debug.cpu#1.cpu_load[3]
1.94 ± 5% -9.9% 1.74 ± 6% perf-profile.cpu-cycles.security_file_free.__fput.____fput.task_work_run.do_notify_resume
58851 ± 1% +20.5% 70942 ± 9% sched_debug.cfs_rq[1]:/.exec_clock
69880 ± 1% -16.9% 58047 ± 11% sched_debug.cpu#7.nr_load_updates
60 ± 5% +9.5% 66 ± 3% sched_debug.cpu#9.cpu_load[4]
12064 ± 7% +12.4% 13561 ± 5% slabinfo.kmalloc-64.num_objs
67 ± 4% +14.9% 77 ± 4% sched_debug.cpu#1.cpu_load[2]
53070 ± 7% +18.3% 62803 ± 5% sched_debug.cfs_rq[9]:/.exec_clock
64 ± 1% +16.3% 74 ± 4% sched_debug.cpu#1.cpu_load[4]
11895 ± 7% +12.9% 13429 ± 5% slabinfo.kmalloc-64.active_objs
67415 ± 1% -16.4% 56345 ± 11% sched_debug.cfs_rq[7]:/.exec_clock
58403 ± 4% +12.0% 65434 ± 6% sched_debug.cpu#9.nr_load_updates
428535 ± 8% +18.3% 507136 ± 6% sched_debug.cfs_rq[9]:/.min_vruntime
505551 ± 1% +16.3% 588207 ± 8% sched_debug.cfs_rq[1]:/.min_vruntime
1818 ± 5% -8.4% 1665 ± 3% slabinfo.proc_inode_cache.active_objs
1818 ± 5% -8.4% 1665 ± 3% slabinfo.proc_inode_cache.num_objs
616 ± 0% +11.4% 686 ± 4% sched_debug.cfs_rq[1]:/.tg_runnable_contrib
2259 ± 14% +23.5% 2790 ± 5% sched_debug.cpu#3.curr->pid
28336 ± 0% +11.2% 31517 ± 4% sched_debug.cfs_rq[1]:/.avg->runnable_avg_sum
576795 ± 1% -14.6% 492717 ± 10% sched_debug.cfs_rq[7]:/.min_vruntime
7176 ± 0% +2.1% 7323 ± 0% vmstat.system.cs
10798 ± 0% -1.1% 10675 ± 0% vmstat.system.in
lkp-a03: Atom
Memory: 8G
wsm: Westmere
Memory: 6G
will-it-scale.time.involuntary_context_switches
70000 ++------------------------------------------------------------------+
| |
60000 ++ O |
O O O O O O O O O O |
50000 ++ O O O O O O O O
| O O O O O O |
40000 ++ O O O O O |
| |
30000 ++ |
| .*. .*.*.. .* |
20000 *+*..* *..*.* *.*.*..*.*.*. *.*.*..*.*.*..*.*.*..* |
| : : |
10000 ++ : : |
| :: |
0 ++--------------*---------------------------------------------------+
[*] bisect-good sample
[O] bisect-bad sample
To reproduce:
apt-get install ruby ruby-oj
git clone git://git.kernel.org/pub/scm/linux/kernel/git/wfg/lkp-tests.git
cd lkp-tests
bin/setup-local job.yaml # the job file attached in this email
bin/run-local job.yaml
Disclaimer:
Results have been estimated based on internal Intel analysis and are provided
for informational purposes only. Any difference in system hardware or software
design or configuration may affect actual performance.
Thanks,
Huang, Ying
View attachment "job.yaml" of type "text/plain" (1582 bytes)
View attachment "reproduce" of type "text/plain" (31 bytes)
_______________________________________________
LKP mailing list
LKP@...ux.intel.com
Powered by blists - more mailing lists