[<prev] [next>] [day] [month] [year] [list]
Message-ID: <1418964602.26403.23.camel@intel.com>
Date: Fri, 19 Dec 2014 12:50:02 +0800
From: Huang Ying <ying.huang@...el.com>
To: Christoph Lameter <cl@...two.org>
Cc: Linus Torvalds <torvalds@...ux-foundation.org>,
LKML <linux-kernel@...r.kernel.org>, LKP ML <lkp@...org>
Subject: [LKP] [vmstat] 7cc36bbddde: +2.6% will-it-scale.scalability -28.7%
time.involuntary_context_switches
FYI, we noticed the below changes on
commit 7cc36bbddde5cd0c98f0c06e3304ab833d662565 ("vmstat: on-demand vmstat workers V8")
testbox/testcase/testparams: lkp-sbx04/will-it-scale/performance-pread2
f0d6d1f6ff6f8525 7cc36bbddde5cd0c98f0c06e33
---------------- --------------------------
%stddev %change %stddev
\ | \
0.00 ± 1% +2.6% 0.00 ± 1% will-it-scale.scalability
533566 ± 0% -1.2% 527362 ± 0% will-it-scale.per_process_ops
486263 ± 0% -0.9% 481764 ± 0% will-it-scale.per_thread_ops
355 ± 14% +163.0% 933 ± 37% sched_debug.cpu#20.curr->pid
6 ± 34% +148.0% 15 ± 11% sched_debug.cfs_rq[52]:/.load
11214 ± 48% -55.6% 4973 ± 14% sched_debug.cfs_rq[11]:/.exec_clock
5 ± 31% -55.0% 2 ± 48% sched_debug.cpu#5.cpu_load[1]
88729 ± 46% -53.2% 41487 ± 3% sched_debug.cfs_rq[11]:/.min_vruntime
408 ± 34% -56.8% 176 ± 3% sched_debug.cpu#43.ttwu_local
7 ± 34% +114.3% 15 ± 12% sched_debug.cpu#52.load
13 ± 46% +70.4% 23 ± 26% sched_debug.cpu#43.load
4 ± 17% +105.3% 9 ± 19% sched_debug.cfs_rq[6]:/.blocked_load_avg
6562 ± 25% +120.1% 14443 ± 16% sched_debug.cfs_rq[7]:/.exec_clock
3 ± 22% -46.7% 2 ± 0% sched_debug.cfs_rq[28]:/.blocked_load_avg
68456 ± 17% -39.7% 41302 ± 47% sched_debug.cfs_rq[46]:/.spread0
187 ± 15% -38.2% 115 ± 18% sched_debug.cpu#40.ttwu_local
941 ± 25% +76.9% 1665 ± 19% sched_debug.cpu#52.curr->pid
11 ± 3% +104.4% 23 ± 49% sched_debug.cfs_rq[48]:/.tg_load_contrib
534 ± 20% +87.9% 1003 ± 48% sched_debug.cpu#6.ttwu_local
51437 ± 28% +55.8% 80132 ± 19% sched_debug.cfs_rq[8]:/.min_vruntime
14311 ± 23% +67.2% 23936 ± 24% sched_debug.cpu#22.sched_goidle
9 ± 8% -35.9% 6 ± 26% sched_debug.cpu#39.cpu_load[4]
14113 ± 24% +66.8% 23537 ± 25% sched_debug.cpu#22.ttwu_count
28742 ± 23% +66.8% 47938 ± 24% sched_debug.cpu#22.nr_switches
7 ± 17% -27.6% 5 ± 20% sched_debug.cpu#42.cpu_load[0]
58556 ± 16% -32.5% 39547 ± 4% sched_debug.cfs_rq[12]:/.min_vruntime
199 ± 9% -37.1% 125 ± 12% sched_debug.cpu#46.ttwu_local
161308 ± 28% -45.6% 87699 ± 30% sched_debug.cpu#2.sched_goidle
322838 ± 28% -45.6% 175601 ± 30% sched_debug.cpu#2.nr_switches
188 ± 8% -30.3% 131 ± 11% sched_debug.cpu#47.ttwu_local
330608 ± 31% -46.5% 176795 ± 29% sched_debug.cpu#2.sched_count
407 ± 12% +39.7% 568 ± 10% sched_debug.cpu#30.ttwu_local
75 ± 21% +31.5% 99 ± 19% sched_debug.cfs_rq[8]:/.tg_runnable_contrib
9 ± 15% -33.3% 6 ± 11% sched_debug.cpu#36.cpu_load[1]
3494 ± 21% +31.0% 4578 ± 18% sched_debug.cfs_rq[8]:/.avg->runnable_avg_sum
304 ± 9% -28.7% 217 ± 9% sched_debug.cpu#39.ttwu_local
118 ± 10% -25.2% 88 ± 11% sched_debug.cpu#48.ttwu_local
171927 ± 32% -45.5% 93673 ± 34% sched_debug.cpu#2.ttwu_count
68928 ± 14% -41.9% 40061 ± 46% sched_debug.cfs_rq[40]:/.spread0
9 ± 11% -33.3% 6 ± 17% sched_debug.cpu#38.cpu_load[4]
9 ± 15% -25.6% 7 ± 15% sched_debug.cpu#38.cpu_load[3]
537 ± 15% -28.3% 385 ± 14% sched_debug.cpu#9.ttwu_local
2694 ± 15% -21.5% 2116 ± 6% sched_debug.cfs_rq[16]:/.avg->runnable_avg_sum
58 ± 15% -21.1% 45 ± 6% sched_debug.cfs_rq[16]:/.tg_runnable_contrib
0.05 ± 8% +33.3% 0.07 ± 10% turbostat.%c3
130948 ± 11% -40.3% 78138 ± 47% sched_debug.cfs_rq[39]:/.spread0
282 ± 10% -22.1% 220 ± 7% sched_debug.cpu#38.ttwu_local
20679 ± 18% -26.9% 15120 ± 17% numa-meminfo.node2.Active(anon)
5170 ± 18% -26.9% 3778 ± 17% numa-vmstat.node2.nr_active_anon
36499 ± 4% -24.1% 27721 ± 8% sched_debug.cfs_rq[39]:/.exec_clock
259406 ± 8% -19.7% 208230 ± 10% sched_debug.cfs_rq[39]:/.min_vruntime
125 ± 5% -20.5% 99 ± 14% sched_debug.cpu#55.ttwu_local
23 ± 10% -26.1% 17 ± 19% sched_debug.cfs_rq[44]:/.load
82250 ± 9% -20.9% 65029 ± 11% sched_debug.cfs_rq[2]:/.min_vruntime
80988 ± 8% +17.1% 94812 ± 7% meminfo.DirectMap4k
47 ± 10% +15.3% 54 ± 8% sched_debug.cfs_rq[20]:/.tg_runnable_contrib
73 ± 7% -22.9% 56 ± 8% sched_debug.cpu#63.ttwu_local
46694 ± 4% +29.7% 60558 ± 7% sched_debug.cpu#7.nr_load_updates
2205 ± 9% +15.1% 2539 ± 8% sched_debug.cfs_rq[20]:/.avg->runnable_avg_sum
772 ± 10% -17.6% 636 ± 4% sched_debug.cpu#33.ttwu_local
1.64 ± 3% +21.3% 1.98 ± 10% perf-profile.cpu-cycles.copy_user_generic_string.copy_page_to_iter.shmem_file_read_iter.new_sync_read.vfs_read
140 ± 25% -34.4% 92 ± 9% sched_debug.cpu#54.ttwu_local
1.27 ± 4% -14.0% 1.09 ± 9% perf-profile.cpu-cycles.put_page.shmem_file_read_iter.new_sync_read.vfs_read.sys_pread64
1.72 ± 7% -14.8% 1.46 ± 9% perf-profile.cpu-cycles.find_get_entry.find_lock_entry.shmem_getpage_gfp.shmem_file_read_iter.new_sync_read
251 ± 1% -17.8% 206 ± 11% sched_debug.cfs_rq[39]:/.tg_runnable_contrib
11533 ± 1% -17.7% 9492 ± 11% sched_debug.cfs_rq[39]:/.avg->runnable_avg_sum
649966 ± 7% +20.0% 779700 ± 5% sched_debug.cpu#36.ttwu_count
3.99 ± 2% +20.4% 4.80 ± 7% perf-profile.cpu-cycles.__wake_up_common.__wake_up.__wake_up_bit.unlock_page.shmem_file_read_iter
613256 ± 9% -9.5% 555274 ± 10% sched_debug.cpu#26.avg_idle
554 ± 22% +55.5% 862 ± 39% sched_debug.cpu#3.ttwu_local
8 ± 4% +51.4% 13 ± 33% sched_debug.cfs_rq[25]:/.tg_load_contrib
637300 ± 8% +15.3% 734754 ± 5% sched_debug.cpu#36.sched_goidle
1275355 ± 8% +15.3% 1470926 ± 5% sched_debug.cpu#36.sched_count
1274943 ± 8% +15.3% 1469761 ± 5% sched_debug.cpu#36.nr_switches
101711 ± 5% -13.8% 87699 ± 7% sched_debug.cpu#39.nr_load_updates
374 ± 19% -31.5% 256 ± 13% sched_debug.cpu#32.ttwu_local
249095 ± 6% -14.2% 213769 ± 7% sched_debug.cfs_rq[38]:/.min_vruntime
34404 ± 5% -15.9% 28935 ± 8% sched_debug.cfs_rq[38]:/.exec_clock
44719 ± 5% -11.1% 39771 ± 6% sched_debug.cpu#62.nr_load_updates
1.06 ± 6% +13.5% 1.20 ± 6% perf-profile.cpu-cycles.wake_bit_function.__wake_up.__wake_up_bit.unlock_page.shmem_file_read_iter
11248 ± 5% -13.2% 9765 ± 9% sched_debug.cfs_rq[38]:/.avg->runnable_avg_sum
2425 ± 14% +18.9% 2883 ± 7% sched_debug.cfs_rq[19]:/.exec_clock
245 ± 5% -13.0% 213 ± 9% sched_debug.cfs_rq[38]:/.tg_runnable_contrib
129 ± 3% -21.3% 101 ± 10% sched_debug.cpu#53.ttwu_local
41965 ± 9% -13.8% 36182 ± 5% numa-meminfo.node2.Active
4000 ± 0% -11.1% 3554 ± 7% proc-vmstat.nr_shmem
16003 ± 0% -11.1% 14220 ± 7% meminfo.Shmem
94 ± 3% +10.6% 104 ± 4% sched_debug.cfs_rq[41]:/.tg_runnable_contrib
5773 ± 11% -15.2% 4893 ± 6% numa-vmstat.node1.nr_slab_unreclaimable
4366 ± 3% +10.4% 4818 ± 4% sched_debug.cfs_rq[41]:/.avg->runnable_avg_sum
23094 ± 11% -15.2% 19576 ± 6% numa-meminfo.node1.SUnreclaim
55562 ± 4% -11.5% 49192 ± 10% sched_debug.cpu#2.nr_load_updates
32616 ± 3% -10.2% 29304 ± 8% sched_debug.cpu#61.ttwu_count
667746 ± 13% +21.4% 810334 ± 3% sched_debug.cpu#34.ttwu_count
34230 ± 6% -11.2% 30408 ± 2% numa-meminfo.node1.Slab
1835 ± 2% +12.1% 2057 ± 6% numa-meminfo.node1.KernelStack
4589 ± 1% -28.7% 3271 ± 2% time.involuntary_context_switches
359 ± 3% +5.2% 378 ± 0% time.percent_of_cpu_this_job_got
1105 ± 3% +5.2% 1163 ± 0% time.system_time
lkp-sbx04: Sandy Bridge-EX
Memory: 64G
time.involuntary_context_switches
5000 ++-------------------------------------------------------------------+
4500 ++*.*.*.*.*.*. .*.*.*. .*.*.*.*.*. *.*.*.* |
* * * * |
4000 ++ |
3500 O+O O O O O O O O O |
| O O O O O O O OO O O O O O O O O O O O O O O O
3000 ++ |
2500 ++ |
2000 ++ |
| |
1500 ++ |
1000 ++ |
| |
500 ++ |
0 ++-------------------------------------O-----O-----------------------+
[*] bisect-good sample
[O] bisect-bad sample
To reproduce:
apt-get install ruby ruby-oj
git clone git://git.kernel.org/pub/scm/linux/kernel/git/wfg/lkp-tests.git
cd lkp-tests
bin/setup-local job.yaml # the job file attached in this email
bin/run-local job.yaml
Disclaimer:
Results have been estimated based on internal Intel analysis and are provided
for informational purposes only. Any difference in system hardware or software
design or configuration may affect actual performance.
Thanks,
Huang, Ying
View attachment "job.yaml" of type "text/plain" (1613 bytes)
View attachment "reproduce" of type "text/plain" (4768 bytes)
_______________________________________________
LKP mailing list
LKP@...ux.intel.com
Powered by blists - more mailing lists