[<prev] [next>] [day] [month] [year] [list]
Message-ID: <87k2smkif7.fsf@yhuang-dev.intel.com>
Date: Sun, 23 Aug 2015 07:33:16 +0800
From: kernel test robot <ying.huang@...el.com>
TO: Jan Kara <jack@...e.com>
CC: Linus Torvalds <torvalds@...ux-foundation.org>
Subject: [lkp] [fsnotify] 8f2f3eb59d: -4.0% will-it-scale.per_thread_ops
FYI, we noticed the below changes on
git://git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux.git master
commit 8f2f3eb59dff4ec538de55f2e0592fec85966aab ("fsnotify: fix oops in fsnotify_clear_marks_by_group_flags()")
=========================================================================================
tbox_group/testcase/rootfs/kconfig/compiler/cpufreq_governor/test:
lkp-sbx04/will-it-scale/debian-x86_64-2015-02-07.cgz/x86_64-rhel/gcc-4.9/performance/read1
commit:
447f6a95a9c80da7faaec3e66e656eab8f262640
8f2f3eb59dff4ec538de55f2e0592fec85966aab
447f6a95a9c80da7 8f2f3eb59dff4ec538de55f2e0
---------------- --------------------------
%stddev %change %stddev
\ | \
1844687 ± 0% -4.0% 1770899 ± 0% will-it-scale.per_thread_ops
283.69 ± 0% +9.5% 310.64 ± 0% will-it-scale.time.user_time
4576 ± 3% -7.3% 4242 ± 6% will-it-scale.time.voluntary_context_switches
7211 ± 10% +54.0% 11101 ± 18% cpuidle.C1E-SNB.usage
10636 ± 36% +69.3% 18003 ± 36% numa-meminfo.node1.Shmem
1.07 ± 4% -13.1% 0.93 ± 9% perf-profile.cpu-cycles.selinux_file_permission.security_file_permission.rw_verify_area.vfs_read.sys_read
4576 ± 3% -7.3% 4242 ± 6% time.voluntary_context_switches
526.75 ±104% -94.2% 30.50 ± 98% numa-numastat.node1.other_node
1540 ± 35% -74.2% 398.00 ± 90% numa-numastat.node2.other_node
32344 ± 5% +7.4% 34722 ± 4% numa-vmstat.node0.numa_other
2658 ± 36% +69.3% 4500 ± 36% numa-vmstat.node1.nr_shmem
935792 ±136% +4247.3% 40682138 ±141% latency_stats.avg.nfs_wait_on_request.nfs_updatepage.nfs_write_end.generic_perform_write.__generic_file_write_iter.generic_file_write_iter.nfs_file_write.__vfs_write.vfs_write.SyS_write.entry_SYSCALL_64_fastpath
935792 ±136% +4247.3% 40682138 ±141% latency_stats.max.nfs_wait_on_request.nfs_updatepage.nfs_write_end.generic_perform_write.__generic_file_write_iter.generic_file_write_iter.nfs_file_write.__vfs_write.vfs_write.SyS_write.entry_SYSCALL_64_fastpath
935792 ±136% +4247.3% 40682138 ±141% latency_stats.sum.nfs_wait_on_request.nfs_updatepage.nfs_write_end.generic_perform_write.__generic_file_write_iter.generic_file_write_iter.nfs_file_write.__vfs_write.vfs_write.SyS_write.entry_SYSCALL_64_fastpath
12893 ± 2% -9.1% 11716 ± 1% slabinfo.kmalloc-192.active_objs
1653 ± 9% -10.3% 1483 ± 5% slabinfo.mnt_cache.active_objs
1653 ± 9% -10.3% 1483 ± 5% slabinfo.mnt_cache.num_objs
1.75 ± 47% -81.0% 0.33 ±141% sched_debug.cfs_rq[10]:/.nr_spread_over
-343206 ±-27% -73.2% -91995 ±-170% sched_debug.cfs_rq[14]:/.spread0
533.25 ± 82% -81.5% 98.75 ± 42% sched_debug.cfs_rq[18]:/.blocked_load_avg
541.75 ± 82% -81.3% 101.25 ± 41% sched_debug.cfs_rq[18]:/.tg_load_contrib
-1217705 ± -5% -30.2% -850080 ±-15% sched_debug.cfs_rq[26]:/.spread0
89722 ± 9% +9.8% 98495 ± 10% sched_debug.cfs_rq[32]:/.exec_clock
101180 ±132% +180.8% 284154 ± 30% sched_debug.cfs_rq[35]:/.spread0
37332 ±473% +725.2% 308082 ± 59% sched_debug.cfs_rq[38]:/.spread0
32054 ±502% +981.6% 346689 ± 39% sched_debug.cfs_rq[39]:/.spread0
1.00 ±100% +100.0% 2.00 ± 50% sched_debug.cfs_rq[42]:/.nr_spread_over
-125980 ±-218% -307.1% 260875 ± 46% sched_debug.cfs_rq[42]:/.spread0
-111501 ±-102% -288.7% 210354 ± 94% sched_debug.cfs_rq[45]:/.spread0
-173363 ±-34% -221.0% 209775 ± 94% sched_debug.cfs_rq[47]:/.spread0
-302090 ±-43% -121.8% 65953 ±322% sched_debug.cfs_rq[4]:/.spread0
-490175 ±-18% -41.1% -288722 ±-31% sched_debug.cfs_rq[50]:/.spread0
-594948 ±-10% -59.7% -239840 ±-33% sched_debug.cfs_rq[51]:/.spread0
1.00 ±100% +6050.0% 61.50 ±141% sched_debug.cfs_rq[53]:/.blocked_load_avg
10.50 ± 8% +614.3% 75.00 ±122% sched_debug.cfs_rq[53]:/.tg_load_contrib
-596043 ±-10% -49.0% -304277 ±-36% sched_debug.cfs_rq[54]:/.spread0
10.00 ± 0% +2062.5% 216.25 ± 40% sched_debug.cfs_rq[56]:/.tg_load_contrib
17.75 ±173% +1302.8% 249.00 ± 26% sched_debug.cfs_rq[60]:/.blocked_load_avg
-809633 ± -9% -36.2% -516886 ±-23% sched_debug.cfs_rq[60]:/.spread0
28.00 ±109% +828.6% 260.00 ± 25% sched_debug.cfs_rq[60]:/.tg_load_contrib
277.75 ± 95% -86.3% 38.00 ±171% sched_debug.cfs_rq[7]:/.blocked_load_avg
293.25 ± 90% -81.8% 53.50 ±121% sched_debug.cfs_rq[7]:/.tg_load_contrib
17.50 ± 2% -28.6% 12.50 ± 34% sched_debug.cpu#0.cpu_load[2]
17.00 ± 4% -25.0% 12.75 ± 35% sched_debug.cpu#0.cpu_load[3]
2907 ± 12% +195.9% 8603 ± 63% sched_debug.cpu#0.sched_goidle
16.50 ± 3% -9.1% 15.00 ± 0% sched_debug.cpu#1.cpu_load[2]
16.50 ± 3% -7.6% 15.25 ± 2% sched_debug.cpu#1.cpu_load[3]
5595 ± 26% -36.4% 3557 ± 11% sched_debug.cpu#11.nr_switches
6885 ± 92% -76.2% 1639 ± 40% sched_debug.cpu#11.ttwu_count
1350 ± 34% -55.0% 608.00 ± 14% sched_debug.cpu#11.ttwu_local
17892 ± 74% -78.3% 3877 ± 18% sched_debug.cpu#12.nr_switches
1288 ± 27% -49.8% 647.50 ± 37% sched_debug.cpu#12.ttwu_local
1405 ± 22% -52.7% 664.50 ± 23% sched_debug.cpu#13.ttwu_local
1.25 ±182% -440.0% -4.25 ±-50% sched_debug.cpu#17.nr_uninterruptible
1976 ± 5% -10.0% 1779 ± 0% sched_debug.cpu#18.curr->pid
983.75 ± 8% +101.6% 1983 ± 32% sched_debug.cpu#18.ttwu_local
-0.25 ±-911% +2300.0% -6.00 ±-28% sched_debug.cpu#21.nr_uninterruptible
2979 ± 49% +159.6% 7734 ± 75% sched_debug.cpu#22.ttwu_count
1111 ± 21% +127.6% 2528 ± 32% sched_debug.cpu#22.ttwu_local
1.00 ±141% -275.0% -1.75 ±-84% sched_debug.cpu#25.nr_uninterruptible
14419 ± 54% -58.2% 6022 ± 84% sched_debug.cpu#25.ttwu_count
14395 ± 70% +252.4% 50729 ± 39% sched_debug.cpu#28.nr_switches
-4.75 ±-17% -115.8% 0.75 ±218% sched_debug.cpu#30.nr_uninterruptible
2335 ±115% -76.6% 547.25 ± 18% sched_debug.cpu#34.ttwu_count
1258 ± 25% -43.3% 713.75 ± 11% sched_debug.cpu#35.nr_switches
1409 ± 23% -39.6% 851.75 ± 9% sched_debug.cpu#35.sched_count
969.50 ± 69% -68.8% 302.00 ± 38% sched_debug.cpu#35.ttwu_count
382.00 ± 37% -66.0% 130.00 ± 14% sched_debug.cpu#35.ttwu_local
808.75 ± 18% +28.3% 1037 ± 15% sched_debug.cpu#38.nr_switches
948.50 ± 16% +23.2% 1168 ± 13% sched_debug.cpu#38.sched_count
70695 ± 2% +6.2% 75047 ± 4% sched_debug.cpu#41.nr_load_updates
1269 ± 13% +55.3% 1970 ± 25% sched_debug.cpu#46.nr_switches
3.25 ± 93% -76.9% 0.75 ±197% sched_debug.cpu#46.nr_uninterruptible
1375 ± 12% +51.1% 2078 ± 23% sched_debug.cpu#46.sched_count
3958 ± 97% +462.9% 22281 ± 25% sched_debug.cpu#50.ttwu_count
457.25 ± 26% +64.3% 751.25 ± 28% sched_debug.cpu#53.ttwu_local
753041 ± 3% -11.1% 669815 ± 5% sched_debug.cpu#58.avg_idle
-1.75 ±-142% -257.1% 2.75 ± 64% sched_debug.cpu#59.nr_uninterruptible
2581 ± 27% +1426.4% 39408 ± 57% sched_debug.cpu#60.nr_switches
2632 ± 27% +1400.2% 39495 ± 57% sched_debug.cpu#60.sched_count
34156 ± 94% -94.8% 1776 ± 15% sched_debug.cpu#61.nr_switches
34250 ± 94% -94.7% 1825 ± 15% sched_debug.cpu#61.sched_count
16821 ± 96% -95.4% 768.50 ± 11% sched_debug.cpu#61.sched_goidle
8128 ±146% -91.7% 676.00 ± 10% sched_debug.cpu#61.ttwu_count
=========================================================================================
tbox_group/testcase/rootfs/kconfig/compiler/cpufreq_governor/test:
ivb42/will-it-scale/debian-x86_64-2015-02-07.cgz/x86_64-rhel/gcc-4.9/performance/readseek1
commit:
447f6a95a9c80da7faaec3e66e656eab8f262640
8f2f3eb59dff4ec538de55f2e0592fec85966aab
447f6a95a9c80da7 8f2f3eb59dff4ec538de55f2e0
---------------- --------------------------
%stddev %change %stddev
\ | \
1915464 ± 0% -2.4% 1869344 ± 0% will-it-scale.per_thread_ops
473.17 ± 0% +6.9% 505.66 ± 0% will-it-scale.time.user_time
0.20 ± 5% -49.4% 0.10 ± 35% turbostat.Pkg%pc6
3.38 ± 0% +34.0% 4.53 ± 1% perf-profile.cpu-cycles.find_get_entry.find_lock_entry.shmem_getpage_gfp.shmem_file_read_iter.__vfs_read
7.42 ± 0% +16.3% 8.62 ± 1% perf-profile.cpu-cycles.find_lock_entry.shmem_getpage_gfp.shmem_file_read_iter.__vfs_read.vfs_read
0.57 ± 6% +72.2% 0.99 ± 6% perf-profile.cpu-cycles.radix_tree_lookup_slot.find_get_entry.find_lock_entry.shmem_getpage_gfp.shmem_file_read_iter
10.58 ± 0% +11.4% 11.79 ± 1% perf-profile.cpu-cycles.shmem_getpage_gfp.shmem_file_read_iter.__vfs_read.vfs_read.sys_read
30.50 ±150% +1140.2% 378.25 ± 49% sched_debug.cfs_rq[22]:/.blocked_load_avg
44.75 ±103% +788.3% 397.50 ± 46% sched_debug.cfs_rq[22]:/.tg_load_contrib
89.50 ±159% +300.3% 358.25 ± 75% sched_debug.cfs_rq[2]:/.blocked_load_avg
115.75 ±123% +231.7% 384.00 ± 70% sched_debug.cfs_rq[2]:/.tg_load_contrib
0.50 ±100% +750.0% 4.25 ± 67% sched_debug.cfs_rq[32]:/.nr_spread_over
499.50 ± 44% -98.2% 9.00 ±101% sched_debug.cfs_rq[40]:/.blocked_load_avg
505.50 ± 44% -95.2% 24.50 ± 73% sched_debug.cfs_rq[40]:/.tg_load_contrib
421.00 ± 56% -85.7% 60.25 ±109% sched_debug.cfs_rq[42]:/.blocked_load_avg
428.75 ± 56% -80.4% 84.00 ± 86% sched_debug.cfs_rq[42]:/.tg_load_contrib
8053 ± 2% +13.4% 9132 ± 5% sched_debug.cfs_rq[47]:/.avg->runnable_avg_sum
175.25 ± 2% +12.7% 197.50 ± 5% sched_debug.cfs_rq[47]:/.tg_runnable_contrib
0.25 ±173% +1500.0% 4.00 ± 77% sched_debug.cfs_rq[8]:/.nr_spread_over
90.75 ± 13% -23.1% 69.75 ± 15% sched_debug.cpu#0.cpu_load[2]
97.00 ± 15% -28.4% 69.50 ± 16% sched_debug.cpu#0.cpu_load[3]
99.50 ± 14% -27.6% 72.00 ± 18% sched_debug.cpu#0.cpu_load[4]
-10.25 ±-14% -73.2% -2.75 ±-180% sched_debug.cpu#1.nr_uninterruptible
8173 ±106% -78.9% 1722 ± 35% sched_debug.cpu#10.nr_switches
3896 ±112% -81.3% 727.50 ± 36% sched_debug.cpu#10.sched_goidle
515.00 ± 40% -47.2% 271.75 ± 49% sched_debug.cpu#10.ttwu_local
2.00 ± 81% -325.0% -4.50 ±-77% sched_debug.cpu#11.nr_uninterruptible
3818 ± 39% -58.2% 1598 ± 68% sched_debug.cpu#15.ttwu_local
0.50 ±331% -650.0% -2.75 ±-74% sched_debug.cpu#16.nr_uninterruptible
12671 ± 30% -58.4% 5270 ± 46% sched_debug.cpu#20.ttwu_count
2285 ± 70% -57.0% 983.50 ± 25% sched_debug.cpu#20.ttwu_local
2722 ± 79% -72.9% 738.75 ± 51% sched_debug.cpu#21.ttwu_local
-2.50 ±-72% -200.0% 2.50 ± 82% sched_debug.cpu#23.nr_uninterruptible
1183 ± 31% +188.4% 3413 ± 22% sched_debug.cpu#24.nr_switches
1384 ± 45% +148.4% 3438 ± 22% sched_debug.cpu#24.sched_count
318.50 ± 54% +347.5% 1425 ± 21% sched_debug.cpu#24.ttwu_local
5255 ± 46% -60.2% 2090 ± 54% sched_debug.cpu#25.nr_switches
5276 ± 46% -59.9% 2114 ± 54% sched_debug.cpu#25.sched_count
1893 ± 42% -66.9% 627.00 ± 75% sched_debug.cpu#25.ttwu_local
1.25 ±142% +240.0% 4.25 ± 45% sched_debug.cpu#27.nr_uninterruptible
0.75 ±272% -322.2% -1.67 ±-28% sched_debug.cpu#31.nr_uninterruptible
1977 ±140% -86.5% 267.25 ± 10% sched_debug.cpu#32.sched_goidle
7.67 ± 78% -122.8% -1.75 ±-84% sched_debug.cpu#34.nr_uninterruptible
3642 ± 37% +205.0% 11108 ± 53% sched_debug.cpu#39.nr_switches
1250 ± 51% +292.0% 4902 ± 52% sched_debug.cpu#39.sched_goidle
3.00 ± 0% +216.7% 9.50 ± 30% sched_debug.cpu#45.cpu_load[0]
3.50 ± 24% +121.4% 7.75 ± 10% sched_debug.cpu#45.cpu_load[1]
3.25 ± 13% +123.1% 7.25 ± 11% sched_debug.cpu#45.cpu_load[2]
3.25 ± 13% +92.3% 6.25 ± 23% sched_debug.cpu#45.cpu_load[3]
3.00 ± 0% +91.7% 5.75 ± 22% sched_debug.cpu#45.cpu_load[4]
1593 ± 19% +63.6% 2605 ± 30% sched_debug.cpu#47.curr->pid
365.75 ± 39% +254.6% 1297 ± 98% sched_debug.cpu#6.ttwu_local
8717 ± 80% -78.7% 1856 ± 45% sched_debug.cpu#8.nr_switches
3992 ± 85% -80.5% 778.50 ± 51% sched_debug.cpu#8.sched_goidle
6221 ±128% -83.9% 998.75 ± 44% sched_debug.cpu#8.ttwu_count
722.00 ± 71% -69.5% 220.25 ± 5% sched_debug.cpu#8.ttwu_local
0.25 ±173% +321.4% 1.05 ± 5% sched_debug.rt_rq[12]:/.rt_time
0.04 ±173% +311.0% 0.17 ± 8% sched_debug.rt_rq[13]:/.rt_time
lkp-sbx04: Sandy Bridge-EX
Memory: 64G
ivb42: Ivytown Ivy Bridge-EP
Memory: 64G
will-it-scale.time.user_time
325 ++--------------------------------------------------------------------+
320 ++ O O O |
| O O O O O O O O |
315 O+ O O |
310 ++ O O O O
| O |
305 ++ |
300 ++ |
295 ++ |
| *.. |
290 ++ .*.. ..*...*.. .. . |
285 *+..*... .. . .*...*. . . *...*...*..*...*... ..*.. |
| * *. * *. .|
280 ++ *
275 ++--------------------------------------------------------------------+
[*] bisect-good sample
[O] bisect-bad sample
To reproduce:
git clone git://git.kernel.org/pub/scm/linux/kernel/git/wfg/lkp-tests.git
cd lkp-tests
bin/lkp install job.yaml # job file is attached in this email
bin/lkp run job.yaml
Disclaimer:
Results have been estimated based on internal Intel analysis and are provided
for informational purposes only. Any difference in system hardware or software
design or configuration may affect actual performance.
Thanks,
Ying Huang
View attachment "job.yaml" of type "text/plain" (3209 bytes)
View attachment "reproduce" of type "text/plain" (4773 bytes)
Powered by blists - more mailing lists