lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [day] [month] [year] [list]
Message-ID: <87k2smkif7.fsf@yhuang-dev.intel.com>
Date:	Sun, 23 Aug 2015 07:33:16 +0800
From:	kernel test robot <ying.huang@...el.com>
TO:	Jan Kara <jack@...e.com>
CC:	Linus Torvalds <torvalds@...ux-foundation.org>
Subject: [lkp] [fsnotify] 8f2f3eb59d: -4.0% will-it-scale.per_thread_ops

FYI, we noticed the below changes on

git://git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux.git master
commit 8f2f3eb59dff4ec538de55f2e0592fec85966aab ("fsnotify: fix oops in fsnotify_clear_marks_by_group_flags()")


=========================================================================================
tbox_group/testcase/rootfs/kconfig/compiler/cpufreq_governor/test:
  lkp-sbx04/will-it-scale/debian-x86_64-2015-02-07.cgz/x86_64-rhel/gcc-4.9/performance/read1

commit: 
  447f6a95a9c80da7faaec3e66e656eab8f262640
  8f2f3eb59dff4ec538de55f2e0592fec85966aab

447f6a95a9c80da7 8f2f3eb59dff4ec538de55f2e0 
---------------- -------------------------- 
         %stddev     %change         %stddev
             \          |                \  
   1844687 ±  0%      -4.0%    1770899 ±  0%  will-it-scale.per_thread_ops
    283.69 ±  0%      +9.5%     310.64 ±  0%  will-it-scale.time.user_time
      4576 ±  3%      -7.3%       4242 ±  6%  will-it-scale.time.voluntary_context_switches
      7211 ± 10%     +54.0%      11101 ± 18%  cpuidle.C1E-SNB.usage
     10636 ± 36%     +69.3%      18003 ± 36%  numa-meminfo.node1.Shmem
      1.07 ±  4%     -13.1%       0.93 ±  9%  perf-profile.cpu-cycles.selinux_file_permission.security_file_permission.rw_verify_area.vfs_read.sys_read
      4576 ±  3%      -7.3%       4242 ±  6%  time.voluntary_context_switches
    526.75 ±104%     -94.2%      30.50 ± 98%  numa-numastat.node1.other_node
      1540 ± 35%     -74.2%     398.00 ± 90%  numa-numastat.node2.other_node
     32344 ±  5%      +7.4%      34722 ±  4%  numa-vmstat.node0.numa_other
      2658 ± 36%     +69.3%       4500 ± 36%  numa-vmstat.node1.nr_shmem
    935792 ±136%   +4247.3%   40682138 ±141%  latency_stats.avg.nfs_wait_on_request.nfs_updatepage.nfs_write_end.generic_perform_write.__generic_file_write_iter.generic_file_write_iter.nfs_file_write.__vfs_write.vfs_write.SyS_write.entry_SYSCALL_64_fastpath
    935792 ±136%   +4247.3%   40682138 ±141%  latency_stats.max.nfs_wait_on_request.nfs_updatepage.nfs_write_end.generic_perform_write.__generic_file_write_iter.generic_file_write_iter.nfs_file_write.__vfs_write.vfs_write.SyS_write.entry_SYSCALL_64_fastpath
    935792 ±136%   +4247.3%   40682138 ±141%  latency_stats.sum.nfs_wait_on_request.nfs_updatepage.nfs_write_end.generic_perform_write.__generic_file_write_iter.generic_file_write_iter.nfs_file_write.__vfs_write.vfs_write.SyS_write.entry_SYSCALL_64_fastpath
     12893 ±  2%      -9.1%      11716 ±  1%  slabinfo.kmalloc-192.active_objs
      1653 ±  9%     -10.3%       1483 ±  5%  slabinfo.mnt_cache.active_objs
      1653 ±  9%     -10.3%       1483 ±  5%  slabinfo.mnt_cache.num_objs
      1.75 ± 47%     -81.0%       0.33 ±141%  sched_debug.cfs_rq[10]:/.nr_spread_over
   -343206 ±-27%     -73.2%     -91995 ±-170%  sched_debug.cfs_rq[14]:/.spread0
    533.25 ± 82%     -81.5%      98.75 ± 42%  sched_debug.cfs_rq[18]:/.blocked_load_avg
    541.75 ± 82%     -81.3%     101.25 ± 41%  sched_debug.cfs_rq[18]:/.tg_load_contrib
  -1217705 ± -5%     -30.2%    -850080 ±-15%  sched_debug.cfs_rq[26]:/.spread0
     89722 ±  9%      +9.8%      98495 ± 10%  sched_debug.cfs_rq[32]:/.exec_clock
    101180 ±132%    +180.8%     284154 ± 30%  sched_debug.cfs_rq[35]:/.spread0
     37332 ±473%    +725.2%     308082 ± 59%  sched_debug.cfs_rq[38]:/.spread0
     32054 ±502%    +981.6%     346689 ± 39%  sched_debug.cfs_rq[39]:/.spread0
      1.00 ±100%    +100.0%       2.00 ± 50%  sched_debug.cfs_rq[42]:/.nr_spread_over
   -125980 ±-218%    -307.1%     260875 ± 46%  sched_debug.cfs_rq[42]:/.spread0
   -111501 ±-102%    -288.7%     210354 ± 94%  sched_debug.cfs_rq[45]:/.spread0
   -173363 ±-34%    -221.0%     209775 ± 94%  sched_debug.cfs_rq[47]:/.spread0
   -302090 ±-43%    -121.8%      65953 ±322%  sched_debug.cfs_rq[4]:/.spread0
   -490175 ±-18%     -41.1%    -288722 ±-31%  sched_debug.cfs_rq[50]:/.spread0
   -594948 ±-10%     -59.7%    -239840 ±-33%  sched_debug.cfs_rq[51]:/.spread0
      1.00 ±100%   +6050.0%      61.50 ±141%  sched_debug.cfs_rq[53]:/.blocked_load_avg
     10.50 ±  8%    +614.3%      75.00 ±122%  sched_debug.cfs_rq[53]:/.tg_load_contrib
   -596043 ±-10%     -49.0%    -304277 ±-36%  sched_debug.cfs_rq[54]:/.spread0
     10.00 ±  0%   +2062.5%     216.25 ± 40%  sched_debug.cfs_rq[56]:/.tg_load_contrib
     17.75 ±173%   +1302.8%     249.00 ± 26%  sched_debug.cfs_rq[60]:/.blocked_load_avg
   -809633 ± -9%     -36.2%    -516886 ±-23%  sched_debug.cfs_rq[60]:/.spread0
     28.00 ±109%    +828.6%     260.00 ± 25%  sched_debug.cfs_rq[60]:/.tg_load_contrib
    277.75 ± 95%     -86.3%      38.00 ±171%  sched_debug.cfs_rq[7]:/.blocked_load_avg
    293.25 ± 90%     -81.8%      53.50 ±121%  sched_debug.cfs_rq[7]:/.tg_load_contrib
     17.50 ±  2%     -28.6%      12.50 ± 34%  sched_debug.cpu#0.cpu_load[2]
     17.00 ±  4%     -25.0%      12.75 ± 35%  sched_debug.cpu#0.cpu_load[3]
      2907 ± 12%    +195.9%       8603 ± 63%  sched_debug.cpu#0.sched_goidle
     16.50 ±  3%      -9.1%      15.00 ±  0%  sched_debug.cpu#1.cpu_load[2]
     16.50 ±  3%      -7.6%      15.25 ±  2%  sched_debug.cpu#1.cpu_load[3]
      5595 ± 26%     -36.4%       3557 ± 11%  sched_debug.cpu#11.nr_switches
      6885 ± 92%     -76.2%       1639 ± 40%  sched_debug.cpu#11.ttwu_count
      1350 ± 34%     -55.0%     608.00 ± 14%  sched_debug.cpu#11.ttwu_local
     17892 ± 74%     -78.3%       3877 ± 18%  sched_debug.cpu#12.nr_switches
      1288 ± 27%     -49.8%     647.50 ± 37%  sched_debug.cpu#12.ttwu_local
      1405 ± 22%     -52.7%     664.50 ± 23%  sched_debug.cpu#13.ttwu_local
      1.25 ±182%    -440.0%      -4.25 ±-50%  sched_debug.cpu#17.nr_uninterruptible
      1976 ±  5%     -10.0%       1779 ±  0%  sched_debug.cpu#18.curr->pid
    983.75 ±  8%    +101.6%       1983 ± 32%  sched_debug.cpu#18.ttwu_local
     -0.25 ±-911%   +2300.0%      -6.00 ±-28%  sched_debug.cpu#21.nr_uninterruptible
      2979 ± 49%    +159.6%       7734 ± 75%  sched_debug.cpu#22.ttwu_count
      1111 ± 21%    +127.6%       2528 ± 32%  sched_debug.cpu#22.ttwu_local
      1.00 ±141%    -275.0%      -1.75 ±-84%  sched_debug.cpu#25.nr_uninterruptible
     14419 ± 54%     -58.2%       6022 ± 84%  sched_debug.cpu#25.ttwu_count
     14395 ± 70%    +252.4%      50729 ± 39%  sched_debug.cpu#28.nr_switches
     -4.75 ±-17%    -115.8%       0.75 ±218%  sched_debug.cpu#30.nr_uninterruptible
      2335 ±115%     -76.6%     547.25 ± 18%  sched_debug.cpu#34.ttwu_count
      1258 ± 25%     -43.3%     713.75 ± 11%  sched_debug.cpu#35.nr_switches
      1409 ± 23%     -39.6%     851.75 ±  9%  sched_debug.cpu#35.sched_count
    969.50 ± 69%     -68.8%     302.00 ± 38%  sched_debug.cpu#35.ttwu_count
    382.00 ± 37%     -66.0%     130.00 ± 14%  sched_debug.cpu#35.ttwu_local
    808.75 ± 18%     +28.3%       1037 ± 15%  sched_debug.cpu#38.nr_switches
    948.50 ± 16%     +23.2%       1168 ± 13%  sched_debug.cpu#38.sched_count
     70695 ±  2%      +6.2%      75047 ±  4%  sched_debug.cpu#41.nr_load_updates
      1269 ± 13%     +55.3%       1970 ± 25%  sched_debug.cpu#46.nr_switches
      3.25 ± 93%     -76.9%       0.75 ±197%  sched_debug.cpu#46.nr_uninterruptible
      1375 ± 12%     +51.1%       2078 ± 23%  sched_debug.cpu#46.sched_count
      3958 ± 97%    +462.9%      22281 ± 25%  sched_debug.cpu#50.ttwu_count
    457.25 ± 26%     +64.3%     751.25 ± 28%  sched_debug.cpu#53.ttwu_local
    753041 ±  3%     -11.1%     669815 ±  5%  sched_debug.cpu#58.avg_idle
     -1.75 ±-142%    -257.1%       2.75 ± 64%  sched_debug.cpu#59.nr_uninterruptible
      2581 ± 27%   +1426.4%      39408 ± 57%  sched_debug.cpu#60.nr_switches
      2632 ± 27%   +1400.2%      39495 ± 57%  sched_debug.cpu#60.sched_count
     34156 ± 94%     -94.8%       1776 ± 15%  sched_debug.cpu#61.nr_switches
     34250 ± 94%     -94.7%       1825 ± 15%  sched_debug.cpu#61.sched_count
     16821 ± 96%     -95.4%     768.50 ± 11%  sched_debug.cpu#61.sched_goidle
      8128 ±146%     -91.7%     676.00 ± 10%  sched_debug.cpu#61.ttwu_count

=========================================================================================
tbox_group/testcase/rootfs/kconfig/compiler/cpufreq_governor/test:
  ivb42/will-it-scale/debian-x86_64-2015-02-07.cgz/x86_64-rhel/gcc-4.9/performance/readseek1

commit: 
  447f6a95a9c80da7faaec3e66e656eab8f262640
  8f2f3eb59dff4ec538de55f2e0592fec85966aab

447f6a95a9c80da7 8f2f3eb59dff4ec538de55f2e0 
---------------- -------------------------- 
         %stddev     %change         %stddev
             \          |                \  
   1915464 ±  0%      -2.4%    1869344 ±  0%  will-it-scale.per_thread_ops
    473.17 ±  0%      +6.9%     505.66 ±  0%  will-it-scale.time.user_time
      0.20 ±  5%     -49.4%       0.10 ± 35%  turbostat.Pkg%pc6
      3.38 ±  0%     +34.0%       4.53 ±  1%  perf-profile.cpu-cycles.find_get_entry.find_lock_entry.shmem_getpage_gfp.shmem_file_read_iter.__vfs_read
      7.42 ±  0%     +16.3%       8.62 ±  1%  perf-profile.cpu-cycles.find_lock_entry.shmem_getpage_gfp.shmem_file_read_iter.__vfs_read.vfs_read
      0.57 ±  6%     +72.2%       0.99 ±  6%  perf-profile.cpu-cycles.radix_tree_lookup_slot.find_get_entry.find_lock_entry.shmem_getpage_gfp.shmem_file_read_iter
     10.58 ±  0%     +11.4%      11.79 ±  1%  perf-profile.cpu-cycles.shmem_getpage_gfp.shmem_file_read_iter.__vfs_read.vfs_read.sys_read
     30.50 ±150%   +1140.2%     378.25 ± 49%  sched_debug.cfs_rq[22]:/.blocked_load_avg
     44.75 ±103%    +788.3%     397.50 ± 46%  sched_debug.cfs_rq[22]:/.tg_load_contrib
     89.50 ±159%    +300.3%     358.25 ± 75%  sched_debug.cfs_rq[2]:/.blocked_load_avg
    115.75 ±123%    +231.7%     384.00 ± 70%  sched_debug.cfs_rq[2]:/.tg_load_contrib
      0.50 ±100%    +750.0%       4.25 ± 67%  sched_debug.cfs_rq[32]:/.nr_spread_over
    499.50 ± 44%     -98.2%       9.00 ±101%  sched_debug.cfs_rq[40]:/.blocked_load_avg
    505.50 ± 44%     -95.2%      24.50 ± 73%  sched_debug.cfs_rq[40]:/.tg_load_contrib
    421.00 ± 56%     -85.7%      60.25 ±109%  sched_debug.cfs_rq[42]:/.blocked_load_avg
    428.75 ± 56%     -80.4%      84.00 ± 86%  sched_debug.cfs_rq[42]:/.tg_load_contrib
      8053 ±  2%     +13.4%       9132 ±  5%  sched_debug.cfs_rq[47]:/.avg->runnable_avg_sum
    175.25 ±  2%     +12.7%     197.50 ±  5%  sched_debug.cfs_rq[47]:/.tg_runnable_contrib
      0.25 ±173%   +1500.0%       4.00 ± 77%  sched_debug.cfs_rq[8]:/.nr_spread_over
     90.75 ± 13%     -23.1%      69.75 ± 15%  sched_debug.cpu#0.cpu_load[2]
     97.00 ± 15%     -28.4%      69.50 ± 16%  sched_debug.cpu#0.cpu_load[3]
     99.50 ± 14%     -27.6%      72.00 ± 18%  sched_debug.cpu#0.cpu_load[4]
    -10.25 ±-14%     -73.2%      -2.75 ±-180%  sched_debug.cpu#1.nr_uninterruptible
      8173 ±106%     -78.9%       1722 ± 35%  sched_debug.cpu#10.nr_switches
      3896 ±112%     -81.3%     727.50 ± 36%  sched_debug.cpu#10.sched_goidle
    515.00 ± 40%     -47.2%     271.75 ± 49%  sched_debug.cpu#10.ttwu_local
      2.00 ± 81%    -325.0%      -4.50 ±-77%  sched_debug.cpu#11.nr_uninterruptible
      3818 ± 39%     -58.2%       1598 ± 68%  sched_debug.cpu#15.ttwu_local
      0.50 ±331%    -650.0%      -2.75 ±-74%  sched_debug.cpu#16.nr_uninterruptible
     12671 ± 30%     -58.4%       5270 ± 46%  sched_debug.cpu#20.ttwu_count
      2285 ± 70%     -57.0%     983.50 ± 25%  sched_debug.cpu#20.ttwu_local
      2722 ± 79%     -72.9%     738.75 ± 51%  sched_debug.cpu#21.ttwu_local
     -2.50 ±-72%    -200.0%       2.50 ± 82%  sched_debug.cpu#23.nr_uninterruptible
      1183 ± 31%    +188.4%       3413 ± 22%  sched_debug.cpu#24.nr_switches
      1384 ± 45%    +148.4%       3438 ± 22%  sched_debug.cpu#24.sched_count
    318.50 ± 54%    +347.5%       1425 ± 21%  sched_debug.cpu#24.ttwu_local
      5255 ± 46%     -60.2%       2090 ± 54%  sched_debug.cpu#25.nr_switches
      5276 ± 46%     -59.9%       2114 ± 54%  sched_debug.cpu#25.sched_count
      1893 ± 42%     -66.9%     627.00 ± 75%  sched_debug.cpu#25.ttwu_local
      1.25 ±142%    +240.0%       4.25 ± 45%  sched_debug.cpu#27.nr_uninterruptible
      0.75 ±272%    -322.2%      -1.67 ±-28%  sched_debug.cpu#31.nr_uninterruptible
      1977 ±140%     -86.5%     267.25 ± 10%  sched_debug.cpu#32.sched_goidle
      7.67 ± 78%    -122.8%      -1.75 ±-84%  sched_debug.cpu#34.nr_uninterruptible
      3642 ± 37%    +205.0%      11108 ± 53%  sched_debug.cpu#39.nr_switches
      1250 ± 51%    +292.0%       4902 ± 52%  sched_debug.cpu#39.sched_goidle
      3.00 ±  0%    +216.7%       9.50 ± 30%  sched_debug.cpu#45.cpu_load[0]
      3.50 ± 24%    +121.4%       7.75 ± 10%  sched_debug.cpu#45.cpu_load[1]
      3.25 ± 13%    +123.1%       7.25 ± 11%  sched_debug.cpu#45.cpu_load[2]
      3.25 ± 13%     +92.3%       6.25 ± 23%  sched_debug.cpu#45.cpu_load[3]
      3.00 ±  0%     +91.7%       5.75 ± 22%  sched_debug.cpu#45.cpu_load[4]
      1593 ± 19%     +63.6%       2605 ± 30%  sched_debug.cpu#47.curr->pid
    365.75 ± 39%    +254.6%       1297 ± 98%  sched_debug.cpu#6.ttwu_local
      8717 ± 80%     -78.7%       1856 ± 45%  sched_debug.cpu#8.nr_switches
      3992 ± 85%     -80.5%     778.50 ± 51%  sched_debug.cpu#8.sched_goidle
      6221 ±128%     -83.9%     998.75 ± 44%  sched_debug.cpu#8.ttwu_count
    722.00 ± 71%     -69.5%     220.25 ±  5%  sched_debug.cpu#8.ttwu_local
      0.25 ±173%    +321.4%       1.05 ±  5%  sched_debug.rt_rq[12]:/.rt_time
      0.04 ±173%    +311.0%       0.17 ±  8%  sched_debug.rt_rq[13]:/.rt_time


lkp-sbx04: Sandy Bridge-EX
Memory: 64G

ivb42: Ivytown Ivy Bridge-EP
Memory: 64G




                           will-it-scale.time.user_time

  325 ++--------------------------------------------------------------------+
  320 ++      O       O          O                                          |
      |   O              O   O       O   O   O   O   O                      |
  315 O+          O                                             O           |
  310 ++                                                 O  O           O   O
      |                                                             O       |
  305 ++                                                                    |
  300 ++                                                                    |
  295 ++                                                                    |
      |                                      *..                            |
  290 ++         .*..          ..*...*..   ..   .                           |
  285 *+..*... ..    .  .*...*.         . .      *...*...*..*...*...  ..*.. |
      |       *       *.                 *                          *.     .|
  280 ++                                                                    *
  275 ++--------------------------------------------------------------------+


	[*] bisect-good sample
	[O] bisect-bad  sample

To reproduce:

        git clone git://git.kernel.org/pub/scm/linux/kernel/git/wfg/lkp-tests.git
        cd lkp-tests
        bin/lkp install job.yaml  # job file is attached in this email
        bin/lkp run     job.yaml


Disclaimer:
Results have been estimated based on internal Intel analysis and are provided
for informational purposes only. Any difference in system hardware or software
design or configuration may affect actual performance.


Thanks,
Ying Huang

View attachment "job.yaml" of type "text/plain" (3209 bytes)

View attachment "reproduce" of type "text/plain" (4773 bytes)

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ