lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [day] [month] [year] [list]
Message-ID: <1427696989.17170.83.camel@intel.com>
Date:	Mon, 30 Mar 2015 14:29:49 +0800
From:	Huang Ying <ying.huang@...el.com>
To:	Trond Myklebust <trond.myklebust@...marydata.com>
Cc:	LKML <linux-kernel@...r.kernel.org>, LKP ML <lkp@...org>
Subject: [LKP] [NFS] a08a8cd375d: +1.5% will-it-scale.per_process_ops

FYI, we noticed the below changes on

git://git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux.git master
commit a08a8cd375db9769588257e7782f6b6b68561b88 ("NFS: Add attribute update barriers to NFS writebacks")


testbox/testcase/testparams: ivb42/will-it-scale/performance-pthread_mutex2

f5062003465c20cf  a08a8cd375db9769588257e778  
----------------  --------------------------  
         %stddev     %change         %stddev
             \          |                \  
 2.012e+08 ±  0%      +1.5%  2.042e+08 ±  0%  will-it-scale.per_process_ops
      3.81 ±  0%      +1.6%       3.88 ±  0%  turbostat.RAMWatt
       387 ±  7%     +13.7%        441 ±  6%  numa-vmstat.node0.nr_page_table_pages
      5639 ± 15%     -15.1%       4788 ±  0%  meminfo.AnonHugePages
      1552 ±  7%     +13.7%       1765 ±  6%  numa-meminfo.node0.PageTables
     29.40 ± 21%     +57.6%      46.32 ±  3%  perf-profile.cpu-cycles.start_secondary
     10.72 ± 30%    -100.0%       0.00 ±  0%  perf-profile.cpu-cycles.rest_init.start_kernel.x86_64_start_reservations.x86_64_start_kernel
     10.72 ± 30%    -100.0%       0.00 ±  0%  perf-profile.cpu-cycles.cpu_startup_entry.rest_init.start_kernel.x86_64_start_reservations.x86_64_start_kernel
      1.23 ± 17%     -29.1%       0.87 ± 19%  perf-profile.cpu-cycles.cmd_record._start.main.__libc_start_main
      1.27 ± 17%     -29.4%       0.89 ± 16%  perf-profile.cpu-cycles.main.__libc_start_main
      1.27 ± 17%     -31.2%       0.87 ± 19%  perf-profile.cpu-cycles._start.main.__libc_start_main
     25.76 ± 19%     +66.1%      42.79 ±  5%  perf-profile.cpu-cycles.cpuidle_enter.cpu_startup_entry.start_secondary
      1.39 ± 16%     -30.3%       0.97 ± 11%  perf-profile.cpu-cycles.__libc_start_main
      1.62 ± 20%     +52.2%       2.47 ± 18%  perf-profile.cpu-cycles.ktime_get_update_offsets_now.hrtimer_interrupt.local_apic_timer_interrupt.smp_apic_timer_interrupt.apic_timer_interrupt
     28.79 ± 21%     +60.5%      46.22 ±  3%  perf-profile.cpu-cycles.cpu_startup_entry.start_secondary
      0.00 ±  0%      +Inf%       0.81 ± 36%  perf-profile.cpu-cycles._raw_spin_unlock_irqrestore.update_blocked_averages.rebalance_domains.run_rebalance_domains.__do_softirq
     10.72 ± 30%    -100.0%       0.00 ±  0%  perf-profile.cpu-cycles.x86_64_start_kernel
     18.38 ± 16%     +43.4%      26.35 ±  3%  perf-profile.cpu-cycles.intel_idle.cpuidle_enter_state.cpuidle_enter.cpu_startup_entry.start_secondary
      1.64 ± 30%     -59.1%       0.67 ± 46%  perf-profile.cpu-cycles.filemap_map_pages.handle_pte_fault.handle_mm_fault.__do_page_fault.do_page_fault
     18.60 ± 17%     +42.4%      26.49 ±  4%  perf-profile.cpu-cycles.cpuidle_enter_state.cpuidle_enter.cpu_startup_entry.start_secondary
     10.42 ± 30%    -100.0%       0.00 ±  0%  perf-profile.cpu-cycles.cpuidle_enter.cpu_startup_entry.rest_init.start_kernel.x86_64_start_reservations
      9.45 ± 34%    -100.0%       0.00 ±  0%  perf-profile.cpu-cycles.smp_reschedule_interrupt.reschedule_interrupt.cpuidle_enter.cpu_startup_entry.rest_init
     10.72 ± 30%    -100.0%       0.00 ±  0%  perf-profile.cpu-cycles.x86_64_start_reservations.x86_64_start_kernel
     13.90 ±  7%     -12.0%      12.22 ±  4%  perf-profile.cpu-cycles.update_process_times.tick_sched_handle.tick_sched_timer.__run_hrtimer.hrtimer_interrupt
     13.92 ±  2%     -22.7%      10.77 ± 18%  perf-profile.cpu-cycles.__do_softirq.irq_exit.scheduler_ipi.smp_reschedule_interrupt.reschedule_interrupt
     10.72 ± 30%    -100.0%       0.00 ±  0%  perf-profile.cpu-cycles.start_kernel.x86_64_start_reservations.x86_64_start_kernel
      9.45 ± 34%    -100.0%       0.00 ±  0%  perf-profile.cpu-cycles.reschedule_interrupt.cpuidle_enter.cpu_startup_entry.rest_init.start_kernel
   1438772 ±  8%     +21.2%    1743708 ±  2%  sched_debug.cfs_rq[0]:/.min_vruntime
     25049 ±  3%     +26.4%      31672 ±  1%  sched_debug.cfs_rq[0]:/.avg->runnable_avg_sum
        17 ± 12%    +121.4%         38 ± 19%  sched_debug.cfs_rq[0]:/.runnable_load_avg
       545 ±  3%     +26.4%        689 ±  2%  sched_debug.cfs_rq[0]:/.tg_runnable_contrib
        48 ± 47%    +147.7%        120 ± 30%  sched_debug.cfs_rq[0]:/.tg_load_contrib
     53474 ±  7%     +63.0%      87186 ±  2%  sched_debug.cfs_rq[0]:/.exec_clock
        61 ±  4%     -43.0%         34 ±  9%  sched_debug.cfs_rq[24]:/.runnable_load_avg
   2241502 ±  4%     -17.8%    1843317 ±  1%  sched_debug.cfs_rq[24]:/.min_vruntime
    802666 ± 24%     -87.6%      99533 ± 40%  sched_debug.cfs_rq[24]:/.spread0
        99 ± 21%     +51.9%        151 ± 31%  sched_debug.cfs_rq[24]:/.tg_load_contrib
        62 ±  5%     -44.6%         34 ± 10%  sched_debug.cfs_rq[24]:/.load
       855 ±  2%     -17.4%        706 ±  2%  sched_debug.cfs_rq[24]:/.tg_runnable_contrib
     39223 ±  2%     -17.4%      32411 ±  2%  sched_debug.cfs_rq[24]:/.avg->runnable_avg_sum
    129807 ±  3%     -26.0%      96066 ±  2%  sched_debug.cfs_rq[24]:/.exec_clock
         1 ±  0%    +300.0%          4 ± 50%  sched_debug.cfs_rq[39]:/.nr_spread_over
         3 ± 33%     -66.7%          1 ±  0%  sched_debug.cfs_rq[3]:/.nr_spread_over
        17 ± 12%    +122.9%         39 ± 18%  sched_debug.cpu#0.cpu_load[0]
     56862 ±  7%     +58.5%      90118 ±  2%  sched_debug.cpu#0.nr_load_updates
      3783 ±  6%     +13.9%       4310 ±  0%  sched_debug.cpu#0.curr->pid
     16708 ± 18%     -18.4%      13637 ±  3%  sched_debug.cpu#0.sched_count
      3907 ± 10%     -24.6%       2945 ± 11%  sched_debug.cpu#0.ttwu_local
      5540 ±  8%     -15.4%       4688 ±  6%  sched_debug.cpu#0.ttwu_count
        19 ± 17%    +182.1%         55 ± 33%  sched_debug.cpu#0.cpu_load[2]
        18 ± 14%    +159.5%         48 ± 29%  sched_debug.cpu#0.cpu_load[1]
        22 ± 27%    +181.8%         62 ± 35%  sched_debug.cpu#0.cpu_load[3]
        24 ± 35%    +175.8%         68 ± 35%  sched_debug.cpu#0.cpu_load[4]
      2253 ± 11%     -23.0%       1734 ±  9%  sched_debug.cpu#0.sched_goidle
      3720 ±  8%     -16.5%       3108 ± 13%  sched_debug.cpu#13.curr->pid
      6862 ± 10%     -32.1%       4659 ± 15%  sched_debug.cpu#15.nr_switches
      3135 ± 11%     -36.6%       1987 ± 15%  sched_debug.cpu#15.sched_goidle
        10 ±  4%     +14.6%         11 ±  7%  sched_debug.cpu#22.cpu_load[4]
        84 ± 16%     -55.2%         38 ± 21%  sched_debug.cpu#24.cpu_load[3]
        94 ± 18%     -55.9%         41 ± 22%  sched_debug.cpu#24.cpu_load[4]
    136876 ±  2%     -20.1%     109402 ±  2%  sched_debug.cpu#24.nr_load_updates
        61 ±  4%     -43.0%         34 ±  9%  sched_debug.cpu#24.cpu_load[0]
        62 ±  5%     -44.6%         34 ± 10%  sched_debug.cpu#24.load
      4589 ±  3%     -10.6%       4101 ±  0%  sched_debug.cpu#24.curr->pid
        65 ±  6%     -45.8%         35 ± 15%  sched_debug.cpu#24.cpu_load[1]
        75 ± 13%     -51.2%         37 ± 20%  sched_debug.cpu#24.cpu_load[2]
       411 ± 29%     +82.3%        749 ± 46%  sched_debug.cpu#26.ttwu_count
       207 ± 26%    +110.2%        436 ± 23%  sched_debug.cpu#32.ttwu_count
      1027 ± 37%     -52.7%        486 ±  2%  sched_debug.cpu#37.ttwu_local
      1550 ± 23%     -41.5%        906 ± 39%  sched_debug.cpu#42.ttwu_count
       773 ± 21%    +152.1%       1950 ± 46%  sched_debug.cpu#45.ttwu_count
      1664 ± 10%    +109.2%       3482 ± 27%  sched_debug.cpu#45.nr_switches
       386 ± 24%    +199.9%       1159 ± 40%  sched_debug.cpu#45.ttwu_local
        10 ±  8%     +59.5%         16 ± 36%  sched_debug.cpu#45.cpu_load[1]
       699 ± 10%     +59.6%       1116 ± 30%  sched_debug.cpu#45.sched_goidle
      1678 ± 10%    +158.7%       4341 ± 48%  sched_debug.cpu#45.sched_count
      2954 ±  4%     +15.7%       3417 ±  6%  sched_debug.cpu#47.curr->pid

ivb42: Ivytown Ivy Bridge-EP
Memory: 64G




                              will-it-scale.per_process_ops

  2.045e+08 ++------------------------O-------------------------------------+
   2.04e+08 ++                           O                         O  O  O  O
            |                   O              O     O                      |
  2.035e+08 O+    O  O             O        O     O     O  O                |
   2.03e+08 ++ O             O                                              |
            |           O  O                                 O  O           |
  2.025e+08 ++                                                              |
   2.02e+08 ++                                                              |
  2.015e+08 ++                                                              |
            |                               *..*..*..*                      |
   2.01e+08 ++               *..     .*.. ..                                |
  2.005e+08 ++ *..*..*..*.. :     .*.    *                                  |
            |..             :   *.                                          |
      2e+08 *+             *                                                |
  1.995e+08 ++--------------------------------------------------------------+

	[*] bisect-good sample
	[O] bisect-bad  sample

To reproduce:

	apt-get install ruby
	git clone git://git.kernel.org/pub/scm/linux/kernel/git/wfg/lkp-tests.git
	cd lkp-tests
	bin/setup-local job.yaml # the job file attached in this email
	bin/run-local   job.yaml


Disclaimer:
Results have been estimated based on internal Intel analysis and are provided
for informational purposes only. Any difference in system hardware or software
design or configuration may affect actual performance.


Thanks,
Ying Huang


View attachment "job.yaml" of type "text/plain" (1596 bytes)

View attachment "reproduce" of type "text/plain" (3592 bytes)

_______________________________________________
LKP mailing list
LKP@...ux.intel.com

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ