lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite for Android: free password hash cracker in your pocket
[<prev] [next>] [day] [month] [year] [list]
Date:	Mon, 16 Feb 2015 13:11:28 +0800
From:	Huang Ying <ying.huang@...el.com>
To:	Ebru Akagunduz <ebru.akagunduz@...il.com>
Cc:	Linus Torvalds <torvalds@...ux-foundation.org>,
	LKML <linux-kernel@...r.kernel.org>, LKP ML <lkp@...org>
Subject: [LKP] [mm] 10359213d05: -3.1% pigz.time.minor_page_faults, +31.3%
 meminfo.AnonHugePages, +1.7% will-it-scale.per_thread_ops

FYI, we noticed the below changes on

git://git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux.git master
commit 10359213d05acf804558bda7cc9b8422a828d1cd ("mm: incorporate read-only pages into transparent huge pages")


testbox/testcase/testparams: xps2/pigz/performance-100%-512K

ba4877b9ca51f80b  10359213d05acf804558bda7cc  
----------------  --------------------------  
       fail:runs  %reproduction    fail:runs
           |             |             |    
         %stddev     %change         %stddev
             \          |                \  
     25068 ±  3%      +8.1%      27104 ±  3%  pigz.time.maximum_resident_set_size
     27037 ±  0%      -3.1%      26197 ±  1%  pigz.time.minor_page_faults
     12558 ±  2%     +31.3%      16493 ±  3%  meminfo.AnonHugePages
      0.92 ± 30%     -45.1%       0.51 ± 25%  perf-profile.cpu-cycles.update_curr.task_tick_fair.scheduler_tick.update_process_times.tick_sched_handle
      1.12 ±  3%     +42.1%       1.59 ± 15%  perf-profile.cpu-cycles.ret_from_fork
      2.46 ± 21%     -29.4%       1.74 ± 18%  perf-profile.cpu-cycles.get_page_from_freelist.__alloc_pages_nodemask.alloc_pages_current.pipe_write.new_sync_write
    595245 ± 11%     -26.6%     436928 ± 23%  sched_debug.cpu#3.avg_idle
    602523 ±  9%     -25.1%     451132 ± 15%  sched_debug.cpu#5.avg_idle
     16.26 ±  9%     -16.0%      13.66 ±  4%  perf-profile.cpu-cycles.tick_program_event.hrtimer_interrupt.hpet_interrupt_handler.handle_irq_event_percpu.handle_irq_event
      0.81 ± 24%     +30.2%       1.05 ±  5%  perf-profile.cpu-cycles.__do_softirq.irq_exit.smp_apic_timer_interrupt.apic_timer_interrupt
      1.68 ±  5%     -22.2%       1.30 ± 22%  perf-profile.cpu-cycles.copy_page_from_iter_iovec.copy_page_from_iter.pipe_write.new_sync_write.vfs_write
       150 ±  2%     -11.8%        132 ±  8%  sched_debug.cfs_rq[3]:/.load
      7785 ±  6%     +11.4%       8675 ±  6%  slabinfo.kmalloc-64.active_objs
      1986 ±  3%     -12.4%       1739 ±  4%  slabinfo.kmalloc-256.active_objs
      8048 ±  4%      +9.7%       8826 ±  5%  slabinfo.kmalloc-64.num_objs
       321 ± 28%     +61.6%        520 ± 28%  sched_debug.cfs_rq[2]:/.blocked_load_avg
       139 ±  3%     +14.0%        159 ±  7%  sched_debug.cpu#1.load
       121 ±  2%      +8.2%        131 ±  3%  sched_debug.cpu#0.cpu_load[4]
       122 ±  2%      +9.2%        134 ±  3%  sched_debug.cpu#0.cpu_load[3]
   1014549 ±  8%      -9.8%     914927 ±  4%  cpuidle.C1E-NHM.time
     25068 ±  3%      +8.1%      27104 ±  3%  time.maximum_resident_set_size
    143231 ±  9%     +11.4%     159600 ±  2%  softirqs.RCU
       150 ±  2%     -12.1%        132 ±  8%  sched_debug.cpu#3.load

testbox/testcase/testparams: lkp-sbx04/will-it-scale/performance-malloc1

ba4877b9ca51f80b  10359213d05acf804558bda7cc  
----------------  --------------------------  
   3954711 ±  4%     +11.8%    4420591 ±  8%  will-it-scale.time.voluntary_context_switches
   7468699 ±  1%      +3.4%    7722740 ±  1%  will-it-scale.time.minor_page_faults
      0.02 ±  0%      -3.7%       0.02 ±  0%  will-it-scale.scalability
     37784 ±  0%      +1.7%      38421 ±  0%  will-it-scale.per_thread_ops
      0.00 ±  0%  +5.8e+12%   58403.14 ± 29%  sched_debug.cfs_rq[49]:/.MIN_vruntime
      0.00 ±  0%  +5.8e+12%   58403.14 ± 29%  sched_debug.cfs_rq[49]:/.max_vruntime
         5 ± 24%    +340.0%         22 ± 13%  sched_debug.cpu#59.cpu_load[1]
         4 ± 17%    +310.5%         19 ± 17%  sched_debug.cpu#59.cpu_load[0]
        34 ± 29%    +198.5%        101 ± 31%  sched_debug.cfs_rq[33]:/.tg_load_contrib
         5 ± 24%    +280.0%         19 ± 15%  sched_debug.cpu#59.cpu_load[2]
        19 ± 41%     -79.5%          4 ± 50%  sched_debug.cfs_rq[26]:/.load
        19 ± 41%     -79.5%          4 ± 50%  sched_debug.cpu#26.load
     33117 ± 39%    +159.0%      85783 ± 15%  sched_debug.cpu#4.sched_goidle
         4 ± 19%    +211.1%         14 ± 14%  sched_debug.cpu#59.cpu_load[3]
     34216 ± 41%    +175.2%      94151 ± 18%  sched_debug.cpu#4.ttwu_count
         7 ± 34%    +160.0%         19 ± 17%  sched_debug.cfs_rq[49]:/.load
       124 ± 21%     -78.8%         26 ± 43%  sched_debug.cfs_rq[48]:/.blocked_load_avg
     66389 ± 39%    +158.7%     171750 ± 15%  sched_debug.cpu#4.nr_switches
       135 ± 23%     -74.1%         35 ± 28%  sched_debug.cfs_rq[48]:/.tg_load_contrib
      0.00 ±  1%    +166.2%       0.00 ± 34%  sched_debug.rt_rq[63]:/.rt_time
         5 ± 47%     -63.6%          2 ±  0%  sched_debug.cpu#11.cpu_load[2]
     67305 ± 38%    +155.4%     171876 ± 15%  sched_debug.cpu#4.sched_count
        47 ± 40%     -68.3%         15 ± 20%  sched_debug.cfs_rq[12]:/.tg_load_contrib
        54 ± 20%    +149.8%        135 ± 44%  sched_debug.cfs_rq[45]:/.blocked_load_avg
      0.00 ±  7%     -15.3%       0.00 ±  4%  sched_debug.rt_rq[7]:/.rt_time
         5 ± 37%     -65.2%          2 ±  0%  sched_debug.cpu#3.cpu_load[1]
         1 ± 34%    +100.0%          2 ± 20%  sched_debug.cfs_rq[19]:/.nr_spread_over
       136 ± 42%     -62.3%         51 ±  8%  sched_debug.cfs_rq[43]:/.tg_load_contrib
        67 ± 16%    +124.6%        150 ± 41%  sched_debug.cfs_rq[45]:/.tg_load_contrib
        69 ± 36%    +126.1%        156 ± 21%  sched_debug.cfs_rq[61]:/.blocked_load_avg
         4 ± 10%    +135.3%         10 ± 10%  sched_debug.cpu#59.cpu_load[4]
         7 ± 14%     -74.2%          2 ± 50%  sched_debug.cpu#49.nr_uninterruptible
       877 ± 32%     +64.3%       1442 ± 30%  sched_debug.cpu#22.curr->pid
         7 ± 29%    +151.6%         19 ± 28%  sched_debug.cpu#50.load
         4 ± 22%     -57.9%          2 ±  0%  sched_debug.cpu#3.cpu_load[2]
    347606 ± 36%     -58.7%     143557 ±  1%  sched_debug.cfs_rq[16]:/.min_vruntime
      0.00 ± 39%     -47.9%       0.00 ±  0%  sched_debug.rt_rq[55]:/.rt_time
      7757 ± 45%     -61.5%       2985 ±  5%  sched_debug.cfs_rq[16]:/.exec_clock
        15 ± 36%     -42.9%          9 ±  0%  sched_debug.cpu#0.cpu_load[0]
        77 ± 32%    +120.7%        170 ± 16%  sched_debug.cfs_rq[61]:/.tg_load_contrib
      8767 ±  9%     +70.0%      14901 ± 33%  sched_debug.cfs_rq[1]:/.exec_clock
       406 ± 36%     +35.5%        550 ± 29%  sched_debug.cpu#46.ttwu_local
     97714 ± 37%     +80.7%     176528 ± 19%  sched_debug.cpu#6.nr_switches
     98697 ± 36%     +79.0%     176658 ± 19%  sched_debug.cpu#6.sched_count
        88 ± 14%     -60.0%         35 ± 23%  sched_debug.cfs_rq[62]:/.blocked_load_avg
      0.00 ± 36%     -48.5%       0.00 ± 10%  sched_debug.rt_rq[54]:/.rt_time
         1 ±  0%    +100.0%          2 ±  0%  sched_debug.cfs_rq[53]:/.nr_spread_over
         3 ± 11%     -46.7%          2 ±  0%  sched_debug.cpu#3.cpu_load[3]
      5714 ± 27%     -37.4%       3577 ±  4%  sched_debug.cpu#28.sched_goidle
       745 ± 26%     +48.1%       1104 ± 20%  sched_debug.cpu#15.ttwu_local
     11686 ± 26%     -37.0%       7366 ±  4%  sched_debug.cpu#28.nr_switches
     90749 ± 11%     +54.3%     140005 ± 29%  sched_debug.cpu#32.sched_goidle
    181994 ± 11%     +54.1%     280377 ± 29%  sched_debug.cpu#32.nr_switches
    182550 ± 10%     +53.8%     280771 ± 29%  sched_debug.cpu#32.sched_count
     48790 ± 37%     +80.7%      88182 ± 19%  sched_debug.cpu#6.sched_goidle
  48101669 ± 16%     +58.7%   76337780 ± 27%  cpuidle.C1-SNB.time
     11390 ± 16%     +62.0%      18455 ± 28%  sched_debug.cpu#11.ttwu_count
     50734 ± 38%     +79.4%      91038 ± 16%  sched_debug.cpu#6.ttwu_count
        15 ± 21%     -34.4%         10 ±  0%  sched_debug.cpu#0.cpu_load[1]
        63 ± 39%     +96.1%        125 ± 12%  sched_debug.cfs_rq[58]:/.blocked_load_avg
       409 ±  9%     +47.9%        605 ± 26%  sched_debug.cpu#59.ttwu_local
      1267 ± 23%     +22.4%       1552 ± 25%  numa-meminfo.node3.AnonHugePages
        72 ± 25%     +92.4%        140 ±  3%  sched_debug.cfs_rq[58]:/.tg_load_contrib
      0.00 ± 32%     -43.6%       0.00 ±  1%  sched_debug.rt_rq[32]:/.rt_time
      1147 ± 24%     +44.2%       1654 ± 10%  sched_debug.cpu#25.ttwu_local
         5 ± 14%     -60.0%          2 ± 50%  sched_debug.cpu#51.nr_uninterruptible
        12 ± 42%     +95.9%         24 ±  4%  sched_debug.cfs_rq[50]:/.load
     14738 ±  6%     +53.2%      22584 ±  7%  sched_debug.cpu#12.sched_goidle
     29667 ±  6%     +52.8%      45321 ±  7%  sched_debug.cpu#12.nr_switches
     14586 ±  8%     +54.0%      22463 ±  7%  sched_debug.cpu#12.ttwu_count
      1025 ± 17%     -41.3%        601 ± 10%  sched_debug.cpu#12.ttwu_local
     26504 ± 33%     +57.3%      41688 ± 19%  sched_debug.cpu#9.sched_count
        73 ± 38%     +68.7%        124 ±  3%  sched_debug.cfs_rq[29]:/.tg_load_contrib
       707 ± 20%     +63.1%       1154 ±  8%  numa-meminfo.node3.Shmem
        94 ± 12%     -50.0%         47 ± 34%  sched_debug.cfs_rq[62]:/.tg_load_contrib
       176 ± 20%     +63.2%        288 ±  9%  numa-vmstat.node3.nr_shmem
      1378 ± 15%     +15.9%       1597 ± 21%  sched_debug.cpu#1.curr->pid
        69 ± 48%     +77.5%        122 ±  2%  sched_debug.cfs_rq[29]:/.blocked_load_avg
       692 ±  2%     +25.0%        865 ± 21%  sched_debug.cpu#30.curr->pid
        17 ± 33%     +52.1%         27 ± 18%  sched_debug.cpu#46.load
         3 ±  0%     -33.3%          2 ±  0%  sched_debug.cpu#3.cpu_load[4]
    390329 ±  0%     +24.1%     484415 ± 19%  sched_debug.cfs_rq[1]:/.min_vruntime
       635 ± 20%     +65.9%       1054 ±  5%  numa-meminfo.node3.Inactive(anon)
       158 ± 20%     +66.2%        263 ±  5%  numa-vmstat.node3.nr_inactive_anon
       140 ±  4%     +26.7%        178 ± 16%  sched_debug.cfs_rq[1]:/.tg_runnable_contrib
      6553 ± 19%     -31.8%       4467 ± 17%  sched_debug.cpu#26.sched_goidle
      6476 ±  4%     +26.5%       8195 ± 16%  sched_debug.cfs_rq[1]:/.avg->runnable_avg_sum
       719 ± 31%     +46.9%       1056 ±  3%  sched_debug.cpu#16.ttwu_local
    278733 ± 11%     +28.3%     357584 ± 11%  cpuidle.C1E-SNB.usage
     12547 ± 10%     +35.0%      16933 ±  6%  sched_debug.cpu#58.sched_count
      6341 ± 10%     +35.5%       8592 ±  2%  sched_debug.cpu#58.ttwu_count
    139557 ± 14%     +18.4%     165256 ± 10%  sched_debug.cfs_rq[59]:/.spread0
      1767 ±  3%     +29.4%       2286 ±  3%  sched_debug.cpu#61.curr->pid
    136520 ± 13%     +16.7%     159351 ±  9%  sched_debug.cfs_rq[63]:/.spread0
    142974 ±  4%     +16.3%     166215 ± 14%  sched_debug.cfs_rq[25]:/.min_vruntime
     13231 ± 23%     +29.8%      17174 ± 12%  sched_debug.cpu#16.sched_count
       113 ± 35%     -34.5%         74 ±  0%  sched_debug.cfs_rq[25]:/.tg_load_contrib
        23 ± 19%     +38.3%         32 ± 10%  sched_debug.cfs_rq[37]:/.load
      1359 ± 21%     -40.1%        814 ± 16%  sched_debug.cpu#16.curr->pid
    141008 ± 16%     +16.6%     164377 ±  9%  sched_debug.cfs_rq[61]:/.spread0
       107 ± 32%     -31.8%         73 ±  1%  sched_debug.cfs_rq[25]:/.blocked_load_avg
     39693 ±  7%     -20.7%      31477 ±  0%  sched_debug.cpu#44.ttwu_count
       960 ± 14%     +38.2%       1327 ±  1%  sched_debug.cpu#22.ttwu_local
       827 ± 11%     +20.2%        995 ±  9%  sched_debug.cpu#23.curr->pid
      1819 ±  5%     +18.8%       2162 ±  7%  sched_debug.cpu#60.curr->pid
      6671 ± 13%     +24.4%       8297 ±  8%  sched_debug.cpu#60.sched_goidle
     13517 ± 13%     +24.2%      16793 ±  8%  sched_debug.cpu#60.nr_switches
     13658 ± 12%     +24.0%      16934 ±  8%  sched_debug.cpu#60.sched_count
     39297 ±  8%     -19.8%      31521 ±  1%  sched_debug.cpu#44.sched_goidle
       835 ± 13%     +30.6%       1090 ±  1%  sched_debug.cpu#18.ttwu_local
     78817 ±  8%     -19.8%      63241 ±  1%  sched_debug.cpu#44.nr_switches
     79478 ±  8%     -20.0%      63595 ±  1%  sched_debug.cpu#44.sched_count
        14 ± 10%     -25.4%         11 ±  9%  sched_debug.cpu#0.cpu_load[2]
      1.43 ±  1%     -20.6%       1.14 ±  2%  perf-profile.cpu-cycles.up_write.vm_munmap.sys_munmap.system_call_fastpath
      6092 ± 10%     +28.1%       7803 ±  0%  sched_debug.cpu#58.sched_goidle
       348 ± 13%     +18.7%        413 ± 10%  sched_debug.cpu#6.ttwu_local
     12410 ± 10%     +27.3%      15796 ±  0%  sched_debug.cpu#58.nr_switches
      7012 ± 15%     +25.7%       8813 ±  7%  sched_debug.cpu#60.ttwu_count
       377 ± 11%     +13.5%        428 ±  7%  sched_debug.cpu#60.ttwu_local
      9050 ±  6%     +15.8%      10476 ±  1%  sched_debug.cfs_rq[4]:/.exec_clock
      1762 ±  2%      +8.1%       1906 ±  8%  sched_debug.cpu#62.curr->pid
   3954711 ±  4%     +11.8%    4420591 ±  8%  time.voluntary_context_switches
      2045 ±  5%     +11.9%       2290 ±  7%  numa-meminfo.node3.KernelStack
        14 ±  8%     -14.3%         12 ±  0%  sched_debug.cpu#43.cpu_load[4]
       371 ±  1%     -17.0%        308 ±  5%  sched_debug.cpu#44.ttwu_local
        15 ± 15%     -20.6%         12 ±  4%  sched_debug.cpu#43.cpu_load[3]
        23 ±  7%     -14.0%         20 ±  0%  sched_debug.cpu#34.cpu_load[3]
      0.00 ±  4%     -22.2%       0.00 ± 14%  sched_debug.rt_rq[50]:/.rt_time
    374746 ± 13%     +18.8%     445099 ±  3%  sched_debug.cfs_rq[50]:/.spread0
    650306 ±  5%     +13.4%     737674 ±  2%  sched_debug.cpu#33.avg_idle
      2796 ±  5%      +3.9%       2905 ±  6%  sched_debug.cpu#36.curr->pid
       388 ±  7%     +14.6%        444 ±  3%  sched_debug.cpu#63.ttwu_local
    641680 ±  2%      +9.2%     700581 ±  5%  sched_debug.cpu#39.avg_idle
      9736 ±  8%     +13.1%      11014 ±  2%  sched_debug.cfs_rq[6]:/.exec_clock
      3356 ±  2%      +8.6%       3646 ±  3%  sched_debug.cfs_rq[29]:/.avg->runnable_avg_sum
        71 ±  3%      +8.7%         78 ±  3%  sched_debug.cfs_rq[29]:/.tg_runnable_contrib
     16875 ±  3%      +9.5%      18477 ±  2%  slabinfo.vm_area_struct.active_objs
     16875 ±  3%      +9.5%      18477 ±  2%  slabinfo.vm_area_struct.num_objs
      1.17 ±  4%      -9.0%       1.06 ±  1%  perf-profile.cpu-cycles.perf_event_mmap.mmap_region.do_mmap_pgoff.vm_mmap_pgoff.sys_mmap_pgoff
     44305 ±  4%     -16.1%      37164 ±  8%  sched_debug.cpu#43.ttwu_count
      6430 ±  1%      +6.4%       6840 ±  5%  sched_debug.cfs_rq[14]:/.avg->runnable_avg_sum
      2768 ±  3%      +3.7%       2870 ±  5%  sched_debug.cpu#38.curr->pid
      9224 ±  2%      +9.8%      10130 ±  1%  slabinfo.anon_vma.num_objs
      9224 ±  2%      +9.8%      10130 ±  1%  slabinfo.anon_vma.active_objs
      1816 ±  4%     +10.1%       2000 ±  2%  sched_debug.cpu#57.curr->pid
     85823 ±  4%     -15.5%      72483 ±  7%  sched_debug.cpu#43.nr_switches
     86191 ±  4%     -15.5%      72848 ±  7%  sched_debug.cpu#43.sched_count
     44120 ± 12%     +18.1%      52091 ±  4%  sched_debug.cpu#1.nr_load_updates
     42455 ±  5%     -14.9%      36139 ±  7%  sched_debug.cpu#43.sched_goidle
     27096 ±  3%     +10.6%      29973 ±  8%  vmstat.system.cs
     18715 ±  0%      +7.0%      20016 ±  5%  vmstat.system.in

testbox/testcase/testparams: snb-drag/phoronix-test-suite/performance-encode-ogg-1.4.1

ba4877b9ca51f80b  10359213d05acf804558bda7cc  
----------------  --------------------------  
     29266 ±  0%     -13.0%      25474 ±  1%  phoronix-test-suite.time.minor_page_faults
       320 ± 35%    +367.6%       1496 ± 44%  sched_debug.cfs_rq[2]:/.tg_load_contrib
        63 ± 35%    +103.1%        129 ± 28%  sched_debug.cpu#2.cpu_load[1]
        68 ± 37%    +148.7%        171 ± 26%  sched_debug.cpu#2.cpu_load[0]
       126 ± 37%     +73.5%        219 ± 38%  sched_debug.cfs_rq[2]:/.runnable_load_avg
       298 ±  3%     -25.5%        222 ± 19%  sched_debug.cfs_rq[3]:/.load
        35 ±  8%     +21.8%         43 ± 10%  sched_debug.cfs_rq[1]:/.nr_spread_over
     29266 ±  0%     -13.0%      25474 ±  1%  time.minor_page_faults

xps2: Nehalem
Memory: 4G

lkp-sbx04: Sandy Bridge-EX
Memory: 64G

snb-drag: Sandy Bridge
Memory: 6G




                             pigz.time.minor_page_faults

  27400 ++------------------------------------------------------------------+
  27200 ++                                 *             **                 |
        | .*             ***    *.*    ** +:   .** .**  +  :     .* *.*  *. *
  27000 **  ***.***.**  +   + **   * .*  *  :**   *   **   *.** *  *   **  *|
  26800 ++            **     *      *       *                  *            |
        |                                                                   |
  26600 ++                                        O                         |
  26400 ++                                    O                             |
  26200 ++ O O    O O    O O  O   O   O OO  O                               |
        O                                                                   |
  26000 +O      O     O   O  O     OO  O        O                           |
  25800 ++    O  O   O         OO                O                          |
        |                                                                   |
  25600 ++  O          O                                                    |
  25400 ++---------------------------------O-O------------------------------+

	[*] bisect-good sample
	[O] bisect-bad  sample

To reproduce:

	apt-get install ruby
	git clone git://git.kernel.org/pub/scm/linux/kernel/git/wfg/lkp-tests.git
	cd lkp-tests
	bin/setup-local job.yaml # the job file attached in this email
	bin/run-local   job.yaml


Disclaimer:
Results have been estimated based on internal Intel analysis and are provided
for informational purposes only. Any difference in system hardware or software
design or configuration may affect actual performance.


Thanks,
Ying Huang


View attachment "job.yaml" of type "text/plain" (1750 bytes)

_______________________________________________
LKP mailing list
LKP@...ux.intel.com

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ