lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [thread-next>] [day] [month] [year] [list]
Date:   Thu, 11 Jun 2020 12:04:53 +0800
From:   kernel test robot <rong.a.chen@...el.com>
To:     Linus Torvalds <torvalds@...ux-foundation.org>
Cc:     Jann Horn <jannh@...gle.com>, Christoph Hellwig <hch@....de>,
        Oleg Nesterov <oleg@...hat.com>,
        Kirill Shutemov <kirill@...temov.name>,
        Jan Kara <jack@...e.cz>,
        Andrea Arcangeli <aarcange@...hat.com>,
        Matthew Wilcox <willy@...radead.org>,
        LKML <linux-kernel@...r.kernel.org>, lkp@...ts.01.org
Subject: [gup] 17839856fd: stress-ng.vm-splice.ops_per_sec 2158.6% improvement

Greeting,

FYI, we noticed a 2158.6% improvement of stress-ng.vm-splice.ops_per_sec due to commit:


commit: 17839856fd588f4ab6b789f482ed3ffd7c403e1f ("gup: document and work around "COW can break either way" issue")
https://git.kernel.org/cgit/linux/kernel/git/torvalds/linux.git master

in testcase: stress-ng
on test machine: 96 threads Intel(R) Xeon(R) Gold 6252 CPU @ 2.10GHz with 192G memory
with following parameters:

	nr_threads: 100%
	disk: 1HDD
	testtime: 30s
	class: pipe
	cpufreq_governor: performance
	ucode: 0x500002c


In addition to that, the commit also has significant impact on the following tests:

+------------------+----------------------------------------------------------------------+
| testcase: change | stress-ng: stress-ng.vm-splice.ops_per_sec 372.8% improvement        |
| test machine     | 96 threads Intel(R) Xeon(R) Gold 6252 CPU @ 2.10GHz with 192G memory |
| test parameters  | class=vm                                                             |
|                  | cpufreq_governor=performance                                         |
|                  | disk=1HDD                                                            |
|                  | fs=ext4                                                              |
|                  | nr_threads=10%                                                       |
|                  | testtime=1s                                                          |
|                  | ucode=0x500002c                                                      |
+------------------+----------------------------------------------------------------------+
| testcase: change | stress-ng: stress-ng.vm-splice.ops_per_sec 2052.7% improvement       |
| test machine     | 96 threads Intel(R) Xeon(R) Gold 6252 CPU @ 2.10GHz with 192G memory |
| test parameters  | class=pipe                                                           |
|                  | cpufreq_governor=performance                                         |
|                  | disk=1HDD                                                            |
|                  | nr_threads=100%                                                      |
|                  | testtime=1s                                                          |
|                  | ucode=0x500002c                                                      |
+------------------+----------------------------------------------------------------------+




Details are as below:
-------------------------------------------------------------------------------------------------->


To reproduce:

        git clone https://github.com/intel/lkp-tests.git
        cd lkp-tests
        bin/lkp install job.yaml  # job file is attached in this email
        bin/lkp run     job.yaml

=========================================================================================
class/compiler/cpufreq_governor/disk/kconfig/nr_threads/rootfs/tbox_group/testcase/testtime/ucode:
  pipe/gcc-9/performance/1HDD/x86_64-rhel-7.6/100%/debian-x86_64-20191114.cgz/lkp-csl-2sp5/stress-ng/30s/0x500002c

commit: 
  f359287765 ("Merge branch 'from-miklos' of git://git.kernel.org/pub/scm/linux/kernel/git/viro/vfs")
  17839856fd ("gup: document and work around "COW can break either way" issue")

f359287765c04711 17839856fd588f4ab6b789f482e 
---------------- --------------------------- 
         %stddev     %change         %stddev
             \          |                \  
     17619            -2.1%      17246        stress-ng.time.system_time
      1529           +24.1%       1898        stress-ng.time.user_time
 1.533e+08         +2158.7%  3.463e+09        stress-ng.vm-splice.ops
   5110873         +2158.6%  1.154e+08        stress-ng.vm-splice.ops_per_sec
     90027 ±  3%     -12.6%      78647 ±  6%  meminfo.PageTables
      7.48            +1.8        9.28        mpstat.cpu.all.usr%
     85.50            -2.0%      83.75        vmstat.cpu.sy
     85.79            -1.8%      84.24        iostat.cpu.system
      7.41           +24.5%       9.22        iostat.cpu.user
      6131 ± 66%    +117.4%      13330 ± 33%  proc-vmstat.numa_hint_faults
      1347 ± 29%    +282.1%       5149 ± 50%  proc-vmstat.numa_hint_faults_local
      1239 ±  9%     +14.4%       1417 ±  5%  slabinfo.kmalloc-rcl-96.active_objs
      1239 ±  9%     +14.4%       1417 ±  5%  slabinfo.kmalloc-rcl-96.num_objs
      0.13 ±  6%      -0.0        0.11 ±  6%  perf-profile.children.cycles-pp.check_preempt_curr
      0.17 ±  6%      +0.0        0.20 ± 10%  perf-profile.children.cycles-pp.rw_verify_area
      0.44 ±  5%      +0.0        0.48 ±  3%  perf-profile.children.cycles-pp.__irqentry_text_start
    339.81 ± 10%     -28.3%     243.75 ± 14%  sched_debug.cfs_rq:/.exec_clock.stddev
      5.95 ± 12%     +24.5%       7.41 ±  6%  sched_debug.cpu.clock.stddev
      5.95 ± 12%     +24.5%       7.41 ±  6%  sched_debug.cpu.clock_task.stddev
     24316 ±  7%     +28.4%      31224 ± 10%  interrupts.CPU15.RES:Rescheduling_interrupts
     38416 ± 26%     -35.0%      24985 ± 23%  interrupts.CPU17.RES:Rescheduling_interrupts
     20612 ± 19%     +85.9%      38315 ± 38%  interrupts.CPU54.RES:Rescheduling_interrupts
     26788 ± 48%     +66.2%      44527 ± 31%  interrupts.CPU57.RES:Rescheduling_interrupts
     21724 ± 15%     +55.8%      33857 ± 23%  interrupts.CPU70.RES:Rescheduling_interrupts
      4845 ± 35%     +59.4%       7723 ±  2%  interrupts.CPU87.NMI:Non-maskable_interrupts
      4845 ± 35%     +59.4%       7723 ±  2%  interrupts.CPU87.PMI:Performance_monitoring_interrupts
      4843 ± 35%     +59.5%       7723 ±  2%  interrupts.CPU89.NMI:Non-maskable_interrupts
      4843 ± 35%     +59.5%       7723 ±  2%  interrupts.CPU89.PMI:Performance_monitoring_interrupts
      4846 ± 35%     +59.3%       7721 ±  2%  interrupts.CPU90.NMI:Non-maskable_interrupts
      4846 ± 35%     +59.3%       7721 ±  2%  interrupts.CPU90.PMI:Performance_monitoring_interrupts
     25860 ± 27%     -27.2%      18820 ± 17%  interrupts.CPU91.RES:Rescheduling_interrupts
      8.59            -9.4%       7.78        perf-stat.i.MPKI
 3.695e+10           +17.1%  4.328e+10        perf-stat.i.branch-instructions
 2.054e+08           +28.4%  2.638e+08        perf-stat.i.branch-misses
     11.31 ±  5%      -2.2        9.07 ±  5%  perf-stat.i.cache-miss-rate%
      3.74           -54.4%       1.70        perf-stat.i.cpi
     54067 ± 66%    +311.0%     222217 ± 18%  perf-stat.i.cycles-between-cache-misses
      0.04 ±  5%      -0.0        0.03 ±  5%  perf-stat.i.dTLB-load-miss-rate%
   6065752 ±  7%     -10.9%    5407417 ±  4%  perf-stat.i.dTLB-load-misses
 4.522e+10           +20.4%  5.444e+10        perf-stat.i.dTLB-loads
 2.697e+10           +20.9%   3.26e+10        perf-stat.i.dTLB-stores
 1.464e+08           +36.3%  1.996e+08 ±  2%  perf-stat.i.iTLB-load-misses
  1.77e+11           +18.5%  2.098e+11        perf-stat.i.instructions
      0.71           +18.5%       0.84        perf-stat.i.ipc
      2.58            -1.9%       2.53        perf-stat.i.metric.GHz
      1131           +18.8%       1344        perf-stat.i.metric.M/sec
     85.95            -2.7       83.28 ±  2%  perf-stat.i.node-load-miss-rate%
     89.18            -0.9       88.25        perf-stat.i.node-store-miss-rate%
      3.29 ±  2%     -17.3%       2.72        perf-stat.overall.MPKI
      0.56            +0.1        0.61        perf-stat.overall.branch-miss-rate%
      1.42           -16.9%       1.18        perf-stat.overall.cpi
      0.01 ±  8%      -0.0        0.01 ±  3%  perf-stat.overall.dTLB-load-miss-rate%
     94.12            +1.5       95.58        perf-stat.overall.iTLB-load-miss-rate%
      1198           -12.7%       1046        perf-stat.overall.instructions-per-iTLB-miss
      0.71           +20.4%       0.85        perf-stat.overall.ipc
  3.62e+10           +18.1%  4.276e+10        perf-stat.ps.branch-instructions
 2.026e+08           +29.2%  2.618e+08        perf-stat.ps.branch-misses
 4.436e+10           +21.4%  5.385e+10        perf-stat.ps.dTLB-loads
 2.644e+10           +21.9%  3.222e+10        perf-stat.ps.dTLB-stores
 1.448e+08           +37.0%  1.983e+08 ±  2%  perf-stat.ps.iTLB-load-misses
 1.734e+11           +19.5%  2.073e+11        perf-stat.ps.instructions
 3.673e+13           +19.5%  4.389e+13        perf-stat.total.instructions
     10749 ±  5%      +7.4%      11548 ±  4%  softirqs.CPU10.RCU
     10512            +8.6%      11414 ±  2%  softirqs.CPU13.RCU
     10476 ±  2%      +9.8%      11503 ±  4%  softirqs.CPU14.RCU
     10790 ±  2%     +10.6%      11938 ±  4%  softirqs.CPU18.RCU
     10721 ±  2%     +10.5%      11842 ±  4%  softirqs.CPU19.RCU
     10754 ±  2%      +9.8%      11804        softirqs.CPU2.RCU
     10792 ±  4%      +9.5%      11819 ±  3%  softirqs.CPU20.RCU
     10722           +10.9%      11893 ±  3%  softirqs.CPU22.RCU
     10968 ±  6%     +10.4%      12110 ±  5%  softirqs.CPU23.RCU
     11012 ±  4%     +22.6%      13503 ± 14%  softirqs.CPU24.RCU
     10584 ±  5%     +12.4%      11899 ±  5%  softirqs.CPU28.RCU
     10512 ±  2%     +13.8%      11966 ±  6%  softirqs.CPU29.RCU
     10826 ±  3%      +8.0%      11696 ±  4%  softirqs.CPU30.RCU
     10482            +8.7%      11390 ±  3%  softirqs.CPU31.RCU
     10400 ±  4%      +8.8%      11315 ±  3%  softirqs.CPU32.RCU
     10356            +9.7%      11361 ±  4%  softirqs.CPU34.RCU
     10193           +10.3%      11239 ±  4%  softirqs.CPU35.RCU
     10159 ±  2%     +11.7%      11346 ±  3%  softirqs.CPU37.RCU
     10262 ±  2%     +10.7%      11360 ±  3%  softirqs.CPU38.RCU
     10077 ±  2%     +10.8%      11163 ±  4%  softirqs.CPU40.RCU
     10204            +9.2%      11142 ±  2%  softirqs.CPU41.RCU
     10523 ±  4%      +6.8%      11238 ±  3%  softirqs.CPU43.RCU
     10015           +10.5%      11064 ±  3%  softirqs.CPU44.RCU
     10376 ±  5%      +8.0%      11202 ±  4%  softirqs.CPU45.RCU
     10164            +9.1%      11090 ±  4%  softirqs.CPU46.RCU
     10176 ±  4%      +8.8%      11071 ±  3%  softirqs.CPU47.RCU
     10739           +10.4%      11853 ±  3%  softirqs.CPU49.RCU
     10569 ±  3%     +14.7%      12120 ±  3%  softirqs.CPU51.RCU
     10627 ±  5%     +12.7%      11976 ±  3%  softirqs.CPU52.RCU
     10507           +12.9%      11858 ±  5%  softirqs.CPU53.RCU
     10413           +10.3%      11482 ±  3%  softirqs.CPU54.RCU
     11050 ±  3%      +6.1%      11720 ±  4%  softirqs.CPU64.RCU
     10663           +10.5%      11781 ±  4%  softirqs.CPU65.RCU
     10461 ±  2%     +10.4%      11544 ±  4%  softirqs.CPU67.RCU
     10565 ±  2%     +11.7%      11804 ±  5%  softirqs.CPU71.RCU
     10200 ±  2%     +22.8%      12529 ± 20%  softirqs.CPU74.RCU
     10078           +23.0%      12396 ± 16%  softirqs.CPU75.RCU
     10182            +9.0%      11101 ±  4%  softirqs.CPU77.RCU
      9994 ±  2%     +12.1%      11201 ±  4%  softirqs.CPU79.RCU
      9848           +11.2%      10954 ±  5%  softirqs.CPU80.RCU
      9802 ±  2%     +12.4%      11017 ±  5%  softirqs.CPU81.RCU
      9924            +9.8%      10893 ±  4%  softirqs.CPU82.RCU
      9921            +9.9%      10906 ±  3%  softirqs.CPU83.RCU
      9905 ±  2%     +10.7%      10963 ±  5%  softirqs.CPU85.RCU
      9901            +8.5%      10746 ±  3%  softirqs.CPU87.RCU
      9953 ±  2%     +10.5%      10997 ±  2%  softirqs.CPU88.RCU
      9944           +11.1%      11047 ±  5%  softirqs.CPU89.RCU
      9986            +8.1%      10796 ±  4%  softirqs.CPU90.RCU
      9796 ±  2%     +10.4%      10812 ±  4%  softirqs.CPU91.RCU
      9750           +10.9%      10816 ±  4%  softirqs.CPU93.RCU
     10018            +9.8%      10997 ±  3%  softirqs.CPU95.RCU
   1009159            +9.1%    1101411 ±  3%  softirqs.RCU


                                                                                
                             stress-ng.time.user_time                           
                                                                                
  2500 +--------------------------------------------------------------------+   
       |                                                                    |   
       |                                                                    |   
  2000 |-O  O O O  O O O O  O O    O   O    O O O  O O O  O O O  O   O    O |   
       |                        O    O                             O   O    |   
       |.+..+.+.  .+.      .+.+.+.. .+.  .+.+.+.  .+.+.+..+. .+..+.+.+.+..  |   
  1500 |-+      +.   +.+.+.        +   +.       +.          +             + |   
       |                                                                    |   
  1000 |-+                                                                  |   
       |                                                                    |   
       |                                                                    |   
   500 |-+                                                                  |   
       |                                                                    |   
       |                                                                    |   
     0 +--------------------------------------------------------------------+   
                                                                                
                                                                                                                                                                
                             stress-ng.time.system_time                         
                                                                                
  18000 +-------------------------------------------------------------------+   
        | O  O O O O  O O O O  O O O    O   O  O O O O  O O O O  O O O O  O |   
  16000 |-+                                                                 |   
  14000 |-+                                                                 |   
        |                                                                   |   
  12000 |-+                                                                 |   
  10000 |-+                                                                 |   
        |                                                                   |   
   8000 |-+                                                                 |   
   6000 |-+                                                                 |   
        |                                                                   |   
   4000 |-+                                                                 |   
   2000 |-+                                                                 |   
        |                                                                   |   
      0 +-------------------------------------------------------------------+   
                                                                                
                                                                                                                                                                
                                stress-ng.vm-splice.ops                         
                                                                                
  3.5e+09 +-----------------------------------------------------------------+   
          |                                                                 |   
    3e+09 |-+                                                               |   
          |                                                                 |   
  2.5e+09 |-+                                                               |   
          |                                                                 |   
    2e+09 |-+                                                               |   
          |                                                                 |   
  1.5e+09 |-+                                                               |   
          |                                                                 |   
    1e+09 |-+                                                               |   
          |                                                                 |   
    5e+08 |-+                                                               |   
          |.+.+..+.+.+.+.+..+.+.+.+.+..+.+.+.+.+..+.+.+.+.+..+.+.+.+.+..+.+ |   
        0 +-----------------------------------------------------------------+   
                                                                                
                                                                                                                                                                
                            stress-ng.vm-splice.ops_per_sec                     
                                                                                
  1.2e+08 +-----------------------------------------------------------------+   
          | O O  O O O O O  O O O   O  O O   O O    O   O    O O O O O  O O |   
    1e+08 |-+                                                               |   
          |                                                                 |   
          |                                                                 |   
    8e+07 |-+                                                               |   
          |                                                                 |   
    6e+07 |-+                                                               |   
          |                                                                 |   
    4e+07 |-+                                                               |   
          |                                                                 |   
          |                                                                 |   
    2e+07 |-+                                                               |   
          |.+.+..+.+.+.+.+..+.+.+.+.+..+.+.+.+.+..+.+.+.+.+..+.+.+.+.+..+.+ |   
        0 +-----------------------------------------------------------------+   
                                                                                
                                                                                
[*] bisect-good sample
[O] bisect-bad  sample

***************************************************************************************************
lkp-cfl-e1: 16 threads Intel(R) Xeon(R) E-2278G CPU @ 3.40GHz with 32G memory


***************************************************************************************************
lkp-csl-2sp5: 96 threads Intel(R) Xeon(R) Gold 6252 CPU @ 2.10GHz with 192G memory
=========================================================================================
class/compiler/cpufreq_governor/disk/fs/kconfig/nr_threads/rootfs/tbox_group/testcase/testtime/ucode:
  vm/gcc-9/performance/1HDD/ext4/x86_64-rhel-7.6/10%/debian-x86_64-20191114.cgz/lkp-csl-2sp5/stress-ng/1s/0x500002c

commit: 
  f359287765 ("Merge branch 'from-miklos' of git://git.kernel.org/pub/scm/linux/kernel/git/viro/vfs")
  17839856fd ("gup: document and work around "COW can break either way" issue")

f359287765c04711 17839856fd588f4ab6b789f482e 
---------------- --------------------------- 
       fail:runs  %reproduction    fail:runs
           |             |             |    
          1:4          -25%            :4     kmsg.Memory_failure:#:recovery_action_for_reserved_kernel_page:Failed
          1:4          -25%            :4     kmsg.Memory_failure:#:reserved_kernel_page_still_referenced_by#users
          0:4           14%           1:4     perf-profile.calltrace.cycles-pp.sync_regs.error_entry
          0:4           15%           1:4     perf-profile.calltrace.cycles-pp.error_entry
          2:4            2%           2:4     perf-profile.children.cycles-pp.error_entry
          0:4            0%           0:4     perf-profile.self.cycles-pp.error_entry
         %stddev     %change         %stddev
             \          |                \  
   3752446 ±  5%    +372.8%   17743420        stress-ng.vm-splice.ops
   3753307 ±  5%    +372.8%   17744113        stress-ng.vm-splice.ops_per_sec
    352732 ± 38%     -37.5%     220415 ± 19%  meminfo.AnonHugePages
      6088 ±  4%      -8.4%       5575 ±  4%  vmstat.io.bo
     26.97            +2.2%      27.56        boot-time.boot
     16.63            +1.1%      16.82        boot-time.dhcp
  11097144 ±  4%      +9.4%   12140802 ±  3%  cpuidle.C1.time
   1306863 ±  5%     +10.0%    1438008 ±  4%  cpuidle.C1.usage
     20928 ± 19%     +40.5%      29401 ± 12%  sched_debug.cfs_rq:/.min_vruntime.max
    621.93 ±  5%     -10.6%     556.11 ±  5%  sched_debug.cpu.curr->pid.stddev
   1413502 ± 10%      +9.9%    1552782        proc-vmstat.nr_foll_pin_acquired
   1413411 ± 10%      +9.9%    1552695        proc-vmstat.nr_foll_pin_released
  83634280 ±  3%      -4.1%   80233229        proc-vmstat.pgalloc_normal
  83604544 ±  3%      -4.1%   80204856        proc-vmstat.pgfree
      2279 ±  3%     +13.7%       2591        slabinfo.khugepaged_mm_slot.active_objs
      2279 ±  3%     +13.7%       2591        slabinfo.khugepaged_mm_slot.num_objs
     51107 ±  4%     -12.7%      44626 ±  7%  slabinfo.lsm_file_cache.active_objs
     53253 ±  3%     -12.1%      46810 ±  7%  slabinfo.lsm_file_cache.num_objs
      0.94 ±  8%      +0.1        1.07 ±  6%  perf-profile.calltrace.cycles-pp.__handle_mm_fault.handle_mm_fault.__get_user_pages.populate_vma_page_range.__mm_populate
      0.96 ±  8%      +0.1        1.09 ±  6%  perf-profile.calltrace.cycles-pp.handle_mm_fault.__get_user_pages.populate_vma_page_range.__mm_populate.vm_mmap_pgoff
      0.11 ± 30%      +0.1        0.16 ± 16%  perf-profile.children.cycles-pp.anon_vma_fork
      0.12 ± 18%      +0.0        0.17 ± 13%  perf-profile.self.cycles-pp.filemap_map_pages
 6.418e+09            +1.2%  6.498e+09        perf-stat.i.branch-instructions
   3572544            +2.1%    3646846        perf-stat.i.iTLB-loads
 3.035e+10            +1.4%  3.078e+10        perf-stat.i.instructions
     20346            -5.6%      19202 ±  2%  perf-stat.i.major-faults
      7.69 ±  4%      -3.5%       7.42 ±  3%  perf-stat.overall.MPKI
 6.243e+09            +1.3%  6.322e+09        perf-stat.ps.branch-instructions
  35907601            +3.8%   37255792 ±  3%  perf-stat.ps.branch-misses
 3.127e+09            +1.2%  3.165e+09        perf-stat.ps.dTLB-stores
   3453004            +2.3%    3531803        perf-stat.ps.iTLB-loads
 2.951e+10            +1.5%  2.994e+10        perf-stat.ps.instructions
     19574            -5.4%      18514 ±  2%  perf-stat.ps.major-faults
     38744 ± 10%     -27.2%      28219 ± 12%  interrupts.CPU24.TLB:TLB_shootdowns
      2974 ± 44%     -77.7%     662.50 ± 29%  interrupts.CPU27.NMI:Non-maskable_interrupts
      2974 ± 44%     -77.7%     662.50 ± 29%  interrupts.CPU27.PMI:Performance_monitoring_interrupts
      2511 ± 40%     -70.7%     736.25 ±  7%  interrupts.CPU28.NMI:Non-maskable_interrupts
      2511 ± 40%     -70.7%     736.25 ±  7%  interrupts.CPU28.PMI:Performance_monitoring_interrupts
     37767 ± 28%     -28.1%      27157 ± 14%  interrupts.CPU28.TLB:TLB_shootdowns
      2729 ± 63%     -60.8%       1070 ± 86%  interrupts.CPU32.NMI:Non-maskable_interrupts
      2729 ± 63%     -60.8%       1070 ± 86%  interrupts.CPU32.PMI:Performance_monitoring_interrupts
      2934 ± 59%     -66.9%     971.75 ± 90%  interrupts.CPU33.NMI:Non-maskable_interrupts
      2934 ± 59%     -66.9%     971.75 ± 90%  interrupts.CPU33.PMI:Performance_monitoring_interrupts
     18203 ±  7%     +17.7%      21419 ±  9%  interrupts.CPU59.CAL:Function_call_interrupts
     24514 ±  8%     +19.0%      29165 ± 10%  interrupts.CPU59.TLB:TLB_shootdowns
     99.25 ± 11%     -19.9%      79.50 ± 13%  interrupts.CPU62.RES:Rescheduling_interrupts
     19246 ±  8%     +19.6%      23023 ±  6%  interrupts.CPU68.CAL:Function_call_interrupts
     26764 ±  7%     +18.4%      31691 ±  8%  interrupts.CPU68.TLB:TLB_shootdowns
      3799 ± 52%     -80.9%     726.25 ± 22%  interrupts.CPU74.NMI:Non-maskable_interrupts
      3799 ± 52%     -80.9%     726.25 ± 22%  interrupts.CPU74.PMI:Performance_monitoring_interrupts
      3149 ± 49%     -79.8%     635.50 ± 21%  interrupts.CPU75.NMI:Non-maskable_interrupts
      3149 ± 49%     -79.8%     635.50 ± 21%  interrupts.CPU75.PMI:Performance_monitoring_interrupts
      2709 ± 42%     -73.4%     719.75 ± 26%  interrupts.CPU76.NMI:Non-maskable_interrupts
      2709 ± 42%     -73.4%     719.75 ± 26%  interrupts.CPU76.PMI:Performance_monitoring_interrupts
      2364 ± 40%     -57.3%       1009 ± 38%  interrupts.CPU79.NMI:Non-maskable_interrupts
      2364 ± 40%     -57.3%       1009 ± 38%  interrupts.CPU79.PMI:Performance_monitoring_interrupts
      3030 ± 48%     -68.1%     967.50 ± 38%  interrupts.CPU81.NMI:Non-maskable_interrupts
      3030 ± 48%     -68.1%     967.50 ± 38%  interrupts.CPU81.PMI:Performance_monitoring_interrupts
      2845 ± 67%     -51.8%       1370 ± 83%  interrupts.CPU83.NMI:Non-maskable_interrupts
      2845 ± 67%     -51.8%       1370 ± 83%  interrupts.CPU83.PMI:Performance_monitoring_interrupts
     27279 ± 11%     +23.0%      33551 ±  9%  interrupts.CPU84.TLB:TLB_shootdowns
     36061 ±  8%     -14.6%      30804 ±  6%  interrupts.CPU92.TLB:TLB_shootdowns
     29829 ±  5%     -14.1%      25619 ± 11%  interrupts.CPU94.TLB:TLB_shootdowns
    160996 ±  9%     -11.9%     141904 ± 11%  interrupts.NMI:Non-maskable_interrupts
    160996 ±  9%     -11.9%     141904 ± 11%  interrupts.PMI:Performance_monitoring_interrupts



***************************************************************************************************
lkp-csl-2sp5: 96 threads Intel(R) Xeon(R) Gold 6252 CPU @ 2.10GHz with 192G memory
=========================================================================================
class/compiler/cpufreq_governor/disk/kconfig/nr_threads/rootfs/tbox_group/testcase/testtime/ucode:
  pipe/gcc-9/performance/1HDD/x86_64-rhel-7.6/100%/debian-x86_64-20191114.cgz/lkp-csl-2sp5/stress-ng/1s/0x500002c

commit: 
  f359287765 ("Merge branch 'from-miklos' of git://git.kernel.org/pub/scm/linux/kernel/git/viro/vfs")
  17839856fd ("gup: document and work around "COW can break either way" issue")

f359287765c04711 17839856fd588f4ab6b789f482e 
---------------- --------------------------- 
         %stddev     %change         %stddev
             \          |                \  
    816942 ±  8%     -12.0%     719276        stress-ng.time.involuntary_context_switches
    235279            -4.0%     225847        stress-ng.time.minor_page_faults
    599.87            -2.8%     583.29        stress-ng.time.system_time
     47.33 ±  2%     +28.2%      60.69        stress-ng.time.user_time
   5412114 ±  5%   +2054.0%  1.166e+08        stress-ng.vm-splice.ops
   5417324 ±  5%   +2052.7%  1.166e+08        stress-ng.vm-splice.ops_per_sec
     16.42            +1.2%      16.62        boot-time.dhcp
      5.94 ±  4%     +24.7%       7.41 ±  4%  iostat.cpu.user
      7.12 ±  5%      +1.4        8.52 ±  2%  mpstat.cpu.all.usr%
  1.61e+08 ± 11%     -50.9%   79016011 ± 47%  cpuidle.C1E.time
    984623 ± 12%     -21.5%     772591 ± 14%  cpuidle.C1E.usage
     19231 ± 11%     +21.8%      23416 ±  8%  numa-meminfo.node0.Mapped
    938.25 ± 85%     -55.9%     413.50 ±173%  numa-meminfo.node1.Inactive(file)
      4850 ± 11%     +18.6%       5754 ± 10%  numa-vmstat.node0.nr_mapped
    234.00 ± 85%     -55.9%     103.25 ±173%  numa-vmstat.node1.nr_inactive_file
    234.00 ± 85%     -55.9%     103.25 ±173%  numa-vmstat.node1.nr_zone_inactive_file
      9049           +13.7%      10287 ± 11%  softirqs.CPU44.TIMER
      9712 ± 10%     -12.2%       8524        softirqs.CPU57.TIMER
      9235 ±  6%      -7.3%       8560 ±  3%  softirqs.CPU82.TIMER
      8160            +5.6%       8617 ±  5%  proc-vmstat.nr_shmem
     57724            +3.1%      59541        proc-vmstat.nr_slab_unreclaimable
     29.50 ± 78%  +11075.4%       3296 ± 69%  proc-vmstat.numa_hint_faults_local
    295784            -2.1%     289478        proc-vmstat.pgfault
      7320 ±  9%     -30.6%       5078 ± 22%  sched_debug.cfs_rq:/.min_vruntime.min
    836.06 ± 11%     +16.2%     971.79 ±  5%  sched_debug.cfs_rq:/.runnable_avg.stddev
     55.75 ± 14%     +18.8%      66.25 ±  6%  sched_debug.cpu.nr_uninterruptible.max
    -59.50           -16.8%     -49.50        sched_debug.cpu.nr_uninterruptible.min
    773.00 ±  6%     +30.9%       1011 ± 14%  interrupts.CPU14.RES:Rescheduling_interrupts
      2901 ± 34%     +43.0%       4147 ± 24%  interrupts.CPU16.NMI:Non-maskable_interrupts
      2901 ± 34%     +43.0%       4147 ± 24%  interrupts.CPU16.PMI:Performance_monitoring_interrupts
    969.25 ± 10%     +74.4%       1690 ± 51%  interrupts.CPU31.RES:Rescheduling_interrupts
    989.25 ± 12%     +39.0%       1375        interrupts.CPU54.RES:Rescheduling_interrupts
      1207 ± 18%     -24.7%     908.75 ± 10%  interrupts.CPU68.RES:Rescheduling_interrupts
    884.75 ± 11%     +33.3%       1179 ± 11%  interrupts.CPU8.RES:Rescheduling_interrupts
      1015 ±  8%     +22.4%       1243 ±  5%  interrupts.CPU81.RES:Rescheduling_interrupts
    865.50 ±  8%     +18.7%       1027 ±  7%  interrupts.CPU9.RES:Rescheduling_interrupts
     80.00 ± 69%     -53.4%      37.25 ± 15%  interrupts.TLB:TLB_shootdowns
      4.42 ± 10%     +34.9%       5.96 ± 25%  perf-stat.i.MPKI
 2.426e+08            +2.0%  2.473e+08        perf-stat.i.branch-misses
 5.281e+08 ±  7%     +10.8%   5.85e+08 ±  4%  perf-stat.i.cache-references
      0.02 ±  7%      +0.0        0.03 ± 25%  perf-stat.i.dTLB-load-miss-rate%
     37835            -3.7%      36434        perf-stat.i.minor-faults
     37835            -3.7%      36434        perf-stat.i.page-faults
      3.25 ±  6%     +10.1%       3.58 ±  3%  perf-stat.overall.MPKI
 2.115e+08            +2.1%  2.158e+08        perf-stat.ps.branch-misses
 4.613e+08 ±  7%     +10.9%  5.116e+08 ±  3%  perf-stat.ps.cache-references
     33390            -2.3%      32620        perf-stat.ps.minor-faults
     33390            -2.3%      32620        perf-stat.ps.page-faults
 1.154e+12            +1.2%  1.168e+12        perf-stat.total.instructions
     11.58           -11.6        0.00        perf-profile.calltrace.cycles-pp.get_user_pages_unlocked.internal_get_user_pages_fast.iov_iter_get_pages.iter_to_pipe.do_vmsplice
     11.26           -11.3        0.00        perf-profile.calltrace.cycles-pp.__get_user_pages.get_user_pages_unlocked.internal_get_user_pages_fast.iov_iter_get_pages.iter_to_pipe
     11.17           -11.2        0.00        perf-profile.calltrace.cycles-pp.follow_page_pte.__get_user_pages.get_user_pages_unlocked.internal_get_user_pages_fast.iov_iter_get_pages
     11.64            -9.4        2.29 ±  3%  perf-profile.calltrace.cycles-pp.internal_get_user_pages_fast.iov_iter_get_pages.iter_to_pipe.do_vmsplice.__do_sys_vmsplice
     11.67            -9.0        2.67 ±  3%  perf-profile.calltrace.cycles-pp.iov_iter_get_pages.iter_to_pipe.do_vmsplice.__do_sys_vmsplice.do_syscall_64
     11.70            -8.1        3.63 ±  3%  perf-profile.calltrace.cycles-pp.iter_to_pipe.do_vmsplice.__do_sys_vmsplice.do_syscall_64.entry_SYSCALL_64_after_hwframe
     12.08            -7.1        4.94 ±  3%  perf-profile.calltrace.cycles-pp.do_vmsplice.__do_sys_vmsplice.do_syscall_64.entry_SYSCALL_64_after_hwframe.vmsplice
      6.36            -6.4        0.00        perf-profile.calltrace.cycles-pp.mark_page_accessed.follow_page_pte.__get_user_pages.get_user_pages_unlocked.internal_get_user_pages_fast
     12.16            -5.1        7.10 ±  3%  perf-profile.calltrace.cycles-pp.__do_sys_vmsplice.do_syscall_64.entry_SYSCALL_64_after_hwframe.vmsplice
     12.18            -4.6        7.54 ±  3%  perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.vmsplice
     12.19            -4.4        7.77 ±  3%  perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.vmsplice
      8.73            -3.0        5.75 ±  3%  perf-profile.calltrace.cycles-pp.__splice_from_pipe.splice_from_pipe.do_splice.__x64_sys_splice.do_syscall_64
     12.25            -2.9        9.32 ±  3%  perf-profile.calltrace.cycles-pp.vmsplice
      9.79            -1.8        7.97 ±  3%  perf-profile.calltrace.cycles-pp.splice_from_pipe.do_splice.__x64_sys_splice.do_syscall_64.entry_SYSCALL_64_after_hwframe
      8.60 ±  7%      -0.8        7.80 ±  5%  perf-profile.calltrace.cycles-pp.do_idle.cpu_startup_entry.start_secondary.secondary_startup_64
      8.60 ±  7%      -0.8        7.80 ±  5%  perf-profile.calltrace.cycles-pp.cpu_startup_entry.start_secondary.secondary_startup_64
      8.60 ±  7%      -0.8        7.80 ±  5%  perf-profile.calltrace.cycles-pp.start_secondary.secondary_startup_64
     12.20 ±  2%      -0.6       11.62        perf-profile.calltrace.cycles-pp.enqueue_task_fair.activate_task.ttwu_do_activate.try_to_wake_up.autoremove_wake_function
     12.22 ±  3%      -0.6       11.64        perf-profile.calltrace.cycles-pp.ttwu_do_activate.try_to_wake_up.autoremove_wake_function.__wake_up_common.__wake_up_common_lock
     12.22 ±  3%      -0.6       11.64        perf-profile.calltrace.cycles-pp.activate_task.ttwu_do_activate.try_to_wake_up.autoremove_wake_function.__wake_up_common
     10.57            -0.6       10.02        perf-profile.calltrace.cycles-pp.autoremove_wake_function.__wake_up_common.__wake_up_common_lock.pipe_write.new_sync_write
     10.53            -0.6        9.98        perf-profile.calltrace.cycles-pp.try_to_wake_up.autoremove_wake_function.__wake_up_common.__wake_up_common_lock.pipe_write
      6.78 ±  2%      -0.4        6.38 ±  2%  perf-profile.calltrace.cycles-pp.prepare_to_wait_event.pipe_read.new_sync_read.vfs_read.ksys_read
      6.57 ±  2%      -0.4        6.20 ±  2%  perf-profile.calltrace.cycles-pp._raw_spin_lock_irqsave.prepare_to_wait_event.pipe_read.new_sync_read.vfs_read
      6.50 ±  2%      -0.4        6.13 ±  2%  perf-profile.calltrace.cycles-pp.native_queued_spin_lock_slowpath._raw_spin_lock_irqsave.prepare_to_wait_event.pipe_read.new_sync_read
      0.55 ±  8%      -0.3        0.26 ±100%  perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.__libc_fork
      0.55 ±  8%      -0.3        0.26 ±100%  perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.__libc_fork
      0.55 ±  8%      -0.3        0.26 ±100%  perf-profile.calltrace.cycles-pp.__do_sys_clone.do_syscall_64.entry_SYSCALL_64_after_hwframe.__libc_fork
      0.55 ±  8%      -0.3        0.26 ±100%  perf-profile.calltrace.cycles-pp._do_fork.__do_sys_clone.do_syscall_64.entry_SYSCALL_64_after_hwframe.__libc_fork
      3.23            -0.3        2.94 ±  3%  perf-profile.calltrace.cycles-pp.direct_splice_actor.splice_direct_to_actor.do_splice_direct.do_sendfile.__x64_sys_sendfile64
      3.19            -0.3        2.90 ±  3%  perf-profile.calltrace.cycles-pp.splice_from_pipe.direct_splice_actor.splice_direct_to_actor.do_splice_direct.do_sendfile
      2.94            -0.3        2.68 ±  3%  perf-profile.calltrace.cycles-pp.__splice_from_pipe.splice_from_pipe.direct_splice_actor.splice_direct_to_actor.do_splice_direct
      1.93 ±  2%      -0.2        1.78 ±  3%  perf-profile.calltrace.cycles-pp.push_pipe.iov_iter_get_pages_alloc.default_file_splice_read.do_splice.__x64_sys_splice
      1.75 ±  2%      -0.1        1.62 ±  3%  perf-profile.calltrace.cycles-pp.xas_load.find_get_entry.pagecache_get_page.generic_file_buffered_read.generic_file_splice_read
      1.39 ±  2%      -0.1        1.30 ±  3%  perf-profile.calltrace.cycles-pp.__alloc_pages_nodemask.push_pipe.iov_iter_get_pages_alloc.default_file_splice_read.do_splice
      0.80 ±  2%      -0.1        0.72 ±  2%  perf-profile.calltrace.cycles-pp.sanity.copy_page_to_iter.generic_file_buffered_read.generic_file_splice_read.splice_direct_to_actor
      0.55 ±  3%      -0.0        0.53        perf-profile.calltrace.cycles-pp.free_unref_page.__splice_from_pipe.splice_from_pipe.do_splice.__x64_sys_splice
      0.58 ±  3%      +0.1        0.64 ±  4%  perf-profile.calltrace.cycles-pp.__kmalloc_node.iov_iter_get_pages_alloc.default_file_splice_read.do_splice.__x64_sys_splice
      1.07 ±  5%      +0.1        1.19 ±  3%  perf-profile.calltrace.cycles-pp.copy_user_enhanced_fast_string.copyout.copy_page_to_iter.pipe_read.new_sync_read
      1.20 ±  4%      +0.1        1.32 ±  2%  perf-profile.calltrace.cycles-pp.copyout.copy_page_to_iter.pipe_read.new_sync_read.vfs_read
      1.54 ±  4%      +0.2        1.74 ±  2%  perf-profile.calltrace.cycles-pp.copy_page_to_iter.pipe_read.new_sync_read.vfs_read.ksys_read
      0.00            +0.6        0.56 ±  4%  perf-profile.calltrace.cycles-pp.mutex_lock.do_vmsplice.__do_sys_vmsplice.do_syscall_64.entry_SYSCALL_64_after_hwframe
      0.58 ± 14%      +0.7        1.33 ±  5%  perf-profile.calltrace.cycles-pp.security_file_permission.do_splice.__x64_sys_splice.do_syscall_64.entry_SYSCALL_64_after_hwframe
      0.96 ±  2%      +0.8        1.72 ±  4%  perf-profile.calltrace.cycles-pp.entry_SYSCALL_64.splice
      0.00            +0.8        0.79 ±  2%  perf-profile.calltrace.cycles-pp.entry_SYSCALL_64.vmsplice
      0.00            +0.8        0.84 ±  4%  perf-profile.calltrace.cycles-pp.__fget_light.__x64_sys_splice.do_syscall_64.entry_SYSCALL_64_after_hwframe.splice
      0.14 ±173%      +0.8        0.98 ±  7%  perf-profile.calltrace.cycles-pp.common_file_perm.security_file_permission.do_splice.__x64_sys_splice.do_syscall_64
      0.00            +0.9        0.87 ±  2%  perf-profile.calltrace.cycles-pp._copy_from_user.rw_copy_check_uvector.import_iovec.__do_sys_vmsplice.do_syscall_64
      0.00            +0.9        0.89 ±  3%  perf-profile.calltrace.cycles-pp.mutex_lock.splice_from_pipe.do_splice.__x64_sys_splice.do_syscall_64
      0.00            +1.3        1.28 ±  2%  perf-profile.calltrace.cycles-pp.rw_copy_check_uvector.import_iovec.__do_sys_vmsplice.do_syscall_64.entry_SYSCALL_64_after_hwframe
      0.00            +1.5        1.53 ±  2%  perf-profile.calltrace.cycles-pp.import_iovec.__do_sys_vmsplice.do_syscall_64.entry_SYSCALL_64_after_hwframe.vmsplice
      0.00            +1.9        1.94 ±  3%  perf-profile.calltrace.cycles-pp.gup_pgd_range.internal_get_user_pages_fast.iov_iter_get_pages.iter_to_pipe.do_vmsplice
     23.74            +3.0       26.75 ±  3%  perf-profile.calltrace.cycles-pp.splice
     11.58           -11.6        0.00        perf-profile.children.cycles-pp.get_user_pages_unlocked
     11.26           -11.3        0.00        perf-profile.children.cycles-pp.__get_user_pages
     11.17           -11.2        0.00        perf-profile.children.cycles-pp.follow_page_pte
     11.64            -9.3        2.32 ±  3%  perf-profile.children.cycles-pp.internal_get_user_pages_fast
     11.67            -9.0        2.68 ±  3%  perf-profile.children.cycles-pp.iov_iter_get_pages
     11.70            -8.0        3.67 ±  3%  perf-profile.children.cycles-pp.iter_to_pipe
     12.08            -7.0        5.04 ±  3%  perf-profile.children.cycles-pp.do_vmsplice
      7.09            -6.4        0.69 ±  4%  perf-profile.children.cycles-pp.mark_page_accessed
     12.16            -5.0        7.15 ±  3%  perf-profile.children.cycles-pp.__do_sys_vmsplice
      5.55            -4.0        1.52 ±  3%  perf-profile.children.cycles-pp.page_cache_pipe_buf_release
     11.81            -3.2        8.62 ±  3%  perf-profile.children.cycles-pp.__splice_from_pipe
     86.84            -2.7       84.19        perf-profile.children.cycles-pp.do_syscall_64
     12.27            -2.5        9.74 ±  3%  perf-profile.children.cycles-pp.vmsplice
     87.27            -2.2       85.05        perf-profile.children.cycles-pp.entry_SYSCALL_64_after_hwframe
     13.00            -2.1       10.91 ±  3%  perf-profile.children.cycles-pp.splice_from_pipe
      8.60 ±  7%      -0.8        7.80 ±  5%  perf-profile.children.cycles-pp.start_secondary
     16.27            -0.6       15.63        perf-profile.children.cycles-pp.autoremove_wake_function
      6.85 ±  2%      -0.4        6.46 ±  2%  perf-profile.children.cycles-pp.prepare_to_wait_event
      3.24            -0.3        2.95 ±  3%  perf-profile.children.cycles-pp.direct_splice_actor
      0.57 ±  2%      -0.2        0.38 ±  4%  perf-profile.children.cycles-pp.page_cache_pipe_buf_confirm
      1.94 ±  2%      -0.1        1.79 ±  2%  perf-profile.children.cycles-pp.push_pipe
      1.80 ±  2%      -0.1        1.67 ±  3%  perf-profile.children.cycles-pp.xas_load
      0.20 ±  7%      -0.1        0.10 ± 19%  perf-profile.children.cycles-pp.kthread
      0.21 ±  6%      -0.1        0.12 ± 17%  perf-profile.children.cycles-pp.ret_from_fork
      0.86 ±  2%      -0.1        0.77 ±  2%  perf-profile.children.cycles-pp.sanity
      0.16 ±  5%      -0.1        0.09 ± 26%  perf-profile.children.cycles-pp.memcpy_erms
      0.15 ±  4%      -0.1        0.09 ± 26%  perf-profile.children.cycles-pp.drm_fb_helper_dirty_work
      0.36            -0.1        0.30        perf-profile.children.cycles-pp.alloc_pages_current
      1.89            -0.1        1.83        perf-profile.children.cycles-pp.__alloc_pages_nodemask
      0.16 ±  5%      -0.1        0.10 ± 23%  perf-profile.children.cycles-pp.worker_thread
      0.16 ±  5%      -0.1        0.10 ± 23%  perf-profile.children.cycles-pp.process_one_work
      0.56 ±  9%      -0.1        0.50 ±  5%  perf-profile.children.cycles-pp.__do_sys_clone
      0.56 ±  9%      -0.1        0.50 ±  5%  perf-profile.children.cycles-pp._do_fork
      0.60 ±  7%      -0.1        0.55 ±  4%  perf-profile.children.cycles-pp.__libc_fork
      0.52 ±  9%      -0.1        0.47 ±  5%  perf-profile.children.cycles-pp.copy_process
      0.10 ±  5%      -0.0        0.07 ±  6%  perf-profile.children.cycles-pp.get_task_policy
      0.11            -0.0        0.09 ±  8%  perf-profile.children.cycles-pp.copy_page_range
      0.07 ±  5%      -0.0        0.06 ±  7%  perf-profile.children.cycles-pp.free_pcp_prepare
      0.11 ±  4%      -0.0        0.09 ±  4%  perf-profile.children.cycles-pp.prep_new_page
      0.17 ±  2%      +0.0        0.18        perf-profile.children.cycles-pp.scheduler_tick
      0.09 ±  4%      +0.0        0.10        perf-profile.children.cycles-pp.ktime_get_coarse_real_ts64
      0.34 ±  2%      +0.0        0.36        perf-profile.children.cycles-pp.__hrtimer_run_queues
      0.37            +0.0        0.39 ±  2%  perf-profile.children.cycles-pp.current_time
      0.05 ±  8%      +0.0        0.08        perf-profile.children.cycles-pp.kill_fasync
      0.24 ±  5%      +0.0        0.27 ±  4%  perf-profile.children.cycles-pp.update_cfs_group
      0.65 ±  2%      +0.0        0.68        perf-profile.children.cycles-pp.smp_apic_timer_interrupt
      0.53 ±  3%      +0.0        0.56        perf-profile.children.cycles-pp.hrtimer_interrupt
      0.60            +0.0        0.64        perf-profile.children.cycles-pp.fsnotify_parent
      0.15 ± 12%      +0.0        0.19 ±  5%  perf-profile.children.cycles-pp.osq_lock
      0.26 ±  7%      +0.0        0.30 ±  4%  perf-profile.children.cycles-pp.__fdget_pos
      0.06 ± 11%      +0.1        0.12 ± 56%  perf-profile.children.cycles-pp.queued_read_lock_slowpath
      0.06 ±  6%      +0.1        0.13 ±  3%  perf-profile.children.cycles-pp.splice_from_pipe_next
      0.00            +0.1        0.07 ± 12%  perf-profile.children.cycles-pp.__fdget
      0.63 ±  3%      +0.1        0.70 ±  4%  perf-profile.children.cycles-pp.__kmalloc_node
      0.05 ±  8%      +0.1        0.13 ±  3%  perf-profile.children.cycles-pp.pipe_unlock
      0.07 ± 10%      +0.1        0.15 ±  7%  perf-profile.children.cycles-pp.check_stack_object
      0.94 ±  3%      +0.1        1.02 ±  2%  perf-profile.children.cycles-pp.fsnotify
      0.06 ±  6%      +0.1        0.15 ±  4%  perf-profile.children.cycles-pp.pipe_lock
      0.00            +0.1        0.11 ±  4%  perf-profile.children.cycles-pp.wait_for_space
      0.30 ±  2%      +0.1        0.41 ±  4%  perf-profile.children.cycles-pp.aa_file_perm
      0.00            +0.1        0.11 ±  4%  perf-profile.children.cycles-pp.__x64_sys_vmsplice
      0.21 ±  3%      +0.1        0.32 ±  3%  perf-profile.children.cycles-pp.pipe_to_null
      0.08 ±  5%      +0.1        0.20 ±  2%  perf-profile.children.cycles-pp.splice_write_null
      0.09 ± 13%      +0.1        0.21 ± 12%  perf-profile.children.cycles-pp.get_pipe_info
      0.34 ±  2%      +0.1        0.46 ±  3%  perf-profile.children.cycles-pp.rcu_all_qs
      0.29 ±  5%      +0.1        0.41 ±  2%  perf-profile.children.cycles-pp.apparmor_file_permission
      0.39            +0.1        0.52        perf-profile.children.cycles-pp.__x86_indirect_thunk_rax
      1.20 ±  4%      +0.1        1.33 ±  2%  perf-profile.children.cycles-pp.copyout
      0.14 ±  3%      +0.1        0.27 ±  3%  perf-profile.children.cycles-pp.generic_pipe_buf_confirm
      0.32 ±  6%      +0.2        0.47        perf-profile.children.cycles-pp.rw_verify_area
      4.91 ±  2%      +0.2        5.06        perf-profile.children.cycles-pp.copy_page_to_iter
      0.15 ±  3%      +0.2        0.33        perf-profile.children.cycles-pp.fpregs_assert_state_consistent
      0.00            +0.2        0.22 ±  4%  perf-profile.children.cycles-pp.add_to_pipe
      0.14 ±  3%      +0.2        0.36 ±  3%  perf-profile.children.cycles-pp.__check_object_size
      0.70            +0.2        0.93        perf-profile.children.cycles-pp._cond_resched
      0.27 ±  5%      +0.2        0.51 ±  3%  perf-profile.children.cycles-pp.copy_user_generic_unrolled
      1.02 ±  3%      +0.3        1.28 ±  3%  perf-profile.children.cycles-pp.mutex_unlock
      0.18 ±  8%      +0.3        0.45 ±  3%  perf-profile.children.cycles-pp.iov_iter_advance
      1.97 ±  3%      +0.3        2.28 ±  2%  perf-profile.children.cycles-pp.copy_user_enhanced_fast_string
      0.00            +0.3        0.32 ±  3%  perf-profile.children.cycles-pp.try_grab_compound_head
      0.61 ±  4%      +0.4        1.00        perf-profile.children.cycles-pp.__might_sleep
      0.29 ±  7%      +0.4        0.68        perf-profile.children.cycles-pp.__might_fault
      0.69            +0.6        1.34 ±  2%  perf-profile.children.cycles-pp.__fget_light
      1.83 ± 13%      +0.7        2.48 ±  9%  perf-profile.children.cycles-pp.common_file_perm
      2.10            +0.7        2.79 ±  2%  perf-profile.children.cycles-pp.___might_sleep
      0.39 ±  3%      +0.9        1.26 ±  3%  perf-profile.children.cycles-pp._copy_from_user
      2.95 ±  9%      +0.9        3.85 ±  5%  perf-profile.children.cycles-pp.security_file_permission
      1.76 ±  5%      +1.1        2.87        perf-profile.children.cycles-pp.mutex_lock
      1.27 ±  2%      +1.2        2.50        perf-profile.children.cycles-pp.syscall_return_via_sysret
      0.64 ±  2%      +1.3        1.91 ±  3%  perf-profile.children.cycles-pp.rw_copy_check_uvector
      0.75 ±  2%      +1.5        2.23 ±  3%  perf-profile.children.cycles-pp.import_iovec
      1.54 ±  4%      +1.6        3.15 ±  2%  perf-profile.children.cycles-pp.entry_SYSCALL_64
      0.06 ±  9%      +1.9        1.97 ±  3%  perf-profile.children.cycles-pp.gup_pgd_range
     24.24            +2.5       26.77 ±  3%  perf-profile.children.cycles-pp.splice
      6.79            -6.1        0.65 ±  5%  perf-profile.self.cycles-pp.mark_page_accessed
      5.48            -4.0        1.47 ±  3%  perf-profile.self.cycles-pp.page_cache_pipe_buf_release
      0.54            -0.2        0.35 ±  6%  perf-profile.self.cycles-pp.page_cache_pipe_buf_confirm
      1.32 ±  3%      -0.1        1.18 ±  3%  perf-profile.self.cycles-pp.xas_load
      0.26 ±  3%      -0.1        0.19 ±  2%  perf-profile.self.cycles-pp._raw_spin_lock
      0.15 ±  5%      -0.1        0.09 ± 26%  perf-profile.self.cycles-pp.memcpy_erms
      0.78 ±  2%      -0.1        0.72 ±  2%  perf-profile.self.cycles-pp.sanity
      0.51 ±  3%      -0.0        0.46 ±  3%  perf-profile.self.cycles-pp.__alloc_pages_nodemask
      0.09 ±  4%      -0.0        0.06        perf-profile.self.cycles-pp.get_task_policy
      0.14 ±  5%      +0.0        0.15 ±  3%  perf-profile.self.cycles-pp.current_time
      0.14 ±  7%      +0.0        0.16 ±  2%  perf-profile.self.cycles-pp.new_sync_write
      0.05            +0.0        0.08 ±  5%  perf-profile.self.cycles-pp.kill_fasync
      0.24 ±  5%      +0.0        0.27 ±  4%  perf-profile.self.cycles-pp.update_cfs_group
      0.16 ±  6%      +0.0        0.19 ±  2%  perf-profile.self.cycles-pp.vfs_write
      0.53 ±  2%      +0.0        0.56        perf-profile.self.cycles-pp.fsnotify_parent
      0.15 ± 12%      +0.0        0.19 ±  6%  perf-profile.self.cycles-pp.osq_lock
      0.21 ±  2%      +0.1        0.27 ±  3%  perf-profile.self.cycles-pp.__x86_indirect_thunk_rax
      0.05 ±  8%      +0.1        0.11 ±  4%  perf-profile.self.cycles-pp.splice_from_pipe_next
      0.06 ±  7%      +0.1        0.12 ±  4%  perf-profile.self.cycles-pp.check_stack_object
      0.91 ±  3%      +0.1        0.99        perf-profile.self.cycles-pp.fsnotify
      0.01 ±173%      +0.1        0.09 ± 11%  perf-profile.self.cycles-pp.pipe_lock
      0.06 ± 11%      +0.1        0.14 ± 17%  perf-profile.self.cycles-pp.get_pipe_info
      0.25 ±  2%      +0.1        0.34 ±  3%  perf-profile.self.cycles-pp.rcu_all_qs
      0.00            +0.1        0.09 ±  4%  perf-profile.self.cycles-pp.pipe_unlock
      0.00            +0.1        0.09 ±  4%  perf-profile.self.cycles-pp.wait_for_space
      0.10 ±  4%      +0.1        0.20 ±  5%  perf-profile.self.cycles-pp.__might_fault
      0.28 ±  2%      +0.1        0.38 ±  4%  perf-profile.self.cycles-pp.aa_file_perm
      0.17 ±  2%      +0.1        0.27 ±  2%  perf-profile.self.cycles-pp.pipe_to_null
      0.00            +0.1        0.10 ±  4%  perf-profile.self.cycles-pp.__x64_sys_vmsplice
      0.08 ±  5%      +0.1        0.18 ±  2%  perf-profile.self.cycles-pp.splice_write_null
      0.28            +0.1        0.38 ±  2%  perf-profile.self.cycles-pp.wakeup_pipe_writers
      0.32 ±  2%      +0.1        0.43 ±  2%  perf-profile.self.cycles-pp._cond_resched
      0.27 ±  4%      +0.1        0.38 ±  3%  perf-profile.self.cycles-pp.apparmor_file_permission
      0.26 ±  6%      +0.1        0.37 ±  3%  perf-profile.self.cycles-pp.wakeup_pipe_readers
      0.51 ±  2%      +0.1        0.63 ±  5%  perf-profile.self.cycles-pp.security_file_permission
      0.07 ±  6%      +0.1        0.20 ±  4%  perf-profile.self.cycles-pp.__check_object_size
      0.08            +0.1        0.21 ±  5%  perf-profile.self.cycles-pp._copy_from_user
      0.29 ±  6%      +0.1        0.43        perf-profile.self.cycles-pp.rw_verify_area
      0.12 ±  3%      +0.1        0.26 ±  4%  perf-profile.self.cycles-pp.generic_pipe_buf_confirm
      0.15 ±  2%      +0.2        0.32        perf-profile.self.cycles-pp.fpregs_assert_state_consistent
      0.11 ±  6%      +0.2        0.32 ±  2%  perf-profile.self.cycles-pp.import_iovec
      0.00            +0.2        0.21 ±  3%  perf-profile.self.cycles-pp.do_vmsplice
      0.00            +0.2        0.21 ±  4%  perf-profile.self.cycles-pp.add_to_pipe
      3.05 ±  2%      +0.2        3.27 ±  3%  perf-profile.self.cycles-pp.find_get_entry
      0.13            +0.2        0.35 ±  3%  perf-profile.self.cycles-pp.rw_copy_check_uvector
      0.26 ±  4%      +0.2        0.49 ±  2%  perf-profile.self.cycles-pp.copy_user_generic_unrolled
      0.99 ±  3%      +0.2        1.24 ±  3%  perf-profile.self.cycles-pp.mutex_unlock
      0.30 ±  2%      +0.2        0.55 ±  4%  perf-profile.self.cycles-pp.__x64_sys_splice
      0.18 ±  6%      +0.3        0.43 ±  3%  perf-profile.self.cycles-pp.iov_iter_advance
      0.00            +0.3        0.29 ±  4%  perf-profile.self.cycles-pp.__do_sys_vmsplice
      1.95 ±  3%      +0.3        2.24 ±  2%  perf-profile.self.cycles-pp.copy_user_enhanced_fast_string
      0.00            +0.3        0.31 ±  3%  perf-profile.self.cycles-pp.try_grab_compound_head
      0.52 ±  4%      +0.3        0.86        perf-profile.self.cycles-pp.__might_sleep
      0.00            +0.3        0.35 ±  5%  perf-profile.self.cycles-pp.iov_iter_get_pages
      0.00            +0.4        0.36 ±  2%  perf-profile.self.cycles-pp.internal_get_user_pages_fast
      0.58 ±  2%      +0.4        0.97        perf-profile.self.cycles-pp.do_syscall_64
      1.06 ±  6%      +0.4        1.46 ±  2%  perf-profile.self.cycles-pp.mutex_lock
      0.54 ±  2%      +0.4        0.98 ±  5%  perf-profile.self.cycles-pp.do_splice
      0.42 ±  2%      +0.5        0.88        perf-profile.self.cycles-pp.entry_SYSCALL_64_after_hwframe
      0.67 ±  3%      +0.5        1.14 ±  2%  perf-profile.self.cycles-pp.splice
      0.00            +0.5        0.49 ±  3%  perf-profile.self.cycles-pp.iter_to_pipe
      0.49 ±  3%      +0.5        0.98 ±  2%  perf-profile.self.cycles-pp.splice_from_pipe
      1.54 ±  2%      +0.5        2.04 ±  2%  perf-profile.self.cycles-pp.__splice_from_pipe
      0.00            +0.5        0.54 ±  3%  perf-profile.self.cycles-pp.vmsplice
      1.52 ± 16%      +0.5        2.06 ± 11%  perf-profile.self.cycles-pp.common_file_perm
      0.67            +0.6        1.30 ±  2%  perf-profile.self.cycles-pp.__fget_light
      2.02            +0.7        2.68 ±  2%  perf-profile.self.cycles-pp.___might_sleep
      0.76 ±  3%      +0.8        1.56 ±  2%  perf-profile.self.cycles-pp.entry_SYSCALL_64
      1.26 ±  2%      +1.2        2.48        perf-profile.self.cycles-pp.syscall_return_via_sysret
      0.05 ±  8%      +1.5        1.58 ±  3%  perf-profile.self.cycles-pp.gup_pgd_range





Disclaimer:
Results have been estimated based on internal Intel analysis and are provided
for informational purposes only. Any difference in system hardware or software
design or configuration may affect actual performance.


Thanks,
Rong Chen


View attachment "config-5.7.0-01306-g17839856fd588" of type "text/plain" (202923 bytes)

View attachment "job-script" of type "text/plain" (7845 bytes)

View attachment "job.yaml" of type "text/plain" (5505 bytes)

View attachment "reproduce" of type "text/plain" (390 bytes)

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ