lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [day] [month] [year] [list]
Message-ID: <202307141335.ac9351d0-oliver.sang@intel.com>
Date:   Fri, 14 Jul 2023 14:05:42 +0800
From:   kernel test robot <oliver.sang@...el.com>
To:     Hugh Dickins <hughd@...gle.com>
CC:     <oe-lkp@...ts.linux.dev>, <lkp@...el.com>,
        <linux-kernel@...r.kernel.org>,
        Andrew Morton <akpm@...ux-foundation.org>,
        Alistair Popple <apopple@...dia.com>,
        Anshuman Khandual <anshuman.khandual@....com>,
        "Axel Rasmussen" <axelrasmussen@...gle.com>,
        Christophe Leroy <christophe.leroy@...roup.eu>,
        Christoph Hellwig <hch@...radead.org>,
        "David Hildenbrand" <david@...hat.com>,
        "Huang, Ying" <ying.huang@...el.com>,
        "Ira Weiny" <ira.weiny@...el.com>, Jason Gunthorpe <jgg@...pe.ca>,
        "Kirill A. Shutemov" <kirill.shutemov@...ux.intel.com>,
        Lorenzo Stoakes <lstoakes@...il.com>,
        Matthew Wilcox <willy@...radead.org>,
        Mel Gorman <mgorman@...hsingularity.net>,
        Miaohe Lin <linmiaohe@...wei.com>,
        "Mike Kravetz" <mike.kravetz@...cle.com>,
        Mike Rapoport <rppt@...nel.org>,
        "Minchan Kim" <minchan@...nel.org>,
        Naoya Horiguchi <naoya.horiguchi@....com>,
        "Pavel Tatashin" <pasha.tatashin@...een.com>,
        Peter Xu <peterx@...hat.com>,
        "Peter Zijlstra" <peterz@...radead.org>,
        Qi Zheng <zhengqi.arch@...edance.com>,
        Ralph Campbell <rcampbell@...dia.com>,
        Ryan Roberts <ryan.roberts@....com>,
        SeongJae Park <sj@...nel.org>, Song Liu <song@...nel.org>,
        Steven Price <steven.price@....com>,
        Suren Baghdasaryan <surenb@...gle.com>,
        Thomas Hellström 
        <thomas.hellstrom@...ux.intel.com>, Will Deacon <will@...nel.org>,
        Yang Shi <shy828301@...il.com>, Yu Zhao <yuzhao@...gle.com>,
        Zack Rusin <zackr@...are.com>, <linux-mm@...ck.org>,
        <feng.tang@...el.com>, <fengwei.yin@...el.com>,
        <oliver.sang@...el.com>
Subject: [linus:master] [mm/memory]  3db82b9374:  vm-scalability.throughput
 39.9% improvement



Hello,

kernel test robot noticed a 39.9% improvement of vm-scalability.throughput on:


commit: 3db82b9374ca921b8b820a75e83809d5c4133d8f ("mm/memory: allow pte_offset_map[_lock]() to fail")
https://git.kernel.org/cgit/linux/kernel/git/torvalds/linux.git master

testcase: vm-scalability
test machine: 96 threads 2 sockets Intel(R) Xeon(R) Platinum 8260L CPU @ 2.40GHz (Cascade Lake) with 128G memory
parameters:

	runtime: 300s
	size: 8T
	test: anon-cow-seq-mt
	cpufreq_governor: performance

test-description: The motivation behind this suite is to exercise functions and regions of the mm/ of the Linux kernel which are of interest to us.
test-url: https://git.kernel.org/cgit/linux/kernel/git/wfg/vm-scalability.git/





Details are as below:
-------------------------------------------------------------------------------------------------->


To reproduce:

        git clone https://github.com/intel/lkp-tests.git
        cd lkp-tests
        sudo bin/lkp install job.yaml           # job file is attached in this email
        bin/lkp split-job --compatible job.yaml # generate the yaml file for lkp run
        sudo bin/lkp run generated-yaml-file

        # if come across any failure that blocks the test,
        # please remove ~/.lkp and /lkp dir to run from a clean state.

=========================================================================================
compiler/cpufreq_governor/kconfig/rootfs/runtime/size/tbox_group/test/testcase:
  gcc-12/performance/x86_64-rhel-8.3/debian-11.1-x86_64-20220510.cgz/300s/8T/lkp-csl-2sp3/anon-cow-seq-mt/vm-scalability

commit: 
  895f5ee464 ("mm/khugepaged: allow pte_offset_map[_lock]() to fail")
  3db82b9374 ("mm/memory: allow pte_offset_map[_lock]() to fail")

895f5ee464cc90a5 3db82b9374ca921b8b820a75e83 
---------------- --------------------------- 
         %stddev     %change         %stddev
             \          |                \  
     44083           +39.9%      61676        vm-scalability.median
   4232016           +39.9%    5920976        vm-scalability.throughput
 2.183e+08           +58.1%  3.451e+08        vm-scalability.time.minor_page_faults
     26859            -6.5%      25105        vm-scalability.time.system_time
    900.08          +185.0%       2565        vm-scalability.time.user_time
 1.271e+09           +39.9%  1.778e+09        vm-scalability.workload
      1.36 ±  4%      +0.4        1.78 ±  3%  mpstat.cpu.all.irq%
      3.16            +5.7        8.89        mpstat.cpu.all.usr%
      1742 ±  3%      -8.9%       1586 ±  2%  vmstat.system.cs
    811989           +50.8%    1224248        vmstat.system.in
 4.639e+08           +54.1%  7.151e+08        turbostat.IRQ
    272.15            +2.2%     278.16        turbostat.PkgWatt
     38.54            +4.1%      40.11        turbostat.RAMWatt
 1.028e+08           +26.1%  1.297e+08 ±  2%  numa-numastat.node0.local_node
 1.029e+08           +26.1%  1.298e+08 ±  2%  numa-numastat.node0.numa_hit
 1.032e+08           +22.9%  1.268e+08        numa-numastat.node1.local_node
 1.032e+08           +22.9%  1.269e+08        numa-numastat.node1.numa_hit
    101.67 ±  7%     +40.0%     142.33 ± 11%  perf-c2c.DRAM.local
      2203 ±  5%     +13.1%       2491 ±  7%  perf-c2c.DRAM.remote
      2821 ±  5%     +40.8%       3973 ±  7%  perf-c2c.HITM.local
      1631 ±  5%     +15.5%       1883 ±  5%  perf-c2c.HITM.remote
   9071420 ±119%    +223.0%   29299562 ± 36%  numa-meminfo.node0.AnonPages
  12743852 ± 77%    +151.1%   32001756 ± 29%  numa-meminfo.node0.AnonPages.max
   9095440 ±119%    +222.6%   29344203 ± 35%  numa-meminfo.node0.Inactive
   9094274 ±119%    +222.7%   29343307 ± 35%  numa-meminfo.node0.Inactive(anon)
  54437513 ± 20%     -38.1%   33706726 ± 31%  numa-meminfo.node0.MemFree
  11366617 ± 99%    +182.4%   32097404 ± 32%  numa-meminfo.node0.MemUsed
     22950 ±204%    +366.5%     107057 ± 43%  numa-meminfo.node0.PageTables
    107196 ± 43%     -78.2%      23373 ±200%  numa-meminfo.node1.PageTables
   2268035 ±119%    +222.7%    7318000 ± 36%  numa-vmstat.node0.nr_anon_pages
  13609253 ± 20%     -38.0%    8436220 ± 31%  numa-vmstat.node0.nr_free_pages
   2273746 ±119%    +222.2%    7326279 ± 35%  numa-vmstat.node0.nr_inactive_anon
      5737 ±204%    +366.1%      26741 ± 43%  numa-vmstat.node0.nr_page_table_pages
   2273746 ±119%    +222.2%    7326277 ± 35%  numa-vmstat.node0.nr_zone_inactive_anon
 1.029e+08           +26.1%  1.298e+08 ±  2%  numa-vmstat.node0.numa_hit
 1.028e+08           +26.1%  1.297e+08 ±  2%  numa-vmstat.node0.numa_local
     26811 ± 43%     -78.2%       5842 ±200%  numa-vmstat.node1.nr_page_table_pages
 1.032e+08           +22.9%  1.269e+08        numa-vmstat.node1.numa_hit
 1.032e+08           +22.9%  1.268e+08        numa-vmstat.node1.numa_local
   2550138 ±221%     -99.1%      23707 ±  7%  sched_debug.cfs_rq:/.load.max
    190971 ± 18%     -25.9%     141526 ± 10%  sched_debug.cfs_rq:/.min_vruntime.stddev
      6.70 ± 18%   +1194.0%      86.64 ±134%  sched_debug.cfs_rq:/.removed.load_avg.avg
     32.66 ±  9%   +1826.5%     629.27 ±135%  sched_debug.cfs_rq:/.removed.load_avg.stddev
      2.64 ± 29%     +67.9%       4.44 ± 14%  sched_debug.cfs_rq:/.removed.runnable_avg.avg
     13.55 ± 17%     +32.4%      17.94 ±  5%  sched_debug.cfs_rq:/.removed.runnable_avg.stddev
      2.64 ± 29%     +67.9%       4.44 ± 14%  sched_debug.cfs_rq:/.removed.util_avg.avg
     13.55 ± 17%     +32.4%      17.94 ±  5%  sched_debug.cfs_rq:/.removed.util_avg.stddev
    190872 ± 18%     -25.9%     141397 ± 10%  sched_debug.cfs_rq:/.spread0.stddev
     21367 ±  8%     +42.1%      30364 ± 53%  sched_debug.cpu.avg_idle.min
   9282723            +7.2%    9947921        proc-vmstat.nr_anon_pages
   2247296            -2.9%    2181210        proc-vmstat.nr_dirty_background_threshold
   4500089            -2.9%    4367755        proc-vmstat.nr_dirty_threshold
  22659550            -2.9%   21997719        proc-vmstat.nr_free_pages
   9304294            +7.1%    9966815        proc-vmstat.nr_inactive_anon
     19877            +1.3%      20136        proc-vmstat.nr_kernel_stack
   9304294            +7.1%    9966815        proc-vmstat.nr_zone_inactive_anon
 2.061e+08           +24.5%  2.566e+08        proc-vmstat.numa_hit
  2.06e+08           +24.5%  2.565e+08        proc-vmstat.numa_local
  10189086 ±  2%     +14.1%   11621781        proc-vmstat.numa_pte_updates
 2.144e+08           +23.6%  2.649e+08        proc-vmstat.pgalloc_normal
 2.191e+08           +57.9%   3.46e+08        proc-vmstat.pgfault
 2.113e+08 ±  2%     +25.1%  2.643e+08        proc-vmstat.pgfree
      4455           +57.9%       7033        proc-vmstat.thp_split_pmd
      3.13 ± 16%     +24.9%       3.90 ±  4%  perf-sched.sch_delay.avg.ms.do_task_dead.do_exit.do_group_exit.__x64_sys_exit_group.do_syscall_64
      0.04 ±  4%     -34.8%       0.03 ±  8%  perf-sched.sch_delay.avg.ms.schedule_timeout.__wait_for_common.wait_for_completion_state.kernel_clone
      0.03 ±  9%     -25.3%       0.02 ± 10%  perf-sched.sch_delay.avg.ms.schedule_timeout.kcompactd.kthread.ret_from_fork
      0.04 ± 25%     -41.0%       0.02 ±  2%  perf-sched.sch_delay.avg.ms.schedule_timeout.rcu_gp_fqs_loop.rcu_gp_kthread.kthread
      0.04 ±  5%     -19.0%       0.03 ±  8%  perf-sched.sch_delay.avg.ms.worker_thread.kthread.ret_from_fork
     18.15 ± 11%     -25.3%      13.57 ± 24%  perf-sched.sch_delay.max.ms.__cond_resched.stop_one_cpu.sched_exec.bprm_execve.part
      0.07 ± 11%     -30.4%       0.05 ± 17%  perf-sched.sch_delay.max.ms.schedule_hrtimeout_range_clock.do_poll.constprop.0.do_sys_poll
      0.11 ±  4%     -29.2%       0.08 ± 11%  perf-sched.sch_delay.max.ms.schedule_timeout.__wait_for_common.wait_for_completion_state.kernel_clone
      4.58 ±219%     -98.4%       0.07 ±  3%  perf-sched.sch_delay.max.ms.schedule_timeout.rcu_gp_fqs_loop.rcu_gp_kthread.kthread
      0.08 ±  8%     -19.4%       0.07 ±  9%  perf-sched.sch_delay.max.ms.smpboot_thread_fn.kthread.ret_from_fork
    127.50 ±  3%     +10.0%     140.24 ±  4%  perf-sched.total_wait_and_delay.average.ms
    127.37 ±  3%     +10.0%     140.10 ±  4%  perf-sched.total_wait_time.average.ms
      4.64 ±  4%     +38.1%       6.41 ±  7%  perf-sched.wait_and_delay.avg.ms.schedule_timeout.rcu_gp_fqs_loop.rcu_gp_kthread.kthread
      1151 ±  3%     -49.0%     587.00 ±  7%  perf-sched.wait_and_delay.count.__cond_resched.__alloc_pages.__folio_alloc.vma_alloc_folio.wp_page_copy
     79.33 ±  8%    +133.2%     185.00 ±  9%  perf-sched.wait_and_delay.count.exit_to_user_mode_loop.exit_to_user_mode_prepare.irqentry_exit_to_user_mode.asm_sysvec_apic_timer_interrupt
      1082 ±  4%     -27.9%     780.33 ±  7%  perf-sched.wait_and_delay.count.schedule_timeout.rcu_gp_fqs_loop.rcu_gp_kthread.kthread
     90.44 ± 80%    +231.2%     299.53 ± 11%  perf-sched.wait_and_delay.max.ms.schedule_timeout.rcu_gp_fqs_loop.rcu_gp_kthread.kthread
      0.83 ± 44%    +231.7%       2.74 ± 10%  perf-sched.wait_time.avg.ms.rcu_gp_kthread.kthread.ret_from_fork
      4.60 ±  4%     +38.7%       6.39 ±  7%  perf-sched.wait_time.avg.ms.schedule_timeout.rcu_gp_fqs_loop.rcu_gp_kthread.kthread
      0.85 ± 44%    +487.1%       5.00        perf-sched.wait_time.max.ms.rcu_gp_kthread.kthread.ret_from_fork
     90.37 ± 80%    +231.4%     299.50 ± 11%  perf-sched.wait_time.max.ms.schedule_timeout.rcu_gp_fqs_loop.rcu_gp_kthread.kthread
      6.11 ±  5%     +34.5%       8.22        perf-stat.i.MPKI
 9.598e+09           +18.8%   1.14e+10        perf-stat.i.branch-instructions
      0.21 ± 11%      -0.0        0.19 ±  4%  perf-stat.i.branch-miss-rate%
  11862691            +1.7%   12069954        perf-stat.i.branch-misses
     32.45            -2.1       30.37        perf-stat.i.cache-miss-rate%
  68318956 ±  2%     +53.7%   1.05e+08        perf-stat.i.cache-misses
 2.101e+08 ±  2%     +64.5%  3.456e+08        perf-stat.i.cache-references
      1654 ±  3%      -9.7%       1493 ±  2%  perf-stat.i.context-switches
      7.76           -13.7%       6.70        perf-stat.i.cpi
    123.01            +3.1%     126.76        perf-stat.i.cpu-migrations
      4168 ±  2%     -34.7%       2724        perf-stat.i.cycles-between-cache-misses
   1099264 ± 17%    +158.9%    2845760 ± 37%  perf-stat.i.dTLB-load-misses
 9.027e+09           +15.7%  1.045e+10        perf-stat.i.dTLB-loads
      0.14            +0.0        0.16 ±  5%  perf-stat.i.dTLB-store-miss-rate%
   2231340           +65.1%    3684026 ±  5%  perf-stat.i.dTLB-store-misses
 1.606e+09           +42.8%  2.292e+09        perf-stat.i.dTLB-stores
     34.26            -8.8       25.48        perf-stat.i.iTLB-load-miss-rate%
   4704797           +59.0%    7482354        perf-stat.i.iTLB-loads
 3.671e+10           +15.9%  4.255e+10        perf-stat.i.instructions
     16351           +12.7%      18425 ±  2%  perf-stat.i.instructions-per-iTLB-miss
      0.14           +14.9%       0.16        perf-stat.i.ipc
    304.16           +74.8%     531.52        perf-stat.i.metric.K/sec
    212.91           +19.8%     255.03        perf-stat.i.metric.M/sec
    725948           +58.0%    1147192        perf-stat.i.minor-faults
   6009534 ±  3%     +21.2%    7284919        perf-stat.i.node-load-misses
    277285 ±  5%     +36.9%     379705 ±  3%  perf-stat.i.node-loads
  13317338          +115.3%   28678842        perf-stat.i.node-store-misses
    725948           +58.0%    1147192        perf-stat.i.page-faults
      5.72 ±  2%     +42.0%       8.12        perf-stat.overall.MPKI
      0.12            -0.0        0.11        perf-stat.overall.branch-miss-rate%
     32.52            -2.1       30.38        perf-stat.overall.cache-miss-rate%
      7.84           -13.6%       6.78        perf-stat.overall.cpi
      4217 ±  2%     -34.8%       2748        perf-stat.overall.cycles-between-cache-misses
      0.01 ± 17%      +0.0        0.03 ± 37%  perf-stat.overall.dTLB-load-miss-rate%
      0.14            +0.0        0.16 ±  5%  perf-stat.overall.dTLB-store-miss-rate%
     34.59            -9.2       25.35        perf-stat.overall.iTLB-load-miss-rate%
     14765           +13.5%      16758 ±  2%  perf-stat.overall.instructions-per-iTLB-miss
      0.13           +15.7%       0.15        perf-stat.overall.ipc
      8706           -17.2%       7210        perf-stat.overall.path-length
 9.566e+09           +18.8%  1.136e+10        perf-stat.ps.branch-instructions
  11819705            +1.7%   12024429        perf-stat.ps.branch-misses
  68088826 ±  2%     +53.7%  1.046e+08        perf-stat.ps.cache-misses
 2.094e+08 ±  2%     +64.5%  3.444e+08        perf-stat.ps.cache-references
      1648 ±  3%      -9.6%       1489 ±  2%  perf-stat.ps.context-switches
    122.76            +3.1%     126.57        perf-stat.ps.cpu-migrations
   1096296 ± 17%    +158.8%    2836984 ± 37%  perf-stat.ps.dTLB-load-misses
 8.998e+09           +15.7%  1.041e+10        perf-stat.ps.dTLB-loads
   2223854           +65.1%    3671206 ±  5%  perf-stat.ps.dTLB-store-misses
   1.6e+09           +42.8%  2.284e+09        perf-stat.ps.dTLB-stores
   4688785           +59.0%    7455950        perf-stat.ps.iTLB-loads
 3.659e+10           +15.9%   4.24e+10        perf-stat.ps.instructions
    723488           +58.0%    1143177        perf-stat.ps.minor-faults
   5989152 ±  3%     +21.2%    7259410        perf-stat.ps.node-load-misses
    277102 ±  5%     +36.7%     378929 ±  3%  perf-stat.ps.node-loads
  13272143          +115.3%   28578623        perf-stat.ps.node-store-misses
    723488           +58.0%    1143178        perf-stat.ps.page-faults
 1.107e+13           +15.9%  1.282e+13        perf-stat.total.instructions
     54.54            -7.2       47.29        perf-profile.calltrace.cycles-pp.native_queued_spin_lock_slowpath._raw_spin_lock.__pte_offset_map_lock.wp_page_copy.__handle_mm_fault
     54.79            -7.2       47.61        perf-profile.calltrace.cycles-pp._raw_spin_lock.__pte_offset_map_lock.wp_page_copy.__handle_mm_fault.handle_mm_fault
     54.80            -7.2       47.62        perf-profile.calltrace.cycles-pp.__pte_offset_map_lock.wp_page_copy.__handle_mm_fault.handle_mm_fault.do_user_addr_fault
     95.87            -6.7       89.13        perf-profile.calltrace.cycles-pp.__handle_mm_fault.handle_mm_fault.do_user_addr_fault.exc_page_fault.asm_exc_page_fault
     55.94            -6.7       49.25        perf-profile.calltrace.cycles-pp.wp_page_copy.__handle_mm_fault.handle_mm_fault.do_user_addr_fault.exc_page_fault
     95.93            -6.7       89.25        perf-profile.calltrace.cycles-pp.handle_mm_fault.do_user_addr_fault.exc_page_fault.asm_exc_page_fault.do_access
     96.36            -6.1       90.26        perf-profile.calltrace.cycles-pp.do_user_addr_fault.exc_page_fault.asm_exc_page_fault.do_access
     96.36            -6.1       90.29        perf-profile.calltrace.cycles-pp.exc_page_fault.asm_exc_page_fault.do_access
     96.83            -5.4       91.46        perf-profile.calltrace.cycles-pp.asm_exc_page_fault.do_access
      0.60 ± 47%      +0.3        0.92 ±  8%  perf-profile.calltrace.cycles-pp.asm_sysvec_call_function.native_queued_spin_lock_slowpath._raw_spin_lock.__pte_offset_map_lock.wp_page_copy
      0.42 ± 71%      +0.3        0.76 ±  9%  perf-profile.calltrace.cycles-pp.sysvec_call_function.asm_sysvec_call_function.native_queued_spin_lock_slowpath._raw_spin_lock.__pte_offset_map_lock
      0.30 ±100%      +0.4        0.70 ±  9%  perf-profile.calltrace.cycles-pp.__flush_smp_call_function_queue.__sysvec_call_function.sysvec_call_function.asm_sysvec_call_function.native_queued_spin_lock_slowpath
      0.30 ±100%      +0.4        0.71 ± 10%  perf-profile.calltrace.cycles-pp.__sysvec_call_function.sysvec_call_function.asm_sysvec_call_function.native_queued_spin_lock_slowpath._raw_spin_lock
      0.00            +0.7        0.71 ±  6%  perf-profile.calltrace.cycles-pp.lock_vma_under_rcu.do_user_addr_fault.exc_page_fault.asm_exc_page_fault.do_access
      1.32 ±  2%      +1.1        2.46 ±  2%  perf-profile.calltrace.cycles-pp.do_rw_once
     99.21            +2.3      101.48        perf-profile.calltrace.cycles-pp.do_access
     94.25            -7.5       86.78        perf-profile.children.cycles-pp.native_queued_spin_lock_slowpath
     94.52            -7.3       87.22        perf-profile.children.cycles-pp._raw_spin_lock
     54.86            -7.2       47.69        perf-profile.children.cycles-pp.__pte_offset_map_lock
     95.88            -6.7       89.13        perf-profile.children.cycles-pp.__handle_mm_fault
     55.94            -6.7       49.25        perf-profile.children.cycles-pp.wp_page_copy
     95.93            -6.7       89.26        perf-profile.children.cycles-pp.handle_mm_fault
     96.36            -6.1       90.27        perf-profile.children.cycles-pp.do_user_addr_fault
     96.37            -6.1       90.29        perf-profile.children.cycles-pp.exc_page_fault
     96.83            -5.5       91.30        perf-profile.children.cycles-pp.asm_exc_page_fault
     98.54            -1.0       97.52        perf-profile.children.cycles-pp.do_access
      0.08 ± 12%      +0.0        0.11 ± 15%  perf-profile.children.cycles-pp.mas_walk
      0.06 ± 11%      +0.0        0.09 ±  4%  perf-profile.children.cycles-pp.__mem_cgroup_charge
      0.06 ±  7%      +0.0        0.10 ±  4%  perf-profile.children.cycles-pp.sync_regs
      0.00            +0.1        0.05        perf-profile.children.cycles-pp.propagate_protected_usage
      0.08 ± 10%      +0.1        0.13 ±  5%  perf-profile.children.cycles-pp.llist_reverse_order
      0.05 ±  8%      +0.1        0.10 ±  4%  perf-profile.children.cycles-pp.___perf_sw_event
      0.00            +0.1        0.05 ±  7%  perf-profile.children.cycles-pp.irqtime_account_irq
      0.08 ±  4%      +0.1        0.13 ±  3%  perf-profile.children.cycles-pp.llist_add_batch
      0.02 ±141%      +0.1        0.07 ±  6%  perf-profile.children.cycles-pp.__count_memcg_events
      0.06 ±  7%      +0.1        0.12 ±  5%  perf-profile.children.cycles-pp.__perf_sw_event
      0.26 ±  3%      +0.1        0.34 ±  3%  perf-profile.children.cycles-pp.copy_mc_fragile
      0.26 ±  4%      +0.1        0.34 ±  4%  perf-profile.children.cycles-pp.__wp_page_copy_user
      0.19 ± 12%      +0.1        0.28 ±  5%  perf-profile.children.cycles-pp.ptep_clear_flush
      0.19 ± 12%      +0.1        0.28 ±  5%  perf-profile.children.cycles-pp.flush_tlb_mm_range
      0.19 ± 12%      +0.1        0.28 ±  5%  perf-profile.children.cycles-pp.on_each_cpu_cond_mask
      0.19 ± 12%      +0.1        0.28 ±  5%  perf-profile.children.cycles-pp.smp_call_function_many_cond
      0.01 ±223%      +0.1        0.10 ±  8%  perf-profile.children.cycles-pp.do_huge_pmd_wp_page
      0.16 ±  4%      +0.1        0.25 ±  4%  perf-profile.children.cycles-pp.native_irq_return_iret
      0.10 ±  3%      +0.1        0.22 ±  4%  perf-profile.children.cycles-pp.up_read
      0.17 ±  6%      +0.2        0.32 ±  4%  perf-profile.children.cycles-pp.page_counter_uncharge
      0.20 ±  5%      +0.2        0.38 ±  3%  perf-profile.children.cycles-pp.uncharge_batch
      0.19 ±  5%      +0.2        0.38 ±  4%  perf-profile.children.cycles-pp.irqentry_exit_to_user_mode
      0.22 ±  6%      +0.2        0.42 ±  4%  perf-profile.children.cycles-pp.__mem_cgroup_uncharge
      0.23 ±  4%      +0.2        0.42 ±  4%  perf-profile.children.cycles-pp.__folio_put
      0.32 ± 31%      +0.3        0.58 ± 11%  perf-profile.children.cycles-pp.flush_tlb_func
      0.14 ±  7%      +0.3        0.42 ±  7%  perf-profile.children.cycles-pp.down_read_trylock
      0.54 ± 22%      +0.4        0.95 ± 10%  perf-profile.children.cycles-pp.__sysvec_call_function
      0.53 ± 23%      +0.4        0.94 ±  9%  perf-profile.children.cycles-pp.__flush_smp_call_function_queue
      0.28 ±  5%      +0.4        0.71 ±  6%  perf-profile.children.cycles-pp.lock_vma_under_rcu
      0.58 ± 22%      +0.4        1.03 ±  9%  perf-profile.children.cycles-pp.sysvec_call_function
      0.72 ± 18%      +0.5        1.25 ±  8%  perf-profile.children.cycles-pp.asm_sysvec_call_function
      2.26 ±  2%      +4.8        7.04 ±  3%  perf-profile.children.cycles-pp.do_rw_once
     93.23            -7.9       85.31        perf-profile.self.cycles-pp.native_queued_spin_lock_slowpath
      0.06 ±  7%      +0.0        0.10 ±  4%  perf-profile.self.cycles-pp.sync_regs
      0.00            +0.1        0.05        perf-profile.self.cycles-pp.propagate_protected_usage
      0.08 ± 10%      +0.1        0.13 ±  5%  perf-profile.self.cycles-pp.llist_reverse_order
      0.00            +0.1        0.05 ±  8%  perf-profile.self.cycles-pp.smp_call_function_many_cond
      0.08 ±  4%      +0.1        0.13 ±  3%  perf-profile.self.cycles-pp.llist_add_batch
      0.00            +0.1        0.06 ±  7%  perf-profile.self.cycles-pp.__count_memcg_events
      0.02 ± 99%      +0.1        0.10 ±  5%  perf-profile.self.cycles-pp.___perf_sw_event
      0.26 ±  3%      +0.1        0.34 ±  3%  perf-profile.self.cycles-pp.copy_mc_fragile
      0.04 ± 71%      +0.1        0.11 ± 14%  perf-profile.self.cycles-pp.__handle_mm_fault
      0.16 ±  5%      +0.1        0.25 ±  4%  perf-profile.self.cycles-pp.native_irq_return_iret
      0.13 ± 12%      +0.1        0.24 ±  9%  perf-profile.self.cycles-pp.__flush_smp_call_function_queue
      0.10 ±  3%      +0.1        0.22 ±  4%  perf-profile.self.cycles-pp.up_read
      0.15 ±  7%      +0.1        0.27 ±  5%  perf-profile.self.cycles-pp.page_counter_uncharge
      0.06 ± 11%      +0.1        0.19 ±  6%  perf-profile.self.cycles-pp.lock_vma_under_rcu
      0.14 ±  4%      +0.2        0.29        perf-profile.self.cycles-pp.wp_page_copy
      0.28 ±  5%      +0.2        0.45        perf-profile.self.cycles-pp._raw_spin_lock
      0.18 ±  5%      +0.2        0.38 ±  4%  perf-profile.self.cycles-pp.irqentry_exit_to_user_mode
      0.22 ± 39%      +0.2        0.45 ± 15%  perf-profile.self.cycles-pp.flush_tlb_func
      0.14 ±  7%      +0.3        0.42 ±  7%  perf-profile.self.cycles-pp.down_read_trylock
      0.85 ±  2%      +1.1        1.93 ±  2%  perf-profile.self.cycles-pp.do_access
      2.19 ±  2%      +4.7        6.91 ±  3%  perf-profile.self.cycles-pp.do_rw_once




Disclaimer:
Results have been estimated based on internal Intel analysis and are provided
for informational purposes only. Any difference in system hardware or software
design or configuration may affect actual performance.


-- 
0-DAY CI Kernel Test Service
https://github.com/intel/lkp-tests/wiki



View attachment "config-6.4.0-rc4-00265-g3db82b9374ca" of type "text/plain" (158813 bytes)

View attachment "job-script" of type "text/plain" (8711 bytes)

View attachment "job.yaml" of type "text/plain" (6417 bytes)

View attachment "reproduce" of type "text/plain" (796 bytes)

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ