lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [thread-next>] [day] [month] [year] [list]
Message-ID: <202303192325.ecbaf968-yujie.liu@intel.com>
Date:   Mon, 20 Mar 2023 09:12:16 +0800
From:   kernel test robot <yujie.liu@...el.com>
To:     Huang Ying <ying.huang@...el.com>
CC:     <oe-lkp@...ts.linux.dev>, <lkp@...el.com>,
        <linux-kernel@...r.kernel.org>,
        Andrew Morton <akpm@...ux-foundation.org>,
        Xin Hao <xhao@...ux.alibaba.com>, Zi Yan <ziy@...dia.com>,
        Yang Shi <shy828301@...il.com>,
        Baolin Wang <baolin.wang@...ux.alibaba.com>,
        Oscar Salvador <osalvador@...e.de>,
        "Matthew Wilcox" <willy@...radead.org>,
        Bharata B Rao <bharata@....com>,
        "Alistair Popple" <apopple@...dia.com>,
        Minchan Kim <minchan@...nel.org>,
        Mike Kravetz <mike.kravetz@...cle.com>,
        Hyeonggon Yoo <42.hyeyoo@...il.com>, <linux-mm@...ck.org>,
        <ying.huang@...el.com>, <feng.tang@...el.com>,
        <zhengjun.xing@...ux.intel.com>, <fengwei.yin@...el.com>
Subject: [linus:master] [migrate_pages] 7e12beb8ca: vm-scalability.throughput
 -3.4% regression

Hello,

FYI, we noticed a -3.4% regression of vm-scalability.throughput due to commit:

commit: 7e12beb8ca2ac98b2ec42e0ea4b76cdc93b58654 ("migrate_pages: batch flushing TLB")
https://git.kernel.org/cgit/linux/kernel/git/torvalds/linux.git master

in testcase: vm-scalability
on test machine: 96 threads 2 sockets Intel(R) Xeon(R) Platinum 8260L CPU @ 2.40GHz (Cascade Lake) with 128G memory
with following parameters:

	runtime: 300s
	size: 512G
	test: anon-cow-rand-mt
	cpufreq_governor: performance

test-description: The motivation behind this suite is to exercise functions and regions of the mm/ of the Linux kernel which are of interest to us.
test-url: https://git.kernel.org/cgit/linux/kernel/git/wfg/vm-scalability.git/


If you fix the issue, kindly add following tag
| Reported-by: kernel test robot <yujie.liu@...el.com>
| Link: https://lore.kernel.org/oe-lkp/202303192325.ecbaf968-yujie.liu@intel.com


Details are as below:

=========================================================================================
compiler/cpufreq_governor/kconfig/rootfs/runtime/size/tbox_group/test/testcase:
  gcc-11/performance/x86_64-rhel-8.3/debian-11.1-x86_64-20220510.cgz/300s/512G/lkp-csl-2sp3/anon-cow-rand-mt/vm-scalability

commit: 
  ebe75e4751 ("migrate_pages: share more code between _unmap and _move")
  7e12beb8ca ("migrate_pages: batch flushing TLB")

ebe75e4751063dce 7e12beb8ca2ac98b2ec42e0ea4b 
---------------- --------------------------- 
         %stddev     %change         %stddev
             \          |                \  
     57522            -3.3%      55603        vm-scalability.median
   5513665            -3.4%    5328506        vm-scalability.throughput
    203067 ±  3%      -8.6%     185675 ±  2%  vm-scalability.time.involuntary_context_switches
  68459282 ±  6%     +42.1%   97269013 ±  2%  vm-scalability.time.minor_page_faults
      9007            -1.8%       8844        vm-scalability.time.percent_of_cpu_this_job_got
      1170 ±  3%     +58.2%       1852 ±  3%  vm-scalability.time.system_time
     26342            -4.6%      25132        vm-scalability.time.user_time
     11275 ±  5%    +364.1%      52332 ±  7%  vm-scalability.time.voluntary_context_switches
 1.658e+09            -3.4%  1.601e+09        vm-scalability.workload
     51013 ± 40%     -67.5%      16584 ±125%  numa-numastat.node1.other_node
     20056 ±  2%      +8.9%      21844 ±  3%  numa-vmstat.node1.nr_slab_unreclaimable
     51013 ± 40%     -67.5%      16584 ±125%  numa-vmstat.node1.numa_other
      2043 ±  3%     +10.5%       2257        vmstat.system.cs
    540820 ±  2%    +186.2%    1547747 ±  8%  vmstat.system.in
      0.00 ±157%      +0.0        0.00 ±  6%  mpstat.cpu.all.iowait%
      2.59            +1.7        4.27 ±  4%  mpstat.cpu.all.irq%
      4.03 ±  3%      +2.3        6.36 ±  3%  mpstat.cpu.all.sys%
      5870 ± 64%     -48.0%       3051 ±  7%  numa-meminfo.node0.Active
    195543 ±  3%      -7.2%     181529 ±  4%  numa-meminfo.node0.Slab
     80226 ±  2%      +8.9%      87378 ±  3%  numa-meminfo.node1.SUnreclaim
  40406018 ±  7%     +66.5%   67272793 ±  2%  proc-vmstat.numa_hint_faults
  20211075 ±  7%     +66.8%   33722069 ±  2%  proc-vmstat.numa_hint_faults_local
  40555366 ±  7%     +66.3%   67430626 ±  2%  proc-vmstat.numa_pte_updates
  69364615 ±  6%     +41.5%   98184580 ±  2%  proc-vmstat.pgfault
    210031 ±  8%    +126.2%     475135 ± 99%  turbostat.C1
 1.382e+09 ±  2%    +140.0%  3.317e+09 ±  5%  turbostat.IRQ
      8771 ±  6%    +466.6%      49695 ±  7%  turbostat.POLL
     87.01            -2.6%      84.76        turbostat.RAMWatt
    145904 ±  2%     -22.2%     113504 ± 11%  sched_debug.cfs_rq:/.min_vruntime.stddev
    841.83 ±  2%     -13.8%     725.47 ±  6%  sched_debug.cfs_rq:/.runnable_avg.min
    549777 ±  9%     -49.2%     279239 ± 34%  sched_debug.cfs_rq:/.spread0.avg
    659447 ±  8%     -36.7%     417735 ± 22%  sched_debug.cfs_rq:/.spread0.max
    145800 ±  2%     -22.1%     113612 ± 11%  sched_debug.cfs_rq:/.spread0.stddev
    785.23 ±  6%     -14.6%     670.61 ± 10%  sched_debug.cfs_rq:/.util_avg.min
     67.96 ±  5%     +22.7%      83.40 ± 10%  sched_debug.cfs_rq:/.util_avg.stddev
    246549 ±  7%     -15.1%     209367 ±  7%  sched_debug.cpu.avg_idle.avg
      1592           +10.8%       1763 ±  3%  sched_debug.cpu.clock_task.stddev
     32106 ± 10%     -17.6%      26468 ±  8%  sched_debug.cpu.nr_switches.max
      1910 ±  6%     +31.0%       2503        sched_debug.cpu.nr_switches.min
      5664 ± 10%     -16.6%       4723 ±  7%  sched_debug.cpu.nr_switches.stddev
      0.18 ±  4%      +0.0        0.23 ±  3%  perf-stat.i.branch-miss-rate%
   8939520 ±  4%     +61.3%   14417578 ±  3%  perf-stat.i.branch-misses
     66.18            -1.7       64.47        perf-stat.i.cache-miss-rate%
      1927 ±  3%     +11.0%       2139        perf-stat.i.context-switches
    158.85           +10.7%     175.92 ±  3%  perf-stat.i.cpu-migrations
      0.04 ±  6%      +0.0        0.05 ± 11%  perf-stat.i.dTLB-load-miss-rate%
   4916471 ±  7%     +39.7%    6870029 ±  9%  perf-stat.i.dTLB-load-misses
      9.10            -0.4        8.71        perf-stat.i.dTLB-store-miss-rate%
 5.311e+08            -4.1%  5.095e+08        perf-stat.i.dTLB-store-misses
   2438160 ±  2%    +161.5%    6374895 ±  7%  perf-stat.i.iTLB-load-misses
    115315 ±  2%     +62.0%     186840 ±  7%  perf-stat.i.iTLB-loads
     43163 ±  5%     -25.7%      32083 ± 26%  perf-stat.i.instructions-per-iTLB-miss
      0.34 ± 37%     -63.2%       0.13 ± 27%  perf-stat.i.major-faults
    226565 ±  6%     +41.4%     320417 ±  2%  perf-stat.i.minor-faults
     50.56            +1.7       52.22        perf-stat.i.node-load-miss-rate%
 1.165e+08            +3.7%  1.208e+08        perf-stat.i.node-load-misses
  1.13e+08            -3.6%  1.089e+08        perf-stat.i.node-loads
 2.678e+08            -3.6%  2.582e+08        perf-stat.i.node-store-misses
 2.655e+08            -4.2%  2.543e+08        perf-stat.i.node-stores
    226565 ±  6%     +41.4%     320418 ±  2%  perf-stat.i.page-faults
      0.08 ±  4%      +0.0        0.12 ±  4%  perf-stat.overall.branch-miss-rate%
     67.13            -1.8       65.28        perf-stat.overall.cache-miss-rate%
    367.93            +2.6%     377.43        perf-stat.overall.cycles-between-cache-misses
      0.04 ±  7%      +0.0        0.05 ± 10%  perf-stat.overall.dTLB-load-miss-rate%
      9.38            -0.4        8.97        perf-stat.overall.dTLB-store-miss-rate%
     95.49            +1.7       97.16        perf-stat.overall.iTLB-load-miss-rate%
     20560 ±  3%     -61.9%       7826 ±  7%  perf-stat.overall.instructions-per-iTLB-miss
     50.76            +1.8       52.60        perf-stat.overall.node-load-miss-rate%
      9205            +3.1%       9485        perf-stat.overall.path-length
   8892515 ±  4%     +62.1%   14412101 ±  3%  perf-stat.ps.branch-misses
      1927 ±  3%     +11.1%       2142        perf-stat.ps.context-switches
    158.37           +11.2%     176.03 ±  3%  perf-stat.ps.cpu-migrations
   4902779 ±  7%     +40.2%    6871859 ±  9%  perf-stat.ps.dTLB-load-misses
 5.295e+08            -4.1%  5.077e+08        perf-stat.ps.dTLB-store-misses
   2428324 ±  2%    +163.0%    6385873 ±  7%  perf-stat.ps.iTLB-load-misses
    114618 ±  2%     +62.5%     186290 ±  7%  perf-stat.ps.iTLB-loads
      0.34 ± 37%     -63.2%       0.13 ± 27%  perf-stat.ps.major-faults
    226036 ±  6%     +41.8%     320615 ±  2%  perf-stat.ps.minor-faults
 1.162e+08            +3.7%  1.205e+08        perf-stat.ps.node-load-misses
 1.127e+08            -3.6%  1.086e+08        perf-stat.ps.node-loads
  2.67e+08            -3.6%  2.573e+08        perf-stat.ps.node-store-misses
 2.647e+08            -4.3%  2.534e+08        perf-stat.ps.node-stores
    226036 ±  6%     +41.8%     320615 ±  2%  perf-stat.ps.page-faults
      0.00            +0.6        0.60 ±  8%  perf-profile.calltrace.cycles-pp.default_send_IPI_mask_sequence_phys.smp_call_function_many_cond.on_each_cpu_cond_mask.arch_tlbbatch_flush.try_to_unmap_flush
      0.00            +0.6        0.64 ±  7%  perf-profile.calltrace.cycles-pp.flush_tlb_func.__flush_smp_call_function_queue.__sysvec_call_function.sysvec_call_function.asm_sysvec_call_function
      0.00            +0.9        0.90 ± 10%  perf-profile.calltrace.cycles-pp.llist_reverse_order.__flush_smp_call_function_queue.__sysvec_call_function.sysvec_call_function.asm_sysvec_call_function
      0.00            +1.9        1.86 ±  9%  perf-profile.calltrace.cycles-pp.__flush_smp_call_function_queue.__sysvec_call_function.sysvec_call_function.asm_sysvec_call_function.do_access
      0.00            +1.9        1.87 ±  8%  perf-profile.calltrace.cycles-pp.__sysvec_call_function.sysvec_call_function.asm_sysvec_call_function.do_access
      0.00            +1.9        1.94 ±  8%  perf-profile.calltrace.cycles-pp.sysvec_call_function.asm_sysvec_call_function.do_access
      0.00            +2.6        2.59 ±  9%  perf-profile.calltrace.cycles-pp.asm_sysvec_call_function.do_access
      0.00            +2.8        2.80 ±  8%  perf-profile.calltrace.cycles-pp.llist_add_batch.smp_call_function_many_cond.on_each_cpu_cond_mask.arch_tlbbatch_flush.try_to_unmap_flush
      3.43 ± 13%      +6.5        9.88 ±  7%  perf-profile.calltrace.cycles-pp.__handle_mm_fault.handle_mm_fault.do_user_addr_fault.exc_page_fault.asm_exc_page_fault
      3.46 ± 13%      +6.5        9.94 ±  7%  perf-profile.calltrace.cycles-pp.handle_mm_fault.do_user_addr_fault.exc_page_fault.asm_exc_page_fault.do_access
      3.15 ± 13%      +6.5        9.69 ±  7%  perf-profile.calltrace.cycles-pp.do_numa_page.__handle_mm_fault.handle_mm_fault.do_user_addr_fault.exc_page_fault
      4.06 ± 11%      +6.7       10.71 ±  7%  perf-profile.calltrace.cycles-pp.asm_exc_page_fault.do_access
      3.82 ± 13%      +6.7       10.48 ±  7%  perf-profile.calltrace.cycles-pp.do_user_addr_fault.exc_page_fault.asm_exc_page_fault.do_access
      3.83 ± 13%      +6.7       10.49 ±  7%  perf-profile.calltrace.cycles-pp.exc_page_fault.asm_exc_page_fault.do_access
      2.57 ± 13%      +6.9        9.46 ±  7%  perf-profile.calltrace.cycles-pp.migrate_misplaced_page.do_numa_page.__handle_mm_fault.handle_mm_fault.do_user_addr_fault
      2.36 ± 13%      +6.9        9.28 ±  7%  perf-profile.calltrace.cycles-pp.migrate_pages_batch.migrate_pages.migrate_misplaced_page.do_numa_page.__handle_mm_fault
      2.36 ± 13%      +6.9        9.29 ±  7%  perf-profile.calltrace.cycles-pp.migrate_pages.migrate_misplaced_page.do_numa_page.__handle_mm_fault.handle_mm_fault
      0.00            +7.5        7.50 ±  7%  perf-profile.calltrace.cycles-pp.smp_call_function_many_cond.on_each_cpu_cond_mask.arch_tlbbatch_flush.try_to_unmap_flush.migrate_pages_batch
      0.00            +7.6        7.56 ±  7%  perf-profile.calltrace.cycles-pp.on_each_cpu_cond_mask.arch_tlbbatch_flush.try_to_unmap_flush.migrate_pages_batch.migrate_pages
      0.00            +7.6        7.57 ±  8%  perf-profile.calltrace.cycles-pp.arch_tlbbatch_flush.try_to_unmap_flush.migrate_pages_batch.migrate_pages.migrate_misplaced_page
      0.00            +7.6        7.57 ±  7%  perf-profile.calltrace.cycles-pp.try_to_unmap_flush.migrate_pages_batch.migrate_pages.migrate_misplaced_page.do_numa_page
      1.55 ± 13%      -1.1        0.42 ±  9%  perf-profile.children.cycles-pp.rmap_walk_anon
      1.30 ± 15%      -1.0        0.30 ±  9%  perf-profile.children.cycles-pp.native_queued_spin_lock_slowpath
      1.11 ± 12%      -0.9        0.18 ±  6%  perf-profile.children.cycles-pp.try_to_migrate_one
      1.17 ± 12%      -0.9        0.26 ±  8%  perf-profile.children.cycles-pp.try_to_migrate
      1.30 ± 12%      -0.9        0.42 ±  8%  perf-profile.children.cycles-pp.migrate_folio_unmap
      1.08 ± 13%      -0.9        0.21 ± 11%  perf-profile.children.cycles-pp._raw_spin_lock
      0.46 ± 14%      -0.3        0.14 ± 13%  perf-profile.children.cycles-pp.page_vma_mapped_walk
      0.35 ± 13%      -0.2        0.11 ± 11%  perf-profile.children.cycles-pp.remove_migration_pte
      0.14 ± 21%      -0.1        0.07 ± 11%  perf-profile.children.cycles-pp.folio_lruvec_lock_irq
      0.14 ± 21%      -0.1        0.08 ± 10%  perf-profile.children.cycles-pp._raw_spin_lock_irq
      0.33 ±  3%      -0.0        0.30        perf-profile.children.cycles-pp.lrand48_r@plt
      0.06 ± 14%      +0.0        0.08 ±  9%  perf-profile.children.cycles-pp.mt_find
      0.06 ± 14%      +0.0        0.08 ± 11%  perf-profile.children.cycles-pp.find_vma
      0.00            +0.1        0.06 ±  9%  perf-profile.children.cycles-pp.folio_migrate_flags
      0.06 ±  8%      +0.1        0.12 ±  8%  perf-profile.children.cycles-pp.exit_to_user_mode_loop
      0.03 ± 81%      +0.1        0.10 ±  8%  perf-profile.children.cycles-pp.uncharge_batch
      0.00            +0.1        0.07 ±  8%  perf-profile.children.cycles-pp.native_sched_clock
      0.06 ± 10%      +0.1        0.13 ±  8%  perf-profile.children.cycles-pp.exit_to_user_mode_prepare
      0.03 ± 81%      +0.1        0.10 ± 10%  perf-profile.children.cycles-pp.__folio_put
      0.03 ± 81%      +0.1        0.10 ± 10%  perf-profile.children.cycles-pp.__mem_cgroup_uncharge
      0.16 ± 12%      +0.1        0.24 ± 10%  perf-profile.children.cycles-pp.up_read
      0.04 ± 50%      +0.1        0.12 ±  8%  perf-profile.children.cycles-pp.task_work_run
      0.01 ±200%      +0.1        0.09 ± 12%  perf-profile.children.cycles-pp.page_counter_uncharge
      0.23 ± 18%      +0.1        0.31 ±  8%  perf-profile.children.cycles-pp.folio_batch_move_lru
      0.00            +0.1        0.08 ± 10%  perf-profile.children.cycles-pp.sched_clock_cpu
      0.23 ± 18%      +0.1        0.31 ±  8%  perf-profile.children.cycles-pp.lru_add_drain
      0.23 ± 18%      +0.1        0.31 ±  8%  perf-profile.children.cycles-pp.lru_add_drain_cpu
      0.19 ± 12%      +0.1        0.28 ± 11%  perf-profile.children.cycles-pp.down_read_trylock
      0.05 ±  7%      +0.1        0.14 ±  8%  perf-profile.children.cycles-pp.mem_cgroup_migrate
      0.00            +0.1        0.09 ± 10%  perf-profile.children.cycles-pp._find_next_bit
      0.03 ± 82%      +0.1        0.12 ±  8%  perf-profile.children.cycles-pp.change_pte_range
      0.03 ± 82%      +0.1        0.12 ±  8%  perf-profile.children.cycles-pp.task_numa_work
      0.03 ± 82%      +0.1        0.12 ±  8%  perf-profile.children.cycles-pp.change_prot_numa
      0.03 ± 82%      +0.1        0.12 ±  8%  perf-profile.children.cycles-pp.change_protection_range
      0.03 ± 82%      +0.1        0.12 ±  8%  perf-profile.children.cycles-pp.change_pmd_range
      0.02 ±123%      +0.1        0.12 ±  6%  perf-profile.children.cycles-pp.irqtime_account_irq
      0.07 ± 13%      +0.1        0.18 ± 24%  perf-profile.children.cycles-pp.__irq_exit_rcu
      0.02 ±122%      +0.1        0.13 ±  6%  perf-profile.children.cycles-pp.page_counter_charge
      0.18 ± 12%      +0.1        0.30 ±  9%  perf-profile.children.cycles-pp.folio_copy
      0.17 ± 13%      +0.1        0.30 ±  9%  perf-profile.children.cycles-pp.copy_page
      0.09 ±  4%      +0.1        0.24 ±  9%  perf-profile.children.cycles-pp.sync_regs
      0.27 ± 11%      +0.2        0.51 ±  8%  perf-profile.children.cycles-pp.move_to_new_folio
      0.27 ± 11%      +0.2        0.51 ±  8%  perf-profile.children.cycles-pp.migrate_folio_extra
      0.10 ±  9%      +0.3        0.40 ±  7%  perf-profile.children.cycles-pp.native_irq_return_iret
      0.07 ± 12%      +0.4        0.47 ±  9%  perf-profile.children.cycles-pp.__default_send_IPI_dest_field
      0.00            +0.4        0.44 ±  9%  perf-profile.children.cycles-pp.native_flush_tlb_local
      0.09 ±  9%      +0.5        0.62 ±  9%  perf-profile.children.cycles-pp.default_send_IPI_mask_sequence_phys
      0.09 ± 10%      +1.2        1.32 ±  7%  perf-profile.children.cycles-pp.flush_tlb_func
      0.26 ± 12%      +1.6        1.85 ±  9%  perf-profile.children.cycles-pp.llist_reverse_order
      0.42 ± 11%      +2.4        2.86 ±  8%  perf-profile.children.cycles-pp.llist_add_batch
      0.42 ± 11%      +3.3        3.76 ±  8%  perf-profile.children.cycles-pp.__sysvec_call_function
      0.42 ± 11%      +3.3        3.76 ±  8%  perf-profile.children.cycles-pp.__flush_smp_call_function_queue
      0.44 ± 11%      +3.5        3.90 ±  8%  perf-profile.children.cycles-pp.sysvec_call_function
      0.58 ±  6%      +4.4        4.95 ±  8%  perf-profile.children.cycles-pp.asm_sysvec_call_function
      3.44 ± 13%      +6.5        9.89 ±  7%  perf-profile.children.cycles-pp.__handle_mm_fault
      3.47 ± 13%      +6.5        9.95 ±  7%  perf-profile.children.cycles-pp.handle_mm_fault
      3.15 ± 13%      +6.5        9.69 ±  7%  perf-profile.children.cycles-pp.do_numa_page
      0.94 ± 12%      +6.7        7.59 ±  7%  perf-profile.children.cycles-pp.on_each_cpu_cond_mask
      0.94 ± 12%      +6.7        7.59 ±  7%  perf-profile.children.cycles-pp.smp_call_function_many_cond
      3.83 ± 13%      +6.7       10.49 ±  7%  perf-profile.children.cycles-pp.do_user_addr_fault
      3.84 ± 13%      +6.7       10.50 ±  7%  perf-profile.children.cycles-pp.exc_page_fault
      4.08 ± 11%      +6.7       10.76 ±  7%  perf-profile.children.cycles-pp.asm_exc_page_fault
      2.57 ± 13%      +6.9        9.46 ±  7%  perf-profile.children.cycles-pp.migrate_misplaced_page
      2.36 ± 13%      +6.9        9.28 ±  7%  perf-profile.children.cycles-pp.migrate_pages_batch
      2.36 ± 13%      +6.9        9.29 ±  7%  perf-profile.children.cycles-pp.migrate_pages
      0.00            +7.6        7.57 ±  7%  perf-profile.children.cycles-pp.try_to_unmap_flush
      0.00            +7.6        7.57 ±  7%  perf-profile.children.cycles-pp.arch_tlbbatch_flush
     67.74 ±  4%      -8.5       59.28 ±  2%  perf-profile.self.cycles-pp.do_access
      1.19 ± 15%      -0.9        0.28 ±  9%  perf-profile.self.cycles-pp.native_queued_spin_lock_slowpath
      0.03 ± 81%      +0.0        0.08 ± 10%  perf-profile.self.cycles-pp.change_pte_range
      0.09 ±  4%      +0.0        0.14 ± 20%  perf-profile.self.cycles-pp._raw_spin_lock
      0.15 ± 12%      +0.0        0.20 ± 10%  perf-profile.self.cycles-pp.up_read
      0.00            +0.1        0.05 ±  8%  perf-profile.self.cycles-pp.try_to_migrate_one
      0.00            +0.1        0.07 ±  8%  perf-profile.self.cycles-pp._find_next_bit
      0.00            +0.1        0.07 ±  8%  perf-profile.self.cycles-pp.native_sched_clock
      0.00            +0.1        0.07 ± 12%  perf-profile.self.cycles-pp.page_counter_uncharge
      0.17 ± 13%      +0.1        0.27 ±  9%  perf-profile.self.cycles-pp.copy_page
      0.01 ±200%      +0.1        0.11 ±  8%  perf-profile.self.cycles-pp.page_counter_charge
      0.09 ±  4%      +0.1        0.24 ±  9%  perf-profile.self.cycles-pp.sync_regs
      0.10 ±  9%      +0.3        0.39 ±  8%  perf-profile.self.cycles-pp.native_irq_return_iret
      0.07 ± 12%      +0.4        0.47 ±  9%  perf-profile.self.cycles-pp.__default_send_IPI_dest_field
      0.00            +0.4        0.44 ± 10%  perf-profile.self.cycles-pp.native_flush_tlb_local
      0.08 ± 13%      +0.5        0.62 ±  7%  perf-profile.self.cycles-pp.__flush_smp_call_function_queue
      0.06 ± 15%      +0.8        0.88 ±  7%  perf-profile.self.cycles-pp.flush_tlb_func
      0.26 ± 12%      +1.6        1.85 ±  9%  perf-profile.self.cycles-pp.llist_reverse_order
      0.36 ± 11%      +2.0        2.40 ±  8%  perf-profile.self.cycles-pp.llist_add_batch
      0.38 ± 13%      +3.1        3.49 ±  7%  perf-profile.self.cycles-pp.smp_call_function_many_cond


To reproduce:

        git clone https://github.com/intel/lkp-tests.git
        cd lkp-tests
        sudo bin/lkp install job.yaml           # job file is attached in this email
        bin/lkp split-job --compatible job.yaml # generate the yaml file for lkp run
        sudo bin/lkp run generated-yaml-file

        # if come across any failure that blocks the test,
        # please remove ~/.lkp and /lkp dir to run from a clean state.


Disclaimer:
Results have been estimated based on internal Intel analysis and are provided
for informational purposes only. Any difference in system hardware or software
design or configuration may affect actual performance.


-- 
0-DAY CI Kernel Test Service
https://github.com/intel/lkp-tests

View attachment "config-6.2.0-rc4-00556-g7e12beb8ca2a" of type "text/plain" (167513 bytes)

View attachment "job-script" of type "text/plain" (8085 bytes)

View attachment "job.yaml" of type "text/plain" (5867 bytes)

View attachment "reproduce" of type "text/plain" (806 bytes)

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ