lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [day] [month] [year] [list]
Message-ID: <202310182223.6ef26fcb-oliver.sang@intel.com>
Date:   Wed, 18 Oct 2023 22:50:46 +0800
From:   kernel test robot <oliver.sang@...el.com>
To:     Mel Gorman <mgorman@...hsingularity.net>
CC:     <oe-lkp@...ts.linux.dev>, <lkp@...el.com>,
        <linux-kernel@...r.kernel.org>, <x86@...nel.org>,
        Ingo Molnar <mingo@...nel.org>,
        Raghavendra K T <raghavendra.kt@....com>,
        <linux-trace-kernel@...r.kernel.org>, <ying.huang@...el.com>,
        <feng.tang@...el.com>, <fengwei.yin@...el.com>,
        <aubrey.li@...ux.intel.com>, <yu.c.chen@...el.com>,
        <oliver.sang@...el.com>
Subject: [tip:sched/core] [sched/numa]  f169c62ff7:
 autonuma-benchmark.numa01.seconds -32.0% improvement



Hello,

kernel test robot noticed a -32.0% improvement of autonuma-benchmark.numa01.seconds on:


commit: f169c62ff7cd1acf8bac8ae17bfeafa307d9e6fa ("sched/numa: Complete scanning of inactive VMAs when there is no alternative")
https://git.kernel.org/cgit/linux/kernel/git/tip/tip.git sched/core

testcase: autonuma-benchmark
test machine: 96 threads 2 sockets Intel(R) Xeon(R) Platinum 8260L CPU @ 2.40GHz (Cascade Lake) with 128G memory
parameters:

	iterations: 4x
	test: _HARD_BIND
	cpufreq_governor: performance






Details are as below:
-------------------------------------------------------------------------------------------------->


The kernel config and materials to reproduce are available at:
https://download.01.org/0day-ci/archive/20231018/202310182223.6ef26fcb-oliver.sang@intel.com

=========================================================================================
compiler/cpufreq_governor/iterations/kconfig/rootfs/tbox_group/test/testcase:
  gcc-12/performance/4x/x86_64-rhel-8.3/debian-11.1-x86_64-20220510.cgz/lkp-csl-2sp3/_HARD_BIND/autonuma-benchmark

commit: 
  b7a5b537c5 ("sched/numa: Complete scanning of partial VMAs regardless of PID activity")
  f169c62ff7 ("sched/numa: Complete scanning of inactive VMAs when there is no alternative")

b7a5b537c55c088d f169c62ff7cd1acf8bac8ae17bf 
---------------- --------------------------- 
         %stddev     %change         %stddev
             \          |                \  
      0.00 ± 58%   +3290.8%       0.06 ± 30%  vmstat.procs.b
 2.101e+10 ±  6%     -47.7%  1.099e+10 ±  5%  cpuidle..time
  21679574 ±  6%     -47.8%   11321089 ±  5%  cpuidle..usage
      2441           -25.2%       1826 ±  2%  uptime.boot
     25666 ±  5%     -40.3%      15332 ±  3%  uptime.idle
    208968 ±  8%     -11.6%     184676 ±  3%  meminfo.Active
    208280 ±  8%     -11.7%     183990 ±  3%  meminfo.Active(anon)
    221680 ±  8%     -11.7%     195741 ±  2%  meminfo.Shmem
   3154708 ±  7%     -12.1%    2773727 ±  4%  numa-numastat.node0.local_node
   3621021 ±  4%     -12.9%    3154008 ±  3%  numa-numastat.node0.numa_hit
   3424075 ±  5%     -19.2%    2767342 ±  3%  numa-numastat.node1.local_node
   3659250 ±  4%     -15.8%    3079449 ±  3%  numa-numastat.node1.numa_hit
   3620012 ±  4%     -12.9%    3153670 ±  3%  numa-vmstat.node0.numa_hit
   3153693 ±  7%     -12.1%    2773388 ±  4%  numa-vmstat.node0.numa_local
   3657708 ±  4%     -15.8%    3078660 ±  3%  numa-vmstat.node1.numa_hit
   3422533 ±  5%     -19.2%    2766552 ±  3%  numa-vmstat.node1.numa_local
      0.03 ± 31%    +840.9%       0.29 ±179%  perf-sched.sch_delay.avg.ms.exit_to_user_mode_loop.exit_to_user_mode_prepare.syscall_exit_to_user_mode.do_syscall_64
      2.34 ± 49%     -83.7%       0.38 ±194%  perf-sched.sch_delay.max.ms.__cond_resched.down_write_killable.setup_arg_pages.load_elf_binary.search_binary_handler
    948.50 ± 16%     -30.3%     661.33 ± 20%  perf-sched.wait_and_delay.count.exit_to_user_mode_loop.exit_to_user_mode_prepare.syscall_exit_to_user_mode.do_syscall_64
      1142 ±  9%     -24.3%     865.00 ±  9%  perf-sched.wait_and_delay.count.pipe_read.vfs_read.ksys_read.do_syscall_64
      9.20 ±  6%      -2.8        6.43 ±  5%  mpstat.cpu.all.idle%
      0.00 ± 14%      +0.0        0.03 ± 28%  mpstat.cpu.all.iowait%
      2.25            -0.5        1.73        mpstat.cpu.all.irq%
      0.08 ±  3%      -0.0        0.06 ±  2%  mpstat.cpu.all.soft%
      2.18 ±  4%      +0.5        2.69 ±  2%  mpstat.cpu.all.sys%
    474.78           -32.0%     322.76 ±  2%  autonuma-benchmark.numa01.seconds
      2386           -25.7%       1774 ±  2%  autonuma-benchmark.time.elapsed_time
      2386           -25.7%       1774 ±  2%  autonuma-benchmark.time.elapsed_time.max
   1298583 ±  2%     -24.3%     983395 ±  3%  autonuma-benchmark.time.involuntary_context_switches
   3120169            -6.0%    2932947        autonuma-benchmark.time.minor_page_faults
      8443            +3.9%       8771        autonuma-benchmark.time.percent_of_cpu_this_job_got
    197252           -23.2%     151486 ±  2%  autonuma-benchmark.time.user_time
     26055           +45.4%      37896 ±  8%  autonuma-benchmark.time.voluntary_context_switches
     42526 ±  9%     -15.6%      35892 ± 10%  turbostat.C1
      0.03            +0.0        0.05 ± 20%  turbostat.C1E%
  21430219 ±  6%     -48.2%   11099622 ±  5%  turbostat.C6
      9.14 ±  6%      -2.7        6.40 ±  5%  turbostat.C6%
      8.80 ±  6%     -31.1%       6.06 ±  5%  turbostat.CPU%c1
  2.31e+08           -24.9%  1.735e+08 ±  2%  turbostat.IRQ
     37846 ±  7%     -29.3%      26768 ±  9%  turbostat.POLL
    283.70            +3.1%     292.44        turbostat.PkgWatt
     63.32           +16.3%      73.65        turbostat.RAMWatt
     52079 ±  8%     -11.6%      46034 ±  3%  proc-vmstat.nr_active_anon
   1560738            -1.8%    1532470        proc-vmstat.nr_anon_pages
      3000            -2.2%       2934        proc-vmstat.nr_anon_transparent_hugepages
    754040            -0.9%     747564        proc-vmstat.nr_file_pages
   1564055            -1.8%    1535343        proc-vmstat.nr_inactive_anon
      3964            -1.4%       3911        proc-vmstat.nr_page_table_pages
     55430 ±  8%     -11.7%      48954 ±  2%  proc-vmstat.nr_shmem
     52079 ±  8%     -11.6%      46034 ±  3%  proc-vmstat.nr_zone_active_anon
   1564055            -1.8%    1535343        proc-vmstat.nr_zone_inactive_anon
     31021 ±  4%    +356.7%     141673 ±  3%  proc-vmstat.numa_hint_faults
     26070 ± 14%    +232.5%      86687 ±  5%  proc-vmstat.numa_hint_faults_local
   7281579           -14.4%    6235148        proc-vmstat.numa_hit
      2255 ±  9%   +3726.0%      86295 ±  3%  proc-vmstat.numa_huge_pte_updates
   6580091           -15.8%    5542759        proc-vmstat.numa_local
    490705         +1551.4%    8103581 ±  3%  proc-vmstat.numa_pages_migrated
   1256422 ±  9%   +3426.6%   44309531 ±  3%  proc-vmstat.numa_pte_updates
   8560551           -16.7%    7127497        proc-vmstat.pgfault
    490705         +1551.4%    8103581 ±  3%  proc-vmstat.pgmigrate_success
    370032 ±  2%     -19.7%     296973 ±  2%  proc-vmstat.pgreuse
    951.67         +1560.5%      15802 ±  3%  proc-vmstat.thp_migration_success
  17842432           -25.5%   13288960 ±  2%  proc-vmstat.unevictable_pgs_scanned
 1.408e+08 ±  2%     -26.2%   1.04e+08 ±  2%  sched_debug.cfs_rq:/.avg_vruntime.avg
 1.476e+08 ±  2%     -27.3%  1.073e+08 ±  2%  sched_debug.cfs_rq:/.avg_vruntime.max
 1.252e+08           -24.9%   94038602 ±  3%  sched_debug.cfs_rq:/.avg_vruntime.min
   4459866 ±  7%     -40.1%    2672703 ±  3%  sched_debug.cfs_rq:/.avg_vruntime.stddev
      0.22 ±  4%     -10.8%       0.20 ±  3%  sched_debug.cfs_rq:/.h_nr_running.stddev
 1.408e+08 ±  2%     -26.2%   1.04e+08 ±  2%  sched_debug.cfs_rq:/.min_vruntime.avg
 1.476e+08 ±  2%     -27.3%  1.073e+08 ±  2%  sched_debug.cfs_rq:/.min_vruntime.max
 1.252e+08           -24.9%   94038602 ±  3%  sched_debug.cfs_rq:/.min_vruntime.min
   4459866 ±  7%     -40.1%    2672703 ±  3%  sched_debug.cfs_rq:/.min_vruntime.stddev
    131.57 ± 11%     -17.7%     108.34 ±  4%  sched_debug.cfs_rq:/.util_avg.stddev
      6.43 ± 26%   +4528.1%     297.67 ±  3%  sched_debug.cfs_rq:/.util_est_enqueued.avg
    289.38 ±  9%    +195.6%     855.43 ±  4%  sched_debug.cfs_rq:/.util_est_enqueued.max
     37.21 ± 11%    +508.1%     226.25 ±  5%  sched_debug.cfs_rq:/.util_est_enqueued.stddev
    827702 ±  3%     -12.7%     722480 ±  6%  sched_debug.cpu.avg_idle.min
    306806 ±  5%     -18.8%     249093 ±  5%  sched_debug.cpu.avg_idle.stddev
   1219552 ±  2%     -25.7%     906296 ±  2%  sched_debug.cpu.clock.avg
   1219905 ±  2%     -25.7%     906607 ±  2%  sched_debug.cpu.clock.max
   1219166 ±  2%     -25.7%     905967 ±  2%  sched_debug.cpu.clock.min
    212.97 ±  6%     -13.3%     184.61 ± 11%  sched_debug.cpu.clock.stddev
   1191500 ±  2%     -25.3%     890437 ±  2%  sched_debug.cpu.clock_task.avg
   1199519 ±  2%     -25.6%     892331 ±  2%  sched_debug.cpu.clock_task.max
   1176815 ±  2%     -25.1%     881433 ±  2%  sched_debug.cpu.clock_task.min
      7162 ± 46%     -77.7%       1593 ± 19%  sched_debug.cpu.clock_task.stddev
     28968 ±  4%     -16.3%      24235 ±  6%  sched_debug.cpu.curr->pid.avg
     36717           -20.7%      29133 ±  2%  sched_debug.cpu.curr->pid.max
      3810 ± 11%     -34.7%       2487 ± 11%  sched_debug.cpu.curr->pid.stddev
   1074237           -17.0%     892137 ±  6%  sched_debug.cpu.max_idle_balance_cost.max
    137289 ±  3%     -43.4%      77712 ± 11%  sched_debug.cpu.max_idle_balance_cost.stddev
      0.00 ±  6%     -13.2%       0.00 ± 10%  sched_debug.cpu.next_balance.stddev
      0.23 ±  3%     -10.5%       0.20 ±  4%  sched_debug.cpu.nr_running.stddev
     22190 ±  2%     -23.8%      16898 ±  3%  sched_debug.cpu.nr_switches.avg
      9201 ±  4%     -15.7%       7756 ±  6%  sched_debug.cpu.nr_switches.min
     10098 ±  3%     -15.0%       8588 ±  7%  sched_debug.cpu.nr_switches.stddev
      7.02 ± 10%     +43.9%      10.10 ± 12%  sched_debug.cpu.nr_uninterruptible.stddev
   1219153 ±  2%     -25.7%     905956 ±  2%  sched_debug.cpu_clk
   1218584 ±  2%     -25.7%     905387 ±  2%  sched_debug.ktime
   1219711 ±  2%     -25.7%     906506 ±  2%  sched_debug.sched_clk
     26.19           +28.8%      33.74        perf-stat.i.MPKI
 1.144e+08            +4.0%   1.19e+08        perf-stat.i.branch-instructions
      1.46            -0.1        1.37        perf-stat.i.branch-miss-rate%
     63.10            +3.1       66.21        perf-stat.i.cache-miss-rate%
  14186186           +29.8%   18412463        perf-stat.i.cache-misses
  22674336           +24.7%   28269331        perf-stat.i.cache-references
    502.02            +2.0%     512.29        perf-stat.i.cpi
 2.643e+11            +3.9%  2.745e+11        perf-stat.i.cpu-cycles
    126.73            +7.2%     135.83        perf-stat.i.cpu-migrations
     21465           -23.4%      16442        perf-stat.i.cycles-between-cache-misses
 1.572e+08            +3.2%  1.623e+08        perf-stat.i.dTLB-loads
      0.27            +0.0        0.29        perf-stat.i.dTLB-store-miss-rate%
    223732            +9.2%     244255        perf-stat.i.dTLB-store-misses
  88029014            +3.0%   90626204        perf-stat.i.dTLB-stores
     80.17            +1.7       81.85        perf-stat.i.iTLB-load-miss-rate%
    314646           -16.1%     263974 ±  5%  perf-stat.i.iTLB-loads
 5.846e+08            +3.7%  6.063e+08        perf-stat.i.instructions
      1246 ±  2%     +13.7%       1416 ±  4%  perf-stat.i.instructions-per-iTLB-miss
      2.75            +3.9%       2.86        perf-stat.i.metric.GHz
      2.57            +9.1%       2.81        perf-stat.i.metric.M/sec
      3527           +10.9%       3913        perf-stat.i.minor-faults
    233066           +31.7%     306835        perf-stat.i.node-load-misses
    238123           +13.9%     271245        perf-stat.i.node-loads
   3112998           +44.6%    4502856 ±  2%  perf-stat.i.node-store-misses
   9385414           +24.5%   11686725 ±  2%  perf-stat.i.node-stores
      3527           +10.9%       3913        perf-stat.i.page-faults
     24.08           +25.2%      30.16        perf-stat.overall.MPKI
      1.54            -0.1        1.49        perf-stat.overall.branch-miss-rate%
     63.01            +2.4       65.42        perf-stat.overall.cache-miss-rate%
     19050           -20.6%      15118        perf-stat.overall.cycles-between-cache-misses
      0.25            +0.0        0.27        perf-stat.overall.dTLB-store-miss-rate%
     62.66            +2.2       64.86        perf-stat.overall.iTLB-load-miss-rate%
      1161 ±  4%     +10.8%       1286 ±  4%  perf-stat.overall.instructions-per-iTLB-miss
     49.19            +3.7       52.85        perf-stat.overall.node-load-miss-rate%
     26.10            +2.6       28.67 ±  2%  perf-stat.overall.node-store-miss-rate%
 1.134e+08            +4.5%  1.185e+08        perf-stat.ps.branch-instructions
  13962667           +30.4%   18209922        perf-stat.ps.cache-misses
  22159412           +25.6%   27835158        perf-stat.ps.cache-references
  2.66e+11            +3.5%  2.752e+11        perf-stat.ps.cpu-cycles
    126.06            +7.4%     135.36        perf-stat.ps.cpu-migrations
  1.56e+08            +3.6%  1.617e+08        perf-stat.ps.dTLB-loads
    222631            +9.1%     242846        perf-stat.ps.dTLB-store-misses
  87353212            +3.2%   90174337        perf-stat.ps.dTLB-stores
    297718 ±  2%     -14.4%     254772 ±  5%  perf-stat.ps.iTLB-loads
 5.797e+08            +4.1%  6.038e+08        perf-stat.ps.instructions
      3454           +11.6%       3856        perf-stat.ps.minor-faults
    231688           +32.7%     307530        perf-stat.ps.node-load-misses
    239364           +14.6%     274365 ±  2%  perf-stat.ps.node-loads
   3208117           +42.8%    4581795 ±  2%  perf-stat.ps.node-store-misses
   9084897           +25.5%   11404755 ±  2%  perf-stat.ps.node-stores
      3455           +11.6%       3856        perf-stat.ps.page-faults
 1.384e+12           -22.6%  1.072e+12        perf-stat.total.instructions
     45.09 ± 10%     -30.4       14.72 ± 55%  perf-profile.calltrace.cycles-pp.asm_sysvec_apic_timer_interrupt
      4.07 ± 30%      -1.9        2.14 ± 40%  perf-profile.calltrace.cycles-pp.ret_from_fork_asm
      4.07 ± 30%      -1.9        2.14 ± 40%  perf-profile.calltrace.cycles-pp.ret_from_fork.ret_from_fork_asm
      4.07 ± 30%      -1.9        2.14 ± 40%  perf-profile.calltrace.cycles-pp.kthread.ret_from_fork.ret_from_fork_asm
      1.74 ± 33%      -1.2        0.57 ± 72%  perf-profile.calltrace.cycles-pp.hpage_collapse_scan_pmd.khugepaged_scan_mm_slot.khugepaged.kthread.ret_from_fork
      1.75 ± 33%      -1.2        0.58 ± 73%  perf-profile.calltrace.cycles-pp.khugepaged.kthread.ret_from_fork.ret_from_fork_asm
      1.74 ± 33%      -1.2        0.58 ± 73%  perf-profile.calltrace.cycles-pp.khugepaged_scan_mm_slot.khugepaged.kthread.ret_from_fork.ret_from_fork_asm
      1.69 ± 33%      -1.1        0.55 ± 72%  perf-profile.calltrace.cycles-pp.collapse_huge_page.hpage_collapse_scan_pmd.khugepaged_scan_mm_slot.khugepaged.kthread
      1.58 ± 33%      -1.1        0.51 ± 72%  perf-profile.calltrace.cycles-pp.__collapse_huge_page_copy.collapse_huge_page.hpage_collapse_scan_pmd.khugepaged_scan_mm_slot.khugepaged
      1.38 ± 32%      -0.9        0.45 ± 72%  perf-profile.calltrace.cycles-pp.copy_mc_fragile.__collapse_huge_page_copy.collapse_huge_page.hpage_collapse_scan_pmd.khugepaged_scan_mm_slot
      1.90 ± 17%      -0.9        0.99 ± 54%  perf-profile.calltrace.cycles-pp.evsel__read_counter.read_counters.process_interval.dispatch_events.cmd_stat
      1.20 ± 18%      -0.6        0.55 ± 75%  perf-profile.calltrace.cycles-pp.readn.evsel__read_counter.read_counters.process_interval.dispatch_events
      1.20 ± 18%      -0.6        0.55 ± 75%  perf-profile.calltrace.cycles-pp.__libc_read.readn.evsel__read_counter.read_counters.process_interval
      1.17 ± 19%      -0.6        0.54 ± 75%  perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.__libc_read.readn.evsel__read_counter.read_counters
      1.17 ± 19%      -0.6        0.54 ± 75%  perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.__libc_read.readn.evsel__read_counter
     46.12 ± 10%     -30.1       15.98 ± 51%  perf-profile.children.cycles-pp.asm_sysvec_apic_timer_interrupt
      4.07 ± 30%      -1.9        2.14 ± 40%  perf-profile.children.cycles-pp.kthread
      1.74 ± 33%      -1.1        0.67 ± 42%  perf-profile.children.cycles-pp.hpage_collapse_scan_pmd
      1.74 ± 33%      -1.1        0.67 ± 43%  perf-profile.children.cycles-pp.khugepaged_scan_mm_slot
      1.75 ± 33%      -1.1        0.68 ± 43%  perf-profile.children.cycles-pp.khugepaged
      1.69 ± 33%      -1.0        0.64 ± 41%  perf-profile.children.cycles-pp.collapse_huge_page
      1.58 ± 33%      -1.0        0.60 ± 41%  perf-profile.children.cycles-pp.__collapse_huge_page_copy
      1.90 ± 17%      -0.9        1.05 ± 42%  perf-profile.children.cycles-pp.evsel__read_counter
      1.92 ± 18%      -0.8        1.10 ± 44%  perf-profile.children.cycles-pp.readn
      1.92 ± 18%      -0.8        1.10 ± 44%  perf-profile.children.cycles-pp.__libc_read
      0.87 ± 15%      -0.4        0.51 ± 38%  perf-profile.children.cycles-pp.do_anonymous_page
      0.65 ± 16%      -0.2        0.42 ± 37%  perf-profile.children.cycles-pp.vma_alloc_folio
      0.46 ± 26%      -0.2        0.23 ± 45%  perf-profile.children.cycles-pp.asm_sysvec_reschedule_ipi
      0.44 ± 22%      -0.2        0.24 ± 46%  perf-profile.children.cycles-pp.mprotect_fixup
      0.40 ± 18%      -0.2        0.21 ± 41%  perf-profile.children.cycles-pp.evlist__id2evsel
      0.24 ± 46%      -0.2        0.07 ± 64%  perf-profile.children.cycles-pp.start_thread
      0.23 ± 45%      -0.2        0.07 ± 64%  perf-profile.children.cycles-pp.perf_evlist__poll_thread
      0.40 ± 17%      -0.2        0.25 ± 34%  perf-profile.children.cycles-pp.mas_store_prealloc
      0.28 ± 19%      -0.1        0.14 ± 40%  perf-profile.children.cycles-pp.__fget_light
      0.26 ± 32%      -0.1        0.13 ± 33%  perf-profile.children.cycles-pp.pte_alloc_one
      0.18 ± 43%      -0.1        0.06 ± 80%  perf-profile.children.cycles-pp.__collapse_huge_page_copy_succeeded
      0.16 ± 30%      -0.1        0.04 ±109%  perf-profile.children.cycles-pp.proc_pid_get_link
      0.30 ± 21%      -0.1        0.19 ± 40%  perf-profile.children.cycles-pp.__mem_cgroup_charge
      0.19 ± 16%      -0.1        0.09 ± 62%  perf-profile.children.cycles-pp.__perf_read_group_add
      0.22 ± 22%      -0.1        0.12 ± 27%  perf-profile.children.cycles-pp.shift_arg_pages
      0.15 ± 33%      -0.1        0.06 ± 71%  perf-profile.children.cycles-pp.xas_find
      0.12 ± 26%      -0.1        0.03 ±101%  perf-profile.children.cycles-pp.folio_add_new_anon_rmap
      0.15 ± 28%      -0.1        0.06 ± 63%  perf-profile.children.cycles-pp.switch_mm_irqs_off
      0.14 ± 27%      -0.1        0.06 ± 71%  perf-profile.children.cycles-pp.try_charge_memcg
      0.19 ± 29%      -0.1        0.11 ± 51%  perf-profile.children.cycles-pp.down_write
      0.17 ± 32%      -0.1        0.09 ± 39%  perf-profile.children.cycles-pp.__pte_alloc
      0.10 ± 32%      -0.1        0.03 ±101%  perf-profile.children.cycles-pp.blk_mq_queue_tag_busy_iter
      0.15 ± 16%      -0.1        0.08 ± 54%  perf-profile.children.cycles-pp.vsnprintf
      0.10 ± 32%      -0.1        0.04 ± 73%  perf-profile.children.cycles-pp.blk_mq_in_flight
      0.12 ± 20%      -0.1        0.06 ± 48%  perf-profile.children.cycles-pp.slab_show
      0.16 ± 24%      -0.1        0.10 ± 38%  perf-profile.children.cycles-pp.__get_free_pages
      0.09 ± 35%      -0.1        0.04 ± 79%  perf-profile.children.cycles-pp.lookup_open
      0.00            +0.3        0.27 ± 57%  perf-profile.children.cycles-pp.__sysvec_call_function
      0.03 ±100%      +0.3        0.33 ± 50%  perf-profile.children.cycles-pp.__flush_smp_call_function_queue
      0.00            +0.4        0.38 ± 53%  perf-profile.children.cycles-pp.sysvec_call_function
      0.04 ± 80%      +1.6        1.67 ± 58%  perf-profile.children.cycles-pp.asm_sysvec_call_function
     43.32 ± 10%     -29.9       13.42 ± 61%  perf-profile.self.cycles-pp.asm_sysvec_apic_timer_interrupt
      0.40 ± 17%      -0.2        0.20 ± 41%  perf-profile.self.cycles-pp.evlist__id2evsel
      0.28 ± 19%      -0.1        0.13 ± 40%  perf-profile.self.cycles-pp.__fget_light
      0.24 ± 30%      -0.1        0.12 ± 29%  perf-profile.self.cycles-pp.zap_pte_range
      0.14 ± 28%      -0.1        0.06 ± 76%  perf-profile.self.cycles-pp.evlist_cpu_iterator__next
      0.13 ± 30%      -0.1        0.05 ± 75%  perf-profile.self.cycles-pp.try_charge_memcg
      0.17 ± 27%      -0.1        0.10 ± 51%  perf-profile.self.cycles-pp.down_write
      0.10 ± 31%      -0.1        0.03 ±102%  perf-profile.self.cycles-pp.blk_mq_queue_tag_busy_iter
      0.13 ± 30%      -0.1        0.06 ± 65%  perf-profile.self.cycles-pp.switch_mm_irqs_off
      0.11 ± 30%      -0.1        0.05 ± 75%  perf-profile.self.cycles-pp.mutex_lock




Disclaimer:
Results have been estimated based on internal Intel analysis and are provided
for informational purposes only. Any difference in system hardware or software
design or configuration may affect actual performance.


-- 
0-DAY CI Kernel Test Service
https://github.com/intel/lkp-tests/wiki

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ