lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [<thread-prev] [day] [month] [year] [list]
Message-ID: <cd078fa1-f997-e9f7-892a-78fb38d30fca@linux.intel.com>
Date:   Tue, 16 Mar 2021 09:29:59 +0800
From:   Xing Zhengjun <zhengjun.xing@...ux.intel.com>
To:     Alex Kogan <alex.kogan@...cle.com>
Cc:     kernel test robot <oliver.sang@...el.com>,
        0day robot <lkp@...el.com>,
        Steven Sistare <steven.sistare@...cle.com>,
        Waiman Long <longman@...hat.com>,
        LKML <linux-kernel@...r.kernel.org>,
        "lkp@...ts.01.org" <lkp@...ts.01.org>,
        "linux@...linux.org.uk" <linux@...linux.org.uk>,
        "peterz@...radead.org" <peterz@...radead.org>,
        "mingo@...hat.com" <mingo@...hat.com>,
        "will.deacon@....com" <will.deacon@....com>,
        "arnd@...db.de" <arnd@...db.de>,
        "linux-arch@...r.kernel.org" <linux-arch@...r.kernel.org>,
        "linux-arm-kernel@...ts.infradead.org" 
        <linux-arm-kernel@...ts.infradead.org>,
        "tglx@...utronix.de" <tglx@...utronix.de>,
        "bp@...en8.de" <bp@...en8.de>, "hpa@...or.com" <hpa@...or.com>,
        "x86@...nel.org" <x86@...nel.org>,
        "guohanjun@...wei.com" <guohanjun@...wei.com>,
        "jglauber@...vell.com" <jglauber@...vell.com>,
        Daniel Jordan <daniel.m.jordan@...cle.com>,
        Dave Dice <dave.dice@...cle.com>
Subject: Re: [External] : Re: [LKP] Re: [locking/qspinlock] 0e8d8f4f12:
 stress-ng.zero.ops_per_sec -9.7% regression



On 3/15/2021 9:30 PM, Alex Kogan wrote:
> Thanks for the update, Xing.
> 
> Is there any way for me to reproduce the regression?
> Last time I checked the job description file had hardcoded strings to identify various components
> (such as HDD partitions, rootfs partitions, etc).
> 

You can try "./stress-ng --timeout 30 --times --verify --metrics-brief 
--zero 48" (our test machine has 48 CPUs, so it is 48, you can change it 
to your test machine CPU numbers).  In the test result, "bogo ops" is 
for "stress-ng.zero.ops",  "bogo ops/s (real time)" is 
"stress-ng.zero.ops_per_sec". The test result may differ from ours, but 
it is easy for you to reproduce and debug the issue.


stress-ng: info:  [2121] dispatching hogs: 48 zero
stress-ng: info:  [2121] successful run completed in 30.03s
stress-ng: info:  [2121] stressor       bogo ops real time  usr time 
sys time   bogo ops/s     bogo ops/s
stress-ng: info:  [2121]                           (secs)    (secs) 
(secs)   (real time) (usr+sys time)
stress-ng: info:  [2121] zero           10209273     30.00    140.46 
1215.98    340307.48        7526.52
stress-ng: info:  [2121] for a 30.03s run time:
stress-ng: info:  [2121]    1441.24s available CPU time
stress-ng: info:  [2121]     140.70s user time   (  9.76%)
stress-ng: info:  [2121]    1216.22s system time ( 84.39%)
stress-ng: info:  [2121]    1356.92s total time  ( 94.15%)



> Regards,
> — Alex
> 
>> On Mar 14, 2021, at 10:28 PM, Xing Zhengjun <zhengjun.xing@...ux.intel.com> wrote:
>>
>>
>>
>> On 3/11/2021 1:17 AM, Alex Kogan wrote:
>>> [ Looks like my previous reply failed to reach the mailing list. Reposting again, after (hopefully)
>>> fixing the problem with the included URL. Apologies if you have received this message in
>>> the past. ]
>>>
>>> A quick update: there is a problem in running this test to reproduce the regression,
>>> and Oliver Sang is aware of that.
>>>
>>> Along with that, I wonder if this is a real issue or some noise in the measurements.
>>> This regression is reported with the “--device" switch, which according to
>>> https://urldefense.com/v3/__https://wiki.ubuntu.com/Kernel/Reference/stress-ng__;!!GqivPVa7Brio!KJgSkso-bLajnACiRoGKa3GJdJO-5X8sZxKAXhK9nojWsD36jGpUmgVSNrZtF7-x$ , enables "raw device driver stressors”.
>>> I wonder if it is relevant to the spin lock performance.
>>>
>>> Any thoughts?
>>>
>>> Also, using this opportunity, any further feedback on the patch would be greatly appreciated.
>>>
>>> Thanks,
>>> — Alex
>>>
>>
>> In the previous test, it test by class "device", it includes a lot of sub-cases except for "zero", they may affect each other, then I split it, only test "zero" test, the test result is as the following, the regression still exists.
>>
>> =========================================================================================
>> tbox_group/testcase/rootfs/kconfig/compiler/nr_threads/disk/testtime/class/test/cpufreq_governor/ucode/debug-setup:
>> lkp-ivb-2ep1/stress-ng/debian-10.4-x86_64-20200603.cgz/x86_64-rhel-8.3/gcc-9/100%/1HDD/30s/device/zero/performance/0x42e/test1
>>
>> commit:
>>    cb45bab007ff0cfc42695d055c4b1bf5a5423d81
>>    0e8d8f4f1214cfbac219d6917b5f6460f818bb7c
>>
>> cb45bab007ff0cfc 0e8d8f4f1214cfbac219d6917b5
>> ---------------- ---------------------------
>>           %stddev     %change         %stddev
>>               \          |                \
>>    10230669 ±  2%     -10.7%    9139816 ±  3%  stress-ng.zero.ops
>>      341020 ±  2%     -10.7%     304650 ±  3% stress-ng.zero.ops_per_sec
>>
>>
>>>
>>> From: kernel test robot <oliver.sang@...el.com>
>>> Sent: Monday, December 28, 2020 3:16 AM
>>> To: Alex Kogan <alex.kogan@...cle.com>
>>> Cc: 0day robot <lkp@...el.com>; Steven Sistare <steven.sistare@...cle.com>; Waiman Long <longman@...hat.com>; LKML <linux-kernel@...r.kernel.org>; lkp@...ts.01.org <lkp@...ts.01.org>; ying.huang@...el.com <ying.huang@...el.com>; feng.tang@...el.com <feng.tang@...el.com>; zhengjun.xing@...el.com <zhengjun.xing@...el.com>; linux@...linux.org.uk <linux@...linux.org.uk>; peterz@...radead.org <peterz@...radead.org>; mingo@...hat.com <mingo@...hat.com>; will.deacon@....com <will.deacon@....com>; arnd@...db.de <arnd@...db.de>; linux-arch@...r.kernel.org <linux-arch@...r.kernel.org>; linux-arm-kernel@...ts.infradead.org <linux-arm-kernel@...ts.infradead.org>; tglx@...utronix.de <tglx@...utronix.de>; bp@...en8.de <bp@...en8.de>; hpa@...or.com <hpa@...or.com>; x86@...nel.org <x86@...nel.org>; guohanjun@...wei.com <guohanjun@...wei.com>; jglauber@...vell.com <jglauber@...vell.com>; Daniel Jordan <daniel.m.jordan@...cle.com>; Alex Kogan <alex.kogan@...cle.com>; Dave Dice <dave.dice@...cle.com>
>>> Subject: [locking/qspinlock] 0e8d8f4f12: stress-ng.zero.ops_per_sec -9.7% regression
>>>   
>>> Greeting,
>>>
>>> FYI, we noticed a -9.7% regression of stress-ng.zero.ops_per_sec due to commit:
>>>
>>>
>>> commit: 0e8d8f4f1214cfbac219d6917b5f6460f818bb7c ("[PATCH v13 3/6] locking/qspinlock: Introduce CNA into the slow path of qspinlock")
>>> url: https://urldefense.com/v3/__https://github.com/0day-ci/linux/commits/Alex-Kogan/Add-NUMA-awareness-to-qspinlock/20201223-135025__;!!GqivPVa7Brio!LF1Vhc6eU7n1kxGSlPpI_wS0LmREqgH6k1226e4w8C7ug0i7368nCfIT44_wbu1g$
>>> base: https://urldefense.com/v3/__https://git.kernel.org/cgit/linux/kernel/git/tip/tip.git__;!!GqivPVa7Brio!LF1Vhc6eU7n1kxGSlPpI_wS0LmREqgH6k1226e4w8C7ug0i7368nCfIT42aMeOW1$  cb262935a166bdef0ccfe6e2adffa00c0f2d038a
>>>
>>> in testcase: stress-ng
>>> on test machine: 48 threads Intel(R) Xeon(R) CPU E5-2697 v2 @ 2.70GHz with 112G memory
>>> with following parameters:
>>>
>>>          nr_threads: 100%
>>>          disk: 1HDD
>>>          testtime: 30s
>>>          class: device
>>>          cpufreq_governor: performance
>>>          ucode: 0x42e
>>>
>>>
>>> In addition to that, the commit also has significant impact on the following tests:
>>>
>>> +------------------+---------------------------------------------------------------------------+
>>> | testcase: change | fsmark: fsmark.files_per_sec 213.9% improvement                           |
>>> | test machine     | 192 threads Intel(R) Xeon(R) Platinum 9242 CPU @ 2.30GHz with 192G memory |
>>> | test parameters  | cpufreq_governor=performance                                              |
>>> |                  | disk=1BRD_48G                                                             |
>>> |                  | filesize=4M                                                               |
>>> |                  | fs=btrfs                                                                  |
>>> |                  | iterations=1x                                                             |
>>> |                  | nr_threads=64t                                                            |
>>> |                  | sync_method=NoSync                                                        |
>>> |                  | test_size=24G                                                             |
>>> |                  | ucode=0x5003003                                                           |
>>> +------------------+---------------------------------------------------------------------------+
>>> | testcase: change | reaim: reaim.jobs_per_min 96.1% improvement                               |
>>> | test machine     | 192 threads Intel(R) Xeon(R) Platinum 9242 CPU @ 2.30GHz with 192G memory |
>>> | test parameters  | cpufreq_governor=performance                                              |
>>> |                  | nr_task=100%                                                              |
>>> |                  | runtime=300s                                                              |
>>> |                  | test=new_fserver                                                          |
>>> |                  | ucode=0x5003003                                                           |
>>> +------------------+---------------------------------------------------------------------------+
>>>
>>>
>>> If you fix the issue, kindly add following tag
>>> Reported-by: kernel test robot <oliver.sang@...el.com>
>>>
>>>
>>> Details are as below:
>>> -------------------------------------------------------------------------------------------------->
>>>
>>>
>>> To reproduce:
>>>
>>>          git clone https://urldefense.com/v3/__https://github.com/intel/lkp-tests.git__;!!GqivPVa7Brio!LF1Vhc6eU7n1kxGSlPpI_wS0LmREqgH6k1226e4w8C7ug0i7368nCfIT4ySp2S6I$
>>>          cd lkp-tests
>>>          bin/lkp install job.yaml  # job file is attached in this email
>>>          bin/lkp run     job.yaml
>>>
>>> =========================================================================================
>>> class/compiler/cpufreq_governor/disk/kconfig/nr_threads/rootfs/tbox_group/testcase/testtime/ucode:
>>>    device/gcc-9/performance/1HDD/x86_64-rhel-8.3/100%/debian-10.4-x86_64-20200603.cgz/lkp-ivb-2ep1/stress-ng/30s/0x42e
>>>
>>> commit:
>>>    cb45bab007 ("locking/qspinlock: Refactor the qspinlock slow path")
>>>    0e8d8f4f12 ("locking/qspinlock: Introduce CNA into the slow path of qspinlock")
>>>
>>> cb45bab007ff0cfc 0e8d8f4f1214cfbac219d6917b5
>>> ---------------- ---------------------------
>>>         fail:runs  %reproduction    fail:runs
>>>             |             |             |
>>>             :4           25%           1:4     last_state.is_incomplete_run
>>>             :4           25%           1:4     dmesg.Kernel_panic-not_syncing:softlockup:hung_tasks
>>>             :4           25%           1:4     dmesg.RIP:smp_call_function_many_cond
>>>             :4           25%           1:4     dmesg.RIP:smp_call_function_single
>>>           %stddev     %change         %stddev
>>>               \          |                \
>>>      481980 ±  9%     -12.0%     424197 ±  5%  stress-ng.time.involuntary_context_switches
>>>   1.198e+08            -4.0%   1.15e+08        stress-ng.time.minor_page_faults
>>>      689.06            -4.9%     655.29        stress-ng.time.user_time
>>>    10626516            -9.7%    9598124        stress-ng.zero.ops
>>>      354216            -9.7%     319937        stress-ng.zero.ops_per_sec
>>>       12871           +19.8%      15424 ± 12%  meminfo.max_used_kB
>>>       29.00            +4.6%      30.33        vmstat.cpu.id
>>>       29.66            +3.7%      30.75 ±  2%  iostat.cpu.idle
>>>        8.21            -7.6%       7.59 ±  2%  iostat.cpu.user
>>>       67636 ± 12%     -21.3%      53263 ± 14%  sched_debug.cpu.nr_switches.max
>>>       70.29 ± 16%     -22.1%      54.73 ± 14%  sched_debug.cpu.nr_uninterruptible.stddev
>>>        7841 ± 58%    +418.8%      40676 ±  2%  numa-meminfo.node0.Active
>>>        6226 ± 71%    +542.5%      40001        numa-meminfo.node0.Active(anon)
>>>       15043 ± 32%    +271.6%      55899        numa-meminfo.node0.Shmem
>>>       77673 ±  3%     -50.0%      38874 ±  9%  numa-meminfo.node1.Active
>>>       77439 ±  3%     -51.0%      37925 ±  8%  numa-meminfo.node1.Active(anon)
>>>      233.75 ±157%    +306.0%     949.00 ± 66%  numa-meminfo.node1.Active(file)
>>>      575517 ±  4%      -8.8%     525050 ±  3%  numa-meminfo.node1.FilePages
>>>       87936 ±  4%     -53.1%      41275 ±  9%  numa-meminfo.node1.Shmem
>>>        2097 ±  5%     +26.8%       2660 ± 21%  slabinfo.dmaengine-unmap-16.active_objs
>>>        2097 ±  5%     +26.8%       2660 ± 21%  slabinfo.dmaengine-unmap-16.num_objs
>>>       40336 ±  6%    +102.0%      81465 ± 50%  slabinfo.filp.active_objs
>>>      657.00 ±  6%     +98.0%       1301 ± 49%  slabinfo.filp.active_slabs
>>>       42081 ±  6%     +97.9%      83296 ± 49%  slabinfo.filp.num_objs
>>>      657.00 ±  6%     +98.0%       1301 ± 49%  slabinfo.filp.num_slabs
>>>       13106 ±  3%     +12.6%      14752 ±  3%  slabinfo.shmem_inode_cache.active_objs
>>>       13224 ±  3%     +12.5%      14873 ±  3%  slabinfo.shmem_inode_cache.num_objs
>>>        1557 ± 71%    +545.2%      10045 ±  2%  numa-vmstat.node0.nr_active_anon
>>>      117.50 ± 57%     -97.4%       3.00 ± 47%  numa-vmstat.node0.nr_mlock
>>>        3761 ± 32%    +273.2%      14037        numa-vmstat.node0.nr_shmem
>>>        1557 ± 71%    +545.2%      10045 ±  2%  numa-vmstat.node0.nr_zone_active_anon
>>>       19391 ±  3%     -50.8%       9533 ±  7%  numa-vmstat.node1.nr_active_anon
>>>       58.25 ±157%    +301.1%     233.67 ± 67%  numa-vmstat.node1.nr_active_file
>>>      143875 ±  3%      -8.7%     131291 ±  3%  numa-vmstat.node1.nr_file_pages
>>>       21981 ±  4%     -52.9%      10346 ±  8%  numa-vmstat.node1.nr_shmem
>>>       19391 ±  3%     -50.8%       9533 ±  7%  numa-vmstat.node1.nr_zone_active_anon
>>>       58.25 ±157%    +301.1%     233.67 ± 67%  numa-vmstat.node1.nr_zone_active_file
>>>      110768 ±  3%     -10.6%      99008 ±  4%  softirqs.CPU0.RCU
>>>      108763 ±  3%     -10.1%      97774 ±  4%  softirqs.CPU1.RCU
>>>      104199 ±  3%      -9.5%      94314 ±  4%  softirqs.CPU12.RCU
>>>      106156 ±  5%     -11.6%      93873 ±  4%  softirqs.CPU13.RCU
>>>      107231 ±  4%     -11.3%      95088 ±  6%  softirqs.CPU14.RCU
>>>      105547 ±  5%     -10.5%      94461 ±  4%  softirqs.CPU15.RCU
>>>      104326 ±  5%      -7.4%      96653 ±  4%  softirqs.CPU22.RCU
>>>      108301 ±  3%      -9.9%      97571 ±  3%  softirqs.CPU25.RCU
>>>      109423 ±  3%     -10.9%      97448 ±  3%  softirqs.CPU26.RCU
>>>      109554 ±  4%      -7.2%     101691 ±  3%  softirqs.CPU31.RCU
>>>      108410 ±  4%     -11.2%      96233 ±  3%  softirqs.CPU37.RCU
>>>      107189 ±  4%      -9.8%      96642 ±  4%  softirqs.CPU38.RCU
>>>      108943 ±  4%     -10.9%      97052 ±  4%  softirqs.CPU39.RCU
>>>        3348 ± 37%    +214.8%      10541 ± 51%  softirqs.NET_RX
>>>       20942 ±  2%      -6.2%      19637 ±  4%  proc-vmstat.nr_active_anon
>>>       67173 ±  4%      -3.7%      64699        proc-vmstat.nr_anon_pages
>>>       10595            -1.7%      10418        proc-vmstat.nr_mapped
>>>       25771 ±  2%      -5.1%      24461 ±  3%  proc-vmstat.nr_shmem
>>>       38683            +5.7%      40878 ±  3%  proc-vmstat.nr_slab_unreclaimable
>>>       20942 ±  2%      -6.2%      19637 ±  4%  proc-vmstat.nr_zone_active_anon
>>>       21010 ± 11%     -23.8%      16020 ± 11%  proc-vmstat.numa_hint_faults
>>>       16666 ± 11%     -28.6%      11904 ±  7%  proc-vmstat.numa_hint_faults_local
>>>       69020 ±  3%      -3.7%      66499        proc-vmstat.numa_other
>>>    21276729            -9.4%   19266740        proc-vmstat.pgactivate
>>>       45217            +4.0%      47044        proc-vmstat.pgalloc_dma32
>>>   1.065e+08            -2.8%  1.035e+08        proc-vmstat.pgalloc_normal
>>>      610.50 ± 10%     +83.9%       1123 ± 28%  proc-vmstat.pgdeactivate
>>>   1.209e+08            -4.1%  1.159e+08        proc-vmstat.pgfault
>>>   1.064e+08            -2.7%  1.035e+08        proc-vmstat.pgfree
>>>        5.50 ± 20%  +71439.4%       3934 ±141%  proc-vmstat.pgmigrate_fail
>>>        3540 ± 29%    +139.9%       8494 ±  7%  proc-vmstat.pgrotated
>>>    10600738            -9.5%    9598142        proc-vmstat.unevictable_pgs_mlocked
>>>    10600295            -9.5%    9598142        proc-vmstat.unevictable_pgs_munlocked
>>>        0.05 ± 97%     -59.4%       0.02 ±  2%  perf-sched.sch_delay.max.ms.__x64_sys_pause.do_syscall_64.entry_SYSCALL_64_after_hwframe.[unknown]
>>>        0.03 ±  6%  +29749.0%       9.03 ± 71%  perf-sched.sch_delay.max.ms.schedule_timeout.rcu_gp_kthread.kthread.ret_from_fork
>>>       20.46 ±  3%      -8.9%      18.64 ±  4%  perf-sched.sch_delay.max.ms.wait_for_partner.fifo_open.do_dentry_open.path_openat
>>>        7526 ± 15%     -25.7%       5592 ±  5%  perf-sched.total_wait_and_delay.max.ms
>>>        7526 ± 15%     -25.7%       5592 ±  5%  perf-sched.total_wait_time.max.ms
>>>        9.67 ±168%     -97.6%       0.24 ± 30%  perf-sched.wait_and_delay.avg.ms.preempt_schedule_common._cond_resched.do_user_addr_fault.exc_page_fault.asm_exc_page_fault
>>>      343.12 ±139%    +236.4%       1154 ± 14%  perf-sched.wait_and_delay.avg.ms.preempt_schedule_common._cond_resched.generic_perform_write.__generic_file_write_iter.generic_file_write_iter
>>>        1.39 ±173%  +26107.3%     364.94        perf-sched.wait_and_delay.avg.ms.schedule_hrtimeout_range_clock.ep_poll.do_epoll_wait.__x64_sys_epoll_wait
>>>        5.59 ±  6%     -20.2%       4.46 ±  6%  perf-sched.wait_and_delay.avg.ms.schedule_timeout.rcu_gp_kthread.kthread.ret_from_fork
>>>        1770 ±  6%     +26.3%       2236 ±  7%  perf-sched.wait_and_delay.count.schedule_timeout.rcu_gp_kthread.kthread.ret_from_fork
>>>       18.50 ± 51%    +272.3%      68.88 ± 22%  perf-sched.wait_and_delay.max.ms.preempt_schedule_common._cond_resched.remove_vma.__do_munmap.__vm_munmap
>>>       20.79 ± 26%    +217.3%      65.97 ±  9%  perf-sched.wait_and_delay.max.ms.preempt_schedule_common._cond_resched.unmap_vmas.unmap_region.__do_munmap
>>>      242.02 ± 18%     -60.5%      95.68 ± 51%  perf-sched.wait_and_delay.max.ms.schedule_timeout.rcu_gp_kthread.kthread.ret_from_fork
>>>        0.12 ± 32%    +266.9%       0.43 ± 78%  perf-sched.wait_time.avg.ms.preempt_schedule_common._cond_resched.__alloc_pages_nodemask.alloc_pages_vma.do_anonymous_page
>>>        1971 ±123%     -99.4%      11.34 ± 10%  perf-sched.wait_time.avg.ms.preempt_schedule_common._cond_resched.__alloc_pages_nodemask.alloc_pages_vma.shmem_alloc_page
>>>        9.67 ±168%     -97.6%       0.24 ± 30%  perf-sched.wait_time.avg.ms.preempt_schedule_common._cond_resched.do_user_addr_fault.exc_page_fault.asm_exc_page_fault
>>>      348.74 ±136%    +231.0%       1154 ± 14%  perf-sched.wait_time.avg.ms.preempt_schedule_common._cond_resched.generic_perform_write.__generic_file_write_iter.generic_file_write_iter
>>>        5.86 ±  4%   +1071.7%      68.64 ± 19%  perf-sched.wait_time.avg.ms.preempt_schedule_common._cond_resched.khugepaged.kthread.ret_from_fork
>>>       38.63 ± 52%    +843.9%     364.62        perf-sched.wait_time.avg.ms.schedule_hrtimeout_range_clock.ep_poll.do_epoll_wait.__x64_sys_epoll_wait
>>>        5.57 ±  6%     -20.3%       4.44 ±  6%  perf-sched.wait_time.avg.ms.schedule_timeout.rcu_gp_kthread.kthread.ret_from_fork
>>>        1.57 ± 14%    +683.9%      12.28 ±100%  perf-sched.wait_time.max.ms.preempt_schedule_common._cond_resched.__alloc_pages_nodemask.alloc_pages_vma.do_anonymous_page
>>>        2468 ±103%     -99.5%      11.34 ± 10%  perf-sched.wait_time.max.ms.preempt_schedule_common._cond_resched.__alloc_pages_nodemask.alloc_pages_vma.shmem_alloc_page
>>>        3.91 ± 14%     -41.4%       2.29 ±  6%  perf-sched.wait_time.max.ms.preempt_schedule_common._cond_resched.down_read.__x64_sys_msync.do_syscall_64
>>>        7.68 ± 17%   +1121.8%      93.81 ± 13%  perf-sched.wait_time.max.ms.preempt_schedule_common._cond_resched.khugepaged.kthread.ret_from_fork
>>>       18.50 ± 51%    +272.3%      68.88 ± 22%  perf-sched.wait_time.max.ms.preempt_schedule_common._cond_resched.remove_vma.__do_munmap.__vm_munmap
>>>       20.79 ± 26%    +217.3%      65.97 ±  9%  perf-sched.wait_time.max.ms.preempt_schedule_common._cond_resched.unmap_vmas.unmap_region.__do_munmap
>>>      242.00 ± 18%     -62.5%      90.66 ± 61%  perf-sched.wait_time.max.ms.schedule_timeout.rcu_gp_kthread.kthread.ret_from_fork
>>>    47358339 ±  7%     -10.5%   42382902 ±  2%  perf-stat.i.branch-misses
>>>    19845519 ±  2%      -8.5%   18156709 ±  3%  perf-stat.i.cache-misses
>>>    1.72e+08            -5.7%  1.621e+08        perf-stat.i.cache-references
>>>      725.90 ±  5%     -13.2%     630.12 ±  5%  perf-stat.i.cpu-migrations
>>>    3.09e+09            -4.3%  2.957e+09        perf-stat.i.dTLB-stores
>>>    19570792            -6.2%   18352176        perf-stat.i.iTLB-load-misses
>>>       19871 ±  4%      +9.8%      21819 ±  3%  perf-stat.i.instructions-per-iTLB-miss
>>>      382.70            -3.0%     371.36 ±  2%  perf-stat.i.metric.M/sec
>>>     7111839 ±  4%     -11.4%    6302363 ±  2%  perf-stat.i.node-load-misses
>>>     8853002 ±  4%      -9.6%    8004893        perf-stat.i.node-loads
>>>     6500622            -7.9%    5986359 ±  4%  perf-stat.i.node-store-misses
>>>     8596412            -6.5%    8036645 ±  3%  perf-stat.i.node-stores
>>>        5.60 ±  2%      -3.9%       5.38        perf-stat.overall.MPKI
>>>        0.63 ±  7%      -0.1        0.57        perf-stat.overall.branch-miss-rate%
>>>        5180 ±  2%      +9.2%       5659 ±  3%  perf-stat.overall.cycles-between-cache-misses
>>>        1575            +4.8%       1651 ±  3%  perf-stat.overall.instructions-per-iTLB-miss
>>>       44.52            -0.5       43.98        perf-stat.overall.node-load-miss-rate%
>>>       43.02            -0.4       42.61        perf-stat.overall.node-store-miss-rate%
>>>    47096030 ±  7%     -11.4%   41713545 ±  3%  perf-stat.ps.branch-misses
>>>    19690067 ±  2%      -9.4%   17836765 ±  3%  perf-stat.ps.cache-misses
>>>     1.7e+08 ±  2%      -6.9%  1.582e+08 ±  2%  perf-stat.ps.cache-references
>>>   1.019e+11            -1.1%  1.008e+11        perf-stat.ps.cpu-cycles
>>>      748.16 ±  5%     -11.3%     663.37 ±  5%  perf-stat.ps.cpu-migrations
>>>   3.052e+09            -5.5%  2.885e+09 ±  2%  perf-stat.ps.dTLB-stores
>>>    19264658            -7.6%   17802646 ±  2%  perf-stat.ps.iTLB-load-misses
>>>     7041022 ±  4%     -12.4%    6164936        perf-stat.ps.node-load-misses
>>>     8774703 ±  4%     -10.5%    7851024        perf-stat.ps.node-loads
>>>     6428583            -9.0%    5851692 ±  4%  perf-stat.ps.node-store-misses
>>>     8513354            -7.5%    7878662 ±  4%  perf-stat.ps.node-stores
>>>      793.75 ±  6%     -16.3%     664.33 ± 12%  interrupts.22:IO-APIC.22-fasteoi.ehci_hcd:usb1
>>>       10846 ± 12%     -22.7%       8387 ±  8%  interrupts.CPU0.CAL:Function_call_interrupts
>>>        4222 ± 14%     -49.6%       2128 ± 43%  interrupts.CPU1.NMI:Non-maskable_interrupts
>>>        4222 ± 14%     -49.6%       2128 ± 43%  interrupts.CPU1.PMI:Performance_monitoring_interrupts
>>>        3219 ± 28%     -21.0%       2544 ± 29%  interrupts.CPU10.NMI:Non-maskable_interrupts
>>>        3219 ± 28%     -21.0%       2544 ± 29%  interrupts.CPU10.PMI:Performance_monitoring_interrupts
>>>       12387 ±  8%     -10.7%      11065 ± 10%  interrupts.CPU10.TLB:TLB_shootdowns
>>>        3883 ± 23%     -35.3%       2511 ± 43%  interrupts.CPU11.NMI:Non-maskable_interrupts
>>>        3883 ± 23%     -35.3%       2511 ± 43%  interrupts.CPU11.PMI:Performance_monitoring_interrupts
>>>       11453 ± 37%     -34.1%       7553 ± 14%  interrupts.CPU12.CAL:Function_call_interrupts
>>>        2074 ±126%     -77.6%     464.00 ±  9%  interrupts.CPU12.RES:Rescheduling_interrupts
>>>       13347 ± 59%     -45.5%       7274 ± 10%  interrupts.CPU12.TLB:TLB_shootdowns
>>>       11138 ± 23%     -31.3%       7657 ± 13%  interrupts.CPU13.CAL:Function_call_interrupts
>>>        3399 ± 34%     -29.1%       2410 ± 50%  interrupts.CPU15.NMI:Non-maskable_interrupts
>>>        3399 ± 34%     -29.1%       2410 ± 50%  interrupts.CPU15.PMI:Performance_monitoring_interrupts
>>>        3318 ± 29%     -25.3%       2479 ± 34%  interrupts.CPU16.NMI:Non-maskable_interrupts
>>>        3318 ± 29%     -25.3%       2479 ± 34%  interrupts.CPU16.PMI:Performance_monitoring_interrupts
>>>        4574 ± 17%     -58.7%       1890 ± 29%  interrupts.CPU20.NMI:Non-maskable_interrupts
>>>        4574 ± 17%     -58.7%       1890 ± 29%  interrupts.CPU20.PMI:Performance_monitoring_interrupts
>>>        2768 ± 37%     -44.7%       1531 ± 30%  interrupts.CPU22.NMI:Non-maskable_interrupts
>>>        2768 ± 37%     -44.7%       1531 ± 30%  interrupts.CPU22.PMI:Performance_monitoring_interrupts
>>>      516.00 ± 23%     -18.9%     418.33 ±  5%  interrupts.CPU22.RES:Rescheduling_interrupts
>>>        3326 ± 27%     -28.2%       2387 ± 34%  interrupts.CPU23.NMI:Non-maskable_interrupts
>>>        3326 ± 27%     -28.2%       2387 ± 34%  interrupts.CPU23.PMI:Performance_monitoring_interrupts
>>>        3850 ± 20%     -39.2%       2340 ± 50%  interrupts.CPU25.NMI:Non-maskable_interrupts
>>>        3850 ± 20%     -39.2%       2340 ± 50%  interrupts.CPU25.PMI:Performance_monitoring_interrupts
>>>      859.25 ± 48%     -46.0%     464.33 ±  5%  interrupts.CPU25.RES:Rescheduling_interrupts
>>>       13154 ± 35%     -25.2%       9841 ± 14%  interrupts.CPU25.TLB:TLB_shootdowns
>>>        3711 ± 23%     -45.6%       2018 ± 45%  interrupts.CPU26.NMI:Non-maskable_interrupts
>>>        3711 ± 23%     -45.6%       2018 ± 45%  interrupts.CPU26.PMI:Performance_monitoring_interrupts
>>>        4507 ±  4%     -43.3%       2556 ± 40%  interrupts.CPU27.NMI:Non-maskable_interrupts
>>>        4507 ±  4%     -43.3%       2556 ± 40%  interrupts.CPU27.PMI:Performance_monitoring_interrupts
>>>       11363 ±  9%     -14.9%       9671 ±  8%  interrupts.CPU27.TLB:TLB_shootdowns
>>>        4430 ±  5%     -51.0%       2172 ± 43%  interrupts.CPU28.NMI:Non-maskable_interrupts
>>>        4430 ±  5%     -51.0%       2172 ± 43%  interrupts.CPU28.PMI:Performance_monitoring_interrupts
>>>        9512 ±  3%     -16.7%       7921 ±  8%  interrupts.CPU29.CAL:Function_call_interrupts
>>>        3914 ± 14%     -59.0%       1606 ± 27%  interrupts.CPU29.NMI:Non-maskable_interrupts
>>>        3914 ± 14%     -59.0%       1606 ± 27%  interrupts.CPU29.PMI:Performance_monitoring_interrupts
>>>      998.00 ± 77%     -55.5%     444.00 ±  2%  interrupts.CPU29.RES:Rescheduling_interrupts
>>>       11508 ± 12%     -22.8%       8881 ± 10%  interrupts.CPU29.TLB:TLB_shootdowns
>>>        3287 ± 22%     -36.6%       2082 ± 43%  interrupts.CPU30.NMI:Non-maskable_interrupts
>>>        3287 ± 22%     -36.6%       2082 ± 43%  interrupts.CPU30.PMI:Performance_monitoring_interrupts
>>>      571.75 ± 12%     -21.0%     451.67 ± 10%  interrupts.CPU30.RES:Rescheduling_interrupts
>>>       10554 ± 20%     -26.5%       7757 ±  6%  interrupts.CPU31.CAL:Function_call_interrupts
>>>        3521 ± 21%     -47.8%       1838 ±  9%  interrupts.CPU31.NMI:Non-maskable_interrupts
>>>        3521 ± 21%     -47.8%       1838 ±  9%  interrupts.CPU31.PMI:Performance_monitoring_interrupts
>>>      604.00 ± 20%     -26.0%     447.00 ± 14%  interrupts.CPU31.RES:Rescheduling_interrupts
>>>        3946 ± 21%     -51.9%       1898 ±  6%  interrupts.CPU32.NMI:Non-maskable_interrupts
>>>        3946 ± 21%     -51.9%       1898 ±  6%  interrupts.CPU32.PMI:Performance_monitoring_interrupts
>>>        2221 ± 74%     -74.2%     573.00 ± 13%  interrupts.CPU32.RES:Rescheduling_interrupts
>>>        2969 ± 22%     -40.2%       1774 ± 11%  interrupts.CPU33.NMI:Non-maskable_interrupts
>>>        2969 ± 22%     -40.2%       1774 ± 11%  interrupts.CPU33.PMI:Performance_monitoring_interrupts
>>>       13653 ± 16%     -40.8%       8079 ±  6%  interrupts.CPU34.CAL:Function_call_interrupts
>>>        3074 ± 20%     -41.7%       1791 ± 18%  interrupts.CPU34.NMI:Non-maskable_interrupts
>>>        3074 ± 20%     -41.7%       1791 ± 18%  interrupts.CPU34.PMI:Performance_monitoring_interrupts
>>>       18328 ± 31%     -48.1%       9519 ±  7%  interrupts.CPU34.TLB:TLB_shootdowns
>>>       11937 ± 21%     -36.1%       7623 ± 15%  interrupts.CPU35.CAL:Function_call_interrupts
>>>        3503 ± 30%     -36.4%       2229 ± 12%  interrupts.CPU35.NMI:Non-maskable_interrupts
>>>        3503 ± 30%     -36.4%       2229 ± 12%  interrupts.CPU35.PMI:Performance_monitoring_interrupts
>>>       16203 ± 41%     -47.9%       8444 ± 15%  interrupts.CPU35.TLB:TLB_shootdowns
>>>        4311 ± 20%     -43.4%       2440 ± 50%  interrupts.CPU37.NMI:Non-maskable_interrupts
>>>        4311 ± 20%     -43.4%       2440 ± 50%  interrupts.CPU37.PMI:Performance_monitoring_interrupts
>>>        1028 ± 98%     -60.6%     404.67 ±  5%  interrupts.CPU37.RES:Rescheduling_interrupts
>>>       11376 ± 22%     -35.6%       7324 ±  2%  interrupts.CPU37.TLB:TLB_shootdowns
>>>        3390 ± 35%     -35.8%       2177 ± 50%  interrupts.CPU38.NMI:Non-maskable_interrupts
>>>        3390 ± 35%     -35.8%       2177 ± 50%  interrupts.CPU38.PMI:Performance_monitoring_interrupts
>>>        2077 ±133%     -80.7%     401.33 ± 12%  interrupts.CPU38.RES:Rescheduling_interrupts
>>>        3760 ± 19%     -54.3%       1716 ± 66%  interrupts.CPU39.NMI:Non-maskable_interrupts
>>>        3760 ± 19%     -54.3%       1716 ± 66%  interrupts.CPU39.PMI:Performance_monitoring_interrupts
>>>        3338 ± 68%     -85.1%     496.67 ± 17%  interrupts.CPU4.RES:Rescheduling_interrupts
>>>        3731 ± 27%     -45.6%       2030 ± 38%  interrupts.CPU40.NMI:Non-maskable_interrupts
>>>        3731 ± 27%     -45.6%       2030 ± 38%  interrupts.CPU40.PMI:Performance_monitoring_interrupts
>>>        4265 ±  5%     -43.5%       2409 ± 45%  interrupts.CPU42.NMI:Non-maskable_interrupts
>>>        4265 ±  5%     -43.5%       2409 ± 45%  interrupts.CPU42.PMI:Performance_monitoring_interrupts
>>>        3425 ± 18%     -41.4%       2007 ± 51%  interrupts.CPU43.NMI:Non-maskable_interrupts
>>>        3425 ± 18%     -41.4%       2007 ± 51%  interrupts.CPU43.PMI:Performance_monitoring_interrupts
>>>      627.00 ± 46%     -43.4%     354.67 ±  2%  interrupts.CPU43.RES:Rescheduling_interrupts
>>>        4679 ± 13%     -71.6%       1331 ± 21%  interrupts.CPU44.NMI:Non-maskable_interrupts
>>>        4679 ± 13%     -71.6%       1331 ± 21%  interrupts.CPU44.PMI:Performance_monitoring_interrupts
>>>      494.50 ± 18%     -25.8%     367.00 ± 15%  interrupts.CPU44.RES:Rescheduling_interrupts
>>>        3975 ± 18%     -43.4%       2249 ± 46%  interrupts.CPU45.NMI:Non-maskable_interrupts
>>>        3975 ± 18%     -43.4%       2249 ± 46%  interrupts.CPU45.PMI:Performance_monitoring_interrupts
>>>        8711 ±  9%     -11.6%       7705        interrupts.CPU46.CAL:Function_call_interrupts
>>>        3650 ± 22%     -44.9%       2010 ± 55%  interrupts.CPU46.NMI:Non-maskable_interrupts
>>>        3650 ± 22%     -44.9%       2010 ± 55%  interrupts.CPU46.PMI:Performance_monitoring_interrupts
>>>      440.75 ±  9%     -16.7%     367.00 ±  7%  interrupts.CPU46.RES:Rescheduling_interrupts
>>>        7891 ± 10%     -25.7%       5860 ± 11%  interrupts.CPU47.CAL:Function_call_interrupts
>>>      793.00 ± 40%     -50.6%     392.00 ±  3%  interrupts.CPU47.RES:Rescheduling_interrupts
>>>        2663 ± 23%     -21.6%       2087 ±  7%  interrupts.CPU5.NMI:Non-maskable_interrupts
>>>        2663 ± 23%     -21.6%       2087 ±  7%  interrupts.CPU5.PMI:Performance_monitoring_interrupts
>>>       12129 ± 30%     -32.7%       8164 ±  2%  interrupts.CPU7.CAL:Function_call_interrupts
>>>       14622 ± 38%     -33.7%       9701 ±  6%  interrupts.CPU7.TLB:TLB_shootdowns
>>>      793.75 ±  6%     -16.3%     664.33 ± 12%  interrupts.CPU8.22:IO-APIC.22-fasteoi.ehci_hcd:usb1
>>>        2516 ±  7%     +26.7%       3188 ± 15%  interrupts.CPU8.NMI:Non-maskable_interrupts
>>>        2516 ±  7%     +26.7%       3188 ± 15%  interrupts.CPU8.PMI:Performance_monitoring_interrupts
>>>        1784 ± 54%     -58.8%     735.00 ±  4%�� interrupts.CPU8.RES:Rescheduling_interrupts
>>>      656.75 ± 18%     +51.0%     992.00 ± 20%  interrupts.CPU9.RES:Rescheduling_interrupts
>>>      168231 ±  7%     -33.4%     112034 ± 19%  interrupts.NMI:Non-maskable_interrupts
>>>      168231 ±  7%     -33.4%     112034 ± 19%  interrupts.PMI:Performance_monitoring_interrupts
>>>        7.01 ± 14%      -3.4        3.60 ± 50%  perf-profile.calltrace.cycles-pp.syscall_exit_to_user_mode.entry_SYSCALL_64_after_hwframe
>>>        4.21 ± 21%      -1.9        2.29 ± 63%  perf-profile.calltrace.cycles-pp.exit_to_user_mode_prepare.syscall_exit_to_user_mode.entry_SYSCALL_64_after_hwframe
>>>        2.84 ± 12%      -1.9        0.96 ± 39%  perf-profile.calltrace.cycles-pp.ksys_mmap_pgoff.do_syscall_64.entry_SYSCALL_64_after_hwframe
>>>        4.06 ± 21%      -1.8        2.23 ± 65%  perf-profile.calltrace.cycles-pp.task_work_run.exit_to_user_mode_prepare.syscall_exit_to_user_mode.entry_SYSCALL_64_after_hwframe
>>>        2.64 ± 13%      -1.8        0.87 ± 40%  perf-profile.calltrace.cycles-pp.vm_mmap_pgoff.ksys_mmap_pgoff.do_syscall_64.entry_SYSCALL_64_after_hwframe
>>>        3.84 ± 22%      -1.7        2.13 ± 67%  perf-profile.calltrace.cycles-pp.__fput.task_work_run.exit_to_user_mode_prepare.syscall_exit_to_user_mode.entry_SYSCALL_64_after_hwframe
>>>        1.49 ±  7%      -0.9        0.54 ± 70%  perf-profile.calltrace.cycles-pp.alloc_empty_file.path_openat.do_filp_open.do_sys_openat2.do_sys_open
>>>        1.46 ±  8%      -0.9        0.53 ± 70%  perf-profile.calltrace.cycles-pp.__alloc_file.alloc_empty_file.path_openat.do_filp_open.do_sys_openat2
>>>        1.28 ±  5%      -0.8        0.51 ± 71%  perf-profile.calltrace.cycles-pp.syscall_return_via_sysret
>>>        1.06 ±  6%      -0.6        0.42 ± 71%  perf-profile.calltrace.cycles-pp.__entry_text_start
>>>        0.77 ± 13%      +0.3        1.05 ± 11%  perf-profile.calltrace.cycles-pp.sysvec_apic_timer_interrupt.asm_sysvec_apic_timer_interrupt.cpuidle_enter_state.cpuidle_enter.do_idle
>>>        0.82 ± 12%      +0.3        1.12 ± 11%  perf-profile.calltrace.cycles-pp.asm_sysvec_apic_timer_interrupt.cpuidle_enter_state.cpuidle_enter.do_idle.cpu_startup_entry
>>>        1.23 ± 13%      +0.4        1.68 ± 18%  perf-profile.calltrace.cycles-pp.wait_for_xmitr.serial8250_console_putchar.uart_console_write.serial8250_console_write.console_unlock
>>>        1.23 ± 13%      +0.4        1.68 ± 18%  perf-profile.calltrace.cycles-pp.serial8250_console_putchar.uart_console_write.serial8250_console_write.console_unlock.vprintk_emit
>>>        1.82 ± 13%      +0.6        2.39 ± 17%  perf-profile.calltrace.cycles-pp.create_basic_memory_bitmaps.snapshot_open.misc_open.chrdev_open.do_dentry_open
>>>        0.00            +0.6        0.61 ± 12%  perf-profile.calltrace.cycles-pp.__sysvec_apic_timer_interrupt.asm_call_sysvec_on_stack.sysvec_apic_timer_interrupt.asm_sysvec_apic_timer_interrupt.cpuidle_enter_state
>>>        0.00            +0.6        0.62 ± 12%  perf-profile.calltrace.cycles-pp.asm_call_sysvec_on_stack.sysvec_apic_timer_interrupt.asm_sysvec_apic_timer_interrupt.cpuidle_enter_state.cpuidle_enter
>>>        4.96 ±  7%      +1.7        6.69 ± 19%  perf-profile.calltrace.cycles-pp.mutex_spin_on_owner.__mutex_lock.misc_open.chrdev_open.do_dentry_open
>>>        7.19 ± 14%      -3.4        3.78 ± 45%  perf-profile.children.cycles-pp.syscall_exit_to_user_mode
>>>        3.03 ± 53%      -2.8        0.24 ± 21%  perf-profile.children.cycles-pp.link_path_walk
>>>        4.28 ± 21%      -1.9        2.35 ± 61%  perf-profile.children.cycles-pp.exit_to_user_mode_prepare
>>>        2.85 ± 12%      -1.9        0.96 ± 40%  perf-profile.children.cycles-pp.ksys_mmap_pgoff
>>>        4.26 ± 21%      -1.9        2.38 ± 62%  perf-profile.children.cycles-pp.task_work_run
>>>        2.65 ± 13%      -1.8        0.88 ± 40%  perf-profile.children.cycles-pp.vm_mmap_pgoff
>>>        4.04 ± 22%      -1.8        2.27 ± 64%  perf-profile.children.cycles-pp.__fput
>>>        1.69 ± 54%      -1.6        0.13 ± 22%  perf-profile.children.cycles-pp.inode_permission
>>>        1.47 ± 51%      -1.3        0.16 ±  3%  perf-profile.children.cycles-pp.walk_component
>>>        1.33 ± 54%      -1.2        0.11 ± 14%  perf-profile.children.cycles-pp.lookup_fast
>>>        1.49 ± 27%      -1.1        0.36 ± 28%  perf-profile.children.cycles-pp._raw_spin_lock
>>>        1.49 ± 31%      -1.1        0.40 ± 48%  perf-profile.children.cycles-pp.rwsem_down_write_slowpath
>>>        1.51 ± 21%      -1.1        0.44 ± 53%  perf-profile.children.cycles-pp.down_write_killable
>>>        0.91 ± 62%      -0.9        0.04 ± 71%  perf-profile.children.cycles-pp.hidraw_release
>>>        1.58 ±  8%      -0.8        0.73 ± 24%  perf-profile.children.cycles-pp.alloc_empty_file
>>>        1.55 ±  9%      -0.8        0.72 ± 24%  perf-profile.children.cycles-pp.__alloc_file
>>>        1.49 ±  5%      -0.7        0.74 ± 28%  perf-profile.children.cycles-pp.syscall_return_via_sysret
>>>        1.26 ±  6%      -0.6        0.63 ± 27%  perf-profile.children.cycles-pp.__entry_text_start
>>>        1.12 ±  9%      -0.6        0.55 ± 24%  perf-profile.children.cycles-pp.kmem_cache_alloc
>>>        0.87 ± 18%      -0.6        0.31 ± 34%  perf-profile.children.cycles-pp.do_mmap
>>>        0.63 ±  7%      -0.4        0.22 ± 25%  perf-profile.children.cycles-pp.dput
>>>        0.55 ± 17%      -0.4        0.15 ± 34%  perf-profile.children.cycles-pp.security_file_open
>>>        0.53 ± 19%      -0.4        0.14 ± 36%  perf-profile.children.cycles-pp.apparmor_file_open
>>>        0.52 ± 32%      -0.4        0.14 ± 55%  perf-profile.children.cycles-pp.rwsem_spin_on_owner
>>>        0.48 ± 22%      -0.4        0.12 ± 13%  perf-profile.children.cycles-pp.ima_file_check
>>>        0.65 ±  7%      -0.4        0.29 ± 35%  perf-profile.children.cycles-pp.__x64_sys_select
>>>        0.47 ± 23%      -0.4        0.11 ± 14%  perf-profile.children.cycles-pp.security_task_getsecid
>>>        0.64 ±  7%      -0.4        0.29 ± 35%  perf-profile.children.cycles-pp.kern_select
>>>        0.62 ± 13%      -0.3        0.27 ± 36%  perf-profile.children.cycles-pp.__x64_sys_exit_group
>>>        0.62 ± 12%      -0.3        0.28 ± 33%  perf-profile.children.cycles-pp.do_group_exit
>>>        0.62 ± 12%      -0.3        0.28 ± 33%  perf-profile.children.cycles-pp.do_exit
>>>        0.45 ± 24%      -0.3        0.11 ± 19%  perf-profile.children.cycles-pp.apparmor_task_getsecid
>>>        0.49 ± 65%      -0.3        0.16 ± 83%  perf-profile.children.cycles-pp.__x64_sys_munmap
>>>        0.49 ± 64%      -0.3        0.16 ± 83%  perf-profile.children.cycles-pp.__vm_munmap
>>>        0.48 ±  5%      -0.3        0.16 ± 34%  perf-profile.children.cycles-pp.lockref_put_or_lock
>>>        0.51 ±  9%      -0.3        0.19 ± 15%  perf-profile.children.cycles-pp.security_file_alloc
>>>        0.51 ±  7%      -0.3        0.20 ± 38%  perf-profile.children.cycles-pp.__x64_sys_close
>>>        0.59 ± 12%      -0.3        0.30 ± 13%  perf-profile.children.cycles-pp.__do_sys_clone
>>>        0.59 ± 12%      -0.3        0.30 ± 13%  perf-profile.children.cycles-pp.kernel_clone
>>>        0.41 ± 11%      -0.3        0.13 ± 18%  perf-profile.children.cycles-pp.apparmor_file_alloc_security
>>>        0.57 ± 12%      -0.3        0.29 ± 12%  perf-profile.children.cycles-pp.copy_process
>>>        0.51 ±  8%      -0.3        0.23 ± 38%  perf-profile.children.cycles-pp.core_sys_select
>>>        0.43 ± 14%      -0.3        0.17 ± 14%  perf-profile.children.cycles-pp.mmput
>>>        0.43 ± 14%      -0.3        0.17 ± 14%  perf-profile.children.cycles-pp.exit_mmap
>>>        0.45 ±  8%      -0.3        0.19 ± 35%  perf-profile.children.cycles-pp.get_unmapped_area
>>>        0.40 ± 17%      -0.3        0.14 ± 20%  perf-profile.children.cycles-pp.__do_sys_newfstat
>>>        0.50 ± 13%      -0.2        0.26 ± 12%  perf-profile.children.cycles-pp.dup_mm
>>>        0.71 ±  9%      -0.2        0.48 ± 11%  perf-profile.children.cycles-pp.do_softirq_own_stack
>>>        0.75 ± 10%      -0.2        0.52 ± 11%  perf-profile.children.cycles-pp.irq_exit_rcu
>>>        0.49 ± 12%      -0.2        0.26 ± 12%  perf-profile.children.cycles-pp.dup_mmap
>>>        0.41 ±  9%      -0.2        0.19 ± 36%  perf-profile.children.cycles-pp.__fget_files
>>>        0.28 ± 52%      -0.2        0.06 ± 75%  perf-profile.children.cycles-pp.mmap_region
>>>        0.35 ±  5%      -0.2        0.14 ± 34%  perf-profile.children.cycles-pp.security_mmap_file
>>>        0.39 ±  7%      -0.2        0.18 ± 15%  perf-profile.children.cycles-pp.getname_flags
>>>        0.32 ±  7%      -0.2        0.12 ± 35%  perf-profile.children.cycles-pp.complete_walk
>>>        0.32 ±  7%      -0.2        0.12 ± 31%  perf-profile.children.cycles-pp.unlazy_walk
>>>        0.29 ± 22%      -0.2        0.09 ± 27%  perf-profile.children.cycles-pp.vfs_fstat
>>>        0.30 ± 17%      -0.2        0.10 ± 12%  perf-profile.children.cycles-pp.security_file_free
>>>        0.28 ± 21%      -0.2        0.08 ± 14%  perf-profile.children.cycles-pp.aa_get_task_label
>>>        0.29 ± 17%      -0.2        0.09 ± 14%  perf-profile.children.cycles-pp.apparmor_file_free_security
>>>        0.25 ± 32%      -0.2        0.06 ± 13%  perf-profile.children.cycles-pp.step_into
>>>        0.36 ± 10%      -0.2        0.17 ± 39%  perf-profile.children.cycles-pp.do_select
>>>        0.33 ±  7%      -0.2        0.14 ± 34%  perf-profile.children.cycles-pp.arch_get_unmapped_area_topdown
>>>        0.29 ±  7%      -0.2        0.10 ± 31%  perf-profile.children.cycles-pp.__legitimize_path
>>>        0.36 ± 17%      -0.2        0.17 ±  7%  perf-profile.children.cycles-pp.__schedule
>>>        0.29 ±  8%      -0.2        0.11 ± 40%  perf-profile.children.cycles-pp.terminate_walk
>>>        0.34 ±  8%      -0.2        0.17 ± 34%  perf-profile.children.cycles-pp.__fget_light
>>>        0.25 ± 25%      -0.2        0.08 ± 24%  perf-profile.children.cycles-pp.may_open
>>>        0.25 ± 18%      -0.2        0.08 ± 35%  perf-profile.children.cycles-pp.mutex_lock
>>>        0.35 ±  7%      -0.2        0.18 ± 50%  perf-profile.children.cycles-pp.__alloc_pages_nodemask
>>>        0.39 ± 13%      -0.2        0.22 ± 24%  perf-profile.children.cycles-pp.ksys_lseek
>>>        0.25 ± 58%      -0.2        0.09 ± 84%  perf-profile.children.cycles-pp.__do_munmap
>>>        0.22 ±  8%      -0.2        0.06 ± 71%  perf-profile.children.cycles-pp.lockref_get_not_dead
>>>        0.21 ± 17%      -0.2        0.06 ± 75%  perf-profile.children.cycles-pp.cdev_put
>>>        0.25 ± 10%      -0.2        0.10 ± 38%  perf-profile.children.cycles-pp.__close_fd
>>>        0.24 ± 10%      -0.1        0.10 ± 38%  perf-profile.children.cycles-pp.pick_file
>>>        0.37 ± 10%      -0.1        0.23 ± 12%  perf-profile.children.cycles-pp._raw_spin_lock_irqsave
>>>        0.27 ±  9%      -0.1        0.13 ± 33%  perf-profile.children.cycles-pp.__alloc_fd
>>>        0.21 ± 23%      -0.1        0.07 ± 20%  perf-profile.children.cycles-pp.free_pgtables
>>>        0.27 ± 20%      -0.1        0.13 ±  9%  perf-profile.children.cycles-pp.schedule
>>>        0.26 ±  7%      -0.1        0.12 ± 18%  perf-profile.children.cycles-pp.strncpy_from_user
>>>        0.23 ± 10%      -0.1        0.09 ± 39%  perf-profile.children.cycles-pp.filp_close
>>>        0.20 ± 11%      -0.1        0.07 ± 71%  perf-profile.children.cycles-pp.vm_unmapped_area
>>>        0.25 ±  4%      -0.1        0.12 ± 29%  perf-profile.children.cycles-pp.entry_SYSCALL_64_safe_stack
>>>        0.17 ± 16%      -0.1        0.04 ± 76%  perf-profile.children.cycles-pp.kobject_put
>>>        0.19 ± 19%      -0.1        0.06 ± 29%  perf-profile.children.cycles-pp.copy_page_range
>>>        0.22 ± 11%      -0.1        0.10 ± 14%  perf-profile.children.cycles-pp.try_to_wake_up
>>>        0.21 ±  9%      -0.1        0.08 ± 34%  perf-profile.children.cycles-pp.common_file_perm
>>>        0.22 ± 10%      -0.1        0.10 ± 32%  perf-profile.children.cycles-pp.fput_many
>>>        0.29 ± 21%      -0.1        0.17 ± 47%  perf-profile.children.cycles-pp.drm_release
>>>        0.29 ± 21%      -0.1        0.17 ± 47%  perf-profile.children.cycles-pp.drm_client_dev_restore
>>>        0.29 ± 21%      -0.1        0.17 ± 47%  perf-profile.children.cycles-pp.drm_fbdev_client_restore
>>>        0.29 ± 21%      -0.1        0.17 ± 47%  perf-profile.children.cycles-pp.__drm_fb_helper_restore_fbdev_mode_unlocked
>>>        0.29 ± 21%      -0.1        0.17 ± 47%  perf-profile.children.cycles-pp.drm_client_modeset_commit
>>>        0.29 ± 21%      -0.1        0.17 ± 47%  perf-profile.children.cycles-pp.drm_client_modeset_commit_locked
>>>        0.29 ± 21%      -0.1        0.17 ± 47%  perf-profile.children.cycles-pp.drm_client_modeset_commit_atomic
>>>        0.28 ±  4%      -0.1        0.16 ± 48%  perf-profile.children.cycles-pp.get_page_from_freelist
>>>        0.20 ± 15%      -0.1        0.08 ± 30%  perf-profile.children.cycles-pp.unmap_vmas
>>>        0.16 ± 21%      -0.1        0.04 ± 76%  perf-profile.children.cycles-pp.copy_pte_range
>>>        0.18 ±  9%      -0.1        0.07 ± 71%  perf-profile.children.cycles-pp.__x64_sys_fcntl
>>>        0.22 ±  7%      -0.1        0.10 ± 36%  perf-profile.children.cycles-pp.__x64_sys_poll
>>>        0.22 ±  7%      -0.1        0.10 ±  9%  perf-profile.children.cycles-pp.__check_object_size
>>>        0.18 ± 12%      -0.1        0.07 ± 25%  perf-profile.children.cycles-pp.unmap_page_range
>>>        0.20 ±  8%      -0.1        0.09 ± 33%  perf-profile.children.cycles-pp.do_sys_poll
>>>        0.15 ±  7%      -0.1        0.05 ± 72%  perf-profile.children.cycles-pp.__might_sleep
>>>        0.15 ±  8%      -0.1        0.05 ± 78%  perf-profile.children.cycles-pp.obj_cgroup_charge
>>>        0.15 ±  9%      -0.1        0.06 ± 16%  perf-profile.children.cycles-pp.zap_pte_range
>>>        0.17 ± 20%      -0.1        0.07 ± 72%  perf-profile.children.cycles-pp.__slab_alloc
>>>        0.16 ± 22%      -0.1        0.07 ± 72%  perf-profile.children.cycles-pp.___slab_alloc
>>>        0.14 ± 19%      -0.1        0.04 ± 71%  perf-profile.children.cycles-pp.unlink_anon_vmas
>>>        0.17 ± 38%      -0.1        0.07 ± 74%  perf-profile.children.cycles-pp.block_ioctl
>>>        0.16 ±  9%      -0.1        0.07 ± 20%  perf-profile.children.cycles-pp.anon_vma_fork
>>>        0.18 ±  5%      -0.1        0.09 ± 18%  perf-profile.children.cycles-pp.___might_sleep
>>>        0.14 ± 32%      -0.1        0.05 ± 81%  perf-profile.children.cycles-pp.drm_stub_open
>>>        0.14 ± 32%      -0.1        0.05 ± 81%  perf-profile.children.cycles-pp.drm_open
>>>        0.14 ± 32%      -0.1        0.05 ± 81%  perf-profile.children.cycles-pp.drm_master_open
>>>        0.16 ±  9%      -0.1        0.07 ± 23%  perf-profile.children.cycles-pp.get_obj_cgroup_from_current
>>>        0.16 ± 38%      -0.1        0.07 ± 74%  perf-profile.children.cycles-pp.blkdev_ioctl
>>>        0.12 ± 12%      -0.1        0.04 ± 71%  perf-profile.children.cycles-pp._cond_resched
>>>        0.12 ±  8%      -0.1        0.04 ± 73%  perf-profile.children.cycles-pp.memset_erms
>>>        0.14 ± 25%      -0.1        0.06 ± 72%  perf-profile.children.cycles-pp.allocate_slab
>>>        0.12 ± 13%      -0.1        0.04 ± 71%  perf-profile.children.cycles-pp.call_rcu
>>>        0.17 ± 25%      -0.1        0.09 ± 50%  perf-profile.children.cycles-pp.path_lookupat
>>>        0.17 ± 30%      -0.1        0.10 ± 61%  perf-profile.children.cycles-pp.__fxstat64
>>>        0.11 ± 12%      -0.1        0.03 ± 70%  perf-profile.children.cycles-pp.anon_vma_clone
>>>        0.11 ± 10%      -0.1        0.03 ± 70%  perf-profile.children.cycles-pp.__mod_memcg_lruvec_state
>>>        0.15 ± 16%      -0.1        0.08 ± 10%  perf-profile.children.cycles-pp.pick_next_task_fair
>>>        0.14 ±  6%      -0.1        0.07 ± 11%  perf-profile.children.cycles-pp.syscall_enter_from_user_mode
>>>        0.10 ± 12%      -0.1        0.03 ± 70%  perf-profile.children.cycles-pp.cp_new_stat
>>>        0.13 ± 10%      -0.1        0.06 ± 19%  perf-profile.children.cycles-pp.__d_lookup_rcu
>>>        0.10 ± 15%      -0.1        0.04 ± 73%  perf-profile.children.cycles-pp.generic_permission
>>>        0.14 ±  9%      -0.1        0.08 ± 20%  perf-profile.children.cycles-pp.__x86_retpoline_rax
>>>        0.08 ±  5%      -0.1        0.03 ± 70%  perf-profile.children.cycles-pp.ktime_get_ts64
>>>        0.09 ±  4%      -0.1        0.04 ± 71%  perf-profile.children.cycles-pp.__might_fault
>>>        0.07 ± 17%      -0.0        0.03 ± 70%  perf-profile.children.cycles-pp.schedule_timeout
>>>        0.11 ±  9%      -0.0        0.08 ± 12%  perf-profile.children.cycles-pp.update_load_avg
>>>        0.09 ±  8%      -0.0        0.06 ±  8%  perf-profile.children.cycles-pp.newidle_balance
>>>        0.10 ± 14%      -0.0        0.07 ±  7%  perf-profile.children.cycles-pp.task_tick_fair
>>>        0.06 ± 15%      +0.0        0.08 ± 16%  perf-profile.children.cycles-pp.tick_irq_enter
>>>        0.04 ± 57%      +0.0        0.06        perf-profile.children.cycles-pp.sched_clock_cpu
>>>        0.08 ± 13%      +0.0        0.11 ±  4%  perf-profile.children.cycles-pp.irq_enter_rcu
>>>        0.07 ±  5%      +0.0        0.12 ± 24%  perf-profile.children.cycles-pp.clockevents_program_event
>>>        0.01 ±173%      +0.1        0.07 ± 18%  perf-profile.children.cycles-pp.rebalance_domains
>>>        0.09 ±  4%      +0.1        0.15 ± 26%  perf-profile.children.cycles-pp.ktime_get
>>>        0.16 ±  9%      +0.1        0.23 ±  7%  perf-profile.children.cycles-pp.menu_select
>>>        0.21 ± 11%      +0.1        0.28 ± 16%  perf-profile.children.cycles-pp.io_serial_out
>>>        0.15 ± 20%      +0.1        0.23 ±  6%  perf-profile.children.cycles-pp.delay_tsc
>>>        1.83 ± 14%      +0.5        2.38 ± 12%  perf-profile.children.cycles-pp.serial8250_console_putchar
>>>        1.85 ± 14%      +0.6        2.40 ± 12%  perf-profile.children.cycles-pp.wait_for_xmitr
>>>        1.99 ± 15%      +0.6        2.56 ± 16%  perf-profile.children.cycles-pp.__dynamic_pr_debug
>>>        1.82 ± 13%      +0.6        2.39 ± 17%  perf-profile.children.cycles-pp.create_basic_memory_bitmaps
>>>        2.13 ± 13%      +0.6        2.74 ± 13%  perf-profile.children.cycles-pp.console_unlock
>>>        2.03 ± 13%      +0.6        2.65 ± 12%  perf-profile.children.cycles-pp.uart_console_write
>>>        2.17 ± 15%      +0.6        2.79 ± 14%  perf-profile.children.cycles-pp.printk
>>>        2.17 ± 15%      +0.6        2.79 ± 14%  perf-profile.children.cycles-pp.vprintk_emit
>>>        2.05 ± 13%      +0.6        2.67 ± 12%  perf-profile.children.cycles-pp.serial8250_console_write
>>>        0.00            +3.7        3.73 ± 66%  perf-profile.children.cycles-pp.__cna_queued_spin_lock_slowpath
>>>        2.92 ±  6%      -1.5        1.44 ± 22%  perf-profile.self.cycles-pp.syscall_exit_to_user_mode
>>>        1.48 ±  5%      -0.7        0.74 ± 28%  perf-profile.self.cycles-pp.syscall_return_via_sysret
>>>        1.25 ±  6%      -0.6        0.62 ± 28%  perf-profile.self.cycles-pp.__entry_text_start
>>>        0.52 ± 18%      -0.4        0.13 ± 34%  perf-profile.self.cycles-pp.apparmor_file_open
>>>        0.50 ± 32%      -0.4        0.13 ± 54%  perf-profile.self.cycles-pp.rwsem_spin_on_owner
>>>        0.44 ±  5%      -0.3        0.13 ± 36%  perf-profile.self.cycles-pp.lockref_put_or_lock
>>>        0.38 ± 12%      -0.3        0.12 ± 18%  perf-profile.self.cycles-pp.apparmor_file_alloc_security
>>>        0.52 ±  4%      -0.3        0.27 ± 25%  perf-profile.self.cycles-pp._raw_spin_lock
>>>        0.28 ± 22%      -0.2        0.05 ± 72%  perf-profile.self.cycles-pp.aa_get_task_label
>>>        0.45 ±  7%      -0.2        0.24 ± 20%  perf-profile.self.cycles-pp.kmem_cache_alloc
>>>        0.39 ± 10%      -0.2        0.19 ± 36%  perf-profile.self.cycles-pp.__fget_files
>>>        0.28 ± 19%      -0.2        0.08 ± 11%  perf-profile.self.cycles-pp.apparmor_file_free_security
>>>        0.30 ±  8%      -0.2        0.11 ± 36%  perf-profile.self.cycles-pp.do_dentry_open
>>>        0.21 ± 21%      -0.2        0.04 ± 76%  perf-profile.self.cycles-pp.mutex_lock
>>>        0.20 ±  3%      -0.1        0.06 ± 71%  perf-profile.self.cycles-pp.lockref_get_not_dead
>>>        0.20 ± 11%      -0.1        0.06 ± 73%  perf-profile.self.cycles-pp.vm_unmapped_area
>>>        0.18 ±  8%      -0.1        0.06 ± 73%  perf-profile.self.cycles-pp.common_file_perm
>>>        0.17 ± 14%      -0.1        0.04 ± 76%  perf-profile.self.cycles-pp.kobject_put
>>>        0.27 ± 15%      -0.1        0.15 ± 25%  perf-profile.self.cycles-pp.__alloc_file
>>>        0.16 ± 39%      -0.1        0.05 ± 78%  perf-profile.self.cycles-pp.rwsem_down_write_slowpath
>>>        0.16 ± 11%      -0.1        0.05 ± 77%  perf-profile.self.cycles-pp.fput_many
>>>        0.14 ±  7%      -0.1        0.04 ± 71%  perf-profile.self.cycles-pp.__might_sleep
>>>        0.18 ±  5%      -0.1        0.09 ± 30%  perf-profile.self.cycles-pp.entry_SYSCALL_64_after_hwframe
>>>        0.14 ± 11%      -0.1        0.04 ± 71%  perf-profile.self.cycles-pp.do_sys_openat2
>>>        0.15 ±  5%      -0.1        0.06 ± 73%  perf-profile.self.cycles-pp.__fput
>>>        0.20 ±  5%      -0.1        0.11 ± 31%  perf-profile.self.cycles-pp._raw_spin_lock_irqsave
>>>        0.13 ± 14%      -0.1        0.04 ± 70%  perf-profile.self.cycles-pp.link_path_walk
>>>        0.18 ±  8%      -0.1        0.09 ± 18%  perf-profile.self.cycles-pp.___might_sleep
>>>        0.13 ± 14%      -0.1        0.04 ± 71%  perf-profile.self.cycles-pp.get_obj_cgroup_from_current
>>>        0.12 ±  6%      -0.1        0.04 ± 73%  perf-profile.self.cycles-pp.do_mmap
>>>        0.12 ± 10%      -0.1        0.05 ± 71%  perf-profile.self.cycles-pp.__x86_retpoline_rax
>>>        0.12 ± 24%      -0.1        0.05 ± 72%  perf-profile.self.cycles-pp.allocate_slab
>>>        0.12 ±  8%      -0.1        0.06 ± 19%  perf-profile.self.cycles-pp.__d_lookup_rcu
>>>        0.10 ± 13%      -0.1        0.04 ± 71%  perf-profile.self.cycles-pp.generic_permission
>>>        0.09 ±  8%      -0.1        0.04 ± 71%  perf-profile.self.cycles-pp.entry_SYSCALL_64_safe_stack
>>>        0.10 ±  4%      -0.0        0.05 ±  8%  perf-profile.self.cycles-pp.syscall_enter_from_user_mode
>>>        0.07 ± 17%      +0.0        0.11 ±  8%  perf-profile.self.cycles-pp.cpuidle_enter_state
>>>        0.10 ±  9%      +0.0        0.14 ± 10%  perf-profile.self.cycles-pp.menu_select
>>>        0.03 ±100%      +0.1        0.09 ± 36%  perf-profile.self.cycles-pp.ktime_get
>>>        0.15 ± 20%      +0.1        0.23 ±  6%  perf-profile.self.cycles-pp.delay_tsc
>>>        0.20 ± 16%      +0.1        0.28 ± 16%  perf-profile.self.cycles-pp.io_serial_out
>>>        0.00            +3.7        3.73 ± 66%  perf-profile.self.cycles-pp.__cna_queued_spin_lock_slowpath
>>>
>>>
>>>                                                                                                                stress-ng.zero.ops_per_sec
>>>                                                                                     400000 +------------------------------------------------------------------+
>>>           |.+.. .+..                     .+..+.  .+.+..       +..            |
>>>    350000 |-+  +    +    +.+..+.+.. .+..+    O +. O O  +.+..O+O  +.+..+.+..+.|
>>>           | O       : O  : O    O  +    O O                 +      O  O O    |
>>>    300000 |-+       :    :    O                                              |
>>>    250000 |-+        :  :                                                    |
>>>           |          :  :                                                    |
>>>    200000 |-+        :  :                                                    |
>>>           |          :  :                                                    |
>>>    150000 |-+        : :                                                     |
>>>    100000 |-+        : :                                                     |
>>>           |           ::                                                     |
>>>     50000 |-+         ::                                                     |
>>>           |           :                                                      |
>>>         0 +------------------------------------------------------------------+
>>>                                                                                                                                                                   [*] bisect-good sample
>>> [O] bisect-bad  sample
>>>
>>> ***************************************************************************************************
>>> lkp-csl-2ap2: 192 threads Intel(R) Xeon(R) Platinum 9242 CPU @ 2.30GHz with 192G memory
>>> =========================================================================================
>>> compiler/cpufreq_governor/disk/filesize/fs/iterations/kconfig/nr_threads/rootfs/sync_method/tbox_group/test_size/testcase/ucode:
>>>    gcc-9/performance/1BRD_48G/4M/btrfs/1x/x86_64-rhel-8.3/64t/debian-10.4-x86_64-20200603.cgz/NoSync/lkp-csl-2ap2/24G/fsmark/0x5003003
>>>
>>> commit:
>>>    cb45bab007 ("locking/qspinlock: Refactor the qspinlock slow path")
>>>    0e8d8f4f12 ("locking/qspinlock: Introduce CNA into the slow path of qspinlock")
>>>
>>> cb45bab007ff0cfc 0e8d8f4f1214cfbac219d6917b5
>>> ---------------- ---------------------------
>>>           %stddev     %change         %stddev
>>>               \          |                \
>>>        1180 ±  2%    +213.9%       3706 ± 11%  fsmark.files_per_sec
>>>      773.00 ±  4%     -30.2%     539.75 ±  7%  fsmark.time.percent_of_cpu_this_job_got
>>>      137.19 ±  3%     -28.5%      98.13 ±  8%  fsmark.time.system_time
>>>       48917 ±  4%      -9.0%      44537 ±  8%  meminfo.AnonHugePages
>>>        4.83 ±  4%      -1.0        3.83 ±  5%  mpstat.cpu.all.sys%
>>>       93.48            +1.1%      94.51        iostat.cpu.idle
>>>        5.76 ±  2%     -16.9%       4.79 ±  6%  iostat.cpu.system
>>>       93.00            +1.1%      94.00        vmstat.cpu.id
>>>       20606 ±  5%     +97.4%      40682 ± 34%  vmstat.system.cs
>>>     4237881 ± 29%    +321.3%   17854499 ±106%  cpuidle.C1.time
>>>       44169 ± 32%    +396.5%     219318 ±113%  cpuidle.C1.usage
>>>       72208 ±  6%    +191.8%     210719 ± 39%  cpuidle.POLL.time
>>>       36708 ±  9%    +306.0%     149034 ± 48%  cpuidle.POLL.usage
>>>        2157 ±  2%     +13.8%       2454 ±  6%  slabinfo.biovec-max.active_objs
>>>        2271 ±  2%     +12.8%       2561 ±  5%  slabinfo.biovec-max.num_objs
>>>        6624           +13.4%       7512 ±  5%  slabinfo.btrfs_delayed_node.active_objs
>>>        6768           +12.1%       7589 ±  5%  slabinfo.btrfs_delayed_node.num_objs
>>>       47815 ± 86%     -89.2%       5149 ±  6%  sched_debug.cfs_rq:/.load.stddev
>>>      142.69 ± 15%     +66.8%     237.99 ± 13%  sched_debug.cfs_rq:/.load_avg.avg
>>>        1598 ± 30%     +92.4%       3075 ± 20%  sched_debug.cfs_rq:/.load_avg.max
>>>      383.38 ± 21%     +90.4%     730.10 ± 13%  sched_debug.cfs_rq:/.load_avg.stddev
>>>       37712 ± 16%     -27.1%      27508 ± 16%  sched_debug.cfs_rq:/.min_vruntime.avg
>>>       58.89 ± 49%     +57.8%      92.96 ±  5%  sched_debug.cfs_rq:/.util_est_enqueued.stddev
>>>        1901 ±  4%      -8.3%       1743 ±  3%  proc-vmstat.nr_active_anon
>>>     5055884            +3.2%    5219972        proc-vmstat.nr_file_pages
>>>     4788961 ±  2%      +3.4%    4953360        proc-vmstat.nr_inactive_file
>>>       36576 ±  2%      -4.1%      35069        proc-vmstat.nr_kernel_stack
>>>        9756            -1.6%       9601        proc-vmstat.nr_mapped
>>>        5205 ±  2%      -5.1%       4939        proc-vmstat.nr_shmem
>>>        1901 ±  4%      -8.3%       1743 ±  3%  proc-vmstat.nr_zone_active_anon
>>>     4788961 ±  2%      +3.4%    4953360        proc-vmstat.nr_zone_inactive_file
>>>        4864 ±  5%      -8.2%       4465        proc-vmstat.pgactivate
>>>   2.507e+09            -8.7%   2.29e+09        perf-stat.i.branch-instructions
>>>    71942409            -5.3%   68125925 ±  2%  perf-stat.i.cache-misses
>>>   4.372e+10           -18.5%  3.565e+10 ±  3%  perf-stat.i.cpu-cycles
>>>   1.231e+10            -7.5%  1.138e+10        perf-stat.i.instructions
>>>        2515 ±  8%     -11.4%       2228 ±  5%  perf-stat.i.instructions-per-iTLB-miss
>>>        0.23           -17.3%       0.19 ±  3%  perf-stat.i.metric.GHz
>>>        0.71 ± 19%     +37.6%       0.97 ± 17%  perf-stat.i.metric.K/sec
>>>       35.43            -6.2%      33.24 ±  2%  perf-stat.i.metric.M/sec
>>>        3.58            -8.5%       3.28        perf-stat.overall.cpi
>>>      612.62            -9.0%     557.36 ±  3%  perf-stat.overall.cycles-between-cache-misses
>>>        2610 ±  6%     -11.9%       2299 ±  4%  perf-stat.overall.instructions-per-iTLB-miss
>>>        0.28            +9.3%       0.31        perf-stat.overall.ipc
>>>    2.39e+09 ±  2%      -6.3%  2.239e+09        perf-stat.ps.branch-instructions
>>>    68534728 ±  2%      -4.6%   65359103 ±  2%  perf-stat.ps.cache-misses
>>>   4.199e+10 ±  2%     -13.3%  3.641e+10        perf-stat.ps.cpu-cycles
>>>   1.173e+10 ±  2%      -5.3%  1.112e+10        perf-stat.ps.instructions
>>>      226946 ± 18%     +44.8%     328685 ± 19%  numa-meminfo.node0.AnonPages
>>>      249710 ± 18%     +41.9%     354296 ± 20%  numa-meminfo.node0.AnonPages.max
>>>      233745 ± 17%     +42.0%     331984 ± 20%  numa-meminfo.node0.Inactive(anon)
>>>        8680 ± 73%    +135.9%      20474 ± 28%  numa-meminfo.node0.PageTables
>>>      250653 ±  2%      +6.8%     267784 ±  5%  numa-meminfo.node0.Unevictable
>>>     2890241 ±  9%     -58.3%    1206212 ± 10%  numa-meminfo.node0.Writeback
>>>      338312 ± 83%    +111.9%     716979 ± 64%  numa-meminfo.node1.Dirty
>>>        8444 ± 82%    +169.8%      22785 ± 37%  numa-meminfo.node2.AnonHugePages
>>>      236489 ± 16%    +142.5%     573382 ± 36%  numa-meminfo.node2.Dirty
>>>     2484617 ± 13%    +117.6%    5406277 ± 50%  numa-meminfo.node2.FilePages
>>>     2459002 ± 19%    +115.1%    5290461 ± 51%  numa-meminfo.node2.Inactive
>>>     2216704 ± 16%    +131.7%    5135596 ± 53%  numa-meminfo.node2.Inactive(file)
>>>      871702 ± 16%     +52.8%    1331684 ±  6%  numa-meminfo.node2.Writeback
>>>        3888 ± 16%     -24.5%       2937 ±  6%  numa-meminfo.node3.Active(anon)
>>>       57796 ± 19%     -49.7%      29078 ± 34%  numa-meminfo.node3.KReclaimable
>>>       57796 ± 19%     -49.7%      29078 ± 34%  numa-meminfo.node3.SReclaimable
>>>      130240 ±  7%     -30.1%      91037 ± 15%  numa-meminfo.node3.Slab
>>>       56616 ± 18%     +45.2%      82199 ± 19%  numa-vmstat.node0.nr_anon_pages
>>>       58314 ± 17%     +42.4%      83024 ± 20%  numa-vmstat.node0.nr_inactive_anon
>>>        2172 ± 73%    +135.7%       5122 ± 28%  numa-vmstat.node0.nr_page_table_pages
>>>       62662 ±  2%      +6.8%      66945 ±  5%  numa-vmstat.node0.nr_unevictable
>>>      738434 ±  9%     -59.3%     300391 ± 10%  numa-vmstat.node0.nr_writeback
>>>       58312 ± 17%     +42.4%      83023 ± 20%  numa-vmstat.node0.nr_zone_inactive_anon
>>>       62662 ±  2%      +6.8%      66945 ±  5%  numa-vmstat.node0.nr_zone_unevictable
>>>      926442 ±  6%     -53.1%     434282 ± 12%  numa-vmstat.node0.nr_zone_write_pending
>>>       84324 ± 81%    +112.0%     178764 ± 64%  numa-vmstat.node1.nr_dirty
>>>      427046 ± 84%     +74.8%     746464 ± 46%  numa-vmstat.node1.nr_zone_write_pending
>>>      627042 ± 18%    +129.0%    1435752 ± 55%  numa-vmstat.node2.nr_dirtied
>>>       59963 ± 18%    +137.8%     142585 ± 36%  numa-vmstat.node2.nr_dirty
>>>      634106 ± 15%    +112.1%    1345185 ± 50%  numa-vmstat.node2.nr_file_pages
>>>      567086 ± 17%    +125.3%    1277507 ± 53%  numa-vmstat.node2.nr_inactive_file
>>>      223407 ± 17%     +48.3%     331299 ±  7%  numa-vmstat.node2.nr_writeback
>>>      567087 ± 17%    +125.3%    1277515 ± 53%  numa-vmstat.node2.nr_zone_inactive_file
>>>      283372 ± 17%     +67.2%     473906 ±  8%  numa-vmstat.node2.nr_zone_write_pending
>>>      922.00 ± 19%     -20.0%     737.25 ±  6%  numa-vmstat.node3.nr_active_anon
>>>       14591 ± 20%     -50.1%       7276 ± 35%  numa-vmstat.node3.nr_slab_reclaimable
>>>      922.00 ± 19%     -20.0%     737.25 ±  6%  numa-vmstat.node3.nr_zone_active_anon
>>>      141466 ± 17%    +523.4%     881964 ± 85%  numa-vmstat.node3.numa_other
>>>      282981 ± 14%     +46.6%     414731 ± 12%  interrupts.CAL:Function_call_interrupts
>>>      770.00 ± 51%     -56.5%     335.00 ± 59%  interrupts.CPU0.NMI:Non-maskable_interrupts
>>>      770.00 ± 51%     -56.5%     335.00 ± 59%  interrupts.CPU0.PMI:Performance_monitoring_interrupts
>>>      346.25 ± 15%     -46.7%     184.50 ± 23%  interrupts.CPU110.NMI:Non-maskable_interrupts
>>>      346.25 ± 15%     -46.7%     184.50 ± 23%  interrupts.CPU110.PMI:Performance_monitoring_interrupts
>>>        1359 ± 14%     +31.7%       1789 ± 17%  interrupts.CPU128.CAL:Function_call_interrupts
>>>        1378 ± 14%   +2294.9%      33020 ±164%  interrupts.CPU153.CAL:Function_call_interrupts
>>>      214.50 ± 34%     +91.7%     411.25 ± 39%  interrupts.CPU16.NMI:Non-maskable_interrupts
>>>      214.50 ± 34%     +91.7%     411.25 ± 39%  interrupts.CPU16.PMI:Performance_monitoring_interrupts
>>>      403.50 ± 35%     -40.3%     241.00 ±  8%  interrupts.CPU169.NMI:Non-maskable_interrupts
>>>      403.50 ± 35%     -40.3%     241.00 ±  8%  interrupts.CPU169.PMI:Performance_monitoring_interrupts
>>>      439.00 ± 50%     -51.5%     212.75 ± 22%  interrupts.CPU170.NMI:Non-maskable_interrupts
>>>      439.00 ± 50%     -51.5%     212.75 ± 22%  interrupts.CPU170.PMI:Performance_monitoring_interrupts
>>>        1062 ±116%     -77.3%     240.75 ± 23%  interrupts.CPU174.NMI:Non-maskable_interrupts
>>>        1062 ±116%     -77.3%     240.75 ± 23%  interrupts.CPU174.PMI:Performance_monitoring_interrupts
>>>      291.25 ± 31%     -34.2%     191.75 ± 30%  interrupts.CPU2.NMI:Non-maskable_interrupts
>>>      291.25 ± 31%     -34.2%     191.75 ± 30%  interrupts.CPU2.PMI:Performance_monitoring_interrupts
>>>      342.00 ± 43%     -58.3%     142.75 ± 16%  interrupts.CPU22.NMI:Non-maskable_interrupts
>>>      342.00 ± 43%     -58.3%     142.75 ± 16%  interrupts.CPU22.PMI:Performance_monitoring_interrupts
>>>      241.25 ± 22%     -42.6%     138.50 ± 22%  interrupts.CPU23.NMI:Non-maskable_interrupts
>>>      241.25 ± 22%     -42.6%     138.50 ± 22%  interrupts.CPU23.PMI:Performance_monitoring_interrupts
>>>        1365 ± 14%     +49.5%       2041 ± 24%  interrupts.CPU35.CAL:Function_call_interrupts
>>>      222.00 ± 94%    +175.1%     610.75 ± 51%  interrupts.CPU63.NMI:Non-maskable_interrupts
>>>      222.00 ± 94%    +175.1%     610.75 ± 51%  interrupts.CPU63.PMI:Performance_monitoring_interrupts
>>>      146.75 ± 39%    +104.6%     300.25 ± 32%  interrupts.CPU64.NMI:Non-maskable_interrupts
>>>      146.75 ± 39%    +104.6%     300.25 ± 32%  interrupts.CPU64.PMI:Performance_monitoring_interrupts
>>>      557.00 ± 88%     -61.3%     215.75 ± 21%  interrupts.CPU74.NMI:Non-maskable_interrupts
>>>      557.00 ± 88%     -61.3%     215.75 ± 21%  interrupts.CPU74.PMI:Performance_monitoring_interrupts
>>>      319.25 ± 22%     -43.9%     179.25 ± 36%  interrupts.CPU75.NMI:Non-maskable_interrupts
>>>      319.25 ± 22%     -43.9%     179.25 ± 36%  interrupts.CPU75.PMI:Performance_monitoring_interrupts
>>>        1343 ± 15%     +23.0%       1652 ±  4%  interrupts.CPU80.CAL:Function_call_interrupts
>>>        1095 ± 66%     -76.5%     257.00 ± 17%  interrupts.CPU96.NMI:Non-maskable_interrupts
>>>        1095 ± 66%     -76.5%     257.00 ± 17%  interrupts.CPU96.PMI:Performance_monitoring_interrupts
>>>      817.25 ± 12%     +62.6%       1329 ± 17%  interrupts.RES:Rescheduling_interrupts
>>>       12.39 ± 23%      -8.2        4.22 ±  6%  perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe
>>>       12.35 ± 24%      -8.1        4.21 ±  6%  perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe
>>>       12.12 ± 24%      -8.0        4.08 ±  6%  perf-profile.calltrace.cycles-pp.ksys_write.do_syscall_64.entry_SYSCALL_64_after_hwframe
>>>       12.11 ± 24%      -8.0        4.07 ±  6%  perf-profile.calltrace.cycles-pp.vfs_write.ksys_write.do_syscall_64.entry_SYSCALL_64_after_hwframe
>>>       12.10 ± 24%      -8.0        4.06 ±  6%  perf-profile.calltrace.cycles-pp.btrfs_file_write_iter.new_sync_write.vfs_write.ksys_write.do_syscall_64
>>>       12.07 ± 24%      -8.0        4.04 ±  6%  perf-profile.calltrace.cycles-pp.btrfs_buffered_write.btrfs_file_write_iter.new_sync_write.vfs_write.ksys_write
>>>       12.28 ± 24%      -7.5        4.83 ±  5%  perf-profile.calltrace.cycles-pp.new_sync_write.vfs_write.ksys_write.do_syscall_64.entry_SYSCALL_64_after_hwframe
>>>        6.07 ± 31%      -6.1        0.00        perf-profile.calltrace.cycles-pp.native_queued_spin_lock_slowpath._raw_spin_lock.__reserve_bytes.btrfs_reserve_metadata_bytes.btrfs_delalloc_reserve_metadata
>>>        6.36 ± 30%      -4.8        1.55 ± 11%  perf-profile.calltrace.cycles-pp.btrfs_delalloc_reserve_metadata.btrfs_buffered_write.btrfs_file_write_iter.new_sync_write.vfs_write
>>>        6.31 ± 30%      -4.8        1.51 ± 12%  perf-profile.calltrace.cycles-pp.__reserve_bytes.btrfs_reserve_metadata_bytes.btrfs_delalloc_reserve_metadata.btrfs_buffered_write.btrfs_file_write_iter
>>>        6.31 ± 30%      -4.8        1.51 ± 12%  perf-profile.calltrace.cycles-pp.btrfs_reserve_metadata_bytes.btrfs_delalloc_reserve_metadata.btrfs_buffered_write.btrfs_file_write_iter.new_sync_write
>>>        6.13 ± 31%      -4.8        1.38 ± 14%  perf-profile.calltrace.cycles-pp._raw_spin_lock.__reserve_bytes.btrfs_reserve_metadata_bytes.btrfs_delalloc_reserve_metadata.btrfs_buffered_write
>>>        4.39 ± 18%      -3.2        1.21 ± 15%  perf-profile.calltrace.cycles-pp.btrfs_inode_rsv_release.btrfs_buffered_write.btrfs_file_write_iter.new_sync_write.vfs_write
>>>        4.38 ± 18%      -3.2        1.20 ± 15%  perf-profile.calltrace.cycles-pp.btrfs_block_rsv_release.btrfs_inode_rsv_release.btrfs_buffered_write.btrfs_file_write_iter.new_sync_write
>>>        4.21 ± 19%      -3.1        1.09 ± 20%  perf-profile.calltrace.cycles-pp._raw_spin_lock.btrfs_block_rsv_release.btrfs_inode_rsv_release.btrfs_buffered_write.btrfs_file_write_iter
>>>        0.13 ±173%      +0.5        0.63 ±  6%  perf-profile.calltrace.cycles-pp.tick_nohz_irq_exit.sysvec_apic_timer_interrupt.asm_sysvec_apic_timer_interrupt.cpuidle_enter_state.cpuidle_enter
>>>        4.51 ±  6%      +0.5        5.04 ±  3%  perf-profile.calltrace.cycles-pp.asm_call_sysvec_on_stack.sysvec_apic_timer_interrupt.asm_sysvec_apic_timer_interrupt.cpuidle_enter_state.cpuidle_enter
>>>        4.48 ±  5%      +0.5        5.02 ±  3%  perf-profile.calltrace.cycles-pp.__sysvec_apic_timer_interrupt.asm_call_sysvec_on_stack.sysvec_apic_timer_interrupt.asm_sysvec_apic_timer_interrupt.cpuidle_enter_state
>>>        4.37 ±  6%      +0.6        4.93 ±  3%  perf-profile.calltrace.cycles-pp.hrtimer_interrupt.__sysvec_apic_timer_interrupt.asm_call_sysvec_on_stack.sysvec_apic_timer_interrupt.asm_sysvec_apic_timer_interrupt
>>>        0.18 ±173%      +0.6        0.75 ± 27%  perf-profile.calltrace.cycles-pp.console_unlock.vprintk_emit.devkmsg_emit.devkmsg_write.cold.new_sync_write
>>>        0.18 ±173%      +0.6        0.76 ± 27%  perf-profile.calltrace.cycles-pp.vprintk_emit.devkmsg_emit.devkmsg_write.cold.new_sync_write.vfs_write
>>>        0.18 ±173%      +0.6        0.76 ± 27%  perf-profile.calltrace.cycles-pp.devkmsg_write.cold.new_sync_write.vfs_write.ksys_write.do_syscall_64
>>>        0.18 ±173%      +0.6        0.76 ± 27%  perf-profile.calltrace.cycles-pp.devkmsg_emit.devkmsg_write.cold.new_sync_write.vfs_write.ksys_write
>>>        0.18 ±173%      +0.6        0.77 ± 27%  perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.write
>>>        0.18 ±173%      +0.6        0.77 ± 27%  perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.write
>>>        0.18 ±173%      +0.6        0.77 ± 27%  perf-profile.calltrace.cycles-pp.ksys_write.do_syscall_64.entry_SYSCALL_64_after_hwframe.write
>>>        0.18 ±173%      +0.6        0.77 ± 27%  perf-profile.calltrace.cycles-pp.vfs_write.ksys_write.do_syscall_64.entry_SYSCALL_64_after_hwframe.write
>>>        0.18 ±173%      +0.6        0.77 ± 27%  perf-profile.calltrace.cycles-pp.write
>>>        0.00            +0.6        0.62 ±  7%  perf-profile.calltrace.cycles-pp.ktime_get.tick_nohz_irq_exit.sysvec_apic_timer_interrupt.asm_sysvec_apic_timer_interrupt.cpuidle_enter_state
>>>        0.00            +1.1        1.06 ± 21%  perf-profile.calltrace.cycles-pp.__cna_queued_spin_lock_slowpath._raw_spin_lock.btrfs_block_rsv_release.btrfs_inode_rsv_release.btrfs_buffered_write
>>>        7.48 ±  5%      +1.2        8.68 ±  2%  perf-profile.calltrace.cycles-pp.sysvec_apic_timer_interrupt.asm_sysvec_apic_timer_interrupt.cpuidle_enter_state.cpuidle_enter.do_idle
>>>        0.00            +1.3        1.33 ± 15%  perf-profile.calltrace.cycles-pp.__cna_queued_spin_lock_slowpath._raw_spin_lock.__reserve_bytes.btrfs_reserve_metadata_bytes.btrfs_delalloc_reserve_metadata
>>>       11.79 ±  8%      +2.8       14.63 ±  8%  perf-profile.calltrace.cycles-pp.asm_sysvec_apic_timer_interrupt.cpuidle_enter_state.cpuidle_enter.do_idle.cpu_startup_entry
>>>        9.11 ± 14%      +4.4       13.51 ± 18%  perf-profile.calltrace.cycles-pp.menu_select.do_idle.cpu_startup_entry.start_secondary.secondary_startup_64_no_verify
>>>       10.58 ± 26%     -10.6        0.00        perf-profile.children.cycles-pp.native_queued_spin_lock_slowpath
>>>       12.10 ± 24%      -8.0        4.06 ±  6%  perf-profile.children.cycles-pp.btrfs_file_write_iter
>>>       12.07 ± 24%      -8.0        4.04 ±  6%  perf-profile.children.cycles-pp.btrfs_buffered_write
>>>       14.00 ± 18%      -7.9        6.09 ±  7%  perf-profile.children.cycles-pp._raw_spin_lock
>>>       12.64 ± 23%      -7.8        4.85 ±  5%  perf-profile.children.cycles-pp.ksys_write
>>>       12.62 ± 24%      -7.8        4.83 ±  5%  perf-profile.children.cycles-pp.new_sync_write
>>>       12.63 ± 23%      -7.8        4.85 ±  5%  perf-profile.children.cycles-pp.vfs_write
>>>       13.17 ± 23%      -7.6        5.60 ±  6%  perf-profile.children.cycles-pp.do_syscall_64
>>>       13.24 ± 23%      -7.3        5.92 ±  6%  perf-profile.children.cycles-pp.entry_SYSCALL_64_after_hwframe
>>>        6.38 ± 30%      -4.8        1.55 ± 12%  perf-profile.children.cycles-pp.__reserve_bytes
>>>        6.33 ± 30%      -4.8        1.52 ± 12%  perf-profile.children.cycles-pp.btrfs_reserve_metadata_bytes
>>>        6.36 ± 30%      -4.8        1.55 ± 11%  perf-profile.children.cycles-pp.btrfs_delalloc_reserve_metadata
>>>        4.49 ± 18%      -3.3        1.23 ± 15%  perf-profile.children.cycles-pp.btrfs_block_rsv_release
>>>        4.39 ± 18%      -3.2        1.21 ± 15%  perf-profile.children.cycles-pp.btrfs_inode_rsv_release
>>>        0.16 ± 77%      -0.1        0.06 ±  7%  perf-profile.children.cycles-pp.do_filp_open
>>>        0.16 ± 77%      -0.1        0.06 ±  7%  perf-profile.children.cycles-pp.path_openat
>>>        0.16 ± 75%      -0.1        0.06 ±  6%  perf-profile.children.cycles-pp.do_sys_open
>>>        0.16 ± 75%      -0.1        0.06 ±  6%  perf-profile.children.cycles-pp.do_sys_openat2
>>>        0.21 ± 32%      -0.1        0.13 ± 29%  perf-profile.children.cycles-pp.rcu_sched_clock_irq
>>>        0.22 ±  9%      -0.1        0.16 ±  6%  perf-profile.children.cycles-pp.update_blocked_averages
>>>        0.10 ± 29%      -0.0        0.05 ± 70%  perf-profile.children.cycles-pp.update_ts_time_stats
>>>        0.07 ± 12%      -0.0        0.04 ± 58%  perf-profile.children.cycles-pp.__pagevec_lru_add_fn
>>>        0.16 ±  9%      +0.0        0.20 ± 12%  perf-profile.children.cycles-pp.brd_lookup_page
>>>        0.13 ± 11%      +0.0        0.17 ± 18%  perf-profile.children.cycles-pp.__radix_tree_lookup
>>>        0.14 ±  7%      +0.1        0.20 ±  9%  perf-profile.children.cycles-pp.__intel_pmu_enable_all
>>>        0.48 ±  8%      +0.1        0.62 ±  8%  perf-profile.children.cycles-pp._raw_spin_lock_irqsave
>>>        0.51 ± 29%      +0.3        0.77 ± 27%  perf-profile.children.cycles-pp.write
>>>        0.39 ± 32%      +0.3        0.65 ±  6%  perf-profile.children.cycles-pp.tick_nohz_irq_exit
>>>        0.50 ± 30%      +0.3        0.76 ± 27%  perf-profile.children.cycles-pp.devkmsg_write.cold
>>>        0.50 ± 30%      +0.3        0.76 ± 27%  perf-profile.children.cycles-pp.devkmsg_emit
>>>        0.03 ±173%      +0.3        0.29 ±108%  perf-profile.children.cycles-pp.osq_lock
>>>        0.03 ±173%      +0.3        0.29 ±107%  perf-profile.children.cycles-pp.__mutex_lock
>>>        0.11 ±130%      +0.3        0.45 ± 68%  perf-profile.children.cycles-pp.__do_sys_finit_module
>>>        0.11 ±130%      +0.3        0.45 ± 68%  perf-profile.children.cycles-pp.load_module
>>>        0.11 ±130%      +0.3        0.45 ± 67%  perf-profile.children.cycles-pp.syscall
>>>        0.14 ± 88%      +0.4        0.54 ± 53%  perf-profile.children.cycles-pp.wb_workfn
>>>        0.14 ± 88%      +0.4        0.54 ± 53%  perf-profile.children.cycles-pp.wb_writeback
>>>        0.14 ± 88%      +0.4        0.54 ± 53%  perf-profile.children.cycles-pp.writeback_sb_inodes
>>>        0.12 ±105%      +0.4        0.53 ± 54%  perf-profile.children.cycles-pp.__writeback_single_inode
>>>        4.76 ±  5%      +0.5        5.21 ±  3%  perf-profile.children.cycles-pp.__sysvec_apic_timer_interrupt
>>>        4.65 ±  5%      +0.5        5.13 ±  2%  perf-profile.children.cycles-pp.hrtimer_interrupt
>>>        6.15 ±  5%      +0.5        6.65 ±  2%  perf-profile.children.cycles-pp.asm_call_sysvec_on_stack
>>>        2.06 ± 28%      +1.0        3.05 ±  9%  perf-profile.children.cycles-pp.ktime_get
>>>        7.80 ±  4%      +1.1        8.94 ±  2%  perf-profile.children.cycles-pp.sysvec_apic_timer_interrupt
>>>       10.10 ±  6%      +2.0       12.05 ±  6%  perf-profile.children.cycles-pp.asm_sysvec_apic_timer_interrupt
>>>        0.00            +2.6        2.56 ± 19%  perf-profile.children.cycles-pp.__cna_queued_spin_lock_slowpath
>>>        9.16 ± 14%      +4.4       13.57 ± 18%  perf-profile.children.cycles-pp.menu_select
>>>       10.50 ± 26%     -10.5        0.00        perf-profile.self.cycles-pp.native_queued_spin_lock_slowpath
>>>        0.18 ± 30%      -0.1        0.11 ± 27%  perf-profile.self.cycles-pp.rcu_sched_clock_irq
>>>        0.10 ± 11%      +0.0        0.13 ± 10%  perf-profile.self.cycles-pp.__extent_writepage
>>>        0.14 ±  7%      +0.1        0.20 ±  9%  perf-profile.self.cycles-pp.__intel_pmu_enable_all
>>>        0.08 ± 10%      +0.1        0.16 ± 22%  perf-profile.self.cycles-pp.end_page_writeback
>>>        0.03 ±173%      +0.2        0.28 ±105%  perf-profile.self.cycles-pp.osq_lock
>>>        1.77 ± 32%      +1.0        2.75 ± 10%  perf-profile.self.cycles-pp.ktime_get
>>>        0.00            +2.5        2.51 ± 19%  perf-profile.self.cycles-pp.__cna_queued_spin_lock_slowpath
>>>        7.68 ± 15%      +4.1       11.79 ± 21%  perf-profile.self.cycles-pp.menu_select
>>>
>>>
>>>
>>> ***************************************************************************************************
>>> lkp-csl-2ap3: 192 threads Intel(R) Xeon(R) Platinum 9242 CPU @ 2.30GHz with 192G memory
>>> =========================================================================================
>>> compiler/cpufreq_governor/kconfig/nr_task/rootfs/runtime/tbox_group/test/testcase/ucode:
>>>    gcc-9/performance/x86_64-rhel-8.3/100%/debian-10.4-x86_64-20200603.cgz/300s/lkp-csl-2ap3/new_fserver/reaim/0x5003003
>>>
>>> commit:
>>>    cb45bab007 ("locking/qspinlock: Refactor the qspinlock slow path")
>>>    0e8d8f4f12 ("locking/qspinlock: Introduce CNA into the slow path of qspinlock")
>>>
>>> cb45bab007ff0cfc 0e8d8f4f1214cfbac219d6917b5
>>> ---------------- ---------------------------
>>>         fail:runs  %reproduction    fail:runs
>>>             |             |             |
>>>            1:4          -25%            :4     last_state.is_incomplete_run
>>>            1:4          -25%            :4     last_state.wget_initrd
>>>             :4            8%           0:4     perf-profile.children.cycles-pp.error_entry
>>>           %stddev     %change         %stddev
>>>               \          |                \
>>>      556.93           -64.1%     200.03        reaim.child_systime
>>>       49.95            +3.6%      51.77        reaim.child_utime
>>>      318145           +96.1%     623983        reaim.jobs_per_min
>>>        1657           +96.1%       3249        reaim.jobs_per_min_child
>>>       92.26            -7.8%      85.11        reaim.jti
>>>      327019          +100.4%     655467        reaim.max_jobs_per_min
>>>        3.55           -49.0%       1.81        reaim.parent_time
>>>        7.19 ±  2%    +100.0%      14.38 ±  4%  reaim.std_dev_percent
>>>        0.24 ±  2%      -7.3%       0.22 ±  2%  reaim.std_dev_time
>>>      462973           +17.9%     545826        reaim.time.involuntary_context_switches
>>>       38288           +98.5%      75996 ±  2%  reaim.time.major_page_faults
>>>   1.041e+08           +44.8%  1.509e+08        reaim.time.minor_page_faults
>>>       10863           -39.7%       6549        reaim.time.percent_of_cpu_this_job_got
>>>       30257           -47.9%      15754        reaim.time.system_time
>>>        2714           +50.2%       4077        reaim.time.user_time
>>>     3000456 ±  2%     +38.9%    4167220 ±  2%  reaim.time.voluntary_context_switches
>>>     1043200           +44.9%    1512000        reaim.workload
>>>       29843           +40.0%      41767        uptime.idle
>>>    20127118 ± 12%     +48.7%   29930674 ±  4%  cpuidle.C1.time
>>>      284217 ±  9%     +66.1%     472177 ±  2%  cpuidle.C1.usage
>>>      308815 ± 13%     +26.0%     389201 ± 16%  cpuidle.POLL.usage
>>>       41.44           +20.8       62.28        mpstat.cpu.all.idle%
>>>        0.69            +1.3        1.98 ±  2%  mpstat.cpu.all.soft%
>>>       51.88           -24.6       27.26        mpstat.cpu.all.sys%
>>>        4.60            +2.3        6.91        mpstat.cpu.all.usr%
>>>       41.33           +50.0%      62.00        vmstat.cpu.id
>>>       53.00           -43.4%      30.00        vmstat.cpu.sy
>>>      110.00           -25.9%      81.50 ±  2%  vmstat.procs.r
>>>       25559 ±  2%     +46.2%      37370        vmstat.system.cs
>>>      269427 ±  4%     -12.8%     234969        meminfo.Active
>>>      269427 ±  4%     -12.8%     234969        meminfo.Active(anon)
>>>       32953            +9.7%      36144 ±  2%  meminfo.KernelStack
>>>       19202           +63.2%      31338        meminfo.PageTables
>>>      483410           +11.8%     540459        meminfo.SUnreclaim
>>>      304770 ±  4%     -12.8%     265628        meminfo.Shmem
>>>      620741           +10.5%     685878        meminfo.Slab
>>>    18788579           +43.8%   27017272 ±  4%  numa-numastat.node0.local_node
>>>    18827783           +43.8%   27074532 ±  4%  numa-numastat.node0.numa_hit
>>>    19046024           +45.8%   27771991 ±  2%  numa-numastat.node1.local_node
>>>    19126483           +45.6%   27844754 ±  2%  numa-numastat.node1.numa_hit
>>>    18849870           +44.4%   27228340 ±  2%  numa-numastat.node2.local_node
>>>    18926181           +44.1%   27271719 ±  2%  numa-numastat.node2.numa_hit
>>>    18622314           +46.2%   27216882 ±  3%  numa-numastat.node3.local_node
>>>    18686267           +46.1%   27303516 ±  3%  numa-numastat.node3.numa_hit
>>>       42977 ± 81%     -79.9%       8629 ± 59%  numa-meminfo.node0.Active
>>>       42977 ± 81%     -79.9%       8629 ± 59%  numa-meminfo.node0.Active(anon)
>>>        7040           +24.4%       8759 ±  8%  numa-meminfo.node1.KernelStack
>>>        2790 ± 11%    +140.4%       6708 ± 26%  numa-meminfo.node1.PageTables
>>>      103698           +19.3%     123717 ±  9%  numa-meminfo.node1.SUnreclaim
>>>      129487 ±  2%     +18.0%     152790 ±  7%  numa-meminfo.node1.Slab
>>>       32541 ± 26%     +46.6%      47695 ± 24%  numa-meminfo.node2.KReclaimable
>>>       32541 ± 26%     +46.6%      47695 ± 24%  numa-meminfo.node2.SReclaimable
>>>      150472 ± 20%     +21.5%     182840 ± 15%  numa-meminfo.node2.Slab
>>>      167884 ± 33%     +71.3%     287506 ± 45%  numa-meminfo.node3.AnonPages.max
>>>        2974 ± 11%    +145.0%       7288 ± 28%  numa-meminfo.node3.PageTables
>>>      109568 ±  4%     +13.7%     124544 ± 12%  numa-meminfo.node3.SUnreclaim
>>>       10666 ± 81%     -79.8%       2152 ± 59%  numa-vmstat.node0.nr_active_anon
>>>       10665 ± 81%     -79.8%       2152 ± 59%  numa-vmstat.node0.nr_zone_active_anon
>>>     9854862           +41.8%   13975810 ±  2%  numa-vmstat.node0.numa_hit
>>>     9773079           +41.9%   13870000 ±  2%  numa-vmstat.node0.numa_local
>>>        7029           +25.1%       8796 ±  8%  numa-vmstat.node1.nr_kernel_stack
>>>      678.67 ±  8%    +152.6%       1714 ± 26%  numa-vmstat.node1.nr_page_table_pages
>>>       25872           +20.1%      31062 ±  9%  numa-vmstat.node1.nr_slab_unreclaimable
>>>     9970820           +42.6%   14214395        numa-vmstat.node1.numa_hit
>>>     9810384           +43.4%   14063434        numa-vmstat.node1.numa_local
>>>        8132 ± 26%     +46.4%      11904 ± 24%  numa-vmstat.node2.nr_slab_reclaimable
>>>     9969168           +43.4%   14294976 ±  2%  numa-vmstat.node2.numa_hit
>>>     9812426           +44.6%   14186949 ±  3%  numa-vmstat.node2.numa_local
>>>      726.67 ± 12%    +158.7%       1879 ± 26%  numa-vmstat.node3.nr_page_table_pages
>>>       27327 ±  4%     +13.9%      31114 ± 12%  numa-vmstat.node3.nr_slab_unreclaimable
>>>     9719726           +44.0%   13991557 ±  2%  numa-vmstat.node3.numa_hit
>>>     9615461           +44.1%   13853485 ±  2%  numa-vmstat.node3.numa_local
>>>      104269 ± 49%     +32.4%     138078 ± 20%  numa-vmstat.node3.numa_other
>>>       67226 ±  4%     -12.4%      58905        proc-vmstat.nr_active_anon
>>>      316527            -3.1%     306823        proc-vmstat.nr_file_pages
>>>       92663            -2.9%      89935        proc-vmstat.nr_inactive_anon
>>>       33012            +9.3%      36075 ±  2%  proc-vmstat.nr_kernel_stack
>>>        4865           +59.3%       7749 ±  4%  proc-vmstat.nr_page_table_pages
>>>       76214 ±  4%     -12.7%      66510        proc-vmstat.nr_shmem
>>>       34346            +5.8%      36334        proc-vmstat.nr_slab_reclaimable
>>>      120937           +11.5%     134852        proc-vmstat.nr_slab_unreclaimable
>>>       67226 ±  4%     -12.4%      58905        proc-vmstat.nr_zone_active_anon
>>>       92663            -2.9%      89935        proc-vmstat.nr_zone_inactive_anon
>>>      152011 ±  9%     -72.0%      42500 ± 30%  proc-vmstat.numa_hint_faults
>>>    75604538           +44.9%  1.095e+08        proc-vmstat.numa_hit
>>>    75344552           +45.0%  1.093e+08        proc-vmstat.numa_local
>>>      129725 ± 13%     -78.5%      27855 ± 33%  proc-vmstat.numa_pages_migrated
>>>      440421 ±  8%     -27.0%     321634 ± 22%  proc-vmstat.numa_pte_updates
>>>      187992 ±  4%      +5.4%     198156        proc-vmstat.pgactivate
>>>    83223343           +45.7%  1.213e+08        proc-vmstat.pgalloc_normal
>>>   1.056e+08           +44.3%  1.524e+08        proc-vmstat.pgfault
>>>    83115624           +45.8%  1.212e+08        proc-vmstat.pgfree
>>>      129725 ± 13%     -78.5%      27855 ± 33%  proc-vmstat.pgmigrate_success
>>>    10628125           +46.1%   15525117        proc-vmstat.pgreuse
>>>       13330 ± 54%    +717.2%     108933 ± 77%  sched_debug.cfs_rq:/.MIN_vruntime.avg
>>>      184231 ± 54%    +190.7%     535537 ± 48%  sched_debug.cfs_rq:/.MIN_vruntime.stddev
>>>        6040 ±  6%    +338.5%      26486 ± 51%  sched_debug.cfs_rq:/.load.avg
>>>      284313 ± 37%    +159.7%     738429 ± 14%  sched_debug.cfs_rq:/.load.max
>>>        1180 ± 35%    -100.0%       0.00        sched_debug.cfs_rq:/.load.min
>>>       22088 ± 35%    +398.6%     110141 ± 30%  sched_debug.cfs_rq:/.load.stddev
>>>       13330 ± 54%    +717.2%     108933 ± 77%  sched_debug.cfs_rq:/.max_vruntime.avg
>>>      184231 ± 54%    +190.7%     535537 ± 48%  sched_debug.cfs_rq:/.max_vruntime.stddev
>>>    16468615           -51.1%    8045534 ± 11%  sched_debug.cfs_rq:/.min_vruntime.avg
>>>    16715699           -50.1%    8341867 ± 11%  sched_debug.cfs_rq:/.min_vruntime.max
>>>    15817769           -51.3%    7697369 ± 11%  sched_debug.cfs_rq:/.min_vruntime.min
>>>        0.71 ± 11%     -47.8%       0.37 ± 14%  sched_debug.cfs_rq:/.nr_running.avg
>>>        1.17           +19.3%       1.39 ±  4%  sched_debug.cfs_rq:/.nr_running.max
>>>        0.22 ± 35%    -100.0%       0.00        sched_debug.cfs_rq:/.nr_running.min
>>>        0.14 ± 18%     +74.2%       0.25 ±  6%  sched_debug.cfs_rq:/.nr_running.stddev
>>>      694.83 ± 15%     -51.9%     334.48 ± 20%  sched_debug.cfs_rq:/.runnable_avg.avg
>>>      110.72 ± 51%     -90.4%      10.63 ±114%  sched_debug.cfs_rq:/.runnable_avg.min
>>>      155.35 ± 11%     +42.5%     221.31 ±  9%  sched_debug.cfs_rq:/.runnable_avg.stddev
>>>       64422 ± 94%    +295.3%     254650 ± 19%  sched_debug.cfs_rq:/.spread0.avg
>>>      311765 ± 14%     +76.7%     550961 ± 14%  sched_debug.cfs_rq:/.spread0.max
>>>     -587893           -84.1%     -93210        sched_debug.cfs_rq:/.spread0.min
>>>      685.90 ± 14%     -56.7%     297.04 ± 23%  sched_debug.cfs_rq:/.util_avg.avg
>>>      100.39 ± 59%     -90.9%       9.12 ±125%  sched_debug.cfs_rq:/.util_avg.min
>>>      153.19 ± 12%     +28.3%     196.53 ± 12%  sched_debug.cfs_rq:/.util_avg.stddev
>>>      398.32 ± 19%     -79.6%      81.24 ± 25%  sched_debug.cfs_rq:/.util_est_enqueued.avg
>>>      932.06 ±  8%     -26.5%     684.67 ± 16%  sched_debug.cfs_rq:/.util_est_enqueued.max
>>>        0.28 ± 28%    -100.0%       0.00        sched_debug.cfs_rq:/.util_est_enqueued.min
>>>      186.70 ± 17%     -31.0%     128.84 ± 18%  sched_debug.cfs_rq:/.util_est_enqueued.stddev
>>>       42223 ± 22%     +72.6%      72869 ± 21%  sched_debug.cpu.avg_idle.min
>>>      513534 ± 67%     -58.1%     214959 ±  3%  sched_debug.cpu.avg_idle.stddev
>>>       69440 ± 18%     -53.1%      32566 ± 23%  sched_debug.cpu.curr->pid.avg
>>>      465.44 ± 54%    -100.0%       0.00        sched_debug.cpu.curr->pid.min
>>>        0.71 ± 11%     -40.6%       0.42 ±  9%  sched_debug.cpu.nr_running.avg
>>>        0.22 ± 35%    -100.0%       0.00        sched_debug.cpu.nr_running.min
>>>        0.18 ± 10%     +68.8%       0.31 ±  5%  sched_debug.cpu.nr_running.stddev
>>>       21723 ±  2%     +35.3%      29395 ±  9%  sched_debug.cpu.nr_switches.avg
>>>       37415 ± 16%     +23.6%      46248 ±  5%  sched_debug.cpu.nr_switches.max
>>>       17608           +44.3%      25417 ± 10%  sched_debug.cpu.nr_switches.min
>>>      161.39 ± 12%     +47.1%     237.34 ± 18%  sched_debug.cpu.nr_uninterruptible.max
>>>      -54.78           +48.6%     -81.42        sched_debug.cpu.nr_uninterruptible.min
>>>       24.07 ±  4%     +58.5%      38.15 ± 11%  sched_debug.cpu.nr_uninterruptible.stddev
>>>       11566           +11.1%      12849        slabinfo.Acpi-State.active_objs
>>>       11566           +11.1%      12849        slabinfo.Acpi-State.num_objs
>>>      240075           +13.6%     272838        slabinfo.dentry.active_objs
>>>        5793           +13.6%       6580        slabinfo.dentry.active_slabs
>>>      243365           +13.6%     276386        slabinfo.dentry.num_objs
>>>        5793           +13.6%       6580        slabinfo.dentry.num_slabs
>>>      955.00 ±  9%     +16.4%       1112 ±  7%  slabinfo.file_lock_cache.active_objs
>>>      955.00 ±  9%     +16.4%       1112 ±  7%  slabinfo.file_lock_cache.num_objs
>>>       14271           +14.7%      16366        slabinfo.files_cache.active_objs
>>>       14271           +14.7%      16366        slabinfo.files_cache.num_objs
>>>      145916           +52.7%     222770 ±  2%  slabinfo.filp.active_objs
>>>        2300           +53.0%       3519 ±  2%  slabinfo.filp.active_slabs
>>>      147261           +53.0%     225274 ±  2%  slabinfo.filp.num_objs
>>>        2300           +53.0%       3519 ±  2%  slabinfo.filp.num_slabs
>>>        1319 ± 14%     +26.2%       1664 ±  8%  slabinfo.khugepaged_mm_slot.active_objs
>>>        1319 ± 14%     +26.2%       1664 ±  8%  slabinfo.khugepaged_mm_slot.num_objs
>>>        1037           +19.6%       1240        slabinfo.kmalloc-512.active_slabs
>>>       66437           +19.5%      79424        slabinfo.kmalloc-512.num_objs
>>>        1037           +19.6%       1240        slabinfo.kmalloc-512.num_slabs
>>>      114959            +9.0%     125273        slabinfo.kmalloc-8.active_objs
>>>        4536 ±  5%      +7.2%       4862 ±  3%  slabinfo.kmalloc-rcl-64.active_objs
>>>        4536 ±  5%      +7.2%       4862 ±  3%  slabinfo.kmalloc-rcl-64.num_objs
>>>       10920           +18.9%      12978        slabinfo.mm_struct.active_objs
>>>       10920           +18.9%      12979        slabinfo.mm_struct.num_objs
>>>        1104 ±  5%     -16.9%     917.00 ± 10%  slabinfo.mnt_cache.active_objs
>>>        1104 ±  5%     -16.9%     917.00 ± 10%  slabinfo.mnt_cache.num_objs
>>>        1551           +14.3%       1773 ±  2%  slabinfo.names_cache.active_objs
>>>        1551           +14.3%       1773 ±  2%  slabinfo.names_cache.num_objs
>>>       35694 ±  2%     +49.3%      53295 ±  4%  slabinfo.pid.active_objs
>>>      557.00 ±  2%     +49.4%     832.25 ±  4%  slabinfo.pid.active_slabs
>>>       35694 ±  2%     +49.3%      53295 ±  4%  slabinfo.pid.num_objs
>>>      557.00 ±  2%     +49.4%     832.25 ±  4%  slabinfo.pid.num_slabs
>>>        8786           +20.2%      10558        slabinfo.sighand_cache.active_objs
>>>      590.00           +20.3%     709.75        slabinfo.sighand_cache.active_slabs
>>>        8858           +20.3%      10654        slabinfo.sighand_cache.num_objs
>>>      590.00           +20.3%     709.75        slabinfo.sighand_cache.num_slabs
>>>        8502           +11.6%       9489 ±  2%  slabinfo.sock_inode_cache.active_objs
>>>        8502           +11.6%       9489 ±  2%  slabinfo.sock_inode_cache.num_objs
>>>       39805 ±  2%     +45.4%      57892 ±  2%  slabinfo.task_delay_info.active_objs
>>>      780.00 ±  2%     +45.5%       1135 ±  2%  slabinfo.task_delay_info.active_slabs
>>>       39805 ±  2%     +45.5%      57905 ±  2%  slabinfo.task_delay_info.num_objs
>>>      780.00 ±  2%     +45.5%       1135 ±  2%  slabinfo.task_delay_info.num_slabs
>>>        3088           +29.3%       3992 ±  2%  slabinfo.task_struct.active_objs
>>>        3098           +29.3%       4005 ±  2%  slabinfo.task_struct.active_slabs
>>>        3098           +29.3%       4005 ±  2%  slabinfo.task_struct.num_objs
>>>        3098           +29.3%       4005 ±  2%  slabinfo.task_struct.num_slabs
>>>   1.353e+10           +11.8%  1.513e+10        perf-stat.i.branch-instructions
>>>    84012986 ±  5%     +46.6%  1.232e+08 ±  6%  perf-stat.i.branch-misses
>>>       31.81 ± 12%     -10.8       21.01 ± 21%  perf-stat.i.cache-miss-rate%
>>>    68723351 ±  2%     -10.6%   61409511 ±  4%  perf-stat.i.cache-misses
>>>   2.121e+08 ± 10%     +57.2%  3.333e+08 ±  8%  perf-stat.i.cache-references
>>>       25583 ±  2%     +51.4%      38727        perf-stat.i.context-switches
>>>        5.21 ±  3%     -25.2%       3.90 ±  5%  perf-stat.i.cpi
>>>      192200            +7.5%     206633        perf-stat.i.cpu-clock
>>>    3.37e+11           -28.3%  2.415e+11        perf-stat.i.cpu-cycles
>>>        3432           +61.5%       5541 ±  3%  perf-stat.i.cpu-migrations
>>>        4302           -29.4%       3039 ±  8%  perf-stat.i.cycles-between-cache-misses
>>>     2726072 ± 23%     +52.3%    4151327 ± 26%  perf-stat.i.dTLB-load-misses
>>>   1.215e+10            +6.3%  1.292e+10        perf-stat.i.dTLB-loads
>>>      941111 ± 11%     +46.9%    1382681 ± 11%  perf-stat.i.dTLB-store-misses
>>>   2.493e+09           +48.4%  3.701e+09        perf-stat.i.dTLB-stores
>>>       71.80 ±  4%      -6.7       65.13        perf-stat.i.iTLB-load-miss-rate%
>>>     8355924 ±  6%     +47.0%   12280660 ±  7%  perf-stat.i.iTLB-load-misses
>>>     3268044 ± 18%     +65.6%    5412376 ±  8%  perf-stat.i.iTLB-loads
>>>   6.464e+10           +19.2%  7.704e+10        perf-stat.i.instructions
>>>        6911 ±  3%     -32.1%       4691 ±  6%  perf-stat.i.instructions-per-iTLB-miss
>>>        0.23           +28.9%       0.30        perf-stat.i.ipc
>>>      124.96          +101.4%     251.73 ±  2%  perf-stat.i.major-faults
>>>        1.76           -41.2%       1.03        perf-stat.i.metric.GHz
>>>        0.56 ±  8%     +32.5%       0.75 ±  7%  perf-stat.i.metric.K/sec
>>>      332691           +49.2%     496370        perf-stat.i.minor-faults
>>>       91.27            -4.0       87.22        perf-stat.i.node-load-miss-rate%
>>>    18748059           -13.0%   16311904 ±  3%  perf-stat.i.node-load-misses
>>>     1587953           +50.0%    2382313        perf-stat.i.node-loads
>>>       86.40            -8.3       78.07 ±  2%  perf-stat.i.node-store-miss-rate%
>>>     9547999           -19.9%    7644037 ±  2%  perf-stat.i.node-store-misses
>>>     1064509           +92.2%    2045699 ±  2%  perf-stat.i.node-stores
>>>      332816           +49.2%     496622        perf-stat.i.page-faults
>>>      192200            +7.5%     206633        perf-stat.i.task-clock
>>>        3.25 ±  9%     +36.0%       4.42 ±  7%  perf-stat.overall.MPKI
>>>        0.62 ±  5%      +0.2        0.83 ±  5%  perf-stat.overall.branch-miss-rate%
>>>       32.96 ± 10%     -13.4       19.58 ± 11%  perf-stat.overall.cache-miss-rate%
>>>        5.25           -42.4%       3.02        perf-stat.overall.cpi
>>>        4953           -28.7%       3533 ±  4%  perf-stat.overall.cycles-between-cache-misses
>>>        0.02 ± 23%      +0.0        0.03 ± 24%  perf-stat.overall.dTLB-load-miss-rate%
>>>       72.34 ±  3%      -3.1       69.29        perf-stat.overall.iTLB-load-miss-rate%
>>>        7868 ±  6%     -20.0%       6292 ±  7%  perf-stat.overall.instructions-per-iTLB-miss
>>>        0.19           +73.5%       0.33        perf-stat.overall.ipc
>>>       92.20            -4.8       87.37        perf-stat.overall.node-load-miss-rate%
>>>       90.00            -9.8       80.24        perf-stat.overall.node-store-miss-rate%
>>>    19255954           -23.1%   14811673        perf-stat.overall.path-length
>>>   1.382e+10            +5.2%  1.454e+10        perf-stat.ps.branch-instructions
>>>    85147082 ±  5%     +42.1%   1.21e+08 ±  5%  perf-stat.ps.branch-misses
>>>    69832184 ±  2%      -9.6%   63144472 ±  4%  perf-stat.ps.cache-misses
>>>   2.141e+08 ±  9%     +52.0%  3.254e+08 ±  8%  perf-stat.ps.cache-references
>>>       25647 ±  2%     +46.2%      37498        perf-stat.ps.context-switches
>>>   3.458e+11           -35.6%  2.227e+11        perf-stat.ps.cpu-cycles
>>>        3456           +55.5%       5375 ±  4%  perf-stat.ps.cpu-migrations
>>>     2731148 ± 22%     +50.7%    4114660 ± 24%  perf-stat.ps.dTLB-load-misses
>>>      951612 ± 11%     +42.4%    1354727 ± 10%  perf-stat.ps.dTLB-store-misses
>>>   2.534e+09           +43.0%  3.624e+09        perf-stat.ps.dTLB-stores
>>>     8410006 ±  6%     +39.9%   11763281 ±  6%  perf-stat.ps.iTLB-load-misses
>>>     3251030 ± 17%     +60.5%    5218236 ±  7%  perf-stat.ps.iTLB-loads
>>>   6.591e+10           +11.7%  7.364e+10        perf-stat.ps.instructions
>>>      127.07           +97.9%     251.44 ±  2%  perf-stat.ps.major-faults
>>>      338752           +44.5%     489644        perf-stat.ps.minor-faults
>>>    19070733           -11.1%   16952678 ±  4%  perf-stat.ps.node-load-misses
>>>     1613838           +51.7%    2448074        perf-stat.ps.node-loads
>>>     9716997           -18.4%    7933032 ±  2%  perf-stat.ps.node-store-misses
>>>     1079544           +80.8%    1952245        perf-stat.ps.node-stores
>>>      338879           +44.6%     489896        perf-stat.ps.page-faults
>>>   2.009e+13           +11.5%  2.239e+13        perf-stat.total.instructions
>>>        0.04 ± 54%    +278.3%       0.14 ± 41%  perf-sched.sch_delay.avg.ms.d_alloc_parallel.__lookup_slow.walk_component.link_path_walk.part
>>>        0.11 ±123%    +813.3%       1.04 ±120%  perf-sched.sch_delay.avg.ms.d_alloc_parallel.__lookup_slow.walk_component.path_lookupat.isra
>>>        0.03 ±  9%    +210.5%       0.10 ± 21%  perf-sched.sch_delay.avg.ms.do_task_dead.do_exit.do_group_exit.__x64_sys_exit_group.do_syscall_64
>>>        0.17 ± 89%    +210.6%       0.52 ± 59%  perf-sched.sch_delay.avg.ms.do_wait.kernel_wait4.__do_sys_wait4.do_syscall_64
>>>        0.05 ± 34%   +1883.2%       0.91 ± 48%  perf-sched.sch_delay.avg.ms.exit_to_user_mode_prepare.irqentry_exit_to_user_mode.asm_sysvec_apic_timer_interrupt.[unknown]
>>>        0.12 ± 45%    +772.8%       1.04 ± 65%  perf-sched.sch_delay.avg.ms.exit_to_user_mode_prepare.irqentry_exit_to_user_mode.asm_sysvec_call_function_single.[unknown]
>>>        0.09 ±  9%   +1004.3%       0.97 ± 90%  perf-sched.sch_delay.avg.ms.exit_to_user_mode_prepare.irqentry_exit_to_user_mode.asm_sysvec_reschedule_ipi.[unknown]
>>>        0.06 ± 28%    +390.8%       0.29 ± 41%  perf-sched.sch_delay.avg.ms.exit_to_user_mode_prepare.syscall_exit_to_user_mode.entry_SYSCALL_64_after_hwframe.[unknown]
>>>        0.63 ±  7%     +80.2%       1.13 ± 21%  perf-sched.sch_delay.avg.ms.exit_to_user_mode_prepare.syscall_exit_to_user_mode.ret_from_fork.[unknown]
>>>        0.07 ± 16%    +113.1%       0.14 ± 21%  perf-sched.sch_delay.avg.ms.io_schedule.__lock_page_killable.filemap_fault.__do_fault
>>>        0.01 ±  7%    +196.2%       0.02 ± 65%  perf-sched.sch_delay.avg.ms.pipe_read.new_sync_read.vfs_read.ksys_read
>>>        0.04 ± 45%    +742.5%       0.35 ± 83%  perf-sched.sch_delay.avg.ms.preempt_schedule_common._cond_resched.__alloc_pages_nodemask.alloc_pages_vma.wp_page_copy
>>>        0.07 ± 78%     -84.3%       0.01 ±142%  perf-sched.sch_delay.avg.ms.preempt_schedule_common._cond_resched.__alloc_pages_nodemask.allocate_slab.___slab_alloc
>>>        0.04 ± 79%   +2194.9%       0.89 ±164%  perf-sched.sch_delay.avg.ms.preempt_schedule_common._cond_resched.__alloc_pages_nodemask.pte_alloc_one.__pte_alloc
>>>        0.07 ± 19%     -86.5%       0.01 ± 59%  perf-sched.sch_delay.avg.ms.preempt_schedule_common._cond_resched.__anon_vma_prepare.do_fault.__handle_mm_fault
>>>        0.25 ±105%     -98.2%       0.00 ±137%  perf-sched.sch_delay.avg.ms.preempt_schedule_common._cond_resched.copy_page_to_iter.generic_file_buffered_read.new_sync_read
>>>        0.40 ±  6%     +91.2%       0.77 ±  5%  perf-sched.sch_delay.avg.ms.preempt_schedule_common._cond_resched.do_user_addr_fault.exc_page_fault.asm_exc_page_fault
>>>        0.08 ±100%     -92.6%       0.01 ± 94%  perf-sched.sch_delay.avg.ms.preempt_schedule_common._cond_resched.down_write.__anon_vma_prepare.do_anonymous_page
>>>        0.03 ± 79%    +277.4%       0.12 ± 73%  perf-sched.sch_delay.avg.ms.preempt_schedule_common._cond_resched.down_write.__vma_adjust.__split_vma
>>>        0.01 ± 16%  +1.5e+05%       8.25 ±103%  perf-sched.sch_delay.avg.ms.preempt_schedule_common._cond_resched.down_write.do_truncate.path_openat
>>>        0.04 ± 19%    +318.9%       0.16 ± 44%  perf-sched.sch_delay.avg.ms.preempt_schedule_common._cond_resched.down_write.unlink_anon_vmas.free_pgtables
>>>        0.04 ±  7%   +1180.0%       0.51 ±145%  perf-sched.sch_delay.avg.ms.preempt_schedule_common._cond_resched.down_write.unlink_file_vma.free_pgtables
>>>        0.00 ± 10%     -89.3%       0.00 ±173%  perf-sched.sch_delay.avg.ms.preempt_schedule_common._cond_resched.down_write_killable.__vm_munmap.__x64_sys_munmap
>>>        0.07 ± 54%    +242.3%       0.26 ± 62%  perf-sched.sch_delay.avg.ms.preempt_schedule_common._cond_resched.down_write_killable.vm_mmap_pgoff.elf_map
>>>        0.03 ± 11%     -87.8%       0.00 ± 13%  perf-sched.sch_delay.avg.ms.preempt_schedule_common._cond_resched.down_write_killable.vm_mmap_pgoff.ksys_mmap_pgoff
>>>        0.01 ± 49%  +1.3e+05%       8.04 ±152%  perf-sched.sch_delay.avg.ms.preempt_schedule_common._cond_resched.dput.do_unlinkat.do_syscall_64
>>>        0.10 ±127%  +16890.4%      16.82 ±163%  perf-sched.sch_delay.avg.ms.preempt_schedule_common._cond_resched.dput.done_path_create.do_linkat
>>>        0.13 ± 18%    +459.3%       0.71 ± 18%  perf-sched.sch_delay.avg.ms.preempt_schedule_common._cond_resched.dput.path_openat.do_filp_open
>>>        0.44 ±118%    +204.8%       1.33 ± 94%  perf-sched.sch_delay.avg.ms.preempt_schedule_common._cond_resched.dput.scan_positives.dcache_readdir
>>>        0.01 ± 36%   +3156.9%       0.43 ± 63%  perf-sched.sch_delay.avg.ms.preempt_schedule_common._cond_resched.dput.step_into.path_openat
>>>        0.07 ± 21%    +554.5%       0.46 ± 93%  perf-sched.sch_delay.avg.ms.preempt_schedule_common._cond_resched.dput.terminate_walk.path_openat
>>>        0.00 ±141%  +19625.0%       0.07 ±150%  perf-sched.sch_delay.avg.ms.preempt_schedule_common._cond_resched.generic_file_buffered_read.__kernel_read.load_elf_binary
>>>        0.08 ±134%    +686.6%       0.66 ± 59%  perf-sched.sch_delay.avg.ms.preempt_schedule_common._cond_resched.kmem_cache_alloc.__d_alloc.d_alloc_cursor
>>>        0.04 ± 37%    +648.3%       0.33 ± 28%  perf-sched.sch_delay.avg.ms.preempt_schedule_common._cond_resched.kmem_cache_alloc.security_file_alloc.__alloc_file
>>>        0.21 ± 61%     -77.6%       0.05 ±139%  perf-sched.sch_delay.avg.ms.preempt_schedule_common._cond_resched.kmem_cache_alloc.vm_area_alloc.__install_special_mapping
>>>        0.30 ± 97%     -99.2%       0.00 ± 34%  perf-sched.sch_delay.avg.ms.preempt_schedule_common._cond_resched.kmem_cache_alloc.vm_area_alloc.do_brk_flags
>>>        3.67 ±140%    -100.0%       0.00 ±173%  perf-sched.sch_delay.avg.ms.preempt_schedule_common._cond_resched.kmem_cache_alloc_node_trace.__get_vm_area_node.__vmalloc_node_range
>>>        0.05 ± 51%     +75.6%       0.09 ± 30%  perf-sched.sch_delay.avg.ms.preempt_schedule_common._cond_resched.kmem_cache_alloc_trace.perf_event_mmap.mmap_region
>>>        0.19 ± 85%     -93.9%       0.01 ±148%  perf-sched.sch_delay.avg.ms.preempt_schedule_common._cond_resched.kmem_cache_alloc_trace.perf_event_mmap.mprotect_fixup
>>>        0.01 ± 66%  +89763.4%       8.39 ±106%  perf-sched.sch_delay.avg.ms.preempt_schedule_common._cond_resched.mnt_want_write.path_openat.do_filp_open
>>>        0.01 ±108%    +903.8%       0.13 ±129%  perf-sched.sch_delay.avg.ms.preempt_schedule_common._cond_resched.mutex_lock.futex_exec_release.exec_mm_release
>>>        0.03 ± 53%     -67.5%       0.01 ± 83%  perf-sched.sch_delay.avg.ms.preempt_schedule_common._cond_resched.mutex_lock.futex_exit_release.exit_mm_release
>>>        0.11 ±125%   +1883.9%       2.25 ± 40%  perf-sched.sch_delay.avg.ms.preempt_schedule_common._cond_resched.smpboot_thread_fn.kthread.ret_from_fork
>>>        0.01 ± 70%    -100.0%       0.00        perf-sched.sch_delay.avg.ms.preempt_schedule_common._cond_resched.stop_one_cpu.migrate_task_to.task_numa_migrate
>>>        0.30 ±  3%    +308.2%       1.22 ± 26%  perf-sched.sch_delay.avg.ms.preempt_schedule_common._cond_resched.stop_one_cpu.sched_exec.bprm_execve
>>>        0.08 ± 68%    -100.0%       0.00        perf-sched.sch_delay.avg.ms.preempt_schedule_common._cond_resched.wait_for_completion.stop_two_cpus.migrate_swap
>>>        0.01 ±103%    -100.0%       0.00        perf-sched.sch_delay.avg.ms.rwsem_down_read_slowpath.page_lock_anon_vma_read.rmap_walk_anon.try_to_unmap
>>>        0.01 ±106%    -100.0%       0.00        perf-sched.sch_delay.avg.ms.rwsem_down_read_slowpath.rmap_walk_anon.remove_migration_ptes.migrate_pages
>>>        0.01 ± 49%   +1115.0%       0.14 ±143%  perf-sched.sch_delay.avg.ms.rwsem_down_write_slowpath.__vma_adjust.__split_vma.__do_munmap
>>>        0.03 ± 52%     -89.4%       0.00 ±107%  perf-sched.sch_delay.avg.ms.rwsem_down_write_slowpath.anon_vma_clone.anon_vma_fork.dup_mmap
>>>        0.02 ± 54%   +1681.5%       0.39 ± 70%  perf-sched.sch_delay.avg.ms.rwsem_down_write_slowpath.do_unlinkat.do_syscall_64.entry_SYSCALL_64_after_hwframe
>>>        0.01 ± 80%    +507.1%       0.09 ± 45%  perf-sched.sch_delay.avg.ms.rwsem_down_write_slowpath.dup_mmap.dup_mm.copy_process
>>>        0.02 ± 24%   +1250.0%       0.26 ± 68%  perf-sched.sch_delay.avg.ms.rwsem_down_write_slowpath.filename_create.do_linkat.__x64_sys_link
>>>        0.07 ± 65%     -87.3%       0.01 ± 75%  perf-sched.sch_delay.avg.ms.rwsem_down_write_slowpath.unlink_anon_vmas.free_pgtables.exit_mmap
>>>        0.02 ± 20%    +232.6%       0.05 ± 33%  perf-sched.sch_delay.avg.ms.rwsem_down_write_slowpath.unlink_file_vma.free_pgtables.exit_mmap
>>>        0.01 ± 27%    +363.9%       0.04 ± 43%  perf-sched.sch_delay.avg.ms.rwsem_down_write_slowpath.vma_link.mmap_region.do_mmap
>>>        0.19 ± 68%     -95.1%       0.01 ± 13%  perf-sched.sch_delay.avg.ms.schedule_hrtimeout_range_clock.ep_poll.do_epoll_wait.__x64_sys_epoll_wait
>>>        0.04 ± 18%   +5366.5%       1.93 ± 40%  perf-sched.sch_delay.avg.ms.schedule_timeout.kcompactd.kthread.ret_from_fork
>>>        0.04 ± 39%    +464.8%       0.24 ± 60%  perf-sched.sch_delay.avg.ms.schedule_timeout.rcu_gp_kthread.kthread.ret_from_fork
>>>        0.00 ±141%  +48300.0%       0.97 ±170%  perf-sched.sch_delay.avg.ms.schedule_timeout.wait_for_completion.__flush_work.lru_add_drain_all
>>>        0.20 ±  7%   +1238.1%       2.63 ± 31%  perf-sched.sch_delay.avg.ms.smpboot_thread_fn.kthread.ret_from_fork
>>>        0.12 ± 27%     -74.7%       0.03 ±149%  perf-sched.sch_delay.avg.ms.wait_for_partner.fifo_open.do_dentry_open.path_openat
>>>        0.01 ±  7%   +9784.2%       1.25 ± 18%  perf-sched.sch_delay.avg.ms.worker_thread.kthread.ret_from_fork
>>>        5.01 ± 31%    +324.7%      21.26 ± 36%  perf-sched.sch_delay.max.ms.d_alloc_parallel.__lookup_slow.walk_component.link_path_walk.part
>>>        3.59 ±132%   +1076.4%      42.25 ± 96%  perf-sched.sch_delay.max.ms.d_alloc_parallel.__lookup_slow.walk_component.path_lookupat.isra
>>>       16.19 ±  7%   +1397.5%     242.51 ± 30%  perf-sched.sch_delay.max.ms.do_task_dead.do_exit.do_group_exit.__x64_sys_exit_group.do_syscall_64
>>>        9.13 ± 27%   +2518.1%     238.94 ± 62%  perf-sched.sch_delay.max.ms.exit_to_user_mode_prepare.irqentry_exit_to_user_mode.asm_sysvec_apic_timer_interrupt.[unknown]
>>>       10.04 ± 65%   +1971.0%     207.85 ± 63%  perf-sched.sch_delay.max.ms.exit_to_user_mode_prepare.irqentry_exit_to_user_mode.asm_sysvec_call_function_single.[unknown]
>>>        9.21 ± 61%   +2000.8%     193.58 ± 63%  perf-sched.sch_delay.max.ms.exit_to_user_mode_prepare.irqentry_exit_to_user_mode.asm_sysvec_reschedule_ipi.[unknown]
>>>       11.01 ± 18%    +799.2%      99.02 ± 86%  perf-sched.sch_delay.max.ms.exit_to_user_mode_prepare.syscall_exit_to_user_mode.entry_SYSCALL_64_after_hwframe.[unknown]
>>>       14.73 ± 13%    +265.9%      53.91 ± 26%  perf-sched.sch_delay.max.ms.exit_to_user_mode_prepare.syscall_exit_to_user_mode.ret_from_fork.[unknown]
>>>       10.65 ± 44%    +151.1%      26.76 ± 19%  perf-sched.sch_delay.max.ms.io_schedule.__lock_page_killable.filemap_fault.__do_fault
>>>        1.21 ± 24%   +2921.4%      36.62 ±124%  perf-sched.sch_delay.max.ms.pipe_read.new_sync_read.vfs_read.ksys_read
>>>        0.53 ± 55%   +1359.3%       7.76 ±113%  perf-sched.sch_delay.max.ms.preempt_schedule_common._cond_resched.__alloc_pages_nodemask.alloc_pages_vma.wp_page_copy
>>>        0.30 ± 92%     -90.7%       0.03 ±155%  perf-sched.sch_delay.max.ms.preempt_schedule_common._cond_resched.__alloc_pages_nodemask.allocate_slab.___slab_alloc
>>>        0.37 ± 69%   +3449.8%      13.07 ±162%  perf-sched.sch_delay.max.ms.preempt_schedule_common._cond_resched.__alloc_pages_nodemask.pte_alloc_one.__pte_alloc
>>>        0.48 ± 37%     -89.3%       0.05 ±102%  perf-sched.sch_delay.max.ms.preempt_schedule_common._cond_resched.__anon_vma_prepare.do_fault.__handle_mm_fault
>>>        0.55 ± 93%     -95.3%       0.03 ±153%  perf-sched.sch_delay.max.ms.preempt_schedule_common._cond_resched.__kmalloc_node.memcg_alloc_page_obj_cgroups.kmem_cache_alloc_node
>>>        0.97 ± 13%    +591.3%       6.72 ±119%  perf-sched.sch_delay.max.ms.preempt_schedule_common._cond_resched.change_p4d_range.change_protection.mprotect_fixup
>>>        0.31 ±106%     -98.6%       0.00 ±137%  perf-sched.sch_delay.max.ms.preempt_schedule_common._cond_resched.copy_page_to_iter.generic_file_buffered_read.new_sync_read
>>>       11.49 ±  5%     +87.8%      21.57 ± 31%  perf-sched.sch_delay.max.ms.preempt_schedule_common._cond_resched.do_user_addr_fault.exc_page_fault.asm_exc_page_fault
>>>        0.32 ±102%     -96.0%       0.01 ± 79%  perf-sched.sch_delay.max.ms.preempt_schedule_common._cond_resched.down_write.__anon_vma_prepare.do_anonymous_page
>>>        1.05 ± 98%    +478.9%       6.10 ± 82%  perf-sched.sch_delay.max.ms.preempt_schedule_common._cond_resched.down_write.__vma_adjust.__split_vma
>>>        0.02 ±  5%  +1.4e+06%     217.29 ±104%  perf-sched.sch_delay.max.ms.preempt_schedule_common._cond_resched.down_write.do_truncate.path_openat
>>>        1.95 ± 26%    +682.5%      15.25 ± 64%  perf-sched.sch_delay.max.ms.preempt_schedule_common._cond_resched.down_write.unlink_anon_vmas.free_pgtables
>>>        0.93 ± 14%   +9721.3%      91.57 ±158%  perf-sched.sch_delay.max.ms.preempt_schedule_common._cond_resched.down_write.unlink_file_vma.free_pgtables
>>>        0.00 ± 10%     -78.6%       0.00 ±173%  perf-sched.sch_delay.max.ms.preempt_schedule_common._cond_resched.down_write_killable.__vm_munmap.__x64_sys_munmap
>>>        0.24 ± 43%     -96.7%       0.01 ± 72%  perf-sched.sch_delay.max.ms.preempt_schedule_common._cond_resched.down_write_killable.vm_mmap_pgoff.ksys_mmap_pgoff
>>>        0.01 ± 28%  +8.3e+05%     113.60 ±164%  perf-sched.sch_delay.max.ms.preempt_schedule_common._cond_resched.dput.do_unlinkat.do_syscall_64
>>>        0.21 ±108%  +34647.0%      72.97 ±145%  perf-sched.sch_delay.max.ms.preempt_schedule_common._cond_resched.dput.done_path_create.do_linkat
>>>        3.03 ± 39%   +1407.5%      45.73 ± 29%  perf-sched.sch_delay.max.ms.preempt_schedule_common._cond_resched.dput.path_openat.do_filp_open
>>>        1.22 ± 39%   +4068.4%      50.91 ± 79%  perf-sched.sch_delay.max.ms.preempt_schedule_common._cond_resched.dput.step_into.path_openat
>>>        3.38 ± 60%   +3700.6%     128.54 ±127%  perf-sched.sch_delay.max.ms.preempt_schedule_common._cond_resched.dput.terminate_walk.path_openat
>>>        0.88 ± 46%    +299.6%       3.50 ± 55%  perf-sched.sch_delay.max.ms.preempt_schedule_common._cond_resched.exit_mmap.mmput.begin_new_exec
>>>        0.00 ±141%  +22250.0%       0.30 ±167%  perf-sched.sch_delay.max.ms.preempt_schedule_common._cond_resched.generic_file_buffered_read.__kernel_read.load_elf_binary
>>>        0.57 ±139%  +12527.6%      71.47 ± 89%  perf-sched.sch_delay.max.ms.preempt_schedule_common._cond_resched.kmem_cache_alloc.__d_alloc.d_alloc_cursor
>>>        0.00 ±141%   +1525.0%       0.03 ±144%  perf-sched.sch_delay.max.ms.preempt_schedule_common._cond_resched.kmem_cache_alloc.__delayacct_tsk_init.copy_process
>>>        1.73 ± 43%    +863.5%      16.71 ± 50%  perf-sched.sch_delay.max.ms.preempt_schedule_common._cond_resched.kmem_cache_alloc.security_file_alloc.__alloc_file
>>>        3.67 ±140%    -100.0%       0.00 ±173%  perf-sched.sch_delay.max.ms.preempt_schedule_common._cond_resched.kmem_cache_alloc_node_trace.__get_vm_area_node.__vmalloc_node_range
>>>        3.19 ± 52%    +345.8%      14.24 ± 84%  perf-sched.sch_delay.max.ms.preempt_schedule_common._cond_resched.kmem_cache_alloc_trace.perf_event_mmap.mmap_region
>>>        1.01 ± 99%     -91.8%       0.08 ±167%  perf-sched.sch_delay.max.ms.preempt_schedule_common._cond_resched.kmem_cache_alloc_trace.perf_event_mmap.mprotect_fixup
>>>        0.11 ± 85%  +1.7e+05%     187.38 ± 84%  perf-sched.sch_delay.max.ms.preempt_schedule_common._cond_resched.mnt_want_write.path_openat.do_filp_open
>>>        0.05 ±130%    +602.6%       0.36 ±142%  perf-sched.sch_delay.max.ms.preempt_schedule_common._cond_resched.mutex_lock.futex_exec_release.exec_mm_release
>>>        0.52 ±136%  +10794.6%      57.09 ± 72%  perf-sched.sch_delay.max.ms.preempt_schedule_common._cond_resched.smpboot_thread_fn.kthread.ret_from_fork
>>>        0.06 ± 92%    -100.0%       0.00        perf-sched.sch_delay.max.ms.preempt_schedule_common._cond_resched.stop_one_cpu.migrate_task_to.task_numa_migrate
>>>       28.50 ± 12%    +645.4%     212.48 ± 39%  perf-sched.sch_delay.max.ms.preempt_schedule_common._cond_resched.stop_one_cpu.sched_exec.bprm_execve
>>>        1.20 ± 76%    -100.0%       0.00        perf-sched.sch_delay.max.ms.preempt_schedule_common._cond_resched.wait_for_completion.stop_two_cpus.migrate_swap
>>>        0.18 ±130%    -100.0%       0.00        perf-sched.sch_delay.max.ms.rwsem_down_read_slowpath.page_lock_anon_vma_read.rmap_walk_anon.try_to_unmap
>>>        0.20 ±132%    -100.0%       0.00        perf-sched.sch_delay.max.ms.rwsem_down_read_slowpath.rmap_walk_anon.remove_migration_ptes.migrate_pages
>>>        1.33 ±103%     -96.9%       0.04 ±173%  perf-sched.sch_delay.max.ms.rwsem_down_write_slowpath.__put_anon_vma.unlink_anon_vmas.free_pgtables
>>>        1.53 ± 79%   +6359.8%      98.51 ±148%  perf-sched.sch_delay.max.ms.rwsem_down_write_slowpath.__vma_adjust.__split_vma.__do_munmap
>>>        2.57 ± 63%    +676.1%      19.92 ± 49%  perf-sched.sch_delay.max.ms.rwsem_down_write_slowpath.__vma_adjust.__split_vma.mprotect_fixup
>>>        0.78 ± 76%     -99.2%       0.01 ±138%  perf-sched.sch_delay.max.ms.rwsem_down_write_slowpath.anon_vma_clone.anon_vma_fork.dup_mmap
>>>        0.19 ± 94%     -97.6%       0.00 ±100%  perf-sched.sch_delay.max.ms.rwsem_down_write_slowpath.anon_vma_fork.dup_mmap.dup_mm
>>>        5.70 ± 45%   +1878.3%     112.68 ± 89%  perf-sched.sch_delay.max.ms.rwsem_down_write_slowpath.do_unlinkat.do_syscall_64.entry_SYSCALL_64_after_hwframe
>>>        3.68 ±111%   +1530.5%      60.07 ± 26%  perf-sched.sch_delay.max.ms.rwsem_down_write_slowpath.dup_mmap.dup_mm.copy_process
>>>        6.33 ± 17%   +2353.2%     155.41 ±105%  perf-sched.sch_delay.max.ms.rwsem_down_write_slowpath.filename_create.do_linkat.__x64_sys_link
>>>        2.23 ± 66%     -98.3%       0.04 ± 65%  perf-sched.sch_delay.max.ms.rwsem_down_write_slowpath.unlink_anon_vmas.free_pgtables.exit_mmap
>>>        6.50 ± 41%    +643.0%      48.28 ± 78%  perf-sched.sch_delay.max.ms.rwsem_down_write_slowpath.unlink_file_vma.free_pgtables.exit_mmap
>>>        2.45 ± 43%   +1688.6%      43.80 ± 88%  perf-sched.sch_delay.max.ms.rwsem_down_write_slowpath.vma_link.mmap_region.do_mmap
>>>        9.31 ± 80%     -99.6%       0.04 ± 54%  perf-sched.sch_delay.max.ms.schedule_hrtimeout_range_clock.ep_poll.do_epoll_wait.__x64_sys_epoll_wait
>>>        0.52 ± 65%  +15185.6%      79.99 ± 38%  perf-sched.sch_delay.max.ms.schedule_timeout.kcompactd.kthread.ret_from_fork
>>>       16.34 ± 98%    +385.7%      79.39 ± 79%  perf-sched.sch_delay.max.ms.schedule_timeout.rcu_gp_kthread.kthread.ret_from_fork
>>>        0.00 ±141%  +72153.1%       1.93 ±171%  perf-sched.sch_delay.max.ms.schedule_timeout.wait_for_completion.__flush_work.lru_add_drain_all
>>>       10.28 ± 58%   +2363.0%     253.08 ± 17%  perf-sched.sch_delay.max.ms.smpboot_thread_fn.kthread.ret_from_fork
>>>        8.41 ± 37%   +3406.7%     295.08 ± 29%  perf-sched.sch_delay.max.ms.worker_thread.kthread.ret_from_fork
>>>        0.12 ± 74%    +358.2%       0.54 ± 29%  perf-sched.total_sch_delay.average.ms
>>>       39.70 ±  3%     -15.2%      33.67 ±  3%  perf-sched.total_wait_and_delay.average.ms
>>>      157875 ±  3%     +29.4%     204320 ±  3%  perf-sched.total_wait_and_delay.count.ms
>>>        5585           +14.5%       6395 ±  3%  perf-sched.total_wait_and_delay.max.ms
>>>       39.58 ±  3%     -16.3%      33.13 ±  3%  perf-sched.total_wait_time.average.ms
>>>        5585           +14.5%       6395 ±  3%  perf-sched.total_wait_time.max.ms
>>>        6.67 ±  2%     +90.8%      12.72 ± 10%  perf-sched.wait_and_delay.avg.ms.do_wait.kernel_wait4.__do_sys_wait4.do_syscall_64
>>>        6.25 ± 22%    +706.2%      50.40 ± 54%  perf-sched.wait_and_delay.avg.ms.exit_to_user_mode_prepare.irqentry_exit_to_user_mode.asm_sysvec_apic_timer_interrupt.[unknown]
>>>        6.09 ± 15%    -100.0%       0.00        perf-sched.wait_and_delay.avg.ms.exit_to_user_mode_prepare.syscall_exit_to_user_mode.entry_SYSCALL_64_after_hwframe.[unknown]
>>>       38.34 ± 13%     +47.7%      56.61 ± 33%  perf-sched.wait_and_delay.avg.ms.pipe_read.new_sync_read.vfs_read.ksys_read
>>>        5.09           +14.7%       5.83 ±  3%  perf-sched.wait_and_delay.avg.ms.preempt_schedule_common._cond_resched.stop_one_cpu.__set_cpus_allowed_ptr.sched_setaffinity
>>>        0.33 ±  3%    +330.6%       1.42 ± 24%  perf-sched.wait_and_delay.avg.ms.preempt_schedule_common._cond_resched.stop_one_cpu.sched_exec.bprm_execve
>>>        1.33 ± 20%    +419.5%       6.93 ±  7%  perf-sched.wait_and_delay.avg.ms.rwsem_down_write_slowpath.do_unlinkat.do_syscall_64.entry_SYSCALL_64_after_hwframe
>>>        0.60 ± 70%    +168.1%       1.61 ± 18%  perf-sched.wait_and_delay.avg.ms.rwsem_down_write_slowpath.dup_mmap.dup_mm.copy_process
>>>        3.71 ± 11%     +29.9%       4.82 ± 12%  perf-sched.wait_and_delay.avg.ms.rwsem_down_write_slowpath.path_openat.do_filp_open.do_sys_openat2
>>>        0.04 ±141%    +440.1%       0.24 ± 12%  perf-sched.wait_and_delay.avg.ms.rwsem_down_write_slowpath.unlink_file_vma.free_pgtables.unmap_region
>>>        0.06 ±141%    +604.1%       0.41 ± 20%  perf-sched.wait_and_delay.avg.ms.rwsem_down_write_slowpath.vma_link.mmap_region.do_mmap
>>>      251.31 ±  7%     -36.4%     159.76 ± 18%  perf-sched.wait_and_delay.avg.ms.schedule_hrtimeout_range_clock.ep_poll.do_epoll_wait.__x64_sys_epoll_wait
>>>       87.92 ± 14%     +50.7%     132.47 ± 22%  perf-sched.wait_and_delay.avg.ms.schedule_hrtimeout_range_clock.poll_schedule_timeout.constprop.0.do_sys_poll
>>>      384.15           -32.3%     260.23 ± 10%  perf-sched.wait_and_delay.avg.ms.smpboot_thread_fn.kthread.ret_from_fork
>>>      142.29           -30.8%      98.47 ±  6%  perf-sched.wait_and_delay.avg.ms.worker_thread.kthread.ret_from_fork
>>>       35.67 ± 11%     +58.4%      56.50        perf-sched.wait_and_delay.count.devkmsg_read.vfs_read.ksys_read.do_syscall_64
>>>       14.67 ±  8%     +19.3%      17.50 ±  4%  perf-sched.wait_and_delay.count.do_nanosleep.hrtimer_nanosleep.__x64_sys_nanosleep.do_syscall_64
>>>       35.67 ±  9%     +58.4%      56.50        perf-sched.wait_and_delay.count.do_syslog.part.0.kmsg_read.vfs_read
>>>       39991           +28.3%      51295 ±  4%  perf-sched.wait_and_delay.count.do_task_dead.do_exit.do_group_exit.__x64_sys_exit_group.do_syscall_64
>>>       41076           +26.8%      52088 ±  4%  perf-sched.wait_and_delay.count.do_wait.kernel_wait4.__do_sys_wait4.do_syscall_64
>>>      914.67 ±  2%     +68.3%       1539 ±  6%  perf-sched.wait_and_delay.count.exit_to_user_mode_prepare.irqentry_exit_to_user_mode.asm_sysvec_apic_timer_interrupt.[unknown]
>>>      122.33 ±141%    +401.1%     613.00 ± 22%  perf-sched.wait_and_delay.count.exit_to_user_mode_prepare.irqentry_exit_to_user_mode.asm_sysvec_reschedule_ipi.[unknown]
>>>        1649 ±  8%    -100.0%       0.00        perf-sched.wait_and_delay.count.exit_to_user_mode_prepare.syscall_exit_to_user_mode.entry_SYSCALL_64_after_hwframe.[unknown]
>>>        4728 ± 13%     -91.3%     412.00 ±100%  perf-sched.wait_and_delay.count.preempt_schedule_common._cond_resched.dput.scan_positives.dcache_readdir
>>>        4.67 ± 44%   +1325.0%      66.50 ± 51%  perf-sched.wait_and_delay.count.preempt_schedule_common._cond_resched.smpboot_thread_fn.kthread.ret_from_fork
>>>        1987 ±  4%     -13.4%       1720 ±  4%  perf-sched.wait_and_delay.count.preempt_schedule_common._cond_resched.stop_one_cpu.__set_cpus_allowed_ptr.sched_setaffinity
>>>        3064 ±  8%     +76.6%       5412 ± 13%  perf-sched.wait_and_delay.count.preempt_schedule_common._cond_resched.stop_one_cpu.sched_exec.bprm_execve
>>>        2046 ±  7%     +68.3%       3443 ± 15%  perf-sched.wait_and_delay.count.rwsem_down_read_slowpath.walk_component.path_lookupat.isra.0
>>>      896.33 ±  5%     +81.3%       1625 ± 12%  perf-sched.wait_and_delay.count.rwsem_down_write_slowpath.do_unlinkat.do_syscall_64.entry_SYSCALL_64_after_hwframe
>>>      776.00 ± 75%    +292.9%       3049 ± 44%  perf-sched.wait_and_delay.count.rwsem_down_write_slowpath.dup_mmap.dup_mm.copy_process
>>>        1649 ± 32%    +198.7%       4926 ± 45%  perf-sched.wait_and_delay.count.rwsem_down_write_slowpath.unlink_file_vma.free_pgtables.exit_mmap
>>>      354.67 ±141%    +482.9%       2067 ± 36%  perf-sched.wait_and_delay.count.rwsem_down_write_slowpath.unlink_file_vma.free_pgtables.unmap_region
>>>      350.00 ±141%    +496.1%       2086 ± 33%  perf-sched.wait_and_delay.count.rwsem_down_write_slowpath.vma_link.mmap_region.do_mmap
>>>       13.00 ± 21%     +61.5%      21.00        perf-sched.wait_and_delay.count.schedule_hrtimeout_range_clock.poll_schedule_timeout.constprop.0.do_select
>>>      163.67 ±  3%     -35.2%     106.00 ± 25%  perf-sched.wait_and_delay.count.schedule_hrtimeout_range_clock.poll_schedule_timeout.constprop.0.do_sys_poll
>>>        1762 ±  4%     -17.5%       1454 ± 12%  perf-sched.wait_and_delay.count.schedule_timeout.rcu_gp_kthread.kthread.ret_from_fork
>>>        9129 ±  2%     +43.1%      13062 ±  9%  perf-sched.wait_and_delay.count.smpboot_thread_fn.kthread.ret_from_fork
>>>       12642           +52.2%      19242 ±  3%  perf-sched.wait_and_delay.count.worker_thread.kthread.ret_from_fork
>>>        5512           -30.9%       3811        perf-sched.wait_and_delay.max.ms.devkmsg_read.vfs_read.ksys_read.do_syscall_64
>>>        5512           -30.9%       3811        perf-sched.wait_and_delay.max.ms.do_syslog.part.0.kmsg_read.vfs_read
>>>      340.62 ±141%    +203.4%       1033        perf-sched.wait_and_delay.max.ms.exit_to_user_mode_prepare.irqentry_exit_to_user_mode.asm_sysvec_reschedule_ipi.[unknown]
>>>        5496          -100.0%       0.00        perf-sched.wait_and_delay.max.ms.exit_to_user_mode_prepare.syscall_exit_to_user_mode.entry_SYSCALL_64_after_hwframe.[unknown]
>>>        5580           -30.1%       3899        perf-sched.wait_and_delay.max.ms.pipe_read.new_sync_read.vfs_read.ksys_read
>>>        1495 ± 31%    +114.1%       3203 ±  4%  perf-sched.wait_and_delay.max.ms.preempt_schedule_common._cond_resched.smpboot_thread_fn.kthread.ret_from_fork
>>>       28.50 ± 12%    +645.4%     212.48 ± 39%  perf-sched.wait_and_delay.max.ms.preempt_schedule_common._cond_resched.stop_one_cpu.sched_exec.bprm_execve
>>>       20.64 ± 78%   +2946.5%     628.69 ± 62%  perf-sched.wait_and_delay.max.ms.rwsem_down_write_slowpath.dup_mmap.dup_mm.copy_process
>>>       20.38 ± 28%    +299.0%      81.32 ± 53%  perf-sched.wait_and_delay.max.ms.rwsem_down_write_slowpath.unlink_file_vma.free_pgtables.exit_mmap
>>>        3.90 ±141%    +526.7%      24.46 ± 34%  perf-sched.wait_and_delay.max.ms.rwsem_down_write_slowpath.unlink_file_vma.free_pgtables.unmap_region
>>>        3.78 ±141%   +1570.8%      63.11 ± 45%  perf-sched.wait_and_delay.max.ms.rwsem_down_write_slowpath.vma_link.mmap_region.do_mmap
>>>        5176           -31.0%       3569        perf-sched.wait_and_delay.max.ms.schedule_hrtimeout_range_clock.ep_poll.do_epoll_wait.__x64_sys_epoll_wait
>>>        5585           -30.2%       3899        perf-sched.wait_and_delay.max.ms.schedule_hrtimeout_range_clock.poll_schedule_timeout.constprop.0.do_select
>>>      505.28           +31.3%     663.68 ±  9%  perf-sched.wait_and_delay.max.ms.schedule_timeout.kcompactd.kthread.ret_from_fork
>>>        4817 ± 15%     -31.2%       3315 ±  3%  perf-sched.wait_and_delay.max.ms.smpboot_thread_fn.kthread.ret_from_fork
>>>        1791           -16.1%       1502 ±  3%  perf-sched.wait_and_delay.max.ms.worker_thread.kthread.ret_from_fork
>>>        0.08 ± 39%    +460.9%       0.43 ± 27%  perf-sched.wait_time.avg.ms.d_alloc_parallel.__lookup_slow.walk_component.link_path_walk.part
>>>        0.17 ± 92%    +182.7%       0.48 ± 59%  perf-sched.wait_time.avg.ms.d_alloc_parallel.__lookup_slow.walk_component.path_lookupat.isra
>>>        6.50 ±  3%     +87.7%      12.20 ±  8%  perf-sched.wait_time.avg.ms.do_wait.kernel_wait4.__do_sys_wait4.do_syscall_64
>>>        0.59 ±123%    +129.1%       1.35 ± 86%  perf-sched.wait_time.avg.ms.exit_to_user_mode_prepare.irqentry_exit_to_user_mode.asm_exc_page_fault.[unknown]
>>>        6.21 ± 22%    +697.5%      49.49 ± 55%  perf-sched.wait_time.avg.ms.exit_to_user_mode_prepare.irqentry_exit_to_user_mode.asm_sysvec_apic_timer_interrupt.[unknown]
>>>        8.60 ± 95%    +175.7%      23.70 ± 34%  perf-sched.wait_time.avg.ms.exit_to_user_mode_prepare.irqentry_exit_to_user_mode.asm_sysvec_call_function_single.[unknown]
>>>        6.03 ± 15%    +354.9%      27.43 ± 13%  perf-sched.wait_time.avg.ms.exit_to_user_mode_prepare.syscall_exit_to_user_mode.entry_SYSCALL_64_after_hwframe.[unknown]
>>>        0.00 ± 88%  +24178.6%       1.13 ±156%  perf-sched.wait_time.avg.ms.io_schedule.__lock_page.__do_fault.do_fault
>>>       38.33 ± 13%     +47.6%      56.59 ± 33%  perf-sched.wait_time.avg.ms.pipe_read.new_sync_read.vfs_read.ksys_read
>>>        0.53 ±120%     -98.4%       0.01 ±102%  perf-sched.wait_time.avg.ms.preempt_schedule_common._cond_resched.__alloc_pages_nodemask.__pmd_alloc.__handle_mm_fault
>>>        0.01 ±115%   +6617.2%       0.36 ±168%  perf-sched.wait_time.avg.ms.preempt_schedule_common._cond_resched.__kmalloc.load_elf_phdrs.load_elf_binary
>>>        0.42 ± 44%    +593.4%       2.89 ± 57%  perf-sched.wait_time.avg.ms.preempt_schedule_common._cond_resched.__kmalloc.security_task_alloc.copy_process
>>>        0.08 ± 57%    +614.6%       0.60 ± 48%  perf-sched.wait_time.avg.ms.preempt_schedule_common._cond_resched.__put_anon_vma.unlink_anon_vmas.free_pgtables
>>>        0.27 ±140%    +525.8%       1.67 ± 86%  perf-sched.wait_time.avg.ms.preempt_schedule_common._cond_resched.cgroup_can_fork.copy_process.kernel_clone
>>>        0.05 ± 25%    +978.8%       0.50 ± 50%  perf-sched.wait_time.avg.ms.preempt_schedule_common._cond_resched.change_p4d_range.change_protection.mprotect_fixup
>>>        0.53 ± 73%    +685.8%       4.14 ± 66%  perf-sched.wait_time.avg.ms.preempt_schedule_common._cond_resched.copy_pte_range.copy_page_range.dup_mmap
>>>        0.02 ±123%   +6830.4%       1.59 ±103%  perf-sched.wait_time.avg.ms.preempt_schedule_common._cond_resched.down_read.path_openat.do_filp_open
>>>        0.06 ± 29%    +796.1%       0.51 ± 93%  perf-sched.wait_time.avg.ms.preempt_schedule_common._cond_resched.down_read.walk_component.link_path_walk
>>>        0.01 ±141%   +2254.2%       0.14 ±161%  perf-sched.wait_time.avg.ms.preempt_schedule_common._cond_resched.down_read_killable.create_elf_tables.isra
>>>        0.01 ± 97%   +9626.4%       1.13 ± 62%  perf-sched.wait_time.avg.ms.preempt_schedule_common._cond_resched.down_write.__anon_vma_prepare.do_fault
>>>        0.02 ± 87%   +4118.8%       0.79 ± 55%  perf-sched.wait_time.avg.ms.preempt_schedule_common._cond_resched.down_write.__vma_adjust.shift_arg_pages
>>>        1.50 ± 68%   +1104.7%      18.05 ±100%  perf-sched.wait_time.avg.ms.preempt_schedule_common._cond_resched.down_write.anon_vma_fork.dup_mmap
>>>        2.02 ± 13%    +454.8%      11.23 ± 69%  perf-sched.wait_time.avg.ms.preempt_schedule_common._cond_resched.down_write.do_truncate.path_openat
>>>        1.78 ± 66%    +587.5%      12.24 ± 99%  perf-sched.wait_time.avg.ms.preempt_schedule_common._cond_resched.down_write.dup_mmap.dup_mm
>>>        0.86 ± 39%     +37.3%       1.18 ± 11%  perf-sched.wait_time.avg.ms.preempt_schedule_common._cond_resched.down_write.unlink_file_vma.free_pgtables
>>>        0.01 ±141%   +3608.3%       0.22 ± 85%  perf-sched.wait_time.avg.ms.preempt_schedule_common._cond_resched.down_write_killable.__vm_munmap.elf_map
>>>        0.01 ±135%    +539.5%       0.08 ± 83%  perf-sched.wait_time.avg.ms.preempt_schedule_common._cond_resched.down_write_killable.vm_brk_flags.load_elf_interp
>>>        0.04 ± 21%    +251.3%       0.14 ± 50%  perf-sched.wait_time.avg.ms.preempt_schedule_common._cond_resched.down_write_killable.vm_mmap_pgoff.elf_map
>>>        0.25 ± 70%   +1512.8%       4.09 ±132%  perf-sched.wait_time.avg.ms.preempt_schedule_common._cond_resched.dput.__fput.task_work_run
>>>        0.18 ±139%   +1040.9%       2.06 ± 93%  perf-sched.wait_time.avg.ms.preempt_schedule_common._cond_resched.dput.dcache_dir_close.__fput
>>>        0.32 ± 75%   +3385.1%      11.30 ±101%  perf-sched.wait_time.avg.ms.preempt_schedule_common._cond_resched.dput.do_unlinkat.do_syscall_64
>>>        9.06 ± 10%    +226.6%      29.60 ± 20%  perf-sched.wait_time.avg.ms.preempt_schedule_common._cond_resched.dput.scan_positives.dcache_readdir
>>>        2.19 ±119%    +216.9%       6.95 ± 21%  perf-sched.wait_time.avg.ms.preempt_schedule_common._cond_resched.dput.step_into.path_openat
>>>        0.16 ± 19%   +1845.7%       3.15 ± 29%  perf-sched.wait_time.avg.ms.preempt_schedule_common._cond_resched.dput.terminate_walk.path_openat
>>>        0.03 ± 70%   +2914.6%       0.82 ± 63%  perf-sched.wait_time.avg.ms.preempt_schedule_common._cond_resched.exit_mmap.mmput.begin_new_exec
>>>        0.97 ± 23%    +131.5%       2.24 ± 56%  perf-sched.wait_time.avg.ms.preempt_schedule_common._cond_resched.exit_mmap.mmput.do_exit
>>>        0.01 ±126%    +319.3%       0.06 ± 53%  perf-sched.wait_time.avg.ms.preempt_schedule_common._cond_resched.kmem_cache_alloc.__anon_vma_prepare.do_anonymous_page
>>>        0.04 ± 15%   +2327.3%       0.89 ± 54%  perf-sched.wait_time.avg.ms.preempt_schedule_common._cond_resched.kmem_cache_alloc.__d_alloc.d_alloc
>>>        0.62 ± 70%   +3428.4%      21.95 ± 35%  perf-sched.wait_time.avg.ms.preempt_schedule_common._cond_resched.kmem_cache_alloc.__d_alloc.d_alloc_cursor
>>>        0.29 ± 72%   +1816.0%       5.56 ± 89%  perf-sched.wait_time.avg.ms.preempt_schedule_common._cond_resched.kmem_cache_alloc.copy_fs_struct.copy_process
>>>        0.05 ±119%   +1698.6%       0.95 ± 91%  perf-sched.wait_time.avg.ms.preempt_schedule_common._cond_resched.kmem_cache_alloc.getname_flags.do_sys_openat2
>>>        0.18 ± 72%    +533.7%       1.16 ± 31%  perf-sched.wait_time.avg.ms.preempt_schedule_common._cond_resched.kmem_cache_alloc.getname_flags.user_path_at_empty
>>>        0.41 ± 34%    +890.7%       4.09 ± 49%  perf-sched.wait_time.avg.ms.preempt_schedule_common._cond_resched.kmem_cache_alloc.prepare_creds.copy_creds
>>>        3.70 ± 18%     +55.5%       5.76 ± 42%  perf-sched.wait_time.avg.ms.preempt_schedule_common._cond_resched.kmem_cache_alloc.vm_area_dup.dup_mmap
>>>       10.64 ±105%    +189.5%      30.81 ± 66%  perf-sched.wait_time.avg.ms.preempt_schedule_common._cond_resched.mnt_want_write.path_openat.do_filp_open
>>>        0.10 ± 68%    -100.0%       0.00        perf-sched.wait_time.avg.ms.preempt_schedule_common._cond_resched.mutex_lock_killable.begin_new_exec.load_elf_binary
>>>        0.05 ± 97%   +1166.8%       0.58 ± 67%  perf-sched.wait_time.avg.ms.preempt_schedule_common._cond_resched.remove_vma.__do_munmap.__vm_munmap
>>>        5.09           +14.7%       5.83 ±  3%  perf-sched.wait_time.avg.ms.preempt_schedule_common._cond_resched.stop_one_cpu.__set_cpus_allowed_ptr.sched_setaffinity
>>>        2.90 ± 20%    -100.0%       0.00        perf-sched.wait_time.avg.ms.preempt_schedule_common._cond_resched.stop_one_cpu.migrate_task_to.task_numa_migrate
>>>        0.03 ±  5%    +539.8%       0.20 ± 26%  perf-sched.wait_time.avg.ms.preempt_schedule_common._cond_resched.stop_one_cpu.sched_exec.bprm_execve
>>>        2.38 ± 18%    -100.0%       0.00        perf-sched.wait_time.avg.ms.preempt_schedule_common._cond_resched.wait_for_completion.stop_two_cpus.migrate_swap
>>>        0.23 ±  2%     +87.1%       0.44 ± 35%  perf-sched.wait_time.avg.ms.preempt_schedule_common._cond_resched.zap_pte_range.unmap_page_range.unmap_vmas
>>>        2.62 ± 44%    -100.0%       0.00        perf-sched.wait_time.avg.ms.rwsem_down_read_slowpath.page_lock_anon_vma_read.rmap_walk_anon.try_to_unmap
>>>        2.41 ± 41%    -100.0%       0.00        perf-sched.wait_time.avg.ms.rwsem_down_read_slowpath.rmap_walk_anon.remove_migration_ptes.migrate_pages
>>>        0.17 ±  4%     +96.1%       0.34 ± 48%  perf-sched.wait_time.avg.ms.rwsem_down_write_slowpath.__vma_adjust.__split_vma.mprotect_fixup
>>>        1.31 ± 21%    +398.6%       6.54 ±  7%  perf-sched.wait_time.avg.ms.rwsem_down_write_slowpath.do_unlinkat.do_syscall_64.entry_SYSCALL_64_after_hwframe
>>>        0.93 ±  7%     +64.0%       1.53 ± 16%  perf-sched.wait_time.avg.ms.rwsem_down_write_slowpath.dup_mmap.dup_mm.copy_process
>>>        0.17 ± 14%     +31.3%       0.22 ± 13%  perf-sched.wait_time.avg.ms.rwsem_down_write_slowpath.unlink_file_vma.free_pgtables.unmap_region
>>>        0.20 ± 15%     +83.9%       0.37 ± 25%  perf-sched.wait_time.avg.ms.rwsem_down_write_slowpath.vma_link.mmap_region.do_mmap
>>>      251.12 ±  7%     -36.4%     159.75 ± 18%  perf-sched.wait_time.avg.ms.schedule_hrtimeout_range_clock.ep_poll.do_epoll_wait.__x64_sys_epoll_wait
>>>       87.92 ± 14%     +50.6%     132.37 ± 23%  perf-sched.wait_time.avg.ms.schedule_hrtimeout_range_clock.poll_schedule_timeout.constprop.0.do_sys_poll
>>>      383.96           -32.9%     257.59 ±  9%  perf-sched.wait_time.avg.ms.smpboot_thread_fn.kthread.ret_from_fork
>>>      142.28           -31.7%      97.22 ±  6%  perf-sched.wait_time.avg.ms.worker_thread.kthread.ret_from_fork
>>>        8.16 ± 46%    +383.8%      39.47 ± 75%  perf-sched.wait_time.max.ms.d_alloc_parallel.__lookup_slow.walk_component.link_path_walk.part
>>>        5512           -30.9%       3811        perf-sched.wait_time.max.ms.devkmsg_read.vfs_read.ksys_read.do_syscall_64
>>>        5512           -30.9%       3811        perf-sched.wait_time.max.ms.do_syslog.part.0.kmsg_read.vfs_read
>>>        5496           -30.6%       3813        perf-sched.wait_time.max.ms.exit_to_user_mode_prepare.syscall_exit_to_user_mode.entry_SYSCALL_64_after_hwframe.[unknown]
>>>        0.07 ±107%  +29156.2%      21.16 ±159%  perf-sched.wait_time.max.ms.io_schedule.__lock_page.__do_fault.do_fault
>>>        5580           -30.1%       3899        perf-sched.wait_time.max.ms.pipe_read.new_sync_read.vfs_read.ksys_read
>>>        4.62 ± 40%    +343.5%      20.48 ± 14%  perf-sched.wait_time.max.ms.preempt_schedule_common._cond_resched.__kmalloc.security_task_alloc.copy_process
>>>        2.87 ± 85%    +498.7%      17.16 ± 34%  perf-sched.wait_time.max.ms.preempt_schedule_common._cond_resched.__put_anon_vma.unlink_anon_vmas.free_pgtables
>>>        1.03 ±140%    +525.1%       6.42 ± 54%  perf-sched.wait_time.max.ms.preempt_schedule_common._cond_resched.cgroup_can_fork.copy_process.kernel_clone
>>>        0.85 ± 22%   +2220.1%      19.71 ± 82%  perf-sched.wait_time.max.ms.preempt_schedule_common._cond_resched.change_p4d_range.change_protection.mprotect_fixup
>>>        7.98 ± 62%    +586.9%      54.83 ±100%  perf-sched.wait_time.max.ms.preempt_schedule_common._cond_resched.copy_pte_range.copy_page_range.dup_mmap
>>>        0.42 ± 73%    +476.2%       2.44 ± 95%  perf-sched.wait_time.max.ms.preempt_schedule_common._cond_resched.copy_strings.isra.0
>>>        0.18 ±131%   +4103.5%       7.45 ± 88%  perf-sched.wait_time.max.ms.preempt_schedule_common._cond_resched.down_read.path_openat.do_filp_open
>>>        2.68 ± 50%   +1905.1%      53.65 ±123%  perf-sched.wait_time.max.ms.preempt_schedule_common._cond_resched.down_read.walk_component.link_path_walk
>>>        0.04 ±141%   +3772.9%       1.39 ±166%  perf-sched.wait_time.max.ms.preempt_schedule_common._cond_resched.down_read_killable.create_elf_tables.isra
>>>        0.03 ±122%  +22698.3%       6.76 ± 88%  perf-sched.wait_time.max.ms.preempt_schedule_common._cond_resched.down_write.__anon_vma_prepare.do_fault
>>>        0.25 ±103%   +3843.4%       9.87 ± 62%  perf-sched.wait_time.max.ms.preempt_schedule_common._cond_resched.down_write.__vma_adjust.shift_arg_pages
>>>       23.74 ± 13%   +1153.9%     297.72 ±136%  perf-sched.wait_time.max.ms.preempt_schedule_common._cond_resched.down_write.anon_vma_clone.anon_vma_fork
>>>        8.46 ± 76%   +3275.9%     285.49 ±146%  perf-sched.wait_time.max.ms.preempt_schedule_common._cond_resched.down_write.anon_vma_fork.dup_mmap
>>>        8.55 ± 14%   +2664.7%     236.39 ± 90%  perf-sched.wait_time.max.ms.preempt_schedule_common._cond_resched.down_write.do_truncate.path_openat
>>>        0.02 ±141%   +3120.9%       0.78 ± 90%  perf-sched.wait_time.max.ms.preempt_schedule_common._cond_resched.down_write_killable.__vm_munmap.elf_map
>>>        0.63 ± 47%    +456.1%       3.51 ± 76%  perf-sched.wait_time.max.ms.preempt_schedule_common._cond_resched.down_write_killable.vm_mmap_pgoff.elf_map
>>>        0.36 ±139%    +758.2%       3.10 ± 66%  perf-sched.wait_time.max.ms.preempt_schedule_common._cond_resched.dput.dcache_dir_close.__fput
>>>        1.91 ± 89%   +6827.3%     132.20 ±137%  perf-sched.wait_time.max.ms.preempt_schedule_common._cond_resched.dput.do_unlinkat.do_syscall_64
>>>      336.18 ±140%     +73.0%     581.60 ± 78%  perf-sched.wait_time.max.ms.preempt_schedule_common._cond_resched.dput.done_path_create.do_linkat
>>>      341.13 ±137%    +196.8%       1012        perf-sched.wait_time.max.ms.preempt_schedule_common._cond_resched.dput.step_into.path_openat
>>>        9.64 ± 52%   +4704.9%     463.29 ± 72%  perf-sched.wait_time.max.ms.preempt_schedule_common._cond_resched.dput.terminate_walk.path_openat
>>>        0.94 ± 71%   +1718.2%      17.05 ± 33%  perf-sched.wait_time.max.ms.preempt_schedule_common._cond_resched.exit_mmap.mmput.begin_new_exec
>>>        6.94 ± 41%    +683.1%      54.36 ±113%  perf-sched.wait_time.max.ms.preempt_schedule_common._cond_resched.exit_mmap.mmput.do_exit
>>>        0.07 ±119%    +888.4%       0.65 ± 80%  perf-sched.wait_time.max.ms.preempt_schedule_common._cond_resched.kmem_cache_alloc.__anon_vma_prepare.do_anonymous_page
>>>        0.42 ± 51%   +3060.9%      13.37 ± 71%  perf-sched.wait_time.max.ms.preempt_schedule_common._cond_resched.kmem_cache_alloc.__d_alloc.d_alloc
>>>        1.61 ± 71%  +29984.9%     484.07 ± 65%  perf-sched.wait_time.max.ms.preempt_schedule_common._cond_resched.kmem_cache_alloc.__d_alloc.d_alloc_cursor
>>>        1.01 ± 76%    +779.3%       8.90 ± 68%  perf-sched.wait_time.max.ms.preempt_schedule_common._cond_resched.kmem_cache_alloc.copy_fs_struct.copy_process
>>>        1.25 ± 73%    +327.5%       5.36 ± 28%  perf-sched.wait_time.max.ms.preempt_schedule_common._cond_resched.kmem_cache_alloc.getname_flags.user_path_at_empty
>>>       23.94 ± 16%    +259.0%      85.92 ± 53%  perf-sched.wait_time.max.ms.preempt_schedule_common._cond_resched.kmem_cache_alloc.vm_area_dup.dup_mmap
>>>        6.01 ± 27%     +52.0%       9.13 ± 24%  perf-sched.wait_time.max.ms.preempt_schedule_common._cond_resched.kmem_cache_alloc_trace.perf_event_mmap.mmap_region
>>>        0.27 ± 77%    -100.0%       0.00        perf-sched.wait_time.max.ms.preempt_schedule_common._cond_resched.mutex_lock_killable.begin_new_exec.load_elf_binary
>>>        0.41 ± 97%   +1470.1%       6.36 ± 97%  perf-sched.wait_time.max.ms.preempt_schedule_common._cond_resched.remove_vma.__do_munmap.__vm_munmap
>>>        2.81 ± 79%    +146.3%       6.92 ± 44%  perf-sched.wait_time.max.ms.preempt_schedule_common._cond_resched.remove_vma.__do_munmap.mmap_region
>>>        7.51 ± 23%    +226.7%      24.54 ± 98%  perf-sched.wait_time.max.ms.preempt_schedule_common._cond_resched.remove_vma.exit_mmap.mmput
>>>        7.36 ±116%    +497.6%      44.00 ± 94%  perf-sched.wait_time.max.ms.preempt_schedule_common._cond_resched.shmem_undo_range.shmem_truncate_range.shmem_evict_inode
>>>        1495 ± 31%    +114.1%       3203 ±  4%  perf-sched.wait_time.max.ms.preempt_schedule_common._cond_resched.smpboot_thread_fn.kthread.ret_from_fork
>>>        9.21 ± 22%    -100.0%       0.00        perf-sched.wait_time.max.ms.preempt_schedule_common._cond_resched.stop_one_cpu.migrate_task_to.task_numa_migrate
>>>       12.64 ±  2%    +300.8%      50.66 ± 43%  perf-sched.wait_time.max.ms.preempt_schedule_common._cond_resched.stop_one_cpu.sched_exec.bprm_execve
>>>        7.61 ± 29%    -100.0%       0.00        perf-sched.wait_time.max.ms.preempt_schedule_common._cond_resched.wait_for_completion.stop_two_cpus.migrate_swap
>>>       20.69 ± 50%    -100.0%       0.00        perf-sched.wait_time.max.ms.rwsem_down_read_slowpath.page_lock_anon_vma_read.rmap_walk_anon.try_to_unmap
>>>       19.13 ± 36%    -100.0%       0.00        perf-sched.wait_time.max.ms.rwsem_down_read_slowpath.rmap_walk_anon.remove_migration_ptes.migrate_pages
>>>        2.03 ± 55%     -96.2%       0.08 ±151%  perf-sched.wait_time.max.ms.rwsem_down_write_slowpath.__put_anon_vma.unlink_anon_vmas.free_pgtables
>>>        7.53 ± 25%    +397.1%      37.43 ± 76%  perf-sched.wait_time.max.ms.rwsem_down_write_slowpath.__vma_adjust.__split_vma.mprotect_fixup
>>>       31.44 ± 23%   +1888.2%     625.17 ± 62%  perf-sched.wait_time.max.ms.rwsem_down_write_slowpath.dup_mmap.dup_mm.copy_process
>>>        2.32 ± 99%     -88.3%       0.27 ± 87%  perf-sched.wait_time.max.ms.rwsem_down_write_slowpath.unlink_anon_vmas.free_pgtables.exit_mmap
>>>       20.37 ± 28%    +184.7%      57.99 ± 80%  perf-sched.wait_time.max.ms.rwsem_down_write_slowpath.unlink_file_vma.free_pgtables.exit_mmap
>>>       11.76 ± 17%    +170.3%      31.80 ± 43%  perf-sched.wait_time.max.ms.rwsem_down_write_slowpath.vma_link.mmap_region.do_mmap
>>>        5176           -31.0%       3569        perf-sched.wait_time.max.ms.schedule_hrtimeout_range_clock.ep_poll.do_epoll_wait.__x64_sys_epoll_wait
>>>        5585           -30.2%       3899        perf-sched.wait_time.max.ms.schedule_hrtimeout_range_clock.poll_schedule_timeout.constprop.0.do_select
>>>      505.14           +15.5%     583.69 ±  5%  perf-sched.wait_time.max.ms.schedule_timeout.kcompactd.kthread.ret_from_fork
>>>        4816 ± 15%     -31.5%       3297 ±  3%  perf-sched.wait_time.max.ms.smpboot_thread_fn.kthread.ret_from_fork
>>>        1791           -20.7%       1420 ±  3%  perf-sched.wait_time.max.ms.worker_thread.kthread.ret_from_fork
>>>       29455 ±  2%     +33.9%      39449 ±  7%  softirqs.CPU0.NET_RX
>>>       25470           +32.0%      33628        softirqs.CPU0.SCHED
>>>       26529 ± 11%     +43.7%      38110 ±  9%  softirqs.CPU1.NET_RX
>>>       22797 ±  3%     +43.8%      32782 ±  3%  softirqs.CPU1.SCHED
>>>       28371 ±  3%     +54.2%      43762 ± 13%  softirqs.CPU10.NET_RX
>>>       22629 ±  2%     +43.4%      32439 ±  2%  softirqs.CPU10.SCHED
>>>       27171 ±  4%     +38.4%      37593 ±  9%  softirqs.CPU100.NET_RX
>>>       23180 ±  2%     +39.1%      32248        softirqs.CPU100.SCHED
>>>       28196 ±  2%     +34.4%      37895 ±  5%  softirqs.CPU101.NET_RX
>>>       23100           +40.7%      32511 ±  2%  softirqs.CPU101.SCHED
>>>       27883 ±  3%     +43.5%      40015 ± 11%  softirqs.CPU102.NET_RX
>>>       22937 ±  2%     +41.4%      32444 ±  2%  softirqs.CPU102.SCHED
>>>       28037           +40.7%      39461 ±  7%  softirqs.CPU103.NET_RX
>>>       23000 ±  2%     +41.0%      32435 ±  2%  softirqs.CPU103.SCHED
>>>       28345           +34.9%      38229 ±  9%  softirqs.CPU104.NET_RX
>>>       22314 ±  3%     +45.3%      32417 ±  2%  softirqs.CPU104.SCHED
>>>       27097 ±  2%     +40.7%      38114 ±  3%  softirqs.CPU105.NET_RX
>>>       22615 ±  2%     +42.8%      32294 ±  2%  softirqs.CPU105.SCHED
>>>       26463 ±  3%     +50.9%      39943 ± 12%  softirqs.CPU106.NET_RX
>>>       22730 ±  2%     +42.2%      32319 ±  2%  softirqs.CPU106.SCHED
>>>       27957 ±  2%     +29.0%      36058 ±  6%  softirqs.CPU107.NET_RX
>>>       22901 ±  2%     +41.3%      32362 ±  2%  softirqs.CPU107.SCHED
>>>       29320 ±  8%     +33.4%      39101 ± 10%  softirqs.CPU108.NET_RX
>>>       22873 ±  2%     +41.2%      32301 ±  2%  softirqs.CPU108.SCHED
>>>       29106 ±  2%     +31.5%      38266 ±  3%  softirqs.CPU109.NET_RX
>>>       22591 ±  2%     +43.1%      32336        softirqs.CPU109.SCHED
>>>       28741 ±  5%     +34.6%      38676 ±  6%  softirqs.CPU11.NET_RX
>>>       22399 ±  3%     +44.5%      32364 ±  2%  softirqs.CPU11.SCHED
>>>       26804 ±  5%     +41.8%      38006 ±  5%  softirqs.CPU110.NET_RX
>>>       22766 ±  2%     +43.3%      32619 ±  2%  softirqs.CPU110.SCHED
>>>       25623 ±  7%     +45.9%      37393 ±  9%  softirqs.CPU111.NET_RX
>>>       22586 ±  2%     +43.4%      32391 ±  2%  softirqs.CPU111.SCHED
>>>       28020 ±  6%     +31.8%      36917 ±  7%  softirqs.CPU112.NET_RX
>>>       22741           +42.2%      32349 ±  2%  softirqs.CPU112.SCHED
>>>       28310 ±  7%     +35.1%      38234 ± 12%  softirqs.CPU113.NET_RX
>>>       22884 ±  2%     +42.4%      32577 ±  2%  softirqs.CPU113.SCHED
>>>       25668 ±  5%     +45.9%      37457 ±  7%  softirqs.CPU114.NET_RX
>>>       22706 ±  3%     +41.7%      32178        softirqs.CPU114.SCHED
>>>       28637 ±  5%     +37.6%      39418 ±  6%  softirqs.CPU115.NET_RX
>>>       22796           +41.0%      32139        softirqs.CPU115.SCHED
>>>       26953 ±  7%     +42.1%      38291 ±  7%  softirqs.CPU116.NET_RX
>>>       22646           +42.1%      32171 ±  2%  softirqs.CPU116.SCHED
>>>       28477 ±  8%     +35.4%      38544 ±  3%  softirqs.CPU117.NET_RX
>>>       22866 ±  2%     +40.8%      32189 ±  2%  softirqs.CPU117.SCHED
>>>       27445 ±  7%     +40.9%      38675 ± 11%  softirqs.CPU118.NET_RX
>>>       22902           +41.4%      32386 ±  2%  softirqs.CPU118.SCHED
>>>       26884 ±  4%     +38.3%      37182 ±  6%  softirqs.CPU119.NET_RX
>>>       22721 ±  2%     +42.0%      32265 ±  2%  softirqs.CPU119.SCHED
>>>       28376 ±  9%     +36.6%      38767 ±  9%  softirqs.CPU12.NET_RX
>>>       22706 ±  3%     +40.4%      31877 ±  2%  softirqs.CPU12.SCHED
>>>       27315 ±  7%     +45.1%      39644 ±  4%  softirqs.CPU120.NET_RX
>>>       23157 ±  2%     +38.9%      32175        softirqs.CPU120.SCHED
>>>       26387 ±  5%     +51.6%      40015 ±  4%  softirqs.CPU121.NET_RX
>>>       22856           +38.5%      31646        softirqs.CPU121.SCHED
>>>       25176 ±  7%     +67.6%      42194 ±  4%  softirqs.CPU122.NET_RX
>>>       22775           +40.3%      31962        softirqs.CPU122.SCHED
>>>       25636 ±  5%     +60.8%      41225 ±  7%  softirqs.CPU123.NET_RX
>>>       23322           +37.7%      32124        softirqs.CPU123.SCHED
>>>       29777 ±  2%     +37.4%      40911 ±  3%  softirqs.CPU124.NET_RX
>>>       23141 ±  2%     +38.7%      32098        softirqs.CPU124.SCHED
>>>       25720 ±  7%     +53.0%      39356 ±  4%  softirqs.CPU125.NET_RX
>>>       23140 ±  3%     +38.7%      32104        softirqs.CPU125.SCHED
>>>       27106 ±  2%     +44.7%      39214 ±  2%  softirqs.CPU126.NET_RX
>>>       33141 ±  3%     +11.4%      36924 ±  7%  softirqs.CPU126.RCU
>>>       23154 ±  2%     +38.6%      32083        softirqs.CPU126.SCHED
>>>       25748 ±  4%     +53.8%      39601 ±  8%  softirqs.CPU127.NET_RX
>>>       23011 ±  2%     +39.4%      32068        softirqs.CPU127.SCHED
>>>       28647 ± 11%     +42.1%      40718        softirqs.CPU128.NET_RX
>>>       34186 ±  3%     +11.6%      38137 ±  7%  softirqs.CPU128.RCU
>>>       23173 ±  2%     +38.2%      32016        softirqs.CPU128.SCHED
>>>       27101 ±  3%     +48.6%      40259 ±  5%  softirqs.CPU129.NET_RX
>>>       22818           +40.4%      32045        softirqs.CPU129.SCHED
>>>       31534 ± 12%     +25.9%      39696 ± 10%  softirqs.CPU13.NET_RX
>>>       22674 ±  2%     +41.6%      32108 ±  2%  softirqs.CPU13.SCHED
>>>       27452           +48.9%      40871 ±  2%  softirqs.CPU130.NET_RX
>>>       23203           +37.1%      31807 ±  2%  softirqs.CPU130.SCHED
>>>       26896 ±  3%     +51.9%      40860 ±  4%  softirqs.CPU131.NET_RX
>>>       23148 ±  2%     +37.4%      31814        softirqs.CPU131.SCHED
>>>       28393 ±  3%     +41.2%      40084 ±  7%  softirqs.CPU132.NET_RX
>>>       23200           +36.7%      31706        softirqs.CPU132.SCHED
>>>       26482 ±  3%     +54.4%      40891 ±  3%  softirqs.CPU133.NET_RX
>>>       22970           +41.0%      32382 ±  2%  softirqs.CPU133.SCHED
>>>       26774 ±  7%     +44.7%      38735 ±  5%  softirqs.CPU134.NET_RX
>>>       33646           +10.7%      37240 ±  7%  softirqs.CPU134.RCU
>>>       23088 ±  2%     +38.6%      32008        softirqs.CPU134.SCHED
>>>       26666           +54.1%      41081 ±  7%  softirqs.CPU135.NET_RX
>>>       33456 ±  4%     +14.0%      38141 ±  8%  softirqs.CPU135.RCU
>>>       22804           +40.0%      31925        softirqs.CPU135.SCHED
>>>       26562 ±  5%     +55.4%      41276 ±  7%  softirqs.CPU136.NET_RX
>>>       23134 ±  3%     +38.9%      32141        softirqs.CPU136.SCHED
>>>       25297 ±  4%     +62.9%      41210 ±  6%  softirqs.CPU137.NET_RX
>>>       22846 ±  2%     +39.9%      31955        softirqs.CPU137.SCHED
>>>       26519 ±  6%     +51.6%      40192 ±  4%  softirqs.CPU138.NET_RX
>>>       22478 ±  2%     +41.7%      31858        softirqs.CPU138.SCHED
>>>       26079 ±  9%     +62.0%      42258 ±  2%  softirqs.CPU139.NET_RX
>>>       22994 ±  3%     +39.2%      32007        softirqs.CPU139.SCHED
>>>       26936 ±  2%     +43.0%      38512 ±  9%  softirqs.CPU14.NET_RX
>>>       22604 ±  3%     +42.3%      32166 ±  3%  softirqs.CPU14.SCHED
>>>       27625 ±  9%     +44.2%      39847 ±  5%  softirqs.CPU140.NET_RX
>>>       33066 ±  5%     +14.6%      37895 ±  8%  softirqs.CPU140.RCU
>>>       22609           +41.6%      32006        softirqs.CPU140.SCHED
>>>       26460 ±  5%     +58.1%      41826 ±  5%  softirqs.CPU141.NET_RX
>>>       34319 ±  4%      +9.6%      37601 ±  8%  softirqs.CPU141.RCU
>>>       22848 ±  2%     +39.5%      31868        softirqs.CPU141.SCHED
>>>       27385 ±  8%     +52.1%      41645 ±  6%  softirqs.CPU142.NET_RX
>>>       22881 ±  2%     +40.0%      32033        softirqs.CPU142.SCHED
>>>       27724 ±  3%     +46.8%      40700 ±  5%  softirqs.CPU143.NET_RX
>>>       34109 ±  4%     +10.8%      37789 ±  8%  softirqs.CPU143.RCU
>>>       23092 ±  2%     +39.1%      32110        softirqs.CPU143.SCHED
>>>       28902 ±  8%     +30.5%      37711        softirqs.CPU144.NET_RX
>>>       34885 ±  7%      +8.2%      37745 ±  7%  softirqs.CPU144.RCU
>>>       22671           +40.8%      31929        softirqs.CPU144.SCHED
>>>       29181 ± 10%     +43.3%      41814 ±  7%  softirqs.CPU145.NET_RX
>>>       22655           +38.5%      31371 ±  2%  softirqs.CPU145.SCHED
>>>       28250 ±  7%     +33.9%      37815 ±  5%  softirqs.CPU146.NET_RX
>>>       22502           +42.7%      32115        softirqs.CPU146.SCHED
>>>       26352 ±  5%     +44.3%      38033 ±  4%  softirqs.CPU147.NET_RX
>>>       34098            +9.1%      37212 ±  5%  softirqs.CPU147.RCU
>>>       22712           +39.8%      31759        softirqs.CPU147.SCHED
>>>       26457 ±  5%     +47.7%      39077 ±  2%  softirqs.CPU148.NET_RX
>>>       34771           +10.8%      38542 ±  5%  softirqs.CPU148.RCU
>>>       22626           +41.1%      31925        softirqs.CPU148.SCHED
>>>       27813 ±  3%     +38.2%      38431 ±  3%  softirqs.CPU149.NET_RX
>>>       34785 ±  3%      +9.4%      38071 ±  6%  softirqs.CPU149.RCU
>>>       22830           +39.5%      31857        softirqs.CPU149.SCHED
>>>       27674           +35.0%      37356 ± 10%  softirqs.CPU15.NET_RX
>>>       22708 ±  3%     +40.8%      31984 ±  2%  softirqs.CPU15.SCHED
>>>       28914 ±  4%     +38.6%      40073 ±  5%  softirqs.CPU150.NET_RX
>>>       22867           +40.2%      32054        softirqs.CPU150.SCHED
>>>       26654           +46.9%      39145 ±  4%  softirqs.CPU151.NET_RX
>>>       35269 ±  5%      +7.1%      37772 ±  5%  softirqs.CPU151.RCU
>>>       22567           +41.7%      31987        softirqs.CPU151.SCHED
>>>       27340 ±  3%     +35.5%      37050 ±  7%  softirqs.CPU152.NET_RX
>>>       22561           +41.1%      31833        softirqs.CPU152.SCHED
>>>       25682 ±  4%     +52.9%      39264 ±  3%  softirqs.CPU153.NET_RX
>>>       33854 ±  3%     +13.6%      38450 ±  5%  softirqs.CPU153.RCU
>>>       22484           +42.2%      31972        softirqs.CPU153.SCHED
>>>       27824 ±  6%     +45.0%      40343        softirqs.CPU154.NET_RX
>>>       34996 ±  2%      +9.1%      38165 ±  5%  softirqs.CPU154.RCU
>>>       22409           +41.5%      31717        softirqs.CPU154.SCHED
>>>       26948 ±  6%     +51.9%      40940 ±  3%  softirqs.CPU155.NET_RX
>>>       35069           +10.6%      38779 ±  5%  softirqs.CPU155.RCU
>>>       22696           +42.2%      32266        softirqs.CPU155.SCHED
>>>       26787 ±  3%     +45.7%      39032 ±  4%  softirqs.CPU156.NET_RX
>>>       22583           +41.3%      31902        softirqs.CPU156.SCHED
>>>       26065 ±  2%     +47.8%      38529 ±  4%  softirqs.CPU157.NET_RX
>>>       22442           +42.0%      31866        softirqs.CPU157.SCHED
>>>       27310 ±  3%     +42.2%      38826 ±  8%  softirqs.CPU158.NET_RX
>>>       22593           +40.5%      31752        softirqs.CPU158.SCHED
>>>       26411 ±  4%     +47.7%      39018        softirqs.CPU159.NET_RX
>>>       22535           +42.5%      32103        softirqs.CPU159.SCHED
>>>       28586 ±  3%     +30.9%      37421 ±  8%  softirqs.CPU16.NET_RX
>>>       22668 ±  3%     +45.0%      32879 ±  3%  softirqs.CPU16.SCHED
>>>       27341 ±  2%     +39.4%      38105 ±  2%  softirqs.CPU160.NET_RX
>>>       22494           +41.5%      31829        softirqs.CPU160.SCHED
>>>       26309           +47.7%      38870 ±  6%  softirqs.CPU161.NET_RX
>>>       22652           +41.1%      31964        softirqs.CPU161.SCHED
>>>       28102 ±  3%     +35.3%      38031 ±  3%  softirqs.CPU162.NET_RX
>>>       34995 ±  4%      +6.5%      37255 ±  7%  softirqs.CPU162.RCU
>>>       22800           +40.4%      32000        softirqs.CPU162.SCHED
>>>       24958 ±  5%     +53.9%      38412 ±  7%  softirqs.CPU163.NET_RX
>>>       22582           +42.1%      32079        softirqs.CPU163.SCHED
>>>       27327 ±  2%     +36.9%      37415 ±  4%  softirqs.CPU164.NET_RX
>>>       22516           +43.0%      32204        softirqs.CPU164.SCHED
>>>       27456 ±  5%     +46.4%      40188 ±  4%  softirqs.CPU165.NET_RX
>>>       22509           +42.6%      32104        softirqs.CPU165.SCHED
>>>       26840 ±  3%     +42.0%      38118 ±  5%  softirqs.CPU166.NET_RX
>>>       22847           +40.1%      32002        softirqs.CPU166.SCHED
>>>       28288 ±  2%     +45.7%      41208 ±  8%  softirqs.CPU167.NET_RX
>>>       22656           +43.6%      32529 ±  2%  softirqs.CPU167.SCHED
>>>       27845           +40.0%      38982 ±  5%  softirqs.CPU168.NET_RX
>>>       22861           +39.9%      31984        softirqs.CPU168.SCHED
>>>       26590 ±  8%     +55.4%      41316 ±  5%  softirqs.CPU169.NET_RX
>>>       22774           +40.5%      31990        softirqs.CPU169.SCHED
>>>       27965 ±  4%     +34.4%      37581 ±  6%  softirqs.CPU17.NET_RX
>>>       22416           +43.0%      32064 ±  3%  softirqs.CPU17.SCHED
>>>       29095 ±  2%     +43.8%      41848 ±  4%  softirqs.CPU170.NET_RX
>>>       22753           +41.9%      32295        softirqs.CPU170.SCHED
>>>       25289           +56.3%      39538 ±  4%  softirqs.CPU171.NET_RX
>>>       33179 ±  2%      +7.3%      35615 ±  7%  softirqs.CPU171.RCU
>>>       22899           +40.5%      32171        softirqs.CPU171.SCHED
>>>       25293 ±  3%     +60.0%      40480 ±  7%  softirqs.CPU172.NET_RX
>>>       22584           +42.6%      32196        softirqs.CPU172.SCHED
>>>       27200 ±  5%     +50.1%      40831 ±  4%  softirqs.CPU173.NET_RX
>>>       22804           +41.1%      32183        softirqs.CPU173.SCHED
>>>       27779 ±  3%     +49.3%      41464 ±  7%  softirqs.CPU174.NET_RX
>>>       22569           +42.0%      32056        softirqs.CPU174.SCHED
>>>       26980 ±  6%     +49.2%      40261 ±  5%  softirqs.CPU175.NET_RX
>>>       22818           +41.2%      32220        softirqs.CPU175.SCHED
>>>       27496 ±  6%     +48.8%      40909 ±  9%  softirqs.CPU176.NET_RX
>>>       22818           +40.9%      32139        softirqs.CPU176.SCHED
>>>       27197           +42.7%      38817 ±  8%  softirqs.CPU177.NET_RX
>>>       22770           +40.4%      31976        softirqs.CPU177.SCHED
>>>       27688 ±  2%     +45.6%      40318 ±  8%  softirqs.CPU178.NET_RX
>>>       22820           +39.9%      31937        softirqs.CPU178.SCHED
>>>       26078 ±  4%     +51.3%      39444 ±  5%  softirqs.CPU179.NET_RX
>>>       22716           +40.5%      31912        softirqs.CPU179.SCHED
>>>       27755 ±  2%     +41.2%      39191 ±  8%  softirqs.CPU18.NET_RX
>>>       22621 ±  2%     +41.0%      31900        softirqs.CPU18.SCHED
>>>       28019 ±  5%     +40.3%      39303 ±  9%  softirqs.CPU180.NET_RX
>>>       22549           +41.3%      31869        softirqs.CPU180.SCHED
>>>       27487 ± 10%     +46.9%      40381 ±  7%  softirqs.CPU181.NET_RX
>>>       22776           +40.2%      31943        softirqs.CPU181.SCHED
>>>       26432 ±  2%     +55.0%      40971 ±  7%  softirqs.CPU182.NET_RX
>>>       22815           +40.7%      32105        softirqs.CPU182.SCHED
>>>       27997 ±  4%     +46.5%      41010 ±  4%  softirqs.CPU183.NET_RX
>>>       22875           +39.9%      32004        softirqs.CPU183.SCHED
>>>       24478 ± 14%     +67.2%      40939 ± 10%  softirqs.CPU184.NET_RX
>>>       22192 ±  3%     +45.0%      32177        softirqs.CPU184.SCHED
>>>       27366           +40.4%      38417 ±  8%  softirqs.CPU185.NET_RX
>>>       22879           +40.4%      32119        softirqs.CPU185.SCHED
>>>       28955 ±  4%     +36.2%      39443 ±  6%  softirqs.CPU186.NET_RX
>>>       22929           +39.2%      31913        softirqs.CPU186.SCHED
>>>       26140 ±  5%     +50.3%      39290 ±  8%  softirqs.CPU187.NET_RX
>>>       23163           +38.6%      32110        softirqs.CPU187.SCHED
>>>       26199           +56.2%      40920 ± 11%  softirqs.CPU188.NET_RX
>>>       22814           +41.5%      32276        softirqs.CPU188.SCHED
>>>       27178 ±  5%     +49.7%      40699 ±  3%  softirqs.CPU189.NET_RX
>>>       22673           +41.5%      32083        softirqs.CPU189.SCHED
>>>       27438 ±  7%     +38.7%      38069 ±  5%  softirqs.CPU19.NET_RX
>>>       22583 ±  3%     +42.0%      32066 ±  2%  softirqs.CPU19.SCHED
>>>       26118 ±  9%     +42.6%      37257 ±  8%  softirqs.CPU190.NET_RX
>>>       22621           +41.6%      32022        softirqs.CPU190.SCHED
>>>       25609 ±  5%     +52.5%      39055 ±  9%  softirqs.CPU191.NET_RX
>>>       22510           +40.5%      31618        softirqs.CPU191.SCHED
>>>       28410 ±  5%     +34.1%      38091 ±  5%  softirqs.CPU2.NET_RX
>>>       22190           +43.9%      31937 ±  2%  softirqs.CPU2.SCHED
>>>       28923 ± 10%     +37.5%      39780 ±  9%  softirqs.CPU20.NET_RX
>>>       22749 ±  4%     +41.2%      32113 ±  2%  softirqs.CPU20.SCHED
>>>       27078 ±  3%     +38.2%      37429 ±  8%  softirqs.CPU21.NET_RX
>>>       22477 ±  3%     +42.5%      32029 ±  2%  softirqs.CPU21.SCHED
>>>       26006 ±  9%     +49.7%      38922 ± 11%  softirqs.CPU22.NET_RX
>>>       22743 ±  3%     +41.2%      32108 ±  2%  softirqs.CPU22.SCHED
>>>       28691 ±  2%     +39.6%      40046 ±  5%  softirqs.CPU23.NET_RX
>>>       22668 ±  3%     +41.1%      31976 ±  2%  softirqs.CPU23.SCHED
>>>       32698 ±  4%     +27.5%      41704 ±  3%  softirqs.CPU24.NET_RX
>>>       23538           +34.9%      31764        softirqs.CPU24.SCHED
>>>       27793           +44.1%      40038 ±  9%  softirqs.CPU25.NET_RX
>>>       22892 ±  2%     +38.0%      31602        softirqs.CPU25.SCHED
>>>       30671 ±  3%     +36.4%      41822 ±  7%  softirqs.CPU26.NET_RX
>>>       23207 ±  2%     +36.8%      31744        softirqs.CPU26.SCHED
>>>       26281 ±  6%     +57.6%      41417 ±  3%  softirqs.CPU27.NET_RX
>>>       22758 ±  2%     +40.4%      31952        softirqs.CPU27.SCHED
>>>       26713 ±  5%     +54.6%      41300 ±  4%  softirqs.CPU28.NET_RX
>>>       33769 ±  2%     +11.3%      37593 ±  7%  softirqs.CPU28.RCU
>>>       22857 ±  3%     +38.2%      31586        softirqs.CPU28.SCHED
>>>       27219 ±  3%     +53.1%      41664 ±  8%  softirqs.CPU29.NET_RX
>>>       22988 ±  3%     +38.2%      31775        softirqs.CPU29.SCHED
>>>       26571 ±  5%     +51.2%      40171 ±  8%  softirqs.CPU3.NET_RX
>>>       23000 ±  3%     +38.9%      31946 ±  3%  softirqs.CPU3.SCHED
>>>       26856 ±  4%     +57.9%      42396 ±  2%  softirqs.CPU30.NET_RX
>>>       22915 ±  3%     +38.0%      31629        softirqs.CPU30.SCHED
>>>       28747 ±  6%     +42.3%      40898 ±  9%  softirqs.CPU31.NET_RX
>>>       22863 ±  2%     +39.4%      31864        softirqs.CPU31.SCHED
>>>       26418 ±  4%     +56.0%      41211 ±  4%  softirqs.CPU32.NET_RX
>>>       22861 ±  3%     +38.3%      31613        softirqs.CPU32.SCHED
>>>       27379 ± 13%     +52.0%      41619 ±  3%  softirqs.CPU33.NET_RX
>>>       22743           +38.7%      31538        softirqs.CPU33.SCHED
>>>       27331 ±  6%     +47.9%      40432 ±  7%  softirqs.CPU34.NET_RX
>>>       23071 ±  2%     +37.5%      31730        softirqs.CPU34.SCHED
>>>       27997 ±  9%     +49.7%      41922 ±  4%  softirqs.CPU35.NET_RX
>>>       34868 ±  3%     +10.6%      38580 ±  7%  softirqs.CPU35.RCU
>>>       23029 ±  3%     +37.4%      31635        softirqs.CPU35.SCHED
>>>       26763 ±  4%     +59.4%      42657 ±  6%  softirqs.CPU36.NET_RX
>>>       22829           +38.7%      31657        softirqs.CPU36.SCHED
>>>       27252           +48.6%      40489 ±  3%  softirqs.CPU37.NET_RX
>>>       22546 ±  2%     +40.2%      31603        softirqs.CPU37.SCHED
>>>       26072 ±  9%     +67.4%      43643 ±  7%  softirqs.CPU38.NET_RX
>>>       22711 ±  4%     +39.6%      31710        softirqs.CPU38.SCHED
>>>       27579 ±  3%     +57.8%      43518 ±  5%  softirqs.CPU39.NET_RX
>>>       22705 ±  2%     +39.2%      31596        softirqs.CPU39.SCHED
>>>       28570 ±  3%     +30.7%      37345 ±  7%  softirqs.CPU4.NET_RX
>>>       22763 ±  3%     +41.4%      32189 ±  3%  softirqs.CPU4.SCHED
>>>       26544 ±  5%     +52.0%      40348 ±  5%  softirqs.CPU40.NET_RX
>>>       22746 ±  3%     +39.5%      31726        softirqs.CPU40.SCHED
>>>       28165 ±  5%     +47.1%      41432 ±  3%  softirqs.CPU41.NET_RX
>>>       22791 ±  3%     +38.8%      31624        softirqs.CPU41.SCHED
>>>       28718 ±  6%     +43.4%      41176 ±  8%  softirqs.CPU42.NET_RX
>>>       22603           +39.3%      31493        softirqs.CPU42.SCHED
>>>       27257 ±  4%     +54.4%      42085 ±  4%  softirqs.CPU43.NET_RX
>>>       22763 ±  3%     +39.4%      31742        softirqs.CPU43.SCHED
>>>       26604 ±  3%     +58.0%      42041 ±  5%  softirqs.CPU44.NET_RX
>>>       33772 ±  5%     +13.2%      38218 ±  7%  softirqs.CPU44.RCU
>>>       22764 ±  2%     +39.6%      31790        softirqs.CPU44.SCHED
>>>       26786 ±  4%     +54.4%      41361 ±  6%  softirqs.CPU45.NET_RX
>>>       34271 ±  3%     +10.2%      37764 ±  8%  softirqs.CPU45.RCU
>>>       22813 ±  3%     +38.4%      31565        softirqs.CPU45.SCHED
>>>       28082 ±  5%     +43.0%      40154 ±  7%  softirqs.CPU46.NET_RX
>>>       34703 ±  4%      +9.0%      37818 ±  6%  softirqs.CPU46.RCU
>>>       22800 ±  4%     +39.4%      31785        softirqs.CPU46.SCHED
>>>       28229 ±  2%     +49.3%      42156 ±  6%  softirqs.CPU47.NET_RX
>>>       35007 ±  3%      +9.5%      38323 ±  7%  softirqs.CPU47.RCU
>>>       22878 ±  2%     +38.2%      31613        softirqs.CPU47.SCHED
>>>       29087 ±  4%     +39.9%      40686 ±  3%  softirqs.CPU48.NET_RX
>>>       22510           +40.5%      31621        softirqs.CPU48.SCHED
>>>       27833 ±  5%     +47.3%      40991 ±  6%  softirqs.CPU49.NET_RX
>>>       22119           +42.7%      31572        softirqs.CPU49.SCHED
>>>       28242           +35.9%      38374 ±  6%  softirqs.CPU5.NET_RX
>>>       22636 ±  3%     +42.1%      32161 ±  2%  softirqs.CPU5.SCHED
>>>       28615 ±  4%     +37.2%      39250 ±  5%  softirqs.CPU50.NET_RX
>>>       34409           +12.2%      38623 ±  5%  softirqs.CPU50.RCU
>>>       22248           +41.5%      31482        softirqs.CPU50.SCHED
>>>       28739 ±  5%     +40.6%      40405        softirqs.CPU51.NET_RX
>>>       22458 ±  2%     +41.1%      31679        softirqs.CPU51.SCHED
>>>       28428 ±  6%     +34.7%      38284        softirqs.CPU52.NET_RX
>>>       22308           +41.7%      31601        softirqs.CPU52.SCHED
>>>       29032 ±  6%     +33.6%      38786 ±  7%  softirqs.CPU53.NET_RX
>>>       34502 ±  5%     +10.0%      37943 ±  8%  softirqs.CPU53.RCU
>>>       22408           +40.7%      31526        softirqs.CPU53.SCHED
>>>       27640 ±  7%     +42.5%      39380 ±  2%  softirqs.CPU54.NET_RX
>>>       22473           +40.7%      31609        softirqs.CPU54.SCHED
>>>       26943 ±  9%     +42.4%      38374 ±  4%  softirqs.CPU55.NET_RX
>>>       22289 ±  2%     +42.7%      31800        softirqs.CPU55.SCHED
>>>       27400 ±  4%     +40.2%      38409 ±  5%  softirqs.CPU56.NET_RX
>>>       22354           +40.0%      31300        softirqs.CPU56.SCHED
>>>       26999 ±  2%     +46.5%      39554 ±  6%  softirqs.CPU57.NET_RX
>>>       34130 ±  3%     +12.9%      38546 ±  7%  softirqs.CPU57.RCU
>>>       22383           +41.0%      31569        softirqs.CPU57.SCHED
>>>       26684 ±  3%     +50.6%      40177 ±  3%  softirqs.CPU58.NET_RX
>>>       34403 ±  3%     +11.9%      38510 ±  7%  softirqs.CPU58.RCU
>>>       21983           +44.1%      31667        softirqs.CPU58.SCHED
>>>       28353           +41.2%      40037 ±  5%  softirqs.CPU59.NET_RX
>>>       22456           +40.7%      31589        softirqs.CPU59.SCHED
>>>       27801 ± 10%     +34.5%      37398 ±  5%  softirqs.CPU6.NET_RX
>>>       22604 ±  3%     +41.8%      32058 ±  2%  softirqs.CPU6.SCHED
>>>       26348 ±  3%     +55.7%      41023        softirqs.CPU60.NET_RX
>>>       22333           +41.1%      31517        softirqs.CPU60.SCHED
>>>       27592 ±  8%     +43.0%      39449 ±  4%  softirqs.CPU61.NET_RX
>>>       22340           +40.8%      31449        softirqs.CPU61.SCHED
>>>       27991 ±  6%     +39.9%      39173 ±  6%  softirqs.CPU62.NET_RX
>>>       22336           +41.0%      31505        softirqs.CPU62.SCHED
>>>       27979 ±  6%     +42.4%      39848 ±  6%  softirqs.CPU63.NET_RX
>>>       22455           +40.7%      31595        softirqs.CPU63.SCHED
>>>       27343 ±  4%     +47.1%      40224 ±  7%  softirqs.CPU64.NET_RX
>>>       22264           +42.4%      31709        softirqs.CPU64.SCHED
>>>       30930 ± 12%     +24.7%      38565 ±  3%  softirqs.CPU65.NET_RX
>>>       22563           +39.9%      31572        softirqs.CPU65.SCHED
>>>       27228 ±  3%     +40.3%      38207 ±  5%  softirqs.CPU66.NET_RX
>>>       33624 ±  2%     +10.9%      37300 ±  6%  softirqs.CPU66.RCU
>>>       22263           +42.0%      31615        softirqs.CPU66.SCHED
>>>       26629 ±  7%     +45.8%      38826 ±  3%  softirqs.CPU67.NET_RX
>>>       22267           +42.1%      31644        softirqs.CPU67.SCHED
>>>       27039 ±  7%     +46.4%      39598 ±  3%  softirqs.CPU68.NET_RX
>>>       22376           +41.0%      31552        softirqs.CPU68.SCHED
>>>       26835 ±  4%     +41.3%      37921 ±  4%  softirqs.CPU69.NET_RX
>>>       22292           +42.2%      31695        softirqs.CPU69.SCHED
>>>       28526 ±  9%     +36.6%      38980 ±  9%  softirqs.CPU7.NET_RX
>>>       22556 ±  3%     +42.0%      32026 ±  3%  softirqs.CPU7.SCHED
>>>       27716 ±  6%     +44.2%      39956 ±  5%  softirqs.CPU70.NET_RX
>>>       22376           +40.4%      31421        softirqs.CPU70.SCHED
>>>       25987 ±  2%     +54.0%      40017 ±  5%  softirqs.CPU71.NET_RX
>>>       22256           +42.9%      31803        softirqs.CPU71.SCHED
>>>       30271 ±  2%     +31.6%      39836 ±  8%  softirqs.CPU72.NET_RX
>>>       22864           +39.9%      31988        softirqs.CPU72.SCHED
>>>       28646 ±  7%     +40.8%      40326 ±  6%  softirqs.CPU73.NET_RX
>>>       22434           +41.4%      31715        softirqs.CPU73.SCHED
>>>       27673 ±  4%     +44.6%      40027 ±  5%  softirqs.CPU74.NET_RX
>>>       22454           +41.1%      31688        softirqs.CPU74.SCHED
>>>       26780 ±  5%     +50.1%      40207 ±  5%  softirqs.CPU75.NET_RX
>>>       22613           +40.8%      31840        softirqs.CPU75.SCHED
>>>       28051           +45.5%      40826 ±  5%  softirqs.CPU76.NET_RX
>>>       22608           +41.0%      31879        softirqs.CPU76.SCHED
>>>       26339 ±  5%     +49.7%      39425 ± 10%  softirqs.CPU77.NET_RX
>>>       22329           +43.4%      32019        softirqs.CPU77.SCHED
>>>       28822           +38.8%      40011 ±  6%  softirqs.CPU78.NET_RX
>>>       22588           +41.2%      31903        softirqs.CPU78.SCHED
>>>       26965 ±  4%     +50.8%      40656 ±  8%  softirqs.CPU79.NET_RX
>>>       22337           +41.9%      31701        softirqs.CPU79.SCHED
>>>       28854 ±  2%     +36.0%      39243 ±  7%  softirqs.CPU8.NET_RX
>>>       33868 ±  2%      +8.2%      36644 ±  4%  softirqs.CPU8.RCU
>>>       22572 ±  2%     +41.6%      31966 ±  3%  softirqs.CPU8.SCHED
>>>       25294 ±  3%     +62.3%      41046 ±  9%  softirqs.CPU80.NET_RX
>>>       22560           +41.6%      31949        softirqs.CPU80.SCHED
>>>       27226 ±  6%     +44.2%      39272 ±  5%  softirqs.CPU81.NET_RX
>>>       22433           +41.8%      31802        softirqs.CPU81.SCHED
>>>       28179 ±  2%     +45.8%      41089 ±  6%  softirqs.CPU82.NET_RX
>>>       22503           +41.1%      31742        softirqs.CPU82.SCHED
>>>       25370 ±  6%     +65.5%      41993 ±  3%  softirqs.CPU83.NET_RX
>>>       33224 ±  3%      +9.5%      36387 ±  5%  softirqs.CPU83.RCU
>>>       22482           +43.4%      32239        softirqs.CPU83.SCHED
>>>       29150 ±  4%     +41.9%      41357 ±  6%  softirqs.CPU84.NET_RX
>>>       22540           +40.3%      31629        softirqs.CPU84.SCHED
>>>       27566 ±  7%     +45.2%      40032 ±  6%  softirqs.CPU85.NET_RX
>>>       22614           +40.5%      31779        softirqs.CPU85.SCHED
>>>       26616 ±  3%     +48.0%      39392 ±  8%  softirqs.CPU86.NET_RX
>>>       22503           +41.8%      31899        softirqs.CPU86.SCHED
>>>       27217           +54.7%      42097 ±  4%  softirqs.CPU87.NET_RX
>>>       32656           +11.1%      36286 ±  8%  softirqs.CPU87.RCU
>>>       22383           +41.8%      31744        softirqs.CPU87.SCHED
>>>       29444 ±  5%     +35.2%      39821 ±  6%  softirqs.CPU88.NET_RX
>>>       22560 ±  2%     +41.3%      31887        softirqs.CPU88.SCHED
>>>       26383 ±  3%     +49.4%      39403 ±  5%  softirqs.CPU89.NET_RX
>>>       22647           +41.2%      31984        softirqs.CPU89.SCHED
>>>       27413           +45.5%      39873 ±  3%  softirqs.CPU9.NET_RX
>>>       22436 ±  3%     +43.3%      32154 ±  2%  softirqs.CPU9.SCHED
>>>       28117 ±  4%     +39.1%      39110 ±  5%  softirqs.CPU90.NET_RX
>>>       22524           +41.6%      31903        softirqs.CPU90.SCHED
>>>       28805 ±  3%     +43.1%      41207 ±  2%  softirqs.CPU91.NET_RX
>>>       22981           +38.0%      31714        softirqs.CPU91.SCHED
>>>       27920 ±  9%     +46.1%      40778 ±  7%  softirqs.CPU92.NET_RX
>>>       33255 ±  4%      +8.9%      36199 ±  7%  softirqs.CPU92.RCU
>>>       22467           +42.2%      31955        softirqs.CPU92.SCHED
>>>       26935 ±  5%     +49.3%      40211 ±  6%  softirqs.CPU93.NET_RX
>>>       22407           +42.4%      31898        softirqs.CPU93.SCHED
>>>       25585 ±  6%     +63.7%      41880 ±  5%  softirqs.CPU94.NET_RX
>>>       33095 ±  4%     +10.4%      36528 ±  9%  softirqs.CPU94.RCU
>>>       22417           +41.7%      31761        softirqs.CPU94.SCHED
>>>       27659 ±  6%     +43.3%      39639 ±  6%  softirqs.CPU95.NET_RX
>>>       21998           +42.6%      31361        softirqs.CPU95.SCHED
>>>       26582 ±  2%     +43.4%      38128 ±  8%  softirqs.CPU96.NET_RX
>>>       22170 ±  3%     +45.7%      32297        softirqs.CPU96.SCHED
>>>       27522 ±  5%     +39.4%      38353 ±  9%  softirqs.CPU97.NET_RX
>>>       22468 ±  3%     +43.4%      32209 ±  2%  softirqs.CPU97.SCHED
>>>       27117 ± 12%     +39.8%      37902 ± 10%  softirqs.CPU98.NET_RX
>>>       22600           +43.2%      32361 ±  2%  softirqs.CPU98.SCHED
>>>       26557 ±  2%     +43.7%      38161 ±  7%  softirqs.CPU99.NET_RX
>>>       23003           +41.6%      32565        softirqs.CPU99.SCHED
>>>     5270210           +44.9%    7635912        softirqs.NET_RX
>>>     4357512           +40.8%    6136169        softirqs.SCHED
>>>       80.01           -80.0        0.00        perf-profile.calltrace.cycles-pp.native_queued_spin_lock_slowpath._raw_spin_lock.scan_positives.dcache_readdir.iterate_dir
>>>       81.90           -68.9       13.02 ±  5%  perf-profile.calltrace.cycles-pp.dcache_readdir.iterate_dir.__x64_sys_getdents64.do_syscall_64.entry_SYSCALL_64_after_hwframe
>>>       81.91           -68.9       13.03 ±  5%  perf-profile.calltrace.cycles-pp.__x64_sys_getdents64.do_syscall_64.entry_SYSCALL_64_after_hwframe
>>>       81.90           -68.9       13.03 ±  5%  perf-profile.calltrace.cycles-pp.iterate_dir.__x64_sys_getdents64.do_syscall_64.entry_SYSCALL_64_after_hwframe
>>>       80.60           -68.8       11.80 ±  5%  perf-profile.calltrace.cycles-pp.scan_positives.dcache_readdir.iterate_dir.__x64_sys_getdents64.do_syscall_64
>>>       80.30           -68.7       11.63 ±  5%  perf-profile.calltrace.cycles-pp._raw_spin_lock.scan_positives.dcache_readdir.iterate_dir.__x64_sys_getdents64
>>>       85.88           -51.1       34.79 ± 13%  perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe
>>>       85.98           -50.8       35.15 ± 13%  perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe
>>>        0.37 ± 70%      +0.4        0.80 ±  4%  perf-profile.calltrace.cycles-pp.do_sys_openat2.do_sys_open.do_syscall_64.entry_SYSCALL_64_after_hwframe.creat64
>>>        0.37 ± 70%      +0.4        0.81 ±  4%  perf-profile.calltrace.cycles-pp.do_sys_open.do_syscall_64.entry_SYSCALL_64_after_hwframe.creat64
>>>        0.37 ± 70%      +0.4        0.81 ±  4%  perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.creat64
>>>        0.39 ± 70%      +0.4        0.82 ±  3%  perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.creat64
>>>        0.39 ± 70%      +0.4        0.84 ±  3%  perf-profile.calltrace.cycles-pp.creat64
>>>        0.37 ± 70%      +0.6        0.93 ± 10%  perf-profile.calltrace.cycles-pp.div_long
>>>        0.00            +0.6        0.62 ± 10%  perf-profile.calltrace.cycles-pp.kmem_cache_free.rcu_do_batch.rcu_core.__softirqentry_text_start.asm_call_sysvec_on_stack
>>>        0.00            +0.6        0.62 ± 10%  perf-profile.calltrace.cycles-pp.__strcat_chk
>>>        0.39 ± 70%      +0.7        1.04 ±  9%  perf-profile.calltrace.cycles-pp.add_int.add_int
>>>        0.38 ± 70%      +0.7        1.04 ±  8%  perf-profile.calltrace.cycles-pp.add_long.add_long
>>>        0.00            +0.7        0.68 ±  9%  perf-profile.calltrace.cycles-pp.rcu_do_batch.rcu_core.__softirqentry_text_start.asm_call_sysvec_on_stack.do_softirq_own_stack
>>>        0.00            +0.7        0.69 ±  9%  perf-profile.calltrace.cycles-pp.rcu_core.__softirqentry_text_start.asm_call_sysvec_on_stack.do_softirq_own_stack.irq_exit_rcu
>>>        0.00            +0.7        0.73 ±  3%  perf-profile.calltrace.cycles-pp.__handle_mm_fault.handle_mm_fault.do_user_addr_fault.exc_page_fault.asm_exc_page_fault
>>>        0.00            +0.8        0.76 ±  9%  perf-profile.calltrace.cycles-pp.__softirqentry_text_start.asm_call_sysvec_on_stack.do_softirq_own_stack.irq_exit_rcu.sysvec_apic_timer_interrupt
>>>        0.00            +0.8        0.76 ±  9%  perf-profile.calltrace.cycles-pp.asm_call_sysvec_on_stack.do_softirq_own_stack.irq_exit_rcu.sysvec_apic_timer_interrupt.asm_sysvec_apic_timer_interrupt
>>>        0.00            +0.8        0.76 ±  9%  perf-profile.calltrace.cycles-pp.do_softirq_own_stack.irq_exit_rcu.sysvec_apic_timer_interrupt.asm_sysvec_apic_timer_interrupt.cpuidle_enter_state
>>>        0.36 ± 70%      +0.8        1.12 ± 11%  perf-profile.calltrace.cycles-pp.__strncat_chk
>>>        0.00            +0.8        0.77 ±  3%  perf-profile.calltrace.cycles-pp.handle_mm_fault.do_user_addr_fault.exc_page_fault.asm_exc_page_fault
>>>        0.00            +0.8        0.77 ±  9%  perf-profile.calltrace.cycles-pp.irq_exit_rcu.sysvec_apic_timer_interrupt.asm_sysvec_apic_timer_interrupt.cpuidle_enter_state.cpuidle_enter
>>>        0.00            +0.8        0.79 ± 23%  perf-profile.calltrace.cycles-pp.osq_lock.rwsem_down_write_slowpath.__vma_adjust.__split_vma.__do_munmap
>>>        0.00            +0.8        0.81 ± 23%  perf-profile.calltrace.cycles-pp.rwsem_down_write_slowpath.__vma_adjust.__split_vma.__do_munmap.mmap_region
>>>        0.00            +0.8        0.81 ± 22%  perf-profile.calltrace.cycles-pp.osq_lock.rwsem_down_write_slowpath.__vma_adjust.__split_vma.mprotect_fixup
>>>        0.00            +0.8        0.82 ±  3%  perf-profile.calltrace.cycles-pp.walk_component.link_path_walk.path_openat.do_filp_open.do_sys_openat2
>>>        0.00            +0.8        0.84 ± 21%  perf-profile.calltrace.cycles-pp.rwsem_down_write_slowpath.__vma_adjust.__split_vma.mprotect_fixup.do_mprotect_pkey
>>>        0.00            +0.8        0.85 ±  3%  perf-profile.calltrace.cycles-pp.do_user_addr_fault.exc_page_fault.asm_exc_page_fault
>>>        0.00            +0.9        0.85 ±  3%  perf-profile.calltrace.cycles-pp.link_path_walk.path_openat.do_filp_open.do_sys_openat2.do_sys_open
>>>        0.00            +0.9        0.86 ±  3%  perf-profile.calltrace.cycles-pp.exc_page_fault.asm_exc_page_fault
>>>        0.00            +0.9        0.86 ± 22%  perf-profile.calltrace.cycles-pp.__vma_adjust.__split_vma.__do_munmap.mmap_region.do_mmap
>>>        0.18 ±141%      +0.9        1.04 ± 24%  perf-profile.calltrace.cycles-pp.osq_lock.rwsem_down_write_slowpath.unlink_file_vma.free_pgtables.unmap_region
>>>        0.17 ±141%      +0.9        1.03 ± 22%  perf-profile.calltrace.cycles-pp.osq_lock.rwsem_down_write_slowpath.vma_link.mmap_region.do_mmap
>>>        0.00            +0.9        0.88 ± 24%  perf-profile.calltrace.cycles-pp.unlink_file_vma.free_pgtables.exit_mmap.mmput.begin_new_exec
>>>        0.00            +0.9        0.88 ± 21%  perf-profile.calltrace.cycles-pp.__split_vma.__do_munmap.mmap_region.do_mmap.vm_mmap_pgoff
>>>        0.18 ±141%      +0.9        1.07 ± 23%  perf-profile.calltrace.cycles-pp.rwsem_down_write_slowpath.unlink_file_vma.free_pgtables.unmap_region.__do_munmap
>>>        0.17 ±141%      +0.9        1.06 ± 21%  perf-profile.calltrace.cycles-pp.rwsem_down_write_slowpath.vma_link.mmap_region.do_mmap.vm_mmap_pgoff
>>>        0.18 ±141%      +0.9        1.08 ± 23%  perf-profile.calltrace.cycles-pp.unlink_file_vma.free_pgtables.unmap_region.__do_munmap.mmap_region
>>>        0.19 ±141%      +0.9        1.09 ± 23%  perf-profile.calltrace.cycles-pp.free_pgtables.unmap_region.__do_munmap.mmap_region.do_mmap
>>>        0.19 ±141%      +0.9        1.12 ± 23%  perf-profile.calltrace.cycles-pp.unmap_region.__do_munmap.mmap_region.do_mmap.vm_mmap_pgoff
>>>        0.00            +0.9        0.93 ±  4%  perf-profile.calltrace.cycles-pp.asm_exc_page_fault
>>>        0.17 ±141%      +0.9        1.10 ± 19%  perf-profile.calltrace.cycles-pp.free_pgtables.exit_mmap.mmput.begin_new_exec.load_elf_binary
>>>        0.18 ±141%      +0.9        1.12 ± 20%  perf-profile.calltrace.cycles-pp.vma_link.mmap_region.do_mmap.vm_mmap_pgoff.ksys_mmap_pgoff
>>>        0.00            +0.9        0.94 ± 19%  perf-profile.calltrace.cycles-pp.__vma_adjust.__split_vma.mprotect_fixup.do_mprotect_pkey.__x64_sys_mprotect
>>>        0.00            +1.0        0.98 ± 18%  perf-profile.calltrace.cycles-pp.__split_vma.mprotect_fixup.do_mprotect_pkey.__x64_sys_mprotect.do_syscall_64
>>>        0.00            +1.0        1.02 ± 18%  perf-profile.calltrace.cycles-pp.mprotect_fixup.do_mprotect_pkey.__x64_sys_mprotect.do_syscall_64.entry_SYSCALL_64_after_hwframe
>>>        0.20 ±141%      +1.0        1.22 ± 24%  perf-profile.calltrace.cycles-pp.unlink_file_vma.free_pgtables.exit_mmap.mmput.do_exit
>>>        0.00            +1.0        1.03 ± 18%  perf-profile.calltrace.cycles-pp.__x64_sys_mprotect.do_syscall_64.entry_SYSCALL_64_after_hwframe
>>>        0.00            +1.0        1.03 ± 18%  perf-profile.calltrace.cycles-pp.do_mprotect_pkey.__x64_sys_mprotect.do_syscall_64.entry_SYSCALL_64_after_hwframe
>>>        0.39 ± 72%      +1.1        1.45 ± 11%  perf-profile.calltrace.cycles-pp.__do_sys_clone.do_syscall_64.entry_SYSCALL_64_after_hwframe
>>>        0.39 ± 72%      +1.1        1.45 ± 11%  perf-profile.calltrace.cycles-pp.kernel_clone.__do_sys_clone.do_syscall_64.entry_SYSCALL_64_after_hwframe
>>>        0.00            +1.1        1.06 ± 26%  perf-profile.calltrace.cycles-pp._raw_spin_lock.lockref_get.d_alloc_cursor.dcache_dir_open.do_dentry_open
>>>        0.00            +1.1        1.06 ± 26%  perf-profile.calltrace.cycles-pp.__cna_queued_spin_lock_slowpath._raw_spin_lock.lockref_get.d_alloc_cursor.dcache_dir_open
>>>        0.00            +1.1        1.06 ± 26%  perf-profile.calltrace.cycles-pp.lockref_get.d_alloc_cursor.dcache_dir_open.do_dentry_open.path_openat
>>>        0.00            +1.1        1.06 ± 26%  perf-profile.calltrace.cycles-pp.dcache_dir_open.do_dentry_open.path_openat.do_filp_open.do_sys_openat2
>>>        0.00            +1.1        1.06 ± 26%  perf-profile.calltrace.cycles-pp.d_alloc_cursor.dcache_dir_open.do_dentry_open.path_openat.do_filp_open
>>>        0.20 ±141%      +1.1        1.27 ± 14%  perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.__libc_fork
>>>        0.20 ±141%      +1.1        1.27 ± 14%  perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.__libc_fork
>>>        0.20 ±141%      +1.1        1.27 ± 14%  perf-profile.calltrace.cycles-pp.__do_sys_clone.do_syscall_64.entry_SYSCALL_64_after_hwframe.__libc_fork
>>>        0.19 ±141%      +1.1        1.27 ± 14%  perf-profile.calltrace.cycles-pp.kernel_clone.__do_sys_clone.do_syscall_64.entry_SYSCALL_64_after_hwframe.__libc_fork
>>>        0.21 ±141%      +1.1        1.33 ± 22%  perf-profile.calltrace.cycles-pp.free_pgtables.exit_mmap.mmput.do_exit.do_group_exit
>>>        0.00            +1.1        1.12 ± 11%  perf-profile.calltrace.cycles-pp.__cna_queued_spin_lock_slowpath._raw_spin_lock.dcache_readdir.iterate_dir.__x64_sys_getdents64
>>>        0.60 ±  6%      +1.1        1.73 ± 24%  perf-profile.calltrace.cycles-pp.exit_mmap.mmput.begin_new_exec.load_elf_binary.exec_binprm
>>>        0.20 ±141%      +1.1        1.33 ± 13%  perf-profile.calltrace.cycles-pp.__libc_fork
>>>        0.60 ±  6%      +1.1        1.73 ± 24%  perf-profile.calltrace.cycles-pp.mmput.begin_new_exec.load_elf_binary.exec_binprm.bprm_execve
>>>        0.17 ±141%      +1.1        1.32 ± 16%  perf-profile.calltrace.cycles-pp.asm_sysvec_apic_timer_interrupt.cpuidle_enter_state.cpuidle_enter.do_idle.cpu_startup_entry
>>>        0.00            +1.1        1.15 ± 18%  perf-profile.calltrace.cycles-pp.drain_obj_stock.refill_obj_stock.kmem_cache_free.rcu_do_batch.rcu_core
>>>        0.62 ±  6%      +1.2        1.78 ± 23%  perf-profile.calltrace.cycles-pp.begin_new_exec.load_elf_binary.exec_binprm.bprm_execve.do_execveat_common
>>>        0.00            +1.2        1.18 ± 21%  perf-profile.calltrace.cycles-pp.sysvec_apic_timer_interrupt.asm_sysvec_apic_timer_interrupt.cpuidle_enter_state.cpuidle_enter.do_idle
>>>        0.80 ± 20%      +1.2        2.03 ± 22%  perf-profile.calltrace.cycles-pp.__do_munmap.mmap_region.do_mmap.vm_mmap_pgoff.ksys_mmap_pgoff
>>>        0.73 ± 17%      +1.2        1.98 ± 14%  perf-profile.calltrace.cycles-pp.exit_mmap.mmput.do_exit.do_group_exit.__x64_sys_exit_group
>>>        0.73 ± 17%      +1.2        1.98 ± 14%  perf-profile.calltrace.cycles-pp.mmput.do_exit.do_group_exit.__x64_sys_exit_group.do_syscall_64
>>>        0.00            +1.3        1.28 ± 18%  perf-profile.calltrace.cycles-pp._raw_spin_lock.lockref_get.do_dentry_open.path_openat.do_filp_open
>>>        0.00            +1.3        1.28 ± 18%  perf-profile.calltrace.cycles-pp.__cna_queued_spin_lock_slowpath._raw_spin_lock.lockref_get.do_dentry_open.path_openat
>>>        0.00            +1.3        1.29 ± 18%  perf-profile.calltrace.cycles-pp.lockref_get.do_dentry_open.path_openat.do_filp_open.do_sys_openat2
>>>        0.00            +1.4        1.36 ± 39%  perf-profile.calltrace.cycles-pp.osq_lock.rwsem_down_write_slowpath.dup_mmap.dup_mm.copy_process
>>>        0.81 ± 16%      +1.4        2.20 ± 13%  perf-profile.calltrace.cycles-pp.do_exit.do_group_exit.__x64_sys_exit_group.do_syscall_64.entry_SYSCALL_64_after_hwframe
>>>        0.81 ± 16%      +1.4        2.21 ± 13%  perf-profile.calltrace.cycles-pp.__x64_sys_exit_group.do_syscall_64.entry_SYSCALL_64_after_hwframe
>>>        0.81 ± 16%      +1.4        2.21 ± 13%  perf-profile.calltrace.cycles-pp.do_group_exit.__x64_sys_exit_group.do_syscall_64.entry_SYSCALL_64_after_hwframe
>>>        0.00            +1.4        1.41 ± 38%  perf-profile.calltrace.cycles-pp.rwsem_down_write_slowpath.dup_mmap.dup_mm.copy_process.kernel_clone
>>>        0.00            +1.4        1.41 ± 17%  perf-profile.calltrace.cycles-pp.refill_obj_stock.kmem_cache_free.rcu_do_batch.rcu_core.__softirqentry_text_start
>>>        0.18 ±141%      +1.7        1.92 ± 26%  perf-profile.calltrace.cycles-pp.osq_lock.rwsem_down_write_slowpath.unlink_file_vma.free_pgtables.exit_mmap
>>>        0.19 ±141%      +1.8        2.01 ± 25%  perf-profile.calltrace.cycles-pp.rwsem_down_write_slowpath.unlink_file_vma.free_pgtables.exit_mmap.mmput
>>>        0.83 ±  5%      +1.8        2.65 ± 14%  perf-profile.calltrace.cycles-pp.load_elf_binary.exec_binprm.bprm_execve.do_execveat_common.__x64_sys_execve
>>>        0.83 ±  4%      +1.8        2.67 ± 13%  perf-profile.calltrace.cycles-pp.exec_binprm.bprm_execve.do_execveat_common.__x64_sys_execve.do_syscall_64
>>>        0.36 ±141%      +2.0        2.32 ± 15%  perf-profile.calltrace.cycles-pp.dup_mm.copy_process.kernel_clone.__do_sys_clone.do_syscall_64
>>>        1.24 ± 20%      +2.0        3.23 ± 21%  perf-profile.calltrace.cycles-pp.mmap_region.do_mmap.vm_mmap_pgoff.ksys_mmap_pgoff.do_syscall_64
>>>        1.25 ± 20%      +2.0        3.23 ± 21%  perf-profile.calltrace.cycles-pp.do_mmap.vm_mmap_pgoff.ksys_mmap_pgoff.do_syscall_64.entry_SYSCALL_64_after_hwframe
>>>        1.26 ± 20%      +2.0        3.25 ± 21%  perf-profile.calltrace.cycles-pp.ksys_mmap_pgoff.do_syscall_64.entry_SYSCALL_64_after_hwframe
>>>        1.26 ± 20%      +2.0        3.25 ± 21%  perf-profile.calltrace.cycles-pp.vm_mmap_pgoff.ksys_mmap_pgoff.do_syscall_64.entry_SYSCALL_64_after_hwframe
>>>        1.68 ± 14%      +2.0        3.72 ± 10%  perf-profile.calltrace.cycles-pp.string_rtns_1
>>>        0.19 ±141%      +2.1        2.29 ± 15%  perf-profile.calltrace.cycles-pp.dup_mmap.dup_mm.copy_process.kernel_clone.__do_sys_clone
>>>        0.39 ±141%      +2.2        2.58 ± 13%  perf-profile.calltrace.cycles-pp.copy_process.kernel_clone.__do_sys_clone.do_syscall_64.entry_SYSCALL_64_after_hwframe
>>>        0.00            +2.6        2.62 ±  6%  perf-profile.calltrace.cycles-pp.do_dentry_open.path_openat.do_filp_open.do_sys_openat2.do_sys_open
>>>        0.40 ± 70%      +3.1        3.52 ± 15%  perf-profile.calltrace.cycles-pp.kmem_cache_free.rcu_do_batch.rcu_core.__softirqentry_text_start.run_ksoftirqd
>>>        0.61 ± 11%      +3.2        3.79 ± 15%  perf-profile.calltrace.cycles-pp.rcu_do_batch.rcu_core.__softirqentry_text_start.run_ksoftirqd.smpboot_thread_fn
>>>        0.61 ± 11%      +3.2        3.79 ± 15%  perf-profile.calltrace.cycles-pp.rcu_core.__softirqentry_text_start.run_ksoftirqd.smpboot_thread_fn.kthread
>>>        0.61 ± 11%      +3.2        3.79 ± 15%  perf-profile.calltrace.cycles-pp.__softirqentry_text_start.run_ksoftirqd.smpboot_thread_fn.kthread.ret_from_fork
>>>        0.61 ± 11%      +3.2        3.79 ± 15%  perf-profile.calltrace.cycles-pp.run_ksoftirqd.smpboot_thread_fn.kthread.ret_from_fork
>>>        0.61 ± 12%      +3.2        3.82 ± 15%  perf-profile.calltrace.cycles-pp.smpboot_thread_fn.kthread.ret_from_fork
>>>        0.65 ± 11%      +3.2        3.88 ± 15%  perf-profile.calltrace.cycles-pp.kthread.ret_from_fork
>>>        0.65 ± 11%      +3.2        3.89 ± 15%  perf-profile.calltrace.cycles-pp.ret_from_fork
>>>        0.22 ±141%      +4.1        4.27 ± 28%  perf-profile.calltrace.cycles-pp.do_sys_openat2.do_sys_open.do_syscall_64.entry_SYSCALL_64_after_hwframe
>>>        0.22 ±141%      +4.1        4.28 ± 28%  perf-profile.calltrace.cycles-pp.do_sys_open.do_syscall_64.entry_SYSCALL_64_after_hwframe
>>>        3.30 ± 10%      +6.1        9.36 ± 10%  perf-profile.calltrace.cycles-pp.intel_idle.cpuidle_enter_state.cpuidle_enter.do_idle.cpu_startup_entry
>>>        0.00            +6.2        6.21 ±  3%  perf-profile.calltrace.cycles-pp._raw_spin_lock.lockref_get_not_dead.__legitimize_path.unlazy_walk.complete_walk
>>>        0.00            +6.2        6.22 ±  3%  perf-profile.calltrace.cycles-pp.lockref_get_not_dead.__legitimize_path.unlazy_walk.complete_walk.path_openat
>>>        0.00            +6.2        6.22 ±  3%  perf-profile.calltrace.cycles-pp.unlazy_walk.complete_walk.path_openat.do_filp_open.do_sys_openat2
>>>        0.00            +6.2        6.22 ±  3%  perf-profile.calltrace.cycles-pp.__legitimize_path.unlazy_walk.complete_walk.path_openat.do_filp_open
>>>        0.00            +6.2        6.22 ±  3%  perf-profile.calltrace.cycles-pp.complete_walk.path_openat.do_filp_open.do_sys_openat2.do_sys_open
>>>        0.00            +6.5        6.47 ±  7%  perf-profile.calltrace.cycles-pp._raw_spin_lock.lockref_put_or_lock.dput.step_into.path_openat
>>>        0.00            +6.5        6.47 ±  7%  perf-profile.calltrace.cycles-pp.__cna_queued_spin_lock_slowpath._raw_spin_lock.lockref_put_or_lock.dput.step_into
>>>        0.00            +6.5        6.47 ±  7%  perf-profile.calltrace.cycles-pp.dput.step_into.path_openat.do_filp_open.do_open_execat
>>>        0.00            +6.5        6.47 ±  7%  perf-profile.calltrace.cycles-pp.lockref_put_or_lock.dput.step_into.path_openat.do_filp_open
>>>        0.00            +7.0        6.99 ± 19%  perf-profile.calltrace.cycles-pp.do_sys_open.do_syscall_64.entry_SYSCALL_64_after_hwframe.__open64_nocancel
>>>        0.00            +7.0        6.99 ± 19%  perf-profile.calltrace.cycles-pp.do_sys_openat2.do_sys_open.do_syscall_64.entry_SYSCALL_64_after_hwframe.__open64_nocancel
>>>        0.00            +7.0        6.99 ± 18%  perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.__open64_nocancel
>>>        0.00            +7.0        7.00 ± 18%  perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.__open64_nocancel
>>>        0.00            +7.0        7.00 ± 18%  perf-profile.calltrace.cycles-pp.__open64_nocancel
>>>        3.79 ± 10%      +7.0       10.82 ±  8%  perf-profile.calltrace.cycles-pp.cpuidle_enter_state.cpuidle_enter.do_idle.cpu_startup_entry.start_secondary
>>>        3.86 ±  9%      +7.1       10.96 ±  8%  perf-profile.calltrace.cycles-pp.cpuidle_enter.do_idle.cpu_startup_entry.start_secondary.secondary_startup_64_no_verify
>>>        4.17 ±  6%      +7.4       11.59 ±  8%  perf-profile.calltrace.cycles-pp.do_idle.cpu_startup_entry.start_secondary.secondary_startup_64_no_verify
>>>        4.17 ±  6%      +7.4       11.59 ±  8%  perf-profile.calltrace.cycles-pp.cpu_startup_entry.start_secondary.secondary_startup_64_no_verify
>>>        4.17 ±  6%      +7.4       11.59 ±  8%  perf-profile.calltrace.cycles-pp.start_secondary.secondary_startup_64_no_verify
>>>        4.20 ±  6%      +7.4       11.63 ±  8%  perf-profile.calltrace.cycles-pp.secondary_startup_64_no_verify
>>>        0.00            +8.6        8.58 ± 38%  perf-profile.calltrace.cycles-pp.do_execveat_common.__x64_sys_execve.do_syscall_64.entry_SYSCALL_64_after_hwframe
>>>        0.00            +8.6        8.59 ± 38%  perf-profile.calltrace.cycles-pp.__x64_sys_execve.do_syscall_64.entry_SYSCALL_64_after_hwframe
>>>        0.00           +11.2       11.23 ±  5%  perf-profile.calltrace.cycles-pp.__cna_queued_spin_lock_slowpath._raw_spin_lock.scan_positives.dcache_readdir.iterate_dir
>>>        0.56 ±  7%     +11.4       11.95 ±  2%  perf-profile.calltrace.cycles-pp.path_openat.do_filp_open.do_sys_openat2.do_sys_open.do_syscall_64
>>>        0.56 ±  7%     +11.4       11.96 ±  2%  perf-profile.calltrace.cycles-pp.do_filp_open.do_sys_openat2.do_sys_open.do_syscall_64.entry_SYSCALL_64_after_hwframe
>>>        0.00           +24.2       24.23 ±  4%  perf-profile.calltrace.cycles-pp._raw_spin_lock.lockref_get_not_dead.__legitimize_path.unlazy_walk.pick_link
>>>        0.00           +24.2       24.25 ±  4%  perf-profile.calltrace.cycles-pp.unlazy_walk.pick_link.step_into.path_openat.do_filp_open
>>>        0.00           +24.2       24.25 ±  4%  perf-profile.calltrace.cycles-pp.__legitimize_path.unlazy_walk.pick_link.step_into.path_openat
>>>        0.00           +24.2       24.25 ±  4%  perf-profile.calltrace.cycles-pp.lockref_get_not_dead.__legitimize_path.unlazy_walk.pick_link.step_into
>>>        0.00           +24.2       24.25 ±  4%  perf-profile.calltrace.cycles-pp.pick_link.step_into.path_openat.do_filp_open.do_open_execat
>>>        1.02 ±  4%     +24.4       25.41 ±  8%  perf-profile.calltrace.cycles-pp.do_execveat_common.__x64_sys_execve.do_syscall_64.entry_SYSCALL_64_after_hwframe.execve
>>>        1.02 ±  4%     +24.4       25.42 ±  8%  perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.execve
>>>        1.02 ±  4%     +24.4       25.42 ±  8%  perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.execve
>>>        1.02 ±  4%     +24.4       25.42 ±  8%  perf-profile.calltrace.cycles-pp.__x64_sys_execve.do_syscall_64.entry_SYSCALL_64_after_hwframe.execve
>>>        1.03 ±  4%     +24.4       25.43 ±  8%  perf-profile.calltrace.cycles-pp.execve
>>>        0.00           +30.4       30.45 ±  3%  perf-profile.calltrace.cycles-pp.__cna_queued_spin_lock_slowpath._raw_spin_lock.lockref_get_not_dead.__legitimize_path.unlazy_walk
>>>        0.00           +30.8       30.82 ±  3%  perf-profile.calltrace.cycles-pp.step_into.path_openat.do_filp_open.do_open_execat.bprm_execve
>>>        0.00           +30.8       30.84 ±  3%  perf-profile.calltrace.cycles-pp.do_filp_open.do_open_execat.bprm_execve.do_execveat_common.__x64_sys_execve
>>>        0.00           +30.8       30.84 ±  3%  perf-profile.calltrace.cycles-pp.path_openat.do_filp_open.do_open_execat.bprm_execve.do_execveat_common
>>>        0.00           +30.8       30.84 ±  3%  perf-profile.calltrace.cycles-pp.do_open_execat.bprm_execve.do_execveat_common.__x64_sys_execve.do_syscall_64
>>>        0.95 ±  4%     +32.8       33.76 ±  3%  perf-profile.calltrace.cycles-pp.bprm_execve.do_execveat_common.__x64_sys_execve.do_syscall_64.entry_SYSCALL_64_after_hwframe
>>>       82.33           -82.3        0.00        perf-profile.children.cycles-pp.native_queued_spin_lock_slowpath
>>>       81.90           -68.9       13.02 ±  5%  perf-profile.children.cycles-pp.dcache_readdir
>>>       81.91           -68.9       13.03 ±  5%  perf-profile.children.cycles-pp.__x64_sys_getdents64
>>>       81.90           -68.9       13.03 ±  5%  perf-profile.children.cycles-pp.iterate_dir
>>>       80.60           -68.8       11.80 ±  5%  perf-profile.children.cycles-pp.scan_positives
>>>       82.71           -27.9       54.83 ±  2%  perf-profile.children.cycles-pp._raw_spin_lock
>>>       89.05           -18.0       71.08        perf-profile.children.cycles-pp.do_syscall_64
>>>       89.35           -17.5       71.82        perf-profile.children.cycles-pp.entry_SYSCALL_64_after_hwframe
>>>        0.54 ± 22%      -0.2        0.32 ± 20%  perf-profile.children.cycles-pp.llseek
>>>        0.62            -0.2        0.45 ±  9%  perf-profile.children.cycles-pp.ksys_lseek
>>>        0.62            -0.2        0.45 ±  9%  perf-profile.children.cycles-pp.dcache_dir_lseek
>>>        0.06 ±  8%      +0.0        0.09 ± 21%  perf-profile.children.cycles-pp.ktime_get_update_offsets_now
>>>        0.06 ± 13%      +0.0        0.10 ± 15%  perf-profile.children.cycles-pp.update_load_avg
>>>        0.16 ±  8%      +0.0        0.21 ± 17%  perf-profile.children.cycles-pp.ktime_get
>>>        0.00            +0.1        0.05        perf-profile.children.cycles-pp.copy_page
>>>        0.00            +0.1        0.05        perf-profile.children.cycles-pp.copy_string_kernel
>>>        0.00            +0.1        0.05        perf-profile.children.cycles-pp.time
>>>        0.00            +0.1        0.05        perf-profile.children.cycles-pp.__x64_sys_rt_sigreturn
>>>        0.00            +0.1        0.05        perf-profile.children.cycles-pp.perf_iterate_sb
>>>        0.00            +0.1        0.05        perf-profile.children.cycles-pp._raw_spin_trylock
>>>        0.00            +0.1        0.05        perf-profile.children.cycles-pp.irqtime_account_irq
>>>        0.00            +0.1        0.05        perf-profile.children.cycles-pp.unlock_page
>>>        0.00            +0.1        0.05 ±  8%  perf-profile.children.cycles-pp.common_file_perm
>>>        0.00            +0.1        0.05 ±  8%  perf-profile.children.cycles-pp.shift_arg_pages
>>>        0.00            +0.1        0.05 ±  8%  perf-profile.children.cycles-pp.__list_del_entry_valid
>>>        0.00            +0.1        0.05 ±  8%  perf-profile.children.cycles-pp.__check_object_size
>>>        0.00            +0.1        0.05 ±  8%  perf-profile.children.cycles-pp.compar1
>>>        0.00            +0.1        0.05 ±  8%  perf-profile.children.cycles-pp.lru_add_drain_cpu
>>>        0.00            +0.1        0.05 ±  8%  perf-profile.children.cycles-pp.do_faccessat
>>>        0.00            +0.1        0.05 ±  8%  perf-profile.children.cycles-pp.call_rcu
>>>        0.00            +0.1        0.05 ±  8%  perf-profile.children.cycles-pp.rebalance_domains
>>>        0.00            +0.1        0.05 ±  9%  perf-profile.children.cycles-pp.__pud_alloc
>>>        0.00            +0.1        0.05 ±  9%  perf-profile.children.cycles-pp.do_task_dead
>>>        0.00            +0.1        0.06 ±  9%  perf-profile.children.cycles-pp.new_sync_read
>>>        0.00            +0.1        0.06 ±  9%  perf-profile.children.cycles-pp.ip_finish_output2
>>>        0.00            +0.1        0.06 ±  9%  perf-profile.children.cycles-pp.sync_regs
>>>        0.00            +0.1        0.06 ±  9%  perf-profile.children.cycles-pp.d_add
>>>        0.00            +0.1        0.06 ±  9%  perf-profile.children.cycles-pp.lru_add_drain
>>>        0.00            +0.1        0.06 ±  7%  perf-profile.children.cycles-pp._cond_resched
>>>        0.00            +0.1        0.06 ±  7%  perf-profile.children.cycles-pp.cfree
>>>        0.00            +0.1        0.06 ±  7%  perf-profile.children.cycles-pp.update_rq_clock
>>>        0.00            +0.1        0.06 ±  7%  perf-profile.children.cycles-pp.pagevec_lru_move_fn
>>>        0.00            +0.1        0.06 ±  7%  perf-profile.children.cycles-pp.__rb_insert_augmented
>>>        0.00            +0.1        0.06 ±  7%  perf-profile.children.cycles-pp.kfree
>>>        0.00            +0.1        0.06 ±  7%  perf-profile.children.cycles-pp.security_cred_free
>>>        0.00            +0.1        0.06 ± 14%  perf-profile.children.cycles-pp.worker_thread
>>>        0.00            +0.1        0.06 ± 11%  perf-profile.children.cycles-pp.read
>>>        0.00            +0.1        0.06        perf-profile.children.cycles-pp.anon_vma_interval_tree_insert
>>>        0.00            +0.1        0.06        perf-profile.children.cycles-pp.__might_sleep
>>>        0.00            +0.1        0.06        perf-profile.children.cycles-pp.simple_lookup
>>>        0.00            +0.1        0.06        perf-profile.children.cycles-pp.create_elf_tables
>>>        0.00            +0.1        0.06        perf-profile.children.cycles-pp.setup_arg_pages
>>>        0.02 ±141%      +0.1        0.08 ± 54%  perf-profile.children.cycles-pp.irq_enter_rcu
>>>        0.00            +0.1        0.06 ±  6%  perf-profile.children.cycles-pp.alloc_pages_vma
>>>        0.00            +0.1        0.06 ±  6%  perf-profile.children.cycles-pp.page_add_file_rmap
>>>        0.00            +0.1        0.06 ± 17%  perf-profile.children.cycles-pp.memcpy_erms
>>>        0.00            +0.1        0.07 ± 17%  perf-profile.children.cycles-pp.do_linkat
>>>        0.00            +0.1        0.07 ± 17%  perf-profile.children.cycles-pp.__x64_sys_link
>>>        0.00            +0.1        0.07 ±  7%  perf-profile.children.cycles-pp.inode_permission
>>>        0.00            +0.1        0.07 ±  7%  perf-profile.children.cycles-pp.ip_output
>>>        0.00            +0.1        0.07 ± 23%  perf-profile.children.cycles-pp.tick_nohz_next_event
>>>        0.05 ± 70%      +0.1        0.12 ± 10%  perf-profile.children.cycles-pp.__close
>>>        0.00            +0.1        0.07 ± 12%  perf-profile.children.cycles-pp.try_charge
>>>        0.00            +0.1        0.07 ±  6%  perf-profile.children.cycles-pp.do_truncate
>>>        0.00            +0.1        0.07 ± 12%  perf-profile.children.cycles-pp.udp_sendmsg
>>>        0.00            +0.1        0.07 ± 12%  perf-profile.children.cycles-pp.put_task_stack
>>>        0.00            +0.1        0.07 ± 12%  perf-profile.children.cycles-pp.free_pages_and_swap_cache
>>>        0.02 ±141%      +0.1        0.08 ±  5%  perf-profile.children.cycles-pp.unlink
>>>        0.07 ± 12%      +0.1        0.14 ± 15%  perf-profile.children.cycles-pp.write
>>>        0.06            +0.1        0.13        perf-profile.children.cycles-pp.wake_up_new_task
>>>        0.06 ± 13%      +0.1        0.13 ±  5%  perf-profile.children.cycles-pp.__d_lookup_rcu
>>>        0.00            +0.1        0.07 ± 10%  perf-profile.children.cycles-pp.uncharge_batch
>>>        0.00            +0.1        0.07        perf-profile.children.cycles-pp.group_send_sig_info
>>>        0.00            +0.1        0.07        perf-profile.children.cycles-pp.__get_user_pages
>>>        0.00            +0.1        0.07        perf-profile.children.cycles-pp.strnlen_user
>>>        0.00            +0.1        0.07        perf-profile.children.cycles-pp.security_task_kill
>>>        0.00            +0.1        0.07        perf-profile.children.cycles-pp.apparmor_task_kill
>>>        0.00            +0.1        0.07 ± 10%  perf-profile.children.cycles-pp.__pmd_alloc
>>>        0.05 ±  8%      +0.1        0.12 ±  4%  perf-profile.children.cycles-pp.apparmor_file_open
>>>        0.09 ± 24%      +0.1        0.16 ±  6%  perf-profile.children.cycles-pp.kill
>>>        0.05            +0.1        0.12 ±  3%  perf-profile.children.cycles-pp.security_file_free
>>>        0.05            +0.1        0.12 ±  3%  perf-profile.children.cycles-pp.apparmor_file_free_security
>>>        0.00            +0.1        0.07 ±  5%  perf-profile.children.cycles-pp.__anon_vma_prepare
>>>        0.00            +0.1        0.07 ±  5%  perf-profile.children.cycles-pp.find_vma
>>>        0.00            +0.1        0.07 ±  5%  perf-profile.children.cycles-pp.get_arg_page
>>>        0.00            +0.1        0.07 ±  5%  perf-profile.children.cycles-pp.__get_user_pages_remote
>>>        0.00            +0.1        0.07 ±  5%  perf-profile.children.cycles-pp.mem_cgroup_charge
>>>        0.00            +0.1        0.07 ±  5%  perf-profile.children.cycles-pp.try_to_wake_up
>>>        0.00            +0.1        0.07 ±  5%  perf-profile.children.cycles-pp.rcu_sched_clock_irq
>>>        0.00            +0.1        0.07 ± 22%  perf-profile.children.cycles-pp.file_free_rcu
>>>        0.05 ±  8%      +0.1        0.13 ±  6%  perf-profile.children.cycles-pp.security_file_open
>>>        0.05 ±  8%      +0.1        0.13 ±  3%  perf-profile.children.cycles-pp.getname_flags
>>>        0.03 ± 70%      +0.1        0.11 ±  4%  perf-profile.children.cycles-pp.kill_pid_info
>>>        0.07            +0.1        0.15 ±  3%  perf-profile.children.cycles-pp.native_irq_return_iret
>>>        0.00            +0.1        0.08 ±  6%  perf-profile.children.cycles-pp.vfs_read
>>>        0.00            +0.1        0.08 ±  6%  perf-profile.children.cycles-pp.__mod_memcg_state
>>>        0.00            +0.1        0.08 ±  6%  perf-profile.children.cycles-pp.mem_cgroup_uncharge_list
>>>        0.03 ± 70%      +0.1        0.11        perf-profile.children.cycles-pp.kill_something_info
>>>        0.00            +0.1        0.08 ±  5%  perf-profile.children.cycles-pp.ksys_read
>>>        0.00            +0.1        0.08 ± 10%  perf-profile.children.cycles-pp.rcu_cblist_dequeue
>>>        0.00            +0.1        0.08 ±  8%  perf-profile.children.cycles-pp.vma_interval_tree_remove
>>>        0.04 ± 71%      +0.1        0.12 ±  3%  perf-profile.children.cycles-pp.__x64_sys_kill
>>>        0.00            +0.1        0.08 ± 10%  perf-profile.children.cycles-pp.filldir64
>>>        0.00            +0.1        0.08 ± 10%  perf-profile.children.cycles-pp.put_cred_rcu
>>>        0.00            +0.1        0.08 ± 26%  perf-profile.children.cycles-pp.tick_nohz_get_sleep_length
>>>        0.00            +0.1        0.08 ± 13%  perf-profile.children.cycles-pp.__slab_free
>>>        0.02 ±141%      +0.1        0.10        perf-profile.children.cycles-pp.up_write
>>>        0.05 ±  8%      +0.1        0.14 ±  3%  perf-profile.children.cycles-pp.copy_pte_range
>>>        0.00            +0.1        0.08 ±  5%  perf-profile.children.cycles-pp.clear_page_erms
>>>        0.06 ±  8%      +0.1        0.14 ±  5%  perf-profile.children.cycles-pp.down_write
>>>        0.02 ±141%      +0.1        0.10 ± 10%  perf-profile.children.cycles-pp.malloc
>>>        0.00            +0.1        0.09 ±  4%  perf-profile.children.cycles-pp.wait4
>>>        0.00            +0.1        0.09 ±  4%  perf-profile.children.cycles-pp.down_read
>>>        0.00            +0.1        0.09 ±  4%  perf-profile.children.cycles-pp.__get_free_pages
>>>        0.00            +0.1        0.09 ±  4%  perf-profile.children.cycles-pp.__put_anon_vma
>>>        0.00            +0.1        0.09 ±  4%  perf-profile.children.cycles-pp.arch_do_signal
>>>        0.00            +0.1        0.09 ±  4%  perf-profile.children.cycles-pp.__dentry_kill
>>>        0.02 ±141%      +0.1        0.11 ±  4%  perf-profile.children.cycles-pp._raw_spin_lock_irqsave
>>>        0.00            +0.1        0.09        perf-profile.children.cycles-pp.strncpy_from_user
>>>        0.00            +0.1        0.09 ±  7%  perf-profile.children.cycles-pp.aa_get_task_label
>>>        0.00            +0.1        0.09 ±  7%  perf-profile.children.cycles-pp.schedule_idle
>>>        0.00            +0.1        0.09 ± 17%  perf-profile.children.cycles-pp.rwsem_wake
>>>        0.02 ±141%      +0.1        0.11 ±  4%  perf-profile.children.cycles-pp.___might_sleep
>>>        0.03 ± 70%      +0.1        0.12 ±  4%  perf-profile.children.cycles-pp.copy_strings
>>>        0.09 ± 13%      +0.1        0.19 ± 14%  perf-profile.children.cycles-pp.update_cfs_group
>>>        0.08 ± 10%      +0.1        0.17 ± 11%  perf-profile.children.cycles-pp.new_sync_write
>>>        0.00            +0.1        0.09 ±  4%  perf-profile.children.cycles-pp.sched_exec
>>>        0.02 ±141%      +0.1        0.11        perf-profile.children.cycles-pp.alloc_set_pte
>>>        0.05 ±  8%      +0.1        0.15 ±  2%  perf-profile.children.cycles-pp.get_page_from_freelist
>>>        0.00            +0.1        0.10 ±  5%  perf-profile.children.cycles-pp.prep_new_page
>>>        0.00            +0.1        0.10 ±  9%  perf-profile.children.cycles-pp.__pte_alloc
>>>        0.04 ± 71%      +0.1        0.13 ±  6%  perf-profile.children.cycles-pp.__mmap
>>>        0.00            +0.1        0.10 ±  8%  perf-profile.children.cycles-pp.sock_sendmsg
>>>        0.00            +0.1        0.10 ±  4%  perf-profile.children.cycles-pp.perf_event_mmap
>>>        0.00            +0.1        0.10 ±  8%  perf-profile.children.cycles-pp.__memcg_kmem_uncharge_page
>>>        0.00            +0.1        0.10 ± 45%  perf-profile.children.cycles-pp.dcache_dir_close
>>>        0.00            +0.1        0.10 ±  7%  perf-profile.children.cycles-pp.sock_write_iter
>>>        0.00            +0.1        0.10 ±  7%  perf-profile.children.cycles-pp.lockref_put_return
>>>        0.09 ± 10%      +0.1        0.19 ±  9%  perf-profile.children.cycles-pp.ksys_write
>>>        0.02 ±141%      +0.1        0.12 ±  3%  perf-profile.children.cycles-pp.apparmor_task_getsecid
>>>        0.08 ±  5%      +0.1        0.19 ±  9%  perf-profile.children.cycles-pp.vfs_write
>>>        0.00            +0.1        0.11 ± 10%  perf-profile.children.cycles-pp.do_unlinkat
>>>        0.08 ± 10%      +0.1        0.19 ±  2%  perf-profile.children.cycles-pp.find_idlest_group
>>>        0.00            +0.1        0.11 ± 33%  perf-profile.children.cycles-pp.__close_nocancel
>>>        0.00            +0.1        0.11 ±  4%  perf-profile.children.cycles-pp.vm_area_alloc
>>>        0.06 ± 16%      +0.1        0.16 ±  5%  perf-profile.children.cycles-pp.do_anonymous_page
>>>        0.02 ±141%      +0.1        0.12 ±  4%  perf-profile.children.cycles-pp.security_task_getsecid
>>>        0.08 ± 14%      +0.1        0.19 ±  6%  perf-profile.children.cycles-pp.__xstat64
>>>        0.00            +0.1        0.11        perf-profile.children.cycles-pp.__mod_memcg_lruvec_state
>>>        0.06 ± 16%      +0.1        0.17 ±  4%  perf-profile.children.cycles-pp.obj_cgroup_charge
>>>        0.07            +0.1        0.18 ±  5%  perf-profile.children.cycles-pp.apparmor_file_alloc_security
>>>        0.00            +0.1        0.11 ±  3%  perf-profile.children.cycles-pp.__clear_user
>>>        0.20 ± 12%      +0.1        0.31 ± 15%  perf-profile.children.cycles-pp.task_tick_fair
>>>        0.02 ±141%      +0.1        0.13 ±  5%  perf-profile.children.cycles-pp.find_busiest_group
>>>        0.02 ±141%      +0.1        0.13 ±  5%  perf-profile.children.cycles-pp.update_sd_lb_stats
>>>        0.00            +0.1        0.11 ±  4%  perf-profile.children.cycles-pp.ima_file_check
>>>        0.08 ± 10%      +0.1        0.20 ±  4%  perf-profile.children.cycles-pp.vma_interval_tree_insert
>>>        0.07 ± 14%      +0.1        0.18 ±  4%  perf-profile.children.cycles-pp.page_remove_rmap
>>>        0.02 ±141%      +0.1        0.14 ±  3%  perf-profile.children.cycles-pp.wp_page_copy
>>>        0.02 ±141%      +0.1        0.14 ±  6%  perf-profile.children.cycles-pp.pte_alloc_one
>>>        0.03 ± 70%      +0.1        0.15 ±  2%  perf-profile.children.cycles-pp.schedule
>>>        0.07 ± 11%      +0.1        0.19        perf-profile.children.cycles-pp._dl_addr
>>>        0.00            +0.1        0.12 ±  8%  perf-profile.children.cycles-pp.finish_task_switch
>>>        0.00            +0.1        0.12 ± 13%  perf-profile.children.cycles-pp.__put_task_struct
>>>        0.08            +0.1        0.20 ±  5%  perf-profile.children.cycles-pp.security_file_alloc
>>>        0.09 ±  9%      +0.1        0.21 ±  2%  perf-profile.children.cycles-pp.select_task_rq_fair
>>>        0.04 ± 71%      +0.1        0.16 ±  5%  perf-profile.children.cycles-pp.load_balance
>>>        0.05 ±  8%      +0.1        0.18 ±  4%  perf-profile.children.cycles-pp.vm_area_dup
>>>        0.04 ± 71%      +0.1        0.17 ±  2%  perf-profile.children.cycles-pp.__d_alloc
>>>        0.02 ±141%      +0.1        0.15 ±  7%  perf-profile.children.cycles-pp.do_wait
>>>        0.02 ±141%      +0.1        0.15 ±  5%  perf-profile.children.cycles-pp.pick_next_task_fair
>>>        0.06 ± 16%      +0.1        0.19 ±  3%  perf-profile.children.cycles-pp.anon_vma_clone
>>>        0.02 ±141%      +0.1        0.15 ±  8%  perf-profile.children.cycles-pp.__do_sys_wait4
>>>        0.02 ±141%      +0.1        0.15 ±  8%  perf-profile.children.cycles-pp.kernel_wait4
>>>        0.05 ±  8%      +0.1        0.19 ±  2%  perf-profile.children.cycles-pp.__vm_munmap
>>>        0.28 ± 13%      +0.1        0.42 ± 12%  perf-profile.children.cycles-pp.scheduler_tick
>>>        0.00            +0.1        0.14 ±  3%  perf-profile.children.cycles-pp.cna_order_queue
>>>        0.08 ± 12%      +0.1        0.22 ±  5%  perf-profile.children.cycles-pp.copy_page_range
>>>        0.06 ± 19%      +0.1        0.21 ±  7%  perf-profile.children.cycles-pp.__memcg_kmem_charge_page
>>>        0.00            +0.1        0.15 ±  5%  perf-profile.children.cycles-pp.newidle_balance
>>>        0.05 ±  8%      +0.2        0.21 ± 19%  perf-profile.children.cycles-pp.__lock_parent
>>>        0.08 ± 14%      +0.2        0.26 ±  4%  perf-profile.children.cycles-pp.rwsem_spin_on_owner
>>>        0.09 ± 15%      +0.2        0.28 ±  4%  perf-profile.children.cycles-pp.release_pages
>>>        0.09 ± 19%      +0.2        0.28 ±  5%  perf-profile.children.cycles-pp.page_counter_try_charge
>>>        0.21 ± 21%      +0.2        0.41 ± 10%  perf-profile.children.cycles-pp.div_int
>>>        0.39 ± 14%      +0.2        0.58 ± 10%  perf-profile.children.cycles-pp.update_process_times
>>>        0.39 ± 14%      +0.2        0.59 ± 10%  perf-profile.children.cycles-pp.tick_sched_handle
>>>        0.09 ± 20%      +0.2        0.29 ±  3%  perf-profile.children.cycles-pp.__memcg_kmem_charge
>>>        0.08 ± 17%      +0.2        0.28 ±  3%  perf-profile.children.cycles-pp.anon_vma_fork
>>>        0.16 ± 12%      +0.2        0.36 ± 10%  perf-profile.children.cycles-pp.mem_rtns_1
>>>        0.25 ± 20%      +0.2        0.46 ± 10%  perf-profile.children.cycles-pp.div_short
>>>        0.07 ± 25%      +0.2        0.28 ±  4%  perf-profile.children.cycles-pp.remove_vma
>>>        0.43 ± 14%      +0.2        0.64 ± 10%  perf-profile.children.cycles-pp.tick_sched_timer
>>>        0.14 ±  3%      +0.2        0.36 ±  5%  perf-profile.children.cycles-pp.setlocale
>>>        0.09 ±  9%      +0.2        0.31        perf-profile.children.cycles-pp.__schedule
>>>        0.11 ± 12%      +0.2        0.33 ±  4%  perf-profile.children.cycles-pp.tlb_flush_mmu
>>>        0.11 ± 16%      +0.2        0.34 ±  5%  perf-profile.children.cycles-pp.tlb_finish_mmu
>>>        0.10 ±  4%      +0.2        0.33 ± 11%  perf-profile.children.cycles-pp.__fput
>>>        0.11 ± 11%      +0.2        0.35 ±  4%  perf-profile.children.cycles-pp.__alloc_pages_nodemask
>>>        0.52 ± 14%      +0.2        0.76 ± 12%  perf-profile.children.cycles-pp.__hrtimer_run_queues
>>>        0.22 ± 19%      +0.2        0.46 ± 11%  perf-profile.children.cycles-pp.add_short
>>>        0.13 ± 10%      +0.3        0.38 ±  5%  perf-profile.children.cycles-pp.lookup_fast
>>>        0.09 ± 15%      +0.3        0.35        perf-profile.children.cycles-pp.unlink_anon_vmas
>>>        0.09 ± 19%      +0.3        0.34 ±  4%  perf-profile.children.cycles-pp.get_obj_cgroup_from_current
>>>        0.12 ±  3%      +0.3        0.38 ± 10%  perf-profile.children.cycles-pp.task_work_run
>>>        0.09 ± 20%      +0.3        0.35        perf-profile.children.cycles-pp.iterate_supers
>>>        0.09 ± 20%      +0.3        0.36 ±  2%  perf-profile.children.cycles-pp.__x64_sys_sync
>>>        0.09 ± 20%      +0.3        0.36 ±  2%  perf-profile.children.cycles-pp.ksys_sync
>>>        0.12 ± 11%      +0.3        0.39 ±  2%  perf-profile.children.cycles-pp.d_alloc
>>>        0.09 ± 20%      +0.3        0.36 ±  2%  perf-profile.children.cycles-pp.sync
>>>        0.75 ± 12%      +0.3        1.03 ± 11%  perf-profile.children.cycles-pp.hrtimer_interrupt
>>>        0.75 ± 12%      +0.3        1.04 ± 12%  perf-profile.children.cycles-pp.__sysvec_apic_timer_interrupt
>>>        0.15 ±  3%      +0.3        0.45 ±  8%  perf-profile.children.cycles-pp.exit_to_user_mode_prepare
>>>        0.09 ± 20%      +0.3        0.39 ±  3%  perf-profile.children.cycles-pp.load_elf_interp
>>>        0.19 ± 12%      +0.3        0.49 ±  3%  perf-profile.children.cycles-pp.zap_pte_range
>>>        0.20 ± 11%      +0.3        0.51 ±  3%  perf-profile.children.cycles-pp.unmap_page_range
>>>        0.52 ± 19%      +0.3        0.84 ±  3%  perf-profile.children.cycles-pp.creat64
>>>        0.21 ±  8%      +0.3        0.54 ±  3%  perf-profile.children.cycles-pp.unmap_vmas
>>>        0.09 ± 24%      +0.3        0.42 ± 11%  perf-profile.children.cycles-pp.propagate_protected_usage
>>>        0.28 ± 14%      +0.3        0.62 ± 10%  perf-profile.children.cycles-pp.__strcat_chk
>>>        0.25 ±  7%      +0.3        0.59 ±  4%  perf-profile.children.cycles-pp.path_lookupat
>>>        0.25 ±  8%      +0.4        0.60 ±  4%  perf-profile.children.cycles-pp.filename_lookup
>>>        0.26 ±  8%      +0.4        0.62 ±  4%  perf-profile.children.cycles-pp.__do_sys_newstat
>>>        0.19 ± 11%      +0.4        0.56 ±  4%  perf-profile.children.cycles-pp.__alloc_file
>>>        0.19 ±  9%      +0.4        0.56 ±  3%  perf-profile.children.cycles-pp.alloc_empty_file
>>>        0.26 ±  8%      +0.4        0.63 ±  4%  perf-profile.children.cycles-pp.vfs_statx
>>>        0.12 ± 15%      +0.4        0.49 ±  2%  perf-profile.children.cycles-pp.elf_map
>>>        0.22 ± 11%      +0.4        0.60 ±  2%  perf-profile.children.cycles-pp.d_alloc_parallel
>>>        0.29 ±  8%      +0.4        0.68 ±  3%  perf-profile.children.cycles-pp.filemap_map_pages
>>>        0.24 ±  9%      +0.4        0.63 ±  2%  perf-profile.children.cycles-pp.__lookup_slow
>>>        0.30 ± 32%      +0.4        0.73 ± 41%  perf-profile.children.cycles-pp.syscall_exit_to_user_mode
>>>        0.49 ± 20%      +0.4        0.93 ± 10%  perf-profile.children.cycles-pp.div_long
>>>        0.32 ± 10%      +0.5        0.78 ±  3%  perf-profile.children.cycles-pp.do_fault
>>>        0.52 ± 17%      +0.5        1.04 ±  9%  perf-profile.children.cycles-pp.add_int
>>>        0.51 ± 18%      +0.5        1.04 ±  8%  perf-profile.children.cycles-pp.add_long
>>>        0.50 ± 14%      +0.6        1.12 ± 11%  perf-profile.children.cycles-pp.__strncat_chk
>>>        0.12 ± 14%      +0.6        0.74 ± 10%  perf-profile.children.cycles-pp.terminate_walk
>>>        0.39 ± 19%      +0.6        1.02 ± 18%  perf-profile.children.cycles-pp.mprotect_fixup
>>>        0.39 ± 18%      +0.6        1.03 ± 18%  perf-profile.children.cycles-pp.__x64_sys_mprotect
>>>        0.39 ± 18%      +0.6        1.03 ± 18%  perf-profile.children.cycles-pp.do_mprotect_pkey
>>>        0.21 ± 23%      +0.7        0.90 ± 13%  perf-profile.children.cycles-pp.__memcg_kmem_uncharge
>>>        0.46 ± 21%      +0.7        1.20 ± 21%  perf-profile.children.cycles-pp.unmap_region
>>>        0.30 ± 17%      +0.7        1.04 ±  3%  perf-profile.children.cycles-pp.kmem_cache_alloc
>>>        0.48 ±  9%      +0.8        1.23 ±  3%  perf-profile.children.cycles-pp.__handle_mm_fault
>>>        0.50 ± 10%      +0.8        1.30 ±  3%  perf-profile.children.cycles-pp.handle_mm_fault
>>>        0.53 ±  9%      +0.8        1.37 ±  3%  perf-profile.children.cycles-pp.do_user_addr_fault
>>>        0.47 ±  9%      +0.8        1.30 ±  2%  perf-profile.children.cycles-pp.link_path_walk
>>>        0.54 ±  8%      +0.8        1.38 ±  3%  perf-profile.children.cycles-pp.exc_page_fault
>>>        0.48 ±  9%      +0.9        1.33 ±  3%  perf-profile.children.cycles-pp.walk_component
>>>        0.49 ± 18%      +0.9        1.34 ± 13%  perf-profile.children.cycles-pp.__libc_fork
>>>        0.50 ± 20%      +0.9        1.41 ± 16%  perf-profile.children.cycles-pp.vma_link
>>>        0.59 ±  4%      +0.9        1.51 ±  4%  perf-profile.children.cycles-pp.asm_exc_page_fault
>>>        0.34 ± 22%      +1.1        1.43 ± 12%  perf-profile.children.cycles-pp.page_counter_cancel
>>>        0.72 ± 19%      +1.2        1.88 ± 19%  perf-profile.children.cycles-pp.__vma_adjust
>>>        0.36 ± 22%      +1.2        1.54 ± 12%  perf-profile.children.cycles-pp.page_counter_uncharge
>>>        0.74 ± 19%      +1.2        1.94 ± 19%  perf-profile.children.cycles-pp.__split_vma
>>>        0.37 ± 21%      +1.2        1.59 ±  5%  perf-profile.children.cycles-pp.irq_exit_rcu
>>>        0.36 ± 22%      +1.2        1.59 ±  5%  perf-profile.children.cycles-pp.do_softirq_own_stack
>>>        0.00            +1.2        1.23 ± 19%  perf-profile.children.cycles-pp.dcache_dir_open
>>>        0.00            +1.2        1.23 ± 19%  perf-profile.children.cycles-pp.d_alloc_cursor
>>>        0.69 ± 15%      +1.3        1.96 ± 14%  perf-profile.children.cycles-pp.begin_new_exec
>>>        0.85 ± 20%      +1.4        2.21 ± 20%  perf-profile.children.cycles-pp.__do_munmap
>>>        0.81 ± 16%      +1.4        2.22 ± 13%  perf-profile.children.cycles-pp.__x64_sys_exit_group
>>>        0.81 ± 16%      +1.4        2.22 ± 13%  perf-profile.children.cycles-pp.do_group_exit
>>>        0.81 ± 16%      +1.4        2.22 ± 13%  perf-profile.children.cycles-pp.do_exit
>>>        1.09 ± 16%      +1.4        2.50 ±  8%  perf-profile.children.cycles-pp.asm_call_sysvec_on_stack
>>>        0.84 ± 20%      +1.4        2.29 ± 15%  perf-profile.children.cycles-pp.dup_mmap
>>>        0.39 ± 20%      +1.5        1.84 ± 12%  perf-profile.children.cycles-pp.drain_obj_stock
>>>        0.85 ± 20%      +1.5        2.32 ± 15%  perf-profile.children.cycles-pp.dup_mm
>>>        1.18 ± 15%      +1.6        2.74 ±  8%  perf-profile.children.cycles-pp.sysvec_apic_timer_interrupt
>>>        0.94 ± 18%      +1.6        2.58 ± 13%  perf-profile.children.cycles-pp.copy_process
>>>        1.29 ± 11%      +1.6        2.94 ±  7%  perf-profile.children.cycles-pp.asm_sysvec_apic_timer_interrupt
>>>        1.00 ± 18%      +1.7        2.72 ± 13%  perf-profile.children.cycles-pp.__do_sys_clone
>>>        1.00 ± 18%      +1.7        2.72 ± 13%  perf-profile.children.cycles-pp.kernel_clone
>>>        0.47 ± 18%      +1.8        2.25 ± 12%  perf-profile.children.cycles-pp.refill_obj_stock
>>>        0.92 ± 14%      +1.9        2.78 ± 10%  perf-profile.children.cycles-pp.load_elf_binary
>>>        0.93 ± 14%      +1.9        2.79 ± 10%  perf-profile.children.cycles-pp.exec_binprm
>>>        1.68 ± 14%      +2.1        3.74 ± 10%  perf-profile.children.cycles-pp.string_rtns_1
>>>        1.30 ± 20%      +2.1        3.37 ± 20%  perf-profile.children.cycles-pp.ksys_mmap_pgoff
>>>        1.35 ± 19%      +2.2        3.51 ± 24%  perf-profile.children.cycles-pp.unlink_file_vma
>>>        1.36 ± 20%      +2.3        3.63 ± 18%  perf-profile.children.cycles-pp.mmap_region
>>>        1.37 ± 20%      +2.3        3.67 ± 18%  perf-profile.children.cycles-pp.do_mmap
>>>        1.39 ± 20%      +2.3        3.72 ± 18%  perf-profile.children.cycles-pp.vm_mmap_pgoff
>>>        1.46 ± 19%      +2.5        3.92 ± 21%  perf-profile.children.cycles-pp.free_pgtables
>>>        1.40 ± 16%      +2.5        3.88 ± 14%  perf-profile.children.cycles-pp.exit_mmap
>>>        1.40 ± 16%      +2.5        3.89 ± 14%  perf-profile.children.cycles-pp.mmput
>>>        0.00            +2.7        2.65 ±  9%  perf-profile.children.cycles-pp.lockref_get
>>>        0.08 ±  5%      +2.7        2.80 ±  8%  perf-profile.children.cycles-pp.do_dentry_open
>>>        0.61 ± 11%      +3.2        3.79 ± 15%  perf-profile.children.cycles-pp.run_ksoftirqd
>>>        0.61 ± 12%      +3.2        3.82 ± 15%  perf-profile.children.cycles-pp.smpboot_thread_fn
>>>        0.65 ± 11%      +3.2        3.88 ± 15%  perf-profile.children.cycles-pp.kthread
>>>        0.66 ± 11%      +3.3        3.92 ± 15%  perf-profile.children.cycles-pp.ret_from_fork
>>>        0.91 ± 14%      +4.3        5.25 ± 12%  perf-profile.children.cycles-pp.rcu_do_batch
>>>        0.92 ± 13%      +4.3        5.25 ± 12%  perf-profile.children.cycles-pp.rcu_core
>>>        0.97 ± 13%      +4.4        5.38 ± 12%  perf-profile.children.cycles-pp.__softirqentry_text_start
>>>        0.99 ± 14%      +4.5        5.52 ± 10%  perf-profile.children.cycles-pp.kmem_cache_free
>>>        3.02 ± 19%      +4.7        7.76 ± 23%  perf-profile.children.cycles-pp.osq_lock
>>>        3.15 ± 19%      +5.0        8.16 ± 22%  perf-profile.children.cycles-pp.rwsem_down_write_slowpath
>>>        3.33 ± 10%      +6.1        9.38 ± 10%  perf-profile.children.cycles-pp.intel_idle
>>>        0.00            +6.4        6.37 ±  2%  perf-profile.children.cycles-pp.complete_walk
>>>        0.13 ±  3%      +7.1        7.19 ± 18%  perf-profile.children.cycles-pp.__open64_nocancel
>>>        3.89 ±  9%      +7.1       10.99 ±  8%  perf-profile.children.cycles-pp.cpuidle_enter_state
>>>        3.89 ±  9%      +7.1       11.00 ±  8%  perf-profile.children.cycles-pp.cpuidle_enter
>>>        0.28 ±  3%      +7.2        7.48 ±  6%  perf-profile.children.cycles-pp.lockref_put_or_lock
>>>        4.17 ±  6%      +7.4       11.59 ±  8%  perf-profile.children.cycles-pp.start_secondary
>>>        4.20 ±  6%      +7.4       11.63 ±  8%  perf-profile.children.cycles-pp.secondary_startup_64_no_verify
>>>        4.20 ±  6%      +7.4       11.63 ±  8%  perf-profile.children.cycles-pp.cpu_startup_entry
>>>        4.20 ±  6%      +7.4       11.63 ±  8%  perf-profile.children.cycles-pp.do_idle
>>>        0.54 ±  4%      +7.5        8.04 ±  6%  perf-profile.children.cycles-pp.dput
>>>        1.12 ±  2%     +11.1       12.26 ±  2%  perf-profile.children.cycles-pp.do_sys_open
>>>        1.12 ±  2%     +11.1       12.25 ±  2%  perf-profile.children.cycles-pp.do_sys_openat2
>>>        0.06 ± 14%     +24.2       24.28 ±  4%  perf-profile.children.cycles-pp.pick_link
>>>        1.03 ±  4%     +24.4       25.43 ±  8%  perf-profile.children.cycles-pp.execve
>>>        0.17 ± 10%     +30.8       30.92 ±  3%  perf-profile.children.cycles-pp.lockref_get_not_dead
>>>        0.10 ±  8%     +30.8       30.87 ±  3%  perf-profile.children.cycles-pp.do_open_execat
>>>        0.18 ±  9%     +30.8       30.95 ±  3%  perf-profile.children.cycles-pp.__legitimize_path
>>>        0.18 ± 12%     +30.8       30.96 ±  3%  perf-profile.children.cycles-pp.unlazy_walk
>>>        0.20 ± 10%     +31.0       31.15 ±  3%  perf-profile.children.cycles-pp.step_into
>>>        1.05 ± 14%     +32.7       33.76 ±  3%  perf-profile.children.cycles-pp.bprm_execve
>>>        1.14 ± 13%     +32.9       33.99 ±  3%  perf-profile.children.cycles-pp.do_execveat_common
>>>        1.15 ± 13%     +32.9       34.01 ±  3%  perf-profile.children.cycles-pp.__x64_sys_execve
>>>        1.16 ±  3%     +41.8       42.98 ±  2%  perf-profile.children.cycles-pp.path_openat
>>>        1.17 ±  3%     +41.8       42.98 ±  2%  perf-profile.children.cycles-pp.do_filp_open
>>>        0.00           +54.2       54.20 ±  2%  perf-profile.children.cycles-pp.__cna_queued_spin_lock_slowpath
>>>       81.68           -81.7        0.00        perf-profile.self.cycles-pp.native_queued_spin_lock_slowpath
>>>        0.05 ±  8%      +0.0        0.08 ± 21%  perf-profile.self.cycles-pp.ktime_get_update_offsets_now
>>>        0.00            +0.1        0.05        perf-profile.self.cycles-pp.link_path_walk
>>>        0.00            +0.1        0.05        perf-profile.self.cycles-pp.apparmor_task_getsecid
>>>        0.00            +0.1        0.05        perf-profile.self.cycles-pp.page_add_file_rmap
>>>        0.00            +0.1        0.05        perf-profile.self.cycles-pp.common_file_perm
>>>        0.00            +0.1        0.05        perf-profile.self.cycles-pp.__list_del_entry_valid
>>>        0.00            +0.1        0.05        perf-profile.self.cycles-pp.copy_page
>>>        0.00            +0.1        0.05        perf-profile.self.cycles-pp._raw_spin_trylock
>>>        0.00            +0.1        0.05 ±  8%  perf-profile.self.cycles-pp.__rb_insert_augmented
>>>        0.00            +0.1        0.05 ±  8%  perf-profile.self.cycles-pp.compar1
>>>        0.00            +0.1        0.06 ±  9%  perf-profile.self.cycles-pp.__might_sleep
>>>        0.00            +0.1        0.06 ±  9%  perf-profile.self.cycles-pp.sync_regs
>>>        0.00            +0.1        0.06 ± 14%  perf-profile.self.cycles-pp.__alloc_file
>>>        0.00            +0.1        0.06 ±  7%  perf-profile.self.cycles-pp.kfree
>>>        0.00            +0.1        0.06        perf-profile.self.cycles-pp.strnlen_user
>>>        0.00            +0.1        0.06        perf-profile.self.cycles-pp.anon_vma_interval_tree_insert
>>>        0.00            +0.1        0.06 ± 11%  perf-profile.self.cycles-pp.memcpy_erms
>>>        0.00            +0.1        0.06 ±  6%  perf-profile.self.cycles-pp.free_pages_and_swap_cache
>>>        0.00            +0.1        0.07 ±  7%  perf-profile.self.cycles-pp.filldir64
>>>        0.00            +0.1        0.07 ±  7%  perf-profile.self.cycles-pp.rcu_sched_clock_irq
>>>        0.06 ± 16%      +0.1        0.12 ±  3%  perf-profile.self.cycles-pp.__d_lookup_rcu
>>>        0.00            +0.1        0.07 ±  6%  perf-profile.self.cycles-pp.copy_pte_range
>>>        0.00            +0.1        0.07 ±  6%  perf-profile.self.cycles-pp._raw_spin_lock_irqsave
>>>        0.00            +0.1        0.07 ± 19%  perf-profile.self.cycles-pp.file_free_rcu
>>>        0.00            +0.1        0.07 ± 10%  perf-profile.self.cycles-pp.rwsem_down_write_slowpath
>>>        0.05 ±  8%      +0.1        0.12 ±  4%  perf-profile.self.cycles-pp.apparmor_file_open
>>>        0.05            +0.1        0.12 ±  3%  perf-profile.self.cycles-pp.apparmor_file_free_security
>>>        0.00            +0.1        0.07 ±  5%  perf-profile.self.cycles-pp.vma_interval_tree_remove
>>>        0.07            +0.1        0.15 ±  3%  perf-profile.self.cycles-pp.native_irq_return_iret
>>>        0.00            +0.1        0.08 ±  6%  perf-profile.self.cycles-pp.__mod_memcg_state
>>>        0.00            +0.1        0.08 ± 11%  perf-profile.self.cycles-pp.rcu_cblist_dequeue
>>>        0.00            +0.1        0.08 ± 14%  perf-profile.self.cycles-pp.__handle_mm_fault
>>>        0.00            +0.1        0.08 ±  5%  perf-profile.self.cycles-pp.down_read
>>>        0.02 ±141%      +0.1        0.10 ±  5%  perf-profile.self.cycles-pp.up_write
>>>        0.00            +0.1        0.08 ± 13%  perf-profile.self.cycles-pp.__slab_free
>>>        0.06 ±  8%      +0.1        0.14 ±  8%  perf-profile.self.cycles-pp.lockref_get_not_dead
>>>        0.00            +0.1        0.08 ±  5%  perf-profile.self.cycles-pp.clear_page_erms
>>>        0.03 ± 70%      +0.1        0.12 ±  6%  perf-profile.self.cycles-pp.down_write
>>>        0.00            +0.1        0.09 ±  7%  perf-profile.self.cycles-pp.malloc
>>>        0.00            +0.1        0.09 ±  7%  perf-profile.self.cycles-pp.aa_get_task_label
>>>        0.02 ±141%      +0.1        0.11 ±  4%  perf-profile.self.cycles-pp.___might_sleep
>>>        0.09 ± 13%      +0.1        0.19 ± 14%  perf-profile.self.cycles-pp.update_cfs_group
>>>        0.06 ± 13%      +0.1        0.15 ±  5%  perf-profile.self.cycles-pp.page_remove_rmap
>>>        0.07 ± 11%      +0.1        0.17 ±  2%  perf-profile.self.cycles-pp.find_idlest_group
>>>        0.00            +0.1        0.10 ±  4%  perf-profile.self.cycles-pp.lockref_put_return
>>>        0.07 ±  7%      +0.1        0.17 ±  6%  perf-profile.self.cycles-pp.apparmor_file_alloc_security
>>>        0.00            +0.1        0.11 ±  7%  perf-profile.self.cycles-pp.update_sd_lb_stats
>>>        0.08 ± 10%      +0.1        0.19 ±  3%  perf-profile.self.cycles-pp.vma_interval_tree_insert
>>>        0.04 ± 71%      +0.1        0.15 ±  5%  perf-profile.self.cycles-pp.release_pages
>>>        0.00            +0.1        0.12 ± 13%  perf-profile.self.cycles-pp.page_counter_uncharge
>>>        0.07 ± 11%      +0.1        0.19 ±  2%  perf-profile.self.cycles-pp._dl_addr
>>>        0.00            +0.1        0.13 ±  3%  perf-profile.self.cycles-pp.cna_order_queue
>>>        0.10 ±  8%      +0.1        0.25 ±  2%  perf-profile.self.cycles-pp.zap_pte_range
>>>        0.08 ± 22%      +0.2        0.23 ±  4%  perf-profile.self.cycles-pp.page_counter_try_charge
>>>        0.08 ± 14%      +0.2        0.25 ±  2%  perf-profile.self.cycles-pp.rwsem_spin_on_owner
>>>        0.21 ± 20%      +0.2        0.40 ± 11%  perf-profile.self.cycles-pp.div_int
>>>        0.15 ±  9%      +0.2        0.35 ± 11%  perf-profile.self.cycles-pp.mem_rtns_1
>>>        0.25 ± 22%      +0.2        0.45 ± 11%  perf-profile.self.cycles-pp.div_short
>>>        0.22 ± 19%      +0.2        0.45 ± 10%  perf-profile.self.cycles-pp.add_short
>>>        0.08 ± 27%      +0.2        0.32 ±  4%  perf-profile.self.cycles-pp.get_obj_cgroup_from_current
>>>        0.42 ±  2%      +0.3        0.69 ±  2%  perf-profile.self.cycles-pp._raw_spin_lock
>>>        0.22 ±  9%      +0.3        0.49 ±  2%  perf-profile.self.cycles-pp.filemap_map_pages
>>>        0.11 ± 19%      +0.3        0.41 ±  3%  perf-profile.self.cycles-pp.kmem_cache_alloc
>>>        0.07 ±  6%      +0.3        0.40 ± 10%  perf-profile.self.cycles-pp.refill_obj_stock
>>>        0.28 ± 14%      +0.3        0.60 ± 11%  perf-profile.self.cycles-pp.__strcat_chk
>>>        0.09 ± 24%      +0.3        0.42 ± 11%  perf-profile.self.cycles-pp.propagate_protected_usage
>>>        0.07 ± 11%      +0.4        0.46 ± 10%  perf-profile.self.cycles-pp.drain_obj_stock
>>>        0.49 ± 19%      +0.4        0.92 ± 10%  perf-profile.self.cycles-pp.div_long
>>>        0.51 ± 17%      +0.5        1.01 ±  8%  perf-profile.self.cycles-pp.add_long
>>>        0.51 ± 17%      +0.5        1.02 ±  9%  perf-profile.self.cycles-pp.add_int
>>>        0.49 ± 14%      +0.6        1.10 ± 11%  perf-profile.self.cycles-pp.__strncat_chk
>>>        0.26 ± 21%      +0.8        1.04 ± 12%  perf-profile.self.cycles-pp.page_counter_cancel
>>>        1.66 ± 14%      +2.0        3.65 ± 10%  perf-profile.self.cycles-pp.string_rtns_1
>>>        0.49 ± 13%      +2.6        3.11 ±  9%  perf-profile.self.cycles-pp.kmem_cache_free
>>>        2.93 ± 19%      +4.6        7.50 ± 22%  perf-profile.self.cycles-pp.osq_lock
>>>        3.33 ± 10%      +6.0        9.38 ± 10%  perf-profile.self.cycles-pp.intel_idle
>>>        0.00           +53.4       53.38 ±  2%  perf-profile.self.cycles-pp.__cna_queued_spin_lock_slowpath
>>>        1579 ± 53%    +572.8%      10624 ± 82%  interrupts.31:PCI-MSI.524289-edge.eth0-TxRx-0
>>>        1203 ±112%     +61.3%       1940 ± 68%  interrupts.32:PCI-MSI.524290-edge.eth0-TxRx-1
>>>        8340 ± 68%     -76.0%       2003 ± 74%  interrupts.34:PCI-MSI.524292-edge.eth0-TxRx-3
>>>      775419 ±  5%     +21.9%     944875 ±  2%  interrupts.CAL:Function_call_interrupts
>>>        5146 ±  4%     -51.2%       2512 ± 34%  interrupts.CPU0.NMI:Non-maskable_interrupts
>>>        5146 ±  4%     -51.2%       2512 ± 34%  interrupts.CPU0.PMI:Performance_monitoring_interrupts
>>>        1113 ±  8%     +13.1%       1259 ±  4%  interrupts.CPU0.RES:Rescheduling_interrupts
>>>      643.67 ±  4%     +53.6%     988.50 ±  7%  interrupts.CPU0.TLB:TLB_shootdowns
>>>        4043 ±  7%     +33.6%       5400 ±  5%  interrupts.CPU1.CAL:Function_call_interrupts
>>>        5177 ±  3%     -41.7%       3020 ± 36%  interrupts.CPU1.NMI:Non-maskable_interrupts
>>>        5177 ±  3%     -41.7%       3020 ± 36%  interrupts.CPU1.PMI:Performance_monitoring_interrupts
>>>      735.33 ±  5%     +48.3%       1090 ±  8%  interrupts.CPU1.RES:Rescheduling_interrupts
>>>      562.00 ± 12%     +47.5%     829.00 ±  6%  interrupts.CPU1.TLB:TLB_shootdowns
>>>        1579 ± 53%    +572.8%      10624 ± 82%  interrupts.CPU10.31:PCI-MSI.524289-edge.eth0-TxRx-0
>>>        3821 ±  7%     +28.3%       4903 ±  6%  interrupts.CPU10.CAL:Function_call_interrupts
>>>        5187 ±  4%     -23.5%       3968 ±  5%  interrupts.CPU10.NMI:Non-maskable_interrupts
>>>        5187 ±  4%     -23.5%       3968 ±  5%  interrupts.CPU10.PMI:Performance_monitoring_interrupts
>>>      738.67 ±  5%     +48.7%       1098 ±  4%  interrupts.CPU10.RES:Rescheduling_interrupts
>>>      522.00 ± 33%     +72.4%     900.00 ± 10%  interrupts.CPU10.TLB:TLB_shootdowns
>>>        3934 ±  9%     +24.9%       4913 ±  6%  interrupts.CPU100.CAL:Function_call_interrupts
>>>        5204 ±  3%     -24.7%       3916 ±  8%  interrupts.CPU100.NMI:Non-maskable_interrupts
>>>        5204 ±  3%     -24.7%       3916 ±  8%  interrupts.CPU100.PMI:Performance_monitoring_interrupts
>>>      743.67 ±  7%     +61.4%       1200 ±  9%  interrupts.CPU100.RES:Rescheduling_interrupts
>>>      474.00 ± 14%     +91.8%     909.00 ± 12%  interrupts.CPU100.TLB:TLB_shootdowns
>>>        3945 ±  4%     +24.3%       4901 ±  4%  interrupts.CPU101.CAL:Function_call_interrupts
>>>        5197 ±  3%     -36.2%       3313 ± 26%  interrupts.CPU101.NMI:Non-maskable_interrupts
>>>        5197 ±  3%     -36.2%       3313 ± 26%  interrupts.CPU101.PMI:Performance_monitoring_interrupts
>>>      720.33 ±  5%     +60.8%       1158 ±  7%  interrupts.CPU101.RES:Rescheduling_interrupts
>>>      496.67 ±  7%     +82.6%     906.75 ± 15%  interrupts.CPU101.TLB:TLB_shootdowns
>>>        3918 ±  5%     +22.8%       4810 ±  6%  interrupts.CPU102.CAL:Function_call_interrupts
>>>        5186 ±  4%     -25.2%       3879 ± 10%  interrupts.CPU102.NMI:Non-maskable_interrupts
>>>        5186 ±  4%     -25.2%       3879 ± 10%  interrupts.CPU102.PMI:Performance_monitoring_interrupts
>>>      699.33 ±  3%     +61.8%       1131 ±  7%  interrupts.CPU102.RES:Rescheduling_interrupts
>>>      406.33 ±  7%    +108.2%     846.00 ± 12%  interrupts.CPU102.TLB:TLB_shootdowns
>>>        3882 ±  5%     +26.2%       4897 ±  4%  interrupts.CPU103.CAL:Function_call_interrupts
>>>        5224 ±  3%     -34.9%       3399 ± 26%  interrupts.CPU103.NMI:Non-maskable_interrupts
>>>        5224 ±  3%     -34.9%       3399 ± 26%  interrupts.CPU103.PMI:Performance_monitoring_interrupts
>>>      658.33 ±  4%     +73.4%       1141 ±  5%  interrupts.CPU103.RES:Rescheduling_interrupts
>>>      449.00 ± 16%     +89.6%     851.50 ±  8%  interrupts.CPU103.TLB:TLB_shootdowns
>>>        3881 ±  3%     +23.9%       4807 ±  3%  interrupts.CPU104.CAL:Function_call_interrupts
>>>        5212 ±  3%     -26.6%       3826 ±  8%  interrupts.CPU104.NMI:Non-maskable_interrupts
>>>        5212 ±  3%     -26.6%       3826 ±  8%  interrupts.CPU104.PMI:Performance_monitoring_interrupts
>>>      637.67 ±  7%     +73.6%       1106 ±  9%  interrupts.CPU104.RES:Rescheduling_interrupts
>>>      423.33 ±  7%     +90.1%     804.75 ± 14%  interrupts.CPU104.TLB:TLB_shootdowns
>>>        3845           +20.2%       4623        interrupts.CPU105.CAL:Function_call_interrupts
>>>      661.33 ±  8%     +68.5%       1114 ±  2%  interrupts.CPU105.RES:Rescheduling_interrupts
>>>      467.33 ± 17%     +87.6%     876.50 ± 21%  interrupts.CPU105.TLB:TLB_shootdowns
>>>        3959 ±  4%     +18.2%       4679        interrupts.CPU106.CAL:Function_call_interrupts
>>>        5251 ±  3%     -25.6%       3905 ± 10%  interrupts.CPU106.NMI:Non-maskable_interrupts
>>>        5251 ±  3%     -25.6%       3905 ± 10%  interrupts.CPU106.PMI:Performance_monitoring_interrupts
>>>      721.33 ±  5%     +59.2%       1148 ±  8%  interrupts.CPU106.RES:Rescheduling_interrupts
>>>      419.00 ± 15%    +150.8%       1050 ± 30%  interrupts.CPU106.TLB:TLB_shootdowns
>>>        3930 ±  5%     +18.9%       4674 ±  4%  interrupts.CPU107.CAL:Function_call_interrupts
>>>        5177 ±  5%     -34.4%       3398 ± 26%  interrupts.CPU107.NMI:Non-maskable_interrupts
>>>        5177 ±  5%     -34.4%       3398 ± 26%  interrupts.CPU107.PMI:Performance_monitoring_interrupts
>>>      665.33 ±  5%     +75.8%       1169 ±  4%  interrupts.CPU107.RES:Rescheduling_interrupts
>>>      445.00 ± 18%     +83.9%     818.50 ± 13%  interrupts.CPU107.TLB:TLB_shootdowns
>>>        3915 ±  8%     +20.9%       4735 ±  3%  interrupts.CPU108.CAL:Function_call_interrupts
>>>        5255 ±  3%     -36.2%       3350 ± 25%  interrupts.CPU108.NMI:Non-maskable_interrupts
>>>        5255 ±  3%     -36.2%       3350 ± 25%  interrupts.CPU108.PMI:Performance_monitoring_interrupts
>>>      684.33 ±  7%     +67.5%       1146 ±  7%  interrupts.CPU108.RES:Rescheduling_interrupts
>>>      509.00 ± 11%     +66.4%     846.75 ±  6%  interrupts.CPU108.TLB:TLB_shootdowns
>>>        3965 ±  3%     +19.8%       4749 ±  4%  interrupts.CPU109.CAL:Function_call_interrupts
>>>        5207 ±  3%     -42.8%       2976 ± 38%  interrupts.CPU109.NMI:Non-maskable_interrupts
>>>        5207 ±  3%     -42.8%       2976 ± 38%  interrupts.CPU109.PMI:Performance_monitoring_interrupts
>>>      701.00 ±  4%     +55.2%       1087 ±  5%  interrupts.CPU109.RES:Rescheduling_interrupts
>>>      510.33 ± 11%     +71.9%     877.25 ±  6%  interrupts.CPU109.TLB:TLB_shootdowns
>>>        1203 ±112%     +61.3%       1940 ± 68%  interrupts.CPU11.32:PCI-MSI.524290-edge.eth0-TxRx-1
>>>        3794 ±  5%     +31.8%       4999 ±  4%  interrupts.CPU11.CAL:Function_call_interrupts
>>>      734.33 ±  6%     +51.7%       1114 ±  3%  interrupts.CPU11.RES:Rescheduling_interrupts
>>>      458.67 ± 14%    +100.9%     921.50 ± 10%  interrupts.CPU11.TLB:TLB_shootdowns
>>>        3949 ±  5%     +27.3%       5029 ±  4%  interrupts.CPU110.CAL:Function_call_interrupts
>>>        5230 ±  3%     -35.3%       3385 ± 24%  interrupts.CPU110.NMI:Non-maskable_interrupts
>>>        5230 ±  3%     -35.3%       3385 ± 24%  interrupts.CPU110.PMI:Performance_monitoring_interrupts
>>>      670.67 ±  6%     +74.5%       1170 ±  7%  interrupts.CPU110.RES:Rescheduling_interrupts
>>>      442.00 ±  8%     +88.5%     833.00 ± 18%  interrupts.CPU110.TLB:TLB_shootdowns
>>>        3865 ±  4%     +24.2%       4802 ±  4%  interrupts.CPU111.CAL:Function_call_interrupts
>>>        5230 ±  3%     -36.3%       3330 ± 25%  interrupts.CPU111.NMI:Non-maskable_interrupts
>>>        5230 ±  3%     -36.3%       3330 ± 25%  interrupts.CPU111.PMI:Performance_monitoring_interrupts
>>>      634.00 ±  9%     +85.7%       1177 ± 10%  interrupts.CPU111.RES:Rescheduling_interrupts
>>>      441.33 ± 13%     +96.4%     866.75 ± 14%  interrupts.CPU111.TLB:TLB_shootdowns
>>>        3873 ±  5%     +24.7%       4829 ±  2%  interrupts.CPU112.CAL:Function_call_interrupts
>>>        5121 ±  4%     -35.0%       3326 ± 24%  interrupts.CPU112.NMI:Non-maskable_interrupts
>>>        5121 ±  4%     -35.0%       3326 ± 24%  interrupts.CPU112.PMI:Performance_monitoring_interrupts
>>>      669.00 ±  2%     +72.3%       1152 ±  7%  interrupts.CPU112.RES:Rescheduling_interrupts
>>>      430.00 ±  2%    +111.9%     911.00 ± 15%  interrupts.CPU112.TLB:TLB_shootdowns
>>>        3921 ±  6%     +24.6%       4885 ±  3%  interrupts.CPU113.CAL:Function_call_interrupts
>>>      672.33 ±  8%     +70.0%       1142 ±  7%  interrupts.CPU113.RES:Rescheduling_interrupts
>>>      431.33 ±  9%     +98.2%     855.00 ± 11%  interrupts.CPU113.TLB:TLB_shootdowns
>>>        3779 ±  7%     +27.7%       4827 ±  4%  interrupts.CPU114.CAL:Function_call_interrupts
>>>      641.00 ± 13%     +79.3%       1149 ±  4%  interrupts.CPU114.RES:Rescheduling_interrupts
>>>      504.00 ± 10%     +77.2%     893.25 ± 20%  interrupts.CPU114.TLB:TLB_shootdowns
>>>        3904 ±  6%     +20.2%       4693        interrupts.CPU115.CAL:Function_call_interrupts
>>>      651.67 ±  3%     +78.9%       1165 ±  2%  interrupts.CPU115.RES:Rescheduling_interrupts
>>>      501.67 ± 29%     +69.9%     852.50 ±  4%  interrupts.CPU115.TLB:TLB_shootdowns
>>>        3818 ±  3%     +23.2%       4703        interrupts.CPU116.CAL:Function_call_interrupts
>>>      630.67 ±  7%     +81.6%       1145 ±  6%  interrupts.CPU116.RES:Rescheduling_interrupts
>>>      439.67 ± 14%     +98.7%     873.50 ± 15%  interrupts.CPU116.TLB:TLB_shootdowns
>>>        4154 ±  4%     +13.8%       4728        interrupts.CPU117.CAL:Function_call_interrupts
>>>      650.67 ±  8%     +70.7%       1110 ±  6%  interrupts.CPU117.RES:Rescheduling_interrupts
>>>      502.33 ± 12%     +77.0%     889.00 ± 10%  interrupts.CPU117.TLB:TLB_shootdowns
>>>        3947 ±  5%     +21.6%       4798 ±  5%  interrupts.CPU118.CAL:Function_call_interrupts
>>>      659.33           +71.4%       1130 ±  4%  interrupts.CPU118.RES:Rescheduling_interrupts
>>>      464.67 ± 11%     +84.5%     857.25 ±  9%  interrupts.CPU118.TLB:TLB_shootdowns
>>>        4065 ±  4%     +19.6%       4860 ±  5%  interrupts.CPU119.CAL:Function_call_interrupts
>>>      614.00 ±  8%     +82.9%       1123 ±  4%  interrupts.CPU119.RES:Rescheduling_interrupts
>>>      480.33 ±  8%     +74.6%     838.50 ±  8%  interrupts.CPU119.TLB:TLB_shootdowns
>>>        4011 ±  3%     +20.9%       4848 ±  7%  interrupts.CPU12.CAL:Function_call_interrupts
>>>      769.33 ±  2%     +44.1%       1108 ±  8%  interrupts.CPU12.RES:Rescheduling_interrupts
>>>      502.00 ±  4%     +96.4%     986.00 ± 12%  interrupts.CPU12.TLB:TLB_shootdowns
>>>        4091 ±  4%     +22.3%       5003 ±  3%  interrupts.CPU120.CAL:Function_call_interrupts
>>>      717.33 ±  9%     +64.2%       1177 ±  6%  interrupts.CPU120.RES:Rescheduling_interrupts
>>>      459.67 ± 14%    +103.1%     933.50 ± 15%  interrupts.CPU120.TLB:TLB_shootdowns
>>>        3923 ±  4%     +25.2%       4913 ±  3%  interrupts.CPU121.CAL:Function_call_interrupts
>>>      617.33 ±  3%     +89.4%       1169 ±  4%  interrupts.CPU121.RES:Rescheduling_interrupts
>>>      401.67 ± 22%    +195.5%       1186 ± 39%  interrupts.CPU121.TLB:TLB_shootdowns
>>>        3867 ±  4%     +34.9%       5218 ±  4%  interrupts.CPU122.CAL:Function_call_interrupts
>>>      596.67 ±  8%    +100.9%       1198 ±  5%  interrupts.CPU122.RES:Rescheduling_interrupts
>>>      450.00 ± 16%    +109.0%     940.50 ±  6%  interrupts.CPU122.TLB:TLB_shootdowns
>>>        4031 ±  8%     +23.2%       4967 ±  3%  interrupts.CPU123.CAL:Function_call_interrupts
>>>      628.00 ±  4%     +94.1%       1218 ±  7%  interrupts.CPU123.RES:Rescheduling_interrupts
>>>      474.67 ± 13%     +96.1%     931.00 ± 12%  interrupts.CPU123.TLB:TLB_shootdowns
>>>        4232 ±  4%     +20.3%       5090 ±  3%  interrupts.CPU124.CAL:Function_call_interrupts
>>>      639.00           +82.9%       1168 ±  7%  interrupts.CPU124.RES:Rescheduling_interrupts
>>>      463.33 ± 17%    +110.1%     973.25 ± 11%  interrupts.CPU124.TLB:TLB_shootdowns
>>>        3972           +21.5%       4826 ±  3%  interrupts.CPU125.CAL:Function_call_interrupts
>>>      658.67 ±  8%     +78.4%       1175 ±  4%  interrupts.CPU125.RES:Rescheduling_interrupts
>>>      480.00 ± 13%     +96.0%     940.75 ± 23%  interrupts.CPU125.TLB:TLB_shootdowns
>>>        4093 ±  3%     +20.8%       4947 ±  4%  interrupts.CPU126.CAL:Function_call_interrupts
>>>      672.67 ±  5%     +75.9%       1183 ±  2%  interrupts.CPU126.RES:Rescheduling_interrupts
>>>      498.00 ± 10%     +90.0%     946.25 ± 11%  interrupts.CPU126.TLB:TLB_shootdowns
>>>        4111 ±  5%     +20.8%       4965 ±  5%  interrupts.CPU127.CAL:Function_call_interrupts
>>>      613.67 ±  2%     +95.2%       1198 ±  5%  interrupts.CPU127.RES:Rescheduling_interrupts
>>>      451.67 ± 21%    +109.8%     947.50 ± 14%  interrupts.CPU127.TLB:TLB_shootdowns
>>>        4066           +20.3%       4891 ±  3%  interrupts.CPU128.CAL:Function_call_interrupts
>>>        5200 ±  2%     -23.3%       3986 ±  9%  interrupts.CPU128.NMI:Non-maskable_interrupts
>>>        5200 ±  2%     -23.3%       3986 ±  9%  interrupts.CPU128.PMI:Performance_monitoring_interrupts
>>>      663.33 ±  4%     +74.6%       1158 ±  6%  interrupts.CPU128.RES:Rescheduling_interrupts
>>>      472.33 ± 17%    +102.4%     956.00 ±  8%  interrupts.CPU128.TLB:TLB_shootdowns
>>>        4075 ±  8%     +20.3%       4900        interrupts.CPU129.CAL:Function_call_interrupts
>>>      640.33 ±  2%     +73.2%       1108 ±  8%  interrupts.CPU129.RES:Rescheduling_interrupts
>>>      445.33 ± 19%    +113.4%     950.25 ± 11%  interrupts.CPU129.TLB:TLB_shootdowns
>>>        8340 ± 68%     -76.0%       2003 ± 74%  interrupts.CPU13.34:PCI-MSI.524292-edge.eth0-TxRx-3
>>>        4049 ±  7%     +20.6%       4885 ±  2%  interrupts.CPU13.CAL:Function_call_interrupts
>>>        5130 ±  3%     -24.2%       3886 ±  5%  interrupts.CPU13.NMI:Non-maskable_interrupts
>>>        5130 ±  3%     -24.2%       3886 ±  5%  interrupts.CPU13.PMI:Performance_monitoring_interrupts
>>>      724.00 ±  5%     +57.8%       1142 ±  5%  interrupts.CPU13.RES:Rescheduling_interrupts
>>>      532.67 ± 23%     +70.5%     908.25 ± 12%  interrupts.CPU13.TLB:TLB_shootdowns
>>>        3881 ±  6%     +30.0%       5044 ±  4%  interrupts.CPU130.CAL:Function_call_interrupts
>>>      621.33 ± 10%     +89.4%       1176 ±  8%  interrupts.CPU130.RES:Rescheduling_interrupts
>>>      516.33 ± 16%     +92.9%     996.25 ± 15%  interrupts.CPU130.TLB:TLB_shootdowns
>>>        4035 ±  2%     +22.0%       4923 ±  3%  interrupts.CPU131.CAL:Function_call_interrupts
>>>      661.67 ±  7%     +69.5%       1121 ±  4%  interrupts.CPU131.RES:Rescheduling_interrupts
>>>      425.67 ± 11%    +116.8%     922.75 ± 15%  interrupts.CPU131.TLB:TLB_shootdowns
>>>        4345 ±  3%     +13.3%       4924 ±  5%  interrupts.CPU132.CAL:Function_call_interrupts
>>>      694.00           +68.9%       1172 ±  5%  interrupts.CPU132.RES:Rescheduling_interrupts
>>>      476.00 ± 11%     +94.5%     926.00 ± 13%  interrupts.CPU132.TLB:TLB_shootdowns
>>>        4091 ±  7%     +22.7%       5018 ±  4%  interrupts.CPU133.CAL:Function_call_interrupts
>>>      663.67 ±  6%     +68.6%       1119 ±  4%  interrupts.CPU133.RES:Rescheduling_interrupts
>>>      534.33 ±  9%     +78.7%     954.75 ± 13%  interrupts.CPU133.TLB:TLB_shootdowns
>>>        3856 ±  7%     +26.7%       4886 ±  3%  interrupts.CPU134.CAL:Function_call_interrupts
>>>        5199 ±  2%     -34.5%       3406 ± 24%  interrupts.CPU134.NMI:Non-maskable_interrupts
>>>        5199 ±  2%     -34.5%       3406 ± 24%  interrupts.CPU134.PMI:Performance_monitoring_interrupts
>>>      667.00 ± 13%     +74.7%       1165 ±  6%  interrupts.CPU134.RES:Rescheduling_interrupts
>>>      447.33 ±  9%    +116.7%     969.25 ±  3%  interrupts.CPU134.TLB:TLB_shootdowns
>>>        4120 ± 11%     +18.6%       4886 ±  3%  interrupts.CPU135.CAL:Function_call_interrupts
>>>        5233 ±  3%     -24.8%       3934 ±  9%  interrupts.CPU135.NMI:Non-maskable_interrupts
>>>        5233 ±  3%     -24.8%       3934 ±  9%  interrupts.CPU135.PMI:Performance_monitoring_interrupts
>>>      643.67 ±  5%     +81.8%       1170 ±  7%  interrupts.CPU135.RES:Rescheduling_interrupts
>>>      490.00 ± 17%     +88.4%     923.00 ± 18%  interrupts.CPU135.TLB:TLB_shootdowns
>>>        4092 ±  3%     +22.8%       5026 ±  3%  interrupts.CPU136.CAL:Function_call_interrupts
>>>        5225 ±  3%     -34.3%       3430 ± 26%  interrupts.CPU136.NMI:Non-maskable_interrupts
>>>        5225 ±  3%     -34.3%       3430 ± 26%  interrupts.CPU136.PMI:Performance_monitoring_interrupts
>>>      673.33 ±  4%     +80.6%       1216 ±  4%  interrupts.CPU136.RES:Rescheduling_interrupts
>>>      488.00 ±  8%     +87.9%     916.75 ± 13%  interrupts.CPU136.TLB:TLB_shootdowns
>>>        4234 ±  9%     +16.3%       4925 ±  4%  interrupts.CPU137.CAL:Function_call_interrupts
>>>        5217 ±  4%     -34.5%       3418 ± 24%  interrupts.CPU137.NMI:Non-maskable_interrupts
>>>        5217 ±  4%     -34.5%       3418 ± 24%  interrupts.CPU137.PMI:Performance_monitoring_interrupts
>>>      625.33 ± 14%     +87.5%       1172 ±  6%  interrupts.CPU137.RES:Rescheduling_interrupts
>>>      441.67 ± 10%    +115.5%     951.75 ± 13%  interrupts.CPU137.TLB:TLB_shootdowns
>>>        3987 ±  6%     +23.1%       4907 ±  2%  interrupts.CPU138.CAL:Function_call_interrupts
>>>        5218 ±  3%     -42.7%       2992 ± 38%  interrupts.CPU138.NMI:Non-maskable_interrupts
>>>        5218 ±  3%     -42.7%       2992 ± 38%  interrupts.CPU138.PMI:Performance_monitoring_interrupts
>>>      672.33 ±  3%     +76.5%       1186 ±  4%  interrupts.CPU138.RES:Rescheduling_interrupts
>>>      439.67 ±  8%    +113.4%     938.25 ± 11%  interrupts.CPU138.TLB:TLB_shootdowns
>>>        4040 ±  7%     +23.1%       4973 ±  4%  interrupts.CPU139.CAL:Function_call_interrupts
>>>      635.33 ±  3%     +85.3%       1177 ±  7%  interrupts.CPU139.RES:Rescheduling_interrupts
>>>      425.33 ± 17%    +110.7%     896.00 ± 10%  interrupts.CPU139.TLB:TLB_shootdowns
>>>        3920 ±  2%     +21.7%       4771 ±  2%  interrupts.CPU14.CAL:Function_call_interrupts
>>>        5123 ±  4%     -43.3%       2903 ± 30%  interrupts.CPU14.NMI:Non-maskable_interrupts
>>>        5123 ±  4%     -43.3%       2903 ± 30%  interrupts.CPU14.PMI:Performance_monitoring_interrupts
>>>      725.00 ± 12%     +56.2%       1132 ±  6%  interrupts.CPU14.RES:Rescheduling_interrupts
>>>      402.00 ±  6%    +110.0%     844.25 ± 10%  interrupts.CPU14.TLB:TLB_shootdowns
>>>        4006 ± 11%     +19.8%       4800 ±  3%  interrupts.CPU140.CAL:Function_call_interrupts
>>>        5230 ±  3%     -24.8%       3932 ± 10%  interrupts.CPU140.NMI:Non-maskable_interrupts
>>>        5230 ±  3%     -24.8%       3932 ± 10%  interrupts.CPU140.PMI:Performance_monitoring_interrupts
>>>      578.33 ±  3%    +102.5%       1171 ±  5%  interrupts.CPU140.RES:Rescheduling_interrupts
>>>      506.00 ±  5%     +90.3%     963.00 ± 11%  interrupts.CPU140.TLB:TLB_shootdowns
>>>        4080 ±  6%     +25.8%       5131 ±  7%  interrupts.CPU141.CAL:Function_call_interrupts
>>>        5148 ±  4%     -33.6%       3418 ± 25%  interrupts.CPU141.NMI:Non-maskable_interrupts
>>>        5148 ±  4%     -33.6%       3418 ± 25%  interrupts.CPU141.PMI:Performance_monitoring_interrupts
>>>      695.00 ±  3%     +74.7%       1214 ±  7%  interrupts.CPU141.RES:Rescheduling_interrupts
>>>      503.33 ± 23%    +150.9%       1263 ± 44%  interrupts.CPU141.TLB:TLB_shootdowns
>>>        4155 ±  8%     +16.6%       4843 ±  4%  interrupts.CPU142.CAL:Function_call_interrupts
>>>      638.00 ±  6%     +82.1%       1161 ±  6%  interrupts.CPU142.RES:Rescheduling_interrupts
>>>      509.33 ±  8%     +89.0%     962.50 ± 14%  interrupts.CPU142.TLB:TLB_shootdowns
>>>        4018 ±  6%     +24.3%       4993        interrupts.CPU143.CAL:Function_call_interrupts
>>>      643.67 ±  2%     +86.6%       1201 ±  8%  interrupts.CPU143.RES:Rescheduling_interrupts
>>>      553.67 ±  7%     +83.0%       1013 ± 10%  interrupts.CPU143.TLB:TLB_shootdowns
>>>        3990 ±  5%     +25.0%       4989 ±  4%  interrupts.CPU144.CAL:Function_call_interrupts
>>>      654.00 ± 10%     +78.0%       1164 ±  4%  interrupts.CPU144.RES:Rescheduling_interrupts
>>>      415.33 ±  4%    +124.9%     934.25 ±  9%  interrupts.CPU144.TLB:TLB_shootdowns
>>>        3875 ±  3%     +30.5%       5056 ±  6%  interrupts.CPU145.CAL:Function_call_interrupts
>>>      647.00 ± 13%     +79.9%       1163 ±  4%  interrupts.CPU145.RES:Rescheduling_interrupts
>>>      445.67 ± 16%    +112.7%     947.75 ± 12%  interrupts.CPU145.TLB:TLB_shootdowns
>>>        3891 ±  7%     +24.8%       4857 ±  5%  interrupts.CPU146.CAL:Function_call_interrupts
>>>      635.00 ±  6%     +80.7%       1147 ±  5%  interrupts.CPU146.RES:Rescheduling_interrupts
>>>      438.67 ± 11%    +127.1%     996.25 ± 10%  interrupts.CPU146.TLB:TLB_shootdowns
>>>        3899 ±  7%     +28.0%       4990 ±  4%  interrupts.CPU147.CAL:Function_call_interrupts
>>>        5189 ±  3%     -24.6%       3915 ± 10%  interrupts.CPU147.NMI:Non-maskable_interrupts
>>>        5189 ±  3%     -24.6%       3915 ± 10%  interrupts.CPU147.PMI:Performance_monitoring_interrupts
>>>      643.33           +80.3%       1160 ±  2%  interrupts.CPU147.RES:Rescheduling_interrupts
>>>      430.00 ± 16%    +116.9%     932.50 ± 17%  interrupts.CPU147.TLB:TLB_shootdowns
>>>        4129 ± 12%     +18.4%       4889        interrupts.CPU148.CAL:Function_call_interrupts
>>>        5219 ±  3%     -24.7%       3932 ±  9%  interrupts.CPU148.NMI:Non-maskable_interrupts
>>>        5219 ±  3%     -24.7%       3932 ±  9%  interrupts.CPU148.PMI:Performance_monitoring_interrupts
>>>      588.33 ±  7%     +93.6%       1139 ±  5%  interrupts.CPU148.RES:Rescheduling_interrupts
>>>      501.00 ± 16%     +78.7%     895.25 ±  8%  interrupts.CPU148.TLB:TLB_shootdowns
>>>        4186 ±  9%     +18.3%       4953 ±  3%  interrupts.CPU149.CAL:Function_call_interrupts
>>>        5185 ±  4%     -32.3%       3509 ± 31%  interrupts.CPU149.NMI:Non-maskable_interrupts
>>>        5185 ±  4%     -32.3%       3509 ± 31%  interrupts.CPU149.PMI:Performance_monitoring_interrupts
>>>      638.67           +70.9%       1091 ±  2%  interrupts.CPU149.RES:Rescheduling_interrupts
>>>      558.33 ± 15%     +71.8%     959.25 ±  7%  interrupts.CPU149.TLB:TLB_shootdowns
>>>        3912 ±  6%     +21.4%       4750 ±  4%  interrupts.CPU15.CAL:Function_call_interrupts
>>>        5197 ±  3%     -22.8%       4013 ±  4%  interrupts.CPU15.NMI:Non-maskable_interrupts
>>>        5197 ±  3%     -22.8%       4013 ±  4%  interrupts.CPU15.PMI:Performance_monitoring_interrupts
>>>      776.33 ±  7%     +39.1%       1080 ±  6%  interrupts.CPU15.RES:Rescheduling_interrupts
>>>      473.00 ± 12%     +98.5%     938.75 ± 11%  interrupts.CPU15.TLB:TLB_shootdowns
>>>        4178 ±  9%     +18.1%       4936 ±  4%  interrupts.CPU150.CAL:Function_call_interrupts
>>>        5192 ±  4%     -25.4%       3873 ± 10%  interrupts.CPU150.NMI:Non-maskable_interrupts
>>>        5192 ±  4%     -25.4%       3873 ± 10%  interrupts.CPU150.PMI:Performance_monitoring_interrupts
>>>      678.67 ±  8%     +64.3%       1115 ±  5%  interrupts.CPU150.RES:Rescheduling_interrupts
>>>      513.33 ± 28%     +73.6%     891.00 ± 21%  interrupts.CPU150.TLB:TLB_shootdowns
>>>        4052 ±  6%     +21.9%       4939 ±  3%  interrupts.CPU151.CAL:Function_call_interrupts
>>>        5260 ±  3%     -25.8%       3903 ± 10%  interrupts.CPU151.NMI:Non-maskable_interrupts
>>>        5260 ±  3%     -25.8%       3903 ± 10%  interrupts.CPU151.PMI:Performance_monitoring_interrupts
>>>      623.33 ±  3%     +86.9%       1164 ±  4%  interrupts.CPU151.RES:Rescheduling_interrupts
>>>      484.67 ± 11%     +98.0%     959.50 ±  8%  interrupts.CPU151.TLB:TLB_shootdowns
>>>        3982 ±  8%     +25.1%       4982 ±  4%  interrupts.CPU152.CAL:Function_call_interrupts
>>>        5183 ±  2%     -32.9%       3475 ± 26%  interrupts.CPU152.NMI:Non-maskable_interrupts
>>>        5183 ±  2%     -32.9%       3475 ± 26%  interrupts.CPU152.PMI:Performance_monitoring_interrupts
>>>      623.33 ±  5%     +79.5%       1119 ±  2%  interrupts.CPU152.RES:Rescheduling_interrupts
>>>      494.00 ± 14%     +89.7%     937.25 ±  9%  interrupts.CPU152.TLB:TLB_shootdowns
>>>        4109 ±  6%     +24.9%       5131 ± 10%  interrupts.CPU153.CAL:Function_call_interrupts
>>>        5179 ±  4%     -31.7%       3536 ± 31%  interrupts.CPU153.NMI:Non-maskable_interrupts
>>>        5179 ±  4%     -31.7%       3536 ± 31%  interrupts.CPU153.PMI:Performance_monitoring_interrupts
>>>      614.67 ±  3%     +94.8%       1197 ±  9%  interrupts.CPU153.RES:Rescheduling_interrupts
>>>      486.00 ± 19%     +83.9%     893.75 ± 13%  interrupts.CPU153.TLB:TLB_shootdowns
>>>        4161 ±  5%     +16.5%       4846 ±  2%  interrupts.CPU154.CAL:Function_call_interrupts
>>>      635.33 ±  3%     +74.6%       1109 ±  3%  interrupts.CPU154.RES:Rescheduling_interrupts
>>>      490.67 ± 21%     +87.4%     919.50 ±  8%  interrupts.CPU154.TLB:TLB_shootdowns
>>>        3946 ±  8%     +28.2%       5057 ±  2%  interrupts.CPU155.CAL:Function_call_interrupts
>>>      631.33 ±  9%     +78.6%       1127 ±  3%  interrupts.CPU155.RES:Rescheduling_interrupts
>>>      489.33 ± 17%    +107.9%       1017 ± 11%  interrupts.CPU155.TLB:TLB_shootdowns
>>>        4028 ±  5%     +21.3%       4887 ±  4%  interrupts.CPU156.CAL:Function_call_interrupts
>>>      627.33 ±  2%     +82.2%       1142 ±  2%  interrupts.CPU156.RES:Rescheduling_interrupts
>>>      452.33 ± 14%    +113.0%     963.25 ± 15%  interrupts.CPU156.TLB:TLB_shootdowns
>>>        3893 ±  8%     +28.2%       4993 ±  4%  interrupts.CPU157.CAL:Function_call_interrupts
>>>      637.33           +80.4%       1149 ±  4%  interrupts.CPU157.RES:Rescheduling_interrupts
>>>      451.67 ± 20%    +105.6%     928.50 ± 11%  interrupts.CPU157.TLB:TLB_shootdowns
>>>        4155 ±  9%     +19.6%       4971 ±  4%  interrupts.CPU158.CAL:Function_call_interrupts
>>>      628.33 ±  6%     +75.8%       1104 ±  2%  interrupts.CPU158.RES:Rescheduling_interrupts
>>>      468.33 ± 11%    +103.4%     952.50 ± 11%  interrupts.CPU158.TLB:TLB_shootdowns
>>>        3965 ±  8%     +23.8%       4910 ±  5%  interrupts.CPU159.CAL:Function_call_interrupts
>>>      597.33 ±  4%     +95.2%       1166 ±  2%  interrupts.CPU159.RES:Rescheduling_interrupts
>>>      436.00 ± 12%    +121.7%     966.50 ± 11%  interrupts.CPU159.TLB:TLB_shootdowns
>>>        3811 ±  3%     +28.3%       4890 ±  6%  interrupts.CPU16.CAL:Function_call_interrupts
>>>        5211 ±  3%     -24.1%       3953 ±  5%  interrupts.CPU16.NMI:Non-maskable_interrupts
>>>        5211 ±  3%     -24.1%       3953 ±  5%  interrupts.CPU16.PMI:Performance_monitoring_interrupts
>>>      726.67 ±  6%     +55.9%       1132 ±  7%  interrupts.CPU16.RES:Rescheduling_interrupts
>>>      505.67 ± 13%     +75.1%     885.50 ± 12%  interrupts.CPU16.TLB:TLB_shootdowns
>>>        3965 ±  8%     +24.8%       4948 ±  6%  interrupts.CPU160.CAL:Function_call_interrupts
>>>      636.67 ±  4%     +74.8%       1113 ±  2%  interrupts.CPU160.RES:Rescheduling_interrupts
>>>      516.67 ± 17%     +88.3%     973.00 ± 14%  interrupts.CPU160.TLB:TLB_shootdowns
>>>        4054 ±  8%     +24.2%       5036 ±  4%  interrupts.CPU161.CAL:Function_call_interrupts
>>>      601.00 ±  6%     +89.6%       1139 ±  3%  interrupts.CPU161.RES:Rescheduling_interrupts
>>>      514.33 ± 13%     +81.9%     935.50 ± 12%  interrupts.CPU161.TLB:TLB_shootdowns
>>>        4039 ± 11%     +19.7%       4836 ±  3%  interrupts.CPU162.CAL:Function_call_interrupts
>>>      643.33 ±  7%     +67.2%       1075        interrupts.CPU162.RES:Rescheduling_interrupts
>>>      469.33 ± 18%    +107.1%     972.00 ±  8%  interrupts.CPU162.TLB:TLB_shootdowns
>>>        4022 ±  7%     +20.9%       4864 ±  3%  interrupts.CPU163.CAL:Function_call_interrupts
>>>      609.33 ±  3%     +92.5%       1172        interrupts.CPU163.RES:Rescheduling_interrupts
>>>      431.00 ± 18%    +146.3%       1061 ±  6%  interrupts.CPU163.TLB:TLB_shootdowns
>>>        4036           +20.7%       4869 ±  4%  interrupts.CPU164.CAL:Function_call_interrupts
>>>      613.00 ±  4%     +86.4%       1142 ±  6%  interrupts.CPU164.RES:Rescheduling_interrupts
>>>      492.33 ± 19%     +87.9%     925.25 ± 10%  interrupts.CPU164.TLB:TLB_shootdowns
>>>        3957 ±  3%     +23.4%       4883 ±  4%  interrupts.CPU165.CAL:Function_call_interrupts
>>>        4307 ± 27%     -42.1%       2492 ± 41%  interrupts.CPU165.NMI:Non-maskable_interrupts
>>>        4307 ± 27%     -42.1%       2492 ± 41%  interrupts.CPU165.PMI:Performance_monitoring_interrupts
>>>      583.00 ± 10%     +88.9%       1101 ±  3%  interrupts.CPU165.RES:Rescheduling_interrupts
>>>      449.00 ± 22%    +108.7%     937.25 ± 13%  interrupts.CPU165.TLB:TLB_shootdowns
>>>        4219 ±  5%     +16.9%       4932 ±  4%  interrupts.CPU166.CAL:Function_call_interrupts
>>>      624.33 ±  4%     +76.9%       1104 ±  5%  interrupts.CPU166.RES:Rescheduling_interrupts
>>>      484.33 ± 17%    +100.6%     971.50 ± 17%  interrupts.CPU166.TLB:TLB_shootdowns
>>>        4064 ± 10%     +21.0%       4918 ±  3%  interrupts.CPU167.CAL:Function_call_interrupts
>>>      616.67 ±  6%     +83.5%       1131 ±  4%  interrupts.CPU167.RES:Rescheduling_interrupts
>>>      452.00 ± 12%     +95.7%     884.50 ±  8%  interrupts.CPU167.TLB:TLB_shootdowns
>>>        4087 ±  6%     +19.4%       4879 ±  6%  interrupts.CPU168.CAL:Function_call_interrupts
>>>      649.67           +69.1%       1098 ±  4%  interrupts.CPU168.RES:Rescheduling_interrupts
>>>      478.33 ±  6%     +92.4%     920.50 ± 12%  interrupts.CPU168.TLB:TLB_shootdowns
>>>        3753 ±  7%     +20.6%       4524 ±  4%  interrupts.CPU169.CAL:Function_call_interrupts
>>>      585.67           +81.9%       1065 ±  2%  interrupts.CPU169.RES:Rescheduling_interrupts
>>>      459.00 ±  4%    +104.2%     937.50 ±  3%  interrupts.CPU169.TLB:TLB_shootdowns
>>>        3764 ±  5%     +28.6%       4841 ±  3%  interrupts.CPU17.CAL:Function_call_interrupts
>>>        5198 ±  3%     -43.0%       2964 ± 35%  interrupts.CPU17.NMI:Non-maskable_interrupts
>>>        5198 ±  3%     -43.0%       2964 ± 35%  interrupts.CPU17.PMI:Performance_monitoring_interrupts
>>>      729.33 ±  5%     +52.1%       1109 ±  7%  interrupts.CPU17.RES:Rescheduling_interrupts
>>>      450.67 ± 10%    +104.4%     921.25 ±  7%  interrupts.CPU17.TLB:TLB_shootdowns
>>>        4099 ±  8%     +24.8%       5118 ±  5%  interrupts.CPU170.CAL:Function_call_interrupts
>>>      590.67 ±  7%     +91.4%       1130        interrupts.CPU170.RES:Rescheduling_interrupts
>>>      552.33 ± 17%     +75.9%     971.75 ± 10%  interrupts.CPU170.TLB:TLB_shootdowns
>>>        4023 ±  8%     +19.5%       4808        interrupts.CPU171.CAL:Function_call_interrupts
>>>      594.67 ±  7%     +82.6%       1085 ±  5%  interrupts.CPU171.RES:Rescheduling_interrupts
>>>      468.33 ± 13%     +94.0%     908.75 ±  7%  interrupts.CPU171.TLB:TLB_shootdowns
>>>        3917 ±  9%     +25.0%       4895 ±  5%  interrupts.CPU172.CAL:Function_call_interrupts
>>>        5248 ±  3%     -25.4%       3915 ± 10%  interrupts.CPU172.NMI:Non-maskable_interrupts
>>>        5248 ±  3%     -25.4%       3915 ± 10%  interrupts.CPU172.PMI:Performance_monitoring_interrupts
>>>      543.00 ±  3%    +103.6%       1105 ±  5%  interrupts.CPU172.RES:Rescheduling_interrupts
>>>      468.67 ± 15%     +94.0%     909.25 ± 17%  interrupts.CPU172.TLB:TLB_shootdowns
>>>        4057 ±  6%     +22.6%       4975 ±  2%  interrupts.CPU173.CAL:Function_call_interrupts
>>>        5158 ±  5%     -31.7%       3521 ± 31%  interrupts.CPU173.NMI:Non-maskable_interrupts
>>>        5158 ±  5%     -31.7%       3521 ± 31%  interrupts.CPU173.PMI:Performance_monitoring_interrupts
>>>      597.67 ±  5%     +87.9%       1123 ±  3%  interrupts.CPU173.RES:Rescheduling_interrupts
>>>      498.67 ± 19%    +133.2%       1163 ± 26%  interrupts.CPU173.TLB:TLB_shootdowns
>>>        3833 ±  8%     +26.8%       4859 ±  3%  interrupts.CPU174.CAL:Function_call_interrupts
>>>        5208 ±  4%     -24.7%       3920 ± 11%  interrupts.CPU174.NMI:Non-maskable_interrupts
>>>        5208 ±  4%     -24.7%       3920 ± 11%  interrupts.CPU174.PMI:Performance_monitoring_interrupts
>>>      570.67 ±  2%     +91.1%       1090 ±  7%  interrupts.CPU174.RES:Rescheduling_interrupts
>>>      470.00 ± 19%    +108.3%     979.00 ± 12%  interrupts.CPU174.TLB:TLB_shootdowns
>>>        4117 ±  7%     +17.7%       4845 ±  3%  interrupts.CPU175.CAL:Function_call_interrupts
>>>        5202 ±  4%     -25.1%       3896 ±  8%  interrupts.CPU175.NMI:Non-maskable_interrupts
>>>        5202 ±  4%     -25.1%       3896 ±  8%  interrupts.CPU175.PMI:Performance_monitoring_interrupts
>>>      638.33 ±  3%     +70.8%       1090 ±  5%  interrupts.CPU175.RES:Rescheduling_interrupts
>>>      463.00 ± 17%    +104.4%     946.50 ±  8%  interrupts.CPU175.TLB:TLB_shootdowns
>>>        4021 ±  4%     +16.9%       4699 ±  2%  interrupts.CPU176.CAL:Function_call_interrupts
>>>      608.67 ±  6%     +79.4%       1091 ±  6%  interrupts.CPU176.RES:Rescheduling_interrupts
>>>      477.33 ± 10%     +97.2%     941.50 ±  7%  interrupts.CPU176.TLB:TLB_shootdowns
>>>        4071 ±  7%     +15.5%       4702        interrupts.CPU177.CAL:Function_call_interrupts
>>>      596.33 ±  4%     +82.2%       1086        interrupts.CPU177.RES:Rescheduling_interrupts
>>>      435.00 ±  5%    +114.7%     934.00 ± 14%  interrupts.CPU177.TLB:TLB_shootdowns
>>>        4130 ±  7%     +16.6%       4815 ±  4%  interrupts.CPU178.CAL:Function_call_interrupts
>>>      633.33 ±  4%     +64.1%       1039 ±  4%  interrupts.CPU178.RES:Rescheduling_interrupts
>>>      478.67 ± 15%     +95.2%     934.50 ± 19%  interrupts.CPU178.TLB:TLB_shootdowns
>>>        3821 ±  4%     +26.8%       4846 ±  3%  interrupts.CPU179.CAL:Function_call_interrupts
>>>      594.67           +76.6%       1050 ±  4%  interrupts.CPU179.RES:Rescheduling_interrupts
>>>      479.00 ±  7%     +96.5%     941.00 ±  6%  interrupts.CPU179.TLB:TLB_shootdowns
>>>        4196 ±  4%     +16.5%       4890 ±  5%  interrupts.CPU18.CAL:Function_call_interrupts
>>>      728.33           +51.1%       1100 ±  3%  interrupts.CPU18.RES:Rescheduling_interrupts
>>>      506.00 ± 11%     +71.7%     868.75 ±  6%  interrupts.CPU18.TLB:TLB_shootdowns
>>>        3950 ±  6%     +23.3%       4870 ±  2%  interrupts.CPU180.CAL:Function_call_interrupts
>>>      606.33 ±  8%     +72.5%       1045 ±  2%  interrupts.CPU180.RES:Rescheduling_interrupts
>>>      498.00 ±  5%     +72.3%     858.00 ± 17%  interrupts.CPU180.TLB:TLB_shootdowns
>>>        4113 ± 11%     +18.3%       4864 ±  5%  interrupts.CPU181.CAL:Function_call_interrupts
>>>      620.33 ±  4%     +73.7%       1077 ±  5%  interrupts.CPU181.RES:Rescheduling_interrupts
>>>      483.33 ±  5%     +90.3%     919.75 ± 11%  interrupts.CPU181.TLB:TLB_shootdowns
>>>        4070 ±  8%     +17.0%       4761 ±  2%  interrupts.CPU182.CAL:Function_call_interrupts
>>>      588.00 ±  7%     +82.8%       1075 ±  3%  interrupts.CPU182.RES:Rescheduling_interrupts
>>>      480.00 ± 17%    +124.7%       1078 ± 10%  interrupts.CPU182.TLB:TLB_shootdowns
>>>        4198 ±  7%     +15.9%       4868 ±  2%  interrupts.CPU183.CAL:Function_call_interrupts
>>>      681.33 ±  6%     +59.5%       1087 ±  6%  interrupts.CPU183.RES:Rescheduling_interrupts
>>>      509.33 ± 16%     +78.5%     909.00 ± 15%  interrupts.CPU183.TLB:TLB_shootdowns
>>>        3955 ±  2%     +25.7%       4970 ±  3%  interrupts.CPU184.CAL:Function_call_interrupts
>>>      565.00 ±  6%     +93.8%       1095 ±  6%  interrupts.CPU184.RES:Rescheduling_interrupts
>>>      443.67 ± 14%    +107.0%     918.25 ±  9%  interrupts.CPU184.TLB:TLB_shootdowns
>>>        4105 ±  6%     +16.5%       4781 ±  5%  interrupts.CPU185.CAL:Function_call_interrupts
>>>      546.67 ±  5%    +105.8%       1125 ±  7%  interrupts.CPU185.RES:Rescheduling_interrupts
>>>      505.00 ± 18%     +86.8%     943.50 ±  8%  interrupts.CPU185.TLB:TLB_shootdowns
>>>        4067 ±  6%     +19.9%       4878 ±  3%  interrupts.CPU186.CAL:Function_call_interrupts
>>>      604.67 ±  6%     +71.2%       1035 ±  7%  interrupts.CPU186.RES:Rescheduling_interrupts
>>>      486.67 ± 12%     +89.7%     923.25 ±  8%  interrupts.CPU186.TLB:TLB_shootdowns
>>>      588.00 ±  6%     +77.9%       1046 ±  5%  interrupts.CPU187.RES:Rescheduling_interrupts
>>>      499.67 ± 15%     +86.2%     930.50 ± 15%  interrupts.CPU187.TLB:TLB_shootdowns
>>>        4186 ±  5%     +17.1%       4901 ±  5%  interrupts.CPU188.CAL:Function_call_interrupts
>>>      616.33 ±  2%     +77.1%       1091 ±  4%  interrupts.CPU188.RES:Rescheduling_interrupts
>>>      461.33 ±  9%    +102.1%     932.25 ± 15%  interrupts.CPU188.TLB:TLB_shootdowns
>>>        3980 ±  6%     +21.8%       4849 ±  2%  interrupts.CPU189.CAL:Function_call_interrupts
>>>      594.33           +87.8%       1116 ±  3%  interrupts.CPU189.RES:Rescheduling_interrupts
>>>      431.33 ± 10%    +126.0%     974.75 ±  5%  interrupts.CPU189.TLB:TLB_shootdowns
>>>        3937 ±  5%     +21.2%       4772 ±  5%  interrupts.CPU19.CAL:Function_call_interrupts
>>>        5173 ±  3%     -31.9%       3525 ± 25%  interrupts.CPU19.NMI:Non-maskable_interrupts
>>>        5173 ±  3%     -31.9%       3525 ± 25%  interrupts.CPU19.PMI:Performance_monitoring_interrupts
>>>      722.67 ±  9%     +48.1%       1070 ±  2%  interrupts.CPU19.RES:Rescheduling_interrupts
>>>      479.67 ± 18%     +95.3%     937.00 ±  9%  interrupts.CPU19.TLB:TLB_shootdowns
>>>        3905 ±  5%     +25.2%       4890 ±  5%  interrupts.CPU190.CAL:Function_call_interrupts
>>>        5230 ±  2%     -24.7%       3937 ± 10%  interrupts.CPU190.NMI:Non-maskable_interrupts
>>>        5230 ±  2%     -24.7%       3937 ± 10%  interrupts.CPU190.PMI:Performance_monitoring_interrupts
>>>      580.33 ±  3%     +84.6%       1071 ±  2%  interrupts.CPU190.RES:Rescheduling_interrupts
>>>      454.33 ± 12%    +111.7%     961.75 ± 14%  interrupts.CPU190.TLB:TLB_shootdowns
>>>        3985 ±  4%     +23.4%       4918 ±  4%  interrupts.CPU191.CAL:Function_call_interrupts
>>>        5184 ±  4%     -34.0%       3422 ± 24%  interrupts.CPU191.NMI:Non-maskable_interrupts
>>>        5184 ±  4%     -34.0%       3422 ± 24%  interrupts.CPU191.PMI:Performance_monitoring_interrupts
>>>      589.67 ±  2%     +91.5%       1129 ±  7%  interrupts.CPU191.RES:Rescheduling_interrupts
>>>      498.67 ± 25%    +102.2%       1008 ± 14%  interrupts.CPU191.TLB:TLB_shootdowns
>>>        3847 ±  6%     +29.1%       4965 ±  8%  interrupts.CPU2.CAL:Function_call_interrupts
>>>        5162 ±  3%     -23.5%       3948 ±  5%  interrupts.CPU2.NMI:Non-maskable_interrupts
>>>        5162 ±  3%     -23.5%       3948 ±  5%  interrupts.CPU2.PMI:Performance_monitoring_interrupts
>>>      691.33 ±  3%     +59.0%       1099 ±  7%  interrupts.CPU2.RES:Rescheduling_interrupts
>>>      486.67 ±  6%     +83.7%     894.25 ± 13%  interrupts.CPU2.TLB:TLB_shootdowns
>>>        4000 ±  3%     +19.6%       4786 ±  5%  interrupts.CPU20.CAL:Function_call_interrupts
>>>        5181 ±  2%     -42.1%       2999 ± 32%  interrupts.CPU20.NMI:Non-maskable_interrupts
>>>        5181 ±  2%     -42.1%       2999 ± 32%  interrupts.CPU20.PMI:Performance_monitoring_interrupts
>>>      774.00 ± 11%     +45.8%       1128 ±  5%  interrupts.CPU20.RES:Rescheduling_interrupts
>>>      458.33 ± 11%    +111.7%     970.25 ± 12%  interrupts.CPU20.TLB:TLB_shootdowns
>>>        4143 ±  4%     +13.1%       4684 ±  3%  interrupts.CPU21.CAL:Function_call_interrupts
>>>      814.67 ± 19%     +35.1%       1100 ±  6%  interrupts.CPU21.RES:Rescheduling_interrupts
>>>      453.67 ± 12%     +97.1%     894.25 ± 13%  interrupts.CPU21.TLB:TLB_shootdowns
>>>        3936           +26.2%       4968 ±  4%  interrupts.CPU22.CAL:Function_call_interrupts
>>>        5194 ±  2%     -22.7%       4017 ±  5%  interrupts.CPU22.NMI:Non-maskable_interrupts
>>>        5194 ±  2%     -22.7%       4017 ±  5%  interrupts.CPU22.PMI:Performance_monitoring_interrupts
>>>      754.67 ± 11%     +45.8%       1100 ±  7%  interrupts.CPU22.RES:Rescheduling_interrupts
>>>      467.67 ± 14%    +100.0%     935.25 ± 17%  interrupts.CPU22.TLB:TLB_shootdowns
>>>        4046 ±  4%     +17.6%       4758 ±  2%  interrupts.CPU23.CAL:Function_call_interrupts
>>>      716.67 ±  2%     +45.6%       1043 ±  5%  interrupts.CPU23.RES:Rescheduling_interrupts
>>>      416.33 ±  7%    +108.4%     867.50 ± 12%  interrupts.CPU23.TLB:TLB_shootdowns
>>>        4323 ±  6%     +17.2%       5066 ±  5%  interrupts.CPU24.CAL:Function_call_interrupts
>>>        5170 ±  3%     -22.1%       4027 ±  6%  interrupts.CPU24.NMI:Non-maskable_interrupts
>>>        5170 ±  3%     -22.1%       4027 ±  6%  interrupts.CPU24.PMI:Performance_monitoring_interrupts
>>>      876.33 ±  4%     +33.1%       1166 ±  6%  interrupts.CPU24.RES:Rescheduling_interrupts
>>>      678.33 ± 16%     +50.4%       1020 ±  9%  interrupts.CPU24.TLB:TLB_shootdowns
>>>        4259 ±  7%     +18.1%       5027 ±  2%  interrupts.CPU25.CAL:Function_call_interrupts
>>>        5240 ±  3%     -23.8%       3993 ±  8%  interrupts.CPU25.NMI:Non-maskable_interrupts
>>>        5240 ±  3%     -23.8%       3993 ±  8%  interrupts.CPU25.PMI:Performance_monitoring_interrupts
>>>      791.67 ±  3%     +44.3%       1142 ±  7%  interrupts.CPU25.RES:Rescheduling_interrupts
>>>      594.00 ± 22%     +62.4%     964.75 ± 15%  interrupts.CPU25.TLB:TLB_shootdowns
>>>        4234 ± 13%     +20.7%       5111 ±  3%  interrupts.CPU26.CAL:Function_call_interrupts
>>>      821.33 ±  3%     +41.4%       1161 ±  6%  interrupts.CPU26.RES:Rescheduling_interrupts
>>>      528.00 ±  5%     +96.6%       1038 ± 18%  interrupts.CPU26.TLB:TLB_shootdowns
>>>        4202 ±  4%     +21.4%       5101 ±  4%  interrupts.CPU27.CAL:Function_call_interrupts
>>>        5218 ±  2%     -22.5%       4045 ±  8%  interrupts.CPU27.NMI:Non-maskable_interrupts
>>>        5218 ±  2%     -22.5%       4045 ±  8%  interrupts.CPU27.PMI:Performance_monitoring_interrupts
>>>      720.00 ±  5%     +58.6%       1142 ±  5%  interrupts.CPU27.RES:Rescheduling_interrupts
>>>      502.67 ± 16%     +89.7%     953.50 ± 12%  interrupts.CPU27.TLB:TLB_shootdowns
>>>        4162 ± 14%     +18.2%       4921 ±  5%  interrupts.CPU28.CAL:Function_call_interrupts
>>>        5157 ±  2%     -23.4%       3951 ±  7%  interrupts.CPU28.NMI:Non-maskable_interrupts
>>>        5157 ±  2%     -23.4%       3951 ±  7%  interrupts.CPU28.PMI:Performance_monitoring_interrupts
>>>      755.33 ±  3%     +47.4%       1113 ±  5%  interrupts.CPU28.RES:Rescheduling_interrupts
>>>      497.00 ± 10%     +91.9%     953.50 ±  3%  interrupts.CPU28.TLB:TLB_shootdowns
>>>        4088 ±  5%     +24.1%       5072 ±  6%  interrupts.CPU29.CAL:Function_call_interrupts
>>>        5188 ±  3%     -22.5%       4021 ±  6%  interrupts.CPU29.NMI:Non-maskable_interrupts
>>>        5188 ±  3%     -22.5%       4021 ±  6%  interrupts.CPU29.PMI:Performance_monitoring_interrupts
>>>      719.33 ±  6%     +57.0%       1129 ±  6%  interrupts.CPU29.RES:Rescheduling_interrupts
>>>      464.33 ±  2%    +131.1%       1073 ± 26%  interrupts.CPU29.TLB:TLB_shootdowns
>>>        3770           +25.7%       4739 ±  5%  interrupts.CPU3.CAL:Function_call_interrupts
>>>      746.33 ±  6%     +50.2%       1120 ± 11%  interrupts.CPU3.RES:Rescheduling_interrupts
>>>      511.67 ± 14%     +74.5%     892.75 ±  6%  interrupts.CPU3.TLB:TLB_shootdowns
>>>        4203 ±  7%     +18.9%       4998 ±  3%  interrupts.CPU30.CAL:Function_call_interrupts
>>>        5183 ±  3%     -41.1%       3055 ± 37%  interrupts.CPU30.NMI:Non-maskable_interrupts
>>>        5183 ±  3%     -41.1%       3055 ± 37%  interrupts.CPU30.PMI:Performance_monitoring_interrupts
>>>      749.33 ±  5%     +52.9%       1146 ±  6%  interrupts.CPU30.RES:Rescheduling_interrupts
>>>      469.33 ± 17%    +108.5%     978.75 ±  8%  interrupts.CPU30.TLB:TLB_shootdowns
>>>        4133 ±  5%     +20.3%       4971 ±  2%  interrupts.CPU31.CAL:Function_call_interrupts
>>>        5174 ±  3%     -31.7%       3536 ± 26%  interrupts.CPU31.NMI:Non-maskable_interrupts
>>>        5174 ±  3%     -31.7%       3536 ± 26%  interrupts.CPU31.PMI:Performance_monitoring_interrupts
>>>      777.33 ±  5%     +53.8%       1195 ±  5%  interrupts.CPU31.RES:Rescheduling_interrupts
>>>      496.00 ±  9%     +96.7%     975.75 ± 16%  interrupts.CPU31.TLB:TLB_shootdowns
>>>        4107 ±  7%     +18.1%       4849 ±  2%  interrupts.CPU32.CAL:Function_call_interrupts
>>>        5163 ±  3%     -21.8%       4038 ±  7%  interrupts.CPU32.NMI:Non-maskable_interrupts
>>>        5163 ±  3%     -21.8%       4038 ±  7%  interrupts.CPU32.PMI:Performance_monitoring_interrupts
>>>      711.67           +52.9%       1088 ±  5%  interrupts.CPU32.RES:Rescheduling_interrupts
>>>      489.33 ± 18%    +115.7%       1055 ± 11%  interrupts.CPU32.TLB:TLB_shootdowns
>>>        4170 ±  4%     +17.4%       4897 ±  2%  interrupts.CPU33.CAL:Function_call_interrupts
>>>        5215 ±  3%     -33.3%       3480 ± 25%  interrupts.CPU33.NMI:Non-maskable_interrupts
>>>        5215 ±  3%     -33.3%       3480 ± 25%  interrupts.CPU33.PMI:Performance_monitoring_interrupts
>>>      785.33 ±  7%     +46.2%       1148 ±  4%  interrupts.CPU33.RES:Rescheduling_interrupts
>>>      547.67 ± 13%     +71.6%     939.75 ± 11%  interrupts.CPU33.TLB:TLB_shootdowns
>>>        4302 ±  6%     +16.0%       4992 ±  4%  interrupts.CPU34.CAL:Function_call_interrupts
>>>        5214 ±  3%     -22.7%       4031 ±  6%  interrupts.CPU34.NMI:Non-maskable_interrupts
>>>        5214 ±  3%     -22.7%       4031 ±  6%  interrupts.CPU34.PMI:Performance_monitoring_interrupts
>>>      787.33 ±  4%     +41.4%       1113 ±  4%  interrupts.CPU34.RES:Rescheduling_interrupts
>>>      590.33 ±  2%     +60.2%     945.50 ± 11%  interrupts.CPU34.TLB:TLB_shootdowns
>>>        4044 ± 11%     +22.9%       4969 ±  4%  interrupts.CPU35.CAL:Function_call_interrupts
>>>        5220 ±  3%     -32.6%       3520 ± 25%  interrupts.CPU35.NMI:Non-maskable_interrupts
>>>        5220 ±  3%     -32.6%       3520 ± 25%  interrupts.CPU35.PMI:Performance_monitoring_interrupts
>>>      735.67 ±  4%     +56.9%       1154 ±  9%  interrupts.CPU35.RES:Rescheduling_interrupts
>>>      485.00 ± 12%    +101.3%     976.50 ± 11%  interrupts.CPU35.TLB:TLB_shootdowns
>>>        4291 ±  8%     +14.9%       4930 ±  5%  interrupts.CPU36.CAL:Function_call_interrupts
>>>      782.67 ±  4%     +51.0%       1182 ±  4%  interrupts.CPU36.RES:Rescheduling_interrupts
>>>      491.00 ± 16%     +84.3%     904.75 ± 16%  interrupts.CPU36.TLB:TLB_shootdowns
>>>        3949 ±  5%     +24.1%       4898 ±  4%  interrupts.CPU37.CAL:Function_call_interrupts
>>>      723.67 ±  3%     +54.0%       1114 ±  7%  interrupts.CPU37.RES:Rescheduling_interrupts
>>>      490.00 ±  9%    +111.4%       1036 ±  8%  interrupts.CPU37.TLB:TLB_shootdowns
>>>        5218 ±  2%     -23.3%       4000 ±  7%  interrupts.CPU38.NMI:Non-maskable_interrupts
>>>        5218 ±  2%     -23.3%       4000 ±  7%  interrupts.CPU38.PMI:Performance_monitoring_interrupts
>>>      797.67 ± 10%     +40.3%       1119 ±  6%  interrupts.CPU38.RES:Rescheduling_interrupts
>>>      500.33 ± 16%     +89.6%     948.50 ± 12%  interrupts.CPU38.TLB:TLB_shootdowns
>>>        3987 ±  5%     +26.0%       5023 ±  5%  interrupts.CPU39.CAL:Function_call_interrupts
>>>      772.00           +47.0%       1134 ±  6%  interrupts.CPU39.RES:Rescheduling_interrupts
>>>      483.00 ± 10%    +102.2%     976.75 ± 17%  interrupts.CPU39.TLB:TLB_shootdowns
>>>        3933 ±  6%     +38.0%       5428 ± 14%  interrupts.CPU4.CAL:Function_call_interrupts
>>>        5183 ±  4%     -43.3%       2939 ± 31%  interrupts.CPU4.NMI:Non-maskable_interrupts
>>>        5183 ±  4%     -43.3%       2939 ± 31%  interrupts.CPU4.PMI:Performance_monitoring_interrupts
>>>      799.33 ±  5%     +35.5%       1083 ±  4%  interrupts.CPU4.RES:Rescheduling_interrupts
>>>      489.67 ± 14%    +100.9%     983.50 ± 11%  interrupts.CPU4.TLB:TLB_shootdowns
>>>        4034 ±  6%     +23.7%       4991        interrupts.CPU40.CAL:Function_call_interrupts
>>>        5212 ±  2%     -24.0%       3963 ±  7%  interrupts.CPU40.NMI:Non-maskable_interrupts
>>>        5212 ±  2%     -24.0%       3963 ±  7%  interrupts.CPU40.PMI:Performance_monitoring_interrupts
>>>      745.67 ±  5%     +53.6%       1145 ±  6%  interrupts.CPU40.RES:Rescheduling_interrupts
>>>      407.67 ± 11%    +140.6%     981.00 ± 13%  interrupts.CPU40.TLB:TLB_shootdowns
>>>        4029 ±  3%     +25.1%       5040 ±  4%  interrupts.CPU41.CAL:Function_call_interrupts
>>>        5237 ±  2%     -24.0%       3980 ±  7%  interrupts.CPU41.NMI:Non-maskable_interrupts
>>>        5237 ±  2%     -24.0%       3980 ±  7%  interrupts.CPU41.PMI:Performance_monitoring_interrupts
>>>      764.00 ± 11%     +49.6%       1143 ± 10%  interrupts.CPU41.RES:Rescheduling_interrupts
>>>      480.33 ± 15%    +102.8%     974.00 ± 14%  interrupts.CPU41.TLB:TLB_shootdowns
>>>        4158 ±  9%     +19.1%       4950 ±  5%  interrupts.CPU42.CAL:Function_call_interrupts
>>>        5205 ±  2%     -23.9%       3963 ±  7%  interrupts.CPU42.NMI:Non-maskable_interrupts
>>>        5205 ±  2%     -23.9%       3963 ±  7%  interrupts.CPU42.PMI:Performance_monitoring_interrupts
>>>      777.00           +48.0%       1150 ±  7%  interrupts.CPU42.RES:Rescheduling_interrupts
>>>      494.67 ±  7%     +85.4%     917.25 ±  6%  interrupts.CPU42.TLB:TLB_shootdowns
>>>        4145 ±  5%     +20.0%       4974 ±  2%  interrupts.CPU43.CAL:Function_call_interrupts
>>>        5199 ±  3%     -23.6%       3971 ±  6%  interrupts.CPU43.NMI:Non-maskable_interrupts
>>>        5199 ±  3%     -23.6%       3971 ±  6%  interrupts.CPU43.PMI:Performance_monitoring_interrupts
>>>      734.00 ±  4%     +54.8%       1136 ±  3%  interrupts.CPU43.RES:Rescheduling_interrupts
>>>      479.33 ± 21%    +116.7%       1038 ± 12%  interrupts.CPU43.TLB:TLB_shootdowns
>>>        4038 ±  6%     +20.5%       4867 ±  2%  interrupts.CPU44.CAL:Function_call_interrupts
>>>        5236 ±  3%     -34.3%       3438 ± 24%  interrupts.CPU44.NMI:Non-maskable_interrupts
>>>        5236 ±  3%     -34.3%       3438 ± 24%  interrupts.CPU44.PMI:Performance_monitoring_interrupts
>>>      721.67           +57.3%       1135 ±  6%  interrupts.CPU44.RES:Rescheduling_interrupts
>>>      486.67 ± 12%     +94.9%     948.50 ± 11%  interrupts.CPU44.TLB:TLB_shootdowns
>>>        4089 ±  7%     +19.7%       4895 ±  4%  interrupts.CPU45.CAL:Function_call_interrupts
>>>      744.67 ±  4%     +45.4%       1083 ±  2%  interrupts.CPU45.RES:Rescheduling_interrupts
>>>      450.33 ± 17%    +114.5%     966.00 ± 19%  interrupts.CPU45.TLB:TLB_shootdowns
>>>        3816 ±  5%     +30.8%       4991 ±  5%  interrupts.CPU46.CAL:Function_call_interrupts
>>>        5219 ±  3%     -32.9%       3502 ± 23%  interrupts.CPU46.NMI:Non-maskable_interrupts
>>>        5219 ±  3%     -32.9%       3502 ± 23%  interrupts.CPU46.PMI:Performance_monitoring_interrupts
>>>      740.33 ±  8%     +61.8%       1197 ±  5%  interrupts.CPU46.RES:Rescheduling_interrupts
>>>      455.00 ±  6%    +107.6%     944.75 ± 15%  interrupts.CPU46.TLB:TLB_shootdowns
>>>        3989 ±  2%     +24.0%       4947 ±  2%  interrupts.CPU47.CAL:Function_call_interrupts
>>>        5204 ±  3%     -32.6%       3506 ± 24%  interrupts.CPU47.NMI:Non-maskable_interrupts
>>>        5204 ±  3%     -32.6%       3506 ± 24%  interrupts.CPU47.PMI:Performance_monitoring_interrupts
>>>      741.67 ±  4%     +53.1%       1135 ± 12%  interrupts.CPU47.RES:Rescheduling_interrupts
>>>      437.33 ± 17%    +120.8%     965.75 ± 16%  interrupts.CPU47.TLB:TLB_shootdowns
>>>        4165 ±  2%     +22.3%       5095 ±  3%  interrupts.CPU48.CAL:Function_call_interrupts
>>>        5176 ±  4%     -21.7%       4053 ±  8%  interrupts.CPU48.NMI:Non-maskable_interrupts
>>>        5176 ±  4%     -21.7%       4053 ±  8%  interrupts.CPU48.PMI:Performance_monitoring_interrupts
>>>      835.00 ±  2%     +29.0%       1077 ±  2%  interrupts.CPU48.RES:Rescheduling_interrupts
>>>      594.67 ± 17%     +62.2%     964.50 ±  5%  interrupts.CPU48.TLB:TLB_shootdowns
>>>        4171 ±  9%     +18.7%       4950 ±  6%  interrupts.CPU49.CAL:Function_call_interrupts
>>>        5186 ±  4%     -33.4%       3452 ± 24%  interrupts.CPU49.NMI:Non-maskable_interrupts
>>>        5186 ±  4%     -33.4%       3452 ± 24%  interrupts.CPU49.PMI:Performance_monitoring_interrupts
>>>      710.67 ±  4%     +55.6%       1105 ±  4%  interrupts.CPU49.RES:Rescheduling_interrupts
>>>      553.67 ±  5%     +74.0%     963.25 ± 10%  interrupts.CPU49.TLB:TLB_shootdowns
>>>        4064 ±  6%     +21.6%       4944 ±  5%  interrupts.CPU5.CAL:Function_call_interrupts
>>>      772.33 ±  2%     +45.7%       1125 ±  4%  interrupts.CPU5.RES:Rescheduling_interrupts
>>>      491.00 ± 14%     +95.2%     958.50 ± 10%  interrupts.CPU5.TLB:TLB_shootdowns
>>>        3929 ±  4%     +26.4%       4966 ±  2%  interrupts.CPU50.CAL:Function_call_interrupts
>>>        5201 ±  4%     -22.9%       4011 ±  7%  interrupts.CPU50.NMI:Non-maskable_interrupts
>>>        5201 ±  4%     -22.9%       4011 ±  7%  interrupts.CPU50.PMI:Performance_monitoring_interrupts
>>>      747.00 ±  4%     +40.6%       1050 ±  3%  interrupts.CPU50.RES:Rescheduling_interrupts
>>>      482.33 ± 17%    +103.1%     979.75 ± 12%  interrupts.CPU50.TLB:TLB_shootdowns
>>>        4192 ±  4%     +20.3%       5043 ±  4%  interrupts.CPU51.CAL:Function_call_interrupts
>>>        5249 ±  3%     -24.0%       3987 ±  7%  interrupts.CPU51.NMI:Non-maskable_interrupts
>>>        5249 ±  3%     -24.0%       3987 ±  7%  interrupts.CPU51.PMI:Performance_monitoring_interrupts
>>>      780.33 ±  7%     +54.9%       1209 ±  9%  interrupts.CPU51.RES:Rescheduling_interrupts
>>>      504.00 ± 11%     +95.9%     987.50 ±  9%  interrupts.CPU51.TLB:TLB_shootdowns
>>>        4084 ±  4%     +21.8%       4972 ±  4%  interrupts.CPU52.CAL:Function_call_interrupts
>>>        5208 ±  4%     -23.1%       4006 ± 11%  interrupts.CPU52.NMI:Non-maskable_interrupts
>>>        5208 ±  4%     -23.1%       4006 ± 11%  interrupts.CPU52.PMI:Performance_monitoring_interrupts
>>>      724.33 ±  3%     +48.6%       1076 ±  4%  interrupts.CPU52.RES:Rescheduling_interrupts
>>>      536.00 ± 13%     +91.7%       1027 ±  7%  interrupts.CPU52.TLB:TLB_shootdowns
>>>        5184 ±  3%     -24.1%       3933 ± 10%  interrupts.CPU53.NMI:Non-maskable_interrupts
>>>        5184 ±  3%     -24.1%       3933 ± 10%  interrupts.CPU53.PMI:Performance_monitoring_interrupts
>>>      722.67 ±  4%     +49.1%       1077 ±  5%  interrupts.CPU53.RES:Rescheduling_interrupts
>>>      447.33 ± 20%    +118.6%     978.00 ± 11%  interrupts.CPU53.TLB:TLB_shootdowns
>>>        4057 ±  3%     +25.0%       5073 ±  5%  interrupts.CPU54.CAL:Function_call_interrupts
>>>        5172 ±  4%     -32.6%       3487 ± 24%  interrupts.CPU54.NMI:Non-maskable_interrupts
>>>        5172 ±  4%     -32.6%       3487 ± 24%  interrupts.CPU54.PMI:Performance_monitoring_interrupts
>>>      722.67           +50.7%       1088 ±  7%  interrupts.CPU54.RES:Rescheduling_interrupts
>>>      491.33 ± 14%    +382.4%       2370 ±102%  interrupts.CPU54.TLB:TLB_shootdowns
>>>        4092 ±  5%     +17.9%       4825        interrupts.CPU55.CAL:Function_call_interrupts
>>>      755.67 ± 10%     +49.2%       1127 ±  4%  interrupts.CPU55.RES:Rescheduling_interrupts
>>>      445.67          +117.2%     968.00 ± 11%  interrupts.CPU55.TLB:TLB_shootdowns
>>>        4116 ±  4%     +20.0%       4940 ±  5%  interrupts.CPU56.CAL:Function_call_interrupts
>>>        5144 ±  5%     -33.4%       3427 ± 23%  interrupts.CPU56.NMI:Non-maskable_interrupts
>>>        5144 ±  5%     -33.4%       3427 ± 23%  interrupts.CPU56.PMI:Performance_monitoring_interrupts
>>>      707.67           +55.5%       1100        interrupts.CPU56.RES:Rescheduling_interrupts
>>>      448.00 ± 16%    +120.1%     986.00 ± 13%  interrupts.CPU56.TLB:TLB_shootdowns
>>>        3861           +26.2%       4871 ±  3%  interrupts.CPU57.CAL:Function_call_interrupts
>>>        5187 ±  3%     -33.0%       3474 ± 25%  interrupts.CPU57.NMI:Non-maskable_interrupts
>>>        5187 ±  3%     -33.0%       3474 ± 25%  interrupts.CPU57.PMI:Performance_monitoring_interrupts
>>>      703.00 ±  3%     +55.6%       1094 ±  4%  interrupts.CPU57.RES:Rescheduling_interrupts
>>>      495.67 ± 19%    +106.5%       1023 ±  8%  interrupts.CPU57.TLB:TLB_shootdowns
>>>        4003 ±  6%     +25.3%       5016 ±  4%  interrupts.CPU58.CAL:Function_call_interrupts
>>>        5267           -32.7%       3543 ± 30%  interrupts.CPU58.NMI:Non-maskable_interrupts
>>>        5267           -32.7%       3543 ± 30%  interrupts.CPU58.PMI:Performance_monitoring_interrupts
>>>      692.00           +61.0%       1114        interrupts.CPU58.RES:Rescheduling_interrupts
>>>      467.67 ± 16%    +109.8%     981.25 ±  9%  interrupts.CPU58.TLB:TLB_shootdowns
>>>        4147 ±  7%     +21.2%       5024 ±  4%  interrupts.CPU59.CAL:Function_call_interrupts
>>>        5175 ±  3%     -33.9%       3419 ± 24%  interrupts.CPU59.NMI:Non-maskable_interrupts
>>>        5175 ±  3%     -33.9%       3419 ± 24%  interrupts.CPU59.PMI:Performance_monitoring_interrupts
>>>      765.00 ±  2%     +40.8%       1077 ±  5%  interrupts.CPU59.RES:Rescheduling_interrupts
>>>      477.67 ± 23%     +97.7%     944.25 ±  9%  interrupts.CPU59.TLB:TLB_shootdowns
>>>        4121 ±  7%     +19.7%       4932 ±  4%  interrupts.CPU6.CAL:Function_call_interrupts
>>>        5158 ±  4%     -31.8%       3518 ± 25%  interrupts.CPU6.NMI:Non-maskable_interrupts
>>>        5158 ±  4%     -31.8%       3518 ± 25%  interrupts.CPU6.PMI:Performance_monitoring_interrupts
>>>      735.33 ± 10%     +48.5%       1091 ±  5%  interrupts.CPU6.RES:Rescheduling_interrupts
>>>      535.33 ± 13%     +72.5%     923.25 ± 14%  interrupts.CPU6.TLB:TLB_shootdowns
>>>        4122 ±  4%     +32.7%       5471 ± 18%  interrupts.CPU60.CAL:Function_call_interrupts
>>>      746.67 ±  4%     +48.7%       1110 ±  4%  interrupts.CPU60.RES:Rescheduling_interrupts
>>>      496.00 ± 23%     +92.2%     953.25 ± 13%  interrupts.CPU60.TLB:TLB_shootdowns
>>>        4022 ±  4%     +33.7%       5376 ± 14%  interrupts.CPU61.CAL:Function_call_interrupts
>>>        5190 ±  4%     -34.9%       3378 ± 24%  interrupts.CPU61.NMI:Non-maskable_interrupts
>>>        5190 ±  4%     -34.9%       3378 ± 24%  interrupts.CPU61.PMI:Performance_monitoring_interrupts
>>>      734.33 ±  4%     +48.2%       1088 ±  2%  interrupts.CPU61.RES:Rescheduling_interrupts
>>>      484.67 ± 15%    +109.0%       1013 ±  9%  interrupts.CPU61.TLB:TLB_shootdowns
>>>        3917 ±  5%     +26.4%       4953 ±  3%  interrupts.CPU62.CAL:Function_call_interrupts
>>>        5189 ±  4%     -32.8%       3489 ± 25%  interrupts.CPU62.NMI:Non-maskable_interrupts
>>>        5189 ±  4%     -32.8%       3489 ± 25%  interrupts.CPU62.PMI:Performance_monitoring_interrupts
>>>      778.00 ±  3%     +42.6%       1109 ±  2%  interrupts.CPU62.RES:Rescheduling_interrupts
>>>      486.33 ± 23%    +110.0%       1021 ± 17%  interrupts.CPU62.TLB:TLB_shootdowns
>>>        4163 ±  8%     +20.8%       5029        interrupts.CPU63.CAL:Function_call_interrupts
>>>        5231 ±  2%     -34.6%       3419 ± 24%  interrupts.CPU63.NMI:Non-maskable_interrupts
>>>        5231 ±  2%     -34.6%       3419 ± 24%  interrupts.CPU63.PMI:Performance_monitoring_interrupts
>>>      775.67 ±  6%     +39.5%       1082        interrupts.CPU63.RES:Rescheduling_interrupts
>>>      491.00 ± 22%     +95.8%     961.25 ± 16%  interrupts.CPU63.TLB:TLB_shootdowns
>>>        4016 ±  2%     +24.2%       4990 ±  3%  interrupts.CPU64.CAL:Function_call_interrupts
>>>        5129 ±  5%     -42.0%       2973 ± 36%  interrupts.CPU64.NMI:Non-maskable_interrupts
>>>        5129 ±  5%     -42.0%       2973 ± 36%  interrupts.CPU64.PMI:Performance_monitoring_interrupts
>>>      757.67 ±  4%     +47.7%       1119 ±  3%  interrupts.CPU64.RES:Rescheduling_interrupts
>>>      453.67 ± 11%    +119.8%     997.00 ± 11%  interrupts.CPU64.TLB:TLB_shootdowns
>>>        3854 ±  2%     +26.3%       4868 ±  7%  interrupts.CPU65.CAL:Function_call_interrupts
>>>      767.33 ±  7%     +37.7%       1056 ±  3%  interrupts.CPU65.RES:Rescheduling_interrupts
>>>      519.67 ± 14%     +85.3%     962.75 ± 13%  interrupts.CPU65.TLB:TLB_shootdowns
>>>        3915 ±  8%     +24.8%       4888 ±  3%  interrupts.CPU66.CAL:Function_call_interrupts
>>>        5199 ±  3%     -34.4%       3409 ± 25%  interrupts.CPU66.NMI:Non-maskable_interrupts
>>>        5199 ±  3%     -34.4%       3409 ± 25%  interrupts.CPU66.PMI:Performance_monitoring_interrupts
>>>      712.67 ±  3%     +56.9%       1118 ±  6%  interrupts.CPU66.RES:Rescheduling_interrupts
>>>      431.67 ± 30%    +138.8%       1030 ± 14%  interrupts.CPU66.TLB:TLB_shootdowns
>>>        4026 ±  9%     +21.0%       4870 ±  3%  interrupts.CPU67.CAL:Function_call_interrupts
>>>        5157 ±  5%     -23.4%       3950 ±  9%  interrupts.CPU67.NMI:Non-maskable_interrupts
>>>        5157 ±  5%     -23.4%       3950 ±  9%  interrupts.CPU67.PMI:Performance_monitoring_interrupts
>>>      671.00           +56.9%       1053 ±  2%  interrupts.CPU67.RES:Rescheduling_interrupts
>>>      489.67 ± 16%     +99.6%     977.50 ± 12%  interrupts.CPU67.TLB:TLB_shootdowns
>>>        4034 ±  4%     +22.0%       4921 ±  2%  interrupts.CPU68.CAL:Function_call_interrupts
>>>        5175 ±  4%     -31.2%       3561 ± 30%  interrupts.CPU68.NMI:Non-maskable_interrupts
>>>        5175 ±  4%     -31.2%       3561 ± 30%  interrupts.CPU68.PMI:Performance_monitoring_interrupts
>>>      709.00 ±  4%     +56.2%       1107 ±  3%  interrupts.CPU68.RES:Rescheduling_interrupts
>>>      493.00 ± 15%     +95.8%     965.25 ± 12%  interrupts.CPU68.TLB:TLB_shootdowns
>>>        4014 ± 10%     +24.6%       5001 ±  2%  interrupts.CPU69.CAL:Function_call_interrupts
>>>        5152 ±  4%     -32.5%       3476 ± 25%  interrupts.CPU69.NMI:Non-maskable_interrupts
>>>        5152 ±  4%     -32.5%       3476 ± 25%  interrupts.CPU69.PMI:Performance_monitoring_interrupts
>>>      717.33 ±  3%     +50.3%       1078 ±  7%  interrupts.CPU69.RES:Rescheduling_interrupts
>>>      434.67 ± 26%    +123.0%     969.25 ±  8%  interrupts.CPU69.TLB:TLB_shootdowns
>>>        3972 ±  8%     +23.1%       4892        interrupts.CPU7.CAL:Function_call_interrupts
>>>      776.67 ±  5%     +40.6%       1091 ±  4%  interrupts.CPU7.RES:Rescheduling_interrupts
>>>      467.33 ± 11%    +102.8%     947.75 ± 16%  interrupts.CPU7.TLB:TLB_shootdowns
>>>        4068 ± 11%     +26.2%       5134        interrupts.CPU70.CAL:Function_call_interrupts
>>>        5188 ±  4%     -42.6%       2975 ± 37%  interrupts.CPU70.NMI:Non-maskable_interrupts
>>>        5188 ±  4%     -42.6%       2975 ± 37%  interrupts.CPU70.PMI:Performance_monitoring_interrupts
>>>      718.00 ±  4%     +54.3%       1107 ±  2%  interrupts.CPU70.RES:Rescheduling_interrupts
>>>      494.00 ± 11%    +104.5%       1010 ±  6%  interrupts.CPU70.TLB:TLB_shootdowns
>>>        3920 ±  4%     +29.5%       5078 ±  4%  interrupts.CPU71.CAL:Function_call_interrupts
>>>        5112 ±  4%     -23.3%       3920 ± 11%  interrupts.CPU71.NMI:Non-maskable_interrupts
>>>        5112 ±  4%     -23.3%       3920 ± 11%  interrupts.CPU71.PMI:Performance_monitoring_interrupts
>>>      667.67           +64.9%       1101 ±  5%  interrupts.CPU71.RES:Rescheduling_interrupts
>>>      473.00 ± 13%    +103.8%     963.75 ± 13%  interrupts.CPU71.TLB:TLB_shootdowns
>>>        3913 ±  3%     +26.8%       4963 ±  2%  interrupts.CPU72.CAL:Function_call_interrupts
>>>        5206 ±  3%     -24.5%       3932 ± 10%  interrupts.CPU72.NMI:Non-maskable_interrupts
>>>        5206 ±  3%     -24.5%       3932 ± 10%  interrupts.CPU72.PMI:Performance_monitoring_interrupts
>>>      811.67 ±  3%     +33.3%       1081 ±  4%  interrupts.CPU72.RES:Rescheduling_interrupts
>>>        3939 ±  3%     +26.0%       4962 ±  6%  interrupts.CPU73.CAL:Function_call_interrupts
>>>        5184 ±  2%     -34.3%       3408 ± 25%  interrupts.CPU73.NMI:Non-maskable_interrupts
>>>        5184 ±  2%     -34.3%       3408 ± 25%  interrupts.CPU73.PMI:Performance_monitoring_interrupts
>>>      763.00 ±  4%     +40.6%       1072 ±  4%  interrupts.CPU73.RES:Rescheduling_interrupts
>>>      570.00 ±  5%     +97.1%       1123 ± 25%  interrupts.CPU73.TLB:TLB_shootdowns
>>>        4069 ±  6%     +21.7%       4954 ±  3%  interrupts.CPU74.CAL:Function_call_interrupts
>>>        5223 ±  3%     -23.9%       3972 ± 10%  interrupts.CPU74.NMI:Non-maskable_interrupts
>>>        5223 ±  3%     -23.9%       3972 ± 10%  interrupts.CPU74.PMI:Performance_monitoring_interrupts
>>>      715.33           +53.9%       1101 ±  5%  interrupts.CPU74.RES:Rescheduling_interrupts
>>>      526.00 ± 14%     +90.9%       1004 ± 11%  interrupts.CPU74.TLB:TLB_shootdowns
>>>        4034 ± 10%     +21.5%       4900 ±  4%  interrupts.CPU75.CAL:Function_call_interrupts
>>>        5241 ±  2%     -42.9%       2995 ± 37%  interrupts.CPU75.NMI:Non-maskable_interrupts
>>>        5241 ±  2%     -42.9%       2995 ± 37%  interrupts.CPU75.PMI:Performance_monitoring_interrupts
>>>      712.00 ±  2%     +49.6%       1065 ±  6%  interrupts.CPU75.RES:Rescheduling_interrupts
>>>        4203 ±  8%     +18.8%       4995 ±  5%  interrupts.CPU76.CAL:Function_call_interrupts
>>>        5216 ±  2%     -32.4%       3525 ± 31%  interrupts.CPU76.NMI:Non-maskable_interrupts
>>>        5216 ±  2%     -32.4%       3525 ± 31%  interrupts.CPU76.PMI:Performance_monitoring_interrupts
>>>      719.00 ±  2%     +44.5%       1039 ±  5%  interrupts.CPU76.RES:Rescheduling_interrupts
>>>      488.00 ± 10%     +87.1%     913.25 ±  9%  interrupts.CPU76.TLB:TLB_shootdowns
>>>        3955 ±  7%     +23.9%       4901 ±  3%  interrupts.CPU77.CAL:Function_call_interrupts
>>>        5224 ±  3%     -24.5%       3944 ± 11%  interrupts.CPU77.NMI:Non-maskable_interrupts
>>>        5224 ±  3%     -24.5%       3944 ± 11%  interrupts.CPU77.PMI:Performance_monitoring_interrupts
>>>      703.67 ±  8%     +50.9%       1062 ±  3%  interrupts.CPU77.RES:Rescheduling_interrupts
>>>      466.67 ±  7%    +125.8%       1053 ± 13%  interrupts.CPU77.TLB:TLB_shootdowns
>>>        4189 ±  6%     +15.8%       4852 ±  6%  interrupts.CPU78.CAL:Function_call_interrupts
>>>        5190 ±  4%     -33.4%       3455 ± 24%  interrupts.CPU78.NMI:Non-maskable_interrupts
>>>        5190 ±  4%     -33.4%       3455 ± 24%  interrupts.CPU78.PMI:Performance_monitoring_interrupts
>>>      706.00           +49.5%       1055 ±  4%  interrupts.CPU78.RES:Rescheduling_interrupts
>>>      442.67 ±  8%    +135.9%       1044 ± 15%  interrupts.CPU78.TLB:TLB_shootdowns
>>>        3902 ±  3%     +24.7%       4865 ±  3%  interrupts.CPU79.CAL:Function_call_interrupts
>>>        5225 ±  3%     -35.1%       3390 ± 24%  interrupts.CPU79.NMI:Non-maskable_interrupts
>>>        5225 ±  3%     -35.1%       3390 ± 24%  interrupts.CPU79.PMI:Performance_monitoring_interrupts
>>>      704.67 ±  4%     +58.3%       1115 ±  7%  interrupts.CPU79.RES:Rescheduling_interrupts
>>>      444.00 ± 13%    +111.1%     937.25 ± 11%  interrupts.CPU79.TLB:TLB_shootdowns
>>>        4089 ±  4%     +19.8%       4898        interrupts.CPU8.CAL:Function_call_interrupts
>>>        5181 ±  4%     -22.3%       4024 ±  5%  interrupts.CPU8.NMI:Non-maskable_interrupts
>>>        5181 ±  4%     -22.3%       4024 ±  5%  interrupts.CPU8.PMI:Performance_monitoring_interrupts
>>>      785.33 ±  4%     +41.3%       1110 ±  6%  interrupts.CPU8.RES:Rescheduling_interrupts
>>>      480.00 ±  9%     +94.0%     931.00 ± 17%  interrupts.CPU8.TLB:TLB_shootdowns
>>>        4093 ±  6%     +22.6%       5018 ±  4%  interrupts.CPU80.CAL:Function_call_interrupts
>>>      721.33 ±  5%     +44.9%       1045 ±  2%  interrupts.CPU80.RES:Rescheduling_interrupts
>>>      485.33 ± 10%    +101.9%     980.00 ±  6%  interrupts.CPU80.TLB:TLB_shootdowns
>>>        4059 ±  9%     +19.3%       4843 ±  4%  interrupts.CPU81.CAL:Function_call_interrupts
>>>        5243 ±  3%     -25.7%       3894 ± 10%  interrupts.CPU81.NMI:Non-maskable_interrupts
>>>        5243 ±  3%     -25.7%       3894 ± 10%  interrupts.CPU81.PMI:Performance_monitoring_interrupts
>>>      689.67 ±  7%     +53.0%       1055 ±  7%  interrupts.CPU81.RES:Rescheduling_interrupts
>>>      553.67 ± 16%     +87.9%       1040 ± 13%  interrupts.CPU81.TLB:TLB_shootdowns
>>>        3950 ±  8%     +23.1%       4863 ±  6%  interrupts.CPU82.CAL:Function_call_interrupts
>>>        5207 ±  3%     -33.8%       3447 ± 25%  interrupts.CPU82.NMI:Non-maskable_interrupts
>>>        5207 ±  3%     -33.8%       3447 ± 25%  interrupts.CPU82.PMI:Performance_monitoring_interrupts
>>>      723.33           +45.0%       1048 ±  2%  interrupts.CPU82.RES:Rescheduling_interrupts
>>>      424.67 ± 12%    +143.4%       1033 ± 11%  interrupts.CPU82.TLB:TLB_shootdowns
>>>        4227 ±  9%     +20.3%       5085 ±  6%  interrupts.CPU83.CAL:Function_call_interrupts
>>>        5224 ±  3%     -42.6%       2998 ± 38%  interrupts.CPU83.NMI:Non-maskable_interrupts
>>>        5224 ±  3%     -42.6%       2998 ± 38%  interrupts.CPU83.PMI:Performance_monitoring_interrupts
>>>      690.00 ±  7%     +57.2%       1084 ±  3%  interrupts.CPU83.RES:Rescheduling_interrupts
>>>      437.33 ±  9%    +170.6%       1183 ± 19%  interrupts.CPU83.TLB:TLB_shootdowns
>>>        4122 ± 10%     +19.3%       4918 ±  3%  interrupts.CPU84.CAL:Function_call_interrupts
>>>        5236 ±  3%     -44.5%       2906 ± 29%  interrupts.CPU84.NMI:Non-maskable_interrupts
>>>        5236 ±  3%     -44.5%       2906 ± 29%  interrupts.CPU84.PMI:Performance_monitoring_interrupts
>>>      714.00 ±  2%     +49.4%       1066 ±  4%  interrupts.CPU84.RES:Rescheduling_interrupts
>>>      460.00 ± 17%    +124.9%       1034 ± 11%  interrupts.CPU84.TLB:TLB_shootdowns
>>>        4072 ±  5%     +19.5%       4864 ±  4%  interrupts.CPU85.CAL:Function_call_interrupts
>>>        5256 ±  3%     -24.5%       3967 ±  9%  interrupts.CPU85.NMI:Non-maskable_interrupts
>>>        5256 ±  3%     -24.5%       3967 ±  9%  interrupts.CPU85.PMI:Performance_monitoring_interrupts
>>>      696.67 ±  5%     +53.2%       1067 ±  5%  interrupts.CPU85.RES:Rescheduling_interrupts
>>>      445.67 ± 13%    +112.0%     945.00 ±  2%  interrupts.CPU85.TLB:TLB_shootdowns
>>>        4091 ±  3%     +21.3%       4964 ±  4%  interrupts.CPU86.CAL:Function_call_interrupts
>>>      699.33 ±  5%     +56.4%       1093 ±  4%  interrupts.CPU86.RES:Rescheduling_interrupts
>>>      466.00 ± 14%    +126.0%       1053 ±  4%  interrupts.CPU86.TLB:TLB_shootdowns
>>>        4003 ±  8%     +25.3%       5015 ±  4%  interrupts.CPU87.CAL:Function_call_interrupts
>>>      687.00 ±  3%     +54.1%       1058 ±  5%  interrupts.CPU87.RES:Rescheduling_interrupts
>>>      474.33 ± 18%    +116.1%       1025 ±  9%  interrupts.CPU87.TLB:TLB_shootdowns
>>>        4087 ±  8%     +21.9%       4982 ±  3%  interrupts.CPU88.CAL:Function_call_interrupts
>>>      695.67           +48.5%       1033 ±  2%  interrupts.CPU88.RES:Rescheduling_interrupts
>>>      482.00 ± 12%    +114.2%       1032 ±  7%  interrupts.CPU88.TLB:TLB_shootdowns
>>>        3940 ±  8%     +25.5%       4944 ±  5%  interrupts.CPU89.CAL:Function_call_interrupts
>>>        5209 ±  3%     -25.2%       3895 ± 10%  interrupts.CPU89.NMI:Non-maskable_interrupts
>>>        5209 ±  3%     -25.2%       3895 ± 10%  interrupts.CPU89.PMI:Performance_monitoring_interrupts
>>>      695.67 ±  4%     +51.6%       1054 ±  4%  interrupts.CPU89.RES:Rescheduling_interrupts
>>>      434.00 ± 15%    +131.6%       1005 ±  9%  interrupts.CPU89.TLB:TLB_shootdowns
>>>        3991 ±  2%     +18.5%       4728 ±  3%  interrupts.CPU9.CAL:Function_call_interrupts
>>>        5149 ±  5%     -33.6%       3418 ± 25%  interrupts.CPU9.NMI:Non-maskable_interrupts
>>>        5149 ±  5%     -33.6%       3418 ± 25%  interrupts.CPU9.PMI:Performance_monitoring_interrupts
>>>      704.67 ±  7%     +56.2%       1100 ±  4%  interrupts.CPU9.RES:Rescheduling_interrupts
>>>      433.67 ±  8%    +107.6%     900.25 ± 12%  interrupts.CPU9.TLB:TLB_shootdowns
>>>        4145 ±  6%     +17.4%       4868 ±  6%  interrupts.CPU90.CAL:Function_call_interrupts
>>>        5217 ±  3%     -42.9%       2977 ± 40%  interrupts.CPU90.NMI:Non-maskable_interrupts
>>>        5217 ±  3%     -42.9%       2977 ± 40%  interrupts.CPU90.PMI:Performance_monitoring_interrupts
>>>      672.33 ±  4%     +61.4%       1085 ±  2%  interrupts.CPU90.RES:Rescheduling_interrupts
>>>      491.33 ± 14%     +90.1%     934.00 ±  9%  interrupts.CPU90.TLB:TLB_shootdowns
>>>        4304 ±  7%     +17.4%       5051 ±  3%  interrupts.CPU91.CAL:Function_call_interrupts
>>>      799.00 ±  4%     +34.7%       1076 ±  3%  interrupts.CPU91.RES:Rescheduling_interrupts
>>>      458.33 ± 13%    +112.6%     974.50 ±  3%  interrupts.CPU91.TLB:TLB_shootdowns
>>>        3978 ±  4%     +21.1%       4816 ±  3%  interrupts.CPU92.CAL:Function_call_interrupts
>>>        5190 ±  4%     -54.3%       2371 ± 21%  interrupts.CPU92.NMI:Non-maskable_interrupts
>>>        5190 ±  4%     -54.3%       2371 ± 21%  interrupts.CPU92.PMI:Performance_monitoring_interrupts
>>>      648.00           +66.6%       1079 ±  4%  interrupts.CPU92.RES:Rescheduling_interrupts
>>>      470.67 ± 14%     +99.7%     940.00 ± 17%  interrupts.CPU92.TLB:TLB_shootdowns
>>>        3937 ±  6%     +23.5%       4863 ±  5%  interrupts.CPU93.CAL:Function_call_interrupts
>>>        5221 ±  3%     -35.6%       3365 ± 23%  interrupts.CPU93.NMI:Non-maskable_interrupts
>>>        5221 ±  3%     -35.6%       3365 ± 23%  interrupts.CPU93.PMI:Performance_monitoring_interrupts
>>>      677.67 ±  6%     +53.6%       1041 ±  5%  interrupts.CPU93.RES:Rescheduling_interrupts
>>>      481.00 ±  4%    +123.9%       1077 ±  4%  interrupts.CPU93.TLB:TLB_shootdowns
>>>        4005 ±  5%     +22.8%       4919 ±  6%  interrupts.CPU94.CAL:Function_call_interrupts
>>>      648.00 ±  2%     +61.7%       1047 ±  6%  interrupts.CPU94.RES:Rescheduling_interrupts
>>>      469.67 ±  9%    +103.2%     954.25 ± 17%  interrupts.CPU94.TLB:TLB_shootdowns
>>>        4039 ±  2%     +26.0%       5088 ±  5%  interrupts.CPU95.CAL:Function_call_interrupts
>>>      664.00 ±  8%     +65.4%       1098 ±  8%  interrupts.CPU95.RES:Rescheduling_interrupts
>>>      458.00 ± 18%    +119.0%       1003 ±  8%  interrupts.CPU95.TLB:TLB_shootdowns
>>>        3941 ±  2%     +20.5%       4750 ±  3%  interrupts.CPU96.CAL:Function_call_interrupts
>>>        5213 ±  5%     -26.6%       3824 ±  9%  interrupts.CPU96.NMI:Non-maskable_interrupts
>>>        5213 ±  5%     -26.6%       3824 ±  9%  interrupts.CPU96.PMI:Performance_monitoring_interrupts
>>>      707.67 ±  4%     +69.0%       1196 ±  4%  interrupts.CPU96.RES:Rescheduling_interrupts
>>>      453.00 ±  5%    +111.2%     956.75 ±  9%  interrupts.CPU96.TLB:TLB_shootdowns
>>>        3892 ±  3%     +29.5%       5041 ±  8%  interrupts.CPU97.CAL:Function_call_interrupts
>>>        5166 ±  4%     -25.1%       3867 ±  9%  interrupts.CPU97.NMI:Non-maskable_interrupts
>>>     ��  5166 ±  4%     -25.1%       3867 ±  9%  interrupts.CPU97.PMI:Performance_monitoring_interrupts
>>>      755.67 ±  8%     +55.6%       1175 ±  5%  interrupts.CPU97.RES:Rescheduling_interrupts
>>>      487.00 ± 10%     +86.1%     906.25 ±  8%  interrupts.CPU97.TLB:TLB_shootdowns
>>>        3869 ±  3%     +23.6%       4782        interrupts.CPU98.CAL:Function_call_interrupts
>>>        5178 ±  5%     -34.9%       3369 ± 24%  interrupts.CPU98.NMI:Non-maskable_interrupts
>>>        5178 ±  5%     -34.9%       3369 ± 24%  interrupts.CPU98.PMI:Performance_monitoring_interrupts
>>>      687.33 ±  5%     +68.1%       1155 ±  4%  interrupts.CPU98.RES:Rescheduling_interrupts
>>>      389.00 ± 13%    +121.1%     860.25 ±  8%  interrupts.CPU98.TLB:TLB_shootdowns
>>>        4000           +21.9%       4876 ±  6%  interrupts.CPU99.CAL:Function_call_interrupts
>>>      719.67           +67.0%       1202 ±  2%  interrupts.CPU99.RES:Rescheduling_interrupts
>>>      463.00 ±  8%     +76.1%     815.50 ± 10%  interrupts.CPU99.TLB:TLB_shootdowns
>>>      913220 ±  2%     -24.1%     693121 ±  9%  interrupts.NMI:Non-maskable_interrupts
>>>      913220 ±  2%     -24.1%     693121 ±  9%  interrupts.PMI:Performance_monitoring_interrupts
>>>      132379           +62.4%     215045        interrupts.RES:Rescheduling_interrupts
>>>       94465 ± 13%     +95.5%     184704 ±  9%  interrupts.TLB:TLB_shootdowns
>>>
>>>
>>>
>>>
>>>
>>> Disclaimer:
>>> Results have been estimated based on internal Intel analysis and are provided
>>> for informational purposes only. Any difference in system hardware or software
>>> design or configuration may affect actual performance.
>>>
>>>
>>> Thanks,
>>> Oliver Sang
>>> _______________________________________________
>>> LKP mailing list -- lkp@...ts.01.org
>>> To unsubscribe send an email to lkp-leave@...ts.01.org
>>
>> -- 
>> Zhengjun Xing
>>
> 

-- 
Zhengjun Xing

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ