lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite for Android: free password hash cracker in your pocket
[<prev] [next>] [thread-next>] [day] [month] [year] [list]
Message-ID: <20200407080036.GA8179@shao2-debian>
Date:   Tue, 7 Apr 2020 16:00:36 +0800
From:   kernel test robot <rong.a.chen@...el.com>
To:     Ritesh Harjani <riteshh@...ux.ibm.com>
Cc:     Theodore Ts'o <tytso@....edu>, kbuild test robot <lkp@...el.com>,
        Jan Kara <jack@...e.cz>,
        "Darrick J. Wong" <darrick.wong@...cle.com>,
        LKML <linux-kernel@...r.kernel.org>, lkp@...ts.01.org
Subject: [ext4] d3b6f23f71: stress-ng.fiemap.ops_per_sec -60.5% regression

Greeting,

FYI, we noticed a -60.5% regression of stress-ng.fiemap.ops_per_sec due to commit:


commit: d3b6f23f71670007817a5d59f3fbafab2b794e8c ("ext4: move ext4_fiemap to use iomap framework")
https://git.kernel.org/cgit/linux/kernel/git/torvalds/linux.git master

in testcase: stress-ng
on test machine: 96 threads Intel(R) Xeon(R) Gold 6252 CPU @ 2.10GHz with 192G memory
with following parameters:

	nr_threads: 10%
	disk: 1HDD
	testtime: 1s
	class: os
	cpufreq_governor: performance
	ucode: 0x500002c
	fs: ext4






Details are as below:
-------------------------------------------------------------------------------------------------->


To reproduce:

        git clone https://github.com/intel/lkp-tests.git
        cd lkp-tests
        bin/lkp install job.yaml  # job file is attached in this email
        bin/lkp run     job.yaml

=========================================================================================
class/compiler/cpufreq_governor/disk/fs/kconfig/nr_threads/rootfs/tbox_group/testcase/testtime/ucode:
  os/gcc-7/performance/1HDD/ext4/x86_64-rhel-7.6/10%/debian-x86_64-20191114.cgz/lkp-csl-2sp5/stress-ng/1s/0x500002c

commit: 
  b2c5764262 ("ext4: make ext4_ind_map_blocks work with fiemap")
  d3b6f23f71 ("ext4: move ext4_fiemap to use iomap framework")

b2c5764262edded1 d3b6f23f71670007817a5d59f3f 
---------------- --------------------------- 
       fail:runs  %reproduction    fail:runs
           |             |             |    
           :4           25%           1:4     dmesg.WARNING:at#for_ip_interrupt_entry/0x
          2:4            5%           2:4     perf-profile.calltrace.cycles-pp.sync_regs.error_entry
          2:4            6%           3:4     perf-profile.calltrace.cycles-pp.error_entry
          3:4            9%           3:4     perf-profile.children.cycles-pp.error_entry
          0:4            1%           0:4     perf-profile.self.cycles-pp.error_entry
         %stddev     %change         %stddev
             \          |                \  
     28623           +28.2%      36703 ± 12%  stress-ng.daemon.ops
     28632           +28.2%      36704 ± 12%  stress-ng.daemon.ops_per_sec
    566.00 ± 22%     -53.2%     265.00 ± 53%  stress-ng.dev.ops
    278.81 ± 22%     -53.0%     131.00 ± 54%  stress-ng.dev.ops_per_sec
     73160           -60.6%      28849 ±  3%  stress-ng.fiemap.ops
     72471           -60.5%      28612 ±  3%  stress-ng.fiemap.ops_per_sec
     23421 ± 12%     +21.2%      28388 ±  6%  stress-ng.filename.ops
     22638 ± 12%     +20.3%      27241 ±  6%  stress-ng.filename.ops_per_sec
     21.25 ±  7%     -10.6%      19.00 ±  3%  stress-ng.iomix.ops
     38.75 ± 49%     -47.7%      20.25 ± 96%  stress-ng.memhotplug.ops
     34.45 ± 52%     -51.8%      16.62 ±106%  stress-ng.memhotplug.ops_per_sec
      1734 ± 10%     +31.4%       2278 ± 10%  stress-ng.resources.ops
    807.56 ±  5%     +35.2%       1091 ±  8%  stress-ng.resources.ops_per_sec
   1007356 ±  3%     -16.5%     840642 ±  9%  stress-ng.revio.ops
   1007692 ±  3%     -16.6%     840711 ±  9%  stress-ng.revio.ops_per_sec
     21812 ±  3%     +16.0%      25294 ±  5%  stress-ng.sysbadaddr.ops
     21821 ±  3%     +15.9%      25294 ±  5%  stress-ng.sysbadaddr.ops_per_sec
    440.75 ±  4%     +21.9%     537.25 ±  9%  stress-ng.sysfs.ops
    440.53 ±  4%     +21.9%     536.86 ±  9%  stress-ng.sysfs.ops_per_sec
  13286582           -11.1%   11805520 ±  6%  stress-ng.time.file_system_outputs
  68253896            +2.4%   69860122        stress-ng.time.minor_page_faults
    197.00 ±  4%     -15.9%     165.75 ± 12%  stress-ng.xattr.ops
    192.45 ±  5%     -16.1%     161.46 ± 11%  stress-ng.xattr.ops_per_sec
     15310           +62.5%      24875 ± 22%  stress-ng.zombie.ops
     15310           +62.5%      24874 ± 22%  stress-ng.zombie.ops_per_sec
    203.50 ± 12%     -47.3%     107.25 ± 49%  vmstat.io.bi
    861318 ± 18%     -29.7%     605884 ±  5%  meminfo.AnonHugePages
   1062742 ± 14%     -20.2%     847853 ±  3%  meminfo.AnonPages
     31093 ±  6%      +9.6%      34090 ±  3%  meminfo.KernelStack
      7151 ± 34%     +55.8%      11145 ±  9%  meminfo.Mlocked
 1.082e+08 ±  5%     -40.2%   64705429 ± 31%  numa-numastat.node0.local_node
 1.082e+08 ±  5%     -40.2%   64739883 ± 31%  numa-numastat.node0.numa_hit
  46032662 ± 21%    +104.3%   94042918 ± 20%  numa-numastat.node1.local_node
  46074205 ± 21%    +104.2%   94072810 ± 20%  numa-numastat.node1.numa_hit
      3942 ±  3%     +14.2%       4501 ±  4%  slabinfo.pool_workqueue.active_objs
      4098 ±  3%     +14.3%       4683 ±  4%  slabinfo.pool_workqueue.num_objs
      4817 ±  7%     +13.3%       5456 ±  8%  slabinfo.proc_dir_entry.active_objs
      5153 ±  6%     +12.5%       5797 ±  8%  slabinfo.proc_dir_entry.num_objs
     18598 ± 13%     -33.1%      12437 ± 20%  sched_debug.cfs_rq:/.load.avg
    452595 ± 56%     -71.4%     129637 ± 76%  sched_debug.cfs_rq:/.load.max
     67675 ± 35%     -55.1%      30377 ± 42%  sched_debug.cfs_rq:/.load.stddev
     18114 ± 12%     -33.7%      12011 ± 20%  sched_debug.cfs_rq:/.runnable_weight.avg
    448215 ± 58%     -72.8%     121789 ± 82%  sched_debug.cfs_rq:/.runnable_weight.max
     67083 ± 37%     -56.3%      29305 ± 43%  sched_debug.cfs_rq:/.runnable_weight.stddev
    -38032          +434.3%    -203212        sched_debug.cfs_rq:/.spread0.avg
   -204466           +95.8%    -400301        sched_debug.cfs_rq:/.spread0.min
     90.02 ± 25%     -58.1%      37.69 ± 52%  sched_debug.cfs_rq:/.util_est_enqueued.avg
    677.54 ±  6%     -39.3%     411.50 ± 22%  sched_debug.cfs_rq:/.util_est_enqueued.max
    196.57 ±  8%     -47.6%     103.05 ± 36%  sched_debug.cfs_rq:/.util_est_enqueued.stddev
      3.34 ± 23%     +34.1%       4.48 ±  4%  sched_debug.cpu.clock.stddev
      3.34 ± 23%     +34.1%       4.48 ±  4%  sched_debug.cpu.clock_task.stddev
    402872 ±  7%     -11.9%     354819 ±  2%  proc-vmstat.nr_active_anon
   1730331            -9.5%    1566418 ±  5%  proc-vmstat.nr_dirtied
     31042 ±  6%      +9.3%      33915 ±  3%  proc-vmstat.nr_kernel_stack
    229047            -2.4%     223615        proc-vmstat.nr_mapped
     74008 ±  7%     +20.5%      89163 ±  8%  proc-vmstat.nr_written
    402872 ±  7%     -11.9%     354819 ±  2%  proc-vmstat.nr_zone_active_anon
     50587 ± 11%     -25.2%      37829 ± 14%  proc-vmstat.numa_pages_migrated
    457500           -23.1%     351918 ± 31%  proc-vmstat.numa_pte_updates
  81382485            +1.9%   82907822        proc-vmstat.pgfault
 2.885e+08 ±  5%     -13.3%  2.502e+08 ±  6%  proc-vmstat.pgfree
     42206 ± 12%     -46.9%      22399 ± 49%  proc-vmstat.pgpgin
    431233 ± 13%     -64.8%     151736 ±109%  proc-vmstat.pgrotated
    176754 ±  7%     -40.2%     105637 ± 31%  proc-vmstat.thp_fault_alloc
    314.50 ± 82%    +341.5%       1388 ± 44%  proc-vmstat.unevictable_pgs_stranded
   1075269 ± 14%     -41.3%     631388 ± 17%  numa-meminfo.node0.Active
    976056 ± 12%     -39.7%     588727 ± 19%  numa-meminfo.node0.Active(anon)
    426857 ± 22%     -36.4%     271375 ± 13%  numa-meminfo.node0.AnonHugePages
    558590 ± 19%     -36.4%     355402 ± 14%  numa-meminfo.node0.AnonPages
   1794824 ±  9%     -28.8%    1277157 ± 20%  numa-meminfo.node0.FilePages
      8517 ± 92%     -82.7%       1473 ± 89%  numa-meminfo.node0.Inactive(file)
    633118 ±  2%     -41.7%     368920 ± 36%  numa-meminfo.node0.Mapped
   2958038 ± 12%     -27.7%    2139271 ± 12%  numa-meminfo.node0.MemUsed
    181401 ±  5%     -13.7%     156561 ±  4%  numa-meminfo.node0.SUnreclaim
    258124 ±  6%     -13.0%     224535 ±  5%  numa-meminfo.node0.Slab
    702083 ± 16%     +31.0%     919406 ± 11%  numa-meminfo.node1.Active
     38663 ±107%    +137.8%      91951 ± 31%  numa-meminfo.node1.Active(file)
   1154975 ±  7%     +41.6%    1635593 ± 12%  numa-meminfo.node1.FilePages
    395813 ± 25%     +62.8%     644533 ± 16%  numa-meminfo.node1.Inactive
    394313 ± 25%     +62.5%     640686 ± 16%  numa-meminfo.node1.Inactive(anon)
    273317           +88.8%     515976 ± 25%  numa-meminfo.node1.Mapped
   2279237 ±  6%     +25.7%    2865582 ±  7%  numa-meminfo.node1.MemUsed
     10830 ± 18%     +29.6%      14033 ±  9%  numa-meminfo.node1.PageTables
    149390 ±  3%     +23.2%     184085 ±  3%  numa-meminfo.node1.SUnreclaim
    569542 ± 16%     +74.8%     995336 ± 21%  numa-meminfo.node1.Shmem
    220774 ±  5%     +20.3%     265656 ±  3%  numa-meminfo.node1.Slab
  35623587 ±  5%     -11.7%   31444514 ±  3%  perf-stat.i.cache-misses
 2.576e+08 ±  5%      -6.8%    2.4e+08 ±  2%  perf-stat.i.cache-references
      3585            -7.3%       3323 ±  5%  perf-stat.i.cpu-migrations
    180139 ±  2%      +4.2%     187668        perf-stat.i.minor-faults
     69.13            +2.6       71.75        perf-stat.i.node-load-miss-rate%
   4313695 ±  2%      -7.4%    3994957 ±  2%  perf-stat.i.node-load-misses
   5466253 ± 11%     -17.3%    4521173 ±  6%  perf-stat.i.node-loads
   2818674 ±  6%     -15.8%    2372542 ±  5%  perf-stat.i.node-stores
    227810            +4.6%     238290        perf-stat.i.page-faults
     12.67 ±  4%      -7.2%      11.76 ±  2%  perf-stat.overall.MPKI
      1.01 ±  4%      -0.0        0.97 ±  3%  perf-stat.overall.branch-miss-rate%
      1044           +13.1%       1181 ±  4%  perf-stat.overall.cycles-between-cache-misses
     40.37 ±  4%      +3.6       44.00 ±  2%  perf-stat.overall.node-store-miss-rate%
  36139526 ±  5%     -12.5%   31625519 ±  3%  perf-stat.ps.cache-misses
 2.566e+08 ±  5%      -6.9%  2.389e+08 ±  2%  perf-stat.ps.cache-references
      3562            -7.2%       3306 ±  5%  perf-stat.ps.cpu-migrations
    179088            +4.2%     186579        perf-stat.ps.minor-faults
   4323383 ±  2%      -7.5%    3999214        perf-stat.ps.node-load-misses
   5607721 ± 10%     -18.5%    4568664 ±  6%  perf-stat.ps.node-loads
   2855134 ±  7%     -16.4%    2387345 ±  5%  perf-stat.ps.node-stores
    226270            +4.6%     236709        perf-stat.ps.page-faults
    242305 ± 10%     -42.4%     139551 ± 18%  numa-vmstat.node0.nr_active_anon
    135983 ± 17%     -37.4%      85189 ± 10%  numa-vmstat.node0.nr_anon_pages
    209.25 ± 16%     -38.1%     129.50 ± 10%  numa-vmstat.node0.nr_anon_transparent_hugepages
    449367 ±  9%     -29.7%     315804 ± 20%  numa-vmstat.node0.nr_file_pages
      2167 ± 90%     -80.6%     419.75 ± 98%  numa-vmstat.node0.nr_inactive_file
    157405 ±  3%     -41.4%      92206 ± 35%  numa-vmstat.node0.nr_mapped
      2022 ± 30%     -73.3%     539.25 ± 91%  numa-vmstat.node0.nr_mlock
      3336 ± 10%     -24.3%       2524 ± 25%  numa-vmstat.node0.nr_page_table_pages
    286158 ± 10%     -41.2%     168337 ± 37%  numa-vmstat.node0.nr_shmem
     45493 ±  5%     -14.1%      39094 ±  4%  numa-vmstat.node0.nr_slab_unreclaimable
    242294 ± 10%     -42.4%     139547 ± 18%  numa-vmstat.node0.nr_zone_active_anon
      2167 ± 90%     -80.6%     419.75 ± 98%  numa-vmstat.node0.nr_zone_inactive_file
  54053924 ±  8%     -39.3%   32786242 ± 34%  numa-vmstat.node0.numa_hit
  53929628 ±  8%     -39.5%   32619715 ± 34%  numa-vmstat.node0.numa_local
      9701 ±107%    +136.9%      22985 ± 31%  numa-vmstat.node1.nr_active_file
    202.50 ± 16%     -25.1%     151.75 ± 23%  numa-vmstat.node1.nr_anon_transparent_hugepages
    284922 ±  7%     +43.3%     408195 ± 13%  numa-vmstat.node1.nr_file_pages
     96002 ± 26%     +67.5%     160850 ± 17%  numa-vmstat.node1.nr_inactive_anon
     68077 ±  2%     +90.3%     129533 ± 25%  numa-vmstat.node1.nr_mapped
    138482 ± 15%     +79.2%     248100 ± 22%  numa-vmstat.node1.nr_shmem
     37396 ±  3%     +23.3%      46094 ±  3%  numa-vmstat.node1.nr_slab_unreclaimable
      9701 ±107%    +136.9%      22985 ± 31%  numa-vmstat.node1.nr_zone_active_file
     96005 ± 26%     +67.5%     160846 ± 17%  numa-vmstat.node1.nr_zone_inactive_anon
  23343661 ± 17%     +99.9%   46664267 ± 23%  numa-vmstat.node1.numa_hit
  23248487 ± 17%    +100.5%   46610447 ± 23%  numa-vmstat.node1.numa_local
    105745 ± 23%    +112.6%     224805 ± 24%  softirqs.CPU0.NET_RX
    133310 ± 36%     -45.3%      72987 ± 52%  softirqs.CPU1.NET_RX
    170110 ± 55%     -66.8%      56407 ±147%  softirqs.CPU11.NET_RX
     91465 ± 36%     -65.2%      31858 ±112%  softirqs.CPU13.NET_RX
    164491 ± 57%     -77.7%      36641 ±121%  softirqs.CPU15.NET_RX
    121069 ± 55%     -99.3%     816.75 ± 96%  softirqs.CPU17.NET_RX
     81019 ±  4%      -8.7%      73967 ±  4%  softirqs.CPU20.RCU
     72143 ± 63%     -89.8%       7360 ±172%  softirqs.CPU22.NET_RX
    270663 ± 17%     -57.9%     113915 ± 45%  softirqs.CPU24.NET_RX
     20149 ± 76%    +474.1%     115680 ± 62%  softirqs.CPU26.NET_RX
     14033 ± 70%    +977.5%     151211 ± 75%  softirqs.CPU27.NET_RX
     27834 ± 94%    +476.1%     160357 ± 28%  softirqs.CPU28.NET_RX
     35346 ± 68%    +212.0%     110290 ± 30%  softirqs.CPU29.NET_RX
     34347 ±103%    +336.5%     149941 ± 32%  softirqs.CPU32.NET_RX
     70077 ±  3%     +10.8%      77624 ±  3%  softirqs.CPU34.RCU
     36453 ± 84%    +339.6%     160253 ± 42%  softirqs.CPU36.NET_RX
     72367 ±  2%     +10.6%      80043        softirqs.CPU37.RCU
     25239 ±118%    +267.7%      92799 ± 45%  softirqs.CPU38.NET_RX
      4995 ±170%   +1155.8%      62734 ± 62%  softirqs.CPU39.NET_RX
      4641 ±145%   +1611.3%      79432 ± 90%  softirqs.CPU42.NET_RX
      7192 ± 65%    +918.0%      73225 ± 66%  softirqs.CPU45.NET_RX
      1772 ±166%   +1837.4%      34344 ± 63%  softirqs.CPU46.NET_RX
     13149 ± 81%    +874.7%     128170 ± 58%  softirqs.CPU47.NET_RX
     86484 ± 94%     -92.6%       6357 ±172%  softirqs.CPU48.NET_RX
    129128 ± 27%     -95.8%       5434 ±172%  softirqs.CPU55.NET_RX
     82772 ± 59%     -91.7%       6891 ±164%  softirqs.CPU56.NET_RX
    145313 ± 57%     -87.8%      17796 ± 88%  softirqs.CPU57.NET_RX
    118160 ± 33%     -86.3%      16226 ±109%  softirqs.CPU58.NET_RX
     94576 ± 56%     -94.1%       5557 ±173%  softirqs.CPU6.NET_RX
     82900 ± 77%     -66.8%      27508 ±171%  softirqs.CPU62.NET_RX
    157291 ± 30%     -81.1%      29656 ±111%  softirqs.CPU64.NET_RX
    135101 ± 28%     -80.2%      26748 ± 90%  softirqs.CPU67.NET_RX
    146574 ± 56%    -100.0%      69.75 ± 98%  softirqs.CPU68.NET_RX
     81347 ±  2%      -9.0%      74024 ±  2%  softirqs.CPU68.RCU
    201729 ± 37%     -99.6%     887.50 ±107%  softirqs.CPU69.NET_RX
    108454 ± 78%     -97.9%       2254 ±169%  softirqs.CPU70.NET_RX
     55289 ±104%     -89.3%       5942 ±172%  softirqs.CPU71.NET_RX
     10112 ±172%    +964.6%     107651 ± 89%  softirqs.CPU72.NET_RX
      3136 ±171%   +1522.2%      50879 ± 66%  softirqs.CPU73.NET_RX
     13353 ± 79%    +809.2%     121407 ±101%  softirqs.CPU74.NET_RX
     75194 ±  3%     +10.3%      82957 ±  5%  softirqs.CPU75.RCU
     11002 ±173%   +1040.8%     125512 ± 61%  softirqs.CPU76.NET_RX
      2463 ±173%   +2567.3%      65708 ± 77%  softirqs.CPU78.NET_RX
     25956 ±  3%      -7.8%      23932 ±  3%  softirqs.CPU78.SCHED
     16366 ±150%    +340.7%      72125 ± 91%  softirqs.CPU82.NET_RX
     14553 ±130%   +1513.4%     234809 ± 27%  softirqs.CPU93.NET_RX
     26314            -9.2%      23884 ±  3%  softirqs.CPU93.SCHED
      4582 ± 88%   +4903.4%     229268 ± 23%  softirqs.CPU94.NET_RX
     11214 ±111%   +1762.5%     208867 ± 18%  softirqs.CPU95.NET_RX
      1.53 ± 27%      -0.5        0.99 ± 17%  perf-profile.calltrace.cycles-pp.exit_to_usermode_loop.do_syscall_64.entry_SYSCALL_64_after_hwframe
      1.52 ± 27%      -0.5        0.99 ± 17%  perf-profile.calltrace.cycles-pp.do_signal.exit_to_usermode_loop.do_syscall_64.entry_SYSCALL_64_after_hwframe
      1.39 ± 29%      -0.5        0.88 ± 21%  perf-profile.calltrace.cycles-pp.do_group_exit.get_signal.do_signal.exit_to_usermode_loop.do_syscall_64
      1.39 ± 29%      -0.5        0.88 ± 21%  perf-profile.calltrace.cycles-pp.get_signal.do_signal.exit_to_usermode_loop.do_syscall_64.entry_SYSCALL_64_after_hwframe
      0.50 ± 59%      +0.3        0.81 ± 13%  perf-profile.calltrace.cycles-pp.filemap_map_pages.handle_pte_fault.__handle_mm_fault.handle_mm_fault.do_page_fault
      5.70 ±  9%      +0.8        6.47 ±  7%  perf-profile.calltrace.cycles-pp.do_exit.do_group_exit.get_signal.do_signal.exit_to_usermode_loop
      5.48 ±  9%      +0.8        6.27 ±  7%  perf-profile.calltrace.cycles-pp.exit_mmap.mmput.do_exit.do_group_exit.get_signal
      5.49 ±  9%      +0.8        6.28 ±  7%  perf-profile.calltrace.cycles-pp.mmput.do_exit.do_group_exit.get_signal.do_signal
      4.30 ±  4%      +1.3        5.60 ±  7%  perf-profile.calltrace.cycles-pp.do_group_exit.get_signal.do_signal.exit_to_usermode_loop.prepare_exit_to_usermode
      4.40 ±  4%      +1.3        5.69 ±  7%  perf-profile.calltrace.cycles-pp.prepare_exit_to_usermode.swapgs_restore_regs_and_return_to_usermode
      4.37 ±  4%      +1.3        5.66 ±  7%  perf-profile.calltrace.cycles-pp.exit_to_usermode_loop.prepare_exit_to_usermode.swapgs_restore_regs_and_return_to_usermode
      4.36 ±  4%      +1.3        5.66 ±  7%  perf-profile.calltrace.cycles-pp.do_signal.exit_to_usermode_loop.prepare_exit_to_usermode.swapgs_restore_regs_and_return_to_usermode
      4.33 ±  4%      +1.3        5.62 ±  7%  perf-profile.calltrace.cycles-pp.get_signal.do_signal.exit_to_usermode_loop.prepare_exit_to_usermode.swapgs_restore_regs_and_return_to_usermode
      4.44 ±  4%      +1.3        5.74 ±  7%  perf-profile.calltrace.cycles-pp.swapgs_restore_regs_and_return_to_usermode
      3.20 ± 10%      -2.4        0.78 ±156%  perf-profile.children.cycles-pp.copy_page
      0.16 ±  9%      -0.1        0.08 ± 64%  perf-profile.children.cycles-pp.irq_work_interrupt
      0.16 ±  9%      -0.1        0.08 ± 64%  perf-profile.children.cycles-pp.smp_irq_work_interrupt
      0.24 ±  5%      -0.1        0.17 ± 18%  perf-profile.children.cycles-pp.irq_work_run_list
      0.16 ±  9%      -0.1        0.10 ± 24%  perf-profile.children.cycles-pp.irq_work_run
      0.16 ±  9%      -0.1        0.10 ± 24%  perf-profile.children.cycles-pp.printk
      0.23 ±  6%      -0.1        0.17 ±  9%  perf-profile.children.cycles-pp.__do_execve_file
      0.08 ± 14%      -0.1        0.03 ±100%  perf-profile.children.cycles-pp.delay_tsc
      0.16 ±  6%      -0.1        0.11 ±  9%  perf-profile.children.cycles-pp.load_elf_binary
      0.16 ±  7%      -0.0        0.12 ± 13%  perf-profile.children.cycles-pp.search_binary_handler
      0.20 ±  7%      -0.0        0.15 ± 10%  perf-profile.children.cycles-pp.call_usermodehelper_exec_async
      0.19 ±  6%      -0.0        0.15 ± 11%  perf-profile.children.cycles-pp.do_execve
      0.08 ± 10%      -0.0        0.04 ± 59%  perf-profile.children.cycles-pp.__vunmap
      0.15 ±  3%      -0.0        0.11 ±  7%  perf-profile.children.cycles-pp.rcu_idle_exit
      0.12 ± 10%      -0.0        0.09 ± 14%  perf-profile.children.cycles-pp.__switch_to_asm
      0.09 ± 13%      -0.0        0.07 ±  5%  perf-profile.children.cycles-pp.des3_ede_encrypt
      0.06 ± 11%      +0.0        0.09 ± 13%  perf-profile.children.cycles-pp.mark_page_accessed
      0.15 ±  5%      +0.0        0.19 ± 12%  perf-profile.children.cycles-pp.apparmor_cred_prepare
      0.22 ±  8%      +0.0        0.27 ± 11%  perf-profile.children.cycles-pp.mem_cgroup_throttle_swaprate
      0.17 ±  2%      +0.0        0.22 ± 12%  perf-profile.children.cycles-pp.security_prepare_creds
      0.95 ± 17%      +0.3        1.22 ± 14%  perf-profile.children.cycles-pp.filemap_map_pages
      5.92 ±  8%      +0.7        6.65 ±  7%  perf-profile.children.cycles-pp.get_signal
      5.66 ±  9%      +0.8        6.44 ±  7%  perf-profile.children.cycles-pp.mmput
      5.65 ±  9%      +0.8        6.43 ±  7%  perf-profile.children.cycles-pp.exit_mmap
      4.40 ±  4%      +1.3        5.70 ±  7%  perf-profile.children.cycles-pp.prepare_exit_to_usermode
      4.45 ±  4%      +1.3        5.75 ±  7%  perf-profile.children.cycles-pp.swapgs_restore_regs_and_return_to_usermode
      3.16 ± 10%      -2.4        0.77 ±155%  perf-profile.self.cycles-pp.copy_page
      0.08 ± 14%      -0.1        0.03 ±100%  perf-profile.self.cycles-pp.delay_tsc
      0.12 ± 10%      -0.0        0.09 ± 14%  perf-profile.self.cycles-pp.__switch_to_asm
      0.08 ± 12%      -0.0        0.06 ± 17%  perf-profile.self.cycles-pp.enqueue_task_fair
      0.09 ± 13%      -0.0        0.07 ±  5%  perf-profile.self.cycles-pp.des3_ede_encrypt
      0.07 ± 13%      +0.0        0.08 ± 19%  perf-profile.self.cycles-pp.__lru_cache_add
      0.19 ±  9%      +0.0        0.22 ± 10%  perf-profile.self.cycles-pp.mem_cgroup_throttle_swaprate
      0.15 ±  5%      +0.0        0.19 ± 11%  perf-profile.self.cycles-pp.apparmor_cred_prepare
      0.05 ± 58%      +0.0        0.09 ± 13%  perf-profile.self.cycles-pp.mark_page_accessed
      0.58 ± 10%      +0.2        0.80 ± 20%  perf-profile.self.cycles-pp.release_pages
      0.75 ±173%  +1.3e+05%       1005 ±100%  interrupts.127:PCI-MSI.31981660-edge.i40e-eth0-TxRx-91
    820.75 ±111%     -99.9%       0.50 ±173%  interrupts.47:PCI-MSI.31981580-edge.i40e-eth0-TxRx-11
    449.25 ± 86%    -100.0%       0.00        interrupts.53:PCI-MSI.31981586-edge.i40e-eth0-TxRx-17
     33.25 ±157%    -100.0%       0.00        interrupts.57:PCI-MSI.31981590-edge.i40e-eth0-TxRx-21
      0.75 ±110%  +63533.3%     477.25 ±162%  interrupts.61:PCI-MSI.31981594-edge.i40e-eth0-TxRx-25
    561.50 ±160%    -100.0%       0.00        interrupts.65:PCI-MSI.31981598-edge.i40e-eth0-TxRx-29
     82921 ±  8%     -11.1%      73748 ±  6%  interrupts.CPU11.CAL:Function_call_interrupts
     66509 ± 30%     -32.6%      44828 ±  8%  interrupts.CPU14.TLB:TLB_shootdowns
     43105 ± 98%     -90.3%       4183 ± 21%  interrupts.CPU17.RES:Rescheduling_interrupts
    148719 ± 70%     -69.4%      45471 ± 16%  interrupts.CPU17.TLB:TLB_shootdowns
     85589 ± 42%     -52.2%      40884 ±  5%  interrupts.CPU20.TLB:TLB_shootdowns
    222472 ± 41%     -98.0%       4360 ± 45%  interrupts.CPU22.RES:Rescheduling_interrupts
      0.50 ±173%  +95350.0%     477.25 ±162%  interrupts.CPU25.61:PCI-MSI.31981594-edge.i40e-eth0-TxRx-25
     76029 ± 10%     +14.9%      87389 ±  5%  interrupts.CPU25.CAL:Function_call_interrupts
    399042 ±  6%     +13.4%     452479 ±  8%  interrupts.CPU27.LOC:Local_timer_interrupts
    561.00 ±161%    -100.0%       0.00        interrupts.CPU29.65:PCI-MSI.31981598-edge.i40e-eth0-TxRx-29
      7034 ± 46%   +1083.8%      83279 ±138%  interrupts.CPU29.RES:Rescheduling_interrupts
     17829 ± 99%     -71.0%       5172 ± 16%  interrupts.CPU30.RES:Rescheduling_interrupts
      5569 ± 15%   +2414.7%     140059 ± 94%  interrupts.CPU31.RES:Rescheduling_interrupts
     37674 ± 16%     +36.6%      51473 ± 25%  interrupts.CPU31.TLB:TLB_shootdowns
     47905 ± 39%     +76.6%      84583 ± 38%  interrupts.CPU34.TLB:TLB_shootdowns
    568.75 ±140%    +224.8%       1847 ± 90%  interrupts.CPU36.NMI:Non-maskable_interrupts
    568.75 ±140%    +224.8%       1847 ± 90%  interrupts.CPU36.PMI:Performance_monitoring_interrupts
      4236 ± 25%   +2168.5%      96092 ± 90%  interrupts.CPU36.RES:Rescheduling_interrupts
     52717 ± 27%     +43.3%      75565 ± 28%  interrupts.CPU37.TLB:TLB_shootdowns
     41418 ±  9%    +136.6%      98010 ± 50%  interrupts.CPU39.TLB:TLB_shootdowns
      5551 ±  8%    +847.8%      52615 ± 66%  interrupts.CPU40.RES:Rescheduling_interrupts
      4746 ± 25%    +865.9%      45841 ± 91%  interrupts.CPU42.RES:Rescheduling_interrupts
     37556 ± 11%     +24.6%      46808 ±  6%  interrupts.CPU42.TLB:TLB_shootdowns
     21846 ±124%     -84.4%       3415 ± 46%  interrupts.CPU48.RES:Rescheduling_interrupts
    891.50 ± 22%     -35.2%     577.25 ± 40%  interrupts.CPU49.NMI:Non-maskable_interrupts
    891.50 ± 22%     -35.2%     577.25 ± 40%  interrupts.CPU49.PMI:Performance_monitoring_interrupts
     20459 ±120%     -79.2%       4263 ± 14%  interrupts.CPU49.RES:Rescheduling_interrupts
     59840 ± 21%     -23.1%      46042 ± 16%  interrupts.CPU5.TLB:TLB_shootdowns
     65200 ± 19%     -34.5%      42678 ±  9%  interrupts.CPU51.TLB:TLB_shootdowns
     70923 ±153%     -94.0%       4270 ± 29%  interrupts.CPU53.RES:Rescheduling_interrupts
     65312 ± 22%     -28.7%      46578 ± 14%  interrupts.CPU56.TLB:TLB_shootdowns
     65828 ± 24%     -33.4%      43846 ±  4%  interrupts.CPU59.TLB:TLB_shootdowns
     72558 ±156%     -93.2%       4906 ±  9%  interrupts.CPU6.RES:Rescheduling_interrupts
     68698 ± 34%     -32.6%      46327 ± 18%  interrupts.CPU61.TLB:TLB_shootdowns
    109745 ± 44%     -57.4%      46711 ± 16%  interrupts.CPU62.TLB:TLB_shootdowns
     89714 ± 44%     -48.5%      46198 ±  7%  interrupts.CPU63.TLB:TLB_shootdowns
     59380 ±136%     -91.5%       5066 ± 13%  interrupts.CPU69.RES:Rescheduling_interrupts
     40094 ± 18%    +133.9%      93798 ± 44%  interrupts.CPU78.TLB:TLB_shootdowns
    129884 ± 72%     -55.3%      58034 ±157%  interrupts.CPU8.RES:Rescheduling_interrupts
     69984 ± 11%     +51.4%     105957 ± 20%  interrupts.CPU80.CAL:Function_call_interrupts
     32857 ± 10%    +128.7%      75131 ± 36%  interrupts.CPU80.TLB:TLB_shootdowns
     35726 ± 16%     +34.6%      48081 ± 12%  interrupts.CPU82.TLB:TLB_shootdowns
     73820 ± 17%     +28.2%      94643 ±  8%  interrupts.CPU84.CAL:Function_call_interrupts
     38829 ± 28%    +190.3%     112736 ± 42%  interrupts.CPU84.TLB:TLB_shootdowns
     36129 ±  4%     +47.6%      53329 ± 13%  interrupts.CPU85.TLB:TLB_shootdowns
      4693 ±  7%   +1323.0%      66793 ±145%  interrupts.CPU86.RES:Rescheduling_interrupts
     38003 ± 11%     +94.8%      74031 ± 43%  interrupts.CPU86.TLB:TLB_shootdowns
     78022 ±  3%      +7.9%      84210 ±  3%  interrupts.CPU87.CAL:Function_call_interrupts
     36359 ±  6%     +54.9%      56304 ± 48%  interrupts.CPU88.TLB:TLB_shootdowns
     89031 ±105%     -95.0%       4475 ± 40%  interrupts.CPU9.RES:Rescheduling_interrupts
     40085 ± 11%     +60.6%      64368 ± 27%  interrupts.CPU91.TLB:TLB_shootdowns
     42244 ± 10%     +44.8%      61162 ± 35%  interrupts.CPU94.TLB:TLB_shootdowns
     40959 ± 15%    +109.4%      85780 ± 41%  interrupts.CPU95.TLB:TLB_shootdowns


                                                                                
                                stress-ng.fiemap.ops                            
                                                                                
  80000 +-------------------------------------------------------------------+   
  75000 |..+.             .+..            .+..+..  .+.  .+..                |   
        |    +..+..+..+.+.     .+..+..  .+       +.   +.    +.+..+..+..+.+..|   
  70000 |-+                   +       +.                                    |   
  65000 |-+                                                                 |   
  60000 |-+                                                                 |   
  55000 |-+                                                                 |   
        |                                                                   |   
  50000 |-+                                                                 |   
  45000 |-+                                                                 |   
  40000 |-+                                                                 |   
  35000 |-+                O                                                |   
        |  O       O                       O     O    O                     |   
  30000 |-+  O  O     O O     O O     O  O    O     O    O  O O  O  O  O O  |   
  25000 +-------------------------------------------------------------------+   
                                                                                
                                                                                                                                                                
                            stress-ng.fiemap.ops_per_sec                        
                                                                                
  80000 +-------------------------------------------------------------------+   
  75000 |..               .+..               .+..                           |   
        |  +.  .+..+..+.+.     .+..+..  .+.+.    +..+.+..+..+.+..+..+..+.+..|   
  70000 |-+  +.               +       +.                                    |   
  65000 |-+                                                                 |   
  60000 |-+                                                                 |   
  55000 |-+                                                                 |   
        |                                                                   |   
  50000 |-+                                                                 |   
  45000 |-+                                                                 |   
  40000 |-+                                                                 |   
  35000 |-+                O                                                |   
        |  O       O                       O     O                          |   
  30000 |-+  O  O     O O     O O     O       O     O O  O  O    O  O  O O  |   
  25000 +-------------------------------------------------------------------+   
                                                                                
                                                                                
[*] bisect-good sample
[O] bisect-bad  sample



Disclaimer:
Results have been estimated based on internal Intel analysis and are provided
for informational purposes only. Any difference in system hardware or software
design or configuration may affect actual performance.


Thanks,
Rong Chen


View attachment "config-5.6.0-rc4-00016-gd3b6f23f71670" of type "text/plain" (203577 bytes)

View attachment "job-script" of type "text/plain" (7899 bytes)

View attachment "job.yaml" of type "text/plain" (5514 bytes)

View attachment "reproduce" of type "text/plain" (609 bytes)

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ