lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite for Android: free password hash cracker in your pocket
[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <f197860d-292f-c270-049d-49c1adbd9817@kernel.dk>
Date:   Fri, 27 Jul 2018 09:14:35 -0600
From:   Jens Axboe <axboe@...nel.dk>
To:     "Wang, Kemi" <kemi.wang@...el.com>,
        SeongJae Park <sj38.park@...il.com>
Cc:     "Ye, Xiaolong" <xiaolong.ye@...el.com>,
        "axboe@...com" <axboe@...com>, "lkp@...org" <lkp@...org>,
        "linux-kernel@...r.kernel.org" <linux-kernel@...r.kernel.org>
Subject: Re: [LKP] [lkp-robot] [brd] 316ba5736c: aim7.jobs-per-min -11.2%
 regression

SeongJae please take a look at this. It's been weeks now and silence
on your end.


On 7/26/18 6:58 AM, Wang, Kemi wrote:
> Hi, SeongJae
> Any update or any info you need from my side?
> 
> -----Original Message-----
> From: SeongJae Park [mailto:sj38.park@...il.com] 
> Sent: Wednesday, July 11, 2018 12:53 AM
> To: Wang, Kemi <kemi.wang@...el.com>
> Cc: Ye, Xiaolong <xiaolong.ye@...el.com>; axboe@...nel.dk; axboe@...com; lkp@...org; linux-kernel@...r.kernel.org
> Subject: Re: [LKP] [lkp-robot] [brd] 316ba5736c: aim7.jobs-per-min -11.2% regression
> 
> Oops, I mistakenly found this mail now.  I will look inside for this though it will take some time because I will not be in office for this week.
> 
> 
> Thanks,
> SeongJae Park
> On Tue, Jul 10, 2018 at 1:30 AM kemi <kemi.wang@...el.com> wrote:
>>
>> Hi, SeongJae
>>   Do you have any input for this regression? thanks
>>
>> On 2018年06月04日 13:52, kernel test robot wrote:
>>>
>>> Greeting,
>>>
>>> FYI, we noticed a -11.2% regression of aim7.jobs-per-min due to commit:
>>>
>>>
>>> commit: 316ba5736c9caa5dbcd84085989862d2df57431d ("brd: Mark as 
>>> non-rotational") 
>>> https://git.kernel.org/cgit/linux/kernel/git/axboe/linux-block.git 
>>> for-4.18/block
>>>
>>> in testcase: aim7
>>> on test machine: 40 threads Intel(R) Xeon(R) CPU E5-2690 v2 @ 
>>> 3.00GHz with 384G memory with following parameters:
>>>
>>>       disk: 1BRD_48G
>>>       fs: btrfs
>>>       test: disk_rw
>>>       load: 1500
>>>       cpufreq_governor: performance
>>>
>>> test-description: AIM7 is a traditional UNIX system level benchmark suite which is used to test and measure the performance of multiuser system.
>>> test-url: 
>>> https://sourceforge.net/projects/aimbench/files/aim-suite7/
>>>
>>>
>>>
>>> Details are as below:
>>> -------------------------------------------------------------------------------------------------->
>>>
>>> ====================================================================
>>> =====================
>>> compiler/cpufreq_governor/disk/fs/kconfig/load/rootfs/tbox_group/test/testcase:
>>>   
>>> gcc-7/performance/1BRD_48G/btrfs/x86_64-rhel-7.2/1500/debian-x86_64-
>>> 2016-08-31.cgz/lkp-ivb-ep01/disk_rw/aim7
>>>
>>> commit:
>>>   522a777566 ("block: consolidate struct request timestamp fields")
>>>   316ba5736c ("brd: Mark as non-rotational")
>>>
>>> 522a777566f56696 316ba5736c9caa5dbcd8408598
>>> ---------------- --------------------------
>>>          %stddev     %change         %stddev
>>>              \          |                \
>>>      28321           -11.2%      25147        aim7.jobs-per-min
>>>     318.19           +12.6%     358.23        aim7.time.elapsed_time
>>>     318.19           +12.6%     358.23        aim7.time.elapsed_time.max
>>>    1437526 ±  2%     +14.6%    1646849 ±  2%  aim7.time.involuntary_context_switches
>>>      11986           +14.2%      13691        aim7.time.system_time
>>>      73.06 ±  2%      -3.6%      70.43        aim7.time.user_time
>>>    2449470 ±  2%     -25.0%    1837521 ±  4%  aim7.time.voluntary_context_switches
>>>      20.25 ± 58%   +1681.5%     360.75 ±109%  numa-meminfo.node1.Mlocked
>>>     456062           -16.3%     381859        softirqs.SCHED
>>>       9015 ±  7%     -21.3%       7098 ± 22%  meminfo.CmaFree
>>>      47.50 ± 58%   +1355.8%     691.50 ± 92%  meminfo.Mlocked
>>>       5.24 ±  3%      -1.2        3.99 ±  2%  mpstat.cpu.idle%
>>>       0.61 ±  2%      -0.1        0.52 ±  2%  mpstat.cpu.usr%
>>>      16627           +12.8%      18762 ±  4%  slabinfo.Acpi-State.active_objs
>>>      16627           +12.9%      18775 ±  4%  slabinfo.Acpi-State.num_objs
>>>      57.00 ±  2%     +17.5%      67.00        vmstat.procs.r
>>>      20936           -24.8%      15752 ±  2%  vmstat.system.cs
>>>      45474            -1.7%      44681        vmstat.system.in
>>>       6.50 ± 59%   +1157.7%      81.75 ± 75%  numa-vmstat.node0.nr_mlock
>>>     242870 ±  3%     +13.2%     274913 ±  7%  numa-vmstat.node0.nr_written
>>>       2278 ±  7%     -22.6%       1763 ± 21%  numa-vmstat.node1.nr_free_cma
>>>       4.75 ± 58%   +1789.5%      89.75 ±109%  numa-vmstat.node1.nr_mlock
>>>   88018135 ±  3%     -48.9%   44980457 ±  7%  cpuidle.C1.time
>>>    1398288 ±  3%     -51.1%     683493 ±  9%  cpuidle.C1.usage
>>>    3499814 ±  2%     -38.5%    2153158 ±  5%  cpuidle.C1E.time
>>>      52722 ±  4%     -45.6%      28692 ±  6%  cpuidle.C1E.usage
>>>    9865857 ±  3%     -40.1%    5905155 ±  5%  cpuidle.C3.time
>>>      69656 ±  2%     -42.6%      39990 ±  5%  cpuidle.C3.usage
>>>     590856 ±  2%     -12.3%     517910        cpuidle.C6.usage
>>>      46160 ±  7%     -53.7%      21372 ± 11%  cpuidle.POLL.time
>>>       1716 ±  7%     -46.6%     916.25 ± 14%  cpuidle.POLL.usage
>>>     197656            +4.1%     205732        proc-vmstat.nr_active_file
>>>     191867            +4.1%     199647        proc-vmstat.nr_dirty
>>>     509282            +1.6%     517318        proc-vmstat.nr_file_pages
>>>       2282 ±  8%     -24.4%       1725 ± 22%  proc-vmstat.nr_free_cma
>>>     357.50           +10.6%     395.25 ±  2%  proc-vmstat.nr_inactive_file
>>>      11.50 ± 58%   +1397.8%     172.25 ± 93%  proc-vmstat.nr_mlock
>>>     970355 ±  4%     +14.6%    1111549 ±  8%  proc-vmstat.nr_written
>>>     197984            +4.1%     206034        proc-vmstat.nr_zone_active_file
>>>     357.50           +10.6%     395.25 ±  2%  proc-vmstat.nr_zone_inactive_file
>>>     192282            +4.1%     200126        proc-vmstat.nr_zone_write_pending
>>>    7901465 ±  3%     -14.0%    6795016 ± 16%  proc-vmstat.pgalloc_movable
>>>     886101           +10.2%     976329        proc-vmstat.pgfault
>>>  2.169e+12           +15.2%  2.497e+12        perf-stat.branch-instructions
>>>       0.41            -0.1        0.35        perf-stat.branch-miss-rate%
>>>      31.19 ±  2%      +1.6       32.82        perf-stat.cache-miss-rate%
>>>  9.116e+09            +8.3%  9.869e+09        perf-stat.cache-misses
>>>  2.924e+10            +2.9%  3.008e+10 ±  2%  perf-stat.cache-references
>>>    6712739 ±  2%     -15.4%    5678643 ±  2%  perf-stat.context-switches
>>>       4.02            +2.7%       4.13        perf-stat.cpi
>>>  3.761e+13           +17.3%  4.413e+13        perf-stat.cpu-cycles
>>>     606958           -13.7%     523758 ±  2%  perf-stat.cpu-migrations
>>>  2.476e+12           +13.4%  2.809e+12        perf-stat.dTLB-loads
>>>       0.18 ±  2%      -0.0        0.16 ±  9%  perf-stat.dTLB-store-miss-rate%
>>>  1.079e+09 ±  2%      -9.6%  9.755e+08 ±  9%  perf-stat.dTLB-store-misses
>>>  5.933e+11            +1.6%  6.029e+11        perf-stat.dTLB-stores
>>>  9.349e+12           +14.2%  1.068e+13        perf-stat.instructions
>>>      11247 ± 11%     +19.8%      13477 ±  9%  perf-stat.instructions-per-iTLB-miss
>>>       0.25            -2.6%       0.24        perf-stat.ipc
>>>     865561           +10.3%     954350        perf-stat.minor-faults
>>>  2.901e+09 ±  3%      +9.8%  3.186e+09 ±  3%  perf-stat.node-load-misses
>>>  3.682e+09 ±  3%     +11.0%  4.088e+09 ±  3%  perf-stat.node-loads
>>>  3.778e+09            +4.8%  3.959e+09 ±  2%  perf-stat.node-store-misses
>>>  5.079e+09            +6.4%  5.402e+09        perf-stat.node-stores
>>>     865565           +10.3%     954352        perf-stat.page-faults
>>>      51.75 ±  5%     -12.5%      45.30 ± 10%  sched_debug.cfs_rq:/.load_avg.avg
>>>     316.35 ±  3%     +17.2%     370.81 ±  8%  sched_debug.cfs_rq:/.util_est_enqueued.stddev
>>>      15294 ± 30%    +234.9%      51219 ± 76%  sched_debug.cpu.avg_idle.min
>>>     299443 ±  3%      -7.3%     277566 ±  5%  sched_debug.cpu.avg_idle.stddev
>>>       1182 ± 19%     -26.3%     872.02 ± 13%  sched_debug.cpu.nr_load_updates.stddev
>>>       1.22 ±  8%     +21.7%       1.48 ±  6%  sched_debug.cpu.nr_running.avg
>>>       2.75 ± 10%     +26.2%       3.47 ±  6%  sched_debug.cpu.nr_running.max
>>>       0.58 ±  7%     +24.2%       0.73 ±  6%  sched_debug.cpu.nr_running.stddev
>>>      77148           -20.0%      61702 ±  7%  sched_debug.cpu.nr_switches.avg
>>>      70024           -24.8%      52647 ±  8%  sched_debug.cpu.nr_switches.min
>>>       6662 ±  6%     +61.9%      10789 ± 24%  sched_debug.cpu.nr_switches.stddev
>>>      80.45 ± 18%     -19.1%      65.05 ±  6%  sched_debug.cpu.nr_uninterruptible.stddev
>>>      76819           -19.3%      62008 ±  8%  sched_debug.cpu.sched_count.avg
>>>      70616           -23.5%      53996 ±  8%  sched_debug.cpu.sched_count.min
>>>       5494 ±  9%     +85.3%      10179 ± 26%  sched_debug.cpu.sched_count.stddev
>>>      16936           -52.9%       7975 ±  9%  sched_debug.cpu.sched_goidle.avg
>>>      19281           -49.9%       9666 ±  7%  sched_debug.cpu.sched_goidle.max
>>>      15417           -54.8%       6962 ± 10%  sched_debug.cpu.sched_goidle.min
>>>     875.00 ±  6%     -35.0%     569.09 ± 13%  sched_debug.cpu.sched_goidle.stddev
>>>      40332           -23.5%      30851 ±  7%  sched_debug.cpu.ttwu_count.avg
>>>      35074           -26.3%      25833 ±  6%  sched_debug.cpu.ttwu_count.min
>>>       3239 ±  8%     +67.4%       5422 ± 28%  sched_debug.cpu.ttwu_count.stddev
>>>       5232           +27.4%       6665 ± 13%  sched_debug.cpu.ttwu_local.avg
>>>      15877 ± 12%     +77.5%      28184 ± 27%  sched_debug.cpu.ttwu_local.max
>>>       2530 ± 10%     +95.9%       4956 ± 27%  sched_debug.cpu.ttwu_local.stddev
>>>       2.52 ±  7%      -0.6        1.95 ±  3%  perf-profile.calltrace.cycles-pp.btrfs_dirty_pages.__btrfs_buffered_write.btrfs_file_write_iter.__vfs_write.vfs_write
>>>       1.48 ± 12%      -0.5        1.01 ±  4%  perf-profile.calltrace.cycles-pp.btrfs_get_extent.btrfs_dirty_pages.__btrfs_buffered_write.btrfs_file_write_iter.__vfs_write
>>>       1.18 ± 16%      -0.4        0.76 ±  7%  perf-profile.calltrace.cycles-pp.btrfs_search_slot.btrfs_lookup_file_extent.btrfs_get_extent.btrfs_dirty_pages.__btrfs_buffered_write
>>>       1.18 ± 16%      -0.4        0.76 ±  7%  perf-profile.calltrace.cycles-pp.btrfs_lookup_file_extent.btrfs_get_extent.btrfs_dirty_pages.__btrfs_buffered_write.btrfs_file_write_iter
>>>       0.90 ± 17%      -0.3        0.56 ±  4%  perf-profile.calltrace.cycles-pp.__dentry_kill.dentry_kill.dput.__fput.task_work_run
>>>       0.90 ± 17%      -0.3        0.56 ±  4%  perf-profile.calltrace.cycles-pp.evict.__dentry_kill.dentry_kill.dput.__fput
>>>       0.90 ± 17%      -0.3        0.56 ±  4%  perf-profile.calltrace.cycles-pp.dentry_kill.dput.__fput.task_work_run.exit_to_usermode_loop
>>>       0.90 ± 18%      -0.3        0.56 ±  4%  perf-profile.calltrace.cycles-pp.btrfs_evict_inode.evict.__dentry_kill.dentry_kill.dput
>>>       0.90 ± 17%      -0.3        0.57 ±  5%  perf-profile.calltrace.cycles-pp.exit_to_usermode_loop.do_syscall_64.entry_SYSCALL_64_after_hwframe
>>>       0.90 ± 17%      -0.3        0.57 ±  5%  perf-profile.calltrace.cycles-pp.task_work_run.exit_to_usermode_loop.do_syscall_64.entry_SYSCALL_64_after_hwframe
>>>       0.90 ± 17%      -0.3        0.57 ±  5%  perf-profile.calltrace.cycles-pp.__fput.task_work_run.exit_to_usermode_loop.do_syscall_64.entry_SYSCALL_64_after_hwframe
>>>       0.90 ± 17%      -0.3        0.57 ±  5%  perf-profile.calltrace.cycles-pp.dput.__fput.task_work_run.exit_to_usermode_loop.do_syscall_64
>>>       1.69            -0.1        1.54 ±  2%  perf-profile.calltrace.cycles-pp.lock_and_cleanup_extent_if_need.__btrfs_buffered_write.btrfs_file_write_iter.__vfs_write.vfs_write
>>>       0.87 ±  4%      -0.1        0.76 ±  2%  perf-profile.calltrace.cycles-pp.__clear_extent_bit.clear_extent_bit.lock_and_cleanup_extent_if_need.__btrfs_buffered_write.btrfs_file_write_iter
>>>       0.87 ±  4%      -0.1        0.76 ±  2%  perf-profile.calltrace.cycles-pp.clear_extent_bit.lock_and_cleanup_extent_if_need.__btrfs_buffered_write.btrfs_file_write_iter.__vfs_write
>>>       0.71 ±  6%      -0.1        0.61 ±  2%  perf-profile.calltrace.cycles-pp.clear_state_bit.__clear_extent_bit.clear_extent_bit.lock_and_cleanup_extent_if_need.__btrfs_buffered_write
>>>       0.69 ±  6%      -0.1        0.60 ±  2%  perf-profile.calltrace.cycles-pp.btrfs_clear_bit_hook.clear_state_bit.__clear_extent_bit.clear_extent_bit.lock_and_cleanup_extent_if_need
>>>      96.77            +0.6       97.33        perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe
>>>       0.00            +0.6        0.56 ±  3%  perf-profile.calltrace.cycles-pp.can_overcommit.reserve_metadata_bytes.btrfs_delalloc_reserve_metadata.__btrfs_buffered_write.btrfs_file_write_iter
>>>      96.72            +0.6       97.29        perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe
>>>      43.13            +0.8       43.91        perf-profile.calltrace.cycles-pp.btrfs_inode_rsv_release.__btrfs_buffered_write.btrfs_file_write_iter.__vfs_write.vfs_write
>>>      42.37            +0.8       43.16        perf-profile.calltrace.cycles-pp.native_queued_spin_lock_slowpath._raw_spin_lock.block_rsv_release_bytes.btrfs_inode_rsv_release.__btrfs_buffered_write
>>>      43.11            +0.8       43.89        perf-profile.calltrace.cycles-pp.block_rsv_release_bytes.btrfs_inode_rsv_release.__btrfs_buffered_write.btrfs_file_write_iter.__vfs_write
>>>      42.96            +0.8       43.77        perf-profile.calltrace.cycles-pp._raw_spin_lock.block_rsv_release_bytes.btrfs_inode_rsv_release.__btrfs_buffered_write.btrfs_file_write_iter
>>>      95.28            +0.9       96.23        perf-profile.calltrace.cycles-pp.ksys_write.do_syscall_64.entry_SYSCALL_64_after_hwframe
>>>      95.22            +1.0       96.18        perf-profile.calltrace.cycles-pp.vfs_write.ksys_write.do_syscall_64.entry_SYSCALL_64_after_hwframe
>>>      94.88            +1.0       95.85        perf-profile.calltrace.cycles-pp.__vfs_write.vfs_write.ksys_write.do_syscall_64.entry_SYSCALL_64_after_hwframe
>>>      94.83            +1.0       95.80        perf-profile.calltrace.cycles-pp.btrfs_file_write_iter.__vfs_write.vfs_write.ksys_write.do_syscall_64
>>>      94.51            +1.0       95.50        perf-profile.calltrace.cycles-pp.__btrfs_buffered_write.btrfs_file_write_iter.__vfs_write.vfs_write.ksys_write
>>>      42.44            +1.1       43.52        perf-profile.calltrace.cycles-pp._raw_spin_lock.reserve_metadata_bytes.btrfs_delalloc_reserve_metadata.__btrfs_buffered_write.btrfs_file_write_iter
>>>      42.09            +1.1       43.18        perf-profile.calltrace.cycles-pp.native_queued_spin_lock_slowpath._raw_spin_lock.reserve_metadata_bytes.btrfs_delalloc_reserve_metadata.__btrfs_buffered_write
>>>      44.07            +1.2       45.29        perf-profile.calltrace.cycles-pp.btrfs_delalloc_reserve_metadata.__btrfs_buffered_write.btrfs_file_write_iter.__vfs_write.vfs_write
>>>      43.42            +1.3       44.69        perf-profile.calltrace.cycles-pp.reserve_metadata_bytes.btrfs_delalloc_reserve_metadata.__btrfs_buffered_write.btrfs_file_write_iter.__vfs_write
>>>       2.06 ± 18%      -0.9        1.21 ±  6%  perf-profile.children.cycles-pp.btrfs_search_slot
>>>       2.54 ±  7%      -0.6        1.96 ±  3%  perf-profile.children.cycles-pp.btrfs_dirty_pages
>>>       1.05 ± 24%      -0.5        0.52 ±  9%  perf-profile.children.cycles-pp._raw_spin_lock_irqsave
>>>       1.50 ± 12%      -0.5        1.03 ±  4%  perf-profile.children.cycles-pp.btrfs_get_extent
>>>       1.22 ± 15%      -0.4        0.79 ±  8%  perf-profile.children.cycles-pp.btrfs_lookup_file_extent
>>>       0.81 ±  5%      -0.4        0.41 ±  6%  perf-profile.children.cycles-pp.btrfs_calc_reclaim_metadata_size
>>>       0.74 ± 24%      -0.4        0.35 ±  9%  perf-profile.children.cycles-pp.btrfs_lock_root_node
>>>       0.74 ± 24%      -0.4        0.35 ±  9%  perf-profile.children.cycles-pp.btrfs_tree_lock
>>>       0.90 ± 17%      -0.3        0.56 ±  4%  perf-profile.children.cycles-pp.__dentry_kill
>>>       0.90 ± 17%      -0.3        0.56 ±  4%  perf-profile.children.cycles-pp.evict
>>>       0.90 ± 17%      -0.3        0.56 ±  4%  perf-profile.children.cycles-pp.dentry_kill
>>>       0.90 ± 18%      -0.3        0.56 ±  4%  perf-profile.children.cycles-pp.btrfs_evict_inode
>>>       0.91 ± 18%      -0.3        0.57 ±  4%  perf-profile.children.cycles-pp.exit_to_usermode_loop
>>>       0.52 ± 20%      -0.3        0.18 ± 14%  perf-profile.children.cycles-pp.do_idle
>>>       0.90 ± 17%      -0.3        0.57 ±  5%  perf-profile.children.cycles-pp.task_work_run
>>>       0.90 ± 17%      -0.3        0.57 ±  5%  perf-profile.children.cycles-pp.__fput
>>>       0.90 ± 18%      -0.3        0.57 ±  4%  perf-profile.children.cycles-pp.dput
>>>       0.51 ± 20%      -0.3        0.18 ± 14%  perf-profile.children.cycles-pp.secondary_startup_64
>>>       0.51 ± 20%      -0.3        0.18 ± 14%  perf-profile.children.cycles-pp.cpu_startup_entry
>>>       0.50 ± 21%      -0.3        0.17 ± 16%  perf-profile.children.cycles-pp.start_secondary
>>>       0.47 ± 20%      -0.3        0.16 ± 13%  perf-profile.children.cycles-pp.cpuidle_enter_state
>>>       0.47 ± 19%      -0.3        0.16 ± 13%  perf-profile.children.cycles-pp.intel_idle
>>>       0.61 ± 20%      -0.3        0.36 ± 11%  perf-profile.children.cycles-pp.btrfs_tree_read_lock
>>>       0.47 ± 26%      -0.3        0.21 ± 10%  perf-profile.children.cycles-pp.prepare_to_wait_event
>>>       0.64 ± 18%      -0.2        0.39 ±  9%  perf-profile.children.cycles-pp.btrfs_read_lock_root_node
>>>       0.40 ± 22%      -0.2        0.21 ±  5%  perf-profile.children.cycles-pp.btrfs_clear_path_blocking
>>>       0.38 ± 23%      -0.2        0.19 ± 13%  perf-profile.children.cycles-pp.finish_wait
>>>       1.51 ±  3%      -0.2        1.35 ±  2%  perf-profile.children.cycles-pp.__clear_extent_bit
>>>       1.71            -0.1        1.56 ±  2%  perf-profile.children.cycles-pp.lock_and_cleanup_extent_if_need
>>>       0.29 ± 25%      -0.1        0.15 ± 10%  perf-profile.children.cycles-pp.btrfs_orphan_del
>>>       0.27 ± 27%      -0.1        0.12 ±  8%  perf-profile.children.cycles-pp.btrfs_del_orphan_item
>>>       0.33 ± 18%      -0.1        0.19 ±  9%  perf-profile.children.cycles-pp.queued_read_lock_slowpath
>>>       0.33 ± 19%      -0.1        0.20 ±  4%  perf-profile.children.cycles-pp.__wake_up_common_lock
>>>       0.45 ± 15%      -0.1        0.34 ±  2%  perf-profile.children.cycles-pp.btrfs_alloc_data_chunk_ondemand
>>>       0.47 ± 16%      -0.1        0.36 ±  4%  perf-profile.children.cycles-pp.btrfs_check_data_free_space
>>>       0.91 ±  4%      -0.1        0.81 ±  3%  perf-profile.children.cycles-pp.clear_extent_bit
>>>       1.07 ±  5%      -0.1        0.97        perf-profile.children.cycles-pp.__set_extent_bit
>>>       0.77 ±  6%      -0.1        0.69 ±  3%  perf-profile.children.cycles-pp.btrfs_clear_bit_hook
>>>       0.17 ± 20%      -0.1        0.08 ± 10%  perf-profile.children.cycles-pp.queued_write_lock_slowpath
>>>       0.16 ± 22%      -0.1        0.08 ± 24%  perf-profile.children.cycles-pp.btrfs_lookup_inode
>>>       0.21 ± 17%      -0.1        0.14 ± 19%  perf-profile.children.cycles-pp.__btrfs_update_delayed_inode
>>>       0.26 ± 12%      -0.1        0.18 ± 13%  perf-profile.children.cycles-pp.btrfs_async_run_delayed_root
>>>       0.52 ±  5%      -0.1        0.45        perf-profile.children.cycles-pp.set_extent_bit
>>>       0.45 ±  5%      -0.1        0.40 ±  3%  perf-profile.children.cycles-pp.alloc_extent_state
>>>       0.11 ± 17%      -0.1        0.06 ± 11%  perf-profile.children.cycles-pp.btrfs_clear_lock_blocking_rw
>>>       0.28 ±  9%      -0.0        0.23 ±  3%  perf-profile.children.cycles-pp.btrfs_drop_pages
>>>       0.07            -0.0        0.03 ±100%  perf-profile.children.cycles-pp.btrfs_set_lock_blocking_rw
>>>       0.39 ±  3%      -0.0        0.34 ±  3%  perf-profile.children.cycles-pp.get_alloc_profile
>>>       0.33 ±  7%      -0.0        0.29        perf-profile.children.cycles-pp.btrfs_set_extent_delalloc
>>>       0.38 ±  2%      -0.0        0.35 ±  4%  perf-profile.children.cycles-pp.__set_page_dirty_nobuffers
>>>       0.49 ±  3%      -0.0        0.46 ±  3%  perf-profile.children.cycles-pp.pagecache_get_page
>>>       0.18 ±  4%      -0.0        0.15 ±  2%  perf-profile.children.cycles-pp.truncate_inode_pages_range
>>>       0.08 ±  5%      -0.0        0.05 ±  9%  perf-profile.children.cycles-pp.btrfs_set_path_blocking
>>>       0.08 ±  6%      -0.0        0.06 ±  6%  perf-profile.children.cycles-pp.truncate_cleanup_page
>>>       0.80 ±  4%      +0.2        0.95 ±  2%  perf-profile.children.cycles-pp.can_overcommit
>>>      96.84            +0.5       97.37        perf-profile.children.cycles-pp.entry_SYSCALL_64_after_hwframe
>>>      96.80            +0.5       97.35        perf-profile.children.cycles-pp.do_syscall_64
>>>      43.34            +0.8       44.17        perf-profile.children.cycles-pp.btrfs_inode_rsv_release
>>>      43.49            +0.8       44.32        perf-profile.children.cycles-pp.block_rsv_release_bytes
>>>      95.32            +0.9       96.26        perf-profile.children.cycles-pp.ksys_write
>>>      95.26            +0.9       96.20        perf-profile.children.cycles-pp.vfs_write
>>>      94.91            +1.0       95.88        perf-profile.children.cycles-pp.__vfs_write
>>>      94.84            +1.0       95.81        perf-profile.children.cycles-pp.btrfs_file_write_iter
>>>      94.55            +1.0       95.55        perf-profile.children.cycles-pp.__btrfs_buffered_write
>>>      86.68            +1.0       87.70        perf-profile.children.cycles-pp.native_queued_spin_lock_slowpath
>>>      44.08            +1.2       45.31        perf-profile.children.cycles-pp.btrfs_delalloc_reserve_metadata
>>>      43.49            +1.3       44.77        perf-profile.children.cycles-pp.reserve_metadata_bytes
>>>      87.59            +1.8       89.38        perf-profile.children.cycles-pp._raw_spin_lock
>>>       0.47 ± 19%      -0.3        0.16 ± 13%  perf-profile.self.cycles-pp.intel_idle
>>>       0.33 ±  6%      -0.1        0.18 ±  6%  perf-profile.self.cycles-pp.get_alloc_profile
>>>       0.27 ±  8%      -0.0        0.22 ±  4%  perf-profile.self.cycles-pp.btrfs_drop_pages
>>>       0.07            -0.0        0.03 ±100%  perf-profile.self.cycles-pp.btrfs_set_lock_blocking_rw
>>>       0.14 ±  5%      -0.0        0.12 ±  6%  perf-profile.self.cycles-pp.clear_page_dirty_for_io
>>>       0.09 ±  5%      -0.0        0.07 ± 10%  perf-profile.self.cycles-pp._raw_spin_lock_irqsave
>>>       0.17 ±  4%      +0.1        0.23 ±  3%  perf-profile.self.cycles-pp.reserve_metadata_bytes
>>>       0.31 ±  7%      +0.1        0.45 ±  2%  perf-profile.self.cycles-pp.can_overcommit
>>>      86.35            +1.0       87.39        perf-profile.self.cycles-pp.native_queued_spin_lock_slowpath
>>>
>>>
>>>
>>>                                   aim7.jobs-per-min
>>>
>>>   29000 +-+-----------------------------------------------------------------+
>>>   28500 +-+               +..   +                           +..+..  +..     |
>>>         |..+    +.+..+.. :    .. +  .+.+..+..+.+.. .+..+.. +       +   +    |
>>>   28000 +-+ + ..         :   +    +.              +       +       +         |
>>>   27500 +-+  +          +                                                   |
>>>         |                                                                   |
>>>   27000 +-+                                                                 |
>>>   26500 +-+                                                                 |
>>>   26000 +-+                                                                 |
>>>         |                                                                   |
>>>   25500 +-+               O       O                               O O  O    |
>>>   25000 +-+                     O    O         O  O O  O  O O               O
>>>         |    O  O O     O    O         O                       O         O  |
>>>   24500 O-+O         O                    O  O                              |
>>>   24000 
>>> +-+-----------------------------------------------------------------
>>> +
>>>
>>>
>>> [*] bisect-good sample
>>> [O] bisect-bad  sample
>>>
>>>
>>> Disclaimer:
>>> Results have been estimated based on internal Intel analysis and are 
>>> provided for informational purposes only. Any difference in system 
>>> hardware or software design or configuration may affect actual performance.
>>>
>>>
>>> Thanks,
>>> Xiaolong
>>>
>>>
>>>
>>> _______________________________________________
>>> LKP mailing list
>>> LKP@...ts.01.org
>>> https://lists.01.org/mailman/listinfo/lkp
>>>


-- 
Jens Axboe

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ