lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <CAOQ4uxhoR8h0Mw=LJGSxMYWC1krjbwgzxbYNDjMmeo3sBAey1w@mail.gmail.com>
Date:   Tue, 11 May 2021 14:13:38 +0300
From:   Amir Goldstein <amir73il@...il.com>
To:     kernel test robot <oliver.sang@...el.com>
Cc:     Jan Kara <jack@...e.cz>, LKML <linux-kernel@...r.kernel.org>,
        lkp@...ts.01.org, kbuild test robot <lkp@...el.com>,
        ying.huang@...el.com, feng.tang@...el.com, zhengjun.xing@...el.com,
        Matthew Bobrowski <mbobrowski@...browski.org>
Subject: Re: [fanotify] 7cea2a3c50: stress-ng.fanotify.ops_per_sec -23.4% regression

On Tue, May 11, 2021 at 9:51 AM kernel test robot <oliver.sang@...el.com> wrote:
>
>
>
> Greeting,
>
> FYI, we noticed a -23.4% regression of stress-ng.fanotify.ops_per_sec due to commit:
>
>
> commit: 7cea2a3c505e87a9d6afc78be4a7f7be636a73a7 ("fanotify: support limited functionality for unprivileged users")
> https://git.kernel.org/cgit/linux/kernel/git/torvalds/linux.git master
>

I am not sure how this is possible.
This commit only changes permissions to perform the operation.
Is the test not being run as root?

Thanks,
Amir.

>
> in testcase: stress-ng
> on test machine: 96 threads 2 sockets Intel(R) Xeon(R) Gold 6252 CPU @ 2.10GHz with 192G memory
> with following parameters:
>
>         nr_threads: 10%
>         disk: 1HDD
>         testtime: 60s
>         fs: ext4
>         class: os
>         test: fanotify
>         cpufreq_governor: performance
>         ucode: 0x5003006
>
>
>
>
> If you fix the issue, kindly add following tag
> Reported-by: kernel test robot <oliver.sang@...el.com>
>
>
> Details are as below:
> -------------------------------------------------------------------------------------------------->
>
>
> To reproduce:
>
>         git clone https://github.com/intel/lkp-tests.git
>         cd lkp-tests
>         bin/lkp install                job.yaml  # job file is attached in this email
>         bin/lkp split-job --compatible job.yaml  # generate the yaml file for lkp run
>         bin/lkp run                    generated-yaml-file
>
> =========================================================================================
> class/compiler/cpufreq_governor/disk/fs/kconfig/nr_threads/rootfs/tbox_group/test/testcase/testtime/ucode:
>   os/gcc-9/performance/1HDD/ext4/x86_64-rhel-8.3/10%/debian-10.4-x86_64-20200603.cgz/lkp-csl-2sp5/fanotify/stress-ng/60s/0x5003006
>
> commit:
>   5b8fea65d1 ("fanotify: configurable limits via sysfs")
>   7cea2a3c50 ("fanotify: support limited functionality for unprivileged users")
>
> 5b8fea65d197f408 7cea2a3c505e87a9d6afc78be4a
> ---------------- ---------------------------
>          %stddev     %change         %stddev
>              \          |                \
>  1.214e+08 ą  3%     -23.4%   93048328 ą  4%  stress-ng.fanotify.ops
>    2023656 ą  3%     -23.4%    1550295 ą  4%  stress-ng.fanotify.ops_per_sec
>   33565053           -10.9%   29899893 ą  3%  stress-ng.time.file_system_outputs
>      19487 ą  3%     -19.0%      15787 ą  3%  stress-ng.time.involuntary_context_switches
>      10868            -1.0%      10755        stress-ng.time.minor_page_faults
>       1648            -1.9%       1618        stress-ng.time.percent_of_cpu_this_job_got
>       1014            -1.6%     998.53        stress-ng.time.system_time
>      38378 ą 97%    +280.5%     146045 ą  7%  stress-ng.time.voluntary_context_switches
>      18.19            -3.9%      17.48 ą  2%  iostat.cpu.system
>       0.96 ą  6%      -0.2        0.72 ą  8%  mpstat.cpu.all.soft%
>       0.21            -0.0        0.18 ą  6%  mpstat.cpu.all.usr%
>     191.47            -2.3%     187.06        turbostat.PkgWatt
>      91.25            -1.4%      89.97        turbostat.RAMWatt
>       3494 ą 39%     +55.6%       5438        vmstat.memory.buff
>       3805 ą 30%     +81.6%       6909 ą  3%  vmstat.system.cs
>     509.50           +23.4%     628.50 ą 18%  interrupts.CPU53.CAL:Function_call_interrupts
>       3119 ą 44%     -62.8%       1159 ą 38%  interrupts.CPU62.NMI:Non-maskable_interrupts
>       3119 ą 44%     -62.8%       1159 ą 38%  interrupts.CPU62.PMI:Performance_monitoring_interrupts
>       4580 ą106%    +357.3%      20942 ą 38%  numa-meminfo.node0.Dirty
>       4290 ą109%    +363.2%      19873 ą 36%  numa-meminfo.node0.Inactive(file)
>       5634 ą 74%    +364.1%      26149 ą 25%  numa-meminfo.node1.Dirty
>       4728 ą 75%    +406.7%      23958 ą 25%  numa-meminfo.node1.Inactive(file)
>      52712 ą 12%     +28.6%      67787 ą  8%  numa-meminfo.node1.KReclaimable
>      52712 ą 12%     +28.6%      67787 ą  8%  numa-meminfo.node1.SReclaimable
>      15433 ą  7%     +15.2%      17772 ą  2%  meminfo.Active
>       3658 ą 37%     +52.2%       5568        meminfo.Active(file)
>       3490 ą 39%     +55.6%       5431        meminfo.Buffers
>      10124 ą 75%    +366.6%      47241 ą 15%  meminfo.Dirty
>     314459 ą  2%     +12.9%     354928        meminfo.Inactive
>       8907 ą 77%    +393.4%      43949 ą 15%  meminfo.Inactive(file)
>     107422           +16.4%     124988 ą  2%  meminfo.KReclaimable
>     107422           +16.4%     124988 ą  2%  meminfo.SReclaimable
>       1110 ą111%    +368.9%       5205 ą 38%  numa-vmstat.node0.nr_dirty
>       1046 ą114%    +372.2%       4940 ą 36%  numa-vmstat.node0.nr_inactive_file
>       1048 ą113%    +371.4%       4941 ą 36%  numa-vmstat.node0.nr_zone_inactive_file
>       1138 ą110%    +360.2%       5241 ą 38%  numa-vmstat.node0.nr_zone_write_pending
>       1385 ą 78%    +368.3%       6486 ą 25%  numa-vmstat.node1.nr_dirty
>       1155 ą 81%    +413.4%       5933 ą 25%  numa-vmstat.node1.nr_inactive_file
>      13198 ą 12%     +28.1%      16909 ą  8%  numa-vmstat.node1.nr_slab_reclaimable
>       1154 ą 81%    +413.8%       5933 ą 25%  numa-vmstat.node1.nr_zone_inactive_file
>       1409 ą 78%    +362.8%       6523 ą 25%  numa-vmstat.node1.nr_zone_write_pending
>       8001 ą  4%     +18.4%       9476 ą  6%  softirqs.CPU3.SCHED
>       7965 ą  6%     +16.4%       9273 ą  7%  softirqs.CPU54.SCHED
>       7857 ą  6%     +18.4%       9306 ą  6%  softirqs.CPU58.SCHED
>       7971 ą  5%     +14.0%       9085 ą  5%  softirqs.CPU62.SCHED
>       7773 ą  5%     +16.7%       9072 ą  7%  softirqs.CPU64.SCHED
>       9103 ą 18%     -33.4%       6060 ą 26%  softirqs.CPU71.RCU
>       7936 ą  7%     +17.7%       9341 ą  7%  softirqs.CPU71.SCHED
>       8074 ą 35%     -50.5%       3997 ą 17%  softirqs.CPU77.RCU
>       7813 ą  6%     +15.7%       9036 ą  7%  softirqs.CPU8.SCHED
>     934447 ą  2%     -10.4%     836914 ą  2%  softirqs.RCU
>     914.67 ą 37%     +52.2%       1392        proc-vmstat.nr_active_file
>    4196621           -10.9%    3740233 ą  3%  proc-vmstat.nr_dirtied
>       2530 ą 75%    +364.4%      11749 ą 15%  proc-vmstat.nr_dirty
>     273393            +3.5%     282850        proc-vmstat.nr_file_pages
>       2239 ą 76%    +388.3%      10932 ą 15%  proc-vmstat.nr_inactive_file
>      26824           +16.3%      31185 ą  3%  proc-vmstat.nr_slab_reclaimable
>       1550 ą 27%    +249.9%       5424 ą 40%  proc-vmstat.nr_written
>     914.67 ą 37%     +52.2%       1392        proc-vmstat.nr_zone_active_file
>       2239 ą 76%    +388.3%      10932 ą 15%  proc-vmstat.nr_zone_inactive_file
>       2547 ą 77%    +363.4%      11805 ą 15%  proc-vmstat.nr_zone_write_pending
>    8695504           -15.1%    7378817 ą  3%  proc-vmstat.numa_hit
>    8608986           -15.3%    7292304 ą  3%  proc-vmstat.numa_local
>   13559136           -18.5%   11047582 ą  4%  proc-vmstat.pgalloc_normal
>   13417362           -18.7%   10904217 ą  4%  proc-vmstat.pgfree
>      27913 ą  8%     +37.7%      38440 ą 10%  slabinfo.buffer_head.active_objs
>     726.67 ą  7%     +37.7%       1000 ą  9%  slabinfo.buffer_head.active_slabs
>      28355 ą  7%     +37.7%      39043 ą  9%  slabinfo.buffer_head.num_objs
>     726.67 ą  7%     +37.7%       1000 ą  9%  slabinfo.buffer_head.num_slabs
>     107206           +11.6%     119642 ą  2%  slabinfo.dentry.active_objs
>       2593           +11.1%       2881 ą  2%  slabinfo.dentry.active_slabs
>     108962           +11.1%     121015 ą  2%  slabinfo.dentry.num_objs
>       2593           +11.1%       2881 ą  2%  slabinfo.dentry.num_slabs
>      66688 ą  9%     +31.8%      87926 ą  8%  slabinfo.ext4_extent_status.active_objs
>     655.00 ą  9%     +31.8%     863.17 ą  8%  slabinfo.ext4_extent_status.active_slabs
>      66862 ą  9%     +31.8%      88097 ą  8%  slabinfo.ext4_extent_status.num_objs
>     655.00 ą  9%     +31.8%     863.17 ą  8%  slabinfo.ext4_extent_status.num_slabs
>      11791 ą 11%     +79.3%      21139 ą  8%  slabinfo.ext4_inode_cache.active_objs
>     440.17 ą 11%     +85.2%     815.00 ą  9%  slabinfo.ext4_inode_cache.active_slabs
>      11889 ą 11%     +85.2%      22019 ą  9%  slabinfo.ext4_inode_cache.num_objs
>     440.17 ą 11%     +85.2%     815.00 ą  9%  slabinfo.ext4_inode_cache.num_slabs
>       1428 ą 25%     +41.9%       2027 ą  5%  slabinfo.jbd2_journal_head.active_objs
>       1428 ą 25%     +41.9%       2027 ą  5%  slabinfo.jbd2_journal_head.num_objs
>       6756 ą 15%     +41.1%       9536 ą  3%  slabinfo.kmalloc-rcl-512.active_objs
>       6792 ą 15%     +42.1%       9649 ą  3%  slabinfo.kmalloc-rcl-512.num_objs
>      72015 ą  5%     +15.2%      82951 ą  4%  slabinfo.vmap_area.active_objs
>       1129 ą  5%     +16.2%       1312 ą  4%  slabinfo.vmap_area.active_slabs
>      72353 ą  5%     +16.2%      84046 ą  4%  slabinfo.vmap_area.num_objs
>       1129 ą  5%     +16.2%       1312 ą  4%  slabinfo.vmap_area.num_slabs
>  4.537e+09           -13.0%  3.948e+09 ą  4%  perf-stat.i.branch-instructions
>   34409487           -12.1%   30235678 ą  2%  perf-stat.i.branch-misses
>      34.06 ą  2%      -2.4       31.67 ą  2%  perf-stat.i.cache-miss-rate%
>   85290547 ą  3%     -12.7%   74474324 ą  5%  perf-stat.i.cache-misses
>       3674 ą 32%     +88.6%       6931 ą  3%  perf-stat.i.context-switches
>       2.35           +12.2%       2.64 ą  4%  perf-stat.i.cpi
>  5.242e+10            -3.5%  5.057e+10        perf-stat.i.cpu-cycles
>     224.46            -9.4%     203.26 ą  2%  perf-stat.i.cpu-migrations
>     656.95 ą  2%     +12.4%     738.24 ą  5%  perf-stat.i.cycles-between-cache-misses
>     300745 ą 11%     +53.4%     461361 ą 19%  perf-stat.i.dTLB-load-misses
>  6.438e+09           -13.9%  5.543e+09 ą  4%  perf-stat.i.dTLB-loads
>     161739 ą  7%     +14.6%     185423 ą  6%  perf-stat.i.dTLB-store-misses
>  3.536e+09           -15.4%  2.991e+09 ą  4%  perf-stat.i.dTLB-stores
>      88.81            -1.9       86.94        perf-stat.i.iTLB-load-miss-rate%
>   20384992 ą  2%     -14.5%   17430423 ą  4%  perf-stat.i.iTLB-load-misses
>  2.218e+10           -13.3%  1.923e+10 ą  4%  perf-stat.i.instructions
>       0.43            -9.5%       0.39 ą  3%  perf-stat.i.ipc
>       0.55            -3.5%       0.53        perf-stat.i.metric.GHz
>     612.45 ą  2%     -11.7%     540.91        perf-stat.i.metric.K/sec
>     153.67           -13.9%     132.37 ą  4%  perf-stat.i.metric.M/sec
>      88.26            -1.5       86.78        perf-stat.i.node-load-miss-rate%
>   31264969 ą  3%     -14.5%   26719326 ą  3%  perf-stat.i.node-load-misses
>      82.85            -2.5       80.30        perf-stat.i.node-store-miss-rate%
>   16368326 ą  3%     -16.1%   13725763 ą  2%  perf-stat.i.node-store-misses
>      11.08 ą  2%      +7.9%      11.95 ą  5%  perf-stat.overall.MPKI
>      34.71 ą  2%      -2.3       32.43 ą  2%  perf-stat.overall.cache-miss-rate%
>       2.36           +11.4%       2.63 ą  2%  perf-stat.overall.cpi
>       0.00 ą 12%      +0.0        0.01 ą 25%  perf-stat.overall.dTLB-load-miss-rate%
>       0.00 ą  8%      +0.0        0.01 ą  9%  perf-stat.overall.dTLB-store-miss-rate%
>      90.08            -1.6       88.49        perf-stat.overall.iTLB-load-miss-rate%
>       0.42           -10.2%       0.38 ą  2%  perf-stat.overall.ipc
>      84.44            -2.2       82.27        perf-stat.overall.node-store-miss-rate%
>  4.465e+09           -13.0%  3.886e+09 ą  4%  perf-stat.ps.branch-instructions
>   33856926           -12.1%   29757385 ą  2%  perf-stat.ps.branch-misses
>   83927962 ą  3%     -12.7%   73296150 ą  5%  perf-stat.ps.cache-misses
>       3620 ą 31%     +88.4%       6819 ą  4%  perf-stat.ps.context-switches
>  5.158e+10            -3.5%  4.977e+10        perf-stat.ps.cpu-cycles
>     221.03            -9.5%     200.02 ą  2%  perf-stat.ps.cpu-migrations
>     295389 ą 11%     +53.7%     454026 ą 19%  perf-stat.ps.dTLB-load-misses
>  6.336e+09           -13.9%  5.456e+09 ą  4%  perf-stat.ps.dTLB-loads
>     159067 ą  7%     +14.7%     182479 ą  6%  perf-stat.ps.dTLB-store-misses
>   3.48e+09           -15.4%  2.944e+09 ą  4%  perf-stat.ps.dTLB-stores
>   20061765 ą  2%     -14.5%   17156375 ą  4%  perf-stat.ps.iTLB-load-misses
>  2.183e+10           -13.3%  1.893e+10 ą  4%  perf-stat.ps.instructions
>   30762758 ą  3%     -14.5%   26295053 ą  3%  perf-stat.ps.node-load-misses
>   16104244 ą  3%     -16.1%   13507363 ą  2%  perf-stat.ps.node-store-misses
>  1.377e+12 ą  2%     -12.2%   1.21e+12 ą  2%  perf-stat.total.instructions
>       2.87 ą 11%      -0.7        2.18 ą 11%  perf-profile.calltrace.cycles-pp.ret_from_fork
>       2.87 ą 11%      -0.7        2.18 ą 11%  perf-profile.calltrace.cycles-pp.kthread.ret_from_fork
>       2.85 ą 11%      -0.7        2.16 ą 11%  perf-profile.calltrace.cycles-pp.smpboot_thread_fn.kthread.ret_from_fork
>       2.84 ą 11%      -0.7        2.15 ą 11%  perf-profile.calltrace.cycles-pp.__softirqentry_text_start.run_ksoftirqd.smpboot_thread_fn.kthread.ret_from_fork
>       2.84 ą 11%      -0.7        2.15 ą 11%  perf-profile.calltrace.cycles-pp.rcu_core.__softirqentry_text_start.run_ksoftirqd.smpboot_thread_fn.kthread
>       2.84 ą 11%      -0.7        2.15 ą 11%  perf-profile.calltrace.cycles-pp.run_ksoftirqd.smpboot_thread_fn.kthread.ret_from_fork
>       2.83 ą 11%      -0.7        2.15 ą 11%  perf-profile.calltrace.cycles-pp.rcu_do_batch.rcu_core.__softirqentry_text_start.run_ksoftirqd.smpboot_thread_fn
>       2.50 ą 10%      -0.6        1.88 ą 11%  perf-profile.calltrace.cycles-pp.kmem_cache_free.rcu_do_batch.rcu_core.__softirqentry_text_start.run_ksoftirqd
>       2.26 ą 10%      -0.6        1.69 ą 14%  perf-profile.calltrace.cycles-pp.kmem_cache_free.fanotify_read.vfs_read.ksys_read.do_syscall_64
>       1.79 ą 12%      -0.5        1.30 ą 14%  perf-profile.calltrace.cycles-pp.kmem_cache_alloc.__alloc_file.alloc_empty_file.dentry_open.copy_event_to_user
>       1.74 ą  9%      -0.5        1.26 ą 13%  perf-profile.calltrace.cycles-pp.fanotify_free_event.fanotify_read.vfs_read.ksys_read.do_syscall_64
>       1.58 ą 12%      -0.4        1.18 ą 11%  perf-profile.calltrace.cycles-pp._raw_spin_lock.fanotify_read.vfs_read.ksys_read.do_syscall_64
>       1.43 ą  8%      -0.4        1.03 ą 29%  perf-profile.calltrace.cycles-pp.lockref_get.fanotify_alloc_event.fanotify_handle_event.fsnotify.__fsnotify_parent
>       2.19 ą  7%      -0.4        1.83 ą  6%  perf-profile.calltrace.cycles-pp.fanotify_alloc_event.fanotify_handle_event.fsnotify.__fsnotify_parent.__fput
>       0.71 ą 11%      -0.3        0.37 ą 70%  perf-profile.calltrace.cycles-pp.put_pid.fanotify_free_event.fanotify_read.vfs_read.ksys_read
>       2.14 ą  7%      -0.3        1.80 ą  7%  perf-profile.calltrace.cycles-pp.kmem_cache_alloc.fanotify_alloc_event.fanotify_handle_event.fsnotify.__fsnotify_parent
>       1.17 ą 10%      -0.3        0.85 ą 12%  perf-profile.calltrace.cycles-pp.ext4_file_open.do_dentry_open.dentry_open.copy_event_to_user.fanotify_read
>       2.37 ą  8%      -0.3        2.07 ą  6%  perf-profile.calltrace.cycles-pp.fanotify_alloc_event.fanotify_handle_event.fsnotify.__fsnotify_parent.do_sys_openat2
>       0.98 ą  9%      -0.3        0.72 ą 14%  perf-profile.calltrace.cycles-pp.path_put.fanotify_free_event.fanotify_read.vfs_read.ksys_read
>       0.96 ą  9%      -0.3        0.71 ą 14%  perf-profile.calltrace.cycles-pp.dput.path_put.fanotify_free_event.fanotify_read.vfs_read
>       0.90 ą  9%      -0.3        0.65 ą 13%  perf-profile.calltrace.cycles-pp.fscrypt_file_open.ext4_file_open.do_dentry_open.dentry_open.copy_event_to_user
>       0.84 ą  8%      -0.2        0.60 ą 14%  perf-profile.calltrace.cycles-pp.lockref_put_return.dput.path_put.fanotify_free_event.fanotify_read
>       0.80 ą  7%      -0.2        0.60 ą  9%  perf-profile.calltrace.cycles-pp.refill_obj_stock.kmem_cache_free.rcu_do_batch.rcu_core.__softirqentry_text_start
>       0.87 ą 15%      -0.2        0.68 ą 10%  perf-profile.calltrace.cycles-pp.native_queued_spin_lock_slowpath._raw_spin_lock.fanotify_read.vfs_read.ksys_read
>       0.89 ą  7%      -0.1        0.76 ą  8%  perf-profile.calltrace.cycles-pp.fanotify_alloc_event.fanotify_handle_event.fsnotify.__fsnotify_parent.vfs_write
>       0.72 ą  9%      -0.1        0.60 ą  5%  perf-profile.calltrace.cycles-pp._raw_spin_lock_irqsave.__wake_up_common_lock.fsnotify_add_event.fanotify_handle_event.fsnotify
>       0.75 ą  6%      -0.1        0.63 ą  8%  perf-profile.calltrace.cycles-pp.fanotify_alloc_event.fanotify_handle_event.fsnotify.__fsnotify_parent.vfs_read
>       0.11 ą223%      +0.7        0.83 ą 24%  perf-profile.calltrace.cycles-pp.fanotify_merge.fsnotify_add_event.fanotify_handle_event.fsnotify.__fsnotify_parent
>       4.92 ą 15%      +1.7        6.61 ą 16%  perf-profile.calltrace.cycles-pp.fsnotify_add_event.fanotify_handle_event.fsnotify.__fsnotify_parent.do_sys_openat2
>       5.69 ą 22%      +3.1        8.80 ą 23%  perf-profile.calltrace.cycles-pp._raw_spin_lock.fsnotify_add_event.fanotify_handle_event.fsnotify.__fsnotify_parent
>       4.00 ą 37%      +3.6        7.60 ą 26%  perf-profile.calltrace.cycles-pp.native_queued_spin_lock_slowpath._raw_spin_lock.fsnotify_add_event.fanotify_handle_event.fsnotify
>       0.00            +4.0        4.02 ą 12%  perf-profile.calltrace.cycles-pp.apparmor_capable.security_capable.ns_capable_common.copy_event_to_user.fanotify_read
>       0.00            +4.1        4.07 ą 12%  perf-profile.calltrace.cycles-pp.security_capable.ns_capable_common.copy_event_to_user.fanotify_read.vfs_read
>       0.00            +4.1        4.14 ą 12%  perf-profile.calltrace.cycles-pp.ns_capable_common.copy_event_to_user.fanotify_read.vfs_read.ksys_read
>       7.67 ą  9%      -1.7        5.97 ą 11%  perf-profile.children.cycles-pp.kmem_cache_free
>       5.74 ą  7%      -1.2        4.59 ą  9%  perf-profile.children.cycles-pp.kmem_cache_alloc
>       4.27 ą 11%      -0.9        3.36 ą 11%  perf-profile.children.cycles-pp.rcu_core
>       6.23 ą  7%      -0.9        5.32 ą  7%  perf-profile.children.cycles-pp.fanotify_alloc_event
>       4.26 ą 11%      -0.9        3.35 ą 11%  perf-profile.children.cycles-pp.rcu_do_batch
>       4.38 ą 10%      -0.9        3.48 ą 11%  perf-profile.children.cycles-pp.__softirqentry_text_start
>       1.09 ą 13%      -0.7        0.35 ą 10%  perf-profile.children.cycles-pp.alloc_fd
>       2.87 ą 11%      -0.7        2.18 ą 11%  perf-profile.children.cycles-pp.ret_from_fork
>       2.87 ą 11%      -0.7        2.18 ą 11%  perf-profile.children.cycles-pp.kthread
>       2.85 ą 11%      -0.7        2.16 ą 11%  perf-profile.children.cycles-pp.smpboot_thread_fn
>       2.84 ą 11%      -0.7        2.15 ą 11%  perf-profile.children.cycles-pp.run_ksoftirqd
>       1.94 ą  8%      -0.5        1.42 ą 10%  perf-profile.children.cycles-pp.get_obj_cgroup_from_current
>       2.52 ą  9%      -0.5        2.01 ą  6%  perf-profile.children.cycles-pp._raw_spin_lock_irqsave
>       3.38 ą  9%      -0.5        2.88 ą  7%  perf-profile.children.cycles-pp.__wake_up_common_lock
>       1.64 ą  8%      -0.4        1.28 ą 10%  perf-profile.children.cycles-pp.refill_obj_stock
>       1.30 ą 11%      -0.3        0.98 ą 12%  perf-profile.children.cycles-pp.ext4_file_open
>       1.53 ą  8%      -0.3        1.23 ą 11%  perf-profile.children.cycles-pp.path_put
>       0.97 ą  9%      -0.3        0.72 ą 12%  perf-profile.children.cycles-pp.fscrypt_file_open
>       0.92 ą  7%      -0.2        0.71 ą 10%  perf-profile.children.cycles-pp.drain_obj_stock
>       0.58 ą  9%      -0.2        0.39 ą 10%  perf-profile.children.cycles-pp.dget_parent
>       0.55 ą  9%      -0.2        0.36 ą  9%  perf-profile.children.cycles-pp.lockref_get_not_zero
>       0.34 ą 25%      -0.1        0.22 ą 12%  perf-profile.children.cycles-pp.pid_vnr
>       0.20 ą 15%      -0.1        0.14 ą 15%  perf-profile.children.cycles-pp.fsnotify_peek_first_event
>       0.25 ą 13%      -0.1        0.20 ą 15%  perf-profile.children.cycles-pp.call_rcu
>       0.26 ą  9%      -0.1        0.20 ą  9%  perf-profile.children.cycles-pp.page_counter_try_charge
>       0.22 ą 12%      -0.1        0.17 ą 15%  perf-profile.children.cycles-pp.kmem_cache_alloc_trace
>       0.24 ą 12%      -0.1        0.19 ą  8%  perf-profile.children.cycles-pp.fanotify_insert_event
>       0.19 ą 10%      -0.1        0.13 ą 17%  perf-profile.children.cycles-pp.close_fd
>       0.23 ą  8%      -0.0        0.19 ą 13%  perf-profile.children.cycles-pp.syscall_return_via_sysret
>       0.15 ą 10%      -0.0        0.11 ą 18%  perf-profile.children.cycles-pp.pick_file
>       0.13 ą 14%      -0.0        0.10 ą 18%  perf-profile.children.cycles-pp.propagate_protected_usage
>       0.08 ą 10%      -0.0        0.05 ą 46%  perf-profile.children.cycles-pp.fanotify_free_group_priv
>       0.16 ą  9%      -0.0        0.13 ą  9%  perf-profile.children.cycles-pp.rcu_all_qs
>       0.09 ą 13%      -0.0        0.06 ą 11%  perf-profile.children.cycles-pp.__fsnotify_alloc_group
>       0.07 ą 11%      -0.0        0.04 ą 45%  perf-profile.children.cycles-pp.alloc_page_buffers
>       0.11 ą 11%      -0.0        0.09 ą  7%  perf-profile.children.cycles-pp.try_to_free_buffers
>       0.10 ą  8%      -0.0        0.08 ą 12%  perf-profile.children.cycles-pp.create_empty_buffers
>       0.07 ą  7%      -0.0        0.04 ą 45%  perf-profile.children.cycles-pp.alloc_buffer_head
>       0.08 ą 14%      -0.0        0.06 ą 11%  perf-profile.children.cycles-pp.__srcu_read_lock
>       0.09 ą 11%      -0.0        0.07 ą 17%  perf-profile.children.cycles-pp.rcu_segcblist_enqueue
>       0.19 ą 60%      +0.3        0.48 ą 30%  perf-profile.children.cycles-pp.__x64_sys_ioctl
>       0.19 ą 61%      +0.3        0.48 ą 30%  perf-profile.children.cycles-pp.do_vfs_ioctl
>       0.19 ą 61%      +0.3        0.48 ą 30%  perf-profile.children.cycles-pp.fanotify_ioctl
>       0.81 ą 17%      +0.5        1.28 ą 15%  perf-profile.children.cycles-pp.fanotify_merge
>       5.66 ą 22%      +3.0        8.69 ą 23%  perf-profile.children.cycles-pp.native_queued_spin_lock_slowpath
>       0.23 ą  9%      +4.0        4.21 ą 12%  perf-profile.children.cycles-pp.apparmor_capable
>       0.24 ą  9%      +4.0        4.26 ą 12%  perf-profile.children.cycles-pp.security_capable
>       0.24 ą  8%      +4.1        4.33 ą 11%  perf-profile.children.cycles-pp.ns_capable_common
>       4.85 ą  9%      -1.1        3.72 ą 11%  perf-profile.self.cycles-pp.kmem_cache_free
>       3.49 ą  9%      -1.1        2.43 ą  9%  perf-profile.self.cycles-pp._raw_spin_lock
>       1.81 ą  8%      -0.5        1.32 ą  9%  perf-profile.self.cycles-pp.get_obj_cgroup_from_current
>       2.04 ą  6%      -0.5        1.57 ą 11%  perf-profile.self.cycles-pp.kmem_cache_alloc
>       2.24 ą  9%      -0.5        1.78 ą  6%  perf-profile.self.cycles-pp._raw_spin_lock_irqsave
>       0.97 ą  9%      -0.2        0.73 ą 11%  perf-profile.self.cycles-pp.do_dentry_open
>       0.51 ą  9%      -0.2        0.33 ą 10%  perf-profile.self.cycles-pp.lockref_get_not_zero
>       0.68 ą  9%      -0.1        0.54 ą  9%  perf-profile.self.cycles-pp.refill_obj_stock
>       0.84 ą  9%      -0.1        0.70 ą  7%  perf-profile.self.cycles-pp.__wake_up_common
>       0.55 ą  6%      -0.1        0.41 ą 11%  perf-profile.self.cycles-pp.drain_obj_stock
>       0.33 ą 25%      -0.1        0.21 ą 12%  perf-profile.self.cycles-pp.pid_vnr
>       0.30 ą  9%      -0.1        0.23 ą 11%  perf-profile.self.cycles-pp.dput
>       0.33 ą 11%      -0.1        0.26 ą 12%  perf-profile.self.cycles-pp.page_counter_cancel
>       0.19 ą 15%      -0.1        0.13 ą 12%  perf-profile.self.cycles-pp.fsnotify_peek_first_event
>       0.24 ą 11%      -0.1        0.19 ą  8%  perf-profile.self.cycles-pp.fanotify_insert_event
>       0.13 ą 14%      -0.1        0.08 ą 16%  perf-profile.self.cycles-pp.alloc_fd
>       0.23 ą  8%      -0.0        0.19 ą 13%  perf-profile.self.cycles-pp.syscall_return_via_sysret
>       0.10 ą 15%      -0.0        0.07 ą  7%  perf-profile.self.cycles-pp.fanotify_read
>       0.14 ą  9%      -0.0        0.11 ą 12%  perf-profile.self.cycles-pp.exit_to_user_mode_prepare
>       0.09 ą 11%      -0.0        0.07 ą 17%  perf-profile.self.cycles-pp.rcu_segcblist_enqueue
>       0.12 ą 16%      +0.1        0.18 ą 16%  perf-profile.self.cycles-pp.___slab_alloc
>       0.00            +0.1        0.07 ą 14%  perf-profile.self.cycles-pp.ns_capable_common
>       0.00            +0.2        0.18 ą 24%  perf-profile.self.cycles-pp.try_to_wake_up
>       0.17 ą 63%      +0.3        0.46 ą 30%  perf-profile.self.cycles-pp.fanotify_ioctl
>       0.79 ą 17%      +0.5        1.27 ą 15%  perf-profile.self.cycles-pp.fanotify_merge
>       5.56 ą 22%      +3.0        8.59 ą 23%  perf-profile.self.cycles-pp.native_queued_spin_lock_slowpath
>       0.22 ą  9%      +3.9        4.08 ą 11%  perf-profile.self.cycles-pp.apparmor_capable
>
>
>
>                             stress-ng.fanotify.ops_per_sec
>
>   2.2e+06 +-----------------------------------------------------------------+
>           | +.    : +         .+.      .+.+.+.+.+.+..+.                     |
>   2.1e+06 |-+ +   :  +.     .+   +..+.+                +.+     +.+.+..+     |
>     2e+06 |-+  + :     +.+.+                              +   +             |
>           |     +                                          +.+              |
>   1.9e+06 |-+                                                               |
>   1.8e+06 |-+                                                               |
>           |                                                                 |
>   1.7e+06 |-+                                                               |
>   1.6e+06 |-+                                                         O     |
>           |                             O                          O      O |
>   1.5e+06 |-+              O   O    O O                O     O          O   |
>   1.4e+06 |-+        O O O   O   O        O O O   O  O   O O   O O          |
>           | O O O  O                            O                           |
>   1.3e+06 +-----------------------------------------------------------------+
>
>
> [*] bisect-good sample
> [O] bisect-bad  sample
>
>
>
> Disclaimer:
> Results have been estimated based on internal Intel analysis and are provided
> for informational purposes only. Any difference in system hardware or software
> design or configuration may affect actual performance.
>
>
> ---
> 0DAY/LKP+ Test Infrastructure                   Open Source Technology Center
> https://lists.01.org/hyperkitty/list/lkp@lists.01.org       Intel Corporation
>
> Thanks,
> Oliver Sang
>

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ