[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <42194ba4-cc3b-c8b3-06e6-e2938df3a87c@intel.com>
Date: Sat, 28 Jan 2023 09:37:17 +0800
From: "Yin, Fengwei" <fengwei.yin@...el.com>
To: kernel test robot <oliver.sang@...el.com>,
John Johansen <john.johansen@...onical.com>
CC: <oe-lkp@...ts.linux.dev>, <lkp@...el.com>,
<linux-kernel@...r.kernel.org>, <apparmor@...ts.ubuntu.com>,
<ying.huang@...el.com>, <feng.tang@...el.com>,
<zhengjun.xing@...ux.intel.com>
Subject: Re: [linus:master] [apparmor] 1ad22fcc4d: stress-ng.kill.ops_per_sec
-42.5% regression
Hi John,
On 12/31/2022 3:18 PM, kernel test robot wrote:
>
> Greeting,
>
> FYI, we noticed a -42.5% regression of stress-ng.kill.ops_per_sec due to commit:
>
>
> commit: 1ad22fcc4d0d2fb2e0f35aed555a86d016d5e590 ("apparmor: rework profile->rules to be a list")
> https://git.kernel.org/cgit/linux/kernel/git/torvalds/linux.git master
>
> in testcase: stress-ng
> on test machine: 96 threads 2 sockets Intel(R) Xeon(R) Gold 6252 CPU @ 2.10GHz (Cascade Lake) with 512G memory
> with following parameters:
>
> nr_threads: 10%
> disk: 1HDD
> testtime: 60s
> fs: ext4
> class: os
> test: kill
> cpufreq_governor: performance
Do you think any other information need be collected for this regression
report? Thanks.
Regards
Yin, Fengwei
>
>
>
>
> If you fix the issue, kindly add following tag
> | Reported-by: kernel test robot <oliver.sang@...el.com>
> | Link: https://lore.kernel.org/oe-lkp/202212311546.755a3ed7-oliver.sang@intel.com
>
>
> Details are as below:
> -------------------------------------------------------------------------------------------------->
>
>
> To reproduce:
>
> git clone https://github.com/intel/lkp-tests.git
> cd lkp-tests
> sudo bin/lkp install job.yaml # job file is attached in this email
> bin/lkp split-job --compatible job.yaml # generate the yaml file for lkp run
> sudo bin/lkp run generated-yaml-file
>
> # if come across any failure that blocks the test,
> # please remove ~/.lkp and /lkp dir to run from a clean state.
>
> =========================================================================================
> class/compiler/cpufreq_governor/disk/fs/kconfig/nr_threads/rootfs/tbox_group/test/testcase/testtime:
> os/gcc-11/performance/1HDD/ext4/x86_64-rhel-8.3/10%/debian-11.1-x86_64-20220510.cgz/lkp-csl-2sp7/kill/stress-ng/60s
>
> commit:
> 217af7e2f4 ("apparmor: refactor profile rules and attachments")
> 1ad22fcc4d ("apparmor: rework profile->rules to be a list")
>
> 217af7e2f4deb629 1ad22fcc4d0d2fb2e0f35aed555
> ---------------- ---------------------------
> %stddev %change %stddev
> \ | \
> 2820 ± 2% -42.7% 1616 stress-ng.kill.kill_calls_per_sec
> 511090 -42.5% 293660 stress-ng.kill.ops
> 8518 -42.5% 4894 stress-ng.kill.ops_per_sec
> 3778 ± 2% -58.4% 1571 ± 3% stress-ng.time.involuntary_context_switches
> 859945 ± 3% -29.7% 604857 ± 2% stress-ng.time.voluntary_context_switches
> 70351 ± 12% +22.0% 85795 ± 8% meminfo.AnonHugePages
> 0.05 ± 7% -22.6% 0.04 turbostat.IPC
> 28627 ± 3% -27.0% 20903 ± 2% vmstat.system.cs
> 128187 ± 92% +78.6% 228921 ± 49% sched_debug.cfs_rq:/.min_vruntime.max
> 20901 ±107% +124.7% 46965 ± 59% sched_debug.cfs_rq:/.min_vruntime.stddev
> -24201 -149.5% 11969 ± 74% sched_debug.cfs_rq:/.spread0.avg
> 69818 ±123% +170.9% 189126 ± 59% sched_debug.cfs_rq:/.spread0.max
> 20904 ±107% +124.7% 46965 ± 59% sched_debug.cfs_rq:/.spread0.stddev
> 1.19 ± 8% +6.3 7.48 ± 10% perf-profile.calltrace.cycles-pp.aa_may_signal.apparmor_task_kill.security_task_kill.kill_something_info.__x64_sys_kill
> 0.58 ± 8% -0.2 0.40 ± 13% perf-profile.children.cycles-pp.check_kill_permission
> 0.30 ± 12% -0.1 0.16 ± 8% perf-profile.children.cycles-pp.profile_signal_perm
> 0.20 ± 10% -0.1 0.12 ± 13% perf-profile.children.cycles-pp.audit_signal_info_syscall
> 0.21 ± 23% -0.1 0.14 ± 12% perf-profile.children.cycles-pp.pause
> 0.18 ± 16% -0.1 0.11 ± 20% perf-profile.children.cycles-pp.__task_pid_nr_ns
> 0.15 ± 10% -0.1 0.09 ± 14% perf-profile.children.cycles-pp.__kill_pgrp_info
> 0.16 ± 13% -0.1 0.10 ± 14% perf-profile.children.cycles-pp.exit_to_user_mode_loop
> 0.16 ± 15% -0.1 0.11 ± 12% perf-profile.children.cycles-pp.exit_to_user_mode_prepare
> 0.10 ± 9% -0.0 0.06 ± 11% perf-profile.children.cycles-pp.audit_signal_info
> 0.07 ± 18% -0.0 0.04 ± 71% perf-profile.children.cycles-pp.alloc_empty_file
> 0.07 ± 18% -0.0 0.04 ± 71% perf-profile.children.cycles-pp.__alloc_file
> 0.11 ± 7% -0.0 0.08 ± 16% perf-profile.children.cycles-pp.do_send_sig_info
> 0.10 ± 12% -0.0 0.06 ± 14% perf-profile.children.cycles-pp.__send_signal_locked
> 0.10 ± 17% -0.0 0.06 ± 19% perf-profile.children.cycles-pp.arch_do_signal_or_restart
> 0.06 ± 14% +0.0 0.11 ± 18% perf-profile.children.cycles-pp.ns_capable
> 0.06 ± 14% +0.0 0.11 ± 18% perf-profile.children.cycles-pp.security_capable
> 0.06 ± 14% +0.0 0.11 ± 18% perf-profile.children.cycles-pp.apparmor_capable
> 0.09 ± 11% +0.0 0.14 ± 23% perf-profile.children.cycles-pp.kernel_clone
> 0.09 ± 8% +0.0 0.14 ± 24% perf-profile.children.cycles-pp.copy_process
> 0.04 ± 71% +0.0 0.08 ± 18% perf-profile.children.cycles-pp.__do_sys_clone
> 0.07 ± 18% +0.1 0.16 ± 9% perf-profile.children.cycles-pp.__x64_sys_exit_group
> 0.07 ± 18% +0.1 0.16 ± 9% perf-profile.children.cycles-pp.do_group_exit
> 0.07 ± 18% +0.1 0.16 ± 9% perf-profile.children.cycles-pp.do_exit
> 0.06 ± 47% +0.1 0.15 ± 8% perf-profile.children.cycles-pp.exit_notify
> 0.29 ± 12% +0.1 0.38 ± 11% perf-profile.children.cycles-pp.queued_read_lock_slowpath
> 0.15 ± 10% +0.1 0.28 ± 13% perf-profile.children.cycles-pp.queued_write_lock_slowpath
> 0.36 ± 10% +0.1 0.50 ± 10% perf-profile.children.cycles-pp.native_queued_spin_lock_slowpath
> 1.24 ± 8% +6.3 7.56 ± 10% perf-profile.children.cycles-pp.aa_may_signal
> 0.28 ± 12% -0.1 0.15 ± 10% perf-profile.self.cycles-pp.profile_signal_perm
> 0.23 ± 16% -0.1 0.12 ± 17% perf-profile.self.cycles-pp.check_kill_permission
> 0.20 ± 10% -0.1 0.12 ± 13% perf-profile.self.cycles-pp.audit_signal_info_syscall
> 0.15 ± 11% -0.1 0.08 ± 13% perf-profile.self.cycles-pp.kill_something_info
> 0.16 ± 6% -0.1 0.10 ± 20% perf-profile.self.cycles-pp.security_task_kill
> 0.14 ± 11% -0.1 0.09 ± 14% perf-profile.self.cycles-pp.__kill_pgrp_info
> 0.16 ± 15% -0.1 0.10 ± 22% perf-profile.self.cycles-pp.__task_pid_nr_ns
> 0.08 ± 13% -0.1 0.04 ± 71% perf-profile.self.cycles-pp.audit_signal_info
> 0.06 ± 14% +0.0 0.11 ± 16% perf-profile.self.cycles-pp.apparmor_capable
> 0.36 ± 10% +0.1 0.50 ± 10% perf-profile.self.cycles-pp.native_queued_spin_lock_slowpath
> 0.96 ± 8% +6.4 7.38 ± 10% perf-profile.self.cycles-pp.aa_may_signal
> 16.17 +18.8% 19.21 perf-stat.i.MPKI
> 8.529e+08 -30.2% 5.952e+08 perf-stat.i.branch-instructions
> 0.68 ± 3% +0.1 0.79 ± 2% perf-stat.i.branch-miss-rate%
> 7256065 ± 3% -7.7% 6698864 ± 2% perf-stat.i.branch-misses
> 13.08 ± 3% +6.0 19.07 perf-stat.i.cache-miss-rate%
> 9361692 ± 2% +16.3% 10884924 ± 2% perf-stat.i.cache-misses
> 71450858 -20.7% 56687571 ± 2% perf-stat.i.cache-references
> 29596 ± 3% -27.5% 21452 ± 2% perf-stat.i.context-switches
> 6.21 +50.4% 9.34 perf-stat.i.cpi
> 2974 ± 2% -14.4% 2547 ± 2% perf-stat.i.cycles-between-cache-misses
> 0.02 ± 13% +0.0 0.03 ± 6% perf-stat.i.dTLB-load-miss-rate%
> 1.408e+09 -32.2% 9.544e+08 perf-stat.i.dTLB-loads
> 7.763e+08 -33.2% 5.183e+08 perf-stat.i.dTLB-stores
> 34.02 -3.0 31.06 perf-stat.i.iTLB-load-miss-rate%
> 1109352 ± 2% -18.6% 903511 perf-stat.i.iTLB-load-misses
> 2150307 -6.8% 2003881 perf-stat.i.iTLB-loads
> 4.552e+09 -30.7% 3.153e+09 perf-stat.i.instructions
> 4134 ± 2% -15.1% 3508 perf-stat.i.instructions-per-iTLB-miss
> 0.19 -26.4% 0.14 perf-stat.i.ipc
> 814.65 -16.5% 680.21 perf-stat.i.metric.K/sec
> 31.63 -31.9% 21.53 perf-stat.i.metric.M/sec
> 80.72 ± 3% +9.8 90.56 perf-stat.i.node-load-miss-rate%
> 325431 ± 5% +970.6% 3483935 ± 2% perf-stat.i.node-load-misses
> 79936 ± 19% +323.7% 338676 ± 6% perf-stat.i.node-loads
> 4260901 ± 2% -34.5% 2792405 perf-stat.i.node-store-misses
> 15.70 +14.6% 18.00 perf-stat.overall.MPKI
> 0.85 ± 2% +0.3 1.12 ± 2% perf-stat.overall.branch-miss-rate%
> 13.11 ± 2% +6.1 19.20 perf-stat.overall.cache-miss-rate%
> 6.03 +44.8% 8.73 perf-stat.overall.cpi
> 2931 -13.8% 2527 ± 2% perf-stat.overall.cycles-between-cache-misses
> 0.02 ± 10% +0.0 0.03 ± 7% perf-stat.overall.dTLB-load-miss-rate%
> 34.03 -3.0 31.07 perf-stat.overall.iTLB-load-miss-rate%
> 4104 ± 2% -15.0% 3487 perf-stat.overall.instructions-per-iTLB-miss
> 0.17 -30.9% 0.11 perf-stat.overall.ipc
> 80.44 ± 3% +10.7 91.14 perf-stat.overall.node-load-miss-rate%
> 8.39e+08 -30.3% 5.852e+08 perf-stat.ps.branch-instructions
> 7111005 ± 2% -7.8% 6555177 ± 2% perf-stat.ps.branch-misses
> 9215998 ± 2% +16.3% 10714798 ± 2% perf-stat.ps.cache-misses
> 70327211 -20.7% 55796363 ± 2% perf-stat.ps.cache-references
> 29136 ± 3% -27.5% 21118 ± 2% perf-stat.ps.context-switches
> 1.385e+09 -32.2% 9.386e+08 perf-stat.ps.dTLB-loads
> 7.64e+08 -33.3% 5.099e+08 perf-stat.ps.dTLB-stores
> 1091797 ± 2% -18.6% 888828 perf-stat.ps.iTLB-load-misses
> 2116297 -6.8% 1971620 perf-stat.ps.iTLB-loads
> 4.478e+09 -30.8% 3.1e+09 perf-stat.ps.instructions
> 320258 ± 5% +970.9% 3429628 ± 2% perf-stat.ps.node-load-misses
> 78661 ± 19% +323.8% 333386 ± 6% perf-stat.ps.node-loads
> 4195116 ± 2% -34.5% 2749032 perf-stat.ps.node-store-misses
> 2.847e+11 -31.2% 1.959e+11 perf-stat.total.instructions
>
>
>
>
> Disclaimer:
> Results have been estimated based on internal Intel analysis and are provided
> for informational purposes only. Any difference in system hardware or software
> design or configuration may affect actual performance.
>
>
Powered by blists - more mailing lists