[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <20220513063308.GA21013@xsang-OptiPlex-9020>
Date: Fri, 13 May 2022 14:33:08 +0800
From: kernel test robot <oliver.sang@...el.com>
To: Waiman Long <longman@...hat.com>
Cc: 0day robot <lkp@...el.com>, LKML <linux-kernel@...r.kernel.org>,
lkp@...ts.01.org, ying.huang@...el.com, feng.tang@...el.com,
zhengjun.xing@...ux.intel.com, fengwei.yin@...el.com,
Tejun Heo <tj@...nel.org>, Zefan Li <lizefan.x@...edance.com>,
Johannes Weiner <hannes@...xchg.org>, cgroups@...r.kernel.org,
linux-mm@...ck.org, Andrew Morton <akpm@...ux-foundation.org>,
Michal Hocko <mhocko@...nel.org>,
Dave Hansen <dave.hansen@...el.com>,
Waiman Long <longman@...hat.com>, stable@...r.kernel.org
Subject: [cgroup/cpuset] 7aaa45f837: kbuild.buildtime_per_iteration 2583.0%
regression
Greeting,
FYI, we noticed a 2583.0% regression of kbuild.buildtime_per_iteration due to commit:
commit: 7aaa45f8374887d98cd479a6bdd3b4c74e6a981a ("[PATCH] cgroup/cpuset: Remove redundant cpu/node masks setup in cpuset_init_smp()")
url: https://github.com/intel-lab-lkp/linux/commits/Waiman-Long/cgroup-cpuset-Remove-redundant-cpu-node-masks-setup-in-cpuset_init_smp/20220425-101126
base: https://git.kernel.org/cgit/linux/kernel/git/tj/cgroup.git for-next
patch link: https://lore.kernel.org/linux-mm/20220425020926.1264611-1-longman@redhat.com
in testcase: kbuild
on test machine: 88 threads 2 sockets Intel(R) Xeon(R) Gold 6238M CPU @ 2.10GHz with 128G memory
with following parameters:
memory.high: 90%
memory.low: 50%
memory.max: max
pids.max: 10000
runtime: 300s
nr_task: 50%
target: autoksyms_recursive
cpufreq_governor: performance
ucode: 0x500320a
If you fix the issue, kindly add following tag
Reported-by: kernel test robot <oliver.sang@...el.com>
Details are as below:
-------------------------------------------------------------------------------------------------->
To reproduce:
git clone https://github.com/intel/lkp-tests.git
cd lkp-tests
sudo bin/lkp install job.yaml # job file is attached in this email
bin/lkp split-job --compatible job.yaml # generate the yaml file for lkp run
sudo bin/lkp run generated-yaml-file
# if come across any failure that blocks the test,
# please remove ~/.lkp and /lkp dir to run from a clean state.
=========================================================================================
compiler/cpufreq_governor/kconfig/memory.high/memory.low/memory.max/nr_task/pids.max/rootfs/runtime/target/tbox_group/testcase/ucode:
gcc-11/performance/x86_64-rhel-8.3/90%/50%/max/50%/10000/debian-10.4-x86_64-20200603.cgz/300s/autoksyms_recursive/lkp-csl-2sp9/kbuild/0x500320a
commit:
4ab93063c8 ("cgroup: Add test_cpucg_weight_underprovisioned() testcase")
7aaa45f837 ("cgroup/cpuset: Remove redundant cpu/node masks setup in cpuset_init_smp()")
4ab93063c83a2478 7aaa45f8374887d98cd479a6bdd
---------------- ---------------------------
%stddev %change %stddev
\ | \
44.43 +2583.0% 1192 kbuild.buildtime_per_iteration
311.14 +283.0% 1191 kbuild.time.elapsed_time
311.14 +283.0% 1191 kbuild.time.elapsed_time.max
363486 -4.7% 346532 kbuild.time.involuntary_context_switches
7528 ± 2% -99.1% 68.00 kbuild.time.major_page_faults
3.854e+08 -85.7% 55047165 kbuild.time.minor_page_faults
2608 -96.4% 95.00 kbuild.time.percent_of_cpu_this_job_got
1361 -89.5% 142.46 kbuild.time.system_time
6756 -85.3% 991.32 kbuild.time.user_time
1763962 -84.6% 270814 kbuild.time.voluntary_context_switches
7.00 -85.7% 1.00 kbuild.workload
1.886e+10 +439.9% 1.018e+11 cpuidle..time
40524113 +397.6% 2.016e+08 ± 9% cpuidle..usage
352.84 +249.7% 1233 uptime.boot
22088 +378.4% 105677 uptime.idle
68.86 +28.8 97.63 mpstat.cpu.all.idle%
0.16 ± 3% -0.1 0.08 ± 8% mpstat.cpu.all.soft%
5.23 -5.1 0.15 mpstat.cpu.all.sys%
24.48 -23.5 0.96 mpstat.cpu.all.usr%
68.83 +42.1% 97.83 vmstat.cpu.id
24.00 -100.0% 0.00 vmstat.cpu.us
26.00 +73.1% 45.00 vmstat.procs.r
17438 -88.3% 2045 vmstat.system.cs
1.749e+08 -70.6% 51378800 numa-numastat.node0.local_node
1.75e+08 -70.6% 51378813 numa-numastat.node0.numa_hit
50974 ± 50% -100.0% 13.50 ± 75% numa-numastat.node0.other_node
1.75e+08 -100.0% 82.83 ±223% numa-numastat.node1.local_node
1.75e+08 -100.0% 79394 numa-numastat.node1.numa_hit
28316 ± 91% +180.1% 79311 numa-numastat.node1.other_node
6700 ± 4% +69.9% 11385 ± 3% meminfo.Active
6540 ± 4% +69.1% 11063 meminfo.Active(anon)
223421 +31.7% 294262 ± 2% meminfo.AnonHugePages
1300267 +31.4% 1708081 meminfo.AnonPages
1576268 +25.7% 1981479 meminfo.Committed_AS
1338713 +28.5% 1719792 meminfo.Inactive
1338565 +28.4% 1719004 meminfo.Inactive(anon)
18332 -11.0% 16310 meminfo.KernelStack
93129 -43.8% 52361 meminfo.Mapped
14656 ± 2% -8.9% 13355 meminfo.PageTables
43576 ± 3% -52.0% 20896 meminfo.Shmem
7581989 -10.5% 6785067 meminfo.max_used_kB
875.33 -100.0% 0.00 turbostat.Avg_MHz
31.74 -31.7 0.00 turbostat.Busy%
2764 -100.0% 0.00 turbostat.Bzy_MHz
173892 ± 10% -100.0% 0.00 turbostat.C1
0.02 ± 17% -0.0 0.00 turbostat.C1%
33508125 ± 20% -100.0% 0.00 turbostat.C1E
48.18 ± 44% -48.2 0.00 turbostat.C1E%
6776884 ±102% -100.0% 0.00 turbostat.C6
20.30 ±104% -20.3 0.00 turbostat.C6%
66.40 ± 4% -100.0% 0.00 turbostat.CPU%c1
1.86 ±139% -100.0% 0.00 turbostat.CPU%c6
59.17 -100.0% 0.00 turbostat.CoreTmp
0.33 -100.0% 0.00 turbostat.IPC
55232878 -100.0% 0.00 turbostat.IRQ
56380 ± 6% -100.0% 0.00 turbostat.POLL
59.00 -100.0% 0.00 turbostat.PkgTmp
215.76 -100.0% 0.00 turbostat.PkgWatt
16.16 -100.0% 0.00 turbostat.RAMWatt
2095 -100.0% 0.00 turbostat.TSC_MHz
813.00 ± 11% +237.1% 2740 numa-vmstat.node0.nr_active_anon
183776 +131.6% 425651 numa-vmstat.node0.nr_anon_pages
186417 +129.7% 428211 numa-vmstat.node0.nr_inactive_anon
1978 +68.0% 3324 numa-vmstat.node0.nr_page_table_pages
3282 ± 7% +53.1% 5023 numa-vmstat.node0.nr_shmem
813.00 ± 11% +237.1% 2740 numa-vmstat.node0.nr_zone_active_anon
186416 +129.7% 428211 numa-vmstat.node0.nr_zone_inactive_anon
1.75e+08 -70.6% 51378302 numa-vmstat.node0.numa_hit
1.749e+08 -70.6% 51378288 numa-vmstat.node0.numa_local
50973 ± 50% -100.0% 13.50 ± 75% numa-vmstat.node0.numa_other
823.33 ± 5% -97.5% 20.33 ± 12% numa-vmstat.node1.nr_active_anon
139750 ± 2% -99.9% 88.17 ±114% numa-vmstat.node1.nr_anon_pages
157716 ±183% -97.6% 3733 ±208% numa-vmstat.node1.nr_file_pages
146649 ± 2% -99.8% 265.00 ± 56% numa-vmstat.node1.nr_inactive_anon
8288 ± 4% -31.8% 5653 ± 5% numa-vmstat.node1.nr_kernel_stack
8789 ± 47% -99.6% 32.00 ±120% numa-vmstat.node1.nr_mapped
1658 -99.8% 4.00 ± 73% numa-vmstat.node1.nr_page_table_pages
7610 ± 3% -97.4% 197.17 ± 27% numa-vmstat.node1.nr_shmem
11225 ±101% -74.3% 2883 ± 9% numa-vmstat.node1.nr_slab_reclaimable
20034 ± 3% -25.8% 14861 ± 4% numa-vmstat.node1.nr_slab_unreclaimable
823.33 ± 5% -97.5% 20.33 ± 12% numa-vmstat.node1.nr_zone_active_anon
146648 ± 2% -99.8% 265.00 ± 56% numa-vmstat.node1.nr_zone_inactive_anon
1.75e+08 -100.0% 79394 numa-vmstat.node1.numa_hit
1.75e+08 -100.0% 82.83 ±223% numa-vmstat.node1.numa_local
28312 ± 91% +180.1% 79311 numa-vmstat.node1.numa_other
1642 ± 4% +68.3% 2763 proc-vmstat.nr_active_anon
322313 +32.5% 427139 proc-vmstat.nr_anon_pages
864421 -1.5% 851593 proc-vmstat.nr_file_pages
332108 +29.4% 429879 proc-vmstat.nr_inactive_anon
18183 -10.3% 16310 proc-vmstat.nr_kernel_stack
23431 -43.6% 13208 proc-vmstat.nr_mapped
3550 ± 2% -6.0% 3337 proc-vmstat.nr_page_table_pages
11164 ± 3% -53.2% 5223 proc-vmstat.nr_shmem
46151 -2.2% 45116 proc-vmstat.nr_slab_reclaimable
46770 -8.6% 42749 proc-vmstat.nr_slab_unreclaimable
1642 ± 4% +68.3% 2763 proc-vmstat.nr_zone_active_anon
332108 +29.4% 429879 proc-vmstat.nr_zone_inactive_anon
360277 ± 30% -99.8% 680.17 ±112% proc-vmstat.numa_hint_faults
357678 ± 31% -99.8% 680.17 ±112% proc-vmstat.numa_hint_faults_local
3.5e+08 -85.3% 51462030 proc-vmstat.numa_hit
3.499e+08 -85.3% 51382705 proc-vmstat.numa_local
10372 ±129% -100.0% 0.00 proc-vmstat.numa_pages_migrated
790289 ± 15% -75.5% 193877 ± 7% proc-vmstat.numa_pte_updates
290045 -86.1% 40379 proc-vmstat.pgactivate
3.498e+08 -85.3% 51462440 proc-vmstat.pgalloc_normal
3.871e+08 -85.1% 57578373 proc-vmstat.pgfault
3.497e+08 -85.3% 51397471 proc-vmstat.pgfree
10372 ±129% -100.0% 0.00 proc-vmstat.pgmigrate_success
2920438 -78.6% 624274 proc-vmstat.pgreuse
18128 -85.7% 2594 proc-vmstat.thp_fault_alloc
771909 -85.7% 110390 proc-vmstat.unevictable_pgs_culled
3411 ± 11% +226.2% 11127 numa-meminfo.node0.Active
3251 ± 11% +237.3% 10966 numa-meminfo.node0.Active(anon)
188175 ± 6% +56.0% 293523 ± 2% numa-meminfo.node0.AnonHugePages
737315 +130.8% 1701834 numa-meminfo.node0.AnonPages
1657848 ± 5% +85.9% 3081876 ± 4% numa-meminfo.node0.AnonPages.max
748023 +129.0% 1712859 numa-meminfo.node0.Inactive
747875 +128.9% 1712108 numa-meminfo.node0.Inactive(anon)
4306840 ± 28% +35.6% 5839514 numa-meminfo.node0.MemUsed
7772 ± 3% +71.1% 13297 numa-meminfo.node0.PageTables
13182 ± 8% +52.5% 20099 numa-meminfo.node0.Shmem
3307 ± 5% -92.7% 242.00 ±147% numa-meminfo.node1.Active
3307 ± 5% -97.5% 81.33 ± 12% numa-meminfo.node1.Active(anon)
34918 ± 34% -100.0% 0.00 numa-meminfo.node1.AnonHugePages
557333 ± 2% -99.9% 353.33 ±114% numa-meminfo.node1.AnonPages
1876230 ± 3% -100.0% 496.67 ± 96% numa-meminfo.node1.AnonPages.max
631836 ±183% -97.6% 14935 ±208% numa-meminfo.node1.FilePages
586137 ± 2% -99.8% 1096 ± 51% numa-meminfo.node1.Inactive
586137 ± 2% -99.8% 1061 ± 56% numa-meminfo.node1.Inactive(anon)
44909 ±101% -74.3% 11533 ± 9% numa-meminfo.node1.KReclaimable
8244 ± 4% -31.4% 5653 ± 5% numa-meminfo.node1.KernelStack
35330 ± 46% -99.6% 129.17 ±119% numa-meminfo.node1.Mapped
1612958 ± 74% -79.4% 333010 ± 7% numa-meminfo.node1.MemUsed
6577 -99.8% 16.17 ± 73% numa-meminfo.node1.PageTables
44909 ±101% -74.3% 11533 ± 9% numa-meminfo.node1.SReclaimable
80135 ± 3% -25.8% 59446 ± 4% numa-meminfo.node1.SUnreclaim
31315 ± 3% -97.5% 789.17 ± 27% numa-meminfo.node1.Shmem
125044 ± 38% -43.2% 70980 ± 4% numa-meminfo.node1.Slab
1.978e+10 -95.6% 8.797e+08 perf-stat.i.branch-instructions
5.616e+08 -95.6% 24754884 ± 8% perf-stat.i.branch-misses
1.495e+08 -94.7% 7868663 ± 5% perf-stat.i.cache-misses
6.176e+08 -94.5% 34193232 ± 36% perf-stat.i.cache-references
17499 -88.4% 2029 perf-stat.i.context-switches
0.93 ± 3% +36.6% 1.27 ± 5% perf-stat.i.cpi
7.611e+10 -93.3% 5.121e+09 ± 4% perf-stat.i.cpu-cycles
1275 -100.0% 0.04 ± 28% perf-stat.i.cpu-migrations
1029 ± 5% -31.3% 707.28 ± 2% perf-stat.i.cycles-between-cache-misses
14172722 -94.1% 836284 ± 47% perf-stat.i.dTLB-load-misses
2.503e+10 -95.5% 1.122e+09 perf-stat.i.dTLB-loads
7917427 -95.8% 332811 ± 15% perf-stat.i.dTLB-store-misses
1.244e+10 -95.4% 5.666e+08 perf-stat.i.dTLB-stores
24631063 -92.3% 1894591 ± 2% perf-stat.i.iTLB-load-misses
71323368 -94.9% 3665231 ± 6% perf-stat.i.iTLB-loads
9.005e+10 -95.5% 4.054e+09 perf-stat.i.instructions
3571 -39.7% 2152 perf-stat.i.instructions-per-iTLB-miss
1.10 ± 2% -28.1% 0.79 ± 5% perf-stat.i.ipc
24.38 ± 2% -99.5% 0.11 ± 23% perf-stat.i.major-faults
0.86 -93.3% 0.06 ± 4% perf-stat.i.metric.GHz
658.74 -95.6% 29.19 perf-stat.i.metric.M/sec
1236978 -96.1% 48019 perf-stat.i.minor-faults
24.93 ± 5% -18.7 6.26 ± 22% perf-stat.i.node-load-miss-rate%
6765575 -97.4% 178075 ± 26% perf-stat.i.node-load-misses
55797319 -94.4% 3120263 perf-stat.i.node-loads
18.32 ± 4% -9.7 8.62 ± 22% perf-stat.i.node-store-miss-rate%
1648658 -97.0% 50137 ± 26% perf-stat.i.node-store-misses
11762943 -95.0% 585083 perf-stat.i.node-stores
1237003 -96.1% 48019 perf-stat.i.page-faults
6.86 +23.4% 8.46 ± 38% perf-stat.overall.MPKI
0.85 +49.5% 1.26 ± 5% perf-stat.overall.cpi
508.96 +27.9% 651.10 perf-stat.overall.cycles-between-cache-misses
25.67 +8.5 34.14 ± 3% perf-stat.overall.iTLB-load-miss-rate%
3655 -41.4% 2140 perf-stat.overall.instructions-per-iTLB-miss
1.18 -32.9% 0.79 ± 5% perf-stat.overall.ipc
10.82 -5.4 5.38 ± 24% perf-stat.overall.node-load-miss-rate%
12.30 -4.4 7.85 ± 23% perf-stat.overall.node-store-miss-rate%
4.002e+12 +20.6% 4.826e+12 perf-stat.overall.path-length
1.972e+10 -95.5% 8.791e+08 perf-stat.ps.branch-instructions
5.599e+08 -95.6% 24737459 ± 8% perf-stat.ps.branch-misses
1.491e+08 -94.7% 7862895 ± 5% perf-stat.ps.cache-misses
6.158e+08 -94.5% 34168138 ± 36% perf-stat.ps.cache-references
17447 -88.4% 2027 perf-stat.ps.context-switches
7.588e+10 -93.3% 5.118e+09 ± 4% perf-stat.ps.cpu-cycles
1272 -100.0% 0.04 ± 28% perf-stat.ps.cpu-migrations
14129808 -94.1% 835661 ± 47% perf-stat.ps.dTLB-load-misses
2.495e+10 -95.5% 1.121e+09 perf-stat.ps.dTLB-loads
7893270 -95.8% 332570 ± 15% perf-stat.ps.dTLB-store-misses
1.24e+10 -95.4% 5.662e+08 perf-stat.ps.dTLB-stores
24556703 -92.3% 1893249 ± 2% perf-stat.ps.iTLB-load-misses
71106326 -94.8% 3662587 ± 6% perf-stat.ps.iTLB-loads
8.978e+10 -95.5% 4.051e+09 perf-stat.ps.instructions
24.32 ± 2% -99.5% 0.11 ± 24% perf-stat.ps.major-faults
1233214 -96.1% 47986 perf-stat.ps.minor-faults
6745825 -97.4% 177950 ± 26% perf-stat.ps.node-load-misses
55627061 -94.4% 3117959 perf-stat.ps.node-loads
1644068 -97.0% 50102 ± 26% perf-stat.ps.node-store-misses
11727184 -95.0% 584651 perf-stat.ps.node-stores
1233238 -96.1% 47986 perf-stat.ps.page-faults
2.802e+13 -82.8% 4.826e+12 perf-stat.total.instructions
0.39 +3586.3% 14.32 ± 4% sched_debug.cfs_rq:/.h_nr_running.avg
1.11 ± 7% +4138.2% 47.09 sched_debug.cfs_rq:/.h_nr_running.max
0.43 +4683.1% 20.72 sched_debug.cfs_rq:/.h_nr_running.stddev
14081 ± 20% +2107.0% 310782 ± 4% sched_debug.cfs_rq:/.load.avg
407170 ± 49% +170.4% 1101004 sched_debug.cfs_rq:/.load.max
52684 ± 42% +768.5% 457567 sched_debug.cfs_rq:/.load.stddev
37.01 ± 46% +723.1% 304.63 ± 4% sched_debug.cfs_rq:/.load_avg.avg
144.31 ± 45% +210.7% 448.32 sched_debug.cfs_rq:/.load_avg.stddev
2488735 -92.5% 187680 ± 3% sched_debug.cfs_rq:/.min_vruntime.avg
2938975 -80.2% 581666 sched_debug.cfs_rq:/.min_vruntime.max
2060745 ± 2% -99.5% 10250 ± 34% sched_debug.cfs_rq:/.min_vruntime.min
182406 ± 6% +39.2% 253824 sched_debug.cfs_rq:/.min_vruntime.stddev
0.39 -23.7% 0.30 ± 4% sched_debug.cfs_rq:/.nr_running.avg
400.60 +3272.5% 13510 ± 5% sched_debug.cfs_rq:/.runnable_avg.avg
971.78 ± 2% +4430.1% 44022 sched_debug.cfs_rq:/.runnable_avg.max
370.10 +5173.5% 19517 sched_debug.cfs_rq:/.runnable_avg.stddev
12579 ±1058% -3228.1% -393486 sched_debug.cfs_rq:/.spread0.avg
462824 ± 24% -99.9% 540.01 ± 92% sched_debug.cfs_rq:/.spread0.max
182401 ± 6% +39.2% 253852 sched_debug.cfs_rq:/.spread0.stddev
400.13 -25.0% 300.07 ± 4% sched_debug.cfs_rq:/.util_avg.avg
369.97 +18.4% 437.88 sched_debug.cfs_rq:/.util_avg.stddev
10.90 ± 10% -87.5% 1.37 ± 48% sched_debug.cfs_rq:/.util_est_enqueued.avg
454.44 ± 14% -90.3% 44.23 ± 3% sched_debug.cfs_rq:/.util_est_enqueued.max
60.82 ± 13% -90.9% 5.53 ± 14% sched_debug.cfs_rq:/.util_est_enqueued.stddev
0.92 ± 36% -90.0% 0.09 ±176% sched_debug.cfs_rq:/init.scope.load_avg.avg
1.08 ± 17% -91.5% 0.09 ±176% sched_debug.cfs_rq:/init.scope.load_avg.max
0.50 ±107% +2506.1% 12.97 ± 57% sched_debug.cfs_rq:/init.scope.min_vruntime.avg
0.54 ±104% +2288.9% 13.00 ± 57% sched_debug.cfs_rq:/init.scope.min_vruntime.max
0.45 ±111% +2768.6% 12.93 ± 57% sched_debug.cfs_rq:/init.scope.min_vruntime.min
0.92 ± 36% -90.0% 0.09 ±176% sched_debug.cfs_rq:/init.scope.runnable_avg.avg
1.08 ± 17% -91.5% 0.09 ±176% sched_debug.cfs_rq:/init.scope.runnable_avg.max
0.51 ±123% +1809.6% 9.78 ± 40% sched_debug.cfs_rq:/init.scope.se->sum_exec_runtime.avg
0.52 ±122% +1769.0% 9.78 ± 40% sched_debug.cfs_rq:/init.scope.se->sum_exec_runtime.max
0.50 ±126% +1852.0% 9.77 ± 40% sched_debug.cfs_rq:/init.scope.se->sum_exec_runtime.min
927421 ±116% -88.6% 105436 ± 29% sched_debug.cfs_rq:/init.scope.se->vruntime.avg
927672 ±116% -88.6% 105945 ± 29% sched_debug.cfs_rq:/init.scope.se->vruntime.max
927171 ±117% -88.7% 104926 ± 29% sched_debug.cfs_rq:/init.scope.se->vruntime.min
-903221 -88.3% -106040 sched_debug.cfs_rq:/init.scope.spread0.avg
-903221 -88.3% -106040 sched_debug.cfs_rq:/init.scope.spread0.max
-903221 -88.3% -106040 sched_debug.cfs_rq:/init.scope.spread0.min
1.08 ± 17% -91.5% 0.09 ±176% sched_debug.cfs_rq:/init.scope.tg_load_avg.avg
1.08 ± 17% -91.5% 0.09 ±176% sched_debug.cfs_rq:/init.scope.tg_load_avg.max
1.08 ± 17% -91.5% 0.09 ±176% sched_debug.cfs_rq:/init.scope.tg_load_avg.min
0.92 ± 36% -90.0% 0.09 ±176% sched_debug.cfs_rq:/init.scope.tg_load_avg_contrib.avg
1.08 ± 17% -91.5% 0.09 ±176% sched_debug.cfs_rq:/init.scope.tg_load_avg_contrib.max
0.92 ± 36% -96.4% 0.03 ±111% sched_debug.cfs_rq:/init.scope.util_avg.avg
1.08 ± 17% -96.9% 0.03 ±111% sched_debug.cfs_rq:/init.scope.util_avg.max
109.53 ±107% +50485.6% 55405 sched_debug.cfs_rq:/kbuild.1.MIN_vruntime.avg
9638 ±107% +475.0% 55414 sched_debug.cfs_rq:/kbuild.1.MIN_vruntime.max
-0.30 -1.9e+07% 55396 sched_debug.cfs_rq:/kbuild.1.MIN_vruntime.min
0.39 ± 2% +11812.2% 46.70 sched_debug.cfs_rq:/kbuild.1.h_nr_running.avg
1.00 +4594.2% 46.94 sched_debug.cfs_rq:/kbuild.1.h_nr_running.max
410733 ± 2% +11823.3% 48972868 sched_debug.cfs_rq:/kbuild.1.load.avg
1048576 +4594.2% 49221905 sched_debug.cfs_rq:/kbuild.1.load.max
659.13 ± 3% +7394.5% 49398 sched_debug.cfs_rq:/kbuild.1.load_avg.avg
2647 ± 11% +1777.5% 49713 sched_debug.cfs_rq:/kbuild.1.load_avg.max
0.03 ±223% +1.8e+08% 49083 ± 2% sched_debug.cfs_rq:/kbuild.1.load_avg.min
109.53 ±107% +50489.2% 55409 sched_debug.cfs_rq:/kbuild.1.max_vruntime.avg
9638 ±107% +475.0% 55418 sched_debug.cfs_rq:/kbuild.1.max_vruntime.max
-0.30 -1.9e+07% 55400 sched_debug.cfs_rq:/kbuild.1.max_vruntime.min
87264 -36.5% 55405 sched_debug.cfs_rq:/kbuild.1.min_vruntime.avg
99270 -44.2% 55414 sched_debug.cfs_rq:/kbuild.1.min_vruntime.max
75767 -26.9% 55396 sched_debug.cfs_rq:/kbuild.1.min_vruntime.min
5024 ± 7% -99.8% 9.08 ±100% sched_debug.cfs_rq:/kbuild.1.min_vruntime.stddev
0.39 ± 2% +11812.2% 46.70 sched_debug.cfs_rq:/kbuild.1.nr_running.avg
1.00 +4594.2% 46.94 sched_debug.cfs_rq:/kbuild.1.nr_running.max
35.29 ± 22% -100.0% 0.00 sched_debug.cfs_rq:/kbuild.1.removed.load_avg.avg
919.11 ± 14% -100.0% 0.00 sched_debug.cfs_rq:/kbuild.1.removed.load_avg.max
156.52 ± 17% -100.0% 0.00 sched_debug.cfs_rq:/kbuild.1.removed.load_avg.stddev
14.27 ± 27% -100.0% 0.00 sched_debug.cfs_rq:/kbuild.1.removed.runnable_avg.avg
419.11 ± 25% -100.0% 0.00 sched_debug.cfs_rq:/kbuild.1.removed.runnable_avg.max
65.99 ± 27% -100.0% 0.00 sched_debug.cfs_rq:/kbuild.1.removed.runnable_avg.stddev
14.27 ± 27% -100.0% 0.00 sched_debug.cfs_rq:/kbuild.1.removed.util_avg.avg
419.11 ± 25% -100.0% 0.00 sched_debug.cfs_rq:/kbuild.1.removed.util_avg.max
65.99 ± 27% -100.0% 0.00 sched_debug.cfs_rq:/kbuild.1.removed.util_avg.stddev
412.27 +10527.3% 43813 sched_debug.cfs_rq:/kbuild.1.runnable_avg.avg
967.14 ± 2% +4443.7% 43944 sched_debug.cfs_rq:/kbuild.1.runnable_avg.max
0.03 ±223% +1.6e+08% 43681 sched_debug.cfs_rq:/kbuild.1.runnable_avg.min
368.39 -64.4% 131.08 ±101% sched_debug.cfs_rq:/kbuild.1.runnable_avg.stddev
13.05 ± 10% +7646.3% 1011 sched_debug.cfs_rq:/kbuild.1.se->avg.load_avg.avg
87.56 ± 16% +1069.5% 1023 sched_debug.cfs_rq:/kbuild.1.se->avg.load_avg.max
412.13 +10524.7% 43787 sched_debug.cfs_rq:/kbuild.1.se->avg.runnable_avg.avg
967.47 ± 2% +4439.4% 43917 sched_debug.cfs_rq:/kbuild.1.se->avg.runnable_avg.max
0.03 ±223% +1.6e+08% 43657 sched_debug.cfs_rq:/kbuild.1.se->avg.runnable_avg.min
368.51 -64.7% 129.95 ±101% sched_debug.cfs_rq:/kbuild.1.se->avg.runnable_avg.stddev
411.92 +141.6% 995.10 sched_debug.cfs_rq:/kbuild.1.se->avg.util_avg.avg
0.03 ±223% +3.5e+06% 983.40 ± 2% sched_debug.cfs_rq:/kbuild.1.se->avg.util_avg.min
368.41 -96.8% 11.70 ±100% sched_debug.cfs_rq:/kbuild.1.se->avg.util_avg.stddev
188343 +220.5% 603616 sched_debug.cfs_rq:/kbuild.1.se->exec_start.avg
188644 +220.2% 604099 sched_debug.cfs_rq:/kbuild.1.se->exec_start.max
186221 +223.9% 603133 sched_debug.cfs_rq:/kbuild.1.se->exec_start.min
51889 ± 7% +1920.8% 1048576 sched_debug.cfs_rq:/kbuild.1.se->load.weight.avg
176539 ± 22% +494.0% 1048576 sched_debug.cfs_rq:/kbuild.1.se->load.weight.max
1173 ± 53% +89267.3% 1048576 sched_debug.cfs_rq:/kbuild.1.se->load.weight.min
41159 ± 21% -100.0% 0.00 sched_debug.cfs_rq:/kbuild.1.se->load.weight.stddev
44377 +1155.3% 557089 sched_debug.cfs_rq:/kbuild.1.se->sum_exec_runtime.avg
52129 +968.7% 557092 sched_debug.cfs_rq:/kbuild.1.se->sum_exec_runtime.max
36817 ± 2% +1413.1% 557086 sched_debug.cfs_rq:/kbuild.1.se->sum_exec_runtime.min
3173 ± 5% -99.9% 3.27 ±100% sched_debug.cfs_rq:/kbuild.1.se->sum_exec_runtime.stddev
2490096 -76.7% 580821 sched_debug.cfs_rq:/kbuild.1.se->vruntime.avg
2932328 -80.2% 581126 sched_debug.cfs_rq:/kbuild.1.se->vruntime.max
2062018 ± 2% -71.8% 580516 sched_debug.cfs_rq:/kbuild.1.se->vruntime.min
182836 ± 7% -99.8% 305.00 ±101% sched_debug.cfs_rq:/kbuild.1.se->vruntime.stddev
-2388892 -78.0% -525720 sched_debug.cfs_rq:/kbuild.1.spread0.avg
-2376905 -77.9% -525711 sched_debug.cfs_rq:/kbuild.1.spread0.max
-2400349 -78.1% -525729 sched_debug.cfs_rq:/kbuild.1.spread0.min
5008 ± 7% -99.8% 9.10 ±100% sched_debug.cfs_rq:/kbuild.1.spread0.stddev
59728 ± 4% -16.5% 49878 sched_debug.cfs_rq:/kbuild.1.tg_load_avg.max
2203 ± 20% -99.9% 1.82 ±223% sched_debug.cfs_rq:/kbuild.1.tg_load_avg.stddev
661.43 ± 3% +7393.3% 49562 sched_debug.cfs_rq:/kbuild.1.tg_load_avg_contrib.avg
2655 ± 11% +1778.5% 49878 sched_debug.cfs_rq:/kbuild.1.tg_load_avg_contrib.max
0.03 ±223% +1.8e+08% 49246 ± 2% sched_debug.cfs_rq:/kbuild.1.tg_load_avg_contrib.min
412.06 +140.8% 992.05 sched_debug.cfs_rq:/kbuild.1.util_avg.avg
0.03 ±223% +3.5e+06% 980.41 ± 2% sched_debug.cfs_rq:/kbuild.1.util_avg.min
368.29 -96.8% 11.64 ±100% sched_debug.cfs_rq:/kbuild.1.util_avg.stddev
0.00 ± 27% +7.1e+08% 6.35 ± 16% sched_debug.cfs_rq:/system.slice.MIN_vruntime.avg
0.00 ± 27% +2.9e+09% 25.72 sched_debug.cfs_rq:/system.slice.MIN_vruntime.max
0.00 ± 27% -70.0% 0.00 ± 25% sched_debug.cfs_rq:/system.slice.MIN_vruntime.min
739.56 ± 54% -90.0% 73.75 ± 36% sched_debug.cfs_rq:/system.slice.load_avg.max
183.19 ± 53% -82.7% 31.64 ± 37% sched_debug.cfs_rq:/system.slice.load_avg.stddev
0.00 ± 27% +7.3e+08% 6.45 ± 17% sched_debug.cfs_rq:/system.slice.max_vruntime.avg
0.00 ± 27% +2.9e+09% 26.12 sched_debug.cfs_rq:/system.slice.max_vruntime.max
0.00 ± 27% -70.0% 0.00 ± 25% sched_debug.cfs_rq:/system.slice.max_vruntime.min
30.65 ± 43% +4649.8% 1455 ± 30% sched_debug.cfs_rq:/system.slice.min_vruntime.avg
97.37 ± 36% +1415.3% 1475 ± 30% sched_debug.cfs_rq:/system.slice.min_vruntime.max
4.88 ±263% +29568.6% 1449 ± 30% sched_debug.cfs_rq:/system.slice.min_vruntime.min
739.56 ± 54% -90.0% 73.75 ± 36% sched_debug.cfs_rq:/system.slice.runnable_avg.max
183.44 ± 53% -82.8% 31.64 ± 37% sched_debug.cfs_rq:/system.slice.runnable_avg.stddev
698.25 ± 54% -92.7% 51.19 sched_debug.cfs_rq:/system.slice.se->avg.load_avg.max
173.10 ± 53% -87.3% 21.91 ± 5% sched_debug.cfs_rq:/system.slice.se->avg.load_avg.stddev
739.56 ± 54% -90.0% 73.76 ± 36% sched_debug.cfs_rq:/system.slice.se->avg.runnable_avg.max
183.45 ± 53% -82.8% 31.64 ± 37% sched_debug.cfs_rq:/system.slice.se->avg.runnable_avg.stddev
50.39 ± 53% -89.1% 5.49 ± 40% sched_debug.cfs_rq:/system.slice.se->avg.util_avg.avg
739.39 ± 54% -97.0% 22.29 ± 38% sched_debug.cfs_rq:/system.slice.se->avg.util_avg.max
183.41 ± 53% -94.8% 9.54 ± 39% sched_debug.cfs_rq:/system.slice.se->avg.util_avg.stddev
10507 ± 54% -92.1% 828.57 ± 5% sched_debug.cfs_rq:/system.slice.se->exec_start.stddev
796179 ± 45% -80.2% 157286 ± 19% sched_debug.cfs_rq:/system.slice.se->load.weight.max
2018 ± 70% +7694.2% 157286 ± 19% sched_debug.cfs_rq:/system.slice.se->load.weight.min
299826 ± 43% -100.0% 0.00 sched_debug.cfs_rq:/system.slice.se->load.weight.stddev
14.57 ±103% +10710.8% 1574 ± 30% sched_debug.cfs_rq:/system.slice.se->sum_exec_runtime.avg
98.32 ± 35% +1522.3% 1595 ± 29% sched_debug.cfs_rq:/system.slice.se->sum_exec_runtime.max
5.78 ±222% +27036.3% 1568 ± 30% sched_debug.cfs_rq:/system.slice.se->sum_exec_runtime.min
10560 ± 60% -95.1% 519.51 ± 12% sched_debug.cfs_rq:/system.slice.se->vruntime.stddev
781.68 ± 54% -89.7% 80.14 ± 30% sched_debug.cfs_rq:/system.slice.tg_load_avg.avg
783.75 ± 54% -88.9% 86.65 ± 31% sched_debug.cfs_rq:/system.slice.tg_load_avg.max
779.86 ± 54% -90.5% 73.87 ± 36% sched_debug.cfs_rq:/system.slice.tg_load_avg.min
739.36 ± 54% -90.0% 73.87 ± 36% sched_debug.cfs_rq:/system.slice.tg_load_avg_contrib.max
183.14 ± 53% -82.7% 31.69 ± 37% sched_debug.cfs_rq:/system.slice.tg_load_avg_contrib.stddev
50.40 ± 53% -89.1% 5.49 ± 40% sched_debug.cfs_rq:/system.slice.util_avg.avg
739.39 ± 54% -97.0% 22.29 ± 38% sched_debug.cfs_rq:/system.slice.util_avg.max
183.40 ± 53% -94.8% 9.54 ± 39% sched_debug.cfs_rq:/system.slice.util_avg.stddev
130276 ± 3% -53.9% 60023 ± 30% sched_debug.cpu.avg_idle.stddev
190561 +223.2% 615836 sched_debug.cpu.clock.avg
190566 +223.2% 615963 sched_debug.cpu.clock.max
190557 +223.1% 615766 sched_debug.cpu.clock.min
2.49 ± 2% +3195.3% 82.17 ± 10% sched_debug.cpu.clock.stddev
188349 +223.1% 608620 sched_debug.cpu.clock_task.avg
188650 +224.3% 611860 sched_debug.cpu.clock_task.max
186038 +224.3% 603373 sched_debug.cpu.clock_task.min
370.07 ± 40% +680.3% 2887 ± 40% sched_debug.cpu.clock_task.stddev
15323 ± 17% -97.1% 451.92 ± 5% sched_debug.cpu.curr->pid.avg
16512 ± 12% -74.5% 4215 ± 5% sched_debug.cpu.curr->pid.stddev
0.00 ± 5% +351.8% 0.00 ± 10% sched_debug.cpu.next_balance.stddev
0.38 +40.2% 0.54 sched_debug.cpu.nr_running.avg
1.11 ± 7% +4138.2% 47.09 sched_debug.cpu.nr_running.max
0.42 +1084.1% 4.99 sched_debug.cpu.nr_running.stddev
34116 -48.2% 17664 sched_debug.cpu.nr_switches.avg
66559 ± 6% +1204.5% 868266 sched_debug.cpu.nr_switches.max
26193 ± 2% -96.9% 813.00 ± 3% sched_debug.cpu.nr_switches.min
7164 ± 3% +1254.8% 97055 sched_debug.cpu.nr_switches.stddev
190557 +223.1% 615766 sched_debug.cpu_clk
189985 +223.8% 615197 sched_debug.ktime
193775 +219.7% 619543 sched_debug.sched_clk
Disclaimer:
Results have been estimated based on internal Intel analysis and are provided
for informational purposes only. Any difference in system hardware or software
design or configuration may affect actual performance.
--
0-DAY CI Kernel Test Service
https://01.org/lkp
View attachment "config-5.18.0-rc3-00100-g7aaa45f83748" of type "text/plain" (162679 bytes)
View attachment "job-script" of type "text/plain" (8669 bytes)
View attachment "job.yaml" of type "text/plain" (5827 bytes)
View attachment "reproduce" of type "text/plain" (1126 bytes)
Powered by blists - more mailing lists