[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <202304230012.4ba9dbe7-yujie.liu@intel.com>
Date: Mon, 24 Apr 2023 14:59:47 +0800
From: kernel test robot <yujie.liu@...el.com>
To: Chen Yu <yu.c.chen@...el.com>
CC: Peter Zijlstra <peterz@...radead.org>,
Vincent Guittot <vincent.guittot@...aro.org>,
Ingo Molnar <mingo@...hat.com>,
Juri Lelli <juri.lelli@...hat.com>,
Mel Gorman <mgorman@...hsingularity.net>,
Tim Chen <tim.c.chen@...el.com>,
Dietmar Eggemann <dietmar.eggemann@....com>,
"Steven Rostedt" <rostedt@...dmis.org>,
Ben Segall <bsegall@...gle.com>,
"K Prateek Nayak" <kprateek.nayak@....com>,
Abel Wu <wuyun.abel@...edance.com>,
"Yicong Yang" <yangyicong@...ilicon.com>,
"Gautham R . Shenoy" <gautham.shenoy@....com>,
Honglei Wang <wanghonglei@...ichuxing.com>,
"Len Brown" <len.brown@...el.com>,
Chen Yu <yu.chen.surf@...il.com>,
Tianchen Ding <dtcccc@...ux.alibaba.com>,
Joel Fernandes <joel@...lfernandes.org>,
Josh Don <joshdon@...gle.com>, Hillf Danton <hdanton@...a.com>,
kernel test robot <yujie.liu@...el.com>,
Arjan Van De Ven <arjan.van.de.ven@...el.com>,
"Aaron Lu" <aaron.lu@...el.com>, <linux-kernel@...r.kernel.org>,
Chen Yu <yu.c.chen@...el.com>
Subject: Re: [PATCH v7 0/2] sched/fair: Introduce SIS_CURRENT to wake up
short task on current CPU
Hello,
kernel test robot noticed a 2250.1% improvement of stress-ng.switch.ops_per_sec on:
patch: "sched/fair: Introduce SIS_CURRENT to wake up short task on current CPU"
testcase: stress-ng
test machine: 224 threads 2 sockets (Sapphire Rapids) with 256G memory
parameters:
nr_threads: 100%
testtime: 60s
sc_pid_max: 4194304
class: scheduler
test: switch
cpufreq_governor: performance
Details are as below:
=========================================================================================
class/compiler/cpufreq_governor/kconfig/nr_threads/rootfs/sc_pid_max/tbox_group/test/testcase/testtime:
scheduler/gcc-11/performance/x86_64-rhel-8.3/100%/debian-11.1-x86_64-20220510.cgz/4194304/lkp-spr-r02/switch/stress-ng/60s
commit:
dac54350b7 ("sched/fair: Record the average duration of a task")
f153e964b7 ("sched/fair: Introduce SIS_CURRENT to wake up short task on current CPU")
dac54350b7363c69 f153e964b7d24fd0375f0efad66
---------------- ---------------------------
%stddev %change %stddev
\ | \
1.49e+08 ± 6% +2250.2% 3.503e+09 stress-ng.switch.ops
2483795 ± 6% +2250.1% 58370891 stress-ng.switch.ops_per_sec
69985 ± 4% +4.4e+06% 3.088e+09 stress-ng.time.involuntary_context_switches
31001 -5.1% 29406 stress-ng.time.minor_page_faults
12849 +60.9% 20680 stress-ng.time.percent_of_cpu_this_job_got
7325 +47.5% 10802 stress-ng.time.system_time
672.80 +209.1% 2079 stress-ng.time.user_time
2.836e+08 ± 6% +1273.7% 3.895e+09 stress-ng.time.voluntary_context_switches
14182 -14.6% 12109 uptime.idle
2.93e+09 -64.8% 1.03e+09 cpuidle..time
1.946e+08 ± 5% -82.0% 34997812 ± 41% cpuidle..usage
52479 ± 14% +90.9% 100199 ± 2% meminfo.Active
52368 ± 14% +91.1% 100095 ± 2% meminfo.Active(anon)
51616 ± 15% +90.0% 98071 ± 2% numa-meminfo.node1.Active
51523 ± 15% +90.3% 98037 ± 2% numa-meminfo.node1.Active(anon)
12896 ± 15% +90.7% 24589 ± 2% numa-vmstat.node1.nr_active_anon
12896 ± 15% +90.7% 24589 ± 2% numa-vmstat.node1.nr_zone_active_anon
29.58 -21.7 7.89 ± 2% mpstat.cpu.all.idle%
6.86 -5.6 1.28 ± 3% mpstat.cpu.all.irq%
0.41 ± 6% -0.4 0.05 ± 5% mpstat.cpu.all.soft%
57.53 +17.9 75.48 mpstat.cpu.all.sys%
5.62 +9.7 15.30 mpstat.cpu.all.usr%
31.00 -68.8% 9.67 ± 4% vmstat.cpu.id
5.00 +183.3% 14.17 ± 2% vmstat.cpu.us
185.00 +59.0% 294.17 vmstat.procs.r
7520613 ± 6% +1324.1% 1.071e+08 vmstat.system.cs
819272 ± 4% -35.1% 531594 ± 20% vmstat.system.in
13105 ± 14% +91.1% 25040 ± 2% proc-vmstat.nr_active_anon
135920 -6.0% 127705 proc-vmstat.nr_inactive_anon
57438 ± 3% +6.0% 60898 proc-vmstat.nr_shmem
13105 ± 14% +91.1% 25040 ± 2% proc-vmstat.nr_zone_active_anon
135920 -6.0% 127705 proc-vmstat.nr_zone_inactive_anon
905140 +1.4% 917932 proc-vmstat.numa_hit
702189 +1.8% 714949 proc-vmstat.numa_local
2252 ± 13% +91.7% 4317 ± 2% proc-vmstat.pgactivate
987449 +1.2% 998848 proc-vmstat.pgalloc_normal
2327 +16.0% 2700 turbostat.Avg_MHz
81.46 +12.1 93.60 turbostat.Busy%
23636974 ± 25% -87.4% 2970987 ± 18% turbostat.C1
1.60 ± 20% -1.4 0.15 ± 14% turbostat.C1%
1.678e+08 ± 3% -93.0% 11676240 ± 5% turbostat.C1E
15.38 ± 2% -12.1 3.25 ± 2% turbostat.C1E%
18.52 -65.5% 6.38 ± 2% turbostat.CPU%c1
0.06 ± 7% +773.7% 0.55 turbostat.IPC
53133659 ± 4% -34.6% 34758470 ± 21% turbostat.IRQ
2707361 ± 2% +633.0% 19845296 ± 69% turbostat.POLL
551.77 +21.5% 670.20 turbostat.PkgWatt
17.13 +8.1% 18.52 turbostat.RAMWatt
606383 ± 18% -97.7% 13786 ± 44% sched_debug.cfs_rq:/.MIN_vruntime.avg
1289912 ± 8% -84.0% 205876 ± 44% sched_debug.cfs_rq:/.MIN_vruntime.stddev
0.55 ± 4% +57.7% 0.87 sched_debug.cfs_rq:/.h_nr_running.avg
3.83 ± 22% -47.8% 2.00 ± 14% sched_debug.cfs_rq:/.h_nr_running.max
0.60 ± 8% -41.3% 0.36 ± 4% sched_debug.cfs_rq:/.h_nr_running.stddev
606383 ± 18% -97.7% 13786 ± 44% sched_debug.cfs_rq:/.max_vruntime.avg
1289912 ± 8% -84.0% 205876 ± 44% sched_debug.cfs_rq:/.max_vruntime.stddev
3441744 +73.7% 5979755 sched_debug.cfs_rq:/.min_vruntime.avg
3673914 +70.4% 6258914 sched_debug.cfs_rq:/.min_vruntime.max
2245166 +65.6% 3719103 sched_debug.cfs_rq:/.min_vruntime.min
88019 ± 2% +106.3% 181555 ± 7% sched_debug.cfs_rq:/.min_vruntime.stddev
0.38 ± 4% +40.3% 0.53 sched_debug.cfs_rq:/.nr_running.avg
0.34 ± 3% -61.9% 0.13 ± 26% sched_debug.cfs_rq:/.nr_running.stddev
622.22 ± 2% +33.3% 829.67 sched_debug.cfs_rq:/.runnable_avg.avg
2083 ± 13% -10.9% 1855 ± 5% sched_debug.cfs_rq:/.runnable_avg.max
320.17 ± 5% -45.6% 174.13 ± 10% sched_debug.cfs_rq:/.runnable_avg.stddev
-1218600 +84.3% -2246378 sched_debug.cfs_rq:/.spread0.min
87973 ± 2% +107.2% 182295 ± 7% sched_debug.cfs_rq:/.spread0.stddev
400.48 ± 2% +45.1% 581.11 sched_debug.cfs_rq:/.util_avg.avg
188.48 ± 6% -20.9% 149.09 ± 9% sched_debug.cfs_rq:/.util_avg.stddev
75.30 ± 8% +411.6% 385.20 ± 2% sched_debug.cfs_rq:/.util_est_enqueued.avg
106.05 ± 7% +53.1% 162.35 ± 6% sched_debug.cfs_rq:/.util_est_enqueued.stddev
648805 ± 6% +55.8% 1010843 ± 13% sched_debug.cpu.avg_idle.max
86024 ± 6% +44.6% 124370 ± 11% sched_debug.cpu.avg_idle.stddev
2332 ± 7% +81.6% 4235 sched_debug.cpu.curr->pid.avg
2566 ± 2% -77.1% 587.02 ± 7% sched_debug.cpu.curr->pid.stddev
0.47 ± 6% +65.4% 0.77 sched_debug.cpu.nr_running.avg
0.56 ± 2% -34.3% 0.37 ± 4% sched_debug.cpu.nr_running.stddev
1031696 ± 6% +1343.5% 14892376 sched_debug.cpu.nr_switches.avg
1140186 ± 4% +1348.9% 16520256 sched_debug.cpu.nr_switches.max
678987 ± 6% +1130.3% 8353556 ± 15% sched_debug.cpu.nr_switches.min
54009 ± 29% +1497.3% 862684 ± 11% sched_debug.cpu.nr_switches.stddev
58611259 +100.0% 1.172e+08 sched_debug.sysctl_sched.sysctl_sched_features
15.59 -90.5% 1.47 ± 4% perf-stat.i.MPKI
1.127e+10 ± 5% +912.3% 1.141e+11 perf-stat.i.branch-instructions
1.43 -0.5 0.95 perf-stat.i.branch-miss-rate%
1.572e+08 ± 6% +529.5% 9.894e+08 perf-stat.i.branch-misses
1.27 ± 14% +29.3 30.61 ± 6% perf-stat.i.cache-miss-rate%
5226056 ± 7% +840.6% 49157630 ± 5% perf-stat.i.cache-misses
9.044e+08 ± 6% -76.3% 2.145e+08 ± 2% perf-stat.i.cache-references
7735617 ± 6% +1348.7% 1.121e+08 perf-stat.i.context-switches
8.96 ± 5% -82.5% 1.56 ± 4% perf-stat.i.cpi
5.207e+11 +17.6% 6.122e+11 perf-stat.i.cpu-cycles
2826254 ± 6% -92.6% 210370 ± 5% perf-stat.i.cpu-migrations
105045 ± 7% -82.7% 18181 ± 6% perf-stat.i.cycles-between-cache-misses
0.37 ± 7% -0.3 0.03 ± 7% perf-stat.i.dTLB-load-miss-rate%
59575414 ± 9% -92.4% 4518176 ± 8% perf-stat.i.dTLB-load-misses
1.541e+10 ± 5% +969.1% 1.647e+11 perf-stat.i.dTLB-loads
0.08 -0.1 0.01 ± 7% perf-stat.i.dTLB-store-miss-rate%
6651717 ± 5% -89.8% 680779 ± 5% perf-stat.i.dTLB-store-misses
8.526e+09 ± 6% +1076.1% 1.003e+11 perf-stat.i.dTLB-stores
5.657e+10 ± 5% +905.0% 5.686e+11 perf-stat.i.instructions
0.14 ± 7% +536.8% 0.91 perf-stat.i.ipc
2.32 +17.7% 2.73 perf-stat.i.metric.GHz
94.15 ± 4% +1221.0% 1243 perf-stat.i.metric.K/sec
160.83 ± 5% +951.6% 1691 perf-stat.i.metric.M/sec
96.30 +2.7 99.03 perf-stat.i.node-load-miss-rate%
2066362 ± 8% +1017.1% 23083138 ± 5% perf-stat.i.node-load-misses
78248 ± 11% -45.9% 42316 ± 11% perf-stat.i.node-loads
16.04 -97.7% 0.37 perf-stat.overall.MPKI
1.39 -0.5 0.87 perf-stat.overall.branch-miss-rate%
0.57 ± 12% +22.7 23.25 ± 7% perf-stat.overall.cache-miss-rate%
9.27 ± 5% -88.4% 1.07 perf-stat.overall.cpi
102426 ± 7% -87.8% 12467 ± 5% perf-stat.overall.cycles-between-cache-misses
0.39 ± 6% -0.4 0.00 ± 9% perf-stat.overall.dTLB-load-miss-rate%
0.08 -0.1 0.00 ± 6% perf-stat.overall.dTLB-store-miss-rate%
0.11 ± 5% +759.8% 0.93 perf-stat.overall.ipc
96.35 +3.5 99.82 perf-stat.overall.node-load-miss-rate%
1.094e+10 ± 5% +927.6% 1.124e+11 perf-stat.ps.branch-instructions
1.522e+08 ± 6% +540.5% 9.746e+08 perf-stat.ps.branch-misses
4975869 ± 6% +873.2% 48427022 ± 5% perf-stat.ps.cache-misses
8.813e+08 ± 6% -76.3% 2.086e+08 ± 2% perf-stat.ps.cache-references
7529020 ± 6% +1367.2% 1.105e+08 perf-stat.ps.context-switches
216929 +1.3% 219851 perf-stat.ps.cpu-clock
5.072e+11 +18.7% 6.019e+11 perf-stat.ps.cpu-cycles
2754210 ± 6% -92.8% 198251 ± 5% perf-stat.ps.cpu-migrations
58032406 ± 9% -92.6% 4269838 ± 8% perf-stat.ps.dTLB-load-misses
1.496e+10 ± 5% +984.7% 1.623e+11 perf-stat.ps.dTLB-loads
6474992 ± 5% -90.0% 647675 ± 5% perf-stat.ps.dTLB-store-misses
8.286e+09 ± 6% +1092.6% 9.882e+10 perf-stat.ps.dTLB-stores
5.491e+10 ± 5% +920.3% 5.602e+11 perf-stat.ps.instructions
1984455 ± 7% +1046.4% 22749741 ± 5% perf-stat.ps.node-load-misses
74637 ± 12% -44.7% 41249 ± 11% perf-stat.ps.node-loads
216929 +1.3% 219851 perf-stat.ps.task-clock
3.397e+12 ± 6% +921.9% 3.471e+13 perf-stat.total.instructions
Disclaimer:
Results have been estimated based on internal Intel analysis and are provided
for informational purposes only. Any difference in system hardware or software
design or configuration may affect actual performance.
--
0-DAY CI Kernel Test Service
https://github.com/intel/lkp-tests
View attachment "config-6.3.0-rc3-00015-gf153e964b7d2" of type "text/plain" (157282 bytes)
View attachment "job-script" of type "text/plain" (8275 bytes)
View attachment "job.yaml" of type "text/plain" (5926 bytes)
View attachment "reproduce" of type "text/plain" (384 bytes)
Powered by blists - more mailing lists