[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <20200612151907.GA11611@vingu-book>
Date: Fri, 12 Jun 2020 17:19:07 +0200
From: Vincent Guittot <vincent.guittot@...aro.org>
To: Xing Zhengjun <zhengjun.xing@...ux.intel.com>
Cc: kernel test robot <rong.a.chen@...el.com>,
Ingo Molnar <mingo@...nel.org>,
Mel Gorman <mgorman@...hsingularity.net>,
Peter Zijlstra <a.p.zijlstra@...llo.nl>,
Juri Lelli <juri.lelli@...hat.com>,
Steven Rostedt <rostedt@...dmis.org>,
Valentin Schneider <valentin.schneider@....com>,
Phil Auld <pauld@...hat.com>, Hillf Danton <hdanton@...a.com>,
LKML <linux-kernel@...r.kernel.org>, lkp@...ts.01.org
Subject: Re: [LKP] [sched/fair] 070f5e860e: reaim.jobs_per_min -10.5%
regression
Le vendredi 12 juin 2020 à 14:36:49 (+0800), Xing Zhengjun a écrit :
> Hi Vincent,
>
> We test the regression still existed in v5.7, do you have time to look at
> it? Thanks.
>
>
The commit 070f5e860ee2 moveis some cases from the state "group has spare capacity" to
to the state "group is overloaded". Typically when util_avg decreases significantly
after a migration but the group is in fact still overloaded.
The current rule uses a fix threshold but has the disavantage of possibly including
some cases with spare capacity but a high runnable_avg (because of tasks running
simultaneously as an example).
It looks like this benchmark is impacted by moving such cases from has_spare_capacity
to is_overloaded. I have a patch in my backlog that tries to fix the problem but I
never sent it because I failed to find a benchmark that will benefit from it.
This patch moves back some cases from overloaded state to "has spare capacity" state.
Could you make it a try ?
---
kernel/sched/fair.c | 12 ++++++++++--
1 file changed, 10 insertions(+), 2 deletions(-)
diff --git a/kernel/sched/fair.c b/kernel/sched/fair.c
index 0ed04d2a8959..c24f85969591 100644
--- a/kernel/sched/fair.c
+++ b/kernel/sched/fair.c
@@ -8213,10 +8213,14 @@ static inline int sg_imbalanced(struct sched_group *group)
static inline bool
group_has_capacity(unsigned int imbalance_pct, struct sg_lb_stats *sgs)
{
+ unsigned long imb;
+
if (sgs->sum_nr_running < sgs->group_weight)
return true;
- if ((sgs->group_capacity * imbalance_pct) <
+ imb = imbalance_pct-100;
+ imb = sgs->sum_nr_running * imb + 100;
+ if ((sgs->group_capacity * imb) <
(sgs->group_runnable * 100))
return false;
@@ -8238,6 +8242,8 @@ group_has_capacity(unsigned int imbalance_pct, struct sg_lb_stats *sgs)
static inline bool
group_is_overloaded(unsigned int imbalance_pct, struct sg_lb_stats *sgs)
{
+ unsigned long imb;
+
if (sgs->sum_nr_running <= sgs->group_weight)
return false;
@@ -8245,7 +8251,9 @@ group_is_overloaded(unsigned int imbalance_pct, struct sg_lb_stats *sgs)
(sgs->group_util * imbalance_pct))
return true;
- if ((sgs->group_capacity * imbalance_pct) <
+ imb = imbalance_pct-100;
+ imb = sgs->sum_nr_running * imb + 100;
+ if ((sgs->group_capacity * imb) <
(sgs->group_runnable * 100))
return true;
--
2.17.1
> =========================================================================================
> tbox_group/testcase/rootfs/kconfig/compiler/runtime/nr_task/debug-setup/test/cpufreq_governor/ucode:
>
> lkp-ivb-d04/reaim/debian-x86_64-20191114.cgz/x86_64-rhel-7.6/gcc-7/300s/100%/test/five_sec/performance/0x21
>
> commit:
> 9f68395333ad7f5bfe2f83473fed363d4229f11c
> 070f5e860ee2bf588c99ef7b4c202451faa48236
> v5.7
>
> 9f68395333ad7f5b 070f5e860ee2bf588c99ef7b4c2 v5.7
> ---------------- --------------------------- ---------------------------
> %stddev %change %stddev %change %stddev
> \ | \ | \
> 0.69 -10.3% 0.62 -9.1% 0.62
> reaim.child_systime
> 0.62 -1.0% 0.61 +0.5% 0.62
> reaim.child_utime
> 66870 -10.0% 60187 -7.6% 61787
> reaim.jobs_per_min
> 16717 -10.0% 15046 -7.6% 15446
> reaim.jobs_per_min_child
> 97.84 -1.1% 96.75 -0.4% 97.43
> reaim.jti
> 72000 -10.8% 64216 -8.3% 66000
> reaim.max_jobs_per_min
> 0.36 +10.6% 0.40 +7.8% 0.39
> reaim.parent_time
> 1.58 ± 2% +71.0% 2.70 ± 2% +26.9% 2.01 ± 2%
> reaim.std_dev_percent
> 0.00 ± 5% +110.4% 0.01 ± 3% +48.8% 0.01 ± 7%
> reaim.std_dev_time
> 50800 -2.4% 49600 -1.6% 50000
> reaim.workload
>
>
> On 3/19/2020 10:38 AM, kernel test robot wrote:
> > Greeting,
> >
> > FYI, we noticed a -10.5% regression of reaim.jobs_per_min due to commit:
> >
> >
> > commit: 070f5e860ee2bf588c99ef7b4c202451faa48236 ("sched/fair: Take into account runnable_avg to classify group")
> > https://git.kernel.org/cgit/linux/kernel/git/next/linux-next.git master
> >
> > in testcase: reaim
> > on test machine: 4 threads Intel(R) Core(TM) i3-3220 CPU @ 3.30GHz with 4G memory
> > with following parameters:
> >
> > runtime: 300s
> > nr_task: 100%
> > test: five_sec
> > cpufreq_governor: performance
> > ucode: 0x21
> >
> > test-description: REAIM is an updated and improved version of AIM 7 benchmark.
> > test-url: https://sourceforge.net/projects/re-aim-7/
> >
> >
> >
> > If you fix the issue, kindly add following tag
> > Reported-by: kernel test robot <rong.a.chen@...el.com>
> >
> >
> > Details are as below:
> > -------------------------------------------------------------------------------------------------->
> >
> >
> > To reproduce:
> >
> > git clone https://github.com/intel/lkp-tests.git
> > cd lkp-tests
> > bin/lkp install job.yaml # job file is attached in this email
> > bin/lkp run job.yaml
> >
> > =========================================================================================
> > compiler/cpufreq_governor/kconfig/nr_task/rootfs/runtime/tbox_group/test/testcase/ucode:
> > gcc-7/performance/x86_64-rhel-7.6/100%/debian-x86_64-20191114.cgz/300s/lkp-ivb-d04/five_sec/reaim/0x21
> >
> > commit:
> > 9f68395333 ("sched/pelt: Add a new runnable average signal")
> > 070f5e860e ("sched/fair: Take into account runnable_avg to classify group")
> >
> > 9f68395333ad7f5b 070f5e860ee2bf588c99ef7b4c2
> > ---------------- ---------------------------
> > fail:runs %reproduction fail:runs
> > | | |
> > 4:4 -18% 3:4 perf-profile.children.cycles-pp.error_entry
> > 3:4 -12% 3:4 perf-profile.self.cycles-pp.error_entry
> > %stddev %change %stddev
> > \ | \
> > 0.68 -10.4% 0.61 reaim.child_systime
> > 67235 -10.5% 60195 reaim.jobs_per_min
> > 16808 -10.5% 15048 reaim.jobs_per_min_child
> > 97.90 -1.2% 96.70 reaim.jti
> > 72000 -10.8% 64216 reaim.max_jobs_per_min
> > 0.36 +11.3% 0.40 reaim.parent_time
> > 1.56 ± 3% +79.1% 2.80 ± 6% reaim.std_dev_percent
> > 0.00 ± 7% +145.9% 0.01 ± 9% reaim.std_dev_time
> > 104276 -16.0% 87616 reaim.time.involuntary_context_switches
> > 15511157 -2.4% 15144312 reaim.time.minor_page_faults
> > 55.00 -7.3% 51.00 reaim.time.percent_of_cpu_this_job_got
> > 88.01 -12.4% 77.12 reaim.time.system_time
> > 79.97 -3.2% 77.38 reaim.time.user_time
> > 216380 -3.4% 208924 reaim.time.voluntary_context_switches
> > 50800 -2.4% 49600 reaim.workload
> > 30.40 ± 2% -4.7% 28.97 ± 2% boot-time.boot
> > 9.38 -0.7 8.66 ± 3% mpstat.cpu.all.sys%
> > 7452 +7.5% 8014 vmstat.system.cs
> > 1457802 ± 16% +49.3% 2176122 ± 13% cpuidle.C1.time
> > 48523684 +43.4% 69570233 ± 22% cpuidle.C1E.time
> > 806543 ± 2% +20.7% 973406 ± 11% cpuidle.C1E.usage
> > 14328 ± 6% +14.5% 16410 ± 8% cpuidle.POLL.time
> > 43300 ± 4% +13.5% 49150 ± 5% softirqs.CPU0.SCHED
> > 118751 -9.3% 107763 softirqs.CPU1.RCU
> > 41679 ± 3% +14.1% 47546 ± 4% softirqs.CPU1.SCHED
> > 42688 ± 3% +12.3% 47931 ± 4% softirqs.CPU2.SCHED
> > 41730 ± 2% +17.7% 49115 ± 4% softirqs.CPU3.SCHED
> > 169399 +14.4% 193744 ± 2% softirqs.SCHED
> > 3419 +1.0% 3453 proc-vmstat.nr_kernel_stack
> > 16365616 -1.8% 16077850 proc-vmstat.numa_hit
> > 16365616 -1.8% 16077850 proc-vmstat.numa_local
> > 93908 -1.6% 92389 proc-vmstat.pgactivate
> > 16269664 -3.9% 15629529 ± 2% proc-vmstat.pgalloc_normal
> > 15918803 -2.3% 15557936 proc-vmstat.pgfault
> > 16644610 -2.0% 16310898 proc-vmstat.pgfree
> > 20125 ±123% +161.7% 52662 ± 30% sched_debug.cfs_rq:/.load.min
> > 348749 ± 10% -11.2% 309562 ± 11% sched_debug.cfs_rq:/.load.stddev
> > 1096 ± 6% -14.4% 938.42 ± 7% sched_debug.cfs_rq:/.load_avg.max
> > 448.46 ± 8% -17.5% 370.19 ± 10% sched_debug.cfs_rq:/.load_avg.stddev
> > 117372 -10.2% 105432 sched_debug.cfs_rq:/.min_vruntime.avg
> > 135242 ± 4% -9.2% 122811 sched_debug.cfs_rq:/.min_vruntime.max
> > 0.53 ± 8% +17.6% 0.62 ± 6% sched_debug.cfs_rq:/.nr_running.avg
> > 29.79 ± 30% -51.0% 14.58 ± 35% sched_debug.cfs_rq:/.nr_spread_over.max
> > 10.21 ± 34% -59.7% 4.12 ± 52% sched_debug.cfs_rq:/.nr_spread_over.stddev
> > 78.25 ± 40% +3304.7% 2664 ± 94% sched_debug.cpu.curr->pid.min
> > 294309 ± 2% +34.3% 395172 ± 12% sched_debug.cpu.nr_switches.min
> > 9.58 ± 35% +84.8% 17.71 ± 40% sched_debug.cpu.nr_uninterruptible.max
> > -6.88 +120.6% -15.17 sched_debug.cpu.nr_uninterruptible.min
> > 6.41 ± 30% +95.2% 12.52 ± 33% sched_debug.cpu.nr_uninterruptible.stddev
> > 286185 +33.4% 381734 ± 13% sched_debug.cpu.sched_count.min
> > 180416 +11.0% 200247 sched_debug.cpu.sched_goidle.avg
> > 116264 ± 3% +44.6% 168090 ± 15% sched_debug.cpu.sched_goidle.min
> > 476.00 ± 8% +92.4% 915.75 ± 3% interrupts.CAL:Function_call_interrupts
> > 110.50 ± 24% +101.1% 222.25 ± 4% interrupts.CPU0.CAL:Function_call_interrupts
> > 1381 ± 29% +23.7% 1709 ± 26% interrupts.CPU0.NMI:Non-maskable_interrupts
> > 1381 ± 29% +23.7% 1709 ± 26% interrupts.CPU0.PMI:Performance_monitoring_interrupts
> > 3319 ± 9% +50.4% 4991 ± 2% interrupts.CPU0.RES:Rescheduling_interrupts
> > 41.25 ± 30% +274.5% 154.50 ± 15% interrupts.CPU0.TLB:TLB_shootdowns
> > 116.25 ± 23% +96.1% 228.00 ± 16% interrupts.CPU1.CAL:Function_call_interrupts
> > 1183 ± 10% +43.1% 1692 ± 23% interrupts.CPU1.NMI:Non-maskable_interrupts
> > 1183 ± 10% +43.1% 1692 ± 23% interrupts.CPU1.PMI:Performance_monitoring_interrupts
> > 3335 ± 7% +60.4% 5350 ± 5% interrupts.CPU1.RES:Rescheduling_interrupts
> > 36.25 ± 30% +344.1% 161.00 ± 8% interrupts.CPU1.TLB:TLB_shootdowns
> > 131.25 ± 11% +81.1% 237.75 ± 11% interrupts.CPU2.CAL:Function_call_interrupts
> > 3247 ± 2% +62.4% 5274 interrupts.CPU2.RES:Rescheduling_interrupts
> > 34.50 ± 36% +357.2% 157.75 ± 7% interrupts.CPU2.TLB:TLB_shootdowns
> > 118.00 ± 13% +93.0% 227.75 ± 9% interrupts.CPU3.CAL:Function_call_interrupts
> > 3155 ± 4% +68.7% 5322 ± 3% interrupts.CPU3.RES:Rescheduling_interrupts
> > 38.50 ± 16% +303.9% 155.50 ± 3% interrupts.CPU3.TLB:TLB_shootdowns
> > 13057 ± 2% +60.4% 20939 interrupts.RES:Rescheduling_interrupts
> > 150.50 ± 27% +317.8% 628.75 ± 3% interrupts.TLB:TLB_shootdowns
> > 2.00 +0.1 2.09 ± 3% perf-stat.i.branch-miss-rate%
> > 10.26 +1.1 11.36 ± 7% perf-stat.i.cache-miss-rate%
> > 2009706 ± 2% +5.4% 2117525 ± 3% perf-stat.i.cache-misses
> > 16867421 -4.5% 16106908 perf-stat.i.cache-references
> > 7514 +7.6% 8083 perf-stat.i.context-switches
> > 1.51 -3.0% 1.47 perf-stat.i.cpi
> > 2.523e+09 ± 3% -8.8% 2.301e+09 ± 2% perf-stat.i.cpu-cycles
> > 124.54 +157.8% 321.08 perf-stat.i.cpu-migrations
> > 1842 ± 10% -18.6% 1498 ± 6% perf-stat.i.cycles-between-cache-misses
> > 752585 ± 2% -4.1% 721714 perf-stat.i.dTLB-store-misses
> > 590441 +2.7% 606399 perf-stat.i.iTLB-load-misses
> > 68766 +4.0% 71488 ± 2% perf-stat.i.iTLB-loads
> > 1.847e+09 ± 3% -4.7% 1.76e+09 ± 2% perf-stat.i.instructions
> > 3490 ± 4% -8.5% 3195 ± 3% perf-stat.i.instructions-per-iTLB-miss
> > 0.68 +3.7% 0.70 perf-stat.i.ipc
> > 51861 -2.1% 50797 perf-stat.i.minor-faults
> > 51861 -2.1% 50797 perf-stat.i.page-faults
> > 2.68 ± 2% +0.1 2.78 perf-stat.overall.branch-miss-rate%
> > 11.91 +1.2 13.14 ± 2% perf-stat.overall.cache-miss-rate%
> > 1.37 -4.3% 1.31 perf-stat.overall.cpi
> > 1255 -13.4% 1087 ± 2% perf-stat.overall.cycles-between-cache-misses
> > 3127 ± 3% -7.2% 2901 ± 2% perf-stat.overall.instructions-per-iTLB-miss
> > 0.73 +4.5% 0.76 perf-stat.overall.ipc
> > 2002763 ± 2% +5.4% 2110303 ± 3% perf-stat.ps.cache-misses
> > 16809816 -4.5% 16051656 perf-stat.ps.cache-references
> > 7489 +7.6% 8055 perf-stat.ps.context-switches
> > 2.514e+09 ± 3% -8.8% 2.293e+09 ± 2% perf-stat.ps.cpu-cycles
> > 124.12 +157.8% 319.95 perf-stat.ps.cpu-migrations
> > 750010 ± 2% -4.1% 719223 perf-stat.ps.dTLB-store-misses
> > 588424 +2.7% 604314 perf-stat.ps.iTLB-load-misses
> > 68533 +4.0% 71246 ± 2% perf-stat.ps.iTLB-loads
> > 1.841e+09 ± 3% -4.7% 1.754e+09 ± 2% perf-stat.ps.instructions
> > 51683 -2.1% 50622 perf-stat.ps.minor-faults
> > 51683 -2.1% 50622 perf-stat.ps.page-faults
> > 5.577e+11 ± 3% -5.1% 5.292e+11 ± 2% perf-stat.total.instructions
> > 7.35 ± 17% -2.7 4.60 ± 10% perf-profile.calltrace.cycles-pp.serial8250_console_putchar.uart_console_write.serial8250_console_write.console_unlock.vprintk_emit
> > 7.74 ± 20% -2.7 5.00 ± 6% perf-profile.calltrace.cycles-pp.wait_for_xmitr.serial8250_console_putchar.uart_console_write.serial8250_console_write.console_unlock
> > 10.14 ± 8% -2.7 7.44 ± 6% perf-profile.calltrace.cycles-pp.new_sync_write.vfs_write.ksys_write.do_syscall_64.entry_SYSCALL_64_after_hwframe
> > 10.66 ± 8% -2.6 8.07 ± 8% perf-profile.calltrace.cycles-pp.vfs_write.ksys_write.do_syscall_64.entry_SYSCALL_64_after_hwframe.write
> > 7.10 ± 17% -2.4 4.69 ± 7% perf-profile.calltrace.cycles-pp.write._fini
> > 7.10 ± 17% -2.4 4.69 ± 7% perf-profile.calltrace.cycles-pp._fini
> > 7.09 ± 17% -2.4 4.69 ± 7% perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.write._fini
> > 7.09 ± 17% -2.4 4.69 ± 7% perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.write._fini
> > 7.09 ± 17% -2.4 4.69 ± 7% perf-profile.calltrace.cycles-pp.ksys_write.do_syscall_64.entry_SYSCALL_64_after_hwframe.write._fini
> > 7.09 ± 17% -2.4 4.69 ± 7% perf-profile.calltrace.cycles-pp.devkmsg_write.new_sync_write.vfs_write.ksys_write.do_syscall_64
> > 7.09 ± 17% -2.4 4.69 ± 7% perf-profile.calltrace.cycles-pp.vprintk_emit.devkmsg_emit.devkmsg_write.new_sync_write.vfs_write
> > 7.09 ± 17% -2.4 4.69 ± 7% perf-profile.calltrace.cycles-pp.devkmsg_emit.devkmsg_write.new_sync_write.vfs_write.ksys_write
> > 6.20 ± 8% -2.1 4.08 ± 5% perf-profile.calltrace.cycles-pp.console_unlock.vprintk_emit.devkmsg_emit.devkmsg_write.new_sync_write
> > 5.15 ± 11% -1.8 3.38 ± 4% perf-profile.calltrace.cycles-pp.serial8250_console_write.console_unlock.vprintk_emit.devkmsg_emit.devkmsg_write
> > 5.05 ± 11% -1.7 3.31 ± 3% perf-profile.calltrace.cycles-pp.uart_console_write.serial8250_console_write.console_unlock.vprintk_emit.devkmsg_emit
> > 7.41 ± 10% -1.1 6.29 ± 5% perf-profile.calltrace.cycles-pp.__do_execve_file.__x64_sys_execve.do_syscall_64.entry_SYSCALL_64_after_hwframe.execve
> > 7.57 ± 11% -1.1 6.46 ± 5% perf-profile.calltrace.cycles-pp.execve
> > 7.46 ± 10% -1.1 6.37 ± 5% perf-profile.calltrace.cycles-pp.__x64_sys_execve.do_syscall_64.entry_SYSCALL_64_after_hwframe.execve
> > 7.46 ± 10% -1.1 6.37 ± 5% perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.execve
> > 7.46 ± 10% -1.1 6.37 ± 5% perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.execve
> > 7.03 ± 5% -1.1 5.95 ± 10% perf-profile.calltrace.cycles-pp.brk
> > 5.90 ± 7% -0.9 4.98 ± 10% perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.brk
> > 5.84 ± 7% -0.9 4.93 ± 9% perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.brk
> > 15.77 ± 2% -0.9 14.88 ± 2% perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe
> > 15.86 ± 2% -0.9 14.97 ± 2% perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe
> > 3.88 ± 6% -0.9 2.99 ± 5% perf-profile.calltrace.cycles-pp.kill
> > 1.70 ± 23% -0.8 0.90 ± 10% perf-profile.calltrace.cycles-pp.delay_tsc.wait_for_xmitr.serial8250_console_putchar.uart_console_write.serial8250_console_write
> > 4.88 ± 8% -0.8 4.08 ± 8% perf-profile.calltrace.cycles-pp.__x64_sys_brk.do_syscall_64.entry_SYSCALL_64_after_hwframe.brk
> > 2.39 ± 27% -0.7 1.67 ± 5% perf-profile.calltrace.cycles-pp.flush_old_exec.load_elf_binary.search_binary_handler.__do_execve_file.__x64_sys_execve
> > 2.29 ± 30% -0.7 1.59 ± 5% perf-profile.calltrace.cycles-pp.mmput.flush_old_exec.load_elf_binary.search_binary_handler.__do_execve_file
> > 2.27 ± 30% -0.7 1.58 ± 5% perf-profile.calltrace.cycles-pp.exit_mmap.mmput.flush_old_exec.load_elf_binary.search_binary_handler
> > 3.11 ± 5% -0.6 2.47 ± 9% perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.kill
> > 3.07 ± 5% -0.6 2.45 ± 9% perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.kill
> > 2.09 ± 18% -0.4 1.67 ± 3% perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.read
> > 2.82 ± 9% -0.4 2.40 ± 12% perf-profile.calltrace.cycles-pp.mmput.do_exit.do_group_exit.__x64_sys_exit_group.do_syscall_64
> > 2.80 ± 9% -0.4 2.38 ± 12% perf-profile.calltrace.cycles-pp.exit_mmap.mmput.do_exit.do_group_exit.__x64_sys_exit_group
> > 1.11 ± 33% -0.4 0.71 ± 10% perf-profile.calltrace.cycles-pp.unmap_vmas.exit_mmap.mmput.flush_old_exec.load_elf_binary
> > 1.05 ± 15% -0.4 0.69 ± 13% perf-profile.calltrace.cycles-pp.vt_console_print.console_unlock.vprintk_emit.devkmsg_emit.devkmsg_write
> > 1.03 ± 17% -0.4 0.68 ± 13% perf-profile.calltrace.cycles-pp.lf.vt_console_print.console_unlock.vprintk_emit.devkmsg_emit
> > 1.03 ± 17% -0.4 0.68 ± 13% perf-profile.calltrace.cycles-pp.con_scroll.lf.vt_console_print.console_unlock.vprintk_emit
> > 1.03 ± 17% -0.4 0.68 ± 13% perf-profile.calltrace.cycles-pp.fbcon_scroll.con_scroll.lf.vt_console_print.console_unlock
> > 0.96 ± 16% -0.3 0.66 ± 12% perf-profile.calltrace.cycles-pp.fbcon_putcs.fbcon_redraw.fbcon_scroll.con_scroll.lf
> > 1.85 ± 4% -0.3 1.58 ± 8% perf-profile.calltrace.cycles-pp.alloc_pages_vma.handle_pte_fault.__handle_mm_fault.handle_mm_fault.do_page_fault
> > 0.89 ± 15% -0.3 0.62 ± 12% perf-profile.calltrace.cycles-pp.kill_pid_info.kill_something_info.__x64_sys_kill.do_syscall_64.entry_SYSCALL_64_after_hwframe
> > 1.67 ± 5% -0.3 1.41 ± 9% perf-profile.calltrace.cycles-pp.__alloc_pages_nodemask.alloc_pages_vma.handle_pte_fault.__handle_mm_fault.handle_mm_fault
> > 1.02 ± 7% -0.3 0.77 ± 12% perf-profile.calltrace.cycles-pp.do_signal.exit_to_usermode_loop.do_syscall_64.entry_SYSCALL_64_after_hwframe.kill
> > 0.94 ± 16% -0.2 0.70 ± 5% perf-profile.calltrace.cycles-pp.clear_page_erms.prep_new_page.get_page_from_freelist.__alloc_pages_nodemask.alloc_pages_vma
> > 0.98 ± 16% -0.2 0.74 ± 7% perf-profile.calltrace.cycles-pp.prep_new_page.get_page_from_freelist.__alloc_pages_nodemask.alloc_pages_vma.handle_pte_fault
> > 1.03 ± 6% -0.2 0.79 ± 10% perf-profile.calltrace.cycles-pp.exit_to_usermode_loop.do_syscall_64.entry_SYSCALL_64_after_hwframe.kill
> > 1.00 ± 10% -0.2 0.77 ± 9% perf-profile.calltrace.cycles-pp.new_sync_read.vfs_read.ksys_read.do_syscall_64.entry_SYSCALL_64_after_hwframe
> > 0.87 ± 10% -0.2 0.66 ± 15% perf-profile.calltrace.cycles-pp.shmem_file_read_iter.new_sync_read.vfs_read.ksys_read.do_syscall_64
> > 1.41 ± 3% -0.2 1.23 ± 7% perf-profile.calltrace.cycles-pp.get_page_from_freelist.__alloc_pages_nodemask.alloc_pages_vma.handle_pte_fault.__handle_mm_fault
> > 1.88 ± 5% -0.1 1.73 perf-profile.calltrace.cycles-pp.__x64_sys_clone.do_syscall_64.entry_SYSCALL_64_after_hwframe
> > 1.87 ± 5% -0.1 1.73 perf-profile.calltrace.cycles-pp._do_fork.__x64_sys_clone.do_syscall_64.entry_SYSCALL_64_after_hwframe
> > 10.34 ± 11% +7.3 17.66 ± 8% perf-profile.calltrace.cycles-pp.cpuidle_enter.do_idle.cpu_startup_entry.start_secondary.secondary_startup_64
> > 10.18 ± 11% +7.3 17.52 ± 8% perf-profile.calltrace.cycles-pp.cpuidle_enter_state.cpuidle_enter.do_idle.cpu_startup_entry.start_secondary
> > 11.32 ± 9% +7.7 19.03 ± 8% perf-profile.calltrace.cycles-pp.do_idle.cpu_startup_entry.start_secondary.secondary_startup_64
> > 11.32 ± 9% +7.7 19.05 ± 8% perf-profile.calltrace.cycles-pp.cpu_startup_entry.start_secondary.secondary_startup_64
> > 11.32 ± 9% +7.7 19.05 ± 8% perf-profile.calltrace.cycles-pp.start_secondary.secondary_startup_64
> > 11.02 ± 5% +8.1 19.14 ± 7% perf-profile.calltrace.cycles-pp.intel_idle.cpuidle_enter_state.cpuidle_enter.do_idle.cpu_startup_entry
> > 16.04 ± 6% +9.1 25.17 ± 8% perf-profile.calltrace.cycles-pp.secondary_startup_64
> > 55.98 -7.0 48.94 ± 4% perf-profile.children.cycles-pp.entry_SYSCALL_64_after_hwframe
> > 55.67 -7.0 48.67 ± 4% perf-profile.children.cycles-pp.do_syscall_64
> > 10.60 ± 16% -3.3 7.30 ± 8% perf-profile.children.cycles-pp.vprintk_emit
> > 13.02 ± 7% -3.0 9.99 ± 7% perf-profile.children.cycles-pp.write
> > 9.92 ± 13% -2.8 7.08 ± 4% perf-profile.children.cycles-pp.console_unlock
> > 10.26 ± 8% -2.7 7.53 ± 6% perf-profile.children.cycles-pp.new_sync_write
> > 10.79 ± 8% -2.6 8.18 ± 8% perf-profile.children.cycles-pp.vfs_write
> > 10.95 ± 8% -2.6 8.36 ± 8% perf-profile.children.cycles-pp.ksys_write
> > 7.17 ± 16% -2.5 4.69 ± 7% perf-profile.children.cycles-pp.devkmsg_write
> > 7.17 ± 16% -2.5 4.69 ± 7% perf-profile.children.cycles-pp.devkmsg_emit
> > 8.65 ± 16% -2.4 6.21 ± 4% perf-profile.children.cycles-pp.serial8250_console_write
> > 8.53 ± 17% -2.4 6.11 ± 4% perf-profile.children.cycles-pp.uart_console_write
> > 7.13 ± 17% -2.4 4.71 ± 6% perf-profile.children.cycles-pp._fini
> > 8.46 ± 16% -2.4 6.07 ± 4% perf-profile.children.cycles-pp.wait_for_xmitr
> > 8.34 ± 16% -2.4 5.97 ± 4% perf-profile.children.cycles-pp.serial8250_console_putchar
> > 5.80 ± 16% -1.6 4.21 ± 6% perf-profile.children.cycles-pp.io_serial_in
> > 7.85 ± 10% -1.2 6.67 ± 5% perf-profile.children.cycles-pp.execve
> > 7.72 ± 11% -1.2 6.55 ± 5% perf-profile.children.cycles-pp.__do_execve_file
> > 5.19 ± 13% -1.1 4.05 ± 8% perf-profile.children.cycles-pp.mmput
> > 5.16 ± 13% -1.1 4.03 ± 8% perf-profile.children.cycles-pp.exit_mmap
> > 7.76 ± 10% -1.1 6.64 ± 5% perf-profile.children.cycles-pp.__x64_sys_execve
> > 7.11 ± 5% -1.1 6.01 ± 10% perf-profile.children.cycles-pp.brk
> > 3.92 ± 6% -0.9 3.03 ± 5% perf-profile.children.cycles-pp.kill
> > 2.63 ± 17% -0.8 1.85 perf-profile.children.cycles-pp.delay_tsc
> > 4.89 ± 8% -0.8 4.12 ± 8% perf-profile.children.cycles-pp.__x64_sys_brk
> > 2.48 ± 27% -0.7 1.74 ± 4% perf-profile.children.cycles-pp.flush_old_exec
> > 3.02 ± 12% -0.7 2.28 ± 12% perf-profile.children.cycles-pp.unmap_page_range
> > 3.15 ± 11% -0.7 2.40 ± 12% perf-profile.children.cycles-pp.unmap_vmas
> > 2.25 ± 19% -0.5 1.75 ± 11% perf-profile.children.cycles-pp.unmap_region
> > 1.27 ± 11% -0.4 0.86 ± 8% perf-profile.children.cycles-pp.vt_console_print
> > 1.24 ± 12% -0.4 0.85 ± 9% perf-profile.children.cycles-pp.lf
> > 1.24 ± 12% -0.4 0.85 ± 9% perf-profile.children.cycles-pp.con_scroll
> > 1.24 ± 12% -0.4 0.85 ± 9% perf-profile.children.cycles-pp.fbcon_scroll
> > 1.79 ± 9% -0.4 1.41 ± 4% perf-profile.children.cycles-pp.release_pages
> > 1.22 ± 11% -0.4 0.85 ± 9% perf-profile.children.cycles-pp.fbcon_redraw
> > 1.17 ± 12% -0.4 0.82 ± 10% perf-profile.children.cycles-pp.fbcon_putcs
> > 1.16 ± 13% -0.3 0.82 ± 10% perf-profile.children.cycles-pp.bit_putcs
> > 0.90 ± 16% -0.3 0.62 ± 12% perf-profile.children.cycles-pp.kill_pid_info
> > 0.95 ± 10% -0.3 0.68 ± 6% perf-profile.children.cycles-pp.drm_fb_helper_cfb_imageblit
> > 0.95 ± 11% -0.3 0.68 ± 6% perf-profile.children.cycles-pp.cfb_imageblit
> > 1.24 ± 7% -0.2 1.01 ± 6% perf-profile.children.cycles-pp.new_sync_read
> > 0.71 ± 4% -0.2 0.49 ± 23% perf-profile.children.cycles-pp.___perf_sw_event
> > 0.55 ± 31% -0.2 0.33 ± 16% perf-profile.children.cycles-pp.unlink_anon_vmas
> > 0.89 ± 11% -0.2 0.67 ± 15% perf-profile.children.cycles-pp.shmem_file_read_iter
> > 0.60 ± 20% -0.2 0.39 ± 20% perf-profile.children.cycles-pp.__send_signal
> > 1.06 ± 6% -0.2 0.85 ± 16% perf-profile.children.cycles-pp.pagevec_lru_move_fn
> > 0.88 -0.2 0.68 ± 6% perf-profile.children.cycles-pp.__perf_sw_event
> > 1.49 ± 5% -0.2 1.29 ± 7% perf-profile.children.cycles-pp.prepare_exit_to_usermode
> > 0.56 ± 12% -0.2 0.37 ± 11% perf-profile.children.cycles-pp.do_send_sig_info
> > 1.65 ± 8% -0.2 1.47 ± 4% perf-profile.children.cycles-pp.perf_event_mmap
> > 0.69 ± 2% -0.2 0.52 ± 16% perf-profile.children.cycles-pp.page_remove_rmap
> > 0.61 ± 5% -0.2 0.44 ± 15% perf-profile.children.cycles-pp.free_unref_page_list
> > 0.60 ± 6% -0.2 0.43 ± 15% perf-profile.children.cycles-pp.__vm_munmap
> > 0.77 ± 12% -0.2 0.62 ± 12% perf-profile.children.cycles-pp.__might_sleep
> > 0.39 ± 12% -0.2 0.24 ± 18% perf-profile.children.cycles-pp.time
> > 0.46 ± 14% -0.1 0.34 ± 14% perf-profile.children.cycles-pp.lru_add_drain_cpu
> > 0.57 ± 8% -0.1 0.47 ± 14% perf-profile.children.cycles-pp.shmem_undo_range
> > 0.41 ± 12% -0.1 0.30 ± 15% perf-profile.children.cycles-pp.copy_fpstate_to_sigframe
> > 0.76 ± 7% -0.1 0.67 ± 8% perf-profile.children.cycles-pp.__x64_sys_rt_sigreturn
> > 0.26 ± 16% -0.1 0.17 ± 17% perf-profile.children.cycles-pp.mark_page_accessed
> > 0.12 ± 47% -0.1 0.04 ±103% perf-profile.children.cycles-pp.sigaction
> > 0.23 ± 12% -0.1 0.15 ± 11% perf-profile.children.cycles-pp.__vm_enough_memory
> > 0.12 ± 18% -0.1 0.05 ±106% perf-profile.children.cycles-pp.__vsprintf_chk
> > 0.23 ± 20% -0.1 0.17 ± 13% perf-profile.children.cycles-pp.d_add
> > 0.13 ± 23% -0.1 0.07 ± 58% perf-profile.children.cycles-pp.fput_many
> > 0.13 ± 14% -0.1 0.08 ± 24% perf-profile.children.cycles-pp.vfs_unlink
> > 0.11 ± 20% -0.0 0.07 ± 7% perf-profile.children.cycles-pp.__update_load_avg_cfs_rq
> > 0.04 ± 63% +0.0 0.08 ± 23% perf-profile.children.cycles-pp.uncharge_page
> > 0.06 ± 22% +0.0 0.10 ± 36% perf-profile.children.cycles-pp.sched_exec
> > 0.44 ± 4% +0.0 0.48 ± 4% perf-profile.children.cycles-pp.close
> > 0.14 ± 22% +0.1 0.21 ± 17% perf-profile.children.cycles-pp.pick_next_task_fair
> > 0.10 ± 17% +0.1 0.17 ± 23% perf-profile.children.cycles-pp.__anon_vma_prepare
> > 0.00 +0.1 0.07 ± 24% perf-profile.children.cycles-pp.update_sd_lb_stats
> > 0.07 ± 34% +0.1 0.15 ± 42% perf-profile.children.cycles-pp.file_free_rcu
> > 0.15 ± 27% +0.1 0.23 ± 21% perf-profile.children.cycles-pp.__strcasecmp
> > 0.20 ± 21% +0.1 0.29 ± 8% perf-profile.children.cycles-pp.__pte_alloc
> > 0.14 ± 47% +0.1 0.23 ± 27% perf-profile.children.cycles-pp.update_blocked_averages
> > 0.09 ± 44% +0.1 0.19 ± 18% perf-profile.children.cycles-pp.schedule_idle
> > 0.00 +0.1 0.10 ± 33% perf-profile.children.cycles-pp.newidle_balance
> > 0.00 +0.1 0.10 ± 18% perf-profile.children.cycles-pp.__vmalloc_node_range
> > 0.21 ± 15% +0.1 0.32 ± 25% perf-profile.children.cycles-pp.__wake_up_common
> > 0.63 ± 8% +0.1 0.77 ± 6% perf-profile.children.cycles-pp.rcu_do_batch
> > 0.76 ± 14% +0.1 0.90 ± 9% perf-profile.children.cycles-pp.rcu_core
> > 0.07 ± 90% +0.2 0.27 ±109% perf-profile.children.cycles-pp.security_mmap_addr
> > 0.46 ± 26% +0.3 0.75 ± 13% perf-profile.children.cycles-pp.__sched_text_start
> > 11.32 ± 9% +7.7 19.05 ± 8% perf-profile.children.cycles-pp.start_secondary
> > 11.03 ± 5% +8.1 19.16 ± 7% perf-profile.children.cycles-pp.intel_idle
> > 14.78 ± 6% +8.5 23.24 ± 8% perf-profile.children.cycles-pp.cpuidle_enter
> > 14.76 ± 6% +8.5 23.24 ± 8% perf-profile.children.cycles-pp.cpuidle_enter_state
> > 16.04 ± 6% +9.1 25.17 ± 8% perf-profile.children.cycles-pp.secondary_startup_64
> > 16.04 ± 6% +9.1 25.17 ± 8% perf-profile.children.cycles-pp.cpu_startup_entry
> > 16.04 ± 6% +9.1 25.19 ± 8% perf-profile.children.cycles-pp.do_idle
> > 5.79 ± 16% -1.6 4.21 ± 6% perf-profile.self.cycles-pp.io_serial_in
> > 2.62 ± 17% -0.8 1.85 perf-profile.self.cycles-pp.delay_tsc
> > 5.11 ± 4% -0.6 4.56 ± 5% perf-profile.self.cycles-pp.do_syscall_64
> > 1.44 ± 6% -0.3 1.15 ± 5% perf-profile.self.cycles-pp.unmap_page_range
> > 0.94 ± 11% -0.3 0.68 ± 6% perf-profile.self.cycles-pp.cfb_imageblit
> > 0.65 ± 6% -0.2 0.42 ± 23% perf-profile.self.cycles-pp.___perf_sw_event
> > 1.42 ± 5% -0.2 1.22 ± 7% perf-profile.self.cycles-pp.prepare_exit_to_usermode
> > 0.65 ± 13% -0.2 0.47 ± 9% perf-profile.self.cycles-pp.do_page_fault
> > 0.65 ± 9% -0.1 0.52 ± 5% perf-profile.self.cycles-pp.release_pages
> > 0.24 ± 20% -0.1 0.15 ± 16% perf-profile.self.cycles-pp.mark_page_accessed
> > 0.16 ± 28% -0.1 0.08 ± 69% perf-profile.self.cycles-pp.free_unref_page_commit
> > 0.12 ± 24% -0.1 0.04 ± 59% perf-profile.self.cycles-pp.__do_munmap
> > 0.10 ± 24% -0.0 0.06 ± 7% perf-profile.self.cycles-pp.__update_load_avg_cfs_rq
> > 0.04 ± 57% +0.0 0.07 ± 19% perf-profile.self.cycles-pp.__sbrk
> > 0.04 ± 57% +0.0 0.08 ± 23% perf-profile.self.cycles-pp.update_load_avg
> > 0.04 ± 57% +0.0 0.08 ± 23% perf-profile.self.cycles-pp.uncharge_page
> > 0.26 ± 11% +0.1 0.39 ± 12% perf-profile.self.cycles-pp.copy_page
> > 0.49 ± 13% +0.1 0.63 ± 13% perf-profile.self.cycles-pp.get_page_from_freelist
> > 11.00 ± 5% +8.1 19.14 ± 7% perf-profile.self.cycles-pp.intel_idle
> >
> >
> > reaim.time.system_time
> > 90 +----------------------------------------------------------------------+
> > | +. .+ ++.++ .+ |
> > 88 |+.+ +.++. + +++ ++.++.+ +.++.+++.+++.+++. +.++ .+ + :.+++.+|
> > | + + + + ++.++ + |
> > 86 |-+ |
> > | |
> > 84 |-+ |
> > | |
> > 82 |-+ O |
> > |O OO OO O |
> > 80 |-+ |
> > | OO OOO O O O O |
> > 78 |-+ O O O O O OO O |
> > | O OOO O |
> > 76 +----------------------------------------------------------------------+
> > reaim.time.percent_of_cpu_this_job_got
> > 55 +--------------------------------------------------------------------+
> > | |
> > 54.5 |-+ |
> > 54 |-+ |
> > | |
> > 53.5 |-+ |
> > | |
> > 53 |-+OOO OOO |
> > | |
> > 52.5 |-+ |
> > 52 |O+ OOO OOO OOO OOO OO |
> > | |
> > 51.5 |-+ |
> > | |
> > 51 +--------------------------------------------------------------------+
> > reaim.parent_time
> > 0.405 +-------------------------------------------------------------------+
> > 0.4 |-+ O |
> > | O O OOO OO |
> > 0.395 |-+ |
> > 0.39 |-+ |
> > 0.385 |-+ OOOO OOO OOO OOO OOOO |
> > 0.38 |O+OOO |
> > | |
> > 0.375 |-+ |
> > 0.37 |-+ |
> > 0.365 |-+ |
> > 0.36 |-+ |
> > | +. .+ +. .+ +. + +. +++. + .+|
> > 0.355 |+. +.+ ++.++ +++. ++.++++ + +++ + +++ + ++.++ + +. : + + |
> > 0.35 +-------------------------------------------------------------------+
> > reaim.child_systime
> > 0.69 +--------------------------------------------------------------------+
> > | +.+ ++. ++.+ .++ ++ +++.+ .++ .+++. + .+ : + .+|
> > 0.68 |+.++ ++.+ + ++ +.+ ++ + + +.+++ +.+ + |
> > 0.67 |-+ |
> > | |
> > 0.66 |-+ |
> > 0.65 |-+ |
> > | O |
> > 0.64 |O+O OOO |
> > 0.63 |-+ O |
> > | O O O O O |
> > 0.62 |-+ OO O O O O O |
> > 0.61 |-+ O OO O OO OOOO |
> > | |
> > 0.6 +--------------------------------------------------------------------+
> > reaim.jobs_per_min
> > 69000 +-------------------------------------------------------------------+
> > 68000 |-.++ + + .++ .+ .++ + .+++.+ |
> > |+ +.+ ++.+++.+ + + +++.+++.+++.+++.++++ +.+ + + +.+ +. |
> > 67000 |-+ + + +|
> > 66000 |-+ |
> > | |
> > 65000 |-+ |
> > 64000 |-+ |
> > 63000 |-+ |
> > |O OOO O O O O |
> > 62000 |-+ O O OOO O OOO OOOO |
> > 61000 |-+ |
> > | O OOO OO |
> > 60000 |-+ OO |
> > 59000 +-------------------------------------------------------------------+
> > reaim.jobs_per_min_child
> > 17500 +-------------------------------------------------------------------+
> > | |
> > 17000 |-.+++. + + .+++.+ .++ + .+++.+ |
> > |+ + ++.+++.+ + +++.+++.+++.+++.++++ +.+ + + +.+++.+|
> > | + |
> > 16500 |-+ |
> > | |
> > 16000 |-+ |
> > | |
> > 15500 |O+OOO OOOO OOO OOO OOO OOOO |
> > | |
> > | O O OO |
> > 15000 |-+ OOO O |
> > | |
> > 14500 +-------------------------------------------------------------------+
> > reaim.max_jobs_per_min
> > 76000 +-------------------------------------------------------------------+
> > | |
> > 74000 |-+ + + + |
> > | :: : :: |
> > | : : : :: : |
> > 72000 |+.++ ++++.+++.+++.+++.++++.+++.+++.+++.++++.+++.+++.+ + +++.+++.+|
> > | |
> > 70000 |-+ |
> > | |
> > 68000 |-+ O |
> > | |
> > | |
> > 66000 |O+O O OOOO OOO OOO OOO OOOO |
> > | |
> > 64000 +-------------------------------------------------------------------+
> > [*] bisect-good sample
> > [O] bisect-bad sample
> >
> >
> >
> > Disclaimer:
> > Results have been estimated based on internal Intel analysis and are provided
> > for informational purposes only. Any difference in system hardware or software
> > design or configuration may affect actual performance.
> >
> >
> > Thanks,
> > Rong Chen
> >
> >
> > _______________________________________________
> > LKP mailing list -- lkp@...ts.01.org
> > To unsubscribe send an email to lkp-leave@...ts.01.org
> >
>
> --
> Zhengjun Xing
Powered by blists - more mailing lists