[<prev] [next>] [<thread-prev] [day] [month] [year] [list]
Message-ID: <tip-6fe0ce1eb04f99a1eb1eb6e7f775666966cf6c80@git.kernel.org>
Date: Tue, 13 Feb 2018 03:28:13 -0800
From: tip-bot for Wen Yang <tipbot@...or.com>
To: linux-tip-commits@...r.kernel.org
Cc: mingo@...nel.org, hpa@...or.com, peterz@...radead.org,
tglx@...utronix.de, jiang.biao2@....com.cn,
linux-kernel@...r.kernel.org, torvalds@...ux-foundation.org,
wen.yang99@....com.cn
Subject: [tip:sched/urgent] sched/deadline: Make update_curr_dl() more
accurate
Commit-ID: 6fe0ce1eb04f99a1eb1eb6e7f775666966cf6c80
Gitweb: https://git.kernel.org/tip/6fe0ce1eb04f99a1eb1eb6e7f775666966cf6c80
Author: Wen Yang <wen.yang99@....com.cn>
AuthorDate: Tue, 6 Feb 2018 09:55:48 +0800
Committer: Ingo Molnar <mingo@...nel.org>
CommitDate: Tue, 13 Feb 2018 11:44:40 +0100
sched/deadline: Make update_curr_dl() more accurate
rq->clock_task may be updated between the two calls of
rq_clock_task() in update_curr_dl(). Calling rq_clock_task() only
once makes it more accurate and efficient, taking update_curr() as
reference.
Suggested-by: Peter Zijlstra <peterz@...radead.org>
Signed-off-by: Wen Yang <wen.yang99@....com.cn>
Signed-off-by: Peter Zijlstra (Intel) <peterz@...radead.org>
Reviewed-by: Jiang Biao <jiang.biao2@....com.cn>
Cc: Linus Torvalds <torvalds@...ux-foundation.org>
Cc: Thomas Gleixner <tglx@...utronix.de>
Cc: zhong.weidong@....com.cn
Link: http://lkml.kernel.org/r/1517882148-44599-1-git-send-email-wen.yang99@zte.com.cn
Signed-off-by: Ingo Molnar <mingo@...nel.org>
---
kernel/sched/deadline.c | 6 ++++--
1 file changed, 4 insertions(+), 2 deletions(-)
diff --git a/kernel/sched/deadline.c b/kernel/sched/deadline.c
index 9bb0e0c..9df0978 100644
--- a/kernel/sched/deadline.c
+++ b/kernel/sched/deadline.c
@@ -1153,6 +1153,7 @@ static void update_curr_dl(struct rq *rq)
struct sched_dl_entity *dl_se = &curr->dl;
u64 delta_exec, scaled_delta_exec;
int cpu = cpu_of(rq);
+ u64 now;
if (!dl_task(curr) || !on_dl_rq(dl_se))
return;
@@ -1165,7 +1166,8 @@ static void update_curr_dl(struct rq *rq)
* natural solution, but the full ramifications of this
* approach need further study.
*/
- delta_exec = rq_clock_task(rq) - curr->se.exec_start;
+ now = rq_clock_task(rq);
+ delta_exec = now - curr->se.exec_start;
if (unlikely((s64)delta_exec <= 0)) {
if (unlikely(dl_se->dl_yielded))
goto throttle;
@@ -1178,7 +1180,7 @@ static void update_curr_dl(struct rq *rq)
curr->se.sum_exec_runtime += delta_exec;
account_group_exec_runtime(curr, delta_exec);
- curr->se.exec_start = rq_clock_task(rq);
+ curr->se.exec_start = now;
cgroup_account_cputime(curr, delta_exec);
sched_rt_avg_update(rq, delta_exec);
Powered by blists - more mailing lists