[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-Id: <1474892393-5095-7-git-send-email-vincent.guittot@linaro.org>
Date: Mon, 26 Sep 2016 14:19:52 +0200
From: Vincent Guittot <vincent.guittot@...aro.org>
To: peterz@...radead.org, mingo@...nel.org,
linux-kernel@...r.kernel.org, dietmar.eggemann@....com,
kernellwp@...il.com
Cc: yuyang.du@...el.com, Morten.Rasmussen@....com,
linaro-kernel@...ts.linaro.org, pjt@...gle.com, bsegall@...gle.com,
Vincent Guittot <vincent.guittot@...aro.org>
Subject: [PATCH 6/7 v4] sched: fix task group initialization
The moves of tasks are now propagated down to root and the utilization
of cfs_rq reflects reality so it doesn't need to be estimated at init.
Signed-off-by: Vincent Guittot <vincent.guittot@...aro.org>
---
kernel/sched/fair.c | 2 +-
1 file changed, 1 insertion(+), 1 deletion(-)
diff --git a/kernel/sched/fair.c b/kernel/sched/fair.c
index bd3b6b9..7acb77fb 100644
--- a/kernel/sched/fair.c
+++ b/kernel/sched/fair.c
@@ -8903,7 +8903,7 @@ void online_fair_sched_group(struct task_group *tg)
se = tg->se[i];
raw_spin_lock_irq(&rq->lock);
- post_init_entity_util_avg(se);
+ attach_entity_cfs_rq(se);
sync_throttle(tg, i);
raw_spin_unlock_irq(&rq->lock);
}
--
1.9.1
Powered by blists - more mailing lists