lists.openwall.net | lists / announce owl-users owl-dev john-users john-dev passwdqc-users yescrypt popa3d-users / oss-security kernel-hardening musl sabotage tlsify passwords / crypt-dev xvendor / Bugtraq Full-Disclosure linux-kernel linux-netdev linux-ext4 linux-hardening linux-cve-announce PHC | |
Open Source and information security mailing list archives
| ||
|
Message-Id: <1410826030-26301-1-git-send-email-morten.rasmussen@arm.com> Date: Tue, 16 Sep 2014 01:07:10 +0100 From: Morten Rasmussen <morten.rasmussen@....com> To: peterz@...radead.org, mingo@...nel.org Cc: linux-kernel@...r.kernel.org, Morten Rasmussen <morten.rasmussen@....com>, Paul Turner <pjt@...gle.com>, Ben Segall <bsegall@...gle.com> Subject: [PATCH] sched: Update task group load contributions during active load-balancing Task group load-contributions are not updated when tasks belonging to task groups are migrated by active load-balancing. If no other task belonging to the same task group is already queued at the destination cpu the group sched_entity will be enqueued with load_avg_contrib=0. Hence, weighted_cpuload() won't reflect the newly added load. The load may remain invisible until the next tick, when the sched_entity load_avg_contrib and task group contributions are reevaluated. The enqueue loop for_each_entity(se) { enqueue_entity(cfs_rq,se) ... enqueue_entity_load_avg(cfs_rq,se) ... update_entity_load_avg(se) ... __update_entity_load_avg_contrib(se) ... ... update_cfs_rq_blocked_load(cfs_rq) ... __update_cfs_rq_tg_load_contrib(cfs_rq) ... } currently skips __update_entity_load_avg_contrib() and __update_cfs_rq_tg_load_contrib() for group entities for active load-balance migrations. The former updates the sched_entity load_avg_contrib, and the latter updates the task group contribution which is needed by the former. They must both be called to ensure that load doesn't temporarily disappear. cc: Paul Turner <pjt@...gle.com> cc: Ben Segall <bsegall@...gle.com> Signed-off-by: Morten Rasmussen <morten.rasmussen@....com> --- kernel/sched/fair.c | 7 ++++++- 1 file changed, 6 insertions(+), 1 deletion(-) diff --git a/kernel/sched/fair.c b/kernel/sched/fair.c index be9e97b..2b6e2eb 100644 --- a/kernel/sched/fair.c +++ b/kernel/sched/fair.c @@ -2521,7 +2521,8 @@ static inline void update_entity_load_avg(struct sched_entity *se, else now = cfs_rq_clock_task(group_cfs_rq(se)); - if (!__update_entity_runnable_avg(now, &se->avg, se->on_rq)) + if (!__update_entity_runnable_avg(now, &se->avg, se->on_rq) && + entity_is_task(se)) return; contrib_delta = __update_entity_load_avg_contrib(se); @@ -2609,6 +2610,10 @@ static inline void enqueue_entity_load_avg(struct cfs_rq *cfs_rq, cfs_rq->runnable_load_avg += se->avg.load_avg_contrib; /* we force update consideration on load-balancer moves */ update_cfs_rq_blocked_load(cfs_rq, !wakeup); + + /* We force update group contributions on load-balancer moves */ + if (wakeup && !entity_is_task(se)) + __update_cfs_rq_tg_load_contrib(cfs_rq, 0); } /* -- 1.7.9.5 -- To unsubscribe from this list: send the line "unsubscribe linux-kernel" in the body of a message to majordomo@...r.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html Please read the FAQ at http://www.tux.org/lkml/
Powered by blists - more mailing lists