[<prev] [next>] [<thread-prev] [day] [month] [year] [list]
Message-ID: <CALZhoSQb9mOm0+rYhgEfF74WkVW2Vkdb+ZE2cAECMs3Z9cqZ+w@mail.gmail.com>
Date: Wed, 3 Jul 2013 20:46:11 +0800
From: Lei Wen <adrian.wenl@...il.com>
To: Lei Wen <leiwen@...vell.com>
Cc: Paul Turner <pjt@...gle.com>, Alex Shi <alex.shi@...el.com>,
Peter Zijlstra <peterz@...radead.org>,
Ingo Molnar <mingo@...e.hu>, mingo@...hat.com,
Kamalesh Babulal <kamalesh@...ux.vnet.ibm.com>,
linux-kernel@...r.kernel.org
Subject: Re: [V3 1/2] sched: add trace events for task and rq usage tracking
Hi Peter,
Do you have some further suggestion for this patch? :)
Thanks,
Lei
On Tue, Jul 2, 2013 at 8:15 PM, Lei Wen <leiwen@...vell.com> wrote:
> Since we could track task in the entity level now, we may want to
> investigate tasks' running status by recording the trace info, so that
> could make some tuning if needed.
>
> Signed-off-by: Lei Wen <leiwen@...vell.com>
> Cc: Alex Shi <alex.shi@...el.com>
> Cc: Peter Zijlstra <peterz@...radead.org>
> Cc: Kamalesh Babulal <kamalesh@...ux.vnet.ibm.com>
> ---
> include/trace/events/sched.h | 76 ++++++++++++++++++++++++++++++++++++++++++
> kernel/sched/fair.c | 15 +++++++--
> 2 files changed, 89 insertions(+), 2 deletions(-)
>
> diff --git a/include/trace/events/sched.h b/include/trace/events/sched.h
> index e5586ca..768b398 100644
> --- a/include/trace/events/sched.h
> +++ b/include/trace/events/sched.h
> @@ -430,6 +430,82 @@ TRACE_EVENT(sched_pi_setprio,
> __entry->oldprio, __entry->newprio)
> );
>
> +#ifdef CONFIG_SMP
> +TRACE_EVENT(sched_task_weighted_load,
> +
> + TP_PROTO(struct sched_entity *se),
> +
> + TP_ARGS(se),
> +
> + TP_STRUCT__entry(
> + __field(pid_t, pid)
> + __field(int, cpu)
> + __field(unsigned long, load)
> + __field(unsigned long, weight)
> + ),
> +
> + TP_fast_assign(
> + __entry->pid = container_of(se, struct task_struct, se)->pid;
> + __entry->cpu = se->cfs_rq->rq->cpu;
> + __entry->load = se->avg.load_avg_contrib;
> + __entry->weight= se->load.weight;
> + ),
> +
> + TP_printk("cpu=%d pid=%d load=%lu weight=%lu",
> + __entry->cpu, __entry->pid,
> + __entry->load, __entry->weight)
> +);
> +
> +TRACE_EVENT(sched_cfs_rq_runnable_load,
> +
> + TP_PROTO(struct cfs_rq *cfs_rq),
> +
> + TP_ARGS(cfs_rq),
> +
> + TP_STRUCT__entry(
> + __field(int, cpu)
> + __field(unsigned long, load)
> + __field(unsigned long, total)
> + ),
> +
> + TP_fast_assign(
> + __entry->cpu = cfs_rq->rq->cpu;
> + __entry->load = cfs_rq->runnable_load_avg;
> + __entry->total = cfs_rq->load.weight;
> + ),
> +
> + TP_printk("cpu=%d avg=%lu total=%lu",
> + __entry->cpu,
> + __entry->load,
> + __entry->total)
> +);
> +
> +TRACE_EVENT(sched_cfs_rq_blocked_load,
> +
> + TP_PROTO(struct cfs_rq *cfs_rq),
> +
> + TP_ARGS(cfs_rq),
> +
> + TP_STRUCT__entry(
> + __field(int, cpu)
> + __field(unsigned long, load)
> + __field(unsigned long, total)
> + ),
> +
> + TP_fast_assign(
> + __entry->cpu = cfs_rq->rq->cpu;
> + __entry->load = cfs_rq->blocked_load_avg;
> + __entry->total = cfs_rq->blocked_load_avg
> + + cfs_rq->runnable_load_avg;
> + ),
> +
> + TP_printk("cpu=%d avg=%lu total=%lu",
> + __entry->cpu,
> + __entry->load,
> + __entry->total)
> +);
> +#endif
> +
> #endif /* _TRACE_SCHED_H */
>
> /* This part must be outside protection */
> diff --git a/kernel/sched/fair.c b/kernel/sched/fair.c
> index f77f9c5..2290469 100644
> --- a/kernel/sched/fair.c
> +++ b/kernel/sched/fair.c
> @@ -1346,6 +1346,7 @@ static inline u64 __synchronize_entity_decay(struct sched_entity *se)
> return 0;
>
> se->avg.load_avg_contrib = decay_load(se->avg.load_avg_contrib, decays);
> + trace_sched_task_weighted_load(se);
> se->avg.decay_count = 0;
>
> return decays;
> @@ -1445,6 +1446,7 @@ static inline void __update_task_entity_contrib(struct sched_entity *se)
> contrib = se->avg.runnable_avg_sum * scale_load_down(se->load.weight);
> contrib /= (se->avg.runnable_avg_period + 1);
> se->avg.load_avg_contrib = scale_load(contrib);
> + trace_sched_task_weighted_load(se);
> }
>
> /* Compute the current contribution to load_avg by se, return any delta */
> @@ -1498,10 +1500,13 @@ static inline void update_entity_load_avg(struct sched_entity *se,
> if (!update_cfs_rq)
> return;
>
> - if (se->on_rq)
> + if (se->on_rq) {
> cfs_rq->runnable_load_avg += contrib_delta;
> - else
> + trace_sched_cfs_rq_runnable_load(cfs_rq);
> + } else {
> subtract_blocked_load_contrib(cfs_rq, -contrib_delta);
> + trace_sched_cfs_rq_blocked_load(cfs_rq);
> + }
> }
>
> /*
> @@ -1531,6 +1536,7 @@ static void update_cfs_rq_blocked_load(struct cfs_rq *cfs_rq, int force_update)
> }
>
> __update_cfs_rq_tg_load_contrib(cfs_rq, force_update);
> + trace_sched_cfs_rq_blocked_load(cfs_rq);
> }
>
> static inline void update_rq_runnable_avg(struct rq *rq, int runnable)
> @@ -1584,10 +1590,12 @@ static inline void enqueue_entity_load_avg(struct cfs_rq *cfs_rq,
> /* migrated tasks did not contribute to our blocked load */
> if (wakeup) {
> subtract_blocked_load_contrib(cfs_rq, se->avg.load_avg_contrib);
> + trace_sched_cfs_rq_blocked_load(cfs_rq);
> update_entity_load_avg(se, 0);
> }
>
> cfs_rq->runnable_load_avg += se->avg.load_avg_contrib;
> + trace_sched_cfs_rq_runnable_load(cfs_rq);
> /* we force update consideration on load-balancer moves */
> update_cfs_rq_blocked_load(cfs_rq, !wakeup);
> }
> @@ -1608,6 +1616,7 @@ static inline void dequeue_entity_load_avg(struct cfs_rq *cfs_rq,
> cfs_rq->runnable_load_avg -= se->avg.load_avg_contrib;
> if (sleep) {
> cfs_rq->blocked_load_avg += se->avg.load_avg_contrib;
> + trace_sched_cfs_rq_blocked_load(cfs_rq);
> se->avg.decay_count = atomic64_read(&cfs_rq->decay_counter);
> } /* migrations, e.g. sleep=0 leave decay_count == 0 */
> }
> @@ -5894,6 +5903,7 @@ static void switched_from_fair(struct rq *rq, struct task_struct *p)
> __synchronize_entity_decay(&p->se);
> subtract_blocked_load_contrib(cfs_rq,
> p->se.avg.load_avg_contrib);
> + trace_sched_cfs_rq_blocked_load(cfs_rq);
> }
> #endif
> }
> @@ -5994,6 +6004,7 @@ static void task_move_group_fair(struct task_struct *p, int on_rq)
> */
> p->se.avg.decay_count = atomic64_read(&cfs_rq->decay_counter);
> cfs_rq->blocked_load_avg += p->se.avg.load_avg_contrib;
> + trace_sched_cfs_rq_blocked_load(cfs_rq);
> #endif
> }
> }
> --
> 1.7.10.4
>
> --
> To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
> the body of a message to majordomo@...r.kernel.org
> More majordomo info at http://vger.kernel.org/majordomo-info.html
> Please read the FAQ at http://www.tux.org/lkml/
--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/
Powered by blists - more mailing lists