[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <20140729123825.GB3935@laptop>
Date: Tue, 29 Jul 2014 14:38:25 +0200
From: Peter Zijlstra <peterz@...radead.org>
To: Kirill Tkhai <ktkhai@...allels.com>
Cc: Kirill Tkhai <tkhai@...dex.ru>, linux-kernel@...r.kernel.org,
nicolas.pitre@...aro.org, pjt@...gle.com, oleg@...hat.com,
rostedt@...dmis.org, umgwanakikbuti@...il.com,
tim.c.chen@...ux.intel.com, mingo@...nel.org
Subject: Re: [PATCH v2 2/5] sched: Teach scheduler to understand
ONRQ_MIGRATING state
On Tue, Jul 29, 2014 at 01:53:02PM +0400, Kirill Tkhai wrote:
> From: Kirill Tkhai <ktkhai@...allels.com>
>
> sched: Teach scheduler to understand ONRQ_MIGRATING state
>
> This is new on_rq state for the cases when task is migrating
> from one src_rq to another dst_rq, and there is no necessity
> to have both RQs locked at the same time.
>
> We will use the state this way:
>
> raw_spin_lock(&src_rq->lock);
> dequeue_task(src_rq, p, 0);
> p->on_rq = ONRQ_MIGRATING;
> set_task_cpu(p, dst_cpu);
> raw_spin_unlock(&src_rq->lock);
>
> raw_spin_lock(&dst_rq->lock);
> p->on_rq = ONRQ_QUEUED;
> enqueue_task(dst_rq, p, 0);
> raw_spin_unlock(&dst_rq->lock);
>
> The profit is that double_rq_lock() is not needed now,
> and this may reduce the latencies in some situations.
You forgot to explain how the spinning on task_migrated() is bounded and
thus doesn't make your beginning and end contradict itself.
> Signed-off-by: Kirill Tkhai <ktkhai@...allels.com>
>
> diff --git a/kernel/sched/core.c b/kernel/sched/core.c
> index 26aa7bc..00d7bcc 100644
> --- a/kernel/sched/core.c
> +++ b/kernel/sched/core.c
> @@ -333,7 +333,8 @@ static inline struct rq *__task_rq_lock(struct task_struct *p)
> for (;;) {
> rq = task_rq(p);
> raw_spin_lock(&rq->lock);
> - if (likely(rq == task_rq(p)))
> + if (likely(rq == task_rq(p) &&
> + !task_migrating(p)))
> return rq;
> raw_spin_unlock(&rq->lock);
> }
I would prefer an extra spin-loop like so, that avoids us spinning on
the rq-lock, which serves no purpose.
diff --git a/kernel/sched/core.c b/kernel/sched/core.c
index 2676866b4394..1e65a0bdbbc3 100644
--- a/kernel/sched/core.c
+++ b/kernel/sched/core.c
@@ -331,9 +331,12 @@ static inline struct rq *__task_rq_lock(struct task_struct *p)
lockdep_assert_held(&p->pi_lock);
for (;;) {
+ while (task_migrating(p))
+ cpu_relax();
+
rq = task_rq(p);
raw_spin_lock(&rq->lock);
- if (likely(rq == task_rq(p)))
+ if (likely(rq == task_rq(p) && !task_migrating(p)))
return rq;
raw_spin_unlock(&rq->lock);
}
> diff --git a/kernel/sched/sched.h b/kernel/sched/sched.h
> index e5a9b6d..f6773d7 100644
> --- a/kernel/sched/sched.h
> +++ b/kernel/sched/sched.h
> @@ -17,6 +17,7 @@ struct rq;
>
> /* .on_rq states of struct task_struct: */
The 'normal' way to write that is: task_struct::on_rq
> #define ONRQ_QUEUED 1
> +#define ONRQ_MIGRATING 2
>
> extern __read_mostly int scheduler_running;
>
--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/
Powered by blists - more mailing lists