[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <0af37d2ec77e3a3c2d8ba5adbe9aab4170dc13e4.camel@redhat.com>
Date: Tue, 17 Sep 2019 12:06:08 -0500
From: Scott Wood <swood@...hat.com>
To: Sebastian Andrzej Siewior <bigeasy@...utronix.de>
Cc: Thomas Gleixner <tglx@...utronix.de>,
Steven Rostedt <rostedt@...dmis.org>,
Peter Zijlstra <peterz@...radead.org>,
Juri Lelli <juri.lelli@...hat.com>,
Daniel Bristot de Oliveira <bristot@...hat.com>,
Clark Williams <williams@...hat.com>,
linux-kernel@...r.kernel.org, linux-rt-users@...r.kernel.org
Subject: Re: [PATCH RT 8/8] sched: Lazy migrate_disable processing
On Tue, 2019-09-17 at 18:50 +0200, Sebastian Andrzej Siewior wrote:
> On 2019-07-27 00:56:38 [-0500], Scott Wood wrote:
> > diff --git a/kernel/cpu.c b/kernel/cpu.c
> > index 885a195dfbe0..0096acf1a692 100644
> > --- a/kernel/cpu.c
> > +++ b/kernel/cpu.c
> > @@ -939,17 +893,34 @@ static int takedown_cpu(unsigned int cpu)
> > */
> > irq_lock_sparse();
> >
> > -#ifdef CONFIG_PREEMPT_RT_FULL
> > - __write_rt_lock(cpuhp_pin);
> > +#ifdef CONFIG_PREEMPT_RT_BASE
> > + WARN_ON_ONCE(takedown_cpu_task);
> > + takedown_cpu_task = current;
> > +
> > +again:
> > + for (;;) {
> > + int nr_pinned;
> > +
> > + set_current_state(TASK_UNINTERRUPTIBLE);
> > + nr_pinned = cpu_nr_pinned(cpu);
> > + if (nr_pinned == 0)
> > + break;
> > + schedule();
> > + }
>
> we used to have cpuhp_pin which ensured that once we own the write lock
> there will be no more tasks that can enter a migrate_disable() section
> on this CPU. It has been placed fairly late to ensure that nothing new
> comes in as part of the shutdown process and that it flushes everything
> out that is still in a migrate_disable() section.
> Now you claim that once the counter reached zero it never increments
> again. I would be happier if there was an explicit check for that :)
I don't claim that. A check is added in take_cpu_down() to see whether it
went back up, and if so, exit with EAGAIN. If *that* check succeeds, it
can't go back up because it's in stop machine, and any tasks will get
migrated to another CPU before they can run again. There's also a WARN in
migrate_tasks() if somehow a migrate-disabled task does get encountered.
> There is no back off and flush mechanism which means on a busy CPU (as
> in heavily lock contended by multiple tasks) this will wait until the
> CPU gets idle again.
Not really any different from the reader-biased rwlock that this replaces...
-Scott
Powered by blists - more mailing lists