[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <20190325094340.GJ9149@xz-x1>
Date: Mon, 25 Mar 2019 17:43:40 +0800
From: Peter Xu <peterx@...hat.com>
To: Thomas Gleixner <tglx@...utronix.de>
Cc: Ming Lei <ming.lei@...hat.com>, Christoph Hellwig <hch@....de>,
Jason Wang <jasowang@...hat.com>,
Luiz Capitulino <lcapitulino@...hat.com>,
Linux Kernel Mailing List <linux-kernel@...r.kernel.org>,
"Michael S. Tsirkin" <mst@...hat.com>, minlei@...hat.com
Subject: Re: Virtio-scsi multiqueue irq affinity
On Mon, Mar 25, 2019 at 09:53:28AM +0100, Thomas Gleixner wrote:
> Ming,
>
> On Mon, 25 Mar 2019, Ming Lei wrote:
> > On Mon, Mar 25, 2019 at 01:02:13PM +0800, Peter Xu wrote:
> > > One thing I can think of is the real-time scenario where "isolcpus="
> > > is provided, then logically we should not allow any isolated CPUs to
> > > be bound to any of the multi-queue IRQs. Though Ming Lei and I had a
> >
> > So far, this behaviour is made by user-space.
> >
> > >From my understanding, IRQ subsystem doesn't handle "isolcpus=", even
> > though the Kconfig help doesn't mention irq affinity affect:
> >
> > Make sure that CPUs running critical tasks are not disturbed by
> > any source of "noise" such as unbound workqueues, timers, kthreads...
> > Unbound jobs get offloaded to housekeeping CPUs. This is driven by
> > the "isolcpus=" boot parameter.
>
> isolcpus has no effect on the interupts. That's what 'irqaffinity=' is for.
>
> > Yeah, some RT application may exclude 'isolcpus=' from some IRQ's
> > affinity via /proc/irq interface, and now it becomes not possible any
> > more to do that for managed IRQ.
> >
> > > discussion offlist before and Ming explained to me that as long as the
> > > isolated CPUs do not generate any IO then there will be no IRQ on
> > > those isolated (real-time) CPUs at all. Can we guarantee that? Now
> >
> > It is only guaranteed for 1:1 mapping.
> >
> > blk-mq uses managed IRQ's affinity to setup queue mapping, for example:
> >
> > 1) single hardware queue
> > - this queue's IRQ affinity includes all CPUs, then the hardware queue's
> > IRQ is only fired on one specific CPU for IO submitted from any CPU
>
> Right. We can special case that for single HW queue to honor the default
> affinity setting. That's not hard to achieve.
>
> > 2) multi hardware queue
> > - there are N hardware queues
> > - for each hardware queue i(i < N), its IRQ's affinity may include N(i) CPUs,
> > then IRQ for this hardware queue i is fired on one specific CPU among N(i).
>
> Correct and that's the sane case where it does not matter much, because if
> your task on an isolated CPU does I/O then redirecting it through some
> other CPU does not make sense. If it doesn't do I/O it wont be affected by
> the dormant queue.
(My thanks to both.)
Now I understand it can be guaranteed so it should not break
determinism of the real-time applications. But again, I'm curious
whether we can specify how to spread the hardware queues of a block
controller (as I asked in my previous post) instead of the default one
(which is to spread the queues upon all the cores)? I'll try to give
a detailed example on this one this time: Let's assume we've had a
host with 2 nodes and 8 cores (Node 0 with CPUs 0-3, Node 1 with CPUs
4-7), and a SCSI controller with 4 queues. We want to take the 2nd
node to run the real-time applications so we do isolcpus=4-7. By
default, IIUC the hardware queues will be allocated like this:
- queue 1: CPU 0,1
- queue 2: CPU 2,3
- queue 3: CPU 4,5
- queue 4: CPU 6,7
And the IRQs of the queues will be bound to the same cpuset that the
queue is bound to.
So my previous question is: since we know that CPU 4-7 won't generate
any IO after all (and they shouldn't), could it be possible that we
configure the system somehow to reflect a mapping like below:
- queue 1: CPU 0
- qeueu 2: CPU 1
- queue 3: CPU 2
- queue 4: CPU 3
Then we disallow the CPUs 4-7 to generate IO and return failure if
they tries to.
Again, I'm pretty uncertain on whether this case can be anything close
to useful... It just came out of my pure curiosity. I think it at
least has some benefits like: we will guarantee that the realtime CPUs
won't send block IO requests (which could be good because it could
simply break real-time determinism), and we'll save two queues from
being totally idle (so if we run non-real-time block applications on
cores 0-3 we still gain 4 hardware queues's throughput rather than 2).
Thanks,
--
Peter Xu
Powered by blists - more mailing lists