[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <029bb2f5-78d5-a3f2-1ae7-97fc7147611d@arm.com>
Date: Fri, 10 Mar 2023 15:57:27 +0000
From: Robin Murphy <robin.murphy@....com>
To: Jason Gunthorpe <jgg@...dia.com>
Cc: Jean-Philippe Brucker <jean-philippe@...aro.org>,
Nicolin Chen <nicolinc@...dia.com>, will@...nel.org,
eric.auger@...hat.com, kevin.tian@...el.com,
baolu.lu@...ux.intel.com, joro@...tes.org,
shameerali.kolothum.thodi@...wei.com,
linux-arm-kernel@...ts.infradead.org, iommu@...ts.linux.dev,
linux-kernel@...r.kernel.org, yi.l.liu@...el.com
Subject: Re: [PATCH v1 02/14] iommufd: Add nesting related data structures for
ARM SMMUv3
On 2023-03-10 15:25, Jason Gunthorpe wrote:
> On Fri, Mar 10, 2023 at 02:52:42PM +0000, Robin Murphy wrote:
>> On 2023-03-09 21:01, Jason Gunthorpe wrote:
>>>> For a lot of SMMUv3 implementations that have a single queue and for
>>>> other architectures, we can do better than hardware emulation.
>>>
>>> How is using a SW emulated virtio formatted queue better than using a
>>> SW emulated SMMUv3 ECMDQ?
>>
>> Since it's not been said, the really big thing is that virtio explicitly
>> informs the host whenever the guest maps something. Emulating SMMUv3 means
>> the host has to chase all the pagetable pointers in guest memory and trap
>> writes such that it has visibility of invalid->valid transitions and can
>> update the physical shadow pagetable correspondingly.
>
> Sorry, I mean in the context of future virtio-iommu that is providing
> nested translation.
Ah, that's probably me missing the context again.
> eg why would anyone want to use virtio to provide SMMUv3 based HW
> accelerated nesting?
>
> Jean suggested that the invalidation flow for virtio-iommu could be
> faster because it is in kernel, but I'm saying that we could also make
> the SMMUv3 invalidation in-kernel with the same basic technique. (and
> actively wondering if we should put more focus on that)
>
> I understand the appeal of the virtio scheme with its current
> map/unmap interface.
>
> I could also see some appeal of a simple virtio-iommu SVA that could
> point map a CPU page table as an option. The guest already has to know
> how to manage these anyhow so it is nicely general.
>
> If iommufd could provide a general cross-driver API to set exactly
> that scenario up then VMM code could also be general. That seems
> prettty interesting.
Indeed, I've always assumed the niche for virtio would be that kind of
in-between use-case using nesting to accelerate simple translation,
where we plug a guest-owned pagetable into a host-owned context. That
way the guest retains the simple virtio interface and only needs to
understand a pagetable format (or as you say, simply share a CPU
pagetable) without having to care about the nitty-gritty of all the
IOMMU-specific moving parts around it. For guests that want to get into
more advanced stuff like managing their own PASID tables, pushing them
towards "native" nesting probably does make more sense.
> But if the plan is to expose more detailed stuff like the CD or GCR3
> PASID tables as something the guest has to manipulate and then a bunch
> of special invalidation to support that, and VMM code to back it, then
> I'm questioning the whole point. We lost the generality.
>
> Just use the normal HW accelerated SMMUv3 nesting model instead.
>
> If virtio-iommu SVA is really important for ARM then I'd suggest
> SMMUv3 should gain a new HW capability to allowed the CD table to be
> in hypervisor memory so it works consistently for virtio-iommu SVA.
Oh, maybe I should have read this far before reasoning the exact same
thing from scratch... oh well, this time I'm not going to go back and
edit :)
Thanks,
Robin.
Powered by blists - more mailing lists