[<prev] [next>] [<thread-prev] [day] [month] [year] [list]
Message-ID: <20211122191816.GH2105516@nvidia.com>
Date: Mon, 22 Nov 2021 15:18:16 -0400
From: Jason Gunthorpe <jgg@...dia.com>
To: Alex Williamson <alex.williamson@...hat.com>
Cc: Shameerali Kolothum Thodi <shameerali.kolothum.thodi@...wei.com>,
Cornelia Huck <cohuck@...hat.com>,
Yishai Hadas <yishaih@...dia.com>, bhelgaas@...gle.com,
saeedm@...dia.com, linux-pci@...r.kernel.org, kvm@...r.kernel.org,
netdev@...r.kernel.org, kuba@...nel.org, leonro@...dia.com,
kwankhede@...dia.com, mgurtovoy@...dia.com, maorg@...dia.com,
"Dr. David Alan Gilbert" <dgilbert@...hat.com>
Subject: Re: [PATCH V2 mlx5-next 12/14] vfio/mlx5: Implement vfio_pci driver
for mlx5 devices
On Thu, Nov 18, 2021 at 11:15:55AM -0700, Alex Williamson wrote:
> On Tue, 16 Nov 2021 21:48:31 -0400
> Jason Gunthorpe <jgg@...dia.com> wrote:
>
> > On Tue, Nov 16, 2021 at 02:10:31PM -0700, Alex Williamson wrote:
> > > On Tue, 16 Nov 2021 15:25:05 -0400
> > > Jason Gunthorpe <jgg@...dia.com> wrote:
> > >
> > > > On Tue, Nov 16, 2021 at 10:57:36AM -0700, Alex Williamson wrote:
> > > >
> > > > > > I think userspace should decide if it wants to use mlx5 built in or
> > > > > > the system IOMMU to do dirty tracking.
> > > > >
> > > > > What information does userspace use to inform such a decision?
> > > >
> > > > Kernel can't know which approach performs better. Operators should
> > > > benchmark and make a choice for their deployment HW. Maybe device
> > > > tracking severely impacts device performance or vice versa.
> > >
> > > I'm all for keeping policy decisions out of the kernel, but it's pretty
> > > absurd to expect a userspace operator to benchmark various combination
> > > and wire various knobs through the user interface for this.
> >
> > Huh? This is the standard operating procedure in netdev land, mm and
> > others. Max performance requires tunables. And here we really do care
> > alot about migration time. I've seen the mm complexity supporting the
> > normal vCPU migration to know this is a high priority for many people.
>
> Per previous reply:
>
> "Maybe start with what uAPI visible knobs really make sense
> and provide a benefit for per-device dirty bitmap tuning and how a
> device agnostic userspace like QEMU is going to make intelligent
> decisions about those knobs."
Well, the main knob is to pick which dirty tracker to use, and a
2ndary knob is to perhaps configure it (eg granual size or something,
but I have nothing concrete right now).
How to do that in the current uapi? I don't know. Beyond some very
ugly 'do not use system iommu' flag that doesn't fit every need, I
don't have any suggestion.
IMHO, generic qemu is fine to do something simple like always prefer
the system IOMMU tracker.
> QEMU is device agnostic and has no visibility to the system IOMMU
> capabilities.
Ah, there is a lot going on here. We are adding iommufd and iommufd's
design has a specific place to expose the iommu_domain, it's unique
capabilities, and uAPI visible connections from the devices. This is
the big uAPI design upgrade that is needed to get to all the features
we want to see on the iommu side.
For what we have now it is quite easy to add a few ioctls for dirty
track query/start/stop/read&clear that wire directly 1:1 to an ops on
the iommu_domain.
So, qemu does a simple 'query' to the iommu_domain, sees dirty track
is supported and esn't even do anything device specific. If that fails
then it queries each vfio_device for a tracker, if that fails it
operates with no dirty track.
Since each domain and device is a uAPI object all policy is trivially
in userspace hands with a simple uAPI design.
> > The current implementation forces no dirty tracking. That is a policy
> > choice that denies userspace the option to run one device with dirty
> > track and simply suspend the other.
>
> "Forces no dirty tracking"? I think you're suggesting that if a device
> within an IOMMU context doesn't support dirty tracking that we're
> forced to always report all mappings within that context as dirty,
> because for some reason we're not allowed to evolve how devices can
> interact with a shared dirty context, but we are allowed to invent a new
> per-device uAPI?
Basically yes. If we are changing the uAPI then let's change it to
something that makes sense for the HW implementations we actually
have. If you have a clever idea how to do this with the current API
I'm interested to hear, but I don't have anything in mind right now
that isn't much more complicated for everyone.
> What if we have a scenario where devices optionally have per device
> dirty page tracking.
>
> A) mlx5 w/ device level dirty page tracking, vGPU w/ page pinning
>
> vs
>
> B) mlx5 w/ device level dirty page tracking, NIC with system IOMMU
> dirty page tracking
>
> Compare and contrast in which case the shared IOMMU context dirty page
> tracking reports both device's versus only the devices without
> per-device tracking. Is this obvious to both userspace and the shared
> dirty page tracking if it's the same or different in both cases?
Neither A nor B have 'shared dirty page tracking'
The only case where we have some kind of shared tracker is when there
are multiple mdevs with migration support, and the mdev HW can support
a shared bitmap.
In iommufd we still have a reasonable place to put a shared mdev dirty
tracker - the "emulated iommu" is also exposed as uAPI object that can
hold it - however I wouldn't want to implement that in iommufd until
we have some intree drivers that need it.
All the out of tree mdev drivers can use the device specific ioctl. It
only means they can't share dirty bitmaps. Which I think is a
reasonable penalty for out of tree drivers.
> If maybe what you're really trying to get at all along is visibility to
> the per-device dirty page contribution, that does sound interesting.
Visibility and control, yes. It is the requests I've had from our
internal teams.
> But I also don't know how that interacts with system IOMMU based
> reporting. Does the IOMMU report to the driver that reports to
> userspace the per-device dirty pages?
No, the system IOMMU reports through iommufd on the
iommu_domain. Completely unconnected from the migration driver.
Completely unconnected is the design simplification because I don't
need to make kernel infrastructure to build that connection and uAPI
to manage it explicitly.
> That depends whether there are other devices in the context and if the
> container dirty context is meant to include all devices or if the
> driver is opt'ing out of the shared tracking...
Right now, *today*, I know of nothing that needs or can make good use
of shared state bitmaps.
> Alternatively, drivers could register callbacks to report their dirty
> pages into the shared tracking for ranges requested by the user. We
> need to figure out how per-device tracking an system IOMMU tracking get
> along if that's where we're headed.
I don't want shared tracking, it is a waste of CPU and memory
resources. With a single API the only thing that would be OK is no
shared state and kernel iterates over all the trackers. This is quite
far from what is there right now.
> do something like that, then we have some devices that can do p2p and
> some devices that cannot... all or none seems like a much more
> deterministic choice for QEMU. How do guest drivers currently test p2p?
There is a 'p2p allowed' call inside the kernel pci_p2p layer. It
could concievably be extended to consult some firmware table provided
by the hypervisor. It is another one of these cases, like IMS, where
guest transparency is not possible :(
However, you can easially see this need arising - eg a GPU is
frequently a P2P target but rarely a P2P initiator.
It is another case where I can see people making custom VMMs deviating
from what makes sense in enterprise qemu.
> Agreed, but I don't see that userspace choosing to use separate
> contexts either negates the value of the kernel aggregating dirty pages
> within a container or clearly makes the case for per-device dirty pages.
I bring it up because I don't see a clear way to support it at all
with the current API design. Happy to hear ideas
> It's only finally here in the thread that we're seeing some of the mlx5
> implementation details that might favor a per-device solution, hints
> that per device page granularity might be useful, and maybe that
> exposing per-device dirty page footprints to userspace is underlying
> this change of course.
Well, yes, I mean we've thought about this quite a lot internally, we
are not suggesting things randomly for "NIH" as you said. We have lots
of different use cases, several customers, multiple devices and more.
> So provide the justification I asked for previously and quoted above,
> what are the things we want to be able to tune that cannot be done
> through reasonable extensions of the current design? I'm not trying to
> be dismissive, I'm lacking facts and evidence of due diligence that the
> current design is incapable of meeting our goals.
It is not incapable or not, it is ROI. I'm sure we can hack the
current design into something, with a lot of work and complexity.
I'm saying we see a much simpler version that has a simpler kernel
side.
> > If we split them it looks quite simple:
> > - new ioctl on vfio device to read & clear dirty bitmap
> > & extension flag to show this is supported
> > - DIRTY_TRACK migration state bit to start/stop
>
> Is this another device_state bit?
Probably, or a device ioctl - semantically the same
> > Simple logic that read is only possible in NDMA/!RUNNING
> > - new ioctl on vfio device to report dirty tracking capability flags:
> > - internal dirty track supported
> > - real dirty track through attached container supported
> > (only mdev can set this today)
>
> How does system IOMMU dirty tracking work?
As above, it is parallel and completely contained in iommufd.
It is split, iommufd only does system iommu tracking, the device FD
only does device integral tracking. They never cross interfaces
internally, so I don't need to make a kernel framework to support,
aggregate and control multiple dirty trackers.
> I don't understand what's being described here, I'm glad an attempt is
> being made to see what this might look like with the current interface,
> but at the same time the outline seems biased towards a complicated
> portrayal.
Well, not understanding is a problem :|
> > a. works the same, kernel turns on both trackers
> > b. works the same, kernel turns on only mlx5
> > c. Hum. We have to disable the 'no tracker report everything as
> > dirty' feature somehow so we can access only the mlx5 tracker
> > without forcing evey page seen as dirty. Needs a details
>
> Doesn't seem as complicated in my rendition.
I don't see your path through the code.
> > And I don't see that it makes userspace really much simpler anyhow. On
> > the other hand it looks like a lot more kernel work..
>
> There's kernel work either way.
Yes, more vs less :)
> > > OTOH, aggregating these features in the IOMMU reduces both overhead
> > > of per device bitmaps and user operations to create their own
> > > consolidated view.
> >
> > I don't understand this, funneling will not reduce overhead, at best
> > with some work we can almost break even by not allocating the SW
> > bitmaps.
>
> The moment we have more than one device that requires a bitmap, where
> the device doesn't have the visibility of the extent of the bitmap, we
> introduce both space and time overhead versus a shared bitmap that can
> be pre-allocated.
The HW designs we can see right now: mlx5, AMD IOMMU, and hns IOMMU
(though less clear on this one) track the full IOVA space, they store
the track in their own memory and they allow reading each page's dirty
bit with an 'atomic test and clear' semantic.
So currently, we have *zero* devices that require the shared bitmap.
Why should I invest in maintaining and extending this code that has no
current user and no known future purpose?
> What's being asked for here is a change from the current plan of
> record, that entails work and justification, including tangible
> benefits versus a fair exploration of how the same might work in the
> current design.
So, we will make patches for our plan to split them.
We'll show an mlx5 implementation using an ioctl on the vfio device.
I'll sketch how a system iommu works through iommu_domain ops in
iommufd (my goal is to get an AMD implementation at least too)
I can write a paragraph how to fit a shared mdev bitmap tracker into
iommufd if an in-tree user comes..
Can someone provide an equally complete view of how to extend the
current API and solve the same set of use cases?
We can revist this in a month or so when we might have a mlx5 patch.
Jason
Powered by blists - more mailing lists