[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <20251002152346.GA3298749@nvidia.com>
Date: Thu, 2 Oct 2025 12:23:46 -0300
From: Jason Gunthorpe <jgg@...dia.com>
To: Danilo Krummrich <dakr@...nel.org>
Cc: John Hubbard <jhubbard@...dia.com>,
Alexandre Courbot <acourbot@...dia.com>,
Joel Fernandes <joelagnelf@...dia.com>,
Timur Tabi <ttabi@...dia.com>, Alistair Popple <apopple@...dia.com>,
Zhi Wang <zhiw@...dia.com>, Surath Mitra <smitra@...dia.com>,
David Airlie <airlied@...il.com>, Simona Vetter <simona@...ll.ch>,
Alex Williamson <alex.williamson@...hat.com>,
Bjorn Helgaas <bhelgaas@...gle.com>,
Krzysztof Wilczyński <kwilczynski@...nel.org>,
Miguel Ojeda <ojeda@...nel.org>,
Alex Gaynor <alex.gaynor@...il.com>,
Boqun Feng <boqun.feng@...il.com>, Gary Guo <gary@...yguo.net>,
Björn Roy Baron <bjorn3_gh@...tonmail.com>,
Benno Lossin <lossin@...nel.org>,
Andreas Hindborg <a.hindborg@...nel.org>,
Alice Ryhl <aliceryhl@...gle.com>, Trevor Gross <tmgross@...ch.edu>,
nouveau@...ts.freedesktop.org, linux-pci@...r.kernel.org,
rust-for-linux@...r.kernel.org, LKML <linux-kernel@...r.kernel.org>
Subject: Re: [PATCH v2 1/2] rust: pci: skip probing VFs if driver doesn't
support VFs
On Thu, Oct 02, 2025 at 05:11:01PM +0200, Danilo Krummrich wrote:
> On Thu Oct 2, 2025 at 3:56 PM CEST, Jason Gunthorpe wrote:
> >> Surely, you can argue that if they have different enough requirements they
> >> should have different device IDs, but "different enough requirements" is pretty
> >> vague and it's not under our control either.
> >
> > If you want two drivers in Linux you need two PCI IDs.
> >
> > We can't reliably select different drivers based on VFness because
> > VFness is wiped out during virtualization.
>
> Sure, but I thought the whole point is that some VFs are not given directly to
> the VM, but have some kind of intermediate layer, such as vGPU. I.e. some kind
> of hybrid approach between full pass-through and mediated devices?
We'd call the intermediate layer "mediation", and it is abnormal to
have something significant there because it impacts performance and
increases the hypervisor attack surface. Usually you just give the
whole VF mostly as-is.
The ideal vfio driver is the common one because the raw VF doesn't
need any meddling.
> >> But, if there is another solution for VFs already, e.g. in the case of nova-core
> >> vGPU, why restrict drivers from opt-out of VFs. (In a previous reply I mentioned
> >> I prefer opt-in, but you convinced me that it should rather be
> >> opt-out.)
> >
> > I think nova-core has a temporary (OOT even!) issue that should be
> > resolved - that doesn't justify adding core kernel infrastructure that
> > will encourage more drivers to go away from our kernel design goals of
> > drivers working equally in host and VM.
>
> My understanding is that vGPU will ensure that the device exposed to the VM will
> be set up to be (at least mostly) compatible with nova-core's PF code paths?
This is not what I've been told, the VF driver has significant
programming model differences in the NVIDIA model, and supports
different commands.
If you look at the VFIO driver RFC it basically does no mediation, it
isn't intercepting MMIO - the guest sees the BARs directly. Most of
the code is "profiling" from what I can tell. Some config space
meddling.
I think Zhi could make it more clear what mediation the actual VFIO
part of the driver is even doing and why..
Keep in mind for NVIDIA vGPU this has changed alot over the years - I
am talking here about the SRIOV driver proposed recently that is
specifically designed to substantially remove the hypervisor
mediation.
> So, there is a semantical difference between vGPU and nova-core that makes a
> differentiation between VF and PF meaningful and justified.
Aa was said earlier, there is a register in the MMIO that identifies
if the vGPU programming model is used, that is what nova core should
key on to determine how to act.
Jason
Powered by blists - more mailing lists