[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <413f59cd3c0a80c5b71a0cd033fdaad082c5a0e7.camel@redhat.com>
Date: Thu, 21 Jul 2022 18:31:56 +0300
From: Maxim Levitsky <mlevitsk@...hat.com>
To: Sean Christopherson <seanjc@...gle.com>
Cc: Santosh Shukla <santosh.shukla@....com>,
Paolo Bonzini <pbonzini@...hat.com>,
Vitaly Kuznetsov <vkuznets@...hat.com>,
Jim Mattson <jmattson@...gle.com>,
Joerg Roedel <joro@...tes.org>,
Tom Lendacky <thomas.lendacky@....com>, kvm@...r.kernel.org,
linux-kernel@...r.kernel.org
Subject: Re: [PATCHv2 4/7] KVM: SVM: Report NMI not allowed when Guest busy
handling VNMI
On Thu, 2022-07-21 at 14:59 +0000, Sean Christopherson wrote:
> On Thu, Jul 21, 2022, Maxim Levitsky wrote:
> > On Wed, 2022-07-20 at 21:54 +0000, Sean Christopherson wrote:
> > > On Sat, Jul 09, 2022, Santosh Shukla wrote:
> > > > @@ -3609,6 +3612,9 @@ static void svm_enable_nmi_window(struct kvm_vcpu *vcpu)
> > > > {
> > > > struct vcpu_svm *svm = to_svm(vcpu);
> > > >
> > > > + if (is_vnmi_enabled(svm))
> > > > + return;
> > >
> > > Ugh, is there really no way to trigger an exit when NMIs become unmasked? Because
> > > if there isn't, this is broken for KVM.
> > >
> > > On bare metal, if two NMIs arrive "simultaneously", so long as NMIs aren't blocked,
> > > the first NMI will be delivered and the second will be pended, i.e. software will
> > > see both NMIs. And if that doesn't hold true, the window for a true collision is
> > > really, really tiny.
> > >
> > > But in KVM, because a vCPU may not be run a long duration, that window becomes
> > > very large. To not drop NMIs and more faithfully emulate hardware, KVM allows two
> > > NMIs to be _pending_. And when that happens, KVM needs to trigger an exit when
> > > NMIs become unmasked _after_ the first NMI is injected.
> >
> > This is how I see this:
> >
> > - When a NMI arrives and neither NMI is injected (V_NMI_PENDING) nor in service (V_NMI_MASK)
> > then all it is needed to inject the NMI will be to set the V_NMI_PENDING bit and do VM entry.
> >
> > - If V_NMI_PENDING is set but not V_NMI_MASK, and another NMI arrives we can make the
> > svm_nmi_allowed return -EBUSY which will cause immediate VM exit,
> >
> > and if hopefully vNMI takes priority over the fake interrupt we raise, it will be injected,
>
> Nit (for other readers following along), it's not a fake interrupt,I would describe
> it as spurious or ignored. It's very much a real IRQ, which matters because it
> factors into event priority.
Yep, 100% agree.
>
> > and upon immediate VM exit we can inject another NMI by setting the V_NMI_PENDING again,
> > and later when the guest is done with first NMI, it will take the second.
>
> Yeaaaah. This depends heavily on the vNMI being prioritized over the IRQ.
>
> > Of course if we get a nested exception, then it will be fun....
> >
> > (the patches don't do it (causing immediate VM exit),
> > but I think we should make the svm_nmi_allowed, check for the case for
> > V_NMI_PENDING && !V_NMI_MASK and make it return -EBUSY).
>
> Yep, though I think there's a wrinkle (see below).
>
> > - If both V_NMI_PENDING and V_NMI_MASK are set, then I guess we lose an NMI.
> > (It means that the guest is handling an NMI, there is a pending NMI, and now
> > another NMI arrived)
> >
> > Sean, this is the problem you mention, right?
>
> Yep. Dropping an NMI in the last case is ok, AFAIK no CPU will pend multiple NMIs
> while another is in-flight. But triggering an immediate exit in svm_nmi_allowed()
> will hang the vCPU as the second pending NMI will never go away since the vCPU
The idea is to trigger the immediate exit only when a NMI was just injected (V_NMI_PENDING=1)
but not masked (that is currently in service, that is V_NMI_MASK=0).
In case both bits are set, the NMI is dropped, that is no immediate exit is requested.
In this case, next VM entry should have no reason to not inject the NMI and then VM exit
on the interrupt we raised, so there should not be a problem with forward progress.
There is an issue still, the NMI could also be masked if we are in SMM (I suggested
setting the V_NMI_MASK manually in this case), thus in this case we won't have more
that one pending NMI, but I guess this is not that big problem.
We can btw also in this case "open" the NMI window by waiting for RSM intercept.
(that is just not inject the NMI, and on RSM inject it, I think that KVM already does this)
I think it should overal work, but no doubt I do expect issues and corner cases,
> won't make forward progress to unmask NMIs. This can also happen if there are
> two pending NMIs and GIF=0, i.e. any time there are multiple pending NMIs and NMIs
> are blocked.
GIF=0 can be dealt with though, if GIF is 0 when 2nd pending NMI arrives, we can
delay its injection to the moment the STGI is executed and intercept STGI.
We I think already do something like that as well.
>
> One other question: what happens if software atomically sets V_NMI_PENDING while
> the VMCB is in use? I assume bad things? I.e. I assume KVM can't "post" NMIs :-)
I can't answer that but I bet that this bit is only checked on VM entry.
Another question about the spec (sorry if I already asked this):
if vGIF is used, and the guest does CLGI, does the CPU set the V_NMI_MASK,
itself, or the CPU considers both the V_GIF and the V_NMI_MASK in the int_ctl,
as a condition of delivering the NMI? I think that the later should be true,
and thus V_NMI_MASK is more like V_NMI_IN_SERVICE.
Best regards,
Maxim Levitsky
>
Powered by blists - more mailing lists