[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <53baeaa7-0fed-d22c-7767-09ae885d13a0@oracle.com>
Date: Wed, 2 Dec 2020 13:12:07 +0000
From: Joao Martins <joao.m.martins@...cle.com>
To: David Woodhouse <dwmw2@...radead.org>,
Ankur Arora <ankur.a.arora@...cle.com>
Cc: Boris Ostrovsky <boris.ostrovsky@...cle.com>,
Paolo Bonzini <pbonzini@...hat.com>,
Radim Krčmář <rkrcmar@...hat.com>,
Thomas Gleixner <tglx@...utronix.de>,
Ingo Molnar <mingo@...hat.com>, Borislav Petkov <bp@...en8.de>,
"H. Peter Anvin" <hpa@...or.com>, x86@...nel.org,
kvm@...r.kernel.org, linux-kernel@...r.kernel.org
Subject: Re: [PATCH RFC 10/39] KVM: x86/xen: support upcall vector
On 12/2/20 11:17 AM, David Woodhouse wrote:
> On Wed, 2019-02-20 at 20:15 +0000, Joao Martins wrote:
>> @@ -176,6 +177,9 @@ int kvm_arch_set_irq_inatomic(struct kvm_kernel_irq_routing_entry *e,
>> int r;
>>
>> switch (e->type) {
>> + case KVM_IRQ_ROUTING_XEN_EVTCHN:
>> + return kvm_xen_set_evtchn(e, kvm, irq_source_id, level,
>> + line_status);
>> case KVM_IRQ_ROUTING_HV_SINT:
>> return kvm_hv_set_sint(e, kvm, irq_source_id, level,
>> line_status);
>> @@ -325,6 +329,13 @@ int kvm_set_routing_entry(struct kvm *kvm,
>> e->hv_sint.vcpu = ue->u.hv_sint.vcpu;
>> e->hv_sint.sint = ue->u.hv_sint.sint;
>> break;
>> + case KVM_IRQ_ROUTING_XEN_EVTCHN:
>> + e->set = kvm_xen_set_evtchn;
>> + e->evtchn.vcpu = ue->u.evtchn.vcpu;
>> + e->evtchn.vector = ue->u.evtchn.vector;
>> + e->evtchn.via = ue->u.evtchn.via;
>> +
>> + return kvm_xen_setup_evtchn(kvm, e);
>> default:
>> return -EINVAL;
>> }
>
>
> Hmm. I'm not sure I've have done it that way.
>
> These IRQ routing entries aren't for individual event channel ports;
> they don't map to kvm_xen_evtchn_send().
>
> They actually represent the upcall to the given vCPU when any event
> channel is signalled, and it's really per-vCPU configuration.
>
Right.
> When the kernel raises (IPI, VIRQ) events on a given CPU, it doesn't
> actually use these routing entries; it just uses the values in
> vcpu_xen->cb.{via,vector} which were cached from the last of these IRQ
> routing entries that happens to have been processed?
>
Correct.
> The VMM is *expected* to set up precisely one of these for each vCPU,
> right?
>
>From guest PoV, the hypercall:
HYPERVISOR_hvm_op(HVMOP_set_param, HVM_PARAM_CALLBACK_IRQ, ...)
(...) is global.
But this one (on more recent versions of Xen, particularly recent Windows guests):
HVMOP_set_evtchn_upcall_vector
Is per-vCPU, and it's a LAPIC vector.
But indeed the VMM ends up changing the @via @vector on a individual CPU.
> Would it not be better to do that via KVM_XEN_HVM_SET_ATTR?
It's definitely an interesting (better?) alternative, considering we use as a vCPU attribute.
I suppose you're suggesting like,
KVM_XEN_ATTR_TYPE_CALLBACK
And passing the vector, and callback type.
> The usage model for userspace is presumably that the VMM should set the
> appropriate bit in evtchn_pending, check evtchn_mask and then call into
> the kernel to do the set_irq() to inject the callback vector to the
> guest?
>
Correct, that's how it works for userspace handled event channels.
> I might be more inclined to go for a model where the kernel handles the
> evtchn_pending/evtchn_mask for us. What would go into the irq routing
> table is { vcpu, port# } which get passed to kvm_xen_evtchn_send().
>
But passing port to the routing and handling the sending of events wouldn't it lead to
unnecessary handling of event channels which aren't handled by the kernel, compared to
just injecting caring about the upcall?
I thought from previous feedback that it was something you wanted to avoid.
> Does that seem reasonable?
>
Otherwise, it seems reasonable to have it.
I'll let Ankur chip in too, as this was something he was more closely modelling after.
> Either way, I do think we need a way for events raised in the kernel to
> be signalled to userspace, if they are targeted at a vCPU which has
> CALLBACK_VIA_INTX that the kernel can't do directly. So we probably
> *do* need that eventfd I was objecting to earlier, except that it's not
> a per-evtchn thing; it's per-vCPU.
>
Ah!
I wanted to mention the GSI callback method too, but wasn't enterily sure if eventfd was
enough.
Joao
Powered by blists - more mailing lists