[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <4B3765D2.5020805@redhat.com>
Date: Sun, 27 Dec 2009 15:49:06 +0200
From: Avi Kivity <avi@...hat.com>
To: Gregory Haskins <gregory.haskins@...il.com>
CC: Ingo Molnar <mingo@...e.hu>,
Anthony Liguori <anthony@...emonkey.ws>,
Bartlomiej Zolnierkiewicz <bzolnier@...il.com>,
Andi Kleen <andi@...stfloor.org>, kvm@...r.kernel.org,
Andrew Morton <akpm@...ux-foundation.org>,
torvalds@...ux-foundation.org,
"linux-kernel@...r.kernel.org" <linux-kernel@...r.kernel.org>,
netdev@...r.kernel.org,
"alacrityvm-devel@...ts.sourceforge.net"
<alacrityvm-devel@...ts.sourceforge.net>
Subject: Re: [GIT PULL] AlacrityVM guest drivers for 2.6.33
On 12/27/2009 03:34 PM, Gregory Haskins wrote:
> On 12/27/09 4:33 AM, Avi Kivity wrote:
>
>> On 12/24/2009 11:36 AM, Gregory Haskins wrote:
>>
>>>> As a twist on this, the VMware paravirt driver interface is so
>>>> hardware-like that they're getting hardware vendors to supply cards that
>>>> implement it. Try that with a pure software approach.
>>>>
>>>>
>>> Any hardware engineer (myself included) will tell you that, generally
>>> speaking, what you can do in hardware you can do in software (think of
>>> what QEMU does today, for instance). It's purely a cost/performance
>>> tradeoff.
>>>
>>> I can at least tell you that is true of vbus. Anything on the vbus side
>>> would be equally eligible for a hardware implementation, though there is
>>> not reason to do this today since we have equivalent functionality in
>>> baremetal already.
>>>
>> There's a huge difference in the probability of vmware getting cards to
>> their spec, or x86 vendors improving interrupt delivery to guests,
>> compared to vbus being implemented in hardware.
>>
> Thats not relevant, however. I said in the original quote that you
> snipped that I made it a software design on purpose, and you tried to
> somehow paint that as a negative because vmware made theirs
> "hardware-like" and you implied it could not be done with my approach
> with the statement "try that with a pure software approach". And the
> bottom line is that the statement is incorrect and/or misleading.
>
It's not incorrect. VMware stuck to the pci specs and as a result they
can have hardware implement their virtual NIC protocol. For vbus this
is much harder to do since you need a side-channel between different
cards to coordinate interrupt delivery. In theory you can do eveything
if you don't consider practicalities.
That's a digression, though, I'm not suggesting we'll see virtio
hardware or that this is a virtio/pci advantage vs. vbus. It's an
anecdote showing that sticking with specs has its advantages.
wrt pci vs vbus, the difference is in the ability to use improvements in
interrupt delivery accelerations in virt hardware. If this happens,
virtio/pci can immediately take advantage of it, while vbus has to stick
with software delivery for backward compatibility, and all that code
becomes a useless support burden.
As an example of what hardware can do when it really sets its mind to
it, s390 can IPI from vcpu to vcpu without exiting to the host.
>>> The only motiviation is if you wanted to preserve
>>> ABI etc, which is what vmware is presumably after. However, I am not
>>> advocating this as necessary at this juncture.
>>>
>>>
>> Maybe AlacrityVM users don't care about compatibility, but my users do.
>>
> Again, not relevant to this thread. Making your interface
> "hardware-like" buys you nothing in the end, as you ultimately need to
> load drivers in the guest either way, and any major OS lets you extend
> both devices and buses with relative ease. The only counter example
> would be if you truly were "hardware-exactly" like e1000 emulation, but
> we already know that this means it is hardware centric and not
> "exit-rate aware" and would perform poorly. Otherwise "compatible" is
> purely a point on the time line (for instance, the moment virtio-pci ABI
> shipped), not an architectural description such as "hardware-like".
>
True, not related to the thread. But it is a problem. The difference
between virtio and vbus here is that virtio is already deployed and its
users expect not to reinstall drivers [1]. Before virtio existed,
people could not deploy performance sensitive applications on kvm. Now
that it exists, we have to support it without requiring users to touch
their guests.
That means that without proof that virtio cannot be scaled, we'll keep
supporting and extending it.
[1] Another difference is the requirement for writing a "bus driver" for
every supported guest, which means dealing with icky bits like hotplug.
--
error compiling committee.c: too many arguments to function
--
To unsubscribe from this list: send the line "unsubscribe netdev" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at http://vger.kernel.org/majordomo-info.html
Powered by blists - more mailing lists