[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <4AB0A070.1050400@redhat.com>
Date: Wed, 16 Sep 2009 11:23:12 +0300
From: Avi Kivity <avi@...hat.com>
To: Gregory Haskins <gregory.haskins@...il.com>
CC: "Michael S. Tsirkin" <mst@...hat.com>,
"Ira W. Snyder" <iws@...o.caltech.edu>, netdev@...r.kernel.org,
virtualization@...ts.linux-foundation.org, kvm@...r.kernel.org,
linux-kernel@...r.kernel.org, mingo@...e.hu, linux-mm@...ck.org,
akpm@...ux-foundation.org, hpa@...or.com,
Rusty Russell <rusty@...tcorp.com.au>, s.hetze@...ux-ag.com,
alacrityvm-devel@...ts.sourceforge.net
Subject: Re: [PATCHv5 3/3] vhost_net: a kernel-level virtio server
On 09/15/2009 11:08 PM, Gregory Haskins wrote:
>
>> There's virtio-console, virtio-blk etc. None of these have kernel-mode
>> servers, but these could be implemented if/when needed.
>>
> IIUC, Ira already needs at least ethernet and console capability.
>
>
He's welcome to pick up the necessary code from qemu.
>>> b) what do you suppose this protocol to aggregate the connections would
>>> look like? (hint: this is what a vbus-connector does).
>>>
>>>
>> You mean multilink? You expose the device as a multiqueue.
>>
> No, what I mean is how do you surface multiple ethernet and consoles to
> the guests? For Ira's case, I think he needs at minimum at least one of
> each, and he mentioned possibly having two unique ethernets at one point.
>
You instantiate multiple vhost-nets. Multiple ethernet NICs is a
supported configuration for kvm.
> His slave boards surface themselves as PCI devices to the x86
> host. So how do you use that to make multiple vhost-based devices (say
> two virtio-nets, and a virtio-console) communicate across the transport?
>
I don't really see the difference between 1 and N here.
> There are multiple ways to do this, but what I am saying is that
> whatever is conceived will start to look eerily like a vbus-connector,
> since this is one of its primary purposes ;)
>
I'm not sure if you're talking about the configuration interface or data
path here.
>>> c) how do you manage the configuration, especially on a per-board basis?
>>>
>>>
>> pci (for kvm/x86).
>>
> Ok, for kvm understood (and I would also add "qemu" to that mix). But
> we are talking about vhost's application in a non-kvm environment here,
> right?.
>
> So if the vhost-X devices are in the "guest",
They aren't in the "guest". The best way to look at it is
- a device side, with a dma engine: vhost-net
- a driver side, only accessing its own memory: virtio-net
Given that Ira's config has the dma engine in the ppc boards, that's
where vhost-net would live (the ppc boards acting as NICs to the x86
board, essentially).
> and the x86 board is just
> a slave...How do you tell each ppc board how many devices and what
> config (e.g. MACs, etc) to instantiate? Do you assume that they should
> all be symmetric and based on positional (e.g. slot) data? What if you
> want asymmetric configurations (if not here, perhaps in a different
> environment)?
>
I have no idea, that's for Ira to solve. If he could fake the PCI
config space as seen by the x86 board, he would just show the normal pci
config and use virtio-pci (multiple channels would show up as a
multifunction device). Given he can't, he needs to tunnel the virtio
config space some other way.
>> Yes. virtio is really virtualization oriented.
>>
> I would say that its vhost in particular that is virtualization
> oriented. virtio, as a concept, generally should work in physical
> systems, if perhaps with some minor modifications. The biggest "limit"
> is having "virt" in its name ;)
>
Let me rephrase. The virtio developers are virtualization oriented. If
it works for non-virt applications, that's good, but not a design goal.
--
error compiling committee.c: too many arguments to function
--
To unsubscribe from this list: send the line "unsubscribe netdev" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at http://vger.kernel.org/majordomo-info.html
Powered by blists - more mailing lists