[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <20180406105422.6tewkkciwerud3tm@sirius.home.kraxel.org>
Date: Fri, 6 Apr 2018 12:54:22 +0200
From: Gerd Hoffmann <kraxel@...hat.com>
To: Daniel Stone <daniel@...ishbar.org>
Cc: dri-devel <dri-devel@...ts.freedesktop.org>,
Tomeu Vizoso <tomeu.vizoso@...labora.com>,
David Airlie <airlied@...ux.ie>,
open list <linux-kernel@...r.kernel.org>,
qemu-devel@...gnu.org,
"moderated list:DMA BUFFER SHARING FRAMEWORK"
<linaro-mm-sig@...ts.linaro.org>,
"open list:DMA BUFFER SHARING FRAMEWORK"
<linux-media@...r.kernel.org>
Subject: Re: [RfC PATCH] Add udmabuf misc device
On Fri, Apr 06, 2018 at 10:52:21AM +0100, Daniel Stone wrote:
> Hi Gerd,
>
> On 14 March 2018 at 08:03, Gerd Hoffmann <kraxel@...hat.com> wrote:
> >> Either mlock account (because it's mlocked defacto), and get_user_pages
> >> won't do that for you.
> >>
> >> Or you write the full-blown userptr implementation, including mmu_notifier
> >> support (see i915 or amdgpu), but that also requires Christian Königs
> >> latest ->invalidate_mapping RFC for dma-buf (since atm exporting userptr
> >> buffers is a no-go).
> >
> > I guess I'll look at mlock accounting for starters then. Easier for
> > now, and leaves the door open to switch to userptr later as this should
> > be transparent to userspace.
>
> Out of interest, do you have usecases for full userptr support? Maybe
> another way would be to allow creation of dmabufs from memfds.
I have two things in mind.
One is vga emulation. I have virtual pci memory bar for the virtual
vga. qemu backs vga memory with anonymous pages right now, switching
that to shmem should be easy though if that makes things easier. Guest
places the framebuffer somewhere in the pci bar, and I want export the
chunk which represents the framebuffer as dma-buf to display it on the
host without copying around data. Framebuffer is linear in guest
physical memory, so a single block only. That is the simpler case.
The more difficuilt one is virtio-gpu ressources. virtio-gpu resources
live in host memory (guest has no direct access). The guest can
optionally specify guest memory pages as backing storage for the
resource. Guest backing storage is allowed to be scattered. Commands
exist to copy both ways between host storage and guest backing.
With virgl (opengl acceleration) enabled the guest will send rendering
commands to fill the framebuffer ressource, so there is no need to copy
content to the framebuffer ressource. The guest may fill other
resources such as textures used for rendering with copy commands.
Without acceleration the guest does software-rendering to the backing
storage, then sends a command to copy the framebuffer content from guest
backing storage to host ressource.
Now it would be useful to allow a shared mapping, so no copying between
guest backing storage and host resource is needed, especially for the
software rendering case (i.e. dumb gem buffers). Being able to export
guest dumb buffers to other host processes would be useful too, for
example to display guest windows seamlessly on the host wayland server.
So getting a dma-buf for the guest backing storage via udmabuf looked
like a useful approach. We can export the guest gem buffers to other
host processes that way. qemu itself could map it too, to get a linear
representation of the scattered guest backing storage.
The other obvious approach would be to do it the other way around and
allow the guest map the host resource somehow. On the host side qemu
could use vgem to allocate resource memory, so it'll be a gem object
already. Mapping that into the guest isn't that straight-forward
though. The guest manages its physical address space, so the guest
would need to find a free spot and ask the host to place the resource
there. Then the guest needs page structs covering the mapped resource,
so it can work with it. Didn't investigate how difficuilt that is. Use
memory hotplug maybe? Can we easily unmap the resource then? Also I
think updating the guests physical memory layout (which we would need to
do on every resource map/unmap) isn't an exactly cheap operation ...
cheers,
Gerd
Powered by blists - more mailing lists