[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <20180409080051.GD31310@phenom.ffwll.local>
Date: Mon, 9 Apr 2018 10:00:51 +0200
From: Daniel Vetter <daniel@...ll.ch>
To: Gerd Hoffmann <kraxel@...hat.com>
Cc: Daniel Stone <daniel@...ishbar.org>,
Tomeu Vizoso <tomeu.vizoso@...labora.com>,
David Airlie <airlied@...ux.ie>, qemu-devel@...gnu.org,
dri-devel <dri-devel@...ts.freedesktop.org>,
open list <linux-kernel@...r.kernel.org>,
"moderated list:DMA BUFFER SHARING FRAMEWORK"
<linaro-mm-sig@...ts.linaro.org>,
"open list:DMA BUFFER SHARING FRAMEWORK"
<linux-media@...r.kernel.org>
Subject: Re: [RfC PATCH] Add udmabuf misc device
On Fri, Apr 06, 2018 at 12:54:22PM +0200, Gerd Hoffmann wrote:
> On Fri, Apr 06, 2018 at 10:52:21AM +0100, Daniel Stone wrote:
> > Hi Gerd,
> >
> > On 14 March 2018 at 08:03, Gerd Hoffmann <kraxel@...hat.com> wrote:
> > >> Either mlock account (because it's mlocked defacto), and get_user_pages
> > >> won't do that for you.
> > >>
> > >> Or you write the full-blown userptr implementation, including mmu_notifier
> > >> support (see i915 or amdgpu), but that also requires Christian Königs
> > >> latest ->invalidate_mapping RFC for dma-buf (since atm exporting userptr
> > >> buffers is a no-go).
> > >
> > > I guess I'll look at mlock accounting for starters then. Easier for
> > > now, and leaves the door open to switch to userptr later as this should
> > > be transparent to userspace.
> >
> > Out of interest, do you have usecases for full userptr support? Maybe
> > another way would be to allow creation of dmabufs from memfds.
>
> I have two things in mind.
>
> One is vga emulation. I have virtual pci memory bar for the virtual
> vga. qemu backs vga memory with anonymous pages right now, switching
> that to shmem should be easy though if that makes things easier. Guest
> places the framebuffer somewhere in the pci bar, and I want export the
> chunk which represents the framebuffer as dma-buf to display it on the
> host without copying around data. Framebuffer is linear in guest
> physical memory, so a single block only. That is the simpler case.
>
> The more difficuilt one is virtio-gpu ressources. virtio-gpu resources
> live in host memory (guest has no direct access). The guest can
> optionally specify guest memory pages as backing storage for the
> resource. Guest backing storage is allowed to be scattered. Commands
> exist to copy both ways between host storage and guest backing.
>
> With virgl (opengl acceleration) enabled the guest will send rendering
> commands to fill the framebuffer ressource, so there is no need to copy
> content to the framebuffer ressource. The guest may fill other
> resources such as textures used for rendering with copy commands.
>
> Without acceleration the guest does software-rendering to the backing
> storage, then sends a command to copy the framebuffer content from guest
> backing storage to host ressource.
>
> Now it would be useful to allow a shared mapping, so no copying between
> guest backing storage and host resource is needed, especially for the
> software rendering case (i.e. dumb gem buffers). Being able to export
> guest dumb buffers to other host processes would be useful too, for
> example to display guest windows seamlessly on the host wayland server.
>
> So getting a dma-buf for the guest backing storage via udmabuf looked
> like a useful approach. We can export the guest gem buffers to other
> host processes that way. qemu itself could map it too, to get a linear
> representation of the scattered guest backing storage.
>
> The other obvious approach would be to do it the other way around and
> allow the guest map the host resource somehow. On the host side qemu
> could use vgem to allocate resource memory, so it'll be a gem object
> already. Mapping that into the guest isn't that straight-forward
> though. The guest manages its physical address space, so the guest
> would need to find a free spot and ask the host to place the resource
> there. Then the guest needs page structs covering the mapped resource,
> so it can work with it. Didn't investigate how difficuilt that is. Use
> memory hotplug maybe? Can we easily unmap the resource then? Also I
> think updating the guests physical memory layout (which we would need to
> do on every resource map/unmap) isn't an exactly cheap operation ...
Generally we try to cache mappings as much as possible. And wrt finding a
slot: Create a sufficiently sized BAR on the virgl device, just for that?
-Daniel
--
Daniel Vetter
Software Engineer, Intel Corporation
http://blog.ffwll.ch
Powered by blists - more mailing lists