|
[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] Re: Why memory lending is needed for GPU acceleration
Hi, On 3/29/26 2:32 PM, Demi Marie Obenour wrote: On 3/24/26 10:17, Demi Marie Obenour wrote:Here is a proposed design document for supporting mapping GPU VRAM and/or file-backed memory into other domains. It's not in the form of a patch because the leading + characters would just make it harder to read for no particular gain, and because this is still RFC right now. Once it is ready to merge, I'll send a proper patch. Nevertheless, you can consider this to be Signed-off-by: Demi Marie Obenour <demiobenour@xxxxxxxxx> This approach is very different from the "frontend-allocates" approach used elsewhere in Xen. It is very much Linux-centric, rather than Xen-centric. In fact, MMU notifiers were invented for KVM, and this approach is exactly the same as the one KVM implements. However, to the best of my understanding, the design described here is the only viable one. Linux MM and GPU drivers require it, and changes to either to relax this requirement will not be accepted upstream. To expand on this: the reason cross-domain Wayland proxies have been doing this SHM copy dance was a deficiency in Linux UAPI. Basically, applications allocate shared memory using local mechanisms like memfd (and good old unlink-of-regular-file, ugh) which weren't compatible with cross-VM sharing. However udmabuf should basically solve it, at least for memfds. (I haven't yet investigated what happens with "unlinked regular files" yet but I don't expect anything good there, welp.) But I have landed a patch in Linux that removes a silly restriction that tied dmabuf import into virtgpu to KMS-only mode: https://git.kernel.org/pub/scm/linux/kernel/git/next/linux-next.git/commit/?id=df4dc947c46bb9f80038f52c6e38cb2d40c10e50And I have experimented with it and got a KVM-based VMM to successfully access and print guest memfd contents that were passed to the host via this mechanism. (Time to actually properly implement it into the full system..)
To recap, how virtio-gpu Host3d memory currently works with KVM is:- the VMM/virtgpu receives a dmabuf over a socket (Wayland/D-Bus/whatever) and registers it internally with some resource ID that's passed to the guest; - When the guest imports that resource, it calls VIRTIO_GPU_CMD_RESOURCE_MAP_BLOB to get a PRIME buffer that can be turned into a dmabuf fd; - the VMM's handler for VIRTIO_GPU_CMD_RESOURCE_MAP_BLOB (referencing libkrun here) literally just calls mmap() on the host dmabuf, using the MAP_FIXED flag to place it correctly inside of the VMM process's guest-exposed VA region (configured via KVM_SET_USER_MEMORY_REGION); - so any resource imported by the guest, even before guest userspace does mmap(), is mapped (as VM_PFNMAP|VM_IO) until the guest releases it. So the generic kernel MM is out of the way, these mappings can't be paged out to swap etc. But accessing them may fault, as the comment for drm_gem_mmap_obj says: * Depending on their requirements, GEM objects can either * provide a fault handler in their vm_ops (in which case any accesses to * the object will be trapped, to perform migration, GTT binding, surface* register allocation, or performance monitoring), or mmap the buffer memory * synchronously after calling drm_gem_mmap_objIt all "just works" in KVM because KVM's resolution of the guest's memory accesses tries to be literally equivalent to what's mapped into the userspace VMM process: hva_to_pfn_remapped explicitly calls fixup_user_fault and eventually gets to the GPU driver's fault handler. Now for Xen this would be… painful, but,we have no need to replicate what KVM does. That's far from the only thing that can be done with a dmabuf. The import-export machinery on the other hand actually does pin the buffers on the driver level, importers are not obligated to support movable buffers (move_notify in dma_buf_attach_ops is entirely optional). Interestingly, there is already XEN_GNTDEV_DMABUF…Wait, do we even have any reason at all to suspect that XEN_GNTDEV_DMABUF doesn't already satisfy all of our buffer-sharing requirements? Thanks, ~val P.S. while I have everyone's attention, can I get some eyes on:https://lore.kernel.org/all/20251126062124.117425-1-val@xxxxxxxxxxxxxxxxxxxxxx/ ?
|
![]() |
Lists.xenproject.org is hosted with RackSpace, monitoring our |