65b847d284
Create sync objects and fences only for dmabufs that are blobs. Once a fence is created (after glFlush) and is signalled, graphic_hw_gl_flushed() will be called and virtio-gpu cmd processing will be resumed. Cc: Gerd Hoffmann <kraxel@redhat.com> Signed-off-by: Vivek Kasireddy <vivek.kasireddy@intel.com> Message-Id: <20210914211837.3229977-4-vivek.kasireddy@intel.com> Signed-off-by: Gerd Hoffmann <kraxel@redhat.com>
225 lines
5.6 KiB
C
225 lines
5.6 KiB
C
/*
|
|
* Virtio GPU Device
|
|
*
|
|
* Copyright Red Hat, Inc. 2013-2014
|
|
*
|
|
* Authors:
|
|
* Dave Airlie <airlied@redhat.com>
|
|
* Gerd Hoffmann <kraxel@redhat.com>
|
|
*
|
|
* This work is licensed under the terms of the GNU GPL, version 2 or later.
|
|
* See the COPYING file in the top-level directory.
|
|
*/
|
|
|
|
#include "qemu/osdep.h"
|
|
#include "qemu/units.h"
|
|
#include "qemu-common.h"
|
|
#include "qemu/iov.h"
|
|
#include "ui/console.h"
|
|
#include "hw/virtio/virtio-gpu.h"
|
|
#include "hw/virtio/virtio-gpu-pixman.h"
|
|
#include "trace.h"
|
|
#include "exec/ramblock.h"
|
|
#include "sysemu/hostmem.h"
|
|
#include <sys/ioctl.h>
|
|
#include <fcntl.h>
|
|
#include <linux/memfd.h>
|
|
#include "qemu/memfd.h"
|
|
#include "standard-headers/linux/udmabuf.h"
|
|
|
|
static void virtio_gpu_create_udmabuf(struct virtio_gpu_simple_resource *res)
|
|
{
|
|
struct udmabuf_create_list *list;
|
|
RAMBlock *rb;
|
|
ram_addr_t offset;
|
|
int udmabuf, i;
|
|
|
|
udmabuf = udmabuf_fd();
|
|
if (udmabuf < 0) {
|
|
return;
|
|
}
|
|
|
|
list = g_malloc0(sizeof(struct udmabuf_create_list) +
|
|
sizeof(struct udmabuf_create_item) * res->iov_cnt);
|
|
|
|
for (i = 0; i < res->iov_cnt; i++) {
|
|
rcu_read_lock();
|
|
rb = qemu_ram_block_from_host(res->iov[i].iov_base, false, &offset);
|
|
rcu_read_unlock();
|
|
|
|
if (!rb || rb->fd < 0) {
|
|
g_free(list);
|
|
return;
|
|
}
|
|
|
|
list->list[i].memfd = rb->fd;
|
|
list->list[i].offset = offset;
|
|
list->list[i].size = res->iov[i].iov_len;
|
|
}
|
|
|
|
list->count = res->iov_cnt;
|
|
list->flags = UDMABUF_FLAGS_CLOEXEC;
|
|
|
|
res->dmabuf_fd = ioctl(udmabuf, UDMABUF_CREATE_LIST, list);
|
|
if (res->dmabuf_fd < 0) {
|
|
warn_report("%s: UDMABUF_CREATE_LIST: %s", __func__,
|
|
strerror(errno));
|
|
}
|
|
g_free(list);
|
|
}
|
|
|
|
static void virtio_gpu_remap_udmabuf(struct virtio_gpu_simple_resource *res)
|
|
{
|
|
res->remapped = mmap(NULL, res->blob_size, PROT_READ,
|
|
MAP_SHARED, res->dmabuf_fd, 0);
|
|
if (res->remapped == MAP_FAILED) {
|
|
warn_report("%s: dmabuf mmap failed: %s", __func__,
|
|
strerror(errno));
|
|
res->remapped = NULL;
|
|
}
|
|
}
|
|
|
|
static void virtio_gpu_destroy_udmabuf(struct virtio_gpu_simple_resource *res)
|
|
{
|
|
if (res->remapped) {
|
|
munmap(res->remapped, res->blob_size);
|
|
res->remapped = NULL;
|
|
}
|
|
if (res->dmabuf_fd >= 0) {
|
|
close(res->dmabuf_fd);
|
|
res->dmabuf_fd = -1;
|
|
}
|
|
}
|
|
|
|
static int find_memory_backend_type(Object *obj, void *opaque)
|
|
{
|
|
bool *memfd_backend = opaque;
|
|
int ret;
|
|
|
|
if (object_dynamic_cast(obj, TYPE_MEMORY_BACKEND)) {
|
|
HostMemoryBackend *backend = MEMORY_BACKEND(obj);
|
|
RAMBlock *rb = backend->mr.ram_block;
|
|
|
|
if (rb && rb->fd > 0) {
|
|
ret = fcntl(rb->fd, F_GET_SEALS);
|
|
if (ret > 0) {
|
|
*memfd_backend = true;
|
|
}
|
|
}
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
bool virtio_gpu_have_udmabuf(void)
|
|
{
|
|
Object *memdev_root;
|
|
int udmabuf;
|
|
bool memfd_backend = false;
|
|
|
|
udmabuf = udmabuf_fd();
|
|
if (udmabuf < 0) {
|
|
return false;
|
|
}
|
|
|
|
memdev_root = object_resolve_path("/objects", NULL);
|
|
object_child_foreach(memdev_root, find_memory_backend_type, &memfd_backend);
|
|
|
|
return memfd_backend;
|
|
}
|
|
|
|
void virtio_gpu_init_udmabuf(struct virtio_gpu_simple_resource *res)
|
|
{
|
|
void *pdata = NULL;
|
|
|
|
res->dmabuf_fd = -1;
|
|
if (res->iov_cnt == 1) {
|
|
pdata = res->iov[0].iov_base;
|
|
} else {
|
|
virtio_gpu_create_udmabuf(res);
|
|
if (res->dmabuf_fd < 0) {
|
|
return;
|
|
}
|
|
virtio_gpu_remap_udmabuf(res);
|
|
if (!res->remapped) {
|
|
return;
|
|
}
|
|
pdata = res->remapped;
|
|
}
|
|
|
|
res->blob = pdata;
|
|
}
|
|
|
|
void virtio_gpu_fini_udmabuf(struct virtio_gpu_simple_resource *res)
|
|
{
|
|
if (res->remapped) {
|
|
virtio_gpu_destroy_udmabuf(res);
|
|
}
|
|
}
|
|
|
|
static void virtio_gpu_free_dmabuf(VirtIOGPU *g, VGPUDMABuf *dmabuf)
|
|
{
|
|
struct virtio_gpu_scanout *scanout;
|
|
|
|
scanout = &g->parent_obj.scanout[dmabuf->scanout_id];
|
|
dpy_gl_release_dmabuf(scanout->con, &dmabuf->buf);
|
|
QTAILQ_REMOVE(&g->dmabuf.bufs, dmabuf, next);
|
|
g_free(dmabuf);
|
|
}
|
|
|
|
static VGPUDMABuf
|
|
*virtio_gpu_create_dmabuf(VirtIOGPU *g,
|
|
uint32_t scanout_id,
|
|
struct virtio_gpu_simple_resource *res,
|
|
struct virtio_gpu_framebuffer *fb)
|
|
{
|
|
VGPUDMABuf *dmabuf;
|
|
|
|
if (res->dmabuf_fd < 0) {
|
|
return NULL;
|
|
}
|
|
|
|
dmabuf = g_new0(VGPUDMABuf, 1);
|
|
dmabuf->buf.width = fb->width;
|
|
dmabuf->buf.height = fb->height;
|
|
dmabuf->buf.stride = fb->stride;
|
|
dmabuf->buf.fourcc = qemu_pixman_to_drm_format(fb->format);
|
|
dmabuf->buf.fd = res->dmabuf_fd;
|
|
dmabuf->buf.allow_fences = true;
|
|
|
|
dmabuf->scanout_id = scanout_id;
|
|
QTAILQ_INSERT_HEAD(&g->dmabuf.bufs, dmabuf, next);
|
|
|
|
return dmabuf;
|
|
}
|
|
|
|
int virtio_gpu_update_dmabuf(VirtIOGPU *g,
|
|
uint32_t scanout_id,
|
|
struct virtio_gpu_simple_resource *res,
|
|
struct virtio_gpu_framebuffer *fb)
|
|
{
|
|
struct virtio_gpu_scanout *scanout = &g->parent_obj.scanout[scanout_id];
|
|
VGPUDMABuf *new_primary, *old_primary = NULL;
|
|
|
|
new_primary = virtio_gpu_create_dmabuf(g, scanout_id, res, fb);
|
|
if (!new_primary) {
|
|
return -EINVAL;
|
|
}
|
|
|
|
if (g->dmabuf.primary) {
|
|
old_primary = g->dmabuf.primary;
|
|
}
|
|
|
|
g->dmabuf.primary = new_primary;
|
|
qemu_console_resize(scanout->con,
|
|
new_primary->buf.width,
|
|
new_primary->buf.height);
|
|
dpy_gl_scanout_dmabuf(scanout->con, &new_primary->buf);
|
|
|
|
if (old_primary) {
|
|
virtio_gpu_free_dmabuf(g, old_primary);
|
|
}
|
|
|
|
return 0;
|
|
}
|