410cefbd5c
Virtio-gpu malloc memory for the queue when it realized, but the queues was not released when it unrealized, which resulting in a memory leak. In addition, vm_change_state_handler is not cleaned up, which is related to vdev and will lead to segmentation fault when VM shutdown. Signed-off-by: wangmeiling <wangmeiling21@huawei.com> Signed-off-by: Binfeng Wu <wubinfeng@huawei.com> Message-Id: <7bbbc0f3-2ad9-83ca-b39b-f976d0837daf@huawei.com> Reviewed-by: Michael S. Tsirkin <mst@redhat.com> Signed-off-by: Michael S. Tsirkin <mst@redhat.com>
317 lines
9.0 KiB
C
317 lines
9.0 KiB
C
/*
|
|
* Virtio GPU Device
|
|
*
|
|
* Copyright Red Hat, Inc. 2013-2014
|
|
*
|
|
* Authors:
|
|
* Dave Airlie <airlied@redhat.com>
|
|
* Gerd Hoffmann <kraxel@redhat.com>
|
|
*
|
|
* This work is licensed under the terms of the GNU GPL, version 2 or later.
|
|
* See the COPYING file in the top-level directory.
|
|
*/
|
|
|
|
#include "qemu/osdep.h"
|
|
|
|
#include "hw/virtio/virtio-gpu.h"
|
|
#include "migration/blocker.h"
|
|
#include "qapi/error.h"
|
|
#include "qemu/error-report.h"
|
|
#include "hw/display/edid.h"
|
|
#include "trace.h"
|
|
|
|
void
|
|
virtio_gpu_base_reset(VirtIOGPUBase *g)
|
|
{
|
|
int i;
|
|
|
|
g->enable = 0;
|
|
|
|
for (i = 0; i < g->conf.max_outputs; i++) {
|
|
g->scanout[i].resource_id = 0;
|
|
g->scanout[i].width = 0;
|
|
g->scanout[i].height = 0;
|
|
g->scanout[i].x = 0;
|
|
g->scanout[i].y = 0;
|
|
g->scanout[i].ds = NULL;
|
|
}
|
|
}
|
|
|
|
void
|
|
virtio_gpu_base_fill_display_info(VirtIOGPUBase *g,
|
|
struct virtio_gpu_resp_display_info *dpy_info)
|
|
{
|
|
int i;
|
|
|
|
for (i = 0; i < g->conf.max_outputs; i++) {
|
|
if (g->enabled_output_bitmask & (1 << i)) {
|
|
dpy_info->pmodes[i].enabled = 1;
|
|
dpy_info->pmodes[i].r.width = cpu_to_le32(g->req_state[i].width);
|
|
dpy_info->pmodes[i].r.height = cpu_to_le32(g->req_state[i].height);
|
|
}
|
|
}
|
|
}
|
|
|
|
void
|
|
virtio_gpu_base_generate_edid(VirtIOGPUBase *g, int scanout,
|
|
struct virtio_gpu_resp_edid *edid)
|
|
{
|
|
qemu_edid_info info = {
|
|
.width_mm = g->req_state[scanout].width_mm,
|
|
.height_mm = g->req_state[scanout].height_mm,
|
|
.prefx = g->req_state[scanout].width,
|
|
.prefy = g->req_state[scanout].height,
|
|
.refresh_rate = g->req_state[scanout].refresh_rate,
|
|
};
|
|
|
|
edid->size = cpu_to_le32(sizeof(edid->edid));
|
|
qemu_edid_generate(edid->edid, sizeof(edid->edid), &info);
|
|
}
|
|
|
|
static void virtio_gpu_invalidate_display(void *opaque)
|
|
{
|
|
}
|
|
|
|
static void virtio_gpu_update_display(void *opaque)
|
|
{
|
|
}
|
|
|
|
static void virtio_gpu_text_update(void *opaque, console_ch_t *chardata)
|
|
{
|
|
}
|
|
|
|
static void virtio_gpu_notify_event(VirtIOGPUBase *g, uint32_t event_type)
|
|
{
|
|
g->virtio_config.events_read |= event_type;
|
|
virtio_notify_config(&g->parent_obj);
|
|
}
|
|
|
|
static void virtio_gpu_ui_info(void *opaque, uint32_t idx, QemuUIInfo *info)
|
|
{
|
|
VirtIOGPUBase *g = opaque;
|
|
|
|
if (idx >= g->conf.max_outputs) {
|
|
return;
|
|
}
|
|
|
|
g->req_state[idx].x = info->xoff;
|
|
g->req_state[idx].y = info->yoff;
|
|
g->req_state[idx].refresh_rate = info->refresh_rate;
|
|
g->req_state[idx].width = info->width;
|
|
g->req_state[idx].height = info->height;
|
|
g->req_state[idx].width_mm = info->width_mm;
|
|
g->req_state[idx].height_mm = info->height_mm;
|
|
|
|
if (info->width && info->height) {
|
|
g->enabled_output_bitmask |= (1 << idx);
|
|
} else {
|
|
g->enabled_output_bitmask &= ~(1 << idx);
|
|
}
|
|
|
|
/* send event to guest */
|
|
virtio_gpu_notify_event(g, VIRTIO_GPU_EVENT_DISPLAY);
|
|
return;
|
|
}
|
|
|
|
static void
|
|
virtio_gpu_gl_flushed(void *opaque)
|
|
{
|
|
VirtIOGPUBase *g = opaque;
|
|
VirtIOGPUBaseClass *vgc = VIRTIO_GPU_BASE_GET_CLASS(g);
|
|
|
|
if (vgc->gl_flushed) {
|
|
vgc->gl_flushed(g);
|
|
}
|
|
}
|
|
|
|
static void
|
|
virtio_gpu_gl_block(void *opaque, bool block)
|
|
{
|
|
VirtIOGPUBase *g = opaque;
|
|
|
|
if (block) {
|
|
g->renderer_blocked++;
|
|
} else {
|
|
g->renderer_blocked--;
|
|
}
|
|
assert(g->renderer_blocked >= 0);
|
|
|
|
if (!block && g->renderer_blocked == 0) {
|
|
virtio_gpu_gl_flushed(g);
|
|
}
|
|
}
|
|
|
|
static int
|
|
virtio_gpu_get_flags(void *opaque)
|
|
{
|
|
VirtIOGPUBase *g = opaque;
|
|
int flags = GRAPHIC_FLAGS_NONE;
|
|
|
|
if (virtio_gpu_virgl_enabled(g->conf)) {
|
|
flags |= GRAPHIC_FLAGS_GL;
|
|
}
|
|
|
|
if (virtio_gpu_dmabuf_enabled(g->conf)) {
|
|
flags |= GRAPHIC_FLAGS_DMABUF;
|
|
}
|
|
|
|
return flags;
|
|
}
|
|
|
|
static const GraphicHwOps virtio_gpu_ops = {
|
|
.get_flags = virtio_gpu_get_flags,
|
|
.invalidate = virtio_gpu_invalidate_display,
|
|
.gfx_update = virtio_gpu_update_display,
|
|
.text_update = virtio_gpu_text_update,
|
|
.ui_info = virtio_gpu_ui_info,
|
|
.gl_block = virtio_gpu_gl_block,
|
|
};
|
|
|
|
bool
|
|
virtio_gpu_base_device_realize(DeviceState *qdev,
|
|
VirtIOHandleOutput ctrl_cb,
|
|
VirtIOHandleOutput cursor_cb,
|
|
Error **errp)
|
|
{
|
|
VirtIODevice *vdev = VIRTIO_DEVICE(qdev);
|
|
VirtIOGPUBase *g = VIRTIO_GPU_BASE(qdev);
|
|
int i;
|
|
|
|
if (g->conf.max_outputs > VIRTIO_GPU_MAX_SCANOUTS) {
|
|
error_setg(errp, "invalid max_outputs > %d", VIRTIO_GPU_MAX_SCANOUTS);
|
|
return false;
|
|
}
|
|
|
|
if (virtio_gpu_virgl_enabled(g->conf)) {
|
|
error_setg(&g->migration_blocker, "virgl is not yet migratable");
|
|
if (migrate_add_blocker(&g->migration_blocker, errp) < 0) {
|
|
return false;
|
|
}
|
|
}
|
|
|
|
g->virtio_config.num_scanouts = cpu_to_le32(g->conf.max_outputs);
|
|
virtio_init(VIRTIO_DEVICE(g), VIRTIO_ID_GPU,
|
|
sizeof(struct virtio_gpu_config));
|
|
|
|
if (virtio_gpu_virgl_enabled(g->conf)) {
|
|
/* use larger control queue in 3d mode */
|
|
virtio_add_queue(vdev, 256, ctrl_cb);
|
|
virtio_add_queue(vdev, 16, cursor_cb);
|
|
} else {
|
|
virtio_add_queue(vdev, 64, ctrl_cb);
|
|
virtio_add_queue(vdev, 16, cursor_cb);
|
|
}
|
|
|
|
g->enabled_output_bitmask = 1;
|
|
|
|
g->req_state[0].width = g->conf.xres;
|
|
g->req_state[0].height = g->conf.yres;
|
|
|
|
g->hw_ops = &virtio_gpu_ops;
|
|
for (i = 0; i < g->conf.max_outputs; i++) {
|
|
g->scanout[i].con =
|
|
graphic_console_init(DEVICE(g), i, &virtio_gpu_ops, g);
|
|
}
|
|
|
|
return true;
|
|
}
|
|
|
|
static uint64_t
|
|
virtio_gpu_base_get_features(VirtIODevice *vdev, uint64_t features,
|
|
Error **errp)
|
|
{
|
|
VirtIOGPUBase *g = VIRTIO_GPU_BASE(vdev);
|
|
|
|
if (virtio_gpu_virgl_enabled(g->conf) ||
|
|
virtio_gpu_rutabaga_enabled(g->conf)) {
|
|
features |= (1 << VIRTIO_GPU_F_VIRGL);
|
|
}
|
|
if (virtio_gpu_edid_enabled(g->conf)) {
|
|
features |= (1 << VIRTIO_GPU_F_EDID);
|
|
}
|
|
if (virtio_gpu_blob_enabled(g->conf)) {
|
|
features |= (1 << VIRTIO_GPU_F_RESOURCE_BLOB);
|
|
}
|
|
if (virtio_gpu_context_init_enabled(g->conf)) {
|
|
features |= (1 << VIRTIO_GPU_F_CONTEXT_INIT);
|
|
}
|
|
|
|
return features;
|
|
}
|
|
|
|
static void
|
|
virtio_gpu_base_set_features(VirtIODevice *vdev, uint64_t features)
|
|
{
|
|
static const uint32_t virgl = (1 << VIRTIO_GPU_F_VIRGL);
|
|
|
|
trace_virtio_gpu_features(((features & virgl) == virgl));
|
|
}
|
|
|
|
void
|
|
virtio_gpu_base_device_unrealize(DeviceState *qdev)
|
|
{
|
|
VirtIOGPUBase *g = VIRTIO_GPU_BASE(qdev);
|
|
VirtIODevice *vdev = VIRTIO_DEVICE(qdev);
|
|
|
|
virtio_del_queue(vdev, 0);
|
|
virtio_del_queue(vdev, 1);
|
|
virtio_cleanup(vdev);
|
|
migrate_del_blocker(&g->migration_blocker);
|
|
}
|
|
|
|
static void
|
|
virtio_gpu_base_class_init(ObjectClass *klass, void *data)
|
|
{
|
|
DeviceClass *dc = DEVICE_CLASS(klass);
|
|
VirtioDeviceClass *vdc = VIRTIO_DEVICE_CLASS(klass);
|
|
|
|
vdc->unrealize = virtio_gpu_base_device_unrealize;
|
|
vdc->get_features = virtio_gpu_base_get_features;
|
|
vdc->set_features = virtio_gpu_base_set_features;
|
|
|
|
set_bit(DEVICE_CATEGORY_DISPLAY, dc->categories);
|
|
dc->hotpluggable = false;
|
|
}
|
|
|
|
static const TypeInfo virtio_gpu_base_info = {
|
|
.name = TYPE_VIRTIO_GPU_BASE,
|
|
.parent = TYPE_VIRTIO_DEVICE,
|
|
.instance_size = sizeof(VirtIOGPUBase),
|
|
.class_size = sizeof(VirtIOGPUBaseClass),
|
|
.class_init = virtio_gpu_base_class_init,
|
|
.abstract = true
|
|
};
|
|
module_obj(TYPE_VIRTIO_GPU_BASE);
|
|
module_kconfig(VIRTIO_GPU);
|
|
|
|
static void
|
|
virtio_register_types(void)
|
|
{
|
|
type_register_static(&virtio_gpu_base_info);
|
|
}
|
|
|
|
type_init(virtio_register_types)
|
|
|
|
QEMU_BUILD_BUG_ON(sizeof(struct virtio_gpu_ctrl_hdr) != 24);
|
|
QEMU_BUILD_BUG_ON(sizeof(struct virtio_gpu_update_cursor) != 56);
|
|
QEMU_BUILD_BUG_ON(sizeof(struct virtio_gpu_resource_unref) != 32);
|
|
QEMU_BUILD_BUG_ON(sizeof(struct virtio_gpu_resource_create_2d) != 40);
|
|
QEMU_BUILD_BUG_ON(sizeof(struct virtio_gpu_set_scanout) != 48);
|
|
QEMU_BUILD_BUG_ON(sizeof(struct virtio_gpu_resource_flush) != 48);
|
|
QEMU_BUILD_BUG_ON(sizeof(struct virtio_gpu_transfer_to_host_2d) != 56);
|
|
QEMU_BUILD_BUG_ON(sizeof(struct virtio_gpu_mem_entry) != 16);
|
|
QEMU_BUILD_BUG_ON(sizeof(struct virtio_gpu_resource_attach_backing) != 32);
|
|
QEMU_BUILD_BUG_ON(sizeof(struct virtio_gpu_resource_detach_backing) != 32);
|
|
QEMU_BUILD_BUG_ON(sizeof(struct virtio_gpu_resp_display_info) != 408);
|
|
|
|
QEMU_BUILD_BUG_ON(sizeof(struct virtio_gpu_transfer_host_3d) != 72);
|
|
QEMU_BUILD_BUG_ON(sizeof(struct virtio_gpu_resource_create_3d) != 72);
|
|
QEMU_BUILD_BUG_ON(sizeof(struct virtio_gpu_ctx_create) != 96);
|
|
QEMU_BUILD_BUG_ON(sizeof(struct virtio_gpu_ctx_destroy) != 24);
|
|
QEMU_BUILD_BUG_ON(sizeof(struct virtio_gpu_ctx_resource) != 32);
|
|
QEMU_BUILD_BUG_ON(sizeof(struct virtio_gpu_cmd_submit) != 32);
|
|
QEMU_BUILD_BUG_ON(sizeof(struct virtio_gpu_get_capset_info) != 32);
|
|
QEMU_BUILD_BUG_ON(sizeof(struct virtio_gpu_resp_capset_info) != 40);
|
|
QEMU_BUILD_BUG_ON(sizeof(struct virtio_gpu_get_capset) != 32);
|
|
QEMU_BUILD_BUG_ON(sizeof(struct virtio_gpu_resp_capset) != 24);
|