366844f3d1
The following MSI-X vectors are required: * VIRTIO Configuration Change * hiprio virtqueue * requests virtqueues Fix the calculation to reserve enough MSI-X vectors. Otherwise guest drivers fall back to a sub-optional configuration where all virtqueues share a single vector. This change does not break live migration compatibility since vhost-user-fs-pci devices are not migratable yet. Reported-by: Vivek Goyal <vgoyal@redhat.com> Signed-off-by: Stefan Hajnoczi <stefanha@redhat.com> Message-Id: <20191209110759.35227-1-stefanha@redhat.com> Reviewed-by: Dr. David Alan Gilbert <dgilbert@redhat.com> Reviewed-by: Michael S. Tsirkin <mst@redhat.com> Signed-off-by: Dr. David Alan Gilbert <dgilbert@redhat.com>
87 lines
2.6 KiB
C
87 lines
2.6 KiB
C
/*
|
|
* Vhost-user filesystem virtio device PCI glue
|
|
*
|
|
* Copyright 2018-2019 Red Hat, Inc.
|
|
*
|
|
* Authors:
|
|
* Dr. David Alan Gilbert <dgilbert@redhat.com>
|
|
*
|
|
* This work is licensed under the terms of the GNU GPL, version 2 or
|
|
* (at your option) any later version. See the COPYING file in the
|
|
* top-level directory.
|
|
*/
|
|
|
|
#include "qemu/osdep.h"
|
|
#include "hw/qdev-properties.h"
|
|
#include "hw/virtio/vhost-user-fs.h"
|
|
#include "virtio-pci.h"
|
|
|
|
struct VHostUserFSPCI {
|
|
VirtIOPCIProxy parent_obj;
|
|
VHostUserFS vdev;
|
|
};
|
|
|
|
typedef struct VHostUserFSPCI VHostUserFSPCI;
|
|
|
|
#define TYPE_VHOST_USER_FS_PCI "vhost-user-fs-pci-base"
|
|
|
|
#define VHOST_USER_FS_PCI(obj) \
|
|
OBJECT_CHECK(VHostUserFSPCI, (obj), TYPE_VHOST_USER_FS_PCI)
|
|
|
|
static Property vhost_user_fs_pci_properties[] = {
|
|
DEFINE_PROP_UINT32("vectors", VirtIOPCIProxy, nvectors,
|
|
DEV_NVECTORS_UNSPECIFIED),
|
|
DEFINE_PROP_END_OF_LIST(),
|
|
};
|
|
|
|
static void vhost_user_fs_pci_realize(VirtIOPCIProxy *vpci_dev, Error **errp)
|
|
{
|
|
VHostUserFSPCI *dev = VHOST_USER_FS_PCI(vpci_dev);
|
|
DeviceState *vdev = DEVICE(&dev->vdev);
|
|
|
|
if (vpci_dev->nvectors == DEV_NVECTORS_UNSPECIFIED) {
|
|
/* Also reserve config change and hiprio queue vectors */
|
|
vpci_dev->nvectors = dev->vdev.conf.num_request_queues + 2;
|
|
}
|
|
|
|
qdev_set_parent_bus(vdev, BUS(&vpci_dev->bus));
|
|
object_property_set_bool(OBJECT(vdev), true, "realized", errp);
|
|
}
|
|
|
|
static void vhost_user_fs_pci_class_init(ObjectClass *klass, void *data)
|
|
{
|
|
DeviceClass *dc = DEVICE_CLASS(klass);
|
|
VirtioPCIClass *k = VIRTIO_PCI_CLASS(klass);
|
|
PCIDeviceClass *pcidev_k = PCI_DEVICE_CLASS(klass);
|
|
k->realize = vhost_user_fs_pci_realize;
|
|
set_bit(DEVICE_CATEGORY_STORAGE, dc->categories);
|
|
dc->props = vhost_user_fs_pci_properties;
|
|
pcidev_k->vendor_id = PCI_VENDOR_ID_REDHAT_QUMRANET;
|
|
pcidev_k->device_id = 0; /* Set by virtio-pci based on virtio id */
|
|
pcidev_k->revision = 0x00;
|
|
pcidev_k->class_id = PCI_CLASS_STORAGE_OTHER;
|
|
}
|
|
|
|
static void vhost_user_fs_pci_instance_init(Object *obj)
|
|
{
|
|
VHostUserFSPCI *dev = VHOST_USER_FS_PCI(obj);
|
|
|
|
virtio_instance_init_common(obj, &dev->vdev, sizeof(dev->vdev),
|
|
TYPE_VHOST_USER_FS);
|
|
}
|
|
|
|
static const VirtioPCIDeviceTypeInfo vhost_user_fs_pci_info = {
|
|
.base_name = TYPE_VHOST_USER_FS_PCI,
|
|
.non_transitional_name = "vhost-user-fs-pci",
|
|
.instance_size = sizeof(VHostUserFSPCI),
|
|
.instance_init = vhost_user_fs_pci_instance_init,
|
|
.class_init = vhost_user_fs_pci_class_init,
|
|
};
|
|
|
|
static void vhost_user_fs_pci_register(void)
|
|
{
|
|
virtio_pci_types_register(&vhost_user_fs_pci_info);
|
|
}
|
|
|
|
type_init(vhost_user_fs_pci_register);
|