/* * vhost-net support * * Copyright Red Hat, Inc. 2010 * * Authors: * Michael S. Tsirkin * * This work is licensed under the terms of the GNU GPL, version 2. See * the COPYING file in the top-level directory. * * Contributions after 2012-01-13 are licensed under the terms of the * GNU GPL, version 2 or (at your option) any later version. */ #include "qemu/osdep.h" #include "net/net.h" #include "net/tap.h" #include "net/vhost-user.h" #include "net/vhost-vdpa.h" #include "standard-headers/linux/vhost_types.h" #include "hw/virtio/virtio-net.h" #include "net/vhost_net.h" #include "qapi/error.h" #include "qemu/error-report.h" #include "qemu/main-loop.h" #include #include #include #include "standard-headers/linux/virtio_ring.h" #include "hw/virtio/vhost.h" #include "hw/virtio/virtio-bus.h" #include "linux-headers/linux/vhost.h" /* Features supported by host kernel. */ static const int kernel_feature_bits[] = { VIRTIO_F_NOTIFY_ON_EMPTY, VIRTIO_RING_F_INDIRECT_DESC, VIRTIO_RING_F_EVENT_IDX, VIRTIO_NET_F_MRG_RXBUF, VIRTIO_F_VERSION_1, VIRTIO_NET_F_MTU, VIRTIO_F_IOMMU_PLATFORM, VIRTIO_F_RING_PACKED, VIRTIO_F_RING_RESET, VIRTIO_NET_F_HASH_REPORT, VHOST_INVALID_FEATURE_BIT }; /* Features supported by others. */ static const int user_feature_bits[] = { VIRTIO_F_NOTIFY_ON_EMPTY, VIRTIO_RING_F_INDIRECT_DESC, VIRTIO_RING_F_EVENT_IDX, VIRTIO_F_ANY_LAYOUT, VIRTIO_F_VERSION_1, VIRTIO_NET_F_CSUM, VIRTIO_NET_F_GUEST_CSUM, VIRTIO_NET_F_GSO, VIRTIO_NET_F_GUEST_TSO4, VIRTIO_NET_F_GUEST_TSO6, VIRTIO_NET_F_GUEST_ECN, VIRTIO_NET_F_GUEST_UFO, VIRTIO_NET_F_HOST_TSO4, VIRTIO_NET_F_HOST_TSO6, VIRTIO_NET_F_HOST_ECN, VIRTIO_NET_F_HOST_UFO, VIRTIO_NET_F_MRG_RXBUF, VIRTIO_NET_F_MTU, VIRTIO_F_IOMMU_PLATFORM, VIRTIO_F_RING_PACKED, VIRTIO_NET_F_RSS, VIRTIO_NET_F_HASH_REPORT, /* This bit implies RARP isn't sent by QEMU out of band */ VIRTIO_NET_F_GUEST_ANNOUNCE, VIRTIO_NET_F_MQ, VHOST_INVALID_FEATURE_BIT }; static const int *vhost_net_get_feature_bits(struct vhost_net *net) { const int *feature_bits = 0; switch (net->nc->info->type) { case NET_CLIENT_DRIVER_TAP: feature_bits = kernel_feature_bits; break; case NET_CLIENT_DRIVER_VHOST_USER: feature_bits = user_feature_bits; break; #ifdef CONFIG_VHOST_NET_VDPA case NET_CLIENT_DRIVER_VHOST_VDPA: feature_bits = vdpa_feature_bits; break; #endif default: error_report("Feature bits not defined for this type: %d", net->nc->info->type); break; } return feature_bits; } uint64_t vhost_net_get_features(struct vhost_net *net, uint64_t features) { return vhost_get_features(&net->dev, vhost_net_get_feature_bits(net), features); } int vhost_net_get_config(struct vhost_net *net, uint8_t *config, uint32_t config_len) { return vhost_dev_get_config(&net->dev, config, config_len, NULL); } int vhost_net_set_config(struct vhost_net *net, const uint8_t *data, uint32_t offset, uint32_t size, uint32_t flags) { return vhost_dev_set_config(&net->dev, data, offset, size, flags); } void vhost_net_ack_features(struct vhost_net *net, uint64_t features) { net->dev.acked_features = net->dev.backend_features; vhost_ack_features(&net->dev, vhost_net_get_feature_bits(net), features); } uint64_t vhost_net_get_max_queues(VHostNetState *net) { return net->dev.max_queues; } uint64_t vhost_net_get_acked_features(VHostNetState *net) { return net->dev.acked_features; } static int vhost_net_get_fd(NetClientState *backend) { switch (backend->info->type) { case NET_CLIENT_DRIVER_TAP: return tap_get_fd(backend); default: fprintf(stderr, "vhost-net requires tap backend\n"); return -ENOSYS; } } struct vhost_net *vhost_net_init(VhostNetOptions *options) { int r; bool backend_kernel = options->backend_type == VHOST_BACKEND_TYPE_KERNEL; struct vhost_net *net = g_new0(struct vhost_net, 1); uint64_t features = 0; Error *local_err = NULL; if (!options->net_backend) { fprintf(stderr, "vhost-net requires net backend to be setup\n"); goto fail; } net->nc = options->net_backend; net->dev.nvqs = options->nvqs; net->dev.max_queues = 1; net->dev.vqs = net->vqs; if (backend_kernel) { r = vhost_net_get_fd(options->net_backend); if (r < 0) { goto fail; } net->dev.backend_features = qemu_has_vnet_hdr(options->net_backend) ? 0 : (1ULL << VHOST_NET_F_VIRTIO_NET_HDR); net->backend = r; net->dev.protocol_features = 0; } else { net->dev.backend_features = 0; net->dev.protocol_features = 0; net->backend = -1; /* vhost-user needs vq_index to initiate a specific queue pair */ net->dev.vq_index = net->nc->queue_index * net->dev.nvqs; } r = vhost_dev_init(&net->dev, options->opaque, options->backend_type, options->busyloop_timeout, &local_err); if (r < 0) { error_report_err(local_err); goto fail; } if (backend_kernel) { if (!qemu_has_vnet_hdr_len(options->net_backend, sizeof(struct virtio_net_hdr_mrg_rxbuf))) { net->dev.features &= ~(1ULL << VIRTIO_NET_F_MRG_RXBUF); } if (~net->dev.features & net->dev.backend_features) { fprintf(stderr, "vhost lacks feature mask 0x%" PRIx64 " for backend\n", (uint64_t)(~net->dev.features & net->dev.backend_features)); goto fail; } } /* Set sane init value. Override when guest acks. */ #ifdef CONFIG_VHOST_NET_USER if (net->nc->info->type == NET_CLIENT_DRIVER_VHOST_USER) { features = vhost_user_get_acked_features(net->nc); if (~net->dev.features & features) { fprintf(stderr, "vhost lacks feature mask 0x%" PRIx64 " for backend\n", (uint64_t)(~net->dev.features & features)); goto fail; } } #endif vhost_net_ack_features(net, features); return net; fail: vhost_dev_cleanup(&net->dev); g_free(net); return NULL; } static void vhost_net_set_vq_index(struct vhost_net *net, int vq_index, int vq_index_end) { net->dev.vq_index = vq_index; net->dev.vq_index_end = vq_index_end; } static int vhost_net_start_one(struct vhost_net *net, VirtIODevice *dev) { struct vhost_vring_file file = { }; int r; if (net->nc->info->start) { r = net->nc->info->start(net->nc); if (r < 0) { return r; } } r = vhost_dev_enable_notifiers(&net->dev, dev); if (r < 0) { goto fail_notifiers; } r = vhost_dev_start(&net->dev, dev); if (r < 0) { goto fail_start; } if (net->nc->info->poll) { net->nc->info->poll(net->nc, false); } if (net->nc->info->type == NET_CLIENT_DRIVER_TAP) { qemu_set_fd_handler(net->backend, NULL, NULL, NULL); file.fd = net->backend; for (file.index = 0; file.index < net->dev.nvqs; ++file.index) { if (!virtio_queue_enabled(dev, net->dev.vq_index + file.index)) { /* Queue might not be ready for start */ continue; } r = vhost_net_set_backend(&net->dev, &file); if (r < 0) { r = -errno; goto fail; } } } if (net->nc->info->load) { r = net->nc->info->load(net->nc); if (r < 0) { goto fail; } } return 0; fail: file.fd = -1; if (net->nc->info->type == NET_CLIENT_DRIVER_TAP) { while (file.index-- > 0) { if (!virtio_queue_enabled(dev, net->dev.vq_index + file.index)) { /* Queue might not be ready for start */ continue; } int r = vhost_net_set_backend(&net->dev, &file); assert(r >= 0); } } if (net->nc->info->poll) { net->nc->info->poll(net->nc, true); } vhost_dev_stop(&net->dev, dev); fail_start: vhost_dev_disable_notifiers(&net->dev, dev); fail_notifiers: return r; } static void vhost_net_stop_one(struct vhost_net *net, VirtIODevice *dev) { struct vhost_vring_file file = { .fd = -1 }; if (net->nc->info->type == NET_CLIENT_DRIVER_TAP) { for (file.index = 0; file.index < net->dev.nvqs; ++file.index) { int r = vhost_net_set_backend(&net->dev, &file); assert(r >= 0); } } if (net->nc->info->poll) { net->nc->info->poll(net->nc, true); } vhost_dev_stop(&net->dev, dev); if (net->nc->info->stop) { net->nc->info->stop(net->nc); } vhost_dev_disable_notifiers(&net->dev, dev); } int vhost_net_start(VirtIODevice *dev, NetClientState *ncs, int data_queue_pairs, int cvq) { BusState *qbus = BUS(qdev_get_parent_bus(DEVICE(dev))); VirtioBusState *vbus = VIRTIO_BUS(qbus); VirtioBusClass *k = VIRTIO_BUS_GET_CLASS(vbus); int total_notifiers = data_queue_pairs * 2 + cvq; VirtIONet *n = VIRTIO_NET(dev); int nvhosts = data_queue_pairs + cvq; struct vhost_net *net; int r, e, i, index_end = data_queue_pairs * 2; NetClientState *peer; if (cvq) { index_end += 1; } if (!k->set_guest_notifiers) { error_report("binding does not support guest notifiers"); return -ENOSYS; } for (i = 0; i < nvhosts; i++) { if (i < data_queue_pairs) { peer = qemu_get_peer(ncs, i); } else { /* Control Virtqueue */ peer = qemu_get_peer(ncs, n->max_queue_pairs); } net = get_vhost_net(peer); vhost_net_set_vq_index(net, i * 2, index_end); /* Suppress the masking guest notifiers on vhost user * because vhost user doesn't interrupt masking/unmasking * properly. */ if (net->nc->info->type == NET_CLIENT_DRIVER_VHOST_USER) { dev->use_guest_notifier_mask = false; } } r = k->set_guest_notifiers(qbus->parent, total_notifiers, true); if (r < 0) { error_report("Error binding guest notifier: %d", -r); goto err; } for (i = 0; i < nvhosts; i++) { if (i < data_queue_pairs) { peer = qemu_get_peer(ncs, i); } else { peer = qemu_get_peer(ncs, n->max_queue_pairs); } r = vhost_net_start_one(get_vhost_net(peer), dev); if (r < 0) { goto err_start; } if (peer->vring_enable) { /* restore vring enable state */ r = vhost_set_vring_enable(peer, peer->vring_enable); if (r < 0) { vhost_net_stop_one(get_vhost_net(peer), dev); goto err_start; } } } return 0; err_start: while (--i >= 0) { peer = qemu_get_peer(ncs, i < data_queue_pairs ? i : n->max_queue_pairs); vhost_net_stop_one(get_vhost_net(peer), dev); } e = k->set_guest_notifiers(qbus->parent, total_notifiers, false); if (e < 0) { fprintf(stderr, "vhost guest notifier cleanup failed: %d\n", e); fflush(stderr); } err: return r; } void vhost_net_stop(VirtIODevice *dev, NetClientState *ncs, int data_queue_pairs, int cvq) { BusState *qbus = BUS(qdev_get_parent_bus(DEVICE(dev))); VirtioBusState *vbus = VIRTIO_BUS(qbus); VirtioBusClass *k = VIRTIO_BUS_GET_CLASS(vbus); VirtIONet *n = VIRTIO_NET(dev); NetClientState *peer; int total_notifiers = data_queue_pairs * 2 + cvq; int nvhosts = data_queue_pairs + cvq; int i, r; for (i = 0; i < nvhosts; i++) { if (i < data_queue_pairs) { peer = qemu_get_peer(ncs, i); } else { peer = qemu_get_peer(ncs, n->max_queue_pairs); } vhost_net_stop_one(get_vhost_net(peer), dev); } r = k->set_guest_notifiers(qbus->parent, total_notifiers, false); if (r < 0) { fprintf(stderr, "vhost guest notifier cleanup failed: %d\n", r); fflush(stderr); } assert(r >= 0); } void vhost_net_cleanup(struct vhost_net *net) { vhost_dev_cleanup(&net->dev); } int vhost_net_notify_migration_done(struct vhost_net *net, char* mac_addr) { const VhostOps *vhost_ops = net->dev.vhost_ops; assert(vhost_ops->backend_type == VHOST_BACKEND_TYPE_USER); assert(vhost_ops->vhost_migration_done); return vhost_ops->vhost_migration_done(&net->dev, mac_addr); } bool vhost_net_virtqueue_pending(VHostNetState *net, int idx) { return vhost_virtqueue_pending(&net->dev, idx); } void vhost_net_virtqueue_mask(VHostNetState *net, VirtIODevice *dev, int idx, bool mask) { vhost_virtqueue_mask(&net->dev, dev, idx, mask); } VHostNetState *get_vhost_net(NetClientState *nc) { VHostNetState *vhost_net = 0; if (!nc) { return 0; } switch (nc->info->type) { case NET_CLIENT_DRIVER_TAP: vhost_net = tap_get_vhost_net(nc); break; #ifdef CONFIG_VHOST_NET_USER case NET_CLIENT_DRIVER_VHOST_USER: vhost_net = vhost_user_get_vhost_net(nc); assert(vhost_net); break; #endif #ifdef CONFIG_VHOST_NET_VDPA case NET_CLIENT_DRIVER_VHOST_VDPA: vhost_net = vhost_vdpa_get_vhost_net(nc); assert(vhost_net); break; #endif default: break; } return vhost_net; } int vhost_set_vring_enable(NetClientState *nc, int enable) { VHostNetState *net = get_vhost_net(nc); const VhostOps *vhost_ops = net->dev.vhost_ops; nc->vring_enable = enable; if (vhost_ops && vhost_ops->vhost_set_vring_enable) { return vhost_ops->vhost_set_vring_enable(&net->dev, enable); } return 0; } int vhost_net_set_mtu(struct vhost_net *net, uint16_t mtu) { const VhostOps *vhost_ops = net->dev.vhost_ops; if (!vhost_ops->vhost_net_set_mtu) { return 0; } return vhost_ops->vhost_net_set_mtu(&net->dev, mtu); } void vhost_net_virtqueue_reset(VirtIODevice *vdev, NetClientState *nc, int vq_index) { VHostNetState *net = get_vhost_net(nc->peer); const VhostOps *vhost_ops = net->dev.vhost_ops; struct vhost_vring_file file = { .fd = -1 }; int idx; /* should only be called after backend is connected */ assert(vhost_ops); idx = vhost_ops->vhost_get_vq_index(&net->dev, vq_index); if (net->nc->info->type == NET_CLIENT_DRIVER_TAP) { file.index = idx; int r = vhost_net_set_backend(&net->dev, &file); assert(r >= 0); } vhost_virtqueue_stop(&net->dev, vdev, net->dev.vqs + idx, net->dev.vq_index + idx); } int vhost_net_virtqueue_restart(VirtIODevice *vdev, NetClientState *nc, int vq_index) { VHostNetState *net = get_vhost_net(nc->peer); const VhostOps *vhost_ops = net->dev.vhost_ops; struct vhost_vring_file file = { }; int idx, r; if (!net->dev.started) { return -EBUSY; } /* should only be called after backend is connected */ assert(vhost_ops); idx = vhost_ops->vhost_get_vq_index(&net->dev, vq_index); r = vhost_virtqueue_start(&net->dev, vdev, net->dev.vqs + idx, net->dev.vq_index + idx); if (r < 0) { goto err_start; } if (net->nc->info->type == NET_CLIENT_DRIVER_TAP) { file.index = idx; file.fd = net->backend; r = vhost_net_set_backend(&net->dev, &file); if (r < 0) { r = -errno; goto err_start; } } return 0; err_start: error_report("Error when restarting the queue."); if (net->nc->info->type == NET_CLIENT_DRIVER_TAP) { file.fd = VHOST_FILE_UNBIND; file.index = idx; int r = vhost_net_set_backend(&net->dev, &file); assert(r >= 0); } vhost_dev_stop(&net->dev, vdev); return r; }