vfio/container: Switch to dma_map|unmap API

No functional change intended.

Signed-off-by: Eric Auger <eric.auger@redhat.com>
Signed-off-by: Yi Liu <yi.l.liu@intel.com>
Signed-off-by: Yi Sun <yi.y.sun@linux.intel.com>
Signed-off-by: Zhenzhong Duan <zhenzhong.duan@intel.com>
Reviewed-by: Cédric Le Goater <clg@redhat.com>
Signed-off-by: Cédric Le Goater <clg@redhat.com>
This commit is contained in:
Eric Auger 2023-11-02 15:12:29 +08:00 committed by Cédric Le Goater
parent d246685791
commit b08501a999
7 changed files with 81 additions and 32 deletions

View File

@ -292,7 +292,7 @@ static bool vfio_get_xlat_addr(IOMMUTLBEntry *iotlb, void **vaddr,
static void vfio_iommu_map_notify(IOMMUNotifier *n, IOMMUTLBEntry *iotlb) static void vfio_iommu_map_notify(IOMMUNotifier *n, IOMMUTLBEntry *iotlb)
{ {
VFIOGuestIOMMU *giommu = container_of(n, VFIOGuestIOMMU, n); VFIOGuestIOMMU *giommu = container_of(n, VFIOGuestIOMMU, n);
VFIOContainer *container = giommu->container; VFIOContainerBase *bcontainer = &giommu->container->bcontainer;
hwaddr iova = iotlb->iova + giommu->iommu_offset; hwaddr iova = iotlb->iova + giommu->iommu_offset;
void *vaddr; void *vaddr;
int ret; int ret;
@ -322,21 +322,22 @@ static void vfio_iommu_map_notify(IOMMUNotifier *n, IOMMUTLBEntry *iotlb)
* of vaddr will always be there, even if the memory object is * of vaddr will always be there, even if the memory object is
* destroyed and its backing memory munmap-ed. * destroyed and its backing memory munmap-ed.
*/ */
ret = vfio_dma_map(container, iova, ret = vfio_container_dma_map(bcontainer, iova,
iotlb->addr_mask + 1, vaddr, iotlb->addr_mask + 1, vaddr,
read_only); read_only);
if (ret) { if (ret) {
error_report("vfio_dma_map(%p, 0x%"HWADDR_PRIx", " error_report("vfio_container_dma_map(%p, 0x%"HWADDR_PRIx", "
"0x%"HWADDR_PRIx", %p) = %d (%s)", "0x%"HWADDR_PRIx", %p) = %d (%s)",
container, iova, bcontainer, iova,
iotlb->addr_mask + 1, vaddr, ret, strerror(-ret)); iotlb->addr_mask + 1, vaddr, ret, strerror(-ret));
} }
} else { } else {
ret = vfio_dma_unmap(container, iova, iotlb->addr_mask + 1, iotlb); ret = vfio_container_dma_unmap(bcontainer, iova,
iotlb->addr_mask + 1, iotlb);
if (ret) { if (ret) {
error_report("vfio_dma_unmap(%p, 0x%"HWADDR_PRIx", " error_report("vfio_container_dma_unmap(%p, 0x%"HWADDR_PRIx", "
"0x%"HWADDR_PRIx") = %d (%s)", "0x%"HWADDR_PRIx") = %d (%s)",
container, iova, bcontainer, iova,
iotlb->addr_mask + 1, ret, strerror(-ret)); iotlb->addr_mask + 1, ret, strerror(-ret));
vfio_set_migration_error(ret); vfio_set_migration_error(ret);
} }
@ -355,9 +356,10 @@ static void vfio_ram_discard_notify_discard(RamDiscardListener *rdl,
int ret; int ret;
/* Unmap with a single call. */ /* Unmap with a single call. */
ret = vfio_dma_unmap(vrdl->container, iova, size , NULL); ret = vfio_container_dma_unmap(&vrdl->container->bcontainer,
iova, size , NULL);
if (ret) { if (ret) {
error_report("%s: vfio_dma_unmap() failed: %s", __func__, error_report("%s: vfio_container_dma_unmap() failed: %s", __func__,
strerror(-ret)); strerror(-ret));
} }
} }
@ -385,8 +387,8 @@ static int vfio_ram_discard_notify_populate(RamDiscardListener *rdl,
section->offset_within_address_space; section->offset_within_address_space;
vaddr = memory_region_get_ram_ptr(section->mr) + start; vaddr = memory_region_get_ram_ptr(section->mr) + start;
ret = vfio_dma_map(vrdl->container, iova, next - start, ret = vfio_container_dma_map(&vrdl->container->bcontainer, iova,
vaddr, section->readonly); next - start, vaddr, section->readonly);
if (ret) { if (ret) {
/* Rollback */ /* Rollback */
vfio_ram_discard_notify_discard(rdl, section); vfio_ram_discard_notify_discard(rdl, section);
@ -684,10 +686,11 @@ static void vfio_listener_region_add(MemoryListener *listener,
} }
} }
ret = vfio_dma_map(container, iova, int128_get64(llsize), ret = vfio_container_dma_map(&container->bcontainer,
vaddr, section->readonly); iova, int128_get64(llsize), vaddr,
section->readonly);
if (ret) { if (ret) {
error_setg(&err, "vfio_dma_map(%p, 0x%"HWADDR_PRIx", " error_setg(&err, "vfio_container_dma_map(%p, 0x%"HWADDR_PRIx", "
"0x%"HWADDR_PRIx", %p) = %d (%s)", "0x%"HWADDR_PRIx", %p) = %d (%s)",
container, iova, int128_get64(llsize), vaddr, ret, container, iova, int128_get64(llsize), vaddr, ret,
strerror(-ret)); strerror(-ret));
@ -784,18 +787,20 @@ static void vfio_listener_region_del(MemoryListener *listener,
if (int128_eq(llsize, int128_2_64())) { if (int128_eq(llsize, int128_2_64())) {
/* The unmap ioctl doesn't accept a full 64-bit span. */ /* The unmap ioctl doesn't accept a full 64-bit span. */
llsize = int128_rshift(llsize, 1); llsize = int128_rshift(llsize, 1);
ret = vfio_dma_unmap(container, iova, int128_get64(llsize), NULL); ret = vfio_container_dma_unmap(&container->bcontainer, iova,
int128_get64(llsize), NULL);
if (ret) { if (ret) {
error_report("vfio_dma_unmap(%p, 0x%"HWADDR_PRIx", " error_report("vfio_container_dma_unmap(%p, 0x%"HWADDR_PRIx", "
"0x%"HWADDR_PRIx") = %d (%s)", "0x%"HWADDR_PRIx") = %d (%s)",
container, iova, int128_get64(llsize), ret, container, iova, int128_get64(llsize), ret,
strerror(-ret)); strerror(-ret));
} }
iova += int128_get64(llsize); iova += int128_get64(llsize);
} }
ret = vfio_dma_unmap(container, iova, int128_get64(llsize), NULL); ret = vfio_container_dma_unmap(&container->bcontainer, iova,
int128_get64(llsize), NULL);
if (ret) { if (ret) {
error_report("vfio_dma_unmap(%p, 0x%"HWADDR_PRIx", " error_report("vfio_container_dma_unmap(%p, 0x%"HWADDR_PRIx", "
"0x%"HWADDR_PRIx") = %d (%s)", "0x%"HWADDR_PRIx") = %d (%s)",
container, iova, int128_get64(llsize), ret, container, iova, int128_get64(llsize), ret,
strerror(-ret)); strerror(-ret));

32
hw/vfio/container-base.c Normal file
View File

@ -0,0 +1,32 @@
/*
* VFIO BASE CONTAINER
*
* Copyright (C) 2023 Intel Corporation.
* Copyright Red Hat, Inc. 2023
*
* Authors: Yi Liu <yi.l.liu@intel.com>
* Eric Auger <eric.auger@redhat.com>
*
* SPDX-License-Identifier: GPL-2.0-or-later
*/
#include "qemu/osdep.h"
#include "qapi/error.h"
#include "qemu/error-report.h"
#include "hw/vfio/vfio-container-base.h"
int vfio_container_dma_map(VFIOContainerBase *bcontainer,
hwaddr iova, ram_addr_t size,
void *vaddr, bool readonly)
{
g_assert(bcontainer->ops->dma_map);
return bcontainer->ops->dma_map(bcontainer, iova, size, vaddr, readonly);
}
int vfio_container_dma_unmap(VFIOContainerBase *bcontainer,
hwaddr iova, ram_addr_t size,
IOMMUTLBEntry *iotlb)
{
g_assert(bcontainer->ops->dma_unmap);
return bcontainer->ops->dma_unmap(bcontainer, iova, size, iotlb);
}

View File

@ -115,9 +115,11 @@ unmap_exit:
/* /*
* DMA - Mapping and unmapping for the "type1" IOMMU interface used on x86 * DMA - Mapping and unmapping for the "type1" IOMMU interface used on x86
*/ */
int vfio_dma_unmap(VFIOContainer *container, hwaddr iova, static int vfio_legacy_dma_unmap(VFIOContainerBase *bcontainer, hwaddr iova,
ram_addr_t size, IOMMUTLBEntry *iotlb) ram_addr_t size, IOMMUTLBEntry *iotlb)
{ {
VFIOContainer *container = container_of(bcontainer, VFIOContainer,
bcontainer);
struct vfio_iommu_type1_dma_unmap unmap = { struct vfio_iommu_type1_dma_unmap unmap = {
.argsz = sizeof(unmap), .argsz = sizeof(unmap),
.flags = 0, .flags = 0,
@ -151,7 +153,7 @@ int vfio_dma_unmap(VFIOContainer *container, hwaddr iova,
*/ */
if (errno == EINVAL && unmap.size && !(unmap.iova + unmap.size) && if (errno == EINVAL && unmap.size && !(unmap.iova + unmap.size) &&
container->iommu_type == VFIO_TYPE1v2_IOMMU) { container->iommu_type == VFIO_TYPE1v2_IOMMU) {
trace_vfio_dma_unmap_overflow_workaround(); trace_vfio_legacy_dma_unmap_overflow_workaround();
unmap.size -= 1ULL << ctz64(container->pgsizes); unmap.size -= 1ULL << ctz64(container->pgsizes);
continue; continue;
} }
@ -170,9 +172,11 @@ int vfio_dma_unmap(VFIOContainer *container, hwaddr iova,
return 0; return 0;
} }
int vfio_dma_map(VFIOContainer *container, hwaddr iova, static int vfio_legacy_dma_map(VFIOContainerBase *bcontainer, hwaddr iova,
ram_addr_t size, void *vaddr, bool readonly) ram_addr_t size, void *vaddr, bool readonly)
{ {
VFIOContainer *container = container_of(bcontainer, VFIOContainer,
bcontainer);
struct vfio_iommu_type1_dma_map map = { struct vfio_iommu_type1_dma_map map = {
.argsz = sizeof(map), .argsz = sizeof(map),
.flags = VFIO_DMA_MAP_FLAG_READ, .flags = VFIO_DMA_MAP_FLAG_READ,
@ -191,7 +195,8 @@ int vfio_dma_map(VFIOContainer *container, hwaddr iova,
* the VGA ROM space. * the VGA ROM space.
*/ */
if (ioctl(container->fd, VFIO_IOMMU_MAP_DMA, &map) == 0 || if (ioctl(container->fd, VFIO_IOMMU_MAP_DMA, &map) == 0 ||
(errno == EBUSY && vfio_dma_unmap(container, iova, size, NULL) == 0 && (errno == EBUSY &&
vfio_legacy_dma_unmap(bcontainer, iova, size, NULL) == 0 &&
ioctl(container->fd, VFIO_IOMMU_MAP_DMA, &map) == 0)) { ioctl(container->fd, VFIO_IOMMU_MAP_DMA, &map) == 0)) {
return 0; return 0;
} }
@ -937,4 +942,7 @@ void vfio_detach_device(VFIODevice *vbasedev)
vfio_put_group(group); vfio_put_group(group);
} }
const VFIOIOMMUOps vfio_legacy_ops; const VFIOIOMMUOps vfio_legacy_ops = {
.dma_map = vfio_legacy_dma_map,
.dma_unmap = vfio_legacy_dma_unmap,
};

View File

@ -2,6 +2,7 @@ vfio_ss = ss.source_set()
vfio_ss.add(files( vfio_ss.add(files(
'helpers.c', 'helpers.c',
'common.c', 'common.c',
'container-base.c',
'container.c', 'container.c',
'spapr.c', 'spapr.c',
'migration.c', 'migration.c',

View File

@ -116,7 +116,7 @@ vfio_region_unmap(const char *name, unsigned long offset, unsigned long end) "Re
vfio_region_sparse_mmap_header(const char *name, int index, int nr_areas) "Device %s region %d: %d sparse mmap entries" vfio_region_sparse_mmap_header(const char *name, int index, int nr_areas) "Device %s region %d: %d sparse mmap entries"
vfio_region_sparse_mmap_entry(int i, unsigned long start, unsigned long end) "sparse entry %d [0x%lx - 0x%lx]" vfio_region_sparse_mmap_entry(int i, unsigned long start, unsigned long end) "sparse entry %d [0x%lx - 0x%lx]"
vfio_get_dev_region(const char *name, int index, uint32_t type, uint32_t subtype) "%s index %d, %08x/%08x" vfio_get_dev_region(const char *name, int index, uint32_t type, uint32_t subtype) "%s index %d, %08x/%08x"
vfio_dma_unmap_overflow_workaround(void) "" vfio_legacy_dma_unmap_overflow_workaround(void) ""
vfio_get_dirty_bitmap(int fd, uint64_t iova, uint64_t size, uint64_t bitmap_size, uint64_t start, uint64_t dirty_pages) "container fd=%d, iova=0x%"PRIx64" size= 0x%"PRIx64" bitmap_size=0x%"PRIx64" start=0x%"PRIx64" dirty_pages=%"PRIu64 vfio_get_dirty_bitmap(int fd, uint64_t iova, uint64_t size, uint64_t bitmap_size, uint64_t start, uint64_t dirty_pages) "container fd=%d, iova=0x%"PRIx64" size= 0x%"PRIx64" bitmap_size=0x%"PRIx64" start=0x%"PRIx64" dirty_pages=%"PRIu64
vfio_iommu_map_dirty_notify(uint64_t iova_start, uint64_t iova_end) "iommu dirty @ 0x%"PRIx64" - 0x%"PRIx64 vfio_iommu_map_dirty_notify(uint64_t iova_start, uint64_t iova_end) "iommu dirty @ 0x%"PRIx64" - 0x%"PRIx64

View File

@ -208,10 +208,6 @@ void vfio_put_address_space(VFIOAddressSpace *space);
bool vfio_devices_all_running_and_saving(VFIOContainer *container); bool vfio_devices_all_running_and_saving(VFIOContainer *container);
/* container->fd */ /* container->fd */
int vfio_dma_unmap(VFIOContainer *container, hwaddr iova,
ram_addr_t size, IOMMUTLBEntry *iotlb);
int vfio_dma_map(VFIOContainer *container, hwaddr iova,
ram_addr_t size, void *vaddr, bool readonly);
int vfio_set_dirty_page_tracking(VFIOContainer *container, bool start); int vfio_set_dirty_page_tracking(VFIOContainer *container, bool start);
int vfio_query_dirty_bitmap(VFIOContainer *container, VFIOBitmap *vbmap, int vfio_query_dirty_bitmap(VFIOContainer *container, VFIOBitmap *vbmap,
hwaddr iova, hwaddr size); hwaddr iova, hwaddr size);

View File

@ -31,6 +31,13 @@ typedef struct VFIOContainerBase {
const VFIOIOMMUOps *ops; const VFIOIOMMUOps *ops;
} VFIOContainerBase; } VFIOContainerBase;
int vfio_container_dma_map(VFIOContainerBase *bcontainer,
hwaddr iova, ram_addr_t size,
void *vaddr, bool readonly);
int vfio_container_dma_unmap(VFIOContainerBase *bcontainer,
hwaddr iova, ram_addr_t size,
IOMMUTLBEntry *iotlb);
struct VFIOIOMMUOps { struct VFIOIOMMUOps {
/* basic feature */ /* basic feature */
int (*dma_map)(VFIOContainerBase *bcontainer, int (*dma_map)(VFIOContainerBase *bcontainer,