vfio/container: Switch to dma_map|unmap API
No functional change intended. Signed-off-by: Eric Auger <eric.auger@redhat.com> Signed-off-by: Yi Liu <yi.l.liu@intel.com> Signed-off-by: Yi Sun <yi.y.sun@linux.intel.com> Signed-off-by: Zhenzhong Duan <zhenzhong.duan@intel.com> Reviewed-by: Cédric Le Goater <clg@redhat.com> Signed-off-by: Cédric Le Goater <clg@redhat.com>
This commit is contained in:
parent
d246685791
commit
b08501a999
@ -292,7 +292,7 @@ static bool vfio_get_xlat_addr(IOMMUTLBEntry *iotlb, void **vaddr,
|
||||
static void vfio_iommu_map_notify(IOMMUNotifier *n, IOMMUTLBEntry *iotlb)
|
||||
{
|
||||
VFIOGuestIOMMU *giommu = container_of(n, VFIOGuestIOMMU, n);
|
||||
VFIOContainer *container = giommu->container;
|
||||
VFIOContainerBase *bcontainer = &giommu->container->bcontainer;
|
||||
hwaddr iova = iotlb->iova + giommu->iommu_offset;
|
||||
void *vaddr;
|
||||
int ret;
|
||||
@ -322,21 +322,22 @@ static void vfio_iommu_map_notify(IOMMUNotifier *n, IOMMUTLBEntry *iotlb)
|
||||
* of vaddr will always be there, even if the memory object is
|
||||
* destroyed and its backing memory munmap-ed.
|
||||
*/
|
||||
ret = vfio_dma_map(container, iova,
|
||||
ret = vfio_container_dma_map(bcontainer, iova,
|
||||
iotlb->addr_mask + 1, vaddr,
|
||||
read_only);
|
||||
if (ret) {
|
||||
error_report("vfio_dma_map(%p, 0x%"HWADDR_PRIx", "
|
||||
error_report("vfio_container_dma_map(%p, 0x%"HWADDR_PRIx", "
|
||||
"0x%"HWADDR_PRIx", %p) = %d (%s)",
|
||||
container, iova,
|
||||
bcontainer, iova,
|
||||
iotlb->addr_mask + 1, vaddr, ret, strerror(-ret));
|
||||
}
|
||||
} else {
|
||||
ret = vfio_dma_unmap(container, iova, iotlb->addr_mask + 1, iotlb);
|
||||
ret = vfio_container_dma_unmap(bcontainer, iova,
|
||||
iotlb->addr_mask + 1, iotlb);
|
||||
if (ret) {
|
||||
error_report("vfio_dma_unmap(%p, 0x%"HWADDR_PRIx", "
|
||||
error_report("vfio_container_dma_unmap(%p, 0x%"HWADDR_PRIx", "
|
||||
"0x%"HWADDR_PRIx") = %d (%s)",
|
||||
container, iova,
|
||||
bcontainer, iova,
|
||||
iotlb->addr_mask + 1, ret, strerror(-ret));
|
||||
vfio_set_migration_error(ret);
|
||||
}
|
||||
@ -355,9 +356,10 @@ static void vfio_ram_discard_notify_discard(RamDiscardListener *rdl,
|
||||
int ret;
|
||||
|
||||
/* Unmap with a single call. */
|
||||
ret = vfio_dma_unmap(vrdl->container, iova, size , NULL);
|
||||
ret = vfio_container_dma_unmap(&vrdl->container->bcontainer,
|
||||
iova, size , NULL);
|
||||
if (ret) {
|
||||
error_report("%s: vfio_dma_unmap() failed: %s", __func__,
|
||||
error_report("%s: vfio_container_dma_unmap() failed: %s", __func__,
|
||||
strerror(-ret));
|
||||
}
|
||||
}
|
||||
@ -385,8 +387,8 @@ static int vfio_ram_discard_notify_populate(RamDiscardListener *rdl,
|
||||
section->offset_within_address_space;
|
||||
vaddr = memory_region_get_ram_ptr(section->mr) + start;
|
||||
|
||||
ret = vfio_dma_map(vrdl->container, iova, next - start,
|
||||
vaddr, section->readonly);
|
||||
ret = vfio_container_dma_map(&vrdl->container->bcontainer, iova,
|
||||
next - start, vaddr, section->readonly);
|
||||
if (ret) {
|
||||
/* Rollback */
|
||||
vfio_ram_discard_notify_discard(rdl, section);
|
||||
@ -684,10 +686,11 @@ static void vfio_listener_region_add(MemoryListener *listener,
|
||||
}
|
||||
}
|
||||
|
||||
ret = vfio_dma_map(container, iova, int128_get64(llsize),
|
||||
vaddr, section->readonly);
|
||||
ret = vfio_container_dma_map(&container->bcontainer,
|
||||
iova, int128_get64(llsize), vaddr,
|
||||
section->readonly);
|
||||
if (ret) {
|
||||
error_setg(&err, "vfio_dma_map(%p, 0x%"HWADDR_PRIx", "
|
||||
error_setg(&err, "vfio_container_dma_map(%p, 0x%"HWADDR_PRIx", "
|
||||
"0x%"HWADDR_PRIx", %p) = %d (%s)",
|
||||
container, iova, int128_get64(llsize), vaddr, ret,
|
||||
strerror(-ret));
|
||||
@ -784,18 +787,20 @@ static void vfio_listener_region_del(MemoryListener *listener,
|
||||
if (int128_eq(llsize, int128_2_64())) {
|
||||
/* The unmap ioctl doesn't accept a full 64-bit span. */
|
||||
llsize = int128_rshift(llsize, 1);
|
||||
ret = vfio_dma_unmap(container, iova, int128_get64(llsize), NULL);
|
||||
ret = vfio_container_dma_unmap(&container->bcontainer, iova,
|
||||
int128_get64(llsize), NULL);
|
||||
if (ret) {
|
||||
error_report("vfio_dma_unmap(%p, 0x%"HWADDR_PRIx", "
|
||||
error_report("vfio_container_dma_unmap(%p, 0x%"HWADDR_PRIx", "
|
||||
"0x%"HWADDR_PRIx") = %d (%s)",
|
||||
container, iova, int128_get64(llsize), ret,
|
||||
strerror(-ret));
|
||||
}
|
||||
iova += int128_get64(llsize);
|
||||
}
|
||||
ret = vfio_dma_unmap(container, iova, int128_get64(llsize), NULL);
|
||||
ret = vfio_container_dma_unmap(&container->bcontainer, iova,
|
||||
int128_get64(llsize), NULL);
|
||||
if (ret) {
|
||||
error_report("vfio_dma_unmap(%p, 0x%"HWADDR_PRIx", "
|
||||
error_report("vfio_container_dma_unmap(%p, 0x%"HWADDR_PRIx", "
|
||||
"0x%"HWADDR_PRIx") = %d (%s)",
|
||||
container, iova, int128_get64(llsize), ret,
|
||||
strerror(-ret));
|
||||
|
32
hw/vfio/container-base.c
Normal file
32
hw/vfio/container-base.c
Normal file
@ -0,0 +1,32 @@
|
||||
/*
|
||||
* VFIO BASE CONTAINER
|
||||
*
|
||||
* Copyright (C) 2023 Intel Corporation.
|
||||
* Copyright Red Hat, Inc. 2023
|
||||
*
|
||||
* Authors: Yi Liu <yi.l.liu@intel.com>
|
||||
* Eric Auger <eric.auger@redhat.com>
|
||||
*
|
||||
* SPDX-License-Identifier: GPL-2.0-or-later
|
||||
*/
|
||||
|
||||
#include "qemu/osdep.h"
|
||||
#include "qapi/error.h"
|
||||
#include "qemu/error-report.h"
|
||||
#include "hw/vfio/vfio-container-base.h"
|
||||
|
||||
int vfio_container_dma_map(VFIOContainerBase *bcontainer,
|
||||
hwaddr iova, ram_addr_t size,
|
||||
void *vaddr, bool readonly)
|
||||
{
|
||||
g_assert(bcontainer->ops->dma_map);
|
||||
return bcontainer->ops->dma_map(bcontainer, iova, size, vaddr, readonly);
|
||||
}
|
||||
|
||||
int vfio_container_dma_unmap(VFIOContainerBase *bcontainer,
|
||||
hwaddr iova, ram_addr_t size,
|
||||
IOMMUTLBEntry *iotlb)
|
||||
{
|
||||
g_assert(bcontainer->ops->dma_unmap);
|
||||
return bcontainer->ops->dma_unmap(bcontainer, iova, size, iotlb);
|
||||
}
|
@ -115,9 +115,11 @@ unmap_exit:
|
||||
/*
|
||||
* DMA - Mapping and unmapping for the "type1" IOMMU interface used on x86
|
||||
*/
|
||||
int vfio_dma_unmap(VFIOContainer *container, hwaddr iova,
|
||||
static int vfio_legacy_dma_unmap(VFIOContainerBase *bcontainer, hwaddr iova,
|
||||
ram_addr_t size, IOMMUTLBEntry *iotlb)
|
||||
{
|
||||
VFIOContainer *container = container_of(bcontainer, VFIOContainer,
|
||||
bcontainer);
|
||||
struct vfio_iommu_type1_dma_unmap unmap = {
|
||||
.argsz = sizeof(unmap),
|
||||
.flags = 0,
|
||||
@ -151,7 +153,7 @@ int vfio_dma_unmap(VFIOContainer *container, hwaddr iova,
|
||||
*/
|
||||
if (errno == EINVAL && unmap.size && !(unmap.iova + unmap.size) &&
|
||||
container->iommu_type == VFIO_TYPE1v2_IOMMU) {
|
||||
trace_vfio_dma_unmap_overflow_workaround();
|
||||
trace_vfio_legacy_dma_unmap_overflow_workaround();
|
||||
unmap.size -= 1ULL << ctz64(container->pgsizes);
|
||||
continue;
|
||||
}
|
||||
@ -170,9 +172,11 @@ int vfio_dma_unmap(VFIOContainer *container, hwaddr iova,
|
||||
return 0;
|
||||
}
|
||||
|
||||
int vfio_dma_map(VFIOContainer *container, hwaddr iova,
|
||||
static int vfio_legacy_dma_map(VFIOContainerBase *bcontainer, hwaddr iova,
|
||||
ram_addr_t size, void *vaddr, bool readonly)
|
||||
{
|
||||
VFIOContainer *container = container_of(bcontainer, VFIOContainer,
|
||||
bcontainer);
|
||||
struct vfio_iommu_type1_dma_map map = {
|
||||
.argsz = sizeof(map),
|
||||
.flags = VFIO_DMA_MAP_FLAG_READ,
|
||||
@ -191,7 +195,8 @@ int vfio_dma_map(VFIOContainer *container, hwaddr iova,
|
||||
* the VGA ROM space.
|
||||
*/
|
||||
if (ioctl(container->fd, VFIO_IOMMU_MAP_DMA, &map) == 0 ||
|
||||
(errno == EBUSY && vfio_dma_unmap(container, iova, size, NULL) == 0 &&
|
||||
(errno == EBUSY &&
|
||||
vfio_legacy_dma_unmap(bcontainer, iova, size, NULL) == 0 &&
|
||||
ioctl(container->fd, VFIO_IOMMU_MAP_DMA, &map) == 0)) {
|
||||
return 0;
|
||||
}
|
||||
@ -937,4 +942,7 @@ void vfio_detach_device(VFIODevice *vbasedev)
|
||||
vfio_put_group(group);
|
||||
}
|
||||
|
||||
const VFIOIOMMUOps vfio_legacy_ops;
|
||||
const VFIOIOMMUOps vfio_legacy_ops = {
|
||||
.dma_map = vfio_legacy_dma_map,
|
||||
.dma_unmap = vfio_legacy_dma_unmap,
|
||||
};
|
||||
|
@ -2,6 +2,7 @@ vfio_ss = ss.source_set()
|
||||
vfio_ss.add(files(
|
||||
'helpers.c',
|
||||
'common.c',
|
||||
'container-base.c',
|
||||
'container.c',
|
||||
'spapr.c',
|
||||
'migration.c',
|
||||
|
@ -116,7 +116,7 @@ vfio_region_unmap(const char *name, unsigned long offset, unsigned long end) "Re
|
||||
vfio_region_sparse_mmap_header(const char *name, int index, int nr_areas) "Device %s region %d: %d sparse mmap entries"
|
||||
vfio_region_sparse_mmap_entry(int i, unsigned long start, unsigned long end) "sparse entry %d [0x%lx - 0x%lx]"
|
||||
vfio_get_dev_region(const char *name, int index, uint32_t type, uint32_t subtype) "%s index %d, %08x/%08x"
|
||||
vfio_dma_unmap_overflow_workaround(void) ""
|
||||
vfio_legacy_dma_unmap_overflow_workaround(void) ""
|
||||
vfio_get_dirty_bitmap(int fd, uint64_t iova, uint64_t size, uint64_t bitmap_size, uint64_t start, uint64_t dirty_pages) "container fd=%d, iova=0x%"PRIx64" size= 0x%"PRIx64" bitmap_size=0x%"PRIx64" start=0x%"PRIx64" dirty_pages=%"PRIu64
|
||||
vfio_iommu_map_dirty_notify(uint64_t iova_start, uint64_t iova_end) "iommu dirty @ 0x%"PRIx64" - 0x%"PRIx64
|
||||
|
||||
|
@ -208,10 +208,6 @@ void vfio_put_address_space(VFIOAddressSpace *space);
|
||||
bool vfio_devices_all_running_and_saving(VFIOContainer *container);
|
||||
|
||||
/* container->fd */
|
||||
int vfio_dma_unmap(VFIOContainer *container, hwaddr iova,
|
||||
ram_addr_t size, IOMMUTLBEntry *iotlb);
|
||||
int vfio_dma_map(VFIOContainer *container, hwaddr iova,
|
||||
ram_addr_t size, void *vaddr, bool readonly);
|
||||
int vfio_set_dirty_page_tracking(VFIOContainer *container, bool start);
|
||||
int vfio_query_dirty_bitmap(VFIOContainer *container, VFIOBitmap *vbmap,
|
||||
hwaddr iova, hwaddr size);
|
||||
|
@ -31,6 +31,13 @@ typedef struct VFIOContainerBase {
|
||||
const VFIOIOMMUOps *ops;
|
||||
} VFIOContainerBase;
|
||||
|
||||
int vfio_container_dma_map(VFIOContainerBase *bcontainer,
|
||||
hwaddr iova, ram_addr_t size,
|
||||
void *vaddr, bool readonly);
|
||||
int vfio_container_dma_unmap(VFIOContainerBase *bcontainer,
|
||||
hwaddr iova, ram_addr_t size,
|
||||
IOMMUTLBEntry *iotlb);
|
||||
|
||||
struct VFIOIOMMUOps {
|
||||
/* basic feature */
|
||||
int (*dma_map)(VFIOContainerBase *bcontainer,
|
||||
|
Loading…
Reference in New Issue
Block a user