a2335113ae
We want to support memory devices that can automatically decide how many memslots they will use. In the worst case, they have to use a single memslot. The target use cases are virtio-mem and the hyper-v balloon. Let's calculate a reasonable limit such a memory device may use, and instruct the device to make a decision based on that limit. Use a simple heuristic that considers: * A memslot soft-limit for all memory devices of 256; also, to not consume too many memslots -- which could harm performance. * Actually still free and unreserved memslots * The percentage of the remaining device memory region that memory device will occupy. Further, while we properly check before plugging a memory device whether there still is are free memslots, we have other memslot consumers (such as boot memory, PCI BARs) that don't perform any checks and might dynamically consume memslots without any prior reservation. So we might succeed in plugging a memory device, but once we dynamically map a PCI BAR we would be in trouble. Doing accounting / reservation / checks for all such users is problematic (e.g., sometimes we might temporarily split boot memory into two memslots, triggered by the BIOS). We use the historic magic memslot number of 509 as orientation to when supporting 256 memory devices -> memslots (leaving 253 for boot memory and other devices) has been proven to work reliable. We'll fallback to suggesting a single memslot if we don't have at least 509 total memslots. Plugging vhost devices with less than 509 memslots available while we have memory devices plugged that consume multiple memslots due to automatic decisions can be problematic. Most configurations might just fail due to "limit < used + reserved", however, it can also happen that these memory devices would suddenly consume memslots that would actually be required by other memslot consumers (boot, PCI BARs) later. Note that this has always been sketchy with vhost devices that support only a small number of memslots; but we don't want to make it any worse.So let's keep it simple and simply reject plugging such vhost devices in such a configuration. Eventually, all vhost devices that want to be fully compatible with such memory devices should support a decent number of memslots (>= 509). Message-ID: <20230926185738.277351-13-david@redhat.com> Reviewed-by: Maciej S. Szmigiero <maciej.szmigiero@oracle.com> Reviewed-by: Michael S. Tsirkin <mst@redhat.com> Signed-off-by: David Hildenbrand <david@redhat.com>
521 lines
17 KiB
C
521 lines
17 KiB
C
/*
|
|
* Memory Device Interface
|
|
*
|
|
* Copyright ProfitBricks GmbH 2012
|
|
* Copyright (C) 2014 Red Hat Inc
|
|
* Copyright (c) 2018 Red Hat Inc
|
|
*
|
|
* This work is licensed under the terms of the GNU GPL, version 2 or later.
|
|
* See the COPYING file in the top-level directory.
|
|
*/
|
|
|
|
#include "qemu/osdep.h"
|
|
#include "qemu/error-report.h"
|
|
#include "hw/mem/memory-device.h"
|
|
#include "qapi/error.h"
|
|
#include "hw/boards.h"
|
|
#include "qemu/range.h"
|
|
#include "hw/virtio/vhost.h"
|
|
#include "sysemu/kvm.h"
|
|
#include "exec/address-spaces.h"
|
|
#include "trace.h"
|
|
|
|
static gint memory_device_addr_sort(gconstpointer a, gconstpointer b)
|
|
{
|
|
const MemoryDeviceState *md_a = MEMORY_DEVICE(a);
|
|
const MemoryDeviceState *md_b = MEMORY_DEVICE(b);
|
|
const MemoryDeviceClass *mdc_a = MEMORY_DEVICE_GET_CLASS(a);
|
|
const MemoryDeviceClass *mdc_b = MEMORY_DEVICE_GET_CLASS(b);
|
|
const uint64_t addr_a = mdc_a->get_addr(md_a);
|
|
const uint64_t addr_b = mdc_b->get_addr(md_b);
|
|
|
|
if (addr_a > addr_b) {
|
|
return 1;
|
|
} else if (addr_a < addr_b) {
|
|
return -1;
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
static int memory_device_build_list(Object *obj, void *opaque)
|
|
{
|
|
GSList **list = opaque;
|
|
|
|
if (object_dynamic_cast(obj, TYPE_MEMORY_DEVICE)) {
|
|
DeviceState *dev = DEVICE(obj);
|
|
if (dev->realized) { /* only realized memory devices matter */
|
|
*list = g_slist_insert_sorted(*list, dev, memory_device_addr_sort);
|
|
}
|
|
}
|
|
|
|
object_child_foreach(obj, memory_device_build_list, opaque);
|
|
return 0;
|
|
}
|
|
|
|
static unsigned int memory_device_get_memslots(MemoryDeviceState *md)
|
|
{
|
|
const MemoryDeviceClass *mdc = MEMORY_DEVICE_GET_CLASS(md);
|
|
|
|
if (mdc->get_memslots) {
|
|
return mdc->get_memslots(md);
|
|
}
|
|
return 1;
|
|
}
|
|
|
|
/*
|
|
* Memslots that are reserved by memory devices (required but still reported
|
|
* as free from KVM / vhost).
|
|
*/
|
|
static unsigned int get_reserved_memslots(MachineState *ms)
|
|
{
|
|
if (ms->device_memory->used_memslots >
|
|
ms->device_memory->required_memslots) {
|
|
/* This is unexpected, and we warned already in the memory notifier. */
|
|
return 0;
|
|
}
|
|
return ms->device_memory->required_memslots -
|
|
ms->device_memory->used_memslots;
|
|
}
|
|
|
|
unsigned int memory_devices_get_reserved_memslots(void)
|
|
{
|
|
if (!current_machine->device_memory) {
|
|
return 0;
|
|
}
|
|
return get_reserved_memslots(current_machine);
|
|
}
|
|
|
|
bool memory_devices_memslot_auto_decision_active(void)
|
|
{
|
|
if (!current_machine->device_memory) {
|
|
return false;
|
|
}
|
|
|
|
return current_machine->device_memory->memslot_auto_decision_active;
|
|
}
|
|
|
|
static unsigned int memory_device_memslot_decision_limit(MachineState *ms,
|
|
MemoryRegion *mr)
|
|
{
|
|
const unsigned int reserved = get_reserved_memslots(ms);
|
|
const uint64_t size = memory_region_size(mr);
|
|
unsigned int max = vhost_get_max_memslots();
|
|
unsigned int free = vhost_get_free_memslots();
|
|
uint64_t available_space;
|
|
unsigned int memslots;
|
|
|
|
if (kvm_enabled()) {
|
|
max = MIN(max, kvm_get_max_memslots());
|
|
free = MIN(free, kvm_get_free_memslots());
|
|
}
|
|
|
|
/*
|
|
* If we only have less overall memslots than what we consider reasonable,
|
|
* just keep it to a minimum.
|
|
*/
|
|
if (max < MEMORY_DEVICES_SAFE_MAX_MEMSLOTS) {
|
|
return 1;
|
|
}
|
|
|
|
/*
|
|
* Consider our soft-limit across all memory devices. We don't really
|
|
* expect to exceed this limit in reasonable configurations.
|
|
*/
|
|
if (MEMORY_DEVICES_SOFT_MEMSLOT_LIMIT <=
|
|
ms->device_memory->required_memslots) {
|
|
return 1;
|
|
}
|
|
memslots = MEMORY_DEVICES_SOFT_MEMSLOT_LIMIT -
|
|
ms->device_memory->required_memslots;
|
|
|
|
/*
|
|
* Consider the actually still free memslots. This is only relevant if
|
|
* other memslot consumers would consume *significantly* more memslots than
|
|
* what we prepared for (> 253). Unlikely, but let's just handle it
|
|
* cleanly.
|
|
*/
|
|
memslots = MIN(memslots, free - reserved);
|
|
if (memslots < 1 || unlikely(free < reserved)) {
|
|
return 1;
|
|
}
|
|
|
|
/* We cannot have any other memory devices? So give all to this device. */
|
|
if (size == ms->maxram_size - ms->ram_size) {
|
|
return memslots;
|
|
}
|
|
|
|
/*
|
|
* Simple heuristic: equally distribute the memslots over the space
|
|
* still available for memory devices.
|
|
*/
|
|
available_space = ms->maxram_size - ms->ram_size -
|
|
ms->device_memory->used_region_size;
|
|
memslots = (double)memslots * size / available_space;
|
|
return memslots < 1 ? 1 : memslots;
|
|
}
|
|
|
|
static void memory_device_check_addable(MachineState *ms, MemoryDeviceState *md,
|
|
MemoryRegion *mr, Error **errp)
|
|
{
|
|
const MemoryDeviceClass *mdc = MEMORY_DEVICE_GET_CLASS(md);
|
|
const uint64_t used_region_size = ms->device_memory->used_region_size;
|
|
const uint64_t size = memory_region_size(mr);
|
|
const unsigned int reserved_memslots = get_reserved_memslots(ms);
|
|
unsigned int required_memslots, memslot_limit;
|
|
|
|
/*
|
|
* Instruct the device to decide how many memslots to use, if applicable,
|
|
* before we query the number of required memslots the first time.
|
|
*/
|
|
if (mdc->decide_memslots) {
|
|
memslot_limit = memory_device_memslot_decision_limit(ms, mr);
|
|
mdc->decide_memslots(md, memslot_limit);
|
|
}
|
|
required_memslots = memory_device_get_memslots(md);
|
|
|
|
/* we will need memory slots for kvm and vhost */
|
|
if (kvm_enabled() &&
|
|
kvm_get_free_memslots() < required_memslots + reserved_memslots) {
|
|
error_setg(errp, "hypervisor has not enough free memory slots left");
|
|
return;
|
|
}
|
|
if (vhost_get_free_memslots() < required_memslots + reserved_memslots) {
|
|
error_setg(errp, "a used vhost backend has not enough free memory slots left");
|
|
return;
|
|
}
|
|
|
|
/* will we exceed the total amount of memory specified */
|
|
if (used_region_size + size < used_region_size ||
|
|
used_region_size + size > ms->maxram_size - ms->ram_size) {
|
|
error_setg(errp, "not enough space, currently 0x%" PRIx64
|
|
" in use of total space for memory devices 0x" RAM_ADDR_FMT,
|
|
used_region_size, ms->maxram_size - ms->ram_size);
|
|
return;
|
|
}
|
|
|
|
}
|
|
|
|
static uint64_t memory_device_get_free_addr(MachineState *ms,
|
|
const uint64_t *hint,
|
|
uint64_t align, uint64_t size,
|
|
Error **errp)
|
|
{
|
|
GSList *list = NULL, *item;
|
|
Range as, new = range_empty;
|
|
|
|
range_init_nofail(&as, ms->device_memory->base,
|
|
memory_region_size(&ms->device_memory->mr));
|
|
|
|
/* start of address space indicates the maximum alignment we expect */
|
|
if (!QEMU_IS_ALIGNED(range_lob(&as), align)) {
|
|
warn_report("the alignment (0x%" PRIx64 ") exceeds the expected"
|
|
" maximum alignment, memory will get fragmented and not"
|
|
" all 'maxmem' might be usable for memory devices.",
|
|
align);
|
|
}
|
|
|
|
if (hint && !QEMU_IS_ALIGNED(*hint, align)) {
|
|
error_setg(errp, "address must be aligned to 0x%" PRIx64 " bytes",
|
|
align);
|
|
return 0;
|
|
}
|
|
|
|
if (!QEMU_IS_ALIGNED(size, align)) {
|
|
error_setg(errp, "backend memory size must be multiple of 0x%"
|
|
PRIx64, align);
|
|
return 0;
|
|
}
|
|
|
|
if (hint) {
|
|
if (range_init(&new, *hint, size) || !range_contains_range(&as, &new)) {
|
|
error_setg(errp, "can't add memory device [0x%" PRIx64 ":0x%" PRIx64
|
|
"], usable range for memory devices [0x%" PRIx64 ":0x%"
|
|
PRIx64 "]", *hint, size, range_lob(&as),
|
|
range_size(&as));
|
|
return 0;
|
|
}
|
|
} else {
|
|
if (range_init(&new, QEMU_ALIGN_UP(range_lob(&as), align), size)) {
|
|
error_setg(errp, "can't add memory device, device too big");
|
|
return 0;
|
|
}
|
|
}
|
|
|
|
/* find address range that will fit new memory device */
|
|
object_child_foreach(OBJECT(ms), memory_device_build_list, &list);
|
|
for (item = list; item; item = g_slist_next(item)) {
|
|
const MemoryDeviceState *md = item->data;
|
|
const MemoryDeviceClass *mdc = MEMORY_DEVICE_GET_CLASS(OBJECT(md));
|
|
uint64_t next_addr;
|
|
Range tmp;
|
|
|
|
range_init_nofail(&tmp, mdc->get_addr(md),
|
|
memory_device_get_region_size(md, &error_abort));
|
|
|
|
if (range_overlaps_range(&tmp, &new)) {
|
|
if (hint) {
|
|
const DeviceState *d = DEVICE(md);
|
|
error_setg(errp, "address range conflicts with memory device"
|
|
" id='%s'", d->id ? d->id : "(unnamed)");
|
|
goto out;
|
|
}
|
|
|
|
next_addr = QEMU_ALIGN_UP(range_upb(&tmp) + 1, align);
|
|
if (!next_addr || range_init(&new, next_addr, range_size(&new))) {
|
|
range_make_empty(&new);
|
|
break;
|
|
}
|
|
} else if (range_lob(&tmp) > range_upb(&new)) {
|
|
break;
|
|
}
|
|
}
|
|
|
|
if (!range_contains_range(&as, &new)) {
|
|
error_setg(errp, "could not find position in guest address space for "
|
|
"memory device - memory fragmented due to alignments");
|
|
}
|
|
out:
|
|
g_slist_free(list);
|
|
return range_lob(&new);
|
|
}
|
|
|
|
MemoryDeviceInfoList *qmp_memory_device_list(void)
|
|
{
|
|
GSList *devices = NULL, *item;
|
|
MemoryDeviceInfoList *list = NULL, **tail = &list;
|
|
|
|
object_child_foreach(qdev_get_machine(), memory_device_build_list,
|
|
&devices);
|
|
|
|
for (item = devices; item; item = g_slist_next(item)) {
|
|
const MemoryDeviceState *md = MEMORY_DEVICE(item->data);
|
|
const MemoryDeviceClass *mdc = MEMORY_DEVICE_GET_CLASS(item->data);
|
|
MemoryDeviceInfo *info = g_new0(MemoryDeviceInfo, 1);
|
|
|
|
mdc->fill_device_info(md, info);
|
|
|
|
QAPI_LIST_APPEND(tail, info);
|
|
}
|
|
|
|
g_slist_free(devices);
|
|
|
|
return list;
|
|
}
|
|
|
|
static int memory_device_plugged_size(Object *obj, void *opaque)
|
|
{
|
|
uint64_t *size = opaque;
|
|
|
|
if (object_dynamic_cast(obj, TYPE_MEMORY_DEVICE)) {
|
|
const DeviceState *dev = DEVICE(obj);
|
|
const MemoryDeviceState *md = MEMORY_DEVICE(obj);
|
|
const MemoryDeviceClass *mdc = MEMORY_DEVICE_GET_CLASS(obj);
|
|
|
|
if (dev->realized) {
|
|
*size += mdc->get_plugged_size(md, &error_abort);
|
|
}
|
|
}
|
|
|
|
object_child_foreach(obj, memory_device_plugged_size, opaque);
|
|
return 0;
|
|
}
|
|
|
|
uint64_t get_plugged_memory_size(void)
|
|
{
|
|
uint64_t size = 0;
|
|
|
|
memory_device_plugged_size(qdev_get_machine(), &size);
|
|
|
|
return size;
|
|
}
|
|
|
|
void memory_device_pre_plug(MemoryDeviceState *md, MachineState *ms,
|
|
const uint64_t *legacy_align, Error **errp)
|
|
{
|
|
const MemoryDeviceClass *mdc = MEMORY_DEVICE_GET_CLASS(md);
|
|
Error *local_err = NULL;
|
|
uint64_t addr, align = 0;
|
|
MemoryRegion *mr;
|
|
|
|
if (!ms->device_memory) {
|
|
error_setg(errp, "the configuration is not prepared for memory devices"
|
|
" (e.g., for memory hotplug), consider specifying the"
|
|
" maxmem option");
|
|
return;
|
|
}
|
|
|
|
mr = mdc->get_memory_region(md, &local_err);
|
|
if (local_err) {
|
|
goto out;
|
|
}
|
|
|
|
memory_device_check_addable(ms, md, mr, &local_err);
|
|
if (local_err) {
|
|
goto out;
|
|
}
|
|
|
|
if (legacy_align) {
|
|
align = *legacy_align;
|
|
} else {
|
|
if (mdc->get_min_alignment) {
|
|
align = mdc->get_min_alignment(md);
|
|
}
|
|
align = MAX(align, memory_region_get_alignment(mr));
|
|
}
|
|
addr = mdc->get_addr(md);
|
|
addr = memory_device_get_free_addr(ms, !addr ? NULL : &addr, align,
|
|
memory_region_size(mr), &local_err);
|
|
if (local_err) {
|
|
goto out;
|
|
}
|
|
mdc->set_addr(md, addr, &local_err);
|
|
if (!local_err) {
|
|
trace_memory_device_pre_plug(DEVICE(md)->id ? DEVICE(md)->id : "",
|
|
addr);
|
|
}
|
|
out:
|
|
error_propagate(errp, local_err);
|
|
}
|
|
|
|
void memory_device_plug(MemoryDeviceState *md, MachineState *ms)
|
|
{
|
|
const MemoryDeviceClass *mdc = MEMORY_DEVICE_GET_CLASS(md);
|
|
const unsigned int memslots = memory_device_get_memslots(md);
|
|
const uint64_t addr = mdc->get_addr(md);
|
|
MemoryRegion *mr;
|
|
|
|
/*
|
|
* We expect that a previous call to memory_device_pre_plug() succeeded, so
|
|
* it can't fail at this point.
|
|
*/
|
|
mr = mdc->get_memory_region(md, &error_abort);
|
|
g_assert(ms->device_memory);
|
|
|
|
ms->device_memory->used_region_size += memory_region_size(mr);
|
|
ms->device_memory->required_memslots += memslots;
|
|
if (mdc->decide_memslots && memslots > 1) {
|
|
ms->device_memory->memslot_auto_decision_active++;
|
|
}
|
|
|
|
memory_region_add_subregion(&ms->device_memory->mr,
|
|
addr - ms->device_memory->base, mr);
|
|
trace_memory_device_plug(DEVICE(md)->id ? DEVICE(md)->id : "", addr);
|
|
}
|
|
|
|
void memory_device_unplug(MemoryDeviceState *md, MachineState *ms)
|
|
{
|
|
const MemoryDeviceClass *mdc = MEMORY_DEVICE_GET_CLASS(md);
|
|
const unsigned int memslots = memory_device_get_memslots(md);
|
|
MemoryRegion *mr;
|
|
|
|
/*
|
|
* We expect that a previous call to memory_device_pre_plug() succeeded, so
|
|
* it can't fail at this point.
|
|
*/
|
|
mr = mdc->get_memory_region(md, &error_abort);
|
|
g_assert(ms->device_memory);
|
|
|
|
memory_region_del_subregion(&ms->device_memory->mr, mr);
|
|
|
|
if (mdc->decide_memslots && memslots > 1) {
|
|
ms->device_memory->memslot_auto_decision_active--;
|
|
}
|
|
ms->device_memory->used_region_size -= memory_region_size(mr);
|
|
ms->device_memory->required_memslots -= memslots;
|
|
trace_memory_device_unplug(DEVICE(md)->id ? DEVICE(md)->id : "",
|
|
mdc->get_addr(md));
|
|
}
|
|
|
|
uint64_t memory_device_get_region_size(const MemoryDeviceState *md,
|
|
Error **errp)
|
|
{
|
|
const MemoryDeviceClass *mdc = MEMORY_DEVICE_GET_CLASS(md);
|
|
MemoryRegion *mr;
|
|
|
|
/* dropping const here is fine as we don't touch the memory region */
|
|
mr = mdc->get_memory_region((MemoryDeviceState *)md, errp);
|
|
if (!mr) {
|
|
return 0;
|
|
}
|
|
|
|
return memory_region_size(mr);
|
|
}
|
|
|
|
static void memory_devices_region_mod(MemoryListener *listener,
|
|
MemoryRegionSection *mrs, bool add)
|
|
{
|
|
DeviceMemoryState *dms = container_of(listener, DeviceMemoryState,
|
|
listener);
|
|
|
|
if (!memory_region_is_ram(mrs->mr)) {
|
|
warn_report("Unexpected memory region mapped into device memory region.");
|
|
return;
|
|
}
|
|
|
|
/*
|
|
* The expectation is that each distinct RAM memory region section in
|
|
* our region for memory devices consumes exactly one memslot in KVM
|
|
* and in vhost. For vhost, this is true, except:
|
|
* * ROM memory regions don't consume a memslot. These get used very
|
|
* rarely for memory devices (R/O NVDIMMs).
|
|
* * Memslots without a fd (memory-backend-ram) don't necessarily
|
|
* consume a memslot. Such setups are quite rare and possibly bogus:
|
|
* the memory would be inaccessible by such vhost devices.
|
|
*
|
|
* So for vhost, in corner cases we might over-estimate the number of
|
|
* memslots that are currently used or that might still be reserved
|
|
* (required - used).
|
|
*/
|
|
dms->used_memslots += add ? 1 : -1;
|
|
|
|
if (dms->used_memslots > dms->required_memslots) {
|
|
warn_report("Memory devices use more memory slots than indicated as required.");
|
|
}
|
|
}
|
|
|
|
static void memory_devices_region_add(MemoryListener *listener,
|
|
MemoryRegionSection *mrs)
|
|
{
|
|
return memory_devices_region_mod(listener, mrs, true);
|
|
}
|
|
|
|
static void memory_devices_region_del(MemoryListener *listener,
|
|
MemoryRegionSection *mrs)
|
|
{
|
|
return memory_devices_region_mod(listener, mrs, false);
|
|
}
|
|
|
|
void machine_memory_devices_init(MachineState *ms, hwaddr base, uint64_t size)
|
|
{
|
|
g_assert(size);
|
|
g_assert(!ms->device_memory);
|
|
ms->device_memory = g_new0(DeviceMemoryState, 1);
|
|
ms->device_memory->base = base;
|
|
|
|
memory_region_init(&ms->device_memory->mr, OBJECT(ms), "device-memory",
|
|
size);
|
|
address_space_init(&ms->device_memory->as, &ms->device_memory->mr,
|
|
"device-memory");
|
|
memory_region_add_subregion(get_system_memory(), ms->device_memory->base,
|
|
&ms->device_memory->mr);
|
|
|
|
/* Track the number of memslots used by memory devices. */
|
|
ms->device_memory->listener.region_add = memory_devices_region_add;
|
|
ms->device_memory->listener.region_del = memory_devices_region_del;
|
|
memory_listener_register(&ms->device_memory->listener,
|
|
&ms->device_memory->as);
|
|
}
|
|
|
|
static const TypeInfo memory_device_info = {
|
|
.name = TYPE_MEMORY_DEVICE,
|
|
.parent = TYPE_INTERFACE,
|
|
.class_size = sizeof(MemoryDeviceClass),
|
|
};
|
|
|
|
static void memory_device_register_types(void)
|
|
{
|
|
type_register_static(&memory_device_info);
|
|
}
|
|
|
|
type_init(memory_device_register_types)
|