ddf1ec2ffe
it replaces a static complied in DSDT MMIO region for CPU hotplug with one created at runtime leaving only truly static CPU hotplug related ASL bits in DSDT. It also puts CPU_HOTPLUG_RESOURCE_DEVICE into PCI0 scope and reserves resources from it, preparing for dropping manual hole punching in PCI0._CRS. Later it also would make easier to reuse current ACPI CPU hotplug on other targets. Also later it would be possible to move remaining CPU hotplug ASL methods into build_ssdt() and add all CPU hotplug related AML into SSDT only when CPU hotplug is enabled, further reducing ACPI tables blob if CPU hotplug isn't used. impl. detail: Windows XP can't handle /BSODs/ OperationRegion declaration in DSDT when variable from SSDT is used for specifying its address/length and also when Field declared in DSDT with OperationRegion from SSDT if DSDT is being parsed before SSDT. But it works just fine when referencing named fields from another table. Hence OperationRegion and Field declaration are moved to SSDT to make XP based editions work. PS: Later Windows editions seem to be fine with above conditions. Signed-off-by: Igor Mammedov <imammedo@redhat.com> Reviewed-by: Michael S. Tsirkin <mst@redhat.com> Signed-off-by: Michael S. Tsirkin <mst@redhat.com>
1757 lines
60 KiB
C
1757 lines
60 KiB
C
/* Support for generating ACPI tables and passing them to Guests
|
|
*
|
|
* Copyright (C) 2008-2010 Kevin O'Connor <kevin@koconnor.net>
|
|
* Copyright (C) 2006 Fabrice Bellard
|
|
* Copyright (C) 2013 Red Hat Inc
|
|
*
|
|
* Author: Michael S. Tsirkin <mst@redhat.com>
|
|
*
|
|
* This program is free software; you can redistribute it and/or modify
|
|
* it under the terms of the GNU General Public License as published by
|
|
* the Free Software Foundation; either version 2 of the License, or
|
|
* (at your option) any later version.
|
|
|
|
* This program is distributed in the hope that it will be useful,
|
|
* but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
* GNU General Public License for more details.
|
|
|
|
* You should have received a copy of the GNU General Public License along
|
|
* with this program; if not, see <http://www.gnu.org/licenses/>.
|
|
*/
|
|
|
|
#include "acpi-build.h"
|
|
#include <stddef.h>
|
|
#include <glib.h>
|
|
#include "qemu-common.h"
|
|
#include "qemu/bitmap.h"
|
|
#include "qemu/osdep.h"
|
|
#include "qemu/range.h"
|
|
#include "qemu/error-report.h"
|
|
#include "hw/pci/pci.h"
|
|
#include "qom/cpu.h"
|
|
#include "hw/i386/pc.h"
|
|
#include "target-i386/cpu.h"
|
|
#include "hw/timer/hpet.h"
|
|
#include "hw/i386/acpi-defs.h"
|
|
#include "hw/acpi/acpi.h"
|
|
#include "hw/nvram/fw_cfg.h"
|
|
#include "hw/acpi/bios-linker-loader.h"
|
|
#include "hw/loader.h"
|
|
#include "hw/isa/isa.h"
|
|
#include "hw/acpi/memory_hotplug.h"
|
|
#include "sysemu/tpm.h"
|
|
#include "hw/acpi/tpm.h"
|
|
|
|
/* Supported chipsets: */
|
|
#include "hw/acpi/piix4.h"
|
|
#include "hw/acpi/pcihp.h"
|
|
#include "hw/i386/ich9.h"
|
|
#include "hw/pci/pci_bus.h"
|
|
#include "hw/pci-host/q35.h"
|
|
#include "hw/i386/intel_iommu.h"
|
|
|
|
#include "hw/i386/q35-acpi-dsdt.hex"
|
|
#include "hw/i386/acpi-dsdt.hex"
|
|
|
|
#include "hw/acpi/aml-build.h"
|
|
|
|
#include "qapi/qmp/qint.h"
|
|
#include "qom/qom-qobject.h"
|
|
#include "exec/ram_addr.h"
|
|
|
|
/* These are used to size the ACPI tables for -M pc-i440fx-1.7 and
|
|
* -M pc-i440fx-2.0. Even if the actual amount of AML generated grows
|
|
* a little bit, there should be plenty of free space since the DSDT
|
|
* shrunk by ~1.5k between QEMU 2.0 and QEMU 2.1.
|
|
*/
|
|
#define ACPI_BUILD_LEGACY_CPU_AML_SIZE 97
|
|
#define ACPI_BUILD_ALIGN_SIZE 0x1000
|
|
|
|
#define ACPI_BUILD_TABLE_SIZE 0x20000
|
|
|
|
/* Reserve RAM space for tables: add another order of magnitude. */
|
|
#define ACPI_BUILD_TABLE_MAX_SIZE 0x200000
|
|
|
|
/* #define DEBUG_ACPI_BUILD */
|
|
#ifdef DEBUG_ACPI_BUILD
|
|
#define ACPI_BUILD_DPRINTF(fmt, ...) \
|
|
do {printf("ACPI_BUILD: " fmt, ## __VA_ARGS__); } while (0)
|
|
#else
|
|
#define ACPI_BUILD_DPRINTF(fmt, ...)
|
|
#endif
|
|
|
|
typedef struct AcpiCpuInfo {
|
|
DECLARE_BITMAP(found_cpus, ACPI_CPU_HOTPLUG_ID_LIMIT);
|
|
} AcpiCpuInfo;
|
|
|
|
typedef struct AcpiMcfgInfo {
|
|
uint64_t mcfg_base;
|
|
uint32_t mcfg_size;
|
|
} AcpiMcfgInfo;
|
|
|
|
typedef struct AcpiPmInfo {
|
|
bool s3_disabled;
|
|
bool s4_disabled;
|
|
bool pcihp_bridge_en;
|
|
uint8_t s4_val;
|
|
uint16_t sci_int;
|
|
uint8_t acpi_enable_cmd;
|
|
uint8_t acpi_disable_cmd;
|
|
uint32_t gpe0_blk;
|
|
uint32_t gpe0_blk_len;
|
|
uint32_t io_base;
|
|
uint16_t cpu_hp_io_base;
|
|
uint16_t cpu_hp_io_len;
|
|
} AcpiPmInfo;
|
|
|
|
typedef struct AcpiMiscInfo {
|
|
bool has_hpet;
|
|
bool has_tpm;
|
|
DECLARE_BITMAP(slot_hotplug_enable, PCI_SLOT_MAX);
|
|
const unsigned char *dsdt_code;
|
|
unsigned dsdt_size;
|
|
uint16_t pvpanic_port;
|
|
} AcpiMiscInfo;
|
|
|
|
typedef struct AcpiBuildPciBusHotplugState {
|
|
GArray *device_table;
|
|
GArray *notify_table;
|
|
struct AcpiBuildPciBusHotplugState *parent;
|
|
bool pcihp_bridge_en;
|
|
} AcpiBuildPciBusHotplugState;
|
|
|
|
static void acpi_get_dsdt(AcpiMiscInfo *info)
|
|
{
|
|
uint16_t *applesmc_sta;
|
|
Object *piix = piix4_pm_find();
|
|
Object *lpc = ich9_lpc_find();
|
|
assert(!!piix != !!lpc);
|
|
|
|
if (piix) {
|
|
info->dsdt_code = AcpiDsdtAmlCode;
|
|
info->dsdt_size = sizeof AcpiDsdtAmlCode;
|
|
applesmc_sta = piix_dsdt_applesmc_sta;
|
|
}
|
|
if (lpc) {
|
|
info->dsdt_code = Q35AcpiDsdtAmlCode;
|
|
info->dsdt_size = sizeof Q35AcpiDsdtAmlCode;
|
|
applesmc_sta = q35_dsdt_applesmc_sta;
|
|
}
|
|
|
|
/* Patch in appropriate value for AppleSMC _STA */
|
|
*(uint8_t *)(info->dsdt_code + *applesmc_sta) =
|
|
applesmc_find() ? 0x0b : 0x00;
|
|
}
|
|
|
|
static
|
|
int acpi_add_cpu_info(Object *o, void *opaque)
|
|
{
|
|
AcpiCpuInfo *cpu = opaque;
|
|
uint64_t apic_id;
|
|
|
|
if (object_dynamic_cast(o, TYPE_CPU)) {
|
|
apic_id = object_property_get_int(o, "apic-id", NULL);
|
|
assert(apic_id < ACPI_CPU_HOTPLUG_ID_LIMIT);
|
|
|
|
set_bit(apic_id, cpu->found_cpus);
|
|
}
|
|
|
|
object_child_foreach(o, acpi_add_cpu_info, opaque);
|
|
return 0;
|
|
}
|
|
|
|
static void acpi_get_cpu_info(AcpiCpuInfo *cpu)
|
|
{
|
|
Object *root = object_get_root();
|
|
|
|
memset(cpu->found_cpus, 0, sizeof cpu->found_cpus);
|
|
object_child_foreach(root, acpi_add_cpu_info, cpu);
|
|
}
|
|
|
|
static void acpi_get_pm_info(AcpiPmInfo *pm)
|
|
{
|
|
Object *piix = piix4_pm_find();
|
|
Object *lpc = ich9_lpc_find();
|
|
Object *obj = NULL;
|
|
QObject *o;
|
|
|
|
if (piix) {
|
|
obj = piix;
|
|
pm->cpu_hp_io_base = PIIX4_CPU_HOTPLUG_IO_BASE;
|
|
}
|
|
if (lpc) {
|
|
obj = lpc;
|
|
pm->cpu_hp_io_base = ICH9_CPU_HOTPLUG_IO_BASE;
|
|
}
|
|
assert(obj);
|
|
|
|
pm->cpu_hp_io_len = ACPI_GPE_PROC_LEN;
|
|
/* Fill in optional s3/s4 related properties */
|
|
o = object_property_get_qobject(obj, ACPI_PM_PROP_S3_DISABLED, NULL);
|
|
if (o) {
|
|
pm->s3_disabled = qint_get_int(qobject_to_qint(o));
|
|
} else {
|
|
pm->s3_disabled = false;
|
|
}
|
|
qobject_decref(o);
|
|
o = object_property_get_qobject(obj, ACPI_PM_PROP_S4_DISABLED, NULL);
|
|
if (o) {
|
|
pm->s4_disabled = qint_get_int(qobject_to_qint(o));
|
|
} else {
|
|
pm->s4_disabled = false;
|
|
}
|
|
qobject_decref(o);
|
|
o = object_property_get_qobject(obj, ACPI_PM_PROP_S4_VAL, NULL);
|
|
if (o) {
|
|
pm->s4_val = qint_get_int(qobject_to_qint(o));
|
|
} else {
|
|
pm->s4_val = false;
|
|
}
|
|
qobject_decref(o);
|
|
|
|
/* Fill in mandatory properties */
|
|
pm->sci_int = object_property_get_int(obj, ACPI_PM_PROP_SCI_INT, NULL);
|
|
|
|
pm->acpi_enable_cmd = object_property_get_int(obj,
|
|
ACPI_PM_PROP_ACPI_ENABLE_CMD,
|
|
NULL);
|
|
pm->acpi_disable_cmd = object_property_get_int(obj,
|
|
ACPI_PM_PROP_ACPI_DISABLE_CMD,
|
|
NULL);
|
|
pm->io_base = object_property_get_int(obj, ACPI_PM_PROP_PM_IO_BASE,
|
|
NULL);
|
|
pm->gpe0_blk = object_property_get_int(obj, ACPI_PM_PROP_GPE0_BLK,
|
|
NULL);
|
|
pm->gpe0_blk_len = object_property_get_int(obj, ACPI_PM_PROP_GPE0_BLK_LEN,
|
|
NULL);
|
|
pm->pcihp_bridge_en =
|
|
object_property_get_bool(obj, "acpi-pci-hotplug-with-bridge-support",
|
|
NULL);
|
|
}
|
|
|
|
static void acpi_get_misc_info(AcpiMiscInfo *info)
|
|
{
|
|
info->has_hpet = hpet_find();
|
|
info->has_tpm = tpm_find();
|
|
info->pvpanic_port = pvpanic_port();
|
|
}
|
|
|
|
static void acpi_get_pci_info(PcPciInfo *info)
|
|
{
|
|
Object *pci_host;
|
|
bool ambiguous;
|
|
|
|
pci_host = object_resolve_path_type("", TYPE_PCI_HOST_BRIDGE, &ambiguous);
|
|
g_assert(!ambiguous);
|
|
g_assert(pci_host);
|
|
|
|
info->w32.begin = object_property_get_int(pci_host,
|
|
PCI_HOST_PROP_PCI_HOLE_START,
|
|
NULL);
|
|
info->w32.end = object_property_get_int(pci_host,
|
|
PCI_HOST_PROP_PCI_HOLE_END,
|
|
NULL);
|
|
info->w64.begin = object_property_get_int(pci_host,
|
|
PCI_HOST_PROP_PCI_HOLE64_START,
|
|
NULL);
|
|
info->w64.end = object_property_get_int(pci_host,
|
|
PCI_HOST_PROP_PCI_HOLE64_END,
|
|
NULL);
|
|
}
|
|
|
|
#define ACPI_BUILD_APPNAME "Bochs"
|
|
#define ACPI_BUILD_APPNAME6 "BOCHS "
|
|
#define ACPI_BUILD_APPNAME4 "BXPC"
|
|
|
|
#define ACPI_BUILD_TABLE_FILE "etc/acpi/tables"
|
|
#define ACPI_BUILD_RSDP_FILE "etc/acpi/rsdp"
|
|
#define ACPI_BUILD_TPMLOG_FILE "etc/tpm/log"
|
|
|
|
static void
|
|
build_header(GArray *linker, GArray *table_data,
|
|
AcpiTableHeader *h, const char *sig, int len, uint8_t rev)
|
|
{
|
|
memcpy(&h->signature, sig, 4);
|
|
h->length = cpu_to_le32(len);
|
|
h->revision = rev;
|
|
memcpy(h->oem_id, ACPI_BUILD_APPNAME6, 6);
|
|
memcpy(h->oem_table_id, ACPI_BUILD_APPNAME4, 4);
|
|
memcpy(h->oem_table_id + 4, sig, 4);
|
|
h->oem_revision = cpu_to_le32(1);
|
|
memcpy(h->asl_compiler_id, ACPI_BUILD_APPNAME4, 4);
|
|
h->asl_compiler_revision = cpu_to_le32(1);
|
|
h->checksum = 0;
|
|
/* Checksum to be filled in by Guest linker */
|
|
bios_linker_loader_add_checksum(linker, ACPI_BUILD_TABLE_FILE,
|
|
table_data->data, h, len, &h->checksum);
|
|
}
|
|
|
|
static GArray *build_alloc_method(const char *name, uint8_t arg_count)
|
|
{
|
|
GArray *method = build_alloc_array();
|
|
|
|
build_append_namestring(method, "%s", name);
|
|
build_append_byte(method, arg_count); /* MethodFlags: ArgCount */
|
|
|
|
return method;
|
|
}
|
|
|
|
static void build_append_and_cleanup_method(GArray *device, GArray *method)
|
|
{
|
|
uint8_t op = 0x14; /* MethodOp */
|
|
|
|
build_package(method, op);
|
|
|
|
build_append_array(device, method);
|
|
build_free_array(method);
|
|
}
|
|
|
|
static void build_append_notify_target_ifequal(GArray *method,
|
|
GArray *target_name,
|
|
uint32_t value)
|
|
{
|
|
GArray *notify = build_alloc_array();
|
|
uint8_t op = 0xA0; /* IfOp */
|
|
|
|
build_append_byte(notify, 0x93); /* LEqualOp */
|
|
build_append_byte(notify, 0x68); /* Arg0Op */
|
|
build_append_int(notify, value);
|
|
build_append_byte(notify, 0x86); /* NotifyOp */
|
|
build_append_array(notify, target_name);
|
|
build_append_byte(notify, 0x69); /* Arg1Op */
|
|
|
|
/* Pack it up */
|
|
build_package(notify, op);
|
|
|
|
build_append_array(method, notify);
|
|
|
|
build_free_array(notify);
|
|
}
|
|
|
|
/* End here */
|
|
#define ACPI_PORT_SMI_CMD 0x00b2 /* TODO: this is APM_CNT_IOPORT */
|
|
|
|
static inline void *acpi_data_push(GArray *table_data, unsigned size)
|
|
{
|
|
unsigned off = table_data->len;
|
|
g_array_set_size(table_data, off + size);
|
|
return table_data->data + off;
|
|
}
|
|
|
|
static unsigned acpi_data_len(GArray *table)
|
|
{
|
|
#if GLIB_CHECK_VERSION(2, 22, 0)
|
|
assert(g_array_get_element_size(table) == 1);
|
|
#endif
|
|
return table->len;
|
|
}
|
|
|
|
static void acpi_align_size(GArray *blob, unsigned align)
|
|
{
|
|
/* Align size to multiple of given size. This reduces the chance
|
|
* we need to change size in the future (breaking cross version migration).
|
|
*/
|
|
g_array_set_size(blob, ROUND_UP(acpi_data_len(blob), align));
|
|
}
|
|
|
|
/* Set a value within table in a safe manner */
|
|
#define ACPI_BUILD_SET_LE(table, size, off, bits, val) \
|
|
do { \
|
|
uint64_t ACPI_BUILD_SET_LE_val = cpu_to_le64(val); \
|
|
memcpy(acpi_data_get_ptr(table, size, off, \
|
|
(bits) / BITS_PER_BYTE), \
|
|
&ACPI_BUILD_SET_LE_val, \
|
|
(bits) / BITS_PER_BYTE); \
|
|
} while (0)
|
|
|
|
static inline void *acpi_data_get_ptr(uint8_t *table_data, unsigned table_size,
|
|
unsigned off, unsigned size)
|
|
{
|
|
assert(off + size > off);
|
|
assert(off + size <= table_size);
|
|
return table_data + off;
|
|
}
|
|
|
|
static inline void acpi_add_table(GArray *table_offsets, GArray *table_data)
|
|
{
|
|
uint32_t offset = cpu_to_le32(table_data->len);
|
|
g_array_append_val(table_offsets, offset);
|
|
}
|
|
|
|
/* FACS */
|
|
static void
|
|
build_facs(GArray *table_data, GArray *linker, PcGuestInfo *guest_info)
|
|
{
|
|
AcpiFacsDescriptorRev1 *facs = acpi_data_push(table_data, sizeof *facs);
|
|
memcpy(&facs->signature, "FACS", 4);
|
|
facs->length = cpu_to_le32(sizeof(*facs));
|
|
}
|
|
|
|
/* Load chipset information in FADT */
|
|
static void fadt_setup(AcpiFadtDescriptorRev1 *fadt, AcpiPmInfo *pm)
|
|
{
|
|
fadt->model = 1;
|
|
fadt->reserved1 = 0;
|
|
fadt->sci_int = cpu_to_le16(pm->sci_int);
|
|
fadt->smi_cmd = cpu_to_le32(ACPI_PORT_SMI_CMD);
|
|
fadt->acpi_enable = pm->acpi_enable_cmd;
|
|
fadt->acpi_disable = pm->acpi_disable_cmd;
|
|
/* EVT, CNT, TMR offset matches hw/acpi/core.c */
|
|
fadt->pm1a_evt_blk = cpu_to_le32(pm->io_base);
|
|
fadt->pm1a_cnt_blk = cpu_to_le32(pm->io_base + 0x04);
|
|
fadt->pm_tmr_blk = cpu_to_le32(pm->io_base + 0x08);
|
|
fadt->gpe0_blk = cpu_to_le32(pm->gpe0_blk);
|
|
/* EVT, CNT, TMR length matches hw/acpi/core.c */
|
|
fadt->pm1_evt_len = 4;
|
|
fadt->pm1_cnt_len = 2;
|
|
fadt->pm_tmr_len = 4;
|
|
fadt->gpe0_blk_len = pm->gpe0_blk_len;
|
|
fadt->plvl2_lat = cpu_to_le16(0xfff); /* C2 state not supported */
|
|
fadt->plvl3_lat = cpu_to_le16(0xfff); /* C3 state not supported */
|
|
fadt->flags = cpu_to_le32((1 << ACPI_FADT_F_WBINVD) |
|
|
(1 << ACPI_FADT_F_PROC_C1) |
|
|
(1 << ACPI_FADT_F_SLP_BUTTON) |
|
|
(1 << ACPI_FADT_F_RTC_S4));
|
|
fadt->flags |= cpu_to_le32(1 << ACPI_FADT_F_USE_PLATFORM_CLOCK);
|
|
/* APIC destination mode ("Flat Logical") has an upper limit of 8 CPUs
|
|
* For more than 8 CPUs, "Clustered Logical" mode has to be used
|
|
*/
|
|
if (max_cpus > 8) {
|
|
fadt->flags |= cpu_to_le32(1 << ACPI_FADT_F_FORCE_APIC_CLUSTER_MODEL);
|
|
}
|
|
}
|
|
|
|
|
|
/* FADT */
|
|
static void
|
|
build_fadt(GArray *table_data, GArray *linker, AcpiPmInfo *pm,
|
|
unsigned facs, unsigned dsdt)
|
|
{
|
|
AcpiFadtDescriptorRev1 *fadt = acpi_data_push(table_data, sizeof(*fadt));
|
|
|
|
fadt->firmware_ctrl = cpu_to_le32(facs);
|
|
/* FACS address to be filled by Guest linker */
|
|
bios_linker_loader_add_pointer(linker, ACPI_BUILD_TABLE_FILE,
|
|
ACPI_BUILD_TABLE_FILE,
|
|
table_data, &fadt->firmware_ctrl,
|
|
sizeof fadt->firmware_ctrl);
|
|
|
|
fadt->dsdt = cpu_to_le32(dsdt);
|
|
/* DSDT address to be filled by Guest linker */
|
|
bios_linker_loader_add_pointer(linker, ACPI_BUILD_TABLE_FILE,
|
|
ACPI_BUILD_TABLE_FILE,
|
|
table_data, &fadt->dsdt,
|
|
sizeof fadt->dsdt);
|
|
|
|
fadt_setup(fadt, pm);
|
|
|
|
build_header(linker, table_data,
|
|
(void *)fadt, "FACP", sizeof(*fadt), 1);
|
|
}
|
|
|
|
static void
|
|
build_madt(GArray *table_data, GArray *linker, AcpiCpuInfo *cpu,
|
|
PcGuestInfo *guest_info)
|
|
{
|
|
int madt_start = table_data->len;
|
|
|
|
AcpiMultipleApicTable *madt;
|
|
AcpiMadtIoApic *io_apic;
|
|
AcpiMadtIntsrcovr *intsrcovr;
|
|
AcpiMadtLocalNmi *local_nmi;
|
|
int i;
|
|
|
|
madt = acpi_data_push(table_data, sizeof *madt);
|
|
madt->local_apic_address = cpu_to_le32(APIC_DEFAULT_ADDRESS);
|
|
madt->flags = cpu_to_le32(1);
|
|
|
|
for (i = 0; i < guest_info->apic_id_limit; i++) {
|
|
AcpiMadtProcessorApic *apic = acpi_data_push(table_data, sizeof *apic);
|
|
apic->type = ACPI_APIC_PROCESSOR;
|
|
apic->length = sizeof(*apic);
|
|
apic->processor_id = i;
|
|
apic->local_apic_id = i;
|
|
if (test_bit(i, cpu->found_cpus)) {
|
|
apic->flags = cpu_to_le32(1);
|
|
} else {
|
|
apic->flags = cpu_to_le32(0);
|
|
}
|
|
}
|
|
io_apic = acpi_data_push(table_data, sizeof *io_apic);
|
|
io_apic->type = ACPI_APIC_IO;
|
|
io_apic->length = sizeof(*io_apic);
|
|
#define ACPI_BUILD_IOAPIC_ID 0x0
|
|
io_apic->io_apic_id = ACPI_BUILD_IOAPIC_ID;
|
|
io_apic->address = cpu_to_le32(IO_APIC_DEFAULT_ADDRESS);
|
|
io_apic->interrupt = cpu_to_le32(0);
|
|
|
|
if (guest_info->apic_xrupt_override) {
|
|
intsrcovr = acpi_data_push(table_data, sizeof *intsrcovr);
|
|
intsrcovr->type = ACPI_APIC_XRUPT_OVERRIDE;
|
|
intsrcovr->length = sizeof(*intsrcovr);
|
|
intsrcovr->source = 0;
|
|
intsrcovr->gsi = cpu_to_le32(2);
|
|
intsrcovr->flags = cpu_to_le16(0); /* conforms to bus specifications */
|
|
}
|
|
for (i = 1; i < 16; i++) {
|
|
#define ACPI_BUILD_PCI_IRQS ((1<<5) | (1<<9) | (1<<10) | (1<<11))
|
|
if (!(ACPI_BUILD_PCI_IRQS & (1 << i))) {
|
|
/* No need for a INT source override structure. */
|
|
continue;
|
|
}
|
|
intsrcovr = acpi_data_push(table_data, sizeof *intsrcovr);
|
|
intsrcovr->type = ACPI_APIC_XRUPT_OVERRIDE;
|
|
intsrcovr->length = sizeof(*intsrcovr);
|
|
intsrcovr->source = i;
|
|
intsrcovr->gsi = cpu_to_le32(i);
|
|
intsrcovr->flags = cpu_to_le16(0xd); /* active high, level triggered */
|
|
}
|
|
|
|
local_nmi = acpi_data_push(table_data, sizeof *local_nmi);
|
|
local_nmi->type = ACPI_APIC_LOCAL_NMI;
|
|
local_nmi->length = sizeof(*local_nmi);
|
|
local_nmi->processor_id = 0xff; /* all processors */
|
|
local_nmi->flags = cpu_to_le16(0);
|
|
local_nmi->lint = 1; /* ACPI_LINT1 */
|
|
|
|
build_header(linker, table_data,
|
|
(void *)(table_data->data + madt_start), "APIC",
|
|
table_data->len - madt_start, 1);
|
|
}
|
|
|
|
/* Encode a hex value */
|
|
static inline char acpi_get_hex(uint32_t val)
|
|
{
|
|
val &= 0x0f;
|
|
return (val <= 9) ? ('0' + val) : ('A' + val - 10);
|
|
}
|
|
|
|
/* 0x5B 0x82 DeviceOp PkgLength NameString */
|
|
#define ACPI_PCIHP_OFFSET_HEX (*ssdt_pcihp_name - *ssdt_pcihp_start + 1)
|
|
#define ACPI_PCIHP_OFFSET_ID (*ssdt_pcihp_id - *ssdt_pcihp_start)
|
|
#define ACPI_PCIHP_OFFSET_ADR (*ssdt_pcihp_adr - *ssdt_pcihp_start)
|
|
#define ACPI_PCIHP_OFFSET_EJ0 (*ssdt_pcihp_ej0 - *ssdt_pcihp_start)
|
|
#define ACPI_PCIHP_SIZEOF (*ssdt_pcihp_end - *ssdt_pcihp_start)
|
|
#define ACPI_PCIHP_AML (ssdp_pcihp_aml + *ssdt_pcihp_start)
|
|
|
|
#define ACPI_PCINOHP_OFFSET_HEX (*ssdt_pcinohp_name - *ssdt_pcinohp_start + 1)
|
|
#define ACPI_PCINOHP_OFFSET_ADR (*ssdt_pcinohp_adr - *ssdt_pcinohp_start)
|
|
#define ACPI_PCINOHP_SIZEOF (*ssdt_pcinohp_end - *ssdt_pcinohp_start)
|
|
#define ACPI_PCINOHP_AML (ssdp_pcihp_aml + *ssdt_pcinohp_start)
|
|
|
|
#define ACPI_PCIVGA_OFFSET_HEX (*ssdt_pcivga_name - *ssdt_pcivga_start + 1)
|
|
#define ACPI_PCIVGA_OFFSET_ADR (*ssdt_pcivga_adr - *ssdt_pcivga_start)
|
|
#define ACPI_PCIVGA_SIZEOF (*ssdt_pcivga_end - *ssdt_pcivga_start)
|
|
#define ACPI_PCIVGA_AML (ssdp_pcihp_aml + *ssdt_pcivga_start)
|
|
|
|
#define ACPI_PCIQXL_OFFSET_HEX (*ssdt_pciqxl_name - *ssdt_pciqxl_start + 1)
|
|
#define ACPI_PCIQXL_OFFSET_ADR (*ssdt_pciqxl_adr - *ssdt_pciqxl_start)
|
|
#define ACPI_PCIQXL_SIZEOF (*ssdt_pciqxl_end - *ssdt_pciqxl_start)
|
|
#define ACPI_PCIQXL_AML (ssdp_pcihp_aml + *ssdt_pciqxl_start)
|
|
|
|
#include "hw/i386/ssdt-mem.hex"
|
|
|
|
/* 0x5B 0x82 DeviceOp PkgLength NameString DimmID */
|
|
#define ACPI_MEM_OFFSET_HEX (*ssdt_mem_name - *ssdt_mem_start + 2)
|
|
#define ACPI_MEM_OFFSET_ID (*ssdt_mem_id - *ssdt_mem_start + 7)
|
|
#define ACPI_MEM_SIZEOF (*ssdt_mem_end - *ssdt_mem_start)
|
|
#define ACPI_MEM_AML (ssdm_mem_aml + *ssdt_mem_start)
|
|
|
|
#define ACPI_SSDT_SIGNATURE 0x54445353 /* SSDT */
|
|
#define ACPI_SSDT_HEADER_LENGTH 36
|
|
|
|
#include "hw/i386/ssdt-misc.hex"
|
|
#include "hw/i386/ssdt-pcihp.hex"
|
|
#include "hw/i386/ssdt-tpm.hex"
|
|
|
|
static void
|
|
build_append_notify_method(GArray *device, const char *name,
|
|
const char *format, int count)
|
|
{
|
|
int i;
|
|
GArray *method = build_alloc_method(name, 2);
|
|
|
|
for (i = 0; i < count; i++) {
|
|
GArray *target = build_alloc_array();
|
|
build_append_namestring(target, format, i);
|
|
assert(i < 256); /* Fits in 1 byte */
|
|
build_append_notify_target_ifequal(method, target, i);
|
|
build_free_array(target);
|
|
}
|
|
|
|
build_append_and_cleanup_method(device, method);
|
|
}
|
|
|
|
static void patch_pcihp(int slot, uint8_t *ssdt_ptr)
|
|
{
|
|
unsigned devfn = PCI_DEVFN(slot, 0);
|
|
|
|
ssdt_ptr[ACPI_PCIHP_OFFSET_HEX] = acpi_get_hex(devfn >> 4);
|
|
ssdt_ptr[ACPI_PCIHP_OFFSET_HEX + 1] = acpi_get_hex(devfn);
|
|
ssdt_ptr[ACPI_PCIHP_OFFSET_ID] = slot;
|
|
ssdt_ptr[ACPI_PCIHP_OFFSET_ADR + 2] = slot;
|
|
}
|
|
|
|
static void patch_pcinohp(int slot, uint8_t *ssdt_ptr)
|
|
{
|
|
unsigned devfn = PCI_DEVFN(slot, 0);
|
|
|
|
ssdt_ptr[ACPI_PCINOHP_OFFSET_HEX] = acpi_get_hex(devfn >> 4);
|
|
ssdt_ptr[ACPI_PCINOHP_OFFSET_HEX + 1] = acpi_get_hex(devfn);
|
|
ssdt_ptr[ACPI_PCINOHP_OFFSET_ADR + 2] = slot;
|
|
}
|
|
|
|
static void patch_pcivga(int slot, uint8_t *ssdt_ptr)
|
|
{
|
|
unsigned devfn = PCI_DEVFN(slot, 0);
|
|
|
|
ssdt_ptr[ACPI_PCIVGA_OFFSET_HEX] = acpi_get_hex(devfn >> 4);
|
|
ssdt_ptr[ACPI_PCIVGA_OFFSET_HEX + 1] = acpi_get_hex(devfn);
|
|
ssdt_ptr[ACPI_PCIVGA_OFFSET_ADR + 2] = slot;
|
|
}
|
|
|
|
static void patch_pciqxl(int slot, uint8_t *ssdt_ptr)
|
|
{
|
|
unsigned devfn = PCI_DEVFN(slot, 0);
|
|
|
|
ssdt_ptr[ACPI_PCIQXL_OFFSET_HEX] = acpi_get_hex(devfn >> 4);
|
|
ssdt_ptr[ACPI_PCIQXL_OFFSET_HEX + 1] = acpi_get_hex(devfn);
|
|
ssdt_ptr[ACPI_PCIQXL_OFFSET_ADR + 2] = slot;
|
|
}
|
|
|
|
/* Assign BSEL property to all buses. In the future, this can be changed
|
|
* to only assign to buses that support hotplug.
|
|
*/
|
|
static void *acpi_set_bsel(PCIBus *bus, void *opaque)
|
|
{
|
|
unsigned *bsel_alloc = opaque;
|
|
unsigned *bus_bsel;
|
|
|
|
if (qbus_is_hotpluggable(BUS(bus))) {
|
|
bus_bsel = g_malloc(sizeof *bus_bsel);
|
|
|
|
*bus_bsel = (*bsel_alloc)++;
|
|
object_property_add_uint32_ptr(OBJECT(bus), ACPI_PCIHP_PROP_BSEL,
|
|
bus_bsel, NULL);
|
|
}
|
|
|
|
return bsel_alloc;
|
|
}
|
|
|
|
static void acpi_set_pci_info(void)
|
|
{
|
|
PCIBus *bus = find_i440fx(); /* TODO: Q35 support */
|
|
unsigned bsel_alloc = 0;
|
|
|
|
if (bus) {
|
|
/* Scan all PCI buses. Set property to enable acpi based hotplug. */
|
|
pci_for_each_bus_depth_first(bus, acpi_set_bsel, NULL, &bsel_alloc);
|
|
}
|
|
}
|
|
|
|
static void build_pci_bus_state_init(AcpiBuildPciBusHotplugState *state,
|
|
AcpiBuildPciBusHotplugState *parent,
|
|
bool pcihp_bridge_en)
|
|
{
|
|
state->parent = parent;
|
|
state->device_table = build_alloc_array();
|
|
state->notify_table = build_alloc_array();
|
|
state->pcihp_bridge_en = pcihp_bridge_en;
|
|
}
|
|
|
|
static void build_pci_bus_state_cleanup(AcpiBuildPciBusHotplugState *state)
|
|
{
|
|
build_free_array(state->device_table);
|
|
build_free_array(state->notify_table);
|
|
}
|
|
|
|
static void *build_pci_bus_begin(PCIBus *bus, void *parent_state)
|
|
{
|
|
AcpiBuildPciBusHotplugState *parent = parent_state;
|
|
AcpiBuildPciBusHotplugState *child = g_malloc(sizeof *child);
|
|
|
|
build_pci_bus_state_init(child, parent, parent->pcihp_bridge_en);
|
|
|
|
return child;
|
|
}
|
|
|
|
static void build_pci_bus_end(PCIBus *bus, void *bus_state)
|
|
{
|
|
AcpiBuildPciBusHotplugState *child = bus_state;
|
|
AcpiBuildPciBusHotplugState *parent = child->parent;
|
|
GArray *bus_table = build_alloc_array();
|
|
DECLARE_BITMAP(slot_hotplug_enable, PCI_SLOT_MAX);
|
|
DECLARE_BITMAP(slot_device_present, PCI_SLOT_MAX);
|
|
DECLARE_BITMAP(slot_device_system, PCI_SLOT_MAX);
|
|
DECLARE_BITMAP(slot_device_vga, PCI_SLOT_MAX);
|
|
DECLARE_BITMAP(slot_device_qxl, PCI_SLOT_MAX);
|
|
uint8_t op;
|
|
int i;
|
|
QObject *bsel;
|
|
GArray *method;
|
|
bool bus_hotplug_support = false;
|
|
|
|
/*
|
|
* Skip bridge subtree creation if bridge hotplug is disabled
|
|
* to make acpi tables compatible with legacy machine types.
|
|
* Skip creation for hotplugged bridges as well.
|
|
*/
|
|
if (bus->parent_dev && (!child->pcihp_bridge_en ||
|
|
DEVICE(bus->parent_dev)->hotplugged)) {
|
|
build_free_array(bus_table);
|
|
build_pci_bus_state_cleanup(child);
|
|
g_free(child);
|
|
return;
|
|
}
|
|
|
|
if (bus->parent_dev) {
|
|
op = 0x82; /* DeviceOp */
|
|
build_append_namestring(bus_table, "S%.02X",
|
|
bus->parent_dev->devfn);
|
|
build_append_byte(bus_table, 0x08); /* NameOp */
|
|
build_append_namestring(bus_table, "_SUN");
|
|
build_append_int(bus_table, PCI_SLOT(bus->parent_dev->devfn));
|
|
build_append_byte(bus_table, 0x08); /* NameOp */
|
|
build_append_namestring(bus_table, "_ADR");
|
|
build_append_int(bus_table, (PCI_SLOT(bus->parent_dev->devfn) << 16) |
|
|
PCI_FUNC(bus->parent_dev->devfn));
|
|
} else {
|
|
op = 0x10; /* ScopeOp */;
|
|
build_append_namestring(bus_table, "PCI0");
|
|
}
|
|
|
|
bsel = object_property_get_qobject(OBJECT(bus), ACPI_PCIHP_PROP_BSEL, NULL);
|
|
if (bsel) {
|
|
build_append_byte(bus_table, 0x08); /* NameOp */
|
|
build_append_namestring(bus_table, "BSEL");
|
|
build_append_int(bus_table, qint_get_int(qobject_to_qint(bsel)));
|
|
memset(slot_hotplug_enable, 0xff, sizeof slot_hotplug_enable);
|
|
} else {
|
|
/* No bsel - no slots are hot-pluggable */
|
|
memset(slot_hotplug_enable, 0x00, sizeof slot_hotplug_enable);
|
|
}
|
|
|
|
memset(slot_device_present, 0x00, sizeof slot_device_present);
|
|
memset(slot_device_system, 0x00, sizeof slot_device_present);
|
|
memset(slot_device_vga, 0x00, sizeof slot_device_vga);
|
|
memset(slot_device_qxl, 0x00, sizeof slot_device_qxl);
|
|
|
|
for (i = 0; i < ARRAY_SIZE(bus->devices); i += PCI_FUNC_MAX) {
|
|
DeviceClass *dc;
|
|
PCIDeviceClass *pc;
|
|
PCIDevice *pdev = bus->devices[i];
|
|
int slot = PCI_SLOT(i);
|
|
bool bridge_in_acpi;
|
|
|
|
if (!pdev) {
|
|
continue;
|
|
}
|
|
|
|
set_bit(slot, slot_device_present);
|
|
pc = PCI_DEVICE_GET_CLASS(pdev);
|
|
dc = DEVICE_GET_CLASS(pdev);
|
|
|
|
/* When hotplug for bridges is enabled, bridges are
|
|
* described in ACPI separately (see build_pci_bus_end).
|
|
* In this case they aren't themselves hot-pluggable.
|
|
* Hotplugged bridges *are* hot-pluggable.
|
|
*/
|
|
bridge_in_acpi = pc->is_bridge && child->pcihp_bridge_en &&
|
|
!DEVICE(pdev)->hotplugged;
|
|
|
|
if (pc->class_id == PCI_CLASS_BRIDGE_ISA || bridge_in_acpi) {
|
|
set_bit(slot, slot_device_system);
|
|
}
|
|
|
|
if (pc->class_id == PCI_CLASS_DISPLAY_VGA) {
|
|
set_bit(slot, slot_device_vga);
|
|
|
|
if (object_dynamic_cast(OBJECT(pdev), "qxl-vga")) {
|
|
set_bit(slot, slot_device_qxl);
|
|
}
|
|
}
|
|
|
|
if (!dc->hotpluggable || bridge_in_acpi) {
|
|
clear_bit(slot, slot_hotplug_enable);
|
|
}
|
|
}
|
|
|
|
/* Append Device object for each slot */
|
|
for (i = 0; i < PCI_SLOT_MAX; i++) {
|
|
bool can_eject = test_bit(i, slot_hotplug_enable);
|
|
bool present = test_bit(i, slot_device_present);
|
|
bool vga = test_bit(i, slot_device_vga);
|
|
bool qxl = test_bit(i, slot_device_qxl);
|
|
bool system = test_bit(i, slot_device_system);
|
|
if (can_eject) {
|
|
void *pcihp = acpi_data_push(bus_table,
|
|
ACPI_PCIHP_SIZEOF);
|
|
memcpy(pcihp, ACPI_PCIHP_AML, ACPI_PCIHP_SIZEOF);
|
|
patch_pcihp(i, pcihp);
|
|
bus_hotplug_support = true;
|
|
} else if (qxl) {
|
|
void *pcihp = acpi_data_push(bus_table,
|
|
ACPI_PCIQXL_SIZEOF);
|
|
memcpy(pcihp, ACPI_PCIQXL_AML, ACPI_PCIQXL_SIZEOF);
|
|
patch_pciqxl(i, pcihp);
|
|
} else if (vga) {
|
|
void *pcihp = acpi_data_push(bus_table,
|
|
ACPI_PCIVGA_SIZEOF);
|
|
memcpy(pcihp, ACPI_PCIVGA_AML, ACPI_PCIVGA_SIZEOF);
|
|
patch_pcivga(i, pcihp);
|
|
} else if (system) {
|
|
/* Nothing to do: system devices are in DSDT or in SSDT above. */
|
|
} else if (present) {
|
|
void *pcihp = acpi_data_push(bus_table,
|
|
ACPI_PCINOHP_SIZEOF);
|
|
memcpy(pcihp, ACPI_PCINOHP_AML, ACPI_PCINOHP_SIZEOF);
|
|
patch_pcinohp(i, pcihp);
|
|
}
|
|
}
|
|
|
|
if (bsel) {
|
|
method = build_alloc_method("DVNT", 2);
|
|
|
|
for (i = 0; i < PCI_SLOT_MAX; i++) {
|
|
GArray *notify;
|
|
uint8_t op;
|
|
|
|
if (!test_bit(i, slot_hotplug_enable)) {
|
|
continue;
|
|
}
|
|
|
|
notify = build_alloc_array();
|
|
op = 0xA0; /* IfOp */
|
|
|
|
build_append_byte(notify, 0x7B); /* AndOp */
|
|
build_append_byte(notify, 0x68); /* Arg0Op */
|
|
build_append_int(notify, 0x1U << i);
|
|
build_append_byte(notify, 0x00); /* NullName */
|
|
build_append_byte(notify, 0x86); /* NotifyOp */
|
|
build_append_namestring(notify, "S%.02X", PCI_DEVFN(i, 0));
|
|
build_append_byte(notify, 0x69); /* Arg1Op */
|
|
|
|
/* Pack it up */
|
|
build_package(notify, op);
|
|
|
|
build_append_array(method, notify);
|
|
|
|
build_free_array(notify);
|
|
}
|
|
|
|
build_append_and_cleanup_method(bus_table, method);
|
|
}
|
|
|
|
/* Append PCNT method to notify about events on local and child buses.
|
|
* Add unconditionally for root since DSDT expects it.
|
|
*/
|
|
if (bus_hotplug_support || child->notify_table->len || !bus->parent_dev) {
|
|
method = build_alloc_method("PCNT", 0);
|
|
|
|
/* If bus supports hotplug select it and notify about local events */
|
|
if (bsel) {
|
|
build_append_byte(method, 0x70); /* StoreOp */
|
|
build_append_int(method, qint_get_int(qobject_to_qint(bsel)));
|
|
build_append_namestring(method, "BNUM");
|
|
build_append_namestring(method, "DVNT");
|
|
build_append_namestring(method, "PCIU");
|
|
build_append_int(method, 1); /* Device Check */
|
|
build_append_namestring(method, "DVNT");
|
|
build_append_namestring(method, "PCID");
|
|
build_append_int(method, 3); /* Eject Request */
|
|
}
|
|
|
|
/* Notify about child bus events in any case */
|
|
build_append_array(method, child->notify_table);
|
|
|
|
build_append_and_cleanup_method(bus_table, method);
|
|
|
|
/* Append description of child buses */
|
|
build_append_array(bus_table, child->device_table);
|
|
|
|
/* Pack it up */
|
|
if (bus->parent_dev) {
|
|
build_extop_package(bus_table, op);
|
|
} else {
|
|
build_package(bus_table, op);
|
|
}
|
|
|
|
/* Append our bus description to parent table */
|
|
build_append_array(parent->device_table, bus_table);
|
|
|
|
/* Also tell parent how to notify us, invoking PCNT method.
|
|
* At the moment this is not needed for root as we have a single root.
|
|
*/
|
|
if (bus->parent_dev) {
|
|
build_append_namestring(parent->notify_table, "^PCNT.S%.02X",
|
|
bus->parent_dev->devfn);
|
|
}
|
|
}
|
|
|
|
qobject_decref(bsel);
|
|
build_free_array(bus_table);
|
|
build_pci_bus_state_cleanup(child);
|
|
g_free(child);
|
|
}
|
|
|
|
static void patch_pci_windows(PcPciInfo *pci, uint8_t *start, unsigned size)
|
|
{
|
|
ACPI_BUILD_SET_LE(start, size, acpi_pci32_start[0], 32, pci->w32.begin);
|
|
|
|
ACPI_BUILD_SET_LE(start, size, acpi_pci32_end[0], 32, pci->w32.end - 1);
|
|
|
|
if (pci->w64.end || pci->w64.begin) {
|
|
ACPI_BUILD_SET_LE(start, size, acpi_pci64_valid[0], 8, 1);
|
|
ACPI_BUILD_SET_LE(start, size, acpi_pci64_start[0], 64, pci->w64.begin);
|
|
ACPI_BUILD_SET_LE(start, size, acpi_pci64_end[0], 64, pci->w64.end - 1);
|
|
ACPI_BUILD_SET_LE(start, size, acpi_pci64_length[0], 64, pci->w64.end - pci->w64.begin);
|
|
} else {
|
|
ACPI_BUILD_SET_LE(start, size, acpi_pci64_valid[0], 8, 0);
|
|
}
|
|
}
|
|
|
|
static void
|
|
build_ssdt(GArray *table_data, GArray *linker,
|
|
AcpiCpuInfo *cpu, AcpiPmInfo *pm, AcpiMiscInfo *misc,
|
|
PcPciInfo *pci, PcGuestInfo *guest_info)
|
|
{
|
|
MachineState *machine = MACHINE(qdev_get_machine());
|
|
uint32_t nr_mem = machine->ram_slots;
|
|
unsigned acpi_cpus = guest_info->apic_id_limit;
|
|
uint8_t *ssdt_ptr;
|
|
Aml *ssdt, *sb_scope, *scope, *pkg, *dev, *method, *crs, *field, *ifctx;
|
|
int i;
|
|
|
|
ssdt = init_aml_allocator();
|
|
/* The current AML generator can cover the APIC ID range [0..255],
|
|
* inclusive, for VCPU hotplug. */
|
|
QEMU_BUILD_BUG_ON(ACPI_CPU_HOTPLUG_ID_LIMIT > 256);
|
|
g_assert(acpi_cpus <= ACPI_CPU_HOTPLUG_ID_LIMIT);
|
|
|
|
/* Copy header and patch values in the S3_ / S4_ / S5_ packages */
|
|
ssdt_ptr = acpi_data_push(ssdt->buf, sizeof(ssdp_misc_aml));
|
|
memcpy(ssdt_ptr, ssdp_misc_aml, sizeof(ssdp_misc_aml));
|
|
|
|
patch_pci_windows(pci, ssdt_ptr, sizeof(ssdp_misc_aml));
|
|
|
|
ACPI_BUILD_SET_LE(ssdt_ptr, sizeof(ssdp_misc_aml),
|
|
ssdt_mctrl_nr_slots[0], 32, nr_mem);
|
|
|
|
/* create S3_ / S4_ / S5_ packages if necessary */
|
|
scope = aml_scope("\\");
|
|
if (!pm->s3_disabled) {
|
|
pkg = aml_package(4);
|
|
aml_append(pkg, aml_int(1)); /* PM1a_CNT.SLP_TYP */
|
|
aml_append(pkg, aml_int(1)); /* PM1b_CNT.SLP_TYP, FIXME: not impl. */
|
|
aml_append(pkg, aml_int(0)); /* reserved */
|
|
aml_append(pkg, aml_int(0)); /* reserved */
|
|
aml_append(scope, aml_name_decl("_S3", pkg));
|
|
}
|
|
|
|
if (!pm->s4_disabled) {
|
|
pkg = aml_package(4);
|
|
aml_append(pkg, aml_int(pm->s4_val)); /* PM1a_CNT.SLP_TYP */
|
|
/* PM1b_CNT.SLP_TYP, FIXME: not impl. */
|
|
aml_append(pkg, aml_int(pm->s4_val));
|
|
aml_append(pkg, aml_int(0)); /* reserved */
|
|
aml_append(pkg, aml_int(0)); /* reserved */
|
|
aml_append(scope, aml_name_decl("_S4", pkg));
|
|
}
|
|
|
|
pkg = aml_package(4);
|
|
aml_append(pkg, aml_int(0)); /* PM1a_CNT.SLP_TYP */
|
|
aml_append(pkg, aml_int(0)); /* PM1b_CNT.SLP_TYP not impl. */
|
|
aml_append(pkg, aml_int(0)); /* reserved */
|
|
aml_append(pkg, aml_int(0)); /* reserved */
|
|
aml_append(scope, aml_name_decl("_S5", pkg));
|
|
aml_append(ssdt, scope);
|
|
|
|
if (misc->pvpanic_port) {
|
|
scope = aml_scope("\\_SB.PCI0.ISA");
|
|
|
|
dev = aml_device("PEVR");
|
|
aml_append(dev, aml_name_decl("_HID", aml_string("QEMU0002")));
|
|
|
|
crs = aml_resource_template();
|
|
aml_append(crs,
|
|
aml_io(aml_decode16, misc->pvpanic_port, misc->pvpanic_port, 1, 1)
|
|
);
|
|
aml_append(dev, aml_name_decl("_CRS", crs));
|
|
|
|
aml_append(dev, aml_operation_region("PEOR", aml_system_io,
|
|
misc->pvpanic_port, 1));
|
|
field = aml_field("PEOR", aml_byte_acc);
|
|
aml_append(field, aml_named_field("PEPT", 8));
|
|
aml_append(dev, field);
|
|
|
|
method = aml_method("RDPT", 0);
|
|
aml_append(method, aml_store(aml_name("PEPT"), aml_local(0)));
|
|
aml_append(method, aml_return(aml_local(0)));
|
|
aml_append(dev, method);
|
|
|
|
method = aml_method("WRPT", 1);
|
|
aml_append(method, aml_store(aml_arg(0), aml_name("PEPT")));
|
|
aml_append(dev, method);
|
|
|
|
aml_append(scope, dev);
|
|
aml_append(ssdt, scope);
|
|
}
|
|
|
|
sb_scope = aml_scope("_SB");
|
|
{
|
|
/* create PCI0.PRES device and its _CRS to reserve CPU hotplug MMIO */
|
|
dev = aml_device("PCI0." stringify(CPU_HOTPLUG_RESOURCE_DEVICE));
|
|
aml_append(dev, aml_name_decl("_HID", aml_eisaid("PNP0A06")));
|
|
aml_append(dev,
|
|
aml_name_decl("_UID", aml_string("CPU Hotplug resources"))
|
|
);
|
|
/* device present, functioning, decoding, not shown in UI */
|
|
aml_append(dev, aml_name_decl("_STA", aml_int(0xB)));
|
|
crs = aml_resource_template();
|
|
aml_append(crs,
|
|
aml_io(aml_decode16, pm->cpu_hp_io_base, pm->cpu_hp_io_base, 1,
|
|
pm->cpu_hp_io_len)
|
|
);
|
|
aml_append(dev, aml_name_decl("_CRS", crs));
|
|
aml_append(sb_scope, dev);
|
|
/* declare CPU hotplug MMIO region and PRS field to access it */
|
|
aml_append(sb_scope, aml_operation_region(
|
|
"PRST", aml_system_io, pm->cpu_hp_io_base, pm->cpu_hp_io_len));
|
|
field = aml_field("PRST", aml_byte_acc);
|
|
aml_append(field, aml_named_field("PRS", 256));
|
|
aml_append(sb_scope, field);
|
|
|
|
/* build Processor object for each processor */
|
|
for (i = 0; i < acpi_cpus; i++) {
|
|
dev = aml_processor(i, 0, 0, "CP%.02X", i);
|
|
|
|
method = aml_method("_MAT", 0);
|
|
aml_append(method, aml_return(aml_call1("CPMA", aml_int(i))));
|
|
aml_append(dev, method);
|
|
|
|
method = aml_method("_STA", 0);
|
|
aml_append(method, aml_return(aml_call1("CPST", aml_int(i))));
|
|
aml_append(dev, method);
|
|
|
|
method = aml_method("_EJ0", 1);
|
|
aml_append(method,
|
|
aml_return(aml_call2("CPEJ", aml_int(i), aml_arg(0)))
|
|
);
|
|
aml_append(dev, method);
|
|
|
|
aml_append(sb_scope, dev);
|
|
}
|
|
|
|
/* build this code:
|
|
* Method(NTFY, 2) {If (LEqual(Arg0, 0x00)) {Notify(CP00, Arg1)} ...}
|
|
*/
|
|
/* Arg0 = Processor ID = APIC ID */
|
|
method = aml_method("NTFY", 2);
|
|
for (i = 0; i < acpi_cpus; i++) {
|
|
ifctx = aml_if(aml_equal(aml_arg(0), aml_int(i)));
|
|
aml_append(ifctx,
|
|
aml_notify(aml_name("CP%.02X", i), aml_arg(1))
|
|
);
|
|
aml_append(method, ifctx);
|
|
}
|
|
aml_append(sb_scope, method);
|
|
|
|
/* build "Name(CPON, Package() { One, One, ..., Zero, Zero, ... })"
|
|
*
|
|
* Note: The ability to create variable-sized packages was first
|
|
* ntroduced in ACPI 2.0. ACPI 1.0 only allowed fixed-size packages
|
|
* ith up to 255 elements. Windows guests up to win2k8 fail when
|
|
* VarPackageOp is used.
|
|
*/
|
|
pkg = acpi_cpus <= 255 ? aml_package(acpi_cpus) :
|
|
aml_varpackage(acpi_cpus);
|
|
|
|
for (i = 0; i < acpi_cpus; i++) {
|
|
uint8_t b = test_bit(i, cpu->found_cpus) ? 0x01 : 0x00;
|
|
aml_append(pkg, aml_int(b));
|
|
}
|
|
aml_append(sb_scope, aml_name_decl("CPON", pkg));
|
|
|
|
if (nr_mem) {
|
|
assert(nr_mem <= ACPI_MAX_RAM_SLOTS);
|
|
/* build memory devices */
|
|
for (i = 0; i < nr_mem; i++) {
|
|
char id[3];
|
|
uint8_t *mem = acpi_data_push(sb_scope->buf, ACPI_MEM_SIZEOF);
|
|
|
|
snprintf(id, sizeof(id), "%02X", i);
|
|
memcpy(mem, ACPI_MEM_AML, ACPI_MEM_SIZEOF);
|
|
memcpy(mem + ACPI_MEM_OFFSET_HEX, id, 2);
|
|
memcpy(mem + ACPI_MEM_OFFSET_ID, id, 2);
|
|
}
|
|
|
|
/* build Method(MEMORY_SLOT_NOTIFY_METHOD, 2) {
|
|
* If (LEqual(Arg0, 0x00)) {Notify(MP00, Arg1)} ...
|
|
*/
|
|
build_append_notify_method(sb_scope->buf,
|
|
stringify(MEMORY_SLOT_NOTIFY_METHOD),
|
|
"MP%0.02X", nr_mem);
|
|
}
|
|
|
|
{
|
|
AcpiBuildPciBusHotplugState hotplug_state;
|
|
Object *pci_host;
|
|
PCIBus *bus = NULL;
|
|
bool ambiguous;
|
|
|
|
pci_host = object_resolve_path_type("", TYPE_PCI_HOST_BRIDGE, &ambiguous);
|
|
if (!ambiguous && pci_host) {
|
|
bus = PCI_HOST_BRIDGE(pci_host)->bus;
|
|
}
|
|
|
|
build_pci_bus_state_init(&hotplug_state, NULL, pm->pcihp_bridge_en);
|
|
|
|
if (bus) {
|
|
/* Scan all PCI buses. Generate tables to support hotplug. */
|
|
pci_for_each_bus_depth_first(bus, build_pci_bus_begin,
|
|
build_pci_bus_end, &hotplug_state);
|
|
}
|
|
|
|
build_append_array(sb_scope->buf, hotplug_state.device_table);
|
|
build_pci_bus_state_cleanup(&hotplug_state);
|
|
}
|
|
aml_append(ssdt, sb_scope);
|
|
}
|
|
|
|
/* copy AML table into ACPI tables blob and patch header there */
|
|
g_array_append_vals(table_data, ssdt->buf->data, ssdt->buf->len);
|
|
build_header(linker, table_data,
|
|
(void *)(table_data->data + table_data->len - ssdt->buf->len),
|
|
"SSDT", ssdt->buf->len, 1);
|
|
free_aml_allocator();
|
|
}
|
|
|
|
static void
|
|
build_hpet(GArray *table_data, GArray *linker)
|
|
{
|
|
Acpi20Hpet *hpet;
|
|
|
|
hpet = acpi_data_push(table_data, sizeof(*hpet));
|
|
/* Note timer_block_id value must be kept in sync with value advertised by
|
|
* emulated hpet
|
|
*/
|
|
hpet->timer_block_id = cpu_to_le32(0x8086a201);
|
|
hpet->addr.address = cpu_to_le64(HPET_BASE);
|
|
build_header(linker, table_data,
|
|
(void *)hpet, "HPET", sizeof(*hpet), 1);
|
|
}
|
|
|
|
static void
|
|
build_tpm_tcpa(GArray *table_data, GArray *linker, GArray *tcpalog)
|
|
{
|
|
Acpi20Tcpa *tcpa = acpi_data_push(table_data, sizeof *tcpa);
|
|
uint64_t log_area_start_address = acpi_data_len(tcpalog);
|
|
|
|
tcpa->platform_class = cpu_to_le16(TPM_TCPA_ACPI_CLASS_CLIENT);
|
|
tcpa->log_area_minimum_length = cpu_to_le32(TPM_LOG_AREA_MINIMUM_SIZE);
|
|
tcpa->log_area_start_address = cpu_to_le64(log_area_start_address);
|
|
|
|
bios_linker_loader_alloc(linker, ACPI_BUILD_TPMLOG_FILE, 1,
|
|
false /* high memory */);
|
|
|
|
/* log area start address to be filled by Guest linker */
|
|
bios_linker_loader_add_pointer(linker, ACPI_BUILD_TABLE_FILE,
|
|
ACPI_BUILD_TPMLOG_FILE,
|
|
table_data, &tcpa->log_area_start_address,
|
|
sizeof(tcpa->log_area_start_address));
|
|
|
|
build_header(linker, table_data,
|
|
(void *)tcpa, "TCPA", sizeof(*tcpa), 2);
|
|
|
|
acpi_data_push(tcpalog, TPM_LOG_AREA_MINIMUM_SIZE);
|
|
}
|
|
|
|
static void
|
|
build_tpm_ssdt(GArray *table_data, GArray *linker)
|
|
{
|
|
void *tpm_ptr;
|
|
|
|
tpm_ptr = acpi_data_push(table_data, sizeof(ssdt_tpm_aml));
|
|
memcpy(tpm_ptr, ssdt_tpm_aml, sizeof(ssdt_tpm_aml));
|
|
}
|
|
|
|
typedef enum {
|
|
MEM_AFFINITY_NOFLAGS = 0,
|
|
MEM_AFFINITY_ENABLED = (1 << 0),
|
|
MEM_AFFINITY_HOTPLUGGABLE = (1 << 1),
|
|
MEM_AFFINITY_NON_VOLATILE = (1 << 2),
|
|
} MemoryAffinityFlags;
|
|
|
|
static void
|
|
acpi_build_srat_memory(AcpiSratMemoryAffinity *numamem, uint64_t base,
|
|
uint64_t len, int node, MemoryAffinityFlags flags)
|
|
{
|
|
numamem->type = ACPI_SRAT_MEMORY;
|
|
numamem->length = sizeof(*numamem);
|
|
memset(numamem->proximity, 0, 4);
|
|
numamem->proximity[0] = node;
|
|
numamem->flags = cpu_to_le32(flags);
|
|
numamem->base_addr = cpu_to_le64(base);
|
|
numamem->range_length = cpu_to_le64(len);
|
|
}
|
|
|
|
static void
|
|
build_srat(GArray *table_data, GArray *linker, PcGuestInfo *guest_info)
|
|
{
|
|
AcpiSystemResourceAffinityTable *srat;
|
|
AcpiSratProcessorAffinity *core;
|
|
AcpiSratMemoryAffinity *numamem;
|
|
|
|
int i;
|
|
uint64_t curnode;
|
|
int srat_start, numa_start, slots;
|
|
uint64_t mem_len, mem_base, next_base;
|
|
PCMachineState *pcms = PC_MACHINE(qdev_get_machine());
|
|
ram_addr_t hotplugabble_address_space_size =
|
|
object_property_get_int(OBJECT(pcms), PC_MACHINE_MEMHP_REGION_SIZE,
|
|
NULL);
|
|
|
|
srat_start = table_data->len;
|
|
|
|
srat = acpi_data_push(table_data, sizeof *srat);
|
|
srat->reserved1 = cpu_to_le32(1);
|
|
core = (void *)(srat + 1);
|
|
|
|
for (i = 0; i < guest_info->apic_id_limit; ++i) {
|
|
core = acpi_data_push(table_data, sizeof *core);
|
|
core->type = ACPI_SRAT_PROCESSOR;
|
|
core->length = sizeof(*core);
|
|
core->local_apic_id = i;
|
|
curnode = guest_info->node_cpu[i];
|
|
core->proximity_lo = curnode;
|
|
memset(core->proximity_hi, 0, 3);
|
|
core->local_sapic_eid = 0;
|
|
core->flags = cpu_to_le32(1);
|
|
}
|
|
|
|
|
|
/* the memory map is a bit tricky, it contains at least one hole
|
|
* from 640k-1M and possibly another one from 3.5G-4G.
|
|
*/
|
|
next_base = 0;
|
|
numa_start = table_data->len;
|
|
|
|
numamem = acpi_data_push(table_data, sizeof *numamem);
|
|
acpi_build_srat_memory(numamem, 0, 640*1024, 0, MEM_AFFINITY_ENABLED);
|
|
next_base = 1024 * 1024;
|
|
for (i = 1; i < guest_info->numa_nodes + 1; ++i) {
|
|
mem_base = next_base;
|
|
mem_len = guest_info->node_mem[i - 1];
|
|
if (i == 1) {
|
|
mem_len -= 1024 * 1024;
|
|
}
|
|
next_base = mem_base + mem_len;
|
|
|
|
/* Cut out the ACPI_PCI hole */
|
|
if (mem_base <= guest_info->ram_size_below_4g &&
|
|
next_base > guest_info->ram_size_below_4g) {
|
|
mem_len -= next_base - guest_info->ram_size_below_4g;
|
|
if (mem_len > 0) {
|
|
numamem = acpi_data_push(table_data, sizeof *numamem);
|
|
acpi_build_srat_memory(numamem, mem_base, mem_len, i - 1,
|
|
MEM_AFFINITY_ENABLED);
|
|
}
|
|
mem_base = 1ULL << 32;
|
|
mem_len = next_base - guest_info->ram_size_below_4g;
|
|
next_base += (1ULL << 32) - guest_info->ram_size_below_4g;
|
|
}
|
|
numamem = acpi_data_push(table_data, sizeof *numamem);
|
|
acpi_build_srat_memory(numamem, mem_base, mem_len, i - 1,
|
|
MEM_AFFINITY_ENABLED);
|
|
}
|
|
slots = (table_data->len - numa_start) / sizeof *numamem;
|
|
for (; slots < guest_info->numa_nodes + 2; slots++) {
|
|
numamem = acpi_data_push(table_data, sizeof *numamem);
|
|
acpi_build_srat_memory(numamem, 0, 0, 0, MEM_AFFINITY_NOFLAGS);
|
|
}
|
|
|
|
/*
|
|
* Entry is required for Windows to enable memory hotplug in OS.
|
|
* Memory devices may override proximity set by this entry,
|
|
* providing _PXM method if necessary.
|
|
*/
|
|
if (hotplugabble_address_space_size) {
|
|
numamem = acpi_data_push(table_data, sizeof *numamem);
|
|
acpi_build_srat_memory(numamem, pcms->hotplug_memory_base,
|
|
hotplugabble_address_space_size, 0,
|
|
MEM_AFFINITY_HOTPLUGGABLE |
|
|
MEM_AFFINITY_ENABLED);
|
|
}
|
|
|
|
build_header(linker, table_data,
|
|
(void *)(table_data->data + srat_start),
|
|
"SRAT",
|
|
table_data->len - srat_start, 1);
|
|
}
|
|
|
|
static void
|
|
build_mcfg_q35(GArray *table_data, GArray *linker, AcpiMcfgInfo *info)
|
|
{
|
|
AcpiTableMcfg *mcfg;
|
|
const char *sig;
|
|
int len = sizeof(*mcfg) + 1 * sizeof(mcfg->allocation[0]);
|
|
|
|
mcfg = acpi_data_push(table_data, len);
|
|
mcfg->allocation[0].address = cpu_to_le64(info->mcfg_base);
|
|
/* Only a single allocation so no need to play with segments */
|
|
mcfg->allocation[0].pci_segment = cpu_to_le16(0);
|
|
mcfg->allocation[0].start_bus_number = 0;
|
|
mcfg->allocation[0].end_bus_number = PCIE_MMCFG_BUS(info->mcfg_size - 1);
|
|
|
|
/* MCFG is used for ECAM which can be enabled or disabled by guest.
|
|
* To avoid table size changes (which create migration issues),
|
|
* always create the table even if there are no allocations,
|
|
* but set the signature to a reserved value in this case.
|
|
* ACPI spec requires OSPMs to ignore such tables.
|
|
*/
|
|
if (info->mcfg_base == PCIE_BASE_ADDR_UNMAPPED) {
|
|
/* Reserved signature: ignored by OSPM */
|
|
sig = "QEMU";
|
|
} else {
|
|
sig = "MCFG";
|
|
}
|
|
build_header(linker, table_data, (void *)mcfg, sig, len, 1);
|
|
}
|
|
|
|
static void
|
|
build_dmar_q35(GArray *table_data, GArray *linker)
|
|
{
|
|
int dmar_start = table_data->len;
|
|
|
|
AcpiTableDmar *dmar;
|
|
AcpiDmarHardwareUnit *drhd;
|
|
|
|
dmar = acpi_data_push(table_data, sizeof(*dmar));
|
|
dmar->host_address_width = VTD_HOST_ADDRESS_WIDTH - 1;
|
|
dmar->flags = 0; /* No intr_remap for now */
|
|
|
|
/* DMAR Remapping Hardware Unit Definition structure */
|
|
drhd = acpi_data_push(table_data, sizeof(*drhd));
|
|
drhd->type = cpu_to_le16(ACPI_DMAR_TYPE_HARDWARE_UNIT);
|
|
drhd->length = cpu_to_le16(sizeof(*drhd)); /* No device scope now */
|
|
drhd->flags = ACPI_DMAR_INCLUDE_PCI_ALL;
|
|
drhd->pci_segment = cpu_to_le16(0);
|
|
drhd->address = cpu_to_le64(Q35_HOST_BRIDGE_IOMMU_ADDR);
|
|
|
|
build_header(linker, table_data, (void *)(table_data->data + dmar_start),
|
|
"DMAR", table_data->len - dmar_start, 1);
|
|
}
|
|
|
|
static void
|
|
build_dsdt(GArray *table_data, GArray *linker, AcpiMiscInfo *misc)
|
|
{
|
|
AcpiTableHeader *dsdt;
|
|
|
|
assert(misc->dsdt_code && misc->dsdt_size);
|
|
|
|
dsdt = acpi_data_push(table_data, misc->dsdt_size);
|
|
memcpy(dsdt, misc->dsdt_code, misc->dsdt_size);
|
|
|
|
memset(dsdt, 0, sizeof *dsdt);
|
|
build_header(linker, table_data, dsdt, "DSDT",
|
|
misc->dsdt_size, 1);
|
|
}
|
|
|
|
/* Build final rsdt table */
|
|
static void
|
|
build_rsdt(GArray *table_data, GArray *linker, GArray *table_offsets)
|
|
{
|
|
AcpiRsdtDescriptorRev1 *rsdt;
|
|
size_t rsdt_len;
|
|
int i;
|
|
|
|
rsdt_len = sizeof(*rsdt) + sizeof(uint32_t) * table_offsets->len;
|
|
rsdt = acpi_data_push(table_data, rsdt_len);
|
|
memcpy(rsdt->table_offset_entry, table_offsets->data,
|
|
sizeof(uint32_t) * table_offsets->len);
|
|
for (i = 0; i < table_offsets->len; ++i) {
|
|
/* rsdt->table_offset_entry to be filled by Guest linker */
|
|
bios_linker_loader_add_pointer(linker,
|
|
ACPI_BUILD_TABLE_FILE,
|
|
ACPI_BUILD_TABLE_FILE,
|
|
table_data, &rsdt->table_offset_entry[i],
|
|
sizeof(uint32_t));
|
|
}
|
|
build_header(linker, table_data,
|
|
(void *)rsdt, "RSDT", rsdt_len, 1);
|
|
}
|
|
|
|
static GArray *
|
|
build_rsdp(GArray *rsdp_table, GArray *linker, unsigned rsdt)
|
|
{
|
|
AcpiRsdpDescriptor *rsdp = acpi_data_push(rsdp_table, sizeof *rsdp);
|
|
|
|
bios_linker_loader_alloc(linker, ACPI_BUILD_RSDP_FILE, 16,
|
|
true /* fseg memory */);
|
|
|
|
memcpy(&rsdp->signature, "RSD PTR ", 8);
|
|
memcpy(rsdp->oem_id, ACPI_BUILD_APPNAME6, 6);
|
|
rsdp->rsdt_physical_address = cpu_to_le32(rsdt);
|
|
/* Address to be filled by Guest linker */
|
|
bios_linker_loader_add_pointer(linker, ACPI_BUILD_RSDP_FILE,
|
|
ACPI_BUILD_TABLE_FILE,
|
|
rsdp_table, &rsdp->rsdt_physical_address,
|
|
sizeof rsdp->rsdt_physical_address);
|
|
rsdp->checksum = 0;
|
|
/* Checksum to be filled by Guest linker */
|
|
bios_linker_loader_add_checksum(linker, ACPI_BUILD_RSDP_FILE,
|
|
rsdp, rsdp, sizeof *rsdp, &rsdp->checksum);
|
|
|
|
return rsdp_table;
|
|
}
|
|
|
|
typedef
|
|
struct AcpiBuildTables {
|
|
GArray *table_data;
|
|
GArray *rsdp;
|
|
GArray *tcpalog;
|
|
GArray *linker;
|
|
} AcpiBuildTables;
|
|
|
|
static inline void acpi_build_tables_init(AcpiBuildTables *tables)
|
|
{
|
|
tables->rsdp = g_array_new(false, true /* clear */, 1);
|
|
tables->table_data = g_array_new(false, true /* clear */, 1);
|
|
tables->tcpalog = g_array_new(false, true /* clear */, 1);
|
|
tables->linker = bios_linker_loader_init();
|
|
}
|
|
|
|
static inline void acpi_build_tables_cleanup(AcpiBuildTables *tables, bool mfre)
|
|
{
|
|
void *linker_data = bios_linker_loader_cleanup(tables->linker);
|
|
g_free(linker_data);
|
|
g_array_free(tables->rsdp, true);
|
|
g_array_free(tables->table_data, true);
|
|
g_array_free(tables->tcpalog, mfre);
|
|
}
|
|
|
|
typedef
|
|
struct AcpiBuildState {
|
|
/* Copy of table in RAM (for patching). */
|
|
ram_addr_t table_ram;
|
|
/* Is table patched? */
|
|
uint8_t patched;
|
|
PcGuestInfo *guest_info;
|
|
void *rsdp;
|
|
ram_addr_t rsdp_ram;
|
|
ram_addr_t linker_ram;
|
|
} AcpiBuildState;
|
|
|
|
static bool acpi_get_mcfg(AcpiMcfgInfo *mcfg)
|
|
{
|
|
Object *pci_host;
|
|
QObject *o;
|
|
bool ambiguous;
|
|
|
|
pci_host = object_resolve_path_type("", TYPE_PCI_HOST_BRIDGE, &ambiguous);
|
|
g_assert(!ambiguous);
|
|
g_assert(pci_host);
|
|
|
|
o = object_property_get_qobject(pci_host, PCIE_HOST_MCFG_BASE, NULL);
|
|
if (!o) {
|
|
return false;
|
|
}
|
|
mcfg->mcfg_base = qint_get_int(qobject_to_qint(o));
|
|
qobject_decref(o);
|
|
|
|
o = object_property_get_qobject(pci_host, PCIE_HOST_MCFG_SIZE, NULL);
|
|
assert(o);
|
|
mcfg->mcfg_size = qint_get_int(qobject_to_qint(o));
|
|
qobject_decref(o);
|
|
return true;
|
|
}
|
|
|
|
static bool acpi_has_iommu(void)
|
|
{
|
|
bool ambiguous;
|
|
Object *intel_iommu;
|
|
|
|
intel_iommu = object_resolve_path_type("", TYPE_INTEL_IOMMU_DEVICE,
|
|
&ambiguous);
|
|
return intel_iommu && !ambiguous;
|
|
}
|
|
|
|
static
|
|
void acpi_build(PcGuestInfo *guest_info, AcpiBuildTables *tables)
|
|
{
|
|
GArray *table_offsets;
|
|
unsigned facs, ssdt, dsdt, rsdt;
|
|
AcpiCpuInfo cpu;
|
|
AcpiPmInfo pm;
|
|
AcpiMiscInfo misc;
|
|
AcpiMcfgInfo mcfg;
|
|
PcPciInfo pci;
|
|
uint8_t *u;
|
|
size_t aml_len = 0;
|
|
GArray *tables_blob = tables->table_data;
|
|
|
|
acpi_get_cpu_info(&cpu);
|
|
acpi_get_pm_info(&pm);
|
|
acpi_get_dsdt(&misc);
|
|
acpi_get_misc_info(&misc);
|
|
acpi_get_pci_info(&pci);
|
|
|
|
table_offsets = g_array_new(false, true /* clear */,
|
|
sizeof(uint32_t));
|
|
ACPI_BUILD_DPRINTF("init ACPI tables\n");
|
|
|
|
bios_linker_loader_alloc(tables->linker, ACPI_BUILD_TABLE_FILE,
|
|
64 /* Ensure FACS is aligned */,
|
|
false /* high memory */);
|
|
|
|
/*
|
|
* FACS is pointed to by FADT.
|
|
* We place it first since it's the only table that has alignment
|
|
* requirements.
|
|
*/
|
|
facs = tables_blob->len;
|
|
build_facs(tables_blob, tables->linker, guest_info);
|
|
|
|
/* DSDT is pointed to by FADT */
|
|
dsdt = tables_blob->len;
|
|
build_dsdt(tables_blob, tables->linker, &misc);
|
|
|
|
/* Count the size of the DSDT and SSDT, we will need it for legacy
|
|
* sizing of ACPI tables.
|
|
*/
|
|
aml_len += tables_blob->len - dsdt;
|
|
|
|
/* ACPI tables pointed to by RSDT */
|
|
acpi_add_table(table_offsets, tables_blob);
|
|
build_fadt(tables_blob, tables->linker, &pm, facs, dsdt);
|
|
|
|
ssdt = tables_blob->len;
|
|
acpi_add_table(table_offsets, tables_blob);
|
|
build_ssdt(tables_blob, tables->linker, &cpu, &pm, &misc, &pci,
|
|
guest_info);
|
|
aml_len += tables_blob->len - ssdt;
|
|
|
|
acpi_add_table(table_offsets, tables_blob);
|
|
build_madt(tables_blob, tables->linker, &cpu, guest_info);
|
|
|
|
if (misc.has_hpet) {
|
|
acpi_add_table(table_offsets, tables_blob);
|
|
build_hpet(tables_blob, tables->linker);
|
|
}
|
|
if (misc.has_tpm) {
|
|
acpi_add_table(table_offsets, tables_blob);
|
|
build_tpm_tcpa(tables_blob, tables->linker, tables->tcpalog);
|
|
|
|
acpi_add_table(table_offsets, tables_blob);
|
|
build_tpm_ssdt(tables_blob, tables->linker);
|
|
}
|
|
if (guest_info->numa_nodes) {
|
|
acpi_add_table(table_offsets, tables_blob);
|
|
build_srat(tables_blob, tables->linker, guest_info);
|
|
}
|
|
if (acpi_get_mcfg(&mcfg)) {
|
|
acpi_add_table(table_offsets, tables_blob);
|
|
build_mcfg_q35(tables_blob, tables->linker, &mcfg);
|
|
}
|
|
if (acpi_has_iommu()) {
|
|
acpi_add_table(table_offsets, tables_blob);
|
|
build_dmar_q35(tables_blob, tables->linker);
|
|
}
|
|
|
|
/* Add tables supplied by user (if any) */
|
|
for (u = acpi_table_first(); u; u = acpi_table_next(u)) {
|
|
unsigned len = acpi_table_len(u);
|
|
|
|
acpi_add_table(table_offsets, tables_blob);
|
|
g_array_append_vals(tables_blob, u, len);
|
|
}
|
|
|
|
/* RSDT is pointed to by RSDP */
|
|
rsdt = tables_blob->len;
|
|
build_rsdt(tables_blob, tables->linker, table_offsets);
|
|
|
|
/* RSDP is in FSEG memory, so allocate it separately */
|
|
build_rsdp(tables->rsdp, tables->linker, rsdt);
|
|
|
|
/* We'll expose it all to Guest so we want to reduce
|
|
* chance of size changes.
|
|
*
|
|
* We used to align the tables to 4k, but of course this would
|
|
* too simple to be enough. 4k turned out to be too small an
|
|
* alignment very soon, and in fact it is almost impossible to
|
|
* keep the table size stable for all (max_cpus, max_memory_slots)
|
|
* combinations. So the table size is always 64k for pc-i440fx-2.1
|
|
* and we give an error if the table grows beyond that limit.
|
|
*
|
|
* We still have the problem of migrating from "-M pc-i440fx-2.0". For
|
|
* that, we exploit the fact that QEMU 2.1 generates _smaller_ tables
|
|
* than 2.0 and we can always pad the smaller tables with zeros. We can
|
|
* then use the exact size of the 2.0 tables.
|
|
*
|
|
* All this is for PIIX4, since QEMU 2.0 didn't support Q35 migration.
|
|
*/
|
|
if (guest_info->legacy_acpi_table_size) {
|
|
/* Subtracting aml_len gives the size of fixed tables. Then add the
|
|
* size of the PIIX4 DSDT/SSDT in QEMU 2.0.
|
|
*/
|
|
int legacy_aml_len =
|
|
guest_info->legacy_acpi_table_size +
|
|
ACPI_BUILD_LEGACY_CPU_AML_SIZE * max_cpus;
|
|
int legacy_table_size =
|
|
ROUND_UP(tables_blob->len - aml_len + legacy_aml_len,
|
|
ACPI_BUILD_ALIGN_SIZE);
|
|
if (tables_blob->len > legacy_table_size) {
|
|
/* Should happen only with PCI bridges and -M pc-i440fx-2.0. */
|
|
error_report("Warning: migration may not work.");
|
|
}
|
|
g_array_set_size(tables_blob, legacy_table_size);
|
|
} else {
|
|
/* Make sure we have a buffer in case we need to resize the tables. */
|
|
if (tables_blob->len > ACPI_BUILD_TABLE_SIZE / 2) {
|
|
/* As of QEMU 2.1, this fires with 160 VCPUs and 255 memory slots. */
|
|
error_report("Warning: ACPI tables are larger than 64k.");
|
|
error_report("Warning: migration may not work.");
|
|
error_report("Warning: please remove CPUs, NUMA nodes, "
|
|
"memory slots or PCI bridges.");
|
|
}
|
|
acpi_align_size(tables_blob, ACPI_BUILD_TABLE_SIZE);
|
|
}
|
|
|
|
acpi_align_size(tables->linker, ACPI_BUILD_ALIGN_SIZE);
|
|
|
|
/* Cleanup memory that's no longer used. */
|
|
g_array_free(table_offsets, true);
|
|
}
|
|
|
|
static void acpi_ram_update(ram_addr_t ram, GArray *data)
|
|
{
|
|
uint32_t size = acpi_data_len(data);
|
|
|
|
/* Make sure RAM size is correct - in case it got changed e.g. by migration */
|
|
qemu_ram_resize(ram, size, &error_abort);
|
|
|
|
memcpy(qemu_get_ram_ptr(ram), data->data, size);
|
|
cpu_physical_memory_set_dirty_range_nocode(ram, size);
|
|
}
|
|
|
|
static void acpi_build_update(void *build_opaque, uint32_t offset)
|
|
{
|
|
AcpiBuildState *build_state = build_opaque;
|
|
AcpiBuildTables tables;
|
|
|
|
/* No state to update or already patched? Nothing to do. */
|
|
if (!build_state || build_state->patched) {
|
|
return;
|
|
}
|
|
build_state->patched = 1;
|
|
|
|
acpi_build_tables_init(&tables);
|
|
|
|
acpi_build(build_state->guest_info, &tables);
|
|
|
|
acpi_ram_update(build_state->table_ram, tables.table_data);
|
|
|
|
if (build_state->rsdp) {
|
|
memcpy(build_state->rsdp, tables.rsdp->data, acpi_data_len(tables.rsdp));
|
|
} else {
|
|
acpi_ram_update(build_state->rsdp_ram, tables.rsdp);
|
|
}
|
|
|
|
acpi_ram_update(build_state->linker_ram, tables.linker);
|
|
acpi_build_tables_cleanup(&tables, true);
|
|
}
|
|
|
|
static void acpi_build_reset(void *build_opaque)
|
|
{
|
|
AcpiBuildState *build_state = build_opaque;
|
|
build_state->patched = 0;
|
|
}
|
|
|
|
static ram_addr_t acpi_add_rom_blob(AcpiBuildState *build_state, GArray *blob,
|
|
const char *name, uint64_t max_size)
|
|
{
|
|
return rom_add_blob(name, blob->data, acpi_data_len(blob), max_size, -1,
|
|
name, acpi_build_update, build_state);
|
|
}
|
|
|
|
static const VMStateDescription vmstate_acpi_build = {
|
|
.name = "acpi_build",
|
|
.version_id = 1,
|
|
.minimum_version_id = 1,
|
|
.fields = (VMStateField[]) {
|
|
VMSTATE_UINT8(patched, AcpiBuildState),
|
|
VMSTATE_END_OF_LIST()
|
|
},
|
|
};
|
|
|
|
void acpi_setup(PcGuestInfo *guest_info)
|
|
{
|
|
AcpiBuildTables tables;
|
|
AcpiBuildState *build_state;
|
|
|
|
if (!guest_info->fw_cfg) {
|
|
ACPI_BUILD_DPRINTF("No fw cfg. Bailing out.\n");
|
|
return;
|
|
}
|
|
|
|
if (!guest_info->has_acpi_build) {
|
|
ACPI_BUILD_DPRINTF("ACPI build disabled. Bailing out.\n");
|
|
return;
|
|
}
|
|
|
|
if (!acpi_enabled) {
|
|
ACPI_BUILD_DPRINTF("ACPI disabled. Bailing out.\n");
|
|
return;
|
|
}
|
|
|
|
build_state = g_malloc0(sizeof *build_state);
|
|
|
|
build_state->guest_info = guest_info;
|
|
|
|
acpi_set_pci_info();
|
|
|
|
acpi_build_tables_init(&tables);
|
|
acpi_build(build_state->guest_info, &tables);
|
|
|
|
/* Now expose it all to Guest */
|
|
build_state->table_ram = acpi_add_rom_blob(build_state, tables.table_data,
|
|
ACPI_BUILD_TABLE_FILE,
|
|
ACPI_BUILD_TABLE_MAX_SIZE);
|
|
assert(build_state->table_ram != RAM_ADDR_MAX);
|
|
|
|
build_state->linker_ram =
|
|
acpi_add_rom_blob(build_state, tables.linker, "etc/table-loader", 0);
|
|
|
|
fw_cfg_add_file(guest_info->fw_cfg, ACPI_BUILD_TPMLOG_FILE,
|
|
tables.tcpalog->data, acpi_data_len(tables.tcpalog));
|
|
|
|
if (!guest_info->rsdp_in_ram) {
|
|
/*
|
|
* Keep for compatibility with old machine types.
|
|
* Though RSDP is small, its contents isn't immutable, so
|
|
* we'll update it along with the rest of tables on guest access.
|
|
*/
|
|
uint32_t rsdp_size = acpi_data_len(tables.rsdp);
|
|
|
|
build_state->rsdp = g_memdup(tables.rsdp->data, rsdp_size);
|
|
fw_cfg_add_file_callback(guest_info->fw_cfg, ACPI_BUILD_RSDP_FILE,
|
|
acpi_build_update, build_state,
|
|
build_state->rsdp, rsdp_size);
|
|
build_state->rsdp_ram = (ram_addr_t)-1;
|
|
} else {
|
|
build_state->rsdp = NULL;
|
|
build_state->rsdp_ram = acpi_add_rom_blob(build_state, tables.rsdp,
|
|
ACPI_BUILD_RSDP_FILE, 0);
|
|
}
|
|
|
|
qemu_register_reset(acpi_build_reset, build_state);
|
|
acpi_build_reset(build_state);
|
|
vmstate_register(NULL, 0, &vmstate_acpi_build, build_state);
|
|
|
|
/* Cleanup tables but don't free the memory: we track it
|
|
* in build_state.
|
|
*/
|
|
acpi_build_tables_cleanup(&tables, false);
|
|
}
|