6410f877f5
- Restrict probe_access*() functions to TCG (Phil) - Extract do_invalidate_device_tlb from vtd_process_device_iotlb_desc (Clément) - Fixes in Loongson IPI model (Bibo & Phil) - Make docs/interop/firmware.json compatible with qapi-gen.py script (Thomas) - Correct MPC I2C MMIO region size (Zoltan) - Remove useless cast in Loongson3 Virt machine (Yao) - Various uses of range overlap API (Yao) - Use ERRP_GUARD macro in nubus_virtio_mmio_realize (Zhao) - Use DMA memory API in Goldfish UART model (Phil) - Expose fifo8_pop_buf and introduce fifo8_drop (Phil) - MAINTAINERS updates (Zhao, Phil) -----BEGIN PGP SIGNATURE----- iQIzBAABCAAdFiEE+qvnXhKRciHc/Wuy4+MsLN6twN4FAmagFF8ACgkQ4+MsLN6t wN5bKg//f5TwUhsy2ff0FJpHheDOj/9Gc2nZ1U/Fp0E5N3sz3A7MGp91wye6Xwi3 XG34YN9LK1AVzuCdrEEs5Uaxs1ZS1R2mV+fZaGHwYYxPDdnXxGyp/2Q0eyRxzbcN zxE2hWscYSZbPVEru4HvZJKfp4XnE1cqA78fJKMAdtq0IPq38tmQNRlJ+gWD9dC6 ZUHXPFf3DnucvVuwqb0JYO/E+uJpcTtgR6pc09Xtv/HFgMiS0vKZ1I/6LChqAUw9 eLMpD/5V2naemVadJe98/dL7gIUnhB8GTjsb4ioblG59AO/uojutwjBSQvFxBUUw U5lX9OSn20ouwcGiqimsz+5ziwhCG0R6r1zeQJFqUxrpZSscq7NQp9ygbvirm+wS edLc8yTPf4MtYOihzPP9jLPcXPZjEV64gSnJISDDFYWANCrysX3suaFEOuVYPl+s ZgQYRVSSYOYHgNqBSRkPKKVUxskSQiqLY3SfGJG4EA9Ktt5lD1cLCXQxhdsqphFm Ws3zkrVVL0EKl4v/4MtCgITIIctN1ZJE9u3oPJjASqSvK6EebFqAJkc2SidzKHz0 F3iYX2AheWNHCQ3HFu023EvFryjlxYk95fs2f6Uj2a9yVbi813qsvd3gcZ8t0kTT +dmQwpu1MxjzZnA6838R6OCMnC+UpMPqQh3dPkU/5AF2fc3NnN8= =J/I2 -----END PGP SIGNATURE----- Merge tag 'hw-misc-20240723' of https://github.com/philmd/qemu into staging Misc HW patch queue - Restrict probe_access*() functions to TCG (Phil) - Extract do_invalidate_device_tlb from vtd_process_device_iotlb_desc (Clément) - Fixes in Loongson IPI model (Bibo & Phil) - Make docs/interop/firmware.json compatible with qapi-gen.py script (Thomas) - Correct MPC I2C MMIO region size (Zoltan) - Remove useless cast in Loongson3 Virt machine (Yao) - Various uses of range overlap API (Yao) - Use ERRP_GUARD macro in nubus_virtio_mmio_realize (Zhao) - Use DMA memory API in Goldfish UART model (Phil) - Expose fifo8_pop_buf and introduce fifo8_drop (Phil) - MAINTAINERS updates (Zhao, Phil) # -----BEGIN PGP SIGNATURE----- # # iQIzBAABCAAdFiEE+qvnXhKRciHc/Wuy4+MsLN6twN4FAmagFF8ACgkQ4+MsLN6t # wN5bKg//f5TwUhsy2ff0FJpHheDOj/9Gc2nZ1U/Fp0E5N3sz3A7MGp91wye6Xwi3 # XG34YN9LK1AVzuCdrEEs5Uaxs1ZS1R2mV+fZaGHwYYxPDdnXxGyp/2Q0eyRxzbcN # zxE2hWscYSZbPVEru4HvZJKfp4XnE1cqA78fJKMAdtq0IPq38tmQNRlJ+gWD9dC6 # ZUHXPFf3DnucvVuwqb0JYO/E+uJpcTtgR6pc09Xtv/HFgMiS0vKZ1I/6LChqAUw9 # eLMpD/5V2naemVadJe98/dL7gIUnhB8GTjsb4ioblG59AO/uojutwjBSQvFxBUUw # U5lX9OSn20ouwcGiqimsz+5ziwhCG0R6r1zeQJFqUxrpZSscq7NQp9ygbvirm+wS # edLc8yTPf4MtYOihzPP9jLPcXPZjEV64gSnJISDDFYWANCrysX3suaFEOuVYPl+s # ZgQYRVSSYOYHgNqBSRkPKKVUxskSQiqLY3SfGJG4EA9Ktt5lD1cLCXQxhdsqphFm # Ws3zkrVVL0EKl4v/4MtCgITIIctN1ZJE9u3oPJjASqSvK6EebFqAJkc2SidzKHz0 # F3iYX2AheWNHCQ3HFu023EvFryjlxYk95fs2f6Uj2a9yVbi813qsvd3gcZ8t0kTT # +dmQwpu1MxjzZnA6838R6OCMnC+UpMPqQh3dPkU/5AF2fc3NnN8= # =J/I2 # -----END PGP SIGNATURE----- # gpg: Signature made Wed 24 Jul 2024 06:36:47 AM AEST # gpg: using RSA key FAABE75E12917221DCFD6BB2E3E32C2CDEADC0DE # gpg: Good signature from "Philippe Mathieu-Daudé (F4BUG) <f4bug@amsat.org>" [full] * tag 'hw-misc-20240723' of https://github.com/philmd/qemu: (28 commits) MAINTAINERS: Add myself as a reviewer of machine core MAINTAINERS: Cover guest-agent in QAPI schema util/fifo8: Introduce fifo8_drop() util/fifo8: Expose fifo8_pop_buf() util/fifo8: Rename fifo8_pop_buf() -> fifo8_pop_bufptr() util/fifo8: Rename fifo8_peek_buf() -> fifo8_peek_bufptr() util/fifo8: Use fifo8_reset() in fifo8_create() util/fifo8: Fix style chardev/char-fe: Document returned value on error hw/char/goldfish: Use DMA memory API hw/nubus/virtio-mmio: Fix missing ERRP_GUARD() in realize handler dump: make range overlap check more readable crypto/block-luks: make range overlap check more readable system/memory_mapping: make range overlap check more readable sparc/ldst_helper: make range overlap check more readable cxl/mailbox: make range overlap check more readable util/range: Make ranges_overlap() return bool hw/mips/loongson3_virt: remove useless type cast hw/i2c/mpc_i2c: Fix mmio region size docs/interop/firmware.json: convert "Example" section ... Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
3016 lines
103 KiB
C
3016 lines
103 KiB
C
/*
|
||
* CXL Utility library for mailbox interface
|
||
*
|
||
* Copyright(C) 2020 Intel Corporation.
|
||
*
|
||
* This work is licensed under the terms of the GNU GPL, version 2. See the
|
||
* COPYING file in the top-level directory.
|
||
*/
|
||
|
||
#include "qemu/osdep.h"
|
||
#include "hw/pci/msi.h"
|
||
#include "hw/pci/msix.h"
|
||
#include "hw/cxl/cxl.h"
|
||
#include "hw/cxl/cxl_events.h"
|
||
#include "hw/cxl/cxl_mailbox.h"
|
||
#include "hw/pci/pci.h"
|
||
#include "hw/pci-bridge/cxl_upstream_port.h"
|
||
#include "qemu/cutils.h"
|
||
#include "qemu/log.h"
|
||
#include "qemu/units.h"
|
||
#include "qemu/uuid.h"
|
||
#include "sysemu/hostmem.h"
|
||
#include "qemu/range.h"
|
||
|
||
#define CXL_CAPACITY_MULTIPLIER (256 * MiB)
|
||
#define CXL_DC_EVENT_LOG_SIZE 8
|
||
#define CXL_NUM_EXTENTS_SUPPORTED 512
|
||
#define CXL_NUM_TAGS_SUPPORTED 0
|
||
|
||
/*
|
||
* How to add a new command, example. The command set FOO, with cmd BAR.
|
||
* 1. Add the command set and cmd to the enum.
|
||
* FOO = 0x7f,
|
||
* #define BAR 0
|
||
* 2. Implement the handler
|
||
* static CXLRetCode cmd_foo_bar(struct cxl_cmd *cmd,
|
||
* CXLDeviceState *cxl_dstate, uint16_t *len)
|
||
* 3. Add the command to the cxl_cmd_set[][]
|
||
* [FOO][BAR] = { "FOO_BAR", cmd_foo_bar, x, y },
|
||
* 4. Implement your handler
|
||
* define_mailbox_handler(FOO_BAR) { ... return CXL_MBOX_SUCCESS; }
|
||
*
|
||
*
|
||
* Writing the handler:
|
||
* The handler will provide the &struct cxl_cmd, the &CXLDeviceState, and the
|
||
* in/out length of the payload. The handler is responsible for consuming the
|
||
* payload from cmd->payload and operating upon it as necessary. It must then
|
||
* fill the output data into cmd->payload (overwriting what was there),
|
||
* setting the length, and returning a valid return code.
|
||
*
|
||
* XXX: The handler need not worry about endianness. The payload is read out of
|
||
* a register interface that already deals with it.
|
||
*/
|
||
|
||
enum {
|
||
INFOSTAT = 0x00,
|
||
#define IS_IDENTIFY 0x1
|
||
#define BACKGROUND_OPERATION_STATUS 0x2
|
||
EVENTS = 0x01,
|
||
#define GET_RECORDS 0x0
|
||
#define CLEAR_RECORDS 0x1
|
||
#define GET_INTERRUPT_POLICY 0x2
|
||
#define SET_INTERRUPT_POLICY 0x3
|
||
FIRMWARE_UPDATE = 0x02,
|
||
#define GET_INFO 0x0
|
||
#define TRANSFER 0x1
|
||
#define ACTIVATE 0x2
|
||
TIMESTAMP = 0x03,
|
||
#define GET 0x0
|
||
#define SET 0x1
|
||
LOGS = 0x04,
|
||
#define GET_SUPPORTED 0x0
|
||
#define GET_LOG 0x1
|
||
FEATURES = 0x05,
|
||
#define GET_SUPPORTED 0x0
|
||
#define GET_FEATURE 0x1
|
||
#define SET_FEATURE 0x2
|
||
IDENTIFY = 0x40,
|
||
#define MEMORY_DEVICE 0x0
|
||
CCLS = 0x41,
|
||
#define GET_PARTITION_INFO 0x0
|
||
#define GET_LSA 0x2
|
||
#define SET_LSA 0x3
|
||
SANITIZE = 0x44,
|
||
#define OVERWRITE 0x0
|
||
#define SECURE_ERASE 0x1
|
||
PERSISTENT_MEM = 0x45,
|
||
#define GET_SECURITY_STATE 0x0
|
||
MEDIA_AND_POISON = 0x43,
|
||
#define GET_POISON_LIST 0x0
|
||
#define INJECT_POISON 0x1
|
||
#define CLEAR_POISON 0x2
|
||
#define GET_SCAN_MEDIA_CAPABILITIES 0x3
|
||
#define SCAN_MEDIA 0x4
|
||
#define GET_SCAN_MEDIA_RESULTS 0x5
|
||
DCD_CONFIG = 0x48,
|
||
#define GET_DC_CONFIG 0x0
|
||
#define GET_DYN_CAP_EXT_LIST 0x1
|
||
#define ADD_DYN_CAP_RSP 0x2
|
||
#define RELEASE_DYN_CAP 0x3
|
||
PHYSICAL_SWITCH = 0x51,
|
||
#define IDENTIFY_SWITCH_DEVICE 0x0
|
||
#define GET_PHYSICAL_PORT_STATE 0x1
|
||
TUNNEL = 0x53,
|
||
#define MANAGEMENT_COMMAND 0x0
|
||
};
|
||
|
||
/* CCI Message Format CXL r3.1 Figure 7-19 */
|
||
typedef struct CXLCCIMessage {
|
||
uint8_t category;
|
||
#define CXL_CCI_CAT_REQ 0
|
||
#define CXL_CCI_CAT_RSP 1
|
||
uint8_t tag;
|
||
uint8_t resv1;
|
||
uint8_t command;
|
||
uint8_t command_set;
|
||
uint8_t pl_length[3];
|
||
uint16_t rc;
|
||
uint16_t vendor_specific;
|
||
uint8_t payload[];
|
||
} QEMU_PACKED CXLCCIMessage;
|
||
|
||
/* This command is only defined to an MLD FM Owned LD or an MHD */
|
||
static CXLRetCode cmd_tunnel_management_cmd(const struct cxl_cmd *cmd,
|
||
uint8_t *payload_in,
|
||
size_t len_in,
|
||
uint8_t *payload_out,
|
||
size_t *len_out,
|
||
CXLCCI *cci)
|
||
{
|
||
PCIDevice *tunnel_target;
|
||
CXLCCI *target_cci;
|
||
struct {
|
||
uint8_t port_or_ld_id;
|
||
uint8_t target_type;
|
||
uint16_t size;
|
||
CXLCCIMessage ccimessage;
|
||
} QEMU_PACKED *in;
|
||
struct {
|
||
uint16_t resp_len;
|
||
uint8_t resv[2];
|
||
CXLCCIMessage ccimessage;
|
||
} QEMU_PACKED *out;
|
||
size_t pl_length, length_out;
|
||
bool bg_started;
|
||
int rc;
|
||
|
||
if (cmd->in < sizeof(*in)) {
|
||
return CXL_MBOX_INVALID_INPUT;
|
||
}
|
||
in = (void *)payload_in;
|
||
out = (void *)payload_out;
|
||
|
||
/* Enough room for minimum sized message - no payload */
|
||
if (in->size < sizeof(in->ccimessage)) {
|
||
return CXL_MBOX_INVALID_PAYLOAD_LENGTH;
|
||
}
|
||
/* Length of input payload should be in->size + a wrapping tunnel header */
|
||
if (in->size != len_in - offsetof(typeof(*out), ccimessage)) {
|
||
return CXL_MBOX_INVALID_PAYLOAD_LENGTH;
|
||
}
|
||
if (in->ccimessage.category != CXL_CCI_CAT_REQ) {
|
||
return CXL_MBOX_INVALID_INPUT;
|
||
}
|
||
|
||
if (in->target_type != 0) {
|
||
qemu_log_mask(LOG_UNIMP,
|
||
"Tunneled Command sent to non existent FM-LD");
|
||
return CXL_MBOX_INVALID_INPUT;
|
||
}
|
||
|
||
/*
|
||
* Target of a tunnel unfortunately depends on type of CCI readint
|
||
* the message.
|
||
* If in a switch, then it's the port number.
|
||
* If in an MLD it is the ld number.
|
||
* If in an MHD target type indicate where we are going.
|
||
*/
|
||
if (object_dynamic_cast(OBJECT(cci->d), TYPE_CXL_TYPE3)) {
|
||
CXLType3Dev *ct3d = CXL_TYPE3(cci->d);
|
||
if (in->port_or_ld_id != 0) {
|
||
/* Only pretending to have one for now! */
|
||
return CXL_MBOX_INVALID_INPUT;
|
||
}
|
||
target_cci = &ct3d->ld0_cci;
|
||
} else if (object_dynamic_cast(OBJECT(cci->d), TYPE_CXL_USP)) {
|
||
CXLUpstreamPort *usp = CXL_USP(cci->d);
|
||
|
||
tunnel_target = pcie_find_port_by_pn(&PCI_BRIDGE(usp)->sec_bus,
|
||
in->port_or_ld_id);
|
||
if (!tunnel_target) {
|
||
return CXL_MBOX_INVALID_INPUT;
|
||
}
|
||
tunnel_target =
|
||
pci_bridge_get_sec_bus(PCI_BRIDGE(tunnel_target))->devices[0];
|
||
if (!tunnel_target) {
|
||
return CXL_MBOX_INVALID_INPUT;
|
||
}
|
||
if (object_dynamic_cast(OBJECT(tunnel_target), TYPE_CXL_TYPE3)) {
|
||
CXLType3Dev *ct3d = CXL_TYPE3(tunnel_target);
|
||
/* Tunneled VDMs always land on FM Owned LD */
|
||
target_cci = &ct3d->vdm_fm_owned_ld_mctp_cci;
|
||
} else {
|
||
return CXL_MBOX_INVALID_INPUT;
|
||
}
|
||
} else {
|
||
return CXL_MBOX_INVALID_INPUT;
|
||
}
|
||
|
||
pl_length = in->ccimessage.pl_length[2] << 16 |
|
||
in->ccimessage.pl_length[1] << 8 | in->ccimessage.pl_length[0];
|
||
rc = cxl_process_cci_message(target_cci,
|
||
in->ccimessage.command_set,
|
||
in->ccimessage.command,
|
||
pl_length, in->ccimessage.payload,
|
||
&length_out, out->ccimessage.payload,
|
||
&bg_started);
|
||
/* Payload should be in place. Rest of CCI header and needs filling */
|
||
out->resp_len = length_out + sizeof(CXLCCIMessage);
|
||
st24_le_p(out->ccimessage.pl_length, length_out);
|
||
out->ccimessage.rc = rc;
|
||
out->ccimessage.category = CXL_CCI_CAT_RSP;
|
||
out->ccimessage.command = in->ccimessage.command;
|
||
out->ccimessage.command_set = in->ccimessage.command_set;
|
||
out->ccimessage.tag = in->ccimessage.tag;
|
||
*len_out = length_out + sizeof(*out);
|
||
|
||
return CXL_MBOX_SUCCESS;
|
||
}
|
||
|
||
static CXLRetCode cmd_events_get_records(const struct cxl_cmd *cmd,
|
||
uint8_t *payload_in, size_t len_in,
|
||
uint8_t *payload_out, size_t *len_out,
|
||
CXLCCI *cci)
|
||
{
|
||
CXLDeviceState *cxlds = &CXL_TYPE3(cci->d)->cxl_dstate;
|
||
CXLGetEventPayload *pl;
|
||
uint8_t log_type;
|
||
int max_recs;
|
||
|
||
if (cmd->in < sizeof(log_type)) {
|
||
return CXL_MBOX_INVALID_INPUT;
|
||
}
|
||
|
||
log_type = payload_in[0];
|
||
|
||
pl = (CXLGetEventPayload *)payload_out;
|
||
|
||
max_recs = (cxlds->payload_size - CXL_EVENT_PAYLOAD_HDR_SIZE) /
|
||
CXL_EVENT_RECORD_SIZE;
|
||
if (max_recs > 0xFFFF) {
|
||
max_recs = 0xFFFF;
|
||
}
|
||
|
||
return cxl_event_get_records(cxlds, pl, log_type, max_recs, len_out);
|
||
}
|
||
|
||
static CXLRetCode cmd_events_clear_records(const struct cxl_cmd *cmd,
|
||
uint8_t *payload_in,
|
||
size_t len_in,
|
||
uint8_t *payload_out,
|
||
size_t *len_out,
|
||
CXLCCI *cci)
|
||
{
|
||
CXLDeviceState *cxlds = &CXL_TYPE3(cci->d)->cxl_dstate;
|
||
CXLClearEventPayload *pl;
|
||
|
||
pl = (CXLClearEventPayload *)payload_in;
|
||
*len_out = 0;
|
||
return cxl_event_clear_records(cxlds, pl);
|
||
}
|
||
|
||
static CXLRetCode cmd_events_get_interrupt_policy(const struct cxl_cmd *cmd,
|
||
uint8_t *payload_in,
|
||
size_t len_in,
|
||
uint8_t *payload_out,
|
||
size_t *len_out,
|
||
CXLCCI *cci)
|
||
{
|
||
CXLDeviceState *cxlds = &CXL_TYPE3(cci->d)->cxl_dstate;
|
||
CXLEventInterruptPolicy *policy;
|
||
CXLEventLog *log;
|
||
|
||
policy = (CXLEventInterruptPolicy *)payload_out;
|
||
|
||
log = &cxlds->event_logs[CXL_EVENT_TYPE_INFO];
|
||
if (log->irq_enabled) {
|
||
policy->info_settings = CXL_EVENT_INT_SETTING(log->irq_vec);
|
||
}
|
||
|
||
log = &cxlds->event_logs[CXL_EVENT_TYPE_WARN];
|
||
if (log->irq_enabled) {
|
||
policy->warn_settings = CXL_EVENT_INT_SETTING(log->irq_vec);
|
||
}
|
||
|
||
log = &cxlds->event_logs[CXL_EVENT_TYPE_FAIL];
|
||
if (log->irq_enabled) {
|
||
policy->failure_settings = CXL_EVENT_INT_SETTING(log->irq_vec);
|
||
}
|
||
|
||
log = &cxlds->event_logs[CXL_EVENT_TYPE_FATAL];
|
||
if (log->irq_enabled) {
|
||
policy->fatal_settings = CXL_EVENT_INT_SETTING(log->irq_vec);
|
||
}
|
||
|
||
log = &cxlds->event_logs[CXL_EVENT_TYPE_DYNAMIC_CAP];
|
||
if (log->irq_enabled) {
|
||
/* Dynamic Capacity borrows the same vector as info */
|
||
policy->dyn_cap_settings = CXL_INT_MSI_MSIX;
|
||
}
|
||
|
||
*len_out = sizeof(*policy);
|
||
return CXL_MBOX_SUCCESS;
|
||
}
|
||
|
||
static CXLRetCode cmd_events_set_interrupt_policy(const struct cxl_cmd *cmd,
|
||
uint8_t *payload_in,
|
||
size_t len_in,
|
||
uint8_t *payload_out,
|
||
size_t *len_out,
|
||
CXLCCI *cci)
|
||
{
|
||
CXLDeviceState *cxlds = &CXL_TYPE3(cci->d)->cxl_dstate;
|
||
CXLEventInterruptPolicy *policy;
|
||
CXLEventLog *log;
|
||
|
||
if (len_in < CXL_EVENT_INT_SETTING_MIN_LEN) {
|
||
return CXL_MBOX_INVALID_PAYLOAD_LENGTH;
|
||
}
|
||
|
||
policy = (CXLEventInterruptPolicy *)payload_in;
|
||
|
||
log = &cxlds->event_logs[CXL_EVENT_TYPE_INFO];
|
||
log->irq_enabled = (policy->info_settings & CXL_EVENT_INT_MODE_MASK) ==
|
||
CXL_INT_MSI_MSIX;
|
||
|
||
log = &cxlds->event_logs[CXL_EVENT_TYPE_WARN];
|
||
log->irq_enabled = (policy->warn_settings & CXL_EVENT_INT_MODE_MASK) ==
|
||
CXL_INT_MSI_MSIX;
|
||
|
||
log = &cxlds->event_logs[CXL_EVENT_TYPE_FAIL];
|
||
log->irq_enabled = (policy->failure_settings & CXL_EVENT_INT_MODE_MASK) ==
|
||
CXL_INT_MSI_MSIX;
|
||
|
||
log = &cxlds->event_logs[CXL_EVENT_TYPE_FATAL];
|
||
log->irq_enabled = (policy->fatal_settings & CXL_EVENT_INT_MODE_MASK) ==
|
||
CXL_INT_MSI_MSIX;
|
||
|
||
/* DCD is optional */
|
||
if (len_in < sizeof(*policy)) {
|
||
return CXL_MBOX_SUCCESS;
|
||
}
|
||
|
||
log = &cxlds->event_logs[CXL_EVENT_TYPE_DYNAMIC_CAP];
|
||
log->irq_enabled = (policy->dyn_cap_settings & CXL_EVENT_INT_MODE_MASK) ==
|
||
CXL_INT_MSI_MSIX;
|
||
|
||
*len_out = 0;
|
||
return CXL_MBOX_SUCCESS;
|
||
}
|
||
|
||
/* CXL r3.1 section 8.2.9.1.1: Identify (Opcode 0001h) */
|
||
static CXLRetCode cmd_infostat_identify(const struct cxl_cmd *cmd,
|
||
uint8_t *payload_in,
|
||
size_t len_in,
|
||
uint8_t *payload_out,
|
||
size_t *len_out,
|
||
CXLCCI *cci)
|
||
{
|
||
PCIDeviceClass *class = PCI_DEVICE_GET_CLASS(cci->d);
|
||
struct {
|
||
uint16_t pcie_vid;
|
||
uint16_t pcie_did;
|
||
uint16_t pcie_subsys_vid;
|
||
uint16_t pcie_subsys_id;
|
||
uint64_t sn;
|
||
uint8_t max_message_size;
|
||
uint8_t component_type;
|
||
} QEMU_PACKED *is_identify;
|
||
QEMU_BUILD_BUG_ON(sizeof(*is_identify) != 18);
|
||
|
||
is_identify = (void *)payload_out;
|
||
is_identify->pcie_vid = class->vendor_id;
|
||
is_identify->pcie_did = class->device_id;
|
||
if (object_dynamic_cast(OBJECT(cci->d), TYPE_CXL_USP)) {
|
||
is_identify->sn = CXL_USP(cci->d)->sn;
|
||
/* Subsystem info not defined for a USP */
|
||
is_identify->pcie_subsys_vid = 0;
|
||
is_identify->pcie_subsys_id = 0;
|
||
is_identify->component_type = 0x0; /* Switch */
|
||
} else if (object_dynamic_cast(OBJECT(cci->d), TYPE_CXL_TYPE3)) {
|
||
PCIDevice *pci_dev = PCI_DEVICE(cci->d);
|
||
|
||
is_identify->sn = CXL_TYPE3(cci->d)->sn;
|
||
/*
|
||
* We can't always use class->subsystem_vendor_id as
|
||
* it is not set if the defaults are used.
|
||
*/
|
||
is_identify->pcie_subsys_vid =
|
||
pci_get_word(pci_dev->config + PCI_SUBSYSTEM_VENDOR_ID);
|
||
is_identify->pcie_subsys_id =
|
||
pci_get_word(pci_dev->config + PCI_SUBSYSTEM_ID);
|
||
is_identify->component_type = 0x3; /* Type 3 */
|
||
}
|
||
|
||
/* TODO: Allow this to vary across different CCIs */
|
||
is_identify->max_message_size = 9; /* 512 bytes - MCTP_CXL_MAILBOX_BYTES */
|
||
*len_out = sizeof(*is_identify);
|
||
return CXL_MBOX_SUCCESS;
|
||
}
|
||
|
||
static void cxl_set_dsp_active_bm(PCIBus *b, PCIDevice *d,
|
||
void *private)
|
||
{
|
||
uint8_t *bm = private;
|
||
if (object_dynamic_cast(OBJECT(d), TYPE_CXL_DSP)) {
|
||
uint8_t port = PCIE_PORT(d)->port;
|
||
bm[port / 8] |= 1 << (port % 8);
|
||
}
|
||
}
|
||
|
||
/* CXL r3.1 Section 7.6.7.1.1: Identify Switch Device (Opcode 5100h) */
|
||
static CXLRetCode cmd_identify_switch_device(const struct cxl_cmd *cmd,
|
||
uint8_t *payload_in,
|
||
size_t len_in,
|
||
uint8_t *payload_out,
|
||
size_t *len_out,
|
||
CXLCCI *cci)
|
||
{
|
||
PCIEPort *usp = PCIE_PORT(cci->d);
|
||
PCIBus *bus = &PCI_BRIDGE(cci->d)->sec_bus;
|
||
int num_phys_ports = pcie_count_ds_ports(bus);
|
||
|
||
struct cxl_fmapi_ident_switch_dev_resp_pl {
|
||
uint8_t ingress_port_id;
|
||
uint8_t rsvd;
|
||
uint8_t num_physical_ports;
|
||
uint8_t num_vcss;
|
||
uint8_t active_port_bitmask[0x20];
|
||
uint8_t active_vcs_bitmask[0x20];
|
||
uint16_t total_vppbs;
|
||
uint16_t bound_vppbs;
|
||
uint8_t num_hdm_decoders_per_usp;
|
||
} QEMU_PACKED *out;
|
||
QEMU_BUILD_BUG_ON(sizeof(*out) != 0x49);
|
||
|
||
out = (struct cxl_fmapi_ident_switch_dev_resp_pl *)payload_out;
|
||
*out = (struct cxl_fmapi_ident_switch_dev_resp_pl) {
|
||
.num_physical_ports = num_phys_ports + 1, /* 1 USP */
|
||
.num_vcss = 1, /* Not yet support multiple VCS - potentially tricky */
|
||
.active_vcs_bitmask[0] = 0x1,
|
||
.total_vppbs = num_phys_ports + 1,
|
||
.bound_vppbs = num_phys_ports + 1,
|
||
.num_hdm_decoders_per_usp = 4,
|
||
};
|
||
|
||
/* Depends on the CCI type */
|
||
if (object_dynamic_cast(OBJECT(cci->intf), TYPE_PCIE_PORT)) {
|
||
out->ingress_port_id = PCIE_PORT(cci->intf)->port;
|
||
} else {
|
||
/* MCTP? */
|
||
out->ingress_port_id = 0;
|
||
}
|
||
|
||
pci_for_each_device_under_bus(bus, cxl_set_dsp_active_bm,
|
||
out->active_port_bitmask);
|
||
out->active_port_bitmask[usp->port / 8] |= (1 << usp->port % 8);
|
||
|
||
*len_out = sizeof(*out);
|
||
|
||
return CXL_MBOX_SUCCESS;
|
||
}
|
||
|
||
/* CXL r3.1 Section 7.6.7.1.2: Get Physical Port State (Opcode 5101h) */
|
||
static CXLRetCode cmd_get_physical_port_state(const struct cxl_cmd *cmd,
|
||
uint8_t *payload_in,
|
||
size_t len_in,
|
||
uint8_t *payload_out,
|
||
size_t *len_out,
|
||
CXLCCI *cci)
|
||
{
|
||
/* CXL r3.1 Table 7-17: Get Physical Port State Request Payload */
|
||
struct cxl_fmapi_get_phys_port_state_req_pl {
|
||
uint8_t num_ports;
|
||
uint8_t ports[];
|
||
} QEMU_PACKED *in;
|
||
|
||
/*
|
||
* CXL r3.1 Table 7-19: Get Physical Port State Port Information Block
|
||
* Format
|
||
*/
|
||
struct cxl_fmapi_port_state_info_block {
|
||
uint8_t port_id;
|
||
uint8_t config_state;
|
||
uint8_t connected_device_cxl_version;
|
||
uint8_t rsv1;
|
||
uint8_t connected_device_type;
|
||
uint8_t port_cxl_version_bitmask;
|
||
uint8_t max_link_width;
|
||
uint8_t negotiated_link_width;
|
||
uint8_t supported_link_speeds_vector;
|
||
uint8_t max_link_speed;
|
||
uint8_t current_link_speed;
|
||
uint8_t ltssm_state;
|
||
uint8_t first_lane_num;
|
||
uint16_t link_state;
|
||
uint8_t supported_ld_count;
|
||
} QEMU_PACKED;
|
||
|
||
/* CXL r3.1 Table 7-18: Get Physical Port State Response Payload */
|
||
struct cxl_fmapi_get_phys_port_state_resp_pl {
|
||
uint8_t num_ports;
|
||
uint8_t rsv1[3];
|
||
struct cxl_fmapi_port_state_info_block ports[];
|
||
} QEMU_PACKED *out;
|
||
PCIBus *bus = &PCI_BRIDGE(cci->d)->sec_bus;
|
||
PCIEPort *usp = PCIE_PORT(cci->d);
|
||
size_t pl_size;
|
||
int i;
|
||
|
||
in = (struct cxl_fmapi_get_phys_port_state_req_pl *)payload_in;
|
||
out = (struct cxl_fmapi_get_phys_port_state_resp_pl *)payload_out;
|
||
|
||
/* Check if what was requested can fit */
|
||
if (sizeof(*out) + sizeof(*out->ports) * in->num_ports > cci->payload_max) {
|
||
return CXL_MBOX_INVALID_INPUT;
|
||
}
|
||
|
||
/* For success there should be a match for each requested */
|
||
out->num_ports = in->num_ports;
|
||
|
||
for (i = 0; i < in->num_ports; i++) {
|
||
struct cxl_fmapi_port_state_info_block *port;
|
||
/* First try to match on downstream port */
|
||
PCIDevice *port_dev;
|
||
uint16_t lnkcap, lnkcap2, lnksta;
|
||
|
||
port = &out->ports[i];
|
||
|
||
port_dev = pcie_find_port_by_pn(bus, in->ports[i]);
|
||
if (port_dev) { /* DSP */
|
||
PCIDevice *ds_dev = pci_bridge_get_sec_bus(PCI_BRIDGE(port_dev))
|
||
->devices[0];
|
||
port->config_state = 3;
|
||
if (ds_dev) {
|
||
if (object_dynamic_cast(OBJECT(ds_dev), TYPE_CXL_TYPE3)) {
|
||
port->connected_device_type = 5; /* Assume MLD for now */
|
||
} else {
|
||
port->connected_device_type = 1;
|
||
}
|
||
} else {
|
||
port->connected_device_type = 0;
|
||
}
|
||
port->supported_ld_count = 3;
|
||
} else if (usp->port == in->ports[i]) { /* USP */
|
||
port_dev = PCI_DEVICE(usp);
|
||
port->config_state = 4;
|
||
port->connected_device_type = 0;
|
||
} else {
|
||
return CXL_MBOX_INVALID_INPUT;
|
||
}
|
||
|
||
port->port_id = in->ports[i];
|
||
/* Information on status of this port in lnksta, lnkcap */
|
||
if (!port_dev->exp.exp_cap) {
|
||
return CXL_MBOX_INTERNAL_ERROR;
|
||
}
|
||
lnksta = port_dev->config_read(port_dev,
|
||
port_dev->exp.exp_cap + PCI_EXP_LNKSTA,
|
||
sizeof(lnksta));
|
||
lnkcap = port_dev->config_read(port_dev,
|
||
port_dev->exp.exp_cap + PCI_EXP_LNKCAP,
|
||
sizeof(lnkcap));
|
||
lnkcap2 = port_dev->config_read(port_dev,
|
||
port_dev->exp.exp_cap + PCI_EXP_LNKCAP2,
|
||
sizeof(lnkcap2));
|
||
|
||
port->max_link_width = (lnkcap & PCI_EXP_LNKCAP_MLW) >> 4;
|
||
port->negotiated_link_width = (lnksta & PCI_EXP_LNKSTA_NLW) >> 4;
|
||
/* No definition for SLS field in linux/pci_regs.h */
|
||
port->supported_link_speeds_vector = (lnkcap2 & 0xFE) >> 1;
|
||
port->max_link_speed = lnkcap & PCI_EXP_LNKCAP_SLS;
|
||
port->current_link_speed = lnksta & PCI_EXP_LNKSTA_CLS;
|
||
/* TODO: Track down if we can get the rest of the info */
|
||
port->ltssm_state = 0x7;
|
||
port->first_lane_num = 0;
|
||
port->link_state = 0;
|
||
port->port_cxl_version_bitmask = 0x2;
|
||
port->connected_device_cxl_version = 0x2;
|
||
}
|
||
|
||
pl_size = sizeof(*out) + sizeof(*out->ports) * in->num_ports;
|
||
*len_out = pl_size;
|
||
|
||
return CXL_MBOX_SUCCESS;
|
||
}
|
||
|
||
/* CXL r3.1 Section 8.2.9.1.2: Background Operation Status (Opcode 0002h) */
|
||
static CXLRetCode cmd_infostat_bg_op_sts(const struct cxl_cmd *cmd,
|
||
uint8_t *payload_in,
|
||
size_t len_in,
|
||
uint8_t *payload_out,
|
||
size_t *len_out,
|
||
CXLCCI *cci)
|
||
{
|
||
struct {
|
||
uint8_t status;
|
||
uint8_t rsvd;
|
||
uint16_t opcode;
|
||
uint16_t returncode;
|
||
uint16_t vendor_ext_status;
|
||
} QEMU_PACKED *bg_op_status;
|
||
QEMU_BUILD_BUG_ON(sizeof(*bg_op_status) != 8);
|
||
|
||
bg_op_status = (void *)payload_out;
|
||
bg_op_status->status = cci->bg.complete_pct << 1;
|
||
if (cci->bg.runtime > 0) {
|
||
bg_op_status->status |= 1U << 0;
|
||
}
|
||
bg_op_status->opcode = cci->bg.opcode;
|
||
bg_op_status->returncode = cci->bg.ret_code;
|
||
*len_out = sizeof(*bg_op_status);
|
||
|
||
return CXL_MBOX_SUCCESS;
|
||
}
|
||
|
||
#define CXL_FW_SLOTS 2
|
||
#define CXL_FW_SIZE 0x02000000 /* 32 mb */
|
||
|
||
/* CXL r3.1 Section 8.2.9.3.1: Get FW Info (Opcode 0200h) */
|
||
static CXLRetCode cmd_firmware_update_get_info(const struct cxl_cmd *cmd,
|
||
uint8_t *payload_in,
|
||
size_t len,
|
||
uint8_t *payload_out,
|
||
size_t *len_out,
|
||
CXLCCI *cci)
|
||
{
|
||
CXLType3Dev *ct3d = CXL_TYPE3(cci->d);
|
||
CXLDeviceState *cxl_dstate = &ct3d->cxl_dstate;
|
||
struct {
|
||
uint8_t slots_supported;
|
||
uint8_t slot_info;
|
||
uint8_t caps;
|
||
uint8_t rsvd[0xd];
|
||
char fw_rev1[0x10];
|
||
char fw_rev2[0x10];
|
||
char fw_rev3[0x10];
|
||
char fw_rev4[0x10];
|
||
} QEMU_PACKED *fw_info;
|
||
QEMU_BUILD_BUG_ON(sizeof(*fw_info) != 0x50);
|
||
|
||
if ((cxl_dstate->vmem_size < CXL_CAPACITY_MULTIPLIER) ||
|
||
(cxl_dstate->pmem_size < CXL_CAPACITY_MULTIPLIER) ||
|
||
(ct3d->dc.total_capacity < CXL_CAPACITY_MULTIPLIER)) {
|
||
return CXL_MBOX_INTERNAL_ERROR;
|
||
}
|
||
|
||
fw_info = (void *)payload_out;
|
||
|
||
fw_info->slots_supported = CXL_FW_SLOTS;
|
||
fw_info->slot_info = (cci->fw.active_slot & 0x7) |
|
||
((cci->fw.staged_slot & 0x7) << 3);
|
||
fw_info->caps = BIT(0); /* online update supported */
|
||
|
||
if (cci->fw.slot[0]) {
|
||
pstrcpy(fw_info->fw_rev1, sizeof(fw_info->fw_rev1), "BWFW VERSION 0");
|
||
}
|
||
if (cci->fw.slot[1]) {
|
||
pstrcpy(fw_info->fw_rev2, sizeof(fw_info->fw_rev2), "BWFW VERSION 1");
|
||
}
|
||
|
||
*len_out = sizeof(*fw_info);
|
||
return CXL_MBOX_SUCCESS;
|
||
}
|
||
|
||
/* CXL r3.1 section 8.2.9.3.2: Transfer FW (Opcode 0201h) */
|
||
#define CXL_FW_XFER_ALIGNMENT 128
|
||
|
||
#define CXL_FW_XFER_ACTION_FULL 0x0
|
||
#define CXL_FW_XFER_ACTION_INIT 0x1
|
||
#define CXL_FW_XFER_ACTION_CONTINUE 0x2
|
||
#define CXL_FW_XFER_ACTION_END 0x3
|
||
#define CXL_FW_XFER_ACTION_ABORT 0x4
|
||
|
||
static CXLRetCode cmd_firmware_update_transfer(const struct cxl_cmd *cmd,
|
||
uint8_t *payload_in,
|
||
size_t len,
|
||
uint8_t *payload_out,
|
||
size_t *len_out,
|
||
CXLCCI *cci)
|
||
{
|
||
struct {
|
||
uint8_t action;
|
||
uint8_t slot;
|
||
uint8_t rsvd1[2];
|
||
uint32_t offset;
|
||
uint8_t rsvd2[0x78];
|
||
uint8_t data[];
|
||
} QEMU_PACKED *fw_transfer = (void *)payload_in;
|
||
size_t offset, length;
|
||
|
||
if (fw_transfer->action == CXL_FW_XFER_ACTION_ABORT) {
|
||
/*
|
||
* At this point there aren't any on-going transfers
|
||
* running in the bg - this is serialized before this
|
||
* call altogether. Just mark the state machine and
|
||
* disregard any other input.
|
||
*/
|
||
cci->fw.transferring = false;
|
||
return CXL_MBOX_SUCCESS;
|
||
}
|
||
|
||
offset = fw_transfer->offset * CXL_FW_XFER_ALIGNMENT;
|
||
length = len - sizeof(*fw_transfer);
|
||
if (offset + length > CXL_FW_SIZE) {
|
||
return CXL_MBOX_INVALID_INPUT;
|
||
}
|
||
|
||
if (cci->fw.transferring) {
|
||
if (fw_transfer->action == CXL_FW_XFER_ACTION_FULL ||
|
||
fw_transfer->action == CXL_FW_XFER_ACTION_INIT) {
|
||
return CXL_MBOX_FW_XFER_IN_PROGRESS;
|
||
}
|
||
/*
|
||
* Abort partitioned package transfer if over 30 secs
|
||
* between parts. As opposed to the explicit ABORT action,
|
||
* semantically treat this condition as an error - as
|
||
* if a part action were passed without a previous INIT.
|
||
*/
|
||
if (difftime(time(NULL), cci->fw.last_partxfer) > 30.0) {
|
||
cci->fw.transferring = false;
|
||
return CXL_MBOX_INVALID_INPUT;
|
||
}
|
||
} else if (fw_transfer->action == CXL_FW_XFER_ACTION_CONTINUE ||
|
||
fw_transfer->action == CXL_FW_XFER_ACTION_END) {
|
||
return CXL_MBOX_INVALID_INPUT;
|
||
}
|
||
|
||
/* allow back-to-back retransmission */
|
||
if ((offset != cci->fw.prev_offset || length != cci->fw.prev_len) &&
|
||
(fw_transfer->action == CXL_FW_XFER_ACTION_CONTINUE ||
|
||
fw_transfer->action == CXL_FW_XFER_ACTION_END)) {
|
||
/* verify no overlaps */
|
||
if (offset < cci->fw.prev_offset + cci->fw.prev_len) {
|
||
return CXL_MBOX_FW_XFER_OUT_OF_ORDER;
|
||
}
|
||
}
|
||
|
||
switch (fw_transfer->action) {
|
||
case CXL_FW_XFER_ACTION_FULL: /* ignores offset */
|
||
case CXL_FW_XFER_ACTION_END:
|
||
if (fw_transfer->slot == 0 ||
|
||
fw_transfer->slot == cci->fw.active_slot ||
|
||
fw_transfer->slot > CXL_FW_SLOTS) {
|
||
return CXL_MBOX_FW_INVALID_SLOT;
|
||
}
|
||
|
||
/* mark the slot used upon bg completion */
|
||
break;
|
||
case CXL_FW_XFER_ACTION_INIT:
|
||
if (offset != 0) {
|
||
return CXL_MBOX_INVALID_INPUT;
|
||
}
|
||
|
||
cci->fw.transferring = true;
|
||
cci->fw.prev_offset = offset;
|
||
cci->fw.prev_len = length;
|
||
break;
|
||
case CXL_FW_XFER_ACTION_CONTINUE:
|
||
cci->fw.prev_offset = offset;
|
||
cci->fw.prev_len = length;
|
||
break;
|
||
default:
|
||
return CXL_MBOX_INVALID_INPUT;
|
||
}
|
||
|
||
if (fw_transfer->action == CXL_FW_XFER_ACTION_FULL) {
|
||
cci->bg.runtime = 10 * 1000UL;
|
||
} else {
|
||
cci->bg.runtime = 2 * 1000UL;
|
||
}
|
||
/* keep relevant context for bg completion */
|
||
cci->fw.curr_action = fw_transfer->action;
|
||
cci->fw.curr_slot = fw_transfer->slot;
|
||
*len_out = 0;
|
||
|
||
return CXL_MBOX_BG_STARTED;
|
||
}
|
||
|
||
static void __do_firmware_xfer(CXLCCI *cci)
|
||
{
|
||
switch (cci->fw.curr_action) {
|
||
case CXL_FW_XFER_ACTION_FULL:
|
||
case CXL_FW_XFER_ACTION_END:
|
||
cci->fw.slot[cci->fw.curr_slot - 1] = true;
|
||
cci->fw.transferring = false;
|
||
break;
|
||
case CXL_FW_XFER_ACTION_INIT:
|
||
case CXL_FW_XFER_ACTION_CONTINUE:
|
||
time(&cci->fw.last_partxfer);
|
||
break;
|
||
default:
|
||
break;
|
||
}
|
||
}
|
||
|
||
/* CXL r3.1 section 8.2.9.3.3: Activate FW (Opcode 0202h) */
|
||
static CXLRetCode cmd_firmware_update_activate(const struct cxl_cmd *cmd,
|
||
uint8_t *payload_in,
|
||
size_t len,
|
||
uint8_t *payload_out,
|
||
size_t *len_out,
|
||
CXLCCI *cci)
|
||
{
|
||
struct {
|
||
uint8_t action;
|
||
uint8_t slot;
|
||
} QEMU_PACKED *fw_activate = (void *)payload_in;
|
||
QEMU_BUILD_BUG_ON(sizeof(*fw_activate) != 0x2);
|
||
|
||
if (fw_activate->slot == 0 ||
|
||
fw_activate->slot == cci->fw.active_slot ||
|
||
fw_activate->slot > CXL_FW_SLOTS) {
|
||
return CXL_MBOX_FW_INVALID_SLOT;
|
||
}
|
||
|
||
/* ensure that an actual fw package is there */
|
||
if (!cci->fw.slot[fw_activate->slot - 1]) {
|
||
return CXL_MBOX_FW_INVALID_SLOT;
|
||
}
|
||
|
||
switch (fw_activate->action) {
|
||
case 0: /* online */
|
||
cci->fw.active_slot = fw_activate->slot;
|
||
break;
|
||
case 1: /* reset */
|
||
cci->fw.staged_slot = fw_activate->slot;
|
||
break;
|
||
default:
|
||
return CXL_MBOX_INVALID_INPUT;
|
||
}
|
||
|
||
return CXL_MBOX_SUCCESS;
|
||
}
|
||
|
||
/* CXL r3.1 Section 8.2.9.4.1: Get Timestamp (Opcode 0300h) */
|
||
static CXLRetCode cmd_timestamp_get(const struct cxl_cmd *cmd,
|
||
uint8_t *payload_in,
|
||
size_t len_in,
|
||
uint8_t *payload_out,
|
||
size_t *len_out,
|
||
CXLCCI *cci)
|
||
{
|
||
CXLDeviceState *cxl_dstate = &CXL_TYPE3(cci->d)->cxl_dstate;
|
||
uint64_t final_time = cxl_device_get_timestamp(cxl_dstate);
|
||
|
||
stq_le_p(payload_out, final_time);
|
||
*len_out = 8;
|
||
|
||
return CXL_MBOX_SUCCESS;
|
||
}
|
||
|
||
/* CXL r3.1 Section 8.2.9.4.2: Set Timestamp (Opcode 0301h) */
|
||
static CXLRetCode cmd_timestamp_set(const struct cxl_cmd *cmd,
|
||
uint8_t *payload_in,
|
||
size_t len_in,
|
||
uint8_t *payload_out,
|
||
size_t *len_out,
|
||
CXLCCI *cci)
|
||
{
|
||
CXLDeviceState *cxl_dstate = &CXL_TYPE3(cci->d)->cxl_dstate;
|
||
|
||
cxl_dstate->timestamp.set = true;
|
||
cxl_dstate->timestamp.last_set = qemu_clock_get_ns(QEMU_CLOCK_VIRTUAL);
|
||
|
||
cxl_dstate->timestamp.host_set = le64_to_cpu(*(uint64_t *)payload_in);
|
||
|
||
*len_out = 0;
|
||
return CXL_MBOX_SUCCESS;
|
||
}
|
||
|
||
/* CXL r3.1 Section 8.2.9.5.2.1: Command Effects Log (CEL) */
|
||
static const QemuUUID cel_uuid = {
|
||
.data = UUID(0x0da9c0b5, 0xbf41, 0x4b78, 0x8f, 0x79,
|
||
0x96, 0xb1, 0x62, 0x3b, 0x3f, 0x17)
|
||
};
|
||
|
||
/* CXL r3.1 Section 8.2.9.5.1: Get Supported Logs (Opcode 0400h) */
|
||
static CXLRetCode cmd_logs_get_supported(const struct cxl_cmd *cmd,
|
||
uint8_t *payload_in,
|
||
size_t len_in,
|
||
uint8_t *payload_out,
|
||
size_t *len_out,
|
||
CXLCCI *cci)
|
||
{
|
||
struct {
|
||
uint16_t entries;
|
||
uint8_t rsvd[6];
|
||
struct {
|
||
QemuUUID uuid;
|
||
uint32_t size;
|
||
} log_entries[1];
|
||
} QEMU_PACKED *supported_logs = (void *)payload_out;
|
||
QEMU_BUILD_BUG_ON(sizeof(*supported_logs) != 0x1c);
|
||
|
||
supported_logs->entries = 1;
|
||
supported_logs->log_entries[0].uuid = cel_uuid;
|
||
supported_logs->log_entries[0].size = 4 * cci->cel_size;
|
||
|
||
*len_out = sizeof(*supported_logs);
|
||
return CXL_MBOX_SUCCESS;
|
||
}
|
||
|
||
/* CXL r3.1 Section 8.2.9.5.2: Get Log (Opcode 0401h) */
|
||
static CXLRetCode cmd_logs_get_log(const struct cxl_cmd *cmd,
|
||
uint8_t *payload_in,
|
||
size_t len_in,
|
||
uint8_t *payload_out,
|
||
size_t *len_out,
|
||
CXLCCI *cci)
|
||
{
|
||
struct {
|
||
QemuUUID uuid;
|
||
uint32_t offset;
|
||
uint32_t length;
|
||
} QEMU_PACKED QEMU_ALIGNED(16) *get_log;
|
||
|
||
get_log = (void *)payload_in;
|
||
|
||
/*
|
||
* CXL r3.1 Section 8.2.9.5.2: Get Log (Opcode 0401h)
|
||
* The device shall return Invalid Input if the Offset or Length
|
||
* fields attempt to access beyond the size of the log as reported by Get
|
||
* Supported Logs.
|
||
*
|
||
* The CEL buffer is large enough to fit all commands in the emulation, so
|
||
* the only possible failure would be if the mailbox itself isn't big
|
||
* enough.
|
||
*/
|
||
if (get_log->offset + get_log->length > cci->payload_max) {
|
||
return CXL_MBOX_INVALID_INPUT;
|
||
}
|
||
|
||
if (!qemu_uuid_is_equal(&get_log->uuid, &cel_uuid)) {
|
||
return CXL_MBOX_INVALID_LOG;
|
||
}
|
||
|
||
/* Store off everything to local variables so we can wipe out the payload */
|
||
*len_out = get_log->length;
|
||
|
||
memmove(payload_out, cci->cel_log + get_log->offset, get_log->length);
|
||
|
||
return CXL_MBOX_SUCCESS;
|
||
}
|
||
|
||
/* CXL r3.1 section 8.2.9.6: Features */
|
||
/*
|
||
* Get Supported Features output payload
|
||
* CXL r3.1 section 8.2.9.6.1 Table 8-96
|
||
*/
|
||
typedef struct CXLSupportedFeatureHeader {
|
||
uint16_t entries;
|
||
uint16_t nsuppfeats_dev;
|
||
uint32_t reserved;
|
||
} QEMU_PACKED CXLSupportedFeatureHeader;
|
||
|
||
/*
|
||
* Get Supported Features Supported Feature Entry
|
||
* CXL r3.1 section 8.2.9.6.1 Table 8-97
|
||
*/
|
||
typedef struct CXLSupportedFeatureEntry {
|
||
QemuUUID uuid;
|
||
uint16_t feat_index;
|
||
uint16_t get_feat_size;
|
||
uint16_t set_feat_size;
|
||
uint32_t attr_flags;
|
||
uint8_t get_feat_version;
|
||
uint8_t set_feat_version;
|
||
uint16_t set_feat_effects;
|
||
uint8_t rsvd[18];
|
||
} QEMU_PACKED CXLSupportedFeatureEntry;
|
||
|
||
/*
|
||
* Get Supported Features Supported Feature Entry
|
||
* CXL rev 3.1 section 8.2.9.6.1 Table 8-97
|
||
*/
|
||
/* Supported Feature Entry : attribute flags */
|
||
#define CXL_FEAT_ENTRY_ATTR_FLAG_CHANGABLE BIT(0)
|
||
#define CXL_FEAT_ENTRY_ATTR_FLAG_DEEPEST_RESET_PERSISTENCE_MASK GENMASK(3, 1)
|
||
#define CXL_FEAT_ENTRY_ATTR_FLAG_PERSIST_ACROSS_FIRMWARE_UPDATE BIT(4)
|
||
#define CXL_FEAT_ENTRY_ATTR_FLAG_SUPPORT_DEFAULT_SELECTION BIT(5)
|
||
#define CXL_FEAT_ENTRY_ATTR_FLAG_SUPPORT_SAVED_SELECTION BIT(6)
|
||
|
||
/* Supported Feature Entry : set feature effects */
|
||
#define CXL_FEAT_ENTRY_SFE_CONFIG_CHANGE_COLD_RESET BIT(0)
|
||
#define CXL_FEAT_ENTRY_SFE_IMMEDIATE_CONFIG_CHANGE BIT(1)
|
||
#define CXL_FEAT_ENTRY_SFE_IMMEDIATE_DATA_CHANGE BIT(2)
|
||
#define CXL_FEAT_ENTRY_SFE_IMMEDIATE_POLICY_CHANGE BIT(3)
|
||
#define CXL_FEAT_ENTRY_SFE_IMMEDIATE_LOG_CHANGE BIT(4)
|
||
#define CXL_FEAT_ENTRY_SFE_SECURITY_STATE_CHANGE BIT(5)
|
||
#define CXL_FEAT_ENTRY_SFE_BACKGROUND_OPERATION BIT(6)
|
||
#define CXL_FEAT_ENTRY_SFE_SUPPORT_SECONDARY_MAILBOX BIT(7)
|
||
#define CXL_FEAT_ENTRY_SFE_SUPPORT_ABORT_BACKGROUND_OPERATION BIT(8)
|
||
#define CXL_FEAT_ENTRY_SFE_CEL_VALID BIT(9)
|
||
#define CXL_FEAT_ENTRY_SFE_CONFIG_CHANGE_CONV_RESET BIT(10)
|
||
#define CXL_FEAT_ENTRY_SFE_CONFIG_CHANGE_CXL_RESET BIT(11)
|
||
|
||
enum CXL_SUPPORTED_FEATURES_LIST {
|
||
CXL_FEATURE_PATROL_SCRUB = 0,
|
||
CXL_FEATURE_ECS,
|
||
CXL_FEATURE_MAX
|
||
};
|
||
|
||
/* Get Feature CXL 3.1 Spec 8.2.9.6.2 */
|
||
/*
|
||
* Get Feature input payload
|
||
* CXL r3.1 section 8.2.9.6.2 Table 8-99
|
||
*/
|
||
/* Get Feature : Payload in selection */
|
||
enum CXL_GET_FEATURE_SELECTION {
|
||
CXL_GET_FEATURE_SEL_CURRENT_VALUE,
|
||
CXL_GET_FEATURE_SEL_DEFAULT_VALUE,
|
||
CXL_GET_FEATURE_SEL_SAVED_VALUE,
|
||
CXL_GET_FEATURE_SEL_MAX
|
||
};
|
||
|
||
/* Set Feature CXL 3.1 Spec 8.2.9.6.3 */
|
||
/*
|
||
* Set Feature input payload
|
||
* CXL r3.1 section 8.2.9.6.3 Table 8-101
|
||
*/
|
||
typedef struct CXLSetFeatureInHeader {
|
||
QemuUUID uuid;
|
||
uint32_t flags;
|
||
uint16_t offset;
|
||
uint8_t version;
|
||
uint8_t rsvd[9];
|
||
} QEMU_PACKED QEMU_ALIGNED(16) CXLSetFeatureInHeader;
|
||
|
||
/* Set Feature : Payload in flags */
|
||
#define CXL_SET_FEATURE_FLAG_DATA_TRANSFER_MASK 0x7
|
||
enum CXL_SET_FEATURE_FLAG_DATA_TRANSFER {
|
||
CXL_SET_FEATURE_FLAG_FULL_DATA_TRANSFER,
|
||
CXL_SET_FEATURE_FLAG_INITIATE_DATA_TRANSFER,
|
||
CXL_SET_FEATURE_FLAG_CONTINUE_DATA_TRANSFER,
|
||
CXL_SET_FEATURE_FLAG_FINISH_DATA_TRANSFER,
|
||
CXL_SET_FEATURE_FLAG_ABORT_DATA_TRANSFER,
|
||
CXL_SET_FEATURE_FLAG_DATA_TRANSFER_MAX
|
||
};
|
||
#define CXL_SET_FEAT_DATA_SAVED_ACROSS_RESET BIT(3)
|
||
|
||
/* CXL r3.1 section 8.2.9.9.11.1: Device Patrol Scrub Control Feature */
|
||
static const QemuUUID patrol_scrub_uuid = {
|
||
.data = UUID(0x96dad7d6, 0xfde8, 0x482b, 0xa7, 0x33,
|
||
0x75, 0x77, 0x4e, 0x06, 0xdb, 0x8a)
|
||
};
|
||
|
||
typedef struct CXLMemPatrolScrubSetFeature {
|
||
CXLSetFeatureInHeader hdr;
|
||
CXLMemPatrolScrubWriteAttrs feat_data;
|
||
} QEMU_PACKED QEMU_ALIGNED(16) CXLMemPatrolScrubSetFeature;
|
||
|
||
/*
|
||
* CXL r3.1 section 8.2.9.9.11.2:
|
||
* DDR5 Error Check Scrub (ECS) Control Feature
|
||
*/
|
||
static const QemuUUID ecs_uuid = {
|
||
.data = UUID(0xe5b13f22, 0x2328, 0x4a14, 0xb8, 0xba,
|
||
0xb9, 0x69, 0x1e, 0x89, 0x33, 0x86)
|
||
};
|
||
|
||
typedef struct CXLMemECSSetFeature {
|
||
CXLSetFeatureInHeader hdr;
|
||
CXLMemECSWriteAttrs feat_data[];
|
||
} QEMU_PACKED QEMU_ALIGNED(16) CXLMemECSSetFeature;
|
||
|
||
/* CXL r3.1 section 8.2.9.6.1: Get Supported Features (Opcode 0500h) */
|
||
static CXLRetCode cmd_features_get_supported(const struct cxl_cmd *cmd,
|
||
uint8_t *payload_in,
|
||
size_t len_in,
|
||
uint8_t *payload_out,
|
||
size_t *len_out,
|
||
CXLCCI *cci)
|
||
{
|
||
struct {
|
||
uint32_t count;
|
||
uint16_t start_index;
|
||
uint16_t reserved;
|
||
} QEMU_PACKED QEMU_ALIGNED(16) * get_feats_in = (void *)payload_in;
|
||
|
||
struct {
|
||
CXLSupportedFeatureHeader hdr;
|
||
CXLSupportedFeatureEntry feat_entries[];
|
||
} QEMU_PACKED QEMU_ALIGNED(16) * get_feats_out = (void *)payload_out;
|
||
uint16_t index, req_entries;
|
||
uint16_t entry;
|
||
|
||
if (!object_dynamic_cast(OBJECT(cci->d), TYPE_CXL_TYPE3)) {
|
||
return CXL_MBOX_UNSUPPORTED;
|
||
}
|
||
if (get_feats_in->count < sizeof(CXLSupportedFeatureHeader) ||
|
||
get_feats_in->start_index >= CXL_FEATURE_MAX) {
|
||
return CXL_MBOX_INVALID_INPUT;
|
||
}
|
||
|
||
req_entries = (get_feats_in->count -
|
||
sizeof(CXLSupportedFeatureHeader)) /
|
||
sizeof(CXLSupportedFeatureEntry);
|
||
req_entries = MIN(req_entries,
|
||
(CXL_FEATURE_MAX - get_feats_in->start_index));
|
||
|
||
for (entry = 0, index = get_feats_in->start_index;
|
||
entry < req_entries; index++) {
|
||
switch (index) {
|
||
case CXL_FEATURE_PATROL_SCRUB:
|
||
/* Fill supported feature entry for device patrol scrub control */
|
||
get_feats_out->feat_entries[entry++] =
|
||
(struct CXLSupportedFeatureEntry) {
|
||
.uuid = patrol_scrub_uuid,
|
||
.feat_index = index,
|
||
.get_feat_size = sizeof(CXLMemPatrolScrubReadAttrs),
|
||
.set_feat_size = sizeof(CXLMemPatrolScrubWriteAttrs),
|
||
.attr_flags = CXL_FEAT_ENTRY_ATTR_FLAG_CHANGABLE,
|
||
.get_feat_version = CXL_MEMDEV_PS_GET_FEATURE_VERSION,
|
||
.set_feat_version = CXL_MEMDEV_PS_SET_FEATURE_VERSION,
|
||
.set_feat_effects = CXL_FEAT_ENTRY_SFE_IMMEDIATE_CONFIG_CHANGE |
|
||
CXL_FEAT_ENTRY_SFE_CEL_VALID,
|
||
};
|
||
break;
|
||
case CXL_FEATURE_ECS:
|
||
/* Fill supported feature entry for device DDR5 ECS control */
|
||
get_feats_out->feat_entries[entry++] =
|
||
(struct CXLSupportedFeatureEntry) {
|
||
.uuid = ecs_uuid,
|
||
.feat_index = index,
|
||
.get_feat_size = CXL_ECS_NUM_MEDIA_FRUS *
|
||
sizeof(CXLMemECSReadAttrs),
|
||
.set_feat_size = CXL_ECS_NUM_MEDIA_FRUS *
|
||
sizeof(CXLMemECSWriteAttrs),
|
||
.attr_flags = CXL_FEAT_ENTRY_ATTR_FLAG_CHANGABLE,
|
||
.get_feat_version = CXL_ECS_GET_FEATURE_VERSION,
|
||
.set_feat_version = CXL_ECS_SET_FEATURE_VERSION,
|
||
.set_feat_effects = CXL_FEAT_ENTRY_SFE_IMMEDIATE_CONFIG_CHANGE |
|
||
CXL_FEAT_ENTRY_SFE_CEL_VALID,
|
||
};
|
||
break;
|
||
default:
|
||
__builtin_unreachable();
|
||
}
|
||
}
|
||
get_feats_out->hdr.nsuppfeats_dev = CXL_FEATURE_MAX;
|
||
get_feats_out->hdr.entries = req_entries;
|
||
*len_out = sizeof(CXLSupportedFeatureHeader) +
|
||
req_entries * sizeof(CXLSupportedFeatureEntry);
|
||
|
||
return CXL_MBOX_SUCCESS;
|
||
}
|
||
|
||
/* CXL r3.1 section 8.2.9.6.2: Get Feature (Opcode 0501h) */
|
||
static CXLRetCode cmd_features_get_feature(const struct cxl_cmd *cmd,
|
||
uint8_t *payload_in,
|
||
size_t len_in,
|
||
uint8_t *payload_out,
|
||
size_t *len_out,
|
||
CXLCCI *cci)
|
||
{
|
||
struct {
|
||
QemuUUID uuid;
|
||
uint16_t offset;
|
||
uint16_t count;
|
||
uint8_t selection;
|
||
} QEMU_PACKED QEMU_ALIGNED(16) * get_feature;
|
||
uint16_t bytes_to_copy = 0;
|
||
CXLType3Dev *ct3d;
|
||
CXLSetFeatureInfo *set_feat_info;
|
||
|
||
if (!object_dynamic_cast(OBJECT(cci->d), TYPE_CXL_TYPE3)) {
|
||
return CXL_MBOX_UNSUPPORTED;
|
||
}
|
||
|
||
ct3d = CXL_TYPE3(cci->d);
|
||
get_feature = (void *)payload_in;
|
||
|
||
set_feat_info = &ct3d->set_feat_info;
|
||
if (qemu_uuid_is_equal(&get_feature->uuid, &set_feat_info->uuid)) {
|
||
return CXL_MBOX_FEATURE_TRANSFER_IN_PROGRESS;
|
||
}
|
||
|
||
if (get_feature->selection != CXL_GET_FEATURE_SEL_CURRENT_VALUE) {
|
||
return CXL_MBOX_UNSUPPORTED;
|
||
}
|
||
if (get_feature->offset + get_feature->count > cci->payload_max) {
|
||
return CXL_MBOX_INVALID_INPUT;
|
||
}
|
||
|
||
if (qemu_uuid_is_equal(&get_feature->uuid, &patrol_scrub_uuid)) {
|
||
if (get_feature->offset >= sizeof(CXLMemPatrolScrubReadAttrs)) {
|
||
return CXL_MBOX_INVALID_INPUT;
|
||
}
|
||
bytes_to_copy = sizeof(CXLMemPatrolScrubReadAttrs) -
|
||
get_feature->offset;
|
||
bytes_to_copy = MIN(bytes_to_copy, get_feature->count);
|
||
memcpy(payload_out,
|
||
(uint8_t *)&ct3d->patrol_scrub_attrs + get_feature->offset,
|
||
bytes_to_copy);
|
||
} else if (qemu_uuid_is_equal(&get_feature->uuid, &ecs_uuid)) {
|
||
if (get_feature->offset >= CXL_ECS_NUM_MEDIA_FRUS *
|
||
sizeof(CXLMemECSReadAttrs)) {
|
||
return CXL_MBOX_INVALID_INPUT;
|
||
}
|
||
bytes_to_copy = CXL_ECS_NUM_MEDIA_FRUS *
|
||
sizeof(CXLMemECSReadAttrs) -
|
||
get_feature->offset;
|
||
bytes_to_copy = MIN(bytes_to_copy, get_feature->count);
|
||
memcpy(payload_out,
|
||
(uint8_t *)&ct3d->ecs_attrs + get_feature->offset,
|
||
bytes_to_copy);
|
||
} else {
|
||
return CXL_MBOX_UNSUPPORTED;
|
||
}
|
||
|
||
*len_out = bytes_to_copy;
|
||
|
||
return CXL_MBOX_SUCCESS;
|
||
}
|
||
|
||
/* CXL r3.1 section 8.2.9.6.3: Set Feature (Opcode 0502h) */
|
||
static CXLRetCode cmd_features_set_feature(const struct cxl_cmd *cmd,
|
||
uint8_t *payload_in,
|
||
size_t len_in,
|
||
uint8_t *payload_out,
|
||
size_t *len_out,
|
||
CXLCCI *cci)
|
||
{
|
||
CXLSetFeatureInHeader *hdr = (void *)payload_in;
|
||
CXLMemPatrolScrubWriteAttrs *ps_write_attrs;
|
||
CXLMemPatrolScrubSetFeature *ps_set_feature;
|
||
CXLMemECSWriteAttrs *ecs_write_attrs;
|
||
CXLMemECSSetFeature *ecs_set_feature;
|
||
CXLSetFeatureInfo *set_feat_info;
|
||
uint16_t bytes_to_copy = 0;
|
||
uint8_t data_transfer_flag;
|
||
CXLType3Dev *ct3d;
|
||
uint16_t count;
|
||
|
||
|
||
if (!object_dynamic_cast(OBJECT(cci->d), TYPE_CXL_TYPE3)) {
|
||
return CXL_MBOX_UNSUPPORTED;
|
||
}
|
||
ct3d = CXL_TYPE3(cci->d);
|
||
set_feat_info = &ct3d->set_feat_info;
|
||
|
||
if (!qemu_uuid_is_null(&set_feat_info->uuid) &&
|
||
!qemu_uuid_is_equal(&hdr->uuid, &set_feat_info->uuid)) {
|
||
return CXL_MBOX_FEATURE_TRANSFER_IN_PROGRESS;
|
||
}
|
||
if (hdr->flags & CXL_SET_FEAT_DATA_SAVED_ACROSS_RESET) {
|
||
set_feat_info->data_saved_across_reset = true;
|
||
} else {
|
||
set_feat_info->data_saved_across_reset = false;
|
||
}
|
||
|
||
data_transfer_flag =
|
||
hdr->flags & CXL_SET_FEATURE_FLAG_DATA_TRANSFER_MASK;
|
||
if (data_transfer_flag == CXL_SET_FEATURE_FLAG_INITIATE_DATA_TRANSFER) {
|
||
set_feat_info->uuid = hdr->uuid;
|
||
set_feat_info->data_size = 0;
|
||
}
|
||
set_feat_info->data_transfer_flag = data_transfer_flag;
|
||
set_feat_info->data_offset = hdr->offset;
|
||
bytes_to_copy = len_in - sizeof(CXLSetFeatureInHeader);
|
||
|
||
if (qemu_uuid_is_equal(&hdr->uuid, &patrol_scrub_uuid)) {
|
||
if (hdr->version != CXL_MEMDEV_PS_SET_FEATURE_VERSION) {
|
||
return CXL_MBOX_UNSUPPORTED;
|
||
}
|
||
|
||
ps_set_feature = (void *)payload_in;
|
||
ps_write_attrs = &ps_set_feature->feat_data;
|
||
memcpy((uint8_t *)&ct3d->patrol_scrub_wr_attrs + hdr->offset,
|
||
ps_write_attrs,
|
||
bytes_to_copy);
|
||
set_feat_info->data_size += bytes_to_copy;
|
||
|
||
if (data_transfer_flag == CXL_SET_FEATURE_FLAG_FULL_DATA_TRANSFER ||
|
||
data_transfer_flag == CXL_SET_FEATURE_FLAG_FINISH_DATA_TRANSFER) {
|
||
ct3d->patrol_scrub_attrs.scrub_cycle &= ~0xFF;
|
||
ct3d->patrol_scrub_attrs.scrub_cycle |=
|
||
ct3d->patrol_scrub_wr_attrs.scrub_cycle_hr & 0xFF;
|
||
ct3d->patrol_scrub_attrs.scrub_flags &= ~0x1;
|
||
ct3d->patrol_scrub_attrs.scrub_flags |=
|
||
ct3d->patrol_scrub_wr_attrs.scrub_flags & 0x1;
|
||
}
|
||
} else if (qemu_uuid_is_equal(&hdr->uuid,
|
||
&ecs_uuid)) {
|
||
if (hdr->version != CXL_ECS_SET_FEATURE_VERSION) {
|
||
return CXL_MBOX_UNSUPPORTED;
|
||
}
|
||
|
||
ecs_set_feature = (void *)payload_in;
|
||
ecs_write_attrs = ecs_set_feature->feat_data;
|
||
memcpy((uint8_t *)ct3d->ecs_wr_attrs + hdr->offset,
|
||
ecs_write_attrs,
|
||
bytes_to_copy);
|
||
set_feat_info->data_size += bytes_to_copy;
|
||
|
||
if (data_transfer_flag == CXL_SET_FEATURE_FLAG_FULL_DATA_TRANSFER ||
|
||
data_transfer_flag == CXL_SET_FEATURE_FLAG_FINISH_DATA_TRANSFER) {
|
||
for (count = 0; count < CXL_ECS_NUM_MEDIA_FRUS; count++) {
|
||
ct3d->ecs_attrs[count].ecs_log_cap =
|
||
ct3d->ecs_wr_attrs[count].ecs_log_cap;
|
||
ct3d->ecs_attrs[count].ecs_config =
|
||
ct3d->ecs_wr_attrs[count].ecs_config & 0x1F;
|
||
}
|
||
}
|
||
} else {
|
||
return CXL_MBOX_UNSUPPORTED;
|
||
}
|
||
|
||
if (data_transfer_flag == CXL_SET_FEATURE_FLAG_FULL_DATA_TRANSFER ||
|
||
data_transfer_flag == CXL_SET_FEATURE_FLAG_FINISH_DATA_TRANSFER ||
|
||
data_transfer_flag == CXL_SET_FEATURE_FLAG_ABORT_DATA_TRANSFER) {
|
||
memset(&set_feat_info->uuid, 0, sizeof(QemuUUID));
|
||
if (qemu_uuid_is_equal(&hdr->uuid, &patrol_scrub_uuid)) {
|
||
memset(&ct3d->patrol_scrub_wr_attrs, 0, set_feat_info->data_size);
|
||
} else if (qemu_uuid_is_equal(&hdr->uuid, &ecs_uuid)) {
|
||
memset(ct3d->ecs_wr_attrs, 0, set_feat_info->data_size);
|
||
}
|
||
set_feat_info->data_transfer_flag = 0;
|
||
set_feat_info->data_saved_across_reset = false;
|
||
set_feat_info->data_offset = 0;
|
||
set_feat_info->data_size = 0;
|
||
}
|
||
|
||
return CXL_MBOX_SUCCESS;
|
||
}
|
||
|
||
/* CXL r3.1 Section 8.2.9.9.1.1: Identify Memory Device (Opcode 4000h) */
|
||
static CXLRetCode cmd_identify_memory_device(const struct cxl_cmd *cmd,
|
||
uint8_t *payload_in,
|
||
size_t len_in,
|
||
uint8_t *payload_out,
|
||
size_t *len_out,
|
||
CXLCCI *cci)
|
||
{
|
||
struct {
|
||
char fw_revision[0x10];
|
||
uint64_t total_capacity;
|
||
uint64_t volatile_capacity;
|
||
uint64_t persistent_capacity;
|
||
uint64_t partition_align;
|
||
uint16_t info_event_log_size;
|
||
uint16_t warning_event_log_size;
|
||
uint16_t failure_event_log_size;
|
||
uint16_t fatal_event_log_size;
|
||
uint32_t lsa_size;
|
||
uint8_t poison_list_max_mer[3];
|
||
uint16_t inject_poison_limit;
|
||
uint8_t poison_caps;
|
||
uint8_t qos_telemetry_caps;
|
||
uint16_t dc_event_log_size;
|
||
} QEMU_PACKED *id;
|
||
QEMU_BUILD_BUG_ON(sizeof(*id) != 0x45);
|
||
CXLType3Dev *ct3d = CXL_TYPE3(cci->d);
|
||
CXLType3Class *cvc = CXL_TYPE3_GET_CLASS(ct3d);
|
||
CXLDeviceState *cxl_dstate = &ct3d->cxl_dstate;
|
||
|
||
if ((!QEMU_IS_ALIGNED(cxl_dstate->vmem_size, CXL_CAPACITY_MULTIPLIER)) ||
|
||
(!QEMU_IS_ALIGNED(cxl_dstate->pmem_size, CXL_CAPACITY_MULTIPLIER)) ||
|
||
(!QEMU_IS_ALIGNED(ct3d->dc.total_capacity, CXL_CAPACITY_MULTIPLIER))) {
|
||
return CXL_MBOX_INTERNAL_ERROR;
|
||
}
|
||
|
||
id = (void *)payload_out;
|
||
|
||
snprintf(id->fw_revision, 0x10, "BWFW VERSION %02d", 0);
|
||
|
||
stq_le_p(&id->total_capacity,
|
||
cxl_dstate->static_mem_size / CXL_CAPACITY_MULTIPLIER);
|
||
stq_le_p(&id->persistent_capacity,
|
||
cxl_dstate->pmem_size / CXL_CAPACITY_MULTIPLIER);
|
||
stq_le_p(&id->volatile_capacity,
|
||
cxl_dstate->vmem_size / CXL_CAPACITY_MULTIPLIER);
|
||
stl_le_p(&id->lsa_size, cvc->get_lsa_size(ct3d));
|
||
/* 256 poison records */
|
||
st24_le_p(id->poison_list_max_mer, 256);
|
||
/* No limit - so limited by main poison record limit */
|
||
stw_le_p(&id->inject_poison_limit, 0);
|
||
stw_le_p(&id->dc_event_log_size, CXL_DC_EVENT_LOG_SIZE);
|
||
|
||
*len_out = sizeof(*id);
|
||
return CXL_MBOX_SUCCESS;
|
||
}
|
||
|
||
/* CXL r3.1 Section 8.2.9.9.2.1: Get Partition Info (Opcode 4100h) */
|
||
static CXLRetCode cmd_ccls_get_partition_info(const struct cxl_cmd *cmd,
|
||
uint8_t *payload_in,
|
||
size_t len_in,
|
||
uint8_t *payload_out,
|
||
size_t *len_out,
|
||
CXLCCI *cci)
|
||
{
|
||
CXLDeviceState *cxl_dstate = &CXL_TYPE3(cci->d)->cxl_dstate;
|
||
struct {
|
||
uint64_t active_vmem;
|
||
uint64_t active_pmem;
|
||
uint64_t next_vmem;
|
||
uint64_t next_pmem;
|
||
} QEMU_PACKED *part_info = (void *)payload_out;
|
||
QEMU_BUILD_BUG_ON(sizeof(*part_info) != 0x20);
|
||
CXLType3Dev *ct3d = container_of(cxl_dstate, CXLType3Dev, cxl_dstate);
|
||
|
||
if ((!QEMU_IS_ALIGNED(cxl_dstate->vmem_size, CXL_CAPACITY_MULTIPLIER)) ||
|
||
(!QEMU_IS_ALIGNED(cxl_dstate->pmem_size, CXL_CAPACITY_MULTIPLIER)) ||
|
||
(!QEMU_IS_ALIGNED(ct3d->dc.total_capacity, CXL_CAPACITY_MULTIPLIER))) {
|
||
return CXL_MBOX_INTERNAL_ERROR;
|
||
}
|
||
|
||
stq_le_p(&part_info->active_vmem,
|
||
cxl_dstate->vmem_size / CXL_CAPACITY_MULTIPLIER);
|
||
/*
|
||
* When both next_vmem and next_pmem are 0, there is no pending change to
|
||
* partitioning.
|
||
*/
|
||
stq_le_p(&part_info->next_vmem, 0);
|
||
stq_le_p(&part_info->active_pmem,
|
||
cxl_dstate->pmem_size / CXL_CAPACITY_MULTIPLIER);
|
||
stq_le_p(&part_info->next_pmem, 0);
|
||
|
||
*len_out = sizeof(*part_info);
|
||
return CXL_MBOX_SUCCESS;
|
||
}
|
||
|
||
/* CXL r3.1 Section 8.2.9.9.2.3: Get LSA (Opcode 4102h) */
|
||
static CXLRetCode cmd_ccls_get_lsa(const struct cxl_cmd *cmd,
|
||
uint8_t *payload_in,
|
||
size_t len_in,
|
||
uint8_t *payload_out,
|
||
size_t *len_out,
|
||
CXLCCI *cci)
|
||
{
|
||
struct {
|
||
uint32_t offset;
|
||
uint32_t length;
|
||
} QEMU_PACKED *get_lsa;
|
||
CXLType3Dev *ct3d = CXL_TYPE3(cci->d);
|
||
CXLType3Class *cvc = CXL_TYPE3_GET_CLASS(ct3d);
|
||
uint32_t offset, length;
|
||
|
||
get_lsa = (void *)payload_in;
|
||
offset = get_lsa->offset;
|
||
length = get_lsa->length;
|
||
|
||
if (offset + length > cvc->get_lsa_size(ct3d)) {
|
||
*len_out = 0;
|
||
return CXL_MBOX_INVALID_INPUT;
|
||
}
|
||
|
||
*len_out = cvc->get_lsa(ct3d, payload_out, length, offset);
|
||
return CXL_MBOX_SUCCESS;
|
||
}
|
||
|
||
/* CXL r3.1 Section 8.2.9.9.2.4: Set LSA (Opcode 4103h) */
|
||
static CXLRetCode cmd_ccls_set_lsa(const struct cxl_cmd *cmd,
|
||
uint8_t *payload_in,
|
||
size_t len_in,
|
||
uint8_t *payload_out,
|
||
size_t *len_out,
|
||
CXLCCI *cci)
|
||
{
|
||
struct set_lsa_pl {
|
||
uint32_t offset;
|
||
uint32_t rsvd;
|
||
uint8_t data[];
|
||
} QEMU_PACKED;
|
||
struct set_lsa_pl *set_lsa_payload = (void *)payload_in;
|
||
CXLType3Dev *ct3d = CXL_TYPE3(cci->d);
|
||
CXLType3Class *cvc = CXL_TYPE3_GET_CLASS(ct3d);
|
||
const size_t hdr_len = offsetof(struct set_lsa_pl, data);
|
||
|
||
*len_out = 0;
|
||
if (!len_in) {
|
||
return CXL_MBOX_SUCCESS;
|
||
}
|
||
|
||
if (set_lsa_payload->offset + len_in > cvc->get_lsa_size(ct3d) + hdr_len) {
|
||
return CXL_MBOX_INVALID_INPUT;
|
||
}
|
||
len_in -= hdr_len;
|
||
|
||
cvc->set_lsa(ct3d, set_lsa_payload->data, len_in, set_lsa_payload->offset);
|
||
return CXL_MBOX_SUCCESS;
|
||
}
|
||
|
||
/* Perform the actual device zeroing */
|
||
static void __do_sanitization(CXLType3Dev *ct3d)
|
||
{
|
||
MemoryRegion *mr;
|
||
|
||
if (ct3d->hostvmem) {
|
||
mr = host_memory_backend_get_memory(ct3d->hostvmem);
|
||
if (mr) {
|
||
void *hostmem = memory_region_get_ram_ptr(mr);
|
||
memset(hostmem, 0, memory_region_size(mr));
|
||
}
|
||
}
|
||
|
||
if (ct3d->hostpmem) {
|
||
mr = host_memory_backend_get_memory(ct3d->hostpmem);
|
||
if (mr) {
|
||
void *hostmem = memory_region_get_ram_ptr(mr);
|
||
memset(hostmem, 0, memory_region_size(mr));
|
||
}
|
||
}
|
||
if (ct3d->lsa) {
|
||
mr = host_memory_backend_get_memory(ct3d->lsa);
|
||
if (mr) {
|
||
void *lsa = memory_region_get_ram_ptr(mr);
|
||
memset(lsa, 0, memory_region_size(mr));
|
||
}
|
||
}
|
||
cxl_discard_all_event_records(&ct3d->cxl_dstate);
|
||
}
|
||
|
||
/*
|
||
* CXL r3.1 Section 8.2.9.9.5.1: Sanitize (Opcode 4400h)
|
||
*
|
||
* Once the Sanitize command has started successfully, the device shall be
|
||
* placed in the media disabled state. If the command fails or is interrupted
|
||
* by a reset or power failure, it shall remain in the media disabled state
|
||
* until a successful Sanitize command has been completed. During this state:
|
||
*
|
||
* 1. Memory writes to the device will have no effect, and all memory reads
|
||
* will return random values (no user data returned, even for locations that
|
||
* the failed Sanitize operation didn’t sanitize yet).
|
||
*
|
||
* 2. Mailbox commands shall still be processed in the disabled state, except
|
||
* that commands that access Sanitized areas shall fail with the Media Disabled
|
||
* error code.
|
||
*/
|
||
static CXLRetCode cmd_sanitize_overwrite(const struct cxl_cmd *cmd,
|
||
uint8_t *payload_in,
|
||
size_t len_in,
|
||
uint8_t *payload_out,
|
||
size_t *len_out,
|
||
CXLCCI *cci)
|
||
{
|
||
CXLType3Dev *ct3d = CXL_TYPE3(cci->d);
|
||
uint64_t total_mem; /* in Mb */
|
||
int secs;
|
||
|
||
total_mem = (ct3d->cxl_dstate.vmem_size + ct3d->cxl_dstate.pmem_size) >> 20;
|
||
if (total_mem <= 512) {
|
||
secs = 4;
|
||
} else if (total_mem <= 1024) {
|
||
secs = 8;
|
||
} else if (total_mem <= 2 * 1024) {
|
||
secs = 15;
|
||
} else if (total_mem <= 4 * 1024) {
|
||
secs = 30;
|
||
} else if (total_mem <= 8 * 1024) {
|
||
secs = 60;
|
||
} else if (total_mem <= 16 * 1024) {
|
||
secs = 2 * 60;
|
||
} else if (total_mem <= 32 * 1024) {
|
||
secs = 4 * 60;
|
||
} else if (total_mem <= 64 * 1024) {
|
||
secs = 8 * 60;
|
||
} else if (total_mem <= 128 * 1024) {
|
||
secs = 15 * 60;
|
||
} else if (total_mem <= 256 * 1024) {
|
||
secs = 30 * 60;
|
||
} else if (total_mem <= 512 * 1024) {
|
||
secs = 60 * 60;
|
||
} else if (total_mem <= 1024 * 1024) {
|
||
secs = 120 * 60;
|
||
} else {
|
||
secs = 240 * 60; /* max 4 hrs */
|
||
}
|
||
|
||
/* EBUSY other bg cmds as of now */
|
||
cci->bg.runtime = secs * 1000UL;
|
||
*len_out = 0;
|
||
|
||
cxl_dev_disable_media(&ct3d->cxl_dstate);
|
||
|
||
/* sanitize when done */
|
||
return CXL_MBOX_BG_STARTED;
|
||
}
|
||
|
||
static CXLRetCode cmd_get_security_state(const struct cxl_cmd *cmd,
|
||
uint8_t *payload_in,
|
||
size_t len_in,
|
||
uint8_t *payload_out,
|
||
size_t *len_out,
|
||
CXLCCI *cci)
|
||
{
|
||
uint32_t *state = (uint32_t *)payload_out;
|
||
|
||
*state = 0;
|
||
*len_out = 4;
|
||
return CXL_MBOX_SUCCESS;
|
||
}
|
||
|
||
/*
|
||
* CXL r3.1 Section 8.2.9.9.4.1: Get Poison List (Opcode 4300h)
|
||
*
|
||
* This is very inefficient, but good enough for now!
|
||
* Also the payload will always fit, so no need to handle the MORE flag and
|
||
* make this stateful. We may want to allow longer poison lists to aid
|
||
* testing that kernel functionality.
|
||
*/
|
||
static CXLRetCode cmd_media_get_poison_list(const struct cxl_cmd *cmd,
|
||
uint8_t *payload_in,
|
||
size_t len_in,
|
||
uint8_t *payload_out,
|
||
size_t *len_out,
|
||
CXLCCI *cci)
|
||
{
|
||
struct get_poison_list_pl {
|
||
uint64_t pa;
|
||
uint64_t length;
|
||
} QEMU_PACKED;
|
||
|
||
struct get_poison_list_out_pl {
|
||
uint8_t flags;
|
||
uint8_t rsvd1;
|
||
uint64_t overflow_timestamp;
|
||
uint16_t count;
|
||
uint8_t rsvd2[0x14];
|
||
struct {
|
||
uint64_t addr;
|
||
uint32_t length;
|
||
uint32_t resv;
|
||
} QEMU_PACKED records[];
|
||
} QEMU_PACKED;
|
||
|
||
struct get_poison_list_pl *in = (void *)payload_in;
|
||
struct get_poison_list_out_pl *out = (void *)payload_out;
|
||
CXLType3Dev *ct3d = CXL_TYPE3(cci->d);
|
||
uint16_t record_count = 0, i = 0;
|
||
uint64_t query_start, query_length;
|
||
CXLPoisonList *poison_list = &ct3d->poison_list;
|
||
CXLPoison *ent;
|
||
uint16_t out_pl_len;
|
||
|
||
query_start = ldq_le_p(&in->pa);
|
||
/* 64 byte alignment required */
|
||
if (query_start & 0x3f) {
|
||
return CXL_MBOX_INVALID_INPUT;
|
||
}
|
||
query_length = ldq_le_p(&in->length) * CXL_CACHE_LINE_SIZE;
|
||
|
||
QLIST_FOREACH(ent, poison_list, node) {
|
||
/* Check for no overlap */
|
||
if (!ranges_overlap(ent->start, ent->length,
|
||
query_start, query_length)) {
|
||
continue;
|
||
}
|
||
record_count++;
|
||
}
|
||
out_pl_len = sizeof(*out) + record_count * sizeof(out->records[0]);
|
||
assert(out_pl_len <= CXL_MAILBOX_MAX_PAYLOAD_SIZE);
|
||
|
||
QLIST_FOREACH(ent, poison_list, node) {
|
||
uint64_t start, stop;
|
||
|
||
/* Check for no overlap */
|
||
if (!ranges_overlap(ent->start, ent->length,
|
||
query_start, query_length)) {
|
||
continue;
|
||
}
|
||
|
||
/* Deal with overlap */
|
||
start = MAX(ROUND_DOWN(ent->start, 64ull), query_start);
|
||
stop = MIN(ROUND_DOWN(ent->start, 64ull) + ent->length,
|
||
query_start + query_length);
|
||
stq_le_p(&out->records[i].addr, start | (ent->type & 0x7));
|
||
stl_le_p(&out->records[i].length, (stop - start) / CXL_CACHE_LINE_SIZE);
|
||
i++;
|
||
}
|
||
if (ct3d->poison_list_overflowed) {
|
||
out->flags = (1 << 1);
|
||
stq_le_p(&out->overflow_timestamp, ct3d->poison_list_overflow_ts);
|
||
}
|
||
if (scan_media_running(cci)) {
|
||
out->flags |= (1 << 2);
|
||
}
|
||
|
||
stw_le_p(&out->count, record_count);
|
||
*len_out = out_pl_len;
|
||
return CXL_MBOX_SUCCESS;
|
||
}
|
||
|
||
/* CXL r3.1 Section 8.2.9.9.4.2: Inject Poison (Opcode 4301h) */
|
||
static CXLRetCode cmd_media_inject_poison(const struct cxl_cmd *cmd,
|
||
uint8_t *payload_in,
|
||
size_t len_in,
|
||
uint8_t *payload_out,
|
||
size_t *len_out,
|
||
CXLCCI *cci)
|
||
{
|
||
CXLType3Dev *ct3d = CXL_TYPE3(cci->d);
|
||
CXLPoisonList *poison_list = &ct3d->poison_list;
|
||
CXLPoison *ent;
|
||
struct inject_poison_pl {
|
||
uint64_t dpa;
|
||
};
|
||
struct inject_poison_pl *in = (void *)payload_in;
|
||
uint64_t dpa = ldq_le_p(&in->dpa);
|
||
CXLPoison *p;
|
||
|
||
QLIST_FOREACH(ent, poison_list, node) {
|
||
if (dpa >= ent->start &&
|
||
dpa + CXL_CACHE_LINE_SIZE <= ent->start + ent->length) {
|
||
return CXL_MBOX_SUCCESS;
|
||
}
|
||
}
|
||
/*
|
||
* Freeze the list if there is an on-going scan media operation.
|
||
*/
|
||
if (scan_media_running(cci)) {
|
||
/*
|
||
* XXX: Spec is ambiguous - is this case considered
|
||
* a successful return despite not adding to the list?
|
||
*/
|
||
goto success;
|
||
}
|
||
|
||
if (ct3d->poison_list_cnt == CXL_POISON_LIST_LIMIT) {
|
||
return CXL_MBOX_INJECT_POISON_LIMIT;
|
||
}
|
||
p = g_new0(CXLPoison, 1);
|
||
|
||
p->length = CXL_CACHE_LINE_SIZE;
|
||
p->start = dpa;
|
||
p->type = CXL_POISON_TYPE_INJECTED;
|
||
|
||
/*
|
||
* Possible todo: Merge with existing entry if next to it and if same type
|
||
*/
|
||
QLIST_INSERT_HEAD(poison_list, p, node);
|
||
ct3d->poison_list_cnt++;
|
||
success:
|
||
*len_out = 0;
|
||
|
||
return CXL_MBOX_SUCCESS;
|
||
}
|
||
|
||
/* CXL r3.1 Section 8.2.9.9.4.3: Clear Poison (Opcode 4302h */
|
||
static CXLRetCode cmd_media_clear_poison(const struct cxl_cmd *cmd,
|
||
uint8_t *payload_in,
|
||
size_t len_in,
|
||
uint8_t *payload_out,
|
||
size_t *len_out,
|
||
CXLCCI *cci)
|
||
{
|
||
CXLType3Dev *ct3d = CXL_TYPE3(cci->d);
|
||
CXLDeviceState *cxl_dstate = &ct3d->cxl_dstate;
|
||
CXLPoisonList *poison_list = &ct3d->poison_list;
|
||
CXLType3Class *cvc = CXL_TYPE3_GET_CLASS(ct3d);
|
||
struct clear_poison_pl {
|
||
uint64_t dpa;
|
||
uint8_t data[64];
|
||
};
|
||
CXLPoison *ent;
|
||
uint64_t dpa;
|
||
|
||
struct clear_poison_pl *in = (void *)payload_in;
|
||
|
||
dpa = ldq_le_p(&in->dpa);
|
||
if (dpa + CXL_CACHE_LINE_SIZE > cxl_dstate->static_mem_size +
|
||
ct3d->dc.total_capacity) {
|
||
return CXL_MBOX_INVALID_PA;
|
||
}
|
||
|
||
/* Clearing a region with no poison is not an error so always do so */
|
||
if (cvc->set_cacheline) {
|
||
if (!cvc->set_cacheline(ct3d, dpa, in->data)) {
|
||
return CXL_MBOX_INTERNAL_ERROR;
|
||
}
|
||
}
|
||
|
||
/*
|
||
* Freeze the list if there is an on-going scan media operation.
|
||
*/
|
||
if (scan_media_running(cci)) {
|
||
/*
|
||
* XXX: Spec is ambiguous - is this case considered
|
||
* a successful return despite not removing from the list?
|
||
*/
|
||
goto success;
|
||
}
|
||
|
||
QLIST_FOREACH(ent, poison_list, node) {
|
||
/*
|
||
* Test for contained in entry. Simpler than general case
|
||
* as clearing 64 bytes and entries 64 byte aligned
|
||
*/
|
||
if ((dpa >= ent->start) && (dpa < ent->start + ent->length)) {
|
||
break;
|
||
}
|
||
}
|
||
if (!ent) {
|
||
goto success;
|
||
}
|
||
|
||
QLIST_REMOVE(ent, node);
|
||
ct3d->poison_list_cnt--;
|
||
|
||
if (dpa > ent->start) {
|
||
CXLPoison *frag;
|
||
/* Cannot overflow as replacing existing entry */
|
||
|
||
frag = g_new0(CXLPoison, 1);
|
||
|
||
frag->start = ent->start;
|
||
frag->length = dpa - ent->start;
|
||
frag->type = ent->type;
|
||
|
||
QLIST_INSERT_HEAD(poison_list, frag, node);
|
||
ct3d->poison_list_cnt++;
|
||
}
|
||
|
||
if (dpa + CXL_CACHE_LINE_SIZE < ent->start + ent->length) {
|
||
CXLPoison *frag;
|
||
|
||
if (ct3d->poison_list_cnt == CXL_POISON_LIST_LIMIT) {
|
||
cxl_set_poison_list_overflowed(ct3d);
|
||
} else {
|
||
frag = g_new0(CXLPoison, 1);
|
||
|
||
frag->start = dpa + CXL_CACHE_LINE_SIZE;
|
||
frag->length = ent->start + ent->length - frag->start;
|
||
frag->type = ent->type;
|
||
QLIST_INSERT_HEAD(poison_list, frag, node);
|
||
ct3d->poison_list_cnt++;
|
||
}
|
||
}
|
||
/* Any fragments have been added, free original entry */
|
||
g_free(ent);
|
||
success:
|
||
*len_out = 0;
|
||
|
||
return CXL_MBOX_SUCCESS;
|
||
}
|
||
|
||
/*
|
||
* CXL r3.1 section 8.2.9.9.4.4: Get Scan Media Capabilities
|
||
*/
|
||
static CXLRetCode
|
||
cmd_media_get_scan_media_capabilities(const struct cxl_cmd *cmd,
|
||
uint8_t *payload_in,
|
||
size_t len_in,
|
||
uint8_t *payload_out,
|
||
size_t *len_out,
|
||
CXLCCI *cci)
|
||
{
|
||
struct get_scan_media_capabilities_pl {
|
||
uint64_t pa;
|
||
uint64_t length;
|
||
} QEMU_PACKED;
|
||
|
||
struct get_scan_media_capabilities_out_pl {
|
||
uint32_t estimated_runtime_ms;
|
||
};
|
||
|
||
CXLType3Dev *ct3d = CXL_TYPE3(cci->d);
|
||
CXLDeviceState *cxl_dstate = &ct3d->cxl_dstate;
|
||
struct get_scan_media_capabilities_pl *in = (void *)payload_in;
|
||
struct get_scan_media_capabilities_out_pl *out = (void *)payload_out;
|
||
uint64_t query_start;
|
||
uint64_t query_length;
|
||
|
||
query_start = ldq_le_p(&in->pa);
|
||
/* 64 byte alignment required */
|
||
if (query_start & 0x3f) {
|
||
return CXL_MBOX_INVALID_INPUT;
|
||
}
|
||
query_length = ldq_le_p(&in->length) * CXL_CACHE_LINE_SIZE;
|
||
|
||
if (query_start + query_length > cxl_dstate->static_mem_size) {
|
||
return CXL_MBOX_INVALID_PA;
|
||
}
|
||
|
||
/*
|
||
* Just use 400 nanosecond access/read latency + 100 ns for
|
||
* the cost of updating the poison list. For small enough
|
||
* chunks return at least 1 ms.
|
||
*/
|
||
stl_le_p(&out->estimated_runtime_ms,
|
||
MAX(1, query_length * (0.0005L / 64)));
|
||
|
||
*len_out = sizeof(*out);
|
||
return CXL_MBOX_SUCCESS;
|
||
}
|
||
|
||
static void __do_scan_media(CXLType3Dev *ct3d)
|
||
{
|
||
CXLPoison *ent;
|
||
unsigned int results_cnt = 0;
|
||
|
||
QLIST_FOREACH(ent, &ct3d->scan_media_results, node) {
|
||
results_cnt++;
|
||
}
|
||
|
||
/* only scan media may clear the overflow */
|
||
if (ct3d->poison_list_overflowed &&
|
||
ct3d->poison_list_cnt == results_cnt) {
|
||
cxl_clear_poison_list_overflowed(ct3d);
|
||
}
|
||
/* scan media has run since last conventional reset */
|
||
ct3d->scan_media_hasrun = true;
|
||
}
|
||
|
||
/*
|
||
* CXL r3.1 section 8.2.9.9.4.5: Scan Media
|
||
*/
|
||
static CXLRetCode cmd_media_scan_media(const struct cxl_cmd *cmd,
|
||
uint8_t *payload_in,
|
||
size_t len_in,
|
||
uint8_t *payload_out,
|
||
size_t *len_out,
|
||
CXLCCI *cci)
|
||
{
|
||
struct scan_media_pl {
|
||
uint64_t pa;
|
||
uint64_t length;
|
||
uint8_t flags;
|
||
} QEMU_PACKED;
|
||
|
||
struct scan_media_pl *in = (void *)payload_in;
|
||
CXLType3Dev *ct3d = CXL_TYPE3(cci->d);
|
||
CXLDeviceState *cxl_dstate = &ct3d->cxl_dstate;
|
||
uint64_t query_start;
|
||
uint64_t query_length;
|
||
CXLPoison *ent, *next;
|
||
|
||
query_start = ldq_le_p(&in->pa);
|
||
/* 64 byte alignment required */
|
||
if (query_start & 0x3f) {
|
||
return CXL_MBOX_INVALID_INPUT;
|
||
}
|
||
query_length = ldq_le_p(&in->length) * CXL_CACHE_LINE_SIZE;
|
||
|
||
if (query_start + query_length > cxl_dstate->static_mem_size) {
|
||
return CXL_MBOX_INVALID_PA;
|
||
}
|
||
if (ct3d->dc.num_regions && query_start + query_length >=
|
||
cxl_dstate->static_mem_size + ct3d->dc.total_capacity) {
|
||
return CXL_MBOX_INVALID_PA;
|
||
}
|
||
|
||
if (in->flags == 0) { /* TODO */
|
||
qemu_log_mask(LOG_UNIMP,
|
||
"Scan Media Event Log is unsupported\n");
|
||
}
|
||
|
||
/* any previous results are discarded upon a new Scan Media */
|
||
QLIST_FOREACH_SAFE(ent, &ct3d->scan_media_results, node, next) {
|
||
QLIST_REMOVE(ent, node);
|
||
g_free(ent);
|
||
}
|
||
|
||
/* kill the poison list - it will be recreated */
|
||
if (ct3d->poison_list_overflowed) {
|
||
QLIST_FOREACH_SAFE(ent, &ct3d->poison_list, node, next) {
|
||
QLIST_REMOVE(ent, node);
|
||
g_free(ent);
|
||
ct3d->poison_list_cnt--;
|
||
}
|
||
}
|
||
|
||
/*
|
||
* Scan the backup list and move corresponding entries
|
||
* into the results list, updating the poison list
|
||
* when possible.
|
||
*/
|
||
QLIST_FOREACH_SAFE(ent, &ct3d->poison_list_bkp, node, next) {
|
||
CXLPoison *res;
|
||
|
||
if (ent->start >= query_start + query_length ||
|
||
ent->start + ent->length <= query_start) {
|
||
continue;
|
||
}
|
||
|
||
/*
|
||
* If a Get Poison List cmd comes in while this
|
||
* scan is being done, it will see the new complete
|
||
* list, while setting the respective flag.
|
||
*/
|
||
if (ct3d->poison_list_cnt < CXL_POISON_LIST_LIMIT) {
|
||
CXLPoison *p = g_new0(CXLPoison, 1);
|
||
|
||
p->start = ent->start;
|
||
p->length = ent->length;
|
||
p->type = ent->type;
|
||
QLIST_INSERT_HEAD(&ct3d->poison_list, p, node);
|
||
ct3d->poison_list_cnt++;
|
||
}
|
||
|
||
res = g_new0(CXLPoison, 1);
|
||
res->start = ent->start;
|
||
res->length = ent->length;
|
||
res->type = ent->type;
|
||
QLIST_INSERT_HEAD(&ct3d->scan_media_results, res, node);
|
||
|
||
QLIST_REMOVE(ent, node);
|
||
g_free(ent);
|
||
}
|
||
|
||
cci->bg.runtime = MAX(1, query_length * (0.0005L / 64));
|
||
*len_out = 0;
|
||
|
||
return CXL_MBOX_BG_STARTED;
|
||
}
|
||
|
||
/*
|
||
* CXL r3.1 section 8.2.9.9.4.6: Get Scan Media Results
|
||
*/
|
||
static CXLRetCode cmd_media_get_scan_media_results(const struct cxl_cmd *cmd,
|
||
uint8_t *payload_in,
|
||
size_t len_in,
|
||
uint8_t *payload_out,
|
||
size_t *len_out,
|
||
CXLCCI *cci)
|
||
{
|
||
struct get_scan_media_results_out_pl {
|
||
uint64_t dpa_restart;
|
||
uint64_t length;
|
||
uint8_t flags;
|
||
uint8_t rsvd1;
|
||
uint16_t count;
|
||
uint8_t rsvd2[0xc];
|
||
struct {
|
||
uint64_t addr;
|
||
uint32_t length;
|
||
uint32_t resv;
|
||
} QEMU_PACKED records[];
|
||
} QEMU_PACKED;
|
||
|
||
struct get_scan_media_results_out_pl *out = (void *)payload_out;
|
||
CXLType3Dev *ct3d = CXL_TYPE3(cci->d);
|
||
CXLPoisonList *scan_media_results = &ct3d->scan_media_results;
|
||
CXLPoison *ent, *next;
|
||
uint16_t total_count = 0, record_count = 0, i = 0;
|
||
uint16_t out_pl_len;
|
||
|
||
if (!ct3d->scan_media_hasrun) {
|
||
return CXL_MBOX_UNSUPPORTED;
|
||
}
|
||
|
||
/*
|
||
* Calculate limits, all entries are within the same address range of the
|
||
* last scan media call.
|
||
*/
|
||
QLIST_FOREACH(ent, scan_media_results, node) {
|
||
size_t rec_size = record_count * sizeof(out->records[0]);
|
||
|
||
if (sizeof(*out) + rec_size < CXL_MAILBOX_MAX_PAYLOAD_SIZE) {
|
||
record_count++;
|
||
}
|
||
total_count++;
|
||
}
|
||
|
||
out_pl_len = sizeof(*out) + record_count * sizeof(out->records[0]);
|
||
assert(out_pl_len <= CXL_MAILBOX_MAX_PAYLOAD_SIZE);
|
||
|
||
memset(out, 0, out_pl_len);
|
||
QLIST_FOREACH_SAFE(ent, scan_media_results, node, next) {
|
||
uint64_t start, stop;
|
||
|
||
if (i == record_count) {
|
||
break;
|
||
}
|
||
|
||
start = ROUND_DOWN(ent->start, 64ull);
|
||
stop = ROUND_DOWN(ent->start, 64ull) + ent->length;
|
||
stq_le_p(&out->records[i].addr, start | (ent->type & 0x7));
|
||
stl_le_p(&out->records[i].length, (stop - start) / CXL_CACHE_LINE_SIZE);
|
||
i++;
|
||
|
||
/* consume the returning entry */
|
||
QLIST_REMOVE(ent, node);
|
||
g_free(ent);
|
||
}
|
||
|
||
stw_le_p(&out->count, record_count);
|
||
if (total_count > record_count) {
|
||
out->flags = (1 << 0); /* More Media Error Records */
|
||
}
|
||
|
||
*len_out = out_pl_len;
|
||
return CXL_MBOX_SUCCESS;
|
||
}
|
||
|
||
/*
|
||
* CXL r3.1 section 8.2.9.9.9.1: Get Dynamic Capacity Configuration
|
||
* (Opcode: 4800h)
|
||
*/
|
||
static CXLRetCode cmd_dcd_get_dyn_cap_config(const struct cxl_cmd *cmd,
|
||
uint8_t *payload_in,
|
||
size_t len_in,
|
||
uint8_t *payload_out,
|
||
size_t *len_out,
|
||
CXLCCI *cci)
|
||
{
|
||
CXLType3Dev *ct3d = CXL_TYPE3(cci->d);
|
||
struct {
|
||
uint8_t region_cnt;
|
||
uint8_t start_rid;
|
||
} QEMU_PACKED *in = (void *)payload_in;
|
||
struct {
|
||
uint8_t num_regions;
|
||
uint8_t regions_returned;
|
||
uint8_t rsvd1[6];
|
||
struct {
|
||
uint64_t base;
|
||
uint64_t decode_len;
|
||
uint64_t region_len;
|
||
uint64_t block_size;
|
||
uint32_t dsmadhandle;
|
||
uint8_t flags;
|
||
uint8_t rsvd2[3];
|
||
} QEMU_PACKED records[];
|
||
} QEMU_PACKED *out = (void *)payload_out;
|
||
struct {
|
||
uint32_t num_extents_supported;
|
||
uint32_t num_extents_available;
|
||
uint32_t num_tags_supported;
|
||
uint32_t num_tags_available;
|
||
} QEMU_PACKED *extra_out;
|
||
uint16_t record_count;
|
||
uint16_t i;
|
||
uint16_t out_pl_len;
|
||
uint8_t start_rid;
|
||
|
||
start_rid = in->start_rid;
|
||
if (start_rid >= ct3d->dc.num_regions) {
|
||
return CXL_MBOX_INVALID_INPUT;
|
||
}
|
||
|
||
record_count = MIN(ct3d->dc.num_regions - in->start_rid, in->region_cnt);
|
||
|
||
out_pl_len = sizeof(*out) + record_count * sizeof(out->records[0]);
|
||
extra_out = (void *)(payload_out + out_pl_len);
|
||
out_pl_len += sizeof(*extra_out);
|
||
assert(out_pl_len <= CXL_MAILBOX_MAX_PAYLOAD_SIZE);
|
||
|
||
out->num_regions = ct3d->dc.num_regions;
|
||
out->regions_returned = record_count;
|
||
for (i = 0; i < record_count; i++) {
|
||
stq_le_p(&out->records[i].base,
|
||
ct3d->dc.regions[start_rid + i].base);
|
||
stq_le_p(&out->records[i].decode_len,
|
||
ct3d->dc.regions[start_rid + i].decode_len /
|
||
CXL_CAPACITY_MULTIPLIER);
|
||
stq_le_p(&out->records[i].region_len,
|
||
ct3d->dc.regions[start_rid + i].len);
|
||
stq_le_p(&out->records[i].block_size,
|
||
ct3d->dc.regions[start_rid + i].block_size);
|
||
stl_le_p(&out->records[i].dsmadhandle,
|
||
ct3d->dc.regions[start_rid + i].dsmadhandle);
|
||
out->records[i].flags = ct3d->dc.regions[start_rid + i].flags;
|
||
}
|
||
/*
|
||
* TODO: Assign values once extents and tags are introduced
|
||
* to use.
|
||
*/
|
||
stl_le_p(&extra_out->num_extents_supported, CXL_NUM_EXTENTS_SUPPORTED);
|
||
stl_le_p(&extra_out->num_extents_available, CXL_NUM_EXTENTS_SUPPORTED -
|
||
ct3d->dc.total_extent_count);
|
||
stl_le_p(&extra_out->num_tags_supported, CXL_NUM_TAGS_SUPPORTED);
|
||
stl_le_p(&extra_out->num_tags_available, CXL_NUM_TAGS_SUPPORTED);
|
||
|
||
*len_out = out_pl_len;
|
||
return CXL_MBOX_SUCCESS;
|
||
}
|
||
|
||
/*
|
||
* CXL r3.1 section 8.2.9.9.9.2:
|
||
* Get Dynamic Capacity Extent List (Opcode 4801h)
|
||
*/
|
||
static CXLRetCode cmd_dcd_get_dyn_cap_ext_list(const struct cxl_cmd *cmd,
|
||
uint8_t *payload_in,
|
||
size_t len_in,
|
||
uint8_t *payload_out,
|
||
size_t *len_out,
|
||
CXLCCI *cci)
|
||
{
|
||
CXLType3Dev *ct3d = CXL_TYPE3(cci->d);
|
||
struct {
|
||
uint32_t extent_cnt;
|
||
uint32_t start_extent_id;
|
||
} QEMU_PACKED *in = (void *)payload_in;
|
||
struct {
|
||
uint32_t count;
|
||
uint32_t total_extents;
|
||
uint32_t generation_num;
|
||
uint8_t rsvd[4];
|
||
CXLDCExtentRaw records[];
|
||
} QEMU_PACKED *out = (void *)payload_out;
|
||
uint32_t start_extent_id = in->start_extent_id;
|
||
CXLDCExtentList *extent_list = &ct3d->dc.extents;
|
||
uint16_t record_count = 0, i = 0, record_done = 0;
|
||
uint16_t out_pl_len, size;
|
||
CXLDCExtent *ent;
|
||
|
||
if (start_extent_id > ct3d->dc.total_extent_count) {
|
||
return CXL_MBOX_INVALID_INPUT;
|
||
}
|
||
|
||
record_count = MIN(in->extent_cnt,
|
||
ct3d->dc.total_extent_count - start_extent_id);
|
||
size = CXL_MAILBOX_MAX_PAYLOAD_SIZE - sizeof(*out);
|
||
record_count = MIN(record_count, size / sizeof(out->records[0]));
|
||
out_pl_len = sizeof(*out) + record_count * sizeof(out->records[0]);
|
||
|
||
stl_le_p(&out->count, record_count);
|
||
stl_le_p(&out->total_extents, ct3d->dc.total_extent_count);
|
||
stl_le_p(&out->generation_num, ct3d->dc.ext_list_gen_seq);
|
||
|
||
if (record_count > 0) {
|
||
CXLDCExtentRaw *out_rec = &out->records[record_done];
|
||
|
||
QTAILQ_FOREACH(ent, extent_list, node) {
|
||
if (i++ < start_extent_id) {
|
||
continue;
|
||
}
|
||
stq_le_p(&out_rec->start_dpa, ent->start_dpa);
|
||
stq_le_p(&out_rec->len, ent->len);
|
||
memcpy(&out_rec->tag, ent->tag, 0x10);
|
||
stw_le_p(&out_rec->shared_seq, ent->shared_seq);
|
||
|
||
record_done++;
|
||
if (record_done == record_count) {
|
||
break;
|
||
}
|
||
}
|
||
}
|
||
|
||
*len_out = out_pl_len;
|
||
return CXL_MBOX_SUCCESS;
|
||
}
|
||
|
||
/*
|
||
* Check whether any bit between addr[nr, nr+size) is set,
|
||
* return true if any bit is set, otherwise return false
|
||
*/
|
||
bool test_any_bits_set(const unsigned long *addr, unsigned long nr,
|
||
unsigned long size)
|
||
{
|
||
unsigned long res = find_next_bit(addr, size + nr, nr);
|
||
|
||
return res < nr + size;
|
||
}
|
||
|
||
CXLDCRegion *cxl_find_dc_region(CXLType3Dev *ct3d, uint64_t dpa, uint64_t len)
|
||
{
|
||
int i;
|
||
CXLDCRegion *region = &ct3d->dc.regions[0];
|
||
|
||
if (dpa < region->base ||
|
||
dpa >= region->base + ct3d->dc.total_capacity) {
|
||
return NULL;
|
||
}
|
||
|
||
/*
|
||
* CXL r3.1 section 9.13.3: Dynamic Capacity Device (DCD)
|
||
*
|
||
* Regions are used in increasing-DPA order, with Region 0 being used for
|
||
* the lowest DPA of Dynamic Capacity and Region 7 for the highest DPA.
|
||
* So check from the last region to find where the dpa belongs. Extents that
|
||
* cross multiple regions are not allowed.
|
||
*/
|
||
for (i = ct3d->dc.num_regions - 1; i >= 0; i--) {
|
||
region = &ct3d->dc.regions[i];
|
||
if (dpa >= region->base) {
|
||
if (dpa + len > region->base + region->len) {
|
||
return NULL;
|
||
}
|
||
return region;
|
||
}
|
||
}
|
||
|
||
return NULL;
|
||
}
|
||
|
||
void cxl_insert_extent_to_extent_list(CXLDCExtentList *list,
|
||
uint64_t dpa,
|
||
uint64_t len,
|
||
uint8_t *tag,
|
||
uint16_t shared_seq)
|
||
{
|
||
CXLDCExtent *extent;
|
||
|
||
extent = g_new0(CXLDCExtent, 1);
|
||
extent->start_dpa = dpa;
|
||
extent->len = len;
|
||
if (tag) {
|
||
memcpy(extent->tag, tag, 0x10);
|
||
}
|
||
extent->shared_seq = shared_seq;
|
||
|
||
QTAILQ_INSERT_TAIL(list, extent, node);
|
||
}
|
||
|
||
void cxl_remove_extent_from_extent_list(CXLDCExtentList *list,
|
||
CXLDCExtent *extent)
|
||
{
|
||
QTAILQ_REMOVE(list, extent, node);
|
||
g_free(extent);
|
||
}
|
||
|
||
/*
|
||
* Add a new extent to the extent "group" if group exists;
|
||
* otherwise, create a new group
|
||
* Return value: the extent group where the extent is inserted.
|
||
*/
|
||
CXLDCExtentGroup *cxl_insert_extent_to_extent_group(CXLDCExtentGroup *group,
|
||
uint64_t dpa,
|
||
uint64_t len,
|
||
uint8_t *tag,
|
||
uint16_t shared_seq)
|
||
{
|
||
if (!group) {
|
||
group = g_new0(CXLDCExtentGroup, 1);
|
||
QTAILQ_INIT(&group->list);
|
||
}
|
||
cxl_insert_extent_to_extent_list(&group->list, dpa, len,
|
||
tag, shared_seq);
|
||
return group;
|
||
}
|
||
|
||
void cxl_extent_group_list_insert_tail(CXLDCExtentGroupList *list,
|
||
CXLDCExtentGroup *group)
|
||
{
|
||
QTAILQ_INSERT_TAIL(list, group, node);
|
||
}
|
||
|
||
void cxl_extent_group_list_delete_front(CXLDCExtentGroupList *list)
|
||
{
|
||
CXLDCExtent *ent, *ent_next;
|
||
CXLDCExtentGroup *group = QTAILQ_FIRST(list);
|
||
|
||
QTAILQ_REMOVE(list, group, node);
|
||
QTAILQ_FOREACH_SAFE(ent, &group->list, node, ent_next) {
|
||
cxl_remove_extent_from_extent_list(&group->list, ent);
|
||
}
|
||
g_free(group);
|
||
}
|
||
|
||
/*
|
||
* CXL r3.1 Table 8-168: Add Dynamic Capacity Response Input Payload
|
||
* CXL r3.1 Table 8-170: Release Dynamic Capacity Input Payload
|
||
*/
|
||
typedef struct CXLUpdateDCExtentListInPl {
|
||
uint32_t num_entries_updated;
|
||
uint8_t flags;
|
||
uint8_t rsvd[3];
|
||
/* CXL r3.1 Table 8-169: Updated Extent */
|
||
struct {
|
||
uint64_t start_dpa;
|
||
uint64_t len;
|
||
uint8_t rsvd[8];
|
||
} QEMU_PACKED updated_entries[];
|
||
} QEMU_PACKED CXLUpdateDCExtentListInPl;
|
||
|
||
/*
|
||
* For the extents in the extent list to operate, check whether they are valid
|
||
* 1. The extent should be in the range of a valid DC region;
|
||
* 2. The extent should not cross multiple regions;
|
||
* 3. The start DPA and the length of the extent should align with the block
|
||
* size of the region;
|
||
* 4. The address range of multiple extents in the list should not overlap.
|
||
*/
|
||
static CXLRetCode cxl_detect_malformed_extent_list(CXLType3Dev *ct3d,
|
||
const CXLUpdateDCExtentListInPl *in)
|
||
{
|
||
uint64_t min_block_size = UINT64_MAX;
|
||
CXLDCRegion *region;
|
||
CXLDCRegion *lastregion = &ct3d->dc.regions[ct3d->dc.num_regions - 1];
|
||
g_autofree unsigned long *blk_bitmap = NULL;
|
||
uint64_t dpa, len;
|
||
uint32_t i;
|
||
|
||
for (i = 0; i < ct3d->dc.num_regions; i++) {
|
||
region = &ct3d->dc.regions[i];
|
||
min_block_size = MIN(min_block_size, region->block_size);
|
||
}
|
||
|
||
blk_bitmap = bitmap_new((lastregion->base + lastregion->len -
|
||
ct3d->dc.regions[0].base) / min_block_size);
|
||
|
||
for (i = 0; i < in->num_entries_updated; i++) {
|
||
dpa = in->updated_entries[i].start_dpa;
|
||
len = in->updated_entries[i].len;
|
||
|
||
region = cxl_find_dc_region(ct3d, dpa, len);
|
||
if (!region) {
|
||
return CXL_MBOX_INVALID_PA;
|
||
}
|
||
|
||
dpa -= ct3d->dc.regions[0].base;
|
||
if (dpa % region->block_size || len % region->block_size) {
|
||
return CXL_MBOX_INVALID_EXTENT_LIST;
|
||
}
|
||
/* the dpa range already covered by some other extents in the list */
|
||
if (test_any_bits_set(blk_bitmap, dpa / min_block_size,
|
||
len / min_block_size)) {
|
||
return CXL_MBOX_INVALID_EXTENT_LIST;
|
||
}
|
||
bitmap_set(blk_bitmap, dpa / min_block_size, len / min_block_size);
|
||
}
|
||
|
||
return CXL_MBOX_SUCCESS;
|
||
}
|
||
|
||
static CXLRetCode cxl_dcd_add_dyn_cap_rsp_dry_run(CXLType3Dev *ct3d,
|
||
const CXLUpdateDCExtentListInPl *in)
|
||
{
|
||
uint32_t i;
|
||
CXLDCExtent *ent;
|
||
CXLDCExtentGroup *ext_group;
|
||
uint64_t dpa, len;
|
||
Range range1, range2;
|
||
|
||
for (i = 0; i < in->num_entries_updated; i++) {
|
||
dpa = in->updated_entries[i].start_dpa;
|
||
len = in->updated_entries[i].len;
|
||
|
||
range_init_nofail(&range1, dpa, len);
|
||
|
||
/*
|
||
* The host-accepted DPA range must be contained by the first extent
|
||
* group in the pending list
|
||
*/
|
||
ext_group = QTAILQ_FIRST(&ct3d->dc.extents_pending);
|
||
if (!cxl_extents_contains_dpa_range(&ext_group->list, dpa, len)) {
|
||
return CXL_MBOX_INVALID_PA;
|
||
}
|
||
|
||
/* to-be-added range should not overlap with range already accepted */
|
||
QTAILQ_FOREACH(ent, &ct3d->dc.extents, node) {
|
||
range_init_nofail(&range2, ent->start_dpa, ent->len);
|
||
if (range_overlaps_range(&range1, &range2)) {
|
||
return CXL_MBOX_INVALID_PA;
|
||
}
|
||
}
|
||
}
|
||
return CXL_MBOX_SUCCESS;
|
||
}
|
||
|
||
/*
|
||
* CXL r3.1 section 8.2.9.9.9.3: Add Dynamic Capacity Response (Opcode 4802h)
|
||
* An extent is added to the extent list and becomes usable only after the
|
||
* response is processed successfully.
|
||
*/
|
||
static CXLRetCode cmd_dcd_add_dyn_cap_rsp(const struct cxl_cmd *cmd,
|
||
uint8_t *payload_in,
|
||
size_t len_in,
|
||
uint8_t *payload_out,
|
||
size_t *len_out,
|
||
CXLCCI *cci)
|
||
{
|
||
CXLUpdateDCExtentListInPl *in = (void *)payload_in;
|
||
CXLType3Dev *ct3d = CXL_TYPE3(cci->d);
|
||
CXLDCExtentList *extent_list = &ct3d->dc.extents;
|
||
uint32_t i;
|
||
uint64_t dpa, len;
|
||
CXLRetCode ret;
|
||
|
||
if (in->num_entries_updated == 0) {
|
||
cxl_extent_group_list_delete_front(&ct3d->dc.extents_pending);
|
||
return CXL_MBOX_SUCCESS;
|
||
}
|
||
|
||
/* Adding extents causes exceeding device's extent tracking ability. */
|
||
if (in->num_entries_updated + ct3d->dc.total_extent_count >
|
||
CXL_NUM_EXTENTS_SUPPORTED) {
|
||
return CXL_MBOX_RESOURCES_EXHAUSTED;
|
||
}
|
||
|
||
ret = cxl_detect_malformed_extent_list(ct3d, in);
|
||
if (ret != CXL_MBOX_SUCCESS) {
|
||
return ret;
|
||
}
|
||
|
||
ret = cxl_dcd_add_dyn_cap_rsp_dry_run(ct3d, in);
|
||
if (ret != CXL_MBOX_SUCCESS) {
|
||
return ret;
|
||
}
|
||
|
||
for (i = 0; i < in->num_entries_updated; i++) {
|
||
dpa = in->updated_entries[i].start_dpa;
|
||
len = in->updated_entries[i].len;
|
||
|
||
cxl_insert_extent_to_extent_list(extent_list, dpa, len, NULL, 0);
|
||
ct3d->dc.total_extent_count += 1;
|
||
ct3_set_region_block_backed(ct3d, dpa, len);
|
||
}
|
||
/* Remove the first extent group in the pending list */
|
||
cxl_extent_group_list_delete_front(&ct3d->dc.extents_pending);
|
||
|
||
return CXL_MBOX_SUCCESS;
|
||
}
|
||
|
||
/*
|
||
* Copy extent list from src to dst
|
||
* Return value: number of extents copied
|
||
*/
|
||
static uint32_t copy_extent_list(CXLDCExtentList *dst,
|
||
const CXLDCExtentList *src)
|
||
{
|
||
uint32_t cnt = 0;
|
||
CXLDCExtent *ent;
|
||
|
||
if (!dst || !src) {
|
||
return 0;
|
||
}
|
||
|
||
QTAILQ_FOREACH(ent, src, node) {
|
||
cxl_insert_extent_to_extent_list(dst, ent->start_dpa, ent->len,
|
||
ent->tag, ent->shared_seq);
|
||
cnt++;
|
||
}
|
||
return cnt;
|
||
}
|
||
|
||
static CXLRetCode cxl_dc_extent_release_dry_run(CXLType3Dev *ct3d,
|
||
const CXLUpdateDCExtentListInPl *in, CXLDCExtentList *updated_list,
|
||
uint32_t *updated_list_size)
|
||
{
|
||
CXLDCExtent *ent, *ent_next;
|
||
uint64_t dpa, len;
|
||
uint32_t i;
|
||
int cnt_delta = 0;
|
||
CXLRetCode ret = CXL_MBOX_SUCCESS;
|
||
|
||
QTAILQ_INIT(updated_list);
|
||
copy_extent_list(updated_list, &ct3d->dc.extents);
|
||
|
||
for (i = 0; i < in->num_entries_updated; i++) {
|
||
Range range;
|
||
|
||
dpa = in->updated_entries[i].start_dpa;
|
||
len = in->updated_entries[i].len;
|
||
|
||
/* Check if the DPA range is not fully backed with valid extents */
|
||
if (!ct3_test_region_block_backed(ct3d, dpa, len)) {
|
||
ret = CXL_MBOX_INVALID_PA;
|
||
goto free_and_exit;
|
||
}
|
||
|
||
/* After this point, extent overflow is the only error can happen */
|
||
while (len > 0) {
|
||
QTAILQ_FOREACH(ent, updated_list, node) {
|
||
range_init_nofail(&range, ent->start_dpa, ent->len);
|
||
|
||
if (range_contains(&range, dpa)) {
|
||
uint64_t len1, len2 = 0, len_done = 0;
|
||
uint64_t ent_start_dpa = ent->start_dpa;
|
||
uint64_t ent_len = ent->len;
|
||
|
||
len1 = dpa - ent->start_dpa;
|
||
/* Found the extent or the subset of an existing extent */
|
||
if (range_contains(&range, dpa + len - 1)) {
|
||
len2 = ent_start_dpa + ent_len - dpa - len;
|
||
} else {
|
||
dpa = ent_start_dpa + ent_len;
|
||
}
|
||
len_done = ent_len - len1 - len2;
|
||
|
||
cxl_remove_extent_from_extent_list(updated_list, ent);
|
||
cnt_delta--;
|
||
|
||
if (len1) {
|
||
cxl_insert_extent_to_extent_list(updated_list,
|
||
ent_start_dpa,
|
||
len1, NULL, 0);
|
||
cnt_delta++;
|
||
}
|
||
if (len2) {
|
||
cxl_insert_extent_to_extent_list(updated_list,
|
||
dpa + len,
|
||
len2, NULL, 0);
|
||
cnt_delta++;
|
||
}
|
||
|
||
if (cnt_delta + ct3d->dc.total_extent_count >
|
||
CXL_NUM_EXTENTS_SUPPORTED) {
|
||
ret = CXL_MBOX_RESOURCES_EXHAUSTED;
|
||
goto free_and_exit;
|
||
}
|
||
|
||
len -= len_done;
|
||
break;
|
||
}
|
||
}
|
||
}
|
||
}
|
||
free_and_exit:
|
||
if (ret != CXL_MBOX_SUCCESS) {
|
||
QTAILQ_FOREACH_SAFE(ent, updated_list, node, ent_next) {
|
||
cxl_remove_extent_from_extent_list(updated_list, ent);
|
||
}
|
||
*updated_list_size = 0;
|
||
} else {
|
||
*updated_list_size = ct3d->dc.total_extent_count + cnt_delta;
|
||
}
|
||
|
||
return ret;
|
||
}
|
||
|
||
/*
|
||
* CXL r3.1 section 8.2.9.9.9.4: Release Dynamic Capacity (Opcode 4803h)
|
||
*/
|
||
static CXLRetCode cmd_dcd_release_dyn_cap(const struct cxl_cmd *cmd,
|
||
uint8_t *payload_in,
|
||
size_t len_in,
|
||
uint8_t *payload_out,
|
||
size_t *len_out,
|
||
CXLCCI *cci)
|
||
{
|
||
CXLUpdateDCExtentListInPl *in = (void *)payload_in;
|
||
CXLType3Dev *ct3d = CXL_TYPE3(cci->d);
|
||
CXLDCExtentList updated_list;
|
||
CXLDCExtent *ent, *ent_next;
|
||
uint32_t updated_list_size;
|
||
CXLRetCode ret;
|
||
|
||
if (in->num_entries_updated == 0) {
|
||
return CXL_MBOX_INVALID_INPUT;
|
||
}
|
||
|
||
ret = cxl_detect_malformed_extent_list(ct3d, in);
|
||
if (ret != CXL_MBOX_SUCCESS) {
|
||
return ret;
|
||
}
|
||
|
||
ret = cxl_dc_extent_release_dry_run(ct3d, in, &updated_list,
|
||
&updated_list_size);
|
||
if (ret != CXL_MBOX_SUCCESS) {
|
||
return ret;
|
||
}
|
||
|
||
/*
|
||
* If the dry run release passes, the returned updated_list will
|
||
* be the updated extent list and we just need to clear the extents
|
||
* in the accepted list and copy extents in the updated_list to accepted
|
||
* list and update the extent count;
|
||
*/
|
||
QTAILQ_FOREACH_SAFE(ent, &ct3d->dc.extents, node, ent_next) {
|
||
ct3_clear_region_block_backed(ct3d, ent->start_dpa, ent->len);
|
||
cxl_remove_extent_from_extent_list(&ct3d->dc.extents, ent);
|
||
}
|
||
copy_extent_list(&ct3d->dc.extents, &updated_list);
|
||
QTAILQ_FOREACH_SAFE(ent, &updated_list, node, ent_next) {
|
||
ct3_set_region_block_backed(ct3d, ent->start_dpa, ent->len);
|
||
cxl_remove_extent_from_extent_list(&updated_list, ent);
|
||
}
|
||
ct3d->dc.total_extent_count = updated_list_size;
|
||
|
||
return CXL_MBOX_SUCCESS;
|
||
}
|
||
|
||
static const struct cxl_cmd cxl_cmd_set[256][256] = {
|
||
[EVENTS][GET_RECORDS] = { "EVENTS_GET_RECORDS",
|
||
cmd_events_get_records, 1, 0 },
|
||
[EVENTS][CLEAR_RECORDS] = { "EVENTS_CLEAR_RECORDS",
|
||
cmd_events_clear_records, ~0, CXL_MBOX_IMMEDIATE_LOG_CHANGE },
|
||
[EVENTS][GET_INTERRUPT_POLICY] = { "EVENTS_GET_INTERRUPT_POLICY",
|
||
cmd_events_get_interrupt_policy, 0, 0 },
|
||
[EVENTS][SET_INTERRUPT_POLICY] = { "EVENTS_SET_INTERRUPT_POLICY",
|
||
cmd_events_set_interrupt_policy,
|
||
~0, CXL_MBOX_IMMEDIATE_CONFIG_CHANGE },
|
||
[FIRMWARE_UPDATE][GET_INFO] = { "FIRMWARE_UPDATE_GET_INFO",
|
||
cmd_firmware_update_get_info, 0, 0 },
|
||
[FIRMWARE_UPDATE][TRANSFER] = { "FIRMWARE_UPDATE_TRANSFER",
|
||
cmd_firmware_update_transfer, ~0, CXL_MBOX_BACKGROUND_OPERATION },
|
||
[FIRMWARE_UPDATE][ACTIVATE] = { "FIRMWARE_UPDATE_ACTIVATE",
|
||
cmd_firmware_update_activate, 2, CXL_MBOX_BACKGROUND_OPERATION },
|
||
[TIMESTAMP][GET] = { "TIMESTAMP_GET", cmd_timestamp_get, 0, 0 },
|
||
[TIMESTAMP][SET] = { "TIMESTAMP_SET", cmd_timestamp_set,
|
||
8, CXL_MBOX_IMMEDIATE_POLICY_CHANGE },
|
||
[LOGS][GET_SUPPORTED] = { "LOGS_GET_SUPPORTED", cmd_logs_get_supported,
|
||
0, 0 },
|
||
[LOGS][GET_LOG] = { "LOGS_GET_LOG", cmd_logs_get_log, 0x18, 0 },
|
||
[FEATURES][GET_SUPPORTED] = { "FEATURES_GET_SUPPORTED",
|
||
cmd_features_get_supported, 0x8, 0 },
|
||
[FEATURES][GET_FEATURE] = { "FEATURES_GET_FEATURE",
|
||
cmd_features_get_feature, 0x15, 0 },
|
||
[FEATURES][SET_FEATURE] = { "FEATURES_SET_FEATURE",
|
||
cmd_features_set_feature,
|
||
~0,
|
||
(CXL_MBOX_IMMEDIATE_CONFIG_CHANGE |
|
||
CXL_MBOX_IMMEDIATE_DATA_CHANGE |
|
||
CXL_MBOX_IMMEDIATE_POLICY_CHANGE |
|
||
CXL_MBOX_IMMEDIATE_LOG_CHANGE |
|
||
CXL_MBOX_SECURITY_STATE_CHANGE)},
|
||
[IDENTIFY][MEMORY_DEVICE] = { "IDENTIFY_MEMORY_DEVICE",
|
||
cmd_identify_memory_device, 0, 0 },
|
||
[CCLS][GET_PARTITION_INFO] = { "CCLS_GET_PARTITION_INFO",
|
||
cmd_ccls_get_partition_info, 0, 0 },
|
||
[CCLS][GET_LSA] = { "CCLS_GET_LSA", cmd_ccls_get_lsa, 8, 0 },
|
||
[CCLS][SET_LSA] = { "CCLS_SET_LSA", cmd_ccls_set_lsa,
|
||
~0, CXL_MBOX_IMMEDIATE_CONFIG_CHANGE | CXL_MBOX_IMMEDIATE_DATA_CHANGE },
|
||
[SANITIZE][OVERWRITE] = { "SANITIZE_OVERWRITE", cmd_sanitize_overwrite, 0,
|
||
(CXL_MBOX_IMMEDIATE_DATA_CHANGE |
|
||
CXL_MBOX_SECURITY_STATE_CHANGE |
|
||
CXL_MBOX_BACKGROUND_OPERATION)},
|
||
[PERSISTENT_MEM][GET_SECURITY_STATE] = { "GET_SECURITY_STATE",
|
||
cmd_get_security_state, 0, 0 },
|
||
[MEDIA_AND_POISON][GET_POISON_LIST] = { "MEDIA_AND_POISON_GET_POISON_LIST",
|
||
cmd_media_get_poison_list, 16, 0 },
|
||
[MEDIA_AND_POISON][INJECT_POISON] = { "MEDIA_AND_POISON_INJECT_POISON",
|
||
cmd_media_inject_poison, 8, 0 },
|
||
[MEDIA_AND_POISON][CLEAR_POISON] = { "MEDIA_AND_POISON_CLEAR_POISON",
|
||
cmd_media_clear_poison, 72, 0 },
|
||
[MEDIA_AND_POISON][GET_SCAN_MEDIA_CAPABILITIES] = {
|
||
"MEDIA_AND_POISON_GET_SCAN_MEDIA_CAPABILITIES",
|
||
cmd_media_get_scan_media_capabilities, 16, 0 },
|
||
[MEDIA_AND_POISON][SCAN_MEDIA] = { "MEDIA_AND_POISON_SCAN_MEDIA",
|
||
cmd_media_scan_media, 17, CXL_MBOX_BACKGROUND_OPERATION },
|
||
[MEDIA_AND_POISON][GET_SCAN_MEDIA_RESULTS] = {
|
||
"MEDIA_AND_POISON_GET_SCAN_MEDIA_RESULTS",
|
||
cmd_media_get_scan_media_results, 0, 0 },
|
||
};
|
||
|
||
static const struct cxl_cmd cxl_cmd_set_dcd[256][256] = {
|
||
[DCD_CONFIG][GET_DC_CONFIG] = { "DCD_GET_DC_CONFIG",
|
||
cmd_dcd_get_dyn_cap_config, 2, 0 },
|
||
[DCD_CONFIG][GET_DYN_CAP_EXT_LIST] = {
|
||
"DCD_GET_DYNAMIC_CAPACITY_EXTENT_LIST", cmd_dcd_get_dyn_cap_ext_list,
|
||
8, 0 },
|
||
[DCD_CONFIG][ADD_DYN_CAP_RSP] = {
|
||
"DCD_ADD_DYNAMIC_CAPACITY_RESPONSE", cmd_dcd_add_dyn_cap_rsp,
|
||
~0, CXL_MBOX_IMMEDIATE_DATA_CHANGE },
|
||
[DCD_CONFIG][RELEASE_DYN_CAP] = {
|
||
"DCD_RELEASE_DYNAMIC_CAPACITY", cmd_dcd_release_dyn_cap,
|
||
~0, CXL_MBOX_IMMEDIATE_DATA_CHANGE },
|
||
};
|
||
|
||
static const struct cxl_cmd cxl_cmd_set_sw[256][256] = {
|
||
[INFOSTAT][IS_IDENTIFY] = { "IDENTIFY", cmd_infostat_identify, 0, 0 },
|
||
[INFOSTAT][BACKGROUND_OPERATION_STATUS] = { "BACKGROUND_OPERATION_STATUS",
|
||
cmd_infostat_bg_op_sts, 0, 0 },
|
||
[TIMESTAMP][GET] = { "TIMESTAMP_GET", cmd_timestamp_get, 0, 0 },
|
||
[TIMESTAMP][SET] = { "TIMESTAMP_SET", cmd_timestamp_set, 8,
|
||
CXL_MBOX_IMMEDIATE_POLICY_CHANGE },
|
||
[LOGS][GET_SUPPORTED] = { "LOGS_GET_SUPPORTED", cmd_logs_get_supported, 0,
|
||
0 },
|
||
[LOGS][GET_LOG] = { "LOGS_GET_LOG", cmd_logs_get_log, 0x18, 0 },
|
||
[PHYSICAL_SWITCH][IDENTIFY_SWITCH_DEVICE] = { "IDENTIFY_SWITCH_DEVICE",
|
||
cmd_identify_switch_device, 0, 0 },
|
||
[PHYSICAL_SWITCH][GET_PHYSICAL_PORT_STATE] = { "SWITCH_PHYSICAL_PORT_STATS",
|
||
cmd_get_physical_port_state, ~0, 0 },
|
||
[TUNNEL][MANAGEMENT_COMMAND] = { "TUNNEL_MANAGEMENT_COMMAND",
|
||
cmd_tunnel_management_cmd, ~0, 0 },
|
||
};
|
||
|
||
/*
|
||
* While the command is executing in the background, the device should
|
||
* update the percentage complete in the Background Command Status Register
|
||
* at least once per second.
|
||
*/
|
||
|
||
#define CXL_MBOX_BG_UPDATE_FREQ 1000UL
|
||
|
||
int cxl_process_cci_message(CXLCCI *cci, uint8_t set, uint8_t cmd,
|
||
size_t len_in, uint8_t *pl_in, size_t *len_out,
|
||
uint8_t *pl_out, bool *bg_started)
|
||
{
|
||
int ret;
|
||
const struct cxl_cmd *cxl_cmd;
|
||
opcode_handler h;
|
||
CXLDeviceState *cxl_dstate;
|
||
|
||
*len_out = 0;
|
||
cxl_cmd = &cci->cxl_cmd_set[set][cmd];
|
||
h = cxl_cmd->handler;
|
||
if (!h) {
|
||
qemu_log_mask(LOG_UNIMP, "Command %04xh not implemented\n",
|
||
set << 8 | cmd);
|
||
return CXL_MBOX_UNSUPPORTED;
|
||
}
|
||
|
||
if (len_in != cxl_cmd->in && cxl_cmd->in != ~0) {
|
||
return CXL_MBOX_INVALID_PAYLOAD_LENGTH;
|
||
}
|
||
|
||
/* Only one bg command at a time */
|
||
if ((cxl_cmd->effect & CXL_MBOX_BACKGROUND_OPERATION) &&
|
||
cci->bg.runtime > 0) {
|
||
return CXL_MBOX_BUSY;
|
||
}
|
||
|
||
/* forbid any selected commands while the media is disabled */
|
||
if (object_dynamic_cast(OBJECT(cci->d), TYPE_CXL_TYPE3)) {
|
||
cxl_dstate = &CXL_TYPE3(cci->d)->cxl_dstate;
|
||
|
||
if (cxl_dev_media_disabled(cxl_dstate)) {
|
||
if (h == cmd_events_get_records ||
|
||
h == cmd_ccls_get_partition_info ||
|
||
h == cmd_ccls_set_lsa ||
|
||
h == cmd_ccls_get_lsa ||
|
||
h == cmd_logs_get_log ||
|
||
h == cmd_media_get_poison_list ||
|
||
h == cmd_media_inject_poison ||
|
||
h == cmd_media_clear_poison ||
|
||
h == cmd_sanitize_overwrite ||
|
||
h == cmd_firmware_update_transfer ||
|
||
h == cmd_firmware_update_activate) {
|
||
return CXL_MBOX_MEDIA_DISABLED;
|
||
}
|
||
}
|
||
}
|
||
|
||
ret = (*h)(cxl_cmd, pl_in, len_in, pl_out, len_out, cci);
|
||
if ((cxl_cmd->effect & CXL_MBOX_BACKGROUND_OPERATION) &&
|
||
ret == CXL_MBOX_BG_STARTED) {
|
||
*bg_started = true;
|
||
} else {
|
||
*bg_started = false;
|
||
}
|
||
|
||
/* Set bg and the return code */
|
||
if (*bg_started) {
|
||
uint64_t now;
|
||
|
||
cci->bg.opcode = (set << 8) | cmd;
|
||
|
||
cci->bg.complete_pct = 0;
|
||
cci->bg.ret_code = 0;
|
||
|
||
now = qemu_clock_get_ms(QEMU_CLOCK_VIRTUAL);
|
||
cci->bg.starttime = now;
|
||
timer_mod(cci->bg.timer, now + CXL_MBOX_BG_UPDATE_FREQ);
|
||
}
|
||
|
||
return ret;
|
||
}
|
||
|
||
static void bg_timercb(void *opaque)
|
||
{
|
||
CXLCCI *cci = opaque;
|
||
uint64_t now = qemu_clock_get_ms(QEMU_CLOCK_VIRTUAL);
|
||
uint64_t total_time = cci->bg.starttime + cci->bg.runtime;
|
||
|
||
assert(cci->bg.runtime > 0);
|
||
|
||
if (now >= total_time) { /* we are done */
|
||
uint16_t ret = CXL_MBOX_SUCCESS;
|
||
|
||
cci->bg.complete_pct = 100;
|
||
cci->bg.ret_code = ret;
|
||
switch (cci->bg.opcode) {
|
||
case 0x0201: /* fw transfer */
|
||
__do_firmware_xfer(cci);
|
||
break;
|
||
case 0x4400: /* sanitize */
|
||
{
|
||
CXLType3Dev *ct3d = CXL_TYPE3(cci->d);
|
||
|
||
__do_sanitization(ct3d);
|
||
cxl_dev_enable_media(&ct3d->cxl_dstate);
|
||
}
|
||
break;
|
||
case 0x4304: /* scan media */
|
||
{
|
||
CXLType3Dev *ct3d = CXL_TYPE3(cci->d);
|
||
|
||
__do_scan_media(ct3d);
|
||
break;
|
||
}
|
||
default:
|
||
__builtin_unreachable();
|
||
break;
|
||
}
|
||
} else {
|
||
/* estimate only */
|
||
cci->bg.complete_pct = 100 * now / total_time;
|
||
timer_mod(cci->bg.timer, now + CXL_MBOX_BG_UPDATE_FREQ);
|
||
}
|
||
|
||
if (cci->bg.complete_pct == 100) {
|
||
/* TODO: generalize to switch CCI */
|
||
CXLType3Dev *ct3d = CXL_TYPE3(cci->d);
|
||
CXLDeviceState *cxl_dstate = &ct3d->cxl_dstate;
|
||
PCIDevice *pdev = PCI_DEVICE(cci->d);
|
||
|
||
cci->bg.starttime = 0;
|
||
/* registers are updated, allow new bg-capable cmds */
|
||
cci->bg.runtime = 0;
|
||
|
||
if (msix_enabled(pdev)) {
|
||
msix_notify(pdev, cxl_dstate->mbox_msi_n);
|
||
} else if (msi_enabled(pdev)) {
|
||
msi_notify(pdev, cxl_dstate->mbox_msi_n);
|
||
}
|
||
}
|
||
}
|
||
|
||
static void cxl_rebuild_cel(CXLCCI *cci)
|
||
{
|
||
cci->cel_size = 0; /* Reset for a fresh build */
|
||
for (int set = 0; set < 256; set++) {
|
||
for (int cmd = 0; cmd < 256; cmd++) {
|
||
if (cci->cxl_cmd_set[set][cmd].handler) {
|
||
const struct cxl_cmd *c = &cci->cxl_cmd_set[set][cmd];
|
||
struct cel_log *log =
|
||
&cci->cel_log[cci->cel_size];
|
||
|
||
log->opcode = (set << 8) | cmd;
|
||
log->effect = c->effect;
|
||
cci->cel_size++;
|
||
}
|
||
}
|
||
}
|
||
}
|
||
|
||
void cxl_init_cci(CXLCCI *cci, size_t payload_max)
|
||
{
|
||
cci->payload_max = payload_max;
|
||
cxl_rebuild_cel(cci);
|
||
|
||
cci->bg.complete_pct = 0;
|
||
cci->bg.starttime = 0;
|
||
cci->bg.runtime = 0;
|
||
cci->bg.timer = timer_new_ms(QEMU_CLOCK_VIRTUAL,
|
||
bg_timercb, cci);
|
||
|
||
memset(&cci->fw, 0, sizeof(cci->fw));
|
||
cci->fw.active_slot = 1;
|
||
cci->fw.slot[cci->fw.active_slot - 1] = true;
|
||
}
|
||
|
||
static void cxl_copy_cci_commands(CXLCCI *cci, const struct cxl_cmd (*cxl_cmds)[256])
|
||
{
|
||
for (int set = 0; set < 256; set++) {
|
||
for (int cmd = 0; cmd < 256; cmd++) {
|
||
if (cxl_cmds[set][cmd].handler) {
|
||
cci->cxl_cmd_set[set][cmd] = cxl_cmds[set][cmd];
|
||
}
|
||
}
|
||
}
|
||
}
|
||
|
||
void cxl_add_cci_commands(CXLCCI *cci, const struct cxl_cmd (*cxl_cmd_set)[256],
|
||
size_t payload_max)
|
||
{
|
||
cci->payload_max = MAX(payload_max, cci->payload_max);
|
||
cxl_copy_cci_commands(cci, cxl_cmd_set);
|
||
cxl_rebuild_cel(cci);
|
||
}
|
||
|
||
void cxl_initialize_mailbox_swcci(CXLCCI *cci, DeviceState *intf,
|
||
DeviceState *d, size_t payload_max)
|
||
{
|
||
cxl_copy_cci_commands(cci, cxl_cmd_set_sw);
|
||
cci->d = d;
|
||
cci->intf = intf;
|
||
cxl_init_cci(cci, payload_max);
|
||
}
|
||
|
||
void cxl_initialize_mailbox_t3(CXLCCI *cci, DeviceState *d, size_t payload_max)
|
||
{
|
||
CXLType3Dev *ct3d = CXL_TYPE3(d);
|
||
|
||
cxl_copy_cci_commands(cci, cxl_cmd_set);
|
||
if (ct3d->dc.num_regions) {
|
||
cxl_copy_cci_commands(cci, cxl_cmd_set_dcd);
|
||
}
|
||
cci->d = d;
|
||
|
||
/* No separation for PCI MB as protocol handled in PCI device */
|
||
cci->intf = d;
|
||
cxl_init_cci(cci, payload_max);
|
||
}
|
||
|
||
static const struct cxl_cmd cxl_cmd_set_t3_ld[256][256] = {
|
||
[INFOSTAT][IS_IDENTIFY] = { "IDENTIFY", cmd_infostat_identify, 0, 0 },
|
||
[LOGS][GET_SUPPORTED] = { "LOGS_GET_SUPPORTED", cmd_logs_get_supported, 0,
|
||
0 },
|
||
[LOGS][GET_LOG] = { "LOGS_GET_LOG", cmd_logs_get_log, 0x18, 0 },
|
||
};
|
||
|
||
void cxl_initialize_t3_ld_cci(CXLCCI *cci, DeviceState *d, DeviceState *intf,
|
||
size_t payload_max)
|
||
{
|
||
cxl_copy_cci_commands(cci, cxl_cmd_set_t3_ld);
|
||
cci->d = d;
|
||
cci->intf = intf;
|
||
cxl_init_cci(cci, payload_max);
|
||
}
|
||
|
||
static const struct cxl_cmd cxl_cmd_set_t3_fm_owned_ld_mctp[256][256] = {
|
||
[INFOSTAT][IS_IDENTIFY] = { "IDENTIFY", cmd_infostat_identify, 0, 0},
|
||
[LOGS][GET_SUPPORTED] = { "LOGS_GET_SUPPORTED", cmd_logs_get_supported, 0,
|
||
0 },
|
||
[LOGS][GET_LOG] = { "LOGS_GET_LOG", cmd_logs_get_log, 0x18, 0 },
|
||
[TIMESTAMP][GET] = { "TIMESTAMP_GET", cmd_timestamp_get, 0, 0 },
|
||
[TUNNEL][MANAGEMENT_COMMAND] = { "TUNNEL_MANAGEMENT_COMMAND",
|
||
cmd_tunnel_management_cmd, ~0, 0 },
|
||
};
|
||
|
||
void cxl_initialize_t3_fm_owned_ld_mctpcci(CXLCCI *cci, DeviceState *d,
|
||
DeviceState *intf,
|
||
size_t payload_max)
|
||
{
|
||
cxl_copy_cci_commands(cci, cxl_cmd_set_t3_fm_owned_ld_mctp);
|
||
cci->d = d;
|
||
cci->intf = intf;
|
||
cxl_init_cci(cci, payload_max);
|
||
}
|