a27bd6c779
In my "build everything" tree, changing hw/qdev-properties.h triggers a recompile of some 2700 out of 6600 objects (not counting tests and objects that don't depend on qemu/osdep.h). Many places including hw/qdev-properties.h (directly or via hw/qdev.h) actually need only hw/qdev-core.h. Include hw/qdev-core.h there instead. hw/qdev.h is actually pointless: all it does is include hw/qdev-core.h and hw/qdev-properties.h, which in turn includes hw/qdev-core.h. Replace the remaining uses of hw/qdev.h by hw/qdev-properties.h. While there, delete a few superfluous inclusions of hw/qdev-core.h. Touching hw/qdev-properties.h now recompiles some 1200 objects. Cc: Paolo Bonzini <pbonzini@redhat.com> Cc: "Daniel P. Berrangé" <berrange@redhat.com> Cc: Eduardo Habkost <ehabkost@redhat.com> Signed-off-by: Markus Armbruster <armbru@redhat.com> Reviewed-by: Eduardo Habkost <ehabkost@redhat.com> Message-Id: <20190812052359.30071-22-armbru@redhat.com>
1543 lines
43 KiB
C
1543 lines
43 KiB
C
/*
|
|
* QEMU rocker switch emulation - PCI device
|
|
*
|
|
* Copyright (c) 2014 Scott Feldman <sfeldma@gmail.com>
|
|
* Copyright (c) 2014 Jiri Pirko <jiri@resnulli.us>
|
|
*
|
|
* This program is free software; you can redistribute it and/or modify
|
|
* it under the terms of the GNU General Public License as published by
|
|
* the Free Software Foundation; either version 2 of the License, or
|
|
* (at your option) any later version.
|
|
*
|
|
* This program is distributed in the hope that it will be useful,
|
|
* but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
* GNU General Public License for more details.
|
|
*/
|
|
|
|
#include "qemu/osdep.h"
|
|
#include "hw/pci/pci.h"
|
|
#include "hw/qdev-properties.h"
|
|
#include "migration/vmstate.h"
|
|
#include "hw/pci/msix.h"
|
|
#include "net/net.h"
|
|
#include "net/eth.h"
|
|
#include "qapi/error.h"
|
|
#include "qapi/qapi-commands-rocker.h"
|
|
#include "qemu/iov.h"
|
|
#include "qemu/module.h"
|
|
#include "qemu/bitops.h"
|
|
|
|
#include "rocker.h"
|
|
#include "rocker_hw.h"
|
|
#include "rocker_fp.h"
|
|
#include "rocker_desc.h"
|
|
#include "rocker_tlv.h"
|
|
#include "rocker_world.h"
|
|
#include "rocker_of_dpa.h"
|
|
|
|
struct rocker {
|
|
/* private */
|
|
PCIDevice parent_obj;
|
|
/* public */
|
|
|
|
MemoryRegion mmio;
|
|
MemoryRegion msix_bar;
|
|
|
|
/* switch configuration */
|
|
char *name; /* switch name */
|
|
char *world_name; /* world name */
|
|
uint32_t fp_ports; /* front-panel port count */
|
|
NICPeers *fp_ports_peers;
|
|
MACAddr fp_start_macaddr; /* front-panel port 0 mac addr */
|
|
uint64_t switch_id; /* switch id */
|
|
|
|
/* front-panel ports */
|
|
FpPort *fp_port[ROCKER_FP_PORTS_MAX];
|
|
|
|
/* register backings */
|
|
uint32_t test_reg;
|
|
uint64_t test_reg64;
|
|
dma_addr_t test_dma_addr;
|
|
uint32_t test_dma_size;
|
|
uint64_t lower32; /* lower 32-bit val in 2-part 64-bit access */
|
|
|
|
/* desc rings */
|
|
DescRing **rings;
|
|
|
|
/* switch worlds */
|
|
World *worlds[ROCKER_WORLD_TYPE_MAX];
|
|
World *world_dflt;
|
|
|
|
QLIST_ENTRY(rocker) next;
|
|
};
|
|
|
|
#define TYPE_ROCKER "rocker"
|
|
|
|
#define ROCKER(obj) \
|
|
OBJECT_CHECK(Rocker, (obj), TYPE_ROCKER)
|
|
|
|
static QLIST_HEAD(, rocker) rockers;
|
|
|
|
Rocker *rocker_find(const char *name)
|
|
{
|
|
Rocker *r;
|
|
|
|
QLIST_FOREACH(r, &rockers, next)
|
|
if (strcmp(r->name, name) == 0) {
|
|
return r;
|
|
}
|
|
|
|
return NULL;
|
|
}
|
|
|
|
World *rocker_get_world(Rocker *r, enum rocker_world_type type)
|
|
{
|
|
if (type < ROCKER_WORLD_TYPE_MAX) {
|
|
return r->worlds[type];
|
|
}
|
|
return NULL;
|
|
}
|
|
|
|
RockerSwitch *qmp_query_rocker(const char *name, Error **errp)
|
|
{
|
|
RockerSwitch *rocker;
|
|
Rocker *r;
|
|
|
|
r = rocker_find(name);
|
|
if (!r) {
|
|
error_setg(errp, "rocker %s not found", name);
|
|
return NULL;
|
|
}
|
|
|
|
rocker = g_new0(RockerSwitch, 1);
|
|
rocker->name = g_strdup(r->name);
|
|
rocker->id = r->switch_id;
|
|
rocker->ports = r->fp_ports;
|
|
|
|
return rocker;
|
|
}
|
|
|
|
RockerPortList *qmp_query_rocker_ports(const char *name, Error **errp)
|
|
{
|
|
RockerPortList *list = NULL;
|
|
Rocker *r;
|
|
int i;
|
|
|
|
r = rocker_find(name);
|
|
if (!r) {
|
|
error_setg(errp, "rocker %s not found", name);
|
|
return NULL;
|
|
}
|
|
|
|
for (i = r->fp_ports - 1; i >= 0; i--) {
|
|
RockerPortList *info = g_malloc0(sizeof(*info));
|
|
info->value = g_malloc0(sizeof(*info->value));
|
|
struct fp_port *port = r->fp_port[i];
|
|
|
|
fp_port_get_info(port, info);
|
|
info->next = list;
|
|
list = info;
|
|
}
|
|
|
|
return list;
|
|
}
|
|
|
|
uint32_t rocker_fp_ports(Rocker *r)
|
|
{
|
|
return r->fp_ports;
|
|
}
|
|
|
|
static uint32_t rocker_get_pport_by_tx_ring(Rocker *r,
|
|
DescRing *ring)
|
|
{
|
|
return (desc_ring_index(ring) - 2) / 2 + 1;
|
|
}
|
|
|
|
static int tx_consume(Rocker *r, DescInfo *info)
|
|
{
|
|
PCIDevice *dev = PCI_DEVICE(r);
|
|
char *buf = desc_get_buf(info, true);
|
|
RockerTlv *tlv_frag;
|
|
RockerTlv *tlvs[ROCKER_TLV_TX_MAX + 1];
|
|
struct iovec iov[ROCKER_TX_FRAGS_MAX] = { { 0, }, };
|
|
uint32_t pport;
|
|
uint32_t port;
|
|
uint16_t tx_offload = ROCKER_TX_OFFLOAD_NONE;
|
|
uint16_t tx_l3_csum_off = 0;
|
|
uint16_t tx_tso_mss = 0;
|
|
uint16_t tx_tso_hdr_len = 0;
|
|
int iovcnt = 0;
|
|
int err = ROCKER_OK;
|
|
int rem;
|
|
int i;
|
|
|
|
if (!buf) {
|
|
return -ROCKER_ENXIO;
|
|
}
|
|
|
|
rocker_tlv_parse(tlvs, ROCKER_TLV_TX_MAX, buf, desc_tlv_size(info));
|
|
|
|
if (!tlvs[ROCKER_TLV_TX_FRAGS]) {
|
|
return -ROCKER_EINVAL;
|
|
}
|
|
|
|
pport = rocker_get_pport_by_tx_ring(r, desc_get_ring(info));
|
|
if (!fp_port_from_pport(pport, &port)) {
|
|
return -ROCKER_EINVAL;
|
|
}
|
|
|
|
if (tlvs[ROCKER_TLV_TX_OFFLOAD]) {
|
|
tx_offload = rocker_tlv_get_u8(tlvs[ROCKER_TLV_TX_OFFLOAD]);
|
|
}
|
|
|
|
switch (tx_offload) {
|
|
case ROCKER_TX_OFFLOAD_L3_CSUM:
|
|
if (!tlvs[ROCKER_TLV_TX_L3_CSUM_OFF]) {
|
|
return -ROCKER_EINVAL;
|
|
}
|
|
break;
|
|
case ROCKER_TX_OFFLOAD_TSO:
|
|
if (!tlvs[ROCKER_TLV_TX_TSO_MSS] ||
|
|
!tlvs[ROCKER_TLV_TX_TSO_HDR_LEN]) {
|
|
return -ROCKER_EINVAL;
|
|
}
|
|
break;
|
|
}
|
|
|
|
if (tlvs[ROCKER_TLV_TX_L3_CSUM_OFF]) {
|
|
tx_l3_csum_off = rocker_tlv_get_le16(tlvs[ROCKER_TLV_TX_L3_CSUM_OFF]);
|
|
}
|
|
|
|
if (tlvs[ROCKER_TLV_TX_TSO_MSS]) {
|
|
tx_tso_mss = rocker_tlv_get_le16(tlvs[ROCKER_TLV_TX_TSO_MSS]);
|
|
}
|
|
|
|
if (tlvs[ROCKER_TLV_TX_TSO_HDR_LEN]) {
|
|
tx_tso_hdr_len = rocker_tlv_get_le16(tlvs[ROCKER_TLV_TX_TSO_HDR_LEN]);
|
|
}
|
|
|
|
rocker_tlv_for_each_nested(tlv_frag, tlvs[ROCKER_TLV_TX_FRAGS], rem) {
|
|
hwaddr frag_addr;
|
|
uint16_t frag_len;
|
|
|
|
if (rocker_tlv_type(tlv_frag) != ROCKER_TLV_TX_FRAG) {
|
|
err = -ROCKER_EINVAL;
|
|
goto err_bad_attr;
|
|
}
|
|
|
|
rocker_tlv_parse_nested(tlvs, ROCKER_TLV_TX_FRAG_ATTR_MAX, tlv_frag);
|
|
|
|
if (!tlvs[ROCKER_TLV_TX_FRAG_ATTR_ADDR] ||
|
|
!tlvs[ROCKER_TLV_TX_FRAG_ATTR_LEN]) {
|
|
err = -ROCKER_EINVAL;
|
|
goto err_bad_attr;
|
|
}
|
|
|
|
frag_addr = rocker_tlv_get_le64(tlvs[ROCKER_TLV_TX_FRAG_ATTR_ADDR]);
|
|
frag_len = rocker_tlv_get_le16(tlvs[ROCKER_TLV_TX_FRAG_ATTR_LEN]);
|
|
|
|
if (iovcnt >= ROCKER_TX_FRAGS_MAX) {
|
|
goto err_too_many_frags;
|
|
}
|
|
iov[iovcnt].iov_len = frag_len;
|
|
iov[iovcnt].iov_base = g_malloc(frag_len);
|
|
|
|
pci_dma_read(dev, frag_addr, iov[iovcnt].iov_base,
|
|
iov[iovcnt].iov_len);
|
|
|
|
iovcnt++;
|
|
}
|
|
|
|
if (iovcnt) {
|
|
/* XXX perform Tx offloads */
|
|
/* XXX silence compiler for now */
|
|
tx_l3_csum_off += tx_tso_mss = tx_tso_hdr_len = 0;
|
|
}
|
|
|
|
err = fp_port_eg(r->fp_port[port], iov, iovcnt);
|
|
|
|
err_too_many_frags:
|
|
err_bad_attr:
|
|
for (i = 0; i < ROCKER_TX_FRAGS_MAX; i++) {
|
|
g_free(iov[i].iov_base);
|
|
}
|
|
|
|
return err;
|
|
}
|
|
|
|
static int cmd_get_port_settings(Rocker *r,
|
|
DescInfo *info, char *buf,
|
|
RockerTlv *cmd_info_tlv)
|
|
{
|
|
RockerTlv *tlvs[ROCKER_TLV_CMD_PORT_SETTINGS_MAX + 1];
|
|
RockerTlv *nest;
|
|
FpPort *fp_port;
|
|
uint32_t pport;
|
|
uint32_t port;
|
|
uint32_t speed;
|
|
uint8_t duplex;
|
|
uint8_t autoneg;
|
|
uint8_t learning;
|
|
char *phys_name;
|
|
MACAddr macaddr;
|
|
enum rocker_world_type mode;
|
|
size_t tlv_size;
|
|
int pos;
|
|
int err;
|
|
|
|
rocker_tlv_parse_nested(tlvs, ROCKER_TLV_CMD_PORT_SETTINGS_MAX,
|
|
cmd_info_tlv);
|
|
|
|
if (!tlvs[ROCKER_TLV_CMD_PORT_SETTINGS_PPORT]) {
|
|
return -ROCKER_EINVAL;
|
|
}
|
|
|
|
pport = rocker_tlv_get_le32(tlvs[ROCKER_TLV_CMD_PORT_SETTINGS_PPORT]);
|
|
if (!fp_port_from_pport(pport, &port)) {
|
|
return -ROCKER_EINVAL;
|
|
}
|
|
fp_port = r->fp_port[port];
|
|
|
|
err = fp_port_get_settings(fp_port, &speed, &duplex, &autoneg);
|
|
if (err) {
|
|
return err;
|
|
}
|
|
|
|
fp_port_get_macaddr(fp_port, &macaddr);
|
|
mode = world_type(fp_port_get_world(fp_port));
|
|
learning = fp_port_get_learning(fp_port);
|
|
phys_name = fp_port_get_name(fp_port);
|
|
|
|
tlv_size = rocker_tlv_total_size(0) + /* nest */
|
|
rocker_tlv_total_size(sizeof(uint32_t)) + /* pport */
|
|
rocker_tlv_total_size(sizeof(uint32_t)) + /* speed */
|
|
rocker_tlv_total_size(sizeof(uint8_t)) + /* duplex */
|
|
rocker_tlv_total_size(sizeof(uint8_t)) + /* autoneg */
|
|
rocker_tlv_total_size(sizeof(macaddr.a)) + /* macaddr */
|
|
rocker_tlv_total_size(sizeof(uint8_t)) + /* mode */
|
|
rocker_tlv_total_size(sizeof(uint8_t)) + /* learning */
|
|
rocker_tlv_total_size(strlen(phys_name));
|
|
|
|
if (tlv_size > desc_buf_size(info)) {
|
|
return -ROCKER_EMSGSIZE;
|
|
}
|
|
|
|
pos = 0;
|
|
nest = rocker_tlv_nest_start(buf, &pos, ROCKER_TLV_CMD_INFO);
|
|
rocker_tlv_put_le32(buf, &pos, ROCKER_TLV_CMD_PORT_SETTINGS_PPORT, pport);
|
|
rocker_tlv_put_le32(buf, &pos, ROCKER_TLV_CMD_PORT_SETTINGS_SPEED, speed);
|
|
rocker_tlv_put_u8(buf, &pos, ROCKER_TLV_CMD_PORT_SETTINGS_DUPLEX, duplex);
|
|
rocker_tlv_put_u8(buf, &pos, ROCKER_TLV_CMD_PORT_SETTINGS_AUTONEG, autoneg);
|
|
rocker_tlv_put(buf, &pos, ROCKER_TLV_CMD_PORT_SETTINGS_MACADDR,
|
|
sizeof(macaddr.a), macaddr.a);
|
|
rocker_tlv_put_u8(buf, &pos, ROCKER_TLV_CMD_PORT_SETTINGS_MODE, mode);
|
|
rocker_tlv_put_u8(buf, &pos, ROCKER_TLV_CMD_PORT_SETTINGS_LEARNING,
|
|
learning);
|
|
rocker_tlv_put(buf, &pos, ROCKER_TLV_CMD_PORT_SETTINGS_PHYS_NAME,
|
|
strlen(phys_name), phys_name);
|
|
rocker_tlv_nest_end(buf, &pos, nest);
|
|
|
|
return desc_set_buf(info, tlv_size);
|
|
}
|
|
|
|
static int cmd_set_port_settings(Rocker *r,
|
|
RockerTlv *cmd_info_tlv)
|
|
{
|
|
RockerTlv *tlvs[ROCKER_TLV_CMD_PORT_SETTINGS_MAX + 1];
|
|
FpPort *fp_port;
|
|
uint32_t pport;
|
|
uint32_t port;
|
|
uint32_t speed;
|
|
uint8_t duplex;
|
|
uint8_t autoneg;
|
|
uint8_t learning;
|
|
MACAddr macaddr;
|
|
enum rocker_world_type mode;
|
|
int err;
|
|
|
|
rocker_tlv_parse_nested(tlvs, ROCKER_TLV_CMD_PORT_SETTINGS_MAX,
|
|
cmd_info_tlv);
|
|
|
|
if (!tlvs[ROCKER_TLV_CMD_PORT_SETTINGS_PPORT]) {
|
|
return -ROCKER_EINVAL;
|
|
}
|
|
|
|
pport = rocker_tlv_get_le32(tlvs[ROCKER_TLV_CMD_PORT_SETTINGS_PPORT]);
|
|
if (!fp_port_from_pport(pport, &port)) {
|
|
return -ROCKER_EINVAL;
|
|
}
|
|
fp_port = r->fp_port[port];
|
|
|
|
if (tlvs[ROCKER_TLV_CMD_PORT_SETTINGS_SPEED] &&
|
|
tlvs[ROCKER_TLV_CMD_PORT_SETTINGS_DUPLEX] &&
|
|
tlvs[ROCKER_TLV_CMD_PORT_SETTINGS_AUTONEG]) {
|
|
|
|
speed = rocker_tlv_get_le32(tlvs[ROCKER_TLV_CMD_PORT_SETTINGS_SPEED]);
|
|
duplex = rocker_tlv_get_u8(tlvs[ROCKER_TLV_CMD_PORT_SETTINGS_DUPLEX]);
|
|
autoneg = rocker_tlv_get_u8(tlvs[ROCKER_TLV_CMD_PORT_SETTINGS_AUTONEG]);
|
|
|
|
err = fp_port_set_settings(fp_port, speed, duplex, autoneg);
|
|
if (err) {
|
|
return err;
|
|
}
|
|
}
|
|
|
|
if (tlvs[ROCKER_TLV_CMD_PORT_SETTINGS_MACADDR]) {
|
|
if (rocker_tlv_len(tlvs[ROCKER_TLV_CMD_PORT_SETTINGS_MACADDR]) !=
|
|
sizeof(macaddr.a)) {
|
|
return -ROCKER_EINVAL;
|
|
}
|
|
memcpy(macaddr.a,
|
|
rocker_tlv_data(tlvs[ROCKER_TLV_CMD_PORT_SETTINGS_MACADDR]),
|
|
sizeof(macaddr.a));
|
|
fp_port_set_macaddr(fp_port, &macaddr);
|
|
}
|
|
|
|
if (tlvs[ROCKER_TLV_CMD_PORT_SETTINGS_MODE]) {
|
|
mode = rocker_tlv_get_u8(tlvs[ROCKER_TLV_CMD_PORT_SETTINGS_MODE]);
|
|
if (mode >= ROCKER_WORLD_TYPE_MAX) {
|
|
return -ROCKER_EINVAL;
|
|
}
|
|
/* We don't support world change. */
|
|
if (!fp_port_check_world(fp_port, r->worlds[mode])) {
|
|
return -ROCKER_EINVAL;
|
|
}
|
|
}
|
|
|
|
if (tlvs[ROCKER_TLV_CMD_PORT_SETTINGS_LEARNING]) {
|
|
learning =
|
|
rocker_tlv_get_u8(tlvs[ROCKER_TLV_CMD_PORT_SETTINGS_LEARNING]);
|
|
fp_port_set_learning(fp_port, learning);
|
|
}
|
|
|
|
return ROCKER_OK;
|
|
}
|
|
|
|
static int cmd_consume(Rocker *r, DescInfo *info)
|
|
{
|
|
char *buf = desc_get_buf(info, false);
|
|
RockerTlv *tlvs[ROCKER_TLV_CMD_MAX + 1];
|
|
RockerTlv *info_tlv;
|
|
World *world;
|
|
uint16_t cmd;
|
|
int err;
|
|
|
|
if (!buf) {
|
|
return -ROCKER_ENXIO;
|
|
}
|
|
|
|
rocker_tlv_parse(tlvs, ROCKER_TLV_CMD_MAX, buf, desc_tlv_size(info));
|
|
|
|
if (!tlvs[ROCKER_TLV_CMD_TYPE] || !tlvs[ROCKER_TLV_CMD_INFO]) {
|
|
return -ROCKER_EINVAL;
|
|
}
|
|
|
|
cmd = rocker_tlv_get_le16(tlvs[ROCKER_TLV_CMD_TYPE]);
|
|
info_tlv = tlvs[ROCKER_TLV_CMD_INFO];
|
|
|
|
/* This might be reworked to something like this:
|
|
* Every world will have an array of command handlers from
|
|
* ROCKER_TLV_CMD_TYPE_UNSPEC to ROCKER_TLV_CMD_TYPE_MAX. There is
|
|
* up to each world to implement whatever command it want.
|
|
* It can reference "generic" commands as cmd_set_port_settings or
|
|
* cmd_get_port_settings
|
|
*/
|
|
|
|
switch (cmd) {
|
|
case ROCKER_TLV_CMD_TYPE_OF_DPA_FLOW_ADD:
|
|
case ROCKER_TLV_CMD_TYPE_OF_DPA_FLOW_MOD:
|
|
case ROCKER_TLV_CMD_TYPE_OF_DPA_FLOW_DEL:
|
|
case ROCKER_TLV_CMD_TYPE_OF_DPA_FLOW_GET_STATS:
|
|
case ROCKER_TLV_CMD_TYPE_OF_DPA_GROUP_ADD:
|
|
case ROCKER_TLV_CMD_TYPE_OF_DPA_GROUP_MOD:
|
|
case ROCKER_TLV_CMD_TYPE_OF_DPA_GROUP_DEL:
|
|
case ROCKER_TLV_CMD_TYPE_OF_DPA_GROUP_GET_STATS:
|
|
world = r->worlds[ROCKER_WORLD_TYPE_OF_DPA];
|
|
err = world_do_cmd(world, info, buf, cmd, info_tlv);
|
|
break;
|
|
case ROCKER_TLV_CMD_TYPE_GET_PORT_SETTINGS:
|
|
err = cmd_get_port_settings(r, info, buf, info_tlv);
|
|
break;
|
|
case ROCKER_TLV_CMD_TYPE_SET_PORT_SETTINGS:
|
|
err = cmd_set_port_settings(r, info_tlv);
|
|
break;
|
|
default:
|
|
err = -ROCKER_EINVAL;
|
|
break;
|
|
}
|
|
|
|
return err;
|
|
}
|
|
|
|
static void rocker_msix_irq(Rocker *r, unsigned vector)
|
|
{
|
|
PCIDevice *dev = PCI_DEVICE(r);
|
|
|
|
DPRINTF("MSI-X notify request for vector %d\n", vector);
|
|
if (vector >= ROCKER_MSIX_VEC_COUNT(r->fp_ports)) {
|
|
DPRINTF("incorrect vector %d\n", vector);
|
|
return;
|
|
}
|
|
msix_notify(dev, vector);
|
|
}
|
|
|
|
int rocker_event_link_changed(Rocker *r, uint32_t pport, bool link_up)
|
|
{
|
|
DescRing *ring = r->rings[ROCKER_RING_EVENT];
|
|
DescInfo *info = desc_ring_fetch_desc(ring);
|
|
RockerTlv *nest;
|
|
char *buf;
|
|
size_t tlv_size;
|
|
int pos;
|
|
int err;
|
|
|
|
if (!info) {
|
|
return -ROCKER_ENOBUFS;
|
|
}
|
|
|
|
tlv_size = rocker_tlv_total_size(sizeof(uint16_t)) + /* event type */
|
|
rocker_tlv_total_size(0) + /* nest */
|
|
rocker_tlv_total_size(sizeof(uint32_t)) + /* pport */
|
|
rocker_tlv_total_size(sizeof(uint8_t)); /* link up */
|
|
|
|
if (tlv_size > desc_buf_size(info)) {
|
|
err = -ROCKER_EMSGSIZE;
|
|
goto err_too_big;
|
|
}
|
|
|
|
buf = desc_get_buf(info, false);
|
|
if (!buf) {
|
|
err = -ROCKER_ENOMEM;
|
|
goto err_no_mem;
|
|
}
|
|
|
|
pos = 0;
|
|
rocker_tlv_put_le32(buf, &pos, ROCKER_TLV_EVENT_TYPE,
|
|
ROCKER_TLV_EVENT_TYPE_LINK_CHANGED);
|
|
nest = rocker_tlv_nest_start(buf, &pos, ROCKER_TLV_EVENT_INFO);
|
|
rocker_tlv_put_le32(buf, &pos, ROCKER_TLV_EVENT_LINK_CHANGED_PPORT, pport);
|
|
rocker_tlv_put_u8(buf, &pos, ROCKER_TLV_EVENT_LINK_CHANGED_LINKUP,
|
|
link_up ? 1 : 0);
|
|
rocker_tlv_nest_end(buf, &pos, nest);
|
|
|
|
err = desc_set_buf(info, tlv_size);
|
|
|
|
err_too_big:
|
|
err_no_mem:
|
|
if (desc_ring_post_desc(ring, err)) {
|
|
rocker_msix_irq(r, ROCKER_MSIX_VEC_EVENT);
|
|
}
|
|
|
|
return err;
|
|
}
|
|
|
|
int rocker_event_mac_vlan_seen(Rocker *r, uint32_t pport, uint8_t *addr,
|
|
uint16_t vlan_id)
|
|
{
|
|
DescRing *ring = r->rings[ROCKER_RING_EVENT];
|
|
DescInfo *info;
|
|
FpPort *fp_port;
|
|
uint32_t port;
|
|
RockerTlv *nest;
|
|
char *buf;
|
|
size_t tlv_size;
|
|
int pos;
|
|
int err;
|
|
|
|
if (!fp_port_from_pport(pport, &port)) {
|
|
return -ROCKER_EINVAL;
|
|
}
|
|
fp_port = r->fp_port[port];
|
|
if (!fp_port_get_learning(fp_port)) {
|
|
return ROCKER_OK;
|
|
}
|
|
|
|
info = desc_ring_fetch_desc(ring);
|
|
if (!info) {
|
|
return -ROCKER_ENOBUFS;
|
|
}
|
|
|
|
tlv_size = rocker_tlv_total_size(sizeof(uint16_t)) + /* event type */
|
|
rocker_tlv_total_size(0) + /* nest */
|
|
rocker_tlv_total_size(sizeof(uint32_t)) + /* pport */
|
|
rocker_tlv_total_size(ETH_ALEN) + /* mac addr */
|
|
rocker_tlv_total_size(sizeof(uint16_t)); /* vlan_id */
|
|
|
|
if (tlv_size > desc_buf_size(info)) {
|
|
err = -ROCKER_EMSGSIZE;
|
|
goto err_too_big;
|
|
}
|
|
|
|
buf = desc_get_buf(info, false);
|
|
if (!buf) {
|
|
err = -ROCKER_ENOMEM;
|
|
goto err_no_mem;
|
|
}
|
|
|
|
pos = 0;
|
|
rocker_tlv_put_le32(buf, &pos, ROCKER_TLV_EVENT_TYPE,
|
|
ROCKER_TLV_EVENT_TYPE_MAC_VLAN_SEEN);
|
|
nest = rocker_tlv_nest_start(buf, &pos, ROCKER_TLV_EVENT_INFO);
|
|
rocker_tlv_put_le32(buf, &pos, ROCKER_TLV_EVENT_MAC_VLAN_PPORT, pport);
|
|
rocker_tlv_put(buf, &pos, ROCKER_TLV_EVENT_MAC_VLAN_MAC, ETH_ALEN, addr);
|
|
rocker_tlv_put_u16(buf, &pos, ROCKER_TLV_EVENT_MAC_VLAN_VLAN_ID, vlan_id);
|
|
rocker_tlv_nest_end(buf, &pos, nest);
|
|
|
|
err = desc_set_buf(info, tlv_size);
|
|
|
|
err_too_big:
|
|
err_no_mem:
|
|
if (desc_ring_post_desc(ring, err)) {
|
|
rocker_msix_irq(r, ROCKER_MSIX_VEC_EVENT);
|
|
}
|
|
|
|
return err;
|
|
}
|
|
|
|
static DescRing *rocker_get_rx_ring_by_pport(Rocker *r,
|
|
uint32_t pport)
|
|
{
|
|
return r->rings[(pport - 1) * 2 + 3];
|
|
}
|
|
|
|
int rx_produce(World *world, uint32_t pport,
|
|
const struct iovec *iov, int iovcnt, uint8_t copy_to_cpu)
|
|
{
|
|
Rocker *r = world_rocker(world);
|
|
PCIDevice *dev = (PCIDevice *)r;
|
|
DescRing *ring = rocker_get_rx_ring_by_pport(r, pport);
|
|
DescInfo *info = desc_ring_fetch_desc(ring);
|
|
char *data;
|
|
size_t data_size = iov_size(iov, iovcnt);
|
|
char *buf;
|
|
uint16_t rx_flags = 0;
|
|
uint16_t rx_csum = 0;
|
|
size_t tlv_size;
|
|
RockerTlv *tlvs[ROCKER_TLV_RX_MAX + 1];
|
|
hwaddr frag_addr;
|
|
uint16_t frag_max_len;
|
|
int pos;
|
|
int err;
|
|
|
|
if (!info) {
|
|
return -ROCKER_ENOBUFS;
|
|
}
|
|
|
|
buf = desc_get_buf(info, false);
|
|
if (!buf) {
|
|
err = -ROCKER_ENXIO;
|
|
goto out;
|
|
}
|
|
rocker_tlv_parse(tlvs, ROCKER_TLV_RX_MAX, buf, desc_tlv_size(info));
|
|
|
|
if (!tlvs[ROCKER_TLV_RX_FRAG_ADDR] ||
|
|
!tlvs[ROCKER_TLV_RX_FRAG_MAX_LEN]) {
|
|
err = -ROCKER_EINVAL;
|
|
goto out;
|
|
}
|
|
|
|
frag_addr = rocker_tlv_get_le64(tlvs[ROCKER_TLV_RX_FRAG_ADDR]);
|
|
frag_max_len = rocker_tlv_get_le16(tlvs[ROCKER_TLV_RX_FRAG_MAX_LEN]);
|
|
|
|
if (data_size > frag_max_len) {
|
|
err = -ROCKER_EMSGSIZE;
|
|
goto out;
|
|
}
|
|
|
|
if (copy_to_cpu) {
|
|
rx_flags |= ROCKER_RX_FLAGS_FWD_OFFLOAD;
|
|
}
|
|
|
|
/* XXX calc rx flags/csum */
|
|
|
|
tlv_size = rocker_tlv_total_size(sizeof(uint16_t)) + /* flags */
|
|
rocker_tlv_total_size(sizeof(uint16_t)) + /* scum */
|
|
rocker_tlv_total_size(sizeof(uint64_t)) + /* frag addr */
|
|
rocker_tlv_total_size(sizeof(uint16_t)) + /* frag max len */
|
|
rocker_tlv_total_size(sizeof(uint16_t)); /* frag len */
|
|
|
|
if (tlv_size > desc_buf_size(info)) {
|
|
err = -ROCKER_EMSGSIZE;
|
|
goto out;
|
|
}
|
|
|
|
/* TODO:
|
|
* iov dma write can be optimized in similar way e1000 does it in
|
|
* e1000_receive_iov. But maybe if would make sense to introduce
|
|
* generic helper iov_dma_write.
|
|
*/
|
|
|
|
data = g_malloc(data_size);
|
|
|
|
iov_to_buf(iov, iovcnt, 0, data, data_size);
|
|
pci_dma_write(dev, frag_addr, data, data_size);
|
|
g_free(data);
|
|
|
|
pos = 0;
|
|
rocker_tlv_put_le16(buf, &pos, ROCKER_TLV_RX_FLAGS, rx_flags);
|
|
rocker_tlv_put_le16(buf, &pos, ROCKER_TLV_RX_CSUM, rx_csum);
|
|
rocker_tlv_put_le64(buf, &pos, ROCKER_TLV_RX_FRAG_ADDR, frag_addr);
|
|
rocker_tlv_put_le16(buf, &pos, ROCKER_TLV_RX_FRAG_MAX_LEN, frag_max_len);
|
|
rocker_tlv_put_le16(buf, &pos, ROCKER_TLV_RX_FRAG_LEN, data_size);
|
|
|
|
err = desc_set_buf(info, tlv_size);
|
|
|
|
out:
|
|
if (desc_ring_post_desc(ring, err)) {
|
|
rocker_msix_irq(r, ROCKER_MSIX_VEC_RX(pport - 1));
|
|
}
|
|
|
|
return err;
|
|
}
|
|
|
|
int rocker_port_eg(Rocker *r, uint32_t pport,
|
|
const struct iovec *iov, int iovcnt)
|
|
{
|
|
FpPort *fp_port;
|
|
uint32_t port;
|
|
|
|
if (!fp_port_from_pport(pport, &port)) {
|
|
return -ROCKER_EINVAL;
|
|
}
|
|
|
|
fp_port = r->fp_port[port];
|
|
|
|
return fp_port_eg(fp_port, iov, iovcnt);
|
|
}
|
|
|
|
static void rocker_test_dma_ctrl(Rocker *r, uint32_t val)
|
|
{
|
|
PCIDevice *dev = PCI_DEVICE(r);
|
|
char *buf;
|
|
int i;
|
|
|
|
buf = g_malloc(r->test_dma_size);
|
|
|
|
switch (val) {
|
|
case ROCKER_TEST_DMA_CTRL_CLEAR:
|
|
memset(buf, 0, r->test_dma_size);
|
|
break;
|
|
case ROCKER_TEST_DMA_CTRL_FILL:
|
|
memset(buf, 0x96, r->test_dma_size);
|
|
break;
|
|
case ROCKER_TEST_DMA_CTRL_INVERT:
|
|
pci_dma_read(dev, r->test_dma_addr, buf, r->test_dma_size);
|
|
for (i = 0; i < r->test_dma_size; i++) {
|
|
buf[i] = ~buf[i];
|
|
}
|
|
break;
|
|
default:
|
|
DPRINTF("not test dma control val=0x%08x\n", val);
|
|
goto err_out;
|
|
}
|
|
pci_dma_write(dev, r->test_dma_addr, buf, r->test_dma_size);
|
|
|
|
rocker_msix_irq(r, ROCKER_MSIX_VEC_TEST);
|
|
|
|
err_out:
|
|
g_free(buf);
|
|
}
|
|
|
|
static void rocker_reset(DeviceState *dev);
|
|
|
|
static void rocker_control(Rocker *r, uint32_t val)
|
|
{
|
|
if (val & ROCKER_CONTROL_RESET) {
|
|
rocker_reset(DEVICE(r));
|
|
}
|
|
}
|
|
|
|
static int rocker_pci_ring_count(Rocker *r)
|
|
{
|
|
/* There are:
|
|
* - command ring
|
|
* - event ring
|
|
* - tx and rx ring per each port
|
|
*/
|
|
return 2 + (2 * r->fp_ports);
|
|
}
|
|
|
|
static bool rocker_addr_is_desc_reg(Rocker *r, hwaddr addr)
|
|
{
|
|
hwaddr start = ROCKER_DMA_DESC_BASE;
|
|
hwaddr end = start + (ROCKER_DMA_DESC_SIZE * rocker_pci_ring_count(r));
|
|
|
|
return addr >= start && addr < end;
|
|
}
|
|
|
|
static void rocker_port_phys_enable_write(Rocker *r, uint64_t new)
|
|
{
|
|
int i;
|
|
bool old_enabled;
|
|
bool new_enabled;
|
|
FpPort *fp_port;
|
|
|
|
for (i = 0; i < r->fp_ports; i++) {
|
|
fp_port = r->fp_port[i];
|
|
old_enabled = fp_port_enabled(fp_port);
|
|
new_enabled = (new >> (i + 1)) & 0x1;
|
|
if (new_enabled == old_enabled) {
|
|
continue;
|
|
}
|
|
if (new_enabled) {
|
|
fp_port_enable(r->fp_port[i]);
|
|
} else {
|
|
fp_port_disable(r->fp_port[i]);
|
|
}
|
|
}
|
|
}
|
|
|
|
static void rocker_io_writel(void *opaque, hwaddr addr, uint32_t val)
|
|
{
|
|
Rocker *r = opaque;
|
|
|
|
if (rocker_addr_is_desc_reg(r, addr)) {
|
|
unsigned index = ROCKER_RING_INDEX(addr);
|
|
unsigned offset = addr & ROCKER_DMA_DESC_MASK;
|
|
|
|
switch (offset) {
|
|
case ROCKER_DMA_DESC_ADDR_OFFSET:
|
|
r->lower32 = (uint64_t)val;
|
|
break;
|
|
case ROCKER_DMA_DESC_ADDR_OFFSET + 4:
|
|
desc_ring_set_base_addr(r->rings[index],
|
|
((uint64_t)val) << 32 | r->lower32);
|
|
r->lower32 = 0;
|
|
break;
|
|
case ROCKER_DMA_DESC_SIZE_OFFSET:
|
|
desc_ring_set_size(r->rings[index], val);
|
|
break;
|
|
case ROCKER_DMA_DESC_HEAD_OFFSET:
|
|
if (desc_ring_set_head(r->rings[index], val)) {
|
|
rocker_msix_irq(r, desc_ring_get_msix_vector(r->rings[index]));
|
|
}
|
|
break;
|
|
case ROCKER_DMA_DESC_CTRL_OFFSET:
|
|
desc_ring_set_ctrl(r->rings[index], val);
|
|
break;
|
|
case ROCKER_DMA_DESC_CREDITS_OFFSET:
|
|
if (desc_ring_ret_credits(r->rings[index], val)) {
|
|
rocker_msix_irq(r, desc_ring_get_msix_vector(r->rings[index]));
|
|
}
|
|
break;
|
|
default:
|
|
DPRINTF("not implemented dma reg write(l) addr=0x" TARGET_FMT_plx
|
|
" val=0x%08x (ring %d, addr=0x%02x)\n",
|
|
addr, val, index, offset);
|
|
break;
|
|
}
|
|
return;
|
|
}
|
|
|
|
switch (addr) {
|
|
case ROCKER_TEST_REG:
|
|
r->test_reg = val;
|
|
break;
|
|
case ROCKER_TEST_REG64:
|
|
case ROCKER_TEST_DMA_ADDR:
|
|
case ROCKER_PORT_PHYS_ENABLE:
|
|
r->lower32 = (uint64_t)val;
|
|
break;
|
|
case ROCKER_TEST_REG64 + 4:
|
|
r->test_reg64 = ((uint64_t)val) << 32 | r->lower32;
|
|
r->lower32 = 0;
|
|
break;
|
|
case ROCKER_TEST_IRQ:
|
|
rocker_msix_irq(r, val);
|
|
break;
|
|
case ROCKER_TEST_DMA_SIZE:
|
|
r->test_dma_size = val & 0xFFFF;
|
|
break;
|
|
case ROCKER_TEST_DMA_ADDR + 4:
|
|
r->test_dma_addr = ((uint64_t)val) << 32 | r->lower32;
|
|
r->lower32 = 0;
|
|
break;
|
|
case ROCKER_TEST_DMA_CTRL:
|
|
rocker_test_dma_ctrl(r, val);
|
|
break;
|
|
case ROCKER_CONTROL:
|
|
rocker_control(r, val);
|
|
break;
|
|
case ROCKER_PORT_PHYS_ENABLE + 4:
|
|
rocker_port_phys_enable_write(r, ((uint64_t)val) << 32 | r->lower32);
|
|
r->lower32 = 0;
|
|
break;
|
|
default:
|
|
DPRINTF("not implemented write(l) addr=0x" TARGET_FMT_plx
|
|
" val=0x%08x\n", addr, val);
|
|
break;
|
|
}
|
|
}
|
|
|
|
static void rocker_io_writeq(void *opaque, hwaddr addr, uint64_t val)
|
|
{
|
|
Rocker *r = opaque;
|
|
|
|
if (rocker_addr_is_desc_reg(r, addr)) {
|
|
unsigned index = ROCKER_RING_INDEX(addr);
|
|
unsigned offset = addr & ROCKER_DMA_DESC_MASK;
|
|
|
|
switch (offset) {
|
|
case ROCKER_DMA_DESC_ADDR_OFFSET:
|
|
desc_ring_set_base_addr(r->rings[index], val);
|
|
break;
|
|
default:
|
|
DPRINTF("not implemented dma reg write(q) addr=0x" TARGET_FMT_plx
|
|
" val=0x" TARGET_FMT_plx " (ring %d, offset=0x%02x)\n",
|
|
addr, val, index, offset);
|
|
break;
|
|
}
|
|
return;
|
|
}
|
|
|
|
switch (addr) {
|
|
case ROCKER_TEST_REG64:
|
|
r->test_reg64 = val;
|
|
break;
|
|
case ROCKER_TEST_DMA_ADDR:
|
|
r->test_dma_addr = val;
|
|
break;
|
|
case ROCKER_PORT_PHYS_ENABLE:
|
|
rocker_port_phys_enable_write(r, val);
|
|
break;
|
|
default:
|
|
DPRINTF("not implemented write(q) addr=0x" TARGET_FMT_plx
|
|
" val=0x" TARGET_FMT_plx "\n", addr, val);
|
|
break;
|
|
}
|
|
}
|
|
|
|
#ifdef DEBUG_ROCKER
|
|
#define regname(reg) case (reg): return #reg
|
|
static const char *rocker_reg_name(void *opaque, hwaddr addr)
|
|
{
|
|
Rocker *r = opaque;
|
|
|
|
if (rocker_addr_is_desc_reg(r, addr)) {
|
|
unsigned index = ROCKER_RING_INDEX(addr);
|
|
unsigned offset = addr & ROCKER_DMA_DESC_MASK;
|
|
static char buf[100];
|
|
char ring_name[10];
|
|
|
|
switch (index) {
|
|
case 0:
|
|
sprintf(ring_name, "cmd");
|
|
break;
|
|
case 1:
|
|
sprintf(ring_name, "event");
|
|
break;
|
|
default:
|
|
sprintf(ring_name, "%s-%d", index % 2 ? "rx" : "tx",
|
|
(index - 2) / 2);
|
|
}
|
|
|
|
switch (offset) {
|
|
case ROCKER_DMA_DESC_ADDR_OFFSET:
|
|
sprintf(buf, "Ring[%s] ADDR", ring_name);
|
|
return buf;
|
|
case ROCKER_DMA_DESC_ADDR_OFFSET+4:
|
|
sprintf(buf, "Ring[%s] ADDR+4", ring_name);
|
|
return buf;
|
|
case ROCKER_DMA_DESC_SIZE_OFFSET:
|
|
sprintf(buf, "Ring[%s] SIZE", ring_name);
|
|
return buf;
|
|
case ROCKER_DMA_DESC_HEAD_OFFSET:
|
|
sprintf(buf, "Ring[%s] HEAD", ring_name);
|
|
return buf;
|
|
case ROCKER_DMA_DESC_TAIL_OFFSET:
|
|
sprintf(buf, "Ring[%s] TAIL", ring_name);
|
|
return buf;
|
|
case ROCKER_DMA_DESC_CTRL_OFFSET:
|
|
sprintf(buf, "Ring[%s] CTRL", ring_name);
|
|
return buf;
|
|
case ROCKER_DMA_DESC_CREDITS_OFFSET:
|
|
sprintf(buf, "Ring[%s] CREDITS", ring_name);
|
|
return buf;
|
|
default:
|
|
sprintf(buf, "Ring[%s] ???", ring_name);
|
|
return buf;
|
|
}
|
|
} else {
|
|
switch (addr) {
|
|
regname(ROCKER_BOGUS_REG0);
|
|
regname(ROCKER_BOGUS_REG1);
|
|
regname(ROCKER_BOGUS_REG2);
|
|
regname(ROCKER_BOGUS_REG3);
|
|
regname(ROCKER_TEST_REG);
|
|
regname(ROCKER_TEST_REG64);
|
|
regname(ROCKER_TEST_REG64+4);
|
|
regname(ROCKER_TEST_IRQ);
|
|
regname(ROCKER_TEST_DMA_ADDR);
|
|
regname(ROCKER_TEST_DMA_ADDR+4);
|
|
regname(ROCKER_TEST_DMA_SIZE);
|
|
regname(ROCKER_TEST_DMA_CTRL);
|
|
regname(ROCKER_CONTROL);
|
|
regname(ROCKER_PORT_PHYS_COUNT);
|
|
regname(ROCKER_PORT_PHYS_LINK_STATUS);
|
|
regname(ROCKER_PORT_PHYS_LINK_STATUS+4);
|
|
regname(ROCKER_PORT_PHYS_ENABLE);
|
|
regname(ROCKER_PORT_PHYS_ENABLE+4);
|
|
regname(ROCKER_SWITCH_ID);
|
|
regname(ROCKER_SWITCH_ID+4);
|
|
}
|
|
}
|
|
return "???";
|
|
}
|
|
#else
|
|
static const char *rocker_reg_name(void *opaque, hwaddr addr)
|
|
{
|
|
return NULL;
|
|
}
|
|
#endif
|
|
|
|
static void rocker_mmio_write(void *opaque, hwaddr addr, uint64_t val,
|
|
unsigned size)
|
|
{
|
|
DPRINTF("Write %s addr " TARGET_FMT_plx
|
|
", size %u, val " TARGET_FMT_plx "\n",
|
|
rocker_reg_name(opaque, addr), addr, size, val);
|
|
|
|
switch (size) {
|
|
case 4:
|
|
rocker_io_writel(opaque, addr, val);
|
|
break;
|
|
case 8:
|
|
rocker_io_writeq(opaque, addr, val);
|
|
break;
|
|
}
|
|
}
|
|
|
|
static uint64_t rocker_port_phys_link_status(Rocker *r)
|
|
{
|
|
int i;
|
|
uint64_t status = 0;
|
|
|
|
for (i = 0; i < r->fp_ports; i++) {
|
|
FpPort *port = r->fp_port[i];
|
|
|
|
if (fp_port_get_link_up(port)) {
|
|
status |= 1 << (i + 1);
|
|
}
|
|
}
|
|
return status;
|
|
}
|
|
|
|
static uint64_t rocker_port_phys_enable_read(Rocker *r)
|
|
{
|
|
int i;
|
|
uint64_t ret = 0;
|
|
|
|
for (i = 0; i < r->fp_ports; i++) {
|
|
FpPort *port = r->fp_port[i];
|
|
|
|
if (fp_port_enabled(port)) {
|
|
ret |= 1 << (i + 1);
|
|
}
|
|
}
|
|
return ret;
|
|
}
|
|
|
|
static uint32_t rocker_io_readl(void *opaque, hwaddr addr)
|
|
{
|
|
Rocker *r = opaque;
|
|
uint32_t ret;
|
|
|
|
if (rocker_addr_is_desc_reg(r, addr)) {
|
|
unsigned index = ROCKER_RING_INDEX(addr);
|
|
unsigned offset = addr & ROCKER_DMA_DESC_MASK;
|
|
|
|
switch (offset) {
|
|
case ROCKER_DMA_DESC_ADDR_OFFSET:
|
|
ret = (uint32_t)desc_ring_get_base_addr(r->rings[index]);
|
|
break;
|
|
case ROCKER_DMA_DESC_ADDR_OFFSET + 4:
|
|
ret = (uint32_t)(desc_ring_get_base_addr(r->rings[index]) >> 32);
|
|
break;
|
|
case ROCKER_DMA_DESC_SIZE_OFFSET:
|
|
ret = desc_ring_get_size(r->rings[index]);
|
|
break;
|
|
case ROCKER_DMA_DESC_HEAD_OFFSET:
|
|
ret = desc_ring_get_head(r->rings[index]);
|
|
break;
|
|
case ROCKER_DMA_DESC_TAIL_OFFSET:
|
|
ret = desc_ring_get_tail(r->rings[index]);
|
|
break;
|
|
case ROCKER_DMA_DESC_CREDITS_OFFSET:
|
|
ret = desc_ring_get_credits(r->rings[index]);
|
|
break;
|
|
default:
|
|
DPRINTF("not implemented dma reg read(l) addr=0x" TARGET_FMT_plx
|
|
" (ring %d, addr=0x%02x)\n", addr, index, offset);
|
|
ret = 0;
|
|
break;
|
|
}
|
|
return ret;
|
|
}
|
|
|
|
switch (addr) {
|
|
case ROCKER_BOGUS_REG0:
|
|
case ROCKER_BOGUS_REG1:
|
|
case ROCKER_BOGUS_REG2:
|
|
case ROCKER_BOGUS_REG3:
|
|
ret = 0xDEADBABE;
|
|
break;
|
|
case ROCKER_TEST_REG:
|
|
ret = r->test_reg * 2;
|
|
break;
|
|
case ROCKER_TEST_REG64:
|
|
ret = (uint32_t)(r->test_reg64 * 2);
|
|
break;
|
|
case ROCKER_TEST_REG64 + 4:
|
|
ret = (uint32_t)((r->test_reg64 * 2) >> 32);
|
|
break;
|
|
case ROCKER_TEST_DMA_SIZE:
|
|
ret = r->test_dma_size;
|
|
break;
|
|
case ROCKER_TEST_DMA_ADDR:
|
|
ret = (uint32_t)r->test_dma_addr;
|
|
break;
|
|
case ROCKER_TEST_DMA_ADDR + 4:
|
|
ret = (uint32_t)(r->test_dma_addr >> 32);
|
|
break;
|
|
case ROCKER_PORT_PHYS_COUNT:
|
|
ret = r->fp_ports;
|
|
break;
|
|
case ROCKER_PORT_PHYS_LINK_STATUS:
|
|
ret = (uint32_t)rocker_port_phys_link_status(r);
|
|
break;
|
|
case ROCKER_PORT_PHYS_LINK_STATUS + 4:
|
|
ret = (uint32_t)(rocker_port_phys_link_status(r) >> 32);
|
|
break;
|
|
case ROCKER_PORT_PHYS_ENABLE:
|
|
ret = (uint32_t)rocker_port_phys_enable_read(r);
|
|
break;
|
|
case ROCKER_PORT_PHYS_ENABLE + 4:
|
|
ret = (uint32_t)(rocker_port_phys_enable_read(r) >> 32);
|
|
break;
|
|
case ROCKER_SWITCH_ID:
|
|
ret = (uint32_t)r->switch_id;
|
|
break;
|
|
case ROCKER_SWITCH_ID + 4:
|
|
ret = (uint32_t)(r->switch_id >> 32);
|
|
break;
|
|
default:
|
|
DPRINTF("not implemented read(l) addr=0x" TARGET_FMT_plx "\n", addr);
|
|
ret = 0;
|
|
break;
|
|
}
|
|
return ret;
|
|
}
|
|
|
|
static uint64_t rocker_io_readq(void *opaque, hwaddr addr)
|
|
{
|
|
Rocker *r = opaque;
|
|
uint64_t ret;
|
|
|
|
if (rocker_addr_is_desc_reg(r, addr)) {
|
|
unsigned index = ROCKER_RING_INDEX(addr);
|
|
unsigned offset = addr & ROCKER_DMA_DESC_MASK;
|
|
|
|
switch (addr & ROCKER_DMA_DESC_MASK) {
|
|
case ROCKER_DMA_DESC_ADDR_OFFSET:
|
|
ret = desc_ring_get_base_addr(r->rings[index]);
|
|
break;
|
|
default:
|
|
DPRINTF("not implemented dma reg read(q) addr=0x" TARGET_FMT_plx
|
|
" (ring %d, addr=0x%02x)\n", addr, index, offset);
|
|
ret = 0;
|
|
break;
|
|
}
|
|
return ret;
|
|
}
|
|
|
|
switch (addr) {
|
|
case ROCKER_BOGUS_REG0:
|
|
case ROCKER_BOGUS_REG2:
|
|
ret = 0xDEADBABEDEADBABEULL;
|
|
break;
|
|
case ROCKER_TEST_REG64:
|
|
ret = r->test_reg64 * 2;
|
|
break;
|
|
case ROCKER_TEST_DMA_ADDR:
|
|
ret = r->test_dma_addr;
|
|
break;
|
|
case ROCKER_PORT_PHYS_LINK_STATUS:
|
|
ret = rocker_port_phys_link_status(r);
|
|
break;
|
|
case ROCKER_PORT_PHYS_ENABLE:
|
|
ret = rocker_port_phys_enable_read(r);
|
|
break;
|
|
case ROCKER_SWITCH_ID:
|
|
ret = r->switch_id;
|
|
break;
|
|
default:
|
|
DPRINTF("not implemented read(q) addr=0x" TARGET_FMT_plx "\n", addr);
|
|
ret = 0;
|
|
break;
|
|
}
|
|
return ret;
|
|
}
|
|
|
|
static uint64_t rocker_mmio_read(void *opaque, hwaddr addr, unsigned size)
|
|
{
|
|
DPRINTF("Read %s addr " TARGET_FMT_plx ", size %u\n",
|
|
rocker_reg_name(opaque, addr), addr, size);
|
|
|
|
switch (size) {
|
|
case 4:
|
|
return rocker_io_readl(opaque, addr);
|
|
case 8:
|
|
return rocker_io_readq(opaque, addr);
|
|
}
|
|
|
|
return -1;
|
|
}
|
|
|
|
static const MemoryRegionOps rocker_mmio_ops = {
|
|
.read = rocker_mmio_read,
|
|
.write = rocker_mmio_write,
|
|
.endianness = DEVICE_LITTLE_ENDIAN,
|
|
.valid = {
|
|
.min_access_size = 4,
|
|
.max_access_size = 8,
|
|
},
|
|
.impl = {
|
|
.min_access_size = 4,
|
|
.max_access_size = 8,
|
|
},
|
|
};
|
|
|
|
static void rocker_msix_vectors_unuse(Rocker *r,
|
|
unsigned int num_vectors)
|
|
{
|
|
PCIDevice *dev = PCI_DEVICE(r);
|
|
int i;
|
|
|
|
for (i = 0; i < num_vectors; i++) {
|
|
msix_vector_unuse(dev, i);
|
|
}
|
|
}
|
|
|
|
static int rocker_msix_vectors_use(Rocker *r,
|
|
unsigned int num_vectors)
|
|
{
|
|
PCIDevice *dev = PCI_DEVICE(r);
|
|
int err;
|
|
int i;
|
|
|
|
for (i = 0; i < num_vectors; i++) {
|
|
err = msix_vector_use(dev, i);
|
|
if (err) {
|
|
goto rollback;
|
|
}
|
|
}
|
|
return 0;
|
|
|
|
rollback:
|
|
rocker_msix_vectors_unuse(r, i);
|
|
return err;
|
|
}
|
|
|
|
static int rocker_msix_init(Rocker *r, Error **errp)
|
|
{
|
|
PCIDevice *dev = PCI_DEVICE(r);
|
|
int err;
|
|
|
|
err = msix_init(dev, ROCKER_MSIX_VEC_COUNT(r->fp_ports),
|
|
&r->msix_bar,
|
|
ROCKER_PCI_MSIX_BAR_IDX, ROCKER_PCI_MSIX_TABLE_OFFSET,
|
|
&r->msix_bar,
|
|
ROCKER_PCI_MSIX_BAR_IDX, ROCKER_PCI_MSIX_PBA_OFFSET,
|
|
0, errp);
|
|
if (err) {
|
|
return err;
|
|
}
|
|
|
|
err = rocker_msix_vectors_use(r, ROCKER_MSIX_VEC_COUNT(r->fp_ports));
|
|
if (err) {
|
|
goto err_msix_vectors_use;
|
|
}
|
|
|
|
return 0;
|
|
|
|
err_msix_vectors_use:
|
|
msix_uninit(dev, &r->msix_bar, &r->msix_bar);
|
|
return err;
|
|
}
|
|
|
|
static void rocker_msix_uninit(Rocker *r)
|
|
{
|
|
PCIDevice *dev = PCI_DEVICE(r);
|
|
|
|
msix_uninit(dev, &r->msix_bar, &r->msix_bar);
|
|
rocker_msix_vectors_unuse(r, ROCKER_MSIX_VEC_COUNT(r->fp_ports));
|
|
}
|
|
|
|
static World *rocker_world_type_by_name(Rocker *r, const char *name)
|
|
{
|
|
int i;
|
|
|
|
for (i = 0; i < ROCKER_WORLD_TYPE_MAX; i++) {
|
|
if (strcmp(name, world_name(r->worlds[i])) == 0) {
|
|
return r->worlds[i];
|
|
}
|
|
}
|
|
return NULL;
|
|
}
|
|
|
|
static void pci_rocker_realize(PCIDevice *dev, Error **errp)
|
|
{
|
|
Rocker *r = ROCKER(dev);
|
|
const MACAddr zero = { .a = { 0, 0, 0, 0, 0, 0 } };
|
|
const MACAddr dflt = { .a = { 0x52, 0x54, 0x00, 0x12, 0x35, 0x01 } };
|
|
static int sw_index;
|
|
int i, err = 0;
|
|
|
|
/* allocate worlds */
|
|
|
|
r->worlds[ROCKER_WORLD_TYPE_OF_DPA] = of_dpa_world_alloc(r);
|
|
|
|
if (!r->world_name) {
|
|
r->world_name = g_strdup(world_name(r->worlds[ROCKER_WORLD_TYPE_OF_DPA]));
|
|
}
|
|
|
|
r->world_dflt = rocker_world_type_by_name(r, r->world_name);
|
|
if (!r->world_dflt) {
|
|
error_setg(errp,
|
|
"invalid argument requested world %s does not exist",
|
|
r->world_name);
|
|
goto err_world_type_by_name;
|
|
}
|
|
|
|
/* set up memory-mapped region at BAR0 */
|
|
|
|
memory_region_init_io(&r->mmio, OBJECT(r), &rocker_mmio_ops, r,
|
|
"rocker-mmio", ROCKER_PCI_BAR0_SIZE);
|
|
pci_register_bar(dev, ROCKER_PCI_BAR0_IDX,
|
|
PCI_BASE_ADDRESS_SPACE_MEMORY, &r->mmio);
|
|
|
|
/* set up memory-mapped region for MSI-X */
|
|
|
|
memory_region_init(&r->msix_bar, OBJECT(r), "rocker-msix-bar",
|
|
ROCKER_PCI_MSIX_BAR_SIZE);
|
|
pci_register_bar(dev, ROCKER_PCI_MSIX_BAR_IDX,
|
|
PCI_BASE_ADDRESS_SPACE_MEMORY, &r->msix_bar);
|
|
|
|
/* MSI-X init */
|
|
|
|
err = rocker_msix_init(r, errp);
|
|
if (err) {
|
|
goto err_msix_init;
|
|
}
|
|
|
|
/* validate switch properties */
|
|
|
|
if (!r->name) {
|
|
r->name = g_strdup(TYPE_ROCKER);
|
|
}
|
|
|
|
if (rocker_find(r->name)) {
|
|
error_setg(errp, "%s already exists", r->name);
|
|
goto err_duplicate;
|
|
}
|
|
|
|
/* Rocker name is passed in port name requests to OS with the intention
|
|
* that the name is used in interface names. Limit the length of the
|
|
* rocker name to avoid naming problems in the OS. Also, adding the
|
|
* port number as p# and unganged breakout b#, where # is at most 2
|
|
* digits, so leave room for it too (-1 for string terminator, -3 for
|
|
* p# and -3 for b#)
|
|
*/
|
|
#define ROCKER_IFNAMSIZ 16
|
|
#define MAX_ROCKER_NAME_LEN (ROCKER_IFNAMSIZ - 1 - 3 - 3)
|
|
if (strlen(r->name) > MAX_ROCKER_NAME_LEN) {
|
|
error_setg(errp,
|
|
"name too long; please shorten to at most %d chars",
|
|
MAX_ROCKER_NAME_LEN);
|
|
goto err_name_too_long;
|
|
}
|
|
|
|
if (memcmp(&r->fp_start_macaddr, &zero, sizeof(zero)) == 0) {
|
|
memcpy(&r->fp_start_macaddr, &dflt, sizeof(dflt));
|
|
r->fp_start_macaddr.a[4] += (sw_index++);
|
|
}
|
|
|
|
if (!r->switch_id) {
|
|
memcpy(&r->switch_id, &r->fp_start_macaddr,
|
|
sizeof(r->fp_start_macaddr));
|
|
}
|
|
|
|
if (r->fp_ports > ROCKER_FP_PORTS_MAX) {
|
|
r->fp_ports = ROCKER_FP_PORTS_MAX;
|
|
}
|
|
|
|
r->rings = g_new(DescRing *, rocker_pci_ring_count(r));
|
|
|
|
/* Rings are ordered like this:
|
|
* - command ring
|
|
* - event ring
|
|
* - port0 tx ring
|
|
* - port0 rx ring
|
|
* - port1 tx ring
|
|
* - port1 rx ring
|
|
* .....
|
|
*/
|
|
|
|
for (i = 0; i < rocker_pci_ring_count(r); i++) {
|
|
DescRing *ring = desc_ring_alloc(r, i);
|
|
|
|
if (i == ROCKER_RING_CMD) {
|
|
desc_ring_set_consume(ring, cmd_consume, ROCKER_MSIX_VEC_CMD);
|
|
} else if (i == ROCKER_RING_EVENT) {
|
|
desc_ring_set_consume(ring, NULL, ROCKER_MSIX_VEC_EVENT);
|
|
} else if (i % 2 == 0) {
|
|
desc_ring_set_consume(ring, tx_consume,
|
|
ROCKER_MSIX_VEC_TX((i - 2) / 2));
|
|
} else if (i % 2 == 1) {
|
|
desc_ring_set_consume(ring, NULL, ROCKER_MSIX_VEC_RX((i - 3) / 2));
|
|
}
|
|
|
|
r->rings[i] = ring;
|
|
}
|
|
|
|
for (i = 0; i < r->fp_ports; i++) {
|
|
FpPort *port =
|
|
fp_port_alloc(r, r->name, &r->fp_start_macaddr,
|
|
i, &r->fp_ports_peers[i]);
|
|
|
|
r->fp_port[i] = port;
|
|
fp_port_set_world(port, r->world_dflt);
|
|
}
|
|
|
|
QLIST_INSERT_HEAD(&rockers, r, next);
|
|
|
|
return;
|
|
|
|
err_name_too_long:
|
|
err_duplicate:
|
|
rocker_msix_uninit(r);
|
|
err_msix_init:
|
|
object_unparent(OBJECT(&r->msix_bar));
|
|
object_unparent(OBJECT(&r->mmio));
|
|
err_world_type_by_name:
|
|
for (i = 0; i < ROCKER_WORLD_TYPE_MAX; i++) {
|
|
if (r->worlds[i]) {
|
|
world_free(r->worlds[i]);
|
|
}
|
|
}
|
|
}
|
|
|
|
static void pci_rocker_uninit(PCIDevice *dev)
|
|
{
|
|
Rocker *r = ROCKER(dev);
|
|
int i;
|
|
|
|
QLIST_REMOVE(r, next);
|
|
|
|
for (i = 0; i < r->fp_ports; i++) {
|
|
FpPort *port = r->fp_port[i];
|
|
|
|
fp_port_free(port);
|
|
r->fp_port[i] = NULL;
|
|
}
|
|
|
|
for (i = 0; i < rocker_pci_ring_count(r); i++) {
|
|
if (r->rings[i]) {
|
|
desc_ring_free(r->rings[i]);
|
|
}
|
|
}
|
|
g_free(r->rings);
|
|
|
|
rocker_msix_uninit(r);
|
|
object_unparent(OBJECT(&r->msix_bar));
|
|
object_unparent(OBJECT(&r->mmio));
|
|
|
|
for (i = 0; i < ROCKER_WORLD_TYPE_MAX; i++) {
|
|
if (r->worlds[i]) {
|
|
world_free(r->worlds[i]);
|
|
}
|
|
}
|
|
g_free(r->fp_ports_peers);
|
|
}
|
|
|
|
static void rocker_reset(DeviceState *dev)
|
|
{
|
|
Rocker *r = ROCKER(dev);
|
|
int i;
|
|
|
|
for (i = 0; i < ROCKER_WORLD_TYPE_MAX; i++) {
|
|
if (r->worlds[i]) {
|
|
world_reset(r->worlds[i]);
|
|
}
|
|
}
|
|
for (i = 0; i < r->fp_ports; i++) {
|
|
fp_port_reset(r->fp_port[i]);
|
|
fp_port_set_world(r->fp_port[i], r->world_dflt);
|
|
}
|
|
|
|
r->test_reg = 0;
|
|
r->test_reg64 = 0;
|
|
r->test_dma_addr = 0;
|
|
r->test_dma_size = 0;
|
|
|
|
for (i = 0; i < rocker_pci_ring_count(r); i++) {
|
|
desc_ring_reset(r->rings[i]);
|
|
}
|
|
|
|
DPRINTF("Reset done\n");
|
|
}
|
|
|
|
static Property rocker_properties[] = {
|
|
DEFINE_PROP_STRING("name", Rocker, name),
|
|
DEFINE_PROP_STRING("world", Rocker, world_name),
|
|
DEFINE_PROP_MACADDR("fp_start_macaddr", Rocker,
|
|
fp_start_macaddr),
|
|
DEFINE_PROP_UINT64("switch_id", Rocker,
|
|
switch_id, 0),
|
|
DEFINE_PROP_ARRAY("ports", Rocker, fp_ports,
|
|
fp_ports_peers, qdev_prop_netdev, NICPeers),
|
|
DEFINE_PROP_END_OF_LIST(),
|
|
};
|
|
|
|
static const VMStateDescription rocker_vmsd = {
|
|
.name = TYPE_ROCKER,
|
|
.unmigratable = 1,
|
|
};
|
|
|
|
static void rocker_class_init(ObjectClass *klass, void *data)
|
|
{
|
|
DeviceClass *dc = DEVICE_CLASS(klass);
|
|
PCIDeviceClass *k = PCI_DEVICE_CLASS(klass);
|
|
|
|
k->realize = pci_rocker_realize;
|
|
k->exit = pci_rocker_uninit;
|
|
k->vendor_id = PCI_VENDOR_ID_REDHAT;
|
|
k->device_id = PCI_DEVICE_ID_REDHAT_ROCKER;
|
|
k->revision = ROCKER_PCI_REVISION;
|
|
k->class_id = PCI_CLASS_NETWORK_OTHER;
|
|
set_bit(DEVICE_CATEGORY_NETWORK, dc->categories);
|
|
dc->desc = "Rocker Switch";
|
|
dc->reset = rocker_reset;
|
|
dc->props = rocker_properties;
|
|
dc->vmsd = &rocker_vmsd;
|
|
}
|
|
|
|
static const TypeInfo rocker_info = {
|
|
.name = TYPE_ROCKER,
|
|
.parent = TYPE_PCI_DEVICE,
|
|
.instance_size = sizeof(Rocker),
|
|
.class_init = rocker_class_init,
|
|
.interfaces = (InterfaceInfo[]) {
|
|
{ INTERFACE_CONVENTIONAL_PCI_DEVICE },
|
|
{ },
|
|
},
|
|
};
|
|
|
|
static void rocker_register_types(void)
|
|
{
|
|
type_register_static(&rocker_info);
|
|
}
|
|
|
|
type_init(rocker_register_types)
|