/************************************************************************** Copyright (c) 2007, Chelsio Inc. All rights reserved. Redistribution and use in source and binary forms, with or without modification, are permitted provided that the following conditions are met: 1. Redistributions of source code must retain the above copyright notice, this list of conditions and the following disclaimer. 2. Neither the name of the Chelsio Corporation nor the names of its contributors may be used to endorse or promote products derived from this software without specific prior written permission. THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. ***************************************************************************/ #include #ifdef __NetBSD__ __KERNEL_RCSID(0, "$NetBSD: cxgb_main.c,v 1.11 2008/02/07 01:21:55 dyoung Exp $"); #endif #ifdef __FreeBSD__ __FBSDID("$FreeBSD: src/sys/dev/cxgb/cxgb_main.c,v 1.36 2007/09/11 23:49:27 kmacy Exp $"); #endif #include #include #include #ifdef __FreeBSD__ #include #include #include #endif #include #include #ifdef __FreeBSD__ #include #include #include #endif #include #include #ifdef __FreeBSD__ #include #include #endif #include #include #ifdef __FreeBSD__ #include #endif #include #include #ifdef __FreeBSD__ #include #endif #include #ifdef __FreeBSD__ #include #endif #include #include #include #include #include #include #include #ifdef __FreeBSD__ #include #endif #include #include #include #include #include #include #include #ifdef __FreeBSD__ #include #endif #ifdef CONFIG_DEFINED #include #else #ifdef __FreeBSD__ #include #endif #ifdef __NetBSD__ #include #endif #endif #ifdef PRIV_SUPPORTED #include #endif #ifdef __NetBSD__ #include #endif static int cxgb_setup_msix(adapter_t *, int); static void cxgb_teardown_msix(adapter_t *); #ifdef __FreeBSD__ static void cxgb_init(void *); #endif #ifdef __NetBSD__ static int cxgb_init(struct ifnet *); #endif static void cxgb_init_locked(struct port_info *); static void cxgb_stop_locked(struct port_info *); static void cxgb_set_rxmode(struct port_info *); #ifdef __FreeBSD__ static int cxgb_ioctl(struct ifnet *, unsigned long, caddr_t); #endif #ifdef __NetBSD__ static int cxgb_ioctl(struct ifnet *, unsigned long, void *); #endif static void cxgb_start(struct ifnet *); #ifdef __NetBSD__ static void cxgb_stop(struct ifnet *, int); #endif #ifdef __FreeBSD__ static void cxgb_start_proc(void *, int ncount); #endif #ifdef __NetBSD__ static void cxgb_start_proc(struct work *, void *); #endif static int cxgb_media_change(struct ifnet *); static void cxgb_media_status(struct ifnet *, struct ifmediareq *); static int setup_sge_qsets(adapter_t *); #ifdef __FreeBSD__ static void cxgb_async_intr(void *); static void cxgb_ext_intr_handler(void *, int); static void cxgb_tick_handler(void *, int); #endif #ifdef __NetBSD__ static int cxgb_async_intr(void *); static void cxgb_ext_intr_handler(struct work *, void *); static void cxgb_tick_handler(struct work *, void *); #endif static void cxgb_down_locked(struct adapter *sc); static void cxgb_tick(void *); static void setup_rss(adapter_t *sc); /* Attachment glue for the PCI controller end of the device. Each port of * the device is attached separately, as defined later. */ #ifdef __FreeBSD__ static int cxgb_controller_probe(device_t); static int cxgb_controller_attach(device_t); static int cxgb_controller_detach(device_t); #endif #ifdef __NetBSD__ static int cxgb_controller_match(device_t dev, struct cfdata *match, void *context); static void cxgb_controller_attach(device_t parent, device_t dev, void *context); static int cxgb_controller_detach(device_t dev, int flags); #endif static void cxgb_free(struct adapter *); static __inline void reg_block_dump(struct adapter *ap, uint8_t *buf, unsigned int start, unsigned int end); #ifdef __FreeBSD__ static void cxgb_get_regs(adapter_t *sc, struct ifconf_regs *regs, uint8_t *buf); static int cxgb_get_regs_len(void); static int offload_open(struct port_info *pi); #endif static void touch_bars(device_t dev); #ifdef notyet static int offload_close(struct toedev *tdev); #endif #ifdef __FreeBSD__ static device_method_t cxgb_controller_methods[] = { DEVMETHOD(device_probe, cxgb_controller_probe), DEVMETHOD(device_attach, cxgb_controller_attach), DEVMETHOD(device_detach, cxgb_controller_detach), /* bus interface */ DEVMETHOD(bus_print_child, bus_generic_print_child), DEVMETHOD(bus_driver_added, bus_generic_driver_added), { 0, 0 } }; static driver_t cxgb_controller_driver = { "cxgbc", cxgb_controller_methods, sizeof(struct adapter) }; static devclass_t cxgb_controller_devclass; DRIVER_MODULE(cxgbc, pci, cxgb_controller_driver, cxgb_controller_devclass, 0, 0); #endif #ifdef __NetBSD__ CFATTACH_DECL(cxgbc, sizeof(struct adapter), cxgb_controller_match, cxgb_controller_attach, cxgb_controller_detach, NULL); #endif /* * Attachment glue for the ports. Attachment is done directly to the * controller device. */ #ifdef __FreeBSD__ static int cxgb_port_probe(device_t); static int cxgb_port_attach(device_t); static int cxgb_port_detach(device_t); #endif #ifdef __NetBSD__ static int cxgb_port_match(device_t dev, struct cfdata *match, void *context); static void cxgb_port_attach(device_t dev, device_t self, void *context); static int cxgb_port_detach(device_t dev, int flags); #endif #ifdef __FreeBSD__ static device_method_t cxgb_port_methods[] = { DEVMETHOD(device_probe, cxgb_port_probe), DEVMETHOD(device_attach, cxgb_port_attach), DEVMETHOD(device_detach, cxgb_port_detach), { 0, 0 } }; static driver_t cxgb_port_driver = { "cxgb", cxgb_port_methods, 0 }; static d_ioctl_t cxgb_extension_ioctl; static d_open_t cxgb_extension_open; static d_close_t cxgb_extension_close; static struct cdevsw cxgb_cdevsw = { .d_version = D_VERSION, .d_flags = 0, .d_open = cxgb_extension_open, .d_close = cxgb_extension_close, .d_ioctl = cxgb_extension_ioctl, .d_name = "cxgb", }; static devclass_t cxgb_port_devclass; DRIVER_MODULE(cxgb, cxgbc, cxgb_port_driver, cxgb_port_devclass, 0, 0); #endif #ifdef __NetBSD__ CFATTACH_DECL(cxgb, sizeof(struct port_device), cxgb_port_match, cxgb_port_attach, cxgb_port_detach, NULL); #endif #define SGE_MSIX_COUNT (SGE_QSETS + 1) extern int collapse_mbufs; #ifdef MSI_SUPPORTED /* * The driver uses the best interrupt scheme available on a platform in the * order MSI-X, MSI, legacy pin interrupts. This parameter determines which * of these schemes the driver may consider as follows: * * msi = 2: choose from among all three options * msi = 1 : only consider MSI and pin interrupts * msi = 0: force pin interrupts */ static int msi_allowed = 2; #endif #ifdef __FreeBSD__ TUNABLE_INT("hw.cxgb.msi_allowed", &msi_allowed); SYSCTL_NODE(_hw, OID_AUTO, cxgb, CTLFLAG_RD, 0, "CXGB driver parameters"); SYSCTL_UINT(_hw_cxgb, OID_AUTO, msi_allowed, CTLFLAG_RDTUN, &msi_allowed, 0, "MSI-X, MSI, INTx selector"); /* * The driver enables offload as a default. * To disable it, use ofld_disable = 1. */ static int ofld_disable = 0; TUNABLE_INT("hw.cxgb.ofld_disable", &ofld_disable); SYSCTL_UINT(_hw_cxgb, OID_AUTO, ofld_disable, CTLFLAG_RDTUN, &ofld_disable, 0, "disable ULP offload"); /* * The driver uses an auto-queue algorithm by default. * To disable it and force a single queue-set per port, use singleq = 1. */ static int singleq = 1; TUNABLE_INT("hw.cxgb.singleq", &singleq); SYSCTL_UINT(_hw_cxgb, OID_AUTO, singleq, CTLFLAG_RDTUN, &singleq, 0, "use a single queue-set per port"); #endif #ifdef __NetBSD__ /* * The driver uses an auto-queue algorithm by default. * To disable it and force a single queue-set per port, use singleq = 1. */ static int singleq = 1; #endif enum { MAX_TXQ_ENTRIES = 16384, MAX_CTRL_TXQ_ENTRIES = 1024, MAX_RSPQ_ENTRIES = 16384, MAX_RX_BUFFERS = 16384, MAX_RX_JUMBO_BUFFERS = 16384, MIN_TXQ_ENTRIES = 4, MIN_CTRL_TXQ_ENTRIES = 4, MIN_RSPQ_ENTRIES = 32, MIN_FL_ENTRIES = 32, MIN_FL_JUMBO_ENTRIES = 32 }; struct filter_info { u32 sip; u32 sip_mask; u32 dip; u16 sport; u16 dport; u32 vlan:12; u32 vlan_prio:3; u32 mac_hit:1; u32 mac_idx:4; u32 mac_vld:1; u32 pkt_type:2; u32 report_filter_id:1; u32 pass:1; u32 rss:1; u32 qset:3; u32 locked:1; u32 valid:1; }; enum { FILTER_NO_VLAN_PRI = 7 }; #define PORT_MASK ((1 << MAX_NPORTS) - 1) /* Table for probing the cards. The desc field isn't actually used */ struct cxgb_ident { uint16_t vendor; uint16_t device; int index; const char *desc; } cxgb_identifiers[] = { {PCI_VENDOR_ID_CHELSIO, 0x0020, 0, "PE9000"}, {PCI_VENDOR_ID_CHELSIO, 0x0021, 1, "T302E"}, {PCI_VENDOR_ID_CHELSIO, 0x0022, 2, "T310E"}, {PCI_VENDOR_ID_CHELSIO, 0x0023, 3, "T320X"}, {PCI_VENDOR_ID_CHELSIO, 0x0024, 1, "T302X"}, {PCI_VENDOR_ID_CHELSIO, 0x0025, 3, "T320E"}, {PCI_VENDOR_ID_CHELSIO, 0x0026, 2, "T310X"}, {PCI_VENDOR_ID_CHELSIO, 0x0030, 2, "T3B10"}, {PCI_VENDOR_ID_CHELSIO, 0x0031, 3, "T3B20"}, {PCI_VENDOR_ID_CHELSIO, 0x0032, 1, "T3B02"}, {PCI_VENDOR_ID_CHELSIO, 0x0033, 4, "T3B04"}, {0, 0, 0, NULL} }; #ifdef __FreeBSD__ static int set_eeprom(struct port_info *pi, const uint8_t *data, int len, int offset); #endif static inline char t3rev2char(struct adapter *adapter) { char rev = 'z'; switch(adapter->params.rev) { case T3_REV_A: rev = 'a'; break; case T3_REV_B: case T3_REV_B2: rev = 'b'; break; case T3_REV_C: rev = 'c'; break; } return rev; } #ifdef __FreeBSD__ static struct cxgb_ident * cxgb_get_ident(device_t dev) { struct cxgb_ident *id; for (id = cxgb_identifiers; id->desc != NULL; id++) { if ((id->vendor == pci_get_vendor(dev)) && (id->device == pci_get_device(dev))) { return (id); } } return (NULL); } static const struct adapter_info * cxgb_get_adapter_info(device_t dev) { struct cxgb_ident *id; const struct adapter_info *ai; id = cxgb_get_ident(dev); if (id == NULL) return (NULL); ai = t3_get_adapter_info(id->index); return (ai); } static int cxgb_controller_probe(device_t dev) { const struct adapter_info *ai; char *ports, buf[80]; int nports; ai = cxgb_get_adapter_info(dev); if (ai == NULL) return (ENXIO); nports = ai->nports0 + ai->nports1; if (nports == 1) ports = "port"; else ports = "ports"; snprintf(buf, sizeof(buf), "%s RNIC, %d %s", ai->desc, nports, ports); device_set_desc_copy(dev, buf); return (BUS_PROBE_DEFAULT); } #endif #ifdef __NetBSD__ static struct cxgb_ident *cxgb_get_ident(struct pci_attach_args *pa) { struct cxgb_ident *id; int vendorid, deviceid; vendorid = PCI_VENDOR(pci_conf_read(pa->pa_pc, pa->pa_tag, PCI_ID_REG)); deviceid = PCI_PRODUCT(pci_conf_read(pa->pa_pc, pa->pa_tag, PCI_ID_REG)); for (id = cxgb_identifiers; id->desc != NULL; id++) { if ((id->vendor == vendorid) && (id->device == deviceid)) { return (id); } } return (NULL); } static const struct adapter_info *cxgb_get_adapter_info(struct pci_attach_args *pa) { struct cxgb_ident *id; const struct adapter_info *ai; id = cxgb_get_ident(pa); if (id == NULL) return (NULL); ai = t3_get_adapter_info(id->index); return (ai); } static int cxgb_controller_match(device_t dev, struct cfdata *match, void *context) { struct pci_attach_args *pa = context; const struct adapter_info *ai; ai = cxgb_get_adapter_info(pa); if (ai == NULL) return (0); return (100); // we ARE the best driver for this card!! } #endif #define FW_FNAME "t3fw%d%d%d" #define TPEEPROM_NAME "t3%ctpe%d%d%d" #define TPSRAM_NAME "t3%cps%d%d%d" #ifdef __FreeBSD__ static int upgrade_fw(adapter_t *sc) { char buf[32]; #ifdef FIRMWARE_LATEST const struct firmware *fw; #else struct firmware *fw; #endif int status; snprintf(&buf[0], sizeof(buf), FW_FNAME, FW_VERSION_MAJOR, FW_VERSION_MINOR, FW_VERSION_MICRO); #ifdef __FreeBSD__ fw = firmware_get(buf); #endif #ifdef __NetBSD__ fw = NULL; #endif if (fw == NULL) { device_printf(sc->dev, "Could not find firmware image %s\n", buf); return (ENOENT); } else device_printf(sc->dev, "updating firmware on card with %s\n", buf); status = t3_load_fw(sc, (const uint8_t *)fw->data, fw->datasize); device_printf(sc->dev, "firmware update returned %s %d\n", (status == 0) ? "success" : "fail", status); firmware_put(fw, FIRMWARE_UNLOAD); return (status); } #endif #ifdef __NetBSD__ int cxgb_cfprint(void *aux, const char *info); int cxgb_cfprint(void *aux, const char *info) { if (info) { printf("cxgb_cfprint(%p, \"%s\")\n", aux, info); INT3; } return (QUIET); } void cxgb_make_task(void *context) { struct cxgb_task *w = (struct cxgb_task *)context; // we can only use workqueue_create() once the system is up and running workqueue_create(&w->wq, w->name, w->func, w->context, PRIBIO, IPL_NET, 0); // printf("======>> create workqueue for %s %p\n", w->name, w->wq); } #endif #ifdef __FreeBSD__ static int cxgb_controller_attach(device_t dev) #endif #ifdef __NetBSD__ static void cxgb_controller_attach(device_t parent, device_t dev, void *context) #endif { device_t child; const struct adapter_info *ai; struct adapter *sc; #ifdef __NetBSD__ struct pci_attach_args *pa = context; struct cxgb_attach_args cxgb_args; int locs[2]; #endif int i, error = 0; uint32_t vers; int port_qsets = 1; int reg; #ifdef MSI_SUPPORTED int msi_needed; #endif #ifdef __FreeBSD__ sc = device_get_softc(dev); #endif #ifdef __NetBSD__ sc = (struct adapter *)dev; // device is first thing in adapter #endif sc->dev = dev; #ifdef __NetBSD__ memcpy(&sc->pa, pa, sizeof(struct pci_attach_args)); #endif sc->msi_count = 0; #ifdef __FreeBSD__ ai = cxgb_get_adapter_info(dev); #endif #ifdef __NetBSD__ ai = cxgb_get_adapter_info(pa); #endif /* * XXX not really related but a recent addition */ #ifdef MSI_SUPPORTED /* find the PCIe link width and set max read request to 4KB*/ if (pci_find_extcap(dev, PCIY_EXPRESS, ®) == 0) { uint16_t lnk, pectl; lnk = pci_read_config(dev, reg + 0x12, 2); sc->link_width = (lnk >> 4) & 0x3f; pectl = pci_read_config(dev, reg + 0x8, 2); pectl = (pectl & ~0x7000) | (5 << 12); pci_write_config(dev, reg + 0x8, pectl, 2); } if (sc->link_width != 0 && sc->link_width <= 4 && (ai->nports0 + ai->nports1) <= 2) { device_printf(sc->dev, "PCIe x%d Link, expect reduced performance\n", sc->link_width); } #endif touch_bars(dev); pci_enable_busmaster(dev); /* * Allocate the registers and make them available to the driver. * The registers that we care about for NIC mode are in BAR 0 */ #ifdef __FreeBSD__ sc->regs_rid = PCIR_BAR(0); if ((sc->regs_res = bus_alloc_resource_any(dev, SYS_RES_MEMORY, &sc->regs_rid, RF_ACTIVE)) == NULL) { device_printf(dev, "Cannot allocate BAR\n"); return (ENXIO); } snprintf(sc->lockbuf, ADAPTER_LOCK_NAME_LEN, "cxgb controller lock %d", device_get_unit(dev)); ADAPTER_LOCK_INIT(sc, sc->lockbuf); snprintf(sc->reglockbuf, ADAPTER_LOCK_NAME_LEN, "SGE reg lock %d", device_get_unit(dev)); snprintf(sc->mdiolockbuf, ADAPTER_LOCK_NAME_LEN, "cxgb mdio lock %d", device_get_unit(dev)); snprintf(sc->elmerlockbuf, ADAPTER_LOCK_NAME_LEN, "cxgb elmer lock %d", device_get_unit(dev)); #endif #ifdef __NetBSD__ sc->regs_rid = PCI_MAPREG_START; t3_os_pci_read_config_4(sc, PCI_MAPREG_START, ®); // call bus_space_map sc->bar0 = reg&0xFFFFF000; bus_space_map(sc->pa.pa_memt, sc->bar0, 4096, 0, &sc->bar0_handle); #endif MTX_INIT(&sc->sge.reg_lock, sc->reglockbuf, NULL, MTX_DEF); MTX_INIT(&sc->mdio_lock, sc->mdiolockbuf, NULL, MTX_DEF); MTX_INIT(&sc->elmer_lock, sc->elmerlockbuf, NULL, MTX_DEF); #ifdef __FreeBSD__ sc->bt = rman_get_bustag(sc->regs_res); sc->bh = rman_get_bushandle(sc->regs_res); sc->mmio_len = rman_get_size(sc->regs_res); #endif #ifdef __NetBSD__ sc->bt = sc->pa.pa_memt; sc->bh = sc->bar0_handle; sc->mmio_len = 4096; #endif if (t3_prep_adapter(sc, ai, 1) < 0) { printf("prep adapter failed\n"); error = ENODEV; goto out; } /* Allocate the BAR for doing MSI-X. If it succeeds, try to allocate * enough messages for the queue sets. If that fails, try falling * back to MSI. If that fails, then try falling back to the legacy * interrupt pin model. */ #ifdef MSI_SUPPORTED sc->msix_regs_rid = 0x20; if ((msi_allowed >= 2) && (sc->msix_regs_res = bus_alloc_resource_any(dev, SYS_RES_MEMORY, &sc->msix_regs_rid, RF_ACTIVE)) != NULL) { msi_needed = sc->msi_count = SGE_MSIX_COUNT; if (((error = pci_alloc_msix(dev, &sc->msi_count)) != 0) || (sc->msi_count != msi_needed)) { device_printf(dev, "msix allocation failed - msi_count = %d" " msi_needed=%d will try msi err=%d\n", sc->msi_count, msi_needed, error); sc->msi_count = 0; pci_release_msi(dev); bus_release_resource(dev, SYS_RES_MEMORY, sc->msix_regs_rid, sc->msix_regs_res); sc->msix_regs_res = NULL; } else { sc->flags |= USING_MSIX; sc->cxgb_intr = t3_intr_msix; } } if ((msi_allowed >= 1) && (sc->msi_count == 0)) { sc->msi_count = 1; if (pci_alloc_msi(dev, &sc->msi_count)) { device_printf(dev, "alloc msi failed - will try INTx\n"); sc->msi_count = 0; pci_release_msi(dev); } else { sc->flags |= USING_MSI; sc->irq_rid = 1; sc->cxgb_intr = t3_intr_msi; } } #endif if (sc->msi_count == 0) { device_printf(dev, "using line interrupts\n"); sc->irq_rid = 0; sc->cxgb_intr = t3b_intr; } #ifdef __FreeBSD__ /* Create a private taskqueue thread for handling driver events */ #ifdef TASKQUEUE_CURRENT sc->tq = taskqueue_create("cxgb_taskq", M_NOWAIT, taskqueue_thread_enqueue, &sc->tq); #else sc->tq = taskqueue_create_fast("cxgb_taskq", M_NOWAIT, taskqueue_thread_enqueue, &sc->tq); #endif if (sc->tq == NULL) { device_printf(dev, "failed to allocate controller task queue\n"); goto out; } taskqueue_start_threads(&sc->tq, 1, PI_NET, "%s taskq", device_get_nameunit(dev)); TASK_INIT(&sc->ext_intr_task, 0, cxgb_ext_intr_handler, sc); TASK_INIT(&sc->tick_task, 0, cxgb_tick_handler, sc); #endif #ifdef __NetBSD__ sc->ext_intr_task.name = "cxgb_ext_intr_handler"; sc->ext_intr_task.func = cxgb_ext_intr_handler; sc->ext_intr_task.context = sc; kthread_create(PRI_NONE, 0, NULL, cxgb_make_task, &sc->ext_intr_task, NULL, "cxgb_make_task"); sc->tick_task.name = "cxgb_tick_handler"; sc->tick_task.func = cxgb_tick_handler; sc->tick_task.context = sc; kthread_create(PRI_NONE, 0, NULL, cxgb_make_task, &sc->tick_task, NULL, "cxgb_make_task"); #endif /* Create a periodic callout for checking adapter status */ #ifdef __FreeBSD__ callout_init(&sc->cxgb_tick_ch, TRUE); #endif #ifdef __NetBSD__ callout_init(&sc->cxgb_tick_ch, 0); #endif if (t3_check_fw_version(sc) != 0) { /* * Warn user that a firmware update will be attempted in init. */ device_printf(dev, "firmware needs to be updated to version %d.%d.%d\n", FW_VERSION_MAJOR, FW_VERSION_MINOR, FW_VERSION_MICRO); sc->flags &= ~FW_UPTODATE; } else { sc->flags |= FW_UPTODATE; } if (t3_check_tpsram_version(sc) != 0) { /* * Warn user that a firmware update will be attempted in init. */ device_printf(dev, "SRAM needs to be updated to version %c-%d.%d.%d\n", t3rev2char(sc), TP_VERSION_MAJOR, TP_VERSION_MINOR, TP_VERSION_MICRO); sc->flags &= ~TPS_UPTODATE; } else { sc->flags |= TPS_UPTODATE; } if ((sc->flags & USING_MSIX) && !singleq) #ifdef __FreeBSD__ port_qsets = min((SGE_QSETS/(sc)->params.nports), mp_ncpus); #endif #ifdef __NetBSD__ port_qsets = (SGE_QSETS/(sc)->params.nports); #endif /* * Create a child device for each MAC. The ethernet attachment * will be done in these children. */ for (i = 0; i < (sc)->params.nports; i++) { struct port_info *pi; #ifdef __FreeBSD__ if ((child = device_add_child(dev, "cxgb", -1)) == NULL) { device_printf(dev, "failed to add child port\n"); error = EINVAL; goto out; } #endif pi = &sc->port[i]; pi->adapter = sc; pi->nqsets = port_qsets; pi->first_qset = i*port_qsets; pi->port_id = i; pi->tx_chan = i >= ai->nports0; pi->txpkt_intf = pi->tx_chan ? 2 * (i - ai->nports0) + 1 : 2 * i; sc->rxpkt_map[pi->txpkt_intf] = i; #ifdef __NetBSD__ cxgb_args.port = i; locs[0] = 1; locs[1] = i; printf("\n"); // for cleaner formatting in dmesg child = config_found_sm_loc(dev, "cxgbc", locs, &cxgb_args, cxgb_cfprint, config_stdsubmatch); printf("\n"); // for cleaner formatting in dmesg #endif sc->portdev[i] = child; #ifdef __FreeBSD__ device_set_softc(child, pi); #endif } #ifdef __FreeBSD__ if ((error = bus_generic_attach(dev)) != 0) goto out; #endif /* * XXX need to poll for link status */ sc->params.stats_update_period = 1; /* initialize sge private state */ t3_sge_init_adapter(sc); t3_led_ready(sc); #ifdef __FreeBSD__ cxgb_offload_init(); if (is_offload(sc)) { setbit(&sc->registered_device_map, OFFLOAD_DEVMAP_BIT); cxgb_adapter_ofld(sc); } #endif error = t3_get_fw_version(sc, &vers); if (error) goto out; snprintf(&sc->fw_version[0], sizeof(sc->fw_version), "%d.%d.%d", G_FW_VERSION_MAJOR(vers), G_FW_VERSION_MINOR(vers), G_FW_VERSION_MICRO(vers)); #ifdef __FreeBSD__ printf("******** firmware rev %s\n", sc->fw_version); #endif #ifdef __FreeBSD__ t3_add_sysctls(sc); #endif out: if (error) { cxgb_free(sc); } #ifdef __FreeBSD__ return (error); #endif } static int #ifdef __FreeBSD__ cxgb_controller_detach(device_t dev) #endif #ifdef __NetBSD__ cxgb_controller_detach(device_t dev, int flags) #endif { struct adapter *sc; #ifdef __FreeBSD__ sc = device_get_softc(dev); #endif #ifdef __NetBSD__ sc = (struct adapter *)dev; #endif cxgb_free(sc); return (0); } static void cxgb_free(struct adapter *sc) { int i; ADAPTER_LOCK(sc); /* * drops the lock */ cxgb_down_locked(sc); #ifdef MSI_SUPPORTED if (sc->flags & (USING_MSI | USING_MSIX)) { device_printf(sc->dev, "releasing msi message(s)\n"); pci_release_msi(sc->dev); } else { device_printf(sc->dev, "no msi message to release\n"); } if (sc->msix_regs_res != NULL) { bus_release_resource(sc->dev, SYS_RES_MEMORY, sc->msix_regs_rid, sc->msix_regs_res); } #endif #ifdef __FreeBSD__ if (sc->tq != NULL) { taskqueue_drain(sc->tq, &sc->ext_intr_task); taskqueue_drain(sc->tq, &sc->tick_task); } #endif t3_sge_deinit_sw(sc); /* * Wait for last callout */ tsleep(&sc, 0, "cxgb unload", 3*hz); for (i = 0; i < (sc)->params.nports; ++i) { if (sc->portdev[i] != NULL) #ifdef __FreeBSD__ device_delete_child(sc->dev, sc->portdev[i]); #endif #ifdef __NetBSD__ { INT3; } #endif } #ifdef __FreeBSD__ bus_generic_detach(sc->dev); if (sc->tq != NULL) taskqueue_free(sc->tq); #endif #ifdef notyet if (is_offload(sc)) { cxgb_adapter_unofld(sc); if (isset(&sc->open_device_map, OFFLOAD_DEVMAP_BIT)) offload_close(&sc->tdev); } #endif t3_free_sge_resources(sc); free(sc->filters, M_DEVBUF); t3_sge_free(sc); #ifdef __FreeBSD__ cxgb_offload_exit(); if (sc->regs_res != NULL) bus_release_resource(sc->dev, SYS_RES_MEMORY, sc->regs_rid, sc->regs_res); #endif MTX_DESTROY(&sc->mdio_lock); MTX_DESTROY(&sc->sge.reg_lock); MTX_DESTROY(&sc->elmer_lock); ADAPTER_LOCK_DEINIT(sc); return; } /** * setup_sge_qsets - configure SGE Tx/Rx/response queues * @sc: the controller softc * * Determines how many sets of SGE queues to use and initializes them. * We support multiple queue sets per port if we have MSI-X, otherwise * just one queue set per port. */ static int setup_sge_qsets(adapter_t *sc) { int i, j, err, irq_idx = 0, qset_idx = 0; u_int ntxq = SGE_TXQ_PER_SET; if ((err = t3_sge_alloc(sc)) != 0) { device_printf(sc->dev, "t3_sge_alloc returned %d\n", err); return (err); } if (sc->params.rev > 0 && !(sc->flags & USING_MSI)) irq_idx = -1; for (i = 0; i < (sc)->params.nports; i++) { struct port_info *pi = &sc->port[i]; for (j = 0; j < pi->nqsets; j++, qset_idx++) { err = t3_sge_alloc_qset(sc, qset_idx, (sc)->params.nports, (sc->flags & USING_MSIX) ? qset_idx + 1 : irq_idx, &sc->params.sge.qset[qset_idx], ntxq, pi); if (err) { t3_free_sge_resources(sc); device_printf(sc->dev, "t3_sge_alloc_qset failed with %d\n", err); return (err); } } } return (0); } static void cxgb_teardown_msix(adapter_t *sc) { int i, nqsets; for (nqsets = i = 0; i < (sc)->params.nports; i++) nqsets += sc->port[i].nqsets; for (i = 0; i < nqsets; i++) { if (sc->msix_intr_tag[i] != NULL) { #ifdef __FreeBSD__ bus_teardown_intr(sc->dev, sc->msix_irq_res[i], sc->msix_intr_tag[i]); #endif sc->msix_intr_tag[i] = NULL; } if (sc->msix_irq_res[i] != NULL) { #ifdef __FreeBSD__ bus_release_resource(sc->dev, SYS_RES_IRQ, sc->msix_irq_rid[i], sc->msix_irq_res[i]); #endif sc->msix_irq_res[i] = NULL; } } } static int cxgb_setup_msix(adapter_t *sc, int msix_count) { int i, j, k, nqsets, rid; /* The first message indicates link changes and error conditions */ sc->irq_rid = 1; #ifdef __FreeBSD__ if ((sc->irq_res = bus_alloc_resource_any(sc->dev, SYS_RES_IRQ, &sc->irq_rid, RF_SHAREABLE | RF_ACTIVE)) == NULL) { device_printf(sc->dev, "Cannot allocate msix interrupt\n"); return (EINVAL); } if (bus_setup_intr(sc->dev, sc->irq_res, INTR_MPSAFE|INTR_TYPE_NET, #ifdef INTR_FILTERS NULL, #endif cxgb_async_intr, sc, &sc->intr_tag)) { device_printf(sc->dev, "Cannot set up interrupt\n"); return (EINVAL); } #endif #ifdef __NetBSD__ /* Allocate PCI interrupt resources. */ if (pci_intr_map(&sc->pa, &sc->intr_handle)) { printf("cxgb_setup_msix(%d): pci_intr_map() failed\n", __LINE__); return (EINVAL); } sc->intr_cookie = pci_intr_establish(sc->pa.pa_pc, sc->intr_handle, IPL_NET, cxgb_async_intr, sc); if (sc->intr_cookie == NULL) { printf("cxgb_setup_msix(%d): pci_intr_establish() failed\n", __LINE__); return (EINVAL); } #endif for (i = k = 0; i < (sc)->params.nports; i++) { nqsets = sc->port[i].nqsets; for (j = 0; j < nqsets; j++, k++) { #ifdef __FreeBSD__ struct sge_qset *qs = &sc->sge.qs[k]; #endif rid = k + 2; if (cxgb_debug) printf("rid=%d ", rid); #ifdef __FreeBSD__ if ((sc->msix_irq_res[k] = bus_alloc_resource_any( sc->dev, SYS_RES_IRQ, &rid, RF_SHAREABLE | RF_ACTIVE)) == NULL) { device_printf(sc->dev, "Cannot allocate " "interrupt for message %d\n", rid); return (EINVAL); } sc->msix_irq_rid[k] = rid; printf("setting up interrupt for port=%d\n", qs->port->port_id); if (bus_setup_intr(sc->dev, sc->msix_irq_res[k], INTR_MPSAFE|INTR_TYPE_NET, #ifdef INTR_FILTERS NULL, #endif t3_intr_msix, qs, &sc->msix_intr_tag[k])) { device_printf(sc->dev, "Cannot set up " "interrupt for message %d\n", rid); return (EINVAL); } #endif #ifdef __NetBSD__ INT3; #endif } } return (0); } #ifdef __FreeBSD__ static int cxgb_port_probe(device_t dev) { struct port_info *p; char buf[80]; p = device_get_softc(dev); snprintf(buf, sizeof(buf), "Port %d %s", p->port_id, p->port_type->desc); device_set_desc_copy(dev, buf); return (0); } #endif #ifdef __NetBSD__ static int cxgb_port_match(device_t dev, struct cfdata *match, void *context) { return (100); } #endif #ifdef __FreeBSD__ static int cxgb_makedev(struct port_info *pi) { pi->port_cdev = make_dev(&cxgb_cdevsw, pi->ifp->if_dunit, UID_ROOT, GID_WHEEL, 0600, if_name(pi->ifp)); if (pi->port_cdev == NULL) return (ENOMEM); pi->port_cdev->si_drv1 = (void *)pi; return (0); } #ifdef TSO_SUPPORTED #define CXGB_CAP (IFCAP_VLAN_HWTAGGING | IFCAP_VLAN_MTU | IFCAP_HWCSUM | IFCAP_VLAN_HWCSUM | IFCAP_TSO | IFCAP_JUMBO_MTU) /* Don't enable TSO6 yet */ #define CXGB_CAP_ENABLE (IFCAP_VLAN_HWTAGGING | IFCAP_VLAN_MTU | IFCAP_HWCSUM | IFCAP_VLAN_HWCSUM | IFCAP_TSO4 | IFCAP_JUMBO_MTU) #else #define CXGB_CAP (IFCAP_VLAN_HWTAGGING | IFCAP_VLAN_MTU | IFCAP_HWCSUM | IFCAP_JUMBO_MTU) /* Don't enable TSO6 yet */ #define CXGB_CAP_ENABLE (IFCAP_VLAN_HWTAGGING | IFCAP_VLAN_MTU | IFCAP_HWCSUM | IFCAP_JUMBO_MTU) #define IFCAP_TSO4 0x0 #define IFCAP_TSO6 0x0 #define CSUM_TSO 0x0 #endif #endif #ifdef __NetBSD__ #define IFCAP_HWCSUM (IFCAP_CSUM_IPv4_Rx | IFCAP_CSUM_IPv4_Tx) #define IFCAP_RXCSUM IFCAP_CSUM_IPv4_Rx #define IFCAP_TXCSUM IFCAP_CSUM_IPv4_Tx #ifdef TSO_SUPPORTED #define CXGB_CAP (IFCAP_HWCSUM | IFCAP_TSO) /* Don't enable TSO6 yet */ #define CXGB_CAP_ENABLE (IFCAP_HWCSUM | IFCAP_TSO4) #else #define CXGB_CAP (IFCAP_HWCSUM) /* Don't enable TSO6 yet */ #define CXGB_CAP_ENABLE (IFCAP_HWCSUM) #define IFCAP_TSO4 0x0 #define IFCAP_TSO6 0x0 #define CSUM_TSO 0x0 #endif #endif #ifdef __FreeBSD__ static int cxgb_port_attach(device_t dev) #endif #ifdef __NetBSD__ static void cxgb_port_attach(device_t dev, device_t self, void *context) #endif { struct port_info *p; #ifdef __NetBSD__ struct port_device *pd; int *port_number = (int *)context; char buf[32]; #endif struct ifnet *ifp; int media_flags; #ifdef __FreeBSD__ int err; p = device_get_softc(dev); #endif #ifdef __NetBSD__ pd = (struct port_device *)self; // device is first element in port_device pd->dev = self; pd->parent = (struct adapter *)dev; pd->port_number = *port_number; p = &pd->parent->port[*port_number]; p->pd = pd; #endif #ifdef __FreeBSD__ snprintf(p->lockbuf, PORT_NAME_LEN, "cxgb port lock %d:%d", device_get_unit(device_get_parent(dev)), p->port_id); #endif PORT_LOCK_INIT(p, p->lockbuf); /* Allocate an ifnet object and set it up */ #ifdef __FreeBSD__ ifp = p->ifp = if_alloc(IFT_ETHER); #endif #ifdef __NetBSD__ ifp = p->ifp = (void *)malloc(sizeof (struct ifnet), M_IFADDR, M_WAITOK); #endif if (ifp == NULL) { device_printf(dev, "Cannot allocate ifnet\n"); #ifdef __FreeBSD__ return (ENOMEM); #endif #ifdef __NetBSD__ return; #endif } #ifdef __NetBSD__ memset(ifp, 0, sizeof(struct ifnet)); #endif /* * Note that there is currently no watchdog timer. */ #ifdef __FreeBSD__ if_initname(ifp, device_get_name(dev), device_get_unit(dev)); #endif #ifdef __NetBSD__ snprintf(buf, sizeof(buf), "cxgb%d", p->port); strcpy(ifp->if_xname, buf); #endif ifp->if_init = cxgb_init; ifp->if_softc = p; ifp->if_flags = IFF_BROADCAST | IFF_SIMPLEX | IFF_MULTICAST; ifp->if_ioctl = cxgb_ioctl; ifp->if_start = cxgb_start; #ifdef __NetBSD__ ifp->if_stop = cxgb_stop; #endif ifp->if_timer = 0; /* Disable ifnet watchdog */ ifp->if_watchdog = NULL; #ifdef __FreeBSD__ ifp->if_snd.ifq_drv_maxlen = TX_ETH_Q_SIZE; IFQ_SET_MAXLEN(&ifp->if_snd, ifp->if_snd.ifq_drv_maxlen); #endif #ifdef __NetBSD__ ifp->if_snd.ifq_maxlen = TX_ETH_Q_SIZE; IFQ_SET_MAXLEN(&ifp->if_snd, ifp->if_snd.ifq_maxlen); #endif IFQ_SET_READY(&ifp->if_snd); #ifdef __FreeBSD__ ifp->if_hwassist = ifp->if_capabilities = ifp->if_capenable = 0; ifp->if_capabilities |= CXGB_CAP; ifp->if_capenable |= CXGB_CAP_ENABLE; ifp->if_hwassist |= (CSUM_TCP | CSUM_UDP | CSUM_IP | CSUM_TSO); #endif #ifdef __NetBSD__ ifp->if_capabilities = ifp->if_capenable = 0; #endif ifp->if_baudrate = 10000000000; // 10 Gbps /* * disable TSO on 4-port - it isn't supported by the firmware yet */ if (p->adapter->params.nports > 2) { ifp->if_capabilities &= ~(IFCAP_TSO4 | IFCAP_TSO6); ifp->if_capenable &= ~(IFCAP_TSO4 | IFCAP_TSO6); #ifdef __FreeBSD__ ifp->if_hwassist &= ~CSUM_TSO; #endif } #ifdef __NetBSD__ if_attach(ifp); #endif ether_ifattach(ifp, p->hw_addr); /* * Only default to jumbo frames on 10GigE */ if (p->adapter->params.nports <= 2) ifp->if_mtu = 9000; #ifdef __FreeBSD__ if ((err = cxgb_makedev(p)) != 0) { printf("makedev failed %d\n", err); return (err); } #endif ifmedia_init(&p->media, IFM_IMASK, cxgb_media_change, cxgb_media_status); if (!strcmp(p->port_type->desc, "10GBASE-CX4")) { media_flags = IFM_ETHER | IFM_10G_CX4 | IFM_FDX; } else if (!strcmp(p->port_type->desc, "10GBASE-SR")) { media_flags = IFM_ETHER | IFM_10G_SR | IFM_FDX; } else if (!strcmp(p->port_type->desc, "10GBASE-XR")) { media_flags = IFM_ETHER | IFM_10G_LR | IFM_FDX; } else if (!strcmp(p->port_type->desc, "10/100/1000BASE-T")) { ifmedia_add(&p->media, IFM_ETHER | IFM_10_T, 0, NULL); ifmedia_add(&p->media, IFM_ETHER | IFM_10_T | IFM_FDX, 0, NULL); ifmedia_add(&p->media, IFM_ETHER | IFM_100_TX, 0, NULL); ifmedia_add(&p->media, IFM_ETHER | IFM_100_TX | IFM_FDX, 0, NULL); ifmedia_add(&p->media, IFM_ETHER | IFM_1000_T | IFM_FDX, 0, NULL); media_flags = 0; } else { printf("unsupported media type %s\n", p->port_type->desc); #ifdef __FreeBSD__ return (ENXIO); #endif #ifdef __NetBSD__ return; #endif } if (media_flags) { ifmedia_add(&p->media, media_flags, 0, NULL); ifmedia_set(&p->media, media_flags); } else { ifmedia_add(&p->media, IFM_ETHER | IFM_AUTO, 0, NULL); ifmedia_set(&p->media, IFM_ETHER | IFM_AUTO); } snprintf(p->taskqbuf, TASKQ_NAME_LEN, "cxgb_port_taskq%d", p->port_id); #ifdef __FreeBSD__ #ifdef TASKQUEUE_CURRENT /* Create a port for handling TX without starvation */ p->tq = taskqueue_create(p->taskqbuf, M_NOWAIT, taskqueue_thread_enqueue, &p->tq); #else /* Create a port for handling TX without starvation */ p->tq = taskqueue_create_fast(p->taskqbuf, M_NOWAIT, taskqueue_thread_enqueue, &p->tq); #endif if (p->tq == NULL) { device_printf(dev, "failed to allocate port task queue\n"); return (ENOMEM); } taskqueue_start_threads(&p->tq, 1, PI_NET, "%s taskq", device_get_nameunit(dev)); TASK_INIT(&p->start_task, 0, cxgb_start_proc, ifp); #endif #ifdef __NetBSD__ p->start_task.name = "cxgb_start_proc"; p->start_task.func = cxgb_start_proc; p->start_task.context = ifp; kthread_create(PRI_NONE, 0, NULL, cxgb_make_task, &p->start_task, NULL, "cxgb_make_task"); #endif t3_sge_init_port(p); #ifdef __FreeBSD__ return (0); #endif } static int #ifdef __FreeBSD__ cxgb_port_detach(device_t dev) #endif #ifdef __NetBSD__ cxgb_port_detach(device_t dev, int flags) #endif { struct port_info *p; #ifdef __FreeBSD__ p = device_get_softc(dev); #endif #ifdef __NetBSD__ p = (struct port_info *)dev; // device is first thing in adapter #endif PORT_LOCK(p); if (p->ifp->if_drv_flags & IFF_DRV_RUNNING) cxgb_stop_locked(p); PORT_UNLOCK(p); #ifdef __FreeBSD__ if (p->tq != NULL) { taskqueue_drain(p->tq, &p->start_task); taskqueue_free(p->tq); p->tq = NULL; } #endif #ifdef __NetBSD__ if (p->start_task.wq != NULL) { workqueue_destroy(p->start_task.wq); p->start_task.wq = NULL; } #endif ether_ifdetach(p->ifp); /* * the lock may be acquired in ifdetach */ PORT_LOCK_DEINIT(p); #ifdef __FreeBSD__ if_free(p->ifp); if (p->port_cdev != NULL) destroy_dev(p->port_cdev); #endif #ifdef __NetBSD__ if_detach(p->ifp); #endif return (0); } void t3_fatal_err(struct adapter *sc) { u_int fw_status[4]; if (sc->flags & FULL_INIT_DONE) { t3_sge_stop(sc); t3_write_reg(sc, A_XGM_TX_CTRL, 0); t3_write_reg(sc, A_XGM_RX_CTRL, 0); t3_write_reg(sc, XGM_REG(A_XGM_TX_CTRL, 1), 0); t3_write_reg(sc, XGM_REG(A_XGM_RX_CTRL, 1), 0); t3_intr_disable(sc); } device_printf(sc->dev,"encountered fatal error, operation suspended\n"); if (!t3_cim_ctl_blk_read(sc, 0xa0, 4, fw_status)) device_printf(sc->dev, "FW_ status: 0x%x, 0x%x, 0x%x, 0x%x\n", fw_status[0], fw_status[1], fw_status[2], fw_status[3]); } int t3_os_find_pci_capability(adapter_t *sc, int cap) { device_t dev; #ifdef __FreeBSD__ struct pci_devinfo *dinfo; pcicfgregs *cfg; uint32_t status; uint8_t ptr; dev = sc->dev; dinfo = device_get_ivars(dev); cfg = &dinfo->cfg; status = pci_read_config(dev, PCIR_STATUS, 2); if (!(status & PCIM_STATUS_CAPPRESENT)) return (0); switch (cfg->hdrtype & PCIM_HDRTYPE) { case 0: case 1: ptr = PCIR_CAP_PTR; break; case 2: ptr = PCIR_CAP_PTR_2; break; default: return (0); break; } ptr = pci_read_config(dev, ptr, 1); while (ptr != 0) { if (pci_read_config(dev, ptr + PCICAP_ID, 1) == cap) return (ptr); ptr = pci_read_config(dev, ptr + PCICAP_NEXTPTR, 1); } #endif #ifdef __NetBSD__ uint32_t status; uint32_t bhlc; uint32_t temp; uint8_t ptr; dev = sc->dev; status = pci_conf_read(sc->pa.pa_pc, sc->pa.pa_tag, PCI_COMMAND_STATUS_REG); if (!(status&PCI_STATUS_CAPLIST_SUPPORT)) return (0); bhlc = pci_conf_read(sc->pa.pa_pc, sc->pa.pa_tag, PCI_BHLC_REG); switch (PCI_HDRTYPE(bhlc)) { case 0: case 1: ptr = PCI_CAPLISTPTR_REG; break; case 2: ptr = PCI_CARDBUS_CAPLISTPTR_REG; break; default: return (0); } temp = pci_conf_read(sc->pa.pa_pc, sc->pa.pa_tag, ptr); ptr = PCI_CAPLIST_PTR(temp); while (ptr != 0) { temp = pci_conf_read(sc->pa.pa_pc, sc->pa.pa_tag, ptr); if (PCI_CAPLIST_CAP(temp) == cap) return (ptr); ptr = PCI_CAPLIST_NEXT(temp); } #endif return (0); } int t3_os_pci_save_state(struct adapter *sc) { #ifdef __FreeBSD__ device_t dev; struct pci_devinfo *dinfo; dev = sc->dev; dinfo = device_get_ivars(dev); pci_cfg_save(dev, dinfo, 0); #endif #ifdef __NetBSD__ INT3; #endif return (0); } int t3_os_pci_restore_state(struct adapter *sc) { #ifdef __FreeBSD__ device_t dev; struct pci_devinfo *dinfo; dev = sc->dev; dinfo = device_get_ivars(dev); pci_cfg_restore(dev, dinfo); #endif #ifdef __NetBSD__ INT3; #endif return (0); } /** * t3_os_link_changed - handle link status changes * @adapter: the adapter associated with the link change * @port_id: the port index whose limk status has changed * @link_stat: the new status of the link * @speed: the new speed setting * @duplex: the new duplex setting * @fc: the new flow-control setting * * This is the OS-dependent handler for link status changes. The OS * neutral handler takes care of most of the processing for these events, * then calls this handler for any OS-specific processing. */ void t3_os_link_changed(adapter_t *adapter, int port_id, int link_status, int speed, int duplex, int fc) { struct port_info *pi = &adapter->port[port_id]; struct cmac *mac = &adapter->port[port_id].mac; if ((pi->ifp->if_flags & IFF_UP) == 0) return; if (link_status) { t3_mac_enable(mac, MAC_DIRECTION_RX); if_link_state_change(pi->ifp, LINK_STATE_UP); } else { if_link_state_change(pi->ifp, LINK_STATE_DOWN); pi->phy.ops->power_down(&pi->phy, 1); t3_mac_disable(mac, MAC_DIRECTION_RX); t3_link_start(&pi->phy, mac, &pi->link_config); } } /* * Interrupt-context handler for external (PHY) interrupts. */ void t3_os_ext_intr_handler(adapter_t *sc) { if (cxgb_debug) printf("t3_os_ext_intr_handler\n"); /* * Schedule a task to handle external interrupts as they may be slow * and we use a mutex to protect MDIO registers. We disable PHY * interrupts in the meantime and let the task reenable them when * it's done. */ ADAPTER_LOCK(sc); if (sc->slow_intr_mask) { sc->slow_intr_mask &= ~F_T3DBG; t3_write_reg(sc, A_PL_INT_ENABLE0, sc->slow_intr_mask); #ifdef __FreeBSD__ taskqueue_enqueue(sc->tq, &sc->ext_intr_task); #endif #ifdef __NetBSD__ workqueue_enqueue(sc->ext_intr_task.wq, &sc->ext_intr_task.w, NULL); #endif } ADAPTER_UNLOCK(sc); } void t3_os_set_hw_addr(adapter_t *adapter, int port_idx, u8 hw_addr[]) { /* * The ifnet might not be allocated before this gets called, * as this is called early on in attach by t3_prep_adapter * save the address off in the port structure */ if (cxgb_debug) #ifdef __FreeBSD__ printf("set_hw_addr on idx %d addr %6D\n", port_idx, hw_addr, ":"); #endif #ifdef __NetBSD__ printf("set_hw_addr on idx %d addr %02x:%02x:%02x:%02x:%02x:%02x\n", port_idx, hw_addr[0], hw_addr[1], hw_addr[2], hw_addr[3], hw_addr[4], hw_addr[5]); #endif bcopy(hw_addr, adapter->port[port_idx].hw_addr, ETHER_ADDR_LEN); } /** * link_start - enable a port * @p: the port to enable * * Performs the MAC and PHY actions needed to enable a port. */ static void cxgb_link_start(struct port_info *p) { struct ifnet *ifp; struct t3_rx_mode rm; struct cmac *mac = &p->mac; ifp = p->ifp; t3_init_rx_mode(&rm, p); if (!mac->multiport) t3_mac_reset(mac); t3_mac_set_mtu(mac, ifp->if_mtu + ETHER_HDR_LEN + ETHER_VLAN_ENCAP_LEN); t3_mac_set_address(mac, 0, p->hw_addr); t3_mac_set_rx_mode(mac, &rm); t3_link_start(&p->phy, mac, &p->link_config); t3_mac_enable(mac, MAC_DIRECTION_RX | MAC_DIRECTION_TX); } /** * setup_rss - configure Receive Side Steering (per-queue connection demux) * @adap: the adapter * * Sets up RSS to distribute packets to multiple receive queues. We * configure the RSS CPU lookup table to distribute to the number of HW * receive queues, and the response queue lookup table to narrow that * down to the response queues actually configured for each port. * We always configure the RSS mapping for two ports since the mapping * table has plenty of entries. */ static void setup_rss(adapter_t *adap) { int i; u_int nq[2]; uint8_t cpus[SGE_QSETS + 1]; uint16_t rspq_map[RSS_TABLE_SIZE]; for (i = 0; i < SGE_QSETS; ++i) cpus[i] = i; cpus[SGE_QSETS] = 0xff; nq[0] = nq[1] = 0; for_each_port(adap, i) { const struct port_info *pi = adap2pinfo(adap, i); nq[pi->tx_chan] += pi->nqsets; } nq[0] = max(nq[0], 1U); nq[1] = max(nq[1], 1U); for (i = 0; i < RSS_TABLE_SIZE / 2; ++i) { rspq_map[i] = i % nq[0]; rspq_map[i + RSS_TABLE_SIZE / 2] = (i % nq[1]) + nq[0]; } /* Calculate the reverse RSS map table */ for (i = 0; i < RSS_TABLE_SIZE; ++i) if (adap->rrss_map[rspq_map[i]] == 0xff) adap->rrss_map[rspq_map[i]] = i; t3_config_rss(adap, F_RQFEEDBACKENABLE | F_TNLLKPEN | F_TNLMAPEN | F_TNLPRTEN | F_TNL2TUPEN | F_TNL4TUPEN | F_OFDMAPEN | V_RRCPLCPUSIZE(6), cpus, rspq_map); } /* * Sends an mbuf to an offload queue driver * after dealing with any active network taps. */ static inline int offload_tx(struct toedev *tdev, struct mbuf *m) { int ret; critical_enter(); ret = t3_offload_tx(tdev, m); critical_exit(); return (ret); } #ifdef __FreeBSD__ static int write_smt_entry(struct adapter *adapter, int idx) { struct port_info *pi = &adapter->port[idx]; struct cpl_smt_write_req *req; struct mbuf *m; if ((m = m_gethdr(M_NOWAIT, MT_DATA)) == NULL) return (ENOMEM); req = mtod(m, struct cpl_smt_write_req *); req->wr.wr_hi = htonl(V_WR_OP(FW_WROPCODE_FORWARD)); OPCODE_TID(req) = htonl(MK_OPCODE_TID(CPL_SMT_WRITE_REQ, idx)); req->mtu_idx = NMTUS - 1; /* should be 0 but there's a T3 bug */ req->iff = idx; memset(req->src_mac1, 0, sizeof(req->src_mac1)); memcpy(req->src_mac0, pi->hw_addr, ETHER_ADDR_LEN); m_set_priority(m, 1); offload_tx(&adapter->tdev, m); return (0); } static int init_smt(struct adapter *adapter) { int i; for_each_port(adapter, i) write_smt_entry(adapter, i); return 0; } static void init_port_mtus(adapter_t *adapter) { unsigned int mtus = adapter->port[0].ifp->if_mtu; if (adapter->port[1].ifp) mtus |= adapter->port[1].ifp->if_mtu << 16; t3_write_reg(adapter, A_TP_MTU_PORT_TABLE, mtus); } #endif static void send_pktsched_cmd(struct adapter *adap, int sched, int qidx, int lo, int hi, int port) { struct mbuf *m; struct mngt_pktsched_wr *req; m = m_gethdr(M_DONTWAIT, MT_DATA); if (m) { req = mtod(m, struct mngt_pktsched_wr *); req->wr_hi = htonl(V_WR_OP(FW_WROPCODE_MNGT)); req->mngt_opcode = FW_MNGTOPCODE_PKTSCHED_SET; req->sched = sched; req->idx = qidx; req->min = lo; req->max = hi; req->binding = port; m->m_len = m->m_pkthdr.len = sizeof(*req); t3_mgmt_tx(adap, m); } } static void bind_qsets(adapter_t *sc) { int i, j; for (i = 0; i < (sc)->params.nports; ++i) { const struct port_info *pi = adap2pinfo(sc, i); for (j = 0; j < pi->nqsets; ++j) { send_pktsched_cmd(sc, 1, pi->first_qset + j, -1, -1, pi->tx_chan); } } } #ifdef __FreeBSD__ static void update_tpeeprom(struct adapter *adap) { #ifdef FIRMWARE_LATEST const struct firmware *tpeeprom; #else struct firmware *tpeeprom; #endif char buf[64]; uint32_t vers; unsigned int major, minor; int ret, len; char rev; t3_seeprom_read(adap, TP_SRAM_OFFSET, &vers); major = G_TP_VERSION_MAJOR(vers); minor = G_TP_VERSION_MINOR(vers); if (major == TP_VERSION_MAJOR && minor == TP_VERSION_MINOR) return; rev = t3rev2char(adap); snprintf(buf, sizeof(buf), TPEEPROM_NAME, rev, TP_VERSION_MAJOR, TP_VERSION_MINOR, TP_VERSION_MICRO); #ifdef __FreeBSD__ tpeeprom = firmware_get(buf); #endif #ifdef __NetBSD__ tpeeprom = NULL; #endif if (tpeeprom == NULL) { #ifdef __FreeBSD__ device_printf(adap->dev, "could not load TP EEPROM: unable to load %s\n", buf); #endif return; } len = tpeeprom->datasize - 4; ret = t3_check_tpsram(adap, tpeeprom->data, tpeeprom->datasize); if (ret) goto release_tpeeprom; if (len != TP_SRAM_LEN) { #ifdef __FreeBSD__ device_printf(adap->dev, "%s length is wrong len=%d expected=%d\n", buf, len, TP_SRAM_LEN); #endif return; } ret = set_eeprom(&adap->port[0], tpeeprom->data, tpeeprom->datasize, TP_SRAM_OFFSET); #ifdef __FreeBSD__ if (!ret) { device_printf(adap->dev, "Protocol SRAM image updated in EEPROM to %d.%d.%d\n", TP_VERSION_MAJOR, TP_VERSION_MINOR, TP_VERSION_MICRO); } else device_printf(adap->dev, "Protocol SRAM image update in EEPROM failed\n"); #endif #ifdef __NetBSD__ if (!ret) printf("Protocol SRAM image updated in EEPROM to %d.%d.%d\n", TP_VERSION_MAJOR, TP_VERSION_MINOR, TP_VERSION_MICRO); else printf("Protocol SRAM image update in EEPROM failed\n"); #endif release_tpeeprom: firmware_put(tpeeprom, FIRMWARE_UNLOAD); return; } static int update_tpsram(struct adapter *adap) { #ifdef FIRMWARE_LATEST const struct firmware *tpsram; #else struct firmware *tpsram; #endif char buf[64]; int ret; char rev; rev = t3rev2char(adap); if (!rev) return 0; update_tpeeprom(adap); snprintf(buf, sizeof(buf), TPSRAM_NAME, rev, TP_VERSION_MAJOR, TP_VERSION_MINOR, TP_VERSION_MICRO); tpsram = firmware_get(buf); if (tpsram == NULL){ #ifdef __FreeBSD__ device_printf(adap->dev, "could not load TP SRAM: unable to load %s\n", buf); #endif #ifdef __NetBSD__ printf("could not load TP SRAM: unable to load %s\n", buf); #endif return (EINVAL); } else #ifdef __FreeBSD__ device_printf(adap->dev, "updating TP SRAM with %s\n", buf); #endif #ifdef __NetBSD__ printf("updating TP SRAM with %s\n", buf); #endif ret = t3_check_tpsram(adap, tpsram->data, tpsram->datasize); if (ret) goto release_tpsram; ret = t3_set_proto_sram(adap, tpsram->data); if (ret) device_printf(adap->dev, "loading protocol SRAM failed\n"); release_tpsram: firmware_put(tpsram, FIRMWARE_UNLOAD); return ret; } #endif /** * cxgb_up - enable the adapter * @adap: adapter being enabled * * Called when the first port is enabled, this function performs the * actions necessary to make an adapter operational, such as completing * the initialization of HW modules, and enabling interrupts. * */ static int cxgb_up(struct adapter *sc) { int err = 0; if ((sc->flags & FULL_INIT_DONE) == 0) { #ifdef __FreeBSD__ if ((sc->flags & FW_UPTODATE) == 0) if ((err = upgrade_fw(sc))) goto out; if ((sc->flags & TPS_UPTODATE) == 0) if ((err = update_tpsram(sc))) goto out; #endif #ifdef __NetBSD__ if ((sc->flags & FW_UPTODATE) == 0) printf("SHOULD UPGRADE FIRMWARE!\n"); if ((sc->flags & TPS_UPTODATE) == 0) printf("SHOULD UPDATE TPSRAM\n"); #endif err = t3_init_hw(sc, 0); if (err) goto out; t3_write_reg(sc, A_ULPRX_TDDP_PSZ, V_HPZ0(PAGE_SHIFT - 12)); err = setup_sge_qsets(sc); if (err) goto out; setup_rss(sc); sc->flags |= FULL_INIT_DONE; } t3_intr_clear(sc); #ifdef __FreeBSD__ /* If it's MSI or INTx, allocate a single interrupt for everything */ if ((sc->flags & USING_MSIX) == 0) { if ((sc->irq_res = bus_alloc_resource_any(sc->dev, SYS_RES_IRQ, &sc->irq_rid, RF_SHAREABLE | RF_ACTIVE)) == NULL) { device_printf(sc->dev, "Cannot allocate interrupt rid=%d\n", sc->irq_rid); err = EINVAL; goto out; } device_printf(sc->dev, "allocated irq_res=%p\n", sc->irq_res); if (bus_setup_intr(sc->dev, sc->irq_res, INTR_MPSAFE|INTR_TYPE_NET, #ifdef INTR_FILTERS NULL, #endif sc->cxgb_intr, sc, &sc->intr_tag)) { device_printf(sc->dev, "Cannot set up interrupt\n"); err = EINVAL; goto irq_err; } } else { cxgb_setup_msix(sc, sc->msi_count); } #endif #ifdef __NetBSD__ /* If it's MSI or INTx, allocate a single interrupt for everything */ if ((sc->flags & USING_MSIX) == 0) { if (pci_intr_map(&sc->pa, &sc->intr_handle)) { device_printf(sc->dev, "Cannot allocate interrupt\n"); err = EINVAL; goto out; } device_printf(sc->dev, "allocated intr_handle=%p\n", sc->intr_handle); sc->intr_cookie = pci_intr_establish(sc->pa.pa_pc, sc->intr_handle, IPL_NET, sc->cxgb_intr, sc); if (sc->intr_cookie == NULL) { device_printf(sc->dev, "Cannot establish interrupt\n"); err = EINVAL; goto irq_err; } } else { printf("Using MSIX?!?!?!\n"); INT3; cxgb_setup_msix(sc, sc->msi_count); } #endif t3_sge_start(sc); t3_intr_enable(sc); if (!(sc->flags & QUEUES_BOUND)) { bind_qsets(sc); sc->flags |= QUEUES_BOUND; } out: return (err); irq_err: CH_ERR(sc, "request_irq failed, err %d\n", err); goto out; } /* * Release resources when all the ports and offloading have been stopped. */ static void cxgb_down_locked(struct adapter *sc) { #ifdef __FreeBSD__ int i; #endif t3_sge_stop(sc); t3_intr_disable(sc); #ifdef __FreeBSD__ if (sc->intr_tag != NULL) { bus_teardown_intr(sc->dev, sc->irq_res, sc->intr_tag); sc->intr_tag = NULL; } if (sc->irq_res != NULL) { device_printf(sc->dev, "de-allocating interrupt irq_rid=%d irq_res=%p\n", sc->irq_rid, sc->irq_res); bus_release_resource(sc->dev, SYS_RES_IRQ, sc->irq_rid, sc->irq_res); sc->irq_res = NULL; } #endif #ifdef __NetBSD__ INT3; // XXXXXXXXXXXXXXXXXX #endif if (sc->flags & USING_MSIX) cxgb_teardown_msix(sc); ADAPTER_UNLOCK(sc); callout_drain(&sc->cxgb_tick_ch); callout_drain(&sc->sge_timer_ch); #ifdef __FreeBSD__ if (sc->tq != NULL) { taskqueue_drain(sc->tq, &sc->slow_intr_task); for (i = 0; i < sc->params.nports; i++) taskqueue_drain(sc->tq, &sc->port[i].timer_reclaim_task); } #endif #ifdef notyet if (sc->port[i].tq != NULL) #endif } #ifdef __FreeBSD__ static int offload_open(struct port_info *pi) { struct adapter *adapter = pi->adapter; struct toedev *tdev = TOEDEV(pi->ifp); int adap_up = adapter->open_device_map & PORT_MASK; int err = 0; if (atomic_cmpset_int(&adapter->open_device_map, (adapter->open_device_map & ~OFFLOAD_DEVMAP_BIT), (adapter->open_device_map | OFFLOAD_DEVMAP_BIT)) == 0) return (0); ADAPTER_LOCK(pi->adapter); if (!adap_up) err = cxgb_up(adapter); ADAPTER_UNLOCK(pi->adapter); if (err) return (err); t3_tp_set_offload_mode(adapter, 1); tdev->lldev = adapter->port[0].ifp; err = cxgb_offload_activate(adapter); if (err) goto out; init_port_mtus(adapter); t3_load_mtus(adapter, adapter->params.mtus, adapter->params.a_wnd, adapter->params.b_wnd, adapter->params.rev == 0 ? adapter->port[0].ifp->if_mtu : 0xffff); init_smt(adapter); /* Call back all registered clients */ cxgb_add_clients(tdev); out: /* restore them in case the offload module has changed them */ if (err) { t3_tp_set_offload_mode(adapter, 0); clrbit(&adapter->open_device_map, OFFLOAD_DEVMAP_BIT); cxgb_set_dummy_ops(tdev); } return (err); } #ifdef notyet static int offload_close(struct toedev *tdev) { struct adapter *adapter = tdev2adap(tdev); if (!isset(&adapter->open_device_map, OFFLOAD_DEVMAP_BIT)) return (0); /* Call back all registered clients */ cxgb_remove_clients(tdev); tdev->lldev = NULL; cxgb_set_dummy_ops(tdev); t3_tp_set_offload_mode(adapter, 0); clrbit(&adapter->open_device_map, OFFLOAD_DEVMAP_BIT); if (!adapter->open_device_map) cxgb_down(adapter); cxgb_offload_deactivate(adapter); return (0); } #endif #endif #ifdef __FreeBSD__ static void cxgb_init(void *arg) #endif #ifdef __NetBSD__ static int cxgb_init(struct ifnet *ifp) #endif { #ifdef __FreeBSD__ struct port_info *p = arg; #endif #ifdef __NetBSD__ struct port_info *p = ifp->if_softc; #endif PORT_LOCK(p); cxgb_init_locked(p); PORT_UNLOCK(p); #ifdef __NetBSD__ return (0); // ???????????? #endif } static void cxgb_init_locked(struct port_info *p) { struct ifnet *ifp; adapter_t *sc = p->adapter; int err; PORT_LOCK_ASSERT_OWNED(p); ifp = p->ifp; ADAPTER_LOCK(p->adapter); if ((sc->open_device_map == 0) && (err = cxgb_up(sc))) { ADAPTER_UNLOCK(p->adapter); cxgb_stop_locked(p); return; } if (p->adapter->open_device_map == 0) { t3_intr_clear(sc); t3_sge_init_adapter(sc); } setbit(&p->adapter->open_device_map, p->port_id); ADAPTER_UNLOCK(p->adapter); #ifdef __FreeBSD__ if (is_offload(sc) && !ofld_disable) { err = offload_open(p); if (err) log(LOG_WARNING, "Could not initialize offload capabilities\n"); } #endif cxgb_link_start(p); t3_link_changed(sc, p->port_id); ifp->if_baudrate = p->link_config.speed * 1000000; device_printf(sc->dev, "enabling interrupts on port=%d\n", p->port_id); t3_port_intr_enable(sc, p->port_id); callout_reset(&sc->cxgb_tick_ch, sc->params.stats_update_period * hz, cxgb_tick, sc); ifp->if_drv_flags |= IFF_DRV_RUNNING; ifp->if_drv_flags &= ~IFF_DRV_OACTIVE; } static void cxgb_set_rxmode(struct port_info *p) { struct t3_rx_mode rm; struct cmac *mac = &p->mac; PORT_LOCK_ASSERT_OWNED(p); t3_init_rx_mode(&rm, p); t3_mac_set_rx_mode(mac, &rm); } static void cxgb_stop_locked(struct port_info *p) { struct ifnet *ifp; PORT_LOCK_ASSERT_OWNED(p); ADAPTER_LOCK_ASSERT_NOTOWNED(p->adapter); ifp = p->ifp; t3_port_intr_disable(p->adapter, p->port_id); ifp->if_drv_flags &= ~(IFF_DRV_RUNNING | IFF_DRV_OACTIVE); p->phy.ops->power_down(&p->phy, 1); t3_mac_disable(&p->mac, MAC_DIRECTION_TX | MAC_DIRECTION_RX); ADAPTER_LOCK(p->adapter); clrbit(&p->adapter->open_device_map, p->port_id); if (p->adapter->open_device_map == 0) { cxgb_down_locked(p->adapter); } else ADAPTER_UNLOCK(p->adapter); } static int cxgb_set_mtu(struct port_info *p, int mtu) { struct ifnet *ifp = p->ifp; struct ifreq ifr; int error = 0; ifr.ifr_mtu = mtu; if ((mtu < ETHERMIN) || (mtu > ETHER_MAX_LEN_JUMBO)) error = EINVAL; else if ((error = ifioctl_common(ifp, SIOCSIFMTU, &ifr)) == ENETRESET) { error = 0; PORT_LOCK(p); if (ifp->if_drv_flags & IFF_DRV_RUNNING) { callout_stop(&p->adapter->cxgb_tick_ch); cxgb_stop_locked(p); cxgb_init_locked(p); } PORT_UNLOCK(p); } return (error); } static int #ifdef __FreeBSD__ cxgb_ioctl(struct ifnet *ifp, unsigned long command, caddr_t data) #endif #ifdef __NetBSD__ cxgb_ioctl(struct ifnet *ifp, unsigned long command, void *data) #endif { struct port_info *p = ifp->if_softc; struct ifaddr *ifa = (struct ifaddr *)data; struct ifreq *ifr = (struct ifreq *)data; int flags, error = 0; uint32_t mask; /* * XXX need to check that we aren't in the middle of an unload */ printf("cxgb_ioctl(%d): command=%08lx\n", __LINE__, command); switch (command) { case SIOCSIFMTU: error = cxgb_set_mtu(p, ifr->ifr_mtu); printf("SIOCSIFMTU: error=%d\n", error); break; case SIOCSIFADDR: printf("SIOCSIFADDR:\n"); case SIOCGIFADDR: printf("SIOCGIFADDR:\n"); PORT_LOCK(p); if (ifa->ifa_addr->sa_family == AF_INET) { ifp->if_flags |= IFF_UP; if (!(ifp->if_drv_flags & IFF_DRV_RUNNING)) cxgb_init_locked(p); arp_ifinit(ifp, ifa); } else error = ether_ioctl(ifp, command, data); PORT_UNLOCK(p); break; case SIOCSIFFLAGS: printf("SIOCSIFFLAGS:\n"); callout_drain(&p->adapter->cxgb_tick_ch); PORT_LOCK(p); if (ifp->if_flags & IFF_UP) { if (ifp->if_drv_flags & IFF_DRV_RUNNING) { flags = p->if_flags; if (((ifp->if_flags ^ flags) & IFF_PROMISC) || ((ifp->if_flags ^ flags) & IFF_ALLMULTI)) cxgb_set_rxmode(p); } else cxgb_init_locked(p); p->if_flags = ifp->if_flags; } else if (ifp->if_drv_flags & IFF_DRV_RUNNING) cxgb_stop_locked(p); if (ifp->if_drv_flags & IFF_DRV_RUNNING) { adapter_t *sc = p->adapter; callout_reset(&sc->cxgb_tick_ch, sc->params.stats_update_period * hz, cxgb_tick, sc); } PORT_UNLOCK(p); break; case SIOCSIFMEDIA: printf("SIOCSIFMEDIA:\n"); case SIOCGIFMEDIA: error = ifmedia_ioctl(ifp, ifr, &p->media, command); printf("SIOCGIFMEDIA: error=%d\n", error); break; #ifdef __FreeBSD__ case SIOCSIFCAP: printf("SIOCSIFCAP:\n"); PORT_LOCK(p); mask = ifr->ifr_reqcap ^ ifp->if_capenable; if (mask & IFCAP_TXCSUM) { if (IFCAP_TXCSUM & ifp->if_capenable) { ifp->if_capenable &= ~(IFCAP_TXCSUM|IFCAP_TSO4); ifp->if_hwassist &= ~(CSUM_TCP | CSUM_UDP | CSUM_TSO); } else { ifp->if_capenable |= IFCAP_TXCSUM; ifp->if_hwassist |= (CSUM_TCP | CSUM_UDP); } } else if (mask & IFCAP_RXCSUM) { if (IFCAP_RXCSUM & ifp->if_capenable) { ifp->if_capenable &= ~IFCAP_RXCSUM; } else { ifp->if_capenable |= IFCAP_RXCSUM; } } if (mask & IFCAP_TSO4) { if (IFCAP_TSO4 & ifp->if_capenable) { ifp->if_capenable &= ~IFCAP_TSO4; ifp->if_hwassist &= ~CSUM_TSO; } else if (IFCAP_TXCSUM & ifp->if_capenable) { ifp->if_capenable |= IFCAP_TSO4; ifp->if_hwassist |= CSUM_TSO; } else { if (cxgb_debug) printf("cxgb requires tx checksum offload" " be enabled to use TSO\n"); error = EINVAL; } } PORT_UNLOCK(p); break; #endif /* __FreeBSD__ */ default: printf("Dir = %x Len = %x Group = '%c' Num = %x\n", (unsigned int)(command&0xe0000000)>>28, (unsigned int)(command&0x1fff0000)>>16, (unsigned int)(command&0xff00)>>8, (unsigned int)command&0xff); if ((error = ether_ioctl(ifp, command, data)) != ENETRESET) break; error = 0; break; } return (error); } static int cxgb_start_tx(struct ifnet *ifp, uint32_t txmax) { struct sge_qset *qs; struct sge_txq *txq; struct port_info *p = ifp->if_softc; struct mbuf *m = NULL; int err, in_use_init, free_it; if (!p->link_config.link_ok) { return (ENXIO); } #ifdef __FreeBSD__ if (IFQ_DRV_IS_EMPTY(&ifp->if_snd)) #endif #ifdef __NetBSD__ if (IFQ_IS_EMPTY(&ifp->if_snd)) #endif { return (ENOBUFS); } qs = &p->adapter->sge.qs[p->first_qset]; txq = &qs->txq[TXQ_ETH]; err = 0; if (txq->flags & TXQ_TRANSMITTING) { return (EINPROGRESS); } mtx_lock(&txq->lock); txq->flags |= TXQ_TRANSMITTING; in_use_init = txq->in_use; while ((txq->in_use - in_use_init < txmax) && (txq->size > txq->in_use + TX_MAX_DESC)) { free_it = 0; #ifdef __FreeBSD__ IFQ_DRV_DEQUEUE(&ifp->if_snd, m); #endif #ifdef __NetBSD__ IFQ_DEQUEUE(&ifp->if_snd, m); #endif if (m == NULL) break; /* * Convert chain to M_IOVEC */ #ifdef __FreeBSD__ KASSERT((m->m_flags & M_IOVEC) == 0, ("IOVEC set too early")); #endif #ifdef __NetBSD__ KASSERT((m->m_flags & M_IOVEC) == 0); #endif #ifdef notyet m0 = m; if (collapse_mbufs && m->m_pkthdr.len > MCLBYTES && m_collapse(m, TX_MAX_SEGS, &m0) == EFBIG) { if ((m0 = m_defrag(m, M_NOWAIT)) != NULL) { m = m0; m_collapse(m, TX_MAX_SEGS, &m0); } else break; } m = m0; #endif if ((err = t3_encap(p, &m, &free_it)) != 0) { printf("t3_encap() returned %d\n", err); break; } #ifdef __FreeBSD__ BPF_MTAP(ifp, m); #endif #ifdef __NetBSD__ // bpf_mtap(ifp, m); #endif if (free_it) { m_freem(m); } } txq->flags &= ~TXQ_TRANSMITTING; mtx_unlock(&txq->lock); if (__predict_false(err)) { if (err == ENOMEM) { ifp->if_drv_flags |= IFF_DRV_OACTIVE; #ifdef __FreeBSD__ IFQ_LOCK(&ifp->if_snd); IFQ_DRV_PREPEND(&ifp->if_snd, m); IFQ_UNLOCK(&ifp->if_snd); #endif #ifdef __NetBSD__ // XXXXXXXXXX lock/unlock?? IF_PREPEND(&ifp->if_snd, m); #endif } } if (err == 0 && m == NULL) err = ENOBUFS; else if ((err == 0) && (txq->size <= txq->in_use + TX_MAX_DESC) && (ifp->if_drv_flags & IFF_DRV_OACTIVE) == 0) { ifp->if_drv_flags |= IFF_DRV_OACTIVE; err = ENOSPC; } return (err); } static void #ifdef __FreeBSD__ cxgb_start_proc(void *arg, int ncount) #endif #ifdef __NetBSD__ cxgb_start_proc(struct work *wk, void *arg) #endif { struct ifnet *ifp = arg; struct port_info *pi = ifp->if_softc; struct sge_qset *qs; struct sge_txq *txq; int error; qs = &pi->adapter->sge.qs[pi->first_qset]; txq = &qs->txq[TXQ_ETH]; do { if (desc_reclaimable(txq) > TX_CLEAN_MAX_DESC >> 2) #ifdef __FreeBSD__ taskqueue_enqueue(pi->tq, &txq->qreclaim_task); #endif #ifdef __NetBSD__ workqueue_enqueue(pi->timer_reclaim_task.wq, &pi->timer_reclaim_task.w, NULL); #endif error = cxgb_start_tx(ifp, TX_START_MAX_DESC); } while (error == 0); } static void cxgb_start(struct ifnet *ifp) { struct port_info *pi = ifp->if_softc; struct sge_qset *qs; struct sge_txq *txq; int err; qs = &pi->adapter->sge.qs[pi->first_qset]; txq = &qs->txq[TXQ_ETH]; if (desc_reclaimable(txq) > TX_CLEAN_MAX_DESC >> 2) #ifdef __FreeBSD__ taskqueue_enqueue(pi->tq, &txq->qreclaim_task); #endif #ifdef __NetBSD__ workqueue_enqueue(pi->timer_reclaim_task.wq, &pi->timer_reclaim_task.w, NULL); #endif err = cxgb_start_tx(ifp, TX_START_MAX_DESC); if (err == 0) #ifdef __FreeBSD__ taskqueue_enqueue(pi->tq, &pi->start_task); #endif #ifdef __NetBSD__ workqueue_enqueue(pi->start_task.wq, &pi->start_task.w, NULL); #endif } #ifdef __NetBSD__ static void cxgb_stop(struct ifnet *ifp, int reason) { struct port_info *pi = ifp->if_softc; printf("cxgb_stop(): pi=%p, reason=%d\n", pi, reason); INT3; } #endif static int cxgb_media_change(struct ifnet *ifp) { #ifdef __FreeBSD__ if_printf(ifp, "media change not supported\n"); #endif #ifdef __NetBSD__ printf("media change not supported: ifp=%p\n", ifp); #endif return (ENXIO); } static void cxgb_media_status(struct ifnet *ifp, struct ifmediareq *ifmr) { struct port_info *p; p = ifp->if_softc; ifmr->ifm_status = IFM_AVALID; ifmr->ifm_active = IFM_ETHER; if (!p->link_config.link_ok) return; ifmr->ifm_status |= IFM_ACTIVE; switch (p->link_config.speed) { case 10: ifmr->ifm_active |= IFM_10_T; break; case 100: ifmr->ifm_active |= IFM_100_TX; break; case 1000: ifmr->ifm_active |= IFM_1000_T; break; } if (p->link_config.duplex) ifmr->ifm_active |= IFM_FDX; else ifmr->ifm_active |= IFM_HDX; } #ifdef __FreeBSD__ static void #endif #ifdef __NetBSD__ static int #endif cxgb_async_intr(void *data) { adapter_t *sc = data; if (cxgb_debug) device_printf(sc->dev, "cxgb_async_intr\n"); #ifdef __FreeBSD__ /* * May need to sleep - defer to taskqueue */ taskqueue_enqueue(sc->tq, &sc->slow_intr_task); #endif #ifdef __NetBSD__ /* * May need to sleep - defer to taskqueue */ workqueue_enqueue(sc->slow_intr_task.wq, &sc->slow_intr_task.w, NULL); return (1); #endif } static void #ifdef __FreeBSD__ cxgb_ext_intr_handler(void *arg, int count) #endif #ifdef __NetBSD__ cxgb_ext_intr_handler(struct work *wk, void *arg) #endif { adapter_t *sc = (adapter_t *)arg; if (cxgb_debug) printf("cxgb_ext_intr_handler\n"); t3_phy_intr_handler(sc); /* Now reenable external interrupts */ ADAPTER_LOCK(sc); if (sc->slow_intr_mask) { sc->slow_intr_mask |= F_T3DBG; t3_write_reg(sc, A_PL_INT_CAUSE0, F_T3DBG); t3_write_reg(sc, A_PL_INT_ENABLE0, sc->slow_intr_mask); } ADAPTER_UNLOCK(sc); } static void check_link_status(adapter_t *sc) { int i; for (i = 0; i < (sc)->params.nports; ++i) { struct port_info *p = &sc->port[i]; if (!(p->port_type->caps & SUPPORTED_IRQ)) t3_link_changed(sc, i); p->ifp->if_baudrate = p->link_config.speed * 1000000; } } static void check_t3b2_mac(struct adapter *adapter) { int i; for_each_port(adapter, i) { struct port_info *p = &adapter->port[i]; struct ifnet *ifp = p->ifp; int status; if ((ifp->if_drv_flags & IFF_DRV_RUNNING) == 0) continue; status = 0; PORT_LOCK(p); if ((ifp->if_drv_flags & IFF_DRV_RUNNING)) status = t3b2_mac_watchdog_task(&p->mac); if (status == 1) p->mac.stats.num_toggled++; else if (status == 2) { struct cmac *mac = &p->mac; t3_mac_set_mtu(mac, ifp->if_mtu + ETHER_HDR_LEN + ETHER_VLAN_ENCAP_LEN); t3_mac_set_address(mac, 0, p->hw_addr); cxgb_set_rxmode(p); t3_link_start(&p->phy, mac, &p->link_config); t3_mac_enable(mac, MAC_DIRECTION_RX | MAC_DIRECTION_TX); t3_port_intr_enable(adapter, p->port_id); p->mac.stats.num_resets++; } PORT_UNLOCK(p); } } static void cxgb_tick(void *arg) { adapter_t *sc = (adapter_t *)arg; #ifdef __FreeBSD__ taskqueue_enqueue(sc->tq, &sc->tick_task); #endif #ifdef __NetBSD__ workqueue_enqueue(sc->tick_task.wq, &sc->tick_task.w, NULL); #endif if (sc->open_device_map != 0) callout_reset(&sc->cxgb_tick_ch, sc->params.stats_update_period * hz, cxgb_tick, sc); } static void #ifdef __FreeBSD__ cxgb_tick_handler(void *arg, int count) #endif #ifdef __NetBSD__ cxgb_tick_handler(struct work *wk, void *arg) #endif { adapter_t *sc = (adapter_t *)arg; const struct adapter_params *p = &sc->params; ADAPTER_LOCK(sc); if (p->linkpoll_period) check_link_status(sc); /* * adapter lock can currently only be acquire after the * port lock */ ADAPTER_UNLOCK(sc); if (p->rev == T3_REV_B2 && p->nports < 4) check_t3b2_mac(sc); } static void touch_bars(device_t dev) { /* * Don't enable yet */ #if !defined(__LP64__) && 0 u32 v; pci_read_config_dword(pdev, PCI_BASE_ADDRESS_1, &v); pci_write_config_dword(pdev, PCI_BASE_ADDRESS_1, v); pci_read_config_dword(pdev, PCI_BASE_ADDRESS_3, &v); pci_write_config_dword(pdev, PCI_BASE_ADDRESS_3, v); pci_read_config_dword(pdev, PCI_BASE_ADDRESS_5, &v); pci_write_config_dword(pdev, PCI_BASE_ADDRESS_5, v); #endif } #ifdef __FreeBSD__ static int set_eeprom(struct port_info *pi, const uint8_t *data, int len, int offset) { uint8_t *buf; int err = 0; u32 aligned_offset, aligned_len, *p; struct adapter *adapter = pi->adapter; aligned_offset = offset & ~3; aligned_len = (len + (offset & 3) + 3) & ~3; if (aligned_offset != offset || aligned_len != len) { buf = malloc(aligned_len, M_DEVBUF, M_WAITOK|M_ZERO); if (!buf) return (ENOMEM); err = t3_seeprom_read(adapter, aligned_offset, (u32 *)buf); if (!err && aligned_len > 4) err = t3_seeprom_read(adapter, aligned_offset + aligned_len - 4, (u32 *)&buf[aligned_len - 4]); if (err) goto out; memcpy(buf + (offset & 3), data, len); } else buf = (uint8_t *)(uintptr_t)data; err = t3_seeprom_wp(adapter, 0); if (err) goto out; for (p = (u32 *)buf; !err && aligned_len; aligned_len -= 4, p++) { err = t3_seeprom_write(adapter, aligned_offset, *p); aligned_offset += 4; } if (!err) err = t3_seeprom_wp(adapter, 1); out: if (buf != data) free(buf, M_DEVBUF); return err; } #endif #ifdef __FreeBSD__ static int in_range(int val, int lo, int hi) { return val < 0 || (val <= hi && val >= lo); } static int cxgb_extension_open(struct cdev *dev, int flags, int fmp, d_thread_t *td) { return (0); } static int cxgb_extension_close(struct cdev *dev, int flags, int fmt, d_thread_t *td) { return (0); } static int cxgb_extension_ioctl(struct cdev *dev, unsigned long cmd, caddr_t data, int fflag, struct thread *td) { int mmd, error = 0; struct port_info *pi = dev->si_drv1; adapter_t *sc = pi->adapter; #ifdef PRIV_SUPPORTED if (priv_check(td, PRIV_DRIVER)) { if (cxgb_debug) printf("user does not have access to privileged ioctls\n"); return (EPERM); } #else if (suser(td)) { if (cxgb_debug) printf("user does not have access to privileged ioctls\n"); return (EPERM); } #endif switch (cmd) { case SIOCGMIIREG: { uint32_t val; struct cphy *phy = &pi->phy; struct mii_data *mid = (struct mii_data *)data; if (!phy->mdio_read) return (EOPNOTSUPP); if (is_10G(sc)) { mmd = mid->phy_id >> 8; if (!mmd) mmd = MDIO_DEV_PCS; else if (mmd > MDIO_DEV_XGXS) return (EINVAL); error = phy->mdio_read(sc, mid->phy_id & 0x1f, mmd, mid->reg_num, &val); } else error = phy->mdio_read(sc, mid->phy_id & 0x1f, 0, mid->reg_num & 0x1f, &val); if (error == 0) mid->val_out = val; break; } case SIOCSMIIREG: { struct cphy *phy = &pi->phy; struct mii_data *mid = (struct mii_data *)data; if (!phy->mdio_write) return (EOPNOTSUPP); if (is_10G(sc)) { mmd = mid->phy_id >> 8; if (!mmd) mmd = MDIO_DEV_PCS; else if (mmd > MDIO_DEV_XGXS) return (EINVAL); error = phy->mdio_write(sc, mid->phy_id & 0x1f, mmd, mid->reg_num, mid->val_in); } else error = phy->mdio_write(sc, mid->phy_id & 0x1f, 0, mid->reg_num & 0x1f, mid->val_in); break; } case CHELSIO_SETREG: { struct ch_reg *edata = (struct ch_reg *)data; if ((edata->addr & 0x3) != 0 || edata->addr >= sc->mmio_len) return (EFAULT); t3_write_reg(sc, edata->addr, edata->val); break; } case CHELSIO_GETREG: { struct ch_reg *edata = (struct ch_reg *)data; if ((edata->addr & 0x3) != 0 || edata->addr >= sc->mmio_len) return (EFAULT); edata->val = t3_read_reg(sc, edata->addr); break; } case CHELSIO_GET_SGE_CONTEXT: { struct ch_cntxt *ecntxt = (struct ch_cntxt *)data; mtx_lock(&sc->sge.reg_lock); switch (ecntxt->cntxt_type) { case CNTXT_TYPE_EGRESS: error = t3_sge_read_ecntxt(sc, ecntxt->cntxt_id, ecntxt->data); break; case CNTXT_TYPE_FL: error = t3_sge_read_fl(sc, ecntxt->cntxt_id, ecntxt->data); break; case CNTXT_TYPE_RSP: error = t3_sge_read_rspq(sc, ecntxt->cntxt_id, ecntxt->data); break; case CNTXT_TYPE_CQ: error = t3_sge_read_cq(sc, ecntxt->cntxt_id, ecntxt->data); break; default: error = EINVAL; break; } mtx_unlock(&sc->sge.reg_lock); break; } case CHELSIO_GET_SGE_DESC: { struct ch_desc *edesc = (struct ch_desc *)data; int ret; if (edesc->queue_num >= SGE_QSETS * 6) return (EINVAL); ret = t3_get_desc(&sc->sge.qs[edesc->queue_num / 6], edesc->queue_num % 6, edesc->idx, edesc->data); if (ret < 0) return (EINVAL); edesc->size = ret; break; } case CHELSIO_SET_QSET_PARAMS: { struct qset_params *q; struct ch_qset_params *t = (struct ch_qset_params *)data; if (t->qset_idx >= SGE_QSETS) return (EINVAL); if (!in_range(t->intr_lat, 0, M_NEWTIMER) || !in_range(t->cong_thres, 0, 255) || !in_range(t->txq_size[0], MIN_TXQ_ENTRIES, MAX_TXQ_ENTRIES) || !in_range(t->txq_size[1], MIN_TXQ_ENTRIES, MAX_TXQ_ENTRIES) || !in_range(t->txq_size[2], MIN_CTRL_TXQ_ENTRIES, MAX_CTRL_TXQ_ENTRIES) || !in_range(t->fl_size[0], MIN_FL_ENTRIES, MAX_RX_BUFFERS) || !in_range(t->fl_size[1], MIN_FL_ENTRIES, MAX_RX_JUMBO_BUFFERS) || !in_range(t->rspq_size, MIN_RSPQ_ENTRIES, MAX_RSPQ_ENTRIES)) return (EINVAL); if ((sc->flags & FULL_INIT_DONE) && (t->rspq_size >= 0 || t->fl_size[0] >= 0 || t->fl_size[1] >= 0 || t->txq_size[0] >= 0 || t->txq_size[1] >= 0 || t->txq_size[2] >= 0 || t->polling >= 0 || t->cong_thres >= 0)) return (EBUSY); q = &sc->params.sge.qset[t->qset_idx]; if (t->rspq_size >= 0) q->rspq_size = t->rspq_size; if (t->fl_size[0] >= 0) q->fl_size = t->fl_size[0]; if (t->fl_size[1] >= 0) q->jumbo_size = t->fl_size[1]; if (t->txq_size[0] >= 0) q->txq_size[0] = t->txq_size[0]; if (t->txq_size[1] >= 0) q->txq_size[1] = t->txq_size[1]; if (t->txq_size[2] >= 0) q->txq_size[2] = t->txq_size[2]; if (t->cong_thres >= 0) q->cong_thres = t->cong_thres; if (t->intr_lat >= 0) { struct sge_qset *qs = &sc->sge.qs[t->qset_idx]; q->coalesce_nsecs = t->intr_lat*1000; t3_update_qset_coalesce(qs, q); } break; } case CHELSIO_GET_QSET_PARAMS: { struct qset_params *q; struct ch_qset_params *t = (struct ch_qset_params *)data; if (t->qset_idx >= SGE_QSETS) return (EINVAL); q = &(sc)->params.sge.qset[t->qset_idx]; t->rspq_size = q->rspq_size; t->txq_size[0] = q->txq_size[0]; t->txq_size[1] = q->txq_size[1]; t->txq_size[2] = q->txq_size[2]; t->fl_size[0] = q->fl_size; t->fl_size[1] = q->jumbo_size; t->polling = q->polling; t->intr_lat = q->coalesce_nsecs / 1000; t->cong_thres = q->cong_thres; break; } case CHELSIO_SET_QSET_NUM: { struct ch_reg *edata = (struct ch_reg *)data; unsigned int port_idx = pi->port_id; if (sc->flags & FULL_INIT_DONE) return (EBUSY); if (edata->val < 1 || (edata->val > 1 && !(sc->flags & USING_MSIX))) return (EINVAL); if (edata->val + sc->port[!port_idx].nqsets > SGE_QSETS) return (EINVAL); sc->port[port_idx].nqsets = edata->val; sc->port[0].first_qset = 0; /* * XXX hardcode ourselves to 2 ports just like LEEENUX */ sc->port[1].first_qset = sc->port[0].nqsets; break; } case CHELSIO_GET_QSET_NUM: { struct ch_reg *edata = (struct ch_reg *)data; edata->val = pi->nqsets; break; } #ifdef notyet case CHELSIO_LOAD_FW: case CHELSIO_GET_PM: case CHELSIO_SET_PM: return (EOPNOTSUPP); break; #endif case CHELSIO_SETMTUTAB: { struct ch_mtus *m = (struct ch_mtus *)data; int i; if (!is_offload(sc)) return (EOPNOTSUPP); if (offload_running(sc)) return (EBUSY); if (m->nmtus != NMTUS) return (EINVAL); if (m->mtus[0] < 81) /* accommodate SACK */ return (EINVAL); /* * MTUs must be in ascending order */ for (i = 1; i < NMTUS; ++i) if (m->mtus[i] < m->mtus[i - 1]) return (EINVAL); memcpy(sc->params.mtus, m->mtus, sizeof(sc->params.mtus)); break; } case CHELSIO_GETMTUTAB: { struct ch_mtus *m = (struct ch_mtus *)data; if (!is_offload(sc)) return (EOPNOTSUPP); memcpy(m->mtus, sc->params.mtus, sizeof(m->mtus)); m->nmtus = NMTUS; break; } case CHELSIO_DEVUP: if (!is_offload(sc)) return (EOPNOTSUPP); return offload_open(pi); break; case CHELSIO_GET_MEM: { struct ch_mem_range *t = (struct ch_mem_range *)data; struct mc7 *mem; uint8_t *useraddr; u64 buf[32]; if (!is_offload(sc)) return (EOPNOTSUPP); if (!(sc->flags & FULL_INIT_DONE)) return (EIO); /* need the memory controllers */ if ((t->addr & 0x7) || (t->len & 0x7)) return (EINVAL); if (t->mem_id == MEM_CM) mem = &sc->cm; else if (t->mem_id == MEM_PMRX) mem = &sc->pmrx; else if (t->mem_id == MEM_PMTX) mem = &sc->pmtx; else return (EINVAL); /* * Version scheme: * bits 0..9: chip version * bits 10..15: chip revision */ t->version = 3 | (sc->params.rev << 10); /* * Read 256 bytes at a time as len can be large and we don't * want to use huge intermediate buffers. */ useraddr = (uint8_t *)(t + 1); /* advance to start of buffer */ while (t->len) { unsigned int chunk = min(t->len, sizeof(buf)); error = t3_mc7_bd_read(mem, t->addr / 8, chunk / 8, buf); if (error) return (-error); if (copyout(buf, useraddr, chunk)) return (EFAULT); useraddr += chunk; t->addr += chunk; t->len -= chunk; } break; } case CHELSIO_READ_TCAM_WORD: { struct ch_tcam_word *t = (struct ch_tcam_word *)data; if (!is_offload(sc)) return (EOPNOTSUPP); if (!(sc->flags & FULL_INIT_DONE)) return (EIO); /* need MC5 */ return -t3_read_mc5_range(&sc->mc5, t->addr, 1, t->buf); break; } case CHELSIO_SET_TRACE_FILTER: { struct ch_trace *t = (struct ch_trace *)data; const struct trace_params *tp; tp = (const struct trace_params *)&t->sip; if (t->config_tx) t3_config_trace_filter(sc, tp, 0, t->invert_match, t->trace_tx); if (t->config_rx) t3_config_trace_filter(sc, tp, 1, t->invert_match, t->trace_rx); break; } case CHELSIO_SET_PKTSCHED: { struct ch_pktsched_params *p = (struct ch_pktsched_params *)data; if (sc->open_device_map == 0) return (EAGAIN); send_pktsched_cmd(sc, p->sched, p->idx, p->min, p->max, p->binding); break; } case CHELSIO_IFCONF_GETREGS: { struct ifconf_regs *regs = (struct ifconf_regs *)data; int reglen = cxgb_get_regs_len(); uint8_t *buf = malloc(REGDUMP_SIZE, M_DEVBUF, M_NOWAIT); if (buf == NULL) { return (ENOMEM); } if (regs->len > reglen) regs->len = reglen; else if (regs->len < reglen) { error = E2BIG; goto done; } cxgb_get_regs(sc, regs, buf); error = copyout(buf, regs->data, reglen); done: free(buf, M_DEVBUF); break; } case CHELSIO_SET_HW_SCHED: { struct ch_hw_sched *t = (struct ch_hw_sched *)data; unsigned int ticks_per_usec = core_ticks_per_usec(sc); if ((sc->flags & FULL_INIT_DONE) == 0) return (EAGAIN); /* need TP to be initialized */ if (t->sched >= NTX_SCHED || !in_range(t->mode, 0, 1) || !in_range(t->channel, 0, 1) || !in_range(t->kbps, 0, 10000000) || !in_range(t->class_ipg, 0, 10000 * 65535 / ticks_per_usec) || !in_range(t->flow_ipg, 0, dack_ticks_to_usec(sc, 0x7ff))) return (EINVAL); if (t->kbps >= 0) { error = t3_config_sched(sc, t->kbps, t->sched); if (error < 0) return (-error); } if (t->class_ipg >= 0) t3_set_sched_ipg(sc, t->sched, t->class_ipg); if (t->flow_ipg >= 0) { t->flow_ipg *= 1000; /* us -> ns */ t3_set_pace_tbl(sc, &t->flow_ipg, t->sched, 1); } if (t->mode >= 0) { int bit = 1 << (S_TX_MOD_TIMER_MODE + t->sched); t3_set_reg_field(sc, A_TP_TX_MOD_QUEUE_REQ_MAP, bit, t->mode ? bit : 0); } if (t->channel >= 0) t3_set_reg_field(sc, A_TP_TX_MOD_QUEUE_REQ_MAP, 1 << t->sched, t->channel << t->sched); break; } default: return (EOPNOTSUPP); break; } return (error); } #endif static __inline void reg_block_dump(struct adapter *ap, uint8_t *buf, unsigned int start, unsigned int end) { uint32_t *p = (uint32_t *)buf + start; for ( ; start <= end; start += sizeof(uint32_t)) *p++ = t3_read_reg(ap, start); } #ifdef __FreeBSD__ #define T3_REGMAP_SIZE (3 * 1024) static int cxgb_get_regs_len(void) { return T3_REGMAP_SIZE; } #undef T3_REGMAP_SIZE static void cxgb_get_regs(adapter_t *sc, struct ifconf_regs *regs, uint8_t *buf) { /* * Version scheme: * bits 0..9: chip version * bits 10..15: chip revision * bit 31: set for PCIe cards */ regs->version = 3 | (sc->params.rev << 10) | (is_pcie(sc) << 31); /* * We skip the MAC statistics registers because they are clear-on-read. * Also reading multi-register stats would need to synchronize with the * periodic mac stats accumulation. Hard to justify the complexity. */ memset(buf, 0, REGDUMP_SIZE); reg_block_dump(sc, buf, 0, A_SG_RSPQ_CREDIT_RETURN); reg_block_dump(sc, buf, A_SG_HI_DRB_HI_THRSH, A_ULPRX_PBL_ULIMIT); reg_block_dump(sc, buf, A_ULPTX_CONFIG, A_MPS_INT_CAUSE); reg_block_dump(sc, buf, A_CPL_SWITCH_CNTRL, A_CPL_MAP_TBL_DATA); reg_block_dump(sc, buf, A_SMB_GLOBAL_TIME_CFG, A_XGM_SERDES_STAT3); reg_block_dump(sc, buf, A_XGM_SERDES_STATUS0, XGM_REG(A_XGM_SERDES_STAT3, 1)); reg_block_dump(sc, buf, XGM_REG(A_XGM_SERDES_STATUS0, 1), XGM_REG(A_XGM_RX_SPI4_SOP_EOP_CNT, 1)); } #endif