1476 lines
36 KiB
C
1476 lines
36 KiB
C
/* $NetBSD: aic6915.c,v 1.13 2005/02/27 00:27:00 perry Exp $ */
|
|
|
|
/*-
|
|
* Copyright (c) 2001 The NetBSD Foundation, Inc.
|
|
* All rights reserved.
|
|
*
|
|
* This code is derived from software contributed to The NetBSD Foundation
|
|
* by Jason R. Thorpe.
|
|
*
|
|
* Redistribution and use in source and binary forms, with or without
|
|
* modification, are permitted provided that the following conditions
|
|
* are met:
|
|
* 1. Redistributions of source code must retain the above copyright
|
|
* notice, this list of conditions and the following disclaimer.
|
|
* 2. Redistributions in binary form must reproduce the above copyright
|
|
* notice, this list of conditions and the following disclaimer in the
|
|
* documentation and/or other materials provided with the distribution.
|
|
* 3. All advertising materials mentioning features or use of this software
|
|
* must display the following acknowledgement:
|
|
* This product includes software developed by the NetBSD
|
|
* Foundation, Inc. and its contributors.
|
|
* 4. Neither the name of The NetBSD Foundation nor the names of its
|
|
* contributors may be used to endorse or promote products derived
|
|
* from this software without specific prior written permission.
|
|
*
|
|
* THIS SOFTWARE IS PROVIDED BY THE NETBSD FOUNDATION, INC. AND CONTRIBUTORS
|
|
* ``AS IS'' AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED
|
|
* TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR
|
|
* PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE FOUNDATION OR CONTRIBUTORS
|
|
* BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
|
|
* CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
|
|
* SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
|
|
* INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
|
|
* CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
|
|
* ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
|
|
* POSSIBILITY OF SUCH DAMAGE.
|
|
*/
|
|
|
|
/*
|
|
* Device driver for the Adaptec AIC-6915 (``Starfire'')
|
|
* 10/100 Ethernet controller.
|
|
*/
|
|
|
|
#include <sys/cdefs.h>
|
|
__KERNEL_RCSID(0, "$NetBSD: aic6915.c,v 1.13 2005/02/27 00:27:00 perry Exp $");
|
|
|
|
#include "bpfilter.h"
|
|
|
|
#include <sys/param.h>
|
|
#include <sys/systm.h>
|
|
#include <sys/callout.h>
|
|
#include <sys/mbuf.h>
|
|
#include <sys/malloc.h>
|
|
#include <sys/kernel.h>
|
|
#include <sys/socket.h>
|
|
#include <sys/ioctl.h>
|
|
#include <sys/errno.h>
|
|
#include <sys/device.h>
|
|
|
|
#include <uvm/uvm_extern.h>
|
|
|
|
#include <net/if.h>
|
|
#include <net/if_dl.h>
|
|
#include <net/if_media.h>
|
|
#include <net/if_ether.h>
|
|
|
|
#if NBPFILTER > 0
|
|
#include <net/bpf.h>
|
|
#endif
|
|
|
|
#include <machine/bus.h>
|
|
#include <machine/intr.h>
|
|
|
|
#include <dev/mii/miivar.h>
|
|
|
|
#include <dev/ic/aic6915reg.h>
|
|
#include <dev/ic/aic6915var.h>
|
|
|
|
static void sf_start(struct ifnet *);
|
|
static void sf_watchdog(struct ifnet *);
|
|
static int sf_ioctl(struct ifnet *, u_long, caddr_t);
|
|
static int sf_init(struct ifnet *);
|
|
static void sf_stop(struct ifnet *, int);
|
|
|
|
static void sf_shutdown(void *);
|
|
|
|
static void sf_txintr(struct sf_softc *);
|
|
static void sf_rxintr(struct sf_softc *);
|
|
static void sf_stats_update(struct sf_softc *);
|
|
|
|
static void sf_reset(struct sf_softc *);
|
|
static void sf_macreset(struct sf_softc *);
|
|
static void sf_rxdrain(struct sf_softc *);
|
|
static int sf_add_rxbuf(struct sf_softc *, int);
|
|
static uint8_t sf_read_eeprom(struct sf_softc *, int);
|
|
static void sf_set_filter(struct sf_softc *);
|
|
|
|
static int sf_mii_read(struct device *, int, int);
|
|
static void sf_mii_write(struct device *, int, int, int);
|
|
static void sf_mii_statchg(struct device *);
|
|
|
|
static void sf_tick(void *);
|
|
|
|
static int sf_mediachange(struct ifnet *);
|
|
static void sf_mediastatus(struct ifnet *, struct ifmediareq *);
|
|
|
|
#define sf_funcreg_read(sc, reg) \
|
|
bus_space_read_4((sc)->sc_st, (sc)->sc_sh_func, (reg))
|
|
#define sf_funcreg_write(sc, reg, val) \
|
|
bus_space_write_4((sc)->sc_st, (sc)->sc_sh_func, (reg), (val))
|
|
|
|
static __inline uint32_t
|
|
sf_reg_read(struct sf_softc *sc, bus_addr_t reg)
|
|
{
|
|
|
|
if (__predict_false(sc->sc_iomapped)) {
|
|
bus_space_write_4(sc->sc_st, sc->sc_sh, SF_IndirectIoAccess,
|
|
reg);
|
|
return (bus_space_read_4(sc->sc_st, sc->sc_sh,
|
|
SF_IndirectIoDataPort));
|
|
}
|
|
|
|
return (bus_space_read_4(sc->sc_st, sc->sc_sh, reg));
|
|
}
|
|
|
|
static __inline void
|
|
sf_reg_write(struct sf_softc *sc, bus_addr_t reg, uint32_t val)
|
|
{
|
|
|
|
if (__predict_false(sc->sc_iomapped)) {
|
|
bus_space_write_4(sc->sc_st, sc->sc_sh, SF_IndirectIoAccess,
|
|
reg);
|
|
bus_space_write_4(sc->sc_st, sc->sc_sh, SF_IndirectIoDataPort,
|
|
val);
|
|
return;
|
|
}
|
|
|
|
bus_space_write_4(sc->sc_st, sc->sc_sh, reg, val);
|
|
}
|
|
|
|
#define sf_genreg_read(sc, reg) \
|
|
sf_reg_read((sc), (reg) + SF_GENREG_OFFSET)
|
|
#define sf_genreg_write(sc, reg, val) \
|
|
sf_reg_write((sc), (reg) + SF_GENREG_OFFSET, (val))
|
|
|
|
/*
|
|
* sf_attach:
|
|
*
|
|
* Attach a Starfire interface to the system.
|
|
*/
|
|
void
|
|
sf_attach(struct sf_softc *sc)
|
|
{
|
|
struct ifnet *ifp = &sc->sc_ethercom.ec_if;
|
|
int i, rseg, error;
|
|
bus_dma_segment_t seg;
|
|
u_int8_t enaddr[ETHER_ADDR_LEN];
|
|
|
|
callout_init(&sc->sc_tick_callout);
|
|
|
|
/*
|
|
* If we're I/O mapped, the functional register handle is
|
|
* the same as the base handle. If we're memory mapped,
|
|
* carve off a chunk of the register space for the functional
|
|
* registers, to save on arithmetic later.
|
|
*/
|
|
if (sc->sc_iomapped)
|
|
sc->sc_sh_func = sc->sc_sh;
|
|
else {
|
|
if ((error = bus_space_subregion(sc->sc_st, sc->sc_sh,
|
|
SF_GENREG_OFFSET, SF_FUNCREG_SIZE, &sc->sc_sh_func)) != 0) {
|
|
printf("%s: unable to sub-region functional "
|
|
"registers, error = %d\n", sc->sc_dev.dv_xname,
|
|
error);
|
|
return;
|
|
}
|
|
}
|
|
|
|
/*
|
|
* Initialize the transmit threshold for this interface. The
|
|
* manual describes the default as 4 * 16 bytes. We start out
|
|
* at 10 * 16 bytes, to avoid a bunch of initial underruns on
|
|
* several platforms.
|
|
*/
|
|
sc->sc_txthresh = 10;
|
|
|
|
/*
|
|
* Allocate the control data structures, and create and load the
|
|
* DMA map for it.
|
|
*/
|
|
if ((error = bus_dmamem_alloc(sc->sc_dmat,
|
|
sizeof(struct sf_control_data), PAGE_SIZE, 0, &seg, 1, &rseg,
|
|
BUS_DMA_NOWAIT)) != 0) {
|
|
printf("%s: unable to allocate control data, error = %d\n",
|
|
sc->sc_dev.dv_xname, error);
|
|
goto fail_0;
|
|
}
|
|
|
|
if ((error = bus_dmamem_map(sc->sc_dmat, &seg, rseg,
|
|
sizeof(struct sf_control_data), (caddr_t *)&sc->sc_control_data,
|
|
BUS_DMA_NOWAIT|BUS_DMA_COHERENT)) != 0) {
|
|
printf("%s: unable to map control data, error = %d\n",
|
|
sc->sc_dev.dv_xname, error);
|
|
goto fail_1;
|
|
}
|
|
|
|
if ((error = bus_dmamap_create(sc->sc_dmat,
|
|
sizeof(struct sf_control_data), 1,
|
|
sizeof(struct sf_control_data), 0, BUS_DMA_NOWAIT,
|
|
&sc->sc_cddmamap)) != 0) {
|
|
printf("%s: unable to create control data DMA map, "
|
|
"error = %d\n", sc->sc_dev.dv_xname, error);
|
|
goto fail_2;
|
|
}
|
|
|
|
if ((error = bus_dmamap_load(sc->sc_dmat, sc->sc_cddmamap,
|
|
sc->sc_control_data, sizeof(struct sf_control_data), NULL,
|
|
BUS_DMA_NOWAIT)) != 0) {
|
|
printf("%s: unable to load control data DMA map, error = %d\n",
|
|
sc->sc_dev.dv_xname, error);
|
|
goto fail_3;
|
|
}
|
|
|
|
/*
|
|
* Create the transmit buffer DMA maps.
|
|
*/
|
|
for (i = 0; i < SF_NTXDESC; i++) {
|
|
if ((error = bus_dmamap_create(sc->sc_dmat, MCLBYTES,
|
|
SF_NTXFRAGS, MCLBYTES, 0, BUS_DMA_NOWAIT,
|
|
&sc->sc_txsoft[i].ds_dmamap)) != 0) {
|
|
printf("%s: unable to create tx DMA map %d, "
|
|
"error = %d\n", sc->sc_dev.dv_xname, i, error);
|
|
goto fail_4;
|
|
}
|
|
}
|
|
|
|
/*
|
|
* Create the receive buffer DMA maps.
|
|
*/
|
|
for (i = 0; i < SF_NRXDESC; i++) {
|
|
if ((error = bus_dmamap_create(sc->sc_dmat, MCLBYTES, 1,
|
|
MCLBYTES, 0, BUS_DMA_NOWAIT,
|
|
&sc->sc_rxsoft[i].ds_dmamap)) != 0) {
|
|
printf("%s: unable to create rx DMA map %d, "
|
|
"error = %d\n", sc->sc_dev.dv_xname, i, error);
|
|
goto fail_5;
|
|
}
|
|
}
|
|
|
|
/*
|
|
* Reset the chip to a known state.
|
|
*/
|
|
sf_reset(sc);
|
|
|
|
/*
|
|
* Read the Ethernet address from the EEPROM.
|
|
*/
|
|
for (i = 0; i < ETHER_ADDR_LEN; i++)
|
|
enaddr[i] = sf_read_eeprom(sc, (15 + (ETHER_ADDR_LEN - 1)) - i);
|
|
|
|
printf("%s: Ethernet address %s\n", sc->sc_dev.dv_xname,
|
|
ether_sprintf(enaddr));
|
|
|
|
if (sf_funcreg_read(sc, SF_PciDeviceConfig) & PDC_System64)
|
|
printf("%s: 64-bit PCI slot detected\n", sc->sc_dev.dv_xname);
|
|
|
|
/*
|
|
* Initialize our media structures and probe the MII.
|
|
*/
|
|
sc->sc_mii.mii_ifp = ifp;
|
|
sc->sc_mii.mii_readreg = sf_mii_read;
|
|
sc->sc_mii.mii_writereg = sf_mii_write;
|
|
sc->sc_mii.mii_statchg = sf_mii_statchg;
|
|
ifmedia_init(&sc->sc_mii.mii_media, IFM_IMASK, sf_mediachange,
|
|
sf_mediastatus);
|
|
mii_attach(&sc->sc_dev, &sc->sc_mii, 0xffffffff, MII_PHY_ANY,
|
|
MII_OFFSET_ANY, 0);
|
|
if (LIST_FIRST(&sc->sc_mii.mii_phys) == NULL) {
|
|
ifmedia_add(&sc->sc_mii.mii_media, IFM_ETHER|IFM_NONE, 0, NULL);
|
|
ifmedia_set(&sc->sc_mii.mii_media, IFM_ETHER|IFM_NONE);
|
|
} else
|
|
ifmedia_set(&sc->sc_mii.mii_media, IFM_ETHER|IFM_AUTO);
|
|
|
|
strcpy(ifp->if_xname, sc->sc_dev.dv_xname);
|
|
ifp->if_softc = sc;
|
|
ifp->if_flags = IFF_BROADCAST | IFF_SIMPLEX | IFF_MULTICAST;
|
|
ifp->if_ioctl = sf_ioctl;
|
|
ifp->if_start = sf_start;
|
|
ifp->if_watchdog = sf_watchdog;
|
|
ifp->if_init = sf_init;
|
|
ifp->if_stop = sf_stop;
|
|
IFQ_SET_READY(&ifp->if_snd);
|
|
|
|
/*
|
|
* Attach the interface.
|
|
*/
|
|
if_attach(ifp);
|
|
ether_ifattach(ifp, enaddr);
|
|
|
|
/*
|
|
* Make sure the interface is shutdown during reboot.
|
|
*/
|
|
sc->sc_sdhook = shutdownhook_establish(sf_shutdown, sc);
|
|
if (sc->sc_sdhook == NULL)
|
|
printf("%s: WARNING: unable to establish shutdown hook\n",
|
|
sc->sc_dev.dv_xname);
|
|
return;
|
|
|
|
/*
|
|
* Free any resources we've allocated during the failed attach
|
|
* attempt. Do this in reverse order an fall through.
|
|
*/
|
|
fail_5:
|
|
for (i = 0; i < SF_NRXDESC; i++) {
|
|
if (sc->sc_rxsoft[i].ds_dmamap != NULL)
|
|
bus_dmamap_destroy(sc->sc_dmat,
|
|
sc->sc_rxsoft[i].ds_dmamap);
|
|
}
|
|
fail_4:
|
|
for (i = 0; i < SF_NTXDESC; i++) {
|
|
if (sc->sc_txsoft[i].ds_dmamap != NULL)
|
|
bus_dmamap_destroy(sc->sc_dmat,
|
|
sc->sc_txsoft[i].ds_dmamap);
|
|
}
|
|
bus_dmamap_unload(sc->sc_dmat, sc->sc_cddmamap);
|
|
fail_3:
|
|
bus_dmamap_destroy(sc->sc_dmat, sc->sc_cddmamap);
|
|
fail_2:
|
|
bus_dmamem_unmap(sc->sc_dmat, (caddr_t) sc->sc_control_data,
|
|
sizeof(struct sf_control_data));
|
|
fail_1:
|
|
bus_dmamem_free(sc->sc_dmat, &seg, rseg);
|
|
fail_0:
|
|
return;
|
|
}
|
|
|
|
/*
|
|
* sf_shutdown:
|
|
*
|
|
* Shutdown hook -- make sure the interface is stopped at reboot.
|
|
*/
|
|
static void
|
|
sf_shutdown(void *arg)
|
|
{
|
|
struct sf_softc *sc = arg;
|
|
|
|
sf_stop(&sc->sc_ethercom.ec_if, 1);
|
|
}
|
|
|
|
/*
|
|
* sf_start: [ifnet interface function]
|
|
*
|
|
* Start packet transmission on the interface.
|
|
*/
|
|
static void
|
|
sf_start(struct ifnet *ifp)
|
|
{
|
|
struct sf_softc *sc = ifp->if_softc;
|
|
struct mbuf *m0, *m;
|
|
struct sf_txdesc0 *txd;
|
|
struct sf_descsoft *ds;
|
|
bus_dmamap_t dmamap;
|
|
int error, producer, last = -1, opending, seg;
|
|
|
|
/*
|
|
* Remember the previous number of pending transmits.
|
|
*/
|
|
opending = sc->sc_txpending;
|
|
|
|
/*
|
|
* Find out where we're sitting.
|
|
*/
|
|
producer = SF_TXDINDEX_TO_HOST(
|
|
TDQPI_HiPrTxProducerIndex_get(
|
|
sf_funcreg_read(sc, SF_TxDescQueueProducerIndex)));
|
|
|
|
/*
|
|
* Loop through the send queue, setting up transmit descriptors
|
|
* until we drain the queue, or use up all available transmit
|
|
* descriptors. Leave a blank one at the end for sanity's sake.
|
|
*/
|
|
while (sc->sc_txpending < (SF_NTXDESC - 1)) {
|
|
/*
|
|
* Grab a packet off the queue.
|
|
*/
|
|
IFQ_POLL(&ifp->if_snd, m0);
|
|
if (m0 == NULL)
|
|
break;
|
|
m = NULL;
|
|
|
|
/*
|
|
* Get the transmit descriptor.
|
|
*/
|
|
txd = &sc->sc_txdescs[producer];
|
|
ds = &sc->sc_txsoft[producer];
|
|
dmamap = ds->ds_dmamap;
|
|
|
|
/*
|
|
* Load the DMA map. If this fails, the packet either
|
|
* didn't fit in the allotted number of frags, or we were
|
|
* short on resources. In this case, we'll copy and try
|
|
* again.
|
|
*/
|
|
if (bus_dmamap_load_mbuf(sc->sc_dmat, dmamap, m0,
|
|
BUS_DMA_WRITE|BUS_DMA_NOWAIT) != 0) {
|
|
MGETHDR(m, M_DONTWAIT, MT_DATA);
|
|
if (m == NULL) {
|
|
printf("%s: unable to allocate Tx mbuf\n",
|
|
sc->sc_dev.dv_xname);
|
|
break;
|
|
}
|
|
if (m0->m_pkthdr.len > MHLEN) {
|
|
MCLGET(m, M_DONTWAIT);
|
|
if ((m->m_flags & M_EXT) == 0) {
|
|
printf("%s: unable to allocate Tx "
|
|
"cluster\n", sc->sc_dev.dv_xname);
|
|
m_freem(m);
|
|
break;
|
|
}
|
|
}
|
|
m_copydata(m0, 0, m0->m_pkthdr.len, mtod(m, caddr_t));
|
|
m->m_pkthdr.len = m->m_len = m0->m_pkthdr.len;
|
|
error = bus_dmamap_load_mbuf(sc->sc_dmat, dmamap,
|
|
m, BUS_DMA_WRITE|BUS_DMA_NOWAIT);
|
|
if (error) {
|
|
printf("%s: unable to load Tx buffer, "
|
|
"error = %d\n", sc->sc_dev.dv_xname, error);
|
|
break;
|
|
}
|
|
}
|
|
|
|
/*
|
|
* WE ARE NOW COMMITTED TO TRANSMITTING THE PACKET.
|
|
*/
|
|
IFQ_DEQUEUE(&ifp->if_snd, m0);
|
|
if (m != NULL) {
|
|
m_freem(m0);
|
|
m0 = m;
|
|
}
|
|
|
|
/* Initialize the descriptor. */
|
|
txd->td_word0 =
|
|
htole32(TD_W0_ID | TD_W0_CRCEN | m0->m_pkthdr.len);
|
|
if (producer == (SF_NTXDESC - 1))
|
|
txd->td_word0 |= TD_W0_END;
|
|
txd->td_word1 = htole32(dmamap->dm_nsegs);
|
|
for (seg = 0; seg < dmamap->dm_nsegs; seg++) {
|
|
txd->td_frags[seg].fr_addr =
|
|
htole32(dmamap->dm_segs[seg].ds_addr);
|
|
txd->td_frags[seg].fr_len =
|
|
htole32(dmamap->dm_segs[seg].ds_len);
|
|
}
|
|
|
|
/* Sync the descriptor and the DMA map. */
|
|
SF_CDTXDSYNC(sc, producer, BUS_DMASYNC_PREWRITE);
|
|
bus_dmamap_sync(sc->sc_dmat, dmamap, 0, dmamap->dm_mapsize,
|
|
BUS_DMASYNC_PREWRITE);
|
|
|
|
/*
|
|
* Store a pointer to the packet so we can free it later.
|
|
*/
|
|
ds->ds_mbuf = m0;
|
|
|
|
/* Advance the Tx pointer. */
|
|
sc->sc_txpending++;
|
|
last = producer;
|
|
producer = SF_NEXTTX(producer);
|
|
|
|
#if NBPFILTER > 0
|
|
/*
|
|
* Pass the packet to any BPF listeners.
|
|
*/
|
|
if (ifp->if_bpf)
|
|
bpf_mtap(ifp->if_bpf, m0);
|
|
#endif
|
|
}
|
|
|
|
if (sc->sc_txpending == (SF_NTXDESC - 1)) {
|
|
/* No more slots left; notify upper layer. */
|
|
ifp->if_flags |= IFF_OACTIVE;
|
|
}
|
|
|
|
if (sc->sc_txpending != opending) {
|
|
KASSERT(last != -1);
|
|
/*
|
|
* We enqueued packets. Cause a transmit interrupt to
|
|
* happen on the last packet we enqueued, and give the
|
|
* new descriptors to the chip by writing the new
|
|
* producer index.
|
|
*/
|
|
sc->sc_txdescs[last].td_word0 |= TD_W0_INTR;
|
|
SF_CDTXDSYNC(sc, last, BUS_DMASYNC_PREWRITE);
|
|
|
|
sf_funcreg_write(sc, SF_TxDescQueueProducerIndex,
|
|
TDQPI_HiPrTxProducerIndex(SF_TXDINDEX_TO_CHIP(producer)));
|
|
|
|
/* Set a watchdog timer in case the chip flakes out. */
|
|
ifp->if_timer = 5;
|
|
}
|
|
}
|
|
|
|
/*
|
|
* sf_watchdog: [ifnet interface function]
|
|
*
|
|
* Watchdog timer handler.
|
|
*/
|
|
static void
|
|
sf_watchdog(struct ifnet *ifp)
|
|
{
|
|
struct sf_softc *sc = ifp->if_softc;
|
|
|
|
printf("%s: device timeout\n", sc->sc_dev.dv_xname);
|
|
ifp->if_oerrors++;
|
|
|
|
(void) sf_init(ifp);
|
|
|
|
/* Try to get more packets going. */
|
|
sf_start(ifp);
|
|
}
|
|
|
|
/*
|
|
* sf_ioctl: [ifnet interface function]
|
|
*
|
|
* Handle control requests from the operator.
|
|
*/
|
|
static int
|
|
sf_ioctl(struct ifnet *ifp, u_long cmd, caddr_t data)
|
|
{
|
|
struct sf_softc *sc = ifp->if_softc;
|
|
struct ifreq *ifr = (struct ifreq *) data;
|
|
int s, error;
|
|
|
|
s = splnet();
|
|
|
|
switch (cmd) {
|
|
case SIOCSIFMEDIA:
|
|
case SIOCGIFMEDIA:
|
|
error = ifmedia_ioctl(ifp, ifr, &sc->sc_mii.mii_media, cmd);
|
|
break;
|
|
|
|
default:
|
|
error = ether_ioctl(ifp, cmd, data);
|
|
if (error == ENETRESET) {
|
|
/*
|
|
* Multicast list has changed; set the hardware filter
|
|
* accordingly.
|
|
*/
|
|
if (ifp->if_flags & IFF_RUNNING)
|
|
sf_set_filter(sc);
|
|
error = 0;
|
|
}
|
|
break;
|
|
}
|
|
|
|
/* Try to get more packets going. */
|
|
sf_start(ifp);
|
|
|
|
splx(s);
|
|
return (error);
|
|
}
|
|
|
|
/*
|
|
* sf_intr:
|
|
*
|
|
* Interrupt service routine.
|
|
*/
|
|
int
|
|
sf_intr(void *arg)
|
|
{
|
|
struct sf_softc *sc = arg;
|
|
uint32_t isr;
|
|
int handled = 0, wantinit = 0;
|
|
|
|
for (;;) {
|
|
/* Reading clears all interrupts we're interested in. */
|
|
isr = sf_funcreg_read(sc, SF_InterruptStatus);
|
|
if ((isr & IS_PCIPadInt) == 0)
|
|
break;
|
|
|
|
handled = 1;
|
|
|
|
/* Handle receive interrupts. */
|
|
if (isr & IS_RxQ1DoneInt)
|
|
sf_rxintr(sc);
|
|
|
|
/* Handle transmit completion interrupts. */
|
|
if (isr & (IS_TxDmaDoneInt|IS_TxQueueDoneInt))
|
|
sf_txintr(sc);
|
|
|
|
/* Handle abnormal interrupts. */
|
|
if (isr & IS_AbnormalInterrupt) {
|
|
/* Statistics. */
|
|
if (isr & IS_StatisticWrapInt)
|
|
sf_stats_update(sc);
|
|
|
|
/* DMA errors. */
|
|
if (isr & IS_DmaErrInt) {
|
|
wantinit = 1;
|
|
printf("%s: WARNING: DMA error\n",
|
|
sc->sc_dev.dv_xname);
|
|
}
|
|
|
|
/* Transmit FIFO underruns. */
|
|
if (isr & IS_TxDataLowInt) {
|
|
if (sc->sc_txthresh < 0xff)
|
|
sc->sc_txthresh++;
|
|
printf("%s: transmit FIFO underrun, new "
|
|
"threshold: %d bytes\n",
|
|
sc->sc_dev.dv_xname,
|
|
sc->sc_txthresh * 16);
|
|
sf_funcreg_write(sc, SF_TransmitFrameCSR,
|
|
sc->sc_TransmitFrameCSR |
|
|
TFCSR_TransmitThreshold(sc->sc_txthresh));
|
|
sf_funcreg_write(sc, SF_TxDescQueueCtrl,
|
|
sc->sc_TxDescQueueCtrl |
|
|
TDQC_TxHighPriorityFifoThreshold(
|
|
sc->sc_txthresh));
|
|
}
|
|
}
|
|
}
|
|
|
|
if (handled) {
|
|
/* Reset the interface, if necessary. */
|
|
if (wantinit)
|
|
sf_init(&sc->sc_ethercom.ec_if);
|
|
|
|
/* Try and get more packets going. */
|
|
sf_start(&sc->sc_ethercom.ec_if);
|
|
}
|
|
|
|
return (handled);
|
|
}
|
|
|
|
/*
|
|
* sf_txintr:
|
|
*
|
|
* Helper -- handle transmit completion interrupts.
|
|
*/
|
|
static void
|
|
sf_txintr(struct sf_softc *sc)
|
|
{
|
|
struct ifnet *ifp = &sc->sc_ethercom.ec_if;
|
|
struct sf_descsoft *ds;
|
|
uint32_t cqci, tcd;
|
|
int consumer, producer, txidx;
|
|
|
|
try_again:
|
|
cqci = sf_funcreg_read(sc, SF_CompletionQueueConsumerIndex);
|
|
|
|
consumer = CQCI_TxCompletionConsumerIndex_get(cqci);
|
|
producer = CQPI_TxCompletionProducerIndex_get(
|
|
sf_funcreg_read(sc, SF_CompletionQueueProducerIndex));
|
|
|
|
if (consumer == producer)
|
|
return;
|
|
|
|
ifp->if_flags &= ~IFF_OACTIVE;
|
|
|
|
while (consumer != producer) {
|
|
SF_CDTXCSYNC(sc, consumer, BUS_DMASYNC_POSTREAD);
|
|
tcd = le32toh(sc->sc_txcomp[consumer].tcd_word0);
|
|
|
|
txidx = SF_TCD_INDEX_TO_HOST(TCD_INDEX(tcd));
|
|
#ifdef DIAGNOSTIC
|
|
if ((tcd & TCD_PR) == 0)
|
|
printf("%s: Tx queue mismatch, index %d\n",
|
|
sc->sc_dev.dv_xname, txidx);
|
|
#endif
|
|
/*
|
|
* NOTE: stats are updated later. We're just
|
|
* releasing packets that have been DMA'd to
|
|
* the chip.
|
|
*/
|
|
ds = &sc->sc_txsoft[txidx];
|
|
SF_CDTXDSYNC(sc, txidx, BUS_DMASYNC_POSTWRITE);
|
|
bus_dmamap_sync(sc->sc_dmat, ds->ds_dmamap,
|
|
0, ds->ds_dmamap->dm_mapsize,
|
|
BUS_DMASYNC_POSTWRITE);
|
|
m_freem(ds->ds_mbuf);
|
|
ds->ds_mbuf = NULL;
|
|
|
|
consumer = SF_NEXTTCD(consumer);
|
|
sc->sc_txpending--;
|
|
}
|
|
|
|
/* XXXJRT -- should be KDASSERT() */
|
|
KASSERT(sc->sc_txpending >= 0);
|
|
|
|
/* If all packets are done, cancel the watchdog timer. */
|
|
if (sc->sc_txpending == 0)
|
|
ifp->if_timer = 0;
|
|
|
|
/* Update the consumer index. */
|
|
sf_funcreg_write(sc, SF_CompletionQueueConsumerIndex,
|
|
(cqci & ~CQCI_TxCompletionConsumerIndex(0x7ff)) |
|
|
CQCI_TxCompletionConsumerIndex(consumer));
|
|
|
|
/* Double check for new completions. */
|
|
goto try_again;
|
|
}
|
|
|
|
/*
|
|
* sf_rxintr:
|
|
*
|
|
* Helper -- handle receive interrupts.
|
|
*/
|
|
static void
|
|
sf_rxintr(struct sf_softc *sc)
|
|
{
|
|
struct ifnet *ifp = &sc->sc_ethercom.ec_if;
|
|
struct sf_descsoft *ds;
|
|
struct sf_rcd_full *rcd;
|
|
struct mbuf *m;
|
|
uint32_t cqci, word0;
|
|
int consumer, producer, bufproducer, rxidx, len;
|
|
|
|
try_again:
|
|
cqci = sf_funcreg_read(sc, SF_CompletionQueueConsumerIndex);
|
|
|
|
consumer = CQCI_RxCompletionQ1ConsumerIndex_get(cqci);
|
|
producer = CQPI_RxCompletionQ1ProducerIndex_get(
|
|
sf_funcreg_read(sc, SF_CompletionQueueProducerIndex));
|
|
bufproducer = RXQ1P_RxDescQ1Producer_get(
|
|
sf_funcreg_read(sc, SF_RxDescQueue1Ptrs));
|
|
|
|
if (consumer == producer)
|
|
return;
|
|
|
|
while (consumer != producer) {
|
|
rcd = &sc->sc_rxcomp[consumer];
|
|
SF_CDRXCSYNC(sc, consumer,
|
|
BUS_DMASYNC_POSTREAD|BUS_DMASYNC_POSTWRITE);
|
|
SF_CDRXCSYNC(sc, consumer,
|
|
BUS_DMASYNC_PREREAD|BUS_DMASYNC_PREWRITE);
|
|
|
|
word0 = le32toh(rcd->rcd_word0);
|
|
rxidx = RCD_W0_EndIndex(word0);
|
|
|
|
ds = &sc->sc_rxsoft[rxidx];
|
|
|
|
consumer = SF_NEXTRCD(consumer);
|
|
bufproducer = SF_NEXTRX(bufproducer);
|
|
|
|
if ((word0 & RCD_W0_OK) == 0) {
|
|
SF_INIT_RXDESC(sc, rxidx);
|
|
continue;
|
|
}
|
|
|
|
bus_dmamap_sync(sc->sc_dmat, ds->ds_dmamap, 0,
|
|
ds->ds_dmamap->dm_mapsize, BUS_DMASYNC_POSTREAD);
|
|
|
|
/*
|
|
* No errors; receive the packet. Note that we have
|
|
* configured the Starfire to NOT transfer the CRC
|
|
* with the packet.
|
|
*/
|
|
len = RCD_W0_Length(word0);
|
|
|
|
#ifdef __NO_STRICT_ALIGNMENT
|
|
/*
|
|
* Allocate a new mbuf cluster. If that fails, we are
|
|
* out of memory, and must drop the packet and recycle
|
|
* the buffer that's already attached to this descriptor.
|
|
*/
|
|
m = ds->ds_mbuf;
|
|
if (sf_add_rxbuf(sc, rxidx) != 0) {
|
|
ifp->if_ierrors++;
|
|
SF_INIT_RXDESC(sc, rxidx);
|
|
bus_dmamap_sync(sc->sc_dmat, ds->ds_dmamap, 0,
|
|
ds->ds_dmamap->dm_mapsize, BUS_DMASYNC_PREREAD);
|
|
continue;
|
|
}
|
|
#else
|
|
/*
|
|
* The Starfire's receive buffer must be 4-byte aligned.
|
|
* But this means that the data after the Ethernet header
|
|
* is misaligned. We must allocate a new buffer and
|
|
* copy the data, shifted forward 2 bytes.
|
|
*/
|
|
MGETHDR(m, M_DONTWAIT, MT_DATA);
|
|
if (m == NULL) {
|
|
dropit:
|
|
ifp->if_ierrors++;
|
|
SF_INIT_RXDESC(sc, rxidx);
|
|
bus_dmamap_sync(sc->sc_dmat, ds->ds_dmamap, 0,
|
|
ds->ds_dmamap->dm_mapsize, BUS_DMASYNC_PREREAD);
|
|
continue;
|
|
}
|
|
if (len > (MHLEN - 2)) {
|
|
MCLGET(m, M_DONTWAIT);
|
|
if ((m->m_flags & M_EXT) == 0) {
|
|
m_freem(m);
|
|
goto dropit;
|
|
}
|
|
}
|
|
m->m_data += 2;
|
|
|
|
/*
|
|
* Note that we use cluster for incoming frames, so the
|
|
* buffer is virtually contiguous.
|
|
*/
|
|
memcpy(mtod(m, caddr_t), mtod(ds->ds_mbuf, caddr_t), len);
|
|
|
|
/* Allow the receive descriptor to continue using its mbuf. */
|
|
SF_INIT_RXDESC(sc, rxidx);
|
|
bus_dmamap_sync(sc->sc_dmat, ds->ds_dmamap, 0,
|
|
ds->ds_dmamap->dm_mapsize, BUS_DMASYNC_PREREAD);
|
|
#endif /* __NO_STRICT_ALIGNMENT */
|
|
|
|
m->m_pkthdr.rcvif = ifp;
|
|
m->m_pkthdr.len = m->m_len = len;
|
|
|
|
#if NBPFILTER > 0
|
|
/*
|
|
* Pass this up to any BPF listeners.
|
|
*/
|
|
if (ifp->if_bpf)
|
|
bpf_mtap(ifp->if_bpf, m);
|
|
#endif /* NBPFILTER > 0 */
|
|
|
|
/* Pass it on. */
|
|
(*ifp->if_input)(ifp, m);
|
|
}
|
|
|
|
/* Update the chip's pointers. */
|
|
sf_funcreg_write(sc, SF_CompletionQueueConsumerIndex,
|
|
(cqci & ~CQCI_RxCompletionQ1ConsumerIndex(0x7ff)) |
|
|
CQCI_RxCompletionQ1ConsumerIndex(consumer));
|
|
sf_funcreg_write(sc, SF_RxDescQueue1Ptrs,
|
|
RXQ1P_RxDescQ1Producer(bufproducer));
|
|
|
|
/* Double-check for any new completions. */
|
|
goto try_again;
|
|
}
|
|
|
|
/*
|
|
* sf_tick:
|
|
*
|
|
* One second timer, used to tick the MII and update stats.
|
|
*/
|
|
static void
|
|
sf_tick(void *arg)
|
|
{
|
|
struct sf_softc *sc = arg;
|
|
int s;
|
|
|
|
s = splnet();
|
|
mii_tick(&sc->sc_mii);
|
|
sf_stats_update(sc);
|
|
splx(s);
|
|
|
|
callout_reset(&sc->sc_tick_callout, hz, sf_tick, sc);
|
|
}
|
|
|
|
/*
|
|
* sf_stats_update:
|
|
*
|
|
* Read the statitistics counters.
|
|
*/
|
|
static void
|
|
sf_stats_update(struct sf_softc *sc)
|
|
{
|
|
struct sf_stats stats;
|
|
struct ifnet *ifp = &sc->sc_ethercom.ec_if;
|
|
uint32_t *p;
|
|
u_int i;
|
|
|
|
p = &stats.TransmitOKFrames;
|
|
for (i = 0; i < (sizeof(stats) / sizeof(uint32_t)); i++) {
|
|
*p++ = sf_genreg_read(sc,
|
|
SF_STATS_BASE + (i * sizeof(uint32_t)));
|
|
sf_genreg_write(sc, SF_STATS_BASE + (i * sizeof(uint32_t)), 0);
|
|
}
|
|
|
|
ifp->if_opackets += stats.TransmitOKFrames;
|
|
|
|
ifp->if_collisions += stats.SingleCollisionFrames +
|
|
stats.MultipleCollisionFrames;
|
|
|
|
ifp->if_oerrors += stats.TransmitAbortDueToExcessiveCollisions +
|
|
stats.TransmitAbortDueToExcessingDeferral +
|
|
stats.FramesLostDueToInternalTransmitErrors;
|
|
|
|
ifp->if_ipackets += stats.ReceiveOKFrames;
|
|
|
|
ifp->if_ierrors += stats.ReceiveCRCErrors + stats.AlignmentErrors +
|
|
stats.ReceiveFramesTooLong + stats.ReceiveFramesTooShort +
|
|
stats.ReceiveFramesJabbersError +
|
|
stats.FramesLostDueToInternalReceiveErrors;
|
|
}
|
|
|
|
/*
|
|
* sf_reset:
|
|
*
|
|
* Perform a soft reset on the Starfire.
|
|
*/
|
|
static void
|
|
sf_reset(struct sf_softc *sc)
|
|
{
|
|
int i;
|
|
|
|
sf_funcreg_write(sc, SF_GeneralEthernetCtrl, 0);
|
|
|
|
sf_macreset(sc);
|
|
|
|
sf_funcreg_write(sc, SF_PciDeviceConfig, PDC_SoftReset);
|
|
for (i = 0; i < 1000; i++) {
|
|
delay(10);
|
|
if ((sf_funcreg_read(sc, SF_PciDeviceConfig) &
|
|
PDC_SoftReset) == 0)
|
|
break;
|
|
}
|
|
|
|
if (i == 1000) {
|
|
printf("%s: reset failed to complete\n", sc->sc_dev.dv_xname);
|
|
sf_funcreg_write(sc, SF_PciDeviceConfig, 0);
|
|
}
|
|
|
|
delay(1000);
|
|
}
|
|
|
|
/*
|
|
* sf_macreset:
|
|
*
|
|
* Reset the MAC portion of the Starfire.
|
|
*/
|
|
static void
|
|
sf_macreset(struct sf_softc *sc)
|
|
{
|
|
|
|
sf_genreg_write(sc, SF_MacConfig1, sc->sc_MacConfig1 | MC1_SoftRst);
|
|
delay(1000);
|
|
sf_genreg_write(sc, SF_MacConfig1, sc->sc_MacConfig1);
|
|
}
|
|
|
|
/*
|
|
* sf_init: [ifnet interface function]
|
|
*
|
|
* Initialize the interface. Must be called at splnet().
|
|
*/
|
|
static int
|
|
sf_init(struct ifnet *ifp)
|
|
{
|
|
struct sf_softc *sc = ifp->if_softc;
|
|
struct sf_descsoft *ds;
|
|
int error = 0;
|
|
u_int i;
|
|
|
|
/*
|
|
* Cancel any pending I/O.
|
|
*/
|
|
sf_stop(ifp, 0);
|
|
|
|
/*
|
|
* Reset the Starfire to a known state.
|
|
*/
|
|
sf_reset(sc);
|
|
|
|
/* Clear the stat counters. */
|
|
for (i = 0; i < sizeof(struct sf_stats); i += sizeof(uint32_t))
|
|
sf_genreg_write(sc, SF_STATS_BASE + i, 0);
|
|
|
|
/*
|
|
* Initialize the transmit descriptor ring.
|
|
*/
|
|
memset(sc->sc_txdescs, 0, sizeof(sc->sc_txdescs));
|
|
sf_funcreg_write(sc, SF_TxDescQueueHighAddr, 0);
|
|
sf_funcreg_write(sc, SF_HiPrTxDescQueueBaseAddr, SF_CDTXDADDR(sc, 0));
|
|
sf_funcreg_write(sc, SF_LoPrTxDescQueueBaseAddr, 0);
|
|
|
|
/*
|
|
* Initialize the transmit completion ring.
|
|
*/
|
|
for (i = 0; i < SF_NTCD; i++) {
|
|
sc->sc_txcomp[i].tcd_word0 = TCD_DMA_ID;
|
|
SF_CDTXCSYNC(sc, i, BUS_DMASYNC_PREREAD|BUS_DMASYNC_PREWRITE);
|
|
}
|
|
sf_funcreg_write(sc, SF_CompletionQueueHighAddr, 0);
|
|
sf_funcreg_write(sc, SF_TxCompletionQueueCtrl, SF_CDTXCADDR(sc, 0));
|
|
|
|
/*
|
|
* Initialize the receive descriptor ring.
|
|
*/
|
|
for (i = 0; i < SF_NRXDESC; i++) {
|
|
ds = &sc->sc_rxsoft[i];
|
|
if (ds->ds_mbuf == NULL) {
|
|
if ((error = sf_add_rxbuf(sc, i)) != 0) {
|
|
printf("%s: unable to allocate or map rx "
|
|
"buffer %d, error = %d\n",
|
|
sc->sc_dev.dv_xname, i, error);
|
|
/*
|
|
* XXX Should attempt to run with fewer receive
|
|
* XXX buffers instead of just failing.
|
|
*/
|
|
sf_rxdrain(sc);
|
|
goto out;
|
|
}
|
|
} else
|
|
SF_INIT_RXDESC(sc, i);
|
|
}
|
|
sf_funcreg_write(sc, SF_RxDescQueueHighAddress, 0);
|
|
sf_funcreg_write(sc, SF_RxDescQueue1LowAddress, SF_CDRXDADDR(sc, 0));
|
|
sf_funcreg_write(sc, SF_RxDescQueue2LowAddress, 0);
|
|
|
|
/*
|
|
* Initialize the receive completion ring.
|
|
*/
|
|
for (i = 0; i < SF_NRCD; i++) {
|
|
sc->sc_rxcomp[i].rcd_word0 = RCD_W0_ID;
|
|
sc->sc_rxcomp[i].rcd_word1 = 0;
|
|
sc->sc_rxcomp[i].rcd_word2 = 0;
|
|
sc->sc_rxcomp[i].rcd_timestamp = 0;
|
|
SF_CDRXCSYNC(sc, i, BUS_DMASYNC_PREREAD|BUS_DMASYNC_PREWRITE);
|
|
}
|
|
sf_funcreg_write(sc, SF_RxCompletionQueue1Ctrl, SF_CDRXCADDR(sc, 0) |
|
|
RCQ1C_RxCompletionQ1Type(3));
|
|
sf_funcreg_write(sc, SF_RxCompletionQueue2Ctrl, 0);
|
|
|
|
/*
|
|
* Initialize the Tx CSR.
|
|
*/
|
|
sc->sc_TransmitFrameCSR = 0;
|
|
sf_funcreg_write(sc, SF_TransmitFrameCSR,
|
|
sc->sc_TransmitFrameCSR |
|
|
TFCSR_TransmitThreshold(sc->sc_txthresh));
|
|
|
|
/*
|
|
* Initialize the Tx descriptor control register.
|
|
*/
|
|
sc->sc_TxDescQueueCtrl = TDQC_SkipLength(0) |
|
|
TDQC_TxDmaBurstSize(4) | /* default */
|
|
TDQC_MinFrameSpacing(3) | /* 128 bytes */
|
|
TDQC_TxDescType(0);
|
|
sf_funcreg_write(sc, SF_TxDescQueueCtrl,
|
|
sc->sc_TxDescQueueCtrl |
|
|
TDQC_TxHighPriorityFifoThreshold(sc->sc_txthresh));
|
|
|
|
/*
|
|
* Initialize the Rx descriptor control registers.
|
|
*/
|
|
sf_funcreg_write(sc, SF_RxDescQueue1Ctrl,
|
|
RDQ1C_RxQ1BufferLength(MCLBYTES) |
|
|
RDQ1C_RxDescSpacing(0));
|
|
sf_funcreg_write(sc, SF_RxDescQueue2Ctrl, 0);
|
|
|
|
/*
|
|
* Initialize the Tx descriptor producer indices.
|
|
*/
|
|
sf_funcreg_write(sc, SF_TxDescQueueProducerIndex,
|
|
TDQPI_HiPrTxProducerIndex(0) |
|
|
TDQPI_LoPrTxProducerIndex(0));
|
|
|
|
/*
|
|
* Initialize the Rx descriptor producer indices.
|
|
*/
|
|
sf_funcreg_write(sc, SF_RxDescQueue1Ptrs,
|
|
RXQ1P_RxDescQ1Producer(SF_NRXDESC - 1));
|
|
sf_funcreg_write(sc, SF_RxDescQueue2Ptrs,
|
|
RXQ2P_RxDescQ2Producer(0));
|
|
|
|
/*
|
|
* Initialize the Tx and Rx completion queue consumer indices.
|
|
*/
|
|
sf_funcreg_write(sc, SF_CompletionQueueConsumerIndex,
|
|
CQCI_TxCompletionConsumerIndex(0) |
|
|
CQCI_RxCompletionQ1ConsumerIndex(0));
|
|
sf_funcreg_write(sc, SF_RxHiPrCompletionPtrs, 0);
|
|
|
|
/*
|
|
* Initialize the Rx DMA control register.
|
|
*/
|
|
sf_funcreg_write(sc, SF_RxDmaCtrl,
|
|
RDC_RxHighPriorityThreshold(6) | /* default */
|
|
RDC_RxBurstSize(4)); /* default */
|
|
|
|
/*
|
|
* Set the receive filter.
|
|
*/
|
|
sc->sc_RxAddressFilteringCtl = 0;
|
|
sf_set_filter(sc);
|
|
|
|
/*
|
|
* Set MacConfig1. When we set the media, MacConfig1 will
|
|
* actually be written and the MAC part reset.
|
|
*/
|
|
sc->sc_MacConfig1 = MC1_PadEn;
|
|
|
|
/*
|
|
* Set the media.
|
|
*/
|
|
mii_mediachg(&sc->sc_mii);
|
|
|
|
/*
|
|
* Initialize the interrupt register.
|
|
*/
|
|
sc->sc_InterruptEn = IS_PCIPadInt | IS_RxQ1DoneInt |
|
|
IS_TxQueueDoneInt | IS_TxDmaDoneInt | IS_DmaErrInt |
|
|
IS_StatisticWrapInt;
|
|
sf_funcreg_write(sc, SF_InterruptEn, sc->sc_InterruptEn);
|
|
|
|
sf_funcreg_write(sc, SF_PciDeviceConfig, PDC_IntEnable |
|
|
PDC_PCIMstDmaEn | (1 << PDC_FifoThreshold_SHIFT));
|
|
|
|
/*
|
|
* Start the transmit and receive processes.
|
|
*/
|
|
sf_funcreg_write(sc, SF_GeneralEthernetCtrl,
|
|
GEC_TxDmaEn|GEC_RxDmaEn|GEC_TransmitEn|GEC_ReceiveEn);
|
|
|
|
/* Start the on second clock. */
|
|
callout_reset(&sc->sc_tick_callout, hz, sf_tick, sc);
|
|
|
|
/*
|
|
* Note that the interface is now running.
|
|
*/
|
|
ifp->if_flags |= IFF_RUNNING;
|
|
ifp->if_flags &= ~IFF_OACTIVE;
|
|
|
|
out:
|
|
if (error) {
|
|
ifp->if_flags &= ~(IFF_RUNNING | IFF_OACTIVE);
|
|
ifp->if_timer = 0;
|
|
printf("%s: interface not running\n", sc->sc_dev.dv_xname);
|
|
}
|
|
return (error);
|
|
}
|
|
|
|
/*
|
|
* sf_rxdrain:
|
|
*
|
|
* Drain the receive queue.
|
|
*/
|
|
static void
|
|
sf_rxdrain(struct sf_softc *sc)
|
|
{
|
|
struct sf_descsoft *ds;
|
|
int i;
|
|
|
|
for (i = 0; i < SF_NRXDESC; i++) {
|
|
ds = &sc->sc_rxsoft[i];
|
|
if (ds->ds_mbuf != NULL) {
|
|
bus_dmamap_unload(sc->sc_dmat, ds->ds_dmamap);
|
|
m_freem(ds->ds_mbuf);
|
|
ds->ds_mbuf = NULL;
|
|
}
|
|
}
|
|
}
|
|
|
|
/*
|
|
* sf_stop: [ifnet interface function]
|
|
*
|
|
* Stop transmission on the interface.
|
|
*/
|
|
static void
|
|
sf_stop(struct ifnet *ifp, int disable)
|
|
{
|
|
struct sf_softc *sc = ifp->if_softc;
|
|
struct sf_descsoft *ds;
|
|
int i;
|
|
|
|
/* Stop the one second clock. */
|
|
callout_stop(&sc->sc_tick_callout);
|
|
|
|
/* Down the MII. */
|
|
mii_down(&sc->sc_mii);
|
|
|
|
/* Disable interrupts. */
|
|
sf_funcreg_write(sc, SF_InterruptEn, 0);
|
|
|
|
/* Stop the transmit and receive processes. */
|
|
sf_funcreg_write(sc, SF_GeneralEthernetCtrl, 0);
|
|
|
|
/*
|
|
* Release any queued transmit buffers.
|
|
*/
|
|
for (i = 0; i < SF_NTXDESC; i++) {
|
|
ds = &sc->sc_txsoft[i];
|
|
if (ds->ds_mbuf != NULL) {
|
|
bus_dmamap_unload(sc->sc_dmat, ds->ds_dmamap);
|
|
m_freem(ds->ds_mbuf);
|
|
ds->ds_mbuf = NULL;
|
|
}
|
|
}
|
|
|
|
if (disable)
|
|
sf_rxdrain(sc);
|
|
|
|
/*
|
|
* Mark the interface down and cancel the watchdog timer.
|
|
*/
|
|
ifp->if_flags &= ~(IFF_RUNNING | IFF_OACTIVE);
|
|
ifp->if_timer = 0;
|
|
}
|
|
|
|
/*
|
|
* sf_read_eeprom:
|
|
*
|
|
* Read from the Starfire EEPROM.
|
|
*/
|
|
static uint8_t
|
|
sf_read_eeprom(struct sf_softc *sc, int offset)
|
|
{
|
|
uint32_t reg;
|
|
|
|
reg = sf_genreg_read(sc, SF_EEPROM_BASE + (offset & ~3));
|
|
|
|
return ((reg >> (8 * (offset & 3))) & 0xff);
|
|
}
|
|
|
|
/*
|
|
* sf_add_rxbuf:
|
|
*
|
|
* Add a receive buffer to the indicated descriptor.
|
|
*/
|
|
static int
|
|
sf_add_rxbuf(struct sf_softc *sc, int idx)
|
|
{
|
|
struct sf_descsoft *ds = &sc->sc_rxsoft[idx];
|
|
struct mbuf *m;
|
|
int error;
|
|
|
|
MGETHDR(m, M_DONTWAIT, MT_DATA);
|
|
if (m == NULL)
|
|
return (ENOBUFS);
|
|
|
|
MCLGET(m, M_DONTWAIT);
|
|
if ((m->m_flags & M_EXT) == 0) {
|
|
m_freem(m);
|
|
return (ENOBUFS);
|
|
}
|
|
|
|
if (ds->ds_mbuf != NULL)
|
|
bus_dmamap_unload(sc->sc_dmat, ds->ds_dmamap);
|
|
|
|
ds->ds_mbuf = m;
|
|
|
|
error = bus_dmamap_load(sc->sc_dmat, ds->ds_dmamap,
|
|
m->m_ext.ext_buf, m->m_ext.ext_size, NULL,
|
|
BUS_DMA_READ|BUS_DMA_NOWAIT);
|
|
if (error) {
|
|
printf("%s: can't load rx DMA map %d, error = %d\n",
|
|
sc->sc_dev.dv_xname, idx, error);
|
|
panic("sf_add_rxbuf"); /* XXX */
|
|
}
|
|
|
|
bus_dmamap_sync(sc->sc_dmat, ds->ds_dmamap, 0,
|
|
ds->ds_dmamap->dm_mapsize, BUS_DMASYNC_PREREAD);
|
|
|
|
SF_INIT_RXDESC(sc, idx);
|
|
|
|
return (0);
|
|
}
|
|
|
|
static void
|
|
sf_set_filter_perfect(struct sf_softc *sc, int slot, uint8_t *enaddr)
|
|
{
|
|
uint32_t reg0, reg1, reg2;
|
|
|
|
reg0 = enaddr[5] | (enaddr[4] << 8);
|
|
reg1 = enaddr[3] | (enaddr[2] << 8);
|
|
reg2 = enaddr[1] | (enaddr[0] << 8);
|
|
|
|
sf_genreg_write(sc, SF_PERFECT_BASE + (slot * 0x10) + 0, reg0);
|
|
sf_genreg_write(sc, SF_PERFECT_BASE + (slot * 0x10) + 4, reg1);
|
|
sf_genreg_write(sc, SF_PERFECT_BASE + (slot * 0x10) + 8, reg2);
|
|
}
|
|
|
|
static void
|
|
sf_set_filter_hash(struct sf_softc *sc, uint8_t *enaddr)
|
|
{
|
|
uint32_t hash, slot, reg;
|
|
|
|
hash = ether_crc32_be(enaddr, ETHER_ADDR_LEN) >> 23;
|
|
slot = hash >> 4;
|
|
|
|
reg = sf_genreg_read(sc, SF_HASH_BASE + (slot * 0x10));
|
|
reg |= 1 << (hash & 0xf);
|
|
sf_genreg_write(sc, SF_HASH_BASE + (slot * 0x10), reg);
|
|
}
|
|
|
|
/*
|
|
* sf_set_filter:
|
|
*
|
|
* Set the Starfire receive filter.
|
|
*/
|
|
static void
|
|
sf_set_filter(struct sf_softc *sc)
|
|
{
|
|
struct ethercom *ec = &sc->sc_ethercom;
|
|
struct ifnet *ifp = &sc->sc_ethercom.ec_if;
|
|
struct ether_multi *enm;
|
|
struct ether_multistep step;
|
|
int i;
|
|
|
|
/* Start by clearing the perfect and hash tables. */
|
|
for (i = 0; i < SF_PERFECT_SIZE; i += sizeof(uint32_t))
|
|
sf_genreg_write(sc, SF_PERFECT_BASE + i, 0);
|
|
|
|
for (i = 0; i < SF_HASH_SIZE; i += sizeof(uint32_t))
|
|
sf_genreg_write(sc, SF_HASH_BASE + i, 0);
|
|
|
|
/*
|
|
* Clear the perfect and hash mode bits.
|
|
*/
|
|
sc->sc_RxAddressFilteringCtl &=
|
|
~(RAFC_PerfectFilteringMode(3) | RAFC_HashFilteringMode(3));
|
|
|
|
if (ifp->if_flags & IFF_BROADCAST)
|
|
sc->sc_RxAddressFilteringCtl |= RAFC_PassBroadcast;
|
|
else
|
|
sc->sc_RxAddressFilteringCtl &= ~RAFC_PassBroadcast;
|
|
|
|
if (ifp->if_flags & IFF_PROMISC) {
|
|
sc->sc_RxAddressFilteringCtl |= RAFC_PromiscuousMode;
|
|
goto allmulti;
|
|
} else
|
|
sc->sc_RxAddressFilteringCtl &= ~RAFC_PromiscuousMode;
|
|
|
|
/*
|
|
* Set normal perfect filtering mode.
|
|
*/
|
|
sc->sc_RxAddressFilteringCtl |= RAFC_PerfectFilteringMode(1);
|
|
|
|
/*
|
|
* First, write the station address to the perfect filter
|
|
* table.
|
|
*/
|
|
sf_set_filter_perfect(sc, 0, LLADDR(ifp->if_sadl));
|
|
|
|
/*
|
|
* Now set the hash bits for each multicast address in our
|
|
* list.
|
|
*/
|
|
ETHER_FIRST_MULTI(step, ec, enm);
|
|
if (enm == NULL)
|
|
goto done;
|
|
while (enm != NULL) {
|
|
if (memcmp(enm->enm_addrlo, enm->enm_addrhi, ETHER_ADDR_LEN)) {
|
|
/*
|
|
* We must listen to a range of multicast addresses.
|
|
* For now, just accept all multicasts, rather than
|
|
* trying to set only those filter bits needed to match
|
|
* the range. (At this time, the only use of address
|
|
* ranges is for IP multicast routing, for which the
|
|
* range is big enough to require all bits set.)
|
|
*/
|
|
goto allmulti;
|
|
}
|
|
sf_set_filter_hash(sc, enm->enm_addrlo);
|
|
ETHER_NEXT_MULTI(step, enm);
|
|
}
|
|
|
|
/*
|
|
* Set "hash only multicast dest, match regardless of VLAN ID".
|
|
*/
|
|
sc->sc_RxAddressFilteringCtl |= RAFC_HashFilteringMode(2);
|
|
goto done;
|
|
|
|
allmulti:
|
|
/*
|
|
* XXX RAFC_PassMulticast is sub-optimal if using VLAN mode.
|
|
*/
|
|
sc->sc_RxAddressFilteringCtl |= RAFC_PassMulticast;
|
|
ifp->if_flags |= IFF_ALLMULTI;
|
|
|
|
done:
|
|
sf_funcreg_write(sc, SF_RxAddressFilteringCtl,
|
|
sc->sc_RxAddressFilteringCtl);
|
|
}
|
|
|
|
/*
|
|
* sf_mii_read: [mii interface function]
|
|
*
|
|
* Read from the MII.
|
|
*/
|
|
static int
|
|
sf_mii_read(struct device *self, int phy, int reg)
|
|
{
|
|
struct sf_softc *sc = (void *) self;
|
|
uint32_t v;
|
|
int i;
|
|
|
|
for (i = 0; i < 1000; i++) {
|
|
v = sf_genreg_read(sc, SF_MII_PHY_REG(phy, reg));
|
|
if (v & MiiDataValid)
|
|
break;
|
|
delay(1);
|
|
}
|
|
|
|
if ((v & MiiDataValid) == 0)
|
|
return (0);
|
|
|
|
if (MiiRegDataPort(v) == 0xffff)
|
|
return (0);
|
|
|
|
return (MiiRegDataPort(v));
|
|
}
|
|
|
|
/*
|
|
* sf_mii_write: [mii interface function]
|
|
*
|
|
* Write to the MII.
|
|
*/
|
|
static void
|
|
sf_mii_write(struct device *self, int phy, int reg, int val)
|
|
{
|
|
struct sf_softc *sc = (void *) self;
|
|
int i;
|
|
|
|
sf_genreg_write(sc, SF_MII_PHY_REG(phy, reg), val);
|
|
|
|
for (i = 0; i < 1000; i++) {
|
|
if ((sf_genreg_read(sc, SF_MII_PHY_REG(phy, reg)) &
|
|
MiiBusy) == 0)
|
|
return;
|
|
delay(1);
|
|
}
|
|
|
|
printf("%s: MII write timed out\n", sc->sc_dev.dv_xname);
|
|
}
|
|
|
|
/*
|
|
* sf_mii_statchg: [mii interface function]
|
|
*
|
|
* Callback from the PHY when the media changes.
|
|
*/
|
|
static void
|
|
sf_mii_statchg(struct device *self)
|
|
{
|
|
struct sf_softc *sc = (void *) self;
|
|
uint32_t ipg;
|
|
|
|
if (sc->sc_mii.mii_media_active & IFM_FDX) {
|
|
sc->sc_MacConfig1 |= MC1_FullDuplex;
|
|
ipg = 0x15;
|
|
} else {
|
|
sc->sc_MacConfig1 &= ~MC1_FullDuplex;
|
|
ipg = 0x11;
|
|
}
|
|
|
|
sf_genreg_write(sc, SF_MacConfig1, sc->sc_MacConfig1);
|
|
sf_macreset(sc);
|
|
|
|
sf_genreg_write(sc, SF_BkToBkIPG, ipg);
|
|
}
|
|
|
|
/*
|
|
* sf_mediastatus: [ifmedia interface function]
|
|
*
|
|
* Callback from ifmedia to request current media status.
|
|
*/
|
|
static void
|
|
sf_mediastatus(struct ifnet *ifp, struct ifmediareq *ifmr)
|
|
{
|
|
struct sf_softc *sc = ifp->if_softc;
|
|
|
|
mii_pollstat(&sc->sc_mii);
|
|
ifmr->ifm_status = sc->sc_mii.mii_media_status;
|
|
ifmr->ifm_active = sc->sc_mii.mii_media_active;
|
|
}
|
|
|
|
/*
|
|
* sf_mediachange: [ifmedia interface function]
|
|
*
|
|
* Callback from ifmedia to request new media setting.
|
|
*/
|
|
static int
|
|
sf_mediachange(struct ifnet *ifp)
|
|
{
|
|
struct sf_softc *sc = ifp->if_softc;
|
|
|
|
if (ifp->if_flags & IFF_UP)
|
|
mii_mediachg(&sc->sc_mii);
|
|
return (0);
|
|
}
|