NetBSD/sys/dev/marvell/if_mvgbe.c

2396 lines
64 KiB
C
Raw Normal View History

/* $NetBSD: if_mvgbe.c,v 1.65 2022/05/21 10:27:30 rin Exp $ */
/*
2013-12-23 06:23:25 +04:00
* Copyright (c) 2007, 2008, 2013 KIYOHARA Takashi
* All rights reserved.
*
* Redistribution and use in source and binary forms, with or without
* modification, are permitted provided that the following conditions
* are met:
* 1. Redistributions of source code must retain the above copyright
* notice, this list of conditions and the following disclaimer.
* 2. Redistributions in binary form must reproduce the above copyright
* notice, this list of conditions and the following disclaimer in the
* documentation and/or other materials provided with the distribution.
*
* THIS SOFTWARE IS PROVIDED BY THE AUTHOR ``AS IS'' AND ANY EXPRESS OR
* IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
* WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
* DISCLAIMED. IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR ANY DIRECT,
* INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
* (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR
* SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
* HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT,
* STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN
* ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
* POSSIBILITY OF SUCH DAMAGE.
*/
#include <sys/cdefs.h>
__KERNEL_RCSID(0, "$NetBSD: if_mvgbe.c,v 1.65 2022/05/21 10:27:30 rin Exp $");
2013-12-23 06:23:25 +04:00
#include "opt_multiprocessor.h"
#if defined MULTIPROCESSOR
#warning Queue Management Method 'Counters' not support. Please use mvxpe instead of this.
2013-12-23 06:23:25 +04:00
#endif
#include <sys/param.h>
#include <sys/bus.h>
2012-10-27 01:03:26 +04:00
#include <sys/callout.h>
#include <sys/device.h>
#include <sys/endian.h>
#include <sys/errno.h>
#include <sys/evcnt.h>
2012-10-27 01:03:26 +04:00
#include <sys/kernel.h>
#include <sys/kmem.h>
#include <sys/mutex.h>
#include <sys/sockio.h>
#include <sys/sysctl.h>
#include <dev/marvell/marvellreg.h>
#include <dev/marvell/marvellvar.h>
#include <dev/marvell/mvgbereg.h>
#include <net/if.h>
#include <net/if_ether.h>
#include <net/if_media.h>
#include <netinet/in.h>
#include <netinet/in_systm.h>
#include <netinet/ip.h>
#include <net/bpf.h>
#include <sys/rndsource.h>
#include <dev/mii/mii.h>
#include <dev/mii/miivar.h>
#include "locators.h"
/* #define MVGBE_DEBUG 3 */
#ifdef MVGBE_DEBUG
#define DPRINTF(x) if (mvgbe_debug) printf x
#define DPRINTFN(n, x) if (mvgbe_debug >= (n)) printf x
int mvgbe_debug = MVGBE_DEBUG;
#else
#define DPRINTF(x)
#define DPRINTFN(n, x)
#endif
#define MVGBE_READ(sc, reg) \
bus_space_read_4((sc)->sc_iot, (sc)->sc_ioh, (reg))
#define MVGBE_WRITE(sc, reg, val) \
bus_space_write_4((sc)->sc_iot, (sc)->sc_ioh, (reg), (val))
#define MVGBE_READ_FILTER(sc, reg, val, c) \
bus_space_read_region_4((sc)->sc_iot, (sc)->sc_dafh, (reg), (val), (c))
#define MVGBE_WRITE_FILTER(sc, reg, val, c) \
bus_space_write_region_4((sc)->sc_iot, (sc)->sc_dafh, (reg), (val), (c))
2013-12-23 06:23:25 +04:00
#define MVGBE_LINKUP_READ(sc) \
bus_space_read_4((sc)->sc_iot, (sc)->sc_linkup.ioh, 0)
#define MVGBE_IS_LINKUP(sc) (MVGBE_LINKUP_READ(sc) & (sc)->sc_linkup.bit)
#define MVGBE_TX_RING_CNT 256
#define MVGBE_TX_RING_MSK (MVGBE_TX_RING_CNT - 1)
#define MVGBE_TX_RING_NEXT(x) (((x) + 1) & MVGBE_TX_RING_MSK)
#define MVGBE_RX_RING_CNT 256
#define MVGBE_RX_RING_MSK (MVGBE_RX_RING_CNT - 1)
#define MVGBE_RX_RING_NEXT(x) (((x) + 1) & MVGBE_RX_RING_MSK)
CTASSERT(MVGBE_TX_RING_CNT > 1 && MVGBE_TX_RING_NEXT(MVGBE_TX_RING_CNT) ==
(MVGBE_TX_RING_CNT + 1) % MVGBE_TX_RING_CNT);
CTASSERT(MVGBE_RX_RING_CNT > 1 && MVGBE_RX_RING_NEXT(MVGBE_RX_RING_CNT) ==
(MVGBE_RX_RING_CNT + 1) % MVGBE_RX_RING_CNT);
#define MVGBE_JSLOTS 384 /* XXXX */
#define MVGBE_JLEN \
((MVGBE_MRU + MVGBE_HWHEADER_SIZE + MVGBE_RXBUF_ALIGN - 1) & \
~MVGBE_RXBUF_MASK)
#define MVGBE_NTXSEG 30
#define MVGBE_JPAGESZ PAGE_SIZE
#define MVGBE_RESID \
(MVGBE_JPAGESZ - (MVGBE_JLEN * MVGBE_JSLOTS) % MVGBE_JPAGESZ)
#define MVGBE_JMEM \
((MVGBE_JLEN * MVGBE_JSLOTS) + MVGBE_RESID)
#define MVGBE_TX_RING_ADDR(sc, i) \
((sc)->sc_ring_map->dm_segs[0].ds_addr + \
offsetof(struct mvgbe_ring_data, mvgbe_tx_ring[(i)]))
#define MVGBE_RX_RING_ADDR(sc, i) \
((sc)->sc_ring_map->dm_segs[0].ds_addr + \
offsetof(struct mvgbe_ring_data, mvgbe_rx_ring[(i)]))
#define MVGBE_CDOFF(x) offsetof(struct mvgbe_ring_data, x)
#define MVGBE_CDTXOFF(x) MVGBE_CDOFF(mvgbe_tx_ring[(x)])
#define MVGBE_CDRXOFF(x) MVGBE_CDOFF(mvgbe_rx_ring[(x)])
#define MVGBE_CDTXSYNC(sc, x, n, ops) \
do { \
int __x, __n; \
const int __descsize = sizeof(struct mvgbe_tx_desc); \
\
__x = (x); \
__n = (n); \
\
/* If it will wrap around, sync to the end of the ring. */ \
if ((__x + __n) > MVGBE_TX_RING_CNT) { \
bus_dmamap_sync((sc)->sc_dmat, \
(sc)->sc_ring_map, MVGBE_CDTXOFF(__x), \
__descsize * (MVGBE_TX_RING_CNT - __x), (ops)); \
__n -= (MVGBE_TX_RING_CNT - __x); \
__x = 0; \
} \
\
/* Now sync whatever is left. */ \
bus_dmamap_sync((sc)->sc_dmat, (sc)->sc_ring_map, \
MVGBE_CDTXOFF((__x)), __descsize * __n, (ops)); \
} while (0 /*CONSTCOND*/)
#define MVGBE_CDRXSYNC(sc, x, ops) \
do { \
bus_dmamap_sync((sc)->sc_dmat, (sc)->sc_ring_map, \
MVGBE_CDRXOFF((x)), sizeof(struct mvgbe_rx_desc), (ops)); \
} while (/*CONSTCOND*/0)
#define MVGBE_IPGINTTX_DEFAULT 768
#define MVGBE_IPGINTRX_DEFAULT 768
#ifdef MVGBE_EVENT_COUNTERS
#define MVGBE_EVCNT_INCR(ev) (ev)->ev_count++
#define MVGBE_EVCNT_ADD(ev, val) (ev)->ev_count += (val)
#else
#define MVGBE_EVCNT_INCR(ev) /* nothing */
#define MVGBE_EVCNT_ADD(ev, val) /* nothing */
#endif
struct mvgbe_jpool_entry {
int slot;
LIST_ENTRY(mvgbe_jpool_entry) jpool_entries;
};
struct mvgbe_chain {
void *mvgbe_desc;
struct mbuf *mvgbe_mbuf;
struct mvgbe_chain *mvgbe_next;
};
struct mvgbe_txmap_entry {
bus_dmamap_t dmamap;
SIMPLEQ_ENTRY(mvgbe_txmap_entry) link;
};
struct mvgbe_chain_data {
struct mvgbe_chain mvgbe_tx_chain[MVGBE_TX_RING_CNT];
struct mvgbe_txmap_entry *mvgbe_tx_map[MVGBE_TX_RING_CNT];
int mvgbe_tx_prod;
int mvgbe_tx_cons;
int mvgbe_tx_cnt;
struct mvgbe_chain mvgbe_rx_chain[MVGBE_RX_RING_CNT];
bus_dmamap_t mvgbe_rx_map[MVGBE_RX_RING_CNT];
bus_dmamap_t mvgbe_rx_jumbo_map;
int mvgbe_rx_prod;
int mvgbe_rx_cons;
int mvgbe_rx_cnt;
/* Stick the jumbo mem management stuff here too. */
void *mvgbe_jslots[MVGBE_JSLOTS];
void *mvgbe_jumbo_buf;
};
struct mvgbe_ring_data {
struct mvgbe_tx_desc mvgbe_tx_ring[MVGBE_TX_RING_CNT];
struct mvgbe_rx_desc mvgbe_rx_ring[MVGBE_RX_RING_CNT];
};
struct mvgbec_softc {
device_t sc_dev;
bus_space_tag_t sc_iot;
bus_space_handle_t sc_ioh;
kmutex_t sc_mtx;
2010-10-02 09:57:42 +04:00
int sc_flags;
};
struct mvgbe_softc {
device_t sc_dev;
2010-10-02 09:57:42 +04:00
int sc_port;
2013-12-23 06:23:25 +04:00
uint32_t sc_version;
bus_space_tag_t sc_iot;
bus_space_handle_t sc_ioh;
2012-10-27 01:03:26 +04:00
bus_space_handle_t sc_dafh; /* dest address filter handle */
bus_dma_tag_t sc_dmat;
struct ethercom sc_ethercom;
struct mii_data sc_mii;
uint8_t sc_enaddr[ETHER_ADDR_LEN]; /* station addr */
2012-10-27 01:03:26 +04:00
callout_t sc_tick_ch; /* tick callout */
struct mvgbe_chain_data sc_cdata;
struct mvgbe_ring_data *sc_rdata;
bus_dmamap_t sc_ring_map;
u_short sc_if_flags;
unsigned int sc_ipginttx;
unsigned int sc_ipgintrx;
int sc_wdogsoft;
LIST_HEAD(__mvgbe_jfreehead, mvgbe_jpool_entry) sc_jfree_listhead;
LIST_HEAD(__mvgbe_jinusehead, mvgbe_jpool_entry) sc_jinuse_listhead;
SIMPLEQ_HEAD(__mvgbe_txmaphead, mvgbe_txmap_entry) sc_txmap_head;
2013-12-23 06:23:25 +04:00
struct {
bus_space_handle_t ioh;
uint32_t bit;
} sc_linkup;
uint32_t sc_cmdsts_opts;
First step of random number subsystem rework described in <20111022023242.BA26F14A158@mail.netbsd.org>. This change includes the following: An initial cleanup and minor reorganization of the entropy pool code in sys/dev/rnd.c and sys/dev/rndpool.c. Several bugs are fixed. Some effort is made to accumulate entropy more quickly at boot time. A generic interface, "rndsink", is added, for stream generators to request that they be re-keyed with good quality entropy from the pool as soon as it is available. The arc4random()/arc4randbytes() implementation in libkern is adjusted to use the rndsink interface for rekeying, which helps address the problem of low-quality keys at boot time. An implementation of the FIPS 140-2 statistical tests for random number generator quality is provided (libkern/rngtest.c). This is based on Greg Rose's implementation from Qualcomm. A new random stream generator, nist_ctr_drbg, is provided. It is based on an implementation of the NIST SP800-90 CTR_DRBG by Henric Jungheim. This generator users AES in a modified counter mode to generate a backtracking-resistant random stream. An abstraction layer, "cprng", is provided for in-kernel consumers of randomness. The arc4random/arc4randbytes API is deprecated for in-kernel use. It is replaced by "cprng_strong". The current cprng_fast implementation wraps the existing arc4random implementation. The current cprng_strong implementation wraps the new CTR_DRBG implementation. Both interfaces are rekeyed from the entropy pool automatically at intervals justifiable from best current cryptographic practice. In some quick tests, cprng_fast() is about the same speed as the old arc4randbytes(), and cprng_strong() is about 20% faster than rnd_extract_data(). Performance is expected to improve. The AES code in src/crypto/rijndael is no longer an optional kernel component, as it is required by cprng_strong, which is not an optional kernel component. The entropy pool output is subjected to the rngtest tests at startup time; if it fails, the system will reboot. There is approximately a 3/10000 chance of a false positive from these tests. Entropy pool _input_ from hardware random numbers is subjected to the rngtest tests at attach time, as well as the FIPS continuous-output test, to detect bad or stuck hardware RNGs; if any are detected, they are detached, but the system continues to run. A problem with rndctl(8) is fixed -- datastructures with pointers in arrays are no longer passed to userspace (this was not a security problem, but rather a major issue for compat32). A new kernel will require a new rndctl. The sysctl kern.arandom() and kern.urandom() nodes are hooked up to the new generators, but the /dev/*random pseudodevices are not, yet. Manual pages for the new kernel interfaces are forthcoming.
2011-11-20 02:51:18 +04:00
krndsource_t sc_rnd_source;
struct sysctllog *mvgbe_clog;
#ifdef MVGBE_EVENT_COUNTERS
struct evcnt sc_ev_rxoverrun;
struct evcnt sc_ev_wdogsoft;
#endif
};
/* Gigabit Ethernet Unit Global part functions */
static int mvgbec_match(device_t, struct cfdata *, void *);
static void mvgbec_attach(device_t, device_t, void *);
static int mvgbec_print(void *, const char *);
static int mvgbec_search(device_t, cfdata_t, const int *, void *);
/* MII funcstions */
Change MII PHY read/write API from: int (*mii_readreg_t)(device_t, int, int); void (*mii_writereg_t)(device_t, int, int, int); to: int (*mii_readreg_t)(device_t, int, int, uint16_t *); int (*mii_writereg_t)(device_t, int, int, uint16_t); Now we can test if a read/write operation failed or not by the return value. In 802.3 spec says that the PHY shall not respond to read/write transaction to the unimplemented register(22.2.4.3). Detecting timeout can be used to check whether a register is implemented or not (if the register conforms to the spec). ukphy(4) can be used this for MII_MMDACR and MII_MMDAADR. Note that I noticed that the following code do infinite loop in the read/wirte function. If it accesses unimplemented PHY register, it will hang. It should be fixed: arm/at91/at91emac.c arm/ep93xx/epe.c arm/omap/omapl1x_emac.c mips/ralink/ralink_eth.c arch/powerpc/booke/dev/pq3etsec.c(read) dev/cadence/if_cemac.c <- hkenken dev/ic/lan9118.c Tested with the following device: axe+ukphy axe+rgephy axen+rgephy (tested by Andrius V) wm+atphy wm+ukphy wm+igphy wm+ihphy wm+makphy sk+makphy sk+brgphy sk+gentbi msk+makphy sip+icsphy sip+ukphy re+rgephy bge+brgphy bnx+brgphy gsip+gphyter rtk+rlphy fxp+inphy (tested by Andrius V) tlp+acphy ex+exphy epic+qsphy vge+ciphy (tested by Andrius V) vr+ukphy (tested by Andrius V) vte+ukphy (tested by Andrius V) Not tested (MAC): arm:at91emac arm:cemac arm:epe arm:geminigmac arm:enet arm:cpsw arm:emac(omac) arm:emac(sunxi) arm:npe evbppc:temac macppc:bm macppc:gm mips:aumac mips:ae mips:cnmac mips:reth mips:sbmac playstation2:smap powerpc:tsec powerpc:emac(ibm4xx) sgimips:mec sparc:be sf ne(ax88190, dl10019) awge ep gem hme smsh mtd sm age alc ale bce cas et jme lii nfe pcn ste stge tl xi aue mue smsc udav url Not tested (PHY): amhphy bmtphy dmphy etphy glxtphy ikphy iophy lxtphy nsphyter pnaphy rdcphy sqphy tlphy tqphy urlphy
2019-01-22 06:42:24 +03:00
static int mvgbec_miibus_readreg(device_t, int, int, uint16_t *);
static int mvgbec_miibus_writereg(device_t, int, int, uint16_t);
static void mvgbec_miibus_statchg(struct ifnet *);
2014-03-15 17:33:48 +04:00
static void mvgbec_wininit(struct mvgbec_softc *, enum marvell_tags *);
/* Gigabit Ethernet Port part functions */
static int mvgbe_match(device_t, struct cfdata *, void *);
static void mvgbe_attach(device_t, device_t, void *);
2012-10-27 01:03:26 +04:00
static void mvgbe_tick(void *);
static int mvgbe_intr(void *);
static void mvgbe_start(struct ifnet *);
static int mvgbe_ioctl(struct ifnet *, u_long, void *);
static int mvgbe_init(struct ifnet *);
static void mvgbe_stop(struct ifnet *, int);
static void mvgbe_watchdog(struct ifnet *);
static int mvgbe_ifflags_cb(struct ethercom *);
static int mvgbe_mediachange(struct ifnet *);
static void mvgbe_mediastatus(struct ifnet *, struct ifmediareq *);
static int mvgbe_init_rx_ring(struct mvgbe_softc *);
static int mvgbe_init_tx_ring(struct mvgbe_softc *);
static int mvgbe_newbuf(struct mvgbe_softc *, int, struct mbuf *, bus_dmamap_t);
static int mvgbe_alloc_jumbo_mem(struct mvgbe_softc *);
static void *mvgbe_jalloc(struct mvgbe_softc *);
static void mvgbe_jfree(struct mbuf *, void *, size_t, void *);
static int mvgbe_encap(struct mvgbe_softc *, struct mbuf *, uint32_t *);
static void mvgbe_rxeof(struct mvgbe_softc *);
static void mvgbe_txeof(struct mvgbe_softc *);
static uint8_t mvgbe_crc8(const uint8_t *, size_t);
static void mvgbe_filter_setup(struct mvgbe_softc *);
#ifdef MVGBE_DEBUG
static void mvgbe_dump_txdesc(struct mvgbe_tx_desc *, int);
#endif
static int mvgbe_ipginttx(struct mvgbec_softc *, struct mvgbe_softc *,
unsigned int);
static int mvgbe_ipgintrx(struct mvgbec_softc *, struct mvgbe_softc *,
unsigned int);
static void sysctl_mvgbe_init(struct mvgbe_softc *);
static int mvgbe_sysctl_ipginttx(SYSCTLFN_PROTO);
static int mvgbe_sysctl_ipgintrx(SYSCTLFN_PROTO);
CFATTACH_DECL_NEW(mvgbec_gt, sizeof(struct mvgbec_softc),
mvgbec_match, mvgbec_attach, NULL, NULL);
CFATTACH_DECL_NEW(mvgbec_mbus, sizeof(struct mvgbec_softc),
mvgbec_match, mvgbec_attach, NULL, NULL);
CFATTACH_DECL_NEW(mvgbe, sizeof(struct mvgbe_softc),
mvgbe_match, mvgbe_attach, NULL, NULL);
device_t mvgbec0 = NULL;
static int mvgbe_root_num;
struct mvgbe_port {
int model;
2010-10-02 09:57:42 +04:00
int unit;
int ports;
2010-10-02 09:57:42 +04:00
int irqs[3];
int flags;
#define FLAGS_FIX_TQTB (1 << 0)
#define FLAGS_FIX_MTU (1 << 1)
#define FLAGS_IPG1 (1 << 2)
#define FLAGS_IPG2 (1 << 3)
2013-12-23 06:23:25 +04:00
#define FLAGS_HAS_PV (1 << 4) /* Has Port Version Register */
} mvgbe_ports[] = {
2010-10-02 09:57:42 +04:00
{ MARVELL_DISCOVERY_II, 0, 3, { 32, 33, 34 }, 0 },
{ MARVELL_DISCOVERY_III, 0, 3, { 32, 33, 34 }, 0 },
#if 0
2010-10-02 09:57:42 +04:00
{ MARVELL_DISCOVERY_LT, 0, ?, { }, 0 },
{ MARVELL_DISCOVERY_V, 0, ?, { }, 0 },
{ MARVELL_DISCOVERY_VI, 0, ?, { }, 0 },
#endif
{ MARVELL_ORION_1_88F5082, 0, 1, { 21 }, FLAGS_FIX_MTU },
{ MARVELL_ORION_1_88F5180N, 0, 1, { 21 }, FLAGS_FIX_MTU },
{ MARVELL_ORION_1_88F5181, 0, 1, { 21 }, FLAGS_FIX_MTU | FLAGS_IPG1 },
{ MARVELL_ORION_1_88F5182, 0, 1, { 21 }, FLAGS_FIX_MTU | FLAGS_IPG1 },
{ MARVELL_ORION_2_88F5281, 0, 1, { 21 }, FLAGS_FIX_MTU | FLAGS_IPG1 },
{ MARVELL_ORION_1_88F6082, 0, 1, { 21 }, FLAGS_FIX_MTU },
{ MARVELL_ORION_1_88W8660, 0, 1, { 21 }, FLAGS_FIX_MTU },
2010-10-02 09:57:42 +04:00
{ MARVELL_KIRKWOOD_88F6180, 0, 1, { 11 }, FLAGS_FIX_TQTB | FLAGS_IPG2 },
{ MARVELL_KIRKWOOD_88F6192, 0, 1, { 11 }, FLAGS_FIX_TQTB | FLAGS_IPG2 },
{ MARVELL_KIRKWOOD_88F6192, 1, 1, { 15 }, FLAGS_FIX_TQTB | FLAGS_IPG2 },
{ MARVELL_KIRKWOOD_88F6281, 0, 1, { 11 }, FLAGS_FIX_TQTB | FLAGS_IPG2 },
{ MARVELL_KIRKWOOD_88F6281, 1, 1, { 15 }, FLAGS_FIX_TQTB | FLAGS_IPG2 },
{ MARVELL_KIRKWOOD_88F6282, 0, 1, { 11 }, FLAGS_FIX_TQTB | FLAGS_IPG2 },
{ MARVELL_KIRKWOOD_88F6282, 1, 1, { 15 }, FLAGS_FIX_TQTB | FLAGS_IPG2 },
{ MARVELL_MV78XX0_MV78100, 0, 1, { 40 }, FLAGS_FIX_TQTB | FLAGS_IPG2 },
{ MARVELL_MV78XX0_MV78100, 1, 1, { 44 }, FLAGS_FIX_TQTB | FLAGS_IPG2 },
{ MARVELL_MV78XX0_MV78200, 0, 1, { 40 }, FLAGS_FIX_TQTB | FLAGS_IPG2 },
{ MARVELL_MV78XX0_MV78200, 1, 1, { 44 }, FLAGS_FIX_TQTB | FLAGS_IPG2 },
{ MARVELL_MV78XX0_MV78200, 2, 1, { 48 }, FLAGS_FIX_TQTB | FLAGS_IPG2 },
{ MARVELL_MV78XX0_MV78200, 3, 1, { 52 }, FLAGS_FIX_TQTB | FLAGS_IPG2 },
2013-12-23 06:23:25 +04:00
2017-01-07 17:24:27 +03:00
{ MARVELL_DOVE_88AP510, 0, 1, { 29 }, FLAGS_FIX_TQTB | FLAGS_IPG2 },
2013-12-23 06:23:25 +04:00
{ MARVELL_ARMADAXP_MV78130, 0, 1, { 66 }, FLAGS_HAS_PV },
{ MARVELL_ARMADAXP_MV78130, 1, 1, { 70 }, FLAGS_HAS_PV },
{ MARVELL_ARMADAXP_MV78130, 2, 1, { 74 }, FLAGS_HAS_PV },
{ MARVELL_ARMADAXP_MV78160, 0, 1, { 66 }, FLAGS_HAS_PV },
{ MARVELL_ARMADAXP_MV78160, 1, 1, { 70 }, FLAGS_HAS_PV },
{ MARVELL_ARMADAXP_MV78160, 2, 1, { 74 }, FLAGS_HAS_PV },
{ MARVELL_ARMADAXP_MV78160, 3, 1, { 78 }, FLAGS_HAS_PV },
{ MARVELL_ARMADAXP_MV78230, 0, 1, { 66 }, FLAGS_HAS_PV },
{ MARVELL_ARMADAXP_MV78230, 1, 1, { 70 }, FLAGS_HAS_PV },
{ MARVELL_ARMADAXP_MV78230, 2, 1, { 74 }, FLAGS_HAS_PV },
{ MARVELL_ARMADAXP_MV78260, 0, 1, { 66 }, FLAGS_HAS_PV },
{ MARVELL_ARMADAXP_MV78260, 1, 1, { 70 }, FLAGS_HAS_PV },
{ MARVELL_ARMADAXP_MV78260, 2, 1, { 74 }, FLAGS_HAS_PV },
{ MARVELL_ARMADAXP_MV78260, 3, 1, { 78 }, FLAGS_HAS_PV },
{ MARVELL_ARMADAXP_MV78460, 0, 1, { 66 }, FLAGS_HAS_PV },
{ MARVELL_ARMADAXP_MV78460, 1, 1, { 70 }, FLAGS_HAS_PV },
{ MARVELL_ARMADAXP_MV78460, 2, 1, { 74 }, FLAGS_HAS_PV },
{ MARVELL_ARMADAXP_MV78460, 3, 1, { 78 }, FLAGS_HAS_PV },
2014-03-15 17:33:48 +04:00
{ MARVELL_ARMADA370_MV6707, 0, 1, { 66 }, FLAGS_HAS_PV },
{ MARVELL_ARMADA370_MV6707, 1, 1, { 70 }, FLAGS_HAS_PV },
{ MARVELL_ARMADA370_MV6710, 0, 1, { 66 }, FLAGS_HAS_PV },
{ MARVELL_ARMADA370_MV6710, 1, 1, { 70 }, FLAGS_HAS_PV },
{ MARVELL_ARMADA370_MV6W11, 0, 1, { 66 }, FLAGS_HAS_PV },
{ MARVELL_ARMADA370_MV6W11, 1, 1, { 70 }, FLAGS_HAS_PV },
};
/* ARGSUSED */
static int
2011-07-28 19:36:47 +04:00
mvgbec_match(device_t parent, cfdata_t match, void *aux)
{
struct marvell_attach_args *mva = aux;
int i;
if (strcmp(mva->mva_name, match->cf_name) != 0)
return 0;
if (mva->mva_offset == MVA_OFFSET_DEFAULT)
return 0;
for (i = 0; i < __arraycount(mvgbe_ports); i++)
if (mva->mva_model == mvgbe_ports[i].model) {
mva->mva_size = MVGBE_SIZE;
return 1;
}
return 0;
}
/* ARGSUSED */
static void
mvgbec_attach(device_t parent, device_t self, void *aux)
{
struct mvgbec_softc *csc = device_private(self);
struct marvell_attach_args *mva = aux, gbea;
struct mvgbe_softc *port;
struct mii_softc *mii;
device_t child;
uint32_t phyaddr;
int i, j;
aprint_naive("\n");
aprint_normal(": Marvell Gigabit Ethernet Controller\n");
csc->sc_dev = self;
csc->sc_iot = mva->mva_iot;
if (bus_space_subregion(mva->mva_iot, mva->mva_ioh, mva->mva_offset,
mva->mva_size, &csc->sc_ioh)) {
aprint_error_dev(self, "Cannot map registers\n");
return;
}
if (mvgbec0 == NULL)
mvgbec0 = self;
phyaddr = 0;
MVGBE_WRITE(csc, MVGBE_PHYADDR, phyaddr);
mutex_init(&csc->sc_mtx, MUTEX_DEFAULT, IPL_NET);
/* Disable and clear Gigabit Ethernet Unit interrupts */
MVGBE_WRITE(csc, MVGBE_EUIM, 0);
MVGBE_WRITE(csc, MVGBE_EUIC, 0);
2014-03-15 17:33:48 +04:00
mvgbec_wininit(csc, mva->mva_tags);
memset(&gbea, 0, sizeof(gbea));
for (i = 0; i < __arraycount(mvgbe_ports); i++) {
2010-10-02 09:57:42 +04:00
if (mvgbe_ports[i].model != mva->mva_model ||
mvgbe_ports[i].unit != mva->mva_unit)
continue;
csc->sc_flags = mvgbe_ports[i].flags;
2010-10-02 09:57:42 +04:00
for (j = 0; j < mvgbe_ports[i].ports; j++) {
gbea.mva_name = "mvgbe";
gbea.mva_model = mva->mva_model;
gbea.mva_iot = csc->sc_iot;
gbea.mva_ioh = csc->sc_ioh;
gbea.mva_unit = j;
gbea.mva_dmat = mva->mva_dmat;
gbea.mva_irq = mvgbe_ports[i].irqs[j];
Merge thorpej-cfargs branch: Simplify and make extensible the config_search() / config_found() / config_attach() interfaces: rather than having different variants for which arguments you want pass along, just have a single call that takes a variadic list of tag-value arguments. Adjust all call sites: - Simplify wherever possible; don't pass along arguments that aren't actually needed. - Don't be explicit about what interface attribute is attaching if the device only has one. (More simplification.) - Add a config_probe() function to be used in indirect configuiration situations, making is visibly easier to see when indirect config is in play, and allowing for future change in semantics. (As of now, this is just a wrapper around config_match(), but that is an implementation detail.) Remove unnecessary or redundant interface attributes where they're not needed. There are currently 5 "cfargs" defined: - CFARG_SUBMATCH (submatch function for direct config) - CFARG_SEARCH (search function for indirect config) - CFARG_IATTR (interface attribte) - CFARG_LOCATORS (locators array) - CFARG_DEVHANDLE (devhandle_t - wraps OFW, ACPI, etc. handles) ...and a sentinel value CFARG_EOL. Add some extra sanity checking to ensure that interface attributes aren't ambiguous. Use CFARG_DEVHANDLE in MI FDT, OFW, and ACPI code, and macppc and shark ports to associate those device handles with device_t instance. This will trickle trough to more places over time (need back-end for pre-OFW Sun OBP; any others?).
2021-04-25 02:36:23 +03:00
child = config_found(csc->sc_dev, &gbea, mvgbec_print,
2021-08-07 19:18:40 +03:00
CFARGS(.submatch = mvgbec_search));
if (child) {
port = device_private(child);
mii = LIST_FIRST(&port->sc_mii.mii_phys);
if (mii != NULL)
phyaddr |= MVGBE_PHYADDR_PHYAD(j,
mii->mii_phy);
}
}
2010-10-02 09:57:42 +04:00
break;
}
MVGBE_WRITE(csc, MVGBE_PHYADDR, phyaddr);
}
static int
mvgbec_print(void *aux, const char *pnp)
{
struct marvell_attach_args *gbea = aux;
if (pnp)
aprint_normal("%s at %s port %d",
gbea->mva_name, pnp, gbea->mva_unit);
else {
if (gbea->mva_unit != MVGBECCF_PORT_DEFAULT)
aprint_normal(" port %d", gbea->mva_unit);
if (gbea->mva_irq != MVGBECCF_IRQ_DEFAULT)
aprint_normal(" irq %d", gbea->mva_irq);
}
return UNCONF;
}
/* ARGSUSED */
static int
mvgbec_search(device_t parent, cfdata_t cf, const int *ldesc, void *aux)
{
struct marvell_attach_args *gbea = aux;
if (cf->cf_loc[MVGBECCF_PORT] == gbea->mva_unit &&
cf->cf_loc[MVGBECCF_IRQ] != MVGBECCF_IRQ_DEFAULT)
gbea->mva_irq = cf->cf_loc[MVGBECCF_IRQ];
return config_match(parent, cf, aux);
}
static int
Change MII PHY read/write API from: int (*mii_readreg_t)(device_t, int, int); void (*mii_writereg_t)(device_t, int, int, int); to: int (*mii_readreg_t)(device_t, int, int, uint16_t *); int (*mii_writereg_t)(device_t, int, int, uint16_t); Now we can test if a read/write operation failed or not by the return value. In 802.3 spec says that the PHY shall not respond to read/write transaction to the unimplemented register(22.2.4.3). Detecting timeout can be used to check whether a register is implemented or not (if the register conforms to the spec). ukphy(4) can be used this for MII_MMDACR and MII_MMDAADR. Note that I noticed that the following code do infinite loop in the read/wirte function. If it accesses unimplemented PHY register, it will hang. It should be fixed: arm/at91/at91emac.c arm/ep93xx/epe.c arm/omap/omapl1x_emac.c mips/ralink/ralink_eth.c arch/powerpc/booke/dev/pq3etsec.c(read) dev/cadence/if_cemac.c <- hkenken dev/ic/lan9118.c Tested with the following device: axe+ukphy axe+rgephy axen+rgephy (tested by Andrius V) wm+atphy wm+ukphy wm+igphy wm+ihphy wm+makphy sk+makphy sk+brgphy sk+gentbi msk+makphy sip+icsphy sip+ukphy re+rgephy bge+brgphy bnx+brgphy gsip+gphyter rtk+rlphy fxp+inphy (tested by Andrius V) tlp+acphy ex+exphy epic+qsphy vge+ciphy (tested by Andrius V) vr+ukphy (tested by Andrius V) vte+ukphy (tested by Andrius V) Not tested (MAC): arm:at91emac arm:cemac arm:epe arm:geminigmac arm:enet arm:cpsw arm:emac(omac) arm:emac(sunxi) arm:npe evbppc:temac macppc:bm macppc:gm mips:aumac mips:ae mips:cnmac mips:reth mips:sbmac playstation2:smap powerpc:tsec powerpc:emac(ibm4xx) sgimips:mec sparc:be sf ne(ax88190, dl10019) awge ep gem hme smsh mtd sm age alc ale bce cas et jme lii nfe pcn ste stge tl xi aue mue smsc udav url Not tested (PHY): amhphy bmtphy dmphy etphy glxtphy ikphy iophy lxtphy nsphyter pnaphy rdcphy sqphy tlphy tqphy urlphy
2019-01-22 06:42:24 +03:00
mvgbec_miibus_readreg(device_t dev, int phy, int reg, uint16_t *val)
{
struct mvgbe_softc *sc = device_private(dev);
struct mvgbec_softc *csc;
struct ifnet *ifp = &sc->sc_ethercom.ec_if;
Change MII PHY read/write API from: int (*mii_readreg_t)(device_t, int, int); void (*mii_writereg_t)(device_t, int, int, int); to: int (*mii_readreg_t)(device_t, int, int, uint16_t *); int (*mii_writereg_t)(device_t, int, int, uint16_t); Now we can test if a read/write operation failed or not by the return value. In 802.3 spec says that the PHY shall not respond to read/write transaction to the unimplemented register(22.2.4.3). Detecting timeout can be used to check whether a register is implemented or not (if the register conforms to the spec). ukphy(4) can be used this for MII_MMDACR and MII_MMDAADR. Note that I noticed that the following code do infinite loop in the read/wirte function. If it accesses unimplemented PHY register, it will hang. It should be fixed: arm/at91/at91emac.c arm/ep93xx/epe.c arm/omap/omapl1x_emac.c mips/ralink/ralink_eth.c arch/powerpc/booke/dev/pq3etsec.c(read) dev/cadence/if_cemac.c <- hkenken dev/ic/lan9118.c Tested with the following device: axe+ukphy axe+rgephy axen+rgephy (tested by Andrius V) wm+atphy wm+ukphy wm+igphy wm+ihphy wm+makphy sk+makphy sk+brgphy sk+gentbi msk+makphy sip+icsphy sip+ukphy re+rgephy bge+brgphy bnx+brgphy gsip+gphyter rtk+rlphy fxp+inphy (tested by Andrius V) tlp+acphy ex+exphy epic+qsphy vge+ciphy (tested by Andrius V) vr+ukphy (tested by Andrius V) vte+ukphy (tested by Andrius V) Not tested (MAC): arm:at91emac arm:cemac arm:epe arm:geminigmac arm:enet arm:cpsw arm:emac(omac) arm:emac(sunxi) arm:npe evbppc:temac macppc:bm macppc:gm mips:aumac mips:ae mips:cnmac mips:reth mips:sbmac playstation2:smap powerpc:tsec powerpc:emac(ibm4xx) sgimips:mec sparc:be sf ne(ax88190, dl10019) awge ep gem hme smsh mtd sm age alc ale bce cas et jme lii nfe pcn ste stge tl xi aue mue smsc udav url Not tested (PHY): amhphy bmtphy dmphy etphy glxtphy ikphy iophy lxtphy nsphyter pnaphy rdcphy sqphy tlphy tqphy urlphy
2019-01-22 06:42:24 +03:00
uint32_t smi;
int i, rv = 0;
if (mvgbec0 == NULL) {
aprint_error_ifnet(ifp, "SMI mvgbec0 not found\n");
return -1;
}
csc = device_private(mvgbec0);
mutex_enter(&csc->sc_mtx);
for (i = 0; i < MVGBE_PHY_TIMEOUT; i++) {
DELAY(1);
if (!(MVGBE_READ(csc, MVGBE_SMI) & MVGBE_SMI_BUSY))
break;
}
if (i == MVGBE_PHY_TIMEOUT) {
aprint_error_ifnet(ifp, "SMI busy timeout\n");
Change MII PHY read/write API from: int (*mii_readreg_t)(device_t, int, int); void (*mii_writereg_t)(device_t, int, int, int); to: int (*mii_readreg_t)(device_t, int, int, uint16_t *); int (*mii_writereg_t)(device_t, int, int, uint16_t); Now we can test if a read/write operation failed or not by the return value. In 802.3 spec says that the PHY shall not respond to read/write transaction to the unimplemented register(22.2.4.3). Detecting timeout can be used to check whether a register is implemented or not (if the register conforms to the spec). ukphy(4) can be used this for MII_MMDACR and MII_MMDAADR. Note that I noticed that the following code do infinite loop in the read/wirte function. If it accesses unimplemented PHY register, it will hang. It should be fixed: arm/at91/at91emac.c arm/ep93xx/epe.c arm/omap/omapl1x_emac.c mips/ralink/ralink_eth.c arch/powerpc/booke/dev/pq3etsec.c(read) dev/cadence/if_cemac.c <- hkenken dev/ic/lan9118.c Tested with the following device: axe+ukphy axe+rgephy axen+rgephy (tested by Andrius V) wm+atphy wm+ukphy wm+igphy wm+ihphy wm+makphy sk+makphy sk+brgphy sk+gentbi msk+makphy sip+icsphy sip+ukphy re+rgephy bge+brgphy bnx+brgphy gsip+gphyter rtk+rlphy fxp+inphy (tested by Andrius V) tlp+acphy ex+exphy epic+qsphy vge+ciphy (tested by Andrius V) vr+ukphy (tested by Andrius V) vte+ukphy (tested by Andrius V) Not tested (MAC): arm:at91emac arm:cemac arm:epe arm:geminigmac arm:enet arm:cpsw arm:emac(omac) arm:emac(sunxi) arm:npe evbppc:temac macppc:bm macppc:gm mips:aumac mips:ae mips:cnmac mips:reth mips:sbmac playstation2:smap powerpc:tsec powerpc:emac(ibm4xx) sgimips:mec sparc:be sf ne(ax88190, dl10019) awge ep gem hme smsh mtd sm age alc ale bce cas et jme lii nfe pcn ste stge tl xi aue mue smsc udav url Not tested (PHY): amhphy bmtphy dmphy etphy glxtphy ikphy iophy lxtphy nsphyter pnaphy rdcphy sqphy tlphy tqphy urlphy
2019-01-22 06:42:24 +03:00
rv = ETIMEDOUT;
goto out;
}
smi =
MVGBE_SMI_PHYAD(phy) | MVGBE_SMI_REGAD(reg) | MVGBE_SMI_OPCODE_READ;
MVGBE_WRITE(csc, MVGBE_SMI, smi);
for (i = 0; i < MVGBE_PHY_TIMEOUT; i++) {
DELAY(1);
smi = MVGBE_READ(csc, MVGBE_SMI);
Change MII PHY read/write API from: int (*mii_readreg_t)(device_t, int, int); void (*mii_writereg_t)(device_t, int, int, int); to: int (*mii_readreg_t)(device_t, int, int, uint16_t *); int (*mii_writereg_t)(device_t, int, int, uint16_t); Now we can test if a read/write operation failed or not by the return value. In 802.3 spec says that the PHY shall not respond to read/write transaction to the unimplemented register(22.2.4.3). Detecting timeout can be used to check whether a register is implemented or not (if the register conforms to the spec). ukphy(4) can be used this for MII_MMDACR and MII_MMDAADR. Note that I noticed that the following code do infinite loop in the read/wirte function. If it accesses unimplemented PHY register, it will hang. It should be fixed: arm/at91/at91emac.c arm/ep93xx/epe.c arm/omap/omapl1x_emac.c mips/ralink/ralink_eth.c arch/powerpc/booke/dev/pq3etsec.c(read) dev/cadence/if_cemac.c <- hkenken dev/ic/lan9118.c Tested with the following device: axe+ukphy axe+rgephy axen+rgephy (tested by Andrius V) wm+atphy wm+ukphy wm+igphy wm+ihphy wm+makphy sk+makphy sk+brgphy sk+gentbi msk+makphy sip+icsphy sip+ukphy re+rgephy bge+brgphy bnx+brgphy gsip+gphyter rtk+rlphy fxp+inphy (tested by Andrius V) tlp+acphy ex+exphy epic+qsphy vge+ciphy (tested by Andrius V) vr+ukphy (tested by Andrius V) vte+ukphy (tested by Andrius V) Not tested (MAC): arm:at91emac arm:cemac arm:epe arm:geminigmac arm:enet arm:cpsw arm:emac(omac) arm:emac(sunxi) arm:npe evbppc:temac macppc:bm macppc:gm mips:aumac mips:ae mips:cnmac mips:reth mips:sbmac playstation2:smap powerpc:tsec powerpc:emac(ibm4xx) sgimips:mec sparc:be sf ne(ax88190, dl10019) awge ep gem hme smsh mtd sm age alc ale bce cas et jme lii nfe pcn ste stge tl xi aue mue smsc udav url Not tested (PHY): amhphy bmtphy dmphy etphy glxtphy ikphy iophy lxtphy nsphyter pnaphy rdcphy sqphy tlphy tqphy urlphy
2019-01-22 06:42:24 +03:00
if (smi & MVGBE_SMI_READVALID) {
*val = smi & MVGBE_SMI_DATA_MASK;
break;
Change MII PHY read/write API from: int (*mii_readreg_t)(device_t, int, int); void (*mii_writereg_t)(device_t, int, int, int); to: int (*mii_readreg_t)(device_t, int, int, uint16_t *); int (*mii_writereg_t)(device_t, int, int, uint16_t); Now we can test if a read/write operation failed or not by the return value. In 802.3 spec says that the PHY shall not respond to read/write transaction to the unimplemented register(22.2.4.3). Detecting timeout can be used to check whether a register is implemented or not (if the register conforms to the spec). ukphy(4) can be used this for MII_MMDACR and MII_MMDAADR. Note that I noticed that the following code do infinite loop in the read/wirte function. If it accesses unimplemented PHY register, it will hang. It should be fixed: arm/at91/at91emac.c arm/ep93xx/epe.c arm/omap/omapl1x_emac.c mips/ralink/ralink_eth.c arch/powerpc/booke/dev/pq3etsec.c(read) dev/cadence/if_cemac.c <- hkenken dev/ic/lan9118.c Tested with the following device: axe+ukphy axe+rgephy axen+rgephy (tested by Andrius V) wm+atphy wm+ukphy wm+igphy wm+ihphy wm+makphy sk+makphy sk+brgphy sk+gentbi msk+makphy sip+icsphy sip+ukphy re+rgephy bge+brgphy bnx+brgphy gsip+gphyter rtk+rlphy fxp+inphy (tested by Andrius V) tlp+acphy ex+exphy epic+qsphy vge+ciphy (tested by Andrius V) vr+ukphy (tested by Andrius V) vte+ukphy (tested by Andrius V) Not tested (MAC): arm:at91emac arm:cemac arm:epe arm:geminigmac arm:enet arm:cpsw arm:emac(omac) arm:emac(sunxi) arm:npe evbppc:temac macppc:bm macppc:gm mips:aumac mips:ae mips:cnmac mips:reth mips:sbmac playstation2:smap powerpc:tsec powerpc:emac(ibm4xx) sgimips:mec sparc:be sf ne(ax88190, dl10019) awge ep gem hme smsh mtd sm age alc ale bce cas et jme lii nfe pcn ste stge tl xi aue mue smsc udav url Not tested (PHY): amhphy bmtphy dmphy etphy glxtphy ikphy iophy lxtphy nsphyter pnaphy rdcphy sqphy tlphy tqphy urlphy
2019-01-22 06:42:24 +03:00
}
}
DPRINTFN(9, ("mvgbec_miibus_readreg: i=%d, timeout=%d\n",
i, MVGBE_PHY_TIMEOUT));
Change MII PHY read/write API from: int (*mii_readreg_t)(device_t, int, int); void (*mii_writereg_t)(device_t, int, int, int); to: int (*mii_readreg_t)(device_t, int, int, uint16_t *); int (*mii_writereg_t)(device_t, int, int, uint16_t); Now we can test if a read/write operation failed or not by the return value. In 802.3 spec says that the PHY shall not respond to read/write transaction to the unimplemented register(22.2.4.3). Detecting timeout can be used to check whether a register is implemented or not (if the register conforms to the spec). ukphy(4) can be used this for MII_MMDACR and MII_MMDAADR. Note that I noticed that the following code do infinite loop in the read/wirte function. If it accesses unimplemented PHY register, it will hang. It should be fixed: arm/at91/at91emac.c arm/ep93xx/epe.c arm/omap/omapl1x_emac.c mips/ralink/ralink_eth.c arch/powerpc/booke/dev/pq3etsec.c(read) dev/cadence/if_cemac.c <- hkenken dev/ic/lan9118.c Tested with the following device: axe+ukphy axe+rgephy axen+rgephy (tested by Andrius V) wm+atphy wm+ukphy wm+igphy wm+ihphy wm+makphy sk+makphy sk+brgphy sk+gentbi msk+makphy sip+icsphy sip+ukphy re+rgephy bge+brgphy bnx+brgphy gsip+gphyter rtk+rlphy fxp+inphy (tested by Andrius V) tlp+acphy ex+exphy epic+qsphy vge+ciphy (tested by Andrius V) vr+ukphy (tested by Andrius V) vte+ukphy (tested by Andrius V) Not tested (MAC): arm:at91emac arm:cemac arm:epe arm:geminigmac arm:enet arm:cpsw arm:emac(omac) arm:emac(sunxi) arm:npe evbppc:temac macppc:bm macppc:gm mips:aumac mips:ae mips:cnmac mips:reth mips:sbmac playstation2:smap powerpc:tsec powerpc:emac(ibm4xx) sgimips:mec sparc:be sf ne(ax88190, dl10019) awge ep gem hme smsh mtd sm age alc ale bce cas et jme lii nfe pcn ste stge tl xi aue mue smsc udav url Not tested (PHY): amhphy bmtphy dmphy etphy glxtphy ikphy iophy lxtphy nsphyter pnaphy rdcphy sqphy tlphy tqphy urlphy
2019-01-22 06:42:24 +03:00
if (i >= MVGBE_PHY_TIMEOUT)
rv = ETIMEDOUT;
Change MII PHY read/write API from: int (*mii_readreg_t)(device_t, int, int); void (*mii_writereg_t)(device_t, int, int, int); to: int (*mii_readreg_t)(device_t, int, int, uint16_t *); int (*mii_writereg_t)(device_t, int, int, uint16_t); Now we can test if a read/write operation failed or not by the return value. In 802.3 spec says that the PHY shall not respond to read/write transaction to the unimplemented register(22.2.4.3). Detecting timeout can be used to check whether a register is implemented or not (if the register conforms to the spec). ukphy(4) can be used this for MII_MMDACR and MII_MMDAADR. Note that I noticed that the following code do infinite loop in the read/wirte function. If it accesses unimplemented PHY register, it will hang. It should be fixed: arm/at91/at91emac.c arm/ep93xx/epe.c arm/omap/omapl1x_emac.c mips/ralink/ralink_eth.c arch/powerpc/booke/dev/pq3etsec.c(read) dev/cadence/if_cemac.c <- hkenken dev/ic/lan9118.c Tested with the following device: axe+ukphy axe+rgephy axen+rgephy (tested by Andrius V) wm+atphy wm+ukphy wm+igphy wm+ihphy wm+makphy sk+makphy sk+brgphy sk+gentbi msk+makphy sip+icsphy sip+ukphy re+rgephy bge+brgphy bnx+brgphy gsip+gphyter rtk+rlphy fxp+inphy (tested by Andrius V) tlp+acphy ex+exphy epic+qsphy vge+ciphy (tested by Andrius V) vr+ukphy (tested by Andrius V) vte+ukphy (tested by Andrius V) Not tested (MAC): arm:at91emac arm:cemac arm:epe arm:geminigmac arm:enet arm:cpsw arm:emac(omac) arm:emac(sunxi) arm:npe evbppc:temac macppc:bm macppc:gm mips:aumac mips:ae mips:cnmac mips:reth mips:sbmac playstation2:smap powerpc:tsec powerpc:emac(ibm4xx) sgimips:mec sparc:be sf ne(ax88190, dl10019) awge ep gem hme smsh mtd sm age alc ale bce cas et jme lii nfe pcn ste stge tl xi aue mue smsc udav url Not tested (PHY): amhphy bmtphy dmphy etphy glxtphy ikphy iophy lxtphy nsphyter pnaphy rdcphy sqphy tlphy tqphy urlphy
2019-01-22 06:42:24 +03:00
out:
mutex_exit(&csc->sc_mtx);
Change MII PHY read/write API from: int (*mii_readreg_t)(device_t, int, int); void (*mii_writereg_t)(device_t, int, int, int); to: int (*mii_readreg_t)(device_t, int, int, uint16_t *); int (*mii_writereg_t)(device_t, int, int, uint16_t); Now we can test if a read/write operation failed or not by the return value. In 802.3 spec says that the PHY shall not respond to read/write transaction to the unimplemented register(22.2.4.3). Detecting timeout can be used to check whether a register is implemented or not (if the register conforms to the spec). ukphy(4) can be used this for MII_MMDACR and MII_MMDAADR. Note that I noticed that the following code do infinite loop in the read/wirte function. If it accesses unimplemented PHY register, it will hang. It should be fixed: arm/at91/at91emac.c arm/ep93xx/epe.c arm/omap/omapl1x_emac.c mips/ralink/ralink_eth.c arch/powerpc/booke/dev/pq3etsec.c(read) dev/cadence/if_cemac.c <- hkenken dev/ic/lan9118.c Tested with the following device: axe+ukphy axe+rgephy axen+rgephy (tested by Andrius V) wm+atphy wm+ukphy wm+igphy wm+ihphy wm+makphy sk+makphy sk+brgphy sk+gentbi msk+makphy sip+icsphy sip+ukphy re+rgephy bge+brgphy bnx+brgphy gsip+gphyter rtk+rlphy fxp+inphy (tested by Andrius V) tlp+acphy ex+exphy epic+qsphy vge+ciphy (tested by Andrius V) vr+ukphy (tested by Andrius V) vte+ukphy (tested by Andrius V) Not tested (MAC): arm:at91emac arm:cemac arm:epe arm:geminigmac arm:enet arm:cpsw arm:emac(omac) arm:emac(sunxi) arm:npe evbppc:temac macppc:bm macppc:gm mips:aumac mips:ae mips:cnmac mips:reth mips:sbmac playstation2:smap powerpc:tsec powerpc:emac(ibm4xx) sgimips:mec sparc:be sf ne(ax88190, dl10019) awge ep gem hme smsh mtd sm age alc ale bce cas et jme lii nfe pcn ste stge tl xi aue mue smsc udav url Not tested (PHY): amhphy bmtphy dmphy etphy glxtphy ikphy iophy lxtphy nsphyter pnaphy rdcphy sqphy tlphy tqphy urlphy
2019-01-22 06:42:24 +03:00
DPRINTFN(9, ("mvgbec_miibus_readreg phy=%d, reg=%#x, val=%#hx\n",
phy, reg, *val));
Change MII PHY read/write API from: int (*mii_readreg_t)(device_t, int, int); void (*mii_writereg_t)(device_t, int, int, int); to: int (*mii_readreg_t)(device_t, int, int, uint16_t *); int (*mii_writereg_t)(device_t, int, int, uint16_t); Now we can test if a read/write operation failed or not by the return value. In 802.3 spec says that the PHY shall not respond to read/write transaction to the unimplemented register(22.2.4.3). Detecting timeout can be used to check whether a register is implemented or not (if the register conforms to the spec). ukphy(4) can be used this for MII_MMDACR and MII_MMDAADR. Note that I noticed that the following code do infinite loop in the read/wirte function. If it accesses unimplemented PHY register, it will hang. It should be fixed: arm/at91/at91emac.c arm/ep93xx/epe.c arm/omap/omapl1x_emac.c mips/ralink/ralink_eth.c arch/powerpc/booke/dev/pq3etsec.c(read) dev/cadence/if_cemac.c <- hkenken dev/ic/lan9118.c Tested with the following device: axe+ukphy axe+rgephy axen+rgephy (tested by Andrius V) wm+atphy wm+ukphy wm+igphy wm+ihphy wm+makphy sk+makphy sk+brgphy sk+gentbi msk+makphy sip+icsphy sip+ukphy re+rgephy bge+brgphy bnx+brgphy gsip+gphyter rtk+rlphy fxp+inphy (tested by Andrius V) tlp+acphy ex+exphy epic+qsphy vge+ciphy (tested by Andrius V) vr+ukphy (tested by Andrius V) vte+ukphy (tested by Andrius V) Not tested (MAC): arm:at91emac arm:cemac arm:epe arm:geminigmac arm:enet arm:cpsw arm:emac(omac) arm:emac(sunxi) arm:npe evbppc:temac macppc:bm macppc:gm mips:aumac mips:ae mips:cnmac mips:reth mips:sbmac playstation2:smap powerpc:tsec powerpc:emac(ibm4xx) sgimips:mec sparc:be sf ne(ax88190, dl10019) awge ep gem hme smsh mtd sm age alc ale bce cas et jme lii nfe pcn ste stge tl xi aue mue smsc udav url Not tested (PHY): amhphy bmtphy dmphy etphy glxtphy ikphy iophy lxtphy nsphyter pnaphy rdcphy sqphy tlphy tqphy urlphy
2019-01-22 06:42:24 +03:00
return rv;
}
Change MII PHY read/write API from: int (*mii_readreg_t)(device_t, int, int); void (*mii_writereg_t)(device_t, int, int, int); to: int (*mii_readreg_t)(device_t, int, int, uint16_t *); int (*mii_writereg_t)(device_t, int, int, uint16_t); Now we can test if a read/write operation failed or not by the return value. In 802.3 spec says that the PHY shall not respond to read/write transaction to the unimplemented register(22.2.4.3). Detecting timeout can be used to check whether a register is implemented or not (if the register conforms to the spec). ukphy(4) can be used this for MII_MMDACR and MII_MMDAADR. Note that I noticed that the following code do infinite loop in the read/wirte function. If it accesses unimplemented PHY register, it will hang. It should be fixed: arm/at91/at91emac.c arm/ep93xx/epe.c arm/omap/omapl1x_emac.c mips/ralink/ralink_eth.c arch/powerpc/booke/dev/pq3etsec.c(read) dev/cadence/if_cemac.c <- hkenken dev/ic/lan9118.c Tested with the following device: axe+ukphy axe+rgephy axen+rgephy (tested by Andrius V) wm+atphy wm+ukphy wm+igphy wm+ihphy wm+makphy sk+makphy sk+brgphy sk+gentbi msk+makphy sip+icsphy sip+ukphy re+rgephy bge+brgphy bnx+brgphy gsip+gphyter rtk+rlphy fxp+inphy (tested by Andrius V) tlp+acphy ex+exphy epic+qsphy vge+ciphy (tested by Andrius V) vr+ukphy (tested by Andrius V) vte+ukphy (tested by Andrius V) Not tested (MAC): arm:at91emac arm:cemac arm:epe arm:geminigmac arm:enet arm:cpsw arm:emac(omac) arm:emac(sunxi) arm:npe evbppc:temac macppc:bm macppc:gm mips:aumac mips:ae mips:cnmac mips:reth mips:sbmac playstation2:smap powerpc:tsec powerpc:emac(ibm4xx) sgimips:mec sparc:be sf ne(ax88190, dl10019) awge ep gem hme smsh mtd sm age alc ale bce cas et jme lii nfe pcn ste stge tl xi aue mue smsc udav url Not tested (PHY): amhphy bmtphy dmphy etphy glxtphy ikphy iophy lxtphy nsphyter pnaphy rdcphy sqphy tlphy tqphy urlphy
2019-01-22 06:42:24 +03:00
static int
mvgbec_miibus_writereg(device_t dev, int phy, int reg, uint16_t val)
{
struct mvgbe_softc *sc = device_private(dev);
struct mvgbec_softc *csc;
struct ifnet *ifp = &sc->sc_ethercom.ec_if;
uint32_t smi;
Change MII PHY read/write API from: int (*mii_readreg_t)(device_t, int, int); void (*mii_writereg_t)(device_t, int, int, int); to: int (*mii_readreg_t)(device_t, int, int, uint16_t *); int (*mii_writereg_t)(device_t, int, int, uint16_t); Now we can test if a read/write operation failed or not by the return value. In 802.3 spec says that the PHY shall not respond to read/write transaction to the unimplemented register(22.2.4.3). Detecting timeout can be used to check whether a register is implemented or not (if the register conforms to the spec). ukphy(4) can be used this for MII_MMDACR and MII_MMDAADR. Note that I noticed that the following code do infinite loop in the read/wirte function. If it accesses unimplemented PHY register, it will hang. It should be fixed: arm/at91/at91emac.c arm/ep93xx/epe.c arm/omap/omapl1x_emac.c mips/ralink/ralink_eth.c arch/powerpc/booke/dev/pq3etsec.c(read) dev/cadence/if_cemac.c <- hkenken dev/ic/lan9118.c Tested with the following device: axe+ukphy axe+rgephy axen+rgephy (tested by Andrius V) wm+atphy wm+ukphy wm+igphy wm+ihphy wm+makphy sk+makphy sk+brgphy sk+gentbi msk+makphy sip+icsphy sip+ukphy re+rgephy bge+brgphy bnx+brgphy gsip+gphyter rtk+rlphy fxp+inphy (tested by Andrius V) tlp+acphy ex+exphy epic+qsphy vge+ciphy (tested by Andrius V) vr+ukphy (tested by Andrius V) vte+ukphy (tested by Andrius V) Not tested (MAC): arm:at91emac arm:cemac arm:epe arm:geminigmac arm:enet arm:cpsw arm:emac(omac) arm:emac(sunxi) arm:npe evbppc:temac macppc:bm macppc:gm mips:aumac mips:ae mips:cnmac mips:reth mips:sbmac playstation2:smap powerpc:tsec powerpc:emac(ibm4xx) sgimips:mec sparc:be sf ne(ax88190, dl10019) awge ep gem hme smsh mtd sm age alc ale bce cas et jme lii nfe pcn ste stge tl xi aue mue smsc udav url Not tested (PHY): amhphy bmtphy dmphy etphy glxtphy ikphy iophy lxtphy nsphyter pnaphy rdcphy sqphy tlphy tqphy urlphy
2019-01-22 06:42:24 +03:00
int i, rv = 0;
if (mvgbec0 == NULL) {
aprint_error_ifnet(ifp, "SMI mvgbec0 not found\n");
Change MII PHY read/write API from: int (*mii_readreg_t)(device_t, int, int); void (*mii_writereg_t)(device_t, int, int, int); to: int (*mii_readreg_t)(device_t, int, int, uint16_t *); int (*mii_writereg_t)(device_t, int, int, uint16_t); Now we can test if a read/write operation failed or not by the return value. In 802.3 spec says that the PHY shall not respond to read/write transaction to the unimplemented register(22.2.4.3). Detecting timeout can be used to check whether a register is implemented or not (if the register conforms to the spec). ukphy(4) can be used this for MII_MMDACR and MII_MMDAADR. Note that I noticed that the following code do infinite loop in the read/wirte function. If it accesses unimplemented PHY register, it will hang. It should be fixed: arm/at91/at91emac.c arm/ep93xx/epe.c arm/omap/omapl1x_emac.c mips/ralink/ralink_eth.c arch/powerpc/booke/dev/pq3etsec.c(read) dev/cadence/if_cemac.c <- hkenken dev/ic/lan9118.c Tested with the following device: axe+ukphy axe+rgephy axen+rgephy (tested by Andrius V) wm+atphy wm+ukphy wm+igphy wm+ihphy wm+makphy sk+makphy sk+brgphy sk+gentbi msk+makphy sip+icsphy sip+ukphy re+rgephy bge+brgphy bnx+brgphy gsip+gphyter rtk+rlphy fxp+inphy (tested by Andrius V) tlp+acphy ex+exphy epic+qsphy vge+ciphy (tested by Andrius V) vr+ukphy (tested by Andrius V) vte+ukphy (tested by Andrius V) Not tested (MAC): arm:at91emac arm:cemac arm:epe arm:geminigmac arm:enet arm:cpsw arm:emac(omac) arm:emac(sunxi) arm:npe evbppc:temac macppc:bm macppc:gm mips:aumac mips:ae mips:cnmac mips:reth mips:sbmac playstation2:smap powerpc:tsec powerpc:emac(ibm4xx) sgimips:mec sparc:be sf ne(ax88190, dl10019) awge ep gem hme smsh mtd sm age alc ale bce cas et jme lii nfe pcn ste stge tl xi aue mue smsc udav url Not tested (PHY): amhphy bmtphy dmphy etphy glxtphy ikphy iophy lxtphy nsphyter pnaphy rdcphy sqphy tlphy tqphy urlphy
2019-01-22 06:42:24 +03:00
return -1;
}
csc = device_private(mvgbec0);
DPRINTFN(9, ("mvgbec_miibus_writereg phy=%d reg=%#x val=%#x\n",
phy, reg, val));
mutex_enter(&csc->sc_mtx);
for (i = 0; i < MVGBE_PHY_TIMEOUT; i++) {
DELAY(1);
if (!(MVGBE_READ(csc, MVGBE_SMI) & MVGBE_SMI_BUSY))
break;
}
if (i == MVGBE_PHY_TIMEOUT) {
aprint_error_ifnet(ifp, "SMI busy timeout\n");
Change MII PHY read/write API from: int (*mii_readreg_t)(device_t, int, int); void (*mii_writereg_t)(device_t, int, int, int); to: int (*mii_readreg_t)(device_t, int, int, uint16_t *); int (*mii_writereg_t)(device_t, int, int, uint16_t); Now we can test if a read/write operation failed or not by the return value. In 802.3 spec says that the PHY shall not respond to read/write transaction to the unimplemented register(22.2.4.3). Detecting timeout can be used to check whether a register is implemented or not (if the register conforms to the spec). ukphy(4) can be used this for MII_MMDACR and MII_MMDAADR. Note that I noticed that the following code do infinite loop in the read/wirte function. If it accesses unimplemented PHY register, it will hang. It should be fixed: arm/at91/at91emac.c arm/ep93xx/epe.c arm/omap/omapl1x_emac.c mips/ralink/ralink_eth.c arch/powerpc/booke/dev/pq3etsec.c(read) dev/cadence/if_cemac.c <- hkenken dev/ic/lan9118.c Tested with the following device: axe+ukphy axe+rgephy axen+rgephy (tested by Andrius V) wm+atphy wm+ukphy wm+igphy wm+ihphy wm+makphy sk+makphy sk+brgphy sk+gentbi msk+makphy sip+icsphy sip+ukphy re+rgephy bge+brgphy bnx+brgphy gsip+gphyter rtk+rlphy fxp+inphy (tested by Andrius V) tlp+acphy ex+exphy epic+qsphy vge+ciphy (tested by Andrius V) vr+ukphy (tested by Andrius V) vte+ukphy (tested by Andrius V) Not tested (MAC): arm:at91emac arm:cemac arm:epe arm:geminigmac arm:enet arm:cpsw arm:emac(omac) arm:emac(sunxi) arm:npe evbppc:temac macppc:bm macppc:gm mips:aumac mips:ae mips:cnmac mips:reth mips:sbmac playstation2:smap powerpc:tsec powerpc:emac(ibm4xx) sgimips:mec sparc:be sf ne(ax88190, dl10019) awge ep gem hme smsh mtd sm age alc ale bce cas et jme lii nfe pcn ste stge tl xi aue mue smsc udav url Not tested (PHY): amhphy bmtphy dmphy etphy glxtphy ikphy iophy lxtphy nsphyter pnaphy rdcphy sqphy tlphy tqphy urlphy
2019-01-22 06:42:24 +03:00
rv = ETIMEDOUT;
goto out;
}
smi = MVGBE_SMI_PHYAD(phy) | MVGBE_SMI_REGAD(reg) |
MVGBE_SMI_OPCODE_WRITE | (val & MVGBE_SMI_DATA_MASK);
MVGBE_WRITE(csc, MVGBE_SMI, smi);
for (i = 0; i < MVGBE_PHY_TIMEOUT; i++) {
DELAY(1);
if (!(MVGBE_READ(csc, MVGBE_SMI) & MVGBE_SMI_BUSY))
break;
}
Change MII PHY read/write API from: int (*mii_readreg_t)(device_t, int, int); void (*mii_writereg_t)(device_t, int, int, int); to: int (*mii_readreg_t)(device_t, int, int, uint16_t *); int (*mii_writereg_t)(device_t, int, int, uint16_t); Now we can test if a read/write operation failed or not by the return value. In 802.3 spec says that the PHY shall not respond to read/write transaction to the unimplemented register(22.2.4.3). Detecting timeout can be used to check whether a register is implemented or not (if the register conforms to the spec). ukphy(4) can be used this for MII_MMDACR and MII_MMDAADR. Note that I noticed that the following code do infinite loop in the read/wirte function. If it accesses unimplemented PHY register, it will hang. It should be fixed: arm/at91/at91emac.c arm/ep93xx/epe.c arm/omap/omapl1x_emac.c mips/ralink/ralink_eth.c arch/powerpc/booke/dev/pq3etsec.c(read) dev/cadence/if_cemac.c <- hkenken dev/ic/lan9118.c Tested with the following device: axe+ukphy axe+rgephy axen+rgephy (tested by Andrius V) wm+atphy wm+ukphy wm+igphy wm+ihphy wm+makphy sk+makphy sk+brgphy sk+gentbi msk+makphy sip+icsphy sip+ukphy re+rgephy bge+brgphy bnx+brgphy gsip+gphyter rtk+rlphy fxp+inphy (tested by Andrius V) tlp+acphy ex+exphy epic+qsphy vge+ciphy (tested by Andrius V) vr+ukphy (tested by Andrius V) vte+ukphy (tested by Andrius V) Not tested (MAC): arm:at91emac arm:cemac arm:epe arm:geminigmac arm:enet arm:cpsw arm:emac(omac) arm:emac(sunxi) arm:npe evbppc:temac macppc:bm macppc:gm mips:aumac mips:ae mips:cnmac mips:reth mips:sbmac playstation2:smap powerpc:tsec powerpc:emac(ibm4xx) sgimips:mec sparc:be sf ne(ax88190, dl10019) awge ep gem hme smsh mtd sm age alc ale bce cas et jme lii nfe pcn ste stge tl xi aue mue smsc udav url Not tested (PHY): amhphy bmtphy dmphy etphy glxtphy ikphy iophy lxtphy nsphyter pnaphy rdcphy sqphy tlphy tqphy urlphy
2019-01-22 06:42:24 +03:00
out:
mutex_exit(&csc->sc_mtx);
Change MII PHY read/write API from: int (*mii_readreg_t)(device_t, int, int); void (*mii_writereg_t)(device_t, int, int, int); to: int (*mii_readreg_t)(device_t, int, int, uint16_t *); int (*mii_writereg_t)(device_t, int, int, uint16_t); Now we can test if a read/write operation failed or not by the return value. In 802.3 spec says that the PHY shall not respond to read/write transaction to the unimplemented register(22.2.4.3). Detecting timeout can be used to check whether a register is implemented or not (if the register conforms to the spec). ukphy(4) can be used this for MII_MMDACR and MII_MMDAADR. Note that I noticed that the following code do infinite loop in the read/wirte function. If it accesses unimplemented PHY register, it will hang. It should be fixed: arm/at91/at91emac.c arm/ep93xx/epe.c arm/omap/omapl1x_emac.c mips/ralink/ralink_eth.c arch/powerpc/booke/dev/pq3etsec.c(read) dev/cadence/if_cemac.c <- hkenken dev/ic/lan9118.c Tested with the following device: axe+ukphy axe+rgephy axen+rgephy (tested by Andrius V) wm+atphy wm+ukphy wm+igphy wm+ihphy wm+makphy sk+makphy sk+brgphy sk+gentbi msk+makphy sip+icsphy sip+ukphy re+rgephy bge+brgphy bnx+brgphy gsip+gphyter rtk+rlphy fxp+inphy (tested by Andrius V) tlp+acphy ex+exphy epic+qsphy vge+ciphy (tested by Andrius V) vr+ukphy (tested by Andrius V) vte+ukphy (tested by Andrius V) Not tested (MAC): arm:at91emac arm:cemac arm:epe arm:geminigmac arm:enet arm:cpsw arm:emac(omac) arm:emac(sunxi) arm:npe evbppc:temac macppc:bm macppc:gm mips:aumac mips:ae mips:cnmac mips:reth mips:sbmac playstation2:smap powerpc:tsec powerpc:emac(ibm4xx) sgimips:mec sparc:be sf ne(ax88190, dl10019) awge ep gem hme smsh mtd sm age alc ale bce cas et jme lii nfe pcn ste stge tl xi aue mue smsc udav url Not tested (PHY): amhphy bmtphy dmphy etphy glxtphy ikphy iophy lxtphy nsphyter pnaphy rdcphy sqphy tlphy tqphy urlphy
2019-01-22 06:42:24 +03:00
if (i == MVGBE_PHY_TIMEOUT) {
aprint_error_ifnet(ifp, "phy write timed out\n");
Change MII PHY read/write API from: int (*mii_readreg_t)(device_t, int, int); void (*mii_writereg_t)(device_t, int, int, int); to: int (*mii_readreg_t)(device_t, int, int, uint16_t *); int (*mii_writereg_t)(device_t, int, int, uint16_t); Now we can test if a read/write operation failed or not by the return value. In 802.3 spec says that the PHY shall not respond to read/write transaction to the unimplemented register(22.2.4.3). Detecting timeout can be used to check whether a register is implemented or not (if the register conforms to the spec). ukphy(4) can be used this for MII_MMDACR and MII_MMDAADR. Note that I noticed that the following code do infinite loop in the read/wirte function. If it accesses unimplemented PHY register, it will hang. It should be fixed: arm/at91/at91emac.c arm/ep93xx/epe.c arm/omap/omapl1x_emac.c mips/ralink/ralink_eth.c arch/powerpc/booke/dev/pq3etsec.c(read) dev/cadence/if_cemac.c <- hkenken dev/ic/lan9118.c Tested with the following device: axe+ukphy axe+rgephy axen+rgephy (tested by Andrius V) wm+atphy wm+ukphy wm+igphy wm+ihphy wm+makphy sk+makphy sk+brgphy sk+gentbi msk+makphy sip+icsphy sip+ukphy re+rgephy bge+brgphy bnx+brgphy gsip+gphyter rtk+rlphy fxp+inphy (tested by Andrius V) tlp+acphy ex+exphy epic+qsphy vge+ciphy (tested by Andrius V) vr+ukphy (tested by Andrius V) vte+ukphy (tested by Andrius V) Not tested (MAC): arm:at91emac arm:cemac arm:epe arm:geminigmac arm:enet arm:cpsw arm:emac(omac) arm:emac(sunxi) arm:npe evbppc:temac macppc:bm macppc:gm mips:aumac mips:ae mips:cnmac mips:reth mips:sbmac playstation2:smap powerpc:tsec powerpc:emac(ibm4xx) sgimips:mec sparc:be sf ne(ax88190, dl10019) awge ep gem hme smsh mtd sm age alc ale bce cas et jme lii nfe pcn ste stge tl xi aue mue smsc udav url Not tested (PHY): amhphy bmtphy dmphy etphy glxtphy ikphy iophy lxtphy nsphyter pnaphy rdcphy sqphy tlphy tqphy urlphy
2019-01-22 06:42:24 +03:00
rv = ETIMEDOUT;
}
return rv;
}
static void
mvgbec_miibus_statchg(struct ifnet *ifp)
{
/* nothing to do */
}
static void
2014-03-15 17:33:48 +04:00
mvgbec_wininit(struct mvgbec_softc *sc, enum marvell_tags *tags)
{
device_t pdev = device_parent(sc->sc_dev);
uint64_t base;
uint32_t en, ac, size;
int window, target, attr, rv, i;
/* First disable all address decode windows */
en = MVGBE_BARE_EN_MASK;
MVGBE_WRITE(sc, MVGBE_BARE, en);
ac = 0;
for (window = 0, i = 0;
tags[i] != MARVELL_TAG_UNDEFINED && window < MVGBE_NWINDOW; i++) {
rv = marvell_winparams_by_tag(pdev, tags[i],
&target, &attr, &base, &size);
if (rv != 0 || size == 0)
continue;
if (base > 0xffffffffULL) {
if (window >= MVGBE_NREMAP) {
aprint_error_dev(sc->sc_dev,
"can't remap window %d\n", window);
continue;
}
MVGBE_WRITE(sc, MVGBE_HA(window),
(base >> 32) & 0xffffffff);
}
MVGBE_WRITE(sc, MVGBE_BASEADDR(window),
MVGBE_BASEADDR_TARGET(target) |
MVGBE_BASEADDR_ATTR(attr) |
MVGBE_BASEADDR_BASE(base));
MVGBE_WRITE(sc, MVGBE_S(window), MVGBE_S_SIZE(size));
en &= ~(1 << window);
/* set full access (r/w) */
ac |= MVGBE_EPAP_EPAR(window, MVGBE_EPAP_AC_FA);
window++;
}
/* allow to access decode window */
MVGBE_WRITE(sc, MVGBE_EPAP, ac);
MVGBE_WRITE(sc, MVGBE_BARE, en);
}
/* ARGSUSED */
static int
2011-07-28 19:36:47 +04:00
mvgbe_match(device_t parent, cfdata_t match, void *aux)
{
struct marvell_attach_args *mva = aux;
uint32_t pbase, maddrh, maddrl;
prop_dictionary_t dict;
dict = device_properties(parent);
if (dict) {
if (prop_dictionary_get(dict, "mac-address"))
return 1;
}
pbase = MVGBE_PORTR_BASE + mva->mva_unit * MVGBE_PORTR_SIZE;
maddrh =
bus_space_read_4(mva->mva_iot, mva->mva_ioh, pbase + MVGBE_MACAH);
maddrl =
bus_space_read_4(mva->mva_iot, mva->mva_ioh, pbase + MVGBE_MACAL);
if ((maddrh | maddrl) == 0)
return 0;
return 1;
}
/* ARGSUSED */
static void
mvgbe_attach(device_t parent, device_t self, void *aux)
{
2013-12-23 06:23:25 +04:00
struct mvgbec_softc *csc = device_private(parent);
struct mvgbe_softc *sc = device_private(self);
struct marvell_attach_args *mva = aux;
struct mvgbe_txmap_entry *entry;
prop_dictionary_t dict;
prop_data_t enaddrp;
struct ifnet *ifp;
struct mii_data * const mii = &sc->sc_mii;
bus_dma_segment_t seg;
bus_dmamap_t dmamap;
int rseg, i;
uint32_t maddrh, maddrl;
uint8_t enaddr[ETHER_ADDR_LEN];
void *kva;
aprint_naive("\n");
aprint_normal("\n");
dict = device_properties(parent);
if (dict)
enaddrp = prop_dictionary_get(dict, "mac-address");
else
enaddrp = NULL;
sc->sc_dev = self;
sc->sc_port = mva->mva_unit;
sc->sc_iot = mva->mva_iot;
2012-10-27 01:03:26 +04:00
callout_init(&sc->sc_tick_ch, 0);
callout_setfunc(&sc->sc_tick_ch, mvgbe_tick, sc);
if (bus_space_subregion(mva->mva_iot, mva->mva_ioh,
MVGBE_PORTR_BASE + mva->mva_unit * MVGBE_PORTR_SIZE,
MVGBE_PORTR_SIZE, &sc->sc_ioh)) {
aprint_error_dev(self, "Cannot map registers\n");
return;
}
if (bus_space_subregion(mva->mva_iot, mva->mva_ioh,
MVGBE_PORTDAFR_BASE + mva->mva_unit * MVGBE_PORTDAFR_SIZE,
MVGBE_PORTDAFR_SIZE, &sc->sc_dafh)) {
aprint_error_dev(self,
"Cannot map destination address filter registers\n");
return;
}
sc->sc_dmat = mva->mva_dmat;
2013-12-23 06:23:25 +04:00
if (csc->sc_flags & FLAGS_HAS_PV) {
/* GbE port has Port Version register. */
sc->sc_version = MVGBE_READ(sc, MVGBE_PV);
aprint_normal_dev(self, "Port Version 0x%x\n", sc->sc_version);
}
if (sc->sc_version >= 0x10) {
/*
* Armada XP
*/
if (bus_space_subregion(mva->mva_iot, mva->mva_ioh,
MVGBE_PS0, sizeof(uint32_t), &sc->sc_linkup.ioh)) {
aprint_error_dev(self, "Cannot map linkup register\n");
return;
}
sc->sc_linkup.bit = MVGBE_PS0_LINKUP;
csc->sc_flags |= FLAGS_IPG2;
} else {
if (bus_space_subregion(mva->mva_iot, sc->sc_ioh,
MVGBE_PS, sizeof(uint32_t), &sc->sc_linkup.ioh)) {
aprint_error_dev(self, "Cannot map linkup register\n");
return;
}
sc->sc_linkup.bit = MVGBE_PS_LINKUP;
}
if (enaddrp) {
memcpy(enaddr, prop_data_data_nocopy(enaddrp), ETHER_ADDR_LEN);
2019-05-23 16:10:50 +03:00
maddrh = enaddr[0] << 24;
maddrh |= enaddr[1] << 16;
maddrh |= enaddr[2] << 8;
maddrh |= enaddr[3];
2019-05-23 16:10:50 +03:00
maddrl = enaddr[4] << 8;
maddrl |= enaddr[5];
MVGBE_WRITE(sc, MVGBE_MACAH, maddrh);
MVGBE_WRITE(sc, MVGBE_MACAL, maddrl);
}
maddrh = MVGBE_READ(sc, MVGBE_MACAH);
maddrl = MVGBE_READ(sc, MVGBE_MACAL);
sc->sc_enaddr[0] = maddrh >> 24;
sc->sc_enaddr[1] = maddrh >> 16;
sc->sc_enaddr[2] = maddrh >> 8;
sc->sc_enaddr[3] = maddrh >> 0;
sc->sc_enaddr[4] = maddrl >> 8;
sc->sc_enaddr[5] = maddrl >> 0;
aprint_normal_dev(self, "Ethernet address %s\n",
ether_sprintf(sc->sc_enaddr));
/* clear all ethernet port interrupts */
MVGBE_WRITE(sc, MVGBE_IC, 0);
MVGBE_WRITE(sc, MVGBE_ICE, 0);
marvell_intr_establish(mva->mva_irq, IPL_NET, mvgbe_intr, sc);
/* Allocate the descriptor queues. */
if (bus_dmamem_alloc(sc->sc_dmat, sizeof(struct mvgbe_ring_data),
PAGE_SIZE, 0, &seg, 1, &rseg, BUS_DMA_NOWAIT)) {
aprint_error_dev(self, "can't alloc rx buffers\n");
return;
}
if (bus_dmamem_map(sc->sc_dmat, &seg, rseg,
sizeof(struct mvgbe_ring_data), &kva, BUS_DMA_NOWAIT)) {
aprint_error_dev(self, "can't map dma buffers (%lu bytes)\n",
(u_long)sizeof(struct mvgbe_ring_data));
goto fail1;
}
if (bus_dmamap_create(sc->sc_dmat, sizeof(struct mvgbe_ring_data), 1,
sizeof(struct mvgbe_ring_data), 0, BUS_DMA_NOWAIT,
&sc->sc_ring_map)) {
aprint_error_dev(self, "can't create dma map\n");
goto fail2;
}
if (bus_dmamap_load(sc->sc_dmat, sc->sc_ring_map, kva,
sizeof(struct mvgbe_ring_data), NULL, BUS_DMA_NOWAIT)) {
aprint_error_dev(self, "can't load dma map\n");
goto fail3;
}
for (i = 0; i < MVGBE_RX_RING_CNT; i++)
sc->sc_cdata.mvgbe_rx_chain[i].mvgbe_mbuf = NULL;
SIMPLEQ_INIT(&sc->sc_txmap_head);
for (i = 0; i < MVGBE_TX_RING_CNT; i++) {
sc->sc_cdata.mvgbe_tx_chain[i].mvgbe_mbuf = NULL;
if (bus_dmamap_create(sc->sc_dmat,
MVGBE_JLEN, MVGBE_NTXSEG, MVGBE_JLEN, 0,
BUS_DMA_NOWAIT, &dmamap)) {
aprint_error_dev(self, "Can't create TX dmamap\n");
goto fail4;
}
entry = kmem_alloc(sizeof(*entry), KM_SLEEP);
entry->dmamap = dmamap;
SIMPLEQ_INSERT_HEAD(&sc->sc_txmap_head, entry, link);
}
sc->sc_rdata = (struct mvgbe_ring_data *)kva;
memset(sc->sc_rdata, 0, sizeof(struct mvgbe_ring_data));
/*
* We can support 802.1Q VLAN-sized frames and jumbo
* Ethernet frames.
*/
sc->sc_ethercom.ec_capabilities |=
ETHERCAP_VLAN_MTU | ETHERCAP_JUMBO_MTU;
/* Try to allocate memory for jumbo buffers. */
if (mvgbe_alloc_jumbo_mem(sc)) {
aprint_error_dev(self, "jumbo buffer allocation failed\n");
goto fail4;
}
ifp = &sc->sc_ethercom.ec_if;
ifp->if_softc = sc;
ifp->if_flags = IFF_BROADCAST | IFF_SIMPLEX | IFF_MULTICAST;
ifp->if_start = mvgbe_start;
ifp->if_ioctl = mvgbe_ioctl;
ifp->if_init = mvgbe_init;
ifp->if_stop = mvgbe_stop;
ifp->if_watchdog = mvgbe_watchdog;
/*
* We can do IPv4/TCPv4/UDPv4 checksums in hardware.
*/
sc->sc_ethercom.ec_if.if_capabilities |=
IFCAP_CSUM_IPv4_Tx | IFCAP_CSUM_IPv4_Rx |
IFCAP_CSUM_TCPv4_Tx | IFCAP_CSUM_TCPv4_Rx |
IFCAP_CSUM_UDPv4_Tx | IFCAP_CSUM_UDPv4_Rx;
/*
* But, IPv6 packets in the stream can cause incorrect TCPv4 Tx sums.
*/
sc->sc_ethercom.ec_if.if_capabilities &= ~IFCAP_CSUM_TCPv4_Tx;
Rename min/max -> uimin/uimax for better honesty. These functions are defined on unsigned int. The generic name min/max should not silently truncate to 32 bits on 64-bit systems. This is purely a name change -- no functional change intended. HOWEVER! Some subsystems have #define min(a, b) ((a) < (b) ? (a) : (b)) #define max(a, b) ((a) > (b) ? (a) : (b)) even though our standard name for that is MIN/MAX. Although these may invite multiple evaluation bugs, these do _not_ cause integer truncation. To avoid `fixing' these cases, I first changed the name in libkern, and then compile-tested every file where min/max occurred in order to confirm that it failed -- and thus confirm that nothing shadowed min/max -- before changing it. I have left a handful of bootloaders that are too annoying to compile-test, and some dead code: cobalt ews4800mips hp300 hppa ia64 luna68k vax acorn32/if_ie.c (not included in any kernels) macppc/if_gm.c (superseded by gem(4)) It should be easy to fix the fallout once identified -- this way of doing things fails safe, and the goal here, after all, is to _avoid_ silent integer truncations, not introduce them. Maybe one day we can reintroduce min/max as type-generic things that never silently truncate. But we should avoid doing that for a while, so that existing code has a chance to be detected by the compiler for conversion to uimin/uimax without changing the semantics until we can properly audit it all. (Who knows, maybe in some cases integer truncation is actually intended!)
2018-09-03 19:29:22 +03:00
IFQ_SET_MAXLEN(&ifp->if_snd, uimax(MVGBE_TX_RING_CNT - 1, IFQ_MAXLEN));
IFQ_SET_READY(&ifp->if_snd);
strcpy(ifp->if_xname, device_xname(sc->sc_dev));
mvgbe_stop(ifp, 0);
/*
* Do MII setup.
*/
mii->mii_ifp = ifp;
mii->mii_readreg = mvgbec_miibus_readreg;
mii->mii_writereg = mvgbec_miibus_writereg;
mii->mii_statchg = mvgbec_miibus_statchg;
sc->sc_ethercom.ec_mii = mii;
ifmedia_init(&mii->mii_media, 0, mvgbe_mediachange, mvgbe_mediastatus);
mii_attach(self, mii, 0xffffffff, MII_PHY_ANY,
parent == mvgbec0 ? 0 : 1, 0);
if (LIST_FIRST(&mii->mii_phys) == NULL) {
aprint_error_dev(self, "no PHY found!\n");
ifmedia_add(&mii->mii_media, IFM_ETHER | IFM_MANUAL, 0, NULL);
ifmedia_set(&mii->mii_media, IFM_ETHER | IFM_MANUAL);
} else
ifmedia_set(&mii->mii_media, IFM_ETHER | IFM_AUTO);
/*
* Call MI attach routines.
*/
if_attach(ifp);
if_deferred_start_init(ifp, NULL);
ether_ifattach(ifp, sc->sc_enaddr);
ether_set_ifflags_cb(&sc->sc_ethercom, mvgbe_ifflags_cb);
sysctl_mvgbe_init(sc);
#ifdef MVGBE_EVENT_COUNTERS
/* Attach event counters. */
evcnt_attach_dynamic(&sc->sc_ev_rxoverrun, EVCNT_TYPE_MISC,
NULL, device_xname(sc->sc_dev), "rxoverrrun");
evcnt_attach_dynamic(&sc->sc_ev_wdogsoft, EVCNT_TYPE_MISC,
NULL, device_xname(sc->sc_dev), "wdogsoft");
#endif
rnd_attach_source(&sc->sc_rnd_source, device_xname(sc->sc_dev),
RND_TYPE_NET, RND_FLAG_DEFAULT);
return;
fail4:
while ((entry = SIMPLEQ_FIRST(&sc->sc_txmap_head)) != NULL) {
SIMPLEQ_REMOVE_HEAD(&sc->sc_txmap_head, link);
bus_dmamap_destroy(sc->sc_dmat, entry->dmamap);
}
bus_dmamap_unload(sc->sc_dmat, sc->sc_ring_map);
fail3:
bus_dmamap_destroy(sc->sc_dmat, sc->sc_ring_map);
fail2:
bus_dmamem_unmap(sc->sc_dmat, kva, sizeof(struct mvgbe_ring_data));
fail1:
bus_dmamem_free(sc->sc_dmat, &seg, rseg);
return;
}
static int
mvgbe_ipginttx(struct mvgbec_softc *csc, struct mvgbe_softc *sc,
unsigned int ipginttx)
{
uint32_t reg;
reg = MVGBE_READ(sc, MVGBE_PTFUT);
if (csc->sc_flags & FLAGS_IPG2) {
if (ipginttx > MVGBE_PTFUT_IPGINTTX_V2_MAX)
return -1;
reg &= ~MVGBE_PTFUT_IPGINTTX_V2_MASK;
reg |= MVGBE_PTFUT_IPGINTTX_V2(ipginttx);
} else if (csc->sc_flags & FLAGS_IPG1) {
if (ipginttx > MVGBE_PTFUT_IPGINTTX_V1_MAX)
return -1;
reg &= ~MVGBE_PTFUT_IPGINTTX_V1_MASK;
reg |= MVGBE_PTFUT_IPGINTTX_V1(ipginttx);
}
MVGBE_WRITE(sc, MVGBE_PTFUT, reg);
return 0;
}
static int
mvgbe_ipgintrx(struct mvgbec_softc *csc, struct mvgbe_softc *sc,
unsigned int ipgintrx)
{
uint32_t reg;
reg = MVGBE_READ(sc, MVGBE_SDC);
if (csc->sc_flags & FLAGS_IPG2) {
if (ipgintrx > MVGBE_SDC_IPGINTRX_V2_MAX)
return -1;
reg &= ~MVGBE_SDC_IPGINTRX_V2_MASK;
reg |= MVGBE_SDC_IPGINTRX_V2(ipgintrx);
} else if (csc->sc_flags & FLAGS_IPG1) {
if (ipgintrx > MVGBE_SDC_IPGINTRX_V1_MAX)
return -1;
reg &= ~MVGBE_SDC_IPGINTRX_V1_MASK;
reg |= MVGBE_SDC_IPGINTRX_V1(ipgintrx);
}
MVGBE_WRITE(sc, MVGBE_SDC, reg);
return 0;
}
2012-10-27 01:03:26 +04:00
static void
mvgbe_tick(void *arg)
{
struct mvgbe_softc *sc = arg;
struct mii_data *mii = &sc->sc_mii;
int s;
s = splnet();
mii_tick(mii);
/* Need more work */
MVGBE_EVCNT_ADD(&sc->sc_ev_rxoverrun, MVGBE_READ(sc, MVGBE_POFC));
2012-10-27 01:03:26 +04:00
splx(s);
callout_schedule(&sc->sc_tick_ch, hz);
}
static int
mvgbe_intr(void *arg)
{
struct mvgbe_softc *sc = arg;
struct ifnet *ifp = &sc->sc_ethercom.ec_if;
uint32_t ic, ice, datum = 0;
int claimed = 0;
for (;;) {
ice = MVGBE_READ(sc, MVGBE_ICE);
ic = MVGBE_READ(sc, MVGBE_IC);
DPRINTFN(3, ("mvgbe_intr: ic=%#x, ice=%#x\n", ic, ice));
if (ic == 0 && ice == 0)
break;
datum = datum ^ ic ^ ice;
MVGBE_WRITE(sc, MVGBE_IC, ~ic);
MVGBE_WRITE(sc, MVGBE_ICE, ~ice);
claimed = 1;
if (!(ifp->if_flags & IFF_RUNNING))
break;
if (ice & MVGBE_ICE_LINKCHG) {
2013-12-23 06:23:25 +04:00
if (MVGBE_IS_LINKUP(sc)) {
/* Enable port RX and TX. */
MVGBE_WRITE(sc, MVGBE_RQC, MVGBE_RQC_ENQ(0));
2013-12-23 06:23:25 +04:00
MVGBE_WRITE(sc, MVGBE_TQC, MVGBE_TQC_ENQ(0));
} else {
MVGBE_WRITE(sc, MVGBE_RQC, MVGBE_RQC_DISQ(0));
2013-12-23 06:23:25 +04:00
MVGBE_WRITE(sc, MVGBE_TQC, MVGBE_TQC_DISQ(0));
}
/* Notify link change event to mii layer */
mii_pollstat(&sc->sc_mii);
}
if (ic & (MVGBE_IC_RXBUF | MVGBE_IC_RXERROR))
mvgbe_rxeof(sc);
2013-12-23 06:23:25 +04:00
if (ice & (MVGBE_ICE_TXBUF_MASK | MVGBE_ICE_TXERR_MASK))
mvgbe_txeof(sc);
}
if_schedule_deferred_start(ifp);
rnd_add_uint32(&sc->sc_rnd_source, datum);
return claimed;
}
static void
mvgbe_start(struct ifnet *ifp)
{
struct mvgbe_softc *sc = ifp->if_softc;
struct mbuf *m_head = NULL;
uint32_t idx = sc->sc_cdata.mvgbe_tx_prod;
int pkts = 0;
DPRINTFN(3, ("mvgbe_start (idx %d, tx_chain[idx] %p)\n", idx,
sc->sc_cdata.mvgbe_tx_chain[idx].mvgbe_mbuf));
if ((ifp->if_flags & (IFF_RUNNING | IFF_OACTIVE)) != IFF_RUNNING)
return;
/* If Link is DOWN, can't start TX */
2013-12-23 06:23:25 +04:00
if (!MVGBE_IS_LINKUP(sc))
return;
while (sc->sc_cdata.mvgbe_tx_chain[idx].mvgbe_mbuf == NULL) {
IFQ_POLL(&ifp->if_snd, m_head);
if (m_head == NULL)
break;
/*
* Pack the data into the transmit ring. If we
* don't have room, set the OACTIVE flag and wait
* for the NIC to drain the ring.
*/
if (mvgbe_encap(sc, m_head, &idx)) {
if (sc->sc_cdata.mvgbe_tx_cnt > 0)
ifp->if_flags |= IFF_OACTIVE;
break;
}
/* now we are committed to transmit the packet */
IFQ_DEQUEUE(&ifp->if_snd, m_head);
pkts++;
/*
* If there's a BPF listener, bounce a copy of this frame
* to him.
*/
bpf_mtap(ifp, m_head, BPF_D_OUT);
}
if (pkts == 0)
return;
/* Transmit at Queue 0 */
if (idx != sc->sc_cdata.mvgbe_tx_prod) {
sc->sc_cdata.mvgbe_tx_prod = idx;
2013-12-23 06:23:25 +04:00
MVGBE_WRITE(sc, MVGBE_TQC, MVGBE_TQC_ENQ(0));
/*
* Set a timeout in case the chip goes out to lunch.
*/
ifp->if_timer = 1;
sc->sc_wdogsoft = 1;
}
}
static int
mvgbe_ioctl(struct ifnet *ifp, u_long cmd, void *data)
{
struct mvgbe_softc *sc = ifp->if_softc;
int s, error = 0;
s = splnet();
switch (cmd) {
default:
DPRINTFN(2, ("mvgbe_ioctl ETHER\n"));
error = ether_ioctl(ifp, cmd, data);
if (error == ENETRESET) {
if (ifp->if_flags & IFF_RUNNING) {
mvgbe_filter_setup(sc);
}
error = 0;
}
break;
}
splx(s);
return error;
}
static int
mvgbe_init(struct ifnet *ifp)
{
struct mvgbe_softc *sc = ifp->if_softc;
2010-10-02 09:57:42 +04:00
struct mvgbec_softc *csc = device_private(device_parent(sc->sc_dev));
struct mii_data *mii = &sc->sc_mii;
2010-10-02 09:57:42 +04:00
uint32_t reg;
int i;
DPRINTFN(2, ("mvgbe_init\n"));
/* Cancel pending I/O and free all RX/TX buffers. */
mvgbe_stop(ifp, 0);
/* clear all ethernet port interrupts */
MVGBE_WRITE(sc, MVGBE_IC, 0);
MVGBE_WRITE(sc, MVGBE_ICE, 0);
/* Init TX/RX descriptors */
if (mvgbe_init_tx_ring(sc) == ENOBUFS) {
aprint_error_ifnet(ifp,
"initialization failed: no memory for tx buffers\n");
return ENOBUFS;
}
if (mvgbe_init_rx_ring(sc) == ENOBUFS) {
aprint_error_ifnet(ifp,
"initialization failed: no memory for rx buffers\n");
return ENOBUFS;
}
if ((csc->sc_flags & FLAGS_IPG1) || (csc->sc_flags & FLAGS_IPG2)) {
sc->sc_ipginttx = MVGBE_IPGINTTX_DEFAULT;
sc->sc_ipgintrx = MVGBE_IPGINTRX_DEFAULT;
}
if (csc->sc_flags & FLAGS_FIX_MTU)
MVGBE_WRITE(sc, MVGBE_MTU, 0); /* hw reset value is wrong */
2013-12-23 06:23:25 +04:00
if (sc->sc_version >= 0x10) {
MVGBE_WRITE(csc, MVGBE_PANC,
MVGBE_PANC_FORCELINKPASS |
MVGBE_PANC_INBANDANBYPASSEN |
MVGBE_PANC_SETMIISPEED |
MVGBE_PANC_SETGMIISPEED |
MVGBE_PANC_ANSPEEDEN |
MVGBE_PANC_SETFCEN |
MVGBE_PANC_PAUSEADV |
MVGBE_PANC_SETFULLDX |
MVGBE_PANC_ANDUPLEXEN |
MVGBE_PANC_RESERVED);
MVGBE_WRITE(csc, MVGBE_PMACC0,
MVGBE_PMACC0_RESERVED |
MVGBE_PMACC0_FRAMESIZELIMIT(1600));
2014-03-15 17:33:48 +04:00
reg = MVGBE_READ(csc, MVGBE_PMACC2);
reg &= MVGBE_PMACC2_PCSEN; /* keep PCSEN bit */
2013-12-23 06:23:25 +04:00
MVGBE_WRITE(csc, MVGBE_PMACC2,
2014-03-15 17:33:48 +04:00
reg | MVGBE_PMACC2_RESERVED | MVGBE_PMACC2_RGMIIEN);
2013-12-23 06:23:25 +04:00
MVGBE_WRITE(sc, MVGBE_PXCX,
MVGBE_READ(sc, MVGBE_PXCX) & ~MVGBE_PXCX_TXCRCDIS);
#ifndef MULTIPROCESSOR
MVGBE_WRITE(sc, MVGBE_PACC, MVGVE_PACC_ACCELERATIONMODE_BM);
#else
MVGBE_WRITE(sc, MVGBE_PACC, MVGVE_PACC_ACCELERATIONMODE_EDM);
#endif
} else {
MVGBE_WRITE(sc, MVGBE_PSC,
MVGBE_PSC_ANFC | /* Enable Auto-Neg Flow Ctrl */
MVGBE_PSC_RESERVED | /* Must be set to 1 */
MVGBE_PSC_FLFAIL | /* Do NOT Force Link Fail */
MVGBE_PSC_MRU(MVGBE_PSC_MRU_9022) | /* we want 9k */
MVGBE_PSC_SETFULLDX); /* Set_FullDx */
/* XXXX: mvgbe(4) always use RGMII. */
MVGBE_WRITE(sc, MVGBE_PSC1,
MVGBE_READ(sc, MVGBE_PSC1) | MVGBE_PSC1_RGMIIEN);
/* XXXX: Also always Weighted Round-Robin Priority Mode */
MVGBE_WRITE(sc, MVGBE_TQFPC, MVGBE_TQFPC_EN(0));
sc->sc_cmdsts_opts = MVGBE_TX_GENERATE_CRC;
}
MVGBE_WRITE(sc, MVGBE_CRDP(0), MVGBE_RX_RING_ADDR(sc, 0));
MVGBE_WRITE(sc, MVGBE_TCQDP, MVGBE_TX_RING_ADDR(sc, 0));
if (csc->sc_flags & FLAGS_FIX_TQTB) {
/*
2010-10-02 09:57:42 +04:00
* Queue 0 (offset 0x72700) must be programmed to 0x3fffffff.
* And offset 0x72704 must be programmed to 0x03ffffff.
* Queue 1 through 7 must be programmed to 0x0.
*/
2010-10-02 09:57:42 +04:00
MVGBE_WRITE(sc, MVGBE_TQTBCOUNT(0), 0x3fffffff);
MVGBE_WRITE(sc, MVGBE_TQTBCONFIG(0), 0x03ffffff);
for (i = 1; i < 8; i++) {
MVGBE_WRITE(sc, MVGBE_TQTBCOUNT(i), 0x0);
MVGBE_WRITE(sc, MVGBE_TQTBCONFIG(i), 0x0);
}
2013-12-23 06:23:25 +04:00
} else if (sc->sc_version < 0x10)
2010-10-02 09:57:42 +04:00
for (i = 1; i < 8; i++) {
MVGBE_WRITE(sc, MVGBE_TQTBCOUNT(i), 0x3fffffff);
MVGBE_WRITE(sc, MVGBE_TQTBCONFIG(i), 0xffff7fff);
MVGBE_WRITE(sc, MVGBE_TQAC(i), 0xfc0000ff);
}
MVGBE_WRITE(sc, MVGBE_PXC, MVGBE_PXC_RXCS);
MVGBE_WRITE(sc, MVGBE_PXCX, 0);
/* Set SDC register except IPGINT bits */
MVGBE_WRITE(sc, MVGBE_SDC,
MVGBE_SDC_RXBSZ_16_64BITWORDS |
#ifndef MVGBE_BIG_ENDIAN
MVGBE_SDC_BLMR | /* Big/Little Endian Receive Mode: No swap */
MVGBE_SDC_BLMT | /* Big/Little Endian Transmit Mode: No swap */
#endif
MVGBE_SDC_TXBSZ_16_64BITWORDS);
/* And then set IPGINT bits */
mvgbe_ipgintrx(csc, sc, sc->sc_ipgintrx);
/* Tx side */
MVGBE_WRITE(sc, MVGBE_PTFUT, 0);
mvgbe_ipginttx(csc, sc, sc->sc_ipginttx);
mvgbe_filter_setup(sc);
mii_mediachg(mii);
/* Enable port */
2013-12-23 06:23:25 +04:00
if (sc->sc_version >= 0x10) {
reg = MVGBE_READ(csc, MVGBE_PMACC0);
MVGBE_WRITE(csc, MVGBE_PMACC0, reg | MVGBE_PMACC0_PORTEN);
} else {
reg = MVGBE_READ(sc, MVGBE_PSC);
MVGBE_WRITE(sc, MVGBE_PSC, reg | MVGBE_PSC_PORTEN);
}
/* If Link is UP, Start RX and TX traffic */
2013-12-23 06:23:25 +04:00
if (MVGBE_IS_LINKUP(sc)) {
/* Enable port RX/TX. */
MVGBE_WRITE(sc, MVGBE_RQC, MVGBE_RQC_ENQ(0));
2013-12-23 06:23:25 +04:00
MVGBE_WRITE(sc, MVGBE_TQC, MVGBE_TQC_ENQ(0));
}
/* Enable interrupt masks */
MVGBE_WRITE(sc, MVGBE_PIM,
MVGBE_IC_RXBUF |
MVGBE_IC_EXTEND |
MVGBE_IC_RXBUFQ_MASK |
MVGBE_IC_RXERROR |
MVGBE_IC_RXERRQ_MASK);
MVGBE_WRITE(sc, MVGBE_PEIM,
2013-12-23 06:23:25 +04:00
MVGBE_ICE_TXBUF_MASK |
MVGBE_ICE_TXERR_MASK |
MVGBE_ICE_LINKCHG);
2012-10-27 01:03:26 +04:00
callout_schedule(&sc->sc_tick_ch, hz);
ifp->if_flags |= IFF_RUNNING;
ifp->if_flags &= ~IFF_OACTIVE;
return 0;
}
/* ARGSUSED */
static void
mvgbe_stop(struct ifnet *ifp, int disable)
{
struct mvgbe_softc *sc = ifp->if_softc;
struct mvgbec_softc *csc = device_private(device_parent(sc->sc_dev));
struct mvgbe_chain_data *cdata = &sc->sc_cdata;
2013-12-23 06:23:25 +04:00
uint32_t reg, txinprog, txfifoemp;
int i, cnt;
DPRINTFN(2, ("mvgbe_stop\n"));
2012-10-27 01:03:26 +04:00
callout_stop(&sc->sc_tick_ch);
/* Stop Rx port activity. Check port Rx activity. */
reg = MVGBE_READ(sc, MVGBE_RQC);
if (reg & MVGBE_RQC_ENQ_MASK)
/* Issue stop command for active channels only */
MVGBE_WRITE(sc, MVGBE_RQC, MVGBE_RQC_DISQ_DISABLE(reg));
/* Stop Tx port activity. Check port Tx activity. */
2013-12-23 06:23:25 +04:00
if (MVGBE_READ(sc, MVGBE_TQC) & MVGBE_TQC_ENQ(0))
MVGBE_WRITE(sc, MVGBE_TQC, MVGBE_TQC_DISQ(0));
/* Force link down */
2013-12-23 06:23:25 +04:00
if (sc->sc_version >= 0x10) {
reg = MVGBE_READ(csc, MVGBE_PANC);
MVGBE_WRITE(csc, MVGBE_PANC, reg | MVGBE_PANC_FORCELINKFAIL);
txinprog = MVGBE_PS_TXINPROG_(0);
txfifoemp = MVGBE_PS_TXFIFOEMP_(0);
} else {
reg = MVGBE_READ(sc, MVGBE_PSC);
MVGBE_WRITE(sc, MVGBE_PSC, reg & ~MVGBE_PSC_FLFAIL);
txinprog = MVGBE_PS_TXINPROG;
txfifoemp = MVGBE_PS_TXFIFOEMP;
}
2019-05-23 16:10:50 +03:00
#define RX_DISABLE_TIMEOUT 0x1000000
#define TX_FIFO_EMPTY_TIMEOUT 0x1000000
/* Wait for all Rx activity to terminate. */
cnt = 0;
do {
if (cnt >= RX_DISABLE_TIMEOUT) {
aprint_error_ifnet(ifp,
"timeout for RX stopped. rqc 0x%x\n", reg);
break;
}
cnt++;
/*
* Check Receive Queue Command register that all Rx queues
* are stopped
*/
reg = MVGBE_READ(sc, MVGBE_RQC);
} while (reg & 0xff);
/* Double check to verify that TX FIFO is empty */
cnt = 0;
while (1) {
do {
if (cnt >= TX_FIFO_EMPTY_TIMEOUT) {
aprint_error_ifnet(ifp,
"timeout for TX FIFO empty. status 0x%x\n",
reg);
break;
}
cnt++;
reg = MVGBE_READ(sc, MVGBE_PS);
2013-12-23 06:23:25 +04:00
} while (!(reg & txfifoemp) || reg & txinprog);
if (cnt >= TX_FIFO_EMPTY_TIMEOUT)
break;
/* Double check */
reg = MVGBE_READ(sc, MVGBE_PS);
2013-12-23 06:23:25 +04:00
if (reg & txfifoemp && !(reg & txinprog))
break;
else
aprint_error_ifnet(ifp,
"TX FIFO empty double check failed."
" %d loops, status 0x%x\n", cnt, reg);
}
2013-12-23 06:23:25 +04:00
/* Reset the Enable bit */
if (sc->sc_version >= 0x10) {
reg = MVGBE_READ(csc, MVGBE_PMACC0);
MVGBE_WRITE(csc, MVGBE_PMACC0, reg & ~MVGBE_PMACC0_PORTEN);
} else {
reg = MVGBE_READ(sc, MVGBE_PSC);
MVGBE_WRITE(sc, MVGBE_PSC, reg & ~MVGBE_PSC_PORTEN);
}
/*
* Disable and clear interrupts
* 0) controller interrupt
* 1) port interrupt cause
* 2) port interrupt mask
*/
MVGBE_WRITE(csc, MVGBE_EUIM, 0);
MVGBE_WRITE(csc, MVGBE_EUIC, 0);
MVGBE_WRITE(sc, MVGBE_IC, 0);
MVGBE_WRITE(sc, MVGBE_ICE, 0);
MVGBE_WRITE(sc, MVGBE_PIM, 0);
MVGBE_WRITE(sc, MVGBE_PEIM, 0);
/* Free RX and TX mbufs still in the queues. */
for (i = 0; i < MVGBE_RX_RING_CNT; i++) {
if (cdata->mvgbe_rx_chain[i].mvgbe_mbuf != NULL) {
m_freem(cdata->mvgbe_rx_chain[i].mvgbe_mbuf);
cdata->mvgbe_rx_chain[i].mvgbe_mbuf = NULL;
}
}
for (i = 0; i < MVGBE_TX_RING_CNT; i++) {
if (cdata->mvgbe_tx_chain[i].mvgbe_mbuf != NULL) {
m_freem(cdata->mvgbe_tx_chain[i].mvgbe_mbuf);
cdata->mvgbe_tx_chain[i].mvgbe_mbuf = NULL;
}
}
ifp->if_flags &= ~(IFF_RUNNING | IFF_OACTIVE);
}
static void
mvgbe_watchdog(struct ifnet *ifp)
{
struct mvgbe_softc *sc = ifp->if_softc;
/*
* Reclaim first as there is a possibility of losing Tx completion
* interrupts.
*/
mvgbe_txeof(sc);
if (sc->sc_cdata.mvgbe_tx_cnt != 0) {
if (sc->sc_wdogsoft) {
/*
* There is race condition between CPU and DMA
* engine. When DMA engine encounters queue end,
* it clears MVGBE_TQC_ENQ bit.
*/
2013-12-23 06:23:25 +04:00
MVGBE_WRITE(sc, MVGBE_TQC, MVGBE_TQC_ENQ(0));
ifp->if_timer = 5;
sc->sc_wdogsoft = 0;
MVGBE_EVCNT_INCR(&sc->sc_ev_wdogsoft);
} else {
aprint_error_ifnet(ifp, "watchdog timeout\n");
2020-02-03 10:47:53 +03:00
if_statinc(ifp, if_oerrors);
mvgbe_init(ifp);
}
}
}
static int
mvgbe_ifflags_cb(struct ethercom *ec)
{
struct ifnet *ifp = &ec->ec_if;
struct mvgbe_softc *sc = ifp->if_softc;
u_short change = ifp->if_flags ^ sc->sc_if_flags;
if (change != 0)
sc->sc_if_flags = ifp->if_flags;
if ((change & ~(IFF_CANTCHANGE | IFF_DEBUG)) != 0)
return ENETRESET;
if ((change & IFF_PROMISC) != 0)
mvgbe_filter_setup(sc);
return 0;
}
/*
* Set media options.
*/
static int
mvgbe_mediachange(struct ifnet *ifp)
{
return ether_mediachange(ifp);
}
/*
* Report current media status.
*/
static void
mvgbe_mediastatus(struct ifnet *ifp, struct ifmediareq *ifmr)
{
ether_mediastatus(ifp, ifmr);
}
static int
mvgbe_init_rx_ring(struct mvgbe_softc *sc)
{
struct mvgbe_chain_data *cd = &sc->sc_cdata;
struct mvgbe_ring_data *rd = sc->sc_rdata;
int i;
memset(rd->mvgbe_rx_ring, 0,
sizeof(struct mvgbe_rx_desc) * MVGBE_RX_RING_CNT);
for (i = 0; i < MVGBE_RX_RING_CNT; i++) {
cd->mvgbe_rx_chain[i].mvgbe_desc =
&rd->mvgbe_rx_ring[i];
if (i == MVGBE_RX_RING_CNT - 1) {
cd->mvgbe_rx_chain[i].mvgbe_next =
&cd->mvgbe_rx_chain[0];
rd->mvgbe_rx_ring[i].nextdescptr =
H2MVGBE32(MVGBE_RX_RING_ADDR(sc, 0));
} else {
cd->mvgbe_rx_chain[i].mvgbe_next =
&cd->mvgbe_rx_chain[i + 1];
rd->mvgbe_rx_ring[i].nextdescptr =
H2MVGBE32(MVGBE_RX_RING_ADDR(sc, i + 1));
}
}
for (i = 0; i < MVGBE_RX_RING_CNT; i++) {
if (mvgbe_newbuf(sc, i, NULL,
sc->sc_cdata.mvgbe_rx_jumbo_map) == ENOBUFS) {
aprint_error_ifnet(&sc->sc_ethercom.ec_if,
"failed alloc of %dth mbuf\n", i);
return ENOBUFS;
}
}
sc->sc_cdata.mvgbe_rx_prod = 0;
sc->sc_cdata.mvgbe_rx_cons = 0;
return 0;
}
static int
mvgbe_init_tx_ring(struct mvgbe_softc *sc)
{
struct mvgbe_chain_data *cd = &sc->sc_cdata;
struct mvgbe_ring_data *rd = sc->sc_rdata;
int i;
memset(sc->sc_rdata->mvgbe_tx_ring, 0,
sizeof(struct mvgbe_tx_desc) * MVGBE_TX_RING_CNT);
for (i = 0; i < MVGBE_TX_RING_CNT; i++) {
cd->mvgbe_tx_chain[i].mvgbe_desc =
&rd->mvgbe_tx_ring[i];
if (i == MVGBE_TX_RING_CNT - 1) {
cd->mvgbe_tx_chain[i].mvgbe_next =
&cd->mvgbe_tx_chain[0];
rd->mvgbe_tx_ring[i].nextdescptr =
H2MVGBE32(MVGBE_TX_RING_ADDR(sc, 0));
} else {
cd->mvgbe_tx_chain[i].mvgbe_next =
&cd->mvgbe_tx_chain[i + 1];
rd->mvgbe_tx_ring[i].nextdescptr =
H2MVGBE32(MVGBE_TX_RING_ADDR(sc, i + 1));
}
rd->mvgbe_tx_ring[i].cmdsts =
H2MVGBE32(MVGBE_BUFFER_OWNED_BY_HOST);
}
sc->sc_cdata.mvgbe_tx_prod = 0;
sc->sc_cdata.mvgbe_tx_cons = 0;
sc->sc_cdata.mvgbe_tx_cnt = 0;
MVGBE_CDTXSYNC(sc, 0, MVGBE_TX_RING_CNT,
BUS_DMASYNC_PREREAD | BUS_DMASYNC_PREWRITE);
return 0;
}
static int
mvgbe_newbuf(struct mvgbe_softc *sc, int i, struct mbuf *m,
bus_dmamap_t dmamap)
{
struct mbuf *m_new = NULL;
struct mvgbe_chain *c;
struct mvgbe_rx_desc *r;
int align;
vaddr_t offset;
if (m == NULL) {
void *buf = NULL;
MGETHDR(m_new, M_DONTWAIT, MT_DATA);
if (m_new == NULL) {
aprint_error_ifnet(&sc->sc_ethercom.ec_if,
"no memory for rx list -- packet dropped!\n");
return ENOBUFS;
}
/* Allocate the jumbo buffer */
buf = mvgbe_jalloc(sc);
if (buf == NULL) {
m_freem(m_new);
DPRINTFN(1, ("%s jumbo allocation failed -- packet "
"dropped!\n", sc->sc_ethercom.ec_if.if_xname));
return ENOBUFS;
}
/* Attach the buffer to the mbuf */
m_new->m_len = m_new->m_pkthdr.len = MVGBE_JLEN;
MEXTADD(m_new, buf, MVGBE_JLEN, 0, mvgbe_jfree, sc);
} else {
/*
* We're re-using a previously allocated mbuf;
* be sure to re-init pointers and lengths to
* default values.
*/
m_new = m;
m_new->m_len = m_new->m_pkthdr.len = MVGBE_JLEN;
m_new->m_data = m_new->m_ext.ext_buf;
}
align = (u_long)m_new->m_data & MVGBE_RXBUF_MASK;
if (align != 0) {
DPRINTFN(1,("align = %d\n", align));
m_adj(m_new, MVGBE_RXBUF_ALIGN - align);
}
c = &sc->sc_cdata.mvgbe_rx_chain[i];
r = c->mvgbe_desc;
c->mvgbe_mbuf = m_new;
offset = (vaddr_t)m_new->m_data - (vaddr_t)sc->sc_cdata.mvgbe_jumbo_buf;
r->bufptr = H2MVGBE32(dmamap->dm_segs[0].ds_addr + offset);
r->bufsize = MVGBE_JLEN & ~MVGBE_RXBUF_MASK;
r->cmdsts =
H2MVGBE32(MVGBE_BUFFER_OWNED_BY_DMA | MVGBE_RX_ENABLE_INTERRUPT);
/* Invalidate RX buffer */
bus_dmamap_sync(sc->sc_dmat, dmamap, offset, r->bufsize,
BUS_DMASYNC_PREREAD);
2010-10-02 09:57:42 +04:00
MVGBE_CDRXSYNC(sc, i, BUS_DMASYNC_PREREAD | BUS_DMASYNC_PREWRITE);
return 0;
}
/*
* Memory management for jumbo frames.
*/
static int
mvgbe_alloc_jumbo_mem(struct mvgbe_softc *sc)
{
char *ptr, *kva;
bus_dma_segment_t seg;
int i, rseg, state, error;
struct mvgbe_jpool_entry *entry;
state = error = 0;
/* Grab a big chunk o' storage. */
if (bus_dmamem_alloc(sc->sc_dmat, MVGBE_JMEM, PAGE_SIZE, 0,
&seg, 1, &rseg, BUS_DMA_NOWAIT)) {
aprint_error_dev(sc->sc_dev, "can't alloc rx buffers\n");
return ENOBUFS;
}
state = 1;
if (bus_dmamem_map(sc->sc_dmat, &seg, rseg, MVGBE_JMEM,
(void **)&kva, BUS_DMA_NOWAIT)) {
aprint_error_dev(sc->sc_dev,
"can't map dma buffers (%d bytes)\n", MVGBE_JMEM);
error = ENOBUFS;
goto out;
}
state = 2;
if (bus_dmamap_create(sc->sc_dmat, MVGBE_JMEM, 1, MVGBE_JMEM, 0,
BUS_DMA_NOWAIT, &sc->sc_cdata.mvgbe_rx_jumbo_map)) {
aprint_error_dev(sc->sc_dev, "can't create dma map\n");
error = ENOBUFS;
goto out;
}
state = 3;
if (bus_dmamap_load(sc->sc_dmat, sc->sc_cdata.mvgbe_rx_jumbo_map,
kva, MVGBE_JMEM, NULL, BUS_DMA_NOWAIT)) {
aprint_error_dev(sc->sc_dev, "can't load dma map\n");
error = ENOBUFS;
goto out;
}
state = 4;
sc->sc_cdata.mvgbe_jumbo_buf = (void *)kva;
DPRINTFN(1,("mvgbe_jumbo_buf = %p\n", sc->sc_cdata.mvgbe_jumbo_buf));
LIST_INIT(&sc->sc_jfree_listhead);
LIST_INIT(&sc->sc_jinuse_listhead);
/*
* Now divide it up into 9K pieces and save the addresses
* in an array.
*/
ptr = sc->sc_cdata.mvgbe_jumbo_buf;
for (i = 0; i < MVGBE_JSLOTS; i++) {
sc->sc_cdata.mvgbe_jslots[i] = ptr;
ptr += MVGBE_JLEN;
entry = kmem_alloc(sizeof(struct mvgbe_jpool_entry), KM_SLEEP);
entry->slot = i;
if (i)
LIST_INSERT_HEAD(&sc->sc_jfree_listhead, entry,
jpool_entries);
else
LIST_INSERT_HEAD(&sc->sc_jinuse_listhead, entry,
jpool_entries);
}
out:
if (error != 0) {
switch (state) {
case 4:
bus_dmamap_unload(sc->sc_dmat,
sc->sc_cdata.mvgbe_rx_jumbo_map);
case 3:
bus_dmamap_destroy(sc->sc_dmat,
sc->sc_cdata.mvgbe_rx_jumbo_map);
case 2:
bus_dmamem_unmap(sc->sc_dmat, kva, MVGBE_JMEM);
case 1:
bus_dmamem_free(sc->sc_dmat, &seg, rseg);
break;
default:
break;
}
}
return error;
}
/*
* Allocate a jumbo buffer.
*/
static void *
mvgbe_jalloc(struct mvgbe_softc *sc)
{
struct mvgbe_jpool_entry *entry;
entry = LIST_FIRST(&sc->sc_jfree_listhead);
if (entry == NULL)
return NULL;
LIST_REMOVE(entry, jpool_entries);
LIST_INSERT_HEAD(&sc->sc_jinuse_listhead, entry, jpool_entries);
return sc->sc_cdata.mvgbe_jslots[entry->slot];
}
/*
* Release a jumbo buffer.
*/
static void
mvgbe_jfree(struct mbuf *m, void *buf, size_t size, void *arg)
{
struct mvgbe_jpool_entry *entry;
struct mvgbe_softc *sc;
int i, s;
/* Extract the softc struct pointer. */
sc = (struct mvgbe_softc *)arg;
if (sc == NULL)
panic("%s: can't find softc pointer!", __func__);
/* calculate the slot this buffer belongs to */
i = ((vaddr_t)buf - (vaddr_t)sc->sc_cdata.mvgbe_jumbo_buf) / MVGBE_JLEN;
if ((i < 0) || (i >= MVGBE_JSLOTS))
panic("%s: asked to free buffer that we don't manage!",
__func__);
s = splvm();
entry = LIST_FIRST(&sc->sc_jinuse_listhead);
if (entry == NULL)
panic("%s: buffer not in use!", __func__);
entry->slot = i;
LIST_REMOVE(entry, jpool_entries);
LIST_INSERT_HEAD(&sc->sc_jfree_listhead, entry, jpool_entries);
if (__predict_true(m != NULL))
pool_cache_put(mb_cache, m);
splx(s);
}
static int
mvgbe_encap(struct mvgbe_softc *sc, struct mbuf *m_head,
uint32_t *txidx)
{
struct mvgbe_tx_desc *f = NULL;
struct mvgbe_txmap_entry *entry;
bus_dma_segment_t *txseg;
bus_dmamap_t txmap;
2013-12-23 06:23:25 +04:00
uint32_t first, current, last, cmdsts;
int m_csumflags, i;
bool needs_defrag = false;
DPRINTFN(3, ("mvgbe_encap\n"));
entry = SIMPLEQ_FIRST(&sc->sc_txmap_head);
if (entry == NULL) {
DPRINTFN(2, ("mvgbe_encap: no txmap available\n"));
return ENOBUFS;
}
txmap = entry->dmamap;
first = current = last = *txidx;
/*
* Preserve m_pkthdr.csum_flags here since m_head might be
* updated by m_defrag()
*/
m_csumflags = m_head->m_pkthdr.csum_flags;
do_defrag:
if (__predict_false(needs_defrag == true)) {
/* A small unaligned segment was detected. */
struct mbuf *m_new;
m_new = m_defrag(m_head, M_DONTWAIT);
if (m_new == NULL) {
DPRINTFN(2, ("mvgbe_encap: defrag failed\n"));
return EFBIG;
}
m_head = m_new;
}
/*
* Start packing the mbufs in this chain into
* the fragment pointers. Stop when we run out
* of fragments or hit the end of the mbuf chain.
*/
if (bus_dmamap_load_mbuf(sc->sc_dmat, txmap, m_head, BUS_DMA_NOWAIT)) {
DPRINTFN(1, ("mvgbe_encap: dmamap failed\n"));
return ENOBUFS;
}
txseg = txmap->dm_segs;
if (__predict_true(needs_defrag == false)) {
/*
* Detect rarely encountered DMA limitation.
*/
for (i = 0; i < txmap->dm_nsegs; i++) {
if (((txseg[i].ds_addr & 7) != 0) &&
(txseg[i].ds_len <= 8) &&
(txseg[i].ds_len >= 1)
) {
txseg = NULL;
bus_dmamap_unload(sc->sc_dmat, txmap);
needs_defrag = true;
goto do_defrag;
}
}
}
/* Sync the DMA map. */
bus_dmamap_sync(sc->sc_dmat, txmap, 0, txmap->dm_mapsize,
BUS_DMASYNC_PREWRITE);
if (sc->sc_cdata.mvgbe_tx_cnt + txmap->dm_nsegs >=
MVGBE_TX_RING_CNT) {
DPRINTFN(2, ("mvgbe_encap: too few descriptors free\n"));
bus_dmamap_unload(sc->sc_dmat, txmap);
return ENOBUFS;
}
DPRINTFN(3, ("mvgbe_encap: dm_nsegs=%d\n", txmap->dm_nsegs));
for (i = 0; i < txmap->dm_nsegs; i++) {
f = &sc->sc_rdata->mvgbe_tx_ring[current];
f->bufptr = H2MVGBE32(txseg[i].ds_addr);
f->bytecnt = H2MVGBE16(txseg[i].ds_len);
if (i != 0)
f->cmdsts = H2MVGBE32(MVGBE_BUFFER_OWNED_BY_DMA);
last = current;
current = MVGBE_TX_RING_NEXT(current);
}
2013-12-23 06:23:25 +04:00
cmdsts = sc->sc_cmdsts_opts;
if (m_csumflags & M_CSUM_IPv4)
cmdsts |= MVGBE_TX_GENERATE_IP_CHKSUM;
if (m_csumflags & M_CSUM_TCPv4)
cmdsts |=
MVGBE_TX_GENERATE_L4_CHKSUM | MVGBE_TX_L4_TYPE_TCP;
if (m_csumflags & M_CSUM_UDPv4)
cmdsts |=
MVGBE_TX_GENERATE_L4_CHKSUM | MVGBE_TX_L4_TYPE_UDP;
if (m_csumflags & (M_CSUM_IPv4 | M_CSUM_TCPv4 | M_CSUM_UDPv4)) {
const int iphdr_unitlen = sizeof(struct ip) / sizeof(uint32_t);
cmdsts |= MVGBE_TX_IP_NO_FRAG |
MVGBE_TX_IP_HEADER_LEN(iphdr_unitlen); /* unit is 4B */
}
if (txmap->dm_nsegs == 1)
f->cmdsts = H2MVGBE32(cmdsts |
MVGBE_TX_ENABLE_INTERRUPT |
MVGBE_TX_ZERO_PADDING |
MVGBE_TX_FIRST_DESC |
MVGBE_TX_LAST_DESC);
else {
f = &sc->sc_rdata->mvgbe_tx_ring[first];
f->cmdsts = H2MVGBE32(cmdsts | MVGBE_TX_FIRST_DESC);
f = &sc->sc_rdata->mvgbe_tx_ring[last];
f->cmdsts = H2MVGBE32(
MVGBE_BUFFER_OWNED_BY_DMA |
MVGBE_TX_ENABLE_INTERRUPT |
MVGBE_TX_ZERO_PADDING |
MVGBE_TX_LAST_DESC);
/* Sync descriptors except first */
MVGBE_CDTXSYNC(sc,
(MVGBE_TX_RING_CNT - 1 == *txidx) ? 0 : (*txidx) + 1,
txmap->dm_nsegs - 1,
BUS_DMASYNC_PREREAD | BUS_DMASYNC_PREWRITE);
}
sc->sc_cdata.mvgbe_tx_chain[last].mvgbe_mbuf = m_head;
SIMPLEQ_REMOVE_HEAD(&sc->sc_txmap_head, link);
sc->sc_cdata.mvgbe_tx_map[last] = entry;
/* Finally, sync first descriptor */
sc->sc_rdata->mvgbe_tx_ring[first].cmdsts |=
H2MVGBE32(MVGBE_BUFFER_OWNED_BY_DMA);
MVGBE_CDTXSYNC(sc, *txidx, 1,
2010-10-02 09:57:42 +04:00
BUS_DMASYNC_PREREAD | BUS_DMASYNC_PREWRITE);
sc->sc_cdata.mvgbe_tx_cnt += i;
*txidx = current;
DPRINTFN(3, ("mvgbe_encap: completed successfully\n"));
return 0;
}
static void
mvgbe_rxeof(struct mvgbe_softc *sc)
{
struct mvgbe_chain_data *cdata = &sc->sc_cdata;
struct mvgbe_rx_desc *cur_rx;
struct ifnet *ifp = &sc->sc_ethercom.ec_if;
struct mbuf *m;
bus_dmamap_t dmamap;
uint32_t rxstat;
uint16_t bufsize;
int idx, cur, total_len;
idx = sc->sc_cdata.mvgbe_rx_prod;
DPRINTFN(3, ("mvgbe_rxeof %d\n", idx));
for (;;) {
cur = idx;
/* Sync the descriptor */
MVGBE_CDRXSYNC(sc, idx,
BUS_DMASYNC_POSTREAD | BUS_DMASYNC_POSTWRITE);
cur_rx = &sc->sc_rdata->mvgbe_rx_ring[idx];
rxstat = MVGBE2H32(cur_rx->cmdsts);
if ((rxstat & MVGBE_BUFFER_OWNED_MASK) ==
MVGBE_BUFFER_OWNED_BY_DMA) {
/* Invalidate the descriptor -- it's not ready yet */
MVGBE_CDRXSYNC(sc, idx, BUS_DMASYNC_PREREAD);
sc->sc_cdata.mvgbe_rx_prod = idx;
break;
}
#ifdef DIAGNOSTIC
if ((rxstat &
(MVGBE_RX_LAST_DESC | MVGBE_RX_FIRST_DESC)) !=
(MVGBE_RX_LAST_DESC | MVGBE_RX_FIRST_DESC))
panic(
"mvgbe_rxeof: buffer size is smaller than packet");
#endif
dmamap = sc->sc_cdata.mvgbe_rx_jumbo_map;
bus_dmamap_sync(sc->sc_dmat, dmamap, 0, dmamap->dm_mapsize,
BUS_DMASYNC_POSTREAD);
m = cdata->mvgbe_rx_chain[idx].mvgbe_mbuf;
cdata->mvgbe_rx_chain[idx].mvgbe_mbuf = NULL;
total_len = MVGBE2H16(cur_rx->bytecnt) - ETHER_CRC_LEN;
bufsize = cur_rx->bufsize;
cdata->mvgbe_rx_map[idx] = NULL;
idx = MVGBE_RX_RING_NEXT(idx);
if (rxstat & MVGBE_ERROR_SUMMARY) {
#if 0
int err = rxstat & MVGBE_RX_ERROR_CODE_MASK;
if (err == MVGBE_RX_CRC_ERROR)
2020-02-03 10:47:53 +03:00
if_statinc(ifp, if_ierrors);
if (err == MVGBE_RX_OVERRUN_ERROR)
2020-02-03 10:47:53 +03:00
if_statinc(ifp, if_ierrors);
if (err == MVGBE_RX_MAX_FRAME_LEN_ERROR)
2020-02-03 10:47:53 +03:00
if_statinc(ifp, if_ierrors);
if (err == MVGBE_RX_RESOURCE_ERROR)
2020-02-03 10:47:53 +03:00
if_statinc(ifp, if_ierrors);
#else
2020-02-03 10:47:53 +03:00
if_statinc(ifp, if_ierrors);
#endif
mvgbe_newbuf(sc, cur, m, dmamap);
continue;
}
if (rxstat & MVGBE_RX_IP_FRAME_TYPE) {
int flgs = 0;
/* Check IPv4 header checksum */
flgs |= M_CSUM_IPv4;
if (!(rxstat & MVGBE_RX_IP_HEADER_OK))
flgs |= M_CSUM_IPv4_BAD;
else if ((bufsize & MVGBE_RX_IP_FRAGMENT) == 0) {
/*
* Check TCPv4/UDPv4 checksum for
* non-fragmented packet only.
*
* It seemd that sometimes
* MVGBE_RX_L4_CHECKSUM_OK bit was set to 0
* even if the checksum is correct and the
* packet was not fragmented. So we don't set
* M_CSUM_TCP_UDP_BAD even if csum bit is 0.
*/
if (((rxstat & MVGBE_RX_L4_TYPE_MASK) ==
MVGBE_RX_L4_TYPE_TCP) &&
((rxstat & MVGBE_RX_L4_CHECKSUM_OK) != 0))
flgs |= M_CSUM_TCPv4;
else if (((rxstat & MVGBE_RX_L4_TYPE_MASK) ==
MVGBE_RX_L4_TYPE_UDP) &&
((rxstat & MVGBE_RX_L4_CHECKSUM_OK) != 0))
flgs |= M_CSUM_UDPv4;
}
m->m_pkthdr.csum_flags = flgs;
}
/*
* Try to allocate a new jumbo buffer. If that
* fails, copy the packet to mbufs and put the
* jumbo buffer back in the ring so it can be
* re-used. If allocating mbufs fails, then we
* have to drop the packet.
*/
if (mvgbe_newbuf(sc, cur, NULL, dmamap) == ENOBUFS) {
struct mbuf *m0;
m0 = m_devget(mtod(m, char *), total_len, 0, ifp);
mvgbe_newbuf(sc, cur, m, dmamap);
if (m0 == NULL) {
aprint_error_ifnet(ifp,
"no receive buffers available --"
" packet dropped!\n");
2020-02-03 10:47:53 +03:00
if_statinc(ifp, if_ierrors);
continue;
}
m = m0;
} else {
m_set_rcvif(m, ifp);
m->m_pkthdr.len = m->m_len = total_len;
}
/* Skip on first 2byte (HW header) */
m_adj(m, MVGBE_HWHEADER_SIZE);
/* pass it on. */
if_percpuq_enqueue(ifp->if_percpuq, m);
}
}
static void
mvgbe_txeof(struct mvgbe_softc *sc)
{
struct mvgbe_chain_data *cdata = &sc->sc_cdata;
struct mvgbe_tx_desc *cur_tx;
struct ifnet *ifp = &sc->sc_ethercom.ec_if;
struct mvgbe_txmap_entry *entry;
uint32_t txstat;
int idx;
DPRINTFN(3, ("mvgbe_txeof\n"));
/*
* Go through our tx ring and free mbufs for those
* frames that have been sent.
*/
idx = cdata->mvgbe_tx_cons;
while (idx != cdata->mvgbe_tx_prod) {
MVGBE_CDTXSYNC(sc, idx, 1,
BUS_DMASYNC_POSTREAD | BUS_DMASYNC_POSTWRITE);
cur_tx = &sc->sc_rdata->mvgbe_tx_ring[idx];
#ifdef MVGBE_DEBUG
if (mvgbe_debug >= 3)
mvgbe_dump_txdesc(cur_tx, idx);
#endif
txstat = MVGBE2H32(cur_tx->cmdsts);
if ((txstat & MVGBE_BUFFER_OWNED_MASK) ==
MVGBE_BUFFER_OWNED_BY_DMA) {
MVGBE_CDTXSYNC(sc, idx, 1, BUS_DMASYNC_PREREAD);
break;
}
if (txstat & MVGBE_TX_LAST_DESC)
2020-02-03 10:47:53 +03:00
if_statinc(ifp, if_opackets);
if (txstat & MVGBE_ERROR_SUMMARY) {
int err = txstat & MVGBE_TX_ERROR_CODE_MASK;
if (err == MVGBE_TX_LATE_COLLISION_ERROR)
2020-02-03 10:47:53 +03:00
if_statinc(ifp, if_collisions);
if (err == MVGBE_TX_UNDERRUN_ERROR)
2020-02-03 10:47:53 +03:00
if_statinc(ifp, if_oerrors);
if (err == MVGBE_TX_EXCESSIVE_COLLISION_ERRO)
2020-02-03 10:47:53 +03:00
if_statinc(ifp, if_collisions);
}
if (cdata->mvgbe_tx_chain[idx].mvgbe_mbuf != NULL) {
entry = cdata->mvgbe_tx_map[idx];
bus_dmamap_sync(sc->sc_dmat, entry->dmamap, 0,
entry->dmamap->dm_mapsize, BUS_DMASYNC_POSTWRITE);
bus_dmamap_unload(sc->sc_dmat, entry->dmamap);
m_freem(cdata->mvgbe_tx_chain[idx].mvgbe_mbuf);
cdata->mvgbe_tx_chain[idx].mvgbe_mbuf = NULL;
SIMPLEQ_INSERT_TAIL(&sc->sc_txmap_head, entry, link);
cdata->mvgbe_tx_map[idx] = NULL;
}
cdata->mvgbe_tx_cnt--;
idx = MVGBE_TX_RING_NEXT(idx);
}
if (cdata->mvgbe_tx_cnt == 0)
ifp->if_timer = 0;
if (cdata->mvgbe_tx_cnt < MVGBE_TX_RING_CNT - 2)
ifp->if_flags &= ~IFF_OACTIVE;
cdata->mvgbe_tx_cons = idx;
}
static uint8_t
mvgbe_crc8(const uint8_t *data, size_t size)
{
int bit;
uint8_t byte;
uint8_t crc = 0;
const uint8_t poly = 0x07;
while (size--)
for (byte = *data++, bit = NBBY-1; bit >= 0; bit--)
crc = (crc << 1) ^ ((((crc >> 7) ^ (byte >> bit)) & 1) ? poly : 0);
return crc;
}
CTASSERT(MVGBE_NDFSMT == MVGBE_NDFOMT);
static void
mvgbe_filter_setup(struct mvgbe_softc *sc)
{
struct ethercom *ec = &sc->sc_ethercom;
struct ifnet *ifp= &sc->sc_ethercom.ec_if;
struct ether_multi *enm;
struct ether_multistep step;
uint32_t dfut[MVGBE_NDFUT], dfsmt[MVGBE_NDFSMT], dfomt[MVGBE_NDFOMT];
uint32_t pxc;
int i;
const uint8_t special[ETHER_ADDR_LEN] = {0x01,0x00,0x5e,0x00,0x00,0x00};
memset(dfut, 0, sizeof(dfut));
memset(dfsmt, 0, sizeof(dfsmt));
memset(dfomt, 0, sizeof(dfomt));
if (ifp->if_flags & (IFF_ALLMULTI | IFF_PROMISC)) {
goto allmulti;
}
ETHER_LOCK(ec);
ETHER_FIRST_MULTI(step, ec, enm);
while (enm != NULL) {
if (memcmp(enm->enm_addrlo, enm->enm_addrhi, ETHER_ADDR_LEN)) {
/* ranges are complex and somewhat rare */
ETHER_UNLOCK(ec);
goto allmulti;
}
/* chip handles some IPv4 multicast specially */
if (memcmp(enm->enm_addrlo, special, 5) == 0) {
i = enm->enm_addrlo[5];
dfsmt[i>>2] |=
MVGBE_DF(i&3, MVGBE_DF_QUEUE(0) | MVGBE_DF_PASS);
} else {
i = mvgbe_crc8(enm->enm_addrlo, ETHER_ADDR_LEN);
dfomt[i>>2] |=
MVGBE_DF(i&3, MVGBE_DF_QUEUE(0) | MVGBE_DF_PASS);
}
ETHER_NEXT_MULTI(step, enm);
}
ETHER_UNLOCK(ec);
goto set;
allmulti:
if (ifp->if_flags & (IFF_ALLMULTI | IFF_PROMISC)) {
for (i = 0; i < MVGBE_NDFSMT; i++) {
dfsmt[i] = dfomt[i] =
MVGBE_DF(0, MVGBE_DF_QUEUE(0) | MVGBE_DF_PASS) |
MVGBE_DF(1, MVGBE_DF_QUEUE(0) | MVGBE_DF_PASS) |
MVGBE_DF(2, MVGBE_DF_QUEUE(0) | MVGBE_DF_PASS) |
MVGBE_DF(3, MVGBE_DF_QUEUE(0) | MVGBE_DF_PASS);
}
}
set:
pxc = MVGBE_READ(sc, MVGBE_PXC);
pxc &= ~MVGBE_PXC_UPM;
pxc |= MVGBE_PXC_RB | MVGBE_PXC_RBIP | MVGBE_PXC_RBARP;
if (ifp->if_flags & IFF_BROADCAST) {
pxc &= ~(MVGBE_PXC_RB | MVGBE_PXC_RBIP | MVGBE_PXC_RBARP);
}
if (ifp->if_flags & IFF_PROMISC) {
pxc |= MVGBE_PXC_UPM;
}
MVGBE_WRITE(sc, MVGBE_PXC, pxc);
/* Set Destination Address Filter Unicast Table */
if (ifp->if_flags & IFF_PROMISC) {
/* pass all unicast addresses */
for (i = 0; i < MVGBE_NDFUT; i++) {
dfut[i] =
MVGBE_DF(0, MVGBE_DF_QUEUE(0) | MVGBE_DF_PASS) |
MVGBE_DF(1, MVGBE_DF_QUEUE(0) | MVGBE_DF_PASS) |
MVGBE_DF(2, MVGBE_DF_QUEUE(0) | MVGBE_DF_PASS) |
MVGBE_DF(3, MVGBE_DF_QUEUE(0) | MVGBE_DF_PASS);
}
} else {
i = sc->sc_enaddr[5] & 0xf; /* last nibble */
dfut[i>>2] = MVGBE_DF(i&3, MVGBE_DF_QUEUE(0) | MVGBE_DF_PASS);
}
MVGBE_WRITE_FILTER(sc, MVGBE_DFUT, dfut, MVGBE_NDFUT);
/* Set Destination Address Filter Multicast Tables */
MVGBE_WRITE_FILTER(sc, MVGBE_DFSMT, dfsmt, MVGBE_NDFSMT);
MVGBE_WRITE_FILTER(sc, MVGBE_DFOMT, dfomt, MVGBE_NDFOMT);
}
#ifdef MVGBE_DEBUG
static void
mvgbe_dump_txdesc(struct mvgbe_tx_desc *desc, int idx)
{
#define DESC_PRINT(X) \
if (X) \
printf("txdesc[%d]." #X "=%#x\n", idx, X);
#ifdef MVGBE_BIG_ENDIAN
DESC_PRINT(desc->bytecnt);
DESC_PRINT(desc->l4ichk);
DESC_PRINT(desc->cmdsts);
DESC_PRINT(desc->nextdescptr);
DESC_PRINT(desc->bufptr);
#else
DESC_PRINT(MVGBE2H32(desc->cmdsts));
DESC_PRINT(MVGBE2H16(desc->l4ichk));
DESC_PRINT(MVGBE2H16(desc->bytecnt));
DESC_PRINT(MVGBE2H32(desc->bufptr));
DESC_PRINT(MVGBE2H32(desc->nextdescptr));
#endif
#undef DESC_PRINT
}
#endif
SYSCTL_SETUP(sysctl_mvgbe, "sysctl mvgbe subtree setup")
{
int rc;
const struct sysctlnode *node;
if ((rc = sysctl_createv(clog, 0, NULL, &node,
0, CTLTYPE_NODE, "mvgbe",
SYSCTL_DESCR("mvgbe interface controls"),
NULL, 0, NULL, 0, CTL_HW, CTL_CREATE, CTL_EOL)) != 0) {
goto err;
}
mvgbe_root_num = node->sysctl_num;
return;
err:
aprint_error("%s: syctl_createv failed (rc = %d)\n", __func__, rc);
}
static void
sysctl_mvgbe_init(struct mvgbe_softc *sc)
{
const struct sysctlnode *node;
int mvgbe_nodenum;
if (sysctl_createv(&sc->mvgbe_clog, 0, NULL, &node,
0, CTLTYPE_NODE, device_xname(sc->sc_dev),
SYSCTL_DESCR("mvgbe per-controller controls"),
NULL, 0, NULL, 0, CTL_HW, mvgbe_root_num, CTL_CREATE,
CTL_EOL) != 0) {
aprint_normal_dev(sc->sc_dev, "couldn't create sysctl node\n");
return;
}
mvgbe_nodenum = node->sysctl_num;
/* interrupt moderation sysctls */
if (sysctl_createv(&sc->mvgbe_clog, 0, NULL, &node,
CTLFLAG_READWRITE, CTLTYPE_INT, "ipginttx",
SYSCTL_DESCR("mvgbe TX interrupt moderation timer"),
mvgbe_sysctl_ipginttx, 0, (void *)sc,
0, CTL_HW, mvgbe_root_num, mvgbe_nodenum, CTL_CREATE,
CTL_EOL) != 0) {
aprint_normal_dev(sc->sc_dev,
"couldn't create ipginttx sysctl node\n");
}
if (sysctl_createv(&sc->mvgbe_clog, 0, NULL, &node,
CTLFLAG_READWRITE, CTLTYPE_INT, "ipgintrx",
SYSCTL_DESCR("mvgbe RX interrupt moderation timer"),
mvgbe_sysctl_ipgintrx, 0, (void *)sc,
0, CTL_HW, mvgbe_root_num, mvgbe_nodenum, CTL_CREATE,
CTL_EOL) != 0) {
aprint_normal_dev(sc->sc_dev,
"couldn't create ipginttx sysctl node\n");
}
}
static int
mvgbe_sysctl_ipginttx(SYSCTLFN_ARGS)
{
int error;
unsigned int t;
struct sysctlnode node;
struct mvgbec_softc *csc;
struct mvgbe_softc *sc;
node = *rnode;
sc = node.sysctl_data;
csc = device_private(device_parent(sc->sc_dev));
t = sc->sc_ipginttx;
node.sysctl_data = &t;
error = sysctl_lookup(SYSCTLFN_CALL(&node));
if (error || newp == NULL)
return error;
if (mvgbe_ipginttx(csc, sc, t) < 0)
return EINVAL;
/*
* update the softc with sysctl-changed value, and mark
* for hardware update
*/
sc->sc_ipginttx = t;
return 0;
}
static int
mvgbe_sysctl_ipgintrx(SYSCTLFN_ARGS)
{
int error;
unsigned int t;
struct sysctlnode node;
struct mvgbec_softc *csc;
struct mvgbe_softc *sc;
node = *rnode;
sc = node.sysctl_data;
csc = device_private(device_parent(sc->sc_dev));
t = sc->sc_ipgintrx;
node.sysctl_data = &t;
error = sysctl_lookup(SYSCTLFN_CALL(&node));
if (error || newp == NULL)
return error;
if (mvgbe_ipgintrx(csc, sc, t) < 0)
return EINVAL;
/*
* update the softc with sysctl-changed value, and mark
* for hardware update
*/
sc->sc_ipgintrx = t;
return 0;
}