diff options
Diffstat (limited to 'sys/dev/sbus/qe.c')
-rw-r--r-- | sys/dev/sbus/qe.c | 1262 |
1 files changed, 1262 insertions, 0 deletions
diff --git a/sys/dev/sbus/qe.c b/sys/dev/sbus/qe.c new file mode 100644 index 00000000000..f5a0c4491be --- /dev/null +++ b/sys/dev/sbus/qe.c @@ -0,0 +1,1262 @@ +/* $OpenBSD: qe.c,v 1.1 2001/08/20 22:09:27 jason Exp $ */ +/* $NetBSD: qe.c,v 1.16 2001/03/30 17:30:18 christos Exp $ */ + +/*- + * Copyright (c) 1999 The NetBSD Foundation, Inc. + * All rights reserved. + * + * This code is derived from software contributed to The NetBSD Foundation + * by Paul Kranenburg. + * + * Redistribution and use in source and binary forms, with or without + * modification, are permitted provided that the following conditions + * are met: + * 1. Redistributions of source code must retain the above copyright + * notice, this list of conditions and the following disclaimer. + * 2. Redistributions in binary form must reproduce the above copyright + * notice, this list of conditions and the following disclaimer in the + * documentation and/or other materials provided with the distribution. + * 3. All advertising materials mentioning features or use of this software + * must display the following acknowledgement: + * This product includes software developed by the NetBSD + * Foundation, Inc. and its contributors. + * 4. Neither the name of The NetBSD Foundation nor the names of its + * contributors may be used to endorse or promote products derived + * from this software without specific prior written permission. + * + * THIS SOFTWARE IS PROVIDED BY THE NETBSD FOUNDATION, INC. AND CONTRIBUTORS + * ``AS IS'' AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED + * TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR + * PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE FOUNDATION OR CONTRIBUTORS + * BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR + * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF + * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS + * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN + * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) + * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE + * POSSIBILITY OF SUCH DAMAGE. + */ + +/* + * Copyright (c) 1998 Jason L. Wright. + * All rights reserved. + * + * Redistribution and use in source and binary forms, with or without + * modification, are permitted provided that the following conditions + * are met: + * 1. Redistributions of source code must retain the above copyright + * notice, this list of conditions and the following disclaimer. + * 2. Redistributions in binary form must reproduce the above copyright + * notice, this list of conditions and the following disclaimer in the + * documentation and/or other materials provided with the distribution. + * 3. The name of the authors may not be used to endorse or promote products + * derived from this software without specific prior written permission. + * + * THIS SOFTWARE IS PROVIDED BY THE AUTHORS ``AS IS'' AND ANY EXPRESS OR + * IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES + * OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED. + * IN NO EVENT SHALL THE AUTHORS BE LIABLE FOR ANY DIRECT, INDIRECT, + * INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT + * NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, + * DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY + * THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT + * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF + * THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. + */ + +/* + * Driver for the SBus qec+qe QuadEthernet board. + * + * This driver was written using the AMD MACE Am79C940 documentation, some + * ideas gleaned from the S/Linux driver for this card, Solaris header files, + * and a loan of a card from Paul Southworth of the Internet Engineering + * Group (www.ieng.com). + */ + +#define QEDEBUG + +#include "bpfilter.h" + +#include <sys/param.h> +#include <sys/systm.h> +#include <sys/kernel.h> +#include <sys/errno.h> +#include <sys/ioctl.h> +#include <sys/mbuf.h> +#include <sys/socket.h> +#include <sys/syslog.h> +#include <sys/device.h> +#include <sys/malloc.h> + +#include <net/if.h> +#include <net/if_dl.h> +#include <net/if_types.h> +#include <net/netisr.h> +#include <net/if_media.h> + +#ifdef INET +#include <netinet/in.h> +#include <netinet/in_systm.h> +#include <netinet/in_var.h> +#include <netinet/ip.h> +#include <netinet/if_ether.h> +#endif + +#ifdef NS +#include <netns/ns.h> +#include <netns/ns_if.h> +#endif + +#if NBPFILTER > 0 +#include <net/bpf.h> +#include <net/bpfdesc.h> +#endif + +#include <machine/bus.h> +#include <machine/intr.h> +#include <machine/autoconf.h> + +#include <dev/sbus/sbusvar.h> +#include <dev/sbus/qecreg.h> +#include <dev/sbus/qecvar.h> +#include <dev/sbus/qereg.h> + +struct qe_softc { + struct device sc_dev; /* base device */ + struct sbusdev sc_sd; /* sbus device */ + bus_space_tag_t sc_bustag; /* bus & dma tags */ + bus_dma_tag_t sc_dmatag; + bus_dmamap_t sc_dmamap; + struct arpcom sc_arpcom; + struct ifmedia sc_ifmedia; /* interface media */ + + struct qec_softc *sc_qec; /* QEC parent */ + + bus_space_handle_t sc_qr; /* QEC registers */ + bus_space_handle_t sc_mr; /* MACE registers */ + bus_space_handle_t sc_cr; /* channel registers */ + + int sc_channel; /* channel number */ + u_int sc_rev; /* board revision */ + + int sc_burst; + + struct qec_ring sc_rb; /* Packet Ring Buffer */ + + /* MAC address */ + u_int8_t sc_enaddr[6]; + +#ifdef QEDEBUG + int sc_debug; +#endif +}; + +int qematch __P((struct device *, void *, void *)); +void qeattach __P((struct device *, struct device *, void *)); + +void qeinit __P((struct qe_softc *)); +void qestart __P((struct ifnet *)); +void qestop __P((struct qe_softc *)); +void qewatchdog __P((struct ifnet *)); +int qeioctl __P((struct ifnet *, u_long, caddr_t)); +void qereset __P((struct qe_softc *)); + +int qeintr __P((void *)); +int qe_eint __P((struct qe_softc *, u_int32_t)); +int qe_rint __P((struct qe_softc *)); +int qe_tint __P((struct qe_softc *)); +void qe_mcreset __P((struct qe_softc *)); + +static int qe_put __P((struct qe_softc *, int, struct mbuf *)); +static void qe_read __P((struct qe_softc *, int, int)); +static struct mbuf *qe_get __P((struct qe_softc *, int, int)); + +/* ifmedia callbacks */ +void qe_ifmedia_sts __P((struct ifnet *, struct ifmediareq *)); +int qe_ifmedia_upd __P((struct ifnet *)); + +struct cfattach qe_ca = { + sizeof(struct qe_softc), qematch, qeattach +}; + +struct cfdriver qe_cd = { + NULL, "qe", DV_IFNET +}; + +int +qematch(parent, vcf, aux) + struct device *parent; + void *vcf; + void *aux; +{ + struct cfdata *cf = vcf; + struct sbus_attach_args *sa = aux; + + return (strcmp(cf->cf_driver->cd_name, sa->sa_name) == 0); +} + +void +qeattach(parent, self, aux) + struct device *parent, *self; + void *aux; +{ + struct sbus_attach_args *sa = aux; + struct qec_softc *qec = (struct qec_softc *)parent; + struct qe_softc *sc = (struct qe_softc *)self; + struct ifnet *ifp = &sc->sc_arpcom.ac_if; + int node = sa->sa_node; + bus_dma_tag_t dmatag = sa->sa_dmatag; + bus_dma_segment_t seg; + bus_size_t size; + int rseg, error; + struct bootpath *bp; + extern void myetheraddr __P((u_char *)); + + if (sa->sa_nreg < 2) { + printf("%s: only %d register sets\n", + self->dv_xname, sa->sa_nreg); + return; + } + + if (bus_space_map2(sa->sa_bustag, + (bus_type_t)sa->sa_reg[0].sbr_slot, + (bus_addr_t)sa->sa_reg[0].sbr_offset, + (bus_size_t)sa->sa_reg[0].sbr_size, + BUS_SPACE_MAP_LINEAR, 0, &sc->sc_cr) != 0) { + printf("%s: cannot map registers\n", self->dv_xname); + return; + } + + if (bus_space_map2(sa->sa_bustag, + (bus_type_t)sa->sa_reg[1].sbr_slot, + (bus_addr_t)sa->sa_reg[1].sbr_offset, + (bus_size_t)sa->sa_reg[1].sbr_size, + BUS_SPACE_MAP_LINEAR, 0, &sc->sc_mr) != 0) { + printf("%s: cannot map registers\n", self->dv_xname); + return; + } + + sc->sc_rev = getpropint(node, "mace-version", -1); + printf(" rev %x", sc->sc_rev); + + sc->sc_qec = qec; + sc->sc_qr = qec->sc_regs; + + sc->sc_channel = getpropint(node, "channel#", -1); + sc->sc_burst = qec->sc_burst; + + qestop(sc); + + /* Note: no interrupt level passed */ + (void)bus_intr_establish(sa->sa_bustag, 0, IPL_NET, 0, qeintr, sc); + myetheraddr(sc->sc_enaddr); + + /* + * Allocate descriptor ring and buffers. + */ + + /* for now, allocate as many bufs as there are ring descriptors */ + sc->sc_rb.rb_ntbuf = QEC_XD_RING_MAXSIZE; + sc->sc_rb.rb_nrbuf = QEC_XD_RING_MAXSIZE; + + size = QEC_XD_RING_MAXSIZE * sizeof(struct qec_xd) + + QEC_XD_RING_MAXSIZE * sizeof(struct qec_xd) + + sc->sc_rb.rb_ntbuf * QE_PKT_BUF_SZ + + sc->sc_rb.rb_nrbuf * QE_PKT_BUF_SZ; + + /* Get a DMA handle */ + if ((error = bus_dmamap_create(dmatag, size, 1, size, 0, + BUS_DMA_NOWAIT, &sc->sc_dmamap)) != 0) { + printf("%s: DMA map create error %d\n", self->dv_xname, error); + return; + } + + /* Allocate DMA buffer */ + if ((error = bus_dmamem_alloc(dmatag, size, 0, 0, + &seg, 1, &rseg, BUS_DMA_NOWAIT)) != 0) { + printf("%s: DMA buffer alloc error %d\n", + self->dv_xname, error); + return; + } + sc->sc_rb.rb_dmabase = sc->sc_dmamap->dm_segs[0].ds_addr; + + /* Map DMA buffer in CPU addressable space */ + if ((error = bus_dmamem_map(dmatag, &seg, rseg, size, + &sc->sc_rb.rb_membase, + BUS_DMA_NOWAIT|BUS_DMA_COHERENT)) != 0) { + printf("%s: DMA buffer map error %d\n", + self->dv_xname, error); + bus_dmamem_free(dmatag, &seg, rseg); + return; + } + + /* Load the buffer */ + if ((error = bus_dmamap_load(dmatag, sc->sc_dmamap, + sc->sc_rb.rb_membase, size, NULL, + BUS_DMA_NOWAIT)) != 0) { + printf("%s: DMA buffer map load error %d\n", + self->dv_xname, error); + bus_dmamem_unmap(dmatag, sc->sc_rb.rb_membase, size); + bus_dmamem_free(dmatag, &seg, rseg); + return; + } + + /* Initialize media properties */ + ifmedia_init(&sc->sc_ifmedia, 0, qe_ifmedia_upd, qe_ifmedia_sts); + ifmedia_add(&sc->sc_ifmedia, + IFM_MAKEWORD(IFM_ETHER,IFM_10_T,0,0), + 0, NULL); + ifmedia_add(&sc->sc_ifmedia, + IFM_MAKEWORD(IFM_ETHER,IFM_10_5,0,0), + 0, NULL); + ifmedia_add(&sc->sc_ifmedia, + IFM_MAKEWORD(IFM_ETHER,IFM_AUTO,0,0), + 0, NULL); + ifmedia_set(&sc->sc_ifmedia, IFM_ETHER|IFM_AUTO); + + bcopy(sc->sc_dev.dv_xname, ifp->if_xname, IFNAMSIZ); + ifp->if_softc = sc; + ifp->if_start = qestart; + ifp->if_ioctl = qeioctl; + ifp->if_watchdog = qewatchdog; + ifp->if_flags = IFF_BROADCAST | IFF_SIMPLEX | IFF_NOTRAILERS | + IFF_MULTICAST; + IFQ_SET_READY(&ifp->if_snd); + + /* Attach the interface. */ + if_attach(ifp); + ether_ifattach(ifp); + + printf(" address %s\n", ether_sprintf(sc->sc_enaddr)); + + bp = sa->sa_bp; + if (bp != NULL && strcmp(bp->name, qe_cd.cd_name) == 0 && + sc->sc_dev.dv_unit == bp->val[1]) + bp->dev = &sc->sc_dev; +} + +/* + * Pull data off an interface. + * Len is the length of data, with local net header stripped. + * We copy the data into mbufs. When full cluster sized units are present, + * we copy into clusters. + */ +static __inline__ struct mbuf * +qe_get(sc, idx, totlen) + struct qe_softc *sc; + int idx, totlen; +{ + struct ifnet *ifp = &sc->sc_arpcom.ac_if; + struct mbuf *m; + struct mbuf *top, **mp; + int len, pad, boff = 0; + caddr_t bp; + + bp = sc->sc_rb.rb_rxbuf + (idx % sc->sc_rb.rb_nrbuf) * QE_PKT_BUF_SZ; + + MGETHDR(m, M_DONTWAIT, MT_DATA); + if (m == NULL) + return (NULL); + m->m_pkthdr.rcvif = ifp; + m->m_pkthdr.len = totlen; + pad = ALIGN(sizeof(struct ether_header)) - sizeof(struct ether_header); + m->m_data += pad; + len = MHLEN - pad; + top = NULL; + mp = ⊤ + + while (totlen > 0) { + if (top) { + MGET(m, M_DONTWAIT, MT_DATA); + if (m == NULL) { + m_freem(top); + return (NULL); + } + len = MLEN; + } + if (top && totlen >= MINCLSIZE) { + MCLGET(m, M_DONTWAIT); + if (m->m_flags & M_EXT) + len = MCLBYTES; + } + m->m_len = len = min(totlen, len); + bcopy(bp + boff, mtod(m, caddr_t), len); + boff += len; + totlen -= len; + *mp = m; + mp = &m->m_next; + } + + return (top); +} + +/* + * Routine to copy from mbuf chain to transmit buffer in + * network buffer memory. + */ +__inline__ int +qe_put(sc, idx, m) + struct qe_softc *sc; + int idx; + struct mbuf *m; +{ + struct mbuf *n; + int len, tlen = 0, boff = 0; + caddr_t bp; + + bp = sc->sc_rb.rb_txbuf + (idx % sc->sc_rb.rb_ntbuf) * QE_PKT_BUF_SZ; + + for (; m; m = n) { + len = m->m_len; + if (len == 0) { + MFREE(m, n); + continue; + } + bcopy(mtod(m, caddr_t), bp+boff, len); + boff += len; + tlen += len; + MFREE(m, n); + } + return (tlen); +} + +/* + * Pass a packet to the higher levels. + */ +__inline__ void +qe_read(sc, idx, len) + struct qe_softc *sc; + int idx, len; +{ + struct ifnet *ifp = &sc->sc_arpcom.ac_if; + struct mbuf *m; + + if (len <= sizeof(struct ether_header) || + len > ETHERMTU + sizeof(struct ether_header)) { + + printf("%s: invalid packet size %d; dropping\n", + ifp->if_xname, len); + + ifp->if_ierrors++; + return; + } + + /* + * Pull packet off interface. + */ + m = qe_get(sc, idx, len); + if (m == NULL) { + ifp->if_ierrors++; + return; + } + ifp->if_ipackets++; + +#if NBPFILTER > 0 + /* + * Check if there's a BPF listener on this interface. + * If so, hand off the raw packet to BPF. + */ + if (ifp->if_bpf) + bpf_mtap(ifp->if_bpf, m); +#endif + /* Pass the packet up. */ + ether_input_mbuf(ifp, m); +} + +/* + * Start output on interface. + * We make two assumptions here: + * 1) that the current priority is set to splnet _before_ this code + * is called *and* is returned to the appropriate priority after + * return + * 2) that the IFF_OACTIVE flag is checked before this code is called + * (i.e. that the output part of the interface is idle) + */ +void +qestart(ifp) + struct ifnet *ifp; +{ + struct qe_softc *sc = (struct qe_softc *)ifp->if_softc; + struct qec_xd *txd = sc->sc_rb.rb_txd; + struct mbuf *m; + unsigned int bix, len; + unsigned int ntbuf = sc->sc_rb.rb_ntbuf; + + if ((ifp->if_flags & (IFF_RUNNING | IFF_OACTIVE)) != IFF_RUNNING) + return; + + bix = sc->sc_rb.rb_tdhead; + + for (;;) { + IFQ_DEQUEUE(&ifp->if_snd, m); + if (m == 0) + break; + +#if NBPFILTER > 0 + /* + * If BPF is listening on this interface, let it see the + * packet before we commit it to the wire. + */ + if (ifp->if_bpf) + bpf_mtap(ifp->if_bpf, m); +#endif + + /* + * Copy the mbuf chain into the transmit buffer. + */ + len = qe_put(sc, bix, m); + + /* + * Initialize transmit registers and start transmission + */ + txd[bix].xd_flags = QEC_XD_OWN | QEC_XD_SOP | QEC_XD_EOP | + (len & QEC_XD_LENGTH); + bus_space_write_4(sc->sc_bustag, sc->sc_cr, QE_CRI_CTRL, + QE_CR_CTRL_TWAKEUP); + + if (++bix == QEC_XD_RING_MAXSIZE) + bix = 0; + + if (++sc->sc_rb.rb_td_nbusy == ntbuf) { + ifp->if_flags |= IFF_OACTIVE; + break; + } + } + + sc->sc_rb.rb_tdhead = bix; +} + +void +qestop(sc) + struct qe_softc *sc; +{ + bus_space_tag_t t = sc->sc_bustag; + bus_space_handle_t mr = sc->sc_mr; + bus_space_handle_t cr = sc->sc_cr; + int n; + +#if defined(SUN4U) || defined(__GNUC__) + (void)&t; +#endif + /* Stop the schwurst */ + bus_space_write_1(t, mr, QE_MRI_BIUCC, QE_MR_BIUCC_SWRST); + for (n = 200; n > 0; n--) { + if ((bus_space_read_1(t, mr, QE_MRI_BIUCC) & + QE_MR_BIUCC_SWRST) == 0) + break; + DELAY(20); + } + + /* then reset */ + bus_space_write_4(t, cr, QE_CRI_CTRL, QE_CR_CTRL_RESET); + for (n = 200; n > 0; n--) { + if ((bus_space_read_4(t, cr, QE_CRI_CTRL) & + QE_CR_CTRL_RESET) == 0) + break; + DELAY(20); + } +} + +/* + * Reset interface. + */ +void +qereset(sc) + struct qe_softc *sc; +{ + int s; + + s = splnet(); + qestop(sc); + qeinit(sc); + splx(s); +} + +void +qewatchdog(ifp) + struct ifnet *ifp; +{ + struct qe_softc *sc = ifp->if_softc; + + log(LOG_ERR, "%s: device timeout\n", sc->sc_dev.dv_xname); + ifp->if_oerrors++; + + qereset(sc); +} + +/* + * Interrupt dispatch. + */ +int +qeintr(arg) + void *arg; +{ + struct qe_softc *sc = (struct qe_softc *)arg; + bus_space_tag_t t = sc->sc_bustag; + u_int32_t qecstat, qestat; + int r = 0; + +#if defined(SUN4U) || defined(__GNUC__) + (void)&t; +#endif + /* Read QEC status and channel status */ + qecstat = bus_space_read_4(t, sc->sc_qr, QEC_QRI_STAT); +#ifdef QEDEBUG + if (sc->sc_debug) { + printf("qe%d: intr: qecstat=%x\n", sc->sc_channel, qecstat); + } +#endif + + /* Filter out status for this channel */ + qecstat = qecstat >> (4 * sc->sc_channel); + if ((qecstat & 0xf) == 0) + return (r); + + qestat = bus_space_read_4(t, sc->sc_cr, QE_CRI_STAT); + +#ifdef QEDEBUG + if (sc->sc_debug) { + int i; + bus_space_tag_t t = sc->sc_bustag; + bus_space_handle_t mr = sc->sc_mr; + + printf("qe%d: intr: qestat=%b\n", sc->sc_channel, + qestat, QE_CR_STAT_BITS); + + printf("MACE registers:\n"); + for (i = 0 ; i < 32; i++) { + printf(" m[%d]=%x,", i, bus_space_read_1(t, mr, i)); + if (((i+1) & 7) == 0) + printf("\n"); + } + } +#endif + + if (qestat & QE_CR_STAT_ALLERRORS) { +#ifdef QEDEBUG + if (sc->sc_debug) + printf("qe%d: eint: qestat=%b\n", sc->sc_channel, + qestat, QE_CR_STAT_BITS); +#endif + r |= qe_eint(sc, qestat); + if (r == -1) + return (1); + } + + if (qestat & QE_CR_STAT_TXIRQ) + r |= qe_tint(sc); + + if (qestat & QE_CR_STAT_RXIRQ) + r |= qe_rint(sc); + + return (r); +} + +/* + * Transmit interrupt. + */ +int +qe_tint(sc) + struct qe_softc *sc; +{ + struct ifnet *ifp = &sc->sc_arpcom.ac_if; + unsigned int bix, txflags; + + bix = sc->sc_rb.rb_tdtail; + + for (;;) { + if (sc->sc_rb.rb_td_nbusy <= 0) + break; + + txflags = sc->sc_rb.rb_txd[bix].xd_flags; + + if (txflags & QEC_XD_OWN) + break; + + ifp->if_flags &= ~IFF_OACTIVE; + ifp->if_opackets++; + + if (++bix == QEC_XD_RING_MAXSIZE) + bix = 0; + + --sc->sc_rb.rb_td_nbusy; + } + + sc->sc_rb.rb_tdtail = bix; + + qestart(ifp); + + if (sc->sc_rb.rb_td_nbusy == 0) + ifp->if_timer = 0; + + return (1); +} + +/* + * Receive interrupt. + */ +int +qe_rint(sc) + struct qe_softc *sc; +{ + struct qec_xd *xd = sc->sc_rb.rb_rxd; + unsigned int bix, len; + unsigned int nrbuf = sc->sc_rb.rb_nrbuf; +#ifdef QEDEBUG + int npackets = 0; +#endif + + bix = sc->sc_rb.rb_rdtail; + + /* + * Process all buffers with valid data. + */ + for (;;) { + len = xd[bix].xd_flags; + if (len & QEC_XD_OWN) + break; + +#ifdef QEDEBUG + npackets++; +#endif + + len &= QEC_XD_LENGTH; + len -= 4; + qe_read(sc, bix, len); + + /* ... */ + xd[(bix+nrbuf) % QEC_XD_RING_MAXSIZE].xd_flags = + QEC_XD_OWN | (QE_PKT_BUF_SZ & QEC_XD_LENGTH); + + if (++bix == QEC_XD_RING_MAXSIZE) + bix = 0; + } +#ifdef QEDEBUG + if (npackets == 0 && sc->sc_debug) + printf("%s: rint: no packets; rb index %d; status 0x%x\n", + sc->sc_dev.dv_xname, bix, len); +#endif + + sc->sc_rb.rb_rdtail = bix; + + return (1); +} + +/* + * Error interrupt. + */ +int +qe_eint(sc, why) + struct qe_softc *sc; + u_int32_t why; +{ + struct ifnet *ifp = &sc->sc_arpcom.ac_if; + int r = 0, rst = 0; + + if (why & QE_CR_STAT_EDEFER) { + printf("%s: excessive tx defers.\n", sc->sc_dev.dv_xname); + r |= 1; + ifp->if_oerrors++; + } + + if (why & QE_CR_STAT_CLOSS) { + printf("%s: no carrier, link down?\n", sc->sc_dev.dv_xname); + ifp->if_oerrors++; + r |= 1; + } + + if (why & QE_CR_STAT_ERETRIES) { + printf("%s: excessive tx retries\n", sc->sc_dev.dv_xname); + ifp->if_oerrors++; + r |= 1; + rst = 1; + } + + + if (why & QE_CR_STAT_LCOLL) { + printf("%s: late tx transmission\n", sc->sc_dev.dv_xname); + ifp->if_oerrors++; + r |= 1; + rst = 1; + } + + if (why & QE_CR_STAT_FUFLOW) { + printf("%s: tx fifo underflow\n", sc->sc_dev.dv_xname); + ifp->if_oerrors++; + r |= 1; + rst = 1; + } + + if (why & QE_CR_STAT_JERROR) { + printf("%s: jabber seen\n", sc->sc_dev.dv_xname); + r |= 1; + } + + if (why & QE_CR_STAT_BERROR) { + printf("%s: babble seen\n", sc->sc_dev.dv_xname); + r |= 1; + } + + if (why & QE_CR_STAT_TCCOFLOW) { + ifp->if_collisions += 256; + ifp->if_oerrors += 256; + r |= 1; + } + + if (why & QE_CR_STAT_TXDERROR) { + printf("%s: tx descriptor is bad\n", sc->sc_dev.dv_xname); + rst = 1; + r |= 1; + } + + if (why & QE_CR_STAT_TXLERR) { + printf("%s: tx late error\n", sc->sc_dev.dv_xname); + ifp->if_oerrors++; + rst = 1; + r |= 1; + } + + if (why & QE_CR_STAT_TXPERR) { + printf("%s: tx dma parity error\n", sc->sc_dev.dv_xname); + ifp->if_oerrors++; + rst = 1; + r |= 1; + } + + if (why & QE_CR_STAT_TXSERR) { + printf("%s: tx dma sbus error ack\n", sc->sc_dev.dv_xname); + ifp->if_oerrors++; + rst = 1; + r |= 1; + } + + if (why & QE_CR_STAT_RCCOFLOW) { + ifp->if_collisions += 256; + ifp->if_ierrors += 256; + r |= 1; + } + + if (why & QE_CR_STAT_RUOFLOW) { + ifp->if_ierrors += 256; + r |= 1; + } + + if (why & QE_CR_STAT_MCOFLOW) { + ifp->if_ierrors += 256; + r |= 1; + } + + if (why & QE_CR_STAT_RXFOFLOW) { + printf("%s: rx fifo overflow\n", sc->sc_dev.dv_xname); + ifp->if_ierrors++; + r |= 1; + } + + if (why & QE_CR_STAT_RLCOLL) { + printf("%s: rx late collision\n", sc->sc_dev.dv_xname); + ifp->if_ierrors++; + ifp->if_collisions++; + r |= 1; + } + + if (why & QE_CR_STAT_FCOFLOW) { + ifp->if_ierrors += 256; + r |= 1; + } + + if (why & QE_CR_STAT_CECOFLOW) { + ifp->if_ierrors += 256; + r |= 1; + } + + if (why & QE_CR_STAT_RXDROP) { + printf("%s: rx packet dropped\n", sc->sc_dev.dv_xname); + ifp->if_ierrors++; + r |= 1; + } + + if (why & QE_CR_STAT_RXSMALL) { + printf("%s: rx buffer too small\n", sc->sc_dev.dv_xname); + ifp->if_ierrors++; + r |= 1; + rst = 1; + } + + if (why & QE_CR_STAT_RXLERR) { + printf("%s: rx late error\n", sc->sc_dev.dv_xname); + ifp->if_ierrors++; + r |= 1; + rst = 1; + } + + if (why & QE_CR_STAT_RXPERR) { + printf("%s: rx dma parity error\n", sc->sc_dev.dv_xname); + ifp->if_ierrors++; + r |= 1; + rst = 1; + } + + if (why & QE_CR_STAT_RXSERR) { + printf("%s: rx dma sbus error ack\n", sc->sc_dev.dv_xname); + ifp->if_ierrors++; + r |= 1; + rst = 1; + } + + if (r == 0) + printf("%s: unexpected interrupt error: %08x\n", + sc->sc_dev.dv_xname, why); + + if (rst) { + printf("%s: resetting...\n", sc->sc_dev.dv_xname); + qereset(sc); + return (-1); + } + + return (r); +} + +int +qeioctl(ifp, cmd, data) + struct ifnet *ifp; + u_long cmd; + caddr_t data; +{ + struct qe_softc *sc = ifp->if_softc; + struct ifaddr *ifa = (struct ifaddr *)data; + struct ifreq *ifr = (struct ifreq *)data; + int s, error = 0; + + s = splnet(); + + switch (cmd) { + case SIOCSIFADDR: + ifp->if_flags |= IFF_UP; + switch (ifa->ifa_addr->sa_family) { +#ifdef INET + case AF_INET: + qeinit(sc); + arp_ifinit(&sc->sc_arpcom, ifa); + break; +#endif /* INET */ +#ifdef NS + case AF_NS: + { + struct ns_addr *ina = &IA_SNS(ifa)->sns_addr; + + if (ns_nullhost(*ina)) + ina->x_host = + *(union ns_host *)LLADDR(ifp->if_sadl); + else + bcopy(ina->x_host.c_host, LLADDR(ifp->if_sadl), + sizeof(sc->sc_enaddr)); + /* Set new address. */ + qeinit(sc); + break; + } +#endif /* NS */ + default: + qeinit(sc); + break; + } + break; + + case SIOCSIFFLAGS: + if ((ifp->if_flags & IFF_UP) == 0 && + (ifp->if_flags & IFF_RUNNING) != 0) { + /* + * If interface is marked down and it is running, then + * stop it. + */ + qestop(sc); + ifp->if_flags &= ~IFF_RUNNING; + + } else if ((ifp->if_flags & IFF_UP) != 0 && + (ifp->if_flags & IFF_RUNNING) == 0) { + /* + * If interface is marked up and it is stopped, then + * start it. + */ + qeinit(sc); + + } else { + /* + * Reset the interface to pick up changes in any other + * flags that affect hardware registers. + */ + qestop(sc); + qeinit(sc); + } +#ifdef QEDEBUG + sc->sc_debug = (ifp->if_flags & IFF_DEBUG) != 0 ? 1 : 0; +#endif + break; + + case SIOCADDMULTI: + case SIOCDELMULTI: + error = (cmd == SIOCADDMULTI) ? + ether_addmulti(ifr, &sc->sc_arpcom): + ether_delmulti(ifr, &sc->sc_arpcom); + + if (error == ENETRESET) { + /* + * Multicast list has changed; set the hardware filter + * accordingly. + */ + qe_mcreset(sc); + error = 0; + } + break; + + case SIOCGIFMEDIA: + case SIOCSIFMEDIA: + error = ifmedia_ioctl(ifp, ifr, &sc->sc_ifmedia, cmd); + break; + + default: + error = EINVAL; + break; + } + + splx(s); + return (error); +} + + +void +qeinit(sc) + struct qe_softc *sc; +{ + struct ifnet *ifp = &sc->sc_arpcom.ac_if; + bus_space_tag_t t = sc->sc_bustag; + bus_space_handle_t cr = sc->sc_cr; + bus_space_handle_t mr = sc->sc_mr; + struct qec_softc *qec = sc->sc_qec; + u_int32_t qecaddr; + u_int8_t *ea; + int s; + +#if defined(SUN4U) || defined(__GNUC__) + (void)&t; +#endif + s = splnet(); + + qestop(sc); + + /* + * Allocate descriptor ring and buffers + */ + qec_meminit(&sc->sc_rb, QE_PKT_BUF_SZ); + + /* Channel registers: */ + bus_space_write_4(t, cr, QE_CRI_RXDS, (u_int32_t)sc->sc_rb.rb_rxddma); + bus_space_write_4(t, cr, QE_CRI_TXDS, (u_int32_t)sc->sc_rb.rb_txddma); + + bus_space_write_4(t, cr, QE_CRI_RIMASK, 0); + bus_space_write_4(t, cr, QE_CRI_TIMASK, 0); + bus_space_write_4(t, cr, QE_CRI_QMASK, 0); + bus_space_write_4(t, cr, QE_CRI_MMASK, QE_CR_MMASK_RXCOLL); + bus_space_write_4(t, cr, QE_CRI_CCNT, 0); + bus_space_write_4(t, cr, QE_CRI_PIPG, 0); + + qecaddr = sc->sc_channel * qec->sc_msize; + bus_space_write_4(t, cr, QE_CRI_RXWBUF, qecaddr); + bus_space_write_4(t, cr, QE_CRI_RXRBUF, qecaddr); + bus_space_write_4(t, cr, QE_CRI_TXWBUF, qecaddr + qec->sc_rsize); + bus_space_write_4(t, cr, QE_CRI_TXRBUF, qecaddr + qec->sc_rsize); + + /* + * When switching from mace<->qec always guarantee an sbus + * turnaround (if last op was read, perform a dummy write, and + * vice versa). + */ + bus_space_read_4(t, cr, QE_CRI_QMASK); + + /* MACE registers: */ + bus_space_write_1(t, mr, QE_MRI_PHYCC, QE_MR_PHYCC_ASEL); + bus_space_write_1(t, mr, QE_MRI_XMTFC, QE_MR_XMTFC_APADXMT); + bus_space_write_1(t, mr, QE_MRI_RCVFC, 0); + + /* + * Mask MACE's receive interrupt, since we're being notified + * by the QEC after DMA completes. + */ + bus_space_write_1(t, mr, QE_MRI_IMR, + QE_MR_IMR_CERRM | QE_MR_IMR_RCVINTM); + + bus_space_write_1(t, mr, QE_MRI_BIUCC, + QE_MR_BIUCC_BSWAP | QE_MR_BIUCC_64TS); + + bus_space_write_1(t, mr, QE_MRI_FIFOFC, + QE_MR_FIFOCC_TXF16 | QE_MR_FIFOCC_RXF32 | + QE_MR_FIFOCC_RFWU | QE_MR_FIFOCC_TFWU); + + bus_space_write_1(t, mr, QE_MRI_PLSCC, QE_MR_PLSCC_TP); + + /* + * Station address + */ + ea = sc->sc_enaddr; + bus_space_write_1(t, mr, QE_MRI_IAC, + QE_MR_IAC_ADDRCHG | QE_MR_IAC_PHYADDR); + bus_space_write_multi_1(t, mr, QE_MRI_PADR, ea, 6); + + /* Apply media settings */ + qe_ifmedia_upd(ifp); + + /* + * Clear Logical address filter + */ + bus_space_write_1(t, mr, QE_MRI_IAC, + QE_MR_IAC_ADDRCHG | QE_MR_IAC_LOGADDR); + bus_space_set_multi_1(t, mr, QE_MRI_LADRF, 0, 8); + bus_space_write_1(t, mr, QE_MRI_IAC, 0); + + /* Clear missed packet count (register cleared on read) */ + (void)bus_space_read_1(t, mr, QE_MRI_MPC); + +#if 0 + /* test register: */ + bus_space_write_1(t, mr, QE_MRI_UTR, 0); +#endif + + /* Reset multicast filter */ + qe_mcreset(sc); + + ifp->if_flags |= IFF_RUNNING; + ifp->if_flags &= ~IFF_OACTIVE; + splx(s); +} + +/* + * Reset multicast filter. + */ +void +qe_mcreset(sc) + struct qe_softc *sc; +{ + struct arpcom *ac = &sc->sc_arpcom; + struct ifnet *ifp = &sc->sc_arpcom.ac_if; + bus_space_tag_t t = sc->sc_bustag; + bus_space_handle_t mr = sc->sc_mr; + struct ether_multi *enm; + struct ether_multistep step; + u_int32_t crc; + u_int16_t hash[4]; + u_int8_t octet, maccc, *ladrp = (u_int8_t *)&hash[0]; + int i, j; + +#if defined(SUN4U) || defined(__GNUC__) + (void)&t; +#endif + + /* We also enable transmitter & receiver here */ + maccc = QE_MR_MACCC_ENXMT | QE_MR_MACCC_ENRCV; + + if (ifp->if_flags & IFF_PROMISC) { + maccc |= QE_MR_MACCC_PROM; + bus_space_write_1(t, mr, QE_MRI_MACCC, maccc); + return; + } + + if (ifp->if_flags & IFF_ALLMULTI) { + bus_space_write_1(t, mr, QE_MRI_IAC, + QE_MR_IAC_ADDRCHG | QE_MR_IAC_LOGADDR); + bus_space_set_multi_1(t, mr, QE_MRI_LADRF, 0xff, 8); + bus_space_write_1(t, mr, QE_MRI_IAC, 0); + bus_space_write_1(t, mr, QE_MRI_MACCC, maccc); + return; + } + + hash[3] = hash[2] = hash[1] = hash[0] = 0; + + ETHER_FIRST_MULTI(step, ac, enm); + while (enm != NULL) { + if (bcmp(enm->enm_addrlo, enm->enm_addrhi, + ETHER_ADDR_LEN) != 0) { + /* + * We must listen to a range of multicast + * addresses. For now, just accept all + * multicasts, rather than trying to set only + * those filter bits needed to match the range. + * (At this time, the only use of address + * ranges is for IP multicast routing, for + * which the range is big enough to require + * all bits set.) + */ + bus_space_write_1(t, mr, QE_MRI_IAC, + QE_MR_IAC_ADDRCHG | QE_MR_IAC_LOGADDR); + bus_space_set_multi_1(t, mr, QE_MRI_LADRF, 0xff, 8); + bus_space_write_1(t, mr, QE_MRI_IAC, 0); + ifp->if_flags |= IFF_ALLMULTI; + break; + } + + crc = 0xffffffff; + + for (i = 0; i < ETHER_ADDR_LEN; i++) { + octet = enm->enm_addrlo[i]; + + for (j = 0; j < 8; j++) { + if ((crc & 1) ^ (octet & 1)) { + crc >>= 1; + crc ^= MC_POLY_LE; + } + else + crc >>= 1; + octet >>= 1; + } + } + + crc >>= 26; + hash[crc >> 4] |= 1 << (crc & 0xf); + ETHER_NEXT_MULTI(step, enm); + } + + bus_space_write_1(t, mr, QE_MRI_IAC, + QE_MR_IAC_ADDRCHG | QE_MR_IAC_LOGADDR); + bus_space_write_multi_1(t, mr, QE_MRI_LADRF, ladrp, 8); + bus_space_write_1(t, mr, QE_MRI_IAC, 0); + bus_space_write_1(t, mr, QE_MRI_MACCC, maccc); +} + +/* + * Get current media settings. + */ +void +qe_ifmedia_sts(ifp, ifmr) + struct ifnet *ifp; + struct ifmediareq *ifmr; +{ + struct qe_softc *sc = ifp->if_softc; + u_int8_t phycc; + + ifmr->ifm_active = IFM_ETHER | IFM_10_T; + phycc = bus_space_read_1(sc->sc_bustag, sc->sc_mr, QE_MRI_PHYCC); + if ((phycc & QE_MR_PHYCC_DLNKTST) == 0) { + ifmr->ifm_status |= IFM_AVALID; + if (phycc & QE_MR_PHYCC_LNKFL) + ifmr->ifm_status &= ~IFM_ACTIVE; + else + ifmr->ifm_status |= IFM_ACTIVE; + } +} + +/* + * Set media options. + */ +int +qe_ifmedia_upd(ifp) + struct ifnet *ifp; +{ + struct qe_softc *sc = ifp->if_softc; + int media = sc->sc_ifmedia.ifm_media; + + if (IFM_TYPE(media) != IFM_ETHER) + return (EINVAL); + + if (IFM_SUBTYPE(media) != IFM_10_T) + return (EINVAL); + + return (0); +} |