From 8cd3a4ecfc9eb826d2634fb9fdce918718fe7cad Mon Sep 17 00:00:00 2001 From: Mark Kettenis Date: Sat, 13 Aug 2016 20:35:58 +0000 Subject: Add the core of the code for dwge(4), a driver for the Synopsis Designware GMAC core which is used on the Allwinner A20 SoCs and later SoCs in the Allwinner family. Since this core is very likely to show up the base of other devices, it gets to live here in dev/ic. Ported from NetBSD (where it is named awge(4)) by patrick@. Some further fixes by me. ok deraadt@ --- sys/dev/ic/dwc_gmac.c | 1484 +++++++++++++++++++++++++++++++++++++++++++++ sys/dev/ic/dwc_gmac_reg.h | 227 +++++++ sys/dev/ic/dwc_gmac_var.h | 95 +++ 3 files changed, 1806 insertions(+) create mode 100644 sys/dev/ic/dwc_gmac.c create mode 100644 sys/dev/ic/dwc_gmac_reg.h create mode 100644 sys/dev/ic/dwc_gmac_var.h (limited to 'sys/dev') diff --git a/sys/dev/ic/dwc_gmac.c b/sys/dev/ic/dwc_gmac.c new file mode 100644 index 00000000000..8d66ec7c669 --- /dev/null +++ b/sys/dev/ic/dwc_gmac.c @@ -0,0 +1,1484 @@ +/* $OpenBSD: dwc_gmac.c,v 1.1 2016/08/13 20:35:57 kettenis Exp $ */ +/* $NetBSD: dwc_gmac.c,v 1.34 2015/08/21 20:12:29 jmcneill Exp $ */ + +/*- + * Copyright (c) 2013, 2014 The NetBSD Foundation, Inc. + * All rights reserved. + * + * This code is derived from software contributed to The NetBSD Foundation + * by Matt Thomas of 3am Software Foundry and Martin Husemann. + * + * Redistribution and use in source and binary forms, with or without + * modification, are permitted provided that the following conditions + * are met: + * 1. Redistributions of source code must retain the above copyright + * notice, this list of conditions and the following disclaimer. + * 2. Redistributions in binary form must reproduce the above copyright + * notice, this list of conditions and the following disclaimer in the + * documentation and/or other materials provided with the distribution. + * + * THIS SOFTWARE IS PROVIDED BY THE NETBSD FOUNDATION, INC. AND CONTRIBUTORS + * ``AS IS'' AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED + * TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR + * PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE FOUNDATION OR CONTRIBUTORS + * BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR + * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF + * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS + * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN + * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) + * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE + * POSSIBILITY OF SUCH DAMAGE. + */ + +/* + * This driver supports the Synopsis Designware GMAC core, as found + * on Allwinner A20 cores and others. + * + * Real documentation seems to not be available, the marketing product + * documents could be found here: + * + * http://www.synopsys.com/dw/ipdir.php?ds=dwc_ether_mac10_100_1000_unive + */ + +/*__KERNEL_RCSID(1, "$NetBSD: dwc_gmac.c,v 1.34 2015/08/21 20:12:29 jmcneill Exp $");*/ + +/* #define DWC_GMAC_DEBUG 1 */ + +#include +#include +#include +#include +#include +#include +#include +#include +#include +#include +#include +#include + +#include "bpfilter.h" + +#include +#include +#if NBPFILTER > 0 +#include +#endif + +#include +#include + +#include +#include + +#include +#include + +int dwc_gmac_ifmedia_upd(struct ifnet *); +void dwc_gmac_ifmedia_sts(struct ifnet *, struct ifmediareq *); + +int dwc_gmac_miibus_read_reg(struct device *, int, int); +void dwc_gmac_miibus_write_reg(struct device *, int, int, int); +void dwc_gmac_miibus_statchg(struct device *); + +int dwc_gmac_reset(struct dwc_gmac_softc *); +void dwc_gmac_write_hwaddr(struct dwc_gmac_softc *, uint8_t *enaddr); +int dwc_gmac_alloc_dma_rings(struct dwc_gmac_softc *); +void dwc_gmac_free_dma_rings(struct dwc_gmac_softc *); +int dwc_gmac_alloc_rx_ring(struct dwc_gmac_softc *, struct dwc_gmac_rx_ring *); +void dwc_gmac_reset_rx_ring(struct dwc_gmac_softc *, struct dwc_gmac_rx_ring *); +void dwc_gmac_free_rx_ring(struct dwc_gmac_softc *, struct dwc_gmac_rx_ring *); +int dwc_gmac_alloc_tx_ring(struct dwc_gmac_softc *, struct dwc_gmac_tx_ring *); +void dwc_gmac_reset_tx_ring(struct dwc_gmac_softc *, struct dwc_gmac_tx_ring *); +void dwc_gmac_free_tx_ring(struct dwc_gmac_softc *, struct dwc_gmac_tx_ring *); +void dwc_gmac_txdesc_sync(struct dwc_gmac_softc *, int, int, int); +int dwc_gmac_init(struct ifnet *); +void dwc_gmac_stop(struct ifnet *, int); +void dwc_gmac_start(struct ifnet *); +int dwc_gmac_queue(struct dwc_gmac_softc *, struct mbuf *); +int dwc_gmac_ioctl(struct ifnet *, u_long, caddr_t); +void dwc_gmac_tx_intr(struct dwc_gmac_softc *); +void dwc_gmac_rx_intr(struct dwc_gmac_softc *); +void dwc_gmac_setmulti(struct dwc_gmac_softc *); +int dwc_gmac_ifflags_cb(struct arpcom *); +static uint32_t bitrev32(uint32_t); + +#define TX_DESC_OFFSET(N) ((DWGE_RX_RING_COUNT+(N)) \ + *sizeof(struct dwc_gmac_dev_dmadesc)) +#define TX_NEXT(N) (((N)+1) & (DWGE_TX_RING_COUNT-1)) + +#define RX_DESC_OFFSET(N) ((N)*sizeof(struct dwc_gmac_dev_dmadesc)) +#define RX_NEXT(N) (((N)+1) & (DWGE_RX_RING_COUNT-1)) + + + +#define GMAC_DEF_DMA_INT_MASK (GMAC_DMA_INT_TIE|GMAC_DMA_INT_RIE| \ + GMAC_DMA_INT_NIE|GMAC_DMA_INT_AIE| \ + GMAC_DMA_INT_FBE|GMAC_DMA_INT_UNE) + +#define GMAC_DMA_INT_ERRORS (GMAC_DMA_INT_AIE|GMAC_DMA_INT_ERE| \ + GMAC_DMA_INT_FBE| \ + GMAC_DMA_INT_RWE|GMAC_DMA_INT_RUE| \ + GMAC_DMA_INT_UNE|GMAC_DMA_INT_OVE| \ + GMAC_DMA_INT_TJE) + +#define AWIN_DEF_MAC_INTRMASK \ + (AWIN_GMAC_MAC_INT_TSI | AWIN_GMAC_MAC_INT_ANEG | \ + AWIN_GMAC_MAC_INT_LINKCHG | AWIN_GMAC_MAC_INT_RGSMII) + + +#ifdef DWC_GMAC_DEBUG +void dwc_gmac_dump_dma(struct dwc_gmac_softc *sc); +void dwc_gmac_dump_tx_desc(struct dwc_gmac_softc *sc); +void dwc_gmac_dump_rx_desc(struct dwc_gmac_softc *sc); +void dwc_dump_and_abort(struct dwc_gmac_softc *sc, const char *msg); +void dwc_dump_status(struct dwc_gmac_softc *sc); +void dwc_gmac_dump_ffilt(struct dwc_gmac_softc *sc, uint32_t ffilt); +#endif + +struct cfdriver dwge_cd = { + NULL, "dwge", DV_IFNET +}; + +void +dwc_gmac_attach(struct dwc_gmac_softc *sc, uint32_t mii_clk) +{ + uint8_t enaddr[ETHER_ADDR_LEN]; + struct mii_data * const mii = &sc->sc_mii; + struct ifnet * const ifp = &sc->sc_ac.ac_if; + uint32_t maclo, machi; + int s; + + mtx_init(&sc->sc_mdio_lock, IPL_NET); + sc->sc_mii_clk = mii_clk & 7; + + /* + * If we did not get an externaly configure address, + * try to read one from the current filter setup, + * before resetting the chip. + */ + maclo = bus_space_read_4(sc->sc_bst, sc->sc_bsh, + AWIN_GMAC_MAC_ADDR0LO); + machi = bus_space_read_4(sc->sc_bst, sc->sc_bsh, + AWIN_GMAC_MAC_ADDR0HI); + + if (maclo == 0xffffffff && (machi & 0xffff) == 0xffff) { + ether_fakeaddr(&sc->sc_ac.ac_if); + memcpy(enaddr, sc->sc_ac.ac_enaddr, ETHER_ADDR_LEN); + } else { + enaddr[0] = maclo & 0x0ff; + enaddr[1] = (maclo >> 8) & 0x0ff; + enaddr[2] = (maclo >> 16) & 0x0ff; + enaddr[3] = (maclo >> 24) & 0x0ff; + enaddr[4] = machi & 0x0ff; + enaddr[5] = (machi >> 8) & 0x0ff; + } + + /* + * Init chip and do initial setup + */ + if (dwc_gmac_reset(sc) != 0) + return; /* not much to cleanup, haven't attached yet */ + dwc_gmac_write_hwaddr(sc, enaddr); + printf("%s: Ethernet address: %s\n", sc->sc_dev.dv_xname, + ether_sprintf(enaddr)); + memcpy(sc->sc_ac.ac_enaddr, enaddr, ETHER_ADDR_LEN); + + /* + * Allocate Tx and Rx rings + */ + if (dwc_gmac_alloc_dma_rings(sc) != 0) { + printf("%s: could not allocate DMA rings\n", + sc->sc_dev.dv_xname); + goto fail; + } + + if (dwc_gmac_alloc_tx_ring(sc, &sc->sc_txq) != 0) { + printf("%s: could not allocate Tx ring\n", + sc->sc_dev.dv_xname); + goto fail; + } + + mtx_init(&sc->sc_rxq.r_mtx, IPL_NET); + if (dwc_gmac_alloc_rx_ring(sc, &sc->sc_rxq) != 0) { + printf("%s: could not allocate Rx ring\n", + sc->sc_dev.dv_xname); + goto fail; + } + + /* + * Prepare interface data + */ + ifp->if_softc = sc; + strlcpy(ifp->if_xname, sc->sc_dev.dv_xname, IFNAMSIZ); + ifp->if_flags = IFF_BROADCAST | IFF_SIMPLEX | IFF_MULTICAST; + ifp->if_ioctl = dwc_gmac_ioctl; + ifp->if_start = dwc_gmac_start; + ifp->if_capabilities = IFCAP_VLAN_MTU; + + IFQ_SET_MAXLEN(&ifp->if_snd, IFQ_MAXLEN); + + /* + * Attach MII subdevices + */ + mii->mii_ifp = ifp; + mii->mii_readreg = dwc_gmac_miibus_read_reg; + mii->mii_writereg = dwc_gmac_miibus_write_reg; + mii->mii_statchg = dwc_gmac_miibus_statchg; + + ifmedia_init(&mii->mii_media, 0, dwc_gmac_ifmedia_upd, + dwc_gmac_ifmedia_sts); + mii_attach((void *)sc, mii, 0xffffffff, MII_PHY_ANY, MII_OFFSET_ANY, + MIIF_DOPAUSE); + + if (LIST_EMPTY(&mii->mii_phys)) { + printf("%s: no PHY found!\n", sc->sc_dev.dv_xname); + ifmedia_add(&mii->mii_media, IFM_ETHER|IFM_MANUAL, 0, NULL); + ifmedia_set(&mii->mii_media, IFM_ETHER|IFM_MANUAL); + } else { + ifmedia_set(&mii->mii_media, IFM_ETHER|IFM_AUTO); + } + + /* + * Ready, attach interface + */ + if_attach(ifp); + ether_ifattach(ifp); + + /* + * Enable interrupts + */ + s = splnet(); + bus_space_write_4(sc->sc_bst, sc->sc_bsh, AWIN_GMAC_MAC_INTMASK, + AWIN_DEF_MAC_INTRMASK); + bus_space_write_4(sc->sc_bst, sc->sc_bsh, AWIN_GMAC_DMA_INTENABLE, + GMAC_DEF_DMA_INT_MASK); + splx(s); + + return; + +fail: + dwc_gmac_free_rx_ring(sc, &sc->sc_rxq); + dwc_gmac_free_tx_ring(sc, &sc->sc_txq); +} + +int +dwc_gmac_ifmedia_upd(struct ifnet *ifp) +{ + struct dwc_gmac_softc *sc = ifp->if_softc; + struct mii_data *mii = &sc->sc_mii; + int err; + if (mii->mii_instance) { + struct mii_softc *miisc; + + LIST_FOREACH(miisc, &mii->mii_phys, mii_list) + mii_phy_reset(miisc); + } + err = mii_mediachg(mii); + return (err); +} + +void +dwc_gmac_ifmedia_sts(struct ifnet *ifp, struct ifmediareq *ifmr) +{ + struct dwc_gmac_softc *sc = ifp->if_softc; + struct mii_data *mii = &sc->sc_mii; + + mii_pollstat(mii); + + ifmr->ifm_active = mii->mii_media_active; + ifmr->ifm_status = mii->mii_media_status; +} + +int +dwc_gmac_reset(struct dwc_gmac_softc *sc) +{ + size_t cnt; + bus_space_write_4(sc->sc_bst, sc->sc_bsh, AWIN_GMAC_DMA_BUSMODE, + bus_space_read_4(sc->sc_bst, sc->sc_bsh, AWIN_GMAC_DMA_BUSMODE) | GMAC_BUSMODE_RESET); + for (cnt = 0; cnt < 3000; cnt++) { + if ((bus_space_read_4(sc->sc_bst, sc->sc_bsh, AWIN_GMAC_DMA_BUSMODE) + & GMAC_BUSMODE_RESET) == 0) + return 0; + delay(10); + } + + printf("%s: reset timed out\n", sc->sc_dev.dv_xname); + return EIO; +} + +void +dwc_gmac_write_hwaddr(struct dwc_gmac_softc *sc, uint8_t *enaddr) +{ + uint32_t lo, hi; + + lo = enaddr[0] | (enaddr[1] << 8) | (enaddr[2] << 16) + | (enaddr[3] << 24); + hi = enaddr[4] | (enaddr[5] << 8); + bus_space_write_4(sc->sc_bst, sc->sc_bsh, AWIN_GMAC_MAC_ADDR0LO, lo); + bus_space_write_4(sc->sc_bst, sc->sc_bsh, AWIN_GMAC_MAC_ADDR0HI, hi); +} + +int +dwc_gmac_miibus_read_reg(struct device *self, int phy, int reg) +{ + struct dwc_gmac_softc * const sc = (struct dwc_gmac_softc *)self; + uint16_t mii; + size_t cnt; + int rv = 0; + + mii = ((phy & GMAC_MII_PHY_MASK) << GMAC_MII_PHY_SHIFT) + | ((reg & GMAC_MII_REG_MASK) << GMAC_MII_REG_SHIFT) + | ((sc->sc_mii_clk & GMAC_MII_CLKMASK_MASK) + << GMAC_MII_CLKMASK_SHIFT) + | GMAC_MII_BUSY; + + mtx_enter(&sc->sc_mdio_lock); + bus_space_write_4(sc->sc_bst, sc->sc_bsh, AWIN_GMAC_MAC_MIIADDR, mii); + + for (cnt = 0; cnt < 1000; cnt++) { + if (!(bus_space_read_4(sc->sc_bst, sc->sc_bsh, + AWIN_GMAC_MAC_MIIADDR) & GMAC_MII_BUSY)) { + rv = bus_space_read_4(sc->sc_bst, sc->sc_bsh, + AWIN_GMAC_MAC_MIIDATA); + break; + } + delay(10); + } + + mtx_leave(&sc->sc_mdio_lock); + + return rv; +} + +void +dwc_gmac_miibus_write_reg(struct device *self, int phy, int reg, int val) +{ + struct dwc_gmac_softc * const sc = (struct dwc_gmac_softc *)self; + uint16_t mii; + size_t cnt; + + mii = ((phy & GMAC_MII_PHY_MASK) << GMAC_MII_PHY_SHIFT) + | ((reg & GMAC_MII_REG_MASK) << GMAC_MII_REG_SHIFT) + | ((sc->sc_mii_clk & GMAC_MII_CLKMASK_MASK) + << GMAC_MII_CLKMASK_SHIFT) + | GMAC_MII_BUSY | GMAC_MII_WRITE; + + mtx_enter(&sc->sc_mdio_lock); + bus_space_write_4(sc->sc_bst, sc->sc_bsh, AWIN_GMAC_MAC_MIIDATA, val); + bus_space_write_4(sc->sc_bst, sc->sc_bsh, AWIN_GMAC_MAC_MIIADDR, mii); + + for (cnt = 0; cnt < 1000; cnt++) { + if (!(bus_space_read_4(sc->sc_bst, sc->sc_bsh, + AWIN_GMAC_MAC_MIIADDR) & GMAC_MII_BUSY)) + break; + delay(10); + } + + mtx_leave(&sc->sc_mdio_lock); +} + +int +dwc_gmac_alloc_rx_ring(struct dwc_gmac_softc *sc, + struct dwc_gmac_rx_ring *ring) +{ + struct dwc_gmac_rx_data *data; + bus_addr_t physaddr; + const size_t descsize = DWGE_RX_RING_COUNT * sizeof(*ring->r_desc); + int error, i, next; + + ring->r_cur = ring->r_next = 0; + memset(ring->r_desc, 0, descsize); + + /* + * Pre-allocate Rx buffers and populate Rx ring. + */ + for (i = 0; i < DWGE_RX_RING_COUNT; i++) { + struct dwc_gmac_dev_dmadesc *desc; + + data = &sc->sc_rxq.r_data[i]; + + MGETHDR(data->rd_m, M_DONTWAIT, MT_DATA); + if (data->rd_m == NULL) { + printf("%s: could not allocate rx mbuf #%d\n", + sc->sc_dev.dv_xname, i); + error = ENOMEM; + goto fail; + } + error = bus_dmamap_create(sc->sc_dmat, MCLBYTES, 1, + MCLBYTES, 0, BUS_DMA_NOWAIT, &data->rd_map); + if (error != 0) { + printf("%s: could not create DMA map\n", + sc->sc_dev.dv_xname); + data->rd_map = NULL; + goto fail; + } + MCLGET(data->rd_m, M_DONTWAIT); + if (!(data->rd_m->m_flags & M_EXT)) { + printf("%s: could not allocate mbuf cluster #%d\n", + sc->sc_dev.dv_xname, i); + error = ENOMEM; + goto fail; + } + + error = bus_dmamap_load(sc->sc_dmat, data->rd_map, + mtod(data->rd_m, void *), MCLBYTES, NULL, + BUS_DMA_READ | BUS_DMA_NOWAIT); + if (error != 0) { + printf("%s: could not load rx buf DMA map #%d", + sc->sc_dev.dv_xname, i); + goto fail; + } + physaddr = data->rd_map->dm_segs[0].ds_addr; + + desc = &sc->sc_rxq.r_desc[i]; + desc->ddesc_data = htole32(physaddr); + next = RX_NEXT(i); + desc->ddesc_next = htole32(ring->r_physaddr + + next * sizeof(*desc)); + desc->ddesc_cntl = htole32( + ((DWGE_MAX_PACKET & DDESC_CNTL_SIZE1MASK) + << DDESC_CNTL_SIZE1SHIFT) | DDESC_CNTL_RXCHAIN); + desc->ddesc_status = htole32(DDESC_STATUS_OWNEDBYDEV); + } + + bus_dmamap_sync(sc->sc_dmat, sc->sc_dma_ring_map, 0, + DWGE_RX_RING_COUNT*sizeof(struct dwc_gmac_dev_dmadesc), + BUS_DMASYNC_PREWRITE|BUS_DMASYNC_PREREAD); + bus_space_write_4(sc->sc_bst, sc->sc_bsh, AWIN_GMAC_DMA_RX_ADDR, + ring->r_physaddr); + + return 0; + +fail: + dwc_gmac_free_rx_ring(sc, ring); + return error; +} + +void +dwc_gmac_reset_rx_ring(struct dwc_gmac_softc *sc, + struct dwc_gmac_rx_ring *ring) +{ + struct dwc_gmac_dev_dmadesc *desc; + int i; + + for (i = 0; i < DWGE_RX_RING_COUNT; i++) { + desc = &sc->sc_rxq.r_desc[i]; + desc->ddesc_cntl = htole32( + ((DWGE_MAX_PACKET & DDESC_CNTL_SIZE1MASK) + << DDESC_CNTL_SIZE1SHIFT) | DDESC_CNTL_RXCHAIN); + desc->ddesc_status = htole32(DDESC_STATUS_OWNEDBYDEV); + } + + bus_dmamap_sync(sc->sc_dmat, sc->sc_dma_ring_map, 0, + DWGE_RX_RING_COUNT*sizeof(struct dwc_gmac_dev_dmadesc), + BUS_DMASYNC_PREREAD|BUS_DMASYNC_PREWRITE); + + ring->r_cur = ring->r_next = 0; + /* reset DMA address to start of ring */ + bus_space_write_4(sc->sc_bst, sc->sc_bsh, AWIN_GMAC_DMA_RX_ADDR, + sc->sc_rxq.r_physaddr); +} + +int +dwc_gmac_alloc_dma_rings(struct dwc_gmac_softc *sc) +{ + const size_t descsize = DWGE_TOTAL_RING_COUNT * + sizeof(struct dwc_gmac_dev_dmadesc); + int error, nsegs; + caddr_t rings; + + error = bus_dmamap_create(sc->sc_dmat, descsize, 1, descsize, 0, + BUS_DMA_NOWAIT, &sc->sc_dma_ring_map); + if (error != 0) { + printf("%s: could not create desc DMA map\n", sc->sc_dev.dv_xname); + sc->sc_dma_ring_map = NULL; + goto fail; + } + + error = bus_dmamem_alloc(sc->sc_dmat, descsize, PAGE_SIZE, 0, + &sc->sc_dma_ring_seg, 1, &nsegs, BUS_DMA_NOWAIT|BUS_DMA_COHERENT); + if (error != 0) { + printf("%s: could not map DMA memory\n", sc->sc_dev.dv_xname); + goto fail; + } + + error = bus_dmamem_map(sc->sc_dmat, &sc->sc_dma_ring_seg, nsegs, + descsize, &rings, BUS_DMA_NOWAIT|BUS_DMA_COHERENT); + if (error != 0) { + printf("%s: could not allocate DMA memory\n", sc->sc_dev.dv_xname); + goto fail; + } + + error = bus_dmamap_load(sc->sc_dmat, sc->sc_dma_ring_map, rings, + descsize, NULL, BUS_DMA_NOWAIT|BUS_DMA_COHERENT); + if (error != 0) { + printf("%s: could not load desc DMA map\n", sc->sc_dev.dv_xname); + goto fail; + } + + /* give first DWGE_RX_RING_COUNT to the RX side */ + sc->sc_rxq.r_desc = (struct dwc_gmac_dev_dmadesc *)rings; + sc->sc_rxq.r_physaddr = sc->sc_dma_ring_map->dm_segs[0].ds_addr; + + /* and next rings to the TX side */ + sc->sc_txq.t_desc = sc->sc_rxq.r_desc + DWGE_RX_RING_COUNT; + sc->sc_txq.t_physaddr = sc->sc_rxq.r_physaddr + + DWGE_RX_RING_COUNT*sizeof(struct dwc_gmac_dev_dmadesc); + + return 0; + +fail: + dwc_gmac_free_dma_rings(sc); + return error; +} + +void +dwc_gmac_free_dma_rings(struct dwc_gmac_softc *sc) +{ + bus_dmamap_sync(sc->sc_dmat, sc->sc_dma_ring_map, 0, + sc->sc_dma_ring_map->dm_mapsize, BUS_DMASYNC_POSTWRITE); + bus_dmamap_unload(sc->sc_dmat, sc->sc_dma_ring_map); + bus_dmamem_unmap(sc->sc_dmat, (caddr_t)sc->sc_rxq.r_desc, + DWGE_TOTAL_RING_COUNT * sizeof(struct dwc_gmac_dev_dmadesc)); + bus_dmamem_free(sc->sc_dmat, &sc->sc_dma_ring_seg, 1); +} + +void +dwc_gmac_free_rx_ring(struct dwc_gmac_softc *sc, struct dwc_gmac_rx_ring *ring) +{ + struct dwc_gmac_rx_data *data; + int i; + + if (ring->r_desc == NULL) + return; + + + for (i = 0; i < DWGE_RX_RING_COUNT; i++) { + data = &ring->r_data[i]; + + if (data->rd_map != NULL) { + bus_dmamap_sync(sc->sc_dmat, data->rd_map, 0, + DWGE_RX_RING_COUNT + *sizeof(struct dwc_gmac_dev_dmadesc), + BUS_DMASYNC_POSTREAD); + bus_dmamap_unload(sc->sc_dmat, data->rd_map); + bus_dmamap_destroy(sc->sc_dmat, data->rd_map); + } + if (data->rd_m != NULL) + m_freem(data->rd_m); + } +} + +int +dwc_gmac_alloc_tx_ring(struct dwc_gmac_softc *sc, + struct dwc_gmac_tx_ring *ring) +{ + int i, error = 0; + + ring->t_queued = 0; + ring->t_cur = ring->t_next = 0; + + memset(ring->t_desc, 0, DWGE_TX_RING_COUNT*sizeof(*ring->t_desc)); + bus_dmamap_sync(sc->sc_dmat, sc->sc_dma_ring_map, + TX_DESC_OFFSET(0), + DWGE_TX_RING_COUNT*sizeof(struct dwc_gmac_dev_dmadesc), + BUS_DMASYNC_POSTWRITE); + + for (i = 0; i < DWGE_TX_RING_COUNT; i++) { + error = bus_dmamap_create(sc->sc_dmat, MCLBYTES, + DWGE_TX_RING_COUNT, MCLBYTES, 0, + BUS_DMA_NOWAIT|BUS_DMA_COHERENT, + &ring->t_data[i].td_map); + if (error != 0) { + printf("%s: could not create TX DMA map #%d\n", + sc->sc_dev.dv_xname, i); + ring->t_data[i].td_map = NULL; + goto fail; + } + ring->t_desc[i].ddesc_next = htole32( + ring->t_physaddr + sizeof(struct dwc_gmac_dev_dmadesc) + *TX_NEXT(i)); + } + + return 0; + +fail: + dwc_gmac_free_tx_ring(sc, ring); + return error; +} + +void +dwc_gmac_txdesc_sync(struct dwc_gmac_softc *sc, int start, int end, int ops) +{ + /* 'end' is pointing one descriptor beyound the last we want to sync */ + if (end > start) { + bus_dmamap_sync(sc->sc_dmat, sc->sc_dma_ring_map, + TX_DESC_OFFSET(start), + TX_DESC_OFFSET(end)-TX_DESC_OFFSET(start), + ops); + return; + } + /* sync from 'start' to end of ring */ + bus_dmamap_sync(sc->sc_dmat, sc->sc_dma_ring_map, + TX_DESC_OFFSET(start), + TX_DESC_OFFSET(DWGE_TX_RING_COUNT)-TX_DESC_OFFSET(start), + ops); + /* sync from start of ring to 'end' */ + bus_dmamap_sync(sc->sc_dmat, sc->sc_dma_ring_map, + TX_DESC_OFFSET(0), + TX_DESC_OFFSET(end)-TX_DESC_OFFSET(0), + ops); +} + +void +dwc_gmac_reset_tx_ring(struct dwc_gmac_softc *sc, + struct dwc_gmac_tx_ring *ring) +{ + int i; + + for (i = 0; i < DWGE_TX_RING_COUNT; i++) { + struct dwc_gmac_tx_data *data = &ring->t_data[i]; + + if (data->td_m != NULL) { + bus_dmamap_sync(sc->sc_dmat, data->td_active, + 0, data->td_active->dm_mapsize, + BUS_DMASYNC_POSTWRITE); + bus_dmamap_unload(sc->sc_dmat, data->td_active); + m_freem(data->td_m); + data->td_m = NULL; + } + } + + bus_dmamap_sync(sc->sc_dmat, sc->sc_dma_ring_map, + TX_DESC_OFFSET(0), + DWGE_TX_RING_COUNT*sizeof(struct dwc_gmac_dev_dmadesc), + BUS_DMASYNC_PREREAD|BUS_DMASYNC_PREWRITE); + bus_space_write_4(sc->sc_bst, sc->sc_bsh, AWIN_GMAC_DMA_TX_ADDR, + sc->sc_txq.t_physaddr); + + ring->t_queued = 0; + ring->t_cur = ring->t_next = 0; +} + +void +dwc_gmac_free_tx_ring(struct dwc_gmac_softc *sc, + struct dwc_gmac_tx_ring *ring) +{ + int i; + + /* unload the maps */ + for (i = 0; i < DWGE_TX_RING_COUNT; i++) { + struct dwc_gmac_tx_data *data = &ring->t_data[i]; + + if (data->td_m != NULL) { + bus_dmamap_sync(sc->sc_dmat, data->td_active, + 0, data->td_map->dm_mapsize, + BUS_DMASYNC_POSTWRITE); + bus_dmamap_unload(sc->sc_dmat, data->td_active); + m_freem(data->td_m); + data->td_m = NULL; + } + } + + /* and actually free them */ + for (i = 0; i < DWGE_TX_RING_COUNT; i++) { + struct dwc_gmac_tx_data *data = &ring->t_data[i]; + + bus_dmamap_destroy(sc->sc_dmat, data->td_map); + } +} + +void +dwc_gmac_miibus_statchg(struct device *dev) +{ + struct dwc_gmac_softc * const sc = (struct dwc_gmac_softc *)dev; + struct mii_data * const mii = &sc->sc_mii; + uint32_t conf, flow; + + /* + * Set MII or GMII interface based on the speed + * negotiated by the PHY. + */ + conf = bus_space_read_4(sc->sc_bst, sc->sc_bsh, AWIN_GMAC_MAC_CONF); + conf &= ~(AWIN_GMAC_MAC_CONF_FES100|AWIN_GMAC_MAC_CONF_MIISEL + |AWIN_GMAC_MAC_CONF_FULLDPLX); + conf |= AWIN_GMAC_MAC_CONF_FRAMEBURST + | AWIN_GMAC_MAC_CONF_DISABLERXOWN + | AWIN_GMAC_MAC_CONF_DISABLEJABBER + | AWIN_GMAC_MAC_CONF_ACS + | AWIN_GMAC_MAC_CONF_RXENABLE + | AWIN_GMAC_MAC_CONF_TXENABLE; + switch (IFM_SUBTYPE(mii->mii_media_active)) { + case IFM_10_T: + conf |= AWIN_GMAC_MAC_CONF_MIISEL; + break; + case IFM_100_TX: + conf |= AWIN_GMAC_MAC_CONF_FES100 | + AWIN_GMAC_MAC_CONF_MIISEL; + break; + case IFM_1000_T: + break; + } + + flow = 0; + if (IFM_OPTIONS(mii->mii_media_active) & IFM_FDX) { + conf |= AWIN_GMAC_MAC_CONF_FULLDPLX; + flow |= ((0x200 & AWIN_GMAC_MAC_FLOWCTRL_PAUSE_MASK) + << AWIN_GMAC_MAC_FLOWCTRL_PAUSE_SHIFT); + } + if (mii->mii_media_active & IFM_ETH_TXPAUSE) { + flow |= AWIN_GMAC_MAC_FLOWCTRL_TFE; + } + if (mii->mii_media_active & IFM_ETH_RXPAUSE) { + flow |= AWIN_GMAC_MAC_FLOWCTRL_RFE; + } + bus_space_write_4(sc->sc_bst, sc->sc_bsh, + AWIN_GMAC_MAC_FLOWCTRL, flow); + +#ifdef DWC_GMAC_DEBUG + printf("%s: setting MAC conf register: %08x\n", + sc->sc_dev.dv_xname, conf); +#endif + + bus_space_write_4(sc->sc_bst, sc->sc_bsh, + AWIN_GMAC_MAC_CONF, conf); +} + +int +dwc_gmac_init(struct ifnet *ifp) +{ + struct dwc_gmac_softc *sc = ifp->if_softc; + uint32_t ffilt; + + if (ifp->if_flags & IFF_RUNNING) + return 0; + + dwc_gmac_stop(ifp, 0); + + /* + * Configure DMA burst/transfer mode and RX/TX priorities. + * XXX - the GMAC_BUSMODE_PRIORXTX bits are undocumented. + */ + bus_space_write_4(sc->sc_bst, sc->sc_bsh, AWIN_GMAC_DMA_BUSMODE, + GMAC_BUSMODE_FIXEDBURST | GMAC_BUSMODE_4PBL | + 2 << GMAC_BUSMODE_RPBL_SHIFT | + 2 << GMAC_BUSMODE_PBL_SHIFT); + + /* + * Set up address filter + */ + ffilt = bus_space_read_4(sc->sc_bst, sc->sc_bsh, AWIN_GMAC_MAC_FFILT); + if (ifp->if_flags & IFF_PROMISC) { + ffilt |= AWIN_GMAC_MAC_FFILT_PR; + } else { + ffilt &= ~AWIN_GMAC_MAC_FFILT_PR; + } + if (ifp->if_flags & IFF_BROADCAST) { + ffilt &= ~AWIN_GMAC_MAC_FFILT_DBF; + } else { + ffilt |= AWIN_GMAC_MAC_FFILT_DBF; + } + bus_space_write_4(sc->sc_bst, sc->sc_bsh, AWIN_GMAC_MAC_FFILT, ffilt); + + /* + * Set up multicast filter + */ + dwc_gmac_setmulti(sc); + + /* + * Set up dma pointer for RX and TX ring + */ + bus_space_write_4(sc->sc_bst, sc->sc_bsh, AWIN_GMAC_DMA_RX_ADDR, + sc->sc_rxq.r_physaddr); + bus_space_write_4(sc->sc_bst, sc->sc_bsh, AWIN_GMAC_DMA_TX_ADDR, + sc->sc_txq.t_physaddr); + + /* + * Start RX/TX part + */ + bus_space_write_4(sc->sc_bst, sc->sc_bsh, + AWIN_GMAC_DMA_OPMODE, GMAC_DMA_OP_RXSTART | GMAC_DMA_OP_TXSTART | + GMAC_DMA_OP_RXSTOREFORWARD | GMAC_DMA_OP_TXSTOREFORWARD); + + ifq_clr_oactive(&ifp->if_snd); + ifp->if_flags |= IFF_RUNNING; + + return 0; +} + +void +dwc_gmac_start(struct ifnet *ifp) +{ + struct dwc_gmac_softc *sc = ifp->if_softc; + int old = sc->sc_txq.t_queued; + int start = sc->sc_txq.t_cur; + struct mbuf *m_head = NULL; + + if (ifq_is_oactive(&ifp->if_snd) || !(ifp->if_flags & IFF_RUNNING)) + return; + + for (;;) { + m_head = ifq_deq_begin(&ifp->if_snd); + if (m_head == NULL) + break; + if (dwc_gmac_queue(sc, m_head) != 0) { + ifq_deq_rollback(&ifp->if_snd, m_head); + ifq_set_oactive(&ifp->if_snd); + break; + } + + ifq_deq_commit(&ifp->if_snd, m_head); + + ifp->if_opackets++; + +#if NBPFILTER > 0 + if (ifp->if_bpf) + bpf_mtap(ifp->if_bpf, m_head, BPF_DIRECTION_OUT); +#endif + + if (sc->sc_txq.t_queued == DWGE_TX_RING_COUNT) { + ifq_set_oactive(&ifp->if_snd); + break; + } + } + + if (sc->sc_txq.t_queued != old) { + /* packets have been queued, kick it off */ + dwc_gmac_txdesc_sync(sc, start, sc->sc_txq.t_cur, + BUS_DMASYNC_PREREAD|BUS_DMASYNC_PREWRITE); + + bus_space_write_4(sc->sc_bst, sc->sc_bsh, + AWIN_GMAC_DMA_TXPOLL, ~0U); +#ifdef DWC_GMAC_DEBUG + dwc_dump_status(sc); +#endif + } +} + +void +dwc_gmac_stop(struct ifnet *ifp, int disable) +{ + struct dwc_gmac_softc *sc = ifp->if_softc; + + ifp->if_flags &= ~IFF_RUNNING; + ifq_clr_oactive(&ifp->if_snd); + + bus_space_write_4(sc->sc_bst, sc->sc_bsh, + AWIN_GMAC_DMA_OPMODE, + bus_space_read_4(sc->sc_bst, sc->sc_bsh, + AWIN_GMAC_DMA_OPMODE) + & ~(GMAC_DMA_OP_TXSTART|GMAC_DMA_OP_RXSTART)); + bus_space_write_4(sc->sc_bst, sc->sc_bsh, + AWIN_GMAC_DMA_OPMODE, + bus_space_read_4(sc->sc_bst, sc->sc_bsh, + AWIN_GMAC_DMA_OPMODE) | GMAC_DMA_OP_FLUSHTX); + + mii_down(&sc->sc_mii); + dwc_gmac_reset_tx_ring(sc, &sc->sc_txq); + dwc_gmac_reset_rx_ring(sc, &sc->sc_rxq); +} + +/* + * Add m0 to the TX ring + */ +int +dwc_gmac_queue(struct dwc_gmac_softc *sc, struct mbuf *m0) +{ + struct dwc_gmac_dev_dmadesc *desc = NULL; + struct dwc_gmac_tx_data *data = NULL; + bus_dmamap_t map; + uint32_t flags, len, status; + int error, i, first; + +#ifdef DWC_GMAC_DEBUG + printf("%s: dwc_gmac_queue: adding mbuf chain %p\n", + sc->sc_dev.dv_xname, m0); +#endif + + first = sc->sc_txq.t_cur; + map = sc->sc_txq.t_data[first].td_map; + + error = bus_dmamap_load_mbuf(sc->sc_dmat, map, m0, + BUS_DMA_WRITE|BUS_DMA_NOWAIT); + if (error != 0) { + printf("%s: could not map mbuf (len: %d, error %d)\n", + sc->sc_dev.dv_xname, m0->m_pkthdr.len, error); + return error; + } + + if (sc->sc_txq.t_queued + map->dm_nsegs > DWGE_TX_RING_COUNT) { + bus_dmamap_unload(sc->sc_dmat, map); + return ENOBUFS; + } + + flags = DDESC_CNTL_TXFIRST|DDESC_CNTL_TXCHAIN; + status = 0; + for (i = 0; i < map->dm_nsegs; i++) { + data = &sc->sc_txq.t_data[sc->sc_txq.t_cur]; + desc = &sc->sc_txq.t_desc[sc->sc_txq.t_cur]; + + desc->ddesc_data = htole32(map->dm_segs[i].ds_addr); + len = (map->dm_segs[i].ds_len & DDESC_CNTL_SIZE1MASK) + << DDESC_CNTL_SIZE1SHIFT; + +#ifdef DWC_GMAC_DEBUG + printf("%s: enqueing desc #%d data %08lx " + "len %lu (flags: %08x, len: %08x)\n", + sc->sc_dev.dv_xname, sc->sc_txq.t_cur, + (unsigned long)map->dm_segs[i].ds_addr, + (unsigned long)map->dm_segs[i].ds_len, + flags, len); +#endif + + desc->ddesc_cntl = htole32(len|flags); + flags &= ~DDESC_CNTL_TXFIRST; + + /* + * Defer passing ownership of the first descriptor + * until we are done. + */ + desc->ddesc_status = htole32(status); + status |= DDESC_STATUS_OWNEDBYDEV; + + sc->sc_txq.t_queued++; + sc->sc_txq.t_cur = TX_NEXT(sc->sc_txq.t_cur); + } + + desc->ddesc_cntl |= htole32(DDESC_CNTL_TXLAST|DDESC_CNTL_TXINT); + + data->td_m = m0; + data->td_active = map; + + bus_dmamap_sync(sc->sc_dmat, map, 0, map->dm_mapsize, + BUS_DMASYNC_PREWRITE); + + /* Pass first to device */ + sc->sc_txq.t_desc[first].ddesc_status = + htole32(DDESC_STATUS_OWNEDBYDEV); + + return 0; +} + +/* + * If the interface is up and running, only modify the receive + * filter when setting promiscuous or debug mode. Otherwise fall + * through to ether_ioctl, which will reset the chip. + */ +int +dwc_gmac_ifflags_cb(struct arpcom *ac) +{ + struct ifnet *ifp = &ac->ac_if; + struct dwc_gmac_softc *sc = ifp->if_softc; + int change = ifp->if_flags ^ sc->sc_if_flags; + + if ((change & ~(IFF_CANTCHANGE|IFF_DEBUG)) != 0) + return ENETRESET; + if ((change & IFF_PROMISC) != 0) + dwc_gmac_setmulti(sc); + return 0; +} + +int +dwc_gmac_ioctl(struct ifnet *ifp, u_long cmd, caddr_t data) +{ + struct dwc_gmac_softc *sc = ifp->if_softc; + struct ifreq *ifr = (struct ifreq *)data; + int s, error = 0; + + s = splnet(); + + switch(cmd) { + case SIOCSIFADDR: + ifp->if_flags |= IFF_UP; + if (!(ifp->if_flags & IFF_RUNNING)) + dwc_gmac_init(ifp); + break; + + case SIOCSIFFLAGS: + if (ifp->if_flags & IFF_UP) { + if (ifp->if_flags & IFF_RUNNING) + error = ENETRESET; + else + dwc_gmac_init(ifp); + } else { + if (ifp->if_flags & IFF_RUNNING) + dwc_gmac_stop(ifp, 0); + } + break; + + case SIOCGIFMEDIA: + case SIOCSIFMEDIA: + error = ifmedia_ioctl(ifp, ifr, &sc->sc_mii.mii_media, cmd); + break; + + default: + error = ether_ioctl(ifp, &sc->sc_ac, cmd, data); + } + + if (error == ENETRESET) { + if (ifp->if_flags & IFF_RUNNING) + dwc_gmac_ifflags_cb(&sc->sc_ac); + error = 0; + } + + sc->sc_if_flags = sc->sc_ac.ac_if.if_flags; + splx(s); + return error; +} + +void +dwc_gmac_tx_intr(struct dwc_gmac_softc *sc) +{ + struct ifnet *ifp = &sc->sc_ac.ac_if; + struct dwc_gmac_tx_data *data; + struct dwc_gmac_dev_dmadesc *desc; + uint32_t status; + int i, nsegs; + + for (i = sc->sc_txq.t_next; sc->sc_txq.t_queued > 0; i = TX_NEXT(i)) { +#ifdef DWC_GMAC_DEBUG + printf("%s: dwc_gmac_tx_intr: checking desc #%d (t_queued: %d)\n", + sc->sc_dev.dv_xname, i, sc->sc_txq.t_queued); +#endif + + /* + * i+1 does not need to be a valid descriptor, + * this is just a special notion to just sync + * a single tx descriptor (i) + */ + dwc_gmac_txdesc_sync(sc, i, i+1, + BUS_DMASYNC_POSTREAD|BUS_DMASYNC_POSTWRITE); + + desc = &sc->sc_txq.t_desc[i]; + status = le32toh(desc->ddesc_status); + if (status & DDESC_STATUS_OWNEDBYDEV) + break; + + data = &sc->sc_txq.t_data[i]; + if (data->td_m == NULL) + continue; + + ifp->if_opackets++; + nsegs = data->td_active->dm_nsegs; + bus_dmamap_sync(sc->sc_dmat, data->td_active, 0, + data->td_active->dm_mapsize, BUS_DMASYNC_POSTWRITE); + bus_dmamap_unload(sc->sc_dmat, data->td_active); + +#ifdef DWC_GMAC_DEBUG + printf("%s: dwc_gmac_tx_intr: done with packet at desc #%d, " + "freeing mbuf %p\n", sc->sc_dev.dv_xname, i, data->td_m); +#endif + + m_freem(data->td_m); + data->td_m = NULL; + + sc->sc_txq.t_queued -= nsegs; + } + + sc->sc_txq.t_next = i; + + if (sc->sc_txq.t_queued < DWGE_TX_RING_COUNT) { + ifq_clr_oactive(&ifp->if_snd); + } +} + +void +dwc_gmac_rx_intr(struct dwc_gmac_softc *sc) +{ + struct ifnet *ifp = &sc->sc_ac.ac_if; + struct dwc_gmac_dev_dmadesc *desc; + struct dwc_gmac_rx_data *data; + bus_addr_t physaddr; + uint32_t status; + struct mbuf *m, *mnew; + int i, len, error; + struct mbuf_list ml = MBUF_LIST_INITIALIZER(); + + for (i = sc->sc_rxq.r_cur; ; i = RX_NEXT(i)) { + bus_dmamap_sync(sc->sc_dmat, sc->sc_dma_ring_map, + RX_DESC_OFFSET(i), sizeof(*desc), + BUS_DMASYNC_POSTREAD|BUS_DMASYNC_POSTWRITE); + desc = &sc->sc_rxq.r_desc[i]; + data = &sc->sc_rxq.r_data[i]; + + status = le32toh(desc->ddesc_status); + if (status & DDESC_STATUS_OWNEDBYDEV) + break; + + if (status & (DDESC_STATUS_RXERROR|DDESC_STATUS_RXTRUNCATED)) { +#ifdef DWC_GMAC_DEBUG + printf("%s: RX error: descriptor status %08x, skipping\n", + sc->sc_dev.dv_xname, status); +#endif + ifp->if_ierrors++; + goto skip; + } + + len = (status >> DDESC_STATUS_FRMLENSHIFT) + & DDESC_STATUS_FRMLENMSK; + +#ifdef DWC_GMAC_DEBUG + printf("%s: rx int: device is done with descriptor #%d, len: %d\n", + sc->sc_dev.dv_xname, i, len); +#endif + + /* + * Try to get a new mbuf before passing this one + * up, if that fails, drop the packet and reuse + * the existing one. + */ + MGETHDR(mnew, M_DONTWAIT, MT_DATA); + if (mnew == NULL) { + ifp->if_ierrors++; + goto skip; + } + MCLGET(mnew, M_DONTWAIT); + if ((mnew->m_flags & M_EXT) == 0) { + m_freem(mnew); + ifp->if_ierrors++; + goto skip; + } + + /* unload old DMA map */ + bus_dmamap_sync(sc->sc_dmat, data->rd_map, 0, + data->rd_map->dm_mapsize, BUS_DMASYNC_POSTREAD); + bus_dmamap_unload(sc->sc_dmat, data->rd_map); + + /* and reload with new mbuf */ + error = bus_dmamap_load(sc->sc_dmat, data->rd_map, + mtod(mnew, void*), MCLBYTES, NULL, + BUS_DMA_READ | BUS_DMA_NOWAIT); + if (error != 0) { + m_freem(mnew); + /* try to reload old mbuf */ + error = bus_dmamap_load(sc->sc_dmat, data->rd_map, + mtod(data->rd_m, void*), MCLBYTES, NULL, + BUS_DMA_READ | BUS_DMA_NOWAIT); + if (error != 0) { + panic("%s: could not load old rx mbuf", + sc->sc_dev.dv_xname); + } + ifp->if_ierrors++; + goto skip; + } + physaddr = data->rd_map->dm_segs[0].ds_addr; + + /* + * New mbuf loaded, update RX ring and continue + */ + m = data->rd_m; + data->rd_m = mnew; + desc->ddesc_data = htole32(physaddr); + + /* finalize mbuf */ +#ifdef __STRICT_ALIGNMENT + { + struct mbuf *m0; + m0 = m_devget(mtod(m, caddr_t), len, ETHER_ALIGN); + m_freem(m); + if (m0 == NULL) { + ifp->if_ierrors++; + goto skip; + } + m = m0; + } +#else + m->m_pkthdr.len = m->m_len = len; +#endif + + ml_enqueue(&ml, m); + +skip: + bus_dmamap_sync(sc->sc_dmat, data->rd_map, 0, + data->rd_map->dm_mapsize, BUS_DMASYNC_PREREAD); + desc->ddesc_cntl = htole32( + ((DWGE_MAX_PACKET & DDESC_CNTL_SIZE1MASK) + << DDESC_CNTL_SIZE1SHIFT) | DDESC_CNTL_RXCHAIN); + desc->ddesc_status = htole32(DDESC_STATUS_OWNEDBYDEV); + bus_dmamap_sync(sc->sc_dmat, sc->sc_dma_ring_map, + RX_DESC_OFFSET(i), sizeof(*desc), + BUS_DMASYNC_PREREAD|BUS_DMASYNC_PREWRITE); + } + + /* update RX pointer */ + sc->sc_rxq.r_cur = i; + + if_input(ifp, &ml); +} + +/* + * Reverse order of bits - http://aggregate.org/MAGIC/#Bit%20Reversal + */ +static uint32_t +bitrev32(uint32_t x) +{ + x = (((x & 0xaaaaaaaa) >> 1) | ((x & 0x55555555) << 1)); + x = (((x & 0xcccccccc) >> 2) | ((x & 0x33333333) << 2)); + x = (((x & 0xf0f0f0f0) >> 4) | ((x & 0x0f0f0f0f) << 4)); + x = (((x & 0xff00ff00) >> 8) | ((x & 0x00ff00ff) << 8)); + + return (x >> 16) | (x << 16); +} + +void +dwc_gmac_setmulti(struct dwc_gmac_softc *sc) +{ + struct ifnet * const ifp = &sc->sc_ac.ac_if; + struct ether_multi *enm; + struct ether_multistep step; + uint32_t hashes[2] = { 0, 0 }; + uint32_t ffilt, h; + int mcnt, s; + + s = splnet(); + + ffilt = bus_space_read_4(sc->sc_bst, sc->sc_bsh, AWIN_GMAC_MAC_FFILT); + + if (ifp->if_flags & IFF_PROMISC) { + ffilt |= AWIN_GMAC_MAC_FFILT_PR; + goto special_filter; + } + + ifp->if_flags &= ~IFF_ALLMULTI; + ffilt &= ~(AWIN_GMAC_MAC_FFILT_PM|AWIN_GMAC_MAC_FFILT_PR); + + bus_space_write_4(sc->sc_bst, sc->sc_bsh, AWIN_GMAC_MAC_HTLOW, 0); + bus_space_write_4(sc->sc_bst, sc->sc_bsh, AWIN_GMAC_MAC_HTHIGH, 0); + + ETHER_FIRST_MULTI(step, &sc->sc_ac, enm); + mcnt = 0; + while (enm != NULL) { + if (memcmp(enm->enm_addrlo, enm->enm_addrhi, + ETHER_ADDR_LEN) != 0) { + ffilt |= AWIN_GMAC_MAC_FFILT_PM; + ifp->if_flags |= IFF_ALLMULTI; + goto special_filter; + } + + h = bitrev32( + ~ether_crc32_le(enm->enm_addrlo, ETHER_ADDR_LEN) + ) >> 26; + hashes[h >> 5] |= (1 << (h & 0x1f)); + + mcnt++; + ETHER_NEXT_MULTI(step, enm); + } + + if (mcnt) + ffilt |= AWIN_GMAC_MAC_FFILT_HMC; + else + ffilt &= ~AWIN_GMAC_MAC_FFILT_HMC; + + bus_space_write_4(sc->sc_bst, sc->sc_bsh, AWIN_GMAC_MAC_FFILT, ffilt); + bus_space_write_4(sc->sc_bst, sc->sc_bsh, AWIN_GMAC_MAC_HTLOW, + hashes[0]); + bus_space_write_4(sc->sc_bst, sc->sc_bsh, AWIN_GMAC_MAC_HTHIGH, + hashes[1]); + sc->sc_if_flags = sc->sc_ac.ac_if.if_flags; + + splx(s); + +#ifdef DWC_GMAC_DEBUG + dwc_gmac_dump_ffilt(sc, ffilt); +#endif + return; + +special_filter: +#ifdef DWC_GMAC_DEBUG + dwc_gmac_dump_ffilt(sc, ffilt); +#endif + /* no MAC hashes, ALLMULTI or PROMISC */ + bus_space_write_4(sc->sc_bst, sc->sc_bsh, AWIN_GMAC_MAC_FFILT, + ffilt); + bus_space_write_4(sc->sc_bst, sc->sc_bsh, AWIN_GMAC_MAC_HTLOW, + 0xffffffff); + bus_space_write_4(sc->sc_bst, sc->sc_bsh, AWIN_GMAC_MAC_HTHIGH, + 0xffffffff); + sc->sc_if_flags = sc->sc_ac.ac_if.if_flags; + splx(s); +} + +int +dwc_gmac_intr(struct dwc_gmac_softc *sc) +{ + uint32_t status, dma_status; + int rv = 0; + + status = bus_space_read_4(sc->sc_bst, sc->sc_bsh, AWIN_GMAC_MAC_INTR); + if (status & AWIN_GMAC_MII_IRQ) { + (void)bus_space_read_4(sc->sc_bst, sc->sc_bsh, + AWIN_GMAC_MII_STATUS); + rv = 1; + mii_pollstat(&sc->sc_mii); + } + + dma_status = bus_space_read_4(sc->sc_bst, sc->sc_bsh, + AWIN_GMAC_DMA_STATUS); + + if (dma_status & (GMAC_DMA_INT_NIE|GMAC_DMA_INT_AIE)) + rv = 1; + + if (dma_status & GMAC_DMA_INT_TIE) + dwc_gmac_tx_intr(sc); + + if (dma_status & GMAC_DMA_INT_RIE) + dwc_gmac_rx_intr(sc); + + /* + * Check error conditions + */ + if (dma_status & GMAC_DMA_INT_ERRORS) { + sc->sc_ac.ac_if.if_oerrors++; +#ifdef DWC_GMAC_DEBUG + dwc_dump_and_abort(sc, "interrupt error condition"); +#endif + } + + /* ack interrupt */ + if (dma_status) + bus_space_write_4(sc->sc_bst, sc->sc_bsh, + AWIN_GMAC_DMA_STATUS, dma_status & GMAC_DMA_INT_MASK); + + /* + * Get more packets + */ + if (rv) + sc->sc_ac.ac_if.if_start(&sc->sc_ac.ac_if); + + return rv; +} + +#ifdef DWC_GMAC_DEBUG +void +dwc_gmac_dump_dma(struct dwc_gmac_softc *sc) +{ + printf("%s: busmode: %08x\n", sc->sc_dev.dv_xname, + bus_space_read_4(sc->sc_bst, sc->sc_bsh, AWIN_GMAC_DMA_BUSMODE)); + printf("%s: tx poll: %08x\n", sc->sc_dev.dv_xname, + bus_space_read_4(sc->sc_bst, sc->sc_bsh, AWIN_GMAC_DMA_TXPOLL)); + printf("%s: rx poll: %08x\n", sc->sc_dev.dv_xname, + bus_space_read_4(sc->sc_bst, sc->sc_bsh, AWIN_GMAC_DMA_RXPOLL)); + printf("%s: rx descriptors: %08x\n", sc->sc_dev.dv_xname, + bus_space_read_4(sc->sc_bst, sc->sc_bsh, AWIN_GMAC_DMA_RX_ADDR)); + printf("%s: tx descriptors: %08x\n", sc->sc_dev.dv_xname, + bus_space_read_4(sc->sc_bst, sc->sc_bsh, AWIN_GMAC_DMA_TX_ADDR)); + printf("%s: status: %08x\n", sc->sc_dev.dv_xname, + bus_space_read_4(sc->sc_bst, sc->sc_bsh, AWIN_GMAC_DMA_STATUS)); + printf("%s: op mode: %08x\n", sc->sc_dev.dv_xname, + bus_space_read_4(sc->sc_bst, sc->sc_bsh, AWIN_GMAC_DMA_OPMODE)); + printf("%s: int enable: %08x\n", sc->sc_dev.dv_xname, + bus_space_read_4(sc->sc_bst, sc->sc_bsh, AWIN_GMAC_DMA_INTENABLE)); + printf("%s: cur tx: %08x\n", sc->sc_dev.dv_xname, + bus_space_read_4(sc->sc_bst, sc->sc_bsh, AWIN_GMAC_DMA_CUR_TX_DESC)); + printf("%s: cur rx: %08x\n", sc->sc_dev.dv_xname, + bus_space_read_4(sc->sc_bst, sc->sc_bsh, AWIN_GMAC_DMA_CUR_RX_DESC)); + printf("%s: cur tx buffer: %08x\n", sc->sc_dev.dv_xname, + bus_space_read_4(sc->sc_bst, sc->sc_bsh, AWIN_GMAC_DMA_CUR_TX_BUFADDR)); + printf("%s: cur rx buffer: %08x\n", sc->sc_dev.dv_xname, + bus_space_read_4(sc->sc_bst, sc->sc_bsh, AWIN_GMAC_DMA_CUR_RX_BUFADDR)); +} + +void +dwc_gmac_dump_tx_desc(struct dwc_gmac_softc *sc) +{ + int i; + + printf("%s: TX queue: cur=%d, next=%d, queued=%d\n", + sc->sc_dev.dv_xname, sc->sc_txq.t_cur, + sc->sc_txq.t_next, sc->sc_txq.t_queued); + printf("%s: TX DMA descriptors:\n", sc->sc_dev.dv_xname); + for (i = 0; i < DWGE_TX_RING_COUNT; i++) { + struct dwc_gmac_dev_dmadesc *desc = &sc->sc_txq.t_desc[i]; + printf("#%d (%08lx): status: %08x cntl: %08x " + "data: %08x next: %08x\n", + i, sc->sc_txq.t_physaddr + + i*sizeof(struct dwc_gmac_dev_dmadesc), + le32toh(desc->ddesc_status), le32toh(desc->ddesc_cntl), + le32toh(desc->ddesc_data), le32toh(desc->ddesc_next)); + } +} + +void +dwc_gmac_dump_rx_desc(struct dwc_gmac_softc *sc) +{ + int i; + + printf("%s: RX queue: cur=%d, next=%d\n", sc->sc_dev.dv_xname, + sc->sc_rxq.r_cur, sc->sc_rxq.r_next); + printf("%s: RX DMA descriptors:\n", sc->sc_dev.dv_xname); + for (i = 0; i < DWGE_RX_RING_COUNT; i++) { + struct dwc_gmac_dev_dmadesc *desc = &sc->sc_rxq.r_desc[i]; + printf("#%d (%08lx): status: %08x cntl: %08x " + "data: %08x next: %08x\n", + i, sc->sc_rxq.r_physaddr + + i*sizeof(struct dwc_gmac_dev_dmadesc), + le32toh(desc->ddesc_status), le32toh(desc->ddesc_cntl), + le32toh(desc->ddesc_data), le32toh(desc->ddesc_next)); + } +} + +void +dwc_dump_status(struct dwc_gmac_softc *sc) +{ + uint32_t status = bus_space_read_4(sc->sc_bst, sc->sc_bsh, + AWIN_GMAC_MAC_INTR); + uint32_t dma_status = bus_space_read_4(sc->sc_bst, sc->sc_bsh, + AWIN_GMAC_DMA_STATUS); + char buf[200]; + + /* print interrupt state */ + snprintb(buf, sizeof(buf), "\177\20" + "b\x10""NI\0" + "b\x0f""AI\0" + "b\x0e""ER\0" + "b\x0d""FB\0" + "b\x0a""ET\0" + "b\x09""RW\0" + "b\x08""RS\0" + "b\x07""RU\0" + "b\x06""RI\0" + "b\x05""UN\0" + "b\x04""OV\0" + "b\x03""TJ\0" + "b\x02""TU\0" + "b\x01""TS\0" + "b\x00""TI\0" + "\0", dma_status); + printf("%s: INTR status: %08x, DMA status: %s\n", + sc->sc_dev.dv_xname, status, buf); +} + +void +dwc_dump_and_abort(struct dwc_gmac_softc *sc, const char *msg) +{ + dwc_dump_status(sc); + dwc_gmac_dump_ffilt(sc, + bus_space_read_4(sc->sc_bst, sc->sc_bsh, AWIN_GMAC_MAC_FFILT)); + dwc_gmac_dump_dma(sc); + dwc_gmac_dump_tx_desc(sc); + dwc_gmac_dump_rx_desc(sc); + + panic("%s", msg); +} + +void dwc_gmac_dump_ffilt(struct dwc_gmac_softc *sc, uint32_t ffilt) +{ + char buf[200]; + + /* print filter setup */ + snprintb(buf, sizeof(buf), "\177\20" + "b\x1f""RA\0" + "b\x0a""HPF\0" + "b\x09""SAF\0" + "b\x08""SAIF\0" + "b\x05""DBF\0" + "b\x04""PM\0" + "b\x03""DAIF\0" + "b\x02""HMC\0" + "b\x01""HUC\0" + "b\x00""PR\0" + "\0", ffilt); + printf("%s: FFILT: %s\n", sc->sc_dev.dv_xname, buf); +} +#endif diff --git a/sys/dev/ic/dwc_gmac_reg.h b/sys/dev/ic/dwc_gmac_reg.h new file mode 100644 index 00000000000..ed8c139bc1a --- /dev/null +++ b/sys/dev/ic/dwc_gmac_reg.h @@ -0,0 +1,227 @@ +/* $OpenBSD: dwc_gmac_reg.h,v 1.1 2016/08/13 20:35:57 kettenis Exp $ */ +/* $NetBSD: dwc_gmac_reg.h,v 1.15 2015/11/21 16:04:11 martin Exp $ */ + +/*- + * Copyright (c) 2013, 2014 The NetBSD Foundation, Inc. + * All rights reserved. + * + * This code is derived from software contributed to The NetBSD Foundation + * by Matt Thomas of 3am Software Foundry and Martin Husemann. + * + * Redistribution and use in source and binary forms, with or without + * modification, are permitted provided that the following conditions + * are met: + * 1. Redistributions of source code must retain the above copyright + * notice, this list of conditions and the following disclaimer. + * 2. Redistributions in binary form must reproduce the above copyright + * notice, this list of conditions and the following disclaimer in the + * documentation and/or other materials provided with the distribution. + * + * THIS SOFTWARE IS PROVIDED BY THE NETBSD FOUNDATION, INC. AND CONTRIBUTORS + * ``AS IS'' AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED + * TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR + * PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE FOUNDATION OR CONTRIBUTORS + * BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR + * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF + * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS + * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN + * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) + * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE + * POSSIBILITY OF SUCH DAMAGE. + */ + +#define AWIN_GMAC_MAC_CONF 0x0000 +#define AWIN_GMAC_MAC_FFILT 0x0004 +#define AWIN_GMAC_MAC_HTHIGH 0x0008 +#define AWIN_GMAC_MAC_HTLOW 0x000c +#define AWIN_GMAC_MAC_MIIADDR 0x0010 +#define AWIN_GMAC_MAC_MIIDATA 0x0014 +#define AWIN_GMAC_MAC_FLOWCTRL 0x0018 +#define AWIN_GMAC_MAC_VLANTAG 0x001c +#define AWIN_GMAC_MAC_VERSION 0x0020 /* not always implemented? */ +#define AWIN_GMAC_MAC_INTR 0x0038 +#define AWIN_GMAC_MAC_INTMASK 0x003c +#define AWIN_GMAC_MAC_ADDR0HI 0x0040 +#define AWIN_GMAC_MAC_ADDR0LO 0x0044 +#define AWIN_GMAC_MII_STATUS 0x00D8 + +#define AWIN_GMAC_MAC_CONF_DISABLEJABBER (1 << 22) /* jabber disable */ +#define AWIN_GMAC_MAC_CONF_FRAMEBURST (1 << 21) /* allow TX frameburst when + in half duplex mode */ +#define AWIN_GMAC_MAC_CONF_MIISEL (1 << 15) /* select MII phy */ +#define AWIN_GMAC_MAC_CONF_FES100 (1 << 14) /* 100 mbit mode */ +#define AWIN_GMAC_MAC_CONF_DISABLERXOWN (1 << 13) /* do not receive our own + TX frames in half duplex + mode */ +#define AWIN_GMAC_MAC_CONF_FULLDPLX (1 << 11) /* select full duplex */ +#define AWIN_GMAC_MAC_CONF_ACS (1 << 7) /* auto pad/CRC stripping */ +#define AWIN_GMAC_MAC_CONF_TXENABLE (1 << 3) /* enable TX dma engine */ +#define AWIN_GMAC_MAC_CONF_RXENABLE (1 << 2) /* enable RX dma engine */ + +#define AWIN_GMAC_MAC_FFILT_RA (1U << 31) /* receive all mode */ +#define AWIN_GMAC_MAC_FFILT_HPF (1 << 10) /* hash or perfect filter */ +#define AWIN_GMAC_MAC_FFILT_SAF (1 << 9) /* source address filter */ +#define AWIN_GMAC_MAC_FFILT_SAIF (1 << 8) /* inverse filtering */ +#define AWIN_GMAC_MAC_FFILT_DBF (1 << 5) /* disable broadcast frames */ +#define AWIN_GMAC_MAC_FFILT_PM (1 << 4) /* promiscious multicast */ +#define AWIN_GMAC_MAC_FFILT_DAIF (1 << 3) /* DA inverse filtering */ +#define AWIN_GMAC_MAC_FFILT_HMC (1 << 2) /* multicast hash compare */ +#define AWIN_GMAC_MAC_FFILT_HUC (1 << 1) /* unicast hash compare */ +#define AWIN_GMAC_MAC_FFILT_PR (1 << 0) /* promiscious mode */ + +#define AWIN_GMAC_MAC_INT_LPI (1 << 10) +#define AWIN_GMAC_MAC_INT_TSI (1 << 9) +#define AWIN_GMAC_MAC_INT_ANEG (1 << 2) +#define AWIN_GMAC_MAC_INT_LINKCHG (1 << 1) +#define AWIN_GMAC_MAC_INT_RGSMII (1 << 0) + +#define AWIN_GMAC_MAC_FLOWCTRL_PAUSE_SHIFT 16 +#define AWIN_GMAC_MAC_FLOWCTRL_PAUSE_MASK 0xffff +#define AWIN_GMAC_MAC_FLOWCTRL_RFE (1 << 2) +#define AWIN_GMAC_MAC_FLOWCTRL_TFE (1 << 1) +#define AWIN_GMAC_MAC_FLOWCTRL_BUSY (1 << 0) + +#define AWIN_GMAC_DMA_BUSMODE 0x1000 +#define AWIN_GMAC_DMA_TXPOLL 0x1004 +#define AWIN_GMAC_DMA_RXPOLL 0x1008 +#define AWIN_GMAC_DMA_RX_ADDR 0x100c +#define AWIN_GMAC_DMA_TX_ADDR 0x1010 +#define AWIN_GMAC_DMA_STATUS 0x1014 +#define AWIN_GMAC_DMA_OPMODE 0x1018 +#define AWIN_GMAC_DMA_INTENABLE 0x101c +#define AWIN_GMAC_DMA_CUR_TX_DESC 0x1048 +#define AWIN_GMAC_DMA_CUR_RX_DESC 0x104c +#define AWIN_GMAC_DMA_CUR_TX_BUFADDR 0x1050 +#define AWIN_GMAC_DMA_CUR_RX_BUFADDR 0x1054 +#define AWIN_GMAC_DMA_HWFEATURES 0x1058 /* not always implemented? */ + +#define GMAC_MII_PHY_SHIFT 11 +#define GMAC_MII_PHY_MASK 0x1f +#define GMAC_MII_REG_SHIFT 6 +#define GMAC_MII_REG_MASK 0x1f + +#define GMAC_MII_BUSY (1 << 0) +#define GMAC_MII_WRITE (1 << 1) +#define GMAC_MII_CLK_60_100M_DIV42 0x0 +#define GMAC_MII_CLK_100_150M_DIV62 0x1 +#define GMAC_MII_CLK_25_35M_DIV16 0x2 +#define GMAC_MII_CLK_35_60M_DIV26 0x3 +#define GMAC_MII_CLK_150_250M_DIV102 0x4 +#define GMAC_MII_CLK_250_300M_DIV124 0x5 +#define GMAC_MII_CLK_DIV4 0x8 +#define GMAC_MII_CLK_DIV6 0x9 +#define GMAC_MII_CLK_DIV8 0xa +#define GMAC_MII_CLK_DIV10 0xb +#define GMAC_MII_CLK_DIV12 0xc +#define GMAC_MII_CLK_DIV14 0xd +#define GMAC_MII_CLK_DIV16 0xe +#define GMAC_MII_CLK_DIV18 0xf +#define GMAC_MII_CLKMASK_SHIFT 2 +#define GMAC_MII_CLKMASK_MASK 0xf + +#define GMAC_BUSMODE_4PBL (1 << 24) +#define GMAC_BUSMODE_RPBL_SHIFT 17 +#define GMAC_BUSMODE_RPBL_MASK 0x3f +#define GMAC_BUSMODE_FIXEDBURST (1 << 16) +#define GMAC_BUSMODE_PRIORXTX_SHIFT 14 +#define GMAC_BUSMODE_PRIORXTX_MASK 0x3 +#define GMAC_BUSMODE_PRIORXTX_41 3 +#define GMAC_BUSMODE_PRIORXTX_31 2 +#define GMAC_BUSMODE_PRIORXTX_21 1 +#define GMAC_BUSMODE_PRIORXTX_11 0 +#define GMAC_BUSMODE_PBL_SHIFT 8 +#define GMAC_BUSMODE_PBL_MASK 0x3f /* possible DMA + burst len */ +#define GMAC_BUSMODE_RESET (1 << 0) + +#define AWIN_GMAC_MII_IRQ (1 << 0) + + +#define GMAC_DMA_OP_DISABLECSDROP (1 << 26) /* disable dropping of + frames with TCP/IP + checksum errors */ +#define GMAC_DMA_OP_RXSTOREFORWARD (1 << 25) /* start RX when a + full frame is available */ +#define GMAC_DMA_OP_DISABLERXFLUSH (1 << 24) /* Do not drop frames + when out of RX descr. */ +#define GMAC_DMA_OP_TXSTOREFORWARD (1 << 21) /* start TX when a + full frame is available */ +#define GMAC_DMA_OP_FLUSHTX (1 << 20) /* flush TX fifo */ +#define GMAC_DMA_OP_TXSTART (1 << 13) /* start TX DMA engine */ +#define GMAC_DMA_OP_RXSTART (1 << 1) /* start RX DMA engine */ + +#define GMAC_DMA_INT_NIE (1 << 16) /* Normal/Summary */ +#define GMAC_DMA_INT_AIE (1 << 15) /* Abnormal/Summary */ +#define GMAC_DMA_INT_ERE (1 << 14) /* Early receive */ +#define GMAC_DMA_INT_FBE (1 << 13) /* Fatal bus error */ +#define GMAC_DMA_INT_ETE (1 << 10) /* Early transmit */ +#define GMAC_DMA_INT_RWE (1 << 9) /* Receive watchdog */ +#define GMAC_DMA_INT_RSE (1 << 8) /* Receive stopped */ +#define GMAC_DMA_INT_RUE (1 << 7) /* Receive buffer unavail. */ +#define GMAC_DMA_INT_RIE (1 << 6) /* Receive interrupt */ +#define GMAC_DMA_INT_UNE (1 << 5) /* Tx underflow */ +#define GMAC_DMA_INT_OVE (1 << 4) /* Receive overflow */ +#define GMAC_DMA_INT_TJE (1 << 3) /* Transmit jabber */ +#define GMAC_DMA_INT_TUE (1 << 2) /* Transmit buffer unavail. */ +#define GMAC_DMA_INT_TSE (1 << 1) /* Transmit stopped */ +#define GMAC_DMA_INT_TIE (1 << 0) /* Transmit interrupt */ + +#define GMAC_DMA_INT_MASK 0x1ffff /* all possible intr bits */ + +struct dwc_gmac_dev_dmadesc { + uint32_t ddesc_status; +/* both: */ +#define DDESC_STATUS_OWNEDBYDEV (1U << 31) + +/* for RX descriptors */ +#define DDESC_STATUS_DAFILTERFAIL (1 << 30) +#define DDESC_STATUS_FRMLENMSK 0x3fff +#define DDESC_STATUS_FRMLENSHIFT 16 +#define DDESC_STATUS_RXERROR (1 << 15) +#define DDESC_STATUS_RXTRUNCATED (1 << 14) +#define DDESC_STATUS_SAFILTERFAIL (1 << 13) +#define DDESC_STATUS_RXIPC_GIANTFRAME (1 << 12) +#define DDESC_STATUS_RXDAMAGED (1 << 11) +#define DDESC_STATUS_RXVLANTAG (1 << 10) +#define DDESC_STATUS_RXFIRST (1 << 9) +#define DDESC_STATUS_RXLAST (1 << 8) +#define DDESC_STATUS_RXIPC_GIANT (1 << 7) +#define DDESC_STATUS_RXCOLLISION (1 << 6) +#define DDESC_STATUS_RXFRAMEETHER (1 << 5) +#define DDESC_STATUS_RXWATCHDOG (1 << 4) +#define DDESC_STATUS_RXMIIERROR (1 << 3) +#define DDESC_STATUS_RXDRIBBLING (1 << 2) +#define DDESC_STATUS_RXCRC (1 << 1) + + uint32_t ddesc_cntl; + +/* for TX descriptors */ +#define DDESC_CNTL_TXINT (1U << 31) +#define DDESC_CNTL_TXLAST (1 << 30) +#define DDESC_CNTL_TXFIRST (1 << 29) +#define DDESC_CNTL_TXCHECKINSCTRL __BITS(27,28) + +#define DDESC_TXCHECK_DISABLED 0 +#define DDESC_TXCHECK_IP 1 +#define DDESC_TXCHECK_IP_NO_PSE 2 +#define DDESC_TXCHECK_FULL 3 + +#define DDESC_CNTL_TXCRCDIS (1 << 26) +#define DDESC_CNTL_TXRINGEND (1 << 25) +#define DDESC_CNTL_TXCHAIN (1 << 24) +#define DDESC_CNTL_TXDISPAD (1 << 23) + +/* for RX descriptors */ +#define DDESC_CNTL_RXINTDIS (1U << 31) +#define DDESC_CNTL_RXRINGEND (1 << 25) +#define DDESC_CNTL_RXCHAIN (1 << 24) + +/* both */ +#define DDESC_CNTL_SIZE1MASK 0x7ff +#define DDESC_CNTL_SIZE1SHIFT 0 +#define DDESC_CNTL_SIZE2MASK 0x7ff +#define DDESC_CNTL_SIZE2SHIFT 11 + + uint32_t ddesc_data; /* pointer to buffer data */ + uint32_t ddesc_next; /* link to next descriptor */ +}; diff --git a/sys/dev/ic/dwc_gmac_var.h b/sys/dev/ic/dwc_gmac_var.h new file mode 100644 index 00000000000..e2b2d2887ca --- /dev/null +++ b/sys/dev/ic/dwc_gmac_var.h @@ -0,0 +1,95 @@ +/* $OpenBSD: dwc_gmac_var.h,v 1.1 2016/08/13 20:35:57 kettenis Exp $ */ +/* $NetBSD: dwc_gmac_var.h,v 1.6 2014/11/22 18:31:03 jmcneill Exp $ */ + +/*- + * Copyright (c) 2013, 2014 The NetBSD Foundation, Inc. + * All rights reserved. + * + * This code is derived from software contributed to The NetBSD Foundation + * by Matt Thomas of 3am Software Foundry and Martin Husemann. + * + * Redistribution and use in source and binary forms, with or without + * modification, are permitted provided that the following conditions + * are met: + * 1. Redistributions of source code must retain the above copyright + * notice, this list of conditions and the following disclaimer. + * 2. Redistributions in binary form must reproduce the above copyright + * notice, this list of conditions and the following disclaimer in the + * documentation and/or other materials provided with the distribution. + * + * THIS SOFTWARE IS PROVIDED BY THE NETBSD FOUNDATION, INC. AND CONTRIBUTORS + * ``AS IS'' AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED + * TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR + * PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE FOUNDATION OR CONTRIBUTORS + * BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR + * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF + * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS + * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN + * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) + * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE + * POSSIBILITY OF SUCH DAMAGE. + */ + + +/* + * We could use 1024 DMA descriptors to fill up an 8k page (each is 16 byte). + * However, on TX we probably will not need that many, and on RX we allocate + * a full mbuf cluster for each, so secondary memory consumption will grow + * rapidly. + * So currently we waste half a page of dma memory and consume 512k Byte of + * RAM for mbuf clusters. + * XXX Maybe fine-tune later, or reconsider unsharing of RX/TX dmamap. + */ +#define DWGE_RX_RING_COUNT 256 +#define DWGE_TX_RING_COUNT 256 +#define DWGE_TOTAL_RING_COUNT \ + (DWGE_RX_RING_COUNT + DWGE_TX_RING_COUNT) + +#define DWGE_MAX_PACKET 0x7ff + + + +struct dwc_gmac_rx_data { + bus_dmamap_t rd_map; + struct mbuf *rd_m; +}; + +struct dwc_gmac_tx_data { + bus_dmamap_t td_map; + bus_dmamap_t td_active; + struct mbuf *td_m; +}; + +struct dwc_gmac_tx_ring { + bus_addr_t t_physaddr; /* PA of TX ring start */ + struct dwc_gmac_dev_dmadesc *t_desc; /* VA of TX ring start */ + struct dwc_gmac_tx_data t_data[DWGE_TX_RING_COUNT]; + int t_cur, t_next, t_queued; +}; + +struct dwc_gmac_rx_ring { + bus_addr_t r_physaddr; /* PA of RX ring start */ + struct dwc_gmac_dev_dmadesc *r_desc; /* VA of RX ring start */ + struct dwc_gmac_rx_data r_data[DWGE_RX_RING_COUNT]; + int r_cur, r_next; + struct mutex r_mtx; +}; + +struct dwc_gmac_softc { + struct device sc_dev; + bus_space_tag_t sc_bst; + bus_space_handle_t sc_bsh; + bus_dma_tag_t sc_dmat; + struct arpcom sc_ac; + struct mii_data sc_mii; + struct mutex sc_mdio_lock; + bus_dmamap_t sc_dma_ring_map; /* common dma memory for RX */ + bus_dma_segment_t sc_dma_ring_seg; /* and TX ring */ + struct dwc_gmac_rx_ring sc_rxq; + struct dwc_gmac_tx_ring sc_txq; + short sc_if_flags; /* shadow of ether flags */ + uint16_t sc_mii_clk; +}; + +void dwc_gmac_attach(struct dwc_gmac_softc*, uint32_t /*mii_clk*/); +int dwc_gmac_intr(struct dwc_gmac_softc*); -- cgit v1.2.3