diff options
author | Mike Larkin <mlarkin@cvs.openbsd.org> | 2021-09-02 10:11:22 +0000 |
---|---|---|
committer | Mike Larkin <mlarkin@cvs.openbsd.org> | 2021-09-02 10:11:22 +0000 |
commit | 46d0ee32ec742b686559f80c1d72fb49685d6a4a (patch) | |
tree | 1182de4ecbe9810d6609d1f4c8c55035faec30eb /sys/dev | |
parent | 5117d1a54dd23a9d5c03888dc09b213ffd0f09e5 (diff) |
aq(4) driver for Aquantia 1/2.5/5/10Gb/s PCIe ethernet adapters
Adds support for Aquantia AQC1xx family of PCIe ethernet adapters. This
driver supports 1Gbps through 10Gbps modes of operation based on the
hardware and media/switch capabilities.
The initial code was ported from NetBSD, with jmatthew@ finishing up
the Tx/Rx ring support and interrupt handler routine.
The driver only supports devices using firmware V2.
This diff enables aq(4) on riscv64 and amd64, the only platforms where
I have tested the driver, but it likely works on other architectures
as well.
Diffstat (limited to 'sys/dev')
-rw-r--r-- | sys/dev/pci/files.pci | 7 | ||||
-rw-r--r-- | sys/dev/pci/if_aq_pci.c | 2796 |
2 files changed, 2802 insertions, 1 deletions
diff --git a/sys/dev/pci/files.pci b/sys/dev/pci/files.pci index a8c82651d3d..0bde7450cc8 100644 --- a/sys/dev/pci/files.pci +++ b/sys/dev/pci/files.pci @@ -1,4 +1,4 @@ -# $OpenBSD: files.pci,v 1.354 2021/04/24 09:37:46 jmatthew Exp $ +# $OpenBSD: files.pci,v 1.355 2021/09/02 10:11:21 mlarkin Exp $ # $NetBSD: files.pci,v 1.20 1996/09/24 17:47:15 christos Exp $ # # Config file and device description for machine-independent PCI code. @@ -230,6 +230,11 @@ file dev/pci/mpii.c mpii attach sili at pci with sili_pci file dev/pci/sili_pci.c sili_pci +# Aquantia/Atlantic 10-Gigabit Ethernet +device aq: ether, ifnet, ifmedia, mii +attach aq at pci +file dev/pci/if_aq_pci.c aq + # Ethernet driver for DC21040-based boards device de: ether, ifnet, ifmedia attach de at pci diff --git a/sys/dev/pci/if_aq_pci.c b/sys/dev/pci/if_aq_pci.c new file mode 100644 index 00000000000..373618581e0 --- /dev/null +++ b/sys/dev/pci/if_aq_pci.c @@ -0,0 +1,2796 @@ +/* $OpenBSD: if_aq_pci.c,v 1.1 2021/09/02 10:11:21 mlarkin Exp $ */ +/* $NetBSD: if_aq.c,v 1.27 2021/06/16 00:21:18 riastradh Exp $ */ + +/* + * Copyright (c) 2021 Jonathan Matthew <jonathan@d14n.org> + * Copyright (c) 2021 Mike Larkin <mlarkin@openbsd.org> + * + * Permission to use, copy, modify, and distribute this software for any + * purpose with or without fee is hereby granted, provided that the above + * copyright notice and this permission notice appear in all copies. + * + * THE SOFTWARE IS PROVIDED "AS IS" AND THE AUTHOR DISCLAIMS ALL WARRANTIES + * WITH REGARD TO THIS SOFTWARE INCLUDING ALL IMPLIED WARRANTIES OF + * MERCHANTABILITY AND FITNESS. IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR + * ANY SPECIAL, DIRECT, INDIRECT, OR CONSEQUENTIAL DAMAGES OR ANY DAMAGES + * WHATSOEVER RESULTING FROM LOSS OF USE, DATA OR PROFITS, WHETHER IN AN + * ACTION OF CONTRACT, NEGLIGENCE OR OTHER TORTIOUS ACTION, ARISING OUT OF + * OR IN CONNECTION WITH THE USE OR PERFORMANCE OF THIS SOFTWARE. + */ + +/** + * aQuantia Corporation Network Driver + * Copyright (C) 2014-2017 aQuantia Corporation. All rights reserved + * + * Redistribution and use in source and binary forms, with or without + * modification, are permitted provided that the following conditions + * are met: + * + * (1) Redistributions of source code must retain the above + * copyright notice, this list of conditions and the following + * disclaimer. + * + * (2) Redistributions in binary form must reproduce the above + * copyright notice, this list of conditions and the following + * disclaimer in the documentation and/or other materials provided + * with the distribution. + * + * (3) The name of the author may not be used to endorse or promote + * products derived from this software without specific prior + * written permission. + * + * THIS SOFTWARE IS PROVIDED BY THE AUTHOR ``AS IS'' AND ANY EXPRESS + * OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED + * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE + * ARE DISCLAIMED. IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR ANY + * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL + * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE + * GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS + * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, + * WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING + * NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS + * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. + * + */ + +/*- + * Copyright (c) 2020 Ryo Shimizu <ryo@nerv.org> + * All rights reserved. + * + * Redistribution and use in source and binary forms, with or without + * modification, are permitted provided that the following conditions + * are met: + * 1. Redistributions of source code must retain the above copyright + * notice, this list of conditions and the following disclaimer. + * 2. Redistributions in binary form must reproduce the above copyright + * notice, this list of conditions and the following disclaimer in the + * documentation and/or other materials provided with the distribution. + * + * THIS SOFTWARE IS PROVIDED BY THE AUTHOR ``AS IS'' AND ANY EXPRESS OR + * IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED + * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE + * DISCLAIMED. IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR ANY DIRECT, + * INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES + * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR + * SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) + * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, + * STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING + * IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE + * POSSIBILITY OF SUCH DAMAGE. + */ +#include <sys/types.h> +#include <sys/device.h> +#include <sys/param.h> +#include <sys/kernel.h> +#include <sys/sockio.h> +#include <sys/systm.h> + +#include <net/if.h> +#include <net/if_media.h> + +#include <netinet/in.h> +#include <netinet/if_ether.h> + +#include <dev/pci/pcireg.h> +#include <dev/pci/pcivar.h> +#include <dev/pci/pcidevs.h> + +/* #define AQ_DEBUG 1 */ +#ifdef AQ_DEBUG +#define DPRINTF(x) printf x +#else +#define DPRINTF(x) +#endif /* AQ_DEBUG */ + +#define DEVNAME(_s) ((_s)->sc_dev.dv_xname) + +#define AQ_BAR0 0x10 +#define AQ_MAXQ 8 + +#define AQ_TXD_NUM 2048 +#define AQ_RXD_NUM 2048 + +#define AQ_TX_MAX_SEGMENTS 1 /* XXX */ + +#define AQ_LINKSTAT_IRQ 31 + +#define RPF_ACTION_HOST 1 + +#define AQ_FW_SOFTRESET_REG 0x0000 +#define AQ_FW_SOFTRESET_DIS (1 << 14) +#define AQ_FW_SOFTRESET_RESET (1 << 15) +#define AQ_FW_VERSION_REG 0x0018 +#define AQ_HW_REVISION_REG 0x001c +#define AQ_GLB_NVR_INTERFACE1_REG 0x0100 +#define AQ_FW_MBOX_CMD_REG 0x0200 +#define AQ_FW_MBOX_CMD_EXECUTE 0x00008000 +#define AQ_FW_MBOX_CMD_BUSY 0x00000100 +#define AQ_FW_MBOX_ADDR_REG 0x0208 +#define AQ_FW_MBOX_VAL_REG 0x020C +#define AQ_FW_GLB_CPU_SEM_REG(i) (0x03a0 + (i) * 4) +#define AQ_FW_SEM_RAM_REG AQ_FW_GLB_CPU_SEM_REG(2) +#define AQ_FW_GLB_CTL2_REG 0x0404 +#define AQ_GLB_GENERAL_PROVISIONING9_REG 0x0520 +#define AQ_GLB_NVR_PROVISIONING2_REG 0x0534 +#define AQ_INTR_STATUS_REG 0x2000 /* intr status */ +#define AQ_INTR_STATUS_CLR_REG 0x2050 /* intr status clear */ +#define AQ_INTR_MASK_REG 0x2060 /* intr mask set */ +#define AQ_INTR_MASK_CLR_REG 0x2070 /* intr mask clear */ +#define AQ_INTR_AUTOMASK_REG 0x2090 + +/* AQ_INTR_IRQ_MAP_TXRX_REG 0x2100-0x2140 */ +#define AQ_INTR_IRQ_MAP_TXRX_REG(i) (0x2100 + ((i) / 2) * 4) +#define AQ_INTR_IRQ_MAP_TX_REG(i) AQ_INTR_IRQ_MAP_TXRX_REG(i) +#define AQ_INTR_IRQ_MAP_TX_IRQMAP(i) (0x1FU << (((i) & 1) ? 16 : 24)) +#define AQ_INTR_IRQ_MAP_TX_EN(i) (1U << (((i) & 1) ? 23 : 31)) +#define AQ_INTR_IRQ_MAP_RX_REG(i) AQ_INTR_IRQ_MAP_TXRX_REG(i) +#define AQ_INTR_IRQ_MAP_RX_IRQMAP(i) (0x1FU << (((i) & 1) ? 0 : 8)) +#define AQ_INTR_IRQ_MAP_RX_EN(i) (1U << (((i) & 1) ? 7 : 15)) + +/* AQ_GEN_INTR_MAP_REG[AQ_RINGS_NUM] 0x2180-0x2200 */ +#define AQ_GEN_INTR_MAP_REG(i) (0x2180 + (i) * 4) +#define AQ_B0_ERR_INT 8U + +#define AQ_INTR_CTRL_REG 0x2300 +#define AQ_INTR_CTRL_IRQMODE ((1 << 0) | (1 << 1)) +#define AQ_INTR_CTRL_IRQMODE_LEGACY 0 +#define AQ_INTR_CTRL_IRQMODE_MSI 1 +#define AQ_INTR_CTRL_IRQMODE_MSIX 2 +#define AQ_INTR_CTRL_MULTIVEC (1 << 2) +#define AQ_INTR_CTRL_RESET_DIS (1 << 29) +#define AQ_INTR_CTRL_RESET_IRQ (1 << 31) +#define AQ_MBOXIF_POWER_GATING_CONTROL_REG 0x32a8 + +#define FW_MPI_MBOX_ADDR_REG 0x0360 +#define FW1X_MPI_INIT1_REG 0x0364 +#define FW1X_MPI_INIT2_REG 0x0370 +#define FW1X_MPI_EFUSEADDR_REG 0x0374 + +#define FW2X_MPI_EFUSEADDR_REG 0x0364 +#define FW2X_MPI_CONTROL_REG 0x0368 /* 64bit */ +#define FW2X_MPI_STATE_REG 0x0370 /* 64bit */ +#define FW_BOOT_EXIT_CODE_REG 0x0388 + +#define FW_BOOT_EXIT_CODE_REG 0x0388 +#define RBL_STATUS_DEAD 0x0000dead +#define RBL_STATUS_SUCCESS 0x0000abba +#define RBL_STATUS_FAILURE 0x00000bad +#define RBL_STATUS_HOST_BOOT 0x0000f1a7 +#define FW_MPI_DAISY_CHAIN_STATUS_REG 0x0704 +#define AQ_PCI_REG_CONTROL_6_REG 0x1014 + +#define FW_MPI_RESETCTRL_REG 0x4000 +#define FW_MPI_RESETCTRL_RESET_DIS (1 << 29) + +#define RX_SYSCONTROL_REG 0x5000 +#define RX_SYSCONTROL_RESET_DIS (1 << 29) + +#define RX_TCP_RSS_HASH_REG 0x5040 + +#define RPF_L2BC_REG 0x5100 +#define RPF_L2BC_EN (1 << 0) +#define RPF_L2BC_PROMISC (1 << 3) +#define RPF_L2BC_ACTION 0x7000 +#define RPF_L2BC_THRESHOLD 0xFFFF0000 + +#define AQ_HW_MAC_OWN 0 + +/* RPF_L2UC_*_REG[34] (actual [38]?) */ +#define RPF_L2UC_LSW_REG(i) (0x5110 + (i) * 8) +#define RPF_L2UC_MSW_REG(i) (0x5114 + (i) * 8) +#define RPF_L2UC_MSW_MACADDR_HI 0xFFFF +#define RPF_L2UC_MSW_ACTION 0x70000 +#define RPF_L2UC_MSW_EN (1 << 31) +#define AQ_HW_MAC_NUM 34 + +/* RPF_MCAST_FILTER_REG[8] 0x5250-0x5270 */ +#define RPF_MCAST_FILTER_REG(i) (0x5250 + (i) * 4) +#define RPF_MCAST_FILTER_EN (1 << 31) +#define RPF_MCAST_FILTER_MASK_REG 0x5270 +#define RPF_MCAST_FILTER_MASK_ALLMULTI (1 << 14) + +#define RPF_VLAN_MODE_REG 0x5280 +#define RPF_VLAN_MODE_PROMISC (1 << 1) +#define RPF_VLAN_MODE_ACCEPT_UNTAGGED (1 << 2) +#define RPF_VLAN_MODE_UNTAGGED_ACTION 0x38 + +#define RPF_VLAN_TPID_REG 0x5284 +#define RPF_VLAN_TPID_OUTER 0xFFFF0000 +#define RPF_VLAN_TPID_INNER 0xFFFF + +/* RPF_ETHERTYPE_FILTER_REG[AQ_RINGS_NUM] 0x5300-0x5380 */ +#define RPF_ETHERTYPE_FILTER_REG(i) (0x5300 + (i) * 4) +#define RPF_ETHERTYPE_FILTER_EN (1 << 31) + +/* RPF_L3_FILTER_REG[8] 0x5380-0x53a0 */ +#define RPF_L3_FILTER_REG(i) (0x5380 + (i) * 4) +#define RPF_L3_FILTER_L4_EN (1 << 31) + +#define RX_FLR_RSS_CONTROL1_REG 0x54c0 +#define RX_FLR_RSS_CONTROL1_EN (1 << 31) + +#define RPF_RPB_RX_TC_UPT_REG 0x54c4 +#define RPF_RPB_RX_TC_UPT_MASK(i) (0x00000007 << ((i) * 4)) + +#define RPB_RPF_RX_REG 0x5700 +#define RPB_RPF_RX_TC_MODE (1 << 8) +#define RPB_RPF_RX_FC_MODE 0x30 +#define RPB_RPF_RX_BUF_EN (1 << 0) + +/* RPB_RXB_BUFSIZE_REG[AQ_TRAFFICCLASS_NUM] 0x5710-0x5790 */ +#define RPB_RXB_BUFSIZE_REG(i) (0x5710 + (i) * 0x10) +#define RPB_RXB_BUFSIZE 0x1FF +#define RPB_RXB_XOFF_REG(i) (0x5714 + (i) * 0x10) +#define RPB_RXB_XOFF_EN (1 << 31) +#define RPB_RXB_XOFF_THRESH_HI 0x3FFF0000 +#define RPB_RXB_XOFF_THRESH_LO 0x3FFF + +#define RX_DMA_INT_DESC_WRWB_EN_REG 0x5a30 +#define RX_DMA_INT_DESC_WRWB_EN (1 << 2) +#define RX_DMA_INT_DESC_MODERATE_EN (1 << 3) + +#define RX_INTR_MODERATION_CTL_REG(i) (0x5a40 + (i) * 4) +#define RX_INTR_MODERATION_CTL_EN (1 << 1) + +#define RX_DMA_DESC_BASE_ADDRLSW_REG(i) (0x5b00 + (i) * 0x20) +#define RX_DMA_DESC_BASE_ADDRMSW_REG(i) (0x5b04 + (i) * 0x20) +#define RX_DMA_DESC_REG(i) (0x5b08 + (i) * 0x20) +#define RX_DMA_DESC_LEN (0x3FF << 3) +#define RX_DMA_DESC_RESET (1 << 25) +#define RX_DMA_DESC_HEADER_SPLIT (1 << 28) +#define RX_DMA_DESC_VLAN_STRIP (1 << 29) +#define RX_DMA_DESC_EN (1 << 31) +#define RX_DMA_DESC_HEAD_PTR_REG(i) (0x5b0c + (i) * 0x20) +#define RX_DMA_DESC_HEAD_PTR 0xFFF +#define RX_DMA_DESC_TAIL_PTR_REG(i) (0x5b10 + (i) * 0x20) +#define RX_DMA_DESC_BUFSIZE_REG(i) (0x5b18 + (i) * 0x20) +#define RX_DMA_DESC_BUFSIZE_DATA 0x000F +#define RX_DMA_DESC_BUFSIZE_HDR 0x0FF0 + +#define RX_DMA_DCAD_REG(i) (0x6100 + (i) * 4) +#define RX_DMA_DCAD_CPUID 0xFF +#define RX_DMA_DCAD_PAYLOAD_EN (1 << 29) +#define RX_DMA_DCAD_HEADER_EN (1 << 30) +#define RX_DMA_DCAD_DESC_EN (1 << 31) + +#define RX_DMA_DCA_REG 0x6180 +#define RX_DMA_DCA_EN (1 << 31) +#define RX_DMA_DCA_MODE 0xF + +#define TX_SYSCONTROL_REG 0x7000 +#define TX_SYSCONTROL_RESET_DIS (1 << 29) + +#define TX_TPO2_REG 0x7040 +#define TX_TPO2_EN (1 << 16) + +#define TPS_DESC_VM_ARB_MODE_REG 0x7300 +#define TPS_DESC_VM_ARB_MODE (1 << 0) +#define TPS_DESC_RATE_REG 0x7310 +#define TPS_DESC_RATE_TA_RST (1 << 31) +#define TPS_DESC_RATE_LIM 0x7FF +#define TPS_DESC_TC_ARB_MODE_REG 0x7200 +#define TPS_DESC_TC_ARB_MODE 0x3 +#define TPS_DATA_TC_ARB_MODE_REG 0x7100 +#define TPS_DATA_TC_ARB_MODE (1 << 0) + +/* TPS_DATA_TCT_REG[AQ_TRAFFICCLASS_NUM] 0x7110-0x7130 */ +#define TPS_DATA_TCT_REG(i) (0x7110 + (i) * 4) +#define TPS_DATA_TCT_CREDIT_MAX 0xFFF0000 +#define TPS_DATA_TCT_WEIGHT 0x1FF +/* TPS_DATA_TCT_REG[AQ_TRAFFICCLASS_NUM] 0x7210-0x7230 */ +#define TPS_DESC_TCT_REG(i) (0x7210 + (i) * 4) +#define TPS_DESC_TCT_CREDIT_MAX 0xFFF0000 +#define TPS_DESC_TCT_WEIGHT 0x1FF + +#define AQ_HW_TXBUF_MAX 160 +#define AQ_HW_RXBUF_MAX 320 + +#define THM_LSO_TCP_FLAG1_REG 0x7820 +#define THM_LSO_TCP_FLAG1_FIRST 0xFFF +#define THM_LSO_TCP_FLAG1_MID 0xFFF0000 +#define THM_LSO_TCP_FLAG2_REG 0x7824 +#define THM_LSO_TCP_FLAG2_LAST 0xFFF + +#define TPB_TX_BUF_REG 0x7900 +#define TPB_TX_BUF_EN (1 << 0) +#define TPB_TX_BUF_SCP_INS_EN (1 << 2) +#define TPB_TX_BUF_TC_MODE_EN (1 << 8) + +/* TPB_TXB_BUFSIZE_REG[AQ_TRAFFICCLASS_NUM] 0x7910-7990 */ +#define TPB_TXB_BUFSIZE_REG(i) (0x7910 + (i) * 0x10) +#define TPB_TXB_BUFSIZE (0xFF) +#define TPB_TXB_THRESH_REG(i) (0x7914 + (i) * 0x10) +#define TPB_TXB_THRESH_HI 0x1FFF0000 +#define TPB_TXB_THRESH_LO 0x1FFF + +#define AQ_HW_TX_DMA_TOTAL_REQ_LIMIT_REG 0x7b20 + +#define TX_DMA_INT_DESC_WRWB_EN_REG 0x7b40 +#define TX_DMA_INT_DESC_WRWB_EN (1 << 1) +#define TX_DMA_INT_DESC_MODERATE_EN (1 << 4) + +#define TX_DMA_DESC_BASE_ADDRLSW_REG(i) (0x7c00 + (i) * 0x40) +#define TX_DMA_DESC_BASE_ADDRMSW_REG(i) (0x7c04 + (i) * 0x40) +#define TX_DMA_DESC_REG(i) (0x7c08 + (i) * 0x40) +#define TX_DMA_DESC_LEN 0x00000FF8 +#define TX_DMA_DESC_EN 0x80000000 +#define TX_DMA_DESC_HEAD_PTR_REG(i) (0x7c0c + (i) * 0x40) +#define TX_DMA_DESC_HEAD_PTR 0x00000FFF +#define TX_DMA_DESC_TAIL_PTR_REG(i) (0x7c10 + (i) * 0x40) +#define TX_DMA_DESC_WRWB_THRESH_REG(i) (0x7c18 + (i) * 0x40) +#define TX_DMA_DESC_WRWB_THRESH 0x00003F00 + +#define TDM_DCAD_REG(i) (0x8400 + (i) * 4) +#define TDM_DCAD_CPUID 0x7F +#define TDM_DCAD_CPUID_EN 0x80000000 + +#define TDM_DCA_REG 0x8480 +#define TDM_DCA_EN (1 << 31) +#define TDM_DCA_MODE 0xF + +#define TX_INTR_MODERATION_CTL_REG(i) (0x8980 + (i) * 4) +#define TX_INTR_MODERATION_CTL_EN (1 << 1) + +#define __LOWEST_SET_BIT(__mask) (((((uint32_t)__mask) - 1) & ((uint32_t)__mask)) ^ ((uint32_t)__mask)) +#define __SHIFTIN(__x, __mask) ((__x) * __LOWEST_SET_BIT(__mask)) + +#if 0 +#define AQ_READ_REG(sc, reg) \ + bus_space_read_4((sc)->sc_iot, (sc)->sc_ioh, (reg)) + +#endif +#define AQ_WRITE_REG(sc, reg, val) \ + bus_space_write_4((sc)->sc_iot, (sc)->sc_ioh, (reg), (val)) + +#define AQ_WRITE_REG_BIT(sc, reg, mask, val) \ + do { \ + uint32_t _v; \ + _v = AQ_READ_REG((sc), (reg)); \ + _v &= ~(mask); \ + if ((val) != 0) \ + _v |= __SHIFTIN((val), (mask)); \ + AQ_WRITE_REG((sc), (reg), _v); \ + } while (/* CONSTCOND */ 0) + +#define AQ_READ64_REG(sc, reg) \ + ((uint64_t)AQ_READ_REG(sc, reg) | \ + (((uint64_t)AQ_READ_REG(sc, (reg) + 4)) << 32)) + +#define AQ_WRITE64_REG(sc, reg, val) \ + do { \ + AQ_WRITE_REG(sc, reg, (uint32_t)val); \ + AQ_WRITE_REG(sc, reg + 4, (uint32_t)(val >> 32)); \ + } while (/* CONSTCOND */0) + +#define WAIT_FOR(expr, us, n, errp) \ + do { \ + unsigned int _n; \ + for (_n = n; (!(expr)) && _n != 0; --_n) { \ + delay((us)); \ + } \ + if ((errp != NULL)) { \ + if (_n == 0) \ + *(errp) = ETIMEDOUT; \ + else \ + *(errp) = 0; \ + } \ + } while (/* CONSTCOND */ 0) + +#define FW_VERSION_MAJOR(sc) (((sc)->sc_fw_version >> 24) & 0xff) +#define FW_VERSION_MINOR(sc) (((sc)->sc_fw_version >> 16) & 0xff) +#define FW_VERSION_BUILD(sc) ((sc)->sc_fw_version & 0xffff) + +#define FEATURES_MIPS 0x00000001 +#define FEATURES_TPO2 0x00000002 +#define FEATURES_RPF2 0x00000004 +#define FEATURES_MPI_AQ 0x00000008 +#define FEATURES_REV_A0 0x10000000 +#define FEATURES_REV_A (FEATURES_REV_A0) +#define FEATURES_REV_B0 0x20000000 +#define FEATURES_REV_B1 0x40000000 +#define FEATURES_REV_B (FEATURES_REV_B0|FEATURES_REV_B1) + +/* lock for FW2X_MPI_{CONTROL,STATE]_REG read-modify-write */ +#define AQ_MPI_LOCK(sc) mtx_enter(&(sc)->sc_mpi_mutex); +#define AQ_MPI_UNLOCK(sc) mtx_leave(&(sc)->sc_mpi_mutex); + +#define FW2X_CTRL_10BASET_HD (1 << 0) +#define FW2X_CTRL_10BASET_FD (1 << 1) +#define FW2X_CTRL_100BASETX_HD (1 << 2) +#define FW2X_CTRL_100BASET4_HD (1 << 3) +#define FW2X_CTRL_100BASET2_HD (1 << 4) +#define FW2X_CTRL_100BASETX_FD (1 << 5) +#define FW2X_CTRL_100BASET2_FD (1 << 6) +#define FW2X_CTRL_1000BASET_HD (1 << 7) +#define FW2X_CTRL_1000BASET_FD (1 << 8) +#define FW2X_CTRL_2P5GBASET_FD (1 << 9) +#define FW2X_CTRL_5GBASET_FD (1 << 10) +#define FW2X_CTRL_10GBASET_FD (1 << 11) +#define FW2X_CTRL_RESERVED1 (1ULL << 32) +#define FW2X_CTRL_10BASET_EEE (1ULL << 33) +#define FW2X_CTRL_RESERVED2 (1ULL << 34) +#define FW2X_CTRL_PAUSE (1ULL << 35) +#define FW2X_CTRL_ASYMMETRIC_PAUSE (1ULL << 36) +#define FW2X_CTRL_100BASETX_EEE (1ULL << 37) +#define FW2X_CTRL_RESERVED3 (1ULL << 38) +#define FW2X_CTRL_RESERVED4 (1ULL << 39) +#define FW2X_CTRL_1000BASET_FD_EEE (1ULL << 40) +#define FW2X_CTRL_2P5GBASET_FD_EEE (1ULL << 41) +#define FW2X_CTRL_5GBASET_FD_EEE (1ULL << 42) +#define FW2X_CTRL_10GBASET_FD_EEE (1ULL << 43) +#define FW2X_CTRL_RESERVED5 (1ULL << 44) +#define FW2X_CTRL_RESERVED6 (1ULL << 45) +#define FW2X_CTRL_RESERVED7 (1ULL << 46) +#define FW2X_CTRL_RESERVED8 (1ULL << 47) +#define FW2X_CTRL_RESERVED9 (1ULL << 48) +#define FW2X_CTRL_CABLE_DIAG (1ULL << 49) +#define FW2X_CTRL_TEMPERATURE (1ULL << 50) +#define FW2X_CTRL_DOWNSHIFT (1ULL << 51) +#define FW2X_CTRL_PTP_AVB_EN (1ULL << 52) +#define FW2X_CTRL_MEDIA_DETECT (1ULL << 53) +#define FW2X_CTRL_LINK_DROP (1ULL << 54) +#define FW2X_CTRL_SLEEP_PROXY (1ULL << 55) +#define FW2X_CTRL_WOL (1ULL << 56) +#define FW2X_CTRL_MAC_STOP (1ULL << 57) +#define FW2X_CTRL_EXT_LOOPBACK (1ULL << 58) +#define FW2X_CTRL_INT_LOOPBACK (1ULL << 59) +#define FW2X_CTRL_EFUSE_AGENT (1ULL << 60) +#define FW2X_CTRL_WOL_TIMER (1ULL << 61) +#define FW2X_CTRL_STATISTICS (1ULL << 62) +#define FW2X_CTRL_TRANSACTION_ID (1ULL << 63) + +#define FW2X_CTRL_RATE_100M FW2X_CTRL_100BASETX_FD +#define FW2X_CTRL_RATE_1G FW2X_CTRL_1000BASET_FD +#define FW2X_CTRL_RATE_2G5 FW2X_CTRL_2P5GBASET_FD +#define FW2X_CTRL_RATE_5G FW2X_CTRL_5GBASET_FD +#define FW2X_CTRL_RATE_10G FW2X_CTRL_10GBASET_FD +#define FW2X_CTRL_RATE_MASK \ + (FW2X_CTRL_RATE_100M | \ + FW2X_CTRL_RATE_1G | \ + FW2X_CTRL_RATE_2G5 | \ + FW2X_CTRL_RATE_5G | \ + FW2X_CTRL_RATE_10G) +#define FW2X_CTRL_EEE_MASK \ + (FW2X_CTRL_10BASET_EEE | \ + FW2X_CTRL_100BASETX_EEE | \ + FW2X_CTRL_1000BASET_FD_EEE | \ + FW2X_CTRL_2P5GBASET_FD_EEE | \ + FW2X_CTRL_5GBASET_FD_EEE | \ + FW2X_CTRL_10GBASET_FD_EEE) + +enum aq_fw_bootloader_mode { + FW_BOOT_MODE_UNKNOWN = 0, + FW_BOOT_MODE_FLB, + FW_BOOT_MODE_RBL_FLASH, + FW_BOOT_MODE_RBL_HOST_BOOTLOAD +}; + +enum aq_media_type { + AQ_MEDIA_TYPE_UNKNOWN = 0, + AQ_MEDIA_TYPE_FIBRE, + AQ_MEDIA_TYPE_TP +}; + +enum aq_link_speed { + AQ_LINK_NONE = 0, + AQ_LINK_100M = (1 << 0), + AQ_LINK_1G = (1 << 1), + AQ_LINK_2G5 = (1 << 2), + AQ_LINK_5G = (1 << 3), + AQ_LINK_10G = (1 << 4) +}; + +#define AQ_LINK_ALL (AQ_LINK_100M | AQ_LINK_1G | AQ_LINK_2G5 | \ + AQ_LINK_5G | AQ_LINK_10G ) +#define AQ_LINK_AUTO AQ_LINK_ALL + +enum aq_link_eee { + AQ_EEE_DISABLE = 0, + AQ_EEE_ENABLE = 1 +}; + +enum aq_hw_fw_mpi_state { + MPI_DEINIT = 0, + MPI_RESET = 1, + MPI_INIT = 2, + MPI_POWER = 4 +}; + +enum aq_link_fc { + AQ_FC_NONE = 0, + AQ_FC_RX = (1 << 0), + AQ_FC_TX = (1 << 1), + AQ_FC_ALL = (AQ_FC_RX | AQ_FC_TX) +}; + +struct aq_dmamem { + bus_dmamap_t aqm_map; + bus_dma_segment_t aqm_seg; + int aqm_nsegs; + size_t aqm_size; + caddr_t aqm_kva; +}; + +#define AQ_DMA_MAP(_aqm) ((_aqm)->aqm_map) +#define AQ_DMA_DVA(_aqm) ((_aqm)->aqm_map->dm_segs[0].ds_addr) +#define AQ_DMA_KVA(_aqm) ((void *)(_aqm)->aqm_kva) +#define AQ_DMA_LEN(_aqm) ((_aqm)->aqm_size) + + +struct aq_mailbox_header { + uint32_t version; + uint32_t transaction_id; + int32_t error; +} __packed __aligned(4); + +struct aq_hw_stats_s { + uint32_t uprc; + uint32_t mprc; + uint32_t bprc; + uint32_t erpt; + uint32_t uptc; + uint32_t mptc; + uint32_t bptc; + uint32_t erpr; + uint32_t mbtc; + uint32_t bbtc; + uint32_t mbrc; + uint32_t bbrc; + uint32_t ubrc; + uint32_t ubtc; + uint32_t ptc; + uint32_t prc; + uint32_t dpc; /* not exists in fw2x_msm_statistics */ + uint32_t cprc; /* not exists in fw2x_msm_statistics */ +} __packed __aligned(4); + +struct aq_fw2x_capabilities { + uint32_t caps_lo; + uint32_t caps_hi; +} __packed __aligned(4); + +struct aq_fw2x_msm_statistics { + uint32_t uprc; + uint32_t mprc; + uint32_t bprc; + uint32_t erpt; + uint32_t uptc; + uint32_t mptc; + uint32_t bptc; + uint32_t erpr; + uint32_t mbtc; + uint32_t bbtc; + uint32_t mbrc; + uint32_t bbrc; + uint32_t ubrc; + uint32_t ubtc; + uint32_t ptc; + uint32_t prc; +} __packed __aligned(4); + +struct aq_fw2x_phy_cable_diag_data { + uint32_t lane_data[4]; +} __packed __aligned(4); + +struct aq_fw2x_mailbox { /* struct fwHostInterface */ + struct aq_mailbox_header header; + struct aq_fw2x_msm_statistics msm; /* msmStatistics_t msm; */ + + uint32_t phy_info1; +#define PHYINFO1_FAULT_CODE __BITS(31,16) +#define PHYINFO1_PHY_H_BIT __BITS(0,15) + uint32_t phy_info2; +#define PHYINFO2_TEMPERATURE __BITS(15,0) +#define PHYINFO2_CABLE_LEN __BITS(23,16) + + struct aq_fw2x_phy_cable_diag_data diag_data; + uint32_t reserved[8]; + + struct aq_fw2x_capabilities caps; + + /* ... */ +} __packed __aligned(4); + +struct aq_rx_desc_read { + uint64_t buf_addr; + uint64_t hdr_addr; +} __packed; + +struct aq_rx_desc_wb { + uint32_t type; +#define AQ_RXDESC_TYPE_RSSTYPE 0x000f +#define AQ_RXDESC_TYPE_ETHER 0x0030 +#define AQ_RXDESC_TYPE_PROTO 0x01c0 +#define AQ_RXDESC_TYPE_VLAN (1 << 9) +#define AQ_RXDESC_TYPE_VLAN2 (1 << 10) +#define AQ_RXDESC_TYPE_DMA_ERR (1 << 12) +#define AQ_RXDESC_TYPE_V4_SUM (1 << 19) +#define AQ_RXDESC_TYPE_TCP_SUM (1 << 20) + uint32_t rss_hash; + uint16_t status; +#define AQ_RXDESC_STATUS_DD (1 << 0) +#define AQ_RXDESC_STATUS_EOP (1 << 1) +#define AQ_RXDESC_STATUS_MACERR (1 << 2) +#define AQ_RXDESC_STATUS_V4_SUM (1 << 3) +#define AQ_RXDESC_STATUS_L4_SUM_ERR (1 << 4) +#define AQ_RXDESC_STATUS_L4_SUM_OK (1 << 5) + uint16_t pkt_len; + uint16_t next_desc_ptr; + uint16_t vlan; +} __packed; + +struct aq_tx_desc { + uint64_t buf_addr; + uint32_t ctl1; +#define AQ_TXDESC_CTL1_TYPE_TXD 0x00000001 +#define AQ_TXDESC_CTL1_TYPE_TXC 0x00000002 +#define AQ_TXDESC_CTL1_BLEN_SHIFT 4 +#define AQ_TXDESC_CTL1_DD (1 << 20) +#define AQ_TXDESC_CTL1_CMD_EOP (1 << 21) +#define AQ_TXDESC_CTL1_CMD_VLAN (1 << 22) +#define AQ_TXDESC_CTL1_CMD_FCS (1 << 23) +#define AQ_TXDESC_CTL1_CMD_IP4CSUM (1 << 24) +#define AQ_TXDESC_CTL1_CMD_L4CSUM (1 << 25) +#define AQ_TXDESC_CTL1_CMD_WB (1 << 27) + +#define AQ_TXDESC_CTL1_VID_SHIFT 4 + uint32_t ctl2; +#define AQ_TXDESC_CTL2_LEN_SHIFT 14 +#define AQ_TXDESC_CTL2_CTX_EN (1 << 13) +} __packed; + +struct aq_slot { + bus_dmamap_t as_map; + struct mbuf *as_m; +}; + +struct aq_rxring { + struct ifiqueue *rx_ifiq; + struct aq_dmamem rx_mem; + struct aq_slot *rx_slots; + int rx_q; + int rx_irq; + + struct timeout rx_refill; + struct if_rxring rx_rxr; + uint32_t rx_prod; + uint32_t rx_cons; +}; + +struct aq_txring { + struct ifqueue *tx_ifq; + struct aq_dmamem tx_mem; + struct aq_slot *tx_slots; + int tx_q; + int tx_irq; + uint32_t tx_prod; + uint32_t tx_cons; +}; + +struct aq_queues { + char q_name[16]; + void *q_ihc; + struct aq_softc *q_sc; + int q_index; + struct aq_rxring q_rx; + struct aq_txring q_tx; +}; + + +struct aq_softc; +struct aq_firmware_ops { + int (*reset)(struct aq_softc *); + int (*set_mode)(struct aq_softc *, enum aq_hw_fw_mpi_state, + enum aq_link_speed, enum aq_link_fc, enum aq_link_eee); + int (*get_mode)(struct aq_softc *, enum aq_hw_fw_mpi_state *, + enum aq_link_speed *, enum aq_link_fc *, enum aq_link_eee *); + int (*get_stats)(struct aq_softc *, struct aq_hw_stats_s *); +}; + +struct aq_softc { + struct device sc_dev; + uint16_t sc_product; + uint16_t sc_revision; + bus_dma_tag_t sc_dmat; + pci_chipset_tag_t sc_pc; + pcitag_t sc_pcitag; + int sc_nqueues; + struct aq_queues sc_queues[AQ_MAXQ]; + struct intrmap *sc_intrmap; + void *sc_ih; + bus_space_handle_t sc_ioh; + bus_space_tag_t sc_iot; + + uint32_t sc_mbox_addr; + int sc_rbl_enabled; + int sc_fast_start_enabled; + int sc_flash_present; + uint32_t sc_fw_version; + const struct aq_firmware_ops *sc_fw_ops; + uint64_t sc_fw_caps; + enum aq_media_type sc_media_type; + enum aq_link_speed sc_available_rates; + uint32_t sc_features; + int sc_linkstat_irq; + struct arpcom sc_arpcom; + struct ifmedia sc_media; + + struct ether_addr sc_enaddr; + struct mutex sc_mpi_mutex; +}; + +const struct pci_matchid aq_devices[] = { + { PCI_VENDOR_AQUANTIA, PCI_PRODUCT_AQUANTIA_AQC100 }, + { PCI_VENDOR_AQUANTIA, PCI_PRODUCT_AQUANTIA_AQC107 }, + { PCI_VENDOR_AQUANTIA, PCI_PRODUCT_AQUANTIA_AQC108 }, + { PCI_VENDOR_AQUANTIA, PCI_PRODUCT_AQUANTIA_AQC109 }, + { PCI_VENDOR_AQUANTIA, PCI_PRODUCT_AQUANTIA_AQC111 }, + { PCI_VENDOR_AQUANTIA, PCI_PRODUCT_AQUANTIA_AQC112 }, + { PCI_VENDOR_AQUANTIA, PCI_PRODUCT_AQUANTIA_AQC100S }, + { PCI_VENDOR_AQUANTIA, PCI_PRODUCT_AQUANTIA_AQC107S }, + { PCI_VENDOR_AQUANTIA, PCI_PRODUCT_AQUANTIA_AQC108S }, + { PCI_VENDOR_AQUANTIA, PCI_PRODUCT_AQUANTIA_AQC109S }, + { PCI_VENDOR_AQUANTIA, PCI_PRODUCT_AQUANTIA_AQC111S }, + { PCI_VENDOR_AQUANTIA, PCI_PRODUCT_AQUANTIA_AQC112S }, + { PCI_VENDOR_AQUANTIA, PCI_PRODUCT_AQUANTIA_D100 }, + { PCI_VENDOR_AQUANTIA, PCI_PRODUCT_AQUANTIA_D107 }, + { PCI_VENDOR_AQUANTIA, PCI_PRODUCT_AQUANTIA_D108 }, + { PCI_VENDOR_AQUANTIA, PCI_PRODUCT_AQUANTIA_D109 }, +}; + +const struct aq_product { + pci_vendor_id_t aq_vendor; + pci_product_id_t aq_product; + enum aq_media_type aq_media_type; + enum aq_link_speed aq_available_rates; +} aq_products[] = { +{ PCI_VENDOR_AQUANTIA, PCI_PRODUCT_AQUANTIA_AQC100, + AQ_MEDIA_TYPE_FIBRE, AQ_LINK_ALL +}, +{ PCI_VENDOR_AQUANTIA, PCI_PRODUCT_AQUANTIA_AQC107, + AQ_MEDIA_TYPE_TP, AQ_LINK_ALL +}, +{ PCI_VENDOR_AQUANTIA, PCI_PRODUCT_AQUANTIA_AQC108, + AQ_MEDIA_TYPE_TP, AQ_LINK_100M | AQ_LINK_1G | AQ_LINK_2G5 | AQ_LINK_5G +}, +{ PCI_VENDOR_AQUANTIA, PCI_PRODUCT_AQUANTIA_AQC109, + AQ_MEDIA_TYPE_TP, AQ_LINK_100M | AQ_LINK_1G | AQ_LINK_2G5 +}, +{ PCI_VENDOR_AQUANTIA, PCI_PRODUCT_AQUANTIA_AQC111, + AQ_MEDIA_TYPE_TP, AQ_LINK_100M | AQ_LINK_1G | AQ_LINK_2G5 | AQ_LINK_5G +}, +{ PCI_VENDOR_AQUANTIA, PCI_PRODUCT_AQUANTIA_AQC112, + AQ_MEDIA_TYPE_TP, AQ_LINK_100M | AQ_LINK_1G | AQ_LINK_2G5 +}, +{ PCI_VENDOR_AQUANTIA, PCI_PRODUCT_AQUANTIA_AQC100S, + AQ_MEDIA_TYPE_FIBRE, AQ_LINK_ALL +}, +{ PCI_VENDOR_AQUANTIA, PCI_PRODUCT_AQUANTIA_AQC107S, + AQ_MEDIA_TYPE_TP, AQ_LINK_ALL +}, +{ PCI_VENDOR_AQUANTIA, PCI_PRODUCT_AQUANTIA_AQC108S, + AQ_MEDIA_TYPE_TP, AQ_LINK_100M | AQ_LINK_1G | AQ_LINK_2G5 | AQ_LINK_5G +}, +{ PCI_VENDOR_AQUANTIA, PCI_PRODUCT_AQUANTIA_AQC109S, + AQ_MEDIA_TYPE_TP, AQ_LINK_100M | AQ_LINK_1G | AQ_LINK_2G5 +}, +{ PCI_VENDOR_AQUANTIA, PCI_PRODUCT_AQUANTIA_AQC111S, + AQ_MEDIA_TYPE_TP, AQ_LINK_100M | AQ_LINK_1G | AQ_LINK_2G5 | AQ_LINK_5G +}, +{ PCI_VENDOR_AQUANTIA, PCI_PRODUCT_AQUANTIA_AQC112S, + AQ_MEDIA_TYPE_TP, AQ_LINK_100M | AQ_LINK_1G | AQ_LINK_2G5 +}, +{ PCI_VENDOR_AQUANTIA, PCI_PRODUCT_AQUANTIA_D100, + AQ_MEDIA_TYPE_FIBRE, AQ_LINK_ALL +}, +{ PCI_VENDOR_AQUANTIA, PCI_PRODUCT_AQUANTIA_D107, + AQ_MEDIA_TYPE_TP, AQ_LINK_ALL +}, +{ PCI_VENDOR_AQUANTIA, PCI_PRODUCT_AQUANTIA_D108, + AQ_MEDIA_TYPE_TP, AQ_LINK_100M | AQ_LINK_1G | AQ_LINK_2G5 | AQ_LINK_5G +}, +{ PCI_VENDOR_AQUANTIA, PCI_PRODUCT_AQUANTIA_D109, + AQ_MEDIA_TYPE_TP, AQ_LINK_100M | AQ_LINK_1G | AQ_LINK_2G5 +} +}; + +int aq_match(struct device *, void *, void *); +void aq_attach(struct device *, struct device *, void *); +int aq_detach(struct device *, int); +int aq_activate(struct device *, int); +int aq_intr(void *); +void aq_global_software_reset(struct aq_softc *); +int aq_fw_reset(struct aq_softc *); +int aq_mac_soft_reset(struct aq_softc *, enum aq_fw_bootloader_mode *); +int aq_mac_soft_reset_rbl(struct aq_softc *, enum aq_fw_bootloader_mode *); +int aq_mac_soft_reset_flb(struct aq_softc *); +int aq_fw_read_version(struct aq_softc *); +int aq_fw_version_init(struct aq_softc *); +int aq_hw_init_ucp(struct aq_softc *); +int aq_fw_downld_dwords(struct aq_softc *, uint32_t, uint32_t *, uint32_t); +int aq_get_mac_addr(struct aq_softc *); +int aq_hw_reset(struct aq_softc *); +int aq_hw_init(struct aq_softc *, int); +void aq_hw_qos_set(struct aq_softc *); +void aq_l3_filter_set(struct aq_softc *); +void aq_hw_init_tx_path(struct aq_softc *); +void aq_hw_init_rx_path(struct aq_softc *); +int aq_set_mac_addr(struct aq_softc *, int, uint8_t *); +int aq_set_linkmode(struct aq_softc *, enum aq_link_speed, + enum aq_link_fc, enum aq_link_eee); +void aq_watchdog(struct ifnet *); +void aq_enable_intr(struct aq_softc *, int, int); +int aq_ioctl(struct ifnet *, u_long, caddr_t); +int aq_up(struct aq_softc *); +void aq_down(struct aq_softc *); +void aq_iff(struct aq_softc *); +void aq_start(struct ifqueue *); +void aq_ifmedia_status(struct ifnet *, struct ifmediareq *); +int aq_ifmedia_change(struct ifnet *); +void aq_update_link_status(struct aq_softc *); + +void aq_refill(void *); +int aq_rx_fill(struct aq_softc *, struct aq_rxring *); +static inline unsigned int aq_rx_fill_slots(struct aq_softc *, + struct aq_rxring *, uint); + +int aq_dmamem_alloc(struct aq_softc *, struct aq_dmamem *, + bus_size_t, u_int); +void aq_dmamem_zero(struct aq_dmamem *); +void aq_dmamem_free(struct aq_softc *, struct aq_dmamem *); + +int aq_fw1x_reset(struct aq_softc *); +int aq_fw1x_get_mode(struct aq_softc *, enum aq_hw_fw_mpi_state *, + enum aq_link_speed *, enum aq_link_fc *, enum aq_link_eee *); +int aq_fw1x_set_mode(struct aq_softc *, enum aq_hw_fw_mpi_state, + enum aq_link_speed, enum aq_link_fc, enum aq_link_eee); +int aq_fw1x_get_stats(struct aq_softc *, struct aq_hw_stats_s *); + +int aq_fw2x_reset(struct aq_softc *); +int aq_fw2x_get_mode(struct aq_softc *, enum aq_hw_fw_mpi_state *, + enum aq_link_speed *, enum aq_link_fc *, enum aq_link_eee *); +int aq_fw2x_set_mode(struct aq_softc *, enum aq_hw_fw_mpi_state, + enum aq_link_speed, enum aq_link_fc, enum aq_link_eee); +int aq_fw2x_get_stats(struct aq_softc *, struct aq_hw_stats_s *); + +const struct aq_firmware_ops aq_fw1x_ops = { + .reset = aq_fw1x_reset, + .set_mode = aq_fw1x_set_mode, + .get_mode = aq_fw1x_get_mode, + .get_stats = aq_fw1x_get_stats, +}; + +const struct aq_firmware_ops aq_fw2x_ops = { + .reset = aq_fw2x_reset, + .set_mode = aq_fw2x_set_mode, + .get_mode = aq_fw2x_get_mode, + .get_stats = aq_fw2x_get_stats, +}; + +struct cfattach aq_ca = { + sizeof(struct aq_softc), aq_match, aq_attach, NULL, + aq_activate +}; + +struct cfdriver aq_cd = { + NULL, "aq", DV_IFNET +}; + +uint32_t +AQ_READ_REG(struct aq_softc *sc, uint32_t reg) +{ + uint32_t res; + + res = bus_space_read_4(sc->sc_iot, sc->sc_ioh, reg); + + return res; +} + + +int +aq_match(struct device *dev, void *match, void *aux) +{ + return pci_matchbyid((struct pci_attach_args *)aux, aq_devices, + sizeof(aq_devices) / sizeof(aq_devices[0])); +} + +const struct aq_product * +aq_lookup(const struct pci_attach_args *pa) +{ + unsigned int i; + + for (i = 0; i < sizeof(aq_products) / sizeof(aq_products[0]); i++) { + if (PCI_VENDOR(pa->pa_id) == aq_products[i].aq_vendor && + PCI_PRODUCT(pa->pa_id) == aq_products[i].aq_product) { + return &aq_products[i]; + } + } + + return NULL; +} + +void +aq_attach(struct device *parent, struct device *self, void *aux) +{ + struct aq_softc *sc = (struct aq_softc *)self; + struct pci_attach_args *pa = aux; + const struct aq_product *aqp; + pcireg_t command, bar, memtype; + pci_chipset_tag_t pc; + pci_intr_handle_t ih; + int (*isr)(void *); + const char *intrstr; + pcitag_t tag; + struct ifnet *ifp = &sc->sc_arpcom.ac_if; + int irqmode; + int i; + + mtx_init(&sc->sc_mpi_mutex, IPL_NET); + + sc->sc_dmat = pa->pa_dmat; + sc->sc_pc = pc = pa->pa_pc; + sc->sc_pcitag = tag = pa->pa_tag; + + command = pci_conf_read(pa->pa_pc, pa->pa_tag, PCI_COMMAND_STATUS_REG); + command |= PCI_COMMAND_MASTER_ENABLE; + pci_conf_write(pa->pa_pc, pa->pa_tag, PCI_COMMAND_STATUS_REG, command); + + sc->sc_product = PCI_PRODUCT(pa->pa_id); + sc->sc_revision = PCI_REVISION(pa->pa_class); + + aqp = aq_lookup(pa); + + bar = pci_conf_read(pc, tag, AQ_BAR0); + if ((PCI_MAPREG_MEM_ADDR(bar) == 0) || + (PCI_MAPREG_TYPE(bar) != PCI_MAPREG_TYPE_MEM)) { + printf("%s: wrong BAR type\n", DEVNAME(sc)); + return; + } + + memtype = pci_mapreg_type(pc, tag, AQ_BAR0); + if (pci_mapreg_map(pa, AQ_BAR0, memtype, 0, &sc->sc_iot, &sc->sc_ioh, + NULL, NULL, 0)) { + printf(": failed to map BAR0\n"); + return; + } + + sc->sc_nqueues = 1; + + if (pci_intr_map_msix(pa, 0, &ih) == 0) { + irqmode = AQ_INTR_CTRL_IRQMODE_MSIX; + } else if (pci_intr_map_msi(pa, &ih) == 0) { + irqmode = AQ_INTR_CTRL_IRQMODE_MSI; + } else if (pci_intr_map(pa, &ih) == 0) { + irqmode = AQ_INTR_CTRL_IRQMODE_LEGACY; + } else { + printf(": failed to map interrupt\n"); + return; + } + + isr = aq_intr; + sc->sc_ih = pci_intr_establish(pa->pa_pc, ih, + IPL_NET | IPL_MPSAFE, isr, sc, self->dv_xname); + intrstr = pci_intr_string(pa->pa_pc, ih); + if (intrstr) + printf(": %s", intrstr); + + if (aq_fw_reset(sc)) + return; + + DPRINTF((", FW version 0x%x", sc->sc_fw_version)); + + if (aq_fw_version_init(sc)) + return; + + if (aq_hw_init_ucp(sc)) + return; + + if (aq_hw_reset(sc)) + return; + + if (aq_get_mac_addr(sc)) + return; + + if (aq_hw_init(sc, irqmode)) + return; + + sc->sc_media_type = aqp->aq_media_type; + sc->sc_available_rates = aqp->aq_available_rates; + + ifmedia_init(&sc->sc_media, IFM_IMASK, aq_ifmedia_change, + aq_ifmedia_status); + + bcopy(sc->sc_enaddr.ether_addr_octet, sc->sc_arpcom.ac_enaddr, 6); + strlcpy(ifp->if_xname, self->dv_xname, IFNAMSIZ); + ifp->if_softc = sc; + ifp->if_flags = IFF_BROADCAST | IFF_MULTICAST | IFF_SIMPLEX; + ifp->if_xflags = IFXF_MPSAFE; + ifp->if_ioctl = aq_ioctl; + ifp->if_qstart = aq_start; + ifp->if_watchdog = aq_watchdog; + ifp->if_hardmtu = 9000; + ifp->if_capabilities = IFCAP_VLAN_MTU; + ifq_set_maxlen(&ifp->if_snd, AQ_TXD_NUM); + + ifmedia_init(&sc->sc_media, IFM_IMASK, aq_ifmedia_change, + aq_ifmedia_status); + if (sc->sc_available_rates & AQ_LINK_100M) { + ifmedia_add(&sc->sc_media, IFM_ETHER|IFM_100_TX, 0, NULL); + ifmedia_add(&sc->sc_media, IFM_ETHER|IFM_100_TX|IFM_FDX, 0, + NULL); + } + + if (sc->sc_available_rates & AQ_LINK_1G) { + ifmedia_add(&sc->sc_media, IFM_ETHER|IFM_1000_T, 0, NULL); + ifmedia_add(&sc->sc_media, IFM_ETHER|IFM_1000_T|IFM_FDX, 0, + NULL); + } + + if (sc->sc_available_rates & AQ_LINK_2G5) { + ifmedia_add(&sc->sc_media, IFM_ETHER | IFM_2500_T, 0, NULL); + ifmedia_add(&sc->sc_media, IFM_ETHER | IFM_2500_T | IFM_FDX, + 0, NULL); + } + + if (sc->sc_available_rates & AQ_LINK_5G) { + ifmedia_add(&sc->sc_media, IFM_ETHER | IFM_5000_T, 0, NULL); + ifmedia_add(&sc->sc_media, IFM_ETHER | IFM_5000_T | IFM_FDX, + 0, NULL); + } + + if (sc->sc_available_rates & AQ_LINK_10G) { + ifmedia_add(&sc->sc_media, IFM_ETHER | IFM_10G_T, 0, NULL); + ifmedia_add(&sc->sc_media, IFM_ETHER | IFM_10G_T | IFM_FDX, + 0, NULL); + } + + ifmedia_add(&sc->sc_media, IFM_ETHER | IFM_AUTO, 0, NULL); + ifmedia_add(&sc->sc_media, IFM_ETHER | IFM_AUTO | IFM_FDX, 0, NULL); + ifmedia_set(&sc->sc_media, IFM_ETHER | IFM_AUTO); + aq_set_linkmode(sc, AQ_LINK_AUTO, AQ_FC_NONE, AQ_EEE_DISABLE); + + if_attach(ifp); + ether_ifattach(ifp); + + if_attach_iqueues(ifp, sc->sc_nqueues); + if_attach_queues(ifp, sc->sc_nqueues); + + for (i = 0; i < sc->sc_nqueues; i++) { + struct aq_queues *aq = &sc->sc_queues[i]; + struct aq_rxring *rx = &aq->q_rx; + struct aq_txring *tx = &aq->q_tx; + + aq->q_sc = sc; + aq->q_index = i; + rx->rx_q = i; + rx->rx_irq = i * 2; + rx->rx_ifiq = ifp->if_iqs[i]; + ifp->if_iqs[i]->ifiq_softc = aq; + timeout_set(&rx->rx_refill, aq_refill, rx); + + tx->tx_q = i; + tx->tx_irq = rx->rx_irq + 1; + tx->tx_ifq = ifp->if_ifqs[i]; + ifp->if_ifqs[i]->ifq_softc = aq; + + if (sc->sc_nqueues > 1) { + /* map msix */ + } + + AQ_WRITE_REG(sc, TX_INTR_MODERATION_CTL_REG(i), 0); + AQ_WRITE_REG(sc, RX_INTR_MODERATION_CTL_REG(i), 0); + } + + AQ_WRITE_REG_BIT(sc, TX_DMA_INT_DESC_WRWB_EN_REG, + TX_DMA_INT_DESC_WRWB_EN, 1); + AQ_WRITE_REG_BIT(sc, TX_DMA_INT_DESC_WRWB_EN_REG, + TX_DMA_INT_DESC_MODERATE_EN, 0); + AQ_WRITE_REG_BIT(sc, RX_DMA_INT_DESC_WRWB_EN_REG, + RX_DMA_INT_DESC_WRWB_EN, 1); + AQ_WRITE_REG_BIT(sc, RX_DMA_INT_DESC_WRWB_EN_REG, + RX_DMA_INT_DESC_MODERATE_EN, 0); + + aq_enable_intr(sc, 1, 0); + printf("\n"); +} + +int +aq_fw_reset(struct aq_softc *sc) +{ + uint32_t ver, v, boot_exit_code; + int i, error; + enum aq_fw_bootloader_mode mode; + + mode = FW_BOOT_MODE_UNKNOWN; + + ver = AQ_READ_REG(sc, AQ_FW_VERSION_REG); + + for (i = 1000; i > 0; i--) { + v = AQ_READ_REG(sc, FW_MPI_DAISY_CHAIN_STATUS_REG); + boot_exit_code = AQ_READ_REG(sc, FW_BOOT_EXIT_CODE_REG); + if (v != 0x06000000 || boot_exit_code != 0) + break; + } + + if (i <= 0) { + printf("%s: F/W reset failed. Neither RBL nor FLB started", + DEVNAME(sc)); + return ETIMEDOUT; + } + + sc->sc_rbl_enabled = (boot_exit_code != 0); + + /* + * Having FW version 0 is an indicator that cold start + * is in progress. This means two things: + * 1) Driver have to wait for FW/HW to finish boot (500ms giveup) + * 2) Driver may skip reset sequence and save time. + */ + if (sc->sc_fast_start_enabled && (ver != 0)) { + error = aq_fw_read_version(sc); + /* Skip reset as it just completed */ + if (error == 0) + return 0; + } + + error = aq_mac_soft_reset(sc, &mode); + if (error != 0) { + printf("%s: MAC reset failed: %d\n", DEVNAME(sc), error); + return error; + } + + switch (mode) { + case FW_BOOT_MODE_FLB: + DPRINTF(("%s: FLB> F/W successfully loaded from flash.", + DEVNAME(sc))); + sc->sc_flash_present = 1; + return aq_fw_read_version(sc); + case FW_BOOT_MODE_RBL_FLASH: + DPRINTF(("%s: RBL> F/W loaded from flash. Host Bootload " + "disabled.", DEVNAME(sc))); + sc->sc_flash_present = 1; + return aq_fw_read_version(sc); + case FW_BOOT_MODE_UNKNOWN: + printf("%s: F/W bootload error: unknown bootloader type", + DEVNAME(sc)); + return ENOTSUP; + case FW_BOOT_MODE_RBL_HOST_BOOTLOAD: + printf("%s: RBL> F/W Host Bootload not implemented", DEVNAME(sc)); + return ENOTSUP; + } + + return ENOTSUP; +} + +int +aq_mac_soft_reset_rbl(struct aq_softc *sc, enum aq_fw_bootloader_mode *mode) +{ + int timo; + + DPRINTF(("%s: RBL> MAC reset STARTED!\n", DEVNAME(sc))); + + AQ_WRITE_REG(sc, AQ_FW_GLB_CTL2_REG, 0x40e1); + AQ_WRITE_REG(sc, AQ_FW_GLB_CPU_SEM_REG(0), 1); + AQ_WRITE_REG(sc, AQ_MBOXIF_POWER_GATING_CONTROL_REG, 0); + + /* MAC FW will reload PHY FW if 1E.1000.3 was cleaned - #undone */ + AQ_WRITE_REG(sc, FW_BOOT_EXIT_CODE_REG, RBL_STATUS_DEAD); + + aq_global_software_reset(sc); + + AQ_WRITE_REG(sc, AQ_FW_GLB_CTL2_REG, 0x40e0); + + /* Wait for RBL to finish boot process. */ +#define RBL_TIMEOUT_MS 10000 + uint16_t rbl_status; + for (timo = RBL_TIMEOUT_MS; timo > 0; timo--) { + rbl_status = AQ_READ_REG(sc, FW_BOOT_EXIT_CODE_REG) & 0xffff; + if (rbl_status != 0 && rbl_status != RBL_STATUS_DEAD) + break; + delay(1000); + } + + if (timo <= 0) { + printf("%s: RBL> RBL restart failed: timeout\n", DEVNAME(sc)); + return EBUSY; + } + + switch (rbl_status) { + case RBL_STATUS_SUCCESS: + if (mode != NULL) + *mode = FW_BOOT_MODE_RBL_FLASH; + DPRINTF(("%s: RBL> reset complete! [Flash]\n", DEVNAME(sc))); + break; + case RBL_STATUS_HOST_BOOT: + if (mode != NULL) + *mode = FW_BOOT_MODE_RBL_HOST_BOOTLOAD; + DPRINTF(("%s: RBL> reset complete! [Host Bootload]\n", + DEVNAME(sc))); + break; + case RBL_STATUS_FAILURE: + default: + printf("%s: unknown RBL status 0x%x\n", DEVNAME(sc), + rbl_status); + return EBUSY; + } + + return 0; +} + +int +aq_mac_soft_reset_flb(struct aq_softc *sc) +{ + uint32_t v; + int timo; + + AQ_WRITE_REG(sc, AQ_FW_GLB_CTL2_REG, 0x40e1); + /* + * Let Felicity hardware to complete SMBUS transaction before + * Global software reset. + */ + delay(50000); + + /* + * If SPI burst transaction was interrupted(before running the script), + * global software reset may not clear SPI interface. + * Clean it up manually before global reset. + */ + AQ_WRITE_REG(sc, AQ_GLB_NVR_PROVISIONING2_REG, 0x00a0); + AQ_WRITE_REG(sc, AQ_GLB_NVR_INTERFACE1_REG, 0x009f); + AQ_WRITE_REG(sc, AQ_GLB_NVR_INTERFACE1_REG, 0x809f); + delay(50000); + + v = AQ_READ_REG(sc, AQ_FW_SOFTRESET_REG); + v &= ~AQ_FW_SOFTRESET_DIS; + v |= AQ_FW_SOFTRESET_RESET; + AQ_WRITE_REG(sc, AQ_FW_SOFTRESET_REG, v); + + /* Kickstart. */ + AQ_WRITE_REG(sc, AQ_FW_GLB_CTL2_REG, 0x80e0); + AQ_WRITE_REG(sc, AQ_MBOXIF_POWER_GATING_CONTROL_REG, 0); + if (!sc->sc_fast_start_enabled) + AQ_WRITE_REG(sc, AQ_GLB_GENERAL_PROVISIONING9_REG, 1); + + /* + * For the case SPI burst transaction was interrupted (by MCP reset + * above), wait until it is completed by hardware. + */ + delay(50000); + + /* MAC Kickstart */ + if (!sc->sc_fast_start_enabled) { + AQ_WRITE_REG(sc, AQ_FW_GLB_CTL2_REG, 0x180e0); + + uint32_t flb_status; + for (timo = 0; timo < 1000; timo++) { + flb_status = AQ_READ_REG(sc, + FW_MPI_DAISY_CHAIN_STATUS_REG) & 0x10; + if (flb_status != 0) + break; + delay(1000); + } + if (flb_status == 0) { + printf("%s: FLB> MAC kickstart failed: timed out\n", + DEVNAME(sc)); + return ETIMEDOUT; + } + DPRINTF(("%s: FLB> MAC kickstart done, %d ms\n", DEVNAME(sc), + timo)); + /* FW reset */ + AQ_WRITE_REG(sc, AQ_FW_GLB_CTL2_REG, 0x80e0); + /* + * Let Felicity hardware complete SMBUS transaction before + * Global software reset. + */ + delay(50000); + sc->sc_fast_start_enabled = true; + } + AQ_WRITE_REG(sc, AQ_FW_GLB_CPU_SEM_REG(0), 1); + + /* PHY Kickstart: #undone */ + aq_global_software_reset(sc); + + for (timo = 0; timo < 1000; timo++) { + if (AQ_READ_REG(sc, AQ_FW_VERSION_REG) != 0) + break; + delay(10000); + } + if (timo >= 1000) { + printf("%s: FLB> Global Soft Reset failed\n", DEVNAME(sc)); + return ETIMEDOUT; + } + DPRINTF(("%s: FLB> F/W restart: %d ms\n", DEVNAME(sc), timo * 10)); + + return 0; + +} + +int +aq_mac_soft_reset(struct aq_softc *sc, enum aq_fw_bootloader_mode *mode) +{ + if (sc->sc_rbl_enabled) + return aq_mac_soft_reset_rbl(sc, mode); + + if (mode != NULL) + *mode = FW_BOOT_MODE_FLB; + return aq_mac_soft_reset_flb(sc); +} + +void +aq_global_software_reset(struct aq_softc *sc) +{ + uint32_t v; + + AQ_WRITE_REG_BIT(sc, RX_SYSCONTROL_REG, RX_SYSCONTROL_RESET_DIS, 0); + AQ_WRITE_REG_BIT(sc, TX_SYSCONTROL_REG, TX_SYSCONTROL_RESET_DIS, 0); + AQ_WRITE_REG_BIT(sc, FW_MPI_RESETCTRL_REG, + FW_MPI_RESETCTRL_RESET_DIS, 0); + + v = AQ_READ_REG(sc, AQ_FW_SOFTRESET_REG); + v &= ~AQ_FW_SOFTRESET_DIS; + v |= AQ_FW_SOFTRESET_RESET; + AQ_WRITE_REG(sc, AQ_FW_SOFTRESET_REG, v); +} + +int +aq_fw_read_version(struct aq_softc *sc) +{ + int i, error = EBUSY; +#define MAC_FW_START_TIMEOUT_MS 10000 + for (i = 0; i < MAC_FW_START_TIMEOUT_MS; i++) { + sc->sc_fw_version = AQ_READ_REG(sc, AQ_FW_VERSION_REG); + if (sc->sc_fw_version != 0) { + error = 0; + break; + } + delay(1000); + } + return error; +} + +int +aq_fw_version_init(struct aq_softc *sc) +{ + int error = 0; + char fw_vers[sizeof("F/W version xxxxx.xxxxx.xxxxx")]; + + if (FW_VERSION_MAJOR(sc) == 1) { + sc->sc_fw_ops = &aq_fw1x_ops; + } else if ((FW_VERSION_MAJOR(sc) == 2) || (FW_VERSION_MAJOR(sc) == 3)) { + sc->sc_fw_ops = &aq_fw2x_ops; + } else { + printf("%s: Unsupported F/W version %d.%d.%d\n", + DEVNAME(sc), + FW_VERSION_MAJOR(sc), FW_VERSION_MINOR(sc), + FW_VERSION_BUILD(sc)); + return ENOTSUP; + } + snprintf(fw_vers, sizeof(fw_vers), "F/W version %d.%d.%d", + FW_VERSION_MAJOR(sc), FW_VERSION_MINOR(sc), FW_VERSION_BUILD(sc)); + + /* detect revision */ + uint32_t hwrev = AQ_READ_REG(sc, AQ_HW_REVISION_REG); + switch (hwrev & 0x0000000f) { + case 0x01: + printf(", revision A0, %s", fw_vers); + sc->sc_features |= FEATURES_REV_A0 | + FEATURES_MPI_AQ | FEATURES_MIPS; + break; + case 0x02: + printf(", revision B0, %s", fw_vers); + sc->sc_features |= FEATURES_REV_B0 | + FEATURES_MPI_AQ | FEATURES_MIPS | + FEATURES_TPO2 | FEATURES_RPF2; + break; + case 0x0A: + printf(", revision B1, %s", fw_vers); + sc->sc_features |= FEATURES_REV_B1 | + FEATURES_MPI_AQ | FEATURES_MIPS | + FEATURES_TPO2 | FEATURES_RPF2; + break; + default: + printf(", Unknown revision (0x%08x)", hwrev); + error = ENOTSUP; + break; + } + return error; +} + +int +aq_hw_init_ucp(struct aq_softc *sc) +{ + int timo; + + if (FW_VERSION_MAJOR(sc) == 1) { + if (AQ_READ_REG(sc, FW1X_MPI_INIT2_REG) == 0) { + uint32_t data; + arc4random_buf(&data, sizeof(data)); + data &= 0xfefefefe; + data |= 0x02020202; + AQ_WRITE_REG(sc, FW1X_MPI_INIT2_REG, data); + } + AQ_WRITE_REG(sc, FW1X_MPI_INIT1_REG, 0); + } + + for (timo = 100; timo > 0; timo--) { + sc->sc_mbox_addr = AQ_READ_REG(sc, FW_MPI_MBOX_ADDR_REG); + if (sc->sc_mbox_addr != 0) + break; + delay(1000); + } + +#define AQ_FW_MIN_VERSION 0x01050006 +#define AQ_FW_MIN_VERSION_STR "1.5.6" + if (sc->sc_fw_version < AQ_FW_MIN_VERSION) { + printf("%s: atlantic: wrong FW version: " AQ_FW_MIN_VERSION_STR + " or later required, this is %d.%d.%d\n", + DEVNAME(sc), + FW_VERSION_MAJOR(sc), + FW_VERSION_MINOR(sc), + FW_VERSION_BUILD(sc)); + return ENOTSUP; + } + + if (sc->sc_mbox_addr == 0) + printf("%s: NULL MBOX!!\n", DEVNAME(sc)); + + return 0; +} + +int +aq_hw_reset(struct aq_softc *sc) +{ + int error; + + /* disable irq */ + AQ_WRITE_REG_BIT(sc, AQ_INTR_CTRL_REG, AQ_INTR_CTRL_RESET_DIS, 0); + + /* apply */ + AQ_WRITE_REG_BIT(sc, AQ_INTR_CTRL_REG, AQ_INTR_CTRL_RESET_IRQ, 1); + + /* wait ack 10 times by 1ms */ + WAIT_FOR( + (AQ_READ_REG(sc, AQ_INTR_CTRL_REG) & AQ_INTR_CTRL_RESET_IRQ) == 0, + 1000, 10, &error); + if (error != 0) { + printf("%s: atlantic: IRQ reset failed: %d\n", DEVNAME(sc), + error); + return error; + } + + return sc->sc_fw_ops->reset(sc); +} + +int +aq_get_mac_addr(struct aq_softc *sc) +{ + uint32_t mac_addr[2]; + uint32_t efuse_shadow_addr; + int err; + + efuse_shadow_addr = 0; + if (FW_VERSION_MAJOR(sc) >= 2) + efuse_shadow_addr = AQ_READ_REG(sc, FW2X_MPI_EFUSEADDR_REG); + else + efuse_shadow_addr = AQ_READ_REG(sc, FW1X_MPI_EFUSEADDR_REG); + + if (efuse_shadow_addr == 0) { + printf("%s: cannot get efuse addr", DEVNAME(sc)); + return ENXIO; + } + + DPRINTF(("%s: efuse_shadow_addr = %x\n", DEVNAME(sc), efuse_shadow_addr)); + + memset(mac_addr, 0, sizeof(mac_addr)); + err = aq_fw_downld_dwords(sc, efuse_shadow_addr + (40 * 4), + mac_addr, 2); + if (err < 0) + return err; + + if (mac_addr[0] == 0 && mac_addr[1] == 0) { + printf("%s: mac address not found", DEVNAME(sc)); + return ENXIO; + } + + DPRINTF(("%s: mac0 %x mac1 %x\n", DEVNAME(sc), mac_addr[0], + mac_addr[1])); + + mac_addr[0] = htobe32(mac_addr[0]); + mac_addr[1] = htobe32(mac_addr[1]); + + DPRINTF(("%s: mac0 %x mac1 %x\n", DEVNAME(sc), mac_addr[0], + mac_addr[1])); + + memcpy(sc->sc_enaddr.ether_addr_octet, + (uint8_t *)mac_addr, ETHER_ADDR_LEN); + DPRINTF((": %s", ether_sprintf(sc->sc_enaddr.ether_addr_octet))); + + return 0; +} + +int +aq_activate(struct device *self, int act) +{ + return 0; +} + +int +aq_fw_downld_dwords(struct aq_softc *sc, uint32_t addr, uint32_t *p, + uint32_t cnt) +{ + uint32_t v; + int error = 0; + + WAIT_FOR(AQ_READ_REG(sc, AQ_FW_SEM_RAM_REG) == 1, 1, 10000, &error); + if (error != 0) { + AQ_WRITE_REG(sc, AQ_FW_SEM_RAM_REG, 1); + v = AQ_READ_REG(sc, AQ_FW_SEM_RAM_REG); + if (v == 0) { + printf("%s: %s:%d: timeout\n", + DEVNAME(sc), __func__, __LINE__); + return ETIMEDOUT; + } + } + + AQ_WRITE_REG(sc, AQ_FW_MBOX_ADDR_REG, addr); + + error = 0; + for (; cnt > 0 && error == 0; cnt--) { + /* execute mailbox interface */ + AQ_WRITE_REG_BIT(sc, AQ_FW_MBOX_CMD_REG, + AQ_FW_MBOX_CMD_EXECUTE, 1); + if (sc->sc_features & FEATURES_REV_B1) { + WAIT_FOR(AQ_READ_REG(sc, AQ_FW_MBOX_ADDR_REG) != addr, + 1, 1000, &error); + } else { + WAIT_FOR((AQ_READ_REG(sc, AQ_FW_MBOX_CMD_REG) & + AQ_FW_MBOX_CMD_BUSY) == 0, + 1, 1000, &error); + } + *p++ = AQ_READ_REG(sc, AQ_FW_MBOX_VAL_REG); + addr += sizeof(uint32_t); + } + AQ_WRITE_REG(sc, AQ_FW_SEM_RAM_REG, 1); + + if (error != 0) + printf("%s: %s:%d: timeout\n", + DEVNAME(sc), __func__, __LINE__); + + return error; +} + +int +aq_fw2x_reset(struct aq_softc *sc) +{ + struct aq_fw2x_capabilities caps = { 0 }; + int error; + + error = aq_fw_downld_dwords(sc, + sc->sc_mbox_addr + offsetof(struct aq_fw2x_mailbox, caps), + (uint32_t *)&caps, sizeof caps / sizeof(uint32_t)); + if (error != 0) { + printf("%s: fw2x> can't get F/W capabilities mask, error %d\n", + DEVNAME(sc), error); + return error; + } + sc->sc_fw_caps = caps.caps_lo | ((uint64_t)caps.caps_hi << 32); + + DPRINTF(("%s: fw2x> F/W capabilities=0x%llx\n", DEVNAME(sc), + sc->sc_fw_caps)); + + return 0; +} + +int +aq_fw1x_reset(struct aq_softc *sc) +{ + printf("%s: unimplemented %s\n", DEVNAME(sc), __func__); + return 0; +} + +int +aq_fw1x_set_mode(struct aq_softc *sc, enum aq_hw_fw_mpi_state w, + enum aq_link_speed x, enum aq_link_fc y, enum aq_link_eee z) +{ + return 0; +} + +int +aq_fw1x_get_mode(struct aq_softc *sc, enum aq_hw_fw_mpi_state *w, + enum aq_link_speed *x, enum aq_link_fc *y, enum aq_link_eee *z) +{ + return 0; +} + +int +aq_fw1x_get_stats(struct aq_softc *sc, struct aq_hw_stats_s *w) +{ + return 0; +} + + +int +aq_fw2x_get_mode(struct aq_softc *sc, enum aq_hw_fw_mpi_state *modep, + enum aq_link_speed *speedp, enum aq_link_fc *fcp, enum aq_link_eee *eeep) +{ + uint64_t mpi_state, mpi_ctrl; + enum aq_link_speed speed; + enum aq_link_fc fc; + + AQ_MPI_LOCK(sc); + + mpi_state = AQ_READ64_REG(sc, FW2X_MPI_STATE_REG); + if (modep != NULL) { + mpi_ctrl = AQ_READ64_REG(sc, FW2X_MPI_CONTROL_REG); + if (mpi_ctrl & FW2X_CTRL_RATE_MASK) + *modep = MPI_INIT; + else + *modep = MPI_DEINIT; + } + + AQ_MPI_UNLOCK(sc); + + if (mpi_state & FW2X_CTRL_RATE_10G) + speed = AQ_LINK_10G; + else if (mpi_state & FW2X_CTRL_RATE_5G) + speed = AQ_LINK_5G; + else if (mpi_state & FW2X_CTRL_RATE_2G5) + speed = AQ_LINK_2G5; + else if (mpi_state & FW2X_CTRL_RATE_1G) + speed = AQ_LINK_1G; + else if (mpi_state & FW2X_CTRL_RATE_100M) + speed = AQ_LINK_100M; + else + speed = AQ_LINK_NONE; + if (speedp != NULL) + *speedp = speed; + + fc = AQ_FC_NONE; + if (mpi_state & FW2X_CTRL_PAUSE) + fc |= AQ_FC_RX; + if (mpi_state & FW2X_CTRL_ASYMMETRIC_PAUSE) + fc |= AQ_FC_TX; + if (fcp != NULL) + *fcp = fc; + + if (eeep != NULL) + *eeep = AQ_EEE_DISABLE; + + return 0; +} + +int +aq_fw2x_get_stats(struct aq_softc *sc, struct aq_hw_stats_s *w) +{ + return 0; +} + +void +aq_hw_l3_filter_set(struct aq_softc *sc) +{ + int i; + + /* clear all filter */ + for (i = 0; i < 8; i++) { + AQ_WRITE_REG_BIT(sc, RPF_L3_FILTER_REG(i), + RPF_L3_FILTER_L4_EN, 0); + } +} + +int +aq_hw_init(struct aq_softc *sc, int irqmode) +{ + uint32_t v; + + /* Force limit MRRS on RDM/TDM to 2K */ + v = AQ_READ_REG(sc, AQ_PCI_REG_CONTROL_6_REG); + AQ_WRITE_REG(sc, AQ_PCI_REG_CONTROL_6_REG, (v & ~0x0707) | 0x0404); + + /* + * TX DMA total request limit. B0 hardware is not capable to + * handle more than (8K-MRRS) incoming DMA data. + * Value 24 in 256byte units + */ + AQ_WRITE_REG(sc, AQ_HW_TX_DMA_TOTAL_REQ_LIMIT_REG, 24); + + aq_hw_init_tx_path(sc); + aq_hw_init_rx_path(sc); + + if (aq_set_mac_addr(sc, AQ_HW_MAC_OWN, sc->sc_enaddr.ether_addr_octet)) + return EINVAL; + + aq_set_linkmode(sc, AQ_LINK_NONE, AQ_FC_NONE, AQ_EEE_DISABLE); + + aq_hw_qos_set(sc); + + /* Enable interrupt */ + AQ_WRITE_REG(sc, AQ_INTR_CTRL_REG, AQ_INTR_CTRL_RESET_DIS); + AQ_WRITE_REG_BIT(sc, AQ_INTR_CTRL_REG, AQ_INTR_CTRL_MULTIVEC, 0); + + AQ_WRITE_REG_BIT(sc, AQ_INTR_CTRL_REG, AQ_INTR_CTRL_IRQMODE, irqmode); + + AQ_WRITE_REG(sc, AQ_INTR_AUTOMASK_REG, 0xffffffff); + + AQ_WRITE_REG(sc, AQ_GEN_INTR_MAP_REG(0), + ((AQ_B0_ERR_INT << 24) | (1U << 31)) | + ((AQ_B0_ERR_INT << 16) | (1 << 23)) + ); + + /* link interrupt */ + sc->sc_linkstat_irq = AQ_LINKSTAT_IRQ; + AQ_WRITE_REG(sc, AQ_GEN_INTR_MAP_REG(3), + (1 << 7) | sc->sc_linkstat_irq); + + return 0; +} + +void +aq_hw_init_tx_path(struct aq_softc *sc) +{ + /* Tx TC/RSS number config */ + AQ_WRITE_REG_BIT(sc, TPB_TX_BUF_REG, TPB_TX_BUF_TC_MODE_EN, 1); + + AQ_WRITE_REG_BIT(sc, THM_LSO_TCP_FLAG1_REG, + THM_LSO_TCP_FLAG1_FIRST, 0x0ff6); + AQ_WRITE_REG_BIT(sc, THM_LSO_TCP_FLAG1_REG, + THM_LSO_TCP_FLAG1_MID, 0x0ff6); + AQ_WRITE_REG_BIT(sc, THM_LSO_TCP_FLAG2_REG, + THM_LSO_TCP_FLAG2_LAST, 0x0f7f); + + /* misc */ + AQ_WRITE_REG(sc, TX_TPO2_REG, + (sc->sc_features & FEATURES_TPO2) ? TX_TPO2_EN : 0); + AQ_WRITE_REG_BIT(sc, TDM_DCA_REG, TDM_DCA_EN, 0); + AQ_WRITE_REG_BIT(sc, TDM_DCA_REG, TDM_DCA_MODE, 0); + + AQ_WRITE_REG_BIT(sc, TPB_TX_BUF_REG, TPB_TX_BUF_SCP_INS_EN, 1); +} + +void +aq_hw_init_rx_path(struct aq_softc *sc) +{ + int i; + + /* clear setting */ + AQ_WRITE_REG_BIT(sc, RPB_RPF_RX_REG, RPB_RPF_RX_TC_MODE, 0); + AQ_WRITE_REG_BIT(sc, RPB_RPF_RX_REG, RPB_RPF_RX_FC_MODE, 0); + AQ_WRITE_REG(sc, RX_FLR_RSS_CONTROL1_REG, 0); + for (i = 0; i < 32; i++) { + AQ_WRITE_REG_BIT(sc, RPF_ETHERTYPE_FILTER_REG(i), + RPF_ETHERTYPE_FILTER_EN, 0); + } + + /* L2 and Multicast filters */ + for (i = 0; i < AQ_HW_MAC_NUM; i++) { + AQ_WRITE_REG_BIT(sc, RPF_L2UC_MSW_REG(i), RPF_L2UC_MSW_EN, 0); + AQ_WRITE_REG_BIT(sc, RPF_L2UC_MSW_REG(i), RPF_L2UC_MSW_ACTION, + RPF_ACTION_HOST); + } + AQ_WRITE_REG(sc, RPF_MCAST_FILTER_MASK_REG, 0); + AQ_WRITE_REG(sc, RPF_MCAST_FILTER_REG(0), 0x00010fff); + + /* Vlan filters */ + AQ_WRITE_REG_BIT(sc, RPF_VLAN_TPID_REG, RPF_VLAN_TPID_OUTER, + ETHERTYPE_QINQ); + AQ_WRITE_REG_BIT(sc, RPF_VLAN_TPID_REG, RPF_VLAN_TPID_INNER, + ETHERTYPE_VLAN); + AQ_WRITE_REG_BIT(sc, RPF_VLAN_MODE_REG, RPF_VLAN_MODE_PROMISC, 0); + + if (sc->sc_features & FEATURES_REV_B) { + AQ_WRITE_REG_BIT(sc, RPF_VLAN_MODE_REG, + RPF_VLAN_MODE_ACCEPT_UNTAGGED, 1); + AQ_WRITE_REG_BIT(sc, RPF_VLAN_MODE_REG, + RPF_VLAN_MODE_UNTAGGED_ACTION, RPF_ACTION_HOST); + } + + AQ_WRITE_REG(sc, RX_TCP_RSS_HASH_REG, 0); + + AQ_WRITE_REG_BIT(sc, RPF_L2BC_REG, RPF_L2BC_EN, 1); + AQ_WRITE_REG_BIT(sc, RPF_L2BC_REG, RPF_L2BC_ACTION, RPF_ACTION_HOST); + AQ_WRITE_REG_BIT(sc, RPF_L2BC_REG, RPF_L2BC_THRESHOLD, 0xffff); + + AQ_WRITE_REG_BIT(sc, RX_DMA_DCA_REG, RX_DMA_DCA_EN, 0); + AQ_WRITE_REG_BIT(sc, RX_DMA_DCA_REG, RX_DMA_DCA_MODE, 0); +} + +/* set multicast filter. index 0 for own address */ +int +aq_set_mac_addr(struct aq_softc *sc, int index, uint8_t *enaddr) +{ + uint32_t h, l; + + if (index >= AQ_HW_MAC_NUM) + return EINVAL; + + if (enaddr == NULL) { + /* disable */ + AQ_WRITE_REG_BIT(sc, + RPF_L2UC_MSW_REG(index), RPF_L2UC_MSW_EN, 0); + return 0; + } + + h = (enaddr[0] << 8) | (enaddr[1]); + l = ((uint32_t)enaddr[2] << 24) | (enaddr[3] << 16) | + (enaddr[4] << 8) | (enaddr[5]); + + /* disable, set, and enable */ + AQ_WRITE_REG_BIT(sc, RPF_L2UC_MSW_REG(index), RPF_L2UC_MSW_EN, 0); + AQ_WRITE_REG(sc, RPF_L2UC_LSW_REG(index), l); + AQ_WRITE_REG_BIT(sc, RPF_L2UC_MSW_REG(index), + RPF_L2UC_MSW_MACADDR_HI, h); + AQ_WRITE_REG_BIT(sc, RPF_L2UC_MSW_REG(index), RPF_L2UC_MSW_ACTION, 1); + AQ_WRITE_REG_BIT(sc, RPF_L2UC_MSW_REG(index), RPF_L2UC_MSW_EN, 1); + + return 0; +} + +int +aq_get_linkmode(struct aq_softc *sc, enum aq_link_speed *speed, + enum aq_link_fc *fc, enum aq_link_eee *eee) +{ + enum aq_hw_fw_mpi_state mode; + int error; + + error = sc->sc_fw_ops->get_mode(sc, &mode, speed, fc, eee); + if (error != 0) + return error; + if (mode != MPI_INIT) + return ENXIO; + + return 0; +} + +int +aq_set_linkmode(struct aq_softc *sc, enum aq_link_speed speed, + enum aq_link_fc fc, enum aq_link_eee eee) +{ + return sc->sc_fw_ops->set_mode(sc, MPI_INIT, speed, fc, eee); +} + +int +aq_fw2x_set_mode(struct aq_softc *sc, enum aq_hw_fw_mpi_state mode, + enum aq_link_speed speed, enum aq_link_fc fc, enum aq_link_eee eee) +{ + uint64_t mpi_ctrl; + int error = 0; + + AQ_MPI_LOCK(sc); + + mpi_ctrl = AQ_READ64_REG(sc, FW2X_MPI_CONTROL_REG); + + switch (mode) { + case MPI_INIT: + mpi_ctrl &= ~FW2X_CTRL_RATE_MASK; + if (speed & AQ_LINK_10G) + mpi_ctrl |= FW2X_CTRL_RATE_10G; + if (speed & AQ_LINK_5G) + mpi_ctrl |= FW2X_CTRL_RATE_5G; + if (speed & AQ_LINK_2G5) + mpi_ctrl |= FW2X_CTRL_RATE_2G5; + if (speed & AQ_LINK_1G) + mpi_ctrl |= FW2X_CTRL_RATE_1G; + if (speed & AQ_LINK_100M) + mpi_ctrl |= FW2X_CTRL_RATE_100M; + + mpi_ctrl &= ~FW2X_CTRL_LINK_DROP; + + mpi_ctrl &= ~FW2X_CTRL_EEE_MASK; + if (eee == AQ_EEE_ENABLE) + mpi_ctrl |= FW2X_CTRL_EEE_MASK; + + mpi_ctrl &= ~(FW2X_CTRL_PAUSE | FW2X_CTRL_ASYMMETRIC_PAUSE); + if (fc & AQ_FC_RX) + mpi_ctrl |= FW2X_CTRL_PAUSE; + if (fc & AQ_FC_TX) + mpi_ctrl |= FW2X_CTRL_ASYMMETRIC_PAUSE; + break; + case MPI_DEINIT: + mpi_ctrl &= ~(FW2X_CTRL_RATE_MASK | FW2X_CTRL_EEE_MASK); + mpi_ctrl &= ~(FW2X_CTRL_PAUSE | FW2X_CTRL_ASYMMETRIC_PAUSE); + break; + default: + printf("%s: fw2x> unknown MPI state %d\n", DEVNAME(sc), mode); + error = EINVAL; + goto failure; + } + AQ_WRITE64_REG(sc, FW2X_MPI_CONTROL_REG, mpi_ctrl); + + failure: + AQ_MPI_UNLOCK(sc); + return error; +} + +void +aq_hw_qos_set(struct aq_softc *sc) +{ + uint32_t tc = 0; + uint32_t buff_size; + + /* TPS Descriptor rate init */ + AQ_WRITE_REG_BIT(sc, TPS_DESC_RATE_REG, TPS_DESC_RATE_TA_RST, 0); + AQ_WRITE_REG_BIT(sc, TPS_DESC_RATE_REG, TPS_DESC_RATE_LIM, 0xa); + + /* TPS VM init */ + AQ_WRITE_REG_BIT(sc, TPS_DESC_VM_ARB_MODE_REG, TPS_DESC_VM_ARB_MODE, 0); + + /* TPS TC credits init */ + AQ_WRITE_REG_BIT(sc, TPS_DESC_TC_ARB_MODE_REG, TPS_DESC_TC_ARB_MODE, 0); + AQ_WRITE_REG_BIT(sc, TPS_DATA_TC_ARB_MODE_REG, TPS_DATA_TC_ARB_MODE, 0); + + AQ_WRITE_REG_BIT(sc, TPS_DATA_TCT_REG(tc), + TPS_DATA_TCT_CREDIT_MAX, 0xfff); + AQ_WRITE_REG_BIT(sc, TPS_DATA_TCT_REG(tc), + TPS_DATA_TCT_WEIGHT, 0x64); + AQ_WRITE_REG_BIT(sc, TPS_DESC_TCT_REG(tc), + TPS_DESC_TCT_CREDIT_MAX, 0x50); + AQ_WRITE_REG_BIT(sc, TPS_DESC_TCT_REG(tc), + TPS_DESC_TCT_WEIGHT, 0x1e); + + /* Tx buf size */ + tc = 0; + buff_size = AQ_HW_TXBUF_MAX; + AQ_WRITE_REG_BIT(sc, TPB_TXB_BUFSIZE_REG(tc), TPB_TXB_BUFSIZE, + buff_size); + AQ_WRITE_REG_BIT(sc, TPB_TXB_THRESH_REG(tc), TPB_TXB_THRESH_HI, + (buff_size * (1024 / 32) * 66) / 100); + AQ_WRITE_REG_BIT(sc, TPB_TXB_THRESH_REG(tc), TPB_TXB_THRESH_LO, + (buff_size * (1024 / 32) * 50) / 100); + + /* QoS Rx buf size per TC */ + tc = 0; + buff_size = AQ_HW_RXBUF_MAX; + AQ_WRITE_REG_BIT(sc, RPB_RXB_BUFSIZE_REG(tc), RPB_RXB_BUFSIZE, + buff_size); + AQ_WRITE_REG_BIT(sc, RPB_RXB_XOFF_REG(tc), RPB_RXB_XOFF_EN, 0); + AQ_WRITE_REG_BIT(sc, RPB_RXB_XOFF_REG(tc), RPB_RXB_XOFF_THRESH_HI, + (buff_size * (1024 / 32) * 66) / 100); + AQ_WRITE_REG_BIT(sc, RPB_RXB_XOFF_REG(tc), RPB_RXB_XOFF_THRESH_LO, + (buff_size * (1024 / 32) * 50) / 100); + + /* QoS 802.1p priority -> TC mapping */ + int i_priority; + for (i_priority = 0; i_priority < 8; i_priority++) { + AQ_WRITE_REG_BIT(sc, RPF_RPB_RX_TC_UPT_REG, + RPF_RPB_RX_TC_UPT_MASK(i_priority), 0); + } +} + +void +aq_txring_reset(struct aq_softc *sc, struct aq_txring *tx, int start) +{ + daddr_t paddr; + + tx->tx_prod = 0; + tx->tx_cons = 0; + + /* empty slots? */ + + AQ_WRITE_REG_BIT(sc, TX_DMA_DESC_REG(tx->tx_q), TX_DMA_DESC_EN, 0); + + if (start == 0) + return; + + paddr = AQ_DMA_DVA(&tx->tx_mem); + AQ_WRITE_REG(sc, TX_DMA_DESC_BASE_ADDRLSW_REG(tx->tx_q), paddr); + AQ_WRITE_REG(sc, TX_DMA_DESC_BASE_ADDRMSW_REG(tx->tx_q), + paddr >> 32); + + AQ_WRITE_REG_BIT(sc, TX_DMA_DESC_REG(tx->tx_q), TX_DMA_DESC_LEN, + AQ_TXD_NUM / 8); + + tx->tx_prod = AQ_READ_REG(sc, TX_DMA_DESC_TAIL_PTR_REG(tx->tx_q)); + tx->tx_cons = tx->tx_prod; + AQ_WRITE_REG(sc, TX_DMA_DESC_WRWB_THRESH_REG(tx->tx_q), 0); + + AQ_WRITE_REG_BIT(sc, AQ_INTR_IRQ_MAP_TX_REG(tx->tx_q), + AQ_INTR_IRQ_MAP_TX_IRQMAP(tx->tx_q), tx->tx_irq); + AQ_WRITE_REG_BIT(sc, AQ_INTR_IRQ_MAP_TX_REG(tx->tx_q), + AQ_INTR_IRQ_MAP_TX_EN(tx->tx_q), 1); + + AQ_WRITE_REG_BIT(sc, TX_DMA_DESC_REG(tx->tx_q), TX_DMA_DESC_EN, 1); + + AQ_WRITE_REG_BIT(sc, TDM_DCAD_REG(tx->tx_q), TDM_DCAD_CPUID, 0); + AQ_WRITE_REG_BIT(sc, TDM_DCAD_REG(tx->tx_q), TDM_DCAD_CPUID_EN, 0); +} + +void +aq_rxring_reset(struct aq_softc *sc, struct aq_rxring *rx, int start) +{ + daddr_t paddr; + + AQ_WRITE_REG_BIT(sc, RX_DMA_DESC_REG(rx->rx_q), RX_DMA_DESC_EN, 0); + /* drain */ + + if (start == 0) + return; + + paddr = AQ_DMA_DVA(&rx->rx_mem); + AQ_WRITE_REG(sc, RX_DMA_DESC_BASE_ADDRLSW_REG(rx->rx_q), paddr); + AQ_WRITE_REG(sc, RX_DMA_DESC_BASE_ADDRMSW_REG(rx->rx_q), + paddr >> 32); + + AQ_WRITE_REG_BIT(sc, RX_DMA_DESC_REG(rx->rx_q), RX_DMA_DESC_LEN, + AQ_RXD_NUM / 8); + + AQ_WRITE_REG_BIT(sc, RX_DMA_DESC_BUFSIZE_REG(rx->rx_q), + RX_DMA_DESC_BUFSIZE_DATA, MCLBYTES / 1024); + AQ_WRITE_REG_BIT(sc, RX_DMA_DESC_BUFSIZE_REG(rx->rx_q), + RX_DMA_DESC_BUFSIZE_HDR, 0); + AQ_WRITE_REG_BIT(sc, RX_DMA_DESC_REG(rx->rx_q), + RX_DMA_DESC_HEADER_SPLIT, 0); + AQ_WRITE_REG_BIT(sc, RX_DMA_DESC_REG(rx->rx_q), + RX_DMA_DESC_VLAN_STRIP, 0); + + rx->rx_cons = AQ_READ_REG(sc, RX_DMA_DESC_HEAD_PTR_REG(rx->rx_q)) & + RX_DMA_DESC_HEAD_PTR; + AQ_WRITE_REG(sc, RX_DMA_DESC_TAIL_PTR_REG(rx->rx_q), rx->rx_cons); + rx->rx_prod = rx->rx_cons; + + AQ_WRITE_REG_BIT(sc, AQ_INTR_IRQ_MAP_RX_REG(rx->rx_q), + AQ_INTR_IRQ_MAP_RX_IRQMAP(rx->rx_q), rx->rx_irq); + AQ_WRITE_REG_BIT(sc, AQ_INTR_IRQ_MAP_RX_REG(rx->rx_q), + AQ_INTR_IRQ_MAP_RX_EN(rx->rx_q), 1); + + AQ_WRITE_REG_BIT(sc, RX_DMA_DCAD_REG(rx->rx_q), + RX_DMA_DCAD_CPUID, 0); + AQ_WRITE_REG_BIT(sc, RX_DMA_DCAD_REG(rx->rx_q), + RX_DMA_DCAD_DESC_EN, 0); + AQ_WRITE_REG_BIT(sc, RX_DMA_DCAD_REG(rx->rx_q), + RX_DMA_DCAD_HEADER_EN, 0); + AQ_WRITE_REG_BIT(sc, RX_DMA_DCAD_REG(rx->rx_q), + RX_DMA_DCAD_PAYLOAD_EN, 0); + + AQ_WRITE_REG_BIT(sc, RX_DMA_DESC_REG(rx->rx_q), RX_DMA_DESC_EN, 1); +} + +static inline unsigned int +aq_rx_fill_slots(struct aq_softc *sc, struct aq_rxring *rx, uint nslots) +{ + struct aq_rx_desc_read *ring, *rd; + struct aq_slot *as; + struct mbuf *m; + uint p, fills; + + ring = AQ_DMA_KVA(&rx->rx_mem); + p = rx->rx_prod; + + bus_dmamap_sync(sc->sc_dmat, AQ_DMA_MAP(&rx->rx_mem), 0, + AQ_DMA_LEN(&rx->rx_mem), BUS_DMASYNC_POSTWRITE); + + for (fills = 0; fills < nslots; fills++) { + as = &rx->rx_slots[p]; + rd = &ring[p]; + + m = MCLGETL(NULL, M_DONTWAIT, MCLBYTES + ETHER_ALIGN); + if (m == NULL) + break; + + m->m_data += (m->m_ext.ext_size - MCLBYTES); + m->m_data += ETHER_ALIGN; + m->m_len = m->m_pkthdr.len = MCLBYTES; + + if (bus_dmamap_load_mbuf(sc->sc_dmat, as->as_map, m, + BUS_DMA_NOWAIT) != 0) { + m_freem(m); + break; + } + as->as_m = m; + + htolem64(&rd->buf_addr, as->as_map->dm_segs[0].ds_addr); + rd->hdr_addr = 0; + p++; + if (p == AQ_RXD_NUM) + p = 0; + } + + bus_dmamap_sync(sc->sc_dmat, AQ_DMA_MAP(&rx->rx_mem), 0, + AQ_DMA_LEN(&rx->rx_mem), BUS_DMASYNC_PREWRITE); + + rx->rx_prod = p; + AQ_WRITE_REG(sc, RX_DMA_DESC_TAIL_PTR_REG(rx->rx_q), rx->rx_prod); + return (nslots - fills); +} + +int +aq_rx_fill(struct aq_softc *sc, struct aq_rxring *rx) +{ + u_int slots; + + slots = if_rxr_get(&rx->rx_rxr, AQ_RXD_NUM); + if (slots == 0) + return 1; + + slots = aq_rx_fill_slots(sc, rx, slots); + if_rxr_put(&rx->rx_rxr, slots); + return 0; +} + +void +aq_refill(void *xq) +{ + struct aq_queues *q = xq; + struct aq_softc *sc = q->q_sc; + + aq_rx_fill(sc, &q->q_rx); + + if (if_rxr_inuse(&q->q_rx.rx_rxr) == 0) + timeout_add(&q->q_rx.rx_refill, 1); +} + +void +aq_rxeof(struct aq_softc *sc, struct aq_rxring *rx) +{ + struct ifnet *ifp = &sc->sc_arpcom.ac_if; + struct aq_rx_desc_wb *rxd; + struct aq_rx_desc_wb *ring; + struct aq_slot *as; + uint32_t end, idx; + uint16_t pktlen, status; + uint32_t rxd_type; + struct mbuf *m; + struct mbuf_list ml = MBUF_LIST_INITIALIZER(); + int rxfree; + + if (!ISSET(ifp->if_flags, IFF_RUNNING)) + return; + + end = AQ_READ_REG(sc, RX_DMA_DESC_HEAD_PTR_REG(rx->rx_q)) & + RX_DMA_DESC_HEAD_PTR; + + bus_dmamap_sync(sc->sc_dmat, AQ_DMA_MAP(&rx->rx_mem), 0, + AQ_DMA_LEN(&rx->rx_mem), BUS_DMASYNC_POSTREAD); + + rxfree = 0; + idx = rx->rx_cons; + ring = AQ_DMA_KVA(&rx->rx_mem); + while (idx != end) { + rxd = &ring[idx]; + as = &rx->rx_slots[idx]; + + bus_dmamap_sync(sc->sc_dmat, as->as_map, 0, + as->as_map->dm_mapsize, BUS_DMASYNC_POSTREAD); + bus_dmamap_unload(sc->sc_dmat, as->as_map); + + status = lemtoh16(&rxd->status); + if ((status & AQ_RXDESC_STATUS_DD) == 0) + break; + + rxfree++; + m = as->as_m; + as->as_m = NULL; + + pktlen = lemtoh16(&rxd->pkt_len); + rxd_type = lemtoh32(&rxd->type); + /* rss hash, vlan */ + + if ((status & AQ_RXDESC_STATUS_MACERR) || + (rxd_type & AQ_RXDESC_TYPE_DMA_ERR)) { + printf("%s:rx: rx error (status %x type %x)\n", + DEVNAME(sc), status, rxd_type); + m_freem(m); + } else { + m->m_pkthdr.len = m->m_len = pktlen; + ml_enqueue(&ml, m); + } + + idx++; + if (idx == AQ_RXD_NUM) + idx = 0; + } + rx->rx_cons = idx; + + if (rxfree > 0) { + if_rxr_put(&rx->rx_rxr, rxfree); + if (ifiq_input(rx->rx_ifiq, &ml)) + if_rxr_livelocked(&rx->rx_rxr); + + aq_rx_fill(sc, rx); + if (if_rxr_inuse(&rx->rx_rxr) == 0) + timeout_add(&rx->rx_refill, 1); + } +} + +void +aq_txeof(struct aq_softc *sc, struct aq_txring *tx) +{ + struct ifnet *ifp = &sc->sc_arpcom.ac_if; + struct aq_slot *as; + uint32_t idx, end, free; + + if (!ISSET(ifp->if_flags, IFF_RUNNING)) + return; + + idx = tx->tx_cons; + end = AQ_READ_REG(sc, TX_DMA_DESC_HEAD_PTR_REG(tx->tx_q)) & + TX_DMA_DESC_HEAD_PTR; + free = 0; + + bus_dmamap_sync(sc->sc_dmat, AQ_DMA_MAP(&tx->tx_mem), 0, + AQ_DMA_LEN(&tx->tx_mem), BUS_DMASYNC_POSTREAD); + + while (idx != end) { + as = &tx->tx_slots[idx]; + bus_dmamap_unload(sc->sc_dmat, as->as_map); + + m_freem(as->as_m); + as->as_m = NULL; + + idx++; + if (idx == AQ_TXD_NUM) + idx = 0; + free++; + } + + tx->tx_cons = idx; + + if (free != 0) { + if (ifq_is_oactive(tx->tx_ifq)) + ifq_restart(tx->tx_ifq); + } +} + +void +aq_start(struct ifqueue *ifq) +{ + struct aq_queues *aq = ifq->ifq_softc; + struct aq_softc *sc = aq->q_sc; + struct aq_txring *tx = &aq->q_tx; + struct aq_tx_desc *ring, *txd; + struct aq_slot *as; + struct mbuf *m; + uint32_t idx, free, used, ctl1, ctl2; + + idx = tx->tx_prod; + free = tx->tx_cons + AQ_TXD_NUM - tx->tx_prod; + used = 0; + + bus_dmamap_sync(sc->sc_dmat, AQ_DMA_MAP(&tx->tx_mem), 0, + AQ_DMA_LEN(&tx->tx_mem), BUS_DMASYNC_POSTWRITE); + ring = (struct aq_tx_desc *)AQ_DMA_KVA(&tx->tx_mem); + + for (;;) { + if (used + AQ_TX_MAX_SEGMENTS >= free) { + ifq_set_oactive(ifq); + break; + } + + m = ifq_dequeue(ifq); + if (m == NULL) + break; + + txd = ring + idx; + as = &tx->tx_slots[idx]; + + if (m_defrag(m, M_DONTWAIT) != 0) { + m_freem(m); + break; + } + + if (bus_dmamap_load_mbuf(sc->sc_dmat, as->as_map, m, + BUS_DMA_STREAMING | BUS_DMA_NOWAIT) != 0) { + m_freem(m); + break; + } + + as->as_m = m; + +#if NBPFILTER > 0 + if (ifp->if_bpf) + bpf_mtap_ether(ifp->if_bpf, m, BPF_DIRECTION_OUT); +#endif + bus_dmamap_sync(sc->sc_dmat, as->as_map, 0, + as->as_map->dm_mapsize, BUS_DMASYNC_PREWRITE); + + ctl1 = AQ_TXDESC_CTL1_TYPE_TXD | (as->as_map->dm_segs[0].ds_len << + AQ_TXDESC_CTL1_BLEN_SHIFT) | AQ_TXDESC_CTL1_CMD_FCS | + AQ_TXDESC_CTL1_CMD_EOP | AQ_TXDESC_CTL1_CMD_WB; + ctl2 = m->m_pkthdr.len << AQ_TXDESC_CTL2_LEN_SHIFT; + + txd->buf_addr = htole64(as->as_map->dm_segs[0].ds_addr); + txd->ctl1 = htole32(ctl1); + txd->ctl2 = htole32(ctl2); + + idx++; + if (idx == AQ_TXD_NUM) + idx = 0; + used++; + } + + bus_dmamap_sync(sc->sc_dmat, AQ_DMA_MAP(&tx->tx_mem), 0, + AQ_DMA_LEN(&tx->tx_mem), BUS_DMASYNC_PREWRITE); + + if (used != 0) { + tx->tx_prod = idx; + AQ_WRITE_REG(sc, TX_DMA_DESC_TAIL_PTR_REG(tx->tx_q), + tx->tx_prod); + } +} + +int +aq_intr(void *arg) +{ + struct aq_softc *sc = arg; + struct aq_queues *aq = &sc->sc_queues[0]; + uint32_t status; + + status = AQ_READ_REG(sc, AQ_INTR_STATUS_REG); + AQ_WRITE_REG(sc, AQ_INTR_STATUS_CLR_REG, 0xffffffff); + + if (status & (1 << sc->sc_linkstat_irq)) + aq_update_link_status(sc); + + if (status & (1 << aq->q_tx.tx_irq)) { + aq_txeof(sc, &aq->q_tx); + AQ_WRITE_REG(sc, AQ_INTR_STATUS_CLR_REG, + (1 << aq->q_tx.tx_irq)); + } + if (status & (1 << aq->q_rx.rx_irq)) { + aq_rxeof(sc, &aq->q_rx); + AQ_WRITE_REG(sc, AQ_INTR_STATUS_CLR_REG, + (1 << aq->q_rx.rx_irq)); + } + + return 1; +} + +void +aq_watchdog(struct ifnet *ifp) +{ + +} + +void +aq_free_slots(struct aq_softc *sc, struct aq_slot *slots, int allocated, + int total) +{ + struct aq_slot *as; + + int i = allocated; + while (i-- > 0) { + as = &slots[i]; + bus_dmamap_destroy(sc->sc_dmat, as->as_map); + if (as->as_m != NULL) + m_freem(as->as_m); + } + free(slots, M_DEVBUF, total * sizeof(*as)); +} + +int +aq_queue_up(struct aq_softc *sc, struct aq_queues *aq) +{ + struct aq_rxring *rx; + struct aq_txring *tx; + struct aq_slot *as; + int i; + + rx = &aq->q_rx; + rx->rx_slots = mallocarray(sizeof(*as), AQ_RXD_NUM, M_DEVBUF, + M_WAITOK | M_ZERO); + if (rx->rx_slots == NULL) { + printf("%s: failed to allocate rx slots %d\n", DEVNAME(sc), + aq->q_index); + return ENOMEM; + } + + for (i = 0; i < AQ_RXD_NUM; i++) { + as = &rx->rx_slots[i]; + if (bus_dmamap_create(sc->sc_dmat, MCLBYTES, 1, MCLBYTES, 0, + BUS_DMA_WAITOK | BUS_DMA_ALLOCNOW | BUS_DMA_64BIT, + &as->as_map) != 0) { + printf("%s: failed to allocate rx dma maps %d\n", + DEVNAME(sc), aq->q_index); + goto destroy_rx_slots; + } + } + + if (aq_dmamem_alloc(sc, &rx->rx_mem, AQ_RXD_NUM * + sizeof(struct aq_rx_desc_read), PAGE_SIZE) != 0) { + printf("%s: unable to allocate rx ring %d\n", DEVNAME(sc), + aq->q_index); + goto destroy_rx_slots; + } + + tx = &aq->q_tx; + tx->tx_slots = mallocarray(sizeof(*as), AQ_TXD_NUM, M_DEVBUF, + M_WAITOK | M_ZERO); + if (tx->tx_slots == NULL) { + printf("%s: failed to allocate tx slots %d\n", DEVNAME(sc), + aq->q_index); + goto destroy_rx_ring; + } + + for (i = 0; i < AQ_TXD_NUM; i++) { + as = &tx->tx_slots[i]; + if (bus_dmamap_create(sc->sc_dmat, MCLBYTES, + AQ_TX_MAX_SEGMENTS, MCLBYTES, 0, + BUS_DMA_WAITOK | BUS_DMA_ALLOCNOW | BUS_DMA_64BIT, + &as->as_map) != 0) { + printf("%s: failed to allocated tx dma maps %d\n", + DEVNAME(sc), aq->q_index); + goto destroy_tx_slots; + } + } + + if (aq_dmamem_alloc(sc, &tx->tx_mem, AQ_TXD_NUM * + sizeof(struct aq_tx_desc), PAGE_SIZE) != 0) { + printf("%s: unable to allocate tx ring %d\n", DEVNAME(sc), + aq->q_index); + goto destroy_tx_slots; + } + + aq_txring_reset(sc, tx, 1); + aq_rxring_reset(sc, rx, 1); + return 0; + +destroy_tx_slots: + aq_free_slots(sc, tx->tx_slots, i, AQ_TXD_NUM); + tx->tx_slots = NULL; + i = AQ_RXD_NUM; + +destroy_rx_ring: + aq_dmamem_free(sc, &rx->rx_mem); +destroy_rx_slots: + aq_free_slots(sc, rx->rx_slots, i, AQ_RXD_NUM); + rx->rx_slots = NULL; + return ENOMEM; +} + +void +aq_queue_down(struct aq_softc *sc, struct aq_queues *aq) +{ + struct aq_txring *tx; + struct aq_rxring *rx; + + tx = &aq->q_tx; + aq_txring_reset(sc, &aq->q_tx, 0); + if (tx->tx_slots != NULL) { + aq_free_slots(sc, tx->tx_slots, AQ_TXD_NUM, AQ_TXD_NUM); + tx->tx_slots = NULL; + } + + aq_dmamem_free(sc, &tx->tx_mem); + + rx = &aq->q_rx; + aq_rxring_reset(sc, &aq->q_rx, 0); + if (rx->rx_slots != NULL) { + aq_free_slots(sc, rx->rx_slots, AQ_RXD_NUM, AQ_RXD_NUM); + rx->rx_slots = NULL; + } + + aq_dmamem_free(sc, &rx->rx_mem); +} + +int +aq_up(struct aq_softc *sc) +{ + struct ifnet *ifp = &sc->sc_arpcom.ac_if; + int i; + + for (i = 0; i < sc->sc_nqueues; i++) { + if (aq_queue_up(sc, &sc->sc_queues[i]) != 0) + goto downqueues; + } + + /* filters? */ + /* enable checksum offload */ + + SET(ifp->if_flags, IFF_RUNNING); + aq_enable_intr(sc, 1, 1); + AQ_WRITE_REG_BIT(sc, TPB_TX_BUF_REG, TPB_TX_BUF_EN, 1); + AQ_WRITE_REG_BIT(sc, RPB_RPF_RX_REG, RPB_RPF_RX_BUF_EN, 1); + + for (i = 0; i < sc->sc_nqueues; i++) { + struct aq_queues *aq = &sc->sc_queues[i]; + + if_rxr_init(&aq->q_rx.rx_rxr, 1, AQ_RXD_NUM - 1); + aq_rx_fill(sc, &aq->q_rx); + + ifq_clr_oactive(aq->q_tx.tx_ifq); + } + + return ENETRESET; + +downqueues: + for (i = 0; i < sc->sc_nqueues; i++) + aq_queue_down(sc, &sc->sc_queues[i]); + return ENOMEM; +} + +void +aq_down(struct aq_softc *sc) +{ + struct ifnet *ifp = &sc->sc_arpcom.ac_if; + int i; + + CLR(ifp->if_flags, IFF_RUNNING); + + aq_enable_intr(sc, 1, 0); + intr_barrier(sc->sc_ih); + + for (i = 0; i < sc->sc_nqueues; i++) { + /* queue intr barrier? */ + aq_queue_down(sc, &sc->sc_queues[i]); + } +} + +void +aq_enable_intr(struct aq_softc *sc, int link, int txrx) +{ + uint32_t imask = 0; + int i; + + if (txrx) { + for (i = 0; i < sc->sc_nqueues; i++) { + imask |= (1 << sc->sc_queues[i].q_tx.tx_irq); + imask |= (1 << sc->sc_queues[i].q_rx.rx_irq); + } + } + + if (link) + imask |= (1 << sc->sc_linkstat_irq); + + AQ_WRITE_REG(sc, AQ_INTR_MASK_REG, imask); + AQ_WRITE_REG(sc, AQ_INTR_STATUS_CLR_REG, 0xffffffff); +} + +void +aq_ifmedia_status(struct ifnet *ifp, struct ifmediareq *ifmr) +{ + struct aq_softc *aq = ifp->if_softc; + enum aq_link_speed speed; + enum aq_link_fc fc; + int media; + int flow; + + if (aq_get_linkmode(aq, &speed, &fc, NULL) != 0) + return; + + switch (speed) { + case AQ_LINK_10G: + media = IFM_10G_T; + break; + case AQ_LINK_5G: + media = IFM_5000_T; + break; + case AQ_LINK_2G5: + media = IFM_2500_T; + break; + case AQ_LINK_1G: + media = IFM_1000_T; + break; + case AQ_LINK_100M: + media = IFM_100_TX; + break; + case AQ_LINK_NONE: + media = 0; + break; + } + + flow = 0; + if (fc & AQ_FC_RX) + flow |= IFM_ETH_RXPAUSE; + if (fc & AQ_FC_TX) + flow |= IFM_ETH_TXPAUSE; + + ifmr->ifm_status = IFM_AVALID; + if (speed != AQ_LINK_NONE) { + ifmr->ifm_status |= IFM_ACTIVE; + ifmr->ifm_active = IFM_ETHER | IFM_AUTO | media | flow; + } +} + +int +aq_ifmedia_change(struct ifnet *ifp) +{ + struct aq_softc *sc = ifp->if_softc; + enum aq_link_speed rate = AQ_LINK_NONE; + enum aq_link_fc fc = AQ_FC_NONE; + + if (IFM_TYPE(sc->sc_media.ifm_media) != IFM_ETHER) + return EINVAL; + + switch (IFM_SUBTYPE(sc->sc_media.ifm_media)) { + case IFM_AUTO: + rate = AQ_LINK_AUTO; + break; + case IFM_NONE: + rate = AQ_LINK_NONE; + break; + case IFM_100_TX: + rate = AQ_LINK_100M; + break; + case IFM_1000_T: + rate = AQ_LINK_1G; + break; + case IFM_2500_T: + rate = AQ_LINK_2G5; + break; + case IFM_5000_T: + rate = AQ_LINK_5G; + break; + case IFM_10G_T: + rate = AQ_LINK_10G; + break; + default: + return ENODEV; + } + + if (sc->sc_media.ifm_media & IFM_FLOW) + fc = AQ_FC_ALL; + + return aq_set_linkmode(sc, rate, fc, AQ_EEE_DISABLE); +} + +void +aq_update_link_status(struct aq_softc *sc) +{ + struct ifnet *ifp = &sc->sc_arpcom.ac_if; + enum aq_link_speed speed; + enum aq_link_fc fc; + + if (aq_get_linkmode(sc, &speed, &fc, NULL) != 0) + return; + + if (speed == AQ_LINK_NONE) { + if (ifp->if_link_state != LINK_STATE_DOWN) { + ifp->if_link_state = LINK_STATE_DOWN; + if_link_state_change(ifp); + } + } else { + if (ifp->if_link_state != LINK_STATE_FULL_DUPLEX) { + ifp->if_link_state = LINK_STATE_FULL_DUPLEX; + if_link_state_change(ifp); + } + } +} + + +int +aq_ioctl(struct ifnet *ifp, u_long cmd, caddr_t data) +{ + struct aq_softc *sc = ifp->if_softc; + struct ifreq *ifr = (struct ifreq *)data; + int error = 0, s; + + s = splnet(); + + switch (cmd) { + case SIOCSIFADDR: + ifp->if_flags |= IFF_UP; + if ((ifp->if_flags & IFF_RUNNING) == 0) + error = aq_up(sc); + break; + case SIOCSIFFLAGS: + if (ifp->if_flags & IFF_UP) { + if (ifp->if_flags & IFF_RUNNING) + error = ENETRESET; + else + error = aq_up(sc); + } else { + if (ifp->if_flags & IFF_RUNNING) + aq_down(sc); + } + break; + case SIOCSIFMEDIA: + case SIOCGIFMEDIA: + error = ifmedia_ioctl(ifp, ifr, &sc->sc_media, cmd); + break; + default: + error = ether_ioctl(ifp, &sc->sc_arpcom, cmd, data); + } + + if (error == ENETRESET) { + if (ifp->if_flags & IFF_RUNNING) + aq_iff(sc); + error = 0; + } + + splx(s); + return error; +} + +void +aq_iff(struct aq_softc *sc) +{ + struct ifnet *ifp = &sc->sc_arpcom.ac_if; + struct arpcom *ac = &sc->sc_arpcom; + struct ether_multi *enm; + struct ether_multistep step; + int idx; + + if (ifp->if_flags & IFF_PROMISC || ac->ac_multirangecnt > 0) { + ifp->if_flags |= IFF_ALLMULTI; + AQ_WRITE_REG_BIT(sc, RPF_L2BC_REG, RPF_L2BC_PROMISC, 1); + } else if (ac->ac_multicnt >= AQ_HW_MAC_NUM || + ISSET(ifp->if_flags, IFF_ALLMULTI)) { + AQ_WRITE_REG_BIT(sc, RPF_L2BC_REG, RPF_L2BC_PROMISC, 0); + AQ_WRITE_REG_BIT(sc, RPF_MCAST_FILTER_MASK_REG, + RPF_MCAST_FILTER_MASK_ALLMULTI, 1); + AQ_WRITE_REG_BIT(sc, RPF_MCAST_FILTER_REG(0), + RPF_MCAST_FILTER_EN, 1); + } else if (ac->ac_multicnt == 0) { + AQ_WRITE_REG_BIT(sc, RPF_L2BC_REG, RPF_L2BC_PROMISC, 0); + AQ_WRITE_REG_BIT(sc, RPF_MCAST_FILTER_REG(0), + RPF_MCAST_FILTER_EN, 0); + } else { + idx = AQ_HW_MAC_OWN + 1; + + /* turn on allmulti while we're rewriting? */ + AQ_WRITE_REG_BIT(sc, RPF_L2BC_REG, RPF_L2BC_PROMISC, 0); + + ETHER_FIRST_MULTI(step, ac, enm); + while (enm != NULL) { + aq_set_mac_addr(sc, idx++, enm->enm_addrlo); + ETHER_NEXT_MULTI(step, enm); + } + + for (; idx < AQ_HW_MAC_NUM; idx++) + aq_set_mac_addr(sc, idx, NULL); + + AQ_WRITE_REG_BIT(sc, RPF_MCAST_FILTER_MASK_REG, + RPF_MCAST_FILTER_MASK_ALLMULTI, 0); + AQ_WRITE_REG_BIT(sc, RPF_MCAST_FILTER_REG(0), + RPF_MCAST_FILTER_EN, 1); + } +} + +int +aq_dmamem_alloc(struct aq_softc *sc, struct aq_dmamem *aqm, + bus_size_t size, u_int align) +{ + aqm->aqm_size = size; + + if (bus_dmamap_create(sc->sc_dmat, aqm->aqm_size, 1, + aqm->aqm_size, 0, + BUS_DMA_WAITOK | BUS_DMA_ALLOCNOW | BUS_DMA_64BIT, + &aqm->aqm_map) != 0) + return (1); + if (bus_dmamem_alloc(sc->sc_dmat, aqm->aqm_size, + align, 0, &aqm->aqm_seg, 1, &aqm->aqm_nsegs, + BUS_DMA_WAITOK | BUS_DMA_ZERO) != 0) + goto destroy; + if (bus_dmamem_map(sc->sc_dmat, &aqm->aqm_seg, aqm->aqm_nsegs, + aqm->aqm_size, &aqm->aqm_kva, BUS_DMA_WAITOK) != 0) + goto free; + if (bus_dmamap_load(sc->sc_dmat, aqm->aqm_map, aqm->aqm_kva, + aqm->aqm_size, NULL, BUS_DMA_WAITOK) != 0) + goto unmap; + + return (0); +unmap: + bus_dmamem_unmap(sc->sc_dmat, aqm->aqm_kva, aqm->aqm_size); +free: + bus_dmamem_free(sc->sc_dmat, &aqm->aqm_seg, 1); +destroy: + bus_dmamap_destroy(sc->sc_dmat, aqm->aqm_map); + return (1); +} + +void +aq_dmamem_free(struct aq_softc *sc, struct aq_dmamem *aqm) +{ + bus_dmamap_unload(sc->sc_dmat, aqm->aqm_map); + bus_dmamem_unmap(sc->sc_dmat, aqm->aqm_kva, aqm->aqm_size); + bus_dmamem_free(sc->sc_dmat, &aqm->aqm_seg, 1); + bus_dmamap_destroy(sc->sc_dmat, aqm->aqm_map); +} |