mbuf: add rte prefix to offload flags
[dpdk.git] / drivers / net / e1000 / em_rxtx.c
index 3291b5e..a5dfe16 100644 (file)
@@ -18,6 +18,7 @@
 #include <rte_log.h>
 #include <rte_debug.h>
 #include <rte_pci.h>
+#include <rte_bus_pci.h>
 #include <rte_memory.h>
 #include <rte_memcpy.h>
 #include <rte_memzone.h>
@@ -31,7 +32,7 @@
 #include <rte_malloc.h>
 #include <rte_mbuf.h>
 #include <rte_ether.h>
-#include <rte_ethdev_driver.h>
+#include <ethdev_driver.h>
 #include <rte_prefetch.h>
 #include <rte_ip.h>
 #include <rte_udp.h>
 
 #define E1000_RXDCTL_GRAN      0x01000000 /* RXDCTL Granularity */
 
-#define E1000_TX_OFFLOAD_MASK ( \
-               PKT_TX_IP_CKSUM |       \
-               PKT_TX_L4_MASK |        \
-               PKT_TX_VLAN_PKT)
+#define E1000_TX_OFFLOAD_MASK (RTE_MBUF_F_TX_IPV6 |           \
+               RTE_MBUF_F_TX_IPV4 |           \
+               RTE_MBUF_F_TX_IP_CKSUM |       \
+               RTE_MBUF_F_TX_L4_MASK |        \
+               RTE_MBUF_F_TX_VLAN_PKT)
 
 #define E1000_TX_OFFLOAD_NOTSUP_MASK \
-               (PKT_TX_OFFLOAD_MASK ^ E1000_TX_OFFLOAD_MASK)
+               (RTE_MBUF_F_TX_OFFLOAD_MASK ^ E1000_TX_OFFLOAD_MASK)
+
+/* PCI offset for querying configuration status register */
+#define PCI_CFG_STATUS_REG                 0x06
+#define FLUSH_DESC_REQUIRED               0x100
+
 
 /**
  * Structure associated with each descriptor of the RX ring of a RX queue.
@@ -164,6 +171,7 @@ struct em_tx_queue {
        uint8_t                wthresh;  /**< Write-back threshold register. */
        struct em_ctx_info ctx_cache;
        /**< Hardware context history.*/
+       uint64_t               offloads; /**< offloads of DEV_TX_OFFLOAD_* */
 };
 
 #if 1
@@ -219,13 +227,13 @@ em_set_xmit_ctx(struct em_tx_queue* txq,
        /* setup IPCS* fields */
        ctx.lower_setup.ip_fields.ipcss = (uint8_t)l2len;
        ctx.lower_setup.ip_fields.ipcso = (uint8_t)(l2len +
-                       offsetof(struct ipv4_hdr, hdr_checksum));
+                       offsetof(struct rte_ipv4_hdr, hdr_checksum));
 
        /*
         * When doing checksum or TCP segmentation with IPv6 headers,
         * IPCSE field should be set t0 0.
         */
-       if (flags & PKT_TX_IP_CKSUM) {
+       if (flags & RTE_MBUF_F_TX_IP_CKSUM) {
                ctx.lower_setup.ip_fields.ipcse =
                        (uint16_t)rte_cpu_to_le_16(ipcse - 1);
                cmd_len |= E1000_TXD_CMD_IP;
@@ -238,15 +246,15 @@ em_set_xmit_ctx(struct em_tx_queue* txq,
        ctx.upper_setup.tcp_fields.tucss = (uint8_t)ipcse;
        ctx.upper_setup.tcp_fields.tucse = 0;
 
-       switch (flags & PKT_TX_L4_MASK) {
-       case PKT_TX_UDP_CKSUM:
+       switch (flags & RTE_MBUF_F_TX_L4_MASK) {
+       case RTE_MBUF_F_TX_UDP_CKSUM:
                ctx.upper_setup.tcp_fields.tucso = (uint8_t)(ipcse +
-                               offsetof(struct udp_hdr, dgram_cksum));
+                               offsetof(struct rte_udp_hdr, dgram_cksum));
                cmp_mask |= TX_MACIP_LEN_CMP_MASK;
                break;
-       case PKT_TX_TCP_CKSUM:
+       case RTE_MBUF_F_TX_TCP_CKSUM:
                ctx.upper_setup.tcp_fields.tucso = (uint8_t)(ipcse +
-                               offsetof(struct tcp_hdr, cksum));
+                               offsetof(struct rte_tcp_hdr, cksum));
                cmd_len |= E1000_TXD_CMD_TCP;
                cmp_mask |= TX_MACIP_LEN_CMP_MASK;
                break;
@@ -302,10 +310,10 @@ em_xmit_cleanup(struct em_tx_queue *txq)
        desc_to_clean_to = sw_ring[desc_to_clean_to].last_id;
        if (! (txr[desc_to_clean_to].upper.fields.status & E1000_TXD_STAT_DD))
        {
-               PMD_TX_FREE_LOG(DEBUG,
-                               "TX descriptor %4u is not done"
-                               "(port=%d queue=%d)", desc_to_clean_to,
-                               txq->port_id, txq->queue_id);
+               PMD_TX_LOG(DEBUG,
+                          "TX descriptor %4u is not done"
+                          "(port=%d queue=%d)", desc_to_clean_to,
+                          txq->port_id, txq->queue_id);
                /* Failed to clean any descriptors, better luck next time */
                return -(1);
        }
@@ -318,11 +326,11 @@ em_xmit_cleanup(struct em_tx_queue *txq)
                nb_tx_to_clean = (uint16_t)(desc_to_clean_to -
                                                last_desc_cleaned);
 
-       PMD_TX_FREE_LOG(DEBUG,
-                       "Cleaning %4u TX descriptors: %4u to %4u "
-                       "(port=%d queue=%d)", nb_tx_to_clean,
-                       last_desc_cleaned, desc_to_clean_to, txq->port_id,
-                       txq->queue_id);
+       PMD_TX_LOG(DEBUG,
+                  "Cleaning %4u TX descriptors: %4u to %4u "
+                  "(port=%d queue=%d)", nb_tx_to_clean,
+                  last_desc_cleaned, desc_to_clean_to, txq->port_id,
+                  txq->queue_id);
 
        /*
         * The last descriptor to clean is done, so that means all the
@@ -347,8 +355,8 @@ tx_desc_cksum_flags_to_upper(uint64_t ol_flags)
        static const uint32_t l3_olinfo[2] = {0, E1000_TXD_POPTS_IXSM << 8};
        uint32_t tmp;
 
-       tmp = l4_olinfo[(ol_flags & PKT_TX_L4_MASK) != PKT_TX_L4_NO_CKSUM];
-       tmp |= l3_olinfo[(ol_flags & PKT_TX_IP_CKSUM) != 0];
+       tmp = l4_olinfo[(ol_flags & RTE_MBUF_F_TX_L4_MASK) != RTE_MBUF_F_TX_L4_NO_CKSUM];
+       tmp |= l3_olinfo[(ol_flags & RTE_MBUF_F_TX_IP_CKSUM) != 0];
        return tmp;
 }
 
@@ -401,7 +409,7 @@ eth_em_xmit_pkts(void *tx_queue, struct rte_mbuf **tx_pkts,
                ol_flags = tx_pkt->ol_flags;
 
                /* If hardware offload required */
-               tx_ol_req = (ol_flags & (PKT_TX_IP_CKSUM | PKT_TX_L4_MASK));
+               tx_ol_req = (ol_flags & (RTE_MBUF_F_TX_IP_CKSUM | RTE_MBUF_F_TX_L4_MASK));
                if (tx_ol_req) {
                        hdrlen.f.vlan_tci = tx_pkt->vlan_tci;
                        hdrlen.f.l2_len = tx_pkt->l2_len;
@@ -448,11 +456,11 @@ eth_em_xmit_pkts(void *tx_queue, struct rte_mbuf **tx_pkts,
                 * nb_used better be less than or equal to txq->tx_rs_thresh
                 */
                while (unlikely (nb_used > txq->nb_tx_free)) {
-                       PMD_TX_FREE_LOG(DEBUG, "Not enough free TX descriptors "
-                                       "nb_used=%4u nb_free=%4u "
-                                       "(port=%d queue=%d)",
-                                       nb_used, txq->nb_tx_free,
-                                       txq->port_id, txq->queue_id);
+                       PMD_TX_LOG(DEBUG, "Not enough free TX descriptors "
+                                  "nb_used=%4u nb_free=%4u "
+                                  "(port=%d queue=%d)",
+                                  nb_used, txq->nb_tx_free,
+                                  txq->port_id, txq->queue_id);
 
                        if (em_xmit_cleanup(txq) != 0) {
                                /* Could not clean any descriptors */
@@ -497,7 +505,7 @@ eth_em_xmit_pkts(void *tx_queue, struct rte_mbuf **tx_pkts,
                popts_spec = 0;
 
                /* Set VLAN Tag offload fields. */
-               if (ol_flags & PKT_TX_VLAN_PKT) {
+               if (ol_flags & RTE_MBUF_F_TX_VLAN_PKT) {
                        cmd_type_len |= E1000_TXD_CMD_VLE;
                        popts_spec = tx_pkt->vlan_tci << E1000_TXD_VLAN_SHIFT;
                }
@@ -570,10 +578,10 @@ eth_em_xmit_pkts(void *tx_queue, struct rte_mbuf **tx_pkts,
 
                /* Set RS bit only on threshold packets' last descriptor */
                if (txq->nb_tx_used >= txq->tx_rs_thresh) {
-                       PMD_TX_FREE_LOG(DEBUG,
-                                       "Setting RS bit on TXD id=%4u "
-                                       "(port=%d queue=%d)",
-                                       tx_last, txq->port_id, txq->queue_id);
+                       PMD_TX_LOG(DEBUG,
+                                  "Setting RS bit on TXD id=%4u "
+                                  "(port=%d queue=%d)",
+                                  tx_last, txq->port_id, txq->queue_id);
 
                        cmd_type_len |= E1000_TXD_CMD_RS;
 
@@ -613,20 +621,20 @@ eth_em_prep_pkts(__rte_unused void *tx_queue, struct rte_mbuf **tx_pkts,
                m = tx_pkts[i];
 
                if (m->ol_flags & E1000_TX_OFFLOAD_NOTSUP_MASK) {
-                       rte_errno = -ENOTSUP;
+                       rte_errno = ENOTSUP;
                        return i;
                }
 
-#ifdef RTE_LIBRTE_ETHDEV_DEBUG
+#ifdef RTE_ETHDEV_DEBUG_TX
                ret = rte_validate_tx_offload(m);
                if (ret != 0) {
-                       rte_errno = ret;
+                       rte_errno = -ret;
                        return i;
                }
 #endif
                ret = rte_net_intel_cksum_prepare(m);
                if (ret != 0) {
-                       rte_errno = ret;
+                       rte_errno = -ret;
                        return i;
                }
        }
@@ -647,7 +655,7 @@ rx_desc_status_to_pkt_flags(uint32_t rx_status)
 
        /* Check if VLAN present */
        pkt_flags = ((rx_status & E1000_RXD_STAT_VP) ?
-               PKT_RX_VLAN | PKT_RX_VLAN_STRIPPED : 0);
+               RTE_MBUF_F_RX_VLAN | RTE_MBUF_F_RX_VLAN_STRIPPED : 0);
 
        return pkt_flags;
 }
@@ -658,9 +666,9 @@ rx_desc_error_to_pkt_flags(uint32_t rx_error)
        uint64_t pkt_flags = 0;
 
        if (rx_error & E1000_RXD_ERR_IPE)
-               pkt_flags |= PKT_RX_IP_CKSUM_BAD;
+               pkt_flags |= RTE_MBUF_F_RX_IP_CKSUM_BAD;
        if (rx_error & E1000_RXD_ERR_TCPE)
-               pkt_flags |= PKT_RX_L4_CKSUM_BAD;
+               pkt_flags |= RTE_MBUF_F_RX_L4_CKSUM_BAD;
        return pkt_flags;
 }
 
@@ -802,7 +810,7 @@ eth_em_recv_pkts(void *rx_queue, struct rte_mbuf **rx_pkts,
                rxm->ol_flags = rxm->ol_flags |
                                rx_desc_error_to_pkt_flags(rxd.errors);
 
-               /* Only valid if PKT_RX_VLAN set in pkt_flags */
+               /* Only valid if RTE_MBUF_F_RX_VLAN set in pkt_flags */
                rxm->vlan_tci = rte_le_to_cpu_16(rxd.special);
 
                /*
@@ -1002,17 +1010,17 @@ eth_em_recv_scattered_pkts(void *rx_queue, struct rte_mbuf **rx_pkts,
                 */
                rxm->next = NULL;
                if (unlikely(rxq->crc_len > 0)) {
-                       first_seg->pkt_len -= ETHER_CRC_LEN;
-                       if (data_len <= ETHER_CRC_LEN) {
+                       first_seg->pkt_len -= RTE_ETHER_CRC_LEN;
+                       if (data_len <= RTE_ETHER_CRC_LEN) {
                                rte_pktmbuf_free_seg(rxm);
                                first_seg->nb_segs--;
                                last_seg->data_len = (uint16_t)
                                        (last_seg->data_len -
-                                        (ETHER_CRC_LEN - data_len));
+                                        (RTE_ETHER_CRC_LEN - data_len));
                                last_seg->next = NULL;
                        } else
-                               rxm->data_len =
-                                       (uint16_t) (data_len - ETHER_CRC_LEN);
+                               rxm->data_len = (uint16_t)
+                                       (data_len - RTE_ETHER_CRC_LEN);
                }
 
                /*
@@ -1028,7 +1036,7 @@ eth_em_recv_scattered_pkts(void *rx_queue, struct rte_mbuf **rx_pkts,
                first_seg->ol_flags = first_seg->ol_flags |
                                        rx_desc_error_to_pkt_flags(rxd.errors);
 
-               /* Only valid if PKT_RX_VLAN set in pkt_flags */
+               /* Only valid if RTE_MBUF_F_RX_VLAN set in pkt_flags */
                rxm->vlan_tci = rte_le_to_cpu_16(rxd.special);
 
                /* Prefetch data of first segment, if configured to do so. */
@@ -1152,6 +1160,37 @@ em_reset_tx_queue(struct em_tx_queue *txq)
        memset((void*)&txq->ctx_cache, 0, sizeof (txq->ctx_cache));
 }
 
+uint64_t
+em_get_tx_port_offloads_capa(struct rte_eth_dev *dev)
+{
+       uint64_t tx_offload_capa;
+
+       RTE_SET_USED(dev);
+       tx_offload_capa =
+               DEV_TX_OFFLOAD_MULTI_SEGS  |
+               DEV_TX_OFFLOAD_VLAN_INSERT |
+               DEV_TX_OFFLOAD_IPV4_CKSUM  |
+               DEV_TX_OFFLOAD_UDP_CKSUM   |
+               DEV_TX_OFFLOAD_TCP_CKSUM;
+
+       return tx_offload_capa;
+}
+
+uint64_t
+em_get_tx_queue_offloads_capa(struct rte_eth_dev *dev)
+{
+       uint64_t tx_queue_offload_capa;
+
+       /*
+        * As only one Tx queue can be used, let per queue offloading
+        * capability be same to per port queue offloading capability
+        * for better convenience.
+        */
+       tx_queue_offload_capa = em_get_tx_port_offloads_capa(dev);
+
+       return tx_queue_offload_capa;
+}
+
 int
 eth_em_tx_queue_setup(struct rte_eth_dev *dev,
                         uint16_t queue_idx,
@@ -1164,9 +1203,12 @@ eth_em_tx_queue_setup(struct rte_eth_dev *dev,
        struct e1000_hw     *hw;
        uint32_t tsize;
        uint16_t tx_rs_thresh, tx_free_thresh;
+       uint64_t offloads;
 
        hw = E1000_DEV_PRIVATE_TO_HW(dev->data->dev_private);
 
+       offloads = tx_conf->offloads | dev->data->dev_conf.txmode.offloads;
+
        /*
         * Validate number of transmit descriptors.
         * It must not exceed hardware maximum, and must be multiple
@@ -1270,6 +1312,7 @@ eth_em_tx_queue_setup(struct rte_eth_dev *dev,
        em_reset_tx_queue(txq);
 
        dev->data->tx_queues[queue_idx] = txq;
+       txq->offloads = offloads;
        return 0;
 }
 
@@ -1328,9 +1371,9 @@ em_get_rx_port_offloads_capa(struct rte_eth_dev *dev)
                DEV_RX_OFFLOAD_IPV4_CKSUM  |
                DEV_RX_OFFLOAD_UDP_CKSUM   |
                DEV_RX_OFFLOAD_TCP_CKSUM   |
-               DEV_RX_OFFLOAD_CRC_STRIP   |
+               DEV_RX_OFFLOAD_KEEP_CRC    |
                DEV_RX_OFFLOAD_SCATTER;
-       if (max_rx_pktlen > ETHER_MAX_LEN)
+       if (max_rx_pktlen > RTE_ETHER_MAX_LEN)
                rx_offload_capa |= DEV_RX_OFFLOAD_JUMBO_FRAME;
 
        return rx_offload_capa;
@@ -1351,22 +1394,6 @@ em_get_rx_queue_offloads_capa(struct rte_eth_dev *dev)
        return rx_queue_offload_capa;
 }
 
-static int
-em_check_rx_queue_offloads(struct rte_eth_dev *dev, uint64_t requested)
-{
-       uint64_t port_offloads = dev->data->dev_conf.rxmode.offloads;
-       uint64_t queue_supported = em_get_rx_queue_offloads_capa(dev);
-       uint64_t port_supported = em_get_rx_port_offloads_capa(dev);
-
-       if ((requested & (queue_supported | port_supported)) != requested)
-               return 0;
-
-       if ((port_offloads ^ requested) & port_supported)
-               return 0;
-
-       return 1;
-}
-
 int
 eth_em_rx_queue_setup(struct rte_eth_dev *dev,
                uint16_t queue_idx,
@@ -1379,21 +1406,11 @@ eth_em_rx_queue_setup(struct rte_eth_dev *dev,
        struct em_rx_queue *rxq;
        struct e1000_hw     *hw;
        uint32_t rsize;
+       uint64_t offloads;
 
        hw = E1000_DEV_PRIVATE_TO_HW(dev->data->dev_private);
 
-       if (!em_check_rx_queue_offloads(dev, rx_conf->offloads)) {
-               PMD_INIT_LOG(ERR, "%p: Rx queue offloads 0x%" PRIx64
-                       " don't match port offloads 0x%" PRIx64
-                       " or supported port offloads 0x%" PRIx64
-                       " or supported queue offloads 0x%" PRIx64,
-                       (void *)dev,
-                       rx_conf->offloads,
-                       dev->data->dev_conf.rxmode.offloads,
-                       em_get_rx_port_offloads_capa(dev),
-                       em_get_rx_queue_offloads_capa(dev));
-               return -ENOTSUP;
-       }
+       offloads = rx_conf->offloads | dev->data->dev_conf.rxmode.offloads;
 
        /*
         * Validate number of receive descriptors.
@@ -1407,12 +1424,13 @@ eth_em_rx_queue_setup(struct rte_eth_dev *dev,
        }
 
        /*
-        * EM devices don't support drop_en functionality
+        * EM devices don't support drop_en functionality.
+        * It's an optimization that does nothing on single-queue devices,
+        * so just log the issue and carry on.
         */
        if (rx_conf->rx_drop_en) {
-               PMD_INIT_LOG(ERR, "drop_en functionality not supported by "
+               PMD_INIT_LOG(NOTICE, "drop_en functionality not supported by "
                             "device");
-               return -EINVAL;
        }
 
        /* Free memory prior to re-allocation if needed. */
@@ -1449,8 +1467,10 @@ eth_em_rx_queue_setup(struct rte_eth_dev *dev,
        rxq->rx_free_thresh = rx_conf->rx_free_thresh;
        rxq->queue_id = queue_idx;
        rxq->port_id = dev->data->port_id;
-       rxq->crc_len = (uint8_t)((dev->data->dev_conf.rxmode.offloads &
-               DEV_RX_OFFLOAD_CRC_STRIP) ? 0 : ETHER_CRC_LEN);
+       if (dev->data->dev_conf.rxmode.offloads & DEV_RX_OFFLOAD_KEEP_CRC)
+               rxq->crc_len = RTE_ETHER_CRC_LEN;
+       else
+               rxq->crc_len = 0;
 
        rxq->rdt_reg_addr = E1000_PCI_REG_ADDR(hw, E1000_RDT(queue_idx));
        rxq->rdh_reg_addr = E1000_PCI_REG_ADDR(hw, E1000_RDH(queue_idx));
@@ -1462,7 +1482,7 @@ eth_em_rx_queue_setup(struct rte_eth_dev *dev,
 
        dev->data->rx_queues[queue_idx] = rxq;
        em_reset_rx_queue(rxq);
-       rxq->offloads = rx_conf->offloads;
+       rxq->offloads = offloads;
 
        return 0;
 }
@@ -1590,12 +1610,14 @@ em_dev_free_queues(struct rte_eth_dev *dev)
        for (i = 0; i < dev->data->nb_rx_queues; i++) {
                eth_em_rx_queue_release(dev->data->rx_queues[i]);
                dev->data->rx_queues[i] = NULL;
+               rte_eth_dma_zone_free(dev, "rx_ring", i);
        }
        dev->data->nb_rx_queues = 0;
 
        for (i = 0; i < dev->data->nb_tx_queues; i++) {
                eth_em_tx_queue_release(dev->data->tx_queues[i]);
                dev->data->tx_queues[i] = NULL;
+               rte_eth_dma_zone_free(dev, "tx_ring", i);
        }
        dev->data->nb_tx_queues = 0;
 }
@@ -1783,9 +1805,10 @@ eth_em_rx_init(struct rte_eth_dev *dev)
                 * Reset crc_len in case it was changed after queue setup by a
                 *  call to configure
                 */
-               rxq->crc_len =
-                       (uint8_t)(dev->data->dev_conf.rxmode.offloads &
-                               DEV_RX_OFFLOAD_CRC_STRIP ? 0 : ETHER_CRC_LEN);
+               if (dev->data->dev_conf.rxmode.offloads & DEV_RX_OFFLOAD_KEEP_CRC)
+                       rxq->crc_len = RTE_ETHER_CRC_LEN;
+               else
+                       rxq->crc_len = 0;
 
                bus_addr = rxq->rx_ring_phys_addr;
                E1000_WRITE_REG(hw, E1000_RDLEN(i),
@@ -1816,7 +1839,7 @@ eth_em_rx_init(struct rte_eth_dev *dev)
                 * one buffer.
                 */
                if (rxmode->offloads & DEV_RX_OFFLOAD_JUMBO_FRAME ||
-                               rctl_bsize < ETHER_MAX_LEN) {
+                               rctl_bsize < RTE_ETHER_MAX_LEN) {
                        if (!dev->data->scattered_rx)
                                PMD_INIT_LOG(DEBUG, "forcing scatter mode");
                        dev->rx_pkt_burst =
@@ -1864,10 +1887,10 @@ eth_em_rx_init(struct rte_eth_dev *dev)
        }
 
        /* Setup the Receive Control Register. */
-       if (dev->data->dev_conf.rxmode.offloads & DEV_RX_OFFLOAD_CRC_STRIP)
-               rctl |= E1000_RCTL_SECRC; /* Strip Ethernet CRC. */
-       else
+       if (dev->data->dev_conf.rxmode.offloads & DEV_RX_OFFLOAD_KEEP_CRC)
                rctl &= ~E1000_RCTL_SECRC; /* Do not Strip Ethernet CRC. */
+       else
+               rctl |= E1000_RCTL_SECRC; /* Strip Ethernet CRC. */
 
        rctl &= ~(3 << E1000_RCTL_MO_SHIFT);
        rctl |= E1000_RCTL_EN | E1000_RCTL_BAM | E1000_RCTL_LBM_NO |
@@ -1948,6 +1971,22 @@ eth_em_tx_init(struct rte_eth_dev *dev)
        tctl |= (E1000_TCTL_PSP | E1000_TCTL_RTLC | E1000_TCTL_EN |
                 (E1000_COLLISION_THRESHOLD << E1000_CT_SHIFT));
 
+       /* SPT and CNP Si errata workaround to avoid data corruption */
+       if (hw->mac.type == e1000_pch_spt) {
+               uint32_t reg_val;
+               reg_val = E1000_READ_REG(hw, E1000_IOSFPC);
+               reg_val |= E1000_RCTL_RDMTS_HEX;
+               E1000_WRITE_REG(hw, E1000_IOSFPC, reg_val);
+
+               /* Dropping the number of outstanding requests from
+                * 3 to 2 in order to avoid a buffer overrun.
+                */
+               reg_val = E1000_READ_REG(hw, E1000_TARC(0));
+               reg_val &= ~E1000_TARC0_CB_MULTIQ_3_REQ;
+               reg_val |= E1000_TARC0_CB_MULTIQ_2_REQ;
+               E1000_WRITE_REG(hw, E1000_TARC(0), reg_val);
+       }
+
        /* This write will effectively turn on the transmit unit. */
        E1000_WRITE_REG(hw, E1000_TCTL, tctl);
 }
@@ -1982,4 +2021,121 @@ em_txq_info_get(struct rte_eth_dev *dev, uint16_t queue_id,
        qinfo->conf.tx_thresh.wthresh = txq->wthresh;
        qinfo->conf.tx_free_thresh = txq->tx_free_thresh;
        qinfo->conf.tx_rs_thresh = txq->tx_rs_thresh;
+       qinfo->conf.offloads = txq->offloads;
+}
+
+static void
+e1000_flush_tx_ring(struct rte_eth_dev *dev)
+{
+       struct e1000_hw *hw = E1000_DEV_PRIVATE_TO_HW(dev->data->dev_private);
+       volatile struct e1000_data_desc *tx_desc;
+       volatile uint32_t *tdt_reg_addr;
+       uint32_t tdt, tctl, txd_lower = E1000_TXD_CMD_IFCS;
+       uint16_t size = 512;
+       struct em_tx_queue *txq;
+       int i;
+
+       if (dev->data->tx_queues == NULL)
+               return;
+       tctl = E1000_READ_REG(hw, E1000_TCTL);
+       E1000_WRITE_REG(hw, E1000_TCTL, tctl | E1000_TCTL_EN);
+       for (i = 0; i < dev->data->nb_tx_queues &&
+               i < E1000_I219_MAX_TX_QUEUE_NUM; i++) {
+               txq = dev->data->tx_queues[i];
+               tdt = E1000_READ_REG(hw, E1000_TDT(i));
+               if (tdt != txq->tx_tail)
+                       return;
+               tx_desc = &txq->tx_ring[txq->tx_tail];
+               tx_desc->buffer_addr = rte_cpu_to_le_64(txq->tx_ring_phys_addr);
+               tx_desc->lower.data = rte_cpu_to_le_32(txd_lower | size);
+               tx_desc->upper.data = 0;
+
+               rte_io_wmb();
+               txq->tx_tail++;
+               if (txq->tx_tail == txq->nb_tx_desc)
+                       txq->tx_tail = 0;
+               tdt_reg_addr = E1000_PCI_REG_ADDR(hw, E1000_TDT(i));
+               E1000_PCI_REG_WRITE(tdt_reg_addr, txq->tx_tail);
+               usec_delay(250);
+       }
+}
+
+static void
+e1000_flush_rx_ring(struct rte_eth_dev *dev)
+{
+       uint32_t rctl, rxdctl;
+       struct e1000_hw *hw = E1000_DEV_PRIVATE_TO_HW(dev->data->dev_private);
+       int i;
+
+       rctl = E1000_READ_REG(hw, E1000_RCTL);
+       E1000_WRITE_REG(hw, E1000_RCTL, rctl & ~E1000_RCTL_EN);
+       E1000_WRITE_FLUSH(hw);
+       usec_delay(150);
+
+       for (i = 0; i < dev->data->nb_rx_queues &&
+               i < E1000_I219_MAX_RX_QUEUE_NUM; i++) {
+               rxdctl = E1000_READ_REG(hw, E1000_RXDCTL(i));
+               /* zero the lower 14 bits (prefetch and host thresholds) */
+               rxdctl &= 0xffffc000;
+
+               /* update thresholds: prefetch threshold to 31,
+                * host threshold to 1 and make sure the granularity
+                * is "descriptors" and not "cache lines"
+                */
+               rxdctl |= (0x1F | (1UL << 8) | E1000_RXDCTL_THRESH_UNIT_DESC);
+
+               E1000_WRITE_REG(hw, E1000_RXDCTL(i), rxdctl);
+       }
+       /* momentarily enable the RX ring for the changes to take effect */
+       E1000_WRITE_REG(hw, E1000_RCTL, rctl | E1000_RCTL_EN);
+       E1000_WRITE_FLUSH(hw);
+       usec_delay(150);
+       E1000_WRITE_REG(hw, E1000_RCTL, rctl & ~E1000_RCTL_EN);
+}
+
+/**
+ * em_flush_desc_rings - remove all descriptors from the descriptor rings
+ *
+ * In i219, the descriptor rings must be emptied before resetting/closing the
+ * HW. Failure to do this will cause the HW to enter a unit hang state which
+ * can only be released by PCI reset on the device
+ *
+ */
+
+void
+em_flush_desc_rings(struct rte_eth_dev *dev)
+{
+       uint32_t fextnvm11, tdlen;
+       struct e1000_hw *hw = E1000_DEV_PRIVATE_TO_HW(dev->data->dev_private);
+       struct rte_pci_device *pci_dev = RTE_ETH_DEV_TO_PCI(dev);
+       uint16_t pci_cfg_status = 0;
+       int ret;
+
+       fextnvm11 = E1000_READ_REG(hw, E1000_FEXTNVM11);
+       E1000_WRITE_REG(hw, E1000_FEXTNVM11,
+                       fextnvm11 | E1000_FEXTNVM11_DISABLE_MULR_FIX);
+       tdlen = E1000_READ_REG(hw, E1000_TDLEN(0));
+       ret = rte_pci_read_config(pci_dev, &pci_cfg_status,
+                  sizeof(pci_cfg_status), PCI_CFG_STATUS_REG);
+       if (ret < 0) {
+               PMD_DRV_LOG(ERR, "Failed to read PCI offset 0x%x",
+                           PCI_CFG_STATUS_REG);
+               return;
+       }
+
+       /* do nothing if we're not in faulty state, or if the queue is empty */
+       if ((pci_cfg_status & FLUSH_DESC_REQUIRED) && tdlen) {
+               /* flush desc ring */
+               e1000_flush_tx_ring(dev);
+               ret = rte_pci_read_config(pci_dev, &pci_cfg_status,
+                               sizeof(pci_cfg_status), PCI_CFG_STATUS_REG);
+               if (ret < 0) {
+                       PMD_DRV_LOG(ERR, "Failed to read PCI offset 0x%x",
+                                       PCI_CFG_STATUS_REG);
+                       return;
+               }
+
+               if (pci_cfg_status & FLUSH_DESC_REQUIRED)
+                       e1000_flush_rx_ring(dev);
+       }
 }