mbuf: add rte prefix to offload flags
[dpdk.git] / drivers / net / cxgbe / sge.c
index 4ea40d1..52c571d 100644 (file)
@@ -20,7 +20,6 @@
 #include <rte_log.h>
 #include <rte_debug.h>
 #include <rte_pci.h>
-#include <rte_atomic.h>
 #include <rte_branch_prediction.h>
 #include <rte_memory.h>
 #include <rte_memzone.h>
 #include <rte_eal.h>
 #include <rte_alarm.h>
 #include <rte_ether.h>
-#include <rte_ethdev_driver.h>
+#include <ethdev_driver.h>
 #include <rte_malloc.h>
 #include <rte_random.h>
 #include <rte_dev.h>
 
-#include "common.h"
-#include "t4_regs.h"
-#include "t4_msg.h"
+#include "base/common.h"
+#include "base/t4_regs.h"
+#include "base/t4_msg.h"
 #include "cxgbe.h"
 
 static inline void ship_tx_pkt_coalesce_wr(struct adapter *adap,
@@ -73,7 +72,7 @@ static inline unsigned int fl_mtu_bufsize(struct adapter *adapter,
 {
        struct sge *s = &adapter->sge;
 
-       return CXGBE_ALIGN(s->pktshift + ETHER_HDR_LEN + VLAN_HLEN + mtu,
+       return CXGBE_ALIGN(s->pktshift + RTE_ETHER_HDR_LEN + VLAN_HLEN + mtu,
                           s->fl_align);
 }
 
@@ -397,7 +396,8 @@ static unsigned int refill_fl_usembufs(struct adapter *adap, struct sge_fl *q,
 
                rte_mbuf_refcnt_set(mbuf, 1);
                mbuf->data_off =
-                       (uint16_t)(RTE_PTR_ALIGN((char *)mbuf->buf_addr +
+                       (uint16_t)((char *)
+                                  RTE_PTR_ALIGN((char *)mbuf->buf_addr +
                                                 RTE_PKTMBUF_HEADROOM,
                                                 adap->sge.fl_align) -
                                   (char *)mbuf->buf_addr);
@@ -539,7 +539,7 @@ static inline unsigned int flits_to_desc(unsigned int n)
  */
 static inline int is_eth_imm(const struct rte_mbuf *m)
 {
-       unsigned int hdrlen = (m->ol_flags & PKT_TX_TCP_SEG) ?
+       unsigned int hdrlen = (m->ol_flags & RTE_MBUF_F_TX_TCP_SEG) ?
                              sizeof(struct cpl_tx_pkt_lso_core) : 0;
 
        hdrlen += sizeof(struct cpl_tx_pkt);
@@ -749,12 +749,12 @@ static u64 hwcsum(enum chip_type chip, const struct rte_mbuf *m)
 {
        int csum_type;
 
-       if (m->ol_flags & PKT_TX_IP_CKSUM) {
-               switch (m->ol_flags & PKT_TX_L4_MASK) {
-               case PKT_TX_TCP_CKSUM:
+       if (m->ol_flags & RTE_MBUF_F_TX_IP_CKSUM) {
+               switch (m->ol_flags & RTE_MBUF_F_TX_L4_MASK) {
+               case RTE_MBUF_F_TX_TCP_CKSUM:
                        csum_type = TX_CSUM_TCPIP;
                        break;
-               case PKT_TX_UDP_CKSUM:
+               case RTE_MBUF_F_TX_UDP_CKSUM:
                        csum_type = TX_CSUM_UDPIP;
                        break;
                default:
@@ -1003,12 +1003,6 @@ static inline int tx_do_packet_coalesce(struct sge_eth_txq *txq,
        struct cpl_tx_pkt_core *cpl;
        struct tx_sw_desc *sd;
        unsigned int idx = q->coalesce.idx, len = mbuf->pkt_len;
-       unsigned int max_coal_pkt_num = is_pf4(adap) ? ETH_COALESCE_PKT_NUM :
-                                                      ETH_COALESCE_VF_PKT_NUM;
-
-#ifdef RTE_LIBRTE_CXGBE_TPUT
-       RTE_SET_USED(nb_pkts);
-#endif
 
        if (q->coalesce.type == 0) {
                mc = (struct ulp_txpkt *)q->coalesce.ptr;
@@ -1035,7 +1029,7 @@ static inline int tx_do_packet_coalesce(struct sge_eth_txq *txq,
        /* fill the cpl message, same as in t4_eth_xmit, this should be kept
         * similar to t4_eth_xmit
         */
-       if (mbuf->ol_flags & PKT_TX_IP_CKSUM) {
+       if (mbuf->ol_flags & RTE_MBUF_F_TX_IP_CKSUM) {
                cntrl = hwcsum(adap->params.chip, mbuf) |
                               F_TXPKT_IPCSUM_DIS;
                txq->stats.tx_cso++;
@@ -1043,7 +1037,7 @@ static inline int tx_do_packet_coalesce(struct sge_eth_txq *txq,
                cntrl = F_TXPKT_L4CSUM_DIS | F_TXPKT_IPCSUM_DIS;
        }
 
-       if (mbuf->ol_flags & PKT_TX_VLAN_PKT) {
+       if (mbuf->ol_flags & RTE_MBUF_F_TX_VLAN_PKT) {
                txq->stats.vlan_ins++;
                cntrl |= F_TXPKT_VLAN_VLD | V_TXPKT_VLAN(mbuf->vlan_tci);
        }
@@ -1081,13 +1075,15 @@ static inline int tx_do_packet_coalesce(struct sge_eth_txq *txq,
        sd->coalesce.sgl[idx & 1] = (struct ulptx_sgl *)(cpl + 1);
        sd->coalesce.idx = (idx & 1) + 1;
 
-       /* send the coaelsced work request if max reached */
-       if (++q->coalesce.idx == max_coal_pkt_num
-#ifndef RTE_LIBRTE_CXGBE_TPUT
-           || q->coalesce.idx >= nb_pkts
-#endif
-           )
+       /* Send the coalesced work request, only if max reached. However,
+        * if lower latency is preferred over throughput, then don't wait
+        * for coalescing the next Tx burst and send the packets now.
+        */
+       q->coalesce.idx++;
+       if (q->coalesce.idx == adap->params.max_tx_coalesce_num ||
+           (adap->devargs.tx_mode_latency && q->coalesce.idx >= nb_pkts))
                ship_tx_pkt_coalesce_wr(adap, txq);
+
        return 0;
 }
 
@@ -1127,24 +1123,25 @@ int t4_eth_xmit(struct sge_eth_txq *txq, struct rte_mbuf *mbuf,
         * The chip min packet length is 10 octets but play safe and reject
         * anything shorter than an Ethernet header.
         */
-       if (unlikely(m->pkt_len < ETHER_HDR_LEN)) {
+       if (unlikely(m->pkt_len < RTE_ETHER_HDR_LEN)) {
 out_free:
                rte_pktmbuf_free(m);
                return 0;
        }
 
-       if ((!(m->ol_flags & PKT_TX_TCP_SEG)) &&
+       if ((!(m->ol_flags & RTE_MBUF_F_TX_TCP_SEG)) &&
            (unlikely(m->pkt_len > max_pkt_len)))
                goto out_free;
 
-       pi = (struct port_info *)txq->data->dev_private;
+       pi = txq->data->dev_private;
        adap = pi->adapter;
 
        cntrl = F_TXPKT_L4CSUM_DIS | F_TXPKT_IPCSUM_DIS;
        /* align the end of coalesce WR to a 512 byte boundary */
        txq->q.coalesce.max = (8 - (txq->q.pidx & 7)) * 8;
 
-       if (!((m->ol_flags & PKT_TX_TCP_SEG) || (m->pkt_len > ETHER_MAX_LEN))) {
+       if (!((m->ol_flags & RTE_MBUF_F_TX_TCP_SEG) ||
+                       m->pkt_len > RTE_ETHER_MAX_LEN)) {
                if (should_tx_packet_coalesce(txq, mbuf, &cflits, adap)) {
                        if (unlikely(map_mbuf(mbuf, addr) < 0)) {
                                dev_warn(adap, "%s: mapping err for coalesce\n",
@@ -1152,7 +1149,6 @@ out_free:
                                txq->stats.mapping_err++;
                                goto out_free;
                        }
-                       rte_prefetch0((volatile void *)addr);
                        return tx_do_packet_coalesce(txq, mbuf, cflits, adap,
                                                     pi, addr, nb_pkts);
                } else {
@@ -1207,7 +1203,7 @@ out_free:
        len += sizeof(*cpl);
 
        /* Coalescing skipped and we send through normal path */
-       if (!(m->ol_flags & PKT_TX_TCP_SEG)) {
+       if (!(m->ol_flags & RTE_MBUF_F_TX_TCP_SEG)) {
                wr->op_immdlen = htonl(V_FW_WR_OP(is_pf4(adap) ?
                                                  FW_ETH_TX_PKT_WR :
                                                  FW_ETH_TX_PKT_VM_WR) |
@@ -1216,7 +1212,7 @@ out_free:
                        cpl = (void *)(wr + 1);
                else
                        cpl = (void *)(vmwr + 1);
-               if (m->ol_flags & PKT_TX_IP_CKSUM) {
+               if (m->ol_flags & RTE_MBUF_F_TX_IP_CKSUM) {
                        cntrl = hwcsum(adap->params.chip, m) |
                                F_TXPKT_IPCSUM_DIS;
                        txq->stats.tx_cso++;
@@ -1226,10 +1222,10 @@ out_free:
                        lso = (void *)(wr + 1);
                else
                        lso = (void *)(vmwr + 1);
-               v6 = (m->ol_flags & PKT_TX_IPV6) != 0;
+               v6 = (m->ol_flags & RTE_MBUF_F_TX_IPV6) != 0;
                l3hdr_len = m->l3_len;
                l4hdr_len = m->l4_len;
-               eth_xtra_len = m->l2_len - ETHER_HDR_LEN;
+               eth_xtra_len = m->l2_len - RTE_ETHER_HDR_LEN;
                len += sizeof(*lso);
                wr->op_immdlen = htonl(V_FW_WR_OP(is_pf4(adap) ?
                                                  FW_ETH_TX_PKT_WR :
@@ -1262,7 +1258,7 @@ out_free:
                txq->stats.tx_cso += m->tso_segsz;
        }
 
-       if (m->ol_flags & PKT_TX_VLAN_PKT) {
+       if (m->ol_flags & RTE_MBUF_F_TX_VLAN_PKT) {
                txq->stats.vlan_ins++;
                cntrl |= F_TXPKT_VLAN_VLD | V_TXPKT_VLAN(m->vlan_tci);
        }
@@ -1426,14 +1422,16 @@ int t4_mgmt_tx(struct sge_ctrl_txq *q, struct rte_mbuf *mbuf)
 
 /**
  * alloc_ring - allocate resources for an SGE descriptor ring
- * @dev: the PCI device's core device
+ * @dev: the port associated with the queue
+ * @z_name: memzone's name
+ * @queue_id: queue index
+ * @socket_id: preferred socket id for memory allocations
  * @nelem: the number of descriptors
  * @elem_size: the size of each descriptor
+ * @stat_size: extra space in HW ring for status information
  * @sw_size: the size of the SW state associated with each ring element
  * @phys: the physical address of the allocated ring
  * @metadata: address of the array holding the SW state for the ring
- * @stat_size: extra space in HW ring for status information
- * @node: preferred node for memory allocations
  *
  * Allocates resources for an SGE descriptor ring, such as Tx queues,
  * free buffer lists, or response queues.  Each SGE ring requires
@@ -1443,39 +1441,34 @@ int t4_mgmt_tx(struct sge_ctrl_txq *q, struct rte_mbuf *mbuf)
  * of the function), the bus address of the HW ring, and the address
  * of the SW ring.
  */
-static void *alloc_ring(size_t nelem, size_t elem_size,
-                       size_t sw_size, dma_addr_t *phys, void *metadata,
-                       size_t stat_size, __rte_unused uint16_t queue_id,
-                       int socket_id, const char *z_name,
-                       const char *z_name_sw)
+static void *alloc_ring(struct rte_eth_dev *dev, const char *z_name,
+                       uint16_t queue_id, int socket_id, size_t nelem,
+                       size_t elem_size, size_t stat_size, size_t sw_size,
+                       dma_addr_t *phys, void *metadata)
 {
        size_t len = CXGBE_MAX_RING_DESC_SIZE * elem_size + stat_size;
+       char z_name_sw[RTE_MEMZONE_NAMESIZE];
        const struct rte_memzone *tz;
        void *s = NULL;
 
+       snprintf(z_name_sw, sizeof(z_name_sw), "eth_p%d_q%d_%s_sw_ring",
+                dev->data->port_id, queue_id, z_name);
+
        dev_debug(adapter, "%s: nelem = %zu; elem_size = %zu; sw_size = %zu; "
                  "stat_size = %zu; queue_id = %u; socket_id = %d; z_name = %s;"
                  " z_name_sw = %s\n", __func__, nelem, elem_size, sw_size,
                  stat_size, queue_id, socket_id, z_name, z_name_sw);
 
-       tz = rte_memzone_lookup(z_name);
-       if (tz) {
-               dev_debug(adapter, "%s: tz exists...returning existing..\n",
-                         __func__);
-               goto alloc_sw_ring;
-       }
-
        /*
         * Allocate TX/RX ring hardware descriptors. A memzone large enough to
         * handle the maximum ring size is allocated in order to allow for
         * resizing in later calls to the queue setup function.
         */
-       tz = rte_memzone_reserve_aligned(z_name, len, socket_id,
-                       RTE_MEMZONE_IOVA_CONTIG, 4096);
+       tz = rte_eth_dma_zone_reserve(dev, z_name, queue_id, len, 4096,
+                                     socket_id);
        if (!tz)
                return NULL;
 
-alloc_sw_ring:
        memset(tz->addr, 0, len);
        if (sw_size) {
                s = rte_zmalloc_socket(z_name_sw, nelem * sw_size,
@@ -1494,98 +1487,6 @@ alloc_sw_ring:
        return tz->addr;
 }
 
-/**
- * t4_pktgl_to_mbuf_usembufs - build an mbuf from a packet gather list
- * @gl: the gather list
- *
- * Builds an mbuf from the given packet gather list.  Returns the mbuf or
- * %NULL if mbuf allocation failed.
- */
-static struct rte_mbuf *t4_pktgl_to_mbuf_usembufs(const struct pkt_gl *gl)
-{
-       /*
-        * If there's only one mbuf fragment, just return that.
-        */
-       if (likely(gl->nfrags == 1))
-               return gl->mbufs[0];
-
-       return NULL;
-}
-
-/**
- * t4_pktgl_to_mbuf - build an mbuf from a packet gather list
- * @gl: the gather list
- *
- * Builds an mbuf from the given packet gather list.  Returns the mbuf or
- * %NULL if mbuf allocation failed.
- */
-static struct rte_mbuf *t4_pktgl_to_mbuf(const struct pkt_gl *gl)
-{
-       return t4_pktgl_to_mbuf_usembufs(gl);
-}
-
-/**
- * t4_ethrx_handler - process an ingress ethernet packet
- * @q: the response queue that received the packet
- * @rsp: the response queue descriptor holding the RX_PKT message
- * @si: the gather list of packet fragments
- *
- * Process an ingress ethernet packet and deliver it to the stack.
- */
-int t4_ethrx_handler(struct sge_rspq *q, const __be64 *rsp,
-                    const struct pkt_gl *si)
-{
-       struct rte_mbuf *mbuf;
-       const struct cpl_rx_pkt *pkt;
-       const struct rss_header *rss_hdr;
-       bool csum_ok;
-       struct sge_eth_rxq *rxq = container_of(q, struct sge_eth_rxq, rspq);
-       u16 err_vec;
-
-       rss_hdr = (const void *)rsp;
-       pkt = (const void *)&rsp[1];
-       /* Compressed error vector is enabled for T6 only */
-       if (q->adapter->params.tp.rx_pkt_encap)
-               err_vec = G_T6_COMPR_RXERR_VEC(ntohs(pkt->err_vec));
-       else
-               err_vec = ntohs(pkt->err_vec);
-       csum_ok = pkt->csum_calc && !err_vec;
-
-       mbuf = t4_pktgl_to_mbuf(si);
-       if (unlikely(!mbuf)) {
-               rxq->stats.rx_drops++;
-               return 0;
-       }
-
-       mbuf->port = pkt->iff;
-       if (pkt->l2info & htonl(F_RXF_IP)) {
-               mbuf->packet_type = RTE_PTYPE_L3_IPV4;
-               if (unlikely(!csum_ok))
-                       mbuf->ol_flags |= PKT_RX_IP_CKSUM_BAD;
-
-               if ((pkt->l2info & htonl(F_RXF_UDP | F_RXF_TCP)) && !csum_ok)
-                       mbuf->ol_flags |= PKT_RX_L4_CKSUM_BAD;
-       } else if (pkt->l2info & htonl(F_RXF_IP6)) {
-               mbuf->packet_type = RTE_PTYPE_L3_IPV6;
-       }
-
-       mbuf->port = pkt->iff;
-
-       if (!rss_hdr->filter_tid && rss_hdr->hash_type) {
-               mbuf->ol_flags |= PKT_RX_RSS_HASH;
-               mbuf->hash.rss = ntohl(rss_hdr->hash_val);
-       }
-
-       if (pkt->vlan_ex) {
-               mbuf->ol_flags |= PKT_RX_VLAN;
-               mbuf->vlan_tci = ntohs(pkt->vlan);
-       }
-       rxq->stats.pkts++;
-       rxq->stats.rx_bytes += mbuf->pkt_len;
-
-       return 0;
-}
-
 #define CXGB4_MSG_AN ((void *)1)
 
 /**
@@ -1604,6 +1505,52 @@ static inline void rspq_next(struct sge_rspq *q)
        }
 }
 
+static inline void cxgbe_set_mbuf_info(struct rte_mbuf *pkt, uint32_t ptype,
+                                      uint64_t ol_flags)
+{
+       pkt->packet_type |= ptype;
+       pkt->ol_flags |= ol_flags;
+}
+
+static inline void cxgbe_fill_mbuf_info(struct adapter *adap,
+                                       const struct cpl_rx_pkt *cpl,
+                                       struct rte_mbuf *pkt)
+{
+       bool csum_ok;
+       u16 err_vec;
+
+       if (adap->params.tp.rx_pkt_encap)
+               err_vec = G_T6_COMPR_RXERR_VEC(ntohs(cpl->err_vec));
+       else
+               err_vec = ntohs(cpl->err_vec);
+
+       csum_ok = cpl->csum_calc && !err_vec;
+
+       if (cpl->vlan_ex)
+               cxgbe_set_mbuf_info(pkt, RTE_PTYPE_L2_ETHER_VLAN,
+                                   RTE_MBUF_F_RX_VLAN | RTE_MBUF_F_RX_VLAN_STRIPPED);
+       else
+               cxgbe_set_mbuf_info(pkt, RTE_PTYPE_L2_ETHER, 0);
+
+       if (cpl->l2info & htonl(F_RXF_IP))
+               cxgbe_set_mbuf_info(pkt, RTE_PTYPE_L3_IPV4,
+                                   csum_ok ? RTE_MBUF_F_RX_IP_CKSUM_GOOD :
+                                   RTE_MBUF_F_RX_IP_CKSUM_BAD);
+       else if (cpl->l2info & htonl(F_RXF_IP6))
+               cxgbe_set_mbuf_info(pkt, RTE_PTYPE_L3_IPV6,
+                                   csum_ok ? RTE_MBUF_F_RX_IP_CKSUM_GOOD :
+                                   RTE_MBUF_F_RX_IP_CKSUM_BAD);
+
+       if (cpl->l2info & htonl(F_RXF_TCP))
+               cxgbe_set_mbuf_info(pkt, RTE_PTYPE_L4_TCP,
+                                   csum_ok ? RTE_MBUF_F_RX_L4_CKSUM_GOOD :
+                                   RTE_MBUF_F_RX_L4_CKSUM_BAD);
+       else if (cpl->l2info & htonl(F_RXF_UDP))
+               cxgbe_set_mbuf_info(pkt, RTE_PTYPE_L4_UDP,
+                                   csum_ok ? RTE_MBUF_F_RX_L4_CKSUM_GOOD :
+                                   RTE_MBUF_F_RX_L4_CKSUM_BAD);
+}
+
 /**
  * process_responses - process responses from an SGE response queue
  * @q: the ingress queue to process
@@ -1655,8 +1602,6 @@ static int process_responses(struct sge_rspq *q, int budget,
                                        (const void *)&q->cur_desc[1];
                                struct rte_mbuf *pkt, *npkt;
                                u32 len, bufsz;
-                               bool csum_ok;
-                               u16 err_vec;
 
                                rc = (const struct rsp_ctrl *)
                                     ((const char *)q->cur_desc +
@@ -1673,16 +1618,6 @@ static int process_responses(struct sge_rspq *q, int budget,
                                len = G_RSPD_LEN(len);
                                pkt->pkt_len = len;
 
-                               /* Compressed error vector is enabled for
-                                * T6 only
-                                */
-                               if (q->adapter->params.tp.rx_pkt_encap)
-                                       err_vec = G_T6_COMPR_RXERR_VEC(
-                                                       ntohs(cpl->err_vec));
-                               else
-                                       err_vec = ntohs(cpl->err_vec);
-                               csum_ok = cpl->csum_calc && !err_vec;
-
                                /* Chain mbufs into len if necessary */
                                while (len) {
                                        struct rte_mbuf *new_pkt = rsd->buf;
@@ -1700,33 +1635,17 @@ static int process_responses(struct sge_rspq *q, int budget,
                                npkt->next = NULL;
                                pkt->nb_segs--;
 
-                               if (cpl->l2info & htonl(F_RXF_IP)) {
-                                       pkt->packet_type = RTE_PTYPE_L3_IPV4;
-                                       if (unlikely(!csum_ok))
-                                               pkt->ol_flags |=
-                                                       PKT_RX_IP_CKSUM_BAD;
-
-                                       if ((cpl->l2info &
-                                            htonl(F_RXF_UDP | F_RXF_TCP)) &&
-                                           !csum_ok)
-                                               pkt->ol_flags |=
-                                                       PKT_RX_L4_CKSUM_BAD;
-                               } else if (cpl->l2info & htonl(F_RXF_IP6)) {
-                                       pkt->packet_type = RTE_PTYPE_L3_IPV6;
-                               }
+                               cxgbe_fill_mbuf_info(q->adapter, cpl, pkt);
 
                                if (!rss_hdr->filter_tid &&
                                    rss_hdr->hash_type) {
-                                       pkt->ol_flags |= PKT_RX_RSS_HASH;
+                                       pkt->ol_flags |= RTE_MBUF_F_RX_RSS_HASH;
                                        pkt->hash.rss =
                                                ntohl(rss_hdr->hash_val);
                                }
 
-                               if (cpl->vlan_ex) {
-                                       pkt->ol_flags |= PKT_RX_VLAN |
-                                                        PKT_RX_VLAN_STRIPPED;
+                               if (cpl->vlan_ex)
                                        pkt->vlan_tci = ntohs(cpl->vlan);
-                               }
 
                                rte_pktmbuf_adj(pkt, s->pktshift);
                                rxq->stats.pkts++;
@@ -1774,6 +1693,11 @@ int cxgbe_poll(struct sge_rspq *q, struct rte_mbuf **rx_pkts,
        unsigned int params;
        u32 val;
 
+       if (unlikely(rxq->flags & IQ_STOPPED)) {
+               *work_done = 0;
+               return 0;
+       }
+
        *work_done = process_responses(q, budget, rx_pkts);
 
        if (*work_done) {
@@ -1834,22 +1758,22 @@ static void __iomem *bar2_address(struct adapter *adapter, unsigned int qid,
        return adapter->bar2 + bar2_qoffset;
 }
 
-int t4_sge_eth_rxq_start(struct adapter *adap, struct sge_rspq *rq)
+int t4_sge_eth_rxq_start(struct adapter *adap, struct sge_eth_rxq *rxq)
 {
-       struct sge_eth_rxq *rxq = container_of(rq, struct sge_eth_rxq, rspq);
        unsigned int fl_id = rxq->fl.size ? rxq->fl.cntxt_id : 0xffff;
 
+       rxq->flags &= ~IQ_STOPPED;
        return t4_iq_start_stop(adap, adap->mbox, true, adap->pf, 0,
-                               rq->cntxt_id, fl_id, 0xffff);
+                               rxq->rspq.cntxt_id, fl_id, 0xffff);
 }
 
-int t4_sge_eth_rxq_stop(struct adapter *adap, struct sge_rspq *rq)
+int t4_sge_eth_rxq_stop(struct adapter *adap, struct sge_eth_rxq *rxq)
 {
-       struct sge_eth_rxq *rxq = container_of(rq, struct sge_eth_rxq, rspq);
        unsigned int fl_id = rxq->fl.size ? rxq->fl.cntxt_id : 0xffff;
 
+       rxq->flags |= IQ_STOPPED;
        return t4_iq_start_stop(adap, adap->mbox, false, adap->pf, 0,
-                               rq->cntxt_id, fl_id, 0xffff);
+                               rxq->rspq.cntxt_id, fl_id, 0xffff);
 }
 
 /*
@@ -1864,23 +1788,16 @@ int t4_sge_alloc_rxq(struct adapter *adap, struct sge_rspq *iq, bool fwevtq,
        int ret, flsz = 0;
        struct fw_iq_cmd c;
        struct sge *s = &adap->sge;
-       struct port_info *pi = (struct port_info *)(eth_dev->data->dev_private);
-       char z_name[RTE_MEMZONE_NAMESIZE];
-       char z_name_sw[RTE_MEMZONE_NAMESIZE];
+       struct port_info *pi = eth_dev->data->dev_private;
        unsigned int nb_refill;
        u8 pciechan;
 
        /* Size needs to be multiple of 16, including status entry. */
        iq->size = cxgbe_roundup(iq->size, 16);
 
-       snprintf(z_name, sizeof(z_name), "%s_%s_%d_%d",
-                eth_dev->device->driver->name,
-                fwevtq ? "fwq_ring" : "rx_ring",
-                eth_dev->data->port_id, queue_id);
-       snprintf(z_name_sw, sizeof(z_name_sw), "%s_sw_ring", z_name);
-
-       iq->desc = alloc_ring(iq->size, iq->iqe_len, 0, &iq->phys_addr, NULL, 0,
-                             queue_id, socket_id, z_name, z_name_sw);
+       iq->desc = alloc_ring(eth_dev, fwevtq ? "fwq_ring" : "rx_ring",
+                             queue_id, socket_id, iq->size, iq->iqe_len,
+                             0, 0, &iq->phys_addr, NULL);
        if (!iq->desc)
                return -ENOMEM;
 
@@ -1938,19 +1855,14 @@ int t4_sge_alloc_rxq(struct adapter *adap, struct sge_rspq *iq, bool fwevtq,
                        fl->size = s->fl_starve_thres - 1 + 2 * 8;
                fl->size = cxgbe_roundup(fl->size, 8);
 
-               snprintf(z_name, sizeof(z_name), "%s_%s_%d_%d",
-                        eth_dev->device->driver->name,
-                        fwevtq ? "fwq_ring" : "fl_ring",
-                        eth_dev->data->port_id, queue_id);
-               snprintf(z_name_sw, sizeof(z_name_sw), "%s_sw_ring", z_name);
-
-               fl->desc = alloc_ring(fl->size, sizeof(__be64),
+               fl->desc = alloc_ring(eth_dev, "fl_ring", queue_id, socket_id,
+                                     fl->size, sizeof(__be64), s->stat_len,
                                      sizeof(struct rx_sw_desc),
-                                     &fl->addr, &fl->sdesc, s->stat_len,
-                                     queue_id, socket_id, z_name, z_name_sw);
-
-               if (!fl->desc)
-                       goto fl_nomem;
+                                     &fl->addr, &fl->sdesc);
+               if (!fl->desc) {
+                       ret = -ENOMEM;
+                       goto err;
+               }
 
                flsz = fl->size / 8 + s->stat_len / sizeof(struct tx_desc);
                c.iqns_to_fl0congen |=
@@ -2041,7 +1953,8 @@ int t4_sge_alloc_rxq(struct adapter *adap, struct sge_rspq *iq, bool fwevtq,
         * simple (and hopefully less wrong).
         */
        if (is_pf4(adap) && !is_t4(adap->params.chip) && cong >= 0) {
-               u32 param, val;
+               u8 cng_ch_bits_log = adap->params.arch.cng_ch_bits_log;
+               u32 param, val, ch_map = 0;
                int i;
 
                param = (V_FW_PARAMS_MNEM(FW_PARAMS_MNEM_DMAQ) |
@@ -2054,9 +1967,9 @@ int t4_sge_alloc_rxq(struct adapter *adap, struct sge_rspq *iq, bool fwevtq,
                                        X_CONMCTXT_CNGTPMODE_CHANNEL);
                        for (i = 0; i < 4; i++) {
                                if (cong & (1 << i))
-                                       val |= V_CONMCTXT_CNGCHMAP(1 <<
-                                                                  (i << 2));
+                                       ch_map |= 1 << (i << cng_ch_bits_log);
                        }
+                       val |= V_CONMCTXT_CNGCHMAP(ch_map);
                }
                ret = t4_set_params(adap, adap->mbox, adap->pf, 0, 1,
                                    &param, &val);
@@ -2070,8 +1983,6 @@ int t4_sge_alloc_rxq(struct adapter *adap, struct sge_rspq *iq, bool fwevtq,
 refill_fl_err:
        t4_iq_free(adap, adap->mbox, adap->pf, 0, FW_IQ_TYPE_FL_INT_CAP,
                   iq->cntxt_id, fl->cntxt_id, 0xffff);
-fl_nomem:
-       ret = -ENOMEM;
 err:
        iq->cntxt_id = 0;
        iq->abs_id = 0;
@@ -2136,23 +2047,16 @@ int t4_sge_alloc_eth_txq(struct adapter *adap, struct sge_eth_txq *txq,
        int ret, nentries;
        struct fw_eq_eth_cmd c;
        struct sge *s = &adap->sge;
-       struct port_info *pi = (struct port_info *)(eth_dev->data->dev_private);
-       char z_name[RTE_MEMZONE_NAMESIZE];
-       char z_name_sw[RTE_MEMZONE_NAMESIZE];
+       struct port_info *pi = eth_dev->data->dev_private;
        u8 pciechan;
 
        /* Add status entries */
        nentries = txq->q.size + s->stat_len / sizeof(struct tx_desc);
 
-       snprintf(z_name, sizeof(z_name), "%s_%s_%d_%d",
-                eth_dev->device->driver->name, "tx_ring",
-                eth_dev->data->port_id, queue_id);
-       snprintf(z_name_sw, sizeof(z_name_sw), "%s_sw_ring", z_name);
-
-       txq->q.desc = alloc_ring(txq->q.size, sizeof(struct tx_desc),
-                                sizeof(struct tx_sw_desc), &txq->q.phys_addr,
-                                &txq->q.sdesc, s->stat_len, queue_id,
-                                socket_id, z_name, z_name_sw);
+       txq->q.desc = alloc_ring(eth_dev, "tx_ring", queue_id, socket_id,
+                                txq->q.size, sizeof(struct tx_desc),
+                                s->stat_len, sizeof(struct tx_sw_desc),
+                                &txq->q.phys_addr, &txq->q.sdesc);
        if (!txq->q.desc)
                return -ENOMEM;
 
@@ -2216,22 +2120,14 @@ int t4_sge_alloc_ctrl_txq(struct adapter *adap, struct sge_ctrl_txq *txq,
        int ret, nentries;
        struct fw_eq_ctrl_cmd c;
        struct sge *s = &adap->sge;
-       struct port_info *pi = (struct port_info *)(eth_dev->data->dev_private);
-       char z_name[RTE_MEMZONE_NAMESIZE];
-       char z_name_sw[RTE_MEMZONE_NAMESIZE];
+       struct port_info *pi = eth_dev->data->dev_private;
 
        /* Add status entries */
        nentries = txq->q.size + s->stat_len / sizeof(struct tx_desc);
 
-       snprintf(z_name, sizeof(z_name), "%s_%s_%d_%d",
-                eth_dev->device->driver->name, "ctrl_tx_ring",
-                eth_dev->data->port_id, queue_id);
-       snprintf(z_name_sw, sizeof(z_name_sw), "%s_sw_ring", z_name);
-
-       txq->q.desc = alloc_ring(txq->q.size, sizeof(struct tx_desc),
-                                0, &txq->q.phys_addr,
-                                NULL, 0, queue_id,
-                                socket_id, z_name, z_name_sw);
+       txq->q.desc = alloc_ring(eth_dev, "ctrl_tx_ring", queue_id,
+                                socket_id, txq->q.size, sizeof(struct tx_desc),
+                                0, 0, &txq->q.phys_addr, NULL);
        if (!txq->q.desc)
                return -ENOMEM;
 
@@ -2302,15 +2198,18 @@ static void free_rspq_fl(struct adapter *adap, struct sge_rspq *rq,
  */
 void t4_sge_eth_clear_queues(struct port_info *pi)
 {
-       int i;
        struct adapter *adap = pi->adapter;
-       struct sge_eth_rxq *rxq = &adap->sge.ethrxq[pi->first_qset];
-       struct sge_eth_txq *txq = &adap->sge.ethtxq[pi->first_qset];
+       struct sge_eth_rxq *rxq;
+       struct sge_eth_txq *txq;
+       int i;
 
+       rxq = &adap->sge.ethrxq[pi->first_rxqset];
        for (i = 0; i < pi->n_rx_qsets; i++, rxq++) {
                if (rxq->rspq.desc)
-                       t4_sge_eth_rxq_stop(adap, &rxq->rspq);
+                       t4_sge_eth_rxq_stop(adap, rxq);
        }
+
+       txq = &adap->sge.ethtxq[pi->first_txqset];
        for (i = 0; i < pi->n_tx_qsets; i++, txq++) {
                if (txq->q.desc) {
                        struct sge_txq *q = &txq->q;
@@ -2326,7 +2225,7 @@ void t4_sge_eth_clear_queues(struct port_info *pi)
 void t4_sge_eth_rxq_release(struct adapter *adap, struct sge_eth_rxq *rxq)
 {
        if (rxq->rspq.desc) {
-               t4_sge_eth_rxq_stop(adap, &rxq->rspq);
+               t4_sge_eth_rxq_stop(adap, rxq);
                free_rspq_fl(adap, &rxq->rspq, rxq->fl.size ? &rxq->fl : NULL);
        }
 }
@@ -2343,6 +2242,36 @@ void t4_sge_eth_txq_release(struct adapter *adap, struct sge_eth_txq *txq)
        }
 }
 
+void t4_sge_eth_release_queues(struct port_info *pi)
+{
+       struct adapter *adap = pi->adapter;
+       struct sge_eth_rxq *rxq;
+       struct sge_eth_txq *txq;
+       unsigned int i;
+
+       rxq = &adap->sge.ethrxq[pi->first_rxqset];
+       /* clean up Ethernet Tx/Rx queues */
+       for (i = 0; i < pi->n_rx_qsets; i++, rxq++) {
+               /* Free only the queues allocated */
+               if (rxq->rspq.desc) {
+                       t4_sge_eth_rxq_release(adap, rxq);
+                       rte_eth_dma_zone_free(rxq->rspq.eth_dev, "fl_ring", i);
+                       rte_eth_dma_zone_free(rxq->rspq.eth_dev, "rx_ring", i);
+                       rxq->rspq.eth_dev = NULL;
+               }
+       }
+
+       txq = &adap->sge.ethtxq[pi->first_txqset];
+       for (i = 0; i < pi->n_tx_qsets; i++, txq++) {
+               /* Free only the queues allocated */
+               if (txq->q.desc) {
+                       t4_sge_eth_txq_release(adap, txq);
+                       rte_eth_dma_zone_free(txq->eth_dev, "tx_ring", i);
+                       txq->eth_dev = NULL;
+               }
+       }
+}
+
 void t4_sge_tx_monitor_start(struct adapter *adap)
 {
        rte_eal_alarm_set(50, tx_timer_cb, (void *)adap);
@@ -2362,21 +2291,6 @@ void t4_sge_tx_monitor_stop(struct adapter *adap)
 void t4_free_sge_resources(struct adapter *adap)
 {
        unsigned int i;
-       struct sge_eth_rxq *rxq = &adap->sge.ethrxq[0];
-       struct sge_eth_txq *txq = &adap->sge.ethtxq[0];
-
-       /* clean up Ethernet Tx/Rx queues */
-       for (i = 0; i < adap->sge.max_ethqsets; i++, rxq++, txq++) {
-               /* Free only the queues allocated */
-               if (rxq->rspq.desc) {
-                       t4_sge_eth_rxq_release(adap, rxq);
-                       rxq->rspq.eth_dev = NULL;
-               }
-               if (txq->q.desc) {
-                       t4_sge_eth_txq_release(adap, txq);
-                       txq->eth_dev = NULL;
-               }
-       }
 
        /* clean up control Tx queues */
        for (i = 0; i < ARRAY_SIZE(adap->sge.ctrlq); i++) {
@@ -2386,12 +2300,17 @@ void t4_free_sge_resources(struct adapter *adap)
                        reclaim_completed_tx_imm(&cq->q);
                        t4_ctrl_eq_free(adap, adap->mbox, adap->pf, 0,
                                        cq->q.cntxt_id);
+                       rte_eth_dma_zone_free(adap->eth_dev, "ctrl_tx_ring", i);
+                       rte_mempool_free(cq->mb_pool);
                        free_txq(&cq->q);
                }
        }
 
-       if (adap->sge.fw_evtq.desc)
+       /* clean up firmware event queue */
+       if (adap->sge.fw_evtq.desc) {
                free_rspq_fl(adap, &adap->sge.fw_evtq, NULL);
+               rte_eth_dma_zone_free(adap->eth_dev, "fwq_ring", 0);
+       }
 }
 
 /**