net/mlx5: remove Rx scatter support
authorNélio Laranjeiro <nelio.laranjeiro@6wind.com>
Fri, 24 Jun 2016 13:17:43 +0000 (15:17 +0200)
committerBruce Richardson <bruce.richardson@intel.com>
Mon, 27 Jun 2016 14:17:51 +0000 (16:17 +0200)
This is done in preparation of bypassing Verbs entirely for the data path
as a performance improvement. RX scatter cannot be maintained during the
transition and will be reimplemented later.

Signed-off-by: Nelio Laranjeiro <nelio.laranjeiro@6wind.com>
Signed-off-by: Adrien Mazarguil <adrien.mazarguil@6wind.com>
drivers/net/mlx5/mlx5_ethdev.c
drivers/net/mlx5/mlx5_rxq.c
drivers/net/mlx5/mlx5_rxtx.c
drivers/net/mlx5/mlx5_rxtx.h

index 280a90a..ca57021 100644 (file)
@@ -623,8 +623,7 @@ mlx5_dev_supported_ptypes_get(struct rte_eth_dev *dev)
 
        };
 
-       if (dev->rx_pkt_burst == mlx5_rx_burst ||
-           dev->rx_pkt_burst == mlx5_rx_burst_sp)
+       if (dev->rx_pkt_burst == mlx5_rx_burst)
                return ptypes;
        return NULL;
 }
@@ -762,19 +761,11 @@ mlx5_dev_set_mtu(struct rte_eth_dev *dev, uint16_t mtu)
                mb_len = rte_pktmbuf_data_room_size(rxq->mp);
                assert(mb_len >= RTE_PKTMBUF_HEADROOM);
                sp = (max_frame_len > (mb_len - RTE_PKTMBUF_HEADROOM));
-               /* Provide new values to rxq_setup(). */
-               dev->data->dev_conf.rxmode.jumbo_frame = sp;
-               dev->data->dev_conf.rxmode.max_rx_pkt_len = max_frame_len;
-               ret = rxq_rehash(dev, rxq);
-               if (ret) {
-                       /* Force SP RX if that queue requires it and abort. */
-                       if (rxq->sp)
-                               rx_func = mlx5_rx_burst_sp;
-                       break;
+               if (sp) {
+                       ERROR("%p: RX scatter is not supported", (void *)dev);
+                       ret = ENOTSUP;
+                       goto out;
                }
-               /* Scattered burst function takes priority. */
-               if (rxq->sp)
-                       rx_func = mlx5_rx_burst_sp;
        }
        /* Burst functions can now be called again. */
        rte_wmb();
@@ -1103,22 +1094,12 @@ priv_set_link(struct priv *priv, int up)
 {
        struct rte_eth_dev *dev = priv->dev;
        int err;
-       unsigned int i;
 
        if (up) {
                err = priv_set_flags(priv, ~IFF_UP, IFF_UP);
                if (err)
                        return err;
-               for (i = 0; i < priv->rxqs_n; i++)
-                       if ((*priv->rxqs)[i]->sp)
-                               break;
-               /* Check if an sp queue exists.
-                * Note: Some old frames might be received.
-                */
-               if (i == priv->rxqs_n)
-                       dev->rx_pkt_burst = mlx5_rx_burst;
-               else
-                       dev->rx_pkt_burst = mlx5_rx_burst_sp;
+               dev->rx_pkt_burst = mlx5_rx_burst;
                dev->tx_pkt_burst = mlx5_tx_burst;
        } else {
                err = priv_set_flags(priv, ~IFF_UP, ~IFF_UP);
index 0bcf55b..38ff9fd 100644 (file)
@@ -633,145 +633,6 @@ priv_rehash_flows(struct priv *priv)
        return 0;
 }
 
-/**
- * Allocate RX queue elements with scattered packets support.
- *
- * @param rxq
- *   Pointer to RX queue structure.
- * @param elts_n
- *   Number of elements to allocate.
- * @param[in] pool
- *   If not NULL, fetch buffers from this array instead of allocating them
- *   with rte_pktmbuf_alloc().
- *
- * @return
- *   0 on success, errno value on failure.
- */
-static int
-rxq_alloc_elts_sp(struct rxq *rxq, unsigned int elts_n,
-                 struct rte_mbuf **pool)
-{
-       unsigned int i;
-       struct rxq_elt_sp (*elts)[elts_n] =
-               rte_calloc_socket("RXQ elements", 1, sizeof(*elts), 0,
-                                 rxq->socket);
-       int ret = 0;
-
-       if (elts == NULL) {
-               ERROR("%p: can't allocate packets array", (void *)rxq);
-               ret = ENOMEM;
-               goto error;
-       }
-       /* For each WR (packet). */
-       for (i = 0; (i != elts_n); ++i) {
-               unsigned int j;
-               struct rxq_elt_sp *elt = &(*elts)[i];
-               struct ibv_sge (*sges)[RTE_DIM(elt->sges)] = &elt->sges;
-
-               /* These two arrays must have the same size. */
-               assert(RTE_DIM(elt->sges) == RTE_DIM(elt->bufs));
-               /* For each SGE (segment). */
-               for (j = 0; (j != RTE_DIM(elt->bufs)); ++j) {
-                       struct ibv_sge *sge = &(*sges)[j];
-                       struct rte_mbuf *buf;
-
-                       if (pool != NULL) {
-                               buf = *(pool++);
-                               assert(buf != NULL);
-                               rte_pktmbuf_reset(buf);
-                       } else
-                               buf = rte_pktmbuf_alloc(rxq->mp);
-                       if (buf == NULL) {
-                               assert(pool == NULL);
-                               ERROR("%p: empty mbuf pool", (void *)rxq);
-                               ret = ENOMEM;
-                               goto error;
-                       }
-                       elt->bufs[j] = buf;
-                       /* Headroom is reserved by rte_pktmbuf_alloc(). */
-                       assert(DATA_OFF(buf) == RTE_PKTMBUF_HEADROOM);
-                       /* Buffer is supposed to be empty. */
-                       assert(rte_pktmbuf_data_len(buf) == 0);
-                       assert(rte_pktmbuf_pkt_len(buf) == 0);
-                       /* sge->addr must be able to store a pointer. */
-                       assert(sizeof(sge->addr) >= sizeof(uintptr_t));
-                       if (j == 0) {
-                               /* The first SGE keeps its headroom. */
-                               sge->addr = rte_pktmbuf_mtod(buf, uintptr_t);
-                               sge->length = (buf->buf_len -
-                                              RTE_PKTMBUF_HEADROOM);
-                       } else {
-                               /* Subsequent SGEs lose theirs. */
-                               assert(DATA_OFF(buf) == RTE_PKTMBUF_HEADROOM);
-                               SET_DATA_OFF(buf, 0);
-                               sge->addr = (uintptr_t)buf->buf_addr;
-                               sge->length = buf->buf_len;
-                       }
-                       sge->lkey = rxq->mr->lkey;
-                       /* Redundant check for tailroom. */
-                       assert(sge->length == rte_pktmbuf_tailroom(buf));
-               }
-       }
-       DEBUG("%p: allocated and configured %u WRs (%zu segments)",
-             (void *)rxq, elts_n, (elts_n * RTE_DIM((*elts)[0].sges)));
-       rxq->elts_n = elts_n;
-       rxq->elts_head = 0;
-       rxq->elts.sp = elts;
-       assert(ret == 0);
-       return 0;
-error:
-       if (elts != NULL) {
-               assert(pool == NULL);
-               for (i = 0; (i != RTE_DIM(*elts)); ++i) {
-                       unsigned int j;
-                       struct rxq_elt_sp *elt = &(*elts)[i];
-
-                       for (j = 0; (j != RTE_DIM(elt->bufs)); ++j) {
-                               struct rte_mbuf *buf = elt->bufs[j];
-
-                               if (buf != NULL)
-                                       rte_pktmbuf_free_seg(buf);
-                       }
-               }
-               rte_free(elts);
-       }
-       DEBUG("%p: failed, freed everything", (void *)rxq);
-       assert(ret > 0);
-       return ret;
-}
-
-/**
- * Free RX queue elements with scattered packets support.
- *
- * @param rxq
- *   Pointer to RX queue structure.
- */
-static void
-rxq_free_elts_sp(struct rxq *rxq)
-{
-       unsigned int i;
-       unsigned int elts_n = rxq->elts_n;
-       struct rxq_elt_sp (*elts)[elts_n] = rxq->elts.sp;
-
-       DEBUG("%p: freeing WRs", (void *)rxq);
-       rxq->elts_n = 0;
-       rxq->elts.sp = NULL;
-       if (elts == NULL)
-               return;
-       for (i = 0; (i != RTE_DIM(*elts)); ++i) {
-               unsigned int j;
-               struct rxq_elt_sp *elt = &(*elts)[i];
-
-               for (j = 0; (j != RTE_DIM(elt->bufs)); ++j) {
-                       struct rte_mbuf *buf = elt->bufs[j];
-
-                       if (buf != NULL)
-                               rte_pktmbuf_free_seg(buf);
-               }
-       }
-       rte_free(elts);
-}
-
 /**
  * Allocate RX queue elements.
  *
@@ -838,7 +699,7 @@ rxq_alloc_elts(struct rxq *rxq, unsigned int elts_n, struct rte_mbuf **pool)
              (void *)rxq, elts_n);
        rxq->elts_n = elts_n;
        rxq->elts_head = 0;
-       rxq->elts.no_sp = elts;
+       rxq->elts = elts;
        assert(ret == 0);
        return 0;
 error:
@@ -869,11 +730,11 @@ rxq_free_elts(struct rxq *rxq)
 {
        unsigned int i;
        unsigned int elts_n = rxq->elts_n;
-       struct rxq_elt (*elts)[elts_n] = rxq->elts.no_sp;
+       struct rxq_elt (*elts)[elts_n] = rxq->elts;
 
        DEBUG("%p: freeing WRs", (void *)rxq);
        rxq->elts_n = 0;
-       rxq->elts.no_sp = NULL;
+       rxq->elts = NULL;
        if (elts == NULL)
                return;
        for (i = 0; (i != RTE_DIM(*elts)); ++i) {
@@ -900,10 +761,7 @@ rxq_cleanup(struct rxq *rxq)
        struct ibv_exp_release_intf_params params;
 
        DEBUG("cleaning up %p", (void *)rxq);
-       if (rxq->sp)
-               rxq_free_elts_sp(rxq);
-       else
-               rxq_free_elts(rxq);
+       rxq_free_elts(rxq);
        rxq->poll = NULL;
        rxq->recv = NULL;
        if (rxq->if_wq != NULL) {
@@ -973,12 +831,12 @@ rxq_rehash(struct rte_eth_dev *dev, struct rxq *rxq)
        struct rte_mbuf **pool;
        unsigned int i, k;
        struct ibv_exp_wq_attr mod;
-       unsigned int mb_len = rte_pktmbuf_data_room_size(rxq->mp);
+       struct rxq_elt (*elts)[tmpl.elts_n];
        int err;
 
        DEBUG("%p: rehashing queue %p", (void *)dev, (void *)rxq);
        /* Number of descriptors and mbufs currently allocated. */
-       desc_n = (tmpl.elts_n * (tmpl.sp ? MLX5_PMD_SGE_WR_N : 1));
+       desc_n = tmpl.elts_n;
        mbuf_n = desc_n;
        /* Toggle RX checksum offload if hardware supports it. */
        if (priv->hw_csum) {
@@ -989,22 +847,6 @@ rxq_rehash(struct rte_eth_dev *dev, struct rxq *rxq)
                tmpl.csum_l2tun = !!dev->data->dev_conf.rxmode.hw_ip_checksum;
                rxq->csum_l2tun = tmpl.csum_l2tun;
        }
-       /* Enable scattered packets support for this queue if necessary. */
-       assert(mb_len >= RTE_PKTMBUF_HEADROOM);
-       if ((dev->data->dev_conf.rxmode.jumbo_frame) &&
-           (dev->data->dev_conf.rxmode.max_rx_pkt_len >
-            (mb_len - RTE_PKTMBUF_HEADROOM))) {
-               tmpl.sp = 1;
-               desc_n /= MLX5_PMD_SGE_WR_N;
-       } else
-               tmpl.sp = 0;
-       DEBUG("%p: %s scattered packets support (%u WRs)",
-             (void *)dev, (tmpl.sp ? "enabling" : "disabling"), desc_n);
-       /* If scatter mode is the same as before, nothing to do. */
-       if (tmpl.sp == rxq->sp) {
-               DEBUG("%p: nothing to do", (void *)dev);
-               return 0;
-       }
        /* From now on, any failure will render the queue unusable.
         * Reinitialize WQ. */
        mod = (struct ibv_exp_wq_attr){
@@ -1025,35 +867,18 @@ rxq_rehash(struct rte_eth_dev *dev, struct rxq *rxq)
        }
        /* Snatch mbufs from original queue. */
        k = 0;
-       if (rxq->sp) {
-               struct rxq_elt_sp (*elts)[rxq->elts_n] = rxq->elts.sp;
-
-               for (i = 0; (i != RTE_DIM(*elts)); ++i) {
-                       struct rxq_elt_sp *elt = &(*elts)[i];
-                       unsigned int j;
-
-                       for (j = 0; (j != RTE_DIM(elt->bufs)); ++j) {
-                               assert(elt->bufs[j] != NULL);
-                               pool[k++] = elt->bufs[j];
-                       }
-               }
-       } else {
-               struct rxq_elt (*elts)[rxq->elts_n] = rxq->elts.no_sp;
-
-               for (i = 0; (i != RTE_DIM(*elts)); ++i) {
-                       struct rxq_elt *elt = &(*elts)[i];
-                       struct rte_mbuf *buf = elt->buf;
+       elts = rxq->elts;
+       for (i = 0; (i != RTE_DIM(*elts)); ++i) {
+               struct rxq_elt *elt = &(*elts)[i];
+               struct rte_mbuf *buf = elt->buf;
 
-                       pool[k++] = buf;
-               }
+               pool[k++] = buf;
        }
        assert(k == mbuf_n);
        tmpl.elts_n = 0;
-       tmpl.elts.sp = NULL;
-       assert((void *)&tmpl.elts.sp == (void *)&tmpl.elts.no_sp);
-       err = ((tmpl.sp) ?
-              rxq_alloc_elts_sp(&tmpl, desc_n, pool) :
-              rxq_alloc_elts(&tmpl, desc_n, pool));
+       tmpl.elts = NULL;
+       assert((void *)&tmpl.elts == NULL);
+       err = rxq_alloc_elts(&tmpl, desc_n, pool);
        if (err) {
                ERROR("%p: cannot reallocate WRs, aborting", (void *)dev);
                rte_free(pool);
@@ -1061,12 +886,11 @@ rxq_rehash(struct rte_eth_dev *dev, struct rxq *rxq)
                return err;
        }
        assert(tmpl.elts_n == desc_n);
-       assert(tmpl.elts.sp != NULL);
        rte_free(pool);
        /* Clean up original data. */
        rxq->elts_n = 0;
-       rte_free(rxq->elts.sp);
-       rxq->elts.sp = NULL;
+       rte_free(rxq->elts);
+       rxq->elts = NULL;
        /* Change queue state to ready. */
        mod = (struct ibv_exp_wq_attr){
                .attr_mask = IBV_EXP_WQ_ATTR_STATE,
@@ -1080,28 +904,14 @@ rxq_rehash(struct rte_eth_dev *dev, struct rxq *rxq)
        }
        /* Post SGEs. */
        assert(tmpl.if_wq != NULL);
-       if (tmpl.sp) {
-               struct rxq_elt_sp (*elts)[tmpl.elts_n] = tmpl.elts.sp;
-
-               for (i = 0; (i != RTE_DIM(*elts)); ++i) {
-                       err = tmpl.if_wq->recv_sg_list
-                               (tmpl.wq,
-                                (*elts)[i].sges,
-                                RTE_DIM((*elts)[i].sges));
-                       if (err)
-                               break;
-               }
-       } else {
-               struct rxq_elt (*elts)[tmpl.elts_n] = tmpl.elts.no_sp;
-
-               for (i = 0; (i != RTE_DIM(*elts)); ++i) {
-                       err = tmpl.if_wq->recv_burst(
-                               tmpl.wq,
-                               &(*elts)[i].sge,
-                               1);
-                       if (err)
-                               break;
-               }
+       elts = tmpl.elts;
+       for (i = 0; (i != RTE_DIM(*elts)); ++i) {
+               err = tmpl.if_wq->recv_burst(
+                       tmpl.wq,
+                       &(*elts)[i].sge,
+                       1);
+               if (err)
+                       break;
        }
        if (err) {
                ERROR("%p: failed to post SGEs with error %d",
@@ -1110,10 +920,7 @@ rxq_rehash(struct rte_eth_dev *dev, struct rxq *rxq)
                err = EIO;
                goto error;
        }
-       if (tmpl.sp)
-               tmpl.recv = tmpl.if_wq->recv_sg_list;
-       else
-               tmpl.recv = tmpl.if_wq->recv_burst;
+       tmpl.recv = tmpl.if_wq->recv_burst;
 error:
        *rxq = tmpl;
        assert(err >= 0);
@@ -1159,31 +966,26 @@ rxq_setup(struct rte_eth_dev *dev, struct rxq *rxq, uint16_t desc,
        } attr;
        enum ibv_exp_query_intf_status status;
        unsigned int mb_len = rte_pktmbuf_data_room_size(mp);
+       struct rxq_elt (*elts)[desc];
        int ret = 0;
        unsigned int i;
        unsigned int cq_size = desc;
 
        (void)conf; /* Thresholds configuration (ignored). */
-       if ((desc == 0) || (desc % MLX5_PMD_SGE_WR_N)) {
-               ERROR("%p: invalid number of RX descriptors (must be a"
-                     " multiple of %d)", (void *)dev, MLX5_PMD_SGE_WR_N);
+       if (desc == 0) {
+               ERROR("%p: invalid number of RX descriptors", (void *)dev);
                return EINVAL;
        }
+       if (MLX5_PMD_SGE_WR_N > 1) {
+               ERROR("%p: RX scatter is not supported", (void *)dev);
+               return ENOTSUP;
+       }
        /* Toggle RX checksum offload if hardware supports it. */
        if (priv->hw_csum)
                tmpl.csum = !!dev->data->dev_conf.rxmode.hw_ip_checksum;
        if (priv->hw_csum_l2tun)
                tmpl.csum_l2tun = !!dev->data->dev_conf.rxmode.hw_ip_checksum;
-       /* Enable scattered packets support for this queue if necessary. */
-       assert(mb_len >= RTE_PKTMBUF_HEADROOM);
-       if ((dev->data->dev_conf.rxmode.jumbo_frame) &&
-           (dev->data->dev_conf.rxmode.max_rx_pkt_len >
-            (mb_len - RTE_PKTMBUF_HEADROOM))) {
-               tmpl.sp = 1;
-               desc /= MLX5_PMD_SGE_WR_N;
-       }
-       DEBUG("%p: %s scattered packets support (%u WRs)",
-             (void *)dev, (tmpl.sp ? "enabling" : "disabling"), desc);
+       (void)mb_len; /* I'll be back! */
        /* Use the entire RX mempool as the memory region. */
        tmpl.mr = mlx5_mp2mr(priv->pd, mp);
        if (tmpl.mr == NULL) {
@@ -1232,10 +1034,7 @@ rxq_setup(struct rte_eth_dev *dev, struct rxq *rxq, uint16_t desc,
                                priv->device_attr.max_qp_wr :
                                (int)cq_size),
                /* Max number of scatter/gather elements in a WR. */
-               .max_recv_sge = ((priv->device_attr.max_sge <
-                                 MLX5_PMD_SGE_WR_N) ?
-                                priv->device_attr.max_sge :
-                                MLX5_PMD_SGE_WR_N),
+               .max_recv_sge = 1,
                .pd = priv->pd,
                .cq = tmpl.cq,
                .comp_mask =
@@ -1297,10 +1096,7 @@ rxq_setup(struct rte_eth_dev *dev, struct rxq *rxq, uint16_t desc,
                      (void *)dev, strerror(ret));
                goto error;
        }
-       if (tmpl.sp)
-               ret = rxq_alloc_elts_sp(&tmpl, desc, NULL);
-       else
-               ret = rxq_alloc_elts(&tmpl, desc, NULL);
+       ret = rxq_alloc_elts(&tmpl, desc, NULL);
        if (ret) {
                ERROR("%p: RXQ allocation failed: %s",
                      (void *)dev, strerror(ret));
@@ -1346,28 +1142,14 @@ rxq_setup(struct rte_eth_dev *dev, struct rxq *rxq, uint16_t desc,
                goto error;
        }
        /* Post SGEs. */
-       if (tmpl.sp) {
-               struct rxq_elt_sp (*elts)[tmpl.elts_n] = tmpl.elts.sp;
-
-               for (i = 0; (i != RTE_DIM(*elts)); ++i) {
-                       ret = tmpl.if_wq->recv_sg_list
-                               (tmpl.wq,
-                                (*elts)[i].sges,
-                                RTE_DIM((*elts)[i].sges));
-                       if (ret)
-                               break;
-               }
-       } else {
-               struct rxq_elt (*elts)[tmpl.elts_n] = tmpl.elts.no_sp;
-
-               for (i = 0; (i != RTE_DIM(*elts)); ++i) {
-                       ret = tmpl.if_wq->recv_burst(
-                               tmpl.wq,
-                               &(*elts)[i].sge,
-                               1);
-                       if (ret)
-                               break;
-               }
+       elts = tmpl.elts;
+       for (i = 0; (i != RTE_DIM(*elts)); ++i) {
+               ret = tmpl.if_wq->recv_burst(
+                       tmpl.wq,
+                       &(*elts)[i].sge,
+                       1);
+               if (ret)
+                       break;
        }
        if (ret) {
                ERROR("%p: failed to post SGEs with error %d",
@@ -1388,10 +1170,7 @@ rxq_setup(struct rte_eth_dev *dev, struct rxq *rxq, uint16_t desc,
 #else /* HAVE_EXP_DEVICE_ATTR_VLAN_OFFLOADS */
        rxq->poll = rxq->if_cq->poll_length_flags;
 #endif /* HAVE_EXP_DEVICE_ATTR_VLAN_OFFLOADS */
-       if (rxq->sp)
-               rxq->recv = rxq->if_wq->recv_sg_list;
-       else
-               rxq->recv = rxq->if_wq->recv_burst;
+       rxq->recv = rxq->if_wq->recv_burst;
        return 0;
 error:
        rxq_cleanup(&tmpl);
@@ -1466,10 +1245,7 @@ mlx5_rx_queue_setup(struct rte_eth_dev *dev, uint16_t idx, uint16_t desc,
                      (void *)dev, (void *)rxq);
                (*priv->rxqs)[idx] = rxq;
                /* Update receive callback. */
-               if (rxq->sp)
-                       dev->rx_pkt_burst = mlx5_rx_burst_sp;
-               else
-                       dev->rx_pkt_burst = mlx5_rx_burst;
+               dev->rx_pkt_burst = mlx5_rx_burst;
        }
        priv_unlock(priv);
        return -ret;
index 9a1f2f7..9f50e74 100644 (file)
@@ -503,217 +503,9 @@ rxq_cq_to_ol_flags(const struct rxq *rxq, uint32_t flags)
        return ol_flags;
 }
 
-/**
- * DPDK callback for RX with scattered packets support.
- *
- * @param dpdk_rxq
- *   Generic pointer to RX queue structure.
- * @param[out] pkts
- *   Array to store received packets.
- * @param pkts_n
- *   Maximum number of packets in array.
- *
- * @return
- *   Number of packets successfully received (<= pkts_n).
- */
-uint16_t
-mlx5_rx_burst_sp(void *dpdk_rxq, struct rte_mbuf **pkts, uint16_t pkts_n)
-{
-       struct rxq *rxq = (struct rxq *)dpdk_rxq;
-       struct rxq_elt_sp (*elts)[rxq->elts_n] = rxq->elts.sp;
-       const unsigned int elts_n = rxq->elts_n;
-       unsigned int elts_head = rxq->elts_head;
-       unsigned int i;
-       unsigned int pkts_ret = 0;
-       int ret;
-
-       if (unlikely(!rxq->sp))
-               return mlx5_rx_burst(dpdk_rxq, pkts, pkts_n);
-       if (unlikely(elts == NULL)) /* See RTE_DEV_CMD_SET_MTU. */
-               return 0;
-       for (i = 0; (i != pkts_n); ++i) {
-               struct rxq_elt_sp *elt = &(*elts)[elts_head];
-               unsigned int len;
-               unsigned int pkt_buf_len;
-               struct rte_mbuf *pkt_buf = NULL; /* Buffer returned in pkts. */
-               struct rte_mbuf **pkt_buf_next = &pkt_buf;
-               unsigned int seg_headroom = RTE_PKTMBUF_HEADROOM;
-               unsigned int j = 0;
-               uint32_t flags;
-               uint16_t vlan_tci;
-
-               /* Sanity checks. */
-               assert(elts_head < rxq->elts_n);
-               assert(rxq->elts_head < rxq->elts_n);
-               ret = rxq->poll(rxq->cq, NULL, NULL, &flags, &vlan_tci);
-               if (unlikely(ret < 0)) {
-                       struct ibv_wc wc;
-                       int wcs_n;
-
-                       DEBUG("rxq=%p, poll_length() failed (ret=%d)",
-                             (void *)rxq, ret);
-                       /* ibv_poll_cq() must be used in case of failure. */
-                       wcs_n = ibv_poll_cq(rxq->cq, 1, &wc);
-                       if (unlikely(wcs_n == 0))
-                               break;
-                       if (unlikely(wcs_n < 0)) {
-                               DEBUG("rxq=%p, ibv_poll_cq() failed (wcs_n=%d)",
-                                     (void *)rxq, wcs_n);
-                               break;
-                       }
-                       assert(wcs_n == 1);
-                       if (unlikely(wc.status != IBV_WC_SUCCESS)) {
-                               /* Whatever, just repost the offending WR. */
-                               DEBUG("rxq=%p, wr_id=%" PRIu64 ": bad work"
-                                     " completion status (%d): %s",
-                                     (void *)rxq, wc.wr_id, wc.status,
-                                     ibv_wc_status_str(wc.status));
-#ifdef MLX5_PMD_SOFT_COUNTERS
-                               /* Increment dropped packets counter. */
-                               ++rxq->stats.idropped;
-#endif
-                               goto repost;
-                       }
-                       ret = wc.byte_len;
-               }
-               if (ret == 0)
-                       break;
-               assert(ret >= (rxq->crc_present << 2));
-               len = ret - (rxq->crc_present << 2);
-               pkt_buf_len = len;
-               /*
-                * Replace spent segments with new ones, concatenate and
-                * return them as pkt_buf.
-                */
-               while (1) {
-                       struct ibv_sge *sge = &elt->sges[j];
-                       struct rte_mbuf *seg = elt->bufs[j];
-                       struct rte_mbuf *rep;
-                       unsigned int seg_tailroom;
-
-                       assert(seg != NULL);
-                       /*
-                        * Fetch initial bytes of packet descriptor into a
-                        * cacheline while allocating rep.
-                        */
-                       rte_prefetch0(seg);
-                       rep = rte_mbuf_raw_alloc(rxq->mp);
-                       if (unlikely(rep == NULL)) {
-                               /*
-                                * Unable to allocate a replacement mbuf,
-                                * repost WR.
-                                */
-                               DEBUG("rxq=%p: can't allocate a new mbuf",
-                                     (void *)rxq);
-                               if (pkt_buf != NULL) {
-                                       *pkt_buf_next = NULL;
-                                       rte_pktmbuf_free(pkt_buf);
-                               }
-                               /* Increment out of memory counters. */
-                               ++rxq->stats.rx_nombuf;
-                               ++rxq->priv->dev->data->rx_mbuf_alloc_failed;
-                               goto repost;
-                       }
-#ifndef NDEBUG
-                       /* Poison user-modifiable fields in rep. */
-                       NEXT(rep) = (void *)((uintptr_t)-1);
-                       SET_DATA_OFF(rep, 0xdead);
-                       DATA_LEN(rep) = 0xd00d;
-                       PKT_LEN(rep) = 0xdeadd00d;
-                       NB_SEGS(rep) = 0x2a;
-                       PORT(rep) = 0x2a;
-                       rep->ol_flags = -1;
-#endif
-                       assert(rep->buf_len == seg->buf_len);
-                       /* Reconfigure sge to use rep instead of seg. */
-                       assert(sge->lkey == rxq->mr->lkey);
-                       sge->addr = ((uintptr_t)rep->buf_addr + seg_headroom);
-                       elt->bufs[j] = rep;
-                       ++j;
-                       /* Update pkt_buf if it's the first segment, or link
-                        * seg to the previous one and update pkt_buf_next. */
-                       *pkt_buf_next = seg;
-                       pkt_buf_next = &NEXT(seg);
-                       /* Update seg information. */
-                       seg_tailroom = (seg->buf_len - seg_headroom);
-                       assert(sge->length == seg_tailroom);
-                       SET_DATA_OFF(seg, seg_headroom);
-                       if (likely(len <= seg_tailroom)) {
-                               /* Last segment. */
-                               DATA_LEN(seg) = len;
-                               PKT_LEN(seg) = len;
-                               /* Sanity check. */
-                               assert(rte_pktmbuf_headroom(seg) ==
-                                      seg_headroom);
-                               assert(rte_pktmbuf_tailroom(seg) ==
-                                      (seg_tailroom - len));
-                               break;
-                       }
-                       DATA_LEN(seg) = seg_tailroom;
-                       PKT_LEN(seg) = seg_tailroom;
-                       /* Sanity check. */
-                       assert(rte_pktmbuf_headroom(seg) == seg_headroom);
-                       assert(rte_pktmbuf_tailroom(seg) == 0);
-                       /* Fix len and clear headroom for next segments. */
-                       len -= seg_tailroom;
-                       seg_headroom = 0;
-               }
-               /* Update head and tail segments. */
-               *pkt_buf_next = NULL;
-               assert(pkt_buf != NULL);
-               assert(j != 0);
-               NB_SEGS(pkt_buf) = j;
-               PORT(pkt_buf) = rxq->port_id;
-               PKT_LEN(pkt_buf) = pkt_buf_len;
-               if (rxq->csum | rxq->csum_l2tun | rxq->vlan_strip) {
-                       pkt_buf->packet_type = rxq_cq_to_pkt_type(flags);
-                       pkt_buf->ol_flags = rxq_cq_to_ol_flags(rxq, flags);
-#ifdef HAVE_EXP_DEVICE_ATTR_VLAN_OFFLOADS
-                       if (flags & IBV_EXP_CQ_RX_CVLAN_STRIPPED_V1) {
-                               pkt_buf->ol_flags |= PKT_RX_VLAN_PKT |
-                                       PKT_RX_VLAN_STRIPPED;
-                               pkt_buf->vlan_tci = vlan_tci;
-                       }
-#endif /* HAVE_EXP_DEVICE_ATTR_VLAN_OFFLOADS */
-               }
-
-               /* Return packet. */
-               *(pkts++) = pkt_buf;
-               ++pkts_ret;
-#ifdef MLX5_PMD_SOFT_COUNTERS
-               /* Increment bytes counter. */
-               rxq->stats.ibytes += pkt_buf_len;
-#endif
-repost:
-               ret = rxq->recv(rxq->wq, elt->sges, RTE_DIM(elt->sges));
-               if (unlikely(ret)) {
-                       /* Inability to repost WRs is fatal. */
-                       DEBUG("%p: recv_sg_list(): failed (ret=%d)",
-                             (void *)rxq->priv,
-                             ret);
-                       abort();
-               }
-               if (++elts_head >= elts_n)
-                       elts_head = 0;
-               continue;
-       }
-       if (unlikely(i == 0))
-               return 0;
-       rxq->elts_head = elts_head;
-#ifdef MLX5_PMD_SOFT_COUNTERS
-       /* Increment packets counter. */
-       rxq->stats.ipackets += pkts_ret;
-#endif
-       return pkts_ret;
-}
-
 /**
  * DPDK callback for RX.
  *
- * The following function is the same as mlx5_rx_burst_sp(), except it doesn't
- * manage scattered packets. Improves performance when MRU is lower than the
- * size of the first segment.
- *
  * @param dpdk_rxq
  *   Generic pointer to RX queue structure.
  * @param[out] pkts
@@ -728,7 +520,7 @@ uint16_t
 mlx5_rx_burst(void *dpdk_rxq, struct rte_mbuf **pkts, uint16_t pkts_n)
 {
        struct rxq *rxq = (struct rxq *)dpdk_rxq;
-       struct rxq_elt (*elts)[rxq->elts_n] = rxq->elts.no_sp;
+       struct rxq_elt (*elts)[rxq->elts_n] = rxq->elts;
        const unsigned int elts_n = rxq->elts_n;
        unsigned int elts_head = rxq->elts_head;
        struct ibv_sge sges[pkts_n];
@@ -736,8 +528,6 @@ mlx5_rx_burst(void *dpdk_rxq, struct rte_mbuf **pkts, uint16_t pkts_n)
        unsigned int pkts_ret = 0;
        int ret;
 
-       if (unlikely(rxq->sp))
-               return mlx5_rx_burst_sp(dpdk_rxq, pkts, pkts_n);
        for (i = 0; (i != pkts_n); ++i) {
                struct rxq_elt *elt = &(*elts)[elts_head];
                unsigned int len;
index 8358ccb..2e1f83b 100644 (file)
@@ -81,12 +81,6 @@ struct mlx5_txq_stats {
        uint64_t odropped; /**< Total of packets not sent when TX ring full. */
 };
 
-/* RX element (scattered packets). */
-struct rxq_elt_sp {
-       struct ibv_sge sges[MLX5_PMD_SGE_WR_N]; /* Scatter/Gather Elements. */
-       struct rte_mbuf *bufs[MLX5_PMD_SGE_WR_N]; /* SGEs buffers. */
-};
-
 /* RX element. */
 struct rxq_elt {
        struct ibv_sge sge; /* Scatter/Gather Element. */
@@ -112,15 +106,11 @@ struct rxq {
        unsigned int port_id; /* Port ID for incoming packets. */
        unsigned int elts_n; /* (*elts)[] length. */
        unsigned int elts_head; /* Current index in (*elts)[]. */
-       unsigned int sp:1; /* Use scattered RX elements. */
        unsigned int csum:1; /* Enable checksum offloading. */
        unsigned int csum_l2tun:1; /* Same for L2 tunnels. */
        unsigned int vlan_strip:1; /* Enable VLAN stripping. */
        unsigned int crc_present:1; /* CRC must be subtracted. */
-       union {
-               struct rxq_elt_sp (*sp)[]; /* Scattered RX elements. */
-               struct rxq_elt (*no_sp)[]; /* RX elements. */
-       } elts;
+       struct rxq_elt (*elts)[]; /* RX elements. */
        unsigned int socket; /* CPU socket ID for allocations. */
        struct mlx5_rxq_stats stats; /* RX queue counters. */
        struct ibv_exp_res_domain *rd; /* Resource Domain. */
@@ -321,7 +311,6 @@ uint16_t mlx5_tx_burst_secondary_setup(void *, struct rte_mbuf **, uint16_t);
 /* mlx5_rxtx.c */
 
 uint16_t mlx5_tx_burst(void *, struct rte_mbuf **, uint16_t);
-uint16_t mlx5_rx_burst_sp(void *, struct rte_mbuf **, uint16_t);
 uint16_t mlx5_rx_burst(void *, struct rte_mbuf **, uint16_t);
 uint16_t removed_tx_burst(void *, struct rte_mbuf **, uint16_t);
 uint16_t removed_rx_burst(void *, struct rte_mbuf **, uint16_t);