net/mlx5: fix reference count on detached indirect action
[dpdk.git] / drivers / net / mlx5 / mlx5_rxq.c
index 8feb3e2..47dc247 100644 (file)
 
 #include <mlx5_glue.h>
 #include <mlx5_malloc.h>
+#include <mlx5_common.h>
 #include <mlx5_common_mr.h>
 
 #include "mlx5_defs.h"
 #include "mlx5.h"
-#include "mlx5_tx.h"
 #include "mlx5_rx.h"
 #include "mlx5_utils.h"
 #include "mlx5_autoconf.h"
@@ -186,7 +186,8 @@ rxq_alloc_elts_sprq(struct mlx5_rxq_ctrl *rxq_ctrl)
                mbuf_init->data_off = RTE_PKTMBUF_HEADROOM;
                rte_mbuf_refcnt_set(mbuf_init, 1);
                mbuf_init->nb_segs = 1;
-               mbuf_init->port = rxq->port_id;
+               /* For shared queues port is provided in CQE */
+               mbuf_init->port = rxq->shared ? 0 : rxq->port_id;
                if (priv->flags & RTE_PKTMBUF_POOL_F_PINNED_EXT_BUF)
                        mbuf_init->ol_flags = RTE_MBUF_F_EXTERNAL;
                /*
@@ -1170,15 +1171,13 @@ mlx5_arm_cq(struct mlx5_rxq_data *rxq, int sq_n_rxq)
        int sq_n = 0;
        uint32_t doorbell_hi;
        uint64_t doorbell;
-       void *cq_db_reg = (char *)rxq->cq_uar + MLX5_CQ_DOORBELL;
 
        sq_n = sq_n_rxq & MLX5_CQ_SQN_MASK;
        doorbell_hi = sq_n << MLX5_CQ_SQN_OFFSET | (rxq->cq_ci & MLX5_CI_MASK);
        doorbell = (uint64_t)doorbell_hi << 32;
        doorbell |= rxq->cqn;
-       rxq->cq_db[MLX5_CQ_ARM_DB] = rte_cpu_to_be_32(doorbell_hi);
-       mlx5_uar_write64(rte_cpu_to_be_64(doorbell),
-                        cq_db_reg, rxq->uar_lock_cq);
+       mlx5_doorbell_ring(&rxq->uar_data, rte_cpu_to_be_64(doorbell),
+                          doorbell_hi, &rxq->cq_db[MLX5_CQ_ARM_DB], 0);
 }
 
 /**
@@ -1459,8 +1458,7 @@ mlx5_mprq_alloc_mp(struct rte_eth_dev *dev)
                rte_errno = ENOMEM;
                return -rte_errno;
        }
-       ret = mlx5_mr_mempool_register(&priv->sh->cdev->mr_scache,
-                                      priv->sh->cdev->pd, mp, &priv->mp_id);
+       ret = mlx5_mr_mempool_register(priv->sh->cdev, mp, false);
        if (ret < 0 && rte_errno != EEXIST) {
                ret = rte_errno;
                DRV_LOG(ERR, "port %u failed to register a mempool for Multi-Packet RQ",
@@ -1843,9 +1841,6 @@ mlx5_rxq_new(struct rte_eth_dev *dev, struct mlx5_rxq_priv *rxq,
                (struct rte_mbuf *(*)[desc_n])(tmpl + 1);
        tmpl->rxq.mprq_bufs =
                (struct mlx5_mprq_buf *(*)[desc])(*tmpl->rxq.elts + desc_n);
-#ifndef RTE_ARCH_64
-       tmpl->rxq.uar_lock_cq = &priv->sh->uar_lock_cq;
-#endif
        tmpl->rxq.idx = idx;
        LIST_INSERT_HEAD(&priv->rxqsctrl, tmpl, next);
        return tmpl;
@@ -2200,6 +2195,9 @@ mlx5_ind_table_obj_get(struct rte_eth_dev *dev, const uint16_t *queues,
  *   Indirection table to release.
  * @param standalone
  *   Indirection table for Standalone queue.
+ * @param deref_rxqs
+ *   If true, then dereference RX queues related to indirection table.
+ *   Otherwise, no additional action will be taken.
  *
  * @return
  *   1 while a reference on it exists, 0 when freed.
@@ -2207,7 +2205,8 @@ mlx5_ind_table_obj_get(struct rte_eth_dev *dev, const uint16_t *queues,
 int
 mlx5_ind_table_obj_release(struct rte_eth_dev *dev,
                           struct mlx5_ind_table_obj *ind_tbl,
-                          bool standalone)
+                          bool standalone,
+                          bool deref_rxqs)
 {
        struct mlx5_priv *priv = dev->data->dev_private;
        unsigned int i, ret;
@@ -2220,8 +2219,10 @@ mlx5_ind_table_obj_release(struct rte_eth_dev *dev,
        if (ret)
                return 1;
        priv->obj_ops.ind_table_destroy(ind_tbl);
-       for (i = 0; i != ind_tbl->queues_n; ++i)
-               claim_nonzero(mlx5_rxq_deref(dev, ind_tbl->queues[i]));
+       if (deref_rxqs) {
+               for (i = 0; i != ind_tbl->queues_n; ++i)
+                       claim_nonzero(mlx5_rxq_deref(dev, ind_tbl->queues[i]));
+       }
        mlx5_free(ind_tbl);
        return 0;
 }
@@ -2444,7 +2445,7 @@ mlx5_ind_table_obj_attach(struct rte_eth_dev *dev,
                return ret;
        }
        for (i = 0; i < ind_tbl->queues_n; i++)
-               mlx5_rxq_get(dev, ind_tbl->queues[i]);
+               mlx5_rxq_ref(dev, ind_tbl->queues[i]);
        return 0;
 }
 
@@ -2578,7 +2579,7 @@ mlx5_hrxq_modify(struct rte_eth_dev *dev, uint32_t hrxq_idx,
        if (ind_tbl != hrxq->ind_table) {
                MLX5_ASSERT(!hrxq->standalone);
                mlx5_ind_table_obj_release(dev, hrxq->ind_table,
-                                          hrxq->standalone);
+                                          hrxq->standalone, true);
                hrxq->ind_table = ind_tbl;
        }
        hrxq->hash_fields = hash_fields;
@@ -2588,7 +2589,8 @@ error:
        err = rte_errno;
        if (ind_tbl != hrxq->ind_table) {
                MLX5_ASSERT(!hrxq->standalone);
-               mlx5_ind_table_obj_release(dev, ind_tbl, hrxq->standalone);
+               mlx5_ind_table_obj_release(dev, ind_tbl, hrxq->standalone,
+                                          true);
        }
        rte_errno = err;
        return -rte_errno;
@@ -2605,7 +2607,7 @@ __mlx5_hrxq_remove(struct rte_eth_dev *dev, struct mlx5_hrxq *hrxq)
        priv->obj_ops.hrxq_destroy(hrxq);
        if (!hrxq->standalone) {
                mlx5_ind_table_obj_release(dev, hrxq->ind_table,
-                                          hrxq->standalone);
+                                          hrxq->standalone, true);
        }
        mlx5_ipool_free(priv->sh->ipool[MLX5_IPOOL_HRXQ], hrxq->idx);
 }
@@ -2671,7 +2673,7 @@ __mlx5_hrxq_create(struct rte_eth_dev *dev,
        return hrxq;
 error:
        if (!rss_desc->ind_tbl)
-               mlx5_ind_table_obj_release(dev, ind_tbl, standalone);
+               mlx5_ind_table_obj_release(dev, ind_tbl, standalone, true);
        if (hrxq)
                mlx5_ipool_free(priv->sh->ipool[MLX5_IPOOL_HRXQ], hrxq_idx);
        return NULL;