net/mlx5: fix xstats reset reinitialization
[dpdk.git] / drivers / net / mlx5 / mlx5_txq.c
index af84f5f..5ab50ec 100644 (file)
@@ -1121,7 +1121,7 @@ mlx5_txq_new(struct rte_eth_dev *dev, uint16_t idx, uint16_t desc,
                rte_errno = ENOMEM;
                goto error;
        }
-       rte_atomic32_inc(&tmpl->refcnt);
+       __atomic_add_fetch(&tmpl->refcnt, 1, __ATOMIC_RELAXED);
        tmpl->type = MLX5_TXQ_TYPE_STANDARD;
        LIST_INSERT_HEAD(&priv->txqsctrl, tmpl, next);
        return tmpl;
@@ -1165,7 +1165,7 @@ mlx5_txq_hairpin_new(struct rte_eth_dev *dev, uint16_t idx, uint16_t desc,
        tmpl->txq.idx = idx;
        tmpl->hairpin_conf = *hairpin_conf;
        tmpl->type = MLX5_TXQ_TYPE_HAIRPIN;
-       rte_atomic32_inc(&tmpl->refcnt);
+       __atomic_add_fetch(&tmpl->refcnt, 1, __ATOMIC_RELAXED);
        LIST_INSERT_HEAD(&priv->txqsctrl, tmpl, next);
        return tmpl;
 }
@@ -1190,7 +1190,7 @@ mlx5_txq_get(struct rte_eth_dev *dev, uint16_t idx)
 
        if (txq_data) {
                ctrl = container_of(txq_data, struct mlx5_txq_ctrl, txq);
-               rte_atomic32_inc(&ctrl->refcnt);
+               __atomic_add_fetch(&ctrl->refcnt, 1, __ATOMIC_RELAXED);
        }
        return ctrl;
 }
@@ -1215,7 +1215,7 @@ mlx5_txq_release(struct rte_eth_dev *dev, uint16_t idx)
        if (!(*priv->txqs)[idx])
                return 0;
        txq_ctrl = container_of((*priv->txqs)[idx], struct mlx5_txq_ctrl, txq);
-       if (!rte_atomic32_dec_and_test(&txq_ctrl->refcnt))
+       if (__atomic_sub_fetch(&txq_ctrl->refcnt, 1, __ATOMIC_RELAXED) > 1)
                return 1;
        if (txq_ctrl->obj) {
                priv->obj_ops.txq_obj_release(txq_ctrl->obj);
@@ -1229,12 +1229,15 @@ mlx5_txq_release(struct rte_eth_dev *dev, uint16_t idx)
                        txq_ctrl->txq.fcqs = NULL;
                }
                txq_free_elts(txq_ctrl);
-               mlx5_mr_btree_free(&txq_ctrl->txq.mr_ctrl.cache_bh);
        }
-       LIST_REMOVE(txq_ctrl, next);
-       mlx5_free(txq_ctrl);
-       (*priv->txqs)[idx] = NULL;
        dev->data->tx_queue_state[idx] = RTE_ETH_QUEUE_STATE_STOPPED;
+       if (!__atomic_load_n(&txq_ctrl->refcnt, __ATOMIC_RELAXED)) {
+               if (txq_ctrl->type == MLX5_TXQ_TYPE_STANDARD)
+                       mlx5_mr_btree_free(&txq_ctrl->txq.mr_ctrl.cache_bh);
+               LIST_REMOVE(txq_ctrl, next);
+               mlx5_free(txq_ctrl);
+               (*priv->txqs)[idx] = NULL;
+       }
        return 0;
 }
 
@@ -1258,7 +1261,7 @@ mlx5_txq_releasable(struct rte_eth_dev *dev, uint16_t idx)
        if (!(*priv->txqs)[idx])
                return -1;
        txq = container_of((*priv->txqs)[idx], struct mlx5_txq_ctrl, txq);
-       return (rte_atomic32_read(&txq->refcnt) == 1);
+       return (__atomic_load_n(&txq->refcnt, __ATOMIC_RELAXED) == 1);
 }
 
 /**
@@ -1305,7 +1308,7 @@ mlx5_txq_dynf_timestamp_set(struct rte_eth_dev *dev)
                                (RTE_MBUF_DYNFLAG_TX_TIMESTAMP_NAME, NULL);
        off = rte_mbuf_dynfield_lookup
                                (RTE_MBUF_DYNFIELD_TIMESTAMP_NAME, NULL);
-       if (nbit > 0 && off >= 0 && sh->txpp.refcnt)
+       if (nbit >= 0 && off >= 0 && sh->txpp.refcnt)
                mask = 1ULL << nbit;
        for (i = 0; i != priv->txqs_n; ++i) {
                data = (*priv->txqs)[i];