X-Git-Url: http://git.droids-corp.org/?a=blobdiff_plain;f=drivers%2Fnet%2Fbnxt%2Fbnxt_txq.c;h=830416af3d01c0c89c35e3f82eda910bcfe4f0d0;hb=cd212c466992535af6e1bc99677cfd2790622b5d;hp=2d7645eeb0d2ae9630ad0986d79a7d9788b92e2c;hpb=f03e66cb64cee0e9b4351a7ac42f467e69d966f2;p=dpdk.git diff --git a/drivers/net/bnxt/bnxt_txq.c b/drivers/net/bnxt/bnxt_txq.c index 2d7645eeb0..830416af3d 100644 --- a/drivers/net/bnxt/bnxt_txq.c +++ b/drivers/net/bnxt/bnxt_txq.c @@ -1,5 +1,5 @@ /* SPDX-License-Identifier: BSD-3-Clause - * Copyright(c) 2014-2018 Broadcom + * Copyright(c) 2014-2021 Broadcom * All rights reserved. */ @@ -8,6 +8,7 @@ #include #include "bnxt.h" +#include "bnxt_hwrm.h" #include "bnxt_ring.h" #include "bnxt_txq.h" #include "bnxt_txr.h" @@ -24,18 +25,18 @@ void bnxt_free_txq_stats(struct bnxt_tx_queue *txq) static void bnxt_tx_queue_release_mbufs(struct bnxt_tx_queue *txq) { - struct bnxt_sw_tx_bd *sw_ring; + struct rte_mbuf **sw_ring; uint16_t i; - if (!txq) + if (!txq || !txq->tx_ring) return; sw_ring = txq->tx_ring->tx_buf_ring; if (sw_ring) { for (i = 0; i < txq->tx_ring->tx_ring_struct->ring_size; i++) { - if (sw_ring[i].mbuf) { - rte_pktmbuf_free_seg(sw_ring[i].mbuf); - sw_ring[i].mbuf = NULL; + if (sw_ring[i]) { + rte_pktmbuf_free_seg(sw_ring[i]); + sw_ring[i] = NULL; } } } @@ -61,11 +62,20 @@ void bnxt_tx_queue_release_op(void *tx_queue) return; /* Free TX ring hardware descriptors */ + bnxt_free_hwrm_tx_ring(txq->bp, txq->queue_id); bnxt_tx_queue_release_mbufs(txq); - bnxt_free_ring(txq->tx_ring->tx_ring_struct); + if (txq->tx_ring) { + bnxt_free_ring(txq->tx_ring->tx_ring_struct); + rte_free(txq->tx_ring->tx_ring_struct); + rte_free(txq->tx_ring); + } /* Free TX completion ring hardware descriptors */ - bnxt_free_ring(txq->cp_ring->cp_ring_struct); + if (txq->cp_ring) { + bnxt_free_ring(txq->cp_ring->cp_ring_struct); + rte_free(txq->cp_ring->cp_ring_struct); + rte_free(txq->cp_ring); + } bnxt_free_txq_stats(txq); rte_memzone_free(txq->mz); @@ -90,17 +100,16 @@ int bnxt_tx_queue_setup_op(struct rte_eth_dev *eth_dev, if (rc) return rc; - if (queue_idx >= BNXT_MAX_RINGS(bp)) { + if (queue_idx >= bnxt_max_rings(bp)) { PMD_DRV_LOG(ERR, "Cannot create Tx ring %d. Only %d rings available\n", queue_idx, bp->max_tx_rings); return -EINVAL; } - if (!nb_desc || nb_desc > MAX_TX_DESC_CNT) { + if (nb_desc < BNXT_MIN_RING_DESC || nb_desc > MAX_TX_DESC_CNT) { PMD_DRV_LOG(ERR, "nb_desc %d is invalid", nb_desc); - rc = -EINVAL; - goto out; + return -EINVAL; } if (eth_dev->data->tx_queues) { @@ -114,8 +123,7 @@ int bnxt_tx_queue_setup_op(struct rte_eth_dev *eth_dev, RTE_CACHE_LINE_SIZE, socket_id); if (!txq) { PMD_DRV_LOG(ERR, "bnxt_tx_queue allocation failed!"); - rc = -ENOMEM; - goto out; + return -ENOMEM; } txq->free = rte_zmalloc_socket(NULL, @@ -123,36 +131,37 @@ int bnxt_tx_queue_setup_op(struct rte_eth_dev *eth_dev, RTE_CACHE_LINE_SIZE, socket_id); if (!txq->free) { PMD_DRV_LOG(ERR, "allocation of tx mbuf free array failed!"); - rte_free(txq); rc = -ENOMEM; - goto out; + goto err; } txq->bp = bp; txq->nb_tx_desc = nb_desc; - txq->tx_free_thresh = tx_conf->tx_free_thresh; + txq->tx_free_thresh = + RTE_MIN(rte_align32pow2(nb_desc) / 4, RTE_BNXT_MAX_TX_BURST); + txq->offloads = eth_dev->data->dev_conf.txmode.offloads | + tx_conf->offloads; + txq->tx_deferred_start = tx_conf->tx_deferred_start; rc = bnxt_init_tx_ring_struct(txq, socket_id); if (rc) - goto out; + goto err; txq->queue_id = queue_idx; txq->port_id = eth_dev->data->port_id; /* Allocate TX ring hardware descriptors */ - if (bnxt_alloc_rings(bp, queue_idx, txq, NULL, txq->cp_ring, NULL, - "txr")) { + if (bnxt_alloc_rings(bp, socket_id, queue_idx, txq, NULL, txq->cp_ring, + NULL, "txr")) { PMD_DRV_LOG(ERR, "ring_dma_zone_reserve for tx_ring failed!"); - bnxt_tx_queue_release_op(txq); rc = -ENOMEM; - goto out; + goto err; } if (bnxt_init_one_tx_ring(txq)) { PMD_DRV_LOG(ERR, "bnxt_init_one_tx_ring failed!"); - bnxt_tx_queue_release_op(txq); rc = -ENOMEM; - goto out; + goto err; } eth_dev->data->tx_queues[queue_idx] = txq; @@ -161,6 +170,9 @@ int bnxt_tx_queue_setup_op(struct rte_eth_dev *eth_dev, txq->tx_started = false; else txq->tx_started = true; -out: + + return 0; +err: + bnxt_tx_queue_release_op(txq); return rc; }