/* SPDX-License-Identifier: BSD-3-Clause
- * Copyright(c) 2014-2018 Broadcom
+ * Copyright(c) 2014-2021 Broadcom
* All rights reserved.
*/
#include <rte_malloc.h>
#include "bnxt.h"
-#include "bnxt_cpr.h"
#include "bnxt_ring.h"
#include "bnxt_txq.h"
#include "bnxt_txr.h"
struct bnxt_sw_tx_bd *sw_ring;
uint16_t i;
- if (!txq)
+ if (!txq || !txq->tx_ring)
return;
sw_ring = txq->tx_ring->tx_buf_ring;
if (sw_ring) {
for (i = 0; i < txq->tx_ring->tx_ring_struct->ring_size; i++) {
if (sw_ring[i].mbuf) {
- rte_pktmbuf_free(sw_ring[i].mbuf);
+ rte_pktmbuf_free_seg(sw_ring[i].mbuf);
sw_ring[i].mbuf = NULL;
}
}
/* Free TX ring hardware descriptors */
bnxt_tx_queue_release_mbufs(txq);
- bnxt_free_ring(txq->tx_ring->tx_ring_struct);
+ if (txq->tx_ring) {
+ bnxt_free_ring(txq->tx_ring->tx_ring_struct);
+ rte_free(txq->tx_ring->tx_ring_struct);
+ rte_free(txq->tx_ring);
+ }
/* Free TX completion ring hardware descriptors */
- bnxt_free_ring(txq->cp_ring->cp_ring_struct);
+ if (txq->cp_ring) {
+ bnxt_free_ring(txq->cp_ring->cp_ring_struct);
+ rte_free(txq->cp_ring->cp_ring_struct);
+ rte_free(txq->cp_ring);
+ }
bnxt_free_txq_stats(txq);
rte_memzone_free(txq->mz);
if (rc)
return rc;
- if (queue_idx >= bp->max_tx_rings) {
+ if (queue_idx >= bnxt_max_rings(bp)) {
PMD_DRV_LOG(ERR,
"Cannot create Tx ring %d. Only %d rings available\n",
queue_idx, bp->max_tx_rings);
return -EINVAL;
}
- if (!nb_desc || nb_desc > MAX_TX_DESC_CNT) {
+ if (nb_desc < BNXT_MIN_RING_DESC || nb_desc > MAX_TX_DESC_CNT) {
PMD_DRV_LOG(ERR, "nb_desc %d is invalid", nb_desc);
- rc = -EINVAL;
- goto out;
+ return -EINVAL;
}
if (eth_dev->data->tx_queues) {
RTE_CACHE_LINE_SIZE, socket_id);
if (!txq) {
PMD_DRV_LOG(ERR, "bnxt_tx_queue allocation failed!");
- rc = -ENOMEM;
- goto out;
+ return -ENOMEM;
}
txq->free = rte_zmalloc_socket(NULL,
RTE_CACHE_LINE_SIZE, socket_id);
if (!txq->free) {
PMD_DRV_LOG(ERR, "allocation of tx mbuf free array failed!");
- rte_free(txq);
rc = -ENOMEM;
- goto out;
+ goto err;
}
txq->bp = bp;
txq->nb_tx_desc = nb_desc;
- txq->tx_free_thresh = tx_conf->tx_free_thresh;
+ txq->tx_free_thresh =
+ RTE_MIN(rte_align32pow2(nb_desc) / 4, RTE_BNXT_MAX_TX_BURST);
+ txq->offloads = eth_dev->data->dev_conf.txmode.offloads |
+ tx_conf->offloads;
+
txq->tx_deferred_start = tx_conf->tx_deferred_start;
rc = bnxt_init_tx_ring_struct(txq, socket_id);
if (rc)
- goto out;
+ goto err;
txq->queue_id = queue_idx;
txq->port_id = eth_dev->data->port_id;
if (bnxt_alloc_rings(bp, queue_idx, txq, NULL, txq->cp_ring, NULL,
"txr")) {
PMD_DRV_LOG(ERR, "ring_dma_zone_reserve for tx_ring failed!");
- bnxt_tx_queue_release_op(txq);
rc = -ENOMEM;
- goto out;
+ goto err;
}
if (bnxt_init_one_tx_ring(txq)) {
PMD_DRV_LOG(ERR, "bnxt_init_one_tx_ring failed!");
- bnxt_tx_queue_release_op(txq);
rc = -ENOMEM;
- goto out;
+ goto err;
}
eth_dev->data->tx_queues[queue_idx] = txq;
txq->tx_started = false;
else
txq->tx_started = true;
-out:
+
+ return 0;
+err:
+ bnxt_tx_queue_release_op(txq);
return rc;
}