ethdev: add namespace
[dpdk.git] / drivers / net / bnxt / bnxt_reps.c
index e5ba090..1c07db3 100644 (file)
@@ -1,5 +1,5 @@
 /* SPDX-License-Identifier: BSD-3-Clause
- * Copyright(c) 2014-2020 Broadcom
+ * Copyright(c) 2014-2021 Broadcom
  * All rights reserved.
  */
 
@@ -29,7 +29,7 @@ static const struct eth_dev_ops bnxt_rep_dev_ops = {
        .dev_stop = bnxt_rep_dev_stop_op,
        .stats_get = bnxt_rep_stats_get_op,
        .stats_reset = bnxt_rep_stats_reset_op,
-       .filter_ctrl = bnxt_filter_ctrl_op
+       .flow_ops_get = bnxt_flow_ops_get_op
 };
 
 uint16_t
@@ -55,17 +55,17 @@ bnxt_vfr_recv(uint16_t port_id, uint16_t queue_id, struct rte_mbuf *mbuf)
        mask = rep_rxr->rx_ring_struct->ring_mask;
 
        /* Put this mbuf on the RxQ of the Representor */
-       prod_rx_buf = &rep_rxr->rx_buf_ring[rep_rxr->rx_prod & mask];
-       if (!*prod_rx_buf) {
+       prod_rx_buf = &rep_rxr->rx_buf_ring[rep_rxr->rx_raw_prod & mask];
+       if (*prod_rx_buf == NULL) {
                *prod_rx_buf = mbuf;
                vfr_bp->rx_bytes[que] += mbuf->pkt_len;
                vfr_bp->rx_pkts[que]++;
-               rep_rxr->rx_prod++;
+               rep_rxr->rx_raw_prod++;
        } else {
                /* Representor Rx ring full, drop pkt */
                vfr_bp->rx_drop_bytes[que] += mbuf->pkt_len;
                vfr_bp->rx_drop_pkts[que]++;
-               rte_pktmbuf_free(mbuf);
+               rte_mbuf_raw_free(mbuf);
        }
 
        return 0;
@@ -104,7 +104,7 @@ bnxt_rep_rx_burst(void *rx_queue,
 static uint16_t
 bnxt_rep_tx_burst(void *tx_queue,
                     struct rte_mbuf **tx_pkts,
-                    __rte_unused uint16_t nb_pkts)
+                    uint16_t nb_pkts)
 {
        struct bnxt_vf_rep_tx_queue *vfr_txq = tx_queue;
        struct bnxt_tx_queue *ptxq;
@@ -187,6 +187,7 @@ int bnxt_representor_init(struct rte_eth_dev *eth_dev, void *params)
        eth_dev->data->dev_flags |= RTE_ETH_DEV_REPRESENTOR |
                                        RTE_ETH_DEV_AUTOFILL_QUEUE_XSTATS;
        eth_dev->data->representor_id = rep_params->vf_id;
+       eth_dev->data->backer_port_id = rep_params->parent_dev->data->port_id;
 
        rte_eth_random_addr(vf_rep_bp->dflt_mac_addr);
        memcpy(vf_rep_bp->mac_addr, vf_rep_bp->dflt_mac_addr,
@@ -536,7 +537,7 @@ int bnxt_rep_dev_info_get_op(struct rte_eth_dev *eth_dev,
 
        dev_info->rx_offload_capa = BNXT_DEV_RX_OFFLOAD_SUPPORT;
        if (parent_bp->flags & BNXT_FLAG_PTP_SUPPORTED)
-               dev_info->rx_offload_capa |= DEV_RX_OFFLOAD_TIMESTAMP;
+               dev_info->rx_offload_capa |= RTE_ETH_RX_OFFLOAD_TIMESTAMP;
        dev_info->tx_offload_capa = BNXT_DEV_TX_OFFLOAD_SUPPORT;
        dev_info->flow_type_rss_offloads = BNXT_ETH_RSS_SUPPORT;
 
@@ -548,7 +549,7 @@ int bnxt_rep_dev_info_get_op(struct rte_eth_dev *eth_dev,
        return 0;
 }
 
-int bnxt_rep_dev_configure_op(__rte_unused struct rte_eth_dev *eth_dev)
+int bnxt_rep_dev_configure_op(struct rte_eth_dev *eth_dev)
 {
        struct bnxt_representor *rep_bp = eth_dev->data->dev_private;
 
@@ -630,7 +631,7 @@ int bnxt_rep_rx_queue_setup_op(struct rte_eth_dev *eth_dev,
        if (eth_dev->data->rx_queues) {
                rxq = eth_dev->data->rx_queues[queue_idx];
                if (rxq)
-                       bnxt_rx_queue_release_op(rxq);
+                       bnxt_rx_queue_release_op(eth_dev, queue_idx);
        }
 
        rxq = rte_zmalloc_socket("bnxt_vfr_rx_queue",
@@ -641,6 +642,8 @@ int bnxt_rep_rx_queue_setup_op(struct rte_eth_dev *eth_dev,
                return -ENOMEM;
        }
 
+       eth_dev->data->rx_queues[queue_idx] = rxq;
+
        rxq->nb_rx_desc = nb_desc;
 
        rc = bnxt_init_rep_rx_ring(rxq, socket_id);
@@ -660,20 +663,19 @@ int bnxt_rep_rx_queue_setup_op(struct rte_eth_dev *eth_dev,
        rxq->rx_ring->rx_buf_ring = buf_ring;
        rxq->queue_id = queue_idx;
        rxq->port_id = eth_dev->data->port_id;
-       eth_dev->data->rx_queues[queue_idx] = rxq;
 
        return 0;
 
 out:
        if (rxq)
-               bnxt_rep_rx_queue_release_op(rxq);
+               bnxt_rep_rx_queue_release_op(eth_dev, queue_idx);
 
        return rc;
 }
 
-void bnxt_rep_rx_queue_release_op(void *rx_queue)
+void bnxt_rep_rx_queue_release_op(struct rte_eth_dev *dev, uint16_t queue_idx)
 {
-       struct bnxt_rx_queue *rxq = (struct bnxt_rx_queue *)rx_queue;
+       struct bnxt_rx_queue *rxq = dev->data->rx_queues[queue_idx];
 
        if (!rxq)
                return;
@@ -728,8 +730,8 @@ int bnxt_rep_tx_queue_setup_op(struct rte_eth_dev *eth_dev,
 
        if (eth_dev->data->tx_queues) {
                vfr_txq = eth_dev->data->tx_queues[queue_idx];
-               bnxt_rep_tx_queue_release_op(vfr_txq);
-               vfr_txq = NULL;
+               if (vfr_txq != NULL)
+                       bnxt_rep_tx_queue_release_op(eth_dev, queue_idx);
        }
 
        vfr_txq = rte_zmalloc_socket("bnxt_vfr_tx_queue",
@@ -758,15 +760,16 @@ int bnxt_rep_tx_queue_setup_op(struct rte_eth_dev *eth_dev,
        return 0;
 }
 
-void bnxt_rep_tx_queue_release_op(void *tx_queue)
+void bnxt_rep_tx_queue_release_op(struct rte_eth_dev *dev, uint16_t queue_idx)
 {
-       struct bnxt_vf_rep_tx_queue *vfr_txq = tx_queue;
+       struct bnxt_vf_rep_tx_queue *vfr_txq = dev->data->tx_queues[queue_idx];
 
        if (!vfr_txq)
                return;
 
        rte_free(vfr_txq->txq);
        rte_free(vfr_txq);
+       dev->data->tx_queues[queue_idx] = NULL;
 }
 
 int bnxt_rep_stats_get_op(struct rte_eth_dev *eth_dev,