ethdev: introduce IP reassembly offload
[dpdk.git] / lib / ethdev / rte_ethdev.c
index 982c1bb..1f63201 100644 (file)
@@ -50,10 +50,10 @@ struct rte_eth_fp_ops rte_eth_fp_ops[RTE_MAX_ETHPORTS];
 /* spinlock for eth device callbacks */
 static rte_spinlock_t eth_dev_cb_lock = RTE_SPINLOCK_INITIALIZER;
 
-/* spinlock for add/remove rx callbacks */
+/* spinlock for add/remove Rx callbacks */
 static rte_spinlock_t eth_dev_rx_cb_lock = RTE_SPINLOCK_INITIALIZER;
 
-/* spinlock for add/remove tx callbacks */
+/* spinlock for add/remove Tx callbacks */
 static rte_spinlock_t eth_dev_tx_cb_lock = RTE_SPINLOCK_INITIALIZER;
 
 /* spinlock for shared data allocation */
@@ -101,9 +101,6 @@ static const struct rte_eth_xstats_name_off eth_dev_txq_stats_strings[] = {
 #define RTE_NB_TXQ_STATS RTE_DIM(eth_dev_txq_stats_strings)
 
 #define RTE_RX_OFFLOAD_BIT2STR(_name)  \
-       { DEV_RX_OFFLOAD_##_name, #_name }
-
-#define RTE_ETH_RX_OFFLOAD_BIT2STR(_name)      \
        { RTE_ETH_RX_OFFLOAD_##_name, #_name }
 
 static const struct {
@@ -128,14 +125,14 @@ static const struct {
        RTE_RX_OFFLOAD_BIT2STR(SCTP_CKSUM),
        RTE_RX_OFFLOAD_BIT2STR(OUTER_UDP_CKSUM),
        RTE_RX_OFFLOAD_BIT2STR(RSS_HASH),
-       RTE_ETH_RX_OFFLOAD_BIT2STR(BUFFER_SPLIT),
+       RTE_RX_OFFLOAD_BIT2STR(BUFFER_SPLIT),
 };
 
 #undef RTE_RX_OFFLOAD_BIT2STR
 #undef RTE_ETH_RX_OFFLOAD_BIT2STR
 
 #define RTE_TX_OFFLOAD_BIT2STR(_name)  \
-       { DEV_TX_OFFLOAD_##_name, #_name }
+       { RTE_ETH_TX_OFFLOAD_##_name, #_name }
 
 static const struct {
        uint64_t offload;
@@ -167,6 +164,17 @@ static const struct {
 
 #undef RTE_TX_OFFLOAD_BIT2STR
 
+static const struct {
+       uint64_t offload;
+       const char *name;
+} rte_eth_dev_capa_names[] = {
+       {RTE_ETH_DEV_CAPA_RUNTIME_RX_QUEUE_SETUP, "RUNTIME_RX_QUEUE_SETUP"},
+       {RTE_ETH_DEV_CAPA_RUNTIME_TX_QUEUE_SETUP, "RUNTIME_TX_QUEUE_SETUP"},
+       {RTE_ETH_DEV_CAPA_RXQ_SHARE, "RXQ_SHARE"},
+       {RTE_ETH_DEV_CAPA_FLOW_RULE_KEEP, "FLOW_RULE_KEEP"},
+       {RTE_ETH_DEV_CAPA_FLOW_SHARED_OBJECT_KEEP, "FLOW_SHARED_OBJECT_KEEP"},
+};
+
 /**
  * The user application callback description.
  *
@@ -538,7 +546,7 @@ unlock:
 
 /*
  * Attach to a port already registered by the primary process, which
- * makes sure that the same device would have the same port id both
+ * makes sure that the same device would have the same port ID both
  * in the primary and secondary process.
  */
 struct rte_eth_dev *
@@ -668,7 +676,7 @@ eth_dev_owner_set(const uint16_t port_id, const uint64_t old_owner_id,
        struct rte_eth_dev_owner *port_owner;
 
        if (port_id >= RTE_MAX_ETHPORTS || !eth_dev_is_allocated(ethdev)) {
-               RTE_ETHDEV_LOG(ERR, "Port id %"PRIu16" is not allocated\n",
+               RTE_ETHDEV_LOG(ERR, "Port ID %"PRIu16" is not allocated\n",
                        port_id);
                return -ENODEV;
        }
@@ -751,16 +759,19 @@ rte_eth_dev_owner_delete(const uint64_t owner_id)
        rte_spinlock_lock(&eth_dev_shared_data->ownership_lock);
 
        if (eth_is_valid_owner_id(owner_id)) {
-               for (port_id = 0; port_id < RTE_MAX_ETHPORTS; port_id++)
-                       if (rte_eth_devices[port_id].data->owner.id == owner_id)
-                               memset(&rte_eth_devices[port_id].data->owner, 0,
+               for (port_id = 0; port_id < RTE_MAX_ETHPORTS; port_id++) {
+                       struct rte_eth_dev_data *data =
+                               rte_eth_devices[port_id].data;
+                       if (data != NULL && data->owner.id == owner_id)
+                               memset(&data->owner, 0,
                                       sizeof(struct rte_eth_dev_owner));
+               }
                RTE_ETHDEV_LOG(NOTICE,
                        "All port owners owned by %016"PRIx64" identifier have removed\n",
                        owner_id);
        } else {
                RTE_ETHDEV_LOG(ERR,
-                              "Invalid owner id=%016"PRIx64"\n",
+                              "Invalid owner ID=%016"PRIx64"\n",
                               owner_id);
                ret = -EINVAL;
        }
@@ -779,7 +790,7 @@ rte_eth_dev_owner_get(const uint16_t port_id, struct rte_eth_dev_owner *owner)
        ethdev = &rte_eth_devices[port_id];
 
        if (!eth_dev_is_allocated(ethdev)) {
-               RTE_ETHDEV_LOG(ERR, "Port id %"PRIu16" is not allocated\n",
+               RTE_ETHDEV_LOG(ERR, "Port ID %"PRIu16" is not allocated\n",
                        port_id);
                return -ENODEV;
        }
@@ -883,6 +894,17 @@ rte_eth_dev_get_port_by_name(const char *name, uint16_t *port_id)
        return -ENODEV;
 }
 
+struct rte_eth_dev *
+rte_eth_dev_get_by_name(const char *name)
+{
+       uint16_t pid;
+
+       if (rte_eth_dev_get_port_by_name(name, &pid))
+               return NULL;
+
+       return &rte_eth_devices[pid];
+}
+
 static int
 eth_err(uint16_t port_id, int ret)
 {
@@ -1173,32 +1195,32 @@ uint32_t
 rte_eth_speed_bitflag(uint32_t speed, int duplex)
 {
        switch (speed) {
-       case ETH_SPEED_NUM_10M:
-               return duplex ? ETH_LINK_SPEED_10M : ETH_LINK_SPEED_10M_HD;
-       case ETH_SPEED_NUM_100M:
-               return duplex ? ETH_LINK_SPEED_100M : ETH_LINK_SPEED_100M_HD;
-       case ETH_SPEED_NUM_1G:
-               return ETH_LINK_SPEED_1G;
-       case ETH_SPEED_NUM_2_5G:
-               return ETH_LINK_SPEED_2_5G;
-       case ETH_SPEED_NUM_5G:
-               return ETH_LINK_SPEED_5G;
-       case ETH_SPEED_NUM_10G:
-               return ETH_LINK_SPEED_10G;
-       case ETH_SPEED_NUM_20G:
-               return ETH_LINK_SPEED_20G;
-       case ETH_SPEED_NUM_25G:
-               return ETH_LINK_SPEED_25G;
-       case ETH_SPEED_NUM_40G:
-               return ETH_LINK_SPEED_40G;
-       case ETH_SPEED_NUM_50G:
-               return ETH_LINK_SPEED_50G;
-       case ETH_SPEED_NUM_56G:
-               return ETH_LINK_SPEED_56G;
-       case ETH_SPEED_NUM_100G:
-               return ETH_LINK_SPEED_100G;
-       case ETH_SPEED_NUM_200G:
-               return ETH_LINK_SPEED_200G;
+       case RTE_ETH_SPEED_NUM_10M:
+               return duplex ? RTE_ETH_LINK_SPEED_10M : RTE_ETH_LINK_SPEED_10M_HD;
+       case RTE_ETH_SPEED_NUM_100M:
+               return duplex ? RTE_ETH_LINK_SPEED_100M : RTE_ETH_LINK_SPEED_100M_HD;
+       case RTE_ETH_SPEED_NUM_1G:
+               return RTE_ETH_LINK_SPEED_1G;
+       case RTE_ETH_SPEED_NUM_2_5G:
+               return RTE_ETH_LINK_SPEED_2_5G;
+       case RTE_ETH_SPEED_NUM_5G:
+               return RTE_ETH_LINK_SPEED_5G;
+       case RTE_ETH_SPEED_NUM_10G:
+               return RTE_ETH_LINK_SPEED_10G;
+       case RTE_ETH_SPEED_NUM_20G:
+               return RTE_ETH_LINK_SPEED_20G;
+       case RTE_ETH_SPEED_NUM_25G:
+               return RTE_ETH_LINK_SPEED_25G;
+       case RTE_ETH_SPEED_NUM_40G:
+               return RTE_ETH_LINK_SPEED_40G;
+       case RTE_ETH_SPEED_NUM_50G:
+               return RTE_ETH_LINK_SPEED_50G;
+       case RTE_ETH_SPEED_NUM_56G:
+               return RTE_ETH_LINK_SPEED_56G;
+       case RTE_ETH_SPEED_NUM_100G:
+               return RTE_ETH_LINK_SPEED_100G;
+       case RTE_ETH_SPEED_NUM_200G:
+               return RTE_ETH_LINK_SPEED_200G;
        default:
                return 0;
        }
@@ -1236,6 +1258,22 @@ rte_eth_dev_tx_offload_name(uint64_t offload)
        return name;
 }
 
+const char *
+rte_eth_dev_capability_name(uint64_t capability)
+{
+       const char *name = "UNKNOWN";
+       unsigned int i;
+
+       for (i = 0; i < RTE_DIM(rte_eth_dev_capa_names); ++i) {
+               if (capability == rte_eth_dev_capa_names[i].offload) {
+                       name = rte_eth_dev_capa_names[i].name;
+                       break;
+               }
+       }
+
+       return name;
+}
+
 static inline int
 eth_dev_check_lro_pkt_size(uint16_t port_id, uint32_t config_size,
                   uint32_t max_rx_pkt_len, uint32_t dev_info_size)
@@ -1265,14 +1303,14 @@ eth_dev_check_lro_pkt_size(uint16_t port_id, uint32_t config_size,
 
 /*
  * Validate offloads that are requested through rte_eth_dev_configure against
- * the offloads successfully set by the ethernet device.
+ * the offloads successfully set by the Ethernet device.
  *
  * @param port_id
  *   The port identifier of the Ethernet device.
  * @param req_offloads
  *   The offloads that have been requested through `rte_eth_dev_configure`.
  * @param set_offloads
- *   The offloads successfully set by the ethernet device.
+ *   The offloads successfully set by the Ethernet device.
  * @param offload_type
  *   The offload type i.e. Rx/Tx string.
  * @param offload_name
@@ -1293,7 +1331,7 @@ eth_dev_validate_offloads(uint16_t port_id, uint64_t req_offloads,
 
        while (offloads_diff != 0) {
                /* Check if any offload is requested but not enabled. */
-               offload = 1ULL << __builtin_ctzll(offloads_diff);
+               offload = RTE_BIT64(__builtin_ctzll(offloads_diff));
                if (offload & req_offloads) {
                        RTE_ETHDEV_LOG(ERR,
                                "Port %u failed to enable %s offload %s\n",
@@ -1327,6 +1365,47 @@ eth_dev_get_overhead_len(uint32_t max_rx_pktlen, uint16_t max_mtu)
        return overhead_len;
 }
 
+/* rte_eth_dev_info_get() should be called prior to this function */
+static int
+eth_dev_validate_mtu(uint16_t port_id, struct rte_eth_dev_info *dev_info,
+               uint16_t mtu)
+{
+       uint32_t overhead_len;
+       uint32_t frame_size;
+
+       if (mtu < dev_info->min_mtu) {
+               RTE_ETHDEV_LOG(ERR,
+                       "MTU (%u) < device min MTU (%u) for port_id %u\n",
+                       mtu, dev_info->min_mtu, port_id);
+               return -EINVAL;
+       }
+       if (mtu > dev_info->max_mtu) {
+               RTE_ETHDEV_LOG(ERR,
+                       "MTU (%u) > device max MTU (%u) for port_id %u\n",
+                       mtu, dev_info->max_mtu, port_id);
+               return -EINVAL;
+       }
+
+       overhead_len = eth_dev_get_overhead_len(dev_info->max_rx_pktlen,
+                       dev_info->max_mtu);
+       frame_size = mtu + overhead_len;
+       if (frame_size < RTE_ETHER_MIN_LEN) {
+               RTE_ETHDEV_LOG(ERR,
+                       "Frame size (%u) < min frame size (%u) for port_id %u\n",
+                       frame_size, RTE_ETHER_MIN_LEN, port_id);
+               return -EINVAL;
+       }
+
+       if (frame_size > dev_info->max_rx_pktlen) {
+               RTE_ETHDEV_LOG(ERR,
+                       "Frame size (%u) > device max frame size (%u) for port_id %u\n",
+                       frame_size, dev_info->max_rx_pktlen, port_id);
+               return -EINVAL;
+       }
+
+       return 0;
+}
+
 int
 rte_eth_dev_configure(uint16_t port_id, uint16_t nb_rx_q, uint16_t nb_tx_q,
                      const struct rte_eth_conf *dev_conf)
@@ -1334,8 +1413,6 @@ rte_eth_dev_configure(uint16_t port_id, uint16_t nb_rx_q, uint16_t nb_tx_q,
        struct rte_eth_dev *dev;
        struct rte_eth_dev_info dev_info;
        struct rte_eth_conf orig_conf;
-       uint32_t max_rx_pktlen;
-       uint32_t overhead_len;
        int diag;
        int ret;
        uint16_t old_mtu;
@@ -1384,10 +1461,6 @@ rte_eth_dev_configure(uint16_t port_id, uint16_t nb_rx_q, uint16_t nb_tx_q,
        if (ret != 0)
                goto rollback;
 
-       /* Get the real Ethernet overhead length */
-       overhead_len = eth_dev_get_overhead_len(dev_info.max_rx_pktlen,
-                       dev_info.max_mtu);
-
        /* If number of queues specified by application for both Rx and Tx is
         * zero, use driver preferred values. This cannot be done individually
         * as it is valid for either Tx or Rx (but not both) to be zero.
@@ -1405,7 +1478,7 @@ rte_eth_dev_configure(uint16_t port_id, uint16_t nb_rx_q, uint16_t nb_tx_q,
 
        if (nb_rx_q > RTE_MAX_QUEUES_PER_PORT) {
                RTE_ETHDEV_LOG(ERR,
-                       "Number of RX queues requested (%u) is greater than max supported(%d)\n",
+                       "Number of Rx queues requested (%u) is greater than max supported(%d)\n",
                        nb_rx_q, RTE_MAX_QUEUES_PER_PORT);
                ret = -EINVAL;
                goto rollback;
@@ -1413,15 +1486,15 @@ rte_eth_dev_configure(uint16_t port_id, uint16_t nb_rx_q, uint16_t nb_tx_q,
 
        if (nb_tx_q > RTE_MAX_QUEUES_PER_PORT) {
                RTE_ETHDEV_LOG(ERR,
-                       "Number of TX queues requested (%u) is greater than max supported(%d)\n",
+                       "Number of Tx queues requested (%u) is greater than max supported(%d)\n",
                        nb_tx_q, RTE_MAX_QUEUES_PER_PORT);
                ret = -EINVAL;
                goto rollback;
        }
 
        /*
-        * Check that the numbers of RX and TX queues are not greater
-        * than the maximum number of RX and TX queues supported by the
+        * Check that the numbers of Rx and Tx queues are not greater
+        * than the maximum number of Rx and Tx queues supported by the
         * configured device.
         */
        if (nb_rx_q > dev_info.max_rx_queues) {
@@ -1454,26 +1527,13 @@ rte_eth_dev_configure(uint16_t port_id, uint16_t nb_rx_q, uint16_t nb_tx_q,
                goto rollback;
        }
 
-       /*
-        * Check that the maximum RX packet length is supported by the
-        * configured device.
-        */
        if (dev_conf->rxmode.mtu == 0)
                dev->data->dev_conf.rxmode.mtu = RTE_ETHER_MTU;
-       max_rx_pktlen = dev->data->dev_conf.rxmode.mtu + overhead_len;
-       if (max_rx_pktlen > dev_info.max_rx_pktlen) {
-               RTE_ETHDEV_LOG(ERR,
-                       "Ethdev port_id=%u max_rx_pktlen %u > max valid value %u\n",
-                       port_id, max_rx_pktlen, dev_info.max_rx_pktlen);
-               ret = -EINVAL;
-               goto rollback;
-       } else if (max_rx_pktlen < RTE_ETHER_MIN_LEN) {
-               RTE_ETHDEV_LOG(ERR,
-                       "Ethdev port_id=%u max_rx_pktlen %u < min valid value %u\n",
-                       port_id, max_rx_pktlen, RTE_ETHER_MIN_LEN);
-               ret = -EINVAL;
+
+       ret = eth_dev_validate_mtu(port_id, &dev_info,
+                       dev->data->dev_conf.rxmode.mtu);
+       if (ret != 0)
                goto rollback;
-       }
 
        dev->data->mtu = dev->data->dev_conf.rxmode.mtu;
 
@@ -1481,7 +1541,13 @@ rte_eth_dev_configure(uint16_t port_id, uint16_t nb_rx_q, uint16_t nb_tx_q,
         * If LRO is enabled, check that the maximum aggregated packet
         * size is supported by the configured device.
         */
-       if (dev_conf->rxmode.offloads & DEV_RX_OFFLOAD_TCP_LRO) {
+       if (dev_conf->rxmode.offloads & RTE_ETH_RX_OFFLOAD_TCP_LRO) {
+               uint32_t max_rx_pktlen;
+               uint32_t overhead_len;
+
+               overhead_len = eth_dev_get_overhead_len(dev_info.max_rx_pktlen,
+                               dev_info.max_mtu);
+               max_rx_pktlen = dev->data->dev_conf.rxmode.mtu + overhead_len;
                if (dev_conf->rxmode.max_lro_pkt_size == 0)
                        dev->data->dev_conf.rxmode.max_lro_pkt_size = max_rx_pktlen;
                ret = eth_dev_check_lro_pkt_size(port_id,
@@ -1532,18 +1598,18 @@ rte_eth_dev_configure(uint16_t port_id, uint16_t nb_rx_q, uint16_t nb_tx_q,
        }
 
        /* Check if Rx RSS distribution is disabled but RSS hash is enabled. */
-       if (((dev_conf->rxmode.mq_mode & ETH_MQ_RX_RSS_FLAG) == 0) &&
-           (dev_conf->rxmode.offloads & DEV_RX_OFFLOAD_RSS_HASH)) {
+       if (((dev_conf->rxmode.mq_mode & RTE_ETH_MQ_RX_RSS_FLAG) == 0) &&
+           (dev_conf->rxmode.offloads & RTE_ETH_RX_OFFLOAD_RSS_HASH)) {
                RTE_ETHDEV_LOG(ERR,
                        "Ethdev port_id=%u config invalid Rx mq_mode without RSS but %s offload is requested\n",
                        port_id,
-                       rte_eth_dev_rx_offload_name(DEV_RX_OFFLOAD_RSS_HASH));
+                       rte_eth_dev_rx_offload_name(RTE_ETH_RX_OFFLOAD_RSS_HASH));
                ret = -EINVAL;
                goto rollback;
        }
 
        /*
-        * Setup new number of RX/TX queues and reconfigure device.
+        * Setup new number of Rx/Tx queues and reconfigure device.
         */
        diag = eth_dev_rx_queue_config(dev, nb_rx_q);
        if (diag != 0) {
@@ -1659,7 +1725,7 @@ eth_dev_mac_restore(struct rte_eth_dev *dev,
                        pool_mask = dev->data->mac_pool_sel[i];
 
                        do {
-                               if (pool_mask & 1ULL)
+                               if (pool_mask & UINT64_C(1))
                                        (*dev->dev_ops->mac_addr_add)(dev,
                                                addr, i, pool);
                                pool_mask >>= 1;
@@ -1865,6 +1931,12 @@ rte_eth_dev_close(uint16_t port_id)
        RTE_ETH_VALID_PORTID_OR_ERR_RET(port_id, -ENODEV);
        dev = &rte_eth_devices[port_id];
 
+       if (dev->data->dev_started) {
+               RTE_ETHDEV_LOG(ERR, "Cannot close started device (port %u)\n",
+                              port_id);
+               return -EINVAL;
+       }
+
        RTE_FUNC_PTR_OR_ERR_RET(*dev->dev_ops->dev_close, -ENOTSUP);
        *lasterr = (*dev->dev_ops->dev_close)(dev);
        if (*lasterr != 0)
@@ -1941,7 +2013,7 @@ rte_eth_rx_queue_check_split(const struct rte_eth_rxseg_split *rx_seg,
         * for each segment specified in extended configuration.
         */
        mp_first = rx_seg[0].mp;
-       offset_mask = (1u << seg_capa->offset_align_log2) - 1;
+       offset_mask = RTE_BIT32(seg_capa->offset_align_log2) - 1;
        for (seg_idx = 0; seg_idx < n_seg; seg_idx++) {
                struct rte_mempool *mpl = rx_seg[seg_idx].mp;
                uint32_t length = rx_seg[seg_idx].length;
@@ -2007,7 +2079,7 @@ rte_eth_rx_queue_setup(uint16_t port_id, uint16_t rx_queue_id,
        dev = &rte_eth_devices[port_id];
 
        if (rx_queue_id >= dev->data->nb_rx_queues) {
-               RTE_ETHDEV_LOG(ERR, "Invalid RX queue_id=%u\n", rx_queue_id);
+               RTE_ETHDEV_LOG(ERR, "Invalid Rx queue_id=%u\n", rx_queue_id);
                return -EINVAL;
        }
 
@@ -2141,12 +2213,20 @@ rte_eth_rx_queue_setup(uint16_t port_id, uint16_t rx_queue_id,
                return -EINVAL;
        }
 
+       if (local_conf.share_group > 0 &&
+           (dev_info.dev_capa & RTE_ETH_DEV_CAPA_RXQ_SHARE) == 0) {
+               RTE_ETHDEV_LOG(ERR,
+                       "Ethdev port_id=%d rx_queue_id=%d, enabled share_group=%hu while device doesn't support Rx queue share\n",
+                       port_id, rx_queue_id, local_conf.share_group);
+               return -EINVAL;
+       }
+
        /*
         * If LRO is enabled, check that the maximum aggregated packet
         * size is supported by the configured device.
         */
        /* Get the real Ethernet overhead length */
-       if (local_conf.offloads & DEV_RX_OFFLOAD_TCP_LRO) {
+       if (local_conf.offloads & RTE_ETH_RX_OFFLOAD_TCP_LRO) {
                uint32_t overhead_len;
                uint32_t max_rx_pktlen;
                int ret;
@@ -2192,7 +2272,7 @@ rte_eth_rx_hairpin_queue_setup(uint16_t port_id, uint16_t rx_queue_id,
        dev = &rte_eth_devices[port_id];
 
        if (rx_queue_id >= dev->data->nb_rx_queues) {
-               RTE_ETHDEV_LOG(ERR, "Invalid RX queue_id=%u\n", rx_queue_id);
+               RTE_ETHDEV_LOG(ERR, "Invalid Rx queue_id=%u\n", rx_queue_id);
                return -EINVAL;
        }
 
@@ -2264,7 +2344,7 @@ rte_eth_tx_queue_setup(uint16_t port_id, uint16_t tx_queue_id,
        dev = &rte_eth_devices[port_id];
 
        if (tx_queue_id >= dev->data->nb_tx_queues) {
-               RTE_ETHDEV_LOG(ERR, "Invalid TX queue_id=%u\n", tx_queue_id);
+               RTE_ETHDEV_LOG(ERR, "Invalid Tx queue_id=%u\n", tx_queue_id);
                return -EINVAL;
        }
 
@@ -2358,7 +2438,7 @@ rte_eth_tx_hairpin_queue_setup(uint16_t port_id, uint16_t tx_queue_id,
        dev = &rte_eth_devices[port_id];
 
        if (tx_queue_id >= dev->data->nb_tx_queues) {
-               RTE_ETHDEV_LOG(ERR, "Invalid TX queue_id=%u\n", tx_queue_id);
+               RTE_ETHDEV_LOG(ERR, "Invalid Tx queue_id=%u\n", tx_queue_id);
                return -EINVAL;
        }
 
@@ -2726,21 +2806,21 @@ const char *
 rte_eth_link_speed_to_str(uint32_t link_speed)
 {
        switch (link_speed) {
-       case ETH_SPEED_NUM_NONE: return "None";
-       case ETH_SPEED_NUM_10M:  return "10 Mbps";
-       case ETH_SPEED_NUM_100M: return "100 Mbps";
-       case ETH_SPEED_NUM_1G:   return "1 Gbps";
-       case ETH_SPEED_NUM_2_5G: return "2.5 Gbps";
-       case ETH_SPEED_NUM_5G:   return "5 Gbps";
-       case ETH_SPEED_NUM_10G:  return "10 Gbps";
-       case ETH_SPEED_NUM_20G:  return "20 Gbps";
-       case ETH_SPEED_NUM_25G:  return "25 Gbps";
-       case ETH_SPEED_NUM_40G:  return "40 Gbps";
-       case ETH_SPEED_NUM_50G:  return "50 Gbps";
-       case ETH_SPEED_NUM_56G:  return "56 Gbps";
-       case ETH_SPEED_NUM_100G: return "100 Gbps";
-       case ETH_SPEED_NUM_200G: return "200 Gbps";
-       case ETH_SPEED_NUM_UNKNOWN: return "Unknown";
+       case RTE_ETH_SPEED_NUM_NONE: return "None";
+       case RTE_ETH_SPEED_NUM_10M:  return "10 Mbps";
+       case RTE_ETH_SPEED_NUM_100M: return "100 Mbps";
+       case RTE_ETH_SPEED_NUM_1G:   return "1 Gbps";
+       case RTE_ETH_SPEED_NUM_2_5G: return "2.5 Gbps";
+       case RTE_ETH_SPEED_NUM_5G:   return "5 Gbps";
+       case RTE_ETH_SPEED_NUM_10G:  return "10 Gbps";
+       case RTE_ETH_SPEED_NUM_20G:  return "20 Gbps";
+       case RTE_ETH_SPEED_NUM_25G:  return "25 Gbps";
+       case RTE_ETH_SPEED_NUM_40G:  return "40 Gbps";
+       case RTE_ETH_SPEED_NUM_50G:  return "50 Gbps";
+       case RTE_ETH_SPEED_NUM_56G:  return "56 Gbps";
+       case RTE_ETH_SPEED_NUM_100G: return "100 Gbps";
+       case RTE_ETH_SPEED_NUM_200G: return "200 Gbps";
+       case RTE_ETH_SPEED_NUM_UNKNOWN: return "Unknown";
        default: return "Invalid";
        }
 }
@@ -2764,14 +2844,14 @@ rte_eth_link_to_str(char *str, size_t len, const struct rte_eth_link *eth_link)
                return -EINVAL;
        }
 
-       if (eth_link->link_status == ETH_LINK_DOWN)
+       if (eth_link->link_status == RTE_ETH_LINK_DOWN)
                return snprintf(str, len, "Link down");
        else
                return snprintf(str, len, "Link up at %s %s %s",
                        rte_eth_link_speed_to_str(eth_link->link_speed),
-                       (eth_link->link_duplex == ETH_LINK_FULL_DUPLEX) ?
+                       (eth_link->link_duplex == RTE_ETH_LINK_FULL_DUPLEX) ?
                        "FDX" : "HDX",
-                       (eth_link->link_autoneg == ETH_LINK_AUTONEG) ?
+                       (eth_link->link_autoneg == RTE_ETH_LINK_AUTONEG) ?
                        "Autoneg" : "Fixed");
 }
 
@@ -3400,7 +3480,8 @@ rte_eth_dev_info_get(uint16_t port_id, struct rte_eth_dev_info *dev_info)
        dev_info->rx_desc_lim = lim;
        dev_info->tx_desc_lim = lim;
        dev_info->device = dev->device;
-       dev_info->min_mtu = RTE_ETHER_MIN_MTU;
+       dev_info->min_mtu = RTE_ETHER_MIN_LEN - RTE_ETHER_HDR_LEN -
+               RTE_ETHER_CRC_LEN;
        dev_info->max_mtu = UINT16_MAX;
 
        RTE_FUNC_PTR_OR_ERR_RET(*dev->dev_ops->dev_infos_get, -ENOTSUP);
@@ -3651,21 +3732,20 @@ rte_eth_dev_set_mtu(uint16_t port_id, uint16_t mtu)
         * which relies on dev->dev_ops->dev_infos_get.
         */
        if (*dev->dev_ops->dev_infos_get != NULL) {
-               uint16_t overhead_len;
-               uint32_t frame_size;
-
                ret = rte_eth_dev_info_get(port_id, &dev_info);
                if (ret != 0)
                        return ret;
 
-               if (mtu < dev_info.min_mtu || mtu > dev_info.max_mtu)
-                       return -EINVAL;
+               ret = eth_dev_validate_mtu(port_id, &dev_info, mtu);
+               if (ret != 0)
+                       return ret;
+       }
 
-               overhead_len = eth_dev_get_overhead_len(dev_info.max_rx_pktlen,
-                               dev_info.max_mtu);
-               frame_size = mtu + overhead_len;
-               if (mtu < RTE_ETHER_MIN_MTU || frame_size > dev_info.max_rx_pktlen)
-                       return -EINVAL;
+       if (dev->data->dev_configured == 0) {
+               RTE_ETHDEV_LOG(ERR,
+                       "Port %u must be configured before MTU set\n",
+                       port_id);
+               return -EINVAL;
        }
 
        ret = (*dev->dev_ops->mtu_set)(dev, mtu);
@@ -3685,8 +3765,8 @@ rte_eth_dev_vlan_filter(uint16_t port_id, uint16_t vlan_id, int on)
        dev = &rte_eth_devices[port_id];
 
        if (!(dev->data->dev_conf.rxmode.offloads &
-             DEV_RX_OFFLOAD_VLAN_FILTER)) {
-               RTE_ETHDEV_LOG(ERR, "Port %u: vlan-filtering disabled\n",
+             RTE_ETH_RX_OFFLOAD_VLAN_FILTER)) {
+               RTE_ETHDEV_LOG(ERR, "Port %u: VLAN-filtering disabled\n",
                        port_id);
                return -ENOSYS;
        }
@@ -3709,9 +3789,9 @@ rte_eth_dev_vlan_filter(uint16_t port_id, uint16_t vlan_id, int on)
                vbit = vlan_id % 64;
 
                if (on)
-                       vfc->ids[vidx] |= UINT64_C(1) << vbit;
+                       vfc->ids[vidx] |= RTE_BIT64(vbit);
                else
-                       vfc->ids[vidx] &= ~(UINT64_C(1) << vbit);
+                       vfc->ids[vidx] &= ~RTE_BIT64(vbit);
        }
 
        return eth_err(port_id, ret);
@@ -3772,44 +3852,44 @@ rte_eth_dev_set_vlan_offload(uint16_t port_id, int offload_mask)
        dev_offloads = orig_offloads;
 
        /* check which option changed by application */
-       cur = !!(offload_mask & ETH_VLAN_STRIP_OFFLOAD);
-       org = !!(dev_offloads & DEV_RX_OFFLOAD_VLAN_STRIP);
+       cur = !!(offload_mask & RTE_ETH_VLAN_STRIP_OFFLOAD);
+       org = !!(dev_offloads & RTE_ETH_RX_OFFLOAD_VLAN_STRIP);
        if (cur != org) {
                if (cur)
-                       dev_offloads |= DEV_RX_OFFLOAD_VLAN_STRIP;
+                       dev_offloads |= RTE_ETH_RX_OFFLOAD_VLAN_STRIP;
                else
-                       dev_offloads &= ~DEV_RX_OFFLOAD_VLAN_STRIP;
-               mask |= ETH_VLAN_STRIP_MASK;
+                       dev_offloads &= ~RTE_ETH_RX_OFFLOAD_VLAN_STRIP;
+               mask |= RTE_ETH_VLAN_STRIP_MASK;
        }
 
-       cur = !!(offload_mask & ETH_VLAN_FILTER_OFFLOAD);
-       org = !!(dev_offloads & DEV_RX_OFFLOAD_VLAN_FILTER);
+       cur = !!(offload_mask & RTE_ETH_VLAN_FILTER_OFFLOAD);
+       org = !!(dev_offloads & RTE_ETH_RX_OFFLOAD_VLAN_FILTER);
        if (cur != org) {
                if (cur)
-                       dev_offloads |= DEV_RX_OFFLOAD_VLAN_FILTER;
+                       dev_offloads |= RTE_ETH_RX_OFFLOAD_VLAN_FILTER;
                else
-                       dev_offloads &= ~DEV_RX_OFFLOAD_VLAN_FILTER;
-               mask |= ETH_VLAN_FILTER_MASK;
+                       dev_offloads &= ~RTE_ETH_RX_OFFLOAD_VLAN_FILTER;
+               mask |= RTE_ETH_VLAN_FILTER_MASK;
        }
 
-       cur = !!(offload_mask & ETH_VLAN_EXTEND_OFFLOAD);
-       org = !!(dev_offloads & DEV_RX_OFFLOAD_VLAN_EXTEND);
+       cur = !!(offload_mask & RTE_ETH_VLAN_EXTEND_OFFLOAD);
+       org = !!(dev_offloads & RTE_ETH_RX_OFFLOAD_VLAN_EXTEND);
        if (cur != org) {
                if (cur)
-                       dev_offloads |= DEV_RX_OFFLOAD_VLAN_EXTEND;
+                       dev_offloads |= RTE_ETH_RX_OFFLOAD_VLAN_EXTEND;
                else
-                       dev_offloads &= ~DEV_RX_OFFLOAD_VLAN_EXTEND;
-               mask |= ETH_VLAN_EXTEND_MASK;
+                       dev_offloads &= ~RTE_ETH_RX_OFFLOAD_VLAN_EXTEND;
+               mask |= RTE_ETH_VLAN_EXTEND_MASK;
        }
 
-       cur = !!(offload_mask & ETH_QINQ_STRIP_OFFLOAD);
-       org = !!(dev_offloads & DEV_RX_OFFLOAD_QINQ_STRIP);
+       cur = !!(offload_mask & RTE_ETH_QINQ_STRIP_OFFLOAD);
+       org = !!(dev_offloads & RTE_ETH_RX_OFFLOAD_QINQ_STRIP);
        if (cur != org) {
                if (cur)
-                       dev_offloads |= DEV_RX_OFFLOAD_QINQ_STRIP;
+                       dev_offloads |= RTE_ETH_RX_OFFLOAD_QINQ_STRIP;
                else
-                       dev_offloads &= ~DEV_RX_OFFLOAD_QINQ_STRIP;
-               mask |= ETH_QINQ_STRIP_MASK;
+                       dev_offloads &= ~RTE_ETH_RX_OFFLOAD_QINQ_STRIP;
+               mask |= RTE_ETH_QINQ_STRIP_MASK;
        }
 
        /*no change*/
@@ -3854,17 +3934,17 @@ rte_eth_dev_get_vlan_offload(uint16_t port_id)
        dev = &rte_eth_devices[port_id];
        dev_offloads = &dev->data->dev_conf.rxmode.offloads;
 
-       if (*dev_offloads & DEV_RX_OFFLOAD_VLAN_STRIP)
-               ret |= ETH_VLAN_STRIP_OFFLOAD;
+       if (*dev_offloads & RTE_ETH_RX_OFFLOAD_VLAN_STRIP)
+               ret |= RTE_ETH_VLAN_STRIP_OFFLOAD;
 
-       if (*dev_offloads & DEV_RX_OFFLOAD_VLAN_FILTER)
-               ret |= ETH_VLAN_FILTER_OFFLOAD;
+       if (*dev_offloads & RTE_ETH_RX_OFFLOAD_VLAN_FILTER)
+               ret |= RTE_ETH_VLAN_FILTER_OFFLOAD;
 
-       if (*dev_offloads & DEV_RX_OFFLOAD_VLAN_EXTEND)
-               ret |= ETH_VLAN_EXTEND_OFFLOAD;
+       if (*dev_offloads & RTE_ETH_RX_OFFLOAD_VLAN_EXTEND)
+               ret |= RTE_ETH_VLAN_EXTEND_OFFLOAD;
 
-       if (*dev_offloads & DEV_RX_OFFLOAD_QINQ_STRIP)
-               ret |= ETH_QINQ_STRIP_OFFLOAD;
+       if (*dev_offloads & RTE_ETH_RX_OFFLOAD_QINQ_STRIP)
+               ret |= RTE_ETH_QINQ_STRIP_OFFLOAD;
 
        return ret;
 }
@@ -3941,7 +4021,7 @@ rte_eth_dev_priority_flow_ctrl_set(uint16_t port_id,
                return -EINVAL;
        }
 
-       if (pfc_conf->priority > (ETH_DCB_NUM_USER_PRIORITIES - 1)) {
+       if (pfc_conf->priority > (RTE_ETH_DCB_NUM_USER_PRIORITIES - 1)) {
                RTE_ETHDEV_LOG(ERR, "Invalid priority, only 0-7 allowed\n");
                return -EINVAL;
        }
@@ -3953,13 +4033,156 @@ rte_eth_dev_priority_flow_ctrl_set(uint16_t port_id,
        return -ENOTSUP;
 }
 
+static int
+validate_rx_pause_config(struct rte_eth_dev_info *dev_info, uint8_t tc_max,
+               struct rte_eth_pfc_queue_conf *pfc_queue_conf)
+{
+       if ((pfc_queue_conf->mode == RTE_ETH_FC_RX_PAUSE) ||
+                       (pfc_queue_conf->mode == RTE_ETH_FC_FULL)) {
+               if (pfc_queue_conf->rx_pause.tx_qid >= dev_info->nb_tx_queues) {
+                       RTE_ETHDEV_LOG(ERR,
+                               "PFC Tx queue not in range for Rx pause requested:%d configured:%d\n",
+                               pfc_queue_conf->rx_pause.tx_qid,
+                               dev_info->nb_tx_queues);
+                       return -EINVAL;
+               }
+
+               if (pfc_queue_conf->rx_pause.tc >= tc_max) {
+                       RTE_ETHDEV_LOG(ERR,
+                               "PFC TC not in range for Rx pause requested:%d max:%d\n",
+                               pfc_queue_conf->rx_pause.tc, tc_max);
+                       return -EINVAL;
+               }
+       }
+
+       return 0;
+}
+
+static int
+validate_tx_pause_config(struct rte_eth_dev_info *dev_info, uint8_t tc_max,
+               struct rte_eth_pfc_queue_conf *pfc_queue_conf)
+{
+       if ((pfc_queue_conf->mode == RTE_ETH_FC_TX_PAUSE) ||
+                       (pfc_queue_conf->mode == RTE_ETH_FC_FULL)) {
+               if (pfc_queue_conf->tx_pause.rx_qid >= dev_info->nb_rx_queues) {
+                       RTE_ETHDEV_LOG(ERR,
+                               "PFC Rx queue not in range for Tx pause requested:%d configured:%d\n",
+                               pfc_queue_conf->tx_pause.rx_qid,
+                               dev_info->nb_rx_queues);
+                       return -EINVAL;
+               }
+
+               if (pfc_queue_conf->tx_pause.tc >= tc_max) {
+                       RTE_ETHDEV_LOG(ERR,
+                               "PFC TC not in range for Tx pause requested:%d max:%d\n",
+                               pfc_queue_conf->tx_pause.tc, tc_max);
+                       return -EINVAL;
+               }
+       }
+
+       return 0;
+}
+
+int
+rte_eth_dev_priority_flow_ctrl_queue_info_get(uint16_t port_id,
+               struct rte_eth_pfc_queue_info *pfc_queue_info)
+{
+       struct rte_eth_dev *dev;
+
+       RTE_ETH_VALID_PORTID_OR_ERR_RET(port_id, -ENODEV);
+       dev = &rte_eth_devices[port_id];
+
+       if (pfc_queue_info == NULL) {
+               RTE_ETHDEV_LOG(ERR, "PFC info param is NULL for port (%u)\n",
+                       port_id);
+               return -EINVAL;
+       }
+
+       if (*dev->dev_ops->priority_flow_ctrl_queue_info_get)
+               return eth_err(port_id, (*dev->dev_ops->priority_flow_ctrl_queue_info_get)
+                       (dev, pfc_queue_info));
+       return -ENOTSUP;
+}
+
+int
+rte_eth_dev_priority_flow_ctrl_queue_configure(uint16_t port_id,
+               struct rte_eth_pfc_queue_conf *pfc_queue_conf)
+{
+       struct rte_eth_pfc_queue_info pfc_info;
+       struct rte_eth_dev_info dev_info;
+       struct rte_eth_dev *dev;
+       int ret;
+
+       RTE_ETH_VALID_PORTID_OR_ERR_RET(port_id, -ENODEV);
+       dev = &rte_eth_devices[port_id];
+
+       if (pfc_queue_conf == NULL) {
+               RTE_ETHDEV_LOG(ERR, "PFC parameters are NULL for port (%u)\n",
+                       port_id);
+               return -EINVAL;
+       }
+
+       ret = rte_eth_dev_info_get(port_id, &dev_info);
+       if (ret != 0)
+               return ret;
+
+       ret = rte_eth_dev_priority_flow_ctrl_queue_info_get(port_id, &pfc_info);
+       if (ret != 0)
+               return ret;
+
+       if (pfc_info.tc_max == 0) {
+               RTE_ETHDEV_LOG(ERR, "Ethdev port %u does not support PFC TC values\n",
+                       port_id);
+               return -ENOTSUP;
+       }
+
+       /* Check requested mode supported or not */
+       if (pfc_info.mode_capa == RTE_ETH_FC_RX_PAUSE &&
+                       pfc_queue_conf->mode == RTE_ETH_FC_TX_PAUSE) {
+               RTE_ETHDEV_LOG(ERR, "PFC Tx pause unsupported for port (%d)\n",
+                       port_id);
+               return -EINVAL;
+       }
+
+       if (pfc_info.mode_capa == RTE_ETH_FC_TX_PAUSE &&
+                       pfc_queue_conf->mode == RTE_ETH_FC_RX_PAUSE) {
+               RTE_ETHDEV_LOG(ERR, "PFC Rx pause unsupported for port (%d)\n",
+                       port_id);
+               return -EINVAL;
+       }
+
+       /* Validate Rx pause parameters */
+       if (pfc_info.mode_capa == RTE_ETH_FC_FULL ||
+                       pfc_info.mode_capa == RTE_ETH_FC_RX_PAUSE) {
+               ret = validate_rx_pause_config(&dev_info, pfc_info.tc_max,
+                               pfc_queue_conf);
+               if (ret != 0)
+                       return ret;
+       }
+
+       /* Validate Tx pause parameters */
+       if (pfc_info.mode_capa == RTE_ETH_FC_FULL ||
+                       pfc_info.mode_capa == RTE_ETH_FC_TX_PAUSE) {
+               ret = validate_tx_pause_config(&dev_info, pfc_info.tc_max,
+                               pfc_queue_conf);
+               if (ret != 0)
+                       return ret;
+       }
+
+       if (*dev->dev_ops->priority_flow_ctrl_queue_config)
+               return eth_err(port_id,
+                              (*dev->dev_ops->priority_flow_ctrl_queue_config)(
+                               dev, pfc_queue_conf));
+       return -ENOTSUP;
+}
+
 static int
 eth_check_reta_mask(struct rte_eth_rss_reta_entry64 *reta_conf,
                        uint16_t reta_size)
 {
        uint16_t i, num;
 
-       num = (reta_size + RTE_RETA_GROUP_SIZE - 1) / RTE_RETA_GROUP_SIZE;
+       num = (reta_size + RTE_ETH_RETA_GROUP_SIZE - 1) / RTE_ETH_RETA_GROUP_SIZE;
        for (i = 0; i < num; i++) {
                if (reta_conf[i].mask)
                        return 0;
@@ -3981,9 +4204,9 @@ eth_check_reta_entry(struct rte_eth_rss_reta_entry64 *reta_conf,
        }
 
        for (i = 0; i < reta_size; i++) {
-               idx = i / RTE_RETA_GROUP_SIZE;
-               shift = i % RTE_RETA_GROUP_SIZE;
-               if ((reta_conf[idx].mask & (1ULL << shift)) &&
+               idx = i / RTE_ETH_RETA_GROUP_SIZE;
+               shift = i % RTE_ETH_RETA_GROUP_SIZE;
+               if ((reta_conf[idx].mask & RTE_BIT64(shift)) &&
                        (reta_conf[idx].reta[shift] >= max_rxq)) {
                        RTE_ETHDEV_LOG(ERR,
                                "reta_conf[%u]->reta[%u]: %u exceeds the maximum rxq index: %u\n",
@@ -4138,7 +4361,7 @@ rte_eth_dev_udp_tunnel_port_add(uint16_t port_id,
                return -EINVAL;
        }
 
-       if (udp_tunnel->prot_type >= RTE_TUNNEL_TYPE_MAX) {
+       if (udp_tunnel->prot_type >= RTE_ETH_TUNNEL_TYPE_MAX) {
                RTE_ETHDEV_LOG(ERR, "Invalid tunnel type\n");
                return -EINVAL;
        }
@@ -4164,7 +4387,7 @@ rte_eth_dev_udp_tunnel_port_delete(uint16_t port_id,
                return -EINVAL;
        }
 
-       if (udp_tunnel->prot_type >= RTE_TUNNEL_TYPE_MAX) {
+       if (udp_tunnel->prot_type >= RTE_ETH_TUNNEL_TYPE_MAX) {
                RTE_ETHDEV_LOG(ERR, "Invalid tunnel type\n");
                return -EINVAL;
        }
@@ -4305,8 +4528,8 @@ rte_eth_dev_mac_addr_add(uint16_t port_id, struct rte_ether_addr *addr,
                        port_id);
                return -EINVAL;
        }
-       if (pool >= ETH_64_POOLS) {
-               RTE_ETHDEV_LOG(ERR, "Pool id must be 0-%d\n", ETH_64_POOLS - 1);
+       if (pool >= RTE_ETH_64_POOLS) {
+               RTE_ETHDEV_LOG(ERR, "Pool ID must be 0-%d\n", RTE_ETH_64_POOLS - 1);
                return -EINVAL;
        }
 
@@ -4322,7 +4545,7 @@ rte_eth_dev_mac_addr_add(uint16_t port_id, struct rte_ether_addr *addr,
                pool_mask = dev->data->mac_pool_sel[index];
 
                /* Check if both MAC address and pool is already there, and do nothing */
-               if (pool_mask & (1ULL << pool))
+               if (pool_mask & RTE_BIT64(pool))
                        return 0;
        }
 
@@ -4334,7 +4557,7 @@ rte_eth_dev_mac_addr_add(uint16_t port_id, struct rte_ether_addr *addr,
                rte_ether_addr_copy(addr, &dev->data->mac_addrs[index]);
 
                /* Update pool bitmap in NIC data structure */
-               dev->data->mac_pool_sel[index] |= (1ULL << pool);
+               dev->data->mac_pool_sel[index] |= RTE_BIT64(pool);
        }
 
        return eth_err(port_id, ret);
@@ -4531,7 +4754,7 @@ int rte_eth_set_queue_rate_limit(uint16_t port_id, uint16_t queue_idx,
 
        if (queue_idx > dev_info.max_tx_queues) {
                RTE_ETHDEV_LOG(ERR,
-                       "Set queue rate limit:port %u: invalid queue id=%u\n",
+                       "Set queue rate limit:port %u: invalid queue ID=%u\n",
                        port_id, queue_idx);
                return -EINVAL;
        }
@@ -4750,22 +4973,22 @@ rte_eth_dev_rx_intr_ctl(uint16_t port_id, int epfd, int op, void *data)
        dev = &rte_eth_devices[port_id];
 
        if (!dev->intr_handle) {
-               RTE_ETHDEV_LOG(ERR, "RX Intr handle unset\n");
+               RTE_ETHDEV_LOG(ERR, "Rx Intr handle unset\n");
                return -ENOTSUP;
        }
 
        intr_handle = dev->intr_handle;
-       if (!intr_handle->intr_vec) {
-               RTE_ETHDEV_LOG(ERR, "RX Intr vector unset\n");
+       if (rte_intr_vec_list_index_get(intr_handle, 0) < 0) {
+               RTE_ETHDEV_LOG(ERR, "Rx Intr vector unset\n");
                return -EPERM;
        }
 
        for (qid = 0; qid < dev->data->nb_rx_queues; qid++) {
-               vec = intr_handle->intr_vec[qid];
+               vec = rte_intr_vec_list_index_get(intr_handle, qid);
                rc = rte_intr_rx_ctl(intr_handle, epfd, op, vec, data);
                if (rc && rc != -EEXIST) {
                        RTE_ETHDEV_LOG(ERR,
-                               "p %u q %u rx ctl error op %d epfd %d vec %u\n",
+                               "p %u q %u Rx ctl error op %d epfd %d vec %u\n",
                                port_id, qid, op, epfd, vec);
                }
        }
@@ -4786,25 +5009,25 @@ rte_eth_dev_rx_intr_ctl_q_get_fd(uint16_t port_id, uint16_t queue_id)
        dev = &rte_eth_devices[port_id];
 
        if (queue_id >= dev->data->nb_rx_queues) {
-               RTE_ETHDEV_LOG(ERR, "Invalid RX queue_id=%u\n", queue_id);
+               RTE_ETHDEV_LOG(ERR, "Invalid Rx queue_id=%u\n", queue_id);
                return -1;
        }
 
        if (!dev->intr_handle) {
-               RTE_ETHDEV_LOG(ERR, "RX Intr handle unset\n");
+               RTE_ETHDEV_LOG(ERR, "Rx Intr handle unset\n");
                return -1;
        }
 
        intr_handle = dev->intr_handle;
-       if (!intr_handle->intr_vec) {
-               RTE_ETHDEV_LOG(ERR, "RX Intr vector unset\n");
+       if (rte_intr_vec_list_index_get(intr_handle, 0) < 0) {
+               RTE_ETHDEV_LOG(ERR, "Rx Intr vector unset\n");
                return -1;
        }
 
-       vec = intr_handle->intr_vec[queue_id];
+       vec = rte_intr_vec_list_index_get(intr_handle, queue_id);
        efd_idx = (vec >= RTE_INTR_VEC_RXTX_OFFSET) ?
                (vec - RTE_INTR_VEC_RXTX_OFFSET) : vec;
-       fd = intr_handle->efds[efd_idx];
+       fd = rte_intr_efds_index_get(intr_handle, efd_idx);
 
        return fd;
 }
@@ -4972,26 +5195,26 @@ rte_eth_dev_rx_intr_ctl_q(uint16_t port_id, uint16_t queue_id,
        dev = &rte_eth_devices[port_id];
 
        if (queue_id >= dev->data->nb_rx_queues) {
-               RTE_ETHDEV_LOG(ERR, "Invalid RX queue_id=%u\n", queue_id);
+               RTE_ETHDEV_LOG(ERR, "Invalid Rx queue_id=%u\n", queue_id);
                return -EINVAL;
        }
 
        if (!dev->intr_handle) {
-               RTE_ETHDEV_LOG(ERR, "RX Intr handle unset\n");
+               RTE_ETHDEV_LOG(ERR, "Rx Intr handle unset\n");
                return -ENOTSUP;
        }
 
        intr_handle = dev->intr_handle;
-       if (!intr_handle->intr_vec) {
-               RTE_ETHDEV_LOG(ERR, "RX Intr vector unset\n");
+       if (rte_intr_vec_list_index_get(intr_handle, 0) < 0) {
+               RTE_ETHDEV_LOG(ERR, "Rx Intr vector unset\n");
                return -EPERM;
        }
 
-       vec = intr_handle->intr_vec[queue_id];
+       vec = rte_intr_vec_list_index_get(intr_handle, queue_id);
        rc = rte_intr_rx_ctl(intr_handle, epfd, op, vec, data);
        if (rc && rc != -EEXIST) {
                RTE_ETHDEV_LOG(ERR,
-                       "p %u q %u rx ctl error op %d epfd %d vec %u\n",
+                       "p %u q %u Rx ctl error op %d epfd %d vec %u\n",
                        port_id, queue_id, op, epfd, vec);
                return rc;
        }
@@ -5269,7 +5492,7 @@ rte_eth_rx_queue_info_get(uint16_t port_id, uint16_t queue_id,
        dev = &rte_eth_devices[port_id];
 
        if (queue_id >= dev->data->nb_rx_queues) {
-               RTE_ETHDEV_LOG(ERR, "Invalid RX queue_id=%u\n", queue_id);
+               RTE_ETHDEV_LOG(ERR, "Invalid Rx queue_id=%u\n", queue_id);
                return -EINVAL;
        }
 
@@ -5314,7 +5537,7 @@ rte_eth_tx_queue_info_get(uint16_t port_id, uint16_t queue_id,
        dev = &rte_eth_devices[port_id];
 
        if (queue_id >= dev->data->nb_tx_queues) {
-               RTE_ETHDEV_LOG(ERR, "Invalid TX queue_id=%u\n", queue_id);
+               RTE_ETHDEV_LOG(ERR, "Invalid Tx queue_id=%u\n", queue_id);
                return -EINVAL;
        }
 
@@ -5359,7 +5582,7 @@ rte_eth_rx_burst_mode_get(uint16_t port_id, uint16_t queue_id,
        dev = &rte_eth_devices[port_id];
 
        if (queue_id >= dev->data->nb_rx_queues) {
-               RTE_ETHDEV_LOG(ERR, "Invalid RX queue_id=%u\n", queue_id);
+               RTE_ETHDEV_LOG(ERR, "Invalid Rx queue_id=%u\n", queue_id);
                return -EINVAL;
        }
 
@@ -5386,7 +5609,7 @@ rte_eth_tx_burst_mode_get(uint16_t port_id, uint16_t queue_id,
        dev = &rte_eth_devices[port_id];
 
        if (queue_id >= dev->data->nb_tx_queues) {
-               RTE_ETHDEV_LOG(ERR, "Invalid TX queue_id=%u\n", queue_id);
+               RTE_ETHDEV_LOG(ERR, "Invalid Tx queue_id=%u\n", queue_id);
                return -EINVAL;
        }
 
@@ -5967,8 +6190,7 @@ rte_eth_devargs_parse(const char *dargs, struct rte_eth_devargs *eth_da)
        }
 
 parse_cleanup:
-       if (args.str)
-               free(args.str);
+       free(args.str);
 
        return result;
 }
@@ -6215,7 +6437,7 @@ eth_dev_handle_port_link_status(const char *cmd __rte_unused,
        rte_tel_data_add_dict_string(d, status_str, "UP");
        rte_tel_data_add_dict_u64(d, "speed", link.link_speed);
        rte_tel_data_add_dict_string(d, "duplex",
-                       (link.link_duplex == ETH_LINK_FULL_DUPLEX) ?
+                       (link.link_duplex == RTE_ETH_LINK_FULL_DUPLEX) ?
                                "full-duplex" : "half-duplex");
        return 0;
 }
@@ -6251,8 +6473,10 @@ eth_dev_handle_port_info(const char *cmd __rte_unused,
                return -ENOMEM;
 
        txq_state = rte_tel_data_alloc();
-       if (!txq_state)
+       if (!txq_state) {
+               rte_tel_data_free(rxq_state);
                return -ENOMEM;
+       }
 
        rte_tel_data_start_dict(d);
        rte_tel_data_add_dict_string(d, "name", eth_dev->data->name);
@@ -6388,7 +6612,7 @@ rte_eth_rx_metadata_negotiate(uint16_t port_id, uint64_t *features)
 
        if (dev->data->dev_configured != 0) {
                RTE_ETHDEV_LOG(ERR,
-                       "The port (id=%"PRIu16") is already configured\n",
+                       "The port (ID=%"PRIu16") is already configured\n",
                        port_id);
                return -EBUSY;
        }
@@ -6403,6 +6627,102 @@ rte_eth_rx_metadata_negotiate(uint16_t port_id, uint64_t *features)
                       (*dev->dev_ops->rx_metadata_negotiate)(dev, features));
 }
 
+int
+rte_eth_ip_reassembly_capability_get(uint16_t port_id,
+               struct rte_eth_ip_reassembly_params *reassembly_capa)
+{
+       struct rte_eth_dev *dev;
+
+       RTE_ETH_VALID_PORTID_OR_ERR_RET(port_id, -ENODEV);
+       dev = &rte_eth_devices[port_id];
+
+       if (dev->data->dev_configured == 0) {
+               RTE_ETHDEV_LOG(ERR,
+                       "Device with port_id=%u is not configured.\n"
+                       "Cannot get IP reassembly capability\n",
+                       port_id);
+               return -EINVAL;
+       }
+
+       if (reassembly_capa == NULL) {
+               RTE_ETHDEV_LOG(ERR, "Cannot get reassembly capability to NULL");
+               return -EINVAL;
+       }
+
+       RTE_FUNC_PTR_OR_ERR_RET(*dev->dev_ops->ip_reassembly_capability_get,
+                               -ENOTSUP);
+       memset(reassembly_capa, 0, sizeof(struct rte_eth_ip_reassembly_params));
+
+       return eth_err(port_id, (*dev->dev_ops->ip_reassembly_capability_get)
+                                       (dev, reassembly_capa));
+}
+
+int
+rte_eth_ip_reassembly_conf_get(uint16_t port_id,
+               struct rte_eth_ip_reassembly_params *conf)
+{
+       struct rte_eth_dev *dev;
+
+       RTE_ETH_VALID_PORTID_OR_ERR_RET(port_id, -ENODEV);
+       dev = &rte_eth_devices[port_id];
+
+       if (dev->data->dev_configured == 0) {
+               RTE_ETHDEV_LOG(ERR,
+                       "Device with port_id=%u is not configured.\n"
+                       "Cannot get IP reassembly configuration\n",
+                       port_id);
+               return -EINVAL;
+       }
+
+       if (conf == NULL) {
+               RTE_ETHDEV_LOG(ERR, "Cannot get reassembly info to NULL");
+               return -EINVAL;
+       }
+
+       RTE_FUNC_PTR_OR_ERR_RET(*dev->dev_ops->ip_reassembly_conf_get,
+                               -ENOTSUP);
+       memset(conf, 0, sizeof(struct rte_eth_ip_reassembly_params));
+       return eth_err(port_id,
+                      (*dev->dev_ops->ip_reassembly_conf_get)(dev, conf));
+}
+
+int
+rte_eth_ip_reassembly_conf_set(uint16_t port_id,
+               const struct rte_eth_ip_reassembly_params *conf)
+{
+       struct rte_eth_dev *dev;
+
+       RTE_ETH_VALID_PORTID_OR_ERR_RET(port_id, -ENODEV);
+       dev = &rte_eth_devices[port_id];
+
+       if (dev->data->dev_configured == 0) {
+               RTE_ETHDEV_LOG(ERR,
+                       "Device with port_id=%u is not configured.\n"
+                       "Cannot set IP reassembly configuration",
+                       port_id);
+               return -EINVAL;
+       }
+
+       if (dev->data->dev_started != 0) {
+               RTE_ETHDEV_LOG(ERR,
+                       "Device with port_id=%u started,\n"
+                       "cannot configure IP reassembly params.\n",
+                       port_id);
+               return -EINVAL;
+       }
+
+       if (conf == NULL) {
+               RTE_ETHDEV_LOG(ERR,
+                               "Invalid IP reassembly configuration (NULL)\n");
+               return -EINVAL;
+       }
+
+       RTE_FUNC_PTR_OR_ERR_RET(*dev->dev_ops->ip_reassembly_conf_set,
+                               -ENOTSUP);
+       return eth_err(port_id,
+                      (*dev->dev_ops->ip_reassembly_conf_set)(dev, conf));
+}
+
 RTE_LOG_REGISTER_DEFAULT(rte_eth_dev_logtype, INFO);
 
 RTE_INIT(ethdev_init_telemetry)