ethdev: add tracepoints
[dpdk.git] / lib / librte_ethdev / rte_ethdev.c
index df7e991..72aed59 100644 (file)
@@ -37,8 +37,9 @@
 #include <rte_string_fns.h>
 #include <rte_kvargs.h>
 #include <rte_class.h>
+#include <rte_ether.h>
 
-#include "rte_ether.h"
+#include "rte_ethdev_trace.h"
 #include "rte_ethdev.h"
 #include "rte_ethdev_driver.h"
 #include "ethdev_profile.h"
@@ -86,7 +87,7 @@ static const struct rte_eth_xstats_name_off rte_stats_strings[] = {
                rx_nombuf)},
 };
 
-#define RTE_NB_STATS (sizeof(rte_stats_strings) / sizeof(rte_stats_strings[0]))
+#define RTE_NB_STATS RTE_DIM(rte_stats_strings)
 
 static const struct rte_eth_xstats_name_off rte_rxq_stats_strings[] = {
        {"packets", offsetof(struct rte_eth_stats, q_ipackets)},
@@ -94,15 +95,13 @@ static const struct rte_eth_xstats_name_off rte_rxq_stats_strings[] = {
        {"errors", offsetof(struct rte_eth_stats, q_errors)},
 };
 
-#define RTE_NB_RXQ_STATS (sizeof(rte_rxq_stats_strings) /      \
-               sizeof(rte_rxq_stats_strings[0]))
+#define RTE_NB_RXQ_STATS RTE_DIM(rte_rxq_stats_strings)
 
 static const struct rte_eth_xstats_name_off rte_txq_stats_strings[] = {
        {"packets", offsetof(struct rte_eth_stats, q_opackets)},
        {"bytes", offsetof(struct rte_eth_stats, q_obytes)},
 };
-#define RTE_NB_TXQ_STATS (sizeof(rte_txq_stats_strings) /      \
-               sizeof(rte_txq_stats_strings[0]))
+#define RTE_NB_TXQ_STATS RTE_DIM(rte_txq_stats_strings)
 
 #define RTE_RX_OFFLOAD_BIT2STR(_name)  \
        { DEV_RX_OFFLOAD_##_name, #_name }
@@ -129,6 +128,7 @@ static const struct {
        RTE_RX_OFFLOAD_BIT2STR(KEEP_CRC),
        RTE_RX_OFFLOAD_BIT2STR(SCTP_CKSUM),
        RTE_RX_OFFLOAD_BIT2STR(OUTER_UDP_CKSUM),
+       RTE_RX_OFFLOAD_BIT2STR(RSS_HASH),
 };
 
 #undef RTE_RX_OFFLOAD_BIT2STR
@@ -161,7 +161,6 @@ static const struct {
        RTE_TX_OFFLOAD_BIT2STR(UDP_TNL_TSO),
        RTE_TX_OFFLOAD_BIT2STR(IP_TNL_TSO),
        RTE_TX_OFFLOAD_BIT2STR(OUTER_UDP_CKSUM),
-       RTE_TX_OFFLOAD_BIT2STR(MATCH_METADATA),
 };
 
 #undef RTE_TX_OFFLOAD_BIT2STR
@@ -753,12 +752,6 @@ rte_eth_dev_get_sec_ctx(uint16_t port_id)
        return rte_eth_devices[port_id].security_ctx;
 }
 
-uint16_t
-rte_eth_dev_count(void)
-{
-       return rte_eth_dev_count_avail();
-}
-
 uint16_t
 rte_eth_dev_count_avail(void)
 {
@@ -904,6 +897,13 @@ rte_eth_dev_rx_queue_start(uint16_t port_id, uint16_t rx_queue_id)
 
        RTE_FUNC_PTR_OR_ERR_RET(*dev->dev_ops->rx_queue_start, -ENOTSUP);
 
+       if (rte_eth_dev_is_rx_hairpin_queue(dev, rx_queue_id)) {
+               RTE_ETHDEV_LOG(INFO,
+                       "Can't start Rx hairpin queue %"PRIu16" of device with port_id=%"PRIu16"\n",
+                       rx_queue_id, port_id);
+               return -EINVAL;
+       }
+
        if (dev->data->rx_queue_state[rx_queue_id] != RTE_ETH_QUEUE_STATE_STOPPED) {
                RTE_ETHDEV_LOG(INFO,
                        "Queue %"PRIu16" of device with port_id=%"PRIu16" already started\n",
@@ -931,6 +931,13 @@ rte_eth_dev_rx_queue_stop(uint16_t port_id, uint16_t rx_queue_id)
 
        RTE_FUNC_PTR_OR_ERR_RET(*dev->dev_ops->rx_queue_stop, -ENOTSUP);
 
+       if (rte_eth_dev_is_rx_hairpin_queue(dev, rx_queue_id)) {
+               RTE_ETHDEV_LOG(INFO,
+                       "Can't stop Rx hairpin queue %"PRIu16" of device with port_id=%"PRIu16"\n",
+                       rx_queue_id, port_id);
+               return -EINVAL;
+       }
+
        if (dev->data->rx_queue_state[rx_queue_id] == RTE_ETH_QUEUE_STATE_STOPPED) {
                RTE_ETHDEV_LOG(INFO,
                        "Queue %"PRIu16" of device with port_id=%"PRIu16" already stopped\n",
@@ -964,6 +971,13 @@ rte_eth_dev_tx_queue_start(uint16_t port_id, uint16_t tx_queue_id)
 
        RTE_FUNC_PTR_OR_ERR_RET(*dev->dev_ops->tx_queue_start, -ENOTSUP);
 
+       if (rte_eth_dev_is_tx_hairpin_queue(dev, tx_queue_id)) {
+               RTE_ETHDEV_LOG(INFO,
+                       "Can't start Tx hairpin queue %"PRIu16" of device with port_id=%"PRIu16"\n",
+                       tx_queue_id, port_id);
+               return -EINVAL;
+       }
+
        if (dev->data->tx_queue_state[tx_queue_id] != RTE_ETH_QUEUE_STATE_STOPPED) {
                RTE_ETHDEV_LOG(INFO,
                        "Queue %"PRIu16" of device with port_id=%"PRIu16" already started\n",
@@ -989,6 +1003,13 @@ rte_eth_dev_tx_queue_stop(uint16_t port_id, uint16_t tx_queue_id)
 
        RTE_FUNC_PTR_OR_ERR_RET(*dev->dev_ops->tx_queue_stop, -ENOTSUP);
 
+       if (rte_eth_dev_is_tx_hairpin_queue(dev, tx_queue_id)) {
+               RTE_ETHDEV_LOG(INFO,
+                       "Can't stop Tx hairpin queue %"PRIu16" of device with port_id=%"PRIu16"\n",
+                       tx_queue_id, port_id);
+               return -EINVAL;
+       }
+
        if (dev->data->tx_queue_state[tx_queue_id] == RTE_ETH_QUEUE_STATE_STOPPED) {
                RTE_ETHDEV_LOG(INFO,
                        "Queue %"PRIu16" of device with port_id=%"PRIu16" already stopped\n",
@@ -1115,6 +1136,84 @@ rte_eth_dev_tx_offload_name(uint64_t offload)
        return name;
 }
 
+static inline int
+check_lro_pkt_size(uint16_t port_id, uint32_t config_size,
+                  uint32_t max_rx_pkt_len, uint32_t dev_info_size)
+{
+       int ret = 0;
+
+       if (dev_info_size == 0) {
+               if (config_size != max_rx_pkt_len) {
+                       RTE_ETHDEV_LOG(ERR, "Ethdev port_id=%d max_lro_pkt_size"
+                                      " %u != %u is not allowed\n",
+                                      port_id, config_size, max_rx_pkt_len);
+                       ret = -EINVAL;
+               }
+       } else if (config_size > dev_info_size) {
+               RTE_ETHDEV_LOG(ERR, "Ethdev port_id=%d max_lro_pkt_size %u "
+                              "> max allowed value %u\n", port_id, config_size,
+                              dev_info_size);
+               ret = -EINVAL;
+       } else if (config_size < RTE_ETHER_MIN_LEN) {
+               RTE_ETHDEV_LOG(ERR, "Ethdev port_id=%d max_lro_pkt_size %u "
+                              "< min allowed value %u\n", port_id, config_size,
+                              (unsigned int)RTE_ETHER_MIN_LEN);
+               ret = -EINVAL;
+       }
+       return ret;
+}
+
+/*
+ * Validate offloads that are requested through rte_eth_dev_configure against
+ * the offloads successfully set by the ethernet device.
+ *
+ * @param port_id
+ *   The port identifier of the Ethernet device.
+ * @param req_offloads
+ *   The offloads that have been requested through `rte_eth_dev_configure`.
+ * @param set_offloads
+ *   The offloads successfully set by the ethernet device.
+ * @param offload_type
+ *   The offload type i.e. Rx/Tx string.
+ * @param offload_name
+ *   The function that prints the offload name.
+ * @return
+ *   - (0) if validation successful.
+ *   - (-EINVAL) if requested offload has been silently disabled.
+ *
+ */
+static int
+validate_offloads(uint16_t port_id, uint64_t req_offloads,
+                 uint64_t set_offloads, const char *offload_type,
+                 const char *(*offload_name)(uint64_t))
+{
+       uint64_t offloads_diff = req_offloads ^ set_offloads;
+       uint64_t offload;
+       int ret = 0;
+
+       while (offloads_diff != 0) {
+               /* Check if any offload is requested but not enabled. */
+               offload = 1ULL << __builtin_ctzll(offloads_diff);
+               if (offload & req_offloads) {
+                       RTE_ETHDEV_LOG(ERR,
+                               "Port %u failed to enable %s offload %s\n",
+                               port_id, offload_type, offload_name(offload));
+                       ret = -EINVAL;
+               }
+
+               /* Check if offload couldn't be disabled. */
+               if (offload & set_offloads) {
+                       RTE_ETHDEV_LOG(DEBUG,
+                               "Port %u %s offload %s is not requested but enabled\n",
+                               port_id, offload_type, offload_name(offload));
+               }
+
+               offloads_diff &= ~offload;
+       }
+
+       return ret;
+}
+
 int
 rte_eth_dev_configure(uint16_t port_id, uint16_t nb_rx_q, uint16_t nb_tx_q,
                      const struct rte_eth_conf *dev_conf)
@@ -1145,7 +1244,9 @@ rte_eth_dev_configure(uint16_t port_id, uint16_t nb_rx_q, uint16_t nb_tx_q,
         * Copy the dev_conf parameter into the dev structure.
         * rte_eth_dev_info_get() requires dev_conf, copy it before dev_info get
         */
-       memcpy(&dev->data->dev_conf, dev_conf, sizeof(dev->data->dev_conf));
+       if (dev_conf != &dev->data->dev_conf)
+               memcpy(&dev->data->dev_conf, dev_conf,
+                      sizeof(dev->data->dev_conf));
 
        ret = rte_eth_dev_info_get(port_id, &dev_info);
        if (ret != 0)
@@ -1245,6 +1346,22 @@ rte_eth_dev_configure(uint16_t port_id, uint16_t nb_rx_q, uint16_t nb_tx_q,
                                                        RTE_ETHER_MAX_LEN;
        }
 
+       /*
+        * If LRO is enabled, check that the maximum aggregated packet
+        * size is supported by the configured device.
+        */
+       if (dev_conf->rxmode.offloads & DEV_RX_OFFLOAD_TCP_LRO) {
+               if (dev_conf->rxmode.max_lro_pkt_size == 0)
+                       dev->data->dev_conf.rxmode.max_lro_pkt_size =
+                               dev->data->dev_conf.rxmode.max_rx_pkt_len;
+               ret = check_lro_pkt_size(port_id,
+                               dev->data->dev_conf.rxmode.max_lro_pkt_size,
+                               dev->data->dev_conf.rxmode.max_rx_pkt_len,
+                               dev_info.max_lro_pkt_size);
+               if (ret != 0)
+                       goto rollback;
+       }
+
        /* Any requested offloading must be within its device capabilities */
        if ((dev_conf->rxmode.offloads & dev_info.rx_offload_capa) !=
             dev_conf->rxmode.offloads) {
@@ -1269,6 +1386,9 @@ rte_eth_dev_configure(uint16_t port_id, uint16_t nb_rx_q, uint16_t nb_tx_q,
                goto rollback;
        }
 
+       dev->data->dev_conf.rx_adv_conf.rss_conf.rss_hf =
+               rte_eth_rss_hf_refine(dev_conf->rx_adv_conf.rss_conf.rss_hf);
+
        /* Check that device supports requested rss hash functions. */
        if ((dev_info.flow_type_rss_offloads |
             dev_conf->rx_adv_conf.rss_conf.rss_hf) !=
@@ -1281,6 +1401,17 @@ rte_eth_dev_configure(uint16_t port_id, uint16_t nb_rx_q, uint16_t nb_tx_q,
                goto rollback;
        }
 
+       /* Check if Rx RSS distribution is disabled but RSS hash is enabled. */
+       if (((dev_conf->rxmode.mq_mode & ETH_MQ_RX_RSS_FLAG) == 0) &&
+           (dev_conf->rxmode.offloads & DEV_RX_OFFLOAD_RSS_HASH)) {
+               RTE_ETHDEV_LOG(ERR,
+                       "Ethdev port_id=%u config invalid Rx mq_mode without RSS but %s offload is requested\n",
+                       port_id,
+                       rte_eth_dev_rx_offload_name(DEV_RX_OFFLOAD_RSS_HASH));
+               ret = -EINVAL;
+               goto rollback;
+       }
+
        /*
         * Setup new number of RX/TX queues and reconfigure device.
         */
@@ -1307,10 +1438,8 @@ rte_eth_dev_configure(uint16_t port_id, uint16_t nb_rx_q, uint16_t nb_tx_q,
        if (diag != 0) {
                RTE_ETHDEV_LOG(ERR, "Port%u dev_configure = %d\n",
                        port_id, diag);
-               rte_eth_dev_rx_queue_config(dev, 0);
-               rte_eth_dev_tx_queue_config(dev, 0);
                ret = eth_err(port_id, diag);
-               goto rollback;
+               goto reset_queues;
        }
 
        /* Initialize Rx profiling if enabled at compilation time. */
@@ -1318,17 +1447,39 @@ rte_eth_dev_configure(uint16_t port_id, uint16_t nb_rx_q, uint16_t nb_tx_q,
        if (diag != 0) {
                RTE_ETHDEV_LOG(ERR, "Port%u __rte_eth_dev_profile_init = %d\n",
                        port_id, diag);
-               rte_eth_dev_rx_queue_config(dev, 0);
-               rte_eth_dev_tx_queue_config(dev, 0);
                ret = eth_err(port_id, diag);
-               goto rollback;
+               goto reset_queues;
        }
 
-       return 0;
+       /* Validate Rx offloads. */
+       diag = validate_offloads(port_id,
+                       dev_conf->rxmode.offloads,
+                       dev->data->dev_conf.rxmode.offloads, "Rx",
+                       rte_eth_dev_rx_offload_name);
+       if (diag != 0) {
+               ret = diag;
+               goto reset_queues;
+       }
+
+       /* Validate Tx offloads. */
+       diag = validate_offloads(port_id,
+                       dev_conf->txmode.offloads,
+                       dev->data->dev_conf.txmode.offloads, "Tx",
+                       rte_eth_dev_tx_offload_name);
+       if (diag != 0) {
+               ret = diag;
+               goto reset_queues;
+       }
 
+       rte_ethdev_trace_configure(port_id, nb_rx_q, nb_tx_q, dev_conf, 0);
+       return 0;
+reset_queues:
+       rte_eth_dev_rx_queue_config(dev, 0);
+       rte_eth_dev_tx_queue_config(dev, 0);
 rollback:
        memcpy(&dev->data->dev_conf, &orig_conf, sizeof(dev->data->dev_conf));
 
+       rte_ethdev_trace_configure(port_id, nb_rx_q, nb_tx_q, dev_conf, ret);
        return ret;
 }
 
@@ -1422,16 +1573,24 @@ rte_eth_dev_config_restore(struct rte_eth_dev *dev,
        }
 
        /* replay all multicast configuration */
-       if (rte_eth_allmulticast_get(port_id) == 1) {
-               ret = rte_eth_allmulticast_enable(port_id);
+       /*
+        * use callbacks directly since we don't need port_id check and
+        * would like to bypass the same value set
+        */
+       if (rte_eth_allmulticast_get(port_id) == 1 &&
+           *dev->dev_ops->allmulticast_enable != NULL) {
+               ret = eth_err(port_id,
+                             (*dev->dev_ops->allmulticast_enable)(dev));
                if (ret != 0 && ret != -ENOTSUP) {
                        RTE_ETHDEV_LOG(ERR,
                                "Failed to enable allmulticast mode for device (port %u): %s\n",
                                port_id, rte_strerror(-ret));
                        return ret;
                }
-       } else if (rte_eth_allmulticast_get(port_id) == 0) {
-               ret = rte_eth_allmulticast_disable(port_id);
+       } else if (rte_eth_allmulticast_get(port_id) == 0 &&
+                  *dev->dev_ops->allmulticast_disable != NULL) {
+               ret = eth_err(port_id,
+                             (*dev->dev_ops->allmulticast_disable)(dev));
                if (ret != 0 && ret != -ENOTSUP) {
                        RTE_ETHDEV_LOG(ERR,
                                "Failed to disable allmulticast mode for device (port %u): %s\n",
@@ -1491,6 +1650,8 @@ rte_eth_dev_start(uint16_t port_id)
                RTE_FUNC_PTR_OR_ERR_RET(*dev->dev_ops->link_update, -ENOTSUP);
                (*dev->dev_ops->link_update)(dev, 0);
        }
+
+       rte_ethdev_trace_start(port_id);
        return 0;
 }
 
@@ -1513,6 +1674,7 @@ rte_eth_dev_stop(uint16_t port_id)
 
        dev->data->dev_started = 0;
        (*dev->dev_ops->dev_stop)(dev);
+       rte_ethdev_trace_stop(port_id);
 }
 
 int
@@ -1553,6 +1715,7 @@ rte_eth_dev_close(uint16_t port_id)
        dev->data->dev_started = 0;
        (*dev->dev_ops->dev_close)(dev);
 
+       rte_ethdev_trace_close(port_id);
        /* check behaviour flag - temporary for PMD migration */
        if ((dev->data->dev_flags & RTE_ETH_DEV_CLOSE_REMOVE) != 0) {
                /* new behaviour: send event + reset state + free all data */
@@ -1738,6 +1901,22 @@ rte_eth_rx_queue_setup(uint16_t port_id, uint16_t rx_queue_id,
                return -EINVAL;
        }
 
+       /*
+        * If LRO is enabled, check that the maximum aggregated packet
+        * size is supported by the configured device.
+        */
+       if (local_conf.offloads & DEV_RX_OFFLOAD_TCP_LRO) {
+               if (dev->data->dev_conf.rxmode.max_lro_pkt_size == 0)
+                       dev->data->dev_conf.rxmode.max_lro_pkt_size =
+                               dev->data->dev_conf.rxmode.max_rx_pkt_len;
+               int ret = check_lro_pkt_size(port_id,
+                               dev->data->dev_conf.rxmode.max_lro_pkt_size,
+                               dev->data->dev_conf.rxmode.max_rx_pkt_len,
+                               dev_info.max_lro_pkt_size);
+               if (ret != 0)
+                       return ret;
+       }
+
        ret = (*dev->dev_ops->rx_queue_setup)(dev, rx_queue_id, nb_rx_desc,
                                              socket_id, &local_conf, mp);
        if (!ret) {
@@ -1746,6 +1925,80 @@ rte_eth_rx_queue_setup(uint16_t port_id, uint16_t rx_queue_id,
                        dev->data->min_rx_buf_size = mbp_buf_size;
        }
 
+       rte_ethdev_trace_rxq_setup(port_id, rx_queue_id, nb_rx_desc, mp,
+               rx_conf, ret);
+       return eth_err(port_id, ret);
+}
+
+int
+rte_eth_rx_hairpin_queue_setup(uint16_t port_id, uint16_t rx_queue_id,
+                              uint16_t nb_rx_desc,
+                              const struct rte_eth_hairpin_conf *conf)
+{
+       int ret;
+       struct rte_eth_dev *dev;
+       struct rte_eth_hairpin_cap cap;
+       void **rxq;
+       int i;
+       int count;
+
+       RTE_ETH_VALID_PORTID_OR_ERR_RET(port_id, -EINVAL);
+
+       dev = &rte_eth_devices[port_id];
+       if (rx_queue_id >= dev->data->nb_rx_queues) {
+               RTE_ETHDEV_LOG(ERR, "Invalid RX queue_id=%u\n", rx_queue_id);
+               return -EINVAL;
+       }
+       ret = rte_eth_dev_hairpin_capability_get(port_id, &cap);
+       if (ret != 0)
+               return ret;
+       RTE_FUNC_PTR_OR_ERR_RET(*dev->dev_ops->rx_hairpin_queue_setup,
+                               -ENOTSUP);
+       /* if nb_rx_desc is zero use max number of desc from the driver. */
+       if (nb_rx_desc == 0)
+               nb_rx_desc = cap.max_nb_desc;
+       if (nb_rx_desc > cap.max_nb_desc) {
+               RTE_ETHDEV_LOG(ERR,
+                       "Invalid value for nb_rx_desc(=%hu), should be: <= %hu",
+                       nb_rx_desc, cap.max_nb_desc);
+               return -EINVAL;
+       }
+       if (conf->peer_count > cap.max_rx_2_tx) {
+               RTE_ETHDEV_LOG(ERR,
+                       "Invalid value for number of peers for Rx queue(=%hu), should be: <= %hu",
+                       conf->peer_count, cap.max_rx_2_tx);
+               return -EINVAL;
+       }
+       if (conf->peer_count == 0) {
+               RTE_ETHDEV_LOG(ERR,
+                       "Invalid value for number of peers for Rx queue(=%hu), should be: > 0",
+                       conf->peer_count);
+               return -EINVAL;
+       }
+       for (i = 0, count = 0; i < dev->data->nb_rx_queues &&
+            cap.max_nb_queues != UINT16_MAX; i++) {
+               if (i == rx_queue_id || rte_eth_dev_is_rx_hairpin_queue(dev, i))
+                       count++;
+       }
+       if (count > cap.max_nb_queues) {
+               RTE_ETHDEV_LOG(ERR, "To many Rx hairpin queues max is %d",
+               cap.max_nb_queues);
+               return -EINVAL;
+       }
+       if (dev->data->dev_started)
+               return -EBUSY;
+       rxq = dev->data->rx_queues;
+       if (rxq[rx_queue_id] != NULL) {
+               RTE_FUNC_PTR_OR_ERR_RET(*dev->dev_ops->rx_queue_release,
+                                       -ENOTSUP);
+               (*dev->dev_ops->rx_queue_release)(rxq[rx_queue_id]);
+               rxq[rx_queue_id] = NULL;
+       }
+       ret = (*dev->dev_ops->rx_hairpin_queue_setup)(dev, rx_queue_id,
+                                                     nb_rx_desc, conf);
+       if (ret == 0)
+               dev->data->rx_queue_state[rx_queue_id] =
+                       RTE_ETH_QUEUE_STATE_HAIRPIN;
        return eth_err(port_id, ret);
 }
 
@@ -1844,10 +2097,82 @@ rte_eth_tx_queue_setup(uint16_t port_id, uint16_t tx_queue_id,
                return -EINVAL;
        }
 
+       rte_ethdev_trace_txq_setup(port_id, tx_queue_id, nb_tx_desc, tx_conf);
        return eth_err(port_id, (*dev->dev_ops->tx_queue_setup)(dev,
                       tx_queue_id, nb_tx_desc, socket_id, &local_conf));
 }
 
+int
+rte_eth_tx_hairpin_queue_setup(uint16_t port_id, uint16_t tx_queue_id,
+                              uint16_t nb_tx_desc,
+                              const struct rte_eth_hairpin_conf *conf)
+{
+       struct rte_eth_dev *dev;
+       struct rte_eth_hairpin_cap cap;
+       void **txq;
+       int i;
+       int count;
+       int ret;
+
+       RTE_ETH_VALID_PORTID_OR_ERR_RET(port_id, -EINVAL);
+       dev = &rte_eth_devices[port_id];
+       if (tx_queue_id >= dev->data->nb_tx_queues) {
+               RTE_ETHDEV_LOG(ERR, "Invalid TX queue_id=%u\n", tx_queue_id);
+               return -EINVAL;
+       }
+       ret = rte_eth_dev_hairpin_capability_get(port_id, &cap);
+       if (ret != 0)
+               return ret;
+       RTE_FUNC_PTR_OR_ERR_RET(*dev->dev_ops->tx_hairpin_queue_setup,
+                               -ENOTSUP);
+       /* if nb_rx_desc is zero use max number of desc from the driver. */
+       if (nb_tx_desc == 0)
+               nb_tx_desc = cap.max_nb_desc;
+       if (nb_tx_desc > cap.max_nb_desc) {
+               RTE_ETHDEV_LOG(ERR,
+                       "Invalid value for nb_tx_desc(=%hu), should be: <= %hu",
+                       nb_tx_desc, cap.max_nb_desc);
+               return -EINVAL;
+       }
+       if (conf->peer_count > cap.max_tx_2_rx) {
+               RTE_ETHDEV_LOG(ERR,
+                       "Invalid value for number of peers for Tx queue(=%hu), should be: <= %hu",
+                       conf->peer_count, cap.max_tx_2_rx);
+               return -EINVAL;
+       }
+       if (conf->peer_count == 0) {
+               RTE_ETHDEV_LOG(ERR,
+                       "Invalid value for number of peers for Tx queue(=%hu), should be: > 0",
+                       conf->peer_count);
+               return -EINVAL;
+       }
+       for (i = 0, count = 0; i < dev->data->nb_tx_queues &&
+            cap.max_nb_queues != UINT16_MAX; i++) {
+               if (i == tx_queue_id || rte_eth_dev_is_tx_hairpin_queue(dev, i))
+                       count++;
+       }
+       if (count > cap.max_nb_queues) {
+               RTE_ETHDEV_LOG(ERR, "To many Tx hairpin queues max is %d",
+               cap.max_nb_queues);
+               return -EINVAL;
+       }
+       if (dev->data->dev_started)
+               return -EBUSY;
+       txq = dev->data->tx_queues;
+       if (txq[tx_queue_id] != NULL) {
+               RTE_FUNC_PTR_OR_ERR_RET(*dev->dev_ops->tx_queue_release,
+                                       -ENOTSUP);
+               (*dev->dev_ops->tx_queue_release)(txq[tx_queue_id]);
+               txq[tx_queue_id] = NULL;
+       }
+       ret = (*dev->dev_ops->tx_hairpin_queue_setup)
+               (dev, tx_queue_id, nb_tx_desc, conf);
+       if (ret == 0)
+               dev->data->tx_queue_state[tx_queue_id] =
+                       RTE_ETH_QUEUE_STATE_HAIRPIN;
+       return eth_err(port_id, ret);
+}
+
 void
 rte_eth_tx_buffer_drop_callback(struct rte_mbuf **pkts, uint16_t unsent,
                void *userdata __rte_unused)
@@ -1922,12 +2247,13 @@ rte_eth_promiscuous_enable(uint16_t port_id)
        RTE_ETH_VALID_PORTID_OR_ERR_RET(port_id, -ENODEV);
        dev = &rte_eth_devices[port_id];
 
+       if (dev->data->promiscuous == 1)
+               return 0;
+
        RTE_FUNC_PTR_OR_ERR_RET(*dev->dev_ops->promiscuous_enable, -ENOTSUP);
 
-       if (dev->data->promiscuous == 0) {
-               diag = (*dev->dev_ops->promiscuous_enable)(dev);
-               dev->data->promiscuous = (diag == 0) ? 1 : 0;
-       }
+       diag = (*dev->dev_ops->promiscuous_enable)(dev);
+       dev->data->promiscuous = (diag == 0) ? 1 : 0;
 
        return eth_err(port_id, diag);
 }
@@ -1941,14 +2267,15 @@ rte_eth_promiscuous_disable(uint16_t port_id)
        RTE_ETH_VALID_PORTID_OR_ERR_RET(port_id, -ENODEV);
        dev = &rte_eth_devices[port_id];
 
+       if (dev->data->promiscuous == 0)
+               return 0;
+
        RTE_FUNC_PTR_OR_ERR_RET(*dev->dev_ops->promiscuous_disable, -ENOTSUP);
 
-       if (dev->data->promiscuous == 1) {
-               dev->data->promiscuous = 0;
-               diag = (*dev->dev_ops->promiscuous_disable)(dev);
-               if (diag != 0)
-                       dev->data->promiscuous = 1;
-       }
+       dev->data->promiscuous = 0;
+       diag = (*dev->dev_ops->promiscuous_disable)(dev);
+       if (diag != 0)
+               dev->data->promiscuous = 1;
 
        return eth_err(port_id, diag);
 }
@@ -1968,30 +2295,40 @@ int
 rte_eth_allmulticast_enable(uint16_t port_id)
 {
        struct rte_eth_dev *dev;
+       int diag;
 
        RTE_ETH_VALID_PORTID_OR_ERR_RET(port_id, -ENODEV);
        dev = &rte_eth_devices[port_id];
 
+       if (dev->data->all_multicast == 1)
+               return 0;
+
        RTE_FUNC_PTR_OR_ERR_RET(*dev->dev_ops->allmulticast_enable, -ENOTSUP);
-       (*dev->dev_ops->allmulticast_enable)(dev);
-       dev->data->all_multicast = 1;
+       diag = (*dev->dev_ops->allmulticast_enable)(dev);
+       dev->data->all_multicast = (diag == 0) ? 1 : 0;
 
-       return 0;
+       return eth_err(port_id, diag);
 }
 
 int
 rte_eth_allmulticast_disable(uint16_t port_id)
 {
        struct rte_eth_dev *dev;
+       int diag;
 
        RTE_ETH_VALID_PORTID_OR_ERR_RET(port_id, -ENODEV);
        dev = &rte_eth_devices[port_id];
 
+       if (dev->data->all_multicast == 0)
+               return 0;
+
        RTE_FUNC_PTR_OR_ERR_RET(*dev->dev_ops->allmulticast_disable, -ENOTSUP);
        dev->data->all_multicast = 0;
-       (*dev->dev_ops->allmulticast_disable)(dev);
+       diag = (*dev->dev_ops->allmulticast_disable)(dev);
+       if (diag != 0)
+               dev->data->all_multicast = 1;
 
-       return 0;
+       return eth_err(port_id, diag);
 }
 
 int
@@ -2639,6 +2976,7 @@ rte_eth_dev_info_get(uint16_t port_id, struct rte_eth_dev_info *dev_info)
         * return status and does not know if get is successful or not.
         */
        memset(dev_info, 0, sizeof(struct rte_eth_dev_info));
+       dev_info->switch_info.domain_id = RTE_ETH_DEV_SWITCH_DOMAIN_ID_INVALID;
 
        RTE_ETH_VALID_PORTID_OR_ERR_RET(port_id, -ENODEV);
        dev = &rte_eth_devices[port_id];
@@ -2657,6 +2995,12 @@ rte_eth_dev_info_get(uint16_t port_id, struct rte_eth_dev_info *dev_info)
                return eth_err(port_id, diag);
        }
 
+       /* Maximum number of queues should be <= RTE_MAX_QUEUES_PER_PORT */
+       dev_info->max_rx_queues = RTE_MIN(dev_info->max_rx_queues,
+                       RTE_MAX_QUEUES_PER_PORT);
+       dev_info->max_tx_queues = RTE_MIN(dev_info->max_tx_queues,
+                       RTE_MAX_QUEUES_PER_PORT);
+
        dev_info->driver_name = dev->device->driver->name;
        dev_info->nb_rx_queues = dev->data->nb_rx_queues;
        dev_info->nb_tx_queues = dev->data->nb_tx_queues;
@@ -2692,6 +3036,92 @@ rte_eth_dev_get_supported_ptypes(uint16_t port_id, uint32_t ptype_mask,
        return j;
 }
 
+int
+rte_eth_dev_set_ptypes(uint16_t port_id, uint32_t ptype_mask,
+                                uint32_t *set_ptypes, unsigned int num)
+{
+       const uint32_t valid_ptype_masks[] = {
+               RTE_PTYPE_L2_MASK,
+               RTE_PTYPE_L3_MASK,
+               RTE_PTYPE_L4_MASK,
+               RTE_PTYPE_TUNNEL_MASK,
+               RTE_PTYPE_INNER_L2_MASK,
+               RTE_PTYPE_INNER_L3_MASK,
+               RTE_PTYPE_INNER_L4_MASK,
+       };
+       const uint32_t *all_ptypes;
+       struct rte_eth_dev *dev;
+       uint32_t unused_mask;
+       unsigned int i, j;
+       int ret;
+
+       RTE_ETH_VALID_PORTID_OR_ERR_RET(port_id, -ENODEV);
+       dev = &rte_eth_devices[port_id];
+
+       if (num > 0 && set_ptypes == NULL)
+               return -EINVAL;
+
+       if (*dev->dev_ops->dev_supported_ptypes_get == NULL ||
+                       *dev->dev_ops->dev_ptypes_set == NULL) {
+               ret = 0;
+               goto ptype_unknown;
+       }
+
+       if (ptype_mask == 0) {
+               ret = (*dev->dev_ops->dev_ptypes_set)(dev,
+                               ptype_mask);
+               goto ptype_unknown;
+       }
+
+       unused_mask = ptype_mask;
+       for (i = 0; i < RTE_DIM(valid_ptype_masks); i++) {
+               uint32_t mask = ptype_mask & valid_ptype_masks[i];
+               if (mask && mask != valid_ptype_masks[i]) {
+                       ret = -EINVAL;
+                       goto ptype_unknown;
+               }
+               unused_mask &= ~valid_ptype_masks[i];
+       }
+
+       if (unused_mask) {
+               ret = -EINVAL;
+               goto ptype_unknown;
+       }
+
+       all_ptypes = (*dev->dev_ops->dev_supported_ptypes_get)(dev);
+       if (all_ptypes == NULL) {
+               ret = 0;
+               goto ptype_unknown;
+       }
+
+       /*
+        * Accommodate as many set_ptypes as possible. If the supplied
+        * set_ptypes array is insufficient fill it partially.
+        */
+       for (i = 0, j = 0; set_ptypes != NULL &&
+                               (all_ptypes[i] != RTE_PTYPE_UNKNOWN); ++i) {
+               if (ptype_mask & all_ptypes[i]) {
+                       if (j < num - 1) {
+                               set_ptypes[j] = all_ptypes[i];
+                               j++;
+                               continue;
+                       }
+                       break;
+               }
+       }
+
+       if (set_ptypes != NULL && j < num)
+               set_ptypes[j] = RTE_PTYPE_UNKNOWN;
+
+       return (*dev->dev_ops->dev_ptypes_set)(dev, ptype_mask);
+
+ptype_unknown:
+       if (num > 0)
+               set_ptypes[0] = RTE_PTYPE_UNKNOWN;
+
+       return ret;
+}
+
 int
 rte_eth_macaddr_get(uint16_t port_id, struct rte_ether_addr *mac_addr)
 {
@@ -2704,7 +3134,6 @@ rte_eth_macaddr_get(uint16_t port_id, struct rte_ether_addr *mac_addr)
        return 0;
 }
 
-
 int
 rte_eth_dev_get_mtu(uint16_t port_id, uint16_t *mtu)
 {
@@ -2833,53 +3262,53 @@ rte_eth_dev_set_vlan_offload(uint16_t port_id, int offload_mask)
        int mask = 0;
        int cur, org = 0;
        uint64_t orig_offloads;
-       uint64_t *dev_offloads;
+       uint64_t dev_offloads;
 
        RTE_ETH_VALID_PORTID_OR_ERR_RET(port_id, -ENODEV);
        dev = &rte_eth_devices[port_id];
 
        /* save original values in case of failure */
        orig_offloads = dev->data->dev_conf.rxmode.offloads;
-       dev_offloads = &dev->data->dev_conf.rxmode.offloads;
+       dev_offloads = orig_offloads;
 
-       /*check which option changed by application*/
+       /* check which option changed by application */
        cur = !!(offload_mask & ETH_VLAN_STRIP_OFFLOAD);
-       org = !!(*dev_offloads & DEV_RX_OFFLOAD_VLAN_STRIP);
+       org = !!(dev_offloads & DEV_RX_OFFLOAD_VLAN_STRIP);
        if (cur != org) {
                if (cur)
-                       *dev_offloads |= DEV_RX_OFFLOAD_VLAN_STRIP;
+                       dev_offloads |= DEV_RX_OFFLOAD_VLAN_STRIP;
                else
-                       *dev_offloads &= ~DEV_RX_OFFLOAD_VLAN_STRIP;
+                       dev_offloads &= ~DEV_RX_OFFLOAD_VLAN_STRIP;
                mask |= ETH_VLAN_STRIP_MASK;
        }
 
        cur = !!(offload_mask & ETH_VLAN_FILTER_OFFLOAD);
-       org = !!(*dev_offloads & DEV_RX_OFFLOAD_VLAN_FILTER);
+       org = !!(dev_offloads & DEV_RX_OFFLOAD_VLAN_FILTER);
        if (cur != org) {
                if (cur)
-                       *dev_offloads |= DEV_RX_OFFLOAD_VLAN_FILTER;
+                       dev_offloads |= DEV_RX_OFFLOAD_VLAN_FILTER;
                else
-                       *dev_offloads &= ~DEV_RX_OFFLOAD_VLAN_FILTER;
+                       dev_offloads &= ~DEV_RX_OFFLOAD_VLAN_FILTER;
                mask |= ETH_VLAN_FILTER_MASK;
        }
 
        cur = !!(offload_mask & ETH_VLAN_EXTEND_OFFLOAD);
-       org = !!(*dev_offloads & DEV_RX_OFFLOAD_VLAN_EXTEND);
+       org = !!(dev_offloads & DEV_RX_OFFLOAD_VLAN_EXTEND);
        if (cur != org) {
                if (cur)
-                       *dev_offloads |= DEV_RX_OFFLOAD_VLAN_EXTEND;
+                       dev_offloads |= DEV_RX_OFFLOAD_VLAN_EXTEND;
                else
-                       *dev_offloads &= ~DEV_RX_OFFLOAD_VLAN_EXTEND;
+                       dev_offloads &= ~DEV_RX_OFFLOAD_VLAN_EXTEND;
                mask |= ETH_VLAN_EXTEND_MASK;
        }
 
        cur = !!(offload_mask & ETH_QINQ_STRIP_OFFLOAD);
-       org = !!(*dev_offloads & DEV_RX_OFFLOAD_QINQ_STRIP);
+       org = !!(dev_offloads & DEV_RX_OFFLOAD_QINQ_STRIP);
        if (cur != org) {
                if (cur)
-                       *dev_offloads |= DEV_RX_OFFLOAD_QINQ_STRIP;
+                       dev_offloads |= DEV_RX_OFFLOAD_QINQ_STRIP;
                else
-                       *dev_offloads &= ~DEV_RX_OFFLOAD_QINQ_STRIP;
+                       dev_offloads &= ~DEV_RX_OFFLOAD_QINQ_STRIP;
                mask |= ETH_QINQ_STRIP_MASK;
        }
 
@@ -2888,10 +3317,11 @@ rte_eth_dev_set_vlan_offload(uint16_t port_id, int offload_mask)
                return ret;
 
        RTE_FUNC_PTR_OR_ERR_RET(*dev->dev_ops->vlan_offload_set, -ENOTSUP);
+       dev->data->dev_conf.rxmode.offloads = dev_offloads;
        ret = (*dev->dev_ops->vlan_offload_set)(dev, mask);
        if (ret) {
                /* hit an error restore  original values */
-               *dev_offloads = orig_offloads;
+               dev->data->dev_conf.rxmode.offloads = orig_offloads;
        }
 
        return eth_err(port_id, ret);
@@ -2918,7 +3348,7 @@ rte_eth_dev_get_vlan_offload(uint16_t port_id)
                ret |= ETH_VLAN_EXTEND_OFFLOAD;
 
        if (*dev_offloads & DEV_RX_OFFLOAD_QINQ_STRIP)
-               ret |= DEV_RX_OFFLOAD_QINQ_STRIP;
+               ret |= ETH_QINQ_STRIP_OFFLOAD;
 
        return ret;
 }
@@ -3094,6 +3524,8 @@ rte_eth_dev_rss_hash_update(uint16_t port_id,
        if (ret != 0)
                return ret;
 
+       rss_conf->rss_hf = rte_eth_rss_hf_refine(rss_conf->rss_hf);
+
        dev = &rte_eth_devices[port_id];
        if ((dev_info.flow_type_rss_offloads | rss_conf->rss_hf) !=
            dev_info.flow_type_rss_offloads) {
@@ -3617,7 +4049,7 @@ rte_eth_dev_callback_unregister(uint16_t port_id,
                        next = TAILQ_NEXT(cb, next);
 
                        if (cb->cb_fn != cb_fn || cb->event != event ||
-                           (cb->cb_arg != (void *)-1 && cb->cb_arg != cb_arg))
+                           (cb_arg != (void *)-1 && cb->cb_arg != cb_arg))
                                continue;
 
                        /*
@@ -3963,12 +4395,19 @@ rte_eth_add_rx_callback(uint16_t port_id, uint16_t queue_id,
        rte_errno = ENOTSUP;
        return NULL;
 #endif
+       struct rte_eth_dev *dev;
+
        /* check input parameters */
        if (!rte_eth_dev_is_valid_port(port_id) || fn == NULL ||
                    queue_id >= rte_eth_devices[port_id].data->nb_rx_queues) {
                rte_errno = EINVAL;
                return NULL;
        }
+       dev = &rte_eth_devices[port_id];
+       if (rte_eth_dev_is_rx_hairpin_queue(dev, queue_id)) {
+               rte_errno = EINVAL;
+               return NULL;
+       }
        struct rte_eth_rxtx_callback *cb = rte_zmalloc(NULL, sizeof(*cb), 0);
 
        if (cb == NULL) {
@@ -4023,7 +4462,7 @@ rte_eth_add_first_rx_callback(uint16_t port_id, uint16_t queue_id,
        cb->param = user_param;
 
        rte_spinlock_lock(&rte_eth_rx_cb_lock);
-       /* Add the callbacks at fisrt position*/
+       /* Add the callbacks at first position */
        cb->next = rte_eth_devices[port_id].post_rx_burst_cbs[queue_id];
        rte_smp_wmb();
        rte_eth_devices[port_id].post_rx_burst_cbs[queue_id] = cb;
@@ -4040,6 +4479,8 @@ rte_eth_add_tx_callback(uint16_t port_id, uint16_t queue_id,
        rte_errno = ENOTSUP;
        return NULL;
 #endif
+       struct rte_eth_dev *dev;
+
        /* check input parameters */
        if (!rte_eth_dev_is_valid_port(port_id) || fn == NULL ||
                    queue_id >= rte_eth_devices[port_id].data->nb_tx_queues) {
@@ -4047,6 +4488,12 @@ rte_eth_add_tx_callback(uint16_t port_id, uint16_t queue_id,
                return NULL;
        }
 
+       dev = &rte_eth_devices[port_id];
+       if (rte_eth_dev_is_tx_hairpin_queue(dev, queue_id)) {
+               rte_errno = EINVAL;
+               return NULL;
+       }
+
        struct rte_eth_rxtx_callback *cb = rte_zmalloc(NULL, sizeof(*cb), 0);
 
        if (cb == NULL) {
@@ -4160,6 +4607,13 @@ rte_eth_rx_queue_info_get(uint16_t port_id, uint16_t queue_id,
                return -EINVAL;
        }
 
+       if (rte_eth_dev_is_rx_hairpin_queue(dev, queue_id)) {
+               RTE_ETHDEV_LOG(INFO,
+                       "Can't get hairpin Rx queue %"PRIu16" info of device with port_id=%"PRIu16"\n",
+                       queue_id, port_id);
+               return -EINVAL;
+       }
+
        RTE_FUNC_PTR_OR_ERR_RET(*dev->dev_ops->rxq_info_get, -ENOTSUP);
 
        memset(qinfo, 0, sizeof(*qinfo));
@@ -4184,6 +4638,13 @@ rte_eth_tx_queue_info_get(uint16_t port_id, uint16_t queue_id,
                return -EINVAL;
        }
 
+       if (rte_eth_dev_is_tx_hairpin_queue(dev, queue_id)) {
+               RTE_ETHDEV_LOG(INFO,
+                       "Can't get hairpin Tx queue %"PRIu16" info of device with port_id=%"PRIu16"\n",
+                       queue_id, port_id);
+               return -EINVAL;
+       }
+
        RTE_FUNC_PTR_OR_ERR_RET(*dev->dev_ops->txq_info_get, -ENOTSUP);
 
        memset(qinfo, 0, sizeof(*qinfo));
@@ -4192,6 +4653,54 @@ rte_eth_tx_queue_info_get(uint16_t port_id, uint16_t queue_id,
        return 0;
 }
 
+int
+rte_eth_rx_burst_mode_get(uint16_t port_id, uint16_t queue_id,
+                         struct rte_eth_burst_mode *mode)
+{
+       struct rte_eth_dev *dev;
+
+       RTE_ETH_VALID_PORTID_OR_ERR_RET(port_id, -ENODEV);
+
+       if (mode == NULL)
+               return -EINVAL;
+
+       dev = &rte_eth_devices[port_id];
+
+       if (queue_id >= dev->data->nb_rx_queues) {
+               RTE_ETHDEV_LOG(ERR, "Invalid RX queue_id=%u\n", queue_id);
+               return -EINVAL;
+       }
+
+       RTE_FUNC_PTR_OR_ERR_RET(*dev->dev_ops->rx_burst_mode_get, -ENOTSUP);
+       memset(mode, 0, sizeof(*mode));
+       return eth_err(port_id,
+                      dev->dev_ops->rx_burst_mode_get(dev, queue_id, mode));
+}
+
+int
+rte_eth_tx_burst_mode_get(uint16_t port_id, uint16_t queue_id,
+                         struct rte_eth_burst_mode *mode)
+{
+       struct rte_eth_dev *dev;
+
+       RTE_ETH_VALID_PORTID_OR_ERR_RET(port_id, -ENODEV);
+
+       if (mode == NULL)
+               return -EINVAL;
+
+       dev = &rte_eth_devices[port_id];
+
+       if (queue_id >= dev->data->nb_tx_queues) {
+               RTE_ETHDEV_LOG(ERR, "Invalid TX queue_id=%u\n", queue_id);
+               return -EINVAL;
+       }
+
+       RTE_FUNC_PTR_OR_ERR_RET(*dev->dev_ops->tx_burst_mode_get, -ENOTSUP);
+       memset(mode, 0, sizeof(*mode));
+       return eth_err(port_id,
+                      dev->dev_ops->tx_burst_mode_get(dev, queue_id, mode));
+}
+
 int
 rte_eth_dev_set_mc_addr_list(uint16_t port_id,
                             struct rte_ether_addr *mc_addr_set,
@@ -4491,6 +5000,38 @@ rte_eth_dev_adjust_nb_rx_tx_desc(uint16_t port_id,
        return 0;
 }
 
+int
+rte_eth_dev_hairpin_capability_get(uint16_t port_id,
+                                  struct rte_eth_hairpin_cap *cap)
+{
+       struct rte_eth_dev *dev;
+
+       RTE_ETH_VALID_PORTID_OR_ERR_RET(port_id, -EINVAL);
+
+       dev = &rte_eth_devices[port_id];
+       RTE_FUNC_PTR_OR_ERR_RET(*dev->dev_ops->hairpin_cap_get, -ENOTSUP);
+       memset(cap, 0, sizeof(*cap));
+       return eth_err(port_id, (*dev->dev_ops->hairpin_cap_get)(dev, cap));
+}
+
+int
+rte_eth_dev_is_rx_hairpin_queue(struct rte_eth_dev *dev, uint16_t queue_id)
+{
+       if (dev->data->rx_queue_state[queue_id] ==
+           RTE_ETH_QUEUE_STATE_HAIRPIN)
+               return 1;
+       return 0;
+}
+
+int
+rte_eth_dev_is_tx_hairpin_queue(struct rte_eth_dev *dev, uint16_t queue_id)
+{
+       if (dev->data->tx_queue_state[queue_id] ==
+           RTE_ETH_QUEUE_STATE_HAIRPIN)
+               return 1;
+       return 0;
+}
+
 int
 rte_eth_dev_pool_ops_supported(uint16_t port_id, const char *pool)
 {
@@ -4533,8 +5074,7 @@ rte_eth_switch_domain_alloc(uint16_t *domain_id)
 
        *domain_id = RTE_ETH_DEV_SWITCH_DOMAIN_ID_INVALID;
 
-       for (i = RTE_ETH_DEV_SWITCH_DOMAIN_ID_INVALID + 1;
-               i < RTE_MAX_ETHPORTS; i++) {
+       for (i = 0; i < RTE_MAX_ETHPORTS; i++) {
                if (rte_eth_switch_domains[i].state ==
                        RTE_ETH_SWITCH_DOMAIN_UNUSED) {
                        rte_eth_switch_domains[i].state =