net/ena: support fast mbuf free
[dpdk.git] / drivers / net / vmxnet3 / vmxnet3_ethdev.c
index 57feb37..2af67dc 100644 (file)
@@ -26,8 +26,8 @@
 #include <rte_eal.h>
 #include <rte_alarm.h>
 #include <rte_ether.h>
-#include <rte_ethdev_driver.h>
-#include <rte_ethdev_pci.h>
+#include <ethdev_driver.h>
+#include <ethdev_pci.h>
 #include <rte_string_fns.h>
 #include <rte_malloc.h>
 #include <rte_dev.h>
 #include "vmxnet3_logs.h"
 #include "vmxnet3_ethdev.h"
 
-#define PROCESS_SYS_EVENTS 0
-
 #define        VMXNET3_TX_MAX_SEG      UINT8_MAX
 
 #define VMXNET3_TX_OFFLOAD_CAP         \
-       (DEV_TX_OFFLOAD_VLAN_INSERT |   \
-        DEV_TX_OFFLOAD_IPV4_CKSUM |    \
-        DEV_TX_OFFLOAD_TCP_CKSUM |     \
-        DEV_TX_OFFLOAD_UDP_CKSUM |     \
-        DEV_TX_OFFLOAD_TCP_TSO |       \
-        DEV_TX_OFFLOAD_MULTI_SEGS)
+       (RTE_ETH_TX_OFFLOAD_VLAN_INSERT |       \
+        RTE_ETH_TX_OFFLOAD_TCP_CKSUM | \
+        RTE_ETH_TX_OFFLOAD_UDP_CKSUM | \
+        RTE_ETH_TX_OFFLOAD_TCP_TSO |   \
+        RTE_ETH_TX_OFFLOAD_MULTI_SEGS)
 
 #define VMXNET3_RX_OFFLOAD_CAP         \
-       (DEV_RX_OFFLOAD_VLAN_STRIP |    \
-        DEV_RX_OFFLOAD_VLAN_FILTER |   \
-        DEV_RX_OFFLOAD_SCATTER |       \
-        DEV_RX_OFFLOAD_IPV4_CKSUM |    \
-        DEV_RX_OFFLOAD_UDP_CKSUM |     \
-        DEV_RX_OFFLOAD_TCP_CKSUM |     \
-        DEV_RX_OFFLOAD_TCP_LRO |       \
-        DEV_RX_OFFLOAD_JUMBO_FRAME)
+       (RTE_ETH_RX_OFFLOAD_VLAN_STRIP |        \
+        RTE_ETH_RX_OFFLOAD_VLAN_FILTER |   \
+        RTE_ETH_RX_OFFLOAD_SCATTER |   \
+        RTE_ETH_RX_OFFLOAD_UDP_CKSUM | \
+        RTE_ETH_RX_OFFLOAD_TCP_CKSUM | \
+        RTE_ETH_RX_OFFLOAD_TCP_LRO |   \
+        RTE_ETH_RX_OFFLOAD_RSS_HASH)
+
+int vmxnet3_segs_dynfield_offset = -1;
 
 static int eth_vmxnet3_dev_init(struct rte_eth_dev *eth_dev);
 static int eth_vmxnet3_dev_uninit(struct rte_eth_dev *eth_dev);
 static int vmxnet3_dev_configure(struct rte_eth_dev *dev);
 static int vmxnet3_dev_start(struct rte_eth_dev *dev);
-static void vmxnet3_dev_stop(struct rte_eth_dev *dev);
-static void vmxnet3_dev_close(struct rte_eth_dev *dev);
+static int vmxnet3_dev_stop(struct rte_eth_dev *dev);
+static int vmxnet3_dev_close(struct rte_eth_dev *dev);
+static int vmxnet3_dev_reset(struct rte_eth_dev *dev);
 static void vmxnet3_dev_set_rxmode(struct vmxnet3_hw *hw, uint32_t feature, int set);
-static void vmxnet3_dev_promiscuous_enable(struct rte_eth_dev *dev);
-static void vmxnet3_dev_promiscuous_disable(struct rte_eth_dev *dev);
-static void vmxnet3_dev_allmulticast_enable(struct rte_eth_dev *dev);
-static void vmxnet3_dev_allmulticast_disable(struct rte_eth_dev *dev);
+static int vmxnet3_dev_promiscuous_enable(struct rte_eth_dev *dev);
+static int vmxnet3_dev_promiscuous_disable(struct rte_eth_dev *dev);
+static int vmxnet3_dev_allmulticast_enable(struct rte_eth_dev *dev);
+static int vmxnet3_dev_allmulticast_disable(struct rte_eth_dev *dev);
 static int __vmxnet3_dev_link_update(struct rte_eth_dev *dev,
                                     int wait_to_complete);
 static int vmxnet3_dev_link_update(struct rte_eth_dev *dev,
@@ -78,25 +77,39 @@ static int vmxnet3_dev_link_update(struct rte_eth_dev *dev,
 static void vmxnet3_hw_stats_save(struct vmxnet3_hw *hw);
 static int vmxnet3_dev_stats_get(struct rte_eth_dev *dev,
                                  struct rte_eth_stats *stats);
-static void vmxnet3_dev_stats_reset(struct rte_eth_dev *dev);
+static int vmxnet3_dev_stats_reset(struct rte_eth_dev *dev);
 static int vmxnet3_dev_xstats_get_names(struct rte_eth_dev *dev,
                                        struct rte_eth_xstat_name *xstats,
                                        unsigned int n);
 static int vmxnet3_dev_xstats_get(struct rte_eth_dev *dev,
                                  struct rte_eth_xstat *xstats, unsigned int n);
-static void vmxnet3_dev_info_get(struct rte_eth_dev *dev,
-                                struct rte_eth_dev_info *dev_info);
+static int vmxnet3_dev_info_get(struct rte_eth_dev *dev,
+                               struct rte_eth_dev_info *dev_info);
+static int vmxnet3_hw_ver_get(struct rte_eth_dev *dev,
+                             char *fw_version, size_t fw_size);
 static const uint32_t *
 vmxnet3_dev_supported_ptypes_get(struct rte_eth_dev *dev);
+static int vmxnet3_dev_mtu_set(struct rte_eth_dev *dev, uint16_t mtu);
 static int vmxnet3_dev_vlan_filter_set(struct rte_eth_dev *dev,
                                       uint16_t vid, int on);
 static int vmxnet3_dev_vlan_offload_set(struct rte_eth_dev *dev, int mask);
 static int vmxnet3_mac_addr_set(struct rte_eth_dev *dev,
                                 struct rte_ether_addr *mac_addr);
+static void vmxnet3_process_events(struct rte_eth_dev *dev);
 static void vmxnet3_interrupt_handler(void *param);
+static int
+vmxnet3_rss_reta_update(struct rte_eth_dev *dev,
+                       struct rte_eth_rss_reta_entry64 *reta_conf,
+                       uint16_t reta_size);
+static int
+vmxnet3_rss_reta_query(struct rte_eth_dev *dev,
+                      struct rte_eth_rss_reta_entry64 *reta_conf,
+                      uint16_t reta_size);
 
-int vmxnet3_logtype_init;
-int vmxnet3_logtype_driver;
+static int vmxnet3_dev_rx_queue_intr_enable(struct rte_eth_dev *dev,
+                                               uint16_t queue_id);
+static int vmxnet3_dev_rx_queue_intr_disable(struct rte_eth_dev *dev,
+                                               uint16_t queue_id);
 
 /*
  * The set of PCI devices this driver supports
@@ -113,24 +126,31 @@ static const struct eth_dev_ops vmxnet3_eth_dev_ops = {
        .dev_start            = vmxnet3_dev_start,
        .dev_stop             = vmxnet3_dev_stop,
        .dev_close            = vmxnet3_dev_close,
+       .dev_reset            = vmxnet3_dev_reset,
+       .link_update          = vmxnet3_dev_link_update,
        .promiscuous_enable   = vmxnet3_dev_promiscuous_enable,
        .promiscuous_disable  = vmxnet3_dev_promiscuous_disable,
        .allmulticast_enable  = vmxnet3_dev_allmulticast_enable,
        .allmulticast_disable = vmxnet3_dev_allmulticast_disable,
-       .link_update          = vmxnet3_dev_link_update,
+       .mac_addr_set         = vmxnet3_mac_addr_set,
+       .mtu_set              = vmxnet3_dev_mtu_set,
        .stats_get            = vmxnet3_dev_stats_get,
-       .xstats_get_names     = vmxnet3_dev_xstats_get_names,
-       .xstats_get           = vmxnet3_dev_xstats_get,
        .stats_reset          = vmxnet3_dev_stats_reset,
-       .mac_addr_set         = vmxnet3_mac_addr_set,
+       .xstats_get           = vmxnet3_dev_xstats_get,
+       .xstats_get_names     = vmxnet3_dev_xstats_get_names,
        .dev_infos_get        = vmxnet3_dev_info_get,
+       .fw_version_get       = vmxnet3_hw_ver_get,
        .dev_supported_ptypes_get = vmxnet3_dev_supported_ptypes_get,
        .vlan_filter_set      = vmxnet3_dev_vlan_filter_set,
        .vlan_offload_set     = vmxnet3_dev_vlan_offload_set,
        .rx_queue_setup       = vmxnet3_dev_rx_queue_setup,
        .rx_queue_release     = vmxnet3_dev_rx_queue_release,
+       .rx_queue_intr_enable = vmxnet3_dev_rx_queue_intr_enable,
+       .rx_queue_intr_disable = vmxnet3_dev_rx_queue_intr_disable,
        .tx_queue_setup       = vmxnet3_dev_tx_queue_setup,
        .tx_queue_release     = vmxnet3_dev_tx_queue_release,
+       .reta_update          = vmxnet3_rss_reta_update,
+       .reta_query           = vmxnet3_rss_reta_query,
 };
 
 struct vmxnet3_xstats_name_off {
@@ -181,30 +201,84 @@ gpa_zone_reserve(struct rte_eth_dev *dev, uint32_t size,
 }
 
 /*
- * This function is based on vmxnet3_disable_intr()
+ * Enable the given interrupt
  */
 static void
-vmxnet3_disable_intr(struct vmxnet3_hw *hw)
+vmxnet3_enable_intr(struct vmxnet3_hw *hw, unsigned int intr_idx)
 {
-       int i;
+       PMD_INIT_FUNC_TRACE();
+       VMXNET3_WRITE_BAR0_REG(hw, VMXNET3_REG_IMR + intr_idx * 8, 0);
+}
 
+/*
+ * Disable the given interrupt
+ */
+static void
+vmxnet3_disable_intr(struct vmxnet3_hw *hw, unsigned int intr_idx)
+{
        PMD_INIT_FUNC_TRACE();
+       VMXNET3_WRITE_BAR0_REG(hw, VMXNET3_REG_IMR + intr_idx * 8, 1);
+}
 
-       hw->shared->devRead.intrConf.intrCtrl |= VMXNET3_IC_DISABLE_ALL;
-       for (i = 0; i < hw->num_intrs; i++)
-               VMXNET3_WRITE_BAR0_REG(hw, VMXNET3_REG_IMR + i * 8, 1);
+/*
+ * Simple helper to get intrCtrl and eventIntrIdx based on config and hw version
+ */
+static void
+vmxnet3_get_intr_ctrl_ev(struct vmxnet3_hw *hw,
+                        uint8 **out_eventIntrIdx,
+                        uint32 **out_intrCtrl)
+{
+
+       if (VMXNET3_VERSION_GE_6(hw) && hw->queuesExtEnabled) {
+               *out_eventIntrIdx = &hw->shared->devReadExt.intrConfExt.eventIntrIdx;
+               *out_intrCtrl = &hw->shared->devReadExt.intrConfExt.intrCtrl;
+       } else {
+               *out_eventIntrIdx = &hw->shared->devRead.intrConf.eventIntrIdx;
+               *out_intrCtrl = &hw->shared->devRead.intrConf.intrCtrl;
+       }
 }
 
+/*
+ * Disable all intrs used by the device
+ */
 static void
-vmxnet3_enable_intr(struct vmxnet3_hw *hw)
+vmxnet3_disable_all_intrs(struct vmxnet3_hw *hw)
 {
        int i;
+       uint8 *eventIntrIdx;
+       uint32 *intrCtrl;
 
        PMD_INIT_FUNC_TRACE();
+       vmxnet3_get_intr_ctrl_ev(hw, &eventIntrIdx, &intrCtrl);
 
-       hw->shared->devRead.intrConf.intrCtrl &= ~VMXNET3_IC_DISABLE_ALL;
-       for (i = 0; i < hw->num_intrs; i++)
-               VMXNET3_WRITE_BAR0_REG(hw, VMXNET3_REG_IMR + i * 8, 0);
+       *intrCtrl |= rte_cpu_to_le_32(VMXNET3_IC_DISABLE_ALL);
+
+       for (i = 0; i < hw->intr.num_intrs; i++)
+               vmxnet3_disable_intr(hw, i);
+}
+
+/*
+ * Enable all intrs used by the device
+ */
+static void
+vmxnet3_enable_all_intrs(struct vmxnet3_hw *hw)
+{
+       uint8 *eventIntrIdx;
+       uint32 *intrCtrl;
+
+       PMD_INIT_FUNC_TRACE();
+       vmxnet3_get_intr_ctrl_ev(hw, &eventIntrIdx, &intrCtrl);
+
+       *intrCtrl &= rte_cpu_to_le_32(~VMXNET3_IC_DISABLE_ALL);
+
+       if (hw->intr.lsc_only) {
+               vmxnet3_enable_intr(hw, *eventIntrIdx);
+       } else {
+               int i;
+
+               for (i = 0; i < hw->intr.num_intrs; i++)
+                       vmxnet3_enable_intr(hw, i);
+       }
 }
 
 /*
@@ -235,6 +309,11 @@ eth_vmxnet3_dev_init(struct rte_eth_dev *eth_dev)
        struct vmxnet3_hw *hw = eth_dev->data->dev_private;
        uint32_t mac_hi, mac_lo, ver;
        struct rte_eth_link link;
+       static const struct rte_mbuf_dynfield vmxnet3_segs_dynfield_desc = {
+               .name = VMXNET3_SEGS_DYNFIELD_NAME,
+               .size = sizeof(vmxnet3_segs_dynfield_t),
+               .align = __alignof__(vmxnet3_segs_dynfield_t),
+       };
 
        PMD_INIT_FUNC_TRACE();
 
@@ -242,8 +321,17 @@ eth_vmxnet3_dev_init(struct rte_eth_dev *eth_dev)
        eth_dev->rx_pkt_burst = &vmxnet3_recv_pkts;
        eth_dev->tx_pkt_burst = &vmxnet3_xmit_pkts;
        eth_dev->tx_pkt_prepare = vmxnet3_prep_pkts;
+       eth_dev->rx_queue_count = vmxnet3_dev_rx_queue_count;
        pci_dev = RTE_ETH_DEV_TO_PCI(eth_dev);
 
+       /* extra mbuf field is required to guess MSS */
+       vmxnet3_segs_dynfield_offset =
+               rte_mbuf_dynfield_register(&vmxnet3_segs_dynfield_desc);
+       if (vmxnet3_segs_dynfield_offset < 0) {
+               PMD_INIT_LOG(ERR, "Cannot register mbuf field.");
+               return -rte_errno;
+       }
+
        /*
         * for secondary processes, we don't initialize any further as primary
         * has already done this work.
@@ -252,6 +340,7 @@ eth_vmxnet3_dev_init(struct rte_eth_dev *eth_dev)
                return 0;
 
        rte_eth_copy_pci_info(eth_dev, pci_dev);
+       eth_dev->data->dev_flags |= RTE_ETH_DEV_AUTOFILL_QUEUE_XSTATS;
 
        /* Vendor and Device ID need to be set before init of shared code */
        hw->device_id = pci_dev->id.device_id;
@@ -265,9 +354,16 @@ eth_vmxnet3_dev_init(struct rte_eth_dev *eth_dev)
 
        /* Check h/w version compatibility with driver. */
        ver = VMXNET3_READ_BAR1_REG(hw, VMXNET3_REG_VRRS);
-       PMD_INIT_LOG(DEBUG, "Hardware version : %d", ver);
 
-       if (ver & (1 << VMXNET3_REV_4)) {
+       if (ver & (1 << VMXNET3_REV_6)) {
+               VMXNET3_WRITE_BAR1_REG(hw, VMXNET3_REG_VRRS,
+                                      1 << VMXNET3_REV_6);
+               hw->version = VMXNET3_REV_6 + 1;
+       } else if (ver & (1 << VMXNET3_REV_5)) {
+               VMXNET3_WRITE_BAR1_REG(hw, VMXNET3_REG_VRRS,
+                                      1 << VMXNET3_REV_5);
+               hw->version = VMXNET3_REV_5 + 1;
+       } else if (ver & (1 << VMXNET3_REV_4)) {
                VMXNET3_WRITE_BAR1_REG(hw, VMXNET3_REG_VRRS,
                                       1 << VMXNET3_REV_4);
                hw->version = VMXNET3_REV_4 + 1;
@@ -288,7 +384,7 @@ eth_vmxnet3_dev_init(struct rte_eth_dev *eth_dev)
                return -EIO;
        }
 
-       PMD_INIT_LOG(DEBUG, "Using device version %d\n", hw->version);
+       PMD_INIT_LOG(INFO, "Using device v%d", hw->version);
 
        /* Check UPT version compatibility with driver. */
        ver = VMXNET3_READ_BAR1_REG(hw, VMXNET3_REG_UVRS);
@@ -319,13 +415,10 @@ eth_vmxnet3_dev_init(struct rte_eth_dev *eth_dev)
        rte_ether_addr_copy((struct rte_ether_addr *)hw->perm_addr,
                        &eth_dev->data->mac_addrs[0]);
 
-       PMD_INIT_LOG(DEBUG, "MAC Address : %02x:%02x:%02x:%02x:%02x:%02x",
+       PMD_INIT_LOG(DEBUG, "MAC Address : " RTE_ETHER_ADDR_PRT_FMT,
                     hw->perm_addr[0], hw->perm_addr[1], hw->perm_addr[2],
                     hw->perm_addr[3], hw->perm_addr[4], hw->perm_addr[5]);
 
-       /* Flag to call rte_eth_dev_release_port() in rte_eth_dev_close(). */
-       eth_dev->data->dev_flags |= RTE_ETH_DEV_CLOSE_REMOVE;
-
        /* Put device in Quiesce Mode */
        VMXNET3_WRITE_BAR1_REG(hw, VMXNET3_REG_CMD, VMXNET3_CMD_QUIESCE_DEV);
 
@@ -350,9 +443,9 @@ eth_vmxnet3_dev_init(struct rte_eth_dev *eth_dev)
 
        /* set the initial link status */
        memset(&link, 0, sizeof(link));
-       link.link_duplex = ETH_LINK_FULL_DUPLEX;
-       link.link_speed = ETH_SPEED_NUM_10G;
-       link.link_autoneg = ETH_LINK_FIXED;
+       link.link_duplex = RTE_ETH_LINK_FULL_DUPLEX;
+       link.link_speed = RTE_ETH_SPEED_NUM_10G;
+       link.link_autoneg = RTE_ETH_LINK_FIXED;
        rte_eth_linkstatus_set(eth_dev, &link);
 
        return 0;
@@ -373,11 +466,6 @@ eth_vmxnet3_dev_uninit(struct rte_eth_dev *eth_dev)
                return -EBUSY;
        }
 
-       eth_dev->dev_ops = NULL;
-       eth_dev->rx_pkt_burst = NULL;
-       eth_dev->tx_pkt_burst = NULL;
-       eth_dev->tx_pkt_prepare = NULL;
-
        return 0;
 }
 
@@ -400,6 +488,40 @@ static struct rte_pci_driver rte_vmxnet3_pmd = {
        .remove = eth_vmxnet3_pci_remove,
 };
 
+static void
+vmxnet3_alloc_intr_resources(struct rte_eth_dev *dev)
+{
+       struct vmxnet3_hw *hw = dev->data->dev_private;
+       uint32_t cfg;
+       int nvec = 1; /* for link event */
+
+       /* intr settings */
+       VMXNET3_WRITE_BAR1_REG(hw, VMXNET3_REG_CMD,
+                              VMXNET3_CMD_GET_CONF_INTR);
+       cfg = VMXNET3_READ_BAR1_REG(hw, VMXNET3_REG_CMD);
+       hw->intr.type = cfg & 0x3;
+       hw->intr.mask_mode = (cfg >> 2) & 0x3;
+
+       if (hw->intr.type == VMXNET3_IT_AUTO)
+               hw->intr.type = VMXNET3_IT_MSIX;
+
+       if (hw->intr.type == VMXNET3_IT_MSIX) {
+               /* only support shared tx/rx intr */
+               if (hw->num_tx_queues != hw->num_rx_queues)
+                       goto msix_err;
+
+               nvec += hw->num_rx_queues;
+               hw->intr.num_intrs = nvec;
+               return;
+       }
+
+msix_err:
+       /* the tx/rx queue interrupt will be disabled */
+       hw->intr.num_intrs = 2;
+       hw->intr.lsc_only = TRUE;
+       PMD_INIT_LOG(INFO, "Enabled MSI-X with %d vectors", hw->intr.num_intrs);
+}
+
 static int
 vmxnet3_dev_configure(struct rte_eth_dev *dev)
 {
@@ -409,15 +531,25 @@ vmxnet3_dev_configure(struct rte_eth_dev *dev)
 
        PMD_INIT_FUNC_TRACE();
 
-       if (dev->data->nb_tx_queues > VMXNET3_MAX_TX_QUEUES ||
-           dev->data->nb_rx_queues > VMXNET3_MAX_RX_QUEUES) {
-               PMD_INIT_LOG(ERR, "ERROR: Number of queues not supported");
-               return -EINVAL;
+       if (dev->data->dev_conf.rxmode.mq_mode & RTE_ETH_MQ_RX_RSS_FLAG)
+               dev->data->dev_conf.rxmode.offloads |= RTE_ETH_RX_OFFLOAD_RSS_HASH;
+
+       if (!VMXNET3_VERSION_GE_6(hw)) {
+               if (!rte_is_power_of_2(dev->data->nb_rx_queues)) {
+                       PMD_INIT_LOG(ERR,
+                                    "ERROR: Number of rx queues not power of 2");
+                       return -EINVAL;
+               }
        }
 
-       if (!rte_is_power_of_2(dev->data->nb_rx_queues)) {
-               PMD_INIT_LOG(ERR, "ERROR: Number of rx queues not power of 2");
-               return -EINVAL;
+       /* At this point, the number of queues requested has already
+        * been validated against dev_infos max queues by EAL
+        */
+       if (dev->data->nb_rx_queues > VMXNET3_MAX_RX_QUEUES ||
+           dev->data->nb_tx_queues > VMXNET3_MAX_TX_QUEUES) {
+               hw->queuesExtEnabled = 1;
+       } else {
+               hw->queuesExtEnabled = 0;
        }
 
        size = dev->data->nb_rx_queues * sizeof(struct Vmxnet3_TxQueueDesc) +
@@ -467,7 +599,7 @@ vmxnet3_dev_configure(struct rte_eth_dev *dev)
        hw->queueDescPA = mz->iova;
        hw->queue_desc_len = (uint16_t)size;
 
-       if (dev->data->dev_conf.rxmode.mq_mode == ETH_MQ_RX_RSS) {
+       if (dev->data->dev_conf.rxmode.mq_mode == RTE_ETH_MQ_RX_RSS) {
                /* Allocate memory structure for UPT1_RSSConf and configure */
                mz = gpa_zone_reserve(dev, sizeof(struct VMXNET3_RSSConf),
                                      "rss_conf", rte_socket_id(),
@@ -483,6 +615,8 @@ vmxnet3_dev_configure(struct rte_eth_dev *dev)
                hw->rss_confPA = mz->iova;
        }
 
+       vmxnet3_alloc_intr_resources(dev);
+
        return 0;
 }
 
@@ -492,7 +626,7 @@ vmxnet3_write_mac(struct vmxnet3_hw *hw, const uint8_t *addr)
        uint32_t val;
 
        PMD_INIT_LOG(DEBUG,
-                    "Writing MAC Address : %02x:%02x:%02x:%02x:%02x:%02x",
+                    "Writing MAC Address : " RTE_ETHER_ADDR_PRT_FMT,
                     addr[0], addr[1], addr[2],
                     addr[3], addr[4], addr[5]);
 
@@ -503,6 +637,84 @@ vmxnet3_write_mac(struct vmxnet3_hw *hw, const uint8_t *addr)
        VMXNET3_WRITE_BAR1_REG(hw, VMXNET3_REG_MACH, val);
 }
 
+/*
+ * Configure the hardware to generate MSI-X interrupts.
+ * If setting up MSIx fails, try setting up MSI (only 1 interrupt vector
+ * which will be disabled to allow lsc to work).
+ *
+ * Returns 0 on success and -1 otherwise.
+ */
+static int
+vmxnet3_configure_msix(struct rte_eth_dev *dev)
+{
+       struct vmxnet3_hw *hw = dev->data->dev_private;
+       struct rte_intr_handle *intr_handle = dev->intr_handle;
+       uint16_t intr_vector;
+       int i;
+
+       hw->intr.event_intr_idx = 0;
+
+       /* only vfio-pci driver can support interrupt mode. */
+       if (!rte_intr_cap_multiple(intr_handle) ||
+           dev->data->dev_conf.intr_conf.rxq == 0)
+               return -1;
+
+       intr_vector = dev->data->nb_rx_queues;
+       if (intr_vector > MAX_RX_QUEUES(hw)) {
+               PMD_INIT_LOG(ERR, "At most %d intr queues supported",
+                            MAX_RX_QUEUES(hw));
+               return -ENOTSUP;
+       }
+
+       if (rte_intr_efd_enable(intr_handle, intr_vector)) {
+               PMD_INIT_LOG(ERR, "Failed to enable fastpath event fd");
+               return -1;
+       }
+
+       if (rte_intr_dp_is_en(intr_handle)) {
+               if (rte_intr_vec_list_alloc(intr_handle, "intr_vec",
+                                                  dev->data->nb_rx_queues)) {
+                       PMD_INIT_LOG(ERR, "Failed to allocate %d Rx queues intr_vec",
+                                       dev->data->nb_rx_queues);
+                       rte_intr_efd_disable(intr_handle);
+                       return -ENOMEM;
+               }
+       }
+
+       if (!rte_intr_allow_others(intr_handle) &&
+           dev->data->dev_conf.intr_conf.lsc != 0) {
+               PMD_INIT_LOG(ERR, "not enough intr vector to support both Rx interrupt and LSC");
+               rte_intr_vec_list_free(intr_handle);
+               rte_intr_efd_disable(intr_handle);
+               return -1;
+       }
+
+       /* if we cannot allocate one MSI-X vector per queue, don't enable
+        * interrupt mode.
+        */
+       if (hw->intr.num_intrs !=
+                               (rte_intr_nb_efd_get(intr_handle) + 1)) {
+               PMD_INIT_LOG(ERR, "Device configured with %d Rx intr vectors, expecting %d",
+                               hw->intr.num_intrs,
+                               rte_intr_nb_efd_get(intr_handle) + 1);
+               rte_intr_vec_list_free(intr_handle);
+               rte_intr_efd_disable(intr_handle);
+               return -1;
+       }
+
+       for (i = 0; i < dev->data->nb_rx_queues; i++)
+               if (rte_intr_vec_list_index_set(intr_handle, i, i + 1))
+                       return -rte_errno;
+
+       for (i = 0; i < hw->intr.num_intrs; i++)
+               hw->intr.mod_levels[i] = UPT1_IML_ADAPTIVE;
+
+       PMD_INIT_LOG(INFO, "intr type %u, mode %u, %u vectors allocated",
+                   hw->intr.type, hw->intr.mask_mode, hw->intr.num_intrs);
+
+       return 0;
+}
+
 static int
 vmxnet3_dev_setup_memreg(struct rte_eth_dev *dev)
 {
@@ -594,9 +806,11 @@ vmxnet3_setup_driver_shared(struct rte_eth_dev *dev)
 {
        struct rte_eth_conf port_conf = dev->data->dev_conf;
        struct vmxnet3_hw *hw = dev->data->dev_private;
+       struct rte_intr_handle *intr_handle = dev->intr_handle;
        uint32_t mtu = dev->data->mtu;
        Vmxnet3_DriverShared *shared = hw->shared;
        Vmxnet3_DSDevRead *devRead = &shared->devRead;
+       struct Vmxnet3_DSDevReadExt *devReadExt = &shared->devReadExt;
        uint64_t rx_offloads = dev->data->dev_conf.rxmode.offloads;
        uint32_t i;
        int ret;
@@ -619,15 +833,6 @@ vmxnet3_setup_driver_shared(struct rte_eth_dev *dev)
        devRead->misc.numTxQueues  = hw->num_tx_queues;
        devRead->misc.numRxQueues  = hw->num_rx_queues;
 
-       /*
-        * Set number of interrupts to 1
-        * PMD by default disables all the interrupts but this is MUST
-        * to activate device. It needs at least one interrupt for
-        * link events to handle
-        */
-       hw->num_intrs = devRead->intrConf.numIntrs = 1;
-       devRead->intrConf.intrCtrl |= VMXNET3_IC_DISABLE_ALL;
-
        for (i = 0; i < hw->num_tx_queues; i++) {
                Vmxnet3_TxQueueDesc *tqd = &hw->tqd_start[i];
                vmxnet3_tx_queue_t *txq  = dev->data->tx_queues[i];
@@ -644,9 +849,15 @@ vmxnet3_setup_driver_shared(struct rte_eth_dev *dev)
                tqd->conf.compRingSize = txq->comp_ring.size;
                tqd->conf.dataRingSize = txq->data_ring.size;
                tqd->conf.txDataRingDescSize = txq->txdata_desc_size;
-               tqd->conf.intrIdx      = txq->comp_ring.intr_idx;
-               tqd->status.stopped    = TRUE;
-               tqd->status.error      = 0;
+
+               if (hw->intr.lsc_only)
+                       tqd->conf.intrIdx = 1;
+               else
+                       tqd->conf.intrIdx =
+                               rte_intr_vec_list_index_get(intr_handle,
+                                                                  i);
+               tqd->status.stopped = TRUE;
+               tqd->status.error   = 0;
                memset(&tqd->stats, 0, sizeof(tqd->stats));
        }
 
@@ -663,29 +874,59 @@ vmxnet3_setup_driver_shared(struct rte_eth_dev *dev)
                rqd->conf.rxRingSize[0]   = rxq->cmd_ring[0].size;
                rqd->conf.rxRingSize[1]   = rxq->cmd_ring[1].size;
                rqd->conf.compRingSize    = rxq->comp_ring.size;
-               rqd->conf.intrIdx         = rxq->comp_ring.intr_idx;
+
                if (VMXNET3_VERSION_GE_3(hw)) {
                        rqd->conf.rxDataRingBasePA = rxq->data_ring.basePA;
                        rqd->conf.rxDataRingDescSize = rxq->data_desc_size;
                }
-               rqd->status.stopped       = TRUE;
-               rqd->status.error         = 0;
+
+               if (hw->intr.lsc_only)
+                       rqd->conf.intrIdx = 1;
+               else
+                       rqd->conf.intrIdx =
+                               rte_intr_vec_list_index_get(intr_handle,
+                                                                  i);
+               rqd->status.stopped = TRUE;
+               rqd->status.error   = 0;
                memset(&rqd->stats, 0, sizeof(rqd->stats));
        }
 
+       /* intr settings */
+       if (VMXNET3_VERSION_GE_6(hw) && hw->queuesExtEnabled) {
+               devReadExt->intrConfExt.autoMask = hw->intr.mask_mode ==
+                                                  VMXNET3_IMM_AUTO;
+               devReadExt->intrConfExt.numIntrs = hw->intr.num_intrs;
+               for (i = 0; i < hw->intr.num_intrs; i++)
+                       devReadExt->intrConfExt.modLevels[i] =
+                               hw->intr.mod_levels[i];
+
+               devReadExt->intrConfExt.eventIntrIdx = hw->intr.event_intr_idx;
+               devReadExt->intrConfExt.intrCtrl |=
+                       rte_cpu_to_le_32(VMXNET3_IC_DISABLE_ALL);
+       } else {
+               devRead->intrConf.autoMask = hw->intr.mask_mode ==
+                                            VMXNET3_IMM_AUTO;
+               devRead->intrConf.numIntrs = hw->intr.num_intrs;
+               for (i = 0; i < hw->intr.num_intrs; i++)
+                       devRead->intrConf.modLevels[i] = hw->intr.mod_levels[i];
+
+               devRead->intrConf.eventIntrIdx = hw->intr.event_intr_idx;
+               devRead->intrConf.intrCtrl |= rte_cpu_to_le_32(VMXNET3_IC_DISABLE_ALL);
+       }
+
        /* RxMode set to 0 of VMXNET3_RXM_xxx */
        devRead->rxFilterConf.rxMode = 0;
 
        /* Setting up feature flags */
-       if (rx_offloads & DEV_RX_OFFLOAD_CHECKSUM)
+       if (rx_offloads & RTE_ETH_RX_OFFLOAD_CHECKSUM)
                devRead->misc.uptFeatures |= VMXNET3_F_RXCSUM;
 
-       if (rx_offloads & DEV_RX_OFFLOAD_TCP_LRO) {
+       if (rx_offloads & RTE_ETH_RX_OFFLOAD_TCP_LRO) {
                devRead->misc.uptFeatures |= VMXNET3_F_LRO;
                devRead->misc.maxNumRxSG = 0;
        }
 
-       if (port_conf.rxmode.mq_mode == ETH_MQ_RX_RSS) {
+       if (port_conf.rxmode.mq_mode == RTE_ETH_MQ_RX_RSS) {
                ret = vmxnet3_rss_configure(dev);
                if (ret != VMXNET3_SUCCESS)
                        return ret;
@@ -697,7 +938,7 @@ vmxnet3_setup_driver_shared(struct rte_eth_dev *dev)
        }
 
        ret = vmxnet3_dev_vlan_offload_set(dev,
-                       ETH_VLAN_STRIP_MASK | ETH_VLAN_FILTER_MASK);
+                       RTE_ETH_VLAN_STRIP_MASK | RTE_ETH_VLAN_FILTER_MASK);
        if (ret)
                return ret;
 
@@ -722,24 +963,18 @@ vmxnet3_dev_start(struct rte_eth_dev *dev)
        /* Save stats before it is reset by CMD_ACTIVATE */
        vmxnet3_hw_stats_save(hw);
 
+       /* configure MSI-X */
+       ret = vmxnet3_configure_msix(dev);
+       if (ret < 0) {
+               /* revert to lsc only */
+               hw->intr.num_intrs = 2;
+               hw->intr.lsc_only = TRUE;
+       }
+
        ret = vmxnet3_setup_driver_shared(dev);
        if (ret != VMXNET3_SUCCESS)
                return ret;
 
-       /* check if lsc interrupt feature is enabled */
-       if (dev->data->dev_conf.intr_conf.lsc) {
-               struct rte_pci_device *pci_dev = RTE_DEV_TO_PCI(dev->device);
-
-               /* Setup interrupt callback  */
-               rte_intr_callback_register(&pci_dev->intr_handle,
-                                          vmxnet3_interrupt_handler, dev);
-
-               if (rte_intr_enable(&pci_dev->intr_handle) < 0) {
-                       PMD_INIT_LOG(ERR, "interrupt enable failed");
-                       return -EIO;
-               }
-       }
-
        /* Exchange shared data with device */
        VMXNET3_WRITE_BAR1_REG(hw, VMXNET3_REG_DSAL,
                               VMXNET3_GET_ADDR_LO(hw->sharedPA));
@@ -755,21 +990,28 @@ vmxnet3_dev_start(struct rte_eth_dev *dev)
                return -EINVAL;
        }
 
-       /* Setup memory region for rx buffers */
-       ret = vmxnet3_dev_setup_memreg(dev);
-       if (ret == 0) {
-               VMXNET3_WRITE_BAR1_REG(hw, VMXNET3_REG_CMD,
-                                      VMXNET3_CMD_REGISTER_MEMREGS);
-               ret = VMXNET3_READ_BAR1_REG(hw, VMXNET3_REG_CMD);
-               if (ret != 0)
-                       PMD_INIT_LOG(DEBUG,
-                                    "Failed in setup memory region cmd\n");
-               ret = 0;
+       /* Check memregs restrictions first */
+       if (dev->data->nb_rx_queues <= VMXNET3_MAX_RX_QUEUES &&
+           dev->data->nb_tx_queues <= VMXNET3_MAX_TX_QUEUES) {
+               ret = vmxnet3_dev_setup_memreg(dev);
+               if (ret == 0) {
+                       VMXNET3_WRITE_BAR1_REG(hw, VMXNET3_REG_CMD,
+                                       VMXNET3_CMD_REGISTER_MEMREGS);
+                       ret = VMXNET3_READ_BAR1_REG(hw, VMXNET3_REG_CMD);
+                       if (ret != 0)
+                               PMD_INIT_LOG(DEBUG,
+                                       "Failed in setup memory region cmd\n");
+                       ret = 0;
+               } else {
+                       PMD_INIT_LOG(DEBUG, "Failed to setup memory region\n");
+               }
        } else {
-               PMD_INIT_LOG(DEBUG, "Failed to setup memory region\n");
+               PMD_INIT_LOG(WARNING, "Memregs can't init (rx: %d, tx: %d)",
+                            dev->data->nb_rx_queues, dev->data->nb_tx_queues);
        }
 
-       if (VMXNET3_VERSION_GE_4(hw)) {
+       if (VMXNET3_VERSION_GE_4(hw) &&
+           dev->data->dev_conf.rxmode.mq_mode == RTE_ETH_MQ_RX_RSS) {
                /* Check for additional RSS  */
                ret = vmxnet3_v4_rss_configure(dev);
                if (ret != VMXNET3_SUCCESS) {
@@ -778,9 +1020,6 @@ vmxnet3_dev_start(struct rte_eth_dev *dev)
                }
        }
 
-       /* Disable interrupts */
-       vmxnet3_disable_intr(hw);
-
        /*
         * Load RX queues with blank mbufs and update next2fill index for device
         * Update RxMode of the device
@@ -796,49 +1035,74 @@ vmxnet3_dev_start(struct rte_eth_dev *dev)
        /* Setting proper Rx Mode and issue Rx Mode Update command */
        vmxnet3_dev_set_rxmode(hw, VMXNET3_RXM_UCAST | VMXNET3_RXM_BCAST, 1);
 
-       if (dev->data->dev_conf.intr_conf.lsc) {
-               vmxnet3_enable_intr(hw);
+       /* Setup interrupt callback  */
+       rte_intr_callback_register(dev->intr_handle,
+                                  vmxnet3_interrupt_handler, dev);
 
-               /*
-                * Update link state from device since this won't be
-                * done upon starting with lsc in use. This is done
-                * only after enabling interrupts to avoid any race
-                * where the link state could change without an
-                * interrupt being fired.
-                */
-               __vmxnet3_dev_link_update(dev, 0);
+       if (rte_intr_enable(dev->intr_handle) < 0) {
+               PMD_INIT_LOG(ERR, "interrupt enable failed");
+               return -EIO;
        }
 
+       /* enable all intrs */
+       vmxnet3_enable_all_intrs(hw);
+
+       vmxnet3_process_events(dev);
+
+       /*
+        * Update link state from device since this won't be
+        * done upon starting with lsc in use. This is done
+        * only after enabling interrupts to avoid any race
+        * where the link state could change without an
+        * interrupt being fired.
+        */
+       __vmxnet3_dev_link_update(dev, 0);
+
        return VMXNET3_SUCCESS;
 }
 
 /*
  * Stop device: disable rx and tx functions to allow for reconfiguring.
  */
-static void
+static int
 vmxnet3_dev_stop(struct rte_eth_dev *dev)
 {
        struct rte_eth_link link;
        struct vmxnet3_hw *hw = dev->data->dev_private;
+       struct rte_intr_handle *intr_handle = dev->intr_handle;
+       int ret;
 
        PMD_INIT_FUNC_TRACE();
 
        if (hw->adapter_stopped == 1) {
                PMD_INIT_LOG(DEBUG, "Device already stopped.");
-               return;
+               return 0;
        }
 
-       /* disable interrupts */
-       vmxnet3_disable_intr(hw);
+       do {
+               /* Unregister has lock to make sure there is no running cb.
+                * This has to happen first since vmxnet3_interrupt_handler
+                * reenables interrupts by calling vmxnet3_enable_intr
+                */
+               ret = rte_intr_callback_unregister(intr_handle,
+                                                  vmxnet3_interrupt_handler,
+                                                  (void *)-1);
+       } while (ret == -EAGAIN);
 
-       if (dev->data->dev_conf.intr_conf.lsc) {
-               struct rte_pci_device *pci_dev = RTE_DEV_TO_PCI(dev->device);
+       if (ret < 0)
+               PMD_DRV_LOG(ERR, "Error attempting to unregister intr cb: %d",
+                           ret);
 
-               rte_intr_disable(&pci_dev->intr_handle);
+       PMD_INIT_LOG(DEBUG, "Disabled %d intr callbacks", ret);
 
-               rte_intr_callback_unregister(&pci_dev->intr_handle,
-                                            vmxnet3_interrupt_handler, dev);
-       }
+       /* disable interrupts */
+       vmxnet3_disable_all_intrs(hw);
+
+       rte_intr_disable(intr_handle);
+
+       /* Clean datapath event and queue/vector mapping */
+       rte_intr_efd_disable(intr_handle);
+       rte_intr_vec_list_free(intr_handle);
 
        /* quiesce the device first */
        VMXNET3_WRITE_BAR1_REG(hw, VMXNET3_REG_CMD, VMXNET3_CMD_QUIESCE_DEV);
@@ -853,12 +1117,15 @@ vmxnet3_dev_stop(struct rte_eth_dev *dev)
 
        /* Clear recorded link status */
        memset(&link, 0, sizeof(link));
-       link.link_duplex = ETH_LINK_FULL_DUPLEX;
-       link.link_speed = ETH_SPEED_NUM_10G;
-       link.link_autoneg = ETH_LINK_FIXED;
+       link.link_duplex = RTE_ETH_LINK_FULL_DUPLEX;
+       link.link_speed = RTE_ETH_SPEED_NUM_10G;
+       link.link_autoneg = RTE_ETH_LINK_FIXED;
        rte_eth_linkstatus_set(dev, &link);
 
        hw->adapter_stopped = 1;
+       dev->data->dev_started = 0;
+
+       return 0;
 }
 
 static void
@@ -868,31 +1135,42 @@ vmxnet3_free_queues(struct rte_eth_dev *dev)
 
        PMD_INIT_FUNC_TRACE();
 
-       for (i = 0; i < dev->data->nb_rx_queues; i++) {
-               void *rxq = dev->data->rx_queues[i];
-
-               vmxnet3_dev_rx_queue_release(rxq);
-       }
+       for (i = 0; i < dev->data->nb_rx_queues; i++)
+               vmxnet3_dev_rx_queue_release(dev, i);
        dev->data->nb_rx_queues = 0;
 
-       for (i = 0; i < dev->data->nb_tx_queues; i++) {
-               void *txq = dev->data->tx_queues[i];
-
-               vmxnet3_dev_tx_queue_release(txq);
-       }
+       for (i = 0; i < dev->data->nb_tx_queues; i++)
+               vmxnet3_dev_tx_queue_release(dev, i);
        dev->data->nb_tx_queues = 0;
 }
 
 /*
  * Reset and stop device.
  */
-static void
+static int
 vmxnet3_dev_close(struct rte_eth_dev *dev)
 {
+       int ret;
        PMD_INIT_FUNC_TRACE();
+       if (rte_eal_process_type() != RTE_PROC_PRIMARY)
+               return 0;
 
-       vmxnet3_dev_stop(dev);
+       ret = vmxnet3_dev_stop(dev);
        vmxnet3_free_queues(dev);
+
+       return ret;
+}
+
+static int
+vmxnet3_dev_reset(struct rte_eth_dev *dev)
+{
+       int ret;
+
+       ret = eth_vmxnet3_dev_uninit(dev);
+       if (ret)
+               return ret;
+       ret = eth_vmxnet3_dev_init(dev);
+       return ret;
 }
 
 static void
@@ -984,8 +1262,6 @@ vmxnet3_hw_stats_save(struct vmxnet3_hw *hw)
 
        VMXNET3_WRITE_BAR1_REG(hw, VMXNET3_REG_CMD, VMXNET3_CMD_GET_STATS);
 
-       RTE_BUILD_BUG_ON(RTE_ETHDEV_QUEUE_STAT_CNTRS < VMXNET3_MAX_TX_QUEUES);
-
        for (i = 0; i < hw->num_tx_queues; i++)
                vmxnet3_hw_tx_stats_get(hw, i, &hw->saved_tx_stats[i]);
        for (i = 0; i < hw->num_rx_queues; i++)
@@ -1087,7 +1363,6 @@ vmxnet3_dev_stats_get(struct rte_eth_dev *dev, struct rte_eth_stats *stats)
 
        VMXNET3_WRITE_BAR1_REG(hw, VMXNET3_REG_CMD, VMXNET3_CMD_GET_STATS);
 
-       RTE_BUILD_BUG_ON(RTE_ETHDEV_QUEUE_STAT_CNTRS < VMXNET3_MAX_TX_QUEUES);
        for (i = 0; i < hw->num_tx_queues; i++) {
                vmxnet3_tx_stats_get(hw, i, &txStats);
 
@@ -1104,7 +1379,6 @@ vmxnet3_dev_stats_get(struct rte_eth_dev *dev, struct rte_eth_stats *stats)
                stats->oerrors += txStats.pktsTxError + txStats.pktsTxDiscard;
        }
 
-       RTE_BUILD_BUG_ON(RTE_ETHDEV_QUEUE_STAT_CNTRS < VMXNET3_MAX_RX_QUEUES);
        for (i = 0; i < hw->num_rx_queues; i++) {
                vmxnet3_rx_stats_get(hw, i, &rxStats);
 
@@ -1127,7 +1401,7 @@ vmxnet3_dev_stats_get(struct rte_eth_dev *dev, struct rte_eth_stats *stats)
        return 0;
 }
 
-static void
+static int
 vmxnet3_dev_stats_reset(struct rte_eth_dev *dev)
 {
        unsigned int i;
@@ -1149,19 +1423,41 @@ vmxnet3_dev_stats_reset(struct rte_eth_dev *dev)
                memcpy(&hw->snapshot_rx_stats[i], &rxStats,
                        sizeof(hw->snapshot_rx_stats[0]));
        }
+
+       return 0;
 }
 
-static void
-vmxnet3_dev_info_get(struct rte_eth_dev *dev __rte_unused,
+static int
+vmxnet3_dev_info_get(struct rte_eth_dev *dev,
                     struct rte_eth_dev_info *dev_info)
 {
        struct vmxnet3_hw *hw = dev->data->dev_private;
+       int queues = 0;
+
+       if (VMXNET3_VERSION_GE_6(hw)) {
+               VMXNET3_WRITE_BAR1_REG(hw, VMXNET3_REG_CMD,
+                                      VMXNET3_CMD_GET_MAX_QUEUES_CONF);
+               queues = VMXNET3_READ_BAR1_REG(hw, VMXNET3_REG_CMD);
+
+               if (queues > 0) {
+                       dev_info->max_rx_queues =
+                         RTE_MIN(VMXNET3_EXT_MAX_RX_QUEUES, ((queues >> 8) & 0xff));
+                       dev_info->max_tx_queues =
+                         RTE_MIN(VMXNET3_EXT_MAX_TX_QUEUES, (queues & 0xff));
+               } else {
+                       dev_info->max_rx_queues = VMXNET3_MAX_RX_QUEUES;
+                       dev_info->max_tx_queues = VMXNET3_MAX_TX_QUEUES;
+               }
+       } else {
+               dev_info->max_rx_queues = VMXNET3_MAX_RX_QUEUES;
+               dev_info->max_tx_queues = VMXNET3_MAX_TX_QUEUES;
+       }
 
-       dev_info->max_rx_queues = VMXNET3_MAX_RX_QUEUES;
-       dev_info->max_tx_queues = VMXNET3_MAX_TX_QUEUES;
        dev_info->min_rx_bufsize = 1518 + RTE_PKTMBUF_HEADROOM;
        dev_info->max_rx_pktlen = 16384; /* includes CRC, cf MAXFRS register */
-       dev_info->speed_capa = ETH_LINK_SPEED_10G;
+       dev_info->min_mtu = VMXNET3_MIN_MTU;
+       dev_info->max_mtu = VMXNET3_MAX_MTU;
+       dev_info->speed_capa = RTE_ETH_LINK_SPEED_10G;
        dev_info->max_mac_addrs = VMXNET3_MAX_MAC_ADDRS;
 
        dev_info->flow_type_rss_offloads = VMXNET3_RSS_OFFLOAD_ALL;
@@ -1188,6 +1484,29 @@ vmxnet3_dev_info_get(struct rte_eth_dev *dev __rte_unused,
        dev_info->rx_queue_offload_capa = 0;
        dev_info->tx_offload_capa = VMXNET3_TX_OFFLOAD_CAP;
        dev_info->tx_queue_offload_capa = 0;
+       if (hw->rss_conf == NULL) {
+               /* RSS not configured */
+               dev_info->reta_size = 0;
+       } else {
+               dev_info->reta_size = hw->rss_conf->indTableSize;
+       }
+       return 0;
+}
+
+static int
+vmxnet3_hw_ver_get(struct rte_eth_dev *dev,
+                  char *fw_version, size_t fw_size)
+{
+       int ret;
+       struct vmxnet3_hw *hw = dev->data->dev_private;
+
+       ret = snprintf(fw_version, fw_size, "v%d", hw->version);
+
+       ret += 1; /* add the size of '\0' */
+       if (fw_size < (uint32_t)ret)
+               return ret;
+       else
+               return 0;
 }
 
 static const uint32_t *
@@ -1214,6 +1533,44 @@ vmxnet3_mac_addr_set(struct rte_eth_dev *dev, struct rte_ether_addr *mac_addr)
        return 0;
 }
 
+static int
+vmxnet3_dev_mtu_set(struct rte_eth_dev *dev, uint16_t mtu)
+{
+       struct vmxnet3_hw *hw = dev->data->dev_private;
+       uint32_t frame_size = mtu + RTE_ETHER_HDR_LEN + RTE_ETHER_CRC_LEN + 4;
+
+       if (mtu < VMXNET3_MIN_MTU)
+               return -EINVAL;
+
+       if (VMXNET3_VERSION_GE_6(hw)) {
+               if (mtu > VMXNET3_V6_MAX_MTU)
+                       return -EINVAL;
+       } else {
+               if (mtu > VMXNET3_MAX_MTU) {
+                       PMD_DRV_LOG(ERR, "MTU %d too large in device version v%d",
+                                   mtu, hw->version);
+                       return -EINVAL;
+               }
+       }
+
+       dev->data->mtu = mtu;
+       /* update max frame size */
+       dev->data->dev_conf.rxmode.mtu = frame_size;
+
+       if (dev->data->dev_started == 0)
+               return 0;
+
+    /* changing mtu for vmxnet3 pmd does not require a restart
+     * as it does not need to repopulate the rx rings to support
+     * different mtu size.  We stop and restart the device here
+     * just to pass the mtu info to the backend.
+     */
+       vmxnet3_dev_stop(dev);
+       vmxnet3_dev_start(dev);
+
+       return 0;
+}
+
 /* return 0 means link status changed, -1 means not changed */
 static int
 __vmxnet3_dev_link_update(struct rte_eth_dev *dev,
@@ -1229,10 +1586,10 @@ __vmxnet3_dev_link_update(struct rte_eth_dev *dev,
        ret = VMXNET3_READ_BAR1_REG(hw, VMXNET3_REG_CMD);
 
        if (ret & 0x1)
-               link.link_status = ETH_LINK_UP;
-       link.link_duplex = ETH_LINK_FULL_DUPLEX;
-       link.link_speed = ETH_SPEED_NUM_10G;
-       link.link_autoneg = ETH_LINK_FIXED;
+               link.link_status = RTE_ETH_LINK_UP;
+       link.link_duplex = RTE_ETH_LINK_FULL_DUPLEX;
+       link.link_speed = RTE_ETH_SPEED_NUM_10G;
+       link.link_autoneg = RTE_ETH_LINK_FIXED;
 
        return rte_eth_linkstatus_set(dev, &link);
 }
@@ -1262,7 +1619,7 @@ vmxnet3_dev_set_rxmode(struct vmxnet3_hw *hw, uint32_t feature, int set)
 }
 
 /* Promiscuous supported only if Vmxnet3_DriverShared is initialized in adapter */
-static void
+static int
 vmxnet3_dev_promiscuous_enable(struct rte_eth_dev *dev)
 {
        struct vmxnet3_hw *hw = dev->data->dev_private;
@@ -1273,41 +1630,49 @@ vmxnet3_dev_promiscuous_enable(struct rte_eth_dev *dev)
 
        VMXNET3_WRITE_BAR1_REG(hw, VMXNET3_REG_CMD,
                               VMXNET3_CMD_UPDATE_VLAN_FILTERS);
+
+       return 0;
 }
 
 /* Promiscuous supported only if Vmxnet3_DriverShared is initialized in adapter */
-static void
+static int
 vmxnet3_dev_promiscuous_disable(struct rte_eth_dev *dev)
 {
        struct vmxnet3_hw *hw = dev->data->dev_private;
        uint32_t *vf_table = hw->shared->devRead.rxFilterConf.vfTable;
        uint64_t rx_offloads = dev->data->dev_conf.rxmode.offloads;
 
-       if (rx_offloads & DEV_RX_OFFLOAD_VLAN_FILTER)
+       if (rx_offloads & RTE_ETH_RX_OFFLOAD_VLAN_FILTER)
                memcpy(vf_table, hw->shadow_vfta, VMXNET3_VFT_TABLE_SIZE);
        else
                memset(vf_table, 0xff, VMXNET3_VFT_TABLE_SIZE);
        vmxnet3_dev_set_rxmode(hw, VMXNET3_RXM_PROMISC, 0);
        VMXNET3_WRITE_BAR1_REG(hw, VMXNET3_REG_CMD,
                               VMXNET3_CMD_UPDATE_VLAN_FILTERS);
+
+       return 0;
 }
 
 /* Allmulticast supported only if Vmxnet3_DriverShared is initialized in adapter */
-static void
+static int
 vmxnet3_dev_allmulticast_enable(struct rte_eth_dev *dev)
 {
        struct vmxnet3_hw *hw = dev->data->dev_private;
 
        vmxnet3_dev_set_rxmode(hw, VMXNET3_RXM_ALL_MULTI, 1);
+
+       return 0;
 }
 
 /* Allmulticast supported only if Vmxnet3_DriverShared is initialized in adapter */
-static void
+static int
 vmxnet3_dev_allmulticast_disable(struct rte_eth_dev *dev)
 {
        struct vmxnet3_hw *hw = dev->data->dev_private;
 
        vmxnet3_dev_set_rxmode(hw, VMXNET3_RXM_ALL_MULTI, 0);
+
+       return 0;
 }
 
 /* Enable/disable filter on vlan */
@@ -1347,8 +1712,8 @@ vmxnet3_dev_vlan_offload_set(struct rte_eth_dev *dev, int mask)
        uint32_t *vf_table = devRead->rxFilterConf.vfTable;
        uint64_t rx_offloads = dev->data->dev_conf.rxmode.offloads;
 
-       if (mask & ETH_VLAN_STRIP_MASK) {
-               if (rx_offloads & DEV_RX_OFFLOAD_VLAN_STRIP)
+       if (mask & RTE_ETH_VLAN_STRIP_MASK) {
+               if (rx_offloads & RTE_ETH_RX_OFFLOAD_VLAN_STRIP)
                        devRead->misc.uptFeatures |= UPT1_F_RXVLAN;
                else
                        devRead->misc.uptFeatures &= ~UPT1_F_RXVLAN;
@@ -1357,8 +1722,8 @@ vmxnet3_dev_vlan_offload_set(struct rte_eth_dev *dev, int mask)
                                       VMXNET3_CMD_UPDATE_FEATURE);
        }
 
-       if (mask & ETH_VLAN_FILTER_MASK) {
-               if (rx_offloads & DEV_RX_OFFLOAD_VLAN_FILTER)
+       if (mask & RTE_ETH_VLAN_FILTER_MASK) {
+               if (rx_offloads & RTE_ETH_RX_OFFLOAD_VLAN_FILTER)
                        memcpy(vf_table, hw->shadow_vfta, VMXNET3_VFT_TABLE_SIZE);
                else
                        memset(vf_table, 0xff, VMXNET3_VFT_TABLE_SIZE);
@@ -1389,9 +1754,9 @@ vmxnet3_process_events(struct rte_eth_dev *dev)
        if (events & VMXNET3_ECR_LINK) {
                PMD_DRV_LOG(DEBUG, "Process events: VMXNET3_ECR_LINK event");
                if (vmxnet3_dev_link_update(dev, 0) == 0)
-                       _rte_eth_dev_callback_process(dev,
-                                                     RTE_ETH_EVENT_INTR_LSC,
-                                                     NULL);
+                       rte_eth_dev_callback_process(dev,
+                                                    RTE_ETH_EVENT_INTR_LSC,
+                                                    NULL);
        }
 
        /* Check if there is an error on xmit/recv queues */
@@ -1422,24 +1787,108 @@ static void
 vmxnet3_interrupt_handler(void *param)
 {
        struct rte_eth_dev *dev = param;
-       struct rte_pci_device *pci_dev = RTE_DEV_TO_PCI(dev->device);
+       struct vmxnet3_hw *hw = dev->data->dev_private;
+       uint32_t events;
+       uint8 *eventIntrIdx;
+       uint32 *intrCtrl;
 
+       PMD_INIT_FUNC_TRACE();
+
+       vmxnet3_get_intr_ctrl_ev(hw, &eventIntrIdx, &intrCtrl);
+       vmxnet3_disable_intr(hw, *eventIntrIdx);
+
+       events = VMXNET3_READ_BAR1_REG(hw, VMXNET3_REG_ECR);
+       if (events == 0)
+               goto done;
+
+       RTE_LOG(DEBUG, PMD, "Reading events: 0x%X", events);
        vmxnet3_process_events(dev);
+done:
+       vmxnet3_enable_intr(hw, *eventIntrIdx);
+}
+
+static int
+vmxnet3_dev_rx_queue_intr_enable(struct rte_eth_dev *dev, uint16_t queue_id)
+{
+       struct vmxnet3_hw *hw = dev->data->dev_private;
+
+       vmxnet3_enable_intr(hw,
+                           rte_intr_vec_list_index_get(dev->intr_handle,
+                                                              queue_id));
 
-       if (rte_intr_ack(&pci_dev->intr_handle) < 0)
-               PMD_DRV_LOG(ERR, "interrupt enable failed");
+       return 0;
+}
+
+static int
+vmxnet3_dev_rx_queue_intr_disable(struct rte_eth_dev *dev, uint16_t queue_id)
+{
+       struct vmxnet3_hw *hw = dev->data->dev_private;
+
+       vmxnet3_disable_intr(hw,
+               rte_intr_vec_list_index_get(dev->intr_handle, queue_id));
+
+       return 0;
 }
 
 RTE_PMD_REGISTER_PCI(net_vmxnet3, rte_vmxnet3_pmd);
 RTE_PMD_REGISTER_PCI_TABLE(net_vmxnet3, pci_id_vmxnet3_map);
 RTE_PMD_REGISTER_KMOD_DEP(net_vmxnet3, "* igb_uio | uio_pci_generic | vfio-pci");
+RTE_LOG_REGISTER_SUFFIX(vmxnet3_logtype_init, init, NOTICE);
+RTE_LOG_REGISTER_SUFFIX(vmxnet3_logtype_driver, driver, NOTICE);
+
+static int
+vmxnet3_rss_reta_update(struct rte_eth_dev *dev,
+                       struct rte_eth_rss_reta_entry64 *reta_conf,
+                       uint16_t reta_size)
+{
+       int i, idx, shift;
+       struct vmxnet3_hw *hw = dev->data->dev_private;
+       struct VMXNET3_RSSConf *dev_rss_conf = hw->rss_conf;
+
+       if (reta_size != dev_rss_conf->indTableSize) {
+               PMD_DRV_LOG(ERR,
+                       "The size of hash lookup table configured (%d) doesn't match "
+                       "the supported number (%d)",
+                       reta_size, dev_rss_conf->indTableSize);
+               return -EINVAL;
+       }
+
+       for (i = 0; i < reta_size; i++) {
+               idx = i / RTE_ETH_RETA_GROUP_SIZE;
+               shift = i % RTE_ETH_RETA_GROUP_SIZE;
+               if (reta_conf[idx].mask & RTE_BIT64(shift))
+                       dev_rss_conf->indTable[i] = (uint8_t)reta_conf[idx].reta[shift];
+       }
+
+       VMXNET3_WRITE_BAR1_REG(hw, VMXNET3_REG_CMD,
+                               VMXNET3_CMD_UPDATE_RSSIDT);
 
-RTE_INIT(vmxnet3_init_log)
+       return 0;
+}
+
+static int
+vmxnet3_rss_reta_query(struct rte_eth_dev *dev,
+                      struct rte_eth_rss_reta_entry64 *reta_conf,
+                      uint16_t reta_size)
 {
-       vmxnet3_logtype_init = rte_log_register("pmd.net.vmxnet3.init");
-       if (vmxnet3_logtype_init >= 0)
-               rte_log_set_level(vmxnet3_logtype_init, RTE_LOG_NOTICE);
-       vmxnet3_logtype_driver = rte_log_register("pmd.net.vmxnet3.driver");
-       if (vmxnet3_logtype_driver >= 0)
-               rte_log_set_level(vmxnet3_logtype_driver, RTE_LOG_NOTICE);
+       int i, idx, shift;
+       struct vmxnet3_hw *hw = dev->data->dev_private;
+       struct VMXNET3_RSSConf *dev_rss_conf = hw->rss_conf;
+
+       if (reta_size != dev_rss_conf->indTableSize) {
+               PMD_DRV_LOG(ERR,
+                       "Size of requested hash lookup table (%d) doesn't "
+                       "match the configured size (%d)",
+                       reta_size, dev_rss_conf->indTableSize);
+               return -EINVAL;
+       }
+
+       for (i = 0; i < reta_size; i++) {
+               idx = i / RTE_ETH_RETA_GROUP_SIZE;
+               shift = i % RTE_ETH_RETA_GROUP_SIZE;
+               if (reta_conf[idx].mask & RTE_BIT64(shift))
+                       reta_conf[idx].reta[shift] = dev_rss_conf->indTable[i];
+       }
+
+       return 0;
 }