net/sfc: use switch port ID as representor ID
[dpdk.git] / drivers / net / sfc / sfc_ethdev.c
index 9f00363..0297e79 100644 (file)
@@ -1,19 +1,20 @@
 /* SPDX-License-Identifier: BSD-3-Clause
  *
- * Copyright (c) 2016-2018 Solarflare Communications Inc.
- * All rights reserved.
+ * Copyright(c) 2019-2021 Xilinx, Inc.
+ * Copyright(c) 2016-2019 Solarflare Communications Inc.
  *
  * This software was jointly developed between OKTET Labs (under contract
  * for Solarflare) and Solarflare Communications, Inc.
  */
 
 #include <rte_dev.h>
-#include <rte_ethdev_driver.h>
-#include <rte_ethdev_pci.h>
+#include <ethdev_driver.h>
+#include <ethdev_pci.h>
 #include <rte_pci.h>
 #include <rte_bus_pci.h>
 #include <rte_errno.h>
 #include <rte_string_fns.h>
+#include <rte_ether.h>
 
 #include "efx.h"
 
 #include "sfc_flow.h"
 #include "sfc_dp.h"
 #include "sfc_dp_rx.h"
+#include "sfc_repr.h"
+#include "sfc_sw_stats.h"
+#include "sfc_switch.h"
+
+#define SFC_XSTAT_ID_INVALID_VAL  UINT64_MAX
+#define SFC_XSTAT_ID_INVALID_NAME '\0'
 
 uint32_t sfc_logtype_driver;
 
 static struct sfc_dp_list sfc_dp_head =
        TAILQ_HEAD_INITIALIZER(sfc_dp_head);
 
+
+static void sfc_eth_dev_clear_ops(struct rte_eth_dev *dev);
+
+
 static int
 sfc_fw_version_get(struct rte_eth_dev *dev, char *fw_version, size_t fw_size)
 {
-       struct sfc_adapter *sa = dev->data->dev_private;
+       struct sfc_adapter *sa = sfc_adapter_by_eth_dev(dev);
        efx_nic_fw_info_t enfi;
        int ret;
        int rc;
 
-       /*
-        * Return value of the callback is likely supposed to be
-        * equal to or greater than 0, nevertheless, if an error
-        * occurs, it will be desirable to pass it to the caller
-        */
-       if ((fw_version == NULL) || (fw_size == 0))
-               return -EINVAL;
-
        rc = efx_nic_get_fw_version(sa->nic, &enfi);
        if (rc != 0)
                return -rc;
@@ -81,31 +84,38 @@ sfc_fw_version_get(struct rte_eth_dev *dev, char *fw_version, size_t fw_size)
                return 0;
 }
 
-static void
+static int
 sfc_dev_infos_get(struct rte_eth_dev *dev, struct rte_eth_dev_info *dev_info)
 {
        const struct sfc_adapter_priv *sap = sfc_adapter_priv_by_eth_dev(dev);
-       struct sfc_adapter *sa = dev->data->dev_private;
-       struct sfc_rss *rss = &sa->rss;
+       struct sfc_adapter_shared *sas = sfc_adapter_shared_by_eth_dev(dev);
+       struct sfc_adapter *sa = sfc_adapter_by_eth_dev(dev);
+       struct sfc_rss *rss = &sas->rss;
+       struct sfc_mae *mae = &sa->mae;
        uint64_t txq_offloads_def = 0;
 
        sfc_log_init(sa, "entry");
 
+       dev_info->min_mtu = RTE_ETHER_MIN_MTU;
+       dev_info->max_mtu = EFX_MAC_SDU_MAX;
+
        dev_info->max_rx_pktlen = EFX_MAC_PDU_MAX;
 
+       dev_info->max_vfs = sa->sriov.num_vfs;
+
        /* Autonegotiation may be disabled */
        dev_info->speed_capa = ETH_LINK_SPEED_FIXED;
-       if (sa->port.phy_adv_cap_mask & EFX_PHY_CAP_1000FDX)
+       if (sa->port.phy_adv_cap_mask & (1u << EFX_PHY_CAP_1000FDX))
                dev_info->speed_capa |= ETH_LINK_SPEED_1G;
-       if (sa->port.phy_adv_cap_mask & EFX_PHY_CAP_10000FDX)
+       if (sa->port.phy_adv_cap_mask & (1u << EFX_PHY_CAP_10000FDX))
                dev_info->speed_capa |= ETH_LINK_SPEED_10G;
-       if (sa->port.phy_adv_cap_mask & EFX_PHY_CAP_25000FDX)
+       if (sa->port.phy_adv_cap_mask & (1u << EFX_PHY_CAP_25000FDX))
                dev_info->speed_capa |= ETH_LINK_SPEED_25G;
-       if (sa->port.phy_adv_cap_mask & EFX_PHY_CAP_40000FDX)
+       if (sa->port.phy_adv_cap_mask & (1u << EFX_PHY_CAP_40000FDX))
                dev_info->speed_capa |= ETH_LINK_SPEED_40G;
-       if (sa->port.phy_adv_cap_mask & EFX_PHY_CAP_50000FDX)
+       if (sa->port.phy_adv_cap_mask & (1u << EFX_PHY_CAP_50000FDX))
                dev_info->speed_capa |= ETH_LINK_SPEED_50G;
-       if (sa->port.phy_adv_cap_mask & EFX_PHY_CAP_100000FDX)
+       if (sa->port.phy_adv_cap_mask & (1u << EFX_PHY_CAP_100000FDX))
                dev_info->speed_capa |= ETH_LINK_SPEED_100G;
 
        dev_info->max_rx_queues = sa->rxq_max;
@@ -152,21 +162,21 @@ sfc_dev_infos_get(struct rte_eth_dev *dev, struct rte_eth_dev_info *dev_info)
        }
 
        /* Initialize to hardware limits */
-       dev_info->rx_desc_lim.nb_max = EFX_RXQ_MAXNDESCS;
-       dev_info->rx_desc_lim.nb_min = EFX_RXQ_MINNDESCS;
+       dev_info->rx_desc_lim.nb_max = sa->rxq_max_entries;
+       dev_info->rx_desc_lim.nb_min = sa->rxq_min_entries;
        /* The RXQ hardware requires that the descriptor count is a power
         * of 2, but rx_desc_lim cannot properly describe that constraint.
         */
-       dev_info->rx_desc_lim.nb_align = EFX_RXQ_MINNDESCS;
+       dev_info->rx_desc_lim.nb_align = sa->rxq_min_entries;
 
        /* Initialize to hardware limits */
        dev_info->tx_desc_lim.nb_max = sa->txq_max_entries;
-       dev_info->tx_desc_lim.nb_min = EFX_TXQ_MINNDESCS;
+       dev_info->tx_desc_lim.nb_min = sa->txq_min_entries;
        /*
         * The TXQ hardware requires that the descriptor count is a power
         * of 2, but tx_desc_lim cannot properly describe that constraint
         */
-       dev_info->tx_desc_lim.nb_align = EFX_TXQ_MINNDESCS;
+       dev_info->tx_desc_lim.nb_align = sa->txq_min_entries;
 
        if (sap->dp_rx->get_dev_info != NULL)
                sap->dp_rx->get_dev_info(dev_info);
@@ -175,24 +185,29 @@ sfc_dev_infos_get(struct rte_eth_dev *dev, struct rte_eth_dev_info *dev_info)
 
        dev_info->dev_capa = RTE_ETH_DEV_CAPA_RUNTIME_RX_QUEUE_SETUP |
                             RTE_ETH_DEV_CAPA_RUNTIME_TX_QUEUE_SETUP;
+
+       if (mae->status == SFC_MAE_STATUS_SUPPORTED) {
+               dev_info->switch_info.name = dev->device->driver->name;
+               dev_info->switch_info.domain_id = mae->switch_domain_id;
+               dev_info->switch_info.port_id = mae->switch_port_id;
+       }
+
+       return 0;
 }
 
 static const uint32_t *
 sfc_dev_supported_ptypes_get(struct rte_eth_dev *dev)
 {
        const struct sfc_adapter_priv *sap = sfc_adapter_priv_by_eth_dev(dev);
-       struct sfc_adapter *sa = dev->data->dev_private;
-       const efx_nic_cfg_t *encp = efx_nic_cfg_get(sa->nic);
-       uint32_t tunnel_encaps = encp->enc_tunnel_encapsulations_supported;
 
-       return sap->dp_rx->supported_ptypes_get(tunnel_encaps);
+       return sap->dp_rx->supported_ptypes_get(sap->shared->tunnel_encaps);
 }
 
 static int
 sfc_dev_configure(struct rte_eth_dev *dev)
 {
        struct rte_eth_dev_data *dev_data = dev->data;
-       struct sfc_adapter *sa = dev_data->dev_private;
+       struct sfc_adapter *sa = sfc_adapter_by_eth_dev(dev);
        int rc;
 
        sfc_log_init(sa, "entry n_rxq=%u n_txq=%u",
@@ -200,9 +215,9 @@ sfc_dev_configure(struct rte_eth_dev *dev)
 
        sfc_adapter_lock(sa);
        switch (sa->state) {
-       case SFC_ADAPTER_CONFIGURED:
+       case SFC_ETHDEV_CONFIGURED:
                /* FALLTHROUGH */
-       case SFC_ADAPTER_INITIALIZED:
+       case SFC_ETHDEV_INITIALIZED:
                rc = sfc_configure(sa);
                break;
        default:
@@ -221,7 +236,7 @@ sfc_dev_configure(struct rte_eth_dev *dev)
 static int
 sfc_dev_start(struct rte_eth_dev *dev)
 {
-       struct sfc_adapter *sa = dev->data->dev_private;
+       struct sfc_adapter *sa = sfc_adapter_by_eth_dev(dev);
        int rc;
 
        sfc_log_init(sa, "entry");
@@ -238,13 +253,13 @@ sfc_dev_start(struct rte_eth_dev *dev)
 static int
 sfc_dev_link_update(struct rte_eth_dev *dev, int wait_to_complete)
 {
-       struct sfc_adapter *sa = dev->data->dev_private;
+       struct sfc_adapter *sa = sfc_adapter_by_eth_dev(dev);
        struct rte_eth_link current_link;
        int ret;
 
        sfc_log_init(sa, "entry");
 
-       if (sa->state != SFC_ADAPTER_STARTED) {
+       if (sa->state != SFC_ETHDEV_STARTED) {
                sfc_port_link_mode_to_info(EFX_LINK_UNKNOWN, &current_link);
        } else if (wait_to_complete) {
                efx_link_mode_t link_mode;
@@ -266,10 +281,10 @@ sfc_dev_link_update(struct rte_eth_dev *dev, int wait_to_complete)
        return ret;
 }
 
-static void
+static int
 sfc_dev_stop(struct rte_eth_dev *dev)
 {
-       struct sfc_adapter *sa = dev->data->dev_private;
+       struct sfc_adapter *sa = sfc_adapter_by_eth_dev(dev);
 
        sfc_log_init(sa, "entry");
 
@@ -278,12 +293,14 @@ sfc_dev_stop(struct rte_eth_dev *dev)
        sfc_adapter_unlock(sa);
 
        sfc_log_init(sa, "done");
+
+       return 0;
 }
 
 static int
 sfc_dev_set_link_up(struct rte_eth_dev *dev)
 {
-       struct sfc_adapter *sa = dev->data->dev_private;
+       struct sfc_adapter *sa = sfc_adapter_by_eth_dev(dev);
        int rc;
 
        sfc_log_init(sa, "entry");
@@ -299,7 +316,7 @@ sfc_dev_set_link_up(struct rte_eth_dev *dev)
 static int
 sfc_dev_set_link_down(struct rte_eth_dev *dev)
 {
-       struct sfc_adapter *sa = dev->data->dev_private;
+       struct sfc_adapter *sa = sfc_adapter_by_eth_dev(dev);
 
        sfc_log_init(sa, "entry");
 
@@ -311,42 +328,78 @@ sfc_dev_set_link_down(struct rte_eth_dev *dev)
 }
 
 static void
+sfc_eth_dev_secondary_clear_ops(struct rte_eth_dev *dev)
+{
+       free(dev->process_private);
+       rte_eth_dev_release_port(dev);
+}
+
+static int
 sfc_dev_close(struct rte_eth_dev *dev)
 {
-       struct sfc_adapter *sa = dev->data->dev_private;
+       struct sfc_adapter *sa = sfc_adapter_by_eth_dev(dev);
 
        sfc_log_init(sa, "entry");
 
+       if (rte_eal_process_type() != RTE_PROC_PRIMARY) {
+               sfc_eth_dev_secondary_clear_ops(dev);
+               return 0;
+       }
+
+       sfc_pre_detach(sa);
+
        sfc_adapter_lock(sa);
        switch (sa->state) {
-       case SFC_ADAPTER_STARTED:
+       case SFC_ETHDEV_STARTED:
                sfc_stop(sa);
-               SFC_ASSERT(sa->state == SFC_ADAPTER_CONFIGURED);
+               SFC_ASSERT(sa->state == SFC_ETHDEV_CONFIGURED);
                /* FALLTHROUGH */
-       case SFC_ADAPTER_CONFIGURED:
+       case SFC_ETHDEV_CONFIGURED:
                sfc_close(sa);
-               SFC_ASSERT(sa->state == SFC_ADAPTER_INITIALIZED);
+               SFC_ASSERT(sa->state == SFC_ETHDEV_INITIALIZED);
                /* FALLTHROUGH */
-       case SFC_ADAPTER_INITIALIZED:
+       case SFC_ETHDEV_INITIALIZED:
                break;
        default:
                sfc_err(sa, "unexpected adapter state %u on close", sa->state);
                break;
        }
+
+       /*
+        * Cleanup all resources.
+        * Rollback primary process sfc_eth_dev_init() below.
+        */
+
+       sfc_eth_dev_clear_ops(dev);
+
+       sfc_detach(sa);
+       sfc_unprobe(sa);
+
+       sfc_kvargs_cleanup(sa);
+
        sfc_adapter_unlock(sa);
+       sfc_adapter_lock_fini(sa);
 
        sfc_log_init(sa, "done");
+
+       /* Required for logging, so cleanup last */
+       sa->eth_dev = NULL;
+
+       free(sa);
+
+       return 0;
 }
 
-static void
+static int
 sfc_dev_filter_set(struct rte_eth_dev *dev, enum sfc_dev_filter_mode mode,
                   boolean_t enabled)
 {
        struct sfc_port *port;
        boolean_t *toggle;
-       struct sfc_adapter *sa = dev->data->dev_private;
+       struct sfc_adapter *sa = sfc_adapter_by_eth_dev(dev);
        boolean_t allmulti = (mode == SFC_DEV_FILTER_MODE_ALLMULTI);
        const char *desc = (allmulti) ? "all-multi" : "promiscuous";
+       int rc = 0;
 
        sfc_adapter_lock(sa);
 
@@ -356,66 +409,93 @@ sfc_dev_filter_set(struct rte_eth_dev *dev, enum sfc_dev_filter_mode mode,
        if (*toggle != enabled) {
                *toggle = enabled;
 
-               if (port->isolated) {
+               if (sfc_sa2shared(sa)->isolated) {
                        sfc_warn(sa, "isolated mode is active on the port");
                        sfc_warn(sa, "the change is to be applied on the next "
                                     "start provided that isolated mode is "
                                     "disabled prior the next start");
-               } else if ((sa->state == SFC_ADAPTER_STARTED) &&
-                          (sfc_set_rx_mode(sa) != 0)) {
+               } else if ((sa->state == SFC_ETHDEV_STARTED) &&
+                          ((rc = sfc_set_rx_mode(sa)) != 0)) {
                        *toggle = !(enabled);
-                       sfc_warn(sa, "Failed to %s %s mode",
-                                ((enabled) ? "enable" : "disable"), desc);
+                       sfc_warn(sa, "Failed to %s %s mode, rc = %d",
+                                ((enabled) ? "enable" : "disable"), desc, rc);
+
+                       /*
+                        * For promiscuous and all-multicast filters a
+                        * permission failure should be reported as an
+                        * unsupported filter.
+                        */
+                       if (rc == EPERM)
+                               rc = ENOTSUP;
                }
        }
 
        sfc_adapter_unlock(sa);
+       return rc;
 }
 
-static void
+static int
 sfc_dev_promisc_enable(struct rte_eth_dev *dev)
 {
-       sfc_dev_filter_set(dev, SFC_DEV_FILTER_MODE_PROMISC, B_TRUE);
+       int rc = sfc_dev_filter_set(dev, SFC_DEV_FILTER_MODE_PROMISC, B_TRUE);
+
+       SFC_ASSERT(rc >= 0);
+       return -rc;
 }
 
-static void
+static int
 sfc_dev_promisc_disable(struct rte_eth_dev *dev)
 {
-       sfc_dev_filter_set(dev, SFC_DEV_FILTER_MODE_PROMISC, B_FALSE);
+       int rc = sfc_dev_filter_set(dev, SFC_DEV_FILTER_MODE_PROMISC, B_FALSE);
+
+       SFC_ASSERT(rc >= 0);
+       return -rc;
 }
 
-static void
+static int
 sfc_dev_allmulti_enable(struct rte_eth_dev *dev)
 {
-       sfc_dev_filter_set(dev, SFC_DEV_FILTER_MODE_ALLMULTI, B_TRUE);
+       int rc = sfc_dev_filter_set(dev, SFC_DEV_FILTER_MODE_ALLMULTI, B_TRUE);
+
+       SFC_ASSERT(rc >= 0);
+       return -rc;
 }
 
-static void
+static int
 sfc_dev_allmulti_disable(struct rte_eth_dev *dev)
 {
-       sfc_dev_filter_set(dev, SFC_DEV_FILTER_MODE_ALLMULTI, B_FALSE);
+       int rc = sfc_dev_filter_set(dev, SFC_DEV_FILTER_MODE_ALLMULTI, B_FALSE);
+
+       SFC_ASSERT(rc >= 0);
+       return -rc;
 }
 
 static int
-sfc_rx_queue_setup(struct rte_eth_dev *dev, uint16_t rx_queue_id,
+sfc_rx_queue_setup(struct rte_eth_dev *dev, uint16_t ethdev_qid,
                   uint16_t nb_rx_desc, unsigned int socket_id,
                   const struct rte_eth_rxconf *rx_conf,
                   struct rte_mempool *mb_pool)
 {
-       struct sfc_adapter *sa = dev->data->dev_private;
+       struct sfc_adapter_shared *sas = sfc_adapter_shared_by_eth_dev(dev);
+       struct sfc_adapter *sa = sfc_adapter_by_eth_dev(dev);
+       sfc_ethdev_qid_t sfc_ethdev_qid = ethdev_qid;
+       struct sfc_rxq_info *rxq_info;
+       sfc_sw_index_t sw_index;
        int rc;
 
        sfc_log_init(sa, "RxQ=%u nb_rx_desc=%u socket_id=%u",
-                    rx_queue_id, nb_rx_desc, socket_id);
+                    ethdev_qid, nb_rx_desc, socket_id);
 
        sfc_adapter_lock(sa);
 
-       rc = sfc_rx_qinit(sa, rx_queue_id, nb_rx_desc, socket_id,
+       sw_index = sfc_rxq_sw_index_by_ethdev_rx_qid(sas, sfc_ethdev_qid);
+       rc = sfc_rx_qinit(sa, sw_index, nb_rx_desc, socket_id,
                          rx_conf, mb_pool);
        if (rc != 0)
                goto fail_rx_qinit;
 
-       dev->data->rx_queues[rx_queue_id] = sa->rxq_info[rx_queue_id].dp;
+       rxq_info = sfc_rxq_info_by_ethdev_qid(sas, sfc_ethdev_qid);
+       dev->data->rx_queues[ethdev_qid] = rxq_info->dp;
 
        sfc_adapter_unlock(sa);
 
@@ -428,12 +508,12 @@ fail_rx_qinit:
 }
 
 static void
-sfc_rx_queue_release(void *queue)
+sfc_rx_queue_release(struct rte_eth_dev *dev, uint16_t qid)
 {
-       struct sfc_dp_rxq *dp_rxq = queue;
+       struct sfc_dp_rxq *dp_rxq = dev->data->rx_queues[qid];
        struct sfc_rxq *rxq;
        struct sfc_adapter *sa;
-       unsigned int sw_index;
+       sfc_sw_index_t sw_index;
 
        if (dp_rxq == NULL)
                return;
@@ -452,23 +532,28 @@ sfc_rx_queue_release(void *queue)
 }
 
 static int
-sfc_tx_queue_setup(struct rte_eth_dev *dev, uint16_t tx_queue_id,
+sfc_tx_queue_setup(struct rte_eth_dev *dev, uint16_t ethdev_qid,
                   uint16_t nb_tx_desc, unsigned int socket_id,
                   const struct rte_eth_txconf *tx_conf)
 {
-       struct sfc_adapter *sa = dev->data->dev_private;
+       struct sfc_adapter_shared *sas = sfc_adapter_shared_by_eth_dev(dev);
+       struct sfc_adapter *sa = sfc_adapter_by_eth_dev(dev);
+       struct sfc_txq_info *txq_info;
+       sfc_sw_index_t sw_index;
        int rc;
 
        sfc_log_init(sa, "TxQ = %u, nb_tx_desc = %u, socket_id = %u",
-                    tx_queue_id, nb_tx_desc, socket_id);
+                    ethdev_qid, nb_tx_desc, socket_id);
 
        sfc_adapter_lock(sa);
 
-       rc = sfc_tx_qinit(sa, tx_queue_id, nb_tx_desc, socket_id, tx_conf);
+       sw_index = sfc_txq_sw_index_by_ethdev_tx_qid(sas, ethdev_qid);
+       rc = sfc_tx_qinit(sa, sw_index, nb_tx_desc, socket_id, tx_conf);
        if (rc != 0)
                goto fail_tx_qinit;
 
-       dev->data->tx_queues[tx_queue_id] = sa->txq_info[tx_queue_id].dp;
+       txq_info = sfc_txq_info_by_ethdev_qid(sas, ethdev_qid);
+       dev->data->tx_queues[ethdev_qid] = txq_info->dp;
 
        sfc_adapter_unlock(sa);
        return 0;
@@ -480,11 +565,11 @@ fail_tx_qinit:
 }
 
 static void
-sfc_tx_queue_release(void *queue)
+sfc_tx_queue_release(struct rte_eth_dev *dev, uint16_t qid)
 {
-       struct sfc_dp_txq *dp_txq = queue;
+       struct sfc_dp_txq *dp_txq = dev->data->tx_queues[qid];
        struct sfc_txq *txq;
-       unsigned int sw_index;
+       sfc_sw_index_t sw_index;
        struct sfc_adapter *sa;
 
        if (dp_txq == NULL)
@@ -505,6 +590,60 @@ sfc_tx_queue_release(void *queue)
        sfc_adapter_unlock(sa);
 }
 
+static void
+sfc_stats_get_dp_rx(struct sfc_adapter *sa, uint64_t *pkts, uint64_t *bytes)
+{
+       struct sfc_adapter_shared *sas = sfc_sa2shared(sa);
+       uint64_t pkts_sum = 0;
+       uint64_t bytes_sum = 0;
+       unsigned int i;
+
+       for (i = 0; i < sas->ethdev_rxq_count; ++i) {
+               struct sfc_rxq_info *rxq_info;
+
+               rxq_info = sfc_rxq_info_by_ethdev_qid(sas, i);
+               if (rxq_info->state & SFC_RXQ_INITIALIZED) {
+                       union sfc_pkts_bytes qstats;
+
+                       sfc_pkts_bytes_get(&rxq_info->dp->dpq.stats, &qstats);
+                       pkts_sum += qstats.pkts -
+                                       sa->sw_stats.reset_rx_pkts[i];
+                       bytes_sum += qstats.bytes -
+                                       sa->sw_stats.reset_rx_bytes[i];
+               }
+       }
+
+       *pkts = pkts_sum;
+       *bytes = bytes_sum;
+}
+
+static void
+sfc_stats_get_dp_tx(struct sfc_adapter *sa, uint64_t *pkts, uint64_t *bytes)
+{
+       struct sfc_adapter_shared *sas = sfc_sa2shared(sa);
+       uint64_t pkts_sum = 0;
+       uint64_t bytes_sum = 0;
+       unsigned int i;
+
+       for (i = 0; i < sas->ethdev_txq_count; ++i) {
+               struct sfc_txq_info *txq_info;
+
+               txq_info = sfc_txq_info_by_ethdev_qid(sas, i);
+               if (txq_info->state & SFC_TXQ_INITIALIZED) {
+                       union sfc_pkts_bytes qstats;
+
+                       sfc_pkts_bytes_get(&txq_info->dp->dpq.stats, &qstats);
+                       pkts_sum += qstats.pkts -
+                                       sa->sw_stats.reset_tx_pkts[i];
+                       bytes_sum += qstats.bytes -
+                                       sa->sw_stats.reset_tx_bytes[i];
+               }
+       }
+
+       *pkts = pkts_sum;
+       *bytes = bytes_sum;
+}
+
 /*
  * Some statistics are computed as A - B where A and B each increase
  * monotonically with some hardware counter(s) and the counters are read
@@ -531,14 +670,22 @@ sfc_update_diff_stat(uint64_t *stat, uint64_t newval)
 static int
 sfc_stats_get(struct rte_eth_dev *dev, struct rte_eth_stats *stats)
 {
-       struct sfc_adapter *sa = dev->data->dev_private;
+       const struct sfc_adapter_priv *sap = sfc_adapter_priv_by_eth_dev(dev);
+       bool have_dp_rx_stats = sap->dp_rx->features & SFC_DP_RX_FEAT_STATS;
+       bool have_dp_tx_stats = sap->dp_tx->features & SFC_DP_TX_FEAT_STATS;
+       struct sfc_adapter *sa = sfc_adapter_by_eth_dev(dev);
        struct sfc_port *port = &sa->port;
        uint64_t *mac_stats;
        int ret;
 
-       rte_spinlock_lock(&port->mac_stats_lock);
+       sfc_adapter_lock(sa);
+
+       if (have_dp_rx_stats)
+               sfc_stats_get_dp_rx(sa, &stats->ipackets, &stats->ibytes);
+       if (have_dp_tx_stats)
+               sfc_stats_get_dp_tx(sa, &stats->opackets, &stats->obytes);
 
-       ret = sfc_port_update_mac_stats(sa);
+       ret = sfc_port_update_mac_stats(sa, B_FALSE);
        if (ret != 0)
                goto unlock;
 
@@ -546,28 +693,41 @@ sfc_stats_get(struct rte_eth_dev *dev, struct rte_eth_stats *stats)
 
        if (EFX_MAC_STAT_SUPPORTED(port->mac_stats_mask,
                                   EFX_MAC_VADAPTER_RX_UNICAST_PACKETS)) {
-               stats->ipackets =
-                       mac_stats[EFX_MAC_VADAPTER_RX_UNICAST_PACKETS] +
-                       mac_stats[EFX_MAC_VADAPTER_RX_MULTICAST_PACKETS] +
-                       mac_stats[EFX_MAC_VADAPTER_RX_BROADCAST_PACKETS];
-               stats->opackets =
-                       mac_stats[EFX_MAC_VADAPTER_TX_UNICAST_PACKETS] +
-                       mac_stats[EFX_MAC_VADAPTER_TX_MULTICAST_PACKETS] +
-                       mac_stats[EFX_MAC_VADAPTER_TX_BROADCAST_PACKETS];
-               stats->ibytes =
-                       mac_stats[EFX_MAC_VADAPTER_RX_UNICAST_BYTES] +
-                       mac_stats[EFX_MAC_VADAPTER_RX_MULTICAST_BYTES] +
-                       mac_stats[EFX_MAC_VADAPTER_RX_BROADCAST_BYTES];
-               stats->obytes =
-                       mac_stats[EFX_MAC_VADAPTER_TX_UNICAST_BYTES] +
-                       mac_stats[EFX_MAC_VADAPTER_TX_MULTICAST_BYTES] +
-                       mac_stats[EFX_MAC_VADAPTER_TX_BROADCAST_BYTES];
+               if (!have_dp_rx_stats) {
+                       stats->ipackets =
+                               mac_stats[EFX_MAC_VADAPTER_RX_UNICAST_PACKETS] +
+                               mac_stats[EFX_MAC_VADAPTER_RX_MULTICAST_PACKETS] +
+                               mac_stats[EFX_MAC_VADAPTER_RX_BROADCAST_PACKETS];
+                       stats->ibytes =
+                               mac_stats[EFX_MAC_VADAPTER_RX_UNICAST_BYTES] +
+                               mac_stats[EFX_MAC_VADAPTER_RX_MULTICAST_BYTES] +
+                               mac_stats[EFX_MAC_VADAPTER_RX_BROADCAST_BYTES];
+
+                       /* CRC is included in these stats, but shouldn't be */
+                       stats->ibytes -= stats->ipackets * RTE_ETHER_CRC_LEN;
+               }
+               if (!have_dp_tx_stats) {
+                       stats->opackets =
+                               mac_stats[EFX_MAC_VADAPTER_TX_UNICAST_PACKETS] +
+                               mac_stats[EFX_MAC_VADAPTER_TX_MULTICAST_PACKETS] +
+                               mac_stats[EFX_MAC_VADAPTER_TX_BROADCAST_PACKETS];
+                       stats->obytes =
+                               mac_stats[EFX_MAC_VADAPTER_TX_UNICAST_BYTES] +
+                               mac_stats[EFX_MAC_VADAPTER_TX_MULTICAST_BYTES] +
+                               mac_stats[EFX_MAC_VADAPTER_TX_BROADCAST_BYTES];
+
+                       /* CRC is included in these stats, but shouldn't be */
+                       stats->obytes -= stats->opackets * RTE_ETHER_CRC_LEN;
+               }
                stats->imissed = mac_stats[EFX_MAC_VADAPTER_RX_BAD_PACKETS];
                stats->oerrors = mac_stats[EFX_MAC_VADAPTER_TX_BAD_PACKETS];
        } else {
-               stats->opackets = mac_stats[EFX_MAC_TX_PKTS];
-               stats->ibytes = mac_stats[EFX_MAC_RX_OCTETS];
-               stats->obytes = mac_stats[EFX_MAC_TX_OCTETS];
+               if (!have_dp_tx_stats) {
+                       stats->opackets = mac_stats[EFX_MAC_TX_PKTS];
+                       stats->obytes = mac_stats[EFX_MAC_TX_OCTETS] -
+                               mac_stats[EFX_MAC_TX_PKTS] * RTE_ETHER_CRC_LEN;
+               }
+
                /*
                 * Take into account stats which are whenever supported
                 * on EF10. If some stat is not supported by current
@@ -591,77 +751,90 @@ sfc_stats_get(struct rte_eth_dev *dev, struct rte_eth_stats *stats)
                        mac_stats[EFX_MAC_RX_JABBER_PKTS];
                /* no oerrors counters supported on EF10 */
 
-               /* Exclude missed, errors and pauses from Rx packets */
-               sfc_update_diff_stat(&port->ipackets,
-                       mac_stats[EFX_MAC_RX_PKTS] -
-                       mac_stats[EFX_MAC_RX_PAUSE_PKTS] -
-                       stats->imissed - stats->ierrors);
-               stats->ipackets = port->ipackets;
+               if (!have_dp_rx_stats) {
+                       /* Exclude missed, errors and pauses from Rx packets */
+                       sfc_update_diff_stat(&port->ipackets,
+                               mac_stats[EFX_MAC_RX_PKTS] -
+                               mac_stats[EFX_MAC_RX_PAUSE_PKTS] -
+                               stats->imissed - stats->ierrors);
+                       stats->ipackets = port->ipackets;
+                       stats->ibytes = mac_stats[EFX_MAC_RX_OCTETS] -
+                               mac_stats[EFX_MAC_RX_PKTS] * RTE_ETHER_CRC_LEN;
+               }
        }
 
 unlock:
-       rte_spinlock_unlock(&port->mac_stats_lock);
+       sfc_adapter_unlock(sa);
        SFC_ASSERT(ret >= 0);
        return -ret;
 }
 
-static void
+static int
 sfc_stats_reset(struct rte_eth_dev *dev)
 {
-       struct sfc_adapter *sa = dev->data->dev_private;
+       struct sfc_adapter *sa = sfc_adapter_by_eth_dev(dev);
        struct sfc_port *port = &sa->port;
        int rc;
 
-       if (sa->state != SFC_ADAPTER_STARTED) {
+       sfc_adapter_lock(sa);
+
+       if (sa->state != SFC_ETHDEV_STARTED) {
                /*
                 * The operation cannot be done if port is not started; it
                 * will be scheduled to be done during the next port start
                 */
                port->mac_stats_reset_pending = B_TRUE;
-               return;
+               sfc_adapter_unlock(sa);
+               return 0;
        }
 
        rc = sfc_port_reset_mac_stats(sa);
        if (rc != 0)
                sfc_err(sa, "failed to reset statistics (rc = %d)", rc);
+
+       sfc_sw_xstats_reset(sa);
+
+       sfc_adapter_unlock(sa);
+
+       SFC_ASSERT(rc >= 0);
+       return -rc;
+}
+
+static unsigned int
+sfc_xstats_get_nb_supported(struct sfc_adapter *sa)
+{
+       struct sfc_port *port = &sa->port;
+       unsigned int nb_supported;
+
+       sfc_adapter_lock(sa);
+       nb_supported = port->mac_stats_nb_supported +
+                      sfc_sw_xstats_get_nb_supported(sa);
+       sfc_adapter_unlock(sa);
+
+       return nb_supported;
 }
 
 static int
 sfc_xstats_get(struct rte_eth_dev *dev, struct rte_eth_xstat *xstats,
               unsigned int xstats_count)
 {
-       struct sfc_adapter *sa = dev->data->dev_private;
-       struct sfc_port *port = &sa->port;
-       uint64_t *mac_stats;
+       struct sfc_adapter *sa = sfc_adapter_by_eth_dev(dev);
+       unsigned int nb_written = 0;
+       unsigned int nb_supported = 0;
        int rc;
-       unsigned int i;
-       int nstats = 0;
-
-       rte_spinlock_lock(&port->mac_stats_lock);
-
-       rc = sfc_port_update_mac_stats(sa);
-       if (rc != 0) {
-               SFC_ASSERT(rc > 0);
-               nstats = -rc;
-               goto unlock;
-       }
 
-       mac_stats = port->mac_stats_buf;
+       if (unlikely(xstats == NULL))
+               return sfc_xstats_get_nb_supported(sa);
 
-       for (i = 0; i < EFX_MAC_NSTATS; ++i) {
-               if (EFX_MAC_STAT_SUPPORTED(port->mac_stats_mask, i)) {
-                       if (xstats != NULL && nstats < (int)xstats_count) {
-                               xstats[nstats].id = nstats;
-                               xstats[nstats].value = mac_stats[i];
-                       }
-                       nstats++;
-               }
-       }
+       rc = sfc_port_get_mac_stats(sa, xstats, xstats_count, &nb_written);
+       if (rc < 0)
+               return rc;
 
-unlock:
-       rte_spinlock_unlock(&port->mac_stats_lock);
+       nb_supported = rc;
+       sfc_sw_xstats_get_vals(sa, xstats, xstats_count, &nb_written,
+                              &nb_supported);
 
-       return nstats;
+       return nb_supported;
 }
 
 static int
@@ -669,21 +842,35 @@ sfc_xstats_get_names(struct rte_eth_dev *dev,
                     struct rte_eth_xstat_name *xstats_names,
                     unsigned int xstats_count)
 {
-       struct sfc_adapter *sa = dev->data->dev_private;
+       struct sfc_adapter *sa = sfc_adapter_by_eth_dev(dev);
        struct sfc_port *port = &sa->port;
        unsigned int i;
        unsigned int nstats = 0;
+       unsigned int nb_written = 0;
+       int ret;
+
+       if (unlikely(xstats_names == NULL))
+               return sfc_xstats_get_nb_supported(sa);
 
        for (i = 0; i < EFX_MAC_NSTATS; ++i) {
                if (EFX_MAC_STAT_SUPPORTED(port->mac_stats_mask, i)) {
-                       if (xstats_names != NULL && nstats < xstats_count)
+                       if (nstats < xstats_count) {
                                strlcpy(xstats_names[nstats].name,
                                        efx_mac_stat_name(sa->nic, i),
                                        sizeof(xstats_names[0].name));
+                               nb_written++;
+                       }
                        nstats++;
                }
        }
 
+       ret = sfc_sw_xstats_get_names(sa, xstats_names, xstats_count,
+                                     &nb_written, &nstats);
+       if (ret != 0) {
+               SFC_ASSERT(ret < 0);
+               return ret;
+       }
+
        return nstats;
 }
 
@@ -691,91 +878,109 @@ static int
 sfc_xstats_get_by_id(struct rte_eth_dev *dev, const uint64_t *ids,
                     uint64_t *values, unsigned int n)
 {
-       struct sfc_adapter *sa = dev->data->dev_private;
+       struct sfc_adapter *sa = sfc_adapter_by_eth_dev(dev);
        struct sfc_port *port = &sa->port;
-       uint64_t *mac_stats;
-       unsigned int nb_supported = 0;
-       unsigned int nb_written = 0;
+       unsigned int nb_supported;
        unsigned int i;
-       int ret;
        int rc;
 
-       if (unlikely(values == NULL) ||
-           unlikely((ids == NULL) && (n < port->mac_stats_nb_supported)))
-               return port->mac_stats_nb_supported;
-
-       rte_spinlock_lock(&port->mac_stats_lock);
-
-       rc = sfc_port_update_mac_stats(sa);
-       if (rc != 0) {
-               SFC_ASSERT(rc > 0);
-               ret = -rc;
-               goto unlock;
-       }
+       if (unlikely(ids == NULL || values == NULL))
+               return -EINVAL;
 
-       mac_stats = port->mac_stats_buf;
+       /*
+        * Values array could be filled in nonsequential order. Fill values with
+        * constant indicating invalid ID first.
+        */
+       for (i = 0; i < n; i++)
+               values[i] = SFC_XSTAT_ID_INVALID_VAL;
 
-       for (i = 0; (i < EFX_MAC_NSTATS) && (nb_written < n); ++i) {
-               if (!EFX_MAC_STAT_SUPPORTED(port->mac_stats_mask, i))
-                       continue;
+       rc = sfc_port_get_mac_stats_by_id(sa, ids, values, n);
+       if (rc != 0)
+               return rc;
 
-               if ((ids == NULL) || (ids[nb_written] == nb_supported))
-                       values[nb_written++] = mac_stats[i];
+       nb_supported = port->mac_stats_nb_supported;
+       sfc_sw_xstats_get_vals_by_id(sa, ids, values, n, &nb_supported);
 
-               ++nb_supported;
+       /* Return number of written stats before invalid ID is encountered. */
+       for (i = 0; i < n; i++) {
+               if (values[i] == SFC_XSTAT_ID_INVALID_VAL)
+                       return i;
        }
 
-       ret = nb_written;
-
-unlock:
-       rte_spinlock_unlock(&port->mac_stats_lock);
-
-       return ret;
+       return n;
 }
 
 static int
 sfc_xstats_get_names_by_id(struct rte_eth_dev *dev,
+                          const uint64_t *ids,
                           struct rte_eth_xstat_name *xstats_names,
-                          const uint64_t *ids, unsigned int size)
+                          unsigned int size)
 {
-       struct sfc_adapter *sa = dev->data->dev_private;
+       struct sfc_adapter *sa = sfc_adapter_by_eth_dev(dev);
        struct sfc_port *port = &sa->port;
-       unsigned int nb_supported = 0;
-       unsigned int nb_written = 0;
+       unsigned int nb_supported;
        unsigned int i;
+       int ret;
 
-       if (unlikely(xstats_names == NULL) ||
-           unlikely((ids == NULL) && (size < port->mac_stats_nb_supported)))
-               return port->mac_stats_nb_supported;
+       if (unlikely(xstats_names == NULL && ids != NULL) ||
+           unlikely(xstats_names != NULL && ids == NULL))
+               return -EINVAL;
 
-       for (i = 0; (i < EFX_MAC_NSTATS) && (nb_written < size); ++i) {
-               if (!EFX_MAC_STAT_SUPPORTED(port->mac_stats_mask, i))
-                       continue;
+       if (unlikely(xstats_names == NULL && ids == NULL))
+               return sfc_xstats_get_nb_supported(sa);
+
+       /*
+        * Names array could be filled in nonsequential order. Fill names with
+        * string indicating invalid ID first.
+        */
+       for (i = 0; i < size; i++)
+               xstats_names[i].name[0] = SFC_XSTAT_ID_INVALID_NAME;
+
+       sfc_adapter_lock(sa);
 
-               if ((ids == NULL) || (ids[nb_written] == nb_supported)) {
-                       char *name = xstats_names[nb_written++].name;
+       SFC_ASSERT(port->mac_stats_nb_supported <=
+                  RTE_DIM(port->mac_stats_by_id));
 
-                       strlcpy(name, efx_mac_stat_name(sa->nic, i),
+       for (i = 0; i < size; i++) {
+               if (ids[i] < port->mac_stats_nb_supported) {
+                       strlcpy(xstats_names[i].name,
+                               efx_mac_stat_name(sa->nic,
+                                                port->mac_stats_by_id[ids[i]]),
                                sizeof(xstats_names[0].name));
                }
+       }
+
+       nb_supported = port->mac_stats_nb_supported;
+
+       sfc_adapter_unlock(sa);
+
+       ret = sfc_sw_xstats_get_names_by_id(sa, ids, xstats_names, size,
+                                           &nb_supported);
+       if (ret != 0) {
+               SFC_ASSERT(ret < 0);
+               return ret;
+       }
 
-               ++nb_supported;
+       /* Return number of written names before invalid ID is encountered. */
+       for (i = 0; i < size; i++) {
+               if (xstats_names[i].name[0] == SFC_XSTAT_ID_INVALID_NAME)
+                       return i;
        }
 
-       return nb_written;
+       return size;
 }
 
 static int
 sfc_flow_ctrl_get(struct rte_eth_dev *dev, struct rte_eth_fc_conf *fc_conf)
 {
-       struct sfc_adapter *sa = dev->data->dev_private;
+       struct sfc_adapter *sa = sfc_adapter_by_eth_dev(dev);
        unsigned int wanted_fc, link_fc;
 
        memset(fc_conf, 0, sizeof(*fc_conf));
 
        sfc_adapter_lock(sa);
 
-       if (sa->state == SFC_ADAPTER_STARTED)
+       if (sa->state == SFC_ETHDEV_STARTED)
                efx_mac_fcntl_get(sa->nic, &wanted_fc, &link_fc);
        else
                link_fc = sa->port.flow_ctrl;
@@ -808,7 +1013,7 @@ sfc_flow_ctrl_get(struct rte_eth_dev *dev, struct rte_eth_fc_conf *fc_conf)
 static int
 sfc_flow_ctrl_set(struct rte_eth_dev *dev, struct rte_eth_fc_conf *fc_conf)
 {
-       struct sfc_adapter *sa = dev->data->dev_private;
+       struct sfc_adapter *sa = sfc_adapter_by_eth_dev(dev);
        struct sfc_port *port = &sa->port;
        unsigned int fcntl;
        int rc;
@@ -841,7 +1046,7 @@ sfc_flow_ctrl_set(struct rte_eth_dev *dev, struct rte_eth_fc_conf *fc_conf)
 
        sfc_adapter_lock(sa);
 
-       if (sa->state == SFC_ADAPTER_STARTED) {
+       if (sa->state == SFC_ETHDEV_STARTED) {
                rc = efx_mac_fcntl_set(sa->nic, fcntl, fc_conf->autoneg);
                if (rc != 0)
                        goto fail_mac_fcntl_set;
@@ -861,10 +1066,39 @@ fail_inval:
        return -rc;
 }
 
+static int
+sfc_check_scatter_on_all_rx_queues(struct sfc_adapter *sa, size_t pdu)
+{
+       struct sfc_adapter_shared * const sas = sfc_sa2shared(sa);
+       const efx_nic_cfg_t *encp = efx_nic_cfg_get(sa->nic);
+       boolean_t scatter_enabled;
+       const char *error;
+       unsigned int i;
+
+       for (i = 0; i < sas->rxq_count; i++) {
+               if ((sas->rxq_info[i].state & SFC_RXQ_INITIALIZED) == 0)
+                       continue;
+
+               scatter_enabled = (sas->rxq_info[i].type_flags &
+                                  EFX_RXQ_FLAG_SCATTER);
+
+               if (!sfc_rx_check_scatter(pdu, sa->rxq_ctrl[i].buf_size,
+                                         encp->enc_rx_prefix_size,
+                                         scatter_enabled,
+                                         encp->enc_rx_scatter_max, &error)) {
+                       sfc_err(sa, "MTU check for RxQ %u failed: %s", i,
+                               error);
+                       return EINVAL;
+               }
+       }
+
+       return 0;
+}
+
 static int
 sfc_dev_set_mtu(struct rte_eth_dev *dev, uint16_t mtu)
 {
-       struct sfc_adapter *sa = dev->data->dev_private;
+       struct sfc_adapter *sa = sfc_adapter_by_eth_dev(dev);
        size_t pdu = EFX_MAC_PDU(mtu);
        size_t old_pdu;
        int rc;
@@ -881,14 +1115,18 @@ sfc_dev_set_mtu(struct rte_eth_dev *dev, uint16_t mtu)
        if (pdu > EFX_MAC_PDU_MAX) {
                sfc_err(sa, "too big MTU %u (PDU size %u greater than max %u)",
                        (unsigned int)mtu, (unsigned int)pdu,
-                       EFX_MAC_PDU_MAX);
+                       (unsigned int)EFX_MAC_PDU_MAX);
                goto fail_inval;
        }
 
        sfc_adapter_lock(sa);
 
+       rc = sfc_check_scatter_on_all_rx_queues(sa, pdu);
+       if (rc != 0)
+               goto fail_check_scatter;
+
        if (pdu != sa->port.pdu) {
-               if (sa->state == SFC_ADAPTER_STARTED) {
+               if (sa->state == SFC_ETHDEV_STARTED) {
                        sfc_stop(sa);
 
                        old_pdu = sa->port.pdu;
@@ -905,7 +1143,7 @@ sfc_dev_set_mtu(struct rte_eth_dev *dev, uint16_t mtu)
         * The driver does not use it, but other PMDs update jumbo frame
         * flag and max_rx_pkt_len when MTU is set.
         */
-       if (mtu > ETHER_MAX_LEN) {
+       if (mtu > RTE_ETHER_MTU) {
                struct rte_eth_rxmode *rxmode = &dev->data->dev_conf.rxmode;
                rxmode->offloads |= DEV_RX_OFFLOAD_JUMBO_FRAME;
        }
@@ -923,6 +1161,8 @@ fail_start:
                sfc_err(sa, "cannot start with neither new (%u) nor old (%u) "
                        "PDU max size - port is stopped",
                        (unsigned int)pdu, (unsigned int)old_pdu);
+
+fail_check_scatter:
        sfc_adapter_unlock(sa);
 
 fail_inval:
@@ -931,21 +1171,24 @@ fail_inval:
        return -rc;
 }
 static int
-sfc_mac_addr_set(struct rte_eth_dev *dev, struct ether_addr *mac_addr)
+sfc_mac_addr_set(struct rte_eth_dev *dev, struct rte_ether_addr *mac_addr)
 {
-       struct sfc_adapter *sa = dev->data->dev_private;
+       struct sfc_adapter *sa = sfc_adapter_by_eth_dev(dev);
        const efx_nic_cfg_t *encp = efx_nic_cfg_get(sa->nic);
        struct sfc_port *port = &sa->port;
-       struct ether_addr *old_addr = &dev->data->mac_addrs[0];
+       struct rte_ether_addr *old_addr = &dev->data->mac_addrs[0];
        int rc = 0;
 
        sfc_adapter_lock(sa);
 
+       if (rte_is_same_ether_addr(mac_addr, &port->default_mac_addr))
+               goto unlock;
+
        /*
         * Copy the address to the device private data so that
         * it could be recalled in the case of adapter restart.
         */
-       ether_addr_copy(mac_addr, &port->default_mac_addr);
+       rte_ether_addr_copy(mac_addr, &port->default_mac_addr);
 
        /*
         * Neither of the two following checks can return
@@ -954,13 +1197,13 @@ sfc_mac_addr_set(struct rte_eth_dev *dev, struct ether_addr *mac_addr)
         * on the next port start if the user prevents
         * isolated mode from being enabled.
         */
-       if (port->isolated) {
+       if (sfc_sa2shared(sa)->isolated) {
                sfc_warn(sa, "isolated mode is active on the port");
                sfc_warn(sa, "will not set MAC address");
                goto unlock;
        }
 
-       if (sa->state != SFC_ADAPTER_STARTED) {
+       if (sa->state != SFC_ETHDEV_STARTED) {
                sfc_notice(sa, "the port is not started");
                sfc_notice(sa, "the new MAC address will be set on port start");
 
@@ -979,12 +1222,12 @@ sfc_mac_addr_set(struct rte_eth_dev *dev, struct ether_addr *mac_addr)
                 * has no effect on received traffic, therefore
                 * we also need to update unicast filters
                 */
-               rc = sfc_set_rx_mode(sa);
+               rc = sfc_set_rx_mode_unchecked(sa);
                if (rc != 0) {
                        sfc_err(sa, "cannot set filter (rc = %u)", rc);
                        /* Rollback the old address */
                        (void)efx_mac_addr_set(sa->nic, old_addr->addr_bytes);
-                       (void)sfc_set_rx_mode(sa);
+                       (void)sfc_set_rx_mode_unchecked(sa);
                }
        } else {
                sfc_warn(sa, "cannot set MAC address with filters installed");
@@ -1005,7 +1248,7 @@ sfc_mac_addr_set(struct rte_eth_dev *dev, struct ether_addr *mac_addr)
 
 unlock:
        if (rc != 0)
-               ether_addr_copy(old_addr, &port->default_mac_addr);
+               rte_ether_addr_copy(old_addr, &port->default_mac_addr);
 
        sfc_adapter_unlock(sa);
 
@@ -1015,16 +1258,16 @@ unlock:
 
 
 static int
-sfc_set_mc_addr_list(struct rte_eth_dev *dev, struct ether_addr *mc_addr_set,
-                    uint32_t nb_mc_addr)
+sfc_set_mc_addr_list(struct rte_eth_dev *dev,
+               struct rte_ether_addr *mc_addr_set, uint32_t nb_mc_addr)
 {
-       struct sfc_adapter *sa = dev->data->dev_private;
+       struct sfc_adapter *sa = sfc_adapter_by_eth_dev(dev);
        struct sfc_port *port = &sa->port;
        uint8_t *mc_addrs = port->mcast_addrs;
        int rc;
        unsigned int i;
 
-       if (port->isolated) {
+       if (sfc_sa2shared(sa)->isolated) {
                sfc_err(sa, "isolated mode is active on the port");
                sfc_err(sa, "will not set multicast address list");
                return -ENOTSUP;
@@ -1047,7 +1290,7 @@ sfc_set_mc_addr_list(struct rte_eth_dev *dev, struct ether_addr *mc_addr_set,
 
        port->nb_mcast_addrs = nb_mc_addr;
 
-       if (sa->state != SFC_ADAPTER_STARTED)
+       if (sa->state != SFC_ETHDEV_STARTED)
                return 0;
 
        rc = efx_mac_multicast_list_set(sa->nic, port->mcast_addrs,
@@ -1064,17 +1307,14 @@ sfc_set_mc_addr_list(struct rte_eth_dev *dev, struct ether_addr *mc_addr_set,
  * use any process-local pointers from the adapter data.
  */
 static void
-sfc_rx_queue_info_get(struct rte_eth_dev *dev, uint16_t rx_queue_id,
+sfc_rx_queue_info_get(struct rte_eth_dev *dev, uint16_t ethdev_qid,
                      struct rte_eth_rxq_info *qinfo)
 {
-       struct sfc_adapter *sa = dev->data->dev_private;
+       struct sfc_adapter_shared *sas = sfc_adapter_shared_by_eth_dev(dev);
+       sfc_ethdev_qid_t sfc_ethdev_qid = ethdev_qid;
        struct sfc_rxq_info *rxq_info;
 
-       sfc_adapter_lock(sa);
-
-       SFC_ASSERT(rx_queue_id < sa->rxq_count);
-
-       rxq_info = &sa->rxq_info[rx_queue_id];
+       rxq_info = sfc_rxq_info_by_ethdev_qid(sas, sfc_ethdev_qid);
 
        qinfo->mp = rxq_info->refill_mb_pool;
        qinfo->conf.rx_free_thresh = rxq_info->refill_threshold;
@@ -1086,8 +1326,6 @@ sfc_rx_queue_info_get(struct rte_eth_dev *dev, uint16_t rx_queue_id,
                qinfo->scattered_rx = 1;
        }
        qinfo->nb_desc = rxq_info->entries;
-
-       sfc_adapter_unlock(sa);
 }
 
 /*
@@ -1095,17 +1333,15 @@ sfc_rx_queue_info_get(struct rte_eth_dev *dev, uint16_t rx_queue_id,
  * use any process-local pointers from the adapter data.
  */
 static void
-sfc_tx_queue_info_get(struct rte_eth_dev *dev, uint16_t tx_queue_id,
+sfc_tx_queue_info_get(struct rte_eth_dev *dev, uint16_t ethdev_qid,
                      struct rte_eth_txq_info *qinfo)
 {
-       struct sfc_adapter *sa = dev->data->dev_private;
+       struct sfc_adapter_shared *sas = sfc_adapter_shared_by_eth_dev(dev);
        struct sfc_txq_info *txq_info;
 
-       sfc_adapter_lock(sa);
-
-       SFC_ASSERT(tx_queue_id < sa->txq_count);
+       SFC_ASSERT(ethdev_qid < sas->ethdev_txq_count);
 
-       txq_info = &sa->txq_info[tx_queue_id];
+       txq_info = sfc_txq_info_by_ethdev_qid(sas, ethdev_qid);
 
        memset(qinfo, 0, sizeof(*qinfo));
 
@@ -1113,8 +1349,6 @@ sfc_tx_queue_info_get(struct rte_eth_dev *dev, uint16_t tx_queue_id,
        qinfo->conf.tx_free_thresh = txq_info->free_thresh;
        qinfo->conf.tx_deferred_start = txq_info->deferred_start;
        qinfo->nb_desc = txq_info->entries;
-
-       sfc_adapter_unlock(sa);
 }
 
 /*
@@ -1122,14 +1356,14 @@ sfc_tx_queue_info_get(struct rte_eth_dev *dev, uint16_t tx_queue_id,
  * use any process-local pointers from the adapter data.
  */
 static uint32_t
-sfc_rx_queue_count(struct rte_eth_dev *dev, uint16_t rx_queue_id)
+sfc_rx_queue_count(struct rte_eth_dev *dev, uint16_t ethdev_qid)
 {
        const struct sfc_adapter_priv *sap = sfc_adapter_priv_by_eth_dev(dev);
-       struct sfc_adapter *sa = dev->data->dev_private;
+       struct sfc_adapter_shared *sas = sfc_adapter_shared_by_eth_dev(dev);
+       sfc_ethdev_qid_t sfc_ethdev_qid = ethdev_qid;
        struct sfc_rxq_info *rxq_info;
 
-       SFC_ASSERT(rx_queue_id < sa->rxq_count);
-       rxq_info = &sa->rxq_info[rx_queue_id];
+       rxq_info = sfc_rxq_info_by_ethdev_qid(sas, sfc_ethdev_qid);
 
        if ((rxq_info->state & SFC_RXQ_STARTED) == 0)
                return 0;
@@ -1137,21 +1371,6 @@ sfc_rx_queue_count(struct rte_eth_dev *dev, uint16_t rx_queue_id)
        return sap->dp_rx->qdesc_npending(rxq_info->dp);
 }
 
-/*
- * The function is used by the secondary process as well. It must not
- * use any process-local pointers from the adapter data.
- */
-static int
-sfc_rx_descriptor_done(void *queue, uint16_t offset)
-{
-       struct sfc_dp_rxq *dp_rxq = queue;
-       const struct sfc_dp_rx *dp_rx;
-
-       dp_rx = sfc_dp_rx_by_dp_rxq(dp_rxq);
-
-       return offset < dp_rx->qdesc_npending(dp_rxq);
-}
-
 /*
  * The function is used by the secondary process as well. It must not
  * use any process-local pointers from the adapter data.
@@ -1183,27 +1402,33 @@ sfc_tx_descriptor_status(void *queue, uint16_t offset)
 }
 
 static int
-sfc_rx_queue_start(struct rte_eth_dev *dev, uint16_t rx_queue_id)
+sfc_rx_queue_start(struct rte_eth_dev *dev, uint16_t ethdev_qid)
 {
-       struct sfc_adapter *sa = dev->data->dev_private;
+       struct sfc_adapter_shared *sas = sfc_adapter_shared_by_eth_dev(dev);
+       struct sfc_adapter *sa = sfc_adapter_by_eth_dev(dev);
+       sfc_ethdev_qid_t sfc_ethdev_qid = ethdev_qid;
+       struct sfc_rxq_info *rxq_info;
+       sfc_sw_index_t sw_index;
        int rc;
 
-       sfc_log_init(sa, "RxQ=%u", rx_queue_id);
+       sfc_log_init(sa, "RxQ=%u", ethdev_qid);
 
        sfc_adapter_lock(sa);
 
        rc = EINVAL;
-       if (sa->state != SFC_ADAPTER_STARTED)
+       if (sa->state != SFC_ETHDEV_STARTED)
                goto fail_not_started;
 
-       if (sa->rxq_info[rx_queue_id].state != SFC_RXQ_INITIALIZED)
+       rxq_info = sfc_rxq_info_by_ethdev_qid(sas, sfc_ethdev_qid);
+       if (rxq_info->state != SFC_RXQ_INITIALIZED)
                goto fail_not_setup;
 
-       rc = sfc_rx_qstart(sa, rx_queue_id);
+       sw_index = sfc_rxq_sw_index_by_ethdev_rx_qid(sas, sfc_ethdev_qid);
+       rc = sfc_rx_qstart(sa, sw_index);
        if (rc != 0)
                goto fail_rx_qstart;
 
-       sa->rxq_info[rx_queue_id].deferred_started = B_TRUE;
+       rxq_info->deferred_started = B_TRUE;
 
        sfc_adapter_unlock(sa);
 
@@ -1218,16 +1443,23 @@ fail_not_started:
 }
 
 static int
-sfc_rx_queue_stop(struct rte_eth_dev *dev, uint16_t rx_queue_id)
+sfc_rx_queue_stop(struct rte_eth_dev *dev, uint16_t ethdev_qid)
 {
-       struct sfc_adapter *sa = dev->data->dev_private;
+       struct sfc_adapter_shared *sas = sfc_adapter_shared_by_eth_dev(dev);
+       struct sfc_adapter *sa = sfc_adapter_by_eth_dev(dev);
+       sfc_ethdev_qid_t sfc_ethdev_qid = ethdev_qid;
+       struct sfc_rxq_info *rxq_info;
+       sfc_sw_index_t sw_index;
 
-       sfc_log_init(sa, "RxQ=%u", rx_queue_id);
+       sfc_log_init(sa, "RxQ=%u", ethdev_qid);
 
        sfc_adapter_lock(sa);
-       sfc_rx_qstop(sa, rx_queue_id);
 
-       sa->rxq_info[rx_queue_id].deferred_started = B_FALSE;
+       sw_index = sfc_rxq_sw_index_by_ethdev_rx_qid(sas, sfc_ethdev_qid);
+       sfc_rx_qstop(sa, sw_index);
+
+       rxq_info = sfc_rxq_info_by_ethdev_qid(sas, sfc_ethdev_qid);
+       rxq_info->deferred_started = B_FALSE;
 
        sfc_adapter_unlock(sa);
 
@@ -1235,27 +1467,32 @@ sfc_rx_queue_stop(struct rte_eth_dev *dev, uint16_t rx_queue_id)
 }
 
 static int
-sfc_tx_queue_start(struct rte_eth_dev *dev, uint16_t tx_queue_id)
+sfc_tx_queue_start(struct rte_eth_dev *dev, uint16_t ethdev_qid)
 {
-       struct sfc_adapter *sa = dev->data->dev_private;
+       struct sfc_adapter_shared *sas = sfc_adapter_shared_by_eth_dev(dev);
+       struct sfc_adapter *sa = sfc_adapter_by_eth_dev(dev);
+       struct sfc_txq_info *txq_info;
+       sfc_sw_index_t sw_index;
        int rc;
 
-       sfc_log_init(sa, "TxQ = %u", tx_queue_id);
+       sfc_log_init(sa, "TxQ = %u", ethdev_qid);
 
        sfc_adapter_lock(sa);
 
        rc = EINVAL;
-       if (sa->state != SFC_ADAPTER_STARTED)
+       if (sa->state != SFC_ETHDEV_STARTED)
                goto fail_not_started;
 
-       if (sa->txq_info[tx_queue_id].state != SFC_TXQ_INITIALIZED)
+       txq_info = sfc_txq_info_by_ethdev_qid(sas, ethdev_qid);
+       if (txq_info->state != SFC_TXQ_INITIALIZED)
                goto fail_not_setup;
 
-       rc = sfc_tx_qstart(sa, tx_queue_id);
+       sw_index = sfc_txq_sw_index_by_ethdev_tx_qid(sas, ethdev_qid);
+       rc = sfc_tx_qstart(sa, sw_index);
        if (rc != 0)
                goto fail_tx_qstart;
 
-       sa->txq_info[tx_queue_id].deferred_started = B_TRUE;
+       txq_info->deferred_started = B_TRUE;
 
        sfc_adapter_unlock(sa);
        return 0;
@@ -1270,17 +1507,22 @@ fail_not_started:
 }
 
 static int
-sfc_tx_queue_stop(struct rte_eth_dev *dev, uint16_t tx_queue_id)
+sfc_tx_queue_stop(struct rte_eth_dev *dev, uint16_t ethdev_qid)
 {
-       struct sfc_adapter *sa = dev->data->dev_private;
+       struct sfc_adapter_shared *sas = sfc_adapter_shared_by_eth_dev(dev);
+       struct sfc_adapter *sa = sfc_adapter_by_eth_dev(dev);
+       struct sfc_txq_info *txq_info;
+       sfc_sw_index_t sw_index;
 
-       sfc_log_init(sa, "TxQ = %u", tx_queue_id);
+       sfc_log_init(sa, "TxQ = %u", ethdev_qid);
 
        sfc_adapter_lock(sa);
 
-       sfc_tx_qstop(sa, tx_queue_id);
+       sw_index = sfc_txq_sw_index_by_ethdev_tx_qid(sas, ethdev_qid);
+       sfc_tx_qstop(sa, sw_index);
 
-       sa->txq_info[tx_queue_id].deferred_started = B_FALSE;
+       txq_info = sfc_txq_info_by_ethdev_qid(sas, ethdev_qid);
+       txq_info->deferred_started = B_FALSE;
 
        sfc_adapter_unlock(sa);
        return 0;
@@ -1309,7 +1551,7 @@ sfc_dev_udp_tunnel_op(struct rte_eth_dev *dev,
                      struct rte_eth_udp_tunnel *tunnel_udp,
                      enum sfc_udp_tunnel_op_e op)
 {
-       struct sfc_adapter *sa = dev->data->dev_private;
+       struct sfc_adapter *sa = sfc_adapter_by_eth_dev(dev);
        efx_tunnel_protocol_t tunnel_proto;
        int rc;
 
@@ -1346,7 +1588,7 @@ sfc_dev_udp_tunnel_op(struct rte_eth_dev *dev,
        if (rc != 0)
                goto fail_op;
 
-       if (sa->state == SFC_ADAPTER_STARTED) {
+       if (sa->state == SFC_ETHDEV_STARTED) {
                rc = efx_tunnel_reconfigure(sa->nic);
                if (rc == EAGAIN) {
                        /*
@@ -1411,27 +1653,23 @@ static int
 sfc_dev_rss_hash_conf_get(struct rte_eth_dev *dev,
                          struct rte_eth_rss_conf *rss_conf)
 {
-       struct sfc_adapter *sa = dev->data->dev_private;
-       struct sfc_rss *rss = &sa->rss;
+       struct sfc_adapter_shared *sas = sfc_adapter_shared_by_eth_dev(dev);
+       struct sfc_rss *rss = &sas->rss;
 
        if (rss->context_type != EFX_RX_SCALE_EXCLUSIVE)
                return -ENOTSUP;
 
-       sfc_adapter_lock(sa);
-
        /*
         * Mapping of hash configuration between RTE and EFX is not one-to-one,
         * hence, conversion is done here to derive a correct set of ETH_RSS
         * flags which corresponds to the active EFX configuration stored
         * locally in 'sfc_adapter' and kept up-to-date
         */
-       rss_conf->rss_hf = sfc_rx_hf_efx_to_rte(sa, rss->hash_types);
+       rss_conf->rss_hf = sfc_rx_hf_efx_to_rte(rss, rss->hash_types);
        rss_conf->rss_key_len = EFX_RSS_KEY_SIZE;
        if (rss_conf->rss_key != NULL)
                rte_memcpy(rss_conf->rss_key, rss->key, EFX_RSS_KEY_SIZE);
 
-       sfc_adapter_unlock(sa);
-
        return 0;
 }
 
@@ -1439,13 +1677,19 @@ static int
 sfc_dev_rss_hash_update(struct rte_eth_dev *dev,
                        struct rte_eth_rss_conf *rss_conf)
 {
-       struct sfc_adapter *sa = dev->data->dev_private;
-       struct sfc_rss *rss = &sa->rss;
-       struct sfc_port *port = &sa->port;
+       struct sfc_adapter *sa = sfc_adapter_by_eth_dev(dev);
+       struct sfc_rss *rss = &sfc_sa2shared(sa)->rss;
        unsigned int efx_hash_types;
+       uint32_t contexts[] = {EFX_RSS_CONTEXT_DEFAULT, rss->dummy_rss_context};
+       unsigned int n_contexts;
+       unsigned int mode_i = 0;
+       unsigned int key_i = 0;
+       unsigned int i = 0;
        int rc = 0;
 
-       if (port->isolated)
+       n_contexts = rss->dummy_rss_context == EFX_RSS_CONTEXT_DEFAULT ? 1 : 2;
+
+       if (sfc_sa2shared(sa)->isolated)
                return -ENOTSUP;
 
        if (rss->context_type != EFX_RX_SCALE_EXCLUSIVE) {
@@ -1460,7 +1704,7 @@ sfc_dev_rss_hash_update(struct rte_eth_dev *dev,
 
        if ((rss_conf->rss_key != NULL) &&
            (rss_conf->rss_key_len != sizeof(rss->key))) {
-               sfc_err(sa, "RSS key size is wrong (should be %lu)",
+               sfc_err(sa, "RSS key size is wrong (should be %zu)",
                        sizeof(rss->key));
                return -EINVAL;
        }
@@ -1471,19 +1715,24 @@ sfc_dev_rss_hash_update(struct rte_eth_dev *dev,
        if (rc != 0)
                goto fail_rx_hf_rte_to_efx;
 
-       rc = efx_rx_scale_mode_set(sa->nic, EFX_RSS_CONTEXT_DEFAULT,
-                                  rss->hash_alg, efx_hash_types, B_TRUE);
-       if (rc != 0)
-               goto fail_scale_mode_set;
+       for (mode_i = 0; mode_i < n_contexts; mode_i++) {
+               rc = efx_rx_scale_mode_set(sa->nic, contexts[mode_i],
+                                          rss->hash_alg, efx_hash_types,
+                                          B_TRUE);
+               if (rc != 0)
+                       goto fail_scale_mode_set;
+       }
 
        if (rss_conf->rss_key != NULL) {
-               if (sa->state == SFC_ADAPTER_STARTED) {
-                       rc = efx_rx_scale_key_set(sa->nic,
-                                                 EFX_RSS_CONTEXT_DEFAULT,
-                                                 rss_conf->rss_key,
-                                                 sizeof(rss->key));
-                       if (rc != 0)
-                               goto fail_scale_key_set;
+               if (sa->state == SFC_ETHDEV_STARTED) {
+                       for (key_i = 0; key_i < n_contexts; key_i++) {
+                               rc = efx_rx_scale_key_set(sa->nic,
+                                                         contexts[key_i],
+                                                         rss_conf->rss_key,
+                                                         sizeof(rss->key));
+                               if (rc != 0)
+                                       goto fail_scale_key_set;
+                       }
                }
 
                rte_memcpy(rss->key, rss_conf->rss_key, sizeof(rss->key));
@@ -1496,12 +1745,20 @@ sfc_dev_rss_hash_update(struct rte_eth_dev *dev,
        return 0;
 
 fail_scale_key_set:
-       if (efx_rx_scale_mode_set(sa->nic, EFX_RSS_CONTEXT_DEFAULT,
-                                 EFX_RX_HASHALG_TOEPLITZ,
-                                 rss->hash_types, B_TRUE) != 0)
-               sfc_err(sa, "failed to restore RSS mode");
+       for (i = 0; i < key_i; i++) {
+               if (efx_rx_scale_key_set(sa->nic, contexts[i], rss->key,
+                                        sizeof(rss->key)) != 0)
+                       sfc_err(sa, "failed to restore RSS key");
+       }
 
 fail_scale_mode_set:
+       for (i = 0; i < mode_i; i++) {
+               if (efx_rx_scale_mode_set(sa->nic, contexts[i],
+                                         EFX_RX_HASHALG_TOEPLITZ,
+                                         rss->hash_types, B_TRUE) != 0)
+                       sfc_err(sa, "failed to restore RSS mode");
+       }
+
 fail_rx_hf_rte_to_efx:
        sfc_adapter_unlock(sa);
        return -rc;
@@ -1516,12 +1773,11 @@ sfc_dev_rss_reta_query(struct rte_eth_dev *dev,
                       struct rte_eth_rss_reta_entry64 *reta_conf,
                       uint16_t reta_size)
 {
-       struct sfc_adapter *sa = dev->data->dev_private;
-       struct sfc_rss *rss = &sa->rss;
-       struct sfc_port *port = &sa->port;
+       struct sfc_adapter_shared *sas = sfc_adapter_shared_by_eth_dev(dev);
+       struct sfc_rss *rss = &sas->rss;
        int entry;
 
-       if (rss->context_type != EFX_RX_SCALE_EXCLUSIVE || port->isolated)
+       if (rss->context_type != EFX_RX_SCALE_EXCLUSIVE || sas->isolated)
                return -ENOTSUP;
 
        if (rss->channels == 0)
@@ -1530,8 +1786,6 @@ sfc_dev_rss_reta_query(struct rte_eth_dev *dev,
        if (reta_size != EFX_RSS_TBL_SIZE)
                return -EINVAL;
 
-       sfc_adapter_lock(sa);
-
        for (entry = 0; entry < reta_size; entry++) {
                int grp = entry / RTE_RETA_GROUP_SIZE;
                int grp_idx = entry % RTE_RETA_GROUP_SIZE;
@@ -1540,8 +1794,6 @@ sfc_dev_rss_reta_query(struct rte_eth_dev *dev,
                        reta_conf[grp].reta[grp_idx] = rss->tbl[entry];
        }
 
-       sfc_adapter_unlock(sa);
-
        return 0;
 }
 
@@ -1550,15 +1802,14 @@ sfc_dev_rss_reta_update(struct rte_eth_dev *dev,
                        struct rte_eth_rss_reta_entry64 *reta_conf,
                        uint16_t reta_size)
 {
-       struct sfc_adapter *sa = dev->data->dev_private;
-       struct sfc_rss *rss = &sa->rss;
-       struct sfc_port *port = &sa->port;
+       struct sfc_adapter *sa = sfc_adapter_by_eth_dev(dev);
+       struct sfc_rss *rss = &sfc_sa2shared(sa)->rss;
        unsigned int *rss_tbl_new;
        uint16_t entry;
        int rc = 0;
 
 
-       if (port->isolated)
+       if (sfc_sa2shared(sa)->isolated)
                return -ENOTSUP;
 
        if (rss->context_type != EFX_RX_SCALE_EXCLUSIVE) {
@@ -1600,7 +1851,7 @@ sfc_dev_rss_reta_update(struct rte_eth_dev *dev,
                }
        }
 
-       if (sa->state == SFC_ADAPTER_STARTED) {
+       if (sa->state == SFC_ETHDEV_STARTED) {
                rc = efx_rx_scale_tbl_set(sa->nic, EFX_RSS_CONTEXT_DEFAULT,
                                          rss_tbl_new, EFX_RSS_TBL_SIZE);
                if (rc != 0)
@@ -1620,59 +1871,11 @@ bad_reta_entry:
 }
 
 static int
-sfc_dev_filter_ctrl(struct rte_eth_dev *dev, enum rte_filter_type filter_type,
-                   enum rte_filter_op filter_op,
-                   void *arg)
+sfc_dev_flow_ops_get(struct rte_eth_dev *dev __rte_unused,
+                    const struct rte_flow_ops **ops)
 {
-       struct sfc_adapter *sa = dev->data->dev_private;
-       int rc = ENOTSUP;
-
-       sfc_log_init(sa, "entry");
-
-       switch (filter_type) {
-       case RTE_ETH_FILTER_NONE:
-               sfc_err(sa, "Global filters configuration not supported");
-               break;
-       case RTE_ETH_FILTER_MACVLAN:
-               sfc_err(sa, "MACVLAN filters not supported");
-               break;
-       case RTE_ETH_FILTER_ETHERTYPE:
-               sfc_err(sa, "EtherType filters not supported");
-               break;
-       case RTE_ETH_FILTER_FLEXIBLE:
-               sfc_err(sa, "Flexible filters not supported");
-               break;
-       case RTE_ETH_FILTER_SYN:
-               sfc_err(sa, "SYN filters not supported");
-               break;
-       case RTE_ETH_FILTER_NTUPLE:
-               sfc_err(sa, "NTUPLE filters not supported");
-               break;
-       case RTE_ETH_FILTER_TUNNEL:
-               sfc_err(sa, "Tunnel filters not supported");
-               break;
-       case RTE_ETH_FILTER_FDIR:
-               sfc_err(sa, "Flow Director filters not supported");
-               break;
-       case RTE_ETH_FILTER_HASH:
-               sfc_err(sa, "Hash filters not supported");
-               break;
-       case RTE_ETH_FILTER_GENERIC:
-               if (filter_op != RTE_ETH_FILTER_GET) {
-                       rc = EINVAL;
-               } else {
-                       *(const void **)arg = &sfc_flow_ops;
-                       rc = 0;
-               }
-               break;
-       default:
-               sfc_err(sa, "Unknown filter type %u", filter_type);
-               break;
-       }
-
-       sfc_log_init(sa, "exit: %d", -rc);
-       SFC_ASSERT(rc >= 0);
-       return -rc;
+       *ops = &sfc_flow_ops;
+       return 0;
 }
 
 static int
@@ -1690,6 +1893,203 @@ sfc_pool_ops_supported(struct rte_eth_dev *dev, const char *pool)
        return sap->dp_rx->pool_ops_supported(pool);
 }
 
+static int
+sfc_rx_queue_intr_enable(struct rte_eth_dev *dev, uint16_t ethdev_qid)
+{
+       const struct sfc_adapter_priv *sap = sfc_adapter_priv_by_eth_dev(dev);
+       struct sfc_adapter_shared *sas = sfc_adapter_shared_by_eth_dev(dev);
+       sfc_ethdev_qid_t sfc_ethdev_qid = ethdev_qid;
+       struct sfc_rxq_info *rxq_info;
+
+       rxq_info = sfc_rxq_info_by_ethdev_qid(sas, sfc_ethdev_qid);
+
+       return sap->dp_rx->intr_enable(rxq_info->dp);
+}
+
+static int
+sfc_rx_queue_intr_disable(struct rte_eth_dev *dev, uint16_t ethdev_qid)
+{
+       const struct sfc_adapter_priv *sap = sfc_adapter_priv_by_eth_dev(dev);
+       struct sfc_adapter_shared *sas = sfc_adapter_shared_by_eth_dev(dev);
+       sfc_ethdev_qid_t sfc_ethdev_qid = ethdev_qid;
+       struct sfc_rxq_info *rxq_info;
+
+       rxq_info = sfc_rxq_info_by_ethdev_qid(sas, sfc_ethdev_qid);
+
+       return sap->dp_rx->intr_disable(rxq_info->dp);
+}
+
+struct sfc_mport_journal_ctx {
+       struct sfc_adapter              *sa;
+       uint16_t                        switch_domain_id;
+       uint32_t                        mcdi_handle;
+       bool                            controllers_assigned;
+       efx_pcie_interface_t            *controllers;
+       size_t                          nb_controllers;
+};
+
+static int
+sfc_journal_ctx_add_controller(struct sfc_mport_journal_ctx *ctx,
+                              efx_pcie_interface_t intf)
+{
+       efx_pcie_interface_t *new_controllers;
+       size_t i, target;
+       size_t new_size;
+
+       if (ctx->controllers == NULL) {
+               ctx->controllers = rte_malloc("sfc_controller_mapping",
+                                             sizeof(ctx->controllers[0]), 0);
+               if (ctx->controllers == NULL)
+                       return ENOMEM;
+
+               ctx->controllers[0] = intf;
+               ctx->nb_controllers = 1;
+
+               return 0;
+       }
+
+       for (i = 0; i < ctx->nb_controllers; i++) {
+               if (ctx->controllers[i] == intf)
+                       return 0;
+               if (ctx->controllers[i] > intf)
+                       break;
+       }
+       target = i;
+
+       ctx->nb_controllers += 1;
+       new_size = ctx->nb_controllers * sizeof(ctx->controllers[0]);
+
+       new_controllers = rte_realloc(ctx->controllers, new_size, 0);
+       if (new_controllers == NULL) {
+               rte_free(ctx->controllers);
+               return ENOMEM;
+       }
+       ctx->controllers = new_controllers;
+
+       for (i = target + 1; i < ctx->nb_controllers; i++)
+               ctx->controllers[i] = ctx->controllers[i - 1];
+
+       ctx->controllers[target] = intf;
+
+       return 0;
+}
+
+static efx_rc_t
+sfc_process_mport_journal_entry(struct sfc_mport_journal_ctx *ctx,
+                               efx_mport_desc_t *mport)
+{
+       efx_mport_sel_t ethdev_mport;
+       int rc;
+
+       sfc_dbg(ctx->sa,
+               "processing mport id %u (controller %u pf %u vf %u)",
+               mport->emd_id.id, mport->emd_vnic.ev_intf,
+               mport->emd_vnic.ev_pf, mport->emd_vnic.ev_vf);
+       efx_mae_mport_invalid(&ethdev_mport);
+
+       if (!ctx->controllers_assigned) {
+               rc = sfc_journal_ctx_add_controller(ctx,
+                                                   mport->emd_vnic.ev_intf);
+               if (rc != 0)
+                       return rc;
+       }
+
+       return 0;
+}
+
+static efx_rc_t
+sfc_process_mport_journal_cb(void *data, efx_mport_desc_t *mport,
+                            size_t mport_len)
+{
+       struct sfc_mport_journal_ctx *ctx = data;
+
+       if (ctx == NULL || ctx->sa == NULL) {
+               sfc_err(ctx->sa, "received NULL context or SFC adapter");
+               return EINVAL;
+       }
+
+       if (mport_len != sizeof(*mport)) {
+               sfc_err(ctx->sa, "actual and expected mport buffer sizes differ");
+               return EINVAL;
+       }
+
+       SFC_ASSERT(sfc_adapter_is_locked(ctx->sa));
+
+       /*
+        * If a zombie flag is set, it means the mport has been marked for
+        * deletion and cannot be used for any new operations. The mport will
+        * be destroyed completely once all references to it are released.
+        */
+       if (mport->emd_zombie) {
+               sfc_dbg(ctx->sa, "mport is a zombie, skipping");
+               return 0;
+       }
+       if (mport->emd_type != EFX_MPORT_TYPE_VNIC) {
+               sfc_dbg(ctx->sa, "mport is not a VNIC, skipping");
+               return 0;
+       }
+       if (mport->emd_vnic.ev_client_type != EFX_MPORT_VNIC_CLIENT_FUNCTION) {
+               sfc_dbg(ctx->sa, "mport is not a function, skipping");
+               return 0;
+       }
+       if (mport->emd_vnic.ev_handle == ctx->mcdi_handle) {
+               sfc_dbg(ctx->sa, "mport is this driver instance, skipping");
+               return 0;
+       }
+
+       return sfc_process_mport_journal_entry(ctx, mport);
+}
+
+static int
+sfc_process_mport_journal(struct sfc_adapter *sa)
+{
+       struct sfc_mport_journal_ctx ctx;
+       const efx_pcie_interface_t *controllers;
+       size_t nb_controllers;
+       efx_rc_t efx_rc;
+       int rc;
+
+       memset(&ctx, 0, sizeof(ctx));
+       ctx.sa = sa;
+       ctx.switch_domain_id = sa->mae.switch_domain_id;
+
+       efx_rc = efx_mcdi_get_own_client_handle(sa->nic, &ctx.mcdi_handle);
+       if (efx_rc != 0) {
+               sfc_err(sa, "failed to get own MCDI handle");
+               SFC_ASSERT(efx_rc > 0);
+               return efx_rc;
+       }
+
+       rc = sfc_mae_switch_domain_controllers(ctx.switch_domain_id,
+                                              &controllers, &nb_controllers);
+       if (rc != 0) {
+               sfc_err(sa, "failed to get controller mapping");
+               return rc;
+       }
+
+       ctx.controllers_assigned = controllers != NULL;
+       ctx.controllers = NULL;
+       ctx.nb_controllers = 0;
+
+       efx_rc = efx_mae_read_mport_journal(sa->nic,
+                                           sfc_process_mport_journal_cb, &ctx);
+       if (efx_rc != 0) {
+               sfc_err(sa, "failed to process MAE mport journal");
+               SFC_ASSERT(efx_rc > 0);
+               return efx_rc;
+       }
+
+       if (controllers == NULL) {
+               rc = sfc_mae_switch_domain_map_controllers(ctx.switch_domain_id,
+                                                          ctx.controllers,
+                                                          ctx.nb_controllers);
+               if (rc != 0)
+                       return rc;
+       }
+
+       return 0;
+}
+
 static const struct eth_dev_ops sfc_eth_dev_ops = {
        .dev_configure                  = sfc_dev_configure,
        .dev_start                      = sfc_dev_start,
@@ -1716,10 +2116,8 @@ static const struct eth_dev_ops sfc_eth_dev_ops = {
        .tx_queue_stop                  = sfc_tx_queue_stop,
        .rx_queue_setup                 = sfc_rx_queue_setup,
        .rx_queue_release               = sfc_rx_queue_release,
-       .rx_queue_count                 = sfc_rx_queue_count,
-       .rx_descriptor_done             = sfc_rx_descriptor_done,
-       .rx_descriptor_status           = sfc_rx_descriptor_status,
-       .tx_descriptor_status           = sfc_tx_descriptor_status,
+       .rx_queue_intr_enable           = sfc_rx_queue_intr_enable,
+       .rx_queue_intr_disable          = sfc_rx_queue_intr_disable,
        .tx_queue_setup                 = sfc_tx_queue_setup,
        .tx_queue_release               = sfc_tx_queue_release,
        .flow_ctrl_get                  = sfc_flow_ctrl_get,
@@ -1731,7 +2129,7 @@ static const struct eth_dev_ops sfc_eth_dev_ops = {
        .reta_query                     = sfc_dev_rss_reta_query,
        .rss_hash_update                = sfc_dev_rss_hash_update,
        .rss_hash_conf_get              = sfc_dev_rss_hash_conf_get,
-       .filter_ctrl                    = sfc_dev_filter_ctrl,
+       .flow_ops_get                   = sfc_dev_flow_ops_get,
        .set_mc_addr_list               = sfc_set_mc_addr_list,
        .rxq_info_get                   = sfc_rx_queue_info_get,
        .txq_info_get                   = sfc_tx_queue_info_get,
@@ -1741,6 +2139,10 @@ static const struct eth_dev_ops sfc_eth_dev_ops = {
        .pool_ops_supported             = sfc_pool_ops_supported,
 };
 
+struct sfc_ethdev_init_data {
+       uint16_t                nb_representors;
+};
+
 /**
  * Duplicate a string in potentially shared memory required for
  * multi-process support.
@@ -1767,7 +2169,7 @@ sfc_strdup(const char *str)
 static int
 sfc_eth_dev_set_ops(struct rte_eth_dev *dev)
 {
-       struct sfc_adapter *sa = dev->data->dev_private;
+       struct sfc_adapter *sa = sfc_adapter_by_eth_dev(dev);
        struct sfc_adapter_shared *sas = sfc_adapter_shared_by_eth_dev(dev);
        const struct sfc_dp_rx *dp_rx;
        const struct sfc_dp_tx *dp_tx;
@@ -1782,6 +2184,11 @@ sfc_eth_dev_set_ops(struct rte_eth_dev *dev)
        case EFX_FAMILY_MEDFORD:
        case EFX_FAMILY_MEDFORD2:
                avail_caps |= SFC_DP_HW_FW_CAP_EF10;
+               avail_caps |= SFC_DP_HW_FW_CAP_RX_EFX;
+               avail_caps |= SFC_DP_HW_FW_CAP_TX_EFX;
+               break;
+       case EFX_FAMILY_RIVERHEAD:
+               avail_caps |= SFC_DP_HW_FW_CAP_EF100;
                break;
        default:
                break;
@@ -1869,8 +2276,12 @@ sfc_eth_dev_set_ops(struct rte_eth_dev *dev)
        sa->priv.dp_tx = dp_tx;
 
        dev->rx_pkt_burst = dp_rx->pkt_burst;
+       dev->tx_pkt_prepare = dp_tx->pkt_prepare;
        dev->tx_pkt_burst = dp_tx->pkt_burst;
 
+       dev->rx_queue_count = sfc_rx_queue_count;
+       dev->rx_descriptor_status = sfc_rx_descriptor_status;
+       dev->tx_descriptor_status = sfc_tx_descriptor_status;
        dev->dev_ops = &sfc_eth_dev_ops;
 
        return 0;
@@ -1892,10 +2303,11 @@ fail_kvarg_rx_datapath:
 static void
 sfc_eth_dev_clear_ops(struct rte_eth_dev *dev)
 {
-       struct sfc_adapter *sa = dev->data->dev_private;
+       struct sfc_adapter *sa = sfc_adapter_by_eth_dev(dev);
        struct sfc_adapter_shared *sas = sfc_adapter_shared_by_eth_dev(dev);
 
        dev->dev_ops = NULL;
+       dev->tx_pkt_prepare = NULL;
        dev->rx_pkt_burst = NULL;
        dev->tx_pkt_burst = NULL;
 
@@ -1909,10 +2321,7 @@ sfc_eth_dev_clear_ops(struct rte_eth_dev *dev)
 }
 
 static const struct eth_dev_ops sfc_eth_dev_secondary_ops = {
-       .rx_queue_count                 = sfc_rx_queue_count,
-       .rx_descriptor_done             = sfc_rx_descriptor_done,
-       .rx_descriptor_status           = sfc_rx_descriptor_status,
-       .tx_descriptor_status           = sfc_tx_descriptor_status,
+       .dev_supported_ptypes_get       = sfc_dev_supported_ptypes_get,
        .reta_query                     = sfc_dev_rss_reta_query,
        .rss_hash_conf_get              = sfc_dev_rss_hash_conf_get,
        .rxq_info_get                   = sfc_rx_queue_info_get,
@@ -1922,12 +2331,6 @@ static const struct eth_dev_ops sfc_eth_dev_secondary_ops = {
 static int
 sfc_eth_dev_secondary_init(struct rte_eth_dev *dev, uint32_t logtype_main)
 {
-       /*
-        * Device private data has really many process-local pointers.
-        * Below code should be extremely careful to use data located
-        * in shared memory only.
-        */
-       struct sfc_adapter *sa = dev->data->dev_private;
        struct sfc_adapter_shared *sas = sfc_adapter_shared_by_eth_dev(dev);
        struct sfc_adapter_priv *sap;
        const struct sfc_dp_rx *dp_rx;
@@ -1948,13 +2351,13 @@ sfc_eth_dev_secondary_init(struct rte_eth_dev *dev, uint32_t logtype_main)
 
        dp_rx = sfc_dp_find_rx_by_name(&sfc_dp_head, sas->dp_rx_name);
        if (dp_rx == NULL) {
-               SFC_LOG(sa, RTE_LOG_ERR, logtype_main,
+               SFC_LOG(sas, RTE_LOG_ERR, logtype_main,
                        "cannot find %s Rx datapath", sas->dp_rx_name);
                rc = ENOENT;
                goto fail_dp_rx;
        }
        if (~dp_rx->features & SFC_DP_RX_FEAT_MULTI_PROCESS) {
-               SFC_LOG(sa, RTE_LOG_ERR, logtype_main,
+               SFC_LOG(sas, RTE_LOG_ERR, logtype_main,
                        "%s Rx datapath does not support multi-process",
                        sas->dp_rx_name);
                rc = EINVAL;
@@ -1963,13 +2366,13 @@ sfc_eth_dev_secondary_init(struct rte_eth_dev *dev, uint32_t logtype_main)
 
        dp_tx = sfc_dp_find_tx_by_name(&sfc_dp_head, sas->dp_tx_name);
        if (dp_tx == NULL) {
-               SFC_LOG(sa, RTE_LOG_ERR, logtype_main,
+               SFC_LOG(sas, RTE_LOG_ERR, logtype_main,
                        "cannot find %s Tx datapath", sas->dp_tx_name);
                rc = ENOENT;
                goto fail_dp_tx;
        }
        if (~dp_tx->features & SFC_DP_TX_FEAT_MULTI_PROCESS) {
-               SFC_LOG(sa, RTE_LOG_ERR, logtype_main,
+               SFC_LOG(sas, RTE_LOG_ERR, logtype_main,
                        "%s Tx datapath does not support multi-process",
                        sas->dp_tx_name);
                rc = EINVAL;
@@ -1981,7 +2384,11 @@ sfc_eth_dev_secondary_init(struct rte_eth_dev *dev, uint32_t logtype_main)
 
        dev->process_private = sap;
        dev->rx_pkt_burst = dp_rx->pkt_burst;
+       dev->tx_pkt_prepare = dp_tx->pkt_prepare;
        dev->tx_pkt_burst = dp_tx->pkt_burst;
+       dev->rx_queue_count = sfc_rx_queue_count;
+       dev->rx_descriptor_status = sfc_rx_descriptor_status;
+       dev->tx_descriptor_status = sfc_tx_descriptor_status;
        dev->dev_ops = &sfc_eth_dev_secondary_ops;
 
        return 0;
@@ -1996,26 +2403,18 @@ fail_alloc_priv:
        return rc;
 }
 
-static void
-sfc_eth_dev_secondary_clear_ops(struct rte_eth_dev *dev)
-{
-       free(dev->process_private);
-       dev->process_private = NULL;
-       dev->dev_ops = NULL;
-       dev->tx_pkt_burst = NULL;
-       dev->rx_pkt_burst = NULL;
-}
-
 static void
 sfc_register_dp(void)
 {
        /* Register once */
        if (TAILQ_EMPTY(&sfc_dp_head)) {
                /* Prefer EF10 datapath */
+               sfc_dp_register(&sfc_dp_head, &sfc_ef100_rx.dp);
                sfc_dp_register(&sfc_dp_head, &sfc_ef10_essb_rx.dp);
                sfc_dp_register(&sfc_dp_head, &sfc_ef10_rx.dp);
                sfc_dp_register(&sfc_dp_head, &sfc_efx_rx.dp);
 
+               sfc_dp_register(&sfc_dp_head, &sfc_ef100_tx.dp);
                sfc_dp_register(&sfc_dp_head, &sfc_ef10_tx.dp);
                sfc_dp_register(&sfc_dp_head, &sfc_efx_tx.dp);
                sfc_dp_register(&sfc_dp_head, &sfc_ef10_simple_tx.dp);
@@ -2023,15 +2422,69 @@ sfc_register_dp(void)
 }
 
 static int
-sfc_eth_dev_init(struct rte_eth_dev *dev)
+sfc_parse_switch_mode(struct sfc_adapter *sa, bool has_representors)
 {
-       struct sfc_adapter *sa = dev->data->dev_private;
-       struct sfc_adapter_shared *sas;
+       const efx_nic_cfg_t *encp = efx_nic_cfg_get(sa->nic);
+       const char *switch_mode = NULL;
+       int rc;
+
+       sfc_log_init(sa, "entry");
+
+       rc = sfc_kvargs_process(sa, SFC_KVARG_SWITCH_MODE,
+                               sfc_kvarg_string_handler, &switch_mode);
+       if (rc != 0)
+               goto fail_kvargs;
+
+       if (switch_mode == NULL) {
+               sa->switchdev = encp->enc_mae_supported &&
+                               (!encp->enc_datapath_cap_evb ||
+                                has_representors);
+       } else if (strcasecmp(switch_mode, SFC_KVARG_SWITCH_MODE_LEGACY) == 0) {
+               sa->switchdev = false;
+       } else if (strcasecmp(switch_mode,
+                             SFC_KVARG_SWITCH_MODE_SWITCHDEV) == 0) {
+               sa->switchdev = true;
+       } else {
+               sfc_err(sa, "invalid switch mode device argument '%s'",
+                       switch_mode);
+               rc = EINVAL;
+               goto fail_mode;
+       }
+
+       sfc_log_init(sa, "done");
+
+       return 0;
+
+fail_mode:
+fail_kvargs:
+       sfc_log_init(sa, "failed: %s", rte_strerror(rc));
+
+       return rc;
+}
+
+static int
+sfc_eth_dev_init(struct rte_eth_dev *dev, void *init_params)
+{
+       struct sfc_adapter_shared *sas = sfc_adapter_shared_by_eth_dev(dev);
        struct rte_pci_device *pci_dev = RTE_ETH_DEV_TO_PCI(dev);
+       struct sfc_ethdev_init_data *init_data = init_params;
        uint32_t logtype_main;
+       struct sfc_adapter *sa;
        int rc;
        const efx_nic_cfg_t *encp;
-       const struct ether_addr *from;
+       const struct rte_ether_addr *from;
+       int ret;
+
+       if (sfc_efx_dev_class_get(pci_dev->device.devargs) !=
+                       SFC_EFX_DEV_CLASS_NET) {
+               SFC_GENERIC_LOG(DEBUG,
+                       "Incompatible device class: skip probing, should be probed by other sfc driver.");
+               return 1;
+       }
+
+       rc = sfc_dp_mport_register();
+       if (rc != 0)
+               return rc;
 
        sfc_register_dp();
 
@@ -2039,29 +2492,47 @@ sfc_eth_dev_init(struct rte_eth_dev *dev)
                                            SFC_LOGTYPE_MAIN_STR,
                                            RTE_LOG_NOTICE);
 
-       sa->priv.shared = &sa->_shared;
-       sas = sa->priv.shared;
-
        if (rte_eal_process_type() != RTE_PROC_PRIMARY)
                return -sfc_eth_dev_secondary_init(dev, logtype_main);
 
+       /* Required for logging */
+       ret = snprintf(sas->log_prefix, sizeof(sas->log_prefix),
+                       "PMD: sfc_efx " PCI_PRI_FMT " #%" PRIu16 ": ",
+                       pci_dev->addr.domain, pci_dev->addr.bus,
+                       pci_dev->addr.devid, pci_dev->addr.function,
+                       dev->data->port_id);
+       if (ret < 0 || ret >= (int)sizeof(sas->log_prefix)) {
+               SFC_GENERIC_LOG(ERR,
+                       "reserved log prefix is too short for " PCI_PRI_FMT,
+                       pci_dev->addr.domain, pci_dev->addr.bus,
+                       pci_dev->addr.devid, pci_dev->addr.function);
+               return -EINVAL;
+       }
+       sas->pci_addr = pci_dev->addr;
+       sas->port_id = dev->data->port_id;
+
        /*
-        * sfc_adapter is a mixture of shared and process private data.
-        * During transition period use it in both kinds. When the
-        * driver becomes ready to separate it, sfc_adapter will become
-        * primary process private only.
+        * Allocate process private data from heap, since it should not
+        * be located in shared memory allocated using rte_malloc() API.
         */
+       sa = calloc(1, sizeof(*sa));
+       if (sa == NULL) {
+               rc = ENOMEM;
+               goto fail_alloc_sa;
+       }
+
        dev->process_private = sa;
 
        /* Required for logging */
-       sas->pci_addr = pci_dev->addr;
-       sas->port_id = dev->data->port_id;
+       sa->priv.shared = sas;
        sa->priv.logtype_main = logtype_main;
 
        sa->eth_dev = dev;
 
        /* Copy PCI device info to the dev->data */
        rte_eth_copy_pci_info(dev, pci_dev);
+       dev->data->dev_flags |= RTE_ETH_DEV_AUTOFILL_QUEUE_XSTATS;
+       dev->data->dev_flags |= RTE_ETH_DEV_FLOW_OPS_THREAD_SAFE;
 
        rc = sfc_kvargs_parse(sa);
        if (rc != 0)
@@ -2069,7 +2540,7 @@ sfc_eth_dev_init(struct rte_eth_dev *dev)
 
        sfc_log_init(sa, "entry");
 
-       dev->data->mac_addrs = rte_zmalloc("sfc", ETHER_ADDR_LEN, 0);
+       dev->data->mac_addrs = rte_zmalloc("sfc", RTE_ETHER_ADDR_LEN, 0);
        if (dev->data->mac_addrs == NULL) {
                rc = ENOMEM;
                goto fail_mac_addrs;
@@ -2083,6 +2554,14 @@ sfc_eth_dev_init(struct rte_eth_dev *dev)
        if (rc != 0)
                goto fail_probe;
 
+       /*
+        * Selecting a default switch mode requires the NIC to be probed and
+        * to have its capabilities filled in.
+        */
+       rc = sfc_parse_switch_mode(sa, init_data->nb_representors > 0);
+       if (rc != 0)
+               goto fail_switch_mode;
+
        sfc_log_init(sa, "set device ops");
        rc = sfc_eth_dev_set_ops(dev);
        if (rc != 0)
@@ -2093,24 +2572,35 @@ sfc_eth_dev_init(struct rte_eth_dev *dev)
        if (rc != 0)
                goto fail_attach;
 
+       if (sa->switchdev && sa->mae.status != SFC_MAE_STATUS_SUPPORTED) {
+               sfc_err(sa,
+                       "failed to enable switchdev mode without MAE support");
+               rc = ENOTSUP;
+               goto fail_switchdev_no_mae;
+       }
+
        encp = efx_nic_cfg_get(sa->nic);
 
        /*
         * The arguments are really reverse order in comparison to
         * Linux kernel. Copy from NIC config to Ethernet device data.
         */
-       from = (const struct ether_addr *)(encp->enc_mac_addr);
-       ether_addr_copy(from, &dev->data->mac_addrs[0]);
+       from = (const struct rte_ether_addr *)(encp->enc_mac_addr);
+       rte_ether_addr_copy(from, &dev->data->mac_addrs[0]);
 
        sfc_adapter_unlock(sa);
 
        sfc_log_init(sa, "done");
        return 0;
 
+fail_switchdev_no_mae:
+       sfc_detach(sa);
+
 fail_attach:
        sfc_eth_dev_clear_ops(dev);
 
 fail_set_ops:
+fail_switch_mode:
        sfc_unprobe(sa);
 
 fail_probe:
@@ -2125,6 +2615,9 @@ fail_mac_addrs:
 fail_kvargs_parse:
        sfc_log_init(sa, "failed %d", rc);
        dev->process_private = NULL;
+       free(sa);
+
+fail_alloc_sa:
        SFC_ASSERT(rc > 0);
        return -rc;
 }
@@ -2132,32 +2625,8 @@ fail_kvargs_parse:
 static int
 sfc_eth_dev_uninit(struct rte_eth_dev *dev)
 {
-       struct sfc_adapter *sa;
-
-       if (rte_eal_process_type() != RTE_PROC_PRIMARY) {
-               sfc_eth_dev_secondary_clear_ops(dev);
-               return 0;
-       }
-
-       sa = dev->data->dev_private;
-       sfc_log_init(sa, "entry");
-
-       sfc_adapter_lock(sa);
-
-       sfc_eth_dev_clear_ops(dev);
-
-       sfc_detach(sa);
-       sfc_unprobe(sa);
-
-       sfc_kvargs_cleanup(sa);
-
-       sfc_adapter_unlock(sa);
-       sfc_adapter_lock_fini(sa);
+       sfc_dev_close(dev);
 
-       sfc_log_init(sa, "done");
-
-       /* Required for logging, so cleanup last */
-       sa->eth_dev = NULL;
        return 0;
 }
 
@@ -2170,14 +2639,301 @@ static const struct rte_pci_id pci_id_sfc_efx_map[] = {
        { RTE_PCI_DEVICE(EFX_PCI_VENID_SFC, EFX_PCI_DEVID_MEDFORD_VF) },
        { RTE_PCI_DEVICE(EFX_PCI_VENID_SFC, EFX_PCI_DEVID_MEDFORD2) },
        { RTE_PCI_DEVICE(EFX_PCI_VENID_SFC, EFX_PCI_DEVID_MEDFORD2_VF) },
+       { RTE_PCI_DEVICE(EFX_PCI_VENID_XILINX, EFX_PCI_DEVID_RIVERHEAD) },
        { .vendor_id = 0 /* sentinel */ }
 };
 
+static int
+sfc_parse_rte_devargs(const char *args, struct rte_eth_devargs *devargs)
+{
+       struct rte_eth_devargs eth_da = { .nb_representor_ports = 0 };
+       int rc;
+
+       if (args != NULL) {
+               rc = rte_eth_devargs_parse(args, &eth_da);
+               if (rc != 0) {
+                       SFC_GENERIC_LOG(ERR,
+                                       "Failed to parse generic devargs '%s'",
+                                       args);
+                       return rc;
+               }
+       }
+
+       *devargs = eth_da;
+
+       return 0;
+}
+
+static int
+sfc_eth_dev_find_or_create(struct rte_pci_device *pci_dev,
+                          struct sfc_ethdev_init_data *init_data,
+                          struct rte_eth_dev **devp,
+                          bool *dev_created)
+{
+       struct rte_eth_dev *dev;
+       bool created = false;
+       int rc;
+
+       dev = rte_eth_dev_allocated(pci_dev->device.name);
+       if (dev == NULL) {
+               rc = rte_eth_dev_create(&pci_dev->device, pci_dev->device.name,
+                                       sizeof(struct sfc_adapter_shared),
+                                       eth_dev_pci_specific_init, pci_dev,
+                                       sfc_eth_dev_init, init_data);
+               if (rc != 0) {
+                       SFC_GENERIC_LOG(ERR, "Failed to create sfc ethdev '%s'",
+                                       pci_dev->device.name);
+                       return rc;
+               }
+
+               created = true;
+
+               dev = rte_eth_dev_allocated(pci_dev->device.name);
+               if (dev == NULL) {
+                       SFC_GENERIC_LOG(ERR,
+                               "Failed to find allocated sfc ethdev '%s'",
+                               pci_dev->device.name);
+                       return -ENODEV;
+               }
+       }
+
+       *devp = dev;
+       *dev_created = created;
+
+       return 0;
+}
+
+static int
+sfc_eth_dev_create_repr(struct sfc_adapter *sa,
+                       efx_pcie_interface_t controller,
+                       uint16_t port,
+                       uint16_t repr_port,
+                       enum rte_eth_representor_type type)
+{
+       struct sfc_repr_entity_info entity;
+       efx_mport_sel_t mport_sel;
+       int rc;
+
+       switch (type) {
+       case RTE_ETH_REPRESENTOR_NONE:
+               return 0;
+       case RTE_ETH_REPRESENTOR_VF:
+       case RTE_ETH_REPRESENTOR_PF:
+               break;
+       case RTE_ETH_REPRESENTOR_SF:
+               sfc_err(sa, "SF representors are not supported");
+               return ENOTSUP;
+       default:
+               sfc_err(sa, "unknown representor type: %d", type);
+               return ENOTSUP;
+       }
+
+       rc = efx_mae_mport_by_pcie_mh_function(controller,
+                                              port,
+                                              repr_port,
+                                              &mport_sel);
+       if (rc != 0) {
+               sfc_err(sa,
+                       "failed to get m-port selector for controller %u port %u repr_port %u: %s",
+                       controller, port, repr_port, rte_strerror(-rc));
+               return rc;
+       }
+
+       memset(&entity, 0, sizeof(entity));
+       entity.type = type;
+       entity.intf = controller;
+       entity.pf = port;
+       entity.vf = repr_port;
+
+       rc = sfc_repr_create(sa->eth_dev, &entity, sa->mae.switch_domain_id,
+                            &mport_sel);
+       if (rc != 0) {
+               sfc_err(sa,
+                       "failed to create representor for controller %u port %u repr_port %u: %s",
+                       controller, port, repr_port, rte_strerror(-rc));
+               return rc;
+       }
+
+       return 0;
+}
+
+static int
+sfc_eth_dev_create_repr_port(struct sfc_adapter *sa,
+                            const struct rte_eth_devargs *eth_da,
+                            efx_pcie_interface_t controller,
+                            uint16_t port)
+{
+       int first_error = 0;
+       uint16_t i;
+       int rc;
+
+       if (eth_da->type == RTE_ETH_REPRESENTOR_PF) {
+               return sfc_eth_dev_create_repr(sa, controller, port,
+                                              EFX_PCI_VF_INVALID,
+                                              eth_da->type);
+       }
+
+       for (i = 0; i < eth_da->nb_representor_ports; i++) {
+               rc = sfc_eth_dev_create_repr(sa, controller, port,
+                                            eth_da->representor_ports[i],
+                                            eth_da->type);
+               if (rc != 0 && first_error == 0)
+                       first_error = rc;
+       }
+
+       return first_error;
+}
+
+static int
+sfc_eth_dev_create_repr_controller(struct sfc_adapter *sa,
+                                  const struct rte_eth_devargs *eth_da,
+                                  efx_pcie_interface_t controller)
+{
+       const efx_nic_cfg_t *encp;
+       int first_error = 0;
+       uint16_t default_port;
+       uint16_t i;
+       int rc;
+
+       if (eth_da->nb_ports == 0) {
+               encp = efx_nic_cfg_get(sa->nic);
+               default_port = encp->enc_intf == controller ? encp->enc_pf : 0;
+               return sfc_eth_dev_create_repr_port(sa, eth_da, controller,
+                                                   default_port);
+       }
+
+       for (i = 0; i < eth_da->nb_ports; i++) {
+               rc = sfc_eth_dev_create_repr_port(sa, eth_da, controller,
+                                                 eth_da->ports[i]);
+               if (rc != 0 && first_error == 0)
+                       first_error = rc;
+       }
+
+       return first_error;
+}
+
+static int
+sfc_eth_dev_create_representors(struct rte_eth_dev *dev,
+                               const struct rte_eth_devargs *eth_da)
+{
+       efx_pcie_interface_t intf;
+       const efx_nic_cfg_t *encp;
+       struct sfc_adapter *sa;
+       uint16_t switch_domain_id;
+       uint16_t i;
+       int rc;
+
+       sa = sfc_adapter_by_eth_dev(dev);
+       switch_domain_id = sa->mae.switch_domain_id;
+
+       switch (eth_da->type) {
+       case RTE_ETH_REPRESENTOR_NONE:
+               return 0;
+       case RTE_ETH_REPRESENTOR_PF:
+       case RTE_ETH_REPRESENTOR_VF:
+               break;
+       case RTE_ETH_REPRESENTOR_SF:
+               sfc_err(sa, "SF representors are not supported");
+               return -ENOTSUP;
+       default:
+               sfc_err(sa, "unknown representor type: %d",
+                       eth_da->type);
+               return -ENOTSUP;
+       }
+
+       if (!sa->switchdev) {
+               sfc_err(sa, "cannot create representors in non-switchdev mode");
+               return -EINVAL;
+       }
+
+       if (!sfc_repr_available(sfc_sa2shared(sa))) {
+               sfc_err(sa, "cannot create representors: unsupported");
+
+               return -ENOTSUP;
+       }
+
+       /*
+        * This is needed to construct the DPDK controller -> EFX interface
+        * mapping.
+        */
+       sfc_adapter_lock(sa);
+       rc = sfc_process_mport_journal(sa);
+       sfc_adapter_unlock(sa);
+       if (rc != 0) {
+               SFC_ASSERT(rc > 0);
+               return -rc;
+       }
+
+       if (eth_da->nb_mh_controllers > 0) {
+               for (i = 0; i < eth_da->nb_mh_controllers; i++) {
+                       rc = sfc_mae_switch_domain_get_intf(switch_domain_id,
+                                               eth_da->mh_controllers[i],
+                                               &intf);
+                       if (rc != 0) {
+                               sfc_err(sa, "failed to get representor");
+                               continue;
+                       }
+                       sfc_eth_dev_create_repr_controller(sa, eth_da, intf);
+               }
+       } else {
+               encp = efx_nic_cfg_get(sa->nic);
+               sfc_eth_dev_create_repr_controller(sa, eth_da, encp->enc_intf);
+       }
+
+       return 0;
+}
+
 static int sfc_eth_dev_pci_probe(struct rte_pci_driver *pci_drv __rte_unused,
        struct rte_pci_device *pci_dev)
 {
-       return rte_eth_dev_pci_generic_probe(pci_dev,
-               sizeof(struct sfc_adapter), sfc_eth_dev_init);
+       struct sfc_ethdev_init_data init_data;
+       struct rte_eth_devargs eth_da;
+       struct rte_eth_dev *dev;
+       bool dev_created;
+       int rc;
+
+       if (pci_dev->device.devargs != NULL) {
+               rc = sfc_parse_rte_devargs(pci_dev->device.devargs->args,
+                                          &eth_da);
+               if (rc != 0)
+                       return rc;
+       } else {
+               memset(&eth_da, 0, sizeof(eth_da));
+       }
+
+       /* If no VF representors specified, check for PF ones */
+       if (eth_da.nb_representor_ports > 0)
+               init_data.nb_representors = eth_da.nb_representor_ports;
+       else
+               init_data.nb_representors = eth_da.nb_ports;
+
+       if (init_data.nb_representors > 0 &&
+           rte_eal_process_type() != RTE_PROC_PRIMARY) {
+               SFC_GENERIC_LOG(ERR,
+                       "Create representors from secondary process not supported, dev '%s'",
+                       pci_dev->device.name);
+               return -ENOTSUP;
+       }
+
+       /*
+        * Driver supports RTE_PCI_DRV_PROBE_AGAIN. Hence create device only
+        * if it does not already exist. Re-probing an existing device is
+        * expected to allow additional representors to be configured.
+        */
+       rc = sfc_eth_dev_find_or_create(pci_dev, &init_data, &dev,
+                                       &dev_created);
+       if (rc != 0)
+               return rc;
+
+       rc = sfc_eth_dev_create_representors(dev, &eth_da);
+       if (rc != 0) {
+               if (dev_created)
+                       (void)rte_eth_dev_destroy(dev, sfc_eth_dev_uninit);
+
+               return rc;
+       }
+
+       return 0;
 }
 
 static int sfc_eth_dev_pci_remove(struct rte_pci_device *pci_dev)
@@ -2189,7 +2945,8 @@ static struct rte_pci_driver sfc_efx_pmd = {
        .id_table = pci_id_sfc_efx_map,
        .drv_flags =
                RTE_PCI_DRV_INTR_LSC |
-               RTE_PCI_DRV_NEED_MAPPING,
+               RTE_PCI_DRV_NEED_MAPPING |
+               RTE_PCI_DRV_PROBE_AGAIN,
        .probe = sfc_eth_dev_pci_probe,
        .remove = sfc_eth_dev_pci_remove,
 };
@@ -2198,6 +2955,7 @@ RTE_PMD_REGISTER_PCI(net_sfc_efx, sfc_efx_pmd);
 RTE_PMD_REGISTER_PCI_TABLE(net_sfc_efx, pci_id_sfc_efx_map);
 RTE_PMD_REGISTER_KMOD_DEP(net_sfc_efx, "* igb_uio | uio_pci_generic | vfio-pci");
 RTE_PMD_REGISTER_PARAM_STRING(net_sfc_efx,
+       SFC_KVARG_SWITCH_MODE "=" SFC_KVARG_VALUES_SWITCH_MODE " "
        SFC_KVARG_RX_DATAPATH "=" SFC_KVARG_VALUES_RX_DATAPATH " "
        SFC_KVARG_TX_DATAPATH "=" SFC_KVARG_VALUES_TX_DATAPATH " "
        SFC_KVARG_PERF_PROFILE "=" SFC_KVARG_VALUES_PERF_PROFILE " "