drivers: advertise kmod dependencies in pmdinfo
[dpdk.git] / drivers / net / szedata2 / rte_eth_szedata2.c
index 5f3a19c..677ba9f 100644 (file)
@@ -1,7 +1,7 @@
 /*-
  *   BSD LICENSE
  *
- *   Copyright (c) 2015 CESNET
+ *   Copyright (c) 2015 - 2016 CESNET
  *   All rights reserved.
  *
  *   Redistribution and use in source and binary forms, with or without
 #include <unistd.h>
 #include <stdbool.h>
 #include <err.h>
+#include <sys/types.h>
+#include <dirent.h>
+#include <sys/stat.h>
+#include <fcntl.h>
+#include <sys/mman.h>
 
 #include <libsze2.h>
 
 #include <rte_memcpy.h>
 #include <rte_kvargs.h>
 #include <rte_dev.h>
+#include <rte_atomic.h>
 
 #include "rte_eth_szedata2.h"
 
-#define RTE_ETH_SZEDATA2_DEV_PATH_ARG "dev_path"
-#define RTE_ETH_SZEDATA2_RX_IFACES_ARG "rx_ifaces"
-#define RTE_ETH_SZEDATA2_TX_IFACES_ARG "tx_ifaces"
-
 #define RTE_ETH_SZEDATA2_MAX_RX_QUEUES 32
 #define RTE_ETH_SZEDATA2_MAX_TX_QUEUES 32
 #define RTE_ETH_SZEDATA2_TX_LOCK_SIZE (32 * 1024 * 1024)
  */
 #define RTE_SZE2_PACKET_HEADER_SIZE_ALIGNED 8
 
+#define RTE_SZEDATA2_DRIVER_NAME net_szedata2
+#define RTE_SZEDATA2_PCI_DRIVER_NAME "rte_szedata2_pmd"
+
+#define SZEDATA2_DEV_PATH_FMT "/dev/szedataII%u"
+
 struct szedata2_rx_queue {
        struct szedata *sze;
        uint8_t rx_channel;
@@ -74,347 +81,947 @@ struct szedata2_tx_queue {
        struct szedata *sze;
        uint8_t tx_channel;
        volatile uint64_t tx_pkts;
-       volatile uint64_t err_pkts;
        volatile uint64_t tx_bytes;
-};
-
-struct rxtx_szedata2 {
-       uint32_t num_of_rx;
-       uint32_t num_of_tx;
-       uint32_t sze_rx_mask_req;
-       uint32_t sze_tx_mask_req;
-       char *sze_dev;
+       volatile uint64_t err_pkts;
 };
 
 struct pmd_internals {
        struct szedata2_rx_queue rx_queue[RTE_ETH_SZEDATA2_MAX_RX_QUEUES];
        struct szedata2_tx_queue tx_queue[RTE_ETH_SZEDATA2_MAX_TX_QUEUES];
-       unsigned nb_rx_queues;
-       unsigned nb_tx_queues;
-       uint32_t num_of_rx;
-       uint32_t num_of_tx;
-       uint32_t sze_rx_req;
-       uint32_t sze_tx_req;
-       int if_index;
-       char *sze_dev;
-};
-
-static const char *valid_arguments[] = {
-       RTE_ETH_SZEDATA2_DEV_PATH_ARG,
-       RTE_ETH_SZEDATA2_RX_IFACES_ARG,
-       RTE_ETH_SZEDATA2_TX_IFACES_ARG,
-       NULL
+       uint16_t max_rx_queues;
+       uint16_t max_tx_queues;
+       char sze_dev[PATH_MAX];
 };
 
 static struct ether_addr eth_addr = {
        .addr_bytes = { 0x00, 0x11, 0x17, 0x00, 0x00, 0x00 }
 };
-static const char *drivername = "SZEdata2 PMD";
-static struct rte_eth_link pmd_link = {
-               .link_speed = ETH_LINK_SPEED_10G,
-               .link_duplex = ETH_LINK_FULL_DUPLEX,
-               .link_status = 0
-};
-
 
-static uint32_t
-count_ones(uint32_t num)
+static uint16_t
+eth_szedata2_rx(void *queue,
+               struct rte_mbuf **bufs,
+               uint16_t nb_pkts)
 {
-       num = num - ((num >> 1) & 0x55555555); /* reuse input as temporary */
-       num = (num & 0x33333333) + ((num >> 2) & 0x33333333);        /* temp */
-       return (((num + (num >> 4)) & 0xF0F0F0F) * 0x1010101) >> 24; /* count */
-}
+       unsigned int i;
+       struct rte_mbuf *mbuf;
+       struct szedata2_rx_queue *sze_q = queue;
+       struct rte_pktmbuf_pool_private *mbp_priv;
+       uint16_t num_rx = 0;
+       uint16_t buf_size;
+       uint16_t sg_size;
+       uint16_t hw_size;
+       uint16_t packet_size;
+       uint64_t num_bytes = 0;
+       struct szedata *sze = sze_q->sze;
+       uint8_t *header_ptr = NULL; /* header of packet */
+       uint8_t *packet_ptr1 = NULL;
+       uint8_t *packet_ptr2 = NULL;
+       uint16_t packet_len1 = 0;
+       uint16_t packet_len2 = 0;
+       uint16_t hw_data_align;
+
+       if (unlikely(sze_q->sze == NULL || nb_pkts == 0))
+               return 0;
 
-static int
-init_rx_channels(struct rte_eth_dev *dev, int v)
-{
-       struct pmd_internals *internals = dev->data->dev_private;
-       int ret;
-       uint32_t i;
-       uint32_t count = internals->num_of_rx;
-       uint32_t num_sub = 0;
-       uint32_t x;
-       uint32_t rx;
-       uint32_t tx;
+       /*
+        * Reads the given number of packets from szedata2 channel given
+        * by queue and copies the packet data into a newly allocated mbuf
+        * to return.
+        */
+       for (i = 0; i < nb_pkts; i++) {
+               mbuf = rte_pktmbuf_alloc(sze_q->mb_pool);
 
-       rx = internals->sze_rx_req;
-       tx = 0;
+               if (unlikely(mbuf == NULL))
+                       break;
 
-       for (i = 0; i < count; i++) {
-               /*
-                * Open, subscribe rx,tx channels and start device
-                */
-               if (v)
-                       RTE_LOG(INFO, PMD, "Opening SZE device %u. time\n", i);
-
-               internals->rx_queue[num_sub].sze =
-                       szedata_open(internals->sze_dev);
-               if (internals->rx_queue[num_sub].sze == NULL)
-                       return -1;
-
-               /* separate least significant non-zero bit */
-               x = rx & ((~rx) + 1);
-
-               if (v)
-                       RTE_LOG(INFO, PMD, "Subscribing rx channel: 0x%x "
-                               "tx channel: 0x%x\n", x, tx);
-
-               ret = szedata_subscribe3(internals->rx_queue[num_sub].sze,
-                               &x, &tx);
-               if (ret) {
-                       szedata_close(internals->rx_queue[num_sub].sze);
-                       internals->rx_queue[num_sub].sze = NULL;
-                       return -1;
+               /* get the next sze packet */
+               if (sze->ct_rx_lck != NULL && !sze->ct_rx_rem_bytes &&
+                               sze->ct_rx_lck->next == NULL) {
+                       /* unlock old data */
+                       szedata_rx_unlock_data(sze_q->sze, sze->ct_rx_lck_orig);
+                       sze->ct_rx_lck_orig = NULL;
+                       sze->ct_rx_lck = NULL;
                }
 
-               if (v)
-                       RTE_LOG(INFO, PMD, "Subscribed rx channel: 0x%x "
-                               "tx channel: 0x%x\n", x, tx);
+               if (!sze->ct_rx_rem_bytes && sze->ct_rx_lck_orig == NULL) {
+                       /* nothing to read, lock new data */
+                       sze->ct_rx_lck = szedata_rx_lock_data(sze_q->sze, ~0U);
+                       sze->ct_rx_lck_orig = sze->ct_rx_lck;
 
-               if (x) {
-                       if (v)
-                               RTE_LOG(INFO, PMD, "Starting SZE device for "
-                                       "rx queue: %u\n", num_sub);
+                       if (sze->ct_rx_lck == NULL) {
+                               /* nothing to lock */
+                               rte_pktmbuf_free(mbuf);
+                               break;
+                       }
 
-                       ret = szedata_start(internals->rx_queue[num_sub].sze);
-                       if (ret) {
-                               szedata_close(internals->rx_queue[num_sub].sze);
-                               internals->rx_queue[num_sub].sze = NULL;
-                               return -1;
+                       sze->ct_rx_cur_ptr = sze->ct_rx_lck->start;
+                       sze->ct_rx_rem_bytes = sze->ct_rx_lck->len;
+
+                       if (!sze->ct_rx_rem_bytes) {
+                               rte_pktmbuf_free(mbuf);
+                               break;
                        }
+               }
 
+               if (sze->ct_rx_rem_bytes < RTE_SZE2_PACKET_HEADER_SIZE) {
                        /*
-                        * set to 1 all bits lower than bit set to 1
-                        * and that bit to 0
+                        * cut in header
+                        * copy parts of header to merge buffer
                         */
-                       x -= 1;
-                       internals->rx_queue[num_sub].rx_channel =
-                               count_ones(x);
-
-                       if (v)
-                               RTE_LOG(INFO, PMD, "Subscribed rx channel "
-                                       "no: %u\n",
-                                       internals->rx_queue[num_sub].rx_channel
-                                       );
-
-                       num_sub++;
-                       internals->nb_rx_queues = num_sub;
+                       if (sze->ct_rx_lck->next == NULL) {
+                               rte_pktmbuf_free(mbuf);
+                               break;
+                       }
+
+                       /* copy first part of header */
+                       rte_memcpy(sze->ct_rx_buffer, sze->ct_rx_cur_ptr,
+                                       sze->ct_rx_rem_bytes);
+
+                       /* copy second part of header */
+                       sze->ct_rx_lck = sze->ct_rx_lck->next;
+                       sze->ct_rx_cur_ptr = sze->ct_rx_lck->start;
+                       rte_memcpy(sze->ct_rx_buffer + sze->ct_rx_rem_bytes,
+                               sze->ct_rx_cur_ptr,
+                               RTE_SZE2_PACKET_HEADER_SIZE -
+                               sze->ct_rx_rem_bytes);
+
+                       sze->ct_rx_cur_ptr += RTE_SZE2_PACKET_HEADER_SIZE -
+                               sze->ct_rx_rem_bytes;
+                       sze->ct_rx_rem_bytes = sze->ct_rx_lck->len -
+                               RTE_SZE2_PACKET_HEADER_SIZE +
+                               sze->ct_rx_rem_bytes;
+
+                       header_ptr = (uint8_t *)sze->ct_rx_buffer;
                } else {
-                       if (v)
-                               RTE_LOG(INFO, PMD,
-                                       "Could not subscribe any rx channel. "
-                                       "Closing SZE device\n");
+                       /* not cut */
+                       header_ptr = (uint8_t *)sze->ct_rx_cur_ptr;
+                       sze->ct_rx_cur_ptr += RTE_SZE2_PACKET_HEADER_SIZE;
+                       sze->ct_rx_rem_bytes -= RTE_SZE2_PACKET_HEADER_SIZE;
+               }
+
+               sg_size = le16toh(*((uint16_t *)header_ptr));
+               hw_size = le16toh(*(((uint16_t *)header_ptr) + 1));
+               packet_size = sg_size -
+                       RTE_SZE2_ALIGN8(RTE_SZE2_PACKET_HEADER_SIZE + hw_size);
+
+
+               /* checks if packet all right */
+               if (!sg_size)
+                       errx(5, "Zero segsize");
 
-                       szedata_close(internals->rx_queue[num_sub].sze);
-                       internals->rx_queue[num_sub].sze = NULL;
+               /* check sg_size and hwsize */
+               if (hw_size > sg_size - RTE_SZE2_PACKET_HEADER_SIZE) {
+                       errx(10, "Hwsize bigger than expected. Segsize: %d, "
+                               "hwsize: %d", sg_size, hw_size);
                }
 
-               /* set least significant non-zero bit to zero */
-               rx = rx & (rx - 1);
-       }
+               hw_data_align =
+                       RTE_SZE2_ALIGN8(RTE_SZE2_PACKET_HEADER_SIZE + hw_size) -
+                       RTE_SZE2_PACKET_HEADER_SIZE;
+
+               if (sze->ct_rx_rem_bytes >=
+                               (uint16_t)(sg_size -
+                               RTE_SZE2_PACKET_HEADER_SIZE)) {
+                       /* no cut */
+                       /* one packet ready - go to another */
+                       packet_ptr1 = sze->ct_rx_cur_ptr + hw_data_align;
+                       packet_len1 = packet_size;
+                       packet_ptr2 = NULL;
+                       packet_len2 = 0;
+
+                       sze->ct_rx_cur_ptr += RTE_SZE2_ALIGN8(sg_size) -
+                               RTE_SZE2_PACKET_HEADER_SIZE;
+                       sze->ct_rx_rem_bytes -= RTE_SZE2_ALIGN8(sg_size) -
+                               RTE_SZE2_PACKET_HEADER_SIZE;
+               } else {
+                       /* cut in data */
+                       if (sze->ct_rx_lck->next == NULL) {
+                               errx(6, "Need \"next\" lock, "
+                                       "but it is missing: %u",
+                                       sze->ct_rx_rem_bytes);
+                       }
+
+                       /* skip hw data */
+                       if (sze->ct_rx_rem_bytes <= hw_data_align) {
+                               uint16_t rem_size = hw_data_align -
+                                       sze->ct_rx_rem_bytes;
+
+                               /* MOVE to next lock */
+                               sze->ct_rx_lck = sze->ct_rx_lck->next;
+                               sze->ct_rx_cur_ptr =
+                                       (void *)(((uint8_t *)
+                                       (sze->ct_rx_lck->start)) + rem_size);
+
+                               packet_ptr1 = sze->ct_rx_cur_ptr;
+                               packet_len1 = packet_size;
+                               packet_ptr2 = NULL;
+                               packet_len2 = 0;
+
+                               sze->ct_rx_cur_ptr +=
+                                       RTE_SZE2_ALIGN8(packet_size);
+                               sze->ct_rx_rem_bytes = sze->ct_rx_lck->len -
+                                       rem_size - RTE_SZE2_ALIGN8(packet_size);
+                       } else {
+                               /* get pointer and length from first part */
+                               packet_ptr1 = sze->ct_rx_cur_ptr +
+                                       hw_data_align;
+                               packet_len1 = sze->ct_rx_rem_bytes -
+                                       hw_data_align;
+
+                               /* MOVE to next lock */
+                               sze->ct_rx_lck = sze->ct_rx_lck->next;
+                               sze->ct_rx_cur_ptr = sze->ct_rx_lck->start;
+
+                               /* get pointer and length from second part */
+                               packet_ptr2 = sze->ct_rx_cur_ptr;
+                               packet_len2 = packet_size - packet_len1;
+
+                               sze->ct_rx_cur_ptr +=
+                                       RTE_SZE2_ALIGN8(packet_size) -
+                                       packet_len1;
+                               sze->ct_rx_rem_bytes = sze->ct_rx_lck->len -
+                                       (RTE_SZE2_ALIGN8(packet_size) -
+                                        packet_len1);
+                       }
+               }
 
-       dev->data->nb_rx_queues = (uint16_t)num_sub;
+               if (unlikely(packet_ptr1 == NULL)) {
+                       rte_pktmbuf_free(mbuf);
+                       break;
+               }
 
-       if (v)
-               RTE_LOG(INFO, PMD, "Successfully opened rx channels: %u\n",
-                       num_sub);
+               /* get the space available for data in the mbuf */
+               mbp_priv = rte_mempool_get_priv(sze_q->mb_pool);
+               buf_size = (uint16_t)(mbp_priv->mbuf_data_room_size -
+                               RTE_PKTMBUF_HEADROOM);
+
+               if (packet_size <= buf_size) {
+                       /* sze packet will fit in one mbuf, go ahead and copy */
+                       rte_memcpy(rte_pktmbuf_mtod(mbuf, void *),
+                                       packet_ptr1, packet_len1);
+                       if (packet_ptr2 != NULL) {
+                               rte_memcpy((void *)(rte_pktmbuf_mtod(mbuf,
+                                       uint8_t *) + packet_len1),
+                                       packet_ptr2, packet_len2);
+                       }
+                       mbuf->data_len = (uint16_t)packet_size;
 
-       return 0;
+                       mbuf->pkt_len = packet_size;
+                       mbuf->port = sze_q->in_port;
+                       bufs[num_rx] = mbuf;
+                       num_rx++;
+                       num_bytes += packet_size;
+               } else {
+                       /*
+                        * sze packet will not fit in one mbuf,
+                        * scattered mode is not enabled, drop packet
+                        */
+                       RTE_LOG(ERR, PMD,
+                               "SZE segment %d bytes will not fit in one mbuf "
+                               "(%d bytes), scattered mode is not enabled, "
+                               "drop packet!!\n",
+                               packet_size, buf_size);
+                       rte_pktmbuf_free(mbuf);
+               }
+       }
+
+       sze_q->rx_pkts += num_rx;
+       sze_q->rx_bytes += num_bytes;
+       return num_rx;
 }
 
-static int
-init_tx_channels(struct rte_eth_dev *dev, int v)
+static uint16_t
+eth_szedata2_rx_scattered(void *queue,
+               struct rte_mbuf **bufs,
+               uint16_t nb_pkts)
 {
-       struct pmd_internals *internals = dev->data->dev_private;
-       int ret;
-       uint32_t i;
-       uint32_t count = internals->num_of_tx;
-       uint32_t num_sub = 0;
-       uint32_t x;
-       uint32_t rx;
-       uint32_t tx;
+       unsigned int i;
+       struct rte_mbuf *mbuf;
+       struct szedata2_rx_queue *sze_q = queue;
+       struct rte_pktmbuf_pool_private *mbp_priv;
+       uint16_t num_rx = 0;
+       uint16_t buf_size;
+       uint16_t sg_size;
+       uint16_t hw_size;
+       uint16_t packet_size;
+       uint64_t num_bytes = 0;
+       struct szedata *sze = sze_q->sze;
+       uint8_t *header_ptr = NULL; /* header of packet */
+       uint8_t *packet_ptr1 = NULL;
+       uint8_t *packet_ptr2 = NULL;
+       uint16_t packet_len1 = 0;
+       uint16_t packet_len2 = 0;
+       uint16_t hw_data_align;
+
+       if (unlikely(sze_q->sze == NULL || nb_pkts == 0))
+               return 0;
 
-       rx = 0;
-       tx = internals->sze_tx_req;
+       /*
+        * Reads the given number of packets from szedata2 channel given
+        * by queue and copies the packet data into a newly allocated mbuf
+        * to return.
+        */
+       for (i = 0; i < nb_pkts; i++) {
+               const struct szedata_lock *ct_rx_lck_backup;
+               unsigned int ct_rx_rem_bytes_backup;
+               unsigned char *ct_rx_cur_ptr_backup;
+
+               /* get the next sze packet */
+               if (sze->ct_rx_lck != NULL && !sze->ct_rx_rem_bytes &&
+                               sze->ct_rx_lck->next == NULL) {
+                       /* unlock old data */
+                       szedata_rx_unlock_data(sze_q->sze, sze->ct_rx_lck_orig);
+                       sze->ct_rx_lck_orig = NULL;
+                       sze->ct_rx_lck = NULL;
+               }
 
-       for (i = 0; i < count; i++) {
                /*
-                * Open, subscribe rx,tx channels and start device
+                * Store items from sze structure which can be changed
+                * before mbuf allocating. Use these items in case of mbuf
+                * allocating failure.
                 */
-               if (v)
-                       RTE_LOG(INFO, PMD, "Opening SZE device %u. time\n",
-                               i + internals->num_of_rx);
-
-               internals->tx_queue[num_sub].sze =
-                       szedata_open(internals->sze_dev);
-               if (internals->tx_queue[num_sub].sze == NULL)
-                       return -1;
-
-               /* separate least significant non-zero bit */
-               x = tx & ((~tx) + 1);
-
-               if (v)
-                       RTE_LOG(INFO, PMD, "Subscribing rx channel: 0x%x "
-                               "tx channel: 0x%x\n", rx, x);
-
-               ret = szedata_subscribe3(internals->tx_queue[num_sub].sze,
-                               &rx, &x);
-               if (ret) {
-                       szedata_close(internals->tx_queue[num_sub].sze);
-                       internals->tx_queue[num_sub].sze = NULL;
-                       return -1;
+               ct_rx_lck_backup = sze->ct_rx_lck;
+               ct_rx_rem_bytes_backup = sze->ct_rx_rem_bytes;
+               ct_rx_cur_ptr_backup = sze->ct_rx_cur_ptr;
+
+               if (!sze->ct_rx_rem_bytes && sze->ct_rx_lck_orig == NULL) {
+                       /* nothing to read, lock new data */
+                       sze->ct_rx_lck = szedata_rx_lock_data(sze_q->sze, ~0U);
+                       sze->ct_rx_lck_orig = sze->ct_rx_lck;
+
+                       /*
+                        * Backup items from sze structure must be updated
+                        * after locking to contain pointers to new locks.
+                        */
+                       ct_rx_lck_backup = sze->ct_rx_lck;
+                       ct_rx_rem_bytes_backup = sze->ct_rx_rem_bytes;
+                       ct_rx_cur_ptr_backup = sze->ct_rx_cur_ptr;
+
+                       if (sze->ct_rx_lck == NULL)
+                               /* nothing to lock */
+                               break;
+
+                       sze->ct_rx_cur_ptr = sze->ct_rx_lck->start;
+                       sze->ct_rx_rem_bytes = sze->ct_rx_lck->len;
+
+                       if (!sze->ct_rx_rem_bytes)
+                               break;
                }
 
-               if (v)
-                       RTE_LOG(INFO, PMD, "Subscribed rx channel: 0x%x "
-                               "tx channel: 0x%x\n", rx, x);
+               if (sze->ct_rx_rem_bytes < RTE_SZE2_PACKET_HEADER_SIZE) {
+                       /*
+                        * cut in header - copy parts of header to merge buffer
+                        */
+                       if (sze->ct_rx_lck->next == NULL)
+                               break;
+
+                       /* copy first part of header */
+                       rte_memcpy(sze->ct_rx_buffer, sze->ct_rx_cur_ptr,
+                                       sze->ct_rx_rem_bytes);
+
+                       /* copy second part of header */
+                       sze->ct_rx_lck = sze->ct_rx_lck->next;
+                       sze->ct_rx_cur_ptr = sze->ct_rx_lck->start;
+                       rte_memcpy(sze->ct_rx_buffer + sze->ct_rx_rem_bytes,
+                               sze->ct_rx_cur_ptr,
+                               RTE_SZE2_PACKET_HEADER_SIZE -
+                               sze->ct_rx_rem_bytes);
+
+                       sze->ct_rx_cur_ptr += RTE_SZE2_PACKET_HEADER_SIZE -
+                               sze->ct_rx_rem_bytes;
+                       sze->ct_rx_rem_bytes = sze->ct_rx_lck->len -
+                               RTE_SZE2_PACKET_HEADER_SIZE +
+                               sze->ct_rx_rem_bytes;
+
+                       header_ptr = (uint8_t *)sze->ct_rx_buffer;
+               } else {
+                       /* not cut */
+                       header_ptr = (uint8_t *)sze->ct_rx_cur_ptr;
+                       sze->ct_rx_cur_ptr += RTE_SZE2_PACKET_HEADER_SIZE;
+                       sze->ct_rx_rem_bytes -= RTE_SZE2_PACKET_HEADER_SIZE;
+               }
+
+               sg_size = le16toh(*((uint16_t *)header_ptr));
+               hw_size = le16toh(*(((uint16_t *)header_ptr) + 1));
+               packet_size = sg_size -
+                       RTE_SZE2_ALIGN8(RTE_SZE2_PACKET_HEADER_SIZE + hw_size);
+
+
+               /* checks if packet all right */
+               if (!sg_size)
+                       errx(5, "Zero segsize");
+
+               /* check sg_size and hwsize */
+               if (hw_size > sg_size - RTE_SZE2_PACKET_HEADER_SIZE) {
+                       errx(10, "Hwsize bigger than expected. Segsize: %d, "
+                                       "hwsize: %d", sg_size, hw_size);
+               }
 
-               if (x) {
-                       if (v)
-                               RTE_LOG(INFO, PMD, "Starting SZE device for "
-                                       "tx queue: %u\n", num_sub);
+               hw_data_align =
+                       RTE_SZE2_ALIGN8((RTE_SZE2_PACKET_HEADER_SIZE +
+                       hw_size)) - RTE_SZE2_PACKET_HEADER_SIZE;
+
+               if (sze->ct_rx_rem_bytes >=
+                               (uint16_t)(sg_size -
+                               RTE_SZE2_PACKET_HEADER_SIZE)) {
+                       /* no cut */
+                       /* one packet ready - go to another */
+                       packet_ptr1 = sze->ct_rx_cur_ptr + hw_data_align;
+                       packet_len1 = packet_size;
+                       packet_ptr2 = NULL;
+                       packet_len2 = 0;
+
+                       sze->ct_rx_cur_ptr += RTE_SZE2_ALIGN8(sg_size) -
+                               RTE_SZE2_PACKET_HEADER_SIZE;
+                       sze->ct_rx_rem_bytes -= RTE_SZE2_ALIGN8(sg_size) -
+                               RTE_SZE2_PACKET_HEADER_SIZE;
+               } else {
+                       /* cut in data */
+                       if (sze->ct_rx_lck->next == NULL) {
+                               errx(6, "Need \"next\" lock, but it is "
+                                       "missing: %u", sze->ct_rx_rem_bytes);
+                       }
 
-                       ret = szedata_start(internals->tx_queue[num_sub].sze);
-                       if (ret) {
-                               szedata_close(internals->tx_queue[num_sub].sze);
-                               internals->tx_queue[num_sub].sze = NULL;
-                               return -1;
+                       /* skip hw data */
+                       if (sze->ct_rx_rem_bytes <= hw_data_align) {
+                               uint16_t rem_size = hw_data_align -
+                                       sze->ct_rx_rem_bytes;
+
+                               /* MOVE to next lock */
+                               sze->ct_rx_lck = sze->ct_rx_lck->next;
+                               sze->ct_rx_cur_ptr =
+                                       (void *)(((uint8_t *)
+                                       (sze->ct_rx_lck->start)) + rem_size);
+
+                               packet_ptr1 = sze->ct_rx_cur_ptr;
+                               packet_len1 = packet_size;
+                               packet_ptr2 = NULL;
+                               packet_len2 = 0;
+
+                               sze->ct_rx_cur_ptr +=
+                                       RTE_SZE2_ALIGN8(packet_size);
+                               sze->ct_rx_rem_bytes = sze->ct_rx_lck->len -
+                                       rem_size - RTE_SZE2_ALIGN8(packet_size);
+                       } else {
+                               /* get pointer and length from first part */
+                               packet_ptr1 = sze->ct_rx_cur_ptr +
+                                       hw_data_align;
+                               packet_len1 = sze->ct_rx_rem_bytes -
+                                       hw_data_align;
+
+                               /* MOVE to next lock */
+                               sze->ct_rx_lck = sze->ct_rx_lck->next;
+                               sze->ct_rx_cur_ptr = sze->ct_rx_lck->start;
+
+                               /* get pointer and length from second part */
+                               packet_ptr2 = sze->ct_rx_cur_ptr;
+                               packet_len2 = packet_size - packet_len1;
+
+                               sze->ct_rx_cur_ptr +=
+                                       RTE_SZE2_ALIGN8(packet_size) -
+                                       packet_len1;
+                               sze->ct_rx_rem_bytes = sze->ct_rx_lck->len -
+                                       (RTE_SZE2_ALIGN8(packet_size) -
+                                        packet_len1);
                        }
+               }
+
+               if (unlikely(packet_ptr1 == NULL))
+                       break;
 
+               mbuf = rte_pktmbuf_alloc(sze_q->mb_pool);
+
+               if (unlikely(mbuf == NULL)) {
                        /*
-                        * set to 1 all bits lower than bit set to 1
-                        * and that bit to 0
+                        * Restore items from sze structure to state after
+                        * unlocking (eventually locking).
                         */
-                       x -= 1;
-                       internals->tx_queue[num_sub].tx_channel =
-                               count_ones(x);
-
-                       if (v)
-                               RTE_LOG(INFO, PMD, "Subscribed tx channel "
-                                       "no: %u\n",
-                                       internals->tx_queue[num_sub].tx_channel
-                                       );
-
-                       num_sub++;
-                       internals->nb_tx_queues = num_sub;
+                       sze->ct_rx_lck = ct_rx_lck_backup;
+                       sze->ct_rx_rem_bytes = ct_rx_rem_bytes_backup;
+                       sze->ct_rx_cur_ptr = ct_rx_cur_ptr_backup;
+                       break;
+               }
+
+               /* get the space available for data in the mbuf */
+               mbp_priv = rte_mempool_get_priv(sze_q->mb_pool);
+               buf_size = (uint16_t)(mbp_priv->mbuf_data_room_size -
+                               RTE_PKTMBUF_HEADROOM);
+
+               if (packet_size <= buf_size) {
+                       /* sze packet will fit in one mbuf, go ahead and copy */
+                       rte_memcpy(rte_pktmbuf_mtod(mbuf, void *),
+                                       packet_ptr1, packet_len1);
+                       if (packet_ptr2 != NULL) {
+                               rte_memcpy((void *)
+                                       (rte_pktmbuf_mtod(mbuf, uint8_t *) +
+                                       packet_len1), packet_ptr2, packet_len2);
+                       }
+                       mbuf->data_len = (uint16_t)packet_size;
                } else {
-                       if (v)
-                               RTE_LOG(INFO, PMD,
-                                       "Could not subscribe any tx channel. "
-                                       "Closing SZE device\n");
+                       /*
+                        * sze packet will not fit in one mbuf,
+                        * scatter packet into more mbufs
+                        */
+                       struct rte_mbuf *m = mbuf;
+                       uint16_t len = rte_pktmbuf_tailroom(mbuf);
+
+                       /* copy first part of packet */
+                       /* fill first mbuf */
+                       rte_memcpy(rte_pktmbuf_append(mbuf, len), packet_ptr1,
+                               len);
+                       packet_len1 -= len;
+                       packet_ptr1 = ((uint8_t *)packet_ptr1) + len;
+
+                       while (packet_len1 > 0) {
+                               /* fill new mbufs */
+                               m->next = rte_pktmbuf_alloc(sze_q->mb_pool);
+
+                               if (unlikely(m->next == NULL)) {
+                                       rte_pktmbuf_free(mbuf);
+                                       /*
+                                        * Restore items from sze structure
+                                        * to state after unlocking (eventually
+                                        * locking).
+                                        */
+                                       sze->ct_rx_lck = ct_rx_lck_backup;
+                                       sze->ct_rx_rem_bytes =
+                                               ct_rx_rem_bytes_backup;
+                                       sze->ct_rx_cur_ptr =
+                                               ct_rx_cur_ptr_backup;
+                                       goto finish;
+                               }
+
+                               m = m->next;
+
+                               len = RTE_MIN(rte_pktmbuf_tailroom(m),
+                                       packet_len1);
+                               rte_memcpy(rte_pktmbuf_append(mbuf, len),
+                                       packet_ptr1, len);
+
+                               (mbuf->nb_segs)++;
+                               packet_len1 -= len;
+                               packet_ptr1 = ((uint8_t *)packet_ptr1) + len;
+                       }
+
+                       if (packet_ptr2 != NULL) {
+                               /* copy second part of packet, if exists */
+                               /* fill the rest of currently last mbuf */
+                               len = rte_pktmbuf_tailroom(m);
+                               rte_memcpy(rte_pktmbuf_append(mbuf, len),
+                                       packet_ptr2, len);
+                               packet_len2 -= len;
+                               packet_ptr2 = ((uint8_t *)packet_ptr2) + len;
+
+                               while (packet_len2 > 0) {
+                                       /* fill new mbufs */
+                                       m->next = rte_pktmbuf_alloc(
+                                                       sze_q->mb_pool);
+
+                                       if (unlikely(m->next == NULL)) {
+                                               rte_pktmbuf_free(mbuf);
+                                               /*
+                                                * Restore items from sze
+                                                * structure to state after
+                                                * unlocking (eventually
+                                                * locking).
+                                                */
+                                               sze->ct_rx_lck =
+                                                       ct_rx_lck_backup;
+                                               sze->ct_rx_rem_bytes =
+                                                       ct_rx_rem_bytes_backup;
+                                               sze->ct_rx_cur_ptr =
+                                                       ct_rx_cur_ptr_backup;
+                                               goto finish;
+                                       }
+
+                                       m = m->next;
+
+                                       len = RTE_MIN(rte_pktmbuf_tailroom(m),
+                                               packet_len2);
+                                       rte_memcpy(
+                                               rte_pktmbuf_append(mbuf, len),
+                                               packet_ptr2, len);
+
+                                       (mbuf->nb_segs)++;
+                                       packet_len2 -= len;
+                                       packet_ptr2 = ((uint8_t *)packet_ptr2) +
+                                               len;
+                               }
+                       }
+               }
+               mbuf->pkt_len = packet_size;
+               mbuf->port = sze_q->in_port;
+               bufs[num_rx] = mbuf;
+               num_rx++;
+               num_bytes += packet_size;
+       }
+
+finish:
+       sze_q->rx_pkts += num_rx;
+       sze_q->rx_bytes += num_bytes;
+       return num_rx;
+}
+
+static uint16_t
+eth_szedata2_tx(void *queue,
+               struct rte_mbuf **bufs,
+               uint16_t nb_pkts)
+{
+       struct rte_mbuf *mbuf;
+       struct szedata2_tx_queue *sze_q = queue;
+       uint16_t num_tx = 0;
+       uint64_t num_bytes = 0;
+
+       const struct szedata_lock *lck;
+       uint32_t lock_size;
+       uint32_t lock_size2;
+       void *dst;
+       uint32_t pkt_len;
+       uint32_t hwpkt_len;
+       uint32_t unlock_size;
+       uint32_t rem_len;
+       uint8_t mbuf_segs;
+       uint16_t pkt_left = nb_pkts;
+
+       if (sze_q->sze == NULL || nb_pkts == 0)
+               return 0;
+
+       while (pkt_left > 0) {
+               unlock_size = 0;
+               lck = szedata_tx_lock_data(sze_q->sze,
+                       RTE_ETH_SZEDATA2_TX_LOCK_SIZE,
+                       sze_q->tx_channel);
+               if (lck == NULL)
+                       continue;
+
+               dst = lck->start;
+               lock_size = lck->len;
+               lock_size2 = lck->next ? lck->next->len : 0;
+
+next_packet:
+               mbuf = bufs[nb_pkts - pkt_left];
+
+               pkt_len = mbuf->pkt_len;
+               mbuf_segs = mbuf->nb_segs;
+
+               hwpkt_len = RTE_SZE2_PACKET_HEADER_SIZE_ALIGNED +
+                       RTE_SZE2_ALIGN8(pkt_len);
+
+               if (lock_size + lock_size2 < hwpkt_len) {
+                       szedata_tx_unlock_data(sze_q->sze, lck, unlock_size);
+                       continue;
+               }
+
+               num_bytes += pkt_len;
+
+               if (lock_size > hwpkt_len) {
+                       void *tmp_dst;
+
+                       rem_len = 0;
+
+                       /* write packet length at first 2 bytes in 8B header */
+                       *((uint16_t *)dst) = htole16(
+                                       RTE_SZE2_PACKET_HEADER_SIZE_ALIGNED +
+                                       pkt_len);
+                       *(((uint16_t *)dst) + 1) = htole16(0);
+
+                       /* copy packet from mbuf */
+                       tmp_dst = ((uint8_t *)(dst)) +
+                               RTE_SZE2_PACKET_HEADER_SIZE_ALIGNED;
+                       if (mbuf_segs == 1) {
+                               /*
+                                * non-scattered packet,
+                                * transmit from one mbuf
+                                */
+                               rte_memcpy(tmp_dst,
+                                       rte_pktmbuf_mtod(mbuf, const void *),
+                                       pkt_len);
+                       } else {
+                               /* scattered packet, transmit from more mbufs */
+                               struct rte_mbuf *m = mbuf;
+                               while (m) {
+                                       rte_memcpy(tmp_dst,
+                                               rte_pktmbuf_mtod(m,
+                                               const void *),
+                                               m->data_len);
+                                       tmp_dst = ((uint8_t *)(tmp_dst)) +
+                                               m->data_len;
+                                       m = m->next;
+                               }
+                       }
+
+
+                       dst = ((uint8_t *)dst) + hwpkt_len;
+                       unlock_size += hwpkt_len;
+                       lock_size -= hwpkt_len;
+
+                       rte_pktmbuf_free(mbuf);
+                       num_tx++;
+                       pkt_left--;
+                       if (pkt_left == 0) {
+                               szedata_tx_unlock_data(sze_q->sze, lck,
+                                       unlock_size);
+                               break;
+                       }
+                       goto next_packet;
+               } else if (lock_size + lock_size2 >= hwpkt_len) {
+                       void *tmp_dst;
+                       uint16_t write_len;
+
+                       /* write packet length at first 2 bytes in 8B header */
+                       *((uint16_t *)dst) =
+                               htole16(RTE_SZE2_PACKET_HEADER_SIZE_ALIGNED +
+                                       pkt_len);
+                       *(((uint16_t *)dst) + 1) = htole16(0);
+
+                       /*
+                        * If the raw packet (pkt_len) is smaller than lock_size
+                        * get the correct length for memcpy
+                        */
+                       write_len =
+                               pkt_len < lock_size -
+                               RTE_SZE2_PACKET_HEADER_SIZE_ALIGNED ?
+                               pkt_len :
+                               lock_size - RTE_SZE2_PACKET_HEADER_SIZE_ALIGNED;
+
+                       rem_len = hwpkt_len - lock_size;
+
+                       tmp_dst = ((uint8_t *)(dst)) +
+                               RTE_SZE2_PACKET_HEADER_SIZE_ALIGNED;
+                       if (mbuf_segs == 1) {
+                               /*
+                                * non-scattered packet,
+                                * transmit from one mbuf
+                                */
+                               /* copy part of packet to first area */
+                               rte_memcpy(tmp_dst,
+                                       rte_pktmbuf_mtod(mbuf, const void *),
+                                       write_len);
+
+                               if (lck->next)
+                                       dst = lck->next->start;
+
+                               /* copy part of packet to second area */
+                               rte_memcpy(dst,
+                                       (const void *)(rte_pktmbuf_mtod(mbuf,
+                                                       const uint8_t *) +
+                                       write_len), pkt_len - write_len);
+                       } else {
+                               /* scattered packet, transmit from more mbufs */
+                               struct rte_mbuf *m = mbuf;
+                               uint16_t written = 0;
+                               uint16_t to_write = 0;
+                               bool new_mbuf = true;
+                               uint16_t write_off = 0;
+
+                               /* copy part of packet to first area */
+                               while (m && written < write_len) {
+                                       to_write = RTE_MIN(m->data_len,
+                                                       write_len - written);
+                                       rte_memcpy(tmp_dst,
+                                               rte_pktmbuf_mtod(m,
+                                                       const void *),
+                                               to_write);
+
+                                       tmp_dst = ((uint8_t *)(tmp_dst)) +
+                                               to_write;
+                                       if (m->data_len <= write_len -
+                                                       written) {
+                                               m = m->next;
+                                               new_mbuf = true;
+                                       } else {
+                                               new_mbuf = false;
+                                       }
+                                       written += to_write;
+                               }
+
+                               if (lck->next)
+                                       dst = lck->next->start;
+
+                               tmp_dst = dst;
+                               written = 0;
+                               write_off = new_mbuf ? 0 : to_write;
+
+                               /* copy part of packet to second area */
+                               while (m && written < pkt_len - write_len) {
+                                       rte_memcpy(tmp_dst, (const void *)
+                                               (rte_pktmbuf_mtod(m,
+                                               uint8_t *) + write_off),
+                                               m->data_len - write_off);
+
+                                       tmp_dst = ((uint8_t *)(tmp_dst)) +
+                                               (m->data_len - write_off);
+                                       written += m->data_len - write_off;
+                                       m = m->next;
+                                       write_off = 0;
+                               }
+                       }
+
+                       dst = ((uint8_t *)dst) + rem_len;
+                       unlock_size += hwpkt_len;
+                       lock_size = lock_size2 - rem_len;
+                       lock_size2 = 0;
 
-                       szedata_close(internals->tx_queue[num_sub].sze);
-                       internals->tx_queue[num_sub].sze = NULL;
+                       rte_pktmbuf_free(mbuf);
+                       num_tx++;
                }
 
-               /* set least significant non-zero bit to zero */
-               tx = tx & (tx - 1);
+               szedata_tx_unlock_data(sze_q->sze, lck, unlock_size);
+               pkt_left--;
        }
 
-       dev->data->nb_tx_queues = (uint16_t)num_sub;
+       sze_q->tx_pkts += num_tx;
+       sze_q->err_pkts += nb_pkts - num_tx;
+       sze_q->tx_bytes += num_bytes;
+       return num_tx;
+}
 
-       if (v)
-               RTE_LOG(INFO, PMD, "Successfully opened tx channels: %u\n",
-                       num_sub);
+static int
+eth_rx_queue_start(struct rte_eth_dev *dev, uint16_t rxq_id)
+{
+       struct szedata2_rx_queue *rxq = dev->data->rx_queues[rxq_id];
+       int ret;
+       struct pmd_internals *internals = (struct pmd_internals *)
+               dev->data->dev_private;
+
+       if (rxq->sze == NULL) {
+               uint32_t rx = 1 << rxq->rx_channel;
+               uint32_t tx = 0;
+               rxq->sze = szedata_open(internals->sze_dev);
+               if (rxq->sze == NULL)
+                       return -EINVAL;
+               ret = szedata_subscribe3(rxq->sze, &rx, &tx);
+               if (ret != 0 || rx == 0)
+                       goto err;
+       }
 
+       ret = szedata_start(rxq->sze);
+       if (ret != 0)
+               goto err;
+       dev->data->rx_queue_state[rxq_id] = RTE_ETH_QUEUE_STATE_STARTED;
        return 0;
+
+err:
+       szedata_close(rxq->sze);
+       rxq->sze = NULL;
+       return -EINVAL;
 }
 
-static void
-close_rx_channels(struct rte_eth_dev *dev)
+static int
+eth_rx_queue_stop(struct rte_eth_dev *dev, uint16_t rxq_id)
 {
-       struct pmd_internals *internals = dev->data->dev_private;
-       uint32_t i;
-       uint32_t num_sub = internals->nb_rx_queues;
+       struct szedata2_rx_queue *rxq = dev->data->rx_queues[rxq_id];
 
-       for (i = 0; i < num_sub; i++) {
-               if (internals->rx_queue[i].sze != NULL) {
-                       szedata_close(internals->rx_queue[i].sze);
-                       internals->rx_queue[i].sze = NULL;
-               }
+       if (rxq->sze != NULL) {
+               szedata_close(rxq->sze);
+               rxq->sze = NULL;
        }
-       /* set number of rx queues to zero */
-       internals->nb_rx_queues = 0;
-       dev->data->nb_rx_queues = (uint16_t)0;
+
+       dev->data->rx_queue_state[rxq_id] = RTE_ETH_QUEUE_STATE_STOPPED;
+       return 0;
 }
 
-static void
-close_tx_channels(struct rte_eth_dev *dev)
+static int
+eth_tx_queue_start(struct rte_eth_dev *dev, uint16_t txq_id)
 {
-       struct pmd_internals *internals = dev->data->dev_private;
-       uint32_t i;
-       uint32_t num_sub = internals->nb_tx_queues;
+       struct szedata2_tx_queue *txq = dev->data->tx_queues[txq_id];
+       int ret;
+       struct pmd_internals *internals = (struct pmd_internals *)
+               dev->data->dev_private;
+
+       if (txq->sze == NULL) {
+               uint32_t rx = 0;
+               uint32_t tx = 1 << txq->tx_channel;
+               txq->sze = szedata_open(internals->sze_dev);
+               if (txq->sze == NULL)
+                       return -EINVAL;
+               ret = szedata_subscribe3(txq->sze, &rx, &tx);
+               if (ret != 0 || tx == 0)
+                       goto err;
+       }
 
-       for (i = 0; i < num_sub; i++) {
-               if (internals->tx_queue[i].sze != NULL) {
-                       szedata_close(internals->tx_queue[i].sze);
-                       internals->tx_queue[i].sze = NULL;
-               }
+       ret = szedata_start(txq->sze);
+       if (ret != 0)
+               goto err;
+       dev->data->tx_queue_state[txq_id] = RTE_ETH_QUEUE_STATE_STARTED;
+       return 0;
+
+err:
+       szedata_close(txq->sze);
+       txq->sze = NULL;
+       return -EINVAL;
+}
+
+static int
+eth_tx_queue_stop(struct rte_eth_dev *dev, uint16_t txq_id)
+{
+       struct szedata2_tx_queue *txq = dev->data->tx_queues[txq_id];
+
+       if (txq->sze != NULL) {
+               szedata_close(txq->sze);
+               txq->sze = NULL;
        }
-       /* set number of rx queues to zero */
-       internals->nb_tx_queues = 0;
-       dev->data->nb_tx_queues = (uint16_t)0;
+
+       dev->data->tx_queue_state[txq_id] = RTE_ETH_QUEUE_STATE_STOPPED;
+       return 0;
 }
 
 static int
 eth_dev_start(struct rte_eth_dev *dev)
 {
-       struct pmd_internals *internals = dev->data->dev_private;
        int ret;
-
-       if (internals->nb_rx_queues == 0) {
-               ret = init_rx_channels(dev, 0);
-               if (ret != 0) {
-                       close_rx_channels(dev);
-                       return -1;
-               }
+       uint16_t i;
+       uint16_t nb_rx = dev->data->nb_rx_queues;
+       uint16_t nb_tx = dev->data->nb_tx_queues;
+
+       for (i = 0; i < nb_rx; i++) {
+               ret = eth_rx_queue_start(dev, i);
+               if (ret != 0)
+                       goto err_rx;
        }
 
-       if (internals->nb_tx_queues == 0) {
-               ret = init_tx_channels(dev, 0);
-               if (ret != 0) {
-                       close_tx_channels(dev);
-                       close_rx_channels(dev);
-                       return -1;
-               }
+       for (i = 0; i < nb_tx; i++) {
+               ret = eth_tx_queue_start(dev, i);
+               if (ret != 0)
+                       goto err_tx;
        }
 
-       dev->data->dev_link.link_status = 1;
        return 0;
+
+err_tx:
+       for (i = 0; i < nb_tx; i++)
+               eth_tx_queue_stop(dev, i);
+err_rx:
+       for (i = 0; i < nb_rx; i++)
+               eth_rx_queue_stop(dev, i);
+       return ret;
 }
 
 static void
 eth_dev_stop(struct rte_eth_dev *dev)
 {
-       unsigned i;
-       struct pmd_internals *internals = dev->data->dev_private;
-
-       for (i = 0; i < internals->nb_rx_queues; i++) {
-               if (internals->rx_queue[i].sze != NULL) {
-                       szedata_close(internals->rx_queue[i].sze);
-                       internals->rx_queue[i].sze = NULL;
-               }
-       }
-
-       for (i = 0; i < internals->nb_tx_queues; i++) {
-               if (internals->tx_queue[i].sze != NULL) {
-                       szedata_close(internals->tx_queue[i].sze);
-                       internals->tx_queue[i].sze = NULL;
-               }
-       }
+       uint16_t i;
+       uint16_t nb_rx = dev->data->nb_rx_queues;
+       uint16_t nb_tx = dev->data->nb_tx_queues;
 
-       internals->nb_rx_queues = 0;
-       internals->nb_tx_queues = 0;
+       for (i = 0; i < nb_tx; i++)
+               eth_tx_queue_stop(dev, i);
 
-       dev->data->nb_rx_queues = (uint16_t)0;
-       dev->data->nb_tx_queues = (uint16_t)0;
-
-       dev->data->dev_link.link_status = 0;
+       for (i = 0; i < nb_rx; i++)
+               eth_rx_queue_stop(dev, i);
 }
 
 static int
-eth_dev_configure(struct rte_eth_dev *dev __rte_unused)
+eth_dev_configure(struct rte_eth_dev *dev)
 {
+       struct rte_eth_dev_data *data = dev->data;
+       if (data->dev_conf.rxmode.enable_scatter == 1) {
+               dev->rx_pkt_burst = eth_szedata2_rx_scattered;
+               data->scattered_rx = 1;
+       } else {
+               dev->rx_pkt_burst = eth_szedata2_rx;
+               data->scattered_rx = 0;
+       }
        return 0;
 }
 
@@ -423,44 +1030,43 @@ eth_dev_info(struct rte_eth_dev *dev,
                struct rte_eth_dev_info *dev_info)
 {
        struct pmd_internals *internals = dev->data->dev_private;
-       dev_info->driver_name = drivername;
-       dev_info->if_index = internals->if_index;
+       dev_info->if_index = 0;
        dev_info->max_mac_addrs = 1;
        dev_info->max_rx_pktlen = (uint32_t)-1;
-       dev_info->max_rx_queues = (uint16_t)internals->nb_rx_queues;
-       dev_info->max_tx_queues = (uint16_t)internals->nb_tx_queues;
+       dev_info->max_rx_queues = internals->max_rx_queues;
+       dev_info->max_tx_queues = internals->max_tx_queues;
        dev_info->min_rx_bufsize = 0;
-       dev_info->pci_dev = NULL;
+       dev_info->speed_capa = ETH_LINK_SPEED_100G;
 }
 
 static void
 eth_stats_get(struct rte_eth_dev *dev,
                struct rte_eth_stats *stats)
 {
-       unsigned i;
+       uint16_t i;
+       uint16_t nb_rx = dev->data->nb_rx_queues;
+       uint16_t nb_tx = dev->data->nb_tx_queues;
        uint64_t rx_total = 0;
        uint64_t tx_total = 0;
        uint64_t tx_err_total = 0;
        uint64_t rx_total_bytes = 0;
        uint64_t tx_total_bytes = 0;
-       const struct pmd_internals *internal = dev->data->dev_private;
+       const struct pmd_internals *internals = dev->data->dev_private;
 
-       for (i = 0; i < RTE_ETHDEV_QUEUE_STAT_CNTRS &&
-                       i < internal->nb_rx_queues; i++) {
-               stats->q_ipackets[i] = internal->rx_queue[i].rx_pkts;
-               stats->q_ibytes[i] = internal->rx_queue[i].rx_bytes;
+       for (i = 0; i < RTE_ETHDEV_QUEUE_STAT_CNTRS && i < nb_rx; i++) {
+               stats->q_ipackets[i] = internals->rx_queue[i].rx_pkts;
+               stats->q_ibytes[i] = internals->rx_queue[i].rx_bytes;
                rx_total += stats->q_ipackets[i];
                rx_total_bytes += stats->q_ibytes[i];
        }
 
-       for (i = 0; i < RTE_ETHDEV_QUEUE_STAT_CNTRS &&
-                       i < internal->nb_tx_queues; i++) {
-               stats->q_opackets[i] = internal->tx_queue[i].tx_pkts;
-               stats->q_errors[i] = internal->tx_queue[i].err_pkts;
-               stats->q_obytes[i] = internal->tx_queue[i].tx_bytes;
+       for (i = 0; i < RTE_ETHDEV_QUEUE_STAT_CNTRS && i < nb_tx; i++) {
+               stats->q_opackets[i] = internals->tx_queue[i].tx_pkts;
+               stats->q_obytes[i] = internals->tx_queue[i].tx_bytes;
+               stats->q_errors[i] = internals->tx_queue[i].err_pkts;
                tx_total += stats->q_opackets[i];
-               tx_err_total += stats->q_errors[i];
                tx_total_bytes += stats->q_obytes[i];
+               tx_err_total += stats->q_errors[i];
        }
 
        stats->ipackets = rx_total;
@@ -473,55 +1079,131 @@ eth_stats_get(struct rte_eth_dev *dev,
 static void
 eth_stats_reset(struct rte_eth_dev *dev)
 {
-       unsigned i;
-       struct pmd_internals *internal = dev->data->dev_private;
-       for (i = 0; i < internal->nb_rx_queues; i++) {
-               internal->rx_queue[i].rx_pkts = 0;
-               internal->rx_queue[i].rx_bytes = 0;
+       uint16_t i;
+       uint16_t nb_rx = dev->data->nb_rx_queues;
+       uint16_t nb_tx = dev->data->nb_tx_queues;
+       struct pmd_internals *internals = dev->data->dev_private;
+
+       for (i = 0; i < nb_rx; i++) {
+               internals->rx_queue[i].rx_pkts = 0;
+               internals->rx_queue[i].rx_bytes = 0;
+               internals->rx_queue[i].err_pkts = 0;
+       }
+       for (i = 0; i < nb_tx; i++) {
+               internals->tx_queue[i].tx_pkts = 0;
+               internals->tx_queue[i].tx_bytes = 0;
+               internals->tx_queue[i].err_pkts = 0;
+       }
+}
+
+static void
+eth_rx_queue_release(void *q)
+{
+       struct szedata2_rx_queue *rxq = (struct szedata2_rx_queue *)q;
+       if (rxq->sze != NULL) {
+               szedata_close(rxq->sze);
+               rxq->sze = NULL;
        }
-       for (i = 0; i < internal->nb_tx_queues; i++) {
-               internal->tx_queue[i].tx_pkts = 0;
-               internal->tx_queue[i].err_pkts = 0;
-               internal->tx_queue[i].tx_bytes = 0;
+}
+
+static void
+eth_tx_queue_release(void *q)
+{
+       struct szedata2_tx_queue *txq = (struct szedata2_tx_queue *)q;
+       if (txq->sze != NULL) {
+               szedata_close(txq->sze);
+               txq->sze = NULL;
        }
 }
 
 static void
 eth_dev_close(struct rte_eth_dev *dev)
 {
-       unsigned i;
-       struct pmd_internals *internals = dev->data->dev_private;
+       uint16_t i;
+       uint16_t nb_rx = dev->data->nb_rx_queues;
+       uint16_t nb_tx = dev->data->nb_tx_queues;
 
-       for (i = 0; i < internals->nb_rx_queues; i++) {
-               if (internals->rx_queue[i].sze != NULL) {
-                       szedata_close(internals->rx_queue[i].sze);
-                       internals->rx_queue[i].sze = NULL;
-               }
+       eth_dev_stop(dev);
+
+       for (i = 0; i < nb_rx; i++) {
+               eth_rx_queue_release(dev->data->rx_queues[i]);
+               dev->data->rx_queues[i] = NULL;
        }
+       dev->data->nb_rx_queues = 0;
+       for (i = 0; i < nb_tx; i++) {
+               eth_tx_queue_release(dev->data->tx_queues[i]);
+               dev->data->tx_queues[i] = NULL;
+       }
+       dev->data->nb_tx_queues = 0;
+}
 
-       for (i = 0; i < internals->nb_tx_queues; i++) {
-               if (internals->tx_queue[i].sze != NULL) {
-                       szedata_close(internals->tx_queue[i].sze);
-                       internals->tx_queue[i].sze = NULL;
-               }
+static int
+eth_link_update(struct rte_eth_dev *dev,
+               int wait_to_complete __rte_unused)
+{
+       struct rte_eth_link link;
+       struct rte_eth_link *link_ptr = &link;
+       struct rte_eth_link *dev_link = &dev->data->dev_link;
+       volatile struct szedata2_cgmii_ibuf *ibuf = SZEDATA2_PCI_RESOURCE_PTR(
+                       dev, SZEDATA2_CGMII_IBUF_BASE_OFF,
+                       volatile struct szedata2_cgmii_ibuf *);
+
+       switch (cgmii_link_speed(ibuf)) {
+       case SZEDATA2_LINK_SPEED_10G:
+               link.link_speed = ETH_SPEED_NUM_10G;
+               break;
+       case SZEDATA2_LINK_SPEED_40G:
+               link.link_speed = ETH_SPEED_NUM_40G;
+               break;
+       case SZEDATA2_LINK_SPEED_100G:
+               link.link_speed = ETH_SPEED_NUM_100G;
+               break;
+       default:
+               link.link_speed = ETH_SPEED_NUM_10G;
+               break;
        }
 
-       internals->nb_rx_queues = 0;
-       internals->nb_tx_queues = 0;
+       /* szedata2 uses only full duplex */
+       link.link_duplex = ETH_LINK_FULL_DUPLEX;
 
-       dev->data->nb_rx_queues = (uint16_t)0;
-       dev->data->nb_tx_queues = (uint16_t)0;
+       link.link_status = (cgmii_ibuf_is_enabled(ibuf) &&
+                       cgmii_ibuf_is_link_up(ibuf)) ? ETH_LINK_UP : ETH_LINK_DOWN;
+
+       link.link_autoneg = ETH_LINK_SPEED_FIXED;
+
+       rte_atomic64_cmpset((uint64_t *)dev_link, *(uint64_t *)dev_link,
+                       *(uint64_t *)link_ptr);
+
+       return 0;
 }
 
-static void
-eth_queue_release(void *q __rte_unused)
+static int
+eth_dev_set_link_up(struct rte_eth_dev *dev)
 {
+       volatile struct szedata2_cgmii_ibuf *ibuf = SZEDATA2_PCI_RESOURCE_PTR(
+                       dev, SZEDATA2_CGMII_IBUF_BASE_OFF,
+                       volatile struct szedata2_cgmii_ibuf *);
+       volatile struct szedata2_cgmii_obuf *obuf = SZEDATA2_PCI_RESOURCE_PTR(
+                       dev, SZEDATA2_CGMII_OBUF_BASE_OFF,
+                       volatile struct szedata2_cgmii_obuf *);
+
+       cgmii_ibuf_enable(ibuf);
+       cgmii_obuf_enable(obuf);
+       return 0;
 }
 
 static int
-eth_link_update(struct rte_eth_dev *dev __rte_unused,
-               int wait_to_complete __rte_unused)
+eth_dev_set_link_down(struct rte_eth_dev *dev)
 {
+       volatile struct szedata2_cgmii_ibuf *ibuf = SZEDATA2_PCI_RESOURCE_PTR(
+                       dev, SZEDATA2_CGMII_IBUF_BASE_OFF,
+                       volatile struct szedata2_cgmii_ibuf *);
+       volatile struct szedata2_cgmii_obuf *obuf = SZEDATA2_PCI_RESOURCE_PTR(
+                       dev, SZEDATA2_CGMII_OBUF_BASE_OFF,
+                       volatile struct szedata2_cgmii_obuf *);
+
+       cgmii_ibuf_disable(ibuf);
+       cgmii_obuf_disable(obuf);
        return 0;
 }
 
@@ -534,11 +1216,28 @@ eth_rx_queue_setup(struct rte_eth_dev *dev,
                struct rte_mempool *mb_pool)
 {
        struct pmd_internals *internals = dev->data->dev_private;
-       struct szedata2_rx_queue *szedata2_q =
-               &internals->rx_queue[rx_queue_id];
-       szedata2_q->mb_pool = mb_pool;
-       dev->data->rx_queues[rx_queue_id] = szedata2_q;
-       szedata2_q->in_port = dev->data->port_id;
+       struct szedata2_rx_queue *rxq = &internals->rx_queue[rx_queue_id];
+       int ret;
+       uint32_t rx = 1 << rx_queue_id;
+       uint32_t tx = 0;
+
+       rxq->sze = szedata_open(internals->sze_dev);
+       if (rxq->sze == NULL)
+               return -EINVAL;
+       ret = szedata_subscribe3(rxq->sze, &rx, &tx);
+       if (ret != 0 || rx == 0) {
+               szedata_close(rxq->sze);
+               rxq->sze = NULL;
+               return -EINVAL;
+       }
+       rxq->rx_channel = rx_queue_id;
+       rxq->in_port = dev->data->port_id;
+       rxq->mb_pool = mb_pool;
+       rxq->rx_pkts = 0;
+       rxq->rx_bytes = 0;
+       rxq->err_pkts = 0;
+
+       dev->data->rx_queues[rx_queue_id] = rxq;
        return 0;
 }
 
@@ -550,7 +1249,26 @@ eth_tx_queue_setup(struct rte_eth_dev *dev,
                const struct rte_eth_txconf *tx_conf __rte_unused)
 {
        struct pmd_internals *internals = dev->data->dev_private;
-       dev->data->tx_queues[tx_queue_id] = &internals->tx_queue[tx_queue_id];
+       struct szedata2_tx_queue *txq = &internals->tx_queue[tx_queue_id];
+       int ret;
+       uint32_t rx = 0;
+       uint32_t tx = 1 << tx_queue_id;
+
+       txq->sze = szedata_open(internals->sze_dev);
+       if (txq->sze == NULL)
+               return -EINVAL;
+       ret = szedata_subscribe3(txq->sze, &rx, &tx);
+       if (ret != 0 || tx == 0) {
+               szedata_close(txq->sze);
+               txq->sze = NULL;
+               return -EINVAL;
+       }
+       txq->tx_channel = tx_queue_id;
+       txq->tx_pkts = 0;
+       txq->tx_bytes = 0;
+       txq->err_pkts = 0;
+
+       dev->data->tx_queues[tx_queue_id] = txq;
        return 0;
 }
 
@@ -560,268 +1278,310 @@ eth_mac_addr_set(struct rte_eth_dev *dev __rte_unused,
 {
 }
 
-static struct eth_dev_ops ops = {
-               .dev_start          = eth_dev_start,
-               .dev_stop           = eth_dev_stop,
-               .dev_close          = eth_dev_close,
-               .dev_configure      = eth_dev_configure,
-               .dev_infos_get      = eth_dev_info,
-               .rx_queue_setup     = eth_rx_queue_setup,
-               .tx_queue_setup     = eth_tx_queue_setup,
-               .rx_queue_release   = eth_queue_release,
-               .tx_queue_release   = eth_queue_release,
-               .link_update        = eth_link_update,
-               .stats_get          = eth_stats_get,
-               .stats_reset        = eth_stats_reset,
-               .mac_addr_set       = eth_mac_addr_set,
-};
-
-static int
-parse_mask(const char *mask_str, uint32_t *mask_num)
+static void
+eth_promiscuous_enable(struct rte_eth_dev *dev)
 {
-       char *endptr;
-       long int value;
-
-       value = strtol(mask_str, &endptr, 0);
-       if (*endptr != '\0' || value > UINT32_MAX || value < 0)
-               return -1;
-
-       *mask_num = (uint32_t)value;
-       return 0;
+       volatile struct szedata2_cgmii_ibuf *ibuf = SZEDATA2_PCI_RESOURCE_PTR(
+                       dev, SZEDATA2_CGMII_IBUF_BASE_OFF,
+                       volatile struct szedata2_cgmii_ibuf *);
+       cgmii_ibuf_mac_mode_write(ibuf, SZEDATA2_MAC_CHMODE_PROMISC);
 }
 
-static int
-add_rx_mask(const char *key __rte_unused, const char *value, void *extra_args)
+static void
+eth_promiscuous_disable(struct rte_eth_dev *dev)
 {
-       struct rxtx_szedata2 *szedata2 = extra_args;
-       uint32_t mask;
-
-       if (parse_mask(value, &mask) != 0)
-               return -1;
-
-       szedata2->sze_rx_mask_req |= mask;
-       return 0;
+       volatile struct szedata2_cgmii_ibuf *ibuf = SZEDATA2_PCI_RESOURCE_PTR(
+                       dev, SZEDATA2_CGMII_IBUF_BASE_OFF,
+                       volatile struct szedata2_cgmii_ibuf *);
+       cgmii_ibuf_mac_mode_write(ibuf, SZEDATA2_MAC_CHMODE_ONLY_VALID);
 }
 
-static int
-add_tx_mask(const char *key __rte_unused, const char *value, void *extra_args)
+static void
+eth_allmulticast_enable(struct rte_eth_dev *dev)
 {
-       struct rxtx_szedata2 *szedata2 = extra_args;
-       uint32_t mask;
-
-       if (parse_mask(value, &mask) != 0)
-               return -1;
-
-       szedata2->sze_tx_mask_req |= mask;
-       return 0;
+       volatile struct szedata2_cgmii_ibuf *ibuf = SZEDATA2_PCI_RESOURCE_PTR(
+                       dev, SZEDATA2_CGMII_IBUF_BASE_OFF,
+                       volatile struct szedata2_cgmii_ibuf *);
+       cgmii_ibuf_mac_mode_write(ibuf, SZEDATA2_MAC_CHMODE_ALL_MULTICAST);
 }
 
-static int
-rte_pmd_init_internals(const char *name, const unsigned nb_rx_queues,
-               const unsigned nb_tx_queues,
-               const unsigned numa_node,
-               struct pmd_internals **internals,
-               struct rte_eth_dev **eth_dev)
+static void
+eth_allmulticast_disable(struct rte_eth_dev *dev)
 {
-       struct rte_eth_dev_data *data = NULL;
-
-       RTE_LOG(INFO, PMD,
-                       "Creating szedata2-backed ethdev on numa socket %u\n",
-                       numa_node);
-
-       /*
-        * now do all data allocation - for eth_dev structure
-        * and internal (private) data
-        */
-       data = rte_zmalloc_socket(name, sizeof(*data), 0, numa_node);
-       if (data == NULL)
-               goto error;
-
-       *internals = rte_zmalloc_socket(name, sizeof(**internals), 0,
-                       numa_node);
-       if (*internals == NULL)
-               goto error;
-
-       /* reserve an ethdev entry */
-       *eth_dev = rte_eth_dev_allocate(name, RTE_ETH_DEV_VIRTUAL);
-       if (*eth_dev == NULL)
-               goto error;
-
-       /*
-        * now put it all together
-        * - store queue data in internals,
-        * - store numa_node info in pci_driver
-        * - point eth_dev_data to internals
-        * - and point eth_dev structure to new eth_dev_data structure
-        *
-        * NOTE: we'll replace the data element, of originally allocated eth_dev
-        * so the rings are local per-process
-        */
-
-       (*internals)->nb_rx_queues = nb_rx_queues;
-       (*internals)->nb_tx_queues = nb_tx_queues;
-
-       (*internals)->if_index = 0;
-
-       data->dev_private = *internals;
-       data->port_id = (*eth_dev)->data->port_id;
-       snprintf(data->name, sizeof(data->name), "%s", (*eth_dev)->data->name);
-       data->nb_rx_queues = (uint16_t)nb_rx_queues;
-       data->nb_tx_queues = (uint16_t)nb_tx_queues;
-       data->dev_link = pmd_link;
-       data->mac_addrs = &eth_addr;
-
-       (*eth_dev)->data = data;
-       (*eth_dev)->dev_ops = &ops;
-       (*eth_dev)->data->dev_flags = RTE_ETH_DEV_DETACHABLE;
-       (*eth_dev)->driver = NULL;
-       (*eth_dev)->data->kdrv = RTE_KDRV_NONE;
-       (*eth_dev)->data->drv_name = drivername;
-       (*eth_dev)->data->numa_node = numa_node;
-
-       return 0;
-
-error:
-       rte_free(data);
-       rte_free(*internals);
-       return -1;
+       volatile struct szedata2_cgmii_ibuf *ibuf = SZEDATA2_PCI_RESOURCE_PTR(
+                       dev, SZEDATA2_CGMII_IBUF_BASE_OFF,
+                       volatile struct szedata2_cgmii_ibuf *);
+       cgmii_ibuf_mac_mode_write(ibuf, SZEDATA2_MAC_CHMODE_ONLY_VALID);
 }
 
+static const struct eth_dev_ops ops = {
+       .dev_start          = eth_dev_start,
+       .dev_stop           = eth_dev_stop,
+       .dev_set_link_up    = eth_dev_set_link_up,
+       .dev_set_link_down  = eth_dev_set_link_down,
+       .dev_close          = eth_dev_close,
+       .dev_configure      = eth_dev_configure,
+       .dev_infos_get      = eth_dev_info,
+       .promiscuous_enable   = eth_promiscuous_enable,
+       .promiscuous_disable  = eth_promiscuous_disable,
+       .allmulticast_enable  = eth_allmulticast_enable,
+       .allmulticast_disable = eth_allmulticast_disable,
+       .rx_queue_start     = eth_rx_queue_start,
+       .rx_queue_stop      = eth_rx_queue_stop,
+       .tx_queue_start     = eth_tx_queue_start,
+       .tx_queue_stop      = eth_tx_queue_stop,
+       .rx_queue_setup     = eth_rx_queue_setup,
+       .tx_queue_setup     = eth_tx_queue_setup,
+       .rx_queue_release   = eth_rx_queue_release,
+       .tx_queue_release   = eth_tx_queue_release,
+       .link_update        = eth_link_update,
+       .stats_get          = eth_stats_get,
+       .stats_reset        = eth_stats_reset,
+       .mac_addr_set       = eth_mac_addr_set,
+};
+
+/*
+ * This function goes through sysfs and looks for an index of szedata2
+ * device file (/dev/szedataIIX, where X is the index).
+ *
+ * @return
+ *           0 on success
+ *          -1 on error
+ */
 static int
-rte_eth_from_szedata2(const char *name,
-               struct rxtx_szedata2 *szedata2,
-               const unsigned numa_node)
+get_szedata2_index(struct rte_eth_dev *dev, uint32_t *index)
 {
-       struct pmd_internals *internals = NULL;
-       struct rte_eth_dev *eth_dev = NULL;
+       DIR *dir;
+       struct dirent *entry;
        int ret;
-
-       if (rte_pmd_init_internals(name, 0, 0, numa_node,
-                       &internals, &eth_dev) < 0)
+       uint32_t tmp_index;
+       FILE *fd;
+       char pcislot_path[PATH_MAX];
+       struct rte_pci_addr pcislot_addr = dev->pci_dev->addr;
+       uint32_t domain;
+       uint32_t bus;
+       uint32_t devid;
+       uint32_t function;
+
+       dir = opendir("/sys/class/combo");
+       if (dir == NULL)
                return -1;
 
-       internals->sze_dev = szedata2->sze_dev;
-       internals->sze_rx_req = szedata2->sze_rx_mask_req;
-       internals->sze_tx_req = szedata2->sze_tx_mask_req;
-       internals->num_of_rx = szedata2->num_of_rx;
-       internals->num_of_tx = szedata2->num_of_tx;
-
-       RTE_LOG(INFO, PMD, "Number of rx channels to open: %u mask: 0x%x\n",
-                       internals->num_of_rx, internals->sze_rx_req);
-       RTE_LOG(INFO, PMD, "Number of tx channels to open: %u mask: 0x%x\n",
-                       internals->num_of_tx, internals->sze_tx_req);
-
-       ret = init_rx_channels(eth_dev, 1);
-       if (ret != 0) {
-               close_rx_channels(eth_dev);
-               return -1;
-       }
-
-       ret = init_tx_channels(eth_dev, 1);
-       if (ret != 0) {
-               close_tx_channels(eth_dev);
-               close_rx_channels(eth_dev);
-               return -1;
+       /*
+        * Iterate through all combosixX directories.
+        * When the value in /sys/class/combo/combosixX/device/pcislot
+        * file is the location of the ethernet device dev, "X" is the
+        * index of the device.
+        */
+       while ((entry = readdir(dir)) != NULL) {
+               ret = sscanf(entry->d_name, "combosix%u", &tmp_index);
+               if (ret != 1)
+                       continue;
+
+               snprintf(pcislot_path, PATH_MAX,
+                       "/sys/class/combo/combosix%u/device/pcislot",
+                       tmp_index);
+
+               fd = fopen(pcislot_path, "r");
+               if (fd == NULL)
+                       continue;
+
+               ret = fscanf(fd, "%4" PRIx16 ":%2" PRIx8 ":%2" PRIx8 ".%" PRIx8,
+                               &domain, &bus, &devid, &function);
+               fclose(fd);
+               if (ret != 4)
+                       continue;
+
+               if (pcislot_addr.domain == domain &&
+                               pcislot_addr.bus == bus &&
+                               pcislot_addr.devid == devid &&
+                               pcislot_addr.function == function) {
+                       *index = tmp_index;
+                       closedir(dir);
+                       return 0;
+               }
        }
 
-       eth_dev->rx_pkt_burst = NULL;
-       eth_dev->tx_pkt_burst = NULL;
-
-       return 0;
+       closedir(dir);
+       return -1;
 }
 
-
 static int
-rte_pmd_szedata2_devinit(const char *name, const char *params)
+rte_szedata2_eth_dev_init(struct rte_eth_dev *dev)
 {
-       unsigned numa_node;
+       struct rte_eth_dev_data *data = dev->data;
+       struct pmd_internals *internals = (struct pmd_internals *)
+               data->dev_private;
+       struct szedata *szedata_temp;
        int ret;
-       struct rte_kvargs *kvlist;
-       unsigned k_idx;
-       struct rte_kvargs_pair *pair = NULL;
-       struct rxtx_szedata2 szedata2 = { 0, 0, 0, 0, NULL };
-       bool dev_path_missing = true;
-
-       RTE_LOG(INFO, PMD, "Initializing pmd_szedata2 for %s\n", name);
-
-       numa_node = rte_socket_id();
+       uint32_t szedata2_index;
+       struct rte_pci_addr *pci_addr = &dev->pci_dev->addr;
+       struct rte_mem_resource *pci_rsc =
+               &dev->pci_dev->mem_resource[PCI_RESOURCE_NUMBER];
+       char rsc_filename[PATH_MAX];
+       void *pci_resource_ptr = NULL;
+       int fd;
+
+       RTE_LOG(INFO, PMD, "Initializing szedata2 device (" PCI_PRI_FMT ")\n",
+                       pci_addr->domain, pci_addr->bus, pci_addr->devid,
+                       pci_addr->function);
+
+       /* Get index of szedata2 device file and create path to device file */
+       ret = get_szedata2_index(dev, &szedata2_index);
+       if (ret != 0) {
+               RTE_LOG(ERR, PMD, "Failed to get szedata2 device index!\n");
+               return -ENODEV;
+       }
+       snprintf(internals->sze_dev, PATH_MAX, SZEDATA2_DEV_PATH_FMT,
+                       szedata2_index);
 
-       kvlist = rte_kvargs_parse(params, valid_arguments);
-       if (kvlist == NULL)
-               return -1;
+       RTE_LOG(INFO, PMD, "SZEDATA2 path: %s\n", internals->sze_dev);
 
        /*
-        * Get szedata2 device path and rx,tx channels from passed arguments.
+        * Get number of available DMA RX and TX channels, which is maximum
+        * number of queues that can be created and store it in private device
+        * data structure.
         */
+       szedata_temp = szedata_open(internals->sze_dev);
+       if (szedata_temp == NULL) {
+               RTE_LOG(ERR, PMD, "szedata_open(): failed to open %s",
+                               internals->sze_dev);
+               return -EINVAL;
+       }
+       internals->max_rx_queues = szedata_ifaces_available(szedata_temp,
+                       SZE2_DIR_RX);
+       internals->max_tx_queues = szedata_ifaces_available(szedata_temp,
+                       SZE2_DIR_TX);
+       szedata_close(szedata_temp);
+
+       RTE_LOG(INFO, PMD, "Available DMA channels RX: %u TX: %u\n",
+                       internals->max_rx_queues, internals->max_tx_queues);
+
+       /* Set rx, tx burst functions */
+       if (data->dev_conf.rxmode.enable_scatter == 1 ||
+               data->scattered_rx == 1) {
+               dev->rx_pkt_burst = eth_szedata2_rx_scattered;
+               data->scattered_rx = 1;
+       } else {
+               dev->rx_pkt_burst = eth_szedata2_rx;
+               data->scattered_rx = 0;
+       }
+       dev->tx_pkt_burst = eth_szedata2_tx;
 
-       if (rte_kvargs_count(kvlist, RTE_ETH_SZEDATA2_DEV_PATH_ARG) != 1)
-               goto err;
-
-       if (rte_kvargs_count(kvlist, RTE_ETH_SZEDATA2_RX_IFACES_ARG) < 1)
-               goto err;
+       /* Set function callbacks for Ethernet API */
+       dev->dev_ops = &ops;
 
-       if (rte_kvargs_count(kvlist, RTE_ETH_SZEDATA2_TX_IFACES_ARG) < 1)
-               goto err;
+       rte_eth_copy_pci_info(dev, dev->pci_dev);
 
-       for (k_idx = 0; k_idx < kvlist->count; k_idx++) {
-               pair = &kvlist->pairs[k_idx];
-               if (strstr(pair->key, RTE_ETH_SZEDATA2_DEV_PATH_ARG) != NULL) {
-                       szedata2.sze_dev = pair->value;
-                       dev_path_missing = false;
-                       break;
-               }
+       /* mmap pci resource0 file to rte_mem_resource structure */
+       if (dev->pci_dev->mem_resource[PCI_RESOURCE_NUMBER].phys_addr ==
+                       0) {
+               RTE_LOG(ERR, PMD, "Missing resource%u file\n",
+                               PCI_RESOURCE_NUMBER);
+               return -EINVAL;
+       }
+       snprintf(rsc_filename, PATH_MAX,
+               "%s/" PCI_PRI_FMT "/resource%u", pci_get_sysfs_path(),
+               pci_addr->domain, pci_addr->bus,
+               pci_addr->devid, pci_addr->function, PCI_RESOURCE_NUMBER);
+       fd = open(rsc_filename, O_RDWR);
+       if (fd < 0) {
+               RTE_LOG(ERR, PMD, "Could not open file %s\n", rsc_filename);
+               return -EINVAL;
        }
 
-       if (dev_path_missing)
-               goto err;
-
-       ret = rte_kvargs_process(kvlist, RTE_ETH_SZEDATA2_RX_IFACES_ARG,
-                       &add_rx_mask, &szedata2);
-       if (ret < 0)
-               goto err;
+       pci_resource_ptr = mmap(0,
+                       dev->pci_dev->mem_resource[PCI_RESOURCE_NUMBER].len,
+                       PROT_READ | PROT_WRITE, MAP_SHARED, fd, 0);
+       close(fd);
+       if (pci_resource_ptr == NULL) {
+               RTE_LOG(ERR, PMD, "Could not mmap file %s (fd = %d)\n",
+                               rsc_filename, fd);
+               return -EINVAL;
+       }
+       dev->pci_dev->mem_resource[PCI_RESOURCE_NUMBER].addr =
+               pci_resource_ptr;
+
+       RTE_LOG(DEBUG, PMD, "resource%u phys_addr = 0x%llx len = %llu "
+                       "virt addr = %llx\n", PCI_RESOURCE_NUMBER,
+                       (unsigned long long)pci_rsc->phys_addr,
+                       (unsigned long long)pci_rsc->len,
+                       (unsigned long long)pci_rsc->addr);
+
+       /* Get link state */
+       eth_link_update(dev, 0);
+
+       /* Allocate space for one mac address */
+       data->mac_addrs = rte_zmalloc(data->name, sizeof(struct ether_addr),
+                       RTE_CACHE_LINE_SIZE);
+       if (data->mac_addrs == NULL) {
+               RTE_LOG(ERR, PMD, "Could not alloc space for MAC address!\n");
+               munmap(dev->pci_dev->mem_resource[PCI_RESOURCE_NUMBER].addr,
+                       dev->pci_dev->mem_resource[PCI_RESOURCE_NUMBER].len);
+               return -EINVAL;
+       }
 
-       ret = rte_kvargs_process(kvlist, RTE_ETH_SZEDATA2_TX_IFACES_ARG,
-                       &add_tx_mask, &szedata2);
-       if (ret < 0)
-               goto err;
+       ether_addr_copy(&eth_addr, data->mac_addrs);
 
-       szedata2.num_of_rx = count_ones(szedata2.sze_rx_mask_req);
-       szedata2.num_of_tx = count_ones(szedata2.sze_tx_mask_req);
+       /* At initial state COMBO card is in promiscuous mode so disable it */
+       eth_promiscuous_disable(dev);
 
-       RTE_LOG(INFO, PMD, "SZE device found at path %s\n", szedata2.sze_dev);
+       RTE_LOG(INFO, PMD, "szedata2 device ("
+                       PCI_PRI_FMT ") successfully initialized\n",
+                       pci_addr->domain, pci_addr->bus, pci_addr->devid,
+                       pci_addr->function);
 
-       return rte_eth_from_szedata2(name, &szedata2, numa_node);
-err:
-       rte_kvargs_free(kvlist);
-       return -1;
+       return 0;
 }
 
 static int
-rte_pmd_szedata2_devuninit(const char *name)
+rte_szedata2_eth_dev_uninit(struct rte_eth_dev *dev)
 {
-       struct rte_eth_dev *dev = NULL;
+       struct rte_pci_addr *pci_addr = &dev->pci_dev->addr;
 
-       RTE_LOG(INFO, PMD, "Uninitializing pmd_szedata2 for %s "
-                       "on numa socket %u\n", name, rte_socket_id());
-
-       if (name == NULL)
-               return -1;
+       rte_free(dev->data->mac_addrs);
+       dev->data->mac_addrs = NULL;
+       munmap(dev->pci_dev->mem_resource[PCI_RESOURCE_NUMBER].addr,
+               dev->pci_dev->mem_resource[PCI_RESOURCE_NUMBER].len);
 
-       dev = rte_eth_dev_allocated(name);
-       if (dev == NULL)
-               return -1;
+       RTE_LOG(INFO, PMD, "szedata2 device ("
+                       PCI_PRI_FMT ") successfully uninitialized\n",
+                       pci_addr->domain, pci_addr->bus, pci_addr->devid,
+                       pci_addr->function);
 
-       rte_free(dev->data->dev_private);
-       rte_free(dev->data);
-       rte_eth_dev_release_port(dev);
        return 0;
 }
 
-static struct rte_driver pmd_szedata2_drv = {
-       .name = "eth_szedata2",
-       .type = PMD_VDEV,
-       .init = rte_pmd_szedata2_devinit,
-       .uninit = rte_pmd_szedata2_devuninit,
+static const struct rte_pci_id rte_szedata2_pci_id_table[] = {
+       {
+               RTE_PCI_DEVICE(PCI_VENDOR_ID_NETCOPE,
+                               PCI_DEVICE_ID_NETCOPE_COMBO80G)
+       },
+       {
+               RTE_PCI_DEVICE(PCI_VENDOR_ID_NETCOPE,
+                               PCI_DEVICE_ID_NETCOPE_COMBO100G)
+       },
+       {
+               RTE_PCI_DEVICE(PCI_VENDOR_ID_NETCOPE,
+                               PCI_DEVICE_ID_NETCOPE_COMBO100G2)
+       },
+       {
+               .vendor_id = 0,
+       }
+};
+
+static struct eth_driver szedata2_eth_driver = {
+       .pci_drv = {
+               .id_table = rte_szedata2_pci_id_table,
+               .probe = rte_eth_dev_pci_probe,
+               .remove = rte_eth_dev_pci_remove,
+       },
+       .eth_dev_init     = rte_szedata2_eth_dev_init,
+       .eth_dev_uninit   = rte_szedata2_eth_dev_uninit,
+       .dev_private_size = sizeof(struct pmd_internals),
 };
 
-PMD_REGISTER_DRIVER(pmd_szedata2_drv);
+RTE_PMD_REGISTER_PCI(RTE_SZEDATA2_DRIVER_NAME, szedata2_eth_driver.pci_drv);
+RTE_PMD_REGISTER_PCI_TABLE(RTE_SZEDATA2_DRIVER_NAME, rte_szedata2_pci_id_table);
+RTE_PMD_REGISTER_KMOD_DEP(RTE_SZEDATA2_DRIVER_NAME,
+       "* combo6core & combov3 & szedata2 & szedata2_cv3");