net: add rte prefix to ether structures
[dpdk.git] / drivers / net / bonding / rte_eth_bond_pmd.c
index 1320cfd..9950635 100644 (file)
 static uint64_t tlb_last_obytets[RTE_MAX_ETHPORTS];
 
 static inline size_t
-get_vlan_offset(struct ether_hdr *eth_hdr, uint16_t *proto)
+get_vlan_offset(struct rte_ether_hdr *eth_hdr, uint16_t *proto)
 {
        size_t vlan_offset = 0;
 
-       if (rte_cpu_to_be_16(ETHER_TYPE_VLAN) == *proto) {
-               struct vlan_hdr *vlan_hdr = (struct vlan_hdr *)(eth_hdr + 1);
+       if (rte_cpu_to_be_16(ETHER_TYPE_VLAN) == *proto ||
+               rte_cpu_to_be_16(ETHER_TYPE_QINQ) == *proto) {
+               struct rte_vlan_hdr *vlan_hdr =
+                       (struct rte_vlan_hdr *)(eth_hdr + 1);
 
-               vlan_offset = sizeof(struct vlan_hdr);
+               vlan_offset = sizeof(struct rte_vlan_hdr);
                *proto = vlan_hdr->eth_proto;
 
                if (rte_cpu_to_be_16(ETHER_TYPE_VLAN) == *proto) {
                        vlan_hdr = vlan_hdr + 1;
                        *proto = vlan_hdr->eth_proto;
-                       vlan_offset += sizeof(struct vlan_hdr);
+                       vlan_offset += sizeof(struct rte_vlan_hdr);
                }
        }
        return vlan_offset;
@@ -57,28 +59,34 @@ bond_ethdev_rx_burst(void *queue, struct rte_mbuf **bufs, uint16_t nb_pkts)
 {
        struct bond_dev_private *internals;
 
-       uint16_t num_rx_slave = 0;
        uint16_t num_rx_total = 0;
-
+       uint16_t slave_count;
+       uint16_t active_slave;
        int i;
 
        /* Cast to structure, containing bonded device's port id and queue id */
        struct bond_rx_queue *bd_rx_q = (struct bond_rx_queue *)queue;
-
        internals = bd_rx_q->dev_private;
+       slave_count = internals->active_slave_count;
+       active_slave = internals->active_slave;
 
+       for (i = 0; i < slave_count && nb_pkts; i++) {
+               uint16_t num_rx_slave;
 
-       for (i = 0; i < internals->active_slave_count && nb_pkts; i++) {
                /* Offset of pointer to *bufs increases as packets are received
                 * from other slaves */
-               num_rx_slave = rte_eth_rx_burst(internals->active_slaves[i],
-                               bd_rx_q->queue_id, bufs + num_rx_total, nb_pkts);
-               if (num_rx_slave) {
-                       num_rx_total += num_rx_slave;
-                       nb_pkts -= num_rx_slave;
-               }
+               num_rx_slave =
+                       rte_eth_rx_burst(internals->active_slaves[active_slave],
+                                        bd_rx_q->queue_id,
+                                        bufs + num_rx_total, nb_pkts);
+               num_rx_total += num_rx_slave;
+               nb_pkts -= num_rx_slave;
+               if (++active_slave == slave_count)
+                       active_slave = 0;
        }
 
+       if (++internals->active_slave >= slave_count)
+               internals->active_slave = 0;
        return num_rx_total;
 }
 
@@ -257,25 +265,32 @@ bond_ethdev_rx_burst_8023ad_fast_queue(void *queue, struct rte_mbuf **bufs,
        uint16_t num_rx_total = 0;      /* Total number of received packets */
        uint16_t slaves[RTE_MAX_ETHPORTS];
        uint16_t slave_count;
-
-       uint16_t i, idx;
+       uint16_t active_slave;
+       uint16_t i;
 
        /* Copy slave list to protect against slave up/down changes during tx
         * bursting */
        slave_count = internals->active_slave_count;
+       active_slave = internals->active_slave;
        memcpy(slaves, internals->active_slaves,
                        sizeof(internals->active_slaves[0]) * slave_count);
 
-       for (i = 0, idx = internals->active_slave;
-                       i < slave_count && num_rx_total < nb_pkts; i++, idx++) {
-               idx = idx % slave_count;
+       for (i = 0; i < slave_count && nb_pkts; i++) {
+               uint16_t num_rx_slave;
 
                /* Read packets from this slave */
-               num_rx_total += rte_eth_rx_burst(slaves[idx], bd_rx_q->queue_id,
-                               &bufs[num_rx_total], nb_pkts - num_rx_total);
+               num_rx_slave = rte_eth_rx_burst(slaves[active_slave],
+                                               bd_rx_q->queue_id,
+                                               bufs + num_rx_total, nb_pkts);
+               num_rx_total += num_rx_slave;
+               nb_pkts -= num_rx_slave;
+
+               if (++active_slave == slave_count)
+                       active_slave = 0;
        }
 
-       internals->active_slave = idx;
+       if (++internals->active_slave >= slave_count)
+               internals->active_slave = 0;
 
        return num_rx_total;
 }
@@ -300,10 +315,10 @@ bond_ethdev_tx_burst_8023ad_fast_queue(void *queue, struct rte_mbuf **bufs,
        /* Mapping array generated by hash function to map mbufs to slaves */
        uint16_t bufs_slave_port_idxs[RTE_MAX_ETHPORTS] = { 0 };
 
-       uint16_t slave_tx_count, slave_tx_fail_count[RTE_MAX_ETHPORTS] = { 0 };
+       uint16_t slave_tx_count;
        uint16_t total_tx_count = 0, total_tx_fail_count = 0;
 
-       uint16_t i, j;
+       uint16_t i;
 
        if (unlikely(nb_bufs == 0))
                return 0;
@@ -320,7 +335,7 @@ bond_ethdev_tx_burst_8023ad_fast_queue(void *queue, struct rte_mbuf **bufs,
 
        dist_slave_count = 0;
        for (i = 0; i < slave_count; i++) {
-               struct port *port = &mode_8023ad_ports[slave_port_ids[i]];
+               struct port *port = &bond_mode_8023ad_ports[slave_port_ids[i]];
 
                if (ACTOR_STATE(port, DISTRIBUTING))
                        dist_slave_port_ids[dist_slave_count++] =
@@ -339,7 +354,7 @@ bond_ethdev_tx_burst_8023ad_fast_queue(void *queue, struct rte_mbuf **bufs,
 
        for (i = 0; i < nb_bufs; i++) {
                /* Populate slave mbuf arrays with mbufs for that slave. */
-               uint8_t slave_idx = bufs_slave_port_idxs[i];
+               uint16_t slave_idx = bufs_slave_port_idxs[i];
 
                slave_bufs[slave_idx][slave_nb_bufs[slave_idx]++] = bufs[i];
        }
@@ -358,34 +373,12 @@ bond_ethdev_tx_burst_8023ad_fast_queue(void *queue, struct rte_mbuf **bufs,
 
                /* If tx burst fails move packets to end of bufs */
                if (unlikely(slave_tx_count < slave_nb_bufs[i])) {
-                       slave_tx_fail_count[i] = slave_nb_bufs[i] -
+                       int slave_tx_fail_count = slave_nb_bufs[i] -
                                        slave_tx_count;
-                       total_tx_fail_count += slave_tx_fail_count[i];
-
-                       /*
-                        * Shift bufs to beginning of array to allow reordering
-                        * later
-                        */
-                       for (j = 0; j < slave_tx_fail_count[i]; j++) {
-                               slave_bufs[i][j] =
-                                       slave_bufs[i][(slave_tx_count - 1) + j];
-                       }
-               }
-       }
-
-       /*
-        * If there are tx burst failures we move packets to end of bufs to
-        * preserve expected PMD behaviour of all failed transmitted being
-        * at the end of the input mbuf array
-        */
-       if (unlikely(total_tx_fail_count > 0)) {
-               int bufs_idx = nb_bufs - total_tx_fail_count - 1;
-
-               for (i = 0; i < slave_count; i++) {
-                       if (slave_tx_fail_count[i] > 0) {
-                               for (j = 0; j < slave_tx_fail_count[i]; j++)
-                                       bufs[bufs_idx++] = slave_bufs[i][j];
-                       }
+                       total_tx_fail_count += slave_tx_fail_count;
+                       memcpy(&bufs[nb_bufs - total_tx_fail_count],
+                              &slave_bufs[i][slave_tx_count],
+                              slave_tx_fail_count * sizeof(bufs[0]));
                }
        }
 
@@ -400,9 +393,10 @@ bond_ethdev_rx_burst_8023ad(void *queue, struct rte_mbuf **bufs,
        /* Cast to structure, containing bonded device's port id and queue id */
        struct bond_rx_queue *bd_rx_q = (struct bond_rx_queue *)queue;
        struct bond_dev_private *internals = bd_rx_q->dev_private;
-       struct ether_addr bond_mac;
-
-       struct ether_hdr *hdr;
+       struct rte_eth_dev *bonded_eth_dev =
+                                       &rte_eth_devices[internals->port_id];
+       struct rte_ether_addr *bond_mac = bonded_eth_dev->data->mac_addrs;
+       struct rte_ether_hdr *hdr;
 
        const uint16_t ether_type_slow_be = rte_be_to_cpu_16(ETHER_TYPE_SLOW);
        uint16_t num_rx_total = 0;      /* Total number of received packets */
@@ -411,10 +405,11 @@ bond_ethdev_rx_burst_8023ad(void *queue, struct rte_mbuf **bufs,
 
        uint8_t collecting;  /* current slave collecting status */
        const uint8_t promisc = internals->promiscuous_en;
-       uint8_t i, j, k;
        uint8_t subtype;
+       uint16_t i;
+       uint16_t j;
+       uint16_t k;
 
-       rte_eth_macaddr_get(internals->port_id, &bond_mac);
        /* Copy slave list to protect against slave up/down changes during tx
         * bursting */
        slave_count = internals->active_slave_count;
@@ -428,7 +423,7 @@ bond_ethdev_rx_burst_8023ad(void *queue, struct rte_mbuf **bufs,
        }
        for (i = 0; i < slave_count && num_rx_total < nb_pkts; i++) {
                j = num_rx_total;
-               collecting = ACTOR_STATE(&mode_8023ad_ports[slaves[idx]],
+               collecting = ACTOR_STATE(&bond_mode_8023ad_ports[slaves[idx]],
                                         COLLECTING);
 
                /* Read packets from this slave */
@@ -450,16 +445,18 @@ bond_ethdev_rx_burst_8023ad(void *queue, struct rte_mbuf **bufs,
                        if (j + 3 < num_rx_total)
                                rte_prefetch0(rte_pktmbuf_mtod(bufs[j + 3], void *));
 
-                       hdr = rte_pktmbuf_mtod(bufs[j], struct ether_hdr *);
+                       hdr = rte_pktmbuf_mtod(bufs[j], struct rte_ether_hdr *);
                        subtype = ((struct slow_protocol_frame *)hdr)->slow_protocol.subtype;
 
                        /* Remove packet from array if it is slow packet or slave is not
                         * in collecting state or bonding interface is not in promiscuous
                         * mode and packet address does not match. */
                        if (unlikely(is_lacp_packets(hdr->ether_type, subtype, bufs[j]) ||
-                               !collecting || (!promisc &&
-                                       !is_multicast_ether_addr(&hdr->d_addr) &&
-                                       !is_same_ether_addr(&bond_mac, &hdr->d_addr)))) {
+                               !collecting ||
+                               (!promisc &&
+                                !is_multicast_ether_addr(&hdr->d_addr) &&
+                                !is_same_ether_addr(bond_mac,
+                                                    &hdr->d_addr)))) {
 
                                if (hdr->ether_type == ether_type_slow_be) {
                                        bond_mode_8023ad_handle_slow_pkt(
@@ -480,7 +477,9 @@ bond_ethdev_rx_burst_8023ad(void *queue, struct rte_mbuf **bufs,
                        idx = 0;
        }
 
-       internals->active_slave = idx;
+       if (++internals->active_slave >= slave_count)
+               internals->active_slave = 0;
+
        return num_rx_total;
 }
 
@@ -491,35 +490,31 @@ uint32_t burstnumberTX;
 #ifdef RTE_LIBRTE_BOND_DEBUG_ALB
 
 static void
-arp_op_name(uint16_t arp_op, char *buf)
+arp_op_name(uint16_t arp_op, char *buf, size_t buf_len)
 {
        switch (arp_op) {
-       case ARP_OP_REQUEST:
-               snprintf(buf, sizeof("ARP Request"), "%s", "ARP Request");
+       case RTE_ARP_OP_REQUEST:
+               strlcpy(buf, "ARP Request", buf_len);
                return;
-       case ARP_OP_REPLY:
-               snprintf(buf, sizeof("ARP Reply"), "%s", "ARP Reply");
+       case RTE_ARP_OP_REPLY:
+               strlcpy(buf, "ARP Reply", buf_len);
                return;
-       case ARP_OP_REVREQUEST:
-               snprintf(buf, sizeof("Reverse ARP Request"), "%s",
-                               "Reverse ARP Request");
+       case RTE_ARP_OP_REVREQUEST:
+               strlcpy(buf, "Reverse ARP Request", buf_len);
                return;
-       case ARP_OP_REVREPLY:
-               snprintf(buf, sizeof("Reverse ARP Reply"), "%s",
-                               "Reverse ARP Reply");
+       case RTE_ARP_OP_REVREPLY:
+               strlcpy(buf, "Reverse ARP Reply", buf_len);
                return;
-       case ARP_OP_INVREQUEST:
-               snprintf(buf, sizeof("Peer Identify Request"), "%s",
-                               "Peer Identify Request");
+       case RTE_ARP_OP_INVREQUEST:
+               strlcpy(buf, "Peer Identify Request", buf_len);
                return;
-       case ARP_OP_INVREPLY:
-               snprintf(buf, sizeof("Peer Identify Reply"), "%s",
-                               "Peer Identify Reply");
+       case RTE_ARP_OP_INVREPLY:
+               strlcpy(buf, "Peer Identify Reply", buf_len);
                return;
        default:
                break;
        }
-       snprintf(buf, sizeof("Unknown"), "%s", "Unknown");
+       strlcpy(buf, "Unknown", buf_len);
        return;
 }
 #endif
@@ -590,12 +585,13 @@ update_client_stats(uint32_t addr, uint16_t port, uint32_t *TXorRXindicator)
 #endif
 
 static void
-mode6_debug(const char __attribute__((unused)) *info, struct ether_hdr *eth_h,
-               uint16_t port, uint32_t __attribute__((unused)) *burstnumber)
+mode6_debug(const char __attribute__((unused)) *info,
+       struct rte_ether_hdr *eth_h, uint16_t port,
+       uint32_t __attribute__((unused)) *burstnumber)
 {
        struct ipv4_hdr *ipv4_h;
 #ifdef RTE_LIBRTE_BOND_DEBUG_ALB
-       struct arp_hdr *arp_h;
+       struct rte_arp_hdr *arp_h;
        char dst_ip[16];
        char ArpOp[24];
        char buf[16];
@@ -620,10 +616,11 @@ mode6_debug(const char __attribute__((unused)) *info, struct ether_hdr *eth_h,
        }
 #ifdef RTE_LIBRTE_BOND_DEBUG_ALB
        else if (ether_type == rte_cpu_to_be_16(ETHER_TYPE_ARP)) {
-               arp_h = (struct arp_hdr *)((char *)(eth_h + 1) + offset);
+               arp_h = (struct rte_arp_hdr *)((char *)(eth_h + 1) + offset);
                ipv4_addr_to_dot(arp_h->arp_data.arp_sip, src_ip, MaxIPv4String);
                ipv4_addr_to_dot(arp_h->arp_data.arp_tip, dst_ip, MaxIPv4String);
-               arp_op_name(rte_be_to_cpu_16(arp_h->arp_op), ArpOp);
+               arp_op_name(rte_be_to_cpu_16(arp_h->arp_opcode),
+                               ArpOp, sizeof(ArpOp));
                MODE6_DEBUG(buf, src_ip, dst_ip, eth_h, ArpOp, port, *burstnumber);
        }
 #endif
@@ -635,7 +632,7 @@ bond_ethdev_rx_burst_alb(void *queue, struct rte_mbuf **bufs, uint16_t nb_pkts)
 {
        struct bond_tx_queue *bd_tx_q = (struct bond_tx_queue *)queue;
        struct bond_dev_private *internals = bd_tx_q->dev_private;
-       struct ether_hdr *eth_h;
+       struct rte_ether_hdr *eth_h;
        uint16_t ether_type, offset;
        uint16_t nb_recv_pkts;
        int i;
@@ -643,7 +640,7 @@ bond_ethdev_rx_burst_alb(void *queue, struct rte_mbuf **bufs, uint16_t nb_pkts)
        nb_recv_pkts = bond_ethdev_rx_burst(queue, bufs, nb_pkts);
 
        for (i = 0; i < nb_recv_pkts; i++) {
-               eth_h = rte_pktmbuf_mtod(bufs[i], struct ether_hdr *);
+               eth_h = rte_pktmbuf_mtod(bufs[i], struct rte_ether_hdr *);
                ether_type = eth_h->ether_type;
                offset = get_vlan_offset(eth_h, &ether_type);
 
@@ -715,8 +712,8 @@ bond_ethdev_tx_burst_round_robin(void *queue, struct rte_mbuf **bufs,
                                tx_fail_total += tx_fail_slave;
 
                                memcpy(&bufs[nb_pkts - tx_fail_total],
-                                               &slave_bufs[i][num_tx_slave],
-                                               tx_fail_slave * sizeof(bufs[0]));
+                                      &slave_bufs[i][num_tx_slave],
+                                      tx_fail_slave * sizeof(bufs[0]));
                        }
                        num_tx_total += num_tx_slave;
                }
@@ -743,7 +740,7 @@ bond_ethdev_tx_burst_active_backup(void *queue,
 }
 
 static inline uint16_t
-ether_hash(struct ether_hdr *eth_hdr)
+ether_hash(struct rte_ether_hdr *eth_hdr)
 {
        unaligned_uint16_t *word_src_addr =
                (unaligned_uint16_t *)eth_hdr->s_addr.addr_bytes;
@@ -778,14 +775,14 @@ ipv6_hash(struct ipv6_hdr *ipv6_hdr)
 
 void
 burst_xmit_l2_hash(struct rte_mbuf **buf, uint16_t nb_pkts,
-               uint8_t slave_count, uint16_t *slaves)
+               uint16_t slave_count, uint16_t *slaves)
 {
-       struct ether_hdr *eth_hdr;
+       struct rte_ether_hdr *eth_hdr;
        uint32_t hash;
        int i;
 
        for (i = 0; i < nb_pkts; i++) {
-               eth_hdr = rte_pktmbuf_mtod(buf[i], struct ether_hdr *);
+               eth_hdr = rte_pktmbuf_mtod(buf[i], struct rte_ether_hdr *);
 
                hash = ether_hash(eth_hdr);
 
@@ -795,16 +792,16 @@ burst_xmit_l2_hash(struct rte_mbuf **buf, uint16_t nb_pkts,
 
 void
 burst_xmit_l23_hash(struct rte_mbuf **buf, uint16_t nb_pkts,
-               uint8_t slave_count, uint16_t *slaves)
+               uint16_t slave_count, uint16_t *slaves)
 {
        uint16_t i;
-       struct ether_hdr *eth_hdr;
+       struct rte_ether_hdr *eth_hdr;
        uint16_t proto;
        size_t vlan_offset;
        uint32_t hash, l3hash;
 
        for (i = 0; i < nb_pkts; i++) {
-               eth_hdr = rte_pktmbuf_mtod(buf[i], struct ether_hdr *);
+               eth_hdr = rte_pktmbuf_mtod(buf[i], struct rte_ether_hdr *);
                l3hash = 0;
 
                proto = eth_hdr->ether_type;
@@ -833,9 +830,9 @@ burst_xmit_l23_hash(struct rte_mbuf **buf, uint16_t nb_pkts,
 
 void
 burst_xmit_l34_hash(struct rte_mbuf **buf, uint16_t nb_pkts,
-               uint8_t slave_count, uint16_t *slaves)
+               uint16_t slave_count, uint16_t *slaves)
 {
-       struct ether_hdr *eth_hdr;
+       struct rte_ether_hdr *eth_hdr;
        uint16_t proto;
        size_t vlan_offset;
        int i;
@@ -845,7 +842,8 @@ burst_xmit_l34_hash(struct rte_mbuf **buf, uint16_t nb_pkts,
        uint32_t hash, l3hash, l4hash;
 
        for (i = 0; i < nb_pkts; i++) {
-               eth_hdr = rte_pktmbuf_mtod(buf[i], struct ether_hdr *);
+               eth_hdr = rte_pktmbuf_mtod(buf[i], struct rte_ether_hdr *);
+               size_t pkt_end = (size_t)eth_hdr + rte_pktmbuf_data_len(buf[i]);
                proto = eth_hdr->ether_type;
                vlan_offset = get_vlan_offset(eth_hdr, &proto);
                l3hash = 0;
@@ -869,13 +867,17 @@ burst_xmit_l34_hash(struct rte_mbuf **buf, uint16_t nb_pkts,
                                        tcp_hdr = (struct tcp_hdr *)
                                                ((char *)ipv4_hdr +
                                                        ip_hdr_offset);
-                                       l4hash = HASH_L4_PORTS(tcp_hdr);
+                                       if ((size_t)tcp_hdr + sizeof(*tcp_hdr)
+                                                       < pkt_end)
+                                               l4hash = HASH_L4_PORTS(tcp_hdr);
                                } else if (ipv4_hdr->next_proto_id ==
                                                                IPPROTO_UDP) {
                                        udp_hdr = (struct udp_hdr *)
                                                ((char *)ipv4_hdr +
                                                        ip_hdr_offset);
-                                       l4hash = HASH_L4_PORTS(udp_hdr);
+                                       if ((size_t)udp_hdr + sizeof(*udp_hdr)
+                                                       < pkt_end)
+                                               l4hash = HASH_L4_PORTS(udp_hdr);
                                }
                        }
                } else if  (rte_cpu_to_be_16(ETHER_TYPE_IPv6) == proto) {
@@ -903,7 +905,7 @@ burst_xmit_l34_hash(struct rte_mbuf **buf, uint16_t nb_pkts,
 struct bwg_slave {
        uint64_t bwg_left_int;
        uint64_t bwg_left_remainder;
-       uint8_t slave;
+       uint16_t slave;
 };
 
 void
@@ -956,11 +958,12 @@ bond_ethdev_update_tlb_slave_cb(void *arg)
        struct bond_dev_private *internals = arg;
        struct rte_eth_stats slave_stats;
        struct bwg_slave bwg_array[RTE_MAX_ETHPORTS];
-       uint8_t slave_count;
+       uint16_t slave_count;
        uint64_t tx_bytes;
 
        uint8_t update_stats = 0;
-       uint8_t i, slave_id;
+       uint16_t slave_id;
+       uint16_t i;
 
        internals->slave_update_idx++;
 
@@ -1007,9 +1010,9 @@ bond_ethdev_tx_burst_tlb(void *queue, struct rte_mbuf **bufs, uint16_t nb_pkts)
        uint16_t num_of_slaves = internals->active_slave_count;
        uint16_t slaves[RTE_MAX_ETHPORTS];
 
-       struct ether_hdr *ether_hdr;
-       struct ether_addr primary_slave_addr;
-       struct ether_addr active_slave_addr;
+       struct rte_ether_hdr *ether_hdr;
+       struct rte_ether_addr primary_slave_addr;
+       struct rte_ether_addr active_slave_addr;
 
        if (num_of_slaves < 1)
                return num_tx_total;
@@ -1031,7 +1034,8 @@ bond_ethdev_tx_burst_tlb(void *queue, struct rte_mbuf **bufs, uint16_t nb_pkts)
                        if (j + 3 < nb_pkts)
                                rte_prefetch0(rte_pktmbuf_mtod(bufs[j+3], void*));
 
-                       ether_hdr = rte_pktmbuf_mtod(bufs[j], struct ether_hdr *);
+                       ether_hdr = rte_pktmbuf_mtod(bufs[j],
+                                               struct rte_ether_hdr *);
                        if (is_same_ether_addr(&ether_hdr->s_addr, &primary_slave_addr))
                                ether_addr_copy(&active_slave_addr, &ether_hdr->s_addr);
 #if defined(RTE_LIBRTE_BOND_DEBUG_ALB) || defined(RTE_LIBRTE_BOND_DEBUG_ALB_L1)
@@ -1067,7 +1071,7 @@ bond_ethdev_tx_burst_alb(void *queue, struct rte_mbuf **bufs, uint16_t nb_pkts)
        struct bond_tx_queue *bd_tx_q = (struct bond_tx_queue *)queue;
        struct bond_dev_private *internals = bd_tx_q->dev_private;
 
-       struct ether_hdr *eth_h;
+       struct rte_ether_hdr *eth_h;
        uint16_t ether_type, offset;
 
        struct client_data *client_info;
@@ -1097,7 +1101,7 @@ bond_ethdev_tx_burst_alb(void *queue, struct rte_mbuf **bufs, uint16_t nb_pkts)
 
        /* Search tx buffer for ARP packets and forward them to alb */
        for (i = 0; i < nb_pkts; i++) {
-               eth_h = rte_pktmbuf_mtod(bufs[i], struct ether_hdr *);
+               eth_h = rte_pktmbuf_mtod(bufs[i], struct rte_ether_hdr *);
                ether_type = eth_h->ether_type;
                offset = get_vlan_offset(eth_h, &ether_type);
 
@@ -1131,8 +1135,10 @@ bond_ethdev_tx_burst_alb(void *queue, struct rte_mbuf **bufs, uint16_t nb_pkts)
                                                     "Failed to allocate ARP packet from pool");
                                        continue;
                                }
-                               pkt_size = sizeof(struct ether_hdr) + sizeof(struct arp_hdr)
-                                               + client_info->vlan_count * sizeof(struct vlan_hdr);
+                               pkt_size = sizeof(struct rte_ether_hdr) +
+                                       sizeof(struct rte_arp_hdr) +
+                                       client_info->vlan_count *
+                                       sizeof(struct rte_vlan_hdr);
                                upd_pkt->data_len = pkt_size;
                                upd_pkt->pkt_len = pkt_size;
 
@@ -1163,7 +1169,8 @@ bond_ethdev_tx_burst_alb(void *queue, struct rte_mbuf **bufs, uint16_t nb_pkts)
 #if defined(RTE_LIBRTE_BOND_DEBUG_ALB) || defined(RTE_LIBRTE_BOND_DEBUG_ALB_L1)
        /* Print TX stats including update packets */
                        for (j = 0; j < slave_bufs_pkts[i]; j++) {
-                               eth_h = rte_pktmbuf_mtod(slave_bufs[i][j], struct ether_hdr *);
+                               eth_h = rte_pktmbuf_mtod(slave_bufs[i][j],
+                                                       struct rte_ether_hdr *);
                                mode6_debug("TX ARP:", eth_h, i, &burstnumberTX);
                        }
 #endif
@@ -1180,7 +1187,8 @@ bond_ethdev_tx_burst_alb(void *queue, struct rte_mbuf **bufs, uint16_t nb_pkts)
                        }
 #if defined(RTE_LIBRTE_BOND_DEBUG_ALB) || defined(RTE_LIBRTE_BOND_DEBUG_ALB_L1)
                        for (j = 0; j < update_bufs_pkts[i]; j++) {
-                               eth_h = rte_pktmbuf_mtod(update_bufs[i][j], struct ether_hdr *);
+                               eth_h = rte_pktmbuf_mtod(update_bufs[i][j],
+                                                       struct rte_ether_hdr *);
                                mode6_debug("TX ARPupd:", eth_h, i, &burstnumberTX);
                        }
 #endif
@@ -1221,10 +1229,10 @@ bond_ethdev_tx_burst_balance(void *queue, struct rte_mbuf **bufs,
        /* Mapping array generated by hash function to map mbufs to slaves */
        uint16_t bufs_slave_port_idxs[nb_bufs];
 
-       uint16_t slave_tx_count, slave_tx_fail_count[RTE_MAX_ETHPORTS] = { 0 };
+       uint16_t slave_tx_count;
        uint16_t total_tx_count = 0, total_tx_fail_count = 0;
 
-       uint16_t i, j;
+       uint16_t i;
 
        if (unlikely(nb_bufs == 0))
                return 0;
@@ -1247,7 +1255,7 @@ bond_ethdev_tx_burst_balance(void *queue, struct rte_mbuf **bufs,
 
        for (i = 0; i < nb_bufs; i++) {
                /* Populate slave mbuf arrays with mbufs for that slave. */
-               uint8_t slave_idx = bufs_slave_port_idxs[i];
+               uint16_t slave_idx = bufs_slave_port_idxs[i];
 
                slave_bufs[slave_idx][slave_nb_bufs[slave_idx]++] = bufs[i];
        }
@@ -1265,34 +1273,12 @@ bond_ethdev_tx_burst_balance(void *queue, struct rte_mbuf **bufs,
 
                /* If tx burst fails move packets to end of bufs */
                if (unlikely(slave_tx_count < slave_nb_bufs[i])) {
-                       slave_tx_fail_count[i] = slave_nb_bufs[i] -
+                       int slave_tx_fail_count = slave_nb_bufs[i] -
                                        slave_tx_count;
-                       total_tx_fail_count += slave_tx_fail_count[i];
-
-                       /*
-                        * Shift bufs to beginning of array to allow reordering
-                        * later
-                        */
-                       for (j = 0; j < slave_tx_fail_count[i]; j++) {
-                               slave_bufs[i][j] =
-                                       slave_bufs[i][(slave_tx_count - 1) + j];
-                       }
-               }
-       }
-
-       /*
-        * If there are tx burst failures we move packets to end of bufs to
-        * preserve expected PMD behaviour of all failed transmitted being
-        * at the end of the input mbuf array
-        */
-       if (unlikely(total_tx_fail_count > 0)) {
-               int bufs_idx = nb_bufs - total_tx_fail_count - 1;
-
-               for (i = 0; i < slave_count; i++) {
-                       if (slave_tx_fail_count[i] > 0) {
-                               for (j = 0; j < slave_tx_fail_count[i]; j++)
-                                       bufs[bufs_idx++] = slave_bufs[i][j];
-                       }
+                       total_tx_fail_count += slave_tx_fail_count;
+                       memcpy(&bufs[nb_bufs - total_tx_fail_count],
+                              &slave_bufs[i][slave_tx_count],
+                              slave_tx_fail_count * sizeof(bufs[0]));
                }
        }
 
@@ -1319,13 +1305,10 @@ bond_ethdev_tx_burst_8023ad(void *queue, struct rte_mbuf **bufs,
        /* Mapping array generated by hash function to map mbufs to slaves */
        uint16_t bufs_slave_port_idxs[RTE_MAX_ETHPORTS] = { 0 };
 
-       uint16_t slave_tx_count, slave_tx_fail_count[RTE_MAX_ETHPORTS] = { 0 };
+       uint16_t slave_tx_count;
        uint16_t total_tx_count = 0, total_tx_fail_count = 0;
 
-       uint16_t i, j;
-
-       if (unlikely(nb_bufs == 0))
-               return 0;
+       uint16_t i;
 
        /* Copy slave list to protect against slave up/down changes during tx
         * bursting */
@@ -1336,16 +1319,40 @@ bond_ethdev_tx_burst_8023ad(void *queue, struct rte_mbuf **bufs,
        memcpy(slave_port_ids, internals->active_slaves,
                        sizeof(slave_port_ids[0]) * slave_count);
 
+       /* Check for LACP control packets and send if available */
+       for (i = 0; i < slave_count; i++) {
+               struct port *port = &bond_mode_8023ad_ports[slave_port_ids[i]];
+               struct rte_mbuf *ctrl_pkt = NULL;
+
+               if (likely(rte_ring_empty(port->tx_ring)))
+                       continue;
+
+               if (rte_ring_dequeue(port->tx_ring,
+                                    (void **)&ctrl_pkt) != -ENOENT) {
+                       slave_tx_count = rte_eth_tx_burst(slave_port_ids[i],
+                                       bd_tx_q->queue_id, &ctrl_pkt, 1);
+                       /*
+                        * re-enqueue LAG control plane packets to buffering
+                        * ring if transmission fails so the packet isn't lost.
+                        */
+                       if (slave_tx_count != 1)
+                               rte_ring_enqueue(port->tx_ring, ctrl_pkt);
+               }
+       }
+
+       if (unlikely(nb_bufs == 0))
+               return 0;
+
        dist_slave_count = 0;
        for (i = 0; i < slave_count; i++) {
-               struct port *port = &mode_8023ad_ports[slave_port_ids[i]];
+               struct port *port = &bond_mode_8023ad_ports[slave_port_ids[i]];
 
                if (ACTOR_STATE(port, DISTRIBUTING))
                        dist_slave_port_ids[dist_slave_count++] =
                                        slave_port_ids[i];
        }
 
-       if (likely(dist_slave_count > 1)) {
+       if (likely(dist_slave_count > 0)) {
 
                /*
                 * Populate slaves mbuf with the packets which are to be sent
@@ -1359,7 +1366,7 @@ bond_ethdev_tx_burst_8023ad(void *queue, struct rte_mbuf **bufs,
                         * Populate slave mbuf arrays with mbufs for that
                         * slave
                         */
-                       uint8_t slave_idx = bufs_slave_port_idxs[i];
+                       uint16_t slave_idx = bufs_slave_port_idxs[i];
 
                        slave_bufs[slave_idx][slave_nb_bufs[slave_idx]++] =
                                        bufs[i];
@@ -1380,64 +1387,17 @@ bond_ethdev_tx_burst_8023ad(void *queue, struct rte_mbuf **bufs,
 
                        /* If tx burst fails move packets to end of bufs */
                        if (unlikely(slave_tx_count < slave_nb_bufs[i])) {
-                               slave_tx_fail_count[i] = slave_nb_bufs[i] -
+                               int slave_tx_fail_count = slave_nb_bufs[i] -
                                                slave_tx_count;
-                               total_tx_fail_count += slave_tx_fail_count[i];
-
-                               /*
-                                * Shift bufs to beginning of array to allow
-                                * reordering later
-                                */
-                               for (j = 0; j < slave_tx_fail_count[i]; j++)
-                                       slave_bufs[i][j] =
-                                               slave_bufs[i]
-                                                       [(slave_tx_count - 1)
-                                                       + j];
-                       }
-               }
+                               total_tx_fail_count += slave_tx_fail_count;
 
-               /*
-                * If there are tx burst failures we move packets to end of
-                * bufs to preserve expected PMD behaviour of all failed
-                * transmitted being at the end of the input mbuf array
-                */
-               if (unlikely(total_tx_fail_count > 0)) {
-                       int bufs_idx = nb_bufs - total_tx_fail_count - 1;
-
-                       for (i = 0; i < slave_count; i++) {
-                               if (slave_tx_fail_count[i] > 0) {
-                                       for (j = 0;
-                                               j < slave_tx_fail_count[i];
-                                               j++) {
-                                               bufs[bufs_idx++] =
-                                                       slave_bufs[i][j];
-                                       }
-                               }
+                               memcpy(&bufs[nb_bufs - total_tx_fail_count],
+                                      &slave_bufs[i][slave_tx_count],
+                                      slave_tx_fail_count * sizeof(bufs[0]));
                        }
                }
        }
 
-       /* Check for LACP control packets and send if available */
-       for (i = 0; i < slave_count; i++) {
-               struct port *port = &mode_8023ad_ports[slave_port_ids[i]];
-               struct rte_mbuf *ctrl_pkt = NULL;
-
-               if (likely(rte_ring_empty(port->tx_ring)))
-                       continue;
-
-               if (rte_ring_dequeue(port->tx_ring,
-                                    (void **)&ctrl_pkt) != -ENOENT) {
-                       slave_tx_count = rte_eth_tx_burst(slave_port_ids[i],
-                                       bd_tx_q->queue_id, &ctrl_pkt, 1);
-                       /*
-                        * re-enqueue LAG control plane packets to buffering
-                        * ring if transmission fails so the packet isn't lost.
-                        */
-                       if (slave_tx_count != 1)
-                               rte_ring_enqueue(port->tx_ring, ctrl_pkt);
-               }
-       }
-
        return total_tx_count;
 }
 
@@ -1448,8 +1408,9 @@ bond_ethdev_tx_burst_broadcast(void *queue, struct rte_mbuf **bufs,
        struct bond_dev_private *internals;
        struct bond_tx_queue *bd_tx_q;
 
-       uint8_t tx_failed_flag = 0, num_of_slaves;
        uint16_t slaves[RTE_MAX_ETHPORTS];
+       uint8_t tx_failed_flag = 0;
+       uint16_t num_of_slaves;
 
        uint16_t max_nb_of_tx_pkts = 0;
 
@@ -1501,7 +1462,7 @@ bond_ethdev_tx_burst_broadcast(void *queue, struct rte_mbuf **bufs,
        return max_nb_of_tx_pkts;
 }
 
-void
+static void
 link_properties_set(struct rte_eth_dev *ethdev, struct rte_eth_link *slave_link)
 {
        struct bond_dev_private *bond_ctx = ethdev->data->dev_private;
@@ -1526,7 +1487,7 @@ link_properties_set(struct rte_eth_dev *ethdev, struct rte_eth_link *slave_link)
        }
 }
 
-int
+static int
 link_properties_valid(struct rte_eth_dev *ethdev,
                struct rte_eth_link *slave_link)
 {
@@ -1545,9 +1506,10 @@ link_properties_valid(struct rte_eth_dev *ethdev,
 }
 
 int
-mac_address_get(struct rte_eth_dev *eth_dev, struct ether_addr *dst_mac_addr)
+mac_address_get(struct rte_eth_dev *eth_dev,
+               struct rte_ether_addr *dst_mac_addr)
 {
-       struct ether_addr *mac_addr;
+       struct rte_ether_addr *mac_addr;
 
        if (eth_dev == NULL) {
                RTE_BOND_LOG(ERR, "NULL pointer eth_dev specified");
@@ -1566,9 +1528,10 @@ mac_address_get(struct rte_eth_dev *eth_dev, struct ether_addr *dst_mac_addr)
 }
 
 int
-mac_address_set(struct rte_eth_dev *eth_dev, struct ether_addr *new_mac_addr)
+mac_address_set(struct rte_eth_dev *eth_dev,
+               struct rte_ether_addr *new_mac_addr)
 {
-       struct ether_addr *mac_addr;
+       struct rte_ether_addr *mac_addr;
 
        if (eth_dev == NULL) {
                RTE_BOND_LOG(ERR, "NULL pointer eth_dev specified");
@@ -1589,7 +1552,7 @@ mac_address_set(struct rte_eth_dev *eth_dev, struct ether_addr *new_mac_addr)
        return 0;
 }
 
-static const struct ether_addr null_mac_addr;
+static const struct rte_ether_addr null_mac_addr;
 
 /*
  * Add additional MAC addresses to the slave
@@ -1599,7 +1562,7 @@ slave_add_mac_addresses(struct rte_eth_dev *bonded_eth_dev,
                uint16_t slave_port_id)
 {
        int i, ret;
-       struct ether_addr *mac_addr;
+       struct rte_ether_addr *mac_addr;
 
        for (i = 1; i < BOND_MAX_MAC_ADDRS; i++) {
                mac_addr = &bonded_eth_dev->data->mac_addrs[i];
@@ -1627,7 +1590,7 @@ slave_remove_mac_addresses(struct rte_eth_dev *bonded_eth_dev,
                uint16_t slave_port_id)
 {
        int i, rc, ret;
-       struct ether_addr *mac_addr;
+       struct rte_ether_addr *mac_addr;
 
        rc = 0;
        for (i = 1; i < BOND_MAX_MAC_ADDRS; i++) {
@@ -1770,7 +1733,7 @@ slave_configure_slow_queue(struct rte_eth_dev *bonded_eth_dev,
        int errval = 0;
        struct bond_dev_private *internals = (struct bond_dev_private *)
                bonded_eth_dev->data->dev_private;
-       struct port *port = &mode_8023ad_ports[slave_eth_dev->data->port_id];
+       struct port *port = &bond_mode_8023ad_ports[slave_eth_dev->data->port_id];
 
        if (port->slow_pool == NULL) {
                char mem_name[256];
@@ -1847,12 +1810,11 @@ slave_configure(struct rte_eth_dev *bonded_eth_dev,
 
        /* If RSS is enabled for bonding, try to enable it for slaves  */
        if (bonded_eth_dev->data->dev_conf.rxmode.mq_mode & ETH_MQ_RX_RSS_FLAG) {
-               if (bonded_eth_dev->data->dev_conf.rx_adv_conf.rss_conf.rss_key_len
-                               != 0) {
+               if (internals->rss_key_len != 0) {
                        slave_eth_dev->data->dev_conf.rx_adv_conf.rss_conf.rss_key_len =
-                                       bonded_eth_dev->data->dev_conf.rx_adv_conf.rss_conf.rss_key_len;
+                                       internals->rss_key_len;
                        slave_eth_dev->data->dev_conf.rx_adv_conf.rss_conf.rss_key =
-                                       bonded_eth_dev->data->dev_conf.rx_adv_conf.rss_conf.rss_key;
+                                       internals->rss_key;
                } else {
                        slave_eth_dev->data->dev_conf.rx_adv_conf.rss_conf.rss_key = NULL;
                }
@@ -2001,7 +1963,7 @@ void
 slave_remove(struct bond_dev_private *internals,
                struct rte_eth_dev *slave_eth_dev)
 {
-       uint8_t i;
+       uint16_t i;
 
        for (i = 0; i < internals->slave_count; i++)
                if (internals->slaves[i].port_id ==
@@ -2051,7 +2013,7 @@ slave_add(struct bond_dev_private *internals,
        slave_details->link_status_wait_to_complete = 0;
        /* clean tlb_last_obytes when adding port for bonding device */
        memcpy(&(slave_details->persisted_mac_addr), slave_eth_dev->data->mac_addrs,
-                       sizeof(struct ether_addr));
+                       sizeof(struct rte_ether_addr));
 }
 
 void
@@ -2097,7 +2059,7 @@ bond_ethdev_start(struct rte_eth_dev *eth_dev)
        }
 
        if (internals->user_defined_mac == 0) {
-               struct ether_addr *new_mac_addr = NULL;
+               struct rte_ether_addr *new_mac_addr = NULL;
 
                for (i = 0; i < internals->slave_count; i++)
                        if (internals->slaves[i].port_id == internals->primary_port)
@@ -2113,10 +2075,6 @@ bond_ethdev_start(struct rte_eth_dev *eth_dev)
                }
        }
 
-       /* Update all slave devices MACs*/
-       if (mac_address_slaves_update(eth_dev) != 0)
-               goto out_err;
-
        /* If bonded device is configure in promiscuous mode then re-apply config */
        if (internals->promiscuous_en)
                bond_ethdev_promiscuous_enable(eth_dev);
@@ -2157,6 +2115,10 @@ bond_ethdev_start(struct rte_eth_dev *eth_dev)
                        (void *)&rte_eth_devices[internals->port_id]);
        }
 
+       /* Update all slave devices MACs*/
+       if (mac_address_slaves_update(eth_dev) != 0)
+               goto out_err;
+
        if (internals->user_defined_primary_port)
                bond_ethdev_primary_set(internals, internals->primary_port);
 
@@ -2177,7 +2139,7 @@ out_err:
 static void
 bond_ethdev_free_queues(struct rte_eth_dev *dev)
 {
-       uint8_t i;
+       uint16_t i;
 
        if (dev->data->rx_queues != NULL) {
                for (i = 0; i < dev->data->nb_rx_queues; i++) {
@@ -2200,7 +2162,7 @@ void
 bond_ethdev_stop(struct rte_eth_dev *eth_dev)
 {
        struct bond_dev_private *internals = eth_dev->data->dev_private;
-       uint8_t i;
+       uint16_t i;
 
        if (internals->mode == BONDING_MODE_8023AD) {
                struct port *port;
@@ -2210,7 +2172,7 @@ bond_ethdev_stop(struct rte_eth_dev *eth_dev)
 
                /* Discard all messages to/from mode 4 state machines */
                for (i = 0; i < internals->active_slave_count; i++) {
-                       port = &mode_8023ad_ports[internals->active_slaves[i]];
+                       port = &bond_mode_8023ad_ports[internals->active_slaves[i]];
 
                        RTE_ASSERT(port->rx_ring != NULL);
                        while (rte_ring_dequeue(port->rx_ring, &pkt) != -ENOENT)
@@ -2229,19 +2191,27 @@ bond_ethdev_stop(struct rte_eth_dev *eth_dev)
                        tlb_last_obytets[internals->active_slaves[i]] = 0;
        }
 
-       internals->link_status_polling_enabled = 0;
-       for (i = 0; i < internals->slave_count; i++)
-               internals->slaves[i].last_link_status = 0;
-
        eth_dev->data->dev_link.link_status = ETH_LINK_DOWN;
        eth_dev->data->dev_started = 0;
+
+       internals->link_status_polling_enabled = 0;
+       for (i = 0; i < internals->slave_count; i++) {
+               uint16_t slave_id = internals->slaves[i].port_id;
+               if (find_slave_by_id(internals->active_slaves,
+                               internals->active_slave_count, slave_id) !=
+                                               internals->active_slave_count) {
+                       internals->slaves[i].last_link_status = 0;
+                       rte_eth_dev_stop(slave_id);
+                       deactivate_slave(eth_dev, slave_id);
+               }
+       }
 }
 
 void
 bond_ethdev_close(struct rte_eth_dev *dev)
 {
        struct bond_dev_private *internals = dev->data->dev_private;
-       uint8_t bond_port_id = internals->port_id;
+       uint16_t bond_port_id = internals->port_id;
        int skipped = 0;
        struct rte_flow_error ferror;
 
@@ -2273,6 +2243,8 @@ bond_ethdev_info(struct rte_eth_dev *dev, struct rte_eth_dev_info *dev_info)
 
        uint16_t max_nb_rx_queues = UINT16_MAX;
        uint16_t max_nb_tx_queues = UINT16_MAX;
+       uint16_t max_rx_desc_lim = UINT16_MAX;
+       uint16_t max_tx_desc_lim = UINT16_MAX;
 
        dev_info->max_mac_addrs = BOND_MAX_MAC_ADDRS;
 
@@ -2286,7 +2258,7 @@ bond_ethdev_info(struct rte_eth_dev *dev, struct rte_eth_dev_info *dev_info)
         */
        if (internals->slave_count > 0) {
                struct rte_eth_dev_info slave_info;
-               uint8_t idx;
+               uint16_t idx;
 
                for (idx = 0; idx < internals->slave_count; idx++) {
                        rte_eth_dev_info_get(internals->slaves[idx].port_id,
@@ -2297,12 +2269,26 @@ bond_ethdev_info(struct rte_eth_dev *dev, struct rte_eth_dev_info *dev_info)
 
                        if (slave_info.max_tx_queues < max_nb_tx_queues)
                                max_nb_tx_queues = slave_info.max_tx_queues;
+
+                       if (slave_info.rx_desc_lim.nb_max < max_rx_desc_lim)
+                               max_rx_desc_lim = slave_info.rx_desc_lim.nb_max;
+
+                       if (slave_info.tx_desc_lim.nb_max < max_tx_desc_lim)
+                               max_tx_desc_lim = slave_info.tx_desc_lim.nb_max;
                }
        }
 
        dev_info->max_rx_queues = max_nb_rx_queues;
        dev_info->max_tx_queues = max_nb_tx_queues;
 
+       memcpy(&dev_info->default_rxconf, &internals->default_rxconf,
+              sizeof(dev_info->default_rxconf));
+       memcpy(&dev_info->default_txconf, &internals->default_txconf,
+              sizeof(dev_info->default_txconf));
+
+       dev_info->rx_desc_lim.nb_max = max_rx_desc_lim;
+       dev_info->tx_desc_lim.nb_max = max_tx_desc_lim;
+
        /**
         * If dedicated hw queues enabled for link bonding device in LACP mode
         * then we need to reduce the maximum number of data path queues by 1.
@@ -2628,6 +2614,9 @@ bond_ethdev_promiscuous_enable(struct rte_eth_dev *eth_dev)
        case BONDING_MODE_TLB:
        case BONDING_MODE_ALB:
        default:
+               /* Do not touch promisc when there cannot be primary ports */
+               if (internals->slave_count == 0)
+                       break;
                rte_eth_promiscuous_enable(internals->current_primary_port);
        }
 }
@@ -2656,6 +2645,9 @@ bond_ethdev_promiscuous_disable(struct rte_eth_dev *dev)
        case BONDING_MODE_TLB:
        case BONDING_MODE_ALB:
        default:
+               /* Do not touch promisc when there cannot be primary ports */
+               if (internals->slave_count == 0)
+                       break;
                rte_eth_promiscuous_disable(internals->current_primary_port);
        }
 }
@@ -2679,14 +2671,15 @@ bond_ethdev_lsc_event_callback(uint16_t port_id, enum rte_eth_event_type type,
        struct rte_eth_link link;
        int rc = -1;
 
-       int i, valid_slave = 0;
-       uint8_t active_pos;
        uint8_t lsc_flag = 0;
+       int valid_slave = 0;
+       uint16_t active_pos;
+       uint16_t i;
 
        if (type != RTE_ETH_EVENT_INTR_LSC || param == NULL)
                return rc;
 
-       bonded_eth_dev = &rte_eth_devices[*(uint8_t *)param];
+       bonded_eth_dev = &rte_eth_devices[*(uint16_t *)param];
 
        if (check_for_bonded_ethdev(bonded_eth_dev))
                return rc;
@@ -2722,10 +2715,24 @@ bond_ethdev_lsc_event_callback(uint16_t port_id, enum rte_eth_event_type type,
                if (active_pos < internals->active_slave_count)
                        goto link_update;
 
-               /* if no active slave ports then set this port to be primary port */
+               /* check link state properties if bonded link is up*/
+               if (bonded_eth_dev->data->dev_link.link_status == ETH_LINK_UP) {
+                       if (link_properties_valid(bonded_eth_dev, &link) != 0)
+                               RTE_BOND_LOG(ERR, "Invalid link properties "
+                                            "for slave %d in bonding mode %d",
+                                            port_id, internals->mode);
+               } else {
+                       /* inherit slave link properties */
+                       link_properties_set(bonded_eth_dev, &link);
+               }
+
+               /* If no active slave ports then set this port to be
+                * the primary port.
+                */
                if (internals->active_slave_count < 1) {
                        /* If first active slave, then change link status */
-                       bonded_eth_dev->data->dev_link.link_status = ETH_LINK_UP;
+                       bonded_eth_dev->data->dev_link.link_status =
+                                                               ETH_LINK_UP;
                        internals->current_primary_port = port_id;
                        lsc_flag = 1;
 
@@ -2734,7 +2741,9 @@ bond_ethdev_lsc_event_callback(uint16_t port_id, enum rte_eth_event_type type,
 
                activate_slave(bonded_eth_dev, port_id);
 
-               /* If user has defined the primary port then default to using it */
+               /* If the user has defined the primary port then default to
+                * using it.
+                */
                if (internals->user_defined_primary_port &&
                                internals->primary_port == port_id)
                        bond_ethdev_primary_set(internals, port_id);
@@ -2936,7 +2945,8 @@ bond_ethdev_mtu_set(struct rte_eth_dev *dev, uint16_t mtu)
 }
 
 static int
-bond_ethdev_mac_address_set(struct rte_eth_dev *dev, struct ether_addr *addr)
+bond_ethdev_mac_address_set(struct rte_eth_dev *dev,
+                       struct rte_ether_addr *addr)
 {
        if (mac_address_set(dev, addr)) {
                RTE_BOND_LOG(ERR, "Failed to update MAC address");
@@ -2958,8 +2968,9 @@ bond_filter_ctrl(struct rte_eth_dev *dev __rte_unused,
 }
 
 static int
-bond_ethdev_mac_addr_add(struct rte_eth_dev *dev, struct ether_addr *mac_addr,
-                               __rte_unused uint32_t index, uint32_t vmdq)
+bond_ethdev_mac_addr_add(struct rte_eth_dev *dev,
+                       struct rte_ether_addr *mac_addr,
+                       __rte_unused uint32_t index, uint32_t vmdq)
 {
        struct rte_eth_dev *slave_eth_dev;
        struct bond_dev_private *internals = dev->data->dev_private;
@@ -3009,7 +3020,7 @@ bond_ethdev_mac_addr_remove(struct rte_eth_dev *dev, uint32_t index)
                        goto end;
        }
 
-       struct ether_addr *mac_addr = &dev->data->mac_addrs[index];
+       struct rte_ether_addr *mac_addr = &dev->data->mac_addrs[index];
 
        for (i = 0; i < internals->slave_count; i++)
                rte_eth_dev_mac_addr_remove(internals->slaves[i].port_id,
@@ -3112,6 +3123,14 @@ bond_alloc(struct rte_vdev_device *dev, uint8_t mode)
        /* Initially allow to choose any offload type */
        internals->flow_type_rss_offloads = ETH_RSS_PROTO_MASK;
 
+       memset(&internals->default_rxconf, 0,
+              sizeof(internals->default_rxconf));
+       memset(&internals->default_txconf, 0,
+              sizeof(internals->default_txconf));
+
+       memset(&internals->rx_desc_lim, 0, sizeof(internals->rx_desc_lim));
+       memset(&internals->tx_desc_lim, 0, sizeof(internals->tx_desc_lim));
+
        memset(internals->active_slaves, 0, sizeof(internals->active_slaves));
        memset(internals->slaves, 0, sizeof(internals->slaves));
 
@@ -3151,10 +3170,9 @@ bond_alloc(struct rte_vdev_device *dev, uint8_t mode)
 
 err:
        rte_free(internals);
-       if (eth_dev != NULL) {
-               rte_free(eth_dev->data->mac_addrs);
-               rte_eth_dev_release_port(eth_dev);
-       }
+       if (eth_dev != NULL)
+               eth_dev->data->dev_private = NULL;
+       rte_eth_dev_release_port(eth_dev);
        return -1;
 }
 
@@ -3175,8 +3193,7 @@ bond_probe(struct rte_vdev_device *dev)
        name = rte_vdev_device_name(dev);
        RTE_BOND_LOG(INFO, "Initializing pmd_bond for %s", name);
 
-       if (rte_eal_process_type() == RTE_PROC_SECONDARY &&
-           strlen(rte_vdev_device_args(dev)) == 0) {
+       if (rte_eal_process_type() == RTE_PROC_SECONDARY) {
                eth_dev = rte_eth_dev_attach_secondary(name);
                if (!eth_dev) {
                        RTE_BOND_LOG(ERR, "Failed to probe %s", name);
@@ -3184,6 +3201,7 @@ bond_probe(struct rte_vdev_device *dev)
                }
                /* TODO: request info from primary to set up Rx and Tx */
                eth_dev->dev_ops = &default_dev_ops;
+               eth_dev->device = &dev->device;
                rte_eth_dev_probing_finish(eth_dev);
                return 0;
        }
@@ -3238,8 +3256,6 @@ bond_probe(struct rte_vdev_device *dev)
        internals = rte_eth_devices[port_id].data->dev_private;
        internals->kvlist = kvlist;
 
-       rte_eth_dev_probing_finish(&rte_eth_devices[port_id]);
-
        if (rte_kvargs_count(kvlist, PMD_BOND_AGG_MODE_KVARG) == 1) {
                if (rte_kvargs_process(kvlist,
                                PMD_BOND_AGG_MODE_KVARG,
@@ -3252,12 +3268,12 @@ bond_probe(struct rte_vdev_device *dev)
                }
 
                if (internals->mode == BONDING_MODE_8023AD)
-                       rte_eth_bond_8023ad_agg_selection_set(port_id,
-                                       agg_mode);
+                       internals->mode4.agg_selection = agg_mode;
        } else {
-               rte_eth_bond_8023ad_agg_selection_set(port_id, AGG_STABLE);
+               internals->mode4.agg_selection = AGG_STABLE;
        }
 
+       rte_eth_dev_probing_finish(&rte_eth_devices[port_id]);
        RTE_BOND_LOG(INFO, "Create bonded device %s on port %d in mode %u on "
                        "socket %u.",   name, port_id, bonding_mode, socket_id);
        return 0;
@@ -3290,6 +3306,9 @@ bond_remove(struct rte_vdev_device *dev)
        if (eth_dev == NULL)
                return -ENODEV;
 
+       if (rte_eal_process_type() != RTE_PROC_PRIMARY)
+               return rte_eth_dev_release_port(eth_dev);
+
        RTE_ASSERT(eth_dev->device == &dev->device);
 
        internals = eth_dev->data->dev_private;
@@ -3312,8 +3331,6 @@ bond_remove(struct rte_vdev_device *dev)
        rte_mempool_free(internals->mode6.mempool);
        rte_bitmap_free(internals->vlan_filter_bmp);
        rte_free(internals->vlan_filter_bmpmem);
-       rte_free(eth_dev->data->dev_private);
-       rte_free(eth_dev->data->mac_addrs);
 
        rte_eth_dev_release_port(eth_dev);
 
@@ -3341,16 +3358,30 @@ bond_ethdev_configure(struct rte_eth_dev *dev)
 
        unsigned i, j;
 
-       /* If RSS is enabled, fill table and key with default values */
+       /*
+        * If RSS is enabled, fill table with default values and
+        * set key to the the value specified in port RSS configuration.
+        * Fall back to default RSS key if the key is not specified
+        */
        if (dev->data->dev_conf.rxmode.mq_mode & ETH_MQ_RX_RSS) {
-               dev->data->dev_conf.rx_adv_conf.rss_conf.rss_key = internals->rss_key;
-               dev->data->dev_conf.rx_adv_conf.rss_conf.rss_key_len = 0;
-               memcpy(internals->rss_key, default_rss_key, 40);
+               if (dev->data->dev_conf.rx_adv_conf.rss_conf.rss_key != NULL) {
+                       internals->rss_key_len =
+                               dev->data->dev_conf.rx_adv_conf.rss_conf.rss_key_len;
+                       memcpy(internals->rss_key,
+                              dev->data->dev_conf.rx_adv_conf.rss_conf.rss_key,
+                              internals->rss_key_len);
+               } else {
+                       internals->rss_key_len = sizeof(default_rss_key);
+                       memcpy(internals->rss_key, default_rss_key,
+                              internals->rss_key_len);
+               }
 
                for (i = 0; i < RTE_DIM(internals->reta_conf); i++) {
                        internals->reta_conf[i].mask = ~0LL;
                        for (j = 0; j < RTE_RETA_GROUP_SIZE; j++)
-                               internals->reta_conf[i].reta[j] = j % dev->data->nb_rx_queues;
+                               internals->reta_conf[i].reta[j] =
+                                               (i * RTE_RETA_GROUP_SIZE + j) %
+                                               dev->data->nb_rx_queues;
                }
        }
 
@@ -3367,7 +3398,7 @@ bond_ethdev_configure(struct rte_eth_dev *dev)
        /* Parse MAC address for bonded device */
        arg_count = rte_kvargs_count(kvlist, PMD_BOND_MAC_ADDR_KVARG);
        if (arg_count == 1) {
-               struct ether_addr bond_mac;
+               struct rte_ether_addr bond_mac;
 
                if (rte_kvargs_process(kvlist, PMD_BOND_MAC_ADDR_KVARG,
                                       &bond_ethdev_parse_bond_mac_addr_kvarg, &bond_mac) < 0) {
@@ -3427,9 +3458,16 @@ bond_ethdev_configure(struct rte_eth_dev *dev)
                                     "Failed to parse agg selection mode for bonded device %s",
                                     name);
                }
-               if (internals->mode == BONDING_MODE_8023AD)
-                       rte_eth_bond_8023ad_agg_selection_set(port_id,
-                                                             agg_mode);
+               if (internals->mode == BONDING_MODE_8023AD) {
+                       int ret = rte_eth_bond_8023ad_agg_selection_set(port_id,
+                                       agg_mode);
+                       if (ret < 0) {
+                               RTE_BOND_LOG(ERR,
+                                       "Invalid args for agg selection set for bonded device %s",
+                                       name);
+                               return -1;
+                       }
+               }
        }
 
        /* Parse/add slave ports to bonded device */
@@ -3606,7 +3644,7 @@ int bond_logtype;
 
 RTE_INIT(bond_init_log)
 {
-       bond_logtype = rte_log_register("pmd.net.bon");
+       bond_logtype = rte_log_register("pmd.net.bond");
        if (bond_logtype >= 0)
                rte_log_set_level(bond_logtype, RTE_LOG_NOTICE);
 }