net/bnxt: implement SSE vector mode
[dpdk.git] / drivers / net / bnxt / bnxt_ethdev.c
index 78ab07d..41771d8 100644 (file)
@@ -214,7 +214,7 @@ static int bnxt_init_chip(struct bnxt *bp)
        /* disable uio/vfio intr/eventfd mapping */
        rte_intr_disable(intr_handle);
 
-       if (bp->eth_dev->data->mtu > ETHER_MTU) {
+       if (bp->eth_dev->data->mtu > RTE_ETHER_MTU) {
                bp->eth_dev->data->dev_conf.rxmode.offloads |=
                        DEV_RX_OFFLOAD_JUMBO_FRAME;
                bp->flags |= BNXT_FLAG_JUMBO;
@@ -462,8 +462,8 @@ static void bnxt_dev_info_get_op(struct rte_eth_dev *eth_dev,
 
        /* Fast path specifics */
        dev_info->min_rx_bufsize = 1;
-       dev_info->max_rx_pktlen = BNXT_MAX_MTU + ETHER_HDR_LEN + ETHER_CRC_LEN
-                                 + VLAN_TAG_SIZE * 2;
+       dev_info->max_rx_pktlen = BNXT_MAX_MTU + RTE_ETHER_HDR_LEN +
+               RTE_ETHER_CRC_LEN + VLAN_TAG_SIZE * 2;
 
        dev_info->rx_offload_capa = BNXT_DEV_RX_OFFLOAD_SUPPORT;
        if (bp->flags & BNXT_FLAG_PTP_SUPPORTED)
@@ -595,9 +595,9 @@ static int bnxt_dev_configure_op(struct rte_eth_dev *eth_dev)
 
        if (rx_offloads & DEV_RX_OFFLOAD_JUMBO_FRAME) {
                eth_dev->data->mtu =
-                               eth_dev->data->dev_conf.rxmode.max_rx_pkt_len -
-                               ETHER_HDR_LEN - ETHER_CRC_LEN - VLAN_TAG_SIZE *
-                               BNXT_NUM_VLANS;
+                       eth_dev->data->dev_conf.rxmode.max_rx_pkt_len -
+                       RTE_ETHER_HDR_LEN - RTE_ETHER_CRC_LEN - VLAN_TAG_SIZE *
+                       BNXT_NUM_VLANS;
                bnxt_mtu_set_op(eth_dev, eth_dev->data->mtu);
        }
        return 0;
@@ -624,6 +624,87 @@ static int bnxt_dev_lsc_intr_setup(struct rte_eth_dev *eth_dev)
        return 0;
 }
 
+/*
+ * Determine whether the current configuration requires support for scattered
+ * receive; return 1 if scattered receive is required and 0 if not.
+ */
+static int bnxt_scattered_rx(struct rte_eth_dev *eth_dev)
+{
+       uint16_t buf_size;
+       int i;
+
+       for (i = 0; i < eth_dev->data->nb_rx_queues; i++) {
+               struct bnxt_rx_queue *rxq = eth_dev->data->rx_queues[i];
+
+               buf_size = (uint16_t)(rte_pktmbuf_data_room_size(rxq->mb_pool) -
+                                     RTE_PKTMBUF_HEADROOM);
+               if (eth_dev->data->dev_conf.rxmode.max_rx_pkt_len > buf_size)
+                       return 1;
+       }
+       return 0;
+}
+
+static eth_rx_burst_t
+bnxt_receive_function(__rte_unused struct rte_eth_dev *eth_dev)
+{
+#ifdef RTE_ARCH_X86
+       /*
+        * Vector mode receive can be enabled only if scatter rx is not
+        * in use and rx offloads are limited to VLAN stripping and
+        * CRC stripping.
+        */
+       if (!eth_dev->data->scattered_rx &&
+           !(eth_dev->data->dev_conf.rxmode.offloads &
+             ~(DEV_RX_OFFLOAD_VLAN_STRIP |
+               DEV_RX_OFFLOAD_KEEP_CRC |
+               DEV_RX_OFFLOAD_JUMBO_FRAME |
+               DEV_RX_OFFLOAD_IPV4_CKSUM |
+               DEV_RX_OFFLOAD_UDP_CKSUM |
+               DEV_RX_OFFLOAD_TCP_CKSUM |
+               DEV_RX_OFFLOAD_OUTER_IPV4_CKSUM |
+               DEV_RX_OFFLOAD_VLAN_FILTER))) {
+               PMD_DRV_LOG(INFO, "Using vector mode receive for port %d\n",
+                           eth_dev->data->port_id);
+               return bnxt_recv_pkts_vec;
+       }
+       PMD_DRV_LOG(INFO, "Vector mode receive disabled for port %d\n",
+                   eth_dev->data->port_id);
+       PMD_DRV_LOG(INFO,
+                   "Port %d scatter: %d rx offload: %" PRIX64 "\n",
+                   eth_dev->data->port_id,
+                   eth_dev->data->scattered_rx,
+                   eth_dev->data->dev_conf.rxmode.offloads);
+#endif
+       return bnxt_recv_pkts;
+}
+
+static eth_tx_burst_t
+bnxt_transmit_function(__rte_unused struct rte_eth_dev *eth_dev)
+{
+#ifdef RTE_ARCH_X86
+       /*
+        * Vector mode receive can be enabled only if scatter tx is not
+        * in use and tx offloads other than VLAN insertion are not
+        * in use.
+        */
+       if (!eth_dev->data->scattered_rx &&
+           !(eth_dev->data->dev_conf.txmode.offloads &
+             ~DEV_TX_OFFLOAD_VLAN_INSERT)) {
+               PMD_DRV_LOG(INFO, "Using vector mode transmit for port %d\n",
+                           eth_dev->data->port_id);
+               return bnxt_xmit_pkts_vec;
+       }
+       PMD_DRV_LOG(INFO, "Vector mode transmit disabled for port %d\n",
+                   eth_dev->data->port_id);
+       PMD_DRV_LOG(INFO,
+                   "Port %d scatter: %d tx offload: %" PRIX64 "\n",
+                   eth_dev->data->port_id,
+                   eth_dev->data->scattered_rx,
+                   eth_dev->data->dev_conf.txmode.offloads);
+#endif
+       return bnxt_xmit_pkts;
+}
+
 static int bnxt_dev_start_op(struct rte_eth_dev *eth_dev)
 {
        struct bnxt *bp = (struct bnxt *)eth_dev->data->dev_private;
@@ -642,6 +723,8 @@ static int bnxt_dev_start_op(struct rte_eth_dev *eth_dev)
        if (rc)
                goto error;
 
+       eth_dev->data->scattered_rx = bnxt_scattered_rx(eth_dev);
+
        bnxt_link_update_op(eth_dev, 1);
 
        if (rx_offloads & DEV_RX_OFFLOAD_VLAN_FILTER)
@@ -652,6 +735,8 @@ static int bnxt_dev_start_op(struct rte_eth_dev *eth_dev)
        if (rc)
                goto error;
 
+       eth_dev->rx_pkt_burst = bnxt_receive_function(eth_dev);
+       eth_dev->tx_pkt_burst = bnxt_transmit_function(eth_dev);
        bp->flags |= BNXT_FLAG_INIT_DONE;
        return 0;
 
@@ -750,7 +835,7 @@ static void bnxt_mac_addr_remove_op(struct rte_eth_dev *eth_dev,
                                                bnxt_filter_info, next);
                                bnxt_hwrm_clear_l2_filter(bp, filter);
                                filter->mac_index = INVALID_MAC_INDEX;
-                               memset(&filter->l2_addr, 0, ETHER_ADDR_LEN);
+                               memset(&filter->l2_addr, 0, RTE_ETHER_ADDR_LEN);
                                STAILQ_INSERT_TAIL(&bp->free_filter_list,
                                                   filter, next);
                        }
@@ -760,7 +845,7 @@ static void bnxt_mac_addr_remove_op(struct rte_eth_dev *eth_dev,
 }
 
 static int bnxt_mac_addr_add_op(struct rte_eth_dev *eth_dev,
-                               struct ether_addr *mac_addr,
+                               struct rte_ether_addr *mac_addr,
                                uint32_t index, uint32_t pool)
 {
        struct bnxt *bp = (struct bnxt *)eth_dev->data->dev_private;
@@ -791,7 +876,7 @@ static int bnxt_mac_addr_add_op(struct rte_eth_dev *eth_dev,
        }
        STAILQ_INSERT_TAIL(&vnic->filter, filter, next);
        filter->mac_index = index;
-       memcpy(filter->l2_addr, mac_addr, ETHER_ADDR_LEN);
+       memcpy(filter->l2_addr, mac_addr, RTE_ETHER_ADDR_LEN);
        return bnxt_hwrm_set_l2_filter(bp, vnic->fw_vnic_id, filter);
 }
 
@@ -1312,7 +1397,7 @@ static int bnxt_del_vlan_filter(struct bnxt *bp, uint16_t vlan_id)
                                new_filter->mac_index =
                                        filter->mac_index;
                                memcpy(new_filter->l2_addr, filter->l2_addr,
-                                      ETHER_ADDR_LEN);
+                                      RTE_ETHER_ADDR_LEN);
                                /* MAC only filter */
                                rc = bnxt_hwrm_set_l2_filter(bp,
                                                             vnic->fw_vnic_id,
@@ -1381,7 +1466,7 @@ static int bnxt_add_vlan_filter(struct bnxt *bp, uint16_t vlan_id)
                        /* Inherit MAC from the previous filter */
                        new_filter->mac_index = filter->mac_index;
                        memcpy(new_filter->l2_addr, filter->l2_addr,
-                              ETHER_ADDR_LEN);
+                              RTE_ETHER_ADDR_LEN);
                        /* MAC + VLAN ID filter */
                        new_filter->l2_ivlan = vlan_id;
                        new_filter->l2_ivlan_mask = 0xF000;
@@ -1451,7 +1536,8 @@ bnxt_vlan_offload_set_op(struct rte_eth_dev *dev, int mask)
 }
 
 static int
-bnxt_set_default_mac_addr_op(struct rte_eth_dev *dev, struct ether_addr *addr)
+bnxt_set_default_mac_addr_op(struct rte_eth_dev *dev,
+                       struct rte_ether_addr *addr)
 {
        struct bnxt *bp = (struct bnxt *)dev->data->dev_private;
        /* Default Filter is tied to VNIC 0 */
@@ -1471,8 +1557,8 @@ bnxt_set_default_mac_addr_op(struct rte_eth_dev *dev, struct ether_addr *addr)
                rc = bnxt_hwrm_clear_l2_filter(bp, filter);
                if (rc)
                        return rc;
-               memcpy(filter->l2_addr, bp->mac_addr, ETHER_ADDR_LEN);
-               memset(filter->l2_addr_mask, 0xff, ETHER_ADDR_LEN);
+               memcpy(filter->l2_addr, bp->mac_addr, RTE_ETHER_ADDR_LEN);
+               memset(filter->l2_addr_mask, 0xff, RTE_ETHER_ADDR_LEN);
                filter->flags |= HWRM_CFA_L2_FILTER_ALLOC_INPUT_FLAGS_PATH_RX;
                filter->enables |=
                        HWRM_CFA_L2_FILTER_ALLOC_INPUT_ENABLES_L2_ADDR |
@@ -1489,7 +1575,7 @@ bnxt_set_default_mac_addr_op(struct rte_eth_dev *dev, struct ether_addr *addr)
 
 static int
 bnxt_dev_set_mc_addr_list_op(struct rte_eth_dev *eth_dev,
-                         struct ether_addr *mc_addr_set,
+                         struct rte_ether_addr *mc_addr_set,
                          uint32_t nb_mc_addr)
 {
        struct bnxt *bp = (struct bnxt *)eth_dev->data->dev_private;
@@ -1507,8 +1593,9 @@ bnxt_dev_set_mc_addr_list_op(struct rte_eth_dev *eth_dev,
        /* TODO Check for Duplicate mcast addresses */
        vnic->flags &= ~BNXT_VNIC_INFO_ALLMULTI;
        for (i = 0; i < nb_mc_addr; i++) {
-               memcpy(vnic->mc_list + off, &mc_addr_list[i], ETHER_ADDR_LEN);
-               off += ETHER_ADDR_LEN;
+               memcpy(vnic->mc_list + off, &mc_addr_list[i],
+                       RTE_ETHER_ADDR_LEN);
+               off += RTE_ETHER_ADDR_LEN;
        }
 
        vnic->mc_addr_cnt = i;
@@ -1576,18 +1663,39 @@ static int bnxt_mtu_set_op(struct rte_eth_dev *eth_dev, uint16_t new_mtu)
 {
        struct bnxt *bp = eth_dev->data->dev_private;
        struct rte_eth_dev_info dev_info;
+       uint32_t new_pkt_size;
        uint32_t rc = 0;
        uint32_t i;
 
+       new_pkt_size = new_mtu + RTE_ETHER_HDR_LEN + RTE_ETHER_CRC_LEN +
+                      VLAN_TAG_SIZE * BNXT_NUM_VLANS;
+
        bnxt_dev_info_get_op(eth_dev, &dev_info);
 
-       if (new_mtu < ETHER_MIN_MTU || new_mtu > BNXT_MAX_MTU) {
+       if (new_mtu < RTE_ETHER_MIN_MTU || new_mtu > BNXT_MAX_MTU) {
                PMD_DRV_LOG(ERR, "MTU requested must be within (%d, %d)\n",
-                       ETHER_MIN_MTU, BNXT_MAX_MTU);
+                       RTE_ETHER_MIN_MTU, BNXT_MAX_MTU);
                return -EINVAL;
        }
 
-       if (new_mtu > ETHER_MTU) {
+#ifdef RTE_ARCH_X86
+       /*
+        * If vector-mode tx/rx is active, disallow any MTU change that would
+        * require scattered receive support.
+        */
+       if (eth_dev->data->dev_started &&
+           (eth_dev->rx_pkt_burst == bnxt_recv_pkts_vec ||
+            eth_dev->tx_pkt_burst == bnxt_xmit_pkts_vec) &&
+           (new_pkt_size >
+            eth_dev->data->min_rx_buf_size - RTE_PKTMBUF_HEADROOM)) {
+               PMD_DRV_LOG(ERR,
+                           "MTU change would require scattered rx support. ");
+               PMD_DRV_LOG(ERR, "Stop port before changing MTU.\n");
+               return -EINVAL;
+       }
+#endif
+
+       if (new_mtu > RTE_ETHER_MTU) {
                bp->flags |= BNXT_FLAG_JUMBO;
                bp->eth_dev->data->dev_conf.rxmode.offloads |=
                        DEV_RX_OFFLOAD_JUMBO_FRAME;
@@ -1597,8 +1705,7 @@ static int bnxt_mtu_set_op(struct rte_eth_dev *eth_dev, uint16_t new_mtu)
                bp->flags &= ~BNXT_FLAG_JUMBO;
        }
 
-       eth_dev->data->dev_conf.rxmode.max_rx_pkt_len =
-               new_mtu + ETHER_HDR_LEN + ETHER_CRC_LEN + VLAN_TAG_SIZE * 2;
+       eth_dev->data->dev_conf.rxmode.max_rx_pkt_len = new_pkt_size;
 
        eth_dev->data->mtu = new_mtu;
        PMD_DRV_LOG(INFO, "New MTU is %d\n", eth_dev->data->mtu);
@@ -1607,8 +1714,8 @@ static int bnxt_mtu_set_op(struct rte_eth_dev *eth_dev, uint16_t new_mtu)
                struct bnxt_vnic_info *vnic = &bp->vnic_info[i];
                uint16_t size = 0;
 
-               vnic->mru = bp->eth_dev->data->mtu + ETHER_HDR_LEN +
-                                       ETHER_CRC_LEN + VLAN_TAG_SIZE * 2;
+               vnic->mru = bp->eth_dev->data->mtu + RTE_ETHER_HDR_LEN +
+                                       RTE_ETHER_CRC_LEN + VLAN_TAG_SIZE * 2;
                rc = bnxt_hwrm_vnic_cfg(bp, vnic);
                if (rc)
                        break;
@@ -1793,8 +1900,8 @@ bnxt_match_and_validate_ether_filter(struct bnxt *bp,
        int match = 0;
        *ret = 0;
 
-       if (efilter->ether_type == ETHER_TYPE_IPv4 ||
-               efilter->ether_type == ETHER_TYPE_IPv6) {
+       if (efilter->ether_type == RTE_ETHER_TYPE_IPV4 ||
+               efilter->ether_type == RTE_ETHER_TYPE_IPV6) {
                PMD_DRV_LOG(ERR, "invalid ether_type(0x%04x) in"
                        " ethertype filter.", efilter->ether_type);
                *ret = -EINVAL;
@@ -1817,7 +1924,7 @@ bnxt_match_and_validate_ether_filter(struct bnxt *bp,
        if (efilter->flags & RTE_ETHTYPE_FLAGS_DROP) {
                STAILQ_FOREACH(mfilter, &vnic0->filter, next) {
                        if ((!memcmp(efilter->mac_addr.addr_bytes,
-                                    mfilter->l2_addr, ETHER_ADDR_LEN) &&
+                                    mfilter->l2_addr, RTE_ETHER_ADDR_LEN) &&
                             mfilter->flags ==
                             HWRM_CFA_NTUPLE_FILTER_ALLOC_INPUT_FLAGS_DROP &&
                             mfilter->ethertype == efilter->ether_type)) {
@@ -1828,7 +1935,7 @@ bnxt_match_and_validate_ether_filter(struct bnxt *bp,
        } else {
                STAILQ_FOREACH(mfilter, &vnic->filter, next)
                        if ((!memcmp(efilter->mac_addr.addr_bytes,
-                                    mfilter->l2_addr, ETHER_ADDR_LEN) &&
+                                    mfilter->l2_addr, RTE_ETHER_ADDR_LEN) &&
                             mfilter->ethertype == efilter->ether_type &&
                             mfilter->flags ==
                             HWRM_CFA_L2_FILTER_CFG_INPUT_FLAGS_PATH_RX)) {
@@ -1883,9 +1990,9 @@ bnxt_ethertype_filter(struct rte_eth_dev *dev,
                }
                bfilter->filter_type = HWRM_CFA_NTUPLE_FILTER;
                memcpy(bfilter->l2_addr, efilter->mac_addr.addr_bytes,
-                      ETHER_ADDR_LEN);
+                      RTE_ETHER_ADDR_LEN);
                memcpy(bfilter->dst_macaddr, efilter->mac_addr.addr_bytes,
-                      ETHER_ADDR_LEN);
+                      RTE_ETHER_ADDR_LEN);
                bfilter->enables |= NTUPLE_FLTR_ALLOC_INPUT_EN_DST_MACADDR;
                bfilter->ethertype = efilter->ether_type;
                bfilter->enables |= NTUPLE_FLTR_ALLOC_INPUT_EN_ETHERTYPE;
@@ -2396,7 +2503,7 @@ bnxt_parse_fdir_filter(struct bnxt *bp,
                //filter1 = bnxt_get_l2_filter(bp, filter, vnic0);
        } else {
                filter->dst_id = vnic->fw_vnic_id;
-               for (i = 0; i < ETHER_ADDR_LEN; i++)
+               for (i = 0; i < RTE_ETHER_ADDR_LEN; i++)
                        if (filter->dst_macaddr[i] == 0x00)
                                filter1 = STAILQ_FIRST(&vnic0->filter);
                        else
@@ -2440,13 +2547,14 @@ bnxt_match_fdir(struct bnxt *bp, struct bnxt_filter_info *nf,
                            mf->l2_ovlan_mask == nf->l2_ovlan_mask &&
                            mf->l2_ivlan == nf->l2_ivlan &&
                            mf->l2_ivlan_mask == nf->l2_ivlan_mask &&
-                           !memcmp(mf->l2_addr, nf->l2_addr, ETHER_ADDR_LEN) &&
+                           !memcmp(mf->l2_addr, nf->l2_addr,
+                                   RTE_ETHER_ADDR_LEN) &&
                            !memcmp(mf->l2_addr_mask, nf->l2_addr_mask,
-                                   ETHER_ADDR_LEN) &&
+                                   RTE_ETHER_ADDR_LEN) &&
                            !memcmp(mf->src_macaddr, nf->src_macaddr,
-                                   ETHER_ADDR_LEN) &&
+                                   RTE_ETHER_ADDR_LEN) &&
                            !memcmp(mf->dst_macaddr, nf->dst_macaddr,
-                                   ETHER_ADDR_LEN) &&
+                                   RTE_ETHER_ADDR_LEN) &&
                            !memcmp(mf->src_ipaddr, nf->src_ipaddr,
                                    sizeof(nf->src_ipaddr)) &&
                            !memcmp(mf->src_ipaddr_mask, nf->src_ipaddr_mask,
@@ -2632,9 +2740,10 @@ bnxt_dev_supported_ptypes_get_op(struct rte_eth_dev *dev)
                RTE_PTYPE_UNKNOWN
        };
 
-       if (dev->rx_pkt_burst == bnxt_recv_pkts)
-               return ptypes;
-       return NULL;
+       if (!dev->rx_pkt_burst)
+               return NULL;
+
+       return ptypes;
 }
 
 static int bnxt_map_regs(struct bnxt *bp, uint32_t *reg_arr, int count,
@@ -2649,7 +2758,7 @@ static int bnxt_map_regs(struct bnxt *bp, uint32_t *reg_arr, int count,
                        return -ERANGE;
        }
        win_off = BNXT_GRCPF_REG_WINDOW_BASE_OUT + (reg_win - 1) * 4;
-       rte_cpu_to_le_32(rte_write32(reg_base, (uint8_t *)bp->bar0 + win_off));
+       rte_write32(reg_base, (uint8_t *)bp->bar0 + win_off);
        return 0;
 }
 
@@ -2680,10 +2789,10 @@ static int bnxt_map_ptp_regs(struct bnxt *bp)
 
 static void bnxt_unmap_ptp_regs(struct bnxt *bp)
 {
-       rte_cpu_to_le_32(rte_write32(0, (uint8_t *)bp->bar0 +
-                        BNXT_GRCPF_REG_WINDOW_BASE_OUT + 16));
-       rte_cpu_to_le_32(rte_write32(0, (uint8_t *)bp->bar0 +
-                        BNXT_GRCPF_REG_WINDOW_BASE_OUT + 20));
+       rte_write32(0, (uint8_t *)bp->bar0 +
+                        BNXT_GRCPF_REG_WINDOW_BASE_OUT + 16);
+       rte_write32(0, (uint8_t *)bp->bar0 +
+                        BNXT_GRCPF_REG_WINDOW_BASE_OUT + 20);
 }
 
 static uint64_t bnxt_cc_read(struct bnxt *bp)
@@ -2733,8 +2842,8 @@ static int bnxt_get_rx_ts(struct bnxt *bp, uint64_t *ts)
                return -EAGAIN;
 
        port_id = pf->port_id;
-       rte_cpu_to_le_32(rte_write32(1 << port_id, (uint8_t *)bp->bar0 +
-              ptp->rx_mapped_regs[BNXT_PTP_RX_FIFO_ADV]));
+       rte_write32(1 << port_id, (uint8_t *)bp->bar0 +
+              ptp->rx_mapped_regs[BNXT_PTP_RX_FIFO_ADV]);
 
        fifo = rte_le_to_cpu_32(rte_read32((uint8_t *)bp->bar0 +
                                   ptp->rx_mapped_regs[BNXT_PTP_RX_FIFO]));
@@ -3353,16 +3462,16 @@ skip_ext_stats:
                goto error_free;
        }
        eth_dev->data->mac_addrs = rte_zmalloc("bnxt_mac_addr_tbl",
-                                       ETHER_ADDR_LEN * bp->max_l2_ctx, 0);
+                                       RTE_ETHER_ADDR_LEN * bp->max_l2_ctx, 0);
        if (eth_dev->data->mac_addrs == NULL) {
                PMD_DRV_LOG(ERR,
                        "Failed to alloc %u bytes needed to store MAC addr tbl",
-                       ETHER_ADDR_LEN * bp->max_l2_ctx);
+                       RTE_ETHER_ADDR_LEN * bp->max_l2_ctx);
                rc = -ENOMEM;
                goto error_free;
        }
 
-       if (bnxt_check_zero_bytes(bp->dflt_mac_addr, ETHER_ADDR_LEN)) {
+       if (bnxt_check_zero_bytes(bp->dflt_mac_addr, RTE_ETHER_ADDR_LEN)) {
                PMD_DRV_LOG(ERR,
                            "Invalid MAC addr %02X:%02X:%02X:%02X:%02X:%02X\n",
                            bp->dflt_mac_addr[0], bp->dflt_mac_addr[1],
@@ -3373,7 +3482,7 @@ skip_ext_stats:
        }
        /* Copy the permanent MAC from the qcap response address now. */
        memcpy(bp->mac_addr, bp->dflt_mac_addr, sizeof(bp->mac_addr));
-       memcpy(&eth_dev->data->mac_addrs[0], bp->mac_addr, ETHER_ADDR_LEN);
+       memcpy(&eth_dev->data->mac_addrs[0], bp->mac_addr, RTE_ETHER_ADDR_LEN);
 
        if (bp->max_ring_grps < bp->rx_cp_nr_rings) {
                /* 1 ring is for default completion ring */