mbuf: mark old VLAN offload flags as deprecated
[dpdk.git] / drivers / net / e1000 / igb_rxtx.c
index e04c2b4..92bcd72 100644 (file)
@@ -54,7 +54,7 @@
                PKT_TX_OUTER_IPV4 |      \
                PKT_TX_IPV6 |            \
                PKT_TX_IPV4 |            \
-               PKT_TX_VLAN_PKT |                \
+               PKT_TX_VLAN |            \
                PKT_TX_IP_CKSUM |                \
                PKT_TX_L4_MASK |                 \
                PKT_TX_TCP_SEG |                 \
@@ -111,7 +111,7 @@ struct igb_rx_queue {
        uint8_t             crc_len;    /**< 0 if CRC stripped, 4 otherwise. */
        uint8_t             drop_en;  /**< If not 0, set SRRCTL.Drop_En. */
        uint32_t            flags;      /**< RX flags. */
-       uint64_t            offloads;   /**< offloads of DEV_RX_OFFLOAD_* */
+       uint64_t            offloads;   /**< offloads of RTE_ETH_RX_OFFLOAD_* */
        const struct rte_memzone *mz;
 };
 
@@ -186,7 +186,7 @@ struct igb_tx_queue {
        /**< Start context position for transmit queue. */
        struct igb_advctx_info ctx_cache[IGB_CTX_NUM];
        /**< Hardware context history.*/
-       uint64_t               offloads; /**< offloads of DEV_TX_OFFLOAD_* */
+       uint64_t               offloads; /**< offloads of RTE_ETH_TX_OFFLOAD_* */
        const struct rte_memzone *mz;
 };
 
@@ -262,7 +262,7 @@ igbe_set_xmit_ctx(struct igb_tx_queue* txq,
        /* Specify which HW CTX to upload. */
        mss_l4len_idx = (ctx_idx << E1000_ADVTXD_IDX_SHIFT);
 
-       if (ol_flags & PKT_TX_VLAN_PKT)
+       if (ol_flags & PKT_TX_VLAN)
                tx_offload_mask.data |= TX_VLAN_CMP_MASK;
 
        /* check if TCP segmentation required for this packet */
@@ -371,7 +371,7 @@ tx_desc_vlan_flags_to_cmdtype(uint64_t ol_flags)
        uint32_t cmdtype;
        static uint32_t vlan_cmd[2] = {0, E1000_ADVTXD_DCMD_VLE};
        static uint32_t tso_cmd[2] = {0, E1000_ADVTXD_DCMD_TSE};
-       cmdtype = vlan_cmd[(ol_flags & PKT_TX_VLAN_PKT) != 0];
+       cmdtype = vlan_cmd[(ol_flags & PKT_TX_VLAN) != 0];
        cmdtype |= tso_cmd[(ol_flags & PKT_TX_TCP_SEG) != 0];
        return cmdtype;
 }
@@ -1459,13 +1459,13 @@ igb_get_tx_port_offloads_capa(struct rte_eth_dev *dev)
        uint64_t tx_offload_capa;
 
        RTE_SET_USED(dev);
-       tx_offload_capa = DEV_TX_OFFLOAD_VLAN_INSERT |
-                         DEV_TX_OFFLOAD_IPV4_CKSUM  |
-                         DEV_TX_OFFLOAD_UDP_CKSUM   |
-                         DEV_TX_OFFLOAD_TCP_CKSUM   |
-                         DEV_TX_OFFLOAD_SCTP_CKSUM  |
-                         DEV_TX_OFFLOAD_TCP_TSO     |
-                         DEV_TX_OFFLOAD_MULTI_SEGS;
+       tx_offload_capa = RTE_ETH_TX_OFFLOAD_VLAN_INSERT |
+                         RTE_ETH_TX_OFFLOAD_IPV4_CKSUM  |
+                         RTE_ETH_TX_OFFLOAD_UDP_CKSUM   |
+                         RTE_ETH_TX_OFFLOAD_TCP_CKSUM   |
+                         RTE_ETH_TX_OFFLOAD_SCTP_CKSUM  |
+                         RTE_ETH_TX_OFFLOAD_TCP_TSO     |
+                         RTE_ETH_TX_OFFLOAD_MULTI_SEGS;
 
        return tx_offload_capa;
 }
@@ -1640,20 +1640,19 @@ igb_get_rx_port_offloads_capa(struct rte_eth_dev *dev)
 
        hw = E1000_DEV_PRIVATE_TO_HW(dev->data->dev_private);
 
-       rx_offload_capa = DEV_RX_OFFLOAD_VLAN_STRIP  |
-                         DEV_RX_OFFLOAD_VLAN_FILTER |
-                         DEV_RX_OFFLOAD_IPV4_CKSUM  |
-                         DEV_RX_OFFLOAD_UDP_CKSUM   |
-                         DEV_RX_OFFLOAD_TCP_CKSUM   |
-                         DEV_RX_OFFLOAD_JUMBO_FRAME |
-                         DEV_RX_OFFLOAD_KEEP_CRC    |
-                         DEV_RX_OFFLOAD_SCATTER     |
-                         DEV_RX_OFFLOAD_RSS_HASH;
+       rx_offload_capa = RTE_ETH_RX_OFFLOAD_VLAN_STRIP  |
+                         RTE_ETH_RX_OFFLOAD_VLAN_FILTER |
+                         RTE_ETH_RX_OFFLOAD_IPV4_CKSUM  |
+                         RTE_ETH_RX_OFFLOAD_UDP_CKSUM   |
+                         RTE_ETH_RX_OFFLOAD_TCP_CKSUM   |
+                         RTE_ETH_RX_OFFLOAD_KEEP_CRC    |
+                         RTE_ETH_RX_OFFLOAD_SCATTER     |
+                         RTE_ETH_RX_OFFLOAD_RSS_HASH;
 
        if (hw->mac.type == e1000_i350 ||
            hw->mac.type == e1000_i210 ||
            hw->mac.type == e1000_i211)
-               rx_offload_capa |= DEV_RX_OFFLOAD_VLAN_EXTEND;
+               rx_offload_capa |= RTE_ETH_RX_OFFLOAD_VLAN_EXTEND;
 
        return rx_offload_capa;
 }
@@ -1734,7 +1733,7 @@ eth_igb_rx_queue_setup(struct rte_eth_dev *dev,
        rxq->reg_idx = (uint16_t)((RTE_ETH_DEV_SRIOV(dev).active == 0) ?
                queue_idx : RTE_ETH_DEV_SRIOV(dev).def_pool_q_idx + queue_idx);
        rxq->port_id = dev->data->port_id;
-       if (dev->data->dev_conf.rxmode.offloads & DEV_RX_OFFLOAD_KEEP_CRC)
+       if (dev->data->dev_conf.rxmode.offloads & RTE_ETH_RX_OFFLOAD_KEEP_CRC)
                rxq->crc_len = RTE_ETHER_CRC_LEN;
        else
                rxq->crc_len = 0;
@@ -1951,23 +1950,23 @@ igb_hw_rss_hash_set(struct e1000_hw *hw, struct rte_eth_rss_conf *rss_conf)
        /* Set configured hashing protocols in MRQC register */
        rss_hf = rss_conf->rss_hf;
        mrqc = E1000_MRQC_ENABLE_RSS_4Q; /* RSS enabled. */
-       if (rss_hf & ETH_RSS_IPV4)
+       if (rss_hf & RTE_ETH_RSS_IPV4)
                mrqc |= E1000_MRQC_RSS_FIELD_IPV4;
-       if (rss_hf & ETH_RSS_NONFRAG_IPV4_TCP)
+       if (rss_hf & RTE_ETH_RSS_NONFRAG_IPV4_TCP)
                mrqc |= E1000_MRQC_RSS_FIELD_IPV4_TCP;
-       if (rss_hf & ETH_RSS_IPV6)
+       if (rss_hf & RTE_ETH_RSS_IPV6)
                mrqc |= E1000_MRQC_RSS_FIELD_IPV6;
-       if (rss_hf & ETH_RSS_IPV6_EX)
+       if (rss_hf & RTE_ETH_RSS_IPV6_EX)
                mrqc |= E1000_MRQC_RSS_FIELD_IPV6_EX;
-       if (rss_hf & ETH_RSS_NONFRAG_IPV6_TCP)
+       if (rss_hf & RTE_ETH_RSS_NONFRAG_IPV6_TCP)
                mrqc |= E1000_MRQC_RSS_FIELD_IPV6_TCP;
-       if (rss_hf & ETH_RSS_IPV6_TCP_EX)
+       if (rss_hf & RTE_ETH_RSS_IPV6_TCP_EX)
                mrqc |= E1000_MRQC_RSS_FIELD_IPV6_TCP_EX;
-       if (rss_hf & ETH_RSS_NONFRAG_IPV4_UDP)
+       if (rss_hf & RTE_ETH_RSS_NONFRAG_IPV4_UDP)
                mrqc |= E1000_MRQC_RSS_FIELD_IPV4_UDP;
-       if (rss_hf & ETH_RSS_NONFRAG_IPV6_UDP)
+       if (rss_hf & RTE_ETH_RSS_NONFRAG_IPV6_UDP)
                mrqc |= E1000_MRQC_RSS_FIELD_IPV6_UDP;
-       if (rss_hf & ETH_RSS_IPV6_UDP_EX)
+       if (rss_hf & RTE_ETH_RSS_IPV6_UDP_EX)
                mrqc |= E1000_MRQC_RSS_FIELD_IPV6_UDP_EX;
        E1000_WRITE_REG(hw, E1000_MRQC, mrqc);
 }
@@ -2033,23 +2032,23 @@ int eth_igb_rss_hash_conf_get(struct rte_eth_dev *dev,
        }
        rss_hf = 0;
        if (mrqc & E1000_MRQC_RSS_FIELD_IPV4)
-               rss_hf |= ETH_RSS_IPV4;
+               rss_hf |= RTE_ETH_RSS_IPV4;
        if (mrqc & E1000_MRQC_RSS_FIELD_IPV4_TCP)
-               rss_hf |= ETH_RSS_NONFRAG_IPV4_TCP;
+               rss_hf |= RTE_ETH_RSS_NONFRAG_IPV4_TCP;
        if (mrqc & E1000_MRQC_RSS_FIELD_IPV6)
-               rss_hf |= ETH_RSS_IPV6;
+               rss_hf |= RTE_ETH_RSS_IPV6;
        if (mrqc & E1000_MRQC_RSS_FIELD_IPV6_EX)
-               rss_hf |= ETH_RSS_IPV6_EX;
+               rss_hf |= RTE_ETH_RSS_IPV6_EX;
        if (mrqc & E1000_MRQC_RSS_FIELD_IPV6_TCP)
-               rss_hf |= ETH_RSS_NONFRAG_IPV6_TCP;
+               rss_hf |= RTE_ETH_RSS_NONFRAG_IPV6_TCP;
        if (mrqc & E1000_MRQC_RSS_FIELD_IPV6_TCP_EX)
-               rss_hf |= ETH_RSS_IPV6_TCP_EX;
+               rss_hf |= RTE_ETH_RSS_IPV6_TCP_EX;
        if (mrqc & E1000_MRQC_RSS_FIELD_IPV4_UDP)
-               rss_hf |= ETH_RSS_NONFRAG_IPV4_UDP;
+               rss_hf |= RTE_ETH_RSS_NONFRAG_IPV4_UDP;
        if (mrqc & E1000_MRQC_RSS_FIELD_IPV6_UDP)
-               rss_hf |= ETH_RSS_NONFRAG_IPV6_UDP;
+               rss_hf |= RTE_ETH_RSS_NONFRAG_IPV6_UDP;
        if (mrqc & E1000_MRQC_RSS_FIELD_IPV6_UDP_EX)
-               rss_hf |= ETH_RSS_IPV6_UDP_EX;
+               rss_hf |= RTE_ETH_RSS_IPV6_UDP_EX;
        rss_conf->rss_hf = rss_hf;
        return 0;
 }
@@ -2171,15 +2170,15 @@ igb_vmdq_rx_hw_configure(struct rte_eth_dev *dev)
                        E1000_VMOLR_ROPE | E1000_VMOLR_BAM |
                        E1000_VMOLR_MPME);
 
-               if (cfg->rx_mode & ETH_VMDQ_ACCEPT_UNTAG)
+               if (cfg->rx_mode & RTE_ETH_VMDQ_ACCEPT_UNTAG)
                        vmolr |= E1000_VMOLR_AUPE;
-               if (cfg->rx_mode & ETH_VMDQ_ACCEPT_HASH_MC)
+               if (cfg->rx_mode & RTE_ETH_VMDQ_ACCEPT_HASH_MC)
                        vmolr |= E1000_VMOLR_ROMPE;
-               if (cfg->rx_mode & ETH_VMDQ_ACCEPT_HASH_UC)
+               if (cfg->rx_mode & RTE_ETH_VMDQ_ACCEPT_HASH_UC)
                        vmolr |= E1000_VMOLR_ROPE;
-               if (cfg->rx_mode & ETH_VMDQ_ACCEPT_BROADCAST)
+               if (cfg->rx_mode & RTE_ETH_VMDQ_ACCEPT_BROADCAST)
                        vmolr |= E1000_VMOLR_BAM;
-               if (cfg->rx_mode & ETH_VMDQ_ACCEPT_MULTICAST)
+               if (cfg->rx_mode & RTE_ETH_VMDQ_ACCEPT_MULTICAST)
                        vmolr |= E1000_VMOLR_MPME;
 
                E1000_WRITE_REG(hw, E1000_VMOLR(i), vmolr);
@@ -2215,9 +2214,9 @@ igb_vmdq_rx_hw_configure(struct rte_eth_dev *dev)
        /* VLVF: set up filters for vlan tags as configured */
        for (i = 0; i < cfg->nb_pool_maps; i++) {
                /* set vlan id in VF register and set the valid bit */
-               E1000_WRITE_REG(hw, E1000_VLVF(i), (E1000_VLVF_VLANID_ENABLE | \
-                        (cfg->pool_map[i].vlan_id & ETH_VLAN_ID_MAX) | \
-                       ((cfg->pool_map[i].pools << E1000_VLVF_POOLSEL_SHIFT ) & \
+               E1000_WRITE_REG(hw, E1000_VLVF(i), (E1000_VLVF_VLANID_ENABLE |
+                       (cfg->pool_map[i].vlan_id & RTE_ETH_VLAN_ID_MAX) |
+                       ((cfg->pool_map[i].pools << E1000_VLVF_POOLSEL_SHIFT) &
                        E1000_VLVF_POOLSEL_MASK)));
        }
 
@@ -2269,7 +2268,7 @@ igb_dev_mq_rx_configure(struct rte_eth_dev *dev)
                E1000_DEV_PRIVATE_TO_HW(dev->data->dev_private);
        uint32_t mrqc;
 
-       if (RTE_ETH_DEV_SRIOV(dev).active == ETH_8_POOLS) {
+       if (RTE_ETH_DEV_SRIOV(dev).active == RTE_ETH_8_POOLS) {
                /*
                 * SRIOV active scheme
                 * FIXME if support RSS together with VMDq & SRIOV
@@ -2283,14 +2282,14 @@ igb_dev_mq_rx_configure(struct rte_eth_dev *dev)
                 * SRIOV inactive scheme
                 */
                switch (dev->data->dev_conf.rxmode.mq_mode) {
-                       case ETH_MQ_RX_RSS:
+                       case RTE_ETH_MQ_RX_RSS:
                                igb_rss_configure(dev);
                                break;
-                       case ETH_MQ_RX_VMDQ_ONLY:
+                       case RTE_ETH_MQ_RX_VMDQ_ONLY:
                                /*Configure general VMDQ only RX parameters*/
                                igb_vmdq_rx_hw_configure(dev);
                                break;
-                       case ETH_MQ_RX_NONE:
+                       case RTE_ETH_MQ_RX_NONE:
                                /* if mq_mode is none, disable rss mode.*/
                        default:
                                igb_rss_disable(dev);
@@ -2312,6 +2311,7 @@ eth_igb_rx_init(struct rte_eth_dev *dev)
        uint32_t srrctl;
        uint16_t buf_size;
        uint16_t rctl_bsize;
+       uint32_t max_len;
        uint16_t i;
        int ret;
 
@@ -2330,16 +2330,15 @@ eth_igb_rx_init(struct rte_eth_dev *dev)
        /*
         * Configure support of jumbo frames, if any.
         */
-       if (dev->data->dev_conf.rxmode.offloads & DEV_RX_OFFLOAD_JUMBO_FRAME) {
-               uint32_t max_len = dev->data->dev_conf.rxmode.max_rx_pkt_len;
-
+       max_len = dev->data->mtu + E1000_ETH_OVERHEAD;
+       if (dev->data->mtu > RTE_ETHER_MTU) {
                rctl |= E1000_RCTL_LPE;
 
                /*
                 * Set maximum packet length by default, and might be updated
                 * together with enabling/disabling dual VLAN.
                 */
-               if (rxmode->offloads & DEV_RX_OFFLOAD_VLAN_EXTEND)
+               if (rxmode->offloads & RTE_ETH_RX_OFFLOAD_VLAN_EXTEND)
                        max_len += VLAN_TAG_SIZE;
 
                E1000_WRITE_REG(hw, E1000_RLPML, max_len);
@@ -2375,7 +2374,7 @@ eth_igb_rx_init(struct rte_eth_dev *dev)
                 * Reset crc_len in case it was changed after queue setup by a
                 *  call to configure
                 */
-               if (dev->data->dev_conf.rxmode.offloads & DEV_RX_OFFLOAD_KEEP_CRC)
+               if (dev->data->dev_conf.rxmode.offloads & RTE_ETH_RX_OFFLOAD_KEEP_CRC)
                        rxq->crc_len = RTE_ETHER_CRC_LEN;
                else
                        rxq->crc_len = 0;
@@ -2410,8 +2409,7 @@ eth_igb_rx_init(struct rte_eth_dev *dev)
                                               E1000_SRRCTL_BSIZEPKT_SHIFT);
 
                        /* It adds dual VLAN length for supporting dual VLAN */
-                       if ((dev->data->dev_conf.rxmode.max_rx_pkt_len +
-                                               2 * VLAN_TAG_SIZE) > buf_size){
+                       if ((max_len + 2 * VLAN_TAG_SIZE) > buf_size) {
                                if (!dev->data->scattered_rx)
                                        PMD_INIT_LOG(DEBUG,
                                                     "forcing scatter mode");
@@ -2446,7 +2444,7 @@ eth_igb_rx_init(struct rte_eth_dev *dev)
                E1000_WRITE_REG(hw, E1000_RXDCTL(rxq->reg_idx), rxdctl);
        }
 
-       if (dev->data->dev_conf.rxmode.offloads & DEV_RX_OFFLOAD_SCATTER) {
+       if (dev->data->dev_conf.rxmode.offloads & RTE_ETH_RX_OFFLOAD_SCATTER) {
                if (!dev->data->scattered_rx)
                        PMD_INIT_LOG(DEBUG, "forcing scatter mode");
                dev->rx_pkt_burst = eth_igb_recv_scattered_pkts;
@@ -2490,16 +2488,16 @@ eth_igb_rx_init(struct rte_eth_dev *dev)
        rxcsum |= E1000_RXCSUM_PCSD;
 
        /* Enable both L3/L4 rx checksum offload */
-       if (rxmode->offloads & DEV_RX_OFFLOAD_IPV4_CKSUM)
+       if (rxmode->offloads & RTE_ETH_RX_OFFLOAD_IPV4_CKSUM)
                rxcsum |= E1000_RXCSUM_IPOFL;
        else
                rxcsum &= ~E1000_RXCSUM_IPOFL;
        if (rxmode->offloads &
-               (DEV_RX_OFFLOAD_TCP_CKSUM | DEV_RX_OFFLOAD_UDP_CKSUM))
+               (RTE_ETH_RX_OFFLOAD_TCP_CKSUM | RTE_ETH_RX_OFFLOAD_UDP_CKSUM))
                rxcsum |= E1000_RXCSUM_TUOFL;
        else
                rxcsum &= ~E1000_RXCSUM_TUOFL;
-       if (rxmode->offloads & DEV_RX_OFFLOAD_CHECKSUM)
+       if (rxmode->offloads & RTE_ETH_RX_OFFLOAD_CHECKSUM)
                rxcsum |= E1000_RXCSUM_CRCOFL;
        else
                rxcsum &= ~E1000_RXCSUM_CRCOFL;
@@ -2507,7 +2505,7 @@ eth_igb_rx_init(struct rte_eth_dev *dev)
        E1000_WRITE_REG(hw, E1000_RXCSUM, rxcsum);
 
        /* Setup the Receive Control Register. */
-       if (dev->data->dev_conf.rxmode.offloads & DEV_RX_OFFLOAD_KEEP_CRC) {
+       if (dev->data->dev_conf.rxmode.offloads & RTE_ETH_RX_OFFLOAD_KEEP_CRC) {
                rctl &= ~E1000_RCTL_SECRC; /* Do not Strip Ethernet CRC. */
 
                /* clear STRCRC bit in all queues */
@@ -2547,7 +2545,7 @@ eth_igb_rx_init(struct rte_eth_dev *dev)
                (hw->mac.mc_filter_type << E1000_RCTL_MO_SHIFT);
 
        /* Make sure VLAN Filters are off. */
-       if (dev->data->dev_conf.rxmode.mq_mode != ETH_MQ_RX_VMDQ_ONLY)
+       if (dev->data->dev_conf.rxmode.mq_mode != RTE_ETH_MQ_RX_VMDQ_ONLY)
                rctl &= ~E1000_RCTL_VFE;
        /* Don't store bad packets. */
        rctl &= ~E1000_RCTL_SBP;
@@ -2635,15 +2633,15 @@ eth_igbvf_rx_init(struct rte_eth_dev *dev)
        uint32_t srrctl;
        uint16_t buf_size;
        uint16_t rctl_bsize;
+       uint32_t max_len;
        uint16_t i;
        int ret;
 
        hw = E1000_DEV_PRIVATE_TO_HW(dev->data->dev_private);
 
        /* setup MTU */
-       e1000_rlpml_set_vf(hw,
-               (uint16_t)(dev->data->dev_conf.rxmode.max_rx_pkt_len +
-               VLAN_TAG_SIZE));
+       max_len = dev->data->mtu + E1000_ETH_OVERHEAD;
+       e1000_rlpml_set_vf(hw, (uint16_t)(max_len + VLAN_TAG_SIZE));
 
        /* Configure and enable each RX queue. */
        rctl_bsize = 0;
@@ -2700,8 +2698,7 @@ eth_igbvf_rx_init(struct rte_eth_dev *dev)
                                               E1000_SRRCTL_BSIZEPKT_SHIFT);
 
                        /* It adds dual VLAN length for supporting dual VLAN */
-                       if ((dev->data->dev_conf.rxmode.max_rx_pkt_len +
-                                               2 * VLAN_TAG_SIZE) > buf_size){
+                       if ((max_len + 2 * VLAN_TAG_SIZE) > buf_size) {
                                if (!dev->data->scattered_rx)
                                        PMD_INIT_LOG(DEBUG,
                                                     "forcing scatter mode");
@@ -2746,7 +2743,7 @@ eth_igbvf_rx_init(struct rte_eth_dev *dev)
                E1000_WRITE_REG(hw, E1000_RXDCTL(i), rxdctl);
        }
 
-       if (dev->data->dev_conf.rxmode.offloads & DEV_RX_OFFLOAD_SCATTER) {
+       if (dev->data->dev_conf.rxmode.offloads & RTE_ETH_RX_OFFLOAD_SCATTER) {
                if (!dev->data->scattered_rx)
                        PMD_INIT_LOG(DEBUG, "forcing scatter mode");
                dev->rx_pkt_burst = eth_igb_recv_scattered_pkts;