update copyright date to 2013
[dpdk.git] / lib / librte_pmd_ixgbe / ixgbe_ethdev.c
index 6d80456..a17cd49 100644 (file)
@@ -1,7 +1,7 @@
 /*-
  *   BSD LICENSE
  * 
- *   Copyright(c) 2010-2012 Intel Corporation. All rights reserved.
+ *   Copyright(c) 2010-2013 Intel Corporation. All rights reserved.
  *   All rights reserved.
  * 
  *   Redistribution and use in source and binary forms, with or without 
 #define IXGBE_LINK_DOWN_CHECK_TIMEOUT 4000 /* ms */
 #define IXGBE_LINK_UP_CHECK_TIMEOUT   1000 /* ms */
 
+#define IXGBE_QUEUE_STAT_COUNTERS (sizeof(hw_stats->qprc) / sizeof(hw_stats->qprc[0]))
+
 static int eth_ixgbe_dev_init(struct eth_driver *eth_drv,
                struct rte_eth_dev *eth_dev);
-static int  ixgbe_dev_configure(struct rte_eth_dev *dev, uint16_t nb_rx_q,
-                               uint16_t nb_tx_q);
+static int  ixgbe_dev_configure(struct rte_eth_dev *dev);
 static int  ixgbe_dev_start(struct rte_eth_dev *dev);
 static void ixgbe_dev_stop(struct rte_eth_dev *dev);
 static void ixgbe_dev_close(struct rte_eth_dev *dev);
@@ -101,40 +102,69 @@ static int ixgbe_dev_link_update(struct rte_eth_dev *dev,
 static void ixgbe_dev_stats_get(struct rte_eth_dev *dev,
                                struct rte_eth_stats *stats);
 static void ixgbe_dev_stats_reset(struct rte_eth_dev *dev);
+static int ixgbe_dev_queue_stats_mapping_set(struct rte_eth_dev *eth_dev,
+                                            uint16_t queue_id,
+                                            uint8_t stat_idx,
+                                            uint8_t is_rx);
 static void ixgbe_dev_info_get(struct rte_eth_dev *dev,
                                struct rte_eth_dev_info *dev_info);
-static void ixgbe_vlan_filter_set(struct rte_eth_dev *dev,
-                                 uint16_t vlan_id,
-                                 int on);
+static int ixgbe_vlan_filter_set(struct rte_eth_dev *dev,
+               uint16_t vlan_id, int on);
+static void ixgbe_vlan_tpid_set(struct rte_eth_dev *dev, uint16_t tpid_id);
+static void ixgbe_vlan_hw_strip_bitmap_set(struct rte_eth_dev *dev, 
+               uint16_t queue, bool on);
+static void ixgbe_vlan_strip_queue_set(struct rte_eth_dev *dev, uint16_t queue,
+               int on);
+static void ixgbe_vlan_offload_set(struct rte_eth_dev *dev, int mask);
+static void ixgbe_vlan_hw_filter_enable(struct rte_eth_dev *dev);
+static void ixgbe_vlan_hw_filter_disable(struct rte_eth_dev *dev);
+static void ixgbe_vlan_hw_strip_enable_all(struct rte_eth_dev *dev);
+static void ixgbe_vlan_hw_strip_disable_all(struct rte_eth_dev *dev);
+static void ixgbe_vlan_hw_strip_enable(struct rte_eth_dev *dev, uint16_t queue);
+static void ixgbe_vlan_hw_strip_disable(struct rte_eth_dev *dev, uint16_t queue);
+static void ixgbe_vlan_hw_extend_enable(struct rte_eth_dev *dev);
+static void ixgbe_vlan_hw_extend_disable(struct rte_eth_dev *dev);
+
 static int ixgbe_dev_led_on(struct rte_eth_dev *dev);
 static int ixgbe_dev_led_off(struct rte_eth_dev *dev);
 static int  ixgbe_flow_ctrl_set(struct rte_eth_dev *dev,
-                               struct rte_eth_fc_conf *fc_conf);
+               struct rte_eth_fc_conf *fc_conf);
+static int ixgbe_priority_flow_ctrl_set(struct rte_eth_dev *dev,
+               struct rte_eth_pfc_conf *pfc_conf);
 static void ixgbe_dev_link_status_print(struct rte_eth_dev *dev);
 static int ixgbe_dev_interrupt_setup(struct rte_eth_dev *dev);
 static int ixgbe_dev_interrupt_get_status(struct rte_eth_dev *dev);
 static int ixgbe_dev_interrupt_action(struct rte_eth_dev *dev);
 static void ixgbe_dev_interrupt_handler(struct rte_intr_handle *handle,
-                                                       void *param);
+               void *param);
 static void ixgbe_dev_interrupt_delayed_handler(void *param);
 static void ixgbe_add_rar(struct rte_eth_dev *dev, struct ether_addr *mac_addr,
-                               uint32_t index, uint32_t pool);
+               uint32_t index, uint32_t pool);
 static void ixgbe_remove_rar(struct rte_eth_dev *dev, uint32_t index);
+static void ixgbe_dcb_init(struct ixgbe_hw *hw,struct ixgbe_dcb_config *dcb_config);
 
 /* For Virtual Function support */
 static int eth_ixgbevf_dev_init(struct eth_driver *eth_drv,
                struct rte_eth_dev *eth_dev);
-static int  ixgbevf_dev_configure(struct rte_eth_dev *dev, uint16_t nb_rx_q,
-               uint16_t nb_tx_q);
+static int  ixgbevf_dev_configure(struct rte_eth_dev *dev);
 static int  ixgbevf_dev_start(struct rte_eth_dev *dev);
 static void ixgbevf_dev_stop(struct rte_eth_dev *dev);
+static void ixgbevf_dev_close(struct rte_eth_dev *dev);
 static void ixgbevf_intr_disable(struct ixgbe_hw *hw);
-static void ixgbevf_dev_stats_get(struct rte_eth_dev *dev, struct rte_eth_stats *stats);
+static void ixgbevf_dev_stats_get(struct rte_eth_dev *dev,
+               struct rte_eth_stats *stats);
 static void ixgbevf_dev_stats_reset(struct rte_eth_dev *dev);
+static int ixgbevf_vlan_filter_set(struct rte_eth_dev *dev, 
+               uint16_t vlan_id, int on);
+static void ixgbevf_vlan_strip_queue_set(struct rte_eth_dev *dev,
+               uint16_t queue, int on);
+static void ixgbevf_vlan_offload_set(struct rte_eth_dev *dev, int mask);
+static void ixgbevf_set_vfta_all(struct rte_eth_dev *dev, bool on);
+
 
 /*
- *  * Define VF Stats MACRO for Non "cleared on read" register
- *   */
+ * Define VF Stats MACRO for Non "cleared on read" register
+ */
 #define UPDATE_VF_STAT(reg, last, cur)                         \
 {                                                               \
        u32 latest = IXGBE_READ_REG(hw, reg);                   \
@@ -151,13 +181,30 @@ static void ixgbevf_dev_stats_reset(struct rte_eth_dev *dev);
        last = latest;                                           \
 }
 
+#define IXGBE_SET_HWSTRIP(h, q) do{\
+               uint32_t idx = (q) / (sizeof ((h)->bitmap[0]) * NBBY); \
+               uint32_t bit = (q) % (sizeof ((h)->bitmap[0]) * NBBY); \
+               (h)->bitmap[idx] |= 1 << bit;\
+       }while(0)
+       
+#define IXGBE_CLEAR_HWSTRIP(h, q) do{\
+               uint32_t idx = (q) / (sizeof ((h)->bitmap[0]) * NBBY); \
+               uint32_t bit = (q) % (sizeof ((h)->bitmap[0]) * NBBY); \
+               (h)->bitmap[idx] &= ~(1 << bit);\
+       }while(0)
+#define IXGBE_GET_HWSTRIP(h, q, r) do{\
+               uint32_t idx = (q) / (sizeof ((h)->bitmap[0]) * NBBY); \
+               uint32_t bit = (q) % (sizeof ((h)->bitmap[0]) * NBBY); \
+               (r) = (h)->bitmap[idx] >> bit & 1;\
+       }while(0)
+
 /*
  * The set of PCI devices this driver supports
  */
 static struct rte_pci_id pci_id_ixgbe_map[] = {
 
-#undef RTE_LIBRTE_IGB_PMD
-#define RTE_PCI_DEV_ID_DECL(vend, dev) {RTE_PCI_DEVICE(vend, dev)},
+#define RTE_PCI_DEV_ID_DECL_IXGBE(vend, dev) {RTE_PCI_DEVICE(vend, dev)},
 #include "rte_pci_dev_ids.h"
 
 { .vendor_id = 0, /* sentinel */ },
@@ -168,13 +215,11 @@ static struct rte_pci_id pci_id_ixgbe_map[] = {
  * The set of PCI devices this driver supports (for 82599 VF)
  */
 static struct rte_pci_id pci_id_ixgbevf_map[] = {
-{
-       .vendor_id = PCI_VENDOR_ID_INTEL,
-       .device_id = IXGBE_DEV_ID_82599_VF,
-       .subsystem_vendor_id = PCI_ANY_ID,
-       .subsystem_device_id = PCI_ANY_ID,
-},
+
+#define RTE_PCI_DEV_ID_DECL_IXGBEVF(vend, dev) {RTE_PCI_DEVICE(vend, dev)},
+#include "rte_pci_dev_ids.h"
 { .vendor_id = 0, /* sentinel */ },
+
 };
 
 static struct eth_dev_ops ixgbe_eth_dev_ops = {
@@ -189,13 +234,20 @@ static struct eth_dev_ops ixgbe_eth_dev_ops = {
        .link_update          = ixgbe_dev_link_update,
        .stats_get            = ixgbe_dev_stats_get,
        .stats_reset          = ixgbe_dev_stats_reset,
+       .queue_stats_mapping_set = ixgbe_dev_queue_stats_mapping_set,
        .dev_infos_get        = ixgbe_dev_info_get,
        .vlan_filter_set      = ixgbe_vlan_filter_set,
+       .vlan_tpid_set        = ixgbe_vlan_tpid_set,
+       .vlan_offload_set     = ixgbe_vlan_offload_set,
+       .vlan_strip_queue_set = ixgbe_vlan_strip_queue_set,
        .rx_queue_setup       = ixgbe_dev_rx_queue_setup,
+       .rx_queue_release     = ixgbe_dev_rx_queue_release,
        .tx_queue_setup       = ixgbe_dev_tx_queue_setup,
+       .tx_queue_release     = ixgbe_dev_tx_queue_release,
        .dev_led_on           = ixgbe_dev_led_on,
        .dev_led_off          = ixgbe_dev_led_off,
        .flow_ctrl_set        = ixgbe_flow_ctrl_set,
+       .priority_flow_ctrl_set = ixgbe_priority_flow_ctrl_set,
        .mac_addr_add         = ixgbe_add_rar,
        .mac_addr_remove      = ixgbe_remove_rar,
        .fdir_add_signature_filter    = ixgbe_fdir_add_signature_filter,
@@ -220,11 +272,15 @@ static struct eth_dev_ops ixgbevf_eth_dev_ops = {
        .link_update          = ixgbe_dev_link_update,
        .stats_get            = ixgbevf_dev_stats_get,
        .stats_reset          = ixgbevf_dev_stats_reset,
-       .dev_close            = ixgbevf_dev_stop,
-
+       .dev_close            = ixgbevf_dev_close,
        .dev_infos_get        = ixgbe_dev_info_get,
+       .vlan_filter_set      = ixgbevf_vlan_filter_set,
+       .vlan_strip_queue_set = ixgbevf_vlan_strip_queue_set,
+       .vlan_offload_set     = ixgbevf_vlan_offload_set,
        .rx_queue_setup       = ixgbe_dev_rx_queue_setup,
+       .rx_queue_release     = ixgbe_dev_rx_queue_release,
        .tx_queue_setup       = ixgbe_dev_tx_queue_setup,
+       .tx_queue_release     = ixgbe_dev_tx_queue_release,
 };
 
 /**
@@ -326,12 +382,137 @@ static void
 ixgbe_reset_qstat_mappings(struct ixgbe_hw *hw)
 {
        uint32_t i;
-       for(i = 0; i != 16; i++) {
+
+       for(i = 0; i != IXGBE_NB_STAT_MAPPING_REGS; i++) {
                IXGBE_WRITE_REG(hw, IXGBE_RQSMR(i), 0);
                IXGBE_WRITE_REG(hw, IXGBE_TQSM(i), 0);
        }
 }
 
+
+static int
+ixgbe_dev_queue_stats_mapping_set(struct rte_eth_dev *eth_dev,
+                                 uint16_t queue_id,
+                                 uint8_t stat_idx,
+                                 uint8_t is_rx)
+{
+#define QSM_REG_NB_BITS_PER_QMAP_FIELD 8
+#define NB_QMAP_FIELDS_PER_QSM_REG 4
+#define QMAP_FIELD_RESERVED_BITS_MASK 0x0f
+
+       struct ixgbe_hw *hw = IXGBE_DEV_PRIVATE_TO_HW(eth_dev->data->dev_private);
+       struct ixgbe_stat_mapping_registers *stat_mappings =
+               IXGBE_DEV_PRIVATE_TO_STAT_MAPPINGS(eth_dev->data->dev_private);
+       uint32_t qsmr_mask = 0;
+       uint32_t clearing_mask = QMAP_FIELD_RESERVED_BITS_MASK;
+       uint32_t q_map;
+       uint8_t n, offset;
+
+       if ((hw->mac.type != ixgbe_mac_82599EB) && (hw->mac.type != ixgbe_mac_X540))
+               return -ENOSYS;
+
+       PMD_INIT_LOG(INFO, "Setting port %d, %s queue_id %d to stat index %d\n",
+                    (int)(eth_dev->data->port_id), is_rx ? "RX" : "TX", queue_id, stat_idx);
+
+       n = queue_id / NB_QMAP_FIELDS_PER_QSM_REG;
+       if (n >= IXGBE_NB_STAT_MAPPING_REGS) {
+               PMD_INIT_LOG(ERR, "Nb of stat mapping registers exceeded\n");
+               return -EIO;
+       }
+       offset = queue_id % NB_QMAP_FIELDS_PER_QSM_REG;
+
+       /* Now clear any previous stat_idx set */
+       clearing_mask <<= (QSM_REG_NB_BITS_PER_QMAP_FIELD * offset);
+       if (!is_rx)
+               stat_mappings->tqsm[n] &= ~clearing_mask;
+       else
+               stat_mappings->rqsmr[n] &= ~clearing_mask;
+
+       q_map = (uint32_t)stat_idx;
+       q_map &= QMAP_FIELD_RESERVED_BITS_MASK;
+       qsmr_mask = q_map << (QSM_REG_NB_BITS_PER_QMAP_FIELD * offset);
+       if (!is_rx)
+               stat_mappings->tqsm[n] |= qsmr_mask;
+       else
+               stat_mappings->rqsmr[n] |= qsmr_mask;
+
+       PMD_INIT_LOG(INFO, "Set port %d, %s queue_id %d to stat index %d\n"
+                    "%s[%d] = 0x%08x\n",
+                    (int)(eth_dev->data->port_id), is_rx ? "RX" : "TX", queue_id, stat_idx,
+                    is_rx ? "RQSMR" : "TQSM",n, is_rx ? stat_mappings->rqsmr[n] : stat_mappings->tqsm[n]);
+
+       /* Now write the mapping in the appropriate register */
+       if (is_rx) {
+               PMD_INIT_LOG(INFO, "Write 0x%x to RX IXGBE stat mapping reg:%d\n",
+                            stat_mappings->rqsmr[n], n);
+               IXGBE_WRITE_REG(hw, IXGBE_RQSMR(n), stat_mappings->rqsmr[n]);
+       }
+       else {
+               PMD_INIT_LOG(INFO, "Write 0x%x to TX IXGBE stat mapping reg:%d\n",
+                            stat_mappings->tqsm[n], n);
+               IXGBE_WRITE_REG(hw, IXGBE_TQSM(n), stat_mappings->tqsm[n]);
+       }
+       return 0;
+}
+
+static void
+ixgbe_restore_statistics_mapping(struct rte_eth_dev * dev)
+{
+       struct ixgbe_stat_mapping_registers *stat_mappings =
+               IXGBE_DEV_PRIVATE_TO_STAT_MAPPINGS(dev->data->dev_private);
+       struct ixgbe_hw *hw = IXGBE_DEV_PRIVATE_TO_HW(dev->data->dev_private);
+       int i;
+
+       /* write whatever was in stat mapping table to the NIC */
+       for (i = 0; i < IXGBE_NB_STAT_MAPPING_REGS; i++) {
+               /* rx */
+               IXGBE_WRITE_REG(hw, IXGBE_RQSMR(i), stat_mappings->rqsmr[i]);
+
+               /* tx */
+               IXGBE_WRITE_REG(hw, IXGBE_TQSM(i), stat_mappings->tqsm[i]);
+       }
+}
+
+static void
+ixgbe_dcb_init(struct ixgbe_hw *hw,struct ixgbe_dcb_config *dcb_config)
+{
+       uint8_t i;
+       struct ixgbe_dcb_tc_config *tc;
+       int dcb_max_tc = IXGBE_DCB_MAX_TRAFFIC_CLASS;
+
+       dcb_config->num_tcs.pg_tcs = dcb_max_tc;
+       dcb_config->num_tcs.pfc_tcs = dcb_max_tc;
+       for (i = 0; i < dcb_max_tc; i++) {
+               tc = &dcb_config->tc_config[i];
+               tc->path[IXGBE_DCB_TX_CONFIG].bwg_id = i;
+               tc->path[IXGBE_DCB_TX_CONFIG].bwg_percent = 100/dcb_max_tc + (i & 1);
+               tc->path[IXGBE_DCB_RX_CONFIG].bwg_id = i;
+               tc->path[IXGBE_DCB_RX_CONFIG].bwg_percent = 100/dcb_max_tc + (i & 1);
+               tc->pfc = ixgbe_dcb_pfc_disabled;
+       }
+
+       /* Initialize default user to priority mapping, UPx->TC0 */
+       tc = &dcb_config->tc_config[0];
+       tc->path[IXGBE_DCB_TX_CONFIG].up_to_tc_bitmap = 0xFF;
+       tc->path[IXGBE_DCB_RX_CONFIG].up_to_tc_bitmap = 0xFF;
+       for (i = 0; i< IXGBE_DCB_MAX_BW_GROUP; i++) {
+               dcb_config->bw_percentage[IXGBE_DCB_TX_CONFIG][i] = 100;
+               dcb_config->bw_percentage[IXGBE_DCB_RX_CONFIG][i] = 100;
+       }
+       dcb_config->rx_pba_cfg = ixgbe_dcb_pba_equal;
+       dcb_config->pfc_mode_enable = false;
+       dcb_config->vt_mode = true;
+       dcb_config->round_robin_enable = false;
+       /* support all DCB capabilities in 82599 */
+       dcb_config->support.capabilities = 0xFF;
+
+       /*we only support 4 Tcs for X540*/              
+       if (hw->mac.type == ixgbe_mac_X540) {
+               dcb_config->num_tcs.pg_tcs = 4;
+               dcb_config->num_tcs.pfc_tcs = 4;
+       }
+} 
+
 /*
  * This function is based on code in ixgbe_attach() in ixgbe/ixgbe.c.
  * It returns 0 on success.
@@ -345,6 +526,10 @@ eth_ixgbe_dev_init(__attribute__((unused)) struct eth_driver *eth_drv,
                IXGBE_DEV_PRIVATE_TO_HW(eth_dev->data->dev_private);
        struct ixgbe_vfta * shadow_vfta =
                IXGBE_DEV_PRIVATE_TO_VFTA(eth_dev->data->dev_private);
+       struct ixgbe_hwstrip *hwstrip = 
+               IXGBE_DEV_PRIVATE_TO_HWSTRIP_BITMAP(eth_dev->data->dev_private);
+       struct ixgbe_dcb_config *dcb_config =
+               IXGBE_DEV_PRIVATE_TO_DCB_CFG(eth_dev->data->dev_private);
        uint32_t ctrl_ext;
        uint16_t csum;
        int diag, i;
@@ -377,13 +562,17 @@ eth_ixgbe_dev_init(__attribute__((unused)) struct eth_driver *eth_drv,
                return -EIO;
        }
 
+       /* Initialize DCB configuration*/
+       memset(dcb_config, 0, sizeof(struct ixgbe_dcb_config));
+       ixgbe_dcb_init(hw,dcb_config);
        /* Get Hardware Flow Control setting */
        hw->fc.requested_mode = ixgbe_fc_full;
        hw->fc.current_mode = ixgbe_fc_full;
        hw->fc.pause_time = IXGBE_FC_PAUSE;
-       hw->fc.low_water = IXGBE_FC_LO;
-       for (i = 0; i < MAX_TRAFFIC_CLASS; i++)
+       for (i = 0; i < IXGBE_DCB_MAX_TRAFFIC_CLASS; i++) {
+               hw->fc.low_water[i] = IXGBE_FC_LO;
                hw->fc.high_water[i] = IXGBE_FC_HI;
+       }
        hw->fc.send_xon = 1;
 
        ixgbe_disable_intr(hw);
@@ -447,6 +636,9 @@ eth_ixgbe_dev_init(__attribute__((unused)) struct eth_driver *eth_drv,
        /* initialize the vfta */
        memset(shadow_vfta, 0, sizeof(*shadow_vfta));
 
+       /* initialize the hw strip bitmap*/
+       memset(hwstrip, 0, sizeof(*hwstrip));
+
        /* let hardware know driver is loaded */
        ctrl_ext = IXGBE_READ_REG(hw, IXGBE_CTRL_EXT);
        ctrl_ext |= IXGBE_CTRL_EXT_DRV_LOAD;
@@ -481,6 +673,10 @@ eth_ixgbevf_dev_init(__attribute__((unused)) struct eth_driver *eth_drv,
        struct rte_pci_device *pci_dev;
        struct ixgbe_hw *hw = IXGBE_DEV_PRIVATE_TO_HW(eth_dev->data->dev_private);
        int diag;
+       struct ixgbe_vfta * shadow_vfta =
+               IXGBE_DEV_PRIVATE_TO_VFTA(eth_dev->data->dev_private);
+       struct ixgbe_hwstrip *hwstrip = 
+               IXGBE_DEV_PRIVATE_TO_HWSTRIP_BITMAP(eth_dev->data->dev_private);
 
        PMD_INIT_LOG(DEBUG, "eth_ixgbevf_dev_init");
 
@@ -491,6 +687,12 @@ eth_ixgbevf_dev_init(__attribute__((unused)) struct eth_driver *eth_drv,
        hw->vendor_id = pci_dev->id.vendor_id;
        hw->hw_addr = (void *)pci_dev->mem_resource.addr;
 
+       /* initialize the vfta */
+       memset(shadow_vfta, 0, sizeof(*shadow_vfta));
+
+       /* initialize the hw strip bitmap*/
+       memset(hwstrip, 0, sizeof(*hwstrip));
+
        /* Initialize the shared code */
        diag = ixgbe_init_shared_code(hw);
        if (diag != IXGBE_SUCCESS) {
@@ -506,6 +708,10 @@ eth_ixgbevf_dev_init(__attribute__((unused)) struct eth_driver *eth_drv,
 
        hw->mac.num_rar_entries = hw->mac.max_rx_queues;
        diag = hw->mac.ops.reset_hw(hw);
+       if (diag != IXGBE_SUCCESS) {
+               PMD_INIT_LOG(ERR, "VF Initialization Failure: %d", diag);
+               return (diag);
+       }
 
        /* Allocate memory for storing MAC addresses */
        eth_dev->data->mac_addrs = rte_zmalloc("ixgbevf", ETHER_ADDR_LEN *
@@ -528,7 +734,7 @@ eth_ixgbevf_dev_init(__attribute__((unused)) struct eth_driver *eth_drv,
 
                default:
                        PMD_INIT_LOG(ERR, "VF Initialization Failure: %d", diag);
-                       return (diag);
+                       return (-EIO);
        }
 
        PMD_INIT_LOG(DEBUG, "\nport %d vendorID=0x%x deviceID=0x%x mac.type=%s\n",
@@ -589,7 +795,7 @@ rte_ixgbevf_pmd_init(void)
        return (0);
 }
 
-static void
+static int
 ixgbe_vlan_filter_set(struct rte_eth_dev *dev, uint16_t vlan_id, int on)
 {
        struct ixgbe_hw *hw =
@@ -611,16 +817,35 @@ ixgbe_vlan_filter_set(struct rte_eth_dev *dev, uint16_t vlan_id, int on)
 
        /* update local VFTA copy */
        shadow_vfta->vfta[vid_idx] = vfta;
+
+       return 0;
+}
+
+static void
+ixgbe_vlan_strip_queue_set(struct rte_eth_dev *dev, uint16_t queue, int on)
+{
+       if (on)
+               ixgbe_vlan_hw_strip_enable(dev, queue);
+       else
+               ixgbe_vlan_hw_strip_disable(dev, queue);
+}
+
+static void
+ixgbe_vlan_tpid_set(struct rte_eth_dev *dev, uint16_t tpid)
+{
+       struct ixgbe_hw *hw =
+               IXGBE_DEV_PRIVATE_TO_HW(dev->data->dev_private);
+
+       /* Only the high 16-bits is valid */
+       IXGBE_WRITE_REG(hw, IXGBE_EXVET, tpid << 16);
 }
 
 static void
-ixgbe_vlan_hw_support_disable(struct rte_eth_dev *dev)
+ixgbe_vlan_hw_filter_disable(struct rte_eth_dev *dev)
 {
        struct ixgbe_hw *hw =
                IXGBE_DEV_PRIVATE_TO_HW(dev->data->dev_private);
        uint32_t vlnctrl;
-       uint32_t rxdctl;
-       uint16_t i;
 
        PMD_INIT_FUNC_TRACE();
 
@@ -628,28 +853,17 @@ ixgbe_vlan_hw_support_disable(struct rte_eth_dev *dev)
        vlnctrl = IXGBE_READ_REG(hw, IXGBE_VLNCTRL);
        vlnctrl &= ~IXGBE_VLNCTRL_VFE;
 
-       if (hw->mac.type == ixgbe_mac_82598EB)
-               vlnctrl &= ~IXGBE_VLNCTRL_VME;
-       else {
-               /* On 82599 the VLAN enable is per/queue in RXDCTL */
-               for (i = 0; i < dev->data->nb_rx_queues; i++) {
-                       rxdctl = IXGBE_READ_REG(hw, IXGBE_RXDCTL(i));
-                       rxdctl &= ~IXGBE_RXDCTL_VME;
-                       IXGBE_WRITE_REG(hw, IXGBE_RXDCTL(i), rxdctl);
-               }
-       }
        IXGBE_WRITE_REG(hw, IXGBE_VLNCTRL, vlnctrl);
 }
 
 static void
-ixgbe_vlan_hw_support_enable(struct rte_eth_dev *dev)
+ixgbe_vlan_hw_filter_enable(struct rte_eth_dev *dev)
 {
        struct ixgbe_hw *hw =
                IXGBE_DEV_PRIVATE_TO_HW(dev->data->dev_private);
        struct ixgbe_vfta * shadow_vfta =
                IXGBE_DEV_PRIVATE_TO_VFTA(dev->data->dev_private);
        uint32_t vlnctrl;
-       uint32_t rxdctl;
        uint16_t i;
 
        PMD_INIT_FUNC_TRACE();
@@ -659,16 +873,6 @@ ixgbe_vlan_hw_support_enable(struct rte_eth_dev *dev)
        vlnctrl &= ~IXGBE_VLNCTRL_CFIEN;
        vlnctrl |= IXGBE_VLNCTRL_VFE;
 
-       if (hw->mac.type == ixgbe_mac_82598EB)
-               vlnctrl |= IXGBE_VLNCTRL_VME;
-       else {
-               /* On 82599 the VLAN enable is per/queue in RXDCTL */
-               for (i = 0; i < dev->data->nb_rx_queues; i++) {
-                       rxdctl = IXGBE_READ_REG(hw, IXGBE_RXDCTL(i));
-                       rxdctl |= IXGBE_RXDCTL_VME;
-                       IXGBE_WRITE_REG(hw, IXGBE_RXDCTL(i), rxdctl);
-               }
-       }
        IXGBE_WRITE_REG(hw, IXGBE_VLNCTRL, vlnctrl);
 
        /* write whatever is in local vfta copy */
@@ -676,33 +880,204 @@ ixgbe_vlan_hw_support_enable(struct rte_eth_dev *dev)
                IXGBE_WRITE_REG(hw, IXGBE_VFTA(i), shadow_vfta->vfta[i]);
 }
 
-static int
-ixgbe_dev_configure(struct rte_eth_dev *dev, uint16_t nb_rx_q, uint16_t nb_tx_q)
+static void 
+ixgbe_vlan_hw_strip_bitmap_set(struct rte_eth_dev *dev, uint16_t queue, bool on)
 {
-       struct ixgbe_interrupt *intr =
-               IXGBE_DEV_PRIVATE_TO_INTR(dev->data->dev_private);
-       int diag;
+       struct ixgbe_hwstrip *hwstrip = 
+               IXGBE_DEV_PRIVATE_TO_HWSTRIP_BITMAP(dev->data->dev_private);
+
+       if(queue >= IXGBE_MAX_RX_QUEUE_NUM)
+               return;
+
+       if (on)
+               IXGBE_SET_HWSTRIP(hwstrip, queue);
+       else
+               IXGBE_CLEAR_HWSTRIP(hwstrip, queue);
+}
+
+static void
+ixgbe_vlan_hw_strip_disable(struct rte_eth_dev *dev, uint16_t queue)
+{
+       struct ixgbe_hw *hw =
+               IXGBE_DEV_PRIVATE_TO_HW(dev->data->dev_private);
+       uint32_t ctrl;
 
        PMD_INIT_FUNC_TRACE();
 
-       /* Allocate the array of pointers to RX queue structures */
-       diag = ixgbe_dev_rx_queue_alloc(dev, nb_rx_q);
-       if (diag != 0) {
-               PMD_INIT_LOG(ERR, "ethdev port_id=%d allocation of array of %d"
-                            "pointers to RX queues failed", dev->data->port_id,
-                            nb_rx_q);
-               return diag;
+       if (hw->mac.type == ixgbe_mac_82598EB) {
+               /* No queue level support */
+               PMD_INIT_LOG(INFO, "82598EB not support queue level hw strip");
+               return;
        }
+       else {
+               /* Other 10G NIC, the VLAN strip can be setup per queue in RXDCTL */
+               ctrl = IXGBE_READ_REG(hw, IXGBE_RXDCTL(queue));
+               ctrl &= ~IXGBE_RXDCTL_VME;
+               IXGBE_WRITE_REG(hw, IXGBE_RXDCTL(queue), ctrl);
+       }
+       /* record those setting for HW strip per queue */
+       ixgbe_vlan_hw_strip_bitmap_set(dev, queue, 0);
+}
 
-       /* Allocate the array of pointers to TX queue structures */
-       diag = ixgbe_dev_tx_queue_alloc(dev, nb_tx_q);
-       if (diag != 0) {
-               PMD_INIT_LOG(ERR, "ethdev port_id=%d allocation of array of %d"
-                            "pointers to TX queues failed", dev->data->port_id,
-                            nb_tx_q);
-               return diag;
+static void
+ixgbe_vlan_hw_strip_enable(struct rte_eth_dev *dev, uint16_t queue)
+{
+       struct ixgbe_hw *hw =
+               IXGBE_DEV_PRIVATE_TO_HW(dev->data->dev_private);
+       uint32_t ctrl;
+
+       PMD_INIT_FUNC_TRACE();
+
+       if (hw->mac.type == ixgbe_mac_82598EB) {
+               /* No queue level supported */
+               PMD_INIT_LOG(INFO, "82598EB not support queue level hw strip");
+               return;
+       }
+       else {
+               /* Other 10G NIC, the VLAN strip can be setup per queue in RXDCTL */
+               ctrl = IXGBE_READ_REG(hw, IXGBE_RXDCTL(queue));
+               ctrl |= IXGBE_RXDCTL_VME;
+               IXGBE_WRITE_REG(hw, IXGBE_RXDCTL(queue), ctrl);
+       }
+       /* record those setting for HW strip per queue */
+       ixgbe_vlan_hw_strip_bitmap_set(dev, queue, 1);
+}
+
+static void
+ixgbe_vlan_hw_strip_disable_all(struct rte_eth_dev *dev)
+{
+       struct ixgbe_hw *hw =
+               IXGBE_DEV_PRIVATE_TO_HW(dev->data->dev_private);
+       uint32_t ctrl;
+       uint16_t i;
+
+       PMD_INIT_FUNC_TRACE();
+
+       if (hw->mac.type == ixgbe_mac_82598EB) {
+               ctrl = IXGBE_READ_REG(hw, IXGBE_VLNCTRL);
+               ctrl &= ~IXGBE_VLNCTRL_VME;
+               IXGBE_WRITE_REG(hw, IXGBE_VLNCTRL, ctrl);
+       }
+       else {
+               /* Other 10G NIC, the VLAN strip can be setup per queue in RXDCTL */
+               for (i = 0; i < dev->data->nb_rx_queues; i++) {
+                       ctrl = IXGBE_READ_REG(hw, IXGBE_RXDCTL(i));
+                       ctrl &= ~IXGBE_RXDCTL_VME;
+                       IXGBE_WRITE_REG(hw, IXGBE_RXDCTL(i), ctrl);
+
+                       /* record those setting for HW strip per queue */
+                       ixgbe_vlan_hw_strip_bitmap_set(dev, i, 0);
+               }
+       }
+}
+
+static void
+ixgbe_vlan_hw_strip_enable_all(struct rte_eth_dev *dev)
+{
+       struct ixgbe_hw *hw =
+               IXGBE_DEV_PRIVATE_TO_HW(dev->data->dev_private);
+       uint32_t ctrl;
+       uint16_t i;
+
+       PMD_INIT_FUNC_TRACE();
+
+       if (hw->mac.type == ixgbe_mac_82598EB) {
+               ctrl = IXGBE_READ_REG(hw, IXGBE_VLNCTRL);
+               ctrl |= IXGBE_VLNCTRL_VME;
+               IXGBE_WRITE_REG(hw, IXGBE_VLNCTRL, ctrl);
+       }
+       else {
+               /* Other 10G NIC, the VLAN strip can be setup per queue in RXDCTL */
+               for (i = 0; i < dev->data->nb_rx_queues; i++) {
+                       ctrl = IXGBE_READ_REG(hw, IXGBE_RXDCTL(i));
+                       ctrl |= IXGBE_RXDCTL_VME;
+                       IXGBE_WRITE_REG(hw, IXGBE_RXDCTL(i), ctrl);
+
+                       /* record those setting for HW strip per queue */
+                       ixgbe_vlan_hw_strip_bitmap_set(dev, i, 1);                      
+               }
+       }
+}
+
+static void
+ixgbe_vlan_hw_extend_disable(struct rte_eth_dev *dev)
+{
+       struct ixgbe_hw *hw =
+               IXGBE_DEV_PRIVATE_TO_HW(dev->data->dev_private);
+       uint32_t ctrl;
+
+       PMD_INIT_FUNC_TRACE();
+
+       /* DMATXCTRL: Geric Double VLAN Disable */
+       ctrl = IXGBE_READ_REG(hw, IXGBE_DMATXCTL);
+       ctrl &= ~IXGBE_DMATXCTL_GDV;
+       IXGBE_WRITE_REG(hw, IXGBE_DMATXCTL, ctrl);
+
+       /* CTRL_EXT: Global Double VLAN Disable */
+       ctrl = IXGBE_READ_REG(hw, IXGBE_CTRL_EXT);
+       ctrl &= ~IXGBE_EXTENDED_VLAN;
+       IXGBE_WRITE_REG(hw, IXGBE_CTRL_EXT, ctrl);
+
+}
+
+static void
+ixgbe_vlan_hw_extend_enable(struct rte_eth_dev *dev)
+{
+       struct ixgbe_hw *hw =
+               IXGBE_DEV_PRIVATE_TO_HW(dev->data->dev_private);
+       uint32_t ctrl;
+
+       PMD_INIT_FUNC_TRACE();
+
+       /* DMATXCTRL: Geric Double VLAN Enable */
+       ctrl  = IXGBE_READ_REG(hw, IXGBE_DMATXCTL);
+       ctrl |= IXGBE_DMATXCTL_GDV;
+       IXGBE_WRITE_REG(hw, IXGBE_DMATXCTL, ctrl);
+
+       /* CTRL_EXT: Global Double VLAN Enable */
+       ctrl  = IXGBE_READ_REG(hw, IXGBE_CTRL_EXT);
+       ctrl |= IXGBE_EXTENDED_VLAN;
+       IXGBE_WRITE_REG(hw, IXGBE_CTRL_EXT, ctrl);
+
+       /*
+        * VET EXT field in the EXVET register = 0x8100 by default
+        * So no need to change. Same to VT field of DMATXCTL register
+        */
+}
+
+static void
+ixgbe_vlan_offload_set(struct rte_eth_dev *dev, int mask)
+{
+       if(mask & ETH_VLAN_STRIP_MASK){
+               if (dev->data->dev_conf.rxmode.hw_vlan_strip)
+                       ixgbe_vlan_hw_strip_enable_all(dev);
+               else
+                       ixgbe_vlan_hw_strip_disable_all(dev);
+       }
+
+       if(mask & ETH_VLAN_FILTER_MASK){
+               if (dev->data->dev_conf.rxmode.hw_vlan_filter)
+                       ixgbe_vlan_hw_filter_enable(dev);
+               else
+                       ixgbe_vlan_hw_filter_disable(dev);
        }
 
+       if(mask & ETH_VLAN_EXTEND_MASK){
+               if (dev->data->dev_conf.rxmode.hw_vlan_extend)
+                       ixgbe_vlan_hw_extend_enable(dev);
+               else
+                       ixgbe_vlan_hw_extend_disable(dev);
+       }
+}
+
+static int
+ixgbe_dev_configure(struct rte_eth_dev *dev)
+{
+       struct ixgbe_interrupt *intr =
+               IXGBE_DEV_PRIVATE_TO_INTR(dev->data->dev_private);
+
+       PMD_INIT_FUNC_TRACE();
+
        /* set flag to update link status after init */
        intr->flags |= IXGBE_FLAG_NEED_LINK_UPDATE;
 
@@ -720,7 +1095,8 @@ ixgbe_dev_start(struct rte_eth_dev *dev)
                IXGBE_DEV_PRIVATE_TO_HW(dev->data->dev_private);
        int err, link_up = 0, negotiate = 0;
        uint32_t speed = 0;
-
+       int mask = 0;
+       
        PMD_INIT_FUNC_TRACE();
 
        /* IXGBE devices don't support half duplex */
@@ -747,7 +1123,7 @@ ixgbe_dev_start(struct rte_eth_dev *dev)
        err = ixgbe_dev_rx_init(dev);
        if (err) {
                PMD_INIT_LOG(ERR, "Unable to initialize RX hardware\n");
-               return err;
+               goto error;
        }
 
        ixgbe_dev_rxtx_start(dev);
@@ -791,7 +1167,7 @@ ixgbe_dev_start(struct rte_eth_dev *dev)
        default:
                PMD_INIT_LOG(ERR, "Invalid link_speed (%u) for port %u\n",
                                dev->data->dev_conf.link_speed, dev->data->port_id);
-               return -EINVAL;
+               goto error;
        }
 
        err = ixgbe_setup_link(hw, speed, negotiate, link_up);
@@ -805,14 +1181,12 @@ ixgbe_dev_start(struct rte_eth_dev *dev)
                        goto error;
        }
 
-       /*
-        * If VLAN filtering is enabled, set up VLAN tag offload and filtering
-        * and restore VFTA.
-        */
-       if (dev->data->dev_conf.rxmode.hw_vlan_filter)
-               ixgbe_vlan_hw_support_enable(dev);
-       else
-               ixgbe_vlan_hw_support_disable(dev);
+       mask = ETH_VLAN_STRIP_MASK | ETH_VLAN_FILTER_MASK | \
+               ETH_VLAN_EXTEND_MASK;
+       ixgbe_vlan_offload_set(dev, mask);
+       
+       /* Configure DCB hw */
+       ixgbe_configure_dcb(dev); 
 
        if (dev->data->dev_conf.fdir_conf.mode != RTE_FDIR_MODE_NONE) {
                err = ixgbe_fdir_configure(dev);
@@ -820,10 +1194,13 @@ ixgbe_dev_start(struct rte_eth_dev *dev)
                        goto error;
        }
 
+       ixgbe_restore_statistics_mapping(dev);
+
        return (0);
 
 error:
        PMD_INIT_LOG(ERR, "failure in ixgbe_dev_start(): %d", err);
+       ixgbe_dev_clear_queues(dev);
        return -EIO;
 }
 
@@ -927,7 +1304,7 @@ ixgbe_dev_stats_get(struct rte_eth_dev *dev, struct rte_eth_stats *stats)
                hw_stats->pxon2offc[i] +=
                    IXGBE_READ_REG(hw, IXGBE_PXON2OFFCNT(i));
        }
-       for (i = 0; i < 16; i++) {
+       for (i = 0; i < IXGBE_QUEUE_STAT_COUNTERS; i++) {
                hw_stats->qprc[i] += IXGBE_READ_REG(hw, IXGBE_QPRC(i));
                hw_stats->qptc[i] += IXGBE_READ_REG(hw, IXGBE_QPTC(i));
                hw_stats->qbrc[i] += IXGBE_READ_REG(hw, IXGBE_QBRC_L(i));
@@ -1034,6 +1411,14 @@ ixgbe_dev_stats_get(struct rte_eth_dev *dev, struct rte_eth_stats *stats)
        stats->obytes = hw_stats->gotc;
        stats->imcasts = hw_stats->mprc;
 
+       for (i = 0; i < IXGBE_QUEUE_STAT_COUNTERS; i++) {
+               stats->q_ipackets[i] = hw_stats->qprc[i];
+               stats->q_opackets[i] = hw_stats->qptc[i];
+               stats->q_ibytes[i] = hw_stats->qbrc[i];
+               stats->q_obytes[i] = hw_stats->qbtc[i];
+               stats->q_errors[i] = hw_stats->qprdc[i];
+       }
+
        /* Rx Errors */
        stats->ierrors = total_missed_rx + hw_stats->crcerrs +
                hw_stats->rlec;
@@ -1492,10 +1877,10 @@ ixgbe_flow_ctrl_set(struct rte_eth_dev *dev, struct rte_eth_fc_conf *fc_conf)
        hw->fc.requested_mode = rte_fcmode_2_ixgbe_fcmode[fc_conf->mode];
        hw->fc.pause_time     = fc_conf->pause_time;
        hw->fc.high_water[0]  = fc_conf->high_water;
-       hw->fc.low_water      = fc_conf->low_water;
+       hw->fc.low_water[0]   = fc_conf->low_water;
        hw->fc.send_xon       = fc_conf->send_xon;
 
-       err = ixgbe_fc_enable(hw, 0);
+       err = ixgbe_fc_enable(hw);
        /* Not negotiated is not an error case */
        if ((err == IXGBE_SUCCESS) || (err == IXGBE_ERR_FC_NOT_NEGOTIATED)) {
                return 0;
@@ -1505,6 +1890,211 @@ ixgbe_flow_ctrl_set(struct rte_eth_dev *dev, struct rte_eth_fc_conf *fc_conf)
        return -EIO;
 }
 
+/**
+ *  ixgbe_pfc_enable_generic - Enable flow control
+ *  @hw: pointer to hardware structure
+ *  @tc_num: traffic class number
+ *  Enable flow control according to the current settings.
+ */
+static int 
+ixgbe_dcb_pfc_enable_generic(struct ixgbe_hw *hw,uint8_t tc_num)
+{
+       int ret_val = 0;
+       uint32_t mflcn_reg, fccfg_reg;
+       uint32_t reg;
+       uint32_t fcrtl, fcrth;
+       uint8_t i;
+       uint8_t nb_rx_en;
+       
+       /* Validate the water mark configuration */
+       if (!hw->fc.pause_time) {
+               ret_val = IXGBE_ERR_INVALID_LINK_SETTINGS;
+               goto out;
+       }
+
+       /* Low water mark of zero causes XOFF floods */
+       if (hw->fc.current_mode & ixgbe_fc_tx_pause) {
+                /* High/Low water can not be 0 */
+               if( (!hw->fc.high_water[tc_num])|| (!hw->fc.low_water[tc_num])) {
+                       PMD_INIT_LOG(ERR,"Invalid water mark configuration\n");
+                       ret_val = IXGBE_ERR_INVALID_LINK_SETTINGS;
+                       goto out;
+               }
+               if(hw->fc.low_water[tc_num] >= hw->fc.high_water[tc_num]) {
+                       PMD_INIT_LOG(ERR,"Invalid water mark configuration\n");
+                       ret_val = IXGBE_ERR_INVALID_LINK_SETTINGS;
+                       goto out;
+               }
+       }
+       /* Negotiate the fc mode to use */
+       ixgbe_fc_autoneg(hw);
+
+       /* Disable any previous flow control settings */
+       mflcn_reg = IXGBE_READ_REG(hw, IXGBE_MFLCN);
+       mflcn_reg &= ~(IXGBE_MFLCN_RPFCE_SHIFT | IXGBE_MFLCN_RFCE|IXGBE_MFLCN_RPFCE);
+
+       fccfg_reg = IXGBE_READ_REG(hw, IXGBE_FCCFG);
+       fccfg_reg &= ~(IXGBE_FCCFG_TFCE_802_3X | IXGBE_FCCFG_TFCE_PRIORITY);
+
+       switch (hw->fc.current_mode) {
+       case ixgbe_fc_none:
+               /*
+                * If the count of enabled RX Priority Flow control >1,
+                * and the TX pause can not be disabled 
+                */
+               nb_rx_en = 0;
+               for (i =0; i < IXGBE_DCB_MAX_TRAFFIC_CLASS; i++) {
+                       reg = IXGBE_READ_REG(hw, IXGBE_FCRTH_82599(i));
+                       if (reg & IXGBE_FCRTH_FCEN)
+                               nb_rx_en++;
+               }
+               if (nb_rx_en > 1)
+                       fccfg_reg |=IXGBE_FCCFG_TFCE_PRIORITY;
+               break;
+       case ixgbe_fc_rx_pause:
+               /*
+                * Rx Flow control is enabled and Tx Flow control is
+                * disabled by software override. Since there really
+                * isn't a way to advertise that we are capable of RX
+                * Pause ONLY, we will advertise that we support both
+                * symmetric and asymmetric Rx PAUSE.  Later, we will
+                * disable the adapter's ability to send PAUSE frames.
+                */
+               mflcn_reg |= IXGBE_MFLCN_RPFCE;
+               /*
+                * If the count of enabled RX Priority Flow control >1,
+                * and the TX pause can not be disabled
+                */
+               nb_rx_en = 0;
+               for (i =0; i < IXGBE_DCB_MAX_TRAFFIC_CLASS; i++) {
+                       reg = IXGBE_READ_REG(hw, IXGBE_FCRTH_82599(i));
+                       if (reg & IXGBE_FCRTH_FCEN)
+                               nb_rx_en++;
+               }
+               if (nb_rx_en > 1)
+                       fccfg_reg |=IXGBE_FCCFG_TFCE_PRIORITY;
+               break;
+       case ixgbe_fc_tx_pause:
+               /*
+                * Tx Flow control is enabled, and Rx Flow control is
+                * disabled by software override.
+                */
+               fccfg_reg |=IXGBE_FCCFG_TFCE_PRIORITY;
+               break;
+       case ixgbe_fc_full:
+               /* Flow control (both Rx and Tx) is enabled by SW override. */
+               mflcn_reg |= IXGBE_MFLCN_RPFCE;
+               fccfg_reg |= IXGBE_FCCFG_TFCE_PRIORITY;
+               break;
+       default:
+               DEBUGOUT("Flow control param set incorrectly\n");
+               ret_val = IXGBE_ERR_CONFIG;
+               goto out;
+               break;
+       }
+
+       /* Set 802.3x based flow control settings. */
+       mflcn_reg |= IXGBE_MFLCN_DPF;
+       IXGBE_WRITE_REG(hw, IXGBE_MFLCN, mflcn_reg);
+       IXGBE_WRITE_REG(hw, IXGBE_FCCFG, fccfg_reg);
+
+       /* Set up and enable Rx high/low water mark thresholds, enable XON. */
+       if ((hw->fc.current_mode & ixgbe_fc_tx_pause) &&
+               hw->fc.high_water[tc_num]) {
+               fcrtl = (hw->fc.low_water[tc_num] << 10) | IXGBE_FCRTL_XONE;
+               IXGBE_WRITE_REG(hw, IXGBE_FCRTL_82599(tc_num), fcrtl);
+               fcrth = (hw->fc.high_water[tc_num] << 10) | IXGBE_FCRTH_FCEN;
+       } else {
+               IXGBE_WRITE_REG(hw, IXGBE_FCRTL_82599(tc_num), 0);
+               /*
+                * In order to prevent Tx hangs when the internal Tx
+                * switch is enabled we must set the high water mark
+                * to the maximum FCRTH value.  This allows the Tx
+                * switch to function even under heavy Rx workloads.
+                */
+               fcrth = IXGBE_READ_REG(hw, IXGBE_RXPBSIZE(tc_num)) - 32;
+       }
+       IXGBE_WRITE_REG(hw, IXGBE_FCRTH_82599(tc_num), fcrth);
+
+       /* Configure pause time (2 TCs per register) */
+       reg = hw->fc.pause_time * 0x00010001;
+       for (i = 0; i < (IXGBE_DCB_MAX_TRAFFIC_CLASS / 2); i++)
+               IXGBE_WRITE_REG(hw, IXGBE_FCTTV(i), reg);
+
+       /* Configure flow control refresh threshold value */
+       IXGBE_WRITE_REG(hw, IXGBE_FCRTV, hw->fc.pause_time / 2);
+
+out:
+       return ret_val;
+}
+
+static int 
+ixgbe_dcb_pfc_enable(struct rte_eth_dev *dev,uint8_t tc_num)
+{
+       struct ixgbe_hw *hw = IXGBE_DEV_PRIVATE_TO_HW(dev->data->dev_private);
+       int32_t ret_val = IXGBE_NOT_IMPLEMENTED;
+
+       if(hw->mac.type != ixgbe_mac_82598EB) {
+               ret_val = ixgbe_dcb_pfc_enable_generic(hw,tc_num);
+       }
+       return ret_val;
+}
+
+static int 
+ixgbe_priority_flow_ctrl_set(struct rte_eth_dev *dev, struct rte_eth_pfc_conf *pfc_conf)
+{
+       int err;
+       uint32_t rx_buf_size;
+       uint32_t max_high_water;
+       uint8_t tc_num;
+       uint8_t  map[IXGBE_DCB_MAX_USER_PRIORITY] = { 0 };
+       struct ixgbe_hw *hw =
+                IXGBE_DEV_PRIVATE_TO_HW(dev->data->dev_private);
+       struct ixgbe_dcb_config *dcb_config =
+                IXGBE_DEV_PRIVATE_TO_DCB_CFG(dev->data->dev_private);
+       
+       enum ixgbe_fc_mode rte_fcmode_2_ixgbe_fcmode[] = {
+               ixgbe_fc_none,
+               ixgbe_fc_rx_pause,
+               ixgbe_fc_tx_pause,
+               ixgbe_fc_full
+       };
+       
+       PMD_INIT_FUNC_TRACE();
+       
+       ixgbe_dcb_unpack_map_cee(dcb_config, IXGBE_DCB_RX_CONFIG, map);
+       tc_num = map[pfc_conf->priority];
+       rx_buf_size = IXGBE_READ_REG(hw, IXGBE_RXPBSIZE(tc_num));
+       PMD_INIT_LOG(DEBUG, "Rx packet buffer size = 0x%x \n", rx_buf_size);
+       /*
+        * At least reserve one Ethernet frame for watermark
+        * high_water/low_water in kilo bytes for ixgbe
+        */
+       max_high_water = (rx_buf_size - ETHER_MAX_LEN) >> IXGBE_RXPBSIZE_SHIFT;
+       if ((pfc_conf->fc.high_water > max_high_water) ||
+               (pfc_conf->fc.high_water <= pfc_conf->fc.low_water)) {
+               PMD_INIT_LOG(ERR, "Invalid high/low water setup value in KB\n");
+               PMD_INIT_LOG(ERR, "High_water must <=  0x%x\n", max_high_water);
+               return (-EINVAL);
+       }
+
+       hw->fc.requested_mode = rte_fcmode_2_ixgbe_fcmode[pfc_conf->fc.mode];
+       hw->fc.pause_time = pfc_conf->fc.pause_time;
+       hw->fc.send_xon = pfc_conf->fc.send_xon;
+       hw->fc.low_water[tc_num] =  pfc_conf->fc.low_water;
+       hw->fc.high_water[tc_num] = pfc_conf->fc.high_water;
+               
+       err = ixgbe_dcb_pfc_enable(dev,tc_num);
+       
+       /* Not negotiated is not an error case */
+       if ((err == IXGBE_SUCCESS) || (err == IXGBE_ERR_FC_NOT_NEGOTIATED)) 
+               return 0;
+
+       PMD_INIT_LOG(ERR, "ixgbe_dcb_pfc_enable = 0x%x \n", err);
+       return -EIO;
+}      
+
 static void
 ixgbe_add_rar(struct rte_eth_dev *dev, struct ether_addr *mac_addr,
                                uint32_t index, uint32_t pool)
@@ -1538,39 +2128,28 @@ ixgbevf_intr_disable(struct ixgbe_hw *hw)
 }
 
 static int
-ixgbevf_dev_configure(struct rte_eth_dev *dev, uint16_t nb_rx_q, uint16_t nb_tx_q)
+ixgbevf_dev_configure(struct rte_eth_dev *dev)
 {
-       int diag;
        struct rte_eth_conf* conf = &dev->data->dev_conf;
 
-       PMD_INIT_FUNC_TRACE();
-
-       /* Allocate the array of pointers to RX queue structures */
-       diag = ixgbe_dev_rx_queue_alloc(dev, nb_rx_q);
-       if (diag != 0) {
-               PMD_INIT_LOG(ERR, "ethdev port_id=%d allocation of array of %d"
-                            "pointers to RX queues failed", dev->data->port_id,
-                            nb_rx_q);
-               return diag;
-       }
-
-       /* Allocate the array of pointers to TX queue structures */
-       diag = ixgbe_dev_tx_queue_alloc(dev, nb_tx_q);
-       if (diag != 0) {
-               PMD_INIT_LOG(ERR, "ethdev port_id=%d allocation of array of %d"
-                            "pointers to TX queues failed", dev->data->port_id,
-                            nb_tx_q);
-               return diag;
-       }
+       PMD_INIT_LOG(DEBUG, "\nConfigured Virtual Function port id: %d\n",
+               dev->data->port_id);
 
+       /*
+        * VF has no ability to enable/disable HW CRC
+        * Keep the persistent behavior the same as Host PF
+        */
+#ifndef RTE_LIBRTE_IXGBE_PF_DISABLE_STRIP_CRC
        if (!conf->rxmode.hw_strip_crc) {
-               /*
-                * VF has no ability to enable/disable HW CRC
-                * Keep the persistent behavior the same as Host PF
-                */
                PMD_INIT_LOG(INFO, "VF can't disable HW CRC Strip\n");
                conf->rxmode.hw_strip_crc = 1;
        }
+#else
+       if (conf->rxmode.hw_strip_crc) {
+               PMD_INIT_LOG(INFO, "VF can't enable HW CRC Strip\n");
+               conf->rxmode.hw_strip_crc = 0;
+       }
+#endif
 
        return 0;
 }
@@ -1578,16 +2157,28 @@ ixgbevf_dev_configure(struct rte_eth_dev *dev, uint16_t nb_rx_q, uint16_t nb_tx_
 static int
 ixgbevf_dev_start(struct rte_eth_dev *dev)
 {
-       int err = 0;
+       int err, mask = 0;
+       
        PMD_INIT_LOG(DEBUG, "ixgbevf_dev_start");
 
        ixgbevf_dev_tx_init(dev);
+
+       /* This can fail when allocating mbufs for descriptor rings */
        err = ixgbevf_dev_rx_init(dev);
-       if(err){
+       if (err) {
+               PMD_INIT_LOG(ERR, "Unable to initialize RX hardware (%d)\n", err);
                ixgbe_dev_clear_queues(dev);
-               PMD_INIT_LOG(ERR,"Unable to initialize RX hardware\n");
                return err;
        }
+       
+       /* Set vfta */
+       ixgbevf_set_vfta_all(dev,1);
+
+       /* Set HW strip */
+       mask = ETH_VLAN_STRIP_MASK | ETH_VLAN_FILTER_MASK | \
+               ETH_VLAN_EXTEND_MASK;
+       ixgbevf_vlan_offload_set(dev, mask);
+
        ixgbevf_dev_rxtx_start(dev);
 
        return 0;
@@ -1599,10 +2190,121 @@ ixgbevf_dev_stop(struct rte_eth_dev *dev)
        struct ixgbe_hw *hw = IXGBE_DEV_PRIVATE_TO_HW(dev->data->dev_private);
 
        PMD_INIT_LOG(DEBUG, "ixgbevf_dev_stop");
+               
+       hw->adapter_stopped = TRUE;
+       ixgbe_stop_adapter(hw);
+
+       /* 
+         * Clear what we set, but we still keep shadow_vfta to 
+         * restore after device starts
+         */
+       ixgbevf_set_vfta_all(dev,0);
+
+       ixgbe_dev_clear_queues(dev);
+}
+
+static void
+ixgbevf_dev_close(struct rte_eth_dev *dev)
+{
+       struct ixgbe_hw *hw = IXGBE_DEV_PRIVATE_TO_HW(dev->data->dev_private);
+
+       PMD_INIT_LOG(DEBUG, "ixgbevf_dev_close");
 
        ixgbe_reset_hw(hw);
-       hw->adapter_stopped = 0;
-       ixgbe_stop_adapter(hw);
+
+       ixgbevf_dev_stop(dev);
+
        /* reprogram the RAR[0] in case user changed it. */
        ixgbe_set_rar(hw, 0, hw->mac.addr, 0, IXGBE_RAH_AV);
 }
+
+static void ixgbevf_set_vfta_all(struct rte_eth_dev *dev, bool on)
+{
+       struct ixgbe_hw *hw = IXGBE_DEV_PRIVATE_TO_HW(dev->data->dev_private);
+       struct ixgbe_vfta * shadow_vfta =
+               IXGBE_DEV_PRIVATE_TO_VFTA(dev->data->dev_private);
+       int i = 0, j = 0, vfta = 0, mask = 1;
+
+       for (i = 0; i < IXGBE_VFTA_SIZE; i++){
+               vfta = shadow_vfta->vfta[i];
+               if(vfta){
+                       mask = 1;
+                       for (j = 0; j < 32; j++){
+                               if(vfta & mask)
+                                       ixgbe_set_vfta(hw, (i<<5)+j, 0, on);
+                               mask<<=1;
+                       }
+               }
+       }
+
+}
+
+static int
+ixgbevf_vlan_filter_set(struct rte_eth_dev *dev, uint16_t vlan_id, int on)
+{
+       struct ixgbe_hw *hw =
+               IXGBE_DEV_PRIVATE_TO_HW(dev->data->dev_private);
+       struct ixgbe_vfta * shadow_vfta =
+               IXGBE_DEV_PRIVATE_TO_VFTA(dev->data->dev_private);
+       uint32_t vid_idx = 0;
+       uint32_t vid_bit = 0;
+       int ret = 0;
+       
+       PMD_INIT_FUNC_TRACE();
+
+       /* vind is not used in VF driver, set to 0, check ixgbe_set_vfta_vf */
+       ret = ixgbe_set_vfta(hw, vlan_id, 0, !!on);
+       if(ret){
+               PMD_INIT_LOG(ERR, "Unable to set VF vlan");
+               return ret;
+       }
+       vid_idx = (uint32_t) ((vlan_id >> 5) & 0x7F);
+       vid_bit = (uint32_t) (1 << (vlan_id & 0x1F));
+
+       /* Save what we set and retore it after device reset */
+       if (on)
+               shadow_vfta->vfta[vid_idx] |= vid_bit;
+       else
+               shadow_vfta->vfta[vid_idx] &= ~vid_bit;
+
+       return 0;
+}
+
+static void
+ixgbevf_vlan_strip_queue_set(struct rte_eth_dev *dev, uint16_t queue, int on)
+{
+       struct ixgbe_hw *hw =
+               IXGBE_DEV_PRIVATE_TO_HW(dev->data->dev_private);
+       uint32_t ctrl;
+
+       PMD_INIT_FUNC_TRACE();
+       
+       if(queue >= hw->mac.max_rx_queues)
+               return;
+
+       ctrl = IXGBE_READ_REG(hw, IXGBE_RXDCTL(queue));
+       if(on)
+               ctrl |= IXGBE_RXDCTL_VME;
+       else 
+               ctrl &= ~IXGBE_RXDCTL_VME;
+       IXGBE_WRITE_REG(hw, IXGBE_RXDCTL(queue), ctrl);
+
+       ixgbe_vlan_hw_strip_bitmap_set( dev, queue, on);
+}
+
+static void
+ixgbevf_vlan_offload_set(struct rte_eth_dev *dev, int mask)
+{
+       struct ixgbe_hw *hw =
+               IXGBE_DEV_PRIVATE_TO_HW(dev->data->dev_private);
+       uint32_t i, on = 0;
+
+       /* VF function only support hw strip feature, others are not support */
+       if(mask & ETH_VLAN_STRIP_MASK){
+               on = !!(dev->data->dev_conf.rxmode.hw_vlan_strip);
+
+               for(i=0; i < hw->mac.max_rx_queues; i++)
+                       ixgbevf_vlan_strip_queue_set(dev,i,on);
+       }
+}
+