ethdev: change queue release callback
[dpdk.git] / drivers / net / fm10k / fm10k_ethdev.c
index dc8bc75..7075d69 100644 (file)
@@ -2,14 +2,15 @@
  * Copyright(c) 2013-2016 Intel Corporation
  */
 
-#include <rte_ethdev_driver.h>
-#include <rte_ethdev_pci.h>
+#include <ethdev_driver.h>
+#include <ethdev_pci.h>
 #include <rte_malloc.h>
 #include <rte_memzone.h>
 #include <rte_string_fns.h>
 #include <rte_dev.h>
 #include <rte_spinlock.h>
 #include <rte_kvargs.h>
+#include <rte_vect.h>
 
 #include "fm10k.h"
 #include "base/fm10k_api.h"
@@ -40,9 +41,6 @@
 #define GLORT_FD_MASK    GLORT_PF_MASK
 #define GLORT_FD_INDEX   GLORT_FD_Q_BASE
 
-int fm10k_logtype_init;
-int fm10k_logtype_driver;
-
 static void fm10k_close_mbx_service(struct fm10k_hw *hw);
 static int fm10k_dev_promiscuous_enable(struct rte_eth_dev *dev);
 static int fm10k_dev_promiscuous_disable(struct rte_eth_dev *dev);
@@ -53,8 +51,8 @@ static int
 fm10k_vlan_filter_set(struct rte_eth_dev *dev, uint16_t vlan_id, int on);
 static void fm10k_MAC_filter_set(struct rte_eth_dev *dev,
        const u8 *mac, bool add, uint32_t pool);
-static void fm10k_tx_queue_release(void *queue);
-static void fm10k_rx_queue_release(void *queue);
+static void fm10k_tx_queue_release(struct rte_eth_dev *dev, uint16_t qid);
+static void fm10k_rx_queue_release(struct rte_eth_dev *dev, uint16_t qid);
 static void fm10k_set_rx_function(struct rte_eth_dev *dev);
 static void fm10k_set_tx_function(struct rte_eth_dev *dev);
 static int fm10k_check_ftag(struct rte_devargs *devargs);
@@ -451,6 +449,9 @@ fm10k_dev_configure(struct rte_eth_dev *dev)
 
        PMD_INIT_FUNC_TRACE();
 
+       if (dev->data->dev_conf.rxmode.mq_mode & ETH_MQ_RX_RSS_FLAG)
+               dev->data->dev_conf.rxmode.offloads |= DEV_RX_OFFLOAD_RSS_HASH;
+
        /* multipe queue mode checking */
        ret  = fm10k_check_mq_mode(dev);
        if (ret != 0) {
@@ -1152,7 +1153,7 @@ fm10k_dev_start(struct rte_eth_dev *dev)
        return 0;
 }
 
-static void
+static int
 fm10k_dev_stop(struct rte_eth_dev *dev)
 {
        struct fm10k_hw *hw = FM10K_DEV_PRIVATE_TO_HW(dev->data->dev_private);
@@ -1161,6 +1162,7 @@ fm10k_dev_stop(struct rte_eth_dev *dev)
        int i;
 
        PMD_INIT_FUNC_TRACE();
+       dev->data->dev_started = 0;
 
        if (dev->data->tx_queues)
                for (i = 0; i < dev->data->nb_tx_queues; i++)
@@ -1187,6 +1189,8 @@ fm10k_dev_stop(struct rte_eth_dev *dev)
        rte_intr_efd_disable(intr_handle);
        rte_free(intr_handle->intr_vec);
        intr_handle->intr_vec = NULL;
+
+       return 0;
 }
 
 static void
@@ -1206,7 +1210,7 @@ fm10k_dev_queue_release(struct rte_eth_dev *dev)
 
        if (dev->data->rx_queues) {
                for (i = 0; i < dev->data->nb_rx_queues; i++)
-                       fm10k_rx_queue_release(dev->data->rx_queues[i]);
+                       fm10k_rx_queue_release(dev, i);
        }
 }
 
@@ -1562,28 +1566,9 @@ fm10k_vlan_filter_set(struct rte_eth_dev *dev, uint16_t vlan_id, int on)
 }
 
 static int
-fm10k_vlan_offload_set(struct rte_eth_dev *dev, int mask)
+fm10k_vlan_offload_set(struct rte_eth_dev *dev __rte_unused,
+                      int mask __rte_unused)
 {
-       if (mask & ETH_VLAN_STRIP_MASK) {
-               if (!(dev->data->dev_conf.rxmode.offloads &
-                       DEV_RX_OFFLOAD_VLAN_STRIP))
-                       PMD_INIT_LOG(ERR, "VLAN stripping is "
-                                       "always on in fm10k");
-       }
-
-       if (mask & ETH_VLAN_EXTEND_MASK) {
-               if (dev->data->dev_conf.rxmode.offloads &
-                       DEV_RX_OFFLOAD_VLAN_EXTEND)
-                       PMD_INIT_LOG(ERR, "VLAN QinQ is not "
-                                       "supported in fm10k");
-       }
-
-       if (mask & ETH_VLAN_FILTER_MASK) {
-               if (!(dev->data->dev_conf.rxmode.offloads &
-                       DEV_RX_OFFLOAD_VLAN_FILTER))
-                       PMD_INIT_LOG(ERR, "VLAN filter is always on in fm10k");
-       }
-
        return 0;
 }
 
@@ -1795,7 +1780,8 @@ static uint64_t fm10k_get_rx_port_offloads_capa(struct rte_eth_dev *dev)
                           DEV_RX_OFFLOAD_UDP_CKSUM   |
                           DEV_RX_OFFLOAD_TCP_CKSUM   |
                           DEV_RX_OFFLOAD_JUMBO_FRAME |
-                          DEV_RX_OFFLOAD_HEADER_SPLIT);
+                          DEV_RX_OFFLOAD_HEADER_SPLIT |
+                          DEV_RX_OFFLOAD_RSS_HASH);
 }
 
 static int
@@ -1859,9 +1845,10 @@ fm10k_rx_queue_setup(struct rte_eth_dev *dev, uint16_t queue_id,
        q->tail_ptr = (volatile uint32_t *)
                &((uint32_t *)hw->hw_addr)[FM10K_RDT(queue_id)];
        q->offloads = offloads;
-       if (handle_rxconf(q, conf))
+       if (handle_rxconf(q, conf)) {
+               rte_free(q);
                return -EINVAL;
-
+       }
        /* allocate memory for the software ring */
        q->sw_ring = rte_zmalloc_socket("fm10k sw ring",
                        (nb_desc + q->nb_fake_desc) * sizeof(struct rte_mbuf *),
@@ -1904,11 +1891,11 @@ fm10k_rx_queue_setup(struct rte_eth_dev *dev, uint16_t queue_id,
 }
 
 static void
-fm10k_rx_queue_release(void *queue)
+fm10k_rx_queue_release(struct rte_eth_dev *dev, uint16_t qid)
 {
        PMD_INIT_FUNC_TRACE();
 
-       rx_queue_free(queue);
+       rx_queue_free(dev->data->rx_queues[qid]);
 }
 
 static inline int
@@ -2041,8 +2028,10 @@ fm10k_tx_queue_setup(struct rte_eth_dev *dev, uint16_t queue_id,
        q->ops = &def_txq_ops;
        q->tail_ptr = (volatile uint32_t *)
                &((uint32_t *)hw->hw_addr)[FM10K_TDT(queue_id)];
-       if (handle_txconf(q, conf))
+       if (handle_txconf(q, conf)) {
+               rte_free(q);
                return -EINVAL;
+       }
 
        /* allocate memory for the software ring */
        q->sw_ring = rte_zmalloc_socket("fm10k sw ring",
@@ -2091,9 +2080,9 @@ fm10k_tx_queue_setup(struct rte_eth_dev *dev, uint16_t queue_id,
 }
 
 static void
-fm10k_tx_queue_release(void *queue)
+fm10k_tx_queue_release(struct rte_eth_dev *dev, uint16_t qid)
 {
-       struct fm10k_tx_queue *q = queue;
+       struct fm10k_tx_queue *q = dev->data->tx_queues[qid];
        PMD_INIT_FUNC_TRACE();
 
        tx_queue_free(q);
@@ -2639,7 +2628,7 @@ fm10k_dev_interrupt_handler_pf(void *param)
                                        true);
 
                        dev_info->sm_down = 0;
-                       _rte_eth_dev_callback_process(dev,
+                       rte_eth_dev_callback_process(dev,
                                        RTE_ETH_EVENT_INTR_LSC,
                                        NULL);
                }
@@ -2653,8 +2642,7 @@ fm10k_dev_interrupt_handler_pf(void *param)
        if (err == FM10K_ERR_RESET_REQUESTED) {
                PMD_INIT_LOG(INFO, "INT: Switch is down");
                dev_info->sm_down = 1;
-               _rte_eth_dev_callback_process(dev, RTE_ETH_EVENT_INTR_LSC,
-                               NULL);
+               rte_eth_dev_callback_process(dev, RTE_ETH_EVENT_INTR_LSC, NULL);
        }
 
        /* Handle SRAM error */
@@ -2721,8 +2709,7 @@ fm10k_dev_interrupt_handler_vf(void *param)
 
                /* Setting reset flag */
                dev_info->sm_down = 1;
-               _rte_eth_dev_callback_process(dev, RTE_ETH_EVENT_INTR_LSC,
-                               NULL);
+               rte_eth_dev_callback_process(dev, RTE_ETH_EVENT_INTR_LSC, NULL);
        }
 
        if (dev_info->sm_down == 1 &&
@@ -2750,8 +2737,7 @@ fm10k_dev_interrupt_handler_vf(void *param)
                fm10k_vlan_filter_set(dev, hw->mac.default_vid, true);
 
                dev_info->sm_down = 0;
-               _rte_eth_dev_callback_process(dev, RTE_ETH_EVENT_INTR_LSC,
-                               NULL);
+               rte_eth_dev_callback_process(dev, RTE_ETH_EVENT_INTR_LSC, NULL);
        }
 
        /* Re-enable interrupt from device side */
@@ -2797,14 +2783,17 @@ fm10k_close_mbx_service(struct fm10k_hw *hw)
        hw->mbx.ops.disconnect(hw, &hw->mbx);
 }
 
-static void
+static int
 fm10k_dev_close(struct rte_eth_dev *dev)
 {
        struct fm10k_hw *hw = FM10K_DEV_PRIVATE_TO_HW(dev->data->dev_private);
        struct rte_pci_device *pdev = RTE_ETH_DEV_TO_PCI(dev);
        struct rte_intr_handle *intr_handle = &pdev->intr_handle;
+       int ret;
 
        PMD_INIT_FUNC_TRACE();
+       if (rte_eal_process_type() != RTE_PROC_PRIMARY)
+               return 0;
 
        fm10k_mbx_lock(hw);
        hw->mac.ops.update_lport_state(hw, hw->mac.dglort_map,
@@ -2816,14 +2805,12 @@ fm10k_dev_close(struct rte_eth_dev *dev)
 
        /* Stop mailbox service first */
        fm10k_close_mbx_service(hw);
-       fm10k_dev_stop(dev);
+
+       ret = fm10k_dev_stop(dev);
+
        fm10k_dev_queue_release(dev);
        fm10k_stop_hw(hw);
 
-       dev->dev_ops = NULL;
-       dev->rx_pkt_burst = NULL;
-       dev->tx_pkt_burst = NULL;
-
        /* disable uio/vfio intr */
        rte_intr_disable(intr_handle);
 
@@ -2842,6 +2829,8 @@ fm10k_dev_close(struct rte_eth_dev *dev)
                rte_intr_callback_unregister(intr_handle,
                        fm10k_dev_interrupt_handler_vf, (void *)dev);
        }
+
+       return ret;
 }
 
 static const struct eth_dev_ops fm10k_eth_dev_ops = {
@@ -2873,10 +2862,6 @@ static const struct eth_dev_ops fm10k_eth_dev_ops = {
        .rx_queue_release       = fm10k_rx_queue_release,
        .tx_queue_setup         = fm10k_tx_queue_setup,
        .tx_queue_release       = fm10k_tx_queue_release,
-       .rx_queue_count         = fm10k_dev_rx_queue_count,
-       .rx_descriptor_done     = fm10k_dev_rx_descriptor_done,
-       .rx_descriptor_status = fm10k_dev_rx_descriptor_status,
-       .tx_descriptor_status = fm10k_dev_tx_descriptor_status,
        .rx_queue_intr_enable   = fm10k_dev_rx_queue_intr_enable,
        .rx_queue_intr_disable  = fm10k_dev_rx_queue_intr_disable,
        .reta_update            = fm10k_reta_update,
@@ -2944,7 +2929,7 @@ fm10k_xmit_pkts_vec(void *tx_queue, struct rte_mbuf **tx_pkts,
        return nb_tx;
 }
 
-static void __attribute__((cold))
+static void __rte_cold
 fm10k_set_tx_function(struct rte_eth_dev *dev)
 {
        struct fm10k_tx_queue *txq;
@@ -2955,7 +2940,8 @@ fm10k_set_tx_function(struct rte_eth_dev *dev)
        if (rte_eal_process_type() != RTE_PROC_PRIMARY) {
                /* primary process has set the ftag flag and offloads */
                txq = dev->data->tx_queues[0];
-               if (fm10k_tx_vec_condition_check(txq)) {
+               if (fm10k_tx_vec_condition_check(txq) ||
+                               rte_vect_get_max_simd_bitwidth() < RTE_VECT_SIMD_128) {
                        dev->tx_pkt_burst = fm10k_xmit_pkts;
                        dev->tx_pkt_prepare = fm10k_prep_pkts;
                        PMD_INIT_LOG(DEBUG, "Use regular Tx func");
@@ -2974,7 +2960,8 @@ fm10k_set_tx_function(struct rte_eth_dev *dev)
                txq = dev->data->tx_queues[i];
                txq->tx_ftag_en = tx_ftag_en;
                /* Check if Vector Tx is satisfied */
-               if (fm10k_tx_vec_condition_check(txq))
+               if (fm10k_tx_vec_condition_check(txq) ||
+                               rte_vect_get_max_simd_bitwidth() < RTE_VECT_SIMD_128)
                        use_sse = 0;
        }
 
@@ -2993,7 +2980,7 @@ fm10k_set_tx_function(struct rte_eth_dev *dev)
        }
 }
 
-static void __attribute__((cold))
+static void __rte_cold
 fm10k_set_rx_function(struct rte_eth_dev *dev)
 {
        struct fm10k_dev_info *dev_info =
@@ -3008,7 +2995,8 @@ fm10k_set_rx_function(struct rte_eth_dev *dev)
         * conditions to be met.
         */
        if (!fm10k_rx_vec_condition_check(dev) &&
-                       dev_info->rx_vec_allowed && !rx_ftag_en) {
+                       dev_info->rx_vec_allowed && !rx_ftag_en &&
+                       rte_vect_get_max_simd_bitwidth() >= RTE_VECT_SIMD_128) {
                if (dev->data->scattered_rx)
                        dev->rx_pkt_burst = fm10k_recv_scattered_pkts_vec;
                else
@@ -3073,6 +3061,10 @@ eth_fm10k_dev_init(struct rte_eth_dev *dev)
        PMD_INIT_FUNC_TRACE();
 
        dev->dev_ops = &fm10k_eth_dev_ops;
+       dev->rx_queue_count = fm10k_dev_rx_queue_count;
+       dev->rx_descriptor_done = fm10k_dev_rx_descriptor_done;
+       dev->rx_descriptor_status = fm10k_dev_rx_descriptor_status;
+       dev->tx_descriptor_status = fm10k_dev_tx_descriptor_status;
        dev->rx_pkt_burst = &fm10k_recv_pkts;
        dev->tx_pkt_burst = &fm10k_xmit_pkts;
        dev->tx_pkt_prepare = &fm10k_prep_pkts;
@@ -3088,6 +3080,7 @@ eth_fm10k_dev_init(struct rte_eth_dev *dev)
        }
 
        rte_eth_copy_pci_info(dev, pdev);
+       dev->data->dev_flags |= RTE_ETH_DEV_AUTOFILL_QUEUE_XSTATS;
 
        macvlan = FM10K_DEV_PRIVATE_TO_MACVLAN(dev->data->dev_private);
        memset(macvlan, 0, sizeof(*macvlan));
@@ -3101,7 +3094,7 @@ eth_fm10k_dev_init(struct rte_eth_dev *dev)
        hw->hw_addr = (void *)pdev->mem_resource[0].addr;
        if (hw->hw_addr == NULL) {
                PMD_INIT_LOG(ERR, "Bad mem resource."
-                       " Try to blacklist unused devices.");
+                       " Try to refuse unused devices.");
                return -EIO;
        }
 
@@ -3148,11 +3141,6 @@ eth_fm10k_dev_init(struct rte_eth_dev *dev)
                &dev->data->mac_addrs[0]);
        }
 
-       /* Pass the information to the rte_eth_dev_close() that it should also
-        * release the private port resources.
-        */
-       dev->data->dev_flags |= RTE_ETH_DEV_CLOSE_REMOVE;
-
        /* Reset the hw statistics */
        diag = fm10k_stats_reset(dev);
        if (diag != 0) {
@@ -3196,19 +3184,19 @@ eth_fm10k_dev_init(struct rte_eth_dev *dev)
 
        /* Make sure Switch Manager is ready before going forward. */
        if (hw->mac.type == fm10k_mac_pf) {
-               int switch_ready = 0;
+               bool switch_ready = false;
 
                for (i = 0; i < MAX_QUERY_SWITCH_STATE_TIMES; i++) {
                        fm10k_mbx_lock(hw);
                        hw->mac.ops.get_host_state(hw, &switch_ready);
                        fm10k_mbx_unlock(hw);
-                       if (switch_ready)
+                       if (switch_ready == true)
                                break;
                        /* Delay some time to acquire async LPORT_MAP info. */
                        rte_delay_us(WAIT_SWITCH_MSG_US);
                }
 
-               if (switch_ready == 0) {
+               if (switch_ready == false) {
                        PMD_INIT_LOG(ERR, "switch is not ready");
                        return -1;
                }
@@ -3260,14 +3248,7 @@ static int
 eth_fm10k_dev_uninit(struct rte_eth_dev *dev)
 {
        PMD_INIT_FUNC_TRACE();
-
-       /* only uninitialize in the primary process */
-       if (rte_eal_process_type() != RTE_PROC_PRIMARY)
-               return 0;
-
-       /* safe to close dev here */
        fm10k_dev_close(dev);
-
        return 0;
 }
 
@@ -3304,13 +3285,11 @@ static struct rte_pci_driver rte_pmd_fm10k = {
 RTE_PMD_REGISTER_PCI(net_fm10k, rte_pmd_fm10k);
 RTE_PMD_REGISTER_PCI_TABLE(net_fm10k, pci_id_fm10k_map);
 RTE_PMD_REGISTER_KMOD_DEP(net_fm10k, "* igb_uio | uio_pci_generic | vfio-pci");
-
-RTE_INIT(fm10k_init_log)
-{
-       fm10k_logtype_init = rte_log_register("pmd.net.fm10k.init");
-       if (fm10k_logtype_init >= 0)
-               rte_log_set_level(fm10k_logtype_init, RTE_LOG_NOTICE);
-       fm10k_logtype_driver = rte_log_register("pmd.net.fm10k.driver");
-       if (fm10k_logtype_driver >= 0)
-               rte_log_set_level(fm10k_logtype_driver, RTE_LOG_NOTICE);
-}
+RTE_LOG_REGISTER_SUFFIX(fm10k_logtype_init, init, NOTICE);
+RTE_LOG_REGISTER_SUFFIX(fm10k_logtype_driver, driver, NOTICE);
+#ifdef RTE_ETHDEV_DEBUG_RX
+RTE_LOG_REGISTER_SUFFIX(fm10k_logtype_rx, rx, DEBUG);
+#endif
+#ifdef RTE_ETHDEV_DEBUG_TX
+RTE_LOG_REGISTER_SUFFIX(fm10k_logtype_tx, tx, DEBUG);
+#endif