net/dpaa2: fix check for key size
[dpdk.git] / drivers / net / dpaa2 / dpaa2_ethdev.c
index b58add9..d862451 100644 (file)
@@ -99,7 +99,6 @@ static const enum rte_filter_op dpaa2_supported_filter_ops[] = {
 };
 
 static struct rte_dpaa2_driver rte_dpaa2_pmd;
-static int dpaa2_dev_uninit(struct rte_eth_dev *eth_dev);
 static int dpaa2_dev_link_update(struct rte_eth_dev *dev,
                                 int wait_to_complete);
 static int dpaa2_dev_set_link_up(struct rte_eth_dev *dev);
@@ -241,8 +240,6 @@ dpaa2_dev_info_get(struct rte_eth_dev *dev, struct rte_eth_dev_info *dev_info)
 
        PMD_INIT_FUNC_TRACE();
 
-       dev_info->if_index = priv->hw_id;
-
        dev_info->max_mac_addrs = priv->max_mac_filters;
        dev_info->max_rx_pktlen = DPAA2_MAX_RX_PKT_LEN;
        dev_info->min_rx_bufsize = DPAA2_MIN_RX_BUF_SIZE;
@@ -677,6 +674,8 @@ dpaa2_dev_rx_queue_setup(struct rte_eth_dev *dev,
        dpaa2_q = (struct dpaa2_queue *)priv->rx_vq[rx_queue_id];
        dpaa2_q->mb_pool = mb_pool; /**< mbuf pool to populate RX ring. */
        dpaa2_q->bp_array = rte_dpaa2_bpid_info;
+       dpaa2_q->nb_desc = UINT16_MAX;
+       dpaa2_q->offloads = rx_conf->offloads;
 
        /*Get the flow id from given VQ id*/
        flow_id = dpaa2_q->flow_id;
@@ -729,7 +728,7 @@ dpaa2_dev_rx_queue_setup(struct rte_eth_dev *dev,
                struct dpni_taildrop taildrop;
 
                taildrop.enable = 1;
-
+               dpaa2_q->nb_desc = nb_rx_desc;
                /* Private CGR will use tail drop length as nb_rx_desc.
                 * for rest cases we can use standard byte based tail drop.
                 * There is no HW restriction, but number of CGRs are limited,
@@ -819,6 +818,9 @@ dpaa2_dev_tx_queue_setup(struct rte_eth_dev *dev,
                return -EINVAL;
        }
 
+       dpaa2_q->nb_desc = UINT16_MAX;
+       dpaa2_q->offloads = tx_conf->offloads;
+
        /* Return if queue already configured */
        if (dpaa2_q->flow_id != 0xffff) {
                dev->data->tx_queues[tx_queue_id] = dpaa2_q;
@@ -872,6 +874,8 @@ dpaa2_dev_tx_queue_setup(struct rte_eth_dev *dev,
        if (!(priv->flags & DPAA2_TX_CGR_OFF)) {
                struct dpni_congestion_notification_cfg cong_notif_cfg = {0};
 
+               dpaa2_q->nb_desc = nb_tx_desc;
+
                cong_notif_cfg.units = DPNI_CONGESTION_UNIT_FRAMES;
                cong_notif_cfg.threshold_entry = nb_tx_desc;
                /* Notify that the queue is not congested when the data in
@@ -1058,8 +1062,7 @@ dpaa2_interrupt_handler(void *param)
                clear = DPNI_IRQ_EVENT_LINK_CHANGED;
                dpaa2_dev_link_update(dev, 0);
                /* calling all the apps registered for link status event */
-               _rte_eth_dev_callback_process(dev, RTE_ETH_EVENT_INTR_LSC,
-                                             NULL);
+               rte_eth_dev_callback_process(dev, RTE_ETH_EVENT_INTR_LSC, NULL);
        }
 out:
        ret = dpni_clear_irq_status(dpni, CMD_PRI_LOW, priv->token,
@@ -1232,27 +1235,61 @@ dpaa2_dev_stop(struct rte_eth_dev *dev)
        rte_eth_linkstatus_set(dev, &link);
 }
 
-static void
+static int
 dpaa2_dev_close(struct rte_eth_dev *dev)
 {
        struct dpaa2_dev_priv *priv = dev->data->dev_private;
        struct fsl_mc_io *dpni = (struct fsl_mc_io *)dev->process_private;
-       int ret;
+       int i, ret;
        struct rte_eth_link link;
 
        PMD_INIT_FUNC_TRACE();
 
-       dpaa2_flow_clean(dev);
+       if (rte_eal_process_type() != RTE_PROC_PRIMARY)
+               return 0;
 
+       if (!dpni) {
+               DPAA2_PMD_WARN("Already closed or not started");
+               return -1;
+       }
+
+       dpaa2_flow_clean(dev);
        /* Clean the device first */
        ret = dpni_reset(dpni, CMD_PRI_LOW, priv->token);
        if (ret) {
                DPAA2_PMD_ERR("Failure cleaning dpni device: err=%d", ret);
-               return;
+               return -1;
        }
 
        memset(&link, 0, sizeof(link));
        rte_eth_linkstatus_set(dev, &link);
+
+       /* Free private queues memory */
+       dpaa2_free_rx_tx_queues(dev);
+       /* Close the device at underlying layer*/
+       ret = dpni_close(dpni, CMD_PRI_LOW, priv->token);
+       if (ret) {
+               DPAA2_PMD_ERR("Failure closing dpni device with err code %d",
+                             ret);
+       }
+
+       /* Free the allocated memory for ethernet private data and dpni*/
+       priv->hw = NULL;
+       dev->process_private = NULL;
+       rte_free(dpni);
+
+       for (i = 0; i < MAX_TCS; i++)
+               rte_free((void *)(size_t)priv->extract.tc_extract_param[i]);
+
+       if (priv->extract.qos_extract_param)
+               rte_free((void *)(size_t)priv->extract.qos_extract_param);
+
+       dev->dev_ops = NULL;
+       dev->rx_pkt_burst = NULL;
+       dev->tx_pkt_burst = NULL;
+
+       DPAA2_PMD_INFO("%s: netdev deleted", dev->data->name);
+       return 0;
 }
 
 static int
@@ -2255,6 +2292,43 @@ dpaa2_dev_flow_ctrl(struct rte_eth_dev *dev,
        return ret;
 }
 
+static void
+dpaa2_rxq_info_get(struct rte_eth_dev *dev, uint16_t queue_id,
+       struct rte_eth_rxq_info *qinfo)
+{
+       struct dpaa2_queue *rxq;
+
+       rxq = (struct dpaa2_queue *)dev->data->rx_queues[queue_id];
+
+       qinfo->mp = rxq->mb_pool;
+       qinfo->scattered_rx = dev->data->scattered_rx;
+       qinfo->nb_desc = rxq->nb_desc;
+
+       qinfo->conf.rx_free_thresh = 1;
+       qinfo->conf.rx_drop_en = 1;
+       qinfo->conf.rx_deferred_start = 0;
+       qinfo->conf.offloads = rxq->offloads;
+}
+
+static void
+dpaa2_txq_info_get(struct rte_eth_dev *dev, uint16_t queue_id,
+       struct rte_eth_txq_info *qinfo)
+{
+       struct dpaa2_queue *txq;
+
+       txq = dev->data->tx_queues[queue_id];
+
+       qinfo->nb_desc = txq->nb_desc;
+       qinfo->conf.tx_thresh.pthresh = 0;
+       qinfo->conf.tx_thresh.hthresh = 0;
+       qinfo->conf.tx_thresh.wthresh = 0;
+
+       qinfo->conf.tx_free_thresh = 0;
+       qinfo->conf.tx_rs_thresh = 0;
+       qinfo->conf.offloads = txq->offloads;
+       qinfo->conf.tx_deferred_start = 0;
+}
+
 static struct eth_dev_ops dpaa2_ethdev_ops = {
        .dev_configure    = dpaa2_eth_dev_configure,
        .dev_start            = dpaa2_dev_start,
@@ -2287,7 +2361,6 @@ static struct eth_dev_ops dpaa2_ethdev_ops = {
        .tx_queue_release  = dpaa2_dev_tx_queue_release,
        .rx_burst_mode_get = dpaa2_dev_rx_burst_mode_get,
        .tx_burst_mode_get = dpaa2_dev_tx_burst_mode_get,
-       .rx_queue_count       = dpaa2_dev_rx_queue_count,
        .flow_ctrl_get        = dpaa2_flow_ctrl_get,
        .flow_ctrl_set        = dpaa2_flow_ctrl_set,
        .mac_addr_add         = dpaa2_dev_add_mac_addr,
@@ -2296,6 +2369,8 @@ static struct eth_dev_ops dpaa2_ethdev_ops = {
        .rss_hash_update      = dpaa2_dev_rss_hash_update,
        .rss_hash_conf_get    = dpaa2_dev_rss_hash_conf_get,
        .filter_ctrl          = dpaa2_dev_flow_ctrl,
+       .rxq_info_get         = dpaa2_rxq_info_get,
+       .txq_info_get         = dpaa2_txq_info_get,
 #if defined(RTE_LIBRTE_IEEE1588)
        .timesync_enable      = dpaa2_timesync_enable,
        .timesync_disable     = dpaa2_timesync_disable,
@@ -2440,6 +2515,7 @@ dpaa2_dev_init(struct rte_eth_dev *eth_dev)
                 * plugged.
                 */
                eth_dev->dev_ops = &dpaa2_ethdev_ops;
+               eth_dev->rx_queue_count = dpaa2_dev_rx_queue_count;
                if (dpaa2_get_devargs(dev->devargs, DRIVER_LOOPBACK_MODE))
                        eth_dev->rx_pkt_burst = dpaa2_dev_loopback_rx;
                else if (dpaa2_get_devargs(dev->devargs,
@@ -2664,56 +2740,9 @@ dpaa2_dev_init(struct rte_eth_dev *eth_dev)
        RTE_LOG(INFO, PMD, "%s: netdev created\n", eth_dev->data->name);
        return 0;
 init_err:
-       dpaa2_dev_uninit(eth_dev);
-       return ret;
-}
-
-static int
-dpaa2_dev_uninit(struct rte_eth_dev *eth_dev)
-{
-       struct dpaa2_dev_priv *priv = eth_dev->data->dev_private;
-       struct fsl_mc_io *dpni = (struct fsl_mc_io *)eth_dev->process_private;
-       int i, ret;
-
-       PMD_INIT_FUNC_TRACE();
-
-       if (rte_eal_process_type() != RTE_PROC_PRIMARY)
-               return 0;
-
-       if (!dpni) {
-               DPAA2_PMD_WARN("Already closed or not started");
-               return -1;
-       }
-
        dpaa2_dev_close(eth_dev);
 
-       dpaa2_free_rx_tx_queues(eth_dev);
-
-       /* Close the device at underlying layer*/
-       ret = dpni_close(dpni, CMD_PRI_LOW, priv->token);
-       if (ret) {
-               DPAA2_PMD_ERR(
-                            "Failure closing dpni device with err code %d",
-                            ret);
-       }
-
-       /* Free the allocated memory for ethernet private data and dpni*/
-       priv->hw = NULL;
-       eth_dev->process_private = NULL;
-       rte_free(dpni);
-
-       for (i = 0; i < MAX_TCS; i++)
-               rte_free((void *)(size_t)priv->extract.tc_extract_param[i]);
-
-       if (priv->extract.qos_extract_param)
-               rte_free((void *)(size_t)priv->extract.qos_extract_param);
-
-       eth_dev->dev_ops = NULL;
-       eth_dev->rx_pkt_burst = NULL;
-       eth_dev->tx_pkt_burst = NULL;
-
-       DPAA2_PMD_INFO("%s: netdev deleted", eth_dev->data->name);
-       return 0;
+       return ret;
 }
 
 static int
@@ -2782,13 +2811,13 @@ static int
 rte_dpaa2_remove(struct rte_dpaa2_device *dpaa2_dev)
 {
        struct rte_eth_dev *eth_dev;
+       int ret;
 
        eth_dev = dpaa2_dev->eth_dev;
-       dpaa2_dev_uninit(eth_dev);
-
-       rte_eth_dev_release_port(eth_dev);
+       dpaa2_dev_close(eth_dev);
+       ret = rte_eth_dev_release_port(eth_dev);
 
-       return 0;
+       return ret;
 }
 
 static struct rte_dpaa2_driver rte_dpaa2_pmd = {