drivers/net: update Rx RSS hash offload capabilities
[dpdk.git] / drivers / net / cxgbe / cxgbe_ethdev.c
index 7d7be69..3dc9351 100644 (file)
@@ -39,6 +39,7 @@
 #include "cxgbe_flow.h"
 
 int cxgbe_logtype;
+int cxgbe_mbox_logtype;
 
 /*
  * Macros needed to support the PCI Device ID Table ...
@@ -67,20 +68,22 @@ uint16_t cxgbe_xmit_pkts(void *tx_queue, struct rte_mbuf **tx_pkts,
        struct sge_eth_txq *txq = (struct sge_eth_txq *)tx_queue;
        uint16_t pkts_sent, pkts_remain;
        uint16_t total_sent = 0;
+       uint16_t idx = 0;
        int ret = 0;
 
-       CXGBE_DEBUG_TX(adapter, "%s: txq = %p; tx_pkts = %p; nb_pkts = %d\n",
-                      __func__, txq, tx_pkts, nb_pkts);
-
        t4_os_lock(&txq->txq_lock);
        /* free up desc from already completed tx */
        reclaim_completed_tx(&txq->q);
+       rte_prefetch0(rte_pktmbuf_mtod(tx_pkts[0], volatile void *));
        while (total_sent < nb_pkts) {
                pkts_remain = nb_pkts - total_sent;
 
                for (pkts_sent = 0; pkts_sent < pkts_remain; pkts_sent++) {
-                       ret = t4_eth_xmit(txq, tx_pkts[total_sent + pkts_sent],
-                                         nb_pkts);
+                       idx = total_sent + pkts_sent;
+                       if ((idx + 1) < nb_pkts)
+                               rte_prefetch0(rte_pktmbuf_mtod(tx_pkts[idx + 1],
+                                                       volatile void *));
+                       ret = t4_eth_xmit(txq, tx_pkts[idx], nb_pkts);
                        if (ret < 0)
                                break;
                }
@@ -101,13 +104,9 @@ uint16_t cxgbe_recv_pkts(void *rx_queue, struct rte_mbuf **rx_pkts,
        struct sge_eth_rxq *rxq = (struct sge_eth_rxq *)rx_queue;
        unsigned int work_done;
 
-       CXGBE_DEBUG_RX(adapter, "%s: rxq->rspq.cntxt_id = %u; nb_pkts = %d\n",
-                      __func__, rxq->rspq.cntxt_id, nb_pkts);
-
        if (cxgbe_poll(&rxq->rspq, rx_pkts, (unsigned int)nb_pkts, &work_done))
                dev_err(adapter, "error in cxgbe poll\n");
 
-       CXGBE_DEBUG_RX(adapter, "%s: work_done = %u\n", __func__, work_done);
        return work_done;
 }
 
@@ -201,6 +200,9 @@ int cxgbe_dev_link_update(struct rte_eth_dev *eth_dev,
        u8 old_link = pi->link_cfg.link_ok;
 
        for (i = 0; i < CXGBE_LINK_STATUS_POLL_CNT; i++) {
+               if (!s->fw_evtq.desc)
+                       break;
+
                cxgbe_poll(&s->fw_evtq, NULL, budget, &work_done);
 
                /* Exit if link status changed or always forced up */
@@ -234,6 +236,9 @@ int cxgbe_dev_set_link_up(struct rte_eth_dev *dev)
        struct sge *s = &adapter->sge;
        int ret;
 
+       if (!s->fw_evtq.desc)
+               return -ENOMEM;
+
        /* Flush all link events */
        cxgbe_poll(&s->fw_evtq, NULL, budget, &work_done);
 
@@ -260,6 +265,9 @@ int cxgbe_dev_set_link_down(struct rte_eth_dev *dev)
        struct sge *s = &adapter->sge;
        int ret;
 
+       if (!s->fw_evtq.desc)
+               return -ENOMEM;
+
        /* Flush all link events */
        cxgbe_poll(&s->fw_evtq, NULL, budget, &work_done);
 
@@ -418,6 +426,8 @@ int cxgbe_dev_configure(struct rte_eth_dev *eth_dev)
 
        CXGBE_FUNC_TRACE();
 
+       eth_dev->data->dev_conf.rxmode.offloads |= DEV_RX_OFFLOAD_RSS_HASH;
+
        if (!(adapter->flags & FW_QUEUE_BOUND)) {
                err = cxgbe_setup_sge_fwevtq(adapter);
                if (err)
@@ -1182,6 +1192,8 @@ static int eth_cxgbe_dev_init(struct rte_eth_dev *eth_dev)
        adapter->eth_dev = eth_dev;
        pi->adapter = adapter;
 
+       cxgbe_process_devargs(adapter);
+
        err = cxgbe_probe(adapter);
        if (err) {
                dev_err(adapter, "%s: cxgbe probe failed with err %d\n",
@@ -1229,12 +1241,15 @@ RTE_PMD_REGISTER_PCI(net_cxgbe, rte_cxgbe_pmd);
 RTE_PMD_REGISTER_PCI_TABLE(net_cxgbe, cxgb4_pci_tbl);
 RTE_PMD_REGISTER_KMOD_DEP(net_cxgbe, "* igb_uio | uio_pci_generic | vfio-pci");
 RTE_PMD_REGISTER_PARAM_STRING(net_cxgbe,
-                             CXGBE_DEVARG_KEEP_OVLAN "=<0|1> "
-                             CXGBE_DEVARG_FORCE_LINK_UP "=<0|1> ");
+                             CXGBE_DEVARG_CMN_KEEP_OVLAN "=<0|1> "
+                             CXGBE_DEVARG_CMN_TX_MODE_LATENCY "=<0|1> ");
 
 RTE_INIT(cxgbe_init_log)
 {
        cxgbe_logtype = rte_log_register("pmd.net.cxgbe");
        if (cxgbe_logtype >= 0)
                rte_log_set_level(cxgbe_logtype, RTE_LOG_NOTICE);
+       cxgbe_mbox_logtype = rte_log_register("pmd.net.cxgbe.mbox");
+       if (cxgbe_mbox_logtype >= 0)
+               rte_log_set_level(cxgbe_mbox_logtype, RTE_LOG_NOTICE);
 }