build: disable experimental API check internally
[dpdk.git] / drivers / net / e1000 / em_rxtx.c
index 482d884..49c5371 100644 (file)
@@ -18,6 +18,7 @@
 #include <rte_log.h>
 #include <rte_debug.h>
 #include <rte_pci.h>
+#include <rte_bus_pci.h>
 #include <rte_memory.h>
 #include <rte_memcpy.h>
 #include <rte_memzone.h>
 #define E1000_TX_OFFLOAD_NOTSUP_MASK \
                (PKT_TX_OFFLOAD_MASK ^ E1000_TX_OFFLOAD_MASK)
 
+/* PCI offset for querying configuration status register */
+#define PCI_CFG_STATUS_REG                 0x06
+#define FLUSH_DESC_REQUIRED               0x100
+
+
 /**
  * Structure associated with each descriptor of the RX ring of a RX queue.
  */
@@ -616,20 +622,20 @@ eth_em_prep_pkts(__rte_unused void *tx_queue, struct rte_mbuf **tx_pkts,
                m = tx_pkts[i];
 
                if (m->ol_flags & E1000_TX_OFFLOAD_NOTSUP_MASK) {
-                       rte_errno = -ENOTSUP;
+                       rte_errno = ENOTSUP;
                        return i;
                }
 
 #ifdef RTE_LIBRTE_ETHDEV_DEBUG
                ret = rte_validate_tx_offload(m);
                if (ret != 0) {
-                       rte_errno = ret;
+                       rte_errno = -ret;
                        return i;
                }
 #endif
                ret = rte_net_intel_cksum_prepare(m);
                if (ret != 0) {
-                       rte_errno = ret;
+                       rte_errno = -ret;
                        return i;
                }
        }
@@ -1964,6 +1970,22 @@ eth_em_tx_init(struct rte_eth_dev *dev)
        tctl |= (E1000_TCTL_PSP | E1000_TCTL_RTLC | E1000_TCTL_EN |
                 (E1000_COLLISION_THRESHOLD << E1000_CT_SHIFT));
 
+       /* SPT and CNP Si errata workaround to avoid data corruption */
+       if (hw->mac.type == e1000_pch_spt) {
+               uint32_t reg_val;
+               reg_val = E1000_READ_REG(hw, E1000_IOSFPC);
+               reg_val |= E1000_RCTL_RDMTS_HEX;
+               E1000_WRITE_REG(hw, E1000_IOSFPC, reg_val);
+
+               /* Dropping the number of outstanding requests from
+                * 3 to 2 in order to avoid a buffer overrun.
+                */
+               reg_val = E1000_READ_REG(hw, E1000_TARC(0));
+               reg_val &= ~E1000_TARC0_CB_MULTIQ_3_REQ;
+               reg_val |= E1000_TARC0_CB_MULTIQ_2_REQ;
+               E1000_WRITE_REG(hw, E1000_TARC(0), reg_val);
+       }
+
        /* This write will effectively turn on the transmit unit. */
        E1000_WRITE_REG(hw, E1000_TCTL, tctl);
 }
@@ -2000,3 +2022,119 @@ em_txq_info_get(struct rte_eth_dev *dev, uint16_t queue_id,
        qinfo->conf.tx_rs_thresh = txq->tx_rs_thresh;
        qinfo->conf.offloads = txq->offloads;
 }
+
+static void
+e1000_flush_tx_ring(struct rte_eth_dev *dev)
+{
+       struct e1000_hw *hw = E1000_DEV_PRIVATE_TO_HW(dev->data->dev_private);
+       volatile struct e1000_data_desc *tx_desc;
+       volatile uint32_t *tdt_reg_addr;
+       uint32_t tdt, tctl, txd_lower = E1000_TXD_CMD_IFCS;
+       uint16_t size = 512;
+       struct em_tx_queue *txq;
+       int i;
+
+       if (dev->data->tx_queues == NULL)
+               return;
+       tctl = E1000_READ_REG(hw, E1000_TCTL);
+       E1000_WRITE_REG(hw, E1000_TCTL, tctl | E1000_TCTL_EN);
+       for (i = 0; i < dev->data->nb_tx_queues &&
+               i < E1000_I219_MAX_TX_QUEUE_NUM; i++) {
+               txq = dev->data->tx_queues[i];
+               tdt = E1000_READ_REG(hw, E1000_TDT(i));
+               if (tdt != txq->tx_tail)
+                       return;
+               tx_desc = &txq->tx_ring[txq->tx_tail];
+               tx_desc->buffer_addr = rte_cpu_to_le_64(txq->tx_ring_phys_addr);
+               tx_desc->lower.data = rte_cpu_to_le_32(txd_lower | size);
+               tx_desc->upper.data = 0;
+
+               rte_cio_wmb();
+               txq->tx_tail++;
+               if (txq->tx_tail == txq->nb_tx_desc)
+                       txq->tx_tail = 0;
+               tdt_reg_addr = E1000_PCI_REG_ADDR(hw, E1000_TDT(i));
+               E1000_PCI_REG_WRITE(tdt_reg_addr, txq->tx_tail);
+               usec_delay(250);
+       }
+}
+
+static void
+e1000_flush_rx_ring(struct rte_eth_dev *dev)
+{
+       uint32_t rctl, rxdctl;
+       struct e1000_hw *hw = E1000_DEV_PRIVATE_TO_HW(dev->data->dev_private);
+       int i;
+
+       rctl = E1000_READ_REG(hw, E1000_RCTL);
+       E1000_WRITE_REG(hw, E1000_RCTL, rctl & ~E1000_RCTL_EN);
+       E1000_WRITE_FLUSH(hw);
+       usec_delay(150);
+
+       for (i = 0; i < dev->data->nb_rx_queues &&
+               i < E1000_I219_MAX_RX_QUEUE_NUM; i++) {
+               rxdctl = E1000_READ_REG(hw, E1000_RXDCTL(i));
+               /* zero the lower 14 bits (prefetch and host thresholds) */
+               rxdctl &= 0xffffc000;
+
+               /* update thresholds: prefetch threshold to 31,
+                * host threshold to 1 and make sure the granularity
+                * is "descriptors" and not "cache lines"
+                */
+               rxdctl |= (0x1F | (1UL << 8) | E1000_RXDCTL_THRESH_UNIT_DESC);
+
+               E1000_WRITE_REG(hw, E1000_RXDCTL(i), rxdctl);
+       }
+       /* momentarily enable the RX ring for the changes to take effect */
+       E1000_WRITE_REG(hw, E1000_RCTL, rctl | E1000_RCTL_EN);
+       E1000_WRITE_FLUSH(hw);
+       usec_delay(150);
+       E1000_WRITE_REG(hw, E1000_RCTL, rctl & ~E1000_RCTL_EN);
+}
+
+/**
+ * em_flush_desc_rings - remove all descriptors from the descriptor rings
+ *
+ * In i219, the descriptor rings must be emptied before resetting/closing the
+ * HW. Failure to do this will cause the HW to enter a unit hang state which
+ * can only be released by PCI reset on the device
+ *
+ */
+
+void
+em_flush_desc_rings(struct rte_eth_dev *dev)
+{
+       uint32_t fextnvm11, tdlen;
+       struct e1000_hw *hw = E1000_DEV_PRIVATE_TO_HW(dev->data->dev_private);
+       struct rte_pci_device *pci_dev = RTE_ETH_DEV_TO_PCI(dev);
+       uint16_t pci_cfg_status = 0;
+       int ret;
+
+       fextnvm11 = E1000_READ_REG(hw, E1000_FEXTNVM11);
+       E1000_WRITE_REG(hw, E1000_FEXTNVM11,
+                       fextnvm11 | E1000_FEXTNVM11_DISABLE_MULR_FIX);
+       tdlen = E1000_READ_REG(hw, E1000_TDLEN(0));
+       ret = rte_pci_read_config(pci_dev, &pci_cfg_status,
+                  sizeof(pci_cfg_status), PCI_CFG_STATUS_REG);
+       if (ret < 0) {
+               PMD_DRV_LOG(ERR, "Failed to read PCI offset 0x%x",
+                           PCI_CFG_STATUS_REG);
+               return;
+       }
+
+       /* do nothing if we're not in faulty state, or if the queue is empty */
+       if ((pci_cfg_status & FLUSH_DESC_REQUIRED) && tdlen) {
+               /* flush desc ring */
+               e1000_flush_tx_ring(dev);
+               ret = rte_pci_read_config(pci_dev, &pci_cfg_status,
+                               sizeof(pci_cfg_status), PCI_CFG_STATUS_REG);
+               if (ret < 0) {
+                       PMD_DRV_LOG(ERR, "Failed to read PCI offset 0x%x",
+                                       PCI_CFG_STATUS_REG);
+                       return;
+               }
+
+               if (pci_cfg_status & FLUSH_DESC_REQUIRED)
+                       e1000_flush_rx_ring(dev);
+       }
+}