net/sfc: add pattern parsing stub to MAE backend
[dpdk.git] / drivers / net / mlx5 / mlx5.c
index 81476e2..74a537b 100644 (file)
@@ -9,9 +9,6 @@
 #include <stdint.h>
 #include <stdlib.h>
 #include <errno.h>
-#include <net/if.h>
-#include <sys/mman.h>
-#include <linux/rtnetlink.h>
 
 #include <rte_malloc.h>
 #include <rte_ethdev_driver.h>
@@ -244,6 +241,28 @@ static const struct mlx5_indexed_pool_config mlx5_ipool_cfg[] = {
                .free = mlx5_free,
                .type = "mlx5_jump_ipool",
        },
+       {
+               .size = sizeof(struct mlx5_flow_dv_sample_resource),
+               .trunk_size = 64,
+               .grow_trunk = 3,
+               .grow_shift = 2,
+               .need_lock = 0,
+               .release_mem_en = 1,
+               .malloc = mlx5_malloc,
+               .free = mlx5_free,
+               .type = "mlx5_sample_ipool",
+       },
+       {
+               .size = sizeof(struct mlx5_flow_dv_dest_array_resource),
+               .trunk_size = 64,
+               .grow_trunk = 3,
+               .grow_shift = 2,
+               .need_lock = 0,
+               .release_mem_en = 1,
+               .malloc = mlx5_malloc,
+               .free = mlx5_free,
+               .type = "mlx5_dest_array_ipool",
+       },
 #endif
        {
                .size = sizeof(struct mlx5_flow_meter),
@@ -707,6 +726,144 @@ mlx5_flex_parser_ecpri_release(struct rte_eth_dev *dev)
        prf->obj = NULL;
 }
 
+/*
+ * Allocate Rx and Tx UARs in robust fashion.
+ * This routine handles the following UAR allocation issues:
+ *
+ *  - tries to allocate the UAR with the most appropriate memory
+ *    mapping type from the ones supported by the host
+ *
+ *  - tries to allocate the UAR with non-NULL base address
+ *    OFED 5.0.x and Upstream rdma_core before v29 returned the NULL as
+ *    UAR base address if UAR was not the first object in the UAR page.
+ *    It caused the PMD failure and we should try to get another UAR
+ *    till we get the first one with non-NULL base address returned.
+ */
+static int
+mlx5_alloc_rxtx_uars(struct mlx5_dev_ctx_shared *sh,
+                    const struct mlx5_dev_config *config)
+{
+       uint32_t uar_mapping, retry;
+       int err = 0;
+       void *base_addr;
+
+       for (retry = 0; retry < MLX5_ALLOC_UAR_RETRY; ++retry) {
+#ifdef MLX5DV_UAR_ALLOC_TYPE_NC
+               /* Control the mapping type according to the settings. */
+               uar_mapping = (config->dbnc == MLX5_TXDB_NCACHED) ?
+                             MLX5DV_UAR_ALLOC_TYPE_NC :
+                             MLX5DV_UAR_ALLOC_TYPE_BF;
+#else
+               RTE_SET_USED(config);
+               /*
+                * It seems we have no way to control the memory mapping type
+                * for the UAR, the default "Write-Combining" type is supposed.
+                * The UAR initialization on queue creation queries the
+                * actual mapping type done by Verbs/kernel and setups the
+                * PMD datapath accordingly.
+                */
+               uar_mapping = 0;
+#endif
+               sh->tx_uar = mlx5_glue->devx_alloc_uar(sh->ctx, uar_mapping);
+#ifdef MLX5DV_UAR_ALLOC_TYPE_NC
+               if (!sh->tx_uar &&
+                   uar_mapping == MLX5DV_UAR_ALLOC_TYPE_BF) {
+                       if (config->dbnc == MLX5_TXDB_CACHED ||
+                           config->dbnc == MLX5_TXDB_HEURISTIC)
+                               DRV_LOG(WARNING, "Devarg tx_db_nc setting "
+                                                "is not supported by DevX");
+                       /*
+                        * In some environments like virtual machine
+                        * the Write Combining mapped might be not supported
+                        * and UAR allocation fails. We try "Non-Cached"
+                        * mapping for the case. The tx_burst routines take
+                        * the UAR mapping type into account on UAR setup
+                        * on queue creation.
+                        */
+                       DRV_LOG(WARNING, "Failed to allocate Tx DevX UAR (BF)");
+                       uar_mapping = MLX5DV_UAR_ALLOC_TYPE_NC;
+                       sh->tx_uar = mlx5_glue->devx_alloc_uar
+                                                       (sh->ctx, uar_mapping);
+               } else if (!sh->tx_uar &&
+                          uar_mapping == MLX5DV_UAR_ALLOC_TYPE_NC) {
+                       if (config->dbnc == MLX5_TXDB_NCACHED)
+                               DRV_LOG(WARNING, "Devarg tx_db_nc settings "
+                                                "is not supported by DevX");
+                       /*
+                        * If Verbs/kernel does not support "Non-Cached"
+                        * try the "Write-Combining".
+                        */
+                       DRV_LOG(WARNING, "Failed to allocate Tx DevX UAR (NC)");
+                       uar_mapping = MLX5DV_UAR_ALLOC_TYPE_BF;
+                       sh->tx_uar = mlx5_glue->devx_alloc_uar
+                                                       (sh->ctx, uar_mapping);
+               }
+#endif
+               if (!sh->tx_uar) {
+                       DRV_LOG(ERR, "Failed to allocate Tx DevX UAR (BF/NC)");
+                       err = ENOMEM;
+                       goto exit;
+               }
+               base_addr = mlx5_os_get_devx_uar_base_addr(sh->tx_uar);
+               if (base_addr)
+                       break;
+               /*
+                * The UARs are allocated by rdma_core within the
+                * IB device context, on context closure all UARs
+                * will be freed, should be no memory/object leakage.
+                */
+               DRV_LOG(WARNING, "Retrying to allocate Tx DevX UAR");
+               sh->tx_uar = NULL;
+       }
+       /* Check whether we finally succeeded with valid UAR allocation. */
+       if (!sh->tx_uar) {
+               DRV_LOG(ERR, "Failed to allocate Tx DevX UAR (NULL base)");
+               err = ENOMEM;
+               goto exit;
+       }
+       for (retry = 0; retry < MLX5_ALLOC_UAR_RETRY; ++retry) {
+               uar_mapping = 0;
+               sh->devx_rx_uar = mlx5_glue->devx_alloc_uar
+                                                       (sh->ctx, uar_mapping);
+#ifdef MLX5DV_UAR_ALLOC_TYPE_NC
+               if (!sh->devx_rx_uar &&
+                   uar_mapping == MLX5DV_UAR_ALLOC_TYPE_BF) {
+                       /*
+                        * Rx UAR is used to control interrupts only,
+                        * should be no datapath noticeable impact,
+                        * can try "Non-Cached" mapping safely.
+                        */
+                       DRV_LOG(WARNING, "Failed to allocate Rx DevX UAR (BF)");
+                       uar_mapping = MLX5DV_UAR_ALLOC_TYPE_NC;
+                       sh->devx_rx_uar = mlx5_glue->devx_alloc_uar
+                                                       (sh->ctx, uar_mapping);
+               }
+#endif
+               if (!sh->devx_rx_uar) {
+                       DRV_LOG(ERR, "Failed to allocate Rx DevX UAR (BF/NC)");
+                       err = ENOMEM;
+                       goto exit;
+               }
+               base_addr = mlx5_os_get_devx_uar_base_addr(sh->devx_rx_uar);
+               if (base_addr)
+                       break;
+               /*
+                * The UARs are allocated by rdma_core within the
+                * IB device context, on context closure all UARs
+                * will be freed, should be no memory/object leakage.
+                */
+               DRV_LOG(WARNING, "Retrying to allocate Rx DevX UAR");
+               sh->devx_rx_uar = NULL;
+       }
+       /* Check whether we finally succeeded with valid UAR allocation. */
+       if (!sh->devx_rx_uar) {
+               DRV_LOG(ERR, "Failed to allocate Rx DevX UAR (NULL base)");
+               err = ENOMEM;
+       }
+exit:
+       return err;
+}
+
 /**
  * Allocate shared device context. If there is multiport device the
  * master and representors will share this context, if there is single
@@ -790,6 +947,14 @@ mlx5_alloc_shared_dev_ctx(const struct mlx5_dev_spawn_data *spawn,
                goto error;
        }
        if (sh->devx) {
+               /* Query the EQN for this core. */
+               err = mlx5_glue->devx_query_eqn(sh->ctx, 0, &sh->eqn);
+               if (err) {
+                       rte_errno = errno;
+                       DRV_LOG(ERR, "Failed to query event queue number %d.",
+                               rte_errno);
+                       goto error;
+               }
                err = mlx5_os_get_pdn(sh->pd, &sh->pdn);
                if (err) {
                        DRV_LOG(ERR, "Fail to extract pdn from PD");
@@ -808,18 +973,14 @@ mlx5_alloc_shared_dev_ctx(const struct mlx5_dev_spawn_data *spawn,
                        err = ENOMEM;
                        goto error;
                }
-               sh->tx_uar = mlx5_glue->devx_alloc_uar(sh->ctx, 0);
-               if (!sh->tx_uar) {
-                       DRV_LOG(ERR, "Failed to allocate DevX UAR.");
-                       err = ENOMEM;
+               err = mlx5_alloc_rxtx_uars(sh, config);
+               if (err)
                        goto error;
-               }
-               sh->devx_rx_uar = mlx5_glue->devx_alloc_uar(sh->ctx, 0);
-               if (!sh->devx_rx_uar) {
-                       DRV_LOG(ERR, "Failed to allocate Rx DevX UAR.");
-                       err = ENOMEM;
-                       goto error;
-               }
+               MLX5_ASSERT(sh->tx_uar);
+               MLX5_ASSERT(mlx5_os_get_devx_uar_base_addr(sh->tx_uar));
+
+               MLX5_ASSERT(sh->devx_rx_uar);
+               MLX5_ASSERT(mlx5_os_get_devx_uar_base_addr(sh->devx_rx_uar));
        }
        sh->flow_id_pool = mlx5_flow_id_pool_alloc
                                        ((1 << HAIRPIN_FLOW_ID_BITS) - 1);
@@ -875,20 +1036,16 @@ error:
        pthread_mutex_destroy(&sh->txpp.mutex);
        pthread_mutex_unlock(&mlx5_dev_ctx_list_mutex);
        MLX5_ASSERT(sh);
-       if (sh->cnt_id_tbl) {
+       if (sh->cnt_id_tbl)
                mlx5_l3t_destroy(sh->cnt_id_tbl);
-               sh->cnt_id_tbl = NULL;
-       }
-       if (sh->tx_uar) {
-               mlx5_glue->devx_free_uar(sh->tx_uar);
-               sh->tx_uar = NULL;
-       }
        if (sh->tis)
                claim_zero(mlx5_devx_cmd_destroy(sh->tis));
        if (sh->td)
                claim_zero(mlx5_devx_cmd_destroy(sh->td));
        if (sh->devx_rx_uar)
                mlx5_glue->devx_free_uar(sh->devx_rx_uar);
+       if (sh->tx_uar)
+               mlx5_glue->devx_free_uar(sh->tx_uar);
        if (sh->pd)
                claim_zero(mlx5_glue->dealloc_pd(sh->pd));
        if (sh->ctx)
@@ -1214,7 +1371,7 @@ mlx5_proc_priv_uninit(struct rte_eth_dev *dev)
  * @param dev
  *   Pointer to Ethernet device structure.
  */
-void
+int
 mlx5_dev_close(struct rte_eth_dev *dev)
 {
        struct mlx5_priv *priv = dev->data->dev_private;
@@ -1224,14 +1381,14 @@ mlx5_dev_close(struct rte_eth_dev *dev)
        if (rte_eal_process_type() == RTE_PROC_SECONDARY) {
                /* Check if process_private released. */
                if (!dev->process_private)
-                       return;
+                       return 0;
                mlx5_tx_uar_uninit_secondary(dev);
                mlx5_proc_priv_uninit(dev);
                rte_eth_dev_release_port(dev);
-               return;
+               return 0;
        }
        if (!priv->sh)
-               return;
+               return 0;
        DRV_LOG(DEBUG, "port %u closing device \"%s\"",
                dev->data->port_id,
                ((priv->sh->ctx != NULL) ?
@@ -1284,9 +1441,7 @@ mlx5_dev_close(struct rte_eth_dev *dev)
        if (priv->reta_idx != NULL)
                mlx5_free(priv->reta_idx);
        if (priv->config.vf)
-               mlx5_nl_mac_addr_flush(priv->nl_socket_route, mlx5_ifindex(dev),
-                                      dev->data->mac_addrs,
-                                      MLX5_MAX_MAC_ADDRESSES, priv->mac_own);
+               mlx5_os_mac_addr_flush(dev);
        if (priv->nl_socket_route >= 0)
                close(priv->nl_socket_route);
        if (priv->nl_socket_rdma >= 0)
@@ -1324,7 +1479,7 @@ mlx5_dev_close(struct rte_eth_dev *dev)
        /*
         * Free the shared context in last turn, because the cleanup
         * routines above may use some shared fields, like
-        * mlx5_nl_mac_addr_flush() uses ibdev_path for retrieveing
+        * mlx5_os_mac_addr_flush() uses ibdev_path for retrieveing
         * ifindex if Netlink fails.
         */
        mlx5_free_shared_dev_ctx(priv->sh);
@@ -1354,6 +1509,7 @@ mlx5_dev_close(struct rte_eth_dev *dev)
         * it is freed when dev_private is freed.
         */
        dev->data->mac_addrs = NULL;
+       return 0;
 }
 
 /**
@@ -1884,6 +2040,7 @@ static int
 mlx5_pci_remove(struct rte_pci_device *pci_dev)
 {
        uint16_t port_id;
+       int ret = 0;
 
        RTE_ETH_FOREACH_DEV_OF(port_id, &pci_dev->device) {
                /*
@@ -1891,11 +2048,11 @@ mlx5_pci_remove(struct rte_pci_device *pci_dev)
                 * call the close function explicitly for secondary process.
                 */
                if (rte_eal_process_type() == RTE_PROC_SECONDARY)
-                       mlx5_dev_close(&rte_eth_devices[port_id]);
+                       ret |= mlx5_dev_close(&rte_eth_devices[port_id]);
                else
-                       rte_eth_dev_close(port_id);
+                       ret |= rte_eth_dev_close(port_id);
        }
-       return 0;
+       return ret == 0 ? 0 : -EIO;
 }
 
 static const struct rte_pci_id mlx5_pci_id_map[] = {