net/mlx5: add flow translate function
[dpdk.git] / drivers / net / mlx5 / mlx5.c
index f17bd08..7493180 100644 (file)
@@ -46,6 +46,7 @@
 #include "mlx5_defs.h"
 #include "mlx5_glue.h"
 #include "mlx5_mr.h"
+#include "mlx5_flow.h"
 
 /* Device parameter to enable RX completion queue compression. */
 #define MLX5_RXQ_CQE_COMP_EN "rxq_cqe_comp_en"
@@ -92,6 +93,9 @@
 /* Activate Netlink support in VF mode. */
 #define MLX5_VF_NL_EN "vf_nl_en"
 
+/* Select port representors to instantiate. */
+#define MLX5_REPRESENTOR "representor"
+
 #ifndef HAVE_IBV_MLX5_MOD_MPW
 #define MLX5DV_CONTEXT_FLAGS_MPW_ALLOWED (1 << 2)
 #define MLX5DV_CONTEXT_FLAGS_ENHANCED_MPW (1 << 3)
@@ -239,6 +243,7 @@ mlx5_dev_close(struct rte_eth_dev *dev)
        /* In case mlx5_dev_stop() has not been called. */
        mlx5_dev_interrupt_handler_uninstall(dev);
        mlx5_traffic_disable(dev);
+       mlx5_flow_flush(dev, NULL);
        /* Prevent crashes when queues are still in use. */
        dev->rx_pkt_burst = removed_rx_burst;
        dev->tx_pkt_burst = removed_tx_burst;
@@ -258,7 +263,6 @@ mlx5_dev_close(struct rte_eth_dev *dev)
                priv->txqs_n = 0;
                priv->txqs = NULL;
        }
-       mlx5_flow_delete_drop_queue(dev);
        mlx5_mprq_free_mp(dev);
        mlx5_mr_release(dev);
        if (priv->pd != NULL) {
@@ -279,6 +283,8 @@ mlx5_dev_close(struct rte_eth_dev *dev)
                close(priv->nl_socket_route);
        if (priv->nl_socket_rdma >= 0)
                close(priv->nl_socket_rdma);
+       if (priv->mnl_socket)
+               mlx5_nl_flow_socket_destroy(priv->mnl_socket);
        ret = mlx5_hrxq_ibv_verify(dev);
        if (ret)
                DRV_LOG(WARNING, "port %u some hash Rx queue still remain",
@@ -394,6 +400,10 @@ const struct eth_dev_ops mlx5_dev_ops_isolate = {
        .dev_set_link_down = mlx5_set_link_down,
        .dev_set_link_up = mlx5_set_link_up,
        .dev_close = mlx5_dev_close,
+       .promiscuous_enable = mlx5_promiscuous_enable,
+       .promiscuous_disable = mlx5_promiscuous_disable,
+       .allmulticast_enable = mlx5_allmulticast_enable,
+       .allmulticast_disable = mlx5_allmulticast_disable,
        .link_update = mlx5_link_update,
        .stats_get = mlx5_stats_get,
        .stats_reset = mlx5_stats_reset,
@@ -443,6 +453,9 @@ mlx5_args_check(const char *key, const char *val, void *opaque)
        struct mlx5_dev_config *config = opaque;
        unsigned long tmp;
 
+       /* No-op, port representors are processed in mlx5_dev_spawn(). */
+       if (!strcmp(MLX5_REPRESENTOR, key))
+               return 0;
        errno = 0;
        tmp = strtoul(val, NULL, 0);
        if (errno) {
@@ -465,7 +478,7 @@ mlx5_args_check(const char *key, const char *val, void *opaque)
        } else if (strcmp(MLX5_TXQS_MIN_INLINE, key) == 0) {
                config->txqs_inline = tmp;
        } else if (strcmp(MLX5_TXQ_MPW_EN, key) == 0) {
-               config->mps = !!tmp ? config->mps : 0;
+               config->mps = !!tmp;
        } else if (strcmp(MLX5_TXQ_MPW_HDR_DSEG_EN, key) == 0) {
                config->mpw_hdr_dseg = !!tmp;
        } else if (strcmp(MLX5_TXQ_MAX_INLINE_LEN, key) == 0) {
@@ -515,6 +528,7 @@ mlx5_args(struct mlx5_dev_config *config, struct rte_devargs *devargs)
                MLX5_RX_VEC_EN,
                MLX5_L3_VXLAN_EN,
                MLX5_VF_NL_EN,
+               MLX5_REPRESENTOR,
                NULL,
        };
        struct rte_kvargs *kvlist;
@@ -555,11 +569,13 @@ static struct rte_pci_driver mlx5_driver;
 static void *uar_base;
 
 static int
-find_lower_va_bound(const struct rte_memseg_list *msl __rte_unused,
+find_lower_va_bound(const struct rte_memseg_list *msl,
                const struct rte_memseg *ms, void *arg)
 {
        void **addr = arg;
 
+       if (msl->external)
+               return 0;
        if (*addr == NULL)
                *addr = ms->addr;
        else
@@ -591,7 +607,7 @@ mlx5_uar_init_primary(struct rte_eth_dev *dev)
        rte_memseg_walk(find_lower_va_bound, &addr);
 
        /* keep distance to hugepages to minimize potential conflicts. */
-       addr = RTE_PTR_SUB(addr, MLX5_UAR_OFFSET + MLX5_UAR_SIZE);
+       addr = RTE_PTR_SUB(addr, (uintptr_t)(MLX5_UAR_OFFSET + MLX5_UAR_SIZE));
        /* anonymous mmap, no real memory consumption. */
        addr = mmap(addr, MLX5_UAR_SIZE,
                    PROT_NONE, MAP_PRIVATE | MAP_ANONYMOUS, -1, 0);
@@ -672,7 +688,9 @@ mlx5_uar_init_secondary(struct rte_eth_dev *dev)
  *
  * @return
  *   A valid Ethernet device object on success, NULL otherwise and rte_errno
- *   is set.
+ *   is set. The following error is defined:
+ *
+ *   EBUSY: device is not supposed to be spawned.
  */
 static struct rte_eth_dev *
 mlx5_dev_spawn(struct rte_device *dpdk_dev,
@@ -687,6 +705,7 @@ mlx5_dev_spawn(struct rte_device *dpdk_dev,
        struct mlx5dv_context dv_attr = { .comp_mask = 0 };
        struct mlx5_dev_config config = {
                .vf = !!vf,
+               .mps = MLX5_ARG_UNSET,
                .tx_vec_en = 1,
                .rx_vec_en = 1,
                .mpw_hdr_dseg = 0,
@@ -709,7 +728,6 @@ mlx5_dev_spawn(struct rte_device *dpdk_dev,
        unsigned int tunnel_en = 0;
        unsigned int mpls_en = 0;
        unsigned int swp = 0;
-       unsigned int verb_priorities = 0;
        unsigned int mprq = 0;
        unsigned int mprq_min_stride_size_n = 0;
        unsigned int mprq_max_stride_size_n = 0;
@@ -723,6 +741,26 @@ mlx5_dev_spawn(struct rte_device *dpdk_dev,
        int own_domain_id = 0;
        unsigned int i;
 
+       /* Determine if this port representor is supposed to be spawned. */
+       if (switch_info->representor && dpdk_dev->devargs) {
+               struct rte_eth_devargs eth_da;
+
+               err = rte_eth_devargs_parse(dpdk_dev->devargs->args, &eth_da);
+               if (err) {
+                       rte_errno = -err;
+                       DRV_LOG(ERR, "failed to process device arguments: %s",
+                               strerror(rte_errno));
+                       return NULL;
+               }
+               for (i = 0; i < eth_da.nb_representor_ports; ++i)
+                       if (eth_da.representor_ports[i] ==
+                           (uint16_t)switch_info->port_name)
+                               break;
+               if (i == eth_da.nb_representor_ports) {
+                       rte_errno = EBUSY;
+                       return NULL;
+               }
+       }
        /* Prepare shared data between primary and secondary process. */
        mlx5_prepare_shared_data();
        errno = 0;
@@ -757,7 +795,6 @@ mlx5_dev_spawn(struct rte_device *dpdk_dev,
                DRV_LOG(DEBUG, "MPW isn't supported");
                mps = MLX5_MPW_DISABLED;
        }
-       config.mps = mps;
 #ifdef HAVE_IBV_MLX5_MOD_SWP
        if (dv_attr.comp_mask & MLX5DV_CONTEXT_MASK_SWP)
                swp = dv_attr.sw_parsing_caps.sw_parsing_offloads;
@@ -911,9 +948,15 @@ mlx5_dev_spawn(struct rte_device *dpdk_dev,
        priv->device_attr = attr;
        priv->pd = pd;
        priv->mtu = ETHER_MTU;
+#ifndef RTE_ARCH_64
+       /* Initialize UAR access locks for 32bit implementations. */
+       rte_spinlock_init(&priv->uar_lock_cq);
+       for (i = 0; i < MLX5_UAR_PAGE_NUM_MAX; i++)
+               rte_spinlock_init(&priv->uar_lock[i]);
+#endif
        /* Some internal functions rely on Netlink sockets, open them now. */
-       priv->nl_socket_rdma = mlx5_nl_init(0, NETLINK_RDMA);
-       priv->nl_socket_route = mlx5_nl_init(RTMGRP_LINK, NETLINK_ROUTE);
+       priv->nl_socket_rdma = mlx5_nl_init(NETLINK_RDMA);
+       priv->nl_socket_route = mlx5_nl_init(NETLINK_ROUTE);
        priv->nl_sn = 0;
        priv->representor = !!switch_info->representor;
        priv->domain_id = RTE_ETH_DEV_SWITCH_DOMAIN_ID_INVALID;
@@ -995,13 +1038,15 @@ mlx5_dev_spawn(struct rte_device *dpdk_dev,
                       (1 << IBV_QPT_RAW_PACKET)));
        if (config.tso)
                config.tso_max_payload_sz = attr.tso_caps.max_tso;
-       if (config.mps && !mps) {
-               DRV_LOG(ERR,
-                       "multi-packet send not supported on this device"
-                       " (" MLX5_TXQ_MPW_EN ")");
-               err = ENOTSUP;
-               goto error;
-       }
+       /*
+        * MPW is disabled by default, while the Enhanced MPW is enabled
+        * by default.
+        */
+       if (config.mps == MLX5_ARG_UNSET)
+               config.mps = (mps == MLX5_MPW_ENHANCED) ? MLX5_MPW_ENHANCED :
+                                                         MLX5_MPW_DISABLED;
+       else
+               config.mps = config.mps ? mps : MLX5_MPW_DISABLED;
        DRV_LOG(INFO, "%sMPS is %s",
                config.mps == MLX5_MPW_ENHANCED ? "enhanced " : "",
                config.mps != MLX5_MPW_DISABLED ? "enabled" : "disabled");
@@ -1088,6 +1133,34 @@ mlx5_dev_spawn(struct rte_device *dpdk_dev,
        claim_zero(mlx5_mac_addr_add(eth_dev, &mac, 0, 0));
        if (vf && config.vf_nl_en)
                mlx5_nl_mac_addr_sync(eth_dev);
+       priv->mnl_socket = mlx5_nl_flow_socket_create();
+       if (!priv->mnl_socket) {
+               err = -rte_errno;
+               DRV_LOG(WARNING,
+                       "flow rules relying on switch offloads will not be"
+                       " supported: cannot open libmnl socket: %s",
+                       strerror(rte_errno));
+       } else {
+               struct rte_flow_error error;
+               unsigned int ifindex = mlx5_ifindex(eth_dev);
+
+               if (!ifindex) {
+                       err = -rte_errno;
+                       error.message =
+                               "cannot retrieve network interface index";
+               } else {
+                       err = mlx5_nl_flow_init(priv->mnl_socket, ifindex,
+                                               &error);
+               }
+               if (err) {
+                       DRV_LOG(WARNING,
+                               "flow rules relying on switch offloads will"
+                               " not be supported: %s: %s",
+                               error.message, strerror(rte_errno));
+                       mlx5_nl_flow_socket_destroy(priv->mnl_socket);
+                       priv->mnl_socket = NULL;
+               }
+       }
        TAILQ_INIT(&priv->flows);
        TAILQ_INIT(&priv->ctrl_flows);
        /* Hint libmlx5 to use PMD allocator for data plane resources */
@@ -1110,24 +1183,12 @@ mlx5_dev_spawn(struct rte_device *dpdk_dev,
        mlx5_link_update(eth_dev, 0);
        /* Store device configuration on private structure. */
        priv->config = config;
-       /* Create drop queue. */
-       err = mlx5_flow_create_drop_queue(eth_dev);
-       if (err) {
-               DRV_LOG(ERR, "port %u drop queue allocation failed: %s",
-                       eth_dev->data->port_id, strerror(rte_errno));
-               err = rte_errno;
-               goto error;
-       }
        /* Supported Verbs flow priority number detection. */
-       if (verb_priorities == 0)
-               verb_priorities = mlx5_get_max_verbs_prio(eth_dev);
-       if (verb_priorities < MLX5_VERBS_FLOW_PRIO_8) {
-               DRV_LOG(ERR, "port %u wrong Verbs flow priorities: %u",
-                       eth_dev->data->port_id, verb_priorities);
-               err = ENOTSUP;
+       err = mlx5_flow_discover_priorities(eth_dev);
+       if (err < 0)
                goto error;
-       }
-       priv->config.max_verbs_prio = verb_priorities;
+       priv->config.flow_prio = err;
+       mlx5_flow_init_driver_ops(eth_dev);
        /*
         * Once the device is added to the list of memory event
         * callback, its global MR cache table cannot be expanded
@@ -1153,6 +1214,8 @@ error:
                        close(priv->nl_socket_route);
                if (priv->nl_socket_rdma >= 0)
                        close(priv->nl_socket_rdma);
+               if (priv->mnl_socket)
+                       mlx5_nl_flow_socket_destroy(priv->mnl_socket);
                if (own_domain_id)
                        claim_zero(rte_eth_switch_domain_free(priv->domain_id));
                rte_free(priv);
@@ -1265,8 +1328,8 @@ mlx5_pci_probe(struct rte_pci_driver *pci_drv __rte_unused,
        ibv_match[n] = NULL;
 
        struct mlx5_dev_spawn_data list[n];
-       int nl_route = n ? mlx5_nl_init(0, NETLINK_ROUTE) : -1;
-       int nl_rdma = n ? mlx5_nl_init(0, NETLINK_RDMA) : -1;
+       int nl_route = n ? mlx5_nl_init(NETLINK_ROUTE) : -1;
+       int nl_rdma = n ? mlx5_nl_init(NETLINK_RDMA) : -1;
        unsigned int i;
        unsigned int u;
 
@@ -1277,7 +1340,8 @@ mlx5_pci_probe(struct rte_pci_driver *pci_drv __rte_unused,
         * Netlink calls assuming kernel drivers are recent enough to
         * support them.
         *
-        * In the event of identification failure through Netlink, either:
+        * In the event of identification failure through Netlink, try again
+        * through sysfs, then either:
         *
         * 1. No device matches (n == 0), complain and bail out.
         * 2. A single IB device matches (n == 1) and is not a representor,
@@ -1296,7 +1360,9 @@ mlx5_pci_probe(struct rte_pci_driver *pci_drv __rte_unused,
                if (nl_route < 0 ||
                    !list[i].ifindex ||
                    mlx5_nl_switch_info(nl_route, list[i].ifindex,
-                                       &list[i].info)) {
+                                       &list[i].info) ||
+                   ((!list[i].info.representor && !list[i].info.master) &&
+                    mlx5_sysfs_switch_info(list[i].ifindex, &list[i].info))) {
                        list[i].ifindex = 0;
                        memset(&list[i].info, 0, sizeof(list[i].info));
                        continue;
@@ -1343,8 +1409,12 @@ mlx5_pci_probe(struct rte_pci_driver *pci_drv __rte_unused,
 
                list[i].eth_dev = mlx5_dev_spawn
                        (&pci_dev->device, list[i].ibv_dev, vf, &list[i].info);
-               if (!list[i].eth_dev)
-                       break;
+               if (!list[i].eth_dev) {
+                       if (rte_errno != EBUSY)
+                               break;
+                       /* Device is disabled, ignore it. */
+                       continue;
+               }
                restore = list[i].eth_dev->data->dev_flags;
                rte_eth_copy_pci_info(list[i].eth_dev, pci_dev);
                /* Restore non-PCI flags cleared by the above call. */
@@ -1370,6 +1440,8 @@ mlx5_pci_probe(struct rte_pci_driver *pci_drv __rte_unused,
                ret = -rte_errno;
                /* Roll back. */
                while (i--) {
+                       if (!list[i].eth_dev)
+                               continue;
                        mlx5_dev_close(list[i].eth_dev);
                        if (rte_eal_process_type() == RTE_PROC_PRIMARY)
                                rte_free(list[i].eth_dev->data->dev_private);
@@ -1420,6 +1492,10 @@ static const struct rte_pci_id mlx5_pci_id_map[] = {
                RTE_PCI_DEVICE(PCI_VENDOR_ID_MELLANOX,
                               PCI_DEVICE_ID_MELLANOX_CONNECTX5BF)
        },
+       {
+               RTE_PCI_DEVICE(PCI_VENDOR_ID_MELLANOX,
+                              PCI_DEVICE_ID_MELLANOX_CONNECTX5BFVF)
+       },
        {
                .vendor_id = 0
        }