drivers: remove direct access to interrupt handle
[dpdk.git] / drivers / net / mlx5 / mlx5_flow_verbs.c
index d04c37f..176d867 100644 (file)
@@ -10,7 +10,7 @@
 
 #include <rte_common.h>
 #include <rte_ether.h>
-#include <rte_ethdev_driver.h>
+#include <ethdev_driver.h>
 #include <rte_flow.h>
 #include <rte_flow_driver.h>
 #include <rte_malloc.h>
@@ -23,7 +23,7 @@
 #include "mlx5_defs.h"
 #include "mlx5.h"
 #include "mlx5_flow.h"
-#include "mlx5_rxtx.h"
+#include "mlx5_rx.h"
 
 #define VERBS_SPEC_INNER(item_flags) \
        (!!((item_flags) & MLX5_FLOW_LAYER_TUNNEL) ? IBV_FLOW_SPEC_INNER : 0)
@@ -39,6 +39,12 @@ static const uint32_t priority_map_5[][MLX5_PRIORITY_MAP_MAX] = {
        { 9, 10, 11 }, { 12, 13, 14 },
 };
 
+/* Verbs specification header. */
+struct ibv_spec_header {
+       enum ibv_flow_spec_type type;
+       uint16_t size;
+};
+
 /**
  * Discover the maximum number of priority available.
  *
@@ -72,12 +78,17 @@ mlx5_flow_discover_priorities(struct rte_eth_dev *dev)
                },
        };
        struct ibv_flow *flow;
-       struct mlx5_hrxq *drop = mlx5_drop_action_create(dev);
+       struct mlx5_hrxq *drop = priv->drop_queue.hrxq;
        uint16_t vprio[] = { 8, 16 };
        int i;
        int priority = 0;
 
-       if (!drop) {
+#if defined(HAVE_MLX5DV_DR_DEVX_PORT) || defined(HAVE_MLX5DV_DR_DEVX_PORT_V35)
+       /* If DevX supported, driver must support 16 verbs flow priorities. */
+       priority = RTE_DIM(priority_map_5);
+       goto out;
+#endif
+       if (!drop->qp) {
                rte_errno = ENOTSUP;
                return -rte_errno;
        }
@@ -89,7 +100,6 @@ mlx5_flow_discover_priorities(struct rte_eth_dev *dev)
                claim_zero(mlx5_glue->destroy_flow(flow));
                priority = vprio[i];
        }
-       mlx5_drop_action_destroy(dev);
        switch (priority) {
        case 8:
                priority = RTE_DIM(priority_map_3);
@@ -104,8 +114,14 @@ mlx5_flow_discover_priorities(struct rte_eth_dev *dev)
                        dev->data->port_id, priority);
                return -rte_errno;
        }
-       DRV_LOG(INFO, "port %u flow maximum priority: %d",
-               dev->data->port_id, priority);
+#if defined(HAVE_MLX5DV_DR_DEVX_PORT) || defined(HAVE_MLX5DV_DR_DEVX_PORT_V35)
+out:
+#endif
+       DRV_LOG(INFO, "port %u supported flow priorities:"
+               " 0-%d for ingress or egress root table,"
+               " 0-%d for non-root table or transfer root table.",
+               dev->data->port_id, priority - 2,
+               MLX5_NON_ROOT_FLOW_MAX_PRIO - 1);
        return priority;
 }
 
@@ -190,7 +206,7 @@ flow_verbs_counter_create(struct rte_eth_dev *dev,
 {
 #if defined(HAVE_IBV_DEVICE_COUNTERS_SET_V42)
        struct mlx5_priv *priv = dev->data->dev_private;
-       struct ibv_context *ctx = priv->sh->ctx;
+       struct ibv_context *ctx = priv->sh->cdev->ctx;
        struct ibv_counter_set_init_attr init = {
                         .counter_set_id = counter->shared_info.id};
 
@@ -202,7 +218,7 @@ flow_verbs_counter_create(struct rte_eth_dev *dev,
        return 0;
 #elif defined(HAVE_IBV_DEVICE_COUNTERS_SET_V45)
        struct mlx5_priv *priv = dev->data->dev_private;
-       struct ibv_context *ctx = priv->sh->ctx;
+       struct ibv_context *ctx = priv->sh->cdev->ctx;
        struct ibv_counters_init_attr init = {0};
        struct ibv_counter_attach_attr attach;
        int ret;
@@ -242,8 +258,6 @@ flow_verbs_counter_create(struct rte_eth_dev *dev,
  *
  * @param[in] dev
  *   Pointer to the Ethernet device structure.
- * @param[in] shared
- *   Indicate if this counter is shared with other flows.
  * @param[in] id
  *   Counter identifier.
  *
@@ -251,21 +265,17 @@ flow_verbs_counter_create(struct rte_eth_dev *dev,
  *   Index to the counter, 0 otherwise and rte_errno is set.
  */
 static uint32_t
-flow_verbs_counter_new(struct rte_eth_dev *dev, uint32_t shared, uint32_t id)
+flow_verbs_counter_new(struct rte_eth_dev *dev, uint32_t id __rte_unused)
 {
        struct mlx5_priv *priv = dev->data->dev_private;
        struct mlx5_flow_counter_mng *cmng = &priv->sh->cmng;
        struct mlx5_flow_counter_pool *pool = NULL;
        struct mlx5_flow_counter *cnt = NULL;
-       union mlx5_l3t_data data;
        uint32_t n_valid = cmng->n_valid;
        uint32_t pool_idx, cnt_idx;
        uint32_t i;
        int ret;
 
-       if (shared && !mlx5_l3t_get_entry(priv->sh->cnt_id_tbl, id, &data) &&
-           data.dword)
-               return data.dword;
        for (pool_idx = 0; pool_idx < n_valid; ++pool_idx) {
                pool = cmng->pools[pool_idx];
                if (!pool)
@@ -312,13 +322,6 @@ flow_verbs_counter_new(struct rte_eth_dev *dev, uint32_t shared, uint32_t id)
        TAILQ_REMOVE(&pool->counters[0], cnt, next);
        i = MLX5_CNT_ARRAY_IDX(pool, cnt);
        cnt_idx = MLX5_MAKE_CNT_IDX(pool_idx, i);
-       if (shared) {
-               data.dword = cnt_idx;
-               if (mlx5_l3t_set_entry(priv->sh->cnt_id_tbl, id, &data))
-                       return 0;
-               cnt->shared_info.id = id;
-               cnt_idx |= MLX5_CNT_SHARED_OFFSET;
-       }
        /* Create counter with Verbs. */
        ret = flow_verbs_counter_create(dev, cnt);
        if (!ret) {
@@ -344,14 +347,10 @@ flow_verbs_counter_new(struct rte_eth_dev *dev, uint32_t shared, uint32_t id)
 static void
 flow_verbs_counter_release(struct rte_eth_dev *dev, uint32_t counter)
 {
-       struct mlx5_priv *priv = dev->data->dev_private;
        struct mlx5_flow_counter_pool *pool;
        struct mlx5_flow_counter *cnt;
 
        cnt = flow_verbs_counter_get_by_idx(dev, counter, &pool);
-       if (IS_SHARED_CNT(counter) &&
-           mlx5_l3t_clear_entry(priv->sh->cnt_id_tbl, cnt->shared_info.id))
-               return;
 #if defined(HAVE_IBV_DEVICE_COUNTERS_SET_V42)
        claim_zero(mlx5_glue->destroy_counter_set
                        ((struct ibv_counter_set *)cnt->dcs_when_active));
@@ -1190,8 +1189,7 @@ flow_verbs_translate_action_count(struct mlx5_flow *dev_flow,
 #endif
 
        if (!flow->counter) {
-               flow->counter = flow_verbs_counter_new(dev, count->shared,
-                                                      count->id);
+               flow->counter = flow_verbs_counter_new(dev, count->id);
                if (!flow->counter)
                        return rte_flow_error_set(error, rte_errno,
                                                  RTE_FLOW_ERROR_TYPE_ACTION,
@@ -1248,7 +1246,8 @@ flow_verbs_validate(struct rte_eth_dev *dev,
        uint64_t last_item = 0;
        uint8_t next_protocol = 0xff;
        uint16_t ether_type = 0;
-       char errstr[32];
+       bool is_empty_vlan = false;
+       uint16_t udp_dport = 0;
 
        if (items == NULL)
                return -1;
@@ -1276,6 +1275,8 @@ flow_verbs_validate(struct rte_eth_dev *dev,
                                ether_type &=
                                        ((const struct rte_flow_item_eth *)
                                         items->mask)->type;
+                               if (ether_type == RTE_BE16(RTE_ETHER_TYPE_VLAN))
+                                       is_empty_vlan = true;
                                ether_type = rte_be_to_cpu_16(ether_type);
                        } else {
                                ether_type = 0;
@@ -1301,6 +1302,7 @@ flow_verbs_validate(struct rte_eth_dev *dev,
                        } else {
                                ether_type = 0;
                        }
+                       is_empty_vlan = false;
                        break;
                case RTE_FLOW_ITEM_TYPE_IPV4:
                        ret = mlx5_flow_validate_item_ipv4
@@ -1353,6 +1355,15 @@ flow_verbs_validate(struct rte_eth_dev *dev,
                        ret = mlx5_flow_validate_item_udp(items, item_flags,
                                                          next_protocol,
                                                          error);
+                       const struct rte_flow_item_udp *spec = items->spec;
+                       const struct rte_flow_item_udp *mask = items->mask;
+                       if (!mask)
+                               mask = &rte_flow_item_udp_mask;
+                       if (spec != NULL)
+                               udp_dport = rte_be_to_cpu_16
+                                               (spec->hdr.dst_port &
+                                                mask->hdr.dst_port);
+
                        if (ret < 0)
                                return ret;
                        last_item = tunnel ? MLX5_FLOW_LAYER_INNER_L4_UDP :
@@ -1370,8 +1381,9 @@ flow_verbs_validate(struct rte_eth_dev *dev,
                                             MLX5_FLOW_LAYER_OUTER_L4_TCP;
                        break;
                case RTE_FLOW_ITEM_TYPE_VXLAN:
-                       ret = mlx5_flow_validate_item_vxlan(items, item_flags,
-                                                           error);
+                       ret = mlx5_flow_validate_item_vxlan(dev, udp_dport,
+                                                           items, item_flags,
+                                                           attr, error);
                        if (ret < 0)
                                return ret;
                        last_item = MLX5_FLOW_LAYER_VXLAN;
@@ -1399,15 +1411,23 @@ flow_verbs_validate(struct rte_eth_dev *dev,
                                return ret;
                        last_item = MLX5_FLOW_LAYER_MPLS;
                        break;
+               case RTE_FLOW_ITEM_TYPE_ICMP:
+               case RTE_FLOW_ITEM_TYPE_ICMP6:
+                       return rte_flow_error_set(error, ENOTSUP,
+                                                 RTE_FLOW_ERROR_TYPE_ITEM,
+                                                 NULL, "ICMP/ICMP6 "
+                                                 "item not supported");
                default:
-                       snprintf(errstr, sizeof(errstr), "item type %d not supported",
-                                items->type);
                        return rte_flow_error_set(error, ENOTSUP,
                                                  RTE_FLOW_ERROR_TYPE_ITEM,
-                                                 NULL, errstr);
+                                                 NULL, "item not supported");
                }
                item_flags |= last_item;
        }
+       if (is_empty_vlan)
+               return rte_flow_error_set(error, ENOTSUP,
+                                                RTE_FLOW_ERROR_TYPE_ITEM, NULL,
+                   "VLAN matching without vid specification is not supported");
        for (; actions->type != RTE_FLOW_ACTION_TYPE_END; actions++) {
                switch (actions->type) {
                case RTE_FLOW_ACTION_TYPE_VOID:
@@ -1626,7 +1646,9 @@ flow_verbs_prepare(struct rte_eth_dev *dev,
        struct mlx5_flow *dev_flow;
        struct mlx5_flow_handle *dev_handle;
        struct mlx5_priv *priv = dev->data->dev_private;
+       struct mlx5_flow_workspace *wks = mlx5_flow_get_thread_workspace();
 
+       MLX5_ASSERT(wks);
        size += flow_verbs_get_actions_size(actions);
        size += flow_verbs_get_items_size(items);
        if (size > MLX5_VERBS_MAX_SPEC_ACT_SIZE) {
@@ -1636,7 +1658,7 @@ flow_verbs_prepare(struct rte_eth_dev *dev,
                return NULL;
        }
        /* In case of corrupting the memory. */
-       if (priv->flow_idx >= MLX5_NUM_MAX_DEV_FLOWS) {
+       if (wks->flow_idx >= MLX5_NUM_MAX_DEV_FLOWS) {
                rte_flow_error_set(error, ENOSPC,
                                   RTE_FLOW_ERROR_TYPE_UNSPECIFIED, NULL,
                                   "not free temporary device flow");
@@ -1650,8 +1672,8 @@ flow_verbs_prepare(struct rte_eth_dev *dev,
                                   "not enough memory to create flow handle");
                return NULL;
        }
-       /* No multi-thread supporting. */
-       dev_flow = &((struct mlx5_flow *)priv->inter_flows)[priv->flow_idx++];
+       MLX5_ASSERT(wks->flow_idx + 1 < RTE_DIM(wks->flows));
+       dev_flow = &wks->flows[wks->flow_idx++];
        dev_flow->handle = dev_handle;
        dev_flow->handle_idx = handle_idx;
        /* Memcpy is used, only size needs to be cleared to 0. */
@@ -1695,12 +1717,12 @@ flow_verbs_translate(struct rte_eth_dev *dev,
        uint64_t priority = attr->priority;
        uint32_t subpriority = 0;
        struct mlx5_priv *priv = dev->data->dev_private;
-       struct mlx5_flow_rss_desc *rss_desc = &((struct mlx5_flow_rss_desc *)
-                                             priv->rss_desc)
-                                             [!!priv->flow_nested_idx];
-       char errstr[32];
+       struct mlx5_flow_workspace *wks = mlx5_flow_get_thread_workspace();
+       struct mlx5_flow_rss_desc *rss_desc;
 
-       if (priority == MLX5_FLOW_PRIO_RSVD)
+       MLX5_ASSERT(wks);
+       rss_desc = &wks->rss_desc;
+       if (priority == MLX5_FLOW_LOWEST_PRIO_INDICATOR)
                priority = priv->config.flow_prio - 1;
        for (; actions->type != RTE_FLOW_ACTION_TYPE_END; actions++) {
                int ret;
@@ -1799,9 +1821,10 @@ flow_verbs_translate(struct rte_eth_dev *dev,
                        flow_verbs_translate_item_tcp(dev_flow, items,
                                                      item_flags);
                        subpriority = MLX5_PRIORITY_MAP_L4;
-                       dev_flow->hash_fields |=
-                               mlx5_flow_hashfields_adjust
-                                       (rss_desc, tunnel, ETH_RSS_TCP,
+                       if (dev_flow->hash_fields != 0)
+                               dev_flow->hash_fields |=
+                                       mlx5_flow_hashfields_adjust
+                                       (rss_desc, tunnel, RTE_ETH_RSS_TCP,
                                         (IBV_RX_HASH_SRC_PORT_TCP |
                                          IBV_RX_HASH_DST_PORT_TCP));
                        item_flags |= tunnel ? MLX5_FLOW_LAYER_INNER_L4_TCP :
@@ -1811,9 +1834,10 @@ flow_verbs_translate(struct rte_eth_dev *dev,
                        flow_verbs_translate_item_udp(dev_flow, items,
                                                      item_flags);
                        subpriority = MLX5_PRIORITY_MAP_L4;
-                       dev_flow->hash_fields |=
-                               mlx5_flow_hashfields_adjust
-                                       (rss_desc, tunnel, ETH_RSS_UDP,
+                       if (dev_flow->hash_fields != 0)
+                               dev_flow->hash_fields |=
+                                       mlx5_flow_hashfields_adjust
+                                       (rss_desc, tunnel, RTE_ETH_RSS_UDP,
                                         (IBV_RX_HASH_SRC_PORT_UDP |
                                          IBV_RX_HASH_DST_PORT_UDP));
                        item_flags |= tunnel ? MLX5_FLOW_LAYER_INNER_L4_UDP :
@@ -1844,11 +1868,9 @@ flow_verbs_translate(struct rte_eth_dev *dev,
                        item_flags |= MLX5_FLOW_LAYER_MPLS;
                        break;
                default:
-                       snprintf(errstr, sizeof(errstr), "item type %d not supported",
-                                items->type);
                        return rte_flow_error_set(error, ENOTSUP,
                                                  RTE_FLOW_ERROR_TYPE_ITEM,
-                                                 NULL, errstr);
+                                                 NULL, "item not supported");
                }
        }
        dev_flow->handle->layers = item_flags;
@@ -1883,15 +1905,10 @@ flow_verbs_remove(struct rte_eth_dev *dev, struct rte_flow *flow)
                        handle->drv_flow = NULL;
                }
                /* hrxq is union, don't touch it only the flag is set. */
-               if (handle->rix_hrxq) {
-                       if (handle->fate_action == MLX5_FLOW_FATE_DROP) {
-                               mlx5_drop_action_destroy(dev);
-                               handle->rix_hrxq = 0;
-                       } else if (handle->fate_action ==
-                                  MLX5_FLOW_FATE_QUEUE) {
-                               mlx5_hrxq_release(dev, handle->rix_hrxq);
-                               handle->rix_hrxq = 0;
-                       }
+               if (handle->rix_hrxq &&
+                   handle->fate_action == MLX5_FLOW_FATE_QUEUE) {
+                       mlx5_hrxq_release(dev, handle->rix_hrxq);
+                       handle->rix_hrxq = 0;
                }
                if (handle->vf_vlan.tag && handle->vf_vlan.created)
                        mlx5_vlan_vmwa_release(dev, &handle->vf_vlan);
@@ -1956,41 +1973,26 @@ flow_verbs_apply(struct rte_eth_dev *dev, struct rte_flow *flow,
        uint32_t dev_handles;
        int err;
        int idx;
+       struct mlx5_flow_workspace *wks = mlx5_flow_get_thread_workspace();
 
-       for (idx = priv->flow_idx - 1; idx >= priv->flow_nested_idx; idx--) {
-               dev_flow = &((struct mlx5_flow *)priv->inter_flows)[idx];
+       MLX5_ASSERT(wks);
+       for (idx = wks->flow_idx - 1; idx >= 0; idx--) {
+               dev_flow = &wks->flows[idx];
                handle = dev_flow->handle;
                if (handle->fate_action == MLX5_FLOW_FATE_DROP) {
-                       hrxq = mlx5_drop_action_create(dev);
-                       if (!hrxq) {
-                               rte_flow_error_set
-                                       (error, errno,
-                                        RTE_FLOW_ERROR_TYPE_UNSPECIFIED, NULL,
-                                        "cannot get drop hash queue");
-                               goto error;
-                       }
+                       MLX5_ASSERT(priv->drop_queue.hrxq);
+                       hrxq = priv->drop_queue.hrxq;
                } else {
                        uint32_t hrxq_idx;
-                       struct mlx5_flow_rss_desc *rss_desc =
-                               &((struct mlx5_flow_rss_desc *)priv->rss_desc)
-                               [!!priv->flow_nested_idx];
+                       struct mlx5_flow_rss_desc *rss_desc = &wks->rss_desc;
 
                        MLX5_ASSERT(rss_desc->queue_num);
-                       hrxq_idx = mlx5_hrxq_get(dev, rss_desc->key,
-                                                MLX5_RSS_HASH_KEY_LEN,
-                                                dev_flow->hash_fields,
-                                                rss_desc->queue,
-                                                rss_desc->queue_num);
-                       if (!hrxq_idx)
-                               hrxq_idx = mlx5_hrxq_new
-                                               (dev, rss_desc->key,
-                                                MLX5_RSS_HASH_KEY_LEN,
-                                                dev_flow->hash_fields,
-                                                rss_desc->queue,
-                                                rss_desc->queue_num,
-                                                !!(handle->layers &
-                                                MLX5_FLOW_LAYER_TUNNEL),
-                                                false);
+                       rss_desc->key_len = MLX5_RSS_HASH_KEY_LEN;
+                       rss_desc->hash_fields = dev_flow->hash_fields;
+                       rss_desc->tunnel = !!(handle->layers &
+                                             MLX5_FLOW_LAYER_TUNNEL);
+                       rss_desc->shared_rss = 0;
+                       hrxq_idx = mlx5_hrxq_get(dev, rss_desc);
                        hrxq = mlx5_ipool_get(priv->sh->ipool[MLX5_IPOOL_HRXQ],
                                              hrxq_idx);
                        if (!hrxq) {
@@ -2029,15 +2031,10 @@ error:
        SILIST_FOREACH(priv->sh->ipool[MLX5_IPOOL_MLX5_FLOW], flow->dev_handles,
                       dev_handles, handle, next) {
                /* hrxq is union, don't touch it only the flag is set. */
-               if (handle->rix_hrxq) {
-                       if (handle->fate_action == MLX5_FLOW_FATE_DROP) {
-                               mlx5_drop_action_destroy(dev);
-                               handle->rix_hrxq = 0;
-                       } else if (handle->fate_action ==
-                                  MLX5_FLOW_FATE_QUEUE) {
-                               mlx5_hrxq_release(dev, handle->rix_hrxq);
-                               handle->rix_hrxq = 0;
-                       }
+               if (handle->rix_hrxq &&
+                   handle->fate_action == MLX5_FLOW_FATE_QUEUE) {
+                       mlx5_hrxq_release(dev, handle->rix_hrxq);
+                       handle->rix_hrxq = 0;
                }
                if (handle->vf_vlan.tag && handle->vf_vlan.created)
                        mlx5_vlan_vmwa_release(dev, &handle->vf_vlan);