net/ice/base: add more macros for FDID priority
[dpdk.git] / drivers / net / mlx5 / mlx5_flow_verbs.c
index c787c98..ccd3395 100644 (file)
 #include <rte_malloc.h>
 #include <rte_ip.h>
 
-#include "mlx5.h"
+#include <mlx5_glue.h>
+#include <mlx5_prm.h>
+
 #include "mlx5_defs.h"
+#include "mlx5.h"
 #include "mlx5_flow.h"
-#include "mlx5_glue.h"
-#include "mlx5_prm.h"
 #include "mlx5_rxtx.h"
 
 #define VERBS_SPEC_INNER(item_flags) \
@@ -252,16 +253,17 @@ flow_verbs_counter_query(struct rte_eth_dev *dev __rte_unused,
  *   Size in bytes of the specification to copy.
  */
 static void
-flow_verbs_spec_add(struct mlx5_flow_verbs *verbs, void *src, unsigned int size)
+flow_verbs_spec_add(struct mlx5_flow_verbs_workspace *verbs,
+                   void *src, unsigned int size)
 {
        void *dst;
 
        if (!verbs)
                return;
-       assert(verbs->specs);
+       MLX5_ASSERT(verbs->specs);
        dst = (void *)(verbs->specs + verbs->size);
        memcpy(dst, src, size);
-       ++verbs->attr->num_of_specs;
+       ++verbs->attr.num_of_specs;
        verbs->size += size;
 }
 
@@ -390,9 +392,9 @@ flow_verbs_translate_item_vlan(struct mlx5_flow *dev_flow,
        if (!(item_flags & l2m))
                flow_verbs_spec_add(&dev_flow->verbs, &eth, size);
        else
-               flow_verbs_item_vlan_update(dev_flow->verbs.attr, &eth);
+               flow_verbs_item_vlan_update(&dev_flow->verbs.attr, &eth);
        if (!tunnel)
-               dev_flow->verbs.vf_vlan.tag =
+               dev_flow->handle->vf_vlan.tag =
                        rte_be_to_cpu_16(spec->tci) & 0x0fff;
 }
 
@@ -493,14 +495,12 @@ flow_verbs_translate_item_ipv6(struct mlx5_flow *dev_flow,
                ipv6.val.traffic_class = (vtc_flow_val & RTE_IPV6_HDR_TC_MASK) >>
                                         RTE_IPV6_HDR_TC_SHIFT;
                ipv6.val.next_hdr = spec->hdr.proto;
-               ipv6.val.hop_limit = spec->hdr.hop_limits;
                ipv6.mask.flow_label =
                        rte_cpu_to_be_32((vtc_flow_mask & RTE_IPV6_HDR_FL_MASK) >>
                                         RTE_IPV6_HDR_FL_SHIFT);
                ipv6.mask.traffic_class = (vtc_flow_mask & RTE_IPV6_HDR_TC_MASK) >>
                                          RTE_IPV6_HDR_TC_SHIFT;
                ipv6.mask.next_hdr = mask->hdr.proto;
-               ipv6.mask.hop_limit = mask->hdr.hop_limits;
                /* Remove unwanted bits from values. */
                for (i = 0; i < RTE_DIM(ipv6.val.src_ip); ++i) {
                        ipv6.val.src_ip[i] &= ipv6.mask.src_ip[i];
@@ -509,7 +509,6 @@ flow_verbs_translate_item_ipv6(struct mlx5_flow *dev_flow,
                ipv6.val.flow_label &= ipv6.mask.flow_label;
                ipv6.val.traffic_class &= ipv6.mask.traffic_class;
                ipv6.val.next_hdr &= ipv6.mask.next_hdr;
-               ipv6.val.hop_limit &= ipv6.mask.hop_limit;
        }
        flow_verbs_spec_add(&dev_flow->verbs, &ipv6, size);
 }
@@ -745,7 +744,7 @@ flow_verbs_translate_item_gre(struct mlx5_flow *dev_flow,
                              const struct rte_flow_item *item __rte_unused,
                              uint64_t item_flags)
 {
-       struct mlx5_flow_verbs *verbs = &dev_flow->verbs;
+       struct mlx5_flow_verbs_workspace *verbs = &dev_flow->verbs;
 #ifndef HAVE_IBV_DEVICE_MPLS_SUPPORT
        unsigned int size = sizeof(struct ibv_flow_spec_tunnel);
        struct ibv_flow_spec_tunnel tunnel = {
@@ -775,11 +774,11 @@ flow_verbs_translate_item_gre(struct mlx5_flow *dev_flow,
        }
 #endif
        if (item_flags & MLX5_FLOW_LAYER_OUTER_L3_IPV4)
-               flow_verbs_item_gre_ip_protocol_update(verbs->attr,
+               flow_verbs_item_gre_ip_protocol_update(&verbs->attr,
                                                       IBV_FLOW_SPEC_IPV4_EXT,
                                                       IPPROTO_GRE);
        else
-               flow_verbs_item_gre_ip_protocol_update(verbs->attr,
+               flow_verbs_item_gre_ip_protocol_update(&verbs->attr,
                                                       IBV_FLOW_SPEC_IPV6,
                                                       IPPROTO_GRE);
        flow_verbs_spec_add(verbs, &tunnel, size);
@@ -1255,6 +1254,18 @@ flow_verbs_validate(struct rte_eth_dev *dev,
                                                  "action not supported");
                }
        }
+       /*
+        * Validate the drop action mutual exclusion with other actions.
+        * Drop action is mutually-exclusive with any other action, except for
+        * Count action.
+        */
+       if ((action_flags & MLX5_FLOW_ACTION_DROP) &&
+           (action_flags & ~(MLX5_FLOW_ACTION_DROP | MLX5_FLOW_ACTION_COUNT)))
+               return rte_flow_error_set(error, EINVAL,
+                                         RTE_FLOW_ERROR_TYPE_ACTION, NULL,
+                                         "Drop action is mutually-exclusive "
+                                         "with any other action, except for "
+                                         "Count action");
        if (!(action_flags & MLX5_FLOW_FATE_ACTIONS))
                return rte_flow_error_set(error, EINVAL,
                                          RTE_FLOW_ERROR_TYPE_ACTION, actions,
@@ -1374,6 +1385,8 @@ flow_verbs_get_items_size(const struct rte_flow_item items[])
  * The required size is calculate based on the actions and items. This function
  * also returns the detected actions and items for later use.
  *
+ * @param[in] dev
+ *   Pointer to Ethernet device.
  * @param[in] attr
  *   Pointer to the flow attributes.
  * @param[in] items
@@ -1388,27 +1401,47 @@ flow_verbs_get_items_size(const struct rte_flow_item items[])
  *   is set.
  */
 static struct mlx5_flow *
-flow_verbs_prepare(const struct rte_flow_attr *attr __rte_unused,
+flow_verbs_prepare(struct rte_eth_dev *dev,
+                  const struct rte_flow_attr *attr __rte_unused,
                   const struct rte_flow_item items[],
                   const struct rte_flow_action actions[],
                   struct rte_flow_error *error)
 {
-       size_t size = sizeof(struct mlx5_flow) + sizeof(struct ibv_flow_attr);
+       size_t size = 0;
        struct mlx5_flow *dev_flow;
+       struct mlx5_flow_handle *dev_handle;
+       struct mlx5_priv *priv = dev->data->dev_private;
 
        size += flow_verbs_get_actions_size(actions);
        size += flow_verbs_get_items_size(items);
-       dev_flow = rte_calloc(__func__, 1, size, 0);
-       if (!dev_flow) {
+       if (size > MLX5_VERBS_MAX_SPEC_ACT_SIZE) {
+               rte_flow_error_set(error, E2BIG,
+                                  RTE_FLOW_ERROR_TYPE_UNSPECIFIED, NULL,
+                                  "Verbs spec/action size too large");
+               return NULL;
+       }
+       /* In case of corrupting the memory. */
+       if (priv->flow_idx >= MLX5_NUM_MAX_DEV_FLOWS) {
+               rte_flow_error_set(error, ENOSPC,
+                                  RTE_FLOW_ERROR_TYPE_UNSPECIFIED, NULL,
+                                  "not free temporary device flow");
+               return NULL;
+       }
+       dev_handle = rte_calloc(__func__, 1, MLX5_FLOW_HANDLE_VERBS_SIZE, 0);
+       if (!dev_handle) {
                rte_flow_error_set(error, ENOMEM,
                                   RTE_FLOW_ERROR_TYPE_UNSPECIFIED, NULL,
-                                  "not enough memory to create flow");
+                                  "not enough memory to create flow handle");
                return NULL;
        }
-       dev_flow->verbs.attr = (void *)(dev_flow + 1);
-       dev_flow->verbs.specs = (void *)(dev_flow->verbs.attr + 1);
+       /* No multi-thread supporting. */
+       dev_flow = &((struct mlx5_flow *)priv->inter_flows)[priv->flow_idx++];
+       dev_flow->handle = dev_handle;
+       /* Memcpy is used, only size needs to be cleared to 0. */
+       dev_flow->verbs.size = 0;
+       dev_flow->verbs.attr.num_of_specs = 0;
        dev_flow->ingress = attr->ingress;
-       dev_flow->transfer = attr->transfer;
+       /* Need to set transfer attribute: not supported in Verbs mode. */
        return dev_flow;
 }
 
@@ -1488,7 +1521,7 @@ flow_verbs_translate(struct rte_eth_dev *dev,
                                                  "action not supported");
                }
        }
-       dev_flow->actions = action_flags;
+       dev_flow->handle->act_flags = action_flags;
        for (; items->type != RTE_FLOW_ITEM_TYPE_END; items++) {
                int tunnel = !!(item_flags & MLX5_FLOW_LAYER_TUNNEL);
 
@@ -1590,10 +1623,11 @@ flow_verbs_translate(struct rte_eth_dev *dev,
                                                  "item not supported");
                }
        }
-       dev_flow->layers = item_flags;
-       dev_flow->verbs.attr->priority =
+       dev_flow->handle->layers = item_flags;
+       /* Other members of attr will be ignored. */
+       dev_flow->verbs.attr.priority =
                mlx5_flow_adjust_priority(dev, priority, subpriority);
-       dev_flow->verbs.attr->port = (uint8_t)priv->ibv_port;
+       dev_flow->verbs.attr.port = (uint8_t)priv->ibv_port;
        return 0;
 }
 
@@ -1608,28 +1642,24 @@ flow_verbs_translate(struct rte_eth_dev *dev,
 static void
 flow_verbs_remove(struct rte_eth_dev *dev, struct rte_flow *flow)
 {
-       struct mlx5_flow_verbs *verbs;
-       struct mlx5_flow *dev_flow;
+       struct mlx5_flow_handle *handle;
 
        if (!flow)
                return;
-       LIST_FOREACH(dev_flow, &flow->dev_flows, next) {
-               verbs = &dev_flow->verbs;
-               if (verbs->flow) {
-                       claim_zero(mlx5_glue->destroy_flow(verbs->flow));
-                       verbs->flow = NULL;
+       LIST_FOREACH(handle, &flow->dev_handles, next) {
+               if (handle->ib_flow) {
+                       claim_zero(mlx5_glue->destroy_flow(handle->ib_flow));
+                       handle->ib_flow = NULL;
                }
-               if (verbs->hrxq) {
-                       if (dev_flow->actions & MLX5_FLOW_ACTION_DROP)
+               if (handle->hrxq) {
+                       if (handle->act_flags & MLX5_FLOW_ACTION_DROP)
                                mlx5_hrxq_drop_release(dev);
                        else
-                               mlx5_hrxq_release(dev, verbs->hrxq);
-                       verbs->hrxq = NULL;
-               }
-               if (dev_flow->verbs.vf_vlan.tag &&
-                   dev_flow->verbs.vf_vlan.created) {
-                       mlx5_vlan_vmwa_release(dev, &dev_flow->verbs.vf_vlan);
+                               mlx5_hrxq_release(dev, handle->hrxq);
+                       handle->hrxq = NULL;
                }
+               if (handle->vf_vlan.tag && handle->vf_vlan.created)
+                       mlx5_vlan_vmwa_release(dev, &handle->vf_vlan);
        }
 }
 
@@ -1644,15 +1674,15 @@ flow_verbs_remove(struct rte_eth_dev *dev, struct rte_flow *flow)
 static void
 flow_verbs_destroy(struct rte_eth_dev *dev, struct rte_flow *flow)
 {
-       struct mlx5_flow *dev_flow;
+       struct mlx5_flow_handle *handle;
 
        if (!flow)
                return;
        flow_verbs_remove(dev, flow);
-       while (!LIST_EMPTY(&flow->dev_flows)) {
-               dev_flow = LIST_FIRST(&flow->dev_flows);
-               LIST_REMOVE(dev_flow, next);
-               rte_free(dev_flow);
+       while (!LIST_EMPTY(&flow->dev_handles)) {
+               handle = LIST_FIRST(&flow->dev_handles);
+               LIST_REMOVE(handle, next);
+               rte_free(handle);
        }
        if (flow->counter) {
                flow_verbs_counter_release(dev, flow->counter);
@@ -1678,15 +1708,17 @@ flow_verbs_apply(struct rte_eth_dev *dev, struct rte_flow *flow,
                 struct rte_flow_error *error)
 {
        struct mlx5_priv *priv = dev->data->dev_private;
-       struct mlx5_flow_verbs *verbs;
+       struct mlx5_flow_handle *handle;
        struct mlx5_flow *dev_flow;
        int err;
-
-       LIST_FOREACH(dev_flow, &flow->dev_flows, next) {
-               verbs = &dev_flow->verbs;
-               if (dev_flow->actions & MLX5_FLOW_ACTION_DROP) {
-                       verbs->hrxq = mlx5_hrxq_drop_new(dev);
-                       if (!verbs->hrxq) {
+       int idx;
+
+       for (idx = priv->flow_idx - 1; idx >= 0; idx--) {
+               dev_flow = &((struct mlx5_flow *)priv->inter_flows)[idx];
+               handle = dev_flow->handle;
+               if (handle->act_flags & MLX5_FLOW_ACTION_DROP) {
+                       handle->hrxq = mlx5_hrxq_drop_new(dev);
+                       if (!handle->hrxq) {
                                rte_flow_error_set
                                        (error, errno,
                                         RTE_FLOW_ERROR_TYPE_UNSPECIFIED, NULL,
@@ -1696,7 +1728,7 @@ flow_verbs_apply(struct rte_eth_dev *dev, struct rte_flow *flow,
                } else {
                        struct mlx5_hrxq *hrxq;
 
-                       assert(flow->rss.queue);
+                       MLX5_ASSERT(flow->rss.queue);
                        hrxq = mlx5_hrxq_get(dev, flow->rss.key,
                                             MLX5_RSS_HASH_KEY_LEN,
                                             dev_flow->hash_fields,
@@ -1704,12 +1736,12 @@ flow_verbs_apply(struct rte_eth_dev *dev, struct rte_flow *flow,
                                             flow->rss.queue_num);
                        if (!hrxq)
                                hrxq = mlx5_hrxq_new(dev, flow->rss.key,
-                                                    MLX5_RSS_HASH_KEY_LEN,
-                                                    dev_flow->hash_fields,
-                                                    (*flow->rss.queue),
-                                                    flow->rss.queue_num,
-                                                    !!(dev_flow->layers &
-                                                      MLX5_FLOW_LAYER_TUNNEL));
+                                               MLX5_RSS_HASH_KEY_LEN,
+                                               dev_flow->hash_fields,
+                                               (*flow->rss.queue),
+                                               flow->rss.queue_num,
+                                               !!(handle->layers &
+                                               MLX5_FLOW_LAYER_TUNNEL));
                        if (!hrxq) {
                                rte_flow_error_set
                                        (error, rte_errno,
@@ -1717,11 +1749,11 @@ flow_verbs_apply(struct rte_eth_dev *dev, struct rte_flow *flow,
                                         "cannot get hash queue");
                                goto error;
                        }
-                       verbs->hrxq = hrxq;
+                       handle->hrxq = hrxq;
                }
-               verbs->flow = mlx5_glue->create_flow(verbs->hrxq->qp,
-                                                    verbs->attr);
-               if (!verbs->flow) {
+               handle->ib_flow = mlx5_glue->create_flow(handle->hrxq->qp,
+                                                    &dev_flow->verbs.attr);
+               if (!handle->ib_flow) {
                        rte_flow_error_set(error, errno,
                                           RTE_FLOW_ERROR_TYPE_UNSPECIFIED,
                                           NULL,
@@ -1729,33 +1761,29 @@ flow_verbs_apply(struct rte_eth_dev *dev, struct rte_flow *flow,
                        goto error;
                }
                if (priv->vmwa_context &&
-                   dev_flow->verbs.vf_vlan.tag &&
-                   !dev_flow->verbs.vf_vlan.created) {
+                   handle->vf_vlan.tag && !handle->vf_vlan.created) {
                        /*
                         * The rule contains the VLAN pattern.
                         * For VF we are going to create VLAN
                         * interface to make hypervisor set correct
                         * e-Switch vport context.
                         */
-                       mlx5_vlan_vmwa_acquire(dev, &dev_flow->verbs.vf_vlan);
+                       mlx5_vlan_vmwa_acquire(dev, &handle->vf_vlan);
                }
        }
        return 0;
 error:
        err = rte_errno; /* Save rte_errno before cleanup. */
-       LIST_FOREACH(dev_flow, &flow->dev_flows, next) {
-               verbs = &dev_flow->verbs;
-               if (verbs->hrxq) {
-                       if (dev_flow->actions & MLX5_FLOW_ACTION_DROP)
+       LIST_FOREACH(handle, &flow->dev_handles, next) {
+               if (handle->hrxq) {
+                       if (handle->act_flags & MLX5_FLOW_ACTION_DROP)
                                mlx5_hrxq_drop_release(dev);
                        else
-                               mlx5_hrxq_release(dev, verbs->hrxq);
-                       verbs->hrxq = NULL;
-               }
-               if (dev_flow->verbs.vf_vlan.tag &&
-                   dev_flow->verbs.vf_vlan.created) {
-                       mlx5_vlan_vmwa_release(dev, &dev_flow->verbs.vf_vlan);
+                               mlx5_hrxq_release(dev, handle->hrxq);
+                       handle->hrxq = NULL;
                }
+               if (handle->vf_vlan.tag && handle->vf_vlan.created)
+                       mlx5_vlan_vmwa_release(dev, &handle->vf_vlan);
        }
        rte_errno = err; /* Restore rte_errno. */
        return -rte_errno;