net: add rte prefix to IP defines
[dpdk.git] / drivers / net / mlx5 / mlx5_flow_verbs.c
index 43fcd0d..2f4c80c 100644 (file)
@@ -20,7 +20,6 @@
 
 #include <rte_common.h>
 #include <rte_ether.h>
-#include <rte_eth_ctrl.h>
 #include <rte_ethdev_driver.h>
 #include <rte_flow.h>
 #include <rte_flow_driver.h>
 
 #include "mlx5.h"
 #include "mlx5_defs.h"
-#include "mlx5_prm.h"
-#include "mlx5_glue.h"
 #include "mlx5_flow.h"
+#include "mlx5_glue.h"
+#include "mlx5_prm.h"
+#include "mlx5_rxtx.h"
 
 #define VERBS_SPEC_INNER(item_flags) \
        (!!((item_flags) & MLX5_FLOW_LAYER_TUNNEL) ? IBV_FLOW_SPEC_INNER : 0)
@@ -55,23 +55,26 @@ flow_verbs_counter_create(struct rte_eth_dev *dev,
                          struct mlx5_flow_counter *counter)
 {
 #if defined(HAVE_IBV_DEVICE_COUNTERS_SET_V42)
-       struct priv *priv = dev->data->dev_private;
+       struct mlx5_priv *priv = dev->data->dev_private;
+       struct ibv_context *ctx = priv->sh->ctx;
        struct ibv_counter_set_init_attr init = {
                         .counter_set_id = counter->id};
 
-       counter->cs = mlx5_glue->create_counter_set(priv->ctx, &init);
+       counter->cs = mlx5_glue->create_counter_set(ctx, &init);
        if (!counter->cs) {
                rte_errno = ENOTSUP;
                return -ENOTSUP;
        }
        return 0;
 #elif defined(HAVE_IBV_DEVICE_COUNTERS_SET_V45)
-       struct priv *priv = dev->data->dev_private;
+       struct mlx5_priv *priv = dev->data->dev_private;
+       struct ibv_context *ctx = priv->sh->ctx;
        struct ibv_counters_init_attr init = {0};
-       struct ibv_counter_attach_attr attach = {0};
+       struct ibv_counter_attach_attr attach;
        int ret;
 
-       counter->cs = mlx5_glue->create_counters(priv->ctx, &init);
+       memset(&attach, 0, sizeof(attach));
+       counter->cs = mlx5_glue->create_counters(ctx, &init);
        if (!counter->cs) {
                rte_errno = ENOTSUP;
                return -ENOTSUP;
@@ -116,17 +119,17 @@ flow_verbs_counter_create(struct rte_eth_dev *dev,
 static struct mlx5_flow_counter *
 flow_verbs_counter_new(struct rte_eth_dev *dev, uint32_t shared, uint32_t id)
 {
-       struct priv *priv = dev->data->dev_private;
+       struct mlx5_priv *priv = dev->data->dev_private;
        struct mlx5_flow_counter *cnt;
        int ret;
 
-       LIST_FOREACH(cnt, &priv->flow_counters, next) {
-               if (!cnt->shared || cnt->shared != shared)
-                       continue;
-               if (cnt->id != id)
-                       continue;
-               cnt->ref_cnt++;
-               return cnt;
+       if (shared) {
+               LIST_FOREACH(cnt, &priv->flow_counters, next) {
+                       if (cnt->shared && cnt->id == id) {
+                               cnt->ref_cnt++;
+                               return cnt;
+                       }
+               }
        }
        cnt = rte_calloc(__func__, 1, sizeof(*cnt), 0);
        if (!cnt) {
@@ -287,14 +290,18 @@ flow_verbs_translate_item_eth(struct mlx5_flow *dev_flow,
        if (spec) {
                unsigned int i;
 
-               memcpy(&eth.val.dst_mac, spec->dst.addr_bytes, ETHER_ADDR_LEN);
-               memcpy(&eth.val.src_mac, spec->src.addr_bytes, ETHER_ADDR_LEN);
+               memcpy(&eth.val.dst_mac, spec->dst.addr_bytes,
+                       RTE_ETHER_ADDR_LEN);
+               memcpy(&eth.val.src_mac, spec->src.addr_bytes,
+                       RTE_ETHER_ADDR_LEN);
                eth.val.ether_type = spec->type;
-               memcpy(&eth.mask.dst_mac, mask->dst.addr_bytes, ETHER_ADDR_LEN);
-               memcpy(&eth.mask.src_mac, mask->src.addr_bytes, ETHER_ADDR_LEN);
+               memcpy(&eth.mask.dst_mac, mask->dst.addr_bytes,
+                       RTE_ETHER_ADDR_LEN);
+               memcpy(&eth.mask.src_mac, mask->src.addr_bytes,
+                       RTE_ETHER_ADDR_LEN);
                eth.mask.ether_type = mask->type;
                /* Remove unwanted bits from values. */
-               for (i = 0; i < ETHER_ADDR_LEN; ++i) {
+               for (i = 0; i < RTE_ETHER_ADDR_LEN; ++i) {
                        eth.val.dst_mac[i] &= eth.mask.dst_mac[i];
                        eth.val.src_mac[i] &= eth.mask.src_mac[i];
                }
@@ -473,17 +480,17 @@ flow_verbs_translate_item_ipv6(struct mlx5_flow *dev_flow,
                vtc_flow_val = rte_be_to_cpu_32(spec->hdr.vtc_flow);
                vtc_flow_mask = rte_be_to_cpu_32(mask->hdr.vtc_flow);
                ipv6.val.flow_label =
-                       rte_cpu_to_be_32((vtc_flow_val & IPV6_HDR_FL_MASK) >>
-                                        IPV6_HDR_FL_SHIFT);
-               ipv6.val.traffic_class = (vtc_flow_val & IPV6_HDR_TC_MASK) >>
-                                        IPV6_HDR_TC_SHIFT;
+                       rte_cpu_to_be_32((vtc_flow_val & RTE_IPV6_HDR_FL_MASK) >>
+                                        RTE_IPV6_HDR_FL_SHIFT);
+               ipv6.val.traffic_class = (vtc_flow_val & RTE_IPV6_HDR_TC_MASK) >>
+                                        RTE_IPV6_HDR_TC_SHIFT;
                ipv6.val.next_hdr = spec->hdr.proto;
                ipv6.val.hop_limit = spec->hdr.hop_limits;
                ipv6.mask.flow_label =
-                       rte_cpu_to_be_32((vtc_flow_mask & IPV6_HDR_FL_MASK) >>
-                                        IPV6_HDR_FL_SHIFT);
-               ipv6.mask.traffic_class = (vtc_flow_mask & IPV6_HDR_TC_MASK) >>
-                                         IPV6_HDR_TC_SHIFT;
+                       rte_cpu_to_be_32((vtc_flow_mask & RTE_IPV6_HDR_FL_MASK) >>
+                                        RTE_IPV6_HDR_FL_SHIFT);
+               ipv6.mask.traffic_class = (vtc_flow_mask & RTE_IPV6_HDR_TC_MASK) >>
+                                         RTE_IPV6_HDR_TC_SHIFT;
                ipv6.mask.next_hdr = mask->hdr.proto;
                ipv6.mask.hop_limit = mask->hdr.hop_limits;
                /* Remove unwanted bits from values. */
@@ -1017,6 +1024,7 @@ flow_verbs_validate(struct rte_eth_dev *dev,
        int ret;
        uint64_t action_flags = 0;
        uint64_t item_flags = 0;
+       uint64_t last_item = 0;
        uint8_t next_protocol = 0xff;
 
        if (items == NULL)
@@ -1036,46 +1044,60 @@ flow_verbs_validate(struct rte_eth_dev *dev,
                                                          error);
                        if (ret < 0)
                                return ret;
-                       item_flags |= tunnel ? MLX5_FLOW_LAYER_INNER_L2 :
-                                              MLX5_FLOW_LAYER_OUTER_L2;
+                       last_item = tunnel ? MLX5_FLOW_LAYER_INNER_L2 :
+                                            MLX5_FLOW_LAYER_OUTER_L2;
                        break;
                case RTE_FLOW_ITEM_TYPE_VLAN:
                        ret = mlx5_flow_validate_item_vlan(items, item_flags,
                                                           error);
                        if (ret < 0)
                                return ret;
-                       item_flags |= tunnel ? (MLX5_FLOW_LAYER_INNER_L2 |
-                                               MLX5_FLOW_LAYER_INNER_VLAN) :
-                                              (MLX5_FLOW_LAYER_OUTER_L2 |
-                                               MLX5_FLOW_LAYER_OUTER_VLAN);
+                       last_item = tunnel ? (MLX5_FLOW_LAYER_INNER_L2 |
+                                             MLX5_FLOW_LAYER_INNER_VLAN) :
+                                            (MLX5_FLOW_LAYER_OUTER_L2 |
+                                             MLX5_FLOW_LAYER_OUTER_VLAN);
                        break;
                case RTE_FLOW_ITEM_TYPE_IPV4:
                        ret = mlx5_flow_validate_item_ipv4(items, item_flags,
-                                                          error);
+                                                          NULL, error);
                        if (ret < 0)
                                return ret;
-                       item_flags |= tunnel ? MLX5_FLOW_LAYER_INNER_L3_IPV4 :
-                                              MLX5_FLOW_LAYER_OUTER_L3_IPV4;
+                       last_item = tunnel ? MLX5_FLOW_LAYER_INNER_L3_IPV4 :
+                                            MLX5_FLOW_LAYER_OUTER_L3_IPV4;
                        if (items->mask != NULL &&
                            ((const struct rte_flow_item_ipv4 *)
-                            items->mask)->hdr.next_proto_id)
+                            items->mask)->hdr.next_proto_id) {
                                next_protocol =
                                        ((const struct rte_flow_item_ipv4 *)
                                         (items->spec))->hdr.next_proto_id;
+                               next_protocol &=
+                                       ((const struct rte_flow_item_ipv4 *)
+                                        (items->mask))->hdr.next_proto_id;
+                       } else {
+                               /* Reset for inner layer. */
+                               next_protocol = 0xff;
+                       }
                        break;
                case RTE_FLOW_ITEM_TYPE_IPV6:
                        ret = mlx5_flow_validate_item_ipv6(items, item_flags,
-                                                          error);
+                                                          NULL, error);
                        if (ret < 0)
                                return ret;
-                       item_flags |= tunnel ? MLX5_FLOW_LAYER_INNER_L3_IPV6 :
-                                              MLX5_FLOW_LAYER_OUTER_L3_IPV6;
+                       last_item = tunnel ? MLX5_FLOW_LAYER_INNER_L3_IPV6 :
+                                            MLX5_FLOW_LAYER_OUTER_L3_IPV6;
                        if (items->mask != NULL &&
                            ((const struct rte_flow_item_ipv6 *)
-                            items->mask)->hdr.proto)
+                            items->mask)->hdr.proto) {
                                next_protocol =
                                        ((const struct rte_flow_item_ipv6 *)
                                         items->spec)->hdr.proto;
+                               next_protocol &=
+                                       ((const struct rte_flow_item_ipv6 *)
+                                        items->mask)->hdr.proto;
+                       } else {
+                               /* Reset for inner layer. */
+                               next_protocol = 0xff;
+                       }
                        break;
                case RTE_FLOW_ITEM_TYPE_UDP:
                        ret = mlx5_flow_validate_item_udp(items, item_flags,
@@ -1083,8 +1105,8 @@ flow_verbs_validate(struct rte_eth_dev *dev,
                                                          error);
                        if (ret < 0)
                                return ret;
-                       item_flags |= tunnel ? MLX5_FLOW_LAYER_INNER_L4_UDP :
-                                              MLX5_FLOW_LAYER_OUTER_L4_UDP;
+                       last_item = tunnel ? MLX5_FLOW_LAYER_INNER_L4_UDP :
+                                            MLX5_FLOW_LAYER_OUTER_L4_UDP;
                        break;
                case RTE_FLOW_ITEM_TYPE_TCP:
                        ret = mlx5_flow_validate_item_tcp
@@ -1094,15 +1116,15 @@ flow_verbs_validate(struct rte_eth_dev *dev,
                                                 error);
                        if (ret < 0)
                                return ret;
-                       item_flags |= tunnel ? MLX5_FLOW_LAYER_INNER_L4_TCP :
-                                              MLX5_FLOW_LAYER_OUTER_L4_TCP;
+                       last_item = tunnel ? MLX5_FLOW_LAYER_INNER_L4_TCP :
+                                            MLX5_FLOW_LAYER_OUTER_L4_TCP;
                        break;
                case RTE_FLOW_ITEM_TYPE_VXLAN:
                        ret = mlx5_flow_validate_item_vxlan(items, item_flags,
                                                            error);
                        if (ret < 0)
                                return ret;
-                       item_flags |= MLX5_FLOW_LAYER_VXLAN;
+                       last_item = MLX5_FLOW_LAYER_VXLAN;
                        break;
                case RTE_FLOW_ITEM_TYPE_VXLAN_GPE:
                        ret = mlx5_flow_validate_item_vxlan_gpe(items,
@@ -1110,35 +1132,29 @@ flow_verbs_validate(struct rte_eth_dev *dev,
                                                                dev, error);
                        if (ret < 0)
                                return ret;
-                       item_flags |= MLX5_FLOW_LAYER_VXLAN_GPE;
+                       last_item = MLX5_FLOW_LAYER_VXLAN_GPE;
                        break;
                case RTE_FLOW_ITEM_TYPE_GRE:
                        ret = mlx5_flow_validate_item_gre(items, item_flags,
                                                          next_protocol, error);
                        if (ret < 0)
                                return ret;
-                       item_flags |= MLX5_FLOW_LAYER_GRE;
+                       last_item = MLX5_FLOW_LAYER_GRE;
                        break;
                case RTE_FLOW_ITEM_TYPE_MPLS:
-                       ret = mlx5_flow_validate_item_mpls(items, item_flags,
-                                                          next_protocol,
-                                                          error);
+                       ret = mlx5_flow_validate_item_mpls(dev, items,
+                                                          item_flags,
+                                                          last_item, error);
                        if (ret < 0)
                                return ret;
-                       if (next_protocol != 0xff &&
-                           next_protocol != IPPROTO_MPLS)
-                               return rte_flow_error_set
-                                       (error, EINVAL,
-                                        RTE_FLOW_ERROR_TYPE_ITEM, items,
-                                        "protocol filtering not compatible"
-                                        " with MPLS layer");
-                       item_flags |= MLX5_FLOW_LAYER_MPLS;
+                       last_item = MLX5_FLOW_LAYER_MPLS;
                        break;
                default:
                        return rte_flow_error_set(error, ENOTSUP,
                                                  RTE_FLOW_ERROR_TYPE_ITEM,
                                                  NULL, "item not supported");
                }
+               item_flags |= last_item;
        }
        for (; actions->type != RTE_FLOW_ACTION_TYPE_END; actions++) {
                switch (actions->type) {
@@ -1181,7 +1197,7 @@ flow_verbs_validate(struct rte_eth_dev *dev,
                case RTE_FLOW_ACTION_TYPE_RSS:
                        ret = mlx5_flow_validate_action_rss(actions,
                                                            action_flags, dev,
-                                                           attr,
+                                                           attr, item_flags,
                                                            error);
                        if (ret < 0)
                                return ret;
@@ -1209,23 +1225,18 @@ flow_verbs_validate(struct rte_eth_dev *dev,
 
 /**
  * Calculate the required bytes that are needed for the action part of the verbs
- * flow, in addtion returns bit-fields with all the detected action, in order to
- * avoid another interation over the actions.
+ * flow.
  *
  * @param[in] actions
  *   Pointer to the list of actions.
- * @param[out] action_flags
- *   Pointer to the detected actions.
  *
  * @return
  *   The size of the memory needed for all actions.
  */
 static int
-flow_verbs_get_actions_and_size(const struct rte_flow_action actions[],
-                               uint64_t *action_flags)
+flow_verbs_get_actions_size(const struct rte_flow_action actions[])
 {
        int size = 0;
-       uint64_t detected_actions = 0;
 
        for (; actions->type != RTE_FLOW_ACTION_TYPE_END; actions++) {
                switch (actions->type) {
@@ -1233,128 +1244,89 @@ flow_verbs_get_actions_and_size(const struct rte_flow_action actions[],
                        break;
                case RTE_FLOW_ACTION_TYPE_FLAG:
                        size += sizeof(struct ibv_flow_spec_action_tag);
-                       detected_actions |= MLX5_FLOW_ACTION_FLAG;
                        break;
                case RTE_FLOW_ACTION_TYPE_MARK:
                        size += sizeof(struct ibv_flow_spec_action_tag);
-                       detected_actions |= MLX5_FLOW_ACTION_MARK;
                        break;
                case RTE_FLOW_ACTION_TYPE_DROP:
                        size += sizeof(struct ibv_flow_spec_action_drop);
-                       detected_actions |= MLX5_FLOW_ACTION_DROP;
                        break;
                case RTE_FLOW_ACTION_TYPE_QUEUE:
-                       detected_actions |= MLX5_FLOW_ACTION_QUEUE;
                        break;
                case RTE_FLOW_ACTION_TYPE_RSS:
-                       detected_actions |= MLX5_FLOW_ACTION_RSS;
                        break;
                case RTE_FLOW_ACTION_TYPE_COUNT:
 #if defined(HAVE_IBV_DEVICE_COUNTERS_SET_V42) || \
        defined(HAVE_IBV_DEVICE_COUNTERS_SET_V45)
                        size += sizeof(struct ibv_flow_spec_counter_action);
 #endif
-                       detected_actions |= MLX5_FLOW_ACTION_COUNT;
                        break;
                default:
                        break;
                }
        }
-       *action_flags = detected_actions;
        return size;
 }
 
 /**
  * Calculate the required bytes that are needed for the item part of the verbs
- * flow, in addtion returns bit-fields with all the detected action, in order to
- * avoid another interation over the actions.
+ * flow.
  *
- * @param[in] actions
+ * @param[in] items
  *   Pointer to the list of items.
- * @param[in, out] item_flags
- *   Pointer to the detected items.
  *
  * @return
  *   The size of the memory needed for all items.
  */
 static int
-flow_verbs_get_items_and_size(const struct rte_flow_item items[],
-                             uint64_t *item_flags)
+flow_verbs_get_items_size(const struct rte_flow_item items[])
 {
        int size = 0;
-       uint64_t detected_items = 0;
 
        for (; items->type != RTE_FLOW_ITEM_TYPE_END; items++) {
-               int tunnel = !!(detected_items & MLX5_FLOW_LAYER_TUNNEL);
-
                switch (items->type) {
                case RTE_FLOW_ITEM_TYPE_VOID:
                        break;
                case RTE_FLOW_ITEM_TYPE_ETH:
                        size += sizeof(struct ibv_flow_spec_eth);
-                       detected_items |= tunnel ? MLX5_FLOW_LAYER_INNER_L2 :
-                                                  MLX5_FLOW_LAYER_OUTER_L2;
                        break;
                case RTE_FLOW_ITEM_TYPE_VLAN:
                        size += sizeof(struct ibv_flow_spec_eth);
-                       detected_items |=
-                               tunnel ? (MLX5_FLOW_LAYER_INNER_L2 |
-                                         MLX5_FLOW_LAYER_INNER_VLAN) :
-                                        (MLX5_FLOW_LAYER_OUTER_L2 |
-                                         MLX5_FLOW_LAYER_OUTER_VLAN);
                        break;
                case RTE_FLOW_ITEM_TYPE_IPV4:
                        size += sizeof(struct ibv_flow_spec_ipv4_ext);
-                       detected_items |= tunnel ?
-                                         MLX5_FLOW_LAYER_INNER_L3_IPV4 :
-                                         MLX5_FLOW_LAYER_OUTER_L3_IPV4;
                        break;
                case RTE_FLOW_ITEM_TYPE_IPV6:
                        size += sizeof(struct ibv_flow_spec_ipv6);
-                       detected_items |= tunnel ?
-                                         MLX5_FLOW_LAYER_INNER_L3_IPV6 :
-                                         MLX5_FLOW_LAYER_OUTER_L3_IPV6;
                        break;
                case RTE_FLOW_ITEM_TYPE_UDP:
                        size += sizeof(struct ibv_flow_spec_tcp_udp);
-                       detected_items |= tunnel ?
-                                         MLX5_FLOW_LAYER_INNER_L4_UDP :
-                                         MLX5_FLOW_LAYER_OUTER_L4_UDP;
                        break;
                case RTE_FLOW_ITEM_TYPE_TCP:
                        size += sizeof(struct ibv_flow_spec_tcp_udp);
-                       detected_items |= tunnel ?
-                                         MLX5_FLOW_LAYER_INNER_L4_TCP :
-                                         MLX5_FLOW_LAYER_OUTER_L4_TCP;
                        break;
                case RTE_FLOW_ITEM_TYPE_VXLAN:
                        size += sizeof(struct ibv_flow_spec_tunnel);
-                       detected_items |= MLX5_FLOW_LAYER_VXLAN;
                        break;
                case RTE_FLOW_ITEM_TYPE_VXLAN_GPE:
                        size += sizeof(struct ibv_flow_spec_tunnel);
-                       detected_items |= MLX5_FLOW_LAYER_VXLAN_GPE;
                        break;
 #ifdef HAVE_IBV_DEVICE_MPLS_SUPPORT
                case RTE_FLOW_ITEM_TYPE_GRE:
                        size += sizeof(struct ibv_flow_spec_gre);
-                       detected_items |= MLX5_FLOW_LAYER_GRE;
                        break;
                case RTE_FLOW_ITEM_TYPE_MPLS:
                        size += sizeof(struct ibv_flow_spec_mpls);
-                       detected_items |= MLX5_FLOW_LAYER_MPLS;
                        break;
 #else
                case RTE_FLOW_ITEM_TYPE_GRE:
                        size += sizeof(struct ibv_flow_spec_tunnel);
-                       detected_items |= MLX5_FLOW_LAYER_TUNNEL;
                        break;
 #endif
                default:
                        break;
                }
        }
-       *item_flags = detected_items;
        return size;
 }
 
@@ -1369,10 +1341,6 @@ flow_verbs_get_items_and_size(const struct rte_flow_item items[],
  *   Pointer to the list of items.
  * @param[in] actions
  *   Pointer to the list of actions.
- * @param[out] item_flags
- *   Pointer to bit mask of all items detected.
- * @param[out] action_flags
- *   Pointer to bit mask of all actions detected.
  * @param[out] error
  *   Pointer to the error structure.
  *
@@ -1384,15 +1352,13 @@ static struct mlx5_flow *
 flow_verbs_prepare(const struct rte_flow_attr *attr __rte_unused,
                   const struct rte_flow_item items[],
                   const struct rte_flow_action actions[],
-                  uint64_t *item_flags,
-                  uint64_t *action_flags,
                   struct rte_flow_error *error)
 {
        uint32_t size = sizeof(struct mlx5_flow) + sizeof(struct ibv_flow_attr);
        struct mlx5_flow *flow;
 
-       size += flow_verbs_get_actions_and_size(actions, action_flags);
-       size += flow_verbs_get_items_and_size(items, item_flags);
+       size += flow_verbs_get_actions_size(actions);
+       size += flow_verbs_get_items_size(items);
        flow = rte_calloc(__func__, 1, size, 0);
        if (!flow) {
                rte_flow_error_set(error, ENOMEM,
@@ -1423,7 +1389,7 @@ flow_verbs_prepare(const struct rte_flow_attr *attr __rte_unused,
  *   Pointer to the error structure.
  *
  * @return
- *   0 on success, else a negative errno value otherwise and rte_ernno is set.
+ *   0 on success, else a negative errno value otherwise and rte_errno is set.
  */
 static int
 flow_verbs_translate(struct rte_eth_dev *dev,
@@ -1438,7 +1404,7 @@ flow_verbs_translate(struct rte_eth_dev *dev,
        uint64_t action_flags = 0;
        uint64_t priority = attr->priority;
        uint32_t subpriority = 0;
-       struct priv *priv = dev->data->dev_private;
+       struct mlx5_priv *priv = dev->data->dev_private;
 
        if (priority == MLX5_FLOW_PRIO_RSVD)
                priority = priv->config.flow_prio - 1;
@@ -1588,6 +1554,7 @@ flow_verbs_translate(struct rte_eth_dev *dev,
        dev_flow->layers = item_flags;
        dev_flow->verbs.attr->priority =
                mlx5_flow_adjust_priority(dev, priority, subpriority);
+       dev_flow->verbs.attr->port = (uint8_t)priv->ibv_port;
        return 0;
 }