net/mlx5: support list of representor PF
[dpdk.git] / drivers / net / mlx5 / mlx5_flow.c
index 5d180fc..c347f81 100644 (file)
@@ -499,7 +499,8 @@ static const struct mlx5_flow_expand_node mlx5_support_expansion[] = {
                        (MLX5_EXPANSION_OUTER_IPV6_UDP,
                         MLX5_EXPANSION_OUTER_IPV6_TCP,
                         MLX5_EXPANSION_IPV4,
-                        MLX5_EXPANSION_IPV6),
+                        MLX5_EXPANSION_IPV6,
+                        MLX5_EXPANSION_GRE),
                .type = RTE_FLOW_ITEM_TYPE_IPV6,
                .rss_types = ETH_RSS_IPV6 | ETH_RSS_FRAG_IPV6 |
                        ETH_RSS_NONFRAG_IPV6_OTHER,
@@ -527,7 +528,8 @@ static const struct mlx5_flow_expand_node mlx5_support_expansion[] = {
                .type = RTE_FLOW_ITEM_TYPE_VXLAN_GPE,
        },
        [MLX5_EXPANSION_GRE] = {
-               .next = MLX5_FLOW_EXPAND_RSS_NEXT(MLX5_EXPANSION_IPV4),
+               .next = MLX5_FLOW_EXPAND_RSS_NEXT(MLX5_EXPANSION_IPV4,
+                                                 MLX5_EXPANSION_IPV6),
                .type = RTE_FLOW_ITEM_TYPE_GRE,
        },
        [MLX5_EXPANSION_MPLS] = {
@@ -855,6 +857,58 @@ mlx5_flow_ext_mreg_supported(struct rte_eth_dev *dev)
        return config->flow_mreg_c[2] != REG_NON;
 }
 
+/**
+ * Get the lowest priority.
+ *
+ * @param[in] dev
+ *   Pointer to the Ethernet device structure.
+ * @param[in] attributes
+ *   Pointer to device flow rule attributes.
+ *
+ * @return
+ *   The value of lowest priority of flow.
+ */
+uint32_t
+mlx5_get_lowest_priority(struct rte_eth_dev *dev,
+                         const struct rte_flow_attr *attr)
+{
+       struct mlx5_priv *priv = dev->data->dev_private;
+
+       if (!attr->group && !attr->transfer)
+               return priv->config.flow_prio - 2;
+       return MLX5_NON_ROOT_FLOW_MAX_PRIO - 1;
+}
+
+/**
+ * Calculate matcher priority of the flow.
+ *
+ * @param[in] dev
+ *   Pointer to the Ethernet device structure.
+ * @param[in] attr
+ *   Pointer to device flow rule attributes.
+ * @param[in] subpriority
+ *   The priority based on the items.
+ * @return
+ *   The matcher priority of the flow.
+ */
+uint16_t
+mlx5_get_matcher_priority(struct rte_eth_dev *dev,
+                         const struct rte_flow_attr *attr,
+                         uint32_t subpriority)
+{
+       uint16_t priority = (uint16_t)attr->priority;
+       struct mlx5_priv *priv = dev->data->dev_private;
+
+       if (!attr->group && !attr->transfer) {
+               if (attr->priority == MLX5_FLOW_LOWEST_PRIO_INDICATOR)
+                       priority = priv->config.flow_prio - 1;
+               return mlx5_os_flow_adjust_priority(dev, priority, subpriority);
+       }
+       if (attr->priority == MLX5_FLOW_LOWEST_PRIO_INDICATOR)
+               priority = MLX5_NON_ROOT_FLOW_MAX_PRIO;
+       return priority * 3 + subpriority;
+}
+
 /**
  * Verify the @p item specifications (spec, last, mask) are compatible with the
  * NIC capabilities.
@@ -1215,10 +1269,14 @@ mlx5_flow_rxq_dynf_metadata_set(struct rte_eth_dev *dev)
                        data->dynf_meta = 0;
                        data->flow_meta_mask = 0;
                        data->flow_meta_offset = -1;
+                       data->flow_meta_port_mask = 0;
                } else {
                        data->dynf_meta = 1;
                        data->flow_meta_mask = rte_flow_dynf_metadata_mask;
                        data->flow_meta_offset = rte_flow_dynf_metadata_offs;
+                       data->flow_meta_port_mask = (uint32_t)~0;
+                       if (priv->config.dv_xmeta_en == MLX5_XMETA_MODE_META16)
+                               data->flow_meta_port_mask >>= 16;
                }
        }
 }
@@ -1669,7 +1727,7 @@ mlx5_flow_validate_attributes(struct rte_eth_dev *dev,
                return rte_flow_error_set(error, ENOTSUP,
                                          RTE_FLOW_ERROR_TYPE_ATTR_GROUP,
                                          NULL, "groups is not supported");
-       if (attributes->priority != MLX5_FLOW_PRIO_RSVD &&
+       if (attributes->priority != MLX5_FLOW_LOWEST_PRIO_INDICATOR &&
            attributes->priority >= priority_max)
                return rte_flow_error_set(error, ENOTSUP,
                                          RTE_FLOW_ERROR_TYPE_ATTR_PRIORITY,
@@ -3837,7 +3895,7 @@ flow_dv_mreg_create_cb(struct mlx5_hlist *list, uint64_t key,
                };
        } else {
                /* Default rule, wildcard match. */
-               attr.priority = MLX5_FLOW_PRIO_RSVD;
+               attr.priority = MLX5_FLOW_LOWEST_PRIO_INDICATOR;
                items[0] = (struct rte_flow_item){
                        .type = RTE_FLOW_ITEM_TYPE_END,
                };
@@ -5685,7 +5743,7 @@ flow_list_create(struct rte_eth_dev *dev, uint32_t *list,
         */
        if (external || dev->data->dev_started ||
            (attr->group == MLX5_FLOW_MREG_CP_TABLE_GROUP &&
-            attr->priority == MLX5_FLOW_PRIO_RSVD)) {
+            attr->priority == MLX5_FLOW_LOWEST_PRIO_INDICATOR)) {
                ret = flow_drv_apply(dev, flow, error);
                if (ret < 0)
                        goto error;
@@ -6185,7 +6243,7 @@ mlx5_ctrl_flow_vlan(struct rte_eth_dev *dev,
        struct mlx5_priv *priv = dev->data->dev_private;
        const struct rte_flow_attr attr = {
                .ingress = 1,
-               .priority = MLX5_FLOW_PRIO_RSVD,
+               .priority = MLX5_FLOW_LOWEST_PRIO_INDICATOR,
        };
        struct rte_flow_item items[] = {
                {
@@ -6440,40 +6498,20 @@ mlx5_flow_query(struct rte_eth_dev *dev,
 }
 
 /**
- * Manage filter operations.
+ * Get rte_flow callbacks.
  *
  * @param dev
  *   Pointer to Ethernet device structure.
- * @param filter_type
- *   Filter type.
- * @param filter_op
- *   Operation to perform.
- * @param arg
+ * @param ops
  *   Pointer to operation-specific structure.
  *
- * @return
- *   0 on success, a negative errno value otherwise and rte_errno is set.
+ * @return 0
  */
 int
-mlx5_dev_filter_ctrl(struct rte_eth_dev *dev,
-                    enum rte_filter_type filter_type,
-                    enum rte_filter_op filter_op,
-                    void *arg)
-{
-       switch (filter_type) {
-       case RTE_ETH_FILTER_GENERIC:
-               if (filter_op != RTE_ETH_FILTER_GET) {
-                       rte_errno = EINVAL;
-                       return -rte_errno;
-               }
-               *(const void **)arg = &mlx5_flow_ops;
-               return 0;
-       default:
-               DRV_LOG(ERR, "port %u filter type (%d) not supported",
-                       dev->data->port_id, filter_type);
-               rte_errno = ENOTSUP;
-               return -rte_errno;
-       }
+mlx5_flow_ops_get(struct rte_eth_dev *dev __rte_unused,
+                 const struct rte_flow_ops **ops)
+{
+       *ops = &mlx5_flow_ops;
        return 0;
 }
 
@@ -7070,7 +7108,7 @@ mlx5_flow_discover_mreg_c(struct rte_eth_dev *dev)
        for (idx = REG_C_2; idx <= REG_C_7; ++idx) {
                struct rte_flow_attr attr = {
                        .group = MLX5_FLOW_MREG_CP_TABLE_GROUP,
-                       .priority = MLX5_FLOW_PRIO_RSVD,
+                       .priority = MLX5_FLOW_LOWEST_PRIO_INDICATOR,
                        .ingress = 1,
                };
                struct rte_flow_item items[] = {
@@ -7400,12 +7438,12 @@ mlx5_shared_action_flush(struct rte_eth_dev *dev)
 {
        struct rte_flow_error error;
        struct mlx5_priv *priv = dev->data->dev_private;
-       struct mlx5_shared_action_rss *action;
+       struct mlx5_shared_action_rss *shared_rss;
        int ret = 0;
        uint32_t idx;
 
        ILIST_FOREACH(priv->sh->ipool[MLX5_IPOOL_RSS_SHARED_ACTIONS],
-                     priv->rss_shared_actions, idx, action, next) {
+                     priv->rss_shared_actions, idx, shared_rss, next) {
                ret |= mlx5_shared_action_destroy(dev,
                       (struct rte_flow_shared_action *)(uintptr_t)idx, &error);
        }
@@ -7843,10 +7881,11 @@ static void get_tunnel_miss(struct rte_eth_dev *dev, void *x)
 
        rte_spinlock_unlock(&thub->sl);
        ctx->tunnel = mlx5_flow_tunnel_allocate(dev, ctx->app_tunnel);
-       ctx->tunnel->refctn = 1;
        rte_spinlock_lock(&thub->sl);
-       if (ctx->tunnel)
+       if (ctx->tunnel) {
+               ctx->tunnel->refctn = 1;
                LIST_INSERT_HEAD(&thub->tunnels, ctx->tunnel, chain);
+       }
 }
 
 
@@ -7872,7 +7911,7 @@ void mlx5_release_tunnel_hub(struct mlx5_dev_ctx_shared *sh, uint16_t port_id)
        if (!thub)
                return;
        if (!LIST_EMPTY(&thub->tunnels))
-               DRV_LOG(WARNING, "port %u tunnels present\n", port_id);
+               DRV_LOG(WARNING, "port %u tunnels present", port_id);
        mlx5_hlist_destroy(thub->groups);
        mlx5_free(thub);
 }
@@ -7888,7 +7927,8 @@ int mlx5_alloc_tunnel_hub(struct mlx5_dev_ctx_shared *sh)
                return -ENOMEM;
        LIST_INIT(&thub->tunnels);
        rte_spinlock_init(&thub->sl);
-       thub->groups = mlx5_hlist_create("flow groups", MLX5_MAX_TABLES, 0,
+       thub->groups = mlx5_hlist_create("flow groups",
+                                        rte_align32pow2(MLX5_MAX_TABLES), 0,
                                         0, mlx5_flow_tunnel_grp2tbl_create_cb,
                                         mlx5_flow_tunnel_grp2tbl_match_cb,
                                         mlx5_flow_tunnel_grp2tbl_remove_cb);