net/mlx5: fix shared RSS and mark actions combination
[dpdk.git] / drivers / net / mlx5 / mlx5_flow_dv.c
index b70d9a8..50673ce 100644 (file)
@@ -35,7 +35,7 @@
 #include "mlx5_rxtx.h"
 #include "rte_pmd_mlx5.h"
 
-#ifdef HAVE_IBV_FLOW_DV_SUPPORT
+#if defined(HAVE_IBV_FLOW_DV_SUPPORT) || !defined(HAVE_INFINIBAND_VERBS_H)
 
 #ifndef HAVE_IBV_FLOW_DEVX_COUNTERS
 #define MLX5DV_FLOW_ACTION_COUNTERS_DEVX 0
@@ -955,7 +955,7 @@ flow_dv_convert_action_set_reg
                                          RTE_FLOW_ERROR_TYPE_ACTION, NULL,
                                          "too many items to modify");
        MLX5_ASSERT(conf->id != REG_NON);
-       MLX5_ASSERT(conf->id < RTE_DIM(reg_to_field));
+       MLX5_ASSERT(conf->id < (enum modify_reg)RTE_DIM(reg_to_field));
        actions[i] = (struct mlx5_modification_cmd) {
                .action_type = MLX5_MODIFICATION_TYPE_SET,
                .field = reg_to_field[conf->id],
@@ -1217,6 +1217,7 @@ flow_dv_convert_action_set_meta
 
        if (reg < 0)
                return reg;
+       MLX5_ASSERT(reg != REG_NON);
        /*
         * In datapath code there is no endianness
         * coversions for perfromance reasons, all
@@ -1438,6 +1439,10 @@ flow_dv_validate_item_meta(struct rte_eth_dev *dev __rte_unused,
                reg = flow_dv_get_metadata_reg(dev, attr, error);
                if (reg < 0)
                        return reg;
+               if (reg == REG_NON)
+                       return rte_flow_error_set(error, ENOTSUP,
+                                       RTE_FLOW_ERROR_TYPE_ITEM, item,
+                                       "unavalable extended metadata register");
                if (reg == REG_B)
                        return rte_flow_error_set(error, ENOTSUP,
                                          RTE_FLOW_ERROR_TYPE_ITEM, item,
@@ -2446,6 +2451,10 @@ flow_dv_validate_action_set_meta(struct rte_eth_dev *dev,
        reg = flow_dv_get_metadata_reg(dev, attr, error);
        if (reg < 0)
                return reg;
+       if (reg == REG_NON)
+               return rte_flow_error_set(error, ENOTSUP,
+                                         RTE_FLOW_ERROR_TYPE_ACTION, action,
+                                         "unavalable extended metadata register");
        if (reg != REG_A && reg != REG_B) {
                struct mlx5_priv *priv = dev->data->dev_private;
 
@@ -2771,8 +2780,7 @@ flow_dv_encap_decap_match_cb(struct mlx5_hlist *list __rte_unused,
        cache_resource = container_of(entry,
                                      struct mlx5_flow_dv_encap_decap_resource,
                                      entry);
-       if (resource->entry.key == cache_resource->entry.key &&
-           resource->reformat_type == cache_resource->reformat_type &&
+       if (resource->reformat_type == cache_resource->reformat_type &&
            resource->ft_type == cache_resource->ft_type &&
            resource->flags == cache_resource->flags &&
            resource->size == cache_resource->size &&
@@ -2890,18 +2898,16 @@ flow_dv_encap_decap_resource_register
                .error = error,
                .data = resource,
        };
+       uint64_t key64;
 
        resource->flags = dev_flow->dv.group ? 0 : 1;
-       resource->entry.key =  __rte_raw_cksum(&encap_decap_key.v32,
-                                              sizeof(encap_decap_key.v32), 0);
+       key64 =  __rte_raw_cksum(&encap_decap_key.v32,
+                                sizeof(encap_decap_key.v32), 0);
        if (resource->reformat_type !=
            MLX5DV_FLOW_ACTION_PACKET_REFORMAT_TYPE_L2_TUNNEL_TO_L2 &&
            resource->size)
-               resource->entry.key = __rte_raw_cksum(resource->buf,
-                                                     resource->size,
-                                                     resource->entry.key);
-       entry = mlx5_hlist_register(sh->encaps_decaps, resource->entry.key,
-                                   &ctx);
+               key64 = __rte_raw_cksum(resource->buf, resource->size, key64);
+       entry = mlx5_hlist_register(sh->encaps_decaps, key64, &ctx);
        if (!entry)
                return -rte_errno;
        resource = container_of(entry, typeof(*resource), entry);
@@ -4525,6 +4531,7 @@ flow_dv_modify_hdr_resource_register
                .error = error,
                .data = resource,
        };
+       uint64_t key64;
 
        resource->flags = dev_flow->dv.group ? 0 :
                          MLX5DV_DR_ACTION_FLAGS_ROOT_LEVEL;
@@ -4533,8 +4540,8 @@ flow_dv_modify_hdr_resource_register
                return rte_flow_error_set(error, EOVERFLOW,
                                          RTE_FLOW_ERROR_TYPE_ACTION, NULL,
                                          "too many modify header items");
-       resource->entry.key = __rte_raw_cksum(&resource->ft_type, key_len, 0);
-       entry = mlx5_hlist_register(sh->modify_cmds, resource->entry.key, &ctx);
+       key64 = __rte_raw_cksum(&resource->ft_type, key_len, 0);
+       entry = mlx5_hlist_register(sh->modify_cmds, key64, &ctx);
        if (!entry)
                return -rte_errno;
        resource = container_of(entry, typeof(*resource), entry);
@@ -5273,6 +5280,11 @@ flow_dv_validate(struct rte_eth_dev *dev, const struct rte_flow_attr *attr,
        } else {
                tunnel = NULL;
        }
+       if (tunnel && priv->representor)
+               return rte_flow_error_set(error, ENOTSUP,
+                                         RTE_FLOW_ERROR_TYPE_UNSPECIFIED, NULL,
+                                         "decap not supported "
+                                         "for VF representor");
        grp_info.std_tbl_fix = tunnel_use_standard_attr_group_translate
                                (dev, tunnel, attr, items, actions);
        ret = flow_dv_validate_attributes(dev, tunnel, attr, &grp_info, error);
@@ -5946,7 +5958,7 @@ flow_dv_validate(struct rte_eth_dev *dev, const struct rte_flow_attr *attr,
                        rw_act_num += MLX5_ACT_NUM_SET_TAG;
                        break;
                case MLX5_RTE_FLOW_ACTION_TYPE_AGE:
-                       if (!attr->group)
+                       if (!attr->transfer && !attr->group)
                                return rte_flow_error_set(error, ENOTSUP,
                                                RTE_FLOW_ERROR_TYPE_UNSPECIFIED,
                                                                           NULL,
@@ -6223,8 +6235,9 @@ flow_dv_prepare(struct rte_eth_dev *dev,
                                   "not enough memory to create flow handle");
                return NULL;
        }
-       MLX5_ASSERT(wks->flow_idx + 1 < RTE_DIM(wks->flows));
+       MLX5_ASSERT(wks->flow_idx < RTE_DIM(wks->flows));
        dev_flow = &wks->flows[wks->flow_idx++];
+       memset(dev_flow, 0, sizeof(*dev_flow));
        dev_flow->handle = dev_handle;
        dev_flow->handle_idx = handle_idx;
        /*
@@ -6236,12 +6249,6 @@ flow_dv_prepare(struct rte_eth_dev *dev,
         */
        dev_flow->dv.value.size = MLX5_ST_SZ_BYTES(fte_match_param) -
                                  MLX5_ST_SZ_BYTES(fte_match_set_misc4);
-       /*
-        * The matching value needs to be cleared to 0 before using. In the
-        * past, it will be automatically cleared when using rte_*alloc
-        * API. The time consumption will be almost the same as before.
-        */
-       memset(dev_flow->dv.value.buf, 0, MLX5_ST_SZ_BYTES(fte_match_param));
        dev_flow->ingress = attr->ingress;
        dev_flow->dv.transfer = attr->transfer;
        return dev_flow;
@@ -7471,6 +7478,7 @@ flow_dv_translate_item_meta(struct rte_eth_dev *dev,
                reg = flow_dv_get_metadata_reg(dev, attr, NULL);
                if (reg < 0)
                        return;
+               MLX5_ASSERT(reg != REG_NON);
                /*
                 * In datapath code there is no endianness
                 * coversions for perfromance reasons, all
@@ -7979,9 +7987,12 @@ flow_dv_tbl_create_cb(struct mlx5_hlist *list, uint64_t key64, void *cb_ctx)
        tbl_data->idx = idx;
        tbl_data->tunnel = tt_prm->tunnel;
        tbl_data->group_id = tt_prm->group_id;
-       tbl_data->external = tt_prm->external;
+       tbl_data->external = !!tt_prm->external;
        tbl_data->tunnel_offload = is_tunnel_offload_active(dev);
        tbl_data->is_egress = !!key.direction;
+       tbl_data->is_transfer = !!key.domain;
+       tbl_data->dummy = !!key.dummy;
+       tbl_data->table_id = key.table_id;
        tbl = &tbl_data->tbl;
        if (key.dummy)
                return &tbl_data->entry;
@@ -8022,6 +8033,21 @@ flow_dv_tbl_create_cb(struct mlx5_hlist *list, uint64_t key64, void *cb_ctx)
        return &tbl_data->entry;
 }
 
+int
+flow_dv_tbl_match_cb(struct mlx5_hlist *list __rte_unused,
+                    struct mlx5_hlist_entry *entry, uint64_t key64,
+                    void *cb_ctx __rte_unused)
+{
+       struct mlx5_flow_tbl_data_entry *tbl_data =
+               container_of(entry, struct mlx5_flow_tbl_data_entry, entry);
+       union mlx5_flow_tbl_key key = { .v64 = key64 };
+
+       return tbl_data->table_id != key.table_id ||
+              tbl_data->dummy != key.dummy ||
+              tbl_data->is_transfer != key.domain ||
+              tbl_data->is_egress != key.direction;
+}
+
 /**
  * Get a flow table.
  *
@@ -8107,10 +8133,7 @@ flow_dv_tbl_remove_cb(struct mlx5_hlist *list,
                                        tbl_data->tunnel->tunnel_id : 0,
                        .group = tbl_data->group_id
                };
-               union mlx5_flow_tbl_key table_key = {
-                       .v64 = entry->key
-               };
-               uint32_t table_id = table_key.table_id;
+               uint32_t table_id = tbl_data->table_id;
 
                tunnel_grp_hash = tbl_data->tunnel ?
                                        tbl_data->tunnel->groups :
@@ -8285,6 +8308,7 @@ flow_dv_tag_create_cb(struct mlx5_hlist *list, uint64_t key, void *ctx)
                return NULL;
        }
        entry->idx = idx;
+       entry->tag_id = key;
        ret = mlx5_flow_os_create_flow_action_tag(key,
                                                  &entry->action);
        if (ret) {
@@ -8297,6 +8321,17 @@ flow_dv_tag_create_cb(struct mlx5_hlist *list, uint64_t key, void *ctx)
        return &entry->entry;
 }
 
+int
+flow_dv_tag_match_cb(struct mlx5_hlist *list __rte_unused,
+                    struct mlx5_hlist_entry *entry, uint64_t key,
+                    void *cb_ctx __rte_unused)
+{
+       struct mlx5_flow_dv_tag_resource *tag =
+               container_of(entry, struct mlx5_flow_dv_tag_resource, entry);
+
+       return key != tag->tag_id;
+}
+
 /**
  * Find existing tag resource or create and register a new one.
  *
@@ -8727,11 +8762,13 @@ flow_dv_sample_create_cb(struct mlx5_cache_list *list __rte_unused,
                                          "for sample");
                goto error;
        }
+       int ret;
+
        cache_resource->normal_path_tbl = tbl;
        if (resource->ft_type == MLX5DV_FLOW_TABLE_TYPE_FDB) {
-               cache_resource->default_miss =
-                               mlx5_glue->dr_create_flow_action_default_miss();
-               if (!cache_resource->default_miss) {
+               ret = mlx5_flow_os_create_flow_action_default_miss
+                       (&cache_resource->default_miss);
+               if (!ret) {
                        rte_flow_error_set(error, ENOMEM,
                                                RTE_FLOW_ERROR_TYPE_UNSPECIFIED,
                                                NULL,
@@ -9469,12 +9506,14 @@ flow_dv_age_pool_create(struct rte_eth_dev *dev,
  *
  * @param[in] dev
  *   Pointer to the Ethernet device structure.
+ * @param[out] error
+ *   Pointer to the error structure.
  *
  * @return
  *   Index to ASO age action on success, 0 otherwise and rte_errno is set.
  */
 static uint32_t
-flow_dv_aso_age_alloc(struct rte_eth_dev *dev)
+flow_dv_aso_age_alloc(struct rte_eth_dev *dev, struct rte_flow_error *error)
 {
        struct mlx5_priv *priv = dev->data->dev_private;
        const struct mlx5_aso_age_pool *pool;
@@ -9489,7 +9528,9 @@ flow_dv_aso_age_alloc(struct rte_eth_dev *dev)
                LIST_REMOVE(age_free, next);
        } else if (!flow_dv_age_pool_create(dev, &age_free)) {
                rte_spinlock_unlock(&mng->free_sl);
-               return 0; /* 0 is an error.*/
+               rte_flow_error_set(error, rte_errno, RTE_FLOW_ERROR_TYPE_ACTION,
+                                  NULL, "failed to create ASO age pool");
+               return 0; /* 0 is an error. */
        }
        rte_spinlock_unlock(&mng->free_sl);
        pool = container_of
@@ -9497,15 +9538,33 @@ flow_dv_aso_age_alloc(struct rte_eth_dev *dev)
                  (age_free - age_free->offset), const struct mlx5_aso_age_pool,
                                                                       actions);
        if (!age_free->dr_action) {
-               age_free->dr_action = mlx5_glue->dr_action_create_flow_hit
-                                               (pool->flow_hit_aso_obj->obj,
-                                                age_free->offset, REG_C_5);
+               int reg_c = mlx5_flow_get_reg_id(dev, MLX5_ASO_FLOW_HIT, 0,
+                                                error);
+
+               if (reg_c < 0) {
+                       rte_flow_error_set(error, rte_errno,
+                                          RTE_FLOW_ERROR_TYPE_ACTION,
+                                          NULL, "failed to get reg_c "
+                                          "for ASO flow hit");
+                       return 0; /* 0 is an error. */
+               }
+#ifdef HAVE_MLX5_DR_CREATE_ACTION_ASO
+               age_free->dr_action = mlx5_glue->dv_create_flow_action_aso
+                               (priv->sh->rx_domain,
+                                pool->flow_hit_aso_obj->obj, age_free->offset,
+                                MLX5DV_DR_ACTION_FLAGS_ASO_FIRST_HIT_SET,
+                                (reg_c - REG_C_0));
+#endif /* HAVE_MLX5_DR_CREATE_ACTION_ASO */
                if (!age_free->dr_action) {
                        rte_errno = errno;
                        rte_spinlock_lock(&mng->free_sl);
                        LIST_INSERT_HEAD(&mng->free, age_free, next);
                        rte_spinlock_unlock(&mng->free_sl);
-                       return 0; /* 0 is an error.*/
+                       rte_flow_error_set(error, rte_errno,
+                                          RTE_FLOW_ERROR_TYPE_ACTION,
+                                          NULL, "failed to create ASO "
+                                          "flow hit action");
+                       return 0; /* 0 is an error. */
                }
        }
        __atomic_store_n(&age_free->refcnt, 1, __ATOMIC_RELAXED);
@@ -9519,18 +9578,21 @@ flow_dv_aso_age_alloc(struct rte_eth_dev *dev)
  *   Pointer to rte_eth_dev structure.
  * @param[in] age
  *   Pointer to the aging action configuration.
+ * @param[out] error
+ *   Pointer to the error structure.
  *
  * @return
  *   Index to flow counter on success, 0 otherwise.
  */
 static uint32_t
 flow_dv_translate_create_aso_age(struct rte_eth_dev *dev,
-                                const struct rte_flow_action_age *age)
+                                const struct rte_flow_action_age *age,
+                                struct rte_flow_error *error)
 {
        uint32_t age_idx = 0;
        struct mlx5_aso_age_action *aso_age;
 
-       age_idx = flow_dv_aso_age_alloc(dev);
+       age_idx = flow_dv_aso_age_alloc(dev, error);
        if (!age_idx)
                return 0;
        aso_age = flow_aso_age_get_by_idx(dev, age_idx);
@@ -9842,7 +9904,7 @@ flow_dv_translate(struct rte_eth_dev *dev,
                case RTE_FLOW_ACTION_TYPE_AGE:
                        if (priv->sh->flow_hit_aso_en && attr->group) {
                                flow->age = flow_dv_translate_create_aso_age
-                                               (dev, action->conf);
+                                               (dev, action->conf, error);
                                if (!flow->age)
                                        return rte_flow_error_set
                                                (error, rte_errno,
@@ -10507,8 +10569,9 @@ flow_dv_translate(struct rte_eth_dev *dev,
        /* Register matcher. */
        matcher.crc = rte_raw_cksum((const void *)matcher.mask.buf,
                                    matcher.mask.size);
-       matcher.priority = mlx5_flow_adjust_priority(dev, priority,
-                                                    matcher.priority);
+       matcher.priority = mlx5_os_flow_adjust_priority(dev,
+                                                       priority,
+                                                       matcher.priority);
        /* reserved field no needs to be set to 0 here. */
        tbl_key.domain = attr->transfer;
        tbl_key.direction = attr->egress;
@@ -10617,47 +10680,6 @@ __flow_dv_action_rss_hrxq_lookup(struct rte_eth_dev *dev, uint32_t idx,
        }
 }
 
-/**
- * Retrieves hash RX queue suitable for the *flow*.
- * If shared action configured for *flow* suitable hash RX queue will be
- * retrieved from attached shared action.
- *
- * @param[in] dev
- *   Pointer to the Ethernet device structure.
- * @param[in] dev_flow
- *   Pointer to the sub flow.
- * @param[in] rss_desc
- *   Pointer to the RSS descriptor.
- * @param[out] hrxq
- *   Pointer to retrieved hash RX queue object.
- *
- * @return
- *   Valid hash RX queue index, otherwise 0 and rte_errno is set.
- */
-static uint32_t
-__flow_dv_rss_get_hrxq(struct rte_eth_dev *dev, struct mlx5_flow *dev_flow,
-                      struct mlx5_flow_rss_desc *rss_desc,
-                      struct mlx5_hrxq **hrxq)
-{
-       struct mlx5_priv *priv = dev->data->dev_private;
-       uint32_t hrxq_idx;
-
-       if (rss_desc->shared_rss) {
-               hrxq_idx = __flow_dv_action_rss_hrxq_lookup
-                               (dev, rss_desc->shared_rss,
-                                dev_flow->hash_fields,
-                                !!(dev_flow->handle->layers &
-                                   MLX5_FLOW_LAYER_TUNNEL));
-               if (hrxq_idx)
-                       *hrxq = mlx5_ipool_get(priv->sh->ipool[MLX5_IPOOL_HRXQ],
-                                              hrxq_idx);
-       } else {
-               *hrxq = flow_dv_hrxq_prepare(dev, dev_flow, rss_desc,
-                                            &hrxq_idx);
-       }
-       return hrxq_idx;
-}
-
 /**
  * Apply the flow to the NIC, lock free,
  * (mutex should be acquired by caller).
@@ -10689,11 +10711,6 @@ flow_dv_apply(struct rte_eth_dev *dev, struct rte_flow *flow,
        struct mlx5_flow_rss_desc *rss_desc = &wks->rss_desc;
 
        MLX5_ASSERT(wks);
-       if (rss_desc->shared_rss) {
-               dh = wks->flows[wks->flow_idx - 1].handle;
-               MLX5_ASSERT(dh->fate_action == MLX5_FLOW_FATE_SHARED_RSS);
-               dh->rix_srss = rss_desc->shared_rss;
-       }
        for (idx = wks->flow_idx - 1; idx >= 0; idx--) {
                dev_flow = &wks->flows[idx];
                dv = &dev_flow->dv;
@@ -10709,11 +10726,34 @@ flow_dv_apply(struct rte_eth_dev *dev, struct rte_flow *flow,
                                                priv->drop_queue.hrxq->action;
                        }
                } else if ((dh->fate_action == MLX5_FLOW_FATE_QUEUE &&
-                          !dv_h->rix_sample && !dv_h->rix_dest_array) ||
-                           (dh->fate_action == MLX5_FLOW_FATE_SHARED_RSS)) {
+                          !dv_h->rix_sample && !dv_h->rix_dest_array)) {
+                       struct mlx5_hrxq *hrxq;
+                       uint32_t hrxq_idx;
+
+                       hrxq = flow_dv_hrxq_prepare(dev, dev_flow, rss_desc,
+                                                   &hrxq_idx);
+                       if (!hrxq) {
+                               rte_flow_error_set
+                                       (error, rte_errno,
+                                        RTE_FLOW_ERROR_TYPE_UNSPECIFIED, NULL,
+                                        "cannot get hash queue");
+                               goto error;
+                       }
+                       dh->rix_hrxq = hrxq_idx;
+                       dv->actions[n++] = hrxq->action;
+               } else if (dh->fate_action == MLX5_FLOW_FATE_SHARED_RSS) {
                        struct mlx5_hrxq *hrxq = NULL;
-                       uint32_t hrxq_idx = __flow_dv_rss_get_hrxq
-                                       (dev, dev_flow, rss_desc, &hrxq);
+                       uint32_t hrxq_idx;
+
+                       hrxq_idx = __flow_dv_action_rss_hrxq_lookup(dev,
+                                               rss_desc->shared_rss,
+                                               dev_flow->hash_fields,
+                                               !!(dh->layers &
+                                               MLX5_FLOW_LAYER_TUNNEL));
+                       if (hrxq_idx)
+                               hrxq = mlx5_ipool_get
+                                       (priv->sh->ipool[MLX5_IPOOL_HRXQ],
+                                        hrxq_idx);
                        if (!hrxq) {
                                rte_flow_error_set
                                        (error, rte_errno,
@@ -10721,8 +10761,7 @@ flow_dv_apply(struct rte_eth_dev *dev, struct rte_flow *flow,
                                         "cannot get hash queue");
                                goto error;
                        }
-                       if (dh->fate_action == MLX5_FLOW_FATE_QUEUE)
-                               dh->rix_hrxq = hrxq_idx;
+                       dh->rix_srss = rss_desc->shared_rss;
                        dv->actions[n++] = hrxq->action;
                } else if (dh->fate_action == MLX5_FLOW_FATE_DEFAULT_MISS) {
                        if (!priv->sh->default_miss_action) {
@@ -10764,12 +10803,12 @@ error:
                if (dh->fate_action == MLX5_FLOW_FATE_QUEUE && dh->rix_hrxq) {
                        mlx5_hrxq_release(dev, dh->rix_hrxq);
                        dh->rix_hrxq = 0;
+               } else if (dh->fate_action == MLX5_FLOW_FATE_SHARED_RSS) {
+                       dh->rix_srss = 0;
                }
                if (dh->vf_vlan.tag && dh->vf_vlan.created)
                        mlx5_vlan_vmwa_release(dev, &dh->vf_vlan);
        }
-       if (rss_desc->shared_rss)
-               wks->flows[wks->flow_idx - 1].handle->rix_srss = 0;
        rte_errno = err; /* Restore rte_errno. */
        return -rte_errno;
 }
@@ -11037,9 +11076,6 @@ flow_dv_fate_resource_release(struct rte_eth_dev *dev,
                flow_dv_port_id_action_resource_release(dev,
                                handle->rix_port_id_action);
                break;
-       case MLX5_FLOW_FATE_SHARED_RSS:
-               flow_dv_shared_rss_action_release(dev, handle->rix_srss);
-               break;
        default:
                DRV_LOG(DEBUG, "Incorrect fate action:%d", handle->fate_action);
                break;
@@ -11057,11 +11093,11 @@ flow_dv_sample_remove_cb(struct mlx5_cache_list *list __rte_unused,
        struct mlx5_priv *priv = dev->data->dev_private;
 
        if (cache_resource->verbs_action)
-               claim_zero(mlx5_glue->destroy_flow_action
+               claim_zero(mlx5_flow_os_destroy_flow_action
                                (cache_resource->verbs_action));
        if (cache_resource->ft_type == MLX5DV_FLOW_TABLE_TYPE_FDB) {
                if (cache_resource->default_miss)
-                       claim_zero(mlx5_glue->destroy_flow_action
+                       claim_zero(mlx5_flow_os_destroy_flow_action
                          (cache_resource->default_miss));
        }
        if (cache_resource->normal_path_tbl)
@@ -11114,7 +11150,7 @@ flow_dv_dest_array_remove_cb(struct mlx5_cache_list *list __rte_unused,
 
        MLX5_ASSERT(cache_resource->action);
        if (cache_resource->action)
-               claim_zero(mlx5_glue->destroy_flow_action
+               claim_zero(mlx5_flow_os_destroy_flow_action
                                        (cache_resource->action));
        for (; i < cache_resource->num_of_dest; i++)
                flow_dv_sample_sub_actions_release(dev,
@@ -11202,6 +11238,7 @@ flow_dv_destroy(struct rte_eth_dev *dev, struct rte_flow *flow)
 {
        struct mlx5_flow_handle *dev_handle;
        struct mlx5_priv *priv = dev->data->dev_private;
+       uint32_t srss = 0;
 
        if (!flow)
                return;
@@ -11246,10 +11283,15 @@ flow_dv_destroy(struct rte_eth_dev *dev, struct rte_flow *flow)
                if (dev_handle->dvh.rix_tag)
                        flow_dv_tag_release(dev,
                                            dev_handle->dvh.rix_tag);
-               flow_dv_fate_resource_release(dev, dev_handle);
+               if (dev_handle->fate_action != MLX5_FLOW_FATE_SHARED_RSS)
+                       flow_dv_fate_resource_release(dev, dev_handle);
+               else if (!srss)
+                       srss = dev_handle->rix_srss;
                mlx5_ipool_free(priv->sh->ipool[MLX5_IPOOL_MLX5_FLOW],
                           tmp_idx);
        }
+       if (srss)
+               flow_dv_shared_rss_action_release(dev, srss);
 }
 
 /**
@@ -11323,20 +11365,26 @@ __flow_dv_action_rss_hrxqs_release(struct rte_eth_dev *dev,
  */
 static int
 __flow_dv_action_rss_setup(struct rte_eth_dev *dev,
-                       uint32_t action_idx,
-                       struct mlx5_shared_action_rss *action,
-                       struct rte_flow_error *error)
+                          uint32_t action_idx,
+                          struct mlx5_shared_action_rss *action,
+                          struct rte_flow_error *error)
 {
        struct mlx5_flow_rss_desc rss_desc = { 0 };
        size_t i;
        int err;
 
+       if (mlx5_ind_table_obj_setup(dev, action->ind_tbl)) {
+               return rte_flow_error_set(error, rte_errno,
+                                         RTE_FLOW_ERROR_TYPE_UNSPECIFIED, NULL,
+                                         "cannot setup indirection table");
+       }
        memcpy(rss_desc.key, action->origin.key, MLX5_RSS_HASH_KEY_LEN);
        rss_desc.key_len = MLX5_RSS_HASH_KEY_LEN;
        rss_desc.const_q = action->origin.queue;
        rss_desc.queue_num = action->origin.queue_num;
        /* Set non-zero value to indicate a shared RSS. */
        rss_desc.shared_rss = action_idx;
+       rss_desc.ind_tbl = action->ind_tbl;
        for (i = 0; i < MLX5_RSS_HASH_FIELDS_LEN; i++) {
                uint32_t hrxq_idx;
                uint64_t hash_fields = mlx5_rss_hash_fields[i];
@@ -11362,6 +11410,8 @@ __flow_dv_action_rss_setup(struct rte_eth_dev *dev,
 error_hrxq_new:
        err = rte_errno;
        __flow_dv_action_rss_hrxqs_release(dev, action);
+       if (!mlx5_ind_table_obj_release(dev, action->ind_tbl, true))
+               action->ind_tbl = NULL;
        rte_errno = err;
        return -rte_errno;
 }
@@ -11414,7 +11464,18 @@ __flow_dv_action_rss_create(struct rte_eth_dev *dev,
                                   "rss action number out of range");
                goto error_rss_init;
        }
-       shared_action->queue = queue;
+       shared_action->ind_tbl = mlx5_malloc(MLX5_MEM_ZERO,
+                                            sizeof(*shared_action->ind_tbl),
+                                            0, SOCKET_ID_ANY);
+       if (!shared_action->ind_tbl) {
+               rte_flow_error_set(error, ENOMEM,
+                                  RTE_FLOW_ERROR_TYPE_UNSPECIFIED, NULL,
+                                  "cannot allocate resource memory");
+               goto error_rss_init;
+       }
+       memcpy(queue, rss->queue, queue_size);
+       shared_action->ind_tbl->queues = queue;
+       shared_action->ind_tbl->queues_n = rss->queue_num;
        origin = &shared_action->origin;
        origin->func = rss->func;
        origin->level = rss->level;
@@ -11425,11 +11486,11 @@ __flow_dv_action_rss_create(struct rte_eth_dev *dev,
        memcpy(shared_action->key, rss_key, MLX5_RSS_HASH_KEY_LEN);
        origin->key = &shared_action->key[0];
        origin->key_len = MLX5_RSS_HASH_KEY_LEN;
-       memcpy(shared_action->queue, rss->queue, queue_size);
-       origin->queue = shared_action->queue;
+       origin->queue = queue;
        origin->queue_num = rss->queue_num;
        if (__flow_dv_action_rss_setup(dev, idx, shared_action, error))
                goto error_rss_init;
+       rte_spinlock_init(&shared_action->action_rss_sl);
        __atomic_add_fetch(&shared_action->refcnt, 1, __ATOMIC_RELAXED);
        rte_spinlock_lock(&priv->shared_act_sl);
        ILIST_INSERT(priv->sh->ipool[MLX5_IPOOL_RSS_SHARED_ACTIONS],
@@ -11437,9 +11498,12 @@ __flow_dv_action_rss_create(struct rte_eth_dev *dev,
        rte_spinlock_unlock(&priv->shared_act_sl);
        return idx;
 error_rss_init:
-       if (shared_action)
+       if (shared_action) {
+               if (shared_action->ind_tbl)
+                       mlx5_free(shared_action->ind_tbl);
                mlx5_ipool_free(priv->sh->ipool[MLX5_IPOOL_RSS_SHARED_ACTIONS],
                                idx);
+       }
        if (queue)
                mlx5_free(queue);
        return 0;
@@ -11469,6 +11533,7 @@ __flow_dv_action_rss_release(struct rte_eth_dev *dev, uint32_t idx,
            mlx5_ipool_get(priv->sh->ipool[MLX5_IPOOL_RSS_SHARED_ACTIONS], idx);
        uint32_t old_refcnt = 1;
        int remaining;
+       uint16_t *queue = NULL;
 
        if (!shared_rss)
                return rte_flow_error_set(error, EINVAL,
@@ -11476,18 +11541,26 @@ __flow_dv_action_rss_release(struct rte_eth_dev *dev, uint32_t idx,
                                          "invalid shared action");
        remaining = __flow_dv_action_rss_hrxqs_release(dev, shared_rss);
        if (remaining)
-               return rte_flow_error_set(error, ETOOMANYREFS,
+               return rte_flow_error_set(error, EBUSY,
                                          RTE_FLOW_ERROR_TYPE_ACTION,
                                          NULL,
                                          "shared rss hrxq has references");
+       queue = shared_rss->ind_tbl->queues;
+       remaining = mlx5_ind_table_obj_release(dev, shared_rss->ind_tbl, true);
+       if (remaining)
+               return rte_flow_error_set(error, EBUSY,
+                                         RTE_FLOW_ERROR_TYPE_ACTION,
+                                         NULL,
+                                         "shared rss indirection table has"
+                                         " references");
        if (!__atomic_compare_exchange_n(&shared_rss->refcnt, &old_refcnt,
                                         0, 0, __ATOMIC_ACQUIRE,
                                         __ATOMIC_RELAXED))
-               return rte_flow_error_set(error, ETOOMANYREFS,
+               return rte_flow_error_set(error, EBUSY,
                                          RTE_FLOW_ERROR_TYPE_ACTION,
                                          NULL,
                                          "shared rss has references");
-       rte_free(shared_rss->queue);
+       mlx5_free(queue);
        rte_spinlock_lock(&priv->shared_act_sl);
        ILIST_REMOVE(priv->sh->ipool[MLX5_IPOOL_RSS_SHARED_ACTIONS],
                     &priv->rss_shared_actions, idx, shared_rss, next);
@@ -11532,7 +11605,7 @@ flow_dv_action_create(struct rte_eth_dev *dev,
                       MLX5_SHARED_ACTION_TYPE_OFFSET) | ret;
                break;
        case RTE_FLOW_ACTION_TYPE_AGE:
-               ret = flow_dv_translate_create_aso_age(dev, action->conf);
+               ret = flow_dv_translate_create_aso_age(dev, action->conf, err);
                idx = (MLX5_SHARED_ACTION_TYPE_AGE <<
                       MLX5_SHARED_ACTION_TYPE_OFFSET) | ret;
                if (ret) {
@@ -11625,11 +11698,9 @@ __flow_dv_action_rss_update(struct rte_eth_dev *dev, uint32_t idx,
        struct mlx5_priv *priv = dev->data->dev_private;
        struct mlx5_shared_action_rss *shared_rss =
            mlx5_ipool_get(priv->sh->ipool[MLX5_IPOOL_RSS_SHARED_ACTIONS], idx);
-       size_t i;
-       int ret;
+       int ret = 0;
        void *queue = NULL;
-       const uint8_t *rss_key;
-       uint32_t rss_key_len;
+       uint16_t *queue_old = NULL;
        uint32_t queue_size = action_conf->queue_num * sizeof(uint16_t);
 
        if (!shared_rss)
@@ -11644,42 +11715,24 @@ __flow_dv_action_rss_update(struct rte_eth_dev *dev, uint32_t idx,
                                          RTE_FLOW_ERROR_TYPE_UNSPECIFIED,
                                          NULL,
                                          "cannot allocate resource memory");
-       if (action_conf->key) {
-               rss_key = action_conf->key;
-               rss_key_len = action_conf->key_len;
+       memcpy(queue, action_conf->queue, queue_size);
+       MLX5_ASSERT(shared_rss->ind_tbl);
+       rte_spinlock_lock(&shared_rss->action_rss_sl);
+       queue_old = shared_rss->ind_tbl->queues;
+       ret = mlx5_ind_table_obj_modify(dev, shared_rss->ind_tbl,
+                                       queue, action_conf->queue_num, true);
+       if (ret) {
+               mlx5_free(queue);
+               ret = rte_flow_error_set(error, rte_errno,
+                                         RTE_FLOW_ERROR_TYPE_ACTION, NULL,
+                                         "cannot update indirection table");
        } else {
-               rss_key = rss_hash_default_key;
-               rss_key_len = MLX5_RSS_HASH_KEY_LEN;
+               mlx5_free(queue_old);
+               shared_rss->origin.queue = queue;
+               shared_rss->origin.queue_num = action_conf->queue_num;
        }
-       for (i = 0; i < MLX5_RSS_HASH_FIELDS_LEN; i++) {
-               uint32_t hrxq_idx;
-               uint64_t hash_fields = mlx5_rss_hash_fields[i];
-               int tunnel;
-
-               for (tunnel = 0; tunnel < 2; tunnel++) {
-                       hrxq_idx = __flow_dv_action_rss_hrxq_lookup
-                                       (dev, idx, hash_fields, tunnel);
-                       MLX5_ASSERT(hrxq_idx);
-                       ret = mlx5_hrxq_modify
-                               (dev, hrxq_idx,
-                                rss_key, rss_key_len,
-                                hash_fields,
-                                action_conf->queue, action_conf->queue_num);
-                       if (ret) {
-                               mlx5_free(queue);
-                               return rte_flow_error_set
-                                       (error, rte_errno,
-                                        RTE_FLOW_ERROR_TYPE_ACTION, NULL,
-                                        "cannot update hash queue");
-                       }
-               }
-       }
-       mlx5_free(shared_rss->queue);
-       shared_rss->queue = queue;
-       memcpy(shared_rss->queue, action_conf->queue, queue_size);
-       shared_rss->origin.queue = shared_rss->queue;
-       shared_rss->origin.queue_num = action_conf->queue_num;
-       return 0;
+       rte_spinlock_unlock(&shared_rss->action_rss_sl);
+       return ret;
 }
 
 /**
@@ -12559,18 +12612,18 @@ flow_dv_sync_domain(struct rte_eth_dev *dev, uint32_t domains, uint32_t flags)
        int ret = 0;
 
        if ((domains & MLX5_DOMAIN_BIT_NIC_RX) && priv->sh->rx_domain != NULL) {
-               ret = mlx5_glue->dr_sync_domain(priv->sh->rx_domain,
+               ret = mlx5_os_flow_dr_sync_domain(priv->sh->rx_domain,
                                                flags);
                if (ret != 0)
                        return ret;
        }
        if ((domains & MLX5_DOMAIN_BIT_NIC_TX) && priv->sh->tx_domain != NULL) {
-               ret = mlx5_glue->dr_sync_domain(priv->sh->tx_domain, flags);
+               ret = mlx5_os_flow_dr_sync_domain(priv->sh->tx_domain, flags);
                if (ret != 0)
                        return ret;
        }
        if ((domains & MLX5_DOMAIN_BIT_FDB) && priv->sh->fdb_domain != NULL) {
-               ret = mlx5_glue->dr_sync_domain(priv->sh->fdb_domain, flags);
+               ret = mlx5_os_flow_dr_sync_domain(priv->sh->fdb_domain, flags);
                if (ret != 0)
                        return ret;
        }