#include "mlx5_autoconf.h"
#include "mlx5_mr.h"
#include "mlx5_flow.h"
+#include "mlx5_flow_os.h"
#include "rte_pmd_mlx5.h"
/* Device parameter to enable RX completion queue compression. */
for (j = 0; j < MLX5_COUNTERS_PER_POOL; ++j)
if (pool->actions[j].dr_action)
claim_zero
- (mlx5_glue->destroy_flow_action
- (pool->actions[j].dr_action));
+ (mlx5_flow_os_destroy_flow_action
+ (pool->actions[j].dr_action));
mlx5_free(pool);
}
mlx5_free(sh->aso_age_mng->pools);
LIST_REMOVE(mng, next);
claim_zero(mlx5_devx_cmd_destroy(mng->dm));
- claim_zero(mlx5_glue->devx_umem_dereg(mng->umem));
+ claim_zero(mlx5_os_umem_dereg(mng->umem));
mlx5_free(mem);
}
if (cnt->action)
claim_zero
- (mlx5_glue->destroy_flow_action
+ (mlx5_flow_os_destroy_flow_action
(cnt->action));
if (fallback && MLX5_POOL_GET_CNT
(pool, j)->dcs_when_free)
sh->port[i].ih_port_id = RTE_MAX_ETHPORTS;
sh->port[i].devx_ih_port_id = RTE_MAX_ETHPORTS;
}
- sh->pd = mlx5_glue->alloc_pd(sh->ctx);
+ sh->pd = mlx5_os_alloc_pd(sh->ctx);
if (sh->pd == NULL) {
DRV_LOG(ERR, "PD allocation failure");
err = ENOMEM;
if (sh->tx_uar)
mlx5_glue->devx_free_uar(sh->tx_uar);
if (sh->pd)
- claim_zero(mlx5_glue->dealloc_pd(sh->pd));
+ claim_zero(mlx5_os_dealloc_pd(sh->pd));
if (sh->ctx)
claim_zero(mlx5_glue->close_device(sh->ctx));
mlx5_free(sh);
sh->tx_uar = NULL;
}
if (sh->pd)
- claim_zero(mlx5_glue->dealloc_pd(sh->pd));
+ claim_zero(mlx5_os_dealloc_pd(sh->pd));
if (sh->tis)
claim_zero(mlx5_devx_cmd_destroy(sh->tis));
if (sh->td)