net/mlx5: reorganize flow API structure
[dpdk.git] / drivers / net / mlx5 / mlx5.c
index ca5bc14..cc13e44 100644 (file)
@@ -6,7 +6,6 @@
 #include <stddef.h>
 #include <unistd.h>
 #include <string.h>
-#include <assert.h>
 #include <stdint.h>
 #include <stdlib.h>
 #include <errno.h>
@@ -30,7 +29,6 @@
 #include <rte_pci.h>
 #include <rte_bus_pci.h>
 #include <rte_common.h>
-#include <rte_config.h>
 #include <rte_kvargs.h>
 #include <rte_rwlock.h>
 #include <rte_spinlock.h>
@@ -40,6 +38,7 @@
 #include <mlx5_glue.h>
 #include <mlx5_devx_cmds.h>
 #include <mlx5_common.h>
+#include <mlx5_common_mp.h>
 
 #include "mlx5_defs.h"
 #include "mlx5.h"
@@ -48,6 +47,7 @@
 #include "mlx5_autoconf.h"
 #include "mlx5_mr.h"
 #include "mlx5_flow.h"
+#include "rte_pmd_mlx5.h"
 
 /* Device parameter to enable RX completion queue compression. */
 #define MLX5_RXQ_CQE_COMP_EN "rxq_cqe_comp_en"
@@ -64,6 +64,9 @@
 /* Device parameter to configure log 2 of the number of strides for MPRQ. */
 #define MLX5_RX_MPRQ_LOG_STRIDE_NUM "mprq_log_stride_num"
 
+/* Device parameter to configure log 2 of the stride size for MPRQ. */
+#define MLX5_RX_MPRQ_LOG_STRIDE_SIZE "mprq_log_stride_size"
+
 /* Device parameter to limit the size of memcpy'd packet for MPRQ. */
 #define MLX5_RX_MPRQ_MAX_MEMCPY_LEN "mprq_max_memcpy_len"
 
 /* Configure timeout of LRO session (in microseconds). */
 #define MLX5_LRO_TIMEOUT_USEC "lro_timeout_usec"
 
+/*
+ * Device parameter to configure the total data buffer size for a single
+ * hairpin queue (logarithm value).
+ */
+#define MLX5_HP_BUF_SIZE "hp_buf_log_sz"
+
 #ifndef HAVE_IBV_MLX5_MOD_MPW
 #define MLX5DV_CONTEXT_FLAGS_MPW_ALLOWED (1 << 2)
 #define MLX5DV_CONTEXT_FLAGS_ENHANCED_MPW (1 << 3)
@@ -189,6 +198,120 @@ struct mlx5_dev_spawn_data {
 static LIST_HEAD(, mlx5_ibv_shared) mlx5_ibv_list = LIST_HEAD_INITIALIZER();
 static pthread_mutex_t mlx5_ibv_list_mutex = PTHREAD_MUTEX_INITIALIZER;
 
+static struct mlx5_indexed_pool_config mlx5_ipool_cfg[] = {
+#ifdef HAVE_IBV_FLOW_DV_SUPPORT
+       {
+               .size = sizeof(struct mlx5_flow_dv_encap_decap_resource),
+               .trunk_size = 64,
+               .grow_trunk = 3,
+               .grow_shift = 2,
+               .need_lock = 0,
+               .release_mem_en = 1,
+               .malloc = rte_malloc_socket,
+               .free = rte_free,
+               .type = "mlx5_encap_decap_ipool",
+       },
+       {
+               .size = sizeof(struct mlx5_flow_dv_push_vlan_action_resource),
+               .trunk_size = 64,
+               .grow_trunk = 3,
+               .grow_shift = 2,
+               .need_lock = 0,
+               .release_mem_en = 1,
+               .malloc = rte_malloc_socket,
+               .free = rte_free,
+               .type = "mlx5_push_vlan_ipool",
+       },
+       {
+               .size = sizeof(struct mlx5_flow_dv_tag_resource),
+               .trunk_size = 64,
+               .grow_trunk = 3,
+               .grow_shift = 2,
+               .need_lock = 0,
+               .release_mem_en = 1,
+               .malloc = rte_malloc_socket,
+               .free = rte_free,
+               .type = "mlx5_tag_ipool",
+       },
+       {
+               .size = sizeof(struct mlx5_flow_dv_port_id_action_resource),
+               .trunk_size = 64,
+               .grow_trunk = 3,
+               .grow_shift = 2,
+               .need_lock = 0,
+               .release_mem_en = 1,
+               .malloc = rte_malloc_socket,
+               .free = rte_free,
+               .type = "mlx5_port_id_ipool",
+       },
+       {
+               .size = sizeof(struct mlx5_flow_tbl_data_entry),
+               .trunk_size = 64,
+               .grow_trunk = 3,
+               .grow_shift = 2,
+               .need_lock = 0,
+               .release_mem_en = 1,
+               .malloc = rte_malloc_socket,
+               .free = rte_free,
+               .type = "mlx5_jump_ipool",
+       },
+#endif
+       {
+               .size = sizeof(struct mlx5_flow_meter),
+               .trunk_size = 64,
+               .grow_trunk = 3,
+               .grow_shift = 2,
+               .need_lock = 0,
+               .release_mem_en = 1,
+               .malloc = rte_malloc_socket,
+               .free = rte_free,
+               .type = "mlx5_meter_ipool",
+       },
+       {
+               .size = sizeof(struct mlx5_flow_mreg_copy_resource),
+               .trunk_size = 64,
+               .grow_trunk = 3,
+               .grow_shift = 2,
+               .need_lock = 0,
+               .release_mem_en = 1,
+               .malloc = rte_malloc_socket,
+               .free = rte_free,
+               .type = "mlx5_mcp_ipool",
+       },
+       {
+               .size = (sizeof(struct mlx5_hrxq) + MLX5_RSS_HASH_KEY_LEN),
+               .trunk_size = 64,
+               .grow_trunk = 3,
+               .grow_shift = 2,
+               .need_lock = 0,
+               .release_mem_en = 1,
+               .malloc = rte_malloc_socket,
+               .free = rte_free,
+               .type = "mlx5_hrxq_ipool",
+       },
+       {
+               .size = sizeof(struct mlx5_flow_handle),
+               .trunk_size = 64,
+               .grow_trunk = 3,
+               .grow_shift = 2,
+               .need_lock = 0,
+               .release_mem_en = 1,
+               .malloc = rte_malloc_socket,
+               .free = rte_free,
+               .type = "mlx5_flow_handle_ipool",
+       },
+       {
+               .size = sizeof(struct rte_flow),
+               .trunk_size = 4096,
+               .need_lock = 1,
+               .release_mem_en = 1,
+               .malloc = rte_malloc_socket,
+               .free = rte_free,
+               .type = "rte_flow_ipool",
+       },
+};
+
+
 #define MLX5_FLOW_MIN_ID_POOL_SIZE 512
 #define MLX5_ID_GENERATION_ARRAY_FACTOR 16
 
@@ -296,7 +419,7 @@ mlx5_flow_id_release(struct mlx5_flow_id_pool *pool, uint32_t id)
        if (pool->curr == pool->last) {
                size = pool->curr - pool->free_arr;
                size2 = size * MLX5_ID_GENERATION_ARRAY_FACTOR;
-               assert(size2 > size);
+               MLX5_ASSERT(size2 > size);
                mem = rte_malloc("", size2 * sizeof(uint32_t), 0);
                if (!mem) {
                        DRV_LOG(ERR, "can't allocate mem for id pool");
@@ -386,9 +509,11 @@ mlx5_flow_counters_mng_close(struct mlx5_ibv_shared *sh)
                                        claim_zero
                                        (mlx5_glue->destroy_flow_action
                                               (pool->counters_raw[j].action));
-                               if (!batch && pool->counters_raw[j].dcs)
+                               if (!batch && MLX5_GET_POOL_CNT_EXT
+                                   (pool, j)->dcs)
                                        claim_zero(mlx5_devx_cmd_destroy
-                                                 (pool->counters_raw[j].dcs));
+                                                 (MLX5_GET_POOL_CNT_EXT
+                                                 (pool, j)->dcs));
                        }
                        TAILQ_REMOVE(&sh->cmng.ccont[i].pool_list, pool,
                                     next);
@@ -405,6 +530,49 @@ mlx5_flow_counters_mng_close(struct mlx5_ibv_shared *sh)
        memset(&sh->cmng, 0, sizeof(sh->cmng));
 }
 
+/**
+ * Initialize the flow resources' indexed mempool.
+ *
+ * @param[in] sh
+ *   Pointer to mlx5_ibv_shared object.
+ * @param[in] sh
+ *   Pointer to user dev config.
+ */
+static void
+mlx5_flow_ipool_create(struct mlx5_ibv_shared *sh,
+                      const struct mlx5_dev_config *config __rte_unused)
+{
+       uint8_t i;
+
+#ifdef HAVE_IBV_FLOW_DV_SUPPORT
+       /*
+        * While DV is supported, user chooses the verbs mode,
+        * the mlx5 flow handle size is different with the
+        * MLX5_FLOW_HANDLE_VERBS_SIZE.
+        */
+       if (!config->dv_flow_en)
+               mlx5_ipool_cfg[MLX5_IPOOL_MLX5_FLOW].size =
+                                       MLX5_FLOW_HANDLE_VERBS_SIZE;
+#endif
+       for (i = 0; i < MLX5_IPOOL_MAX; ++i)
+               sh->ipool[i] = mlx5_ipool_create(&mlx5_ipool_cfg[i]);
+}
+
+/**
+ * Release the flow resources' indexed mempool.
+ *
+ * @param[in] sh
+ *   Pointer to mlx5_ibv_shared object.
+ */
+static void
+mlx5_flow_ipool_destroy(struct mlx5_ibv_shared *sh)
+{
+       uint8_t i;
+
+       for (i = 0; i < MLX5_IPOOL_MAX; ++i)
+               mlx5_ipool_destroy(sh->ipool[i]);
+}
+
 /**
  * Extract pdn of PD object using DV API.
  *
@@ -442,7 +610,7 @@ mlx5_config_doorbell_mapping_env(const struct mlx5_dev_config *config)
        char *env;
        int value;
 
-       assert(rte_eal_process_type() == RTE_PROC_PRIMARY);
+       MLX5_ASSERT(rte_eal_process_type() == RTE_PROC_PRIMARY);
        /* Get environment variable to store. */
        env = getenv(MLX5_SHUT_UP_BF);
        value = env ? !!strcmp(env, "0") : MLX5_ARG_UNSET;
@@ -457,7 +625,7 @@ mlx5_config_doorbell_mapping_env(const struct mlx5_dev_config *config)
 static void
 mlx5_restore_doorbell_mapping_env(int value)
 {
-       assert(rte_eal_process_type() == RTE_PROC_PRIMARY);
+       MLX5_ASSERT(rte_eal_process_type() == RTE_PROC_PRIMARY);
        /* Restore the original environment variable state. */
        if (value == MLX5_ARG_UNSET)
                unsetenv(MLX5_SHUT_UP_BF);
@@ -497,9 +665,9 @@ mlx5_alloc_shared_ibctx(const struct mlx5_dev_spawn_data *spawn,
        struct mlx5_devx_tis_attr tis_attr = { 0 };
 #endif
 
-       assert(spawn);
+       MLX5_ASSERT(spawn);
        /* Secondary process should not create the shared context. */
-       assert(rte_eal_process_type() == RTE_PROC_PRIMARY);
+       MLX5_ASSERT(rte_eal_process_type() == RTE_PROC_PRIMARY);
        pthread_mutex_lock(&mlx5_ibv_list_mutex);
        /* Search for IB context by device name. */
        LIST_FOREACH(sh, &mlx5_ibv_list, next) {
@@ -509,7 +677,7 @@ mlx5_alloc_shared_ibctx(const struct mlx5_dev_spawn_data *spawn,
                }
        }
        /* No device found, we have to create new shared context. */
-       assert(spawn->max_port);
+       MLX5_ASSERT(spawn->max_port);
        sh = rte_zmalloc("ethdev shared ib context",
                         sizeof(struct mlx5_ibv_shared) +
                         spawn->max_port *
@@ -596,7 +764,8 @@ mlx5_alloc_shared_ibctx(const struct mlx5_dev_spawn_data *spawn,
                        goto error;
                }
        }
-       sh->flow_id_pool = mlx5_flow_id_pool_alloc(UINT32_MAX);
+       sh->flow_id_pool = mlx5_flow_id_pool_alloc
+                                       ((1 << HAIRPIN_FLOW_ID_BITS) - 1);
        if (!sh->flow_id_pool) {
                DRV_LOG(ERR, "can't create flow id pool");
                err = ENOMEM;
@@ -612,7 +781,7 @@ mlx5_alloc_shared_ibctx(const struct mlx5_dev_spawn_data *spawn,
         * At this point the device is not added to the memory
         * event list yet, context is just being created.
         */
-       err = mlx5_mr_btree_init(&sh->mr.cache,
+       err = mlx5_mr_btree_init(&sh->share_cache.cache,
                                 MLX5_MR_BTREE_CACHE_N * 2,
                                 spawn->pci_dev->device.numa_node);
        if (err) {
@@ -620,6 +789,7 @@ mlx5_alloc_shared_ibctx(const struct mlx5_dev_spawn_data *spawn,
                goto error;
        }
        mlx5_flow_counters_mng_init(sh);
+       mlx5_flow_ipool_create(sh, config);
        /* Add device to memory callback list. */
        rte_rwlock_write_lock(&mlx5_shared_data->mem_event_rwlock);
        LIST_INSERT_HEAD(&mlx5_shared_data->mem_event_cb_list,
@@ -632,7 +802,7 @@ exit:
        return sh;
 error:
        pthread_mutex_unlock(&mlx5_ibv_list_mutex);
-       assert(sh);
+       MLX5_ASSERT(sh);
        if (sh->tis)
                claim_zero(mlx5_devx_cmd_destroy(sh->tis));
        if (sh->td)
@@ -644,7 +814,7 @@ error:
        if (sh->flow_id_pool)
                mlx5_flow_id_pool_release(sh->flow_id_pool);
        rte_free(sh);
-       assert(err > 0);
+       MLX5_ASSERT(err > 0);
        rte_errno = err;
        return NULL;
 }
@@ -660,31 +830,31 @@ static void
 mlx5_free_shared_ibctx(struct mlx5_ibv_shared *sh)
 {
        pthread_mutex_lock(&mlx5_ibv_list_mutex);
-#ifndef NDEBUG
+#ifdef RTE_LIBRTE_MLX5_DEBUG
        /* Check the object presence in the list. */
        struct mlx5_ibv_shared *lctx;
 
        LIST_FOREACH(lctx, &mlx5_ibv_list, next)
                if (lctx == sh)
                        break;
-       assert(lctx);
+       MLX5_ASSERT(lctx);
        if (lctx != sh) {
                DRV_LOG(ERR, "Freeing non-existing shared IB context");
                goto exit;
        }
 #endif
-       assert(sh);
-       assert(sh->refcnt);
+       MLX5_ASSERT(sh);
+       MLX5_ASSERT(sh->refcnt);
        /* Secondary process should not free the shared context. */
-       assert(rte_eal_process_type() == RTE_PROC_PRIMARY);
+       MLX5_ASSERT(rte_eal_process_type() == RTE_PROC_PRIMARY);
        if (--sh->refcnt)
                goto exit;
-       /* Release created Memory Regions. */
-       mlx5_mr_release(sh);
        /* Remove from memory callback device list. */
        rte_rwlock_write_lock(&mlx5_shared_data->mem_event_rwlock);
        LIST_REMOVE(sh, mem_event_cb);
        rte_rwlock_write_unlock(&mlx5_shared_data->mem_event_rwlock);
+       /* Release created Memory Regions. */
+       mlx5_mr_release_cache(&sh->share_cache);
        /* Remove context from the global device list. */
        LIST_REMOVE(sh, next);
        /*
@@ -692,7 +862,8 @@ mlx5_free_shared_ibctx(struct mlx5_ibv_shared *sh)
         *  Only primary process handles async device events.
         **/
        mlx5_flow_counters_mng_close(sh);
-       assert(!sh->intr_cnt);
+       mlx5_flow_ipool_destroy(sh);
+       MLX5_ASSERT(!sh->intr_cnt);
        if (sh->intr_cnt)
                mlx5_intr_callback_unregister
                        (&sh->intr_handle, mlx5_dev_interrupt_handler, sh);
@@ -748,7 +919,7 @@ mlx5_free_table_hash_list(struct mlx5_priv *priv)
        if (pos) {
                tbl_data = container_of(pos, struct mlx5_flow_tbl_data_entry,
                                        entry);
-               assert(tbl_data);
+               MLX5_ASSERT(tbl_data);
                mlx5_hlist_remove(sh->flow_tbls, pos);
                rte_free(tbl_data);
        }
@@ -757,7 +928,7 @@ mlx5_free_table_hash_list(struct mlx5_priv *priv)
        if (pos) {
                tbl_data = container_of(pos, struct mlx5_flow_tbl_data_entry,
                                        entry);
-               assert(tbl_data);
+               MLX5_ASSERT(tbl_data);
                mlx5_hlist_remove(sh->flow_tbls, pos);
                rte_free(tbl_data);
        }
@@ -767,7 +938,7 @@ mlx5_free_table_hash_list(struct mlx5_priv *priv)
        if (pos) {
                tbl_data = container_of(pos, struct mlx5_flow_tbl_data_entry,
                                        entry);
-               assert(tbl_data);
+               MLX5_ASSERT(tbl_data);
                mlx5_hlist_remove(sh->flow_tbls, pos);
                rte_free(tbl_data);
        }
@@ -791,7 +962,7 @@ mlx5_alloc_table_hash_list(struct mlx5_priv *priv)
        char s[MLX5_HLIST_NAMESIZE];
        int err = 0;
 
-       assert(sh);
+       MLX5_ASSERT(sh);
        snprintf(s, sizeof(s), "%s_flow_table", priv->sh->ibdev_name);
        sh->flow_tbls = mlx5_hlist_create(s, MLX5_FLOW_TABLE_HLIST_ARRAY_SIZE);
        if (!sh->flow_tbls) {
@@ -982,9 +1153,9 @@ mlx5_free_shared_dr(struct mlx5_priv *priv)
                return;
        priv->dr_shared = 0;
        sh = priv->sh;
-       assert(sh);
+       MLX5_ASSERT(sh);
 #ifdef HAVE_MLX5DV_DR
-       assert(sh->dv_refcnt);
+       MLX5_ASSERT(sh->dv_refcnt);
        if (sh->dv_refcnt && --sh->dv_refcnt)
                return;
        if (sh->rx_domain) {
@@ -1119,7 +1290,7 @@ mlx5_alloc_verbs_buf(size_t size, void *data)
 
                socket = ctrl->socket;
        }
-       assert(data != NULL);
+       MLX5_ASSERT(data != NULL);
        ret = rte_malloc_socket(__func__, size, alignment, socket);
        if (!ret && size)
                rte_errno = ENOMEM;
@@ -1137,7 +1308,7 @@ mlx5_alloc_verbs_buf(size_t size, void *data)
 static void
 mlx5_free_verbs_buf(void *ptr, void *data __rte_unused)
 {
-       assert(data != NULL);
+       MLX5_ASSERT(data != NULL);
        rte_free(ptr);
 }
 
@@ -1156,7 +1327,7 @@ int
 mlx5_udp_tunnel_port_add(struct rte_eth_dev *dev __rte_unused,
                         struct rte_eth_udp_tunnel *udp_tunnel)
 {
-       assert(udp_tunnel != NULL);
+       MLX5_ASSERT(udp_tunnel != NULL);
        if (udp_tunnel->prot_type == RTE_TUNNEL_TYPE_VXLAN &&
            udp_tunnel->udp_port == 4789)
                return 0;
@@ -1235,9 +1406,20 @@ mlx5_dev_close(struct rte_eth_dev *dev)
        /* In case mlx5_dev_stop() has not been called. */
        mlx5_dev_interrupt_handler_uninstall(dev);
        mlx5_dev_interrupt_handler_devx_uninstall(dev);
+       /*
+        * If default mreg copy action is removed at the stop stage,
+        * the search will return none and nothing will be done anymore.
+        */
+       mlx5_flow_stop_default(dev);
        mlx5_traffic_disable(dev);
-       mlx5_flow_flush(dev, NULL);
+       /*
+        * If all the flows are already flushed in the device stop stage,
+        * then this will return directly without any action.
+        */
+       mlx5_flow_list_flush(dev, &priv->flows, true);
        mlx5_flow_meter_flush(dev, NULL);
+       /* Free the intermediate buffers for flow creation. */
+       mlx5_flow_free_intermediate(dev);
        /* Prevent crashes when queues are still in use. */
        dev->rx_pkt_burst = removed_rx_burst;
        dev->tx_pkt_burst = removed_tx_burst;
@@ -1270,23 +1452,15 @@ mlx5_dev_close(struct rte_eth_dev *dev)
        if (priv->reta_idx != NULL)
                rte_free(priv->reta_idx);
        if (priv->config.vf)
-               mlx5_nl_mac_addr_flush(dev);
+               mlx5_nl_mac_addr_flush(priv->nl_socket_route, mlx5_ifindex(dev),
+                                      dev->data->mac_addrs,
+                                      MLX5_MAX_MAC_ADDRESSES, priv->mac_own);
        if (priv->nl_socket_route >= 0)
                close(priv->nl_socket_route);
        if (priv->nl_socket_rdma >= 0)
                close(priv->nl_socket_rdma);
        if (priv->vmwa_context)
                mlx5_vlan_vmwa_exit(priv->vmwa_context);
-       if (priv->sh) {
-               /*
-                * Free the shared context in last turn, because the cleanup
-                * routines above may use some shared fields, like
-                * mlx5_nl_mac_addr_flush() uses ibdev_path for retrieveing
-                * ifindex if Netlink fails.
-                */
-               mlx5_free_shared_ibctx(priv->sh);
-               priv->sh = NULL;
-       }
        ret = mlx5_hrxq_verify(dev);
        if (ret)
                DRV_LOG(WARNING, "port %u some hash Rx queue still remain",
@@ -1315,6 +1489,16 @@ mlx5_dev_close(struct rte_eth_dev *dev)
        if (ret)
                DRV_LOG(WARNING, "port %u some flows still remain",
                        dev->data->port_id);
+       if (priv->sh) {
+               /*
+                * Free the shared context in last turn, because the cleanup
+                * routines above may use some shared fields, like
+                * mlx5_nl_mac_addr_flush() uses ibdev_path for retrieveing
+                * ifindex if Netlink fails.
+                */
+               mlx5_free_shared_ibctx(priv->sh);
+               priv->sh = NULL;
+       }
        if (priv->domain_id != RTE_ETH_DEV_SWITCH_DOMAIN_ID_INVALID) {
                unsigned int c = 0;
                uint16_t port_id;
@@ -1387,6 +1571,10 @@ const struct eth_dev_ops mlx5_dev_ops = {
        .filter_ctrl = mlx5_dev_filter_ctrl,
        .rx_descriptor_status = mlx5_rx_descriptor_status,
        .tx_descriptor_status = mlx5_tx_descriptor_status,
+       .rxq_info_get = mlx5_rxq_info_get,
+       .txq_info_get = mlx5_txq_info_get,
+       .rx_burst_mode_get = mlx5_rx_burst_mode_get,
+       .tx_burst_mode_get = mlx5_tx_burst_mode_get,
        .rx_queue_count = mlx5_rx_queue_count,
        .rx_queue_intr_enable = mlx5_rx_intr_enable,
        .rx_queue_intr_disable = mlx5_rx_intr_disable,
@@ -1409,6 +1597,10 @@ static const struct eth_dev_ops mlx5_dev_sec_ops = {
        .dev_infos_get = mlx5_dev_infos_get,
        .rx_descriptor_status = mlx5_rx_descriptor_status,
        .tx_descriptor_status = mlx5_tx_descriptor_status,
+       .rxq_info_get = mlx5_rxq_info_get,
+       .txq_info_get = mlx5_txq_info_get,
+       .rx_burst_mode_get = mlx5_rx_burst_mode_get,
+       .tx_burst_mode_get = mlx5_tx_burst_mode_get,
        .get_module_info = mlx5_get_module_info,
        .get_module_eeprom = mlx5_get_module_eeprom,
 };
@@ -1453,6 +1645,10 @@ const struct eth_dev_ops mlx5_dev_ops_isolate = {
        .filter_ctrl = mlx5_dev_filter_ctrl,
        .rx_descriptor_status = mlx5_rx_descriptor_status,
        .tx_descriptor_status = mlx5_tx_descriptor_status,
+       .rxq_info_get = mlx5_rxq_info_get,
+       .txq_info_get = mlx5_txq_info_get,
+       .rx_burst_mode_get = mlx5_rx_burst_mode_get,
+       .tx_burst_mode_get = mlx5_tx_burst_mode_get,
        .rx_queue_intr_enable = mlx5_rx_intr_enable,
        .rx_queue_intr_disable = mlx5_rx_intr_disable,
        .is_removed = mlx5_is_removed,
@@ -1501,6 +1697,8 @@ mlx5_args_check(const char *key, const char *val, void *opaque)
                config->mprq.enabled = !!tmp;
        } else if (strcmp(MLX5_RX_MPRQ_LOG_STRIDE_NUM, key) == 0) {
                config->mprq.stride_num_n = tmp;
+       } else if (strcmp(MLX5_RX_MPRQ_LOG_STRIDE_SIZE, key) == 0) {
+               config->mprq.stride_size_n = tmp;
        } else if (strcmp(MLX5_RX_MPRQ_MAX_MEMCPY_LEN, key) == 0) {
                config->mprq.max_memcpy_len = tmp;
        } else if (strcmp(MLX5_RXQS_MIN_MPRQ, key) == 0) {
@@ -1567,6 +1765,8 @@ mlx5_args_check(const char *key, const char *val, void *opaque)
                config->lro.timeout = tmp;
        } else if (strcmp(MLX5_CLASS_ARG_NAME, key) == 0) {
                DRV_LOG(DEBUG, "class argument is %s.", val);
+       } else if (strcmp(MLX5_HP_BUF_SIZE, key) == 0) {
+               config->log_hp_size = tmp;
        } else {
                DRV_LOG(WARNING, "%s: unknown parameter", key);
                rte_errno = EINVAL;
@@ -1595,6 +1795,7 @@ mlx5_args(struct mlx5_dev_config *config, struct rte_devargs *devargs)
                MLX5_RXQ_PKT_PAD_EN,
                MLX5_RX_MPRQ_EN,
                MLX5_RX_MPRQ_LOG_STRIDE_NUM,
+               MLX5_RX_MPRQ_LOG_STRIDE_SIZE,
                MLX5_RX_MPRQ_MAX_MEMCPY_LEN,
                MLX5_RXQS_MIN_MPRQ,
                MLX5_TXQ_INLINE,
@@ -1619,6 +1820,7 @@ mlx5_args(struct mlx5_dev_config *config, struct rte_devargs *devargs)
                MLX5_MAX_DUMP_FILES_NUM,
                MLX5_LRO_TIMEOUT_USEC,
                MLX5_CLASS_ARG_NAME,
+               MLX5_HP_BUF_SIZE,
                NULL,
        };
        struct rte_kvargs *kvlist;
@@ -1671,7 +1873,7 @@ mlx5_init_once(void)
        if (mlx5_init_shared_data())
                return -rte_errno;
        sd = mlx5_shared_data;
-       assert(sd);
+       MLX5_ASSERT(sd);
        rte_spinlock_lock(&sd->lock);
        switch (rte_eal_process_type()) {
        case RTE_PROC_PRIMARY:
@@ -1681,7 +1883,8 @@ mlx5_init_once(void)
                rte_rwlock_init(&sd->mem_event_rwlock);
                rte_mem_event_callback_register("MLX5_MEM_EVENT_CB",
                                                mlx5_mr_mem_event_cb, NULL);
-               ret = mlx5_mp_init_primary();
+               ret = mlx5_mp_init_primary(MLX5_MP_NAME,
+                                          mlx5_mp_primary_handle);
                if (ret)
                        goto out;
                sd->init_done = true;
@@ -1689,7 +1892,8 @@ mlx5_init_once(void)
        case RTE_PROC_SECONDARY:
                if (ld->init_done)
                        break;
-               ret = mlx5_mp_init_secondary();
+               ret = mlx5_mp_init_secondary(MLX5_MP_NAME,
+                                            mlx5_mp_secondary_handle);
                if (ret)
                        goto out;
                ++sd->secondary_cnt;
@@ -1711,7 +1915,7 @@ out:
  *   key is specified in devargs
  * - if DevX is enabled the inline mode is queried from the
  *   device (HCA attributes and NIC vport context if needed).
- * - otherwise L2 mode (18 bytes) is assumed for ConnectX-4/4LX
+ * - otherwise L2 mode (18 bytes) is assumed for ConnectX-4/4 Lx
  *   and none (0 bytes) for other NICs
  *
  * @param spawn
@@ -1853,7 +2057,7 @@ mlx5_set_metadata_mask(struct rte_eth_dev *dev)
        default:
                meta = 0;
                mark = 0;
-               assert(false);
+               MLX5_ASSERT(false);
                break;
        }
        if (sh->dv_mark_mask && sh->dv_mark_mask != mark)
@@ -1946,7 +2150,7 @@ mlx5_get_dbr(struct rte_eth_dev *dev, struct mlx5_devx_dbr_page **dbr_page)
                ; /* Empty. */
        /* Find the first clear bit. */
        j = rte_bsf64(~page->dbr_bitmap[i]);
-       assert(i < (MLX5_DBR_PER_PAGE / 64));
+       MLX5_ASSERT(i < (MLX5_DBR_PER_PAGE / 64));
        page->dbr_bitmap[i] |= (1 << j);
        page->dbr_count++;
        *dbr_page = page;
@@ -1997,6 +2201,25 @@ mlx5_release_dbr(struct rte_eth_dev *dev, uint32_t umem_id, uint64_t offset)
        return ret;
 }
 
+int
+rte_pmd_mlx5_get_dyn_flag_names(char *names[], unsigned int n)
+{
+       static const char *const dynf_names[] = {
+               RTE_PMD_MLX5_FINE_GRANULARITY_INLINE,
+               RTE_MBUF_DYNFLAG_METADATA_NAME
+       };
+       unsigned int i;
+
+       if (n < RTE_DIM(dynf_names))
+               return -ENOMEM;
+       for (i = 0; i < RTE_DIM(dynf_names); i++) {
+               if (names[i] == NULL)
+                       return -EINVAL;
+               strcpy(names[i], dynf_names[i]);
+       }
+       return RTE_DIM(dynf_names);
+}
+
 /**
  * Check sibling device configurations.
  *
@@ -2020,7 +2243,7 @@ mlx5_dev_check_sibling_config(struct mlx5_priv *priv,
        struct mlx5_dev_config *sh_conf = NULL;
        uint16_t port_id;
 
-       assert(sh);
+       MLX5_ASSERT(sh);
        /* Nothing to compare for the single/first device. */
        if (sh->refcnt == 1)
                return 0;
@@ -2145,6 +2368,8 @@ mlx5_dev_spawn(struct rte_device *dpdk_dev,
        }
        DRV_LOG(DEBUG, "naming Ethernet device \"%s\"", name);
        if (rte_eal_process_type() == RTE_PROC_SECONDARY) {
+               struct mlx5_mp_id mp_id;
+
                eth_dev = rte_eth_dev_attach_secondary(name);
                if (eth_dev == NULL) {
                        DRV_LOG(ERR, "can not attach rte ethdev");
@@ -2156,8 +2381,10 @@ mlx5_dev_spawn(struct rte_device *dpdk_dev,
                err = mlx5_proc_priv_init(eth_dev);
                if (err)
                        return NULL;
+               mp_id.port_id = eth_dev->data->port_id;
+               strlcpy(mp_id.name, MLX5_MP_NAME, RTE_MP_MAX_NAME_LEN);
                /* Receive command fd from primary process */
-               err = mlx5_mp_req_verbs_cmd_fd(eth_dev);
+               err = mlx5_mp_req_verbs_cmd_fd(&mp_id);
                if (err < 0)
                        return NULL;
                /* Remap UAR for Tx queues. */
@@ -2250,8 +2477,6 @@ mlx5_dev_spawn(struct rte_device *dpdk_dev,
                        mprq_caps.min_single_wqe_log_num_of_strides;
                mprq_max_stride_num_n =
                        mprq_caps.max_single_wqe_log_num_of_strides;
-               config.mprq.stride_num_n = RTE_MAX(MLX5_MPRQ_STRIDE_NUM_N,
-                                                  mprq_min_stride_num_n);
        }
 #endif
        if (RTE_CACHE_LINE_SIZE == 128 &&
@@ -2321,6 +2546,8 @@ mlx5_dev_spawn(struct rte_device *dpdk_dev,
        priv->ibv_port = spawn->ibv_port;
        priv->pci_dev = spawn->pci_dev;
        priv->mtu = RTE_ETHER_MTU;
+       priv->mp_id.port_id = port_id;
+       strlcpy(priv->mp_id.name, MLX5_MP_NAME, RTE_MP_MAX_NAME_LEN);
 #ifndef RTE_ARCH_64
        /* Initialize UAR access locks for 32bit implementations. */
        rte_spinlock_init(&priv->uar_lock_cq);
@@ -2330,7 +2557,6 @@ mlx5_dev_spawn(struct rte_device *dpdk_dev,
        /* Some internal functions rely on Netlink sockets, open them now. */
        priv->nl_socket_rdma = mlx5_nl_init(NETLINK_RDMA);
        priv->nl_socket_route = mlx5_nl_init(NETLINK_ROUTE);
-       priv->nl_sn = 0;
        priv->representor = !!switch_info->representor;
        priv->master = !!switch_info->master;
        priv->domain_id = RTE_ETH_DEV_SWITCH_DOMAIN_ID_INVALID;
@@ -2566,17 +2792,32 @@ mlx5_dev_spawn(struct rte_device *dpdk_dev,
 #endif
        }
        if (config.mprq.enabled && mprq) {
-               if (config.mprq.stride_num_n > mprq_max_stride_num_n ||
-                   config.mprq.stride_num_n < mprq_min_stride_num_n) {
+               if (config.mprq.stride_num_n &&
+                   (config.mprq.stride_num_n > mprq_max_stride_num_n ||
+                    config.mprq.stride_num_n < mprq_min_stride_num_n)) {
                        config.mprq.stride_num_n =
-                               RTE_MAX(MLX5_MPRQ_STRIDE_NUM_N,
-                                       mprq_min_stride_num_n);
+                               RTE_MIN(RTE_MAX(MLX5_MPRQ_STRIDE_NUM_N,
+                                               mprq_min_stride_num_n),
+                                       mprq_max_stride_num_n);
                        DRV_LOG(WARNING,
                                "the number of strides"
                                " for Multi-Packet RQ is out of range,"
                                " setting default value (%u)",
                                1 << config.mprq.stride_num_n);
                }
+               if (config.mprq.stride_size_n &&
+                   (config.mprq.stride_size_n > mprq_max_stride_size_n ||
+                    config.mprq.stride_size_n < mprq_min_stride_size_n)) {
+                       config.mprq.stride_size_n =
+                               RTE_MIN(RTE_MAX(MLX5_MPRQ_STRIDE_SIZE_N,
+                                               mprq_min_stride_size_n),
+                                       mprq_max_stride_size_n);
+                       DRV_LOG(WARNING,
+                               "the size of a stride"
+                               " for Multi-Packet RQ is out of range,"
+                               " setting default value (%u)",
+                               1 << config.mprq.stride_size_n);
+               }
                config.mprq.min_stride_size_n = mprq_min_stride_size_n;
                config.mprq.max_stride_size_n = mprq_max_stride_size_n;
        } else if (config.mprq.enabled && !mprq) {
@@ -2602,7 +2843,7 @@ mlx5_dev_spawn(struct rte_device *dpdk_dev,
         * is permanent throughout the lifetime of device. So, we may store
         * the ifindex here and use the cached value further.
         */
-       assert(spawn->ifindex);
+       MLX5_ASSERT(spawn->ifindex);
        priv->if_index = spawn->ifindex;
        eth_dev->data->dev_private = priv;
        priv->dev_data = eth_dev->data;
@@ -2623,7 +2864,7 @@ mlx5_dev_spawn(struct rte_device *dpdk_dev,
                mac.addr_bytes[0], mac.addr_bytes[1],
                mac.addr_bytes[2], mac.addr_bytes[3],
                mac.addr_bytes[4], mac.addr_bytes[5]);
-#ifndef NDEBUG
+#ifdef RTE_LIBRTE_MLX5_DEBUG
        {
                char ifname[IF_NAMESIZE];
 
@@ -2650,9 +2891,12 @@ mlx5_dev_spawn(struct rte_device *dpdk_dev,
        /* Register MAC address. */
        claim_zero(mlx5_mac_addr_add(eth_dev, &mac, 0, 0));
        if (config.vf && config.vf_nl_en)
-               mlx5_nl_mac_addr_sync(eth_dev);
-       TAILQ_INIT(&priv->flows);
-       TAILQ_INIT(&priv->ctrl_flows);
+               mlx5_nl_mac_addr_sync(priv->nl_socket_route,
+                                     mlx5_ifindex(eth_dev),
+                                     eth_dev->data->mac_addrs,
+                                     MLX5_MAX_MAC_ADDRESSES);
+       priv->flows = 0;
+       priv->ctrl_flows = 0;
        TAILQ_INIT(&priv->flow_meters);
        TAILQ_INIT(&priv->flow_meter_profiles);
        /* Hint libmlx5 to use PMD allocator for data plane resources */
@@ -2725,6 +2969,11 @@ mlx5_dev_spawn(struct rte_device *dpdk_dev,
                        err = ENOTSUP;
                        goto error;
        }
+       /*
+        * Allocate the buffer for flow creating, just once.
+        * The allocation must be done before any flow creating.
+        */
+       mlx5_flow_alloc_intermediate(eth_dev);
        /* Query availibility of metadata reg_c's. */
        err = mlx5_flow_discover_mreg_c(eth_dev);
        if (err < 0) {
@@ -2782,7 +3031,7 @@ error:
        }
        if (sh)
                mlx5_free_shared_ibctx(sh);
-       assert(err > 0);
+       MLX5_ASSERT(err > 0);
        rte_errno = err;
        return NULL;
 }
@@ -2885,7 +3134,7 @@ mlx5_device_bond_pci_match(const struct ibv_device *ibv_dev,
        if (!file)
                return -1;
        /* Use safe format to check maximal buffer length. */
-       assert(atol(RTE_STR(IF_NAMESIZE)) == IF_NAMESIZE);
+       MLX5_ASSERT(atol(RTE_STR(IF_NAMESIZE)) == IF_NAMESIZE);
        while (fscanf(file, "%" RTE_STR(IF_NAMESIZE) "s", ifname) == 1) {
                char tmp_str[IF_NAMESIZE + 32];
                struct rte_pci_addr pci_addr;
@@ -2983,7 +3232,7 @@ mlx5_pci_probe(struct rte_pci_driver *pci_drv __rte_unused,
                        strerror(rte_errno));
                return -rte_errno;
        }
-       assert(pci_drv == &mlx5_driver);
+       MLX5_ASSERT(pci_drv == &mlx5_driver);
        errno = 0;
        ibv_list = mlx5_glue->get_device_list(&ret);
        if (!ibv_list) {
@@ -3104,10 +3353,10 @@ mlx5_pci_probe(struct rte_pci_driver *pci_drv __rte_unused,
                 * it may be E-Switch master device and representors.
                 * We have to perform identification trough the ports.
                 */
-               assert(nl_rdma >= 0);
-               assert(ns == 0);
-               assert(nd == 1);
-               assert(np);
+               MLX5_ASSERT(nl_rdma >= 0);
+               MLX5_ASSERT(ns == 0);
+               MLX5_ASSERT(nd == 1);
+               MLX5_ASSERT(np);
                for (i = 1; i <= np; ++i) {
                        list[ns].max_port = np;
                        list[ns].ibv_port = i;
@@ -3282,7 +3531,7 @@ mlx5_pci_probe(struct rte_pci_driver *pci_drv __rte_unused,
                        goto exit;
                }
        }
-       assert(ns);
+       MLX5_ASSERT(ns);
        /*
         * Sort list to probe devices in natural order for users convenience
         * (i.e. master first, then representors from lowest to highest ID).
@@ -3302,12 +3551,14 @@ mlx5_pci_probe(struct rte_pci_driver *pci_drv __rte_unused,
                .mr_ext_memseg_en = 1,
                .mprq = {
                        .enabled = 0, /* Disabled by default. */
-                       .stride_num_n = MLX5_MPRQ_STRIDE_NUM_N,
+                       .stride_num_n = 0,
+                       .stride_size_n = 0,
                        .max_memcpy_len = MLX5_MPRQ_MEMCPY_DEFAULT_LEN,
                        .min_rxqs_num = MLX5_MPRQ_MIN_RXQS,
                },
                .dv_esw_en = 1,
                .dv_flow_en = 1,
+               .log_hp_size = MLX5_ARG_UNSET,
        };
        /* Device specific configuration. */
        switch (pci_dev->id.device_id) {
@@ -3377,7 +3628,7 @@ exit:
                close(nl_route);
        if (list)
                rte_free(list);
-       assert(ibv_list);
+       MLX5_ASSERT(ibv_list);
        mlx5_glue->free_device_list(ibv_list);
        return ret;
 }
@@ -3495,6 +3746,10 @@ static const struct rte_pci_id mlx5_pci_id_map[] = {
                RTE_PCI_DEVICE(PCI_VENDOR_ID_MELLANOX,
                                PCI_DEVICE_ID_MELLANOX_CONNECTX6DXVF)
        },
+       {
+               RTE_PCI_DEVICE(PCI_VENDOR_ID_MELLANOX,
+                               PCI_DEVICE_ID_MELLANOX_CONNECTX6DXBF)
+       },
        {
                .vendor_id = 0
        }