common/mlx5: fix PCI driver name
[dpdk.git] / drivers / net / mlx5 / linux / mlx5_os.c
index d95082f..ed3f020 100644 (file)
@@ -226,13 +226,12 @@ mlx5_alloc_shared_dr(struct mlx5_priv *priv)
 {
        struct mlx5_dev_ctx_shared *sh = priv->sh;
        char s[MLX5_HLIST_NAMESIZE];
-       int err = 0;
+       int err;
 
-       if (!sh->flow_tbls)
-               err = mlx5_alloc_table_hash_list(priv);
-       else
-               DRV_LOG(DEBUG, "sh->flow_tbls[%p] already created, reuse\n",
-                       (void *)sh->flow_tbls);
+       MLX5_ASSERT(sh && sh->refcnt);
+       if (sh->refcnt > 1)
+               return 0;
+       err = mlx5_alloc_table_hash_list(priv);
        if (err)
                return err;
        /* Create tags hash list table. */
@@ -261,12 +260,6 @@ mlx5_alloc_shared_dr(struct mlx5_priv *priv)
 #ifdef HAVE_MLX5DV_DR
        void *domain;
 
-       if (sh->dv_refcnt) {
-               /* Shared DV/DR structures is already initialized. */
-               sh->dv_refcnt++;
-               priv->dr_shared = 1;
-               return 0;
-       }
        /* Reference counter is zero, we should initialize structures. */
        domain = mlx5_glue->dr_create_domain(sh->ctx,
                                             MLX5DV_DR_DOMAIN_TYPE_NIC_RX);
@@ -298,6 +291,12 @@ mlx5_alloc_shared_dr(struct mlx5_priv *priv)
                sh->esw_drop_action = mlx5_glue->dr_create_flow_action_drop();
        }
 #endif
+       if (!sh->tunnel_hub)
+               err = mlx5_alloc_tunnel_hub(sh);
+       if (err) {
+               DRV_LOG(ERR, "mlx5_alloc_tunnel_hub failed err=%d", err);
+               goto error;
+       }
        if (priv->config.reclaim_mode == MLX5_RCM_AGGR) {
                mlx5_glue->dr_reclaim_domain_memory(sh->rx_domain, 1);
                mlx5_glue->dr_reclaim_domain_memory(sh->tx_domain, 1);
@@ -306,8 +305,6 @@ mlx5_alloc_shared_dr(struct mlx5_priv *priv)
        }
        sh->pop_vlan_action = mlx5_glue->dr_create_flow_action_pop_vlan();
 #endif /* HAVE_MLX5DV_DR */
-       sh->dv_refcnt++;
-       priv->dr_shared = 1;
        return 0;
 error:
        /* Rollback the created objects. */
@@ -344,6 +341,10 @@ error:
                mlx5_hlist_destroy(sh->tag_table, NULL, NULL);
                sh->tag_table = NULL;
        }
+       if (sh->tunnel_hub) {
+               mlx5_release_tunnel_hub(sh, priv->dev_port);
+               sh->tunnel_hub = NULL;
+       }
        mlx5_free_table_hash_list(priv);
        return err;
 }
@@ -357,17 +358,12 @@ error:
 void
 mlx5_os_free_shared_dr(struct mlx5_priv *priv)
 {
-       struct mlx5_dev_ctx_shared *sh;
+       struct mlx5_dev_ctx_shared *sh = priv->sh;
 
-       if (!priv->dr_shared)
+       MLX5_ASSERT(sh && sh->refcnt);
+       if (sh->refcnt > 1)
                return;
-       priv->dr_shared = 0;
-       sh = priv->sh;
-       MLX5_ASSERT(sh);
 #ifdef HAVE_MLX5DV_DR
-       MLX5_ASSERT(sh->dv_refcnt);
-       if (sh->dv_refcnt && --sh->dv_refcnt)
-               return;
        if (sh->rx_domain) {
                mlx5_glue->dr_destroy_domain(sh->rx_domain);
                sh->rx_domain = NULL;
@@ -405,6 +401,10 @@ mlx5_os_free_shared_dr(struct mlx5_priv *priv)
                mlx5_hlist_destroy(sh->tag_table, NULL, NULL);
                sh->tag_table = NULL;
        }
+       if (sh->tunnel_hub) {
+               mlx5_release_tunnel_hub(sh, priv->dev_port);
+               sh->tunnel_hub = NULL;
+       }
        mlx5_free_table_hash_list(priv);
 }
 
@@ -561,6 +561,45 @@ mlx5_os_txq_obj_release(struct mlx5_txq_obj *txq_obj)
        mlx5_txq_ibv_obj_release(txq_obj);
 }
 
+/**
+ * DV flow counter mode detect and config.
+ *
+ * @param dev
+ *   Pointer to rte_eth_dev structure.
+ *
+ */
+static void
+mlx5_flow_counter_mode_config(struct rte_eth_dev *dev __rte_unused)
+{
+#ifdef HAVE_IBV_FLOW_DV_SUPPORT
+       struct mlx5_priv *priv = dev->data->dev_private;
+       struct mlx5_dev_ctx_shared *sh = priv->sh;
+       bool fallback;
+
+#ifndef HAVE_IBV_DEVX_ASYNC
+       fallback = true;
+#else
+       fallback = false;
+       if (!priv->config.devx || !priv->config.dv_flow_en ||
+           !priv->config.hca_attr.flow_counters_dump ||
+           !(priv->config.hca_attr.flow_counter_bulk_alloc_bitmap & 0x4) ||
+           (mlx5_flow_dv_discover_counter_offset_support(dev) == -ENOTSUP))
+               fallback = true;
+#endif
+       if (fallback)
+               DRV_LOG(INFO, "Use fall-back DV counter management. Flow "
+                       "counter dump:%d, bulk_alloc_bitmap:0x%hhx.",
+                       priv->config.hca_attr.flow_counters_dump,
+                       priv->config.hca_attr.flow_counter_bulk_alloc_bitmap);
+       /* Initialize fallback mode only on the port initializes sh. */
+       if (sh->refcnt == 1)
+               sh->cmng.counter_fallback = fallback;
+       else if (fallback != sh->cmng.counter_fallback)
+               DRV_LOG(WARNING, "Port %d in sh has different fallback mode "
+                       "with others:%d.", PORT_ID(priv), fallback);
+#endif
+}
+
 /**
  * Spawn an Ethernet device from Verbs information.
  *
@@ -708,6 +747,10 @@ err_secondary:
                        strerror(rte_errno));
                goto error;
        }
+       if (config->dv_miss_info) {
+               if (switch_info->master || switch_info->representor)
+                       config->dv_xmeta_en = MLX5_XMETA_MODE_META16;
+       }
        mlx5_malloc_mem_select(config->sys_mem_en);
        sh = mlx5_alloc_shared_dev_ctx(spawn, config);
        if (!sh)
@@ -1029,19 +1072,16 @@ err_secondary:
                DRV_LOG(INFO, "Rx CQE padding is enabled");
        }
        if (config->devx) {
-               priv->counter_fallback = 0;
                err = mlx5_devx_cmd_query_hca_attr(sh->ctx, &config->hca_attr);
                if (err) {
                        err = -err;
                        goto error;
                }
-               if (!config->hca_attr.flow_counters_dump)
-                       priv->counter_fallback = 1;
-#ifndef HAVE_IBV_DEVX_ASYNC
-               priv->counter_fallback = 1;
-#endif
-               if (priv->counter_fallback)
-                       DRV_LOG(INFO, "Use fall-back DV counter management");
+               /* Check relax ordering support. */
+               if (config->hca_attr.relaxed_ordering_write &&
+                   config->hca_attr.relaxed_ordering_read  &&
+                   !haswell_broadwell_cpu)
+                       sh->cmng.relaxed_ordering = 1;
                /* Check for LRO support. */
                if (config->dest_tir && config->hca_attr.lro_cap &&
                    config->dv_flow_en) {
@@ -1443,6 +1483,7 @@ err_secondary:
                        goto error;
                }
        }
+       mlx5_flow_counter_mode_config(eth_dev);
        return eth_dev;
 error:
        if (priv) {