net/txgbe: add queue stats mapping
[dpdk.git] / drivers / net / mlx5 / linux / mlx5_os.c
index 69123e1..40f9446 100644 (file)
@@ -45,6 +45,8 @@
 #include "mlx5_flow.h"
 #include "rte_pmd_mlx5.h"
 #include "mlx5_verbs.h"
+#include "mlx5_nl.h"
+#include "mlx5_devx.h"
 
 #define MLX5_TAGS_HLIST_ARRAY_SIZE 8192
 
@@ -224,13 +226,12 @@ mlx5_alloc_shared_dr(struct mlx5_priv *priv)
 {
        struct mlx5_dev_ctx_shared *sh = priv->sh;
        char s[MLX5_HLIST_NAMESIZE];
-       int err = 0;
+       int err;
 
-       if (!sh->flow_tbls)
-               err = mlx5_alloc_table_hash_list(priv);
-       else
-               DRV_LOG(DEBUG, "sh->flow_tbls[%p] already created, reuse\n",
-                       (void *)sh->flow_tbls);
+       MLX5_ASSERT(sh && sh->refcnt);
+       if (sh->refcnt > 1)
+               return 0;
+       err = mlx5_alloc_table_hash_list(priv);
        if (err)
                return err;
        /* Create tags hash list table. */
@@ -241,15 +242,24 @@ mlx5_alloc_shared_dr(struct mlx5_priv *priv)
                err = ENOMEM;
                goto error;
        }
+       snprintf(s, sizeof(s), "%s_hdr_modify", sh->ibdev_name);
+       sh->modify_cmds = mlx5_hlist_create(s, MLX5_FLOW_HDR_MODIFY_HTABLE_SZ);
+       if (!sh->modify_cmds) {
+               DRV_LOG(ERR, "hdr modify hash creation failed");
+               err = ENOMEM;
+               goto error;
+       }
+       snprintf(s, sizeof(s), "%s_encaps_decaps", sh->ibdev_name);
+       sh->encaps_decaps = mlx5_hlist_create(s,
+                                             MLX5_FLOW_ENCAP_DECAP_HTABLE_SZ);
+       if (!sh->encaps_decaps) {
+               DRV_LOG(ERR, "encap decap hash creation failed");
+               err = ENOMEM;
+               goto error;
+       }
 #ifdef HAVE_MLX5DV_DR
        void *domain;
 
-       if (sh->dv_refcnt) {
-               /* Shared DV/DR structures is already initialized. */
-               sh->dv_refcnt++;
-               priv->dr_shared = 1;
-               return 0;
-       }
        /* Reference counter is zero, we should initialize structures. */
        domain = mlx5_glue->dr_create_domain(sh->ctx,
                                             MLX5DV_DR_DOMAIN_TYPE_NIC_RX);
@@ -289,8 +299,6 @@ mlx5_alloc_shared_dr(struct mlx5_priv *priv)
        }
        sh->pop_vlan_action = mlx5_glue->dr_create_flow_action_pop_vlan();
 #endif /* HAVE_MLX5DV_DR */
-       sh->dv_refcnt++;
-       priv->dr_shared = 1;
        return 0;
 error:
        /* Rollback the created objects. */
@@ -314,6 +322,14 @@ error:
                mlx5_glue->destroy_flow_action(sh->pop_vlan_action);
                sh->pop_vlan_action = NULL;
        }
+       if (sh->encaps_decaps) {
+               mlx5_hlist_destroy(sh->encaps_decaps, NULL, NULL);
+               sh->encaps_decaps = NULL;
+       }
+       if (sh->modify_cmds) {
+               mlx5_hlist_destroy(sh->modify_cmds, NULL, NULL);
+               sh->modify_cmds = NULL;
+       }
        if (sh->tag_table) {
                /* tags should be destroyed with flow before. */
                mlx5_hlist_destroy(sh->tag_table, NULL, NULL);
@@ -332,17 +348,12 @@ error:
 void
 mlx5_os_free_shared_dr(struct mlx5_priv *priv)
 {
-       struct mlx5_dev_ctx_shared *sh;
+       struct mlx5_dev_ctx_shared *sh = priv->sh;
 
-       if (!priv->dr_shared)
+       MLX5_ASSERT(sh && sh->refcnt);
+       if (sh->refcnt > 1)
                return;
-       priv->dr_shared = 0;
-       sh = priv->sh;
-       MLX5_ASSERT(sh);
 #ifdef HAVE_MLX5DV_DR
-       MLX5_ASSERT(sh->dv_refcnt);
-       if (sh->dv_refcnt && --sh->dv_refcnt)
-               return;
        if (sh->rx_domain) {
                mlx5_glue->dr_destroy_domain(sh->rx_domain);
                sh->rx_domain = NULL;
@@ -367,6 +378,14 @@ mlx5_os_free_shared_dr(struct mlx5_priv *priv)
        }
        pthread_mutex_destroy(&sh->dv_mutex);
 #endif /* HAVE_MLX5DV_DR */
+       if (sh->encaps_decaps) {
+               mlx5_hlist_destroy(sh->encaps_decaps, NULL, NULL);
+               sh->encaps_decaps = NULL;
+       }
+       if (sh->modify_cmds) {
+               mlx5_hlist_destroy(sh->modify_cmds, NULL, NULL);
+               sh->modify_cmds = NULL;
+       }
        if (sh->tag_table) {
                /* tags should be destroyed with flow before. */
                mlx5_hlist_destroy(sh->tag_table, NULL, NULL);
@@ -478,6 +497,95 @@ out:
        return ret;
 }
 
+/**
+ * Create the Tx queue DevX/Verbs object.
+ *
+ * @param dev
+ *   Pointer to Ethernet device.
+ * @param idx
+ *   Queue index in DPDK Tx queue array.
+ *
+ * @return
+ *   0 on success, a negative errno value otherwise and rte_errno is set.
+ */
+static int
+mlx5_os_txq_obj_new(struct rte_eth_dev *dev, uint16_t idx)
+{
+       struct mlx5_priv *priv = dev->data->dev_private;
+       struct mlx5_txq_data *txq_data = (*priv->txqs)[idx];
+       struct mlx5_txq_ctrl *txq_ctrl =
+                       container_of(txq_data, struct mlx5_txq_ctrl, txq);
+
+       if (txq_ctrl->type == MLX5_TXQ_TYPE_HAIRPIN)
+               return mlx5_txq_devx_obj_new(dev, idx);
+#ifdef HAVE_MLX5DV_DEVX_UAR_OFFSET
+       if (!priv->config.dv_esw_en)
+               return mlx5_txq_devx_obj_new(dev, idx);
+#endif
+       return mlx5_txq_ibv_obj_new(dev, idx);
+}
+
+/**
+ * Release an Tx DevX/verbs queue object.
+ *
+ * @param txq_obj
+ *   DevX/Verbs Tx queue object.
+ */
+static void
+mlx5_os_txq_obj_release(struct mlx5_txq_obj *txq_obj)
+{
+       if (txq_obj->txq_ctrl->type == MLX5_TXQ_TYPE_HAIRPIN) {
+               mlx5_txq_devx_obj_release(txq_obj);
+               return;
+       }
+#ifdef HAVE_MLX5DV_DEVX_UAR_OFFSET
+       if (!txq_obj->txq_ctrl->priv->config.dv_esw_en) {
+               mlx5_txq_devx_obj_release(txq_obj);
+               return;
+       }
+#endif
+       mlx5_txq_ibv_obj_release(txq_obj);
+}
+
+/**
+ * DV flow counter mode detect and config.
+ *
+ * @param dev
+ *   Pointer to rte_eth_dev structure.
+ *
+ */
+static void
+mlx5_flow_counter_mode_config(struct rte_eth_dev *dev __rte_unused)
+{
+#ifdef HAVE_IBV_FLOW_DV_SUPPORT
+       struct mlx5_priv *priv = dev->data->dev_private;
+       struct mlx5_dev_ctx_shared *sh = priv->sh;
+       bool fallback;
+
+#ifndef HAVE_IBV_DEVX_ASYNC
+       fallback = true;
+#else
+       fallback = false;
+       if (!priv->config.devx || !priv->config.dv_flow_en ||
+           !priv->config.hca_attr.flow_counters_dump ||
+           !(priv->config.hca_attr.flow_counter_bulk_alloc_bitmap & 0x4) ||
+           (mlx5_flow_dv_discover_counter_offset_support(dev) == -ENOTSUP))
+               fallback = true;
+#endif
+       if (fallback)
+               DRV_LOG(INFO, "Use fall-back DV counter management. Flow "
+                       "counter dump:%d, bulk_alloc_bitmap:0x%hhx.",
+                       priv->config.hca_attr.flow_counters_dump,
+                       priv->config.hca_attr.flow_counter_bulk_alloc_bitmap);
+       /* Initialize fallback mode only on the port initializes sh. */
+       if (sh->refcnt == 1)
+               sh->cmng.counter_fallback = fallback;
+       else if (fallback != sh->cmng.counter_fallback)
+               DRV_LOG(WARNING, "Port %d in sh has different fallback mode "
+                       "with others:%d.", PORT_ID(priv), fallback);
+#endif
+}
+
 /**
  * Spawn an Ethernet device from Verbs information.
  *
@@ -585,6 +693,8 @@ mlx5_dev_spawn(struct rte_device *dpdk_dev,
                }
                eth_dev->device = dpdk_dev;
                eth_dev->dev_ops = &mlx5_os_dev_sec_ops;
+               eth_dev->rx_descriptor_status = mlx5_rx_descriptor_status;
+               eth_dev->tx_descriptor_status = mlx5_tx_descriptor_status;
                err = mlx5_proc_priv_init(eth_dev);
                if (err)
                        return NULL;
@@ -944,19 +1054,16 @@ err_secondary:
                DRV_LOG(INFO, "Rx CQE padding is enabled");
        }
        if (config->devx) {
-               priv->counter_fallback = 0;
                err = mlx5_devx_cmd_query_hca_attr(sh->ctx, &config->hca_attr);
                if (err) {
                        err = -err;
                        goto error;
                }
-               if (!config->hca_attr.flow_counters_dump)
-                       priv->counter_fallback = 1;
-#ifndef HAVE_IBV_DEVX_ASYNC
-               priv->counter_fallback = 1;
-#endif
-               if (priv->counter_fallback)
-                       DRV_LOG(INFO, "Use fall-back DV counter management");
+               /* Check relax ordering support. */
+               if (config->hca_attr.relaxed_ordering_write &&
+                   config->hca_attr.relaxed_ordering_read  &&
+                   !haswell_broadwell_cpu)
+                       sh->cmng.relaxed_ordering = 1;
                /* Check for LRO support. */
                if (config->dest_tir && config->hca_attr.lro_cap &&
                    config->dv_flow_en) {
@@ -972,6 +1079,9 @@ err_secondary:
                                config->hca_attr.lro_timer_supported_periods[0];
                        DRV_LOG(DEBUG, "LRO session timeout set to %d usec",
                                config->lro.timeout);
+                       DRV_LOG(DEBUG, "LRO minimal size of TCP segment "
+                               "required for coalescing is %d bytes",
+                               config->hca_attr.lro_min_mss_size);
                }
 #if defined(HAVE_MLX5DV_DR) && defined(HAVE_MLX5_DR_CREATE_ACTION_FLOW_METER)
                if (config->hca_attr.qos.sup &&
@@ -999,6 +1109,20 @@ err_secondary:
                                        priv->mtr_color_reg);
                        }
                }
+#endif
+#if defined(HAVE_MLX5DV_DR) && defined(HAVE_MLX5_DR_CREATE_ACTION_FLOW_SAMPLE)
+               if (config->hca_attr.log_max_ft_sampler_num > 0  &&
+                   config->dv_flow_en) {
+                       priv->sampler_en = 1;
+                       DRV_LOG(DEBUG, "The Sampler enabled!\n");
+               } else {
+                       priv->sampler_en = 0;
+                       if (!config->hca_attr.log_max_ft_sampler_num)
+                               DRV_LOG(WARNING, "No available register for"
+                                               " Sampler.");
+                       else
+                               DRV_LOG(DEBUG, "DV flow is not supported!\n");
+               }
 #endif
        }
        if (config->tx_pp) {
@@ -1136,8 +1260,6 @@ err_secondary:
                err = ENOMEM;
                goto error;
        }
-       /* Flag to call rte_eth_dev_release_port() in rte_eth_dev_close(). */
-       eth_dev->data->dev_flags |= RTE_ETH_DEV_CLOSE_REMOVE;
        if (priv->representor) {
                eth_dev->data->dev_flags |= RTE_ETH_DEV_REPRESENTOR;
                eth_dev->data->representor_id = priv->representor_id;
@@ -1149,10 +1271,24 @@ err_secondary:
         */
        MLX5_ASSERT(spawn->ifindex);
        priv->if_index = spawn->ifindex;
+       if (priv->pf_bond >= 0 && priv->master) {
+               /* Get bond interface info */
+               err = mlx5_sysfs_bond_info(priv->if_index,
+                                    &priv->bond_ifindex,
+                                    priv->bond_name);
+               if (err)
+                       DRV_LOG(ERR, "unable to get bond info: %s",
+                               strerror(rte_errno));
+               else
+                       DRV_LOG(INFO, "PF device %u, bond device %u(%s)",
+                               priv->if_index, priv->bond_ifindex,
+                               priv->bond_name);
+       }
        eth_dev->data->dev_private = priv;
        priv->dev_data = eth_dev->data;
        eth_dev->data->mac_addrs = priv->mac;
        eth_dev->device = dpdk_dev;
+       eth_dev->data->dev_flags |= RTE_ETH_DEV_AUTOFILL_QUEUE_XSTATS;
        /* Configure the first MAC address by default. */
        if (mlx5_get_mac(eth_dev, &mac.addr_bytes)) {
                DRV_LOG(ERR,
@@ -1192,6 +1328,9 @@ err_secondary:
        eth_dev->rx_pkt_burst = removed_rx_burst;
        eth_dev->tx_pkt_burst = removed_tx_burst;
        eth_dev->dev_ops = &mlx5_os_dev_ops;
+       eth_dev->rx_descriptor_status = mlx5_rx_descriptor_status;
+       eth_dev->tx_descriptor_status = mlx5_tx_descriptor_status;
+       eth_dev->rx_queue_count = mlx5_rx_queue_count;
        /* Register MAC address. */
        claim_zero(mlx5_mac_addr_add(eth_dev, &mac, 0, 0));
        if (config->vf && config->vf_nl_en)
@@ -1250,6 +1389,26 @@ err_secondary:
                        goto error;
                }
        }
+       if (config->devx && config->dv_flow_en && config->dest_tir) {
+               priv->obj_ops = devx_obj_ops;
+               priv->obj_ops.drop_action_create =
+                                               ibv_obj_ops.drop_action_create;
+               priv->obj_ops.drop_action_destroy =
+                                               ibv_obj_ops.drop_action_destroy;
+#ifndef HAVE_MLX5DV_DEVX_UAR_OFFSET
+               priv->obj_ops.txq_obj_modify = ibv_obj_ops.txq_obj_modify;
+#else
+               if (config->dv_esw_en)
+                       priv->obj_ops.txq_obj_modify =
+                                               ibv_obj_ops.txq_obj_modify;
+#endif
+               /* Use specific wrappers for Tx object. */
+               priv->obj_ops.txq_obj_new = mlx5_os_txq_obj_new;
+               priv->obj_ops.txq_obj_release = mlx5_os_txq_obj_release;
+
+       } else {
+               priv->obj_ops = ibv_obj_ops;
+       }
        /* Supported Verbs flow priority number detection. */
        err = mlx5_flow_discover_priorities(eth_dev);
        if (err < 0) {
@@ -1306,6 +1465,7 @@ err_secondary:
                        goto error;
                }
        }
+       mlx5_flow_counter_mode_config(eth_dev);
        return eth_dev;
 error:
        if (priv) {
@@ -2317,6 +2477,23 @@ mlx5_os_set_allmulti(struct rte_eth_dev *dev, int enable)
                                mlx5_ifindex(dev), !!enable);
 }
 
+/**
+ * Flush device MAC addresses
+ *
+ * @param dev
+ *   Pointer to Ethernet device structure.
+ *
+ */
+void
+mlx5_os_mac_addr_flush(struct rte_eth_dev *dev)
+{
+       struct mlx5_priv *priv = dev->data->dev_private;
+
+       mlx5_nl_mac_addr_flush(priv->nl_socket_route, mlx5_ifindex(dev),
+                              dev->data->mac_addrs,
+                              MLX5_MAX_MAC_ADDRESSES, priv->mac_own);
+}
+
 const struct eth_dev_ops mlx5_os_dev_ops = {
        .dev_configure = mlx5_dev_configure,
        .dev_start = mlx5_dev_start,
@@ -2363,13 +2540,10 @@ const struct eth_dev_ops mlx5_os_dev_ops = {
        .rss_hash_update = mlx5_rss_hash_update,
        .rss_hash_conf_get = mlx5_rss_hash_conf_get,
        .filter_ctrl = mlx5_dev_filter_ctrl,
-       .rx_descriptor_status = mlx5_rx_descriptor_status,
-       .tx_descriptor_status = mlx5_tx_descriptor_status,
        .rxq_info_get = mlx5_rxq_info_get,
        .txq_info_get = mlx5_txq_info_get,
        .rx_burst_mode_get = mlx5_rx_burst_mode_get,
        .tx_burst_mode_get = mlx5_tx_burst_mode_get,
-       .rx_queue_count = mlx5_rx_queue_count,
        .rx_queue_intr_enable = mlx5_rx_intr_enable,
        .rx_queue_intr_disable = mlx5_rx_intr_disable,
        .is_removed = mlx5_is_removed,
@@ -2394,8 +2568,6 @@ const struct eth_dev_ops mlx5_os_dev_sec_ops = {
        .rx_queue_stop = mlx5_rx_queue_stop,
        .tx_queue_start = mlx5_tx_queue_start,
        .tx_queue_stop = mlx5_tx_queue_stop,
-       .rx_descriptor_status = mlx5_rx_descriptor_status,
-       .tx_descriptor_status = mlx5_tx_descriptor_status,
        .rxq_info_get = mlx5_rxq_info_get,
        .txq_info_get = mlx5_txq_info_get,
        .rx_burst_mode_get = mlx5_rx_burst_mode_get,
@@ -2447,8 +2619,6 @@ const struct eth_dev_ops mlx5_os_dev_ops_isolate = {
        .vlan_strip_queue_set = mlx5_vlan_strip_queue_set,
        .vlan_offload_set = mlx5_vlan_offload_set,
        .filter_ctrl = mlx5_dev_filter_ctrl,
-       .rx_descriptor_status = mlx5_rx_descriptor_status,
-       .tx_descriptor_status = mlx5_tx_descriptor_status,
        .rxq_info_get = mlx5_rxq_info_get,
        .txq_info_get = mlx5_txq_info_get,
        .rx_burst_mode_get = mlx5_rx_burst_mode_get,