common/mlx5: add Direct Verbs constants for Windows
[dpdk.git] / drivers / net / bnxt / tf_core / tf_rm.c
index 50f6b1e..dd537aa 100644 (file)
@@ -37,6 +37,12 @@ struct tf_rm_element {
         */
        uint16_t hcapi_type;
 
+       /**
+        * Resource slices.  How many slices will fit in the
+        * resource pool chunk size.
+        */
+       uint8_t slices;
+
        /**
         * HCAPI RM allocated range information for the element.
         */
@@ -221,6 +227,7 @@ tf_rm_adjust_index(struct tf_rm_element *db,
  *   the DB in which this function is used. Each entry holds residual
  *   value for that entry.
  */
+#if (TF_RM_DEBUG == 1)
 static void
 tf_rm_log_residuals(enum tf_dir dir,
                    enum tf_module_type module,
@@ -234,14 +241,14 @@ tf_rm_log_residuals(enum tf_dir dir,
         */
        for (i = 0; i < count; i++) {
                if (residuals[i] != 0)
-                       TFP_DRV_LOG(ERR,
+                       TFP_DRV_LOG(INFO,
                                "%s, %s was not cleaned up, %d outstanding\n",
                                tf_dir_2_str(dir),
                                tf_module_subtype_2_str(module, i),
                                residuals[i]);
        }
 }
-
+#endif /* TF_RM_DEBUG == 1 */
 /**
  * Performs a check of the passed in DB for any lingering elements. If
  * a resource type was found to not have been cleaned up by the caller
@@ -357,11 +364,12 @@ tf_rm_check_residuals(struct tf_rm_new_db *rm_db,
                *resv_size = found;
        }
 
+#if (TF_RM_DEBUG == 1)
        tf_rm_log_residuals(rm_db->dir,
                            rm_db->module,
                            rm_db->num_entries,
                            residuals);
-
+#endif
        tfp_free((void *)residuals);
        *resv = local_resv;
 
@@ -402,12 +410,16 @@ tf_rm_check_residuals(struct tf_rm_new_db *rm_db,
  *     -          - Failure if negative
  */
 static int
-tf_rm_update_parent_reservations(struct tf_rm_element_cfg *cfg,
+tf_rm_update_parent_reservations(struct tf *tfp,
+                                struct tf_dev_info *dev,
+                                struct tf_rm_element_cfg *cfg,
                                 uint16_t *alloc_cnt,
                                 uint16_t num_elements,
-                                uint16_t *req_cnt)
+                                uint16_t *req_cnt,
+                                bool shared_session)
 {
        int parent, child;
+       const char *type_str;
 
        /* Search through all the elements */
        for (parent = 0; parent < num_elements; parent++) {
@@ -415,31 +427,62 @@ tf_rm_update_parent_reservations(struct tf_rm_element_cfg *cfg,
 
                /* If I am a parent */
                if (cfg[parent].cfg_type == TF_RM_ELEM_CFG_HCAPI_BA_PARENT) {
-                       /* start with my own count */
-                       RTE_ASSERT(cfg[parent].slices);
-                       combined_cnt =
-                               alloc_cnt[parent] / cfg[parent].slices;
+                       uint8_t p_slices = 1;
+
+                       /* Shared session doesn't support slices */
+                       if (!shared_session)
+                               p_slices = cfg[parent].slices;
+
+                       RTE_ASSERT(p_slices);
+
+                       combined_cnt = alloc_cnt[parent] / p_slices;
 
-                       if (alloc_cnt[parent] % cfg[parent].slices)
+                       if (alloc_cnt[parent] % p_slices)
                                combined_cnt++;
 
+                       if (alloc_cnt[parent]) {
+                               dev->ops->tf_dev_get_resource_str(tfp,
+                                                        cfg[parent].hcapi_type,
+                                                        &type_str);
+#if (TF_RM_DEBUG == 1)
+                               printf("%s:%s cnt(%d) slices(%d)\n",
+                                      type_str, tf_tbl_type_2_str(parent),
+                                      alloc_cnt[parent], p_slices);
+#endif /* (TF_RM_DEBUG == 1) */
+                       }
+
                        /* Search again through all the elements */
                        for (child = 0; child < num_elements; child++) {
                                /* If this is one of my children */
                                if (cfg[child].cfg_type ==
                                    TF_RM_ELEM_CFG_HCAPI_BA_CHILD &&
-                                   cfg[child].parent_subtype == parent) {
+                                   cfg[child].parent_subtype == parent &&
+                                   alloc_cnt[child]) {
+                                       uint8_t c_slices = 1;
                                        uint16_t cnt = 0;
-                                       RTE_ASSERT(cfg[child].slices);
 
+                                       if (!shared_session)
+                                               c_slices = cfg[child].slices;
+
+                                       RTE_ASSERT(c_slices);
+
+                                       dev->ops->tf_dev_get_resource_str(tfp,
+                                                         cfg[child].hcapi_type,
+                                                          &type_str);
+#if (TF_RM_DEBUG == 1)
+                                       printf("%s:%s cnt(%d) slices(%d)\n",
+                                              type_str,
+                                              tf_tbl_type_2_str(child),
+                                              alloc_cnt[child],
+                                              c_slices);
+#endif /* (TF_RM_DEBUG == 1) */
                                        /* Increment the parents combined count
                                         * with each child's count adjusted for
-                                        * number of slices per RM allocated item.
+                                        * number of slices per RM alloc item.
                                         */
-                                       cnt =
-                                        alloc_cnt[child] / cfg[child].slices;
+                                       cnt = alloc_cnt[child] / c_slices;
 
-                                       if (alloc_cnt[child] % cfg[child].slices)
+                                       if (alloc_cnt[child] % c_slices)
                                                cnt++;
 
                                        combined_cnt += cnt;
@@ -449,6 +492,10 @@ tf_rm_update_parent_reservations(struct tf_rm_element_cfg *cfg,
                        }
                        /* Save the parent count to be requested */
                        req_cnt[parent] = combined_cnt;
+#if (TF_RM_DEBUG == 1)
+                       printf("%s calculated total:%d\n\n",
+                              type_str, req_cnt[parent]);
+#endif /* (TF_RM_DEBUG == 1) */
                }
        }
        return 0;
@@ -471,6 +518,7 @@ tf_rm_create_db(struct tf *tfp,
        struct tf_rm_new_db *rm_db;
        struct tf_rm_element *db;
        uint32_t pool_size;
+       bool shared_session = 0;
 
        TF_CHECK_PARMS2(tfp, parms);
 
@@ -487,7 +535,6 @@ tf_rm_create_db(struct tf *tfp,
        /* Need device max number of elements for the RM QCAPS */
        rc = dev->ops->tf_dev_get_max_types(tfp, &max_types);
 
-
        /* Allocate memory for RM QCAPS request */
        cparms.nitems = max_types;
        cparms.size = sizeof(struct tf_rm_resc_req_entry);
@@ -523,12 +570,15 @@ tf_rm_create_db(struct tf *tfp,
        tfp_memcpy(req_cnt, parms->alloc_cnt,
                   parms->num_elements * sizeof(uint16_t));
 
+       shared_session = tf_session_is_shared_session(tfs);
+
        /* Update the req_cnt based upon the element configuration
         */
-       tf_rm_update_parent_reservations(parms->cfg,
+       tf_rm_update_parent_reservations(tfp, dev, parms->cfg,
                                         parms->alloc_cnt,
                                         parms->num_elements,
-                                        req_cnt);
+                                        req_cnt,
+                                        shared_session);
 
        /* Process capabilities against DB requirements. However, as a
         * DB can hold elements that are not HCAPI we can reduce the
@@ -544,11 +594,12 @@ tf_rm_create_db(struct tf *tfp,
                                       &hcapi_items);
 
        if (hcapi_items == 0) {
-               TFP_DRV_LOG(ERR,
-                           "%s: module:%s Empty RM DB create request\n",
-                           tf_dir_2_str(parms->dir),
-                           tf_module_2_str(parms->module));
-
+#if (TF_RM_DEBUG == 1)
+               TFP_DRV_LOG(INFO,
+                       "%s: module: %s Empty RM DB create request\n",
+                       tf_dir_2_str(parms->dir),
+                       tf_module_2_str(parms->module));
+#endif
                parms->rm_db = NULL;
                return -ENOMEM;
        }
@@ -597,11 +648,11 @@ tf_rm_create_db(struct tf *tfp,
                                                              hcapi_type,
                                                              &type_str);
                                TFP_DRV_LOG(ERR,
-                                           "Failure, %s:%d:%s req:%d avail:%d\n",
-                                           tf_dir_2_str(parms->dir),
-                                           hcapi_type, type_str,
-                                           req_cnt[i],
-                                           query[hcapi_type].max);
+                                       "Failure, %s:%d:%s req:%d avail:%d\n",
+                                       tf_dir_2_str(parms->dir),
+                                       hcapi_type, type_str,
+                                       req_cnt[i],
+                                       query[hcapi_type].max);
                                return -EINVAL;
                        }
                }
@@ -645,6 +696,7 @@ tf_rm_create_db(struct tf *tfp,
 
                db[i].cfg_type = cfg->cfg_type;
                db[i].hcapi_type = cfg->hcapi_type;
+               db[i].slices = cfg->slices;
 
                /* Save the parent subtype for later use to find the pool
                 */
@@ -675,17 +727,6 @@ tf_rm_create_db(struct tf *tfp,
                        /* Only allocate BA pool if a BA type not a child */
                        if (cfg->cfg_type == TF_RM_ELEM_CFG_HCAPI_BA ||
                            cfg->cfg_type == TF_RM_ELEM_CFG_HCAPI_BA_PARENT) {
-                               if (cfg->divider) {
-                                       resv[j].stride =
-                                               resv[j].stride / cfg->divider;
-                                       if (resv[j].stride <= 0) {
-                                               TFP_DRV_LOG(ERR,
-                                                    "%s:Divide fails:%d:%s\n",
-                                                    tf_dir_2_str(parms->dir),
-                                                    cfg->hcapi_type, type_str);
-                                               goto fail;
-                                       }
-                               }
                                /* Create pool */
                                pool_size = (BITALLOC_SIZEOF(resv[j].stride) /
                                             sizeof(struct bitalloc));
@@ -702,7 +743,9 @@ tf_rm_create_db(struct tf *tfp,
                                }
                                db[i].pool = (struct bitalloc *)cparms.mem_va;
 
-                               rc = ba_init(db[i].pool, resv[j].stride);
+                               rc = ba_init(db[i].pool,
+                                            resv[j].stride,
+                                            !tf_session_is_shared_session(tfs));
                                if (rc) {
                                        TFP_DRV_LOG(ERR,
                                          "%s: Pool init failed, type:%d:%s\n",
@@ -729,6 +772,13 @@ tf_rm_create_db(struct tf *tfp,
        rm_db->module = parms->module;
        *parms->rm_db = (void *)rm_db;
 
+#if (TF_RM_DEBUG == 1)
+
+       printf("%s: module:%s\n",
+              tf_dir_2_str(parms->dir),
+              tf_module_2_str(parms->module));
+#endif /* (TF_RM_DEBUG == 1) */
+
        tfp_free((void *)req);
        tfp_free((void *)resv);
        tfp_free((void *)req_cnt);
@@ -746,6 +796,245 @@ tf_rm_create_db(struct tf *tfp,
        return -EINVAL;
 }
 
+int
+tf_rm_create_db_no_reservation(struct tf *tfp,
+                              struct tf_rm_create_db_parms *parms)
+{
+       int rc;
+       struct tf_session *tfs;
+       struct tf_dev_info *dev;
+       int i, j;
+       uint16_t hcapi_items, *req_cnt;
+       struct tfp_calloc_parms cparms;
+       struct tf_rm_resc_req_entry *req;
+       struct tf_rm_resc_entry *resv;
+       struct tf_rm_new_db *rm_db;
+       struct tf_rm_element *db;
+       uint32_t pool_size;
+
+       TF_CHECK_PARMS2(tfp, parms);
+
+       /* Retrieve the session information */
+       rc = tf_session_get_session_internal(tfp, &tfs);
+       if (rc)
+               return rc;
+
+       /* Retrieve device information */
+       rc = tf_session_get_device(tfs, &dev);
+       if (rc)
+               return rc;
+
+       /* Copy requested counts (alloc_cnt) from tf_open_session() to local
+        * copy (req_cnt) so that it can be updated if required.
+        */
+
+       cparms.nitems = parms->num_elements;
+       cparms.size = sizeof(uint16_t);
+       cparms.alignment = 0;
+       rc = tfp_calloc(&cparms);
+       if (rc)
+               return rc;
+
+       req_cnt = (uint16_t *)cparms.mem_va;
+
+       tfp_memcpy(req_cnt, parms->alloc_cnt,
+                  parms->num_elements * sizeof(uint16_t));
+
+       /* Process capabilities against DB requirements. However, as a
+        * DB can hold elements that are not HCAPI we can reduce the
+        * req msg content by removing those out of the request yet
+        * the DB holds them all as to give a fast lookup. We can also
+        * remove entries where there are no request for elements.
+        */
+       tf_rm_count_hcapi_reservations(parms->dir,
+                                      parms->module,
+                                      parms->cfg,
+                                      req_cnt,
+                                      parms->num_elements,
+                                      &hcapi_items);
+
+       if (hcapi_items == 0) {
+               TFP_DRV_LOG(ERR,
+                       "%s: module:%s Empty RM DB create request\n",
+                       tf_dir_2_str(parms->dir),
+                       tf_module_2_str(parms->module));
+
+               parms->rm_db = NULL;
+               return -ENOMEM;
+       }
+
+       /* Alloc request, alignment already set */
+       cparms.nitems = (size_t)hcapi_items;
+       cparms.size = sizeof(struct tf_rm_resc_req_entry);
+       rc = tfp_calloc(&cparms);
+       if (rc)
+               return rc;
+       req = (struct tf_rm_resc_req_entry *)cparms.mem_va;
+
+       /* Alloc reservation, alignment and nitems already set */
+       cparms.size = sizeof(struct tf_rm_resc_entry);
+       rc = tfp_calloc(&cparms);
+       if (rc)
+               return rc;
+       resv = (struct tf_rm_resc_entry *)cparms.mem_va;
+
+       /* Build the request */
+       for (i = 0, j = 0; i < parms->num_elements; i++) {
+               struct tf_rm_element_cfg *cfg = &parms->cfg[i];
+               uint16_t hcapi_type = cfg->hcapi_type;
+
+               /* Only perform reservation for requested entries
+                */
+               if (req_cnt[i] == 0)
+                       continue;
+
+               /* Skip any children in the request */
+               if (cfg->cfg_type == TF_RM_ELEM_CFG_HCAPI ||
+                   cfg->cfg_type == TF_RM_ELEM_CFG_HCAPI_BA ||
+                   cfg->cfg_type == TF_RM_ELEM_CFG_HCAPI_BA_PARENT) {
+                       req[j].type = hcapi_type;
+                       req[j].min = req_cnt[i];
+                       req[j].max = req_cnt[i];
+                       j++;
+               }
+       }
+
+       /* Get all resources info for the module type
+        */
+       rc = tf_msg_session_resc_info(tfp,
+                                     dev,
+                                     parms->dir,
+                                     hcapi_items,
+                                     req,
+                                     resv);
+       if (rc)
+               return rc;
+
+       /* Build the RM DB per the request */
+       cparms.nitems = 1;
+       cparms.size = sizeof(struct tf_rm_new_db);
+       rc = tfp_calloc(&cparms);
+       if (rc)
+               return rc;
+       rm_db = (void *)cparms.mem_va;
+
+       /* Build the DB within RM DB */
+       cparms.nitems = parms->num_elements;
+       cparms.size = sizeof(struct tf_rm_element);
+       rc = tfp_calloc(&cparms);
+       if (rc)
+               return rc;
+       rm_db->db = (struct tf_rm_element *)cparms.mem_va;
+
+       db = rm_db->db;
+       for (i = 0, j = 0; i < parms->num_elements; i++) {
+               struct tf_rm_element_cfg *cfg = &parms->cfg[i];
+               const char *type_str;
+
+               dev->ops->tf_dev_get_resource_str(tfp,
+                                                 cfg->hcapi_type,
+                                                 &type_str);
+
+               db[i].cfg_type = cfg->cfg_type;
+               db[i].hcapi_type = cfg->hcapi_type;
+
+               /* Save the parent subtype for later use to find the pool
+                */
+               if (cfg->cfg_type == TF_RM_ELEM_CFG_HCAPI_BA_CHILD)
+                       db[i].parent_subtype = cfg->parent_subtype;
+
+               /* If the element didn't request an allocation no need
+                * to create a pool nor verify if we got a reservation.
+                */
+               if (req_cnt[i] == 0)
+                       continue;
+
+               /* Skip any children or invalid
+                */
+               if (cfg->cfg_type != TF_RM_ELEM_CFG_HCAPI &&
+                   cfg->cfg_type != TF_RM_ELEM_CFG_HCAPI_BA &&
+                   cfg->cfg_type != TF_RM_ELEM_CFG_HCAPI_BA_PARENT)
+                       continue;
+
+               /* If the element had requested an allocation and that
+                * allocation was a success (full amount) then
+                * allocate the pool.
+                */
+               if (req_cnt[i] == resv[j].stride) {
+                       db[i].alloc.entry.start = resv[j].start;
+                       db[i].alloc.entry.stride = resv[j].stride;
+
+                       /* Only allocate BA pool if a BA type not a child */
+                       if (cfg->cfg_type == TF_RM_ELEM_CFG_HCAPI_BA ||
+                           cfg->cfg_type == TF_RM_ELEM_CFG_HCAPI_BA_PARENT) {
+                               /* Create pool */
+                               pool_size = (BITALLOC_SIZEOF(resv[j].stride) /
+                                            sizeof(struct bitalloc));
+                               /* Alloc request, alignment already set */
+                               cparms.nitems = pool_size;
+                               cparms.size = sizeof(struct bitalloc);
+                               rc = tfp_calloc(&cparms);
+                               if (rc) {
+                                       TFP_DRV_LOG(ERR,
+                                        "%s: Pool alloc failed, type:%d:%s\n",
+                                        tf_dir_2_str(parms->dir),
+                                        cfg->hcapi_type, type_str);
+                                       goto fail;
+                               }
+                               db[i].pool = (struct bitalloc *)cparms.mem_va;
+
+                               rc = ba_init(db[i].pool,
+                                            resv[j].stride,
+                                            !tf_session_is_shared_session(tfs));
+                               if (rc) {
+                                       TFP_DRV_LOG(ERR,
+                                         "%s: Pool init failed, type:%d:%s\n",
+                                         tf_dir_2_str(parms->dir),
+                                         cfg->hcapi_type, type_str);
+                                       goto fail;
+                               }
+                       }
+                       j++;
+               } else {
+                       /* Bail out as we want what we requested for
+                        * all elements, not any less.
+                        */
+                       TFP_DRV_LOG(ERR,
+                                   "%s: Alloc failed %d:%s req:%d, alloc:%d\n",
+                                   tf_dir_2_str(parms->dir), cfg->hcapi_type,
+                                   type_str, req_cnt[i], resv[j].stride);
+                       goto fail;
+               }
+       }
+
+       rm_db->num_entries = parms->num_elements;
+       rm_db->dir = parms->dir;
+       rm_db->module = parms->module;
+       *parms->rm_db = (void *)rm_db;
+
+#if (TF_RM_DEBUG == 1)
+
+       printf("%s: module:%s\n",
+              tf_dir_2_str(parms->dir),
+              tf_module_2_str(parms->module));
+#endif /* (TF_RM_DEBUG == 1) */
+
+       tfp_free((void *)req);
+       tfp_free((void *)resv);
+       tfp_free((void *)req_cnt);
+       return 0;
+
+ fail:
+       tfp_free((void *)req);
+       tfp_free((void *)resv);
+       tfp_free((void *)db->pool);
+       tfp_free((void *)db);
+       tfp_free((void *)rm_db);
+       tfp_free((void *)req_cnt);
+       parms->rm_db = NULL;
+
+       return -EINVAL;
+}
 int
 tf_rm_free_db(struct tf *tfp,
              struct tf_rm_free_db_parms *parms)
@@ -993,7 +1282,6 @@ tf_rm_is_allocated(struct tf_rm_is_allocated_parms *parms)
 
        cfg_type = rm_db->db[parms->subtype].cfg_type;
 
-
        /* Bail out if not controlled by RM */
        if (cfg_type != TF_RM_ELEM_CFG_HCAPI_BA &&
            cfg_type != TF_RM_ELEM_CFG_HCAPI_BA_PARENT &&
@@ -1043,6 +1331,42 @@ tf_rm_get_info(struct tf_rm_get_alloc_info_parms *parms)
        return 0;
 }
 
+int
+tf_rm_get_all_info(struct tf_rm_get_alloc_info_parms *parms, int size)
+{
+       struct tf_rm_new_db *rm_db;
+       enum tf_rm_elem_cfg_type cfg_type;
+       struct tf_rm_alloc_info *info = parms->info;
+       int i;
+
+       TF_CHECK_PARMS1(parms);
+
+       /* No rm info available for this module type
+        */
+       if (!parms->rm_db)
+               return -ENOMEM;
+
+       rm_db = (struct tf_rm_new_db *)parms->rm_db;
+       TF_CHECK_PARMS1(rm_db->db);
+
+       for (i = 0; i < size; i++) {
+               cfg_type = rm_db->db[i].cfg_type;
+
+               /* Bail out if not controlled by HCAPI */
+               if (cfg_type == TF_RM_ELEM_CFG_NULL) {
+                       info++;
+                       continue;
+               }
+
+               memcpy(info,
+                      &rm_db->db[i].alloc,
+                      sizeof(struct tf_rm_alloc_info));
+               info++;
+       }
+
+       return 0;
+}
+
 int
 tf_rm_get_hcapi_type(struct tf_rm_get_hcapi_parms *parms)
 {
@@ -1063,6 +1387,26 @@ tf_rm_get_hcapi_type(struct tf_rm_get_hcapi_parms *parms)
 
        return 0;
 }
+int
+tf_rm_get_slices(struct tf_rm_get_slices_parms *parms)
+{
+       struct tf_rm_new_db *rm_db;
+       enum tf_rm_elem_cfg_type cfg_type;
+
+       TF_CHECK_PARMS2(parms, parms->rm_db);
+       rm_db = (struct tf_rm_new_db *)parms->rm_db;
+       TF_CHECK_PARMS1(rm_db->db);
+
+       cfg_type = rm_db->db[parms->subtype].cfg_type;
+
+       /* Bail out if not controlled by HCAPI */
+       if (cfg_type == TF_RM_ELEM_CFG_NULL)
+               return -ENOTSUP;
+
+       *parms->slices = rm_db->db[parms->subtype].slices;
+
+       return 0;
+}
 
 int
 tf_rm_get_inuse_count(struct tf_rm_get_inuse_count_parms *parms)