/* SPDX-License-Identifier: BSD-3-Clause
- * Copyright(c) 2014-2020 Broadcom
+ * Copyright(c) 2014-2021 Broadcom
* All rights reserved.
*/
#include <rte_common.h>
+#include <rte_cycles.h>
#include <rte_malloc.h>
#include <rte_log.h>
#include <rte_alarm.h>
#include "bnxt_ulp.h"
#include "bnxt_tf_common.h"
#include "ulp_fc_mgr.h"
+#include "ulp_flow_db.h"
#include "ulp_template_db_enum.h"
#include "ulp_template_struct.h"
#include "tf_tbl.h"
ulp_fc_mgr_shadow_mem_alloc(struct hw_fc_mem_info *parms, int size)
{
/* Allocate memory*/
- if (parms == NULL)
+ if (!parms)
return -EINVAL;
parms->mem_va = rte_zmalloc("ulp_fc_info",
RTE_CACHE_LINE_ROUNDUP(size),
4096);
- if (parms->mem_va == NULL) {
+ if (!parms->mem_va) {
BNXT_TF_DBG(ERR, "Allocate failed mem_va\n");
return -ENOMEM;
}
/* Add the FC info tbl to the ulp context. */
bnxt_ulp_cntxt_ptr2_fc_info_set(ctxt, ulp_fc_info);
+ ulp_fc_info->num_counters = dparms->flow_count_db_entries;
+ if (!ulp_fc_info->num_counters) {
+ /* No need for software counters, call fw directly */
+ BNXT_TF_DBG(DEBUG, "Sw flow counter support not enabled\n");
+ return 0;
+ }
+
sw_acc_cntr_tbl_sz = sizeof(struct sw_acc_counter) *
dparms->flow_count_db_entries;
ulp_fc_mgr_deinit(struct bnxt_ulp_context *ctxt)
{
struct bnxt_ulp_fc_info *ulp_fc_info;
+ struct hw_fc_mem_info *shd_info;
int i;
ulp_fc_info = bnxt_ulp_cntxt_ptr2_fc_info_get(ctxt);
pthread_mutex_destroy(&ulp_fc_info->fc_lock);
- for (i = 0; i < TF_DIR_MAX; i++)
- rte_free(ulp_fc_info->sw_acc_tbl[i]);
-
- for (i = 0; i < TF_DIR_MAX; i++)
- ulp_fc_mgr_shadow_mem_free(&ulp_fc_info->shadow_hw_tbl[i]);
+ if (ulp_fc_info->num_counters) {
+ for (i = 0; i < TF_DIR_MAX; i++)
+ rte_free(ulp_fc_info->sw_acc_tbl[i]);
+ for (i = 0; i < TF_DIR_MAX; i++) {
+ shd_info = &ulp_fc_info->shadow_hw_tbl[i];
+ ulp_fc_mgr_shadow_mem_free(shd_info);
+ }
+ }
rte_free(ulp_fc_info);
ulp_fc_info = bnxt_ulp_cntxt_ptr2_fc_info_get(ctxt);
- return !!(ulp_fc_info->flags & ULP_FLAG_FC_THREAD);
+ if (ulp_fc_info)
+ return !!(ulp_fc_info->flags & ULP_FLAG_FC_THREAD);
+
+ return false;
}
/*
ulp_fc_info = bnxt_ulp_cntxt_ptr2_fc_info_get(ctxt);
- if (!(ulp_fc_info->flags & ULP_FLAG_FC_THREAD)) {
+ if (ulp_fc_info && !(ulp_fc_info->flags & ULP_FLAG_FC_THREAD)) {
rte_eal_alarm_set(US_PER_S * ULP_FC_TIMER,
- ulp_fc_mgr_alarm_cb,
- (void *)ctxt);
+ ulp_fc_mgr_alarm_cb, NULL);
ulp_fc_info->flags |= ULP_FLAG_FC_THREAD;
}
return;
ulp_fc_info->flags &= ~ULP_FLAG_FC_THREAD;
- rte_eal_alarm_cancel(ulp_fc_mgr_alarm_cb, (void *)ctxt);
+ rte_eal_alarm_cancel(ulp_fc_mgr_alarm_cb, NULL);
}
/*
* num_counters [in] The number of counters
*
*/
-static int32_t ulp_bulk_get_flow_stats(struct tf *tfp,
- struct bnxt_ulp_fc_info *fc_info,
- enum tf_dir dir, uint32_t num_counters)
+__rte_unused static int32_t
+ulp_bulk_get_flow_stats(struct tf *tfp,
+ struct bnxt_ulp_fc_info *fc_info,
+ enum tf_dir dir,
+ struct bnxt_ulp_device_params *dparms)
+/* MARK AS UNUSED FOR NOW TO AVOID COMPILATION ERRORS TILL API is RESOLVED */
{
int rc = 0;
struct tf_tbl_get_bulk_parms parms = { 0 };
parms.dir = dir;
parms.type = stype;
parms.starting_idx = fc_info->shadow_hw_tbl[dir].start_idx;
- parms.num_entries = num_counters;
+ parms.num_entries = dparms->flow_count_db_entries / 2; /* direction */
/*
* TODO:
* Size of an entry needs to obtained from template
*/
parms.entry_sz_in_bytes = sizeof(uint64_t);
stats = (uint64_t *)fc_info->shadow_hw_tbl[dir].mem_va;
- parms.physical_mem_addr = (uintptr_t)fc_info->shadow_hw_tbl[dir].mem_pa;
+ parms.physical_mem_addr = (uint64_t)
+ ((uintptr_t)(fc_info->shadow_hw_tbl[dir].mem_pa));
- if (stats == NULL) {
+ if (!stats) {
PMD_DRV_LOG(ERR,
"BULK: Memory not initialized id:0x%x dir:%d\n",
parms.starting_idx, dir);
return rc;
}
- for (i = 0; i < num_counters; i++) {
+ for (i = 0; i < parms.num_entries; i++) {
/* TBD - Get PKT/BYTE COUNT SHIFT/MASK from Template */
sw_acc_tbl_entry = &fc_info->sw_acc_tbl[dir][i];
if (!sw_acc_tbl_entry->valid)
continue;
- sw_acc_tbl_entry->pkt_count += FLOW_CNTR_PKTS(stats[i]);
- sw_acc_tbl_entry->byte_count += FLOW_CNTR_BYTES(stats[i]);
+ sw_acc_tbl_entry->pkt_count += FLOW_CNTR_PKTS(stats[i],
+ dparms);
+ sw_acc_tbl_entry->byte_count += FLOW_CNTR_BYTES(stats[i],
+ dparms);
+ }
+
+ return rc;
+}
+
+static int32_t
+ulp_fc_tf_flow_stat_get(struct bnxt_ulp_context *ctxt,
+ struct ulp_flow_db_res_params *res,
+ struct rte_flow_query_count *qcount)
+{
+ struct tf *tfp;
+ struct bnxt_ulp_device_params *dparms;
+ struct tf_get_tbl_entry_parms parms = { 0 };
+ struct tf_set_tbl_entry_parms sparms = { 0 };
+ enum tf_tbl_type stype = TF_TBL_TYPE_ACT_STATS_64;
+ uint64_t stats = 0;
+ uint32_t dev_id = 0;
+ int32_t rc = 0;
+
+ tfp = bnxt_ulp_cntxt_tfp_get(ctxt, BNXT_ULP_SHARED_SESSION_NO);
+ if (!tfp) {
+ BNXT_TF_DBG(ERR, "Failed to get the truflow pointer\n");
+ return -EINVAL;
+ }
+
+ if (bnxt_ulp_cntxt_dev_id_get(ctxt, &dev_id)) {
+ BNXT_TF_DBG(DEBUG, "Failed to get device id\n");
+ bnxt_ulp_cntxt_entry_release();
+ return -EINVAL;
+ }
+
+ dparms = bnxt_ulp_device_params_get(dev_id);
+ if (!dparms) {
+ BNXT_TF_DBG(DEBUG, "Failed to device parms\n");
+ bnxt_ulp_cntxt_entry_release();
+ return -EINVAL;
+ }
+ parms.dir = res->direction;
+ parms.type = stype;
+ parms.idx = res->resource_hndl;
+ parms.data_sz_in_bytes = sizeof(uint64_t);
+ parms.data = (uint8_t *)&stats;
+ rc = tf_get_tbl_entry(tfp, &parms);
+ if (rc) {
+ PMD_DRV_LOG(ERR,
+ "Get failed for id:0x%x rc:%d\n",
+ parms.idx, rc);
+ return rc;
+ }
+ qcount->hits = FLOW_CNTR_PKTS(stats, dparms);
+ if (qcount->hits)
+ qcount->hits_set = 1;
+ qcount->bytes = FLOW_CNTR_BYTES(stats, dparms);
+ if (qcount->bytes)
+ qcount->bytes_set = 1;
+
+ if (qcount->reset) {
+ stats = 0;
+ sparms.dir = res->direction;
+ sparms.type = stype;
+ sparms.idx = res->resource_hndl;
+ sparms.data = (uint8_t *)&stats;
+ sparms.data_sz_in_bytes = sizeof(uint64_t);
+ rc = tf_set_tbl_entry(tfp, &sparms);
+ if (rc) {
+ PMD_DRV_LOG(ERR, "Set failed for id:0x%x rc:%d\n",
+ sparms.idx, rc);
+ return rc;
+ }
+ }
+ return rc;
+}
+
+static int ulp_get_single_flow_stat(struct bnxt_ulp_context *ctxt,
+ struct tf *tfp,
+ struct bnxt_ulp_fc_info *fc_info,
+ enum tf_dir dir,
+ uint32_t hw_cntr_id,
+ struct bnxt_ulp_device_params *dparms)
+{
+ int rc = 0;
+ struct tf_get_tbl_entry_parms parms = { 0 };
+ enum tf_tbl_type stype = TF_TBL_TYPE_ACT_STATS_64; /* TBD:Template? */
+ struct sw_acc_counter *sw_acc_tbl_entry = NULL, *t_sw;
+ uint64_t stats = 0;
+ uint32_t sw_cntr_indx = 0;
+
+ parms.dir = dir;
+ parms.type = stype;
+ parms.idx = hw_cntr_id;
+ /*
+ * TODO:
+ * Size of an entry needs to obtained from template
+ */
+ parms.data_sz_in_bytes = sizeof(uint64_t);
+ parms.data = (uint8_t *)&stats;
+ rc = tf_get_tbl_entry(tfp, &parms);
+ if (rc) {
+ PMD_DRV_LOG(ERR,
+ "Get failed for id:0x%x rc:%d\n",
+ parms.idx, rc);
+ return rc;
+ }
+
+ /* TBD - Get PKT/BYTE COUNT SHIFT/MASK from Template */
+ sw_cntr_indx = hw_cntr_id - fc_info->shadow_hw_tbl[dir].start_idx;
+ sw_acc_tbl_entry = &fc_info->sw_acc_tbl[dir][sw_cntr_indx];
+ /* Some dpdk applications may accumulate the flow counters while some
+ * may not. In cases where the application is accumulating the counters
+ * the PMD need not do the accumulation itself and viceversa to report
+ * the correct flow counters.
+ */
+ if (ctxt->cfg_data->accum_stats) {
+ sw_acc_tbl_entry->pkt_count += FLOW_CNTR_PKTS(stats, dparms);
+ sw_acc_tbl_entry->byte_count += FLOW_CNTR_BYTES(stats, dparms);
+ } else {
+ sw_acc_tbl_entry->pkt_count = FLOW_CNTR_PKTS(stats, dparms);
+ sw_acc_tbl_entry->byte_count = FLOW_CNTR_BYTES(stats, dparms);
+ }
+
+ /* Update the parent counters if it is child flow */
+ if (sw_acc_tbl_entry->parent_flow_id) {
+ /* Update the parent counters */
+ t_sw = sw_acc_tbl_entry;
+ if (ulp_flow_db_parent_flow_count_update(ctxt,
+ t_sw->parent_flow_id,
+ t_sw->pkt_count,
+ t_sw->byte_count)) {
+ PMD_DRV_LOG(ERR, "Error updating parent counters\n");
+ }
}
return rc;
}
+
/*
* Alarm handler that will issue the TF-Core API to fetch
* data from the chip's internal flow counters
* ctxt [in] The ulp context for the flow counter manager
*
*/
+
void
-ulp_fc_mgr_alarm_cb(void *arg)
+ulp_fc_mgr_alarm_cb(void *arg __rte_unused)
{
- int rc = 0, i;
- struct bnxt_ulp_context *ctxt = arg;
+ int rc = 0;
+ unsigned int j;
+ enum tf_dir i;
+ struct bnxt_ulp_context *ctxt;
struct bnxt_ulp_fc_info *ulp_fc_info;
struct bnxt_ulp_device_params *dparms;
struct tf *tfp;
- uint32_t dev_id;
+ uint32_t dev_id, hw_cntr_id = 0, num_entries = 0;
+
+ ctxt = bnxt_ulp_cntxt_entry_acquire();
+ if (ctxt == NULL) {
+ BNXT_TF_DBG(INFO, "could not get the ulp context lock\n");
+ rte_eal_alarm_set(US_PER_S * ULP_FC_TIMER,
+ ulp_fc_mgr_alarm_cb, NULL);
+ return;
+ }
ulp_fc_info = bnxt_ulp_cntxt_ptr2_fc_info_get(ctxt);
- if (!ulp_fc_info)
+ if (!ulp_fc_info) {
+ bnxt_ulp_cntxt_entry_release();
return;
+ }
if (bnxt_ulp_cntxt_dev_id_get(ctxt, &dev_id)) {
BNXT_TF_DBG(DEBUG, "Failed to get device id\n");
+ bnxt_ulp_cntxt_entry_release();
return;
}
dparms = bnxt_ulp_device_params_get(dev_id);
if (!dparms) {
BNXT_TF_DBG(DEBUG, "Failed to device parms\n");
+ bnxt_ulp_cntxt_entry_release();
return;
}
- tfp = bnxt_ulp_cntxt_tfp_get(ctxt);
+ tfp = bnxt_ulp_cntxt_tfp_get(ctxt, BNXT_ULP_SHARED_SESSION_NO);
if (!tfp) {
BNXT_TF_DBG(ERR, "Failed to get the truflow pointer\n");
+ bnxt_ulp_cntxt_entry_release();
return;
}
if (!ulp_fc_info->num_entries) {
pthread_mutex_unlock(&ulp_fc_info->fc_lock);
ulp_fc_mgr_thread_cancel(ctxt);
+ bnxt_ulp_cntxt_entry_release();
return;
}
-
- for (i = 0; i < TF_DIR_MAX; i++) {
+ /*
+ * Commented for now till GET_BULK is resolved, just get the first flow
+ * stat for now
+ for (i = 0; i < TF_DIR_MAX; i++) {
rc = ulp_bulk_get_flow_stats(tfp, ulp_fc_info, i,
dparms->flow_count_db_entries);
if (rc)
break;
}
+ */
+
+ /* reset the parent accumulation counters before accumulation if any */
+ ulp_flow_db_parent_flow_count_reset(ctxt);
+
+ num_entries = dparms->flow_count_db_entries / 2;
+ for (i = 0; i < TF_DIR_MAX; i++) {
+ for (j = 0; j < num_entries; j++) {
+ if (!ulp_fc_info->sw_acc_tbl[i][j].valid)
+ continue;
+ hw_cntr_id = ulp_fc_info->sw_acc_tbl[i][j].hw_cntr_id;
+ rc = ulp_get_single_flow_stat(ctxt, tfp, ulp_fc_info, i,
+ hw_cntr_id, dparms);
+ if (rc)
+ break;
+ }
+ }
pthread_mutex_unlock(&ulp_fc_info->fc_lock);
if (rc) {
ulp_fc_mgr_thread_cancel(ctxt);
+ bnxt_ulp_cntxt_entry_release();
return;
}
out:
+ bnxt_ulp_cntxt_entry_release();
rte_eal_alarm_set(US_PER_S * ULP_FC_TIMER,
- ulp_fc_mgr_alarm_cb,
- (void *)ctxt);
+ ulp_fc_mgr_alarm_cb, NULL);
}
/*
ulp_fc_info = bnxt_ulp_cntxt_ptr2_fc_info_get(ctxt);
- /* Assuming start_idx of 0 is invalid */
- return (ulp_fc_info->shadow_hw_tbl[dir].start_idx != 0);
+ if (ulp_fc_info)
+ return ulp_fc_info->shadow_hw_tbl[dir].start_idx_is_set;
+
+ return false;
}
/*
if (!ulp_fc_info)
return -EIO;
- /* Assuming that 0 is an invalid counter ID ? */
- if (ulp_fc_info->shadow_hw_tbl[dir].start_idx == 0)
+ if (!ulp_fc_info->shadow_hw_tbl[dir].start_idx_is_set) {
ulp_fc_info->shadow_hw_tbl[dir].start_idx = start_idx;
+ ulp_fc_info->shadow_hw_tbl[dir].start_idx_is_set = true;
+ }
return 0;
}
if (!ulp_fc_info)
return -EIO;
+ if (!ulp_fc_info->num_counters)
+ return 0;
+
pthread_mutex_lock(&ulp_fc_info->fc_lock);
sw_cntr_idx = hw_cntr_id - ulp_fc_info->shadow_hw_tbl[dir].start_idx;
ulp_fc_info->sw_acc_tbl[dir][sw_cntr_idx].valid = true;
+ ulp_fc_info->sw_acc_tbl[dir][sw_cntr_idx].hw_cntr_id = hw_cntr_id;
ulp_fc_info->num_entries++;
pthread_mutex_unlock(&ulp_fc_info->fc_lock);
if (!ulp_fc_info)
return -EIO;
+ if (!ulp_fc_info->num_counters)
+ return 0;
+
pthread_mutex_lock(&ulp_fc_info->fc_lock);
sw_cntr_idx = hw_cntr_id - ulp_fc_info->shadow_hw_tbl[dir].start_idx;
ulp_fc_info->sw_acc_tbl[dir][sw_cntr_idx].valid = false;
+ ulp_fc_info->sw_acc_tbl[dir][sw_cntr_idx].hw_cntr_id = 0;
ulp_fc_info->sw_acc_tbl[dir][sw_cntr_idx].pkt_count = 0;
ulp_fc_info->sw_acc_tbl[dir][sw_cntr_idx].byte_count = 0;
ulp_fc_info->num_entries--;
return 0;
}
+
+/*
+ * Fill the rte_flow_query_count 'data' argument passed
+ * in the rte_flow_query() with the values obtained and
+ * accumulated locally.
+ *
+ * ctxt [in] The ulp context for the flow counter manager
+ *
+ * flow_id [in] The HW flow ID
+ *
+ * count [out] The rte_flow_query_count 'data' that is set
+ *
+ */
+int ulp_fc_mgr_query_count_get(struct bnxt_ulp_context *ctxt,
+ uint32_t flow_id,
+ struct rte_flow_query_count *count)
+{
+ int rc = 0;
+ uint32_t nxt_resource_index = 0;
+ struct bnxt_ulp_fc_info *ulp_fc_info;
+ struct ulp_flow_db_res_params params;
+ enum tf_dir dir;
+ uint32_t hw_cntr_id = 0, sw_cntr_idx = 0;
+ struct sw_acc_counter *sw_acc_tbl_entry;
+ bool found_cntr_resource = false;
+
+ ulp_fc_info = bnxt_ulp_cntxt_ptr2_fc_info_get(ctxt);
+ if (!ulp_fc_info)
+ return -ENODEV;
+
+ if (bnxt_ulp_cntxt_acquire_fdb_lock(ctxt))
+ return -EIO;
+
+ do {
+ rc = ulp_flow_db_resource_get(ctxt,
+ BNXT_ULP_FDB_TYPE_REGULAR,
+ flow_id,
+ &nxt_resource_index,
+ ¶ms);
+ if (params.resource_func ==
+ BNXT_ULP_RESOURCE_FUNC_INDEX_TABLE &&
+ (params.resource_sub_type ==
+ BNXT_ULP_RESOURCE_SUB_TYPE_INDEX_TABLE_INT_COUNT ||
+ params.resource_sub_type ==
+ BNXT_ULP_RESOURCE_SUB_TYPE_INDEX_TABLE_EXT_COUNT ||
+ params.resource_sub_type ==
+ BNXT_ULP_RESOURCE_SUB_TYPE_INDEX_TABLE_INT_COUNT_ACC)) {
+ found_cntr_resource = true;
+ break;
+ }
+ } while (!rc && nxt_resource_index);
+
+ bnxt_ulp_cntxt_release_fdb_lock(ctxt);
+
+ if (rc || !found_cntr_resource)
+ return rc;
+
+ dir = params.direction;
+ hw_cntr_id = params.resource_hndl;
+ if (params.resource_sub_type ==
+ BNXT_ULP_RESOURCE_SUB_TYPE_INDEX_TABLE_INT_COUNT) {
+ if (!ulp_fc_info->num_counters)
+ return ulp_fc_tf_flow_stat_get(ctxt, ¶ms, count);
+
+ /* TODO:
+ * Think about optimizing with try_lock later
+ */
+ pthread_mutex_lock(&ulp_fc_info->fc_lock);
+ sw_cntr_idx = hw_cntr_id -
+ ulp_fc_info->shadow_hw_tbl[dir].start_idx;
+ sw_acc_tbl_entry = &ulp_fc_info->sw_acc_tbl[dir][sw_cntr_idx];
+ if (sw_acc_tbl_entry->pkt_count) {
+ count->hits_set = 1;
+ count->bytes_set = 1;
+ count->hits = sw_acc_tbl_entry->pkt_count;
+ count->bytes = sw_acc_tbl_entry->byte_count;
+ }
+ if (count->reset) {
+ sw_acc_tbl_entry->pkt_count = 0;
+ sw_acc_tbl_entry->byte_count = 0;
+ }
+ pthread_mutex_unlock(&ulp_fc_info->fc_lock);
+ } else if (params.resource_sub_type ==
+ BNXT_ULP_RESOURCE_SUB_TYPE_INDEX_TABLE_INT_COUNT_ACC) {
+ /* Get stats from the parent child table */
+ ulp_flow_db_parent_flow_count_get(ctxt, flow_id,
+ &count->hits, &count->bytes,
+ count->reset);
+ count->hits_set = 1;
+ count->bytes_set = 1;
+ } else {
+ /* TBD: Handle External counters */
+ rc = -EINVAL;
+ }
+
+ return rc;
+}
+
+/*
+ * Set the parent flow if it is SW accumulation counter entry.
+ *
+ * ctxt [in] The ulp context for the flow counter manager
+ *
+ * dir [in] The direction of the flow
+ *
+ * hw_cntr_id [in] The HW flow counter ID
+ *
+ * fid [in] parent flow id
+ *
+ */
+int32_t ulp_fc_mgr_cntr_parent_flow_set(struct bnxt_ulp_context *ctxt,
+ enum tf_dir dir,
+ uint32_t hw_cntr_id,
+ uint32_t fid)
+{
+ struct bnxt_ulp_fc_info *ulp_fc_info;
+ uint32_t sw_cntr_idx;
+ int32_t rc = 0;
+
+ ulp_fc_info = bnxt_ulp_cntxt_ptr2_fc_info_get(ctxt);
+ if (!ulp_fc_info)
+ return -EIO;
+
+ pthread_mutex_lock(&ulp_fc_info->fc_lock);
+ sw_cntr_idx = hw_cntr_id - ulp_fc_info->shadow_hw_tbl[dir].start_idx;
+ if (ulp_fc_info->sw_acc_tbl[dir][sw_cntr_idx].valid) {
+ ulp_fc_info->sw_acc_tbl[dir][sw_cntr_idx].parent_flow_id = fid;
+ } else {
+ BNXT_TF_DBG(ERR, "Failed to set parent flow id %x:%x\n",
+ hw_cntr_id, fid);
+ rc = -ENOENT;
+ }
+ pthread_mutex_unlock(&ulp_fc_info->fc_lock);
+
+ return rc;
+}