From: Michael Wildt Date: Thu, 2 Jul 2020 23:28:10 +0000 (-0700) Subject: net/bnxt: update table get to use new design X-Git-Url: http://git.droids-corp.org/?a=commitdiff_plain;h=ced3cded44928094f9d9ec7ee06fede936d53f84;p=dpdk.git net/bnxt: update table get to use new design - Move bulk table get implementation to new Tbl Module design. - Update messages for bulk table get - Retrieve specified table element using bulk mechanism - Remove deprecated resource definitions - Update device type configuration for P4. - Update RM DB HCAPI count check and fix EM internal and host code such that EM DBs can be created correctly. - Update error logging to be info on unbind in the different modules. - Move RTE RSVD out of tf_resources.h Signed-off-by: Michael Wildt Signed-off-by: Venkat Duvvuru Reviewed-by: Randy Schacher Reviewed-by: Ajit Khaparde --- diff --git a/drivers/net/bnxt/hcapi/cfa_p40_tbl.h b/drivers/net/bnxt/hcapi/cfa_p40_tbl.h new file mode 100644 index 0000000000..76fb54ee1b --- /dev/null +++ b/drivers/net/bnxt/hcapi/cfa_p40_tbl.h @@ -0,0 +1,250 @@ +/* SPDX-License-Identifier: BSD-3-Clause + * Copyright(c) 2019-2020 Broadcom + * All rights reserved. + */ +/* + * Name: cfa_p40_tbl.h + * + * Description: header for SWE based on Truflow + * + * Date: 12/16/19 17:18:12 + * + * Note: This file was originally generated by tflib_decode.py. + * Remainder is hand coded due to lack of availability of xml for + * additional tables at this time (EEM Record and union fields) + * + **/ +#ifndef _CFA_P40_TBL_H_ +#define _CFA_P40_TBL_H_ + +#include "cfa_p40_hw.h" + +#include "hcapi_cfa_defs.h" + +const struct hcapi_cfa_field cfa_p40_prof_l2_ctxt_tcam_layout[] = { + {CFA_P40_PROF_L2_CTXT_TCAM_VALID_BITPOS, + CFA_P40_PROF_L2_CTXT_TCAM_VALID_NUM_BITS}, + {CFA_P40_PROF_L2_CTXT_TCAM_KEY_TYPE_BITPOS, + CFA_P40_PROF_L2_CTXT_TCAM_KEY_TYPE_NUM_BITS}, + {CFA_P40_PROF_L2_CTXT_TCAM_TUN_HDR_TYPE_BITPOS, + CFA_P40_PROF_L2_CTXT_TCAM_TUN_HDR_TYPE_NUM_BITS}, + {CFA_P40_PROF_L2_CTXT_TCAM_T_L2_NUMTAGS_BITPOS, + CFA_P40_PROF_L2_CTXT_TCAM_T_L2_NUMTAGS_NUM_BITS}, + {CFA_P40_PROF_L2_CTXT_TCAM_L2_NUMTAGS_BITPOS, + CFA_P40_PROF_L2_CTXT_TCAM_L2_NUMTAGS_NUM_BITS}, + {CFA_P40_PROF_L2_CTXT_TCAM_MAC1_BITPOS, + CFA_P40_PROF_L2_CTXT_TCAM_MAC1_NUM_BITS}, + {CFA_P40_PROF_L2_CTXT_TCAM_T_OVID_BITPOS, + CFA_P40_PROF_L2_CTXT_TCAM_T_OVID_NUM_BITS}, + {CFA_P40_PROF_L2_CTXT_TCAM_T_IVID_BITPOS, + CFA_P40_PROF_L2_CTXT_TCAM_T_IVID_NUM_BITS}, + {CFA_P40_PROF_L2_CTXT_TCAM_SPARIF_BITPOS, + CFA_P40_PROF_L2_CTXT_TCAM_SPARIF_NUM_BITS}, + {CFA_P40_PROF_L2_CTXT_TCAM_SVIF_BITPOS, + CFA_P40_PROF_L2_CTXT_TCAM_SVIF_NUM_BITS}, + {CFA_P40_PROF_L2_CTXT_TCAM_MAC0_BITPOS, + CFA_P40_PROF_L2_CTXT_TCAM_MAC0_NUM_BITS}, + {CFA_P40_PROF_L2_CTXT_TCAM_OVID_BITPOS, + CFA_P40_PROF_L2_CTXT_TCAM_OVID_NUM_BITS}, + {CFA_P40_PROF_L2_CTXT_TCAM_IVID_BITPOS, + CFA_P40_PROF_L2_CTXT_TCAM_IVID_NUM_BITS}, +}; + +const struct hcapi_cfa_field cfa_p40_act_veb_tcam_layout[] = { + {CFA_P40_ACT_VEB_TCAM_VALID_BITPOS, + CFA_P40_ACT_VEB_TCAM_VALID_NUM_BITS}, + {CFA_P40_ACT_VEB_TCAM_RESERVED_BITPOS, + CFA_P40_ACT_VEB_TCAM_RESERVED_NUM_BITS}, + {CFA_P40_ACT_VEB_TCAM_PARIF_IN_BITPOS, + CFA_P40_ACT_VEB_TCAM_PARIF_IN_NUM_BITS}, + {CFA_P40_ACT_VEB_TCAM_NUM_VTAGS_BITPOS, + CFA_P40_ACT_VEB_TCAM_NUM_VTAGS_NUM_BITS}, + {CFA_P40_ACT_VEB_TCAM_MAC_BITPOS, + CFA_P40_ACT_VEB_TCAM_MAC_NUM_BITS}, + {CFA_P40_ACT_VEB_TCAM_OVID_BITPOS, + CFA_P40_ACT_VEB_TCAM_OVID_NUM_BITS}, + {CFA_P40_ACT_VEB_TCAM_IVID_BITPOS, + CFA_P40_ACT_VEB_TCAM_IVID_NUM_BITS}, +}; + +const struct hcapi_cfa_field cfa_p40_lkup_tcam_record_mem_layout[] = { + {CFA_P40_LKUP_TCAM_RECORD_MEM_VALID_BITPOS, + CFA_P40_LKUP_TCAM_RECORD_MEM_VALID_NUM_BITS}, + {CFA_P40_LKUP_TCAM_RECORD_MEM_ACT_REC_PTR_BITPOS, + CFA_P40_LKUP_TCAM_RECORD_MEM_ACT_REC_PTR_NUM_BITS}, + {CFA_P40_LKUP_TCAM_RECORD_MEM_STRENGTH_BITPOS, + CFA_P40_LKUP_TCAM_RECORD_MEM_STRENGTH_NUM_BITS}, +}; + +const struct hcapi_cfa_field cfa_p40_prof_ctxt_remap_mem_layout[] = { + {CFA_P40_PROF_CTXT_REMAP_MEM_TPID_ANTI_SPOOF_CTL_BITPOS, + CFA_P40_PROF_CTXT_REMAP_MEM_TPID_ANTI_SPOOF_CTL_NUM_BITS}, + {CFA_P40_PROF_CTXT_REMAP_MEM_PRI_ANTI_SPOOF_CTL_BITPOS, + CFA_P40_PROF_CTXT_REMAP_MEM_PRI_ANTI_SPOOF_CTL_NUM_BITS}, + {CFA_P40_PROF_CTXT_REMAP_MEM_BYP_SP_LKUP_BITPOS, + CFA_P40_PROF_CTXT_REMAP_MEM_BYP_SP_LKUP_NUM_BITS}, + {CFA_P40_PROF_CTXT_REMAP_MEM_SP_REC_PTR_BITPOS, + CFA_P40_PROF_CTXT_REMAP_MEM_SP_REC_PTR_NUM_BITS}, + {CFA_P40_PROF_CTXT_REMAP_MEM_BD_ACT_EN_BITPOS, + CFA_P40_PROF_CTXT_REMAP_MEM_BD_ACT_EN_NUM_BITS}, + {CFA_P40_PROF_CTXT_REMAP_MEM_DEFAULT_TPID_BITPOS, + CFA_P40_PROF_CTXT_REMAP_MEM_DEFAULT_TPID_NUM_BITS}, + {CFA_P40_PROF_CTXT_REMAP_MEM_ALLOWED_TPID_BITPOS, + CFA_P40_PROF_CTXT_REMAP_MEM_ALLOWED_TPID_NUM_BITS}, + {CFA_P40_PROF_CTXT_REMAP_MEM_DEFAULT_PRI_BITPOS, + CFA_P40_PROF_CTXT_REMAP_MEM_DEFAULT_PRI_NUM_BITS}, + {CFA_P40_PROF_CTXT_REMAP_MEM_ALLOWED_PRI_BITPOS, + CFA_P40_PROF_CTXT_REMAP_MEM_ALLOWED_PRI_NUM_BITS}, + {CFA_P40_PROF_CTXT_REMAP_MEM_PARIF_BITPOS, + CFA_P40_PROF_CTXT_REMAP_MEM_PARIF_NUM_BITS}, + {CFA_P40_PROF_CTXT_REMAP_MEM_BYP_LKUP_EN_BITPOS, + CFA_P40_PROF_CTXT_REMAP_MEM_BYP_LKUP_EN_NUM_BITS}, + /* Fields below not generated through automation */ + {CFA_P40_PROF_CTXT_REMAP_MEM_PROF_VNIC_BITPOS, + CFA_P40_PROF_CTXT_REMAP_MEM_PROF_VNIC_NUM_BITS}, + {CFA_P40_PROF_CTXT_REMAP_MEM_PROF_FUNC_BITPOS, + CFA_P40_PROF_CTXT_REMAP_MEM_PROF_FUNC_NUM_BITS}, + {CFA_P40_PROF_CTXT_REMAP_MEM_L2_CTXT_BITPOS, + CFA_P40_PROF_CTXT_REMAP_MEM_L2_CTXT_NUM_BITS}, + {CFA_P40_PROF_CTXT_REMAP_MEM_ARP_BITPOS, + CFA_P40_PROF_CTXT_REMAP_MEM_ARP_NUM_BITS}, +}; + +const struct hcapi_cfa_field cfa_p40_prof_profile_tcam_remap_mem_layout[] = { + {CFA_P40_PROF_PROFILE_TCAM_REMAP_MEM_PL_BYP_LKUP_EN_BITPOS, + CFA_P40_PROF_PROFILE_TCAM_REMAP_MEM_PL_BYP_LKUP_EN_NUM_BITS}, + {CFA_P40_PROF_PROFILE_TCAM_REMAP_MEM_EM_SEARCH_ENB_BITPOS, + CFA_P40_PROF_PROFILE_TCAM_REMAP_MEM_EM_SEARCH_ENB_NUM_BITS}, + {CFA_P40_PROF_PROFILE_TCAM_REMAP_MEM_EM_PROFILE_ID_BITPOS, + CFA_P40_PROF_PROFILE_TCAM_REMAP_MEM_EM_PROFILE_ID_NUM_BITS}, + {CFA_P40_PROF_PROFILE_TCAM_REMAP_MEM_EM_KEY_ID_BITPOS, + CFA_P40_PROF_PROFILE_TCAM_REMAP_MEM_EM_KEY_ID_NUM_BITS}, + {CFA_P40_PROF_PROFILE_TCAM_REMAP_MEM_EM_KEY_MASK_BITPOS, + CFA_P40_PROF_PROFILE_TCAM_REMAP_MEM_EM_KEY_MASK_NUM_BITS}, + {CFA_P40_PROF_PROFILE_TCAM_REMAP_MEM_TCAM_SEARCH_ENB_BITPOS, + CFA_P40_PROF_PROFILE_TCAM_REMAP_MEM_TCAM_SEARCH_ENB_NUM_BITS}, + {CFA_P40_PROF_PROFILE_TCAM_REMAP_MEM_TCAM_PROFILE_ID_BITPOS, + CFA_P40_PROF_PROFILE_TCAM_REMAP_MEM_TCAM_PROFILE_ID_NUM_BITS}, + {CFA_P40_PROF_PROFILE_TCAM_REMAP_MEM_TCAM_KEY_ID_BITPOS, + CFA_P40_PROF_PROFILE_TCAM_REMAP_MEM_TCAM_KEY_ID_NUM_BITS}, +}; + +const struct hcapi_cfa_field cfa_p40_prof_profile_tcam_layout[] = { + {CFA_P40_PROF_PROFILE_TCAM_VALID_BITPOS, + CFA_P40_PROF_PROFILE_TCAM_VALID_NUM_BITS}, + {CFA_P40_PROF_PROFILE_TCAM_PKT_TYPE_BITPOS, + CFA_P40_PROF_PROFILE_TCAM_PKT_TYPE_NUM_BITS}, + {CFA_P40_PROF_PROFILE_TCAM_RECYCLE_CNT_BITPOS, + CFA_P40_PROF_PROFILE_TCAM_RECYCLE_CNT_NUM_BITS}, + {CFA_P40_PROF_PROFILE_TCAM_AGG_ERROR_BITPOS, + CFA_P40_PROF_PROFILE_TCAM_AGG_ERROR_NUM_BITS}, + {CFA_P40_PROF_PROFILE_TCAM_PROF_FUNC_BITPOS, + CFA_P40_PROF_PROFILE_TCAM_PROF_FUNC_NUM_BITS}, + {CFA_P40_PROF_PROFILE_TCAM_RESERVED_BITPOS, + CFA_P40_PROF_PROFILE_TCAM_RESERVED_NUM_BITS}, + {CFA_P40_PROF_PROFILE_TCAM_HREC_NEXT_BITPOS, + CFA_P40_PROF_PROFILE_TCAM_HREC_NEXT_NUM_BITS}, + {CFA_P40_PROF_PROFILE_TCAM_TL2_HDR_VALID_BITPOS, + CFA_P40_PROF_PROFILE_TCAM_TL2_HDR_VALID_NUM_BITS}, + {CFA_P40_PROF_PROFILE_TCAM_TL2_HDR_TYPE_BITPOS, + CFA_P40_PROF_PROFILE_TCAM_TL2_HDR_TYPE_NUM_BITS}, + {CFA_P40_PROF_PROFILE_TCAM_TL2_UC_MC_BC_BITPOS, + CFA_P40_PROF_PROFILE_TCAM_TL2_UC_MC_BC_NUM_BITS}, + {CFA_P40_PROF_PROFILE_TCAM_TL2_VTAG_PRESENT_BITPOS, + CFA_P40_PROF_PROFILE_TCAM_TL2_VTAG_PRESENT_NUM_BITS}, + {CFA_P40_PROF_PROFILE_TCAM_TL2_TWO_VTAGS_BITPOS, + CFA_P40_PROF_PROFILE_TCAM_TL2_TWO_VTAGS_NUM_BITS}, + {CFA_P40_PROF_PROFILE_TCAM_TL3_VALID_BITPOS, + CFA_P40_PROF_PROFILE_TCAM_TL3_VALID_NUM_BITS}, + {CFA_P40_PROF_PROFILE_TCAM_TL3_ERROR_BITPOS, + CFA_P40_PROF_PROFILE_TCAM_TL3_ERROR_NUM_BITS}, + {CFA_P40_PROF_PROFILE_TCAM_TL3_HDR_TYPE_BITPOS, + CFA_P40_PROF_PROFILE_TCAM_TL3_HDR_TYPE_NUM_BITS}, + {CFA_P40_PROF_PROFILE_TCAM_TL3_HDR_ISIP_BITPOS, + CFA_P40_PROF_PROFILE_TCAM_TL3_HDR_ISIP_NUM_BITS}, + {CFA_P40_PROF_PROFILE_TCAM_TL3_IPV6_CMP_SRC_BITPOS, + CFA_P40_PROF_PROFILE_TCAM_TL3_IPV6_CMP_SRC_NUM_BITS}, + {CFA_P40_PROF_PROFILE_TCAM_TL3_IPV6_CMP_DEST_BITPOS, + CFA_P40_PROF_PROFILE_TCAM_TL3_IPV6_CMP_DEST_NUM_BITS}, + {CFA_P40_PROF_PROFILE_TCAM_TL4_HDR_VALID_BITPOS, + CFA_P40_PROF_PROFILE_TCAM_TL4_HDR_VALID_NUM_BITS}, + {CFA_P40_PROF_PROFILE_TCAM_TL4_HDR_ERROR_BITPOS, + CFA_P40_PROF_PROFILE_TCAM_TL4_HDR_ERROR_NUM_BITS}, + {CFA_P40_PROF_PROFILE_TCAM_TL4_HDR_TYPE_BITPOS, + CFA_P40_PROF_PROFILE_TCAM_TL4_HDR_TYPE_NUM_BITS}, + {CFA_P40_PROF_PROFILE_TCAM_TL4_HDR_IS_UDP_TCP_BITPOS, + CFA_P40_PROF_PROFILE_TCAM_TL4_HDR_IS_UDP_TCP_NUM_BITS}, + {CFA_P40_PROF_PROFILE_TCAM_TUN_HDR_VALID_BITPOS, + CFA_P40_PROF_PROFILE_TCAM_TUN_HDR_VALID_NUM_BITS}, + {CFA_P40_PROF_PROFILE_TCAM_TUN_HDR_ERR_BITPOS, + CFA_P40_PROF_PROFILE_TCAM_TUN_HDR_ERR_NUM_BITS}, + {CFA_P40_PROF_PROFILE_TCAM_TUN_HDR_TYPE_BITPOS, + CFA_P40_PROF_PROFILE_TCAM_TUN_HDR_TYPE_NUM_BITS}, + {CFA_P40_PROF_PROFILE_TCAM_TUN_HDR_FLAGS_BITPOS, + CFA_P40_PROF_PROFILE_TCAM_TUN_HDR_FLAGS_NUM_BITS}, + {CFA_P40_PROF_PROFILE_TCAM_L2_HDR_VALID_BITPOS, + CFA_P40_PROF_PROFILE_TCAM_L2_HDR_VALID_NUM_BITS}, + {CFA_P40_PROF_PROFILE_TCAM_L2_HDR_ERROR_BITPOS, + CFA_P40_PROF_PROFILE_TCAM_L2_HDR_ERROR_NUM_BITS}, + {CFA_P40_PROF_PROFILE_TCAM_L2_HDR_TYPE_BITPOS, + CFA_P40_PROF_PROFILE_TCAM_L2_HDR_TYPE_NUM_BITS}, + {CFA_P40_PROF_PROFILE_TCAM_L2_UC_MC_BC_BITPOS, + CFA_P40_PROF_PROFILE_TCAM_L2_UC_MC_BC_NUM_BITS}, + {CFA_P40_PROF_PROFILE_TCAM_L2_VTAG_PRESENT_BITPOS, + CFA_P40_PROF_PROFILE_TCAM_L2_VTAG_PRESENT_NUM_BITS}, + {CFA_P40_PROF_PROFILE_TCAM_L2_TWO_VTAGS_BITPOS, + CFA_P40_PROF_PROFILE_TCAM_L2_TWO_VTAGS_NUM_BITS}, + {CFA_P40_PROF_PROFILE_TCAM_L3_VALID_BITPOS, + CFA_P40_PROF_PROFILE_TCAM_L3_VALID_NUM_BITS}, + {CFA_P40_PROF_PROFILE_TCAM_L3_ERROR_BITPOS, + CFA_P40_PROF_PROFILE_TCAM_L3_ERROR_NUM_BITS}, + {CFA_P40_PROF_PROFILE_TCAM_L3_HDR_TYPE_BITPOS, + CFA_P40_PROF_PROFILE_TCAM_L3_HDR_TYPE_NUM_BITS}, + {CFA_P40_PROF_PROFILE_TCAM_L3_HDR_ISIP_BITPOS, + CFA_P40_PROF_PROFILE_TCAM_L3_HDR_ISIP_NUM_BITS}, + {CFA_P40_PROF_PROFILE_TCAM_L3_IPV6_CMP_SRC_BITPOS, + CFA_P40_PROF_PROFILE_TCAM_L3_IPV6_CMP_SRC_NUM_BITS}, + {CFA_P40_PROF_PROFILE_TCAM_L3_IPV6_CMP_DEST_BITPOS, + CFA_P40_PROF_PROFILE_TCAM_L3_IPV6_CMP_DEST_NUM_BITS}, + {CFA_P40_PROF_PROFILE_TCAM_L4_HDR_VALID_BITPOS, + CFA_P40_PROF_PROFILE_TCAM_L4_HDR_VALID_NUM_BITS}, + {CFA_P40_PROF_PROFILE_TCAM_L4_HDR_ERROR_BITPOS, + CFA_P40_PROF_PROFILE_TCAM_L4_HDR_ERROR_NUM_BITS}, + {CFA_P40_PROF_PROFILE_TCAM_L4_HDR_TYPE_BITPOS, + CFA_P40_PROF_PROFILE_TCAM_L4_HDR_TYPE_NUM_BITS}, + {CFA_P40_PROF_PROFILE_TCAM_L4_HDR_IS_UDP_TCP_BITPOS, + CFA_P40_PROF_PROFILE_TCAM_L4_HDR_IS_UDP_TCP_NUM_BITS}, +}; + +/**************************************************************************/ +/** + * Non-autogenerated fields + */ + +const struct hcapi_cfa_field cfa_p40_eem_key_tbl_layout[] = { + {CFA_P40_EEM_KEY_TBL_VALID_BITPOS, + CFA_P40_EEM_KEY_TBL_VALID_NUM_BITS}, + + {CFA_P40_EEM_KEY_TBL_L1_CACHEABLE_BITPOS, + CFA_P40_EEM_KEY_TBL_L1_CACHEABLE_NUM_BITS}, + + {CFA_P40_EEM_KEY_TBL_STRENGTH_BITPOS, + CFA_P40_EEM_KEY_TBL_STRENGTH_NUM_BITS}, + + {CFA_P40_EEM_KEY_TBL_KEY_SZ_BITPOS, + CFA_P40_EEM_KEY_TBL_KEY_SZ_NUM_BITS}, + + {CFA_P40_EEM_KEY_TBL_REC_SZ_BITPOS, + CFA_P40_EEM_KEY_TBL_REC_SZ_NUM_BITS}, + + {CFA_P40_EEM_KEY_TBL_ACT_REC_INT_BITPOS, + CFA_P40_EEM_KEY_TBL_ACT_REC_INT_NUM_BITS}, + + {CFA_P40_EEM_KEY_TBL_EXT_FLOW_CTR_BITPOS, + CFA_P40_EEM_KEY_TBL_EXT_FLOW_CTR_NUM_BITS}, + + {CFA_P40_EEM_KEY_TBL_AR_PTR_BITPOS, + CFA_P40_EEM_KEY_TBL_AR_PTR_NUM_BITS}, + +}; +#endif /* _CFA_P40_TBL_H_ */ diff --git a/drivers/net/bnxt/hcapi/hcapi_cfa.h b/drivers/net/bnxt/hcapi/hcapi_cfa.h index 0724865db1..89b2315a82 100644 --- a/drivers/net/bnxt/hcapi/hcapi_cfa.h +++ b/drivers/net/bnxt/hcapi/hcapi_cfa.h @@ -243,6 +243,8 @@ int hcapi_cfa_p4_wc_tcam_hwop(struct hcapi_cfa_hwop *op, struct hcapi_cfa_data *obj_data); int hcapi_cfa_p4_wc_tcam_rec_hwop(struct hcapi_cfa_hwop *op, struct hcapi_cfa_data *obj_data); +int hcapi_cfa_p4_mirror_hwop(struct hcapi_cfa_hwop *op, + struct hcapi_cfa_data *mirror); #endif /* SUPPORT_CFA_HW_P4 */ /** * HCAPI CFA device HW operation function callback definition diff --git a/drivers/net/bnxt/meson.build b/drivers/net/bnxt/meson.build index 35038dc8b3..7f3ec62043 100644 --- a/drivers/net/bnxt/meson.build +++ b/drivers/net/bnxt/meson.build @@ -41,10 +41,9 @@ sources = files('bnxt_cpr.c', 'tf_core/tf_identifier.c', 'tf_core/tf_shadow_tbl.c', 'tf_core/tf_shadow_tcam.c', - 'tf_core/tf_tbl_type.c', 'tf_core/tf_tcam.c', 'tf_core/tf_util.c', - 'tf_core/tf_rm_new.c', + 'tf_core/tf_rm.c', 'hcapi/hcapi_cfa_p4.c', diff --git a/drivers/net/bnxt/tf_core/Makefile b/drivers/net/bnxt/tf_core/Makefile index 7191a2a21a..a48f150baf 100644 --- a/drivers/net/bnxt/tf_core/Makefile +++ b/drivers/net/bnxt/tf_core/Makefile @@ -23,7 +23,5 @@ SRCS-$(CONFIG_RTE_LIBRTE_BNXT_PMD) += tf_core/tf_device_p4.c SRCS-$(CONFIG_RTE_LIBRTE_BNXT_PMD) += tf_core/tf_identifier.c SRCS-$(CONFIG_RTE_LIBRTE_BNXT_PMD) += tf_core/tf_shadow_tbl.c SRCS-$(CONFIG_RTE_LIBRTE_BNXT_PMD) += tf_core/tf_shadow_tcam.c -SRCS-$(CONFIG_RTE_LIBRTE_BNXT_PMD) += tf_core/tf_tbl_type.c SRCS-$(CONFIG_RTE_LIBRTE_BNXT_PMD) += tf_core/tf_tcam.c SRCS-$(CONFIG_RTE_LIBRTE_BNXT_PMD) += tf_core/tf_util.c -SRCS-$(CONFIG_RTE_LIBRTE_BNXT_PMD) += tf_core/tf_rm_new.c diff --git a/drivers/net/bnxt/tf_core/tf_common.h b/drivers/net/bnxt/tf_core/tf_common.h index ec3bca835d..b982203db7 100644 --- a/drivers/net/bnxt/tf_core/tf_common.h +++ b/drivers/net/bnxt/tf_core/tf_common.h @@ -6,52 +6,11 @@ #ifndef _TF_COMMON_H_ #define _TF_COMMON_H_ -/* Helper to check the parms */ -#define TF_CHECK_PARMS_SESSION(tfp, parms) do { \ - if ((parms) == NULL || (tfp) == NULL) { \ - TFP_DRV_LOG(ERR, "Invalid Argument(s)\n"); \ - return -EINVAL; \ - } \ - if ((tfp)->session == NULL || \ - (tfp)->session->core_data == NULL) { \ - TFP_DRV_LOG(ERR, "%s: session error\n", \ - tf_dir_2_str((parms)->dir)); \ - return -EINVAL; \ - } \ - } while (0) - -#define TF_CHECK_PARMS_SESSION_NO_DIR(tfp, parms) do { \ - if ((parms) == NULL || (tfp) == NULL) { \ - TFP_DRV_LOG(ERR, "Invalid Argument(s)\n"); \ - return -EINVAL; \ - } \ - if ((tfp)->session == NULL || \ - (tfp)->session->core_data == NULL) { \ - TFP_DRV_LOG(ERR, "Session error\n"); \ - return -EINVAL; \ - } \ - } while (0) - -#define TF_CHECK_PARMS(tfp, parms) do { \ - if ((parms) == NULL || (tfp) == NULL) { \ - TFP_DRV_LOG(ERR, "Invalid Argument(s)\n"); \ - return -EINVAL; \ - } \ - } while (0) - -#define TF_CHECK_TFP_SESSION(tfp) do { \ - if ((tfp) == NULL) { \ - TFP_DRV_LOG(ERR, "Invalid Argument(s)\n"); \ - return -EINVAL; \ - } \ - if ((tfp)->session == NULL || \ - (tfp)->session->core_data == NULL) { \ - TFP_DRV_LOG(ERR, "Session error\n"); \ - return -EINVAL; \ - } \ - } while (0) - +/* Helpers to performs parameter check */ +/** + * Checks 1 parameter against NULL. + */ #define TF_CHECK_PARMS1(parms) do { \ if ((parms) == NULL) { \ TFP_DRV_LOG(ERR, "Invalid Argument(s)\n"); \ @@ -59,6 +18,9 @@ } \ } while (0) +/** + * Checks 2 parameters against NULL. + */ #define TF_CHECK_PARMS2(parms1, parms2) do { \ if ((parms1) == NULL || (parms2) == NULL) { \ TFP_DRV_LOG(ERR, "Invalid Argument(s)\n"); \ @@ -66,6 +28,9 @@ } \ } while (0) +/** + * Checks 3 parameters against NULL. + */ #define TF_CHECK_PARMS3(parms1, parms2, parms3) do { \ if ((parms1) == NULL || \ (parms2) == NULL || \ diff --git a/drivers/net/bnxt/tf_core/tf_core.c b/drivers/net/bnxt/tf_core/tf_core.c index 8b3e15c8ab..8727900c4d 100644 --- a/drivers/net/bnxt/tf_core/tf_core.c +++ b/drivers/net/bnxt/tf_core/tf_core.c @@ -186,7 +186,7 @@ int tf_insert_em_entry(struct tf *tfp, struct tf_dev_info *dev; int rc; - TF_CHECK_PARMS_SESSION(tfp, parms); + TF_CHECK_PARMS2(tfp, parms); /* Retrieve the session information */ rc = tf_session_get_session(tfp, &tfs); @@ -241,7 +241,7 @@ int tf_delete_em_entry(struct tf *tfp, struct tf_dev_info *dev; int rc; - TF_CHECK_PARMS_SESSION(tfp, parms); + TF_CHECK_PARMS2(tfp, parms); /* Retrieve the session information */ rc = tf_session_get_session(tfp, &tfs); @@ -523,7 +523,7 @@ int tf_get_tcam_entry(struct tf *tfp __rte_unused, struct tf_get_tcam_entry_parms *parms __rte_unused) { - TF_CHECK_PARMS_SESSION(tfp, parms); + TF_CHECK_PARMS2(tfp, parms); return -EOPNOTSUPP; } @@ -821,7 +821,80 @@ tf_get_tbl_entry(struct tf *tfp, return rc; } -/* API defined in tf_core.h */ +int +tf_bulk_get_tbl_entry(struct tf *tfp, + struct tf_bulk_get_tbl_entry_parms *parms) +{ + int rc = 0; + struct tf_session *tfs; + struct tf_dev_info *dev; + struct tf_tbl_get_bulk_parms bparms; + + TF_CHECK_PARMS2(tfp, parms); + + /* Can't do static initialization due to UT enum check */ + memset(&bparms, 0, sizeof(struct tf_tbl_get_bulk_parms)); + + /* Retrieve the session information */ + rc = tf_session_get_session(tfp, &tfs); + if (rc) { + TFP_DRV_LOG(ERR, + "%s: Failed to lookup session, rc:%s\n", + tf_dir_2_str(parms->dir), + strerror(-rc)); + return rc; + } + + /* Retrieve the device information */ + rc = tf_session_get_device(tfs, &dev); + if (rc) { + TFP_DRV_LOG(ERR, + "%s: Failed to lookup device, rc:%s\n", + tf_dir_2_str(parms->dir), + strerror(-rc)); + return rc; + } + + if (parms->type == TF_TBL_TYPE_EXT) { + /* Not supported, yet */ + rc = -EOPNOTSUPP; + TFP_DRV_LOG(ERR, + "%s, External table type not supported, rc:%s\n", + tf_dir_2_str(parms->dir), + strerror(-rc)); + + return rc; + } + + /* Internal table type processing */ + + if (dev->ops->tf_dev_get_bulk_tbl == NULL) { + rc = -EOPNOTSUPP; + TFP_DRV_LOG(ERR, + "%s: Operation not supported, rc:%s\n", + tf_dir_2_str(parms->dir), + strerror(-rc)); + return -EOPNOTSUPP; + } + + bparms.dir = parms->dir; + bparms.type = parms->type; + bparms.starting_idx = parms->starting_idx; + bparms.num_entries = parms->num_entries; + bparms.entry_sz_in_bytes = parms->entry_sz_in_bytes; + bparms.physical_mem_addr = parms->physical_mem_addr; + rc = dev->ops->tf_dev_get_bulk_tbl(tfp, &bparms); + if (rc) { + TFP_DRV_LOG(ERR, + "%s: Table get bulk failed, rc:%s\n", + tf_dir_2_str(parms->dir), + strerror(-rc)); + return rc; + } + + return rc; +} + int tf_alloc_tbl_scope(struct tf *tfp, struct tf_alloc_tbl_scope_parms *parms) @@ -830,7 +903,7 @@ tf_alloc_tbl_scope(struct tf *tfp, struct tf_dev_info *dev; int rc; - TF_CHECK_PARMS_SESSION_NO_DIR(tfp, parms); + TF_CHECK_PARMS2(tfp, parms); /* Retrieve the session information */ rc = tf_session_get_session(tfp, &tfs); @@ -861,7 +934,6 @@ tf_alloc_tbl_scope(struct tf *tfp, return rc; } -/* API defined in tf_core.h */ int tf_free_tbl_scope(struct tf *tfp, struct tf_free_tbl_scope_parms *parms) @@ -870,7 +942,7 @@ tf_free_tbl_scope(struct tf *tfp, struct tf_dev_info *dev; int rc; - TF_CHECK_PARMS_SESSION_NO_DIR(tfp, parms); + TF_CHECK_PARMS2(tfp, parms); /* Retrieve the session information */ rc = tf_session_get_session(tfp, &tfs); diff --git a/drivers/net/bnxt/tf_core/tf_device.h b/drivers/net/bnxt/tf_core/tf_device.h index 3c56ab3488..00bd996e5e 100644 --- a/drivers/net/bnxt/tf_core/tf_device.h +++ b/drivers/net/bnxt/tf_core/tf_device.h @@ -8,7 +8,7 @@ #include "tf_core.h" #include "tf_identifier.h" -#include "tf_tbl_type.h" +#include "tf_tbl.h" #include "tf_tcam.h" struct tf; @@ -293,7 +293,27 @@ struct tf_dev_ops { * - (-EINVAL) on failure. */ int (*tf_dev_get_tbl)(struct tf *tfp, - struct tf_tbl_get_parms *parms); + struct tf_tbl_get_parms *parms); + + /** + * Retrieves the specified table type element using 'bulk' + * mechanism. + * + * This API retrieves the specified element data by invoking the + * firmware. + * + * [in] tfp + * Pointer to TF handle + * + * [in] parms + * Pointer to table get bulk parameters + * + * Returns + * - (0) if successful. + * - (-EINVAL) on failure. + */ + int (*tf_dev_get_bulk_tbl)(struct tf *tfp, + struct tf_tbl_get_bulk_parms *parms); /** * Allocation of a tcam element. diff --git a/drivers/net/bnxt/tf_core/tf_device_p4.c b/drivers/net/bnxt/tf_core/tf_device_p4.c index 127c655a64..e3526672f0 100644 --- a/drivers/net/bnxt/tf_core/tf_device_p4.c +++ b/drivers/net/bnxt/tf_core/tf_device_p4.c @@ -8,7 +8,7 @@ #include "tf_device.h" #include "tf_identifier.h" -#include "tf_tbl_type.h" +#include "tf_tbl.h" #include "tf_tcam.h" #include "tf_em.h" @@ -88,6 +88,7 @@ const struct tf_dev_ops tf_dev_ops_p4_init = { .tf_dev_alloc_search_tbl = NULL, .tf_dev_set_tbl = NULL, .tf_dev_get_tbl = NULL, + .tf_dev_get_bulk_tbl = NULL, .tf_dev_alloc_tcam = NULL, .tf_dev_free_tcam = NULL, .tf_dev_alloc_search_tcam = NULL, @@ -114,6 +115,7 @@ const struct tf_dev_ops tf_dev_ops_p4 = { .tf_dev_alloc_search_tbl = NULL, .tf_dev_set_tbl = tf_tbl_set, .tf_dev_get_tbl = tf_tbl_get, + .tf_dev_get_bulk_tbl = tf_tbl_bulk_get, .tf_dev_alloc_tcam = tf_tcam_alloc, .tf_dev_free_tcam = tf_tcam_free, .tf_dev_alloc_search_tcam = NULL, diff --git a/drivers/net/bnxt/tf_core/tf_device_p4.h b/drivers/net/bnxt/tf_core/tf_device_p4.h index da6dd65a3f..473e4eae55 100644 --- a/drivers/net/bnxt/tf_core/tf_device_p4.h +++ b/drivers/net/bnxt/tf_core/tf_device_p4.h @@ -9,7 +9,7 @@ #include #include "tf_core.h" -#include "tf_rm_new.h" +#include "tf_rm.h" struct tf_rm_element_cfg tf_ident_p4[TF_IDENT_TYPE_MAX] = { { TF_RM_ELEM_CFG_HCAPI, CFA_RESOURCE_TYPE_P4_L2_CTXT_REMAP }, @@ -41,8 +41,7 @@ struct tf_rm_element_cfg tf_tbl_p4[TF_TBL_TYPE_MAX] = { { TF_RM_ELEM_CFG_HCAPI, CFA_RESOURCE_TYPE_P4_ENCAP_64B }, { TF_RM_ELEM_CFG_HCAPI, CFA_RESOURCE_TYPE_P4_SP_MAC }, { TF_RM_ELEM_CFG_HCAPI, CFA_RESOURCE_TYPE_P4_SP_MAC_IPV4 }, - /* CFA_RESOURCE_TYPE_P4_SP_MAC_IPV6 */ - { TF_RM_ELEM_CFG_NULL, CFA_RESOURCE_TYPE_INVALID }, + { TF_RM_ELEM_CFG_HCAPI, CFA_RESOURCE_TYPE_P4_SP_MAC_IPV6 }, { TF_RM_ELEM_CFG_HCAPI, CFA_RESOURCE_TYPE_P4_COUNTER_64B }, { TF_RM_ELEM_CFG_HCAPI, CFA_RESOURCE_TYPE_P4_NAT_SPORT }, { TF_RM_ELEM_CFG_HCAPI, CFA_RESOURCE_TYPE_P4_NAT_DPORT }, diff --git a/drivers/net/bnxt/tf_core/tf_em.h b/drivers/net/bnxt/tf_core/tf_em.h index cf799c200d..6bfcbd59e5 100644 --- a/drivers/net/bnxt/tf_core/tf_em.h +++ b/drivers/net/bnxt/tf_core/tf_em.h @@ -23,6 +23,56 @@ #define TF_EM_MAX_MASK 0x7FFF #define TF_EM_MAX_ENTRY (128 * 1024 * 1024) +/** + * Hardware Page sizes supported for EEM: + * 4K, 8K, 64K, 256K, 1M, 2M, 4M, 1G. + * + * Round-down other page sizes to the lower hardware page + * size supported. + */ +#define TF_EM_PAGE_SIZE_4K 12 +#define TF_EM_PAGE_SIZE_8K 13 +#define TF_EM_PAGE_SIZE_64K 16 +#define TF_EM_PAGE_SIZE_256K 18 +#define TF_EM_PAGE_SIZE_1M 20 +#define TF_EM_PAGE_SIZE_2M 21 +#define TF_EM_PAGE_SIZE_4M 22 +#define TF_EM_PAGE_SIZE_1G 30 + +/* Set page size */ +#define BNXT_TF_PAGE_SIZE TF_EM_PAGE_SIZE_2M + +#if (BNXT_TF_PAGE_SIZE == TF_EM_PAGE_SIZE_4K) /** 4K */ +#define TF_EM_PAGE_SHIFT TF_EM_PAGE_SIZE_4K +#define TF_EM_PAGE_SIZE_ENUM HWRM_TF_CTXT_MEM_RGTR_INPUT_PAGE_SIZE_4K +#elif (BNXT_TF_PAGE_SIZE == TF_EM_PAGE_SIZE_8K) /** 8K */ +#define TF_EM_PAGE_SHIFT TF_EM_PAGE_SIZE_8K +#define TF_EM_PAGE_SIZE_ENUM HWRM_TF_CTXT_MEM_RGTR_INPUT_PAGE_SIZE_8K +#elif (BNXT_TF_PAGE_SIZE == TF_EM_PAGE_SIZE_64K) /** 64K */ +#define TF_EM_PAGE_SHIFT TF_EM_PAGE_SIZE_64K +#define TF_EM_PAGE_SIZE_ENUM HWRM_TF_CTXT_MEM_RGTR_INPUT_PAGE_SIZE_64K +#elif (BNXT_TF_PAGE_SIZE == TF_EM_PAGE_SIZE_256K) /** 256K */ +#define TF_EM_PAGE_SHIFT TF_EM_PAGE_SIZE_256K +#define TF_EM_PAGE_SIZE_ENUM HWRM_TF_CTXT_MEM_RGTR_INPUT_PAGE_SIZE_256K +#elif (BNXT_TF_PAGE_SIZE == TF_EM_PAGE_SIZE_1M) /** 1M */ +#define TF_EM_PAGE_SHIFT TF_EM_PAGE_SIZE_1M +#define TF_EM_PAGE_SIZE_ENUM HWRM_TF_CTXT_MEM_RGTR_INPUT_PAGE_SIZE_1M +#elif (BNXT_TF_PAGE_SIZE == TF_EM_PAGE_SIZE_2M) /** 2M */ +#define TF_EM_PAGE_SHIFT TF_EM_PAGE_SIZE_2M +#define TF_EM_PAGE_SIZE_ENUM HWRM_TF_CTXT_MEM_RGTR_INPUT_PAGE_SIZE_2M +#elif (BNXT_TF_PAGE_SIZE == TF_EM_PAGE_SIZE_4M) /** 4M */ +#define TF_EM_PAGE_SHIFT TF_EM_PAGE_SIZE_4M +#define TF_EM_PAGE_SIZE_ENUM HWRM_TF_CTXT_MEM_RGTR_INPUT_PAGE_SIZE_4M +#elif (BNXT_TF_PAGE_SIZE == TF_EM_PAGE_SIZE_1G) /** 1G */ +#define TF_EM_PAGE_SHIFT TF_EM_PAGE_SIZE_1G +#define TF_EM_PAGE_SIZE_ENUM HWRM_TF_CTXT_MEM_RGTR_INPUT_PAGE_SIZE_1G +#else +#error "Invalid Page Size specified. Please use a TF_EM_PAGE_SIZE_n define" +#endif + +#define TF_EM_PAGE_SIZE (1 << TF_EM_PAGE_SHIFT) +#define TF_EM_PAGE_ALIGNMENT (1 << TF_EM_PAGE_SHIFT) + /* * Used to build GFID: * @@ -80,13 +130,43 @@ struct tf_em_cfg_parms { }; /** - * @page table Table + * @page em EM * * @ref tf_alloc_eem_tbl_scope * * @ref tf_free_eem_tbl_scope_cb * - * @ref tbl_scope_cb_find + * @ref tf_em_insert_int_entry + * + * @ref tf_em_delete_int_entry + * + * @ref tf_em_insert_ext_entry + * + * @ref tf_em_delete_ext_entry + * + * @ref tf_em_insert_ext_sys_entry + * + * @ref tf_em_delete_ext_sys_entry + * + * @ref tf_em_int_bind + * + * @ref tf_em_int_unbind + * + * @ref tf_em_ext_common_bind + * + * @ref tf_em_ext_common_unbind + * + * @ref tf_em_ext_host_alloc + * + * @ref tf_em_ext_host_free + * + * @ref tf_em_ext_system_alloc + * + * @ref tf_em_ext_system_free + * + * @ref tf_em_ext_common_free + * + * @ref tf_em_ext_common_alloc */ /** @@ -328,7 +408,7 @@ int tf_em_ext_host_free(struct tf *tfp, * -EINVAL - Parameter error */ int tf_em_ext_system_alloc(struct tf *tfp, - struct tf_alloc_tbl_scope_parms *parms); + struct tf_alloc_tbl_scope_parms *parms); /** * Free for external EEM using system memory @@ -344,7 +424,7 @@ int tf_em_ext_system_alloc(struct tf *tfp, * -EINVAL - Parameter error */ int tf_em_ext_system_free(struct tf *tfp, - struct tf_free_tbl_scope_parms *parms); + struct tf_free_tbl_scope_parms *parms); /** * Common free for external EEM using host or system memory diff --git a/drivers/net/bnxt/tf_core/tf_em_common.c b/drivers/net/bnxt/tf_core/tf_em_common.c index ba6aa7ac15..d0d80daeb0 100644 --- a/drivers/net/bnxt/tf_core/tf_em_common.c +++ b/drivers/net/bnxt/tf_core/tf_em_common.c @@ -194,12 +194,13 @@ tf_em_ext_common_bind(struct tf *tfp, int rc; int i; struct tf_rm_create_db_parms db_cfg = { 0 }; + uint8_t db_exists = 0; TF_CHECK_PARMS2(tfp, parms); if (init) { TFP_DRV_LOG(ERR, - "Identifier already initialized\n"); + "EM Ext DB already initialized\n"); return -EINVAL; } @@ -210,19 +211,29 @@ tf_em_ext_common_bind(struct tf *tfp, for (i = 0; i < TF_DIR_MAX; i++) { db_cfg.dir = i; db_cfg.alloc_cnt = parms->resources->em_cnt[i].cnt; + + /* Check if we got any request to support EEM, if so + * we build an EM Ext DB holding Table Scopes. + */ + if (db_cfg.alloc_cnt[TF_EM_TBL_TYPE_TBL_SCOPE] == 0) + continue; + db_cfg.rm_db = &eem_db[i]; rc = tf_rm_create_db(tfp, &db_cfg); if (rc) { TFP_DRV_LOG(ERR, - "%s: EM DB creation failed\n", + "%s: EM Ext DB creation failed\n", tf_dir_2_str(i)); return rc; } + db_exists = 1; } - mem_type = parms->mem_type; - init = 1; + if (db_exists) { + mem_type = parms->mem_type; + init = 1; + } return 0; } @@ -236,13 +247,11 @@ tf_em_ext_common_unbind(struct tf *tfp) TF_CHECK_PARMS1(tfp); - /* Bail if nothing has been initialized done silent as to - * allow for creation cleanup. - */ + /* Bail if nothing has been initialized */ if (!init) { - TFP_DRV_LOG(ERR, - "No EM DBs created\n"); - return -EINVAL; + TFP_DRV_LOG(INFO, + "No EM Ext DBs created\n"); + return 0; } for (i = 0; i < TF_DIR_MAX; i++) { diff --git a/drivers/net/bnxt/tf_core/tf_em_internal.c b/drivers/net/bnxt/tf_core/tf_em_internal.c index 9be91ad5dd..1c514747d8 100644 --- a/drivers/net/bnxt/tf_core/tf_em_internal.c +++ b/drivers/net/bnxt/tf_core/tf_em_internal.c @@ -225,12 +225,13 @@ tf_em_int_bind(struct tf *tfp, int i; struct tf_rm_create_db_parms db_cfg = { 0 }; struct tf_session *session; + uint8_t db_exists = 0; TF_CHECK_PARMS2(tfp, parms); if (init) { TFP_DRV_LOG(ERR, - "Identifier already initialized\n"); + "EM Int DB already initialized\n"); return -EINVAL; } @@ -242,31 +243,35 @@ tf_em_int_bind(struct tf *tfp, TF_SESSION_EM_POOL_SIZE); } - /* - * I'm not sure that this code is needed. - * leaving for now until resolved - */ - if (parms->num_elements) { - db_cfg.type = TF_DEVICE_MODULE_TYPE_EM; - db_cfg.num_elements = parms->num_elements; - db_cfg.cfg = parms->cfg; - - for (i = 0; i < TF_DIR_MAX; i++) { - db_cfg.dir = i; - db_cfg.alloc_cnt = parms->resources->em_cnt[i].cnt; - db_cfg.rm_db = &em_db[i]; - rc = tf_rm_create_db(tfp, &db_cfg); - if (rc) { - TFP_DRV_LOG(ERR, - "%s: EM DB creation failed\n", - tf_dir_2_str(i)); + db_cfg.type = TF_DEVICE_MODULE_TYPE_EM; + db_cfg.num_elements = parms->num_elements; + db_cfg.cfg = parms->cfg; - return rc; - } + for (i = 0; i < TF_DIR_MAX; i++) { + db_cfg.dir = i; + db_cfg.alloc_cnt = parms->resources->em_cnt[i].cnt; + + /* Check if we got any request to support EEM, if so + * we build an EM Int DB holding Table Scopes. + */ + if (db_cfg.alloc_cnt[TF_EM_TBL_TYPE_EM_RECORD] == 0) + continue; + + db_cfg.rm_db = &em_db[i]; + rc = tf_rm_create_db(tfp, &db_cfg); + if (rc) { + TFP_DRV_LOG(ERR, + "%s: EM Int DB creation failed\n", + tf_dir_2_str(i)); + + return rc; } + db_exists = 1; } - init = 1; + if (db_exists) + init = 1; + return 0; } @@ -280,13 +285,11 @@ tf_em_int_unbind(struct tf *tfp) TF_CHECK_PARMS1(tfp); - /* Bail if nothing has been initialized done silent as to - * allow for creation cleanup. - */ + /* Bail if nothing has been initialized */ if (!init) { - TFP_DRV_LOG(ERR, - "No EM DBs created\n"); - return -EINVAL; + TFP_DRV_LOG(INFO, + "No EM Int DBs created\n"); + return 0; } session = (struct tf_session *)tfp->session->core_data; diff --git a/drivers/net/bnxt/tf_core/tf_identifier.c b/drivers/net/bnxt/tf_core/tf_identifier.c index b197bb2711..2113710814 100644 --- a/drivers/net/bnxt/tf_core/tf_identifier.c +++ b/drivers/net/bnxt/tf_core/tf_identifier.c @@ -7,7 +7,7 @@ #include "tf_identifier.h" #include "tf_common.h" -#include "tf_rm_new.h" +#include "tf_rm.h" #include "tf_util.h" #include "tfp.h" @@ -35,7 +35,7 @@ tf_ident_bind(struct tf *tfp, if (init) { TFP_DRV_LOG(ERR, - "Identifier already initialized\n"); + "Identifier DB already initialized\n"); return -EINVAL; } @@ -65,7 +65,7 @@ tf_ident_bind(struct tf *tfp, } int -tf_ident_unbind(struct tf *tfp __rte_unused) +tf_ident_unbind(struct tf *tfp) { int rc; int i; @@ -73,13 +73,11 @@ tf_ident_unbind(struct tf *tfp __rte_unused) TF_CHECK_PARMS1(tfp); - /* Bail if nothing has been initialized done silent as to - * allow for creation cleanup. - */ + /* Bail if nothing has been initialized */ if (!init) { - TFP_DRV_LOG(ERR, + TFP_DRV_LOG(INFO, "No Identifier DBs created\n"); - return -EINVAL; + return 0; } for (i = 0; i < TF_DIR_MAX; i++) { diff --git a/drivers/net/bnxt/tf_core/tf_msg.c b/drivers/net/bnxt/tf_core/tf_msg.c index d8b80bc84c..02d8a49713 100644 --- a/drivers/net/bnxt/tf_core/tf_msg.c +++ b/drivers/net/bnxt/tf_core/tf_msg.c @@ -871,26 +871,41 @@ tf_msg_get_tbl_entry(struct tf *tfp, int tf_msg_bulk_get_tbl_entry(struct tf *tfp, - struct tf_bulk_get_tbl_entry_parms *params) + enum tf_dir dir, + uint16_t hcapi_type, + uint32_t starting_idx, + uint16_t num_entries, + uint16_t entry_sz_in_bytes, + uint64_t physical_mem_addr) { int rc; struct tfp_send_msg_parms parms = { 0 }; struct tf_tbl_type_bulk_get_input req = { 0 }; struct tf_tbl_type_bulk_get_output resp = { 0 }; - struct tf_session *tfs = (struct tf_session *)(tfp->session->core_data); + struct tf_session *tfs; int data_size = 0; + /* Retrieve the session information */ + rc = tf_session_get_session(tfp, &tfs); + if (rc) { + TFP_DRV_LOG(ERR, + "%s: Failed to lookup session, rc:%s\n", + tf_dir_2_str(dir), + strerror(-rc)); + return rc; + } + /* Populate the request */ req.fw_session_id = tfp_cpu_to_le_32(tfs->session_id.internal.fw_session_id); - req.flags = tfp_cpu_to_le_16(params->dir); - req.type = tfp_cpu_to_le_32(params->type); - req.start_index = tfp_cpu_to_le_32(params->starting_idx); - req.num_entries = tfp_cpu_to_le_32(params->num_entries); + req.flags = tfp_cpu_to_le_16(dir); + req.type = tfp_cpu_to_le_32(hcapi_type); + req.start_index = tfp_cpu_to_le_32(starting_idx); + req.num_entries = tfp_cpu_to_le_32(num_entries); - data_size = params->num_entries * params->entry_sz_in_bytes; + data_size = num_entries * entry_sz_in_bytes; - req.host_addr = tfp_cpu_to_le_64(params->physical_mem_addr); + req.host_addr = tfp_cpu_to_le_64(physical_mem_addr); MSG_PREP(parms, TF_KONG_MB, diff --git a/drivers/net/bnxt/tf_core/tf_msg.h b/drivers/net/bnxt/tf_core/tf_msg.h index 8e276d4c01..7432873d7e 100644 --- a/drivers/net/bnxt/tf_core/tf_msg.h +++ b/drivers/net/bnxt/tf_core/tf_msg.h @@ -11,7 +11,6 @@ #include "tf_tbl.h" #include "tf_rm.h" -#include "tf_rm_new.h" #include "tf_tcam.h" struct tf; @@ -422,6 +421,11 @@ int tf_msg_get_tbl_entry(struct tf *tfp, * 0 on Success else internal Truflow error */ int tf_msg_bulk_get_tbl_entry(struct tf *tfp, - struct tf_bulk_get_tbl_entry_parms *parms); + enum tf_dir dir, + uint16_t hcapi_type, + uint32_t starting_idx, + uint16_t num_entries, + uint16_t entry_sz_in_bytes, + uint64_t physical_mem_addr); #endif /* _TF_MSG_H_ */ diff --git a/drivers/net/bnxt/tf_core/tf_resources.h b/drivers/net/bnxt/tf_core/tf_resources.h index b7b4451020..4688514fcf 100644 --- a/drivers/net/bnxt/tf_core/tf_resources.h +++ b/drivers/net/bnxt/tf_core/tf_resources.h @@ -6,535 +6,6 @@ #ifndef _TF_RESOURCES_H_ #define _TF_RESOURCES_H_ -/* - * Hardware specific MAX values - * NOTE: Should really come from the chip_cfg.h in some MAX form or HCAPI - */ - -/* Common HW resources for all chip variants */ -#define TF_NUM_L2_CTXT_TCAM 1024 /* < Number of L2 context TCAM - * entries - */ -#define TF_NUM_PROF_FUNC 128 /* < Number prof_func ID */ -#define TF_NUM_PROF_TCAM 1024 /* < Number entries in profile - * TCAM - */ -#define TF_NUM_EM_PROF_ID 64 /* < Number software EM Profile - * IDs - */ -#define TF_NUM_WC_PROF_ID 256 /* < Number WC profile IDs */ -#define TF_NUM_WC_TCAM_ROW 512 /* < Number of rows in WC TCAM */ -#define TF_NUM_METER_PROF 256 /* < Number of meter profiles */ -#define TF_NUM_METER 1024 /* < Number of meter instances */ -#define TF_NUM_MIRROR 2 /* < Number of mirror instances */ -#define TF_NUM_UPAR 2 /* < Number of UPAR instances */ - -/* Wh+/SR specific HW resources */ -#define TF_NUM_SP_TCAM 512 /* < Number of Source Property TCAM - * entries - */ - -/* SR/SR2 specific HW resources */ -#define TF_NUM_L2_FUNC 256 /* < Number of L2 Func */ - - -/* Thor, SR2 common HW resources */ -#define TF_NUM_FKB 1 /* < Number of Flexible Key Builder - * templates - */ - -/* SR2 specific HW resources */ #define TF_NUM_TBL_SCOPE 16 /* < Number of TBL scopes */ -#define TF_NUM_EPOCH0 1 /* < Number of Epoch0 */ -#define TF_NUM_EPOCH1 1 /* < Number of Epoch1 */ -#define TF_NUM_METADATA 8 /* < Number of MetaData Profiles */ -#define TF_NUM_CT_STATE 32 /* < Number of Connection Tracking - * States - */ -#define TF_NUM_RANGE_PROF 16 /* < Number of Range Profiles */ -#define TF_NUM_RANGE_ENTRY (64 * 1024) /* < Number of Range Entries */ -#define TF_NUM_LAG_ENTRY 256 /* < Number of LAG Entries */ - -/* - * Common for the Reserved Resource defines below: - * - * - HW Resources - * For resources where a priority level plays a role, i.e. l2 ctx - * tcam entries, both a number of resources and a begin/end pair is - * required. The begin/end is used to assure TFLIB gets the correct - * priority setting for that resource. - * - * For EM records there is no priority required thus a number of - * resources is sufficient. - * - * Example, TCAM: - * 64 L2 CTXT TCAM entries would in a max 1024 pool be entry - * 0-63 as HW presents 0 as the highest priority entry. - * - * - SRAM Resources - * Handled as regular resources as there is no priority required. - * - * Common for these resources is that they are handled per direction, - * rx/tx. - */ - -/* HW Resources */ - -/* L2 CTX */ -#define TF_RSVD_L2_CTXT_TCAM_RX 64 -#define TF_RSVD_L2_CTXT_TCAM_BEGIN_IDX_RX 0 -#define TF_RSVD_L2_CTXT_TCAM_END_IDX_RX (TF_RSVD_L2_CTXT_RX - 1) -#define TF_RSVD_L2_CTXT_TCAM_TX 960 -#define TF_RSVD_L2_CTXT_TCAM_BEGIN_IDX_TX 0 -#define TF_RSVD_L2_CTXT_TCAM_END_IDX_TX (TF_RSVD_L2_CTXT_TX - 1) - -/* Profiler */ -#define TF_RSVD_PROF_FUNC_RX 64 -#define TF_RSVD_PROF_FUNC_BEGIN_IDX_RX 64 -#define TF_RSVD_PROF_FUNC_END_IDX_RX 127 -#define TF_RSVD_PROF_FUNC_TX 64 -#define TF_RSVD_PROF_FUNC_BEGIN_IDX_TX 64 -#define TF_RSVD_PROF_FUNC_END_IDX_TX 127 - -#define TF_RSVD_PROF_TCAM_RX 64 -#define TF_RSVD_PROF_TCAM_BEGIN_IDX_RX 960 -#define TF_RSVD_PROF_TCAM_END_IDX_RX 1023 -#define TF_RSVD_PROF_TCAM_TX 64 -#define TF_RSVD_PROF_TCAM_BEGIN_IDX_TX 960 -#define TF_RSVD_PROF_TCAM_END_IDX_TX 1023 - -/* EM Profiles IDs */ -#define TF_RSVD_EM_PROF_ID_RX 64 -#define TF_RSVD_EM_PROF_ID_BEGIN_IDX_RX 0 -#define TF_RSVD_EM_PROF_ID_END_IDX_RX 63 /* Less on CU+ then SR */ -#define TF_RSVD_EM_PROF_ID_TX 64 -#define TF_RSVD_EM_PROF_ID_BEGIN_IDX_TX 0 -#define TF_RSVD_EM_PROF_ID_END_IDX_TX 63 /* Less on CU+ then SR */ - -/* EM Records */ -#define TF_RSVD_EM_REC_RX 16000 -#define TF_RSVD_EM_REC_BEGIN_IDX_RX 0 -#define TF_RSVD_EM_REC_TX 16000 -#define TF_RSVD_EM_REC_BEGIN_IDX_TX 0 - -/* Wildcard */ -#define TF_RSVD_WC_TCAM_PROF_ID_RX 128 -#define TF_RSVD_WC_TCAM_PROF_ID_BEGIN_IDX_RX 128 -#define TF_RSVD_WC_TCAM_PROF_ID_END_IDX_RX 255 -#define TF_RSVD_WC_TCAM_PROF_ID_TX 128 -#define TF_RSVD_WC_TCAM_PROF_ID_BEGIN_IDX_TX 128 -#define TF_RSVD_WC_TCAM_PROF_ID_END_IDX_TX 255 - -#define TF_RSVD_WC_TCAM_RX 64 -#define TF_RSVD_WC_TCAM_BEGIN_IDX_RX 0 -#define TF_RSVD_WC_TCAM_END_IDX_RX 63 -#define TF_RSVD_WC_TCAM_TX 64 -#define TF_RSVD_WC_TCAM_BEGIN_IDX_TX 0 -#define TF_RSVD_WC_TCAM_END_IDX_TX 63 - -#define TF_RSVD_METER_PROF_RX 0 -#define TF_RSVD_METER_PROF_BEGIN_IDX_RX 0 -#define TF_RSVD_METER_PROF_END_IDX_RX 0 -#define TF_RSVD_METER_PROF_TX 0 -#define TF_RSVD_METER_PROF_BEGIN_IDX_TX 0 -#define TF_RSVD_METER_PROF_END_IDX_TX 0 - -#define TF_RSVD_METER_INST_RX 0 -#define TF_RSVD_METER_INST_BEGIN_IDX_RX 0 -#define TF_RSVD_METER_INST_END_IDX_RX 0 -#define TF_RSVD_METER_INST_TX 0 -#define TF_RSVD_METER_INST_BEGIN_IDX_TX 0 -#define TF_RSVD_METER_INST_END_IDX_TX 0 - -/* Mirror */ -/* Not yet supported fully in the infra */ -#define TF_RSVD_MIRROR_RX 0 -#define TF_RSVD_MIRROR_BEGIN_IDX_RX 0 -#define TF_RSVD_MIRROR_END_IDX_RX 0 -#define TF_RSVD_MIRROR_TX 0 -#define TF_RSVD_MIRROR_BEGIN_IDX_TX 0 -#define TF_RSVD_MIRROR_END_IDX_TX 0 - -/* UPAR */ -/* Not yet supported fully in the infra */ -#define TF_RSVD_UPAR_RX 0 -#define TF_RSVD_UPAR_BEGIN_IDX_RX 0 -#define TF_RSVD_UPAR_END_IDX_RX 0 -#define TF_RSVD_UPAR_TX 0 -#define TF_RSVD_UPAR_BEGIN_IDX_TX 0 -#define TF_RSVD_UPAR_END_IDX_TX 0 - -/* Source Properties */ -/* Not yet supported fully in the infra */ -#define TF_RSVD_SP_TCAM_RX 0 -#define TF_RSVD_SP_TCAM_BEGIN_IDX_RX 0 -#define TF_RSVD_SP_TCAM_END_IDX_RX 0 -#define TF_RSVD_SP_TCAM_TX 0 -#define TF_RSVD_SP_TCAM_BEGIN_IDX_TX 0 -#define TF_RSVD_SP_TCAM_END_IDX_TX 0 - -/* L2 Func */ -#define TF_RSVD_L2_FUNC_RX 0 -#define TF_RSVD_L2_FUNC_BEGIN_IDX_RX 0 -#define TF_RSVD_L2_FUNC_END_IDX_RX 0 -#define TF_RSVD_L2_FUNC_TX 0 -#define TF_RSVD_L2_FUNC_BEGIN_IDX_TX 0 -#define TF_RSVD_L2_FUNC_END_IDX_TX 0 - -/* FKB */ -#define TF_RSVD_FKB_RX 0 -#define TF_RSVD_FKB_BEGIN_IDX_RX 0 -#define TF_RSVD_FKB_END_IDX_RX 0 -#define TF_RSVD_FKB_TX 0 -#define TF_RSVD_FKB_BEGIN_IDX_TX 0 -#define TF_RSVD_FKB_END_IDX_TX 0 - -/* TBL Scope */ -#define TF_RSVD_TBL_SCOPE_RX 1 -#define TF_RSVD_TBL_SCOPE_BEGIN_IDX_RX 0 -#define TF_RSVD_TBL_SCOPE_END_IDX_RX 1 -#define TF_RSVD_TBL_SCOPE_TX 1 -#define TF_RSVD_TBL_SCOPE_BEGIN_IDX_TX 0 -#define TF_RSVD_TBL_SCOPE_END_IDX_TX 1 - -/* EPOCH0 */ -/* Not yet supported fully in the infra */ -#define TF_RSVD_EPOCH0_RX 0 -#define TF_RSVD_EPOCH0_BEGIN_IDX_RX 0 -#define TF_RSVD_EPOCH0_END_IDX_RX 0 -#define TF_RSVD_EPOCH0_TX 0 -#define TF_RSVD_EPOCH0_BEGIN_IDX_TX 0 -#define TF_RSVD_EPOCH0_END_IDX_TX 0 - -/* EPOCH1 */ -/* Not yet supported fully in the infra */ -#define TF_RSVD_EPOCH1_RX 0 -#define TF_RSVD_EPOCH1_BEGIN_IDX_RX 0 -#define TF_RSVD_EPOCH1_END_IDX_RX 0 -#define TF_RSVD_EPOCH1_TX 0 -#define TF_RSVD_EPOCH1_BEGIN_IDX_TX 0 -#define TF_RSVD_EPOCH1_END_IDX_TX 0 - -/* METADATA */ -/* Not yet supported fully in the infra */ -#define TF_RSVD_METADATA_RX 0 -#define TF_RSVD_METADATA_BEGIN_IDX_RX 0 -#define TF_RSVD_METADATA_END_IDX_RX 0 -#define TF_RSVD_METADATA_TX 0 -#define TF_RSVD_METADATA_BEGIN_IDX_TX 0 -#define TF_RSVD_METADATA_END_IDX_TX 0 - -/* CT_STATE */ -/* Not yet supported fully in the infra */ -#define TF_RSVD_CT_STATE_RX 0 -#define TF_RSVD_CT_STATE_BEGIN_IDX_RX 0 -#define TF_RSVD_CT_STATE_END_IDX_RX 0 -#define TF_RSVD_CT_STATE_TX 0 -#define TF_RSVD_CT_STATE_BEGIN_IDX_TX 0 -#define TF_RSVD_CT_STATE_END_IDX_TX 0 - -/* RANGE_PROF */ -/* Not yet supported fully in the infra */ -#define TF_RSVD_RANGE_PROF_RX 0 -#define TF_RSVD_RANGE_PROF_BEGIN_IDX_RX 0 -#define TF_RSVD_RANGE_PROF_END_IDX_RX 0 -#define TF_RSVD_RANGE_PROF_TX 0 -#define TF_RSVD_RANGE_PROF_BEGIN_IDX_TX 0 -#define TF_RSVD_RANGE_PROF_END_IDX_TX 0 - -/* RANGE_ENTRY */ -/* Not yet supported fully in the infra */ -#define TF_RSVD_RANGE_ENTRY_RX 0 -#define TF_RSVD_RANGE_ENTRY_BEGIN_IDX_RX 0 -#define TF_RSVD_RANGE_ENTRY_END_IDX_RX 0 -#define TF_RSVD_RANGE_ENTRY_TX 0 -#define TF_RSVD_RANGE_ENTRY_BEGIN_IDX_TX 0 -#define TF_RSVD_RANGE_ENTRY_END_IDX_TX 0 - -/* LAG_ENTRY */ -/* Not yet supported fully in the infra */ -#define TF_RSVD_LAG_ENTRY_RX 0 -#define TF_RSVD_LAG_ENTRY_BEGIN_IDX_RX 0 -#define TF_RSVD_LAG_ENTRY_END_IDX_RX 0 -#define TF_RSVD_LAG_ENTRY_TX 0 -#define TF_RSVD_LAG_ENTRY_BEGIN_IDX_TX 0 -#define TF_RSVD_LAG_ENTRY_END_IDX_TX 0 - - -/* SRAM - Resources - * Limited to the types that CFA provides. - */ -#define TF_RSVD_SRAM_FULL_ACTION_RX 8001 -#define TF_RSVD_SRAM_FULL_ACTION_BEGIN_IDX_RX 0 -#define TF_RSVD_SRAM_FULL_ACTION_TX 8001 -#define TF_RSVD_SRAM_FULL_ACTION_BEGIN_IDX_TX 0 - -/* Not yet supported fully in the infra */ -#define TF_RSVD_SRAM_MCG_RX 0 -#define TF_RSVD_SRAM_MCG_BEGIN_IDX_RX 0 -/* Multicast Group on TX is not supported */ -#define TF_RSVD_SRAM_MCG_TX 0 -#define TF_RSVD_SRAM_MCG_BEGIN_IDX_TX 0 - -/* First encap of 8B RX is reserved by CFA */ -#define TF_RSVD_SRAM_ENCAP_8B_RX 32 -#define TF_RSVD_SRAM_ENCAP_8B_BEGIN_IDX_RX 0 -/* First encap of 8B TX is reserved by CFA */ -#define TF_RSVD_SRAM_ENCAP_8B_TX 0 -#define TF_RSVD_SRAM_ENCAP_8B_BEGIN_IDX_TX 0 - -#define TF_RSVD_SRAM_ENCAP_16B_RX 16 -#define TF_RSVD_SRAM_ENCAP_16B_BEGIN_IDX_RX 0 -/* First encap of 16B TX is reserved by CFA */ -#define TF_RSVD_SRAM_ENCAP_16B_TX 20 -#define TF_RSVD_SRAM_ENCAP_16B_BEGIN_IDX_TX 0 - -/* Encap of 64B on RX is not supported */ -#define TF_RSVD_SRAM_ENCAP_64B_RX 0 -#define TF_RSVD_SRAM_ENCAP_64B_BEGIN_IDX_RX 0 -/* First encap of 64B TX is reserved by CFA */ -#define TF_RSVD_SRAM_ENCAP_64B_TX 1007 -#define TF_RSVD_SRAM_ENCAP_64B_BEGIN_IDX_TX 0 - -#define TF_RSVD_SRAM_SP_SMAC_RX 0 -#define TF_RSVD_SRAM_SP_SMAC_BEGIN_IDX_RX 0 -#define TF_RSVD_SRAM_SP_SMAC_TX 0 -#define TF_RSVD_SRAM_SP_SMAC_BEGIN_IDX_TX 0 - -/* SRAM SP IPV4 on RX is not supported */ -#define TF_RSVD_SRAM_SP_SMAC_IPV4_RX 0 -#define TF_RSVD_SRAM_SP_SMAC_IPV4_BEGIN_IDX_RX 0 -#define TF_RSVD_SRAM_SP_SMAC_IPV4_TX 511 -#define TF_RSVD_SRAM_SP_SMAC_IPV4_BEGIN_IDX_TX 0 - -/* SRAM SP IPV6 on RX is not supported */ -#define TF_RSVD_SRAM_SP_SMAC_IPV6_RX 0 -#define TF_RSVD_SRAM_SP_SMAC_IPV6_BEGIN_IDX_RX 0 -/* Not yet supported fully in infra */ -#define TF_RSVD_SRAM_SP_SMAC_IPV6_TX 0 -#define TF_RSVD_SRAM_SP_SMAC_IPV6_BEGIN_IDX_TX 0 - -#define TF_RSVD_SRAM_COUNTER_64B_RX 160 -#define TF_RSVD_SRAM_COUNTER_64B_BEGIN_IDX_RX 0 -#define TF_RSVD_SRAM_COUNTER_64B_TX 160 -#define TF_RSVD_SRAM_COUNTER_64B_BEGIN_IDX_TX 0 - -#define TF_RSVD_SRAM_NAT_SPORT_RX 0 -#define TF_RSVD_SRAM_NAT_SPORT_BEGIN_IDX_RX 0 -#define TF_RSVD_SRAM_NAT_SPORT_TX 0 -#define TF_RSVD_SRAM_NAT_SPORT_BEGIN_IDX_TX 0 - -#define TF_RSVD_SRAM_NAT_DPORT_RX 0 -#define TF_RSVD_SRAM_NAT_DPORT_BEGIN_IDX_RX 0 -#define TF_RSVD_SRAM_NAT_DPORT_TX 0 -#define TF_RSVD_SRAM_NAT_DPORT_BEGIN_IDX_TX 0 - -#define TF_RSVD_SRAM_NAT_S_IPV4_RX 0 -#define TF_RSVD_SRAM_NAT_S_IPV4_BEGIN_IDX_RX 0 -#define TF_RSVD_SRAM_NAT_S_IPV4_TX 0 -#define TF_RSVD_SRAM_NAT_S_IPV4_BEGIN_IDX_TX 0 - -#define TF_RSVD_SRAM_NAT_D_IPV4_RX 0 -#define TF_RSVD_SRAM_NAT_D_IPV4_BEGIN_IDX_RX 0 -#define TF_RSVD_SRAM_NAT_D_IPV4_TX 0 -#define TF_RSVD_SRAM_NAT_D_IPV4_BEGIN_IDX_TX 0 - -/* HW Resource Pool names */ - -#define TF_L2_CTXT_TCAM_POOL_NAME l2_ctxt_tcam_pool -#define TF_L2_CTXT_TCAM_POOL_NAME_RX l2_ctxt_tcam_pool_rx -#define TF_L2_CTXT_TCAM_POOL_NAME_TX l2_ctxt_tcam_pool_tx - -#define TF_PROF_FUNC_POOL_NAME prof_func_pool -#define TF_PROF_FUNC_POOL_NAME_RX prof_func_pool_rx -#define TF_PROF_FUNC_POOL_NAME_TX prof_func_pool_tx - -#define TF_PROF_TCAM_POOL_NAME prof_tcam_pool -#define TF_PROF_TCAM_POOL_NAME_RX prof_tcam_pool_rx -#define TF_PROF_TCAM_POOL_NAME_TX prof_tcam_pool_tx - -#define TF_EM_PROF_ID_POOL_NAME em_prof_id_pool -#define TF_EM_PROF_ID_POOL_NAME_RX em_prof_id_pool_rx -#define TF_EM_PROF_ID_POOL_NAME_TX em_prof_id_pool_tx - -#define TF_WC_TCAM_PROF_ID_POOL_NAME wc_tcam_prof_id_pool -#define TF_WC_TCAM_PROF_ID_POOL_NAME_RX wc_tcam_prof_id_pool_rx -#define TF_WC_TCAM_PROF_ID_POOL_NAME_TX wc_tcam_prof_id_pool_tx - -#define TF_WC_TCAM_POOL_NAME wc_tcam_pool -#define TF_WC_TCAM_POOL_NAME_RX wc_tcam_pool_rx -#define TF_WC_TCAM_POOL_NAME_TX wc_tcam_pool_tx - -#define TF_METER_PROF_POOL_NAME meter_prof_pool -#define TF_METER_PROF_POOL_NAME_RX meter_prof_pool_rx -#define TF_METER_PROF_POOL_NAME_TX meter_prof_pool_tx - -#define TF_METER_INST_POOL_NAME meter_inst_pool -#define TF_METER_INST_POOL_NAME_RX meter_inst_pool_rx -#define TF_METER_INST_POOL_NAME_TX meter_inst_pool_tx - -#define TF_MIRROR_POOL_NAME mirror_pool -#define TF_MIRROR_POOL_NAME_RX mirror_pool_rx -#define TF_MIRROR_POOL_NAME_TX mirror_pool_tx - -#define TF_UPAR_POOL_NAME upar_pool -#define TF_UPAR_POOL_NAME_RX upar_pool_rx -#define TF_UPAR_POOL_NAME_TX upar_pool_tx - -#define TF_SP_TCAM_POOL_NAME sp_tcam_pool -#define TF_SP_TCAM_POOL_NAME_RX sp_tcam_pool_rx -#define TF_SP_TCAM_POOL_NAME_TX sp_tcam_pool_tx - -#define TF_FKB_POOL_NAME fkb_pool -#define TF_FKB_POOL_NAME_RX fkb_pool_rx -#define TF_FKB_POOL_NAME_TX fkb_pool_tx - -#define TF_TBL_SCOPE_POOL_NAME tbl_scope_pool -#define TF_TBL_SCOPE_POOL_NAME_RX tbl_scope_pool_rx -#define TF_TBL_SCOPE_POOL_NAME_TX tbl_scope_pool_tx - -#define TF_L2_FUNC_POOL_NAME l2_func_pool -#define TF_L2_FUNC_POOL_NAME_RX l2_func_pool_rx -#define TF_L2_FUNC_POOL_NAME_TX l2_func_pool_tx - -#define TF_EPOCH0_POOL_NAME epoch0_pool -#define TF_EPOCH0_POOL_NAME_RX epoch0_pool_rx -#define TF_EPOCH0_POOL_NAME_TX epoch0_pool_tx - -#define TF_EPOCH1_POOL_NAME epoch1_pool -#define TF_EPOCH1_POOL_NAME_RX epoch1_pool_rx -#define TF_EPOCH1_POOL_NAME_TX epoch1_pool_tx - -#define TF_METADATA_POOL_NAME metadata_pool -#define TF_METADATA_POOL_NAME_RX metadata_pool_rx -#define TF_METADATA_POOL_NAME_TX metadata_pool_tx - -#define TF_CT_STATE_POOL_NAME ct_state_pool -#define TF_CT_STATE_POOL_NAME_RX ct_state_pool_rx -#define TF_CT_STATE_POOL_NAME_TX ct_state_pool_tx - -#define TF_RANGE_PROF_POOL_NAME range_prof_pool -#define TF_RANGE_PROF_POOL_NAME_RX range_prof_pool_rx -#define TF_RANGE_PROF_POOL_NAME_TX range_prof_pool_tx - -#define TF_RANGE_ENTRY_POOL_NAME range_entry_pool -#define TF_RANGE_ENTRY_POOL_NAME_RX range_entry_pool_rx -#define TF_RANGE_ENTRY_POOL_NAME_TX range_entry_pool_tx - -#define TF_LAG_ENTRY_POOL_NAME lag_entry_pool -#define TF_LAG_ENTRY_POOL_NAME_RX lag_entry_pool_rx -#define TF_LAG_ENTRY_POOL_NAME_TX lag_entry_pool_tx - -/* SRAM Resource Pool names */ -#define TF_SRAM_FULL_ACTION_POOL_NAME sram_full_action_pool -#define TF_SRAM_FULL_ACTION_POOL_NAME_RX sram_full_action_pool_rx -#define TF_SRAM_FULL_ACTION_POOL_NAME_TX sram_full_action_pool_tx - -#define TF_SRAM_MCG_POOL_NAME sram_mcg_pool -#define TF_SRAM_MCG_POOL_NAME_RX sram_mcg_pool_rx -#define TF_SRAM_MCG_POOL_NAME_TX sram_mcg_pool_tx - -#define TF_SRAM_ENCAP_8B_POOL_NAME sram_encap_8b_pool -#define TF_SRAM_ENCAP_8B_POOL_NAME_RX sram_encap_8b_pool_rx -#define TF_SRAM_ENCAP_8B_POOL_NAME_TX sram_encap_8b_pool_tx - -#define TF_SRAM_ENCAP_16B_POOL_NAME sram_encap_16b_pool -#define TF_SRAM_ENCAP_16B_POOL_NAME_RX sram_encap_16b_pool_rx -#define TF_SRAM_ENCAP_16B_POOL_NAME_TX sram_encap_16b_pool_tx - -#define TF_SRAM_ENCAP_64B_POOL_NAME sram_encap_64b_pool -#define TF_SRAM_ENCAP_64B_POOL_NAME_RX sram_encap_64b_pool_rx -#define TF_SRAM_ENCAP_64B_POOL_NAME_TX sram_encap_64b_pool_tx - -#define TF_SRAM_SP_SMAC_POOL_NAME sram_sp_smac_pool -#define TF_SRAM_SP_SMAC_POOL_NAME_RX sram_sp_smac_pool_rx -#define TF_SRAM_SP_SMAC_POOL_NAME_TX sram_sp_smac_pool_tx - -#define TF_SRAM_SP_SMAC_IPV4_POOL_NAME sram_sp_smac_ipv4_pool -#define TF_SRAM_SP_SMAC_IPV4_POOL_NAME_RX sram_sp_smac_ipv4_pool_rx -#define TF_SRAM_SP_SMAC_IPV4_POOL_NAME_TX sram_sp_smac_ipv4_pool_tx - -#define TF_SRAM_SP_SMAC_IPV6_POOL_NAME sram_sp_smac_ipv6_pool -#define TF_SRAM_SP_SMAC_IPV6_POOL_NAME_RX sram_sp_smac_ipv6_pool_rx -#define TF_SRAM_SP_SMAC_IPV6_POOL_NAME_TX sram_sp_smac_ipv6_pool_tx - -#define TF_SRAM_STATS_64B_POOL_NAME sram_stats_64b_pool -#define TF_SRAM_STATS_64B_POOL_NAME_RX sram_stats_64b_pool_rx -#define TF_SRAM_STATS_64B_POOL_NAME_TX sram_stats_64b_pool_tx - -#define TF_SRAM_NAT_SPORT_POOL_NAME sram_nat_sport_pool -#define TF_SRAM_NAT_SPORT_POOL_NAME_RX sram_nat_sport_pool_rx -#define TF_SRAM_NAT_SPORT_POOL_NAME_TX sram_nat_sport_pool_tx - -#define TF_SRAM_NAT_DPORT_POOL_NAME sram_nat_dport_pool -#define TF_SRAM_NAT_DPORT_POOL_NAME_RX sram_nat_dport_pool_rx -#define TF_SRAM_NAT_DPORT_POOL_NAME_TX sram_nat_dport_pool_tx - -#define TF_SRAM_NAT_S_IPV4_POOL_NAME sram_nat_s_ipv4_pool -#define TF_SRAM_NAT_S_IPV4_POOL_NAME_RX sram_nat_s_ipv4_pool_rx -#define TF_SRAM_NAT_S_IPV4_POOL_NAME_TX sram_nat_s_ipv4_pool_tx - -#define TF_SRAM_NAT_D_IPV4_POOL_NAME sram_nat_d_ipv4_pool -#define TF_SRAM_NAT_D_IPV4_POOL_NAME_RX sram_nat_d_ipv4_pool_rx -#define TF_SRAM_NAT_D_IPV4_POOL_NAME_TX sram_nat_d_ipv4_pool_tx - -/* Sw Resource Pool Names */ - -#define TF_L2_CTXT_REMAP_POOL_NAME l2_ctxt_remap_pool -#define TF_L2_CTXT_REMAP_POOL_NAME_RX l2_ctxt_remap_pool_rx -#define TF_L2_CTXT_REMAP_POOL_NAME_TX l2_ctxt_remap_pool_tx - - -/** HW Resource types - */ -enum tf_resource_type_hw { - /* Common HW resources for all chip variants */ - TF_RESC_TYPE_HW_L2_CTXT_TCAM, - TF_RESC_TYPE_HW_PROF_FUNC, - TF_RESC_TYPE_HW_PROF_TCAM, - TF_RESC_TYPE_HW_EM_PROF_ID, - TF_RESC_TYPE_HW_EM_REC, - TF_RESC_TYPE_HW_WC_TCAM_PROF_ID, - TF_RESC_TYPE_HW_WC_TCAM, - TF_RESC_TYPE_HW_METER_PROF, - TF_RESC_TYPE_HW_METER_INST, - TF_RESC_TYPE_HW_MIRROR, - TF_RESC_TYPE_HW_UPAR, - /* Wh+/SR specific HW resources */ - TF_RESC_TYPE_HW_SP_TCAM, - /* SR/SR2 specific HW resources */ - TF_RESC_TYPE_HW_L2_FUNC, - /* Thor, SR2 common HW resources */ - TF_RESC_TYPE_HW_FKB, - /* SR2 specific HW resources */ - TF_RESC_TYPE_HW_TBL_SCOPE, - TF_RESC_TYPE_HW_EPOCH0, - TF_RESC_TYPE_HW_EPOCH1, - TF_RESC_TYPE_HW_METADATA, - TF_RESC_TYPE_HW_CT_STATE, - TF_RESC_TYPE_HW_RANGE_PROF, - TF_RESC_TYPE_HW_RANGE_ENTRY, - TF_RESC_TYPE_HW_LAG_ENTRY, - TF_RESC_TYPE_HW_MAX -}; - -/** HW Resource types - */ -enum tf_resource_type_sram { - TF_RESC_TYPE_SRAM_FULL_ACTION, - TF_RESC_TYPE_SRAM_MCG, - TF_RESC_TYPE_SRAM_ENCAP_8B, - TF_RESC_TYPE_SRAM_ENCAP_16B, - TF_RESC_TYPE_SRAM_ENCAP_64B, - TF_RESC_TYPE_SRAM_SP_SMAC, - TF_RESC_TYPE_SRAM_SP_SMAC_IPV4, - TF_RESC_TYPE_SRAM_SP_SMAC_IPV6, - TF_RESC_TYPE_SRAM_COUNTER_64B, - TF_RESC_TYPE_SRAM_NAT_SPORT, - TF_RESC_TYPE_SRAM_NAT_DPORT, - TF_RESC_TYPE_SRAM_NAT_S_IPV4, - TF_RESC_TYPE_SRAM_NAT_D_IPV4, - TF_RESC_TYPE_SRAM_MAX -}; #endif /* _TF_RESOURCES_H_ */ diff --git a/drivers/net/bnxt/tf_core/tf_rm.c b/drivers/net/bnxt/tf_core/tf_rm.c index e0a84e64d3..e0469b6532 100644 --- a/drivers/net/bnxt/tf_core/tf_rm.c +++ b/drivers/net/bnxt/tf_core/tf_rm.c @@ -7,3171 +7,916 @@ #include +#include + #include "tf_rm.h" -#include "tf_core.h" +#include "tf_common.h" #include "tf_util.h" #include "tf_session.h" -#include "tf_resources.h" -#include "tf_msg.h" -#include "bnxt.h" +#include "tf_device.h" #include "tfp.h" +#include "tf_msg.h" /** - * Internal macro to perform HW resource allocation check between what - * firmware reports vs what was statically requested. - * - * Parameters: - * struct tf_rm_hw_query *hquery - Pointer to the hw query result - * enum tf_dir dir - Direction to process - * enum tf_resource_type_hw hcapi_type - HCAPI type, the index element - * in the hw query structure - * define def_value - Define value to check against - * uint32_t *eflag - Result of the check - */ -#define TF_RM_CHECK_HW_ALLOC(hquery, dir, hcapi_type, def_value, eflag) do { \ - if ((dir) == TF_DIR_RX) { \ - if ((hquery)->hw_query[(hcapi_type)].max != def_value ## _RX) \ - *(eflag) |= 1 << (hcapi_type); \ - } else { \ - if ((hquery)->hw_query[(hcapi_type)].max != def_value ## _TX) \ - *(eflag) |= 1 << (hcapi_type); \ - } \ -} while (0) - -/** - * Internal macro to perform HW resource allocation check between what - * firmware reports vs what was statically requested. - * - * Parameters: - * struct tf_rm_sram_query *squery - Pointer to the sram query result - * enum tf_dir dir - Direction to process - * enum tf_resource_type_sram hcapi_type - HCAPI type, the index element - * in the hw query structure - * define def_value - Define value to check against - * uint32_t *eflag - Result of the check - */ -#define TF_RM_CHECK_SRAM_ALLOC(squery, dir, hcapi_type, def_value, eflag) do { \ - if ((dir) == TF_DIR_RX) { \ - if ((squery)->sram_query[(hcapi_type)].max != def_value ## _RX)\ - *(eflag) |= 1 << (hcapi_type); \ - } else { \ - if ((squery)->sram_query[(hcapi_type)].max != def_value ## _TX)\ - *(eflag) |= 1 << (hcapi_type); \ - } \ -} while (0) - -/** - * Internal macro to convert a reserved resource define name to be - * direction specific. - * - * Parameters: - * enum tf_dir dir - Direction to process - * string type - Type name to append RX or TX to - * string dtype - Direction specific type - * - * + * Generic RM Element data type that an RM DB is build upon. */ -#define TF_RESC_RSVD(dir, type, dtype) do { \ - if ((dir) == TF_DIR_RX) \ - (dtype) = type ## _RX; \ - else \ - (dtype) = type ## _TX; \ - } while (0) - -const char -*tf_hcapi_hw_2_str(enum tf_resource_type_hw hw_type) -{ - switch (hw_type) { - case TF_RESC_TYPE_HW_L2_CTXT_TCAM: - return "L2 ctxt tcam"; - case TF_RESC_TYPE_HW_PROF_FUNC: - return "Profile Func"; - case TF_RESC_TYPE_HW_PROF_TCAM: - return "Profile tcam"; - case TF_RESC_TYPE_HW_EM_PROF_ID: - return "EM profile id"; - case TF_RESC_TYPE_HW_EM_REC: - return "EM record"; - case TF_RESC_TYPE_HW_WC_TCAM_PROF_ID: - return "WC tcam profile id"; - case TF_RESC_TYPE_HW_WC_TCAM: - return "WC tcam"; - case TF_RESC_TYPE_HW_METER_PROF: - return "Meter profile"; - case TF_RESC_TYPE_HW_METER_INST: - return "Meter instance"; - case TF_RESC_TYPE_HW_MIRROR: - return "Mirror"; - case TF_RESC_TYPE_HW_UPAR: - return "UPAR"; - case TF_RESC_TYPE_HW_SP_TCAM: - return "Source properties tcam"; - case TF_RESC_TYPE_HW_L2_FUNC: - return "L2 Function"; - case TF_RESC_TYPE_HW_FKB: - return "FKB"; - case TF_RESC_TYPE_HW_TBL_SCOPE: - return "Table scope"; - case TF_RESC_TYPE_HW_EPOCH0: - return "EPOCH0"; - case TF_RESC_TYPE_HW_EPOCH1: - return "EPOCH1"; - case TF_RESC_TYPE_HW_METADATA: - return "Metadata"; - case TF_RESC_TYPE_HW_CT_STATE: - return "Connection tracking state"; - case TF_RESC_TYPE_HW_RANGE_PROF: - return "Range profile"; - case TF_RESC_TYPE_HW_RANGE_ENTRY: - return "Range entry"; - case TF_RESC_TYPE_HW_LAG_ENTRY: - return "LAG"; - default: - return "Invalid identifier"; - } -} - -const char -*tf_hcapi_sram_2_str(enum tf_resource_type_sram sram_type) -{ - switch (sram_type) { - case TF_RESC_TYPE_SRAM_FULL_ACTION: - return "Full action"; - case TF_RESC_TYPE_SRAM_MCG: - return "MCG"; - case TF_RESC_TYPE_SRAM_ENCAP_8B: - return "Encap 8B"; - case TF_RESC_TYPE_SRAM_ENCAP_16B: - return "Encap 16B"; - case TF_RESC_TYPE_SRAM_ENCAP_64B: - return "Encap 64B"; - case TF_RESC_TYPE_SRAM_SP_SMAC: - return "Source properties SMAC"; - case TF_RESC_TYPE_SRAM_SP_SMAC_IPV4: - return "Source properties SMAC IPv4"; - case TF_RESC_TYPE_SRAM_SP_SMAC_IPV6: - return "Source properties IPv6"; - case TF_RESC_TYPE_SRAM_COUNTER_64B: - return "Counter 64B"; - case TF_RESC_TYPE_SRAM_NAT_SPORT: - return "NAT source port"; - case TF_RESC_TYPE_SRAM_NAT_DPORT: - return "NAT destination port"; - case TF_RESC_TYPE_SRAM_NAT_S_IPV4: - return "NAT source IPv4"; - case TF_RESC_TYPE_SRAM_NAT_D_IPV4: - return "NAT destination IPv4"; - default: - return "Invalid identifier"; - } -} +struct tf_rm_element { + /** + * RM Element configuration type. If Private then the + * hcapi_type can be ignored. If Null then the element is not + * valid for the device. + */ + enum tf_rm_elem_cfg_type cfg_type; -/** - * Helper function to perform a HW HCAPI resource type lookup against - * the reserved value of the same static type. - * - * Returns: - * -EOPNOTSUPP - Reserved resource type not supported - * Value - Integer value of the reserved value for the requested type - */ -static int -tf_rm_rsvd_hw_value(enum tf_dir dir, enum tf_resource_type_hw index) -{ - uint32_t value = -EOPNOTSUPP; + /** + * HCAPI RM Type for the element. + */ + uint16_t hcapi_type; - switch (index) { - case TF_RESC_TYPE_HW_L2_CTXT_TCAM: - TF_RESC_RSVD(dir, TF_RSVD_L2_CTXT_TCAM, value); - break; - case TF_RESC_TYPE_HW_PROF_FUNC: - TF_RESC_RSVD(dir, TF_RSVD_PROF_FUNC, value); - break; - case TF_RESC_TYPE_HW_PROF_TCAM: - TF_RESC_RSVD(dir, TF_RSVD_PROF_TCAM, value); - break; - case TF_RESC_TYPE_HW_EM_PROF_ID: - TF_RESC_RSVD(dir, TF_RSVD_EM_PROF_ID, value); - break; - case TF_RESC_TYPE_HW_EM_REC: - TF_RESC_RSVD(dir, TF_RSVD_EM_REC, value); - break; - case TF_RESC_TYPE_HW_WC_TCAM_PROF_ID: - TF_RESC_RSVD(dir, TF_RSVD_WC_TCAM_PROF_ID, value); - break; - case TF_RESC_TYPE_HW_WC_TCAM: - TF_RESC_RSVD(dir, TF_RSVD_WC_TCAM, value); - break; - case TF_RESC_TYPE_HW_METER_PROF: - TF_RESC_RSVD(dir, TF_RSVD_METER_PROF, value); - break; - case TF_RESC_TYPE_HW_METER_INST: - TF_RESC_RSVD(dir, TF_RSVD_METER_INST, value); - break; - case TF_RESC_TYPE_HW_MIRROR: - TF_RESC_RSVD(dir, TF_RSVD_MIRROR, value); - break; - case TF_RESC_TYPE_HW_UPAR: - TF_RESC_RSVD(dir, TF_RSVD_UPAR, value); - break; - case TF_RESC_TYPE_HW_SP_TCAM: - TF_RESC_RSVD(dir, TF_RSVD_SP_TCAM, value); - break; - case TF_RESC_TYPE_HW_L2_FUNC: - TF_RESC_RSVD(dir, TF_RSVD_L2_FUNC, value); - break; - case TF_RESC_TYPE_HW_FKB: - TF_RESC_RSVD(dir, TF_RSVD_FKB, value); - break; - case TF_RESC_TYPE_HW_TBL_SCOPE: - TF_RESC_RSVD(dir, TF_RSVD_TBL_SCOPE, value); - break; - case TF_RESC_TYPE_HW_EPOCH0: - TF_RESC_RSVD(dir, TF_RSVD_EPOCH0, value); - break; - case TF_RESC_TYPE_HW_EPOCH1: - TF_RESC_RSVD(dir, TF_RSVD_EPOCH1, value); - break; - case TF_RESC_TYPE_HW_METADATA: - TF_RESC_RSVD(dir, TF_RSVD_METADATA, value); - break; - case TF_RESC_TYPE_HW_CT_STATE: - TF_RESC_RSVD(dir, TF_RSVD_CT_STATE, value); - break; - case TF_RESC_TYPE_HW_RANGE_PROF: - TF_RESC_RSVD(dir, TF_RSVD_RANGE_PROF, value); - break; - case TF_RESC_TYPE_HW_RANGE_ENTRY: - TF_RESC_RSVD(dir, TF_RSVD_RANGE_ENTRY, value); - break; - case TF_RESC_TYPE_HW_LAG_ENTRY: - TF_RESC_RSVD(dir, TF_RSVD_LAG_ENTRY, value); - break; - default: - break; - } + /** + * HCAPI RM allocated range information for the element. + */ + struct tf_rm_alloc_info alloc; - return value; -} + /** + * Bit allocator pool for the element. Pool size is controlled + * by the struct tf_session_resources at time of session creation. + * Null indicates that the element is not used for the device. + */ + struct bitalloc *pool; +}; /** - * Helper function to perform a SRAM HCAPI resource type lookup - * against the reserved value of the same static type. - * - * Returns: - * -EOPNOTSUPP - Reserved resource type not supported - * Value - Integer value of the reserved value for the requested type + * TF RM DB definition */ -static int -tf_rm_rsvd_sram_value(enum tf_dir dir, enum tf_resource_type_sram index) -{ - uint32_t value = -EOPNOTSUPP; - - switch (index) { - case TF_RESC_TYPE_SRAM_FULL_ACTION: - TF_RESC_RSVD(dir, TF_RSVD_SRAM_FULL_ACTION, value); - break; - case TF_RESC_TYPE_SRAM_MCG: - TF_RESC_RSVD(dir, TF_RSVD_SRAM_MCG, value); - break; - case TF_RESC_TYPE_SRAM_ENCAP_8B: - TF_RESC_RSVD(dir, TF_RSVD_SRAM_ENCAP_8B, value); - break; - case TF_RESC_TYPE_SRAM_ENCAP_16B: - TF_RESC_RSVD(dir, TF_RSVD_SRAM_ENCAP_16B, value); - break; - case TF_RESC_TYPE_SRAM_ENCAP_64B: - TF_RESC_RSVD(dir, TF_RSVD_SRAM_ENCAP_64B, value); - break; - case TF_RESC_TYPE_SRAM_SP_SMAC: - TF_RESC_RSVD(dir, TF_RSVD_SRAM_SP_SMAC, value); - break; - case TF_RESC_TYPE_SRAM_SP_SMAC_IPV4: - TF_RESC_RSVD(dir, TF_RSVD_SRAM_SP_SMAC_IPV4, value); - break; - case TF_RESC_TYPE_SRAM_SP_SMAC_IPV6: - TF_RESC_RSVD(dir, TF_RSVD_SRAM_SP_SMAC_IPV6, value); - break; - case TF_RESC_TYPE_SRAM_COUNTER_64B: - TF_RESC_RSVD(dir, TF_RSVD_SRAM_COUNTER_64B, value); - break; - case TF_RESC_TYPE_SRAM_NAT_SPORT: - TF_RESC_RSVD(dir, TF_RSVD_SRAM_NAT_SPORT, value); - break; - case TF_RESC_TYPE_SRAM_NAT_DPORT: - TF_RESC_RSVD(dir, TF_RSVD_SRAM_NAT_DPORT, value); - break; - case TF_RESC_TYPE_SRAM_NAT_S_IPV4: - TF_RESC_RSVD(dir, TF_RSVD_SRAM_NAT_S_IPV4, value); - break; - case TF_RESC_TYPE_SRAM_NAT_D_IPV4: - TF_RESC_RSVD(dir, TF_RSVD_SRAM_NAT_D_IPV4, value); - break; - default: - break; - } - - return value; -} +struct tf_rm_new_db { + /** + * Number of elements in the DB + */ + uint16_t num_entries; -/** - * Helper function to print all the HW resource qcaps errors reported - * in the error_flag. - * - * [in] dir - * Receive or transmit direction - * - * [in] error_flag - * Pointer to the hw error flags created at time of the query check - */ -static void -tf_rm_print_hw_qcaps_error(enum tf_dir dir, - struct tf_rm_hw_query *hw_query, - uint32_t *error_flag) -{ - int i; + /** + * Direction this DB controls. + */ + enum tf_dir dir; - TFP_DRV_LOG(ERR, "QCAPS errors HW\n"); - TFP_DRV_LOG(ERR, " Direction: %s\n", tf_dir_2_str(dir)); - TFP_DRV_LOG(ERR, " Elements:\n"); + /** + * Module type, used for logging purposes. + */ + enum tf_device_module_type type; - for (i = 0; i < TF_RESC_TYPE_HW_MAX; i++) { - if (*error_flag & 1 << i) - TFP_DRV_LOG(ERR, " %s, %d elem available, req:%d\n", - tf_hcapi_hw_2_str(i), - hw_query->hw_query[i].max, - tf_rm_rsvd_hw_value(dir, i)); - } -} + /** + * The DB consists of an array of elements + */ + struct tf_rm_element *db; +}; /** - * Helper function to print all the SRAM resource qcaps errors - * reported in the error_flag. + * Adjust an index according to the allocation information. * - * [in] dir - * Receive or transmit direction + * All resources are controlled in a 0 based pool. Some resources, by + * design, are not 0 based, i.e. Full Action Records (SRAM) thus they + * need to be adjusted before they are handed out. * - * [in] error_flag - * Pointer to the sram error flags created at time of the query check - */ -static void -tf_rm_print_sram_qcaps_error(enum tf_dir dir, - struct tf_rm_sram_query *sram_query, - uint32_t *error_flag) -{ - int i; - - TFP_DRV_LOG(ERR, "QCAPS errors SRAM\n"); - TFP_DRV_LOG(ERR, " Direction: %s\n", tf_dir_2_str(dir)); - TFP_DRV_LOG(ERR, " Elements:\n"); - - for (i = 0; i < TF_RESC_TYPE_SRAM_MAX; i++) { - if (*error_flag & 1 << i) - TFP_DRV_LOG(ERR, " %s, %d elem available, req:%d\n", - tf_hcapi_sram_2_str(i), - sram_query->sram_query[i].max, - tf_rm_rsvd_sram_value(dir, i)); - } -} - -/** - * Performs a HW resource check between what firmware capability - * reports and what the core expects is available. + * [in] cfg + * Pointer to the DB configuration * - * Firmware performs the resource carving at AFM init time and the - * resource capability is reported in the TruFlow qcaps msg. + * [in] reservations + * Pointer to the allocation values associated with the module * - * [in] query - * Pointer to HW Query data structure. Query holds what the firmware - * offers of the HW resources. + * [in] count + * Number of DB configuration elements * - * [in] dir - * Receive or transmit direction - * - * [in/out] error_flag - * Pointer to a bit array indicating the error of a single HCAPI - * resource type. When a bit is set to 1, the HCAPI resource type - * failed static allocation. + * [out] valid_count + * Number of HCAPI entries with a reservation value greater than 0 * * Returns: - * 0 - Success - * -ENOMEM - Failure on one of the allocated resources. Check the - * error_flag for what types are flagged errored. - */ -static int -tf_rm_check_hw_qcaps_static(struct tf_rm_hw_query *query, - enum tf_dir dir, - uint32_t *error_flag) -{ - *error_flag = 0; - - TF_RM_CHECK_HW_ALLOC(query, - dir, - TF_RESC_TYPE_HW_L2_CTXT_TCAM, - TF_RSVD_L2_CTXT_TCAM, - error_flag); - - TF_RM_CHECK_HW_ALLOC(query, - dir, - TF_RESC_TYPE_HW_PROF_FUNC, - TF_RSVD_PROF_FUNC, - error_flag); - - TF_RM_CHECK_HW_ALLOC(query, - dir, - TF_RESC_TYPE_HW_PROF_TCAM, - TF_RSVD_PROF_TCAM, - error_flag); - - TF_RM_CHECK_HW_ALLOC(query, - dir, - TF_RESC_TYPE_HW_EM_PROF_ID, - TF_RSVD_EM_PROF_ID, - error_flag); - - TF_RM_CHECK_HW_ALLOC(query, - dir, - TF_RESC_TYPE_HW_EM_REC, - TF_RSVD_EM_REC, - error_flag); - - TF_RM_CHECK_HW_ALLOC(query, - dir, - TF_RESC_TYPE_HW_WC_TCAM_PROF_ID, - TF_RSVD_WC_TCAM_PROF_ID, - error_flag); - - TF_RM_CHECK_HW_ALLOC(query, - dir, - TF_RESC_TYPE_HW_WC_TCAM, - TF_RSVD_WC_TCAM, - error_flag); - - TF_RM_CHECK_HW_ALLOC(query, - dir, - TF_RESC_TYPE_HW_METER_PROF, - TF_RSVD_METER_PROF, - error_flag); - - TF_RM_CHECK_HW_ALLOC(query, - dir, - TF_RESC_TYPE_HW_METER_INST, - TF_RSVD_METER_INST, - error_flag); - - TF_RM_CHECK_HW_ALLOC(query, - dir, - TF_RESC_TYPE_HW_MIRROR, - TF_RSVD_MIRROR, - error_flag); - - TF_RM_CHECK_HW_ALLOC(query, - dir, - TF_RESC_TYPE_HW_UPAR, - TF_RSVD_UPAR, - error_flag); - - TF_RM_CHECK_HW_ALLOC(query, - dir, - TF_RESC_TYPE_HW_SP_TCAM, - TF_RSVD_SP_TCAM, - error_flag); - - TF_RM_CHECK_HW_ALLOC(query, - dir, - TF_RESC_TYPE_HW_L2_FUNC, - TF_RSVD_L2_FUNC, - error_flag); - - TF_RM_CHECK_HW_ALLOC(query, - dir, - TF_RESC_TYPE_HW_FKB, - TF_RSVD_FKB, - error_flag); - - TF_RM_CHECK_HW_ALLOC(query, - dir, - TF_RESC_TYPE_HW_TBL_SCOPE, - TF_RSVD_TBL_SCOPE, - error_flag); - - TF_RM_CHECK_HW_ALLOC(query, - dir, - TF_RESC_TYPE_HW_EPOCH0, - TF_RSVD_EPOCH0, - error_flag); - - TF_RM_CHECK_HW_ALLOC(query, - dir, - TF_RESC_TYPE_HW_EPOCH1, - TF_RSVD_EPOCH1, - error_flag); - - TF_RM_CHECK_HW_ALLOC(query, - dir, - TF_RESC_TYPE_HW_METADATA, - TF_RSVD_METADATA, - error_flag); - - TF_RM_CHECK_HW_ALLOC(query, - dir, - TF_RESC_TYPE_HW_CT_STATE, - TF_RSVD_CT_STATE, - error_flag); - - TF_RM_CHECK_HW_ALLOC(query, - dir, - TF_RESC_TYPE_HW_RANGE_PROF, - TF_RSVD_RANGE_PROF, - error_flag); - - TF_RM_CHECK_HW_ALLOC(query, - dir, - TF_RESC_TYPE_HW_RANGE_ENTRY, - TF_RSVD_RANGE_ENTRY, - error_flag); - - TF_RM_CHECK_HW_ALLOC(query, - dir, - TF_RESC_TYPE_HW_LAG_ENTRY, - TF_RSVD_LAG_ENTRY, - error_flag); - - if (*error_flag != 0) - return -ENOMEM; - - return 0; -} - -/** - * Performs a SRAM resource check between what firmware capability - * reports and what the core expects is available. - * - * Firmware performs the resource carving at AFM init time and the - * resource capability is reported in the TruFlow qcaps msg. - * - * [in] query - * Pointer to SRAM Query data structure. Query holds what the - * firmware offers of the SRAM resources. - * - * [in] dir - * Receive or transmit direction - * - * [in/out] error_flag - * Pointer to a bit array indicating the error of a single HCAPI - * resource type. When a bit is set to 1, the HCAPI resource type - * failed static allocation. - * - * Returns: - * 0 - Success - * -ENOMEM - Failure on one of the allocated resources. Check the - * error_flag for what types are flagged errored. - */ -static int -tf_rm_check_sram_qcaps_static(struct tf_rm_sram_query *query, - enum tf_dir dir, - uint32_t *error_flag) -{ - *error_flag = 0; - - TF_RM_CHECK_SRAM_ALLOC(query, - dir, - TF_RESC_TYPE_SRAM_FULL_ACTION, - TF_RSVD_SRAM_FULL_ACTION, - error_flag); - - TF_RM_CHECK_SRAM_ALLOC(query, - dir, - TF_RESC_TYPE_SRAM_MCG, - TF_RSVD_SRAM_MCG, - error_flag); - - TF_RM_CHECK_SRAM_ALLOC(query, - dir, - TF_RESC_TYPE_SRAM_ENCAP_8B, - TF_RSVD_SRAM_ENCAP_8B, - error_flag); - - TF_RM_CHECK_SRAM_ALLOC(query, - dir, - TF_RESC_TYPE_SRAM_ENCAP_16B, - TF_RSVD_SRAM_ENCAP_16B, - error_flag); - - TF_RM_CHECK_SRAM_ALLOC(query, - dir, - TF_RESC_TYPE_SRAM_ENCAP_64B, - TF_RSVD_SRAM_ENCAP_64B, - error_flag); - - TF_RM_CHECK_SRAM_ALLOC(query, - dir, - TF_RESC_TYPE_SRAM_SP_SMAC, - TF_RSVD_SRAM_SP_SMAC, - error_flag); - - TF_RM_CHECK_SRAM_ALLOC(query, - dir, - TF_RESC_TYPE_SRAM_SP_SMAC_IPV4, - TF_RSVD_SRAM_SP_SMAC_IPV4, - error_flag); - - TF_RM_CHECK_SRAM_ALLOC(query, - dir, - TF_RESC_TYPE_SRAM_SP_SMAC_IPV6, - TF_RSVD_SRAM_SP_SMAC_IPV6, - error_flag); - - TF_RM_CHECK_SRAM_ALLOC(query, - dir, - TF_RESC_TYPE_SRAM_COUNTER_64B, - TF_RSVD_SRAM_COUNTER_64B, - error_flag); - - TF_RM_CHECK_SRAM_ALLOC(query, - dir, - TF_RESC_TYPE_SRAM_NAT_SPORT, - TF_RSVD_SRAM_NAT_SPORT, - error_flag); - - TF_RM_CHECK_SRAM_ALLOC(query, - dir, - TF_RESC_TYPE_SRAM_NAT_DPORT, - TF_RSVD_SRAM_NAT_DPORT, - error_flag); - - TF_RM_CHECK_SRAM_ALLOC(query, - dir, - TF_RESC_TYPE_SRAM_NAT_S_IPV4, - TF_RSVD_SRAM_NAT_S_IPV4, - error_flag); - - TF_RM_CHECK_SRAM_ALLOC(query, - dir, - TF_RESC_TYPE_SRAM_NAT_D_IPV4, - TF_RSVD_SRAM_NAT_D_IPV4, - error_flag); - - if (*error_flag != 0) - return -ENOMEM; - - return 0; -} - -/** - * Internal function to mark pool entries used. + * 0 - Success + * - EOPNOTSUPP - Operation not supported */ static void -tf_rm_reserve_range(uint32_t count, - uint32_t rsv_begin, - uint32_t rsv_end, - uint32_t max, - struct bitalloc *pool) +tf_rm_count_hcapi_reservations(enum tf_dir dir, + enum tf_device_module_type type, + struct tf_rm_element_cfg *cfg, + uint16_t *reservations, + uint16_t count, + uint16_t *valid_count) { - uint32_t i; + int i; + uint16_t cnt = 0; - /* If no resources has been requested we mark everything - * 'used' - */ - if (count == 0) { - for (i = 0; i < max; i++) - ba_alloc_index(pool, i); - } else { - /* Support 2 main modes - * Reserved range starts from bottom up (with - * pre-reserved value or not) - * - begin = 0 to end xx - * - begin = 1 to end xx - * - * Reserved range starts from top down - * - begin = yy to end max - */ + for (i = 0; i < count; i++) { + if (cfg[i].cfg_type == TF_RM_ELEM_CFG_HCAPI && + reservations[i] > 0) + cnt++; - /* Bottom up check, start from 0 */ - if (rsv_begin == 0) { - for (i = rsv_end + 1; i < max; i++) - ba_alloc_index(pool, i); - } - - /* Bottom up check, start from 1 or higher OR - * Top Down + /* Only log msg if a type is attempted reserved and + * not supported. We ignore EM module as its using a + * split configuration array thus it would fail for + * this type of check. */ - if (rsv_begin >= 1) { - /* Allocate from 0 until start */ - for (i = 0; i < rsv_begin; i++) - ba_alloc_index(pool, i); - - /* Skip and then do the remaining */ - if (rsv_end < max - 1) { - for (i = rsv_end; i < max; i++) - ba_alloc_index(pool, i); - } - } - } -} - -/** - * Internal function to mark all the l2 ctxt allocated that Truflow - * does not own. - */ -static void -tf_rm_rsvd_l2_ctxt(struct tf_session *tfs) -{ - uint32_t index = TF_RESC_TYPE_HW_L2_CTXT_TCAM; - uint32_t end = 0; - - /* l2 ctxt rx direction */ - if (tfs->resc.rx.hw_entry[index].stride > 0) - end = tfs->resc.rx.hw_entry[index].start + - tfs->resc.rx.hw_entry[index].stride - 1; - - tf_rm_reserve_range(tfs->resc.rx.hw_entry[index].stride, - tfs->resc.rx.hw_entry[index].start, - end, - TF_NUM_L2_CTXT_TCAM, - tfs->TF_L2_CTXT_TCAM_POOL_NAME_RX); - - /* l2 ctxt tx direction */ - if (tfs->resc.tx.hw_entry[index].stride > 0) - end = tfs->resc.tx.hw_entry[index].start + - tfs->resc.tx.hw_entry[index].stride - 1; - - tf_rm_reserve_range(tfs->resc.tx.hw_entry[index].stride, - tfs->resc.tx.hw_entry[index].start, - end, - TF_NUM_L2_CTXT_TCAM, - tfs->TF_L2_CTXT_TCAM_POOL_NAME_TX); -} - -/** - * Internal function to mark all the profile tcam and profile func - * resources that Truflow does not own. - */ -static void -tf_rm_rsvd_prof(struct tf_session *tfs) -{ - uint32_t index = TF_RESC_TYPE_HW_PROF_FUNC; - uint32_t end = 0; - - /* profile func rx direction */ - if (tfs->resc.rx.hw_entry[index].stride > 0) - end = tfs->resc.rx.hw_entry[index].start + - tfs->resc.rx.hw_entry[index].stride - 1; - - tf_rm_reserve_range(tfs->resc.rx.hw_entry[index].stride, - tfs->resc.rx.hw_entry[index].start, - end, - TF_NUM_PROF_FUNC, - tfs->TF_PROF_FUNC_POOL_NAME_RX); - - /* profile func tx direction */ - if (tfs->resc.tx.hw_entry[index].stride > 0) - end = tfs->resc.tx.hw_entry[index].start + - tfs->resc.tx.hw_entry[index].stride - 1; - - tf_rm_reserve_range(tfs->resc.tx.hw_entry[index].stride, - tfs->resc.tx.hw_entry[index].start, - end, - TF_NUM_PROF_FUNC, - tfs->TF_PROF_FUNC_POOL_NAME_TX); - - index = TF_RESC_TYPE_HW_PROF_TCAM; - - /* profile tcam rx direction */ - if (tfs->resc.rx.hw_entry[index].stride > 0) - end = tfs->resc.rx.hw_entry[index].start + - tfs->resc.rx.hw_entry[index].stride - 1; - - tf_rm_reserve_range(tfs->resc.rx.hw_entry[index].stride, - tfs->resc.rx.hw_entry[index].start, - end, - TF_NUM_PROF_TCAM, - tfs->TF_PROF_TCAM_POOL_NAME_RX); - - /* profile tcam tx direction */ - if (tfs->resc.tx.hw_entry[index].stride > 0) - end = tfs->resc.tx.hw_entry[index].start + - tfs->resc.tx.hw_entry[index].stride - 1; - - tf_rm_reserve_range(tfs->resc.tx.hw_entry[index].stride, - tfs->resc.tx.hw_entry[index].start, - end, - TF_NUM_PROF_TCAM, - tfs->TF_PROF_TCAM_POOL_NAME_TX); -} - -/** - * Internal function to mark all the em profile id allocated that - * Truflow does not own. - */ -static void -tf_rm_rsvd_em_prof(struct tf_session *tfs) -{ - uint32_t index = TF_RESC_TYPE_HW_EM_PROF_ID; - uint32_t end = 0; - - /* em prof id rx direction */ - if (tfs->resc.rx.hw_entry[index].stride > 0) - end = tfs->resc.rx.hw_entry[index].start + - tfs->resc.rx.hw_entry[index].stride - 1; - - tf_rm_reserve_range(tfs->resc.rx.hw_entry[index].stride, - tfs->resc.rx.hw_entry[index].start, - end, - TF_NUM_EM_PROF_ID, - tfs->TF_EM_PROF_ID_POOL_NAME_RX); - - /* em prof id tx direction */ - if (tfs->resc.tx.hw_entry[index].stride > 0) - end = tfs->resc.tx.hw_entry[index].start + - tfs->resc.tx.hw_entry[index].stride - 1; - - tf_rm_reserve_range(tfs->resc.tx.hw_entry[index].stride, - tfs->resc.tx.hw_entry[index].start, - end, - TF_NUM_EM_PROF_ID, - tfs->TF_EM_PROF_ID_POOL_NAME_TX); -} - -/** - * Internal function to mark all the wildcard tcam and profile id - * resources that Truflow does not own. - */ -static void -tf_rm_rsvd_wc(struct tf_session *tfs) -{ - uint32_t index = TF_RESC_TYPE_HW_WC_TCAM_PROF_ID; - uint32_t end = 0; - - /* wc profile id rx direction */ - if (tfs->resc.rx.hw_entry[index].stride > 0) - end = tfs->resc.rx.hw_entry[index].start + - tfs->resc.rx.hw_entry[index].stride - 1; - - tf_rm_reserve_range(tfs->resc.rx.hw_entry[index].stride, - tfs->resc.rx.hw_entry[index].start, - end, - TF_NUM_WC_PROF_ID, - tfs->TF_WC_TCAM_PROF_ID_POOL_NAME_RX); - - /* wc profile id tx direction */ - if (tfs->resc.tx.hw_entry[index].stride > 0) - end = tfs->resc.tx.hw_entry[index].start + - tfs->resc.tx.hw_entry[index].stride - 1; - - tf_rm_reserve_range(tfs->resc.tx.hw_entry[index].stride, - tfs->resc.tx.hw_entry[index].start, - end, - TF_NUM_WC_PROF_ID, - tfs->TF_WC_TCAM_PROF_ID_POOL_NAME_TX); - - index = TF_RESC_TYPE_HW_WC_TCAM; - - /* wc tcam rx direction */ - if (tfs->resc.rx.hw_entry[index].stride > 0) - end = tfs->resc.rx.hw_entry[index].start + - tfs->resc.rx.hw_entry[index].stride - 1; - - tf_rm_reserve_range(tfs->resc.rx.hw_entry[index].stride, - tfs->resc.rx.hw_entry[index].start, - end, - TF_NUM_WC_TCAM_ROW, - tfs->TF_WC_TCAM_POOL_NAME_RX); - - /* wc tcam tx direction */ - if (tfs->resc.tx.hw_entry[index].stride > 0) - end = tfs->resc.tx.hw_entry[index].start + - tfs->resc.tx.hw_entry[index].stride - 1; - - tf_rm_reserve_range(tfs->resc.tx.hw_entry[index].stride, - tfs->resc.tx.hw_entry[index].start, - end, - TF_NUM_WC_TCAM_ROW, - tfs->TF_WC_TCAM_POOL_NAME_TX); -} - -/** - * Internal function to mark all the meter resources allocated that - * Truflow does not own. - */ -static void -tf_rm_rsvd_meter(struct tf_session *tfs) -{ - uint32_t index = TF_RESC_TYPE_HW_METER_PROF; - uint32_t end = 0; - - /* meter profiles rx direction */ - if (tfs->resc.rx.hw_entry[index].stride > 0) - end = tfs->resc.rx.hw_entry[index].start + - tfs->resc.rx.hw_entry[index].stride - 1; - - tf_rm_reserve_range(tfs->resc.rx.hw_entry[index].stride, - tfs->resc.rx.hw_entry[index].start, - end, - TF_NUM_METER_PROF, - tfs->TF_METER_PROF_POOL_NAME_RX); - - /* meter profiles tx direction */ - if (tfs->resc.tx.hw_entry[index].stride > 0) - end = tfs->resc.tx.hw_entry[index].start + - tfs->resc.tx.hw_entry[index].stride - 1; - - tf_rm_reserve_range(tfs->resc.tx.hw_entry[index].stride, - tfs->resc.tx.hw_entry[index].start, - end, - TF_NUM_METER_PROF, - tfs->TF_METER_PROF_POOL_NAME_TX); - - index = TF_RESC_TYPE_HW_METER_INST; - - /* meter rx direction */ - if (tfs->resc.rx.hw_entry[index].stride > 0) - end = tfs->resc.rx.hw_entry[index].start + - tfs->resc.rx.hw_entry[index].stride - 1; - - tf_rm_reserve_range(tfs->resc.rx.hw_entry[index].stride, - tfs->resc.rx.hw_entry[index].start, - end, - TF_NUM_METER, - tfs->TF_METER_INST_POOL_NAME_RX); - - /* meter tx direction */ - if (tfs->resc.tx.hw_entry[index].stride > 0) - end = tfs->resc.tx.hw_entry[index].start + - tfs->resc.tx.hw_entry[index].stride - 1; - - tf_rm_reserve_range(tfs->resc.tx.hw_entry[index].stride, - tfs->resc.tx.hw_entry[index].start, - end, - TF_NUM_METER, - tfs->TF_METER_INST_POOL_NAME_TX); -} - -/** - * Internal function to mark all the mirror resources allocated that - * Truflow does not own. - */ -static void -tf_rm_rsvd_mirror(struct tf_session *tfs) -{ - uint32_t index = TF_RESC_TYPE_HW_MIRROR; - uint32_t end = 0; - - /* mirror rx direction */ - if (tfs->resc.rx.hw_entry[index].stride > 0) - end = tfs->resc.rx.hw_entry[index].start + - tfs->resc.rx.hw_entry[index].stride - 1; - - tf_rm_reserve_range(tfs->resc.rx.hw_entry[index].stride, - tfs->resc.rx.hw_entry[index].start, - end, - TF_NUM_MIRROR, - tfs->TF_MIRROR_POOL_NAME_RX); - - /* mirror tx direction */ - if (tfs->resc.tx.hw_entry[index].stride > 0) - end = tfs->resc.tx.hw_entry[index].start + - tfs->resc.tx.hw_entry[index].stride - 1; - - tf_rm_reserve_range(tfs->resc.tx.hw_entry[index].stride, - tfs->resc.tx.hw_entry[index].start, - end, - TF_NUM_MIRROR, - tfs->TF_MIRROR_POOL_NAME_TX); -} - -/** - * Internal function to mark all the upar resources allocated that - * Truflow does not own. - */ -static void -tf_rm_rsvd_upar(struct tf_session *tfs) -{ - uint32_t index = TF_RESC_TYPE_HW_UPAR; - uint32_t end = 0; - - /* upar rx direction */ - if (tfs->resc.rx.hw_entry[index].stride > 0) - end = tfs->resc.rx.hw_entry[index].start + - tfs->resc.rx.hw_entry[index].stride - 1; - - tf_rm_reserve_range(tfs->resc.rx.hw_entry[index].stride, - tfs->resc.rx.hw_entry[index].start, - end, - TF_NUM_UPAR, - tfs->TF_UPAR_POOL_NAME_RX); - - /* upar tx direction */ - if (tfs->resc.tx.hw_entry[index].stride > 0) - end = tfs->resc.tx.hw_entry[index].start + - tfs->resc.tx.hw_entry[index].stride - 1; - - tf_rm_reserve_range(tfs->resc.tx.hw_entry[index].stride, - tfs->resc.tx.hw_entry[index].start, - end, - TF_NUM_UPAR, - tfs->TF_UPAR_POOL_NAME_TX); -} - -/** - * Internal function to mark all the sp tcam resources allocated that - * Truflow does not own. - */ -static void -tf_rm_rsvd_sp_tcam(struct tf_session *tfs) -{ - uint32_t index = TF_RESC_TYPE_HW_SP_TCAM; - uint32_t end = 0; - - /* sp tcam rx direction */ - if (tfs->resc.rx.hw_entry[index].stride > 0) - end = tfs->resc.rx.hw_entry[index].start + - tfs->resc.rx.hw_entry[index].stride - 1; - - tf_rm_reserve_range(tfs->resc.rx.hw_entry[index].stride, - tfs->resc.rx.hw_entry[index].start, - end, - TF_NUM_SP_TCAM, - tfs->TF_SP_TCAM_POOL_NAME_RX); - - /* sp tcam tx direction */ - if (tfs->resc.tx.hw_entry[index].stride > 0) - end = tfs->resc.tx.hw_entry[index].start + - tfs->resc.tx.hw_entry[index].stride - 1; - - tf_rm_reserve_range(tfs->resc.tx.hw_entry[index].stride, - tfs->resc.tx.hw_entry[index].start, - end, - TF_NUM_SP_TCAM, - tfs->TF_SP_TCAM_POOL_NAME_TX); -} - -/** - * Internal function to mark all the l2 func resources allocated that - * Truflow does not own. - */ -static void -tf_rm_rsvd_l2_func(struct tf_session *tfs) -{ - uint32_t index = TF_RESC_TYPE_HW_L2_FUNC; - uint32_t end = 0; - - /* l2 func rx direction */ - if (tfs->resc.rx.hw_entry[index].stride > 0) - end = tfs->resc.rx.hw_entry[index].start + - tfs->resc.rx.hw_entry[index].stride - 1; - - tf_rm_reserve_range(tfs->resc.rx.hw_entry[index].stride, - tfs->resc.rx.hw_entry[index].start, - end, - TF_NUM_L2_FUNC, - tfs->TF_L2_FUNC_POOL_NAME_RX); - - /* l2 func tx direction */ - if (tfs->resc.tx.hw_entry[index].stride > 0) - end = tfs->resc.tx.hw_entry[index].start + - tfs->resc.tx.hw_entry[index].stride - 1; - - tf_rm_reserve_range(tfs->resc.tx.hw_entry[index].stride, - tfs->resc.tx.hw_entry[index].start, - end, - TF_NUM_L2_FUNC, - tfs->TF_L2_FUNC_POOL_NAME_TX); -} - -/** - * Internal function to mark all the fkb resources allocated that - * Truflow does not own. - */ -static void -tf_rm_rsvd_fkb(struct tf_session *tfs) -{ - uint32_t index = TF_RESC_TYPE_HW_FKB; - uint32_t end = 0; - - /* fkb rx direction */ - if (tfs->resc.rx.hw_entry[index].stride > 0) - end = tfs->resc.rx.hw_entry[index].start + - tfs->resc.rx.hw_entry[index].stride - 1; - - tf_rm_reserve_range(tfs->resc.rx.hw_entry[index].stride, - tfs->resc.rx.hw_entry[index].start, - end, - TF_NUM_FKB, - tfs->TF_FKB_POOL_NAME_RX); - - /* fkb tx direction */ - if (tfs->resc.tx.hw_entry[index].stride > 0) - end = tfs->resc.tx.hw_entry[index].start + - tfs->resc.tx.hw_entry[index].stride - 1; - - tf_rm_reserve_range(tfs->resc.tx.hw_entry[index].stride, - tfs->resc.tx.hw_entry[index].start, - end, - TF_NUM_FKB, - tfs->TF_FKB_POOL_NAME_TX); -} - -/** - * Internal function to mark all the tbld scope resources allocated - * that Truflow does not own. - */ -static void -tf_rm_rsvd_tbl_scope(struct tf_session *tfs) -{ - uint32_t index = TF_RESC_TYPE_HW_TBL_SCOPE; - uint32_t end = 0; - - /* tbl scope rx direction */ - if (tfs->resc.rx.hw_entry[index].stride > 0) - end = tfs->resc.rx.hw_entry[index].start + - tfs->resc.rx.hw_entry[index].stride - 1; - - tf_rm_reserve_range(tfs->resc.rx.hw_entry[index].stride, - tfs->resc.rx.hw_entry[index].start, - end, - TF_NUM_TBL_SCOPE, - tfs->TF_TBL_SCOPE_POOL_NAME_RX); - - /* tbl scope tx direction */ - if (tfs->resc.tx.hw_entry[index].stride > 0) - end = tfs->resc.tx.hw_entry[index].start + - tfs->resc.tx.hw_entry[index].stride - 1; - - tf_rm_reserve_range(tfs->resc.tx.hw_entry[index].stride, - tfs->resc.tx.hw_entry[index].start, - end, - TF_NUM_TBL_SCOPE, - tfs->TF_TBL_SCOPE_POOL_NAME_TX); -} - -/** - * Internal function to mark all the l2 epoch resources allocated that - * Truflow does not own. - */ -static void -tf_rm_rsvd_epoch(struct tf_session *tfs) -{ - uint32_t index = TF_RESC_TYPE_HW_EPOCH0; - uint32_t end = 0; - - /* epoch0 rx direction */ - if (tfs->resc.rx.hw_entry[index].stride > 0) - end = tfs->resc.rx.hw_entry[index].start + - tfs->resc.rx.hw_entry[index].stride - 1; - - tf_rm_reserve_range(tfs->resc.rx.hw_entry[index].stride, - tfs->resc.rx.hw_entry[index].start, - end, - TF_NUM_EPOCH0, - tfs->TF_EPOCH0_POOL_NAME_RX); - - /* epoch0 tx direction */ - if (tfs->resc.tx.hw_entry[index].stride > 0) - end = tfs->resc.tx.hw_entry[index].start + - tfs->resc.tx.hw_entry[index].stride - 1; - - tf_rm_reserve_range(tfs->resc.tx.hw_entry[index].stride, - tfs->resc.tx.hw_entry[index].start, - end, - TF_NUM_EPOCH0, - tfs->TF_EPOCH0_POOL_NAME_TX); - - index = TF_RESC_TYPE_HW_EPOCH1; - - /* epoch1 rx direction */ - if (tfs->resc.rx.hw_entry[index].stride > 0) - end = tfs->resc.rx.hw_entry[index].start + - tfs->resc.rx.hw_entry[index].stride - 1; - - tf_rm_reserve_range(tfs->resc.rx.hw_entry[index].stride, - tfs->resc.rx.hw_entry[index].start, - end, - TF_NUM_EPOCH1, - tfs->TF_EPOCH1_POOL_NAME_RX); - - /* epoch1 tx direction */ - if (tfs->resc.tx.hw_entry[index].stride > 0) - end = tfs->resc.tx.hw_entry[index].start + - tfs->resc.tx.hw_entry[index].stride - 1; - - tf_rm_reserve_range(tfs->resc.tx.hw_entry[index].stride, - tfs->resc.tx.hw_entry[index].start, - end, - TF_NUM_EPOCH1, - tfs->TF_EPOCH1_POOL_NAME_TX); -} - -/** - * Internal function to mark all the metadata resources allocated that - * Truflow does not own. - */ -static void -tf_rm_rsvd_metadata(struct tf_session *tfs) -{ - uint32_t index = TF_RESC_TYPE_HW_METADATA; - uint32_t end = 0; - - /* metadata rx direction */ - if (tfs->resc.rx.hw_entry[index].stride > 0) - end = tfs->resc.rx.hw_entry[index].start + - tfs->resc.rx.hw_entry[index].stride - 1; - - tf_rm_reserve_range(tfs->resc.rx.hw_entry[index].stride, - tfs->resc.rx.hw_entry[index].start, - end, - TF_NUM_METADATA, - tfs->TF_METADATA_POOL_NAME_RX); - - /* metadata tx direction */ - if (tfs->resc.tx.hw_entry[index].stride > 0) - end = tfs->resc.tx.hw_entry[index].start + - tfs->resc.tx.hw_entry[index].stride - 1; - - tf_rm_reserve_range(tfs->resc.tx.hw_entry[index].stride, - tfs->resc.tx.hw_entry[index].start, - end, - TF_NUM_METADATA, - tfs->TF_METADATA_POOL_NAME_TX); -} - -/** - * Internal function to mark all the ct state resources allocated that - * Truflow does not own. - */ -static void -tf_rm_rsvd_ct_state(struct tf_session *tfs) -{ - uint32_t index = TF_RESC_TYPE_HW_CT_STATE; - uint32_t end = 0; - - /* ct state rx direction */ - if (tfs->resc.rx.hw_entry[index].stride > 0) - end = tfs->resc.rx.hw_entry[index].start + - tfs->resc.rx.hw_entry[index].stride - 1; - - tf_rm_reserve_range(tfs->resc.rx.hw_entry[index].stride, - tfs->resc.rx.hw_entry[index].start, - end, - TF_NUM_CT_STATE, - tfs->TF_CT_STATE_POOL_NAME_RX); - - /* ct state tx direction */ - if (tfs->resc.tx.hw_entry[index].stride > 0) - end = tfs->resc.tx.hw_entry[index].start + - tfs->resc.tx.hw_entry[index].stride - 1; - - tf_rm_reserve_range(tfs->resc.tx.hw_entry[index].stride, - tfs->resc.tx.hw_entry[index].start, - end, - TF_NUM_CT_STATE, - tfs->TF_CT_STATE_POOL_NAME_TX); -} - -/** - * Internal function to mark all the range resources allocated that - * Truflow does not own. - */ -static void -tf_rm_rsvd_range(struct tf_session *tfs) -{ - uint32_t index = TF_RESC_TYPE_HW_RANGE_PROF; - uint32_t end = 0; - - /* range profile rx direction */ - if (tfs->resc.rx.hw_entry[index].stride > 0) - end = tfs->resc.rx.hw_entry[index].start + - tfs->resc.rx.hw_entry[index].stride - 1; - - tf_rm_reserve_range(tfs->resc.rx.hw_entry[index].stride, - tfs->resc.rx.hw_entry[index].start, - end, - TF_NUM_RANGE_PROF, - tfs->TF_RANGE_PROF_POOL_NAME_RX); - - /* range profile tx direction */ - if (tfs->resc.tx.hw_entry[index].stride > 0) - end = tfs->resc.tx.hw_entry[index].start + - tfs->resc.tx.hw_entry[index].stride - 1; - - tf_rm_reserve_range(tfs->resc.tx.hw_entry[index].stride, - tfs->resc.tx.hw_entry[index].start, - end, - TF_NUM_RANGE_PROF, - tfs->TF_RANGE_PROF_POOL_NAME_TX); - - index = TF_RESC_TYPE_HW_RANGE_ENTRY; - - /* range entry rx direction */ - if (tfs->resc.rx.hw_entry[index].stride > 0) - end = tfs->resc.rx.hw_entry[index].start + - tfs->resc.rx.hw_entry[index].stride - 1; - - tf_rm_reserve_range(tfs->resc.rx.hw_entry[index].stride, - tfs->resc.rx.hw_entry[index].start, - end, - TF_NUM_RANGE_ENTRY, - tfs->TF_RANGE_ENTRY_POOL_NAME_RX); - - /* range entry tx direction */ - if (tfs->resc.tx.hw_entry[index].stride > 0) - end = tfs->resc.tx.hw_entry[index].start + - tfs->resc.tx.hw_entry[index].stride - 1; - - tf_rm_reserve_range(tfs->resc.tx.hw_entry[index].stride, - tfs->resc.tx.hw_entry[index].start, - end, - TF_NUM_RANGE_ENTRY, - tfs->TF_RANGE_ENTRY_POOL_NAME_TX); -} - -/** - * Internal function to mark all the lag resources allocated that - * Truflow does not own. - */ -static void -tf_rm_rsvd_lag_entry(struct tf_session *tfs) -{ - uint32_t index = TF_RESC_TYPE_HW_LAG_ENTRY; - uint32_t end = 0; - - /* lag entry rx direction */ - if (tfs->resc.rx.hw_entry[index].stride > 0) - end = tfs->resc.rx.hw_entry[index].start + - tfs->resc.rx.hw_entry[index].stride - 1; - - tf_rm_reserve_range(tfs->resc.rx.hw_entry[index].stride, - tfs->resc.rx.hw_entry[index].start, - end, - TF_NUM_LAG_ENTRY, - tfs->TF_LAG_ENTRY_POOL_NAME_RX); - - /* lag entry tx direction */ - if (tfs->resc.tx.hw_entry[index].stride > 0) - end = tfs->resc.tx.hw_entry[index].start + - tfs->resc.tx.hw_entry[index].stride - 1; - - tf_rm_reserve_range(tfs->resc.tx.hw_entry[index].stride, - tfs->resc.tx.hw_entry[index].start, - end, - TF_NUM_LAG_ENTRY, - tfs->TF_LAG_ENTRY_POOL_NAME_TX); -} - -/** - * Internal function to mark all the full action resources allocated - * that Truflow does not own. - */ -static void -tf_rm_rsvd_sram_full_action(struct tf_session *tfs) -{ - uint32_t index = TF_RESC_TYPE_SRAM_FULL_ACTION; - uint16_t end = 0; - - /* full action rx direction */ - if (tfs->resc.rx.sram_entry[index].stride > 0) - end = tfs->resc.rx.sram_entry[index].start + - tfs->resc.rx.sram_entry[index].stride - 1; - - tf_rm_reserve_range(tfs->resc.rx.sram_entry[index].stride, - TF_RSVD_SRAM_FULL_ACTION_BEGIN_IDX_RX, - end, - TF_RSVD_SRAM_FULL_ACTION_RX, - tfs->TF_SRAM_FULL_ACTION_POOL_NAME_RX); - - /* full action tx direction */ - if (tfs->resc.tx.sram_entry[index].stride > 0) - end = tfs->resc.tx.sram_entry[index].start + - tfs->resc.tx.sram_entry[index].stride - 1; - - tf_rm_reserve_range(tfs->resc.tx.sram_entry[index].stride, - TF_RSVD_SRAM_FULL_ACTION_BEGIN_IDX_TX, - end, - TF_RSVD_SRAM_FULL_ACTION_TX, - tfs->TF_SRAM_FULL_ACTION_POOL_NAME_TX); -} - -/** - * Internal function to mark all the multicast group resources - * allocated that Truflow does not own. - */ -static void -tf_rm_rsvd_sram_mcg(struct tf_session *tfs) -{ - uint32_t index = TF_RESC_TYPE_SRAM_MCG; - uint16_t end = 0; - - /* multicast group rx direction */ - if (tfs->resc.rx.sram_entry[index].stride > 0) - end = tfs->resc.rx.sram_entry[index].start + - tfs->resc.rx.sram_entry[index].stride - 1; - - tf_rm_reserve_range(tfs->resc.rx.sram_entry[index].stride, - TF_RSVD_SRAM_MCG_BEGIN_IDX_RX, - end, - TF_RSVD_SRAM_MCG_RX, - tfs->TF_SRAM_MCG_POOL_NAME_RX); - - /* Multicast Group on TX is not supported */ -} - -/** - * Internal function to mark all the encap resources allocated that - * Truflow does not own. - */ -static void -tf_rm_rsvd_sram_encap(struct tf_session *tfs) -{ - uint32_t index = TF_RESC_TYPE_SRAM_ENCAP_8B; - uint16_t end = 0; - - /* encap 8b rx direction */ - if (tfs->resc.rx.sram_entry[index].stride > 0) - end = tfs->resc.rx.sram_entry[index].start + - tfs->resc.rx.sram_entry[index].stride - 1; - - tf_rm_reserve_range(tfs->resc.rx.sram_entry[index].stride, - TF_RSVD_SRAM_ENCAP_8B_BEGIN_IDX_RX, - end, - TF_RSVD_SRAM_ENCAP_8B_RX, - tfs->TF_SRAM_ENCAP_8B_POOL_NAME_RX); - - /* encap 8b tx direction */ - if (tfs->resc.tx.sram_entry[index].stride > 0) - end = tfs->resc.tx.sram_entry[index].start + - tfs->resc.tx.sram_entry[index].stride - 1; - - tf_rm_reserve_range(tfs->resc.tx.sram_entry[index].stride, - TF_RSVD_SRAM_ENCAP_8B_BEGIN_IDX_TX, - end, - TF_RSVD_SRAM_ENCAP_8B_TX, - tfs->TF_SRAM_ENCAP_8B_POOL_NAME_TX); - - index = TF_RESC_TYPE_SRAM_ENCAP_16B; - - /* encap 16b rx direction */ - if (tfs->resc.rx.sram_entry[index].stride > 0) - end = tfs->resc.rx.sram_entry[index].start + - tfs->resc.rx.sram_entry[index].stride - 1; - - tf_rm_reserve_range(tfs->resc.rx.sram_entry[index].stride, - TF_RSVD_SRAM_ENCAP_16B_BEGIN_IDX_RX, - end, - TF_RSVD_SRAM_ENCAP_16B_RX, - tfs->TF_SRAM_ENCAP_16B_POOL_NAME_RX); - - /* encap 16b tx direction */ - if (tfs->resc.tx.sram_entry[index].stride > 0) - end = tfs->resc.tx.sram_entry[index].start + - tfs->resc.tx.sram_entry[index].stride - 1; - - tf_rm_reserve_range(tfs->resc.tx.sram_entry[index].stride, - TF_RSVD_SRAM_ENCAP_16B_BEGIN_IDX_TX, - end, - TF_RSVD_SRAM_ENCAP_16B_TX, - tfs->TF_SRAM_ENCAP_16B_POOL_NAME_TX); - - index = TF_RESC_TYPE_SRAM_ENCAP_64B; - - /* Encap 64B not supported on RX */ - - /* Encap 64b tx direction */ - if (tfs->resc.tx.sram_entry[index].stride > 0) - end = tfs->resc.tx.sram_entry[index].start + - tfs->resc.tx.sram_entry[index].stride - 1; - - tf_rm_reserve_range(tfs->resc.tx.sram_entry[index].stride, - TF_RSVD_SRAM_ENCAP_64B_BEGIN_IDX_TX, - end, - TF_RSVD_SRAM_ENCAP_64B_TX, - tfs->TF_SRAM_ENCAP_64B_POOL_NAME_TX); -} - -/** - * Internal function to mark all the sp resources allocated that - * Truflow does not own. - */ -static void -tf_rm_rsvd_sram_sp(struct tf_session *tfs) -{ - uint32_t index = TF_RESC_TYPE_SRAM_SP_SMAC; - uint16_t end = 0; - - /* sp smac rx direction */ - if (tfs->resc.rx.sram_entry[index].stride > 0) - end = tfs->resc.rx.sram_entry[index].start + - tfs->resc.rx.sram_entry[index].stride - 1; - - tf_rm_reserve_range(tfs->resc.rx.sram_entry[index].stride, - TF_RSVD_SRAM_SP_SMAC_BEGIN_IDX_RX, - end, - TF_RSVD_SRAM_SP_SMAC_RX, - tfs->TF_SRAM_SP_SMAC_POOL_NAME_RX); - - /* sp smac tx direction */ - if (tfs->resc.tx.sram_entry[index].stride > 0) - end = tfs->resc.tx.sram_entry[index].start + - tfs->resc.tx.sram_entry[index].stride - 1; - - tf_rm_reserve_range(tfs->resc.tx.sram_entry[index].stride, - TF_RSVD_SRAM_SP_SMAC_BEGIN_IDX_TX, - end, - TF_RSVD_SRAM_SP_SMAC_TX, - tfs->TF_SRAM_SP_SMAC_POOL_NAME_TX); - - index = TF_RESC_TYPE_SRAM_SP_SMAC_IPV4; - - /* SP SMAC IPv4 not supported on RX */ - - /* sp smac ipv4 tx direction */ - if (tfs->resc.tx.sram_entry[index].stride > 0) - end = tfs->resc.tx.sram_entry[index].start + - tfs->resc.tx.sram_entry[index].stride - 1; - - tf_rm_reserve_range(tfs->resc.tx.sram_entry[index].stride, - TF_RSVD_SRAM_SP_SMAC_IPV4_BEGIN_IDX_TX, - end, - TF_RSVD_SRAM_SP_SMAC_IPV4_TX, - tfs->TF_SRAM_SP_SMAC_IPV4_POOL_NAME_TX); - - index = TF_RESC_TYPE_SRAM_SP_SMAC_IPV6; - - /* SP SMAC IPv6 not supported on RX */ - - /* sp smac ipv6 tx direction */ - if (tfs->resc.tx.sram_entry[index].stride > 0) - end = tfs->resc.tx.sram_entry[index].start + - tfs->resc.tx.sram_entry[index].stride - 1; - - tf_rm_reserve_range(tfs->resc.tx.sram_entry[index].stride, - TF_RSVD_SRAM_SP_SMAC_IPV6_BEGIN_IDX_TX, - end, - TF_RSVD_SRAM_SP_SMAC_IPV6_TX, - tfs->TF_SRAM_SP_SMAC_IPV6_POOL_NAME_TX); -} - -/** - * Internal function to mark all the stat resources allocated that - * Truflow does not own. - */ -static void -tf_rm_rsvd_sram_stats(struct tf_session *tfs) -{ - uint32_t index = TF_RESC_TYPE_SRAM_COUNTER_64B; - uint16_t end = 0; - - /* counter 64b rx direction */ - if (tfs->resc.rx.sram_entry[index].stride > 0) - end = tfs->resc.rx.sram_entry[index].start + - tfs->resc.rx.sram_entry[index].stride - 1; - - tf_rm_reserve_range(tfs->resc.rx.sram_entry[index].stride, - TF_RSVD_SRAM_COUNTER_64B_BEGIN_IDX_RX, - end, - TF_RSVD_SRAM_COUNTER_64B_RX, - tfs->TF_SRAM_STATS_64B_POOL_NAME_RX); - - /* counter 64b tx direction */ - if (tfs->resc.tx.sram_entry[index].stride > 0) - end = tfs->resc.tx.sram_entry[index].start + - tfs->resc.tx.sram_entry[index].stride - 1; - - tf_rm_reserve_range(tfs->resc.tx.sram_entry[index].stride, - TF_RSVD_SRAM_COUNTER_64B_BEGIN_IDX_TX, - end, - TF_RSVD_SRAM_COUNTER_64B_TX, - tfs->TF_SRAM_STATS_64B_POOL_NAME_TX); -} - -/** - * Internal function to mark all the nat resources allocated that - * Truflow does not own. - */ -static void -tf_rm_rsvd_sram_nat(struct tf_session *tfs) -{ - uint32_t index = TF_RESC_TYPE_SRAM_NAT_SPORT; - uint16_t end = 0; - - /* nat source port rx direction */ - if (tfs->resc.rx.sram_entry[index].stride > 0) - end = tfs->resc.rx.sram_entry[index].start + - tfs->resc.rx.sram_entry[index].stride - 1; - - tf_rm_reserve_range(tfs->resc.rx.sram_entry[index].stride, - TF_RSVD_SRAM_NAT_SPORT_BEGIN_IDX_RX, - end, - TF_RSVD_SRAM_NAT_SPORT_RX, - tfs->TF_SRAM_NAT_SPORT_POOL_NAME_RX); - - /* nat source port tx direction */ - if (tfs->resc.tx.sram_entry[index].stride > 0) - end = tfs->resc.tx.sram_entry[index].start + - tfs->resc.tx.sram_entry[index].stride - 1; - - tf_rm_reserve_range(tfs->resc.tx.sram_entry[index].stride, - TF_RSVD_SRAM_NAT_SPORT_BEGIN_IDX_TX, - end, - TF_RSVD_SRAM_NAT_SPORT_TX, - tfs->TF_SRAM_NAT_SPORT_POOL_NAME_TX); - - index = TF_RESC_TYPE_SRAM_NAT_DPORT; - - /* nat destination port rx direction */ - if (tfs->resc.rx.sram_entry[index].stride > 0) - end = tfs->resc.rx.sram_entry[index].start + - tfs->resc.rx.sram_entry[index].stride - 1; - - tf_rm_reserve_range(tfs->resc.rx.sram_entry[index].stride, - TF_RSVD_SRAM_NAT_DPORT_BEGIN_IDX_RX, - end, - TF_RSVD_SRAM_NAT_DPORT_RX, - tfs->TF_SRAM_NAT_DPORT_POOL_NAME_RX); - - /* nat destination port tx direction */ - if (tfs->resc.tx.sram_entry[index].stride > 0) - end = tfs->resc.tx.sram_entry[index].start + - tfs->resc.tx.sram_entry[index].stride - 1; - - tf_rm_reserve_range(tfs->resc.tx.sram_entry[index].stride, - TF_RSVD_SRAM_NAT_DPORT_BEGIN_IDX_TX, - end, - TF_RSVD_SRAM_NAT_DPORT_TX, - tfs->TF_SRAM_NAT_DPORT_POOL_NAME_TX); - - index = TF_RESC_TYPE_SRAM_NAT_S_IPV4; - - /* nat source port ipv4 rx direction */ - if (tfs->resc.rx.sram_entry[index].stride > 0) - end = tfs->resc.rx.sram_entry[index].start + - tfs->resc.rx.sram_entry[index].stride - 1; - - tf_rm_reserve_range(tfs->resc.rx.sram_entry[index].stride, - TF_RSVD_SRAM_NAT_S_IPV4_BEGIN_IDX_RX, - end, - TF_RSVD_SRAM_NAT_S_IPV4_RX, - tfs->TF_SRAM_NAT_S_IPV4_POOL_NAME_RX); - - /* nat source ipv4 port tx direction */ - if (tfs->resc.tx.sram_entry[index].stride > 0) - end = tfs->resc.tx.sram_entry[index].start + - tfs->resc.tx.sram_entry[index].stride - 1; - - tf_rm_reserve_range(tfs->resc.tx.sram_entry[index].stride, - TF_RSVD_SRAM_NAT_S_IPV4_BEGIN_IDX_TX, - end, - TF_RSVD_SRAM_NAT_S_IPV4_TX, - tfs->TF_SRAM_NAT_S_IPV4_POOL_NAME_TX); - - index = TF_RESC_TYPE_SRAM_NAT_D_IPV4; - - /* nat destination port ipv4 rx direction */ - if (tfs->resc.rx.sram_entry[index].stride > 0) - end = tfs->resc.rx.sram_entry[index].start + - tfs->resc.rx.sram_entry[index].stride - 1; - - tf_rm_reserve_range(tfs->resc.rx.sram_entry[index].stride, - TF_RSVD_SRAM_NAT_D_IPV4_BEGIN_IDX_RX, - end, - TF_RSVD_SRAM_NAT_D_IPV4_RX, - tfs->TF_SRAM_NAT_D_IPV4_POOL_NAME_RX); - - /* nat destination ipv4 port tx direction */ - if (tfs->resc.tx.sram_entry[index].stride > 0) - end = tfs->resc.tx.sram_entry[index].start + - tfs->resc.tx.sram_entry[index].stride - 1; - - tf_rm_reserve_range(tfs->resc.tx.sram_entry[index].stride, - TF_RSVD_SRAM_NAT_D_IPV4_BEGIN_IDX_TX, - end, - TF_RSVD_SRAM_NAT_D_IPV4_TX, - tfs->TF_SRAM_NAT_D_IPV4_POOL_NAME_TX); -} - -/** - * Internal function used to validate the HW allocated resources - * against the requested values. - */ -static int -tf_rm_hw_alloc_validate(enum tf_dir dir, - struct tf_rm_hw_alloc *hw_alloc, - struct tf_rm_entry *hw_entry) -{ - int error = 0; - int i; - - for (i = 0; i < TF_RESC_TYPE_HW_MAX; i++) { - if (hw_entry[i].stride != hw_alloc->hw_num[i]) { + if (type != TF_DEVICE_MODULE_TYPE_EM && + cfg[i].cfg_type == TF_RM_ELEM_CFG_NULL && + reservations[i] > 0) { TFP_DRV_LOG(ERR, - "%s, Alloc failed id:%d expect:%d got:%d\n", + "%s, %s, %s allocation not supported\n", + tf_device_module_type_2_str(type), tf_dir_2_str(dir), - i, - hw_alloc->hw_num[i], - hw_entry[i].stride); - error = -1; - } - } - - return error; -} - -/** - * Internal function used to validate the SRAM allocated resources - * against the requested values. - */ -static int -tf_rm_sram_alloc_validate(enum tf_dir dir __rte_unused, - struct tf_rm_sram_alloc *sram_alloc, - struct tf_rm_entry *sram_entry) -{ - int error = 0; - int i; - - for (i = 0; i < TF_RESC_TYPE_SRAM_MAX; i++) { - if (sram_entry[i].stride != sram_alloc->sram_num[i]) { - TFP_DRV_LOG(ERR, - "%s, Alloc failed idx:%d expect:%d got:%d\n", + tf_device_module_type_subtype_2_str(type, i)); + printf("%s, %s, %s allocation of %d not supported\n", + tf_device_module_type_2_str(type), tf_dir_2_str(dir), - i, - sram_alloc->sram_num[i], - sram_entry[i].stride); - error = -1; + tf_device_module_type_subtype_2_str(type, i), + reservations[i]); + } } - return error; + *valid_count = cnt; } /** - * Internal function used to mark all the HW resources allocated that - * Truflow does not own. + * Resource Manager Adjust of base index definitions. */ -static void -tf_rm_reserve_hw(struct tf *tfp) -{ - struct tf_session *tfs = (struct tf_session *)(tfp->session->core_data); - - /* TBD - * There is no direct AFM resource allocation as it is carved - * statically at AFM boot time. Thus the bit allocators work - * on the full HW resource amount and we just mark everything - * used except the resources that Truflow took ownership off. - */ - tf_rm_rsvd_l2_ctxt(tfs); - tf_rm_rsvd_prof(tfs); - tf_rm_rsvd_em_prof(tfs); - tf_rm_rsvd_wc(tfs); - tf_rm_rsvd_mirror(tfs); - tf_rm_rsvd_meter(tfs); - tf_rm_rsvd_upar(tfs); - tf_rm_rsvd_sp_tcam(tfs); - tf_rm_rsvd_l2_func(tfs); - tf_rm_rsvd_fkb(tfs); - tf_rm_rsvd_tbl_scope(tfs); - tf_rm_rsvd_epoch(tfs); - tf_rm_rsvd_metadata(tfs); - tf_rm_rsvd_ct_state(tfs); - tf_rm_rsvd_range(tfs); - tf_rm_rsvd_lag_entry(tfs); -} - -/** - * Internal function used to mark all the SRAM resources allocated - * that Truflow does not own. - */ -static void -tf_rm_reserve_sram(struct tf *tfp) -{ - struct tf_session *tfs = (struct tf_session *)(tfp->session->core_data); - - /* TBD - * There is no direct AFM resource allocation as it is carved - * statically at AFM boot time. Thus the bit allocators work - * on the full HW resource amount and we just mark everything - * used except the resources that Truflow took ownership off. - */ - tf_rm_rsvd_sram_full_action(tfs); - tf_rm_rsvd_sram_mcg(tfs); - tf_rm_rsvd_sram_encap(tfs); - tf_rm_rsvd_sram_sp(tfs); - tf_rm_rsvd_sram_stats(tfs); - tf_rm_rsvd_sram_nat(tfs); -} - -/** - * Internal function used to allocate and validate all HW resources. - */ -static int -tf_rm_allocate_validate_hw(struct tf *tfp, - enum tf_dir dir) -{ - int rc; - int i; - struct tf_rm_hw_query hw_query; - struct tf_rm_hw_alloc hw_alloc; - struct tf_session *tfs = (struct tf_session *)(tfp->session->core_data); - struct tf_rm_entry *hw_entries; - uint32_t error_flag; - - if (dir == TF_DIR_RX) - hw_entries = tfs->resc.rx.hw_entry; - else - hw_entries = tfs->resc.tx.hw_entry; - - /* Query for Session HW Resources */ - - memset(&hw_query, 0, sizeof(hw_query)); /* RSXX */ - rc = tf_rm_check_hw_qcaps_static(&hw_query, dir, &error_flag); - if (rc) { - /* Log error */ - TFP_DRV_LOG(ERR, - "%s, HW QCAPS validation failed," - "error_flag:0x%x, rc:%s\n", - tf_dir_2_str(dir), - error_flag, - strerror(-rc)); - tf_rm_print_hw_qcaps_error(dir, &hw_query, &error_flag); - goto cleanup; - } - - /* Post process HW capability */ - for (i = 0; i < TF_RESC_TYPE_HW_MAX; i++) - hw_alloc.hw_num[i] = hw_query.hw_query[i].max; - - /* Allocate Session HW Resources */ - /* Perform HW allocation validation as its possible the - * resource availability changed between qcaps and alloc - */ - rc = tf_rm_hw_alloc_validate(dir, &hw_alloc, hw_entries); - if (rc) { - /* Log error */ - TFP_DRV_LOG(ERR, - "%s, HW Resource validation failed, rc:%s\n", - tf_dir_2_str(dir), - strerror(-rc)); - goto cleanup; - } - - return 0; - - cleanup: - - return -1; -} +enum tf_rm_adjust_type { + TF_RM_ADJUST_ADD_BASE, /**< Adds base to the index */ + TF_RM_ADJUST_RM_BASE /**< Removes base from the index */ +}; /** - * Internal function used to allocate and validate all SRAM resources. + * Adjust an index according to the allocation information. * - * [in] tfp - * Pointer to TF handle + * All resources are controlled in a 0 based pool. Some resources, by + * design, are not 0 based, i.e. Full Action Records (SRAM) thus they + * need to be adjusted before they are handed out. * - * [in] dir - * Receive or transmit direction + * [in] db + * Pointer to the db, used for the lookup + * + * [in] action + * Adjust action + * + * [in] db_index + * DB index for the element type + * + * [in] index + * Index to convert + * + * [out] adj_index + * Adjusted index * * Returns: - * 0 - Success - * -1 - Internal error + * 0 - Success + * - EOPNOTSUPP - Operation not supported */ static int -tf_rm_allocate_validate_sram(struct tf *tfp, - enum tf_dir dir) +tf_rm_adjust_index(struct tf_rm_element *db, + enum tf_rm_adjust_type action, + uint32_t db_index, + uint32_t index, + uint32_t *adj_index) { - int rc; - int i; - struct tf_rm_sram_query sram_query; - struct tf_rm_sram_alloc sram_alloc; - struct tf_session *tfs = (struct tf_session *)(tfp->session->core_data); - struct tf_rm_entry *sram_entries; - uint32_t error_flag; - - if (dir == TF_DIR_RX) - sram_entries = tfs->resc.rx.sram_entry; - else - sram_entries = tfs->resc.tx.sram_entry; - - memset(&sram_query, 0, sizeof(sram_query)); /* RSXX */ - rc = tf_rm_check_sram_qcaps_static(&sram_query, dir, &error_flag); - if (rc) { - /* Log error */ - TFP_DRV_LOG(ERR, - "%s, SRAM QCAPS validation failed," - "error_flag:%x, rc:%s\n", - tf_dir_2_str(dir), - error_flag, - strerror(-rc)); - tf_rm_print_sram_qcaps_error(dir, &sram_query, &error_flag); - goto cleanup; - } + int rc = 0; + uint32_t base_index; - /* Post process SRAM capability */ - for (i = 0; i < TF_RESC_TYPE_SRAM_MAX; i++) - sram_alloc.sram_num[i] = sram_query.sram_query[i].max; + base_index = db[db_index].alloc.entry.start; - /* Perform SRAM allocation validation as its possible the - * resource availability changed between qcaps and alloc - */ - rc = tf_rm_sram_alloc_validate(dir, &sram_alloc, sram_entries); - if (rc) { - /* Log error */ - TFP_DRV_LOG(ERR, - "%s, SRAM Resource allocation validation failed," - " rc:%s\n", - tf_dir_2_str(dir), - strerror(-rc)); - goto cleanup; + switch (action) { + case TF_RM_ADJUST_RM_BASE: + *adj_index = index - base_index; + break; + case TF_RM_ADJUST_ADD_BASE: + *adj_index = index + base_index; + break; + default: + return -EOPNOTSUPP; } - return 0; - - cleanup: - - return -1; + return rc; } /** - * Helper function used to prune a HW resource array to only hold - * elements that needs to be flushed. - * - * [in] tfs - * Session handle + * Logs an array of found residual entries to the console. * * [in] dir * Receive or transmit direction * - * [in] hw_entries - * Master HW Resource database + * [in] type + * Type of Device Module * - * [in/out] flush_entries - * Pruned HW Resource database of entries to be flushed. This - * array should be passed in as a complete copy of the master HW - * Resource database. The outgoing result will be a pruned version - * based on the result of the requested checking + * [in] count + * Number of entries in the residual array * - * Returns: - * 0 - Success, no flush required - * 1 - Success, flush required - * -1 - Internal error + * [in] residuals + * Pointer to an array of residual entries. Array is index same as + * the DB in which this function is used. Each entry holds residual + * value for that entry. */ -static int -tf_rm_hw_to_flush(struct tf_session *tfs, - enum tf_dir dir, - struct tf_rm_entry *hw_entries, - struct tf_rm_entry *flush_entries) +static void +tf_rm_log_residuals(enum tf_dir dir, + enum tf_device_module_type type, + uint16_t count, + uint16_t *residuals) { - int rc; - int flush_rc = 0; - int free_cnt; - struct bitalloc *pool; + int i; - /* Check all the hw resource pools and check for left over - * elements. Any found will result in the complete pool of a - * type to get invalidated. + /* Walk the residual array and log the types that wasn't + * cleaned up to the console. */ - - TF_RM_GET_POOLS(tfs, dir, &pool, - TF_L2_CTXT_TCAM_POOL_NAME, - rc); - if (rc) - return rc; - free_cnt = ba_free_count(pool); - if (free_cnt == hw_entries[TF_RESC_TYPE_HW_L2_CTXT_TCAM].stride) { - flush_entries[TF_RESC_TYPE_HW_L2_CTXT_TCAM].start = 0; - flush_entries[TF_RESC_TYPE_HW_L2_CTXT_TCAM].stride = 0; - } else { - flush_rc = 1; - } - - TF_RM_GET_POOLS(tfs, dir, &pool, - TF_PROF_FUNC_POOL_NAME, - rc); - if (rc) - return rc; - free_cnt = ba_free_count(pool); - if (free_cnt == hw_entries[TF_RESC_TYPE_HW_PROF_FUNC].stride) { - flush_entries[TF_RESC_TYPE_HW_PROF_FUNC].start = 0; - flush_entries[TF_RESC_TYPE_HW_PROF_FUNC].stride = 0; - } else { - flush_rc = 1; - } - - TF_RM_GET_POOLS(tfs, dir, &pool, - TF_PROF_TCAM_POOL_NAME, - rc); - if (rc) - return rc; - free_cnt = ba_free_count(pool); - if (free_cnt == hw_entries[TF_RESC_TYPE_HW_PROF_TCAM].stride) { - flush_entries[TF_RESC_TYPE_HW_PROF_TCAM].start = 0; - flush_entries[TF_RESC_TYPE_HW_PROF_TCAM].stride = 0; - } else { - flush_rc = 1; - } - - TF_RM_GET_POOLS(tfs, dir, &pool, - TF_EM_PROF_ID_POOL_NAME, - rc); - if (rc) - return rc; - free_cnt = ba_free_count(pool); - if (free_cnt == hw_entries[TF_RESC_TYPE_HW_EM_PROF_ID].stride) { - flush_entries[TF_RESC_TYPE_HW_EM_PROF_ID].start = 0; - flush_entries[TF_RESC_TYPE_HW_EM_PROF_ID].stride = 0; - } else { - flush_rc = 1; - } - - flush_entries[TF_RESC_TYPE_HW_EM_REC].start = 0; - flush_entries[TF_RESC_TYPE_HW_EM_REC].stride = 0; - - TF_RM_GET_POOLS(tfs, dir, &pool, - TF_WC_TCAM_PROF_ID_POOL_NAME, - rc); - if (rc) - return rc; - free_cnt = ba_free_count(pool); - if (free_cnt == hw_entries[TF_RESC_TYPE_HW_WC_TCAM_PROF_ID].stride) { - flush_entries[TF_RESC_TYPE_HW_WC_TCAM_PROF_ID].start = 0; - flush_entries[TF_RESC_TYPE_HW_WC_TCAM_PROF_ID].stride = 0; - } else { - flush_rc = 1; - } - - TF_RM_GET_POOLS(tfs, dir, &pool, - TF_WC_TCAM_POOL_NAME, - rc); - if (rc) - return rc; - free_cnt = ba_free_count(pool); - if (free_cnt == hw_entries[TF_RESC_TYPE_HW_WC_TCAM].stride) { - flush_entries[TF_RESC_TYPE_HW_WC_TCAM].start = 0; - flush_entries[TF_RESC_TYPE_HW_WC_TCAM].stride = 0; - } else { - flush_rc = 1; - } - - TF_RM_GET_POOLS(tfs, dir, &pool, - TF_METER_PROF_POOL_NAME, - rc); - if (rc) - return rc; - free_cnt = ba_free_count(pool); - if (free_cnt == hw_entries[TF_RESC_TYPE_HW_METER_PROF].stride) { - flush_entries[TF_RESC_TYPE_HW_METER_PROF].start = 0; - flush_entries[TF_RESC_TYPE_HW_METER_PROF].stride = 0; - } else { - flush_rc = 1; - } - - TF_RM_GET_POOLS(tfs, dir, &pool, - TF_METER_INST_POOL_NAME, - rc); - if (rc) - return rc; - free_cnt = ba_free_count(pool); - if (free_cnt == hw_entries[TF_RESC_TYPE_HW_METER_INST].stride) { - flush_entries[TF_RESC_TYPE_HW_METER_INST].start = 0; - flush_entries[TF_RESC_TYPE_HW_METER_INST].stride = 0; - } else { - flush_rc = 1; - } - - TF_RM_GET_POOLS(tfs, dir, &pool, - TF_MIRROR_POOL_NAME, - rc); - if (rc) - return rc; - free_cnt = ba_free_count(pool); - if (free_cnt == hw_entries[TF_RESC_TYPE_HW_MIRROR].stride) { - flush_entries[TF_RESC_TYPE_HW_MIRROR].start = 0; - flush_entries[TF_RESC_TYPE_HW_MIRROR].stride = 0; - } else { - flush_rc = 1; - } - - TF_RM_GET_POOLS(tfs, dir, &pool, - TF_UPAR_POOL_NAME, - rc); - if (rc) - return rc; - free_cnt = ba_free_count(pool); - if (free_cnt == hw_entries[TF_RESC_TYPE_HW_UPAR].stride) { - flush_entries[TF_RESC_TYPE_HW_UPAR].start = 0; - flush_entries[TF_RESC_TYPE_HW_UPAR].stride = 0; - } else { - flush_rc = 1; - } - - TF_RM_GET_POOLS(tfs, dir, &pool, - TF_SP_TCAM_POOL_NAME, - rc); - if (rc) - return rc; - free_cnt = ba_free_count(pool); - if (free_cnt == hw_entries[TF_RESC_TYPE_HW_SP_TCAM].stride) { - flush_entries[TF_RESC_TYPE_HW_SP_TCAM].start = 0; - flush_entries[TF_RESC_TYPE_HW_SP_TCAM].stride = 0; - } else { - flush_rc = 1; - } - - TF_RM_GET_POOLS(tfs, dir, &pool, - TF_L2_FUNC_POOL_NAME, - rc); - if (rc) - return rc; - free_cnt = ba_free_count(pool); - if (free_cnt == hw_entries[TF_RESC_TYPE_HW_L2_FUNC].stride) { - flush_entries[TF_RESC_TYPE_HW_L2_FUNC].start = 0; - flush_entries[TF_RESC_TYPE_HW_L2_FUNC].stride = 0; - } else { - flush_rc = 1; - } - - TF_RM_GET_POOLS(tfs, dir, &pool, - TF_FKB_POOL_NAME, - rc); - if (rc) - return rc; - free_cnt = ba_free_count(pool); - if (free_cnt == hw_entries[TF_RESC_TYPE_HW_FKB].stride) { - flush_entries[TF_RESC_TYPE_HW_FKB].start = 0; - flush_entries[TF_RESC_TYPE_HW_FKB].stride = 0; - } else { - flush_rc = 1; - } - - TF_RM_GET_POOLS(tfs, dir, &pool, - TF_TBL_SCOPE_POOL_NAME, - rc); - if (rc) - return rc; - free_cnt = ba_free_count(pool); - if (free_cnt == hw_entries[TF_RESC_TYPE_HW_TBL_SCOPE].stride) { - flush_entries[TF_RESC_TYPE_HW_TBL_SCOPE].start = 0; - flush_entries[TF_RESC_TYPE_HW_TBL_SCOPE].stride = 0; - } else { - TFP_DRV_LOG(ERR, "%s, TBL_SCOPE free_cnt:%d, entries:%d\n", - tf_dir_2_str(dir), - free_cnt, - hw_entries[TF_RESC_TYPE_HW_TBL_SCOPE].stride); - flush_rc = 1; - } - - TF_RM_GET_POOLS(tfs, dir, &pool, - TF_EPOCH0_POOL_NAME, - rc); - if (rc) - return rc; - free_cnt = ba_free_count(pool); - if (free_cnt == hw_entries[TF_RESC_TYPE_HW_EPOCH0].stride) { - flush_entries[TF_RESC_TYPE_HW_EPOCH0].start = 0; - flush_entries[TF_RESC_TYPE_HW_EPOCH0].stride = 0; - } else { - flush_rc = 1; - } - - TF_RM_GET_POOLS(tfs, dir, &pool, - TF_EPOCH1_POOL_NAME, - rc); - if (rc) - return rc; - free_cnt = ba_free_count(pool); - if (free_cnt == hw_entries[TF_RESC_TYPE_HW_EPOCH1].stride) { - flush_entries[TF_RESC_TYPE_HW_EPOCH1].start = 0; - flush_entries[TF_RESC_TYPE_HW_EPOCH1].stride = 0; - } else { - flush_rc = 1; - } - - TF_RM_GET_POOLS(tfs, dir, &pool, - TF_METADATA_POOL_NAME, - rc); - if (rc) - return rc; - free_cnt = ba_free_count(pool); - if (free_cnt == hw_entries[TF_RESC_TYPE_HW_METADATA].stride) { - flush_entries[TF_RESC_TYPE_HW_METADATA].start = 0; - flush_entries[TF_RESC_TYPE_HW_METADATA].stride = 0; - } else { - flush_rc = 1; - } - - TF_RM_GET_POOLS(tfs, dir, &pool, - TF_CT_STATE_POOL_NAME, - rc); - if (rc) - return rc; - free_cnt = ba_free_count(pool); - if (free_cnt == hw_entries[TF_RESC_TYPE_HW_CT_STATE].stride) { - flush_entries[TF_RESC_TYPE_HW_CT_STATE].start = 0; - flush_entries[TF_RESC_TYPE_HW_CT_STATE].stride = 0; - } else { - flush_rc = 1; - } - - TF_RM_GET_POOLS(tfs, dir, &pool, - TF_RANGE_PROF_POOL_NAME, - rc); - if (rc) - return rc; - free_cnt = ba_free_count(pool); - if (free_cnt == hw_entries[TF_RESC_TYPE_HW_RANGE_PROF].stride) { - flush_entries[TF_RESC_TYPE_HW_RANGE_PROF].start = 0; - flush_entries[TF_RESC_TYPE_HW_RANGE_PROF].stride = 0; - } else { - flush_rc = 1; - } - - TF_RM_GET_POOLS(tfs, dir, &pool, - TF_RANGE_ENTRY_POOL_NAME, - rc); - if (rc) - return rc; - free_cnt = ba_free_count(pool); - if (free_cnt == hw_entries[TF_RESC_TYPE_HW_RANGE_ENTRY].stride) { - flush_entries[TF_RESC_TYPE_HW_RANGE_ENTRY].start = 0; - flush_entries[TF_RESC_TYPE_HW_RANGE_ENTRY].stride = 0; - } else { - flush_rc = 1; - } - - TF_RM_GET_POOLS(tfs, dir, &pool, - TF_LAG_ENTRY_POOL_NAME, - rc); - if (rc) - return rc; - free_cnt = ba_free_count(pool); - if (free_cnt == hw_entries[TF_RESC_TYPE_HW_LAG_ENTRY].stride) { - flush_entries[TF_RESC_TYPE_HW_LAG_ENTRY].start = 0; - flush_entries[TF_RESC_TYPE_HW_LAG_ENTRY].stride = 0; - } else { - flush_rc = 1; + for (i = 0; i < count; i++) { + if (residuals[i] != 0) + TFP_DRV_LOG(ERR, + "%s, %s was not cleaned up, %d outstanding\n", + tf_dir_2_str(dir), + tf_device_module_type_subtype_2_str(type, i), + residuals[i]); } - - return flush_rc; } /** - * Helper function used to prune a SRAM resource array to only hold - * elements that needs to be flushed. + * Performs a check of the passed in DB for any lingering elements. If + * a resource type was found to not have been cleaned up by the caller + * then its residual values are recorded, logged and passed back in an + * allocate reservation array that the caller can pass to the FW for + * cleanup. * - * [in] tfs - * Session handle - * - * [in] dir - * Receive or transmit direction + * [in] db + * Pointer to the db, used for the lookup * - * [in] hw_entries - * Master SRAM Resource data base + * [out] resv_size + * Pointer to the reservation size of the generated reservation + * array. * - * [in/out] flush_entries - * Pruned SRAM Resource database of entries to be flushed. This - * array should be passed in as a complete copy of the master SRAM - * Resource database. The outgoing result will be a pruned version - * based on the result of the requested checking + * [in/out] resv + * Pointer Pointer to a reservation array. The reservation array is + * allocated after the residual scan and holds any found residual + * entries. Thus it can be smaller than the DB that the check was + * performed on. Array must be freed by the caller. + * + * [out] residuals_present + * Pointer to a bool flag indicating if residual was present in the + * DB * * Returns: - * 0 - Success, no flush required - * 1 - Success, flush required - * -1 - Internal error + * 0 - Success + * - EOPNOTSUPP - Operation not supported */ static int -tf_rm_sram_to_flush(struct tf_session *tfs, - enum tf_dir dir, - struct tf_rm_entry *sram_entries, - struct tf_rm_entry *flush_entries) +tf_rm_check_residuals(struct tf_rm_new_db *rm_db, + uint16_t *resv_size, + struct tf_rm_resc_entry **resv, + bool *residuals_present) { int rc; - int flush_rc = 0; - int free_cnt; - struct bitalloc *pool; - - /* Check all the sram resource pools and check for left over - * elements. Any found will result in the complete pool of a - * type to get invalidated. - */ - - TF_RM_GET_POOLS(tfs, dir, &pool, - TF_SRAM_FULL_ACTION_POOL_NAME, - rc); + int i; + int f; + uint16_t count; + uint16_t found; + uint16_t *residuals = NULL; + uint16_t hcapi_type; + struct tf_rm_get_inuse_count_parms iparms; + struct tf_rm_get_alloc_info_parms aparms; + struct tf_rm_get_hcapi_parms hparms; + struct tf_rm_alloc_info info; + struct tfp_calloc_parms cparms; + struct tf_rm_resc_entry *local_resv = NULL; + + /* Create array to hold the entries that have residuals */ + cparms.nitems = rm_db->num_entries; + cparms.size = sizeof(uint16_t); + cparms.alignment = 0; + rc = tfp_calloc(&cparms); if (rc) return rc; - free_cnt = ba_free_count(pool); - if (free_cnt == sram_entries[TF_RESC_TYPE_SRAM_FULL_ACTION].stride) { - flush_entries[TF_RESC_TYPE_SRAM_FULL_ACTION].start = 0; - flush_entries[TF_RESC_TYPE_SRAM_FULL_ACTION].stride = 0; - } else { - flush_rc = 1; + + residuals = (uint16_t *)cparms.mem_va; + + /* Traverse the DB and collect any residual elements */ + iparms.rm_db = rm_db; + iparms.count = &count; + for (i = 0, found = 0; i < rm_db->num_entries; i++) { + iparms.db_index = i; + rc = tf_rm_get_inuse_count(&iparms); + /* Not a device supported entry, just skip */ + if (rc == -ENOTSUP) + continue; + if (rc) + goto cleanup_residuals; + + if (count) { + found++; + residuals[i] = count; + *residuals_present = true; + } } - /* Only pools for RX direction */ - if (dir == TF_DIR_RX) { - TF_RM_GET_POOLS_RX(tfs, &pool, - TF_SRAM_MCG_POOL_NAME); + if (*residuals_present) { + /* Populate a reduced resv array with only the entries + * that have residuals. + */ + cparms.nitems = found; + cparms.size = sizeof(struct tf_rm_resc_entry); + cparms.alignment = 0; + rc = tfp_calloc(&cparms); if (rc) return rc; - free_cnt = ba_free_count(pool); - if (free_cnt == sram_entries[TF_RESC_TYPE_SRAM_MCG].stride) { - flush_entries[TF_RESC_TYPE_SRAM_MCG].start = 0; - flush_entries[TF_RESC_TYPE_SRAM_MCG].stride = 0; - } else { - flush_rc = 1; + + local_resv = (struct tf_rm_resc_entry *)cparms.mem_va; + + aparms.rm_db = rm_db; + hparms.rm_db = rm_db; + hparms.hcapi_type = &hcapi_type; + for (i = 0, f = 0; i < rm_db->num_entries; i++) { + if (residuals[i] == 0) + continue; + aparms.db_index = i; + aparms.info = &info; + rc = tf_rm_get_info(&aparms); + if (rc) + goto cleanup_all; + + hparms.db_index = i; + rc = tf_rm_get_hcapi_type(&hparms); + if (rc) + goto cleanup_all; + + local_resv[f].type = hcapi_type; + local_resv[f].start = info.entry.start; + local_resv[f].stride = info.entry.stride; + f++; } - } else { - /* Always prune TX direction */ - flush_entries[TF_RESC_TYPE_SRAM_MCG].start = 0; - flush_entries[TF_RESC_TYPE_SRAM_MCG].stride = 0; + *resv_size = found; } - TF_RM_GET_POOLS(tfs, dir, &pool, - TF_SRAM_ENCAP_8B_POOL_NAME, - rc); + tf_rm_log_residuals(rm_db->dir, + rm_db->type, + rm_db->num_entries, + residuals); + + tfp_free((void *)residuals); + *resv = local_resv; + + return 0; + + cleanup_all: + tfp_free((void *)local_resv); + *resv = NULL; + cleanup_residuals: + tfp_free((void *)residuals); + + return rc; +} + +int +tf_rm_create_db(struct tf *tfp, + struct tf_rm_create_db_parms *parms) +{ + int rc; + int i; + int j; + struct tf_session *tfs; + struct tf_dev_info *dev; + uint16_t max_types; + struct tfp_calloc_parms cparms; + struct tf_rm_resc_req_entry *query; + enum tf_rm_resc_resv_strategy resv_strategy; + struct tf_rm_resc_req_entry *req; + struct tf_rm_resc_entry *resv; + struct tf_rm_new_db *rm_db; + struct tf_rm_element *db; + uint32_t pool_size; + uint16_t hcapi_items; + + TF_CHECK_PARMS2(tfp, parms); + + /* Retrieve the session information */ + rc = tf_session_get_session(tfp, &tfs); if (rc) return rc; - free_cnt = ba_free_count(pool); - if (free_cnt == sram_entries[TF_RESC_TYPE_SRAM_ENCAP_8B].stride) { - flush_entries[TF_RESC_TYPE_SRAM_ENCAP_8B].start = 0; - flush_entries[TF_RESC_TYPE_SRAM_ENCAP_8B].stride = 0; - } else { - flush_rc = 1; - } - TF_RM_GET_POOLS(tfs, dir, &pool, - TF_SRAM_ENCAP_16B_POOL_NAME, - rc); + /* Retrieve device information */ + rc = tf_session_get_device(tfs, &dev); if (rc) return rc; - free_cnt = ba_free_count(pool); - if (free_cnt == sram_entries[TF_RESC_TYPE_SRAM_ENCAP_16B].stride) { - flush_entries[TF_RESC_TYPE_SRAM_ENCAP_16B].start = 0; - flush_entries[TF_RESC_TYPE_SRAM_ENCAP_16B].stride = 0; - } else { - flush_rc = 1; - } - /* Only pools for TX direction */ - if (dir == TF_DIR_TX) { - TF_RM_GET_POOLS_TX(tfs, &pool, - TF_SRAM_ENCAP_64B_POOL_NAME); - if (rc) - return rc; - free_cnt = ba_free_count(pool); - if (free_cnt == - sram_entries[TF_RESC_TYPE_SRAM_ENCAP_64B].stride) { - flush_entries[TF_RESC_TYPE_SRAM_ENCAP_64B].start = 0; - flush_entries[TF_RESC_TYPE_SRAM_ENCAP_64B].stride = 0; - } else { - flush_rc = 1; - } - } else { - /* Always prune RX direction */ - flush_entries[TF_RESC_TYPE_SRAM_ENCAP_64B].start = 0; - flush_entries[TF_RESC_TYPE_SRAM_ENCAP_64B].stride = 0; - } + /* Need device max number of elements for the RM QCAPS */ + rc = dev->ops->tf_dev_get_max_types(tfp, &max_types); + if (rc) + return rc; - TF_RM_GET_POOLS(tfs, dir, &pool, - TF_SRAM_SP_SMAC_POOL_NAME, - rc); + cparms.nitems = max_types; + cparms.size = sizeof(struct tf_rm_resc_req_entry); + cparms.alignment = 0; + rc = tfp_calloc(&cparms); if (rc) return rc; - free_cnt = ba_free_count(pool); - if (free_cnt == sram_entries[TF_RESC_TYPE_SRAM_SP_SMAC].stride) { - flush_entries[TF_RESC_TYPE_SRAM_SP_SMAC].start = 0; - flush_entries[TF_RESC_TYPE_SRAM_SP_SMAC].stride = 0; - } else { - flush_rc = 1; - } - /* Only pools for TX direction */ - if (dir == TF_DIR_TX) { - TF_RM_GET_POOLS_TX(tfs, &pool, - TF_SRAM_SP_SMAC_IPV4_POOL_NAME); - if (rc) - return rc; - free_cnt = ba_free_count(pool); - if (free_cnt == - sram_entries[TF_RESC_TYPE_SRAM_SP_SMAC_IPV4].stride) { - flush_entries[TF_RESC_TYPE_SRAM_SP_SMAC_IPV4].start = 0; - flush_entries[TF_RESC_TYPE_SRAM_SP_SMAC_IPV4].stride = - 0; - } else { - flush_rc = 1; - } - } else { - /* Always prune RX direction */ - flush_entries[TF_RESC_TYPE_SRAM_SP_SMAC_IPV4].start = 0; - flush_entries[TF_RESC_TYPE_SRAM_SP_SMAC_IPV4].stride = 0; - } + query = (struct tf_rm_resc_req_entry *)cparms.mem_va; - /* Only pools for TX direction */ - if (dir == TF_DIR_TX) { - TF_RM_GET_POOLS_TX(tfs, &pool, - TF_SRAM_SP_SMAC_IPV6_POOL_NAME); - if (rc) - return rc; - free_cnt = ba_free_count(pool); - if (free_cnt == - sram_entries[TF_RESC_TYPE_SRAM_SP_SMAC_IPV6].stride) { - flush_entries[TF_RESC_TYPE_SRAM_SP_SMAC_IPV6].start = 0; - flush_entries[TF_RESC_TYPE_SRAM_SP_SMAC_IPV6].stride = - 0; - } else { - flush_rc = 1; - } - } else { - /* Always prune RX direction */ - flush_entries[TF_RESC_TYPE_SRAM_SP_SMAC_IPV6].start = 0; - flush_entries[TF_RESC_TYPE_SRAM_SP_SMAC_IPV6].stride = 0; + /* Get Firmware Capabilities */ + rc = tf_msg_session_resc_qcaps(tfp, + parms->dir, + max_types, + query, + &resv_strategy); + if (rc) + return rc; + + /* Process capabilities against DB requirements. However, as a + * DB can hold elements that are not HCAPI we can reduce the + * req msg content by removing those out of the request yet + * the DB holds them all as to give a fast lookup. We can also + * remove entries where there are no request for elements. + */ + tf_rm_count_hcapi_reservations(parms->dir, + parms->type, + parms->cfg, + parms->alloc_cnt, + parms->num_elements, + &hcapi_items); + + /* Handle the case where a DB create request really ends up + * being empty. Unsupported (if not rare) case but possible + * that no resources are necessary for a 'direction'. + */ + if (hcapi_items == 0) { + TFP_DRV_LOG(ERR, + "%s: DB create request for Zero elements, DB Type:%s\n", + tf_dir_2_str(parms->dir), + tf_device_module_type_2_str(parms->type)); + + parms->rm_db = NULL; + return -ENOMEM; } - TF_RM_GET_POOLS(tfs, dir, &pool, - TF_SRAM_STATS_64B_POOL_NAME, - rc); + /* Alloc request, alignment already set */ + cparms.nitems = (size_t)hcapi_items; + cparms.size = sizeof(struct tf_rm_resc_req_entry); + rc = tfp_calloc(&cparms); if (rc) return rc; - free_cnt = ba_free_count(pool); - if (free_cnt == sram_entries[TF_RESC_TYPE_SRAM_COUNTER_64B].stride) { - flush_entries[TF_RESC_TYPE_SRAM_COUNTER_64B].start = 0; - flush_entries[TF_RESC_TYPE_SRAM_COUNTER_64B].stride = 0; - } else { - flush_rc = 1; - } + req = (struct tf_rm_resc_req_entry *)cparms.mem_va; - TF_RM_GET_POOLS(tfs, dir, &pool, - TF_SRAM_NAT_SPORT_POOL_NAME, - rc); + /* Alloc reservation, alignment and nitems already set */ + cparms.size = sizeof(struct tf_rm_resc_entry); + rc = tfp_calloc(&cparms); if (rc) return rc; - free_cnt = ba_free_count(pool); - if (free_cnt == sram_entries[TF_RESC_TYPE_SRAM_NAT_SPORT].stride) { - flush_entries[TF_RESC_TYPE_SRAM_NAT_SPORT].start = 0; - flush_entries[TF_RESC_TYPE_SRAM_NAT_SPORT].stride = 0; - } else { - flush_rc = 1; + resv = (struct tf_rm_resc_entry *)cparms.mem_va; + + /* Build the request */ + for (i = 0, j = 0; i < parms->num_elements; i++) { + /* Skip any non HCAPI cfg elements */ + if (parms->cfg[i].cfg_type == TF_RM_ELEM_CFG_HCAPI) { + /* Only perform reservation for entries that + * has been requested + */ + if (parms->alloc_cnt[i] == 0) + continue; + + /* Verify that we can get the full amount + * allocated per the qcaps availability. + */ + if (parms->alloc_cnt[i] <= + query[parms->cfg[i].hcapi_type].max) { + req[j].type = parms->cfg[i].hcapi_type; + req[j].min = parms->alloc_cnt[i]; + req[j].max = parms->alloc_cnt[i]; + j++; + } else { + TFP_DRV_LOG(ERR, + "%s: Resource failure, type:%d\n", + tf_dir_2_str(parms->dir), + parms->cfg[i].hcapi_type); + TFP_DRV_LOG(ERR, + "req:%d, avail:%d\n", + parms->alloc_cnt[i], + query[parms->cfg[i].hcapi_type].max); + return -EINVAL; + } + } } - TF_RM_GET_POOLS(tfs, dir, &pool, - TF_SRAM_NAT_DPORT_POOL_NAME, - rc); + rc = tf_msg_session_resc_alloc(tfp, + parms->dir, + hcapi_items, + req, + resv); if (rc) return rc; - free_cnt = ba_free_count(pool); - if (free_cnt == sram_entries[TF_RESC_TYPE_SRAM_NAT_DPORT].stride) { - flush_entries[TF_RESC_TYPE_SRAM_NAT_DPORT].start = 0; - flush_entries[TF_RESC_TYPE_SRAM_NAT_DPORT].stride = 0; - } else { - flush_rc = 1; - } - TF_RM_GET_POOLS(tfs, dir, &pool, - TF_SRAM_NAT_S_IPV4_POOL_NAME, - rc); + /* Build the RM DB per the request */ + cparms.nitems = 1; + cparms.size = sizeof(struct tf_rm_new_db); + rc = tfp_calloc(&cparms); if (rc) return rc; - free_cnt = ba_free_count(pool); - if (free_cnt == sram_entries[TF_RESC_TYPE_SRAM_NAT_S_IPV4].stride) { - flush_entries[TF_RESC_TYPE_SRAM_NAT_S_IPV4].start = 0; - flush_entries[TF_RESC_TYPE_SRAM_NAT_S_IPV4].stride = 0; - } else { - flush_rc = 1; - } + rm_db = (void *)cparms.mem_va; - TF_RM_GET_POOLS(tfs, dir, &pool, - TF_SRAM_NAT_D_IPV4_POOL_NAME, - rc); + /* Build the DB within RM DB */ + cparms.nitems = parms->num_elements; + cparms.size = sizeof(struct tf_rm_element); + rc = tfp_calloc(&cparms); if (rc) return rc; - free_cnt = ba_free_count(pool); - if (free_cnt == sram_entries[TF_RESC_TYPE_SRAM_NAT_D_IPV4].stride) { - flush_entries[TF_RESC_TYPE_SRAM_NAT_D_IPV4].start = 0; - flush_entries[TF_RESC_TYPE_SRAM_NAT_D_IPV4].stride = 0; - } else { - flush_rc = 1; - } + rm_db->db = (struct tf_rm_element *)cparms.mem_va; - return flush_rc; -} + db = rm_db->db; + for (i = 0, j = 0; i < parms->num_elements; i++) { + db[i].cfg_type = parms->cfg[i].cfg_type; + db[i].hcapi_type = parms->cfg[i].hcapi_type; -/** - * Helper function used to generate an error log for the HW types that - * needs to be flushed. The types should have been cleaned up ahead of - * invoking tf_close_session. - * - * [in] hw_entries - * HW Resource database holding elements to be flushed - */ -static void -tf_rm_log_hw_flush(enum tf_dir dir, - struct tf_rm_entry *hw_entries) -{ - int i; + /* Skip any non HCAPI types as we didn't include them + * in the reservation request. + */ + if (parms->cfg[i].cfg_type != TF_RM_ELEM_CFG_HCAPI) + continue; - /* Walk the hw flush array and log the types that wasn't - * cleaned up. - */ - for (i = 0; i < TF_RESC_TYPE_HW_MAX; i++) { - if (hw_entries[i].stride != 0) + /* If the element didn't request an allocation no need + * to create a pool nor verify if we got a reservation. + */ + if (parms->alloc_cnt[i] == 0) + continue; + + /* If the element had requested an allocation and that + * allocation was a success (full amount) then + * allocate the pool. + */ + if (parms->alloc_cnt[i] == resv[j].stride) { + db[i].alloc.entry.start = resv[j].start; + db[i].alloc.entry.stride = resv[j].stride; + + printf("Entry:%d Start:%d Stride:%d\n", + i, + resv[j].start, + resv[j].stride); + + /* Create pool */ + pool_size = (BITALLOC_SIZEOF(resv[j].stride) / + sizeof(struct bitalloc)); + /* Alloc request, alignment already set */ + cparms.nitems = pool_size; + cparms.size = sizeof(struct bitalloc); + rc = tfp_calloc(&cparms); + if (rc) { + TFP_DRV_LOG(ERR, + "%s: Pool alloc failed, type:%d\n", + tf_dir_2_str(parms->dir), + db[i].cfg_type); + goto fail; + } + db[i].pool = (struct bitalloc *)cparms.mem_va; + + rc = ba_init(db[i].pool, resv[j].stride); + if (rc) { + TFP_DRV_LOG(ERR, + "%s: Pool init failed, type:%d\n", + tf_dir_2_str(parms->dir), + db[i].cfg_type); + goto fail; + } + j++; + } else { + /* Bail out as we want what we requested for + * all elements, not any less. + */ TFP_DRV_LOG(ERR, - "%s, %s was not cleaned up\n", - tf_dir_2_str(dir), - tf_hcapi_hw_2_str(i)); + "%s: Alloc failed, type:%d\n", + tf_dir_2_str(parms->dir), + db[i].cfg_type); + TFP_DRV_LOG(ERR, + "req:%d, alloc:%d\n", + parms->alloc_cnt[i], + resv[j].stride); + goto fail; + } } + + rm_db->num_entries = parms->num_elements; + rm_db->dir = parms->dir; + rm_db->type = parms->type; + *parms->rm_db = (void *)rm_db; + + printf("%s: type:%d num_entries:%d\n", + tf_dir_2_str(parms->dir), + parms->type, + i); + + tfp_free((void *)req); + tfp_free((void *)resv); + + return 0; + + fail: + tfp_free((void *)req); + tfp_free((void *)resv); + tfp_free((void *)db->pool); + tfp_free((void *)db); + tfp_free((void *)rm_db); + parms->rm_db = NULL; + + return -EINVAL; } -/** - * Helper function used to generate an error log for the SRAM types - * that needs to be flushed. The types should have been cleaned up - * ahead of invoking tf_close_session. - * - * [in] sram_entries - * SRAM Resource database holding elements to be flushed - */ -static void -tf_rm_log_sram_flush(enum tf_dir dir, - struct tf_rm_entry *sram_entries) +int +tf_rm_free_db(struct tf *tfp, + struct tf_rm_free_db_parms *parms) { + int rc; int i; + uint16_t resv_size = 0; + struct tf_rm_new_db *rm_db; + struct tf_rm_resc_entry *resv; + bool residuals_found = false; + + TF_CHECK_PARMS2(parms, parms->rm_db); + + /* Device unbind happens when the TF Session is closed and the + * session ref count is 0. Device unbind will cleanup each of + * its support modules, i.e. Identifier, thus we're ending up + * here to close the DB. + * + * On TF Session close it is assumed that the session has already + * cleaned up all its resources, individually, while + * destroying its flows. + * + * To assist in the 'cleanup checking' the DB is checked for any + * remaining elements and logged if found to be the case. + * + * Any such elements will need to be 'cleared' ahead of + * returning the resources to the HCAPI RM. + * + * RM will signal FW to flush the DB resources. FW will + * perform the invalidation. TF Session close will return the + * previous allocated elements to the RM and then close the + * HCAPI RM registration. That then saves several 'free' msgs + * from being required. + */ - /* Walk the sram flush array and log the types that wasn't - * cleaned up. + rm_db = (struct tf_rm_new_db *)parms->rm_db; + + /* Check for residuals that the client didn't clean up */ + rc = tf_rm_check_residuals(rm_db, + &resv_size, + &resv, + &residuals_found); + if (rc) + return rc; + + /* Invalidate any residuals followed by a DB traversal for + * pool cleanup. */ - for (i = 0; i < TF_RESC_TYPE_SRAM_MAX; i++) { - if (sram_entries[i].stride != 0) + if (residuals_found) { + rc = tf_msg_session_resc_flush(tfp, + parms->dir, + resv_size, + resv); + tfp_free((void *)resv); + /* On failure we still have to cleanup so we can only + * log that FW failed. + */ + if (rc) TFP_DRV_LOG(ERR, - "%s, %s was not cleaned up\n", - tf_dir_2_str(dir), - tf_hcapi_sram_2_str(i)); + "%s: Internal Flush error, module:%s\n", + tf_dir_2_str(parms->dir), + tf_device_module_type_2_str(rm_db->type)); } -} -void -tf_rm_init(struct tf *tfp __rte_unused) -{ - struct tf_session *tfs = - (struct tf_session *)(tfp->session->core_data); + for (i = 0; i < rm_db->num_entries; i++) + tfp_free((void *)rm_db->db[i].pool); - /* This version is host specific and should be checked against - * when attaching as there is no guarantee that a secondary - * would run from same image version. - */ - tfs->ver.major = TF_SESSION_VER_MAJOR; - tfs->ver.minor = TF_SESSION_VER_MINOR; - tfs->ver.update = TF_SESSION_VER_UPDATE; - - tfs->session_id.id = 0; - tfs->ref_count = 0; - - /* Initialization of Table Scopes */ - /* ll_init(&tfs->tbl_scope_ll); */ - - /* Initialization of HW and SRAM resource DB */ - memset(&tfs->resc, 0, sizeof(struct tf_rm_db)); - - /* Initialization of HW Resource Pools */ - ba_init(tfs->TF_L2_CTXT_TCAM_POOL_NAME_RX, TF_NUM_L2_CTXT_TCAM); - ba_init(tfs->TF_L2_CTXT_TCAM_POOL_NAME_TX, TF_NUM_L2_CTXT_TCAM); - ba_init(tfs->TF_PROF_FUNC_POOL_NAME_RX, TF_NUM_PROF_FUNC); - ba_init(tfs->TF_PROF_FUNC_POOL_NAME_TX, TF_NUM_PROF_FUNC); - ba_init(tfs->TF_PROF_TCAM_POOL_NAME_RX, TF_NUM_PROF_TCAM); - ba_init(tfs->TF_PROF_TCAM_POOL_NAME_TX, TF_NUM_PROF_TCAM); - ba_init(tfs->TF_EM_PROF_ID_POOL_NAME_RX, TF_NUM_EM_PROF_ID); - ba_init(tfs->TF_EM_PROF_ID_POOL_NAME_TX, TF_NUM_EM_PROF_ID); - - /* TBD, how do we want to handle EM records ?*/ - /* EM Records should not be controlled by way of a pool */ - - ba_init(tfs->TF_WC_TCAM_PROF_ID_POOL_NAME_RX, TF_NUM_WC_PROF_ID); - ba_init(tfs->TF_WC_TCAM_PROF_ID_POOL_NAME_TX, TF_NUM_WC_PROF_ID); - ba_init(tfs->TF_WC_TCAM_POOL_NAME_RX, TF_NUM_WC_TCAM_ROW); - ba_init(tfs->TF_WC_TCAM_POOL_NAME_TX, TF_NUM_WC_TCAM_ROW); - ba_init(tfs->TF_METER_PROF_POOL_NAME_RX, TF_NUM_METER_PROF); - ba_init(tfs->TF_METER_PROF_POOL_NAME_TX, TF_NUM_METER_PROF); - ba_init(tfs->TF_METER_INST_POOL_NAME_RX, TF_NUM_METER); - ba_init(tfs->TF_METER_INST_POOL_NAME_TX, TF_NUM_METER); - ba_init(tfs->TF_MIRROR_POOL_NAME_RX, TF_NUM_MIRROR); - ba_init(tfs->TF_MIRROR_POOL_NAME_TX, TF_NUM_MIRROR); - ba_init(tfs->TF_UPAR_POOL_NAME_RX, TF_NUM_UPAR); - ba_init(tfs->TF_UPAR_POOL_NAME_TX, TF_NUM_UPAR); - - ba_init(tfs->TF_SP_TCAM_POOL_NAME_RX, TF_NUM_SP_TCAM); - ba_init(tfs->TF_SP_TCAM_POOL_NAME_TX, TF_NUM_SP_TCAM); - - ba_init(tfs->TF_FKB_POOL_NAME_RX, TF_NUM_FKB); - ba_init(tfs->TF_FKB_POOL_NAME_TX, TF_NUM_FKB); - - ba_init(tfs->TF_TBL_SCOPE_POOL_NAME_RX, TF_NUM_TBL_SCOPE); - ba_init(tfs->TF_TBL_SCOPE_POOL_NAME_TX, TF_NUM_TBL_SCOPE); - ba_init(tfs->TF_L2_FUNC_POOL_NAME_RX, TF_NUM_L2_FUNC); - ba_init(tfs->TF_L2_FUNC_POOL_NAME_TX, TF_NUM_L2_FUNC); - ba_init(tfs->TF_EPOCH0_POOL_NAME_RX, TF_NUM_EPOCH0); - ba_init(tfs->TF_EPOCH0_POOL_NAME_TX, TF_NUM_EPOCH0); - ba_init(tfs->TF_EPOCH1_POOL_NAME_RX, TF_NUM_EPOCH1); - ba_init(tfs->TF_EPOCH1_POOL_NAME_TX, TF_NUM_EPOCH1); - ba_init(tfs->TF_METADATA_POOL_NAME_RX, TF_NUM_METADATA); - ba_init(tfs->TF_METADATA_POOL_NAME_TX, TF_NUM_METADATA); - ba_init(tfs->TF_CT_STATE_POOL_NAME_RX, TF_NUM_CT_STATE); - ba_init(tfs->TF_CT_STATE_POOL_NAME_TX, TF_NUM_CT_STATE); - ba_init(tfs->TF_RANGE_PROF_POOL_NAME_RX, TF_NUM_RANGE_PROF); - ba_init(tfs->TF_RANGE_PROF_POOL_NAME_TX, TF_NUM_RANGE_PROF); - ba_init(tfs->TF_RANGE_ENTRY_POOL_NAME_RX, TF_NUM_RANGE_ENTRY); - ba_init(tfs->TF_RANGE_ENTRY_POOL_NAME_TX, TF_NUM_RANGE_ENTRY); - ba_init(tfs->TF_LAG_ENTRY_POOL_NAME_RX, TF_NUM_LAG_ENTRY); - ba_init(tfs->TF_LAG_ENTRY_POOL_NAME_TX, TF_NUM_LAG_ENTRY); - - /* Initialization of SRAM Resource Pools - * These pools are set to the TFLIB defined MAX sizes not - * AFM's HW max as to limit the memory consumption - */ - ba_init(tfs->TF_SRAM_FULL_ACTION_POOL_NAME_RX, - TF_RSVD_SRAM_FULL_ACTION_RX); - ba_init(tfs->TF_SRAM_FULL_ACTION_POOL_NAME_TX, - TF_RSVD_SRAM_FULL_ACTION_TX); - /* Only Multicast Group on RX is supported */ - ba_init(tfs->TF_SRAM_MCG_POOL_NAME_RX, - TF_RSVD_SRAM_MCG_RX); - ba_init(tfs->TF_SRAM_ENCAP_8B_POOL_NAME_RX, - TF_RSVD_SRAM_ENCAP_8B_RX); - ba_init(tfs->TF_SRAM_ENCAP_8B_POOL_NAME_TX, - TF_RSVD_SRAM_ENCAP_8B_TX); - ba_init(tfs->TF_SRAM_ENCAP_16B_POOL_NAME_RX, - TF_RSVD_SRAM_ENCAP_16B_RX); - ba_init(tfs->TF_SRAM_ENCAP_16B_POOL_NAME_TX, - TF_RSVD_SRAM_ENCAP_16B_TX); - /* Only Encap 64B on TX is supported */ - ba_init(tfs->TF_SRAM_ENCAP_64B_POOL_NAME_TX, - TF_RSVD_SRAM_ENCAP_64B_TX); - ba_init(tfs->TF_SRAM_SP_SMAC_POOL_NAME_RX, - TF_RSVD_SRAM_SP_SMAC_RX); - ba_init(tfs->TF_SRAM_SP_SMAC_POOL_NAME_TX, - TF_RSVD_SRAM_SP_SMAC_TX); - /* Only SP SMAC IPv4 on TX is supported */ - ba_init(tfs->TF_SRAM_SP_SMAC_IPV4_POOL_NAME_TX, - TF_RSVD_SRAM_SP_SMAC_IPV4_TX); - /* Only SP SMAC IPv6 on TX is supported */ - ba_init(tfs->TF_SRAM_SP_SMAC_IPV6_POOL_NAME_TX, - TF_RSVD_SRAM_SP_SMAC_IPV6_TX); - ba_init(tfs->TF_SRAM_STATS_64B_POOL_NAME_RX, - TF_RSVD_SRAM_COUNTER_64B_RX); - ba_init(tfs->TF_SRAM_STATS_64B_POOL_NAME_TX, - TF_RSVD_SRAM_COUNTER_64B_TX); - ba_init(tfs->TF_SRAM_NAT_SPORT_POOL_NAME_RX, - TF_RSVD_SRAM_NAT_SPORT_RX); - ba_init(tfs->TF_SRAM_NAT_SPORT_POOL_NAME_TX, - TF_RSVD_SRAM_NAT_SPORT_TX); - ba_init(tfs->TF_SRAM_NAT_DPORT_POOL_NAME_RX, - TF_RSVD_SRAM_NAT_DPORT_RX); - ba_init(tfs->TF_SRAM_NAT_DPORT_POOL_NAME_TX, - TF_RSVD_SRAM_NAT_DPORT_TX); - ba_init(tfs->TF_SRAM_NAT_S_IPV4_POOL_NAME_RX, - TF_RSVD_SRAM_NAT_S_IPV4_RX); - ba_init(tfs->TF_SRAM_NAT_S_IPV4_POOL_NAME_TX, - TF_RSVD_SRAM_NAT_S_IPV4_TX); - ba_init(tfs->TF_SRAM_NAT_D_IPV4_POOL_NAME_RX, - TF_RSVD_SRAM_NAT_D_IPV4_RX); - ba_init(tfs->TF_SRAM_NAT_D_IPV4_POOL_NAME_TX, - TF_RSVD_SRAM_NAT_D_IPV4_TX); - - /* Initialization of pools local to TF Core */ - ba_init(tfs->TF_L2_CTXT_REMAP_POOL_NAME_RX, TF_NUM_L2_CTXT_TCAM); - ba_init(tfs->TF_L2_CTXT_REMAP_POOL_NAME_TX, TF_NUM_L2_CTXT_TCAM); + tfp_free((void *)parms->rm_db); + + return rc; } int -tf_rm_allocate_validate(struct tf *tfp) +tf_rm_allocate(struct tf_rm_allocate_parms *parms) { int rc; - int i; + int id; + uint32_t index; + struct tf_rm_new_db *rm_db; + enum tf_rm_elem_cfg_type cfg_type; - for (i = 0; i < TF_DIR_MAX; i++) { - rc = tf_rm_allocate_validate_hw(tfp, i); - if (rc) - return rc; - rc = tf_rm_allocate_validate_sram(tfp, i); - if (rc) - return rc; + TF_CHECK_PARMS2(parms, parms->rm_db); + + rm_db = (struct tf_rm_new_db *)parms->rm_db; + cfg_type = rm_db->db[parms->db_index].cfg_type; + + /* Bail out if not controlled by RM */ + if (cfg_type != TF_RM_ELEM_CFG_HCAPI && + cfg_type != TF_RM_ELEM_CFG_PRIVATE) + return -ENOTSUP; + + /* Bail out if the pool is not valid, should never happen */ + if (rm_db->db[parms->db_index].pool == NULL) { + rc = -ENOTSUP; + TFP_DRV_LOG(ERR, + "%s: Invalid pool for this type:%d, rc:%s\n", + tf_dir_2_str(rm_db->dir), + parms->db_index, + strerror(-rc)); + return rc; } - /* With both HW and SRAM allocated and validated we can - * 'scrub' the reservation on the pools. + /* + * priority 0: allocate from top of the tcam i.e. high + * priority !0: allocate index from bottom i.e lowest */ - tf_rm_reserve_hw(tfp); - tf_rm_reserve_sram(tfp); + if (parms->priority) + id = ba_alloc_reverse(rm_db->db[parms->db_index].pool); + else + id = ba_alloc(rm_db->db[parms->db_index].pool); + if (id == BA_FAIL) { + rc = -ENOMEM; + TFP_DRV_LOG(ERR, + "%s: Allocation failed, rc:%s\n", + tf_dir_2_str(rm_db->dir), + strerror(-rc)); + return rc; + } + + /* Adjust for any non zero start value */ + rc = tf_rm_adjust_index(rm_db->db, + TF_RM_ADJUST_ADD_BASE, + parms->db_index, + id, + &index); + if (rc) { + TFP_DRV_LOG(ERR, + "%s: Alloc adjust of base index failed, rc:%s\n", + tf_dir_2_str(rm_db->dir), + strerror(-rc)); + return -EINVAL; + } + + *parms->index = index; return rc; } int -tf_rm_close(struct tf *tfp) +tf_rm_free(struct tf_rm_free_parms *parms) { int rc; - int rc_close = 0; - int i; - struct tf_rm_entry *hw_entries; - struct tf_rm_entry *hw_flush_entries; - struct tf_rm_entry *sram_entries; - struct tf_rm_entry *sram_flush_entries; - struct tf_session *tfs __rte_unused = - (struct tf_session *)(tfp->session->core_data); - - struct tf_rm_db flush_resc = tfs->resc; - - /* On close it is assumed that the session has already cleaned - * up all its resources, individually, while destroying its - * flows. No checking is performed thus the behavior is as - * follows. - * - * Session RM will signal FW to release session resources. FW - * will perform invalidation of all the allocated entries - * (assures any outstanding resources has been cleared, then - * free the FW RM instance. - * - * Session will then be freed by tf_close_session() thus there - * is no need to clean each resource pool as the whole session - * is going away. - */ - - for (i = 0; i < TF_DIR_MAX; i++) { - if (i == TF_DIR_RX) { - hw_entries = tfs->resc.rx.hw_entry; - hw_flush_entries = flush_resc.rx.hw_entry; - sram_entries = tfs->resc.rx.sram_entry; - sram_flush_entries = flush_resc.rx.sram_entry; - } else { - hw_entries = tfs->resc.tx.hw_entry; - hw_flush_entries = flush_resc.tx.hw_entry; - sram_entries = tfs->resc.tx.sram_entry; - sram_flush_entries = flush_resc.tx.sram_entry; - } + uint32_t adj_index; + struct tf_rm_new_db *rm_db; + enum tf_rm_elem_cfg_type cfg_type; - /* Check for any not previously freed HW resources and - * flush if required. - */ - rc = tf_rm_hw_to_flush(tfs, i, hw_entries, hw_flush_entries); - if (rc) { - rc_close = -ENOTEMPTY; - /* Log error */ - TFP_DRV_LOG(ERR, - "%s, lingering HW resources, rc:%s\n", - tf_dir_2_str(i), - strerror(-rc)); + TF_CHECK_PARMS2(parms, parms->rm_db); - /* Log the entries to be flushed */ - tf_rm_log_hw_flush(i, hw_flush_entries); - } + rm_db = (struct tf_rm_new_db *)parms->rm_db; + cfg_type = rm_db->db[parms->db_index].cfg_type; - /* Check for any not previously freed SRAM resources - * and flush if required. - */ - rc = tf_rm_sram_to_flush(tfs, - i, - sram_entries, - sram_flush_entries); - if (rc) { - rc_close = -ENOTEMPTY; - /* Log error */ - TFP_DRV_LOG(ERR, - "%s, lingering SRAM resources, rc:%s\n", - tf_dir_2_str(i), - strerror(-rc)); + /* Bail out if not controlled by RM */ + if (cfg_type != TF_RM_ELEM_CFG_HCAPI && + cfg_type != TF_RM_ELEM_CFG_PRIVATE) + return -ENOTSUP; - /* Log the entries to be flushed */ - tf_rm_log_sram_flush(i, sram_flush_entries); - } + /* Bail out if the pool is not valid, should never happen */ + if (rm_db->db[parms->db_index].pool == NULL) { + rc = -ENOTSUP; + TFP_DRV_LOG(ERR, + "%s: Invalid pool for this type:%d, rc:%s\n", + tf_dir_2_str(rm_db->dir), + parms->db_index, + strerror(-rc)); + return rc; } - return rc_close; -} + /* Adjust for any non zero start value */ + rc = tf_rm_adjust_index(rm_db->db, + TF_RM_ADJUST_RM_BASE, + parms->db_index, + parms->index, + &adj_index); + if (rc) + return rc; -#if (TF_SHADOW == 1) -int -tf_rm_shadow_db_init(struct tf_session *tfs) -{ - rc = 1; + rc = ba_free(rm_db->db[parms->db_index].pool, adj_index); + /* No logging direction matters and that is not available here */ + if (rc) + return rc; return rc; } -#endif /* TF_SHADOW */ int -tf_rm_lookup_tcam_type_pool(struct tf_session *tfs, - enum tf_dir dir, - enum tf_tcam_tbl_type type, - struct bitalloc **pool) +tf_rm_is_allocated(struct tf_rm_is_allocated_parms *parms) { - int rc = -EOPNOTSUPP; + int rc; + uint32_t adj_index; + struct tf_rm_new_db *rm_db; + enum tf_rm_elem_cfg_type cfg_type; - *pool = NULL; + TF_CHECK_PARMS2(parms, parms->rm_db); - switch (type) { - case TF_TCAM_TBL_TYPE_L2_CTXT_TCAM: - TF_RM_GET_POOLS(tfs, dir, pool, - TF_L2_CTXT_TCAM_POOL_NAME, - rc); - break; - case TF_TCAM_TBL_TYPE_PROF_TCAM: - TF_RM_GET_POOLS(tfs, dir, pool, - TF_PROF_TCAM_POOL_NAME, - rc); - break; - case TF_TCAM_TBL_TYPE_WC_TCAM: - TF_RM_GET_POOLS(tfs, dir, pool, - TF_WC_TCAM_POOL_NAME, - rc); - break; - case TF_TCAM_TBL_TYPE_VEB_TCAM: - case TF_TCAM_TBL_TYPE_SP_TCAM: - case TF_TCAM_TBL_TYPE_CT_RULE_TCAM: - default: - break; - } + rm_db = (struct tf_rm_new_db *)parms->rm_db; + cfg_type = rm_db->db[parms->db_index].cfg_type; - if (rc == -EOPNOTSUPP) { - TFP_DRV_LOG(ERR, - "%s, Tcam type not supported, type:%d\n", - tf_dir_2_str(dir), - type); - return rc; - } else if (rc == -1) { + /* Bail out if not controlled by RM */ + if (cfg_type != TF_RM_ELEM_CFG_HCAPI && + cfg_type != TF_RM_ELEM_CFG_PRIVATE) + return -ENOTSUP; + + /* Bail out if the pool is not valid, should never happen */ + if (rm_db->db[parms->db_index].pool == NULL) { + rc = -ENOTSUP; TFP_DRV_LOG(ERR, - "%s, Tcam type lookup failed, type:%d\n", - tf_dir_2_str(dir), - type); + "%s: Invalid pool for this type:%d, rc:%s\n", + tf_dir_2_str(rm_db->dir), + parms->db_index, + strerror(-rc)); return rc; } - return 0; + /* Adjust for any non zero start value */ + rc = tf_rm_adjust_index(rm_db->db, + TF_RM_ADJUST_RM_BASE, + parms->db_index, + parms->index, + &adj_index); + if (rc) + return rc; + + *parms->allocated = ba_inuse(rm_db->db[parms->db_index].pool, + adj_index); + + return rc; } int -tf_rm_lookup_tbl_type_pool(struct tf_session *tfs, - enum tf_dir dir, - enum tf_tbl_type type, - struct bitalloc **pool) +tf_rm_get_info(struct tf_rm_get_alloc_info_parms *parms) { - int rc = -EOPNOTSUPP; + struct tf_rm_new_db *rm_db; + enum tf_rm_elem_cfg_type cfg_type; - *pool = NULL; + TF_CHECK_PARMS2(parms, parms->rm_db); - switch (type) { - case TF_TBL_TYPE_FULL_ACT_RECORD: - TF_RM_GET_POOLS(tfs, dir, pool, - TF_SRAM_FULL_ACTION_POOL_NAME, - rc); - break; - case TF_TBL_TYPE_MCAST_GROUPS: - /* No pools for TX direction, so bail out */ - if (dir == TF_DIR_TX) - break; - TF_RM_GET_POOLS_RX(tfs, pool, - TF_SRAM_MCG_POOL_NAME); - rc = 0; - break; - case TF_TBL_TYPE_ACT_ENCAP_8B: - TF_RM_GET_POOLS(tfs, dir, pool, - TF_SRAM_ENCAP_8B_POOL_NAME, - rc); - break; - case TF_TBL_TYPE_ACT_ENCAP_16B: - TF_RM_GET_POOLS(tfs, dir, pool, - TF_SRAM_ENCAP_16B_POOL_NAME, - rc); - break; - case TF_TBL_TYPE_ACT_ENCAP_64B: - /* No pools for RX direction, so bail out */ - if (dir == TF_DIR_RX) - break; - TF_RM_GET_POOLS_TX(tfs, pool, - TF_SRAM_ENCAP_64B_POOL_NAME); - rc = 0; - break; - case TF_TBL_TYPE_ACT_SP_SMAC: - TF_RM_GET_POOLS(tfs, dir, pool, - TF_SRAM_SP_SMAC_POOL_NAME, - rc); - break; - case TF_TBL_TYPE_ACT_SP_SMAC_IPV4: - /* No pools for TX direction, so bail out */ - if (dir == TF_DIR_RX) - break; - TF_RM_GET_POOLS_TX(tfs, pool, - TF_SRAM_SP_SMAC_IPV4_POOL_NAME); - rc = 0; - break; - case TF_TBL_TYPE_ACT_SP_SMAC_IPV6: - /* No pools for TX direction, so bail out */ - if (dir == TF_DIR_RX) - break; - TF_RM_GET_POOLS_TX(tfs, pool, - TF_SRAM_SP_SMAC_IPV6_POOL_NAME); - rc = 0; - break; - case TF_TBL_TYPE_ACT_STATS_64: - TF_RM_GET_POOLS(tfs, dir, pool, - TF_SRAM_STATS_64B_POOL_NAME, - rc); - break; - case TF_TBL_TYPE_ACT_MODIFY_SPORT: - TF_RM_GET_POOLS(tfs, dir, pool, - TF_SRAM_NAT_SPORT_POOL_NAME, - rc); - break; - case TF_TBL_TYPE_ACT_MODIFY_IPV4_SRC: - TF_RM_GET_POOLS(tfs, dir, pool, - TF_SRAM_NAT_S_IPV4_POOL_NAME, - rc); - break; - case TF_TBL_TYPE_ACT_MODIFY_IPV4_DEST: - TF_RM_GET_POOLS(tfs, dir, pool, - TF_SRAM_NAT_D_IPV4_POOL_NAME, - rc); - break; - case TF_TBL_TYPE_METER_PROF: - TF_RM_GET_POOLS(tfs, dir, pool, - TF_METER_PROF_POOL_NAME, - rc); - break; - case TF_TBL_TYPE_METER_INST: - TF_RM_GET_POOLS(tfs, dir, pool, - TF_METER_INST_POOL_NAME, - rc); - break; - case TF_TBL_TYPE_MIRROR_CONFIG: - TF_RM_GET_POOLS(tfs, dir, pool, - TF_MIRROR_POOL_NAME, - rc); - break; - case TF_TBL_TYPE_UPAR: - TF_RM_GET_POOLS(tfs, dir, pool, - TF_UPAR_POOL_NAME, - rc); - break; - case TF_TBL_TYPE_EPOCH0: - TF_RM_GET_POOLS(tfs, dir, pool, - TF_EPOCH0_POOL_NAME, - rc); - break; - case TF_TBL_TYPE_EPOCH1: - TF_RM_GET_POOLS(tfs, dir, pool, - TF_EPOCH1_POOL_NAME, - rc); - break; - case TF_TBL_TYPE_METADATA: - TF_RM_GET_POOLS(tfs, dir, pool, - TF_METADATA_POOL_NAME, - rc); - break; - case TF_TBL_TYPE_CT_STATE: - TF_RM_GET_POOLS(tfs, dir, pool, - TF_CT_STATE_POOL_NAME, - rc); - break; - case TF_TBL_TYPE_RANGE_PROF: - TF_RM_GET_POOLS(tfs, dir, pool, - TF_RANGE_PROF_POOL_NAME, - rc); - break; - case TF_TBL_TYPE_RANGE_ENTRY: - TF_RM_GET_POOLS(tfs, dir, pool, - TF_RANGE_ENTRY_POOL_NAME, - rc); - break; - case TF_TBL_TYPE_LAG: - TF_RM_GET_POOLS(tfs, dir, pool, - TF_LAG_ENTRY_POOL_NAME, - rc); - break; - /* Not yet supported */ - case TF_TBL_TYPE_ACT_ENCAP_32B: - case TF_TBL_TYPE_ACT_MODIFY_IPV6_DEST: - case TF_TBL_TYPE_ACT_MODIFY_IPV6_SRC: - case TF_TBL_TYPE_VNIC_SVIF: - break; - /* No bitalloc pools for these types */ - case TF_TBL_TYPE_EXT: - default: - break; - } + rm_db = (struct tf_rm_new_db *)parms->rm_db; + cfg_type = rm_db->db[parms->db_index].cfg_type; - if (rc == -EOPNOTSUPP) { - TFP_DRV_LOG(ERR, - "%s, Table type not supported, type:%d\n", - tf_dir_2_str(dir), - type); - return rc; - } else if (rc == -1) { - TFP_DRV_LOG(ERR, - "%s, Table type lookup failed, type:%d\n", - tf_dir_2_str(dir), - type); - return rc; - } + /* Bail out if not controlled by RM */ + if (cfg_type != TF_RM_ELEM_CFG_HCAPI && + cfg_type != TF_RM_ELEM_CFG_PRIVATE) + return -ENOTSUP; + + memcpy(parms->info, + &rm_db->db[parms->db_index].alloc, + sizeof(struct tf_rm_alloc_info)); return 0; } int -tf_rm_convert_tbl_type(enum tf_tbl_type type, - uint32_t *hcapi_type) +tf_rm_get_hcapi_type(struct tf_rm_get_hcapi_parms *parms) { - int rc = 0; + struct tf_rm_new_db *rm_db; + enum tf_rm_elem_cfg_type cfg_type; - switch (type) { - case TF_TBL_TYPE_FULL_ACT_RECORD: - *hcapi_type = TF_RESC_TYPE_SRAM_FULL_ACTION; - break; - case TF_TBL_TYPE_MCAST_GROUPS: - *hcapi_type = TF_RESC_TYPE_SRAM_MCG; - break; - case TF_TBL_TYPE_ACT_ENCAP_8B: - *hcapi_type = TF_RESC_TYPE_SRAM_ENCAP_8B; - break; - case TF_TBL_TYPE_ACT_ENCAP_16B: - *hcapi_type = TF_RESC_TYPE_SRAM_ENCAP_16B; - break; - case TF_TBL_TYPE_ACT_ENCAP_64B: - *hcapi_type = TF_RESC_TYPE_SRAM_ENCAP_64B; - break; - case TF_TBL_TYPE_ACT_SP_SMAC: - *hcapi_type = TF_RESC_TYPE_SRAM_SP_SMAC; - break; - case TF_TBL_TYPE_ACT_SP_SMAC_IPV4: - *hcapi_type = TF_RESC_TYPE_SRAM_SP_SMAC_IPV4; - break; - case TF_TBL_TYPE_ACT_SP_SMAC_IPV6: - *hcapi_type = TF_RESC_TYPE_SRAM_SP_SMAC_IPV6; - break; - case TF_TBL_TYPE_ACT_STATS_64: - *hcapi_type = TF_RESC_TYPE_SRAM_COUNTER_64B; - break; - case TF_TBL_TYPE_ACT_MODIFY_SPORT: - *hcapi_type = TF_RESC_TYPE_SRAM_NAT_SPORT; - break; - case TF_TBL_TYPE_ACT_MODIFY_DPORT: - *hcapi_type = TF_RESC_TYPE_SRAM_NAT_DPORT; - break; - case TF_TBL_TYPE_ACT_MODIFY_IPV4_SRC: - *hcapi_type = TF_RESC_TYPE_SRAM_NAT_S_IPV4; - break; - case TF_TBL_TYPE_ACT_MODIFY_IPV4_DEST: - *hcapi_type = TF_RESC_TYPE_SRAM_NAT_D_IPV4; - break; - case TF_TBL_TYPE_METER_PROF: - *hcapi_type = TF_RESC_TYPE_HW_METER_PROF; - break; - case TF_TBL_TYPE_METER_INST: - *hcapi_type = TF_RESC_TYPE_HW_METER_INST; - break; - case TF_TBL_TYPE_MIRROR_CONFIG: - *hcapi_type = TF_RESC_TYPE_HW_MIRROR; - break; - case TF_TBL_TYPE_UPAR: - *hcapi_type = TF_RESC_TYPE_HW_UPAR; - break; - case TF_TBL_TYPE_EPOCH0: - *hcapi_type = TF_RESC_TYPE_HW_EPOCH0; - break; - case TF_TBL_TYPE_EPOCH1: - *hcapi_type = TF_RESC_TYPE_HW_EPOCH1; - break; - case TF_TBL_TYPE_METADATA: - *hcapi_type = TF_RESC_TYPE_HW_METADATA; - break; - case TF_TBL_TYPE_CT_STATE: - *hcapi_type = TF_RESC_TYPE_HW_CT_STATE; - break; - case TF_TBL_TYPE_RANGE_PROF: - *hcapi_type = TF_RESC_TYPE_HW_RANGE_PROF; - break; - case TF_TBL_TYPE_RANGE_ENTRY: - *hcapi_type = TF_RESC_TYPE_HW_RANGE_ENTRY; - break; - case TF_TBL_TYPE_LAG: - *hcapi_type = TF_RESC_TYPE_HW_LAG_ENTRY; - break; - /* Not yet supported */ - case TF_TBL_TYPE_ACT_ENCAP_32B: - case TF_TBL_TYPE_ACT_MODIFY_IPV6_DEST: - case TF_TBL_TYPE_ACT_MODIFY_IPV6_SRC: - case TF_TBL_TYPE_VNIC_SVIF: - case TF_TBL_TYPE_EXT: /* No pools for this type */ - default: - *hcapi_type = -1; - rc = -EOPNOTSUPP; - } + TF_CHECK_PARMS2(parms, parms->rm_db); - return rc; + rm_db = (struct tf_rm_new_db *)parms->rm_db; + cfg_type = rm_db->db[parms->db_index].cfg_type; + + /* Bail out if not controlled by RM */ + if (cfg_type != TF_RM_ELEM_CFG_HCAPI && + cfg_type != TF_RM_ELEM_CFG_PRIVATE) + return -ENOTSUP; + + *parms->hcapi_type = rm_db->db[parms->db_index].hcapi_type; + + return 0; } int -tf_rm_convert_index(struct tf_session *tfs, - enum tf_dir dir, - enum tf_tbl_type type, - enum tf_rm_convert_type c_type, - uint32_t index, - uint32_t *convert_index) +tf_rm_get_inuse_count(struct tf_rm_get_inuse_count_parms *parms) { - int rc; - struct tf_rm_resc *resc; - uint32_t hcapi_type; - uint32_t base_index; + int rc = 0; + struct tf_rm_new_db *rm_db; + enum tf_rm_elem_cfg_type cfg_type; - if (dir == TF_DIR_RX) - resc = &tfs->resc.rx; - else if (dir == TF_DIR_TX) - resc = &tfs->resc.tx; - else - return -EOPNOTSUPP; + TF_CHECK_PARMS2(parms, parms->rm_db); - rc = tf_rm_convert_tbl_type(type, &hcapi_type); - if (rc) - return -1; - - switch (type) { - case TF_TBL_TYPE_FULL_ACT_RECORD: - case TF_TBL_TYPE_MCAST_GROUPS: - case TF_TBL_TYPE_ACT_ENCAP_8B: - case TF_TBL_TYPE_ACT_ENCAP_16B: - case TF_TBL_TYPE_ACT_ENCAP_32B: - case TF_TBL_TYPE_ACT_ENCAP_64B: - case TF_TBL_TYPE_ACT_SP_SMAC: - case TF_TBL_TYPE_ACT_SP_SMAC_IPV4: - case TF_TBL_TYPE_ACT_SP_SMAC_IPV6: - case TF_TBL_TYPE_ACT_STATS_64: - case TF_TBL_TYPE_ACT_MODIFY_SPORT: - case TF_TBL_TYPE_ACT_MODIFY_DPORT: - case TF_TBL_TYPE_ACT_MODIFY_IPV4_SRC: - case TF_TBL_TYPE_ACT_MODIFY_IPV4_DEST: - base_index = resc->sram_entry[hcapi_type].start; - break; - case TF_TBL_TYPE_MIRROR_CONFIG: - case TF_TBL_TYPE_METER_PROF: - case TF_TBL_TYPE_METER_INST: - case TF_TBL_TYPE_UPAR: - case TF_TBL_TYPE_EPOCH0: - case TF_TBL_TYPE_EPOCH1: - case TF_TBL_TYPE_METADATA: - case TF_TBL_TYPE_CT_STATE: - case TF_TBL_TYPE_RANGE_PROF: - case TF_TBL_TYPE_RANGE_ENTRY: - case TF_TBL_TYPE_LAG: - base_index = resc->hw_entry[hcapi_type].start; - break; - /* Not yet supported */ - case TF_TBL_TYPE_VNIC_SVIF: - case TF_TBL_TYPE_EXT: /* No pools for this type */ - default: - return -EOPNOTSUPP; - } + rm_db = (struct tf_rm_new_db *)parms->rm_db; + cfg_type = rm_db->db[parms->db_index].cfg_type; - switch (c_type) { - case TF_RM_CONVERT_RM_BASE: - *convert_index = index - base_index; - break; - case TF_RM_CONVERT_ADD_BASE: - *convert_index = index + base_index; - break; - default: - return -EOPNOTSUPP; + /* Bail out if not controlled by RM */ + if (cfg_type != TF_RM_ELEM_CFG_HCAPI && + cfg_type != TF_RM_ELEM_CFG_PRIVATE) + return -ENOTSUP; + + /* Bail silently (no logging), if the pool is not valid there + * was no elements allocated for it. + */ + if (rm_db->db[parms->db_index].pool == NULL) { + *parms->count = 0; + return 0; } - return 0; + *parms->count = ba_inuse_count(rm_db->db[parms->db_index].pool); + + return rc; + } diff --git a/drivers/net/bnxt/tf_core/tf_rm.h b/drivers/net/bnxt/tf_core/tf_rm.h index 1a09f13a78..3b818f2af5 100644 --- a/drivers/net/bnxt/tf_core/tf_rm.h +++ b/drivers/net/bnxt/tf_core/tf_rm.h @@ -3,301 +3,444 @@ * All rights reserved. */ -#ifndef TF_RM_H_ -#define TF_RM_H_ +#ifndef TF_RM_NEW_H_ +#define TF_RM_NEW_H_ -#include "tf_resources.h" #include "tf_core.h" #include "bitalloc.h" +#include "tf_device.h" struct tf; -struct tf_session; -/* Internal macro to determine appropriate allocation pools based on - * DIRECTION parm, also performs error checking for DIRECTION parm. The - * SESSION_POOL and SESSION pointers are set appropriately upon - * successful return (the GLOBAL_POOL is used to globally manage - * resource allocation and the SESSION_POOL is used to track the - * resources that have been allocated to the session) +/** + * The Resource Manager (RM) module provides basic DB handling for + * internal resources. These resources exists within the actual device + * and are controlled by the HCAPI Resource Manager running on the + * firmware. + * + * The RM DBs are all intended to be indexed using TF types there for + * a lookup requires no additional conversion. The DB configuration + * specifies the TF Type to HCAPI Type mapping and it becomes the + * responsibility of the DB initialization to handle this static + * mapping. + * + * Accessor functions are providing access to the DB, thus hiding the + * implementation. * - * parameters: - * struct tfp *tfp - * enum tf_dir direction - * struct bitalloc **session_pool - * string base_pool_name - used to form pointers to the - * appropriate bit allocation - * pools, both directions of the - * session pools must have same - * base name, for example if - * POOL_NAME is feat_pool: - the - * ptr's to the session pools - * are feat_pool_rx feat_pool_tx + * The RM DB will work on its initial allocated sizes so the + * capability of dynamically growing a particular resource is not + * possible. If this capability later becomes a requirement then the + * MAX pool size of the Chip œneeds to be added to the tf_rm_elem_info + * structure and several new APIs would need to be added to allow for + * growth of a single TF resource type. * - * int rc - return code - * 0 - Success - * -1 - invalid DIRECTION parm + * The access functions does not check for NULL pointers as it's a + * support module, not called directly. */ -#define TF_RM_GET_POOLS(tfs, direction, session_pool, pool_name, rc) do { \ - (rc) = 0; \ - if ((direction) == TF_DIR_RX) { \ - *(session_pool) = (tfs)->pool_name ## _RX; \ - } else if ((direction) == TF_DIR_TX) { \ - *(session_pool) = (tfs)->pool_name ## _TX; \ - } else { \ - rc = -1; \ - } \ - } while (0) -#define TF_RM_GET_POOLS_RX(tfs, session_pool, pool_name) \ - (*(session_pool) = (tfs)->pool_name ## _RX) +/** + * Resource reservation single entry result. Used when accessing HCAPI + * RM on the firmware. + */ +struct tf_rm_new_entry { + /** Starting index of the allocated resource */ + uint16_t start; + /** Number of allocated elements */ + uint16_t stride; +}; -#define TF_RM_GET_POOLS_TX(tfs, session_pool, pool_name) \ - (*(session_pool) = (tfs)->pool_name ## _TX) +/** + * RM Element configuration enumeration. Used by the Device to + * indicate how the RM elements the DB consists off, are to be + * configured at time of DB creation. The TF may present types to the + * ULP layer that is not controlled by HCAPI within the Firmware. + */ +enum tf_rm_elem_cfg_type { + /** No configuration */ + TF_RM_ELEM_CFG_NULL, + /** HCAPI 'controlled', uses a Pool for internal storage */ + TF_RM_ELEM_CFG_HCAPI, + /** Private thus not HCAPI 'controlled', creates a Pool for storage */ + TF_RM_ELEM_CFG_PRIVATE, + /** + * Shared element thus it belongs to a shared FW Session and + * is not controlled by the Host. + */ + TF_RM_ELEM_CFG_SHARED, + TF_RM_TYPE_MAX +}; /** - * Resource query single entry + * RM Reservation strategy enumeration. Type of strategy comes from + * the HCAPI RM QCAPS handshake. */ -struct tf_rm_query_entry { - /** Minimum guaranteed number of elements */ - uint16_t min; - /** Maximum non-guaranteed number of elements */ - uint16_t max; +enum tf_rm_resc_resv_strategy { + TF_RM_RESC_RESV_STATIC_PARTITION, + TF_RM_RESC_RESV_STRATEGY_1, + TF_RM_RESC_RESV_STRATEGY_2, + TF_RM_RESC_RESV_STRATEGY_3, + TF_RM_RESC_RESV_MAX }; /** - * Resource single entry + * RM Element configuration structure, used by the Device to configure + * how an individual TF type is configured in regard to the HCAPI RM + * of same type. */ -struct tf_rm_entry { - /** Starting index of the allocated resource */ - uint16_t start; - /** Number of allocated elements */ - uint16_t stride; +struct tf_rm_element_cfg { + /** + * RM Element config controls how the DB for that element is + * processed. + */ + enum tf_rm_elem_cfg_type cfg_type; + + /* If a HCAPI to TF type conversion is required then TF type + * can be added here. + */ + + /** + * HCAPI RM Type for the element. Used for TF to HCAPI type + * conversion. + */ + uint16_t hcapi_type; }; /** - * Resource query array of HW entities + * Allocation information for a single element. */ -struct tf_rm_hw_query { - /** array of HW resource entries */ - struct tf_rm_query_entry hw_query[TF_RESC_TYPE_HW_MAX]; +struct tf_rm_alloc_info { + /** + * HCAPI RM allocated range information. + * + * NOTE: + * In case of dynamic allocation support this would have + * to be changed to linked list of tf_rm_entry instead. + */ + struct tf_rm_new_entry entry; }; /** - * Resource allocation array of HW entities + * Create RM DB parameters */ -struct tf_rm_hw_alloc { - /** array of HW resource entries */ - uint16_t hw_num[TF_RESC_TYPE_HW_MAX]; +struct tf_rm_create_db_parms { + /** + * [in] Device module type. Used for logging purposes. + */ + enum tf_device_module_type type; + /** + * [in] Receive or transmit direction. + */ + enum tf_dir dir; + /** + * [in] Number of elements. + */ + uint16_t num_elements; + /** + * [in] Parameter structure array. Array size is num_elements. + */ + struct tf_rm_element_cfg *cfg; + /** + * Resource allocation count array. This array content + * originates from the tf_session_resources that is passed in + * on session open. + * Array size is num_elements. + */ + uint16_t *alloc_cnt; + /** + * [out] RM DB Handle + */ + void **rm_db; }; /** - * Resource query array of SRAM entities + * Free RM DB parameters */ -struct tf_rm_sram_query { - /** array of SRAM resource entries */ - struct tf_rm_query_entry sram_query[TF_RESC_TYPE_SRAM_MAX]; +struct tf_rm_free_db_parms { + /** + * [in] Receive or transmit direction + */ + enum tf_dir dir; + /** + * [in] RM DB Handle + */ + void *rm_db; }; /** - * Resource allocation array of SRAM entities + * Allocate RM parameters for a single element */ -struct tf_rm_sram_alloc { - /** array of SRAM resource entries */ - uint16_t sram_num[TF_RESC_TYPE_SRAM_MAX]; +struct tf_rm_allocate_parms { + /** + * [in] RM DB Handle + */ + void *rm_db; + /** + * [in] DB Index, indicates which DB entry to perform the + * action on. + */ + uint16_t db_index; + /** + * [in] Pointer to the allocated index in normalized + * form. Normalized means the index has been adjusted, + * i.e. Full Action Record offsets. + */ + uint32_t *index; + /** + * [in] Priority, indicates the priority of the entry + * priority 0: allocate from top of the tcam (from index 0 + * or lowest available index) + * priority !0: allocate from bottom of the tcam (from highest + * available index) + */ + uint32_t priority; }; /** - * Resource Manager arrays for a single direction + * Free RM parameters for a single element */ -struct tf_rm_resc { - /** array of HW resource entries */ - struct tf_rm_entry hw_entry[TF_RESC_TYPE_HW_MAX]; - /** array of SRAM resource entries */ - struct tf_rm_entry sram_entry[TF_RESC_TYPE_SRAM_MAX]; +struct tf_rm_free_parms { + /** + * [in] RM DB Handle + */ + void *rm_db; + /** + * [in] DB Index, indicates which DB entry to perform the + * action on. + */ + uint16_t db_index; + /** + * [in] Index to free + */ + uint16_t index; }; /** - * Resource Manager Database + * Is Allocated parameters for a single element */ -struct tf_rm_db { - struct tf_rm_resc rx; - struct tf_rm_resc tx; +struct tf_rm_is_allocated_parms { + /** + * [in] RM DB Handle + */ + void *rm_db; + /** + * [in] DB Index, indicates which DB entry to perform the + * action on. + */ + uint16_t db_index; + /** + * [in] Index to free + */ + uint32_t index; + /** + * [in] Pointer to flag that indicates the state of the query + */ + int *allocated; }; /** - * Helper function used to convert HW HCAPI resource type to a string. + * Get Allocation information for a single element */ -const char -*tf_hcapi_hw_2_str(enum tf_resource_type_hw hw_type); +struct tf_rm_get_alloc_info_parms { + /** + * [in] RM DB Handle + */ + void *rm_db; + /** + * [in] DB Index, indicates which DB entry to perform the + * action on. + */ + uint16_t db_index; + /** + * [out] Pointer to the requested allocation information for + * the specified db_index + */ + struct tf_rm_alloc_info *info; +}; /** - * Helper function used to convert SRAM HCAPI resource type to a string. + * Get HCAPI type parameters for a single element */ -const char -*tf_hcapi_sram_2_str(enum tf_resource_type_sram sram_type); +struct tf_rm_get_hcapi_parms { + /** + * [in] RM DB Handle + */ + void *rm_db; + /** + * [in] DB Index, indicates which DB entry to perform the + * action on. + */ + uint16_t db_index; + /** + * [out] Pointer to the hcapi type for the specified db_index + */ + uint16_t *hcapi_type; +}; /** - * Initializes the Resource Manager and the associated database - * entries for HW and SRAM resources. Must be called before any other - * Resource Manager functions. + * Get InUse count parameters for single element + */ +struct tf_rm_get_inuse_count_parms { + /** + * [in] RM DB Handle + */ + void *rm_db; + /** + * [in] DB Index, indicates which DB entry to perform the + * action on. + */ + uint16_t db_index; + /** + * [out] Pointer to the inuse count for the specified db_index + */ + uint16_t *count; +}; + +/** + * @page rm Resource Manager * - * [in] tfp - * Pointer to TF handle + * @ref tf_rm_create_db + * + * @ref tf_rm_free_db + * + * @ref tf_rm_allocate + * + * @ref tf_rm_free + * + * @ref tf_rm_is_allocated + * + * @ref tf_rm_get_info + * + * @ref tf_rm_get_hcapi_type + * + * @ref tf_rm_get_inuse_count */ -void tf_rm_init(struct tf *tfp); /** - * Allocates and validates both HW and SRAM resources per the NVM - * configuration. If any allocation fails all resources for the - * session is deallocated. + * Creates and fills a Resource Manager (RM) DB with requested + * elements. The DB is indexed per the parms structure. * * [in] tfp - * Pointer to TF handle + * Pointer to TF handle, used for HCAPI communication + * + * [in] parms + * Pointer to create parameters * * Returns * - (0) if successful. * - (-EINVAL) on failure. */ -int tf_rm_allocate_validate(struct tf *tfp); +/* + * NOTE: + * - Fail on parameter check + * - Fail on DB creation, i.e. alloc amount is not possible or validation fails + * - Fail on DB creation if DB already exist + * + * - Allocs local DB + * - Does hcapi qcaps + * - Does hcapi reservation + * - Populates the pool with allocated elements + * - Returns handle to the created DB + */ +int tf_rm_create_db(struct tf *tfp, + struct tf_rm_create_db_parms *parms); /** - * Closes the Resource Manager and frees all allocated resources per - * the associated database. + * Closes the Resource Manager (RM) DB and frees all allocated + * resources per the associated database. * * [in] tfp - * Pointer to TF handle + * Pointer to TF handle, used for HCAPI communication + * + * [in] parms + * Pointer to free parameters * * Returns * - (0) if successful. * - (-EINVAL) on failure. - * - (-ENOTEMPTY) if resources are not cleaned up before close */ -int tf_rm_close(struct tf *tfp); +int tf_rm_free_db(struct tf *tfp, + struct tf_rm_free_db_parms *parms); -#if (TF_SHADOW == 1) /** - * Initializes Shadow DB of configuration elements + * Allocates a single element for the type specified, within the DB. * - * [in] tfs - * Pointer to TF Session + * [in] parms + * Pointer to allocate parameters * - * Returns: - * 0 - Success + * Returns + * - (0) if successful. + * - (-EINVAL) on failure. + * - (-ENOMEM) if pool is empty */ -int tf_rm_shadow_db_init(struct tf_session *tfs); -#endif /* TF_SHADOW */ +int tf_rm_allocate(struct tf_rm_allocate_parms *parms); /** - * Perform a Session Pool lookup using the Tcam table type. - * - * Function will print error msg if tcam type is unsupported or lookup - * failed. + * Free's a single element for the type specified, within the DB. * - * [in] tfs - * Pointer to TF Session + * [in] parms + * Pointer to free parameters * - * [in] type - * Type of the object - * - * [in] dir - * Receive or transmit direction - * - * [in/out] session_pool - * Session pool - * - * Returns: - * 0 - Success will set the **pool - * -EOPNOTSUPP - Type is not supported + * Returns + * - (0) if successful. + * - (-EINVAL) on failure. */ -int -tf_rm_lookup_tcam_type_pool(struct tf_session *tfs, - enum tf_dir dir, - enum tf_tcam_tbl_type type, - struct bitalloc **pool); +int tf_rm_free(struct tf_rm_free_parms *parms); /** - * Perform a Session Pool lookup using the Table type. - * - * Function will print error msg if table type is unsupported or - * lookup failed. - * - * [in] tfs - * Pointer to TF Session - * - * [in] type - * Type of the object + * Performs an allocation verification check on a specified element. * - * [in] dir - * Receive or transmit direction + * [in] parms + * Pointer to is allocated parameters * - * [in/out] session_pool - * Session pool - * - * Returns: - * 0 - Success will set the **pool - * -EOPNOTSUPP - Type is not supported + * Returns + * - (0) if successful. + * - (-EINVAL) on failure. */ -int -tf_rm_lookup_tbl_type_pool(struct tf_session *tfs, - enum tf_dir dir, - enum tf_tbl_type type, - struct bitalloc **pool); +/* + * NOTE: + * - If pool is set to Chip MAX, then the query index must be checked + * against the allocated range and query index must be allocated as well. + * - If pool is allocated size only, then check if query index is allocated. + */ +int tf_rm_is_allocated(struct tf_rm_is_allocated_parms *parms); /** - * Converts the TF Table Type to internal HCAPI_TYPE - * - * [in] type - * Type to be converted + * Retrieves an elements allocation information from the Resource + * Manager (RM) DB. * - * [in/out] hcapi_type - * Converted type + * [in] parms + * Pointer to get info parameters * - * Returns: - * 0 - Success will set the *hcapi_type - * -EOPNOTSUPP - Type is not supported + * Returns + * - (0) if successful. + * - (-EINVAL) on failure. */ -int -tf_rm_convert_tbl_type(enum tf_tbl_type type, - uint32_t *hcapi_type); +int tf_rm_get_info(struct tf_rm_get_alloc_info_parms *parms); /** - * TF RM Convert of index methods. + * Performs a lookup in the Resource Manager DB and retrieves the + * requested HCAPI RM type. + * + * [in] parms + * Pointer to get hcapi parameters + * + * Returns + * - (0) if successful. + * - (-EINVAL) on failure. */ -enum tf_rm_convert_type { - /** Adds the base of the Session Pool to the index */ - TF_RM_CONVERT_ADD_BASE, - /** Removes the Session Pool base from the index */ - TF_RM_CONVERT_RM_BASE -}; +int tf_rm_get_hcapi_type(struct tf_rm_get_hcapi_parms *parms); /** - * Provides conversion of the Table Type index in relation to the - * Session Pool base. - * - * [in] tfs - * Pointer to TF Session - * - * [in] dir - * Receive or transmit direction - * - * [in] type - * Type of the object + * Performs a lookup in the Resource Manager DB and retrieves the + * requested HCAPI RM type inuse count. * - * [in] c_type - * Type of conversion to perform + * [in] parms + * Pointer to get inuse parameters * - * [in] index - * Index to be converted - * - * [in/out] convert_index - * Pointer to the converted index + * Returns + * - (0) if successful. + * - (-EINVAL) on failure. */ -int -tf_rm_convert_index(struct tf_session *tfs, - enum tf_dir dir, - enum tf_tbl_type type, - enum tf_rm_convert_type c_type, - uint32_t index, - uint32_t *convert_index); +int tf_rm_get_inuse_count(struct tf_rm_get_inuse_count_parms *parms); -#endif /* TF_RM_H_ */ +#endif /* TF_RM_NEW_H_ */ diff --git a/drivers/net/bnxt/tf_core/tf_rm_new.c b/drivers/net/bnxt/tf_core/tf_rm_new.c deleted file mode 100644 index 2d9be654a5..0000000000 --- a/drivers/net/bnxt/tf_core/tf_rm_new.c +++ /dev/null @@ -1,907 +0,0 @@ -/* SPDX-License-Identifier: BSD-3-Clause - * Copyright(c) 2019-2020 Broadcom - * All rights reserved. - */ - -#include - -#include - -#include - -#include "tf_rm_new.h" -#include "tf_common.h" -#include "tf_util.h" -#include "tf_session.h" -#include "tf_device.h" -#include "tfp.h" -#include "tf_msg.h" - -/** - * Generic RM Element data type that an RM DB is build upon. - */ -struct tf_rm_element { - /** - * RM Element configuration type. If Private then the - * hcapi_type can be ignored. If Null then the element is not - * valid for the device. - */ - enum tf_rm_elem_cfg_type cfg_type; - - /** - * HCAPI RM Type for the element. - */ - uint16_t hcapi_type; - - /** - * HCAPI RM allocated range information for the element. - */ - struct tf_rm_alloc_info alloc; - - /** - * Bit allocator pool for the element. Pool size is controlled - * by the struct tf_session_resources at time of session creation. - * Null indicates that the element is not used for the device. - */ - struct bitalloc *pool; -}; - -/** - * TF RM DB definition - */ -struct tf_rm_new_db { - /** - * Number of elements in the DB - */ - uint16_t num_entries; - - /** - * Direction this DB controls. - */ - enum tf_dir dir; - - /** - * Module type, used for logging purposes. - */ - enum tf_device_module_type type; - - /** - * The DB consists of an array of elements - */ - struct tf_rm_element *db; -}; - -/** - * Adjust an index according to the allocation information. - * - * All resources are controlled in a 0 based pool. Some resources, by - * design, are not 0 based, i.e. Full Action Records (SRAM) thus they - * need to be adjusted before they are handed out. - * - * [in] cfg - * Pointer to the DB configuration - * - * [in] reservations - * Pointer to the allocation values associated with the module - * - * [in] count - * Number of DB configuration elements - * - * [out] valid_count - * Number of HCAPI entries with a reservation value greater than 0 - * - * Returns: - * 0 - Success - * - EOPNOTSUPP - Operation not supported - */ -static void -tf_rm_count_hcapi_reservations(enum tf_dir dir, - enum tf_device_module_type type, - struct tf_rm_element_cfg *cfg, - uint16_t *reservations, - uint16_t count, - uint16_t *valid_count) -{ - int i; - uint16_t cnt = 0; - - for (i = 0; i < count; i++) { - if (cfg[i].cfg_type == TF_RM_ELEM_CFG_HCAPI && - reservations[i] > 0) - cnt++; - - /* Only log msg if a type is attempted reserved and - * not supported. We ignore EM module as its using a - * split configuration array thus it would fail for - * this type of check. - */ - if (type != TF_DEVICE_MODULE_TYPE_EM && - cfg[i].cfg_type == TF_RM_ELEM_CFG_NULL && - reservations[i] > 0) { - TFP_DRV_LOG(ERR, - "%s, %s, %s allocation not supported\n", - tf_device_module_type_2_str(type), - tf_dir_2_str(dir), - tf_device_module_type_subtype_2_str(type, i)); - printf("%s, %s, %s allocation of %d not supported\n", - tf_device_module_type_2_str(type), - tf_dir_2_str(dir), - tf_device_module_type_subtype_2_str(type, i), - reservations[i]); - } - } - - *valid_count = cnt; -} - -/** - * Resource Manager Adjust of base index definitions. - */ -enum tf_rm_adjust_type { - TF_RM_ADJUST_ADD_BASE, /**< Adds base to the index */ - TF_RM_ADJUST_RM_BASE /**< Removes base from the index */ -}; - -/** - * Adjust an index according to the allocation information. - * - * All resources are controlled in a 0 based pool. Some resources, by - * design, are not 0 based, i.e. Full Action Records (SRAM) thus they - * need to be adjusted before they are handed out. - * - * [in] db - * Pointer to the db, used for the lookup - * - * [in] action - * Adjust action - * - * [in] db_index - * DB index for the element type - * - * [in] index - * Index to convert - * - * [out] adj_index - * Adjusted index - * - * Returns: - * 0 - Success - * - EOPNOTSUPP - Operation not supported - */ -static int -tf_rm_adjust_index(struct tf_rm_element *db, - enum tf_rm_adjust_type action, - uint32_t db_index, - uint32_t index, - uint32_t *adj_index) -{ - int rc = 0; - uint32_t base_index; - - base_index = db[db_index].alloc.entry.start; - - switch (action) { - case TF_RM_ADJUST_RM_BASE: - *adj_index = index - base_index; - break; - case TF_RM_ADJUST_ADD_BASE: - *adj_index = index + base_index; - break; - default: - return -EOPNOTSUPP; - } - - return rc; -} - -/** - * Logs an array of found residual entries to the console. - * - * [in] dir - * Receive or transmit direction - * - * [in] type - * Type of Device Module - * - * [in] count - * Number of entries in the residual array - * - * [in] residuals - * Pointer to an array of residual entries. Array is index same as - * the DB in which this function is used. Each entry holds residual - * value for that entry. - */ -static void -tf_rm_log_residuals(enum tf_dir dir, - enum tf_device_module_type type, - uint16_t count, - uint16_t *residuals) -{ - int i; - - /* Walk the residual array and log the types that wasn't - * cleaned up to the console. - */ - for (i = 0; i < count; i++) { - if (residuals[i] != 0) - TFP_DRV_LOG(ERR, - "%s, %s was not cleaned up, %d outstanding\n", - tf_dir_2_str(dir), - tf_device_module_type_subtype_2_str(type, i), - residuals[i]); - } -} - -/** - * Performs a check of the passed in DB for any lingering elements. If - * a resource type was found to not have been cleaned up by the caller - * then its residual values are recorded, logged and passed back in an - * allocate reservation array that the caller can pass to the FW for - * cleanup. - * - * [in] db - * Pointer to the db, used for the lookup - * - * [out] resv_size - * Pointer to the reservation size of the generated reservation - * array. - * - * [in/out] resv - * Pointer Pointer to a reservation array. The reservation array is - * allocated after the residual scan and holds any found residual - * entries. Thus it can be smaller than the DB that the check was - * performed on. Array must be freed by the caller. - * - * [out] residuals_present - * Pointer to a bool flag indicating if residual was present in the - * DB - * - * Returns: - * 0 - Success - * - EOPNOTSUPP - Operation not supported - */ -static int -tf_rm_check_residuals(struct tf_rm_new_db *rm_db, - uint16_t *resv_size, - struct tf_rm_resc_entry **resv, - bool *residuals_present) -{ - int rc; - int i; - int f; - uint16_t count; - uint16_t found; - uint16_t *residuals = NULL; - uint16_t hcapi_type; - struct tf_rm_get_inuse_count_parms iparms; - struct tf_rm_get_alloc_info_parms aparms; - struct tf_rm_get_hcapi_parms hparms; - struct tf_rm_alloc_info info; - struct tfp_calloc_parms cparms; - struct tf_rm_resc_entry *local_resv = NULL; - - /* Create array to hold the entries that have residuals */ - cparms.nitems = rm_db->num_entries; - cparms.size = sizeof(uint16_t); - cparms.alignment = 0; - rc = tfp_calloc(&cparms); - if (rc) - return rc; - - residuals = (uint16_t *)cparms.mem_va; - - /* Traverse the DB and collect any residual elements */ - iparms.rm_db = rm_db; - iparms.count = &count; - for (i = 0, found = 0; i < rm_db->num_entries; i++) { - iparms.db_index = i; - rc = tf_rm_get_inuse_count(&iparms); - /* Not a device supported entry, just skip */ - if (rc == -ENOTSUP) - continue; - if (rc) - goto cleanup_residuals; - - if (count) { - found++; - residuals[i] = count; - *residuals_present = true; - } - } - - if (*residuals_present) { - /* Populate a reduced resv array with only the entries - * that have residuals. - */ - cparms.nitems = found; - cparms.size = sizeof(struct tf_rm_resc_entry); - cparms.alignment = 0; - rc = tfp_calloc(&cparms); - if (rc) - return rc; - - local_resv = (struct tf_rm_resc_entry *)cparms.mem_va; - - aparms.rm_db = rm_db; - hparms.rm_db = rm_db; - hparms.hcapi_type = &hcapi_type; - for (i = 0, f = 0; i < rm_db->num_entries; i++) { - if (residuals[i] == 0) - continue; - aparms.db_index = i; - aparms.info = &info; - rc = tf_rm_get_info(&aparms); - if (rc) - goto cleanup_all; - - hparms.db_index = i; - rc = tf_rm_get_hcapi_type(&hparms); - if (rc) - goto cleanup_all; - - local_resv[f].type = hcapi_type; - local_resv[f].start = info.entry.start; - local_resv[f].stride = info.entry.stride; - f++; - } - *resv_size = found; - } - - tf_rm_log_residuals(rm_db->dir, - rm_db->type, - rm_db->num_entries, - residuals); - - tfp_free((void *)residuals); - *resv = local_resv; - - return 0; - - cleanup_all: - tfp_free((void *)local_resv); - *resv = NULL; - cleanup_residuals: - tfp_free((void *)residuals); - - return rc; -} - -int -tf_rm_create_db(struct tf *tfp, - struct tf_rm_create_db_parms *parms) -{ - int rc; - int i; - int j; - struct tf_session *tfs; - struct tf_dev_info *dev; - uint16_t max_types; - struct tfp_calloc_parms cparms; - struct tf_rm_resc_req_entry *query; - enum tf_rm_resc_resv_strategy resv_strategy; - struct tf_rm_resc_req_entry *req; - struct tf_rm_resc_entry *resv; - struct tf_rm_new_db *rm_db; - struct tf_rm_element *db; - uint32_t pool_size; - uint16_t hcapi_items; - - TF_CHECK_PARMS2(tfp, parms); - - /* Retrieve the session information */ - rc = tf_session_get_session(tfp, &tfs); - if (rc) - return rc; - - /* Retrieve device information */ - rc = tf_session_get_device(tfs, &dev); - if (rc) - return rc; - - /* Need device max number of elements for the RM QCAPS */ - rc = dev->ops->tf_dev_get_max_types(tfp, &max_types); - if (rc) - return rc; - - cparms.nitems = max_types; - cparms.size = sizeof(struct tf_rm_resc_req_entry); - cparms.alignment = 0; - rc = tfp_calloc(&cparms); - if (rc) - return rc; - - query = (struct tf_rm_resc_req_entry *)cparms.mem_va; - - /* Get Firmware Capabilities */ - rc = tf_msg_session_resc_qcaps(tfp, - parms->dir, - max_types, - query, - &resv_strategy); - if (rc) - return rc; - - /* Process capabilities against DB requirements. However, as a - * DB can hold elements that are not HCAPI we can reduce the - * req msg content by removing those out of the request yet - * the DB holds them all as to give a fast lookup. We can also - * remove entries where there are no request for elements. - */ - tf_rm_count_hcapi_reservations(parms->dir, - parms->type, - parms->cfg, - parms->alloc_cnt, - parms->num_elements, - &hcapi_items); - - /* Alloc request, alignment already set */ - cparms.nitems = (size_t)hcapi_items; - cparms.size = sizeof(struct tf_rm_resc_req_entry); - rc = tfp_calloc(&cparms); - if (rc) - return rc; - req = (struct tf_rm_resc_req_entry *)cparms.mem_va; - - /* Alloc reservation, alignment and nitems already set */ - cparms.size = sizeof(struct tf_rm_resc_entry); - rc = tfp_calloc(&cparms); - if (rc) - return rc; - resv = (struct tf_rm_resc_entry *)cparms.mem_va; - - /* Build the request */ - for (i = 0, j = 0; i < parms->num_elements; i++) { - /* Skip any non HCAPI cfg elements */ - if (parms->cfg[i].cfg_type == TF_RM_ELEM_CFG_HCAPI) { - /* Only perform reservation for entries that - * has been requested - */ - if (parms->alloc_cnt[i] == 0) - continue; - - /* Verify that we can get the full amount - * allocated per the qcaps availability. - */ - if (parms->alloc_cnt[i] <= - query[parms->cfg[i].hcapi_type].max) { - req[j].type = parms->cfg[i].hcapi_type; - req[j].min = parms->alloc_cnt[i]; - req[j].max = parms->alloc_cnt[i]; - j++; - } else { - TFP_DRV_LOG(ERR, - "%s: Resource failure, type:%d\n", - tf_dir_2_str(parms->dir), - parms->cfg[i].hcapi_type); - TFP_DRV_LOG(ERR, - "req:%d, avail:%d\n", - parms->alloc_cnt[i], - query[parms->cfg[i].hcapi_type].max); - return -EINVAL; - } - } - } - - rc = tf_msg_session_resc_alloc(tfp, - parms->dir, - hcapi_items, - req, - resv); - if (rc) - return rc; - - /* Build the RM DB per the request */ - cparms.nitems = 1; - cparms.size = sizeof(struct tf_rm_new_db); - rc = tfp_calloc(&cparms); - if (rc) - return rc; - rm_db = (void *)cparms.mem_va; - - /* Build the DB within RM DB */ - cparms.nitems = parms->num_elements; - cparms.size = sizeof(struct tf_rm_element); - rc = tfp_calloc(&cparms); - if (rc) - return rc; - rm_db->db = (struct tf_rm_element *)cparms.mem_va; - - db = rm_db->db; - for (i = 0, j = 0; i < parms->num_elements; i++) { - db[i].cfg_type = parms->cfg[i].cfg_type; - db[i].hcapi_type = parms->cfg[i].hcapi_type; - - /* Skip any non HCAPI types as we didn't include them - * in the reservation request. - */ - if (parms->cfg[i].cfg_type != TF_RM_ELEM_CFG_HCAPI) - continue; - - /* If the element didn't request an allocation no need - * to create a pool nor verify if we got a reservation. - */ - if (parms->alloc_cnt[i] == 0) - continue; - - /* If the element had requested an allocation and that - * allocation was a success (full amount) then - * allocate the pool. - */ - if (parms->alloc_cnt[i] == resv[j].stride) { - db[i].alloc.entry.start = resv[j].start; - db[i].alloc.entry.stride = resv[j].stride; - - printf("Entry:%d Start:%d Stride:%d\n", - i, - resv[j].start, - resv[j].stride); - - /* Create pool */ - pool_size = (BITALLOC_SIZEOF(resv[j].stride) / - sizeof(struct bitalloc)); - /* Alloc request, alignment already set */ - cparms.nitems = pool_size; - cparms.size = sizeof(struct bitalloc); - rc = tfp_calloc(&cparms); - if (rc) { - TFP_DRV_LOG(ERR, - "%s: Pool alloc failed, type:%d\n", - tf_dir_2_str(parms->dir), - db[i].cfg_type); - goto fail; - } - db[i].pool = (struct bitalloc *)cparms.mem_va; - - rc = ba_init(db[i].pool, resv[j].stride); - if (rc) { - TFP_DRV_LOG(ERR, - "%s: Pool init failed, type:%d\n", - tf_dir_2_str(parms->dir), - db[i].cfg_type); - goto fail; - } - j++; - } else { - /* Bail out as we want what we requested for - * all elements, not any less. - */ - TFP_DRV_LOG(ERR, - "%s: Alloc failed, type:%d\n", - tf_dir_2_str(parms->dir), - db[i].cfg_type); - TFP_DRV_LOG(ERR, - "req:%d, alloc:%d\n", - parms->alloc_cnt[i], - resv[j].stride); - goto fail; - } - } - - rm_db->num_entries = parms->num_elements; - rm_db->dir = parms->dir; - rm_db->type = parms->type; - *parms->rm_db = (void *)rm_db; - - printf("%s: type:%d num_entries:%d\n", - tf_dir_2_str(parms->dir), - parms->type, - i); - - tfp_free((void *)req); - tfp_free((void *)resv); - - return 0; - - fail: - tfp_free((void *)req); - tfp_free((void *)resv); - tfp_free((void *)db->pool); - tfp_free((void *)db); - tfp_free((void *)rm_db); - parms->rm_db = NULL; - - return -EINVAL; -} - -int -tf_rm_free_db(struct tf *tfp, - struct tf_rm_free_db_parms *parms) -{ - int rc; - int i; - uint16_t resv_size = 0; - struct tf_rm_new_db *rm_db; - struct tf_rm_resc_entry *resv; - bool residuals_found = false; - - TF_CHECK_PARMS2(parms, parms->rm_db); - - /* Device unbind happens when the TF Session is closed and the - * session ref count is 0. Device unbind will cleanup each of - * its support modules, i.e. Identifier, thus we're ending up - * here to close the DB. - * - * On TF Session close it is assumed that the session has already - * cleaned up all its resources, individually, while - * destroying its flows. - * - * To assist in the 'cleanup checking' the DB is checked for any - * remaining elements and logged if found to be the case. - * - * Any such elements will need to be 'cleared' ahead of - * returning the resources to the HCAPI RM. - * - * RM will signal FW to flush the DB resources. FW will - * perform the invalidation. TF Session close will return the - * previous allocated elements to the RM and then close the - * HCAPI RM registration. That then saves several 'free' msgs - * from being required. - */ - - rm_db = (struct tf_rm_new_db *)parms->rm_db; - - /* Check for residuals that the client didn't clean up */ - rc = tf_rm_check_residuals(rm_db, - &resv_size, - &resv, - &residuals_found); - if (rc) - return rc; - - /* Invalidate any residuals followed by a DB traversal for - * pool cleanup. - */ - if (residuals_found) { - rc = tf_msg_session_resc_flush(tfp, - parms->dir, - resv_size, - resv); - tfp_free((void *)resv); - /* On failure we still have to cleanup so we can only - * log that FW failed. - */ - if (rc) - TFP_DRV_LOG(ERR, - "%s: Internal Flush error, module:%s\n", - tf_dir_2_str(parms->dir), - tf_device_module_type_2_str(rm_db->type)); - } - - for (i = 0; i < rm_db->num_entries; i++) - tfp_free((void *)rm_db->db[i].pool); - - tfp_free((void *)parms->rm_db); - - return rc; -} - -int -tf_rm_allocate(struct tf_rm_allocate_parms *parms) -{ - int rc; - int id; - uint32_t index; - struct tf_rm_new_db *rm_db; - enum tf_rm_elem_cfg_type cfg_type; - - TF_CHECK_PARMS2(parms, parms->rm_db); - - rm_db = (struct tf_rm_new_db *)parms->rm_db; - cfg_type = rm_db->db[parms->db_index].cfg_type; - - /* Bail out if not controlled by RM */ - if (cfg_type != TF_RM_ELEM_CFG_HCAPI && - cfg_type != TF_RM_ELEM_CFG_PRIVATE) - return -ENOTSUP; - - /* Bail out if the pool is not valid, should never happen */ - if (rm_db->db[parms->db_index].pool == NULL) { - rc = -ENOTSUP; - TFP_DRV_LOG(ERR, - "%s: Invalid pool for this type:%d, rc:%s\n", - tf_dir_2_str(rm_db->dir), - parms->db_index, - strerror(-rc)); - return rc; - } - - /* - * priority 0: allocate from top of the tcam i.e. high - * priority !0: allocate index from bottom i.e lowest - */ - if (parms->priority) - id = ba_alloc_reverse(rm_db->db[parms->db_index].pool); - else - id = ba_alloc(rm_db->db[parms->db_index].pool); - if (id == BA_FAIL) { - rc = -ENOMEM; - TFP_DRV_LOG(ERR, - "%s: Allocation failed, rc:%s\n", - tf_dir_2_str(rm_db->dir), - strerror(-rc)); - return rc; - } - - /* Adjust for any non zero start value */ - rc = tf_rm_adjust_index(rm_db->db, - TF_RM_ADJUST_ADD_BASE, - parms->db_index, - id, - &index); - if (rc) { - TFP_DRV_LOG(ERR, - "%s: Alloc adjust of base index failed, rc:%s\n", - tf_dir_2_str(rm_db->dir), - strerror(-rc)); - return -EINVAL; - } - - *parms->index = index; - - return rc; -} - -int -tf_rm_free(struct tf_rm_free_parms *parms) -{ - int rc; - uint32_t adj_index; - struct tf_rm_new_db *rm_db; - enum tf_rm_elem_cfg_type cfg_type; - - TF_CHECK_PARMS2(parms, parms->rm_db); - - rm_db = (struct tf_rm_new_db *)parms->rm_db; - cfg_type = rm_db->db[parms->db_index].cfg_type; - - /* Bail out if not controlled by RM */ - if (cfg_type != TF_RM_ELEM_CFG_HCAPI && - cfg_type != TF_RM_ELEM_CFG_PRIVATE) - return -ENOTSUP; - - /* Bail out if the pool is not valid, should never happen */ - if (rm_db->db[parms->db_index].pool == NULL) { - rc = -ENOTSUP; - TFP_DRV_LOG(ERR, - "%s: Invalid pool for this type:%d, rc:%s\n", - tf_dir_2_str(rm_db->dir), - parms->db_index, - strerror(-rc)); - return rc; - } - - /* Adjust for any non zero start value */ - rc = tf_rm_adjust_index(rm_db->db, - TF_RM_ADJUST_RM_BASE, - parms->db_index, - parms->index, - &adj_index); - if (rc) - return rc; - - rc = ba_free(rm_db->db[parms->db_index].pool, adj_index); - /* No logging direction matters and that is not available here */ - if (rc) - return rc; - - return rc; -} - -int -tf_rm_is_allocated(struct tf_rm_is_allocated_parms *parms) -{ - int rc; - uint32_t adj_index; - struct tf_rm_new_db *rm_db; - enum tf_rm_elem_cfg_type cfg_type; - - TF_CHECK_PARMS2(parms, parms->rm_db); - - rm_db = (struct tf_rm_new_db *)parms->rm_db; - cfg_type = rm_db->db[parms->db_index].cfg_type; - - /* Bail out if not controlled by RM */ - if (cfg_type != TF_RM_ELEM_CFG_HCAPI && - cfg_type != TF_RM_ELEM_CFG_PRIVATE) - return -ENOTSUP; - - /* Bail out if the pool is not valid, should never happen */ - if (rm_db->db[parms->db_index].pool == NULL) { - rc = -ENOTSUP; - TFP_DRV_LOG(ERR, - "%s: Invalid pool for this type:%d, rc:%s\n", - tf_dir_2_str(rm_db->dir), - parms->db_index, - strerror(-rc)); - return rc; - } - - /* Adjust for any non zero start value */ - rc = tf_rm_adjust_index(rm_db->db, - TF_RM_ADJUST_RM_BASE, - parms->db_index, - parms->index, - &adj_index); - if (rc) - return rc; - - *parms->allocated = ba_inuse(rm_db->db[parms->db_index].pool, - adj_index); - - return rc; -} - -int -tf_rm_get_info(struct tf_rm_get_alloc_info_parms *parms) -{ - struct tf_rm_new_db *rm_db; - enum tf_rm_elem_cfg_type cfg_type; - - TF_CHECK_PARMS2(parms, parms->rm_db); - - rm_db = (struct tf_rm_new_db *)parms->rm_db; - cfg_type = rm_db->db[parms->db_index].cfg_type; - - /* Bail out if not controlled by RM */ - if (cfg_type != TF_RM_ELEM_CFG_HCAPI && - cfg_type != TF_RM_ELEM_CFG_PRIVATE) - return -ENOTSUP; - - memcpy(parms->info, - &rm_db->db[parms->db_index].alloc, - sizeof(struct tf_rm_alloc_info)); - - return 0; -} - -int -tf_rm_get_hcapi_type(struct tf_rm_get_hcapi_parms *parms) -{ - struct tf_rm_new_db *rm_db; - enum tf_rm_elem_cfg_type cfg_type; - - TF_CHECK_PARMS2(parms, parms->rm_db); - - rm_db = (struct tf_rm_new_db *)parms->rm_db; - cfg_type = rm_db->db[parms->db_index].cfg_type; - - /* Bail out if not controlled by RM */ - if (cfg_type != TF_RM_ELEM_CFG_HCAPI && - cfg_type != TF_RM_ELEM_CFG_PRIVATE) - return -ENOTSUP; - - *parms->hcapi_type = rm_db->db[parms->db_index].hcapi_type; - - return 0; -} - -int -tf_rm_get_inuse_count(struct tf_rm_get_inuse_count_parms *parms) -{ - int rc = 0; - struct tf_rm_new_db *rm_db; - enum tf_rm_elem_cfg_type cfg_type; - - TF_CHECK_PARMS2(parms, parms->rm_db); - - rm_db = (struct tf_rm_new_db *)parms->rm_db; - cfg_type = rm_db->db[parms->db_index].cfg_type; - - /* Bail out if not controlled by RM */ - if (cfg_type != TF_RM_ELEM_CFG_HCAPI && - cfg_type != TF_RM_ELEM_CFG_PRIVATE) - return -ENOTSUP; - - /* Bail silently (no logging), if the pool is not valid there - * was no elements allocated for it. - */ - if (rm_db->db[parms->db_index].pool == NULL) { - *parms->count = 0; - return 0; - } - - *parms->count = ba_inuse_count(rm_db->db[parms->db_index].pool); - - return rc; - -} diff --git a/drivers/net/bnxt/tf_core/tf_rm_new.h b/drivers/net/bnxt/tf_core/tf_rm_new.h deleted file mode 100644 index 3b818f2af5..0000000000 --- a/drivers/net/bnxt/tf_core/tf_rm_new.h +++ /dev/null @@ -1,446 +0,0 @@ -/* SPDX-License-Identifier: BSD-3-Clause - * Copyright(c) 2019-2020 Broadcom - * All rights reserved. - */ - -#ifndef TF_RM_NEW_H_ -#define TF_RM_NEW_H_ - -#include "tf_core.h" -#include "bitalloc.h" -#include "tf_device.h" - -struct tf; - -/** - * The Resource Manager (RM) module provides basic DB handling for - * internal resources. These resources exists within the actual device - * and are controlled by the HCAPI Resource Manager running on the - * firmware. - * - * The RM DBs are all intended to be indexed using TF types there for - * a lookup requires no additional conversion. The DB configuration - * specifies the TF Type to HCAPI Type mapping and it becomes the - * responsibility of the DB initialization to handle this static - * mapping. - * - * Accessor functions are providing access to the DB, thus hiding the - * implementation. - * - * The RM DB will work on its initial allocated sizes so the - * capability of dynamically growing a particular resource is not - * possible. If this capability later becomes a requirement then the - * MAX pool size of the Chip œneeds to be added to the tf_rm_elem_info - * structure and several new APIs would need to be added to allow for - * growth of a single TF resource type. - * - * The access functions does not check for NULL pointers as it's a - * support module, not called directly. - */ - -/** - * Resource reservation single entry result. Used when accessing HCAPI - * RM on the firmware. - */ -struct tf_rm_new_entry { - /** Starting index of the allocated resource */ - uint16_t start; - /** Number of allocated elements */ - uint16_t stride; -}; - -/** - * RM Element configuration enumeration. Used by the Device to - * indicate how the RM elements the DB consists off, are to be - * configured at time of DB creation. The TF may present types to the - * ULP layer that is not controlled by HCAPI within the Firmware. - */ -enum tf_rm_elem_cfg_type { - /** No configuration */ - TF_RM_ELEM_CFG_NULL, - /** HCAPI 'controlled', uses a Pool for internal storage */ - TF_RM_ELEM_CFG_HCAPI, - /** Private thus not HCAPI 'controlled', creates a Pool for storage */ - TF_RM_ELEM_CFG_PRIVATE, - /** - * Shared element thus it belongs to a shared FW Session and - * is not controlled by the Host. - */ - TF_RM_ELEM_CFG_SHARED, - TF_RM_TYPE_MAX -}; - -/** - * RM Reservation strategy enumeration. Type of strategy comes from - * the HCAPI RM QCAPS handshake. - */ -enum tf_rm_resc_resv_strategy { - TF_RM_RESC_RESV_STATIC_PARTITION, - TF_RM_RESC_RESV_STRATEGY_1, - TF_RM_RESC_RESV_STRATEGY_2, - TF_RM_RESC_RESV_STRATEGY_3, - TF_RM_RESC_RESV_MAX -}; - -/** - * RM Element configuration structure, used by the Device to configure - * how an individual TF type is configured in regard to the HCAPI RM - * of same type. - */ -struct tf_rm_element_cfg { - /** - * RM Element config controls how the DB for that element is - * processed. - */ - enum tf_rm_elem_cfg_type cfg_type; - - /* If a HCAPI to TF type conversion is required then TF type - * can be added here. - */ - - /** - * HCAPI RM Type for the element. Used for TF to HCAPI type - * conversion. - */ - uint16_t hcapi_type; -}; - -/** - * Allocation information for a single element. - */ -struct tf_rm_alloc_info { - /** - * HCAPI RM allocated range information. - * - * NOTE: - * In case of dynamic allocation support this would have - * to be changed to linked list of tf_rm_entry instead. - */ - struct tf_rm_new_entry entry; -}; - -/** - * Create RM DB parameters - */ -struct tf_rm_create_db_parms { - /** - * [in] Device module type. Used for logging purposes. - */ - enum tf_device_module_type type; - /** - * [in] Receive or transmit direction. - */ - enum tf_dir dir; - /** - * [in] Number of elements. - */ - uint16_t num_elements; - /** - * [in] Parameter structure array. Array size is num_elements. - */ - struct tf_rm_element_cfg *cfg; - /** - * Resource allocation count array. This array content - * originates from the tf_session_resources that is passed in - * on session open. - * Array size is num_elements. - */ - uint16_t *alloc_cnt; - /** - * [out] RM DB Handle - */ - void **rm_db; -}; - -/** - * Free RM DB parameters - */ -struct tf_rm_free_db_parms { - /** - * [in] Receive or transmit direction - */ - enum tf_dir dir; - /** - * [in] RM DB Handle - */ - void *rm_db; -}; - -/** - * Allocate RM parameters for a single element - */ -struct tf_rm_allocate_parms { - /** - * [in] RM DB Handle - */ - void *rm_db; - /** - * [in] DB Index, indicates which DB entry to perform the - * action on. - */ - uint16_t db_index; - /** - * [in] Pointer to the allocated index in normalized - * form. Normalized means the index has been adjusted, - * i.e. Full Action Record offsets. - */ - uint32_t *index; - /** - * [in] Priority, indicates the priority of the entry - * priority 0: allocate from top of the tcam (from index 0 - * or lowest available index) - * priority !0: allocate from bottom of the tcam (from highest - * available index) - */ - uint32_t priority; -}; - -/** - * Free RM parameters for a single element - */ -struct tf_rm_free_parms { - /** - * [in] RM DB Handle - */ - void *rm_db; - /** - * [in] DB Index, indicates which DB entry to perform the - * action on. - */ - uint16_t db_index; - /** - * [in] Index to free - */ - uint16_t index; -}; - -/** - * Is Allocated parameters for a single element - */ -struct tf_rm_is_allocated_parms { - /** - * [in] RM DB Handle - */ - void *rm_db; - /** - * [in] DB Index, indicates which DB entry to perform the - * action on. - */ - uint16_t db_index; - /** - * [in] Index to free - */ - uint32_t index; - /** - * [in] Pointer to flag that indicates the state of the query - */ - int *allocated; -}; - -/** - * Get Allocation information for a single element - */ -struct tf_rm_get_alloc_info_parms { - /** - * [in] RM DB Handle - */ - void *rm_db; - /** - * [in] DB Index, indicates which DB entry to perform the - * action on. - */ - uint16_t db_index; - /** - * [out] Pointer to the requested allocation information for - * the specified db_index - */ - struct tf_rm_alloc_info *info; -}; - -/** - * Get HCAPI type parameters for a single element - */ -struct tf_rm_get_hcapi_parms { - /** - * [in] RM DB Handle - */ - void *rm_db; - /** - * [in] DB Index, indicates which DB entry to perform the - * action on. - */ - uint16_t db_index; - /** - * [out] Pointer to the hcapi type for the specified db_index - */ - uint16_t *hcapi_type; -}; - -/** - * Get InUse count parameters for single element - */ -struct tf_rm_get_inuse_count_parms { - /** - * [in] RM DB Handle - */ - void *rm_db; - /** - * [in] DB Index, indicates which DB entry to perform the - * action on. - */ - uint16_t db_index; - /** - * [out] Pointer to the inuse count for the specified db_index - */ - uint16_t *count; -}; - -/** - * @page rm Resource Manager - * - * @ref tf_rm_create_db - * - * @ref tf_rm_free_db - * - * @ref tf_rm_allocate - * - * @ref tf_rm_free - * - * @ref tf_rm_is_allocated - * - * @ref tf_rm_get_info - * - * @ref tf_rm_get_hcapi_type - * - * @ref tf_rm_get_inuse_count - */ - -/** - * Creates and fills a Resource Manager (RM) DB with requested - * elements. The DB is indexed per the parms structure. - * - * [in] tfp - * Pointer to TF handle, used for HCAPI communication - * - * [in] parms - * Pointer to create parameters - * - * Returns - * - (0) if successful. - * - (-EINVAL) on failure. - */ -/* - * NOTE: - * - Fail on parameter check - * - Fail on DB creation, i.e. alloc amount is not possible or validation fails - * - Fail on DB creation if DB already exist - * - * - Allocs local DB - * - Does hcapi qcaps - * - Does hcapi reservation - * - Populates the pool with allocated elements - * - Returns handle to the created DB - */ -int tf_rm_create_db(struct tf *tfp, - struct tf_rm_create_db_parms *parms); - -/** - * Closes the Resource Manager (RM) DB and frees all allocated - * resources per the associated database. - * - * [in] tfp - * Pointer to TF handle, used for HCAPI communication - * - * [in] parms - * Pointer to free parameters - * - * Returns - * - (0) if successful. - * - (-EINVAL) on failure. - */ -int tf_rm_free_db(struct tf *tfp, - struct tf_rm_free_db_parms *parms); - -/** - * Allocates a single element for the type specified, within the DB. - * - * [in] parms - * Pointer to allocate parameters - * - * Returns - * - (0) if successful. - * - (-EINVAL) on failure. - * - (-ENOMEM) if pool is empty - */ -int tf_rm_allocate(struct tf_rm_allocate_parms *parms); - -/** - * Free's a single element for the type specified, within the DB. - * - * [in] parms - * Pointer to free parameters - * - * Returns - * - (0) if successful. - * - (-EINVAL) on failure. - */ -int tf_rm_free(struct tf_rm_free_parms *parms); - -/** - * Performs an allocation verification check on a specified element. - * - * [in] parms - * Pointer to is allocated parameters - * - * Returns - * - (0) if successful. - * - (-EINVAL) on failure. - */ -/* - * NOTE: - * - If pool is set to Chip MAX, then the query index must be checked - * against the allocated range and query index must be allocated as well. - * - If pool is allocated size only, then check if query index is allocated. - */ -int tf_rm_is_allocated(struct tf_rm_is_allocated_parms *parms); - -/** - * Retrieves an elements allocation information from the Resource - * Manager (RM) DB. - * - * [in] parms - * Pointer to get info parameters - * - * Returns - * - (0) if successful. - * - (-EINVAL) on failure. - */ -int tf_rm_get_info(struct tf_rm_get_alloc_info_parms *parms); - -/** - * Performs a lookup in the Resource Manager DB and retrieves the - * requested HCAPI RM type. - * - * [in] parms - * Pointer to get hcapi parameters - * - * Returns - * - (0) if successful. - * - (-EINVAL) on failure. - */ -int tf_rm_get_hcapi_type(struct tf_rm_get_hcapi_parms *parms); - -/** - * Performs a lookup in the Resource Manager DB and retrieves the - * requested HCAPI RM type inuse count. - * - * [in] parms - * Pointer to get inuse parameters - * - * Returns - * - (0) if successful. - * - (-EINVAL) on failure. - */ -int tf_rm_get_inuse_count(struct tf_rm_get_inuse_count_parms *parms); - -#endif /* TF_RM_NEW_H_ */ diff --git a/drivers/net/bnxt/tf_core/tf_session.h b/drivers/net/bnxt/tf_core/tf_session.h index 705bb0955c..e4472ed7f5 100644 --- a/drivers/net/bnxt/tf_core/tf_session.h +++ b/drivers/net/bnxt/tf_core/tf_session.h @@ -14,6 +14,7 @@ #include "tf_device.h" #include "tf_rm.h" #include "tf_tbl.h" +#include "tf_resources.h" #include "stack.h" /** @@ -43,7 +44,8 @@ #define TF_SESSION_EM_POOL_SIZE \ (TF_SESSION_TOTAL_FN_BLOCKS / TF_SESSION_EM_ENTRY_SIZE) -/** Session +/** + * Session * * Shared memory containing private TruFlow session information. * Through this structure the session can keep track of resource @@ -99,216 +101,6 @@ struct tf_session { /** Device handle */ struct tf_dev_info dev; - /** Session HW and SRAM resources */ - struct tf_rm_db resc; - - /* Session HW resource pools */ - - /** RX L2 CTXT TCAM Pool */ - BITALLOC_INST(TF_L2_CTXT_TCAM_POOL_NAME_RX, TF_NUM_L2_CTXT_TCAM); - /** TX L2 CTXT TCAM Pool */ - BITALLOC_INST(TF_L2_CTXT_TCAM_POOL_NAME_TX, TF_NUM_L2_CTXT_TCAM); - - /** RX Profile Func Pool */ - BITALLOC_INST(TF_PROF_FUNC_POOL_NAME_RX, TF_NUM_PROF_FUNC); - /** TX Profile Func Pool */ - BITALLOC_INST(TF_PROF_FUNC_POOL_NAME_TX, TF_NUM_PROF_FUNC); - - /** RX Profile TCAM Pool */ - BITALLOC_INST(TF_PROF_TCAM_POOL_NAME_RX, TF_NUM_PROF_TCAM); - /** TX Profile TCAM Pool */ - BITALLOC_INST(TF_PROF_TCAM_POOL_NAME_TX, TF_NUM_PROF_TCAM); - - /** RX EM Profile ID Pool */ - BITALLOC_INST(TF_EM_PROF_ID_POOL_NAME_RX, TF_NUM_EM_PROF_ID); - /** TX EM Key Pool */ - BITALLOC_INST(TF_EM_PROF_ID_POOL_NAME_TX, TF_NUM_EM_PROF_ID); - - /** RX WC Profile Pool */ - BITALLOC_INST(TF_WC_TCAM_PROF_ID_POOL_NAME_RX, TF_NUM_WC_PROF_ID); - /** TX WC Profile Pool */ - BITALLOC_INST(TF_WC_TCAM_PROF_ID_POOL_NAME_TX, TF_NUM_WC_PROF_ID); - - /* TBD, how do we want to handle EM records ?*/ - /* EM Records are not controlled by way of a pool */ - - /** RX WC TCAM Pool */ - BITALLOC_INST(TF_WC_TCAM_POOL_NAME_RX, TF_NUM_WC_TCAM_ROW); - /** TX WC TCAM Pool */ - BITALLOC_INST(TF_WC_TCAM_POOL_NAME_TX, TF_NUM_WC_TCAM_ROW); - - /** RX Meter Profile Pool */ - BITALLOC_INST(TF_METER_PROF_POOL_NAME_RX, TF_NUM_METER_PROF); - /** TX Meter Profile Pool */ - BITALLOC_INST(TF_METER_PROF_POOL_NAME_TX, TF_NUM_METER_PROF); - - /** RX Meter Instance Pool */ - BITALLOC_INST(TF_METER_INST_POOL_NAME_RX, TF_NUM_METER); - /** TX Meter Pool */ - BITALLOC_INST(TF_METER_INST_POOL_NAME_TX, TF_NUM_METER); - - /** RX Mirror Configuration Pool*/ - BITALLOC_INST(TF_MIRROR_POOL_NAME_RX, TF_NUM_MIRROR); - /** RX Mirror Configuration Pool */ - BITALLOC_INST(TF_MIRROR_POOL_NAME_TX, TF_NUM_MIRROR); - - /** RX UPAR Pool */ - BITALLOC_INST(TF_UPAR_POOL_NAME_RX, TF_NUM_UPAR); - /** TX UPAR Pool */ - BITALLOC_INST(TF_UPAR_POOL_NAME_TX, TF_NUM_UPAR); - - /** RX SP TCAM Pool */ - BITALLOC_INST(TF_SP_TCAM_POOL_NAME_RX, TF_NUM_SP_TCAM); - /** TX SP TCAM Pool */ - BITALLOC_INST(TF_SP_TCAM_POOL_NAME_TX, TF_NUM_SP_TCAM); - - /** RX FKB Pool */ - BITALLOC_INST(TF_FKB_POOL_NAME_RX, TF_NUM_FKB); - /** TX FKB Pool */ - BITALLOC_INST(TF_FKB_POOL_NAME_TX, TF_NUM_FKB); - - /** RX Table Scope Pool */ - BITALLOC_INST(TF_TBL_SCOPE_POOL_NAME_RX, TF_NUM_TBL_SCOPE); - /** TX Table Scope Pool */ - BITALLOC_INST(TF_TBL_SCOPE_POOL_NAME_TX, TF_NUM_TBL_SCOPE); - - /** RX L2 Func Pool */ - BITALLOC_INST(TF_L2_FUNC_POOL_NAME_RX, TF_NUM_L2_FUNC); - /** TX L2 Func Pool */ - BITALLOC_INST(TF_L2_FUNC_POOL_NAME_TX, TF_NUM_L2_FUNC); - - /** RX Epoch0 Pool */ - BITALLOC_INST(TF_EPOCH0_POOL_NAME_RX, TF_NUM_EPOCH0); - /** TX Epoch0 Pool */ - BITALLOC_INST(TF_EPOCH0_POOL_NAME_TX, TF_NUM_EPOCH0); - - /** TX Epoch1 Pool */ - BITALLOC_INST(TF_EPOCH1_POOL_NAME_RX, TF_NUM_EPOCH1); - /** TX Epoch1 Pool */ - BITALLOC_INST(TF_EPOCH1_POOL_NAME_TX, TF_NUM_EPOCH1); - - /** RX MetaData Profile Pool */ - BITALLOC_INST(TF_METADATA_POOL_NAME_RX, TF_NUM_METADATA); - /** TX MetaData Profile Pool */ - BITALLOC_INST(TF_METADATA_POOL_NAME_TX, TF_NUM_METADATA); - - /** RX Connection Tracking State Pool */ - BITALLOC_INST(TF_CT_STATE_POOL_NAME_RX, TF_NUM_CT_STATE); - /** TX Connection Tracking State Pool */ - BITALLOC_INST(TF_CT_STATE_POOL_NAME_TX, TF_NUM_CT_STATE); - - /** RX Range Profile Pool */ - BITALLOC_INST(TF_RANGE_PROF_POOL_NAME_RX, TF_NUM_RANGE_PROF); - /** TX Range Profile Pool */ - BITALLOC_INST(TF_RANGE_PROF_POOL_NAME_TX, TF_NUM_RANGE_PROF); - - /** RX Range Pool */ - BITALLOC_INST(TF_RANGE_ENTRY_POOL_NAME_RX, TF_NUM_RANGE_ENTRY); - /** TX Range Pool */ - BITALLOC_INST(TF_RANGE_ENTRY_POOL_NAME_TX, TF_NUM_RANGE_ENTRY); - - /** RX LAG Pool */ - BITALLOC_INST(TF_LAG_ENTRY_POOL_NAME_RX, TF_NUM_LAG_ENTRY); - /** TX LAG Pool */ - BITALLOC_INST(TF_LAG_ENTRY_POOL_NAME_TX, TF_NUM_LAG_ENTRY); - - /* Session SRAM pools */ - - /** RX Full Action Record Pool */ - BITALLOC_INST(TF_SRAM_FULL_ACTION_POOL_NAME_RX, - TF_RSVD_SRAM_FULL_ACTION_RX); - /** TX Full Action Record Pool */ - BITALLOC_INST(TF_SRAM_FULL_ACTION_POOL_NAME_TX, - TF_RSVD_SRAM_FULL_ACTION_TX); - - /** RX Multicast Group Pool, only RX is supported */ - BITALLOC_INST(TF_SRAM_MCG_POOL_NAME_RX, - TF_RSVD_SRAM_MCG_RX); - - /** RX Encap 8B Pool*/ - BITALLOC_INST(TF_SRAM_ENCAP_8B_POOL_NAME_RX, - TF_RSVD_SRAM_ENCAP_8B_RX); - /** TX Encap 8B Pool*/ - BITALLOC_INST(TF_SRAM_ENCAP_8B_POOL_NAME_TX, - TF_RSVD_SRAM_ENCAP_8B_TX); - - /** RX Encap 16B Pool */ - BITALLOC_INST(TF_SRAM_ENCAP_16B_POOL_NAME_RX, - TF_RSVD_SRAM_ENCAP_16B_RX); - /** TX Encap 16B Pool */ - BITALLOC_INST(TF_SRAM_ENCAP_16B_POOL_NAME_TX, - TF_RSVD_SRAM_ENCAP_16B_TX); - - /** TX Encap 64B Pool, only TX is supported */ - BITALLOC_INST(TF_SRAM_ENCAP_64B_POOL_NAME_TX, - TF_RSVD_SRAM_ENCAP_64B_TX); - - /** RX Source Properties SMAC Pool */ - BITALLOC_INST(TF_SRAM_SP_SMAC_POOL_NAME_RX, - TF_RSVD_SRAM_SP_SMAC_RX); - /** TX Source Properties SMAC Pool */ - BITALLOC_INST(TF_SRAM_SP_SMAC_POOL_NAME_TX, - TF_RSVD_SRAM_SP_SMAC_TX); - - /** TX Source Properties SMAC IPv4 Pool, only TX is supported */ - BITALLOC_INST(TF_SRAM_SP_SMAC_IPV4_POOL_NAME_TX, - TF_RSVD_SRAM_SP_SMAC_IPV4_TX); - - /** TX Source Properties SMAC IPv6 Pool, only TX is supported */ - BITALLOC_INST(TF_SRAM_SP_SMAC_IPV6_POOL_NAME_TX, - TF_RSVD_SRAM_SP_SMAC_IPV6_TX); - - /** RX Counter 64B Pool */ - BITALLOC_INST(TF_SRAM_STATS_64B_POOL_NAME_RX, - TF_RSVD_SRAM_COUNTER_64B_RX); - /** TX Counter 64B Pool */ - BITALLOC_INST(TF_SRAM_STATS_64B_POOL_NAME_TX, - TF_RSVD_SRAM_COUNTER_64B_TX); - - /** RX NAT Source Port Pool */ - BITALLOC_INST(TF_SRAM_NAT_SPORT_POOL_NAME_RX, - TF_RSVD_SRAM_NAT_SPORT_RX); - /** TX NAT Source Port Pool */ - BITALLOC_INST(TF_SRAM_NAT_SPORT_POOL_NAME_TX, - TF_RSVD_SRAM_NAT_SPORT_TX); - - /** RX NAT Destination Port Pool */ - BITALLOC_INST(TF_SRAM_NAT_DPORT_POOL_NAME_RX, - TF_RSVD_SRAM_NAT_DPORT_RX); - /** TX NAT Destination Port Pool */ - BITALLOC_INST(TF_SRAM_NAT_DPORT_POOL_NAME_TX, - TF_RSVD_SRAM_NAT_DPORT_TX); - - /** RX NAT Source IPv4 Pool */ - BITALLOC_INST(TF_SRAM_NAT_S_IPV4_POOL_NAME_RX, - TF_RSVD_SRAM_NAT_S_IPV4_RX); - /** TX NAT Source IPv4 Pool */ - BITALLOC_INST(TF_SRAM_NAT_S_IPV4_POOL_NAME_TX, - TF_RSVD_SRAM_NAT_S_IPV4_TX); - - /** RX NAT Destination IPv4 Pool */ - BITALLOC_INST(TF_SRAM_NAT_D_IPV4_POOL_NAME_RX, - TF_RSVD_SRAM_NAT_D_IPV4_RX); - /** TX NAT IPv4 Destination Pool */ - BITALLOC_INST(TF_SRAM_NAT_D_IPV4_POOL_NAME_TX, - TF_RSVD_SRAM_NAT_D_IPV4_TX); - - /** - * Pools not allocated from HCAPI RM - */ - - /** RX L2 Ctx Remap ID Pool */ - BITALLOC_INST(TF_L2_CTXT_REMAP_POOL_NAME_RX, TF_NUM_L2_CTXT_TCAM); - /** TX L2 Ctx Remap ID Pool */ - BITALLOC_INST(TF_L2_CTXT_REMAP_POOL_NAME_TX, TF_NUM_L2_CTXT_TCAM); - - /** CRC32 seed table */ -#define TF_LKUP_SEED_MEM_SIZE 512 - uint32_t lkup_em_seed_mem[TF_DIR_MAX][TF_LKUP_SEED_MEM_SIZE]; - - /** Lookup3 init values */ - uint32_t lkup_lkup3_init_cfg[TF_DIR_MAX]; - /** Table scope array */ struct tf_tbl_scope_cb tbl_scopes[TF_NUM_TBL_SCOPE]; diff --git a/drivers/net/bnxt/tf_core/tf_tbl.c b/drivers/net/bnxt/tf_core/tf_tbl.c index d7f5de4c40..05e866dc6c 100644 --- a/drivers/net/bnxt/tf_core/tf_tbl.c +++ b/drivers/net/bnxt/tf_core/tf_tbl.c @@ -5,175 +5,413 @@ /* Truflow Table APIs and supporting code */ -#include -#include -#include -#include -#include #include -#include -#include "hsi_struct_def_dpdk.h" -#include "tf_core.h" +#include "tf_tbl.h" +#include "tf_common.h" +#include "tf_rm.h" #include "tf_util.h" -#include "tf_em.h" #include "tf_msg.h" #include "tfp.h" -#include "hwrm_tf.h" -#include "bnxt.h" -#include "tf_resources.h" -#include "tf_rm.h" -#include "stack.h" -#include "tf_common.h" + + +struct tf; + +/** + * Table DBs. + */ +static void *tbl_db[TF_DIR_MAX]; + +/** + * Table Shadow DBs + */ +/* static void *shadow_tbl_db[TF_DIR_MAX]; */ + +/** + * Init flag, set on bind and cleared on unbind + */ +static uint8_t init; /** - * Internal function to get a Table Entry. Supports all Table Types - * except the TF_TBL_TYPE_EXT as that is handled as a table scope. - * - * [in] tfp - * Pointer to TruFlow handle - * - * [in] parms - * Pointer to input parameters - * - * Returns: - * 0 - Success - * -EINVAL - Parameter error + * Shadow init flag, set on bind and cleared on unbind */ -static int -tf_bulk_get_tbl_entry_internal(struct tf *tfp, - struct tf_bulk_get_tbl_entry_parms *parms) +/* static uint8_t shadow_init; */ + +int +tf_tbl_bind(struct tf *tfp, + struct tf_tbl_cfg_parms *parms) +{ + int rc; + int i; + struct tf_rm_create_db_parms db_cfg = { 0 }; + + TF_CHECK_PARMS2(tfp, parms); + + if (init) { + TFP_DRV_LOG(ERR, + "Table DB already initialized\n"); + return -EINVAL; + } + + db_cfg.num_elements = parms->num_elements; + db_cfg.type = TF_DEVICE_MODULE_TYPE_TABLE; + db_cfg.num_elements = parms->num_elements; + db_cfg.cfg = parms->cfg; + + for (i = 0; i < TF_DIR_MAX; i++) { + db_cfg.dir = i; + db_cfg.alloc_cnt = parms->resources->tbl_cnt[i].cnt; + db_cfg.rm_db = &tbl_db[i]; + rc = tf_rm_create_db(tfp, &db_cfg); + if (rc) { + TFP_DRV_LOG(ERR, + "%s: Table DB creation failed\n", + tf_dir_2_str(i)); + + return rc; + } + } + + init = 1; + + printf("Table Type - initialized\n"); + + return 0; +} + +int +tf_tbl_unbind(struct tf *tfp) { int rc; - int id; - uint32_t index; - struct bitalloc *session_pool; - struct tf_session *tfs = (struct tf_session *)(tfp->session->core_data); - - /* Lookup the pool using the table type of the element */ - rc = tf_rm_lookup_tbl_type_pool(tfs, - parms->dir, - parms->type, - &session_pool); - /* Error logging handled by tf_rm_lookup_tbl_type_pool */ + int i; + struct tf_rm_free_db_parms fparms = { 0 }; + + TF_CHECK_PARMS1(tfp); + + /* Bail if nothing has been initialized */ + if (!init) { + TFP_DRV_LOG(INFO, + "No Table DBs created\n"); + return 0; + } + + for (i = 0; i < TF_DIR_MAX; i++) { + fparms.dir = i; + fparms.rm_db = tbl_db[i]; + rc = tf_rm_free_db(tfp, &fparms); + if (rc) + return rc; + + tbl_db[i] = NULL; + } + + init = 0; + + return 0; +} + +int +tf_tbl_alloc(struct tf *tfp __rte_unused, + struct tf_tbl_alloc_parms *parms) +{ + int rc; + uint32_t idx; + struct tf_rm_allocate_parms aparms = { 0 }; + + TF_CHECK_PARMS2(tfp, parms); + + if (!init) { + TFP_DRV_LOG(ERR, + "%s: No Table DBs created\n", + tf_dir_2_str(parms->dir)); + return -EINVAL; + } + + /* Allocate requested element */ + aparms.rm_db = tbl_db[parms->dir]; + aparms.db_index = parms->type; + aparms.index = &idx; + rc = tf_rm_allocate(&aparms); + if (rc) { + TFP_DRV_LOG(ERR, + "%s: Failed allocate, type:%d\n", + tf_dir_2_str(parms->dir), + parms->type); + return rc; + } + + *parms->idx = idx; + + return 0; +} + +int +tf_tbl_free(struct tf *tfp __rte_unused, + struct tf_tbl_free_parms *parms) +{ + int rc; + struct tf_rm_is_allocated_parms aparms = { 0 }; + struct tf_rm_free_parms fparms = { 0 }; + int allocated = 0; + + TF_CHECK_PARMS2(tfp, parms); + + if (!init) { + TFP_DRV_LOG(ERR, + "%s: No Table DBs created\n", + tf_dir_2_str(parms->dir)); + return -EINVAL; + } + + /* Check if element is in use */ + aparms.rm_db = tbl_db[parms->dir]; + aparms.db_index = parms->type; + aparms.index = parms->idx; + aparms.allocated = &allocated; + rc = tf_rm_is_allocated(&aparms); if (rc) return rc; - index = parms->starting_idx; + if (!allocated) { + TFP_DRV_LOG(ERR, + "%s: Entry already free, type:%d, index:%d\n", + tf_dir_2_str(parms->dir), + parms->type, + parms->idx); + return rc; + } - /* - * Adjust the returned index/offset as there is no guarantee - * that the start is 0 at time of RM allocation - */ - tf_rm_convert_index(tfs, - parms->dir, + /* Free requested element */ + fparms.rm_db = tbl_db[parms->dir]; + fparms.db_index = parms->type; + fparms.index = parms->idx; + rc = tf_rm_free(&fparms); + if (rc) { + TFP_DRV_LOG(ERR, + "%s: Free failed, type:%d, index:%d\n", + tf_dir_2_str(parms->dir), parms->type, - TF_RM_CONVERT_RM_BASE, - parms->starting_idx, - &index); + parms->idx); + return rc; + } + + return 0; +} + +int +tf_tbl_alloc_search(struct tf *tfp __rte_unused, + struct tf_tbl_alloc_search_parms *parms __rte_unused) +{ + return 0; +} + +int +tf_tbl_set(struct tf *tfp, + struct tf_tbl_set_parms *parms) +{ + int rc; + int allocated = 0; + uint16_t hcapi_type; + struct tf_rm_is_allocated_parms aparms = { 0 }; + struct tf_rm_get_hcapi_parms hparms = { 0 }; + + TF_CHECK_PARMS3(tfp, parms, parms->data); + + if (!init) { + TFP_DRV_LOG(ERR, + "%s: No Table DBs created\n", + tf_dir_2_str(parms->dir)); + return -EINVAL; + } /* Verify that the entry has been previously allocated */ - id = ba_inuse(session_pool, index); - if (id != 1) { + aparms.rm_db = tbl_db[parms->dir]; + aparms.db_index = parms->type; + aparms.index = parms->idx; + aparms.allocated = &allocated; + rc = tf_rm_is_allocated(&aparms); + if (rc) + return rc; + + if (!allocated) { TFP_DRV_LOG(ERR, - "%s, Invalid or not allocated index, type:%d, starting_idx:%d\n", + "%s, Invalid or not allocated index, type:%d, idx:%d\n", tf_dir_2_str(parms->dir), parms->type, - index); + parms->idx); return -EINVAL; } - /* Get the entry */ - rc = tf_msg_bulk_get_tbl_entry(tfp, parms); + /* Set the entry */ + hparms.rm_db = tbl_db[parms->dir]; + hparms.db_index = parms->type; + hparms.hcapi_type = &hcapi_type; + rc = tf_rm_get_hcapi_type(&hparms); if (rc) { TFP_DRV_LOG(ERR, - "%s, Bulk get failed, type:%d, rc:%s\n", + "%s, Failed type lookup, type:%d, rc:%s\n", tf_dir_2_str(parms->dir), parms->type, strerror(-rc)); + return rc; } - return rc; + rc = tf_msg_set_tbl_entry(tfp, + parms->dir, + hcapi_type, + parms->data_sz_in_bytes, + parms->data, + parms->idx); + if (rc) { + TFP_DRV_LOG(ERR, + "%s, Set failed, type:%d, rc:%s\n", + tf_dir_2_str(parms->dir), + parms->type, + strerror(-rc)); + } + + return 0; } -#if (TF_SHADOW == 1) -/** - * Allocate Tbl entry from the Shadow DB. Shadow DB is searched for - * the requested entry. If found the ref count is incremente and - * returned. - * - * [in] tfs - * Pointer to session - * [in] parms - * Allocation parameters - * - * Return: - * 0 - Success, entry found and ref count incremented - * -ENOENT - Failure, entry not found - */ -static int -tf_alloc_tbl_entry_shadow(struct tf_session *tfs __rte_unused, - struct tf_alloc_tbl_entry_parms *parms __rte_unused) +int +tf_tbl_get(struct tf *tfp, + struct tf_tbl_get_parms *parms) { - TFP_DRV_LOG(ERR, - "%s, Entry Alloc with search not supported\n", - tf_dir_2_str(parms->dir)); + int rc; + uint16_t hcapi_type; + int allocated = 0; + struct tf_rm_is_allocated_parms aparms = { 0 }; + struct tf_rm_get_hcapi_parms hparms = { 0 }; - return -EOPNOTSUPP; -} + TF_CHECK_PARMS3(tfp, parms, parms->data); -/** - * Free Tbl entry from the Shadow DB. Shadow DB is searched for - * the requested entry. If found the ref count is decremente and - * new ref_count returned. - * - * [in] tfs - * Pointer to session - * [in] parms - * Allocation parameters - * - * Return: - * 0 - Success, entry found and ref count decremented - * -ENOENT - Failure, entry not found - */ -static int -tf_free_tbl_entry_shadow(struct tf_session *tfs, - struct tf_free_tbl_entry_parms *parms) -{ - TFP_DRV_LOG(ERR, - "%s, Entry Free with search not supported\n", - tf_dir_2_str(parms->dir)); + if (!init) { + TFP_DRV_LOG(ERR, + "%s: No Table DBs created\n", + tf_dir_2_str(parms->dir)); + return -EINVAL; + } - return -EOPNOTSUPP; -} -#endif /* TF_SHADOW */ + /* Verify that the entry has been previously allocated */ + aparms.rm_db = tbl_db[parms->dir]; + aparms.db_index = parms->type; + aparms.index = parms->idx; + aparms.allocated = &allocated; + rc = tf_rm_is_allocated(&aparms); + if (rc) + return rc; + if (!allocated) { + TFP_DRV_LOG(ERR, + "%s, Invalid or not allocated index, type:%d, idx:%d\n", + tf_dir_2_str(parms->dir), + parms->type, + parms->idx); + return -EINVAL; + } + + /* Set the entry */ + hparms.rm_db = tbl_db[parms->dir]; + hparms.db_index = parms->type; + hparms.hcapi_type = &hcapi_type; + rc = tf_rm_get_hcapi_type(&hparms); + if (rc) { + TFP_DRV_LOG(ERR, + "%s, Failed type lookup, type:%d, rc:%s\n", + tf_dir_2_str(parms->dir), + parms->type, + strerror(-rc)); + return rc; + } + + /* Get the entry */ + rc = tf_msg_get_tbl_entry(tfp, + parms->dir, + hcapi_type, + parms->data_sz_in_bytes, + parms->data, + parms->idx); + if (rc) { + TFP_DRV_LOG(ERR, + "%s, Get failed, type:%d, rc:%s\n", + tf_dir_2_str(parms->dir), + parms->type, + strerror(-rc)); + } + + return 0; +} - /* API defined in tf_core.h */ int -tf_bulk_get_tbl_entry(struct tf *tfp, - struct tf_bulk_get_tbl_entry_parms *parms) +tf_tbl_bulk_get(struct tf *tfp, + struct tf_tbl_get_bulk_parms *parms) { - int rc = 0; + int rc; + int i; + uint16_t hcapi_type; + uint32_t idx; + int allocated = 0; + struct tf_rm_is_allocated_parms aparms = { 0 }; + struct tf_rm_get_hcapi_parms hparms = { 0 }; - TF_CHECK_PARMS_SESSION(tfp, parms); + TF_CHECK_PARMS2(tfp, parms); - if (parms->type == TF_TBL_TYPE_EXT) { - /* Not supported, yet */ + if (!init) { TFP_DRV_LOG(ERR, - "%s, External table type not supported\n", + "%s: No Table DBs created\n", tf_dir_2_str(parms->dir)); - rc = -EOPNOTSUPP; - } else { - /* Internal table type processing */ - rc = tf_bulk_get_tbl_entry_internal(tfp, parms); + return -EINVAL; + } + /* Verify that the entries has been previously allocated */ + aparms.rm_db = tbl_db[parms->dir]; + aparms.db_index = parms->type; + aparms.allocated = &allocated; + idx = parms->starting_idx; + for (i = 0; i < parms->num_entries; i++) { + aparms.index = idx; + rc = tf_rm_is_allocated(&aparms); if (rc) + return rc; + + if (!allocated) { TFP_DRV_LOG(ERR, - "%s, Bulk get failed, type:%d, rc:%s\n", + "%s, Invalid or not allocated index, type:%d, idx:%d\n", tf_dir_2_str(parms->dir), parms->type, - strerror(-rc)); + idx); + return -EINVAL; + } + idx++; + } + + hparms.rm_db = tbl_db[parms->dir]; + hparms.db_index = parms->type; + hparms.hcapi_type = &hcapi_type; + rc = tf_rm_get_hcapi_type(&hparms); + if (rc) { + TFP_DRV_LOG(ERR, + "%s, Failed type lookup, type:%d, rc:%s\n", + tf_dir_2_str(parms->dir), + parms->type, + strerror(-rc)); + return rc; + } + + /* Get the entries */ + rc = tf_msg_bulk_get_tbl_entry(tfp, + parms->dir, + hcapi_type, + parms->starting_idx, + parms->num_entries, + parms->entry_sz_in_bytes, + parms->physical_mem_addr); + if (rc) { + TFP_DRV_LOG(ERR, + "%s, Bulk get failed, type:%d, rc:%s\n", + tf_dir_2_str(parms->dir), + parms->type, + strerror(-rc)); } return rc; diff --git a/drivers/net/bnxt/tf_core/tf_tbl.h b/drivers/net/bnxt/tf_core/tf_tbl.h index b175573457..eb560ffa72 100644 --- a/drivers/net/bnxt/tf_core/tf_tbl.h +++ b/drivers/net/bnxt/tf_core/tf_tbl.h @@ -3,17 +3,21 @@ * All rights reserved. */ -#ifndef _TF_TBL_H_ -#define _TF_TBL_H_ - -#include +#ifndef TF_TBL_TYPE_H_ +#define TF_TBL_TYPE_H_ #include "tf_core.h" #include "stack.h" -struct tf_session; +struct tf; + +/** + * The Table module provides processing of Internal TF table types. + */ -/** table scope control block content */ +/** + * Table scope control block content + */ struct tf_em_caps { uint32_t flags; uint32_t supported; @@ -35,66 +39,364 @@ struct tf_em_caps { struct tf_tbl_scope_cb { uint32_t tbl_scope_id; int index; - struct hcapi_cfa_em_ctx_mem_info em_ctx_info[TF_DIR_MAX]; - struct tf_em_caps em_caps[TF_DIR_MAX]; - struct stack ext_act_pool[TF_DIR_MAX]; - uint32_t *ext_act_pool_mem[TF_DIR_MAX]; + struct hcapi_cfa_em_ctx_mem_info em_ctx_info[TF_DIR_MAX]; + struct tf_em_caps em_caps[TF_DIR_MAX]; + struct stack ext_act_pool[TF_DIR_MAX]; + uint32_t *ext_act_pool_mem[TF_DIR_MAX]; +}; + +/** + * Table configuration parameters + */ +struct tf_tbl_cfg_parms { + /** + * Number of table types in each of the configuration arrays + */ + uint16_t num_elements; + /** + * Table Type element configuration array + */ + struct tf_rm_element_cfg *cfg; + /** + * Shadow table type configuration array + */ + struct tf_shadow_tbl_cfg *shadow_cfg; + /** + * Boolean controlling the request shadow copy. + */ + bool shadow_copy; + /** + * Session resource allocations + */ + struct tf_session_resources *resources; +}; + +/** + * Table allocation parameters + */ +struct tf_tbl_alloc_parms { + /** + * [in] Receive or transmit direction + */ + enum tf_dir dir; + /** + * [in] Type of the allocation + */ + enum tf_tbl_type type; + /** + * [out] Idx of allocated entry or found entry (if search_enable) + */ + uint32_t *idx; +}; + +/** + * Table free parameters + */ +struct tf_tbl_free_parms { + /** + * [in] Receive or transmit direction + */ + enum tf_dir dir; + /** + * [in] Type of the allocation type + */ + enum tf_tbl_type type; + /** + * [in] Index to free + */ + uint32_t idx; + /** + * [out] Reference count after free, only valid if session has been + * created with shadow_copy. + */ + uint16_t ref_cnt; }; -/** Hardware Page sizes supported for EEM: 4K, 8K, 64K, 256K, 1M, 2M, 4M, 1G. - * Round-down other page sizes to the lower hardware page size supported. - */ -#define TF_EM_PAGE_SIZE_4K 12 -#define TF_EM_PAGE_SIZE_8K 13 -#define TF_EM_PAGE_SIZE_64K 16 -#define TF_EM_PAGE_SIZE_256K 18 -#define TF_EM_PAGE_SIZE_1M 20 -#define TF_EM_PAGE_SIZE_2M 21 -#define TF_EM_PAGE_SIZE_4M 22 -#define TF_EM_PAGE_SIZE_1G 30 - -/* Set page size */ -#define BNXT_TF_PAGE_SIZE TF_EM_PAGE_SIZE_2M - -#if (BNXT_TF_PAGE_SIZE == TF_EM_PAGE_SIZE_4K) /** 4K */ -#define TF_EM_PAGE_SHIFT TF_EM_PAGE_SIZE_4K -#define TF_EM_PAGE_SIZE_ENUM HWRM_TF_CTXT_MEM_RGTR_INPUT_PAGE_SIZE_4K -#elif (BNXT_TF_PAGE_SIZE == TF_EM_PAGE_SIZE_8K) /** 8K */ -#define TF_EM_PAGE_SHIFT TF_EM_PAGE_SIZE_8K -#define TF_EM_PAGE_SIZE_ENUM HWRM_TF_CTXT_MEM_RGTR_INPUT_PAGE_SIZE_8K -#elif (BNXT_TF_PAGE_SIZE == TF_EM_PAGE_SIZE_64K) /** 64K */ -#define TF_EM_PAGE_SHIFT TF_EM_PAGE_SIZE_64K -#define TF_EM_PAGE_SIZE_ENUM HWRM_TF_CTXT_MEM_RGTR_INPUT_PAGE_SIZE_64K -#elif (BNXT_TF_PAGE_SIZE == TF_EM_PAGE_SIZE_256K) /** 256K */ -#define TF_EM_PAGE_SHIFT TF_EM_PAGE_SIZE_256K -#define TF_EM_PAGE_SIZE_ENUM HWRM_TF_CTXT_MEM_RGTR_INPUT_PAGE_SIZE_256K -#elif (BNXT_TF_PAGE_SIZE == TF_EM_PAGE_SIZE_1M) /** 1M */ -#define TF_EM_PAGE_SHIFT TF_EM_PAGE_SIZE_1M -#define TF_EM_PAGE_SIZE_ENUM HWRM_TF_CTXT_MEM_RGTR_INPUT_PAGE_SIZE_1M -#elif (BNXT_TF_PAGE_SIZE == TF_EM_PAGE_SIZE_2M) /** 2M */ -#define TF_EM_PAGE_SHIFT TF_EM_PAGE_SIZE_2M -#define TF_EM_PAGE_SIZE_ENUM HWRM_TF_CTXT_MEM_RGTR_INPUT_PAGE_SIZE_2M -#elif (BNXT_TF_PAGE_SIZE == TF_EM_PAGE_SIZE_4M) /** 4M */ -#define TF_EM_PAGE_SHIFT TF_EM_PAGE_SIZE_4M -#define TF_EM_PAGE_SIZE_ENUM HWRM_TF_CTXT_MEM_RGTR_INPUT_PAGE_SIZE_4M -#elif (BNXT_TF_PAGE_SIZE == TF_EM_PAGE_SIZE_1G) /** 1G */ -#define TF_EM_PAGE_SHIFT TF_EM_PAGE_SIZE_1G -#define TF_EM_PAGE_SIZE_ENUM HWRM_TF_CTXT_MEM_RGTR_INPUT_PAGE_SIZE_1G -#else -#error "Invalid Page Size specified. Please use a TF_EM_PAGE_SIZE_n define" -#endif - -#define TF_EM_PAGE_SIZE (1 << TF_EM_PAGE_SHIFT) -#define TF_EM_PAGE_ALIGNMENT (1 << TF_EM_PAGE_SHIFT) - -/** - * Initialize table pool structure to indicate - * no table scope has been associated with the - * external pool of indexes. - * - * [in] session - */ -void -tf_init_tbl_pool(struct tf_session *session); - -#endif /* _TF_TBL_H_ */ +/** + * Table allocate search parameters + */ +struct tf_tbl_alloc_search_parms { + /** + * [in] Receive or transmit direction + */ + enum tf_dir dir; + /** + * [in] Type of the allocation + */ + enum tf_tbl_type type; + /** + * [in] Table scope identifier (ignored unless TF_TBL_TYPE_EXT) + */ + uint32_t tbl_scope_id; + /** + * [in] Enable search for matching entry. If the table type is + * internal the shadow copy will be searched before + * alloc. Session must be configured with shadow copy enabled. + */ + uint8_t search_enable; + /** + * [in] Result data to search for (if search_enable) + */ + uint8_t *result; + /** + * [in] Result data size in bytes (if search_enable) + */ + uint16_t result_sz_in_bytes; + /** + * [out] If search_enable, set if matching entry found + */ + uint8_t hit; + /** + * [out] Current ref count after allocation (if search_enable) + */ + uint16_t ref_cnt; + /** + * [out] Idx of allocated entry or found entry (if search_enable) + */ + uint32_t idx; +}; + +/** + * Table set parameters + */ +struct tf_tbl_set_parms { + /** + * [in] Receive or transmit direction + */ + enum tf_dir dir; + /** + * [in] Type of object to set + */ + enum tf_tbl_type type; + /** + * [in] Entry data + */ + uint8_t *data; + /** + * [in] Entry size + */ + uint16_t data_sz_in_bytes; + /** + * [in] Entry index to write to + */ + uint32_t idx; +}; + +/** + * Table get parameters + */ +struct tf_tbl_get_parms { + /** + * [in] Receive or transmit direction + */ + enum tf_dir dir; + /** + * [in] Type of object to get + */ + enum tf_tbl_type type; + /** + * [out] Entry data + */ + uint8_t *data; + /** + * [out] Entry size + */ + uint16_t data_sz_in_bytes; + /** + * [in] Entry index to read + */ + uint32_t idx; +}; + +/** + * Table get bulk parameters + */ +struct tf_tbl_get_bulk_parms { + /** + * [in] Receive or transmit direction + */ + enum tf_dir dir; + /** + * [in] Type of object to get + */ + enum tf_tbl_type type; + /** + * [in] Starting index to read from + */ + uint32_t starting_idx; + /** + * [in] Number of sequential entries + */ + uint16_t num_entries; + /** + * [in] Size of the single entry + */ + uint16_t entry_sz_in_bytes; + /** + * [out] Host physical address, where the data + * will be copied to by the firmware. + * Use tfp_calloc() API and mem_pa + * variable of the tfp_calloc_parms + * structure for the physical address. + */ + uint64_t physical_mem_addr; +}; + +/** + * @page tbl Table + * + * @ref tf_tbl_bind + * + * @ref tf_tbl_unbind + * + * @ref tf_tbl_alloc + * + * @ref tf_tbl_free + * + * @ref tf_tbl_alloc_search + * + * @ref tf_tbl_set + * + * @ref tf_tbl_get + * + * @ref tf_tbl_bulk_get + */ + +/** + * Initializes the Table module with the requested DBs. Must be + * invoked as the first thing before any of the access functions. + * + * [in] tfp + * Pointer to TF handle, used for HCAPI communication + * + * [in] parms + * Pointer to Table configuration parameters + * + * Returns + * - (0) if successful. + * - (-EINVAL) on failure. + */ +int tf_tbl_bind(struct tf *tfp, + struct tf_tbl_cfg_parms *parms); + +/** + * Cleans up the private DBs and releases all the data. + * + * [in] tfp + * Pointer to TF handle, used for HCAPI communication + * + * [in] parms + * Pointer to parameters + * + * Returns + * - (0) if successful. + * - (-EINVAL) on failure. + */ +int tf_tbl_unbind(struct tf *tfp); + +/** + * Allocates the requested table type from the internal RM DB. + * + * [in] tfp + * Pointer to TF handle, used for HCAPI communication + * + * [in] parms + * Pointer to Table allocation parameters + * + * Returns + * - (0) if successful. + * - (-EINVAL) on failure. + */ +int tf_tbl_alloc(struct tf *tfp, + struct tf_tbl_alloc_parms *parms); + +/** + * Free's the requested table type and returns it to the DB. If shadow + * DB is enabled its searched first and if found the element refcount + * is decremented. If refcount goes to 0 then its returned to the + * table type DB. + * + * [in] tfp + * Pointer to TF handle, used for HCAPI communication + * + * [in] parms + * Pointer to Table free parameters + * + * Returns + * - (0) if successful. + * - (-EINVAL) on failure. + */ +int tf_tbl_free(struct tf *tfp, + struct tf_tbl_free_parms *parms); + +/** + * Supported if Shadow DB is configured. Searches the Shadow DB for + * any matching element. If found the refcount in the shadow DB is + * updated accordingly. If not found a new element is allocated and + * installed into the shadow DB. + * + * [in] tfp + * Pointer to TF handle, used for HCAPI communication + * + * [in] parms + * Pointer to parameters + * + * Returns + * - (0) if successful. + * - (-EINVAL) on failure. + */ +int tf_tbl_alloc_search(struct tf *tfp, + struct tf_tbl_alloc_search_parms *parms); + +/** + * Configures the requested element by sending a firmware request which + * then installs it into the device internal structures. + * + * [in] tfp + * Pointer to TF handle, used for HCAPI communication + * + * [in] parms + * Pointer to Table set parameters + * + * Returns + * - (0) if successful. + * - (-EINVAL) on failure. + */ +int tf_tbl_set(struct tf *tfp, + struct tf_tbl_set_parms *parms); + +/** + * Retrieves the requested element by sending a firmware request to get + * the element. + * + * [in] tfp + * Pointer to TF handle, used for HCAPI communication + * + * [in] parms + * Pointer to Table get parameters + * + * Returns + * - (0) if successful. + * - (-EINVAL) on failure. + */ +int tf_tbl_get(struct tf *tfp, + struct tf_tbl_get_parms *parms); + +/** + * Retrieves bulk block of elements by sending a firmware request to + * get the elements. + * + * [in] tfp + * Pointer to TF handle, used for HCAPI communication + * + * [in] parms + * Pointer to Table get bulk parameters + * + * Returns + * - (0) if successful. + * - (-EINVAL) on failure. + */ +int tf_tbl_bulk_get(struct tf *tfp, + struct tf_tbl_get_bulk_parms *parms); + +#endif /* TF_TBL_TYPE_H */ diff --git a/drivers/net/bnxt/tf_core/tf_tbl_type.c b/drivers/net/bnxt/tf_core/tf_tbl_type.c deleted file mode 100644 index 2f5af6060a..0000000000 --- a/drivers/net/bnxt/tf_core/tf_tbl_type.c +++ /dev/null @@ -1,342 +0,0 @@ -/* SPDX-License-Identifier: BSD-3-Clause - * Copyright(c) 2019-2020 Broadcom - * All rights reserved. - */ - -#include - -#include "tf_tbl_type.h" -#include "tf_common.h" -#include "tf_rm_new.h" -#include "tf_util.h" -#include "tf_msg.h" -#include "tfp.h" - -struct tf; - -/** - * Table DBs. - */ -static void *tbl_db[TF_DIR_MAX]; - -/** - * Table Shadow DBs - */ -/* static void *shadow_tbl_db[TF_DIR_MAX]; */ - -/** - * Init flag, set on bind and cleared on unbind - */ -static uint8_t init; - -/** - * Shadow init flag, set on bind and cleared on unbind - */ -/* static uint8_t shadow_init; */ - -int -tf_tbl_bind(struct tf *tfp, - struct tf_tbl_cfg_parms *parms) -{ - int rc; - int i; - struct tf_rm_create_db_parms db_cfg = { 0 }; - - TF_CHECK_PARMS2(tfp, parms); - - if (init) { - TFP_DRV_LOG(ERR, - "Table already initialized\n"); - return -EINVAL; - } - - db_cfg.num_elements = parms->num_elements; - db_cfg.type = TF_DEVICE_MODULE_TYPE_TABLE; - db_cfg.num_elements = parms->num_elements; - db_cfg.cfg = parms->cfg; - - for (i = 0; i < TF_DIR_MAX; i++) { - db_cfg.dir = i; - db_cfg.alloc_cnt = parms->resources->tbl_cnt[i].cnt; - db_cfg.rm_db = &tbl_db[i]; - rc = tf_rm_create_db(tfp, &db_cfg); - if (rc) { - TFP_DRV_LOG(ERR, - "%s: Table DB creation failed\n", - tf_dir_2_str(i)); - - return rc; - } - } - - init = 1; - - printf("Table Type - initialized\n"); - - return 0; -} - -int -tf_tbl_unbind(struct tf *tfp __rte_unused) -{ - int rc; - int i; - struct tf_rm_free_db_parms fparms = { 0 }; - - TF_CHECK_PARMS1(tfp); - - /* Bail if nothing has been initialized done silent as to - * allow for creation cleanup. - */ - if (!init) { - TFP_DRV_LOG(ERR, - "No Table DBs created\n"); - return -EINVAL; - } - - for (i = 0; i < TF_DIR_MAX; i++) { - fparms.dir = i; - fparms.rm_db = tbl_db[i]; - rc = tf_rm_free_db(tfp, &fparms); - if (rc) - return rc; - - tbl_db[i] = NULL; - } - - init = 0; - - return 0; -} - -int -tf_tbl_alloc(struct tf *tfp __rte_unused, - struct tf_tbl_alloc_parms *parms) -{ - int rc; - uint32_t idx; - struct tf_rm_allocate_parms aparms = { 0 }; - - TF_CHECK_PARMS2(tfp, parms); - - if (!init) { - TFP_DRV_LOG(ERR, - "%s: No Table DBs created\n", - tf_dir_2_str(parms->dir)); - return -EINVAL; - } - - /* Allocate requested element */ - aparms.rm_db = tbl_db[parms->dir]; - aparms.db_index = parms->type; - aparms.index = &idx; - rc = tf_rm_allocate(&aparms); - if (rc) { - TFP_DRV_LOG(ERR, - "%s: Failed allocate, type:%d\n", - tf_dir_2_str(parms->dir), - parms->type); - return rc; - } - - *parms->idx = idx; - - return 0; -} - -int -tf_tbl_free(struct tf *tfp __rte_unused, - struct tf_tbl_free_parms *parms) -{ - int rc; - struct tf_rm_is_allocated_parms aparms = { 0 }; - struct tf_rm_free_parms fparms = { 0 }; - int allocated = 0; - - TF_CHECK_PARMS2(tfp, parms); - - if (!init) { - TFP_DRV_LOG(ERR, - "%s: No Table DBs created\n", - tf_dir_2_str(parms->dir)); - return -EINVAL; - } - - /* Check if element is in use */ - aparms.rm_db = tbl_db[parms->dir]; - aparms.db_index = parms->type; - aparms.index = parms->idx; - aparms.allocated = &allocated; - rc = tf_rm_is_allocated(&aparms); - if (rc) - return rc; - - if (!allocated) { - TFP_DRV_LOG(ERR, - "%s: Entry already free, type:%d, index:%d\n", - tf_dir_2_str(parms->dir), - parms->type, - parms->idx); - return rc; - } - - /* Free requested element */ - fparms.rm_db = tbl_db[parms->dir]; - fparms.db_index = parms->type; - fparms.index = parms->idx; - rc = tf_rm_free(&fparms); - if (rc) { - TFP_DRV_LOG(ERR, - "%s: Free failed, type:%d, index:%d\n", - tf_dir_2_str(parms->dir), - parms->type, - parms->idx); - return rc; - } - - return 0; -} - -int -tf_tbl_alloc_search(struct tf *tfp __rte_unused, - struct tf_tbl_alloc_search_parms *parms __rte_unused) -{ - return 0; -} - -int -tf_tbl_set(struct tf *tfp, - struct tf_tbl_set_parms *parms) -{ - int rc; - int allocated = 0; - uint16_t hcapi_type; - struct tf_rm_is_allocated_parms aparms = { 0 }; - struct tf_rm_get_hcapi_parms hparms = { 0 }; - - TF_CHECK_PARMS3(tfp, parms, parms->data); - - if (!init) { - TFP_DRV_LOG(ERR, - "%s: No Table DBs created\n", - tf_dir_2_str(parms->dir)); - return -EINVAL; - } - - /* Verify that the entry has been previously allocated */ - aparms.rm_db = tbl_db[parms->dir]; - aparms.db_index = parms->type; - aparms.index = parms->idx; - aparms.allocated = &allocated; - rc = tf_rm_is_allocated(&aparms); - if (rc) - return rc; - - if (!allocated) { - TFP_DRV_LOG(ERR, - "%s, Invalid or not allocated index, type:%d, idx:%d\n", - tf_dir_2_str(parms->dir), - parms->type, - parms->idx); - return -EINVAL; - } - - /* Set the entry */ - hparms.rm_db = tbl_db[parms->dir]; - hparms.db_index = parms->type; - hparms.hcapi_type = &hcapi_type; - rc = tf_rm_get_hcapi_type(&hparms); - if (rc) { - TFP_DRV_LOG(ERR, - "%s, Failed type lookup, type:%d, rc:%s\n", - tf_dir_2_str(parms->dir), - parms->type, - strerror(-rc)); - return rc; - } - - rc = tf_msg_set_tbl_entry(tfp, - parms->dir, - hcapi_type, - parms->data_sz_in_bytes, - parms->data, - parms->idx); - if (rc) { - TFP_DRV_LOG(ERR, - "%s, Set failed, type:%d, rc:%s\n", - tf_dir_2_str(parms->dir), - parms->type, - strerror(-rc)); - } - - return 0; -} - -int -tf_tbl_get(struct tf *tfp, - struct tf_tbl_get_parms *parms) -{ - int rc; - uint16_t hcapi_type; - int allocated = 0; - struct tf_rm_is_allocated_parms aparms = { 0 }; - struct tf_rm_get_hcapi_parms hparms = { 0 }; - - TF_CHECK_PARMS3(tfp, parms, parms->data); - - if (!init) { - TFP_DRV_LOG(ERR, - "%s: No Table DBs created\n", - tf_dir_2_str(parms->dir)); - return -EINVAL; - } - - /* Verify that the entry has been previously allocated */ - aparms.rm_db = tbl_db[parms->dir]; - aparms.db_index = parms->type; - aparms.index = parms->idx; - aparms.allocated = &allocated; - rc = tf_rm_is_allocated(&aparms); - if (rc) - return rc; - - if (!allocated) { - TFP_DRV_LOG(ERR, - "%s, Invalid or not allocated index, type:%d, idx:%d\n", - tf_dir_2_str(parms->dir), - parms->type, - parms->idx); - return -EINVAL; - } - - /* Set the entry */ - hparms.rm_db = tbl_db[parms->dir]; - hparms.db_index = parms->type; - hparms.hcapi_type = &hcapi_type; - rc = tf_rm_get_hcapi_type(&hparms); - if (rc) { - TFP_DRV_LOG(ERR, - "%s, Failed type lookup, type:%d, rc:%s\n", - tf_dir_2_str(parms->dir), - parms->type, - strerror(-rc)); - return rc; - } - - /* Get the entry */ - rc = tf_msg_get_tbl_entry(tfp, - parms->dir, - hcapi_type, - parms->data_sz_in_bytes, - parms->data, - parms->idx); - if (rc) { - TFP_DRV_LOG(ERR, - "%s, Get failed, type:%d, rc:%s\n", - tf_dir_2_str(parms->dir), - parms->type, - strerror(-rc)); - } - - return 0; -} diff --git a/drivers/net/bnxt/tf_core/tf_tbl_type.h b/drivers/net/bnxt/tf_core/tf_tbl_type.h deleted file mode 100644 index 3474489a6c..0000000000 --- a/drivers/net/bnxt/tf_core/tf_tbl_type.h +++ /dev/null @@ -1,318 +0,0 @@ -/* SPDX-License-Identifier: BSD-3-Clause - * Copyright(c) 2019-2020 Broadcom - * All rights reserved. - */ - -#ifndef TF_TBL_TYPE_H_ -#define TF_TBL_TYPE_H_ - -#include "tf_core.h" - -struct tf; - -/** - * The Table module provides processing of Internal TF table types. - */ - -/** - * Table configuration parameters - */ -struct tf_tbl_cfg_parms { - /** - * Number of table types in each of the configuration arrays - */ - uint16_t num_elements; - /** - * Table Type element configuration array - */ - struct tf_rm_element_cfg *cfg; - /** - * Shadow table type configuration array - */ - struct tf_shadow_tbl_cfg *shadow_cfg; - /** - * Boolean controlling the request shadow copy. - */ - bool shadow_copy; - /** - * Session resource allocations - */ - struct tf_session_resources *resources; -}; - -/** - * Table allocation parameters - */ -struct tf_tbl_alloc_parms { - /** - * [in] Receive or transmit direction - */ - enum tf_dir dir; - /** - * [in] Type of the allocation - */ - enum tf_tbl_type type; - /** - * [out] Idx of allocated entry or found entry (if search_enable) - */ - uint32_t *idx; -}; - -/** - * Table free parameters - */ -struct tf_tbl_free_parms { - /** - * [in] Receive or transmit direction - */ - enum tf_dir dir; - /** - * [in] Type of the allocation type - */ - enum tf_tbl_type type; - /** - * [in] Index to free - */ - uint32_t idx; - /** - * [out] Reference count after free, only valid if session has been - * created with shadow_copy. - */ - uint16_t ref_cnt; -}; - -/** - * Table allocate search parameters - */ -struct tf_tbl_alloc_search_parms { - /** - * [in] Receive or transmit direction - */ - enum tf_dir dir; - /** - * [in] Type of the allocation - */ - enum tf_tbl_type type; - /** - * [in] Table scope identifier (ignored unless TF_TBL_TYPE_EXT) - */ - uint32_t tbl_scope_id; - /** - * [in] Enable search for matching entry. If the table type is - * internal the shadow copy will be searched before - * alloc. Session must be configured with shadow copy enabled. - */ - uint8_t search_enable; - /** - * [in] Result data to search for (if search_enable) - */ - uint8_t *result; - /** - * [in] Result data size in bytes (if search_enable) - */ - uint16_t result_sz_in_bytes; - /** - * [out] If search_enable, set if matching entry found - */ - uint8_t hit; - /** - * [out] Current ref count after allocation (if search_enable) - */ - uint16_t ref_cnt; - /** - * [out] Idx of allocated entry or found entry (if search_enable) - */ - uint32_t idx; -}; - -/** - * Table set parameters - */ -struct tf_tbl_set_parms { - /** - * [in] Receive or transmit direction - */ - enum tf_dir dir; - /** - * [in] Type of object to set - */ - enum tf_tbl_type type; - /** - * [in] Entry data - */ - uint8_t *data; - /** - * [in] Entry size - */ - uint16_t data_sz_in_bytes; - /** - * [in] Entry index to write to - */ - uint32_t idx; -}; - -/** - * Table get parameters - */ -struct tf_tbl_get_parms { - /** - * [in] Receive or transmit direction - */ - enum tf_dir dir; - /** - * [in] Type of object to get - */ - enum tf_tbl_type type; - /** - * [out] Entry data - */ - uint8_t *data; - /** - * [out] Entry size - */ - uint16_t data_sz_in_bytes; - /** - * [in] Entry index to read - */ - uint32_t idx; -}; - -/** - * @page tbl Table - * - * @ref tf_tbl_bind - * - * @ref tf_tbl_unbind - * - * @ref tf_tbl_alloc - * - * @ref tf_tbl_free - * - * @ref tf_tbl_alloc_search - * - * @ref tf_tbl_set - * - * @ref tf_tbl_get - */ - -/** - * Initializes the Table module with the requested DBs. Must be - * invoked as the first thing before any of the access functions. - * - * [in] tfp - * Pointer to TF handle, used for HCAPI communication - * - * [in] parms - * Pointer to Table configuration parameters - * - * Returns - * - (0) if successful. - * - (-EINVAL) on failure. - */ -int tf_tbl_bind(struct tf *tfp, - struct tf_tbl_cfg_parms *parms); - -/** - * Cleans up the private DBs and releases all the data. - * - * [in] tfp - * Pointer to TF handle, used for HCAPI communication - * - * [in] parms - * Pointer to parameters - * - * Returns - * - (0) if successful. - * - (-EINVAL) on failure. - */ -int tf_tbl_unbind(struct tf *tfp); - -/** - * Allocates the requested table type from the internal RM DB. - * - * [in] tfp - * Pointer to TF handle, used for HCAPI communication - * - * [in] parms - * Pointer to Table allocation parameters - * - * Returns - * - (0) if successful. - * - (-EINVAL) on failure. - */ -int tf_tbl_alloc(struct tf *tfp, - struct tf_tbl_alloc_parms *parms); - -/** - * Free's the requested table type and returns it to the DB. If shadow - * DB is enabled its searched first and if found the element refcount - * is decremented. If refcount goes to 0 then its returned to the - * table type DB. - * - * [in] tfp - * Pointer to TF handle, used for HCAPI communication - * - * [in] parms - * Pointer to Table free parameters - * - * Returns - * - (0) if successful. - * - (-EINVAL) on failure. - */ -int tf_tbl_free(struct tf *tfp, - struct tf_tbl_free_parms *parms); - -/** - * Supported if Shadow DB is configured. Searches the Shadow DB for - * any matching element. If found the refcount in the shadow DB is - * updated accordingly. If not found a new element is allocated and - * installed into the shadow DB. - * - * [in] tfp - * Pointer to TF handle, used for HCAPI communication - * - * [in] parms - * Pointer to parameters - * - * Returns - * - (0) if successful. - * - (-EINVAL) on failure. - */ -int tf_tbl_alloc_search(struct tf *tfp, - struct tf_tbl_alloc_search_parms *parms); - -/** - * Configures the requested element by sending a firmware request which - * then installs it into the device internal structures. - * - * [in] tfp - * Pointer to TF handle, used for HCAPI communication - * - * [in] parms - * Pointer to Table set parameters - * - * Returns - * - (0) if successful. - * - (-EINVAL) on failure. - */ -int tf_tbl_set(struct tf *tfp, - struct tf_tbl_set_parms *parms); - -/** - * Retrieves the requested element by sending a firmware request to get - * the element. - * - * [in] tfp - * Pointer to TF handle, used for HCAPI communication - * - * [in] parms - * Pointer to Table get parameters - * - * Returns - * - (0) if successful. - * - (-EINVAL) on failure. - */ -int tf_tbl_get(struct tf *tfp, - struct tf_tbl_get_parms *parms); - -#endif /* TF_TBL_TYPE_H */ diff --git a/drivers/net/bnxt/tf_core/tf_tcam.c b/drivers/net/bnxt/tf_core/tf_tcam.c index a1761ad567..fc047f8f80 100644 --- a/drivers/net/bnxt/tf_core/tf_tcam.c +++ b/drivers/net/bnxt/tf_core/tf_tcam.c @@ -9,7 +9,7 @@ #include "tf_tcam.h" #include "tf_common.h" #include "tf_util.h" -#include "tf_rm_new.h" +#include "tf_rm.h" #include "tf_device.h" #include "tfp.h" #include "tf_session.h" @@ -49,7 +49,7 @@ tf_tcam_bind(struct tf *tfp, if (init) { TFP_DRV_LOG(ERR, - "TCAM already initialized\n"); + "TCAM DB already initialized\n"); return -EINVAL; } @@ -86,11 +86,12 @@ tf_tcam_unbind(struct tf *tfp) TF_CHECK_PARMS1(tfp); - /* Bail if nothing has been initialized done silent as to - * allow for creation cleanup. - */ - if (!init) - return -EINVAL; + /* Bail if nothing has been initialized */ + if (!init) { + TFP_DRV_LOG(INFO, + "No TCAM DBs created\n"); + return 0; + } for (i = 0; i < TF_DIR_MAX; i++) { fparms.dir = i;