net/qede/base: add UFP support
[dpdk.git] / drivers / net / qede / base / ecore_dev.c
index 1608b19..283c65b 100644 (file)
 static osal_spinlock_t qm_lock;
 static bool qm_lock_init;
 
+/******************** Doorbell Recovery *******************/
+/* The doorbell recovery mechanism consists of a list of entries which represent
+ * doorbelling entities (l2 queues, roce sq/rq/cqs, the slowpath spq, etc). Each
+ * entity needs to register with the mechanism and provide the parameters
+ * describing it's doorbell, including a location where last used doorbell data
+ * can be found. The doorbell execute function will traverse the list and
+ * doorbell all of the registered entries.
+ */
+struct ecore_db_recovery_entry {
+       osal_list_entry_t       list_entry;
+       void OSAL_IOMEM         *db_addr;
+       void                    *db_data;
+       enum ecore_db_rec_width db_width;
+       enum ecore_db_rec_space db_space;
+       u8                      hwfn_idx;
+};
+
+/* display a single doorbell recovery entry */
+void ecore_db_recovery_dp_entry(struct ecore_hwfn *p_hwfn,
+                               struct ecore_db_recovery_entry *db_entry,
+                               const char *action)
+{
+       DP_VERBOSE(p_hwfn, ECORE_MSG_SPQ, "(%s: db_entry %p, addr %p, data %p, width %s, %s space, hwfn %d)\n",
+                  action, db_entry, db_entry->db_addr, db_entry->db_data,
+                  db_entry->db_width == DB_REC_WIDTH_32B ? "32b" : "64b",
+                  db_entry->db_space == DB_REC_USER ? "user" : "kernel",
+                  db_entry->hwfn_idx);
+}
+
+/* doorbell address sanity (address within doorbell bar range) */
+bool ecore_db_rec_sanity(struct ecore_dev *p_dev, void OSAL_IOMEM *db_addr,
+                        void *db_data)
+{
+       /* make sure doorbell address  is within the doorbell bar */
+       if (db_addr < p_dev->doorbells || (u8 *)db_addr >
+                       (u8 *)p_dev->doorbells + p_dev->db_size) {
+               OSAL_WARN(true,
+                         "Illegal doorbell address: %p. Legal range for doorbell addresses is [%p..%p]\n",
+                         db_addr, p_dev->doorbells,
+                         (u8 *)p_dev->doorbells + p_dev->db_size);
+               return false;
+       }
+
+       /* make sure doorbell data pointer is not null */
+       if (!db_data) {
+               OSAL_WARN(true, "Illegal doorbell data pointer: %p", db_data);
+               return false;
+       }
+
+       return true;
+}
+
+/* find hwfn according to the doorbell address */
+struct ecore_hwfn *ecore_db_rec_find_hwfn(struct ecore_dev *p_dev,
+                                         void OSAL_IOMEM *db_addr)
+{
+       struct ecore_hwfn *p_hwfn;
+
+       /* In CMT doorbell bar is split down the middle between engine 0 and
+        * enigne 1
+        */
+       if (ECORE_IS_CMT(p_dev))
+               p_hwfn = db_addr < p_dev->hwfns[1].doorbells ?
+                       &p_dev->hwfns[0] : &p_dev->hwfns[1];
+       else
+               p_hwfn = ECORE_LEADING_HWFN(p_dev);
+
+       return p_hwfn;
+}
+
+/* add a new entry to the doorbell recovery mechanism */
+enum _ecore_status_t ecore_db_recovery_add(struct ecore_dev *p_dev,
+                                          void OSAL_IOMEM *db_addr,
+                                          void *db_data,
+                                          enum ecore_db_rec_width db_width,
+                                          enum ecore_db_rec_space db_space)
+{
+       struct ecore_db_recovery_entry *db_entry;
+       struct ecore_hwfn *p_hwfn;
+
+       /* shortcircuit VFs, for now */
+       if (IS_VF(p_dev)) {
+               DP_VERBOSE(p_dev, ECORE_MSG_IOV, "db recovery - skipping VF doorbell\n");
+               return ECORE_SUCCESS;
+       }
+
+       /* sanitize doorbell address */
+       if (!ecore_db_rec_sanity(p_dev, db_addr, db_data))
+               return ECORE_INVAL;
+
+       /* obtain hwfn from doorbell address */
+       p_hwfn = ecore_db_rec_find_hwfn(p_dev, db_addr);
+
+       /* create entry */
+       db_entry = OSAL_ZALLOC(p_hwfn->p_dev, GFP_KERNEL, sizeof(*db_entry));
+       if (!db_entry) {
+               DP_NOTICE(p_dev, false, "Failed to allocate a db recovery entry\n");
+               return ECORE_NOMEM;
+       }
+
+       /* populate entry */
+       db_entry->db_addr = db_addr;
+       db_entry->db_data = db_data;
+       db_entry->db_width = db_width;
+       db_entry->db_space = db_space;
+       db_entry->hwfn_idx = p_hwfn->my_id;
+
+       /* display */
+       ecore_db_recovery_dp_entry(p_hwfn, db_entry, "Adding");
+
+       /* protect the list */
+       OSAL_SPIN_LOCK(&p_hwfn->db_recovery_info.lock);
+       OSAL_LIST_PUSH_TAIL(&db_entry->list_entry,
+                           &p_hwfn->db_recovery_info.list);
+       OSAL_SPIN_UNLOCK(&p_hwfn->db_recovery_info.lock);
+
+       return ECORE_SUCCESS;
+}
+
+/* remove an entry from the doorbell recovery mechanism */
+enum _ecore_status_t ecore_db_recovery_del(struct ecore_dev *p_dev,
+                                          void OSAL_IOMEM *db_addr,
+                                          void *db_data)
+{
+       struct ecore_db_recovery_entry *db_entry = OSAL_NULL;
+       enum _ecore_status_t rc = ECORE_INVAL;
+       struct ecore_hwfn *p_hwfn;
+
+       /* shortcircuit VFs, for now */
+       if (IS_VF(p_dev)) {
+               DP_VERBOSE(p_dev, ECORE_MSG_IOV, "db recovery - skipping VF doorbell\n");
+               return ECORE_SUCCESS;
+       }
+
+       /* sanitize doorbell address */
+       if (!ecore_db_rec_sanity(p_dev, db_addr, db_data))
+               return ECORE_INVAL;
+
+       /* obtain hwfn from doorbell address */
+       p_hwfn = ecore_db_rec_find_hwfn(p_dev, db_addr);
+
+       /* protect the list */
+       OSAL_SPIN_LOCK(&p_hwfn->db_recovery_info.lock);
+       OSAL_LIST_FOR_EACH_ENTRY(db_entry,
+                                &p_hwfn->db_recovery_info.list,
+                                list_entry,
+                                struct ecore_db_recovery_entry) {
+               /* search according to db_data addr since db_addr is not unique
+                * (roce)
+                */
+               if (db_entry->db_data == db_data) {
+                       ecore_db_recovery_dp_entry(p_hwfn, db_entry,
+                                                  "Deleting");
+                       OSAL_LIST_REMOVE_ENTRY(&db_entry->list_entry,
+                                              &p_hwfn->db_recovery_info.list);
+                       rc = ECORE_SUCCESS;
+                       break;
+               }
+       }
+
+       OSAL_SPIN_UNLOCK(&p_hwfn->db_recovery_info.lock);
+
+       if (rc == ECORE_INVAL)
+               /*OSAL_WARN(true,*/
+               DP_NOTICE(p_hwfn, false,
+                         "Failed to find element in list. Key (db_data addr) was %p. db_addr was %p\n",
+                         db_data, db_addr);
+       else
+               OSAL_FREE(p_dev, db_entry);
+
+       return rc;
+}
+
+/* initialize the doorbell recovery mechanism */
+enum _ecore_status_t ecore_db_recovery_setup(struct ecore_hwfn *p_hwfn)
+{
+       DP_VERBOSE(p_hwfn, ECORE_MSG_SPQ, "Setting up db recovery\n");
+
+       /* make sure db_size was set in p_dev */
+       if (!p_hwfn->p_dev->db_size) {
+               DP_ERR(p_hwfn->p_dev, "db_size not set\n");
+               return ECORE_INVAL;
+       }
+
+       OSAL_LIST_INIT(&p_hwfn->db_recovery_info.list);
+#ifdef CONFIG_ECORE_LOCK_ALLOC
+       OSAL_SPIN_LOCK_ALLOC(p_hwfn, &p_hwfn->db_recovery_info.lock);
+#endif
+       OSAL_SPIN_LOCK_INIT(&p_hwfn->db_recovery_info.lock);
+       p_hwfn->db_recovery_info.db_recovery_counter = 0;
+
+       return ECORE_SUCCESS;
+}
+
+/* destroy the doorbell recovery mechanism */
+void ecore_db_recovery_teardown(struct ecore_hwfn *p_hwfn)
+{
+       struct ecore_db_recovery_entry *db_entry = OSAL_NULL;
+
+       DP_VERBOSE(p_hwfn, ECORE_MSG_SPQ, "Tearing down db recovery\n");
+       if (!OSAL_LIST_IS_EMPTY(&p_hwfn->db_recovery_info.list)) {
+               DP_VERBOSE(p_hwfn, false, "Doorbell Recovery teardown found the doorbell recovery list was not empty (Expected in disorderly driver unload (e.g. recovery) otherwise this probably means some flow forgot to db_recovery_del). Prepare to purge doorbell recovery list...\n");
+               while (!OSAL_LIST_IS_EMPTY(&p_hwfn->db_recovery_info.list)) {
+                       db_entry = OSAL_LIST_FIRST_ENTRY(
+                                               &p_hwfn->db_recovery_info.list,
+                                               struct ecore_db_recovery_entry,
+                                               list_entry);
+                       ecore_db_recovery_dp_entry(p_hwfn, db_entry, "Purging");
+                       OSAL_LIST_REMOVE_ENTRY(&db_entry->list_entry,
+                                              &p_hwfn->db_recovery_info.list);
+                       OSAL_FREE(p_hwfn->p_dev, db_entry);
+               }
+       }
+#ifdef CONFIG_ECORE_LOCK_ALLOC
+       OSAL_SPIN_LOCK_DEALLOC(&p_hwfn->db_recovery_info.lock);
+#endif
+       p_hwfn->db_recovery_info.db_recovery_counter = 0;
+}
+
+/* print the content of the doorbell recovery mechanism */
+void ecore_db_recovery_dp(struct ecore_hwfn *p_hwfn)
+{
+       struct ecore_db_recovery_entry *db_entry = OSAL_NULL;
+
+       DP_NOTICE(p_hwfn, false,
+                 "Dispalying doorbell recovery database. Counter was %d\n",
+                 p_hwfn->db_recovery_info.db_recovery_counter);
+
+       /* protect the list */
+       OSAL_SPIN_LOCK(&p_hwfn->db_recovery_info.lock);
+       OSAL_LIST_FOR_EACH_ENTRY(db_entry,
+                                &p_hwfn->db_recovery_info.list,
+                                list_entry,
+                                struct ecore_db_recovery_entry) {
+               ecore_db_recovery_dp_entry(p_hwfn, db_entry, "Printing");
+       }
+
+       OSAL_SPIN_UNLOCK(&p_hwfn->db_recovery_info.lock);
+}
+
+/* ring the doorbell of a single doorbell recovery entry */
+void ecore_db_recovery_ring(struct ecore_hwfn *p_hwfn,
+                           struct ecore_db_recovery_entry *db_entry,
+                           enum ecore_db_rec_exec db_exec)
+{
+       /* Print according to width */
+       if (db_entry->db_width == DB_REC_WIDTH_32B)
+               DP_VERBOSE(p_hwfn, ECORE_MSG_SPQ, "%s doorbell address %p data %x\n",
+                          db_exec == DB_REC_DRY_RUN ? "would have rung" : "ringing",
+                          db_entry->db_addr, *(u32 *)db_entry->db_data);
+       else
+               DP_VERBOSE(p_hwfn, ECORE_MSG_SPQ, "%s doorbell address %p data %lx\n",
+                          db_exec == DB_REC_DRY_RUN ? "would have rung" : "ringing",
+                          db_entry->db_addr,
+                          *(unsigned long *)(db_entry->db_data));
+
+       /* Sanity */
+       if (!ecore_db_rec_sanity(p_hwfn->p_dev, db_entry->db_addr,
+                                db_entry->db_data))
+               return;
+
+       /* Flush the write combined buffer. Since there are multiple doorbelling
+        * entities using the same address, if we don't flush, a transaction
+        * could be lost.
+        */
+       OSAL_WMB(p_hwfn->p_dev);
+
+       /* Ring the doorbell */
+       if (db_exec == DB_REC_REAL_DEAL || db_exec == DB_REC_ONCE) {
+               if (db_entry->db_width == DB_REC_WIDTH_32B)
+                       DIRECT_REG_WR(p_hwfn, db_entry->db_addr,
+                                     *(u32 *)(db_entry->db_data));
+               else
+                       DIRECT_REG_WR64(p_hwfn, db_entry->db_addr,
+                                       *(u64 *)(db_entry->db_data));
+       }
+
+       /* Flush the write combined buffer. Next doorbell may come from a
+        * different entity to the same address...
+        */
+       OSAL_WMB(p_hwfn->p_dev);
+}
+
+/* traverse the doorbell recovery entry list and ring all the doorbells */
+void ecore_db_recovery_execute(struct ecore_hwfn *p_hwfn,
+                              enum ecore_db_rec_exec db_exec)
+{
+       struct ecore_db_recovery_entry *db_entry = OSAL_NULL;
+
+       if (db_exec != DB_REC_ONCE) {
+               DP_NOTICE(p_hwfn, false, "Executing doorbell recovery. Counter was %d\n",
+                         p_hwfn->db_recovery_info.db_recovery_counter);
+
+               /* track amount of times recovery was executed */
+               p_hwfn->db_recovery_info.db_recovery_counter++;
+       }
+
+       /* protect the list */
+       OSAL_SPIN_LOCK(&p_hwfn->db_recovery_info.lock);
+       OSAL_LIST_FOR_EACH_ENTRY(db_entry,
+                                &p_hwfn->db_recovery_info.list,
+                                list_entry,
+                                struct ecore_db_recovery_entry) {
+               ecore_db_recovery_ring(p_hwfn, db_entry, db_exec);
+               if (db_exec == DB_REC_ONCE)
+                       break;
+       }
+
+       OSAL_SPIN_UNLOCK(&p_hwfn->db_recovery_info.lock);
+}
+/******************** Doorbell Recovery end ****************/
+
 /* Configurable */
 #define ECORE_MIN_DPIS         (4)     /* The minimal num of DPIs required to
                                         * load the driver. The number was
@@ -49,7 +361,7 @@ static bool qm_lock_init;
                                         */
 
 /* Derived */
-#define ECORE_MIN_PWM_REGION   ((ECORE_WID_SIZE) * (ECORE_MIN_DPIS))
+#define ECORE_MIN_PWM_REGION   (ECORE_WID_SIZE * ECORE_MIN_DPIS)
 
 enum BAR_ID {
        BAR_ID_0,               /* used for GRC */
@@ -80,7 +392,7 @@ static u32 ecore_hw_bar_size(struct ecore_hwfn *p_hwfn,
         * they were found to be useful MFW started updating them from 8.7.7.0.
         * In older MFW versions they are set to 0 which means disabled.
         */
-       if (p_hwfn->p_dev->num_hwfns > 1) {
+       if (ECORE_IS_CMT(p_hwfn->p_dev)) {
                DP_INFO(p_hwfn,
                        "BAR size not configured. Assuming BAR size of 256kB for GRC and 512kB for DB\n");
                val = BAR_ID_0 ? 256 * 1024 : 512 * 1024;
@@ -172,6 +484,9 @@ void ecore_resc_free(struct ecore_dev *p_dev)
                ecore_dmae_info_free(p_hwfn);
                ecore_dcbx_info_free(p_hwfn, p_hwfn->p_dcbx_info);
                /* @@@TBD Flush work-queue ? */
+
+               /* destroy doorbell recovery mechanism */
+               ecore_db_recovery_teardown(p_hwfn);
        }
 }
 
@@ -309,7 +624,7 @@ static void ecore_init_qm_params(struct ecore_hwfn *p_hwfn)
        qm_info->vport_wfq_en = 1;
 
        /* TC config is different for AH 4 port */
-       four_port = p_hwfn->p_dev->num_ports_in_engines == MAX_NUM_PORTS_K2;
+       four_port = p_hwfn->p_dev->num_ports_in_engine == MAX_NUM_PORTS_K2;
 
        /* in AH 4 port we have fewer TCs per port */
        qm_info->max_phys_tcs_per_port = four_port ? NUM_PHYS_TCS_4PORT_K2 :
@@ -338,7 +653,7 @@ static void ecore_init_qm_vport_params(struct ecore_hwfn *p_hwfn)
 static void ecore_init_qm_port_params(struct ecore_hwfn *p_hwfn)
 {
        /* Initialize qm port parameters */
-       u8 i, active_phys_tcs, num_ports = p_hwfn->p_dev->num_ports_in_engines;
+       u8 i, active_phys_tcs, num_ports = p_hwfn->p_dev->num_ports_in_engine;
 
        /* indicate how ooo and high pri traffic is dealt with */
        active_phys_tcs = num_ports == MAX_NUM_PORTS_K2 ?
@@ -350,7 +665,7 @@ static void ecore_init_qm_port_params(struct ecore_hwfn *p_hwfn)
 
                p_qm_port->active = 1;
                p_qm_port->active_phys_tcs = active_phys_tcs;
-               p_qm_port->num_pbf_cmd_lines = PBF_MAX_CMD_LINES / num_ports;
+               p_qm_port->num_pbf_cmd_lines = PBF_MAX_CMD_LINES_E4 / num_ports;
                p_qm_port->num_btb_blocks = BTB_MAX_BLOCKS / num_ports;
        }
 }
@@ -692,7 +1007,7 @@ static void ecore_dp_init_qm_params(struct ecore_hwfn *p_hwfn)
                   qm_info->num_pf_rls, ecore_get_pq_flags(p_hwfn));
 
        /* port table */
-       for (i = 0; i < p_hwfn->p_dev->num_ports_in_engines; i++) {
+       for (i = 0; i < p_hwfn->p_dev->num_ports_in_engine; i++) {
                port = &qm_info->qm_port_params[i];
                DP_VERBOSE(p_hwfn, ECORE_MSG_HW,
                           "port idx %d, active %d, active_phys_tcs %d,"
@@ -821,7 +1136,7 @@ static enum _ecore_status_t ecore_alloc_qm_data(struct ecore_hwfn *p_hwfn)
 
        qm_info->qm_port_params = OSAL_ZALLOC(p_hwfn->p_dev, GFP_KERNEL,
                                      sizeof(struct init_qm_port_params) *
-                                     p_hwfn->p_dev->num_ports_in_engines);
+                                     p_hwfn->p_dev->num_ports_in_engine);
        if (!qm_info->qm_port_params)
                goto alloc_err;
 
@@ -863,12 +1178,17 @@ enum _ecore_status_t ecore_resc_alloc(struct ecore_dev *p_dev)
                struct ecore_hwfn *p_hwfn = &p_dev->hwfns[i];
                u32 n_eqes, num_cons;
 
+               /* initialize the doorbell recovery mechanism */
+               rc = ecore_db_recovery_setup(p_hwfn);
+               if (rc)
+                       goto alloc_err;
+
                /* First allocate the context manager structure */
                rc = ecore_cxt_mngr_alloc(p_hwfn);
                if (rc)
                        goto alloc_err;
 
-               /* Set the HW cid/tid numbers (in the contest manager)
+               /* Set the HW cid/tid numbers (in the context manager)
                 * Must be done prior to any further computations.
                 */
                rc = ecore_cxt_set_pf_params(p_hwfn);
@@ -1118,7 +1438,7 @@ static enum _ecore_status_t ecore_calc_hw_mode(struct ecore_hwfn *p_hwfn)
        }
 
        /* Ports per engine is based on the values in CNIG_REG_NW_PORT_MODE */
-       switch (p_hwfn->p_dev->num_ports_in_engines) {
+       switch (p_hwfn->p_dev->num_ports_in_engine) {
        case 1:
                hw_mode |= 1 << MODE_PORTS_PER_ENG_1;
                break;
@@ -1131,23 +1451,15 @@ static enum _ecore_status_t ecore_calc_hw_mode(struct ecore_hwfn *p_hwfn)
        default:
                DP_NOTICE(p_hwfn, true,
                          "num_ports_in_engine = %d not supported\n",
-                         p_hwfn->p_dev->num_ports_in_engines);
+                         p_hwfn->p_dev->num_ports_in_engine);
                return ECORE_INVAL;
        }
 
-       switch (p_hwfn->p_dev->mf_mode) {
-       case ECORE_MF_DEFAULT:
-       case ECORE_MF_NPAR:
-               hw_mode |= 1 << MODE_MF_SI;
-               break;
-       case ECORE_MF_OVLAN:
+       if (OSAL_TEST_BIT(ECORE_MF_OVLAN_CLSS,
+                         &p_hwfn->p_dev->mf_bits))
                hw_mode |= 1 << MODE_MF_SD;
-               break;
-       default:
-               DP_NOTICE(p_hwfn, true,
-                         "Unsupported MF mode, init as DEFAULT\n");
+       else
                hw_mode |= 1 << MODE_MF_SI;
-       }
 
 #ifndef ASIC_ONLY
        if (CHIP_REV_IS_SLOW(p_hwfn->p_dev)) {
@@ -1163,7 +1475,7 @@ static enum _ecore_status_t ecore_calc_hw_mode(struct ecore_hwfn *p_hwfn)
 #endif
                hw_mode |= 1 << MODE_ASIC;
 
-       if (p_hwfn->p_dev->num_hwfns > 1)
+       if (ECORE_IS_CMT(p_hwfn->p_dev))
                hw_mode |= 1 << MODE_100G;
 
        p_hwfn->hw_info.hw_mode = hw_mode;
@@ -1205,10 +1517,10 @@ static enum _ecore_status_t ecore_hw_init_chip(struct ecore_hwfn *p_hwfn,
                if (ECORE_IS_AH(p_dev)) {
                        /* 2 for 4-port, 1 for 2-port, 0 for 1-port */
                        ecore_wr(p_hwfn, p_ptt, MISC_REG_PORT_MODE,
-                                (p_dev->num_ports_in_engines >> 1));
+                                (p_dev->num_ports_in_engine >> 1));
 
                        ecore_wr(p_hwfn, p_ptt, MISC_REG_BLOCK_256B_EN,
-                                p_dev->num_ports_in_engines == 4 ? 0 : 3);
+                                p_dev->num_ports_in_engine == 4 ? 0 : 3);
                }
        }
 
@@ -1347,7 +1659,7 @@ static enum _ecore_status_t ecore_hw_init_common(struct ecore_hwfn *p_hwfn,
        }
 
        ecore_qm_common_rt_init(p_hwfn,
-                               p_dev->num_ports_in_engines,
+                               p_dev->num_ports_in_engine,
                                qm_info->max_phys_tcs_per_port,
                                qm_info->pf_rl_en, qm_info->pf_wfq_en,
                                qm_info->vport_rl_en, qm_info->vport_wfq_en,
@@ -1572,9 +1884,9 @@ static enum _ecore_status_t
 ecore_hw_init_dpi_size(struct ecore_hwfn *p_hwfn,
                       struct ecore_ptt *p_ptt, u32 pwm_region_size, u32 n_cpus)
 {
-       u32 dpi_page_size_1, dpi_page_size_2, dpi_page_size;
-       u32 dpi_bit_shift, dpi_count;
+       u32 dpi_bit_shift, dpi_count, dpi_page_size;
        u32 min_dpis;
+       u32 n_wids;
 
        /* Calculate DPI size
         * ------------------
@@ -1597,12 +1909,11 @@ ecore_hw_init_dpi_size(struct ecore_hwfn *p_hwfn,
         * 0 is 4kB, 1 is 8kB and etc. Hence the minimum size is 4,096
         * containing 4 WIDs.
         */
-       dpi_page_size_1 = ECORE_WID_SIZE * n_cpus;
-       dpi_page_size_2 = OSAL_MAX_T(u32, ECORE_WID_SIZE, OSAL_PAGE_SIZE);
-       dpi_page_size = OSAL_MAX_T(u32, dpi_page_size_1, dpi_page_size_2);
-       dpi_page_size = OSAL_ROUNDUP_POW_OF_TWO(dpi_page_size);
+       n_wids = OSAL_MAX_T(u32, ECORE_MIN_WIDS, n_cpus);
+       dpi_page_size = ECORE_WID_SIZE * OSAL_ROUNDUP_POW_OF_TWO(n_wids);
+       dpi_page_size = (dpi_page_size + OSAL_PAGE_SIZE - 1) &
+                       ~(OSAL_PAGE_SIZE - 1);
        dpi_bit_shift = OSAL_LOG2(dpi_page_size / 4096);
-
        dpi_count = pwm_region_size / dpi_page_size;
 
        min_dpis = p_hwfn->pf_params.rdma_pf_params.min_dpis;
@@ -1640,7 +1951,7 @@ ecore_hw_init_pf_doorbell_bar(struct ecore_hwfn *p_hwfn,
        u8 cond;
 
        db_bar_size = ecore_hw_bar_size(p_hwfn, p_ptt, BAR_ID_1);
-       if (p_hwfn->p_dev->num_hwfns > 1)
+       if (ECORE_IS_CMT(p_hwfn->p_dev))
                db_bar_size /= 2;
 
        /* Calculate doorbell regions
@@ -1661,7 +1972,8 @@ ecore_hw_init_pf_doorbell_bar(struct ecore_hwfn *p_hwfn,
            ecore_cxt_get_proto_cid_count(p_hwfn, PROTOCOLID_CORE,
                                          OSAL_NULL) +
            ecore_cxt_get_proto_cid_count(p_hwfn, PROTOCOLID_ETH, OSAL_NULL);
-       norm_regsize = ROUNDUP(ECORE_PF_DEMS_SIZE * non_pwm_conn, 4096);
+       norm_regsize = ROUNDUP(ECORE_PF_DEMS_SIZE * non_pwm_conn,
+                              OSAL_PAGE_SIZE);
        min_addr_reg1 = norm_regsize / 4096;
        pwm_regsize = db_bar_size - norm_regsize;
 
@@ -1739,7 +2051,21 @@ static enum _ecore_status_t ecore_hw_init_port(struct ecore_hwfn *p_hwfn,
                                               struct ecore_ptt *p_ptt,
                                               int hw_mode)
 {
+       u32 ppf_to_eng_sel[NIG_REG_PPF_TO_ENGINE_SEL_RT_SIZE];
+       u32 val;
        enum _ecore_status_t rc = ECORE_SUCCESS;
+       u8 i;
+
+       /* In CMT for non-RoCE packets - use connection based classification */
+       val = ECORE_IS_CMT(p_hwfn->p_dev) ? 0x8 : 0x0;
+       for (i = 0; i < NIG_REG_PPF_TO_ENGINE_SEL_RT_SIZE; i++)
+               ppf_to_eng_sel[i] = val;
+       STORE_RT_REG_AGG(p_hwfn, NIG_REG_PPF_TO_ENGINE_SEL_RT_OFFSET,
+                        ppf_to_eng_sel);
+
+       /* In CMT the gate should be cleared by the 2nd hwfn */
+       if (!ECORE_IS_CMT(p_hwfn->p_dev) || !IS_LEAD_HWFN(p_hwfn))
+               STORE_RT_REG(p_hwfn, NIG_REG_BRB_GATE_DNTFWD_PORT_RT_OFFSET, 0);
 
        rc = ecore_init_run(p_hwfn, p_ptt, PHASE_PORT, p_hwfn->port_id,
                            hw_mode);
@@ -1758,7 +2084,7 @@ static enum _ecore_status_t ecore_hw_init_port(struct ecore_hwfn *p_hwfn,
                else if (ECORE_IS_BB(p_hwfn->p_dev))
                        ecore_link_init_bb(p_hwfn, p_ptt, p_hwfn->port_id);
        } else if (CHIP_REV_IS_EMUL(p_hwfn->p_dev)) {
-               if (p_hwfn->p_dev->num_hwfns > 1) {
+               if (ECORE_IS_CMT(p_hwfn->p_dev)) {
                        /* Activate OPTE in CMT */
                        u32 val;
 
@@ -1820,6 +2146,11 @@ ecore_hw_init_pf(struct ecore_hwfn *p_hwfn,
                STORE_RT_REG(p_hwfn, NIG_REG_LLH_FUNC_TAG_EN_RT_OFFSET, 1);
                STORE_RT_REG(p_hwfn, NIG_REG_LLH_FUNC_TAG_VALUE_RT_OFFSET,
                             p_hwfn->hw_info.ovlan);
+
+               DP_VERBOSE(p_hwfn, ECORE_MSG_HW,
+                          "Configuring LLH_FUNC_FILTER_HDR_SEL\n");
+               STORE_RT_REG(p_hwfn, NIG_REG_LLH_FUNC_FILTER_HDR_SEL_RT_OFFSET,
+                            1);
        }
 
        /* Enable classification by MAC if needed */
@@ -1880,7 +2211,6 @@ ecore_hw_init_pf(struct ecore_hwfn *p_hwfn,
 
                /* send function start command */
                rc = ecore_sp_pf_start(p_hwfn, p_ptt, p_tunn,
-                                      p_hwfn->p_dev->mf_mode,
                                       allow_npar_tx_switch);
                if (rc) {
                        DP_NOTICE(p_hwfn, true,
@@ -2017,8 +2347,7 @@ enum _ecore_status_t ecore_hw_init(struct ecore_dev *p_dev,
        enum _ecore_status_t rc = ECORE_SUCCESS;
        int i;
 
-       if ((p_params->int_mode == ECORE_INT_MODE_MSI) &&
-           (p_dev->num_hwfns > 1)) {
+       if ((p_params->int_mode == ECORE_INT_MODE_MSI) && ECORE_IS_CMT(p_dev)) {
                DP_NOTICE(p_dev, false,
                          "MSI mode is not supported for CMT devices\n");
                return ECORE_INVAL;
@@ -2159,7 +2488,7 @@ enum _ecore_status_t ecore_hw_init(struct ecore_dev *p_dev,
                           "sending phony dcbx set command to trigger DCBx attention handling\n");
                rc = ecore_mcp_cmd(p_hwfn, p_hwfn->p_main_ptt,
                                   DRV_MSG_CODE_SET_DCBX,
-                                  1 << DRV_MB_PARAM_DCBX_NOTIFY_SHIFT, &resp,
+                                  1 << DRV_MB_PARAM_DCBX_NOTIFY_OFFSET, &resp,
                                   &param);
                if (rc != ECORE_SUCCESS) {
                        DP_NOTICE(p_hwfn, true,
@@ -2836,9 +3165,6 @@ static enum _ecore_status_t ecore_hw_set_resc_info(struct ecore_hwfn *p_hwfn,
        return ECORE_SUCCESS;
 }
 
-#define ECORE_RESC_ALLOC_LOCK_RETRY_CNT                10
-#define ECORE_RESC_ALLOC_LOCK_RETRY_INTVL_US   10000 /* 10 msec */
-
 static enum _ecore_status_t ecore_hw_get_resc(struct ecore_hwfn *p_hwfn,
                                              struct ecore_ptt *p_ptt,
                                              bool drv_resc_alloc)
@@ -2870,13 +3196,8 @@ static enum _ecore_status_t ecore_hw_get_resc(struct ecore_hwfn *p_hwfn,
         * Old drivers that don't acquire the lock can run in parallel, and
         * their allocation values won't be affected by the updated max values.
         */
-       OSAL_MEM_ZERO(&resc_lock_params, sizeof(resc_lock_params));
-       resc_lock_params.resource = ECORE_RESC_LOCK_RESC_ALLOC;
-       resc_lock_params.retry_num = ECORE_RESC_ALLOC_LOCK_RETRY_CNT;
-       resc_lock_params.retry_interval = ECORE_RESC_ALLOC_LOCK_RETRY_INTVL_US;
-       resc_lock_params.sleep_b4_retry = true;
-       OSAL_MEM_ZERO(&resc_unlock_params, sizeof(resc_unlock_params));
-       resc_unlock_params.resource = ECORE_RESC_LOCK_RESC_ALLOC;
+       ecore_mcp_resc_lock_default_init(&resc_lock_params, &resc_unlock_params,
+                                        ECORE_RESC_LOCK_RESC_ALLOC, false);
 
        rc = ecore_mcp_resc_lock(p_hwfn, p_ptt, &resc_lock_params);
        if (rc != ECORE_SUCCESS && rc != ECORE_NOTIMPL) {
@@ -3178,6 +3499,37 @@ ecore_hw_get_nvm_info(struct ecore_hwfn *p_hwfn,
            NVM_CFG1_GLOB_MF_MODE_OFFSET;
 
        switch (mf_mode) {
+       case NVM_CFG1_GLOB_MF_MODE_MF_ALLOWED:
+               p_hwfn->p_dev->mf_bits = 1 << ECORE_MF_OVLAN_CLSS;
+               break;
+       case NVM_CFG1_GLOB_MF_MODE_UFP:
+               p_hwfn->p_dev->mf_bits = 1 << ECORE_MF_OVLAN_CLSS |
+                                        1 << ECORE_MF_UFP_SPECIFIC;
+               break;
+
+       case NVM_CFG1_GLOB_MF_MODE_NPAR1_0:
+               p_hwfn->p_dev->mf_bits = 1 << ECORE_MF_LLH_MAC_CLSS |
+                                        1 << ECORE_MF_LLH_PROTO_CLSS |
+                                        1 << ECORE_MF_LL2_NON_UNICAST |
+                                        1 << ECORE_MF_INTER_PF_SWITCH;
+               break;
+       case NVM_CFG1_GLOB_MF_MODE_DEFAULT:
+               p_hwfn->p_dev->mf_bits = 1 << ECORE_MF_LLH_MAC_CLSS |
+                                        1 << ECORE_MF_LLH_PROTO_CLSS |
+                                        1 << ECORE_MF_LL2_NON_UNICAST;
+               if (ECORE_IS_BB(p_hwfn->p_dev))
+                       p_hwfn->p_dev->mf_bits |= 1 << ECORE_MF_NEED_DEF_PF;
+               break;
+       }
+       DP_INFO(p_hwfn, "Multi function mode is 0x%lx\n",
+               p_hwfn->p_dev->mf_bits);
+
+       /* It's funny since we have another switch, but it's easier
+        * to throw this away in linux this way. Long term, it might be
+        * better to have have getters for needed ECORE_MF_* fields,
+        * convert client code and eliminate this.
+        */
+       switch (mf_mode) {
        case NVM_CFG1_GLOB_MF_MODE_MF_ALLOWED:
                p_hwfn->p_dev->mf_mode = ECORE_MF_OVLAN;
                break;
@@ -3187,9 +3539,10 @@ ecore_hw_get_nvm_info(struct ecore_hwfn *p_hwfn,
        case NVM_CFG1_GLOB_MF_MODE_DEFAULT:
                p_hwfn->p_dev->mf_mode = ECORE_MF_DEFAULT;
                break;
+       case NVM_CFG1_GLOB_MF_MODE_UFP:
+               p_hwfn->p_dev->mf_mode = ECORE_MF_UFP;
+               break;
        }
-       DP_INFO(p_hwfn, "Multi function mode is %08x\n",
-               p_hwfn->p_dev->mf_mode);
 
        /* Read Multi-function information from shmem */
        addr = MCP_REG_SCRATCH + nvm_cfg1_offset +
@@ -3248,7 +3601,7 @@ static void ecore_get_num_funcs(struct ecore_hwfn *p_hwfn,
 
        if (reg_function_hide & 0x1) {
                if (ECORE_IS_BB(p_dev)) {
-                       if (ECORE_PATH_ID(p_hwfn) && p_dev->num_hwfns == 1) {
+                       if (ECORE_PATH_ID(p_hwfn) && !ECORE_IS_CMT(p_dev)) {
                                num_funcs = 0;
                                eng_mask = 0xaaaa;
                        } else {
@@ -3298,14 +3651,14 @@ static void ecore_get_num_funcs(struct ecore_hwfn *p_hwfn,
 static void ecore_hw_info_port_num_bb(struct ecore_hwfn *p_hwfn,
                                      struct ecore_ptt *p_ptt)
 {
+       struct ecore_dev *p_dev = p_hwfn->p_dev;
        u32 port_mode;
 
 #ifndef ASIC_ONLY
        /* Read the port mode */
-       if (CHIP_REV_IS_FPGA(p_hwfn->p_dev))
+       if (CHIP_REV_IS_FPGA(p_dev))
                port_mode = 4;
-       else if (CHIP_REV_IS_EMUL(p_hwfn->p_dev) &&
-                (p_hwfn->p_dev->num_hwfns > 1))
+       else if (CHIP_REV_IS_EMUL(p_dev) && ECORE_IS_CMT(p_dev))
                /* In CMT on emulation, assume 1 port */
                port_mode = 1;
        else
@@ -3313,38 +3666,39 @@ static void ecore_hw_info_port_num_bb(struct ecore_hwfn *p_hwfn,
        port_mode = ecore_rd(p_hwfn, p_ptt, CNIG_REG_NW_PORT_MODE_BB);
 
        if (port_mode < 3) {
-               p_hwfn->p_dev->num_ports_in_engines = 1;
+               p_dev->num_ports_in_engine = 1;
        } else if (port_mode <= 5) {
-               p_hwfn->p_dev->num_ports_in_engines = 2;
+               p_dev->num_ports_in_engine = 2;
        } else {
                DP_NOTICE(p_hwfn, true, "PORT MODE: %d not supported\n",
-                         p_hwfn->p_dev->num_ports_in_engines);
+                         p_dev->num_ports_in_engine);
 
-               /* Default num_ports_in_engines to something */
-               p_hwfn->p_dev->num_ports_in_engines = 1;
+               /* Default num_ports_in_engine to something */
+               p_dev->num_ports_in_engine = 1;
        }
 }
 
 static void ecore_hw_info_port_num_ah_e5(struct ecore_hwfn *p_hwfn,
                                         struct ecore_ptt *p_ptt)
 {
+       struct ecore_dev *p_dev = p_hwfn->p_dev;
        u32 port;
        int i;
 
-       p_hwfn->p_dev->num_ports_in_engines = 0;
+       p_dev->num_ports_in_engine = 0;
 
 #ifndef ASIC_ONLY
-       if (CHIP_REV_IS_EMUL(p_hwfn->p_dev)) {
+       if (CHIP_REV_IS_EMUL(p_dev)) {
                port = ecore_rd(p_hwfn, p_ptt, MISCS_REG_ECO_RESERVED);
                switch ((port & 0xf000) >> 12) {
                case 1:
-                       p_hwfn->p_dev->num_ports_in_engines = 1;
+                       p_dev->num_ports_in_engine = 1;
                        break;
                case 3:
-                       p_hwfn->p_dev->num_ports_in_engines = 2;
+                       p_dev->num_ports_in_engine = 2;
                        break;
                case 0xf:
-                       p_hwfn->p_dev->num_ports_in_engines = 4;
+                       p_dev->num_ports_in_engine = 4;
                        break;
                default:
                        DP_NOTICE(p_hwfn, false,
@@ -3358,17 +3712,47 @@ static void ecore_hw_info_port_num_ah_e5(struct ecore_hwfn *p_hwfn,
                                        CNIG_REG_NIG_PORT0_CONF_K2_E5 +
                                        (i * 4));
                        if (port & 1)
-                               p_hwfn->p_dev->num_ports_in_engines++;
+                               p_dev->num_ports_in_engine++;
                }
+
+       if (!p_dev->num_ports_in_engine) {
+               DP_NOTICE(p_hwfn, true, "All NIG ports are inactive\n");
+
+               /* Default num_ports_in_engine to something */
+               p_dev->num_ports_in_engine = 1;
+       }
 }
 
 static void ecore_hw_info_port_num(struct ecore_hwfn *p_hwfn,
                                   struct ecore_ptt *p_ptt)
 {
-       if (ECORE_IS_BB(p_hwfn->p_dev))
+       struct ecore_dev *p_dev = p_hwfn->p_dev;
+
+       /* Determine the number of ports per engine */
+       if (ECORE_IS_BB(p_dev))
                ecore_hw_info_port_num_bb(p_hwfn, p_ptt);
        else
                ecore_hw_info_port_num_ah_e5(p_hwfn, p_ptt);
+
+       /* Get the total number of ports of the device */
+       if (ECORE_IS_CMT(p_dev)) {
+               /* In CMT there is always only one port */
+               p_dev->num_ports = 1;
+#ifndef ASIC_ONLY
+       } else if (CHIP_REV_IS_EMUL(p_dev) || CHIP_REV_IS_TEDIBEAR(p_dev)) {
+               p_dev->num_ports = p_dev->num_ports_in_engine *
+                                  ecore_device_num_engines(p_dev);
+#endif
+       } else {
+               u32 addr, global_offsize, global_addr;
+
+               addr = SECTION_OFFSIZE_ADDR(p_hwfn->mcp_info->public_base,
+                                           PUBLIC_GLOBAL);
+               global_offsize = ecore_rd(p_hwfn, p_ptt, addr);
+               global_addr = SECTION_ADDR(global_offsize, 0);
+               addr = global_addr + OFFSETOF(struct public_global, max_ports);
+               p_dev->num_ports = (u8)ecore_rd(p_hwfn, p_ptt, addr);
+       }
 }
 
 static void ecore_mcp_get_eee_caps(struct ecore_hwfn *p_hwfn,
@@ -3412,14 +3796,8 @@ ecore_get_hw_info(struct ecore_hwfn *p_hwfn, struct ecore_ptt *p_ptt,
                }
        }
 
-       /* TODO In get_hw_info, amoungst others:
-        * Get MCP FW revision and determine according to it the supported
-        * featrues (e.g. DCB)
-        * Get boot mode
-        * ecore_get_pcie_width_speed, WOL capability.
-        * Number of global CQ-s (for storage
-        */
-       ecore_hw_info_port_num(p_hwfn, p_ptt);
+       if (IS_LEAD_HWFN(p_hwfn))
+               ecore_hw_info_port_num(p_hwfn, p_ptt);
 
        ecore_mcp_get_capabilities(p_hwfn, p_ptt);
 
@@ -3463,6 +3841,8 @@ ecore_get_hw_info(struct ecore_hwfn *p_hwfn, struct ecore_ptt *p_ptt,
                ecore_mcp_cmd_port_init(p_hwfn, p_ptt);
 
                ecore_mcp_get_eee_caps(p_hwfn, p_ptt);
+
+               ecore_mcp_read_ufp_config(p_hwfn, p_ptt);
        }
 
        if (personality != ECORE_PCI_DEFAULT) {
@@ -3545,12 +3925,10 @@ static enum _ecore_status_t ecore_get_dev_info(struct ecore_hwfn *p_hwfn,
                return ECORE_ABORTED;
        }
 
-       p_dev->chip_num = (u16)ecore_rd(p_hwfn, p_ptt,
-                                               MISCS_REG_CHIP_NUM);
-       p_dev->chip_rev = (u16)ecore_rd(p_hwfn, p_ptt,
-                                               MISCS_REG_CHIP_REV);
-
-       MASK_FIELD(CHIP_REV, p_dev->chip_rev);
+       tmp = ecore_rd(p_hwfn, p_ptt, MISCS_REG_CHIP_NUM);
+       p_dev->chip_num = (u16)GET_FIELD(tmp, CHIP_NUM);
+       tmp = ecore_rd(p_hwfn, p_ptt, MISCS_REG_CHIP_REV);
+       p_dev->chip_rev = (u8)GET_FIELD(tmp, CHIP_REV);
 
        /* Learn number of HW-functions */
        tmp = ecore_rd(p_hwfn, p_ptt, MISCS_REG_CMT_ENABLED_FOR_PAIR);
@@ -3573,20 +3951,19 @@ static enum _ecore_status_t ecore_get_dev_info(struct ecore_hwfn *p_hwfn,
        }
 #endif
 
-       p_dev->chip_bond_id = ecore_rd(p_hwfn, p_ptt,
-                                      MISCS_REG_CHIP_TEST_REG) >> 4;
-       MASK_FIELD(CHIP_BOND_ID, p_dev->chip_bond_id);
-       p_dev->chip_metal = (u16)ecore_rd(p_hwfn, p_ptt,
-                                          MISCS_REG_CHIP_METAL);
-       MASK_FIELD(CHIP_METAL, p_dev->chip_metal);
+       tmp = ecore_rd(p_hwfn, p_ptt, MISCS_REG_CHIP_TEST_REG);
+       p_dev->chip_bond_id = (u8)GET_FIELD(tmp, CHIP_BOND_ID);
+       tmp = ecore_rd(p_hwfn, p_ptt, MISCS_REG_CHIP_METAL);
+       p_dev->chip_metal = (u8)GET_FIELD(tmp, CHIP_METAL);
+
        DP_INFO(p_dev->hwfns,
-               "Chip details - %s %c%d, Num: %04x Rev: %04x Bond id: %04x Metal: %04x\n",
+               "Chip details - %s %c%d, Num: %04x Rev: %02x Bond id: %02x Metal: %02x\n",
                ECORE_IS_BB(p_dev) ? "BB" : "AH",
                'A' + p_dev->chip_rev, (int)p_dev->chip_metal,
                p_dev->chip_num, p_dev->chip_rev, p_dev->chip_bond_id,
                p_dev->chip_metal);
 
-       if (ECORE_IS_BB(p_dev) && CHIP_REV_IS_A0(p_dev)) {
+       if (ECORE_IS_BB_A0(p_dev)) {
                DP_NOTICE(p_dev->hwfns, false,
                          "The chip type/rev (BB A0) is not supported!\n");
                return ECORE_ABORTED;
@@ -3626,7 +4003,6 @@ void ecore_prepare_hibernate(struct ecore_dev *p_dev)
                           "Mark hw/fw uninitialized\n");
 
                p_hwfn->hw_init_done = false;
-               p_hwfn->first_on_engine = false;
 
                ecore_ptt_invalidate(p_hwfn);
        }
@@ -3799,7 +4175,7 @@ enum _ecore_status_t ecore_hw_prepare(struct ecore_dev *p_dev,
        p_params->personality = p_hwfn->hw_info.personality;
 
        /* initilalize 2nd hwfn if necessary */
-       if (p_dev->num_hwfns > 1) {
+       if (ECORE_IS_CMT(p_dev)) {
                void OSAL_IOMEM *p_regview, *p_doorbell;
                u8 OSAL_IOMEM *addr;
 
@@ -4263,7 +4639,8 @@ enum _ecore_status_t ecore_llh_add_mac_filter(struct ecore_hwfn *p_hwfn,
        u32 high, low, entry_num;
        enum _ecore_status_t rc;
 
-       if (!(IS_MF_SI(p_hwfn) || IS_MF_DEFAULT(p_hwfn)))
+       if (!OSAL_TEST_BIT(ECORE_MF_LLH_MAC_CLSS,
+                          &p_hwfn->p_dev->mf_bits))
                return ECORE_SUCCESS;
 
        high = p_filter[1] | (p_filter[0] << 8);
@@ -4330,7 +4707,8 @@ void ecore_llh_remove_mac_filter(struct ecore_hwfn *p_hwfn,
        u32 high, low, entry_num;
        enum _ecore_status_t rc;
 
-       if (!(IS_MF_SI(p_hwfn) || IS_MF_DEFAULT(p_hwfn)))
+       if (!OSAL_TEST_BIT(ECORE_MF_LLH_MAC_CLSS,
+                          &p_hwfn->p_dev->mf_bits))
                return;
 
        high = p_filter[1] | (p_filter[0] << 8);
@@ -4404,7 +4782,8 @@ ecore_llh_add_protocol_filter(struct ecore_hwfn *p_hwfn,
        u32 high, low, entry_num;
        enum _ecore_status_t rc;
 
-       if (!(IS_MF_SI(p_hwfn) || IS_MF_DEFAULT(p_hwfn)))
+       if (!OSAL_TEST_BIT(ECORE_MF_LLH_PROTO_CLSS,
+                          &p_hwfn->p_dev->mf_bits))
                return ECORE_SUCCESS;
 
        high = 0;
@@ -4547,7 +4926,8 @@ ecore_llh_remove_protocol_filter(struct ecore_hwfn *p_hwfn,
        u32 high, low, entry_num;
        enum _ecore_status_t rc;
 
-       if (!(IS_MF_SI(p_hwfn) || IS_MF_DEFAULT(p_hwfn)))
+       if (!OSAL_TEST_BIT(ECORE_MF_LLH_PROTO_CLSS,
+                          &p_hwfn->p_dev->mf_bits))
                return;
 
        high = 0;
@@ -4615,7 +4995,10 @@ static void ecore_llh_clear_all_filters_bb_ah(struct ecore_hwfn *p_hwfn,
 void ecore_llh_clear_all_filters(struct ecore_hwfn *p_hwfn,
                             struct ecore_ptt *p_ptt)
 {
-       if (!(IS_MF_SI(p_hwfn) || IS_MF_DEFAULT(p_hwfn)))
+       if (!OSAL_TEST_BIT(ECORE_MF_LLH_PROTO_CLSS,
+                          &p_hwfn->p_dev->mf_bits) &&
+           !OSAL_TEST_BIT(ECORE_MF_LLH_MAC_CLSS,
+                          &p_hwfn->p_dev->mf_bits))
                return;
 
        if (ECORE_IS_BB(p_hwfn->p_dev) || ECORE_IS_AH(p_hwfn->p_dev))
@@ -4626,7 +5009,7 @@ enum _ecore_status_t
 ecore_llh_set_function_as_default(struct ecore_hwfn *p_hwfn,
                                  struct ecore_ptt *p_ptt)
 {
-       if (IS_MF_DEFAULT(p_hwfn) && ECORE_IS_BB(p_hwfn->p_dev)) {
+       if (OSAL_TEST_BIT(ECORE_MF_NEED_DEF_PF, &p_hwfn->p_dev->mf_bits)) {
                ecore_wr(p_hwfn, p_ptt,
                         NIG_REG_LLH_TAGMAC_DEF_PF_VECTOR,
                         1 << p_hwfn->abs_pf_id / 2);
@@ -4804,8 +5187,7 @@ static void ecore_configure_wfq_for_all_vports(struct ecore_hwfn *p_hwfn,
        }
 }
 
-static void
-ecore_init_wfq_default_param(struct ecore_hwfn *p_hwfn, u32 min_pf_rate)
+static void ecore_init_wfq_default_param(struct ecore_hwfn *p_hwfn)
 {
        int i;
 
@@ -4814,8 +5196,7 @@ ecore_init_wfq_default_param(struct ecore_hwfn *p_hwfn, u32 min_pf_rate)
 }
 
 static void ecore_disable_wfq_for_all_vports(struct ecore_hwfn *p_hwfn,
-                                            struct ecore_ptt *p_ptt,
-                                            u32 min_pf_rate)
+                                            struct ecore_ptt *p_ptt)
 {
        struct init_qm_vport_params *vport_params;
        int i;
@@ -4823,7 +5204,7 @@ static void ecore_disable_wfq_for_all_vports(struct ecore_hwfn *p_hwfn,
        vport_params = p_hwfn->qm_info.qm_vport_params;
 
        for (i = 0; i < p_hwfn->qm_info.num_vports; i++) {
-               ecore_init_wfq_default_param(p_hwfn, min_pf_rate);
+               ecore_init_wfq_default_param(p_hwfn);
                ecore_init_vport_wfq(p_hwfn, p_ptt,
                                     vport_params[i].first_tx_pq_id,
                                     vport_params[i].vport_wfq);
@@ -4977,7 +5358,7 @@ static int __ecore_configure_vp_wfq_on_link_change(struct ecore_hwfn *p_hwfn,
        if (rc == ECORE_SUCCESS && use_wfq)
                ecore_configure_wfq_for_all_vports(p_hwfn, p_ptt, min_pf_rate);
        else
-               ecore_disable_wfq_for_all_vports(p_hwfn, p_ptt, min_pf_rate);
+               ecore_disable_wfq_for_all_vports(p_hwfn, p_ptt);
 
        return rc;
 }
@@ -4991,7 +5372,7 @@ int ecore_configure_vport_wfq(struct ecore_dev *p_dev, u16 vp_id, u32 rate)
        int i, rc = ECORE_INVAL;
 
        /* TBD - for multiple hardware functions - that is 100 gig */
-       if (p_dev->num_hwfns > 1) {
+       if (ECORE_IS_CMT(p_dev)) {
                DP_NOTICE(p_dev, false,
                          "WFQ configuration is not supported for this device\n");
                return rc;
@@ -5026,7 +5407,7 @@ void ecore_configure_vp_wfq_on_link_change(struct ecore_dev *p_dev,
        int i;
 
        /* TBD - for multiple hardware functions - that is 100 gig */
-       if (p_dev->num_hwfns > 1) {
+       if (ECORE_IS_CMT(p_dev)) {
                DP_VERBOSE(p_dev, ECORE_MSG_LINK,
                           "WFQ configuration is not supported for this device\n");
                return;
@@ -5180,8 +5561,7 @@ void ecore_clean_wfq_db(struct ecore_hwfn *p_hwfn, struct ecore_ptt *p_ptt)
        p_link = &p_hwfn->mcp_info->link_output;
 
        if (p_link->min_pf_rate)
-               ecore_disable_wfq_for_all_vports(p_hwfn, p_ptt,
-                                                p_link->min_pf_rate);
+               ecore_disable_wfq_for_all_vports(p_hwfn, p_ptt);
 
        OSAL_MEMSET(p_hwfn->qm_info.wfq_data, 0,
                    sizeof(*p_hwfn->qm_info.wfq_data) *
@@ -5195,11 +5575,7 @@ int ecore_device_num_engines(struct ecore_dev *p_dev)
 
 int ecore_device_num_ports(struct ecore_dev *p_dev)
 {
-       /* in CMT always only one port */
-       if (p_dev->num_hwfns > 1)
-               return 1;
-
-       return p_dev->num_ports_in_engines * ecore_device_num_engines(p_dev);
+       return p_dev->num_ports;
 }
 
 void ecore_set_fw_mac_addr(__le16 *fw_msb,