net/qede/base: refine error handling
[dpdk.git] / drivers / net / qede / base / ecore_dev.c
index edf2896..a85d26d 100644 (file)
@@ -28,7 +28,6 @@
 #include "mcp_public.h"
 #include "ecore_iro.h"
 #include "nvm_cfg.h"
-#include "ecore_dev_api.h"
 #include "ecore_dcbx.h"
 #include "ecore_l2.h"
 
@@ -40,7 +39,7 @@
  * there's more than a single compiled ecore component in system].
  */
 static osal_spinlock_t qm_lock;
-static bool qm_lock_init;
+static u32 qm_lock_ref_cnt;
 
 /******************** Doorbell Recovery *******************/
 /* The doorbell recovery mechanism consists of a list of entries which represent
@@ -228,7 +227,8 @@ enum _ecore_status_t ecore_db_recovery_setup(struct ecore_hwfn *p_hwfn)
 
        OSAL_LIST_INIT(&p_hwfn->db_recovery_info.list);
 #ifdef CONFIG_ECORE_LOCK_ALLOC
-       OSAL_SPIN_LOCK_ALLOC(p_hwfn, &p_hwfn->db_recovery_info.lock);
+       if (OSAL_SPIN_LOCK_ALLOC(p_hwfn, &p_hwfn->db_recovery_info.lock))
+               return ECORE_NOMEM;
 #endif
        OSAL_SPIN_LOCK_INIT(&p_hwfn->db_recovery_info.lock);
        p_hwfn->db_recovery_info.db_recovery_counter = 0;
@@ -363,11 +363,6 @@ void ecore_db_recovery_execute(struct ecore_hwfn *p_hwfn,
 /* Derived */
 #define ECORE_MIN_PWM_REGION   (ECORE_WID_SIZE * ECORE_MIN_DPIS)
 
-enum BAR_ID {
-       BAR_ID_0,               /* used for GRC */
-       BAR_ID_1                /* Used for doorbells */
-};
-
 static u32 ecore_hw_bar_size(struct ecore_hwfn *p_hwfn,
                             struct ecore_ptt *p_ptt,
                             enum BAR_ID bar_id)
@@ -376,13 +371,8 @@ static u32 ecore_hw_bar_size(struct ecore_hwfn *p_hwfn,
                       PGLUE_B_REG_PF_BAR0_SIZE : PGLUE_B_REG_PF_BAR1_SIZE);
        u32 val;
 
-       if (IS_VF(p_hwfn->p_dev)) {
-               /* TODO - assume each VF hwfn has 64Kb for Bar0; Bar1 can be
-                * read from actual register, but we're currently not using
-                * it for actual doorbelling.
-                */
-               return 1 << 17;
-       }
+       if (IS_VF(p_hwfn->p_dev))
+               return ecore_vf_hw_bar_size(p_hwfn, bar_id);
 
        val = ecore_rd(p_hwfn, p_ptt, bar_reg);
        if (val)
@@ -422,7 +412,7 @@ void ecore_init_dp(struct ecore_dev *p_dev,
        }
 }
 
-void ecore_init_struct(struct ecore_dev *p_dev)
+enum _ecore_status_t ecore_init_struct(struct ecore_dev *p_dev)
 {
        u8 i;
 
@@ -434,9 +424,10 @@ void ecore_init_struct(struct ecore_dev *p_dev)
                p_hwfn->b_active = false;
 
 #ifdef CONFIG_ECORE_LOCK_ALLOC
-               OSAL_MUTEX_ALLOC(p_hwfn, &p_hwfn->dmae_info.mutex);
+               if (OSAL_SPIN_LOCK_ALLOC(p_hwfn, &p_hwfn->dmae_info.lock))
+                       goto handle_err;
 #endif
-               OSAL_MUTEX_INIT(&p_hwfn->dmae_info.mutex);
+               OSAL_SPIN_LOCK_INIT(&p_hwfn->dmae_info.lock);
        }
 
        /* hwfn 0 is always active */
@@ -444,6 +435,17 @@ void ecore_init_struct(struct ecore_dev *p_dev)
 
        /* set the default cache alignment to 128 (may be overridden later) */
        p_dev->cache_shift = 7;
+       return ECORE_SUCCESS;
+#ifdef CONFIG_ECORE_LOCK_ALLOC
+handle_err:
+       while (--i) {
+               struct ecore_hwfn *p_hwfn = OSAL_NULL;
+
+               p_hwfn = &p_dev->hwfns[i];
+               OSAL_SPIN_LOCK_DEALLOC(&p_hwfn->dmae_info.lock);
+       }
+       return ECORE_NOMEM;
+#endif
 }
 
 static void ecore_qm_info_free(struct ecore_hwfn *p_hwfn)
@@ -482,7 +484,7 @@ void ecore_resc_free(struct ecore_dev *p_dev)
                ecore_iov_free(p_hwfn);
                ecore_l2_free(p_hwfn);
                ecore_dmae_info_free(p_hwfn);
-               ecore_dcbx_info_free(p_hwfn, p_hwfn->p_dcbx_info);
+               ecore_dcbx_info_free(p_hwfn);
                /* @@@TBD Flush work-queue ? */
 
                /* destroy doorbell recovery mechanism */
@@ -732,6 +734,7 @@ static void ecore_init_qm_pq(struct ecore_hwfn *p_hwfn,
                       "pq overflow! pq %d, max pq %d\n", pq_idx, max_pq);
 
        /* init pq params */
+       qm_info->qm_pq_params[pq_idx].port_id = p_hwfn->port_id;
        qm_info->qm_pq_params[pq_idx].vport_id = qm_info->start_vport +
                                                 qm_info->num_vports;
        qm_info->qm_pq_params[pq_idx].tc_id = tc;
@@ -1036,10 +1039,9 @@ static void ecore_dp_init_qm_params(struct ecore_hwfn *p_hwfn)
        for (i = 0; i < qm_info->num_pqs; i++) {
                pq = &qm_info->qm_pq_params[i];
                DP_VERBOSE(p_hwfn, ECORE_MSG_HW,
-                          "pq idx %d, vport_id %d, tc %d, wrr_grp %d,"
-                          " rl_valid %d\n",
-                          qm_info->start_pq + i, pq->vport_id, pq->tc_id,
-                          pq->wrr_group, pq->rl_valid);
+                          "pq idx %d, port %d, vport_id %d, tc %d, wrr_grp %d, rl_valid %d\n",
+                          qm_info->start_pq + i, pq->port_id, pq->vport_id,
+                          pq->tc_id, pq->wrr_group, pq->rl_valid);
        }
 }
 
@@ -1094,7 +1096,7 @@ enum _ecore_status_t ecore_qm_reconf(struct ecore_hwfn *p_hwfn,
        ecore_init_clear_rt_data(p_hwfn);
 
        /* prepare QM portion of runtime array */
-       ecore_qm_init_pf(p_hwfn, p_ptt);
+       ecore_qm_init_pf(p_hwfn, p_ptt, false);
 
        /* activate init tool on runtime array */
        rc = ecore_init_run(p_hwfn, p_ptt, PHASE_QM_PF, p_hwfn->rel_pf_id,
@@ -1300,16 +1302,14 @@ enum _ecore_status_t ecore_resc_alloc(struct ecore_dev *p_dev)
                /* DMA info initialization */
                rc = ecore_dmae_info_alloc(p_hwfn);
                if (rc) {
-                       DP_NOTICE(p_hwfn, true,
-                                 "Failed to allocate memory for dmae_info"
-                                 " structure\n");
+                       DP_NOTICE(p_hwfn, false, "Failed to allocate memory for dmae_info structure\n");
                        goto alloc_err;
                }
 
                /* DCBX initialization */
                rc = ecore_dcbx_info_alloc(p_hwfn);
                if (rc) {
-                       DP_NOTICE(p_hwfn, true,
+                       DP_NOTICE(p_hwfn, false,
                                  "Failed to allocate memory for dcbx structure\n");
                        goto alloc_err;
                }
@@ -1318,7 +1318,7 @@ enum _ecore_status_t ecore_resc_alloc(struct ecore_dev *p_dev)
        p_dev->reset_stats = OSAL_ZALLOC(p_dev, GFP_KERNEL,
                                         sizeof(*p_dev->reset_stats));
        if (!p_dev->reset_stats) {
-               DP_NOTICE(p_dev, true, "Failed to allocate reset statistics\n");
+               DP_NOTICE(p_dev, false, "Failed to allocate reset statistics\n");
                goto alloc_no_mem;
        }
 
@@ -1455,19 +1455,11 @@ static enum _ecore_status_t ecore_calc_hw_mode(struct ecore_hwfn *p_hwfn)
                return ECORE_INVAL;
        }
 
-       switch (p_hwfn->p_dev->mf_mode) {
-       case ECORE_MF_DEFAULT:
-       case ECORE_MF_NPAR:
-               hw_mode |= 1 << MODE_MF_SI;
-               break;
-       case ECORE_MF_OVLAN:
+       if (OSAL_TEST_BIT(ECORE_MF_OVLAN_CLSS,
+                         &p_hwfn->p_dev->mf_bits))
                hw_mode |= 1 << MODE_MF_SD;
-               break;
-       default:
-               DP_NOTICE(p_hwfn, true,
-                         "Unsupported MF mode, init as DEFAULT\n");
+       else
                hw_mode |= 1 << MODE_MF_SI;
-       }
 
 #ifndef ASIC_ONLY
        if (CHIP_REV_IS_SLOW(p_hwfn->p_dev)) {
@@ -1677,7 +1669,8 @@ static enum _ecore_status_t ecore_hw_init_common(struct ecore_hwfn *p_hwfn,
 
        ecore_init_cache_line_size(p_hwfn, p_ptt);
 
-       rc = ecore_init_run(p_hwfn, p_ptt, PHASE_ENGINE, ANY_PHASE_ID, hw_mode);
+       rc = ecore_init_run(p_hwfn, p_ptt, PHASE_ENGINE, ECORE_PATH_ID(p_hwfn),
+                           hw_mode);
        if (rc != ECORE_SUCCESS)
                return rc;
 
@@ -2007,7 +2000,7 @@ ecore_hw_init_pf_doorbell_bar(struct ecore_hwfn *p_hwfn,
                /* Either EDPM is mandatory, or we are attempting to allocate a
                 * WID per CPU.
                 */
-               n_cpus = OSAL_NUM_ACTIVE_CPU();
+               n_cpus = OSAL_NUM_CPUS();
                rc = ecore_hw_init_dpi_size(p_hwfn, p_ptt, pwm_regsize, n_cpus);
        }
 
@@ -2154,6 +2147,11 @@ ecore_hw_init_pf(struct ecore_hwfn *p_hwfn,
                STORE_RT_REG(p_hwfn, NIG_REG_LLH_FUNC_TAG_EN_RT_OFFSET, 1);
                STORE_RT_REG(p_hwfn, NIG_REG_LLH_FUNC_TAG_VALUE_RT_OFFSET,
                             p_hwfn->hw_info.ovlan);
+
+               DP_VERBOSE(p_hwfn, ECORE_MSG_HW,
+                          "Configuring LLH_FUNC_FILTER_HDR_SEL\n");
+               STORE_RT_REG(p_hwfn, NIG_REG_LLH_FUNC_FILTER_HDR_SEL_RT_OFFSET,
+                            1);
        }
 
        /* Enable classification by MAC if needed */
@@ -2174,6 +2172,11 @@ ecore_hw_init_pf(struct ecore_hwfn *p_hwfn,
        /* perform debug configuration when chip is out of reset */
        OSAL_BEFORE_PF_START((void *)p_hwfn->p_dev, p_hwfn->my_id);
 
+       /* Sanity check before the PF init sequence that uses DMAE */
+       rc = ecore_dmae_sanity(p_hwfn, p_ptt, "pf_phase");
+       if (rc)
+               return rc;
+
        /* PF Init sequence */
        rc = ecore_init_run(p_hwfn, p_ptt, PHASE_PF, rel_pf_id, hw_mode);
        if (rc)
@@ -2214,48 +2217,48 @@ ecore_hw_init_pf(struct ecore_hwfn *p_hwfn,
 
                /* send function start command */
                rc = ecore_sp_pf_start(p_hwfn, p_ptt, p_tunn,
-                                      p_hwfn->p_dev->mf_mode,
                                       allow_npar_tx_switch);
                if (rc) {
                        DP_NOTICE(p_hwfn, true,
                                  "Function start ramrod failed\n");
                } else {
-                       prs_reg = ecore_rd(p_hwfn, p_ptt, PRS_REG_SEARCH_TAG1);
-                       DP_VERBOSE(p_hwfn, ECORE_MSG_STORAGE,
-                                  "PRS_REG_SEARCH_TAG1: %x\n", prs_reg);
-
-                       if (p_hwfn->hw_info.personality == ECORE_PCI_FCOE) {
-                               ecore_wr(p_hwfn, p_ptt, PRS_REG_SEARCH_TAG1,
-                                        (1 << 2));
-                               ecore_wr(p_hwfn, p_ptt,
-                                   PRS_REG_PKT_LEN_STAT_TAGS_NOT_COUNTED_FIRST,
-                                   0x100);
-                       }
-                       DP_VERBOSE(p_hwfn, ECORE_MSG_STORAGE,
-                                  "PRS_REG_SEARCH registers after start PFn\n");
-                       prs_reg = ecore_rd(p_hwfn, p_ptt, PRS_REG_SEARCH_TCP);
-                       DP_VERBOSE(p_hwfn, ECORE_MSG_STORAGE,
-                                  "PRS_REG_SEARCH_TCP: %x\n", prs_reg);
-                       prs_reg = ecore_rd(p_hwfn, p_ptt, PRS_REG_SEARCH_UDP);
-                       DP_VERBOSE(p_hwfn, ECORE_MSG_STORAGE,
-                                  "PRS_REG_SEARCH_UDP: %x\n", prs_reg);
-                       prs_reg = ecore_rd(p_hwfn, p_ptt, PRS_REG_SEARCH_FCOE);
-                       DP_VERBOSE(p_hwfn, ECORE_MSG_STORAGE,
-                                  "PRS_REG_SEARCH_FCOE: %x\n", prs_reg);
-                       prs_reg = ecore_rd(p_hwfn, p_ptt, PRS_REG_SEARCH_ROCE);
-                       DP_VERBOSE(p_hwfn, ECORE_MSG_STORAGE,
-                                  "PRS_REG_SEARCH_ROCE: %x\n", prs_reg);
-                       prs_reg = ecore_rd(p_hwfn, p_ptt,
-                                          PRS_REG_SEARCH_TCP_FIRST_FRAG);
-                       DP_VERBOSE(p_hwfn, ECORE_MSG_STORAGE,
-                                  "PRS_REG_SEARCH_TCP_FIRST_FRAG: %x\n",
-                                  prs_reg);
-                       prs_reg = ecore_rd(p_hwfn, p_ptt, PRS_REG_SEARCH_TAG1);
-                       DP_VERBOSE(p_hwfn, ECORE_MSG_STORAGE,
-                                  "PRS_REG_SEARCH_TAG1: %x\n", prs_reg);
+                       return rc;
                }
+               prs_reg = ecore_rd(p_hwfn, p_ptt, PRS_REG_SEARCH_TAG1);
+               DP_VERBOSE(p_hwfn, ECORE_MSG_STORAGE,
+                               "PRS_REG_SEARCH_TAG1: %x\n", prs_reg);
+
+               if (p_hwfn->hw_info.personality == ECORE_PCI_FCOE) {
+                       ecore_wr(p_hwfn, p_ptt, PRS_REG_SEARCH_TAG1,
+                                       (1 << 2));
+                       ecore_wr(p_hwfn, p_ptt,
+                                PRS_REG_PKT_LEN_STAT_TAGS_NOT_COUNTED_FIRST,
+                                0x100);
+               }
+               DP_VERBOSE(p_hwfn, ECORE_MSG_STORAGE,
+                               "PRS_REG_SEARCH registers after start PFn\n");
+               prs_reg = ecore_rd(p_hwfn, p_ptt, PRS_REG_SEARCH_TCP);
+               DP_VERBOSE(p_hwfn, ECORE_MSG_STORAGE,
+                               "PRS_REG_SEARCH_TCP: %x\n", prs_reg);
+               prs_reg = ecore_rd(p_hwfn, p_ptt, PRS_REG_SEARCH_UDP);
+               DP_VERBOSE(p_hwfn, ECORE_MSG_STORAGE,
+                               "PRS_REG_SEARCH_UDP: %x\n", prs_reg);
+               prs_reg = ecore_rd(p_hwfn, p_ptt, PRS_REG_SEARCH_FCOE);
+               DP_VERBOSE(p_hwfn, ECORE_MSG_STORAGE,
+                               "PRS_REG_SEARCH_FCOE: %x\n", prs_reg);
+               prs_reg = ecore_rd(p_hwfn, p_ptt, PRS_REG_SEARCH_ROCE);
+               DP_VERBOSE(p_hwfn, ECORE_MSG_STORAGE,
+                               "PRS_REG_SEARCH_ROCE: %x\n", prs_reg);
+               prs_reg = ecore_rd(p_hwfn, p_ptt,
+                               PRS_REG_SEARCH_TCP_FIRST_FRAG);
+               DP_VERBOSE(p_hwfn, ECORE_MSG_STORAGE,
+                               "PRS_REG_SEARCH_TCP_FIRST_FRAG: %x\n",
+                               prs_reg);
+               prs_reg = ecore_rd(p_hwfn, p_ptt, PRS_REG_SEARCH_TAG1);
+               DP_VERBOSE(p_hwfn, ECORE_MSG_STORAGE,
+                               "PRS_REG_SEARCH_TAG1: %x\n", prs_reg);
        }
-       return rc;
+       return ECORE_SUCCESS;
 }
 
 enum _ecore_status_t ecore_pglueb_set_pfid_enable(struct ecore_hwfn *p_hwfn,
@@ -2297,28 +2300,16 @@ static void ecore_reset_mb_shadow(struct ecore_hwfn *p_hwfn,
                    p_hwfn->mcp_info->mfw_mb_length);
 }
 
-enum _ecore_status_t ecore_vf_start(struct ecore_hwfn *p_hwfn,
-                                   struct ecore_hw_init_params *p_params)
-{
-       if (p_params->p_tunn) {
-               ecore_vf_set_vf_start_tunn_update_param(p_params->p_tunn);
-               ecore_vf_pf_tunnel_param_update(p_hwfn, p_params->p_tunn);
-       }
-
-       p_hwfn->b_int_enabled = 1;
-
-       return ECORE_SUCCESS;
-}
-
 static void ecore_pglueb_clear_err(struct ecore_hwfn *p_hwfn,
-                                    struct ecore_ptt *p_ptt)
+                                  struct ecore_ptt *p_ptt)
 {
        ecore_wr(p_hwfn, p_ptt, PGLUE_B_REG_WAS_ERROR_PF_31_0_CLR,
                 1 << p_hwfn->abs_pf_id);
 }
 
-static void
-ecore_fill_load_req_params(struct ecore_load_req_params *p_load_req,
+static enum _ecore_status_t
+ecore_fill_load_req_params(struct ecore_hwfn *p_hwfn,
+                          struct ecore_load_req_params *p_load_req,
                           struct ecore_drv_load_params *p_drv_load)
 {
        /* Make sure that if ecore-client didn't provide inputs, all the
@@ -2330,15 +2321,64 @@ ecore_fill_load_req_params(struct ecore_load_req_params *p_load_req,
 
        OSAL_MEM_ZERO(p_load_req, sizeof(*p_load_req));
 
-       if (p_drv_load != OSAL_NULL) {
-               p_load_req->drv_role = p_drv_load->is_crash_kernel ?
-                                      ECORE_DRV_ROLE_KDUMP :
-                                      ECORE_DRV_ROLE_OS;
+       if (p_drv_load == OSAL_NULL)
+               goto out;
+
+       p_load_req->drv_role = p_drv_load->is_crash_kernel ?
+                              ECORE_DRV_ROLE_KDUMP :
+                              ECORE_DRV_ROLE_OS;
+       p_load_req->avoid_eng_reset = p_drv_load->avoid_eng_reset;
+       p_load_req->override_force_load = p_drv_load->override_force_load;
+
+       /* Old MFW versions don't support timeout values other than default and
+        * none, so these values are replaced according to the fall-back action.
+        */
+
+       if (p_drv_load->mfw_timeout_val == ECORE_LOAD_REQ_LOCK_TO_DEFAULT ||
+           p_drv_load->mfw_timeout_val == ECORE_LOAD_REQ_LOCK_TO_NONE ||
+           (p_hwfn->mcp_info->capabilities &
+            FW_MB_PARAM_FEATURE_SUPPORT_DRV_LOAD_TO)) {
                p_load_req->timeout_val = p_drv_load->mfw_timeout_val;
-               p_load_req->avoid_eng_reset = p_drv_load->avoid_eng_reset;
-               p_load_req->override_force_load =
-                       p_drv_load->override_force_load;
+               goto out;
+       }
+
+       switch (p_drv_load->mfw_timeout_fallback) {
+       case ECORE_TO_FALLBACK_TO_NONE:
+               p_load_req->timeout_val = ECORE_LOAD_REQ_LOCK_TO_NONE;
+               break;
+       case ECORE_TO_FALLBACK_TO_DEFAULT:
+               p_load_req->timeout_val = ECORE_LOAD_REQ_LOCK_TO_DEFAULT;
+               break;
+       case ECORE_TO_FALLBACK_FAIL_LOAD:
+               DP_NOTICE(p_hwfn, false,
+                         "Received %d as a value for MFW timeout while the MFW supports only default [%d] or none [%d]. Abort.\n",
+                         p_drv_load->mfw_timeout_val,
+                         ECORE_LOAD_REQ_LOCK_TO_DEFAULT,
+                         ECORE_LOAD_REQ_LOCK_TO_NONE);
+               return ECORE_ABORTED;
        }
+
+       DP_INFO(p_hwfn,
+               "Modified the MFW timeout value from %d to %s [%d] due to lack of MFW support\n",
+               p_drv_load->mfw_timeout_val,
+               (p_load_req->timeout_val == ECORE_LOAD_REQ_LOCK_TO_DEFAULT) ?
+               "default" : "none",
+               p_load_req->timeout_val);
+out:
+       return ECORE_SUCCESS;
+}
+
+enum _ecore_status_t ecore_vf_start(struct ecore_hwfn *p_hwfn,
+                                   struct ecore_hw_init_params *p_params)
+{
+       if (p_params->p_tunn) {
+               ecore_vf_set_vf_start_tunn_update_param(p_params->p_tunn);
+               ecore_vf_pf_tunnel_param_update(p_hwfn, p_params->p_tunn);
+       }
+
+       p_hwfn->b_int_enabled = 1;
+
+       return ECORE_SUCCESS;
 }
 
 enum _ecore_status_t ecore_hw_init(struct ecore_dev *p_dev,
@@ -2381,12 +2421,17 @@ enum _ecore_status_t ecore_hw_init(struct ecore_dev *p_dev,
                if (rc != ECORE_SUCCESS)
                        return rc;
 
-               ecore_fill_load_req_params(&load_req_params,
-                                          p_params->p_drv_load_params);
+               ecore_set_spq_block_timeout(p_hwfn, p_params->spq_timeout_ms);
+
+               rc = ecore_fill_load_req_params(p_hwfn, &load_req_params,
+                                               p_params->p_drv_load_params);
+               if (rc != ECORE_SUCCESS)
+                       return rc;
+
                rc = ecore_mcp_load_req(p_hwfn, p_hwfn->p_main_ptt,
                                        &load_req_params);
                if (rc != ECORE_SUCCESS) {
-                       DP_NOTICE(p_hwfn, true,
+                       DP_NOTICE(p_hwfn, false,
                                  "Failed sending a LOAD_REQ command\n");
                        return rc;
                }
@@ -2419,10 +2464,17 @@ enum _ecore_status_t ecore_hw_init(struct ecore_dev *p_dev,
                p_hwfn->first_on_engine = (load_code ==
                                           FW_MSG_CODE_DRV_LOAD_ENGINE);
 
-               if (!qm_lock_init) {
+               if (!qm_lock_ref_cnt) {
+#ifdef CONFIG_ECORE_LOCK_ALLOC
+                       rc = OSAL_SPIN_LOCK_ALLOC(p_hwfn, &qm_lock);
+                       if (rc) {
+                               DP_ERR(p_hwfn, "qm_lock allocation failed\n");
+                               goto qm_lock_fail;
+                       }
+#endif
                        OSAL_SPIN_LOCK_INIT(&qm_lock);
-                       qm_lock_init = true;
                }
+               ++qm_lock_ref_cnt;
 
                /* Clean up chip from previous driver if such remains exist.
                 * This is not needed when the PF is the first one on the
@@ -2477,15 +2529,23 @@ enum _ecore_status_t ecore_hw_init(struct ecore_dev *p_dev,
                }
 
                if (rc != ECORE_SUCCESS) {
-                       DP_NOTICE(p_hwfn, true,
+                       DP_NOTICE(p_hwfn, false,
                                  "init phase failed for loadcode 0x%x (rc %d)\n",
                                  load_code, rc);
                        goto load_err;
                }
 
                rc = ecore_mcp_load_done(p_hwfn, p_hwfn->p_main_ptt);
-               if (rc != ECORE_SUCCESS)
+               if (rc != ECORE_SUCCESS) {
+                       DP_NOTICE(p_hwfn, false,
+                                 "Sending load done failed, rc = %d\n", rc);
+                       if (rc == ECORE_NOMEM) {
+                               DP_NOTICE(p_hwfn, false,
+                                         "Sending load done was failed due to memory allocation failure\n");
+                               goto load_err;
+                       }
                        return rc;
+               }
 
                /* send DCBX attention request command */
                DP_VERBOSE(p_hwfn, ECORE_MSG_DCB,
@@ -2495,7 +2555,7 @@ enum _ecore_status_t ecore_hw_init(struct ecore_dev *p_dev,
                                   1 << DRV_MB_PARAM_DCBX_NOTIFY_OFFSET, &resp,
                                   &param);
                if (rc != ECORE_SUCCESS) {
-                       DP_NOTICE(p_hwfn, true,
+                       DP_NOTICE(p_hwfn, false,
                                  "Failed to send DCBX attention request\n");
                        return rc;
                }
@@ -2528,6 +2588,12 @@ enum _ecore_status_t ecore_hw_init(struct ecore_dev *p_dev,
        return rc;
 
 load_err:
+       --qm_lock_ref_cnt;
+#ifdef CONFIG_ECORE_LOCK_ALLOC
+       if (!qm_lock_ref_cnt)
+               OSAL_SPIN_LOCK_DEALLOC(&qm_lock);
+qm_lock_fail:
+#endif
        /* The MFW load lock should be released regardless of success or failure
         * of initialization.
         * TODO: replace this with an attempt to send cancel_load.
@@ -2562,8 +2628,8 @@ static void ecore_hw_timers_stop(struct ecore_dev *p_dev,
        if (i < ECORE_HW_STOP_RETRY_LIMIT)
                return;
 
-       DP_NOTICE(p_hwfn, true, "Timers linear scans are not over"
-                 " [Connection %02x Tasks %02x]\n",
+       DP_NOTICE(p_hwfn, false,
+                 "Timers linear scans are not over [Connection %02x Tasks %02x]\n",
                  (u8)ecore_rd(p_hwfn, p_ptt, TM_REG_PF_SCAN_ACTIVE_CONN),
                  (u8)ecore_rd(p_hwfn, p_ptt, TM_REG_PF_SCAN_ACTIVE_TASK));
 }
@@ -2628,7 +2694,7 @@ enum _ecore_status_t ecore_hw_stop(struct ecore_dev *p_dev)
                if (!p_dev->recov_in_prog) {
                        rc = ecore_mcp_unload_req(p_hwfn, p_ptt);
                        if (rc != ECORE_SUCCESS) {
-                               DP_NOTICE(p_hwfn, true,
+                               DP_NOTICE(p_hwfn, false,
                                          "Failed sending a UNLOAD_REQ command. rc = %d.\n",
                                          rc);
                                rc2 = ECORE_UNKNOWN_ERROR;
@@ -2643,7 +2709,7 @@ enum _ecore_status_t ecore_hw_stop(struct ecore_dev *p_dev)
 
                rc = ecore_sp_pf_stop(p_hwfn);
                if (rc != ECORE_SUCCESS) {
-                       DP_NOTICE(p_hwfn, true,
+                       DP_NOTICE(p_hwfn, false,
                                  "Failed to close PF against FW [rc = %d]. Continue to stop HW to prevent illegal host access by the device.\n",
                                  rc);
                        rc2 = ECORE_UNKNOWN_ERROR;
@@ -2697,10 +2763,21 @@ enum _ecore_status_t ecore_hw_stop(struct ecore_dev *p_dev)
                ecore_wr(p_hwfn, p_ptt, DORQ_REG_PF_DB_ENABLE, 0);
                ecore_wr(p_hwfn, p_ptt, QM_REG_PF_EN, 0);
 
+               --qm_lock_ref_cnt;
+#ifdef CONFIG_ECORE_LOCK_ALLOC
+               if (!qm_lock_ref_cnt)
+                       OSAL_SPIN_LOCK_DEALLOC(&qm_lock);
+#endif
+
                if (!p_dev->recov_in_prog) {
-                       ecore_mcp_unload_done(p_hwfn, p_ptt);
+                       rc = ecore_mcp_unload_done(p_hwfn, p_ptt);
+                       if (rc == ECORE_NOMEM) {
+                               DP_NOTICE(p_hwfn, false,
+                                        "Failed sending an UNLOAD_DONE command due to a memory allocation failure. Resending.\n");
+                               rc = ecore_mcp_unload_done(p_hwfn, p_ptt);
+                       }
                        if (rc != ECORE_SUCCESS) {
-                               DP_NOTICE(p_hwfn, true,
+                               DP_NOTICE(p_hwfn, false,
                                          "Failed sending a UNLOAD_DONE command. rc = %d.\n",
                                          rc);
                                rc2 = ECORE_UNKNOWN_ERROR;
@@ -2883,12 +2960,15 @@ static void ecore_hw_set_feat(struct ecore_hwfn *p_hwfn)
                                   FEAT_NUM(p_hwfn, ECORE_VF_L2_QUE));
        }
 
-       feat_num[ECORE_FCOE_CQ] = OSAL_MIN_T(u32, sb_cnt.cnt,
-                                            RESC_NUM(p_hwfn,
-                                                     ECORE_CMDQS_CQS));
-       feat_num[ECORE_ISCSI_CQ] = OSAL_MIN_T(u32, sb_cnt.cnt,
-                                             RESC_NUM(p_hwfn,
-                                                      ECORE_CMDQS_CQS));
+       if (ECORE_IS_FCOE_PERSONALITY(p_hwfn))
+               feat_num[ECORE_FCOE_CQ] =
+                       OSAL_MIN_T(u32, sb_cnt.cnt, RESC_NUM(p_hwfn,
+                                                            ECORE_CMDQS_CQS));
+
+       if (ECORE_IS_ISCSI_PERSONALITY(p_hwfn))
+               feat_num[ECORE_ISCSI_CQ] =
+                       OSAL_MIN_T(u32, sb_cnt.cnt, RESC_NUM(p_hwfn,
+                                                            ECORE_CMDQS_CQS));
 
        DP_VERBOSE(p_hwfn, ECORE_MSG_PROBE,
                   "#PF_L2_QUEUE=%d VF_L2_QUEUES=%d #ROCE_CNQ=%d #FCOE_CQ=%d #ISCSI_CQ=%d #SB=%d\n",
@@ -2948,7 +3028,7 @@ __ecore_hw_set_soft_resc_size(struct ecore_hwfn *p_hwfn,
        rc = ecore_mcp_set_resc_max_val(p_hwfn, p_ptt, res_id,
                                        resc_max_val, p_mcp_resp);
        if (rc != ECORE_SUCCESS) {
-               DP_NOTICE(p_hwfn, true,
+               DP_NOTICE(p_hwfn, false,
                          "MFW response failure for a max value setting of resource %d [%s]\n",
                          res_id, ecore_hw_get_resc_name(res_id));
                return rc;
@@ -3194,7 +3274,7 @@ static enum _ecore_status_t ecore_hw_get_resc(struct ecore_hwfn *p_hwfn,
         * resources allocation queries should be atomic. Since several PFs can
         * run in parallel - a resource lock is needed.
         * If either the resource lock or resource set value commands are not
-        * supported - skip the the max values setting, release the lock if
+        * supported - skip the max values setting, release the lock if
         * needed, and proceed to the queries. Other failures, including a
         * failure to acquire the lock, will cause this function to fail.
         * Old drivers that don't acquire the lock can run in parallel, and
@@ -3332,8 +3412,8 @@ ecore_hw_get_nvm_info(struct ecore_hwfn *p_hwfn,
        nvm_cfg1_offset = ecore_rd(p_hwfn, p_ptt, nvm_cfg_addr + 4);
 
        addr = MCP_REG_SCRATCH + nvm_cfg1_offset +
-           OFFSETOF(struct nvm_cfg1, glob) + OFFSETOF(struct nvm_cfg1_glob,
-                                                      core_cfg);
+                  OFFSETOF(struct nvm_cfg1, glob) +
+                  OFFSETOF(struct nvm_cfg1_glob, core_cfg);
 
        core_cfg = ecore_rd(p_hwfn, p_ptt, addr);
 
@@ -3413,8 +3493,8 @@ ecore_hw_get_nvm_info(struct ecore_hwfn *p_hwfn,
        p_caps->speed_capabilities = link->speed.advertised_speeds;
 
        link_temp = ecore_rd(p_hwfn, p_ptt,
-                            port_cfg_addr +
-                            OFFSETOF(struct nvm_cfg1_port, link_settings));
+                                port_cfg_addr +
+                                OFFSETOF(struct nvm_cfg1_port, link_settings));
        switch ((link_temp & NVM_CFG1_PORT_DRV_LINK_SPEED_MASK) >>
                NVM_CFG1_PORT_DRV_LINK_SPEED_OFFSET) {
        case NVM_CFG1_PORT_DRV_LINK_SPEED_AUTONEG:
@@ -3494,8 +3574,8 @@ ecore_hw_get_nvm_info(struct ecore_hwfn *p_hwfn,
 
        /* Read Multi-function information from shmem */
        addr = MCP_REG_SCRATCH + nvm_cfg1_offset +
-           OFFSETOF(struct nvm_cfg1, glob) +
-           OFFSETOF(struct nvm_cfg1_glob, generic_cont0);
+                  OFFSETOF(struct nvm_cfg1, glob) +
+                  OFFSETOF(struct nvm_cfg1_glob, generic_cont0);
 
        generic_cont0 = ecore_rd(p_hwfn, p_ptt, addr);
 
@@ -3503,6 +3583,41 @@ ecore_hw_get_nvm_info(struct ecore_hwfn *p_hwfn,
            NVM_CFG1_GLOB_MF_MODE_OFFSET;
 
        switch (mf_mode) {
+       case NVM_CFG1_GLOB_MF_MODE_MF_ALLOWED:
+               p_hwfn->p_dev->mf_bits = 1 << ECORE_MF_OVLAN_CLSS;
+               break;
+       case NVM_CFG1_GLOB_MF_MODE_UFP:
+               p_hwfn->p_dev->mf_bits = 1 << ECORE_MF_OVLAN_CLSS |
+                                        1 << ECORE_MF_UFP_SPECIFIC;
+               break;
+
+       case NVM_CFG1_GLOB_MF_MODE_NPAR1_0:
+               p_hwfn->p_dev->mf_bits = 1 << ECORE_MF_LLH_MAC_CLSS |
+                                        1 << ECORE_MF_LLH_PROTO_CLSS |
+                                        1 << ECORE_MF_LL2_NON_UNICAST |
+                                        1 << ECORE_MF_INTER_PF_SWITCH |
+                                        1 << ECORE_MF_DISABLE_ARFS;
+               break;
+       case NVM_CFG1_GLOB_MF_MODE_DEFAULT:
+               p_hwfn->p_dev->mf_bits = 1 << ECORE_MF_LLH_MAC_CLSS |
+                                        1 << ECORE_MF_LLH_PROTO_CLSS |
+                                        1 << ECORE_MF_LL2_NON_UNICAST;
+               if (ECORE_IS_BB(p_hwfn->p_dev))
+                       p_hwfn->p_dev->mf_bits |= 1 << ECORE_MF_NEED_DEF_PF;
+               break;
+       }
+       DP_INFO(p_hwfn, "Multi function mode is 0x%lx\n",
+               p_hwfn->p_dev->mf_bits);
+
+       if (ECORE_IS_CMT(p_hwfn->p_dev))
+               p_hwfn->p_dev->mf_bits |= (1 << ECORE_MF_DISABLE_ARFS);
+
+       /* It's funny since we have another switch, but it's easier
+        * to throw this away in linux this way. Long term, it might be
+        * better to have have getters for needed ECORE_MF_* fields,
+        * convert client code and eliminate this.
+        */
+       switch (mf_mode) {
        case NVM_CFG1_GLOB_MF_MODE_MF_ALLOWED:
                p_hwfn->p_dev->mf_mode = ECORE_MF_OVLAN;
                break;
@@ -3512,31 +3627,32 @@ ecore_hw_get_nvm_info(struct ecore_hwfn *p_hwfn,
        case NVM_CFG1_GLOB_MF_MODE_DEFAULT:
                p_hwfn->p_dev->mf_mode = ECORE_MF_DEFAULT;
                break;
+       case NVM_CFG1_GLOB_MF_MODE_UFP:
+               p_hwfn->p_dev->mf_mode = ECORE_MF_UFP;
+               break;
        }
-       DP_INFO(p_hwfn, "Multi function mode is %08x\n",
-               p_hwfn->p_dev->mf_mode);
 
        /* Read Multi-function information from shmem */
        addr = MCP_REG_SCRATCH + nvm_cfg1_offset +
-           OFFSETOF(struct nvm_cfg1, glob) +
-           OFFSETOF(struct nvm_cfg1_glob, device_capabilities);
+                  OFFSETOF(struct nvm_cfg1, glob) +
+                  OFFSETOF(struct nvm_cfg1_glob, device_capabilities);
 
        device_capabilities = ecore_rd(p_hwfn, p_ptt, addr);
        if (device_capabilities & NVM_CFG1_GLOB_DEVICE_CAPABILITIES_ETHERNET)
                OSAL_SET_BIT(ECORE_DEV_CAP_ETH,
-                            &p_hwfn->hw_info.device_capabilities);
+                               &p_hwfn->hw_info.device_capabilities);
        if (device_capabilities & NVM_CFG1_GLOB_DEVICE_CAPABILITIES_FCOE)
                OSAL_SET_BIT(ECORE_DEV_CAP_FCOE,
-                            &p_hwfn->hw_info.device_capabilities);
+                               &p_hwfn->hw_info.device_capabilities);
        if (device_capabilities & NVM_CFG1_GLOB_DEVICE_CAPABILITIES_ISCSI)
                OSAL_SET_BIT(ECORE_DEV_CAP_ISCSI,
-                            &p_hwfn->hw_info.device_capabilities);
+                               &p_hwfn->hw_info.device_capabilities);
        if (device_capabilities & NVM_CFG1_GLOB_DEVICE_CAPABILITIES_ROCE)
                OSAL_SET_BIT(ECORE_DEV_CAP_ROCE,
-                            &p_hwfn->hw_info.device_capabilities);
+                               &p_hwfn->hw_info.device_capabilities);
        if (device_capabilities & NVM_CFG1_GLOB_DEVICE_CAPABILITIES_IWARP)
                OSAL_SET_BIT(ECORE_DEV_CAP_IWARP,
-                            &p_hwfn->hw_info.device_capabilities);
+                               &p_hwfn->hw_info.device_capabilities);
 
        rc = ecore_mcp_fill_shmem_func_info(p_hwfn, p_ptt);
        if (rc != ECORE_SUCCESS && p_params->b_relaxed_probe) {
@@ -3813,6 +3929,8 @@ ecore_get_hw_info(struct ecore_hwfn *p_hwfn, struct ecore_ptt *p_ptt,
                ecore_mcp_cmd_port_init(p_hwfn, p_ptt);
 
                ecore_mcp_get_eee_caps(p_hwfn, p_ptt);
+
+               ecore_mcp_read_ufp_config(p_hwfn, p_ptt);
        }
 
        if (personality != ECORE_PCI_DEFAULT) {
@@ -4011,7 +4129,7 @@ ecore_hw_prepare_single(struct ecore_hwfn *p_hwfn,
        /* Allocate PTT pool */
        rc = ecore_ptt_pool_alloc(p_hwfn);
        if (rc) {
-               DP_NOTICE(p_hwfn, true, "Failed to prepare hwfn's hw\n");
+               DP_NOTICE(p_hwfn, false, "Failed to prepare hwfn's hw\n");
                if (p_params->b_relaxed_probe)
                        p_params->p_relaxed_res = ECORE_HW_PREPARE_FAILED_MEM;
                goto err0;
@@ -4036,7 +4154,7 @@ ecore_hw_prepare_single(struct ecore_hwfn *p_hwfn,
        /* Initialize MCP structure */
        rc = ecore_mcp_cmd_init(p_hwfn, p_hwfn->p_main_ptt);
        if (rc) {
-               DP_NOTICE(p_hwfn, true, "Failed initializing mcp command\n");
+               DP_NOTICE(p_hwfn, false, "Failed initializing mcp command\n");
                if (p_params->b_relaxed_probe)
                        p_params->p_relaxed_res = ECORE_HW_PREPARE_FAILED_MEM;
                goto err1;
@@ -4046,7 +4164,7 @@ ecore_hw_prepare_single(struct ecore_hwfn *p_hwfn,
        rc = ecore_get_hw_info(p_hwfn, p_hwfn->p_main_ptt,
                               p_params->personality, p_params);
        if (rc) {
-               DP_NOTICE(p_hwfn, true, "Failed to get HW information\n");
+               DP_NOTICE(p_hwfn, false, "Failed to get HW information\n");
                goto err2;
        }
 
@@ -4089,7 +4207,7 @@ ecore_hw_prepare_single(struct ecore_hwfn *p_hwfn,
        /* Allocate the init RT array and initialize the init-ops engine */
        rc = ecore_init_alloc(p_hwfn);
        if (rc) {
-               DP_NOTICE(p_hwfn, true, "Failed to allocate the init array\n");
+               DP_NOTICE(p_hwfn, false, "Failed to allocate the init array\n");
                if (p_params->b_relaxed_probe)
                        p_params->p_relaxed_res = ECORE_HW_PREPARE_FAILED_MEM;
                goto err2;
@@ -4179,8 +4297,7 @@ enum _ecore_status_t ecore_hw_prepare(struct ecore_dev *p_dev,
                                ecore_mcp_free(p_hwfn);
                                ecore_hw_hwfn_free(p_hwfn);
                        } else {
-                               DP_NOTICE(p_dev, true,
-                                         "What do we need to free when VF hwfn1 init fails\n");
+                               DP_NOTICE(p_dev, false, "What do we need to free when VF hwfn1 init fails\n");
                        }
                        return rc;
                }
@@ -4211,7 +4328,7 @@ void ecore_hw_remove(struct ecore_dev *p_dev)
                ecore_mcp_free(p_hwfn);
 
 #ifdef CONFIG_ECORE_LOCK_ALLOC
-               OSAL_MUTEX_DEALLOC(&p_hwfn->dmae_info.mutex);
+               OSAL_SPIN_LOCK_DEALLOC(&p_hwfn->dmae_info.lock);
 #endif
        }
 
@@ -4342,7 +4459,7 @@ ecore_chain_alloc_next_ptr(struct ecore_dev *p_dev, struct ecore_chain *p_chain)
                p_virt = OSAL_DMA_ALLOC_COHERENT(p_dev, &p_phys,
                                                 ECORE_CHAIN_PAGE_SIZE);
                if (!p_virt) {
-                       DP_NOTICE(p_dev, true,
+                       DP_NOTICE(p_dev, false,
                                  "Failed to allocate chain memory\n");
                        return ECORE_NOMEM;
                }
@@ -4375,7 +4492,7 @@ ecore_chain_alloc_single(struct ecore_dev *p_dev, struct ecore_chain *p_chain)
 
        p_virt = OSAL_DMA_ALLOC_COHERENT(p_dev, &p_phys, ECORE_CHAIN_PAGE_SIZE);
        if (!p_virt) {
-               DP_NOTICE(p_dev, true, "Failed to allocate chain memory\n");
+               DP_NOTICE(p_dev, false, "Failed to allocate chain memory\n");
                return ECORE_NOMEM;
        }
 
@@ -4390,16 +4507,16 @@ ecore_chain_alloc_pbl(struct ecore_dev *p_dev,
                      struct ecore_chain *p_chain,
                      struct ecore_chain_ext_pbl *ext_pbl)
 {
-       void *p_virt = OSAL_NULL;
-       u8 *p_pbl_virt = OSAL_NULL;
-       void **pp_virt_addr_tbl = OSAL_NULL;
-       dma_addr_t p_phys = 0, p_pbl_phys = 0;
        u32 page_cnt = p_chain->page_cnt, size, i;
+       dma_addr_t p_phys = 0, p_pbl_phys = 0;
+       void **pp_virt_addr_tbl = OSAL_NULL;
+       u8 *p_pbl_virt = OSAL_NULL;
+       void *p_virt = OSAL_NULL;
 
        size = page_cnt * sizeof(*pp_virt_addr_tbl);
        pp_virt_addr_tbl = (void **)OSAL_VZALLOC(p_dev, size);
        if (!pp_virt_addr_tbl) {
-               DP_NOTICE(p_dev, true,
+               DP_NOTICE(p_dev, false,
                          "Failed to allocate memory for the chain virtual addresses table\n");
                return ECORE_NOMEM;
        }
@@ -4423,7 +4540,7 @@ ecore_chain_alloc_pbl(struct ecore_dev *p_dev,
        ecore_chain_init_pbl_mem(p_chain, p_pbl_virt, p_pbl_phys,
                                 pp_virt_addr_tbl);
        if (!p_pbl_virt) {
-               DP_NOTICE(p_dev, true, "Failed to allocate chain pbl memory\n");
+               DP_NOTICE(p_dev, false, "Failed to allocate chain pbl memory\n");
                return ECORE_NOMEM;
        }
 
@@ -4431,7 +4548,7 @@ ecore_chain_alloc_pbl(struct ecore_dev *p_dev,
                p_virt = OSAL_DMA_ALLOC_COHERENT(p_dev, &p_phys,
                                                 ECORE_CHAIN_PAGE_SIZE);
                if (!p_virt) {
-                       DP_NOTICE(p_dev, true,
+                       DP_NOTICE(p_dev, false,
                                  "Failed to allocate chain memory\n");
                        return ECORE_NOMEM;
                }
@@ -4471,7 +4588,7 @@ enum _ecore_status_t ecore_chain_alloc(struct ecore_dev *p_dev,
        rc = ecore_chain_alloc_sanity_check(p_dev, cnt_type, elem_size,
                                            page_cnt);
        if (rc) {
-               DP_NOTICE(p_dev, true,
+               DP_NOTICE(p_dev, false,
                          "Cannot allocate a chain with the given arguments:\n"
                          "[use_mode %d, mode %d, cnt_type %d, num_elems %d, elem_size %zu]\n",
                          intended_use, mode, cnt_type, num_elems, elem_size);
@@ -4607,9 +4724,10 @@ enum _ecore_status_t ecore_llh_add_mac_filter(struct ecore_hwfn *p_hwfn,
                                          struct ecore_ptt *p_ptt, u8 *p_filter)
 {
        u32 high, low, entry_num;
-       enum _ecore_status_t rc;
+       enum _ecore_status_t rc = ECORE_SUCCESS;
 
-       if (!(IS_MF_SI(p_hwfn) || IS_MF_DEFAULT(p_hwfn)))
+       if (!OSAL_TEST_BIT(ECORE_MF_LLH_MAC_CLSS,
+                          &p_hwfn->p_dev->mf_bits))
                return ECORE_SUCCESS;
 
        high = p_filter[1] | (p_filter[0] << 8);
@@ -4630,7 +4748,7 @@ enum _ecore_status_t ecore_llh_add_mac_filter(struct ecore_hwfn *p_hwfn,
                   p_filter[0], p_filter[1], p_filter[2], p_filter[3],
                   p_filter[4], p_filter[5], entry_num);
 
-       return ECORE_SUCCESS;
+       return rc;
 }
 
 static enum _ecore_status_t
@@ -4674,9 +4792,10 @@ void ecore_llh_remove_mac_filter(struct ecore_hwfn *p_hwfn,
                             struct ecore_ptt *p_ptt, u8 *p_filter)
 {
        u32 high, low, entry_num;
-       enum _ecore_status_t rc;
+       enum _ecore_status_t rc = ECORE_SUCCESS;
 
-       if (!(IS_MF_SI(p_hwfn) || IS_MF_DEFAULT(p_hwfn)))
+       if (!OSAL_TEST_BIT(ECORE_MF_LLH_MAC_CLSS,
+                          &p_hwfn->p_dev->mf_bits))
                return;
 
        high = p_filter[1] | (p_filter[0] << 8);
@@ -4748,10 +4867,11 @@ ecore_llh_add_protocol_filter(struct ecore_hwfn *p_hwfn,
                              enum ecore_llh_port_filter_type_t type)
 {
        u32 high, low, entry_num;
-       enum _ecore_status_t rc;
+       enum _ecore_status_t rc = ECORE_SUCCESS;
 
-       if (!(IS_MF_SI(p_hwfn) || IS_MF_DEFAULT(p_hwfn)))
-               return ECORE_SUCCESS;
+       if (!OSAL_TEST_BIT(ECORE_MF_LLH_PROTO_CLSS,
+                          &p_hwfn->p_dev->mf_bits))
+               return rc;
 
        high = 0;
        low = 0;
@@ -4824,7 +4944,7 @@ ecore_llh_add_protocol_filter(struct ecore_hwfn *p_hwfn,
                break;
        }
 
-       return ECORE_SUCCESS;
+       return rc;
 }
 
 static enum _ecore_status_t
@@ -4891,9 +5011,10 @@ ecore_llh_remove_protocol_filter(struct ecore_hwfn *p_hwfn,
                                 enum ecore_llh_port_filter_type_t type)
 {
        u32 high, low, entry_num;
-       enum _ecore_status_t rc;
+       enum _ecore_status_t rc = ECORE_SUCCESS;
 
-       if (!(IS_MF_SI(p_hwfn) || IS_MF_DEFAULT(p_hwfn)))
+       if (!OSAL_TEST_BIT(ECORE_MF_LLH_PROTO_CLSS,
+                          &p_hwfn->p_dev->mf_bits))
                return;
 
        high = 0;
@@ -4961,7 +5082,10 @@ static void ecore_llh_clear_all_filters_bb_ah(struct ecore_hwfn *p_hwfn,
 void ecore_llh_clear_all_filters(struct ecore_hwfn *p_hwfn,
                             struct ecore_ptt *p_ptt)
 {
-       if (!(IS_MF_SI(p_hwfn) || IS_MF_DEFAULT(p_hwfn)))
+       if (!OSAL_TEST_BIT(ECORE_MF_LLH_PROTO_CLSS,
+                          &p_hwfn->p_dev->mf_bits) &&
+           !OSAL_TEST_BIT(ECORE_MF_LLH_MAC_CLSS,
+                          &p_hwfn->p_dev->mf_bits))
                return;
 
        if (ECORE_IS_BB(p_hwfn->p_dev) || ECORE_IS_AH(p_hwfn->p_dev))
@@ -4972,7 +5096,7 @@ enum _ecore_status_t
 ecore_llh_set_function_as_default(struct ecore_hwfn *p_hwfn,
                                  struct ecore_ptt *p_ptt)
 {
-       if (IS_MF_DEFAULT(p_hwfn) && ECORE_IS_BB(p_hwfn->p_dev)) {
+       if (OSAL_TEST_BIT(ECORE_MF_NEED_DEF_PF, &p_hwfn->p_dev->mf_bits)) {
                ecore_wr(p_hwfn, p_ptt,
                         NIG_REG_LLH_TAGMAC_DEF_PF_VECTOR,
                         1 << p_hwfn->abs_pf_id / 2);
@@ -5208,13 +5332,6 @@ static enum _ecore_status_t ecore_init_wfq_param(struct ecore_hwfn *p_hwfn,
        non_requested_count = num_vports - req_count;
 
        /* validate possible error cases */
-       if (req_rate > min_pf_rate) {
-               DP_VERBOSE(p_hwfn, ECORE_MSG_LINK,
-                          "Vport [%d] - Requested rate[%d Mbps] is greater than configured PF min rate[%d Mbps]\n",
-                          vport_id, req_rate, min_pf_rate);
-               return ECORE_INVAL;
-       }
-
        if (req_rate < min_pf_rate / ECORE_WFQ_UNIT) {
                DP_VERBOSE(p_hwfn, ECORE_MSG_LINK,
                           "Vport [%d] - Requested rate[%d Mbps] is less than one percent of configured PF min rate[%d Mbps]\n",