net/qede/base: prevent race condition during unload
[dpdk.git] / drivers / net / qede / base / ecore_dev.c
index 7fce4fd..d8e4ca2 100644 (file)
@@ -30,6 +30,7 @@
 #include "nvm_cfg.h"
 #include "ecore_dev_api.h"
 #include "ecore_dcbx.h"
+#include "ecore_l2.h"
 
 /* TODO - there's a bug in DCBx re-configuration flows in MF, as the QM
  * registers involved are not split and thus configuration is a race where
@@ -145,33 +146,27 @@ void ecore_resc_free(struct ecore_dev *p_dev)
 {
        int i;
 
-       if (IS_VF(p_dev))
+       if (IS_VF(p_dev)) {
+               for_each_hwfn(p_dev, i)
+                       ecore_l2_free(&p_dev->hwfns[i]);
                return;
+       }
 
        OSAL_FREE(p_dev, p_dev->fw_data);
 
        OSAL_FREE(p_dev, p_dev->reset_stats);
 
-       for_each_hwfn(p_dev, i) {
-               struct ecore_hwfn *p_hwfn = &p_dev->hwfns[i];
-
-               OSAL_FREE(p_dev, p_hwfn->p_tx_cids);
-               OSAL_FREE(p_dev, p_hwfn->p_rx_cids);
-       }
-
        for_each_hwfn(p_dev, i) {
                struct ecore_hwfn *p_hwfn = &p_dev->hwfns[i];
 
                ecore_cxt_mngr_free(p_hwfn);
                ecore_qm_info_free(p_hwfn);
                ecore_spq_free(p_hwfn);
-               ecore_eq_free(p_hwfn, p_hwfn->p_eq);
-               ecore_consq_free(p_hwfn, p_hwfn->p_consq);
+               ecore_eq_free(p_hwfn);
+               ecore_consq_free(p_hwfn);
                ecore_int_free(p_hwfn);
-#ifdef CONFIG_ECORE_LL2
-               ecore_ll2_free(p_hwfn, p_hwfn->p_ll2_info);
-#endif
                ecore_iov_free(p_hwfn);
+               ecore_l2_free(p_hwfn);
                ecore_dmae_info_free(p_hwfn);
                ecore_dcbx_info_free(p_hwfn, p_hwfn->p_dcbx_info);
                /* @@@TBD Flush work-queue ? */
@@ -301,6 +296,7 @@ u16 ecore_init_qm_get_num_pqs(struct ecore_hwfn *p_hwfn)
 static void ecore_init_qm_params(struct ecore_hwfn *p_hwfn)
 {
        struct ecore_qm_info *qm_info = &p_hwfn->qm_info;
+       bool four_port;
 
        /* pq and vport bases for this PF */
        qm_info->start_pq = (u16)RESC_START(p_hwfn, ECORE_PQ);
@@ -310,10 +306,19 @@ static void ecore_init_qm_params(struct ecore_hwfn *p_hwfn)
        qm_info->vport_rl_en = 1;
        qm_info->vport_wfq_en = 1;
 
+       /* TC config is different for AH 4 port */
+       four_port = p_hwfn->p_dev->num_ports_in_engines == MAX_NUM_PORTS_K2;
+
        /* in AH 4 port we have fewer TCs per port */
-       qm_info->max_phys_tcs_per_port =
-               p_hwfn->p_dev->num_ports_in_engines == MAX_NUM_PORTS_K2 ?
-                       NUM_PHYS_TCS_4PORT_K2 : NUM_OF_PHYS_TCS;
+       qm_info->max_phys_tcs_per_port = four_port ? NUM_PHYS_TCS_4PORT_K2 :
+                                                    NUM_OF_PHYS_TCS;
+
+       /* unless MFW indicated otherwise, ooo_tc should be 3 for AH 4 port and
+        * 4 otherwise
+        */
+       if (!qm_info->ooo_tc)
+               qm_info->ooo_tc = four_port ? DCBX_TCP_OOO_K2_4PORT_TC :
+                                             DCBX_TCP_OOO_TC;
 }
 
 /* initialize qm vport params */
@@ -542,8 +547,7 @@ static void ecore_init_qm_ooo_pq(struct ecore_hwfn *p_hwfn)
                return;
 
        ecore_init_qm_set_idx(p_hwfn, PQ_FLAGS_OOO, qm_info->num_pqs);
-       ecore_init_qm_pq(p_hwfn, qm_info, DCBX_ISCSI_OOO_TC,
-                        PQ_INIT_SHARE_VPORT);
+       ecore_init_qm_pq(p_hwfn, qm_info, qm_info->ooo_tc, PQ_INIT_SHARE_VPORT);
 }
 
 static void ecore_init_qm_pure_ack_pq(struct ecore_hwfn *p_hwfn)
@@ -836,52 +840,23 @@ alloc_err:
 
 enum _ecore_status_t ecore_resc_alloc(struct ecore_dev *p_dev)
 {
-       struct ecore_consq *p_consq;
-       struct ecore_eq *p_eq;
-#ifdef CONFIG_ECORE_LL2
-       struct ecore_ll2_info *p_ll2_info;
-#endif
        enum _ecore_status_t rc = ECORE_SUCCESS;
        int i;
 
-       if (IS_VF(p_dev))
+       if (IS_VF(p_dev)) {
+               for_each_hwfn(p_dev, i) {
+                       rc = ecore_l2_alloc(&p_dev->hwfns[i]);
+                       if (rc != ECORE_SUCCESS)
+                               return rc;
+               }
                return rc;
+       }
 
        p_dev->fw_data = OSAL_ZALLOC(p_dev, GFP_KERNEL,
                                     sizeof(*p_dev->fw_data));
        if (!p_dev->fw_data)
                return ECORE_NOMEM;
 
-       /* Allocate Memory for the Queue->CID mapping */
-       for_each_hwfn(p_dev, i) {
-               struct ecore_hwfn *p_hwfn = &p_dev->hwfns[i];
-               u32 num_tx_conns = RESC_NUM(p_hwfn, ECORE_L2_QUEUE);
-               int tx_size, rx_size;
-
-               /* @@@TMP - resc management, change to actual required size */
-               if (p_hwfn->pf_params.eth_pf_params.num_cons > num_tx_conns)
-                       num_tx_conns = p_hwfn->pf_params.eth_pf_params.num_cons;
-               tx_size = sizeof(struct ecore_hw_cid_data) * num_tx_conns;
-               rx_size = sizeof(struct ecore_hw_cid_data) *
-                   RESC_NUM(p_hwfn, ECORE_L2_QUEUE);
-
-               p_hwfn->p_tx_cids = OSAL_ZALLOC(p_hwfn->p_dev, GFP_KERNEL,
-                                               tx_size);
-               if (!p_hwfn->p_tx_cids) {
-                       DP_NOTICE(p_hwfn, true,
-                                 "Failed to allocate memory for Tx Cids\n");
-                       goto alloc_no_mem;
-               }
-
-               p_hwfn->p_rx_cids = OSAL_ZALLOC(p_hwfn->p_dev, GFP_KERNEL,
-                                               rx_size);
-               if (!p_hwfn->p_rx_cids) {
-                       DP_NOTICE(p_hwfn, true,
-                                 "Failed to allocate memory for Rx Cids\n");
-                       goto alloc_no_mem;
-               }
-       }
-
        for_each_hwfn(p_dev, i) {
                struct ecore_hwfn *p_hwfn = &p_dev->hwfns[i];
                u32 n_eqes, num_cons;
@@ -988,24 +963,17 @@ enum _ecore_status_t ecore_resc_alloc(struct ecore_dev *p_dev)
                        goto alloc_no_mem;
                }
 
-               p_eq = ecore_eq_alloc(p_hwfn, (u16)n_eqes);
-               if (!p_eq)
-                       goto alloc_no_mem;
-               p_hwfn->p_eq = p_eq;
+               rc = ecore_eq_alloc(p_hwfn, (u16)n_eqes);
+               if (rc)
+                       goto alloc_err;
 
-               p_consq = ecore_consq_alloc(p_hwfn);
-               if (!p_consq)
-                       goto alloc_no_mem;
-               p_hwfn->p_consq = p_consq;
-
-#ifdef CONFIG_ECORE_LL2
-               if (p_hwfn->using_ll2) {
-                       p_ll2_info = ecore_ll2_alloc(p_hwfn);
-                       if (!p_ll2_info)
-                               goto alloc_no_mem;
-                       p_hwfn->p_ll2_info = p_ll2_info;
-               }
-#endif
+               rc = ecore_consq_alloc(p_hwfn);
+               if (rc)
+                       goto alloc_err;
+
+               rc = ecore_l2_alloc(p_hwfn);
+               if (rc != ECORE_SUCCESS)
+                       goto alloc_err;
 
                /* DMA info initialization */
                rc = ecore_dmae_info_alloc(p_hwfn);
@@ -1045,16 +1013,19 @@ void ecore_resc_setup(struct ecore_dev *p_dev)
 {
        int i;
 
-       if (IS_VF(p_dev))
+       if (IS_VF(p_dev)) {
+               for_each_hwfn(p_dev, i)
+                       ecore_l2_setup(&p_dev->hwfns[i]);
                return;
+       }
 
        for_each_hwfn(p_dev, i) {
                struct ecore_hwfn *p_hwfn = &p_dev->hwfns[i];
 
                ecore_cxt_mngr_setup(p_hwfn);
                ecore_spq_setup(p_hwfn);
-               ecore_eq_setup(p_hwfn, p_hwfn->p_eq);
-               ecore_consq_setup(p_hwfn, p_hwfn->p_consq);
+               ecore_eq_setup(p_hwfn);
+               ecore_consq_setup(p_hwfn);
 
                /* Read shadow of current MFW mailbox */
                ecore_mcp_read_mb(p_hwfn, p_hwfn->p_main_ptt);
@@ -1064,11 +1035,8 @@ void ecore_resc_setup(struct ecore_dev *p_dev)
 
                ecore_int_setup(p_hwfn, p_hwfn->p_main_ptt);
 
+               ecore_l2_setup(p_hwfn);
                ecore_iov_setup(p_hwfn, p_hwfn->p_main_ptt);
-#ifdef CONFIG_ECORE_LL2
-               if (p_hwfn->using_ll2)
-                       ecore_ll2_setup(p_hwfn, p_hwfn->p_ll2_info);
-#endif
        }
 }
 
@@ -1756,7 +1724,7 @@ ecore_hw_init_pf_doorbell_bar(struct ecore_hwfn *p_hwfn,
 static enum _ecore_status_t
 ecore_hw_init_pf(struct ecore_hwfn *p_hwfn,
                 struct ecore_ptt *p_ptt,
-                struct ecore_tunn_start_params *p_tunn,
+                struct ecore_tunnel_info *p_tunn,
                 int hw_mode,
                 bool b_hw_start,
                 enum ecore_int_mode int_mode, bool allow_npar_tx_switch)
@@ -1936,13 +1904,27 @@ static void ecore_reset_mb_shadow(struct ecore_hwfn *p_hwfn,
                    p_hwfn->mcp_info->mfw_mb_length);
 }
 
+enum _ecore_status_t ecore_vf_start(struct ecore_hwfn *p_hwfn,
+                                   struct ecore_hw_init_params *p_params)
+{
+       if (p_params->p_tunn) {
+               ecore_vf_set_vf_start_tunn_update_param(p_params->p_tunn);
+               ecore_vf_pf_tunnel_param_update(p_hwfn, p_params->p_tunn);
+       }
+
+       p_hwfn->b_int_enabled = 1;
+
+       return ECORE_SUCCESS;
+}
+
 enum _ecore_status_t ecore_hw_init(struct ecore_dev *p_dev,
                                   struct ecore_hw_init_params *p_params)
 {
-       enum _ecore_status_t rc = ECORE_SUCCESS, mfw_rc;
+       struct ecore_load_req_params load_req_params;
        u32 load_code, param, drv_mb_param;
-       bool b_default_mtu = true;
        struct ecore_hwfn *p_hwfn;
+       bool b_default_mtu = true;
+       enum _ecore_status_t rc = ECORE_SUCCESS, mfw_rc;
        int i;
 
        if ((p_params->int_mode == ECORE_INT_MODE_MSI) &&
@@ -1968,7 +1950,7 @@ enum _ecore_status_t ecore_hw_init(struct ecore_dev *p_dev,
                }
 
                if (IS_VF(p_dev)) {
-                       p_hwfn->b_int_enabled = 1;
+                       ecore_vf_start(p_hwfn, p_params);
                        continue;
                }
 
@@ -1981,17 +1963,25 @@ enum _ecore_status_t ecore_hw_init(struct ecore_dev *p_dev,
                if (rc != ECORE_SUCCESS)
                        return rc;
 
-               /* @@@TBD need to add here:
-                * Check for fan failure
-                * Prev_unload
-                */
-               rc = ecore_mcp_load_req(p_hwfn, p_hwfn->p_main_ptt, &load_code);
-               if (rc) {
+               OSAL_MEM_ZERO(&load_req_params, sizeof(load_req_params));
+               load_req_params.drv_role = p_params->is_crash_kernel ?
+                                          ECORE_DRV_ROLE_KDUMP :
+                                          ECORE_DRV_ROLE_OS;
+               load_req_params.timeout_val = p_params->mfw_timeout_val;
+               load_req_params.avoid_eng_reset = p_params->avoid_eng_reset;
+               rc = ecore_mcp_load_req(p_hwfn, p_hwfn->p_main_ptt,
+                                       &load_req_params);
+               if (rc != ECORE_SUCCESS) {
                        DP_NOTICE(p_hwfn, true,
-                                 "Failed sending LOAD_REQ command\n");
+                                 "Failed sending LOAD_REQ command\n");
                        return rc;
                }
 
+               load_code = load_req_params.load_code;
+               DP_VERBOSE(p_hwfn, ECORE_MSG_SP,
+                          "Load request was sent. Load code: 0x%x\n",
+                          load_code);
+
                /* CQ75580:
                 * When coming back from hiberbate state, the registers from
                 * which shadow is read initially are not initialized. It turns
@@ -2004,10 +1994,6 @@ enum _ecore_status_t ecore_hw_init(struct ecore_dev *p_dev,
                 */
                ecore_reset_mb_shadow(p_hwfn, p_hwfn->p_main_ptt);
 
-               DP_VERBOSE(p_hwfn, ECORE_MSG_SP,
-                          "Load request was sent. Resp:0x%x, Load code: 0x%x\n",
-                          rc, load_code);
-
                /* Only relevant for recovery:
                 * Clear the indication after the LOAD_REQ command is responded
                 * by the MFW.
@@ -2026,13 +2012,13 @@ enum _ecore_status_t ecore_hw_init(struct ecore_dev *p_dev,
                case FW_MSG_CODE_DRV_LOAD_ENGINE:
                        rc = ecore_hw_init_common(p_hwfn, p_hwfn->p_main_ptt,
                                                  p_hwfn->hw_info.hw_mode);
-                       if (rc)
+                       if (rc != ECORE_SUCCESS)
                                break;
                        /* Fall into */
                case FW_MSG_CODE_DRV_LOAD_PORT:
                        rc = ecore_hw_init_port(p_hwfn, p_hwfn->p_main_ptt,
                                                p_hwfn->hw_info.hw_mode);
-                       if (rc)
+                       if (rc != ECORE_SUCCESS)
                                break;
                        /* Fall into */
                case FW_MSG_CODE_DRV_LOAD_FUNCTION:
@@ -2044,6 +2030,8 @@ enum _ecore_status_t ecore_hw_init(struct ecore_dev *p_dev,
                                              p_params->allow_npar_tx_switch);
                        break;
                default:
+                       DP_NOTICE(p_hwfn, false,
+                                 "Unexpected load code [0x%08x]", load_code);
                        rc = ECORE_NOTIMPL;
                        break;
                }
@@ -2059,9 +2047,10 @@ enum _ecore_status_t ecore_hw_init(struct ecore_dev *p_dev,
                                       0, &load_code, &param);
                if (rc != ECORE_SUCCESS)
                        return rc;
+
                if (mfw_rc != ECORE_SUCCESS) {
                        DP_NOTICE(p_hwfn, true,
-                                 "Failed sending LOAD_DONE command\n");
+                                 "Failed sending LOAD_DONE command\n");
                        return mfw_rc;
                }
 
@@ -2083,10 +2072,7 @@ enum _ecore_status_t ecore_hw_init(struct ecore_dev *p_dev,
 
        if (IS_PF(p_dev)) {
                p_hwfn = ECORE_LEADING_HWFN(p_dev);
-               drv_mb_param = (FW_MAJOR_VERSION << 24) |
-                              (FW_MINOR_VERSION << 16) |
-                              (FW_REVISION_VERSION << 8) |
-                              (FW_ENGINEERING_VERSION);
+               drv_mb_param = STORM_FW_VERSION;
                rc = ecore_mcp_cmd(p_hwfn, p_hwfn->p_main_ptt,
                                   DRV_MSG_CODE_OV_UPDATE_STORM_FW_VER,
                                   drv_mb_param, &load_code, &param);
@@ -2153,32 +2139,77 @@ void ecore_hw_timers_stop_all(struct ecore_dev *p_dev)
        }
 }
 
+static enum _ecore_status_t ecore_verify_reg_val(struct ecore_hwfn *p_hwfn,
+                                                struct ecore_ptt *p_ptt,
+                                                u32 addr, u32 expected_val)
+{
+       u32 val = ecore_rd(p_hwfn, p_ptt, addr);
+
+       if (val != expected_val) {
+               DP_NOTICE(p_hwfn, true,
+                         "Value at address 0x%08x is 0x%08x while the expected value is 0x%08x\n",
+                         addr, val, expected_val);
+               return ECORE_UNKNOWN_ERROR;
+       }
+
+       return ECORE_SUCCESS;
+}
+
 enum _ecore_status_t ecore_hw_stop(struct ecore_dev *p_dev)
 {
-       enum _ecore_status_t rc = ECORE_SUCCESS, t_rc;
+       struct ecore_hwfn *p_hwfn;
+       struct ecore_ptt *p_ptt;
+       enum _ecore_status_t rc, rc2 = ECORE_SUCCESS;
        int j;
 
        for_each_hwfn(p_dev, j) {
-               struct ecore_hwfn *p_hwfn = &p_dev->hwfns[j];
-               struct ecore_ptt *p_ptt = p_hwfn->p_main_ptt;
+               p_hwfn = &p_dev->hwfns[j];
+               p_ptt = p_hwfn->p_main_ptt;
 
                DP_VERBOSE(p_hwfn, ECORE_MSG_IFDOWN, "Stopping hw/fw\n");
 
                if (IS_VF(p_dev)) {
                        ecore_vf_pf_int_cleanup(p_hwfn);
+                       rc = ecore_vf_pf_reset(p_hwfn);
+                       if (rc != ECORE_SUCCESS) {
+                               DP_NOTICE(p_hwfn, true,
+                                         "ecore_vf_pf_reset failed. rc = %d.\n",
+                                         rc);
+                               rc2 = ECORE_UNKNOWN_ERROR;
+                       }
                        continue;
                }
 
                /* mark the hw as uninitialized... */
                p_hwfn->hw_init_done = false;
 
+               /* Send unload command to MCP */
+               if (!p_dev->recov_in_prog) {
+                       rc = ecore_mcp_unload_req(p_hwfn, p_ptt);
+                       if (rc != ECORE_SUCCESS) {
+                               DP_NOTICE(p_hwfn, true,
+                                         "Failed sending a UNLOAD_REQ command. rc = %d.\n",
+                                         rc);
+                               rc2 = ECORE_UNKNOWN_ERROR;
+                       }
+               }
+
+               OSAL_DPC_SYNC(p_hwfn);
+
+               /* After this point no MFW attentions are expected, e.g. prevent
+                * race between pf stop and dcbx pf update.
+                */
+
                rc = ecore_sp_pf_stop(p_hwfn);
-               if (rc)
+               if (rc != ECORE_SUCCESS) {
                        DP_NOTICE(p_hwfn, true,
-                                 "Failed to close PF against FW. Continue to stop HW to prevent illegal host access by the device\n");
+                                 "Failed to close PF against FW [rc = %d]. Continue to stop HW to prevent illegal host access by the device.\n",
+                                 rc);
+                       rc2 = ECORE_UNKNOWN_ERROR;
+               }
 
                /* perform debug action after PF stop was sent */
-               OSAL_AFTER_PF_STOP((void *)p_hwfn->p_dev, p_hwfn->my_id);
+               OSAL_AFTER_PF_STOP((void *)p_dev, p_hwfn->my_id);
 
                /* close NIG to BRB gate */
                ecore_wr(p_hwfn, p_ptt,
@@ -2205,20 +2236,48 @@ enum _ecore_status_t ecore_hw_stop(struct ecore_dev *p_dev)
                ecore_int_igu_init_pure_rt(p_hwfn, p_ptt, false, true);
                /* Need to wait 1ms to guarantee SBs are cleared */
                OSAL_MSLEEP(1);
-       }
+
+               if (!p_dev->recov_in_prog) {
+                       ecore_verify_reg_val(p_hwfn, p_ptt,
+                                            QM_REG_USG_CNT_PF_TX, 0);
+                       ecore_verify_reg_val(p_hwfn, p_ptt,
+                                            QM_REG_USG_CNT_PF_OTHER, 0);
+                       /* @@@TBD - assert on incorrect xCFC values (10.b) */
+               }
+
+               /* Disable PF in HW blocks */
+               ecore_wr(p_hwfn, p_ptt, DORQ_REG_PF_DB_ENABLE, 0);
+               ecore_wr(p_hwfn, p_ptt, QM_REG_PF_EN, 0);
+
+               if (!p_dev->recov_in_prog) {
+                       ecore_mcp_unload_done(p_hwfn, p_ptt);
+                       if (rc != ECORE_SUCCESS) {
+                               DP_NOTICE(p_hwfn, true,
+                                         "Failed sending a UNLOAD_DONE command. rc = %d.\n",
+                                         rc);
+                               rc2 = ECORE_UNKNOWN_ERROR;
+                       }
+               }
+       } /* hwfn loop */
 
        if (IS_PF(p_dev)) {
+               p_hwfn = ECORE_LEADING_HWFN(p_dev);
+               p_ptt = ECORE_LEADING_HWFN(p_dev)->p_main_ptt;
+
                /* Disable DMAE in PXP - in CMT, this should only be done for
                 * first hw-function, and only after all transactions have
                 * stopped for all active hw-functions.
                 */
-               t_rc = ecore_change_pci_hwfn(&p_dev->hwfns[0],
-                                            p_dev->hwfns[0].p_main_ptt, false);
-               if (t_rc != ECORE_SUCCESS)
-                       rc = t_rc;
+               rc = ecore_change_pci_hwfn(p_hwfn, p_ptt, false);
+               if (rc != ECORE_SUCCESS) {
+                       DP_NOTICE(p_hwfn, true,
+                                 "ecore_change_pci_hwfn failed. rc = %d.\n",
+                                 rc);
+                       rc2 = ECORE_UNKNOWN_ERROR;
+               }
        }
 
-       return rc;
+       return rc2;
 }
 
 void ecore_hw_stop_fastpath(struct ecore_dev *p_dev)
@@ -2279,84 +2338,6 @@ void ecore_hw_start_fastpath(struct ecore_hwfn *p_hwfn)
                 NIG_REG_RX_LLH_BRB_GATE_DNTFWD_PERPF, 0x0);
 }
 
-static enum _ecore_status_t ecore_reg_assert(struct ecore_hwfn *p_hwfn,
-                                            struct ecore_ptt *p_ptt, u32 reg,
-                                            bool expected)
-{
-       u32 assert_val = ecore_rd(p_hwfn, p_ptt, reg);
-
-       if (assert_val != expected) {
-               DP_NOTICE(p_hwfn, true, "Value at address 0x%08x != 0x%08x\n",
-                         reg, expected);
-               return ECORE_UNKNOWN_ERROR;
-       }
-
-       return 0;
-}
-
-enum _ecore_status_t ecore_hw_reset(struct ecore_dev *p_dev)
-{
-       enum _ecore_status_t rc = ECORE_SUCCESS;
-       u32 unload_resp, unload_param;
-       int i;
-
-       for_each_hwfn(p_dev, i) {
-               struct ecore_hwfn *p_hwfn = &p_dev->hwfns[i];
-
-               if (IS_VF(p_dev)) {
-                       rc = ecore_vf_pf_reset(p_hwfn);
-                       if (rc)
-                               return rc;
-                       continue;
-               }
-
-               DP_VERBOSE(p_hwfn, ECORE_MSG_IFDOWN, "Resetting hw/fw\n");
-
-               /* Check for incorrect states */
-               if (!p_dev->recov_in_prog) {
-                       ecore_reg_assert(p_hwfn, p_hwfn->p_main_ptt,
-                                        QM_REG_USG_CNT_PF_TX, 0);
-                       ecore_reg_assert(p_hwfn, p_hwfn->p_main_ptt,
-                                        QM_REG_USG_CNT_PF_OTHER, 0);
-                       /* @@@TBD - assert on incorrect xCFC values (10.b) */
-               }
-
-               /* Disable PF in HW blocks */
-               ecore_wr(p_hwfn, p_hwfn->p_main_ptt, DORQ_REG_PF_DB_ENABLE, 0);
-               ecore_wr(p_hwfn, p_hwfn->p_main_ptt, QM_REG_PF_EN, 0);
-
-               if (p_dev->recov_in_prog) {
-                       DP_VERBOSE(p_hwfn, ECORE_MSG_IFDOWN,
-                                  "Recovery is in progress -> skip sending unload_req/done\n");
-                       break;
-               }
-
-               /* Send unload command to MCP */
-               rc = ecore_mcp_cmd(p_hwfn, p_hwfn->p_main_ptt,
-                                  DRV_MSG_CODE_UNLOAD_REQ,
-                                  DRV_MB_PARAM_UNLOAD_WOL_MCP,
-                                  &unload_resp, &unload_param);
-               if (rc != ECORE_SUCCESS) {
-                       DP_NOTICE(p_hwfn, true,
-                                 "ecore_hw_reset: UNLOAD_REQ failed\n");
-                       /* @@TBD - what to do? for now, assume ENG. */
-                       unload_resp = FW_MSG_CODE_DRV_UNLOAD_ENGINE;
-               }
-
-               rc = ecore_mcp_cmd(p_hwfn, p_hwfn->p_main_ptt,
-                                  DRV_MSG_CODE_UNLOAD_DONE,
-                                  0, &unload_resp, &unload_param);
-               if (rc != ECORE_SUCCESS) {
-                       DP_NOTICE(p_hwfn,
-                                 true, "ecore_hw_reset: UNLOAD_DONE failed\n");
-                       /* @@@TBD - Should it really ASSERT here ? */
-                       return rc;
-               }
-       }
-
-       return rc;
-}
-
 /* Free hwfn memory and resources acquired in hw_hwfn_prepare */
 static void ecore_hw_hwfn_free(struct ecore_hwfn *p_hwfn)
 {
@@ -2452,173 +2433,203 @@ static void ecore_hw_set_feat(struct ecore_hwfn *p_hwfn)
                   RESC_NUM(p_hwfn, ECORE_SB));
 }
 
-static enum resource_id_enum
-ecore_hw_get_mfw_res_id(enum ecore_resources res_id)
+const char *ecore_hw_get_resc_name(enum ecore_resources res_id)
 {
-       enum resource_id_enum mfw_res_id = RESOURCE_NUM_INVALID;
-
        switch (res_id) {
        case ECORE_SB:
-               mfw_res_id = RESOURCE_NUM_SB_E;
-               break;
+               return "SB";
        case ECORE_L2_QUEUE:
-               mfw_res_id = RESOURCE_NUM_L2_QUEUE_E;
-               break;
+               return "L2_QUEUE";
        case ECORE_VPORT:
-               mfw_res_id = RESOURCE_NUM_VPORT_E;
-               break;
+               return "VPORT";
        case ECORE_RSS_ENG:
-               mfw_res_id = RESOURCE_NUM_RSS_ENGINES_E;
-               break;
+               return "RSS_ENG";
        case ECORE_PQ:
-               mfw_res_id = RESOURCE_NUM_PQ_E;
-               break;
+               return "PQ";
        case ECORE_RL:
-               mfw_res_id = RESOURCE_NUM_RL_E;
-               break;
+               return "RL";
        case ECORE_MAC:
+               return "MAC";
        case ECORE_VLAN:
-               /* Each VFC resource can accommodate both a MAC and a VLAN */
-               mfw_res_id = RESOURCE_VFC_FILTER_E;
-               break;
+               return "VLAN";
+       case ECORE_RDMA_CNQ_RAM:
+               return "RDMA_CNQ_RAM";
        case ECORE_ILT:
-               mfw_res_id = RESOURCE_ILT_E;
-               break;
+               return "ILT";
        case ECORE_LL2_QUEUE:
-               mfw_res_id = RESOURCE_LL2_QUEUE_E;
-               break;
-       case ECORE_RDMA_CNQ_RAM:
+               return "LL2_QUEUE";
        case ECORE_CMDQS_CQS:
-               /* CNQ/CMDQS are the same resource */
-               mfw_res_id = RESOURCE_CQS_E;
-               break;
+               return "CMDQS_CQS";
        case ECORE_RDMA_STATS_QUEUE:
-               mfw_res_id = RESOURCE_RDMA_STATS_QUEUE_E;
-               break;
+               return "RDMA_STATS_QUEUE";
+       case ECORE_BDQ:
+               return "BDQ";
        default:
-               break;
+               return "UNKNOWN_RESOURCE";
+       }
+}
+
+static enum _ecore_status_t
+__ecore_hw_set_soft_resc_size(struct ecore_hwfn *p_hwfn,
+                             enum ecore_resources res_id, u32 resc_max_val,
+                             u32 *p_mcp_resp)
+{
+       enum _ecore_status_t rc;
+
+       rc = ecore_mcp_set_resc_max_val(p_hwfn, p_hwfn->p_main_ptt, res_id,
+                                       resc_max_val, p_mcp_resp);
+       if (rc != ECORE_SUCCESS) {
+               DP_NOTICE(p_hwfn, true,
+                         "MFW response failure for a max value setting of resource %d [%s]\n",
+                         res_id, ecore_hw_get_resc_name(res_id));
+               return rc;
+       }
+
+       if (*p_mcp_resp != FW_MSG_CODE_RESOURCE_ALLOC_OK)
+               DP_INFO(p_hwfn,
+                       "Failed to set the max value of resource %d [%s]. mcp_resp = 0x%08x.\n",
+                       res_id, ecore_hw_get_resc_name(res_id), *p_mcp_resp);
+
+       return ECORE_SUCCESS;
+}
+
+static enum _ecore_status_t
+ecore_hw_set_soft_resc_size(struct ecore_hwfn *p_hwfn)
+{
+       bool b_ah = ECORE_IS_AH(p_hwfn->p_dev);
+       u32 resc_max_val, mcp_resp;
+       u8 res_id;
+       enum _ecore_status_t rc;
+
+       for (res_id = 0; res_id < ECORE_MAX_RESC; res_id++) {
+               /* @DPDK */
+               switch (res_id) {
+               case ECORE_LL2_QUEUE:
+               case ECORE_RDMA_CNQ_RAM:
+               case ECORE_RDMA_STATS_QUEUE:
+               case ECORE_BDQ:
+                       resc_max_val = 0;
+                       break;
+               default:
+                       continue;
+               }
+
+               rc = __ecore_hw_set_soft_resc_size(p_hwfn, res_id,
+                                                  resc_max_val, &mcp_resp);
+               if (rc != ECORE_SUCCESS)
+                       return rc;
+
+               /* There's no point to continue to the next resource if the
+                * command is not supported by the MFW.
+                * We do continue if the command is supported but the resource
+                * is unknown to the MFW. Such a resource will be later
+                * configured with the default allocation values.
+                */
+               if (mcp_resp == FW_MSG_CODE_UNSUPPORTED)
+                       return ECORE_NOTIMPL;
        }
 
-       return mfw_res_id;
+       return ECORE_SUCCESS;
 }
 
-static u32 ecore_hw_get_dflt_resc_num(struct ecore_hwfn *p_hwfn,
-                                     enum ecore_resources res_id)
+static
+enum _ecore_status_t ecore_hw_get_dflt_resc(struct ecore_hwfn *p_hwfn,
+                                           enum ecore_resources res_id,
+                                           u32 *p_resc_num, u32 *p_resc_start)
 {
        u8 num_funcs = p_hwfn->num_funcs_on_engine;
        bool b_ah = ECORE_IS_AH(p_hwfn->p_dev);
        struct ecore_sb_cnt_info sb_cnt_info;
-       u32 dflt_resc_num = 0;
 
        switch (res_id) {
        case ECORE_SB:
                OSAL_MEM_ZERO(&sb_cnt_info, sizeof(sb_cnt_info));
                ecore_int_get_num_sbs(p_hwfn, &sb_cnt_info);
-               dflt_resc_num = sb_cnt_info.sb_cnt;
+               *p_resc_num = sb_cnt_info.sb_cnt;
                break;
        case ECORE_L2_QUEUE:
-               dflt_resc_num = (b_ah ? MAX_NUM_L2_QUEUES_K2 :
+               *p_resc_num = (b_ah ? MAX_NUM_L2_QUEUES_K2 :
                                 MAX_NUM_L2_QUEUES_BB) / num_funcs;
                break;
        case ECORE_VPORT:
-               dflt_resc_num = (b_ah ? MAX_NUM_VPORTS_K2 :
+               *p_resc_num = (b_ah ? MAX_NUM_VPORTS_K2 :
                                 MAX_NUM_VPORTS_BB) / num_funcs;
                break;
        case ECORE_RSS_ENG:
-               dflt_resc_num = (b_ah ? ETH_RSS_ENGINE_NUM_K2 :
+               *p_resc_num = (b_ah ? ETH_RSS_ENGINE_NUM_K2 :
                                 ETH_RSS_ENGINE_NUM_BB) / num_funcs;
                break;
        case ECORE_PQ:
-               dflt_resc_num = (b_ah ? MAX_QM_TX_QUEUES_K2 :
+               *p_resc_num = (b_ah ? MAX_QM_TX_QUEUES_K2 :
                                 MAX_QM_TX_QUEUES_BB) / num_funcs;
                break;
        case ECORE_RL:
-               dflt_resc_num = MAX_QM_GLOBAL_RLS / num_funcs;
+               *p_resc_num = MAX_QM_GLOBAL_RLS / num_funcs;
                break;
        case ECORE_MAC:
        case ECORE_VLAN:
                /* Each VFC resource can accommodate both a MAC and a VLAN */
-               dflt_resc_num = ETH_NUM_MAC_FILTERS / num_funcs;
+               *p_resc_num = ETH_NUM_MAC_FILTERS / num_funcs;
                break;
        case ECORE_ILT:
-               dflt_resc_num = (b_ah ? PXP_NUM_ILT_RECORDS_K2 :
+               *p_resc_num = (b_ah ? PXP_NUM_ILT_RECORDS_K2 :
                                 PXP_NUM_ILT_RECORDS_BB) / num_funcs;
                break;
        case ECORE_LL2_QUEUE:
-               dflt_resc_num = MAX_NUM_LL2_RX_QUEUES / num_funcs;
+               *p_resc_num = MAX_NUM_LL2_RX_QUEUES / num_funcs;
                break;
        case ECORE_RDMA_CNQ_RAM:
        case ECORE_CMDQS_CQS:
                /* CNQ/CMDQS are the same resource */
                /* @DPDK */
-               dflt_resc_num = (NUM_OF_GLOBAL_QUEUES / 2) / num_funcs;
+               *p_resc_num = (NUM_OF_GLOBAL_QUEUES / 2) / num_funcs;
                break;
        case ECORE_RDMA_STATS_QUEUE:
                /* @DPDK */
-               dflt_resc_num = (b_ah ? MAX_NUM_VPORTS_K2 :
+               *p_resc_num = (b_ah ? MAX_NUM_VPORTS_K2 :
                                 MAX_NUM_VPORTS_BB) / num_funcs;
                break;
+       case ECORE_BDQ:
+               /* @DPDK */
+               *p_resc_num = 0;
+               break;
        default:
                break;
        }
 
-       return dflt_resc_num;
-}
 
-static const char *ecore_hw_get_resc_name(enum ecore_resources res_id)
-{
        switch (res_id) {
-       case ECORE_SB:
-               return "SB";
-       case ECORE_L2_QUEUE:
-               return "L2_QUEUE";
-       case ECORE_VPORT:
-               return "VPORT";
-       case ECORE_RSS_ENG:
-               return "RSS_ENG";
-       case ECORE_PQ:
-               return "PQ";
-       case ECORE_RL:
-               return "RL";
-       case ECORE_MAC:
-               return "MAC";
-       case ECORE_VLAN:
-               return "VLAN";
-       case ECORE_RDMA_CNQ_RAM:
-               return "RDMA_CNQ_RAM";
-       case ECORE_ILT:
-               return "ILT";
-       case ECORE_LL2_QUEUE:
-               return "LL2_QUEUE";
-       case ECORE_CMDQS_CQS:
-               return "CMDQS_CQS";
-       case ECORE_RDMA_STATS_QUEUE:
-               return "RDMA_STATS_QUEUE";
+       case ECORE_BDQ:
+               if (!*p_resc_num)
+                       *p_resc_start = 0;
+               break;
        default:
-               return "UNKNOWN_RESOURCE";
+               *p_resc_start = *p_resc_num * p_hwfn->enabled_func_idx;
+               break;
        }
+
+       return ECORE_SUCCESS;
 }
 
-static enum _ecore_status_t ecore_hw_set_resc_info(struct ecore_hwfn *p_hwfn,
-                                                  enum ecore_resources res_id,
-                                                  bool drv_resc_alloc)
+static enum _ecore_status_t
+__ecore_hw_set_resc_info(struct ecore_hwfn *p_hwfn, enum ecore_resources res_id,
+                        bool drv_resc_alloc)
 {
-       u32 dflt_resc_num = 0, dflt_resc_start = 0, mcp_resp, mcp_param;
-       u32 *p_resc_num, *p_resc_start;
-       struct resource_info resc_info;
+       u32 dflt_resc_num = 0, dflt_resc_start = 0;
+       u32 mcp_resp, *p_resc_num, *p_resc_start;
        enum _ecore_status_t rc;
 
        p_resc_num = &RESC_NUM(p_hwfn, res_id);
        p_resc_start = &RESC_START(p_hwfn, res_id);
 
-       dflt_resc_num = ecore_hw_get_dflt_resc_num(p_hwfn, res_id);
-       if (!dflt_resc_num) {
+       rc = ecore_hw_get_dflt_resc(p_hwfn, res_id, &dflt_resc_num,
+                                   &dflt_resc_start);
+       if (rc != ECORE_SUCCESS) {
                DP_ERR(p_hwfn,
                       "Failed to get default amount for resource %d [%s]\n",
                        res_id, ecore_hw_get_resc_name(res_id));
-               return ECORE_INVAL;
+               return rc;
        }
-       dflt_resc_start = dflt_resc_num * p_hwfn->enabled_func_idx;
 
 #ifndef ASIC_ONLY
        if (CHIP_REV_IS_SLOW(p_hwfn->p_dev)) {
@@ -2628,17 +2639,8 @@ static enum _ecore_status_t ecore_hw_set_resc_info(struct ecore_hwfn *p_hwfn,
        }
 #endif
 
-       OSAL_MEM_ZERO(&resc_info, sizeof(resc_info));
-       resc_info.res_id = ecore_hw_get_mfw_res_id(res_id);
-       if (resc_info.res_id == RESOURCE_NUM_INVALID) {
-               DP_ERR(p_hwfn,
-                      "Failed to match resource %d with MFW resources\n",
-                      res_id);
-               return ECORE_INVAL;
-       }
-
-       rc = ecore_mcp_get_resc_info(p_hwfn, p_hwfn->p_main_ptt, &resc_info,
-                                    &mcp_resp, &mcp_param);
+       rc = ecore_mcp_get_resc_info(p_hwfn, p_hwfn->p_main_ptt, res_id,
+                                    &mcp_resp, p_resc_num, p_resc_start);
        if (rc != ECORE_SUCCESS) {
                DP_NOTICE(p_hwfn, true,
                          "MFW response failure for an allocation request for"
@@ -2652,13 +2654,11 @@ static enum _ecore_status_t ecore_hw_set_resc_info(struct ecore_hwfn *p_hwfn,
         * - There is an internal error in the MFW while processing the request
         * - The resource ID is unknown to the MFW
         */
-       if (mcp_resp != FW_MSG_CODE_RESOURCE_ALLOC_OK &&
-           mcp_resp != FW_MSG_CODE_RESOURCE_ALLOC_DEPRECATED) {
-               /* @DPDK */
+       if (mcp_resp != FW_MSG_CODE_RESOURCE_ALLOC_OK) {
                DP_INFO(p_hwfn,
-                       "Resource %d [%s]: No allocation info was received"
-                       " [mcp_resp 0x%x]. Applying default values"
-                       " [num %d, start %d].\n",
+                       "Failed to receive allocation info for resource %d [%s]."
+                       " mcp_resp = 0x%x. Applying default values"
+                       " [%d,%d].\n",
                        res_id, ecore_hw_get_resc_name(res_id), mcp_resp,
                        dflt_resc_num, dflt_resc_start);
 
@@ -2670,16 +2670,13 @@ static enum _ecore_status_t ecore_hw_set_resc_info(struct ecore_hwfn *p_hwfn,
        /* TBD - remove this when revising the handling of the SB resource */
        if (res_id == ECORE_SB) {
                /* Excluding the slowpath SB */
-               resc_info.size -= 1;
-               resc_info.offset -= p_hwfn->enabled_func_idx;
+               *p_resc_num -= 1;
+               *p_resc_start -= p_hwfn->enabled_func_idx;
        }
 
-       *p_resc_num = resc_info.size;
-       *p_resc_start = resc_info.offset;
-
        if (*p_resc_num != dflt_resc_num || *p_resc_start != dflt_resc_start) {
                DP_INFO(p_hwfn,
-                       "Resource %d [%s]: MFW allocation [num %d, start %d] differs from default values [num %d, start %d]%s\n",
+                       "MFW allocation for resource %d [%s] differs from default values [%d,%d vs. %d,%d]%s\n",
                        res_id, ecore_hw_get_resc_name(res_id), *p_resc_num,
                        *p_resc_start, dflt_resc_num, dflt_resc_start,
                        drv_resc_alloc ? " - Applying default values" : "");
@@ -2692,12 +2689,32 @@ out:
        return ECORE_SUCCESS;
 }
 
+static enum _ecore_status_t ecore_hw_set_resc_info(struct ecore_hwfn *p_hwfn,
+                                                  bool drv_resc_alloc)
+{
+       enum _ecore_status_t rc;
+       u8 res_id;
+
+       for (res_id = 0; res_id < ECORE_MAX_RESC; res_id++) {
+               rc = __ecore_hw_set_resc_info(p_hwfn, res_id, drv_resc_alloc);
+               if (rc != ECORE_SUCCESS)
+                       return rc;
+       }
+
+       return ECORE_SUCCESS;
+}
+
+#define ECORE_RESC_ALLOC_LOCK_RETRY_CNT                10
+#define ECORE_RESC_ALLOC_LOCK_RETRY_INTVL_US   10000 /* 10 msec */
+
 static enum _ecore_status_t ecore_hw_get_resc(struct ecore_hwfn *p_hwfn,
                                              bool drv_resc_alloc)
 {
+       struct ecore_resc_unlock_params resc_unlock_params;
+       struct ecore_resc_lock_params resc_lock_params;
        bool b_ah = ECORE_IS_AH(p_hwfn->p_dev);
-       enum _ecore_status_t rc;
        u8 res_id;
+       enum _ecore_status_t rc;
 #ifndef ASIC_ONLY
        u32 *resc_start = p_hwfn->hw_info.resc_start;
        u32 *resc_num = p_hwfn->hw_info.resc_num;
@@ -2710,10 +2727,62 @@ static enum _ecore_status_t ecore_hw_get_resc(struct ecore_hwfn *p_hwfn,
        u32 roce_min_ilt_lines = PXP_NUM_ILT_RECORDS_BB / MAX_NUM_PFS_BB;
 #endif
 
-       for (res_id = 0; res_id < ECORE_MAX_RESC; res_id++) {
-               rc = ecore_hw_set_resc_info(p_hwfn, res_id, drv_resc_alloc);
+       /* Setting the max values of the soft resources and the following
+        * resources allocation queries should be atomic. Since several PFs can
+        * run in parallel - a resource lock is needed.
+        * If either the resource lock or resource set value commands are not
+        * supported - skip the the max values setting, release the lock if
+        * needed, and proceed to the queries. Other failures, including a
+        * failure to acquire the lock, will cause this function to fail.
+        * Old drivers that don't acquire the lock can run in parallel, and
+        * their allocation values won't be affected by the updated max values.
+        */
+       OSAL_MEM_ZERO(&resc_lock_params, sizeof(resc_lock_params));
+       resc_lock_params.resource = ECORE_RESC_LOCK_RESC_ALLOC;
+       resc_lock_params.retry_num = ECORE_RESC_ALLOC_LOCK_RETRY_CNT;
+       resc_lock_params.retry_interval = ECORE_RESC_ALLOC_LOCK_RETRY_INTVL_US;
+       resc_lock_params.sleep_b4_retry = true;
+       OSAL_MEM_ZERO(&resc_unlock_params, sizeof(resc_unlock_params));
+       resc_unlock_params.resource = ECORE_RESC_LOCK_RESC_ALLOC;
+
+       rc = ecore_mcp_resc_lock(p_hwfn, p_hwfn->p_main_ptt, &resc_lock_params);
+       if (rc != ECORE_SUCCESS && rc != ECORE_NOTIMPL) {
+               return rc;
+       } else if (rc == ECORE_NOTIMPL) {
+               DP_INFO(p_hwfn,
+                       "Skip the max values setting of the soft resources since the resource lock is not supported by the MFW\n");
+       } else if (rc == ECORE_SUCCESS && !resc_lock_params.b_granted) {
+               DP_NOTICE(p_hwfn, false,
+                         "Failed to acquire the resource lock for the resource allocation commands\n");
+               rc = ECORE_BUSY;
+               goto unlock_and_exit;
+       } else {
+               rc = ecore_hw_set_soft_resc_size(p_hwfn);
+               if (rc != ECORE_SUCCESS && rc != ECORE_NOTIMPL) {
+                       DP_NOTICE(p_hwfn, false,
+                                 "Failed to set the max values of the soft resources\n");
+                       goto unlock_and_exit;
+               } else if (rc == ECORE_NOTIMPL) {
+                       DP_INFO(p_hwfn,
+                               "Skip the max values setting of the soft resources since it is not supported by the MFW\n");
+                       rc = ecore_mcp_resc_unlock(p_hwfn, p_hwfn->p_main_ptt,
+                                                  &resc_unlock_params);
+                       if (rc != ECORE_SUCCESS)
+                               DP_INFO(p_hwfn,
+                                       "Failed to release the resource lock for the resource allocation commands\n");
+               }
+       }
+
+       rc = ecore_hw_set_resc_info(p_hwfn, drv_resc_alloc);
+       if (rc != ECORE_SUCCESS)
+               goto unlock_and_exit;
+
+       if (resc_lock_params.b_granted && !resc_unlock_params.b_released) {
+               rc = ecore_mcp_resc_unlock(p_hwfn, p_hwfn->p_main_ptt,
+                                          &resc_unlock_params);
                if (rc != ECORE_SUCCESS)
-                       return rc;
+                       DP_INFO(p_hwfn,
+                               "Failed to release the resource lock for the resource allocation commands\n");
        }
 
 #ifndef ASIC_ONLY
@@ -2766,6 +2835,10 @@ static enum _ecore_status_t ecore_hw_get_resc(struct ecore_hwfn *p_hwfn,
                           RESC_START(p_hwfn, res_id));
 
        return ECORE_SUCCESS;
+
+unlock_and_exit:
+       ecore_mcp_resc_unlock(p_hwfn, p_hwfn->p_main_ptt, &resc_unlock_params);
+       return rc;
 }
 
 static enum _ecore_status_t
@@ -3619,13 +3692,13 @@ static void ecore_chain_free_pbl(struct ecore_dev *p_dev,
                                 struct ecore_chain *p_chain)
 {
        void **pp_virt_addr_tbl = p_chain->pbl.pp_virt_addr_tbl;
-       u8 *p_pbl_virt = (u8 *)p_chain->pbl.p_virt_table;
+       u8 *p_pbl_virt = (u8 *)p_chain->pbl_sp.p_virt_table;
        u32 page_cnt = p_chain->page_cnt, i, pbl_size;
 
        if (!pp_virt_addr_tbl)
                return;
 
-       if (!p_chain->pbl.p_virt_table)
+       if (!p_pbl_virt)
                goto out;
 
        for (i = 0; i < page_cnt; i++) {
@@ -3641,10 +3714,10 @@ static void ecore_chain_free_pbl(struct ecore_dev *p_dev,
 
        pbl_size = page_cnt * ECORE_CHAIN_PBL_ENTRY_SIZE;
 
-       if (!p_chain->pbl.external)
-               OSAL_DMA_FREE_COHERENT(p_dev, p_chain->pbl.p_virt_table,
-                                      p_chain->pbl.p_phys_table, pbl_size);
-out:
+       if (!p_chain->b_external_pbl)
+               OSAL_DMA_FREE_COHERENT(p_dev, p_chain->pbl_sp.p_virt_table,
+                                      p_chain->pbl_sp.p_phys_table, pbl_size);
+ out:
        OSAL_VFREE(p_dev, p_chain->pbl.pp_virt_addr_tbl);
 }
 
@@ -3755,13 +3828,12 @@ ecore_chain_alloc_pbl(struct ecore_dev *p_dev,
        u32 page_cnt = p_chain->page_cnt, size, i;
 
        size = page_cnt * sizeof(*pp_virt_addr_tbl);
-       pp_virt_addr_tbl = (void **)OSAL_VALLOC(p_dev, size);
+       pp_virt_addr_tbl = (void **)OSAL_VZALLOC(p_dev, size);
        if (!pp_virt_addr_tbl) {
                DP_NOTICE(p_dev, true,
                          "Failed to allocate memory for the chain virtual addresses table\n");
                return ECORE_NOMEM;
        }
-       OSAL_MEM_ZERO(pp_virt_addr_tbl, size);
 
        /* The allocation of the PBL table is done with its full size, since it
         * is expected to be successive.
@@ -3776,7 +3848,7 @@ ecore_chain_alloc_pbl(struct ecore_dev *p_dev,
        } else {
                p_pbl_virt = ext_pbl->p_pbl_virt;
                p_pbl_phys = ext_pbl->p_pbl_phys;
-               p_chain->pbl.external = true;
+               p_chain->b_external_pbl = true;
        }
 
        ecore_chain_init_pbl_mem(p_chain, p_pbl_virt, p_pbl_phys,
@@ -4232,11 +4304,6 @@ static enum _ecore_status_t ecore_set_coalesce(struct ecore_hwfn *p_hwfn,
 {
        struct coalescing_timeset *p_coal_timeset;
 
-       if (IS_VF(p_hwfn->p_dev)) {
-               DP_NOTICE(p_hwfn, true, "VF coalescing config not supported\n");
-               return ECORE_INVAL;
-       }
-
        if (p_hwfn->p_dev->int_coalescing_mode != ECORE_COAL_MODE_ENABLE) {
                DP_NOTICE(p_hwfn, true,
                          "Coalescing configuration not enabled\n");
@@ -4252,13 +4319,53 @@ static enum _ecore_status_t ecore_set_coalesce(struct ecore_hwfn *p_hwfn,
        return ECORE_SUCCESS;
 }
 
+enum _ecore_status_t ecore_set_queue_coalesce(struct ecore_hwfn *p_hwfn,
+                                             u16 rx_coal, u16 tx_coal,
+                                             void *p_handle)
+{
+       struct ecore_queue_cid *p_cid = (struct ecore_queue_cid *)p_handle;
+       enum _ecore_status_t rc = ECORE_SUCCESS;
+       struct ecore_ptt *p_ptt;
+
+       /* TODO - Configuring a single queue's coalescing but
+        * claiming all queues are abiding same configuration
+        * for PF and VF both.
+        */
+
+       if (IS_VF(p_hwfn->p_dev))
+               return ecore_vf_pf_set_coalesce(p_hwfn, rx_coal,
+                                               tx_coal, p_cid);
+
+       p_ptt = ecore_ptt_acquire(p_hwfn);
+       if (!p_ptt)
+               return ECORE_AGAIN;
+
+       if (rx_coal) {
+               rc = ecore_set_rxq_coalesce(p_hwfn, p_ptt, rx_coal, p_cid);
+               if (rc)
+                       goto out;
+               p_hwfn->p_dev->rx_coalesce_usecs = rx_coal;
+       }
+
+       if (tx_coal) {
+               rc = ecore_set_txq_coalesce(p_hwfn, p_ptt, tx_coal, p_cid);
+               if (rc)
+                       goto out;
+               p_hwfn->p_dev->tx_coalesce_usecs = tx_coal;
+       }
+out:
+       ecore_ptt_release(p_hwfn, p_ptt);
+
+       return rc;
+}
+
 enum _ecore_status_t ecore_set_rxq_coalesce(struct ecore_hwfn *p_hwfn,
                                            struct ecore_ptt *p_ptt,
-                                           u16 coalesce, u16 qid, u16 sb_id)
+                                           u16 coalesce,
+                                           struct ecore_queue_cid *p_cid)
 {
        struct ustorm_eth_queue_zone eth_qzone;
        u8 timeset, timer_res;
-       u16 fw_qid = 0;
        u32 address;
        enum _ecore_status_t rc;
 
@@ -4275,33 +4382,30 @@ enum _ecore_status_t ecore_set_rxq_coalesce(struct ecore_hwfn *p_hwfn,
        }
        timeset = (u8)(coalesce >> timer_res);
 
-       rc = ecore_fw_l2_queue(p_hwfn, qid, &fw_qid);
-       if (rc != ECORE_SUCCESS)
-               return rc;
-
-       rc = ecore_int_set_timer_res(p_hwfn, p_ptt, timer_res, sb_id, false);
+       rc = ecore_int_set_timer_res(p_hwfn, p_ptt, timer_res,
+                                    p_cid->abs.sb_idx, false);
        if (rc != ECORE_SUCCESS)
                goto out;
 
-       address = BAR0_MAP_REG_USDM_RAM + USTORM_ETH_QUEUE_ZONE_OFFSET(fw_qid);
+       address = BAR0_MAP_REG_USDM_RAM +
+                 USTORM_ETH_QUEUE_ZONE_OFFSET(p_cid->abs.queue_id);
 
        rc = ecore_set_coalesce(p_hwfn, p_ptt, address, &eth_qzone,
                                sizeof(struct ustorm_eth_queue_zone), timeset);
        if (rc != ECORE_SUCCESS)
                goto out;
 
-       p_hwfn->p_dev->rx_coalesce_usecs = coalesce;
-out:
+ out:
        return rc;
 }
 
 enum _ecore_status_t ecore_set_txq_coalesce(struct ecore_hwfn *p_hwfn,
                                            struct ecore_ptt *p_ptt,
-                                           u16 coalesce, u16 qid, u16 sb_id)
+                                           u16 coalesce,
+                                           struct ecore_queue_cid *p_cid)
 {
        struct xstorm_eth_queue_zone eth_qzone;
        u8 timeset, timer_res;
-       u16 fw_qid = 0;
        u32 address;
        enum _ecore_status_t rc;
 
@@ -4319,23 +4423,17 @@ enum _ecore_status_t ecore_set_txq_coalesce(struct ecore_hwfn *p_hwfn,
 
        timeset = (u8)(coalesce >> timer_res);
 
-       rc = ecore_fw_l2_queue(p_hwfn, qid, &fw_qid);
-       if (rc != ECORE_SUCCESS)
-               return rc;
-
-       rc = ecore_int_set_timer_res(p_hwfn, p_ptt, timer_res, sb_id, true);
+       rc = ecore_int_set_timer_res(p_hwfn, p_ptt, timer_res,
+                                    p_cid->abs.sb_idx, true);
        if (rc != ECORE_SUCCESS)
                goto out;
 
-       address = BAR0_MAP_REG_XSDM_RAM + XSTORM_ETH_QUEUE_ZONE_OFFSET(fw_qid);
+       address = BAR0_MAP_REG_XSDM_RAM +
+                 XSTORM_ETH_QUEUE_ZONE_OFFSET(p_cid->abs.queue_id);
 
        rc = ecore_set_coalesce(p_hwfn, p_ptt, address, &eth_qzone,
                                sizeof(struct xstorm_eth_queue_zone), timeset);
-       if (rc != ECORE_SUCCESS)
-               goto out;
-
-       p_hwfn->p_dev->tx_coalesce_usecs = coalesce;
-out:
+ out:
        return rc;
 }