struct nix_bp_cfg_rsp *rsp;
int rc;
- if (otx2_dev_is_vf(dev))
- return 0;
-
if (enb) {
req = otx2_mbox_alloc_msg_nix_bp_enable(mbox);
req->chan_base = 0;
struct otx2_mbox *mbox = dev->mbox;
int rc;
- if (otx2_dev_is_vf(dev))
- return -ENOTSUP;
+ if (otx2_dev_is_lbk(dev)) {
+ fc_conf->mode = RTE_FC_NONE;
+ return 0;
+ }
req = otx2_mbox_alloc_msg_cgx_cfg_pause_frm(mbox);
req->set = 0;
if (enb) {
aq->cq.bpid = fc->bpid[0];
aq->cq_mask.bpid = ~(aq->cq_mask.bpid);
- aq->cq.bp = NIX_CQ_BP_LEVEL;
+ aq->cq.bp = rxq->cq_drop;
aq->cq_mask.bp = ~(aq->cq_mask.bp);
}
uint8_t tx_pause, rx_pause;
int rc = 0;
- if (otx2_dev_is_vf(dev))
+ if (otx2_dev_is_lbk(dev)) {
+ otx2_info("No flow control support for LBK bound ethports");
return -ENOTSUP;
+ }
if (fc_conf->high_water || fc_conf->low_water || fc_conf->pause_time ||
fc_conf->mac_ctrl_frame_fwd || fc_conf->autoneg) {
/* Check if TX pause frame is already enabled or not */
if (fc->tx_pause ^ tx_pause) {
- if (otx2_dev_is_A0(dev) && eth_dev->data->dev_started) {
- /* on A0, CQ should be in disabled state
+ if (otx2_dev_is_Ax(dev) && eth_dev->data->dev_started) {
+ /* on Ax, CQ should be in disabled state
* while setting flow control configuration.
*/
otx2_info("Stop the port=%d for setting flow control\n",
struct otx2_eth_dev *dev = otx2_eth_pmd_priv(eth_dev);
struct rte_eth_fc_conf fc_conf;
- if (otx2_dev_is_vf(dev))
+ if (otx2_dev_is_lbk(dev))
return 0;
memset(&fc_conf, 0, sizeof(struct rte_eth_fc_conf));
*/
otx2_nix_flow_ctrl_get(eth_dev, &fc_conf);
- /* To avoid Link credit deadlock on A0, disable Tx FC if it's enabled */
- if (otx2_dev_is_A0(dev) &&
+ /* To avoid Link credit deadlock on Ax, disable Tx FC if it's enabled */
+ if (otx2_dev_is_Ax(dev) &&
(fc_conf.mode == RTE_FC_FULL || fc_conf.mode == RTE_FC_RX_PAUSE)) {
fc_conf.mode =
(fc_conf.mode == RTE_FC_FULL ||