From: Chen Jing D(Mark) Date: Thu, 14 Aug 2014 07:35:01 +0000 (+0800) Subject: i40e: queue start and stop X-Git-Tag: spdx-start~10464 X-Git-Url: http://git.droids-corp.org/?a=commitdiff_plain;h=e406719449cb73b5ae6b9fb3b01ea67d9f62651f;p=dpdk.git i40e: queue start and stop Add functions to start/stop specific RX/TX queue. Support fields start_rx_per_q and start_tx_per_q. In the meanwhile, change dev_start/stop functions to call per-queue functions. Signed-off-by: Chen Jing D(Mark) Reviewed-by: Konstantin Ananyev Reviewed-by: Changchun Ouyang Reviewed-by: Huawei Xie [Thomas: reword comments and merge 2 patches] --- diff --git a/lib/librte_pmd_i40e/i40e_ethdev.c b/lib/librte_pmd_i40e/i40e_ethdev.c index f5a3ae7bf2..4e65ca4a16 100644 --- a/lib/librte_pmd_i40e/i40e_ethdev.c +++ b/lib/librte_pmd_i40e/i40e_ethdev.c @@ -236,6 +236,10 @@ static struct eth_dev_ops i40e_eth_dev_ops = { .vlan_offload_set = i40e_vlan_offload_set, .vlan_strip_queue_set = i40e_vlan_strip_queue_set, .vlan_pvid_set = i40e_vlan_pvid_set, + .rx_queue_start = i40e_dev_rx_queue_start, + .rx_queue_stop = i40e_dev_rx_queue_stop, + .tx_queue_start = i40e_dev_tx_queue_start, + .tx_queue_stop = i40e_dev_tx_queue_stop, .rx_queue_setup = i40e_dev_rx_queue_setup, .rx_queue_release = i40e_dev_rx_queue_release, .rx_queue_count = i40e_dev_rx_queue_count, @@ -801,7 +805,6 @@ i40e_dev_start(struct rte_eth_dev *dev) err_up: i40e_vsi_switch_queues(vsi, FALSE); - i40e_dev_clear_queues(dev); return ret; } @@ -815,9 +818,6 @@ i40e_dev_stop(struct rte_eth_dev *dev) /* Disable all queues */ i40e_vsi_switch_queues(vsi, FALSE); - /* Clear all queues and release memory */ - i40e_dev_clear_queues(dev); - /* Set link down */ i40e_dev_set_link_down(dev); @@ -3005,17 +3005,21 @@ static int i40e_vsi_switch_tx_queues(struct i40e_vsi *vsi, bool on) { struct rte_eth_dev_data *dev_data = I40E_VSI_TO_DEV_DATA(vsi); - struct i40e_hw *hw = I40E_VSI_TO_HW(vsi); struct i40e_tx_queue *txq; - uint16_t i, pf_q; + struct rte_eth_dev *dev = I40E_VSI_TO_ETH_DEV(vsi); + uint16_t i; int ret; - pf_q = vsi->base_queue; - for (i = 0; i < dev_data->nb_tx_queues; i++, pf_q++) { + for (i = 0; i < dev_data->nb_tx_queues; i++) { txq = dev_data->tx_queues[i]; - if (!txq->q_set) - continue; /* Queue not configured */ - ret = i40e_switch_tx_queue(hw, pf_q, on); + /* Don't operate the queue if not configured or + * if starting only per queue */ + if (!txq->q_set || (on && txq->start_tx_per_q)) + continue; + if (on) + ret = i40e_dev_tx_queue_start(dev, i); + else + ret = i40e_dev_tx_queue_stop(dev, i); if ( ret != I40E_SUCCESS) return ret; } @@ -3079,18 +3083,22 @@ static int i40e_vsi_switch_rx_queues(struct i40e_vsi *vsi, bool on) { struct rte_eth_dev_data *dev_data = I40E_VSI_TO_DEV_DATA(vsi); - struct i40e_hw *hw = I40E_VSI_TO_HW(vsi); struct i40e_rx_queue *rxq; - uint16_t i, pf_q; + struct rte_eth_dev *dev = I40E_VSI_TO_ETH_DEV(vsi); + uint16_t i; int ret; - pf_q = vsi->base_queue; - for (i = 0; i < dev_data->nb_rx_queues; i++, pf_q++) { + for (i = 0; i < dev_data->nb_rx_queues; i++) { rxq = dev_data->rx_queues[i]; - if (!rxq->q_set) - continue; /* Queue not configured */ - ret = i40e_switch_rx_queue(hw, pf_q, on); - if ( ret != I40E_SUCCESS) + /* Don't operate the queue if not configured or + * if starting only per queue */ + if (!rxq->q_set || (on && rxq->start_rx_per_q)) + continue; + if (on) + ret = i40e_dev_rx_queue_start(dev, i); + else + ret = i40e_dev_rx_queue_stop(dev, i); + if (ret != I40E_SUCCESS) return ret; } diff --git a/lib/librte_pmd_i40e/i40e_rxtx.c b/lib/librte_pmd_i40e/i40e_rxtx.c index 83b9462732..f15384483c 100644 --- a/lib/librte_pmd_i40e/i40e_rxtx.c +++ b/lib/librte_pmd_i40e/i40e_rxtx.c @@ -88,9 +88,6 @@ i40e_ring_dma_zone_reserve(struct rte_eth_dev *dev, uint16_t queue_id, uint32_t ring_size, int socket_id); -static void i40e_reset_rx_queue(struct i40e_rx_queue *rxq); -static void i40e_reset_tx_queue(struct i40e_tx_queue *txq); -static void i40e_tx_queue_release_mbufs(struct i40e_tx_queue *txq); static uint16_t i40e_xmit_pkts_simple(void *tx_queue, struct rte_mbuf **tx_pkts, uint16_t nb_pkts); @@ -1428,6 +1425,118 @@ i40e_xmit_pkts_simple(void *tx_queue, return nb_tx; } +int +i40e_dev_rx_queue_start(struct rte_eth_dev *dev, uint16_t rx_queue_id) +{ + struct i40e_vsi *vsi = I40E_DEV_PRIVATE_TO_VSI(dev->data->dev_private); + struct i40e_rx_queue *rxq; + int err = -1; + struct i40e_hw *hw = I40E_VSI_TO_HW(vsi); + uint16_t q_base = vsi->base_queue; + + PMD_INIT_FUNC_TRACE(); + + if (rx_queue_id < dev->data->nb_rx_queues) { + rxq = dev->data->rx_queues[rx_queue_id]; + + err = i40e_alloc_rx_queue_mbufs(rxq); + if (err) { + PMD_DRV_LOG(ERR, "Failed to allocate RX queue mbuf\n"); + return err; + } + + rte_wmb(); + + /* Init the RX tail regieter. */ + I40E_PCI_REG_WRITE(rxq->qrx_tail, rxq->nb_rx_desc - 1); + + err = i40e_switch_rx_queue(hw, rx_queue_id + q_base, TRUE); + + if (err) { + PMD_DRV_LOG(ERR, "Failed to switch RX queue %u on\n", + rx_queue_id); + + i40e_rx_queue_release_mbufs(rxq); + i40e_reset_rx_queue(rxq); + } + } + + return err; +} + +int +i40e_dev_rx_queue_stop(struct rte_eth_dev *dev, uint16_t rx_queue_id) +{ + struct i40e_vsi *vsi = I40E_DEV_PRIVATE_TO_VSI(dev->data->dev_private); + struct i40e_rx_queue *rxq; + int err; + struct i40e_hw *hw = I40E_VSI_TO_HW(vsi); + uint16_t q_base = vsi->base_queue; + + if (rx_queue_id < dev->data->nb_rx_queues) { + rxq = dev->data->rx_queues[rx_queue_id]; + + err = i40e_switch_rx_queue(hw, rx_queue_id + q_base, FALSE); + + if (err) { + PMD_DRV_LOG(ERR, "Failed to switch RX queue %u off\n", + rx_queue_id); + return err; + } + i40e_rx_queue_release_mbufs(rxq); + i40e_reset_rx_queue(rxq); + } + + return 0; +} + +int +i40e_dev_tx_queue_start(struct rte_eth_dev *dev, uint16_t tx_queue_id) +{ + struct i40e_vsi *vsi = I40E_DEV_PRIVATE_TO_VSI(dev->data->dev_private); + int err = -1; + struct i40e_hw *hw = I40E_VSI_TO_HW(vsi); + uint16_t q_base = vsi->base_queue; + + PMD_INIT_FUNC_TRACE(); + + if (tx_queue_id < dev->data->nb_tx_queues) { + err = i40e_switch_tx_queue(hw, tx_queue_id + q_base, TRUE); + if (err) + PMD_DRV_LOG(ERR, "Failed to switch TX queue %u on\n", + tx_queue_id); + } + + return err; +} + +int +i40e_dev_tx_queue_stop(struct rte_eth_dev *dev, uint16_t tx_queue_id) +{ + struct i40e_vsi *vsi = I40E_DEV_PRIVATE_TO_VSI(dev->data->dev_private); + struct i40e_tx_queue *txq; + int err; + struct i40e_hw *hw = I40E_VSI_TO_HW(vsi); + uint16_t q_base = vsi->base_queue; + + if (tx_queue_id < dev->data->nb_tx_queues) { + txq = dev->data->tx_queues[tx_queue_id]; + + err = i40e_switch_tx_queue(hw, tx_queue_id + q_base, FALSE); + + if (err) { + PMD_DRV_LOG(ERR, "Failed to switch TX queue %u of\n", + tx_queue_id); + return err; + } + + i40e_tx_queue_release_mbufs(txq); + i40e_reset_tx_queue(txq); + } + + return 0; +} + int i40e_dev_rx_queue_setup(struct rte_eth_dev *dev, uint16_t queue_idx, @@ -1482,6 +1591,7 @@ i40e_dev_rx_queue_setup(struct rte_eth_dev *dev, 0 : ETHER_CRC_LEN); rxq->drop_en = rx_conf->rx_drop_en; rxq->vsi = vsi; + rxq->start_rx_per_q = rx_conf->start_rx_per_q; /* Allocate the maximun number of RX ring hardware descriptor. */ ring_size = sizeof(union i40e_rx_desc) * I40E_MAX_RING_DESC; @@ -1767,6 +1877,7 @@ i40e_dev_tx_queue_setup(struct rte_eth_dev *dev, txq->port_id = dev->data->port_id; txq->txq_flags = tx_conf->txq_flags; txq->vsi = vsi; + txq->start_tx_per_q = tx_conf->start_tx_per_q; #ifdef RTE_LIBRTE_XEN_DOM0 txq->tx_ring_phys_addr = rte_mem_phy2mch(tz->memseg_id, tz->phys_addr); @@ -1874,7 +1985,7 @@ i40e_rx_queue_release_mbufs(struct i40e_rx_queue *rxq) #endif /* RTE_LIBRTE_I40E_RX_ALLOW_BULK_ALLOC */ } -static void +void i40e_reset_rx_queue(struct i40e_rx_queue *rxq) { unsigned i; @@ -1905,7 +2016,7 @@ i40e_reset_rx_queue(struct i40e_rx_queue *rxq) rxq->pkt_last_seg = NULL; } -static void +void i40e_tx_queue_release_mbufs(struct i40e_tx_queue *txq) { uint16_t i; @@ -1923,7 +2034,7 @@ i40e_tx_queue_release_mbufs(struct i40e_tx_queue *txq) } } -static void +void i40e_reset_tx_queue(struct i40e_tx_queue *txq) { struct i40e_tx_entry *txe; @@ -2161,7 +2272,7 @@ i40e_rx_queue_init(struct i40e_rx_queue *rxq) } rxq->qrx_tail = hw->hw_addr + I40E_QRX_TAIL(pf_q); - err = i40e_alloc_rx_queue_mbufs(rxq); + mbp_priv = rte_mempool_get_priv(rxq->mp); buf_size = (uint16_t)(mbp_priv->mbuf_data_room_size - RTE_PKTMBUF_HEADROOM); @@ -2172,16 +2283,10 @@ i40e_rx_queue_init(struct i40e_rx_queue *rxq) dev->rx_pkt_burst = i40e_recv_scattered_pkts; } - rte_wmb(); - /* Init the RX tail regieter. */ - I40E_PCI_REG_WRITE(rxq->qrx_tail, 0); I40E_PCI_REG_WRITE(rxq->qrx_tail, rxq->nb_rx_desc - 1); - if (err) - PMD_DRV_LOG(ERR, "Failed to allocate RX queue mbuf\n"); - - return err; + return 0; } void diff --git a/lib/librte_pmd_i40e/i40e_rxtx.h b/lib/librte_pmd_i40e/i40e_rxtx.h index 6db2faf17b..4478592be6 100644 --- a/lib/librte_pmd_i40e/i40e_rxtx.h +++ b/lib/librte_pmd_i40e/i40e_rxtx.h @@ -112,6 +112,7 @@ struct i40e_rx_queue { uint16_t max_pkt_len; /* Maximum packet length */ uint8_t hs_mode; /* Header Split mode */ bool q_set; /**< indicate if rx queue has been configured */ + bool start_rx_per_q; /**< don't start this queue in dev start */ }; struct i40e_tx_entry { @@ -150,8 +151,13 @@ struct i40e_tx_queue { uint16_t tx_next_dd; uint16_t tx_next_rs; bool q_set; /**< indicate if tx queue has been configured */ + bool start_tx_per_q; /**< don't start this queue in dev start */ }; +int i40e_dev_rx_queue_start(struct rte_eth_dev *dev, uint16_t rx_queue_id); +int i40e_dev_rx_queue_stop(struct rte_eth_dev *dev, uint16_t rx_queue_id); +int i40e_dev_tx_queue_start(struct rte_eth_dev *dev, uint16_t tx_queue_id); +int i40e_dev_tx_queue_stop(struct rte_eth_dev *dev, uint16_t tx_queue_id); int i40e_dev_rx_queue_setup(struct rte_eth_dev *dev, uint16_t queue_idx, uint16_t nb_desc, @@ -179,6 +185,9 @@ int i40e_rx_queue_init(struct i40e_rx_queue *rxq); void i40e_free_tx_resources(struct i40e_tx_queue *txq); void i40e_free_rx_resources(struct i40e_rx_queue *rxq); void i40e_dev_clear_queues(struct rte_eth_dev *dev); +void i40e_reset_rx_queue(struct i40e_rx_queue *rxq); +void i40e_reset_tx_queue(struct i40e_tx_queue *txq); +void i40e_tx_queue_release_mbufs(struct i40e_tx_queue *txq); int i40e_alloc_rx_queue_mbufs(struct i40e_rx_queue *rxq); void i40e_rx_queue_release_mbufs(struct i40e_rx_queue *rxq);