eth_dev->rx_pkt_burst = virtio_recv_pkts_vec;
} else if (hw->use_inorder_rx) {
PMD_INIT_LOG(INFO,
- "virtio: using inorder mergeable buffer Rx path on port %u",
+ "virtio: using inorder Rx path on port %u",
eth_dev->data->port_id);
- eth_dev->rx_pkt_burst =
- &virtio_recv_mergeable_pkts_inorder;
+ eth_dev->rx_pkt_burst = &virtio_recv_pkts_inorder;
} else if (vtpci_with_feature(hw, VIRTIO_NET_F_MRG_RXBUF)) {
PMD_INIT_LOG(INFO,
"virtio: using mergeable buffer Rx path on port %u",
if (vtpci_with_feature(hw, VIRTIO_F_IN_ORDER)) {
hw->use_inorder_tx = 1;
- if (vtpci_with_feature(hw, VIRTIO_NET_F_MRG_RXBUF) &&
- !vtpci_packed_queue(hw)) {
- hw->use_inorder_rx = 1;
- hw->use_simple_rx = 0;
- } else {
- hw->use_inorder_rx = 0;
- }
+ hw->use_inorder_rx = 1;
+ hw->use_simple_rx = 0;
}
if (vtpci_packed_queue(hw)) {
uint16_t virtio_recv_mergeable_pkts_packed(void *rx_queue,
struct rte_mbuf **rx_pkts, uint16_t nb_pkts);
-uint16_t virtio_recv_mergeable_pkts_inorder(void *rx_queue,
+uint16_t virtio_recv_pkts_inorder(void *rx_queue,
struct rte_mbuf **rx_pkts, uint16_t nb_pkts);
uint16_t virtio_xmit_pkts(void *tx_queue, struct rte_mbuf **tx_pkts,
uint16_t
-virtio_recv_mergeable_pkts_inorder(void *rx_queue,
+virtio_recv_pkts_inorder(void *rx_queue,
struct rte_mbuf **rx_pkts,
uint16_t nb_pkts)
{
header = (struct virtio_net_hdr_mrg_rxbuf *)
((char *)rxm->buf_addr + RTE_PKTMBUF_HEADROOM
- hdr_size);
- seg_num = header->num_buffers;
- if (seg_num == 0)
+ if (vtpci_with_feature(hw, VIRTIO_NET_F_MRG_RXBUF)) {
+ seg_num = header->num_buffers;
+ if (seg_num == 0)
+ seg_num = 1;
+ } else {
seg_num = 1;
+ }
rxm->data_off = RTE_PKTMBUF_HEADROOM;
rxm->nb_segs = seg_num;