/* SPDX-License-Identifier: BSD-3-Clause
- * Copyright(c) 2020 Hisilicon Limited.
+ * Copyright(c) 2020-2021 HiSilicon Limited.
*/
#include <rte_io.h>
{
struct hns3_rx_queue *rxq = rx_queue;
struct hns3_desc *rxdp = &rxq->rx_ring[rxq->next_to_use];
- uint64_t bd_err_mask; /* bit mask indicate whick pkts is error */
+ uint64_t pkt_err_mask; /* bit mask indicate whick pkts is error */
uint16_t nb_rx;
- nb_pkts = RTE_MIN(nb_pkts, HNS3_DEFAULT_RX_BURST);
- nb_pkts = RTE_ALIGN_FLOOR(nb_pkts, HNS3_DEFAULT_DESCS_PER_LOOP);
-
rte_prefetch_non_temporal(rxdp);
+ nb_pkts = RTE_ALIGN_FLOOR(nb_pkts, HNS3_DEFAULT_DESCS_PER_LOOP);
+
if (rxq->rx_rearm_nb > HNS3_DEFAULT_RXQ_REARM_THRESH)
hns3_rxq_rearm_mbuf(rxq);
rte_prefetch0(rxq->sw_ring[rxq->next_to_use + 2].mbuf);
rte_prefetch0(rxq->sw_ring[rxq->next_to_use + 3].mbuf);
- bd_err_mask = 0;
- nb_rx = hns3_recv_burst_vec(rxq, rx_pkts, nb_pkts, &bd_err_mask);
- if (unlikely(bd_err_mask))
- nb_rx = hns3_rx_reassemble_pkts(rx_pkts, nb_rx, bd_err_mask);
+ if (likely(nb_pkts <= HNS3_DEFAULT_RX_BURST)) {
+ pkt_err_mask = 0;
+ nb_rx = hns3_recv_burst_vec(rxq, rx_pkts, nb_pkts,
+ &pkt_err_mask);
+ nb_rx = hns3_rx_reassemble_pkts(rx_pkts, nb_rx, pkt_err_mask);
+ return nb_rx;
+ }
+
+ nb_rx = 0;
+ while (nb_pkts > 0) {
+ uint16_t ret, n;
+
+ n = RTE_MIN(nb_pkts, HNS3_DEFAULT_RX_BURST);
+ pkt_err_mask = 0;
+ ret = hns3_recv_burst_vec(rxq, &rx_pkts[nb_rx], n,
+ &pkt_err_mask);
+ nb_pkts -= ret;
+ nb_rx += hns3_rx_reassemble_pkts(&rx_pkts[nb_rx], ret,
+ pkt_err_mask);
+ if (ret < n)
+ break;
+
+ if (rxq->rx_rearm_nb > HNS3_DEFAULT_RXQ_REARM_THRESH)
+ hns3_rxq_rearm_mbuf(rxq);
+ }
return nb_rx;
}
mb_def.port = rxq->port_id;
rte_mbuf_refcnt_set(&mb_def, 1);
+ /* compile-time verifies the rearm_data first 8bytes */
+ RTE_BUILD_BUG_ON(offsetof(struct rte_mbuf, data_off) <
+ offsetof(struct rte_mbuf, rearm_data));
+ RTE_BUILD_BUG_ON(offsetof(struct rte_mbuf, refcnt) <
+ offsetof(struct rte_mbuf, rearm_data));
+ RTE_BUILD_BUG_ON(offsetof(struct rte_mbuf, nb_segs) <
+ offsetof(struct rte_mbuf, rearm_data));
+ RTE_BUILD_BUG_ON(offsetof(struct rte_mbuf, port) <
+ offsetof(struct rte_mbuf, rearm_data));
+ RTE_BUILD_BUG_ON(offsetof(struct rte_mbuf, data_off) -
+ offsetof(struct rte_mbuf, rearm_data) > 6);
+ RTE_BUILD_BUG_ON(offsetof(struct rte_mbuf, refcnt) -
+ offsetof(struct rte_mbuf, rearm_data) > 6);
+ RTE_BUILD_BUG_ON(offsetof(struct rte_mbuf, nb_segs) -
+ offsetof(struct rte_mbuf, rearm_data) > 6);
+ RTE_BUILD_BUG_ON(offsetof(struct rte_mbuf, port) -
+ offsetof(struct rte_mbuf, rearm_data) > 6);
+
/* prevent compiler reordering: rearm_data covers previous fields */
rte_compiler_barrier();
p = (uintptr_t)&mb_def.rearm_data;