dev_info->max_rx_queues = ETH_MEMIF_MAX_NUM_Q_PAIRS;
dev_info->max_tx_queues = ETH_MEMIF_MAX_NUM_Q_PAIRS;
dev_info->min_rx_bufsize = 0;
+ dev_info->tx_offload_capa = DEV_TX_OFFLOAD_MULTI_SEGS;
return 0;
}
if (mbuf != mbuf_head)
rte_pktmbuf_pkt_len(mbuf_head) += cp_len;
- memcpy(rte_pktmbuf_mtod_offset(mbuf, void *, dst_off),
- (uint8_t *)memif_get_buffer(proc_private, d0) + src_off,
- cp_len);
+ rte_memcpy(rte_pktmbuf_mtod_offset(mbuf, void *,
+ dst_off),
+ (uint8_t *)memif_get_buffer(proc_private, d0) +
+ src_off, cp_len);
src_off += cp_len;
dst_off += cp_len;
rte_eth_devices[mq->in_port].process_private;
memif_ring_t *ring = memif_get_ring_from_queue(proc_private, mq);
uint16_t slot, saved_slot, n_free, ring_size, mask, n_tx_pkts = 0;
- uint16_t src_len, src_off, dst_len, dst_off, cp_len;
+ uint16_t src_len, src_off, dst_len, dst_off, cp_len, nb_segs;
memif_ring_type_t type = mq->type;
memif_desc_t *d0;
struct rte_mbuf *mbuf;
while (n_tx_pkts < nb_pkts && n_free) {
mbuf_head = *bufs++;
+ nb_segs = mbuf_head->nb_segs;
mbuf = mbuf_head;
saved_slot = slot;
}
cp_len = RTE_MIN(dst_len, src_len);
- memcpy((uint8_t *)memif_get_buffer(proc_private, d0) + dst_off,
- rte_pktmbuf_mtod_offset(mbuf, void *, src_off),
- cp_len);
+ rte_memcpy((uint8_t *)memif_get_buffer(proc_private,
+ d0) + dst_off,
+ rte_pktmbuf_mtod_offset(mbuf, void *, src_off),
+ cp_len);
mq->n_bytes += cp_len;
src_off += cp_len;
d0->length = dst_off;
}
- if (rte_pktmbuf_is_contiguous(mbuf) == 0) {
+ if (--nb_segs > 0) {
mbuf = mbuf->next;
goto next_in_chain;
}
uint16_t slot, uint16_t n_free)
{
memif_desc_t *d0;
+ uint16_t nb_segs = mbuf->nb_segs;
int used_slots = 1;
next_in_chain:
d0->flags = 0;
/* check if buffer is chained */
- if (rte_pktmbuf_is_contiguous(mbuf) == 0) {
+ if (--nb_segs > 0) {
if (n_free < 2)
return 0;
/* mark buffer as chained */