X-Git-Url: http://git.droids-corp.org/?a=blobdiff_plain;f=lib%2Fvhost%2Fvirtio_net.c;h=b93482587c76ff225aad15321d4ef9e19b182721;hb=786963fdcf3eed83a74607e001b46089b2792c35;hp=8e36f4c34031ff2177c2293fcf09bee57c4d3418;hpb=ca7036b4af3a82d258cca914e71171434b3d0320;p=dpdk.git diff --git a/lib/vhost/virtio_net.c b/lib/vhost/virtio_net.c index 8e36f4c340..b93482587c 100644 --- a/lib/vhost/virtio_net.c +++ b/lib/vhost/virtio_net.c @@ -218,14 +218,17 @@ vhost_flush_enqueue_batch_packed(struct virtio_net *dev, { uint16_t i; uint16_t flags; - uint16_t last_used_idx = vq->last_used_idx; - struct vring_packed_desc *desc_base = &vq->desc_packed[last_used_idx]; + uint16_t last_used_idx; + struct vring_packed_desc *desc_base; if (vq->shadow_used_idx) { do_data_copy_enqueue(dev, vq); vhost_flush_enqueue_shadow_packed(dev, vq); } + last_used_idx = vq->last_used_idx; + desc_base = &vq->desc_packed[last_used_idx]; + flags = PACKED_DESC_ENQUEUE_USED_FLAG(vq->used_wrap_counter); vhost_for_each_try_unroll(i, 0, PACKED_BATCH_SIZE) { @@ -1474,7 +1477,7 @@ static __rte_always_inline void store_dma_desc_info_split(struct vring_used_elem *s_ring, struct vring_used_elem *d_ring, uint16_t ring_size, uint16_t s_idx, uint16_t d_idx, uint16_t count) { - uint16_t elem_size = sizeof(struct vring_used_elem); + size_t elem_size = sizeof(struct vring_used_elem); if (d_idx + count <= ring_size) { rte_memcpy(d_ring + d_idx, s_ring + s_idx, count * elem_size); @@ -1491,7 +1494,7 @@ store_dma_desc_info_packed(struct vring_used_elem_packed *s_ring, struct vring_used_elem_packed *d_ring, uint16_t ring_size, uint16_t s_idx, uint16_t d_idx, uint16_t count) { - uint16_t elem_size = sizeof(struct vring_used_elem_packed); + size_t elem_size = sizeof(struct vring_used_elem_packed); if (d_idx + count <= ring_size) { rte_memcpy(d_ring + d_idx, s_ring + s_idx, count * elem_size); @@ -2649,6 +2652,9 @@ virtio_dev_extbuf_alloc(struct rte_mbuf *pkt, uint32_t size) return 0; } +/* + * Prepare a host supported pktmbuf. + */ static __rte_always_inline int virtio_dev_pktmbuf_prep(struct virtio_net *dev, struct rte_mbuf *pkt, uint32_t data_len) @@ -2667,32 +2673,6 @@ virtio_dev_pktmbuf_prep(struct virtio_net *dev, struct rte_mbuf *pkt, return -1; } -/* - * Allocate a host supported pktmbuf. - */ -static __rte_always_inline struct rte_mbuf * -virtio_dev_pktmbuf_alloc(struct virtio_net *dev, struct rte_mempool *mp, - uint32_t data_len) -{ - struct rte_mbuf *pkt = rte_pktmbuf_alloc(mp); - - if (unlikely(pkt == NULL)) { - VHOST_LOG_DATA(ERR, - "Failed to allocate memory for mbuf.\n"); - return NULL; - } - - if (virtio_dev_pktmbuf_prep(dev, pkt, data_len)) { - /* Data doesn't fit into the buffer and the host supports - * only linear buffers - */ - rte_pktmbuf_free(pkt); - return NULL; - } - - return pkt; -} - __rte_always_inline static uint16_t virtio_dev_tx_split(struct virtio_net *dev, struct vhost_virtqueue *vq, @@ -2722,6 +2702,9 @@ virtio_dev_tx_split(struct virtio_net *dev, struct vhost_virtqueue *vq, VHOST_LOG_DATA(DEBUG, "(%d) about to dequeue %u buffers\n", dev->vid, count); + if (rte_pktmbuf_alloc_bulk(mbuf_pool, pkts, count)) + return 0; + for (i = 0; i < count; i++) { struct buf_vector buf_vec[BUF_VECTOR_MAX]; uint16_t head_idx; @@ -2738,8 +2721,8 @@ virtio_dev_tx_split(struct virtio_net *dev, struct vhost_virtqueue *vq, update_shadow_used_ring_split(vq, head_idx, 0); - pkts[i] = virtio_dev_pktmbuf_alloc(dev, mbuf_pool, buf_len); - if (unlikely(pkts[i] == NULL)) { + err = virtio_dev_pktmbuf_prep(dev, pkts[i], buf_len); + if (unlikely(err)) { /* * mbuf allocation fails for jumbo packets when external * buffer allocation is not allowed and linear buffer @@ -2759,7 +2742,6 @@ virtio_dev_tx_split(struct virtio_net *dev, struct vhost_virtqueue *vq, err = copy_desc_to_mbuf(dev, vq, buf_vec, nr_vec, pkts[i], mbuf_pool, legacy_ol_flags); if (unlikely(err)) { - rte_pktmbuf_free(pkts[i]); if (!allocerr_warned) { VHOST_LOG_DATA(ERR, "Failed to copy desc to mbuf on %s.\n", @@ -2772,6 +2754,9 @@ virtio_dev_tx_split(struct virtio_net *dev, struct vhost_virtqueue *vq, } } + if (dropped) + rte_pktmbuf_free_bulk(&pkts[i - 1], count - i + 1); + vq->last_avail_idx += i; do_data_copy_dequeue(vq);