vhost: prepare async for mbuf to descriptor refactoring
authorMaxime Coquelin <maxime.coquelin@redhat.com>
Tue, 26 Oct 2021 16:29:01 +0000 (18:29 +0200)
committerMaxime Coquelin <maxime.coquelin@redhat.com>
Fri, 29 Oct 2021 10:32:30 +0000 (12:32 +0200)
This patch extracts the IO vectors filling from
async_mbuf_to_desc() into a dedicated function as a
preliminary step of merging copy_mubf_to_desc() and
async_mbuf_to_desc().

Signed-off-by: Maxime Coquelin <maxime.coquelin@redhat.com>
Reviewed-by: Jiayu Hu <jiayu.hu@intel.com>
lib/vhost/virtio_net.c

index 2f76523..a901e29 100644 (file)
@@ -791,6 +791,109 @@ copy_vnet_hdr_to_desc(struct virtio_net *dev, struct vhost_virtqueue *vq,
        }
 }
 
+static __rte_always_inline int
+async_iter_initialize(struct vhost_async *async)
+{
+       struct rte_vhost_iov_iter *iter;
+
+       if (unlikely(async->iovec_idx >= VHOST_MAX_ASYNC_VEC)) {
+               VHOST_LOG_DATA(ERR, "no more async iovec available\n");
+               return -1;
+       }
+
+       iter = async->iov_iter + async->iter_idx;
+       iter->iov = async->iovec + async->iovec_idx;
+       iter->nr_segs = 0;
+
+       return 0;
+}
+
+static __rte_always_inline int
+async_iter_add_iovec(struct vhost_async *async, void *src, void *dst, size_t len)
+{
+       struct rte_vhost_iov_iter *iter;
+       struct rte_vhost_iovec *iovec;
+
+       if (unlikely(async->iovec_idx >= VHOST_MAX_ASYNC_VEC)) {
+               static bool vhost_max_async_vec_log;
+
+               if (!vhost_max_async_vec_log) {
+                       VHOST_LOG_DATA(ERR, "no more async iovec available\n");
+                       vhost_max_async_vec_log = true;
+               }
+
+               return -1;
+       }
+
+       iter = async->iov_iter + async->iter_idx;
+       iovec = async->iovec + async->iovec_idx;
+
+       iovec->src_addr = src;
+       iovec->dst_addr = dst;
+       iovec->len = len;
+
+       iter->nr_segs++;
+       async->iovec_idx++;
+
+       return 0;
+}
+
+static __rte_always_inline void
+async_iter_finalize(struct vhost_async *async)
+{
+       async->iter_idx++;
+}
+
+static __rte_always_inline void
+async_iter_cancel(struct vhost_async *async)
+{
+       struct rte_vhost_iov_iter *iter;
+
+       iter = async->iov_iter + async->iter_idx;
+       async->iovec_idx -= iter->nr_segs;
+       iter->nr_segs = 0;
+       iter->iov = NULL;
+}
+
+static __rte_always_inline void
+async_iter_reset(struct vhost_async *async)
+{
+       async->iter_idx = 0;
+       async->iovec_idx = 0;
+}
+
+static __rte_always_inline int
+async_mbuf_to_desc_seg(struct virtio_net *dev, struct vhost_virtqueue *vq,
+               struct rte_mbuf *m, uint32_t mbuf_offset,
+               uint64_t buf_iova, uint32_t cpy_len)
+{
+       struct vhost_async *async = vq->async;
+       uint64_t mapped_len;
+       uint32_t buf_offset = 0;
+       void *hpa;
+
+       while (cpy_len) {
+               hpa = (void *)(uintptr_t)gpa_to_first_hpa(dev,
+                               buf_iova + buf_offset, cpy_len, &mapped_len);
+               if (unlikely(!hpa)) {
+                       VHOST_LOG_DATA(ERR, "(%d) %s: failed to get hpa.\n", dev->vid, __func__);
+                       return -1;
+               }
+
+               if (unlikely(async_iter_add_iovec(async,
+                                               (void *)(uintptr_t)rte_pktmbuf_iova_offset(m,
+                                                       mbuf_offset),
+                                               hpa, (size_t)mapped_len)))
+                       return -1;
+
+               cpy_len -= (uint32_t)mapped_len;
+               mbuf_offset += (uint32_t)mapped_len;
+               buf_offset += (uint32_t)mapped_len;
+       }
+
+       return 0;
+}
+
 static __rte_always_inline int
 copy_mbuf_to_desc(struct virtio_net *dev, struct vhost_virtqueue *vq,
                            struct rte_mbuf *m, struct buf_vector *buf_vec,
@@ -924,77 +1027,6 @@ out:
        return error;
 }
 
-static __rte_always_inline int
-async_iter_initialize(struct vhost_async *async)
-{
-       struct rte_vhost_iov_iter *iter;
-
-       if (unlikely(async->iovec_idx >= VHOST_MAX_ASYNC_VEC)) {
-               VHOST_LOG_DATA(ERR, "no more async iovec available\n");
-               return -1;
-       }
-
-       iter = async->iov_iter + async->iter_idx;
-       iter->iov = async->iovec + async->iovec_idx;
-       iter->nr_segs = 0;
-
-       return 0;
-}
-
-static __rte_always_inline int
-async_iter_add_iovec(struct vhost_async *async, void *src, void *dst, size_t len)
-{
-       struct rte_vhost_iov_iter *iter;
-       struct rte_vhost_iovec *iovec;
-
-       if (unlikely(async->iovec_idx >= VHOST_MAX_ASYNC_VEC)) {
-               static bool vhost_max_async_vec_log;
-
-               if (!vhost_max_async_vec_log) {
-                       VHOST_LOG_DATA(ERR, "no more async iovec available\n");
-                       vhost_max_async_vec_log = true;
-               }
-
-               return -1;
-       }
-
-       iter = async->iov_iter + async->iter_idx;
-       iovec = async->iovec + async->iovec_idx;
-
-       iovec->src_addr = src;
-       iovec->dst_addr = dst;
-       iovec->len = len;
-
-       iter->nr_segs++;
-       async->iovec_idx++;
-
-       return 0;
-}
-
-static __rte_always_inline void
-async_iter_finalize(struct vhost_async *async)
-{
-       async->iter_idx++;
-}
-
-static __rte_always_inline void
-async_iter_cancel(struct vhost_async *async)
-{
-       struct rte_vhost_iov_iter *iter;
-
-       iter = async->iov_iter + async->iter_idx;
-       async->iovec_idx -= iter->nr_segs;
-       iter->nr_segs = 0;
-       iter->iov = NULL;
-}
-
-static __rte_always_inline void
-async_iter_reset(struct vhost_async *async)
-{
-       async->iter_idx = 0;
-       async->iovec_idx = 0;
-}
-
 static __rte_always_inline int
 async_mbuf_to_desc(struct virtio_net *dev, struct vhost_virtqueue *vq,
                        struct rte_mbuf *m, struct buf_vector *buf_vec,
@@ -1005,14 +1037,11 @@ async_mbuf_to_desc(struct virtio_net *dev, struct vhost_virtqueue *vq,
        struct virtio_net_hdr_mrg_rxbuf tmp_hdr, *hdr = NULL;
        uint64_t buf_addr, buf_iova;
        uint64_t hdr_addr;
-       uint64_t mapped_len;
        uint32_t vec_idx = 0;
        uint32_t mbuf_offset, mbuf_avail;
        uint32_t buf_offset, buf_avail;
        uint32_t cpy_len, buf_len;
 
-       void *hpa;
-
        if (unlikely(m == NULL))
                return -1;
 
@@ -1096,28 +1125,15 @@ async_mbuf_to_desc(struct virtio_net *dev, struct vhost_virtqueue *vq,
 
                cpy_len = RTE_MIN(buf_avail, mbuf_avail);
 
-               while (unlikely(cpy_len)) {
-                       hpa = (void *)(uintptr_t)gpa_to_first_hpa(dev,
-                                       buf_iova + buf_offset,
-                                       cpy_len, &mapped_len);
-                       if (unlikely(!hpa)) {
-                               VHOST_LOG_DATA(ERR, "(%d) %s: failed to get hpa.\n",
-                               dev->vid, __func__);
-                               goto error;
-                       }
-
-                       if (unlikely(async_iter_add_iovec(async,
-                                       (void *)(uintptr_t)rte_pktmbuf_iova_offset(m,
-                                               mbuf_offset),
-                                       hpa, (size_t)mapped_len)))
-                               goto error;
-
-                       cpy_len -= (uint32_t)mapped_len;
-                       mbuf_avail  -= (uint32_t)mapped_len;
-                       mbuf_offset += (uint32_t)mapped_len;
-                       buf_avail  -= (uint32_t)mapped_len;
-                       buf_offset += (uint32_t)mapped_len;
+               if (async_mbuf_to_desc_seg(dev, vq, m, mbuf_offset,
+                                       buf_iova + buf_offset, cpy_len) < 0) {
+                       goto error;
                }
+
+               mbuf_avail  -= cpy_len;
+               mbuf_offset += cpy_len;
+               buf_avail  -= cpy_len;
+               buf_offset += cpy_len;
        }
 
        async_iter_finalize(async);