vhost: introduce async enqueue registration API
[dpdk.git] / lib / librte_vhost / socket.c
index 0a66ef9..c4626d2 100644 (file)
@@ -42,6 +42,7 @@ struct vhost_user_socket {
        bool use_builtin_virtio_net;
        bool extbuf;
        bool linearbuf;
+       bool async_copy;
 
        /*
         * The "supported_features" indicates the feature bits the
@@ -55,12 +56,7 @@ struct vhost_user_socket {
 
        uint64_t protocol_features;
 
-       /*
-        * Device id to identify a specific backend device.
-        * It's set to -1 for the default software implementation.
-        * If valid, one socket can have 1 connection only.
-        */
-       int vdpa_dev_id;
+       struct rte_vdpa_device *vdpa_dev;
 
        struct vhost_device_ops const *notify_ops;
 };
@@ -210,6 +206,7 @@ vhost_user_add_connection(int fd, struct vhost_user_socket *vsocket)
        size_t size;
        struct vhost_user_connection *conn;
        int ret;
+       struct virtio_net *dev;
 
        if (vsocket == NULL)
                return;
@@ -230,7 +227,7 @@ vhost_user_add_connection(int fd, struct vhost_user_socket *vsocket)
 
        vhost_set_builtin_virtio_net(vid, vsocket->use_builtin_virtio_net);
 
-       vhost_attach_vdpa_device(vid, vsocket->vdpa_dev_id);
+       vhost_attach_vdpa_device(vid, vsocket->vdpa_dev);
 
        if (vsocket->dequeue_zero_copy)
                vhost_enable_dequeue_zero_copy(vid);
@@ -241,6 +238,13 @@ vhost_user_add_connection(int fd, struct vhost_user_socket *vsocket)
        if (vsocket->linearbuf)
                vhost_enable_linearbuf(vid);
 
+       if (vsocket->async_copy) {
+               dev = get_device(vid);
+
+               if (dev)
+                       dev->async_copy = 1;
+       }
+
        VHOST_LOG_CONFIG(INFO, "new device, handle is %d\n", vid);
 
        if (vsocket->notify_ops->new_connection) {
@@ -578,17 +582,18 @@ find_vhost_user_socket(const char *path)
 }
 
 int
-rte_vhost_driver_attach_vdpa_device(const char *path, int did)
+rte_vhost_driver_attach_vdpa_device(const char *path,
+               struct rte_vdpa_device *dev)
 {
        struct vhost_user_socket *vsocket;
 
-       if (rte_vdpa_get_device(did) == NULL || path == NULL)
+       if (dev == NULL || path == NULL)
                return -1;
 
        pthread_mutex_lock(&vhost_user.mutex);
        vsocket = find_vhost_user_socket(path);
        if (vsocket)
-               vsocket->vdpa_dev_id = did;
+               vsocket->vdpa_dev = dev;
        pthread_mutex_unlock(&vhost_user.mutex);
 
        return vsocket ? 0 : -1;
@@ -602,25 +607,25 @@ rte_vhost_driver_detach_vdpa_device(const char *path)
        pthread_mutex_lock(&vhost_user.mutex);
        vsocket = find_vhost_user_socket(path);
        if (vsocket)
-               vsocket->vdpa_dev_id = -1;
+               vsocket->vdpa_dev = NULL;
        pthread_mutex_unlock(&vhost_user.mutex);
 
        return vsocket ? 0 : -1;
 }
 
-int
-rte_vhost_driver_get_vdpa_device_id(const char *path)
+struct rte_vdpa_device *
+rte_vhost_driver_get_vdpa_device(const char *path)
 {
        struct vhost_user_socket *vsocket;
-       int did = -1;
+       struct rte_vdpa_device *dev = NULL;
 
        pthread_mutex_lock(&vhost_user.mutex);
        vsocket = find_vhost_user_socket(path);
        if (vsocket)
-               did = vsocket->vdpa_dev_id;
+               dev = vsocket->vdpa_dev;
        pthread_mutex_unlock(&vhost_user.mutex);
 
-       return did;
+       return dev;
 }
 
 int
@@ -693,7 +698,6 @@ rte_vhost_driver_get_features(const char *path, uint64_t *features)
        struct vhost_user_socket *vsocket;
        uint64_t vdpa_features;
        struct rte_vdpa_device *vdpa_dev;
-       int did = -1;
        int ret = 0;
 
        pthread_mutex_lock(&vhost_user.mutex);
@@ -705,14 +709,13 @@ rte_vhost_driver_get_features(const char *path, uint64_t *features)
                goto unlock_exit;
        }
 
-       did = vsocket->vdpa_dev_id;
-       vdpa_dev = rte_vdpa_get_device(did);
+       vdpa_dev = vsocket->vdpa_dev;
        if (!vdpa_dev || !vdpa_dev->ops->get_features) {
                *features = vsocket->features;
                goto unlock_exit;
        }
 
-       if (vdpa_dev->ops->get_features(did, &vdpa_features) < 0) {
+       if (vdpa_dev->ops->get_features(vdpa_dev, &vdpa_features) < 0) {
                VHOST_LOG_CONFIG(ERR,
                                "failed to get vdpa features "
                                "for socket file %s.\n", path);
@@ -748,7 +751,6 @@ rte_vhost_driver_get_protocol_features(const char *path,
        struct vhost_user_socket *vsocket;
        uint64_t vdpa_protocol_features;
        struct rte_vdpa_device *vdpa_dev;
-       int did = -1;
        int ret = 0;
 
        pthread_mutex_lock(&vhost_user.mutex);
@@ -760,14 +762,13 @@ rte_vhost_driver_get_protocol_features(const char *path,
                goto unlock_exit;
        }
 
-       did = vsocket->vdpa_dev_id;
-       vdpa_dev = rte_vdpa_get_device(did);
+       vdpa_dev = vsocket->vdpa_dev;
        if (!vdpa_dev || !vdpa_dev->ops->get_protocol_features) {
                *protocol_features = vsocket->protocol_features;
                goto unlock_exit;
        }
 
-       if (vdpa_dev->ops->get_protocol_features(did,
+       if (vdpa_dev->ops->get_protocol_features(vdpa_dev,
                                &vdpa_protocol_features) < 0) {
                VHOST_LOG_CONFIG(ERR,
                                "failed to get vdpa protocol features "
@@ -790,7 +791,6 @@ rte_vhost_driver_get_queue_num(const char *path, uint32_t *queue_num)
        struct vhost_user_socket *vsocket;
        uint32_t vdpa_queue_num;
        struct rte_vdpa_device *vdpa_dev;
-       int did = -1;
        int ret = 0;
 
        pthread_mutex_lock(&vhost_user.mutex);
@@ -802,14 +802,13 @@ rte_vhost_driver_get_queue_num(const char *path, uint32_t *queue_num)
                goto unlock_exit;
        }
 
-       did = vsocket->vdpa_dev_id;
-       vdpa_dev = rte_vdpa_get_device(did);
+       vdpa_dev = vsocket->vdpa_dev;
        if (!vdpa_dev || !vdpa_dev->ops->get_queue_num) {
                *queue_num = VHOST_MAX_QUEUE_PAIRS;
                goto unlock_exit;
        }
 
-       if (vdpa_dev->ops->get_queue_num(did, &vdpa_queue_num) < 0) {
+       if (vdpa_dev->ops->get_queue_num(vdpa_dev, &vdpa_queue_num) < 0) {
                VHOST_LOG_CONFIG(ERR,
                                "failed to get vdpa queue number "
                                "for socket file %s.\n", path);
@@ -878,7 +877,7 @@ rte_vhost_driver_register(const char *path, uint64_t flags)
                        "error: failed to init connection mutex\n");
                goto out_free;
        }
-       vsocket->vdpa_dev_id = -1;
+       vsocket->vdpa_dev = NULL;
        vsocket->dequeue_zero_copy = flags & RTE_VHOST_USER_DEQUEUE_ZERO_COPY;
        vsocket->extbuf = flags & RTE_VHOST_USER_EXTBUF_SUPPORT;
        vsocket->linearbuf = flags & RTE_VHOST_USER_LINEARBUF_SUPPORT;
@@ -891,6 +890,17 @@ rte_vhost_driver_register(const char *path, uint64_t flags)
                goto out_mutex;
        }
 
+       vsocket->async_copy = flags & RTE_VHOST_USER_ASYNC_COPY;
+
+       if (vsocket->async_copy &&
+               (flags & (RTE_VHOST_USER_IOMMU_SUPPORT |
+               RTE_VHOST_USER_POSTCOPY_SUPPORT))) {
+               VHOST_LOG_CONFIG(ERR, "error: enabling async copy and IOMMU "
+                       "or post-copy feature simultaneously is not "
+                       "supported\n");
+               goto out_mutex;
+       }
+
        /*
         * Set the supported features correctly for the builtin vhost-user
         * net driver.
@@ -941,6 +951,13 @@ rte_vhost_driver_register(const char *path, uint64_t flags)
                        ~(1ULL << VHOST_USER_PROTOCOL_F_PAGEFAULT);
        }
 
+       if (vsocket->async_copy) {
+               vsocket->supported_features &= ~(1ULL << VHOST_F_LOG_ALL);
+               vsocket->features &= ~(1ULL << VHOST_F_LOG_ALL);
+               VHOST_LOG_CONFIG(INFO,
+                       "Logging feature is disabled in async copy mode\n");
+       }
+
        /*
         * We'll not be able to receive a buffer from guest in linear mode
         * without external buffer if it will not fit in a single mbuf, which is