vhost: adapt library for selective datapath
authorZhihong Wang <zhihong.wang@intel.com>
Mon, 2 Apr 2018 11:46:55 +0000 (19:46 +0800)
committerFerruh Yigit <ferruh.yigit@intel.com>
Fri, 13 Apr 2018 22:40:21 +0000 (00:40 +0200)
This patch adapts vhost lib for selective datapath by calling device ops
at the corresponding stage.

Signed-off-by: Zhihong Wang <zhihong.wang@intel.com>
Reviewed-by: Maxime Coquelin <maxime.coquelin@redhat.com>
lib/librte_vhost/rte_vhost.h
lib/librte_vhost/rte_vhost_version.map
lib/librte_vhost/socket.c
lib/librte_vhost/vhost.c
lib/librte_vhost/vhost.h
lib/librte_vhost/vhost_user.c

index 8f35167..fe0338d 100644 (file)
@@ -289,6 +289,33 @@ int rte_vhost_driver_disable_features(const char *path, uint64_t features);
  */
 int rte_vhost_driver_get_features(const char *path, uint64_t *features);
 
+/**
+ * Get the protocol feature bits before feature negotiation.
+ *
+ * @param path
+ *  The vhost-user socket file path
+ * @param protocol_features
+ *  A pointer to store the queried protocol feature bits
+ * @return
+ *  0 on success, -1 on failure
+ */
+int __rte_experimental
+rte_vhost_driver_get_protocol_features(const char *path,
+               uint64_t *protocol_features);
+
+/**
+ * Get the queue number bits before feature negotiation.
+ *
+ * @param path
+ *  The vhost-user socket file path
+ * @param queue_num
+ *  A pointer to store the queried queue number bits
+ * @return
+ *  0 on success, -1 on failure
+ */
+int __rte_experimental
+rte_vhost_driver_get_queue_num(const char *path, uint32_t *queue_num);
+
 /**
  * Get the feature bits after negotiation
  *
index e30285d..55e0af7 100644 (file)
@@ -69,4 +69,6 @@ EXPERIMENTAL {
        rte_vhost_driver_detach_vdpa_device;
        rte_vhost_driver_get_vdpa_device_id;
        rte_vhost_get_vdpa_device_id;
+       rte_vhost_driver_get_protocol_features;
+       rte_vhost_driver_get_queue_num;
 } DPDK_18.02;
index b26bdb0..636fc25 100644 (file)
@@ -215,6 +215,8 @@ vhost_user_add_connection(int fd, struct vhost_user_socket *vsocket)
 
        vhost_set_builtin_virtio_net(vid, vsocket->use_builtin_virtio_net);
 
+       vhost_attach_vdpa_device(vid, vsocket->vdpa_dev_id);
+
        if (vsocket->dequeue_zero_copy)
                vhost_enable_dequeue_zero_copy(vid);
 
@@ -666,20 +668,123 @@ int
 rte_vhost_driver_get_features(const char *path, uint64_t *features)
 {
        struct vhost_user_socket *vsocket;
+       uint64_t vdpa_features;
+       struct rte_vdpa_device *vdpa_dev;
+       int did = -1;
+       int ret = 0;
 
        pthread_mutex_lock(&vhost_user.mutex);
        vsocket = find_vhost_user_socket(path);
-       if (vsocket)
+       if (!vsocket) {
+               RTE_LOG(ERR, VHOST_CONFIG,
+                       "socket file %s is not registered yet.\n", path);
+               ret = -1;
+               goto unlock_exit;
+       }
+
+       did = vsocket->vdpa_dev_id;
+       vdpa_dev = rte_vdpa_get_device(did);
+       if (!vdpa_dev || !vdpa_dev->ops->get_features) {
                *features = vsocket->features;
+               goto unlock_exit;
+       }
+
+       if (vdpa_dev->ops->get_features(did, &vdpa_features) < 0) {
+               RTE_LOG(ERR, VHOST_CONFIG,
+                               "failed to get vdpa features "
+                               "for socket file %s.\n", path);
+               ret = -1;
+               goto unlock_exit;
+       }
+
+       *features = vsocket->features & vdpa_features;
+
+unlock_exit:
        pthread_mutex_unlock(&vhost_user.mutex);
+       return ret;
+}
 
+int
+rte_vhost_driver_get_protocol_features(const char *path,
+               uint64_t *protocol_features)
+{
+       struct vhost_user_socket *vsocket;
+       uint64_t vdpa_protocol_features;
+       struct rte_vdpa_device *vdpa_dev;
+       int did = -1;
+       int ret = 0;
+
+       pthread_mutex_lock(&vhost_user.mutex);
+       vsocket = find_vhost_user_socket(path);
        if (!vsocket) {
                RTE_LOG(ERR, VHOST_CONFIG,
                        "socket file %s is not registered yet.\n", path);
-               return -1;
-       } else {
-               return 0;
+               ret = -1;
+               goto unlock_exit;
        }
+
+       did = vsocket->vdpa_dev_id;
+       vdpa_dev = rte_vdpa_get_device(did);
+       if (!vdpa_dev || !vdpa_dev->ops->get_protocol_features) {
+               *protocol_features = VHOST_USER_PROTOCOL_FEATURES;
+               goto unlock_exit;
+       }
+
+       if (vdpa_dev->ops->get_protocol_features(did,
+                               &vdpa_protocol_features) < 0) {
+               RTE_LOG(ERR, VHOST_CONFIG,
+                               "failed to get vdpa protocol features "
+                               "for socket file %s.\n", path);
+               ret = -1;
+               goto unlock_exit;
+       }
+
+       *protocol_features = VHOST_USER_PROTOCOL_FEATURES
+               & vdpa_protocol_features;
+
+unlock_exit:
+       pthread_mutex_unlock(&vhost_user.mutex);
+       return ret;
+}
+
+int
+rte_vhost_driver_get_queue_num(const char *path, uint32_t *queue_num)
+{
+       struct vhost_user_socket *vsocket;
+       uint32_t vdpa_queue_num;
+       struct rte_vdpa_device *vdpa_dev;
+       int did = -1;
+       int ret = 0;
+
+       pthread_mutex_lock(&vhost_user.mutex);
+       vsocket = find_vhost_user_socket(path);
+       if (!vsocket) {
+               RTE_LOG(ERR, VHOST_CONFIG,
+                       "socket file %s is not registered yet.\n", path);
+               ret = -1;
+               goto unlock_exit;
+       }
+
+       did = vsocket->vdpa_dev_id;
+       vdpa_dev = rte_vdpa_get_device(did);
+       if (!vdpa_dev || !vdpa_dev->ops->get_queue_num) {
+               *queue_num = VHOST_MAX_QUEUE_PAIRS;
+               goto unlock_exit;
+       }
+
+       if (vdpa_dev->ops->get_queue_num(did, &vdpa_queue_num) < 0) {
+               RTE_LOG(ERR, VHOST_CONFIG,
+                               "failed to get vdpa queue number "
+                               "for socket file %s.\n", path);
+               ret = -1;
+               goto unlock_exit;
+       }
+
+       *queue_num = RTE_MIN((uint32_t)VHOST_MAX_QUEUE_PAIRS, vdpa_queue_num);
+
+unlock_exit:
+       pthread_mutex_unlock(&vhost_user.mutex);
+       return ret;
 }
 
 /*
index 16b0f9a..b1afd69 100644 (file)
@@ -296,11 +296,17 @@ void
 vhost_destroy_device(int vid)
 {
        struct virtio_net *dev = get_device(vid);
+       struct rte_vdpa_device *vdpa_dev;
+       int did = -1;
 
        if (dev == NULL)
                return;
 
        if (dev->flags & VIRTIO_DEV_RUNNING) {
+               did = dev->vdpa_dev_id;
+               vdpa_dev = rte_vdpa_get_device(did);
+               if (vdpa_dev && vdpa_dev->ops->dev_close)
+                       vdpa_dev->ops->dev_close(dev->vid);
                dev->flags &= ~VIRTIO_DEV_RUNNING;
                dev->notify_ops->destroy_device(vid);
        }
index 453d9a3..c14a905 100644 (file)
@@ -28,6 +28,8 @@
 #define VIRTIO_DEV_READY 2
 /* Used to indicate that the built-in vhost net device backend is enabled */
 #define VIRTIO_DEV_BUILTIN_VIRTIO_NET 4
+/* Used to indicate that the device has its own data path and configured */
+#define VIRTIO_DEV_VDPA_CONFIGURED 8
 
 /* Backend value set by guest. */
 #define VIRTIO_DEV_STOPPED -1
index 66e1b82..157cf2f 100644 (file)
@@ -133,7 +133,14 @@ vhost_user_set_owner(void)
 static int
 vhost_user_reset_owner(struct virtio_net *dev)
 {
+       struct rte_vdpa_device *vdpa_dev;
+       int did = -1;
+
        if (dev->flags & VIRTIO_DEV_RUNNING) {
+               did = dev->vdpa_dev_id;
+               vdpa_dev = rte_vdpa_get_device(did);
+               if (vdpa_dev && vdpa_dev->ops->dev_close)
+                       vdpa_dev->ops->dev_close(dev->vid);
                dev->flags &= ~VIRTIO_DEV_RUNNING;
                dev->notify_ops->destroy_device(dev->vid);
        }
@@ -155,6 +162,18 @@ vhost_user_get_features(struct virtio_net *dev)
        return features;
 }
 
+/*
+ * The queue number that we support are requested.
+ */
+static uint32_t
+vhost_user_get_queue_num(struct virtio_net *dev)
+{
+       uint32_t queue_num = 0;
+
+       rte_vhost_driver_get_queue_num(dev->ifname, &queue_num);
+       return queue_num;
+}
+
 /*
  * We receive the negotiated features supported by us and the virtio device.
  */
@@ -162,6 +181,8 @@ static int
 vhost_user_set_features(struct virtio_net *dev, uint64_t features)
 {
        uint64_t vhost_features = 0;
+       struct rte_vdpa_device *vdpa_dev;
+       int did = -1;
 
        rte_vhost_driver_get_features(dev->ifname, &vhost_features);
        if (features & ~vhost_features) {
@@ -191,6 +212,11 @@ vhost_user_set_features(struct virtio_net *dev, uint64_t features)
                        dev->notify_ops->features_changed(dev->vid, features);
        }
 
+       did = dev->vdpa_dev_id;
+       vdpa_dev = rte_vdpa_get_device(did);
+       if (vdpa_dev && vdpa_dev->ops->set_features)
+               vdpa_dev->ops->set_features(dev->vid);
+
        dev->features = features;
        if (dev->features &
                ((1 << VIRTIO_NET_F_MRG_RXBUF) | (1ULL << VIRTIO_F_VERSION_1))) {
@@ -935,14 +961,21 @@ vhost_user_get_vring_base(struct virtio_net *dev,
                          VhostUserMsg *msg)
 {
        struct vhost_virtqueue *vq = dev->virtqueue[msg->payload.state.index];
+       struct rte_vdpa_device *vdpa_dev;
+       int did = -1;
 
        /* We have to stop the queue (virtio) if it is running. */
        if (dev->flags & VIRTIO_DEV_RUNNING) {
+               did = dev->vdpa_dev_id;
+               vdpa_dev = rte_vdpa_get_device(did);
+               if (vdpa_dev && vdpa_dev->ops->dev_close)
+                       vdpa_dev->ops->dev_close(dev->vid);
                dev->flags &= ~VIRTIO_DEV_RUNNING;
                dev->notify_ops->destroy_device(dev->vid);
        }
 
        dev->flags &= ~VIRTIO_DEV_READY;
+       dev->flags &= ~VIRTIO_DEV_VDPA_CONFIGURED;
 
        /* Here we are safe to get the last avail index */
        msg->payload.state.num = vq->last_avail_idx;
@@ -985,16 +1018,24 @@ vhost_user_set_vring_enable(struct virtio_net *dev,
                            VhostUserMsg *msg)
 {
        int enable = (int)msg->payload.state.num;
+       int index = (int)msg->payload.state.index;
+       struct rte_vdpa_device *vdpa_dev;
+       int did = -1;
 
        RTE_LOG(INFO, VHOST_CONFIG,
                "set queue enable: %d to qp idx: %d\n",
-               enable, msg->payload.state.index);
+               enable, index);
+
+       did = dev->vdpa_dev_id;
+       vdpa_dev = rte_vdpa_get_device(did);
+       if (vdpa_dev && vdpa_dev->ops->set_vring_state)
+               vdpa_dev->ops->set_vring_state(dev->vid, index, enable);
 
        if (dev->notify_ops->vring_state_changed)
                dev->notify_ops->vring_state_changed(dev->vid,
-                               msg->payload.state.index, enable);
+                               index, enable);
 
-       dev->virtqueue[msg->payload.state.index]->enabled = enable;
+       dev->virtqueue[index]->enabled = enable;
 
        return 0;
 }
@@ -1003,9 +1044,10 @@ static void
 vhost_user_get_protocol_features(struct virtio_net *dev,
                                 struct VhostUserMsg *msg)
 {
-       uint64_t features, protocol_features = VHOST_USER_PROTOCOL_FEATURES;
+       uint64_t features, protocol_features;
 
        rte_vhost_driver_get_features(dev->ifname, &features);
+       rte_vhost_driver_get_protocol_features(dev->ifname, &protocol_features);
 
        /*
         * REPLY_ACK protocol feature is only mandatory for now
@@ -1101,6 +1143,8 @@ static int
 vhost_user_send_rarp(struct virtio_net *dev, struct VhostUserMsg *msg)
 {
        uint8_t *mac = (uint8_t *)&msg->payload.u64;
+       struct rte_vdpa_device *vdpa_dev;
+       int did = -1;
 
        RTE_LOG(DEBUG, VHOST_CONFIG,
                ":: mac: %02x:%02x:%02x:%02x:%02x:%02x\n",
@@ -1116,6 +1160,10 @@ vhost_user_send_rarp(struct virtio_net *dev, struct VhostUserMsg *msg)
         */
        rte_smp_wmb();
        rte_atomic16_set(&dev->broadcast_rarp, 1);
+       did = dev->vdpa_dev_id;
+       vdpa_dev = rte_vdpa_get_device(did);
+       if (vdpa_dev && vdpa_dev->ops->migration_done)
+               vdpa_dev->ops->migration_done(dev->vid);
 
        return 0;
 }
@@ -1377,6 +1425,8 @@ vhost_user_msg_handler(int vid, int fd)
 {
        struct virtio_net *dev;
        struct VhostUserMsg msg;
+       struct rte_vdpa_device *vdpa_dev;
+       int did = -1;
        int ret;
        int unlock_required = 0;
 
@@ -1529,7 +1579,7 @@ vhost_user_msg_handler(int vid, int fd)
                break;
 
        case VHOST_USER_GET_QUEUE_NUM:
-               msg.payload.u64 = VHOST_MAX_QUEUE_PAIRS;
+               msg.payload.u64 = (uint64_t)vhost_user_get_queue_num(dev);
                msg.size = sizeof(msg.payload.u64);
                send_vhost_reply(fd, &msg);
                break;
@@ -1582,6 +1632,16 @@ vhost_user_msg_handler(int vid, int fd)
                }
        }
 
+       did = dev->vdpa_dev_id;
+       vdpa_dev = rte_vdpa_get_device(did);
+       if (vdpa_dev && virtio_is_ready(dev) &&
+                       !(dev->flags & VIRTIO_DEV_VDPA_CONFIGURED) &&
+                       msg.request.master == VHOST_USER_SET_VRING_ENABLE) {
+               if (vdpa_dev->ops->dev_conf)
+                       vdpa_dev->ops->dev_conf(dev->vid);
+               dev->flags |= VIRTIO_DEV_VDPA_CONFIGURED;
+       }
+
        return 0;
 }