doc: support IPsec Multi-buffer lib v0.54
[dpdk.git] / drivers / net / virtio / virtio_user / virtio_user_dev.c
index c3ab9a2..7fb135f 100644 (file)
@@ -224,17 +224,13 @@ out:
 static inline void
 parse_mac(struct virtio_user_dev *dev, const char *mac)
 {
-       int i, r;
-       uint32_t tmp[RTE_ETHER_ADDR_LEN];
+       struct rte_ether_addr tmp;
 
        if (!mac)
                return;
 
-       r = sscanf(mac, "%x:%x:%x:%x:%x:%x", &tmp[0],
-                       &tmp[1], &tmp[2], &tmp[3], &tmp[4], &tmp[5]);
-       if (r == RTE_ETHER_ADDR_LEN) {
-               for (i = 0; i < RTE_ETHER_ADDR_LEN; ++i)
-                       dev->mac_addr[i] = (uint8_t)tmp[i];
+       if (rte_ether_unformat_addr(mac, &tmp) == 0) {
+               memcpy(dev->mac_addr, &tmp, RTE_ETHER_ADDR_LEN);
                dev->mac_specified = 1;
        } else {
                /* ignore the wrong mac, use random mac */
@@ -319,9 +315,9 @@ virtio_user_fill_intr_handle(struct virtio_user_dev *dev)
 
 static void
 virtio_user_mem_event_cb(enum rte_mem_event type __rte_unused,
-                                                const void *addr __rte_unused,
-                                                size_t len __rte_unused,
-                                                void *arg)
+                        const void *addr,
+                        size_t len __rte_unused,
+                        void *arg)
 {
        struct virtio_user_dev *dev = arg;
        struct rte_memseg_list *msl;
@@ -541,7 +537,8 @@ virtio_user_dev_uninit(struct virtio_user_dev *dev)
                close(dev->kickfds[i]);
        }
 
-       close(dev->vhostfd);
+       if (dev->vhostfd >= 0)
+               close(dev->vhostfd);
 
        if (dev->is_server && dev->listenfd >= 0) {
                close(dev->listenfd);
@@ -549,8 +546,11 @@ virtio_user_dev_uninit(struct virtio_user_dev *dev)
        }
 
        if (dev->vhostfds) {
-               for (i = 0; i < dev->max_queue_pairs; ++i)
+               for (i = 0; i < dev->max_queue_pairs; ++i) {
                        close(dev->vhostfds[i]);
+                       if (dev->tapfds[i] >= 0)
+                               close(dev->tapfds[i]);
+               }
                free(dev->vhostfds);
                free(dev->tapfds);
        }
@@ -617,6 +617,10 @@ virtio_user_handle_ctrl_msg(struct virtio_user_dev *dev, struct vring *vring,
 
                queues = *(uint16_t *)(uintptr_t)vring->desc[idx_data].addr;
                status = virtio_user_handle_mq(dev, queues);
+       } else if (hdr->class == VIRTIO_NET_CTRL_RX  ||
+                  hdr->class == VIRTIO_NET_CTRL_MAC ||
+                  hdr->class == VIRTIO_NET_CTRL_VLAN) {
+               status = 0;
        }
 
        /* Update status */
@@ -628,7 +632,7 @@ virtio_user_handle_ctrl_msg(struct virtio_user_dev *dev, struct vring *vring,
 static inline int
 desc_is_avail(struct vring_packed_desc *desc, bool wrap_counter)
 {
-       uint16_t flags = desc->flags;
+       uint16_t flags = __atomic_load_n(&desc->flags, __ATOMIC_ACQUIRE);
 
        return wrap_counter == !!(flags & VRING_PACKED_DESC_F_AVAIL) &&
                wrap_counter != !!(flags & VRING_PACKED_DESC_F_USED);
@@ -668,6 +672,10 @@ virtio_user_handle_ctrl_msg_packed(struct virtio_user_dev *dev,
                queues = *(uint16_t *)(uintptr_t)
                                vring->desc[idx_data].addr;
                status = virtio_user_handle_mq(dev, queues);
+       } else if (hdr->class == VIRTIO_NET_CTRL_RX  ||
+                  hdr->class == VIRTIO_NET_CTRL_MAC ||
+                  hdr->class == VIRTIO_NET_CTRL_VLAN) {
+               status = 0;
        }
 
        /* Update status */
@@ -688,6 +696,10 @@ virtio_user_handle_cq_packed(struct virtio_user_dev *dev, uint16_t queue_idx)
        struct vring_packed *vring = &dev->packed_vrings[queue_idx];
        uint16_t n_descs, flags;
 
+       /* Perform a load-acquire barrier in desc_is_avail to
+        * enforce the ordering between desc flags and desc
+        * content.
+        */
        while (desc_is_avail(&vring->desc[vq->used_idx],
                             vq->used_wrap_counter)) {
 
@@ -698,8 +710,8 @@ virtio_user_handle_cq_packed(struct virtio_user_dev *dev, uint16_t queue_idx)
                if (vq->used_wrap_counter)
                        flags |= VRING_PACKED_DESC_F_AVAIL_USED;
 
-               rte_smp_wmb();
-               vring->desc[vq->used_idx].flags = flags;
+               __atomic_store_n(&vring->desc[vq->used_idx].flags, flags,
+                                __ATOMIC_RELEASE);
 
                vq->used_idx += n_descs;
                if (vq->used_idx >= dev->queue_size) {
@@ -718,8 +730,10 @@ virtio_user_handle_cq(struct virtio_user_dev *dev, uint16_t queue_idx)
        struct vring *vring = &dev->vrings[queue_idx];
 
        /* Consume avail ring, using used ring idx as first one */
-       while (vring->used->idx != vring->avail->idx) {
-               avail_idx = (vring->used->idx) & (vring->num - 1);
+       while (__atomic_load_n(&vring->used->idx, __ATOMIC_RELAXED)
+              != vring->avail->idx) {
+               avail_idx = __atomic_load_n(&vring->used->idx, __ATOMIC_RELAXED)
+                           & (vring->num - 1);
                desc_idx = vring->avail->ring[avail_idx];
 
                n_descs = virtio_user_handle_ctrl_msg(dev, vring, desc_idx);
@@ -729,6 +743,6 @@ virtio_user_handle_cq(struct virtio_user_dev *dev, uint16_t queue_idx)
                uep->id = desc_idx;
                uep->len = n_descs;
 
-               vring->used->idx++;
+               __atomic_add_fetch(&vring->used->idx, 1, __ATOMIC_RELAXED);
        }
 }