X-Git-Url: http://git.droids-corp.org/?a=blobdiff_plain;ds=sidebyside;f=lib%2Flibrte_vhost%2Fvdpa.c;h=ae6fdd24edbece10206ec5c27b961868977c797a;hb=83c2047c5fe82fb25674c3053d4871838e7455cb;hp=c82fd437098cbcec182398c89297e3e6c5a926f4;hpb=d7280c9fffcb54cc7fdde9a001c650ed01eeb0fa;p=dpdk.git diff --git a/lib/librte_vhost/vdpa.c b/lib/librte_vhost/vdpa.c index c82fd43709..ae6fdd24ed 100644 --- a/lib/librte_vhost/vdpa.c +++ b/lib/librte_vhost/vdpa.c @@ -9,107 +9,345 @@ */ #include +#include +#include #include +#include +#include + #include "rte_vdpa.h" +#include "rte_vdpa_dev.h" #include "vhost.h" -static struct rte_vdpa_device *vdpa_devices[MAX_VHOST_DEVICE]; -static uint32_t vdpa_device_num; +/** Double linked list of vDPA devices. */ +TAILQ_HEAD(vdpa_device_list, rte_vdpa_device); + +static struct vdpa_device_list vdpa_device_list = + TAILQ_HEAD_INITIALIZER(vdpa_device_list); +static rte_spinlock_t vdpa_device_list_lock = RTE_SPINLOCK_INITIALIZER; -static bool -is_same_vdpa_device(struct rte_vdpa_dev_addr *a, - struct rte_vdpa_dev_addr *b) + +/* Unsafe, needs to be called with vdpa_device_list_lock held */ +static struct rte_vdpa_device * +__vdpa_find_device_by_name(const char *name) { - bool ret = true; - - if (a->type != b->type) - return false; - - switch (a->type) { - case PCI_ADDR: - if (a->pci_addr.domain != b->pci_addr.domain || - a->pci_addr.bus != b->pci_addr.bus || - a->pci_addr.devid != b->pci_addr.devid || - a->pci_addr.function != b->pci_addr.function) - ret = false; - break; - default: - break; + struct rte_vdpa_device *dev, *ret = NULL; + + if (name == NULL) + return NULL; + + TAILQ_FOREACH(dev, &vdpa_device_list, next) { + if (!strncmp(dev->device->name, name, RTE_DEV_NAME_MAX_LEN)) { + ret = dev; + break; + } } return ret; } -int -rte_vdpa_register_device(struct rte_vdpa_dev_addr *addr, +struct rte_vdpa_device * +rte_vdpa_find_device_by_name(const char *name) +{ + struct rte_vdpa_device *dev; + + rte_spinlock_lock(&vdpa_device_list_lock); + dev = __vdpa_find_device_by_name(name); + rte_spinlock_unlock(&vdpa_device_list_lock); + + return dev; +} + +struct rte_device * +rte_vdpa_get_rte_device(struct rte_vdpa_device *vdpa_dev) +{ + if (vdpa_dev == NULL) + return NULL; + + return vdpa_dev->device; +} + +struct rte_vdpa_device * +rte_vdpa_register_device(struct rte_device *rte_dev, struct rte_vdpa_dev_ops *ops) { struct rte_vdpa_device *dev; - char device_name[MAX_VDPA_NAME_LEN]; - int i; - if (vdpa_device_num >= MAX_VHOST_DEVICE) - return -1; + if (ops == NULL) + return NULL; - for (i = 0; i < MAX_VHOST_DEVICE; i++) { - dev = vdpa_devices[i]; - if (dev && is_same_vdpa_device(&dev->addr, addr)) - return -1; + /* Check mandatory ops are implemented */ + if (!ops->get_queue_num || !ops->get_features || + !ops->get_protocol_features || !ops->dev_conf || + !ops->dev_close || !ops->set_vring_state || + !ops->set_features) { + VHOST_LOG_CONFIG(ERR, + "Some mandatory vDPA ops aren't implemented\n"); + return NULL; } - for (i = 0; i < MAX_VHOST_DEVICE; i++) { - if (vdpa_devices[i] == NULL) - break; + rte_spinlock_lock(&vdpa_device_list_lock); + /* Check the device hasn't been register already */ + dev = __vdpa_find_device_by_name(rte_dev->name); + if (dev) { + dev = NULL; + goto out_unlock; } - sprintf(device_name, "vdpa-dev-%d", i); - dev = rte_zmalloc(device_name, sizeof(struct rte_vdpa_device), - RTE_CACHE_LINE_SIZE); + dev = rte_zmalloc(NULL, sizeof(*dev), 0); if (!dev) - return -1; + goto out_unlock; - memcpy(&dev->addr, addr, sizeof(struct rte_vdpa_dev_addr)); + dev->device = rte_dev; dev->ops = ops; - vdpa_devices[i] = dev; - vdpa_device_num++; + TAILQ_INSERT_TAIL(&vdpa_device_list, dev, next); +out_unlock: + rte_spinlock_unlock(&vdpa_device_list_lock); - return i; + return dev; } int -rte_vdpa_unregister_device(int did) +rte_vdpa_unregister_device(struct rte_vdpa_device *dev) { - if (did < 0 || did >= MAX_VHOST_DEVICE || vdpa_devices[did] == NULL) + struct rte_vdpa_device *cur_dev, *tmp_dev; + int ret = -1; + + rte_spinlock_lock(&vdpa_device_list_lock); + TAILQ_FOREACH_SAFE(cur_dev, &vdpa_device_list, next, tmp_dev) { + if (dev != cur_dev) + continue; + + TAILQ_REMOVE(&vdpa_device_list, dev, next); + rte_free(dev); + ret = 0; + break; + } + rte_spinlock_unlock(&vdpa_device_list_lock); + + return ret; +} + +int +rte_vdpa_relay_vring_used(int vid, uint16_t qid, void *vring_m) +{ + struct virtio_net *dev = get_device(vid); + uint16_t idx, idx_m, desc_id; + struct vhost_virtqueue *vq; + struct vring_desc desc; + struct vring_desc *desc_ring; + struct vring_desc *idesc = NULL; + struct vring *s_vring; + uint64_t dlen; + uint32_t nr_descs; + int ret; + + if (!dev || !vring_m) + return -1; + + if (qid >= dev->nr_vring) + return -1; + + if (vq_is_packed(dev)) return -1; - rte_free(vdpa_devices[did]); - vdpa_devices[did] = NULL; - vdpa_device_num--; + s_vring = (struct vring *)vring_m; + vq = dev->virtqueue[qid]; + idx = vq->used->idx; + idx_m = s_vring->used->idx; + ret = (uint16_t)(idx_m - idx); + + while (idx != idx_m) { + /* copy used entry, used ring logging is not covered here */ + vq->used->ring[idx & (vq->size - 1)] = + s_vring->used->ring[idx & (vq->size - 1)]; + + desc_id = vq->used->ring[idx & (vq->size - 1)].id; + desc_ring = vq->desc; + nr_descs = vq->size; + + if (unlikely(desc_id >= vq->size)) + return -1; + + if (vq->desc[desc_id].flags & VRING_DESC_F_INDIRECT) { + dlen = vq->desc[desc_id].len; + nr_descs = dlen / sizeof(struct vring_desc); + if (unlikely(nr_descs > vq->size)) + return -1; + + desc_ring = (struct vring_desc *)(uintptr_t) + vhost_iova_to_vva(dev, vq, + vq->desc[desc_id].addr, &dlen, + VHOST_ACCESS_RO); + if (unlikely(!desc_ring)) + return -1; + + if (unlikely(dlen < vq->desc[desc_id].len)) { + idesc = vhost_alloc_copy_ind_table(dev, vq, + vq->desc[desc_id].addr, + vq->desc[desc_id].len); + if (unlikely(!idesc)) + return -1; + + desc_ring = idesc; + } + + desc_id = 0; + } + + /* dirty page logging for DMA writeable buffer */ + do { + if (unlikely(desc_id >= vq->size)) + goto fail; + if (unlikely(nr_descs-- == 0)) + goto fail; + desc = desc_ring[desc_id]; + if (desc.flags & VRING_DESC_F_WRITE) + vhost_log_write_iova(dev, vq, desc.addr, + desc.len); + desc_id = desc.next; + } while (desc.flags & VRING_DESC_F_NEXT); + + if (unlikely(idesc)) { + free_ind_table(idesc); + idesc = NULL; + } + + idx++; + } + + rte_smp_wmb(); + vq->used->idx = idx_m; - return did; + if (dev->features & (1ULL << VIRTIO_RING_F_EVENT_IDX)) + vring_used_event(s_vring) = idx_m; + + return ret; + +fail: + if (unlikely(idesc)) + free_ind_table(idesc); + return -1; +} + +int +rte_vdpa_get_queue_num(struct rte_vdpa_device *dev, uint32_t *queue_num) +{ + if (dev == NULL || dev->ops == NULL || dev->ops->get_queue_num == NULL) + return -1; + + return dev->ops->get_queue_num(dev, queue_num); +} + +int +rte_vdpa_get_features(struct rte_vdpa_device *dev, uint64_t *features) +{ + if (dev == NULL || dev->ops == NULL || dev->ops->get_features == NULL) + return -1; + + return dev->ops->get_features(dev, features); +} + +int +rte_vdpa_get_protocol_features(struct rte_vdpa_device *dev, uint64_t *features) +{ + if (dev == NULL || dev->ops == NULL || + dev->ops->get_protocol_features == NULL) + return -1; + + return dev->ops->get_protocol_features(dev, features); +} + +int +rte_vdpa_get_stats_names(struct rte_vdpa_device *dev, + struct rte_vdpa_stat_name *stats_names, + unsigned int size) +{ + if (!dev) + return -EINVAL; + + RTE_FUNC_PTR_OR_ERR_RET(dev->ops->get_stats_names, -ENOTSUP); + + return dev->ops->get_stats_names(dev, stats_names, size); } int -rte_vdpa_find_device_id(struct rte_vdpa_dev_addr *addr) +rte_vdpa_get_stats(struct rte_vdpa_device *dev, uint16_t qid, + struct rte_vdpa_stat *stats, unsigned int n) +{ + if (!dev || !stats || !n) + return -EINVAL; + + RTE_FUNC_PTR_OR_ERR_RET(dev->ops->get_stats, -ENOTSUP); + + return dev->ops->get_stats(dev, qid, stats, n); +} + +int +rte_vdpa_reset_stats(struct rte_vdpa_device *dev, uint16_t qid) +{ + if (!dev) + return -EINVAL; + + RTE_FUNC_PTR_OR_ERR_RET(dev->ops->reset_stats, -ENOTSUP); + + return dev->ops->reset_stats(dev, qid); +} + +static int +vdpa_dev_match(struct rte_vdpa_device *dev, + const struct rte_device *rte_dev) +{ + if (dev->device == rte_dev) + return 0; + + return -1; +} + +/* Generic rte_vdpa_dev comparison function. */ +typedef int (*rte_vdpa_cmp_t)(struct rte_vdpa_device *, + const struct rte_device *rte_dev); + +static struct rte_vdpa_device * +vdpa_find_device(const struct rte_vdpa_device *start, rte_vdpa_cmp_t cmp, + struct rte_device *rte_dev) { struct rte_vdpa_device *dev; - int i; - for (i = 0; i < MAX_VHOST_DEVICE; ++i) { - dev = vdpa_devices[i]; - if (dev && is_same_vdpa_device(&dev->addr, addr)) - return i; + rte_spinlock_lock(&vdpa_device_list_lock); + if (start == NULL) + dev = TAILQ_FIRST(&vdpa_device_list); + else + dev = TAILQ_NEXT(start, next); + + while (dev != NULL) { + if (cmp(dev, rte_dev) == 0) + break; + + dev = TAILQ_NEXT(dev, next); } + rte_spinlock_unlock(&vdpa_device_list_lock); - return -1; + return dev; } -struct rte_vdpa_device * -rte_vdpa_get_device(int did) +static void * +vdpa_dev_iterate(const void *start, + const char *str, + const struct rte_dev_iterator *it) { - if (did < 0 || did >= MAX_VHOST_DEVICE) - return NULL; + struct rte_vdpa_device *vdpa_dev = NULL; + + RTE_SET_USED(str); - return vdpa_devices[did]; + vdpa_dev = vdpa_find_device(start, vdpa_dev_match, it->device); + + return vdpa_dev; } + +static struct rte_class rte_class_vdpa = { + .dev_iterate = vdpa_dev_iterate, +}; + +RTE_REGISTER_CLASS(vdpa, rte_class_vdpa);