static int virtio_dev_link_update(struct rte_eth_dev *dev,
int wait_to_complete);
static int virtio_dev_vlan_offload_set(struct rte_eth_dev *dev, int mask);
+static int virtio_dev_rss_hash_update(struct rte_eth_dev *dev,
+ struct rte_eth_rss_conf *rss_conf);
+static int virtio_dev_rss_hash_conf_get(struct rte_eth_dev *dev,
+ struct rte_eth_rss_conf *rss_conf);
+static int virtio_dev_rss_reta_update(struct rte_eth_dev *dev,
+ struct rte_eth_rss_reta_entry64 *reta_conf,
+ uint16_t reta_size);
+static int virtio_dev_rss_reta_query(struct rte_eth_dev *dev,
+ struct rte_eth_rss_reta_entry64 *reta_conf,
+ uint16_t reta_size);
static void virtio_set_hwaddr(struct virtio_hw *hw);
static void virtio_get_hwaddr(struct virtio_hw *hw);
struct rte_ether_addr *mac_addr);
static int virtio_intr_disable(struct rte_eth_dev *dev);
+static int virtio_get_monitor_addr(void *rx_queue,
+ struct rte_power_monitor_cond *pmc);
static int virtio_dev_queue_stats_mapping_set(
struct rte_eth_dev *eth_dev,
}
static int
-virtio_set_multiple_queues(struct rte_eth_dev *dev, uint16_t nb_queues)
+virtio_set_multiple_queues_rss(struct rte_eth_dev *dev, uint16_t nb_queues)
{
struct virtio_hw *hw = dev->data->dev_private;
struct virtio_pmd_ctrl ctrl;
- int dlen[1];
+ struct virtio_net_ctrl_rss rss;
+ int dlen, ret;
+
+ rss.hash_types = hw->rss_hash_types & VIRTIO_NET_HASH_TYPE_MASK;
+ RTE_BUILD_BUG_ON(!RTE_IS_POWER_OF_2(VIRTIO_NET_RSS_RETA_SIZE));
+ rss.indirection_table_mask = VIRTIO_NET_RSS_RETA_SIZE - 1;
+ rss.unclassified_queue = 0;
+ memcpy(rss.indirection_table, hw->rss_reta, VIRTIO_NET_RSS_RETA_SIZE * sizeof(uint16_t));
+ rss.max_tx_vq = nb_queues;
+ rss.hash_key_length = VIRTIO_NET_RSS_KEY_SIZE;
+ memcpy(rss.hash_key_data, hw->rss_key, VIRTIO_NET_RSS_KEY_SIZE);
+
+ ctrl.hdr.class = VIRTIO_NET_CTRL_MQ;
+ ctrl.hdr.cmd = VIRTIO_NET_CTRL_MQ_RSS_CONFIG;
+ memcpy(ctrl.data, &rss, sizeof(rss));
+
+ dlen = sizeof(rss);
+
+ ret = virtio_send_command(hw->cvq, &ctrl, &dlen, 1);
+ if (ret) {
+ PMD_INIT_LOG(ERR, "RSS multiqueue configured but send command failed");
+ return -EINVAL;
+ }
+
+ return 0;
+}
+
+static int
+virtio_set_multiple_queues_auto(struct rte_eth_dev *dev, uint16_t nb_queues)
+{
+ struct virtio_hw *hw = dev->data->dev_private;
+ struct virtio_pmd_ctrl ctrl;
+ int dlen;
int ret;
ctrl.hdr.class = VIRTIO_NET_CTRL_MQ;
ctrl.hdr.cmd = VIRTIO_NET_CTRL_MQ_VQ_PAIRS_SET;
memcpy(ctrl.data, &nb_queues, sizeof(uint16_t));
- dlen[0] = sizeof(uint16_t);
+ dlen = sizeof(uint16_t);
- ret = virtio_send_command(hw->cvq, &ctrl, dlen, 1);
+ ret = virtio_send_command(hw->cvq, &ctrl, &dlen, 1);
if (ret) {
PMD_INIT_LOG(ERR, "Multiqueue configured but send command "
"failed, this is too late now...");
return 0;
}
+static int
+virtio_set_multiple_queues(struct rte_eth_dev *dev, uint16_t nb_queues)
+{
+ struct virtio_hw *hw = dev->data->dev_private;
+
+ if (virtio_with_feature(hw, VIRTIO_NET_F_RSS))
+ return virtio_set_multiple_queues_rss(dev, nb_queues);
+ else
+ return virtio_set_multiple_queues_auto(dev, nb_queues);
+}
+
static uint16_t
virtio_get_nr_vq(struct virtio_hw *hw)
{
static void virtio_queues_unbind_intr(struct rte_eth_dev *dev);
+static void
+virtio_free_rss(struct virtio_hw *hw)
+{
+ rte_free(hw->rss_key);
+ hw->rss_key = NULL;
+
+ rte_free(hw->rss_reta);
+ hw->rss_reta = NULL;
+}
+
int
virtio_dev_close(struct rte_eth_dev *dev)
{
if (intr_conf->lsc || intr_conf->rxq) {
virtio_intr_disable(dev);
rte_intr_efd_disable(dev->intr_handle);
- rte_free(dev->intr_handle->intr_vec);
- dev->intr_handle->intr_vec = NULL;
+ rte_intr_vec_list_free(dev->intr_handle);
}
virtio_reset(hw);
virtio_dev_free_mbufs(dev);
virtio_free_queues(hw);
+ virtio_free_rss(hw);
return VIRTIO_OPS(hw)->dev_close(hw);
}
.rx_queue_intr_enable = virtio_dev_rx_queue_intr_enable,
.rx_queue_intr_disable = virtio_dev_rx_queue_intr_disable,
.tx_queue_setup = virtio_dev_tx_queue_setup,
+ .rss_hash_update = virtio_dev_rss_hash_update,
+ .rss_hash_conf_get = virtio_dev_rss_hash_conf_get,
+ .reta_update = virtio_dev_rss_reta_update,
+ .reta_query = virtio_dev_rss_reta_query,
/* collect stats per queue */
.queue_stats_mapping_set = virtio_dev_queue_stats_mapping_set,
.vlan_filter_set = virtio_vlan_filter_set,
.mac_addr_add = virtio_mac_addr_add,
.mac_addr_remove = virtio_mac_addr_remove,
.mac_addr_set = virtio_mac_addr_set,
+ .get_monitor_addr = virtio_get_monitor_addr,
};
/*
return 0;
}
+#define CLB_VAL_IDX 0
+#define CLB_MSK_IDX 1
+#define CLB_MATCH_IDX 2
+static int
+virtio_monitor_callback(const uint64_t value,
+ const uint64_t opaque[RTE_POWER_MONITOR_OPAQUE_SZ])
+{
+ const uint64_t m = opaque[CLB_MSK_IDX];
+ const uint64_t v = opaque[CLB_VAL_IDX];
+ const uint64_t c = opaque[CLB_MATCH_IDX];
+
+ if (c)
+ return (value & m) == v ? -1 : 0;
+ else
+ return (value & m) == v ? 0 : -1;
+}
+
+static int
+virtio_get_monitor_addr(void *rx_queue, struct rte_power_monitor_cond *pmc)
+{
+ struct virtnet_rx *rxvq = rx_queue;
+ struct virtqueue *vq = virtnet_rxq_to_vq(rxvq);
+ struct virtio_hw *hw;
+
+ if (vq == NULL)
+ return -EINVAL;
+
+ hw = vq->hw;
+ if (virtio_with_packed_queue(hw)) {
+ struct vring_packed_desc *desc;
+ desc = vq->vq_packed.ring.desc;
+ pmc->addr = &desc[vq->vq_used_cons_idx].flags;
+ if (vq->vq_packed.used_wrap_counter)
+ pmc->opaque[CLB_VAL_IDX] =
+ VRING_PACKED_DESC_F_AVAIL_USED;
+ else
+ pmc->opaque[CLB_VAL_IDX] = 0;
+ pmc->opaque[CLB_MSK_IDX] = VRING_PACKED_DESC_F_AVAIL_USED;
+ pmc->opaque[CLB_MATCH_IDX] = 1;
+ pmc->size = sizeof(desc[vq->vq_used_cons_idx].flags);
+ } else {
+ pmc->addr = &vq->vq_split.ring.used->idx;
+ pmc->opaque[CLB_VAL_IDX] = vq->vq_used_cons_idx
+ & (vq->vq_nentries - 1);
+ pmc->opaque[CLB_MSK_IDX] = vq->vq_nentries - 1;
+ pmc->opaque[CLB_MATCH_IDX] = 0;
+ pmc->size = sizeof(vq->vq_split.ring.used->idx);
+ }
+ pmc->fn = virtio_monitor_callback;
+
+ return 0;
+}
+
static int
virtio_vlan_filter_set(struct rte_eth_dev *dev, uint16_t vlan_id, int on)
{
PMD_INIT_LOG(INFO, "queue/interrupt binding");
for (i = 0; i < dev->data->nb_rx_queues; ++i) {
- dev->intr_handle->intr_vec[i] = i + 1;
+ if (rte_intr_vec_list_index_set(dev->intr_handle, i,
+ i + 1))
+ return -rte_errno;
if (VIRTIO_OPS(hw)->set_queue_irq(hw, hw->vqs[i * 2], i + 1) ==
VIRTIO_MSI_NO_VECTOR) {
PMD_DRV_LOG(ERR, "failed to set queue vector");
return -1;
}
- if (!dev->intr_handle->intr_vec) {
- dev->intr_handle->intr_vec =
- rte_zmalloc("intr_vec",
- hw->max_queue_pairs * sizeof(int), 0);
- if (!dev->intr_handle->intr_vec) {
- PMD_INIT_LOG(ERR, "Failed to allocate %u rxq vectors",
- hw->max_queue_pairs);
- return -ENOMEM;
- }
+ if (rte_intr_vec_list_alloc(dev->intr_handle, "intr_vec",
+ hw->max_queue_pairs)) {
+ PMD_INIT_LOG(ERR, "Failed to allocate %u rxq vectors",
+ hw->max_queue_pairs);
+ return -ENOMEM;
}
if (dev->data->dev_flags & RTE_ETH_DEV_INTR_LSC) {
return 0;
}
+
+static void
+virtio_get_speed_duplex(struct rte_eth_dev *eth_dev,
+ struct rte_eth_link *link)
+{
+ struct virtio_hw *hw = eth_dev->data->dev_private;
+ struct virtio_net_config *config;
+ struct virtio_net_config local_config;
+
+ config = &local_config;
+ virtio_read_dev_config(hw,
+ offsetof(struct virtio_net_config, speed),
+ &config->speed, sizeof(config->speed));
+ virtio_read_dev_config(hw,
+ offsetof(struct virtio_net_config, duplex),
+ &config->duplex, sizeof(config->duplex));
+ hw->speed = config->speed;
+ hw->duplex = config->duplex;
+ if (link != NULL) {
+ link->link_duplex = hw->duplex;
+ link->link_speed = hw->speed;
+ }
+ PMD_INIT_LOG(DEBUG, "link speed = %d, duplex = %d",
+ hw->speed, hw->duplex);
+}
+
+static uint64_t
+ethdev_to_virtio_rss_offloads(uint64_t ethdev_hash_types)
+{
+ uint64_t virtio_hash_types = 0;
+
+ if (ethdev_hash_types & (RTE_ETH_RSS_IPV4 | RTE_ETH_RSS_FRAG_IPV4 |
+ RTE_ETH_RSS_NONFRAG_IPV4_OTHER))
+ virtio_hash_types |= VIRTIO_NET_HASH_TYPE_IPV4;
+
+ if (ethdev_hash_types & RTE_ETH_RSS_NONFRAG_IPV4_TCP)
+ virtio_hash_types |= VIRTIO_NET_HASH_TYPE_TCPV4;
+
+ if (ethdev_hash_types & RTE_ETH_RSS_NONFRAG_IPV4_UDP)
+ virtio_hash_types |= VIRTIO_NET_HASH_TYPE_UDPV4;
+
+ if (ethdev_hash_types & (RTE_ETH_RSS_IPV6 | RTE_ETH_RSS_FRAG_IPV6 |
+ RTE_ETH_RSS_NONFRAG_IPV6_OTHER))
+ virtio_hash_types |= VIRTIO_NET_HASH_TYPE_IPV6;
+
+ if (ethdev_hash_types & RTE_ETH_RSS_NONFRAG_IPV6_TCP)
+ virtio_hash_types |= VIRTIO_NET_HASH_TYPE_TCPV6;
+
+ if (ethdev_hash_types & RTE_ETH_RSS_NONFRAG_IPV6_UDP)
+ virtio_hash_types |= VIRTIO_NET_HASH_TYPE_UDPV6;
+
+ if (ethdev_hash_types & RTE_ETH_RSS_IPV6_EX)
+ virtio_hash_types |= VIRTIO_NET_HASH_TYPE_IP_EX;
+
+ if (ethdev_hash_types & RTE_ETH_RSS_IPV6_TCP_EX)
+ virtio_hash_types |= VIRTIO_NET_HASH_TYPE_TCP_EX;
+
+ if (ethdev_hash_types & RTE_ETH_RSS_IPV6_UDP_EX)
+ virtio_hash_types |= VIRTIO_NET_HASH_TYPE_UDP_EX;
+
+ return virtio_hash_types;
+}
+
+static uint64_t
+virtio_to_ethdev_rss_offloads(uint64_t virtio_hash_types)
+{
+ uint64_t rss_offloads = 0;
+
+ if (virtio_hash_types & VIRTIO_NET_HASH_TYPE_IPV4)
+ rss_offloads |= RTE_ETH_RSS_IPV4 | RTE_ETH_RSS_FRAG_IPV4 |
+ RTE_ETH_RSS_NONFRAG_IPV4_OTHER;
+
+ if (virtio_hash_types & VIRTIO_NET_HASH_TYPE_TCPV4)
+ rss_offloads |= RTE_ETH_RSS_NONFRAG_IPV4_TCP;
+
+ if (virtio_hash_types & VIRTIO_NET_HASH_TYPE_UDPV4)
+ rss_offloads |= RTE_ETH_RSS_NONFRAG_IPV4_UDP;
+
+ if (virtio_hash_types & VIRTIO_NET_HASH_TYPE_IPV6)
+ rss_offloads |= RTE_ETH_RSS_IPV6 | RTE_ETH_RSS_FRAG_IPV6 |
+ RTE_ETH_RSS_NONFRAG_IPV6_OTHER;
+
+ if (virtio_hash_types & VIRTIO_NET_HASH_TYPE_TCPV6)
+ rss_offloads |= RTE_ETH_RSS_NONFRAG_IPV6_TCP;
+
+ if (virtio_hash_types & VIRTIO_NET_HASH_TYPE_UDPV6)
+ rss_offloads |= RTE_ETH_RSS_NONFRAG_IPV6_UDP;
+
+ if (virtio_hash_types & VIRTIO_NET_HASH_TYPE_IP_EX)
+ rss_offloads |= RTE_ETH_RSS_IPV6_EX;
+
+ if (virtio_hash_types & VIRTIO_NET_HASH_TYPE_TCP_EX)
+ rss_offloads |= RTE_ETH_RSS_IPV6_TCP_EX;
+
+ if (virtio_hash_types & VIRTIO_NET_HASH_TYPE_UDP_EX)
+ rss_offloads |= RTE_ETH_RSS_IPV6_UDP_EX;
+
+ return rss_offloads;
+}
+
+static int
+virtio_dev_get_rss_config(struct virtio_hw *hw, uint32_t *rss_hash_types)
+{
+ struct virtio_net_config local_config;
+ struct virtio_net_config *config = &local_config;
+
+ virtio_read_dev_config(hw,
+ offsetof(struct virtio_net_config, rss_max_key_size),
+ &config->rss_max_key_size,
+ sizeof(config->rss_max_key_size));
+ if (config->rss_max_key_size < VIRTIO_NET_RSS_KEY_SIZE) {
+ PMD_INIT_LOG(ERR, "Invalid device RSS max key size (%u)",
+ config->rss_max_key_size);
+ return -EINVAL;
+ }
+
+ virtio_read_dev_config(hw,
+ offsetof(struct virtio_net_config,
+ rss_max_indirection_table_length),
+ &config->rss_max_indirection_table_length,
+ sizeof(config->rss_max_indirection_table_length));
+ if (config->rss_max_indirection_table_length < VIRTIO_NET_RSS_RETA_SIZE) {
+ PMD_INIT_LOG(ERR, "Invalid device RSS max reta size (%u)",
+ config->rss_max_indirection_table_length);
+ return -EINVAL;
+ }
+
+ virtio_read_dev_config(hw,
+ offsetof(struct virtio_net_config, supported_hash_types),
+ &config->supported_hash_types,
+ sizeof(config->supported_hash_types));
+ if ((config->supported_hash_types & VIRTIO_NET_HASH_TYPE_MASK) == 0) {
+ PMD_INIT_LOG(ERR, "Invalid device RSS hash types (0x%x)",
+ config->supported_hash_types);
+ return -EINVAL;
+ }
+
+ *rss_hash_types = config->supported_hash_types & VIRTIO_NET_HASH_TYPE_MASK;
+
+ PMD_INIT_LOG(DEBUG, "Device RSS config:");
+ PMD_INIT_LOG(DEBUG, "\t-Max key size: %u", config->rss_max_key_size);
+ PMD_INIT_LOG(DEBUG, "\t-Max reta size: %u", config->rss_max_indirection_table_length);
+ PMD_INIT_LOG(DEBUG, "\t-Supported hash types: 0x%x", *rss_hash_types);
+
+ return 0;
+}
+
+static int
+virtio_dev_rss_hash_update(struct rte_eth_dev *dev,
+ struct rte_eth_rss_conf *rss_conf)
+{
+ struct virtio_hw *hw = dev->data->dev_private;
+ char old_rss_key[VIRTIO_NET_RSS_KEY_SIZE];
+ uint32_t old_hash_types;
+ uint16_t nb_queues;
+ int ret;
+
+ if (!virtio_with_feature(hw, VIRTIO_NET_F_RSS))
+ return -ENOTSUP;
+
+ if (rss_conf->rss_hf & ~virtio_to_ethdev_rss_offloads(VIRTIO_NET_HASH_TYPE_MASK))
+ return -EINVAL;
+
+ old_hash_types = hw->rss_hash_types;
+ hw->rss_hash_types = ethdev_to_virtio_rss_offloads(rss_conf->rss_hf);
+
+ if (rss_conf->rss_key && rss_conf->rss_key_len) {
+ if (rss_conf->rss_key_len != VIRTIO_NET_RSS_KEY_SIZE) {
+ PMD_INIT_LOG(ERR, "Driver only supports %u RSS key length",
+ VIRTIO_NET_RSS_KEY_SIZE);
+ ret = -EINVAL;
+ goto restore_types;
+ }
+ memcpy(old_rss_key, hw->rss_key, VIRTIO_NET_RSS_KEY_SIZE);
+ memcpy(hw->rss_key, rss_conf->rss_key, VIRTIO_NET_RSS_KEY_SIZE);
+ }
+
+ nb_queues = RTE_MAX(dev->data->nb_rx_queues, dev->data->nb_tx_queues);
+ ret = virtio_set_multiple_queues_rss(dev, nb_queues);
+ if (ret < 0) {
+ PMD_INIT_LOG(ERR, "Failed to apply new RSS config to the device");
+ goto restore_key;
+ }
+
+ return 0;
+restore_key:
+ if (rss_conf->rss_key && rss_conf->rss_key_len)
+ memcpy(hw->rss_key, old_rss_key, VIRTIO_NET_RSS_KEY_SIZE);
+restore_types:
+ hw->rss_hash_types = old_hash_types;
+
+ return ret;
+}
+
+static int
+virtio_dev_rss_hash_conf_get(struct rte_eth_dev *dev,
+ struct rte_eth_rss_conf *rss_conf)
+{
+ struct virtio_hw *hw = dev->data->dev_private;
+
+ if (!virtio_with_feature(hw, VIRTIO_NET_F_RSS))
+ return -ENOTSUP;
+
+ if (rss_conf->rss_key && rss_conf->rss_key_len >= VIRTIO_NET_RSS_KEY_SIZE)
+ memcpy(rss_conf->rss_key, hw->rss_key, VIRTIO_NET_RSS_KEY_SIZE);
+ rss_conf->rss_key_len = VIRTIO_NET_RSS_KEY_SIZE;
+ rss_conf->rss_hf = virtio_to_ethdev_rss_offloads(hw->rss_hash_types);
+
+ return 0;
+}
+
+static int virtio_dev_rss_reta_update(struct rte_eth_dev *dev,
+ struct rte_eth_rss_reta_entry64 *reta_conf,
+ uint16_t reta_size)
+{
+ struct virtio_hw *hw = dev->data->dev_private;
+ uint16_t nb_queues;
+ uint16_t old_reta[VIRTIO_NET_RSS_RETA_SIZE];
+ int idx, pos, i, ret;
+
+ if (!virtio_with_feature(hw, VIRTIO_NET_F_RSS))
+ return -ENOTSUP;
+
+ if (reta_size != VIRTIO_NET_RSS_RETA_SIZE)
+ return -EINVAL;
+
+ memcpy(old_reta, hw->rss_reta, sizeof(old_reta));
+
+ for (i = 0; i < reta_size; i++) {
+ idx = i / RTE_ETH_RETA_GROUP_SIZE;
+ pos = i % RTE_ETH_RETA_GROUP_SIZE;
+
+ if (((reta_conf[idx].mask >> pos) & 0x1) == 0)
+ continue;
+
+ hw->rss_reta[i] = reta_conf[idx].reta[pos];
+ }
+
+ nb_queues = RTE_MAX(dev->data->nb_rx_queues, dev->data->nb_tx_queues);
+ ret = virtio_set_multiple_queues_rss(dev, nb_queues);
+ if (ret < 0) {
+ PMD_INIT_LOG(ERR, "Failed to apply new RETA to the device");
+ memcpy(hw->rss_reta, old_reta, sizeof(old_reta));
+ }
+
+ hw->rss_rx_queues = dev->data->nb_rx_queues;
+
+ return ret;
+}
+
+static int virtio_dev_rss_reta_query(struct rte_eth_dev *dev,
+ struct rte_eth_rss_reta_entry64 *reta_conf,
+ uint16_t reta_size)
+{
+ struct virtio_hw *hw = dev->data->dev_private;
+ int idx, i;
+
+ if (!virtio_with_feature(hw, VIRTIO_NET_F_RSS))
+ return -ENOTSUP;
+
+ if (reta_size != VIRTIO_NET_RSS_RETA_SIZE)
+ return -EINVAL;
+
+ for (i = 0; i < reta_size; i++) {
+ idx = i / RTE_ETH_RETA_GROUP_SIZE;
+ reta_conf[idx].reta[i % RTE_ETH_RETA_GROUP_SIZE] = hw->rss_reta[i];
+ }
+
+ return 0;
+}
+
+/*
+ * As default RSS hash key, it uses the default key of the
+ * Intel IXGBE devices. It can be updated by the application
+ * with any 40B key value.
+ */
+static uint8_t rss_intel_key[VIRTIO_NET_RSS_KEY_SIZE] = {
+ 0x6D, 0x5A, 0x56, 0xDA, 0x25, 0x5B, 0x0E, 0xC2,
+ 0x41, 0x67, 0x25, 0x3D, 0x43, 0xA3, 0x8F, 0xB0,
+ 0xD0, 0xCA, 0x2B, 0xCB, 0xAE, 0x7B, 0x30, 0xB4,
+ 0x77, 0xCB, 0x2D, 0xA3, 0x80, 0x30, 0xF2, 0x0C,
+ 0x6A, 0x42, 0xB7, 0x3B, 0xBE, 0xAC, 0x01, 0xFA,
+};
+
+static int
+virtio_dev_rss_init(struct rte_eth_dev *eth_dev)
+{
+ struct virtio_hw *hw = eth_dev->data->dev_private;
+ uint16_t nb_rx_queues = eth_dev->data->nb_rx_queues;
+ struct rte_eth_rss_conf *rss_conf;
+ int ret, i;
+
+ if (!nb_rx_queues) {
+ PMD_INIT_LOG(ERR, "Cannot init RSS if no Rx queues");
+ return -EINVAL;
+ }
+
+ rss_conf = ð_dev->data->dev_conf.rx_adv_conf.rss_conf;
+
+ ret = virtio_dev_get_rss_config(hw, &hw->rss_hash_types);
+ if (ret)
+ return ret;
+
+ if (rss_conf->rss_hf) {
+ /* Ensure requested hash types are supported by the device */
+ if (rss_conf->rss_hf & ~virtio_to_ethdev_rss_offloads(hw->rss_hash_types))
+ return -EINVAL;
+
+ hw->rss_hash_types = ethdev_to_virtio_rss_offloads(rss_conf->rss_hf);
+ }
+
+ if (!hw->rss_key) {
+ /* Setup default RSS key if not already setup by the user */
+ hw->rss_key = rte_malloc_socket("rss_key",
+ VIRTIO_NET_RSS_KEY_SIZE, 0,
+ eth_dev->device->numa_node);
+ if (!hw->rss_key) {
+ PMD_INIT_LOG(ERR, "Failed to allocate RSS key");
+ return -1;
+ }
+ }
+
+ if (rss_conf->rss_key && rss_conf->rss_key_len) {
+ if (rss_conf->rss_key_len != VIRTIO_NET_RSS_KEY_SIZE) {
+ PMD_INIT_LOG(ERR, "Driver only supports %u RSS key length",
+ VIRTIO_NET_RSS_KEY_SIZE);
+ return -EINVAL;
+ }
+ memcpy(hw->rss_key, rss_conf->rss_key, VIRTIO_NET_RSS_KEY_SIZE);
+ } else {
+ memcpy(hw->rss_key, rss_intel_key, VIRTIO_NET_RSS_KEY_SIZE);
+ }
+
+ if (!hw->rss_reta) {
+ /* Setup default RSS reta if not already setup by the user */
+ hw->rss_reta = rte_zmalloc_socket("rss_reta",
+ VIRTIO_NET_RSS_RETA_SIZE * sizeof(uint16_t), 0,
+ eth_dev->device->numa_node);
+ if (!hw->rss_reta) {
+ PMD_INIT_LOG(ERR, "Failed to allocate RSS reta");
+ return -1;
+ }
+
+ hw->rss_rx_queues = 0;
+ }
+
+ /* Re-initialize the RSS reta if the number of RX queues has changed */
+ if (hw->rss_rx_queues != nb_rx_queues) {
+ for (i = 0; i < VIRTIO_NET_RSS_RETA_SIZE; i++)
+ hw->rss_reta[i] = i % nb_rx_queues;
+ hw->rss_rx_queues = nb_rx_queues;
+ }
+
+ return 0;
+}
+
#define DUPLEX_UNKNOWN 0xff
/* reset device and renegotiate features if needed */
static int
hw->mac_addr[0], hw->mac_addr[1], hw->mac_addr[2],
hw->mac_addr[3], hw->mac_addr[4], hw->mac_addr[5]);
- if (hw->speed == RTE_ETH_SPEED_NUM_UNKNOWN) {
- if (virtio_with_feature(hw, VIRTIO_NET_F_SPEED_DUPLEX)) {
- config = &local_config;
- virtio_read_dev_config(hw,
- offsetof(struct virtio_net_config, speed),
- &config->speed, sizeof(config->speed));
- virtio_read_dev_config(hw,
- offsetof(struct virtio_net_config, duplex),
- &config->duplex, sizeof(config->duplex));
- hw->speed = config->speed;
- hw->duplex = config->duplex;
- }
- }
+ hw->get_speed_via_feat = hw->speed == RTE_ETH_SPEED_NUM_UNKNOWN &&
+ virtio_with_feature(hw, VIRTIO_NET_F_SPEED_DUPLEX);
+ if (hw->get_speed_via_feat)
+ virtio_get_speed_duplex(eth_dev, NULL);
if (hw->duplex == DUPLEX_UNKNOWN)
hw->duplex = RTE_ETH_LINK_FULL_DUPLEX;
PMD_INIT_LOG(DEBUG, "link speed = %d, duplex = %d",
config->status = 0;
}
- if (virtio_with_feature(hw, VIRTIO_NET_F_MQ)) {
+ if (virtio_with_feature(hw, VIRTIO_NET_F_MQ) ||
+ virtio_with_feature(hw, VIRTIO_NET_F_RSS)) {
virtio_read_dev_config(hw,
offsetof(struct virtio_net_config, max_virtqueue_pairs),
&config->max_virtqueue_pairs,
sizeof(config->max_virtqueue_pairs));
} else {
PMD_INIT_LOG(DEBUG,
- "VIRTIO_NET_F_MQ is not supported");
+ "Neither VIRTIO_NET_F_MQ nor VIRTIO_NET_F_RSS are supported");
config->max_virtqueue_pairs = 1;
}
VLAN_TAG_LEN - hw->vtnet_hdr_size;
}
+ hw->rss_hash_types = 0;
+ if (virtio_with_feature(hw, VIRTIO_NET_F_RSS))
+ if (virtio_dev_rss_init(eth_dev))
+ return -1;
+
PMD_INIT_LOG(DEBUG, "config->max_virtqueue_pairs=%d",
config->max_virtqueue_pairs);
PMD_INIT_LOG(DEBUG, "config->status=%d", config->status);
PMD_INIT_LOG(DEBUG, "configure");
req_features = VIRTIO_PMD_DEFAULT_GUEST_FEATURES;
- if (rxmode->mq_mode != RTE_ETH_MQ_RX_NONE) {
+ if (rxmode->mq_mode != RTE_ETH_MQ_RX_NONE && rxmode->mq_mode != RTE_ETH_MQ_RX_RSS) {
PMD_DRV_LOG(ERR,
"Unsupported Rx multi queue mode %d",
rxmode->mq_mode);
return ret;
}
+ if (rxmode->mq_mode == RTE_ETH_MQ_RX_RSS)
+ req_features |= (1ULL << VIRTIO_NET_F_RSS);
+
if (rxmode->mtu > hw->max_mtu)
req_features &= ~(1ULL << VIRTIO_NET_F_MTU);
return ret;
}
+ if ((rxmode->mq_mode & RTE_ETH_MQ_RX_RSS_FLAG) &&
+ !virtio_with_feature(hw, VIRTIO_NET_F_RSS)) {
+ PMD_DRV_LOG(ERR, "RSS support requested but not supported by the device");
+ return -ENOTSUP;
+ }
+
if ((rx_offloads & (RTE_ETH_RX_OFFLOAD_UDP_CKSUM |
RTE_ETH_RX_OFFLOAD_TCP_CKSUM)) &&
!virtio_with_feature(hw, VIRTIO_NET_F_GUEST_CSUM)) {
hw->has_rx_offload = rx_offload_enabled(hw);
if (dev->data->dev_flags & RTE_ETH_DEV_INTR_LSC)
- /* Enable vector (0) for Link State Intrerrupt */
+ /* Enable vector (0) for Link State Interrupt */
if (VIRTIO_OPS(hw)->set_config_irq(hw, 0) ==
VIRTIO_MSI_NO_VECTOR) {
PMD_DRV_LOG(ERR, "failed to set config vector");
}
}
- /* Enable uio/vfio intr/eventfd mapping: althrough we already did that
+ /* Enable uio/vfio intr/eventfd mapping: although we already did that
* in device configure, but it could be unmapped when device is
* stopped.
*/
dev->data->port_id);
} else {
link.link_status = RTE_ETH_LINK_UP;
+ if (hw->get_speed_via_feat)
+ virtio_get_speed_duplex(dev, &link);
PMD_INIT_LOG(DEBUG, "Port %d is up",
dev->data->port_id);
}
} else {
link.link_status = RTE_ETH_LINK_UP;
+ if (hw->get_speed_via_feat)
+ virtio_get_speed_duplex(dev, &link);
}
return rte_eth_linkstatus_set(dev, &link);
virtio_dev_info_get(struct rte_eth_dev *dev, struct rte_eth_dev_info *dev_info)
{
uint64_t tso_mask, host_features;
+ uint32_t rss_hash_types = 0;
struct virtio_hw *hw = dev->data->dev_private;
dev_info->speed_capa = virtio_dev_speed_capa_get(hw->speed);
if ((host_features & tso_mask) == tso_mask)
dev_info->tx_offload_capa |= RTE_ETH_TX_OFFLOAD_TCP_TSO;
+ if (host_features & (1ULL << VIRTIO_NET_F_RSS)) {
+ virtio_dev_get_rss_config(hw, &rss_hash_types);
+ dev_info->hash_key_size = VIRTIO_NET_RSS_KEY_SIZE;
+ dev_info->reta_size = VIRTIO_NET_RSS_RETA_SIZE;
+ dev_info->flow_type_rss_offloads =
+ virtio_to_ethdev_rss_offloads(rss_hash_types);
+ } else {
+ dev_info->hash_key_size = 0;
+ dev_info->reta_size = 0;
+ dev_info->flow_type_rss_offloads = 0;
+ }
+
if (host_features & (1ULL << VIRTIO_F_RING_PACKED)) {
/*
* According to 2.7 Packed Virtqueues,