X-Git-Url: http://git.droids-corp.org/?a=blobdiff_plain;f=drivers%2Fnet%2Ffailsafe%2Ffailsafe_ops.c;h=d36096587513f9dafd02fdf2d0504386527b824e;hb=9a0752f498d2725428480b4b004feeedd852c395;hp=223f26a3c7969f00f41c712d78cee929e6c32c9c;hpb=b737a1ee57b6d7a868f71ab97894398fac61963b;p=dpdk.git diff --git a/drivers/net/failsafe/failsafe_ops.c b/drivers/net/failsafe/failsafe_ops.c index 223f26a3c7..d360965875 100644 --- a/drivers/net/failsafe/failsafe_ops.c +++ b/drivers/net/failsafe/failsafe_ops.c @@ -34,6 +34,7 @@ #include #include +#include #include #include #include @@ -63,23 +64,175 @@ static struct rte_eth_dev_info default_infos = { .nb_seg_max = UINT16_MAX, .nb_mtu_seg_max = UINT16_MAX, }, - /* Set of understood capabilities */ - .rx_offload_capa = 0x0, + /* + * Set of capabilities that can be verified upon + * configuring a sub-device. + */ + .rx_offload_capa = + DEV_RX_OFFLOAD_VLAN_STRIP | + DEV_RX_OFFLOAD_QINQ_STRIP | + DEV_RX_OFFLOAD_IPV4_CKSUM | + DEV_RX_OFFLOAD_UDP_CKSUM | + DEV_RX_OFFLOAD_TCP_CKSUM | + DEV_RX_OFFLOAD_TCP_LRO, .tx_offload_capa = 0x0, .flow_type_rss_offloads = 0x0, }; +/** + * Check whether a specific offloading capability + * is supported by a sub_device. + * + * @return + * 0: all requested capabilities are supported by the sub_device + * positive value: This flag at least is not supported by the sub_device + */ +static int +fs_port_offload_validate(struct rte_eth_dev *dev, + struct sub_device *sdev) +{ + struct rte_eth_dev_info infos = {0}; + struct rte_eth_conf *cf; + uint32_t cap; + + cf = &dev->data->dev_conf; + SUBOPS(sdev, dev_infos_get)(ETH(sdev), &infos); + /* RX capabilities */ + cap = infos.rx_offload_capa; + if (cf->rxmode.hw_vlan_strip && + ((cap & DEV_RX_OFFLOAD_VLAN_STRIP) == 0)) { + WARN("VLAN stripping offload requested but not supported by sub_device %d", + SUB_ID(sdev)); + return DEV_RX_OFFLOAD_VLAN_STRIP; + } + if (cf->rxmode.hw_ip_checksum && + ((cap & (DEV_RX_OFFLOAD_IPV4_CKSUM | + DEV_RX_OFFLOAD_UDP_CKSUM | + DEV_RX_OFFLOAD_TCP_CKSUM)) != + (DEV_RX_OFFLOAD_IPV4_CKSUM | + DEV_RX_OFFLOAD_UDP_CKSUM | + DEV_RX_OFFLOAD_TCP_CKSUM))) { + WARN("IP checksum offload requested but not supported by sub_device %d", + SUB_ID(sdev)); + return DEV_RX_OFFLOAD_IPV4_CKSUM | + DEV_RX_OFFLOAD_UDP_CKSUM | + DEV_RX_OFFLOAD_TCP_CKSUM; + } + if (cf->rxmode.enable_lro && + ((cap & DEV_RX_OFFLOAD_TCP_LRO) == 0)) { + WARN("TCP LRO offload requested but not supported by sub_device %d", + SUB_ID(sdev)); + return DEV_RX_OFFLOAD_TCP_LRO; + } + if (cf->rxmode.hw_vlan_extend && + ((cap & DEV_RX_OFFLOAD_QINQ_STRIP) == 0)) { + WARN("Stacked VLAN stripping offload requested but not supported by sub_device %d", + SUB_ID(sdev)); + return DEV_RX_OFFLOAD_QINQ_STRIP; + } + /* TX capabilities */ + /* Nothing to do, no tx capa supported */ + return 0; +} + +/* + * Disable the dev_conf flag related to an offload capability flag + * within an ethdev configuration. + */ +static int +fs_port_disable_offload(struct rte_eth_conf *cf, + uint32_t ol_cap) +{ + switch (ol_cap) { + case DEV_RX_OFFLOAD_VLAN_STRIP: + INFO("Disabling VLAN stripping offload"); + cf->rxmode.hw_vlan_strip = 0; + break; + case DEV_RX_OFFLOAD_IPV4_CKSUM: + case DEV_RX_OFFLOAD_UDP_CKSUM: + case DEV_RX_OFFLOAD_TCP_CKSUM: + case (DEV_RX_OFFLOAD_IPV4_CKSUM | + DEV_RX_OFFLOAD_UDP_CKSUM | + DEV_RX_OFFLOAD_TCP_CKSUM): + INFO("Disabling IP checksum offload"); + cf->rxmode.hw_ip_checksum = 0; + break; + case DEV_RX_OFFLOAD_TCP_LRO: + INFO("Disabling TCP LRO offload"); + cf->rxmode.enable_lro = 0; + break; + case DEV_RX_OFFLOAD_QINQ_STRIP: + INFO("Disabling stacked VLAN stripping offload"); + cf->rxmode.hw_vlan_extend = 0; + break; + default: + DEBUG("Unable to disable offload capability: %" PRIx32, + ol_cap); + return -1; + } + return 0; +} + static int fs_dev_configure(struct rte_eth_dev *dev) { struct sub_device *sdev; uint8_t i; + int capa_flag; int ret; FOREACH_SUBDEV(sdev, i, dev) { if (sdev->state != DEV_PROBED) continue; + DEBUG("Checking capabilities for sub_device %d", i); + while ((capa_flag = fs_port_offload_validate(dev, sdev))) { + /* + * Refuse to change configuration if multiple devices + * are present and we already have configured at least + * some of them. + */ + if (PRIV(dev)->state >= DEV_ACTIVE && + PRIV(dev)->subs_tail > 1) { + ERROR("device already configured, cannot fix live configuration"); + return -1; + } + ret = fs_port_disable_offload(&dev->data->dev_conf, + capa_flag); + if (ret) { + ERROR("Unable to disable offload capability"); + return ret; + } + } + } + FOREACH_SUBDEV(sdev, i, dev) { + int rmv_interrupt = 0; + int lsc_interrupt = 0; + int lsc_enabled; + + if (sdev->state != DEV_PROBED) + continue; + + rmv_interrupt = ETH(sdev)->data->dev_flags & + RTE_ETH_DEV_INTR_RMV; + if (rmv_interrupt) { + DEBUG("Enabling RMV interrupts for sub_device %d", i); + dev->data->dev_conf.intr_conf.rmv = 1; + } else { + DEBUG("sub_device %d does not support RMV event", i); + } + lsc_enabled = dev->data->dev_conf.intr_conf.lsc; + lsc_interrupt = lsc_enabled && + (ETH(sdev)->data->dev_flags & + RTE_ETH_DEV_INTR_LSC); + if (lsc_interrupt) { + DEBUG("Enabling LSC interrupts for sub_device %d", i); + dev->data->dev_conf.intr_conf.lsc = 1; + } else if (lsc_enabled && !lsc_interrupt) { + DEBUG("Disabling LSC interrupts for sub_device %d", i); + dev->data->dev_conf.intr_conf.lsc = 0; + } DEBUG("Configuring sub-device %d", i); + sdev->remove = 0; ret = rte_eth_dev_configure(PORT_ID(sdev), dev->data->nb_rx_queues, dev->data->nb_tx_queues, @@ -88,6 +241,26 @@ fs_dev_configure(struct rte_eth_dev *dev) ERROR("Could not configure sub_device %d", i); return ret; } + if (rmv_interrupt) { + ret = rte_eth_dev_callback_register(PORT_ID(sdev), + RTE_ETH_EVENT_INTR_RMV, + failsafe_eth_rmv_event_callback, + sdev); + if (ret) + WARN("Failed to register RMV callback for sub_device %d", + SUB_ID(sdev)); + } + dev->data->dev_conf.intr_conf.rmv = 0; + if (lsc_interrupt) { + ret = rte_eth_dev_callback_register(PORT_ID(sdev), + RTE_ETH_EVENT_INTR_LSC, + failsafe_eth_lsc_event_callback, + dev); + if (ret) + WARN("Failed to register LSC callback for sub_device %d", + SUB_ID(sdev)); + } + dev->data->dev_conf.intr_conf.lsc = lsc_enabled; sdev->state = DEV_ACTIVE; } if (PRIV(dev)->state < DEV_ACTIVE) @@ -113,7 +286,7 @@ fs_dev_start(struct rte_eth_dev *dev) } if (PRIV(dev)->state < DEV_STARTED) PRIV(dev)->state = DEV_STARTED; - fs_switch_dev(dev); + fs_switch_dev(dev, NULL); return 0; } @@ -224,10 +397,14 @@ fs_rx_queue_setup(struct rte_eth_dev *dev, fs_rx_queue_release(rxq); dev->data->rx_queues[rx_queue_id] = NULL; } - rxq = rte_zmalloc(NULL, sizeof(*rxq), + rxq = rte_zmalloc(NULL, + sizeof(*rxq) + + sizeof(rte_atomic64_t) * PRIV(dev)->subs_tail, RTE_CACHE_LINE_SIZE); if (rxq == NULL) return -ENOMEM; + FOREACH_SUBDEV(sdev, i, dev) + rte_atomic64_init(&rxq->refcnt[i]); rxq->qid = rx_queue_id; rxq->socket_id = socket_id; rxq->info.mp = mb_pool; @@ -287,10 +464,14 @@ fs_tx_queue_setup(struct rte_eth_dev *dev, fs_tx_queue_release(txq); dev->data->tx_queues[tx_queue_id] = NULL; } - txq = rte_zmalloc("ethdev TX queue", sizeof(*txq), + txq = rte_zmalloc("ethdev TX queue", + sizeof(*txq) + + sizeof(rte_atomic64_t) * PRIV(dev)->subs_tail, RTE_CACHE_LINE_SIZE); if (txq == NULL) return -ENOMEM; + FOREACH_SUBDEV(sdev, i, dev) + rte_atomic64_init(&txq->refcnt[i]); txq->qid = tx_queue_id; txq->socket_id = socket_id; txq->info.conf = *tx_conf; @@ -401,13 +582,25 @@ fs_link_update(struct rte_eth_dev *dev, return -1; } -static void +static int fs_stats_get(struct rte_eth_dev *dev, struct rte_eth_stats *stats) { - if (TX_SUBDEV(dev) == NULL) - return; - rte_eth_stats_get(PORT_ID(TX_SUBDEV(dev)), stats); + struct sub_device *sdev; + uint8_t i; + int ret; + + rte_memcpy(stats, &PRIV(dev)->stats_accumulator, sizeof(*stats)); + FOREACH_SUBDEV_STATE(sdev, i, dev, DEV_ACTIVE) { + ret = rte_eth_stats_get(PORT_ID(sdev), &sdev->stats_snapshot); + if (ret) { + ERROR("Operation rte_eth_stats_get failed for sub_device %d with error %d", + i, ret); + return ret; + } + failsafe_stats_increment(stats, &sdev->stats_snapshot); + } + return 0; } static void @@ -416,8 +609,11 @@ fs_stats_reset(struct rte_eth_dev *dev) struct sub_device *sdev; uint8_t i; - FOREACH_SUBDEV_STATE(sdev, i, dev, DEV_ACTIVE) + FOREACH_SUBDEV_STATE(sdev, i, dev, DEV_ACTIVE) { rte_eth_stats_reset(PORT_ID(sdev)); + memset(&sdev->stats_snapshot, 0, sizeof(struct rte_eth_stats)); + } + memset(&PRIV(dev)->stats_accumulator, 0, sizeof(struct rte_eth_stats)); } /**