X-Git-Url: http://git.droids-corp.org/?a=blobdiff_plain;f=drivers%2Fnet%2Fnull%2Frte_eth_null.c;h=ca03469d0e6dc9e5d28a4d304a17dcaa9f2bd581;hb=7e8d5583f7d34bcf7fe8fe9a5806b7d8366ded27;hp=2c08ebf8cdc35d661ab419684ce480adcd7fda91;hpb=276bb4cec184a1ea88fe2e8361ee5aaa68ed168c;p=dpdk.git diff --git a/drivers/net/null/rte_eth_null.c b/drivers/net/null/rte_eth_null.c index 2c08ebf8cd..ca03469d0e 100644 --- a/drivers/net/null/rte_eth_null.c +++ b/drivers/net/null/rte_eth_null.c @@ -4,8 +4,8 @@ */ #include -#include -#include +#include +#include #include #include #include @@ -14,13 +14,16 @@ #define ETH_NULL_PACKET_SIZE_ARG "size" #define ETH_NULL_PACKET_COPY_ARG "copy" +#define ETH_NULL_PACKET_NO_RX_ARG "no-rx" static unsigned int default_packet_size = 64; static unsigned int default_packet_copy; +static unsigned int default_no_rx; static const char *valid_arguments[] = { ETH_NULL_PACKET_SIZE_ARG, ETH_NULL_PACKET_COPY_ARG, + ETH_NULL_PACKET_NO_RX_ARG, NULL }; @@ -39,11 +42,13 @@ struct null_queue { struct pmd_options { unsigned int packet_copy; unsigned int packet_size; + unsigned int no_rx; }; struct pmd_internals { unsigned int packet_size; unsigned int packet_copy; + unsigned int no_rx; uint16_t port_id; struct null_queue rx_null_queues[RTE_MAX_QUEUES_PER_PORT]; @@ -56,19 +61,19 @@ struct pmd_internals { rte_spinlock_t rss_lock; uint16_t reta_size; - struct rte_eth_rss_reta_entry64 reta_conf[ETH_RSS_RETA_SIZE_128 / - RTE_RETA_GROUP_SIZE]; + struct rte_eth_rss_reta_entry64 reta_conf[RTE_ETH_RSS_RETA_SIZE_128 / + RTE_ETH_RETA_GROUP_SIZE]; uint8_t rss_key[40]; /**< 40-byte hash key. */ }; static struct rte_eth_link pmd_link = { - .link_speed = ETH_SPEED_NUM_10G, - .link_duplex = ETH_LINK_FULL_DUPLEX, - .link_status = ETH_LINK_DOWN, - .link_autoneg = ETH_LINK_FIXED, + .link_speed = RTE_ETH_SPEED_NUM_10G, + .link_duplex = RTE_ETH_LINK_FULL_DUPLEX, + .link_status = RTE_ETH_LINK_DOWN, + .link_autoneg = RTE_ETH_LINK_FIXED, }; -static int eth_null_logtype; +RTE_LOG_REGISTER_DEFAULT(eth_null_logtype, NOTICE); #define PMD_LOG(level, fmt, args...) \ rte_log(RTE_LOG_ ## level, eth_null_logtype, \ @@ -126,6 +131,13 @@ eth_null_copy_rx(void *q, struct rte_mbuf **bufs, uint16_t nb_bufs) return i; } +static uint16_t +eth_null_no_rx(void *q __rte_unused, struct rte_mbuf **bufs __rte_unused, + uint16_t nb_bufs __rte_unused) +{ + return 0; +} + static uint16_t eth_null_tx(void *q, struct rte_mbuf **bufs, uint16_t nb_bufs) { @@ -177,17 +189,19 @@ eth_dev_start(struct rte_eth_dev *dev) if (dev == NULL) return -EINVAL; - dev->data->dev_link.link_status = ETH_LINK_UP; + dev->data->dev_link.link_status = RTE_ETH_LINK_UP; return 0; } -static void +static int eth_dev_stop(struct rte_eth_dev *dev) { if (dev == NULL) - return; + return 0; + + dev->data->dev_link.link_status = RTE_ETH_LINK_DOWN; - dev->data->dev_link.link_status = ETH_LINK_DOWN; + return 0; } static int @@ -339,14 +353,24 @@ eth_stats_reset(struct rte_eth_dev *dev) } static void -eth_queue_release(void *q) +eth_rx_queue_release(struct rte_eth_dev *dev, uint16_t qid) { - struct null_queue *nq; + struct null_queue *nq = dev->data->rx_queues[qid]; - if (q == NULL) + if (nq == NULL) + return; + + rte_free(nq->dummy_packet); +} + +static void +eth_tx_queue_release(struct rte_eth_dev *dev, uint16_t qid) +{ + struct null_queue *nq = dev->data->tx_queues[qid]; + + if (nq == NULL) return; - nq = q; rte_free(nq->dummy_packet); } @@ -367,9 +391,9 @@ eth_rss_reta_update(struct rte_eth_dev *dev, rte_spinlock_lock(&internal->rss_lock); /* Copy RETA table */ - for (i = 0; i < (internal->reta_size / RTE_RETA_GROUP_SIZE); i++) { + for (i = 0; i < (internal->reta_size / RTE_ETH_RETA_GROUP_SIZE); i++) { internal->reta_conf[i].mask = reta_conf[i].mask; - for (j = 0; j < RTE_RETA_GROUP_SIZE; j++) + for (j = 0; j < RTE_ETH_RETA_GROUP_SIZE; j++) if ((reta_conf[i].mask >> j) & 0x01) internal->reta_conf[i].reta[j] = reta_conf[i].reta[j]; } @@ -392,8 +416,8 @@ eth_rss_reta_query(struct rte_eth_dev *dev, rte_spinlock_lock(&internal->rss_lock); /* Copy RETA table */ - for (i = 0; i < (internal->reta_size / RTE_RETA_GROUP_SIZE); i++) { - for (j = 0; j < RTE_RETA_GROUP_SIZE; j++) + for (i = 0; i < (internal->reta_size / RTE_ETH_RETA_GROUP_SIZE); i++) { + for (j = 0; j < RTE_ETH_RETA_GROUP_SIZE; j++) if ((reta_conf[i].mask >> j) & 0x01) reta_conf[i].reta[j] = internal->reta_conf[i].reta[j]; } @@ -446,15 +470,31 @@ eth_mac_address_set(__rte_unused struct rte_eth_dev *dev, return 0; } +static int +eth_dev_close(struct rte_eth_dev *dev) +{ + PMD_LOG(INFO, "Closing null ethdev on NUMA socket %u", + rte_socket_id()); + + if (rte_eal_process_type() != RTE_PROC_PRIMARY) + return 0; + + /* mac_addrs must not be freed alone because part of dev_private */ + dev->data->mac_addrs = NULL; + + return 0; +} + static const struct eth_dev_ops ops = { + .dev_close = eth_dev_close, .dev_start = eth_dev_start, .dev_stop = eth_dev_stop, .dev_configure = eth_dev_configure, .dev_infos_get = eth_dev_info, .rx_queue_setup = eth_rx_queue_setup, .tx_queue_setup = eth_tx_queue_setup, - .rx_queue_release = eth_queue_release, - .tx_queue_release = eth_queue_release, + .rx_queue_release = eth_rx_queue_release, + .tx_queue_release = eth_tx_queue_release, .mtu_set = eth_mtu_set, .link_update = eth_link_update, .mac_addr_set = eth_mac_address_set, @@ -504,11 +544,12 @@ eth_dev_null_create(struct rte_vdev_device *dev, struct pmd_options *args) internals = eth_dev->data->dev_private; internals->packet_size = args->packet_size; internals->packet_copy = args->packet_copy; + internals->no_rx = args->no_rx; internals->port_id = eth_dev->data->port_id; rte_eth_random_addr(internals->eth_addr.addr_bytes); - internals->flow_type_rss_offloads = ETH_RSS_PROTO_MASK; - internals->reta_size = RTE_DIM(internals->reta_conf) * RTE_RETA_GROUP_SIZE; + internals->flow_type_rss_offloads = RTE_ETH_RSS_PROTO_MASK; + internals->reta_size = RTE_DIM(internals->reta_conf) * RTE_ETH_RETA_GROUP_SIZE; rte_memcpy(internals->rss_key, default_rss_key, 40); @@ -519,6 +560,7 @@ eth_dev_null_create(struct rte_vdev_device *dev, struct pmd_options *args) data->mac_addrs = &internals->eth_addr; data->promiscuous = 1; data->all_multicast = 1; + data->dev_flags |= RTE_ETH_DEV_AUTOFILL_QUEUE_XSTATS; eth_dev->dev_ops = &ops; @@ -526,6 +568,9 @@ eth_dev_null_create(struct rte_vdev_device *dev, struct pmd_options *args) if (internals->packet_copy) { eth_dev->rx_pkt_burst = eth_null_copy_rx; eth_dev->tx_pkt_burst = eth_null_copy_tx; + } else if (internals->no_rx) { + eth_dev->rx_pkt_burst = eth_null_no_rx; + eth_dev->tx_pkt_burst = eth_null_tx; } else { eth_dev->rx_pkt_burst = eth_null_rx; eth_dev->tx_pkt_burst = eth_null_tx; @@ -569,6 +614,24 @@ get_packet_copy_arg(const char *key __rte_unused, return 0; } +static int +get_packet_no_rx_arg(const char *key __rte_unused, + const char *value, void *extra_args) +{ + const char *a = value; + unsigned int no_rx; + + if (value == NULL || extra_args == NULL) + return -EINVAL; + + no_rx = (unsigned int)strtoul(a, NULL, 0); + if (no_rx != 0 && no_rx != 1) + return -1; + + *(unsigned int *)extra_args = no_rx; + return 0; +} + static int rte_pmd_null_probe(struct rte_vdev_device *dev) { @@ -576,6 +639,7 @@ rte_pmd_null_probe(struct rte_vdev_device *dev) struct pmd_options args = { .packet_copy = default_packet_copy, .packet_size = default_packet_size, + .no_rx = default_no_rx, }; struct rte_kvargs *kvlist = NULL; struct rte_eth_dev *eth_dev; @@ -602,6 +666,9 @@ rte_pmd_null_probe(struct rte_vdev_device *dev) if (internals->packet_copy) { eth_dev->rx_pkt_burst = eth_null_copy_rx; eth_dev->tx_pkt_burst = eth_null_copy_tx; + } else if (internals->no_rx) { + eth_dev->rx_pkt_burst = eth_null_no_rx; + eth_dev->tx_pkt_burst = eth_null_tx; } else { eth_dev->rx_pkt_burst = eth_null_rx; eth_dev->tx_pkt_burst = eth_null_tx; @@ -627,6 +694,20 @@ rte_pmd_null_probe(struct rte_vdev_device *dev) &get_packet_copy_arg, &args.packet_copy); if (ret < 0) goto free_kvlist; + + ret = rte_kvargs_process(kvlist, + ETH_NULL_PACKET_NO_RX_ARG, + &get_packet_no_rx_arg, &args.no_rx); + if (ret < 0) + goto free_kvlist; + + if (args.no_rx && args.packet_copy) { + PMD_LOG(ERR, + "Both %s and %s arguments at the same time not supported", + ETH_NULL_PACKET_COPY_ARG, + ETH_NULL_PACKET_NO_RX_ARG); + goto free_kvlist; + } } PMD_LOG(INFO, "Configure pmd_null: packet size is %d, " @@ -649,18 +730,12 @@ rte_pmd_null_remove(struct rte_vdev_device *dev) if (!dev) return -EINVAL; - PMD_LOG(INFO, "Closing null ethdev on numa socket %u", - rte_socket_id()); - /* find the ethdev entry */ eth_dev = rte_eth_dev_allocated(rte_vdev_device_name(dev)); if (eth_dev == NULL) - return -1; - - if (rte_eal_process_type() == RTE_PROC_PRIMARY) - /* mac_addrs must not be freed alone because part of dev_private */ - eth_dev->data->mac_addrs = NULL; + return 0; /* port already released */ + eth_dev_close(eth_dev); rte_eth_dev_release_port(eth_dev); return 0; @@ -675,11 +750,5 @@ RTE_PMD_REGISTER_VDEV(net_null, pmd_null_drv); RTE_PMD_REGISTER_ALIAS(net_null, eth_null); RTE_PMD_REGISTER_PARAM_STRING(net_null, "size= " - "copy="); - -RTE_INIT(eth_null_init_log) -{ - eth_null_logtype = rte_log_register("pmd.net.null"); - if (eth_null_logtype >= 0) - rte_log_set_level(eth_null_logtype, RTE_LOG_NOTICE); -} + "copy= " + ETH_NULL_PACKET_NO_RX_ARG "=0|1");