X-Git-Url: http://git.droids-corp.org/?a=blobdiff_plain;f=app%2Ftest-pmd%2Ftestpmd.c;h=6d5bbc82404e7279c3d19128d97a0faea6b255c7;hb=5984037501c144ee2535ed57e27029554aa97dd7;hp=99bacddbfdcafa763ba5fe63ad265f76a519fd26;hpb=99e040d3adc9ee2284e315219f2a6baee680f4e3;p=dpdk.git diff --git a/app/test-pmd/testpmd.c b/app/test-pmd/testpmd.c index 99bacddbfd..6d5bbc8240 100644 --- a/app/test-pmd/testpmd.c +++ b/app/test-pmd/testpmd.c @@ -9,7 +9,9 @@ #include #include #include +#ifndef RTE_EXEC_ENV_WINDOWS #include +#endif #include #include #include @@ -46,20 +48,23 @@ #include #include #include -#ifdef RTE_LIBRTE_IXGBE_PMD +#ifdef RTE_NET_IXGBE #include #endif -#ifdef RTE_LIBRTE_PDUMP +#ifdef RTE_LIB_PDUMP #include #endif #include #include -#ifdef RTE_LIBRTE_BITRATE +#ifdef RTE_LIB_BITRATESTATS #include #endif -#ifdef RTE_LIBRTE_LATENCY_STATS +#ifdef RTE_LIB_LATENCYSTATS #include #endif +#ifdef RTE_EXEC_ENV_WINDOWS +#include +#endif #include "testpmd.h" @@ -83,7 +88,7 @@ uint16_t verbose_level = 0; /**< Silent by default. */ int testpmd_logtype; /**< Log type for testpmd logs */ -/* use master core for command line ? */ +/* use main core for command line ? */ uint8_t interactive = 0; uint8_t auto_start = 0; uint8_t tx_first; @@ -179,16 +184,15 @@ struct fwd_engine * fwd_engines[] = { &csum_fwd_engine, &icmp_echo_engine, &noisy_vnf_engine, -#if defined RTE_LIBRTE_PMD_SOFTNIC - &softnic_fwd_engine, -#endif + &five_tuple_swap_fwd_engine, #ifdef RTE_LIBRTE_IEEE1588 &ieee1588_fwd_engine, #endif + &shared_rxq_engine, NULL, }; -struct rte_mempool *mempools[RTE_MAX_NUMA_NODES]; +struct rte_mempool *mempools[RTE_MAX_NUMA_NODES * MAX_SEGS_BUFFER_SPLIT]; uint16_t mempool_flags; struct fwd_config cur_fwd_config; @@ -197,17 +201,39 @@ uint32_t retry_enabled; uint32_t burst_tx_delay_time = BURST_TX_WAIT_US; uint32_t burst_tx_retry_num = BURST_TX_RETRIES; -uint16_t mbuf_data_size = DEFAULT_MBUF_DATA_SIZE; /**< Mbuf data space size. */ +uint32_t mbuf_data_size_n = 1; /* Number of specified mbuf sizes. */ +uint16_t mbuf_data_size[MAX_SEGS_BUFFER_SPLIT] = { + DEFAULT_MBUF_DATA_SIZE +}; /**< Mbuf data space size. */ uint32_t param_total_num_mbufs = 0; /**< number of mbufs in all pools - if * specified on command-line. */ uint16_t stats_period; /**< Period to show statistics (disabled by default) */ +/** Extended statistics to show. */ +struct rte_eth_xstat_name *xstats_display; + +unsigned int xstats_display_num; /**< Size of extended statistics to show */ + /* * In container, it cannot terminate the process which running with 'stats-period' * option. Set flag to exit stats period loop after received SIGINT/SIGTERM. */ uint8_t f_quit; +/* + * Max Rx frame size, set by '--max-pkt-len' parameter. + */ +uint32_t max_rx_pkt_len; + +/* + * Configuration of packet segments used to scatter received packets + * if some of split features is configured. + */ +uint16_t rx_pkt_seg_lengths[MAX_SEGS_BUFFER_SPLIT]; +uint8_t rx_pkt_nb_segs; /**< Number of segments to split */ +uint16_t rx_pkt_seg_offsets[MAX_SEGS_BUFFER_SPLIT]; +uint8_t rx_pkt_nb_offs; /**< Number of specified offsets */ + /* * Configuration of packet segments used by the "txonly" processing engine. */ @@ -223,15 +249,20 @@ enum tx_pkt_split tx_pkt_split = TX_PKT_SPLIT_OFF; uint8_t txonly_multi_flow; /**< Whether multiple flows are generated in TXONLY mode. */ +uint32_t tx_pkt_times_inter; +/**< Timings for send scheduling in TXONLY mode, time between bursts. */ + +uint32_t tx_pkt_times_intra; +/**< Timings for send scheduling in TXONLY mode, time between packets. */ + uint16_t nb_pkt_per_burst = DEF_PKT_BURST; /**< Number of packets per burst. */ +uint16_t nb_pkt_flowgen_clones; /**< Number of Tx packet clones to send in flowgen mode. */ +int nb_flows_flowgen = 1024; /**< Number of flows in flowgen mode. */ uint16_t mb_mempool_cache = DEF_MBUF_CACHE; /**< Size of mbuf mempool cache. */ /* current configuration is in DCB or not,0 means it is not in DCB mode */ uint8_t dcb_config = 0; -/* Whether the dcb is in testing status */ -uint8_t dcb_test = 0; - /* * Configurable number of RX/TX queues. */ @@ -363,6 +394,9 @@ bool setup_on_probe_event = true; /* Clear ptypes on port initialization. */ uint8_t clear_ptypes = true; +/* Hairpin ports configuration mode. */ +uint16_t hairpin_mode; + /* Pretty printing of ethdev events */ static const char * const eth_event_desc[] = { [RTE_ETH_EVENT_UNKNOWN] = "unknown", @@ -375,6 +409,7 @@ static const char * const eth_event_desc[] = { [RTE_ETH_EVENT_INTR_RMV] = "device removal", [RTE_ETH_EVENT_NEW] = "device probed", [RTE_ETH_EVENT_DESTROY] = "device released", + [RTE_ETH_EVENT_FLOW_AGED] = "flow aged", [RTE_ETH_EVENT_MAX] = NULL, }; @@ -388,7 +423,8 @@ uint32_t event_print_mask = (UINT32_C(1) << RTE_ETH_EVENT_UNKNOWN) | (UINT32_C(1) << RTE_ETH_EVENT_INTR_RESET) | (UINT32_C(1) << RTE_ETH_EVENT_IPSEC) | (UINT32_C(1) << RTE_ETH_EVENT_MACSEC) | - (UINT32_C(1) << RTE_ETH_EVENT_INTR_RMV); + (UINT32_C(1) << RTE_ETH_EVENT_INTR_RMV) | + (UINT32_C(1) << RTE_ETH_EVENT_FLOW_AGED); /* * Decide if all memory are locked for performance. */ @@ -398,13 +434,13 @@ int do_mlockall = 0; * NIC bypass mode configuration options. */ -#if defined RTE_LIBRTE_IXGBE_PMD && defined RTE_LIBRTE_IXGBE_BYPASS +#if defined RTE_NET_IXGBE && defined RTE_LIBRTE_IXGBE_BYPASS /* The NIC bypass watchdog timeout. */ uint32_t bypass_timeout = RTE_PMD_IXGBE_BYPASS_TMT_OFF; #endif -#ifdef RTE_LIBRTE_LATENCY_STATS +#ifdef RTE_LIB_LATENCYSTATS /* * Set when latency stats is enabled in the commandline @@ -421,10 +457,7 @@ lcoreid_t latencystats_lcore_id = -1; /* * Ethernet device configuration. */ -struct rte_eth_rxmode rx_mode = { - .max_rx_pkt_len = RTE_ETHER_MAX_LEN, - /**< Default maximum frame length. */ -}; +struct rte_eth_rxmode rx_mode; struct rte_eth_txmode tx_mode = { .offloads = DEV_TX_OFFLOAD_MBUF_FAST_FREE, @@ -455,24 +488,30 @@ struct rte_fdir_conf fdir_conf = { volatile int test_done = 1; /* stop packet forwarding when set to 1. */ -struct queue_stats_mappings tx_queue_stats_mappings_array[MAX_TX_QUEUE_STATS_MAPPINGS]; -struct queue_stats_mappings rx_queue_stats_mappings_array[MAX_RX_QUEUE_STATS_MAPPINGS]; +/* + * Display zero values by default for xstats + */ +uint8_t xstats_hide_zero; -struct queue_stats_mappings *tx_queue_stats_mappings = tx_queue_stats_mappings_array; -struct queue_stats_mappings *rx_queue_stats_mappings = rx_queue_stats_mappings_array; +/* + * Measure of CPU cycles disabled by default + */ +uint8_t record_core_cycles; -uint16_t nb_tx_queue_stats_mappings = 0; -uint16_t nb_rx_queue_stats_mappings = 0; +/* + * Display of RX and TX bursts disabled by default + */ +uint8_t record_burst_stats; /* - * Display zero values by default for xstats + * Number of ports per shared Rx queue group, 0 disable. */ -uint8_t xstats_hide_zero; +uint32_t rxq_share; unsigned int num_sockets = 0; unsigned int socket_ids[RTE_MAX_NUMA_NODES]; -#ifdef RTE_LIBRTE_BITRATE +#ifdef RTE_LIB_BITRATESTATS /* Bitrate statistics */ struct rte_stats_bitrates *bitrate_data; lcoreid_t bitrate_lcore_id; @@ -482,10 +521,128 @@ uint8_t bitrate_enabled; struct gro_status gro_ports[RTE_MAX_ETHPORTS]; uint8_t gro_flush_cycles = GRO_DEFAULT_FLUSH_CYCLES; +/* + * hexadecimal bitmask of RX mq mode can be enabled. + */ +enum rte_eth_rx_mq_mode rx_mq_mode = ETH_MQ_RX_VMDQ_DCB_RSS; + +/* + * Used to set forced link speed + */ +uint32_t eth_link_speed; + +/* + * ID of the current process in multi-process, used to + * configure the queues to be polled. + */ +int proc_id; + +/* + * Number of processes in multi-process, used to + * configure the queues to be polled. + */ +unsigned int num_procs = 1; + +static void +eth_rx_metadata_negotiate_mp(uint16_t port_id) +{ + uint64_t rx_meta_features = 0; + int ret; + + if (!is_proc_primary()) + return; + + rx_meta_features |= RTE_ETH_RX_METADATA_USER_FLAG; + rx_meta_features |= RTE_ETH_RX_METADATA_USER_MARK; + rx_meta_features |= RTE_ETH_RX_METADATA_TUNNEL_ID; + + ret = rte_eth_rx_metadata_negotiate(port_id, &rx_meta_features); + if (ret == 0) { + if (!(rx_meta_features & RTE_ETH_RX_METADATA_USER_FLAG)) { + TESTPMD_LOG(DEBUG, "Flow action FLAG will not affect Rx mbufs on port %u\n", + port_id); + } + + if (!(rx_meta_features & RTE_ETH_RX_METADATA_USER_MARK)) { + TESTPMD_LOG(DEBUG, "Flow action MARK will not affect Rx mbufs on port %u\n", + port_id); + } + + if (!(rx_meta_features & RTE_ETH_RX_METADATA_TUNNEL_ID)) { + TESTPMD_LOG(DEBUG, "Flow tunnel offload support might be limited or unavailable on port %u\n", + port_id); + } + } else if (ret != -ENOTSUP) { + rte_exit(EXIT_FAILURE, "Error when negotiating Rx meta features on port %u: %s\n", + port_id, rte_strerror(-ret)); + } +} + +static void +flow_pick_transfer_proxy_mp(uint16_t port_id) +{ + struct rte_port *port = &ports[port_id]; + int ret; + + port->flow_transfer_proxy = port_id; + + if (!is_proc_primary()) + return; + + ret = rte_flow_pick_transfer_proxy(port_id, &port->flow_transfer_proxy, + NULL); + if (ret != 0) { + fprintf(stderr, "Error picking flow transfer proxy for port %u: %s - ignore\n", + port_id, rte_strerror(-ret)); + } +} + +static int +eth_dev_configure_mp(uint16_t port_id, uint16_t nb_rx_q, uint16_t nb_tx_q, + const struct rte_eth_conf *dev_conf) +{ + if (is_proc_primary()) + return rte_eth_dev_configure(port_id, nb_rx_q, nb_tx_q, + dev_conf); + return 0; +} + +static int +eth_dev_start_mp(uint16_t port_id) +{ + if (is_proc_primary()) + return rte_eth_dev_start(port_id); + + return 0; +} + +static int +eth_dev_stop_mp(uint16_t port_id) +{ + if (is_proc_primary()) + return rte_eth_dev_stop(port_id); + + return 0; +} + +static void +mempool_free_mp(struct rte_mempool *mp) +{ + if (is_proc_primary()) + rte_mempool_free(mp); +} + +static int +eth_dev_set_mtu_mp(uint16_t port_id, uint16_t mtu) +{ + if (is_proc_primary()) + return rte_eth_dev_set_mtu(port_id, mtu); + + return 0; +} + /* Forward function declarations */ static void setup_attached_port(portid_t pi); -static void map_port_queue_stats_mapping_registers(portid_t pi, - struct rte_port *port); static void check_all_ports_link_status(uint32_t port_mask); static int eth_event_callback(portid_t port_id, enum rte_eth_event_type type, @@ -493,6 +650,7 @@ static int eth_event_callback(portid_t port_id, static void dev_event_callback(const char *device_name, enum rte_dev_event_type type, void *param); +static void fill_xstats_display_info(void); /* * Check if all the ports are started. @@ -506,6 +664,7 @@ uint16_t gso_max_segment_size = RTE_ETHER_MAX_LEN - RTE_ETHER_CRC_LEN; /* Holds the registered mbuf dynamic flags names. */ char dynf_names[64][RTE_MBUF_DYN_NAMESIZE]; + /* * Helper function to check if socket is already discovered. * If yes, return positive value. If not, return zero. @@ -545,7 +704,7 @@ set_default_fwd_lcores_config(void) } socket_ids[num_sockets++] = sock_num; } - if (i == rte_get_master_lcore()) + if (i == rte_get_main_lcore()) continue; fwd_lcores_cpuids[nb_lc++] = i; } @@ -598,6 +757,7 @@ set_def_fwd_config(void) set_default_fwd_ports_config(); } +#ifndef RTE_EXEC_ENV_WINDOWS /* extremely pessimistic estimation of memory required to create a mempool */ static int calc_mem_size(uint32_t nb_mbufs, uint32_t mbuf_sz, size_t pgsz, size_t *out) @@ -820,16 +980,23 @@ dma_unmap_cb(struct rte_mempool *mp __rte_unused, void *opaque __rte_unused, int ret; RTE_ETH_FOREACH_DEV(pid) { - struct rte_eth_dev *dev = - &rte_eth_devices[pid]; + struct rte_eth_dev_info dev_info; + + ret = eth_dev_info_get_print_err(pid, &dev_info); + if (ret != 0) { + TESTPMD_LOG(DEBUG, + "unable to get device info for port %d on addr 0x%p," + "mempool unmapping will not be performed\n", + pid, memhdr->addr); + continue; + } - ret = rte_dev_dma_unmap(dev->device, memhdr->addr, 0, - memhdr->len); + ret = rte_dev_dma_unmap(dev_info.device, memhdr->addr, 0, memhdr->len); if (ret) { TESTPMD_LOG(DEBUG, "unable to DMA unmap addr 0x%p " "for device %s\n", - memhdr->addr, dev->data->name); + memhdr->addr, dev_info.device->name); } } ret = rte_extmem_unregister(memhdr->addr, memhdr->len); @@ -855,19 +1022,26 @@ dma_map_cb(struct rte_mempool *mp __rte_unused, void *opaque __rte_unused, return; } RTE_ETH_FOREACH_DEV(pid) { - struct rte_eth_dev *dev = - &rte_eth_devices[pid]; + struct rte_eth_dev_info dev_info; - ret = rte_dev_dma_map(dev->device, memhdr->addr, 0, - memhdr->len); + ret = eth_dev_info_get_print_err(pid, &dev_info); + if (ret != 0) { + TESTPMD_LOG(DEBUG, + "unable to get device info for port %d on addr 0x%p," + "mempool mapping will not be performed\n", + pid, memhdr->addr); + continue; + } + ret = rte_dev_dma_map(dev_info.device, memhdr->addr, 0, memhdr->len); if (ret) { TESTPMD_LOG(DEBUG, "unable to DMA map addr 0x%p " "for device %s\n", - memhdr->addr, dev->data->name); + memhdr->addr, dev_info.device->name); } } } +#endif static unsigned int setup_extbuf(uint32_t nb_mbufs, uint16_t mbuf_sz, unsigned int socket_id, @@ -934,14 +1108,24 @@ setup_extbuf(uint32_t nb_mbufs, uint16_t mbuf_sz, unsigned int socket_id, */ static struct rte_mempool * mbuf_pool_create(uint16_t mbuf_seg_size, unsigned nb_mbuf, - unsigned int socket_id) + unsigned int socket_id, uint16_t size_idx) { char pool_name[RTE_MEMPOOL_NAMESIZE]; struct rte_mempool *rte_mp = NULL; +#ifndef RTE_EXEC_ENV_WINDOWS uint32_t mb_size; mb_size = sizeof(struct rte_mbuf) + mbuf_seg_size; - mbuf_poolname_build(socket_id, pool_name, sizeof(pool_name)); +#endif + mbuf_poolname_build(socket_id, pool_name, sizeof(pool_name), size_idx); + if (!is_proc_primary()) { + rte_mp = rte_mempool_lookup(pool_name); + if (rte_mp == NULL) + rte_exit(EXIT_FAILURE, + "Get mbuf pool for socket %u failed: %s\n", + socket_id, rte_strerror(rte_errno)); + return rte_mp; + } TESTPMD_LOG(INFO, "create a new mbuf pool <%s>: n=%u, size=%u, socket=%u\n", @@ -957,6 +1141,7 @@ mbuf_pool_create(uint16_t mbuf_seg_size, unsigned nb_mbuf, mb_mempool_cache, 0, mbuf_seg_size, socket_id); break; } +#ifndef RTE_EXEC_ENV_WINDOWS case MP_ALLOC_ANON: { rte_mp = rte_mempool_create_empty(pool_name, nb_mbuf, @@ -997,6 +1182,7 @@ mbuf_pool_create(uint16_t mbuf_seg_size, unsigned nb_mbuf, heap_socket); break; } +#endif case MP_ALLOC_XBUF: { struct rte_pktmbuf_extmem *ext_mem; @@ -1023,7 +1209,9 @@ mbuf_pool_create(uint16_t mbuf_seg_size, unsigned nb_mbuf, } } +#ifndef RTE_EXEC_ENV_WINDOWS err: +#endif if (rte_mp == NULL) { rte_exit(EXIT_FAILURE, "Creation of mbuf pool for socket %u failed: %s\n", @@ -1045,10 +1233,8 @@ check_socket_id(const unsigned int socket_id) if (new_socket_id(socket_id)) { if (!warning_once && numa_support) - printf("Warning: NUMA should be configured manually by" - " using --port-numa-config and" - " --ring-numa-config parameters along with" - " --numa.\n"); + fprintf(stderr, + "Warning: NUMA should be configured manually by using --port-numa-config and --ring-numa-config parameters along with --numa.\n"); warning_once = 1; return -1; } @@ -1095,11 +1281,9 @@ check_nb_rxq(queueid_t rxq) allowed_max_rxq = get_allowed_max_nb_rxq(&pid); if (rxq > allowed_max_rxq) { - printf("Fail: input rxq (%u) can't be greater " - "than max_rx_queues (%u) of port %u\n", - rxq, - allowed_max_rxq, - pid); + fprintf(stderr, + "Fail: input rxq (%u) can't be greater than max_rx_queues (%u) of port %u\n", + rxq, allowed_max_rxq, pid); return -1; } return 0; @@ -1145,11 +1329,9 @@ check_nb_txq(queueid_t txq) allowed_max_txq = get_allowed_max_nb_txq(&pid); if (txq > allowed_max_txq) { - printf("Fail: input txq (%u) can't be greater " - "than max_tx_queues (%u) of port %u\n", - txq, - allowed_max_txq, - pid); + fprintf(stderr, + "Fail: input txq (%u) can't be greater than max_tx_queues (%u) of port %u\n", + txq, allowed_max_txq, pid); return -1; } return 0; @@ -1220,21 +1402,17 @@ check_nb_rxd(queueid_t rxd) allowed_max_rxd = get_allowed_max_nb_rxd(&pid); if (rxd > allowed_max_rxd) { - printf("Fail: input rxd (%u) can't be greater " - "than max_rxds (%u) of port %u\n", - rxd, - allowed_max_rxd, - pid); + fprintf(stderr, + "Fail: input rxd (%u) can't be greater than max_rxds (%u) of port %u\n", + rxd, allowed_max_rxd, pid); return -1; } allowed_min_rxd = get_allowed_min_nb_rxd(&pid); if (rxd < allowed_min_rxd) { - printf("Fail: input rxd (%u) can't be less " - "than min_rxds (%u) of port %u\n", - rxd, - allowed_min_rxd, - pid); + fprintf(stderr, + "Fail: input rxd (%u) can't be less than min_rxds (%u) of port %u\n", + rxd, allowed_min_rxd, pid); return -1; } @@ -1305,21 +1483,17 @@ check_nb_txd(queueid_t txd) allowed_max_txd = get_allowed_max_nb_txd(&pid); if (txd > allowed_max_txd) { - printf("Fail: input txd (%u) can't be greater " - "than max_txds (%u) of port %u\n", - txd, - allowed_max_txd, - pid); + fprintf(stderr, + "Fail: input txd (%u) can't be greater than max_txds (%u) of port %u\n", + txd, allowed_max_txd, pid); return -1; } allowed_min_txd = get_allowed_min_nb_txd(&pid); if (txd < allowed_min_txd) { - printf("Fail: input txd (%u) can't be less " - "than min_txds (%u) of port %u\n", - txd, - allowed_min_txd, - pid); + fprintf(stderr, + "Fail: input txd (%u) can't be less than min_txds (%u) of port %u\n", + txd, allowed_min_txd, pid); return -1; } return 0; @@ -1365,31 +1539,102 @@ check_nb_hairpinq(queueid_t hairpinq) allowed_max_hairpinq = get_allowed_max_nb_hairpinq(&pid); if (hairpinq > allowed_max_hairpinq) { - printf("Fail: input hairpin (%u) can't be greater " - "than max_hairpin_queues (%u) of port %u\n", - hairpinq, allowed_max_hairpinq, pid); + fprintf(stderr, + "Fail: input hairpin (%u) can't be greater than max_hairpin_queues (%u) of port %u\n", + hairpinq, allowed_max_hairpinq, pid); return -1; } return 0; } +static int +get_eth_overhead(struct rte_eth_dev_info *dev_info) +{ + uint32_t eth_overhead; + + if (dev_info->max_mtu != UINT16_MAX && + dev_info->max_rx_pktlen > dev_info->max_mtu) + eth_overhead = dev_info->max_rx_pktlen - dev_info->max_mtu; + else + eth_overhead = RTE_ETHER_HDR_LEN + RTE_ETHER_CRC_LEN; + + return eth_overhead; +} + +static void +init_config_port_offloads(portid_t pid, uint32_t socket_id) +{ + struct rte_port *port = &ports[pid]; + int ret; + int i; + + eth_rx_metadata_negotiate_mp(pid); + flow_pick_transfer_proxy_mp(pid); + + port->dev_conf.txmode = tx_mode; + port->dev_conf.rxmode = rx_mode; + + ret = eth_dev_info_get_print_err(pid, &port->dev_info); + if (ret != 0) + rte_exit(EXIT_FAILURE, "rte_eth_dev_info_get() failed\n"); + + if (!(port->dev_info.tx_offload_capa & DEV_TX_OFFLOAD_MBUF_FAST_FREE)) + port->dev_conf.txmode.offloads &= + ~DEV_TX_OFFLOAD_MBUF_FAST_FREE; + + /* Apply Rx offloads configuration */ + for (i = 0; i < port->dev_info.max_rx_queues; i++) + port->rx_conf[i].offloads = port->dev_conf.rxmode.offloads; + /* Apply Tx offloads configuration */ + for (i = 0; i < port->dev_info.max_tx_queues; i++) + port->tx_conf[i].offloads = port->dev_conf.txmode.offloads; + + if (eth_link_speed) + port->dev_conf.link_speeds = eth_link_speed; + + if (max_rx_pkt_len) + port->dev_conf.rxmode.mtu = max_rx_pkt_len - + get_eth_overhead(&port->dev_info); + + /* set flag to initialize port/queue */ + port->need_reconfig = 1; + port->need_reconfig_queues = 1; + port->socket_id = socket_id; + port->tx_metadata = 0; + + /* + * Check for maximum number of segments per MTU. + * Accordingly update the mbuf data size. + */ + if (port->dev_info.rx_desc_lim.nb_mtu_seg_max != UINT16_MAX && + port->dev_info.rx_desc_lim.nb_mtu_seg_max != 0) { + uint32_t eth_overhead = get_eth_overhead(&port->dev_info); + uint16_t mtu; + + if (rte_eth_dev_get_mtu(pid, &mtu) == 0) { + uint16_t data_size = (mtu + eth_overhead) / + port->dev_info.rx_desc_lim.nb_mtu_seg_max; + uint16_t buffer_size = data_size + RTE_PKTMBUF_HEADROOM; + + if (buffer_size > mbuf_data_size[0]) { + mbuf_data_size[0] = buffer_size; + TESTPMD_LOG(WARNING, + "Configured mbuf size of the first segment %hu\n", + mbuf_data_size[0]); + } + } + } +} + static void init_config(void) { portid_t pid; - struct rte_port *port; struct rte_mempool *mbp; unsigned int nb_mbuf_per_pool; lcoreid_t lc_id; - uint8_t port_per_socket[RTE_MAX_NUMA_NODES]; struct rte_gro_param gro_param; uint32_t gso_types; - uint16_t data_size; - bool warning = 0; - int k; - int ret; - - memset(port_per_socket,0,RTE_MAX_NUMA_NODES); /* Configuration of logical cores. */ fwd_lcores = rte_zmalloc("testpmd: fwd_lcores", @@ -1411,25 +1656,12 @@ init_config(void) } RTE_ETH_FOREACH_DEV(pid) { - port = &ports[pid]; - /* Apply default TxRx configuration for all ports */ - port->dev_conf.txmode = tx_mode; - port->dev_conf.rxmode = rx_mode; - - ret = eth_dev_info_get_print_err(pid, &port->dev_info); - if (ret != 0) - rte_exit(EXIT_FAILURE, - "rte_eth_dev_info_get() failed\n"); + uint32_t socket_id; - if (!(port->dev_info.tx_offload_capa & - DEV_TX_OFFLOAD_MBUF_FAST_FREE)) - port->dev_conf.txmode.offloads &= - ~DEV_TX_OFFLOAD_MBUF_FAST_FREE; if (numa_support) { - if (port_numa[pid] != NUMA_NO_CONFIG) - port_per_socket[port_numa[pid]]++; - else { - uint32_t socket_id = rte_eth_dev_socket_id(pid); + socket_id = port_numa[pid]; + if (port_numa[pid] == NUMA_NO_CONFIG) { + socket_id = rte_eth_dev_socket_id(pid); /* * if socket_id is invalid, @@ -1437,45 +1669,14 @@ init_config(void) */ if (check_socket_id(socket_id) < 0) socket_id = socket_ids[0]; - port_per_socket[socket_id]++; - } - } - - /* Apply Rx offloads configuration */ - for (k = 0; k < port->dev_info.max_rx_queues; k++) - port->rx_conf[k].offloads = - port->dev_conf.rxmode.offloads; - /* Apply Tx offloads configuration */ - for (k = 0; k < port->dev_info.max_tx_queues; k++) - port->tx_conf[k].offloads = - port->dev_conf.txmode.offloads; - - /* set flag to initialize port/queue */ - port->need_reconfig = 1; - port->need_reconfig_queues = 1; - port->tx_metadata = 0; - - /* Check for maximum number of segments per MTU. Accordingly - * update the mbuf data size. - */ - if (port->dev_info.rx_desc_lim.nb_mtu_seg_max != UINT16_MAX && - port->dev_info.rx_desc_lim.nb_mtu_seg_max != 0) { - data_size = rx_mode.max_rx_pkt_len / - port->dev_info.rx_desc_lim.nb_mtu_seg_max; - - if ((data_size + RTE_PKTMBUF_HEADROOM) > - mbuf_data_size) { - mbuf_data_size = data_size + - RTE_PKTMBUF_HEADROOM; - warning = 1; } + } else { + socket_id = (socket_num == UMA_NO_CONFIG) ? + 0 : socket_num; } + /* Apply default TxRx configuration for all ports */ + init_config_port_offloads(pid, socket_id); } - - if (warning) - TESTPMD_LOG(WARNING, "Configured mbuf size %hu\n", - mbuf_data_size); - /* * Create pools of mbuf. * If NUMA support is disabled, create a single pool of mbuf in @@ -1495,21 +1696,23 @@ init_config(void) } if (numa_support) { - uint8_t i; + uint8_t i, j; for (i = 0; i < num_sockets; i++) - mempools[i] = mbuf_pool_create(mbuf_data_size, - nb_mbuf_per_pool, - socket_ids[i]); + for (j = 0; j < mbuf_data_size_n; j++) + mempools[i * MAX_SEGS_BUFFER_SPLIT + j] = + mbuf_pool_create(mbuf_data_size[j], + nb_mbuf_per_pool, + socket_ids[i], j); } else { - if (socket_num == UMA_NO_CONFIG) - mempools[0] = mbuf_pool_create(mbuf_data_size, - nb_mbuf_per_pool, 0); - else - mempools[socket_num] = mbuf_pool_create - (mbuf_data_size, - nb_mbuf_per_pool, - socket_num); + uint8_t i; + + for (i = 0; i < mbuf_data_size_n; i++) + mempools[i] = mbuf_pool_create + (mbuf_data_size[i], + nb_mbuf_per_pool, + socket_num == UMA_NO_CONFIG ? + 0 : socket_num, i); } init_port_config(); @@ -1521,10 +1724,10 @@ init_config(void) */ for (lc_id = 0; lc_id < nb_lcores; lc_id++) { mbp = mbuf_pool_find( - rte_lcore_to_socket_id(fwd_lcores_cpuids[lc_id])); + rte_lcore_to_socket_id(fwd_lcores_cpuids[lc_id]), 0); if (mbp == NULL) - mbp = mbuf_pool_find(0); + mbp = mbuf_pool_find(0, 0); fwd_lcores[lc_id]->mbp = mbp; /* initialize GSO context */ fwd_lcores[lc_id]->gso_ctx.direct_pool = mbp; @@ -1535,10 +1738,6 @@ init_config(void) fwd_lcores[lc_id]->gso_ctx.flag = 0; } - /* Configuration of packet forwarding streams. */ - if (init_fwd_streams() < 0) - rte_exit(EXIT_FAILURE, "FAIL from init_fwd_streams()\n"); - fwd_config_setup(); /* create a gro context for each lcore */ @@ -1554,40 +1753,14 @@ init_config(void) "rte_gro_ctx_create() failed\n"); } } - -#if defined RTE_LIBRTE_PMD_SOFTNIC - if (strcmp(cur_fwd_eng->fwd_mode_name, "softnic") == 0) { - RTE_ETH_FOREACH_DEV(pid) { - port = &ports[pid]; - const char *driver = port->dev_info.driver_name; - - if (strcmp(driver, "net_softnic") == 0) - port->softport.fwd_lcore_arg = fwd_lcores; - } - } -#endif - } void reconfig(portid_t new_port_id, unsigned socket_id) { - struct rte_port *port; - int ret; - /* Reconfiguration of Ethernet ports. */ - port = &ports[new_port_id]; - - ret = eth_dev_info_get_print_err(new_port_id, &port->dev_info); - if (ret != 0) - return; - - /* set flag to initialize port/queue */ - port->need_reconfig = 1; - port->need_reconfig_queues = 1; - port->socket_id = socket_id; - + init_config_port_offloads(new_port_id, socket_id); init_port_config(); } @@ -1604,15 +1777,15 @@ init_fwd_streams(void) RTE_ETH_FOREACH_DEV(pid) { port = &ports[pid]; if (nb_rxq > port->dev_info.max_rx_queues) { - printf("Fail: nb_rxq(%d) is greater than " - "max_rx_queues(%d)\n", nb_rxq, - port->dev_info.max_rx_queues); + fprintf(stderr, + "Fail: nb_rxq(%d) is greater than max_rx_queues(%d)\n", + nb_rxq, port->dev_info.max_rx_queues); return -1; } if (nb_txq > port->dev_info.max_tx_queues) { - printf("Fail: nb_txq(%d) is greater than " - "max_tx_queues(%d)\n", nb_txq, - port->dev_info.max_tx_queues); + fprintf(stderr, + "Fail: nb_txq(%d) is greater than max_tx_queues(%d)\n", + nb_txq, port->dev_info.max_tx_queues); return -1; } if (numa_support) { @@ -1639,7 +1812,8 @@ init_fwd_streams(void) q = RTE_MAX(nb_rxq, nb_txq); if (q == 0) { - printf("Fail: Cannot allocate fwd streams as number of queues is 0\n"); + fprintf(stderr, + "Fail: Cannot allocate fwd streams as number of queues is 0\n"); return -1; } nb_fwd_streams_new = (streamid_t)(nb_ports * q); @@ -1681,63 +1855,72 @@ init_fwd_streams(void) return 0; } -#ifdef RTE_TEST_PMD_RECORD_BURST_STATS static void pkt_burst_stats_display(const char *rx_tx, struct pkt_burst_stats *pbs) { - unsigned int total_burst; - unsigned int nb_burst; - unsigned int burst_stats[3]; - uint16_t pktnb_stats[3]; + uint64_t total_burst, sburst; + uint64_t nb_burst; + uint64_t burst_stats[4]; + uint16_t pktnb_stats[4]; uint16_t nb_pkt; - int burst_percent[3]; + int burst_percent[4], sburstp; + int i; /* * First compute the total number of packet bursts and the * two highest numbers of bursts of the same number of packets. */ - total_burst = 0; - burst_stats[0] = burst_stats[1] = burst_stats[2] = 0; - pktnb_stats[0] = pktnb_stats[1] = pktnb_stats[2] = 0; - for (nb_pkt = 0; nb_pkt < MAX_PKT_BURST; nb_pkt++) { + memset(&burst_stats, 0x0, sizeof(burst_stats)); + memset(&pktnb_stats, 0x0, sizeof(pktnb_stats)); + + /* Show stats for 0 burst size always */ + total_burst = pbs->pkt_burst_spread[0]; + burst_stats[0] = pbs->pkt_burst_spread[0]; + pktnb_stats[0] = 0; + + /* Find the next 2 burst sizes with highest occurrences. */ + for (nb_pkt = 1; nb_pkt < MAX_PKT_BURST; nb_pkt++) { nb_burst = pbs->pkt_burst_spread[nb_pkt]; + if (nb_burst == 0) continue; + total_burst += nb_burst; - if (nb_burst > burst_stats[0]) { - burst_stats[1] = burst_stats[0]; - pktnb_stats[1] = pktnb_stats[0]; - burst_stats[0] = nb_burst; - pktnb_stats[0] = nb_pkt; - } else if (nb_burst > burst_stats[1]) { + + if (nb_burst > burst_stats[1]) { + burst_stats[2] = burst_stats[1]; + pktnb_stats[2] = pktnb_stats[1]; burst_stats[1] = nb_burst; pktnb_stats[1] = nb_pkt; + } else if (nb_burst > burst_stats[2]) { + burst_stats[2] = nb_burst; + pktnb_stats[2] = nb_pkt; } } if (total_burst == 0) return; - burst_percent[0] = (burst_stats[0] * 100) / total_burst; - printf(" %s-bursts : %u [%d%% of %d pkts", rx_tx, total_burst, - burst_percent[0], (int) pktnb_stats[0]); - if (burst_stats[0] == total_burst) { - printf("]\n"); - return; - } - if (burst_stats[0] + burst_stats[1] == total_burst) { - printf(" + %d%% of %d pkts]\n", - 100 - burst_percent[0], pktnb_stats[1]); - return; - } - burst_percent[1] = (burst_stats[1] * 100) / total_burst; - burst_percent[2] = 100 - (burst_percent[0] + burst_percent[1]); - if ((burst_percent[1] == 0) || (burst_percent[2] == 0)) { - printf(" + %d%% of others]\n", 100 - burst_percent[0]); - return; + + printf(" %s-bursts : %"PRIu64" [", rx_tx, total_burst); + for (i = 0, sburst = 0, sburstp = 0; i < 4; i++) { + if (i == 3) { + printf("%d%% of other]\n", 100 - sburstp); + return; + } + + sburst += burst_stats[i]; + if (sburst == total_burst) { + printf("%d%% of %d pkts]\n", + 100 - sburstp, (int) pktnb_stats[i]); + return; + } + + burst_percent[i] = + (double)burst_stats[i] / total_burst * 100; + printf("%d%% of %d pkts + ", + burst_percent[i], (int) pktnb_stats[i]); + sburstp += burst_percent[i]; } - printf(" + %d%% of %d pkts + %d%% of others]\n", - burst_percent[1], (int) pktnb_stats[1], burst_percent[2]); } -#endif /* RTE_TEST_PMD_RECORD_BURST_STATS */ static void fwd_stream_stats_display(streamid_t stream_id) @@ -1764,14 +1947,16 @@ fwd_stream_stats_display(streamid_t stream_id) " Rx- bad outer L4 checksum: %-14"PRIu64"\n", fs->rx_bad_ip_csum, fs->rx_bad_l4_csum, fs->rx_bad_outer_l4_csum); + printf(" RX- bad outer IP checksum: %-14"PRIu64"\n", + fs->rx_bad_outer_ip_csum); } else { printf("\n"); } -#ifdef RTE_TEST_PMD_RECORD_BURST_STATS - pkt_burst_stats_display("RX", &fs->rx_burst_stats); - pkt_burst_stats_display("TX", &fs->tx_burst_stats); -#endif + if (record_burst_stats) { + pkt_burst_stats_display("RX", &fs->rx_burst_stats); + pkt_burst_stats_display("TX", &fs->tx_burst_stats); + } } void @@ -1786,14 +1971,13 @@ fwd_stats_display(void) uint64_t rx_bad_ip_csum; uint64_t rx_bad_l4_csum; uint64_t rx_bad_outer_l4_csum; + uint64_t rx_bad_outer_ip_csum; } ports_stats[RTE_MAX_ETHPORTS]; uint64_t total_rx_dropped = 0; uint64_t total_tx_dropped = 0; uint64_t total_rx_nombuf = 0; struct rte_eth_stats stats; -#ifdef RTE_TEST_PMD_RECORD_CORE_CYCLES uint64_t fwd_cycles = 0; -#endif uint64_t total_recv = 0; uint64_t total_xmit = 0; struct rte_port *port; @@ -1820,14 +2004,13 @@ fwd_stats_display(void) ports_stats[fs->rx_port].rx_bad_l4_csum += fs->rx_bad_l4_csum; ports_stats[fs->rx_port].rx_bad_outer_l4_csum += fs->rx_bad_outer_l4_csum; + ports_stats[fs->rx_port].rx_bad_outer_ip_csum += + fs->rx_bad_outer_ip_csum; -#ifdef RTE_TEST_PMD_RECORD_CORE_CYCLES - fwd_cycles += fs->core_cycles; -#endif + if (record_core_cycles) + fwd_cycles += fs->core_cycles; } for (i = 0; i < cur_fwd_config.nb_fwd_ports; i++) { - uint8_t j; - pt_id = fwd_ports_ids[i]; port = &ports[pt_id]; @@ -1850,88 +2033,37 @@ fwd_stats_display(void) printf("\n %s Forward statistics for port %-2d %s\n", fwd_stats_border, pt_id, fwd_stats_border); - if (!port->rx_queue_stats_mapping_enabled && - !port->tx_queue_stats_mapping_enabled) { - printf(" RX-packets: %-14"PRIu64 - " RX-dropped: %-14"PRIu64 - "RX-total: %-"PRIu64"\n", - stats.ipackets, stats.imissed, - stats.ipackets + stats.imissed); - - if (cur_fwd_eng == &csum_fwd_engine) - printf(" Bad-ipcsum: %-14"PRIu64 - " Bad-l4csum: %-14"PRIu64 - "Bad-outer-l4csum: %-14"PRIu64"\n", - ports_stats[pt_id].rx_bad_ip_csum, - ports_stats[pt_id].rx_bad_l4_csum, - ports_stats[pt_id].rx_bad_outer_l4_csum); - if (stats.ierrors + stats.rx_nombuf > 0) { - printf(" RX-error: %-"PRIu64"\n", - stats.ierrors); - printf(" RX-nombufs: %-14"PRIu64"\n", - stats.rx_nombuf); - } - - printf(" TX-packets: %-14"PRIu64 - " TX-dropped: %-14"PRIu64 - "TX-total: %-"PRIu64"\n", - stats.opackets, ports_stats[pt_id].tx_dropped, - stats.opackets + ports_stats[pt_id].tx_dropped); - } else { - printf(" RX-packets: %14"PRIu64 - " RX-dropped:%14"PRIu64 - " RX-total:%14"PRIu64"\n", - stats.ipackets, stats.imissed, - stats.ipackets + stats.imissed); - - if (cur_fwd_eng == &csum_fwd_engine) - printf(" Bad-ipcsum:%14"PRIu64 - " Bad-l4csum:%14"PRIu64 - " Bad-outer-l4csum: %-14"PRIu64"\n", - ports_stats[pt_id].rx_bad_ip_csum, - ports_stats[pt_id].rx_bad_l4_csum, - ports_stats[pt_id].rx_bad_outer_l4_csum); - if ((stats.ierrors + stats.rx_nombuf) > 0) { - printf(" RX-error:%"PRIu64"\n", stats.ierrors); - printf(" RX-nombufs: %14"PRIu64"\n", - stats.rx_nombuf); - } - - printf(" TX-packets: %14"PRIu64 - " TX-dropped:%14"PRIu64 - " TX-total:%14"PRIu64"\n", - stats.opackets, ports_stats[pt_id].tx_dropped, - stats.opackets + ports_stats[pt_id].tx_dropped); + printf(" RX-packets: %-14"PRIu64" RX-dropped: %-14"PRIu64 + "RX-total: %-"PRIu64"\n", stats.ipackets, stats.imissed, + stats.ipackets + stats.imissed); + + if (cur_fwd_eng == &csum_fwd_engine) { + printf(" Bad-ipcsum: %-14"PRIu64 + " Bad-l4csum: %-14"PRIu64 + "Bad-outer-l4csum: %-14"PRIu64"\n", + ports_stats[pt_id].rx_bad_ip_csum, + ports_stats[pt_id].rx_bad_l4_csum, + ports_stats[pt_id].rx_bad_outer_l4_csum); + printf(" Bad-outer-ipcsum: %-14"PRIu64"\n", + ports_stats[pt_id].rx_bad_outer_ip_csum); + } + if (stats.ierrors + stats.rx_nombuf > 0) { + printf(" RX-error: %-"PRIu64"\n", stats.ierrors); + printf(" RX-nombufs: %-14"PRIu64"\n", stats.rx_nombuf); } -#ifdef RTE_TEST_PMD_RECORD_BURST_STATS - if (ports_stats[pt_id].rx_stream) - pkt_burst_stats_display("RX", - &ports_stats[pt_id].rx_stream->rx_burst_stats); - if (ports_stats[pt_id].tx_stream) - pkt_burst_stats_display("TX", + printf(" TX-packets: %-14"PRIu64" TX-dropped: %-14"PRIu64 + "TX-total: %-"PRIu64"\n", + stats.opackets, ports_stats[pt_id].tx_dropped, + stats.opackets + ports_stats[pt_id].tx_dropped); + + if (record_burst_stats) { + if (ports_stats[pt_id].rx_stream) + pkt_burst_stats_display("RX", + &ports_stats[pt_id].rx_stream->rx_burst_stats); + if (ports_stats[pt_id].tx_stream) + pkt_burst_stats_display("TX", &ports_stats[pt_id].tx_stream->tx_burst_stats); -#endif - - if (port->rx_queue_stats_mapping_enabled) { - printf("\n"); - for (j = 0; j < RTE_ETHDEV_QUEUE_STAT_CNTRS; j++) { - printf(" Stats reg %2d RX-packets:%14"PRIu64 - " RX-errors:%14"PRIu64 - " RX-bytes:%14"PRIu64"\n", - j, stats.q_ipackets[j], - stats.q_errors[j], stats.q_ibytes[j]); - } - printf("\n"); - } - if (port->tx_queue_stats_mapping_enabled) { - for (j = 0; j < RTE_ETHDEV_QUEUE_STAT_CNTRS; j++) { - printf(" Stats reg %2d TX-packets:%14"PRIu64 - " TX-bytes:%14" - PRIu64"\n", - j, stats.q_opackets[j], - stats.q_obytes[j]); - } } printf(" %s--------------------------------%s\n", @@ -1952,13 +2084,24 @@ fwd_stats_display(void) printf(" %s++++++++++++++++++++++++++++++++++++++++++++++" "%s\n", acc_stats_border, acc_stats_border); -#ifdef RTE_TEST_PMD_RECORD_CORE_CYCLES - if (total_recv > 0) - printf("\n CPU cycles/packet=%u (total cycles=" - "%"PRIu64" / total RX packets=%"PRIu64")\n", - (unsigned int)(fwd_cycles / total_recv), - fwd_cycles, total_recv); -#endif + if (record_core_cycles) { +#define CYC_PER_MHZ 1E6 + if (total_recv > 0 || total_xmit > 0) { + uint64_t total_pkts = 0; + if (strcmp(cur_fwd_eng->fwd_mode_name, "txonly") == 0 || + strcmp(cur_fwd_eng->fwd_mode_name, "flowgen") == 0) + total_pkts = total_xmit; + else + total_pkts = total_recv; + + printf("\n CPU cycles/packet=%.2F (total cycles=" + "%"PRIu64" / total %s packets=%"PRIu64") at %"PRIu64 + " MHz Clock\n", + (double) fwd_cycles / total_pkts, + fwd_cycles, cur_fwd_eng->fwd_mode_name, total_pkts, + (uint64_t)(rte_get_tsc_hz() / CYC_PER_MHZ)); + } + } } void @@ -1981,14 +2124,11 @@ fwd_stats_reset(void) fs->rx_bad_ip_csum = 0; fs->rx_bad_l4_csum = 0; fs->rx_bad_outer_l4_csum = 0; + fs->rx_bad_outer_ip_csum = 0; -#ifdef RTE_TEST_PMD_RECORD_BURST_STATS memset(&fs->rx_burst_stats, 0, sizeof(fs->rx_burst_stats)); memset(&fs->tx_burst_stats, 0, sizeof(fs->tx_burst_stats)); -#endif -#ifdef RTE_TEST_PMD_RECORD_CORE_CYCLES fs->core_cycles = 0; -#endif } } @@ -2005,6 +2145,11 @@ flush_fwd_rx_queues(void) uint64_t prev_tsc = 0, diff_tsc, cur_tsc, timer_tsc = 0; uint64_t timer_period; + if (num_procs > 1) { + printf("multi-process not support for flushing fwd Rx queues, skip the below lines and return.\n"); + return; + } + /* convert to number of cycles */ timer_period = rte_get_timer_hz(); /* 1 second timeout */ @@ -2043,7 +2188,7 @@ run_pkt_fwd_on_lcore(struct fwd_lcore *fc, packet_fwd_t pkt_fwd) struct fwd_stream **fsm; streamid_t nb_fs; streamid_t sm_id; -#ifdef RTE_LIBRTE_BITRATE +#ifdef RTE_LIB_BITRATESTATS uint64_t tics_per_1sec; uint64_t tics_datum; uint64_t tics_current; @@ -2058,7 +2203,7 @@ run_pkt_fwd_on_lcore(struct fwd_lcore *fc, packet_fwd_t pkt_fwd) do { for (sm_id = 0; sm_id < nb_fs; sm_id++) (*pkt_fwd)(fsm[sm_id]); -#ifdef RTE_LIBRTE_BITRATE +#ifdef RTE_LIB_BITRATESTATS if (bitrate_enabled != 0 && bitrate_lcore_id == rte_lcore_id()) { tics_current = rte_rdtsc(); @@ -2071,7 +2216,7 @@ run_pkt_fwd_on_lcore(struct fwd_lcore *fc, packet_fwd_t pkt_fwd) } } #endif -#ifdef RTE_LIBRTE_LATENCY_STATS +#ifdef RTE_LIB_LATENCYSTATS if (latencystats_enabled != 0 && latencystats_lcore_id == rte_lcore_id()) rte_latencystats_update(); @@ -2113,16 +2258,10 @@ run_one_txonly_burst_on_core(void *fwd_arg) static void launch_packet_forwarding(lcore_function_t *pkt_fwd_on_lcore) { - port_fwd_begin_t port_fwd_begin; unsigned int i; unsigned int lc_id; int diag; - port_fwd_begin = cur_fwd_config.fwd_eng->port_fwd_begin; - if (port_fwd_begin != NULL) { - for (i = 0; i < cur_fwd_config.nb_fwd_ports; i++) - (*port_fwd_begin)(fwd_ports_ids[i]); - } for (i = 0; i < cur_fwd_config.nb_fwd_lcores; i++) { lc_id = fwd_lcores_cpuids[i]; if ((interactive == 0) || (lc_id != rte_lcore_id())) { @@ -2130,8 +2269,9 @@ launch_packet_forwarding(lcore_function_t *pkt_fwd_on_lcore) diag = rte_eal_remote_launch(pkt_fwd_on_lcore, fwd_lcores[i], lc_id); if (diag != 0) - printf("launch lcore %u failed - diag=%d\n", - lc_id, diag); + fprintf(stderr, + "launch lcore %u failed - diag=%d\n", + lc_id, diag); } } } @@ -2144,9 +2284,7 @@ start_packet_forwarding(int with_tx_first) { port_fwd_begin_t port_fwd_begin; port_fwd_end_t port_fwd_end; - struct rte_port *port; unsigned int i; - portid_t pt_id; if (strcmp(cur_fwd_eng->fwd_mode_name, "rxonly") == 0 && !nb_rxq) rte_exit(EXIT_FAILURE, "rxq are 0, cannot use rxonly fwd mode\n"); @@ -2162,53 +2300,53 @@ start_packet_forwarding(int with_tx_first) cur_fwd_eng->fwd_mode_name); if (all_ports_started() == 0) { - printf("Not all ports were started\n"); + fprintf(stderr, "Not all ports were started\n"); return; } if (test_done == 0) { - printf("Packet forwarding already started\n"); + fprintf(stderr, "Packet forwarding already started\n"); return; } + fwd_config_setup(); - if(dcb_test) { - for (i = 0; i < nb_fwd_ports; i++) { - pt_id = fwd_ports_ids[i]; - port = &ports[pt_id]; - if (!port->dcb_flag) { - printf("In DCB mode, all forwarding ports must " - "be configured in this mode.\n"); + pkt_fwd_config_display(&cur_fwd_config); + if (!pkt_fwd_shared_rxq_check()) + return; + + port_fwd_begin = cur_fwd_config.fwd_eng->port_fwd_begin; + if (port_fwd_begin != NULL) { + for (i = 0; i < cur_fwd_config.nb_fwd_ports; i++) { + if (port_fwd_begin(fwd_ports_ids[i])) { + fprintf(stderr, + "Packet forwarding is not ready\n"); return; } } - if (nb_fwd_lcores == 1) { - printf("In DCB mode,the nb forwarding cores " - "should be larger than 1.\n"); - return; + } + + if (with_tx_first) { + port_fwd_begin = tx_only_engine.port_fwd_begin; + if (port_fwd_begin != NULL) { + for (i = 0; i < cur_fwd_config.nb_fwd_ports; i++) { + if (port_fwd_begin(fwd_ports_ids[i])) { + fprintf(stderr, + "Packet forwarding is not ready\n"); + return; + } + } } } - test_done = 0; - fwd_config_setup(); + test_done = 0; if(!no_flush_rx) flush_fwd_rx_queues(); - pkt_fwd_config_display(&cur_fwd_config); rxtx_config_display(); fwd_stats_reset(); - for (i = 0; i < cur_fwd_config.nb_fwd_ports; i++) { - pt_id = fwd_ports_ids[i]; - port = &ports[pt_id]; - map_port_queue_stats_mapping_registers(pt_id, port); - } if (with_tx_first) { - port_fwd_begin = tx_only_engine.port_fwd_begin; - if (port_fwd_begin != NULL) { - for (i = 0; i < cur_fwd_config.nb_fwd_ports; i++) - (*port_fwd_begin)(fwd_ports_ids[i]); - } while (with_tx_first--) { launch_packet_forwarding( run_one_txonly_burst_on_core); @@ -2232,7 +2370,7 @@ stop_packet_forwarding(void) int i; if (test_done) { - printf("Packet forwarding not started\n"); + fprintf(stderr, "Packet forwarding not started\n"); return; } printf("Telling cores to stop..."); @@ -2258,14 +2396,14 @@ void dev_set_link_up(portid_t pid) { if (rte_eth_dev_set_link_up(pid) < 0) - printf("\nSet link up fail.\n"); + fprintf(stderr, "\nSet link up fail.\n"); } void dev_set_link_down(portid_t pid) { if (rte_eth_dev_set_link_down(pid) < 0) - printf("\nSet link down fail.\n"); + fprintf(stderr, "\nSet link down fail.\n"); } static int @@ -2324,7 +2462,7 @@ port_is_started(portid_t port_id) /* Configure the Rx and Tx hairpin queues for the selected port. */ static int -setup_hairpin_queues(portid_t pi) +setup_hairpin_queues(portid_t pi, portid_t p_pi, uint16_t cnt_pi) { queueid_t qi; struct rte_eth_hairpin_conf hairpin_conf = { @@ -2333,10 +2471,49 @@ setup_hairpin_queues(portid_t pi) int i; int diag; struct rte_port *port = &ports[pi]; + uint16_t peer_rx_port = pi; + uint16_t peer_tx_port = pi; + uint32_t manual = 1; + uint32_t tx_exp = hairpin_mode & 0x10; + + if (!(hairpin_mode & 0xf)) { + peer_rx_port = pi; + peer_tx_port = pi; + manual = 0; + } else if (hairpin_mode & 0x1) { + peer_tx_port = rte_eth_find_next_owned_by(pi + 1, + RTE_ETH_DEV_NO_OWNER); + if (peer_tx_port >= RTE_MAX_ETHPORTS) + peer_tx_port = rte_eth_find_next_owned_by(0, + RTE_ETH_DEV_NO_OWNER); + if (p_pi != RTE_MAX_ETHPORTS) { + peer_rx_port = p_pi; + } else { + uint16_t next_pi; + + /* Last port will be the peer RX port of the first. */ + RTE_ETH_FOREACH_DEV(next_pi) + peer_rx_port = next_pi; + } + manual = 1; + } else if (hairpin_mode & 0x2) { + if (cnt_pi & 0x1) { + peer_rx_port = p_pi; + } else { + peer_rx_port = rte_eth_find_next_owned_by(pi + 1, + RTE_ETH_DEV_NO_OWNER); + if (peer_rx_port >= RTE_MAX_ETHPORTS) + peer_rx_port = pi; + } + peer_tx_port = peer_rx_port; + manual = 1; + } for (qi = nb_txq, i = 0; qi < nb_hairpinq + nb_txq; qi++) { - hairpin_conf.peers[0].port = pi; + hairpin_conf.peers[0].port = peer_rx_port; hairpin_conf.peers[0].queue = i + nb_rxq; + hairpin_conf.manual_bind = !!manual; + hairpin_conf.tx_explicit = !!tx_exp; diag = rte_eth_tx_hairpin_queue_setup (pi, qi, nb_txd, &hairpin_conf); i++; @@ -2347,17 +2524,19 @@ setup_hairpin_queues(portid_t pi) if (rte_atomic16_cmpset(&(port->port_status), RTE_PORT_HANDLING, RTE_PORT_STOPPED) == 0) - printf("Port %d can not be set back " - "to stopped\n", pi); - printf("Fail to configure port %d hairpin " - "queues\n", pi); + fprintf(stderr, + "Port %d can not be set back to stopped\n", pi); + fprintf(stderr, "Fail to configure port %d hairpin queues\n", + pi); /* try to reconfigure queues next time */ port->need_reconfig_queues = 1; return -1; } for (qi = nb_rxq, i = 0; qi < nb_hairpinq + nb_rxq; qi++) { - hairpin_conf.peers[0].port = pi; + hairpin_conf.peers[0].port = peer_tx_port; hairpin_conf.peers[0].queue = i + nb_txq; + hairpin_conf.manual_bind = !!manual; + hairpin_conf.tx_explicit = !!tx_exp; diag = rte_eth_rx_hairpin_queue_setup (pi, qi, nb_rxd, &hairpin_conf); i++; @@ -2368,10 +2547,10 @@ setup_hairpin_queues(portid_t pi) if (rte_atomic16_cmpset(&(port->port_status), RTE_PORT_HANDLING, RTE_PORT_STOPPED) == 0) - printf("Port %d can not be set back " - "to stopped\n", pi); - printf("Fail to configure port %d hairpin " - "queues\n", pi); + fprintf(stderr, + "Port %d can not be set back to stopped\n", pi); + fprintf(stderr, "Fail to configure port %d hairpin queues\n", + pi); /* try to reconfigure queues next time */ port->need_reconfig_queues = 1; return -1; @@ -2379,21 +2558,171 @@ setup_hairpin_queues(portid_t pi) return 0; } +/* Configure the Rx with optional split. */ +int +rx_queue_setup(uint16_t port_id, uint16_t rx_queue_id, + uint16_t nb_rx_desc, unsigned int socket_id, + struct rte_eth_rxconf *rx_conf, struct rte_mempool *mp) +{ + union rte_eth_rxseg rx_useg[MAX_SEGS_BUFFER_SPLIT] = {}; + unsigned int i, mp_n; + int ret; + + if (rx_pkt_nb_segs <= 1 || + (rx_conf->offloads & RTE_ETH_RX_OFFLOAD_BUFFER_SPLIT) == 0) { + rx_conf->rx_seg = NULL; + rx_conf->rx_nseg = 0; + ret = rte_eth_rx_queue_setup(port_id, rx_queue_id, + nb_rx_desc, socket_id, + rx_conf, mp); + return ret; + } + for (i = 0; i < rx_pkt_nb_segs; i++) { + struct rte_eth_rxseg_split *rx_seg = &rx_useg[i].split; + struct rte_mempool *mpx; + /* + * Use last valid pool for the segments with number + * exceeding the pool index. + */ + mp_n = (i > mbuf_data_size_n) ? mbuf_data_size_n - 1 : i; + mpx = mbuf_pool_find(socket_id, mp_n); + /* Handle zero as mbuf data buffer size. */ + rx_seg->length = rx_pkt_seg_lengths[i] ? + rx_pkt_seg_lengths[i] : + mbuf_data_size[mp_n]; + rx_seg->offset = i < rx_pkt_nb_offs ? + rx_pkt_seg_offsets[i] : 0; + rx_seg->mp = mpx ? mpx : mp; + } + rx_conf->rx_nseg = rx_pkt_nb_segs; + rx_conf->rx_seg = rx_useg; + ret = rte_eth_rx_queue_setup(port_id, rx_queue_id, nb_rx_desc, + socket_id, rx_conf, NULL); + rx_conf->rx_seg = NULL; + rx_conf->rx_nseg = 0; + return ret; +} + +static int +alloc_xstats_display_info(portid_t pi) +{ + uint64_t **ids_supp = &ports[pi].xstats_info.ids_supp; + uint64_t **prev_values = &ports[pi].xstats_info.prev_values; + uint64_t **curr_values = &ports[pi].xstats_info.curr_values; + + if (xstats_display_num == 0) + return 0; + + *ids_supp = calloc(xstats_display_num, sizeof(**ids_supp)); + if (*ids_supp == NULL) + goto fail_ids_supp; + + *prev_values = calloc(xstats_display_num, + sizeof(**prev_values)); + if (*prev_values == NULL) + goto fail_prev_values; + + *curr_values = calloc(xstats_display_num, + sizeof(**curr_values)); + if (*curr_values == NULL) + goto fail_curr_values; + + ports[pi].xstats_info.allocated = true; + + return 0; + +fail_curr_values: + free(*prev_values); +fail_prev_values: + free(*ids_supp); +fail_ids_supp: + return -ENOMEM; +} + +static void +free_xstats_display_info(portid_t pi) +{ + if (!ports[pi].xstats_info.allocated) + return; + free(ports[pi].xstats_info.ids_supp); + free(ports[pi].xstats_info.prev_values); + free(ports[pi].xstats_info.curr_values); + ports[pi].xstats_info.allocated = false; +} + +/** Fill helper structures for specified port to show extended statistics. */ +static void +fill_xstats_display_info_for_port(portid_t pi) +{ + unsigned int stat, stat_supp; + const char *xstat_name; + struct rte_port *port; + uint64_t *ids_supp; + int rc; + + if (xstats_display_num == 0) + return; + + if (pi == (portid_t)RTE_PORT_ALL) { + fill_xstats_display_info(); + return; + } + + port = &ports[pi]; + if (port->port_status != RTE_PORT_STARTED) + return; + + if (!port->xstats_info.allocated && alloc_xstats_display_info(pi) != 0) + rte_exit(EXIT_FAILURE, + "Failed to allocate xstats display memory\n"); + + ids_supp = port->xstats_info.ids_supp; + for (stat = stat_supp = 0; stat < xstats_display_num; stat++) { + xstat_name = xstats_display[stat].name; + rc = rte_eth_xstats_get_id_by_name(pi, xstat_name, + ids_supp + stat_supp); + if (rc != 0) { + fprintf(stderr, "No xstat '%s' on port %u - skip it %u\n", + xstat_name, pi, stat); + continue; + } + stat_supp++; + } + + port->xstats_info.ids_supp_sz = stat_supp; +} + +/** Fill helper structures for all ports to show extended statistics. */ +static void +fill_xstats_display_info(void) +{ + portid_t pi; + + if (xstats_display_num == 0) + return; + + RTE_ETH_FOREACH_DEV(pi) + fill_xstats_display_info_for_port(pi); +} + int start_port(portid_t pid) { int diag, need_check_link_status = -1; portid_t pi; + portid_t p_pi = RTE_MAX_ETHPORTS; + portid_t pl[RTE_MAX_ETHPORTS]; + portid_t peer_pl[RTE_MAX_ETHPORTS]; + uint16_t cnt_pi = 0; + uint16_t cfg_pi = 0; + int peer_pi; queueid_t qi; struct rte_port *port; - struct rte_ether_addr mac_addr; struct rte_eth_hairpin_cap cap; if (port_id_is_invalid(pid, ENABLED_WARN)) return 0; - if(dcb_config) - dcb_test = 1; RTE_ETH_FOREACH_DEV(pi) { if (pid != pi && pid != (portid_t)RTE_PORT_ALL) continue; @@ -2402,18 +2731,22 @@ start_port(portid_t pid) port = &ports[pi]; if (rte_atomic16_cmpset(&(port->port_status), RTE_PORT_STOPPED, RTE_PORT_HANDLING) == 0) { - printf("Port %d is now not stopped\n", pi); + fprintf(stderr, "Port %d is now not stopped\n", pi); continue; } if (port->need_reconfig > 0) { + struct rte_eth_conf dev_conf; + int k; + port->need_reconfig = 0; if (flow_isolate_all) { int ret = port_flow_isolate(pi, 1); if (ret) { - printf("Failed to apply isolated" - " mode on port %d\n", pi); + fprintf(stderr, + "Failed to apply isolated mode on port %d\n", + pi); return -1; } } @@ -2422,26 +2755,60 @@ start_port(portid_t pid) port->socket_id); if (nb_hairpinq > 0 && rte_eth_dev_hairpin_capability_get(pi, &cap)) { - printf("Port %d doesn't support hairpin " - "queues\n", pi); + fprintf(stderr, + "Port %d doesn't support hairpin queues\n", + pi); return -1; } + /* configure port */ - diag = rte_eth_dev_configure(pi, nb_rxq + nb_hairpinq, + diag = eth_dev_configure_mp(pi, nb_rxq + nb_hairpinq, nb_txq + nb_hairpinq, &(port->dev_conf)); if (diag != 0) { if (rte_atomic16_cmpset(&(port->port_status), RTE_PORT_HANDLING, RTE_PORT_STOPPED) == 0) - printf("Port %d can not be set back " - "to stopped\n", pi); - printf("Fail to configure port %d\n", pi); + fprintf(stderr, + "Port %d can not be set back to stopped\n", + pi); + fprintf(stderr, "Fail to configure port %d\n", + pi); /* try to reconfigure port next time */ port->need_reconfig = 1; return -1; } + /* get device configuration*/ + if (0 != + eth_dev_conf_get_print_err(pi, &dev_conf)) { + fprintf(stderr, + "port %d can not get device configuration\n", + pi); + return -1; + } + /* Apply Rx offloads configuration */ + if (dev_conf.rxmode.offloads != + port->dev_conf.rxmode.offloads) { + port->dev_conf.rxmode.offloads |= + dev_conf.rxmode.offloads; + for (k = 0; + k < port->dev_info.max_rx_queues; + k++) + port->rx_conf[k].offloads |= + dev_conf.rxmode.offloads; + } + /* Apply Tx offloads configuration */ + if (dev_conf.txmode.offloads != + port->dev_conf.txmode.offloads) { + port->dev_conf.txmode.offloads |= + dev_conf.txmode.offloads; + for (k = 0; + k < port->dev_info.max_tx_queues; + k++) + port->tx_conf[k].offloads |= + dev_conf.txmode.offloads; + } } - if (port->need_reconfig_queues > 0) { + if (port->need_reconfig_queues > 0 && is_proc_primary()) { port->need_reconfig_queues = 0; /* setup tx queues */ for (qi = 0; qi < nb_txq; qi++) { @@ -2464,10 +2831,12 @@ start_port(portid_t pid) if (rte_atomic16_cmpset(&(port->port_status), RTE_PORT_HANDLING, RTE_PORT_STOPPED) == 0) - printf("Port %d can not be set back " - "to stopped\n", pi); - printf("Fail to configure port %d tx queues\n", - pi); + fprintf(stderr, + "Port %d can not be set back to stopped\n", + pi); + fprintf(stderr, + "Fail to configure port %d tx queues\n", + pi); /* try to reconfigure queues next time */ port->need_reconfig_queues = 1; return -1; @@ -2477,31 +2846,31 @@ start_port(portid_t pid) if ((numa_support) && (rxring_numa[pi] != NUMA_NO_CONFIG)) { struct rte_mempool * mp = - mbuf_pool_find(rxring_numa[pi]); + mbuf_pool_find + (rxring_numa[pi], 0); if (mp == NULL) { - printf("Failed to setup RX queue:" - "No mempool allocation" - " on the socket %d\n", + fprintf(stderr, + "Failed to setup RX queue: No mempool allocation on the socket %d\n", rxring_numa[pi]); return -1; } - diag = rte_eth_rx_queue_setup(pi, qi, + diag = rx_queue_setup(pi, qi, port->nb_rx_desc[qi], rxring_numa[pi], &(port->rx_conf[qi]), mp); } else { struct rte_mempool *mp = - mbuf_pool_find(port->socket_id); + mbuf_pool_find + (port->socket_id, 0); if (mp == NULL) { - printf("Failed to setup RX queue:" - "No mempool allocation" - " on the socket %d\n", + fprintf(stderr, + "Failed to setup RX queue: No mempool allocation on the socket %d\n", port->socket_id); return -1; } - diag = rte_eth_rx_queue_setup(pi, qi, + diag = rx_queue_setup(pi, qi, port->nb_rx_desc[qi], port->socket_id, &(port->rx_conf[qi]), @@ -2514,16 +2883,18 @@ start_port(portid_t pid) if (rte_atomic16_cmpset(&(port->port_status), RTE_PORT_HANDLING, RTE_PORT_STOPPED) == 0) - printf("Port %d can not be set back " - "to stopped\n", pi); - printf("Fail to configure port %d rx queues\n", - pi); + fprintf(stderr, + "Port %d can not be set back to stopped\n", + pi); + fprintf(stderr, + "Fail to configure port %d rx queues\n", + pi); /* try to reconfigure queues next time */ port->need_reconfig_queues = 1; return -1; } /* setup hairpin queues */ - if (setup_hairpin_queues(pi) != 0) + if (setup_hairpin_queues(pi, p_pi, cnt_pi) != 0) return -1; } configure_rxtx_dump_callbacks(verbose_level); @@ -2531,41 +2902,94 @@ start_port(portid_t pid) diag = rte_eth_dev_set_ptypes(pi, RTE_PTYPE_UNKNOWN, NULL, 0); if (diag < 0) - printf( - "Port %d: Failed to disable Ptype parsing\n", - pi); + fprintf(stderr, + "Port %d: Failed to disable Ptype parsing\n", + pi); } + p_pi = pi; + cnt_pi++; + /* start port */ - if (rte_eth_dev_start(pi) < 0) { - printf("Fail to start port %d\n", pi); + diag = eth_dev_start_mp(pi); + if (diag < 0) { + fprintf(stderr, "Fail to start port %d: %s\n", + pi, rte_strerror(-diag)); /* Fail to setup rx queue, return */ if (rte_atomic16_cmpset(&(port->port_status), RTE_PORT_HANDLING, RTE_PORT_STOPPED) == 0) - printf("Port %d can not be set back to " - "stopped\n", pi); + fprintf(stderr, + "Port %d can not be set back to stopped\n", + pi); continue; } if (rte_atomic16_cmpset(&(port->port_status), RTE_PORT_HANDLING, RTE_PORT_STARTED) == 0) - printf("Port %d can not be set into started\n", pi); + fprintf(stderr, "Port %d can not be set into started\n", + pi); - if (eth_macaddr_get_print_err(pi, &mac_addr) == 0) - printf("Port %d: %02X:%02X:%02X:%02X:%02X:%02X\n", pi, - mac_addr.addr_bytes[0], mac_addr.addr_bytes[1], - mac_addr.addr_bytes[2], mac_addr.addr_bytes[3], - mac_addr.addr_bytes[4], mac_addr.addr_bytes[5]); + if (eth_macaddr_get_print_err(pi, &port->eth_addr) == 0) + printf("Port %d: " RTE_ETHER_ADDR_PRT_FMT "\n", pi, + RTE_ETHER_ADDR_BYTES(&port->eth_addr)); /* at least one port started, need checking link status */ need_check_link_status = 1; + + pl[cfg_pi++] = pi; } if (need_check_link_status == 1 && !no_link_check) check_all_ports_link_status(RTE_PORT_ALL); else if (need_check_link_status == 0) - printf("Please stop the ports first\n"); + fprintf(stderr, "Please stop the ports first\n"); + + if (hairpin_mode & 0xf) { + uint16_t i; + int j; + + /* bind all started hairpin ports */ + for (i = 0; i < cfg_pi; i++) { + pi = pl[i]; + /* bind current Tx to all peer Rx */ + peer_pi = rte_eth_hairpin_get_peer_ports(pi, peer_pl, + RTE_MAX_ETHPORTS, 1); + if (peer_pi < 0) + return peer_pi; + for (j = 0; j < peer_pi; j++) { + if (!port_is_started(peer_pl[j])) + continue; + diag = rte_eth_hairpin_bind(pi, peer_pl[j]); + if (diag < 0) { + fprintf(stderr, + "Error during binding hairpin Tx port %u to %u: %s\n", + pi, peer_pl[j], + rte_strerror(-diag)); + return -1; + } + } + /* bind all peer Tx to current Rx */ + peer_pi = rte_eth_hairpin_get_peer_ports(pi, peer_pl, + RTE_MAX_ETHPORTS, 0); + if (peer_pi < 0) + return peer_pi; + for (j = 0; j < peer_pi; j++) { + if (!port_is_started(peer_pl[j])) + continue; + diag = rte_eth_hairpin_bind(peer_pl[j], pi); + if (diag < 0) { + fprintf(stderr, + "Error during binding hairpin Tx port %u to %u: %s\n", + peer_pl[j], pi, + rte_strerror(-diag)); + return -1; + } + } + } + } + + fill_xstats_display_info_for_port(pid); printf("Done\n"); return 0; @@ -2577,11 +3001,8 @@ stop_port(portid_t pid) portid_t pi; struct rte_port *port; int need_check_link_status = 0; - - if (dcb_test) { - dcb_test = 0; - dcb_config = 0; - } + portid_t peer_pl[RTE_MAX_ETHPORTS]; + int peer_pi; if (port_id_is_invalid(pid, ENABLED_WARN)) return; @@ -2593,12 +3014,16 @@ stop_port(portid_t pid) continue; if (port_is_forwarding(pi) != 0 && test_done == 0) { - printf("Please remove port %d from forwarding configuration.\n", pi); + fprintf(stderr, + "Please remove port %d from forwarding configuration.\n", + pi); continue; } if (port_is_bonding_slave(pi)) { - printf("Please remove port %d from bonded device.\n", pi); + fprintf(stderr, + "Please remove port %d from bonded device.\n", + pi); continue; } @@ -2607,11 +3032,33 @@ stop_port(portid_t pid) RTE_PORT_HANDLING) == 0) continue; - rte_eth_dev_stop(pi); + if (hairpin_mode & 0xf) { + int j; + + rte_eth_hairpin_unbind(pi, RTE_MAX_ETHPORTS); + /* unbind all peer Tx from current Rx */ + peer_pi = rte_eth_hairpin_get_peer_ports(pi, peer_pl, + RTE_MAX_ETHPORTS, 0); + if (peer_pi < 0) + continue; + for (j = 0; j < peer_pi; j++) { + if (!port_is_started(peer_pl[j])) + continue; + rte_eth_hairpin_unbind(peer_pl[j], pi); + } + } + + if (port->flow_list) + port_flow_flush(pi); + + if (eth_dev_stop_mp(pi) != 0) + RTE_LOG(ERR, EAL, "rte_eth_dev_stop failed for port %u\n", + pi); if (rte_atomic16_cmpset(&(port->port_status), RTE_PORT_HANDLING, RTE_PORT_STOPPED) == 0) - printf("Port %d can not be set into stopped\n", pi); + fprintf(stderr, "Port %d can not be set into stopped\n", + pi); need_check_link_status = 1; } if (need_check_link_status && !no_link_check) @@ -2658,39 +3105,36 @@ close_port(portid_t pid) continue; if (port_is_forwarding(pi) != 0 && test_done == 0) { - printf("Please remove port %d from forwarding configuration.\n", pi); + fprintf(stderr, + "Please remove port %d from forwarding configuration.\n", + pi); continue; } if (port_is_bonding_slave(pi)) { - printf("Please remove port %d from bonded device.\n", pi); + fprintf(stderr, + "Please remove port %d from bonded device.\n", + pi); continue; } port = &ports[pi]; if (rte_atomic16_cmpset(&(port->port_status), RTE_PORT_CLOSED, RTE_PORT_CLOSED) == 1) { - printf("Port %d is already closed\n", pi); + fprintf(stderr, "Port %d is already closed\n", pi); continue; } - if (rte_atomic16_cmpset(&(port->port_status), - RTE_PORT_STOPPED, RTE_PORT_HANDLING) == 0) { - printf("Port %d is now not stopped\n", pi); - continue; - } - - if (port->flow_list) + if (is_proc_primary()) { port_flow_flush(pi); - rte_eth_dev_close(pi); - - remove_invalid_ports(); + port_flex_item_flush(pi); + rte_eth_dev_close(pi); + } - if (rte_atomic16_cmpset(&(port->port_status), - RTE_PORT_HANDLING, RTE_PORT_CLOSED) == 0) - printf("Port %d cannot be set to closed\n", pi); + free_xstats_display_info(pi); } + remove_invalid_ports(); printf("Done\n"); } @@ -2706,7 +3150,8 @@ reset_port(portid_t pid) if ((pid == (portid_t)RTE_PORT_ALL && !all_ports_stopped()) || (pid != (portid_t)RTE_PORT_ALL && !port_is_stopped(pid))) { - printf("Can not reset port(s), please stop port(s) first.\n"); + fprintf(stderr, + "Can not reset port(s), please stop port(s) first.\n"); return; } @@ -2717,14 +3162,16 @@ reset_port(portid_t pid) continue; if (port_is_forwarding(pi) != 0 && test_done == 0) { - printf("Please remove port %d from forwarding " - "configuration.\n", pi); + fprintf(stderr, + "Please remove port %d from forwarding configuration.\n", + pi); continue; } if (port_is_bonding_slave(pi)) { - printf("Please remove port %d from bonded device.\n", - pi); + fprintf(stderr, + "Please remove port %d from bonded device.\n", + pi); continue; } @@ -2734,7 +3181,8 @@ reset_port(portid_t pid) port->need_reconfig = 1; port->need_reconfig_queues = 1; } else { - printf("Failed to reset port %d. diag=%d\n", pi, diag); + fprintf(stderr, "Failed to reset port %d. diag=%d\n", + pi, diag); } } @@ -2750,7 +3198,7 @@ attach_port(char *identifier) printf("Attaching a new port...\n"); if (identifier == NULL) { - printf("Invalid parameters are specified\n"); + fprintf(stderr, "Invalid parameters are specified\n"); return; } @@ -2791,7 +3239,8 @@ setup_attached_port(portid_t pi) reconfig(pi, socket_id); ret = rte_eth_promiscuous_enable(pi); if (ret != 0) - printf("Error during enabling promiscuous mode for port %u: %s - ignore\n", + fprintf(stderr, + "Error during enabling promiscuous mode for port %u: %s - ignore\n", pi, rte_strerror(-ret)); ports_ids[nb_ports++] = pi; @@ -2810,26 +3259,27 @@ detach_device(struct rte_device *dev) portid_t sibling; if (dev == NULL) { - printf("Device already removed\n"); + fprintf(stderr, "Device already removed\n"); return; } printf("Removing a device...\n"); - if (rte_dev_remove(dev) < 0) { - TESTPMD_LOG(ERR, "Failed to detach device %s\n", dev->name); - return; - } RTE_ETH_FOREACH_DEV_OF(sibling, dev) { - /* reset mapping between old ports and removed device */ - rte_eth_devices[sibling].device = NULL; if (ports[sibling].port_status != RTE_PORT_CLOSED) { - /* sibling ports are forced to be closed */ - ports[sibling].port_status = RTE_PORT_CLOSED; - printf("Port %u is closed\n", sibling); + if (ports[sibling].port_status != RTE_PORT_STOPPED) { + fprintf(stderr, "Port %u not stopped\n", + sibling); + return; + } + port_flow_flush(sibling); } } + if (rte_dev_remove(dev) < 0) { + TESTPMD_LOG(ERR, "Failed to detach device %s\n", dev->name); + return; + } remove_invalid_ports(); printf("Device is detached\n"); @@ -2841,20 +3291,28 @@ detach_device(struct rte_device *dev) void detach_port_device(portid_t port_id) { + int ret; + struct rte_eth_dev_info dev_info; + if (port_id_is_invalid(port_id, ENABLED_WARN)) return; if (ports[port_id].port_status != RTE_PORT_CLOSED) { if (ports[port_id].port_status != RTE_PORT_STOPPED) { - printf("Port not stopped\n"); + fprintf(stderr, "Port not stopped\n"); return; } - printf("Port was not closed\n"); - if (ports[port_id].flow_list) - port_flow_flush(port_id); + fprintf(stderr, "Port was not closed\n"); } - detach_device(rte_eth_devices[port_id].device); + ret = eth_dev_info_get_print_err(port_id, &dev_info); + if (ret != 0) { + TESTPMD_LOG(ERR, + "Failed to get device info for port %d, not detaching\n", + port_id); + return; + } + detach_device(dev_info.device); } void @@ -2868,31 +3326,27 @@ detach_devargs(char *identifier) memset(&da, 0, sizeof(da)); if (rte_devargs_parsef(&da, "%s", identifier)) { - printf("cannot parse identifier\n"); - if (da.args) - free(da.args); + fprintf(stderr, "cannot parse identifier\n"); return; } RTE_ETH_FOREACH_MATCHING_DEV(port_id, identifier, &iterator) { if (ports[port_id].port_status != RTE_PORT_CLOSED) { if (ports[port_id].port_status != RTE_PORT_STOPPED) { - printf("Port %u not stopped\n", port_id); + fprintf(stderr, "Port %u not stopped\n", + port_id); rte_eth_iterator_cleanup(&iterator); + rte_devargs_reset(&da); return; } - - /* sibling ports are forced to be closed */ - if (ports[port_id].flow_list) - port_flow_flush(port_id); - ports[port_id].port_status = RTE_PORT_CLOSED; - printf("Port %u is now closed\n", port_id); + port_flow_flush(port_id); } } if (rte_eal_hotplug_remove(da.bus->name, da.name) != 0) { TESTPMD_LOG(ERR, "Failed to detach device %s(%s)\n", da.name, da.bus->name); + rte_devargs_reset(&da); return; } @@ -2901,25 +3355,28 @@ detach_devargs(char *identifier) printf("Device %s is detached\n", identifier); printf("Now total ports is %d\n", nb_ports); printf("Done\n"); + rte_devargs_reset(&da); } void pmd_test_exit(void) { portid_t pt_id; + unsigned int i; int ret; - int i; if (test_done == 0) stop_packet_forwarding(); - for (i = 0 ; i < RTE_MAX_NUMA_NODES ; i++) { +#ifndef RTE_EXEC_ENV_WINDOWS + for (i = 0 ; i < RTE_DIM(mempools) ; i++) { if (mempools[i]) { if (mp_alloc_type == MP_ALLOC_ANON) rte_mempool_mem_iter(mempools[i], dma_unmap_cb, NULL); } } +#endif if (ports != NULL) { no_link_check = 1; RTE_ETH_FOREACH_DEV(pt_id) { @@ -2957,10 +3414,11 @@ pmd_test_exit(void) return; } } - for (i = 0 ; i < RTE_MAX_NUMA_NODES ; i++) { + for (i = 0 ; i < RTE_DIM(mempools) ; i++) { if (mempools[i]) - rte_mempool_free(mempools[i]); + mempool_free_mp(mempools[i]); } + free(xstats_display); printf("\nBye...\n"); } @@ -2981,6 +3439,7 @@ check_all_ports_link_status(uint32_t port_mask) uint8_t count, all_ports_up, print_flag = 0; struct rte_eth_link link; int ret; + char link_status[RTE_ETH_LINK_MAX_STR_LEN]; printf("Checking link statuses...\n"); fflush(stdout); @@ -2994,20 +3453,16 @@ check_all_ports_link_status(uint32_t port_mask) if (ret < 0) { all_ports_up = 0; if (print_flag == 1) - printf("Port %u link get failed: %s\n", + fprintf(stderr, + "Port %u link get failed: %s\n", portid, rte_strerror(-ret)); continue; } /* print link status if flag set */ if (print_flag == 1) { - if (link.link_status) - printf( - "Port%d Link Up. speed %u Mbps- %s\n", - portid, link.link_speed, - (link.link_duplex == ETH_LINK_FULL_DUPLEX) ? - ("full-duplex") : ("half-duplex\n")); - else - printf("Port %d Link Down\n", portid); + rte_eth_link_to_str(link_status, + sizeof(link_status), &link); + printf("Port %d %s\n", portid, link_status); continue; } /* clear all_ports_up flag if any link down */ @@ -3035,19 +3490,14 @@ check_all_ports_link_status(uint32_t port_mask) } } -/* - * This callback is for remove a port for a device. It has limitation because - * it is not for multiple port removal for a device. - * TODO: the device detach invoke will plan to be removed from user side to - * eal. And convert all PMDs to free port resources on ether device closing. - */ static void rmv_port_callback(void *arg) { int need_to_start = 0; int org_no_link_check = no_link_check; portid_t port_id = (intptr_t)arg; - struct rte_device *dev; + struct rte_eth_dev_info dev_info; + int ret; RTE_ETH_VALID_PORTID_OR_RET(port_id); @@ -3059,11 +3509,16 @@ rmv_port_callback(void *arg) stop_port(port_id); no_link_check = org_no_link_check; - /* Save rte_device pointer before closing ethdev port */ - dev = rte_eth_devices[port_id].device; - close_port(port_id); - detach_device(dev); /* might be already removed or have more ports */ - + ret = eth_dev_info_get_print_err(port_id, &dev_info); + if (ret != 0) + TESTPMD_LOG(ERR, + "Failed to get device info for port %d, not detaching\n", + port_id); + else { + struct rte_device *device = dev_info.device; + close_port(port_id); + detach_device(device); /* might be already removed or have more ports */ + } if (need_to_start) start_packet_forwarding(0); } @@ -3077,7 +3532,8 @@ eth_event_callback(portid_t port_id, enum rte_eth_event_type type, void *param, RTE_SET_USED(ret_param); if (type >= RTE_ETH_EVENT_MAX) { - fprintf(stderr, "\nPort %" PRIu16 ": %s called upon invalid event %d\n", + fprintf(stderr, + "\nPort %" PRIu16 ": %s called upon invalid event %d\n", port_id, __func__, type); fflush(stderr); } else if (event_print_mask & (UINT32_C(1) << type)) { @@ -3096,7 +3552,12 @@ eth_event_callback(portid_t port_id, enum rte_eth_event_type type, void *param, break; if (rte_eal_alarm_set(100000, rmv_port_callback, (void *)(intptr_t)port_id)) - fprintf(stderr, "Could not set up deferred device removal\n"); + fprintf(stderr, + "Could not set up deferred device removal\n"); + break; + case RTE_ETH_EVENT_DESTROY: + ports[port_id].port_status = RTE_PORT_CLOSED; + printf("Port %u is closed\n", port_id); break; default: break; @@ -3176,93 +3637,24 @@ dev_event_callback(const char *device_name, enum rte_dev_event_type type, } } -static int -set_tx_queue_stats_mapping_registers(portid_t port_id, struct rte_port *port) -{ - uint16_t i; - int diag; - uint8_t mapping_found = 0; - - for (i = 0; i < nb_tx_queue_stats_mappings; i++) { - if ((tx_queue_stats_mappings[i].port_id == port_id) && - (tx_queue_stats_mappings[i].queue_id < nb_txq )) { - diag = rte_eth_dev_set_tx_queue_stats_mapping(port_id, - tx_queue_stats_mappings[i].queue_id, - tx_queue_stats_mappings[i].stats_counter_id); - if (diag != 0) - return diag; - mapping_found = 1; - } - } - if (mapping_found) - port->tx_queue_stats_mapping_enabled = 1; - return 0; -} - -static int -set_rx_queue_stats_mapping_registers(portid_t port_id, struct rte_port *port) -{ - uint16_t i; - int diag; - uint8_t mapping_found = 0; - - for (i = 0; i < nb_rx_queue_stats_mappings; i++) { - if ((rx_queue_stats_mappings[i].port_id == port_id) && - (rx_queue_stats_mappings[i].queue_id < nb_rxq )) { - diag = rte_eth_dev_set_rx_queue_stats_mapping(port_id, - rx_queue_stats_mappings[i].queue_id, - rx_queue_stats_mappings[i].stats_counter_id); - if (diag != 0) - return diag; - mapping_found = 1; - } - } - if (mapping_found) - port->rx_queue_stats_mapping_enabled = 1; - return 0; -} - static void -map_port_queue_stats_mapping_registers(portid_t pi, struct rte_port *port) -{ - int diag = 0; - - diag = set_tx_queue_stats_mapping_registers(pi, port); - if (diag != 0) { - if (diag == -ENOTSUP) { - port->tx_queue_stats_mapping_enabled = 0; - printf("TX queue stats mapping not supported port id=%d\n", pi); - } - else - rte_exit(EXIT_FAILURE, - "set_tx_queue_stats_mapping_registers " - "failed for port id=%d diag=%d\n", - pi, diag); - } - - diag = set_rx_queue_stats_mapping_registers(pi, port); - if (diag != 0) { - if (diag == -ENOTSUP) { - port->rx_queue_stats_mapping_enabled = 0; - printf("RX queue stats mapping not supported port id=%d\n", pi); - } - else - rte_exit(EXIT_FAILURE, - "set_rx_queue_stats_mapping_registers " - "failed for port id=%d diag=%d\n", - pi, diag); - } -} - -static void -rxtx_port_config(struct rte_port *port) +rxtx_port_config(portid_t pid) { uint16_t qid; uint64_t offloads; + struct rte_port *port = &ports[pid]; for (qid = 0; qid < nb_rxq; qid++) { offloads = port->rx_conf[qid].offloads; port->rx_conf[qid] = port->dev_info.default_rxconf; + + if (rxq_share > 0 && + (port->dev_info.dev_capa & RTE_ETH_DEV_CAPA_RXQ_SHARE)) { + /* Non-zero share group to enable RxQ share. */ + port->rx_conf[qid].share_group = pid / rxq_share + 1; + port->rx_conf[qid].share_qid = qid; /* Equal mapping. */ + } + if (offloads != 0) port->rx_conf[qid].offloads = offloads; @@ -3311,12 +3703,50 @@ rxtx_port_config(struct rte_port *port) } } +/* + * Helper function to set MTU from frame size + * + * port->dev_info should be set before calling this function. + * + * return 0 on success, negative on error + */ +int +update_mtu_from_frame_size(portid_t portid, uint32_t max_rx_pktlen) +{ + struct rte_port *port = &ports[portid]; + uint32_t eth_overhead; + uint16_t mtu, new_mtu; + + eth_overhead = get_eth_overhead(&port->dev_info); + + if (rte_eth_dev_get_mtu(portid, &mtu) != 0) { + printf("Failed to get MTU for port %u\n", portid); + return -1; + } + + new_mtu = max_rx_pktlen - eth_overhead; + + if (mtu == new_mtu) + return 0; + + if (eth_dev_set_mtu_mp(portid, new_mtu) != 0) { + fprintf(stderr, + "Failed to set MTU to %u for port %u\n", + new_mtu, portid); + return -1; + } + + port->dev_conf.rxmode.mtu = new_mtu; + + return 0; +} + void init_port_config(void) { portid_t pid; struct rte_port *port; - int ret; + int ret, i; RTE_ETH_FOREACH_DEV(pid) { port = &ports[pid]; @@ -3336,30 +3766,36 @@ init_port_config(void) } if (port->dcb_flag == 0) { - if( port->dev_conf.rx_adv_conf.rss_conf.rss_hf != 0) - port->dev_conf.rxmode.mq_mode = ETH_MQ_RX_RSS; - else + if (port->dev_conf.rx_adv_conf.rss_conf.rss_hf != 0) { + port->dev_conf.rxmode.mq_mode = + (enum rte_eth_rx_mq_mode) + (rx_mq_mode & ETH_MQ_RX_RSS); + } else { port->dev_conf.rxmode.mq_mode = ETH_MQ_RX_NONE; + port->dev_conf.rxmode.offloads &= + ~DEV_RX_OFFLOAD_RSS_HASH; + + for (i = 0; + i < port->dev_info.nb_rx_queues; + i++) + port->rx_conf[i].offloads &= + ~DEV_RX_OFFLOAD_RSS_HASH; + } } - rxtx_port_config(port); + rxtx_port_config(pid); ret = eth_macaddr_get_print_err(pid, &port->eth_addr); if (ret != 0) return; - map_port_queue_stats_mapping_registers(pid, port); -#if defined RTE_LIBRTE_IXGBE_PMD && defined RTE_LIBRTE_IXGBE_BYPASS +#if defined RTE_NET_IXGBE && defined RTE_LIBRTE_IXGBE_BYPASS rte_pmd_ixgbe_bypass_init(pid); #endif - if (lsc_interrupt && - (rte_eth_devices[pid].data->dev_flags & - RTE_ETH_DEV_INTR_LSC)) + if (lsc_interrupt && (*port->dev_info.dev_flags & RTE_ETH_DEV_INTR_LSC)) port->dev_conf.intr_conf.lsc = 1; - if (rmv_interrupt && - (rte_eth_devices[pid].data->dev_flags & - RTE_ETH_DEV_INTR_RMV)) + if (rmv_interrupt && (*port->dev_info.dev_flags & RTE_ETH_DEV_INTR_RMV)) port->dev_conf.intr_conf.rmv = 1; } } @@ -3383,10 +3819,19 @@ void clear_port_slave_flag(portid_t slave_pid) uint8_t port_is_bonding_slave(portid_t slave_pid) { struct rte_port *port; + struct rte_eth_dev_info dev_info; + int ret; port = &ports[slave_pid]; - if ((rte_eth_devices[slave_pid].data->dev_flags & - RTE_ETH_DEV_BONDED_SLAVE) || (port->slave_flag == 1)) + ret = eth_dev_info_get_print_err(slave_pid, &dev_info); + if (ret != 0) { + TESTPMD_LOG(ERR, + "Failed to get device info for port id %d," + "cannot determine if the port is a bonded slave", + slave_pid); + return 0; + } + if ((*dev_info.dev_flags & RTE_ETH_DEV_BONDED_SLAVE) || (port->slave_flag == 1)) return 1; return 0; } @@ -3438,7 +3883,9 @@ get_eth_dcb_conf(portid_t pid, struct rte_eth_conf *eth_conf, } /* set DCB mode of RX and TX of multiple queues */ - eth_conf->rxmode.mq_mode = ETH_MQ_RX_VMDQ_DCB; + eth_conf->rxmode.mq_mode = + (enum rte_eth_rx_mq_mode) + (rx_mq_mode & ETH_MQ_RX_VMDQ_DCB); eth_conf->txmode.mq_mode = ETH_MQ_TX_VMDQ_DCB; } else { struct rte_eth_dcb_rx_conf *rx_conf = @@ -3446,6 +3893,8 @@ get_eth_dcb_conf(portid_t pid, struct rte_eth_conf *eth_conf, struct rte_eth_dcb_tx_conf *tx_conf = ð_conf->tx_adv_conf.dcb_tx_conf; + memset(&rss_conf, 0, sizeof(struct rte_eth_rss_conf)); + rc = rte_eth_dev_rss_hash_conf_get(pid, &rss_conf); if (rc != 0) return rc; @@ -3458,7 +3907,9 @@ get_eth_dcb_conf(portid_t pid, struct rte_eth_conf *eth_conf, tx_conf->dcb_tc[i] = i % num_tcs; } - eth_conf->rxmode.mq_mode = ETH_MQ_RX_DCB_RSS; + eth_conf->rxmode.mq_mode = + (enum rte_eth_rx_mq_mode) + (rx_mq_mode & ETH_MQ_RX_DCB_RSS); eth_conf->rx_adv_conf.rss_conf = rss_conf; eth_conf->txmode.mq_mode = ETH_MQ_TX_DCB; } @@ -3483,14 +3934,14 @@ init_port_dcb_config(portid_t pid, int retval; uint16_t i; + if (num_procs > 1) { + printf("The multi-process feature doesn't support dcb.\n"); + return -ENOTSUP; + } rte_port = &ports[pid]; - memset(&port_conf, 0, sizeof(struct rte_eth_conf)); - /* Enter DCB configuration status */ - dcb_config = 1; - - port_conf.rxmode = rte_port->dev_conf.rxmode; - port_conf.txmode = rte_port->dev_conf.txmode; + /* retain the original device configuration. */ + memcpy(&port_conf, &rte_port->dev_conf, sizeof(struct rte_eth_conf)); /*set configuration of DCB in vt mode and DCB in non-vt mode*/ retval = get_eth_dcb_conf(pid, &port_conf, dcb_mode, num_tcs, pfc_en); @@ -3512,8 +3963,9 @@ init_port_dcb_config(portid_t pid, */ if (dcb_mode == DCB_VT_ENABLED && rte_port->dev_info.vmdq_pool_base > 0) { - printf("VMDQ_DCB multi-queue mode is nonsensical" - " for port %d.", pid); + fprintf(stderr, + "VMDQ_DCB multi-queue mode is nonsensical for port %d.\n", + pid); return -1; } @@ -3543,7 +3995,7 @@ init_port_dcb_config(portid_t pid, memcpy(&rte_port->dev_conf, &port_conf, sizeof(struct rte_eth_conf)); - rxtx_port_config(rte_port); + rxtx_port_config(pid); /* VLAN filter */ rte_port->dev_conf.rxmode.offloads |= DEV_RX_OFFLOAD_VLAN_FILTER; for (i = 0; i < RTE_DIM(vlan_tags); i++) @@ -3553,16 +4005,19 @@ init_port_dcb_config(portid_t pid, if (retval != 0) return retval; - map_port_queue_stats_mapping_registers(pid, rte_port); - rte_port->dcb_flag = 1; + /* Enter DCB configuration status */ + dcb_config = 1; + return 0; } static void init_port(void) { + int i; + /* Configuration of Ethernet ports. */ ports = rte_zmalloc("testpmd: ports", sizeof(struct rte_port) * RTE_MAX_ETHPORTS, @@ -3572,7 +4027,10 @@ init_port(void) "rte_zmalloc(%d struct rte_port) failed\n", RTE_MAX_ETHPORTS); } - + for (i = 0; i < RTE_MAX_ETHPORTS; i++) + ports[i].xstats_info.allocated = false; + for (i = 0; i < RTE_MAX_ETHPORTS; i++) + LIST_INIT(&ports[i].flow_tunnel_list); /* Initialize ports NUMA structures */ memset(port_numa, NUMA_NO_CONFIG, RTE_MAX_ETHPORTS); memset(rxring_numa, NUMA_NO_CONFIG, RTE_MAX_ETHPORTS); @@ -3607,13 +4065,13 @@ static void signal_handler(int signum) { if (signum == SIGINT || signum == SIGTERM) { - printf("\nSignal %d received, preparing to exit...\n", - signum); -#ifdef RTE_LIBRTE_PDUMP + fprintf(stderr, "\nSignal %d received, preparing to exit...\n", + signum); +#ifdef RTE_LIB_PDUMP /* uninitialize packet capture framework */ rte_pdump_uninit(); #endif -#ifdef RTE_LIBRTE_LATENCY_STATS +#ifdef RTE_LIB_LATENCYSTATS if (latencystats_enabled != 0) rte_latencystats_uninit(); #endif @@ -3621,8 +4079,10 @@ signal_handler(int signum) /* Set flag to indicate the force termination. */ f_quit = 1; /* exit with the expected status */ +#ifndef RTE_EXEC_ENV_WINDOWS signal(signum, SIG_DFL); kill(getpid(), signum); +#endif } } @@ -3647,15 +4107,11 @@ main(int argc, char** argv) rte_exit(EXIT_FAILURE, "Cannot init EAL: %s\n", rte_strerror(rte_errno)); - if (rte_eal_process_type() == RTE_PROC_SECONDARY) - rte_exit(EXIT_FAILURE, - "Secondary process type not supported.\n"); - ret = register_eth_event_callback(); if (ret != 0) rte_exit(EXIT_FAILURE, "Cannot register for ethdev events"); -#ifdef RTE_LIBRTE_PDUMP +#ifdef RTE_LIB_PDUMP /* initialize packet capture framework */ rte_pdump_init(); #endif @@ -3678,10 +4134,10 @@ main(int argc, char** argv) "Check the core mask argument\n"); /* Bitrate/latency stats disabled by default */ -#ifdef RTE_LIBRTE_BITRATE +#ifdef RTE_LIB_BITRATESTATS bitrate_enabled = 0; #endif -#ifdef RTE_LIBRTE_LATENCY_STATS +#ifdef RTE_LIB_LATENCYSTATS latencystats_enabled = 0; #endif @@ -3697,28 +4153,31 @@ main(int argc, char** argv) if (argc > 1) launch_args_parse(argc, argv); +#ifndef RTE_EXEC_ENV_WINDOWS if (do_mlockall && mlockall(MCL_CURRENT | MCL_FUTURE)) { TESTPMD_LOG(NOTICE, "mlockall() failed with error \"%s\"\n", strerror(errno)); } +#endif if (tx_first && interactive) rte_exit(EXIT_FAILURE, "--tx-first cannot be used on " "interactive mode.\n"); if (tx_first && lsc_interrupt) { - printf("Warning: lsc_interrupt needs to be off when " - " using tx_first. Disabling.\n"); + fprintf(stderr, + "Warning: lsc_interrupt needs to be off when using tx_first. Disabling.\n"); lsc_interrupt = 0; } if (!nb_rxq && !nb_txq) - printf("Warning: Either rx or tx queues should be non-zero\n"); + fprintf(stderr, + "Warning: Either rx or tx queues should be non-zero\n"); if (nb_rxq > 1 && nb_rxq > nb_txq) - printf("Warning: nb_rxq=%d enables RSS configuration, " - "but nb_txq=%d will prevent to fully test it.\n", - nb_rxq, nb_txq); + fprintf(stderr, + "Warning: nb_rxq=%d enables RSS configuration, but nb_txq=%d will prevent to fully test it.\n", + nb_rxq, nb_txq); init_config(); @@ -3753,26 +4212,28 @@ main(int argc, char** argv) RTE_ETH_FOREACH_DEV(port_id) { ret = rte_eth_promiscuous_enable(port_id); if (ret != 0) - printf("Error during enabling promiscuous mode for port %u: %s - ignore\n", + fprintf(stderr, + "Error during enabling promiscuous mode for port %u: %s - ignore\n", port_id, rte_strerror(-ret)); } /* Init metrics library */ rte_metrics_init(rte_socket_id()); -#ifdef RTE_LIBRTE_LATENCY_STATS +#ifdef RTE_LIB_LATENCYSTATS if (latencystats_enabled != 0) { int ret = rte_latencystats_init(1, NULL); if (ret) - printf("Warning: latencystats init()" - " returned error %d\n", ret); - printf("Latencystats running on lcore %d\n", + fprintf(stderr, + "Warning: latencystats init() returned error %d\n", + ret); + fprintf(stderr, "Latencystats running on lcore %d\n", latencystats_lcore_id); } #endif /* Setup bitrate stats */ -#ifdef RTE_LIBRTE_BITRATE +#ifdef RTE_LIB_BITRATESTATS if (bitrate_enabled != 0) { bitrate_data = rte_stats_bitrate_create(); if (bitrate_data == NULL) @@ -3781,8 +4242,7 @@ main(int argc, char** argv) rte_stats_bitrate_reg(bitrate_data); } #endif - -#ifdef RTE_LIBRTE_CMDLINE +#ifdef RTE_LIB_CMDLINE if (strlen(cmdline_filename) != 0) cmdline_read_from_file(cmdline_filename); @@ -3821,7 +4281,7 @@ main(int argc, char** argv) } /* Sleep to avoid unnecessary checks */ prev_time = cur_time; - sleep(1); + rte_delay_us_sleep(US_PER_S); } }