X-Git-Url: http://git.droids-corp.org/?a=blobdiff_plain;f=app%2Ftest-pmd%2Ftestpmd.h;h=e016931376af22b68b97b0043a6565d4c1888a4c;hb=2be70bcdb84002f479fae5a49f22d514e31e545d;hp=857a11f8de74ba0d83835f03e304943cd6367510;hpb=e1d44d0ad623825b4719e56ace937c368e7944c0;p=dpdk.git diff --git a/app/test-pmd/testpmd.h b/app/test-pmd/testpmd.h index 857a11f8de..e016931376 100644 --- a/app/test-pmd/testpmd.h +++ b/app/test-pmd/testpmd.h @@ -12,6 +12,7 @@ #include #include #include +#include #define RTE_PORT_ALL (~(portid_t)0x0) @@ -42,6 +43,16 @@ */ #define RTE_MAX_SEGS_PER_PKT 255 /**< nb_segs is a 8-bit unsigned char. */ +/* + * The maximum number of segments per packet is used to configure + * buffer split feature, also specifies the maximum amount of + * optional Rx pools to allocate mbufs to split. + */ +#define MAX_SEGS_BUFFER_SPLIT 8 /**< nb_segs is a 8-bit unsigned char. */ + +/* The prefix of the mbuf pool names created by the application. */ +#define MBUF_POOL_NAME_PFX "mb_pool" + #define MAX_PKT_BURST 512 #define DEF_PKT_BURST 32 @@ -58,12 +69,6 @@ typedef uint16_t portid_t; typedef uint16_t queueid_t; typedef uint16_t streamid_t; -#if defined RTE_LIBRTE_PMD_SOFTNIC -#define SOFTNIC 1 -#else -#define SOFTNIC 0 -#endif - enum { PORT_TOPOLOGY_PAIRED, PORT_TOPOLOGY_CHAINED, @@ -76,11 +81,12 @@ enum { /**< allocate mempool natively, but populate using anonymous memory */ MP_ALLOC_XMEM, /**< allocate and populate mempool using anonymous memory */ - MP_ALLOC_XMEM_HUGE + MP_ALLOC_XMEM_HUGE, /**< allocate and populate mempool using anonymous hugepage memory */ + MP_ALLOC_XBUF + /**< allocate mempool natively, use rte_pktmbuf_pool_create_extbuf */ }; -#ifdef RTE_TEST_PMD_RECORD_BURST_STATS /** * The data structure associated with RX and TX packet burst statistics * that are recorded for each forwarding stream. @@ -88,7 +94,6 @@ enum { struct pkt_burst_stats { unsigned int pkt_burst_spread[MAX_PKT_BURST]; }; -#endif /** Information for a given RSS type. */ struct rss_type_info { @@ -103,6 +108,13 @@ struct rss_type_info { */ extern const struct rss_type_info rss_type_table[]; +/** + * Dynf name array. + * + * Array that holds the name for each dynf. + */ +extern char dynf_names[64][RTE_MBUF_DYN_NAMESIZE]; + /** * The data structure associated with a forwarding stream between a receive * port/queue and a transmit port/queue. @@ -126,13 +138,9 @@ struct fwd_stream { uint64_t rx_bad_outer_l4_csum; /**< received packets has bad outer l4 checksum */ unsigned int gro_times; /**< GRO operation times */ -#ifdef RTE_TEST_PMD_RECORD_CORE_CYCLES uint64_t core_cycles; /**< used for RX and TX processing */ -#endif -#ifdef RTE_TEST_PMD_RECORD_BURST_STATS struct pkt_burst_stats rx_burst_stats; struct pkt_burst_stats tx_burst_stats; -#endif }; /** Descriptor for a single flow. */ @@ -145,15 +153,33 @@ struct port_flow { uint8_t data[]; /**< Storage for flow rule description */ }; -#ifdef SOFTNIC -/** - * The data structure associate with softnic port - */ -struct softnic_port { - uint32_t default_tm_hierarchy_enable; /**< default tm hierarchy */ - struct fwd_lcore **fwd_lcore_arg; /**< softnic fwd core parameters */ +/* Descriptor for shared action */ +struct port_shared_action { + struct port_shared_action *next; /**< Next flow in list. */ + uint32_t id; /**< Shared action ID. */ + enum rte_flow_action_type type; /**< Action type. */ + struct rte_flow_shared_action *action; /**< Shared action handle. */ +}; + +struct port_flow_tunnel { + LIST_ENTRY(port_flow_tunnel) chain; + struct rte_flow_action *pmd_actions; + struct rte_flow_item *pmd_items; + uint32_t id; + uint32_t num_pmd_actions; + uint32_t num_pmd_items; + struct rte_flow_tunnel tunnel; + struct rte_flow_action *actions; + struct rte_flow_item *items; +}; + +struct tunnel_ops { + uint32_t id; + char type[16]; + uint32_t enabled:1; + uint32_t actions:1; + uint32_t items:1; }; -#endif /** * The data structure associated with each port. @@ -185,14 +211,17 @@ struct rte_port { uint32_t mc_addr_nb; /**< nb. of addr. in mc_addr_pool */ uint8_t slave_flag; /**< bonding slave port */ struct port_flow *flow_list; /**< Associated flows. */ + struct port_shared_action *actions_list; + /**< Associated shared actions. */ + LIST_HEAD(, port_flow_tunnel) flow_tunnel_list; const struct rte_eth_rxtx_callback *rx_dump_cb[RTE_MAX_QUEUES_PER_PORT+1]; const struct rte_eth_rxtx_callback *tx_dump_cb[RTE_MAX_QUEUES_PER_PORT+1]; -#ifdef SOFTNIC - struct softnic_port softport; /**< softnic params */ -#endif /**< metadata value to insert in Tx packets. */ uint32_t tx_metadata; const struct rte_eth_rxtx_callback *tx_set_md_cb[RTE_MAX_QUEUES_PER_PORT+1]; + /**< dynamic flags. */ + uint64_t mbuf_dynf; + const struct rte_eth_rxtx_callback *tx_set_dynf_cb[RTE_MAX_QUEUES_PER_PORT+1]; }; /** @@ -254,9 +283,7 @@ extern struct fwd_engine tx_only_engine; extern struct fwd_engine csum_fwd_engine; extern struct fwd_engine icmp_echo_engine; extern struct fwd_engine noisy_vnf_engine; -#ifdef SOFTNIC -extern struct fwd_engine softnic_fwd_engine; -#endif +extern struct fwd_engine five_tuple_swap_fwd_engine; #ifdef RTE_LIBRTE_IEEE1588 extern struct fwd_engine ieee1588_fwd_engine; #endif @@ -310,6 +337,8 @@ extern uint16_t nb_rx_queue_stats_mappings; extern uint8_t xstats_hide_zero; /**< Hide zero values for xstats display */ /* globals used for configuration */ +extern uint8_t record_core_cycles; /**< Enables measurement of CPU cycles */ +extern uint8_t record_burst_stats; /**< Enables display of RX and TX bursts */ extern uint16_t verbose_level; /**< Drives messages being displayed, if any. */ extern int testpmd_logtype; /**< Log type for testpmd logs */ extern uint8_t interactive; @@ -406,23 +435,36 @@ extern uint64_t noisy_lkup_num_reads_writes; extern uint8_t dcb_config; extern uint8_t dcb_test; -extern uint16_t mbuf_data_size; /**< Mbuf data space size. */ +extern uint32_t mbuf_data_size_n; +extern uint16_t mbuf_data_size[MAX_SEGS_BUFFER_SPLIT]; +/**< Mbuf data space size. */ extern uint32_t param_total_num_mbufs; extern uint16_t stats_period; -#ifdef RTE_LIBRTE_LATENCY_STATS +extern uint16_t hairpin_mode; + +#ifdef RTE_LIB_LATENCYSTATS extern uint8_t latencystats_enabled; extern lcoreid_t latencystats_lcore_id; #endif -#ifdef RTE_LIBRTE_BITRATE +#ifdef RTE_LIB_BITRATESTATS extern lcoreid_t bitrate_lcore_id; extern uint8_t bitrate_enabled; #endif extern struct rte_fdir_conf fdir_conf; +/* + * Configuration of packet segments used to scatter received packets + * if some of split features is configured. + */ +extern uint16_t rx_pkt_seg_lengths[MAX_SEGS_BUFFER_SPLIT]; +extern uint8_t rx_pkt_nb_segs; /**< Number of segments to split */ +extern uint16_t rx_pkt_seg_offsets[MAX_SEGS_BUFFER_SPLIT]; +extern uint8_t rx_pkt_nb_offs; /**< Number of specified offsets */ + /* * Configuration of packet segments used by the "txonly" processing engine. */ @@ -430,6 +472,8 @@ extern struct rte_fdir_conf fdir_conf; extern uint16_t tx_pkt_length; /**< Length of TXONLY packet */ extern uint16_t tx_pkt_seg_lengths[RTE_MAX_SEGS_PER_PKT]; /**< Seg. lengths */ extern uint8_t tx_pkt_nb_segs; /**< Number of segments in TX packets */ +extern uint32_t tx_pkt_times_intra; +extern uint32_t tx_pkt_times_inter; enum tx_pkt_split { TX_PKT_SPLIT_OFF, @@ -463,6 +507,7 @@ extern struct fwd_lcore **fwd_lcores; extern struct fwd_stream **fwd_streams; extern uint16_t vxlan_gpe_udp_port; /**< UDP port of tunnel VXLAN-GPE. */ +extern uint16_t geneve_udp_port; /**< UDP port of tunnel GENEVE. */ extern portid_t nb_peer_eth_addrs; /**< Number of peer ethernet addresses. */ extern struct rte_ether_addr peer_eth_addrs[RTE_MAX_ETHPORTS]; @@ -590,6 +635,8 @@ struct mplsoudp_decap_conf { }; extern struct mplsoudp_decap_conf mplsoudp_decap_conf; +extern enum rte_eth_rx_mq_mode rx_mq_mode; + static inline unsigned int lcore_num(void) { @@ -602,6 +649,9 @@ lcore_num(void) rte_panic("lcore_id of current thread not found in fwd_lcores_cpuids\n"); } +void +parse_fwd_portlist(const char *port); + static inline struct fwd_lcore * current_fwd_lcore(void) { @@ -610,17 +660,23 @@ current_fwd_lcore(void) /* Mbuf Pools */ static inline void -mbuf_poolname_build(unsigned int sock_id, char* mp_name, int name_size) +mbuf_poolname_build(unsigned int sock_id, char *mp_name, + int name_size, uint16_t idx) { - snprintf(mp_name, name_size, "mbuf_pool_socket_%u", sock_id); + if (!idx) + snprintf(mp_name, name_size, + MBUF_POOL_NAME_PFX "_%u", sock_id); + else + snprintf(mp_name, name_size, + MBUF_POOL_NAME_PFX "_%hu_%hu", (uint16_t)sock_id, idx); } static inline struct rte_mempool * -mbuf_pool_find(unsigned int sock_id) +mbuf_pool_find(unsigned int sock_id, uint16_t idx) { char pool_name[RTE_MEMPOOL_NAMESIZE]; - mbuf_poolname_build(sock_id, pool_name, sizeof(pool_name)); + mbuf_poolname_build(sock_id, pool_name, sizeof(pool_name), idx); return rte_mempool_lookup((const char *)pool_name); } @@ -683,6 +739,34 @@ port_pci_reg_write(struct rte_port *port, uint32_t reg_off, uint32_t reg_v) #define port_id_pci_reg_write(pt_id, reg_off, reg_value) \ port_pci_reg_write(&ports[(pt_id)], (reg_off), (reg_value)) +static inline void +get_start_cycles(uint64_t *start_tsc) +{ + if (record_core_cycles) + *start_tsc = rte_rdtsc(); +} + +static inline void +get_end_cycles(struct fwd_stream *fs, uint64_t start_tsc) +{ + if (record_core_cycles) + fs->core_cycles += rte_rdtsc() - start_tsc; +} + +static inline void +inc_rx_burst_stats(struct fwd_stream *fs, uint16_t nb_rx) +{ + if (record_burst_stats) + fs->rx_burst_stats.pkt_burst_spread[nb_rx]++; +} + +static inline void +inc_tx_burst_stats(struct fwd_stream *fs, uint16_t nb_tx) +{ + if (record_burst_stats) + fs->tx_burst_stats.pkt_burst_spread[nb_tx]++; +} + /* Prototypes */ unsigned int parse_item_list(char* str, const char* item_name, unsigned int max_items, @@ -699,6 +783,8 @@ void nic_stats_mapping_display(portid_t port_id); void device_infos_display(const char *identifier); void port_infos_display(portid_t port_id); void port_summary_display(portid_t port_id); +void port_eeprom_display(portid_t port_id); +void port_module_eeprom_display(portid_t port_id); void port_summary_header_display(void); void port_offload_cap_display(portid_t port_id); void rx_queue_infos_display(portid_t port_idi, uint16_t queue_id); @@ -724,19 +810,41 @@ void port_reg_bit_field_set(portid_t port_id, uint32_t reg_off, uint8_t bit1_pos, uint8_t bit2_pos, uint32_t value); void port_reg_display(portid_t port_id, uint32_t reg_off); void port_reg_set(portid_t port_id, uint32_t reg_off, uint32_t value); +int port_shared_action_create(portid_t port_id, uint32_t id, + const struct rte_flow_shared_action_conf *conf, + const struct rte_flow_action *action); +int port_shared_action_destroy(portid_t port_id, + uint32_t n, const uint32_t *action); +struct rte_flow_shared_action *port_shared_action_get_by_id(portid_t port_id, + uint32_t id); +int port_shared_action_update(portid_t port_id, uint32_t id, + const struct rte_flow_action *action); int port_flow_validate(portid_t port_id, const struct rte_flow_attr *attr, const struct rte_flow_item *pattern, - const struct rte_flow_action *actions); + const struct rte_flow_action *actions, + const struct tunnel_ops *tunnel_ops); int port_flow_create(portid_t port_id, const struct rte_flow_attr *attr, const struct rte_flow_item *pattern, - const struct rte_flow_action *actions); + const struct rte_flow_action *actions, + const struct tunnel_ops *tunnel_ops); +int port_shared_action_query(portid_t port_id, uint32_t id); +void update_age_action_context(const struct rte_flow_action *actions, + struct port_flow *pf); int port_flow_destroy(portid_t port_id, uint32_t n, const uint32_t *rule); int port_flow_flush(portid_t port_id); +int port_flow_dump(portid_t port_id, const char *file_name); int port_flow_query(portid_t port_id, uint32_t rule, const struct rte_flow_action *action); void port_flow_list(portid_t port_id, uint32_t n, const uint32_t *group); +void port_flow_aged(portid_t port_id, uint8_t destroy); +const char *port_flow_tunnel_type(struct rte_flow_tunnel *tunnel); +struct port_flow_tunnel * +port_flow_locate_tunnel(uint16_t port_id, struct rte_flow_tunnel *tun); +void port_flow_tunnel_list(portid_t port_id); +void port_flow_tunnel_destroy(portid_t port_id, uint32_t tunnel_id); +void port_flow_tunnel_create(portid_t port_id, const struct tunnel_ops *ops); int port_flow_isolate(portid_t port_id, int set); void rx_ring_desc_display(portid_t port_id, queueid_t rxq_id, uint16_t rxd_id); @@ -770,10 +878,20 @@ void set_qmap(portid_t port_id, uint8_t is_rx, uint16_t queue_id, uint8_t map_va void set_xstats_hide_zero(uint8_t on_off); +void set_record_core_cycles(uint8_t on_off); +void set_record_burst_stats(uint8_t on_off); void set_verbose_level(uint16_t vb_level); -void set_tx_pkt_segments(unsigned *seg_lengths, unsigned nb_segs); +void set_rx_pkt_segments(unsigned int *seg_lengths, unsigned int nb_segs); +void show_rx_pkt_segments(void); +void set_rx_pkt_offsets(unsigned int *seg_offsets, unsigned int nb_offs); +void show_rx_pkt_offsets(void); +void set_tx_pkt_segments(unsigned int *seg_lengths, unsigned int nb_segs); void show_tx_pkt_segments(void); +void set_tx_pkt_times(unsigned int *tx_times); +void show_tx_pkt_times(void); void set_tx_pkt_split(const char *name); +int parse_fec_mode(const char *name, enum rte_eth_fec_mode *mode); +void show_fec_capability(uint32_t num, struct rte_eth_fec_capa *speed_fec_capa); void set_nb_pkt_per_burst(uint16_t pkt_burst); char *list_pkt_forwarding_modes(void); char *list_pkt_forwarding_retry_modes(void); @@ -797,7 +915,7 @@ void stop_port(portid_t pid); void close_port(portid_t pid); void reset_port(portid_t pid); void attach_port(char *identifier); -void detach_device(char *identifier); +void detach_devargs(char *identifier); void detach_port_device(portid_t port_id); int all_ports_stopped(void); int port_is_stopped(portid_t port_id); @@ -814,6 +932,11 @@ void port_rss_reta_info(portid_t port_id, void set_vf_traffic(portid_t port_id, uint8_t is_rx, uint16_t vf, uint8_t on); +int +rx_queue_setup(uint16_t port_id, uint16_t rx_queue_id, + uint16_t nb_rx_desc, unsigned int socket_id, + struct rte_eth_rxconf *rx_conf, struct rte_mempool *mp); + int set_queue_rate_limit(portid_t port_id, uint16_t queue_idx, uint16_t rate); int set_vf_rate_limit(portid_t port_id, uint16_t vf, uint16_t rate, uint64_t q_msk); @@ -862,6 +985,8 @@ queueid_t get_allowed_max_nb_rxq(portid_t *pid); int check_nb_rxq(queueid_t rxq); queueid_t get_allowed_max_nb_txq(portid_t *pid); int check_nb_txq(queueid_t txq); +int check_nb_rxd(queueid_t rxd); +int check_nb_txd(queueid_t txd); queueid_t get_allowed_max_nb_hairpinq(portid_t *pid); int check_nb_hairpinq(queueid_t hairpinq); @@ -884,6 +1009,12 @@ uint16_t tx_pkt_set_md(uint16_t port_id, __rte_unused uint16_t queue, void add_tx_md_callback(portid_t portid); void remove_tx_md_callback(portid_t portid); +uint16_t tx_pkt_set_dynf(uint16_t port_id, __rte_unused uint16_t queue, + struct rte_mbuf *pkts[], uint16_t nb_pkts, + __rte_unused void *user_param); +void add_tx_dynf_callback(portid_t portid); +void remove_tx_dynf_callback(portid_t portid); + /* * Work-around of a compilation error with ICC on invocations of the * rte_be_to_cpu_16() function.