1 /* SPDX-License-Identifier: BSD-3-Clause
2 * Copyright(c) 2010-2014 Intel Corporation
3 * Copyright 2018 Mellanox Technologies, Ltd
10 #include <rte_ether.h>
11 #include <rte_vxlan.h>
12 #include <rte_ethdev.h>
18 print_ether_addr(const char *what, const struct rte_ether_addr *eth_addr)
20 char buf[RTE_ETHER_ADDR_FMT_SIZE];
21 rte_ether_format_addr(buf, RTE_ETHER_ADDR_FMT_SIZE, eth_addr);
22 printf("%s%s", what, buf);
26 dump_pkt_burst(uint16_t port_id, uint16_t queue, struct rte_mbuf *pkts[],
27 uint16_t nb_pkts, int is_rx)
30 const struct rte_ether_hdr *eth_hdr;
31 struct rte_ether_hdr _eth_hdr;
34 uint16_t i, packet_type;
35 uint16_t is_encapsulation;
37 struct rte_net_hdr_lens hdr_lens;
38 uint32_t sw_packet_type;
46 printf("port %u/queue %u: %s %u packets\n",
48 is_rx ? "received" : "sent",
49 (unsigned int) nb_pkts);
50 for (i = 0; i < nb_pkts; i++) {
52 struct rte_flow_error error;
53 struct rte_flow_restore_info info = { 0, };
56 eth_hdr = rte_pktmbuf_read(mb, 0, sizeof(_eth_hdr), &_eth_hdr);
57 eth_type = RTE_BE_TO_CPU_16(eth_hdr->ether_type);
58 packet_type = mb->packet_type;
59 is_encapsulation = RTE_ETH_IS_TUNNEL_PKT(packet_type);
60 ret = rte_flow_get_restore_info(port_id, mb, &info, &error);
62 printf("restore info:");
63 if (info.flags & RTE_FLOW_RESTORE_INFO_TUNNEL) {
64 struct port_flow_tunnel *port_tunnel;
66 port_tunnel = port_flow_locate_tunnel
67 (port_id, &info.tunnel);
70 printf(" #%u", port_tunnel->id);
72 printf(" %s", "-none-");
74 port_flow_tunnel_type(&info.tunnel));
76 printf(" - no tunnel info");
78 if (info.flags & RTE_FLOW_RESTORE_INFO_ENCAPSULATED)
79 printf(" - outer header present");
81 printf(" - no outer header");
82 if (info.flags & RTE_FLOW_RESTORE_INFO_GROUP_ID)
83 printf(" - miss group %u", info.group_id);
85 printf(" - no miss group");
88 print_ether_addr(" src=", ð_hdr->s_addr);
89 print_ether_addr(" - dst=", ð_hdr->d_addr);
90 printf(" - type=0x%04x - length=%u - nb_segs=%d",
91 eth_type, (unsigned int) mb->pkt_len,
93 ol_flags = mb->ol_flags;
94 if (ol_flags & PKT_RX_RSS_HASH) {
95 printf(" - RSS hash=0x%x", (unsigned int) mb->hash.rss);
96 printf(" - RSS queue=0x%x", (unsigned int) queue);
98 if (ol_flags & PKT_RX_FDIR) {
99 printf(" - FDIR matched ");
100 if (ol_flags & PKT_RX_FDIR_ID)
103 else if (ol_flags & PKT_RX_FDIR_FLX)
104 printf("flex bytes=0x%08x %08x",
105 mb->hash.fdir.hi, mb->hash.fdir.lo);
107 printf("hash=0x%x ID=0x%x ",
108 mb->hash.fdir.hash, mb->hash.fdir.id);
110 if (ol_flags & PKT_RX_TIMESTAMP)
111 printf(" - timestamp %"PRIu64" ", mb->timestamp);
112 if (ol_flags & PKT_RX_QINQ)
113 printf(" - QinQ VLAN tci=0x%x, VLAN tci outer=0x%x",
114 mb->vlan_tci, mb->vlan_tci_outer);
115 else if (ol_flags & PKT_RX_VLAN)
116 printf(" - VLAN tci=0x%x", mb->vlan_tci);
117 if (!is_rx && (ol_flags & PKT_TX_DYNF_METADATA))
118 printf(" - Tx metadata: 0x%x",
119 *RTE_FLOW_DYNF_METADATA(mb));
120 if (is_rx && (ol_flags & PKT_RX_DYNF_METADATA))
121 printf(" - Rx metadata: 0x%x",
122 *RTE_FLOW_DYNF_METADATA(mb));
123 for (dynf_index = 0; dynf_index < 64; dynf_index++) {
124 if (dynf_names[dynf_index][0] != '\0')
125 printf(" - dynf %s: %d",
126 dynf_names[dynf_index],
127 !!(ol_flags & (1UL << dynf_index)));
129 if (mb->packet_type) {
130 rte_get_ptype_name(mb->packet_type, buf, sizeof(buf));
131 printf(" - hw ptype: %s", buf);
133 sw_packet_type = rte_net_get_ptype(mb, &hdr_lens,
135 rte_get_ptype_name(sw_packet_type, buf, sizeof(buf));
136 printf(" - sw ptype: %s", buf);
137 if (sw_packet_type & RTE_PTYPE_L2_MASK)
138 printf(" - l2_len=%d", hdr_lens.l2_len);
139 if (sw_packet_type & RTE_PTYPE_L3_MASK)
140 printf(" - l3_len=%d", hdr_lens.l3_len);
141 if (sw_packet_type & RTE_PTYPE_L4_MASK)
142 printf(" - l4_len=%d", hdr_lens.l4_len);
143 if (sw_packet_type & RTE_PTYPE_TUNNEL_MASK)
144 printf(" - tunnel_len=%d", hdr_lens.tunnel_len);
145 if (sw_packet_type & RTE_PTYPE_INNER_L2_MASK)
146 printf(" - inner_l2_len=%d", hdr_lens.inner_l2_len);
147 if (sw_packet_type & RTE_PTYPE_INNER_L3_MASK)
148 printf(" - inner_l3_len=%d", hdr_lens.inner_l3_len);
149 if (sw_packet_type & RTE_PTYPE_INNER_L4_MASK)
150 printf(" - inner_l4_len=%d", hdr_lens.inner_l4_len);
151 if (is_encapsulation) {
152 struct rte_ipv4_hdr *ipv4_hdr;
153 struct rte_ipv6_hdr *ipv6_hdr;
154 struct rte_udp_hdr *udp_hdr;
159 struct rte_vxlan_hdr *vxlan_hdr;
161 l2_len = sizeof(struct rte_ether_hdr);
163 /* Do not support ipv4 option field */
164 if (RTE_ETH_IS_IPV4_HDR(packet_type)) {
165 l3_len = sizeof(struct rte_ipv4_hdr);
166 ipv4_hdr = rte_pktmbuf_mtod_offset(mb,
167 struct rte_ipv4_hdr *,
169 l4_proto = ipv4_hdr->next_proto_id;
171 l3_len = sizeof(struct rte_ipv6_hdr);
172 ipv6_hdr = rte_pktmbuf_mtod_offset(mb,
173 struct rte_ipv6_hdr *,
175 l4_proto = ipv6_hdr->proto;
177 if (l4_proto == IPPROTO_UDP) {
178 udp_hdr = rte_pktmbuf_mtod_offset(mb,
179 struct rte_udp_hdr *,
181 l4_len = sizeof(struct rte_udp_hdr);
182 vxlan_hdr = rte_pktmbuf_mtod_offset(mb,
183 struct rte_vxlan_hdr *,
184 l2_len + l3_len + l4_len);
185 udp_port = RTE_BE_TO_CPU_16(udp_hdr->dst_port);
186 vx_vni = rte_be_to_cpu_32(vxlan_hdr->vx_vni);
187 printf(" - VXLAN packet: packet type =%d, "
188 "Destination UDP port =%d, VNI = %d",
189 packet_type, udp_port, vx_vni >> 8);
192 printf(" - %s queue=0x%x", is_rx ? "Receive" : "Send",
193 (unsigned int) queue);
195 rte_get_rx_ol_flag_list(mb->ol_flags, buf, sizeof(buf));
196 printf(" ol_flags: %s\n", buf);
197 if (rte_mbuf_check(mb, 1, &reason) < 0)
198 printf("INVALID mbuf: %s\n", reason);
203 dump_rx_pkts(uint16_t port_id, uint16_t queue, struct rte_mbuf *pkts[],
204 uint16_t nb_pkts, __rte_unused uint16_t max_pkts,
205 __rte_unused void *user_param)
207 dump_pkt_burst(port_id, queue, pkts, nb_pkts, 1);
212 dump_tx_pkts(uint16_t port_id, uint16_t queue, struct rte_mbuf *pkts[],
213 uint16_t nb_pkts, __rte_unused void *user_param)
215 dump_pkt_burst(port_id, queue, pkts, nb_pkts, 0);
220 tx_pkt_set_md(uint16_t port_id, __rte_unused uint16_t queue,
221 struct rte_mbuf *pkts[], uint16_t nb_pkts,
222 __rte_unused void *user_param)
227 * Add metadata value to every Tx packet,
228 * and set ol_flags accordingly.
230 if (rte_flow_dynf_metadata_avail())
231 for (i = 0; i < nb_pkts; i++) {
232 *RTE_FLOW_DYNF_METADATA(pkts[i]) =
233 ports[port_id].tx_metadata;
234 pkts[i]->ol_flags |= PKT_TX_DYNF_METADATA;
240 add_tx_md_callback(portid_t portid)
242 struct rte_eth_dev_info dev_info;
246 if (port_id_is_invalid(portid, ENABLED_WARN))
249 ret = eth_dev_info_get_print_err(portid, &dev_info);
253 for (queue = 0; queue < dev_info.nb_tx_queues; queue++)
254 if (!ports[portid].tx_set_md_cb[queue])
255 ports[portid].tx_set_md_cb[queue] =
256 rte_eth_add_tx_callback(portid, queue,
257 tx_pkt_set_md, NULL);
261 remove_tx_md_callback(portid_t portid)
263 struct rte_eth_dev_info dev_info;
267 if (port_id_is_invalid(portid, ENABLED_WARN))
270 ret = eth_dev_info_get_print_err(portid, &dev_info);
274 for (queue = 0; queue < dev_info.nb_tx_queues; queue++)
275 if (ports[portid].tx_set_md_cb[queue]) {
276 rte_eth_remove_tx_callback(portid, queue,
277 ports[portid].tx_set_md_cb[queue]);
278 ports[portid].tx_set_md_cb[queue] = NULL;
283 tx_pkt_set_dynf(uint16_t port_id, __rte_unused uint16_t queue,
284 struct rte_mbuf *pkts[], uint16_t nb_pkts,
285 __rte_unused void *user_param)
289 if (ports[port_id].mbuf_dynf)
290 for (i = 0; i < nb_pkts; i++)
291 pkts[i]->ol_flags |= ports[port_id].mbuf_dynf;
296 add_tx_dynf_callback(portid_t portid)
298 struct rte_eth_dev_info dev_info;
302 if (port_id_is_invalid(portid, ENABLED_WARN))
305 ret = eth_dev_info_get_print_err(portid, &dev_info);
309 for (queue = 0; queue < dev_info.nb_tx_queues; queue++)
310 if (!ports[portid].tx_set_dynf_cb[queue])
311 ports[portid].tx_set_dynf_cb[queue] =
312 rte_eth_add_tx_callback(portid, queue,
313 tx_pkt_set_dynf, NULL);
317 remove_tx_dynf_callback(portid_t portid)
319 struct rte_eth_dev_info dev_info;
323 if (port_id_is_invalid(portid, ENABLED_WARN))
326 ret = eth_dev_info_get_print_err(portid, &dev_info);
330 for (queue = 0; queue < dev_info.nb_tx_queues; queue++)
331 if (ports[portid].tx_set_dynf_cb[queue]) {
332 rte_eth_remove_tx_callback(portid, queue,
333 ports[portid].tx_set_dynf_cb[queue]);
334 ports[portid].tx_set_dynf_cb[queue] = NULL;
339 eth_dev_info_get_print_err(uint16_t port_id,
340 struct rte_eth_dev_info *dev_info)
344 ret = rte_eth_dev_info_get(port_id, dev_info);
346 printf("Error during getting device (port %u) info: %s\n",
347 port_id, strerror(-ret));
353 eth_set_promisc_mode(uint16_t port, int enable)
358 ret = rte_eth_promiscuous_enable(port);
360 ret = rte_eth_promiscuous_disable(port);
363 printf("Error during %s promiscuous mode for port %u: %s\n",
364 enable ? "enabling" : "disabling",
365 port, rte_strerror(-ret));
369 eth_set_allmulticast_mode(uint16_t port, int enable)
374 ret = rte_eth_allmulticast_enable(port);
376 ret = rte_eth_allmulticast_disable(port);
379 printf("Error during %s all-multicast mode for port %u: %s\n",
380 enable ? "enabling" : "disabling",
381 port, rte_strerror(-ret));
385 eth_link_get_nowait_print_err(uint16_t port_id, struct rte_eth_link *link)
389 ret = rte_eth_link_get_nowait(port_id, link);
391 printf("Device (port %u) link get (without wait) failed: %s\n",
392 port_id, rte_strerror(-ret));
398 eth_macaddr_get_print_err(uint16_t port_id, struct rte_ether_addr *mac_addr)
402 ret = rte_eth_macaddr_get(port_id, mac_addr);
404 printf("Error getting device (port %u) mac address: %s\n",
405 port_id, rte_strerror(-ret));