#define _htons(x) (x)
#endif
-uint16_t vxlan_gpe_udp_port = 4790;
+uint16_t vxlan_gpe_udp_port = RTE_VXLAN_GPE_DEFAULT_PORT;
uint16_t geneve_udp_port = RTE_GENEVE_DEFAULT_PORT;
/* structure that caches offload info for the current packet */
{
struct rte_tcp_hdr *tcp_hdr;
- info->l3_len = (ipv4_hdr->version_ihl & 0x0f) * 4;
+ info->l3_len = rte_ipv4_hdr_len(ipv4_hdr);
info->l4_proto = ipv4_hdr->next_proto_id;
/* only fill l4_len for TCP, it's useful for TSO */
}
}
+/* Fill in outer layers length */
+static void
+update_tunnel_outer(struct testpmd_offload_info *info)
+{
+ info->is_tunnel = 1;
+ info->outer_ethertype = info->ethertype;
+ info->outer_l2_len = info->l2_len;
+ info->outer_l3_len = info->l3_len;
+ info->outer_l4_proto = info->l4_proto;
+}
+
/*
* Parse a GTP protocol header.
* No optional fields and next extension header type.
udp_hdr->dst_port != _htons(RTE_GTPU_UDP_PORT))
return;
- info->is_tunnel = 1;
- info->outer_ethertype = info->ethertype;
- info->outer_l2_len = info->l2_len;
- info->outer_l3_len = info->l3_len;
- info->outer_l4_proto = info->l4_proto;
+ update_tunnel_outer(info);
info->l2_len = 0;
gtp_hdr = (struct rte_gtp_hdr *)((char *)udp_hdr +
{
struct rte_ether_hdr *eth_hdr;
- /* check udp destination port, 4789 is the default vxlan port
- * (rfc7348) or that the rx offload flag is set (i40e only
- * currently) */
- if (udp_hdr->dst_port != _htons(4789) &&
+ /* check udp destination port, RTE_VXLAN_DEFAULT_PORT (4789) is the
+ * default vxlan port (rfc7348) or that the rx offload flag is set
+ * (i40e only currently)
+ */
+ if (udp_hdr->dst_port != _htons(RTE_VXLAN_DEFAULT_PORT) &&
RTE_ETH_IS_TUNNEL_PKT(pkt_type) == 0)
return;
- info->is_tunnel = 1;
- info->outer_ethertype = info->ethertype;
- info->outer_l2_len = info->l2_len;
- info->outer_l3_len = info->l3_len;
- info->outer_l4_proto = info->l4_proto;
+ update_tunnel_outer(info);
eth_hdr = (struct rte_ether_hdr *)((char *)udp_hdr +
sizeof(struct rte_udp_hdr) +
if (!vxlan_gpe_hdr->proto || vxlan_gpe_hdr->proto ==
RTE_VXLAN_GPE_TYPE_IPV4) {
- info->is_tunnel = 1;
- info->outer_ethertype = info->ethertype;
- info->outer_l2_len = info->l2_len;
- info->outer_l3_len = info->l3_len;
- info->outer_l4_proto = info->l4_proto;
+ update_tunnel_outer(info);
ipv4_hdr = (struct rte_ipv4_hdr *)((char *)vxlan_gpe_hdr +
vxlan_gpe_len);
info->l2_len = 0;
} else if (vxlan_gpe_hdr->proto == RTE_VXLAN_GPE_TYPE_IPV6) {
- info->is_tunnel = 1;
- info->outer_ethertype = info->ethertype;
- info->outer_l2_len = info->l2_len;
- info->outer_l3_len = info->l3_len;
- info->outer_l4_proto = info->l4_proto;
+ update_tunnel_outer(info);
ipv6_hdr = (struct rte_ipv6_hdr *)((char *)vxlan_gpe_hdr +
vxlan_gpe_len);
info->l2_len = 0;
} else if (vxlan_gpe_hdr->proto == RTE_VXLAN_GPE_TYPE_ETH) {
- info->is_tunnel = 1;
- info->outer_ethertype = info->ethertype;
- info->outer_l2_len = info->l2_len;
- info->outer_l3_len = info->l3_len;
- info->outer_l4_proto = info->l4_proto;
+ update_tunnel_outer(info);
eth_hdr = (struct rte_ether_hdr *)((char *)vxlan_gpe_hdr +
vxlan_gpe_len);
info->l2_len += RTE_ETHER_VXLAN_GPE_HLEN;
}
-/* Fill in outer layers length */
-static void
-update_tunnel_outer(struct testpmd_offload_info *info)
-{
- info->is_tunnel = 1;
- info->outer_ethertype = info->ethertype;
- info->outer_l2_len = info->l2_len;
- info->outer_l3_len = info->l3_len;
- info->outer_l4_proto = info->l4_proto;
-}
-
/* Parse a geneve header */
static void
parse_geneve(struct rte_udp_hdr *udp_hdr,
gre_len += GRE_EXT_LEN;
if (gre_hdr->proto == _htons(RTE_ETHER_TYPE_IPV4)) {
- info->is_tunnel = 1;
- info->outer_ethertype = info->ethertype;
- info->outer_l2_len = info->l2_len;
- info->outer_l3_len = info->l3_len;
- info->outer_l4_proto = info->l4_proto;
+ update_tunnel_outer(info);
ipv4_hdr = (struct rte_ipv4_hdr *)((char *)gre_hdr + gre_len);
info->l2_len = 0;
} else if (gre_hdr->proto == _htons(RTE_ETHER_TYPE_IPV6)) {
- info->is_tunnel = 1;
- info->outer_ethertype = info->ethertype;
- info->outer_l2_len = info->l2_len;
- info->outer_l3_len = info->l3_len;
- info->outer_l4_proto = info->l4_proto;
+ update_tunnel_outer(info);
ipv6_hdr = (struct rte_ipv6_hdr *)((char *)gre_hdr + gre_len);
info->l2_len = 0;
} else if (gre_hdr->proto == _htons(RTE_ETHER_TYPE_TEB)) {
- info->is_tunnel = 1;
- info->outer_ethertype = info->ethertype;
- info->outer_l2_len = info->l2_len;
- info->outer_l3_len = info->l3_len;
- info->outer_l4_proto = info->l4_proto;
+ update_tunnel_outer(info);
eth_hdr = (struct rte_ether_hdr *)((char *)gre_hdr + gre_len);
mp = current_fwd_lcore()->mbp;
if (tx_pkt_split == TX_PKT_SPLIT_RND)
- nb_seg = random() % tx_pkt_nb_segs + 1;
+ nb_seg = rte_rand() % tx_pkt_nb_segs + 1;
else
nb_seg = tx_pkt_nb_segs;
uint32_t rx_bad_ip_csum;
uint32_t rx_bad_l4_csum;
uint32_t rx_bad_outer_l4_csum;
+ uint32_t rx_bad_outer_ip_csum;
struct testpmd_offload_info info;
uint16_t nb_segments = 0;
int ret;
rx_bad_ip_csum = 0;
rx_bad_l4_csum = 0;
rx_bad_outer_l4_csum = 0;
+ rx_bad_outer_ip_csum = 0;
gro_enable = gro_ports[fs->rx_port].enable;
txp = &ports[fs->tx_port];
rx_bad_l4_csum += 1;
if (rx_ol_flags & PKT_RX_OUTER_L4_CKSUM_BAD)
rx_bad_outer_l4_csum += 1;
+ if (rx_ol_flags & PKT_RX_OUTER_IP_CKSUM_BAD)
+ rx_bad_outer_ip_csum += 1;
/* step 1: dissect packet, parsing optional vlan, ip4/ip6, vxlan
* and inner headers */
ret = rte_gso_segment(pkts_burst[i], gso_ctx,
&gso_segments[nb_segments],
GSO_MAX_PKT_BURST - nb_segments);
- if (ret >= 0)
+ if (ret >= 1) {
+ /* pkts_burst[i] can be freed safely here. */
+ rte_pktmbuf_free(pkts_burst[i]);
nb_segments += ret;
- else {
+ } else if (ret == 0) {
+ /* 0 means it can be transmitted directly
+ * without gso.
+ */
+ gso_segments[nb_segments] = pkts_burst[i];
+ nb_segments += 1;
+ } else {
TESTPMD_LOG(DEBUG, "Unable to segment packet");
rte_pktmbuf_free(pkts_burst[i]);
}
nb_prep = rte_eth_tx_prepare(fs->tx_port, fs->tx_queue,
tx_pkts_burst, nb_rx);
if (nb_prep != nb_rx)
- printf("Preparing packet burst to transmit failed: %s\n",
- rte_strerror(rte_errno));
+ fprintf(stderr,
+ "Preparing packet burst to transmit failed: %s\n",
+ rte_strerror(rte_errno));
nb_tx = rte_eth_tx_burst(fs->tx_port, fs->tx_queue, tx_pkts_burst,
nb_prep);
fs->rx_bad_ip_csum += rx_bad_ip_csum;
fs->rx_bad_l4_csum += rx_bad_l4_csum;
fs->rx_bad_outer_l4_csum += rx_bad_outer_l4_csum;
+ fs->rx_bad_outer_ip_csum += rx_bad_outer_ip_csum;
inc_tx_burst_stats(fs, nb_tx);
if (unlikely(nb_tx < nb_rx)) {