static uint16_t
get_psd_sum(void *l3_hdr, uint16_t ethertype, uint64_t ol_flags)
{
- if (ethertype == RTE_ETHER_TYPE_IPv4)
+ if (ethertype == RTE_ETHER_TYPE_IPV4)
return rte_ipv4_phdr_cksum(l3_hdr, ol_flags);
- else /* assume ethertype == RTE_ETHER_TYPE_IPv6 */
+ else /* assume ethertype == RTE_ETHER_TYPE_IPV6 */
return rte_ipv6_phdr_cksum(l3_hdr, ol_flags);
}
}
switch (ethertype) {
- case RTE_ETHER_TYPE_IPv4:
+ case RTE_ETHER_TYPE_IPV4:
ipv4_hdr = (struct rte_ipv4_hdr *)
((char *)eth_hdr + info->outer_l2_len);
info->outer_l3_len = sizeof(struct rte_ipv4_hdr);
*l4_proto = ipv4_hdr->next_proto_id;
break;
- case RTE_ETHER_TYPE_IPv6:
+ case RTE_ETHER_TYPE_IPV6:
ipv6_hdr = (struct rte_ipv6_hdr *)
((char *)eth_hdr + info->outer_l2_len);
info->outer_l3_len = sizeof(struct rte_ipv6_hdr);
uint16_t ethertype;
struct rte_ipv4_hdr *ipv4_hdr;
struct rte_ipv6_hdr *ipv6_hdr;
- struct udp_hdr *udp_hdr;
- struct tcp_hdr *tcp_hdr;
+ struct rte_udp_hdr *udp_hdr;
+ struct rte_tcp_hdr *tcp_hdr;
struct rte_sctp_hdr *sctp_hdr;
uint64_t ol_flags = 0;
l3_hdr = (char *)eth_hdr + info->l2_len;
- if (ethertype == RTE_ETHER_TYPE_IPv4) {
+ if (ethertype == RTE_ETHER_TYPE_IPV4) {
ipv4_hdr = (struct rte_ipv4_hdr *)l3_hdr;
ipv4_hdr->hdr_checksum = 0;
ol_flags |= PKT_TX_IPV4;
ol_flags |= PKT_TX_IP_CKSUM;
info->l3_len = sizeof(struct rte_ipv4_hdr);
l4_proto = ipv4_hdr->next_proto_id;
- } else if (ethertype == RTE_ETHER_TYPE_IPv6) {
+ } else if (ethertype == RTE_ETHER_TYPE_IPV6) {
ipv6_hdr = (struct rte_ipv6_hdr *)l3_hdr;
info->l3_len = sizeof(struct rte_ipv6_hdr);
l4_proto = ipv6_hdr->proto;
return 0; /* packet type not supported, nothing to do */
if (l4_proto == IPPROTO_UDP) {
- udp_hdr = (struct udp_hdr *)((char *)l3_hdr + info->l3_len);
+ udp_hdr = (struct rte_udp_hdr *)((char *)l3_hdr + info->l3_len);
ol_flags |= PKT_TX_UDP_CKSUM;
udp_hdr->dgram_cksum = get_psd_sum(l3_hdr,
ethertype, ol_flags);
} else if (l4_proto == IPPROTO_TCP) {
- tcp_hdr = (struct tcp_hdr *)((char *)l3_hdr + info->l3_len);
+ tcp_hdr = (struct rte_tcp_hdr *)((char *)l3_hdr + info->l3_len);
/* Put PKT_TX_TCP_SEG bit setting before get_psd_sum(), because
* it depends on PKT_TX_TCP_SEG to calculate pseudo-header
* checksum.
{
uint8_t l4_proto = 0;
uint16_t outer_header_len;
- struct udp_hdr *udp_hdr;
+ struct rte_udp_hdr *udp_hdr;
union tunnel_offload_info info = { .data = 0 };
struct rte_ether_hdr *phdr =
rte_pktmbuf_mtod(pkt, struct rte_ether_hdr *);
if (l4_proto != IPPROTO_UDP)
return -1;
- udp_hdr = (struct udp_hdr *)((char *)phdr +
+ udp_hdr = (struct rte_udp_hdr *)((char *)phdr +
info.outer_l2_len + info.outer_l3_len);
/** check udp destination port, 4789 is the default vxlan port
* (rfc7348) or that the rx offload flag is set (i40e only
* currently)*/
- if (udp_hdr->dst_port != rte_cpu_to_be_16(DEFAULT_VXLAN_PORT) &&
+ if (udp_hdr->dst_port != rte_cpu_to_be_16(RTE_VXLAN_DEFAULT_PORT) &&
(pkt->packet_type & RTE_PTYPE_TUNNEL_MASK) == 0)
return -1;
outer_header_len = info.outer_l2_len + info.outer_l3_len
- + sizeof(struct udp_hdr) + sizeof(struct rte_vxlan_hdr);
+ + sizeof(struct rte_udp_hdr) + sizeof(struct rte_vxlan_hdr);
rte_pktmbuf_adj(pkt, outer_header_len);
struct rte_ether_hdr *pneth =
(struct rte_ether_hdr *) rte_pktmbuf_prepend(m,
sizeof(struct rte_ether_hdr) + sizeof(struct rte_ipv4_hdr)
- + sizeof(struct udp_hdr) + sizeof(struct rte_vxlan_hdr));
+ + sizeof(struct rte_udp_hdr) + sizeof(struct rte_vxlan_hdr));
struct rte_ipv4_hdr *ip = (struct rte_ipv4_hdr *) &pneth[1];
- struct udp_hdr *udp = (struct udp_hdr *) &ip[1];
+ struct rte_udp_hdr *udp = (struct rte_udp_hdr *) &ip[1];
struct rte_vxlan_hdr *vxlan = (struct rte_vxlan_hdr *) &udp[1];
/* convert TX queue ID to vport ID */
/*UDP HEADER*/
udp->dgram_cksum = 0;
udp->dgram_len = rte_cpu_to_be_16(old_len
- + sizeof(struct udp_hdr)
+ + sizeof(struct rte_udp_hdr)
+ sizeof(struct rte_vxlan_hdr));
udp->dst_port = rte_cpu_to_be_16(vxdev.dst_port);