X-Git-Url: http://git.droids-corp.org/?a=blobdiff_plain;f=lib%2Flibrte_net%2Frte_net.h;h=94b06d9ee4a5688be14f8cf7a22220d260368c26;hb=4a67e71816199cf8992a1a02a5aedd8b648382ef;hp=2173f95a92b063453e0fb8262013d7df514096c8;hpb=a7c528e5d71ff3f569898d268f9de129fdfc152b;p=dpdk.git diff --git a/lib/librte_net/rte_net.h b/lib/librte_net/rte_net.h index 2173f95a92..94b06d9ee4 100644 --- a/lib/librte_net/rte_net.h +++ b/lib/librte_net/rte_net.h @@ -51,7 +51,8 @@ struct rte_net_hdr_lens { * Protocol that follows IPv6 header. * -1 if an error occurs during mbuf parsing. */ -int __rte_experimental +__rte_experimental +int rte_net_skip_ip6_ext(uint16_t proto, const struct rte_mbuf *m, uint32_t *off, int *frag); @@ -112,27 +113,24 @@ uint32_t rte_net_get_ptype(const struct rte_mbuf *m, static inline int rte_net_intel_cksum_flags_prepare(struct rte_mbuf *m, uint64_t ol_flags) { - struct rte_ipv4_hdr *ipv4_hdr; + /* Initialise ipv4_hdr to avoid false positive compiler warnings. */ + struct rte_ipv4_hdr *ipv4_hdr = NULL; struct rte_ipv6_hdr *ipv6_hdr; - struct tcp_hdr *tcp_hdr; - struct udp_hdr *udp_hdr; + struct rte_tcp_hdr *tcp_hdr; + struct rte_udp_hdr *udp_hdr; uint64_t inner_l3_offset = m->l2_len; -#ifdef RTE_LIBRTE_ETHDEV_DEBUG /* * Does packet set any of available offloads? * Mainly it is required to avoid fragmented headers check if * no offloads are requested. */ - if (!(ol_flags & PKT_TX_OFFLOAD_MASK)) + if (!(ol_flags & (PKT_TX_IP_CKSUM | PKT_TX_L4_MASK | PKT_TX_TCP_SEG))) return 0; -#endif - if ((ol_flags & PKT_TX_OUTER_IP_CKSUM) || - (ol_flags & PKT_TX_OUTER_IPV6)) + if (ol_flags & (PKT_TX_OUTER_IPV4 | PKT_TX_OUTER_IPV6)) inner_l3_offset += m->outer_l2_len + m->outer_l3_len; -#ifdef RTE_LIBRTE_ETHDEV_DEBUG /* * Check if headers are fragmented. * The check could be less strict depending on which offloads are @@ -141,7 +139,6 @@ rte_net_intel_cksum_flags_prepare(struct rte_mbuf *m, uint64_t ol_flags) if (unlikely(rte_pktmbuf_data_len(m) < inner_l3_offset + m->l3_len + m->l4_len)) return -ENOTSUP; -#endif if (ol_flags & PKT_TX_IPV4) { ipv4_hdr = rte_pktmbuf_mtod_offset(m, struct rte_ipv4_hdr *, @@ -151,9 +148,9 @@ rte_net_intel_cksum_flags_prepare(struct rte_mbuf *m, uint64_t ol_flags) ipv4_hdr->hdr_checksum = 0; } - if ((ol_flags & PKT_TX_UDP_CKSUM) == PKT_TX_UDP_CKSUM) { + if ((ol_flags & PKT_TX_L4_MASK) == PKT_TX_UDP_CKSUM) { if (ol_flags & PKT_TX_IPV4) { - udp_hdr = (struct udp_hdr *)((char *)ipv4_hdr + + udp_hdr = (struct rte_udp_hdr *)((char *)ipv4_hdr + m->l3_len); udp_hdr->dgram_cksum = rte_ipv4_phdr_cksum(ipv4_hdr, ol_flags); @@ -161,16 +158,17 @@ rte_net_intel_cksum_flags_prepare(struct rte_mbuf *m, uint64_t ol_flags) ipv6_hdr = rte_pktmbuf_mtod_offset(m, struct rte_ipv6_hdr *, inner_l3_offset); /* non-TSO udp */ - udp_hdr = rte_pktmbuf_mtod_offset(m, struct udp_hdr *, + udp_hdr = rte_pktmbuf_mtod_offset(m, + struct rte_udp_hdr *, inner_l3_offset + m->l3_len); udp_hdr->dgram_cksum = rte_ipv6_phdr_cksum(ipv6_hdr, ol_flags); } - } else if ((ol_flags & PKT_TX_TCP_CKSUM) || + } else if ((ol_flags & PKT_TX_L4_MASK) == PKT_TX_TCP_CKSUM || (ol_flags & PKT_TX_TCP_SEG)) { if (ol_flags & PKT_TX_IPV4) { /* non-TSO tcp or TSO */ - tcp_hdr = (struct tcp_hdr *)((char *)ipv4_hdr + + tcp_hdr = (struct rte_tcp_hdr *)((char *)ipv4_hdr + m->l3_len); tcp_hdr->cksum = rte_ipv4_phdr_cksum(ipv4_hdr, ol_flags); @@ -178,7 +176,8 @@ rte_net_intel_cksum_flags_prepare(struct rte_mbuf *m, uint64_t ol_flags) ipv6_hdr = rte_pktmbuf_mtod_offset(m, struct rte_ipv6_hdr *, inner_l3_offset); /* non-TSO tcp or TSO */ - tcp_hdr = rte_pktmbuf_mtod_offset(m, struct tcp_hdr *, + tcp_hdr = rte_pktmbuf_mtod_offset(m, + struct rte_tcp_hdr *, inner_l3_offset + m->l3_len); tcp_hdr->cksum = rte_ipv6_phdr_cksum(ipv6_hdr, ol_flags);