git.droids-corp.org
/
dpdk.git
/ blobdiff
commit
grep
author
committer
pickaxe
?
search:
re
summary
|
shortlog
|
log
|
commit
|
commitdiff
|
tree
raw
|
inline
| side by side
remove experimental tags from all symbol definitions
[dpdk.git]
/
lib
/
librte_net
/
rte_net.c
diff --git
a/lib/librte_net/rte_net.c
b/lib/librte_net/rte_net.c
index
7123834
..
dfccbbf
100644
(file)
--- a/
lib/librte_net/rte_net.c
+++ b/
lib/librte_net/rte_net.c
@@
-161,10
+161,10
@@
ptype_tunnel(uint16_t *proto, const struct rte_mbuf *m,
return RTE_PTYPE_TUNNEL_GRE;
}
case IPPROTO_IPIP:
return RTE_PTYPE_TUNNEL_GRE;
}
case IPPROTO_IPIP:
- *proto = rte_cpu_to_be_16(RTE_ETHER_TYPE_IP
v
4);
+ *proto = rte_cpu_to_be_16(RTE_ETHER_TYPE_IP
V
4);
return RTE_PTYPE_TUNNEL_IP;
case IPPROTO_IPV6:
return RTE_PTYPE_TUNNEL_IP;
case IPPROTO_IPV6:
- *proto = rte_cpu_to_be_16(RTE_ETHER_TYPE_IP
v
6);
+ *proto = rte_cpu_to_be_16(RTE_ETHER_TYPE_IP
V
6);
return RTE_PTYPE_TUNNEL_IP; /* IP is also valid for IPv6 */
default:
return 0;
return RTE_PTYPE_TUNNEL_IP; /* IP is also valid for IPv6 */
default:
return 0;
@@
-179,7
+179,7
@@
ip4_hlen(const struct rte_ipv4_hdr *hdr)
}
/* parse ipv6 extended headers, update offset and return next proto */
}
/* parse ipv6 extended headers, update offset and return next proto */
-int
__rte_experimental
+int
rte_net_skip_ip6_ext(uint16_t proto, const struct rte_mbuf *m, uint32_t *off,
int *frag)
{
rte_net_skip_ip6_ext(uint16_t proto, const struct rte_mbuf *m, uint32_t *off,
int *frag)
{
@@
-249,7
+249,7
@@
uint32_t rte_net_get_ptype(const struct rte_mbuf *m,
if ((layers & RTE_PTYPE_L2_MASK) == 0)
return 0;
if ((layers & RTE_PTYPE_L2_MASK) == 0)
return 0;
- if (proto == rte_cpu_to_be_16(RTE_ETHER_TYPE_IP
v
4))
+ if (proto == rte_cpu_to_be_16(RTE_ETHER_TYPE_IP
V
4))
goto l3; /* fast path if packet is IPv4 */
if (proto == rte_cpu_to_be_16(RTE_ETHER_TYPE_VLAN)) {
goto l3; /* fast path if packet is IPv4 */
if (proto == rte_cpu_to_be_16(RTE_ETHER_TYPE_VLAN)) {
@@
-299,7
+299,7
@@
l3:
if ((layers & RTE_PTYPE_L3_MASK) == 0)
return pkt_type;
if ((layers & RTE_PTYPE_L3_MASK) == 0)
return pkt_type;
- if (proto == rte_cpu_to_be_16(RTE_ETHER_TYPE_IP
v
4)) {
+ if (proto == rte_cpu_to_be_16(RTE_ETHER_TYPE_IP
V
4)) {
const struct rte_ipv4_hdr *ip4h;
struct rte_ipv4_hdr ip4h_copy;
const struct rte_ipv4_hdr *ip4h;
struct rte_ipv4_hdr ip4h_copy;
@@
-322,7
+322,7
@@
l3:
}
proto = ip4h->next_proto_id;
pkt_type |= ptype_l4(proto);
}
proto = ip4h->next_proto_id;
pkt_type |= ptype_l4(proto);
- } else if (proto == rte_cpu_to_be_16(RTE_ETHER_TYPE_IP
v
6)) {
+ } else if (proto == rte_cpu_to_be_16(RTE_ETHER_TYPE_IP
V
6)) {
const struct rte_ipv6_hdr *ip6h;
struct rte_ipv6_hdr ip6h_copy;
int frag = 0;
const struct rte_ipv6_hdr *ip6h;
struct rte_ipv6_hdr ip6h_copy;
int frag = 0;
@@
-357,11
+357,11
@@
l3:
}
if ((pkt_type & RTE_PTYPE_L4_MASK) == RTE_PTYPE_L4_UDP) {
}
if ((pkt_type & RTE_PTYPE_L4_MASK) == RTE_PTYPE_L4_UDP) {
- hdr_lens->l4_len = sizeof(struct udp_hdr);
+ hdr_lens->l4_len = sizeof(struct
rte_
udp_hdr);
return pkt_type;
} else if ((pkt_type & RTE_PTYPE_L4_MASK) == RTE_PTYPE_L4_TCP) {
return pkt_type;
} else if ((pkt_type & RTE_PTYPE_L4_MASK) == RTE_PTYPE_L4_TCP) {
- const struct tcp_hdr *th;
- struct tcp_hdr th_copy;
+ const struct
rte_
tcp_hdr *th;
+ struct
rte_
tcp_hdr th_copy;
th = rte_pktmbuf_read(m, off, sizeof(*th), &th_copy);
if (unlikely(th == NULL))
th = rte_pktmbuf_read(m, off, sizeof(*th), &th_copy);
if (unlikely(th == NULL))
@@
-370,7
+370,7
@@
l3:
hdr_lens->l4_len = (th->data_off & 0xf0) >> 2;
return pkt_type;
} else if ((pkt_type & RTE_PTYPE_L4_MASK) == RTE_PTYPE_L4_SCTP) {
hdr_lens->l4_len = (th->data_off & 0xf0) >> 2;
return pkt_type;
} else if ((pkt_type & RTE_PTYPE_L4_MASK) == RTE_PTYPE_L4_SCTP) {
- hdr_lens->l4_len = sizeof(struct sctp_hdr);
+ hdr_lens->l4_len = sizeof(struct
rte_
sctp_hdr);
return pkt_type;
} else {
uint32_t prev_off = off;
return pkt_type;
} else {
uint32_t prev_off = off;
@@
-431,7
+431,7
@@
l3:
if ((layers & RTE_PTYPE_INNER_L3_MASK) == 0)
return pkt_type;
if ((layers & RTE_PTYPE_INNER_L3_MASK) == 0)
return pkt_type;
- if (proto == rte_cpu_to_be_16(RTE_ETHER_TYPE_IP
v
4)) {
+ if (proto == rte_cpu_to_be_16(RTE_ETHER_TYPE_IP
V
4)) {
const struct rte_ipv4_hdr *ip4h;
struct rte_ipv4_hdr ip4h_copy;
const struct rte_ipv4_hdr *ip4h;
struct rte_ipv4_hdr ip4h_copy;
@@
-454,7
+454,7
@@
l3:
}
proto = ip4h->next_proto_id;
pkt_type |= ptype_inner_l4(proto);
}
proto = ip4h->next_proto_id;
pkt_type |= ptype_inner_l4(proto);
- } else if (proto == rte_cpu_to_be_16(RTE_ETHER_TYPE_IP
v
6)) {
+ } else if (proto == rte_cpu_to_be_16(RTE_ETHER_TYPE_IP
V
6)) {
const struct rte_ipv6_hdr *ip6h;
struct rte_ipv6_hdr ip6h_copy;
int frag = 0;
const struct rte_ipv6_hdr *ip6h;
struct rte_ipv6_hdr ip6h_copy;
int frag = 0;
@@
-493,11
+493,11
@@
l3:
}
if ((pkt_type & RTE_PTYPE_INNER_L4_MASK) == RTE_PTYPE_INNER_L4_UDP) {
}
if ((pkt_type & RTE_PTYPE_INNER_L4_MASK) == RTE_PTYPE_INNER_L4_UDP) {
- hdr_lens->inner_l4_len = sizeof(struct udp_hdr);
+ hdr_lens->inner_l4_len = sizeof(struct
rte_
udp_hdr);
} else if ((pkt_type & RTE_PTYPE_INNER_L4_MASK) ==
RTE_PTYPE_INNER_L4_TCP) {
} else if ((pkt_type & RTE_PTYPE_INNER_L4_MASK) ==
RTE_PTYPE_INNER_L4_TCP) {
- const struct tcp_hdr *th;
- struct tcp_hdr th_copy;
+ const struct
rte_
tcp_hdr *th;
+ struct
rte_
tcp_hdr th_copy;
th = rte_pktmbuf_read(m, off, sizeof(*th), &th_copy);
if (unlikely(th == NULL))
th = rte_pktmbuf_read(m, off, sizeof(*th), &th_copy);
if (unlikely(th == NULL))
@@
-506,7
+506,7
@@
l3:
hdr_lens->inner_l4_len = (th->data_off & 0xf0) >> 2;
} else if ((pkt_type & RTE_PTYPE_INNER_L4_MASK) ==
RTE_PTYPE_INNER_L4_SCTP) {
hdr_lens->inner_l4_len = (th->data_off & 0xf0) >> 2;
} else if ((pkt_type & RTE_PTYPE_INNER_L4_MASK) ==
RTE_PTYPE_INNER_L4_SCTP) {
- hdr_lens->inner_l4_len = sizeof(struct sctp_hdr);
+ hdr_lens->inner_l4_len = sizeof(struct
rte_
sctp_hdr);
} else {
hdr_lens->inner_l4_len = 0;
}
} else {
hdr_lens->inner_l4_len = 0;
}