git.droids-corp.org
/
dpdk.git
/ blobdiff
commit
grep
author
committer
pickaxe
?
search:
re
summary
|
shortlog
|
log
|
commit
|
commitdiff
|
tree
raw
|
inline
| side by side
net: add rte prefix to ether defines
[dpdk.git]
/
lib
/
librte_net
/
rte_net.c
diff --git
a/lib/librte_net/rte_net.c
b/lib/librte_net/rte_net.c
index
315c37c
..
5551cce
100644
(file)
--- a/
lib/librte_net/rte_net.c
+++ b/
lib/librte_net/rte_net.c
@@
-155,16
+155,16
@@
ptype_tunnel(uint16_t *proto, const struct rte_mbuf *m,
*off += opt_len[flags];
*proto = gh->proto;
*off += opt_len[flags];
*proto = gh->proto;
- if (*proto == rte_cpu_to_be_16(ETHER_TYPE_TEB))
+ if (*proto == rte_cpu_to_be_16(
RTE_
ETHER_TYPE_TEB))
return RTE_PTYPE_TUNNEL_NVGRE;
else
return RTE_PTYPE_TUNNEL_GRE;
}
case IPPROTO_IPIP:
return RTE_PTYPE_TUNNEL_NVGRE;
else
return RTE_PTYPE_TUNNEL_GRE;
}
case IPPROTO_IPIP:
- *proto = rte_cpu_to_be_16(ETHER_TYPE_IPv4);
+ *proto = rte_cpu_to_be_16(
RTE_
ETHER_TYPE_IPv4);
return RTE_PTYPE_TUNNEL_IP;
case IPPROTO_IPV6:
return RTE_PTYPE_TUNNEL_IP;
case IPPROTO_IPV6:
- *proto = rte_cpu_to_be_16(ETHER_TYPE_IPv6);
+ *proto = rte_cpu_to_be_16(
RTE_
ETHER_TYPE_IPv6);
return RTE_PTYPE_TUNNEL_IP; /* IP is also valid for IPv6 */
default:
return 0;
return RTE_PTYPE_TUNNEL_IP; /* IP is also valid for IPv6 */
default:
return 0;
@@
-249,10
+249,10
@@
uint32_t rte_net_get_ptype(const struct rte_mbuf *m,
if ((layers & RTE_PTYPE_L2_MASK) == 0)
return 0;
if ((layers & RTE_PTYPE_L2_MASK) == 0)
return 0;
- if (proto == rte_cpu_to_be_16(ETHER_TYPE_IPv4))
+ if (proto == rte_cpu_to_be_16(
RTE_
ETHER_TYPE_IPv4))
goto l3; /* fast path if packet is IPv4 */
goto l3; /* fast path if packet is IPv4 */
- if (proto == rte_cpu_to_be_16(ETHER_TYPE_VLAN)) {
+ if (proto == rte_cpu_to_be_16(
RTE_
ETHER_TYPE_VLAN)) {
const struct rte_vlan_hdr *vh;
struct rte_vlan_hdr vh_copy;
const struct rte_vlan_hdr *vh;
struct rte_vlan_hdr vh_copy;
@@
-263,7
+263,7
@@
uint32_t rte_net_get_ptype(const struct rte_mbuf *m,
off += sizeof(*vh);
hdr_lens->l2_len += sizeof(*vh);
proto = vh->eth_proto;
off += sizeof(*vh);
hdr_lens->l2_len += sizeof(*vh);
proto = vh->eth_proto;
- } else if (proto == rte_cpu_to_be_16(ETHER_TYPE_QINQ)) {
+ } else if (proto == rte_cpu_to_be_16(
RTE_
ETHER_TYPE_QINQ)) {
const struct rte_vlan_hdr *vh;
struct rte_vlan_hdr vh_copy;
const struct rte_vlan_hdr *vh;
struct rte_vlan_hdr vh_copy;
@@
-275,8
+275,8
@@
uint32_t rte_net_get_ptype(const struct rte_mbuf *m,
off += 2 * sizeof(*vh);
hdr_lens->l2_len += 2 * sizeof(*vh);
proto = vh->eth_proto;
off += 2 * sizeof(*vh);
hdr_lens->l2_len += 2 * sizeof(*vh);
proto = vh->eth_proto;
- } else if ((proto == rte_cpu_to_be_16(ETHER_TYPE_MPLS)) ||
- (proto == rte_cpu_to_be_16(ETHER_TYPE_MPLSM))) {
+ } else if ((proto == rte_cpu_to_be_16(
RTE_
ETHER_TYPE_MPLS)) ||
+ (proto == rte_cpu_to_be_16(
RTE_
ETHER_TYPE_MPLSM))) {
unsigned int i;
const struct mpls_hdr *mh;
struct mpls_hdr mh_copy;
unsigned int i;
const struct mpls_hdr *mh;
struct mpls_hdr mh_copy;
@@
-299,7
+299,7
@@
l3:
if ((layers & RTE_PTYPE_L3_MASK) == 0)
return pkt_type;
if ((layers & RTE_PTYPE_L3_MASK) == 0)
return pkt_type;
- if (proto == rte_cpu_to_be_16(ETHER_TYPE_IPv4)) {
+ if (proto == rte_cpu_to_be_16(
RTE_
ETHER_TYPE_IPv4)) {
const struct ipv4_hdr *ip4h;
struct ipv4_hdr ip4h_copy;
const struct ipv4_hdr *ip4h;
struct ipv4_hdr ip4h_copy;
@@
-322,7
+322,7
@@
l3:
}
proto = ip4h->next_proto_id;
pkt_type |= ptype_l4(proto);
}
proto = ip4h->next_proto_id;
pkt_type |= ptype_l4(proto);
- } else if (proto == rte_cpu_to_be_16(ETHER_TYPE_IPv6)) {
+ } else if (proto == rte_cpu_to_be_16(
RTE_
ETHER_TYPE_IPv6)) {
const struct ipv6_hdr *ip6h;
struct ipv6_hdr ip6h_copy;
int frag = 0;
const struct ipv6_hdr *ip6h;
struct ipv6_hdr ip6h_copy;
int frag = 0;
@@
-391,7
+391,7
@@
l3:
return pkt_type;
hdr_lens->inner_l2_len = 0;
return pkt_type;
hdr_lens->inner_l2_len = 0;
- if (proto == rte_cpu_to_be_16(ETHER_TYPE_TEB)) {
+ if (proto == rte_cpu_to_be_16(
RTE_
ETHER_TYPE_TEB)) {
eh = rte_pktmbuf_read(m, off, sizeof(*eh), &eh_copy);
if (unlikely(eh == NULL))
return pkt_type;
eh = rte_pktmbuf_read(m, off, sizeof(*eh), &eh_copy);
if (unlikely(eh == NULL))
return pkt_type;
@@
-401,7
+401,7
@@
l3:
hdr_lens->inner_l2_len = sizeof(*eh);
}
hdr_lens->inner_l2_len = sizeof(*eh);
}
- if (proto == rte_cpu_to_be_16(ETHER_TYPE_VLAN)) {
+ if (proto == rte_cpu_to_be_16(
RTE_
ETHER_TYPE_VLAN)) {
const struct rte_vlan_hdr *vh;
struct rte_vlan_hdr vh_copy;
const struct rte_vlan_hdr *vh;
struct rte_vlan_hdr vh_copy;
@@
-413,7
+413,7
@@
l3:
off += sizeof(*vh);
hdr_lens->inner_l2_len += sizeof(*vh);
proto = vh->eth_proto;
off += sizeof(*vh);
hdr_lens->inner_l2_len += sizeof(*vh);
proto = vh->eth_proto;
- } else if (proto == rte_cpu_to_be_16(ETHER_TYPE_QINQ)) {
+ } else if (proto == rte_cpu_to_be_16(
RTE_
ETHER_TYPE_QINQ)) {
const struct rte_vlan_hdr *vh;
struct rte_vlan_hdr vh_copy;
const struct rte_vlan_hdr *vh;
struct rte_vlan_hdr vh_copy;
@@
-431,7
+431,7
@@
l3:
if ((layers & RTE_PTYPE_INNER_L3_MASK) == 0)
return pkt_type;
if ((layers & RTE_PTYPE_INNER_L3_MASK) == 0)
return pkt_type;
- if (proto == rte_cpu_to_be_16(ETHER_TYPE_IPv4)) {
+ if (proto == rte_cpu_to_be_16(
RTE_
ETHER_TYPE_IPv4)) {
const struct ipv4_hdr *ip4h;
struct ipv4_hdr ip4h_copy;
const struct ipv4_hdr *ip4h;
struct ipv4_hdr ip4h_copy;
@@
-454,7
+454,7
@@
l3:
}
proto = ip4h->next_proto_id;
pkt_type |= ptype_inner_l4(proto);
}
proto = ip4h->next_proto_id;
pkt_type |= ptype_inner_l4(proto);
- } else if (proto == rte_cpu_to_be_16(ETHER_TYPE_IPv6)) {
+ } else if (proto == rte_cpu_to_be_16(
RTE_
ETHER_TYPE_IPv6)) {
const struct ipv6_hdr *ip6h;
struct ipv6_hdr ip6h_copy;
int frag = 0;
const struct ipv6_hdr *ip6h;
struct ipv6_hdr ip6h_copy;
int frag = 0;