X-Git-Url: http://git.droids-corp.org/?a=blobdiff_plain;f=lib%2Flibrte_pipeline%2Frte_table_action.c;h=1a295a19c6ece8c268d93f0aa21633aabb82a9c7;hb=a7c528e5d71ff3f569898d268f9de129fdfc152b;hp=edb33402153324687f65b7d280a1853db8ed28e5;hpb=96303217a6063d9085655e5b22a4cb8389b0f240;p=dpdk.git diff --git a/lib/librte_pipeline/rte_table_action.c b/lib/librte_pipeline/rte_table_action.c index edb3340215..1a295a19c6 100644 --- a/lib/librte_pipeline/rte_table_action.c +++ b/lib/librte_pipeline/rte_table_action.c @@ -107,39 +107,31 @@ mtr_cfg_check(struct rte_table_action_mtr_config *mtr) return 0; } -#define MBUF_SCHED_QUEUE_TC_COLOR(queue, tc, color) \ - ((uint16_t)((((uint64_t)(queue)) & 0x3) | \ - ((((uint64_t)(tc)) & 0x3) << 2) | \ - ((((uint64_t)(color)) & 0x3) << 4))) - -#define MBUF_SCHED_COLOR(sched, color) \ - (((sched) & (~0x30LLU)) | ((color) << 4)) - struct mtr_trtcm_data { struct rte_meter_trtcm trtcm; - uint64_t stats[e_RTE_METER_COLORS]; + uint64_t stats[RTE_COLORS]; } __attribute__((__packed__)); #define MTR_TRTCM_DATA_METER_PROFILE_ID_GET(data) \ - (((data)->stats[e_RTE_METER_GREEN] & 0xF8LLU) >> 3) + (((data)->stats[RTE_COLOR_GREEN] & 0xF8LLU) >> 3) static void mtr_trtcm_data_meter_profile_id_set(struct mtr_trtcm_data *data, uint32_t profile_id) { - data->stats[e_RTE_METER_GREEN] &= ~0xF8LLU; - data->stats[e_RTE_METER_GREEN] |= (profile_id % 32) << 3; + data->stats[RTE_COLOR_GREEN] &= ~0xF8LLU; + data->stats[RTE_COLOR_GREEN] |= (profile_id % 32) << 3; } #define MTR_TRTCM_DATA_POLICER_ACTION_DROP_GET(data, color)\ (((data)->stats[(color)] & 4LLU) >> 2) #define MTR_TRTCM_DATA_POLICER_ACTION_COLOR_GET(data, color)\ - ((enum rte_meter_color)((data)->stats[(color)] & 3LLU)) + ((enum rte_color)((data)->stats[(color)] & 3LLU)) static void mtr_trtcm_data_policer_action_set(struct mtr_trtcm_data *data, - enum rte_meter_color color, + enum rte_color color, enum rte_table_action_policer action) { if (action == RTE_TABLE_ACTION_POLICER_DROP) { @@ -152,14 +144,14 @@ mtr_trtcm_data_policer_action_set(struct mtr_trtcm_data *data, static uint64_t mtr_trtcm_data_stats_get(struct mtr_trtcm_data *data, - enum rte_meter_color color) + enum rte_color color) { return data->stats[color] >> 8; } static void mtr_trtcm_data_stats_reset(struct mtr_trtcm_data *data, - enum rte_meter_color color) + enum rte_color color) { data->stats[color] &= 0xFFLU; } @@ -174,9 +166,9 @@ mtr_data_size(struct rte_table_action_mtr_config *mtr) } struct dscp_table_entry_data { - enum rte_meter_color color; + enum rte_color color; uint16_t tc; - uint16_t queue_tc_color; + uint16_t tc_queue; }; struct dscp_table_data { @@ -295,16 +287,16 @@ mtr_apply(struct mtr_trtcm_data *data, /* Policer actions */ mtr_trtcm_data_policer_action_set(data_tc, - e_RTE_METER_GREEN, - p_tc->policer[e_RTE_METER_GREEN]); + RTE_COLOR_GREEN, + p_tc->policer[RTE_COLOR_GREEN]); mtr_trtcm_data_policer_action_set(data_tc, - e_RTE_METER_YELLOW, - p_tc->policer[e_RTE_METER_YELLOW]); + RTE_COLOR_YELLOW, + p_tc->policer[RTE_COLOR_YELLOW]); mtr_trtcm_data_policer_action_set(data_tc, - e_RTE_METER_RED, - p_tc->policer[e_RTE_METER_RED]); + RTE_COLOR_RED, + p_tc->policer[RTE_COLOR_RED]); } return 0; @@ -319,17 +311,15 @@ pkt_work_mtr(struct rte_mbuf *mbuf, uint32_t dscp, uint16_t total_length) { - uint64_t drop_mask, sched; - uint64_t *sched_ptr = (uint64_t *) &mbuf->hash.sched; + uint64_t drop_mask; struct dscp_table_entry_data *dscp_entry = &dscp_table->entry[dscp]; - enum rte_meter_color color_in, color_meter, color_policer; + enum rte_color color_in, color_meter, color_policer; uint32_t tc, mp_id; tc = dscp_entry->tc; color_in = dscp_entry->color; data += tc; mp_id = MTR_TRTCM_DATA_METER_PROFILE_ID_GET(data); - sched = *sched_ptr; /* Meter */ color_meter = rte_meter_trtcm_color_aware_check( @@ -346,7 +336,7 @@ pkt_work_mtr(struct rte_mbuf *mbuf, drop_mask = MTR_TRTCM_DATA_POLICER_ACTION_DROP_GET(data, color_meter); color_policer = MTR_TRTCM_DATA_POLICER_ACTION_COLOR_GET(data, color_meter); - *sched_ptr = MBUF_SCHED_COLOR(sched, color_policer); + rte_mbuf_sched_color_set(mbuf, (uint8_t)color_policer); return drop_mask; } @@ -368,9 +358,8 @@ tm_cfg_check(struct rte_table_action_tm_config *tm) } struct tm_data { - uint16_t queue_tc_color; - uint16_t subport; - uint32_t pipe; + uint32_t queue_id; + uint32_t reserved; } __attribute__((__packed__)); static int @@ -397,9 +386,9 @@ tm_apply(struct tm_data *data, return status; /* Apply */ - data->queue_tc_color = 0; - data->subport = (uint16_t) p->subport_id; - data->pipe = p->pipe_id; + data->queue_id = p->subport_id << + (__builtin_ctz(cfg->n_pipes_per_subport) + 4) | + p->pipe_id << 4; return 0; } @@ -411,12 +400,11 @@ pkt_work_tm(struct rte_mbuf *mbuf, uint32_t dscp) { struct dscp_table_entry_data *dscp_entry = &dscp_table->entry[dscp]; - struct tm_data *sched_ptr = (struct tm_data *) &mbuf->hash.sched; - struct tm_data sched; - - sched = *data; - sched.queue_tc_color = dscp_entry->queue_tc_color; - *sched_ptr = sched; + uint32_t queue_id = data->queue_id | + (dscp_entry->tc << 2) | + dscp_entry->tc_queue; + rte_mbuf_sched_set(mbuf, queue_id, dscp_entry->tc, + (uint8_t)dscp_entry->color); } /** @@ -432,6 +420,7 @@ encap_valid(enum rte_table_action_encap_type encap) case RTE_TABLE_ACTION_ENCAP_MPLS: case RTE_TABLE_ACTION_ENCAP_PPPOE: case RTE_TABLE_ACTION_ENCAP_VXLAN: + case RTE_TABLE_ACTION_ENCAP_QINQ_PPPOE: return 1; default: return 0; @@ -449,7 +438,7 @@ encap_cfg_check(struct rte_table_action_encap_config *encap) } struct encap_ether_data { - struct ether_hdr ether; + struct rte_ether_hdr ether; } __attribute__((__packed__)); #define VLAN(pcp, dei, vid) \ @@ -458,14 +447,14 @@ struct encap_ether_data { (((uint64_t)(vid)) & 0xFFFLLU)) \ struct encap_vlan_data { - struct ether_hdr ether; - struct vlan_hdr vlan; + struct rte_ether_hdr ether; + struct rte_vlan_hdr vlan; } __attribute__((__packed__)); struct encap_qinq_data { - struct ether_hdr ether; - struct vlan_hdr svlan; - struct vlan_hdr cvlan; + struct rte_ether_hdr ether; + struct rte_vlan_hdr svlan; + struct rte_vlan_hdr cvlan; } __attribute__((__packed__)); #define ETHER_TYPE_MPLS_UNICAST 0x8847 @@ -479,13 +468,11 @@ struct encap_qinq_data { (((uint64_t)(ttl)) & 0xFFLLU))) struct encap_mpls_data { - struct ether_hdr ether; + struct rte_ether_hdr ether; uint32_t mpls[RTE_TABLE_ACTION_MPLS_LABELS_MAX]; uint32_t mpls_count; } __attribute__((__packed__)); -#define ETHER_TYPE_PPPOE_SESSION 0x8864 - #define PPP_PROTOCOL_IP 0x0021 struct pppoe_ppp_hdr { @@ -496,40 +483,47 @@ struct pppoe_ppp_hdr { } __attribute__((__packed__)); struct encap_pppoe_data { - struct ether_hdr ether; + struct rte_ether_hdr ether; struct pppoe_ppp_hdr pppoe_ppp; } __attribute__((__packed__)); #define IP_PROTO_UDP 17 struct encap_vxlan_ipv4_data { - struct ether_hdr ether; - struct ipv4_hdr ipv4; + struct rte_ether_hdr ether; + struct rte_ipv4_hdr ipv4; struct udp_hdr udp; - struct vxlan_hdr vxlan; + struct rte_vxlan_hdr vxlan; } __attribute__((__packed__)); struct encap_vxlan_ipv4_vlan_data { - struct ether_hdr ether; - struct vlan_hdr vlan; - struct ipv4_hdr ipv4; + struct rte_ether_hdr ether; + struct rte_vlan_hdr vlan; + struct rte_ipv4_hdr ipv4; struct udp_hdr udp; - struct vxlan_hdr vxlan; + struct rte_vxlan_hdr vxlan; } __attribute__((__packed__)); struct encap_vxlan_ipv6_data { - struct ether_hdr ether; - struct ipv6_hdr ipv6; + struct rte_ether_hdr ether; + struct rte_ipv6_hdr ipv6; struct udp_hdr udp; - struct vxlan_hdr vxlan; + struct rte_vxlan_hdr vxlan; } __attribute__((__packed__)); struct encap_vxlan_ipv6_vlan_data { - struct ether_hdr ether; - struct vlan_hdr vlan; - struct ipv6_hdr ipv6; + struct rte_ether_hdr ether; + struct rte_vlan_hdr vlan; + struct rte_ipv6_hdr ipv6; struct udp_hdr udp; - struct vxlan_hdr vxlan; + struct rte_vxlan_hdr vxlan; +} __attribute__((__packed__)); + +struct encap_qinq_pppoe_data { + struct rte_ether_hdr ether; + struct rte_vlan_hdr svlan; + struct rte_vlan_hdr cvlan; + struct pppoe_ppp_hdr pppoe_ppp; } __attribute__((__packed__)); static size_t @@ -563,6 +557,9 @@ encap_data_size(struct rte_table_action_encap_config *encap) else return sizeof(struct encap_vxlan_ipv6_data); + case 1LLU << RTE_TABLE_ACTION_ENCAP_QINQ_PPPOE: + return sizeof(struct encap_qinq_pppoe_data); + default: return 0; } @@ -599,6 +596,9 @@ encap_apply_check(struct rte_table_action_encap_params *p, case RTE_TABLE_ACTION_ENCAP_VXLAN: return 0; + case RTE_TABLE_ACTION_ENCAP_QINQ_PPPOE: + return 0; + default: return -EINVAL; } @@ -611,12 +611,12 @@ encap_ether_apply(void *data, { struct encap_ether_data *d = data; uint16_t ethertype = (common_cfg->ip_version) ? - ETHER_TYPE_IPv4 : - ETHER_TYPE_IPv6; + RTE_ETHER_TYPE_IPv4 : + RTE_ETHER_TYPE_IPv6; /* Ethernet */ - ether_addr_copy(&p->ether.ether.da, &d->ether.d_addr); - ether_addr_copy(&p->ether.ether.sa, &d->ether.s_addr); + rte_ether_addr_copy(&p->ether.ether.da, &d->ether.d_addr); + rte_ether_addr_copy(&p->ether.ether.sa, &d->ether.s_addr); d->ether.ether_type = rte_htons(ethertype); return 0; @@ -629,13 +629,13 @@ encap_vlan_apply(void *data, { struct encap_vlan_data *d = data; uint16_t ethertype = (common_cfg->ip_version) ? - ETHER_TYPE_IPv4 : - ETHER_TYPE_IPv6; + RTE_ETHER_TYPE_IPv4 : + RTE_ETHER_TYPE_IPv6; /* Ethernet */ - ether_addr_copy(&p->vlan.ether.da, &d->ether.d_addr); - ether_addr_copy(&p->vlan.ether.sa, &d->ether.s_addr); - d->ether.ether_type = rte_htons(ETHER_TYPE_VLAN); + rte_ether_addr_copy(&p->vlan.ether.da, &d->ether.d_addr); + rte_ether_addr_copy(&p->vlan.ether.sa, &d->ether.s_addr); + d->ether.ether_type = rte_htons(RTE_ETHER_TYPE_VLAN); /* VLAN */ d->vlan.vlan_tci = rte_htons(VLAN(p->vlan.vlan.pcp, @@ -653,19 +653,19 @@ encap_qinq_apply(void *data, { struct encap_qinq_data *d = data; uint16_t ethertype = (common_cfg->ip_version) ? - ETHER_TYPE_IPv4 : - ETHER_TYPE_IPv6; + RTE_ETHER_TYPE_IPv4 : + RTE_ETHER_TYPE_IPv6; /* Ethernet */ - ether_addr_copy(&p->qinq.ether.da, &d->ether.d_addr); - ether_addr_copy(&p->qinq.ether.sa, &d->ether.s_addr); - d->ether.ether_type = rte_htons(ETHER_TYPE_QINQ); + rte_ether_addr_copy(&p->qinq.ether.da, &d->ether.d_addr); + rte_ether_addr_copy(&p->qinq.ether.sa, &d->ether.s_addr); + d->ether.ether_type = rte_htons(RTE_ETHER_TYPE_QINQ); /* SVLAN */ d->svlan.vlan_tci = rte_htons(VLAN(p->qinq.svlan.pcp, p->qinq.svlan.dei, p->qinq.svlan.vid)); - d->svlan.eth_proto = rte_htons(ETHER_TYPE_VLAN); + d->svlan.eth_proto = rte_htons(RTE_ETHER_TYPE_VLAN); /* CVLAN */ d->cvlan.vlan_tci = rte_htons(VLAN(p->qinq.cvlan.pcp, @@ -676,6 +676,38 @@ encap_qinq_apply(void *data, return 0; } +static int +encap_qinq_pppoe_apply(void *data, + struct rte_table_action_encap_params *p) +{ + struct encap_qinq_pppoe_data *d = data; + + /* Ethernet */ + rte_ether_addr_copy(&p->qinq.ether.da, &d->ether.d_addr); + rte_ether_addr_copy(&p->qinq.ether.sa, &d->ether.s_addr); + d->ether.ether_type = rte_htons(RTE_ETHER_TYPE_VLAN); + + /* SVLAN */ + d->svlan.vlan_tci = rte_htons(VLAN(p->qinq.svlan.pcp, + p->qinq.svlan.dei, + p->qinq.svlan.vid)); + d->svlan.eth_proto = rte_htons(RTE_ETHER_TYPE_VLAN); + + /* CVLAN */ + d->cvlan.vlan_tci = rte_htons(VLAN(p->qinq.cvlan.pcp, + p->qinq.cvlan.dei, + p->qinq.cvlan.vid)); + d->cvlan.eth_proto = rte_htons(ETHER_TYPE_PPPOE_SESSION); + + /* PPPoE and PPP*/ + d->pppoe_ppp.ver_type_code = rte_htons(0x1100); + d->pppoe_ppp.session_id = rte_htons(p->qinq_pppoe.pppoe.session_id); + d->pppoe_ppp.length = 0; /* not pre-computed */ + d->pppoe_ppp.protocol = rte_htons(PPP_PROTOCOL_IP); + + return 0; +} + static int encap_mpls_apply(void *data, struct rte_table_action_encap_params *p) @@ -687,8 +719,8 @@ encap_mpls_apply(void *data, uint32_t i; /* Ethernet */ - ether_addr_copy(&p->mpls.ether.da, &d->ether.d_addr); - ether_addr_copy(&p->mpls.ether.sa, &d->ether.s_addr); + rte_ether_addr_copy(&p->mpls.ether.da, &d->ether.d_addr); + rte_ether_addr_copy(&p->mpls.ether.sa, &d->ether.s_addr); d->ether.ether_type = rte_htons(ethertype); /* MPLS */ @@ -714,8 +746,8 @@ encap_pppoe_apply(void *data, struct encap_pppoe_data *d = data; /* Ethernet */ - ether_addr_copy(&p->pppoe.ether.da, &d->ether.d_addr); - ether_addr_copy(&p->pppoe.ether.sa, &d->ether.s_addr); + rte_ether_addr_copy(&p->pppoe.ether.da, &d->ether.d_addr); + rte_ether_addr_copy(&p->pppoe.ether.sa, &d->ether.s_addr); d->ether.ether_type = rte_htons(ETHER_TYPE_PPPOE_SESSION); /* PPPoE and PPP*/ @@ -744,15 +776,17 @@ encap_vxlan_apply(void *data, struct encap_vxlan_ipv4_vlan_data *d = data; /* Ethernet */ - ether_addr_copy(&p->vxlan.ether.da, &d->ether.d_addr); - ether_addr_copy(&p->vxlan.ether.sa, &d->ether.s_addr); - d->ether.ether_type = rte_htons(ETHER_TYPE_VLAN); + rte_ether_addr_copy(&p->vxlan.ether.da, + &d->ether.d_addr); + rte_ether_addr_copy(&p->vxlan.ether.sa, + &d->ether.s_addr); + d->ether.ether_type = rte_htons(RTE_ETHER_TYPE_VLAN); /* VLAN */ d->vlan.vlan_tci = rte_htons(VLAN(p->vxlan.vlan.pcp, p->vxlan.vlan.dei, p->vxlan.vlan.vid)); - d->vlan.eth_proto = rte_htons(ETHER_TYPE_IPv4); + d->vlan.eth_proto = rte_htons(RTE_ETHER_TYPE_IPv4); /* IPv4*/ d->ipv4.version_ihl = 0x45; @@ -783,9 +817,11 @@ encap_vxlan_apply(void *data, struct encap_vxlan_ipv4_data *d = data; /* Ethernet */ - ether_addr_copy(&p->vxlan.ether.da, &d->ether.d_addr); - ether_addr_copy(&p->vxlan.ether.sa, &d->ether.s_addr); - d->ether.ether_type = rte_htons(ETHER_TYPE_IPv4); + rte_ether_addr_copy(&p->vxlan.ether.da, + &d->ether.d_addr); + rte_ether_addr_copy(&p->vxlan.ether.sa, + &d->ether.s_addr); + d->ether.ether_type = rte_htons(RTE_ETHER_TYPE_IPv4); /* IPv4*/ d->ipv4.version_ihl = 0x45; @@ -818,15 +854,17 @@ encap_vxlan_apply(void *data, struct encap_vxlan_ipv6_vlan_data *d = data; /* Ethernet */ - ether_addr_copy(&p->vxlan.ether.da, &d->ether.d_addr); - ether_addr_copy(&p->vxlan.ether.sa, &d->ether.s_addr); - d->ether.ether_type = rte_htons(ETHER_TYPE_VLAN); + rte_ether_addr_copy(&p->vxlan.ether.da, + &d->ether.d_addr); + rte_ether_addr_copy(&p->vxlan.ether.sa, + &d->ether.s_addr); + d->ether.ether_type = rte_htons(RTE_ETHER_TYPE_VLAN); /* VLAN */ d->vlan.vlan_tci = rte_htons(VLAN(p->vxlan.vlan.pcp, p->vxlan.vlan.dei, p->vxlan.vlan.vid)); - d->vlan.eth_proto = rte_htons(ETHER_TYPE_IPv6); + d->vlan.eth_proto = rte_htons(RTE_ETHER_TYPE_IPv6); /* IPv6*/ d->ipv6.vtc_flow = rte_htonl((6 << 28) | @@ -857,9 +895,11 @@ encap_vxlan_apply(void *data, struct encap_vxlan_ipv6_data *d = data; /* Ethernet */ - ether_addr_copy(&p->vxlan.ether.da, &d->ether.d_addr); - ether_addr_copy(&p->vxlan.ether.sa, &d->ether.s_addr); - d->ether.ether_type = rte_htons(ETHER_TYPE_IPv6); + rte_ether_addr_copy(&p->vxlan.ether.da, + &d->ether.d_addr); + rte_ether_addr_copy(&p->vxlan.ether.sa, + &d->ether.s_addr); + d->ether.ether_type = rte_htons(RTE_ETHER_TYPE_IPv6); /* IPv6*/ d->ipv6.vtc_flow = rte_htonl((6 << 28) | @@ -921,6 +961,9 @@ encap_apply(void *data, case RTE_TABLE_ACTION_ENCAP_VXLAN: return encap_vxlan_apply(data, p, cfg); + case RTE_TABLE_ACTION_ENCAP_QINQ_PPPOE: + return encap_qinq_pppoe_apply(data, p); + default: return -EINVAL; } @@ -962,13 +1005,13 @@ pkt_work_encap_vxlan_ipv4(struct rte_mbuf *mbuf, ether_length = (uint16_t)mbuf->pkt_len; ipv4_total_length = ether_length + - (sizeof(struct vxlan_hdr) + + (sizeof(struct rte_vxlan_hdr) + sizeof(struct udp_hdr) + - sizeof(struct ipv4_hdr)); + sizeof(struct rte_ipv4_hdr)); ipv4_hdr_cksum = encap_vxlan_ipv4_checksum_update(vxlan_tbl->ipv4.hdr_checksum, rte_htons(ipv4_total_length)); udp_length = ether_length + - (sizeof(struct vxlan_hdr) + + (sizeof(struct rte_vxlan_hdr) + sizeof(struct udp_hdr)); vxlan_pkt = encap(ether, vxlan_tbl, sizeof(*vxlan_tbl)); @@ -992,13 +1035,13 @@ pkt_work_encap_vxlan_ipv4_vlan(struct rte_mbuf *mbuf, ether_length = (uint16_t)mbuf->pkt_len; ipv4_total_length = ether_length + - (sizeof(struct vxlan_hdr) + + (sizeof(struct rte_vxlan_hdr) + sizeof(struct udp_hdr) + - sizeof(struct ipv4_hdr)); + sizeof(struct rte_ipv4_hdr)); ipv4_hdr_cksum = encap_vxlan_ipv4_checksum_update(vxlan_tbl->ipv4.hdr_checksum, rte_htons(ipv4_total_length)); udp_length = ether_length + - (sizeof(struct vxlan_hdr) + + (sizeof(struct rte_vxlan_hdr) + sizeof(struct udp_hdr)); vxlan_pkt = encap(ether, vxlan_tbl, sizeof(*vxlan_tbl)); @@ -1022,10 +1065,10 @@ pkt_work_encap_vxlan_ipv6(struct rte_mbuf *mbuf, ether_length = (uint16_t)mbuf->pkt_len; ipv6_payload_length = ether_length + - (sizeof(struct vxlan_hdr) + + (sizeof(struct rte_vxlan_hdr) + sizeof(struct udp_hdr)); udp_length = ether_length + - (sizeof(struct vxlan_hdr) + + (sizeof(struct rte_vxlan_hdr) + sizeof(struct udp_hdr)); vxlan_pkt = encap(ether, vxlan_tbl, sizeof(*vxlan_tbl)); @@ -1048,10 +1091,10 @@ pkt_work_encap_vxlan_ipv6_vlan(struct rte_mbuf *mbuf, ether_length = (uint16_t)mbuf->pkt_len; ipv6_payload_length = ether_length + - (sizeof(struct vxlan_hdr) + + (sizeof(struct rte_vxlan_hdr) + sizeof(struct udp_hdr)); udp_length = ether_length + - (sizeof(struct vxlan_hdr) + + (sizeof(struct rte_vxlan_hdr) + sizeof(struct udp_hdr)); vxlan_pkt = encap(ether, vxlan_tbl, sizeof(*vxlan_tbl)); @@ -1098,7 +1141,7 @@ pkt_work_encap(struct rte_mbuf *mbuf, case 1LLU << RTE_TABLE_ACTION_ENCAP_MPLS: { struct encap_mpls_data *mpls = data; - size_t size = sizeof(struct ether_hdr) + + size_t size = sizeof(struct rte_ether_hdr) + mpls->mpls_count * 4; encap(ip, data, size); @@ -1119,6 +1162,18 @@ pkt_work_encap(struct rte_mbuf *mbuf, break; } + case 1LLU << RTE_TABLE_ACTION_ENCAP_QINQ_PPPOE: + { + struct encap_qinq_pppoe_data *qinq_pppoe = + encap(ip, data, sizeof(struct encap_qinq_pppoe_data)); + qinq_pppoe->pppoe_ppp.length = rte_htons(total_length + 2); + mbuf->data_off = ip_offset - (sizeof(struct rte_mbuf) + + sizeof(struct encap_qinq_pppoe_data)); + mbuf->pkt_len = mbuf->data_len = total_length + + sizeof(struct encap_qinq_pppoe_data); + break; + } + case 1LLU << RTE_TABLE_ACTION_ENCAP_VXLAN: { if (cfg->vxlan.ip_version) @@ -1287,7 +1342,7 @@ nat_ipv6_tcp_udp_checksum_update(uint16_t cksum0, } static __rte_always_inline void -pkt_ipv4_work_nat(struct ipv4_hdr *ip, +pkt_ipv4_work_nat(struct rte_ipv4_hdr *ip, struct nat_ipv4_data *data, struct rte_table_action_nat_config *cfg) { @@ -1373,7 +1428,7 @@ pkt_ipv4_work_nat(struct ipv4_hdr *ip, } static __rte_always_inline void -pkt_ipv6_work_nat(struct ipv6_hdr *ip, +pkt_ipv6_work_nat(struct rte_ipv6_hdr *ip, struct nat_ipv6_data *data, struct rte_table_action_nat_config *cfg) { @@ -1481,7 +1536,7 @@ ttl_apply(void *data, } static __rte_always_inline uint64_t -pkt_ipv4_work_ttl(struct ipv4_hdr *ip, +pkt_ipv4_work_ttl(struct rte_ipv4_hdr *ip, struct ttl_data *data) { uint32_t drop; @@ -1502,7 +1557,7 @@ pkt_ipv4_work_ttl(struct ipv4_hdr *ip, } static __rte_always_inline uint64_t -pkt_ipv6_work_ttl(struct ipv6_hdr *ip, +pkt_ipv6_work_ttl(struct rte_ipv6_hdr *ip, struct ttl_data *data) { uint32_t drop; @@ -1694,10 +1749,9 @@ get_block_size(const struct rte_crypto_sym_xform *xform, uint8_t cdev_id) rte_cryptodev_info_get(cdev_id, &dev_info); - for (i = 0;; i++) { + for (i = 0; dev_info.capabilities[i].op != RTE_CRYPTO_OP_TYPE_UNDEFINED; + i++) { cap = &dev_info.capabilities[i]; - if (!cap) - break; if (cap->sym.xform_type != xform->type) continue; @@ -2011,6 +2065,127 @@ pkt_work_sym_crypto(struct rte_mbuf *mbuf, struct sym_crypto_data *data, return 0; } +/** + * RTE_TABLE_ACTION_TAG + */ +struct tag_data { + uint32_t tag; +} __attribute__((__packed__)); + +static int +tag_apply(struct tag_data *data, + struct rte_table_action_tag_params *p) +{ + data->tag = p->tag; + return 0; +} + +static __rte_always_inline void +pkt_work_tag(struct rte_mbuf *mbuf, + struct tag_data *data) +{ + mbuf->hash.fdir.hi = data->tag; + mbuf->ol_flags |= PKT_RX_FDIR | PKT_RX_FDIR_ID; +} + +static __rte_always_inline void +pkt4_work_tag(struct rte_mbuf *mbuf0, + struct rte_mbuf *mbuf1, + struct rte_mbuf *mbuf2, + struct rte_mbuf *mbuf3, + struct tag_data *data0, + struct tag_data *data1, + struct tag_data *data2, + struct tag_data *data3) +{ + mbuf0->hash.fdir.hi = data0->tag; + mbuf1->hash.fdir.hi = data1->tag; + mbuf2->hash.fdir.hi = data2->tag; + mbuf3->hash.fdir.hi = data3->tag; + + mbuf0->ol_flags |= PKT_RX_FDIR | PKT_RX_FDIR_ID; + mbuf1->ol_flags |= PKT_RX_FDIR | PKT_RX_FDIR_ID; + mbuf2->ol_flags |= PKT_RX_FDIR | PKT_RX_FDIR_ID; + mbuf3->ol_flags |= PKT_RX_FDIR | PKT_RX_FDIR_ID; +} + +/** + * RTE_TABLE_ACTION_DECAP + */ +struct decap_data { + uint16_t n; +} __attribute__((__packed__)); + +static int +decap_apply(struct decap_data *data, + struct rte_table_action_decap_params *p) +{ + data->n = p->n; + return 0; +} + +static __rte_always_inline void +pkt_work_decap(struct rte_mbuf *mbuf, + struct decap_data *data) +{ + uint16_t data_off = mbuf->data_off; + uint16_t data_len = mbuf->data_len; + uint32_t pkt_len = mbuf->pkt_len; + uint16_t n = data->n; + + mbuf->data_off = data_off + n; + mbuf->data_len = data_len - n; + mbuf->pkt_len = pkt_len - n; +} + +static __rte_always_inline void +pkt4_work_decap(struct rte_mbuf *mbuf0, + struct rte_mbuf *mbuf1, + struct rte_mbuf *mbuf2, + struct rte_mbuf *mbuf3, + struct decap_data *data0, + struct decap_data *data1, + struct decap_data *data2, + struct decap_data *data3) +{ + uint16_t data_off0 = mbuf0->data_off; + uint16_t data_len0 = mbuf0->data_len; + uint32_t pkt_len0 = mbuf0->pkt_len; + + uint16_t data_off1 = mbuf1->data_off; + uint16_t data_len1 = mbuf1->data_len; + uint32_t pkt_len1 = mbuf1->pkt_len; + + uint16_t data_off2 = mbuf2->data_off; + uint16_t data_len2 = mbuf2->data_len; + uint32_t pkt_len2 = mbuf2->pkt_len; + + uint16_t data_off3 = mbuf3->data_off; + uint16_t data_len3 = mbuf3->data_len; + uint32_t pkt_len3 = mbuf3->pkt_len; + + uint16_t n0 = data0->n; + uint16_t n1 = data1->n; + uint16_t n2 = data2->n; + uint16_t n3 = data3->n; + + mbuf0->data_off = data_off0 + n0; + mbuf0->data_len = data_len0 - n0; + mbuf0->pkt_len = pkt_len0 - n0; + + mbuf1->data_off = data_off1 + n1; + mbuf1->data_len = data_len1 - n1; + mbuf1->pkt_len = pkt_len1 - n1; + + mbuf2->data_off = data_off2 + n2; + mbuf2->data_len = data_len2 - n2; + mbuf2->pkt_len = pkt_len2 - n2; + + mbuf3->data_off = data_off3 + n3; + mbuf3->data_len = data_len3 - n3; + mbuf3->pkt_len = pkt_len3 - n3; +} + /** * Action profile */ @@ -2028,6 +2203,8 @@ action_valid(enum rte_table_action_type action) case RTE_TABLE_ACTION_STATS: case RTE_TABLE_ACTION_TIME: case RTE_TABLE_ACTION_SYM_CRYPTO: + case RTE_TABLE_ACTION_TAG: + case RTE_TABLE_ACTION_DECAP: return 1; default: return 0; @@ -2162,6 +2339,12 @@ action_data_size(enum rte_table_action_type action, case RTE_TABLE_ACTION_SYM_CRYPTO: return (sizeof(struct sym_crypto_data)); + case RTE_TABLE_ACTION_TAG: + return sizeof(struct tag_data); + + case RTE_TABLE_ACTION_DECAP: + return sizeof(struct decap_data); + default: return 0; } @@ -2419,6 +2602,14 @@ rte_table_action_apply(struct rte_table_action *action, &action->cfg.sym_crypto, action_params); + case RTE_TABLE_ACTION_TAG: + return tag_apply(action_data, + action_params); + + case RTE_TABLE_ACTION_DECAP: + return decap_apply(action_data, + action_params); + default: return -EINVAL; } @@ -2444,17 +2635,13 @@ rte_table_action_dscp_table_update(struct rte_table_action *action, &action->dscp_table.entry[i]; struct rte_table_action_dscp_table_entry *entry = &table->entry[i]; - uint16_t queue_tc_color = - MBUF_SCHED_QUEUE_TC_COLOR(entry->tc_queue_id, - entry->tc_id, - entry->color); if ((dscp_mask & (1LLU << i)) == 0) continue; data->color = entry->color; data->tc = entry->tc_id; - data->queue_tc_color = queue_tc_color; + data->tc_queue = entry->tc_queue_id; } return 0; @@ -2552,14 +2739,14 @@ rte_table_action_meter_read(struct rte_table_action *action, if ((tc_mask & (1 << i)) == 0) continue; - dst->n_packets[e_RTE_METER_GREEN] = - mtr_trtcm_data_stats_get(src, e_RTE_METER_GREEN); + dst->n_packets[RTE_COLOR_GREEN] = + mtr_trtcm_data_stats_get(src, RTE_COLOR_GREEN); - dst->n_packets[e_RTE_METER_YELLOW] = - mtr_trtcm_data_stats_get(src, e_RTE_METER_YELLOW); + dst->n_packets[RTE_COLOR_YELLOW] = + mtr_trtcm_data_stats_get(src, RTE_COLOR_YELLOW); - dst->n_packets[e_RTE_METER_RED] = - mtr_trtcm_data_stats_get(src, e_RTE_METER_RED); + dst->n_packets[RTE_COLOR_RED] = + mtr_trtcm_data_stats_get(src, RTE_COLOR_RED); dst->n_packets_valid = 1; dst->n_bytes_valid = 0; @@ -2576,9 +2763,9 @@ rte_table_action_meter_read(struct rte_table_action *action, if ((tc_mask & (1 << i)) == 0) continue; - mtr_trtcm_data_stats_reset(src, e_RTE_METER_GREEN); - mtr_trtcm_data_stats_reset(src, e_RTE_METER_YELLOW); - mtr_trtcm_data_stats_reset(src, e_RTE_METER_RED); + mtr_trtcm_data_stats_reset(src, RTE_COLOR_GREEN); + mtr_trtcm_data_stats_reset(src, RTE_COLOR_YELLOW); + mtr_trtcm_data_stats_reset(src, RTE_COLOR_RED); } @@ -2706,16 +2893,16 @@ pkt_work(struct rte_mbuf *mbuf, uint16_t total_length; if (cfg->common.ip_version) { - struct ipv4_hdr *hdr = ip; + struct rte_ipv4_hdr *hdr = ip; dscp = hdr->type_of_service >> 2; total_length = rte_ntohs(hdr->total_length); } else { - struct ipv6_hdr *hdr = ip; + struct rte_ipv6_hdr *hdr = ip; dscp = (rte_ntohl(hdr->vtc_flow) & 0x0F600000) >> 18; - total_length = - rte_ntohs(hdr->payload_len) + sizeof(struct ipv6_hdr); + total_length = rte_ntohs(hdr->payload_len) + + sizeof(struct rte_ipv6_hdr); } if (cfg->action_mask & (1LLU << RTE_TABLE_ACTION_LB)) { @@ -2749,6 +2936,14 @@ pkt_work(struct rte_mbuf *mbuf, dscp); } + if (cfg->action_mask & (1LLU << RTE_TABLE_ACTION_DECAP)) { + void *data = action_data_get(table_entry, + action, + RTE_TABLE_ACTION_DECAP); + + pkt_work_decap(mbuf, data); + } + if (cfg->action_mask & (1LLU << RTE_TABLE_ACTION_ENCAP)) { void *data = action_data_get(table_entry, action, RTE_TABLE_ACTION_ENCAP); @@ -2803,6 +2998,14 @@ pkt_work(struct rte_mbuf *mbuf, ip_offset); } + if (cfg->action_mask & (1LLU << RTE_TABLE_ACTION_TAG)) { + void *data = action_data_get(table_entry, + action, + RTE_TABLE_ACTION_TAG); + + pkt_work_tag(mbuf, data); + } + return drop_mask; } @@ -2838,10 +3041,10 @@ pkt4_work(struct rte_mbuf **mbufs, uint16_t total_length0, total_length1, total_length2, total_length3; if (cfg->common.ip_version) { - struct ipv4_hdr *hdr0 = ip0; - struct ipv4_hdr *hdr1 = ip1; - struct ipv4_hdr *hdr2 = ip2; - struct ipv4_hdr *hdr3 = ip3; + struct rte_ipv4_hdr *hdr0 = ip0; + struct rte_ipv4_hdr *hdr1 = ip1; + struct rte_ipv4_hdr *hdr2 = ip2; + struct rte_ipv4_hdr *hdr3 = ip3; dscp0 = hdr0->type_of_service >> 2; dscp1 = hdr1->type_of_service >> 2; @@ -2853,24 +3056,24 @@ pkt4_work(struct rte_mbuf **mbufs, total_length2 = rte_ntohs(hdr2->total_length); total_length3 = rte_ntohs(hdr3->total_length); } else { - struct ipv6_hdr *hdr0 = ip0; - struct ipv6_hdr *hdr1 = ip1; - struct ipv6_hdr *hdr2 = ip2; - struct ipv6_hdr *hdr3 = ip3; + struct rte_ipv6_hdr *hdr0 = ip0; + struct rte_ipv6_hdr *hdr1 = ip1; + struct rte_ipv6_hdr *hdr2 = ip2; + struct rte_ipv6_hdr *hdr3 = ip3; dscp0 = (rte_ntohl(hdr0->vtc_flow) & 0x0F600000) >> 18; dscp1 = (rte_ntohl(hdr1->vtc_flow) & 0x0F600000) >> 18; dscp2 = (rte_ntohl(hdr2->vtc_flow) & 0x0F600000) >> 18; dscp3 = (rte_ntohl(hdr3->vtc_flow) & 0x0F600000) >> 18; - total_length0 = - rte_ntohs(hdr0->payload_len) + sizeof(struct ipv6_hdr); - total_length1 = - rte_ntohs(hdr1->payload_len) + sizeof(struct ipv6_hdr); - total_length2 = - rte_ntohs(hdr2->payload_len) + sizeof(struct ipv6_hdr); - total_length3 = - rte_ntohs(hdr3->payload_len) + sizeof(struct ipv6_hdr); + total_length0 = rte_ntohs(hdr0->payload_len) + + sizeof(struct rte_ipv6_hdr); + total_length1 = rte_ntohs(hdr1->payload_len) + + sizeof(struct rte_ipv6_hdr); + total_length2 = rte_ntohs(hdr2->payload_len) + + sizeof(struct rte_ipv6_hdr); + total_length3 = rte_ntohs(hdr3->payload_len) + + sizeof(struct rte_ipv6_hdr); } if (cfg->action_mask & (1LLU << RTE_TABLE_ACTION_LB)) { @@ -2974,6 +3177,24 @@ pkt4_work(struct rte_mbuf **mbufs, dscp3); } + if (cfg->action_mask & (1LLU << RTE_TABLE_ACTION_DECAP)) { + void *data0 = action_data_get(table_entry0, + action, + RTE_TABLE_ACTION_DECAP); + void *data1 = action_data_get(table_entry1, + action, + RTE_TABLE_ACTION_DECAP); + void *data2 = action_data_get(table_entry2, + action, + RTE_TABLE_ACTION_DECAP); + void *data3 = action_data_get(table_entry3, + action, + RTE_TABLE_ACTION_DECAP); + + pkt4_work_decap(mbuf0, mbuf1, mbuf2, mbuf3, + data0, data1, data2, data3); + } + if (cfg->action_mask & (1LLU << RTE_TABLE_ACTION_ENCAP)) { void *data0 = action_data_get(table_entry0, action, RTE_TABLE_ACTION_ENCAP); @@ -3111,6 +3332,24 @@ pkt4_work(struct rte_mbuf **mbufs, ip_offset); } + if (cfg->action_mask & (1LLU << RTE_TABLE_ACTION_TAG)) { + void *data0 = action_data_get(table_entry0, + action, + RTE_TABLE_ACTION_TAG); + void *data1 = action_data_get(table_entry1, + action, + RTE_TABLE_ACTION_TAG); + void *data2 = action_data_get(table_entry2, + action, + RTE_TABLE_ACTION_TAG); + void *data3 = action_data_get(table_entry3, + action, + RTE_TABLE_ACTION_TAG); + + pkt4_work_tag(mbuf0, mbuf1, mbuf2, mbuf3, + data0, data1, data2, data3); + } + return drop_mask0 | (drop_mask1 << 1) | (drop_mask2 << 2) |