#include <rte_malloc.h>
#include <rte_flow.h>
#include <rte_ethdev.h>
+#include <rte_vxlan.h>
+#include <rte_gtp.h>
#include "actions_gen.h"
#include "flow_gen.h"
#include "config.h"
+
/* Storage for additional parameters for actions */
struct additional_para {
uint16_t queue;
uint16_t next_table;
uint16_t *queues;
uint16_t queues_number;
+ uint32_t counter;
+ uint64_t encap_data;
+ uint64_t decap_data;
+};
+
+/* Storage for struct rte_flow_action_raw_encap including external data. */
+struct action_raw_encap_data {
+ struct rte_flow_action_raw_encap conf;
+ uint8_t data[128];
+ uint8_t preserve[128];
+ uint16_t idx;
+};
+
+/* Storage for struct rte_flow_action_raw_decap including external data. */
+struct action_raw_decap_data {
+ struct rte_flow_action_raw_decap conf;
+ uint8_t data[128];
+ uint16_t idx;
};
/* Storage for struct rte_flow_action_rss including external data. */
static void
add_mark(struct rte_flow_action *actions,
uint8_t actions_counter,
- __rte_unused struct additional_para para)
+ struct additional_para para)
{
static struct rte_flow_action_mark mark_action;
+ uint32_t counter = para.counter;
do {
- mark_action.id = MARK_ID;
+ /* Random values from 1 to 256 */
+ mark_action.id = (counter % 255) + 1;
} while (0);
actions[actions_counter].type = RTE_FLOW_ACTION_TYPE_MARK;
uint16_t queue;
- rss_data = rte_malloc("rss_data",
- sizeof(struct action_rss_data), 0);
+ if (rss_data == NULL)
+ rss_data = rte_malloc("rss_data",
+ sizeof(struct action_rss_data), 0);
if (rss_data == NULL)
rte_exit(EXIT_FAILURE, "No Memory available!");
actions[actions_counter].conf = &count_action;
}
+static void
+add_set_src_mac(struct rte_flow_action *actions,
+ uint8_t actions_counter,
+ __rte_unused struct additional_para para)
+{
+ static struct rte_flow_action_set_mac set_mac;
+ uint32_t mac = para.counter;
+ uint16_t i;
+
+ /* Fixed value */
+ if (FIXED_VALUES)
+ mac = 1;
+
+ /* Mac address to be set is random each time */
+ for (i = 0; i < RTE_ETHER_ADDR_LEN; i++) {
+ set_mac.mac_addr[i] = mac & 0xff;
+ mac = mac >> 8;
+ }
+
+ actions[actions_counter].type = RTE_FLOW_ACTION_TYPE_SET_MAC_SRC;
+ actions[actions_counter].conf = &set_mac;
+}
+
+static void
+add_set_dst_mac(struct rte_flow_action *actions,
+ uint8_t actions_counter,
+ __rte_unused struct additional_para para)
+{
+ static struct rte_flow_action_set_mac set_mac;
+ uint32_t mac = para.counter;
+ uint16_t i;
+
+ /* Fixed value */
+ if (FIXED_VALUES)
+ mac = 1;
+
+ /* Mac address to be set is random each time */
+ for (i = 0; i < RTE_ETHER_ADDR_LEN; i++) {
+ set_mac.mac_addr[i] = mac & 0xff;
+ mac = mac >> 8;
+ }
+
+ actions[actions_counter].type = RTE_FLOW_ACTION_TYPE_SET_MAC_DST;
+ actions[actions_counter].conf = &set_mac;
+}
+
+static void
+add_set_src_ipv4(struct rte_flow_action *actions,
+ uint8_t actions_counter,
+ __rte_unused struct additional_para para)
+{
+ static struct rte_flow_action_set_ipv4 set_ipv4;
+ uint32_t ip = para.counter;
+
+ /* Fixed value */
+ if (FIXED_VALUES)
+ ip = 1;
+
+ /* IPv4 value to be set is random each time */
+ set_ipv4.ipv4_addr = RTE_BE32(ip + 1);
+
+ actions[actions_counter].type = RTE_FLOW_ACTION_TYPE_SET_IPV4_SRC;
+ actions[actions_counter].conf = &set_ipv4;
+}
+
+static void
+add_set_dst_ipv4(struct rte_flow_action *actions,
+ uint8_t actions_counter,
+ __rte_unused struct additional_para para)
+{
+ static struct rte_flow_action_set_ipv4 set_ipv4;
+ uint32_t ip = para.counter;
+
+ /* Fixed value */
+ if (FIXED_VALUES)
+ ip = 1;
+
+ /* IPv4 value to be set is random each time */
+ set_ipv4.ipv4_addr = RTE_BE32(ip + 1);
+
+ actions[actions_counter].type = RTE_FLOW_ACTION_TYPE_SET_IPV4_DST;
+ actions[actions_counter].conf = &set_ipv4;
+}
+
+static void
+add_set_src_ipv6(struct rte_flow_action *actions,
+ uint8_t actions_counter,
+ __rte_unused struct additional_para para)
+{
+ static struct rte_flow_action_set_ipv6 set_ipv6;
+ uint32_t ipv6 = para.counter;
+ uint8_t i;
+
+ /* Fixed value */
+ if (FIXED_VALUES)
+ ipv6 = 1;
+
+ /* IPv6 value to set is random each time */
+ for (i = 0; i < 16; i++) {
+ set_ipv6.ipv6_addr[i] = ipv6 & 0xff;
+ ipv6 = ipv6 >> 8;
+ }
+
+ actions[actions_counter].type = RTE_FLOW_ACTION_TYPE_SET_IPV6_SRC;
+ actions[actions_counter].conf = &set_ipv6;
+}
+
+static void
+add_set_dst_ipv6(struct rte_flow_action *actions,
+ uint8_t actions_counter,
+ __rte_unused struct additional_para para)
+{
+ static struct rte_flow_action_set_ipv6 set_ipv6;
+ uint32_t ipv6 = para.counter;
+ uint8_t i;
+
+ /* Fixed value */
+ if (FIXED_VALUES)
+ ipv6 = 1;
+
+ /* IPv6 value to set is random each time */
+ for (i = 0; i < 16; i++) {
+ set_ipv6.ipv6_addr[i] = ipv6 & 0xff;
+ ipv6 = ipv6 >> 8;
+ }
+
+ actions[actions_counter].type = RTE_FLOW_ACTION_TYPE_SET_IPV6_DST;
+ actions[actions_counter].conf = &set_ipv6;
+}
+
+static void
+add_set_src_tp(struct rte_flow_action *actions,
+ uint8_t actions_counter,
+ __rte_unused struct additional_para para)
+{
+ static struct rte_flow_action_set_tp set_tp;
+ uint32_t tp = para.counter;
+
+ /* Fixed value */
+ if (FIXED_VALUES)
+ tp = 100;
+
+ /* TP src port is random each time */
+ tp = tp % 0xffff;
+
+ set_tp.port = RTE_BE16(tp & 0xffff);
+
+ actions[actions_counter].type = RTE_FLOW_ACTION_TYPE_SET_TP_SRC;
+ actions[actions_counter].conf = &set_tp;
+}
+
+static void
+add_set_dst_tp(struct rte_flow_action *actions,
+ uint8_t actions_counter,
+ __rte_unused struct additional_para para)
+{
+ static struct rte_flow_action_set_tp set_tp;
+ uint32_t tp = para.counter;
+
+ /* Fixed value */
+ if (FIXED_VALUES)
+ tp = 100;
+
+ /* TP src port is random each time */
+ if (tp > 0xffff)
+ tp = tp >> 16;
+
+ set_tp.port = RTE_BE16(tp & 0xffff);
+
+ actions[actions_counter].type = RTE_FLOW_ACTION_TYPE_SET_TP_DST;
+ actions[actions_counter].conf = &set_tp;
+}
+
+static void
+add_inc_tcp_ack(struct rte_flow_action *actions,
+ uint8_t actions_counter,
+ __rte_unused struct additional_para para)
+{
+ static rte_be32_t value;
+ uint32_t ack_value = para.counter;
+
+ /* Fixed value */
+ if (FIXED_VALUES)
+ ack_value = 1;
+
+ value = RTE_BE32(ack_value);
+
+ actions[actions_counter].type = RTE_FLOW_ACTION_TYPE_INC_TCP_ACK;
+ actions[actions_counter].conf = &value;
+}
+
+static void
+add_dec_tcp_ack(struct rte_flow_action *actions,
+ uint8_t actions_counter,
+ __rte_unused struct additional_para para)
+{
+ static rte_be32_t value;
+ uint32_t ack_value = para.counter;
+
+ /* Fixed value */
+ if (FIXED_VALUES)
+ ack_value = 1;
+
+ value = RTE_BE32(ack_value);
+
+ actions[actions_counter].type = RTE_FLOW_ACTION_TYPE_DEC_TCP_ACK;
+ actions[actions_counter].conf = &value;
+}
+
+static void
+add_inc_tcp_seq(struct rte_flow_action *actions,
+ uint8_t actions_counter,
+ __rte_unused struct additional_para para)
+{
+ static rte_be32_t value;
+ uint32_t seq_value = para.counter;
+
+ /* Fixed value */
+ if (FIXED_VALUES)
+ seq_value = 1;
+
+ value = RTE_BE32(seq_value);
+
+ actions[actions_counter].type = RTE_FLOW_ACTION_TYPE_INC_TCP_SEQ;
+ actions[actions_counter].conf = &value;
+}
+
+static void
+add_dec_tcp_seq(struct rte_flow_action *actions,
+ uint8_t actions_counter,
+ __rte_unused struct additional_para para)
+{
+ static rte_be32_t value;
+ uint32_t seq_value = para.counter;
+
+ /* Fixed value */
+ if (FIXED_VALUES)
+ seq_value = 1;
+
+ value = RTE_BE32(seq_value);
+
+ actions[actions_counter].type = RTE_FLOW_ACTION_TYPE_DEC_TCP_SEQ;
+ actions[actions_counter].conf = &value;
+}
+
+static void
+add_set_ttl(struct rte_flow_action *actions,
+ uint8_t actions_counter,
+ __rte_unused struct additional_para para)
+{
+ static struct rte_flow_action_set_ttl set_ttl;
+ uint32_t ttl_value = para.counter;
+
+ /* Fixed value */
+ if (FIXED_VALUES)
+ ttl_value = 1;
+
+ /* Set ttl to random value each time */
+ ttl_value = ttl_value % 0xff;
+
+ set_ttl.ttl_value = ttl_value;
+
+ actions[actions_counter].type = RTE_FLOW_ACTION_TYPE_SET_TTL;
+ actions[actions_counter].conf = &set_ttl;
+}
+
+static void
+add_dec_ttl(struct rte_flow_action *actions,
+ uint8_t actions_counter,
+ __rte_unused struct additional_para para)
+{
+ actions[actions_counter].type = RTE_FLOW_ACTION_TYPE_DEC_TTL;
+}
+
+static void
+add_set_ipv4_dscp(struct rte_flow_action *actions,
+ uint8_t actions_counter,
+ __rte_unused struct additional_para para)
+{
+ static struct rte_flow_action_set_dscp set_dscp;
+ uint32_t dscp_value = para.counter;
+
+ /* Fixed value */
+ if (FIXED_VALUES)
+ dscp_value = 1;
+
+ /* Set dscp to random value each time */
+ dscp_value = dscp_value % 0xff;
+
+ set_dscp.dscp = dscp_value;
+
+ actions[actions_counter].type = RTE_FLOW_ACTION_TYPE_SET_IPV4_DSCP;
+ actions[actions_counter].conf = &set_dscp;
+}
+
+static void
+add_set_ipv6_dscp(struct rte_flow_action *actions,
+ uint8_t actions_counter,
+ __rte_unused struct additional_para para)
+{
+ static struct rte_flow_action_set_dscp set_dscp;
+ uint32_t dscp_value = para.counter;
+
+ /* Fixed value */
+ if (FIXED_VALUES)
+ dscp_value = 1;
+
+ /* Set dscp to random value each time */
+ dscp_value = dscp_value % 0xff;
+
+ set_dscp.dscp = dscp_value;
+
+ actions[actions_counter].type = RTE_FLOW_ACTION_TYPE_SET_IPV6_DSCP;
+ actions[actions_counter].conf = &set_dscp;
+}
+
+static void
+add_flag(struct rte_flow_action *actions,
+ uint8_t actions_counter,
+ __rte_unused struct additional_para para)
+{
+ actions[actions_counter].type = RTE_FLOW_ACTION_TYPE_FLAG;
+}
+
+static void
+add_ether_header(uint8_t **header, uint64_t data,
+ __rte_unused struct additional_para para)
+{
+ struct rte_flow_item_eth eth_item;
+
+ if (!(data & FLOW_ITEM_MASK(RTE_FLOW_ITEM_TYPE_ETH)))
+ return;
+
+ memset(ð_item, 0, sizeof(struct rte_flow_item_eth));
+ if (data & FLOW_ITEM_MASK(RTE_FLOW_ITEM_TYPE_VLAN))
+ eth_item.type = RTE_BE16(RTE_ETHER_TYPE_VLAN);
+ else if (data & FLOW_ITEM_MASK(RTE_FLOW_ITEM_TYPE_IPV4))
+ eth_item.type = RTE_BE16(RTE_ETHER_TYPE_IPV4);
+ else if (data & FLOW_ITEM_MASK(RTE_FLOW_ITEM_TYPE_IPV6))
+ eth_item.type = RTE_BE16(RTE_ETHER_TYPE_IPV6);
+ memcpy(*header, ð_item, sizeof(eth_item));
+ *header += sizeof(eth_item);
+}
+
+static void
+add_vlan_header(uint8_t **header, uint64_t data,
+ __rte_unused struct additional_para para)
+{
+ struct rte_flow_item_vlan vlan_item;
+ uint16_t vlan_value;
+
+ if (!(data & FLOW_ITEM_MASK(RTE_FLOW_ITEM_TYPE_VLAN)))
+ return;
+
+ vlan_value = VLAN_VALUE;
+
+ memset(&vlan_item, 0, sizeof(struct rte_flow_item_vlan));
+ vlan_item.tci = RTE_BE16(vlan_value);
+
+ if (data & FLOW_ITEM_MASK(RTE_FLOW_ITEM_TYPE_IPV4))
+ vlan_item.inner_type = RTE_BE16(RTE_ETHER_TYPE_IPV4);
+ if (data & FLOW_ITEM_MASK(RTE_FLOW_ITEM_TYPE_IPV6))
+ vlan_item.inner_type = RTE_BE16(RTE_ETHER_TYPE_IPV6);
+ memcpy(*header, &vlan_item, sizeof(vlan_item));
+ *header += sizeof(vlan_item);
+}
+
+static void
+add_ipv4_header(uint8_t **header, uint64_t data,
+ struct additional_para para)
+{
+ struct rte_flow_item_ipv4 ipv4_item;
+ uint32_t ip_dst = para.counter;
+
+ if (!(data & FLOW_ITEM_MASK(RTE_FLOW_ITEM_TYPE_IPV4)))
+ return;
+
+ /* Fixed value */
+ if (FIXED_VALUES)
+ ip_dst = 1;
+
+ memset(&ipv4_item, 0, sizeof(struct rte_flow_item_ipv4));
+ ipv4_item.hdr.src_addr = RTE_IPV4(127, 0, 0, 1);
+ ipv4_item.hdr.dst_addr = RTE_BE32(ip_dst);
+ ipv4_item.hdr.version_ihl = RTE_IPV4_VHL_DEF;
+ if (data & FLOW_ITEM_MASK(RTE_FLOW_ITEM_TYPE_UDP))
+ ipv4_item.hdr.next_proto_id = RTE_IP_TYPE_UDP;
+ if (data & FLOW_ITEM_MASK(RTE_FLOW_ITEM_TYPE_GRE))
+ ipv4_item.hdr.next_proto_id = RTE_IP_TYPE_GRE;
+ memcpy(*header, &ipv4_item, sizeof(ipv4_item));
+ *header += sizeof(ipv4_item);
+}
+
+static void
+add_ipv6_header(uint8_t **header, uint64_t data,
+ __rte_unused struct additional_para para)
+{
+ struct rte_flow_item_ipv6 ipv6_item;
+
+ if (!(data & FLOW_ITEM_MASK(RTE_FLOW_ITEM_TYPE_IPV6)))
+ return;
+
+ memset(&ipv6_item, 0, sizeof(struct rte_flow_item_ipv6));
+ if (data & FLOW_ITEM_MASK(RTE_FLOW_ITEM_TYPE_UDP))
+ ipv6_item.hdr.proto = RTE_IP_TYPE_UDP;
+ if (data & FLOW_ITEM_MASK(RTE_FLOW_ITEM_TYPE_GRE))
+ ipv6_item.hdr.proto = RTE_IP_TYPE_GRE;
+ memcpy(*header, &ipv6_item, sizeof(ipv6_item));
+ *header += sizeof(ipv6_item);
+}
+
+static void
+add_udp_header(uint8_t **header, uint64_t data,
+ __rte_unused struct additional_para para)
+{
+ struct rte_flow_item_udp udp_item;
+
+ if (!(data & FLOW_ITEM_MASK(RTE_FLOW_ITEM_TYPE_UDP)))
+ return;
+
+ memset(&udp_item, 0, sizeof(struct rte_flow_item_udp));
+ if (data & FLOW_ITEM_MASK(RTE_FLOW_ITEM_TYPE_VXLAN))
+ udp_item.hdr.dst_port = RTE_BE16(RTE_VXLAN_DEFAULT_PORT);
+ if (data & FLOW_ITEM_MASK(RTE_FLOW_ITEM_TYPE_VXLAN_GPE))
+ udp_item.hdr.dst_port = RTE_BE16(RTE_VXLAN_GPE_UDP_PORT);
+ if (data & FLOW_ITEM_MASK(RTE_FLOW_ITEM_TYPE_GENEVE))
+ udp_item.hdr.dst_port = RTE_BE16(RTE_GENEVE_UDP_PORT);
+ if (data & FLOW_ITEM_MASK(RTE_FLOW_ITEM_TYPE_GTP))
+ udp_item.hdr.dst_port = RTE_BE16(RTE_GTPU_UDP_PORT);
+ memcpy(*header, &udp_item, sizeof(udp_item));
+ *header += sizeof(udp_item);
+}
+
+static void
+add_vxlan_header(uint8_t **header, uint64_t data,
+ struct additional_para para)
+{
+ struct rte_flow_item_vxlan vxlan_item;
+ uint32_t vni_value = para.counter;
+ uint8_t i;
+
+ if (!(data & FLOW_ITEM_MASK(RTE_FLOW_ITEM_TYPE_VXLAN)))
+ return;
+
+ /* Fixed value */
+ if (FIXED_VALUES)
+ vni_value = 1;
+
+ memset(&vxlan_item, 0, sizeof(struct rte_flow_item_vxlan));
+
+ for (i = 0; i < 3; i++)
+ vxlan_item.vni[2 - i] = vni_value >> (i * 8);
+ vxlan_item.flags = 0x8;
+
+ memcpy(*header, &vxlan_item, sizeof(vxlan_item));
+ *header += sizeof(vxlan_item);
+}
+
+static void
+add_vxlan_gpe_header(uint8_t **header, uint64_t data,
+ struct additional_para para)
+{
+ struct rte_flow_item_vxlan_gpe vxlan_gpe_item;
+ uint32_t vni_value = para.counter;
+ uint8_t i;
+
+ if (!(data & FLOW_ITEM_MASK(RTE_FLOW_ITEM_TYPE_VXLAN_GPE)))
+ return;
+
+ /* Fixed value */
+ if (FIXED_VALUES)
+ vni_value = 1;
+
+ memset(&vxlan_gpe_item, 0, sizeof(struct rte_flow_item_vxlan_gpe));
+
+ for (i = 0; i < 3; i++)
+ vxlan_gpe_item.vni[2 - i] = vni_value >> (i * 8);
+ vxlan_gpe_item.flags = 0x0c;
+
+ memcpy(*header, &vxlan_gpe_item, sizeof(vxlan_gpe_item));
+ *header += sizeof(vxlan_gpe_item);
+}
+
+static void
+add_gre_header(uint8_t **header, uint64_t data,
+ __rte_unused struct additional_para para)
+{
+ struct rte_flow_item_gre gre_item;
+
+ if (!(data & FLOW_ITEM_MASK(RTE_FLOW_ITEM_TYPE_GRE)))
+ return;
+
+ memset(&gre_item, 0, sizeof(struct rte_flow_item_gre));
+
+ gre_item.protocol = RTE_BE16(RTE_ETHER_TYPE_TEB);
+
+ memcpy(*header, &gre_item, sizeof(gre_item));
+ *header += sizeof(gre_item);
+}
+
+static void
+add_geneve_header(uint8_t **header, uint64_t data,
+ struct additional_para para)
+{
+ struct rte_flow_item_geneve geneve_item;
+ uint32_t vni_value = para.counter;
+ uint8_t i;
+
+ if (!(data & FLOW_ITEM_MASK(RTE_FLOW_ITEM_TYPE_GENEVE)))
+ return;
+
+ /* Fixed value */
+ if (FIXED_VALUES)
+ vni_value = 1;
+
+ memset(&geneve_item, 0, sizeof(struct rte_flow_item_geneve));
+
+ for (i = 0; i < 3; i++)
+ geneve_item.vni[2 - i] = vni_value >> (i * 8);
+
+ memcpy(*header, &geneve_item, sizeof(geneve_item));
+ *header += sizeof(geneve_item);
+}
+
+static void
+add_gtp_header(uint8_t **header, uint64_t data,
+ struct additional_para para)
+{
+ struct rte_flow_item_gtp gtp_item;
+ uint32_t teid_value = para.counter;
+
+ if (!(data & FLOW_ITEM_MASK(RTE_FLOW_ITEM_TYPE_GTP)))
+ return;
+
+ /* Fixed value */
+ if (FIXED_VALUES)
+ teid_value = 1;
+
+ memset(>p_item, 0, sizeof(struct rte_flow_item_gtp));
+
+ gtp_item.teid = RTE_BE32(teid_value);
+ gtp_item.msg_type = 255;
+
+ memcpy(*header, >p_item, sizeof(gtp_item));
+ *header += sizeof(gtp_item);
+}
+
+static const struct encap_decap_headers {
+ void (*funct)(
+ uint8_t **header,
+ uint64_t data,
+ struct additional_para para
+ );
+} headers[] = {
+ {.funct = add_ether_header},
+ {.funct = add_vlan_header},
+ {.funct = add_ipv4_header},
+ {.funct = add_ipv6_header},
+ {.funct = add_udp_header},
+ {.funct = add_vxlan_header},
+ {.funct = add_vxlan_gpe_header},
+ {.funct = add_gre_header},
+ {.funct = add_geneve_header},
+ {.funct = add_gtp_header},
+};
+
+static void
+add_raw_encap(struct rte_flow_action *actions,
+ uint8_t actions_counter,
+ struct additional_para para)
+{
+ static struct action_raw_encap_data *action_encap_data;
+ uint64_t encap_data = para.encap_data;
+ uint8_t *header;
+ uint8_t i;
+
+ /* Avoid double allocation. */
+ if (action_encap_data == NULL)
+ action_encap_data = rte_malloc("encap_data",
+ sizeof(struct action_raw_encap_data), 0);
+
+ /* Check if allocation failed. */
+ if (action_encap_data == NULL)
+ rte_exit(EXIT_FAILURE, "No Memory available!");
+
+ *action_encap_data = (struct action_raw_encap_data) {
+ .conf = (struct rte_flow_action_raw_encap) {
+ .data = action_encap_data->data,
+ },
+ .data = {},
+ };
+ header = action_encap_data->data;
+
+ for (i = 0; i < RTE_DIM(headers); i++)
+ headers[i].funct(&header, encap_data, para);
+
+ action_encap_data->conf.size = header -
+ action_encap_data->data;
+
+ actions[actions_counter].type = RTE_FLOW_ACTION_TYPE_RAW_ENCAP;
+ actions[actions_counter].conf = &action_encap_data->conf;
+}
+
+static void
+add_raw_decap(struct rte_flow_action *actions,
+ uint8_t actions_counter,
+ struct additional_para para)
+{
+ static struct action_raw_decap_data *action_decap_data;
+ uint64_t decap_data = para.decap_data;
+ uint8_t *header;
+ uint8_t i;
+
+ /* Avoid double allocation. */
+ if (action_decap_data == NULL)
+ action_decap_data = rte_malloc("decap_data",
+ sizeof(struct action_raw_decap_data), 0);
+
+ /* Check if allocation failed. */
+ if (action_decap_data == NULL)
+ rte_exit(EXIT_FAILURE, "No Memory available!");
+
+ *action_decap_data = (struct action_raw_decap_data) {
+ .conf = (struct rte_flow_action_raw_decap) {
+ .data = action_decap_data->data,
+ },
+ .data = {},
+ };
+ header = action_decap_data->data;
+
+ for (i = 0; i < RTE_DIM(headers); i++)
+ headers[i].funct(&header, decap_data, para);
+
+ action_decap_data->conf.size = header -
+ action_decap_data->data;
+
+ actions[actions_counter].type = RTE_FLOW_ACTION_TYPE_RAW_DECAP;
+ actions[actions_counter].conf = &action_decap_data->conf;
+}
+
+static void
+add_vxlan_encap(struct rte_flow_action *actions,
+ uint8_t actions_counter,
+ __rte_unused struct additional_para para)
+{
+ static struct rte_flow_action_vxlan_encap vxlan_encap;
+ static struct rte_flow_item items[5];
+ static struct rte_flow_item_eth item_eth;
+ static struct rte_flow_item_ipv4 item_ipv4;
+ static struct rte_flow_item_udp item_udp;
+ static struct rte_flow_item_vxlan item_vxlan;
+ uint32_t ip_dst = para.counter;
+
+ /* Fixed value */
+ if (FIXED_VALUES)
+ ip_dst = 1;
+
+ items[0].spec = &item_eth;
+ items[0].mask = &item_eth;
+ items[0].type = RTE_FLOW_ITEM_TYPE_ETH;
+
+ item_ipv4.hdr.src_addr = RTE_IPV4(127, 0, 0, 1);
+ item_ipv4.hdr.dst_addr = RTE_BE32(ip_dst);
+ item_ipv4.hdr.version_ihl = RTE_IPV4_VHL_DEF;
+ items[1].spec = &item_ipv4;
+ items[1].mask = &item_ipv4;
+ items[1].type = RTE_FLOW_ITEM_TYPE_IPV4;
+
+
+ item_udp.hdr.dst_port = RTE_BE16(RTE_VXLAN_DEFAULT_PORT);
+ items[2].spec = &item_udp;
+ items[2].mask = &item_udp;
+ items[2].type = RTE_FLOW_ITEM_TYPE_UDP;
+
+
+ item_vxlan.vni[2] = 1;
+ items[3].spec = &item_vxlan;
+ items[3].mask = &item_vxlan;
+ items[3].type = RTE_FLOW_ITEM_TYPE_VXLAN;
+
+ items[4].type = RTE_FLOW_ITEM_TYPE_END;
+
+ vxlan_encap.definition = items;
+
+ actions[actions_counter].type = RTE_FLOW_ACTION_TYPE_VXLAN_ENCAP;
+ actions[actions_counter].conf = &vxlan_encap;
+}
+
+static void
+add_vxlan_decap(struct rte_flow_action *actions,
+ uint8_t actions_counter,
+ __rte_unused struct additional_para para)
+{
+ actions[actions_counter].type = RTE_FLOW_ACTION_TYPE_VXLAN_DECAP;
+}
+
void
fill_actions(struct rte_flow_action *actions, uint64_t *flow_actions,
- uint32_t counter, uint16_t next_table, uint16_t hairpinq)
+ uint32_t counter, uint16_t next_table, uint16_t hairpinq,
+ uint64_t encap_data, uint64_t decap_data)
{
struct additional_para additional_para_data;
uint8_t actions_counter = 0;
.next_table = next_table,
.queues = queues,
.queues_number = RXQ_NUM,
+ .counter = counter,
+ .encap_data = encap_data,
+ .decap_data = decap_data,
};
if (hairpinq != 0) {
.mask = FLOW_ACTION_MASK(RTE_FLOW_ACTION_TYPE_SET_TAG),
.funct = add_set_tag,
},
+ {
+ .mask = FLOW_ACTION_MASK(
+ RTE_FLOW_ACTION_TYPE_FLAG
+ ),
+ .funct = add_flag,
+ },
+ {
+ .mask = FLOW_ACTION_MASK(
+ RTE_FLOW_ACTION_TYPE_SET_MAC_SRC
+ ),
+ .funct = add_set_src_mac,
+ },
+ {
+ .mask = FLOW_ACTION_MASK(
+ RTE_FLOW_ACTION_TYPE_SET_MAC_DST
+ ),
+ .funct = add_set_dst_mac,
+ },
+ {
+ .mask = FLOW_ACTION_MASK(
+ RTE_FLOW_ACTION_TYPE_SET_IPV4_SRC
+ ),
+ .funct = add_set_src_ipv4,
+ },
+ {
+ .mask = FLOW_ACTION_MASK(
+ RTE_FLOW_ACTION_TYPE_SET_IPV4_DST
+ ),
+ .funct = add_set_dst_ipv4,
+ },
+ {
+ .mask = FLOW_ACTION_MASK(
+ RTE_FLOW_ACTION_TYPE_SET_IPV6_SRC
+ ),
+ .funct = add_set_src_ipv6,
+ },
+ {
+ .mask = FLOW_ACTION_MASK(
+ RTE_FLOW_ACTION_TYPE_SET_IPV6_DST
+ ),
+ .funct = add_set_dst_ipv6,
+ },
+ {
+ .mask = FLOW_ACTION_MASK(
+ RTE_FLOW_ACTION_TYPE_SET_TP_SRC
+ ),
+ .funct = add_set_src_tp,
+ },
+ {
+ .mask = FLOW_ACTION_MASK(
+ RTE_FLOW_ACTION_TYPE_SET_TP_DST
+ ),
+ .funct = add_set_dst_tp,
+ },
+ {
+ .mask = FLOW_ACTION_MASK(
+ RTE_FLOW_ACTION_TYPE_INC_TCP_ACK
+ ),
+ .funct = add_inc_tcp_ack,
+ },
+ {
+ .mask = FLOW_ACTION_MASK(
+ RTE_FLOW_ACTION_TYPE_DEC_TCP_ACK
+ ),
+ .funct = add_dec_tcp_ack,
+ },
+ {
+ .mask = FLOW_ACTION_MASK(
+ RTE_FLOW_ACTION_TYPE_INC_TCP_SEQ
+ ),
+ .funct = add_inc_tcp_seq,
+ },
+ {
+ .mask = FLOW_ACTION_MASK(
+ RTE_FLOW_ACTION_TYPE_DEC_TCP_SEQ
+ ),
+ .funct = add_dec_tcp_seq,
+ },
+ {
+ .mask = FLOW_ACTION_MASK(
+ RTE_FLOW_ACTION_TYPE_SET_TTL
+ ),
+ .funct = add_set_ttl,
+ },
+ {
+ .mask = FLOW_ACTION_MASK(
+ RTE_FLOW_ACTION_TYPE_DEC_TTL
+ ),
+ .funct = add_dec_ttl,
+ },
+ {
+ .mask = FLOW_ACTION_MASK(
+ RTE_FLOW_ACTION_TYPE_SET_IPV4_DSCP
+ ),
+ .funct = add_set_ipv4_dscp,
+ },
+ {
+ .mask = FLOW_ACTION_MASK(
+ RTE_FLOW_ACTION_TYPE_SET_IPV6_DSCP
+ ),
+ .funct = add_set_ipv6_dscp,
+ },
{
.mask = FLOW_ACTION_MASK(RTE_FLOW_ACTION_TYPE_QUEUE),
.funct = add_queue,
.mask = HAIRPIN_RSS_ACTION,
.funct = add_rss,
},
+ {
+ .mask = FLOW_ACTION_MASK(
+ RTE_FLOW_ACTION_TYPE_RAW_ENCAP
+ ),
+ .funct = add_raw_encap,
+ },
+ {
+ .mask = FLOW_ACTION_MASK(
+ RTE_FLOW_ACTION_TYPE_RAW_DECAP
+ ),
+ .funct = add_raw_decap,
+ },
+ {
+ .mask = FLOW_ACTION_MASK(
+ RTE_FLOW_ACTION_TYPE_VXLAN_ENCAP
+ ),
+ .funct = add_vxlan_encap,
+ },
+ {
+ .mask = FLOW_ACTION_MASK(
+ RTE_FLOW_ACTION_TYPE_VXLAN_DECAP
+ ),
+ .funct = add_vxlan_decap,
+ },
};
for (j = 0; j < MAX_ACTIONS_NUM; j++) {