net/iavf: simplify flow director rules for IP fragment
[dpdk.git] / drivers / net / iavf / iavf_fdir.c
index d683a46..e81416d 100644 (file)
@@ -11,7 +11,7 @@
 #include <stdarg.h>
 
 #include <rte_ether.h>
-#include <rte_ethdev_driver.h>
+#include <ethdev_driver.h>
 #include <rte_malloc.h>
 #include <rte_tailq.h>
 
 #define IAVF_FDIR_IPV6_TC_OFFSET 20
 #define IAVF_IPV6_TC_MASK  (0xFF << IAVF_FDIR_IPV6_TC_OFFSET)
 
+#define IAVF_GTPU_EH_DWLINK 0
+#define IAVF_GTPU_EH_UPLINK 1
+
 #define IAVF_FDIR_INSET_ETH (\
        IAVF_INSET_ETHERTYPE)
 
 #define IAVF_FDIR_INSET_ETH_IPV4 (\
        IAVF_INSET_IPV4_SRC | IAVF_INSET_IPV4_DST | \
        IAVF_INSET_IPV4_PROTO | IAVF_INSET_IPV4_TOS | \
-       IAVF_INSET_IPV4_TTL)
+       IAVF_INSET_IPV4_TTL | IAVF_INSET_IPV4_ID)
 
 #define IAVF_FDIR_INSET_ETH_IPV4_UDP (\
        IAVF_INSET_IPV4_SRC | IAVF_INSET_IPV4_DST | \
@@ -53,6 +56,9 @@
        IAVF_INSET_IPV6_NEXT_HDR | IAVF_INSET_IPV6_TC | \
        IAVF_INSET_IPV6_HOP_LIMIT)
 
+#define IAVF_FDIR_INSET_ETH_IPV6_FRAG_EXT (\
+       IAVF_INSET_IPV6_ID)
+
 #define IAVF_FDIR_INSET_ETH_IPV6_UDP (\
        IAVF_INSET_IPV6_SRC | IAVF_INSET_IPV6_DST | \
        IAVF_INSET_IPV6_TC | IAVF_INSET_IPV6_HOP_LIMIT | \
        IAVF_INSET_IPV4_SRC | IAVF_INSET_IPV4_DST | \
        IAVF_INSET_GTPU_TEID)
 
+#define IAVF_FDIR_INSET_GTPU_IPV4 (\
+       IAVF_INSET_TUN_IPV4_SRC | IAVF_INSET_TUN_IPV4_DST | \
+       IAVF_INSET_TUN_IPV4_PROTO | IAVF_INSET_TUN_IPV4_TOS | \
+       IAVF_INSET_TUN_IPV4_TTL)
+
+#define IAVF_FDIR_INSET_GTPU_IPV4_UDP (\
+       IAVF_FDIR_INSET_GTPU_IPV4 | \
+       IAVF_INSET_TUN_UDP_SRC_PORT | IAVF_INSET_TUN_UDP_DST_PORT)
+
+#define IAVF_FDIR_INSET_GTPU_IPV4_TCP (\
+       IAVF_FDIR_INSET_GTPU_IPV4 | \
+       IAVF_INSET_TUN_TCP_SRC_PORT | IAVF_INSET_TUN_TCP_DST_PORT)
+
 #define IAVF_FDIR_INSET_IPV4_GTPU_EH (\
        IAVF_INSET_IPV4_SRC | IAVF_INSET_IPV4_DST | \
        IAVF_INSET_GTPU_TEID | IAVF_INSET_GTPU_QFI)
        IAVF_INSET_IPV6_SRC | IAVF_INSET_IPV6_DST | \
        IAVF_INSET_GTPU_TEID)
 
+#define IAVF_FDIR_INSET_GTPU_IPV6 (\
+       IAVF_INSET_TUN_IPV6_SRC | IAVF_INSET_TUN_IPV6_DST | \
+       IAVF_INSET_TUN_IPV6_NEXT_HDR | IAVF_INSET_TUN_IPV6_TC | \
+       IAVF_INSET_TUN_IPV6_HOP_LIMIT)
+
+#define IAVF_FDIR_INSET_GTPU_IPV6_UDP (\
+       IAVF_FDIR_INSET_GTPU_IPV6 | \
+       IAVF_INSET_TUN_UDP_SRC_PORT | IAVF_INSET_TUN_UDP_DST_PORT)
+
+#define IAVF_FDIR_INSET_GTPU_IPV6_TCP (\
+       IAVF_FDIR_INSET_GTPU_IPV6 | \
+       IAVF_INSET_TUN_TCP_SRC_PORT | IAVF_INSET_TUN_TCP_DST_PORT)
+
 #define IAVF_FDIR_INSET_IPV6_GTPU_EH (\
        IAVF_INSET_IPV6_SRC | IAVF_INSET_IPV6_DST | \
        IAVF_INSET_GTPU_TEID | IAVF_INSET_GTPU_QFI)
 #define IAVF_FDIR_INSET_L2TPV3OIP (\
        IAVF_L2TPV3OIP_SESSION_ID)
 
-#define IAVF_FDIR_INSET_ESP (\
+#define IAVF_FDIR_INSET_IPV4_ESP (\
+       IAVF_INSET_IPV4_SRC | IAVF_INSET_IPV4_DST | \
+       IAVF_INSET_ESP_SPI)
+
+#define IAVF_FDIR_INSET_IPV6_ESP (\
+       IAVF_INSET_IPV6_SRC | IAVF_INSET_IPV6_DST | \
        IAVF_INSET_ESP_SPI)
 
 #define IAVF_FDIR_INSET_AH (\
 #define IAVF_FDIR_INSET_PFCP (\
        IAVF_INSET_PFCP_S_FIELD)
 
+#define IAVF_FDIR_INSET_ECPRI (\
+       IAVF_INSET_ECPRI)
+
+#define IAVF_FDIR_INSET_GRE_IPV4 (\
+       IAVF_INSET_TUN_IPV4_SRC | IAVF_INSET_TUN_IPV4_DST | \
+       IAVF_INSET_TUN_IPV4_TOS | IAVF_INSET_TUN_IPV4_PROTO)
+
+#define IAVF_FDIR_INSET_GRE_IPV4_TCP (\
+       IAVF_FDIR_INSET_GRE_IPV4 | IAVF_INSET_TUN_TCP_SRC_PORT | \
+       IAVF_INSET_TUN_TCP_DST_PORT)
+
+#define IAVF_FDIR_INSET_GRE_IPV4_UDP (\
+       IAVF_FDIR_INSET_GRE_IPV4 | IAVF_INSET_TUN_UDP_SRC_PORT | \
+       IAVF_INSET_TUN_UDP_DST_PORT)
+
+#define IAVF_FDIR_INSET_GRE_IPV6 (\
+       IAVF_INSET_TUN_IPV6_SRC | IAVF_INSET_TUN_IPV6_DST | \
+       IAVF_INSET_TUN_IPV6_TC | IAVF_INSET_TUN_IPV6_NEXT_HDR)
+
+#define IAVF_FDIR_INSET_GRE_IPV6_TCP (\
+       IAVF_FDIR_INSET_GRE_IPV6 | IAVF_INSET_TUN_TCP_SRC_PORT | \
+       IAVF_INSET_TUN_TCP_DST_PORT)
+
+#define IAVF_FDIR_INSET_GRE_IPV6_UDP (\
+       IAVF_FDIR_INSET_GRE_IPV6 | IAVF_INSET_TUN_UDP_SRC_PORT | \
+       IAVF_INSET_TUN_UDP_DST_PORT)
+
 static struct iavf_pattern_match_item iavf_fdir_pattern[] = {
-       {iavf_pattern_ethertype,                IAVF_FDIR_INSET_ETH,                    IAVF_INSET_NONE},
-       {iavf_pattern_eth_ipv4,                 IAVF_FDIR_INSET_ETH_IPV4,               IAVF_INSET_NONE},
-       {iavf_pattern_eth_ipv4_udp,             IAVF_FDIR_INSET_ETH_IPV4_UDP,           IAVF_INSET_NONE},
-       {iavf_pattern_eth_ipv4_tcp,             IAVF_FDIR_INSET_ETH_IPV4_TCP,           IAVF_INSET_NONE},
-       {iavf_pattern_eth_ipv4_sctp,            IAVF_FDIR_INSET_ETH_IPV4_SCTP,          IAVF_INSET_NONE},
-       {iavf_pattern_eth_ipv6,                 IAVF_FDIR_INSET_ETH_IPV6,               IAVF_INSET_NONE},
-       {iavf_pattern_eth_ipv6_udp,             IAVF_FDIR_INSET_ETH_IPV6_UDP,           IAVF_INSET_NONE},
-       {iavf_pattern_eth_ipv6_tcp,             IAVF_FDIR_INSET_ETH_IPV6_TCP,           IAVF_INSET_NONE},
-       {iavf_pattern_eth_ipv6_sctp,            IAVF_FDIR_INSET_ETH_IPV6_SCTP,          IAVF_INSET_NONE},
-       {iavf_pattern_eth_ipv4_gtpu,            IAVF_FDIR_INSET_IPV4_GTPU,              IAVF_INSET_NONE},
-       {iavf_pattern_eth_ipv4_gtpu_eh,         IAVF_FDIR_INSET_IPV4_GTPU_EH,           IAVF_INSET_NONE},
-       {iavf_pattern_eth_ipv6_gtpu,            IAVF_FDIR_INSET_IPV6_GTPU,              IAVF_INSET_NONE},
-       {iavf_pattern_eth_ipv6_gtpu_eh,         IAVF_FDIR_INSET_IPV6_GTPU_EH,           IAVF_INSET_NONE},
-       {iavf_pattern_eth_ipv4_l2tpv3,          IAVF_FDIR_INSET_L2TPV3OIP,              IAVF_INSET_NONE},
-       {iavf_pattern_eth_ipv6_l2tpv3,          IAVF_FDIR_INSET_L2TPV3OIP,              IAVF_INSET_NONE},
-       {iavf_pattern_eth_ipv4_esp,             IAVF_FDIR_INSET_ESP,                    IAVF_INSET_NONE},
-       {iavf_pattern_eth_ipv6_esp,             IAVF_FDIR_INSET_ESP,                    IAVF_INSET_NONE},
-       {iavf_pattern_eth_ipv4_ah,              IAVF_FDIR_INSET_AH,                     IAVF_INSET_NONE},
-       {iavf_pattern_eth_ipv6_ah,              IAVF_FDIR_INSET_AH,                     IAVF_INSET_NONE},
-       {iavf_pattern_eth_ipv4_udp_esp,         IAVF_FDIR_INSET_IPV4_NATT_ESP,          IAVF_INSET_NONE},
-       {iavf_pattern_eth_ipv6_udp_esp,         IAVF_FDIR_INSET_IPV6_NATT_ESP,          IAVF_INSET_NONE},
-       {iavf_pattern_eth_ipv4_pfcp,            IAVF_FDIR_INSET_PFCP,                   IAVF_INSET_NONE},
-       {iavf_pattern_eth_ipv6_pfcp,            IAVF_FDIR_INSET_PFCP,                   IAVF_INSET_NONE},
+       {iavf_pattern_ethertype,                 IAVF_FDIR_INSET_ETH,           IAVF_INSET_NONE},
+       {iavf_pattern_eth_ipv4,                  IAVF_FDIR_INSET_ETH_IPV4,      IAVF_INSET_NONE},
+       {iavf_pattern_eth_ipv4_udp,              IAVF_FDIR_INSET_ETH_IPV4_UDP,  IAVF_INSET_NONE},
+       {iavf_pattern_eth_ipv4_tcp,              IAVF_FDIR_INSET_ETH_IPV4_TCP,  IAVF_INSET_NONE},
+       {iavf_pattern_eth_ipv4_sctp,             IAVF_FDIR_INSET_ETH_IPV4_SCTP, IAVF_INSET_NONE},
+       {iavf_pattern_eth_ipv6,                  IAVF_FDIR_INSET_ETH_IPV6,      IAVF_INSET_NONE},
+       {iavf_pattern_eth_ipv6_frag_ext,        IAVF_FDIR_INSET_ETH_IPV6_FRAG_EXT,      IAVF_INSET_NONE},
+       {iavf_pattern_eth_ipv6_udp,              IAVF_FDIR_INSET_ETH_IPV6_UDP,  IAVF_INSET_NONE},
+       {iavf_pattern_eth_ipv6_tcp,              IAVF_FDIR_INSET_ETH_IPV6_TCP,  IAVF_INSET_NONE},
+       {iavf_pattern_eth_ipv6_sctp,             IAVF_FDIR_INSET_ETH_IPV6_SCTP, IAVF_INSET_NONE},
+       {iavf_pattern_eth_ipv4_gtpu,             IAVF_FDIR_INSET_IPV4_GTPU,     IAVF_INSET_NONE},
+       {iavf_pattern_eth_ipv4_gtpu_ipv4,        IAVF_FDIR_INSET_GTPU_IPV4,     IAVF_INSET_NONE},
+       {iavf_pattern_eth_ipv4_gtpu_ipv4_udp,    IAVF_FDIR_INSET_GTPU_IPV4_UDP, IAVF_INSET_NONE},
+       {iavf_pattern_eth_ipv4_gtpu_ipv4_tcp,    IAVF_FDIR_INSET_GTPU_IPV4_TCP, IAVF_INSET_NONE},
+       {iavf_pattern_eth_ipv4_gtpu_ipv6,        IAVF_FDIR_INSET_GTPU_IPV6,     IAVF_INSET_NONE},
+       {iavf_pattern_eth_ipv4_gtpu_ipv6_udp,    IAVF_FDIR_INSET_GTPU_IPV6_UDP, IAVF_INSET_NONE},
+       {iavf_pattern_eth_ipv4_gtpu_ipv6_tcp,    IAVF_FDIR_INSET_GTPU_IPV6_TCP, IAVF_INSET_NONE},
+       {iavf_pattern_eth_ipv4_gtpu_eh,          IAVF_FDIR_INSET_IPV4_GTPU_EH,  IAVF_INSET_NONE},
+       {iavf_pattern_eth_ipv4_gtpu_eh_ipv4,     IAVF_FDIR_INSET_GTPU_IPV4,     IAVF_INSET_NONE},
+       {iavf_pattern_eth_ipv4_gtpu_eh_ipv4_udp, IAVF_FDIR_INSET_GTPU_IPV4_UDP, IAVF_INSET_NONE},
+       {iavf_pattern_eth_ipv4_gtpu_eh_ipv4_tcp, IAVF_FDIR_INSET_GTPU_IPV4_TCP, IAVF_INSET_NONE},
+       {iavf_pattern_eth_ipv4_gtpu_eh_ipv6,     IAVF_FDIR_INSET_GTPU_IPV6,     IAVF_INSET_NONE},
+       {iavf_pattern_eth_ipv4_gtpu_eh_ipv6_udp, IAVF_FDIR_INSET_GTPU_IPV6_UDP, IAVF_INSET_NONE},
+       {iavf_pattern_eth_ipv4_gtpu_eh_ipv6_tcp, IAVF_FDIR_INSET_GTPU_IPV6_TCP, IAVF_INSET_NONE},
+       {iavf_pattern_eth_ipv6_gtpu,             IAVF_FDIR_INSET_IPV6_GTPU,     IAVF_INSET_NONE},
+       {iavf_pattern_eth_ipv6_gtpu_eh,          IAVF_FDIR_INSET_IPV6_GTPU_EH,  IAVF_INSET_NONE},
+       {iavf_pattern_eth_ipv4_l2tpv3,           IAVF_FDIR_INSET_L2TPV3OIP,     IAVF_INSET_NONE},
+       {iavf_pattern_eth_ipv6_l2tpv3,           IAVF_FDIR_INSET_L2TPV3OIP,     IAVF_INSET_NONE},
+       {iavf_pattern_eth_ipv4_esp,              IAVF_FDIR_INSET_IPV4_ESP,      IAVF_INSET_NONE},
+       {iavf_pattern_eth_ipv6_esp,              IAVF_FDIR_INSET_IPV6_ESP,      IAVF_INSET_NONE},
+       {iavf_pattern_eth_ipv4_ah,               IAVF_FDIR_INSET_AH,            IAVF_INSET_NONE},
+       {iavf_pattern_eth_ipv6_ah,               IAVF_FDIR_INSET_AH,            IAVF_INSET_NONE},
+       {iavf_pattern_eth_ipv4_udp_esp,          IAVF_FDIR_INSET_IPV4_NATT_ESP, IAVF_INSET_NONE},
+       {iavf_pattern_eth_ipv6_udp_esp,          IAVF_FDIR_INSET_IPV6_NATT_ESP, IAVF_INSET_NONE},
+       {iavf_pattern_eth_ipv4_pfcp,             IAVF_FDIR_INSET_PFCP,          IAVF_INSET_NONE},
+       {iavf_pattern_eth_ipv6_pfcp,             IAVF_FDIR_INSET_PFCP,          IAVF_INSET_NONE},
+       {iavf_pattern_eth_ecpri,                 IAVF_FDIR_INSET_ECPRI,         IAVF_INSET_NONE},
+       {iavf_pattern_eth_ipv4_ecpri,            IAVF_FDIR_INSET_ECPRI,         IAVF_INSET_NONE},
+       {iavf_pattern_eth_ipv4_gre_ipv4,        IAVF_FDIR_INSET_GRE_IPV4,       IAVF_INSET_NONE},
+       {iavf_pattern_eth_ipv4_gre_ipv4_tcp,    IAVF_FDIR_INSET_GRE_IPV4_TCP,   IAVF_INSET_NONE},
+       {iavf_pattern_eth_ipv4_gre_ipv4_udp,    IAVF_FDIR_INSET_GRE_IPV4_UDP,   IAVF_INSET_NONE},
+       {iavf_pattern_eth_ipv4_gre_ipv6,        IAVF_FDIR_INSET_GRE_IPV6,       IAVF_INSET_NONE},
+       {iavf_pattern_eth_ipv4_gre_ipv6_tcp,    IAVF_FDIR_INSET_GRE_IPV6_TCP,   IAVF_INSET_NONE},
+       {iavf_pattern_eth_ipv4_gre_ipv6_udp,    IAVF_FDIR_INSET_GRE_IPV6_UDP,   IAVF_INSET_NONE},
+       {iavf_pattern_eth_ipv6_gre_ipv4,        IAVF_FDIR_INSET_GRE_IPV4,       IAVF_INSET_NONE},
+       {iavf_pattern_eth_ipv6_gre_ipv4_tcp,    IAVF_FDIR_INSET_GRE_IPV4_TCP,   IAVF_INSET_NONE},
+       {iavf_pattern_eth_ipv6_gre_ipv4_udp,    IAVF_FDIR_INSET_GRE_IPV4_UDP,   IAVF_INSET_NONE},
+       {iavf_pattern_eth_ipv6_gre_ipv6,        IAVF_FDIR_INSET_GRE_IPV6,       IAVF_INSET_NONE},
+       {iavf_pattern_eth_ipv6_gre_ipv6_tcp,    IAVF_FDIR_INSET_GRE_IPV6_TCP,   IAVF_INSET_NONE},
+       {iavf_pattern_eth_ipv6_gre_ipv6_udp,    IAVF_FDIR_INSET_GRE_IPV6_UDP,   IAVF_INSET_NONE},
 };
 
 static struct iavf_flow_parser iavf_fdir_parser;
@@ -448,18 +539,93 @@ iavf_fdir_parse_action(struct iavf_adapter *ad,
        return 0;
 }
 
+static bool
+iavf_fdir_refine_input_set(const uint64_t input_set,
+                          const uint64_t input_set_mask,
+                          struct iavf_fdir_conf *filter)
+{
+       struct virtchnl_proto_hdr *hdr, *hdr_last;
+       struct rte_flow_item_ipv4 ipv4_spec;
+       struct rte_flow_item_ipv6 ipv6_spec;
+       int last_layer;
+       uint8_t proto_id;
+
+       if (input_set & ~input_set_mask)
+               return false;
+       else if (input_set)
+               return true;
+
+       last_layer = filter->add_fltr.rule_cfg.proto_hdrs.count - 1;
+       /* Last layer of TCP/UDP pattern isn't less than 2. */
+       if (last_layer < 2)
+               return false;
+       hdr_last = &filter->add_fltr.rule_cfg.proto_hdrs.proto_hdr[last_layer];
+       if (hdr_last->type == VIRTCHNL_PROTO_HDR_TCP)
+               proto_id = 6;
+       else if (hdr_last->type == VIRTCHNL_PROTO_HDR_UDP)
+               proto_id = 17;
+       else
+               return false;
+
+       hdr = &filter->add_fltr.rule_cfg.proto_hdrs.proto_hdr[last_layer - 1];
+       switch (hdr->type) {
+       case VIRTCHNL_PROTO_HDR_IPV4:
+               VIRTCHNL_ADD_PROTO_HDR_FIELD_BIT(hdr, IPV4, PROT);
+               memset(&ipv4_spec, 0, sizeof(ipv4_spec));
+               ipv4_spec.hdr.next_proto_id = proto_id;
+               rte_memcpy(hdr->buffer, &ipv4_spec.hdr,
+                          sizeof(ipv4_spec.hdr));
+               return true;
+       case VIRTCHNL_PROTO_HDR_IPV6:
+               VIRTCHNL_ADD_PROTO_HDR_FIELD_BIT(hdr, IPV6, PROT);
+               memset(&ipv6_spec, 0, sizeof(ipv6_spec));
+               ipv6_spec.hdr.proto = proto_id;
+               rte_memcpy(hdr->buffer, &ipv6_spec.hdr,
+                          sizeof(ipv6_spec.hdr));
+               return true;
+       default:
+               return false;
+       }
+}
+
+static void
+iavf_fdir_add_fragment_hdr(struct virtchnl_proto_hdrs *hdrs, int layer)
+{
+       struct virtchnl_proto_hdr *hdr1;
+       struct virtchnl_proto_hdr *hdr2;
+       int i;
+
+       if (layer < 0 || layer > hdrs->count)
+               return;
+
+       /* shift headers layer */
+       for (i = hdrs->count; i >= layer; i--) {
+               hdr1 = &hdrs->proto_hdr[i];
+               hdr2 = &hdrs->proto_hdr[i - 1];
+               *hdr1 = *hdr2;
+       }
+
+       /* adding dummy fragment header */
+       hdr1 = &hdrs->proto_hdr[layer];
+       VIRTCHNL_SET_PROTO_HDR_TYPE(hdr1, IPV4_FRAG);
+       hdrs->count = ++layer;
+}
+
 static int
 iavf_fdir_parse_pattern(__rte_unused struct iavf_adapter *ad,
                        const struct rte_flow_item pattern[],
+                       const uint64_t input_set_mask,
                        struct rte_flow_error *error,
                        struct iavf_fdir_conf *filter)
 {
-       const struct rte_flow_item *item = pattern;
-       enum rte_flow_item_type item_type;
+       struct virtchnl_proto_hdrs *hdrs =
+                       &filter->add_fltr.rule_cfg.proto_hdrs;
        enum rte_flow_item_type l3 = RTE_FLOW_ITEM_TYPE_END;
        const struct rte_flow_item_eth *eth_spec, *eth_mask;
-       const struct rte_flow_item_ipv4 *ipv4_spec, *ipv4_mask;
+       const struct rte_flow_item_ipv4 *ipv4_spec, *ipv4_last, *ipv4_mask;
        const struct rte_flow_item_ipv6 *ipv6_spec, *ipv6_mask;
+       const struct rte_flow_item_ipv6_frag_ext *ipv6_frag_spec;
+       const struct rte_flow_item_ipv6_frag_ext *ipv6_frag_mask;
        const struct rte_flow_item_udp *udp_spec, *udp_mask;
        const struct rte_flow_item_tcp *tcp_spec, *tcp_mask;
        const struct rte_flow_item_sctp *sctp_spec, *sctp_mask;
@@ -469,13 +635,17 @@ iavf_fdir_parse_pattern(__rte_unused struct iavf_adapter *ad,
        const struct rte_flow_item_esp *esp_spec, *esp_mask;
        const struct rte_flow_item_ah *ah_spec, *ah_mask;
        const struct rte_flow_item_pfcp *pfcp_spec, *pfcp_mask;
+       const struct rte_flow_item_ecpri *ecpri_spec, *ecpri_mask;
+       const struct rte_flow_item_gre *gre_spec, *gre_mask;
+       const struct rte_flow_item *item = pattern;
+       struct virtchnl_proto_hdr *hdr, *hdr1 = NULL;
+       struct rte_ecpri_common_hdr ecpri_common;
        uint64_t input_set = IAVF_INSET_NONE;
-
+       enum rte_flow_item_type item_type;
        enum rte_flow_item_type next_type;
+       uint8_t tun_inner = 0;
        uint16_t ether_type;
-
        int layer = 0;
-       struct virtchnl_proto_hdr *hdr;
 
        uint8_t  ipv6_addr_mask[16] = {
                0xFF, 0xFF, 0xFF, 0xFF, 0xFF, 0xFF, 0xFF, 0xFF,
@@ -483,26 +653,28 @@ iavf_fdir_parse_pattern(__rte_unused struct iavf_adapter *ad,
        };
 
        for (item = pattern; item->type != RTE_FLOW_ITEM_TYPE_END; item++) {
-               if (item->last) {
+               item_type = item->type;
+
+               if (item->last && !(item_type == RTE_FLOW_ITEM_TYPE_IPV4 ||
+                                   item_type ==
+                                   RTE_FLOW_ITEM_TYPE_IPV6_FRAG_EXT)) {
                        rte_flow_error_set(error, EINVAL,
-                                       RTE_FLOW_ERROR_TYPE_ITEM, item,
-                                       "Not support range");
+                                          RTE_FLOW_ERROR_TYPE_ITEM, item,
+                                          "Not support range");
                }
 
-               item_type = item->type;
-
                switch (item_type) {
                case RTE_FLOW_ITEM_TYPE_ETH:
                        eth_spec = item->spec;
                        eth_mask = item->mask;
                        next_type = (item + 1)->type;
 
-                       hdr = &filter->add_fltr.rule_cfg.proto_hdrs.proto_hdr[layer];
+                       hdr1 = &hdrs->proto_hdr[layer];
 
-                       VIRTCHNL_SET_PROTO_HDR_TYPE(hdr, ETH);
+                       VIRTCHNL_SET_PROTO_HDR_TYPE(hdr1, ETH);
 
                        if (next_type == RTE_FLOW_ITEM_TYPE_END &&
-                               (!eth_spec || !eth_mask)) {
+                           (!eth_spec || !eth_mask)) {
                                rte_flow_error_set(error, EINVAL,
                                                RTE_FLOW_ERROR_TYPE_ITEM,
                                                item, "NULL eth spec/mask.");
@@ -538,64 +710,121 @@ iavf_fdir_parse_pattern(__rte_unused struct iavf_adapter *ad,
                                }
 
                                input_set |= IAVF_INSET_ETHERTYPE;
-                               VIRTCHNL_ADD_PROTO_HDR_FIELD_BIT(hdr, ETH, ETHERTYPE);
+                               VIRTCHNL_ADD_PROTO_HDR_FIELD_BIT(hdr1, ETH,
+                                                                ETHERTYPE);
 
-                               rte_memcpy(hdr->buffer,
-                                       eth_spec, sizeof(*eth_spec));
+                               rte_memcpy(hdr1->buffer, eth_spec,
+                                          sizeof(struct rte_ether_hdr));
                        }
 
-                       filter->add_fltr.rule_cfg.proto_hdrs.count = ++layer;
+                       hdrs->count = ++layer;
                        break;
 
                case RTE_FLOW_ITEM_TYPE_IPV4:
                        l3 = RTE_FLOW_ITEM_TYPE_IPV4;
                        ipv4_spec = item->spec;
+                       ipv4_last = item->last;
                        ipv4_mask = item->mask;
+                       next_type = (item + 1)->type;
 
-                       hdr = &filter->add_fltr.rule_cfg.proto_hdrs.proto_hdr[layer];
+                       hdr = &hdrs->proto_hdr[layer];
 
                        VIRTCHNL_SET_PROTO_HDR_TYPE(hdr, IPV4);
 
-                       if (ipv4_spec && ipv4_mask) {
-                               if (ipv4_mask->hdr.version_ihl ||
-                                       ipv4_mask->hdr.total_length ||
-                                       ipv4_mask->hdr.packet_id ||
-                                       ipv4_mask->hdr.fragment_offset ||
-                                       ipv4_mask->hdr.hdr_checksum) {
-                                       rte_flow_error_set(error, EINVAL,
-                                               RTE_FLOW_ERROR_TYPE_ITEM,
-                                               item, "Invalid IPv4 mask.");
-                                       return -rte_errno;
-                               }
+                       if (!(ipv4_spec && ipv4_mask)) {
+                               hdrs->count = ++layer;
+                               break;
+                       }
 
-                               if (ipv4_mask->hdr.type_of_service ==
-                                                               UINT8_MAX) {
-                                       input_set |= IAVF_INSET_IPV4_TOS;
-                                       VIRTCHNL_ADD_PROTO_HDR_FIELD_BIT(hdr, IPV4, DSCP);
-                               }
-                               if (ipv4_mask->hdr.next_proto_id == UINT8_MAX) {
-                                       input_set |= IAVF_INSET_IPV4_PROTO;
-                                       VIRTCHNL_ADD_PROTO_HDR_FIELD_BIT(hdr, IPV4, PROT);
-                               }
-                               if (ipv4_mask->hdr.time_to_live == UINT8_MAX) {
-                                       input_set |= IAVF_INSET_IPV4_TTL;
-                                       VIRTCHNL_ADD_PROTO_HDR_FIELD_BIT(hdr, IPV4, TTL);
-                               }
-                               if (ipv4_mask->hdr.src_addr == UINT32_MAX) {
-                                       input_set |= IAVF_INSET_IPV4_SRC;
-                                       VIRTCHNL_ADD_PROTO_HDR_FIELD_BIT(hdr, IPV4, SRC);
-                               }
-                               if (ipv4_mask->hdr.dst_addr == UINT32_MAX) {
-                                       input_set |= IAVF_INSET_IPV4_DST;
-                                       VIRTCHNL_ADD_PROTO_HDR_FIELD_BIT(hdr, IPV4, DST);
-                               }
+                       if (ipv4_mask->hdr.version_ihl ||
+                           ipv4_mask->hdr.total_length ||
+                           ipv4_mask->hdr.hdr_checksum) {
+                               rte_flow_error_set(error, EINVAL,
+                                                  RTE_FLOW_ERROR_TYPE_ITEM,
+                                                  item, "Invalid IPv4 mask.");
+                               return -rte_errno;
+                       }
 
-                               rte_memcpy(hdr->buffer,
-                                       &ipv4_spec->hdr,
-                                       sizeof(ipv4_spec->hdr));
+                       if (ipv4_last &&
+                           (ipv4_last->hdr.version_ihl ||
+                            ipv4_last->hdr.type_of_service ||
+                            ipv4_last->hdr.time_to_live ||
+                            ipv4_last->hdr.total_length |
+                            ipv4_last->hdr.next_proto_id ||
+                            ipv4_last->hdr.hdr_checksum ||
+                            ipv4_last->hdr.src_addr ||
+                            ipv4_last->hdr.dst_addr)) {
+                               rte_flow_error_set(error, EINVAL,
+                                                  RTE_FLOW_ERROR_TYPE_ITEM,
+                                                  item, "Invalid IPv4 last.");
+                               return -rte_errno;
+                       }
+
+                       if (ipv4_mask->hdr.type_of_service ==
+                           UINT8_MAX) {
+                               input_set |= IAVF_INSET_IPV4_TOS;
+                               VIRTCHNL_ADD_PROTO_HDR_FIELD_BIT(hdr, IPV4,
+                                                                DSCP);
+                       }
+
+                       if (ipv4_mask->hdr.next_proto_id == UINT8_MAX) {
+                               input_set |= IAVF_INSET_IPV4_PROTO;
+                               VIRTCHNL_ADD_PROTO_HDR_FIELD_BIT(hdr, IPV4,
+                                                                PROT);
+                       }
+
+                       if (ipv4_mask->hdr.time_to_live == UINT8_MAX) {
+                               input_set |= IAVF_INSET_IPV4_TTL;
+                               VIRTCHNL_ADD_PROTO_HDR_FIELD_BIT(hdr, IPV4,
+                                                                TTL);
+                       }
+
+                       if (ipv4_mask->hdr.src_addr == UINT32_MAX) {
+                               input_set |= IAVF_INSET_IPV4_SRC;
+                               VIRTCHNL_ADD_PROTO_HDR_FIELD_BIT(hdr, IPV4,
+                                                                SRC);
+                       }
+
+                       if (ipv4_mask->hdr.dst_addr == UINT32_MAX) {
+                               input_set |= IAVF_INSET_IPV4_DST;
+                               VIRTCHNL_ADD_PROTO_HDR_FIELD_BIT(hdr, IPV4,
+                                                                DST);
+                       }
+
+                       if (tun_inner) {
+                               input_set &= ~IAVF_PROT_IPV4_OUTER;
+                               input_set |= IAVF_PROT_IPV4_INNER;
+                       }
+
+                       rte_memcpy(hdr->buffer, &ipv4_spec->hdr,
+                                  sizeof(ipv4_spec->hdr));
+
+                       hdrs->count = ++layer;
+
+                       /* fragment Ipv4:
+                        * spec is 0x2000, mask is 0x2000
+                        */
+                       if (ipv4_spec->hdr.fragment_offset ==
+                           rte_cpu_to_be_16(RTE_IPV4_HDR_MF_FLAG) &&
+                           ipv4_mask->hdr.fragment_offset ==
+                           rte_cpu_to_be_16(RTE_IPV4_HDR_MF_FLAG)) {
+                               /* all IPv4 fragment packet has the same
+                                * ethertype, if the spec and mask is valid,
+                                * set ethertype into input set.
+                                */
+                               input_set |= IAVF_INSET_ETHERTYPE;
+                               VIRTCHNL_ADD_PROTO_HDR_FIELD_BIT(hdr1, ETH,
+                                                                ETHERTYPE);
+
+                               /* add dummy header for IPv4 Fragment */
+                               iavf_fdir_add_fragment_hdr(hdrs, layer);
+                       } else if (ipv4_mask->hdr.packet_id == UINT16_MAX) {
+                               rte_flow_error_set(error, EINVAL,
+                                                  RTE_FLOW_ERROR_TYPE_ITEM,
+                                                  item, "Invalid IPv4 mask.");
+                               return -rte_errno;
                        }
 
-                       filter->add_fltr.rule_cfg.proto_hdrs.count = ++layer;
                        break;
 
                case RTE_FLOW_ITEM_TYPE_IPV6:
@@ -603,58 +832,112 @@ iavf_fdir_parse_pattern(__rte_unused struct iavf_adapter *ad,
                        ipv6_spec = item->spec;
                        ipv6_mask = item->mask;
 
-                       hdr = &filter->add_fltr.rule_cfg.proto_hdrs.proto_hdr[layer];
+                       hdr = &hdrs->proto_hdr[layer];
 
                        VIRTCHNL_SET_PROTO_HDR_TYPE(hdr, IPV6);
 
-                       if (ipv6_spec && ipv6_mask) {
-                               if (ipv6_mask->hdr.payload_len) {
-                                       rte_flow_error_set(error, EINVAL,
-                                               RTE_FLOW_ERROR_TYPE_ITEM,
-                                               item, "Invalid IPv6 mask");
-                                       return -rte_errno;
-                               }
+                       if (!(ipv6_spec && ipv6_mask)) {
+                               hdrs->count = ++layer;
+                               break;
+                       }
 
-                               if ((ipv6_mask->hdr.vtc_flow &
-                                       rte_cpu_to_be_32(IAVF_IPV6_TC_MASK))
-                                       == rte_cpu_to_be_32(IAVF_IPV6_TC_MASK)) {
-                                       input_set |= IAVF_INSET_IPV6_TC;
-                                       VIRTCHNL_ADD_PROTO_HDR_FIELD_BIT(hdr, IPV6, TC);
-                               }
-                               if (ipv6_mask->hdr.proto == UINT8_MAX) {
-                                       input_set |= IAVF_INSET_IPV6_NEXT_HDR;
-                                       VIRTCHNL_ADD_PROTO_HDR_FIELD_BIT(hdr, IPV6, PROT);
-                               }
-                               if (ipv6_mask->hdr.hop_limits == UINT8_MAX) {
-                                       input_set |= IAVF_INSET_IPV6_HOP_LIMIT;
-                                       VIRTCHNL_ADD_PROTO_HDR_FIELD_BIT(hdr, IPV6, HOP_LIMIT);
-                               }
-                               if (!memcmp(ipv6_mask->hdr.src_addr,
-                                       ipv6_addr_mask,
-                                       RTE_DIM(ipv6_mask->hdr.src_addr))) {
-                                       input_set |= IAVF_INSET_IPV6_SRC;
-                                       VIRTCHNL_ADD_PROTO_HDR_FIELD_BIT(hdr, IPV6, SRC);
-                               }
-                               if (!memcmp(ipv6_mask->hdr.dst_addr,
-                                       ipv6_addr_mask,
-                                       RTE_DIM(ipv6_mask->hdr.dst_addr))) {
-                                       input_set |= IAVF_INSET_IPV6_DST;
-                                       VIRTCHNL_ADD_PROTO_HDR_FIELD_BIT(hdr, IPV6, DST);
-                               }
+                       if (ipv6_mask->hdr.payload_len) {
+                               rte_flow_error_set(error, EINVAL,
+                                                  RTE_FLOW_ERROR_TYPE_ITEM,
+                                                  item, "Invalid IPv6 mask");
+                               return -rte_errno;
+                       }
 
-                               rte_memcpy(hdr->buffer,
-                                       &ipv6_spec->hdr,
-                                       sizeof(ipv6_spec->hdr));
+                       if ((ipv6_mask->hdr.vtc_flow &
+                             rte_cpu_to_be_32(IAVF_IPV6_TC_MASK))
+                            == rte_cpu_to_be_32(IAVF_IPV6_TC_MASK)) {
+                               input_set |= IAVF_INSET_IPV6_TC;
+                               VIRTCHNL_ADD_PROTO_HDR_FIELD_BIT(hdr, IPV6,
+                                                                TC);
                        }
 
-                       filter->add_fltr.rule_cfg.proto_hdrs.count = ++layer;
+                       if (ipv6_mask->hdr.proto == UINT8_MAX) {
+                               input_set |= IAVF_INSET_IPV6_NEXT_HDR;
+                               VIRTCHNL_ADD_PROTO_HDR_FIELD_BIT(hdr, IPV6,
+                                                                PROT);
+                       }
+
+                       if (ipv6_mask->hdr.hop_limits == UINT8_MAX) {
+                               input_set |= IAVF_INSET_IPV6_HOP_LIMIT;
+                               VIRTCHNL_ADD_PROTO_HDR_FIELD_BIT(hdr, IPV6,
+                                                                HOP_LIMIT);
+                       }
+
+                       if (!memcmp(ipv6_mask->hdr.src_addr, ipv6_addr_mask,
+                                   RTE_DIM(ipv6_mask->hdr.src_addr))) {
+                               input_set |= IAVF_INSET_IPV6_SRC;
+                               VIRTCHNL_ADD_PROTO_HDR_FIELD_BIT(hdr, IPV6,
+                                                                SRC);
+                       }
+                       if (!memcmp(ipv6_mask->hdr.dst_addr, ipv6_addr_mask,
+                                   RTE_DIM(ipv6_mask->hdr.dst_addr))) {
+                               input_set |= IAVF_INSET_IPV6_DST;
+                               VIRTCHNL_ADD_PROTO_HDR_FIELD_BIT(hdr, IPV6,
+                                                                DST);
+                       }
+
+                       if (tun_inner) {
+                               input_set &= ~IAVF_PROT_IPV6_OUTER;
+                               input_set |= IAVF_PROT_IPV6_INNER;
+                       }
+
+                       rte_memcpy(hdr->buffer, &ipv6_spec->hdr,
+                                  sizeof(ipv6_spec->hdr));
+
+                       hdrs->count = ++layer;
+                       break;
+
+               case RTE_FLOW_ITEM_TYPE_IPV6_FRAG_EXT:
+                       ipv6_frag_spec = item->spec;
+                       ipv6_frag_mask = item->mask;
+                       next_type = (item + 1)->type;
+
+                       hdr = &hdrs->proto_hdr[layer];
+
+                       VIRTCHNL_SET_PROTO_HDR_TYPE(hdr, IPV6_EH_FRAG);
+
+                       if (!(ipv6_frag_spec && ipv6_frag_mask)) {
+                               hdrs->count = ++layer;
+                               break;
+                       }
+
+                       /* fragment Ipv6:
+                        * spec is 0x1, mask is 0x1
+                        */
+                       if (ipv6_frag_spec->hdr.frag_data ==
+                           rte_cpu_to_be_16(1) &&
+                           ipv6_frag_mask->hdr.frag_data ==
+                           rte_cpu_to_be_16(1)) {
+                               /* all IPv6 fragment packet has the same
+                                * ethertype, if the spec and mask is valid,
+                                * set ethertype into input set.
+                                */
+                               input_set |= IAVF_INSET_ETHERTYPE;
+                               VIRTCHNL_ADD_PROTO_HDR_FIELD_BIT(hdr1, ETH,
+                                                                ETHERTYPE);
+
+                               rte_memcpy(hdr->buffer, &ipv6_frag_spec->hdr,
+                                          sizeof(ipv6_frag_spec->hdr));
+                       } else if (ipv6_frag_mask->hdr.id == UINT32_MAX) {
+                               rte_flow_error_set(error, EINVAL,
+                                                  RTE_FLOW_ERROR_TYPE_ITEM,
+                                                  item, "Invalid IPv6 mask.");
+                               return -rte_errno;
+                       }
+
+                       hdrs->count = ++layer;
                        break;
 
                case RTE_FLOW_ITEM_TYPE_UDP:
                        udp_spec = item->spec;
                        udp_mask = item->mask;
 
-                       hdr = &filter->add_fltr.rule_cfg.proto_hdrs.proto_hdr[layer];
+                       hdr = &hdrs->proto_hdr[layer];
 
                        VIRTCHNL_SET_PROTO_HDR_TYPE(hdr, UDP);
 
@@ -676,6 +959,11 @@ iavf_fdir_parse_pattern(__rte_unused struct iavf_adapter *ad,
                                        VIRTCHNL_ADD_PROTO_HDR_FIELD_BIT(hdr, UDP, DST_PORT);
                                }
 
+                               if (tun_inner) {
+                                       input_set &= ~IAVF_PROT_UDP_OUTER;
+                                       input_set |= IAVF_PROT_UDP_INNER;
+                               }
+
                                if (l3 == RTE_FLOW_ITEM_TYPE_IPV4)
                                        rte_memcpy(hdr->buffer,
                                                &udp_spec->hdr,
@@ -686,14 +974,14 @@ iavf_fdir_parse_pattern(__rte_unused struct iavf_adapter *ad,
                                                sizeof(udp_spec->hdr));
                        }
 
-                       filter->add_fltr.rule_cfg.proto_hdrs.count = ++layer;
+                       hdrs->count = ++layer;
                        break;
 
                case RTE_FLOW_ITEM_TYPE_TCP:
                        tcp_spec = item->spec;
                        tcp_mask = item->mask;
 
-                       hdr = &filter->add_fltr.rule_cfg.proto_hdrs.proto_hdr[layer];
+                       hdr = &hdrs->proto_hdr[layer];
 
                        VIRTCHNL_SET_PROTO_HDR_TYPE(hdr, TCP);
 
@@ -720,6 +1008,11 @@ iavf_fdir_parse_pattern(__rte_unused struct iavf_adapter *ad,
                                        VIRTCHNL_ADD_PROTO_HDR_FIELD_BIT(hdr, TCP, DST_PORT);
                                }
 
+                               if (tun_inner) {
+                                       input_set &= ~IAVF_PROT_TCP_OUTER;
+                                       input_set |= IAVF_PROT_TCP_INNER;
+                               }
+
                                if (l3 == RTE_FLOW_ITEM_TYPE_IPV4)
                                        rte_memcpy(hdr->buffer,
                                                &tcp_spec->hdr,
@@ -730,14 +1023,14 @@ iavf_fdir_parse_pattern(__rte_unused struct iavf_adapter *ad,
                                                sizeof(tcp_spec->hdr));
                        }
 
-                       filter->add_fltr.rule_cfg.proto_hdrs.count = ++layer;
+                       hdrs->count = ++layer;
                        break;
 
                case RTE_FLOW_ITEM_TYPE_SCTP:
                        sctp_spec = item->spec;
                        sctp_mask = item->mask;
 
-                       hdr = &filter->add_fltr.rule_cfg.proto_hdrs.proto_hdr[layer];
+                       hdr = &hdrs->proto_hdr[layer];
 
                        VIRTCHNL_SET_PROTO_HDR_TYPE(hdr, SCTP);
 
@@ -768,14 +1061,14 @@ iavf_fdir_parse_pattern(__rte_unused struct iavf_adapter *ad,
                                                sizeof(sctp_spec->hdr));
                        }
 
-                       filter->add_fltr.rule_cfg.proto_hdrs.count = ++layer;
+                       hdrs->count = ++layer;
                        break;
 
                case RTE_FLOW_ITEM_TYPE_GTPU:
                        gtp_spec = item->spec;
                        gtp_mask = item->mask;
 
-                       hdr = &filter->add_fltr.rule_cfg.proto_hdrs.proto_hdr[layer];
+                       hdr = &hdrs->proto_hdr[layer];
 
                        VIRTCHNL_SET_PROTO_HDR_TYPE(hdr, GTPU_IP);
 
@@ -798,16 +1091,25 @@ iavf_fdir_parse_pattern(__rte_unused struct iavf_adapter *ad,
                                        gtp_spec, sizeof(*gtp_spec));
                        }
 
-                       filter->add_fltr.rule_cfg.proto_hdrs.count = ++layer;
+                       tun_inner = 1;
+
+                       hdrs->count = ++layer;
                        break;
 
                case RTE_FLOW_ITEM_TYPE_GTP_PSC:
                        gtp_psc_spec = item->spec;
                        gtp_psc_mask = item->mask;
 
-                       hdr = &filter->add_fltr.rule_cfg.proto_hdrs.proto_hdr[layer];
+                       hdr = &hdrs->proto_hdr[layer];
 
-                       VIRTCHNL_SET_PROTO_HDR_TYPE(hdr, GTPU_EH);
+                       if (!gtp_psc_spec)
+                               VIRTCHNL_SET_PROTO_HDR_TYPE(hdr, GTPU_EH);
+                       else if ((gtp_psc_mask->qfi) && !(gtp_psc_mask->pdu_type))
+                               VIRTCHNL_SET_PROTO_HDR_TYPE(hdr, GTPU_EH);
+                       else if (gtp_psc_spec->pdu_type == IAVF_GTPU_EH_UPLINK)
+                               VIRTCHNL_SET_PROTO_HDR_TYPE(hdr, GTPU_EH_PDU_UP);
+                       else if (gtp_psc_spec->pdu_type == IAVF_GTPU_EH_DWLINK)
+                               VIRTCHNL_SET_PROTO_HDR_TYPE(hdr, GTPU_EH_PDU_DWN);
 
                        if (gtp_psc_spec && gtp_psc_mask) {
                                if (gtp_psc_mask->qfi == UINT8_MAX) {
@@ -819,14 +1121,14 @@ iavf_fdir_parse_pattern(__rte_unused struct iavf_adapter *ad,
                                        sizeof(*gtp_psc_spec));
                        }
 
-                       filter->add_fltr.rule_cfg.proto_hdrs.count = ++layer;
+                       hdrs->count = ++layer;
                        break;
 
                case RTE_FLOW_ITEM_TYPE_L2TPV3OIP:
                        l2tpv3oip_spec = item->spec;
                        l2tpv3oip_mask = item->mask;
 
-                       hdr = &filter->add_fltr.rule_cfg.proto_hdrs.proto_hdr[layer];
+                       hdr = &hdrs->proto_hdr[layer];
 
                        VIRTCHNL_SET_PROTO_HDR_TYPE(hdr, L2TPV3);
 
@@ -840,14 +1142,14 @@ iavf_fdir_parse_pattern(__rte_unused struct iavf_adapter *ad,
                                        sizeof(*l2tpv3oip_spec));
                        }
 
-                       filter->add_fltr.rule_cfg.proto_hdrs.count = ++layer;
+                       hdrs->count = ++layer;
                        break;
 
                case RTE_FLOW_ITEM_TYPE_ESP:
                        esp_spec = item->spec;
                        esp_mask = item->mask;
 
-                       hdr = &filter->add_fltr.rule_cfg.proto_hdrs.proto_hdr[layer];
+                       hdr = &hdrs->proto_hdr[layer];
 
                        VIRTCHNL_SET_PROTO_HDR_TYPE(hdr, ESP);
 
@@ -861,14 +1163,14 @@ iavf_fdir_parse_pattern(__rte_unused struct iavf_adapter *ad,
                                        sizeof(esp_spec->hdr));
                        }
 
-                       filter->add_fltr.rule_cfg.proto_hdrs.count = ++layer;
+                       hdrs->count = ++layer;
                        break;
 
                case RTE_FLOW_ITEM_TYPE_AH:
                        ah_spec = item->spec;
                        ah_mask = item->mask;
 
-                       hdr = &filter->add_fltr.rule_cfg.proto_hdrs.proto_hdr[layer];
+                       hdr = &hdrs->proto_hdr[layer];
 
                        VIRTCHNL_SET_PROTO_HDR_TYPE(hdr, AH);
 
@@ -882,14 +1184,14 @@ iavf_fdir_parse_pattern(__rte_unused struct iavf_adapter *ad,
                                        sizeof(*ah_spec));
                        }
 
-                       filter->add_fltr.rule_cfg.proto_hdrs.count = ++layer;
+                       hdrs->count = ++layer;
                        break;
 
                case RTE_FLOW_ITEM_TYPE_PFCP:
                        pfcp_spec = item->spec;
                        pfcp_mask = item->mask;
 
-                       hdr = &filter->add_fltr.rule_cfg.proto_hdrs.proto_hdr[layer];
+                       hdr = &hdrs->proto_hdr[layer];
 
                        VIRTCHNL_SET_PROTO_HDR_TYPE(hdr, PFCP);
 
@@ -903,7 +1205,50 @@ iavf_fdir_parse_pattern(__rte_unused struct iavf_adapter *ad,
                                        sizeof(*pfcp_spec));
                        }
 
-                       filter->add_fltr.rule_cfg.proto_hdrs.count = ++layer;
+                       hdrs->count = ++layer;
+                       break;
+
+               case RTE_FLOW_ITEM_TYPE_ECPRI:
+                       ecpri_spec = item->spec;
+                       ecpri_mask = item->mask;
+
+                       ecpri_common.u32 = rte_be_to_cpu_32(ecpri_spec->hdr.common.u32);
+
+                       hdr = &hdrs->proto_hdr[layer];
+
+                       VIRTCHNL_SET_PROTO_HDR_TYPE(hdr, ECPRI);
+
+                       if (ecpri_spec && ecpri_mask) {
+                               if (ecpri_common.type == RTE_ECPRI_MSG_TYPE_IQ_DATA &&
+                                               ecpri_mask->hdr.type0.pc_id == UINT16_MAX) {
+                                       input_set |= IAVF_ECPRI_PC_RTC_ID;
+                                       VIRTCHNL_ADD_PROTO_HDR_FIELD_BIT(hdr, ECPRI,
+                                                                        PC_RTC_ID);
+                               }
+
+                               rte_memcpy(hdr->buffer, ecpri_spec,
+                                       sizeof(*ecpri_spec));
+                       }
+
+                       hdrs->count = ++layer;
+                       break;
+
+               case RTE_FLOW_ITEM_TYPE_GRE:
+                       gre_spec = item->spec;
+                       gre_mask = item->mask;
+
+                       hdr = &hdrs->proto_hdr[layer];
+
+                       VIRTCHNL_SET_PROTO_HDR_TYPE(hdr, GRE);
+
+                       if (gre_spec && gre_mask) {
+                               rte_memcpy(hdr->buffer, gre_spec,
+                                          sizeof(*gre_spec));
+                       }
+
+                       tun_inner = 1;
+
+                       hdrs->count = ++layer;
                        break;
 
                case RTE_FLOW_ITEM_TYPE_VOID:
@@ -924,6 +1269,15 @@ iavf_fdir_parse_pattern(__rte_unused struct iavf_adapter *ad,
                return -rte_errno;
        }
 
+       if (!iavf_fdir_refine_input_set(input_set,
+                                       input_set_mask | IAVF_INSET_ETHERTYPE,
+                                       filter)) {
+               rte_flow_error_set(error, EINVAL,
+                                  RTE_FLOW_ERROR_TYPE_ITEM_SPEC, pattern,
+                                  "Invalid input set");
+               return -rte_errno;
+       }
+
        filter->input_set = input_set;
 
        return 0;
@@ -941,7 +1295,6 @@ iavf_fdir_parse(struct iavf_adapter *ad,
        struct iavf_info *vf = IAVF_DEV_PRIVATE_TO_VF(ad);
        struct iavf_fdir_conf *filter = &vf->fdir.conf;
        struct iavf_pattern_match_item *item = NULL;
-       uint64_t input_set;
        int ret;
 
        memset(filter, 0, sizeof(*filter));
@@ -950,19 +1303,11 @@ iavf_fdir_parse(struct iavf_adapter *ad,
        if (!item)
                return -rte_errno;
 
-       ret = iavf_fdir_parse_pattern(ad, pattern, error, filter);
+       ret = iavf_fdir_parse_pattern(ad, pattern, item->input_set_mask,
+                                     error, filter);
        if (ret)
                goto error;
 
-       input_set = filter->input_set;
-       if (!input_set || input_set & ~item->input_set_mask) {
-               rte_flow_error_set(error, EINVAL,
-                               RTE_FLOW_ERROR_TYPE_ITEM_SPEC, pattern,
-                               "Invalid input set");
-               ret = -rte_errno;
-               goto error;
-       }
-
        ret = iavf_fdir_parse_action(ad, actions, error, filter);
        if (ret)
                goto error;