net: add rte prefix to IP structure
[dpdk.git] / drivers / net / enic / enic_clsf.c
index 7d2bb78..65d0be3 100644 (file)
@@ -1,43 +1,16 @@
-/*
- * Copyright 2008-2014 Cisco Systems, Inc.  All rights reserved.
+/* SPDX-License-Identifier: BSD-3-Clause
+ * Copyright 2008-2017 Cisco Systems, Inc.  All rights reserved.
  * Copyright 2007 Nuova Systems, Inc.  All rights reserved.
- *
- * Copyright (c) 2014, Cisco Systems, Inc.
- * All rights reserved.
- *
- * Redistribution and use in source and binary forms, with or without
- * modification, are permitted provided that the following conditions
- * are met:
- *
- * 1. Redistributions of source code must retain the above copyright
- * notice, this list of conditions and the following disclaimer.
- *
- * 2. Redistributions in binary form must reproduce the above copyright
- * notice, this list of conditions and the following disclaimer in
- * the documentation and/or other materials provided with the
- * distribution.
- *
- * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
- * "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
- * LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS
- * FOR A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE
- * COPYRIGHT HOLDER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT,
- * INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING,
- * BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
- * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER
- * CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
- * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN
- * ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
- * POSSIBILITY OF SUCH DAMAGE.
- *
  */
 
-#include <libgen.h>
-
-#include <rte_ethdev.h>
+#include <rte_ethdev_driver.h>
 #include <rte_malloc.h>
 #include <rte_hash.h>
 #include <rte_byteorder.h>
+#include <rte_ip.h>
+#include <rte_tcp.h>
+#include <rte_udp.h>
+#include <rte_sctp.h>
 
 #include "enic_compat.h"
 #include "enic.h"
@@ -52,7 +25,7 @@
 #include "vnic_intr.h"
 #include "vnic_nic.h"
 
-#ifdef RTE_MACHINE_CPUFLAG_SSE4_2
+#ifdef RTE_ARCH_X86
 #include <rte_hash_crc.h>
 #define DEFAULT_HASH_FUNC       rte_hash_crc
 #else
 
 #define ENICPMD_CLSF_HASH_ENTRIES       ENICPMD_FDIR_MAX
 
+static void copy_fltr_v1(struct filter_v2 *fltr,
+               const struct rte_eth_fdir_input *input,
+               const struct rte_eth_fdir_masks *masks);
+static void copy_fltr_v2(struct filter_v2 *fltr,
+               const struct rte_eth_fdir_input *input,
+               const struct rte_eth_fdir_masks *masks);
+
 void enic_fdir_stats_get(struct enic *enic, struct rte_eth_fdir_stats *stats)
 {
        *stats = enic->fdir.stats;
 }
 
+void enic_fdir_info_get(struct enic *enic, struct rte_eth_fdir_info *info)
+{
+       info->mode = (enum rte_fdir_mode)enic->fdir.modes;
+       info->flow_types_mask[0] = enic->fdir.types_mask;
+}
+
+void enic_fdir_info(struct enic *enic)
+{
+       enic->fdir.modes = (u32)RTE_FDIR_MODE_PERFECT;
+       enic->fdir.types_mask  = 1 << RTE_ETH_FLOW_NONFRAG_IPV4_UDP |
+                                1 << RTE_ETH_FLOW_NONFRAG_IPV4_TCP;
+       if (enic->adv_filters) {
+               enic->fdir.types_mask |= 1 << RTE_ETH_FLOW_NONFRAG_IPV4_OTHER |
+                                        1 << RTE_ETH_FLOW_NONFRAG_IPV4_SCTP |
+                                        1 << RTE_ETH_FLOW_NONFRAG_IPV6_UDP |
+                                        1 << RTE_ETH_FLOW_NONFRAG_IPV6_TCP |
+                                        1 << RTE_ETH_FLOW_NONFRAG_IPV6_SCTP |
+                                        1 << RTE_ETH_FLOW_NONFRAG_IPV6_OTHER;
+               enic->fdir.copy_fltr_fn = copy_fltr_v2;
+       } else {
+               enic->fdir.copy_fltr_fn = copy_fltr_v1;
+       }
+}
+
+static void
+enic_set_layer(struct filter_generic_1 *gp, unsigned int flag,
+              enum filter_generic_1_layer layer, void *mask, void *val,
+              unsigned int len)
+{
+       gp->mask_flags |= flag;
+       gp->val_flags |= gp->mask_flags;
+       memcpy(gp->layer[layer].mask, mask, len);
+       memcpy(gp->layer[layer].val, val, len);
+}
+
+/* Copy Flow Director filter to a VIC ipv4 filter (for Cisco VICs
+ * without advanced filter support.
+ */
+static void
+copy_fltr_v1(struct filter_v2 *fltr, const struct rte_eth_fdir_input *input,
+            __rte_unused const struct rte_eth_fdir_masks *masks)
+{
+       fltr->type = FILTER_IPV4_5TUPLE;
+       fltr->u.ipv4.src_addr = rte_be_to_cpu_32(
+               input->flow.ip4_flow.src_ip);
+       fltr->u.ipv4.dst_addr = rte_be_to_cpu_32(
+               input->flow.ip4_flow.dst_ip);
+       fltr->u.ipv4.src_port = rte_be_to_cpu_16(
+               input->flow.udp4_flow.src_port);
+       fltr->u.ipv4.dst_port = rte_be_to_cpu_16(
+               input->flow.udp4_flow.dst_port);
+
+       if (input->flow_type == RTE_ETH_FLOW_NONFRAG_IPV4_TCP)
+               fltr->u.ipv4.protocol = PROTO_TCP;
+       else
+               fltr->u.ipv4.protocol = PROTO_UDP;
+
+       fltr->u.ipv4.flags = FILTER_FIELDS_IPV4_5TUPLE;
+}
+
+/* Copy Flow Director filter to a VIC generic filter (requires advanced
+ * filter support.
+ */
+static void
+copy_fltr_v2(struct filter_v2 *fltr, const struct rte_eth_fdir_input *input,
+            const struct rte_eth_fdir_masks *masks)
+{
+       struct filter_generic_1 *gp = &fltr->u.generic_1;
+
+       fltr->type = FILTER_DPDK_1;
+       memset(gp, 0, sizeof(*gp));
+
+       if (input->flow_type == RTE_ETH_FLOW_NONFRAG_IPV4_UDP) {
+               struct udp_hdr udp_mask, udp_val;
+               memset(&udp_mask, 0, sizeof(udp_mask));
+               memset(&udp_val, 0, sizeof(udp_val));
+
+               if (input->flow.udp4_flow.src_port) {
+                       udp_mask.src_port = masks->src_port_mask;
+                       udp_val.src_port = input->flow.udp4_flow.src_port;
+               }
+               if (input->flow.udp4_flow.dst_port) {
+                       udp_mask.dst_port = masks->dst_port_mask;
+                       udp_val.dst_port = input->flow.udp4_flow.dst_port;
+               }
+
+               enic_set_layer(gp, FILTER_GENERIC_1_UDP, FILTER_GENERIC_1_L4,
+                              &udp_mask, &udp_val, sizeof(struct udp_hdr));
+       } else if (input->flow_type == RTE_ETH_FLOW_NONFRAG_IPV4_TCP) {
+               struct tcp_hdr tcp_mask, tcp_val;
+               memset(&tcp_mask, 0, sizeof(tcp_mask));
+               memset(&tcp_val, 0, sizeof(tcp_val));
+
+               if (input->flow.tcp4_flow.src_port) {
+                       tcp_mask.src_port = masks->src_port_mask;
+                       tcp_val.src_port = input->flow.tcp4_flow.src_port;
+               }
+               if (input->flow.tcp4_flow.dst_port) {
+                       tcp_mask.dst_port = masks->dst_port_mask;
+                       tcp_val.dst_port = input->flow.tcp4_flow.dst_port;
+               }
+
+               enic_set_layer(gp, FILTER_GENERIC_1_TCP, FILTER_GENERIC_1_L4,
+                              &tcp_mask, &tcp_val, sizeof(struct tcp_hdr));
+       } else if (input->flow_type == RTE_ETH_FLOW_NONFRAG_IPV4_SCTP) {
+               struct sctp_hdr sctp_mask, sctp_val;
+               memset(&sctp_mask, 0, sizeof(sctp_mask));
+               memset(&sctp_val, 0, sizeof(sctp_val));
+
+               if (input->flow.sctp4_flow.src_port) {
+                       sctp_mask.src_port = masks->src_port_mask;
+                       sctp_val.src_port = input->flow.sctp4_flow.src_port;
+               }
+               if (input->flow.sctp4_flow.dst_port) {
+                       sctp_mask.dst_port = masks->dst_port_mask;
+                       sctp_val.dst_port = input->flow.sctp4_flow.dst_port;
+               }
+               if (input->flow.sctp4_flow.verify_tag) {
+                       sctp_mask.tag = 0xffffffff;
+                       sctp_val.tag = input->flow.sctp4_flow.verify_tag;
+               }
+
+               /*
+                * Unlike UDP/TCP (FILTER_GENERIC_1_{UDP,TCP}), the firmware
+                * has no "packet is SCTP" flag. Use flag=0 (generic L4) and
+                * manually set proto_id=sctp below.
+                */
+               enic_set_layer(gp, 0, FILTER_GENERIC_1_L4, &sctp_mask,
+                              &sctp_val, sizeof(struct sctp_hdr));
+       }
+
+       if (input->flow_type == RTE_ETH_FLOW_NONFRAG_IPV4_UDP ||
+           input->flow_type == RTE_ETH_FLOW_NONFRAG_IPV4_TCP ||
+           input->flow_type == RTE_ETH_FLOW_NONFRAG_IPV4_SCTP ||
+           input->flow_type == RTE_ETH_FLOW_NONFRAG_IPV4_OTHER) {
+               struct rte_ipv4_hdr ip4_mask, ip4_val;
+               memset(&ip4_mask, 0, sizeof(struct rte_ipv4_hdr));
+               memset(&ip4_val, 0, sizeof(struct rte_ipv4_hdr));
+
+               if (input->flow.ip4_flow.tos) {
+                       ip4_mask.type_of_service = masks->ipv4_mask.tos;
+                       ip4_val.type_of_service = input->flow.ip4_flow.tos;
+               }
+               if (input->flow.ip4_flow.ttl) {
+                       ip4_mask.time_to_live = masks->ipv4_mask.ttl;
+                       ip4_val.time_to_live = input->flow.ip4_flow.ttl;
+               }
+               if (input->flow.ip4_flow.proto) {
+                       ip4_mask.next_proto_id = masks->ipv4_mask.proto;
+                       ip4_val.next_proto_id = input->flow.ip4_flow.proto;
+               } else if (input->flow_type == RTE_ETH_FLOW_NONFRAG_IPV4_SCTP) {
+                       /* Explicitly match the SCTP protocol number */
+                       ip4_mask.next_proto_id = 0xff;
+                       ip4_val.next_proto_id = IPPROTO_SCTP;
+               }
+               if (input->flow.ip4_flow.src_ip) {
+                       ip4_mask.src_addr =  masks->ipv4_mask.src_ip;
+                       ip4_val.src_addr = input->flow.ip4_flow.src_ip;
+               }
+               if (input->flow.ip4_flow.dst_ip) {
+                       ip4_mask.dst_addr =  masks->ipv4_mask.dst_ip;
+                       ip4_val.dst_addr = input->flow.ip4_flow.dst_ip;
+               }
+
+               enic_set_layer(gp, FILTER_GENERIC_1_IPV4, FILTER_GENERIC_1_L3,
+                       &ip4_mask, &ip4_val, sizeof(struct rte_ipv4_hdr));
+       }
+
+       if (input->flow_type == RTE_ETH_FLOW_NONFRAG_IPV6_UDP) {
+               struct udp_hdr udp_mask, udp_val;
+               memset(&udp_mask, 0, sizeof(udp_mask));
+               memset(&udp_val, 0, sizeof(udp_val));
+
+               if (input->flow.udp6_flow.src_port) {
+                       udp_mask.src_port = masks->src_port_mask;
+                       udp_val.src_port = input->flow.udp6_flow.src_port;
+               }
+               if (input->flow.udp6_flow.dst_port) {
+                       udp_mask.dst_port = masks->dst_port_mask;
+                       udp_val.dst_port = input->flow.udp6_flow.dst_port;
+               }
+               enic_set_layer(gp, FILTER_GENERIC_1_UDP, FILTER_GENERIC_1_L4,
+                              &udp_mask, &udp_val, sizeof(struct udp_hdr));
+       } else if (input->flow_type == RTE_ETH_FLOW_NONFRAG_IPV6_TCP) {
+               struct tcp_hdr tcp_mask, tcp_val;
+               memset(&tcp_mask, 0, sizeof(tcp_mask));
+               memset(&tcp_val, 0, sizeof(tcp_val));
+
+               if (input->flow.tcp6_flow.src_port) {
+                       tcp_mask.src_port = masks->src_port_mask;
+                       tcp_val.src_port = input->flow.tcp6_flow.src_port;
+               }
+               if (input->flow.tcp6_flow.dst_port) {
+                       tcp_mask.dst_port = masks->dst_port_mask;
+                       tcp_val.dst_port = input->flow.tcp6_flow.dst_port;
+               }
+               enic_set_layer(gp, FILTER_GENERIC_1_TCP, FILTER_GENERIC_1_L4,
+                              &tcp_mask, &tcp_val, sizeof(struct tcp_hdr));
+       } else if (input->flow_type == RTE_ETH_FLOW_NONFRAG_IPV6_SCTP) {
+               struct sctp_hdr sctp_mask, sctp_val;
+               memset(&sctp_mask, 0, sizeof(sctp_mask));
+               memset(&sctp_val, 0, sizeof(sctp_val));
+
+               if (input->flow.sctp6_flow.src_port) {
+                       sctp_mask.src_port = masks->src_port_mask;
+                       sctp_val.src_port = input->flow.sctp6_flow.src_port;
+               }
+               if (input->flow.sctp6_flow.dst_port) {
+                       sctp_mask.dst_port = masks->dst_port_mask;
+                       sctp_val.dst_port = input->flow.sctp6_flow.dst_port;
+               }
+               if (input->flow.sctp6_flow.verify_tag) {
+                       sctp_mask.tag = 0xffffffff;
+                       sctp_val.tag = input->flow.sctp6_flow.verify_tag;
+               }
+
+               enic_set_layer(gp, 0, FILTER_GENERIC_1_L4, &sctp_mask,
+                              &sctp_val, sizeof(struct sctp_hdr));
+       }
+
+       if (input->flow_type == RTE_ETH_FLOW_NONFRAG_IPV6_UDP ||
+           input->flow_type == RTE_ETH_FLOW_NONFRAG_IPV6_TCP ||
+           input->flow_type == RTE_ETH_FLOW_NONFRAG_IPV6_SCTP ||
+           input->flow_type == RTE_ETH_FLOW_NONFRAG_IPV6_OTHER) {
+               struct rte_ipv6_hdr ipv6_mask, ipv6_val;
+               memset(&ipv6_mask, 0, sizeof(struct rte_ipv6_hdr));
+               memset(&ipv6_val, 0, sizeof(struct rte_ipv6_hdr));
+
+               if (input->flow.ipv6_flow.proto) {
+                       ipv6_mask.proto = masks->ipv6_mask.proto;
+                       ipv6_val.proto = input->flow.ipv6_flow.proto;
+               } else if (input->flow_type == RTE_ETH_FLOW_NONFRAG_IPV6_SCTP) {
+                       /* See comments for IPv4 SCTP above. */
+                       ipv6_mask.proto = 0xff;
+                       ipv6_val.proto = IPPROTO_SCTP;
+               }
+               memcpy(ipv6_mask.src_addr, masks->ipv6_mask.src_ip,
+                      sizeof(ipv6_mask.src_addr));
+               memcpy(ipv6_val.src_addr, input->flow.ipv6_flow.src_ip,
+                      sizeof(ipv6_val.src_addr));
+               memcpy(ipv6_mask.dst_addr, masks->ipv6_mask.dst_ip,
+                      sizeof(ipv6_mask.dst_addr));
+               memcpy(ipv6_val.dst_addr, input->flow.ipv6_flow.dst_ip,
+                      sizeof(ipv6_val.dst_addr));
+               if (input->flow.ipv6_flow.tc) {
+                       ipv6_mask.vtc_flow = masks->ipv6_mask.tc << 12;
+                       ipv6_val.vtc_flow = input->flow.ipv6_flow.tc << 12;
+               }
+               if (input->flow.ipv6_flow.hop_limits) {
+                       ipv6_mask.hop_limits = masks->ipv6_mask.hop_limits;
+                       ipv6_val.hop_limits = input->flow.ipv6_flow.hop_limits;
+               }
+
+               enic_set_layer(gp, FILTER_GENERIC_1_IPV6, FILTER_GENERIC_1_L3,
+                       &ipv6_mask, &ipv6_val, sizeof(struct rte_ipv6_hdr));
+       }
+}
+
 int enic_fdir_del_fltr(struct enic *enic, struct rte_eth_fdir_filter *params)
 {
        int32_t pos;
@@ -84,7 +322,7 @@ int enic_fdir_del_fltr(struct enic *enic, struct rte_eth_fdir_filter *params)
 
                /* Delete the filter */
                vnic_dev_classifier(enic->vdev, CLSF_DEL,
-                       &key->fltr_id, NULL);
+                       &key->fltr_id, NULL, NULL);
                rte_free(key);
                enic->fdir.nodes[pos] = NULL;
                enic->fdir.stats.free++;
@@ -97,17 +335,19 @@ int enic_fdir_del_fltr(struct enic *enic, struct rte_eth_fdir_filter *params)
 int enic_fdir_add_fltr(struct enic *enic, struct rte_eth_fdir_filter *params)
 {
        struct enic_fdir_node *key;
-       struct filter fltr = {0};
+       struct filter_v2 fltr;
        int32_t pos;
        u8 do_free = 0;
        u16 old_fltr_id = 0;
        u32 flowtype_supported;
        u16 flex_bytes;
        u16 queue;
+       struct filter_action_v2 action;
 
-       flowtype_supported = (
-               (RTE_ETH_FLOW_NONFRAG_IPV4_TCP == params->input.flow_type) ||
-               (RTE_ETH_FLOW_NONFRAG_IPV4_UDP == params->input.flow_type));
+       memset(&fltr, 0, sizeof(fltr));
+       memset(&action, 0, sizeof(action));
+       flowtype_supported = enic->fdir.types_mask
+                            & (1 << params->input.flow_type);
 
        flex_bytes = ((params->input.flow_ext.flexbytes[1] << 8 & 0xFF00) |
                (params->input.flow_ext.flexbytes[0] & 0xFF));
@@ -120,7 +360,12 @@ int enic_fdir_add_fltr(struct enic *enic, struct rte_eth_fdir_filter *params)
                return -ENOTSUP;
        }
 
-       queue = params->action.rx_queue;
+       /* Get the enicpmd RQ from the DPDK Rx queue */
+       queue = enic_rte_rq_idx_to_sop_idx(params->action.rx_queue);
+
+       if (!enic->rq[queue].in_use)
+               return -EINVAL;
+
        /* See if the key is already there in the table */
        pos = rte_hash_del_key(enic->fdir.hash, params);
        switch (pos) {
@@ -173,7 +418,7 @@ int enic_fdir_add_fltr(struct enic *enic, struct rte_eth_fdir_filter *params)
                         * Delete the filter and add the modified one later
                         */
                        vnic_dev_classifier(enic->vdev, CLSF_DEL,
-                               &key->fltr_id, NULL);
+                               &key->fltr_id, NULL, NULL);
                        enic->fdir.stats.free++;
                }
 
@@ -183,24 +428,13 @@ int enic_fdir_add_fltr(struct enic *enic, struct rte_eth_fdir_filter *params)
        key->filter = *params;
        key->rq_index = queue;
 
-       fltr.type = FILTER_IPV4_5TUPLE;
-       fltr.u.ipv4.src_addr = rte_be_to_cpu_32(
-               params->input.flow.ip4_flow.src_ip);
-       fltr.u.ipv4.dst_addr = rte_be_to_cpu_32(
-               params->input.flow.ip4_flow.dst_ip);
-       fltr.u.ipv4.src_port = rte_be_to_cpu_16(
-               params->input.flow.udp4_flow.src_port);
-       fltr.u.ipv4.dst_port = rte_be_to_cpu_16(
-               params->input.flow.udp4_flow.dst_port);
-
-       if (RTE_ETH_FLOW_NONFRAG_IPV4_TCP == params->input.flow_type)
-               fltr.u.ipv4.protocol = PROTO_TCP;
-       else
-               fltr.u.ipv4.protocol = PROTO_UDP;
-
-       fltr.u.ipv4.flags = FILTER_FIELDS_IPV4_5TUPLE;
+       enic->fdir.copy_fltr_fn(&fltr, &params->input,
+                               &enic->rte_dev->data->dev_conf.fdir_conf.mask);
+       action.type = FILTER_ACTION_RQ_STEERING;
+       action.rq_idx = queue;
 
-       if (!vnic_dev_classifier(enic->vdev, CLSF_ADD, &queue, &fltr)) {
+       if (!vnic_dev_classifier(enic->vdev, CLSF_ADD, &queue, &fltr,
+           &action)) {
                key->fltr_id = queue;
        } else {
                dev_err(enic, "Add classifier entry failed\n");
@@ -210,7 +444,8 @@ int enic_fdir_add_fltr(struct enic *enic, struct rte_eth_fdir_filter *params)
        }
 
        if (do_free)
-               vnic_dev_classifier(enic->vdev, CLSF_DEL, &old_fltr_id, NULL);
+               vnic_dev_classifier(enic->vdev, CLSF_DEL, &old_fltr_id, NULL,
+                                   NULL);
        else{
                enic->fdir.stats.free--;
                enic->fdir.stats.add++;
@@ -218,6 +453,7 @@ int enic_fdir_add_fltr(struct enic *enic, struct rte_eth_fdir_filter *params)
 
        pos = rte_hash_add_key(enic->fdir.hash, params);
        if (pos < 0) {
+               enic->fdir.stats.f_add++;
                dev_err(enic, "Add hash key failed\n");
                return pos;
        }
@@ -235,8 +471,9 @@ void enic_clsf_destroy(struct enic *enic)
                key = enic->fdir.nodes[index];
                if (key) {
                        vnic_dev_classifier(enic->vdev, CLSF_DEL,
-                               &key->fltr_id, NULL);
+                               &key->fltr_id, NULL, NULL);
                        rte_free(key);
+                       enic->fdir.nodes[index] = NULL;
                }
        }
 
@@ -248,15 +485,16 @@ void enic_clsf_destroy(struct enic *enic)
 
 int enic_clsf_init(struct enic *enic)
 {
+       char clsf_name[RTE_HASH_NAMESIZE];
        struct rte_hash_parameters hash_params = {
-               .name = "enicpmd_clsf_hash",
+               .name = clsf_name,
                .entries = ENICPMD_CLSF_HASH_ENTRIES,
                .key_len = sizeof(struct rte_eth_fdir_filter),
                .hash_func = DEFAULT_HASH_FUNC,
                .hash_func_init_val = 0,
                .socket_id = SOCKET_ID_ANY,
        };
-
+       snprintf(clsf_name, RTE_HASH_NAMESIZE, "enic_clsf_%s", enic->bdf_name);
        enic->fdir.hash = rte_hash_create(&hash_params);
        memset(&enic->fdir.stats, 0, sizeof(enic->fdir.stats));
        enic->fdir.stats.free = ENICPMD_FDIR_MAX;