net/kni: set packet input port in Rx
[dpdk.git] / drivers / net / qede / qede_filter.c
index fd49858..86a2e0d 100644 (file)
@@ -221,7 +221,7 @@ qede_fdir_to_arfs_filter(struct rte_eth_dev *eth_dev,
        case RTE_ETH_FLOW_NONFRAG_IPV4_TCP:
        case RTE_ETH_FLOW_NONFRAG_IPV4_UDP:
                /* fill the common ip header */
-               arfs->tuple.eth_proto = RTE_ETHER_TYPE_IPv4;
+               arfs->tuple.eth_proto = RTE_ETHER_TYPE_IPV4;
                arfs->tuple.dst_ipv4 = input->flow.ip4_flow.dst_ip;
                arfs->tuple.src_ipv4 = input->flow.ip4_flow.src_ip;
                arfs->tuple.ip_proto = next_proto[input->flow_type];
@@ -237,7 +237,7 @@ qede_fdir_to_arfs_filter(struct rte_eth_dev *eth_dev,
                break;
        case RTE_ETH_FLOW_NONFRAG_IPV6_TCP:
        case RTE_ETH_FLOW_NONFRAG_IPV6_UDP:
-               arfs->tuple.eth_proto = RTE_ETHER_TYPE_IPv6;
+               arfs->tuple.eth_proto = RTE_ETHER_TYPE_IPV6;
                arfs->tuple.ip_proto = next_proto[input->flow_type];
                rte_memcpy(arfs->tuple.dst_ipv6,
                           &input->flow.ipv6_flow.dst_ip,
@@ -272,6 +272,7 @@ qede_config_arfs_filter(struct rte_eth_dev *eth_dev,
 {
        struct qede_dev *qdev = QEDE_INIT_QDEV(eth_dev);
        struct ecore_dev *edev = QEDE_INIT_EDEV(qdev);
+       struct ecore_ntuple_filter_params params;
        char mz_name[RTE_MEMZONE_NAMESIZE] = {0};
        struct qede_arfs_entry *tmp = NULL;
        const struct rte_memzone *mz;
@@ -344,12 +345,18 @@ qede_config_arfs_filter(struct rte_eth_dev *eth_dev,
                ecore_arfs_mode_configure(p_hwfn, p_hwfn->p_arfs_ptt,
                                          &qdev->arfs_info.arfs);
        }
+
+       memset(&params, 0, sizeof(params));
+       params.addr = (dma_addr_t)mz->iova;
+       params.length = pkt_len;
+       params.qid = arfs->rx_queue;
+       params.vport_id = 0;
+       params.b_is_add = add;
+       params.b_is_drop = arfs->is_drop;
+
        /* configure filter with ECORE_SPQ_MODE_EBLOCK */
        rc = ecore_configure_rfs_ntuple_filter(p_hwfn, NULL,
-                                              (dma_addr_t)mz->iova,
-                                              pkt_len,
-                                              arfs->rx_queue,
-                                              0, add);
+                                              &params);
        if (rc == ECORE_SUCCESS) {
                if (add) {
                        arfs->pkt_len = pkt_len;
@@ -431,7 +438,7 @@ qede_fdir_filter_add(struct rte_eth_dev *eth_dev,
                return -EINVAL;
        }
 
-       if (fdir->action.rx_queue >= QEDE_RSS_COUNT(qdev)) {
+       if (fdir->action.rx_queue >= QEDE_RSS_COUNT(eth_dev)) {
                DP_ERR(edev, "invalid queue number %u\n",
                       fdir->action.rx_queue);
                return -EINVAL;
@@ -473,7 +480,7 @@ qede_arfs_construct_pkt(struct rte_eth_dev *eth_dev,
 
        *ether_type = rte_cpu_to_be_16(arfs->tuple.eth_proto);
        switch (arfs->tuple.eth_proto) {
-       case RTE_ETHER_TYPE_IPv4:
+       case RTE_ETHER_TYPE_IPV4:
                ip = (struct rte_ipv4_hdr *)raw_pkt;
                ip->version_ihl = QEDE_FDIR_IP_DEFAULT_VERSION_IHL;
                ip->total_length = sizeof(struct rte_ipv4_hdr);
@@ -506,7 +513,7 @@ qede_arfs_construct_pkt(struct rte_eth_dev *eth_dev,
                        params->tcp = true;
                }
                break;
-       case RTE_ETHER_TYPE_IPv6:
+       case RTE_ETHER_TYPE_IPV6:
                ip6 = (struct rte_ipv6_hdr *)raw_pkt;
                ip6->proto = arfs->tuple.ip_proto;
                ip6->vtc_flow =
@@ -1023,7 +1030,7 @@ qede_set_ucast_tunn_cmn_param(struct ecore_filter_ucast *ucast,
 static int
 _qede_tunn_filter_config(struct rte_eth_dev *eth_dev,
                         const struct rte_eth_tunnel_filter_conf *conf,
-                        __attribute__((unused)) enum rte_filter_op filter_op,
+                        __rte_unused enum rte_filter_op filter_op,
                         enum ecore_tunn_clss *clss,
                         bool add)
 {
@@ -1161,7 +1168,7 @@ qede_tunn_filter_config(struct rte_eth_dev *eth_dev,
 }
 
 static int
-qede_flow_validate_attr(__attribute__((unused))struct rte_eth_dev *dev,
+qede_flow_validate_attr(__rte_unused struct rte_eth_dev *dev,
                        const struct rte_flow_attr *attr,
                        struct rte_flow_error *error)
 {
@@ -1211,7 +1218,7 @@ qede_flow_validate_attr(__attribute__((unused))struct rte_eth_dev *dev,
 }
 
 static int
-qede_flow_parse_pattern(__attribute__((unused))struct rte_eth_dev *dev,
+qede_flow_parse_pattern(__rte_unused struct rte_eth_dev *dev,
                        const struct rte_flow_item pattern[],
                        struct rte_flow_error *error,
                        struct rte_flow *flow)
@@ -1267,7 +1274,7 @@ qede_flow_parse_pattern(__attribute__((unused))struct rte_eth_dev *dev,
                                flow->entry.tuple.src_ipv4 = spec->hdr.src_addr;
                                flow->entry.tuple.dst_ipv4 = spec->hdr.dst_addr;
                                flow->entry.tuple.eth_proto =
-                                       RTE_ETHER_TYPE_IPv4;
+                                       RTE_ETHER_TYPE_IPV4;
                        }
                        break;
 
@@ -1285,7 +1292,7 @@ qede_flow_parse_pattern(__attribute__((unused))struct rte_eth_dev *dev,
                                           spec->hdr.dst_addr,
                                           IPV6_ADDR_LEN);
                                flow->entry.tuple.eth_proto =
-                                       RTE_ETHER_TYPE_IPv6;
+                                       RTE_ETHER_TYPE_IPV6;
                        }
                        break;
 
@@ -1345,7 +1352,6 @@ qede_flow_parse_actions(struct rte_eth_dev *dev,
                        struct rte_flow_error *error,
                        struct rte_flow *flow)
 {
-       struct qede_dev *qdev = QEDE_INIT_QDEV(dev);
        const struct rte_flow_action_queue *queue;
 
        if (actions == NULL) {
@@ -1360,7 +1366,7 @@ qede_flow_parse_actions(struct rte_eth_dev *dev,
                case RTE_FLOW_ACTION_TYPE_QUEUE:
                        queue = actions->conf;
 
-                       if (queue->index >= QEDE_RSS_COUNT(qdev)) {
+                       if (queue->index >= QEDE_RSS_COUNT(dev)) {
                                rte_flow_error_set(error, EINVAL,
                                                   RTE_FLOW_ERROR_TYPE_ACTION,
                                                   actions,
@@ -1372,12 +1378,15 @@ qede_flow_parse_actions(struct rte_eth_dev *dev,
                                flow->entry.rx_queue = queue->index;
 
                        break;
-
+               case RTE_FLOW_ACTION_TYPE_DROP:
+                       if (flow)
+                               flow->entry.is_drop = true;
+                       break;
                default:
                        rte_flow_error_set(error, ENOTSUP,
                                           RTE_FLOW_ERROR_TYPE_ACTION,
                                           actions,
-                                          "Action is not supported - only ACTION_TYPE_QUEUE supported");
+                                          "Action is not supported - only ACTION_TYPE_QUEUE and ACTION_TYPE_DROP supported");
                        return -rte_errno;
                }
        }
@@ -1477,10 +1486,32 @@ qede_flow_destroy(struct rte_eth_dev *eth_dev,
        return rc;
 }
 
+static int
+qede_flow_flush(struct rte_eth_dev *eth_dev,
+               struct rte_flow_error *error)
+{
+       struct qede_dev *qdev = QEDE_INIT_QDEV(eth_dev);
+       struct qede_arfs_entry *tmp = NULL;
+       int rc = 0;
+
+       while (!SLIST_EMPTY(&qdev->arfs_info.arfs_list_head)) {
+               tmp = SLIST_FIRST(&qdev->arfs_info.arfs_list_head);
+
+               rc = qede_config_arfs_filter(eth_dev, tmp, false);
+               if (rc < 0)
+                       rte_flow_error_set(error, rc,
+                                          RTE_FLOW_ERROR_TYPE_HANDLE, NULL,
+                                          "Failed to flush flow filter");
+       }
+
+       return rc;
+}
+
 const struct rte_flow_ops qede_flow_ops = {
        .validate = qede_flow_validate,
        .create = qede_flow_create,
        .destroy = qede_flow_destroy,
+       .flush = qede_flow_flush,
 };
 
 int qede_dev_filter_ctrl(struct rte_eth_dev *eth_dev,