net/hns3: fix configurations of port-level scheduling rate
[dpdk.git] / drivers / net / hns3 / hns3_flow.c
index 05cc95e..0d5dd1a 100644 (file)
@@ -168,9 +168,9 @@ hns3_counter_new(struct rte_eth_dev *dev, uint32_t shared, uint32_t id,
        if (cnt) {
                if (!cnt->shared || cnt->shared != shared)
                        return rte_flow_error_set(error, ENOTSUP,
-                                                 RTE_FLOW_ERROR_TYPE_ACTION,
-                                                 cnt,
-                                                 "Counter id is used,shared flag not match");
+                               RTE_FLOW_ERROR_TYPE_ACTION_CONF,
+                               cnt,
+                               "Counter id is used, shared flag not match");
                cnt->ref_cnt++;
                return 0;
        }
@@ -178,7 +178,7 @@ hns3_counter_new(struct rte_eth_dev *dev, uint32_t shared, uint32_t id,
        cnt = rte_zmalloc("hns3 counter", sizeof(*cnt), 0);
        if (cnt == NULL)
                return rte_flow_error_set(error, ENOMEM,
-                                         RTE_FLOW_ERROR_TYPE_ACTION, cnt,
+                                         RTE_FLOW_ERROR_TYPE_HANDLE, cnt,
                                          "Alloc mem for counter failed");
        cnt->id = id;
        cnt->shared = shared;
@@ -206,13 +206,13 @@ hns3_counter_query(struct rte_eth_dev *dev, struct rte_flow *flow,
        cnt = hns3_counter_lookup(dev, flow->counter_id);
        if (cnt == NULL)
                return rte_flow_error_set(error, EINVAL,
-                                         RTE_FLOW_ERROR_TYPE_UNSPECIFIED, NULL,
+                                         RTE_FLOW_ERROR_TYPE_HANDLE, NULL,
                                          "Can't find counter id");
 
        ret = hns3_get_count(&hns->hw, flow->counter_id, &value);
        if (ret) {
                rte_flow_error_set(error, -ret,
-                                  RTE_FLOW_ERROR_TYPE_UNSPECIFIED,
+                                  RTE_FLOW_ERROR_TYPE_HANDLE,
                                   NULL, "Read counter fail.");
                return ret;
        }
@@ -374,9 +374,9 @@ hns3_handle_actions(struct rte_eth_dev *dev,
                            (const struct rte_flow_action_mark *)actions->conf;
                        if (mark->id >= HNS3_MAX_FILTER_ID)
                                return rte_flow_error_set(error, EINVAL,
-                                                    RTE_FLOW_ERROR_TYPE_ACTION,
-                                                    actions,
-                                                    "Invalid Mark ID");
+                                               RTE_FLOW_ERROR_TYPE_ACTION_CONF,
+                                               actions,
+                                               "Invalid Mark ID");
                        rule->fd_id = mark->id;
                        rule->flags |= HNS3_RULE_FLAG_FDID;
                        break;
@@ -390,9 +390,9 @@ hns3_handle_actions(struct rte_eth_dev *dev,
                        counter_num = pf->fdir.fd_cfg.cnt_num[HNS3_FD_STAGE_1];
                        if (act_count->id >= counter_num)
                                return rte_flow_error_set(error, EINVAL,
-                                                    RTE_FLOW_ERROR_TYPE_ACTION,
-                                                    actions,
-                                                    "Invalid counter id");
+                                               RTE_FLOW_ERROR_TYPE_ACTION_CONF,
+                                               actions,
+                                               "Invalid counter id");
                        rule->act_cnt = *act_count;
                        rule->flags |= HNS3_RULE_FLAG_COUNTER;
                        break;
@@ -556,7 +556,7 @@ hns3_parse_ipv4(const struct rte_flow_item *item, struct hns3_fdir_rule *rule,
                    ipv4_mask->hdr.time_to_live ||
                    ipv4_mask->hdr.hdr_checksum) {
                        return rte_flow_error_set(error, EINVAL,
-                                                 RTE_FLOW_ERROR_TYPE_ITEM,
+                                                 RTE_FLOW_ERROR_TYPE_ITEM_MASK,
                                                  item,
                                                  "Only support src & dst ip,tos,proto in IPV4");
                }
@@ -621,7 +621,7 @@ hns3_parse_ipv6(const struct rte_flow_item *item, struct hns3_fdir_rule *rule,
                if (ipv6_mask->hdr.vtc_flow ||
                    ipv6_mask->hdr.payload_len || ipv6_mask->hdr.hop_limits) {
                        return rte_flow_error_set(error, EINVAL,
-                                                 RTE_FLOW_ERROR_TYPE_ITEM,
+                                                 RTE_FLOW_ERROR_TYPE_ITEM_MASK,
                                                  item,
                                                  "Only support src & dst ip,proto in IPV6");
                }
@@ -681,7 +681,7 @@ hns3_parse_tcp(const struct rte_flow_item *item, struct hns3_fdir_rule *rule,
                    tcp_mask->hdr.rx_win ||
                    tcp_mask->hdr.cksum || tcp_mask->hdr.tcp_urp) {
                        return rte_flow_error_set(error, EINVAL,
-                                                 RTE_FLOW_ERROR_TYPE_ITEM,
+                                                 RTE_FLOW_ERROR_TYPE_ITEM_MASK,
                                                  item,
                                                  "Only support src & dst port in TCP");
                }
@@ -728,7 +728,7 @@ hns3_parse_udp(const struct rte_flow_item *item, struct hns3_fdir_rule *rule,
                udp_mask = item->mask;
                if (udp_mask->hdr.dgram_len || udp_mask->hdr.dgram_cksum) {
                        return rte_flow_error_set(error, EINVAL,
-                                                 RTE_FLOW_ERROR_TYPE_ITEM,
+                                                 RTE_FLOW_ERROR_TYPE_ITEM_MASK,
                                                  item,
                                                  "Only support src & dst port in UDP");
                }
@@ -775,7 +775,7 @@ hns3_parse_sctp(const struct rte_flow_item *item, struct hns3_fdir_rule *rule,
                sctp_mask = item->mask;
                if (sctp_mask->hdr.cksum)
                        return rte_flow_error_set(error, EINVAL,
-                                                 RTE_FLOW_ERROR_TYPE_ITEM,
+                                                 RTE_FLOW_ERROR_TYPE_ITEM_MASK,
                                                  item,
                                                  "Only support src & dst port in SCTP");
 
@@ -920,14 +920,14 @@ hns3_parse_vxlan(const struct rte_flow_item *item, struct hns3_fdir_rule *rule,
 
        if (vxlan_mask->flags)
                return rte_flow_error_set(error, EINVAL,
-                                         RTE_FLOW_ERROR_TYPE_ITEM, item,
+                                         RTE_FLOW_ERROR_TYPE_ITEM_MASK, item,
                                          "Flags is not supported in VxLAN");
 
        /* VNI must be totally masked or not. */
        if (memcmp(vxlan_mask->vni, full_mask, VNI_OR_TNI_LEN) &&
            memcmp(vxlan_mask->vni, zero_mask, VNI_OR_TNI_LEN))
                return rte_flow_error_set(error, EINVAL,
-                                         RTE_FLOW_ERROR_TYPE_ITEM, item,
+                                         RTE_FLOW_ERROR_TYPE_ITEM_MASK, item,
                                          "VNI must be totally masked or not in VxLAN");
        if (vxlan_mask->vni[0]) {
                hns3_set_bit(rule->input_set, OUTER_TUN_VNI, 1);
@@ -971,14 +971,14 @@ hns3_parse_nvgre(const struct rte_flow_item *item, struct hns3_fdir_rule *rule,
 
        if (nvgre_mask->protocol || nvgre_mask->c_k_s_rsvd0_ver)
                return rte_flow_error_set(error, EINVAL,
-                                         RTE_FLOW_ERROR_TYPE_ITEM, item,
+                                         RTE_FLOW_ERROR_TYPE_ITEM_MASK, item,
                                          "Ver/protocal is not supported in NVGRE");
 
        /* TNI must be totally masked or not. */
        if (memcmp(nvgre_mask->tni, full_mask, VNI_OR_TNI_LEN) &&
            memcmp(nvgre_mask->tni, zero_mask, VNI_OR_TNI_LEN))
                return rte_flow_error_set(error, EINVAL,
-                                         RTE_FLOW_ERROR_TYPE_ITEM, item,
+                                         RTE_FLOW_ERROR_TYPE_ITEM_MASK, item,
                                          "TNI must be totally masked or not in NVGRE");
 
        if (nvgre_mask->tni[0]) {
@@ -1025,13 +1025,13 @@ hns3_parse_geneve(const struct rte_flow_item *item, struct hns3_fdir_rule *rule,
 
        if (geneve_mask->ver_opt_len_o_c_rsvd0 || geneve_mask->protocol)
                return rte_flow_error_set(error, EINVAL,
-                                         RTE_FLOW_ERROR_TYPE_ITEM, item,
+                                         RTE_FLOW_ERROR_TYPE_ITEM_MASK, item,
                                          "Ver/protocal is not supported in GENEVE");
        /* VNI must be totally masked or not. */
        if (memcmp(geneve_mask->vni, full_mask, VNI_OR_TNI_LEN) &&
            memcmp(geneve_mask->vni, zero_mask, VNI_OR_TNI_LEN))
                return rte_flow_error_set(error, EINVAL,
-                                         RTE_FLOW_ERROR_TYPE_ITEM, item,
+                                         RTE_FLOW_ERROR_TYPE_ITEM_MASK, item,
                                          "VNI must be totally masked or not in GENEVE");
        if (geneve_mask->vni[0]) {
                hns3_set_bit(rule->input_set, OUTER_TUN_VNI, 1);
@@ -1062,7 +1062,7 @@ hns3_parse_tunnel(const struct rte_flow_item *item, struct hns3_fdir_rule *rule,
                break;
        default:
                return rte_flow_error_set(error, ENOTSUP,
-                                         RTE_FLOW_ERROR_TYPE_HANDLE,
+                                         RTE_FLOW_ERROR_TYPE_ITEM,
                                          NULL, "Unsupported tunnel type!");
        }
        if (ret)
@@ -1116,7 +1116,7 @@ hns3_parse_normal(const struct rte_flow_item *item,
                break;
        default:
                return rte_flow_error_set(error, ENOTSUP,
-                                         RTE_FLOW_ERROR_TYPE_HANDLE,
+                                         RTE_FLOW_ERROR_TYPE_ITEM,
                                          NULL, "Unsupported normal type!");
        }
 
@@ -1132,7 +1132,7 @@ hns3_validate_item(const struct rte_flow_item *item,
 
        if (item->last)
                return rte_flow_error_set(error, ENOTSUP,
-                                         RTE_FLOW_ERROR_TYPE_UNSPECIFIED, item,
+                                         RTE_FLOW_ERROR_TYPE_ITEM_LAST, item,
                                          "Not supported last point for range");
 
        for (i = 0; i < step_mngr.count; i++) {
@@ -1218,7 +1218,7 @@ hns3_parse_fdir_filter(struct rte_eth_dev *dev,
 
        if (dev->data->dev_conf.fdir_conf.mode != RTE_FDIR_MODE_PERFECT)
                return rte_flow_error_set(error, ENOTSUP,
-                                         RTE_FLOW_ERROR_TYPE_ITEM_NUM, NULL,
+                                         RTE_FLOW_ERROR_TYPE_HANDLE, NULL,
                                          "fdir_conf.mode isn't perfect");
 
        step_mngr.items = first_items;
@@ -1356,7 +1356,6 @@ hns3_parse_rss_filter(struct rte_eth_dev *dev,
        const struct rte_flow_action_rss *rss;
        const struct rte_flow_action *act;
        uint32_t act_index = 0;
-       uint64_t flow_types;
        uint16_t n;
 
        NEXT_ITEM_OF_ACTION(act, actions, act_index);
@@ -1364,60 +1363,55 @@ hns3_parse_rss_filter(struct rte_eth_dev *dev,
 
        if (rss == NULL) {
                return rte_flow_error_set(error, EINVAL,
-                                         RTE_FLOW_ERROR_TYPE_ACTION,
+                                         RTE_FLOW_ERROR_TYPE_ACTION_CONF,
                                          act, "no valid queues");
        }
 
+       if (rss->queue_num > RTE_DIM(rss_conf->queue))
+               return rte_flow_error_set(error, ENOTSUP,
+                                         RTE_FLOW_ERROR_TYPE_ACTION_CONF, act,
+                                         "queue number configured exceeds "
+                                         "queue buffer size driver supported");
+
        for (n = 0; n < rss->queue_num; n++) {
-               if (rss->queue[n] < dev->data->nb_rx_queues)
+               if (rss->queue[n] < hw->alloc_rss_size)
                        continue;
                return rte_flow_error_set(error, EINVAL,
-                                         RTE_FLOW_ERROR_TYPE_ACTION,
-                                         act,
-                                         "queue id > max number of queues");
+                                         RTE_FLOW_ERROR_TYPE_ACTION_CONF, act,
+                                         "queue id must be less than queue number allocated to a TC");
        }
 
-       /* Parse flow types of RSS */
        if (!(rss->types & HNS3_ETH_RSS_SUPPORT) && rss->types)
                return rte_flow_error_set(error, EINVAL,
-                                         RTE_FLOW_ERROR_TYPE_ACTION,
+                                         RTE_FLOW_ERROR_TYPE_ACTION_CONF,
                                          act,
                                          "Flow types is unsupported by "
                                          "hns3's RSS");
-
-       flow_types = rss->types & HNS3_ETH_RSS_SUPPORT;
-       if (flow_types != rss->types)
-               hns3_warn(hw, "RSS flow types(%" PRIx64 ") include unsupported "
-                         "flow types", rss->types);
-
-       /* Parse RSS related parameters from RSS configuration */
-       switch (rss->func) {
-       case RTE_ETH_HASH_FUNCTION_DEFAULT:
-       case RTE_ETH_HASH_FUNCTION_TOEPLITZ:
-       case RTE_ETH_HASH_FUNCTION_SIMPLE_XOR:
-       case RTE_ETH_HASH_FUNCTION_SYMMETRIC_TOEPLITZ:
-               break;
-       default:
+       if (rss->func >= RTE_ETH_HASH_FUNCTION_MAX)
                return rte_flow_error_set(error, ENOTSUP,
-                                         RTE_FLOW_ERROR_TYPE_ACTION, act,
-                                         "input RSS hash functions are not supported");
-       }
-
+                                         RTE_FLOW_ERROR_TYPE_ACTION_CONF, act,
+                                         "RSS hash func are not supported");
        if (rss->level)
                return rte_flow_error_set(error, ENOTSUP,
-                                         RTE_FLOW_ERROR_TYPE_ACTION, act,
+                                         RTE_FLOW_ERROR_TYPE_ACTION_CONF, act,
                                          "a nonzero RSS encapsulation level is not supported");
        if (rss->key_len && rss->key_len != RTE_DIM(rss_conf->key))
                return rte_flow_error_set(error, ENOTSUP,
-                                         RTE_FLOW_ERROR_TYPE_ACTION, act,
+                                         RTE_FLOW_ERROR_TYPE_ACTION_CONF, act,
                                          "RSS hash key must be exactly 40 bytes");
-       if (rss->queue_num > RTE_DIM(rss_conf->queue))
-               return rte_flow_error_set(error, ENOTSUP,
-                                         RTE_FLOW_ERROR_TYPE_ACTION, act,
-                                         "too many queues for RSS context");
 
+       /*
+        * For Kunpeng920 and Kunpeng930 NIC hardware, it is not supported to
+        * use dst port/src port fields to RSS hash for the following packet
+        * types.
+        * - IPV4 FRAG | IPV4 NONFRAG | IPV6 FRAG | IPV6 NONFRAG
+        * Besides, for Kunpeng920, The NIC hardware is not supported to use
+        * src/dst port fields to RSS hash for IPV6 SCTP packet type.
+        */
        if (rss->types & (ETH_RSS_L4_DST_ONLY | ETH_RSS_L4_SRC_ONLY) &&
-           (rss->types & ETH_RSS_IP))
+          (rss->types & ETH_RSS_IP ||
+          (!hw->rss_info.ipv6_sctp_offload_supported &&
+          rss->types & ETH_RSS_NONFRAG_IPV6_SCTP)))
                return rte_flow_error_set(error, EINVAL,
                                          RTE_FLOW_ERROR_TYPE_ACTION_CONF,
                                          &rss->types,
@@ -1457,9 +1451,8 @@ hns3_disable_rss(struct hns3_hw *hw)
 static void
 hns3_parse_rss_key(struct hns3_hw *hw, struct rte_flow_action_rss *rss_conf)
 {
-       if (rss_conf->key == NULL ||
-           rss_conf->key_len < HNS3_RSS_KEY_SIZE) {
-               hns3_info(hw, "Default RSS hash key to be set");
+       if (rss_conf->key == NULL || rss_conf->key_len < HNS3_RSS_KEY_SIZE) {
+               hns3_warn(hw, "Default RSS hash key to be set");
                rss_conf->key = hns3_hash_key;
                rss_conf->key_len = HNS3_RSS_KEY_SIZE;
        }
@@ -1500,10 +1493,8 @@ hns3_hw_rss_hash_set(struct hns3_hw *hw, struct rte_flow_action_rss *rss_config)
        struct hns3_rss_tuple_cfg *tuple;
        int ret;
 
-       /* Parse hash key */
        hns3_parse_rss_key(hw, rss_config);
 
-       /* Parse hash algorithm */
        ret = hns3_parse_rss_algorithm(hw, &rss_config->func,
                                       &hw->rss_info.hash_algo);
        if (ret)
@@ -1531,26 +1522,22 @@ hns3_update_indir_table(struct rte_eth_dev *dev,
 {
        struct hns3_adapter *hns = dev->data->dev_private;
        struct hns3_hw *hw = &hns->hw;
-       uint8_t indir_tbl[HNS3_RSS_IND_TBL_SIZE];
-       uint16_t j, allow_rss_queues;
-       uint8_t queue_id;
+       uint16_t indir_tbl[HNS3_RSS_IND_TBL_SIZE];
+       uint16_t j;
        uint32_t i;
 
-       allow_rss_queues = RTE_MIN(dev->data->nb_rx_queues, hw->rss_size_max);
        /* Fill in redirection table */
        memcpy(indir_tbl, hw->rss_info.rss_indirection_tbl,
-              HNS3_RSS_IND_TBL_SIZE);
+              sizeof(hw->rss_info.rss_indirection_tbl));
        for (i = 0, j = 0; i < HNS3_RSS_IND_TBL_SIZE; i++, j++) {
                j %= num;
-               if (conf->queue[j] >= allow_rss_queues) {
-                       hns3_err(hw, "Invalid queue id(%u) to be set in "
-                                    "redirection table, max number of rss "
-                                    "queues: %u", conf->queue[j],
-                                allow_rss_queues);
+               if (conf->queue[j] >= hw->alloc_rss_size) {
+                       hns3_err(hw, "queue id(%u) set to redirection table "
+                                "exceeds queue number(%u) allocated to a TC.",
+                                conf->queue[j], hw->alloc_rss_size);
                        return -EINVAL;
                }
-               queue_id = conf->queue[j];
-               indir_tbl[i] = queue_id;
+               indir_tbl[i] = conf->queue[j];
        }
 
        return hns3_set_rss_indir_table(hw, indir_tbl, HNS3_RSS_IND_TBL_SIZE);
@@ -1616,11 +1603,8 @@ hns3_config_rss_filter(struct rte_eth_dev *dev,
                return 0;
        }
 
-       /* Get rx queues num */
-       num = dev->data->nb_rx_queues;
-
        /* Set rx queues to use */
-       num = RTE_MIN(num, rss_flow_conf.queue_num);
+       num = RTE_MIN(dev->data->nb_rx_queues, rss_flow_conf.queue_num);
        if (rss_flow_conf.queue_num > num)
                hns3_warn(hw, "Config queue numbers %u are beyond the scope of truncated",
                          rss_flow_conf.queue_num);
@@ -1657,7 +1641,6 @@ rss_config_err:
        return ret;
 }
 
-/* Remove the rss filter */
 static int
 hns3_clear_rss_filter(struct rte_eth_dev *dev)
 {
@@ -1693,7 +1676,6 @@ hns3_clear_rss_filter(struct rte_eth_dev *dev)
        return ret;
 }
 
-/* Restore the rss filter */
 int
 hns3_restore_rss_filter(struct rte_eth_dev *dev)
 {
@@ -1715,7 +1697,6 @@ hns3_flow_parse_rss(struct rte_eth_dev *dev,
        struct hns3_hw *hw = &hns->hw;
        bool ret;
 
-       /* Action rss same */
        ret = hns3_action_rss_same(&hw->rss_info.conf, &conf->conf);
        if (ret) {
                hns3_err(hw, "Enter duplicate RSS configuration : %d", ret);
@@ -1839,8 +1820,8 @@ hns3_flow_create(struct rte_eth_dev *dev, const struct rte_flow_attr *attr,
                        ret = -ENOMEM;
                        goto err;
                }
-               memcpy(&rss_filter_ptr->filter_info, rss_conf,
-                       sizeof(struct hns3_rss_conf));
+               hns3_rss_conf_copy(&rss_filter_ptr->filter_info,
+                                  &rss_conf->conf);
                rss_filter_ptr->filter_info.valid = true;
                TAILQ_INSERT_TAIL(&process_list->filter_rss_list,
                                  rss_filter_ptr, entries);
@@ -1873,6 +1854,7 @@ hns3_flow_create(struct rte_eth_dev *dev, const struct rte_flow_attr *attr,
                        ret = -ENOMEM;
                        goto err_fdir;
                }
+
                memcpy(&fdir_rule_ptr->fdir_conf, &fdir_rule,
                        sizeof(struct hns3_fdir_rule));
                TAILQ_INSERT_TAIL(&process_list->fdir_list,