return 0;
}
-#ifdef RTE_SCHED_RED
+#ifdef RTE_SCHED_CMAN
#define WRED_SUPPORTED 1
#else
#define WRED_SUPPORTED 0
.sched_sp_n_priorities_max = 1,
.sched_wfq_n_children_per_group_max = UINT32_MAX,
.sched_wfq_n_groups_max = 1,
-#ifdef RTE_SCHED_SUBPORT_TC_OV
.sched_wfq_weight_max = UINT32_MAX,
.sched_wfq_packet_mode_supported = 0,
.sched_wfq_byte_mode_supported = 1,
-#else
- .sched_wfq_weight_max = 1,
- .sched_wfq_packet_mode_supported = 0,
- .sched_wfq_byte_mode_supported = 1,
-#endif
.stats_mask = STATS_MASK_DEFAULT,
} },
return NULL;
}
-#ifdef RTE_SCHED_RED
+#ifdef RTE_SCHED_CMAN
static void
wred_profiles_set(struct rte_eth_dev *dev, uint32_t subport_id)
for (tc_id = 0; tc_id < RTE_SCHED_TRAFFIC_CLASSES_PER_PIPE; tc_id++)
for (color = RTE_COLOR_GREEN; color < RTE_COLORS; color++) {
struct rte_red_params *dst =
- &pp->red_params[tc_id][color];
+ &pp->cman_params->red_params[tc_id][color];
struct tm_wred_profile *src_wp =
tm_tc_wred_profile_get(dev, tc_id);
struct rte_tm_red_params *src =
return 0;
}
-#ifdef RTE_SCHED_SUBPORT_TC_OV
-
static int
update_pipe_weight(struct rte_eth_dev *dev, struct tm_node *np, uint32_t weight)
{
return 0;
}
-#endif
-
static int
update_queue_weight(struct rte_eth_dev *dev,
struct tm_node *nq, uint32_t weight)
rte_strerror(EINVAL));
/* fall-through */
case TM_NODE_LEVEL_PIPE:
-#ifdef RTE_SCHED_SUBPORT_TC_OV
if (update_pipe_weight(dev, n, weight))
return -rte_tm_error_set(error,
EINVAL,
NULL,
rte_strerror(EINVAL));
return 0;
-#else
- return -rte_tm_error_set(error,
- EINVAL,
- RTE_TM_ERROR_TYPE_NODE_WEIGHT,
- NULL,
- rte_strerror(EINVAL));
-#endif
/* fall-through */
case TM_NODE_LEVEL_TC:
return -rte_tm_error_set(error,