-/*
- * BSD LICENSE
- *
- * Copyright (C) Cavium, Inc 2017.
- *
- * Redistribution and use in source and binary forms, with or without
- * modification, are permitted provided that the following conditions
- * are met:
- *
- * * Redistributions of source code must retain the above copyright
- * notice, this list of conditions and the following disclaimer.
- * * Redistributions in binary form must reproduce the above copyright
- * notice, this list of conditions and the following disclaimer in
- * the documentation and/or other materials provided with the
- * distribution.
- * * Neither the name of Cavium, Inc nor the names of its
- * contributors may be used to endorse or promote products derived
- * from this software without specific prior written permission.
- *
- * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
- * "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
- * LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
- * A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
- * OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
- * SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
- * LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
- * DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
- * THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
- * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
- * OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
+/* SPDX-License-Identifier: BSD-3-Clause
+ * Copyright(c) 2017 Cavium, Inc
*/
#include "test_order_common.h"
if (m == NULL)
continue;
- const uint32_t flow = (uintptr_t)m % nb_flows;
+ const flow_id_t flow = (uintptr_t)m % nb_flows;
/* Maintain seq number per flow */
- m->seqn = producer_flow_seq[flow]++;
-
- ev.flow_id = flow;
- ev.mbuf = m;
+ *order_mbuf_seqn(t, m) = producer_flow_seq[flow]++;
+ order_flow_id_save(t, flow, m, &ev);
while (rte_event_enqueue_burst(dev_id, port, &ev, 1) != 1) {
if (t->err)
int
order_opt_check(struct evt_options *opt)
{
- /* 1 producer + N workers + 1 master */
+ if (opt->prod_type != EVT_PROD_TYPE_SYNT) {
+ evt_err("Invalid producer type '%s' valid producer '%s'",
+ evt_prod_id_to_name(opt->prod_type),
+ evt_prod_id_to_name(EVT_PROD_TYPE_SYNT));
+ return -1;
+ }
+
+ /* 1 producer + N workers + main */
if (rte_lcore_count() < 3) {
evt_err("test need minimum 3 lcores");
return -1;
}
/* Validate worker lcores */
- if (evt_lcores_has_overlap(opt->wlcores, rte_get_master_lcore())) {
- evt_err("worker lcores overlaps with master lcore");
+ if (evt_lcores_has_overlap(opt->wlcores, rte_get_main_lcore())) {
+ evt_err("worker lcores overlaps with main lcore");
return -1;
}
}
/* Validate producer lcore */
- if (plcore == (int)rte_get_master_lcore()) {
- evt_err("producer lcore and master lcore should be different");
+ if (plcore == (int)rte_get_main_lcore()) {
+ evt_err("producer lcore and main lcore should be different");
return -1;
}
if (!rte_lcore_is_enabled(plcore)) {
order_test_setup(struct evt_test *test, struct evt_options *opt)
{
void *test_order;
+ struct test_order *t;
+ static const struct rte_mbuf_dynfield flow_id_dynfield_desc = {
+ .name = "test_event_dynfield_flow_id",
+ .size = sizeof(flow_id_t),
+ .align = __alignof__(flow_id_t),
+ };
+ static const struct rte_mbuf_dynfield seqn_dynfield_desc = {
+ .name = "test_event_dynfield_seqn",
+ .size = sizeof(seqn_t),
+ .align = __alignof__(seqn_t),
+ };
test_order = rte_zmalloc_socket(test->name, sizeof(struct test_order),
RTE_CACHE_LINE_SIZE, opt->socket_id);
goto nomem;
}
test->test_priv = test_order;
+ t = evt_test_priv(test);
- struct test_order *t = evt_test_priv(test);
+ t->flow_id_dynfield_offset =
+ rte_mbuf_dynfield_register(&flow_id_dynfield_desc);
+ if (t->flow_id_dynfield_offset < 0) {
+ evt_err("failed to register mbuf field");
+ return -rte_errno;
+ }
+
+ t->seqn_dynfield_offset =
+ rte_mbuf_dynfield_register(&seqn_dynfield_desc);
+ if (t->seqn_dynfield_offset < 0) {
+ evt_err("failed to register mbuf field");
+ return -rte_errno;
+ }
t->producer_flow_seq = rte_zmalloc_socket("test_producer_flow_seq",
sizeof(*t->producer_flow_seq) * opt->nb_flows,
int wkr_idx = 0;
/* launch workers */
- RTE_LCORE_FOREACH_SLAVE(lcore_id) {
+ RTE_LCORE_FOREACH_WORKER(lcore_id) {
if (!(opt->wlcores[lcore_id]))
continue;
rte_event_dev_dump(opt->dev_id, stdout);
evt_err("No schedules for seconds, deadlock");
t->err = true;
- rte_smp_wmb();
break;
}
old_remaining = remaining;
int ret;
uint8_t port;
struct test_order *t = evt_test_priv(test);
+ struct rte_event_dev_info dev_info;
+
+ memset(&dev_info, 0, sizeof(struct rte_event_dev_info));
+ ret = rte_event_dev_info_get(opt->dev_id, &dev_info);
+ if (ret) {
+ evt_err("failed to get eventdev info %d", opt->dev_id);
+ return ret;
+ }
+
+ if (opt->wkr_deq_dep > dev_info.max_event_port_dequeue_depth)
+ opt->wkr_deq_dep = dev_info.max_event_port_dequeue_depth;
/* port configuration */
- const struct rte_event_port_conf wkr_p_conf = {
+ const struct rte_event_port_conf p_conf = {
.dequeue_depth = opt->wkr_deq_dep,
- .enqueue_depth = 64,
- .new_event_threshold = 4096,
+ .enqueue_depth = dev_info.max_event_port_dequeue_depth,
+ .new_event_threshold = dev_info.max_num_events,
};
/* setup one port per worker, linking to all queues */
w->port_id = port;
w->t = t;
- ret = rte_event_port_setup(opt->dev_id, port, &wkr_p_conf);
+ ret = rte_event_port_setup(opt->dev_id, port, &p_conf);
if (ret) {
evt_err("failed to setup port %d", port);
return ret;
return -EINVAL;
}
}
- /* port for producer, no links */
- const struct rte_event_port_conf prod_conf = {
- .dequeue_depth = 8,
- .enqueue_depth = 32,
- .new_event_threshold = 1200,
- };
struct prod_data *p = &t->prod;
p->dev_id = opt->dev_id;
p->queue_id = 0;
p->t = t;
- ret = rte_event_port_setup(opt->dev_id, port, &prod_conf);
+ ret = rte_event_port_setup(opt->dev_id, port, &p_conf);
if (ret) {
evt_err("failed to setup producer port %d", port);
return ret;