From: Pavan Nikhilesh Date: Wed, 15 Sep 2021 13:15:20 +0000 (+0530) Subject: eventdev/eth_rx: simplify event vector config X-Git-Url: http://git.droids-corp.org/?a=commitdiff_plain;h=929ebdd543094ab7df018c9c7178b624b5709406;p=dpdk.git eventdev/eth_rx: simplify event vector config Include vector configuration into the structure ``rte_event_eth_rx_adapter_queue_conf`` that is used to configure Rx adapter ethernet device Rx queue parameters. This simplifies event vector configuration as it avoids splitting configuration per Rx queue. Signed-off-by: Pavan Nikhilesh Acked-by: Jay Jayatheerthan Acked-by: Ray Kinsella Acked-by: Jerin Jacob --- diff --git a/app/test-eventdev/test_pipeline_common.c b/app/test-eventdev/test_pipeline_common.c index 2775e72c58..40c06ad652 100644 --- a/app/test-eventdev/test_pipeline_common.c +++ b/app/test-eventdev/test_pipeline_common.c @@ -331,7 +331,6 @@ pipeline_event_rx_adapter_setup(struct evt_options *opt, uint8_t stride, uint16_t prod; struct rte_mempool *vector_pool = NULL; struct rte_event_eth_rx_adapter_queue_conf queue_conf; - struct rte_event_eth_rx_adapter_event_vector_config vec_conf; memset(&queue_conf, 0, sizeof(struct rte_event_eth_rx_adapter_queue_conf)); @@ -397,8 +396,12 @@ pipeline_event_rx_adapter_setup(struct evt_options *opt, uint8_t stride, } if (cap & RTE_EVENT_ETH_RX_ADAPTER_CAP_EVENT_VECTOR) { + queue_conf.vector_sz = opt->vector_size; + queue_conf.vector_timeout_ns = + opt->vector_tmo_nsec; queue_conf.rx_queue_flags |= RTE_EVENT_ETH_RX_ADAPTER_QUEUE_EVENT_VECTOR; + queue_conf.vector_mp = vector_pool; } else { evt_err("Rx adapter doesn't support event vector"); return -EINVAL; @@ -418,17 +421,6 @@ pipeline_event_rx_adapter_setup(struct evt_options *opt, uint8_t stride, return ret; } - if (opt->ena_vector) { - vec_conf.vector_sz = opt->vector_size; - vec_conf.vector_timeout_ns = opt->vector_tmo_nsec; - vec_conf.vector_mp = vector_pool; - if (rte_event_eth_rx_adapter_queue_event_vector_config( - prod, prod, -1, &vec_conf) < 0) { - evt_err("Failed to configure event vectorization for Rx adapter"); - return -EINVAL; - } - } - if (!(cap & RTE_EVENT_ETH_RX_ADAPTER_CAP_INTERNAL_PORT)) { uint32_t service_id = -1U; diff --git a/doc/guides/prog_guide/event_ethernet_rx_adapter.rst b/doc/guides/prog_guide/event_ethernet_rx_adapter.rst index c01e5a9666..0780b6f711 100644 --- a/doc/guides/prog_guide/event_ethernet_rx_adapter.rst +++ b/doc/guides/prog_guide/event_ethernet_rx_adapter.rst @@ -195,12 +195,17 @@ The event devices, ethernet device pairs which support the capability flow characteristics and generate a ``rte_event`` containing ``rte_event_vector`` whose event type is either ``RTE_EVENT_TYPE_ETHDEV_VECTOR`` or ``RTE_EVENT_TYPE_ETH_RX_ADAPTER_VECTOR``. -The aggregation size and timeout are configurable at a queue level and the -maximum, minimum vector sizes and timeouts vary based on the device capability -and can be queried using ``rte_event_eth_rx_adapter_vector_limits_get``. +The maximum, minimum vector sizes and timeouts vary based on the device +capability and can be queried using +``rte_event_eth_rx_adapter_vector_limits_get``. The Rx adapter additionally might include useful data such as ethernet device port and queue identifier in the ``rte_event_vector::port`` and ``rte_event_vector::queue`` and mark ``rte_event_vector::attr_valid`` as true. +The aggregation size and timeout are configurable at a queue level by setting +``rte_event_eth_rx_adapter_queue_conf::vector_sz``, +``rte_event_eth_rx_adapter_queue_conf::vector_timeout_ns`` and +``rte_event_eth_rx_adapter_queue_conf::vector_mp`` when adding queues using +``rte_event_eth_rx_adapter_queue_add``. A loop processing ``rte_event_vector`` containing mbufs is shown below. diff --git a/doc/guides/rel_notes/deprecation.rst b/doc/guides/rel_notes/deprecation.rst index cc2b89850b..32fac9b999 100644 --- a/doc/guides/rel_notes/deprecation.rst +++ b/doc/guides/rel_notes/deprecation.rst @@ -166,6 +166,12 @@ Deprecation Notices values to the function ``rte_event_eth_rx_adapter_queue_add`` using the structure ``rte_event_eth_rx_adapter_queue_add``. +* eventdev: Reserved bytes of ``rte_event_crypto_request`` is a space holder + for ``response_info``. Both should be decoupled for better clarity in + DPDK 21.11. + New space for ``response_info`` can be made by changing + ``rte_event_crypto_metadata`` type to structure from union. + * metrics: The function ``rte_metrics_init`` will have a non-void return in order to notify errors instead of calling ``rte_exit``. diff --git a/drivers/event/cnxk/cn10k_eventdev.c b/drivers/event/cnxk/cn10k_eventdev.c index dec165346c..bfd470cffd 100644 --- a/drivers/event/cnxk/cn10k_eventdev.c +++ b/drivers/event/cnxk/cn10k_eventdev.c @@ -695,81 +695,6 @@ cn10k_sso_rx_adapter_vector_limits( return 0; } -static int -cnxk_sso_rx_adapter_vwqe_enable(struct cnxk_eth_dev *cnxk_eth_dev, - uint16_t port_id, uint16_t rq_id, uint16_t sz, - uint64_t tmo_ns, struct rte_mempool *vmp) -{ - struct roc_nix_rq *rq; - - rq = &cnxk_eth_dev->rqs[rq_id]; - - if (!rq->sso_ena) - return -EINVAL; - if (rq->flow_tag_width == 0) - return -EINVAL; - - rq->vwqe_ena = 1; - rq->vwqe_first_skip = 0; - rq->vwqe_aura_handle = roc_npa_aura_handle_to_aura(vmp->pool_id); - rq->vwqe_max_sz_exp = rte_log2_u32(sz); - rq->vwqe_wait_tmo = - tmo_ns / - ((roc_nix_get_vwqe_interval(&cnxk_eth_dev->nix) + 1) * 100); - rq->tag_mask = (port_id & 0xF) << 20; - rq->tag_mask |= - (((port_id >> 4) & 0xF) | (RTE_EVENT_TYPE_ETHDEV_VECTOR << 4)) - << 24; - - return roc_nix_rq_modify(&cnxk_eth_dev->nix, rq, 0); -} - -static int -cn10k_sso_rx_adapter_vector_config( - const struct rte_eventdev *event_dev, const struct rte_eth_dev *eth_dev, - int32_t rx_queue_id, - const struct rte_event_eth_rx_adapter_event_vector_config *config) -{ - struct cnxk_eth_dev *cnxk_eth_dev; - struct cnxk_sso_evdev *dev; - int i, rc; - - rc = strncmp(eth_dev->device->driver->name, "net_cn10k", 8); - if (rc) - return -EINVAL; - - dev = cnxk_sso_pmd_priv(event_dev); - cnxk_eth_dev = cnxk_eth_pmd_priv(eth_dev); - if (rx_queue_id < 0) { - for (i = 0; i < eth_dev->data->nb_rx_queues; i++) { - cnxk_sso_updt_xae_cnt(dev, config->vector_mp, - RTE_EVENT_TYPE_ETHDEV_VECTOR); - rc = cnxk_sso_xae_reconfigure( - (struct rte_eventdev *)(uintptr_t)event_dev); - rc = cnxk_sso_rx_adapter_vwqe_enable( - cnxk_eth_dev, eth_dev->data->port_id, i, - config->vector_sz, config->vector_timeout_ns, - config->vector_mp); - if (rc) - return -EINVAL; - } - } else { - - cnxk_sso_updt_xae_cnt(dev, config->vector_mp, - RTE_EVENT_TYPE_ETHDEV_VECTOR); - rc = cnxk_sso_xae_reconfigure( - (struct rte_eventdev *)(uintptr_t)event_dev); - rc = cnxk_sso_rx_adapter_vwqe_enable( - cnxk_eth_dev, eth_dev->data->port_id, rx_queue_id, - config->vector_sz, config->vector_timeout_ns, - config->vector_mp); - if (rc) - return -EINVAL; - } - - return 0; -} - static int cn10k_sso_tx_adapter_caps_get(const struct rte_eventdev *dev, const struct rte_eth_dev *eth_dev, uint32_t *caps) @@ -883,8 +808,6 @@ static struct rte_eventdev_ops cn10k_sso_dev_ops = { .eth_rx_adapter_stop = cnxk_sso_rx_adapter_stop, .eth_rx_adapter_vector_limits_get = cn10k_sso_rx_adapter_vector_limits, - .eth_rx_adapter_event_vector_config = - cn10k_sso_rx_adapter_vector_config, .eth_tx_adapter_caps_get = cn10k_sso_tx_adapter_caps_get, .eth_tx_adapter_queue_add = cn10k_sso_tx_adapter_queue_add, diff --git a/drivers/event/cnxk/cnxk_eventdev_adptr.c b/drivers/event/cnxk/cnxk_eventdev_adptr.c index a34efbbfaa..b9a5df0136 100644 --- a/drivers/event/cnxk/cnxk_eventdev_adptr.c +++ b/drivers/event/cnxk/cnxk_eventdev_adptr.c @@ -175,6 +175,35 @@ cnxk_sso_rxq_disable(struct cnxk_eth_dev *cnxk_eth_dev, uint16_t rq_id) return roc_nix_rq_modify(&cnxk_eth_dev->nix, rq, 0); } +static int +cnxk_sso_rx_adapter_vwqe_enable(struct cnxk_eth_dev *cnxk_eth_dev, + uint16_t port_id, uint16_t rq_id, uint16_t sz, + uint64_t tmo_ns, struct rte_mempool *vmp) +{ + struct roc_nix_rq *rq; + + rq = &cnxk_eth_dev->rqs[rq_id]; + + if (!rq->sso_ena) + return -EINVAL; + if (rq->flow_tag_width == 0) + return -EINVAL; + + rq->vwqe_ena = 1; + rq->vwqe_first_skip = 0; + rq->vwqe_aura_handle = roc_npa_aura_handle_to_aura(vmp->pool_id); + rq->vwqe_max_sz_exp = rte_log2_u32(sz); + rq->vwqe_wait_tmo = + tmo_ns / + ((roc_nix_get_vwqe_interval(&cnxk_eth_dev->nix) + 1) * 100); + rq->tag_mask = (port_id & 0xF) << 20; + rq->tag_mask |= + (((port_id >> 4) & 0xF) | (RTE_EVENT_TYPE_ETHDEV_VECTOR << 4)) + << 24; + + return roc_nix_rq_modify(&cnxk_eth_dev->nix, rq, 0); +} + int cnxk_sso_rx_adapter_queue_add( const struct rte_eventdev *event_dev, const struct rte_eth_dev *eth_dev, @@ -202,6 +231,18 @@ cnxk_sso_rx_adapter_queue_add( &queue_conf->ev, !!(queue_conf->rx_queue_flags & RTE_EVENT_ETH_RX_ADAPTER_CAP_OVERRIDE_FLOW_ID)); + if (queue_conf->rx_queue_flags & + RTE_EVENT_ETH_RX_ADAPTER_QUEUE_EVENT_VECTOR) { + cnxk_sso_updt_xae_cnt(dev, queue_conf->vector_mp, + RTE_EVENT_TYPE_ETHDEV_VECTOR); + rc |= cnxk_sso_xae_reconfigure( + (struct rte_eventdev *)(uintptr_t)event_dev); + rc |= cnxk_sso_rx_adapter_vwqe_enable( + cnxk_eth_dev, port, rx_queue_id, + queue_conf->vector_sz, + queue_conf->vector_timeout_ns, + queue_conf->vector_mp); + } rox_nix_fc_npa_bp_cfg(&cnxk_eth_dev->nix, rxq_sp->qconf.mp->pool_id, true, dev->force_ena_bp); diff --git a/lib/eventdev/eventdev_pmd.h b/lib/eventdev/eventdev_pmd.h index 94d99f4903..cc1afe37d5 100644 --- a/lib/eventdev/eventdev_pmd.h +++ b/lib/eventdev/eventdev_pmd.h @@ -663,32 +663,6 @@ typedef int (*eventdev_eth_rx_adapter_vector_limits_get_t)( const struct rte_eventdev *dev, const struct rte_eth_dev *eth_dev, struct rte_event_eth_rx_adapter_vector_limits *limits); -struct rte_event_eth_rx_adapter_event_vector_config; -/** - * Enable event vector on an given Rx queue of a ethernet devices belonging to - * the Rx adapter. - * - * @param dev - * Event device pointer - * - * @param eth_dev - * Ethernet device pointer - * - * @param rx_queue_id - * The Rx queue identifier - * - * @param config - * Pointer to the event vector configuration structure. - * - * @return - * - 0: Success. - * - <0: Error code returned by the driver function. - */ -typedef int (*eventdev_eth_rx_adapter_event_vector_config_t)( - const struct rte_eventdev *dev, const struct rte_eth_dev *eth_dev, - int32_t rx_queue_id, - const struct rte_event_eth_rx_adapter_event_vector_config *config); - typedef uint32_t rte_event_pmd_selftest_seqn_t; extern int rte_event_pmd_selftest_seqn_dynfield_offset; @@ -1114,9 +1088,6 @@ struct rte_eventdev_ops { eventdev_eth_rx_adapter_vector_limits_get_t eth_rx_adapter_vector_limits_get; /**< Get event vector limits for the Rx adapter */ - eventdev_eth_rx_adapter_event_vector_config_t - eth_rx_adapter_event_vector_config; - /**< Configure Rx adapter with event vector */ eventdev_timer_adapter_caps_get_t timer_adapter_caps_get; /**< Get timer adapter capabilities */ diff --git a/lib/eventdev/rte_event_eth_rx_adapter.c b/lib/eventdev/rte_event_eth_rx_adapter.c index b3d94162bb..cc93fcc10b 100644 --- a/lib/eventdev/rte_event_eth_rx_adapter.c +++ b/lib/eventdev/rte_event_eth_rx_adapter.c @@ -1947,6 +1947,24 @@ rxa_add_queue(struct rte_event_eth_rx_adapter *rx_adapter, } else qi_ev->flow_id = 0; + if (conf->rx_queue_flags & + RTE_EVENT_ETH_RX_ADAPTER_QUEUE_EVENT_VECTOR) { + queue_info->ena_vector = 1; + qi_ev->event_type = RTE_EVENT_TYPE_ETH_RX_ADAPTER_VECTOR; + rxa_set_vector_data(queue_info, conf->vector_sz, + conf->vector_timeout_ns, conf->vector_mp, + rx_queue_id, dev_info->dev->data->port_id); + rx_adapter->ena_vector = 1; + rx_adapter->vector_tmo_ticks = + rx_adapter->vector_tmo_ticks ? + RTE_MIN(queue_info->vector_data + .vector_timeout_ticks >> + 1, + rx_adapter->vector_tmo_ticks) : + queue_info->vector_data.vector_timeout_ticks >> + 1; + } + rxa_update_queue(rx_adapter, dev_info, rx_queue_id, 1); if (rxa_polled_queue(dev_info, rx_queue_id)) { rx_adapter->num_rx_polled += !pollq; @@ -1972,42 +1990,6 @@ rxa_add_queue(struct rte_event_eth_rx_adapter *rx_adapter, } } -static void -rxa_sw_event_vector_configure( - struct rte_event_eth_rx_adapter *rx_adapter, uint16_t eth_dev_id, - int rx_queue_id, - const struct rte_event_eth_rx_adapter_event_vector_config *config) -{ - struct eth_device_info *dev_info = &rx_adapter->eth_devices[eth_dev_id]; - struct eth_rx_queue_info *queue_info; - struct rte_event *qi_ev; - - if (rx_queue_id == -1) { - uint16_t nb_rx_queues; - uint16_t i; - - nb_rx_queues = dev_info->dev->data->nb_rx_queues; - for (i = 0; i < nb_rx_queues; i++) - rxa_sw_event_vector_configure(rx_adapter, eth_dev_id, i, - config); - return; - } - - queue_info = &dev_info->rx_queue[rx_queue_id]; - qi_ev = (struct rte_event *)&queue_info->event; - queue_info->ena_vector = 1; - qi_ev->event_type = RTE_EVENT_TYPE_ETH_RX_ADAPTER_VECTOR; - rxa_set_vector_data(queue_info, config->vector_sz, - config->vector_timeout_ns, config->vector_mp, - rx_queue_id, dev_info->dev->data->port_id); - rx_adapter->ena_vector = 1; - rx_adapter->vector_tmo_ticks = - rx_adapter->vector_tmo_ticks ? - RTE_MIN(config->vector_timeout_ns >> 1, - rx_adapter->vector_tmo_ticks) : - config->vector_timeout_ns >> 1; -} - static int rxa_sw_add(struct rte_event_eth_rx_adapter *rx_adapter, uint16_t eth_dev_id, int rx_queue_id, @@ -2322,6 +2304,7 @@ rte_event_eth_rx_adapter_queue_add(uint8_t id, struct rte_event_eth_rx_adapter *rx_adapter; struct rte_eventdev *dev; struct eth_device_info *dev_info; + struct rte_event_eth_rx_adapter_vector_limits limits; RTE_EVENT_ETH_RX_ADAPTER_ID_VALID_OR_ERR_RET(id, -EINVAL); RTE_ETH_VALID_PORTID_OR_ERR_RET(eth_dev_id, -EINVAL); @@ -2349,13 +2332,46 @@ rte_event_eth_rx_adapter_queue_add(uint8_t id, return -EINVAL; } - if ((cap & RTE_EVENT_ETH_RX_ADAPTER_CAP_EVENT_VECTOR) == 0 && - (queue_conf->rx_queue_flags & - RTE_EVENT_ETH_RX_ADAPTER_QUEUE_EVENT_VECTOR)) { - RTE_EDEV_LOG_ERR("Event vectorization is not supported," - " eth port: %" PRIu16 " adapter id: %" PRIu8, - eth_dev_id, id); - return -EINVAL; + if (queue_conf->rx_queue_flags & + RTE_EVENT_ETH_RX_ADAPTER_QUEUE_EVENT_VECTOR) { + + if ((cap & RTE_EVENT_ETH_RX_ADAPTER_CAP_EVENT_VECTOR) == 0) { + RTE_EDEV_LOG_ERR("Event vectorization is not supported," + " eth port: %" PRIu16 + " adapter id: %" PRIu8, + eth_dev_id, id); + return -EINVAL; + } + + ret = rte_event_eth_rx_adapter_vector_limits_get( + rx_adapter->eventdev_id, eth_dev_id, &limits); + if (ret < 0) { + RTE_EDEV_LOG_ERR("Failed to get event device vector limits," + " eth port: %" PRIu16 + " adapter id: %" PRIu8, + eth_dev_id, id); + return -EINVAL; + } + if (queue_conf->vector_sz < limits.min_sz || + queue_conf->vector_sz > limits.max_sz || + queue_conf->vector_timeout_ns < limits.min_timeout_ns || + queue_conf->vector_timeout_ns > limits.max_timeout_ns || + queue_conf->vector_mp == NULL) { + RTE_EDEV_LOG_ERR("Invalid event vector configuration," + " eth port: %" PRIu16 + " adapter id: %" PRIu8, + eth_dev_id, id); + return -EINVAL; + } + if (queue_conf->vector_mp->elt_size < + (sizeof(struct rte_event_vector) + + (sizeof(uintptr_t) * queue_conf->vector_sz))) { + RTE_EDEV_LOG_ERR("Invalid event vector configuration," + " eth port: %" PRIu16 + " adapter id: %" PRIu8, + eth_dev_id, id); + return -EINVAL; + } } if ((cap & RTE_EVENT_ETH_RX_ADAPTER_CAP_MULTI_EVENTQ) == 0 && @@ -2551,83 +2567,6 @@ unlock_ret: return ret; } -int -rte_event_eth_rx_adapter_queue_event_vector_config( - uint8_t id, uint16_t eth_dev_id, int32_t rx_queue_id, - struct rte_event_eth_rx_adapter_event_vector_config *config) -{ - struct rte_event_eth_rx_adapter_vector_limits limits; - struct rte_event_eth_rx_adapter *rx_adapter; - struct rte_eventdev *dev; - uint32_t cap; - int ret; - - RTE_EVENT_ETH_RX_ADAPTER_ID_VALID_OR_ERR_RET(id, -EINVAL); - RTE_ETH_VALID_PORTID_OR_ERR_RET(eth_dev_id, -EINVAL); - - rx_adapter = rxa_id_to_adapter(id); - if ((rx_adapter == NULL) || (config == NULL)) - return -EINVAL; - - dev = &rte_eventdevs[rx_adapter->eventdev_id]; - ret = rte_event_eth_rx_adapter_caps_get(rx_adapter->eventdev_id, - eth_dev_id, &cap); - if (ret) { - RTE_EDEV_LOG_ERR("Failed to get adapter caps edev %" PRIu8 - "eth port %" PRIu16, - id, eth_dev_id); - return ret; - } - - if (!(cap & RTE_EVENT_ETH_RX_ADAPTER_CAP_EVENT_VECTOR)) { - RTE_EDEV_LOG_ERR("Event vectorization is not supported," - " eth port: %" PRIu16 " adapter id: %" PRIu8, - eth_dev_id, id); - return -EINVAL; - } - - ret = rte_event_eth_rx_adapter_vector_limits_get( - rx_adapter->eventdev_id, eth_dev_id, &limits); - if (ret) { - RTE_EDEV_LOG_ERR("Failed to get vector limits edev %" PRIu8 - "eth port %" PRIu16, - rx_adapter->eventdev_id, eth_dev_id); - return ret; - } - - if (config->vector_sz < limits.min_sz || - config->vector_sz > limits.max_sz || - config->vector_timeout_ns < limits.min_timeout_ns || - config->vector_timeout_ns > limits.max_timeout_ns || - config->vector_mp == NULL) { - RTE_EDEV_LOG_ERR("Invalid event vector configuration," - " eth port: %" PRIu16 " adapter id: %" PRIu8, - eth_dev_id, id); - return -EINVAL; - } - if (config->vector_mp->elt_size < - (sizeof(struct rte_event_vector) + - (sizeof(uintptr_t) * config->vector_sz))) { - RTE_EDEV_LOG_ERR("Invalid event vector configuration," - " eth port: %" PRIu16 " adapter id: %" PRIu8, - eth_dev_id, id); - return -EINVAL; - } - - if (cap & RTE_EVENT_ETH_RX_ADAPTER_CAP_INTERNAL_PORT) { - RTE_FUNC_PTR_OR_ERR_RET( - *dev->dev_ops->eth_rx_adapter_event_vector_config, - -ENOTSUP); - ret = dev->dev_ops->eth_rx_adapter_event_vector_config( - dev, &rte_eth_devices[eth_dev_id], rx_queue_id, config); - } else { - rxa_sw_event_vector_configure(rx_adapter, eth_dev_id, - rx_queue_id, config); - } - - return ret; -} - int rte_event_eth_rx_adapter_vector_limits_get( uint8_t dev_id, uint16_t eth_port_id, diff --git a/lib/eventdev/rte_event_eth_rx_adapter.h b/lib/eventdev/rte_event_eth_rx_adapter.h index 182dd2e5dd..3f8b362295 100644 --- a/lib/eventdev/rte_event_eth_rx_adapter.h +++ b/lib/eventdev/rte_event_eth_rx_adapter.h @@ -171,9 +171,6 @@ struct rte_event_eth_rx_adapter_queue_conf { * The event adapter sets ev.event_type to RTE_EVENT_TYPE_ETHDEV in the * enqueued event. */ -}; - -struct rte_event_eth_rx_adapter_event_vector_config { uint16_t vector_sz; /**< * Indicates the maximum number for mbufs to combine and form a vector. @@ -548,33 +545,6 @@ int rte_event_eth_rx_adapter_vector_limits_get( uint8_t dev_id, uint16_t eth_port_id, struct rte_event_eth_rx_adapter_vector_limits *limits); -/** - * Configure event vectorization for a given ethernet device queue, that has - * been added to a event eth Rx adapter. - * - * @param id - * The identifier of the ethernet Rx event adapter. - * - * @param eth_dev_id - * The identifier of the ethernet device. - * - * @param rx_queue_id - * Ethernet device receive queue index. - * If rx_queue_id is -1, then all Rx queues configured for the ethernet device - * are configured with event vectorization. - * - * @param config - * Event vector configuration structure. - * - * @return - * - 0: Success, Receive queue configured correctly. - * - <0: Error code on failure. - */ -__rte_experimental -int rte_event_eth_rx_adapter_queue_event_vector_config( - uint8_t id, uint16_t eth_dev_id, int32_t rx_queue_id, - struct rte_event_eth_rx_adapter_event_vector_config *config); - #ifdef __cplusplus } #endif diff --git a/lib/eventdev/version.map b/lib/eventdev/version.map index 88625621ec..cd86d2d908 100644 --- a/lib/eventdev/version.map +++ b/lib/eventdev/version.map @@ -142,7 +142,6 @@ EXPERIMENTAL { #added in 21.05 rte_event_vector_pool_create; rte_event_eth_rx_adapter_vector_limits_get; - rte_event_eth_rx_adapter_queue_event_vector_config; __rte_eventdev_trace_crypto_adapter_enqueue; };