struct rte_event_port_conf *port_conf = arg;
struct rte_event_crypto_adapter *adapter = eca_id_to_adapter(id);
+ if (adapter == NULL)
+ return -EINVAL;
+
dev = &rte_eventdevs[adapter->eventdev_id];
dev_conf = dev->data->dev_conf;
return ret;
}
-int __rte_experimental
+int
rte_event_crypto_adapter_create_ext(uint8_t id, uint8_t dev_id,
rte_event_crypto_adapter_conf_cb conf_cb,
enum rte_event_crypto_adapter_mode mode,
}
-int __rte_experimental
+int
rte_event_crypto_adapter_create(uint8_t id, uint8_t dev_id,
struct rte_event_port_conf *port_config,
enum rte_event_crypto_adapter_mode mode)
return ret;
}
-int __rte_experimental
+int
rte_event_crypto_adapter_free(uint8_t id)
{
struct rte_event_crypto_adapter *adapter;
cdev_id = m_data->request_info.cdev_id;
qp_id = m_data->request_info.queue_pair_id;
qp_info = &adapter->cdevs[cdev_id].qpairs[qp_id];
- if (qp_info == NULL) {
+ if (!qp_info->qp_enabled) {
rte_pktmbuf_free(crypto_op->sym->m_src);
rte_crypto_op_free(crypto_op);
continue;
cdev_id = m_data->request_info.cdev_id;
qp_id = m_data->request_info.queue_pair_id;
qp_info = &adapter->cdevs[cdev_id].qpairs[qp_id];
- if (qp_info == NULL) {
+ if (!qp_info->qp_enabled) {
rte_pktmbuf_free(crypto_op->sym->m_src);
rte_crypto_op_free(crypto_op);
continue;
ret = 0;
for (cdev_id = 0; cdev_id < num_cdev; cdev_id++) {
curr_dev = &adapter->cdevs[cdev_id];
- if (curr_dev == NULL)
- continue;
dev = curr_dev->dev;
-
+ if (dev == NULL)
+ continue;
for (qp = 0; qp < dev->data->nb_queue_pairs; qp++) {
curr_queue = &curr_dev->qpairs[qp];
for (cdev_id = adapter->next_cdev_id;
cdev_id < num_cdev; cdev_id++) {
curr_dev = &adapter->cdevs[cdev_id];
- if (curr_dev == NULL)
- continue;
dev = curr_dev->dev;
+ if (dev == NULL)
+ continue;
dev_qps = dev->data->nb_queue_pairs;
for (qp = curr_dev->next_queue_pair_id;
return 0;
}
-int __rte_experimental
+int
rte_event_crypto_adapter_queue_pair_add(uint8_t id,
uint8_t cdev_id,
int32_t queue_pair_id,
return 0;
}
-int __rte_experimental
+int
rte_event_crypto_adapter_queue_pair_del(uint8_t id, uint8_t cdev_id,
int32_t queue_pair_id)
{
return 0;
}
-int __rte_experimental
+int
rte_event_crypto_adapter_start(uint8_t id)
{
struct rte_event_crypto_adapter *adapter;
return eca_adapter_ctrl(id, 1);
}
-int __rte_experimental
+int
rte_event_crypto_adapter_stop(uint8_t id)
{
return eca_adapter_ctrl(id, 0);
}
-int __rte_experimental
+int
rte_event_crypto_adapter_stats_get(uint8_t id,
struct rte_event_crypto_adapter_stats *stats)
{
return 0;
}
-int __rte_experimental
+int
rte_event_crypto_adapter_stats_reset(uint8_t id)
{
struct rte_event_crypto_adapter *adapter;
return 0;
}
-int __rte_experimental
+int
rte_event_crypto_adapter_service_id_get(uint8_t id, uint32_t *service_id)
{
struct rte_event_crypto_adapter *adapter;
return adapter->service_inited ? 0 : -ESRCH;
}
-int __rte_experimental
+int
rte_event_crypto_adapter_event_port_get(uint8_t id, uint8_t *event_port_id)
{
struct rte_event_crypto_adapter *adapter;