X-Git-Url: http://git.droids-corp.org/?a=blobdiff_plain;f=drivers%2Fcrypto%2Fscheduler%2Fscheduler_pmd_ops.c;h=cb125e8027a60dd2a98ef21e3b7fdcf82f3cc2fa;hb=a2999c7bfe395708a6a68ebb6f3199af7d25cffe;hp=fef686ff752c0d8b58651d2a4ef91872f87c3fb3;hpb=f2f020d21049cb6439c35276b767d1c42e047618;p=dpdk.git diff --git a/drivers/crypto/scheduler/scheduler_pmd_ops.c b/drivers/crypto/scheduler/scheduler_pmd_ops.c index fef686ff75..cb125e8027 100644 --- a/drivers/crypto/scheduler/scheduler_pmd_ops.c +++ b/drivers/crypto/scheduler/scheduler_pmd_ops.c @@ -1,37 +1,8 @@ -/*- - * BSD LICENSE - * - * Copyright(c) 2017 Intel Corporation. All rights reserved. - * - * Redistribution and use in source and binary forms, with or without - * modification, are permitted provided that the following conditions - * are met: - * - * * Redistributions of source code must retain the above copyright - * notice, this list of conditions and the following disclaimer. - * * Redistributions in binary form must reproduce the above copyright - * notice, this list of conditions and the following disclaimer in - * the documentation and/or other materials provided with the - * distribution. - * * Neither the name of Intel Corporation nor the names of its - * contributors may be used to endorse or promote products derived - * from this software without specific prior written permission. - * - * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS - * "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT - * LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR - * A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT - * OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, - * SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT - * LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, - * DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY - * THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT - * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE - * OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. +/* SPDX-License-Identifier: BSD-3-Clause + * Copyright(c) 2017 Intel Corporation */ #include -#include #include #include #include @@ -41,42 +12,43 @@ #include "scheduler_pmd_private.h" -/** attaching the slaves predefined by scheduler's EAL options */ +/** attaching the workers predefined by scheduler's EAL options */ static int -scheduler_attach_init_slave(struct rte_cryptodev *dev) +scheduler_attach_init_worker(struct rte_cryptodev *dev) { struct scheduler_ctx *sched_ctx = dev->data->dev_private; uint8_t scheduler_id = dev->data->dev_id; int i; - for (i = sched_ctx->nb_init_slaves - 1; i >= 0; i--) { - const char *dev_name = sched_ctx->init_slave_names[i]; - struct rte_cryptodev *slave_dev = + for (i = sched_ctx->nb_init_workers - 1; i >= 0; i--) { + const char *dev_name = sched_ctx->init_worker_names[i]; + struct rte_cryptodev *worker_dev = rte_cryptodev_pmd_get_named_dev(dev_name); int status; - if (!slave_dev) { - CS_LOG_ERR("Failed to locate slave dev %s", + if (!worker_dev) { + CR_SCHED_LOG(ERR, "Failed to locate worker dev %s", dev_name); return -EINVAL; } - status = rte_cryptodev_scheduler_slave_attach( - scheduler_id, slave_dev->data->dev_id); + status = rte_cryptodev_scheduler_worker_attach( + scheduler_id, worker_dev->data->dev_id); if (status < 0) { - CS_LOG_ERR("Failed to attach slave cryptodev %u", - slave_dev->data->dev_id); + CR_SCHED_LOG(ERR, "Failed to attach worker cryptodev %u", + worker_dev->data->dev_id); return status; } - CS_LOG_INFO("Scheduler %s attached slave %s\n", + CR_SCHED_LOG(INFO, "Scheduler %s attached worker %s", dev->data->name, - sched_ctx->init_slave_names[i]); + sched_ctx->init_worker_names[i]); - rte_free(sched_ctx->init_slave_names[i]); + rte_free(sched_ctx->init_worker_names[i]); + sched_ctx->init_worker_names[i] = NULL; - sched_ctx->nb_init_slaves -= 1; + sched_ctx->nb_init_workers -= 1; } return 0; @@ -90,17 +62,17 @@ scheduler_pmd_config(struct rte_cryptodev *dev, uint32_t i; int ret; - /* although scheduler_attach_init_slave presents multiple times, + /* although scheduler_attach_init_worker presents multiple times, * there will be only 1 meaningful execution. */ - ret = scheduler_attach_init_slave(dev); + ret = scheduler_attach_init_worker(dev); if (ret < 0) return ret; - for (i = 0; i < sched_ctx->nb_slaves; i++) { - uint8_t slave_dev_id = sched_ctx->slaves[i].dev_id; + for (i = 0; i < sched_ctx->nb_workers; i++) { + uint8_t worker_dev_id = sched_ctx->workers[i].dev_id; - ret = rte_cryptodev_configure(slave_dev_id, config); + ret = rte_cryptodev_configure(worker_dev_id, config); if (ret < 0) break; } @@ -117,7 +89,7 @@ update_order_ring(struct rte_cryptodev *dev, uint16_t qp_id) if (sched_ctx->reordering_enabled) { char order_ring_name[RTE_CRYPTODEV_NAME_MAX_LEN]; uint32_t buff_size = rte_align32pow2( - sched_ctx->nb_slaves * PER_SLAVE_BUFF_SIZE); + sched_ctx->nb_workers * PER_WORKER_BUFF_SIZE); if (qp_ctx->order_ring) { rte_ring_free(qp_ctx->order_ring); @@ -130,7 +102,7 @@ update_order_ring(struct rte_cryptodev *dev, uint16_t qp_id) if (snprintf(order_ring_name, RTE_CRYPTODEV_NAME_MAX_LEN, "%s_rb_%u_%u", RTE_STR(CRYPTODEV_NAME_SCHEDULER_PMD), dev->data->dev_id, qp_id) < 0) { - CS_LOG_ERR("failed to create unique reorder buffer " + CR_SCHED_LOG(ERR, "failed to create unique reorder buffer" "name"); return -ENOMEM; } @@ -139,7 +111,7 @@ update_order_ring(struct rte_cryptodev *dev, uint16_t qp_id) buff_size, rte_socket_id(), RING_F_SP_ENQ | RING_F_SC_DEQ); if (!qp_ctx->order_ring) { - CS_LOG_ERR("failed to create order ring"); + CR_SCHED_LOG(ERR, "failed to create order ring"); return -ENOMEM; } } else { @@ -163,38 +135,38 @@ scheduler_pmd_start(struct rte_cryptodev *dev) if (dev->data->dev_started) return 0; - /* although scheduler_attach_init_slave presents multiple times, + /* although scheduler_attach_init_worker presents multiple times, * there will be only 1 meaningful execution. */ - ret = scheduler_attach_init_slave(dev); + ret = scheduler_attach_init_worker(dev); if (ret < 0) return ret; for (i = 0; i < dev->data->nb_queue_pairs; i++) { ret = update_order_ring(dev, i); if (ret < 0) { - CS_LOG_ERR("Failed to update reorder buffer"); + CR_SCHED_LOG(ERR, "Failed to update reorder buffer"); return ret; } } if (sched_ctx->mode == CDEV_SCHED_MODE_NOT_SET) { - CS_LOG_ERR("Scheduler mode is not set"); + CR_SCHED_LOG(ERR, "Scheduler mode is not set"); return -1; } - if (!sched_ctx->nb_slaves) { - CS_LOG_ERR("No slave in the scheduler"); + if (!sched_ctx->nb_workers) { + CR_SCHED_LOG(ERR, "No worker in the scheduler"); return -1; } - RTE_FUNC_PTR_OR_ERR_RET(*sched_ctx->ops.slave_attach, -ENOTSUP); + RTE_FUNC_PTR_OR_ERR_RET(*sched_ctx->ops.worker_attach, -ENOTSUP); - for (i = 0; i < sched_ctx->nb_slaves; i++) { - uint8_t slave_dev_id = sched_ctx->slaves[i].dev_id; + for (i = 0; i < sched_ctx->nb_workers; i++) { + uint8_t worker_dev_id = sched_ctx->workers[i].dev_id; - if ((*sched_ctx->ops.slave_attach)(dev, slave_dev_id) < 0) { - CS_LOG_ERR("Failed to attach slave"); + if ((*sched_ctx->ops.worker_attach)(dev, worker_dev_id) < 0) { + CR_SCHED_LOG(ERR, "Failed to attach worker"); return -ENOTSUP; } } @@ -202,20 +174,20 @@ scheduler_pmd_start(struct rte_cryptodev *dev) RTE_FUNC_PTR_OR_ERR_RET(*sched_ctx->ops.scheduler_start, -ENOTSUP); if ((*sched_ctx->ops.scheduler_start)(dev) < 0) { - CS_LOG_ERR("Scheduler start failed"); + CR_SCHED_LOG(ERR, "Scheduler start failed"); return -1; } - /* start all slaves */ - for (i = 0; i < sched_ctx->nb_slaves; i++) { - uint8_t slave_dev_id = sched_ctx->slaves[i].dev_id; - struct rte_cryptodev *slave_dev = - rte_cryptodev_pmd_get_dev(slave_dev_id); + /* start all workers */ + for (i = 0; i < sched_ctx->nb_workers; i++) { + uint8_t worker_dev_id = sched_ctx->workers[i].dev_id; + struct rte_cryptodev *worker_dev = + rte_cryptodev_pmd_get_dev(worker_dev_id); - ret = (*slave_dev->dev_ops->dev_start)(slave_dev); + ret = (*worker_dev->dev_ops->dev_start)(worker_dev); if (ret < 0) { - CS_LOG_ERR("Failed to start slave dev %u", - slave_dev_id); + CR_SCHED_LOG(ERR, "Failed to start worker dev %u", + worker_dev_id); return ret; } } @@ -233,23 +205,23 @@ scheduler_pmd_stop(struct rte_cryptodev *dev) if (!dev->data->dev_started) return; - /* stop all slaves first */ - for (i = 0; i < sched_ctx->nb_slaves; i++) { - uint8_t slave_dev_id = sched_ctx->slaves[i].dev_id; - struct rte_cryptodev *slave_dev = - rte_cryptodev_pmd_get_dev(slave_dev_id); + /* stop all workers first */ + for (i = 0; i < sched_ctx->nb_workers; i++) { + uint8_t worker_dev_id = sched_ctx->workers[i].dev_id; + struct rte_cryptodev *worker_dev = + rte_cryptodev_pmd_get_dev(worker_dev_id); - (*slave_dev->dev_ops->dev_stop)(slave_dev); + (*worker_dev->dev_ops->dev_stop)(worker_dev); } if (*sched_ctx->ops.scheduler_stop) (*sched_ctx->ops.scheduler_stop)(dev); - for (i = 0; i < sched_ctx->nb_slaves; i++) { - uint8_t slave_dev_id = sched_ctx->slaves[i].dev_id; + for (i = 0; i < sched_ctx->nb_workers; i++) { + uint8_t worker_dev_id = sched_ctx->workers[i].dev_id; - if (*sched_ctx->ops.slave_detach) - (*sched_ctx->ops.slave_detach)(dev, slave_dev_id); + if (*sched_ctx->ops.worker_detach) + (*sched_ctx->ops.worker_detach)(dev, worker_dev_id); } } @@ -265,13 +237,13 @@ scheduler_pmd_close(struct rte_cryptodev *dev) if (dev->data->dev_started) return -EBUSY; - /* close all slaves first */ - for (i = 0; i < sched_ctx->nb_slaves; i++) { - uint8_t slave_dev_id = sched_ctx->slaves[i].dev_id; - struct rte_cryptodev *slave_dev = - rte_cryptodev_pmd_get_dev(slave_dev_id); + /* close all workers first */ + for (i = 0; i < sched_ctx->nb_workers; i++) { + uint8_t worker_dev_id = sched_ctx->workers[i].dev_id; + struct rte_cryptodev *worker_dev = + rte_cryptodev_pmd_get_dev(worker_dev_id); - ret = (*slave_dev->dev_ops->dev_close)(slave_dev); + ret = (*worker_dev->dev_ops->dev_close)(worker_dev); if (ret < 0) return ret; } @@ -290,11 +262,15 @@ scheduler_pmd_close(struct rte_cryptodev *dev) } } - if (sched_ctx->private_ctx) + if (sched_ctx->private_ctx) { rte_free(sched_ctx->private_ctx); + sched_ctx->private_ctx = NULL; + } - if (sched_ctx->capabilities) + if (sched_ctx->capabilities) { rte_free(sched_ctx->capabilities); + sched_ctx->capabilities = NULL; + } return 0; } @@ -307,19 +283,19 @@ scheduler_pmd_stats_get(struct rte_cryptodev *dev, struct scheduler_ctx *sched_ctx = dev->data->dev_private; uint32_t i; - for (i = 0; i < sched_ctx->nb_slaves; i++) { - uint8_t slave_dev_id = sched_ctx->slaves[i].dev_id; - struct rte_cryptodev *slave_dev = - rte_cryptodev_pmd_get_dev(slave_dev_id); - struct rte_cryptodev_stats slave_stats = {0}; + for (i = 0; i < sched_ctx->nb_workers; i++) { + uint8_t worker_dev_id = sched_ctx->workers[i].dev_id; + struct rte_cryptodev *worker_dev = + rte_cryptodev_pmd_get_dev(worker_dev_id); + struct rte_cryptodev_stats worker_stats = {0}; - (*slave_dev->dev_ops->stats_get)(slave_dev, &slave_stats); + (*worker_dev->dev_ops->stats_get)(worker_dev, &worker_stats); - stats->enqueued_count += slave_stats.enqueued_count; - stats->dequeued_count += slave_stats.dequeued_count; + stats->enqueued_count += worker_stats.enqueued_count; + stats->dequeued_count += worker_stats.dequeued_count; - stats->enqueue_err_count += slave_stats.enqueue_err_count; - stats->dequeue_err_count += slave_stats.dequeue_err_count; + stats->enqueue_err_count += worker_stats.enqueue_err_count; + stats->dequeue_err_count += worker_stats.dequeue_err_count; } } @@ -330,12 +306,12 @@ scheduler_pmd_stats_reset(struct rte_cryptodev *dev) struct scheduler_ctx *sched_ctx = dev->data->dev_private; uint32_t i; - for (i = 0; i < sched_ctx->nb_slaves; i++) { - uint8_t slave_dev_id = sched_ctx->slaves[i].dev_id; - struct rte_cryptodev *slave_dev = - rte_cryptodev_pmd_get_dev(slave_dev_id); + for (i = 0; i < sched_ctx->nb_workers; i++) { + uint8_t worker_dev_id = sched_ctx->workers[i].dev_id; + struct rte_cryptodev *worker_dev = + rte_cryptodev_pmd_get_dev(worker_dev_id); - (*slave_dev->dev_ops->stats_reset)(slave_dev); + (*worker_dev->dev_ops->stats_reset)(worker_dev); } } @@ -345,34 +321,50 @@ scheduler_pmd_info_get(struct rte_cryptodev *dev, struct rte_cryptodev_info *dev_info) { struct scheduler_ctx *sched_ctx = dev->data->dev_private; - uint32_t max_nb_sessions = sched_ctx->nb_slaves ? - UINT32_MAX : RTE_CRYPTODEV_PMD_DEFAULT_MAX_NB_SESSIONS; + uint32_t max_nb_sess = 0; + uint16_t headroom_sz = 0; + uint16_t tailroom_sz = 0; uint32_t i; if (!dev_info) return; - /* although scheduler_attach_init_slave presents multiple times, + /* although scheduler_attach_init_worker presents multiple times, * there will be only 1 meaningful execution. */ - scheduler_attach_init_slave(dev); + scheduler_attach_init_worker(dev); - for (i = 0; i < sched_ctx->nb_slaves; i++) { - uint8_t slave_dev_id = sched_ctx->slaves[i].dev_id; - struct rte_cryptodev_info slave_info; + for (i = 0; i < sched_ctx->nb_workers; i++) { + uint8_t worker_dev_id = sched_ctx->workers[i].dev_id; + struct rte_cryptodev_info worker_info; + + rte_cryptodev_info_get(worker_dev_id, &worker_info); + uint32_t dev_max_sess = worker_info.sym.max_nb_sessions; + if (dev_max_sess != 0) { + if (max_nb_sess == 0 || dev_max_sess < max_nb_sess) + max_nb_sess = worker_info.sym.max_nb_sessions; + } - rte_cryptodev_info_get(slave_dev_id, &slave_info); - max_nb_sessions = slave_info.sym.max_nb_sessions < - max_nb_sessions ? - slave_info.sym.max_nb_sessions : - max_nb_sessions; + /* Get the max headroom requirement among worker PMDs */ + headroom_sz = worker_info.min_mbuf_headroom_req > + headroom_sz ? + worker_info.min_mbuf_headroom_req : + headroom_sz; + + /* Get the max tailroom requirement among worker PMDs */ + tailroom_sz = worker_info.min_mbuf_tailroom_req > + tailroom_sz ? + worker_info.min_mbuf_tailroom_req : + tailroom_sz; } dev_info->driver_id = dev->driver_id; dev_info->feature_flags = dev->feature_flags; dev_info->capabilities = sched_ctx->capabilities; dev_info->max_nb_queue_pairs = sched_ctx->max_nb_queue_pairs; - dev_info->sym.max_nb_sessions = max_nb_sessions; + dev_info->min_mbuf_headroom_req = headroom_sz; + dev_info->min_mbuf_tailroom_req = tailroom_sz; + dev_info->sym.max_nb_sessions = max_nb_sess; } /** Release queue pair */ @@ -398,8 +390,7 @@ scheduler_pmd_qp_release(struct rte_cryptodev *dev, uint16_t qp_id) /** Setup a queue pair */ static int scheduler_pmd_qp_setup(struct rte_cryptodev *dev, uint16_t qp_id, - const struct rte_cryptodev_qp_conf *qp_conf, int socket_id, - struct rte_mempool *session_pool) + const struct rte_cryptodev_qp_conf *qp_conf, int socket_id) { struct scheduler_ctx *sched_ctx = dev->data->dev_private; struct scheduler_qp_ctx *qp_ctx; @@ -410,7 +401,7 @@ scheduler_pmd_qp_setup(struct rte_cryptodev *dev, uint16_t qp_id, if (snprintf(name, RTE_CRYPTODEV_NAME_MAX_LEN, "CRYTO_SCHE PMD %u QP %u", dev->data->dev_id, qp_id) < 0) { - CS_LOG_ERR("Failed to create unique queue pair name"); + CR_SCHED_LOG(ERR, "Failed to create unique queue pair name"); return -EFAULT; } @@ -418,16 +409,16 @@ scheduler_pmd_qp_setup(struct rte_cryptodev *dev, uint16_t qp_id, if (dev->data->queue_pairs[qp_id] != NULL) scheduler_pmd_qp_release(dev, qp_id); - for (i = 0; i < sched_ctx->nb_slaves; i++) { - uint8_t slave_id = sched_ctx->slaves[i].dev_id; + for (i = 0; i < sched_ctx->nb_workers; i++) { + uint8_t worker_id = sched_ctx->workers[i].dev_id; /* - * All slaves will share the same session mempool + * All workers will share the same session mempool * for session-less operations, so the objects * must be big enough for all the drivers used. */ - ret = rte_cryptodev_queue_pair_setup(slave_id, qp_id, - qp_conf, socket_id, session_pool); + ret = rte_cryptodev_queue_pair_setup(worker_id, qp_id, + qp_conf, socket_id); if (ret < 0) return ret; } @@ -443,19 +434,19 @@ scheduler_pmd_qp_setup(struct rte_cryptodev *dev, uint16_t qp_id, dev->data->queue_pairs[qp_id] = qp_ctx; - /* although scheduler_attach_init_slave presents multiple times, + /* although scheduler_attach_init_worker presents multiple times, * there will be only 1 meaningful execution. */ - ret = scheduler_attach_init_slave(dev); + ret = scheduler_attach_init_worker(dev); if (ret < 0) { - CS_LOG_ERR("Failed to attach slave"); + CR_SCHED_LOG(ERR, "Failed to attach worker"); scheduler_pmd_qp_release(dev, qp_id); return ret; } if (*sched_ctx->ops.config_queue_pair) { if ((*sched_ctx->ops.config_queue_pair)(dev, qp_id) < 0) { - CS_LOG_ERR("Unable to configure queue pair"); + CR_SCHED_LOG(ERR, "Unable to configure queue pair"); return -1; } } @@ -463,41 +454,18 @@ scheduler_pmd_qp_setup(struct rte_cryptodev *dev, uint16_t qp_id, return 0; } -/** Start queue pair */ -static int -scheduler_pmd_qp_start(__rte_unused struct rte_cryptodev *dev, - __rte_unused uint16_t queue_pair_id) -{ - return -ENOTSUP; -} - -/** Stop queue pair */ -static int -scheduler_pmd_qp_stop(__rte_unused struct rte_cryptodev *dev, - __rte_unused uint16_t queue_pair_id) -{ - return -ENOTSUP; -} - -/** Return the number of allocated queue pairs */ -static uint32_t -scheduler_pmd_qp_count(struct rte_cryptodev *dev) -{ - return dev->data->nb_queue_pairs; -} - static uint32_t -scheduler_pmd_session_get_size(struct rte_cryptodev *dev __rte_unused) +scheduler_pmd_sym_session_get_size(struct rte_cryptodev *dev __rte_unused) { struct scheduler_ctx *sched_ctx = dev->data->dev_private; uint8_t i = 0; uint32_t max_priv_sess_size = 0; - /* Check what is the maximum private session size for all slaves */ - for (i = 0; i < sched_ctx->nb_slaves; i++) { - uint8_t slave_dev_id = sched_ctx->slaves[i].dev_id; - struct rte_cryptodev *dev = &rte_cryptodevs[slave_dev_id]; - uint32_t priv_sess_size = (*dev->dev_ops->session_get_size)(dev); + /* Check what is the maximum private session size for all workers */ + for (i = 0; i < sched_ctx->nb_workers; i++) { + uint8_t worker_dev_id = sched_ctx->workers[i].dev_id; + struct rte_cryptodev *dev = &rte_cryptodevs[worker_dev_id]; + uint32_t priv_sess_size = (*dev->dev_ops->sym_session_get_size)(dev); if (max_priv_sess_size < priv_sess_size) max_priv_sess_size = priv_sess_size; @@ -507,7 +475,7 @@ scheduler_pmd_session_get_size(struct rte_cryptodev *dev __rte_unused) } static int -scheduler_pmd_session_configure(struct rte_cryptodev *dev, +scheduler_pmd_sym_session_configure(struct rte_cryptodev *dev, struct rte_crypto_sym_xform *xform, struct rte_cryptodev_sym_session *sess, struct rte_mempool *mempool) @@ -516,13 +484,13 @@ scheduler_pmd_session_configure(struct rte_cryptodev *dev, uint32_t i; int ret; - for (i = 0; i < sched_ctx->nb_slaves; i++) { - struct scheduler_slave *slave = &sched_ctx->slaves[i]; + for (i = 0; i < sched_ctx->nb_workers; i++) { + struct scheduler_worker *worker = &sched_ctx->workers[i]; - ret = rte_cryptodev_sym_session_init(slave->dev_id, sess, + ret = rte_cryptodev_sym_session_init(worker->dev_id, sess, xform, mempool); if (ret < 0) { - CS_LOG_ERR("unabled to config sym session"); + CR_SCHED_LOG(ERR, "unable to config sym session"); return ret; } } @@ -532,21 +500,21 @@ scheduler_pmd_session_configure(struct rte_cryptodev *dev, /** Clear the memory of session so it doesn't leave key material behind */ static void -scheduler_pmd_session_clear(struct rte_cryptodev *dev, +scheduler_pmd_sym_session_clear(struct rte_cryptodev *dev, struct rte_cryptodev_sym_session *sess) { struct scheduler_ctx *sched_ctx = dev->data->dev_private; uint32_t i; - /* Clear private data of slaves */ - for (i = 0; i < sched_ctx->nb_slaves; i++) { - struct scheduler_slave *slave = &sched_ctx->slaves[i]; + /* Clear private data of workers */ + for (i = 0; i < sched_ctx->nb_workers; i++) { + struct scheduler_worker *worker = &sched_ctx->workers[i]; - rte_cryptodev_sym_session_clear(slave->dev_id, sess); + rte_cryptodev_sym_session_clear(worker->dev_id, sess); } } -struct rte_cryptodev_ops scheduler_pmd_ops = { +static struct rte_cryptodev_ops scheduler_pmd_ops = { .dev_configure = scheduler_pmd_config, .dev_start = scheduler_pmd_start, .dev_stop = scheduler_pmd_stop, @@ -559,13 +527,10 @@ struct rte_cryptodev_ops scheduler_pmd_ops = { .queue_pair_setup = scheduler_pmd_qp_setup, .queue_pair_release = scheduler_pmd_qp_release, - .queue_pair_start = scheduler_pmd_qp_start, - .queue_pair_stop = scheduler_pmd_qp_stop, - .queue_pair_count = scheduler_pmd_qp_count, - .session_get_size = scheduler_pmd_session_get_size, - .session_configure = scheduler_pmd_session_configure, - .session_clear = scheduler_pmd_session_clear, + .sym_session_get_size = scheduler_pmd_sym_session_get_size, + .sym_session_configure = scheduler_pmd_sym_session_configure, + .sym_session_clear = scheduler_pmd_sym_session_clear, }; struct rte_cryptodev_ops *rte_crypto_scheduler_pmd_ops = &scheduler_pmd_ops;