4 * Copyright(c) 2017 Intel Corporation. All rights reserved.
6 * Redistribution and use in source and binary forms, with or without
7 * modification, are permitted provided that the following conditions
10 * * Redistributions of source code must retain the above copyright
11 * notice, this list of conditions and the following disclaimer.
12 * * Redistributions in binary form must reproduce the above copyright
13 * notice, this list of conditions and the following disclaimer in
14 * the documentation and/or other materials provided with the
16 * * Neither the name of Intel Corporation nor the names of its
17 * contributors may be used to endorse or promote products derived
18 * from this software without specific prior written permission.
20 * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
21 * "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
22 * LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
23 * A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
24 * OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
25 * SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
26 * LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
27 * DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
28 * THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
29 * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
30 * OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
32 #include <rte_common.h>
33 #include <rte_hexdump.h>
34 #include <rte_cryptodev.h>
35 #include <rte_cryptodev_pmd.h>
36 #include <rte_cryptodev_vdev.h>
38 #include <rte_malloc.h>
39 #include <rte_cpuflags.h>
40 #include <rte_reorder.h>
42 #include "rte_cryptodev_scheduler.h"
43 #include "scheduler_pmd_private.h"
45 uint8_t cryptodev_driver_id;
47 struct scheduler_init_params {
48 struct rte_crypto_vdev_init_params def_p;
50 enum rte_cryptodev_scheduler_mode mode;
51 uint32_t enable_ordering;
53 char slave_names[RTE_CRYPTODEV_SCHEDULER_MAX_NB_SLAVES]
54 [RTE_CRYPTODEV_SCHEDULER_NAME_MAX_LEN];
57 #define RTE_CRYPTODEV_VDEV_NAME ("name")
58 #define RTE_CRYPTODEV_VDEV_SLAVE ("slave")
59 #define RTE_CRYPTODEV_VDEV_MODE ("mode")
60 #define RTE_CRYPTODEV_VDEV_ORDERING ("ordering")
61 #define RTE_CRYPTODEV_VDEV_MAX_NB_QP_ARG ("max_nb_queue_pairs")
62 #define RTE_CRYPTODEV_VDEV_MAX_NB_SESS_ARG ("max_nb_sessions")
63 #define RTE_CRYPTODEV_VDEV_SOCKET_ID ("socket_id")
64 #define RTE_CRYPTODEV_VDEV_COREMASK ("coremask")
65 #define RTE_CRYPTODEV_VDEV_CORELIST ("corelist")
67 const char *scheduler_valid_params[] = {
68 RTE_CRYPTODEV_VDEV_NAME,
69 RTE_CRYPTODEV_VDEV_SLAVE,
70 RTE_CRYPTODEV_VDEV_MODE,
71 RTE_CRYPTODEV_VDEV_ORDERING,
72 RTE_CRYPTODEV_VDEV_MAX_NB_QP_ARG,
73 RTE_CRYPTODEV_VDEV_MAX_NB_SESS_ARG,
74 RTE_CRYPTODEV_VDEV_SOCKET_ID,
75 RTE_CRYPTODEV_VDEV_COREMASK,
76 RTE_CRYPTODEV_VDEV_CORELIST
79 struct scheduler_parse_map {
84 const struct scheduler_parse_map scheduler_mode_map[] = {
85 {RTE_STR(SCHEDULER_MODE_NAME_ROUND_ROBIN),
86 CDEV_SCHED_MODE_ROUNDROBIN},
87 {RTE_STR(SCHEDULER_MODE_NAME_PKT_SIZE_DISTR),
88 CDEV_SCHED_MODE_PKT_SIZE_DISTR},
89 {RTE_STR(SCHEDULER_MODE_NAME_FAIL_OVER),
90 CDEV_SCHED_MODE_FAILOVER},
91 {RTE_STR(SCHEDULER_MODE_NAME_MULTI_CORE),
92 CDEV_SCHED_MODE_MULTICORE}
95 const struct scheduler_parse_map scheduler_ordering_map[] = {
101 cryptodev_scheduler_create(const char *name,
102 struct rte_vdev_device *vdev,
103 struct scheduler_init_params *init_params)
105 struct rte_cryptodev *dev;
106 struct scheduler_ctx *sched_ctx;
110 if (init_params->def_p.name[0] == '\0')
111 snprintf(init_params->def_p.name,
112 sizeof(init_params->def_p.name),
115 dev = rte_cryptodev_vdev_pmd_init(init_params->def_p.name,
116 sizeof(struct scheduler_ctx),
117 init_params->def_p.socket_id,
120 CS_LOG_ERR("driver %s: failed to create cryptodev vdev",
125 dev->driver_id = cryptodev_driver_id;
126 dev->dev_ops = rte_crypto_scheduler_pmd_ops;
128 sched_ctx = dev->data->dev_private;
129 sched_ctx->max_nb_queue_pairs =
130 init_params->def_p.max_nb_queue_pairs;
132 if (init_params->mode == CDEV_SCHED_MODE_MULTICORE) {
135 sched_ctx->nb_wc = 0;
137 for (i = 0; i < RTE_CRYPTODEV_SCHEDULER_MAX_NB_WORKER_CORES; i++) {
138 if (init_params->wcmask & (1ULL << i)) {
139 sched_ctx->wc_pool[sched_ctx->nb_wc++] = i;
141 " Worker core[%u]=%u added\n",
142 sched_ctx->nb_wc-1, i);
147 if (init_params->mode > CDEV_SCHED_MODE_USERDEFINED &&
148 init_params->mode < CDEV_SCHED_MODE_COUNT) {
149 ret = rte_cryptodev_scheduler_mode_set(dev->data->dev_id,
152 rte_cryptodev_pmd_release_device(dev);
156 for (i = 0; i < RTE_DIM(scheduler_mode_map); i++) {
157 if (scheduler_mode_map[i].val != sched_ctx->mode)
160 RTE_LOG(INFO, PMD, " Scheduling mode = %s\n",
161 scheduler_mode_map[i].name);
166 sched_ctx->reordering_enabled = init_params->enable_ordering;
168 for (i = 0; i < RTE_DIM(scheduler_ordering_map); i++) {
169 if (scheduler_ordering_map[i].val !=
170 sched_ctx->reordering_enabled)
173 RTE_LOG(INFO, PMD, " Packet ordering = %s\n",
174 scheduler_ordering_map[i].name);
179 for (i = 0; i < init_params->nb_slaves; i++) {
180 sched_ctx->init_slave_names[sched_ctx->nb_init_slaves] =
183 RTE_CRYPTODEV_SCHEDULER_NAME_MAX_LEN, 0,
186 if (!sched_ctx->init_slave_names[
187 sched_ctx->nb_init_slaves]) {
188 CS_LOG_ERR("driver %s: Insufficient memory",
193 strncpy(sched_ctx->init_slave_names[
194 sched_ctx->nb_init_slaves],
195 init_params->slave_names[i],
196 RTE_CRYPTODEV_SCHEDULER_NAME_MAX_LEN - 1);
198 sched_ctx->nb_init_slaves++;
202 * Initialize capabilities structure as an empty structure,
203 * in case device information is requested when no slaves are attached
205 sched_ctx->capabilities = rte_zmalloc_socket(NULL,
206 sizeof(struct rte_cryptodev_capabilities),
209 if (!sched_ctx->capabilities) {
210 RTE_LOG(ERR, PMD, "Not enough memory for capability "
219 cryptodev_scheduler_remove(struct rte_vdev_device *vdev)
222 struct rte_cryptodev *dev;
223 struct scheduler_ctx *sched_ctx;
228 name = rte_vdev_device_name(vdev);
229 dev = rte_cryptodev_pmd_get_named_dev(name);
233 sched_ctx = dev->data->dev_private;
235 if (sched_ctx->nb_slaves) {
238 for (i = 0; i < sched_ctx->nb_slaves; i++)
239 rte_cryptodev_scheduler_slave_detach(dev->data->dev_id,
240 sched_ctx->slaves[i].dev_id);
243 RTE_LOG(INFO, PMD, "Closing Crypto Scheduler device %s on numa "
244 "socket %u\n", name, rte_socket_id());
249 /** Parse integer from integer argument */
251 parse_integer_arg(const char *key __rte_unused,
252 const char *value, void *extra_args)
254 int *i = (int *) extra_args;
258 CS_LOG_ERR("Argument has to be positive.\n");
265 /** Parse integer from hexadecimal integer argument */
267 parse_coremask_arg(const char *key __rte_unused,
268 const char *value, void *extra_args)
270 struct scheduler_init_params *params = extra_args;
272 params->wcmask = strtoull(value, NULL, 16);
277 /** Parse integer from list of integers argument */
279 parse_corelist_arg(const char *key __rte_unused,
280 const char *value, void *extra_args)
282 struct scheduler_init_params *params = extra_args;
284 params->wcmask = 0ULL;
286 const char *token = value;
288 while (isdigit(token[0])) {
290 unsigned int core = strtoul(token, &rval, 10);
292 params->wcmask |= 1ULL << core;
293 token = (const char *)rval;
294 if (token[0] == '\0')
304 parse_name_arg(const char *key __rte_unused,
305 const char *value, void *extra_args)
307 struct rte_crypto_vdev_init_params *params = extra_args;
309 if (strlen(value) >= RTE_CRYPTODEV_NAME_MAX_LEN - 1) {
310 CS_LOG_ERR("Invalid name %s, should be less than "
311 "%u bytes.\n", value,
312 RTE_CRYPTODEV_NAME_MAX_LEN - 1);
316 strncpy(params->name, value, RTE_CRYPTODEV_NAME_MAX_LEN);
323 parse_slave_arg(const char *key __rte_unused,
324 const char *value, void *extra_args)
326 struct scheduler_init_params *param = extra_args;
328 if (param->nb_slaves >= RTE_CRYPTODEV_SCHEDULER_MAX_NB_SLAVES) {
329 CS_LOG_ERR("Too many slaves.\n");
333 strncpy(param->slave_names[param->nb_slaves++], value,
334 RTE_CRYPTODEV_SCHEDULER_NAME_MAX_LEN - 1);
340 parse_mode_arg(const char *key __rte_unused,
341 const char *value, void *extra_args)
343 struct scheduler_init_params *param = extra_args;
346 for (i = 0; i < RTE_DIM(scheduler_mode_map); i++) {
347 if (strcmp(value, scheduler_mode_map[i].name) == 0) {
348 param->mode = (enum rte_cryptodev_scheduler_mode)
349 scheduler_mode_map[i].val;
354 if (i == RTE_DIM(scheduler_mode_map)) {
355 CS_LOG_ERR("Unrecognized input.\n");
363 parse_ordering_arg(const char *key __rte_unused,
364 const char *value, void *extra_args)
366 struct scheduler_init_params *param = extra_args;
369 for (i = 0; i < RTE_DIM(scheduler_ordering_map); i++) {
370 if (strcmp(value, scheduler_ordering_map[i].name) == 0) {
371 param->enable_ordering =
372 scheduler_ordering_map[i].val;
377 if (i == RTE_DIM(scheduler_ordering_map)) {
378 CS_LOG_ERR("Unrecognized input.\n");
386 scheduler_parse_init_params(struct scheduler_init_params *params,
387 const char *input_args)
389 struct rte_kvargs *kvlist = NULL;
396 kvlist = rte_kvargs_parse(input_args,
397 scheduler_valid_params);
401 ret = rte_kvargs_process(kvlist,
402 RTE_CRYPTODEV_VDEV_MAX_NB_QP_ARG,
404 ¶ms->def_p.max_nb_queue_pairs);
408 ret = rte_kvargs_process(kvlist,
409 RTE_CRYPTODEV_VDEV_MAX_NB_SESS_ARG,
411 ¶ms->def_p.max_nb_sessions);
415 ret = rte_kvargs_process(kvlist, RTE_CRYPTODEV_VDEV_SOCKET_ID,
417 ¶ms->def_p.socket_id);
421 ret = rte_kvargs_process(kvlist, RTE_CRYPTODEV_VDEV_COREMASK,
427 ret = rte_kvargs_process(kvlist, RTE_CRYPTODEV_VDEV_CORELIST,
433 ret = rte_kvargs_process(kvlist, RTE_CRYPTODEV_VDEV_NAME,
439 ret = rte_kvargs_process(kvlist, RTE_CRYPTODEV_VDEV_SLAVE,
440 &parse_slave_arg, params);
444 ret = rte_kvargs_process(kvlist, RTE_CRYPTODEV_VDEV_MODE,
445 &parse_mode_arg, params);
449 ret = rte_kvargs_process(kvlist, RTE_CRYPTODEV_VDEV_ORDERING,
450 &parse_ordering_arg, params);
456 rte_kvargs_free(kvlist);
461 cryptodev_scheduler_probe(struct rte_vdev_device *vdev)
463 struct scheduler_init_params init_params = {
465 RTE_CRYPTODEV_VDEV_DEFAULT_MAX_NB_QUEUE_PAIRS,
466 RTE_CRYPTODEV_VDEV_DEFAULT_MAX_NB_SESSIONS,
471 .mode = CDEV_SCHED_MODE_NOT_SET,
472 .enable_ordering = 0,
473 .slave_names = { {0} }
477 name = rte_vdev_device_name(vdev);
481 scheduler_parse_init_params(&init_params,
482 rte_vdev_device_args(vdev));
484 RTE_LOG(INFO, PMD, "Initialising %s on NUMA node %d\n",
486 init_params.def_p.socket_id);
487 RTE_LOG(INFO, PMD, " Max number of queue pairs = %d\n",
488 init_params.def_p.max_nb_queue_pairs);
489 RTE_LOG(INFO, PMD, " Max number of sessions = %d\n",
490 init_params.def_p.max_nb_sessions);
491 if (init_params.def_p.name[0] != '\0')
492 RTE_LOG(INFO, PMD, " User defined name = %s\n",
493 init_params.def_p.name);
494 if (init_params.wcmask != 0)
495 RTE_LOG(INFO, PMD, " workers core mask = %"PRIx64"\n",
498 return cryptodev_scheduler_create(name,
503 static struct rte_vdev_driver cryptodev_scheduler_pmd_drv = {
504 .probe = cryptodev_scheduler_probe,
505 .remove = cryptodev_scheduler_remove
508 static struct cryptodev_driver scheduler_crypto_drv;
510 RTE_PMD_REGISTER_VDEV(CRYPTODEV_NAME_SCHEDULER_PMD,
511 cryptodev_scheduler_pmd_drv);
512 RTE_PMD_REGISTER_PARAM_STRING(CRYPTODEV_NAME_SCHEDULER_PMD,
513 "max_nb_queue_pairs=<int> "
514 "max_nb_sessions=<int> "
517 RTE_PMD_REGISTER_CRYPTO_DRIVER(scheduler_crypto_drv,
518 cryptodev_scheduler_pmd_drv,
519 cryptodev_driver_id);