net/mlx5: enable more shared code on Windows
[dpdk.git] / app / test-crypto-perf / main.c
index 62ae604..99f86e9 100644 (file)
@@ -9,7 +9,7 @@
 #include <rte_random.h>
 #include <rte_eal.h>
 #include <rte_cryptodev.h>
-#ifdef RTE_LIBRTE_PMD_CRYPTO_SCHEDULER
+#ifdef RTE_CRYPTO_SCHEDULER
 #include <rte_cryptodev_scheduler.h>
 #endif
 
@@ -156,7 +156,14 @@ cperf_initialize_cryptodev(struct cperf_options *opts, uint8_t *enabled_cdevs)
                if (sess_size > max_sess_size)
                        max_sess_size = sess_size;
        }
-
+#ifdef RTE_LIB_SECURITY
+       for (cdev_id = 0; cdev_id < rte_cryptodev_count(); cdev_id++) {
+               sess_size = rte_security_session_get_size(
+                               rte_cryptodev_get_sec_ctx(cdev_id));
+               if (sess_size > max_sess_size)
+                       max_sess_size = sess_size;
+       }
+#endif
        /*
         * Calculate number of needed queue pairs, based on the amount
         * of available number of logical cores and crypto devices.
@@ -170,7 +177,7 @@ cperf_initialize_cryptodev(struct cperf_options *opts, uint8_t *enabled_cdevs)
        for (i = 0; i < enabled_cdev_count &&
                        i < RTE_CRYPTO_MAX_DEVS; i++) {
                cdev_id = enabled_cdevs[i];
-#ifdef RTE_LIBRTE_PMD_CRYPTO_SCHEDULER
+#ifdef RTE_CRYPTO_SCHEDULER
                /*
                 * If multi-core scheduler is used, limit the number
                 * of queue pairs to 1, as there is no way to know
@@ -238,7 +245,7 @@ cperf_initialize_cryptodev(struct cperf_options *opts, uint8_t *enabled_cdevs)
                 */
                if (!strcmp((const char *)opts->device_type,
                                        "crypto_scheduler")) {
-#ifdef RTE_LIBRTE_PMD_CRYPTO_SCHEDULER
+#ifdef RTE_CRYPTO_SCHEDULER
                        uint32_t nb_slaves =
                                rte_cryptodev_scheduler_workers_get(cdev_id,
                                                                NULL);
@@ -247,8 +254,7 @@ cperf_initialize_cryptodev(struct cperf_options *opts, uint8_t *enabled_cdevs)
                                opts->nb_qps * nb_slaves;
 #endif
                } else
-                       sessions_needed = enabled_cdev_count *
-                                               opts->nb_qps * 2;
+                       sessions_needed = enabled_cdev_count * opts->nb_qps;
 
                /*
                 * A single session is required per queue pair
@@ -384,7 +390,7 @@ cperf_check_test_vector(struct cperf_options *opts,
                if (opts->cipher_algo == RTE_CRYPTO_CIPHER_NULL) {
                        if (test_vec->plaintext.data == NULL)
                                return -1;
-               } else if (opts->cipher_algo != RTE_CRYPTO_CIPHER_NULL) {
+               } else {
                        if (test_vec->plaintext.data == NULL)
                                return -1;
                        if (test_vec->plaintext.length < opts->max_buffer_size)
@@ -434,7 +440,7 @@ cperf_check_test_vector(struct cperf_options *opts,
                                return -1;
                        if (test_vec->plaintext.length < opts->max_buffer_size)
                                return -1;
-               } else if (opts->cipher_algo != RTE_CRYPTO_CIPHER_NULL) {
+               } else {
                        if (test_vec->plaintext.data == NULL)
                                return -1;
                        if (test_vec->plaintext.length < opts->max_buffer_size)
@@ -594,7 +600,7 @@ main(int argc, char **argv)
 
        i = 0;
        uint8_t qp_id = 0, cdev_index = 0;
-       RTE_LCORE_FOREACH_SLAVE(lcore_id) {
+       RTE_LCORE_FOREACH_WORKER(lcore_id) {
 
                if (i == total_nb_qps)
                        break;
@@ -658,7 +664,7 @@ main(int argc, char **argv)
                                distribution_total[buffer_size_count - 1];
 
                i = 0;
-               RTE_LCORE_FOREACH_SLAVE(lcore_id) {
+               RTE_LCORE_FOREACH_WORKER(lcore_id) {
 
                        if (i == total_nb_qps)
                                break;
@@ -668,7 +674,7 @@ main(int argc, char **argv)
                        i++;
                }
                i = 0;
-               RTE_LCORE_FOREACH_SLAVE(lcore_id) {
+               RTE_LCORE_FOREACH_WORKER(lcore_id) {
 
                        if (i == total_nb_qps)
                                break;
@@ -688,7 +694,7 @@ main(int argc, char **argv)
 
                while (opts.test_buffer_size <= opts.max_buffer_size) {
                        i = 0;
-                       RTE_LCORE_FOREACH_SLAVE(lcore_id) {
+                       RTE_LCORE_FOREACH_WORKER(lcore_id) {
 
                                if (i == total_nb_qps)
                                        break;
@@ -698,7 +704,7 @@ main(int argc, char **argv)
                                i++;
                        }
                        i = 0;
-                       RTE_LCORE_FOREACH_SLAVE(lcore_id) {
+                       RTE_LCORE_FOREACH_WORKER(lcore_id) {
 
                                if (i == total_nb_qps)
                                        break;
@@ -722,7 +728,7 @@ main(int argc, char **argv)
        }
 
        i = 0;
-       RTE_LCORE_FOREACH_SLAVE(lcore_id) {
+       RTE_LCORE_FOREACH_WORKER(lcore_id) {
 
                if (i == total_nb_qps)
                        break;
@@ -742,7 +748,7 @@ main(int argc, char **argv)
 
 err:
        i = 0;
-       RTE_LCORE_FOREACH_SLAVE(lcore_id) {
+       RTE_LCORE_FOREACH_WORKER(lcore_id) {
                if (i == total_nb_qps)
                        break;