ethdev: add encap level to RSS flow API action
[dpdk.git] / app / test-crypto-perf / cperf_test_pmd_cyclecount.c
index 2cc459e..8f76160 100644 (file)
@@ -1,33 +1,5 @@
-/*-
- *   BSD LICENSE
- *
- *   Copyright(c) 2017 Intel Corporation. All rights reserved.
- *
- *   Redistribution and use in source and binary forms, with or without
- *   modification, are permitted provided that the following conditions
- *   are met:
- *
- *     * Redistributions of source code must retain the above copyright
- *       notice, this list of conditions and the following disclaimer.
- *     * Redistributions in binary form must reproduce the above copyright
- *       notice, this list of conditions and the following disclaimer in
- *       the documentation and/or other materials provided with the
- *       distribution.
- *     * Neither the name of Intel Corporation nor the names of its
- *       contributors may be used to endorse or promote products derived
- *       from this software without specific prior written permission.
- *
- *   THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
- *   "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
- *   LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
- *   A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
- *   OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
- *   SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
- *   LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
- *   DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
- *   THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
- *   (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
- *   OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
+/* SPDX-License-Identifier: BSD-3-Clause
+ * Copyright(c) 2017 Intel Corporation
  */
 
 #include <stdbool.h>
@@ -51,12 +23,7 @@ struct cperf_pmd_cyclecount_ctx {
        uint16_t qp_id;
        uint8_t lcore_id;
 
-       struct rte_mempool *pkt_mbuf_pool_in;
-       struct rte_mempool *pkt_mbuf_pool_out;
-       struct rte_mbuf **mbufs_in;
-       struct rte_mbuf **mbufs_out;
-
-       struct rte_mempool *crypto_op_pool;
+       struct rte_mempool *pool;
        struct rte_crypto_op **ops;
        struct rte_crypto_op **ops_processed;
 
@@ -64,6 +31,9 @@ struct cperf_pmd_cyclecount_ctx {
 
        cperf_populate_ops_t populate_ops;
 
+       uint32_t src_buf_offset;
+       uint32_t dst_buf_offset;
+
        const struct cperf_options *options;
        const struct cperf_test_vector *test_vector;
 };
@@ -95,11 +65,9 @@ cperf_pmd_cyclecount_test_free(struct cperf_pmd_cyclecount_ctx *ctx)
                        rte_cryptodev_sym_session_free(ctx->sess);
                }
 
-               cperf_free_common_memory(ctx->options,
-                               ctx->pkt_mbuf_pool_in,
-                               ctx->pkt_mbuf_pool_out,
-                               ctx->mbufs_in, ctx->mbufs_out,
-                               ctx->crypto_op_pool);
+               if (ctx->pool)
+                       rte_mempool_free(ctx->pool);
+
                if (ctx->ops)
                        rte_free(ctx->ops);
 
@@ -143,10 +111,9 @@ cperf_pmd_cyclecount_test_constructor(struct rte_mempool *sess_mp,
        if (ctx->sess == NULL)
                goto err;
 
-       if (cperf_alloc_common_memory(options, test_vector, dev_id, 0,
-                       &ctx->pkt_mbuf_pool_in, &ctx->pkt_mbuf_pool_out,
-                       &ctx->mbufs_in, &ctx->mbufs_out,
-                       &ctx->crypto_op_pool) < 0)
+       if (cperf_alloc_common_memory(options, test_vector, dev_id, qp_id, 0,
+                       &ctx->src_buf_offset, &ctx->dst_buf_offset,
+                       &ctx->pool) < 0)
                goto err;
 
        ctx->ops = rte_malloc("ops", alloc_sz, 0);
@@ -174,6 +141,7 @@ pmd_cyclecount_bench_ops(struct pmd_cyclecount_state *state, uint32_t cur_op,
        uint32_t iter_ops_needed =
                        RTE_MIN(state->opts->nb_descriptors, iter_ops_left);
        uint32_t cur_iter_op;
+       uint32_t imix_idx = 0;
 
        for (cur_iter_op = 0; cur_iter_op < iter_ops_needed;
                        cur_iter_op += test_burst_size) {
@@ -181,18 +149,25 @@ pmd_cyclecount_bench_ops(struct pmd_cyclecount_state *state, uint32_t cur_op,
                                test_burst_size);
                struct rte_crypto_op **ops = &state->ctx->ops[cur_iter_op];
 
-               if (burst_size != rte_crypto_op_bulk_alloc(
-                               state->ctx->crypto_op_pool,
-                               RTE_CRYPTO_OP_TYPE_SYMMETRIC,
-                               ops, burst_size))
-                       return -1;
+               /* Allocate objects containing crypto operations and mbufs */
+               if (rte_mempool_get_bulk(state->ctx->pool, (void **)ops,
+                                       burst_size) != 0) {
+                       RTE_LOG(ERR, USER1,
+                                       "Failed to allocate more crypto operations "
+                                       "from the crypto operation pool.\n"
+                                       "Consider increasing the pool size "
+                                       "with --pool-sz\n");
+                               return -1;
+               }
 
                /* Setup crypto op, attach mbuf etc */
                (state->ctx->populate_ops)(ops,
-                               &state->ctx->mbufs_in[cur_iter_op],
-                               &state->ctx->mbufs_out[cur_iter_op], burst_size,
+                               state->ctx->src_buf_offset,
+                               state->ctx->dst_buf_offset,
+                               burst_size,
                                state->ctx->sess, state->opts,
-                               state->ctx->test_vector, iv_offset);
+                               state->ctx->test_vector, iv_offset,
+                               &imix_idx);
 
 #ifdef CPERF_LINEARIZATION_ENABLE
                /* Check if source mbufs require coalescing */
@@ -204,7 +179,7 @@ pmd_cyclecount_bench_ops(struct pmd_cyclecount_state *state, uint32_t cur_op,
                        }
                }
 #endif /* CPERF_LINEARIZATION_ENABLE */
-               rte_mempool_put_bulk(state->ctx->crypto_op_pool, (void **)ops,
+               rte_mempool_put_bulk(state->ctx->pool, (void **)ops,
                                burst_size);
        }
 
@@ -217,6 +192,7 @@ pmd_cyclecount_build_ops(struct pmd_cyclecount_state *state,
                uint32_t iter_ops_needed, uint16_t test_burst_size)
 {
        uint32_t cur_iter_op;
+       uint32_t imix_idx = 0;
 
        for (cur_iter_op = 0; cur_iter_op < iter_ops_needed;
                        cur_iter_op += test_burst_size) {
@@ -224,18 +200,25 @@ pmd_cyclecount_build_ops(struct pmd_cyclecount_state *state,
                                iter_ops_needed - cur_iter_op, test_burst_size);
                struct rte_crypto_op **ops = &state->ctx->ops[cur_iter_op];
 
-               if (burst_size != rte_crypto_op_bulk_alloc(
-                               state->ctx->crypto_op_pool,
-                               RTE_CRYPTO_OP_TYPE_SYMMETRIC,
-                               ops, burst_size))
-                       return -1;
+               /* Allocate objects containing crypto operations and mbufs */
+               if (rte_mempool_get_bulk(state->ctx->pool, (void **)ops,
+                                       burst_size) != 0) {
+                       RTE_LOG(ERR, USER1,
+                                       "Failed to allocate more crypto operations "
+                                       "from the crypto operation pool.\n"
+                                       "Consider increasing the pool size "
+                                       "with --pool-sz\n");
+                               return -1;
+               }
 
                /* Setup crypto op, attach mbuf etc */
                (state->ctx->populate_ops)(ops,
-                               &state->ctx->mbufs_in[cur_iter_op],
-                               &state->ctx->mbufs_out[cur_iter_op], burst_size,
+                               state->ctx->src_buf_offset,
+                               state->ctx->dst_buf_offset,
+                               burst_size,
                                state->ctx->sess, state->opts,
-                               state->ctx->test_vector, iv_offset);
+                               state->ctx->test_vector, iv_offset,
+                               &imix_idx);
        }
        return 0;
 }
@@ -382,7 +365,7 @@ pmd_cyclecount_bench_burst_sz(
                 * we may not have processed all ops that we allocated, so
                 * free everything we've allocated.
                 */
-               rte_mempool_put_bulk(state->ctx->crypto_op_pool,
+               rte_mempool_put_bulk(state->ctx->pool,
                                (void **)state->ctx->ops, iter_ops_allocd);
        }
 
@@ -421,7 +404,7 @@ cperf_pmd_cyclecount_test_runner(void *test_ctx)
        struct rte_cryptodev_info dev_info;
 
        /* Check if source mbufs require coalescing */
-       if (opts->segments_nb > 1) {
+       if (opts->segments_sz < ctx->options->max_buffer_size) {
                rte_cryptodev_info_get(state.ctx->dev_id, &dev_info);
                if ((dev_info.feature_flags &
                                    RTE_CRYPTODEV_FF_MBUF_SCATTER_GATHER) ==