X-Git-Url: http://git.droids-corp.org/?a=blobdiff_plain;f=app%2Ftest%2Ftest_mempool_perf.c;h=ce7c6241ab00a098fd200a69cf3f8cfd3486d3aa;hb=ed579e50c66db90485593c6c79f5366a57145f2a;hp=cdc02a007ca8c37733e591ff37f51301b3427f8f;hpb=2465980ad2e7ccaa223050b00d97b403477b2394;p=dpdk.git diff --git a/app/test/test_mempool_perf.c b/app/test/test_mempool_perf.c index cdc02a007c..ce7c6241ab 100644 --- a/app/test/test_mempool_perf.c +++ b/app/test/test_mempool_perf.c @@ -1,34 +1,6 @@ -/*- - * BSD LICENSE - * - * Copyright(c) 2010-2014 Intel Corporation. All rights reserved. - * All rights reserved. - * - * Redistribution and use in source and binary forms, with or without - * modification, are permitted provided that the following conditions - * are met: - * - * * Redistributions of source code must retain the above copyright - * notice, this list of conditions and the following disclaimer. - * * Redistributions in binary form must reproduce the above copyright - * notice, this list of conditions and the following disclaimer in - * the documentation and/or other materials provided with the - * distribution. - * * Neither the name of Intel Corporation nor the names of its - * contributors may be used to endorse or promote products derived - * from this software without specific prior written permission. - * - * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS - * "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT - * LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR - * A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT - * OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, - * SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT - * LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, - * DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY - * THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT - * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE - * OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. +/* SPDX-License-Identifier: BSD-3-Clause + * Copyright(c) 2010-2014 Intel Corporation + * Copyright(c) 2022 SmartShare Systems */ #include @@ -44,18 +16,16 @@ #include #include #include -#include #include #include #include #include #include -#include #include -#include #include #include #include +#include #include "test.h" @@ -78,39 +48,61 @@ * - One core without cache * - Two cores without cache * - Max. cores without cache + * - One core with user-owned cache + * - Two cores with user-owned cache + * - Max. cores with user-owned cache * * - Bulk size (*n_get_bulk*, *n_put_bulk*) * * - Bulk get from 1 to 32 * - Bulk put from 1 to 32 + * - Bulk get and put from 1 to 32, compile time constant * * - Number of kept objects (*n_keep*) * * - 32 * - 128 + * - 512 */ #define N 65536 #define TIME_S 5 #define MEMPOOL_ELT_SIZE 2048 -#define MAX_KEEP 128 +#define MAX_KEEP 512 #define MEMPOOL_SIZE ((rte_lcore_count()*(MAX_KEEP+RTE_MEMPOOL_CACHE_MAX_SIZE))-1) -static struct rte_mempool *mp; -static struct rte_mempool *mp_cache, *mp_nocache; +/* Number of pointers fitting into one cache line. */ +#define CACHE_LINE_BURST (RTE_CACHE_LINE_SIZE / sizeof(uintptr_t)) + +#define LOG_ERR() printf("test failed at %s():%d\n", __func__, __LINE__) +#define RET_ERR() do { \ + LOG_ERR(); \ + return -1; \ + } while (0) +#define GOTO_ERR(var, label) do { \ + LOG_ERR(); \ + var = -1; \ + goto label; \ + } while (0) -static rte_atomic32_t synchro; +static int use_external_cache; +static unsigned external_cache_size = RTE_MEMPOOL_CACHE_MAX_SIZE; + +static uint32_t synchro; /* number of objects in one bulk operation (get or put) */ static unsigned n_get_bulk; static unsigned n_put_bulk; -/* number of objects retrived from mempool before putting them back */ +/* number of objects retrieved from mempool before putting them back */ static unsigned n_keep; +/* true if we want to test with constant n_get_bulk and n_put_bulk */ +static int use_constant_values; + /* number of enqueues / dequeues */ struct mempool_test_stats { - unsigned enq_count; + uint64_t enq_count; } __rte_cache_aligned; static struct mempool_test_stats stats[RTE_MAX_LCORE]; @@ -120,7 +112,7 @@ static struct mempool_test_stats stats[RTE_MAX_LCORE]; * other bytes are set to 0. */ static void -my_obj_init(struct rte_mempool *mp, __attribute__((unused)) void *arg, +my_obj_init(struct rte_mempool *mp, __rte_unused void *arg, void *obj, unsigned i) { uint32_t *objnum = obj; @@ -128,100 +120,150 @@ my_obj_init(struct rte_mempool *mp, __attribute__((unused)) void *arg, *objnum = i; } +static __rte_always_inline int +test_loop(struct rte_mempool *mp, struct rte_mempool_cache *cache, + unsigned int x_keep, unsigned int x_get_bulk, unsigned int x_put_bulk) +{ + void *obj_table[MAX_KEEP] __rte_cache_aligned; + unsigned int idx; + unsigned int i; + int ret; + + for (i = 0; likely(i < (N / x_keep)); i++) { + /* get x_keep objects by bulk of x_get_bulk */ + for (idx = 0; idx < x_keep; idx += x_get_bulk) { + ret = rte_mempool_generic_get(mp, + &obj_table[idx], + x_get_bulk, + cache); + if (unlikely(ret < 0)) { + rte_mempool_dump(stdout, mp); + return ret; + } + } + + /* put the objects back by bulk of x_put_bulk */ + for (idx = 0; idx < x_keep; idx += x_put_bulk) { + rte_mempool_generic_put(mp, + &obj_table[idx], + x_put_bulk, + cache); + } + } + + return 0; +} + static int -per_lcore_mempool_test(__attribute__((unused)) void *arg) +per_lcore_mempool_test(void *arg) { - void *obj_table[MAX_KEEP]; - unsigned i, idx; + struct rte_mempool *mp = arg; unsigned lcore_id = rte_lcore_id(); - int ret; + int ret = 0; uint64_t start_cycles, end_cycles; uint64_t time_diff = 0, hz = rte_get_timer_hz(); + struct rte_mempool_cache *cache; + + if (use_external_cache) { + /* Create a user-owned mempool cache. */ + cache = rte_mempool_cache_create(external_cache_size, + SOCKET_ID_ANY); + if (cache == NULL) + RET_ERR(); + } else { + /* May be NULL if cache is disabled. */ + cache = rte_mempool_default_cache(mp, lcore_id); + } /* n_get_bulk and n_put_bulk must be divisors of n_keep */ if (((n_keep / n_get_bulk) * n_get_bulk) != n_keep) - return -1; + GOTO_ERR(ret, out); if (((n_keep / n_put_bulk) * n_put_bulk) != n_keep) - return -1; + GOTO_ERR(ret, out); + /* for constant n, n_get_bulk and n_put_bulk must be the same */ + if (use_constant_values && n_put_bulk != n_get_bulk) + GOTO_ERR(ret, out); stats[lcore_id].enq_count = 0; - /* wait synchro for slaves */ - if (lcore_id != rte_get_master_lcore()) - while (rte_atomic32_read(&synchro) == 0); + /* wait synchro for workers */ + if (lcore_id != rte_get_main_lcore()) + rte_wait_until_equal_32(&synchro, 1, __ATOMIC_RELAXED); start_cycles = rte_get_timer_cycles(); while (time_diff/hz < TIME_S) { - for (i = 0; likely(i < (N/n_keep)); i++) { - /* get n_keep objects by bulk of n_bulk */ - idx = 0; - while (idx < n_keep) { - ret = rte_mempool_get_bulk(mp, &obj_table[idx], - n_get_bulk); - if (unlikely(ret < 0)) { - rte_mempool_dump(stdout, mp); - rte_ring_dump(stdout, mp->ring); - /* in this case, objects are lost... */ - return -1; - } - idx += n_get_bulk; - } + if (!use_constant_values) + ret = test_loop(mp, cache, n_keep, n_get_bulk, n_put_bulk); + else if (n_get_bulk == 1) + ret = test_loop(mp, cache, n_keep, 1, 1); + else if (n_get_bulk == 4) + ret = test_loop(mp, cache, n_keep, 4, 4); + else if (n_get_bulk == CACHE_LINE_BURST) + ret = test_loop(mp, cache, n_keep, + CACHE_LINE_BURST, CACHE_LINE_BURST); + else if (n_get_bulk == 32) + ret = test_loop(mp, cache, n_keep, 32, 32); + else + ret = -1; + + if (ret < 0) + GOTO_ERR(ret, out); - /* put the objects back */ - idx = 0; - while (idx < n_keep) { - rte_mempool_put_bulk(mp, &obj_table[idx], - n_put_bulk); - idx += n_put_bulk; - } - } end_cycles = rte_get_timer_cycles(); time_diff = end_cycles - start_cycles; stats[lcore_id].enq_count += N; } - return 0; +out: + if (use_external_cache) { + rte_mempool_cache_flush(cache, mp); + rte_mempool_cache_free(cache); + } + + return ret; } /* launch all the per-lcore test, and display the result */ static int -launch_cores(unsigned cores) +launch_cores(struct rte_mempool *mp, unsigned int cores) { unsigned lcore_id; - unsigned rate; + uint64_t rate; int ret; unsigned cores_save = cores; - rte_atomic32_set(&synchro, 0); + __atomic_store_n(&synchro, 0, __ATOMIC_RELAXED); /* reset stats */ memset(stats, 0, sizeof(stats)); printf("mempool_autotest cache=%u cores=%u n_get_bulk=%u " - "n_put_bulk=%u n_keep=%u ", - (unsigned) mp->cache_size, cores, n_get_bulk, n_put_bulk, n_keep); + "n_put_bulk=%u n_keep=%u constant_n=%u ", + use_external_cache ? + external_cache_size : (unsigned) mp->cache_size, + cores, n_get_bulk, n_put_bulk, n_keep, use_constant_values); - if (rte_mempool_count(mp) != MEMPOOL_SIZE) { + if (rte_mempool_avail_count(mp) != MEMPOOL_SIZE) { printf("mempool is not full\n"); return -1; } - RTE_LCORE_FOREACH_SLAVE(lcore_id) { + RTE_LCORE_FOREACH_WORKER(lcore_id) { if (cores == 1) break; cores--; rte_eal_remote_launch(per_lcore_mempool_test, - NULL, lcore_id); + mp, lcore_id); } - /* start synchro and launch test on master */ - rte_atomic32_set(&synchro, 1); + /* start synchro and launch test on main */ + __atomic_store_n(&synchro, 1, __ATOMIC_RELAXED); - ret = per_lcore_mempool_test(NULL); + ret = per_lcore_mempool_test(mp); cores = cores_save; - RTE_LCORE_FOREACH_SLAVE(lcore_id) { + RTE_LCORE_FOREACH_WORKER(lcore_id) { if (cores == 1) break; cores--; @@ -238,18 +280,18 @@ launch_cores(unsigned cores) for (lcore_id = 0; lcore_id < RTE_MAX_LCORE; lcore_id++) rate += (stats[lcore_id].enq_count / TIME_S); - printf("rate_persec=%u\n", rate); + printf("rate_persec=%" PRIu64 "\n", rate); return 0; } /* for a given number of core, launch all test cases */ static int -do_one_mempool_test(unsigned cores) +do_one_mempool_test(struct rte_mempool *mp, unsigned int cores) { - unsigned bulk_tab_get[] = { 1, 4, 32, 0 }; - unsigned bulk_tab_put[] = { 1, 4, 32, 0 }; - unsigned keep_tab[] = { 32, 128, 0 }; + unsigned int bulk_tab_get[] = { 1, 4, CACHE_LINE_BURST, 32, 0 }; + unsigned int bulk_tab_put[] = { 1, 4, CACHE_LINE_BURST, 32, 0 }; + unsigned int keep_tab[] = { 32, 128, 512, 0 }; unsigned *get_bulk_ptr; unsigned *put_bulk_ptr; unsigned *keep_ptr; @@ -259,13 +301,21 @@ do_one_mempool_test(unsigned cores) for (put_bulk_ptr = bulk_tab_put; *put_bulk_ptr; put_bulk_ptr++) { for (keep_ptr = keep_tab; *keep_ptr; keep_ptr++) { + use_constant_values = 0; n_get_bulk = *get_bulk_ptr; n_put_bulk = *put_bulk_ptr; n_keep = *keep_ptr; - ret = launch_cores(cores); - + ret = launch_cores(mp, cores); if (ret < 0) return -1; + + /* replay test with constant values */ + if (n_get_bulk == n_put_bulk) { + use_constant_values = 1; + ret = launch_cores(mp, cores); + if (ret < 0) + return -1; + } } } } @@ -275,62 +325,116 @@ do_one_mempool_test(unsigned cores) static int test_mempool_perf(void) { - rte_atomic32_init(&synchro); + struct rte_mempool *mp_cache = NULL; + struct rte_mempool *mp_nocache = NULL; + struct rte_mempool *default_pool = NULL; + const char *default_pool_ops; + int ret = -1; /* create a mempool (without cache) */ + mp_nocache = rte_mempool_create("perf_test_nocache", MEMPOOL_SIZE, + MEMPOOL_ELT_SIZE, 0, 0, + NULL, NULL, + my_obj_init, NULL, + SOCKET_ID_ANY, 0); if (mp_nocache == NULL) - mp_nocache = rte_mempool_create("perf_test_nocache", MEMPOOL_SIZE, - MEMPOOL_ELT_SIZE, 0, 0, - NULL, NULL, - my_obj_init, NULL, - SOCKET_ID_ANY, 0); - if (mp_nocache == NULL) - return -1; + goto err; /* create a mempool (with cache) */ + mp_cache = rte_mempool_create("perf_test_cache", MEMPOOL_SIZE, + MEMPOOL_ELT_SIZE, + RTE_MEMPOOL_CACHE_MAX_SIZE, 0, + NULL, NULL, + my_obj_init, NULL, + SOCKET_ID_ANY, 0); if (mp_cache == NULL) - mp_cache = rte_mempool_create("perf_test_cache", MEMPOOL_SIZE, - MEMPOOL_ELT_SIZE, - RTE_MEMPOOL_CACHE_MAX_SIZE, 0, - NULL, NULL, - my_obj_init, NULL, - SOCKET_ID_ANY, 0); - if (mp_cache == NULL) - return -1; + goto err; + + default_pool_ops = rte_mbuf_best_mempool_ops(); + /* Create a mempool based on Default handler */ + default_pool = rte_mempool_create_empty("default_pool", + MEMPOOL_SIZE, + MEMPOOL_ELT_SIZE, + 0, 0, + SOCKET_ID_ANY, 0); + + if (default_pool == NULL) { + printf("cannot allocate %s mempool\n", default_pool_ops); + goto err; + } + + if (rte_mempool_set_ops_byname(default_pool, default_pool_ops, NULL) + < 0) { + printf("cannot set %s handler\n", default_pool_ops); + goto err; + } + + if (rte_mempool_populate_default(default_pool) < 0) { + printf("cannot populate %s mempool\n", default_pool_ops); + goto err; + } + + rte_mempool_obj_iter(default_pool, my_obj_init, NULL); /* performance test with 1, 2 and max cores */ printf("start performance test (without cache)\n"); - mp = mp_nocache; - if (do_one_mempool_test(1) < 0) - return -1; + if (do_one_mempool_test(mp_nocache, 1) < 0) + goto err; - if (do_one_mempool_test(2) < 0) - return -1; + if (do_one_mempool_test(mp_nocache, 2) < 0) + goto err; - if (do_one_mempool_test(rte_lcore_count()) < 0) - return -1; + if (do_one_mempool_test(mp_nocache, rte_lcore_count()) < 0) + goto err; + + /* performance test with 1, 2 and max cores */ + printf("start performance test for %s (without cache)\n", + default_pool_ops); + + if (do_one_mempool_test(default_pool, 1) < 0) + goto err; + + if (do_one_mempool_test(default_pool, 2) < 0) + goto err; + + if (do_one_mempool_test(default_pool, rte_lcore_count()) < 0) + goto err; /* performance test with 1, 2 and max cores */ printf("start performance test (with cache)\n"); - mp = mp_cache; - if (do_one_mempool_test(1) < 0) - return -1; + if (do_one_mempool_test(mp_cache, 1) < 0) + goto err; - if (do_one_mempool_test(2) < 0) - return -1; + if (do_one_mempool_test(mp_cache, 2) < 0) + goto err; - if (do_one_mempool_test(rte_lcore_count()) < 0) - return -1; + if (do_one_mempool_test(mp_cache, rte_lcore_count()) < 0) + goto err; + + /* performance test with 1, 2 and max cores */ + printf("start performance test (with user-owned cache)\n"); + use_external_cache = 1; + + if (do_one_mempool_test(mp_nocache, 1) < 0) + goto err; + + if (do_one_mempool_test(mp_nocache, 2) < 0) + goto err; + + if (do_one_mempool_test(mp_nocache, rte_lcore_count()) < 0) + goto err; rte_mempool_list_dump(stdout); - return 0; + ret = 0; + +err: + rte_mempool_free(mp_cache); + rte_mempool_free(mp_nocache); + rte_mempool_free(default_pool); + return ret; } -static struct test_command mempool_perf_cmd = { - .command = "mempool_perf_autotest", - .callback = test_mempool_perf, -}; -REGISTER_TEST_COMMAND(mempool_perf_cmd); +REGISTER_TEST_COMMAND(mempool_perf_autotest, test_mempool_perf);