X-Git-Url: http://git.droids-corp.org/?a=blobdiff_plain;f=app%2Ftest%2Ftest_func_reentrancy.c;h=baa01ffc2c00a2b35bb0a6247ca0a800afc9aa60;hb=dd0eedb1cfcf0cb7423d859177c5bc6f931eaf8a;hp=15d4dae153fac880d8418f9f3d7eeb84866b77c6;hpb=1c1d4d7a923d4804f1926fc5264f9ecdd8977b04;p=dpdk.git diff --git a/app/test/test_func_reentrancy.c b/app/test/test_func_reentrancy.c index 15d4dae153..baa01ffc2c 100644 --- a/app/test/test_func_reentrancy.c +++ b/app/test/test_func_reentrancy.c @@ -1,13 +1,13 @@ /*- * BSD LICENSE - * - * Copyright(c) 2010-2013 Intel Corporation. All rights reserved. + * + * Copyright(c) 2010-2014 Intel Corporation. All rights reserved. * All rights reserved. - * + * * Redistribution and use in source and binary forms, with or without * modification, are permitted provided that the following conditions * are met: - * + * * * Redistributions of source code must retain the above copyright * notice, this list of conditions and the following disclaimer. * * Redistributions in binary form must reproduce the above copyright @@ -17,7 +17,7 @@ * * Neither the name of Intel Corporation nor the names of its * contributors may be used to endorse or promote products derived * from this software without specific prior written permission. - * + * * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS * "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT * LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR @@ -47,7 +47,6 @@ #include #include #include -#include #include #include #include @@ -70,8 +69,6 @@ #include -#include - #include "test.h" typedef int (*case_func_t)(void* arg); @@ -81,11 +78,12 @@ typedef void (*case_clean_t)(unsigned lcore_id); #define MAX_ITER_TIMES (16) #define MAX_LPM_ITER_TIMES (8) -#define MEMPOOL_ELT_SIZE (0) +#define MEMPOOL_ELT_SIZE (sizeof(uint32_t)) #define MEMPOOL_SIZE (4) #define MAX_LCORES RTE_MAX_MEMZONE / (MAX_ITER_TIMES * 4U) +static rte_atomic32_t obj_count = RTE_ATOMIC32_INIT(0); static rte_atomic32_t synchro = RTE_ATOMIC32_INIT(0); #define WAIT_SYNCHRO_FOR_SLAVES() do{ \ @@ -94,7 +92,7 @@ static rte_atomic32_t synchro = RTE_ATOMIC32_INIT(0); } while(0) /* - * rte_eal_init only init once + * rte_eal_init only init once */ static int test_eal_init_once(__attribute__((unused)) void *arg) @@ -103,6 +101,7 @@ test_eal_init_once(__attribute__((unused)) void *arg) WAIT_SYNCHRO_FOR_SLAVES(); + rte_atomic32_set(&obj_count, 1); /* silent the check in the caller */ if (rte_eal_init(0, NULL) != -1) return -1; @@ -121,17 +120,17 @@ ring_create_lookup(__attribute__((unused)) void *arg) int i; WAIT_SYNCHRO_FOR_SLAVES(); - + /* create the same ring simultaneously on all threads */ for (i = 0; i < MAX_ITER_TIMES; i++) { rp = rte_ring_create("fr_test_once", 4096, SOCKET_ID_ANY, 0); - if ((NULL == rp) && (rte_ring_lookup("fr_test_once") == NULL)) - return -1; + if (rp != NULL) + rte_atomic32_inc(&obj_count); } /* create/lookup new ring several times */ for (i = 0; i < MAX_ITER_TIMES; i++) { - rte_snprintf(ring_name, sizeof(ring_name), "fr_test_%d_%d", lcore_self, i); + snprintf(ring_name, sizeof(ring_name), "fr_test_%d_%d", lcore_self, i); rp = rte_ring_create(ring_name, 4096, SOCKET_ID_ANY, 0); if (NULL == rp) return -1; @@ -141,7 +140,7 @@ ring_create_lookup(__attribute__((unused)) void *arg) /* verify all ring created sucessful */ for (i = 0; i < MAX_ITER_TIMES; i++) { - rte_snprintf(ring_name, sizeof(ring_name), "fr_test_%d_%d", lcore_self, i); + snprintf(ring_name, sizeof(ring_name), "fr_test_%d_%d", lcore_self, i); if (rte_ring_lookup(ring_name) == NULL) return -1; } @@ -175,13 +174,13 @@ mempool_create_lookup(__attribute__((unused)) void *arg) NULL, NULL, my_obj_init, NULL, SOCKET_ID_ANY, 0); - if ((NULL == mp) && (rte_mempool_lookup("fr_test_once") == NULL)) - return -1; + if (mp != NULL) + rte_atomic32_inc(&obj_count); } /* create/lookup new ring several times */ for (i = 0; i < MAX_ITER_TIMES; i++) { - rte_snprintf(mempool_name, sizeof(mempool_name), "fr_test_%d_%d", lcore_self, i); + snprintf(mempool_name, sizeof(mempool_name), "fr_test_%d_%d", lcore_self, i); mp = rte_mempool_create(mempool_name, MEMPOOL_SIZE, MEMPOOL_ELT_SIZE, 0, 0, NULL, NULL, @@ -195,7 +194,7 @@ mempool_create_lookup(__attribute__((unused)) void *arg) /* verify all ring created sucessful */ for (i = 0; i < MAX_ITER_TIMES; i++) { - rte_snprintf(mempool_name, sizeof(mempool_name), "fr_test_%d_%d", lcore_self, i); + snprintf(mempool_name, sizeof(mempool_name), "fr_test_%d_%d", lcore_self, i); if (rte_mempool_lookup(mempool_name) == NULL) return -1; } @@ -212,7 +211,7 @@ hash_clean(unsigned lcore_id) int i; for (i = 0; i < MAX_ITER_TIMES; i++) { - rte_snprintf(hash_name, sizeof(hash_name), "fr_test_%d_%d", lcore_id, i); + snprintf(hash_name, sizeof(hash_name), "fr_test_%d_%d", lcore_id, i); if ((handle = rte_hash_find_existing(hash_name)) != NULL) rte_hash_free(handle); @@ -229,30 +228,29 @@ hash_create_free(__attribute__((unused)) void *arg) struct rte_hash_parameters hash_params = { .name = NULL, .entries = 16, - .bucket_entries = 4, .key_len = 4, - .hash_func = (rte_hash_function)rte_jhash2, + .hash_func = (rte_hash_function)rte_jhash_32b, .hash_func_init_val = 0, .socket_id = 0, }; WAIT_SYNCHRO_FOR_SLAVES(); - + /* create the same hash simultaneously on all threads */ hash_params.name = "fr_test_once"; for (i = 0; i < MAX_ITER_TIMES; i++) { handle = rte_hash_create(&hash_params); - if ((NULL == handle) && (rte_hash_find_existing("fr_test_once") == NULL)) - return -1; + if (handle != NULL) + rte_atomic32_inc(&obj_count); } /* create mutiple times simultaneously */ for (i = 0; i < MAX_ITER_TIMES; i++) { - rte_snprintf(hash_name, sizeof(hash_name), "fr_test_%d_%d", lcore_self, i); + snprintf(hash_name, sizeof(hash_name), "fr_test_%d_%d", lcore_self, i); hash_params.name = hash_name; handle = rte_hash_create(&hash_params); - if (NULL == handle) + if (NULL == handle) return -1; /* verify correct existing and then free all */ @@ -264,7 +262,7 @@ hash_create_free(__attribute__((unused)) void *arg) /* verify free correct */ for (i = 0; i < MAX_ITER_TIMES; i++) { - rte_snprintf(hash_name, sizeof(hash_name), "fr_test_%d_%d", lcore_self, i); + snprintf(hash_name, sizeof(hash_name), "fr_test_%d_%d", lcore_self, i); if (NULL != rte_hash_find_existing(hash_name)) return -1; @@ -281,7 +279,7 @@ fbk_clean(unsigned lcore_id) int i; for (i = 0; i < MAX_ITER_TIMES; i++) { - rte_snprintf(fbk_name, sizeof(fbk_name), "fr_test_%d_%d", lcore_id, i); + snprintf(fbk_name, sizeof(fbk_name), "fr_test_%d_%d", lcore_id, i); if ((handle = rte_fbk_hash_find_existing(fbk_name)) != NULL) rte_fbk_hash_free(handle); @@ -305,23 +303,23 @@ fbk_create_free(__attribute__((unused)) void *arg) }; WAIT_SYNCHRO_FOR_SLAVES(); - + /* create the same fbk hash table simultaneously on all threads */ fbk_params.name = "fr_test_once"; for (i = 0; i < MAX_ITER_TIMES; i++) { handle = rte_fbk_hash_create(&fbk_params); - if ((NULL == handle) && (rte_fbk_hash_find_existing("fr_test_once") == NULL)) - return -1; + if (handle != NULL) + rte_atomic32_inc(&obj_count); } /* create mutiple fbk tables simultaneously */ for (i = 0; i < MAX_ITER_TIMES; i++) { - rte_snprintf(fbk_name, sizeof(fbk_name), "fr_test_%d_%d", lcore_self, i); + snprintf(fbk_name, sizeof(fbk_name), "fr_test_%d_%d", lcore_self, i); fbk_params.name = fbk_name; handle = rte_fbk_hash_create(&fbk_params); if (NULL == handle) - return -1; + return -1; /* verify correct existing and then free all */ if (handle != rte_fbk_hash_find_existing(fbk_name)) @@ -332,7 +330,7 @@ fbk_create_free(__attribute__((unused)) void *arg) /* verify free correct */ for (i = 0; i < MAX_ITER_TIMES; i++) { - rte_snprintf(fbk_name, sizeof(fbk_name), "fr_test_%d_%d", lcore_self, i); + snprintf(fbk_name, sizeof(fbk_name), "fr_test_%d_%d", lcore_self, i); if (NULL != rte_fbk_hash_find_existing(fbk_name)) return -1; @@ -351,7 +349,7 @@ lpm_clean(unsigned lcore_id) int i; for (i = 0; i < MAX_LPM_ITER_TIMES; i++) { - rte_snprintf(lpm_name, sizeof(lpm_name), "fr_test_%d_%d", lcore_id, i); + snprintf(lpm_name, sizeof(lpm_name), "fr_test_%d_%d", lcore_id, i); if ((lpm = rte_lpm_find_existing(lpm_name)) != NULL) rte_lpm_free(lpm); @@ -363,24 +361,29 @@ lpm_create_free(__attribute__((unused)) void *arg) { unsigned lcore_self = rte_lcore_id(); struct rte_lpm *lpm; + struct rte_lpm_config config; + + config.max_rules = 4; + config.number_tbl8s = 256; + config.flags = 0; char lpm_name[MAX_STRING_SIZE]; int i; WAIT_SYNCHRO_FOR_SLAVES(); - + /* create the same lpm simultaneously on all threads */ for (i = 0; i < MAX_ITER_TIMES; i++) { - lpm = rte_lpm_create("fr_test_once", SOCKET_ID_ANY, 4, RTE_LPM_HEAP); - if ((NULL == lpm) && (rte_lpm_find_existing("fr_test_once") == NULL)) - return -1; + lpm = rte_lpm_create("fr_test_once", SOCKET_ID_ANY, &config); + if (lpm != NULL) + rte_atomic32_inc(&obj_count); } /* create mutiple fbk tables simultaneously */ for (i = 0; i < MAX_LPM_ITER_TIMES; i++) { - rte_snprintf(lpm_name, sizeof(lpm_name), "fr_test_%d_%d", lcore_self, i); - lpm = rte_lpm_create(lpm_name, SOCKET_ID_ANY, 4, RTE_LPM_HEAP); + snprintf(lpm_name, sizeof(lpm_name), "fr_test_%d_%d", lcore_self, i); + lpm = rte_lpm_create(lpm_name, SOCKET_ID_ANY, &config); if (NULL == lpm) - return -1; + return -1; /* verify correct existing and then free all */ if (lpm != rte_lpm_find_existing(lpm_name)) @@ -391,7 +394,7 @@ lpm_create_free(__attribute__((unused)) void *arg) /* verify free correct */ for (i = 0; i < MAX_LPM_ITER_TIMES; i++) { - rte_snprintf(lpm_name, sizeof(lpm_name), "fr_test_%d_%d", lcore_self, i); + snprintf(lpm_name, sizeof(lpm_name), "fr_test_%d_%d", lcore_self, i); if (NULL != rte_lpm_find_existing(lpm_name)) return -1; } @@ -421,8 +424,8 @@ struct test_case test_cases[] = { #endif /* RTE_LIBRTE_LPM */ }; -/** - * launch test case in two separate thread +/** + * launch test case in two separate thread */ static int launch_test(struct test_case *pt_case) @@ -431,10 +434,12 @@ launch_test(struct test_case *pt_case) unsigned lcore_id; unsigned cores_save = rte_lcore_count(); unsigned cores = RTE_MIN(cores_save, MAX_LCORES); + unsigned count; if (pt_case->func == NULL) return -1; + rte_atomic32_set(&obj_count, 0); rte_atomic32_set(&synchro, 0); RTE_LCORE_FOREACH_SLAVE(lcore_id) { @@ -456,18 +461,25 @@ launch_test(struct test_case *pt_case) cores--; if (rte_eal_wait_lcore(lcore_id) < 0) ret = -1; - + if (pt_case->clean != NULL) pt_case->clean(lcore_id); } - + + count = rte_atomic32_read(&obj_count); + if (count != 1) { + printf("%s: common object allocated %d times (should be 1)\n", + pt_case->name, count); + ret = -1; + } + return ret; } /** * Main entry of func_reentrancy test */ -int +static int test_func_reentrancy(void) { uint32_t case_id; @@ -484,7 +496,7 @@ test_func_reentrancy(void) pt_case = &test_cases[case_id]; if (pt_case->func == NULL) continue; - + if (launch_test(pt_case) < 0) { printf("Func-ReEnt CASE %"PRIu32": %s FAIL\n", case_id, pt_case->name); return -1; @@ -494,3 +506,5 @@ test_func_reentrancy(void) return 0; } + +REGISTER_TEST_COMMAND(func_reentrancy_autotest, test_func_reentrancy);