X-Git-Url: http://git.droids-corp.org/?a=blobdiff_plain;f=app%2Ftest%2Ftest_mempool.c;h=b9880b32b8f95ffa330595488044db1265ecf37e;hb=dd0eedb1cfcf0cb7423d859177c5bc6f931eaf8a;hp=31582d825cc6ccfe81c2e74325b16ae3c05258e9;hpb=99ca3b7a82b0fa0db8cda1f097901376dc86fdf5;p=dpdk.git diff --git a/app/test/test_mempool.c b/app/test/test_mempool.c index 31582d825c..b9880b32b8 100644 --- a/app/test/test_mempool.c +++ b/app/test/test_mempool.c @@ -52,7 +52,6 @@ #include #include #include -#include #include #include #include @@ -75,106 +74,19 @@ #define MAX_KEEP 16 #define MEMPOOL_SIZE ((rte_lcore_count()*(MAX_KEEP+RTE_MEMPOOL_CACHE_MAX_SIZE))-1) +#define LOG_ERR() printf("test failed at %s():%d\n", __func__, __LINE__) #define RET_ERR() do { \ - printf("test failed at %s():%d\n", __func__, __LINE__); \ + LOG_ERR(); \ return -1; \ } while (0) +#define GOTO_ERR(var, label) do { \ + LOG_ERR(); \ + var = -1; \ + goto label; \ + } while (0) static rte_atomic32_t synchro; -/* - * Simple example of custom mempool structure. Holds pointers to all the - * elements which are simply malloc'd in this example. - */ -struct custom_mempool { - rte_spinlock_t lock; - unsigned count; - unsigned size; - void *elts[]; -}; - -/* - * Loop through all the element pointers and allocate a chunk of memory, then - * insert that memory into the ring. - */ -static int -custom_mempool_alloc(struct rte_mempool *mp) -{ - struct custom_mempool *cm; - - cm = rte_zmalloc("custom_mempool", - sizeof(struct custom_mempool) + mp->size * sizeof(void *), 0); - if (cm == NULL) - return -ENOMEM; - - rte_spinlock_init(&cm->lock); - cm->count = 0; - cm->size = mp->size; - mp->pool_data = cm; - return 0; -} - -static void -custom_mempool_free(struct rte_mempool *mp) -{ - rte_free((void *)(mp->pool_data)); -} - -static int -custom_mempool_enqueue(struct rte_mempool *mp, void * const *obj_table, - unsigned n) -{ - struct custom_mempool *cm = (struct custom_mempool *)(mp->pool_data); - int ret = 0; - - rte_spinlock_lock(&cm->lock); - if (cm->count + n > cm->size) { - ret = -ENOBUFS; - } else { - memcpy(&cm->elts[cm->count], obj_table, sizeof(void *) * n); - cm->count += n; - } - rte_spinlock_unlock(&cm->lock); - return ret; -} - - -static int -custom_mempool_dequeue(struct rte_mempool *mp, void **obj_table, unsigned n) -{ - struct custom_mempool *cm = (struct custom_mempool *)(mp->pool_data); - int ret = 0; - - rte_spinlock_lock(&cm->lock); - if (n > cm->count) { - ret = -ENOENT; - } else { - cm->count -= n; - memcpy(obj_table, &cm->elts[cm->count], sizeof(void *) * n); - } - rte_spinlock_unlock(&cm->lock); - return ret; -} - -static unsigned -custom_mempool_get_count(const struct rte_mempool *mp) -{ - struct custom_mempool *cm = (struct custom_mempool *)(mp->pool_data); - - return cm->count; -} - -static struct rte_mempool_ops mempool_ops_custom = { - .name = "custom_handler", - .alloc = custom_mempool_alloc, - .free = custom_mempool_free, - .enqueue = custom_mempool_enqueue, - .dequeue = custom_mempool_dequeue, - .get_count = custom_mempool_get_count, -}; - -MEMPOOL_REGISTER_OPS(mempool_ops_custom); - /* * save the object number in the first 4 bytes of object data. All * other bytes are set to 0. @@ -191,7 +103,7 @@ my_obj_init(struct rte_mempool *mp, __attribute__((unused)) void *arg, /* basic tests (done on one core) */ static int -test_mempool_basic(struct rte_mempool *mp) +test_mempool_basic(struct rte_mempool *mp, int use_external_cache) { uint32_t *objnum; void **objtable; @@ -199,47 +111,62 @@ test_mempool_basic(struct rte_mempool *mp) char *obj_data; int ret = 0; unsigned i, j; + int offset; + struct rte_mempool_cache *cache; + + if (use_external_cache) { + /* Create a user-owned mempool cache. */ + cache = rte_mempool_cache_create(RTE_MEMPOOL_CACHE_MAX_SIZE, + SOCKET_ID_ANY); + if (cache == NULL) + RET_ERR(); + } else { + /* May be NULL if cache is disabled. */ + cache = rte_mempool_default_cache(mp, rte_lcore_id()); + } /* dump the mempool status */ rte_mempool_dump(stdout, mp); printf("get an object\n"); - if (rte_mempool_get(mp, &obj) < 0) - RET_ERR(); + if (rte_mempool_generic_get(mp, &obj, 1, cache, 0) < 0) + GOTO_ERR(ret, out); rte_mempool_dump(stdout, mp); /* tests that improve coverage */ printf("get object count\n"); - if (rte_mempool_count(mp) != MEMPOOL_SIZE - 1) - RET_ERR(); + /* We have to count the extra caches, one in this case. */ + offset = use_external_cache ? 1 * cache->len : 0; + if (rte_mempool_avail_count(mp) + offset != MEMPOOL_SIZE - 1) + GOTO_ERR(ret, out); printf("get private data\n"); if (rte_mempool_get_priv(mp) != (char *)mp + MEMPOOL_HEADER_SIZE(mp, mp->cache_size)) - RET_ERR(); + GOTO_ERR(ret, out); #ifndef RTE_EXEC_ENV_BSDAPP /* rte_mem_virt2phy() not supported on bsd */ printf("get physical address of an object\n"); if (rte_mempool_virt2phy(mp, obj) != rte_mem_virt2phy(obj)) - RET_ERR(); + GOTO_ERR(ret, out); #endif printf("put the object back\n"); - rte_mempool_put(mp, obj); + rte_mempool_generic_put(mp, &obj, 1, cache, 0); rte_mempool_dump(stdout, mp); printf("get 2 objects\n"); - if (rte_mempool_get(mp, &obj) < 0) - RET_ERR(); - if (rte_mempool_get(mp, &obj2) < 0) { - rte_mempool_put(mp, obj); - RET_ERR(); + if (rte_mempool_generic_get(mp, &obj, 1, cache, 0) < 0) + GOTO_ERR(ret, out); + if (rte_mempool_generic_get(mp, &obj2, 1, cache, 0) < 0) { + rte_mempool_generic_put(mp, &obj, 1, cache, 0); + GOTO_ERR(ret, out); } rte_mempool_dump(stdout, mp); printf("put the objects back\n"); - rte_mempool_put(mp, obj); - rte_mempool_put(mp, obj2); + rte_mempool_generic_put(mp, &obj, 1, cache, 0); + rte_mempool_generic_put(mp, &obj2, 1, cache, 0); rte_mempool_dump(stdout, mp); /* @@ -248,10 +175,10 @@ test_mempool_basic(struct rte_mempool *mp) */ objtable = malloc(MEMPOOL_SIZE * sizeof(void *)); if (objtable == NULL) - RET_ERR(); + GOTO_ERR(ret, out); for (i = 0; i < MEMPOOL_SIZE; i++) { - if (rte_mempool_get(mp, &objtable[i]) < 0) + if (rte_mempool_generic_get(mp, &objtable[i], 1, cache, 0) < 0) break; } @@ -273,13 +200,19 @@ test_mempool_basic(struct rte_mempool *mp) ret = -1; } - rte_mempool_put(mp, objtable[i]); + rte_mempool_generic_put(mp, &objtable[i], 1, cache, 0); } free(objtable); if (ret == -1) printf("objects were modified!\n"); +out: + if (use_external_cache) { + rte_mempool_cache_flush(cache, mp); + rte_mempool_cache_free(cache); + } + return ret; } @@ -338,7 +271,7 @@ static int test_mempool_single_producer(void) printf("obj not owned by this mempool\n"); RET_ERR(); } - rte_mempool_sp_put(mp_spsc, obj); + rte_mempool_put(mp_spsc, obj); rte_spinlock_lock(&scsp_spinlock); scsp_obj_table[i] = NULL; rte_spinlock_unlock(&scsp_spinlock); @@ -371,7 +304,7 @@ static int test_mempool_single_consumer(void) rte_spinlock_unlock(&scsp_spinlock); if (i >= MAX_KEEP) continue; - if (rte_mempool_sc_get(mp_spsc, &obj) < 0) + if (rte_mempool_get(mp_spsc, &obj) < 0) break; rte_spinlock_lock(&scsp_spinlock); scsp_obj_table[i] = obj; @@ -470,20 +403,20 @@ test_mempool_basic_ex(struct rte_mempool *mp) return ret; } printf("test_mempool_basic_ex now mempool (%s) has %u free entries\n", - mp->name, rte_mempool_free_count(mp)); + mp->name, rte_mempool_in_use_count(mp)); if (rte_mempool_full(mp) != 1) { printf("test_mempool_basic_ex the mempool should be full\n"); goto fail_mp_basic_ex; } for (i = 0; i < MEMPOOL_SIZE; i ++) { - if (rte_mempool_mc_get(mp, &obj[i]) < 0) { + if (rte_mempool_get(mp, &obj[i]) < 0) { printf("test_mp_basic_ex fail to get object for [%u]\n", i); goto fail_mp_basic_ex; } } - if (rte_mempool_mc_get(mp, &err_obj) == 0) { + if (rte_mempool_get(mp, &err_obj) == 0) { printf("test_mempool_basic_ex get an impossible obj\n"); goto fail_mp_basic_ex; } @@ -494,7 +427,7 @@ test_mempool_basic_ex(struct rte_mempool *mp) } for (i = 0; i < MEMPOOL_SIZE; i++) - rte_mempool_mp_put(mp, obj[i]); + rte_mempool_put(mp, obj[i]); if (rte_mempool_full(mp) != 1) { printf("test_mempool_basic_ex the mempool should be full\n"); @@ -567,12 +500,18 @@ test_mempool_xmem_misc(void) return 0; } +static void +walk_cb(struct rte_mempool *mp, void *userdata __rte_unused) +{ + printf("\t%s\n", mp->name); +} + static int test_mempool(void) { struct rte_mempool *mp_cache = NULL; struct rte_mempool *mp_nocache = NULL; - struct rte_mempool *mp_ext = NULL; + struct rte_mempool *mp_stack = NULL; rte_atomic32_init(&synchro); @@ -602,25 +541,25 @@ test_mempool(void) } /* create a mempool with an external handler */ - mp_ext = rte_mempool_create_empty("test_ext", + mp_stack = rte_mempool_create_empty("test_stack", MEMPOOL_SIZE, MEMPOOL_ELT_SIZE, RTE_MEMPOOL_CACHE_MAX_SIZE, 0, SOCKET_ID_ANY, 0); - if (mp_ext == NULL) { - printf("cannot allocate mp_ext mempool\n"); + if (mp_stack == NULL) { + printf("cannot allocate mp_stack mempool\n"); goto err; } - if (rte_mempool_set_ops_byname(mp_ext, "custom_handler", NULL) < 0) { - printf("cannot set custom handler\n"); + if (rte_mempool_set_ops_byname(mp_stack, "stack", NULL) < 0) { + printf("cannot set stack handler\n"); goto err; } - if (rte_mempool_populate_default(mp_ext) < 0) { - printf("cannot populate mp_ext mempool\n"); + if (rte_mempool_populate_default(mp_stack) < 0) { + printf("cannot populate mp_stack mempool\n"); goto err; } - rte_mempool_obj_iter(mp_ext, my_obj_init, NULL); + rte_mempool_obj_iter(mp_stack, my_obj_init, NULL); /* retrieve the mempool from its name */ if (rte_mempool_lookup("test_nocache") != mp_nocache) { @@ -628,14 +567,21 @@ test_mempool(void) goto err; } + printf("Walk into mempools:\n"); + rte_mempool_walk(walk_cb, NULL); + rte_mempool_list_dump(stdout); /* basic tests without cache */ - if (test_mempool_basic(mp_nocache) < 0) + if (test_mempool_basic(mp_nocache, 0) < 0) goto err; /* basic tests with cache */ - if (test_mempool_basic(mp_cache) < 0) + if (test_mempool_basic(mp_cache, 0) < 0) + goto err; + + /* basic tests with user-owned cache */ + if (test_mempool_basic(mp_nocache, 1) < 0) goto err; /* more basic tests without cache */ @@ -655,6 +601,10 @@ test_mempool(void) if (test_mempool_xmem_misc() < 0) goto err; + /* test the stack handler */ + if (test_mempool_basic(mp_stack, 1) < 0) + goto err; + rte_mempool_list_dump(stdout); return 0; @@ -662,12 +612,8 @@ test_mempool(void) err: rte_mempool_free(mp_nocache); rte_mempool_free(mp_cache); - rte_mempool_free(mp_ext); + rte_mempool_free(mp_stack); return -1; } -static struct test_command mempool_cmd = { - .command = "mempool_autotest", - .callback = test_mempool, -}; -REGISTER_TEST_COMMAND(mempool_cmd); +REGISTER_TEST_COMMAND(mempool_autotest, test_mempool);