X-Git-Url: http://git.droids-corp.org/?a=blobdiff_plain;ds=inline;f=app%2Ftest%2Ftest_mempool.c;h=8e493eda47f4c2050836aeecd93fa82daf7b2b2c;hb=1f1e4b7cbaad919ebb865fdb75c99e7929c2801f;hp=66bc8d86b7e59cf997e0937082937beb9eae4bfb;hpb=b240af8b10f961298149a86dbb53823334c5870d;p=dpdk.git diff --git a/app/test/test_mempool.c b/app/test/test_mempool.c index 66bc8d86b7..8e493eda47 100644 --- a/app/test/test_mempool.c +++ b/app/test/test_mempool.c @@ -12,8 +12,10 @@ #include #include +#include #include #include +#include #include #include #include @@ -111,7 +113,7 @@ test_mempool_basic(struct rte_mempool *mp, int use_external_cache) printf("get private data\n"); if (rte_mempool_get_priv(mp) != (char *)mp + - MEMPOOL_HEADER_SIZE(mp, mp->cache_size)) + RTE_MEMPOOL_HEADER_SIZE(mp, mp->cache_size)) GOTO_ERR(ret, out); #ifndef RTE_EXEC_ENV_FREEBSD /* rte_mem_virt2iova() not supported on bsd */ @@ -205,15 +207,15 @@ static int test_mempool_creation_with_exceeded_cache_size(void) return 0; } -static int test_mempool_creation_with_unknown_flag(void) +static int test_mempool_creation_with_invalid_flags(void) { struct rte_mempool *mp_cov; - mp_cov = rte_mempool_create("test_mempool_unknown_flag", MEMPOOL_SIZE, + mp_cov = rte_mempool_create("test_mempool_invalid_flags", MEMPOOL_SIZE, MEMPOOL_ELT_SIZE, 0, 0, NULL, NULL, NULL, NULL, - SOCKET_ID_ANY, MEMPOOL_F_NO_IOVA_CONTIG << 1); + SOCKET_ID_ANY, ~RTE_MEMPOOL_VALID_USER_FLAGS); if (mp_cov != NULL) { rte_mempool_free(mp_cov); @@ -302,7 +304,7 @@ static int test_mempool_single_consumer(void) } /* - * test function for mempool test based on singple consumer and single producer, + * test function for mempool test based on single consumer and single producer, * can run on one lcore only */ static int @@ -320,7 +322,7 @@ my_mp_init(struct rte_mempool *mp, __rte_unused void *arg) } /* - * it tests the mempool operations based on singple producer and single consumer + * it tests the mempool operations based on single producer and single consumer */ static int test_mempool_sp_sc(void) @@ -336,8 +338,8 @@ test_mempool_sp_sc(void) my_mp_init, NULL, my_obj_init, NULL, SOCKET_ID_ANY, - MEMPOOL_F_NO_CACHE_ALIGN | MEMPOOL_F_SP_PUT | - MEMPOOL_F_SC_GET); + RTE_MEMPOOL_F_NO_CACHE_ALIGN | RTE_MEMPOOL_F_SP_PUT | + RTE_MEMPOOL_F_SC_GET); if (mp_spsc == NULL) RET_ERR(); } @@ -489,6 +491,352 @@ test_mp_mem_init(struct rte_mempool *mp, data->ret = 0; } +struct test_mempool_events_data { + struct rte_mempool *mp; + enum rte_mempool_event event; + bool invoked; +}; + +static void +test_mempool_events_cb(enum rte_mempool_event event, + struct rte_mempool *mp, void *user_data) +{ + struct test_mempool_events_data *data = user_data; + + data->mp = mp; + data->event = event; + data->invoked = true; +} + +static int +test_mempool_events(int (*populate)(struct rte_mempool *mp)) +{ +#pragma push_macro("RTE_TEST_TRACE_FAILURE") +#undef RTE_TEST_TRACE_FAILURE +#define RTE_TEST_TRACE_FAILURE(...) do { goto fail; } while (0) + + static const size_t callback_num = 3; + static const size_t mempool_num = 2; + static const unsigned int mempool_elt_size = 64; + static const unsigned int mempool_size = 64; + + struct test_mempool_events_data data[callback_num]; + struct rte_mempool *mp[mempool_num], *freed; + char name[RTE_MEMPOOL_NAMESIZE]; + size_t i, j; + int ret; + + memset(mp, 0, sizeof(mp)); + for (i = 0; i < callback_num; i++) { + ret = rte_mempool_event_callback_register + (test_mempool_events_cb, &data[i]); + RTE_TEST_ASSERT_EQUAL(ret, 0, "Failed to register the callback %zu: %s", + i, rte_strerror(rte_errno)); + } + ret = rte_mempool_event_callback_unregister(test_mempool_events_cb, mp); + RTE_TEST_ASSERT_NOT_EQUAL(ret, 0, "Unregistered a non-registered callback"); + /* NULL argument has no special meaning in this API. */ + ret = rte_mempool_event_callback_unregister(test_mempool_events_cb, + NULL); + RTE_TEST_ASSERT_NOT_EQUAL(ret, 0, "Unregistered a non-registered callback with NULL argument"); + + /* Create mempool 0 that will be observed by all callbacks. */ + memset(&data, 0, sizeof(data)); + strcpy(name, "empty0"); + mp[0] = rte_mempool_create_empty(name, mempool_size, + mempool_elt_size, 0, 0, + SOCKET_ID_ANY, 0); + RTE_TEST_ASSERT_NOT_NULL(mp[0], "Cannot create mempool %s: %s", + name, rte_strerror(rte_errno)); + for (j = 0; j < callback_num; j++) + RTE_TEST_ASSERT_EQUAL(data[j].invoked, false, + "Callback %zu invoked on %s mempool creation", + j, name); + + rte_mempool_set_ops_byname(mp[0], rte_mbuf_best_mempool_ops(), NULL); + ret = populate(mp[0]); + RTE_TEST_ASSERT_EQUAL(ret, (int)mp[0]->size, "Failed to populate mempool %s: %s", + name, rte_strerror(-ret)); + for (j = 0; j < callback_num; j++) { + RTE_TEST_ASSERT_EQUAL(data[j].invoked, true, + "Callback %zu not invoked on mempool %s population", + j, name); + RTE_TEST_ASSERT_EQUAL(data[j].event, + RTE_MEMPOOL_EVENT_READY, + "Wrong callback invoked, expected READY"); + RTE_TEST_ASSERT_EQUAL(data[j].mp, mp[0], + "Callback %zu invoked for a wrong mempool instead of %s", + j, name); + } + + /* Check that unregistered callback 0 observes no events. */ + ret = rte_mempool_event_callback_unregister(test_mempool_events_cb, + &data[0]); + RTE_TEST_ASSERT_EQUAL(ret, 0, "Failed to unregister callback 0: %s", + rte_strerror(rte_errno)); + memset(&data, 0, sizeof(data)); + strcpy(name, "empty1"); + mp[1] = rte_mempool_create_empty(name, mempool_size, + mempool_elt_size, 0, 0, + SOCKET_ID_ANY, 0); + RTE_TEST_ASSERT_NOT_NULL(mp[1], "Cannot create mempool %s: %s", + name, rte_strerror(rte_errno)); + rte_mempool_set_ops_byname(mp[1], rte_mbuf_best_mempool_ops(), NULL); + ret = populate(mp[1]); + RTE_TEST_ASSERT_EQUAL(ret, (int)mp[1]->size, "Failed to populate mempool %s: %s", + name, rte_strerror(-ret)); + RTE_TEST_ASSERT_EQUAL(data[0].invoked, false, + "Unregistered callback 0 invoked on %s mempool populaton", + name); + + for (i = 0; i < mempool_num; i++) { + memset(&data, 0, sizeof(data)); + sprintf(name, "empty%zu", i); + rte_mempool_free(mp[i]); + /* + * Save pointer to check that it was passed to the callback, + * but put NULL into the array in case cleanup is called early. + */ + freed = mp[i]; + mp[i] = NULL; + for (j = 1; j < callback_num; j++) { + RTE_TEST_ASSERT_EQUAL(data[j].invoked, true, + "Callback %zu not invoked on mempool %s destruction", + j, name); + RTE_TEST_ASSERT_EQUAL(data[j].event, + RTE_MEMPOOL_EVENT_DESTROY, + "Wrong callback invoked, expected DESTROY"); + RTE_TEST_ASSERT_EQUAL(data[j].mp, freed, + "Callback %zu invoked for a wrong mempool instead of %s", + j, name); + } + RTE_TEST_ASSERT_EQUAL(data[0].invoked, false, + "Unregistered callback 0 invoked on %s mempool destruction", + name); + } + + for (j = 1; j < callback_num; j++) { + ret = rte_mempool_event_callback_unregister + (test_mempool_events_cb, &data[j]); + RTE_TEST_ASSERT_EQUAL(ret, 0, "Failed to unregister the callback %zu: %s", + j, rte_strerror(rte_errno)); + } + return TEST_SUCCESS; + +fail: + for (j = 0; j < callback_num; j++) + rte_mempool_event_callback_unregister + (test_mempool_events_cb, &data[j]); + for (i = 0; i < mempool_num; i++) + rte_mempool_free(mp[i]); + return TEST_FAILED; + +#pragma pop_macro("RTE_TEST_TRACE_FAILURE") +} + +struct test_mempool_events_safety_data { + bool invoked; + int (*api_func)(rte_mempool_event_callback *func, void *user_data); + rte_mempool_event_callback *cb_func; + void *cb_user_data; + int ret; +}; + +static void +test_mempool_events_safety_cb(enum rte_mempool_event event, + struct rte_mempool *mp, void *user_data) +{ + struct test_mempool_events_safety_data *data = user_data; + + RTE_SET_USED(event); + RTE_SET_USED(mp); + data->invoked = true; + data->ret = data->api_func(data->cb_func, data->cb_user_data); +} + +static int +test_mempool_events_safety(void) +{ +#pragma push_macro("RTE_TEST_TRACE_FAILURE") +#undef RTE_TEST_TRACE_FAILURE +#define RTE_TEST_TRACE_FAILURE(...) do { \ + ret = TEST_FAILED; \ + goto exit; \ + } while (0) + + struct test_mempool_events_data data; + struct test_mempool_events_safety_data sdata[2]; + struct rte_mempool *mp; + size_t i; + int ret; + + /* removes itself */ + sdata[0].api_func = rte_mempool_event_callback_unregister; + sdata[0].cb_func = test_mempool_events_safety_cb; + sdata[0].cb_user_data = &sdata[0]; + sdata[0].ret = -1; + rte_mempool_event_callback_register(test_mempool_events_safety_cb, + &sdata[0]); + /* inserts a callback after itself */ + sdata[1].api_func = rte_mempool_event_callback_register; + sdata[1].cb_func = test_mempool_events_cb; + sdata[1].cb_user_data = &data; + sdata[1].ret = -1; + rte_mempool_event_callback_register(test_mempool_events_safety_cb, + &sdata[1]); + + mp = rte_mempool_create_empty("empty", MEMPOOL_SIZE, + MEMPOOL_ELT_SIZE, 0, 0, + SOCKET_ID_ANY, 0); + RTE_TEST_ASSERT_NOT_NULL(mp, "Cannot create mempool: %s", + rte_strerror(rte_errno)); + memset(&data, 0, sizeof(data)); + ret = rte_mempool_populate_default(mp); + RTE_TEST_ASSERT_EQUAL(ret, (int)mp->size, "Failed to populate mempool: %s", + rte_strerror(-ret)); + + RTE_TEST_ASSERT_EQUAL(sdata[0].ret, 0, "Callback failed to unregister itself: %s", + rte_strerror(rte_errno)); + RTE_TEST_ASSERT_EQUAL(sdata[1].ret, 0, "Failed to insert a new callback: %s", + rte_strerror(rte_errno)); + RTE_TEST_ASSERT_EQUAL(data.invoked, false, + "Inserted callback is invoked on mempool population"); + + memset(&data, 0, sizeof(data)); + sdata[0].invoked = false; + rte_mempool_free(mp); + mp = NULL; + RTE_TEST_ASSERT_EQUAL(sdata[0].invoked, false, + "Callback that unregistered itself was called"); + RTE_TEST_ASSERT_EQUAL(sdata[1].ret, -EEXIST, + "New callback inserted twice"); + RTE_TEST_ASSERT_EQUAL(data.invoked, true, + "Inserted callback is not invoked on mempool destruction"); + + rte_mempool_event_callback_unregister(test_mempool_events_cb, &data); + for (i = 0; i < RTE_DIM(sdata); i++) + rte_mempool_event_callback_unregister + (test_mempool_events_safety_cb, &sdata[i]); + ret = TEST_SUCCESS; + +exit: + /* cleanup, don't care which callbacks are already removed */ + rte_mempool_event_callback_unregister(test_mempool_events_cb, &data); + for (i = 0; i < RTE_DIM(sdata); i++) + rte_mempool_event_callback_unregister + (test_mempool_events_safety_cb, &sdata[i]); + /* in case of failure before the planned destruction */ + rte_mempool_free(mp); + return ret; + +#pragma pop_macro("RTE_TEST_TRACE_FAILURE") +} + +#pragma push_macro("RTE_TEST_TRACE_FAILURE") +#undef RTE_TEST_TRACE_FAILURE +#define RTE_TEST_TRACE_FAILURE(...) do { \ + ret = TEST_FAILED; \ + goto exit; \ + } while (0) + +static int +test_mempool_flag_non_io_set_when_no_iova_contig_set(void) +{ + const struct rte_memzone *mz = NULL; + void *virt; + rte_iova_t iova; + size_t size = MEMPOOL_ELT_SIZE * 16; + struct rte_mempool *mp = NULL; + int ret; + + mz = rte_memzone_reserve("test_mempool", size, SOCKET_ID_ANY, 0); + RTE_TEST_ASSERT_NOT_NULL(mz, "Cannot allocate memory"); + virt = mz->addr; + iova = mz->iova; + mp = rte_mempool_create_empty("empty", MEMPOOL_SIZE, + MEMPOOL_ELT_SIZE, 0, 0, + SOCKET_ID_ANY, RTE_MEMPOOL_F_NO_IOVA_CONTIG); + RTE_TEST_ASSERT_NOT_NULL(mp, "Cannot create mempool: %s", + rte_strerror(rte_errno)); + rte_mempool_set_ops_byname(mp, rte_mbuf_best_mempool_ops(), NULL); + + RTE_TEST_ASSERT(mp->flags & RTE_MEMPOOL_F_NON_IO, + "NON_IO flag is not set on an empty mempool"); + + /* + * Always use valid IOVA so that populate() has no other reason + * to infer that the mempool cannot be used for IO. + */ + ret = rte_mempool_populate_iova(mp, virt, iova, size, NULL, NULL); + RTE_TEST_ASSERT(ret > 0, "Failed to populate mempool: %s", + rte_strerror(-ret)); + RTE_TEST_ASSERT(mp->flags & RTE_MEMPOOL_F_NON_IO, + "NON_IO flag is not set when NO_IOVA_CONTIG is set"); + ret = TEST_SUCCESS; +exit: + rte_mempool_free(mp); + rte_memzone_free(mz); + return ret; +} + +static int +test_mempool_flag_non_io_unset_when_populated_with_valid_iova(void) +{ + const struct rte_memzone *mz = NULL; + void *virt; + rte_iova_t iova; + size_t total_size = MEMPOOL_ELT_SIZE * MEMPOOL_SIZE; + size_t block_size = total_size / 3; + struct rte_mempool *mp = NULL; + int ret; + + /* + * Since objects from the pool are never used in the test, + * we don't care for contiguous IOVA, on the other hand, + * requiring it could cause spurious test failures. + */ + mz = rte_memzone_reserve("test_mempool", total_size, SOCKET_ID_ANY, 0); + RTE_TEST_ASSERT_NOT_NULL(mz, "Cannot allocate memory"); + virt = mz->addr; + iova = mz->iova; + mp = rte_mempool_create_empty("empty", MEMPOOL_SIZE, + MEMPOOL_ELT_SIZE, 0, 0, + SOCKET_ID_ANY, 0); + RTE_TEST_ASSERT_NOT_NULL(mp, "Cannot create mempool: %s", + rte_strerror(rte_errno)); + + RTE_TEST_ASSERT(mp->flags & RTE_MEMPOOL_F_NON_IO, + "NON_IO flag is not set on an empty mempool"); + + ret = rte_mempool_populate_iova(mp, RTE_PTR_ADD(virt, 1 * block_size), + RTE_BAD_IOVA, block_size, NULL, NULL); + RTE_TEST_ASSERT(ret > 0, "Failed to populate mempool: %s", + rte_strerror(-ret)); + RTE_TEST_ASSERT(mp->flags & RTE_MEMPOOL_F_NON_IO, + "NON_IO flag is not set when mempool is populated with only RTE_BAD_IOVA"); + + ret = rte_mempool_populate_iova(mp, virt, iova, block_size, NULL, NULL); + RTE_TEST_ASSERT(ret > 0, "Failed to populate mempool: %s", + rte_strerror(-ret)); + RTE_TEST_ASSERT(!(mp->flags & RTE_MEMPOOL_F_NON_IO), + "NON_IO flag is not unset when mempool is populated with valid IOVA"); + + ret = rte_mempool_populate_iova(mp, RTE_PTR_ADD(virt, 2 * block_size), + RTE_BAD_IOVA, block_size, NULL, NULL); + RTE_TEST_ASSERT(ret > 0, "Failed to populate mempool: %s", + rte_strerror(-ret)); + RTE_TEST_ASSERT(!(mp->flags & RTE_MEMPOOL_F_NON_IO), + "NON_IO flag is set even when some objects have valid IOVA"); + ret = TEST_SUCCESS; + +exit: + rte_mempool_free(mp); + rte_memzone_free(mz); + return ret; +} + +#pragma pop_macro("RTE_TEST_TRACE_FAILURE") + static int test_mempool(void) { @@ -646,14 +994,14 @@ test_mempool(void) if (test_mempool_basic_ex(mp_nocache) < 0) GOTO_ERR(ret, err); - /* mempool operation test based on single producer and single comsumer */ + /* mempool operation test based on single producer and single consumer */ if (test_mempool_sp_sc() < 0) GOTO_ERR(ret, err); if (test_mempool_creation_with_exceeded_cache_size() < 0) GOTO_ERR(ret, err); - if (test_mempool_creation_with_unknown_flag() < 0) + if (test_mempool_creation_with_invalid_flags() < 0) GOTO_ERR(ret, err); if (test_mempool_same_name_twice_creation() < 0) @@ -666,6 +1014,20 @@ test_mempool(void) if (test_mempool_basic(default_pool, 1) < 0) GOTO_ERR(ret, err); + /* test mempool event callbacks */ + if (test_mempool_events(rte_mempool_populate_default) < 0) + GOTO_ERR(ret, err); + if (test_mempool_events(rte_mempool_populate_anon) < 0) + GOTO_ERR(ret, err); + if (test_mempool_events_safety() < 0) + GOTO_ERR(ret, err); + + /* test NON_IO flag inference */ + if (test_mempool_flag_non_io_set_when_no_iova_contig_set() < 0) + GOTO_ERR(ret, err); + if (test_mempool_flag_non_io_unset_when_populated_with_valid_iova() < 0) + GOTO_ERR(ret, err); + rte_mempool_list_dump(stdout); ret = 0;