X-Git-Url: http://git.droids-corp.org/?a=blobdiff_plain;f=app%2Ftest%2Ftest_memzone.c;h=7edfd06c4d48cbf1dec0f14de716cdb77bb0aaad;hb=923ceaeac140e1579abafb56f4131db542add903;hp=221acad4646e91f83faf204a6f4ba8f316afaad3;hpb=591a9d7985c1230652d9f7ea1f9221e8c66ec188;p=dpdk.git diff --git a/app/test/test_memzone.c b/app/test/test_memzone.c index 221acad464..7edfd06c4d 100644 --- a/app/test/test_memzone.c +++ b/app/test/test_memzone.c @@ -1,38 +1,10 @@ -/*- - * BSD LICENSE - * - * Copyright(c) 2010-2014 Intel Corporation. All rights reserved. - * All rights reserved. - * - * Redistribution and use in source and binary forms, with or without - * modification, are permitted provided that the following conditions - * are met: - * - * * Redistributions of source code must retain the above copyright - * notice, this list of conditions and the following disclaimer. - * * Redistributions in binary form must reproduce the above copyright - * notice, this list of conditions and the following disclaimer in - * the documentation and/or other materials provided with the - * distribution. - * * Neither the name of Intel Corporation nor the names of its - * contributors may be used to endorse or promote products derived - * from this software without specific prior written permission. - * - * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS - * "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT - * LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR - * A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT - * OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, - * SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT - * LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, - * DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY - * THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT - * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE - * OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. +/* SPDX-License-Identifier: BSD-3-Clause + * Copyright(c) 2010-2014 Intel Corporation */ #include #include +#include #include #include @@ -40,11 +12,13 @@ #include #include #include -#include #include -#include #include #include +#include +#include +#include "../../lib/librte_eal/common/malloc_elem.h" +#include "../../lib/librte_eal/common/eal_memcfg.h" #include "test.h" @@ -74,9 +48,11 @@ * - Check flags for specific huge page size reservation */ +#define TEST_MEMZONE_NAME(suffix) "MZ_TEST_" suffix + /* Test if memory overlaps: return 1 if true, or 0 if false. */ static int -is_memory_overlap(phys_addr_t ptr1, size_t len1, phys_addr_t ptr2, size_t len2) +is_memory_overlap(rte_iova_t ptr1, size_t len1, rte_iova_t ptr2, size_t len2) { if (ptr2 >= ptr1 && (ptr2 - ptr1) < len1) return 1; @@ -90,14 +66,14 @@ test_memzone_invalid_alignment(void) { const struct rte_memzone * mz; - mz = rte_memzone_lookup("invalid_alignment"); + mz = rte_memzone_lookup(TEST_MEMZONE_NAME("invalid_alignment")); if (mz != NULL) { printf("Zone with invalid alignment has been reserved\n"); return -1; } - mz = rte_memzone_reserve_aligned("invalid_alignment", 100, - SOCKET_ID_ANY, 0, 100); + mz = rte_memzone_reserve_aligned(TEST_MEMZONE_NAME("invalid_alignment"), + 100, SOCKET_ID_ANY, 0, 100); if (mz != NULL) { printf("Zone with invalid alignment has been reserved\n"); return -1; @@ -110,14 +86,16 @@ test_memzone_reserving_zone_size_bigger_than_the_maximum(void) { const struct rte_memzone * mz; - mz = rte_memzone_lookup("zone_size_bigger_than_the_maximum"); + mz = rte_memzone_lookup( + TEST_MEMZONE_NAME("zone_size_bigger_than_the_maximum")); if (mz != NULL) { printf("zone_size_bigger_than_the_maximum has been reserved\n"); return -1; } - mz = rte_memzone_reserve("zone_size_bigger_than_the_maximum", (size_t)-1, - SOCKET_ID_ANY, 0); + mz = rte_memzone_reserve( + TEST_MEMZONE_NAME("zone_size_bigger_than_the_maximum"), + (size_t)-1, SOCKET_ID_ANY, 0); if (mz != NULL) { printf("It is impossible to reserve such big a memzone\n"); return -1; @@ -126,27 +104,59 @@ test_memzone_reserving_zone_size_bigger_than_the_maximum(void) return 0; } +struct walk_arg { + int hugepage_2MB_avail; + int hugepage_1GB_avail; + int hugepage_16MB_avail; + int hugepage_16GB_avail; +}; +static int +find_available_pagesz(const struct rte_memseg_list *msl, void *arg) +{ + struct walk_arg *wa = arg; + + if (msl->external) + return 0; + + if (msl->page_sz == RTE_PGSIZE_2M) + wa->hugepage_2MB_avail = 1; + if (msl->page_sz == RTE_PGSIZE_1G) + wa->hugepage_1GB_avail = 1; + if (msl->page_sz == RTE_PGSIZE_16M) + wa->hugepage_16MB_avail = 1; + if (msl->page_sz == RTE_PGSIZE_16G) + wa->hugepage_16GB_avail = 1; + + return 0; +} + static int test_memzone_reserve_flags(void) { const struct rte_memzone *mz; - const struct rte_memseg *ms; - int hugepage_2MB_avail = 0; - int hugepage_1GB_avail = 0; + struct walk_arg wa; + int hugepage_2MB_avail, hugepage_1GB_avail; + int hugepage_16MB_avail, hugepage_16GB_avail; const size_t size = 100; - int i = 0; - ms = rte_eal_get_physmem_layout(); - for (i = 0; i < RTE_MAX_MEMSEG; i++) { - if (ms[i].hugepage_sz == RTE_PGSIZE_2M) - hugepage_2MB_avail = 1; - if (ms[i].hugepage_sz == RTE_PGSIZE_1G) - hugepage_1GB_avail = 1; - } - /* Display the availability of 2MB and 1GB pages */ + + memset(&wa, 0, sizeof(wa)); + + rte_memseg_list_walk(find_available_pagesz, &wa); + + hugepage_2MB_avail = wa.hugepage_2MB_avail; + hugepage_1GB_avail = wa.hugepage_1GB_avail; + hugepage_16MB_avail = wa.hugepage_16MB_avail; + hugepage_16GB_avail = wa.hugepage_16GB_avail; + + /* Display the availability of 2MB ,1GB, 16MB, 16GB pages */ if (hugepage_2MB_avail) printf("2MB Huge pages available\n"); if (hugepage_1GB_avail) printf("1GB Huge pages available\n"); + if (hugepage_16MB_avail) + printf("16MB Huge pages available\n"); + if (hugepage_16GB_avail) + printf("16GB Huge pages available\n"); /* * If 2MB pages available, check that a small memzone is correctly * reserved from 2MB huge pages when requested by the RTE_MEMZONE_2MB flag. @@ -154,8 +164,8 @@ test_memzone_reserve_flags(void) * available page size (i.e 1GB ) when 2MB pages are unavailable. */ if (hugepage_2MB_avail) { - mz = rte_memzone_reserve("flag_zone_2M", size, SOCKET_ID_ANY, - RTE_MEMZONE_2MB); + mz = rte_memzone_reserve(TEST_MEMZONE_NAME("flag_zone_2M"), + size, SOCKET_ID_ANY, RTE_MEMZONE_2MB); if (mz == NULL) { printf("MEMZONE FLAG 2MB\n"); return -1; @@ -164,8 +174,13 @@ test_memzone_reserve_flags(void) printf("hugepage_sz not equal 2M\n"); return -1; } + if (rte_memzone_free(mz)) { + printf("Fail memzone free\n"); + return -1; + } - mz = rte_memzone_reserve("flag_zone_2M_HINT", size, SOCKET_ID_ANY, + mz = rte_memzone_reserve(TEST_MEMZONE_NAME("flag_zone_2M_HINT"), + size, SOCKET_ID_ANY, RTE_MEMZONE_2MB|RTE_MEMZONE_SIZE_HINT_ONLY); if (mz == NULL) { printf("MEMZONE FLAG 2MB\n"); @@ -175,12 +190,18 @@ test_memzone_reserve_flags(void) printf("hugepage_sz not equal 2M\n"); return -1; } + if (rte_memzone_free(mz)) { + printf("Fail memzone free\n"); + return -1; + } /* Check if 1GB huge pages are unavailable, that function fails unless * HINT flag is indicated */ if (!hugepage_1GB_avail) { - mz = rte_memzone_reserve("flag_zone_1G_HINT", size, SOCKET_ID_ANY, + mz = rte_memzone_reserve( + TEST_MEMZONE_NAME("flag_zone_1G_HINT"), + size, SOCKET_ID_ANY, RTE_MEMZONE_1GB|RTE_MEMZONE_SIZE_HINT_ONLY); if (mz == NULL) { printf("MEMZONE FLAG 1GB & HINT\n"); @@ -190,9 +211,14 @@ test_memzone_reserve_flags(void) printf("hugepage_sz not equal 2M\n"); return -1; } + if (rte_memzone_free(mz)) { + printf("Fail memzone free\n"); + return -1; + } - mz = rte_memzone_reserve("flag_zone_1G", size, SOCKET_ID_ANY, - RTE_MEMZONE_1GB); + mz = rte_memzone_reserve( + TEST_MEMZONE_NAME("flag_zone_1G"), size, + SOCKET_ID_ANY, RTE_MEMZONE_1GB); if (mz != NULL) { printf("MEMZONE FLAG 1GB\n"); return -1; @@ -202,8 +228,8 @@ test_memzone_reserve_flags(void) /*As with 2MB tests above for 1GB huge page requests*/ if (hugepage_1GB_avail) { - mz = rte_memzone_reserve("flag_zone_1G", size, SOCKET_ID_ANY, - RTE_MEMZONE_1GB); + mz = rte_memzone_reserve(TEST_MEMZONE_NAME("flag_zone_1G"), + size, SOCKET_ID_ANY, RTE_MEMZONE_1GB); if (mz == NULL) { printf("MEMZONE FLAG 1GB\n"); return -1; @@ -212,8 +238,13 @@ test_memzone_reserve_flags(void) printf("hugepage_sz not equal 1G\n"); return -1; } + if (rte_memzone_free(mz)) { + printf("Fail memzone free\n"); + return -1; + } - mz = rte_memzone_reserve("flag_zone_1G_HINT", size, SOCKET_ID_ANY, + mz = rte_memzone_reserve(TEST_MEMZONE_NAME("flag_zone_1G_HINT"), + size, SOCKET_ID_ANY, RTE_MEMZONE_1GB|RTE_MEMZONE_SIZE_HINT_ONLY); if (mz == NULL) { printf("MEMZONE FLAG 1GB\n"); @@ -223,12 +254,18 @@ test_memzone_reserve_flags(void) printf("hugepage_sz not equal 1G\n"); return -1; } + if (rte_memzone_free(mz)) { + printf("Fail memzone free\n"); + return -1; + } /* Check if 1GB huge pages are unavailable, that function fails unless * HINT flag is indicated */ if (!hugepage_2MB_avail) { - mz = rte_memzone_reserve("flag_zone_2M_HINT", size, SOCKET_ID_ANY, + mz = rte_memzone_reserve( + TEST_MEMZONE_NAME("flag_zone_2M_HINT"), + size, SOCKET_ID_ANY, RTE_MEMZONE_2MB|RTE_MEMZONE_SIZE_HINT_ONLY); if (mz == NULL){ printf("MEMZONE FLAG 2MB & HINT\n"); @@ -238,8 +275,13 @@ test_memzone_reserve_flags(void) printf("hugepage_sz not equal 1G\n"); return -1; } - mz = rte_memzone_reserve("flag_zone_2M", size, SOCKET_ID_ANY, - RTE_MEMZONE_2MB); + if (rte_memzone_free(mz)) { + printf("Fail memzone free\n"); + return -1; + } + mz = rte_memzone_reserve( + TEST_MEMZONE_NAME("flag_zone_2M"), size, + SOCKET_ID_ANY, RTE_MEMZONE_2MB); if (mz != NULL) { printf("MEMZONE FLAG 2MB\n"); return -1; @@ -247,195 +289,313 @@ test_memzone_reserve_flags(void) } if (hugepage_2MB_avail && hugepage_1GB_avail) { - mz = rte_memzone_reserve("flag_zone_2M_HINT", size, SOCKET_ID_ANY, - RTE_MEMZONE_2MB|RTE_MEMZONE_1GB); + mz = rte_memzone_reserve( + TEST_MEMZONE_NAME("flag_zone_2M_HINT"), + size, SOCKET_ID_ANY, + RTE_MEMZONE_2MB|RTE_MEMZONE_1GB); + if (mz == NULL) { + printf("BOTH SIZES SET\n"); + return -1; + } + if (mz->hugepage_sz != RTE_PGSIZE_1G && + mz->hugepage_sz != RTE_PGSIZE_2M) { + printf("Wrong size when both sizes set\n"); + return -1; + } + if (rte_memzone_free(mz)) { + printf("Fail memzone free\n"); + return -1; + } + } + } + /* + * This option is for IBM Power. If 16MB pages available, check + * that a small memzone is correctly reserved from 16MB huge pages + * when requested by the RTE_MEMZONE_16MB flag. Also check that + * RTE_MEMZONE_SIZE_HINT_ONLY flag only defaults to an available + * page size (i.e 16GB ) when 16MB pages are unavailable. + */ + if (hugepage_16MB_avail) { + mz = rte_memzone_reserve(TEST_MEMZONE_NAME("flag_zone_16M"), + size, SOCKET_ID_ANY, RTE_MEMZONE_16MB); + if (mz == NULL) { + printf("MEMZONE FLAG 16MB\n"); + return -1; + } + if (mz->hugepage_sz != RTE_PGSIZE_16M) { + printf("hugepage_sz not equal 16M\n"); + return -1; + } + if (rte_memzone_free(mz)) { + printf("Fail memzone free\n"); + return -1; + } + + mz = rte_memzone_reserve( + TEST_MEMZONE_NAME("flag_zone_16M_HINT"), size, + SOCKET_ID_ANY, + RTE_MEMZONE_16MB|RTE_MEMZONE_SIZE_HINT_ONLY); + if (mz == NULL) { + printf("MEMZONE FLAG 16MB\n"); + return -1; + } + if (mz->hugepage_sz != RTE_PGSIZE_16M) { + printf("hugepage_sz not equal 16M\n"); + return -1; + } + if (rte_memzone_free(mz)) { + printf("Fail memzone free\n"); + return -1; + } + + /* Check if 1GB huge pages are unavailable, that function fails + * unless HINT flag is indicated + */ + if (!hugepage_16GB_avail) { + mz = rte_memzone_reserve( + TEST_MEMZONE_NAME("flag_zone_16G_HINT"), + size, SOCKET_ID_ANY, + RTE_MEMZONE_16GB | + RTE_MEMZONE_SIZE_HINT_ONLY); + if (mz == NULL) { + printf("MEMZONE FLAG 16GB & HINT\n"); + return -1; + } + if (mz->hugepage_sz != RTE_PGSIZE_16M) { + printf("hugepage_sz not equal 16M\n"); + return -1; + } + if (rte_memzone_free(mz)) { + printf("Fail memzone free\n"); + return -1; + } + + mz = rte_memzone_reserve( + TEST_MEMZONE_NAME("flag_zone_16G"), + size, + SOCKET_ID_ANY, RTE_MEMZONE_16GB); + if (mz != NULL) { + printf("MEMZONE FLAG 16GB\n"); + return -1; + } + } + } + /*As with 16MB tests above for 16GB huge page requests*/ + if (hugepage_16GB_avail) { + mz = rte_memzone_reserve(TEST_MEMZONE_NAME("flag_zone_16G"), + size, SOCKET_ID_ANY, RTE_MEMZONE_16GB); + if (mz == NULL) { + printf("MEMZONE FLAG 16GB\n"); + return -1; + } + if (mz->hugepage_sz != RTE_PGSIZE_16G) { + printf("hugepage_sz not equal 16G\n"); + return -1; + } + if (rte_memzone_free(mz)) { + printf("Fail memzone free\n"); + return -1; + } + + mz = rte_memzone_reserve( + TEST_MEMZONE_NAME("flag_zone_16G_HINT"), size, + SOCKET_ID_ANY, + RTE_MEMZONE_16GB|RTE_MEMZONE_SIZE_HINT_ONLY); + if (mz == NULL) { + printf("MEMZONE FLAG 16GB\n"); + return -1; + } + if (mz->hugepage_sz != RTE_PGSIZE_16G) { + printf("hugepage_sz not equal 16G\n"); + return -1; + } + if (rte_memzone_free(mz)) { + printf("Fail memzone free\n"); + return -1; + } + + /* Check if 1GB huge pages are unavailable, that function fails + * unless HINT flag is indicated + */ + if (!hugepage_16MB_avail) { + mz = rte_memzone_reserve( + TEST_MEMZONE_NAME("flag_zone_16M_HINT"), + size, SOCKET_ID_ANY, + RTE_MEMZONE_16MB | + RTE_MEMZONE_SIZE_HINT_ONLY); + if (mz == NULL) { + printf("MEMZONE FLAG 16MB & HINT\n"); + return -1; + } + if (mz->hugepage_sz != RTE_PGSIZE_16G) { + printf("hugepage_sz not equal 16G\n"); + return -1; + } + if (rte_memzone_free(mz)) { + printf("Fail memzone free\n"); + return -1; + } + mz = rte_memzone_reserve( + TEST_MEMZONE_NAME("flag_zone_16M"), + size, SOCKET_ID_ANY, RTE_MEMZONE_16MB); if (mz != NULL) { + printf("MEMZONE FLAG 16MB\n"); + return -1; + } + } + + if (hugepage_16MB_avail && hugepage_16GB_avail) { + mz = rte_memzone_reserve( + TEST_MEMZONE_NAME("flag_zone_16M_HINT"), + size, SOCKET_ID_ANY, + RTE_MEMZONE_16MB|RTE_MEMZONE_16GB); + if (mz == NULL) { printf("BOTH SIZES SET\n"); return -1; } + if (mz->hugepage_sz != RTE_PGSIZE_16G && + mz->hugepage_sz != RTE_PGSIZE_16M) { + printf("Wrong size when both sizes set\n"); + return -1; + } + if (rte_memzone_free(mz)) { + printf("Fail memzone free\n"); + return -1; + } } } return 0; } + +/* Find the heap with the greatest free block size */ +static size_t +find_max_block_free_size(unsigned int align, unsigned int socket_id) +{ + struct rte_malloc_socket_stats stats; + size_t len, overhead; + + rte_malloc_get_socket_stats(socket_id, &stats); + + len = stats.greatest_free_size; + overhead = MALLOC_ELEM_OVERHEAD; + + if (len == 0) + return 0; + + align = RTE_CACHE_LINE_ROUNDUP(align); + overhead += align; + + if (len < overhead) + return 0; + + return len - overhead; +} + static int test_memzone_reserve_max(void) { - const struct rte_memzone *mz; - const struct rte_config *config; - const struct rte_memseg *ms; - int memseg_idx = 0; - int memzone_idx = 0; - size_t len = 0; - void* last_addr; - size_t maxlen = 0; - - /* get pointer to global configuration */ - config = rte_eal_get_configuration(); - - ms = rte_eal_get_physmem_layout(); - - for (memseg_idx = 0; memseg_idx < RTE_MAX_MEMSEG; memseg_idx++){ - /* ignore smaller memsegs as they can only get smaller */ - if (ms[memseg_idx].len < maxlen) - continue; - - /* align everything */ - last_addr = RTE_PTR_ALIGN_CEIL(ms[memseg_idx].addr, CACHE_LINE_SIZE); - len = ms[memseg_idx].len - RTE_PTR_DIFF(last_addr, ms[memseg_idx].addr); - len &= ~((size_t) CACHE_LINE_MASK); - - /* cycle through all memzones */ - for (memzone_idx = 0; memzone_idx < RTE_MAX_MEMZONE; memzone_idx++) { - - /* stop when reaching last allocated memzone */ - if (config->mem_config->memzone[memzone_idx].addr == NULL) - break; - - /* check if the memzone is in our memseg and subtract length */ - if ((config->mem_config->memzone[memzone_idx].addr >= - ms[memseg_idx].addr) && - (config->mem_config->memzone[memzone_idx].addr < - (RTE_PTR_ADD(ms[memseg_idx].addr, ms[memseg_idx].len)))) { - /* since the zones can now be aligned and occasionally skip - * some space, we should calculate the length based on - * reported length and start addresses difference. Addresses - * are allocated sequentially so we don't need to worry about - * them being in the right order. - */ - len -= RTE_PTR_DIFF( - config->mem_config->memzone[memzone_idx].addr, - last_addr); - len -= config->mem_config->memzone[memzone_idx].len; - last_addr = RTE_PTR_ADD(config->mem_config->memzone[memzone_idx].addr, - (size_t) config->mem_config->memzone[memzone_idx].len); - } - } + unsigned int i; - /* we don't need to calculate offset here since length - * is always cache-aligned */ - if (len > maxlen) - maxlen = len; - } + for (i = 0; i < rte_socket_count(); i++) { + const struct rte_memzone *mz; + size_t maxlen; + int socket; - if (maxlen == 0) { - printf("There is no space left!\n"); - return 0; - } + socket = rte_socket_id_by_idx(i); + maxlen = find_max_block_free_size(0, socket); - mz = rte_memzone_reserve("max_zone", 0, SOCKET_ID_ANY, 0); - if (mz == NULL){ - printf("Failed to reserve a big chunk of memory\n"); - rte_dump_physmem_layout(stdout); - rte_memzone_dump(stdout); - return -1; - } + if (maxlen == 0) { + printf("There is no space left!\n"); + return 0; + } - if (mz->len != maxlen) { - printf("Memzone reserve with 0 size did not return bigest block\n"); - printf("Expected size = %zu, actual size = %zu\n", - maxlen, mz->len); - rte_dump_physmem_layout(stdout); - rte_memzone_dump(stdout); + mz = rte_memzone_reserve(TEST_MEMZONE_NAME("max_zone"), 0, + socket, 0); + if (mz == NULL) { + printf("Failed to reserve a big chunk of memory - %s\n", + rte_strerror(rte_errno)); + rte_dump_physmem_layout(stdout); + rte_memzone_dump(stdout); + return -1; + } - return -1; + if (mz->len != maxlen) { + printf("Memzone reserve with 0 size did not return bigest block\n"); + printf("Expected size = %zu, actual size = %zu\n", + maxlen, mz->len); + rte_dump_physmem_layout(stdout); + rte_memzone_dump(stdout); + return -1; + } + + if (rte_memzone_free(mz)) { + printf("Fail memzone free\n"); + return -1; + } } + return 0; } static int test_memzone_reserve_max_aligned(void) { - const struct rte_memzone *mz; - const struct rte_config *config; - const struct rte_memseg *ms; - int memseg_idx = 0; - int memzone_idx = 0; - uintptr_t addr_offset; - size_t len = 0; - void* last_addr; - size_t maxlen = 0; - - /* random alignment */ - rte_srand((unsigned)rte_rdtsc()); - const unsigned align = 1 << ((rte_rand() % 8) + 5); /* from 128 up to 4k alignment */ - - /* get pointer to global configuration */ - config = rte_eal_get_configuration(); - - ms = rte_eal_get_physmem_layout(); - - addr_offset = 0; - - for (memseg_idx = 0; memseg_idx < RTE_MAX_MEMSEG; memseg_idx++){ - - /* ignore smaller memsegs as they can only get smaller */ - if (ms[memseg_idx].len < maxlen) - continue; - - /* align everything */ - last_addr = RTE_PTR_ALIGN_CEIL(ms[memseg_idx].addr, CACHE_LINE_SIZE); - len = ms[memseg_idx].len - RTE_PTR_DIFF(last_addr, ms[memseg_idx].addr); - len &= ~((size_t) CACHE_LINE_MASK); - - /* cycle through all memzones */ - for (memzone_idx = 0; memzone_idx < RTE_MAX_MEMZONE; memzone_idx++) { - - /* stop when reaching last allocated memzone */ - if (config->mem_config->memzone[memzone_idx].addr == NULL) - break; - - /* check if the memzone is in our memseg and subtract length */ - if ((config->mem_config->memzone[memzone_idx].addr >= - ms[memseg_idx].addr) && - (config->mem_config->memzone[memzone_idx].addr < - (RTE_PTR_ADD(ms[memseg_idx].addr, ms[memseg_idx].len)))) { - /* since the zones can now be aligned and occasionally skip - * some space, we should calculate the length based on - * reported length and start addresses difference. - */ - len -= (uintptr_t) RTE_PTR_SUB( - config->mem_config->memzone[memzone_idx].addr, - (uintptr_t) last_addr); - len -= config->mem_config->memzone[memzone_idx].len; - last_addr = - RTE_PTR_ADD(config->mem_config->memzone[memzone_idx].addr, - (size_t) config->mem_config->memzone[memzone_idx].len); - } - } + unsigned int i; - /* make sure we get the alignment offset */ - if (len > maxlen) { - addr_offset = RTE_PTR_ALIGN_CEIL((uintptr_t) last_addr, align) - (uintptr_t) last_addr; - maxlen = len; - } - } + for (i = 0; i < rte_socket_count(); i++) { + const struct rte_memzone *mz; + size_t maxlen, minlen = 0; + int socket; - if (maxlen == 0 || maxlen == addr_offset) { - printf("There is no space left for biggest %u-aligned memzone!\n", align); - return 0; - } + socket = rte_socket_id_by_idx(i); - maxlen -= addr_offset; + /* random alignment */ + rte_srand((unsigned int)rte_rdtsc()); + const unsigned int align = 1 << ((rte_rand() % 8) + 5); /* from 128 up to 4k alignment */ - mz = rte_memzone_reserve_aligned("max_zone_aligned", 0, - SOCKET_ID_ANY, 0, align); - if (mz == NULL){ - printf("Failed to reserve a big chunk of memory\n"); - rte_dump_physmem_layout(stdout); - rte_memzone_dump(stdout); - return -1; - } + /* memzone size may be between size and size - align */ + minlen = find_max_block_free_size(align, socket); + maxlen = find_max_block_free_size(0, socket); - if (mz->len != maxlen) { - printf("Memzone reserve with 0 size and alignment %u did not return" - " bigest block\n", align); - printf("Expected size = %zu, actual size = %zu\n", - maxlen, mz->len); - rte_dump_physmem_layout(stdout); - rte_memzone_dump(stdout); + if (minlen == 0 || maxlen == 0) { + printf("There is no space left for biggest %u-aligned memzone!\n", + align); + return 0; + } - return -1; + mz = rte_memzone_reserve_aligned( + TEST_MEMZONE_NAME("max_zone_aligned"), + 0, socket, 0, align); + if (mz == NULL) { + printf("Failed to reserve a big chunk of memory - %s\n", + rte_strerror(rte_errno)); + rte_dump_physmem_layout(stdout); + rte_memzone_dump(stdout); + return -1; + } + if (mz->addr != RTE_PTR_ALIGN(mz->addr, align)) { + printf("Memzone reserve with 0 size and alignment %u did not return aligned block\n", + align); + rte_dump_physmem_layout(stdout); + rte_memzone_dump(stdout); + return -1; + } + + if (mz->len < minlen || mz->len > maxlen) { + printf("Memzone reserve with 0 size and alignment %u did not return" + " bigest block\n", align); + printf("Expected size = %zu-%zu, actual size = %zu\n", + minlen, maxlen, mz->len); + rte_dump_physmem_layout(stdout); + rte_memzone_dump(stdout); + return -1; + } + + if (rte_memzone_free(mz)) { + printf("Fail memzone free\n"); + return -1; + } } return 0; } @@ -450,113 +610,140 @@ test_memzone_aligned(void) const struct rte_memzone *memzone_aligned_1024; /* memzone that should automatically be adjusted to align on 64 bytes */ - memzone_aligned_32 = rte_memzone_reserve_aligned("aligned_32", 100, - SOCKET_ID_ANY, 0, 32); + memzone_aligned_32 = rte_memzone_reserve_aligned( + TEST_MEMZONE_NAME("aligned_32"), 100, SOCKET_ID_ANY, 0, + 32); /* memzone that is supposed to be aligned on a 128 byte boundary */ - memzone_aligned_128 = rte_memzone_reserve_aligned("aligned_128", 100, - SOCKET_ID_ANY, 0, 128); + memzone_aligned_128 = rte_memzone_reserve_aligned( + TEST_MEMZONE_NAME("aligned_128"), 100, SOCKET_ID_ANY, 0, + 128); /* memzone that is supposed to be aligned on a 256 byte boundary */ - memzone_aligned_256 = rte_memzone_reserve_aligned("aligned_256", 100, - SOCKET_ID_ANY, 0, 256); + memzone_aligned_256 = rte_memzone_reserve_aligned( + TEST_MEMZONE_NAME("aligned_256"), 100, SOCKET_ID_ANY, 0, + 256); /* memzone that is supposed to be aligned on a 512 byte boundary */ - memzone_aligned_512 = rte_memzone_reserve_aligned("aligned_512", 100, - SOCKET_ID_ANY, 0, 512); + memzone_aligned_512 = rte_memzone_reserve_aligned( + TEST_MEMZONE_NAME("aligned_512"), 100, SOCKET_ID_ANY, 0, + 512); /* memzone that is supposed to be aligned on a 1024 byte boundary */ - memzone_aligned_1024 = rte_memzone_reserve_aligned("aligned_1024", 100, - SOCKET_ID_ANY, 0, 1024); + memzone_aligned_1024 = rte_memzone_reserve_aligned( + TEST_MEMZONE_NAME("aligned_1024"), 100, SOCKET_ID_ANY, + 0, 1024); printf("check alignments and lengths\n"); if (memzone_aligned_32 == NULL) { printf("Unable to reserve 64-byte aligned memzone!\n"); return -1; } - if ((memzone_aligned_32->phys_addr & CACHE_LINE_MASK) != 0) + if ((memzone_aligned_32->iova & RTE_CACHE_LINE_MASK) != 0) return -1; - if (((uintptr_t) memzone_aligned_32->addr & CACHE_LINE_MASK) != 0) + if (((uintptr_t) memzone_aligned_32->addr & RTE_CACHE_LINE_MASK) != 0) return -1; - if ((memzone_aligned_32->len & CACHE_LINE_MASK) != 0) + if ((memzone_aligned_32->len & RTE_CACHE_LINE_MASK) != 0) return -1; if (memzone_aligned_128 == NULL) { printf("Unable to reserve 128-byte aligned memzone!\n"); return -1; } - if ((memzone_aligned_128->phys_addr & 127) != 0) + if ((memzone_aligned_128->iova & 127) != 0) return -1; if (((uintptr_t) memzone_aligned_128->addr & 127) != 0) return -1; - if ((memzone_aligned_128->len & CACHE_LINE_MASK) != 0) + if ((memzone_aligned_128->len & RTE_CACHE_LINE_MASK) != 0) return -1; if (memzone_aligned_256 == NULL) { printf("Unable to reserve 256-byte aligned memzone!\n"); return -1; } - if ((memzone_aligned_256->phys_addr & 255) != 0) + if ((memzone_aligned_256->iova & 255) != 0) return -1; if (((uintptr_t) memzone_aligned_256->addr & 255) != 0) return -1; - if ((memzone_aligned_256->len & CACHE_LINE_MASK) != 0) + if ((memzone_aligned_256->len & RTE_CACHE_LINE_MASK) != 0) return -1; if (memzone_aligned_512 == NULL) { printf("Unable to reserve 512-byte aligned memzone!\n"); return -1; } - if ((memzone_aligned_512->phys_addr & 511) != 0) + if ((memzone_aligned_512->iova & 511) != 0) return -1; if (((uintptr_t) memzone_aligned_512->addr & 511) != 0) return -1; - if ((memzone_aligned_512->len & CACHE_LINE_MASK) != 0) + if ((memzone_aligned_512->len & RTE_CACHE_LINE_MASK) != 0) return -1; if (memzone_aligned_1024 == NULL) { printf("Unable to reserve 1024-byte aligned memzone!\n"); return -1; } - if ((memzone_aligned_1024->phys_addr & 1023) != 0) + if ((memzone_aligned_1024->iova & 1023) != 0) return -1; if (((uintptr_t) memzone_aligned_1024->addr & 1023) != 0) return -1; - if ((memzone_aligned_1024->len & CACHE_LINE_MASK) != 0) + if ((memzone_aligned_1024->len & RTE_CACHE_LINE_MASK) != 0) return -1; /* check that zones don't overlap */ printf("check overlapping\n"); - if (is_memory_overlap(memzone_aligned_32->phys_addr, memzone_aligned_32->len, - memzone_aligned_128->phys_addr, memzone_aligned_128->len)) + if (is_memory_overlap(memzone_aligned_32->iova, memzone_aligned_32->len, + memzone_aligned_128->iova, memzone_aligned_128->len)) return -1; - if (is_memory_overlap(memzone_aligned_32->phys_addr, memzone_aligned_32->len, - memzone_aligned_256->phys_addr, memzone_aligned_256->len)) + if (is_memory_overlap(memzone_aligned_32->iova, memzone_aligned_32->len, + memzone_aligned_256->iova, memzone_aligned_256->len)) return -1; - if (is_memory_overlap(memzone_aligned_32->phys_addr, memzone_aligned_32->len, - memzone_aligned_512->phys_addr, memzone_aligned_512->len)) + if (is_memory_overlap(memzone_aligned_32->iova, memzone_aligned_32->len, + memzone_aligned_512->iova, memzone_aligned_512->len)) return -1; - if (is_memory_overlap(memzone_aligned_32->phys_addr, memzone_aligned_32->len, - memzone_aligned_1024->phys_addr, memzone_aligned_1024->len)) + if (is_memory_overlap(memzone_aligned_32->iova, memzone_aligned_32->len, + memzone_aligned_1024->iova, memzone_aligned_1024->len)) return -1; - if (is_memory_overlap(memzone_aligned_128->phys_addr, memzone_aligned_128->len, - memzone_aligned_256->phys_addr, memzone_aligned_256->len)) + if (is_memory_overlap(memzone_aligned_128->iova, memzone_aligned_128->len, + memzone_aligned_256->iova, memzone_aligned_256->len)) return -1; - if (is_memory_overlap(memzone_aligned_128->phys_addr, memzone_aligned_128->len, - memzone_aligned_512->phys_addr, memzone_aligned_512->len)) + if (is_memory_overlap(memzone_aligned_128->iova, memzone_aligned_128->len, + memzone_aligned_512->iova, memzone_aligned_512->len)) return -1; - if (is_memory_overlap(memzone_aligned_128->phys_addr, memzone_aligned_128->len, - memzone_aligned_1024->phys_addr, memzone_aligned_1024->len)) + if (is_memory_overlap(memzone_aligned_128->iova, memzone_aligned_128->len, + memzone_aligned_1024->iova, memzone_aligned_1024->len)) return -1; - if (is_memory_overlap(memzone_aligned_256->phys_addr, memzone_aligned_256->len, - memzone_aligned_512->phys_addr, memzone_aligned_512->len)) + if (is_memory_overlap(memzone_aligned_256->iova, memzone_aligned_256->len, + memzone_aligned_512->iova, memzone_aligned_512->len)) return -1; - if (is_memory_overlap(memzone_aligned_256->phys_addr, memzone_aligned_256->len, - memzone_aligned_1024->phys_addr, memzone_aligned_1024->len)) + if (is_memory_overlap(memzone_aligned_256->iova, memzone_aligned_256->len, + memzone_aligned_1024->iova, memzone_aligned_1024->len)) return -1; - if (is_memory_overlap(memzone_aligned_512->phys_addr, memzone_aligned_512->len, - memzone_aligned_1024->phys_addr, memzone_aligned_1024->len)) + if (is_memory_overlap(memzone_aligned_512->iova, memzone_aligned_512->len, + memzone_aligned_1024->iova, memzone_aligned_1024->len)) return -1; + + /* free all used zones */ + if (rte_memzone_free(memzone_aligned_32)) { + printf("Fail memzone free\n"); + return -1; + } + if (rte_memzone_free(memzone_aligned_128)) { + printf("Fail memzone free\n"); + return -1; + } + if (rte_memzone_free(memzone_aligned_256)) { + printf("Fail memzone free\n"); + return -1; + } + if (rte_memzone_free(memzone_aligned_512)) { + printf("Fail memzone free\n"); + return -1; + } + if (rte_memzone_free(memzone_aligned_1024)) { + printf("Fail memzone free\n"); + return -1; + } return 0; } @@ -565,404 +752,237 @@ check_memzone_bounded(const char *name, uint32_t len, uint32_t align, uint32_t bound) { const struct rte_memzone *mz; - phys_addr_t bmask; + rte_iova_t bmask; - bmask = ~((phys_addr_t)bound - 1); + bmask = ~((rte_iova_t)bound - 1); - if ((mz = rte_memzone_reserve_bounded(name, len, SOCKET_ID_ANY, 0, + if ((mz = rte_memzone_reserve_bounded(name, len, SOCKET_ID_ANY, 0, align, bound)) == NULL) { printf("%s(%s): memzone creation failed\n", __func__, name); - return (-1); + return -1; } - if ((mz->phys_addr & ((phys_addr_t)align - 1)) != 0) { + if ((mz->iova & ((rte_iova_t)align - 1)) != 0) { printf("%s(%s): invalid phys addr alignment\n", __func__, mz->name); - return (-1); + return -1; } if (((uintptr_t) mz->addr & ((uintptr_t)align - 1)) != 0) { printf("%s(%s): invalid virtual addr alignment\n", __func__, mz->name); - return (-1); + return -1; } - if ((mz->len & CACHE_LINE_MASK) != 0 || mz->len < len || - mz->len < CACHE_LINE_SIZE) { + if ((mz->len & RTE_CACHE_LINE_MASK) != 0 || mz->len < len || + mz->len < RTE_CACHE_LINE_SIZE) { printf("%s(%s): invalid length\n", __func__, mz->name); - return (-1); + return -1; } - if ((mz->phys_addr & bmask) != - ((mz->phys_addr + mz->len - 1) & bmask)) { + if ((mz->iova & bmask) != + ((mz->iova + mz->len - 1) & bmask)) { printf("%s(%s): invalid memzone boundary %u crossed\n", __func__, mz->name, bound); - return (-1); + return -1; } - return (0); + if (rte_memzone_free(mz)) { + printf("Fail memzone free\n"); + return -1; + } + + return 0; } static int test_memzone_bounded(void) { const struct rte_memzone *memzone_err; - const char *name; int rc; /* should fail as boundary is not power of two */ - name = "bounded_error_31"; - if ((memzone_err = rte_memzone_reserve_bounded(name, - 100, SOCKET_ID_ANY, 0, 32, UINT32_MAX)) != NULL) { + memzone_err = rte_memzone_reserve_bounded( + TEST_MEMZONE_NAME("bounded_error_31"), 100, + SOCKET_ID_ANY, 0, 32, UINT32_MAX); + if (memzone_err != NULL) { printf("%s(%s)created a memzone with invalid boundary " "conditions\n", __func__, memzone_err->name); - return (-1); + return -1; } - + /* should fail as len is greater then boundary */ - name = "bounded_error_32"; - if ((memzone_err = rte_memzone_reserve_bounded(name, - 100, SOCKET_ID_ANY, 0, 32, 32)) != NULL) { + memzone_err = rte_memzone_reserve_bounded( + TEST_MEMZONE_NAME("bounded_error_32"), 100, + SOCKET_ID_ANY, 0, 32, 32); + if (memzone_err != NULL) { printf("%s(%s)created a memzone with invalid boundary " "conditions\n", __func__, memzone_err->name); - return (-1); + return -1; } - if ((rc = check_memzone_bounded("bounded_128", 100, 128, 128)) != 0) - return (rc); + rc = check_memzone_bounded(TEST_MEMZONE_NAME("bounded_128"), 100, 128, + 128); + if (rc != 0) + return rc; - if ((rc = check_memzone_bounded("bounded_256", 100, 256, 128)) != 0) - return (rc); + rc = check_memzone_bounded(TEST_MEMZONE_NAME("bounded_256"), 100, 256, + 128); + if (rc != 0) + return rc; - if ((rc = check_memzone_bounded("bounded_1K", 100, 64, 1024)) != 0) - return (rc); + rc = check_memzone_bounded(TEST_MEMZONE_NAME("bounded_1K"), 100, 64, + 1024); + if (rc != 0) + return rc; - if ((rc = check_memzone_bounded("bounded_1K_MAX", 0, 64, 1024)) != 0) - return (rc); + rc = check_memzone_bounded(TEST_MEMZONE_NAME("bounded_1K_MAX"), 0, 64, + 1024); + if (rc != 0) + return rc; - return (0); + return 0; } static int -test_memzone_reserve_memory_in_smallest_segment(void) +test_memzone_free(void) { - const struct rte_memzone *mz; - const struct rte_memseg *ms, *min_ms, *prev_min_ms; - size_t min_len, prev_min_len; - const struct rte_config *config; + const struct rte_memzone *mz[RTE_MAX_MEMZONE + 1]; int i; + char name[20]; - config = rte_eal_get_configuration(); - - min_ms = NULL; /*< smallest segment */ - prev_min_ms = NULL; /*< second smallest segment */ - - /* find two smallest segments */ - for (i = 0; i < RTE_MAX_MEMSEG; i++) { - ms = &config->mem_config->free_memseg[i]; - - if (ms->addr == NULL) - break; - if (ms->len == 0) - continue; - - if (min_ms == NULL) - min_ms = ms; - else if (min_ms->len > ms->len) { - /* set last smallest to second last */ - prev_min_ms = min_ms; - - /* set new smallest */ - min_ms = ms; - } - else if (prev_min_ms == NULL) { - prev_min_ms = ms; - } - } + mz[0] = rte_memzone_reserve(TEST_MEMZONE_NAME("tempzone0"), 2000, + SOCKET_ID_ANY, 0); + mz[1] = rte_memzone_reserve(TEST_MEMZONE_NAME("tempzone1"), 4000, + SOCKET_ID_ANY, 0); - if (min_ms == NULL || prev_min_ms == NULL) { - printf("Smallest segments not found!\n"); + if (mz[0] > mz[1]) return -1; - } - - min_len = min_ms->len; - prev_min_len = prev_min_ms->len; - - /* try reserving a memzone in the smallest memseg */ - mz = rte_memzone_reserve("smallest_mz", CACHE_LINE_SIZE, - SOCKET_ID_ANY, 0); - if (mz == NULL) { - printf("Failed to reserve memory from smallest memseg!\n"); + if (!rte_memzone_lookup(TEST_MEMZONE_NAME("tempzone0"))) return -1; - } - if (prev_min_ms->len != prev_min_len && - min_ms->len != min_len - CACHE_LINE_SIZE) { - printf("Reserved memory from wrong memseg!\n"); + if (!rte_memzone_lookup(TEST_MEMZONE_NAME("tempzone1"))) return -1; - } - - return 0; -} -/* this test is a bit tricky, and thus warrants explanation. - * - * first, we find two smallest memsegs to conduct our experiments on. - * - * then, we bring them within alignment from each other: if second segment is - * twice+ as big as the first, reserve memory from that segment; if second - * segment is comparable in length to the first, then cut the first segment - * down until it becomes less than half of second segment, and then cut down - * the second segment to be within alignment of the first. - * - * then, we have to pass the following test: if segments are within alignment - * of each other (that is, the difference is less than 256 bytes, which is what - * our alignment will be), segment with smallest offset should be picked. - * - * we know that min_ms will be our smallest segment, so we need to make sure - * that we adjust the alignments so that the bigger segment has smallest - * alignment (in our case, smallest segment will have 64-byte alignment, while - * bigger segment will have 128-byte alignment). - */ -static int -test_memzone_reserve_memory_with_smallest_offset(void) -{ - const struct rte_memseg *ms, *min_ms, *prev_min_ms; - size_t len, min_len, prev_min_len; - const struct rte_config *config; - int i, align; - - config = rte_eal_get_configuration(); - - min_ms = NULL; /*< smallest segment */ - prev_min_ms = NULL; /*< second smallest segment */ - align = CACHE_LINE_SIZE * 4; - - /* find two smallest segments */ - for (i = 0; i < RTE_MAX_MEMSEG; i++) { - ms = &config->mem_config->free_memseg[i]; - - if (ms->addr == NULL) - break; - if (ms->len == 0) - continue; - - if (min_ms == NULL) - min_ms = ms; - else if (min_ms->len > ms->len) { - /* set last smallest to second last */ - prev_min_ms = min_ms; - - /* set new smallest */ - min_ms = ms; - } - else if (prev_min_ms == NULL) { - prev_min_ms = ms; - } + if (rte_memzone_free(mz[0])) { + printf("Fail memzone free - tempzone0\n"); + return -1; } - - if (min_ms == NULL || prev_min_ms == NULL) { - printf("Smallest segments not found!\n"); + if (rte_memzone_lookup(TEST_MEMZONE_NAME("tempzone0"))) { + printf("Found previously free memzone - tempzone0\n"); return -1; } + mz[2] = rte_memzone_reserve(TEST_MEMZONE_NAME("tempzone2"), 2000, + SOCKET_ID_ANY, 0); - prev_min_len = prev_min_ms->len; - min_len = min_ms->len; - - /* if smallest segment is bigger than half of bigger segment */ - if (prev_min_ms->len - min_ms->len <= min_ms->len) { - - len = (min_ms->len * 2) - prev_min_ms->len; - - /* make sure final length is *not* aligned */ - while (((min_ms->addr_64 + len) & (align-1)) == 0) - len += CACHE_LINE_SIZE; - - if (rte_memzone_reserve("dummy_mz1", len, SOCKET_ID_ANY, 0) == NULL) { - printf("Cannot reserve memory!\n"); - return -1; - } - - /* check if we got memory from correct segment */ - if (min_ms->len != min_len - len) { - printf("Reserved memory from wrong segment!\n"); - return -1; - } + if (mz[2] > mz[1]) { + printf("tempzone2 should have gotten the free entry from tempzone0\n"); + return -1; } - /* if we don't need to touch smallest segment but it's aligned */ - else if ((min_ms->addr_64 & (align-1)) == 0) { - if (rte_memzone_reserve("align_mz1", CACHE_LINE_SIZE, - SOCKET_ID_ANY, 0) == NULL) { - printf("Cannot reserve memory!\n"); - return -1; - } - if (min_ms->len != min_len - CACHE_LINE_SIZE) { - printf("Reserved memory from wrong segment!\n"); - return -1; - } - } - - /* if smallest segment is less than half of bigger segment */ - if (prev_min_ms->len - min_ms->len > min_ms->len) { - len = prev_min_ms->len - min_ms->len - align; - - /* make sure final length is aligned */ - while (((prev_min_ms->addr_64 + len) & (align-1)) != 0) - len += CACHE_LINE_SIZE; - - if (rte_memzone_reserve("dummy_mz2", len, SOCKET_ID_ANY, 0) == NULL) { - printf("Cannot reserve memory!\n"); - return -1; - } - - /* check if we got memory from correct segment */ - if (prev_min_ms->len != prev_min_len - len) { - printf("Reserved memory from wrong segment!\n"); - return -1; - } + if (rte_memzone_free(mz[2])) { + printf("Fail memzone free - tempzone2\n"); + return -1; } - len = CACHE_LINE_SIZE; - - - - prev_min_len = prev_min_ms->len; - min_len = min_ms->len; - - if (min_len >= prev_min_len || prev_min_len - min_len > (unsigned) align) { - printf("Segments are of wrong lengths!\n"); + if (rte_memzone_lookup(TEST_MEMZONE_NAME("tempzone2"))) { + printf("Found previously free memzone - tempzone2\n"); return -1; } - - /* try reserving from a bigger segment */ - if (rte_memzone_reserve_aligned("smallest_offset", len, SOCKET_ID_ANY, 0, align) == - NULL) { - printf("Cannot reserve memory!\n"); + if (rte_memzone_free(mz[1])) { + printf("Fail memzone free - tempzone1\n"); return -1; } - - /* check if we got memory from correct segment */ - if (min_ms->len != min_len && prev_min_ms->len != (prev_min_len - len)) { - printf("Reserved memory from segment with smaller offset!\n"); + if (rte_memzone_lookup(TEST_MEMZONE_NAME("tempzone1"))) { + printf("Found previously free memzone - tempzone1\n"); return -1; } - return 0; -} + i = 0; + do { + snprintf(name, sizeof(name), TEST_MEMZONE_NAME("tempzone%u"), + i); + mz[i] = rte_memzone_reserve(name, 1, SOCKET_ID_ANY, 0); + } while (mz[i++] != NULL); -static int -test_memzone_reserve_remainder(void) -{ - const struct rte_memzone *mz1, *mz2; - const struct rte_memseg *ms, *min_ms = NULL; - size_t min_len; - const struct rte_config *config; - int i, align; - - min_len = 0; - align = CACHE_LINE_SIZE; - - config = rte_eal_get_configuration(); - - /* find minimum free contiguous length */ - for (i = 0; i < RTE_MAX_MEMSEG; i++) { - ms = &config->mem_config->free_memseg[i]; - - if (ms->addr == NULL) - break; - if (ms->len == 0) - continue; - - if (min_len == 0 || ms->len < min_len) { - min_len = ms->len; - min_ms = ms; - - /* find maximum alignment this segment is able to hold */ - align = CACHE_LINE_SIZE; - while ((ms->addr_64 & (align-1)) == 0) { - align <<= 1; - } - } - } - - if (min_ms == NULL) { - printf("Minimal sized segment not found!\n"); + if (rte_memzone_free(mz[0])) { + printf("Fail memzone free - tempzone0\n"); return -1; } + mz[0] = rte_memzone_reserve(TEST_MEMZONE_NAME("tempzone0new"), 0, + SOCKET_ID_ANY, 0); - /* try reserving min_len bytes with alignment - this should not affect our - * memseg, the memory will be taken from a different one. - */ - mz1 = rte_memzone_reserve_aligned("reserve_remainder_1", min_len, - SOCKET_ID_ANY, 0, align); - if (mz1 == NULL) { - printf("Failed to reserve %zu bytes aligned on %i bytes\n", min_len, - align); - return -1; - } - if (min_ms->len != min_len) { - printf("Memseg memory should not have been reserved!\n"); + if (mz[0] == NULL) { + printf("Fail to create memzone - tempzone0new - when MAX memzones were " + "created and one was free\n"); return -1; } - /* try reserving min_len bytes with less alignment - this should fill up - * the segment. - */ - mz2 = rte_memzone_reserve("reserve_remainder_2", min_len, - SOCKET_ID_ANY, 0); - if (mz2 == NULL) { - printf("Failed to reserve %zu bytes\n", min_len); - return -1; - } - if (min_ms->len != 0) { - printf("Memseg memory should have been reserved!\n"); - return -1; + for (i = i - 2; i >= 0; i--) { + if (rte_memzone_free(mz[i])) { + printf("Fail memzone free - tempzone%d\n", i); + return -1; + } } return 0; } -int -test_memzone(void) +static int +test_memzone_basic(void) { const struct rte_memzone *memzone1; const struct rte_memzone *memzone2; const struct rte_memzone *memzone3; const struct rte_memzone *memzone4; const struct rte_memzone *mz; + int memzone_cnt_after, memzone_cnt_expected; + int memzone_cnt_before = + rte_eal_get_configuration()->mem_config->memzones.count; - memzone1 = rte_memzone_reserve("testzone1", 100, + memzone1 = rte_memzone_reserve(TEST_MEMZONE_NAME("testzone1"), 100, SOCKET_ID_ANY, 0); - memzone2 = rte_memzone_reserve("testzone2", 1000, + memzone2 = rte_memzone_reserve(TEST_MEMZONE_NAME("testzone2"), 1000, 0, 0); - memzone3 = rte_memzone_reserve("testzone3", 1000, + memzone3 = rte_memzone_reserve(TEST_MEMZONE_NAME("testzone3"), 1000, 1, 0); - memzone4 = rte_memzone_reserve("testzone4", 1024, + memzone4 = rte_memzone_reserve(TEST_MEMZONE_NAME("testzone4"), 1024, SOCKET_ID_ANY, 0); /* memzone3 may be NULL if we don't have NUMA */ if (memzone1 == NULL || memzone2 == NULL || memzone4 == NULL) return -1; + /* check how many memzones we are expecting */ + memzone_cnt_expected = memzone_cnt_before + + (memzone1 != NULL) + (memzone2 != NULL) + + (memzone3 != NULL) + (memzone4 != NULL); + + memzone_cnt_after = + rte_eal_get_configuration()->mem_config->memzones.count; + + if (memzone_cnt_after != memzone_cnt_expected) + return -1; + + rte_memzone_dump(stdout); /* check cache-line alignments */ printf("check alignments and lengths\n"); - if ((memzone1->phys_addr & CACHE_LINE_MASK) != 0) + if ((memzone1->iova & RTE_CACHE_LINE_MASK) != 0) return -1; - if ((memzone2->phys_addr & CACHE_LINE_MASK) != 0) + if ((memzone2->iova & RTE_CACHE_LINE_MASK) != 0) return -1; - if (memzone3 != NULL && (memzone3->phys_addr & CACHE_LINE_MASK) != 0) + if (memzone3 != NULL && (memzone3->iova & RTE_CACHE_LINE_MASK) != 0) return -1; - if ((memzone1->len & CACHE_LINE_MASK) != 0 || memzone1->len == 0) + if ((memzone1->len & RTE_CACHE_LINE_MASK) != 0 || memzone1->len == 0) return -1; - if ((memzone2->len & CACHE_LINE_MASK) != 0 || memzone2->len == 0) + if ((memzone2->len & RTE_CACHE_LINE_MASK) != 0 || memzone2->len == 0) return -1; - if (memzone3 != NULL && ((memzone3->len & CACHE_LINE_MASK) != 0 || + if (memzone3 != NULL && ((memzone3->len & RTE_CACHE_LINE_MASK) != 0 || memzone3->len == 0)) return -1; if (memzone4->len != 1024) @@ -971,16 +991,16 @@ test_memzone(void) /* check that zones don't overlap */ printf("check overlapping\n"); - if (is_memory_overlap(memzone1->phys_addr, memzone1->len, - memzone2->phys_addr, memzone2->len)) + if (is_memory_overlap(memzone1->iova, memzone1->len, + memzone2->iova, memzone2->len)) return -1; if (memzone3 != NULL && - is_memory_overlap(memzone1->phys_addr, memzone1->len, - memzone3->phys_addr, memzone3->len)) + is_memory_overlap(memzone1->iova, memzone1->len, + memzone3->iova, memzone3->len)) return -1; if (memzone3 != NULL && - is_memory_overlap(memzone2->phys_addr, memzone2->len, - memzone3->phys_addr, memzone3->len)) + is_memory_overlap(memzone2->iova, memzone2->len, + memzone3->iova, memzone3->len)) return -1; printf("check socket ID\n"); @@ -992,26 +1012,68 @@ test_memzone(void) return -1; printf("test zone lookup\n"); - mz = rte_memzone_lookup("testzone1"); + mz = rte_memzone_lookup(TEST_MEMZONE_NAME("testzone1")); if (mz != memzone1) return -1; printf("test duplcate zone name\n"); - mz = rte_memzone_reserve("testzone1", 100, + mz = rte_memzone_reserve(TEST_MEMZONE_NAME("testzone1"), 100, SOCKET_ID_ANY, 0); if (mz != NULL) return -1; - printf("test reserving memzone with bigger size than the maximum\n"); - if (test_memzone_reserving_zone_size_bigger_than_the_maximum() < 0) + if (rte_memzone_free(memzone1)) { + printf("Fail memzone free - memzone1\n"); return -1; + } + if (rte_memzone_free(memzone2)) { + printf("Fail memzone free - memzone2\n"); + return -1; + } + if (memzone3 && rte_memzone_free(memzone3)) { + printf("Fail memzone free - memzone3\n"); + return -1; + } + if (rte_memzone_free(memzone4)) { + printf("Fail memzone free - memzone4\n"); + return -1; + } - printf("test reserving memory in smallest segments\n"); - if (test_memzone_reserve_memory_in_smallest_segment() < 0) + memzone_cnt_after = + rte_eal_get_configuration()->mem_config->memzones.count; + if (memzone_cnt_after != memzone_cnt_before) return -1; - printf("test reserving memory in segments with smallest offsets\n"); - if (test_memzone_reserve_memory_with_smallest_offset() < 0) + return 0; +} + +static int test_memzones_left; +static int memzone_walk_cnt; +static void memzone_walk_clb(const struct rte_memzone *mz, + void *arg __rte_unused) +{ + memzone_walk_cnt++; + if (!strncmp(TEST_MEMZONE_NAME(""), mz->name, RTE_MEMZONE_NAMESIZE)) + test_memzones_left++; +} + +static int +test_memzone(void) +{ + /* take note of how many memzones were allocated before running */ + int memzone_cnt = + rte_eal_get_configuration()->mem_config->memzones.count; + + printf("test basic memzone API\n"); + if (test_memzone_basic() < 0) + return -1; + + printf("test free memzone\n"); + if (test_memzone_free() < 0) + return -1; + + printf("test reserving memzone with bigger size than the maximum\n"); + if (test_memzone_reserving_zone_size_bigger_than_the_maximum() < 0) return -1; printf("test memzone_reserve flags\n"); @@ -1030,10 +1092,6 @@ test_memzone(void) if (test_memzone_invalid_alignment() < 0) return -1; - printf("test reserving amounts of memory equal to segment's length\n"); - if (test_memzone_reserve_remainder() < 0) - return -1; - printf("test reserving the largest size memzone possible\n"); if (test_memzone_reserve_max() < 0) return -1; @@ -1042,5 +1100,17 @@ test_memzone(void) if (test_memzone_reserve_max_aligned() < 0) return -1; + printf("check memzone cleanup\n"); + memzone_walk_cnt = 0; + test_memzones_left = 0; + rte_memzone_walk(memzone_walk_clb, NULL); + if (memzone_walk_cnt != memzone_cnt || test_memzones_left > 0) { + printf("there are some memzones left after test\n"); + rte_memzone_dump(stdout); + return -1; + } + return 0; } + +REGISTER_TEST_COMMAND(memzone_autotest, test_memzone);