4 * Copyright(c) 2010-2014 Intel Corporation. All rights reserved.
7 * Redistribution and use in source and binary forms, with or without
8 * modification, are permitted provided that the following conditions
11 * * Redistributions of source code must retain the above copyright
12 * notice, this list of conditions and the following disclaimer.
13 * * Redistributions in binary form must reproduce the above copyright
14 * notice, this list of conditions and the following disclaimer in
15 * the documentation and/or other materials provided with the
17 * * Neither the name of Intel Corporation nor the names of its
18 * contributors may be used to endorse or promote products derived
19 * from this software without specific prior written permission.
21 * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
22 * "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
23 * LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
24 * A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
25 * OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
26 * SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
27 * LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
28 * DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
29 * THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
30 * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
31 * OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
39 #include <sys/queue.h>
41 #include <rte_memory.h>
43 #include <rte_eal_memconfig.h>
44 #include <rte_launch.h>
45 #include <rte_per_lcore.h>
46 #include <rte_lcore.h>
47 #include <rte_common.h>
48 #include <rte_string_fns.h>
49 #include <rte_spinlock.h>
50 #include <rte_memcpy.h>
51 #include <rte_atomic.h>
53 #include "malloc_elem.h"
54 #include "malloc_heap.h"
57 check_hugepage_sz(unsigned flags, uint64_t hugepage_sz)
59 unsigned check_flag = 0;
61 if (!(flags & ~RTE_MEMZONE_SIZE_HINT_ONLY))
64 switch (hugepage_sz) {
66 check_flag = RTE_MEMZONE_256KB;
69 check_flag = RTE_MEMZONE_2MB;
72 check_flag = RTE_MEMZONE_16MB;
75 check_flag = RTE_MEMZONE_256MB;
78 check_flag = RTE_MEMZONE_512MB;
81 check_flag = RTE_MEMZONE_1GB;
84 check_flag = RTE_MEMZONE_4GB;
87 check_flag = RTE_MEMZONE_16GB;
90 return check_flag & flags;
94 * Expand the heap with a memseg.
95 * This reserves the zone and sets a dummy malloc_elem header at the end
96 * to prevent overflow. The rest of the zone is added to free list as a single
100 malloc_heap_add_memseg(struct malloc_heap *heap, struct rte_memseg *ms)
102 /* allocate the memory block headers, one at end, one at start */
103 struct malloc_elem *start_elem = (struct malloc_elem *)ms->addr;
104 struct malloc_elem *end_elem = RTE_PTR_ADD(ms->addr,
105 ms->len - MALLOC_ELEM_OVERHEAD);
106 end_elem = RTE_PTR_ALIGN_FLOOR(end_elem, RTE_CACHE_LINE_SIZE);
107 const size_t elem_size = (uintptr_t)end_elem - (uintptr_t)start_elem;
109 malloc_elem_init(start_elem, heap, ms, elem_size);
110 malloc_elem_mkend(end_elem, start_elem);
111 malloc_elem_free_list_insert(start_elem);
113 heap->total_size += elem_size;
117 * Iterates through the freelist for a heap to find a free element
118 * which can store data of the required size and with the requested alignment.
119 * If size is 0, find the biggest available elem.
120 * Returns null on failure, or pointer to element on success.
122 static struct malloc_elem *
123 find_suitable_element(struct malloc_heap *heap, size_t size,
124 unsigned flags, size_t align, size_t bound)
127 struct malloc_elem *elem, *alt_elem = NULL;
129 for (idx = malloc_elem_free_list_index(size);
130 idx < RTE_HEAP_NUM_FREELISTS; idx++) {
131 for (elem = LIST_FIRST(&heap->free_head[idx]);
132 !!elem; elem = LIST_NEXT(elem, free_list)) {
133 if (malloc_elem_can_hold(elem, size, align, bound)) {
134 if (check_hugepage_sz(flags, elem->ms->hugepage_sz))
136 if (alt_elem == NULL)
142 if ((alt_elem != NULL) && (flags & RTE_MEMZONE_SIZE_HINT_ONLY))
149 * Main function to allocate a block of memory from the heap.
150 * It locks the free list, scans it, and adds a new memseg if the
151 * scan fails. Once the new memseg is added, it re-scans and should return
152 * the new element after releasing the lock.
155 malloc_heap_alloc(struct malloc_heap *heap,
156 const char *type __attribute__((unused)), size_t size, unsigned flags,
157 size_t align, size_t bound)
159 struct malloc_elem *elem;
161 size = RTE_CACHE_LINE_ROUNDUP(size);
162 align = RTE_CACHE_LINE_ROUNDUP(align);
164 rte_spinlock_lock(&heap->lock);
166 elem = find_suitable_element(heap, size, flags, align, bound);
168 elem = malloc_elem_alloc(elem, size, align, bound);
169 /* increase heap's count of allocated elements */
172 rte_spinlock_unlock(&heap->lock);
174 return elem == NULL ? NULL : (void *)(&elem[1]);
178 * Function to retrieve data for heap on given socket
181 malloc_heap_get_stats(const struct malloc_heap *heap,
182 struct rte_malloc_socket_stats *socket_stats)
185 struct malloc_elem *elem;
187 /* Initialise variables for heap */
188 socket_stats->free_count = 0;
189 socket_stats->heap_freesz_bytes = 0;
190 socket_stats->greatest_free_size = 0;
192 /* Iterate through free list */
193 for (idx = 0; idx < RTE_HEAP_NUM_FREELISTS; idx++) {
194 for (elem = LIST_FIRST(&heap->free_head[idx]);
195 !!elem; elem = LIST_NEXT(elem, free_list))
197 socket_stats->free_count++;
198 socket_stats->heap_freesz_bytes += elem->size;
199 if (elem->size > socket_stats->greatest_free_size)
200 socket_stats->greatest_free_size = elem->size;
203 /* Get stats on overall heap and allocated memory on this heap */
204 socket_stats->heap_totalsz_bytes = heap->total_size;
205 socket_stats->heap_allocsz_bytes = (socket_stats->heap_totalsz_bytes -
206 socket_stats->heap_freesz_bytes);
207 socket_stats->alloc_count = heap->alloc_count;
212 rte_eal_malloc_heap_init(void)
214 struct rte_mem_config *mcfg = rte_eal_get_configuration()->mem_config;
216 struct rte_memseg *ms;
221 for (ms = &mcfg->memseg[0], ms_cnt = 0;
222 (ms_cnt < RTE_MAX_MEMSEG) && (ms->len > 0);
224 malloc_heap_add_memseg(&mcfg->malloc_heaps[ms->socket_id], ms);