regex/mlx5: fix leak after probing failure
[dpdk.git] / app / test / test_spinlock.c
1 /* SPDX-License-Identifier: BSD-3-Clause
2  * Copyright(c) 2010-2014 Intel Corporation
3  */
4
5 #include <stdio.h>
6 #include <stdint.h>
7 #include <inttypes.h>
8 #include <string.h>
9 #include <unistd.h>
10 #include <sys/queue.h>
11
12 #include <rte_common.h>
13 #include <rte_memory.h>
14 #include <rte_per_lcore.h>
15 #include <rte_launch.h>
16 #include <rte_eal.h>
17 #include <rte_lcore.h>
18 #include <rte_cycles.h>
19 #include <rte_spinlock.h>
20
21 #include "test.h"
22
23 /*
24  * Spinlock test
25  * =============
26  *
27  * - There is a global spinlock and a table of spinlocks (one per lcore).
28  *
29  * - The test function takes all of these locks and launches the
30  *   ``test_spinlock_per_core()`` function on each core (except the main).
31  *
32  *   - The function takes the global lock, display something, then releases
33  *     the global lock.
34  *   - The function takes the per-lcore lock, display something, then releases
35  *     the per-core lock.
36  *
37  * - The main function unlocks the per-lcore locks sequentially and
38  *   waits between each lock. This triggers the display of a message
39  *   for each core, in the correct order. The autotest script checks that
40  *   this order is correct.
41  *
42  * - A load test is carried out, with all cores attempting to lock a single lock
43  *   multiple times
44  */
45
46 static rte_spinlock_t sl, sl_try;
47 static rte_spinlock_t sl_tab[RTE_MAX_LCORE];
48 static rte_spinlock_recursive_t slr;
49 static unsigned count = 0;
50
51 static uint32_t synchro;
52
53 static int
54 test_spinlock_per_core(__rte_unused void *arg)
55 {
56         rte_spinlock_lock(&sl);
57         printf("Global lock taken on core %u\n", rte_lcore_id());
58         rte_spinlock_unlock(&sl);
59
60         rte_spinlock_lock(&sl_tab[rte_lcore_id()]);
61         printf("Hello from core %u !\n", rte_lcore_id());
62         rte_spinlock_unlock(&sl_tab[rte_lcore_id()]);
63
64         return 0;
65 }
66
67 static int
68 test_spinlock_recursive_per_core(__rte_unused void *arg)
69 {
70         unsigned id = rte_lcore_id();
71
72         rte_spinlock_recursive_lock(&slr);
73         printf("Global recursive lock taken on core %u - count = %d\n",
74                id, slr.count);
75         rte_spinlock_recursive_lock(&slr);
76         printf("Global recursive lock taken on core %u - count = %d\n",
77                id, slr.count);
78         rte_spinlock_recursive_lock(&slr);
79         printf("Global recursive lock taken on core %u - count = %d\n",
80                id, slr.count);
81
82         printf("Hello from within recursive locks from core %u !\n", id);
83
84         rte_spinlock_recursive_unlock(&slr);
85         printf("Global recursive lock released on core %u - count = %d\n",
86                id, slr.count);
87         rte_spinlock_recursive_unlock(&slr);
88         printf("Global recursive lock released on core %u - count = %d\n",
89                id, slr.count);
90         rte_spinlock_recursive_unlock(&slr);
91         printf("Global recursive lock released on core %u - count = %d\n",
92                id, slr.count);
93
94         return 0;
95 }
96
97 static rte_spinlock_t lk = RTE_SPINLOCK_INITIALIZER;
98 static uint64_t time_count[RTE_MAX_LCORE] = {0};
99
100 #define MAX_LOOP 10000
101
102 static int
103 load_loop_fn(void *func_param)
104 {
105         uint64_t time_diff = 0, begin;
106         uint64_t hz = rte_get_timer_hz();
107         volatile uint64_t lcount = 0;
108         const int use_lock = *(int*)func_param;
109         const unsigned lcore = rte_lcore_id();
110
111         /* wait synchro for workers */
112         if (lcore != rte_get_main_lcore())
113                 rte_wait_until_equal_32(&synchro, 1, __ATOMIC_RELAXED);
114
115         begin = rte_get_timer_cycles();
116         while (lcount < MAX_LOOP) {
117                 if (use_lock)
118                         rte_spinlock_lock(&lk);
119                 lcount++;
120                 if (use_lock)
121                         rte_spinlock_unlock(&lk);
122         }
123         time_diff = rte_get_timer_cycles() - begin;
124         time_count[lcore] = time_diff * 1000000 / hz;
125         return 0;
126 }
127
128 static int
129 test_spinlock_perf(void)
130 {
131         unsigned int i;
132         uint64_t total = 0;
133         int lock = 0;
134         const unsigned lcore = rte_lcore_id();
135
136         printf("\nTest with no lock on single core...\n");
137         load_loop_fn(&lock);
138         printf("Core [%u] Cost Time = %"PRIu64" us\n", lcore,
139                                                 time_count[lcore]);
140         memset(time_count, 0, sizeof(time_count));
141
142         printf("\nTest with lock on single core...\n");
143         lock = 1;
144         load_loop_fn(&lock);
145         printf("Core [%u] Cost Time = %"PRIu64" us\n", lcore,
146                                                 time_count[lcore]);
147         memset(time_count, 0, sizeof(time_count));
148
149         printf("\nTest with lock on %u cores...\n", rte_lcore_count());
150
151         /* Clear synchro and start workers */
152         __atomic_store_n(&synchro, 0, __ATOMIC_RELAXED);
153         rte_eal_mp_remote_launch(load_loop_fn, &lock, SKIP_MAIN);
154
155         /* start synchro and launch test on main */
156         __atomic_store_n(&synchro, 1, __ATOMIC_RELAXED);
157         load_loop_fn(&lock);
158
159         rte_eal_mp_wait_lcore();
160
161         RTE_LCORE_FOREACH(i) {
162                 printf("Core [%u] Cost Time = %"PRIu64" us\n", i,
163                                                 time_count[i]);
164                 total += time_count[i];
165         }
166
167         printf("Total Cost Time = %"PRIu64" us\n", total);
168
169         return 0;
170 }
171
172 /*
173  * Use rte_spinlock_trylock() to trylock a spinlock object,
174  * If it could not lock the object successfully, it would
175  * return immediately and the variable of "count" would be
176  * increased by one per times. the value of "count" could be
177  * checked as the result later.
178  */
179 static int
180 test_spinlock_try(__rte_unused void *arg)
181 {
182         if (rte_spinlock_trylock(&sl_try) == 0) {
183                 rte_spinlock_lock(&sl);
184                 count ++;
185                 rte_spinlock_unlock(&sl);
186         }
187
188         return 0;
189 }
190
191
192 /*
193  * Test rte_eal_get_lcore_state() in addition to spinlocks
194  * as we have "waiting" then "running" lcores.
195  */
196 static int
197 test_spinlock(void)
198 {
199         int ret = 0;
200         int i;
201
202         /* worker cores should be waiting: print it */
203         RTE_LCORE_FOREACH_WORKER(i) {
204                 printf("lcore %d state: %d\n", i,
205                        (int) rte_eal_get_lcore_state(i));
206         }
207
208         rte_spinlock_init(&sl);
209         rte_spinlock_init(&sl_try);
210         rte_spinlock_recursive_init(&slr);
211         for (i=0; i<RTE_MAX_LCORE; i++)
212                 rte_spinlock_init(&sl_tab[i]);
213
214         rte_spinlock_lock(&sl);
215
216         RTE_LCORE_FOREACH_WORKER(i) {
217                 rte_spinlock_lock(&sl_tab[i]);
218                 rte_eal_remote_launch(test_spinlock_per_core, NULL, i);
219         }
220
221         /* worker cores should be busy: print it */
222         RTE_LCORE_FOREACH_WORKER(i) {
223                 printf("lcore %d state: %d\n", i,
224                        (int) rte_eal_get_lcore_state(i));
225         }
226         rte_spinlock_unlock(&sl);
227
228         RTE_LCORE_FOREACH_WORKER(i) {
229                 rte_spinlock_unlock(&sl_tab[i]);
230                 rte_delay_ms(10);
231         }
232
233         rte_eal_mp_wait_lcore();
234
235         rte_spinlock_recursive_lock(&slr);
236
237         /*
238          * Try to acquire a lock that we already own
239          */
240         if(!rte_spinlock_recursive_trylock(&slr)) {
241                 printf("rte_spinlock_recursive_trylock failed on a lock that "
242                        "we already own\n");
243                 ret = -1;
244         } else
245                 rte_spinlock_recursive_unlock(&slr);
246
247         RTE_LCORE_FOREACH_WORKER(i) {
248                 rte_eal_remote_launch(test_spinlock_recursive_per_core, NULL, i);
249         }
250         rte_spinlock_recursive_unlock(&slr);
251         rte_eal_mp_wait_lcore();
252
253         /*
254          * Test if it could return immediately from try-locking a locked object.
255          * Here it will lock the spinlock object first, then launch all the worker
256          * lcores to trylock the same spinlock object.
257          * All the worker lcores should give up try-locking a locked object and
258          * return immediately, and then increase the "count" initialized with zero
259          * by one per times.
260          * We can check if the "count" is finally equal to the number of all worker
261          * lcores to see if the behavior of try-locking a locked spinlock object
262          * is correct.
263          */
264         if (rte_spinlock_trylock(&sl_try) == 0) {
265                 return -1;
266         }
267         count = 0;
268         RTE_LCORE_FOREACH_WORKER(i) {
269                 rte_eal_remote_launch(test_spinlock_try, NULL, i);
270         }
271         rte_eal_mp_wait_lcore();
272         rte_spinlock_unlock(&sl_try);
273         if (rte_spinlock_is_locked(&sl)) {
274                 printf("spinlock is locked but it should not be\n");
275                 return -1;
276         }
277         rte_spinlock_lock(&sl);
278         if (count != ( rte_lcore_count() - 1)) {
279                 ret = -1;
280         }
281         rte_spinlock_unlock(&sl);
282
283         /*
284          * Test if it can trylock recursively.
285          * Use rte_spinlock_recursive_trylock() to check if it can lock a spinlock
286          * object recursively. Here it will try to lock a spinlock object twice.
287          */
288         if (rte_spinlock_recursive_trylock(&slr) == 0) {
289                 printf("It failed to do the first spinlock_recursive_trylock but it should able to do\n");
290                 return -1;
291         }
292         if (rte_spinlock_recursive_trylock(&slr) == 0) {
293                 printf("It failed to do the second spinlock_recursive_trylock but it should able to do\n");
294                 return -1;
295         }
296         rte_spinlock_recursive_unlock(&slr);
297         rte_spinlock_recursive_unlock(&slr);
298
299         if (test_spinlock_perf() < 0)
300                 return -1;
301
302         return ret;
303 }
304
305 REGISTER_TEST_COMMAND(spinlock_autotest, test_spinlock);