1 /* SPDX-License-Identifier: BSD-3-Clause
2 * Copyright(C) 2020 Marvell International Ltd.
9 #include <rte_common.h>
10 #include <rte_errno.h>
11 #include <rte_lcore.h>
12 #include <rte_per_lcore.h>
13 #include <rte_string_fns.h>
15 #include "eal_trace.h"
17 RTE_DEFINE_PER_LCORE(volatile int, trace_point_sz);
18 RTE_DEFINE_PER_LCORE(void *, trace_mem);
19 static RTE_DEFINE_PER_LCORE(char *, ctf_field);
21 static struct trace_point_head tp_list = STAILQ_HEAD_INITIALIZER(tp_list);
22 static struct trace trace = { .args = STAILQ_HEAD_INITIALIZER(trace.args), };
30 struct trace_point_head *
31 trace_list_head_get(void)
39 struct trace_arg *arg;
41 /* Trace memory should start with 8B aligned for natural alignment */
42 RTE_BUILD_BUG_ON((offsetof(struct __rte_trace_header, mem) % 8) != 0);
44 /* One of the trace point registration failed */
45 if (trace.register_errno) {
46 rte_errno = trace.register_errno;
50 if (!STAILQ_EMPTY(&trace.args))
53 if (!rte_trace_is_enabled())
56 rte_spinlock_init(&trace.lock);
58 /* Is duplicate trace name registered */
59 if (trace_has_duplicate_entry())
62 /* Generate UUID ver 4 with total size of events and number of
65 trace_uuid_generate();
67 /* Apply buffer size configuration for trace output */
68 trace_bufsz_args_apply();
70 /* Generate CTF TDSL metadata */
71 if (trace_metadata_create() < 0)
74 /* Create trace directory */
78 /* Save current epoch timestamp for future use */
79 if (trace_epoch_time_save() < 0)
82 /* Apply global configurations */
83 STAILQ_FOREACH(arg, &trace.args, next)
84 trace_args_apply(arg->val);
86 rte_trace_mode_set(trace.mode);
91 trace_metadata_destroy();
93 trace_err("failed to initialize trace [%s]", rte_strerror(rte_errno));
100 if (!rte_trace_is_enabled())
103 trace_metadata_destroy();
104 eal_trace_args_free();
108 rte_trace_is_enabled(void)
114 trace_mode_set(rte_trace_point_t *trace, enum rte_trace_mode mode)
116 if (mode == RTE_TRACE_MODE_OVERWRITE)
117 __atomic_and_fetch(trace, ~__RTE_TRACE_FIELD_ENABLE_DISCARD,
120 __atomic_or_fetch(trace, __RTE_TRACE_FIELD_ENABLE_DISCARD,
125 rte_trace_mode_set(enum rte_trace_mode mode)
127 struct trace_point *tp;
129 if (!rte_trace_is_enabled())
132 STAILQ_FOREACH(tp, &tp_list, next)
133 trace_mode_set(tp->handle, mode);
139 rte_trace_mode rte_trace_mode_get(void)
145 trace_point_is_invalid(rte_trace_point_t *t)
147 return (t == NULL) || (trace_id_get(t) >= trace.nb_trace_points);
151 rte_trace_point_is_enabled(rte_trace_point_t *trace)
155 if (trace_point_is_invalid(trace))
158 val = __atomic_load_n(trace, __ATOMIC_ACQUIRE);
159 return (val & __RTE_TRACE_FIELD_ENABLE_MASK) != 0;
163 rte_trace_point_enable(rte_trace_point_t *trace)
165 if (trace_point_is_invalid(trace))
168 __atomic_or_fetch(trace, __RTE_TRACE_FIELD_ENABLE_MASK,
174 rte_trace_point_disable(rte_trace_point_t *trace)
176 if (trace_point_is_invalid(trace))
179 __atomic_and_fetch(trace, ~__RTE_TRACE_FIELD_ENABLE_MASK,
185 rte_trace_pattern(const char *pattern, bool enable)
187 struct trace_point *tp;
188 int rc = 0, found = 0;
190 STAILQ_FOREACH(tp, &tp_list, next) {
191 if (fnmatch(pattern, tp->name, 0) == 0) {
193 rc = rte_trace_point_enable(tp->handle);
195 rc = rte_trace_point_disable(tp->handle);
206 rte_trace_regexp(const char *regex, bool enable)
208 struct trace_point *tp;
209 int rc = 0, found = 0;
212 if (regcomp(&r, regex, 0) != 0)
215 STAILQ_FOREACH(tp, &tp_list, next) {
216 if (regexec(&r, tp->name, 0, NULL, 0) == 0) {
218 rc = rte_trace_point_enable(tp->handle);
220 rc = rte_trace_point_disable(tp->handle);
232 rte_trace_point_lookup(const char *name)
234 struct trace_point *tp;
239 STAILQ_FOREACH(tp, &tp_list, next)
240 if (strncmp(tp->name, name, TRACE_POINT_NAME_SIZE) == 0)
247 trace_point_dump(FILE *f, struct trace_point *tp)
249 rte_trace_point_t *handle = tp->handle;
251 fprintf(f, "\tid %d, %s, size is %d, %s\n",
252 trace_id_get(handle), tp->name,
253 (uint16_t)(*handle & __RTE_TRACE_FIELD_SIZE_MASK),
254 rte_trace_point_is_enabled(handle) ? "enabled" : "disabled");
258 trace_lcore_mem_dump(FILE *f)
260 struct trace *trace = trace_obj_get();
261 struct __rte_trace_header *header;
264 if (trace->nb_trace_mem_list == 0)
267 rte_spinlock_lock(&trace->lock);
268 fprintf(f, "nb_trace_mem_list = %d\n", trace->nb_trace_mem_list);
269 fprintf(f, "\nTrace mem info\n--------------\n");
270 for (count = 0; count < trace->nb_trace_mem_list; count++) {
271 header = trace->lcore_meta[count].mem;
272 fprintf(f, "\tid %d, mem=%p, area=%s, lcore_id=%d, name=%s\n",
274 trace_area_to_string(trace->lcore_meta[count].area),
275 header->stream_header.lcore_id,
276 header->stream_header.thread_name);
278 rte_spinlock_unlock(&trace->lock);
282 rte_trace_dump(FILE *f)
284 struct trace_point_head *tp_list = trace_list_head_get();
285 struct trace *trace = trace_obj_get();
286 struct trace_point *tp;
288 fprintf(f, "\nGlobal info\n-----------\n");
289 fprintf(f, "status = %s\n",
290 rte_trace_is_enabled() ? "enabled" : "disabled");
291 fprintf(f, "mode = %s\n",
292 trace_mode_to_string(rte_trace_mode_get()));
293 fprintf(f, "dir = %s\n", trace->dir);
294 fprintf(f, "buffer len = %d\n", trace->buff_len);
295 fprintf(f, "number of trace points = %d\n", trace->nb_trace_points);
297 trace_lcore_mem_dump(f);
298 fprintf(f, "\nTrace point info\n----------------\n");
299 STAILQ_FOREACH(tp, tp_list, next)
300 trace_point_dump(f, tp);
304 __rte_trace_mem_per_thread_alloc(void)
306 struct trace *trace = trace_obj_get();
307 struct __rte_trace_header *header;
310 if (!rte_trace_is_enabled())
313 if (RTE_PER_LCORE(trace_mem))
316 rte_spinlock_lock(&trace->lock);
318 count = trace->nb_trace_mem_list;
320 /* Allocate room for storing the thread trace mem meta */
321 trace->lcore_meta = realloc(trace->lcore_meta,
322 sizeof(trace->lcore_meta[0]) * (count + 1));
324 /* Provide dummy space for fast path to consume */
325 if (trace->lcore_meta == NULL) {
326 trace_crit("trace mem meta memory realloc failed");
331 /* First attempt from huge page */
332 header = eal_malloc_no_trace(NULL, trace_mem_sz(trace->buff_len), 8);
334 trace->lcore_meta[count].area = TRACE_AREA_HUGEPAGE;
338 /* Second attempt from heap */
339 header = malloc(trace_mem_sz(trace->buff_len));
340 if (header == NULL) {
341 trace_crit("trace mem malloc attempt failed");
347 /* Second attempt from heap is success */
348 trace->lcore_meta[count].area = TRACE_AREA_HEAP;
350 /* Initialize the trace header */
353 header->len = trace->buff_len;
354 header->stream_header.magic = TRACE_CTF_MAGIC;
355 rte_uuid_copy(header->stream_header.uuid, trace->uuid);
356 header->stream_header.lcore_id = rte_lcore_id();
358 /* Store the thread name */
359 char *name = header->stream_header.thread_name;
360 memset(name, 0, __RTE_TRACE_EMIT_STRING_LEN_MAX);
361 rte_thread_getname(pthread_self(), name,
362 __RTE_TRACE_EMIT_STRING_LEN_MAX);
364 trace->lcore_meta[count].mem = header;
365 trace->nb_trace_mem_list++;
367 RTE_PER_LCORE(trace_mem) = header;
368 rte_spinlock_unlock(&trace->lock);
372 trace_mem_per_thread_free_unlocked(struct thread_mem_meta *meta)
374 if (meta->area == TRACE_AREA_HUGEPAGE)
375 eal_free_no_trace(meta->mem);
376 else if (meta->area == TRACE_AREA_HEAP)
381 trace_mem_per_thread_free(void)
383 struct trace *trace = trace_obj_get();
384 struct __rte_trace_header *header;
387 header = RTE_PER_LCORE(trace_mem);
391 rte_spinlock_lock(&trace->lock);
392 for (count = 0; count < trace->nb_trace_mem_list; count++) {
393 if (trace->lcore_meta[count].mem == header)
396 if (count != trace->nb_trace_mem_list) {
397 struct thread_mem_meta *meta = &trace->lcore_meta[count];
399 trace_mem_per_thread_free_unlocked(meta);
400 if (count != trace->nb_trace_mem_list - 1) {
401 memmove(meta, meta + 1,
403 (trace->nb_trace_mem_list - count - 1));
405 trace->nb_trace_mem_list--;
407 rte_spinlock_unlock(&trace->lock);
413 struct trace *trace = trace_obj_get();
416 if (!rte_trace_is_enabled())
419 rte_spinlock_lock(&trace->lock);
420 for (count = 0; count < trace->nb_trace_mem_list; count++) {
421 trace_mem_per_thread_free_unlocked(&trace->lcore_meta[count]);
423 trace->nb_trace_mem_list = 0;
424 rte_spinlock_unlock(&trace->lock);
428 __rte_trace_point_emit_field(size_t sz, const char *in, const char *datatype)
434 fixup = trace_metadata_fixup_field(in);
437 rc = asprintf(&field, "%s %s %s;\n",
438 RTE_PER_LCORE(ctf_field) != NULL ?
439 RTE_PER_LCORE(ctf_field) : "",
441 free(RTE_PER_LCORE(ctf_field));
444 RTE_PER_LCORE(trace_point_sz) = 0;
445 RTE_PER_LCORE(ctf_field) = NULL;
446 trace_crit("could not allocate CTF field");
449 RTE_PER_LCORE(trace_point_sz) += sz;
450 RTE_PER_LCORE(ctf_field) = field;
454 __rte_trace_point_register(rte_trace_point_t *handle, const char *name,
455 void (*register_fn)(void))
457 struct trace_point *tp;
460 /* Sanity checks of arguments */
461 if (name == NULL || register_fn == NULL || handle == NULL) {
462 trace_err("invalid arguments");
467 /* Check the size of the trace point object */
468 RTE_PER_LCORE(trace_point_sz) = 0;
470 if (RTE_PER_LCORE(trace_point_sz) == 0) {
471 trace_err("missing rte_trace_emit_header() in register fn");
476 /* Is size overflowed */
477 if (RTE_PER_LCORE(trace_point_sz) > UINT16_MAX) {
478 trace_err("trace point size overflowed");
483 /* Are we running out of space to store trace points? */
484 if (trace.nb_trace_points > UINT16_MAX) {
485 trace_err("trace point exceeds the max count");
490 /* Get the size of the trace point */
491 sz = RTE_PER_LCORE(trace_point_sz);
492 tp = calloc(1, sizeof(struct trace_point));
494 trace_err("fail to allocate trace point memory");
499 /* Initialize the trace point */
500 if (rte_strscpy(tp->name, name, TRACE_POINT_NAME_SIZE) < 0) {
501 trace_err("name is too long");
505 /* Copy the accumulated fields description and clear it for the next
508 tp->ctf_field = RTE_PER_LCORE(ctf_field);
509 RTE_PER_LCORE(ctf_field) = NULL;
511 /* Form the trace handle */
513 *handle |= trace.nb_trace_points << __RTE_TRACE_FIELD_ID_SHIFT;
515 trace.nb_trace_points++;
518 /* Add the trace point at tail */
519 STAILQ_INSERT_TAIL(&tp_list, tp, next);
520 __atomic_thread_fence(__ATOMIC_RELEASE);
527 if (trace.register_errno == 0)
528 trace.register_errno = rte_errno;