RTE_DEFINE_PER_LCORE(volatile int, trace_point_sz);
RTE_DEFINE_PER_LCORE(void *, trace_mem);
-static RTE_DEFINE_PER_LCORE(char, ctf_field[TRACE_CTF_FIELD_SIZE]);
-static RTE_DEFINE_PER_LCORE(int, ctf_count);
+static RTE_DEFINE_PER_LCORE(char *, ctf_field);
static struct trace_point_head tp_list = STAILQ_HEAD_INITIALIZER(tp_list);
static struct trace trace = { .args = STAILQ_HEAD_INITIALIZER(trace.args), };
{
if (!rte_trace_is_enabled())
return;
- trace_mem_per_thread_free();
+ trace_mem_free();
trace_metadata_destroy();
eal_trace_args_free();
}
rte_spinlock_unlock(&trace->lock);
}
+static void
+trace_mem_per_thread_free_unlocked(struct thread_mem_meta *meta)
+{
+ if (meta->area == TRACE_AREA_HUGEPAGE)
+ eal_free_no_trace(meta->mem);
+ else if (meta->area == TRACE_AREA_HEAP)
+ free(meta->mem);
+}
+
void
trace_mem_per_thread_free(void)
+{
+ struct trace *trace = trace_obj_get();
+ struct __rte_trace_header *header;
+ uint32_t count;
+
+ header = RTE_PER_LCORE(trace_mem);
+ if (header == NULL)
+ return;
+
+ rte_spinlock_lock(&trace->lock);
+ for (count = 0; count < trace->nb_trace_mem_list; count++) {
+ if (trace->lcore_meta[count].mem == header)
+ break;
+ }
+ if (count != trace->nb_trace_mem_list) {
+ struct thread_mem_meta *meta = &trace->lcore_meta[count];
+
+ trace_mem_per_thread_free_unlocked(meta);
+ if (count != trace->nb_trace_mem_list - 1) {
+ memmove(meta, meta + 1,
+ sizeof(*meta) *
+ (trace->nb_trace_mem_list - count - 1));
+ }
+ trace->nb_trace_mem_list--;
+ }
+ rte_spinlock_unlock(&trace->lock);
+}
+
+void
+trace_mem_free(void)
{
struct trace *trace = trace_obj_get();
uint32_t count;
- void *mem;
if (!rte_trace_is_enabled())
return;
rte_spinlock_lock(&trace->lock);
for (count = 0; count < trace->nb_trace_mem_list; count++) {
- mem = trace->lcore_meta[count].mem;
- if (trace->lcore_meta[count].area == TRACE_AREA_HUGEPAGE)
- eal_free_no_trace(mem);
- else if (trace->lcore_meta[count].area == TRACE_AREA_HEAP)
- free(mem);
+ trace_mem_per_thread_free_unlocked(&trace->lcore_meta[count]);
}
+ trace->nb_trace_mem_list = 0;
rte_spinlock_unlock(&trace->lock);
}
void
__rte_trace_point_emit_field(size_t sz, const char *in, const char *datatype)
{
- char *field = RTE_PER_LCORE(ctf_field);
- int count = RTE_PER_LCORE(ctf_count);
- size_t size;
+ char *field;
+ char *fixup;
int rc;
- size = RTE_MAX(0, TRACE_CTF_FIELD_SIZE - 1 - count);
- RTE_PER_LCORE(trace_point_sz) += sz;
- rc = snprintf(RTE_PTR_ADD(field, count), size, "%s %s;", datatype, in);
- if (rc <= 0 || (size_t)rc >= size) {
+ fixup = trace_metadata_fixup_field(in);
+ if (fixup != NULL)
+ in = fixup;
+ rc = asprintf(&field, "%s %s %s;\n",
+ RTE_PER_LCORE(ctf_field) != NULL ?
+ RTE_PER_LCORE(ctf_field) : "",
+ datatype, in);
+ free(RTE_PER_LCORE(ctf_field));
+ free(fixup);
+ if (rc == -1) {
RTE_PER_LCORE(trace_point_sz) = 0;
- trace_crit("CTF field is too long");
+ RTE_PER_LCORE(ctf_field) = NULL;
+ trace_crit("could not allocate CTF field");
return;
}
- RTE_PER_LCORE(ctf_count) += rc;
+ RTE_PER_LCORE(trace_point_sz) += sz;
+ RTE_PER_LCORE(ctf_field) = field;
}
int
__rte_trace_point_register(rte_trace_point_t *handle, const char *name,
void (*register_fn)(void))
{
- char *field = RTE_PER_LCORE(ctf_field);
struct trace_point *tp;
uint16_t sz;
/* Check the size of the trace point object */
RTE_PER_LCORE(trace_point_sz) = 0;
- RTE_PER_LCORE(ctf_count) = 0;
register_fn();
if (RTE_PER_LCORE(trace_point_sz) == 0) {
trace_err("missing rte_trace_emit_header() in register fn");
goto free;
}
- /* Copy the field data for future use */
- if (rte_strscpy(tp->ctf_field, field, TRACE_CTF_FIELD_SIZE) < 0) {
- trace_err("CTF field size is too long");
- rte_errno = E2BIG;
- goto free;
- }
-
- /* Clear field memory for the next event */
- memset(field, 0, TRACE_CTF_FIELD_SIZE);
+ /* Copy the accumulated fields description and clear it for the next
+ * trace point.
+ */
+ tp->ctf_field = RTE_PER_LCORE(ctf_field);
+ RTE_PER_LCORE(ctf_field) = NULL;
/* Form the trace handle */
*handle = sz;