// SPDX-License-Identifier: GPL-2.0-only /* * CTF writing support via babeltrace. * * Copyright (C) 2014, Jiri Olsa <jolsa@redhat.com> * Copyright (C) 2014, Sebastian Andrzej Siewior <bigeasy@linutronix.de> */ #include <errno.h> #include <inttypes.h> #include <linux/compiler.h> #include <linux/kernel.h> #include <linux/zalloc.h> #include <babeltrace/ctf-writer/writer.h> #include <babeltrace/ctf-writer/clock.h> #include <babeltrace/ctf-writer/stream.h> #include <babeltrace/ctf-writer/event.h> #include <babeltrace/ctf-writer/event-types.h> #include <babeltrace/ctf-writer/event-fields.h> #include <babeltrace/ctf-ir/utils.h> #include <babeltrace/ctf/events.h> #include "asm/bug.h" #include "data-convert.h" #include "session.h" #include "debug.h" #include "tool.h" #include "evlist.h" #include "evsel.h" #include "machine.h" #include "config.h" #include <linux/ctype.h> #include <linux/err.h> #include <linux/time64.h> #include "util.h" #include "clockid.h" #include "util/sample.h" #ifdef HAVE_LIBTRACEEVENT #include <traceevent/event-parse.h> #endif #define pr_N(n, fmt, ...) \ eprintf(n, debug_data_convert, fmt, ##__VA_ARGS__) #define pr(fmt, ...) pr_N(1, pr_fmt(fmt), ##__VA_ARGS__) #define pr2(fmt, ...) pr_N(2, pr_fmt(fmt), ##__VA_ARGS__) #define pr_time2(t, fmt, ...) pr_time_N(2, debug_data_convert, t, pr_fmt(fmt), ##__VA_ARGS__) struct evsel_priv { struct bt_ctf_event_class *event_class; }; #define MAX_CPUS 4096 struct ctf_stream { struct bt_ctf_stream *stream; int cpu; u32 count; }; struct ctf_writer { /* writer primitives */ struct bt_ctf_writer *writer; struct ctf_stream **stream; int stream_cnt; struct bt_ctf_stream_class *stream_class; struct bt_ctf_clock *clock; /* data types */ union { struct { struct bt_ctf_field_type *s64; struct bt_ctf_field_type *u64; struct bt_ctf_field_type *s32; struct bt_ctf_field_type *u32; struct bt_ctf_field_type *string; struct bt_ctf_field_type *u32_hex; struct bt_ctf_field_type *u64_hex; }; struct bt_ctf_field_type *array[6]; } data; struct bt_ctf_event_class *comm_class; struct bt_ctf_event_class *exit_class; struct bt_ctf_event_class *fork_class; struct bt_ctf_event_class *mmap_class; struct bt_ctf_event_class *mmap2_class; }; struct convert { struct perf_tool tool; struct ctf_writer writer; u64 events_size; u64 events_count; u64 non_sample_count; /* Ordered events configured queue size. */ u64 queue_size; }; static int value_set(struct bt_ctf_field_type *type, struct bt_ctf_event *event, const char *name, u64 val) { struct bt_ctf_field *field; bool sign = bt_ctf_field_type_integer_get_signed(type); int ret; field = bt_ctf_field_create(type); if (!field) { pr_err("failed to create a field %s\n", name); return -1; } if (sign) { ret = bt_ctf_field_signed_integer_set_value(field, val); if (ret) { pr_err("failed to set field value %s\n", name); goto err; } } else { ret = bt_ctf_field_unsigned_integer_set_value(field, val); if (ret) { pr_err("failed to set field value %s\n", name); goto err; } } ret = bt_ctf_event_set_payload(event, name, field); if (ret) { pr_err("failed to set payload %s\n", name); goto err; } pr2(" SET [%s = %" PRIu64 "]\n", name, val); err: bt_ctf_field_put(field); return ret; } #define __FUNC_VALUE_SET(_name, _val_type) \ static __maybe_unused int value_set_##_name(struct ctf_writer *cw, \ struct bt_ctf_event *event, \ const char *name, \ _val_type val) \ { \ struct bt_ctf_field_type *type = cw->data._name; \ return value_set(type, event, name, (u64) val); \ } #define FUNC_VALUE_SET(_name) __FUNC_VALUE_SET(_name, _name) FUNC_VALUE_SET(s32) FUNC_VALUE_SET(u32) FUNC_VALUE_SET(s64) FUNC_VALUE_SET(u64) __FUNC_VALUE_SET(u64_hex, u64) static int string_set_value(struct bt_ctf_field *field, const char *string); static __maybe_unused int value_set_string(struct ctf_writer *cw, struct bt_ctf_event *event, const char *name, const char *string) { struct bt_ctf_field_type *type = cw->data.string; struct bt_ctf_field *field; int ret = 0; field = bt_ctf_field_create(type); if (!field) { pr_err("failed to create a field %s\n", name); return -1; } ret = string_set_value(field, string); if (ret) { pr_err("failed to set value %s\n", name); goto err_put_field; } ret = bt_ctf_event_set_payload(event, name, field); if (ret) pr_err("failed to set payload %s\n", name); err_put_field: bt_ctf_field_put(field); return ret; } static struct bt_ctf_field_type* get_tracepoint_field_type(struct ctf_writer *cw, struct tep_format_field *field) { unsigned long flags = field->flags; if (flags & TEP_FIELD_IS_STRING) return cw->data.string; if (!(flags & TEP_FIELD_IS_SIGNED)) { /* unsigned long are mostly pointers */ if (flags & TEP_FIELD_IS_LONG || flags & TEP_FIELD_IS_POINTER) return cw->data.u64_hex; } if (flags & TEP_FIELD_IS_SIGNED) { if (field->size == 8) return cw->data.s64; else return cw->data.s32; } if (field->size == 8) return cw->data.u64; else return cw->data.u32; } static unsigned long long adjust_signedness(unsigned long long value_int, int size) { unsigned long long value_mask; /* * value_mask = (1 << (size * 8 - 1)) - 1. * Directly set value_mask for code readers. */ switch (size) { case 1: value_mask = 0x7fULL; break; case 2: value_mask = 0x7fffULL; break; case 4: value_mask = 0x7fffffffULL; break; case 8: /* * For 64 bit value, return it self. There is no need * to fill high bit. */ /* Fall through */ default: /* BUG! */ return value_int; } /* If it is a positive value, don't adjust. */ if ((value_int & (~0ULL - value_mask)) == 0) return value_int; /* Fill upper part of value_int with 1 to make it a negative long long. */ return (value_int & value_mask) | ~value_mask; } static int string_set_value(struct bt_ctf_field *field, const char *string) { char *buffer = NULL; size_t len = strlen(string), i, p; int err; for (i = p = 0; i < len; i++, p++) { if (isprint(string[i])) { if (!buffer) continue; buffer[p] = string[i]; } else { char numstr[5]; snprintf(numstr, sizeof(numstr), "\\x%02x", (unsigned int)(string[i]) & 0xff); if (!buffer) { buffer = zalloc(i + (len - i) * 4 + 2); if (!buffer) { pr_err("failed to set unprintable string '%s'\n", string); return bt_ctf_field_string_set_value(field, "UNPRINTABLE-STRING"); } if (i > 0) strncpy(buffer, string, i); } memcpy(buffer + p, numstr, 4); p += 3; } } if (!buffer) return bt_ctf_field_string_set_value(field, string); err = bt_ctf_field_string_set_value(field, buffer); free(buffer); return err; } static int add_tracepoint_field_value(struct ctf_writer *cw, struct bt_ctf_event_class *event_class, struct bt_ctf_event *event, struct perf_sample *sample, struct tep_format_field *fmtf) { struct bt_ctf_field_type *type; struct bt_ctf_field *array_field; struct bt_ctf_field *field; const char *name = fmtf->name; void *data = sample->raw_data; unsigned long flags = fmtf->flags; unsigned int n_items; unsigned int i; unsigned int offset; unsigned int len; int ret; name = fmtf->alias; offset = fmtf->offset; len = fmtf->size; if (flags & TEP_FIELD_IS_STRING) flags &= ~TEP_FIELD_IS_ARRAY; if (flags & TEP_FIELD_IS_DYNAMIC) { unsigned long long tmp_val; tmp_val = tep_read_number(fmtf->event->tep, data + offset, len); offset = tmp_val; len = offset >> 16; offset &= 0xffff; if (tep_field_is_relative(flags)) offset += fmtf->offset + fmtf->size; } if (flags & TEP_FIELD_IS_ARRAY) { type = bt_ctf_event_class_get_field_by_name( event_class, name); array_field = bt_ctf_field_create(type); bt_ctf_field_type_put(type); if (!array_field) { pr_err("Failed to create array type %s\n", name); return -1; } len = fmtf->size / fmtf->arraylen; n_items = fmtf->arraylen; } else { n_items = 1; array_field = NULL; } type = get_tracepoint_field_type(cw, fmtf); for (i = 0; i < n_items; i++) { if (flags & TEP_FIELD_IS_ARRAY) field = bt_ctf_field_array_get_field(array_field, i); else field = bt_ctf_field_create(type); if (!field) { pr_err("failed to create a field %s\n", name); return -1; } if (flags & TEP_FIELD_IS_STRING) ret = string_set_value(field, data + offset + i * len); else { unsigned long long value_int; value_int = tep_read_number( fmtf->event->tep, data + offset + i * len, len); if (!(flags & TEP_FIELD_IS_SIGNED)) ret = bt_ctf_field_unsigned_integer_set_value( field, value_int); else ret = bt_ctf_field_signed_integer_set_value( field, adjust_signedness(value_int, len)); } if (ret) { pr_err("failed to set file value %s\n", name); goto err_put_field; } if (!(flags & TEP_FIELD_IS_ARRAY)) { ret = bt_ctf_event_set_payload(event, name, field); if (ret) { pr_err("failed to set payload %s\n", name); goto err_put_field; } } bt_ctf_field_put(field); } if (flags & TEP_FIELD_IS_ARRAY) { ret = bt_ctf_event_set_payload(event, name, array_field); if (ret) { pr_err("Failed add payload array %s\n", name); return -1; } bt_ctf_field_put(array_field); } return 0; err_put_field: bt_ctf_field_put(field); return -1; } static int add_tracepoint_fields_values(struct ctf_writer *cw, struct bt_ctf_event_class *event_class, struct bt_ctf_event *event, struct tep_format_field *fields, struct perf_sample *sample) { struct tep_format_field *field; int ret; for (field = fields; field; field = field->next) { ret = add_tracepoint_field_value(cw, event_class, event, sample, field); if (ret) return -1; } return 0; } static int add_tracepoint_values(struct ctf_writer *cw, struct bt_ctf_event_class *event_class, struct bt_ctf_event *event, struct evsel *evsel, struct perf_sample *sample) { struct tep_format_field *common_fields = evsel->tp_format->format.common_fields; struct tep_format_field *fields = evsel->tp_format->format.fields; int ret; ret = add_tracepoint_fields_values(cw, event_class, event, common_fields, sample); if (!ret) ret = add_tracepoint_fields_values(cw, event_class, event, fields, sample); return ret; } static int add_bpf_output_values(struct bt_ctf_event_class *event_class, struct bt_ctf_event *event, struct perf_sample *sample) { struct bt_ctf_field_type *len_type, *seq_type; struct bt_ctf_field *len_field, *seq_field; unsigned int raw_size = sample->raw_size; unsigned int nr_elements = raw_size / sizeof(u32); unsigned int i; int ret; if (nr_elements * sizeof(u32) != raw_size) pr_warning("Incorrect raw_size (%u) in bpf output event, skip %zu bytes\n", raw_size, nr_elements * sizeof(u32) - raw_size); len_type = bt_ctf_event_class_get_field_by_name(event_class, "raw_len"); len_field = bt_ctf_field_create(len_type); if (!len_field) { pr_err("failed to create 'raw_len' for bpf output event\n"); ret = -1; goto put_len_type; } ret = bt_ctf_field_unsigned_integer_set_value(len_field, nr_elements); if (ret) { pr_err("failed to set field value for raw_len\n"); goto put_len_field; } ret = bt_ctf_event_set_payload(event, "raw_len", len_field); if (ret) { pr_err("failed to set payload to raw_len\n"); goto put_len_field; } seq_type = bt_ctf_event_class_get_field_by_name(event_class, "raw_data"); seq_field = bt_ctf_field_create(seq_type); if (!seq_field) { pr_err("failed to create 'raw_data' for bpf output event\n"); ret = -1; goto put_seq_type; } ret = bt_ctf_field_sequence_set_length(seq_field, len_field); if (ret) { pr_err("failed to set length of 'raw_data'\n"); goto put_seq_field; } for (i = 0; i < nr_elements; i++) { struct bt_ctf_field *elem_field = bt_ctf_field_sequence_get_field(seq_field, i); ret = bt_ctf_field_unsigned_integer_set_value(elem_field, ((u32 *)(sample->raw_data))[i]); bt_ctf_field_put(elem_field); if (ret) { pr_err("failed to set raw_data[%d]\n", i); goto put_seq_field; } } ret = bt_ctf_event_set_payload(event, "raw_data", seq_field); if (ret) pr_err("failed to set payload for raw_data\n"); put_seq_field: bt_ctf_field_put(seq_field); put_seq_type: bt_ctf_field_type_put(seq_type); put_len_field: bt_ctf_field_put(len_field); put_len_type: bt_ctf_field_type_put(len_type); return ret; } static int add_callchain_output_values(struct bt_ctf_event_class *event_class, struct bt_ctf_event *event, struct ip_callchain *callchain) { struct bt_ctf_field_type *len_type, *seq_type; struct bt_ctf_field *len_field, *seq_field; unsigned int nr_elements = callchain->nr; unsigned int i; int ret; len_type = bt_ctf_event_class_get_field_by_name( event_class, "perf_callchain_size"); len_field = bt_ctf_field_create(len_type); if (!len_field) { pr_err("failed to create 'perf_callchain_size' for callchain output event\n"); ret = -1; goto put_len_type; } ret = bt_ctf_field_unsigned_integer_set_value(len_field, nr_elements); if (ret) { pr_err("failed to set field value for perf_callchain_size\n"); goto put_len_field; } ret = bt_ctf_event_set_payload(event, "perf_callchain_size", len_field); if (ret) { pr_err("failed to set payload to perf_callchain_size\n"); goto put_len_field; } seq_type = bt_ctf_event_class_get_field_by_name( event_class, "perf_callchain"); seq_field = bt_ctf_field_create(seq_type); if (!seq_field) { pr_err("failed to create 'perf_callchain' for callchain output event\n"); ret = -1; goto put_seq_type; } ret = bt_ctf_field_sequence_set_length(seq_field, len_field); if (ret) { pr_err("failed to set length of 'perf_callchain'\n"); goto put_seq_field; } for (i = 0; i < nr_elements; i++) { struct bt_ctf_field *elem_field = bt_ctf_field_sequence_get_field(seq_field, i); ret = bt_ctf_field_unsigned_integer_set_value(elem_field, ((u64 *)(callchain->ips))[i]); bt_ctf_field_put(elem_field); if (ret) { pr_err("failed to set callchain[%d]\n", i); goto put_seq_field; } } ret = bt_ctf_event_set_payload(event, "perf_callchain", seq_field); if (ret) pr_err("failed to set payload for raw_data\n"); put_seq_field: bt_ctf_field_put(seq_field); put_seq_type: bt_ctf_field_type_put(seq_type); put_len_field: bt_ctf_field_put(len_field); put_len_type: bt_ctf_field_type_put(len_type); return ret; } static int add_generic_values(struct ctf_writer *cw, struct bt_ctf_event *event, struct evsel *evsel, struct perf_sample *sample) { u64 type = evsel->core.attr.sample_type; int ret; /* * missing: * PERF_SAMPLE_TIME - not needed as we have it in * ctf event header * PERF_SAMPLE_READ - TODO * PERF_SAMPLE_RAW - tracepoint fields are handled separately * PERF_SAMPLE_BRANCH_STACK - TODO * PERF_SAMPLE_REGS_USER - TODO * PERF_SAMPLE_STACK_USER - TODO */ if (type & PERF_SAMPLE_IP) { ret = value_set_u64_hex(cw, event, "perf_ip", sample->ip); if (ret) return -1; } if (type & PERF_SAMPLE_TID) { ret = value_set_s32(cw, event, "perf_tid", sample->tid); if (ret) return -1; ret = value_set_s32(cw, event, "perf_pid", sample->pid); if (ret) return -1; } if ((type & PERF_SAMPLE_ID) || (type & PERF_SAMPLE_IDENTIFIER)) { ret = value_set_u64(cw, event, "perf_id", sample->id); if (ret) return -1; } if (type & PERF_SAMPLE_STREAM_ID) { ret = value_set_u64(cw, event, "perf_stream_id", sample->stream_id); if (ret) return -1; } if (type & PERF_SAMPLE_PERIOD) { ret = value_set_u64(cw, event, "perf_period", sample->period); if (ret) return -1; } if (type & PERF_SAMPLE_WEIGHT) { ret = value_set_u64(cw, event, "perf_weight", sample->weight); if (ret) return -1; } if (type & PERF_SAMPLE_DATA_SRC) { ret = value_set_u64(cw, event, "perf_data_src", sample->data_src); if (ret) return -1; } if (type & PERF_SAMPLE_TRANSACTION) { ret = value_set_u64(cw, event, "perf_transaction", sample->transaction); if (ret) return -1; } return 0; } static int ctf_stream__flush(struct ctf_stream *cs) { int err = 0; if (cs) { err = bt_ctf_stream_flush(cs->stream); if (err) pr_err("CTF stream %d flush failed\n", cs->cpu); pr("Flush stream for cpu %d (%u samples)\n", cs->cpu, cs->count); cs->count = 0; } return err; } static struct ctf_stream *ctf_stream__create(struct ctf_writer *cw, int cpu) { struct ctf_stream *cs; struct bt_ctf_field *pkt_ctx = NULL; struct bt_ctf_field *cpu_field = NULL; struct bt_ctf_stream *stream = NULL; int ret; cs = zalloc(sizeof(*cs)); if (!cs) { pr_err("Failed to allocate ctf stream\n"); return NULL; } stream = bt_ctf_writer_create_stream(cw->writer, cw->stream_class); if (!stream) { pr_err("Failed to create CTF stream\n"); goto out; } pkt_ctx = bt_ctf_stream_get_packet_context(stream); if (!pkt_ctx) { pr_err("Failed to obtain packet context\n"); goto out; } cpu_field = bt_ctf_field_structure_get_field(pkt_ctx, "cpu_id"); bt_ctf_field_put(pkt_ctx); if (!cpu_field) { pr_err("Failed to obtain cpu field\n"); goto out; } ret = bt_ctf_field_unsigned_integer_set_value(cpu_field, (u32) cpu); if (ret) { pr_err("Failed to update CPU number\n"); goto out; } bt_ctf_field_put(cpu_field); cs->cpu = cpu; cs->stream = stream; return cs; out: if (cpu_field) bt_ctf_field_put(cpu_field); if (stream) bt_ctf_stream_put(stream); free(cs); return NULL; } static void ctf_stream__delete(struct ctf_stream *cs) { if (cs) { bt_ctf_stream_put(cs->stream); free(cs); } } static struct ctf_stream *ctf_stream(struct ctf_writer *cw, int cpu) { struct ctf_stream *cs = cw->stream[cpu]; if (!cs) { cs = ctf_stream__create(cw, cpu); cw->stream[cpu] = cs; } return cs; } static int get_sample_cpu(struct ctf_writer *cw, struct perf_sample *sample, struct evsel *evsel) { int cpu = 0; if (evsel->core.attr.sample_type & PERF_SAMPLE_CPU) cpu = sample->cpu; if (cpu > cw->stream_cnt) { pr_err("Event was recorded for CPU %d, limit is at %d.\n", cpu, cw->stream_cnt); cpu = 0; } return cpu; } #define STREAM_FLUSH_COUNT 100000 /* * Currently we have no other way to determine the * time for the stream flush other than keep track * of the number of events and check it against * threshold. */ static bool is_flush_needed(struct ctf_stream *cs) { return cs->count >= STREAM_FLUSH_COUNT; } static int process_sample_event(struct perf_tool *tool, union perf_event *_event, struct perf_sample *sample, struct evsel *evsel, struct machine *machine __maybe_unused) { struct convert *c = container_of(tool, struct convert, tool); struct evsel_priv *priv = evsel->priv; struct ctf_writer *cw = &c->writer; struct ctf_stream *cs; struct bt_ctf_event_class *event_class; struct bt_ctf_event *event; int ret; unsigned long type = evsel->core.attr.sample_type; if (WARN_ONCE(!priv, "Failed to setup all events.\n")) return 0; event_class = priv->event_class; /* update stats */ c->events_count++; c->events_size += _event->header.size; pr_time2(sample->time, "sample %" PRIu64 "\n", c->events_count); event = bt_ctf_event_create(event_class); if (!event) { pr_err("Failed to create an CTF event\n"); return -1; } bt_ctf_clock_set_time(cw->clock, sample->time); ret = add_generic_values(cw, event, evsel, sample); if (ret) return -1; if (evsel->core.attr.type == PERF_TYPE_TRACEPOINT) { ret = add_tracepoint_values(cw, event_class, event, evsel, sample); if (ret) return -1; } if (type & PERF_SAMPLE_CALLCHAIN) { ret = add_callchain_output_values(event_class, event, sample->callchain); if (ret) return -1; } if (evsel__is_bpf_output(evsel)) { ret = add_bpf_output_values(event_class, event, sample); if (ret) return -1; } cs = ctf_stream(cw, get_sample_cpu(cw, sample, evsel)); if (cs) { if (is_flush_needed(cs)) ctf_stream__flush(cs); cs->count++; bt_ctf_stream_append_event(cs->stream, event); } bt_ctf_event_put(event); return cs ? 0 : -1; } #define __NON_SAMPLE_SET_FIELD(_name, _type, _field) \ do { \ ret = value_set_##_type(cw, event, #_field, _event->_name._field);\ if (ret) \ return -1; \ } while(0) #define __FUNC_PROCESS_NON_SAMPLE(_name, body) \ static int process_##_name##_event(struct perf_tool *tool, \ union perf_event *_event, \ struct perf_sample *sample, \ struct machine *machine) \ { \ struct convert *c = container_of(tool, struct convert, tool);\ struct ctf_writer *cw = &c->writer; \ struct bt_ctf_event_class *event_class = cw->_name##_class;\ struct bt_ctf_event *event; \ struct ctf_stream *cs; \ int ret; \ \ c->non_sample_count++; \ c->events_size += _event->header.size; \ event = bt_ctf_event_create(event_class); \ if (!event) { \ pr_err("Failed to create an CTF event\n"); \ return -1; \ } \ \ bt_ctf_clock_set_time(cw->clock, sample->time); \ body \ cs = ctf_stream(cw, 0); \ if (cs) { \ if (is_flush_needed(cs)) \ ctf_stream__flush(cs); \ \ cs->count++; \ bt_ctf_stream_append_event(cs->stream, event); \ } \ bt_ctf_event_put(event); \ \ return perf_event__process_##_name(tool, _event, sample, machine);\ } __FUNC_PROCESS_NON_SAMPLE(comm, __NON_SAMPLE_SET_FIELD(comm, u32, pid); __NON_SAMPLE_SET_FIELD(comm, u32, tid); __NON_SAMPLE_SET_FIELD(comm, string, comm); ) __FUNC_PROCESS_NON_SAMPLE(fork, __NON_SAMPLE_SET_FIELD(fork, u32, pid); __NON_SAMPLE_SET_FIELD(fork, u32, ppid); __NON_SAMPLE_SET_FIELD(fork, u32, tid); __NON_SAMPLE_SET_FIELD(fork, u32, ptid); __NON_SAMPLE_SET_FIELD(fork, u64, time); ) __FUNC_PROCESS_NON_SAMPLE(exit, __NON_SAMPLE_SET_FIELD(fork, u32, pid); __NON_SAMPLE_SET_FIELD(fork, u32, ppid); __NON_SAMPLE_SET_FIELD(fork, u32, tid); __NON_SAMPLE_SET_FIELD(fork, u32, ptid); __NON_SAMPLE_SET_FIELD(fork, u64, time); ) __FUNC_PROCESS_NON_SAMPLE(mmap, __NON_SAMPLE_SET_FIELD(mmap, u32, pid); __NON_SAMPLE_SET_FIELD(mmap, u32, tid); __NON_SAMPLE_SET_FIELD(mmap, u64_hex, start); __NON_SAMPLE_SET_FIELD(mmap, string, filename); ) __FUNC_PROCESS_NON_SAMPLE(mmap2, __NON_SAMPLE_SET_FIELD(mmap2, u32, pid); __NON_SAMPLE_SET_FIELD(mmap2, u32, tid); __NON_SAMPLE_SET_FIELD(mmap2, u64_hex, start); __NON_SAMPLE_SET_FIELD(mmap2, string, filename); ) #undef __NON_SAMPLE_SET_FIELD #undef __FUNC_PROCESS_NON_SAMPLE /* If dup < 0, add a prefix. Else, add _dupl_X suffix. */ static char *change_name(char *name, char *orig_name, int dup) { char *new_name = NULL; size_t len; if (!name) name = orig_name; if (dup >= 10) goto out; /* * Add '_' prefix to potential keywork. According to * Mathieu Desnoyers (https://lore.kernel.org/lkml/1074266107.40857.1422045946295.JavaMail.zimbra@efficios.com), * further CTF spec updating may require us to use '$'. */ if (dup < 0) len = strlen(name) + sizeof("_"); else len = strlen(orig_name) + sizeof("_dupl_X"); new_name = malloc(len); if (!new_name) goto out; if (dup < 0) snprintf(new_name, len, "_%s", name); else snprintf(new_name, len, "%s_dupl_%d", orig_name, dup); out: if (name != orig_name) free(name); return new_name; } static int event_class_add_field(struct bt_ctf_event_class *event_class, struct bt_ctf_field_type *type, struct tep_format_field *field) { struct bt_ctf_field_type *t = NULL; char *name; int dup = 1; int ret; /* alias was already assigned */ if (field->alias != field->name) return bt_ctf_event_class_add_field(event_class, type, (char *)field->alias); name = field->name; /* If 'name' is a keywork, add prefix. */ if (bt_ctf_validate_identifier(name)) name = change_name(name, field->name, -1); if (!name) { pr_err("Failed to fix invalid identifier."); return -1; } while ((t = bt_ctf_event_class_get_field_by_name(event_class, name))) { bt_ctf_field_type_put(t); name = change_name(name, field->name, dup++); if (!name) { pr_err("Failed to create dup name for '%s'\n", field->name); return -1; } } ret = bt_ctf_event_class_add_field(event_class, type, name); if (!ret) field->alias = name; return ret; } static int add_tracepoint_fields_types(struct ctf_writer *cw, struct tep_format_field *fields, struct bt_ctf_event_class *event_class) { struct tep_format_field *field; int ret; for (field = fields; field; field = field->next) { struct bt_ctf_field_type *type; unsigned long flags = field->flags; pr2(" field '%s'\n", field->name); type = get_tracepoint_field_type(cw, field); if (!type) return -1; /* * A string is an array of chars. For this we use the string * type and don't care that it is an array. What we don't * support is an array of strings. */ if (flags & TEP_FIELD_IS_STRING) flags &= ~TEP_FIELD_IS_ARRAY; if (flags & TEP_FIELD_IS_ARRAY) type = bt_ctf_field_type_array_create(type, field->arraylen); ret = event_class_add_field(event_class, type, field); if (flags & TEP_FIELD_IS_ARRAY) bt_ctf_field_type_put(type); if (ret) { pr_err("Failed to add field '%s': %d\n", field->name, ret); return -1; } } return 0; } static int add_tracepoint_types(struct ctf_writer *cw, struct evsel *evsel, struct bt_ctf_event_class *class) { struct tep_format_field *common_fields = evsel->tp_format->format.common_fields; struct tep_format_field *fields = evsel->tp_format->format.fields; int ret; ret = add_tracepoint_fields_types(cw, common_fields, class); if (!ret) ret = add_tracepoint_fields_types(cw, fields, class); return ret; } static int add_bpf_output_types(struct ctf_writer *cw, struct bt_ctf_event_class *class) { struct bt_ctf_field_type *len_type = cw->data.u32; struct bt_ctf_field_type *seq_base_type = cw->data.u32_hex; struct bt_ctf_field_type *seq_type; int ret; ret = bt_ctf_event_class_add_field(class, len_type, "raw_len"); if (ret) return ret; seq_type = bt_ctf_field_type_sequence_create(seq_base_type, "raw_len"); if (!seq_type) return -1; return bt_ctf_event_class_add_field(class, seq_type, "raw_data"); } static int add_generic_types(struct ctf_writer *cw, struct evsel *evsel, struct bt_ctf_event_class *event_class) { u64 type = evsel->core.attr.sample_type; /* * missing: * PERF_SAMPLE_TIME - not needed as we have it in * ctf event header * PERF_SAMPLE_READ - TODO * PERF_SAMPLE_CALLCHAIN - TODO * PERF_SAMPLE_RAW - tracepoint fields and BPF output * are handled separately * PERF_SAMPLE_BRANCH_STACK - TODO * PERF_SAMPLE_REGS_USER - TODO * PERF_SAMPLE_STACK_USER - TODO */ #define ADD_FIELD(cl, t, n) \ do { \ pr2(" field '%s'\n", n); \ if (bt_ctf_event_class_add_field(cl, t, n)) { \ pr_err("Failed to add field '%s';\n", n); \ return -1; \ } \ } while (0) if (type & PERF_SAMPLE_IP) ADD_FIELD(event_class, cw->data.u64_hex, "perf_ip"); if (type & PERF_SAMPLE_TID) { ADD_FIELD(event_class, cw->data.s32, "perf_tid"); ADD_FIELD(event_class, cw->data.s32, "perf_pid"); } if ((type & PERF_SAMPLE_ID) || (type & PERF_SAMPLE_IDENTIFIER)) ADD_FIELD(event_class, cw->data.u64, "perf_id"); if (type & PERF_SAMPLE_STREAM_ID) ADD_FIELD(event_class, cw->data.u64, "perf_stream_id"); if (type & PERF_SAMPLE_PERIOD) ADD_FIELD(event_class, cw->data.u64, "perf_period"); if (type & PERF_SAMPLE_WEIGHT) ADD_FIELD(event_class, cw->data.u64, "perf_weight"); if (type & PERF_SAMPLE_DATA_SRC) ADD_FIELD(event_class, cw->data.u64, "perf_data_src"); if (type & PERF_SAMPLE_TRANSACTION) ADD_FIELD(event_class, cw->data.u64, "perf_transaction"); if (type & PERF_SAMPLE_CALLCHAIN) { ADD_FIELD(event_class, cw->data.u32, "perf_callchain_size"); ADD_FIELD(event_class, bt_ctf_field_type_sequence_create( cw->data.u64_hex, "perf_callchain_size"), "perf_callchain"); } #undef ADD_FIELD return 0; } static int add_event(struct ctf_writer *cw, struct evsel *evsel) { struct bt_ctf_event_class *event_class; struct evsel_priv *priv; const char *name = evsel__name(evsel); int ret; pr("Adding event '%s' (type %d)\n", name, evsel->core.attr.type); event_class = bt_ctf_event_class_create(name); if (!event_class) return -1; ret = add_generic_types(cw, evsel, event_class); if (ret) goto err; if (evsel->core.attr.type == PERF_TYPE_TRACEPOINT) { ret = add_tracepoint_types(cw, evsel, event_class); if (ret) goto err; } if (evsel__is_bpf_output(evsel)) { ret = add_bpf_output_types(cw, event_class); if (ret) goto err; } ret = bt_ctf_stream_class_add_event_class(cw->stream_class, event_class); if (ret) { pr("Failed to add event class into stream.\n"); goto err; } priv = malloc(sizeof(*priv)); if (!priv) goto err; priv->event_class = event_class; evsel->priv = priv; return 0; err: bt_ctf_event_class_put(event_class); pr_err("Failed to add event '%s'.\n", name); return -1; } static int setup_events(struct ctf_writer *cw, struct perf_session *session) { struct evlist *evlist = session->evlist; struct evsel *evsel; int ret; evlist__for_each_entry(evlist, evsel) { ret = add_event(cw, evsel); if (ret) return ret; } return 0; } #define __NON_SAMPLE_ADD_FIELD(t, n) \ do { \ pr2(" field '%s'\n", #n); \ if (bt_ctf_event_class_add_field(event_class, cw->data.t, #n)) {\ pr_err("Failed to add field '%s';\n", #n);\ return -1; \ } \ } while(0) #define __FUNC_ADD_NON_SAMPLE_EVENT_CLASS(_name, body) \ static int add_##_name##_event(struct ctf_writer *cw) \ { \ struct bt_ctf_event_class *event_class; \ int ret; \ \ pr("Adding "#_name" event\n"); \ event_class = bt_ctf_event_class_create("perf_" #_name);\ if (!event_class) \ return -1; \ body \ \ ret = bt_ctf_stream_class_add_event_class(cw->stream_class, event_class);\ if (ret) { \ pr("Failed to add event class '"#_name"' into stream.\n");\ return ret; \ } \ \ cw->_name##_class = event_class; \ bt_ctf_event_class_put(event_class); \ return 0; \ } __FUNC_ADD_NON_SAMPLE_EVENT_CLASS(comm, __NON_SAMPLE_ADD_FIELD(u32, pid); __NON_SAMPLE_ADD_FIELD(u32, tid); __NON_SAMPLE_ADD_FIELD(string, comm); ) __FUNC_ADD_NON_SAMPLE_EVENT_CLASS(fork, __NON_SAMPLE_ADD_FIELD(u32, pid); __NON_SAMPLE_ADD_FIELD(u32, ppid); __NON_SAMPLE_ADD_FIELD(u32, tid); __NON_SAMPLE_ADD_FIELD(u32, ptid); __NON_SAMPLE_ADD_FIELD(u64, time); ) __FUNC_ADD_NON_SAMPLE_EVENT_CLASS(exit, __NON_SAMPLE_ADD_FIELD(u32, pid); __NON_SAMPLE_ADD_FIELD(u32, ppid); __NON_SAMPLE_ADD_FIELD(u32, tid); __NON_SAMPLE_ADD_FIELD(u32, ptid); __NON_SAMPLE_ADD_FIELD(u64, time); ) __FUNC_ADD_NON_SAMPLE_EVENT_CLASS(mmap, __NON_SAMPLE_ADD_FIELD(u32, pid); __NON_SAMPLE_ADD_FIELD(u32, tid); __NON_SAMPLE_ADD_FIELD(u64_hex, start); __NON_SAMPLE_ADD_FIELD(string, filename); ) __FUNC_ADD_NON_SAMPLE_EVENT_CLASS(mmap2, __NON_SAMPLE_ADD_FIELD(u32, pid); __NON_SAMPLE_ADD_FIELD(u32, tid); __NON_SAMPLE_ADD_FIELD(u64_hex, start); __NON_SAMPLE_ADD_FIELD(string, filename); ) #undef __NON_SAMPLE_ADD_FIELD #undef __FUNC_ADD_NON_SAMPLE_EVENT_CLASS static int setup_non_sample_events(struct ctf_writer *cw, struct perf_session *session __maybe_unused) { int ret; ret = add_comm_event(cw); if (ret) return ret; ret = add_exit_event(cw); if (ret) return ret; ret = add_fork_event(cw); if (ret) return ret; ret = add_mmap_event(cw); if (ret) return ret; ret = add_mmap2_event(cw); if (ret) return ret; return 0; } static void cleanup_events(struct perf_session *session) { struct evlist *evlist = session->evlist; struct evsel *evsel; evlist__for_each_entry(evlist, evsel) { struct evsel_priv *priv; priv = evsel->priv; bt_ctf_event_class_put(priv->event_class); zfree(&evsel->priv); } evlist__delete(evlist); session->evlist = NULL; } static int setup_streams(struct ctf_writer *cw, struct perf_session *session) { struct ctf_stream **stream; struct perf_header *ph = &session->header; int ncpus; /* * Try to get the number of cpus used in the data file, * if not present fallback to the MAX_CPUS. */ ncpus = ph->env.nr_cpus_avail ?: MAX_CPUS; stream = zalloc(sizeof(*stream) * ncpus); if (!stream) { pr_err("Failed to allocate streams.\n"); return -ENOMEM; } cw->stream = stream; cw->stream_cnt = ncpus; return 0; } static void free_streams(struct ctf_writer *cw) { int cpu; for (cpu = 0; cpu < cw->stream_cnt; cpu++) ctf_stream__delete(cw->stream[cpu]); zfree(&cw->stream); } static int ctf_writer__setup_env(struct ctf_writer *cw, struct perf_session *session) { struct perf_header *header = &session->header; struct bt_ctf_writer *writer = cw->writer; #define ADD(__n, __v) \ do { \ if (bt_ctf_writer_add_environment_field(writer, __n, __v)) \ return -1; \ } while (0) ADD("host", header->env.hostname); ADD("sysname", "Linux"); ADD("release", header->env.os_release); ADD("version", header->env.version); ADD("machine", header->env.arch); ADD("domain", "kernel"); ADD("tracer_name", "perf"); #undef ADD return 0; } static int ctf_writer__setup_clock(struct ctf_writer *cw, struct perf_session *session, bool tod) { struct bt_ctf_clock *clock = cw->clock; const char *desc = "perf clock"; int64_t offset = 0; if (tod) { struct perf_env *env = &session->header.env; if (!env->clock.enabled) { pr_err("Can't provide --tod time, missing clock data. " "Please record with -k/--clockid option.\n"); return -1; } desc = clockid_name(env->clock.clockid); offset = env->clock.tod_ns - env->clock.clockid_ns; } #define SET(__n, __v) \ do { \ if (bt_ctf_clock_set_##__n(clock, __v)) \ return -1; \ } while (0) SET(frequency, 1000000000); SET(offset, offset); SET(description, desc); SET(precision, 10); SET(is_absolute, 0); #undef SET return 0; } static struct bt_ctf_field_type *create_int_type(int size, bool sign, bool hex) { struct bt_ctf_field_type *type; type = bt_ctf_field_type_integer_create(size); if (!type) return NULL; if (sign && bt_ctf_field_type_integer_set_signed(type, 1)) goto err; if (hex && bt_ctf_field_type_integer_set_base(type, BT_CTF_INTEGER_BASE_HEXADECIMAL)) goto err; #if __BYTE_ORDER__ == __ORDER_BIG_ENDIAN__ bt_ctf_field_type_set_byte_order(type, BT_CTF_BYTE_ORDER_BIG_ENDIAN); #else bt_ctf_field_type_set_byte_order(type, BT_CTF_BYTE_ORDER_LITTLE_ENDIAN); #endif pr2("Created type: INTEGER %d-bit %ssigned %s\n", size, sign ? "un" : "", hex ? "hex" : ""); return type; err: bt_ctf_field_type_put(type); return NULL; } static void ctf_writer__cleanup_data(struct ctf_writer *cw) { unsigned int i; for (i = 0; i < ARRAY_SIZE(cw->data.array); i++) bt_ctf_field_type_put(cw->data.array[i]); } static int ctf_writer__init_data(struct ctf_writer *cw) { #define CREATE_INT_TYPE(type, size, sign, hex) \ do { \ (type) = create_int_type(size, sign, hex); \ if (!(type)) \ goto err; \ } while (0) CREATE_INT_TYPE(cw->data.s64, 64, true, false); CREATE_INT_TYPE(cw->data.u64, 64, false, false); CREATE_INT_TYPE(cw->data.s32, 32, true, false); CREATE_INT_TYPE(cw->data.u32, 32, false, false); CREATE_INT_TYPE(cw->data.u32_hex, 32, false, true); CREATE_INT_TYPE(cw->data.u64_hex, 64, false, true); cw->data.string = bt_ctf_field_type_string_create(); if (cw->data.string) return 0; err: ctf_writer__cleanup_data(cw); pr_err("Failed to create data types.\n"); return -1; } static void ctf_writer__cleanup(struct ctf_writer *cw) { ctf_writer__cleanup_data(cw); bt_ctf_clock_put(cw->clock); free_streams(cw); bt_ctf_stream_class_put(cw->stream_class); bt_ctf_writer_put(cw->writer); /* and NULL all the pointers */ memset(cw, 0, sizeof(*cw)); } static int ctf_writer__init(struct ctf_writer *cw, const char *path, struct perf_session *session, bool tod) { struct bt_ctf_writer *writer; struct bt_ctf_stream_class *stream_class; struct bt_ctf_clock *clock; struct bt_ctf_field_type *pkt_ctx_type; int ret; /* CTF writer */ writer = bt_ctf_writer_create(path); if (!writer) goto err; cw->writer = writer; /* CTF clock */ clock = bt_ctf_clock_create("perf_clock"); if (!clock) { pr("Failed to create CTF clock.\n"); goto err_cleanup; } cw->clock = clock; if (ctf_writer__setup_clock(cw, session, tod)) { pr("Failed to setup CTF clock.\n"); goto err_cleanup; } /* CTF stream class */ stream_class = bt_ctf_stream_class_create("perf_stream"); if (!stream_class) { pr("Failed to create CTF stream class.\n"); goto err_cleanup; } cw->stream_class = stream_class; /* CTF clock stream setup */ if (bt_ctf_stream_class_set_clock(stream_class, clock)) { pr("Failed to assign CTF clock to stream class.\n"); goto err_cleanup; } if (ctf_writer__init_data(cw)) goto err_cleanup; /* Add cpu_id for packet context */ pkt_ctx_type = bt_ctf_stream_class_get_packet_context_type(stream_class); if (!pkt_ctx_type) goto err_cleanup; ret = bt_ctf_field_type_structure_add_field(pkt_ctx_type, cw->data.u32, "cpu_id"); bt_ctf_field_type_put(pkt_ctx_type); if (ret) goto err_cleanup; /* CTF clock writer setup */ if (bt_ctf_writer_add_clock(writer, clock)) { pr("Failed to assign CTF clock to writer.\n"); goto err_cleanup; } return 0; err_cleanup: ctf_writer__cleanup(cw); err: pr_err("Failed to setup CTF writer.\n"); return -1; } static int ctf_writer__flush_streams(struct ctf_writer *cw) { int cpu, ret = 0; for (cpu = 0; cpu < cw->stream_cnt && !ret; cpu++) ret = ctf_stream__flush(cw->stream[cpu]); return ret; } static int convert__config(const char *var, const char *value, void *cb) { struct convert *c = cb; if (!strcmp(var, "convert.queue-size")) return perf_config_u64(&c->queue_size, var, value); return 0; } int bt_convert__perf2ctf(const char *input, const char *path, struct perf_data_convert_opts *opts) { struct perf_session *session; struct perf_data data = { .path = input, .mode = PERF_DATA_MODE_READ, .force = opts->force, }; struct convert c = { .tool = { .sample = process_sample_event, .mmap = perf_event__process_mmap, .mmap2 = perf_event__process_mmap2, .comm = perf_event__process_comm, .exit = perf_event__process_exit, .fork = perf_event__process_fork, .lost = perf_event__process_lost, .tracing_data = perf_event__process_tracing_data, .build_id = perf_event__process_build_id, .namespaces = perf_event__process_namespaces, .ordered_events = true, .ordering_requires_timestamps = true, }, }; struct ctf_writer *cw = &c.writer; int err; if (opts->all) { c.tool.comm = process_comm_event; c.tool.exit = process_exit_event; c.tool.fork = process_fork_event; c.tool.mmap = process_mmap_event; c.tool.mmap2 = process_mmap2_event; } err = perf_config(convert__config, &c); if (err) return err; err = -1; /* perf.data session */ session = perf_session__new(&data, &c.tool); if (IS_ERR(session)) return PTR_ERR(session); /* CTF writer */ if (ctf_writer__init(cw, path, session, opts->tod)) goto free_session; if (c.queue_size) { ordered_events__set_alloc_size(&session->ordered_events, c.queue_size); } /* CTF writer env/clock setup */ if (ctf_writer__setup_env(cw, session)) goto free_writer; /* CTF events setup */ if (setup_events(cw, session)) goto free_writer; if (opts->all && setup_non_sample_events(cw, session)) goto free_writer; if (setup_streams(cw, session)) goto free_writer; err = perf_session__process_events(session); if (!err) err = ctf_writer__flush_streams(cw); else pr_err("Error during conversion.\n"); fprintf(stderr, "[ perf data convert: Converted '%s' into CTF data '%s' ]\n", data.path, path); fprintf(stderr, "[ perf data convert: Converted and wrote %.3f MB (%" PRIu64 " samples", (double) c.events_size / 1024.0 / 1024.0, c.events_count); if (!c.non_sample_count) fprintf(stderr, ") ]\n"); else fprintf(stderr, ", %" PRIu64 " non-samples) ]\n", c.non_sample_count); cleanup_events(session); perf_session__delete(session); ctf_writer__cleanup(cw); return err; free_writer: ctf_writer__cleanup(cw); free_session: perf_session__delete(session); pr_err("Error during conversion setup.\n"); return err; }