2 * CTF writing support via babeltrace.
4 * Copyright (C) 2014, Jiri Olsa <jolsa@redhat.com>
5 * Copyright (C) 2014, Sebastian Andrzej Siewior <bigeasy@linutronix.de>
7 * Released under the GPL v2. (and only v2, not any later version)
10 #include <linux/compiler.h>
11 #include <babeltrace/ctf-writer/writer.h>
12 #include <babeltrace/ctf-writer/clock.h>
13 #include <babeltrace/ctf-writer/stream.h>
14 #include <babeltrace/ctf-writer/event.h>
15 #include <babeltrace/ctf-writer/event-types.h>
16 #include <babeltrace/ctf-writer/event-fields.h>
17 #include <babeltrace/ctf-ir/utils.h>
18 #include <babeltrace/ctf/events.h>
19 #include <traceevent/event-parse.h>
21 #include "data-convert-bt.h"
30 #define pr_N(n, fmt, ...) \
31 eprintf(n, debug_data_convert, fmt, ##__VA_ARGS__)
33 #define pr(fmt, ...) pr_N(1, pr_fmt(fmt), ##__VA_ARGS__)
34 #define pr2(fmt, ...) pr_N(2, pr_fmt(fmt), ##__VA_ARGS__)
36 #define pr_time2(t, fmt, ...) pr_time_N(2, debug_data_convert, t, pr_fmt(fmt), ##__VA_ARGS__)
39 struct bt_ctf_event_class *event_class;
45 struct bt_ctf_stream *stream;
51 /* writer primitives */
52 struct bt_ctf_writer *writer;
53 struct ctf_stream **stream;
55 struct bt_ctf_stream_class *stream_class;
56 struct bt_ctf_clock *clock;
61 struct bt_ctf_field_type *s64;
62 struct bt_ctf_field_type *u64;
63 struct bt_ctf_field_type *s32;
64 struct bt_ctf_field_type *u32;
65 struct bt_ctf_field_type *string;
66 struct bt_ctf_field_type *u32_hex;
67 struct bt_ctf_field_type *u64_hex;
69 struct bt_ctf_field_type *array[6];
74 struct perf_tool tool;
75 struct ctf_writer writer;
80 /* Ordered events configured queue size. */
84 static int value_set(struct bt_ctf_field_type *type,
85 struct bt_ctf_event *event,
86 const char *name, u64 val)
88 struct bt_ctf_field *field;
89 bool sign = bt_ctf_field_type_integer_get_signed(type);
92 field = bt_ctf_field_create(type);
94 pr_err("failed to create a field %s\n", name);
99 ret = bt_ctf_field_signed_integer_set_value(field, val);
101 pr_err("failed to set field value %s\n", name);
105 ret = bt_ctf_field_unsigned_integer_set_value(field, val);
107 pr_err("failed to set field value %s\n", name);
112 ret = bt_ctf_event_set_payload(event, name, field);
114 pr_err("failed to set payload %s\n", name);
118 pr2(" SET [%s = %" PRIu64 "]\n", name, val);
121 bt_ctf_field_put(field);
125 #define __FUNC_VALUE_SET(_name, _val_type) \
126 static __maybe_unused int value_set_##_name(struct ctf_writer *cw, \
127 struct bt_ctf_event *event, \
131 struct bt_ctf_field_type *type = cw->data._name; \
132 return value_set(type, event, name, (u64) val); \
135 #define FUNC_VALUE_SET(_name) __FUNC_VALUE_SET(_name, _name)
141 __FUNC_VALUE_SET(u64_hex, u64)
143 static struct bt_ctf_field_type*
144 get_tracepoint_field_type(struct ctf_writer *cw, struct format_field *field)
146 unsigned long flags = field->flags;
148 if (flags & FIELD_IS_STRING)
149 return cw->data.string;
151 if (!(flags & FIELD_IS_SIGNED)) {
152 /* unsigned long are mostly pointers */
153 if (flags & FIELD_IS_LONG || flags & FIELD_IS_POINTER)
154 return cw->data.u64_hex;
157 if (flags & FIELD_IS_SIGNED) {
158 if (field->size == 8)
164 if (field->size == 8)
170 static unsigned long long adjust_signedness(unsigned long long value_int, int size)
172 unsigned long long value_mask;
175 * value_mask = (1 << (size * 8 - 1)) - 1.
176 * Directly set value_mask for code readers.
180 value_mask = 0x7fULL;
183 value_mask = 0x7fffULL;
186 value_mask = 0x7fffffffULL;
190 * For 64 bit value, return it self. There is no need
199 /* If it is a positive value, don't adjust. */
200 if ((value_int & (~0ULL - value_mask)) == 0)
203 /* Fill upper part of value_int with 1 to make it a negative long long. */
204 return (value_int & value_mask) | ~value_mask;
207 static int string_set_value(struct bt_ctf_field *field, const char *string)
210 size_t len = strlen(string), i, p;
213 for (i = p = 0; i < len; i++, p++) {
214 if (isprint(string[i])) {
217 buffer[p] = string[i];
221 snprintf(numstr, sizeof(numstr), "\\x%02x",
222 (unsigned int)(string[i]) & 0xff);
225 buffer = zalloc(i + (len - i) * 4 + 2);
227 pr_err("failed to set unprintable string '%s'\n", string);
228 return bt_ctf_field_string_set_value(field, "UNPRINTABLE-STRING");
231 strncpy(buffer, string, i);
233 strncat(buffer + p, numstr, 4);
239 return bt_ctf_field_string_set_value(field, string);
240 err = bt_ctf_field_string_set_value(field, buffer);
245 static int add_tracepoint_field_value(struct ctf_writer *cw,
246 struct bt_ctf_event_class *event_class,
247 struct bt_ctf_event *event,
248 struct perf_sample *sample,
249 struct format_field *fmtf)
251 struct bt_ctf_field_type *type;
252 struct bt_ctf_field *array_field;
253 struct bt_ctf_field *field;
254 const char *name = fmtf->name;
255 void *data = sample->raw_data;
256 unsigned long flags = fmtf->flags;
257 unsigned int n_items;
264 offset = fmtf->offset;
266 if (flags & FIELD_IS_STRING)
267 flags &= ~FIELD_IS_ARRAY;
269 if (flags & FIELD_IS_DYNAMIC) {
270 unsigned long long tmp_val;
272 tmp_val = pevent_read_number(fmtf->event->pevent,
279 if (flags & FIELD_IS_ARRAY) {
281 type = bt_ctf_event_class_get_field_by_name(
283 array_field = bt_ctf_field_create(type);
284 bt_ctf_field_type_put(type);
286 pr_err("Failed to create array type %s\n", name);
290 len = fmtf->size / fmtf->arraylen;
291 n_items = fmtf->arraylen;
297 type = get_tracepoint_field_type(cw, fmtf);
299 for (i = 0; i < n_items; i++) {
300 if (flags & FIELD_IS_ARRAY)
301 field = bt_ctf_field_array_get_field(array_field, i);
303 field = bt_ctf_field_create(type);
306 pr_err("failed to create a field %s\n", name);
310 if (flags & FIELD_IS_STRING)
311 ret = string_set_value(field, data + offset + i * len);
313 unsigned long long value_int;
315 value_int = pevent_read_number(
317 data + offset + i * len, len);
319 if (!(flags & FIELD_IS_SIGNED))
320 ret = bt_ctf_field_unsigned_integer_set_value(
323 ret = bt_ctf_field_signed_integer_set_value(
324 field, adjust_signedness(value_int, len));
328 pr_err("failed to set file value %s\n", name);
331 if (!(flags & FIELD_IS_ARRAY)) {
332 ret = bt_ctf_event_set_payload(event, name, field);
334 pr_err("failed to set payload %s\n", name);
338 bt_ctf_field_put(field);
340 if (flags & FIELD_IS_ARRAY) {
341 ret = bt_ctf_event_set_payload(event, name, array_field);
343 pr_err("Failed add payload array %s\n", name);
346 bt_ctf_field_put(array_field);
351 bt_ctf_field_put(field);
355 static int add_tracepoint_fields_values(struct ctf_writer *cw,
356 struct bt_ctf_event_class *event_class,
357 struct bt_ctf_event *event,
358 struct format_field *fields,
359 struct perf_sample *sample)
361 struct format_field *field;
364 for (field = fields; field; field = field->next) {
365 ret = add_tracepoint_field_value(cw, event_class, event, sample,
373 static int add_tracepoint_values(struct ctf_writer *cw,
374 struct bt_ctf_event_class *event_class,
375 struct bt_ctf_event *event,
376 struct perf_evsel *evsel,
377 struct perf_sample *sample)
379 struct format_field *common_fields = evsel->tp_format->format.common_fields;
380 struct format_field *fields = evsel->tp_format->format.fields;
383 ret = add_tracepoint_fields_values(cw, event_class, event,
384 common_fields, sample);
386 ret = add_tracepoint_fields_values(cw, event_class, event,
393 add_bpf_output_values(struct bt_ctf_event_class *event_class,
394 struct bt_ctf_event *event,
395 struct perf_sample *sample)
397 struct bt_ctf_field_type *len_type, *seq_type;
398 struct bt_ctf_field *len_field, *seq_field;
399 unsigned int raw_size = sample->raw_size;
400 unsigned int nr_elements = raw_size / sizeof(u32);
404 if (nr_elements * sizeof(u32) != raw_size)
405 pr_warning("Incorrect raw_size (%u) in bpf output event, skip %lu bytes\n",
406 raw_size, nr_elements * sizeof(u32) - raw_size);
408 len_type = bt_ctf_event_class_get_field_by_name(event_class, "raw_len");
409 len_field = bt_ctf_field_create(len_type);
411 pr_err("failed to create 'raw_len' for bpf output event\n");
416 ret = bt_ctf_field_unsigned_integer_set_value(len_field, nr_elements);
418 pr_err("failed to set field value for raw_len\n");
421 ret = bt_ctf_event_set_payload(event, "raw_len", len_field);
423 pr_err("failed to set payload to raw_len\n");
427 seq_type = bt_ctf_event_class_get_field_by_name(event_class, "raw_data");
428 seq_field = bt_ctf_field_create(seq_type);
430 pr_err("failed to create 'raw_data' for bpf output event\n");
435 ret = bt_ctf_field_sequence_set_length(seq_field, len_field);
437 pr_err("failed to set length of 'raw_data'\n");
441 for (i = 0; i < nr_elements; i++) {
442 struct bt_ctf_field *elem_field =
443 bt_ctf_field_sequence_get_field(seq_field, i);
445 ret = bt_ctf_field_unsigned_integer_set_value(elem_field,
446 ((u32 *)(sample->raw_data))[i]);
448 bt_ctf_field_put(elem_field);
450 pr_err("failed to set raw_data[%d]\n", i);
455 ret = bt_ctf_event_set_payload(event, "raw_data", seq_field);
457 pr_err("failed to set payload for raw_data\n");
460 bt_ctf_field_put(seq_field);
462 bt_ctf_field_type_put(seq_type);
464 bt_ctf_field_put(len_field);
466 bt_ctf_field_type_put(len_type);
470 static int add_generic_values(struct ctf_writer *cw,
471 struct bt_ctf_event *event,
472 struct perf_evsel *evsel,
473 struct perf_sample *sample)
475 u64 type = evsel->attr.sample_type;
480 * PERF_SAMPLE_TIME - not needed as we have it in
482 * PERF_SAMPLE_READ - TODO
483 * PERF_SAMPLE_CALLCHAIN - TODO
484 * PERF_SAMPLE_RAW - tracepoint fields are handled separately
485 * PERF_SAMPLE_BRANCH_STACK - TODO
486 * PERF_SAMPLE_REGS_USER - TODO
487 * PERF_SAMPLE_STACK_USER - TODO
490 if (type & PERF_SAMPLE_IP) {
491 ret = value_set_u64_hex(cw, event, "perf_ip", sample->ip);
496 if (type & PERF_SAMPLE_TID) {
497 ret = value_set_s32(cw, event, "perf_tid", sample->tid);
501 ret = value_set_s32(cw, event, "perf_pid", sample->pid);
506 if ((type & PERF_SAMPLE_ID) ||
507 (type & PERF_SAMPLE_IDENTIFIER)) {
508 ret = value_set_u64(cw, event, "perf_id", sample->id);
513 if (type & PERF_SAMPLE_STREAM_ID) {
514 ret = value_set_u64(cw, event, "perf_stream_id", sample->stream_id);
519 if (type & PERF_SAMPLE_PERIOD) {
520 ret = value_set_u64(cw, event, "perf_period", sample->period);
525 if (type & PERF_SAMPLE_WEIGHT) {
526 ret = value_set_u64(cw, event, "perf_weight", sample->weight);
531 if (type & PERF_SAMPLE_DATA_SRC) {
532 ret = value_set_u64(cw, event, "perf_data_src",
538 if (type & PERF_SAMPLE_TRANSACTION) {
539 ret = value_set_u64(cw, event, "perf_transaction",
540 sample->transaction);
548 static int ctf_stream__flush(struct ctf_stream *cs)
553 err = bt_ctf_stream_flush(cs->stream);
555 pr_err("CTF stream %d flush failed\n", cs->cpu);
557 pr("Flush stream for cpu %d (%u samples)\n",
566 static struct ctf_stream *ctf_stream__create(struct ctf_writer *cw, int cpu)
568 struct ctf_stream *cs;
569 struct bt_ctf_field *pkt_ctx = NULL;
570 struct bt_ctf_field *cpu_field = NULL;
571 struct bt_ctf_stream *stream = NULL;
574 cs = zalloc(sizeof(*cs));
576 pr_err("Failed to allocate ctf stream\n");
580 stream = bt_ctf_writer_create_stream(cw->writer, cw->stream_class);
582 pr_err("Failed to create CTF stream\n");
586 pkt_ctx = bt_ctf_stream_get_packet_context(stream);
588 pr_err("Failed to obtain packet context\n");
592 cpu_field = bt_ctf_field_structure_get_field(pkt_ctx, "cpu_id");
593 bt_ctf_field_put(pkt_ctx);
595 pr_err("Failed to obtain cpu field\n");
599 ret = bt_ctf_field_unsigned_integer_set_value(cpu_field, (u32) cpu);
601 pr_err("Failed to update CPU number\n");
605 bt_ctf_field_put(cpu_field);
613 bt_ctf_field_put(cpu_field);
615 bt_ctf_stream_put(stream);
621 static void ctf_stream__delete(struct ctf_stream *cs)
624 bt_ctf_stream_put(cs->stream);
629 static struct ctf_stream *ctf_stream(struct ctf_writer *cw, int cpu)
631 struct ctf_stream *cs = cw->stream[cpu];
634 cs = ctf_stream__create(cw, cpu);
635 cw->stream[cpu] = cs;
641 static int get_sample_cpu(struct ctf_writer *cw, struct perf_sample *sample,
642 struct perf_evsel *evsel)
646 if (evsel->attr.sample_type & PERF_SAMPLE_CPU)
649 if (cpu > cw->stream_cnt) {
650 pr_err("Event was recorded for CPU %d, limit is at %d.\n",
651 cpu, cw->stream_cnt);
658 #define STREAM_FLUSH_COUNT 100000
661 * Currently we have no other way to determine the
662 * time for the stream flush other than keep track
663 * of the number of events and check it against
666 static bool is_flush_needed(struct ctf_stream *cs)
668 return cs->count >= STREAM_FLUSH_COUNT;
671 static int process_sample_event(struct perf_tool *tool,
672 union perf_event *_event,
673 struct perf_sample *sample,
674 struct perf_evsel *evsel,
675 struct machine *machine __maybe_unused)
677 struct convert *c = container_of(tool, struct convert, tool);
678 struct evsel_priv *priv = evsel->priv;
679 struct ctf_writer *cw = &c->writer;
680 struct ctf_stream *cs;
681 struct bt_ctf_event_class *event_class;
682 struct bt_ctf_event *event;
685 if (WARN_ONCE(!priv, "Failed to setup all events.\n"))
688 event_class = priv->event_class;
692 c->events_size += _event->header.size;
694 pr_time2(sample->time, "sample %" PRIu64 "\n", c->events_count);
696 event = bt_ctf_event_create(event_class);
698 pr_err("Failed to create an CTF event\n");
702 bt_ctf_clock_set_time(cw->clock, sample->time);
704 ret = add_generic_values(cw, event, evsel, sample);
708 if (evsel->attr.type == PERF_TYPE_TRACEPOINT) {
709 ret = add_tracepoint_values(cw, event_class, event,
715 if (perf_evsel__is_bpf_output(evsel)) {
716 ret = add_bpf_output_values(event_class, event, sample);
721 cs = ctf_stream(cw, get_sample_cpu(cw, sample, evsel));
723 if (is_flush_needed(cs))
724 ctf_stream__flush(cs);
727 bt_ctf_stream_append_event(cs->stream, event);
730 bt_ctf_event_put(event);
734 /* If dup < 0, add a prefix. Else, add _dupl_X suffix. */
735 static char *change_name(char *name, char *orig_name, int dup)
737 char *new_name = NULL;
746 * Add '_' prefix to potential keywork. According to
747 * Mathieu Desnoyers (https://lkml.org/lkml/2015/1/23/652),
748 * futher CTF spec updating may require us to use '$'.
751 len = strlen(name) + sizeof("_");
753 len = strlen(orig_name) + sizeof("_dupl_X");
755 new_name = malloc(len);
760 snprintf(new_name, len, "_%s", name);
762 snprintf(new_name, len, "%s_dupl_%d", orig_name, dup);
765 if (name != orig_name)
770 static int event_class_add_field(struct bt_ctf_event_class *event_class,
771 struct bt_ctf_field_type *type,
772 struct format_field *field)
774 struct bt_ctf_field_type *t = NULL;
779 /* alias was already assigned */
780 if (field->alias != field->name)
781 return bt_ctf_event_class_add_field(event_class, type,
782 (char *)field->alias);
786 /* If 'name' is a keywork, add prefix. */
787 if (bt_ctf_validate_identifier(name))
788 name = change_name(name, field->name, -1);
791 pr_err("Failed to fix invalid identifier.");
794 while ((t = bt_ctf_event_class_get_field_by_name(event_class, name))) {
795 bt_ctf_field_type_put(t);
796 name = change_name(name, field->name, dup++);
798 pr_err("Failed to create dup name for '%s'\n", field->name);
803 ret = bt_ctf_event_class_add_field(event_class, type, name);
810 static int add_tracepoint_fields_types(struct ctf_writer *cw,
811 struct format_field *fields,
812 struct bt_ctf_event_class *event_class)
814 struct format_field *field;
817 for (field = fields; field; field = field->next) {
818 struct bt_ctf_field_type *type;
819 unsigned long flags = field->flags;
821 pr2(" field '%s'\n", field->name);
823 type = get_tracepoint_field_type(cw, field);
828 * A string is an array of chars. For this we use the string
829 * type and don't care that it is an array. What we don't
830 * support is an array of strings.
832 if (flags & FIELD_IS_STRING)
833 flags &= ~FIELD_IS_ARRAY;
835 if (flags & FIELD_IS_ARRAY)
836 type = bt_ctf_field_type_array_create(type, field->arraylen);
838 ret = event_class_add_field(event_class, type, field);
840 if (flags & FIELD_IS_ARRAY)
841 bt_ctf_field_type_put(type);
844 pr_err("Failed to add field '%s': %d\n",
853 static int add_tracepoint_types(struct ctf_writer *cw,
854 struct perf_evsel *evsel,
855 struct bt_ctf_event_class *class)
857 struct format_field *common_fields = evsel->tp_format->format.common_fields;
858 struct format_field *fields = evsel->tp_format->format.fields;
861 ret = add_tracepoint_fields_types(cw, common_fields, class);
863 ret = add_tracepoint_fields_types(cw, fields, class);
868 static int add_bpf_output_types(struct ctf_writer *cw,
869 struct bt_ctf_event_class *class)
871 struct bt_ctf_field_type *len_type = cw->data.u32;
872 struct bt_ctf_field_type *seq_base_type = cw->data.u32_hex;
873 struct bt_ctf_field_type *seq_type;
876 ret = bt_ctf_event_class_add_field(class, len_type, "raw_len");
880 seq_type = bt_ctf_field_type_sequence_create(seq_base_type, "raw_len");
884 return bt_ctf_event_class_add_field(class, seq_type, "raw_data");
887 static int add_generic_types(struct ctf_writer *cw, struct perf_evsel *evsel,
888 struct bt_ctf_event_class *event_class)
890 u64 type = evsel->attr.sample_type;
894 * PERF_SAMPLE_TIME - not needed as we have it in
896 * PERF_SAMPLE_READ - TODO
897 * PERF_SAMPLE_CALLCHAIN - TODO
898 * PERF_SAMPLE_RAW - tracepoint fields and BPF output
899 * are handled separately
900 * PERF_SAMPLE_BRANCH_STACK - TODO
901 * PERF_SAMPLE_REGS_USER - TODO
902 * PERF_SAMPLE_STACK_USER - TODO
905 #define ADD_FIELD(cl, t, n) \
907 pr2(" field '%s'\n", n); \
908 if (bt_ctf_event_class_add_field(cl, t, n)) { \
909 pr_err("Failed to add field '%s';\n", n); \
914 if (type & PERF_SAMPLE_IP)
915 ADD_FIELD(event_class, cw->data.u64_hex, "perf_ip");
917 if (type & PERF_SAMPLE_TID) {
918 ADD_FIELD(event_class, cw->data.s32, "perf_tid");
919 ADD_FIELD(event_class, cw->data.s32, "perf_pid");
922 if ((type & PERF_SAMPLE_ID) ||
923 (type & PERF_SAMPLE_IDENTIFIER))
924 ADD_FIELD(event_class, cw->data.u64, "perf_id");
926 if (type & PERF_SAMPLE_STREAM_ID)
927 ADD_FIELD(event_class, cw->data.u64, "perf_stream_id");
929 if (type & PERF_SAMPLE_PERIOD)
930 ADD_FIELD(event_class, cw->data.u64, "perf_period");
932 if (type & PERF_SAMPLE_WEIGHT)
933 ADD_FIELD(event_class, cw->data.u64, "perf_weight");
935 if (type & PERF_SAMPLE_DATA_SRC)
936 ADD_FIELD(event_class, cw->data.u64, "perf_data_src");
938 if (type & PERF_SAMPLE_TRANSACTION)
939 ADD_FIELD(event_class, cw->data.u64, "perf_transaction");
945 static int add_event(struct ctf_writer *cw, struct perf_evsel *evsel)
947 struct bt_ctf_event_class *event_class;
948 struct evsel_priv *priv;
949 const char *name = perf_evsel__name(evsel);
952 pr("Adding event '%s' (type %d)\n", name, evsel->attr.type);
954 event_class = bt_ctf_event_class_create(name);
958 ret = add_generic_types(cw, evsel, event_class);
962 if (evsel->attr.type == PERF_TYPE_TRACEPOINT) {
963 ret = add_tracepoint_types(cw, evsel, event_class);
968 if (perf_evsel__is_bpf_output(evsel)) {
969 ret = add_bpf_output_types(cw, event_class);
974 ret = bt_ctf_stream_class_add_event_class(cw->stream_class, event_class);
976 pr("Failed to add event class into stream.\n");
980 priv = malloc(sizeof(*priv));
984 priv->event_class = event_class;
989 bt_ctf_event_class_put(event_class);
990 pr_err("Failed to add event '%s'.\n", name);
994 static int setup_events(struct ctf_writer *cw, struct perf_session *session)
996 struct perf_evlist *evlist = session->evlist;
997 struct perf_evsel *evsel;
1000 evlist__for_each(evlist, evsel) {
1001 ret = add_event(cw, evsel);
1008 static void cleanup_events(struct perf_session *session)
1010 struct perf_evlist *evlist = session->evlist;
1011 struct perf_evsel *evsel;
1013 evlist__for_each(evlist, evsel) {
1014 struct evsel_priv *priv;
1017 bt_ctf_event_class_put(priv->event_class);
1018 zfree(&evsel->priv);
1021 perf_evlist__delete(evlist);
1022 session->evlist = NULL;
1025 static int setup_streams(struct ctf_writer *cw, struct perf_session *session)
1027 struct ctf_stream **stream;
1028 struct perf_header *ph = &session->header;
1032 * Try to get the number of cpus used in the data file,
1033 * if not present fallback to the MAX_CPUS.
1035 ncpus = ph->env.nr_cpus_avail ?: MAX_CPUS;
1037 stream = zalloc(sizeof(*stream) * ncpus);
1039 pr_err("Failed to allocate streams.\n");
1043 cw->stream = stream;
1044 cw->stream_cnt = ncpus;
1048 static void free_streams(struct ctf_writer *cw)
1052 for (cpu = 0; cpu < cw->stream_cnt; cpu++)
1053 ctf_stream__delete(cw->stream[cpu]);
1058 static int ctf_writer__setup_env(struct ctf_writer *cw,
1059 struct perf_session *session)
1061 struct perf_header *header = &session->header;
1062 struct bt_ctf_writer *writer = cw->writer;
1064 #define ADD(__n, __v) \
1066 if (bt_ctf_writer_add_environment_field(writer, __n, __v)) \
1070 ADD("host", header->env.hostname);
1071 ADD("sysname", "Linux");
1072 ADD("release", header->env.os_release);
1073 ADD("version", header->env.version);
1074 ADD("machine", header->env.arch);
1075 ADD("domain", "kernel");
1076 ADD("tracer_name", "perf");
1082 static int ctf_writer__setup_clock(struct ctf_writer *cw)
1084 struct bt_ctf_clock *clock = cw->clock;
1086 bt_ctf_clock_set_description(clock, "perf clock");
1088 #define SET(__n, __v) \
1090 if (bt_ctf_clock_set_##__n(clock, __v)) \
1094 SET(frequency, 1000000000);
1098 SET(is_absolute, 0);
1104 static struct bt_ctf_field_type *create_int_type(int size, bool sign, bool hex)
1106 struct bt_ctf_field_type *type;
1108 type = bt_ctf_field_type_integer_create(size);
1113 bt_ctf_field_type_integer_set_signed(type, 1))
1117 bt_ctf_field_type_integer_set_base(type, BT_CTF_INTEGER_BASE_HEXADECIMAL))
1120 #if __BYTE_ORDER == __BIG_ENDIAN
1121 bt_ctf_field_type_set_byte_order(type, BT_CTF_BYTE_ORDER_BIG_ENDIAN);
1123 bt_ctf_field_type_set_byte_order(type, BT_CTF_BYTE_ORDER_LITTLE_ENDIAN);
1126 pr2("Created type: INTEGER %d-bit %ssigned %s\n",
1127 size, sign ? "un" : "", hex ? "hex" : "");
1131 bt_ctf_field_type_put(type);
1135 static void ctf_writer__cleanup_data(struct ctf_writer *cw)
1139 for (i = 0; i < ARRAY_SIZE(cw->data.array); i++)
1140 bt_ctf_field_type_put(cw->data.array[i]);
1143 static int ctf_writer__init_data(struct ctf_writer *cw)
1145 #define CREATE_INT_TYPE(type, size, sign, hex) \
1147 (type) = create_int_type(size, sign, hex); \
1152 CREATE_INT_TYPE(cw->data.s64, 64, true, false);
1153 CREATE_INT_TYPE(cw->data.u64, 64, false, false);
1154 CREATE_INT_TYPE(cw->data.s32, 32, true, false);
1155 CREATE_INT_TYPE(cw->data.u32, 32, false, false);
1156 CREATE_INT_TYPE(cw->data.u32_hex, 32, false, true);
1157 CREATE_INT_TYPE(cw->data.u64_hex, 64, false, true);
1159 cw->data.string = bt_ctf_field_type_string_create();
1160 if (cw->data.string)
1164 ctf_writer__cleanup_data(cw);
1165 pr_err("Failed to create data types.\n");
1169 static void ctf_writer__cleanup(struct ctf_writer *cw)
1171 ctf_writer__cleanup_data(cw);
1173 bt_ctf_clock_put(cw->clock);
1175 bt_ctf_stream_class_put(cw->stream_class);
1176 bt_ctf_writer_put(cw->writer);
1178 /* and NULL all the pointers */
1179 memset(cw, 0, sizeof(*cw));
1182 static int ctf_writer__init(struct ctf_writer *cw, const char *path)
1184 struct bt_ctf_writer *writer;
1185 struct bt_ctf_stream_class *stream_class;
1186 struct bt_ctf_clock *clock;
1187 struct bt_ctf_field_type *pkt_ctx_type;
1191 writer = bt_ctf_writer_create(path);
1195 cw->writer = writer;
1198 clock = bt_ctf_clock_create("perf_clock");
1200 pr("Failed to create CTF clock.\n");
1206 if (ctf_writer__setup_clock(cw)) {
1207 pr("Failed to setup CTF clock.\n");
1211 /* CTF stream class */
1212 stream_class = bt_ctf_stream_class_create("perf_stream");
1213 if (!stream_class) {
1214 pr("Failed to create CTF stream class.\n");
1218 cw->stream_class = stream_class;
1220 /* CTF clock stream setup */
1221 if (bt_ctf_stream_class_set_clock(stream_class, clock)) {
1222 pr("Failed to assign CTF clock to stream class.\n");
1226 if (ctf_writer__init_data(cw))
1229 /* Add cpu_id for packet context */
1230 pkt_ctx_type = bt_ctf_stream_class_get_packet_context_type(stream_class);
1234 ret = bt_ctf_field_type_structure_add_field(pkt_ctx_type, cw->data.u32, "cpu_id");
1235 bt_ctf_field_type_put(pkt_ctx_type);
1239 /* CTF clock writer setup */
1240 if (bt_ctf_writer_add_clock(writer, clock)) {
1241 pr("Failed to assign CTF clock to writer.\n");
1248 ctf_writer__cleanup(cw);
1250 pr_err("Failed to setup CTF writer.\n");
1254 static int ctf_writer__flush_streams(struct ctf_writer *cw)
1258 for (cpu = 0; cpu < cw->stream_cnt && !ret; cpu++)
1259 ret = ctf_stream__flush(cw->stream[cpu]);
1264 static int convert__config(const char *var, const char *value, void *cb)
1266 struct convert *c = cb;
1268 if (!strcmp(var, "convert.queue-size")) {
1269 c->queue_size = perf_config_u64(var, value);
1276 int bt_convert__perf2ctf(const char *input, const char *path, bool force)
1278 struct perf_session *session;
1279 struct perf_data_file file = {
1281 .mode = PERF_DATA_MODE_READ,
1284 struct convert c = {
1286 .sample = process_sample_event,
1287 .mmap = perf_event__process_mmap,
1288 .mmap2 = perf_event__process_mmap2,
1289 .comm = perf_event__process_comm,
1290 .exit = perf_event__process_exit,
1291 .fork = perf_event__process_fork,
1292 .lost = perf_event__process_lost,
1293 .tracing_data = perf_event__process_tracing_data,
1294 .build_id = perf_event__process_build_id,
1295 .ordered_events = true,
1296 .ordering_requires_timestamps = true,
1299 struct ctf_writer *cw = &c.writer;
1302 perf_config(convert__config, &c);
1305 if (ctf_writer__init(cw, path))
1308 /* perf.data session */
1309 session = perf_session__new(&file, 0, &c.tool);
1314 ordered_events__set_alloc_size(&session->ordered_events,
1318 /* CTF writer env/clock setup */
1319 if (ctf_writer__setup_env(cw, session))
1322 /* CTF events setup */
1323 if (setup_events(cw, session))
1326 if (setup_streams(cw, session))
1329 err = perf_session__process_events(session);
1331 err = ctf_writer__flush_streams(cw);
1333 pr_err("Error during conversion.\n");
1336 "[ perf data convert: Converted '%s' into CTF data '%s' ]\n",
1340 "[ perf data convert: Converted and wrote %.3f MB (%" PRIu64 " samples) ]\n",
1341 (double) c.events_size / 1024.0 / 1024.0,
1344 cleanup_events(session);
1345 perf_session__delete(session);
1346 ctf_writer__cleanup(cw);
1351 perf_session__delete(session);
1353 ctf_writer__cleanup(cw);
1354 pr_err("Error during conversion setup.\n");