]> asedeno.scripts.mit.edu Git - linux.git/blob - tools/perf/builtin-trace.c
shmem: avoid maybe-uninitialized warning
[linux.git] / tools / perf / builtin-trace.c
1 /*
2  * builtin-trace.c
3  *
4  * Builtin 'trace' command:
5  *
6  * Display a continuously updated trace of any workload, CPU, specific PID,
7  * system wide, etc.  Default format is loosely strace like, but any other
8  * event may be specified using --event.
9  *
10  * Copyright (C) 2012, 2013, 2014, 2015 Red Hat Inc, Arnaldo Carvalho de Melo <acme@redhat.com>
11  *
12  * Initially based on the 'trace' prototype by Thomas Gleixner:
13  *
14  * http://lwn.net/Articles/415728/ ("Announcing a new utility: 'trace'")
15  *
16  * Released under the GPL v2. (and only v2, not any later version)
17  */
18
19 #include <traceevent/event-parse.h>
20 #include <api/fs/tracing_path.h>
21 #include "builtin.h"
22 #include "util/color.h"
23 #include "util/debug.h"
24 #include "util/evlist.h"
25 #include <subcmd/exec-cmd.h>
26 #include "util/machine.h"
27 #include "util/session.h"
28 #include "util/thread.h"
29 #include <subcmd/parse-options.h>
30 #include "util/strlist.h"
31 #include "util/intlist.h"
32 #include "util/thread_map.h"
33 #include "util/stat.h"
34 #include "trace-event.h"
35 #include "util/parse-events.h"
36 #include "util/bpf-loader.h"
37 #include "callchain.h"
38 #include "syscalltbl.h"
39 #include "rb_resort.h"
40
41 #include <libaudit.h> /* FIXME: Still needed for audit_errno_to_name */
42 #include <stdlib.h>
43 #include <linux/err.h>
44 #include <linux/filter.h>
45 #include <linux/audit.h>
46 #include <linux/random.h>
47 #include <linux/stringify.h>
48 #include <linux/time64.h>
49
50 #ifndef O_CLOEXEC
51 # define O_CLOEXEC              02000000
52 #endif
53
54 struct trace {
55         struct perf_tool        tool;
56         struct syscalltbl       *sctbl;
57         struct {
58                 int             max;
59                 struct syscall  *table;
60                 struct {
61                         struct perf_evsel *sys_enter,
62                                           *sys_exit;
63                 }               events;
64         } syscalls;
65         struct record_opts      opts;
66         struct perf_evlist      *evlist;
67         struct machine          *host;
68         struct thread           *current;
69         u64                     base_time;
70         FILE                    *output;
71         unsigned long           nr_events;
72         struct strlist          *ev_qualifier;
73         struct {
74                 size_t          nr;
75                 int             *entries;
76         }                       ev_qualifier_ids;
77         struct {
78                 size_t          nr;
79                 pid_t           *entries;
80         }                       filter_pids;
81         double                  duration_filter;
82         double                  runtime_ms;
83         struct {
84                 u64             vfs_getname,
85                                 proc_getname;
86         } stats;
87         unsigned int            max_stack;
88         unsigned int            min_stack;
89         bool                    not_ev_qualifier;
90         bool                    live;
91         bool                    full_time;
92         bool                    sched;
93         bool                    multiple_threads;
94         bool                    summary;
95         bool                    summary_only;
96         bool                    show_comm;
97         bool                    show_tool_stats;
98         bool                    trace_syscalls;
99         bool                    kernel_syscallchains;
100         bool                    force;
101         bool                    vfs_getname;
102         int                     trace_pgfaults;
103         int                     open_id;
104 };
105
106 struct tp_field {
107         int offset;
108         union {
109                 u64 (*integer)(struct tp_field *field, struct perf_sample *sample);
110                 void *(*pointer)(struct tp_field *field, struct perf_sample *sample);
111         };
112 };
113
114 #define TP_UINT_FIELD(bits) \
115 static u64 tp_field__u##bits(struct tp_field *field, struct perf_sample *sample) \
116 { \
117         u##bits value; \
118         memcpy(&value, sample->raw_data + field->offset, sizeof(value)); \
119         return value;  \
120 }
121
122 TP_UINT_FIELD(8);
123 TP_UINT_FIELD(16);
124 TP_UINT_FIELD(32);
125 TP_UINT_FIELD(64);
126
127 #define TP_UINT_FIELD__SWAPPED(bits) \
128 static u64 tp_field__swapped_u##bits(struct tp_field *field, struct perf_sample *sample) \
129 { \
130         u##bits value; \
131         memcpy(&value, sample->raw_data + field->offset, sizeof(value)); \
132         return bswap_##bits(value);\
133 }
134
135 TP_UINT_FIELD__SWAPPED(16);
136 TP_UINT_FIELD__SWAPPED(32);
137 TP_UINT_FIELD__SWAPPED(64);
138
139 static int tp_field__init_uint(struct tp_field *field,
140                                struct format_field *format_field,
141                                bool needs_swap)
142 {
143         field->offset = format_field->offset;
144
145         switch (format_field->size) {
146         case 1:
147                 field->integer = tp_field__u8;
148                 break;
149         case 2:
150                 field->integer = needs_swap ? tp_field__swapped_u16 : tp_field__u16;
151                 break;
152         case 4:
153                 field->integer = needs_swap ? tp_field__swapped_u32 : tp_field__u32;
154                 break;
155         case 8:
156                 field->integer = needs_swap ? tp_field__swapped_u64 : tp_field__u64;
157                 break;
158         default:
159                 return -1;
160         }
161
162         return 0;
163 }
164
165 static void *tp_field__ptr(struct tp_field *field, struct perf_sample *sample)
166 {
167         return sample->raw_data + field->offset;
168 }
169
170 static int tp_field__init_ptr(struct tp_field *field, struct format_field *format_field)
171 {
172         field->offset = format_field->offset;
173         field->pointer = tp_field__ptr;
174         return 0;
175 }
176
177 struct syscall_tp {
178         struct tp_field id;
179         union {
180                 struct tp_field args, ret;
181         };
182 };
183
184 static int perf_evsel__init_tp_uint_field(struct perf_evsel *evsel,
185                                           struct tp_field *field,
186                                           const char *name)
187 {
188         struct format_field *format_field = perf_evsel__field(evsel, name);
189
190         if (format_field == NULL)
191                 return -1;
192
193         return tp_field__init_uint(field, format_field, evsel->needs_swap);
194 }
195
196 #define perf_evsel__init_sc_tp_uint_field(evsel, name) \
197         ({ struct syscall_tp *sc = evsel->priv;\
198            perf_evsel__init_tp_uint_field(evsel, &sc->name, #name); })
199
200 static int perf_evsel__init_tp_ptr_field(struct perf_evsel *evsel,
201                                          struct tp_field *field,
202                                          const char *name)
203 {
204         struct format_field *format_field = perf_evsel__field(evsel, name);
205
206         if (format_field == NULL)
207                 return -1;
208
209         return tp_field__init_ptr(field, format_field);
210 }
211
212 #define perf_evsel__init_sc_tp_ptr_field(evsel, name) \
213         ({ struct syscall_tp *sc = evsel->priv;\
214            perf_evsel__init_tp_ptr_field(evsel, &sc->name, #name); })
215
216 static void perf_evsel__delete_priv(struct perf_evsel *evsel)
217 {
218         zfree(&evsel->priv);
219         perf_evsel__delete(evsel);
220 }
221
222 static int perf_evsel__init_syscall_tp(struct perf_evsel *evsel, void *handler)
223 {
224         evsel->priv = malloc(sizeof(struct syscall_tp));
225         if (evsel->priv != NULL) {
226                 if (perf_evsel__init_sc_tp_uint_field(evsel, id))
227                         goto out_delete;
228
229                 evsel->handler = handler;
230                 return 0;
231         }
232
233         return -ENOMEM;
234
235 out_delete:
236         zfree(&evsel->priv);
237         return -ENOENT;
238 }
239
240 static struct perf_evsel *perf_evsel__syscall_newtp(const char *direction, void *handler)
241 {
242         struct perf_evsel *evsel = perf_evsel__newtp("raw_syscalls", direction);
243
244         /* older kernel (e.g., RHEL6) use syscalls:{enter,exit} */
245         if (IS_ERR(evsel))
246                 evsel = perf_evsel__newtp("syscalls", direction);
247
248         if (IS_ERR(evsel))
249                 return NULL;
250
251         if (perf_evsel__init_syscall_tp(evsel, handler))
252                 goto out_delete;
253
254         return evsel;
255
256 out_delete:
257         perf_evsel__delete_priv(evsel);
258         return NULL;
259 }
260
261 #define perf_evsel__sc_tp_uint(evsel, name, sample) \
262         ({ struct syscall_tp *fields = evsel->priv; \
263            fields->name.integer(&fields->name, sample); })
264
265 #define perf_evsel__sc_tp_ptr(evsel, name, sample) \
266         ({ struct syscall_tp *fields = evsel->priv; \
267            fields->name.pointer(&fields->name, sample); })
268
269 struct syscall_arg {
270         unsigned long val;
271         struct thread *thread;
272         struct trace  *trace;
273         void          *parm;
274         u8            idx;
275         u8            mask;
276 };
277
278 struct strarray {
279         int         offset;
280         int         nr_entries;
281         const char **entries;
282 };
283
284 #define DEFINE_STRARRAY(array) struct strarray strarray__##array = { \
285         .nr_entries = ARRAY_SIZE(array), \
286         .entries = array, \
287 }
288
289 #define DEFINE_STRARRAY_OFFSET(array, off) struct strarray strarray__##array = { \
290         .offset     = off, \
291         .nr_entries = ARRAY_SIZE(array), \
292         .entries = array, \
293 }
294
295 static size_t __syscall_arg__scnprintf_strarray(char *bf, size_t size,
296                                                 const char *intfmt,
297                                                 struct syscall_arg *arg)
298 {
299         struct strarray *sa = arg->parm;
300         int idx = arg->val - sa->offset;
301
302         if (idx < 0 || idx >= sa->nr_entries)
303                 return scnprintf(bf, size, intfmt, arg->val);
304
305         return scnprintf(bf, size, "%s", sa->entries[idx]);
306 }
307
308 static size_t syscall_arg__scnprintf_strarray(char *bf, size_t size,
309                                               struct syscall_arg *arg)
310 {
311         return __syscall_arg__scnprintf_strarray(bf, size, "%d", arg);
312 }
313
314 #define SCA_STRARRAY syscall_arg__scnprintf_strarray
315
316 #if defined(__i386__) || defined(__x86_64__)
317 /*
318  * FIXME: Make this available to all arches as soon as the ioctl beautifier
319  *        gets rewritten to support all arches.
320  */
321 static size_t syscall_arg__scnprintf_strhexarray(char *bf, size_t size,
322                                                  struct syscall_arg *arg)
323 {
324         return __syscall_arg__scnprintf_strarray(bf, size, "%#x", arg);
325 }
326
327 #define SCA_STRHEXARRAY syscall_arg__scnprintf_strhexarray
328 #endif /* defined(__i386__) || defined(__x86_64__) */
329
330 static size_t syscall_arg__scnprintf_fd(char *bf, size_t size,
331                                         struct syscall_arg *arg);
332
333 #define SCA_FD syscall_arg__scnprintf_fd
334
335 #ifndef AT_FDCWD
336 #define AT_FDCWD        -100
337 #endif
338
339 static size_t syscall_arg__scnprintf_fd_at(char *bf, size_t size,
340                                            struct syscall_arg *arg)
341 {
342         int fd = arg->val;
343
344         if (fd == AT_FDCWD)
345                 return scnprintf(bf, size, "CWD");
346
347         return syscall_arg__scnprintf_fd(bf, size, arg);
348 }
349
350 #define SCA_FDAT syscall_arg__scnprintf_fd_at
351
352 static size_t syscall_arg__scnprintf_close_fd(char *bf, size_t size,
353                                               struct syscall_arg *arg);
354
355 #define SCA_CLOSE_FD syscall_arg__scnprintf_close_fd
356
357 static size_t syscall_arg__scnprintf_hex(char *bf, size_t size,
358                                          struct syscall_arg *arg)
359 {
360         return scnprintf(bf, size, "%#lx", arg->val);
361 }
362
363 #define SCA_HEX syscall_arg__scnprintf_hex
364
365 static size_t syscall_arg__scnprintf_int(char *bf, size_t size,
366                                          struct syscall_arg *arg)
367 {
368         return scnprintf(bf, size, "%d", arg->val);
369 }
370
371 #define SCA_INT syscall_arg__scnprintf_int
372
373 static const char *bpf_cmd[] = {
374         "MAP_CREATE", "MAP_LOOKUP_ELEM", "MAP_UPDATE_ELEM", "MAP_DELETE_ELEM",
375         "MAP_GET_NEXT_KEY", "PROG_LOAD",
376 };
377 static DEFINE_STRARRAY(bpf_cmd);
378
379 static const char *epoll_ctl_ops[] = { "ADD", "DEL", "MOD", };
380 static DEFINE_STRARRAY_OFFSET(epoll_ctl_ops, 1);
381
382 static const char *itimers[] = { "REAL", "VIRTUAL", "PROF", };
383 static DEFINE_STRARRAY(itimers);
384
385 static const char *keyctl_options[] = {
386         "GET_KEYRING_ID", "JOIN_SESSION_KEYRING", "UPDATE", "REVOKE", "CHOWN",
387         "SETPERM", "DESCRIBE", "CLEAR", "LINK", "UNLINK", "SEARCH", "READ",
388         "INSTANTIATE", "NEGATE", "SET_REQKEY_KEYRING", "SET_TIMEOUT",
389         "ASSUME_AUTHORITY", "GET_SECURITY", "SESSION_TO_PARENT", "REJECT",
390         "INSTANTIATE_IOV", "INVALIDATE", "GET_PERSISTENT",
391 };
392 static DEFINE_STRARRAY(keyctl_options);
393
394 static const char *whences[] = { "SET", "CUR", "END",
395 #ifdef SEEK_DATA
396 "DATA",
397 #endif
398 #ifdef SEEK_HOLE
399 "HOLE",
400 #endif
401 };
402 static DEFINE_STRARRAY(whences);
403
404 static const char *fcntl_cmds[] = {
405         "DUPFD", "GETFD", "SETFD", "GETFL", "SETFL", "GETLK", "SETLK",
406         "SETLKW", "SETOWN", "GETOWN", "SETSIG", "GETSIG", "F_GETLK64",
407         "F_SETLK64", "F_SETLKW64", "F_SETOWN_EX", "F_GETOWN_EX",
408         "F_GETOWNER_UIDS",
409 };
410 static DEFINE_STRARRAY(fcntl_cmds);
411
412 static const char *rlimit_resources[] = {
413         "CPU", "FSIZE", "DATA", "STACK", "CORE", "RSS", "NPROC", "NOFILE",
414         "MEMLOCK", "AS", "LOCKS", "SIGPENDING", "MSGQUEUE", "NICE", "RTPRIO",
415         "RTTIME",
416 };
417 static DEFINE_STRARRAY(rlimit_resources);
418
419 static const char *sighow[] = { "BLOCK", "UNBLOCK", "SETMASK", };
420 static DEFINE_STRARRAY(sighow);
421
422 static const char *clockid[] = {
423         "REALTIME", "MONOTONIC", "PROCESS_CPUTIME_ID", "THREAD_CPUTIME_ID",
424         "MONOTONIC_RAW", "REALTIME_COARSE", "MONOTONIC_COARSE", "BOOTTIME",
425         "REALTIME_ALARM", "BOOTTIME_ALARM", "SGI_CYCLE", "TAI"
426 };
427 static DEFINE_STRARRAY(clockid);
428
429 static const char *socket_families[] = {
430         "UNSPEC", "LOCAL", "INET", "AX25", "IPX", "APPLETALK", "NETROM",
431         "BRIDGE", "ATMPVC", "X25", "INET6", "ROSE", "DECnet", "NETBEUI",
432         "SECURITY", "KEY", "NETLINK", "PACKET", "ASH", "ECONET", "ATMSVC",
433         "RDS", "SNA", "IRDA", "PPPOX", "WANPIPE", "LLC", "IB", "CAN", "TIPC",
434         "BLUETOOTH", "IUCV", "RXRPC", "ISDN", "PHONET", "IEEE802154", "CAIF",
435         "ALG", "NFC", "VSOCK",
436 };
437 static DEFINE_STRARRAY(socket_families);
438
439 static size_t syscall_arg__scnprintf_access_mode(char *bf, size_t size,
440                                                  struct syscall_arg *arg)
441 {
442         size_t printed = 0;
443         int mode = arg->val;
444
445         if (mode == F_OK) /* 0 */
446                 return scnprintf(bf, size, "F");
447 #define P_MODE(n) \
448         if (mode & n##_OK) { \
449                 printed += scnprintf(bf + printed, size - printed, "%s", #n); \
450                 mode &= ~n##_OK; \
451         }
452
453         P_MODE(R);
454         P_MODE(W);
455         P_MODE(X);
456 #undef P_MODE
457
458         if (mode)
459                 printed += scnprintf(bf + printed, size - printed, "|%#x", mode);
460
461         return printed;
462 }
463
464 #define SCA_ACCMODE syscall_arg__scnprintf_access_mode
465
466 static size_t syscall_arg__scnprintf_filename(char *bf, size_t size,
467                                               struct syscall_arg *arg);
468
469 #define SCA_FILENAME syscall_arg__scnprintf_filename
470
471 static size_t syscall_arg__scnprintf_pipe_flags(char *bf, size_t size,
472                                                 struct syscall_arg *arg)
473 {
474         int printed = 0, flags = arg->val;
475
476 #define P_FLAG(n) \
477         if (flags & O_##n) { \
478                 printed += scnprintf(bf + printed, size - printed, "%s%s", printed ? "|" : "", #n); \
479                 flags &= ~O_##n; \
480         }
481
482         P_FLAG(CLOEXEC);
483         P_FLAG(NONBLOCK);
484 #undef P_FLAG
485
486         if (flags)
487                 printed += scnprintf(bf + printed, size - printed, "%s%#x", printed ? "|" : "", flags);
488
489         return printed;
490 }
491
492 #define SCA_PIPE_FLAGS syscall_arg__scnprintf_pipe_flags
493
494 #if defined(__i386__) || defined(__x86_64__)
495 /*
496  * FIXME: Make this available to all arches.
497  */
498 #define TCGETS          0x5401
499
500 static const char *tioctls[] = {
501         "TCGETS", "TCSETS", "TCSETSW", "TCSETSF", "TCGETA", "TCSETA", "TCSETAW",
502         "TCSETAF", "TCSBRK", "TCXONC", "TCFLSH", "TIOCEXCL", "TIOCNXCL",
503         "TIOCSCTTY", "TIOCGPGRP", "TIOCSPGRP", "TIOCOUTQ", "TIOCSTI",
504         "TIOCGWINSZ", "TIOCSWINSZ", "TIOCMGET", "TIOCMBIS", "TIOCMBIC",
505         "TIOCMSET", "TIOCGSOFTCAR", "TIOCSSOFTCAR", "FIONREAD", "TIOCLINUX",
506         "TIOCCONS", "TIOCGSERIAL", "TIOCSSERIAL", "TIOCPKT", "FIONBIO",
507         "TIOCNOTTY", "TIOCSETD", "TIOCGETD", "TCSBRKP", [0x27] = "TIOCSBRK",
508         "TIOCCBRK", "TIOCGSID", "TCGETS2", "TCSETS2", "TCSETSW2", "TCSETSF2",
509         "TIOCGRS485", "TIOCSRS485", "TIOCGPTN", "TIOCSPTLCK",
510         "TIOCGDEV||TCGETX", "TCSETX", "TCSETXF", "TCSETXW", "TIOCSIG",
511         "TIOCVHANGUP", "TIOCGPKT", "TIOCGPTLCK", "TIOCGEXCL",
512         [0x50] = "FIONCLEX", "FIOCLEX", "FIOASYNC", "TIOCSERCONFIG",
513         "TIOCSERGWILD", "TIOCSERSWILD", "TIOCGLCKTRMIOS", "TIOCSLCKTRMIOS",
514         "TIOCSERGSTRUCT", "TIOCSERGETLSR", "TIOCSERGETMULTI", "TIOCSERSETMULTI",
515         "TIOCMIWAIT", "TIOCGICOUNT", [0x60] = "FIOQSIZE",
516 };
517
518 static DEFINE_STRARRAY_OFFSET(tioctls, 0x5401);
519 #endif /* defined(__i386__) || defined(__x86_64__) */
520
521 #ifndef GRND_NONBLOCK
522 #define GRND_NONBLOCK   0x0001
523 #endif
524 #ifndef GRND_RANDOM
525 #define GRND_RANDOM     0x0002
526 #endif
527
528 static size_t syscall_arg__scnprintf_getrandom_flags(char *bf, size_t size,
529                                                    struct syscall_arg *arg)
530 {
531         int printed = 0, flags = arg->val;
532
533 #define P_FLAG(n) \
534         if (flags & GRND_##n) { \
535                 printed += scnprintf(bf + printed, size - printed, "%s%s", printed ? "|" : "", #n); \
536                 flags &= ~GRND_##n; \
537         }
538
539         P_FLAG(RANDOM);
540         P_FLAG(NONBLOCK);
541 #undef P_FLAG
542
543         if (flags)
544                 printed += scnprintf(bf + printed, size - printed, "%s%#x", printed ? "|" : "", flags);
545
546         return printed;
547 }
548
549 #define SCA_GETRANDOM_FLAGS syscall_arg__scnprintf_getrandom_flags
550
551 #define STRARRAY(arg, name, array) \
552           .arg_scnprintf = { [arg] = SCA_STRARRAY, }, \
553           .arg_parm      = { [arg] = &strarray__##array, }
554
555 #include "trace/beauty/eventfd.c"
556 #include "trace/beauty/flock.c"
557 #include "trace/beauty/futex_op.c"
558 #include "trace/beauty/mmap.c"
559 #include "trace/beauty/mode_t.c"
560 #include "trace/beauty/msg_flags.c"
561 #include "trace/beauty/open_flags.c"
562 #include "trace/beauty/perf_event_open.c"
563 #include "trace/beauty/pid.c"
564 #include "trace/beauty/sched_policy.c"
565 #include "trace/beauty/seccomp.c"
566 #include "trace/beauty/signum.c"
567 #include "trace/beauty/socket_type.c"
568 #include "trace/beauty/waitid_options.c"
569
570 static struct syscall_fmt {
571         const char *name;
572         const char *alias;
573         size_t     (*arg_scnprintf[6])(char *bf, size_t size, struct syscall_arg *arg);
574         void       *arg_parm[6];
575         bool       errmsg;
576         bool       errpid;
577         bool       timeout;
578         bool       hexret;
579 } syscall_fmts[] = {
580         { .name     = "access",     .errmsg = true,
581           .arg_scnprintf = { [1] = SCA_ACCMODE,  /* mode */ }, },
582         { .name     = "arch_prctl", .errmsg = true, .alias = "prctl", },
583         { .name     = "bpf",        .errmsg = true, STRARRAY(0, cmd, bpf_cmd), },
584         { .name     = "brk",        .hexret = true,
585           .arg_scnprintf = { [0] = SCA_HEX, /* brk */ }, },
586         { .name     = "chdir",      .errmsg = true, },
587         { .name     = "chmod",      .errmsg = true, },
588         { .name     = "chroot",     .errmsg = true, },
589         { .name     = "clock_gettime",  .errmsg = true, STRARRAY(0, clk_id, clockid), },
590         { .name     = "clone",      .errpid = true, },
591         { .name     = "close",      .errmsg = true,
592           .arg_scnprintf = { [0] = SCA_CLOSE_FD, /* fd */ }, },
593         { .name     = "connect",    .errmsg = true, },
594         { .name     = "creat",      .errmsg = true, },
595         { .name     = "dup",        .errmsg = true, },
596         { .name     = "dup2",       .errmsg = true, },
597         { .name     = "dup3",       .errmsg = true, },
598         { .name     = "epoll_ctl",  .errmsg = true, STRARRAY(1, op, epoll_ctl_ops), },
599         { .name     = "eventfd2",   .errmsg = true,
600           .arg_scnprintf = { [1] = SCA_EFD_FLAGS, /* flags */ }, },
601         { .name     = "faccessat",  .errmsg = true, },
602         { .name     = "fadvise64",  .errmsg = true, },
603         { .name     = "fallocate",  .errmsg = true, },
604         { .name     = "fchdir",     .errmsg = true, },
605         { .name     = "fchmod",     .errmsg = true, },
606         { .name     = "fchmodat",   .errmsg = true,
607           .arg_scnprintf = { [0] = SCA_FDAT, /* fd */ }, },
608         { .name     = "fchown",     .errmsg = true, },
609         { .name     = "fchownat",   .errmsg = true,
610           .arg_scnprintf = { [0] = SCA_FDAT, /* fd */ }, },
611         { .name     = "fcntl",      .errmsg = true,
612           .arg_scnprintf = { [1] = SCA_STRARRAY, /* cmd */ },
613           .arg_parm      = { [1] = &strarray__fcntl_cmds, /* cmd */ }, },
614         { .name     = "fdatasync",  .errmsg = true, },
615         { .name     = "flock",      .errmsg = true,
616           .arg_scnprintf = { [1] = SCA_FLOCK, /* cmd */ }, },
617         { .name     = "fsetxattr",  .errmsg = true, },
618         { .name     = "fstat",      .errmsg = true, .alias = "newfstat", },
619         { .name     = "fstatat",    .errmsg = true, .alias = "newfstatat", },
620         { .name     = "fstatfs",    .errmsg = true, },
621         { .name     = "fsync",    .errmsg = true, },
622         { .name     = "ftruncate", .errmsg = true, },
623         { .name     = "futex",      .errmsg = true,
624           .arg_scnprintf = { [1] = SCA_FUTEX_OP, /* op */ }, },
625         { .name     = "futimesat", .errmsg = true,
626           .arg_scnprintf = { [0] = SCA_FDAT, /* fd */ }, },
627         { .name     = "getdents",   .errmsg = true, },
628         { .name     = "getdents64", .errmsg = true, },
629         { .name     = "getitimer",  .errmsg = true, STRARRAY(0, which, itimers), },
630         { .name     = "getpid",     .errpid = true, },
631         { .name     = "getpgid",    .errpid = true, },
632         { .name     = "getppid",    .errpid = true, },
633         { .name     = "getrandom",  .errmsg = true,
634           .arg_scnprintf = { [2] = SCA_GETRANDOM_FLAGS, /* flags */ }, },
635         { .name     = "getrlimit",  .errmsg = true, STRARRAY(0, resource, rlimit_resources), },
636         { .name     = "getxattr",   .errmsg = true, },
637         { .name     = "inotify_add_watch",          .errmsg = true, },
638         { .name     = "ioctl",      .errmsg = true,
639           .arg_scnprintf = {
640 #if defined(__i386__) || defined(__x86_64__)
641 /*
642  * FIXME: Make this available to all arches.
643  */
644                              [1] = SCA_STRHEXARRAY, /* cmd */
645                              [2] = SCA_HEX, /* arg */ },
646           .arg_parm      = { [1] = &strarray__tioctls, /* cmd */ }, },
647 #else
648                              [2] = SCA_HEX, /* arg */ }, },
649 #endif
650         { .name     = "keyctl",     .errmsg = true, STRARRAY(0, option, keyctl_options), },
651         { .name     = "kill",       .errmsg = true,
652           .arg_scnprintf = { [1] = SCA_SIGNUM, /* sig */ }, },
653         { .name     = "lchown",    .errmsg = true, },
654         { .name     = "lgetxattr",  .errmsg = true, },
655         { .name     = "linkat",     .errmsg = true,
656           .arg_scnprintf = { [0] = SCA_FDAT, /* fd */ }, },
657         { .name     = "listxattr",  .errmsg = true, },
658         { .name     = "llistxattr", .errmsg = true, },
659         { .name     = "lremovexattr",  .errmsg = true, },
660         { .name     = "lseek",      .errmsg = true,
661           .arg_scnprintf = { [2] = SCA_STRARRAY, /* whence */ },
662           .arg_parm      = { [2] = &strarray__whences, /* whence */ }, },
663         { .name     = "lsetxattr",  .errmsg = true, },
664         { .name     = "lstat",      .errmsg = true, .alias = "newlstat", },
665         { .name     = "lsxattr",    .errmsg = true, },
666         { .name     = "madvise",    .errmsg = true,
667           .arg_scnprintf = { [0] = SCA_HEX,      /* start */
668                              [2] = SCA_MADV_BHV, /* behavior */ }, },
669         { .name     = "mkdir",    .errmsg = true, },
670         { .name     = "mkdirat",    .errmsg = true,
671           .arg_scnprintf = { [0] = SCA_FDAT, /* fd */ }, },
672         { .name     = "mknod",      .errmsg = true, },
673         { .name     = "mknodat",    .errmsg = true,
674           .arg_scnprintf = { [0] = SCA_FDAT, /* fd */ }, },
675         { .name     = "mlock",      .errmsg = true,
676           .arg_scnprintf = { [0] = SCA_HEX, /* addr */ }, },
677         { .name     = "mlockall",   .errmsg = true,
678           .arg_scnprintf = { [0] = SCA_HEX, /* addr */ }, },
679         { .name     = "mmap",       .hexret = true,
680           .arg_scnprintf = { [0] = SCA_HEX,       /* addr */
681                              [2] = SCA_MMAP_PROT, /* prot */
682                              [3] = SCA_MMAP_FLAGS, /* flags */ }, },
683         { .name     = "mprotect",   .errmsg = true,
684           .arg_scnprintf = { [0] = SCA_HEX, /* start */
685                              [2] = SCA_MMAP_PROT, /* prot */ }, },
686         { .name     = "mq_unlink", .errmsg = true,
687           .arg_scnprintf = { [0] = SCA_FILENAME, /* u_name */ }, },
688         { .name     = "mremap",     .hexret = true,
689           .arg_scnprintf = { [0] = SCA_HEX, /* addr */
690                              [3] = SCA_MREMAP_FLAGS, /* flags */
691                              [4] = SCA_HEX, /* new_addr */ }, },
692         { .name     = "munlock",    .errmsg = true,
693           .arg_scnprintf = { [0] = SCA_HEX, /* addr */ }, },
694         { .name     = "munmap",     .errmsg = true,
695           .arg_scnprintf = { [0] = SCA_HEX, /* addr */ }, },
696         { .name     = "name_to_handle_at", .errmsg = true,
697           .arg_scnprintf = { [0] = SCA_FDAT, /* dfd */ }, },
698         { .name     = "newfstatat", .errmsg = true,
699           .arg_scnprintf = { [0] = SCA_FDAT, /* dfd */ }, },
700         { .name     = "open",       .errmsg = true,
701           .arg_scnprintf = { [1] = SCA_OPEN_FLAGS, /* flags */ }, },
702         { .name     = "open_by_handle_at", .errmsg = true,
703           .arg_scnprintf = { [0] = SCA_FDAT, /* dfd */
704                              [2] = SCA_OPEN_FLAGS, /* flags */ }, },
705         { .name     = "openat",     .errmsg = true,
706           .arg_scnprintf = { [0] = SCA_FDAT, /* dfd */
707                              [2] = SCA_OPEN_FLAGS, /* flags */ }, },
708         { .name     = "perf_event_open", .errmsg = true,
709           .arg_scnprintf = { [2] = SCA_INT, /* cpu */
710                              [3] = SCA_FD,  /* group_fd */
711                              [4] = SCA_PERF_FLAGS,  /* flags */ }, },
712         { .name     = "pipe2",      .errmsg = true,
713           .arg_scnprintf = { [1] = SCA_PIPE_FLAGS, /* flags */ }, },
714         { .name     = "poll",       .errmsg = true, .timeout = true, },
715         { .name     = "ppoll",      .errmsg = true, .timeout = true, },
716         { .name     = "pread",      .errmsg = true, .alias = "pread64", },
717         { .name     = "preadv",     .errmsg = true, .alias = "pread", },
718         { .name     = "prlimit64",  .errmsg = true, STRARRAY(1, resource, rlimit_resources), },
719         { .name     = "pwrite",     .errmsg = true, .alias = "pwrite64", },
720         { .name     = "pwritev",    .errmsg = true, },
721         { .name     = "read",       .errmsg = true, },
722         { .name     = "readlink",   .errmsg = true, },
723         { .name     = "readlinkat", .errmsg = true,
724           .arg_scnprintf = { [0] = SCA_FDAT, /* dfd */ }, },
725         { .name     = "readv",      .errmsg = true, },
726         { .name     = "recvfrom",   .errmsg = true,
727           .arg_scnprintf = { [3] = SCA_MSG_FLAGS, /* flags */ }, },
728         { .name     = "recvmmsg",   .errmsg = true,
729           .arg_scnprintf = { [3] = SCA_MSG_FLAGS, /* flags */ }, },
730         { .name     = "recvmsg",    .errmsg = true,
731           .arg_scnprintf = { [2] = SCA_MSG_FLAGS, /* flags */ }, },
732         { .name     = "removexattr", .errmsg = true, },
733         { .name     = "renameat",   .errmsg = true,
734           .arg_scnprintf = { [0] = SCA_FDAT, /* dfd */ }, },
735         { .name     = "rmdir",    .errmsg = true, },
736         { .name     = "rt_sigaction", .errmsg = true,
737           .arg_scnprintf = { [0] = SCA_SIGNUM, /* sig */ }, },
738         { .name     = "rt_sigprocmask",  .errmsg = true, STRARRAY(0, how, sighow), },
739         { .name     = "rt_sigqueueinfo", .errmsg = true,
740           .arg_scnprintf = { [1] = SCA_SIGNUM, /* sig */ }, },
741         { .name     = "rt_tgsigqueueinfo", .errmsg = true,
742           .arg_scnprintf = { [2] = SCA_SIGNUM, /* sig */ }, },
743         { .name     = "sched_getattr",        .errmsg = true, },
744         { .name     = "sched_setattr",        .errmsg = true, },
745         { .name     = "sched_setscheduler",   .errmsg = true,
746           .arg_scnprintf = { [1] = SCA_SCHED_POLICY, /* policy */ }, },
747         { .name     = "seccomp", .errmsg = true,
748           .arg_scnprintf = { [0] = SCA_SECCOMP_OP, /* op */
749                              [1] = SCA_SECCOMP_FLAGS, /* flags */ }, },
750         { .name     = "select",     .errmsg = true, .timeout = true, },
751         { .name     = "sendmmsg",    .errmsg = true,
752           .arg_scnprintf = { [3] = SCA_MSG_FLAGS, /* flags */ }, },
753         { .name     = "sendmsg",    .errmsg = true,
754           .arg_scnprintf = { [2] = SCA_MSG_FLAGS, /* flags */ }, },
755         { .name     = "sendto",     .errmsg = true,
756           .arg_scnprintf = { [3] = SCA_MSG_FLAGS, /* flags */ }, },
757         { .name     = "set_tid_address", .errpid = true, },
758         { .name     = "setitimer",  .errmsg = true, STRARRAY(0, which, itimers), },
759         { .name     = "setpgid",    .errmsg = true, },
760         { .name     = "setrlimit",  .errmsg = true, STRARRAY(0, resource, rlimit_resources), },
761         { .name     = "setxattr",   .errmsg = true, },
762         { .name     = "shutdown",   .errmsg = true, },
763         { .name     = "socket",     .errmsg = true,
764           .arg_scnprintf = { [0] = SCA_STRARRAY, /* family */
765                              [1] = SCA_SK_TYPE, /* type */ },
766           .arg_parm      = { [0] = &strarray__socket_families, /* family */ }, },
767         { .name     = "socketpair", .errmsg = true,
768           .arg_scnprintf = { [0] = SCA_STRARRAY, /* family */
769                              [1] = SCA_SK_TYPE, /* type */ },
770           .arg_parm      = { [0] = &strarray__socket_families, /* family */ }, },
771         { .name     = "stat",       .errmsg = true, .alias = "newstat", },
772         { .name     = "statfs",     .errmsg = true, },
773         { .name     = "swapoff",    .errmsg = true,
774           .arg_scnprintf = { [0] = SCA_FILENAME, /* specialfile */ }, },
775         { .name     = "swapon",     .errmsg = true,
776           .arg_scnprintf = { [0] = SCA_FILENAME, /* specialfile */ }, },
777         { .name     = "symlinkat",  .errmsg = true,
778           .arg_scnprintf = { [0] = SCA_FDAT, /* dfd */ }, },
779         { .name     = "tgkill",     .errmsg = true,
780           .arg_scnprintf = { [2] = SCA_SIGNUM, /* sig */ }, },
781         { .name     = "tkill",      .errmsg = true,
782           .arg_scnprintf = { [1] = SCA_SIGNUM, /* sig */ }, },
783         { .name     = "truncate",   .errmsg = true, },
784         { .name     = "uname",      .errmsg = true, .alias = "newuname", },
785         { .name     = "unlinkat",   .errmsg = true,
786           .arg_scnprintf = { [0] = SCA_FDAT, /* dfd */ }, },
787         { .name     = "utime",  .errmsg = true, },
788         { .name     = "utimensat",  .errmsg = true,
789           .arg_scnprintf = { [0] = SCA_FDAT, /* dirfd */ }, },
790         { .name     = "utimes",  .errmsg = true, },
791         { .name     = "vmsplice",  .errmsg = true, },
792         { .name     = "wait4",      .errpid = true,
793           .arg_scnprintf = { [2] = SCA_WAITID_OPTIONS, /* options */ }, },
794         { .name     = "waitid",     .errpid = true,
795           .arg_scnprintf = { [3] = SCA_WAITID_OPTIONS, /* options */ }, },
796         { .name     = "write",      .errmsg = true, },
797         { .name     = "writev",     .errmsg = true, },
798 };
799
800 static int syscall_fmt__cmp(const void *name, const void *fmtp)
801 {
802         const struct syscall_fmt *fmt = fmtp;
803         return strcmp(name, fmt->name);
804 }
805
806 static struct syscall_fmt *syscall_fmt__find(const char *name)
807 {
808         const int nmemb = ARRAY_SIZE(syscall_fmts);
809         return bsearch(name, syscall_fmts, nmemb, sizeof(struct syscall_fmt), syscall_fmt__cmp);
810 }
811
812 struct syscall {
813         struct event_format *tp_format;
814         int                 nr_args;
815         struct format_field *args;
816         const char          *name;
817         bool                is_exit;
818         struct syscall_fmt  *fmt;
819         size_t              (**arg_scnprintf)(char *bf, size_t size, struct syscall_arg *arg);
820         void                **arg_parm;
821 };
822
823 static size_t fprintf_duration(unsigned long t, FILE *fp)
824 {
825         double duration = (double)t / NSEC_PER_MSEC;
826         size_t printed = fprintf(fp, "(");
827
828         if (duration >= 1.0)
829                 printed += color_fprintf(fp, PERF_COLOR_RED, "%6.3f ms", duration);
830         else if (duration >= 0.01)
831                 printed += color_fprintf(fp, PERF_COLOR_YELLOW, "%6.3f ms", duration);
832         else
833                 printed += color_fprintf(fp, PERF_COLOR_NORMAL, "%6.3f ms", duration);
834         return printed + fprintf(fp, "): ");
835 }
836
837 /**
838  * filename.ptr: The filename char pointer that will be vfs_getname'd
839  * filename.entry_str_pos: Where to insert the string translated from
840  *                         filename.ptr by the vfs_getname tracepoint/kprobe.
841  */
842 struct thread_trace {
843         u64               entry_time;
844         bool              entry_pending;
845         unsigned long     nr_events;
846         unsigned long     pfmaj, pfmin;
847         char              *entry_str;
848         double            runtime_ms;
849         struct {
850                 unsigned long ptr;
851                 short int     entry_str_pos;
852                 bool          pending_open;
853                 unsigned int  namelen;
854                 char          *name;
855         } filename;
856         struct {
857                 int       max;
858                 char      **table;
859         } paths;
860
861         struct intlist *syscall_stats;
862 };
863
864 static struct thread_trace *thread_trace__new(void)
865 {
866         struct thread_trace *ttrace =  zalloc(sizeof(struct thread_trace));
867
868         if (ttrace)
869                 ttrace->paths.max = -1;
870
871         ttrace->syscall_stats = intlist__new(NULL);
872
873         return ttrace;
874 }
875
876 static struct thread_trace *thread__trace(struct thread *thread, FILE *fp)
877 {
878         struct thread_trace *ttrace;
879
880         if (thread == NULL)
881                 goto fail;
882
883         if (thread__priv(thread) == NULL)
884                 thread__set_priv(thread, thread_trace__new());
885
886         if (thread__priv(thread) == NULL)
887                 goto fail;
888
889         ttrace = thread__priv(thread);
890         ++ttrace->nr_events;
891
892         return ttrace;
893 fail:
894         color_fprintf(fp, PERF_COLOR_RED,
895                       "WARNING: not enough memory, dropping samples!\n");
896         return NULL;
897 }
898
899 #define TRACE_PFMAJ             (1 << 0)
900 #define TRACE_PFMIN             (1 << 1)
901
902 static const size_t trace__entry_str_size = 2048;
903
904 static int trace__set_fd_pathname(struct thread *thread, int fd, const char *pathname)
905 {
906         struct thread_trace *ttrace = thread__priv(thread);
907
908         if (fd > ttrace->paths.max) {
909                 char **npath = realloc(ttrace->paths.table, (fd + 1) * sizeof(char *));
910
911                 if (npath == NULL)
912                         return -1;
913
914                 if (ttrace->paths.max != -1) {
915                         memset(npath + ttrace->paths.max + 1, 0,
916                                (fd - ttrace->paths.max) * sizeof(char *));
917                 } else {
918                         memset(npath, 0, (fd + 1) * sizeof(char *));
919                 }
920
921                 ttrace->paths.table = npath;
922                 ttrace->paths.max   = fd;
923         }
924
925         ttrace->paths.table[fd] = strdup(pathname);
926
927         return ttrace->paths.table[fd] != NULL ? 0 : -1;
928 }
929
930 static int thread__read_fd_path(struct thread *thread, int fd)
931 {
932         char linkname[PATH_MAX], pathname[PATH_MAX];
933         struct stat st;
934         int ret;
935
936         if (thread->pid_ == thread->tid) {
937                 scnprintf(linkname, sizeof(linkname),
938                           "/proc/%d/fd/%d", thread->pid_, fd);
939         } else {
940                 scnprintf(linkname, sizeof(linkname),
941                           "/proc/%d/task/%d/fd/%d", thread->pid_, thread->tid, fd);
942         }
943
944         if (lstat(linkname, &st) < 0 || st.st_size + 1 > (off_t)sizeof(pathname))
945                 return -1;
946
947         ret = readlink(linkname, pathname, sizeof(pathname));
948
949         if (ret < 0 || ret > st.st_size)
950                 return -1;
951
952         pathname[ret] = '\0';
953         return trace__set_fd_pathname(thread, fd, pathname);
954 }
955
956 static const char *thread__fd_path(struct thread *thread, int fd,
957                                    struct trace *trace)
958 {
959         struct thread_trace *ttrace = thread__priv(thread);
960
961         if (ttrace == NULL)
962                 return NULL;
963
964         if (fd < 0)
965                 return NULL;
966
967         if ((fd > ttrace->paths.max || ttrace->paths.table[fd] == NULL)) {
968                 if (!trace->live)
969                         return NULL;
970                 ++trace->stats.proc_getname;
971                 if (thread__read_fd_path(thread, fd))
972                         return NULL;
973         }
974
975         return ttrace->paths.table[fd];
976 }
977
978 static size_t syscall_arg__scnprintf_fd(char *bf, size_t size,
979                                         struct syscall_arg *arg)
980 {
981         int fd = arg->val;
982         size_t printed = scnprintf(bf, size, "%d", fd);
983         const char *path = thread__fd_path(arg->thread, fd, arg->trace);
984
985         if (path)
986                 printed += scnprintf(bf + printed, size - printed, "<%s>", path);
987
988         return printed;
989 }
990
991 static size_t syscall_arg__scnprintf_close_fd(char *bf, size_t size,
992                                               struct syscall_arg *arg)
993 {
994         int fd = arg->val;
995         size_t printed = syscall_arg__scnprintf_fd(bf, size, arg);
996         struct thread_trace *ttrace = thread__priv(arg->thread);
997
998         if (ttrace && fd >= 0 && fd <= ttrace->paths.max)
999                 zfree(&ttrace->paths.table[fd]);
1000
1001         return printed;
1002 }
1003
1004 static void thread__set_filename_pos(struct thread *thread, const char *bf,
1005                                      unsigned long ptr)
1006 {
1007         struct thread_trace *ttrace = thread__priv(thread);
1008
1009         ttrace->filename.ptr = ptr;
1010         ttrace->filename.entry_str_pos = bf - ttrace->entry_str;
1011 }
1012
1013 static size_t syscall_arg__scnprintf_filename(char *bf, size_t size,
1014                                               struct syscall_arg *arg)
1015 {
1016         unsigned long ptr = arg->val;
1017
1018         if (!arg->trace->vfs_getname)
1019                 return scnprintf(bf, size, "%#x", ptr);
1020
1021         thread__set_filename_pos(arg->thread, bf, ptr);
1022         return 0;
1023 }
1024
1025 static bool trace__filter_duration(struct trace *trace, double t)
1026 {
1027         return t < (trace->duration_filter * NSEC_PER_MSEC);
1028 }
1029
1030 static size_t trace__fprintf_tstamp(struct trace *trace, u64 tstamp, FILE *fp)
1031 {
1032         double ts = (double)(tstamp - trace->base_time) / NSEC_PER_MSEC;
1033
1034         return fprintf(fp, "%10.3f ", ts);
1035 }
1036
1037 static bool done = false;
1038 static bool interrupted = false;
1039
1040 static void sig_handler(int sig)
1041 {
1042         done = true;
1043         interrupted = sig == SIGINT;
1044 }
1045
1046 static size_t trace__fprintf_entry_head(struct trace *trace, struct thread *thread,
1047                                         u64 duration, u64 tstamp, FILE *fp)
1048 {
1049         size_t printed = trace__fprintf_tstamp(trace, tstamp, fp);
1050         printed += fprintf_duration(duration, fp);
1051
1052         if (trace->multiple_threads) {
1053                 if (trace->show_comm)
1054                         printed += fprintf(fp, "%.14s/", thread__comm_str(thread));
1055                 printed += fprintf(fp, "%d ", thread->tid);
1056         }
1057
1058         return printed;
1059 }
1060
1061 static int trace__process_event(struct trace *trace, struct machine *machine,
1062                                 union perf_event *event, struct perf_sample *sample)
1063 {
1064         int ret = 0;
1065
1066         switch (event->header.type) {
1067         case PERF_RECORD_LOST:
1068                 color_fprintf(trace->output, PERF_COLOR_RED,
1069                               "LOST %" PRIu64 " events!\n", event->lost.lost);
1070                 ret = machine__process_lost_event(machine, event, sample);
1071                 break;
1072         default:
1073                 ret = machine__process_event(machine, event, sample);
1074                 break;
1075         }
1076
1077         return ret;
1078 }
1079
1080 static int trace__tool_process(struct perf_tool *tool,
1081                                union perf_event *event,
1082                                struct perf_sample *sample,
1083                                struct machine *machine)
1084 {
1085         struct trace *trace = container_of(tool, struct trace, tool);
1086         return trace__process_event(trace, machine, event, sample);
1087 }
1088
1089 static char *trace__machine__resolve_kernel_addr(void *vmachine, unsigned long long *addrp, char **modp)
1090 {
1091         struct machine *machine = vmachine;
1092
1093         if (machine->kptr_restrict_warned)
1094                 return NULL;
1095
1096         if (symbol_conf.kptr_restrict) {
1097                 pr_warning("Kernel address maps (/proc/{kallsyms,modules}) are restricted.\n\n"
1098                            "Check /proc/sys/kernel/kptr_restrict.\n\n"
1099                            "Kernel samples will not be resolved.\n");
1100                 machine->kptr_restrict_warned = true;
1101                 return NULL;
1102         }
1103
1104         return machine__resolve_kernel_addr(vmachine, addrp, modp);
1105 }
1106
1107 static int trace__symbols_init(struct trace *trace, struct perf_evlist *evlist)
1108 {
1109         int err = symbol__init(NULL);
1110
1111         if (err)
1112                 return err;
1113
1114         trace->host = machine__new_host();
1115         if (trace->host == NULL)
1116                 return -ENOMEM;
1117
1118         if (trace_event__register_resolver(trace->host, trace__machine__resolve_kernel_addr) < 0)
1119                 return -errno;
1120
1121         err = __machine__synthesize_threads(trace->host, &trace->tool, &trace->opts.target,
1122                                             evlist->threads, trace__tool_process, false,
1123                                             trace->opts.proc_map_timeout);
1124         if (err)
1125                 symbol__exit();
1126
1127         return err;
1128 }
1129
1130 static int syscall__set_arg_fmts(struct syscall *sc)
1131 {
1132         struct format_field *field;
1133         int idx = 0, len;
1134
1135         sc->arg_scnprintf = calloc(sc->nr_args, sizeof(void *));
1136         if (sc->arg_scnprintf == NULL)
1137                 return -1;
1138
1139         if (sc->fmt)
1140                 sc->arg_parm = sc->fmt->arg_parm;
1141
1142         for (field = sc->args; field; field = field->next) {
1143                 if (sc->fmt && sc->fmt->arg_scnprintf[idx])
1144                         sc->arg_scnprintf[idx] = sc->fmt->arg_scnprintf[idx];
1145                 else if (strcmp(field->type, "const char *") == 0 &&
1146                          (strcmp(field->name, "filename") == 0 ||
1147                           strcmp(field->name, "path") == 0 ||
1148                           strcmp(field->name, "pathname") == 0))
1149                         sc->arg_scnprintf[idx] = SCA_FILENAME;
1150                 else if (field->flags & FIELD_IS_POINTER)
1151                         sc->arg_scnprintf[idx] = syscall_arg__scnprintf_hex;
1152                 else if (strcmp(field->type, "pid_t") == 0)
1153                         sc->arg_scnprintf[idx] = SCA_PID;
1154                 else if (strcmp(field->type, "umode_t") == 0)
1155                         sc->arg_scnprintf[idx] = SCA_MODE_T;
1156                 else if ((strcmp(field->type, "int") == 0 ||
1157                           strcmp(field->type, "unsigned int") == 0 ||
1158                           strcmp(field->type, "long") == 0) &&
1159                          (len = strlen(field->name)) >= 2 &&
1160                          strcmp(field->name + len - 2, "fd") == 0) {
1161                         /*
1162                          * /sys/kernel/tracing/events/syscalls/sys_enter*
1163                          * egrep 'field:.*fd;' .../format|sed -r 's/.*field:([a-z ]+) [a-z_]*fd.+/\1/g'|sort|uniq -c
1164                          * 65 int
1165                          * 23 unsigned int
1166                          * 7 unsigned long
1167                          */
1168                         sc->arg_scnprintf[idx] = SCA_FD;
1169                 }
1170                 ++idx;
1171         }
1172
1173         return 0;
1174 }
1175
1176 static int trace__read_syscall_info(struct trace *trace, int id)
1177 {
1178         char tp_name[128];
1179         struct syscall *sc;
1180         const char *name = syscalltbl__name(trace->sctbl, id);
1181
1182         if (name == NULL)
1183                 return -1;
1184
1185         if (id > trace->syscalls.max) {
1186                 struct syscall *nsyscalls = realloc(trace->syscalls.table, (id + 1) * sizeof(*sc));
1187
1188                 if (nsyscalls == NULL)
1189                         return -1;
1190
1191                 if (trace->syscalls.max != -1) {
1192                         memset(nsyscalls + trace->syscalls.max + 1, 0,
1193                                (id - trace->syscalls.max) * sizeof(*sc));
1194                 } else {
1195                         memset(nsyscalls, 0, (id + 1) * sizeof(*sc));
1196                 }
1197
1198                 trace->syscalls.table = nsyscalls;
1199                 trace->syscalls.max   = id;
1200         }
1201
1202         sc = trace->syscalls.table + id;
1203         sc->name = name;
1204
1205         sc->fmt  = syscall_fmt__find(sc->name);
1206
1207         snprintf(tp_name, sizeof(tp_name), "sys_enter_%s", sc->name);
1208         sc->tp_format = trace_event__tp_format("syscalls", tp_name);
1209
1210         if (IS_ERR(sc->tp_format) && sc->fmt && sc->fmt->alias) {
1211                 snprintf(tp_name, sizeof(tp_name), "sys_enter_%s", sc->fmt->alias);
1212                 sc->tp_format = trace_event__tp_format("syscalls", tp_name);
1213         }
1214
1215         if (IS_ERR(sc->tp_format))
1216                 return -1;
1217
1218         sc->args = sc->tp_format->format.fields;
1219         sc->nr_args = sc->tp_format->format.nr_fields;
1220         /*
1221          * We need to check and discard the first variable '__syscall_nr'
1222          * or 'nr' that mean the syscall number. It is needless here.
1223          * So drop '__syscall_nr' or 'nr' field but does not exist on older kernels.
1224          */
1225         if (sc->args && (!strcmp(sc->args->name, "__syscall_nr") || !strcmp(sc->args->name, "nr"))) {
1226                 sc->args = sc->args->next;
1227                 --sc->nr_args;
1228         }
1229
1230         sc->is_exit = !strcmp(name, "exit_group") || !strcmp(name, "exit");
1231
1232         return syscall__set_arg_fmts(sc);
1233 }
1234
1235 static int trace__validate_ev_qualifier(struct trace *trace)
1236 {
1237         int err = 0, i;
1238         struct str_node *pos;
1239
1240         trace->ev_qualifier_ids.nr = strlist__nr_entries(trace->ev_qualifier);
1241         trace->ev_qualifier_ids.entries = malloc(trace->ev_qualifier_ids.nr *
1242                                                  sizeof(trace->ev_qualifier_ids.entries[0]));
1243
1244         if (trace->ev_qualifier_ids.entries == NULL) {
1245                 fputs("Error:\tNot enough memory for allocating events qualifier ids\n",
1246                        trace->output);
1247                 err = -EINVAL;
1248                 goto out;
1249         }
1250
1251         i = 0;
1252
1253         strlist__for_each_entry(pos, trace->ev_qualifier) {
1254                 const char *sc = pos->s;
1255                 int id = syscalltbl__id(trace->sctbl, sc);
1256
1257                 if (id < 0) {
1258                         if (err == 0) {
1259                                 fputs("Error:\tInvalid syscall ", trace->output);
1260                                 err = -EINVAL;
1261                         } else {
1262                                 fputs(", ", trace->output);
1263                         }
1264
1265                         fputs(sc, trace->output);
1266                 }
1267
1268                 trace->ev_qualifier_ids.entries[i++] = id;
1269         }
1270
1271         if (err < 0) {
1272                 fputs("\nHint:\ttry 'perf list syscalls:sys_enter_*'"
1273                       "\nHint:\tand: 'man syscalls'\n", trace->output);
1274                 zfree(&trace->ev_qualifier_ids.entries);
1275                 trace->ev_qualifier_ids.nr = 0;
1276         }
1277 out:
1278         return err;
1279 }
1280
1281 /*
1282  * args is to be interpreted as a series of longs but we need to handle
1283  * 8-byte unaligned accesses. args points to raw_data within the event
1284  * and raw_data is guaranteed to be 8-byte unaligned because it is
1285  * preceded by raw_size which is a u32. So we need to copy args to a temp
1286  * variable to read it. Most notably this avoids extended load instructions
1287  * on unaligned addresses
1288  */
1289
1290 static size_t syscall__scnprintf_args(struct syscall *sc, char *bf, size_t size,
1291                                       unsigned char *args, struct trace *trace,
1292                                       struct thread *thread)
1293 {
1294         size_t printed = 0;
1295         unsigned char *p;
1296         unsigned long val;
1297
1298         if (sc->args != NULL) {
1299                 struct format_field *field;
1300                 u8 bit = 1;
1301                 struct syscall_arg arg = {
1302                         .idx    = 0,
1303                         .mask   = 0,
1304                         .trace  = trace,
1305                         .thread = thread,
1306                 };
1307
1308                 for (field = sc->args; field;
1309                      field = field->next, ++arg.idx, bit <<= 1) {
1310                         if (arg.mask & bit)
1311                                 continue;
1312
1313                         /* special care for unaligned accesses */
1314                         p = args + sizeof(unsigned long) * arg.idx;
1315                         memcpy(&val, p, sizeof(val));
1316
1317                         /*
1318                          * Suppress this argument if its value is zero and
1319                          * and we don't have a string associated in an
1320                          * strarray for it.
1321                          */
1322                         if (val == 0 &&
1323                             !(sc->arg_scnprintf &&
1324                               sc->arg_scnprintf[arg.idx] == SCA_STRARRAY &&
1325                               sc->arg_parm[arg.idx]))
1326                                 continue;
1327
1328                         printed += scnprintf(bf + printed, size - printed,
1329                                              "%s%s: ", printed ? ", " : "", field->name);
1330                         if (sc->arg_scnprintf && sc->arg_scnprintf[arg.idx]) {
1331                                 arg.val = val;
1332                                 if (sc->arg_parm)
1333                                         arg.parm = sc->arg_parm[arg.idx];
1334                                 printed += sc->arg_scnprintf[arg.idx](bf + printed,
1335                                                                       size - printed, &arg);
1336                         } else {
1337                                 printed += scnprintf(bf + printed, size - printed,
1338                                                      "%ld", val);
1339                         }
1340                 }
1341         } else if (IS_ERR(sc->tp_format)) {
1342                 /*
1343                  * If we managed to read the tracepoint /format file, then we
1344                  * may end up not having any args, like with gettid(), so only
1345                  * print the raw args when we didn't manage to read it.
1346                  */
1347                 int i = 0;
1348
1349                 while (i < 6) {
1350                         /* special care for unaligned accesses */
1351                         p = args + sizeof(unsigned long) * i;
1352                         memcpy(&val, p, sizeof(val));
1353                         printed += scnprintf(bf + printed, size - printed,
1354                                              "%sarg%d: %ld",
1355                                              printed ? ", " : "", i, val);
1356                         ++i;
1357                 }
1358         }
1359
1360         return printed;
1361 }
1362
1363 typedef int (*tracepoint_handler)(struct trace *trace, struct perf_evsel *evsel,
1364                                   union perf_event *event,
1365                                   struct perf_sample *sample);
1366
1367 static struct syscall *trace__syscall_info(struct trace *trace,
1368                                            struct perf_evsel *evsel, int id)
1369 {
1370
1371         if (id < 0) {
1372
1373                 /*
1374                  * XXX: Noticed on x86_64, reproduced as far back as 3.0.36, haven't tried
1375                  * before that, leaving at a higher verbosity level till that is
1376                  * explained. Reproduced with plain ftrace with:
1377                  *
1378                  * echo 1 > /t/events/raw_syscalls/sys_exit/enable
1379                  * grep "NR -1 " /t/trace_pipe
1380                  *
1381                  * After generating some load on the machine.
1382                  */
1383                 if (verbose > 1) {
1384                         static u64 n;
1385                         fprintf(trace->output, "Invalid syscall %d id, skipping (%s, %" PRIu64 ") ...\n",
1386                                 id, perf_evsel__name(evsel), ++n);
1387                 }
1388                 return NULL;
1389         }
1390
1391         if ((id > trace->syscalls.max || trace->syscalls.table[id].name == NULL) &&
1392             trace__read_syscall_info(trace, id))
1393                 goto out_cant_read;
1394
1395         if ((id > trace->syscalls.max || trace->syscalls.table[id].name == NULL))
1396                 goto out_cant_read;
1397
1398         return &trace->syscalls.table[id];
1399
1400 out_cant_read:
1401         if (verbose) {
1402                 fprintf(trace->output, "Problems reading syscall %d", id);
1403                 if (id <= trace->syscalls.max && trace->syscalls.table[id].name != NULL)
1404                         fprintf(trace->output, "(%s)", trace->syscalls.table[id].name);
1405                 fputs(" information\n", trace->output);
1406         }
1407         return NULL;
1408 }
1409
1410 static void thread__update_stats(struct thread_trace *ttrace,
1411                                  int id, struct perf_sample *sample)
1412 {
1413         struct int_node *inode;
1414         struct stats *stats;
1415         u64 duration = 0;
1416
1417         inode = intlist__findnew(ttrace->syscall_stats, id);
1418         if (inode == NULL)
1419                 return;
1420
1421         stats = inode->priv;
1422         if (stats == NULL) {
1423                 stats = malloc(sizeof(struct stats));
1424                 if (stats == NULL)
1425                         return;
1426                 init_stats(stats);
1427                 inode->priv = stats;
1428         }
1429
1430         if (ttrace->entry_time && sample->time > ttrace->entry_time)
1431                 duration = sample->time - ttrace->entry_time;
1432
1433         update_stats(stats, duration);
1434 }
1435
1436 static int trace__printf_interrupted_entry(struct trace *trace, struct perf_sample *sample)
1437 {
1438         struct thread_trace *ttrace;
1439         u64 duration;
1440         size_t printed;
1441
1442         if (trace->current == NULL)
1443                 return 0;
1444
1445         ttrace = thread__priv(trace->current);
1446
1447         if (!ttrace->entry_pending)
1448                 return 0;
1449
1450         duration = sample->time - ttrace->entry_time;
1451
1452         printed  = trace__fprintf_entry_head(trace, trace->current, duration, ttrace->entry_time, trace->output);
1453         printed += fprintf(trace->output, "%-70s) ...\n", ttrace->entry_str);
1454         ttrace->entry_pending = false;
1455
1456         return printed;
1457 }
1458
1459 static int trace__sys_enter(struct trace *trace, struct perf_evsel *evsel,
1460                             union perf_event *event __maybe_unused,
1461                             struct perf_sample *sample)
1462 {
1463         char *msg;
1464         void *args;
1465         size_t printed = 0;
1466         struct thread *thread;
1467         int id = perf_evsel__sc_tp_uint(evsel, id, sample), err = -1;
1468         struct syscall *sc = trace__syscall_info(trace, evsel, id);
1469         struct thread_trace *ttrace;
1470
1471         if (sc == NULL)
1472                 return -1;
1473
1474         thread = machine__findnew_thread(trace->host, sample->pid, sample->tid);
1475         ttrace = thread__trace(thread, trace->output);
1476         if (ttrace == NULL)
1477                 goto out_put;
1478
1479         args = perf_evsel__sc_tp_ptr(evsel, args, sample);
1480
1481         if (ttrace->entry_str == NULL) {
1482                 ttrace->entry_str = malloc(trace__entry_str_size);
1483                 if (!ttrace->entry_str)
1484                         goto out_put;
1485         }
1486
1487         if (!(trace->duration_filter || trace->summary_only || trace->min_stack))
1488                 trace__printf_interrupted_entry(trace, sample);
1489
1490         ttrace->entry_time = sample->time;
1491         msg = ttrace->entry_str;
1492         printed += scnprintf(msg + printed, trace__entry_str_size - printed, "%s(", sc->name);
1493
1494         printed += syscall__scnprintf_args(sc, msg + printed, trace__entry_str_size - printed,
1495                                            args, trace, thread);
1496
1497         if (sc->is_exit) {
1498                 if (!(trace->duration_filter || trace->summary_only || trace->min_stack)) {
1499                         trace__fprintf_entry_head(trace, thread, 1, ttrace->entry_time, trace->output);
1500                         fprintf(trace->output, "%-70s)\n", ttrace->entry_str);
1501                 }
1502         } else {
1503                 ttrace->entry_pending = true;
1504                 /* See trace__vfs_getname & trace__sys_exit */
1505                 ttrace->filename.pending_open = false;
1506         }
1507
1508         if (trace->current != thread) {
1509                 thread__put(trace->current);
1510                 trace->current = thread__get(thread);
1511         }
1512         err = 0;
1513 out_put:
1514         thread__put(thread);
1515         return err;
1516 }
1517
1518 static int trace__resolve_callchain(struct trace *trace, struct perf_evsel *evsel,
1519                                     struct perf_sample *sample,
1520                                     struct callchain_cursor *cursor)
1521 {
1522         struct addr_location al;
1523
1524         if (machine__resolve(trace->host, &al, sample) < 0 ||
1525             thread__resolve_callchain(al.thread, cursor, evsel, sample, NULL, NULL, trace->max_stack))
1526                 return -1;
1527
1528         return 0;
1529 }
1530
1531 static int trace__fprintf_callchain(struct trace *trace, struct perf_sample *sample)
1532 {
1533         /* TODO: user-configurable print_opts */
1534         const unsigned int print_opts = EVSEL__PRINT_SYM |
1535                                         EVSEL__PRINT_DSO |
1536                                         EVSEL__PRINT_UNKNOWN_AS_ADDR;
1537
1538         return sample__fprintf_callchain(sample, 38, print_opts, &callchain_cursor, trace->output);
1539 }
1540
1541 static int trace__sys_exit(struct trace *trace, struct perf_evsel *evsel,
1542                            union perf_event *event __maybe_unused,
1543                            struct perf_sample *sample)
1544 {
1545         long ret;
1546         u64 duration = 0;
1547         struct thread *thread;
1548         int id = perf_evsel__sc_tp_uint(evsel, id, sample), err = -1, callchain_ret = 0;
1549         struct syscall *sc = trace__syscall_info(trace, evsel, id);
1550         struct thread_trace *ttrace;
1551
1552         if (sc == NULL)
1553                 return -1;
1554
1555         thread = machine__findnew_thread(trace->host, sample->pid, sample->tid);
1556         ttrace = thread__trace(thread, trace->output);
1557         if (ttrace == NULL)
1558                 goto out_put;
1559
1560         if (trace->summary)
1561                 thread__update_stats(ttrace, id, sample);
1562
1563         ret = perf_evsel__sc_tp_uint(evsel, ret, sample);
1564
1565         if (id == trace->open_id && ret >= 0 && ttrace->filename.pending_open) {
1566                 trace__set_fd_pathname(thread, ret, ttrace->filename.name);
1567                 ttrace->filename.pending_open = false;
1568                 ++trace->stats.vfs_getname;
1569         }
1570
1571         if (ttrace->entry_time) {
1572                 duration = sample->time - ttrace->entry_time;
1573                 if (trace__filter_duration(trace, duration))
1574                         goto out;
1575         } else if (trace->duration_filter)
1576                 goto out;
1577
1578         if (sample->callchain) {
1579                 callchain_ret = trace__resolve_callchain(trace, evsel, sample, &callchain_cursor);
1580                 if (callchain_ret == 0) {
1581                         if (callchain_cursor.nr < trace->min_stack)
1582                                 goto out;
1583                         callchain_ret = 1;
1584                 }
1585         }
1586
1587         if (trace->summary_only)
1588                 goto out;
1589
1590         trace__fprintf_entry_head(trace, thread, duration, ttrace->entry_time, trace->output);
1591
1592         if (ttrace->entry_pending) {
1593                 fprintf(trace->output, "%-70s", ttrace->entry_str);
1594         } else {
1595                 fprintf(trace->output, " ... [");
1596                 color_fprintf(trace->output, PERF_COLOR_YELLOW, "continued");
1597                 fprintf(trace->output, "]: %s()", sc->name);
1598         }
1599
1600         if (sc->fmt == NULL) {
1601 signed_print:
1602                 fprintf(trace->output, ") = %ld", ret);
1603         } else if (ret < 0 && (sc->fmt->errmsg || sc->fmt->errpid)) {
1604                 char bf[STRERR_BUFSIZE];
1605                 const char *emsg = str_error_r(-ret, bf, sizeof(bf)),
1606                            *e = audit_errno_to_name(-ret);
1607
1608                 fprintf(trace->output, ") = -1 %s %s", e, emsg);
1609         } else if (ret == 0 && sc->fmt->timeout)
1610                 fprintf(trace->output, ") = 0 Timeout");
1611         else if (sc->fmt->hexret)
1612                 fprintf(trace->output, ") = %#lx", ret);
1613         else if (sc->fmt->errpid) {
1614                 struct thread *child = machine__find_thread(trace->host, ret, ret);
1615
1616                 if (child != NULL) {
1617                         fprintf(trace->output, ") = %ld", ret);
1618                         if (child->comm_set)
1619                                 fprintf(trace->output, " (%s)", thread__comm_str(child));
1620                         thread__put(child);
1621                 }
1622         } else
1623                 goto signed_print;
1624
1625         fputc('\n', trace->output);
1626
1627         if (callchain_ret > 0)
1628                 trace__fprintf_callchain(trace, sample);
1629         else if (callchain_ret < 0)
1630                 pr_err("Problem processing %s callchain, skipping...\n", perf_evsel__name(evsel));
1631 out:
1632         ttrace->entry_pending = false;
1633         err = 0;
1634 out_put:
1635         thread__put(thread);
1636         return err;
1637 }
1638
1639 static int trace__vfs_getname(struct trace *trace, struct perf_evsel *evsel,
1640                               union perf_event *event __maybe_unused,
1641                               struct perf_sample *sample)
1642 {
1643         struct thread *thread = machine__findnew_thread(trace->host, sample->pid, sample->tid);
1644         struct thread_trace *ttrace;
1645         size_t filename_len, entry_str_len, to_move;
1646         ssize_t remaining_space;
1647         char *pos;
1648         const char *filename = perf_evsel__rawptr(evsel, sample, "pathname");
1649
1650         if (!thread)
1651                 goto out;
1652
1653         ttrace = thread__priv(thread);
1654         if (!ttrace)
1655                 goto out;
1656
1657         filename_len = strlen(filename);
1658
1659         if (ttrace->filename.namelen < filename_len) {
1660                 char *f = realloc(ttrace->filename.name, filename_len + 1);
1661
1662                 if (f == NULL)
1663                                 goto out;
1664
1665                 ttrace->filename.namelen = filename_len;
1666                 ttrace->filename.name = f;
1667         }
1668
1669         strcpy(ttrace->filename.name, filename);
1670         ttrace->filename.pending_open = true;
1671
1672         if (!ttrace->filename.ptr)
1673                 goto out;
1674
1675         entry_str_len = strlen(ttrace->entry_str);
1676         remaining_space = trace__entry_str_size - entry_str_len - 1; /* \0 */
1677         if (remaining_space <= 0)
1678                 goto out;
1679
1680         if (filename_len > (size_t)remaining_space) {
1681                 filename += filename_len - remaining_space;
1682                 filename_len = remaining_space;
1683         }
1684
1685         to_move = entry_str_len - ttrace->filename.entry_str_pos + 1; /* \0 */
1686         pos = ttrace->entry_str + ttrace->filename.entry_str_pos;
1687         memmove(pos + filename_len, pos, to_move);
1688         memcpy(pos, filename, filename_len);
1689
1690         ttrace->filename.ptr = 0;
1691         ttrace->filename.entry_str_pos = 0;
1692 out:
1693         return 0;
1694 }
1695
1696 static int trace__sched_stat_runtime(struct trace *trace, struct perf_evsel *evsel,
1697                                      union perf_event *event __maybe_unused,
1698                                      struct perf_sample *sample)
1699 {
1700         u64 runtime = perf_evsel__intval(evsel, sample, "runtime");
1701         double runtime_ms = (double)runtime / NSEC_PER_MSEC;
1702         struct thread *thread = machine__findnew_thread(trace->host,
1703                                                         sample->pid,
1704                                                         sample->tid);
1705         struct thread_trace *ttrace = thread__trace(thread, trace->output);
1706
1707         if (ttrace == NULL)
1708                 goto out_dump;
1709
1710         ttrace->runtime_ms += runtime_ms;
1711         trace->runtime_ms += runtime_ms;
1712         thread__put(thread);
1713         return 0;
1714
1715 out_dump:
1716         fprintf(trace->output, "%s: comm=%s,pid=%u,runtime=%" PRIu64 ",vruntime=%" PRIu64 ")\n",
1717                evsel->name,
1718                perf_evsel__strval(evsel, sample, "comm"),
1719                (pid_t)perf_evsel__intval(evsel, sample, "pid"),
1720                runtime,
1721                perf_evsel__intval(evsel, sample, "vruntime"));
1722         thread__put(thread);
1723         return 0;
1724 }
1725
1726 static void bpf_output__printer(enum binary_printer_ops op,
1727                                 unsigned int val, void *extra)
1728 {
1729         FILE *output = extra;
1730         unsigned char ch = (unsigned char)val;
1731
1732         switch (op) {
1733         case BINARY_PRINT_CHAR_DATA:
1734                 fprintf(output, "%c", isprint(ch) ? ch : '.');
1735                 break;
1736         case BINARY_PRINT_DATA_BEGIN:
1737         case BINARY_PRINT_LINE_BEGIN:
1738         case BINARY_PRINT_ADDR:
1739         case BINARY_PRINT_NUM_DATA:
1740         case BINARY_PRINT_NUM_PAD:
1741         case BINARY_PRINT_SEP:
1742         case BINARY_PRINT_CHAR_PAD:
1743         case BINARY_PRINT_LINE_END:
1744         case BINARY_PRINT_DATA_END:
1745         default:
1746                 break;
1747         }
1748 }
1749
1750 static void bpf_output__fprintf(struct trace *trace,
1751                                 struct perf_sample *sample)
1752 {
1753         print_binary(sample->raw_data, sample->raw_size, 8,
1754                      bpf_output__printer, trace->output);
1755 }
1756
1757 static int trace__event_handler(struct trace *trace, struct perf_evsel *evsel,
1758                                 union perf_event *event __maybe_unused,
1759                                 struct perf_sample *sample)
1760 {
1761         int callchain_ret = 0;
1762
1763         if (sample->callchain) {
1764                 callchain_ret = trace__resolve_callchain(trace, evsel, sample, &callchain_cursor);
1765                 if (callchain_ret == 0) {
1766                         if (callchain_cursor.nr < trace->min_stack)
1767                                 goto out;
1768                         callchain_ret = 1;
1769                 }
1770         }
1771
1772         trace__printf_interrupted_entry(trace, sample);
1773         trace__fprintf_tstamp(trace, sample->time, trace->output);
1774
1775         if (trace->trace_syscalls)
1776                 fprintf(trace->output, "(         ): ");
1777
1778         fprintf(trace->output, "%s:", evsel->name);
1779
1780         if (perf_evsel__is_bpf_output(evsel)) {
1781                 bpf_output__fprintf(trace, sample);
1782         } else if (evsel->tp_format) {
1783                 event_format__fprintf(evsel->tp_format, sample->cpu,
1784                                       sample->raw_data, sample->raw_size,
1785                                       trace->output);
1786         }
1787
1788         fprintf(trace->output, ")\n");
1789
1790         if (callchain_ret > 0)
1791                 trace__fprintf_callchain(trace, sample);
1792         else if (callchain_ret < 0)
1793                 pr_err("Problem processing %s callchain, skipping...\n", perf_evsel__name(evsel));
1794 out:
1795         return 0;
1796 }
1797
1798 static void print_location(FILE *f, struct perf_sample *sample,
1799                            struct addr_location *al,
1800                            bool print_dso, bool print_sym)
1801 {
1802
1803         if ((verbose || print_dso) && al->map)
1804                 fprintf(f, "%s@", al->map->dso->long_name);
1805
1806         if ((verbose || print_sym) && al->sym)
1807                 fprintf(f, "%s+0x%" PRIx64, al->sym->name,
1808                         al->addr - al->sym->start);
1809         else if (al->map)
1810                 fprintf(f, "0x%" PRIx64, al->addr);
1811         else
1812                 fprintf(f, "0x%" PRIx64, sample->addr);
1813 }
1814
1815 static int trace__pgfault(struct trace *trace,
1816                           struct perf_evsel *evsel,
1817                           union perf_event *event __maybe_unused,
1818                           struct perf_sample *sample)
1819 {
1820         struct thread *thread;
1821         struct addr_location al;
1822         char map_type = 'd';
1823         struct thread_trace *ttrace;
1824         int err = -1;
1825         int callchain_ret = 0;
1826
1827         thread = machine__findnew_thread(trace->host, sample->pid, sample->tid);
1828
1829         if (sample->callchain) {
1830                 callchain_ret = trace__resolve_callchain(trace, evsel, sample, &callchain_cursor);
1831                 if (callchain_ret == 0) {
1832                         if (callchain_cursor.nr < trace->min_stack)
1833                                 goto out_put;
1834                         callchain_ret = 1;
1835                 }
1836         }
1837
1838         ttrace = thread__trace(thread, trace->output);
1839         if (ttrace == NULL)
1840                 goto out_put;
1841
1842         if (evsel->attr.config == PERF_COUNT_SW_PAGE_FAULTS_MAJ)
1843                 ttrace->pfmaj++;
1844         else
1845                 ttrace->pfmin++;
1846
1847         if (trace->summary_only)
1848                 goto out;
1849
1850         thread__find_addr_location(thread, sample->cpumode, MAP__FUNCTION,
1851                               sample->ip, &al);
1852
1853         trace__fprintf_entry_head(trace, thread, 0, sample->time, trace->output);
1854
1855         fprintf(trace->output, "%sfault [",
1856                 evsel->attr.config == PERF_COUNT_SW_PAGE_FAULTS_MAJ ?
1857                 "maj" : "min");
1858
1859         print_location(trace->output, sample, &al, false, true);
1860
1861         fprintf(trace->output, "] => ");
1862
1863         thread__find_addr_location(thread, sample->cpumode, MAP__VARIABLE,
1864                                    sample->addr, &al);
1865
1866         if (!al.map) {
1867                 thread__find_addr_location(thread, sample->cpumode,
1868                                            MAP__FUNCTION, sample->addr, &al);
1869
1870                 if (al.map)
1871                         map_type = 'x';
1872                 else
1873                         map_type = '?';
1874         }
1875
1876         print_location(trace->output, sample, &al, true, false);
1877
1878         fprintf(trace->output, " (%c%c)\n", map_type, al.level);
1879
1880         if (callchain_ret > 0)
1881                 trace__fprintf_callchain(trace, sample);
1882         else if (callchain_ret < 0)
1883                 pr_err("Problem processing %s callchain, skipping...\n", perf_evsel__name(evsel));
1884 out:
1885         err = 0;
1886 out_put:
1887         thread__put(thread);
1888         return err;
1889 }
1890
1891 static void trace__set_base_time(struct trace *trace,
1892                                  struct perf_evsel *evsel,
1893                                  struct perf_sample *sample)
1894 {
1895         /*
1896          * BPF events were not setting PERF_SAMPLE_TIME, so be more robust
1897          * and don't use sample->time unconditionally, we may end up having
1898          * some other event in the future without PERF_SAMPLE_TIME for good
1899          * reason, i.e. we may not be interested in its timestamps, just in
1900          * it taking place, picking some piece of information when it
1901          * appears in our event stream (vfs_getname comes to mind).
1902          */
1903         if (trace->base_time == 0 && !trace->full_time &&
1904             (evsel->attr.sample_type & PERF_SAMPLE_TIME))
1905                 trace->base_time = sample->time;
1906 }
1907
1908 static int trace__process_sample(struct perf_tool *tool,
1909                                  union perf_event *event,
1910                                  struct perf_sample *sample,
1911                                  struct perf_evsel *evsel,
1912                                  struct machine *machine __maybe_unused)
1913 {
1914         struct trace *trace = container_of(tool, struct trace, tool);
1915         struct thread *thread;
1916         int err = 0;
1917
1918         tracepoint_handler handler = evsel->handler;
1919
1920         thread = machine__findnew_thread(trace->host, sample->pid, sample->tid);
1921         if (thread && thread__is_filtered(thread))
1922                 return 0;
1923
1924         trace__set_base_time(trace, evsel, sample);
1925
1926         if (handler) {
1927                 ++trace->nr_events;
1928                 handler(trace, evsel, event, sample);
1929         }
1930
1931         return err;
1932 }
1933
1934 static int trace__record(struct trace *trace, int argc, const char **argv)
1935 {
1936         unsigned int rec_argc, i, j;
1937         const char **rec_argv;
1938         const char * const record_args[] = {
1939                 "record",
1940                 "-R",
1941                 "-m", "1024",
1942                 "-c", "1",
1943         };
1944
1945         const char * const sc_args[] = { "-e", };
1946         unsigned int sc_args_nr = ARRAY_SIZE(sc_args);
1947         const char * const majpf_args[] = { "-e", "major-faults" };
1948         unsigned int majpf_args_nr = ARRAY_SIZE(majpf_args);
1949         const char * const minpf_args[] = { "-e", "minor-faults" };
1950         unsigned int minpf_args_nr = ARRAY_SIZE(minpf_args);
1951
1952         /* +1 is for the event string below */
1953         rec_argc = ARRAY_SIZE(record_args) + sc_args_nr + 1 +
1954                 majpf_args_nr + minpf_args_nr + argc;
1955         rec_argv = calloc(rec_argc + 1, sizeof(char *));
1956
1957         if (rec_argv == NULL)
1958                 return -ENOMEM;
1959
1960         j = 0;
1961         for (i = 0; i < ARRAY_SIZE(record_args); i++)
1962                 rec_argv[j++] = record_args[i];
1963
1964         if (trace->trace_syscalls) {
1965                 for (i = 0; i < sc_args_nr; i++)
1966                         rec_argv[j++] = sc_args[i];
1967
1968                 /* event string may be different for older kernels - e.g., RHEL6 */
1969                 if (is_valid_tracepoint("raw_syscalls:sys_enter"))
1970                         rec_argv[j++] = "raw_syscalls:sys_enter,raw_syscalls:sys_exit";
1971                 else if (is_valid_tracepoint("syscalls:sys_enter"))
1972                         rec_argv[j++] = "syscalls:sys_enter,syscalls:sys_exit";
1973                 else {
1974                         pr_err("Neither raw_syscalls nor syscalls events exist.\n");
1975                         return -1;
1976                 }
1977         }
1978
1979         if (trace->trace_pgfaults & TRACE_PFMAJ)
1980                 for (i = 0; i < majpf_args_nr; i++)
1981                         rec_argv[j++] = majpf_args[i];
1982
1983         if (trace->trace_pgfaults & TRACE_PFMIN)
1984                 for (i = 0; i < minpf_args_nr; i++)
1985                         rec_argv[j++] = minpf_args[i];
1986
1987         for (i = 0; i < (unsigned int)argc; i++)
1988                 rec_argv[j++] = argv[i];
1989
1990         return cmd_record(j, rec_argv, NULL);
1991 }
1992
1993 static size_t trace__fprintf_thread_summary(struct trace *trace, FILE *fp);
1994
1995 static bool perf_evlist__add_vfs_getname(struct perf_evlist *evlist)
1996 {
1997         struct perf_evsel *evsel = perf_evsel__newtp("probe", "vfs_getname");
1998
1999         if (IS_ERR(evsel))
2000                 return false;
2001
2002         if (perf_evsel__field(evsel, "pathname") == NULL) {
2003                 perf_evsel__delete(evsel);
2004                 return false;
2005         }
2006
2007         evsel->handler = trace__vfs_getname;
2008         perf_evlist__add(evlist, evsel);
2009         return true;
2010 }
2011
2012 static struct perf_evsel *perf_evsel__new_pgfault(u64 config)
2013 {
2014         struct perf_evsel *evsel;
2015         struct perf_event_attr attr = {
2016                 .type = PERF_TYPE_SOFTWARE,
2017                 .mmap_data = 1,
2018         };
2019
2020         attr.config = config;
2021         attr.sample_period = 1;
2022
2023         event_attr_init(&attr);
2024
2025         evsel = perf_evsel__new(&attr);
2026         if (evsel)
2027                 evsel->handler = trace__pgfault;
2028
2029         return evsel;
2030 }
2031
2032 static void trace__handle_event(struct trace *trace, union perf_event *event, struct perf_sample *sample)
2033 {
2034         const u32 type = event->header.type;
2035         struct perf_evsel *evsel;
2036
2037         if (type != PERF_RECORD_SAMPLE) {
2038                 trace__process_event(trace, trace->host, event, sample);
2039                 return;
2040         }
2041
2042         evsel = perf_evlist__id2evsel(trace->evlist, sample->id);
2043         if (evsel == NULL) {
2044                 fprintf(trace->output, "Unknown tp ID %" PRIu64 ", skipping...\n", sample->id);
2045                 return;
2046         }
2047
2048         trace__set_base_time(trace, evsel, sample);
2049
2050         if (evsel->attr.type == PERF_TYPE_TRACEPOINT &&
2051             sample->raw_data == NULL) {
2052                 fprintf(trace->output, "%s sample with no payload for tid: %d, cpu %d, raw_size=%d, skipping...\n",
2053                        perf_evsel__name(evsel), sample->tid,
2054                        sample->cpu, sample->raw_size);
2055         } else {
2056                 tracepoint_handler handler = evsel->handler;
2057                 handler(trace, evsel, event, sample);
2058         }
2059 }
2060
2061 static int trace__add_syscall_newtp(struct trace *trace)
2062 {
2063         int ret = -1;
2064         struct perf_evlist *evlist = trace->evlist;
2065         struct perf_evsel *sys_enter, *sys_exit;
2066
2067         sys_enter = perf_evsel__syscall_newtp("sys_enter", trace__sys_enter);
2068         if (sys_enter == NULL)
2069                 goto out;
2070
2071         if (perf_evsel__init_sc_tp_ptr_field(sys_enter, args))
2072                 goto out_delete_sys_enter;
2073
2074         sys_exit = perf_evsel__syscall_newtp("sys_exit", trace__sys_exit);
2075         if (sys_exit == NULL)
2076                 goto out_delete_sys_enter;
2077
2078         if (perf_evsel__init_sc_tp_uint_field(sys_exit, ret))
2079                 goto out_delete_sys_exit;
2080
2081         perf_evlist__add(evlist, sys_enter);
2082         perf_evlist__add(evlist, sys_exit);
2083
2084         if (callchain_param.enabled && !trace->kernel_syscallchains) {
2085                 /*
2086                  * We're interested only in the user space callchain
2087                  * leading to the syscall, allow overriding that for
2088                  * debugging reasons using --kernel_syscall_callchains
2089                  */
2090                 sys_exit->attr.exclude_callchain_kernel = 1;
2091         }
2092
2093         trace->syscalls.events.sys_enter = sys_enter;
2094         trace->syscalls.events.sys_exit  = sys_exit;
2095
2096         ret = 0;
2097 out:
2098         return ret;
2099
2100 out_delete_sys_exit:
2101         perf_evsel__delete_priv(sys_exit);
2102 out_delete_sys_enter:
2103         perf_evsel__delete_priv(sys_enter);
2104         goto out;
2105 }
2106
2107 static int trace__set_ev_qualifier_filter(struct trace *trace)
2108 {
2109         int err = -1;
2110         struct perf_evsel *sys_exit;
2111         char *filter = asprintf_expr_inout_ints("id", !trace->not_ev_qualifier,
2112                                                 trace->ev_qualifier_ids.nr,
2113                                                 trace->ev_qualifier_ids.entries);
2114
2115         if (filter == NULL)
2116                 goto out_enomem;
2117
2118         if (!perf_evsel__append_tp_filter(trace->syscalls.events.sys_enter,
2119                                           filter)) {
2120                 sys_exit = trace->syscalls.events.sys_exit;
2121                 err = perf_evsel__append_tp_filter(sys_exit, filter);
2122         }
2123
2124         free(filter);
2125 out:
2126         return err;
2127 out_enomem:
2128         errno = ENOMEM;
2129         goto out;
2130 }
2131
2132 static int trace__run(struct trace *trace, int argc, const char **argv)
2133 {
2134         struct perf_evlist *evlist = trace->evlist;
2135         struct perf_evsel *evsel, *pgfault_maj = NULL, *pgfault_min = NULL;
2136         int err = -1, i;
2137         unsigned long before;
2138         const bool forks = argc > 0;
2139         bool draining = false;
2140
2141         trace->live = true;
2142
2143         if (trace->trace_syscalls && trace__add_syscall_newtp(trace))
2144                 goto out_error_raw_syscalls;
2145
2146         if (trace->trace_syscalls)
2147                 trace->vfs_getname = perf_evlist__add_vfs_getname(evlist);
2148
2149         if ((trace->trace_pgfaults & TRACE_PFMAJ)) {
2150                 pgfault_maj = perf_evsel__new_pgfault(PERF_COUNT_SW_PAGE_FAULTS_MAJ);
2151                 if (pgfault_maj == NULL)
2152                         goto out_error_mem;
2153                 perf_evlist__add(evlist, pgfault_maj);
2154         }
2155
2156         if ((trace->trace_pgfaults & TRACE_PFMIN)) {
2157                 pgfault_min = perf_evsel__new_pgfault(PERF_COUNT_SW_PAGE_FAULTS_MIN);
2158                 if (pgfault_min == NULL)
2159                         goto out_error_mem;
2160                 perf_evlist__add(evlist, pgfault_min);
2161         }
2162
2163         if (trace->sched &&
2164             perf_evlist__add_newtp(evlist, "sched", "sched_stat_runtime",
2165                                    trace__sched_stat_runtime))
2166                 goto out_error_sched_stat_runtime;
2167
2168         err = perf_evlist__create_maps(evlist, &trace->opts.target);
2169         if (err < 0) {
2170                 fprintf(trace->output, "Problems parsing the target to trace, check your options!\n");
2171                 goto out_delete_evlist;
2172         }
2173
2174         err = trace__symbols_init(trace, evlist);
2175         if (err < 0) {
2176                 fprintf(trace->output, "Problems initializing symbol libraries!\n");
2177                 goto out_delete_evlist;
2178         }
2179
2180         perf_evlist__config(evlist, &trace->opts, NULL);
2181
2182         if (callchain_param.enabled) {
2183                 bool use_identifier = false;
2184
2185                 if (trace->syscalls.events.sys_exit) {
2186                         perf_evsel__config_callchain(trace->syscalls.events.sys_exit,
2187                                                      &trace->opts, &callchain_param);
2188                         use_identifier = true;
2189                 }
2190
2191                 if (pgfault_maj) {
2192                         perf_evsel__config_callchain(pgfault_maj, &trace->opts, &callchain_param);
2193                         use_identifier = true;
2194                 }
2195
2196                 if (pgfault_min) {
2197                         perf_evsel__config_callchain(pgfault_min, &trace->opts, &callchain_param);
2198                         use_identifier = true;
2199                 }
2200
2201                 if (use_identifier) {
2202                        /*
2203                         * Now we have evsels with different sample_ids, use
2204                         * PERF_SAMPLE_IDENTIFIER to map from sample to evsel
2205                         * from a fixed position in each ring buffer record.
2206                         *
2207                         * As of this the changeset introducing this comment, this
2208                         * isn't strictly needed, as the fields that can come before
2209                         * PERF_SAMPLE_ID are all used, but we'll probably disable
2210                         * some of those for things like copying the payload of
2211                         * pointer syscall arguments, and for vfs_getname we don't
2212                         * need PERF_SAMPLE_ADDR and PERF_SAMPLE_IP, so do this
2213                         * here as a warning we need to use PERF_SAMPLE_IDENTIFIER.
2214                         */
2215                         perf_evlist__set_sample_bit(evlist, IDENTIFIER);
2216                         perf_evlist__reset_sample_bit(evlist, ID);
2217                 }
2218         }
2219
2220         signal(SIGCHLD, sig_handler);
2221         signal(SIGINT, sig_handler);
2222
2223         if (forks) {
2224                 err = perf_evlist__prepare_workload(evlist, &trace->opts.target,
2225                                                     argv, false, NULL);
2226                 if (err < 0) {
2227                         fprintf(trace->output, "Couldn't run the workload!\n");
2228                         goto out_delete_evlist;
2229                 }
2230         }
2231
2232         err = perf_evlist__open(evlist);
2233         if (err < 0)
2234                 goto out_error_open;
2235
2236         err = bpf__apply_obj_config();
2237         if (err) {
2238                 char errbuf[BUFSIZ];
2239
2240                 bpf__strerror_apply_obj_config(err, errbuf, sizeof(errbuf));
2241                 pr_err("ERROR: Apply config to BPF failed: %s\n",
2242                          errbuf);
2243                 goto out_error_open;
2244         }
2245
2246         /*
2247          * Better not use !target__has_task() here because we need to cover the
2248          * case where no threads were specified in the command line, but a
2249          * workload was, and in that case we will fill in the thread_map when
2250          * we fork the workload in perf_evlist__prepare_workload.
2251          */
2252         if (trace->filter_pids.nr > 0)
2253                 err = perf_evlist__set_filter_pids(evlist, trace->filter_pids.nr, trace->filter_pids.entries);
2254         else if (thread_map__pid(evlist->threads, 0) == -1)
2255                 err = perf_evlist__set_filter_pid(evlist, getpid());
2256
2257         if (err < 0)
2258                 goto out_error_mem;
2259
2260         if (trace->ev_qualifier_ids.nr > 0) {
2261                 err = trace__set_ev_qualifier_filter(trace);
2262                 if (err < 0)
2263                         goto out_errno;
2264
2265                 pr_debug("event qualifier tracepoint filter: %s\n",
2266                          trace->syscalls.events.sys_exit->filter);
2267         }
2268
2269         err = perf_evlist__apply_filters(evlist, &evsel);
2270         if (err < 0)
2271                 goto out_error_apply_filters;
2272
2273         err = perf_evlist__mmap(evlist, trace->opts.mmap_pages, false);
2274         if (err < 0)
2275                 goto out_error_mmap;
2276
2277         if (!target__none(&trace->opts.target) && !trace->opts.initial_delay)
2278                 perf_evlist__enable(evlist);
2279
2280         if (forks)
2281                 perf_evlist__start_workload(evlist);
2282
2283         if (trace->opts.initial_delay) {
2284                 usleep(trace->opts.initial_delay * 1000);
2285                 perf_evlist__enable(evlist);
2286         }
2287
2288         trace->multiple_threads = thread_map__pid(evlist->threads, 0) == -1 ||
2289                                   evlist->threads->nr > 1 ||
2290                                   perf_evlist__first(evlist)->attr.inherit;
2291 again:
2292         before = trace->nr_events;
2293
2294         for (i = 0; i < evlist->nr_mmaps; i++) {
2295                 union perf_event *event;
2296
2297                 while ((event = perf_evlist__mmap_read(evlist, i)) != NULL) {
2298                         struct perf_sample sample;
2299
2300                         ++trace->nr_events;
2301
2302                         err = perf_evlist__parse_sample(evlist, event, &sample);
2303                         if (err) {
2304                                 fprintf(trace->output, "Can't parse sample, err = %d, skipping...\n", err);
2305                                 goto next_event;
2306                         }
2307
2308                         trace__handle_event(trace, event, &sample);
2309 next_event:
2310                         perf_evlist__mmap_consume(evlist, i);
2311
2312                         if (interrupted)
2313                                 goto out_disable;
2314
2315                         if (done && !draining) {
2316                                 perf_evlist__disable(evlist);
2317                                 draining = true;
2318                         }
2319                 }
2320         }
2321
2322         if (trace->nr_events == before) {
2323                 int timeout = done ? 100 : -1;
2324
2325                 if (!draining && perf_evlist__poll(evlist, timeout) > 0) {
2326                         if (perf_evlist__filter_pollfd(evlist, POLLERR | POLLHUP) == 0)
2327                                 draining = true;
2328
2329                         goto again;
2330                 }
2331         } else {
2332                 goto again;
2333         }
2334
2335 out_disable:
2336         thread__zput(trace->current);
2337
2338         perf_evlist__disable(evlist);
2339
2340         if (!err) {
2341                 if (trace->summary)
2342                         trace__fprintf_thread_summary(trace, trace->output);
2343
2344                 if (trace->show_tool_stats) {
2345                         fprintf(trace->output, "Stats:\n "
2346                                                " vfs_getname : %" PRIu64 "\n"
2347                                                " proc_getname: %" PRIu64 "\n",
2348                                 trace->stats.vfs_getname,
2349                                 trace->stats.proc_getname);
2350                 }
2351         }
2352
2353 out_delete_evlist:
2354         perf_evlist__delete(evlist);
2355         trace->evlist = NULL;
2356         trace->live = false;
2357         return err;
2358 {
2359         char errbuf[BUFSIZ];
2360
2361 out_error_sched_stat_runtime:
2362         tracing_path__strerror_open_tp(errno, errbuf, sizeof(errbuf), "sched", "sched_stat_runtime");
2363         goto out_error;
2364
2365 out_error_raw_syscalls:
2366         tracing_path__strerror_open_tp(errno, errbuf, sizeof(errbuf), "raw_syscalls", "sys_(enter|exit)");
2367         goto out_error;
2368
2369 out_error_mmap:
2370         perf_evlist__strerror_mmap(evlist, errno, errbuf, sizeof(errbuf));
2371         goto out_error;
2372
2373 out_error_open:
2374         perf_evlist__strerror_open(evlist, errno, errbuf, sizeof(errbuf));
2375
2376 out_error:
2377         fprintf(trace->output, "%s\n", errbuf);
2378         goto out_delete_evlist;
2379
2380 out_error_apply_filters:
2381         fprintf(trace->output,
2382                 "Failed to set filter \"%s\" on event %s with %d (%s)\n",
2383                 evsel->filter, perf_evsel__name(evsel), errno,
2384                 str_error_r(errno, errbuf, sizeof(errbuf)));
2385         goto out_delete_evlist;
2386 }
2387 out_error_mem:
2388         fprintf(trace->output, "Not enough memory to run!\n");
2389         goto out_delete_evlist;
2390
2391 out_errno:
2392         fprintf(trace->output, "errno=%d,%s\n", errno, strerror(errno));
2393         goto out_delete_evlist;
2394 }
2395
2396 static int trace__replay(struct trace *trace)
2397 {
2398         const struct perf_evsel_str_handler handlers[] = {
2399                 { "probe:vfs_getname",       trace__vfs_getname, },
2400         };
2401         struct perf_data_file file = {
2402                 .path  = input_name,
2403                 .mode  = PERF_DATA_MODE_READ,
2404                 .force = trace->force,
2405         };
2406         struct perf_session *session;
2407         struct perf_evsel *evsel;
2408         int err = -1;
2409
2410         trace->tool.sample        = trace__process_sample;
2411         trace->tool.mmap          = perf_event__process_mmap;
2412         trace->tool.mmap2         = perf_event__process_mmap2;
2413         trace->tool.comm          = perf_event__process_comm;
2414         trace->tool.exit          = perf_event__process_exit;
2415         trace->tool.fork          = perf_event__process_fork;
2416         trace->tool.attr          = perf_event__process_attr;
2417         trace->tool.tracing_data = perf_event__process_tracing_data;
2418         trace->tool.build_id      = perf_event__process_build_id;
2419
2420         trace->tool.ordered_events = true;
2421         trace->tool.ordering_requires_timestamps = true;
2422
2423         /* add tid to output */
2424         trace->multiple_threads = true;
2425
2426         session = perf_session__new(&file, false, &trace->tool);
2427         if (session == NULL)
2428                 return -1;
2429
2430         if (trace->opts.target.pid)
2431                 symbol_conf.pid_list_str = strdup(trace->opts.target.pid);
2432
2433         if (trace->opts.target.tid)
2434                 symbol_conf.tid_list_str = strdup(trace->opts.target.tid);
2435
2436         if (symbol__init(&session->header.env) < 0)
2437                 goto out;
2438
2439         trace->host = &session->machines.host;
2440
2441         err = perf_session__set_tracepoints_handlers(session, handlers);
2442         if (err)
2443                 goto out;
2444
2445         evsel = perf_evlist__find_tracepoint_by_name(session->evlist,
2446                                                      "raw_syscalls:sys_enter");
2447         /* older kernels have syscalls tp versus raw_syscalls */
2448         if (evsel == NULL)
2449                 evsel = perf_evlist__find_tracepoint_by_name(session->evlist,
2450                                                              "syscalls:sys_enter");
2451
2452         if (evsel &&
2453             (perf_evsel__init_syscall_tp(evsel, trace__sys_enter) < 0 ||
2454             perf_evsel__init_sc_tp_ptr_field(evsel, args))) {
2455                 pr_err("Error during initialize raw_syscalls:sys_enter event\n");
2456                 goto out;
2457         }
2458
2459         evsel = perf_evlist__find_tracepoint_by_name(session->evlist,
2460                                                      "raw_syscalls:sys_exit");
2461         if (evsel == NULL)
2462                 evsel = perf_evlist__find_tracepoint_by_name(session->evlist,
2463                                                              "syscalls:sys_exit");
2464         if (evsel &&
2465             (perf_evsel__init_syscall_tp(evsel, trace__sys_exit) < 0 ||
2466             perf_evsel__init_sc_tp_uint_field(evsel, ret))) {
2467                 pr_err("Error during initialize raw_syscalls:sys_exit event\n");
2468                 goto out;
2469         }
2470
2471         evlist__for_each_entry(session->evlist, evsel) {
2472                 if (evsel->attr.type == PERF_TYPE_SOFTWARE &&
2473                     (evsel->attr.config == PERF_COUNT_SW_PAGE_FAULTS_MAJ ||
2474                      evsel->attr.config == PERF_COUNT_SW_PAGE_FAULTS_MIN ||
2475                      evsel->attr.config == PERF_COUNT_SW_PAGE_FAULTS))
2476                         evsel->handler = trace__pgfault;
2477         }
2478
2479         setup_pager();
2480
2481         err = perf_session__process_events(session);
2482         if (err)
2483                 pr_err("Failed to process events, error %d", err);
2484
2485         else if (trace->summary)
2486                 trace__fprintf_thread_summary(trace, trace->output);
2487
2488 out:
2489         perf_session__delete(session);
2490
2491         return err;
2492 }
2493
2494 static size_t trace__fprintf_threads_header(FILE *fp)
2495 {
2496         size_t printed;
2497
2498         printed  = fprintf(fp, "\n Summary of events:\n\n");
2499
2500         return printed;
2501 }
2502
2503 DEFINE_RESORT_RB(syscall_stats, a->msecs > b->msecs,
2504         struct stats    *stats;
2505         double          msecs;
2506         int             syscall;
2507 )
2508 {
2509         struct int_node *source = rb_entry(nd, struct int_node, rb_node);
2510         struct stats *stats = source->priv;
2511
2512         entry->syscall = source->i;
2513         entry->stats   = stats;
2514         entry->msecs   = stats ? (u64)stats->n * (avg_stats(stats) / NSEC_PER_MSEC) : 0;
2515 }
2516
2517 static size_t thread__dump_stats(struct thread_trace *ttrace,
2518                                  struct trace *trace, FILE *fp)
2519 {
2520         size_t printed = 0;
2521         struct syscall *sc;
2522         struct rb_node *nd;
2523         DECLARE_RESORT_RB_INTLIST(syscall_stats, ttrace->syscall_stats);
2524
2525         if (syscall_stats == NULL)
2526                 return 0;
2527
2528         printed += fprintf(fp, "\n");
2529
2530         printed += fprintf(fp, "   syscall            calls    total       min       avg       max      stddev\n");
2531         printed += fprintf(fp, "                               (msec)    (msec)    (msec)    (msec)        (%%)\n");
2532         printed += fprintf(fp, "   --------------- -------- --------- --------- --------- ---------     ------\n");
2533
2534         resort_rb__for_each_entry(nd, syscall_stats) {
2535                 struct stats *stats = syscall_stats_entry->stats;
2536                 if (stats) {
2537                         double min = (double)(stats->min) / NSEC_PER_MSEC;
2538                         double max = (double)(stats->max) / NSEC_PER_MSEC;
2539                         double avg = avg_stats(stats);
2540                         double pct;
2541                         u64 n = (u64) stats->n;
2542
2543                         pct = avg ? 100.0 * stddev_stats(stats)/avg : 0.0;
2544                         avg /= NSEC_PER_MSEC;
2545
2546                         sc = &trace->syscalls.table[syscall_stats_entry->syscall];
2547                         printed += fprintf(fp, "   %-15s", sc->name);
2548                         printed += fprintf(fp, " %8" PRIu64 " %9.3f %9.3f %9.3f",
2549                                            n, syscall_stats_entry->msecs, min, avg);
2550                         printed += fprintf(fp, " %9.3f %9.2f%%\n", max, pct);
2551                 }
2552         }
2553
2554         resort_rb__delete(syscall_stats);
2555         printed += fprintf(fp, "\n\n");
2556
2557         return printed;
2558 }
2559
2560 static size_t trace__fprintf_thread(FILE *fp, struct thread *thread, struct trace *trace)
2561 {
2562         size_t printed = 0;
2563         struct thread_trace *ttrace = thread__priv(thread);
2564         double ratio;
2565
2566         if (ttrace == NULL)
2567                 return 0;
2568
2569         ratio = (double)ttrace->nr_events / trace->nr_events * 100.0;
2570
2571         printed += fprintf(fp, " %s (%d), ", thread__comm_str(thread), thread->tid);
2572         printed += fprintf(fp, "%lu events, ", ttrace->nr_events);
2573         printed += fprintf(fp, "%.1f%%", ratio);
2574         if (ttrace->pfmaj)
2575                 printed += fprintf(fp, ", %lu majfaults", ttrace->pfmaj);
2576         if (ttrace->pfmin)
2577                 printed += fprintf(fp, ", %lu minfaults", ttrace->pfmin);
2578         if (trace->sched)
2579                 printed += fprintf(fp, ", %.3f msec\n", ttrace->runtime_ms);
2580         else if (fputc('\n', fp) != EOF)
2581                 ++printed;
2582
2583         printed += thread__dump_stats(ttrace, trace, fp);
2584
2585         return printed;
2586 }
2587
2588 static unsigned long thread__nr_events(struct thread_trace *ttrace)
2589 {
2590         return ttrace ? ttrace->nr_events : 0;
2591 }
2592
2593 DEFINE_RESORT_RB(threads, (thread__nr_events(a->thread->priv) < thread__nr_events(b->thread->priv)),
2594         struct thread *thread;
2595 )
2596 {
2597         entry->thread = rb_entry(nd, struct thread, rb_node);
2598 }
2599
2600 static size_t trace__fprintf_thread_summary(struct trace *trace, FILE *fp)
2601 {
2602         DECLARE_RESORT_RB_MACHINE_THREADS(threads, trace->host);
2603         size_t printed = trace__fprintf_threads_header(fp);
2604         struct rb_node *nd;
2605
2606         if (threads == NULL) {
2607                 fprintf(fp, "%s", "Error sorting output by nr_events!\n");
2608                 return 0;
2609         }
2610
2611         resort_rb__for_each_entry(nd, threads)
2612                 printed += trace__fprintf_thread(fp, threads_entry->thread, trace);
2613
2614         resort_rb__delete(threads);
2615
2616         return printed;
2617 }
2618
2619 static int trace__set_duration(const struct option *opt, const char *str,
2620                                int unset __maybe_unused)
2621 {
2622         struct trace *trace = opt->value;
2623
2624         trace->duration_filter = atof(str);
2625         return 0;
2626 }
2627
2628 static int trace__set_filter_pids(const struct option *opt, const char *str,
2629                                   int unset __maybe_unused)
2630 {
2631         int ret = -1;
2632         size_t i;
2633         struct trace *trace = opt->value;
2634         /*
2635          * FIXME: introduce a intarray class, plain parse csv and create a
2636          * { int nr, int entries[] } struct...
2637          */
2638         struct intlist *list = intlist__new(str);
2639
2640         if (list == NULL)
2641                 return -1;
2642
2643         i = trace->filter_pids.nr = intlist__nr_entries(list) + 1;
2644         trace->filter_pids.entries = calloc(i, sizeof(pid_t));
2645
2646         if (trace->filter_pids.entries == NULL)
2647                 goto out;
2648
2649         trace->filter_pids.entries[0] = getpid();
2650
2651         for (i = 1; i < trace->filter_pids.nr; ++i)
2652                 trace->filter_pids.entries[i] = intlist__entry(list, i - 1)->i;
2653
2654         intlist__delete(list);
2655         ret = 0;
2656 out:
2657         return ret;
2658 }
2659
2660 static int trace__open_output(struct trace *trace, const char *filename)
2661 {
2662         struct stat st;
2663
2664         if (!stat(filename, &st) && st.st_size) {
2665                 char oldname[PATH_MAX];
2666
2667                 scnprintf(oldname, sizeof(oldname), "%s.old", filename);
2668                 unlink(oldname);
2669                 rename(filename, oldname);
2670         }
2671
2672         trace->output = fopen(filename, "w");
2673
2674         return trace->output == NULL ? -errno : 0;
2675 }
2676
2677 static int parse_pagefaults(const struct option *opt, const char *str,
2678                             int unset __maybe_unused)
2679 {
2680         int *trace_pgfaults = opt->value;
2681
2682         if (strcmp(str, "all") == 0)
2683                 *trace_pgfaults |= TRACE_PFMAJ | TRACE_PFMIN;
2684         else if (strcmp(str, "maj") == 0)
2685                 *trace_pgfaults |= TRACE_PFMAJ;
2686         else if (strcmp(str, "min") == 0)
2687                 *trace_pgfaults |= TRACE_PFMIN;
2688         else
2689                 return -1;
2690
2691         return 0;
2692 }
2693
2694 static void evlist__set_evsel_handler(struct perf_evlist *evlist, void *handler)
2695 {
2696         struct perf_evsel *evsel;
2697
2698         evlist__for_each_entry(evlist, evsel)
2699                 evsel->handler = handler;
2700 }
2701
2702 int cmd_trace(int argc, const char **argv, const char *prefix __maybe_unused)
2703 {
2704         const char *trace_usage[] = {
2705                 "perf trace [<options>] [<command>]",
2706                 "perf trace [<options>] -- <command> [<options>]",
2707                 "perf trace record [<options>] [<command>]",
2708                 "perf trace record [<options>] -- <command> [<options>]",
2709                 NULL
2710         };
2711         struct trace trace = {
2712                 .syscalls = {
2713                         . max = -1,
2714                 },
2715                 .opts = {
2716                         .target = {
2717                                 .uid       = UINT_MAX,
2718                                 .uses_mmap = true,
2719                         },
2720                         .user_freq     = UINT_MAX,
2721                         .user_interval = ULLONG_MAX,
2722                         .no_buffering  = true,
2723                         .mmap_pages    = UINT_MAX,
2724                         .proc_map_timeout  = 500,
2725                 },
2726                 .output = stderr,
2727                 .show_comm = true,
2728                 .trace_syscalls = true,
2729                 .kernel_syscallchains = false,
2730                 .max_stack = UINT_MAX,
2731         };
2732         const char *output_name = NULL;
2733         const char *ev_qualifier_str = NULL;
2734         const struct option trace_options[] = {
2735         OPT_CALLBACK(0, "event", &trace.evlist, "event",
2736                      "event selector. use 'perf list' to list available events",
2737                      parse_events_option),
2738         OPT_BOOLEAN(0, "comm", &trace.show_comm,
2739                     "show the thread COMM next to its id"),
2740         OPT_BOOLEAN(0, "tool_stats", &trace.show_tool_stats, "show tool stats"),
2741         OPT_STRING('e', "expr", &ev_qualifier_str, "expr", "list of syscalls to trace"),
2742         OPT_STRING('o', "output", &output_name, "file", "output file name"),
2743         OPT_STRING('i', "input", &input_name, "file", "Analyze events in file"),
2744         OPT_STRING('p', "pid", &trace.opts.target.pid, "pid",
2745                     "trace events on existing process id"),
2746         OPT_STRING('t', "tid", &trace.opts.target.tid, "tid",
2747                     "trace events on existing thread id"),
2748         OPT_CALLBACK(0, "filter-pids", &trace, "CSV list of pids",
2749                      "pids to filter (by the kernel)", trace__set_filter_pids),
2750         OPT_BOOLEAN('a', "all-cpus", &trace.opts.target.system_wide,
2751                     "system-wide collection from all CPUs"),
2752         OPT_STRING('C', "cpu", &trace.opts.target.cpu_list, "cpu",
2753                     "list of cpus to monitor"),
2754         OPT_BOOLEAN(0, "no-inherit", &trace.opts.no_inherit,
2755                     "child tasks do not inherit counters"),
2756         OPT_CALLBACK('m', "mmap-pages", &trace.opts.mmap_pages, "pages",
2757                      "number of mmap data pages",
2758                      perf_evlist__parse_mmap_pages),
2759         OPT_STRING('u', "uid", &trace.opts.target.uid_str, "user",
2760                    "user to profile"),
2761         OPT_CALLBACK(0, "duration", &trace, "float",
2762                      "show only events with duration > N.M ms",
2763                      trace__set_duration),
2764         OPT_BOOLEAN(0, "sched", &trace.sched, "show blocking scheduler events"),
2765         OPT_INCR('v', "verbose", &verbose, "be more verbose"),
2766         OPT_BOOLEAN('T', "time", &trace.full_time,
2767                     "Show full timestamp, not time relative to first start"),
2768         OPT_BOOLEAN('s', "summary", &trace.summary_only,
2769                     "Show only syscall summary with statistics"),
2770         OPT_BOOLEAN('S', "with-summary", &trace.summary,
2771                     "Show all syscalls and summary with statistics"),
2772         OPT_CALLBACK_DEFAULT('F', "pf", &trace.trace_pgfaults, "all|maj|min",
2773                      "Trace pagefaults", parse_pagefaults, "maj"),
2774         OPT_BOOLEAN(0, "syscalls", &trace.trace_syscalls, "Trace syscalls"),
2775         OPT_BOOLEAN('f', "force", &trace.force, "don't complain, do it"),
2776         OPT_CALLBACK(0, "call-graph", &trace.opts,
2777                      "record_mode[,record_size]", record_callchain_help,
2778                      &record_parse_callchain_opt),
2779         OPT_BOOLEAN(0, "kernel-syscall-graph", &trace.kernel_syscallchains,
2780                     "Show the kernel callchains on the syscall exit path"),
2781         OPT_UINTEGER(0, "min-stack", &trace.min_stack,
2782                      "Set the minimum stack depth when parsing the callchain, "
2783                      "anything below the specified depth will be ignored."),
2784         OPT_UINTEGER(0, "max-stack", &trace.max_stack,
2785                      "Set the maximum stack depth when parsing the callchain, "
2786                      "anything beyond the specified depth will be ignored. "
2787                      "Default: kernel.perf_event_max_stack or " __stringify(PERF_MAX_STACK_DEPTH)),
2788         OPT_UINTEGER(0, "proc-map-timeout", &trace.opts.proc_map_timeout,
2789                         "per thread proc mmap processing timeout in ms"),
2790         OPT_UINTEGER('D', "delay", &trace.opts.initial_delay,
2791                      "ms to wait before starting measurement after program "
2792                      "start"),
2793         OPT_END()
2794         };
2795         bool __maybe_unused max_stack_user_set = true;
2796         bool mmap_pages_user_set = true;
2797         const char * const trace_subcommands[] = { "record", NULL };
2798         int err;
2799         char bf[BUFSIZ];
2800
2801         signal(SIGSEGV, sighandler_dump_stack);
2802         signal(SIGFPE, sighandler_dump_stack);
2803
2804         trace.evlist = perf_evlist__new();
2805         trace.sctbl = syscalltbl__new();
2806
2807         if (trace.evlist == NULL || trace.sctbl == NULL) {
2808                 pr_err("Not enough memory to run!\n");
2809                 err = -ENOMEM;
2810                 goto out;
2811         }
2812
2813         argc = parse_options_subcommand(argc, argv, trace_options, trace_subcommands,
2814                                  trace_usage, PARSE_OPT_STOP_AT_NON_OPTION);
2815
2816         err = bpf__setup_stdout(trace.evlist);
2817         if (err) {
2818                 bpf__strerror_setup_stdout(trace.evlist, err, bf, sizeof(bf));
2819                 pr_err("ERROR: Setup BPF stdout failed: %s\n", bf);
2820                 goto out;
2821         }
2822
2823         err = -1;
2824
2825         if (trace.trace_pgfaults) {
2826                 trace.opts.sample_address = true;
2827                 trace.opts.sample_time = true;
2828         }
2829
2830         if (trace.opts.mmap_pages == UINT_MAX)
2831                 mmap_pages_user_set = false;
2832
2833         if (trace.max_stack == UINT_MAX) {
2834                 trace.max_stack = input_name ? PERF_MAX_STACK_DEPTH : sysctl_perf_event_max_stack;
2835                 max_stack_user_set = false;
2836         }
2837
2838 #ifdef HAVE_DWARF_UNWIND_SUPPORT
2839         if ((trace.min_stack || max_stack_user_set) && !callchain_param.enabled && trace.trace_syscalls)
2840                 record_opts__parse_callchain(&trace.opts, &callchain_param, "dwarf", false);
2841 #endif
2842
2843         if (callchain_param.enabled) {
2844                 if (!mmap_pages_user_set && geteuid() == 0)
2845                         trace.opts.mmap_pages = perf_event_mlock_kb_in_pages() * 4;
2846
2847                 symbol_conf.use_callchain = true;
2848         }
2849
2850         if (trace.evlist->nr_entries > 0)
2851                 evlist__set_evsel_handler(trace.evlist, trace__event_handler);
2852
2853         if ((argc >= 1) && (strcmp(argv[0], "record") == 0))
2854                 return trace__record(&trace, argc-1, &argv[1]);
2855
2856         /* summary_only implies summary option, but don't overwrite summary if set */
2857         if (trace.summary_only)
2858                 trace.summary = trace.summary_only;
2859
2860         if (!trace.trace_syscalls && !trace.trace_pgfaults &&
2861             trace.evlist->nr_entries == 0 /* Was --events used? */) {
2862                 pr_err("Please specify something to trace.\n");
2863                 return -1;
2864         }
2865
2866         if (!trace.trace_syscalls && ev_qualifier_str) {
2867                 pr_err("The -e option can't be used with --no-syscalls.\n");
2868                 goto out;
2869         }
2870
2871         if (output_name != NULL) {
2872                 err = trace__open_output(&trace, output_name);
2873                 if (err < 0) {
2874                         perror("failed to create output file");
2875                         goto out;
2876                 }
2877         }
2878
2879         trace.open_id = syscalltbl__id(trace.sctbl, "open");
2880
2881         if (ev_qualifier_str != NULL) {
2882                 const char *s = ev_qualifier_str;
2883                 struct strlist_config slist_config = {
2884                         .dirname = system_path(STRACE_GROUPS_DIR),
2885                 };
2886
2887                 trace.not_ev_qualifier = *s == '!';
2888                 if (trace.not_ev_qualifier)
2889                         ++s;
2890                 trace.ev_qualifier = strlist__new(s, &slist_config);
2891                 if (trace.ev_qualifier == NULL) {
2892                         fputs("Not enough memory to parse event qualifier",
2893                               trace.output);
2894                         err = -ENOMEM;
2895                         goto out_close;
2896                 }
2897
2898                 err = trace__validate_ev_qualifier(&trace);
2899                 if (err)
2900                         goto out_close;
2901         }
2902
2903         err = target__validate(&trace.opts.target);
2904         if (err) {
2905                 target__strerror(&trace.opts.target, err, bf, sizeof(bf));
2906                 fprintf(trace.output, "%s", bf);
2907                 goto out_close;
2908         }
2909
2910         err = target__parse_uid(&trace.opts.target);
2911         if (err) {
2912                 target__strerror(&trace.opts.target, err, bf, sizeof(bf));
2913                 fprintf(trace.output, "%s", bf);
2914                 goto out_close;
2915         }
2916
2917         if (!argc && target__none(&trace.opts.target))
2918                 trace.opts.target.system_wide = true;
2919
2920         if (input_name)
2921                 err = trace__replay(&trace);
2922         else
2923                 err = trace__run(&trace, argc, argv);
2924
2925 out_close:
2926         if (output_name != NULL)
2927                 fclose(trace.output);
2928 out:
2929         return err;
2930 }