1 #include <linux/types.h>
9 #include "thread_map.h"
11 static const char *perf_event__names
[] = {
13 [PERF_RECORD_MMAP
] = "MMAP",
14 [PERF_RECORD_MMAP2
] = "MMAP2",
15 [PERF_RECORD_LOST
] = "LOST",
16 [PERF_RECORD_COMM
] = "COMM",
17 [PERF_RECORD_EXIT
] = "EXIT",
18 [PERF_RECORD_THROTTLE
] = "THROTTLE",
19 [PERF_RECORD_UNTHROTTLE
] = "UNTHROTTLE",
20 [PERF_RECORD_FORK
] = "FORK",
21 [PERF_RECORD_READ
] = "READ",
22 [PERF_RECORD_SAMPLE
] = "SAMPLE",
23 [PERF_RECORD_HEADER_ATTR
] = "ATTR",
24 [PERF_RECORD_HEADER_EVENT_TYPE
] = "EVENT_TYPE",
25 [PERF_RECORD_HEADER_TRACING_DATA
] = "TRACING_DATA",
26 [PERF_RECORD_HEADER_BUILD_ID
] = "BUILD_ID",
27 [PERF_RECORD_FINISHED_ROUND
] = "FINISHED_ROUND",
30 const char *perf_event__name(unsigned int id
)
32 if (id
>= ARRAY_SIZE(perf_event__names
))
34 if (!perf_event__names
[id
])
36 return perf_event__names
[id
];
39 static struct perf_sample synth_sample
= {
48 static pid_t
perf_event__get_comm_tgid(pid_t pid
, char *comm
, size_t len
)
50 char filename
[PATH_MAX
];
56 snprintf(filename
, sizeof(filename
), "/proc/%d/status", pid
);
58 fp
= fopen(filename
, "r");
60 pr_debug("couldn't open %s\n", filename
);
64 while (!comm
[0] || (tgid
< 0)) {
65 if (fgets(bf
, sizeof(bf
), fp
) == NULL
) {
66 pr_warning("couldn't get COMM and pgid, malformed %s\n",
71 if (memcmp(bf
, "Name:", 5) == 0) {
73 while (*name
&& isspace(*name
))
75 size
= strlen(name
) - 1;
78 memcpy(comm
, name
, size
);
81 } else if (memcmp(bf
, "Tgid:", 5) == 0) {
83 while (*tgids
&& isspace(*tgids
))
94 static pid_t
perf_event__synthesize_comm(struct perf_tool
*tool
,
95 union perf_event
*event
, pid_t pid
,
97 perf_event__handler_t process
,
98 struct machine
*machine
)
100 char filename
[PATH_MAX
];
103 struct dirent dirent
, *next
;
106 memset(&event
->comm
, 0, sizeof(event
->comm
));
108 tgid
= perf_event__get_comm_tgid(pid
, event
->comm
.comm
,
109 sizeof(event
->comm
.comm
));
113 event
->comm
.pid
= tgid
;
114 event
->comm
.header
.type
= PERF_RECORD_COMM
;
116 size
= strlen(event
->comm
.comm
) + 1;
117 size
= PERF_ALIGN(size
, sizeof(u64
));
118 memset(event
->comm
.comm
+ size
, 0, machine
->id_hdr_size
);
119 event
->comm
.header
.size
= (sizeof(event
->comm
) -
120 (sizeof(event
->comm
.comm
) - size
) +
121 machine
->id_hdr_size
);
123 event
->comm
.tid
= pid
;
125 if (process(tool
, event
, &synth_sample
, machine
) != 0)
131 snprintf(filename
, sizeof(filename
), "/proc/%d/task", pid
);
133 tasks
= opendir(filename
);
135 pr_debug("couldn't open %s\n", filename
);
139 while (!readdir_r(tasks
, &dirent
, &next
) && next
) {
141 pid
= strtol(dirent
.d_name
, &end
, 10);
145 /* already have tgid; jut want to update the comm */
146 (void) perf_event__get_comm_tgid(pid
, event
->comm
.comm
,
147 sizeof(event
->comm
.comm
));
149 size
= strlen(event
->comm
.comm
) + 1;
150 size
= PERF_ALIGN(size
, sizeof(u64
));
151 memset(event
->comm
.comm
+ size
, 0, machine
->id_hdr_size
);
152 event
->comm
.header
.size
= (sizeof(event
->comm
) -
153 (sizeof(event
->comm
.comm
) - size
) +
154 machine
->id_hdr_size
);
156 event
->comm
.tid
= pid
;
158 if (process(tool
, event
, &synth_sample
, machine
) != 0) {
169 static int perf_event__synthesize_mmap_events(struct perf_tool
*tool
,
170 union perf_event
*event
,
171 pid_t pid
, pid_t tgid
,
172 perf_event__handler_t process
,
173 struct machine
*machine
)
175 char filename
[PATH_MAX
];
179 snprintf(filename
, sizeof(filename
), "/proc/%d/maps", pid
);
181 fp
= fopen(filename
, "r");
184 * We raced with a task exiting - just return:
186 pr_debug("couldn't open %s\n", filename
);
190 event
->header
.type
= PERF_RECORD_MMAP
;
192 * Just like the kernel, see __perf_event_mmap in kernel/perf_event.c
194 event
->header
.misc
= PERF_RECORD_MISC_USER
;
199 char execname
[PATH_MAX
];
200 char anonstr
[] = "//anon";
204 if (fgets(bf
, sizeof(bf
), fp
) == NULL
)
207 /* ensure null termination since stack will be reused. */
208 strcpy(execname
, "");
210 /* 00400000-0040c000 r-xp 00000000 fd:01 41038 /bin/cat */
211 n
= sscanf(bf
, "%"PRIx64
"-%"PRIx64
" %s %"PRIx64
" %*x:%*x %*u %s\n",
212 &event
->mmap
.start
, &event
->mmap
.len
, prot
,
222 if (!strcmp(execname
, ""))
223 strcpy(execname
, anonstr
);
225 size
= strlen(execname
) + 1;
226 memcpy(event
->mmap
.filename
, execname
, size
);
227 size
= PERF_ALIGN(size
, sizeof(u64
));
228 event
->mmap
.len
-= event
->mmap
.start
;
229 event
->mmap
.header
.size
= (sizeof(event
->mmap
) -
230 (sizeof(event
->mmap
.filename
) - size
));
231 memset(event
->mmap
.filename
+ size
, 0, machine
->id_hdr_size
);
232 event
->mmap
.header
.size
+= machine
->id_hdr_size
;
233 event
->mmap
.pid
= tgid
;
234 event
->mmap
.tid
= pid
;
236 if (process(tool
, event
, &synth_sample
, machine
) != 0) {
246 int perf_event__synthesize_modules(struct perf_tool
*tool
,
247 perf_event__handler_t process
,
248 struct machine
*machine
)
252 struct map_groups
*kmaps
= &machine
->kmaps
;
253 union perf_event
*event
= zalloc((sizeof(event
->mmap
) +
254 machine
->id_hdr_size
));
256 pr_debug("Not enough memory synthesizing mmap event "
257 "for kernel modules\n");
261 event
->header
.type
= PERF_RECORD_MMAP
;
264 * kernel uses 0 for user space maps, see kernel/perf_event.c
267 if (machine__is_host(machine
))
268 event
->header
.misc
= PERF_RECORD_MISC_KERNEL
;
270 event
->header
.misc
= PERF_RECORD_MISC_GUEST_KERNEL
;
272 for (nd
= rb_first(&kmaps
->maps
[MAP__FUNCTION
]);
273 nd
; nd
= rb_next(nd
)) {
275 struct map
*pos
= rb_entry(nd
, struct map
, rb_node
);
277 if (pos
->dso
->kernel
)
280 size
= PERF_ALIGN(pos
->dso
->long_name_len
+ 1, sizeof(u64
));
281 event
->mmap
.header
.type
= PERF_RECORD_MMAP
;
282 event
->mmap
.header
.size
= (sizeof(event
->mmap
) -
283 (sizeof(event
->mmap
.filename
) - size
));
284 memset(event
->mmap
.filename
+ size
, 0, machine
->id_hdr_size
);
285 event
->mmap
.header
.size
+= machine
->id_hdr_size
;
286 event
->mmap
.start
= pos
->start
;
287 event
->mmap
.len
= pos
->end
- pos
->start
;
288 event
->mmap
.pid
= machine
->pid
;
290 memcpy(event
->mmap
.filename
, pos
->dso
->long_name
,
291 pos
->dso
->long_name_len
+ 1);
292 if (process(tool
, event
, &synth_sample
, machine
) != 0) {
302 static int __event__synthesize_thread(union perf_event
*comm_event
,
303 union perf_event
*mmap_event
,
305 perf_event__handler_t process
,
306 struct perf_tool
*tool
,
307 struct machine
*machine
)
309 pid_t tgid
= perf_event__synthesize_comm(tool
, comm_event
, pid
, full
,
313 return perf_event__synthesize_mmap_events(tool
, mmap_event
, pid
, tgid
,
317 int perf_event__synthesize_thread_map(struct perf_tool
*tool
,
318 struct thread_map
*threads
,
319 perf_event__handler_t process
,
320 struct machine
*machine
)
322 union perf_event
*comm_event
, *mmap_event
;
323 int err
= -1, thread
, j
;
325 comm_event
= malloc(sizeof(comm_event
->comm
) + machine
->id_hdr_size
);
326 if (comm_event
== NULL
)
329 mmap_event
= malloc(sizeof(mmap_event
->mmap
) + machine
->id_hdr_size
);
330 if (mmap_event
== NULL
)
334 for (thread
= 0; thread
< threads
->nr
; ++thread
) {
335 if (__event__synthesize_thread(comm_event
, mmap_event
,
336 threads
->map
[thread
], 0,
337 process
, tool
, machine
)) {
343 * comm.pid is set to thread group id by
344 * perf_event__synthesize_comm
346 if ((int) comm_event
->comm
.pid
!= threads
->map
[thread
]) {
347 bool need_leader
= true;
349 /* is thread group leader in thread_map? */
350 for (j
= 0; j
< threads
->nr
; ++j
) {
351 if ((int) comm_event
->comm
.pid
== threads
->map
[j
]) {
357 /* if not, generate events for it */
359 __event__synthesize_thread(comm_event
,
361 comm_event
->comm
.pid
, 0,
362 process
, tool
, machine
)) {
375 int perf_event__synthesize_threads(struct perf_tool
*tool
,
376 perf_event__handler_t process
,
377 struct machine
*machine
)
380 struct dirent dirent
, *next
;
381 union perf_event
*comm_event
, *mmap_event
;
384 comm_event
= malloc(sizeof(comm_event
->comm
) + machine
->id_hdr_size
);
385 if (comm_event
== NULL
)
388 mmap_event
= malloc(sizeof(mmap_event
->mmap
) + machine
->id_hdr_size
);
389 if (mmap_event
== NULL
)
392 proc
= opendir("/proc");
396 while (!readdir_r(proc
, &dirent
, &next
) && next
) {
398 pid_t pid
= strtol(dirent
.d_name
, &end
, 10);
400 if (*end
) /* only interested in proper numerical dirents */
403 * We may race with exiting thread, so don't stop just because
404 * one thread couldn't be synthesized.
406 __event__synthesize_thread(comm_event
, mmap_event
, pid
, 1,
407 process
, tool
, machine
);
420 struct process_symbol_args
{
425 static int find_symbol_cb(void *arg
, const char *name
, char type
,
428 struct process_symbol_args
*args
= arg
;
431 * Must be a function or at least an alias, as in PARISC64, where "_text" is
432 * an 'A' to the same address as "_stext".
434 if (!(symbol_type__is_a(type
, MAP__FUNCTION
) ||
435 type
== 'A') || strcmp(name
, args
->name
))
442 int perf_event__synthesize_kernel_mmap(struct perf_tool
*tool
,
443 perf_event__handler_t process
,
444 struct machine
*machine
,
445 const char *symbol_name
)
448 const char *filename
, *mmap_name
;
450 char name_buff
[PATH_MAX
];
454 * We should get this from /sys/kernel/sections/.text, but till that is
455 * available use this, and after it is use this as a fallback for older
458 struct process_symbol_args args
= { .name
= symbol_name
, };
459 union perf_event
*event
= zalloc((sizeof(event
->mmap
) +
460 machine
->id_hdr_size
));
462 pr_debug("Not enough memory synthesizing mmap event "
463 "for kernel modules\n");
467 mmap_name
= machine__mmap_name(machine
, name_buff
, sizeof(name_buff
));
468 if (machine__is_host(machine
)) {
470 * kernel uses PERF_RECORD_MISC_USER for user space maps,
471 * see kernel/perf_event.c __perf_event_mmap
473 event
->header
.misc
= PERF_RECORD_MISC_KERNEL
;
474 filename
= "/proc/kallsyms";
476 event
->header
.misc
= PERF_RECORD_MISC_GUEST_KERNEL
;
477 if (machine__is_default_guest(machine
))
478 filename
= (char *) symbol_conf
.default_guest_kallsyms
;
480 sprintf(path
, "%s/proc/kallsyms", machine
->root_dir
);
485 if (kallsyms__parse(filename
, &args
, find_symbol_cb
) <= 0) {
490 map
= machine
->vmlinux_maps
[MAP__FUNCTION
];
491 size
= snprintf(event
->mmap
.filename
, sizeof(event
->mmap
.filename
),
492 "%s%s", mmap_name
, symbol_name
) + 1;
493 size
= PERF_ALIGN(size
, sizeof(u64
));
494 event
->mmap
.header
.type
= PERF_RECORD_MMAP
;
495 event
->mmap
.header
.size
= (sizeof(event
->mmap
) -
496 (sizeof(event
->mmap
.filename
) - size
) + machine
->id_hdr_size
);
497 event
->mmap
.pgoff
= args
.start
;
498 event
->mmap
.start
= map
->start
;
499 event
->mmap
.len
= map
->end
- event
->mmap
.start
;
500 event
->mmap
.pid
= machine
->pid
;
502 err
= process(tool
, event
, &synth_sample
, machine
);
508 size_t perf_event__fprintf_comm(union perf_event
*event
, FILE *fp
)
510 return fprintf(fp
, ": %s:%d\n", event
->comm
.comm
, event
->comm
.tid
);
513 int perf_event__process_comm(struct perf_tool
*tool __maybe_unused
,
514 union perf_event
*event
,
515 struct perf_sample
*sample
,
516 struct machine
*machine
)
518 return machine__process_comm_event(machine
, event
, sample
);
521 int perf_event__process_lost(struct perf_tool
*tool __maybe_unused
,
522 union perf_event
*event
,
523 struct perf_sample
*sample
,
524 struct machine
*machine
)
526 return machine__process_lost_event(machine
, event
, sample
);
529 size_t perf_event__fprintf_mmap(union perf_event
*event
, FILE *fp
)
531 return fprintf(fp
, " %d/%d: [%#" PRIx64
"(%#" PRIx64
") @ %#" PRIx64
"]: %s\n",
532 event
->mmap
.pid
, event
->mmap
.tid
, event
->mmap
.start
,
533 event
->mmap
.len
, event
->mmap
.pgoff
, event
->mmap
.filename
);
536 size_t perf_event__fprintf_mmap2(union perf_event
*event
, FILE *fp
)
538 return fprintf(fp
, " %d/%d: [%#" PRIx64
"(%#" PRIx64
") @ %#" PRIx64
539 " %02x:%02x %"PRIu64
" %"PRIu64
"]: %s\n",
540 event
->mmap2
.pid
, event
->mmap2
.tid
, event
->mmap2
.start
,
541 event
->mmap2
.len
, event
->mmap2
.pgoff
, event
->mmap2
.maj
,
542 event
->mmap2
.min
, event
->mmap2
.ino
,
543 event
->mmap2
.ino_generation
,
544 event
->mmap2
.filename
);
547 int perf_event__process_mmap(struct perf_tool
*tool __maybe_unused
,
548 union perf_event
*event
,
549 struct perf_sample
*sample
,
550 struct machine
*machine
)
552 return machine__process_mmap_event(machine
, event
, sample
);
555 int perf_event__process_mmap2(struct perf_tool
*tool __maybe_unused
,
556 union perf_event
*event
,
557 struct perf_sample
*sample
,
558 struct machine
*machine
)
560 return machine__process_mmap2_event(machine
, event
, sample
);
563 size_t perf_event__fprintf_task(union perf_event
*event
, FILE *fp
)
565 return fprintf(fp
, "(%d:%d):(%d:%d)\n",
566 event
->fork
.pid
, event
->fork
.tid
,
567 event
->fork
.ppid
, event
->fork
.ptid
);
570 int perf_event__process_fork(struct perf_tool
*tool __maybe_unused
,
571 union perf_event
*event
,
572 struct perf_sample
*sample
,
573 struct machine
*machine
)
575 return machine__process_fork_event(machine
, event
, sample
);
578 int perf_event__process_exit(struct perf_tool
*tool __maybe_unused
,
579 union perf_event
*event
,
580 struct perf_sample
*sample
,
581 struct machine
*machine
)
583 return machine__process_exit_event(machine
, event
, sample
);
586 size_t perf_event__fprintf(union perf_event
*event
, FILE *fp
)
588 size_t ret
= fprintf(fp
, "PERF_RECORD_%s",
589 perf_event__name(event
->header
.type
));
591 switch (event
->header
.type
) {
592 case PERF_RECORD_COMM
:
593 ret
+= perf_event__fprintf_comm(event
, fp
);
595 case PERF_RECORD_FORK
:
596 case PERF_RECORD_EXIT
:
597 ret
+= perf_event__fprintf_task(event
, fp
);
599 case PERF_RECORD_MMAP
:
600 ret
+= perf_event__fprintf_mmap(event
, fp
);
602 case PERF_RECORD_MMAP2
:
603 ret
+= perf_event__fprintf_mmap2(event
, fp
);
606 ret
+= fprintf(fp
, "\n");
612 int perf_event__process(struct perf_tool
*tool __maybe_unused
,
613 union perf_event
*event
,
614 struct perf_sample
*sample
,
615 struct machine
*machine
)
617 return machine__process_event(machine
, event
, sample
);
620 void thread__find_addr_map(struct thread
*thread
,
621 struct machine
*machine
, u8 cpumode
,
622 enum map_type type
, u64 addr
,
623 struct addr_location
*al
)
625 struct map_groups
*mg
= &thread
->mg
;
626 bool load_map
= false;
630 al
->cpumode
= cpumode
;
631 al
->filtered
= false;
633 if (machine
== NULL
) {
638 if (cpumode
== PERF_RECORD_MISC_KERNEL
&& perf_host
) {
640 mg
= &machine
->kmaps
;
642 } else if (cpumode
== PERF_RECORD_MISC_USER
&& perf_host
) {
644 } else if (cpumode
== PERF_RECORD_MISC_GUEST_KERNEL
&& perf_guest
) {
646 mg
= &machine
->kmaps
;
650 * 'u' means guest os user space.
651 * TODO: We don't support guest user space. Might support late.
653 if (cpumode
== PERF_RECORD_MISC_GUEST_USER
&& perf_guest
)
659 if ((cpumode
== PERF_RECORD_MISC_GUEST_USER
||
660 cpumode
== PERF_RECORD_MISC_GUEST_KERNEL
) &&
663 if ((cpumode
== PERF_RECORD_MISC_USER
||
664 cpumode
== PERF_RECORD_MISC_KERNEL
) &&
671 al
->map
= map_groups__find(mg
, type
, al
->addr
);
672 if (al
->map
== NULL
) {
674 * If this is outside of all known maps, and is a negative
675 * address, try to look it up in the kernel dso, as it might be
676 * a vsyscall or vdso (which executes in user-mode).
678 * XXX This is nasty, we should have a symbol list in the
679 * "[vdso]" dso, but for now lets use the old trick of looking
680 * in the whole kernel symbol list.
682 if ((long long)al
->addr
< 0 &&
683 cpumode
== PERF_RECORD_MISC_USER
&&
684 machine
&& mg
!= &machine
->kmaps
) {
685 mg
= &machine
->kmaps
;
690 * Kernel maps might be changed when loading symbols so loading
691 * must be done prior to using kernel maps.
694 map__load(al
->map
, machine
->symbol_filter
);
695 al
->addr
= al
->map
->map_ip(al
->map
, al
->addr
);
699 void thread__find_addr_location(struct thread
*thread
, struct machine
*machine
,
700 u8 cpumode
, enum map_type type
, u64 addr
,
701 struct addr_location
*al
)
703 thread__find_addr_map(thread
, machine
, cpumode
, type
, addr
, al
);
705 al
->sym
= map__find_symbol(al
->map
, al
->addr
,
706 machine
->symbol_filter
);
711 int perf_event__preprocess_sample(const union perf_event
*event
,
712 struct machine
*machine
,
713 struct addr_location
*al
,
714 struct perf_sample
*sample
)
716 u8 cpumode
= event
->header
.misc
& PERF_RECORD_MISC_CPUMODE_MASK
;
717 struct thread
*thread
= machine__findnew_thread(machine
, sample
->pid
,
723 if (symbol_conf
.comm_list
&&
724 !strlist__has_entry(symbol_conf
.comm_list
, thread__comm_str(thread
)))
727 dump_printf(" ... thread: %s:%d\n", thread__comm_str(thread
), thread
->tid
);
729 * Have we already created the kernel maps for this machine?
731 * This should have happened earlier, when we processed the kernel MMAP
732 * events, but for older perf.data files there was no such thing, so do
735 if (cpumode
== PERF_RECORD_MISC_KERNEL
&&
736 machine
->vmlinux_maps
[MAP__FUNCTION
] == NULL
)
737 machine__create_kernel_maps(machine
);
739 thread__find_addr_map(thread
, machine
, cpumode
, MAP__FUNCTION
,
741 dump_printf(" ...... dso: %s\n",
742 al
->map
? al
->map
->dso
->long_name
:
743 al
->level
== 'H' ? "[hypervisor]" : "<not found>");
745 al
->cpu
= sample
->cpu
;
748 struct dso
*dso
= al
->map
->dso
;
750 if (symbol_conf
.dso_list
&&
751 (!dso
|| !(strlist__has_entry(symbol_conf
.dso_list
,
753 (dso
->short_name
!= dso
->long_name
&&
754 strlist__has_entry(symbol_conf
.dso_list
,
758 al
->sym
= map__find_symbol(al
->map
, al
->addr
,
759 machine
->symbol_filter
);
762 if (symbol_conf
.sym_list
&&
763 (!al
->sym
|| !strlist__has_entry(symbol_conf
.sym_list
,