1 #include <linux/types.h>
9 #include "thread_map.h"
11 static const char *perf_event__names
[] = {
13 [PERF_RECORD_MMAP
] = "MMAP",
14 [PERF_RECORD_MMAP2
] = "MMAP2",
15 [PERF_RECORD_LOST
] = "LOST",
16 [PERF_RECORD_COMM
] = "COMM",
17 [PERF_RECORD_EXIT
] = "EXIT",
18 [PERF_RECORD_THROTTLE
] = "THROTTLE",
19 [PERF_RECORD_UNTHROTTLE
] = "UNTHROTTLE",
20 [PERF_RECORD_FORK
] = "FORK",
21 [PERF_RECORD_READ
] = "READ",
22 [PERF_RECORD_SAMPLE
] = "SAMPLE",
23 [PERF_RECORD_HEADER_ATTR
] = "ATTR",
24 [PERF_RECORD_HEADER_EVENT_TYPE
] = "EVENT_TYPE",
25 [PERF_RECORD_HEADER_TRACING_DATA
] = "TRACING_DATA",
26 [PERF_RECORD_HEADER_BUILD_ID
] = "BUILD_ID",
27 [PERF_RECORD_FINISHED_ROUND
] = "FINISHED_ROUND",
30 const char *perf_event__name(unsigned int id
)
32 if (id
>= ARRAY_SIZE(perf_event__names
))
34 if (!perf_event__names
[id
])
36 return perf_event__names
[id
];
39 static struct perf_sample synth_sample
= {
48 static pid_t
perf_event__get_comm_tgid(pid_t pid
, char *comm
, size_t len
)
50 char filename
[PATH_MAX
];
56 snprintf(filename
, sizeof(filename
), "/proc/%d/status", pid
);
58 fp
= fopen(filename
, "r");
60 pr_debug("couldn't open %s\n", filename
);
64 while (!comm
[0] || (tgid
< 0)) {
65 if (fgets(bf
, sizeof(bf
), fp
) == NULL
) {
66 pr_warning("couldn't get COMM and pgid, malformed %s\n",
71 if (memcmp(bf
, "Name:", 5) == 0) {
73 while (*name
&& isspace(*name
))
75 size
= strlen(name
) - 1;
78 memcpy(comm
, name
, size
);
81 } else if (memcmp(bf
, "Tgid:", 5) == 0) {
83 while (*tgids
&& isspace(*tgids
))
94 static pid_t
perf_event__synthesize_comm(struct perf_tool
*tool
,
95 union perf_event
*event
, pid_t pid
,
97 perf_event__handler_t process
,
98 struct machine
*machine
)
100 char filename
[PATH_MAX
];
103 struct dirent dirent
, *next
;
106 memset(&event
->comm
, 0, sizeof(event
->comm
));
108 tgid
= perf_event__get_comm_tgid(pid
, event
->comm
.comm
,
109 sizeof(event
->comm
.comm
));
113 event
->comm
.pid
= tgid
;
114 event
->comm
.header
.type
= PERF_RECORD_COMM
;
116 size
= strlen(event
->comm
.comm
) + 1;
117 size
= PERF_ALIGN(size
, sizeof(u64
));
118 memset(event
->comm
.comm
+ size
, 0, machine
->id_hdr_size
);
119 event
->comm
.header
.size
= (sizeof(event
->comm
) -
120 (sizeof(event
->comm
.comm
) - size
) +
121 machine
->id_hdr_size
);
123 event
->comm
.tid
= pid
;
125 if (process(tool
, event
, &synth_sample
, machine
) != 0)
131 snprintf(filename
, sizeof(filename
), "/proc/%d/task", pid
);
133 tasks
= opendir(filename
);
135 pr_debug("couldn't open %s\n", filename
);
139 while (!readdir_r(tasks
, &dirent
, &next
) && next
) {
141 pid
= strtol(dirent
.d_name
, &end
, 10);
145 /* already have tgid; jut want to update the comm */
146 (void) perf_event__get_comm_tgid(pid
, event
->comm
.comm
,
147 sizeof(event
->comm
.comm
));
149 size
= strlen(event
->comm
.comm
) + 1;
150 size
= PERF_ALIGN(size
, sizeof(u64
));
151 memset(event
->comm
.comm
+ size
, 0, machine
->id_hdr_size
);
152 event
->comm
.header
.size
= (sizeof(event
->comm
) -
153 (sizeof(event
->comm
.comm
) - size
) +
154 machine
->id_hdr_size
);
156 event
->comm
.tid
= pid
;
158 if (process(tool
, event
, &synth_sample
, machine
) != 0) {
169 static int perf_event__synthesize_mmap_events(struct perf_tool
*tool
,
170 union perf_event
*event
,
171 pid_t pid
, pid_t tgid
,
172 perf_event__handler_t process
,
173 struct machine
*machine
,
176 char filename
[PATH_MAX
];
180 snprintf(filename
, sizeof(filename
), "/proc/%d/maps", pid
);
182 fp
= fopen(filename
, "r");
185 * We raced with a task exiting - just return:
187 pr_debug("couldn't open %s\n", filename
);
191 event
->header
.type
= PERF_RECORD_MMAP
;
196 char execname
[PATH_MAX
];
197 char anonstr
[] = "//anon";
201 if (fgets(bf
, sizeof(bf
), fp
) == NULL
)
204 /* ensure null termination since stack will be reused. */
205 strcpy(execname
, "");
207 /* 00400000-0040c000 r-xp 00000000 fd:01 41038 /bin/cat */
208 n
= sscanf(bf
, "%"PRIx64
"-%"PRIx64
" %s %"PRIx64
" %*x:%*x %*u %s\n",
209 &event
->mmap
.start
, &event
->mmap
.len
, prot
,
213 * Anon maps don't have the execname.
218 * Just like the kernel, see __perf_event_mmap in kernel/perf_event.c
220 event
->header
.misc
= PERF_RECORD_MISC_USER
;
222 if (prot
[2] != 'x') {
223 if (!mmap_data
|| prot
[0] != 'r')
226 event
->header
.misc
|= PERF_RECORD_MISC_MMAP_DATA
;
229 if (!strcmp(execname
, ""))
230 strcpy(execname
, anonstr
);
232 size
= strlen(execname
) + 1;
233 memcpy(event
->mmap
.filename
, execname
, size
);
234 size
= PERF_ALIGN(size
, sizeof(u64
));
235 event
->mmap
.len
-= event
->mmap
.start
;
236 event
->mmap
.header
.size
= (sizeof(event
->mmap
) -
237 (sizeof(event
->mmap
.filename
) - size
));
238 memset(event
->mmap
.filename
+ size
, 0, machine
->id_hdr_size
);
239 event
->mmap
.header
.size
+= machine
->id_hdr_size
;
240 event
->mmap
.pid
= tgid
;
241 event
->mmap
.tid
= pid
;
243 if (process(tool
, event
, &synth_sample
, machine
) != 0) {
253 int perf_event__synthesize_modules(struct perf_tool
*tool
,
254 perf_event__handler_t process
,
255 struct machine
*machine
)
259 struct map_groups
*kmaps
= &machine
->kmaps
;
260 union perf_event
*event
= zalloc((sizeof(event
->mmap
) +
261 machine
->id_hdr_size
));
263 pr_debug("Not enough memory synthesizing mmap event "
264 "for kernel modules\n");
268 event
->header
.type
= PERF_RECORD_MMAP
;
271 * kernel uses 0 for user space maps, see kernel/perf_event.c
274 if (machine__is_host(machine
))
275 event
->header
.misc
= PERF_RECORD_MISC_KERNEL
;
277 event
->header
.misc
= PERF_RECORD_MISC_GUEST_KERNEL
;
279 for (nd
= rb_first(&kmaps
->maps
[MAP__FUNCTION
]);
280 nd
; nd
= rb_next(nd
)) {
282 struct map
*pos
= rb_entry(nd
, struct map
, rb_node
);
284 if (pos
->dso
->kernel
)
287 size
= PERF_ALIGN(pos
->dso
->long_name_len
+ 1, sizeof(u64
));
288 event
->mmap
.header
.type
= PERF_RECORD_MMAP
;
289 event
->mmap
.header
.size
= (sizeof(event
->mmap
) -
290 (sizeof(event
->mmap
.filename
) - size
));
291 memset(event
->mmap
.filename
+ size
, 0, machine
->id_hdr_size
);
292 event
->mmap
.header
.size
+= machine
->id_hdr_size
;
293 event
->mmap
.start
= pos
->start
;
294 event
->mmap
.len
= pos
->end
- pos
->start
;
295 event
->mmap
.pid
= machine
->pid
;
297 memcpy(event
->mmap
.filename
, pos
->dso
->long_name
,
298 pos
->dso
->long_name_len
+ 1);
299 if (process(tool
, event
, &synth_sample
, machine
) != 0) {
309 static int __event__synthesize_thread(union perf_event
*comm_event
,
310 union perf_event
*mmap_event
,
312 perf_event__handler_t process
,
313 struct perf_tool
*tool
,
314 struct machine
*machine
, bool mmap_data
)
316 pid_t tgid
= perf_event__synthesize_comm(tool
, comm_event
, pid
, full
,
320 return perf_event__synthesize_mmap_events(tool
, mmap_event
, pid
, tgid
,
321 process
, machine
, mmap_data
);
324 int perf_event__synthesize_thread_map(struct perf_tool
*tool
,
325 struct thread_map
*threads
,
326 perf_event__handler_t process
,
327 struct machine
*machine
,
330 union perf_event
*comm_event
, *mmap_event
;
331 int err
= -1, thread
, j
;
333 comm_event
= malloc(sizeof(comm_event
->comm
) + machine
->id_hdr_size
);
334 if (comm_event
== NULL
)
337 mmap_event
= malloc(sizeof(mmap_event
->mmap
) + machine
->id_hdr_size
);
338 if (mmap_event
== NULL
)
342 for (thread
= 0; thread
< threads
->nr
; ++thread
) {
343 if (__event__synthesize_thread(comm_event
, mmap_event
,
344 threads
->map
[thread
], 0,
345 process
, tool
, machine
,
352 * comm.pid is set to thread group id by
353 * perf_event__synthesize_comm
355 if ((int) comm_event
->comm
.pid
!= threads
->map
[thread
]) {
356 bool need_leader
= true;
358 /* is thread group leader in thread_map? */
359 for (j
= 0; j
< threads
->nr
; ++j
) {
360 if ((int) comm_event
->comm
.pid
== threads
->map
[j
]) {
366 /* if not, generate events for it */
368 __event__synthesize_thread(comm_event
, mmap_event
,
369 comm_event
->comm
.pid
, 0,
370 process
, tool
, machine
,
384 int perf_event__synthesize_threads(struct perf_tool
*tool
,
385 perf_event__handler_t process
,
386 struct machine
*machine
, bool mmap_data
)
389 struct dirent dirent
, *next
;
390 union perf_event
*comm_event
, *mmap_event
;
393 comm_event
= malloc(sizeof(comm_event
->comm
) + machine
->id_hdr_size
);
394 if (comm_event
== NULL
)
397 mmap_event
= malloc(sizeof(mmap_event
->mmap
) + machine
->id_hdr_size
);
398 if (mmap_event
== NULL
)
401 proc
= opendir("/proc");
405 while (!readdir_r(proc
, &dirent
, &next
) && next
) {
407 pid_t pid
= strtol(dirent
.d_name
, &end
, 10);
409 if (*end
) /* only interested in proper numerical dirents */
412 * We may race with exiting thread, so don't stop just because
413 * one thread couldn't be synthesized.
415 __event__synthesize_thread(comm_event
, mmap_event
, pid
, 1,
416 process
, tool
, machine
, mmap_data
);
429 struct process_symbol_args
{
434 static int find_symbol_cb(void *arg
, const char *name
, char type
,
437 struct process_symbol_args
*args
= arg
;
440 * Must be a function or at least an alias, as in PARISC64, where "_text" is
441 * an 'A' to the same address as "_stext".
443 if (!(symbol_type__is_a(type
, MAP__FUNCTION
) ||
444 type
== 'A') || strcmp(name
, args
->name
))
451 int perf_event__synthesize_kernel_mmap(struct perf_tool
*tool
,
452 perf_event__handler_t process
,
453 struct machine
*machine
,
454 const char *symbol_name
)
457 const char *filename
, *mmap_name
;
459 char name_buff
[PATH_MAX
];
463 * We should get this from /sys/kernel/sections/.text, but till that is
464 * available use this, and after it is use this as a fallback for older
467 struct process_symbol_args args
= { .name
= symbol_name
, };
468 union perf_event
*event
= zalloc((sizeof(event
->mmap
) +
469 machine
->id_hdr_size
));
471 pr_debug("Not enough memory synthesizing mmap event "
472 "for kernel modules\n");
476 mmap_name
= machine__mmap_name(machine
, name_buff
, sizeof(name_buff
));
477 if (machine__is_host(machine
)) {
479 * kernel uses PERF_RECORD_MISC_USER for user space maps,
480 * see kernel/perf_event.c __perf_event_mmap
482 event
->header
.misc
= PERF_RECORD_MISC_KERNEL
;
483 filename
= "/proc/kallsyms";
485 event
->header
.misc
= PERF_RECORD_MISC_GUEST_KERNEL
;
486 if (machine__is_default_guest(machine
))
487 filename
= (char *) symbol_conf
.default_guest_kallsyms
;
489 sprintf(path
, "%s/proc/kallsyms", machine
->root_dir
);
494 if (kallsyms__parse(filename
, &args
, find_symbol_cb
) <= 0) {
499 map
= machine
->vmlinux_maps
[MAP__FUNCTION
];
500 size
= snprintf(event
->mmap
.filename
, sizeof(event
->mmap
.filename
),
501 "%s%s", mmap_name
, symbol_name
) + 1;
502 size
= PERF_ALIGN(size
, sizeof(u64
));
503 event
->mmap
.header
.type
= PERF_RECORD_MMAP
;
504 event
->mmap
.header
.size
= (sizeof(event
->mmap
) -
505 (sizeof(event
->mmap
.filename
) - size
) + machine
->id_hdr_size
);
506 event
->mmap
.pgoff
= args
.start
;
507 event
->mmap
.start
= map
->start
;
508 event
->mmap
.len
= map
->end
- event
->mmap
.start
;
509 event
->mmap
.pid
= machine
->pid
;
511 err
= process(tool
, event
, &synth_sample
, machine
);
517 size_t perf_event__fprintf_comm(union perf_event
*event
, FILE *fp
)
519 return fprintf(fp
, ": %s:%d\n", event
->comm
.comm
, event
->comm
.tid
);
522 int perf_event__process_comm(struct perf_tool
*tool __maybe_unused
,
523 union perf_event
*event
,
524 struct perf_sample
*sample
,
525 struct machine
*machine
)
527 return machine__process_comm_event(machine
, event
, sample
);
530 int perf_event__process_lost(struct perf_tool
*tool __maybe_unused
,
531 union perf_event
*event
,
532 struct perf_sample
*sample
,
533 struct machine
*machine
)
535 return machine__process_lost_event(machine
, event
, sample
);
538 size_t perf_event__fprintf_mmap(union perf_event
*event
, FILE *fp
)
540 return fprintf(fp
, " %d/%d: [%#" PRIx64
"(%#" PRIx64
") @ %#" PRIx64
"]: %c %s\n",
541 event
->mmap
.pid
, event
->mmap
.tid
, event
->mmap
.start
,
542 event
->mmap
.len
, event
->mmap
.pgoff
,
543 (event
->header
.misc
& PERF_RECORD_MISC_MMAP_DATA
) ? 'r' : 'x',
544 event
->mmap
.filename
);
547 size_t perf_event__fprintf_mmap2(union perf_event
*event
, FILE *fp
)
549 return fprintf(fp
, " %d/%d: [%#" PRIx64
"(%#" PRIx64
") @ %#" PRIx64
550 " %02x:%02x %"PRIu64
" %"PRIu64
"]: %c %s\n",
551 event
->mmap2
.pid
, event
->mmap2
.tid
, event
->mmap2
.start
,
552 event
->mmap2
.len
, event
->mmap2
.pgoff
, event
->mmap2
.maj
,
553 event
->mmap2
.min
, event
->mmap2
.ino
,
554 event
->mmap2
.ino_generation
,
555 (event
->header
.misc
& PERF_RECORD_MISC_MMAP_DATA
) ? 'r' : 'x',
556 event
->mmap2
.filename
);
559 int perf_event__process_mmap(struct perf_tool
*tool __maybe_unused
,
560 union perf_event
*event
,
561 struct perf_sample
*sample
,
562 struct machine
*machine
)
564 return machine__process_mmap_event(machine
, event
, sample
);
567 int perf_event__process_mmap2(struct perf_tool
*tool __maybe_unused
,
568 union perf_event
*event
,
569 struct perf_sample
*sample
,
570 struct machine
*machine
)
572 return machine__process_mmap2_event(machine
, event
, sample
);
575 size_t perf_event__fprintf_task(union perf_event
*event
, FILE *fp
)
577 return fprintf(fp
, "(%d:%d):(%d:%d)\n",
578 event
->fork
.pid
, event
->fork
.tid
,
579 event
->fork
.ppid
, event
->fork
.ptid
);
582 int perf_event__process_fork(struct perf_tool
*tool __maybe_unused
,
583 union perf_event
*event
,
584 struct perf_sample
*sample
,
585 struct machine
*machine
)
587 return machine__process_fork_event(machine
, event
, sample
);
590 int perf_event__process_exit(struct perf_tool
*tool __maybe_unused
,
591 union perf_event
*event
,
592 struct perf_sample
*sample
,
593 struct machine
*machine
)
595 return machine__process_exit_event(machine
, event
, sample
);
598 size_t perf_event__fprintf(union perf_event
*event
, FILE *fp
)
600 size_t ret
= fprintf(fp
, "PERF_RECORD_%s",
601 perf_event__name(event
->header
.type
));
603 switch (event
->header
.type
) {
604 case PERF_RECORD_COMM
:
605 ret
+= perf_event__fprintf_comm(event
, fp
);
607 case PERF_RECORD_FORK
:
608 case PERF_RECORD_EXIT
:
609 ret
+= perf_event__fprintf_task(event
, fp
);
611 case PERF_RECORD_MMAP
:
612 ret
+= perf_event__fprintf_mmap(event
, fp
);
614 case PERF_RECORD_MMAP2
:
615 ret
+= perf_event__fprintf_mmap2(event
, fp
);
618 ret
+= fprintf(fp
, "\n");
624 int perf_event__process(struct perf_tool
*tool __maybe_unused
,
625 union perf_event
*event
,
626 struct perf_sample
*sample
,
627 struct machine
*machine
)
629 return machine__process_event(machine
, event
, sample
);
632 void thread__find_addr_map(struct thread
*thread
,
633 struct machine
*machine
, u8 cpumode
,
634 enum map_type type
, u64 addr
,
635 struct addr_location
*al
)
637 struct map_groups
*mg
= &thread
->mg
;
638 bool load_map
= false;
642 al
->cpumode
= cpumode
;
643 al
->filtered
= false;
645 if (machine
== NULL
) {
650 if (cpumode
== PERF_RECORD_MISC_KERNEL
&& perf_host
) {
652 mg
= &machine
->kmaps
;
654 } else if (cpumode
== PERF_RECORD_MISC_USER
&& perf_host
) {
656 } else if (cpumode
== PERF_RECORD_MISC_GUEST_KERNEL
&& perf_guest
) {
658 mg
= &machine
->kmaps
;
662 * 'u' means guest os user space.
663 * TODO: We don't support guest user space. Might support late.
665 if (cpumode
== PERF_RECORD_MISC_GUEST_USER
&& perf_guest
)
671 if ((cpumode
== PERF_RECORD_MISC_GUEST_USER
||
672 cpumode
== PERF_RECORD_MISC_GUEST_KERNEL
) &&
675 if ((cpumode
== PERF_RECORD_MISC_USER
||
676 cpumode
== PERF_RECORD_MISC_KERNEL
) &&
683 al
->map
= map_groups__find(mg
, type
, al
->addr
);
684 if (al
->map
== NULL
) {
686 * If this is outside of all known maps, and is a negative
687 * address, try to look it up in the kernel dso, as it might be
688 * a vsyscall or vdso (which executes in user-mode).
690 * XXX This is nasty, we should have a symbol list in the
691 * "[vdso]" dso, but for now lets use the old trick of looking
692 * in the whole kernel symbol list.
694 if ((long long)al
->addr
< 0 &&
695 cpumode
== PERF_RECORD_MISC_USER
&&
696 machine
&& mg
!= &machine
->kmaps
) {
697 mg
= &machine
->kmaps
;
702 * Kernel maps might be changed when loading symbols so loading
703 * must be done prior to using kernel maps.
706 map__load(al
->map
, machine
->symbol_filter
);
707 al
->addr
= al
->map
->map_ip(al
->map
, al
->addr
);
711 void thread__find_addr_location(struct thread
*thread
, struct machine
*machine
,
712 u8 cpumode
, enum map_type type
, u64 addr
,
713 struct addr_location
*al
)
715 thread__find_addr_map(thread
, machine
, cpumode
, type
, addr
, al
);
717 al
->sym
= map__find_symbol(al
->map
, al
->addr
,
718 machine
->symbol_filter
);
723 int perf_event__preprocess_sample(const union perf_event
*event
,
724 struct machine
*machine
,
725 struct addr_location
*al
,
726 struct perf_sample
*sample
)
728 u8 cpumode
= event
->header
.misc
& PERF_RECORD_MISC_CPUMODE_MASK
;
729 struct thread
*thread
= machine__findnew_thread(machine
, sample
->pid
,
735 if (symbol_conf
.comm_list
&&
736 !strlist__has_entry(symbol_conf
.comm_list
, thread__comm_str(thread
)))
739 dump_printf(" ... thread: %s:%d\n", thread__comm_str(thread
), thread
->tid
);
741 * Have we already created the kernel maps for this machine?
743 * This should have happened earlier, when we processed the kernel MMAP
744 * events, but for older perf.data files there was no such thing, so do
747 if (cpumode
== PERF_RECORD_MISC_KERNEL
&&
748 machine
->vmlinux_maps
[MAP__FUNCTION
] == NULL
)
749 machine__create_kernel_maps(machine
);
751 thread__find_addr_map(thread
, machine
, cpumode
, MAP__FUNCTION
,
753 dump_printf(" ...... dso: %s\n",
754 al
->map
? al
->map
->dso
->long_name
:
755 al
->level
== 'H' ? "[hypervisor]" : "<not found>");
757 al
->cpu
= sample
->cpu
;
760 struct dso
*dso
= al
->map
->dso
;
762 if (symbol_conf
.dso_list
&&
763 (!dso
|| !(strlist__has_entry(symbol_conf
.dso_list
,
765 (dso
->short_name
!= dso
->long_name
&&
766 strlist__has_entry(symbol_conf
.dso_list
,
770 al
->sym
= map__find_symbol(al
->map
, al
->addr
,
771 machine
->symbol_filter
);
774 if (symbol_conf
.sym_list
&&
775 (!al
->sym
|| !strlist__has_entry(symbol_conf
.sym_list
,