1 // SPDX-License-Identifier: GPL-2.0-only
3 #include "util/debug.h"
5 #include "util/event.h"
6 #include "util/evlist.h"
7 #include "util/machine.h"
9 #include "util/map_symbol.h"
10 #include "util/branch.h"
11 #include "util/memswap.h"
12 #include "util/namespaces.h"
13 #include "util/session.h"
14 #include "util/stat.h"
15 #include "util/symbol.h"
16 #include "util/synthetic-events.h"
17 #include "util/target.h"
18 #include "util/time-utils.h"
19 #include "util/cgroup.h"
20 #include <linux/bitops.h>
21 #include <linux/kernel.h>
22 #include <linux/string.h>
23 #include <linux/zalloc.h>
24 #include <linux/perf_event.h>
26 #include <perf/evsel.h>
27 #include <internal/cpumap.h>
28 #include <perf/cpumap.h>
29 #include <internal/lib.h> // page_size
30 #include <internal/threadmap.h>
31 #include <perf/threadmap.h>
32 #include <symbol/kallsyms.h>
38 #include <uapi/linux/mman.h> /* To get things like MAP_HUGETLB even on older libc headers */
39 #include <api/fs/fs.h>
41 #include <sys/types.h>
46 #define DEFAULT_PROC_MAP_PARSE_TIMEOUT 500
48 unsigned int proc_map_timeout
= DEFAULT_PROC_MAP_PARSE_TIMEOUT
;
50 int perf_tool__process_synth_event(struct perf_tool
*tool
,
51 union perf_event
*event
,
52 struct machine
*machine
,
53 perf_event__handler_t process
)
55 struct perf_sample synth_sample
= {
62 .cpumode
= event
->header
.misc
& PERF_RECORD_MISC_CPUMODE_MASK
,
65 return process(tool
, event
, &synth_sample
, machine
);
69 * Assumes that the first 4095 bytes of /proc/pid/stat contains
70 * the comm, tgid and ppid.
72 static int perf_event__get_comm_ids(pid_t pid
, char *comm
, size_t len
,
73 pid_t
*tgid
, pid_t
*ppid
)
79 char *name
, *tgids
, *ppids
;
84 snprintf(bf
, sizeof(bf
), "/proc/%d/status", pid
);
86 fd
= open(bf
, O_RDONLY
);
88 pr_debug("couldn't open %s\n", bf
);
92 n
= read(fd
, bf
, sizeof(bf
) - 1);
95 pr_warning("Couldn't get COMM, tigd and ppid for pid %d\n",
101 name
= strstr(bf
, "Name:");
102 tgids
= strstr(bf
, "Tgid:");
103 ppids
= strstr(bf
, "PPid:");
108 name
= skip_spaces(name
+ 5); /* strlen("Name:") */
109 nl
= strchr(name
, '\n');
116 memcpy(comm
, name
, size
);
119 pr_debug("Name: string not found for pid %d\n", pid
);
123 tgids
+= 5; /* strlen("Tgid:") */
126 pr_debug("Tgid: string not found for pid %d\n", pid
);
130 ppids
+= 5; /* strlen("PPid:") */
133 pr_debug("PPid: string not found for pid %d\n", pid
);
139 static int perf_event__prepare_comm(union perf_event
*event
, pid_t pid
,
140 struct machine
*machine
,
141 pid_t
*tgid
, pid_t
*ppid
)
147 memset(&event
->comm
, 0, sizeof(event
->comm
));
149 if (machine__is_host(machine
)) {
150 if (perf_event__get_comm_ids(pid
, event
->comm
.comm
,
151 sizeof(event
->comm
.comm
),
156 *tgid
= machine
->pid
;
162 event
->comm
.pid
= *tgid
;
163 event
->comm
.header
.type
= PERF_RECORD_COMM
;
165 size
= strlen(event
->comm
.comm
) + 1;
166 size
= PERF_ALIGN(size
, sizeof(u64
));
167 memset(event
->comm
.comm
+ size
, 0, machine
->id_hdr_size
);
168 event
->comm
.header
.size
= (sizeof(event
->comm
) -
169 (sizeof(event
->comm
.comm
) - size
) +
170 machine
->id_hdr_size
);
171 event
->comm
.tid
= pid
;
176 pid_t
perf_event__synthesize_comm(struct perf_tool
*tool
,
177 union perf_event
*event
, pid_t pid
,
178 perf_event__handler_t process
,
179 struct machine
*machine
)
183 if (perf_event__prepare_comm(event
, pid
, machine
, &tgid
, &ppid
) != 0)
186 if (perf_tool__process_synth_event(tool
, event
, machine
, process
) != 0)
192 static void perf_event__get_ns_link_info(pid_t pid
, const char *ns
,
193 struct perf_ns_link_info
*ns_link_info
)
198 sprintf(proc_ns
, "/proc/%u/ns/%s", pid
, ns
);
199 if (stat64(proc_ns
, &st
) == 0) {
200 ns_link_info
->dev
= st
.st_dev
;
201 ns_link_info
->ino
= st
.st_ino
;
205 int perf_event__synthesize_namespaces(struct perf_tool
*tool
,
206 union perf_event
*event
,
207 pid_t pid
, pid_t tgid
,
208 perf_event__handler_t process
,
209 struct machine
*machine
)
212 struct perf_ns_link_info
*ns_link_info
;
214 if (!tool
|| !tool
->namespace_events
)
217 memset(&event
->namespaces
, 0, (sizeof(event
->namespaces
) +
218 (NR_NAMESPACES
* sizeof(struct perf_ns_link_info
)) +
219 machine
->id_hdr_size
));
221 event
->namespaces
.pid
= tgid
;
222 event
->namespaces
.tid
= pid
;
224 event
->namespaces
.nr_namespaces
= NR_NAMESPACES
;
226 ns_link_info
= event
->namespaces
.link_info
;
228 for (idx
= 0; idx
< event
->namespaces
.nr_namespaces
; idx
++)
229 perf_event__get_ns_link_info(pid
, perf_ns__name(idx
),
232 event
->namespaces
.header
.type
= PERF_RECORD_NAMESPACES
;
234 event
->namespaces
.header
.size
= (sizeof(event
->namespaces
) +
235 (NR_NAMESPACES
* sizeof(struct perf_ns_link_info
)) +
236 machine
->id_hdr_size
);
238 if (perf_tool__process_synth_event(tool
, event
, machine
, process
) != 0)
244 static int perf_event__synthesize_fork(struct perf_tool
*tool
,
245 union perf_event
*event
,
246 pid_t pid
, pid_t tgid
, pid_t ppid
,
247 perf_event__handler_t process
,
248 struct machine
*machine
)
250 memset(&event
->fork
, 0, sizeof(event
->fork
) + machine
->id_hdr_size
);
253 * for main thread set parent to ppid from status file. For other
254 * threads set parent pid to main thread. ie., assume main thread
255 * spawns all threads in a process
258 event
->fork
.ppid
= ppid
;
259 event
->fork
.ptid
= ppid
;
261 event
->fork
.ppid
= tgid
;
262 event
->fork
.ptid
= tgid
;
264 event
->fork
.pid
= tgid
;
265 event
->fork
.tid
= pid
;
266 event
->fork
.header
.type
= PERF_RECORD_FORK
;
267 event
->fork
.header
.misc
= PERF_RECORD_MISC_FORK_EXEC
;
269 event
->fork
.header
.size
= (sizeof(event
->fork
) + machine
->id_hdr_size
);
271 if (perf_tool__process_synth_event(tool
, event
, machine
, process
) != 0)
277 static bool read_proc_maps_line(struct io
*io
, __u64
*start
, __u64
*end
,
278 u32
*prot
, u32
*flags
, __u64
*offset
,
281 ssize_t pathname_size
, char *pathname
)
285 char *start_pathname
= pathname
;
287 if (io__get_hex(io
, start
) != '-')
289 if (io__get_hex(io
, end
) != ' ')
292 /* map protection and flags bits */
294 ch
= io__get_char(io
);
299 ch
= io__get_char(io
);
304 ch
= io__get_char(io
);
309 ch
= io__get_char(io
);
313 *flags
= MAP_PRIVATE
;
316 if (io__get_char(io
) != ' ')
319 if (io__get_hex(io
, offset
) != ' ')
322 if (io__get_hex(io
, &temp
) != ':')
325 if (io__get_hex(io
, &temp
) != ' ')
329 ch
= io__get_dec(io
, inode
);
335 ch
= io__get_char(io
);
340 if (ch
== '\0' || ch
== '\n' ||
341 (pathname
+ 1 - start_pathname
) >= pathname_size
) {
346 ch
= io__get_char(io
);
350 int perf_event__synthesize_mmap_events(struct perf_tool
*tool
,
351 union perf_event
*event
,
352 pid_t pid
, pid_t tgid
,
353 perf_event__handler_t process
,
354 struct machine
*machine
,
357 unsigned long long t
;
360 bool truncation
= false;
361 unsigned long long timeout
= proc_map_timeout
* 1000000ULL;
363 const char *hugetlbfs_mnt
= hugetlbfs__mountpoint();
364 int hugetlbfs_mnt_len
= hugetlbfs_mnt
? strlen(hugetlbfs_mnt
) : 0;
366 if (machine__is_default_guest(machine
))
369 snprintf(bf
, sizeof(bf
), "%s/proc/%d/task/%d/maps",
370 machine
->root_dir
, pid
, pid
);
372 io
.fd
= open(bf
, O_RDONLY
, 0);
375 * We raced with a task exiting - just return:
377 pr_debug("couldn't open %s\n", bf
);
380 io__init(&io
, io
.fd
, bf
, sizeof(bf
));
382 event
->header
.type
= PERF_RECORD_MMAP2
;
386 static const char anonstr
[] = "//anon";
389 /* ensure null termination since stack will be reused. */
390 event
->mmap2
.filename
[0] = '\0';
392 /* 00400000-0040c000 r-xp 00000000 fd:01 41038 /bin/cat */
393 if (!read_proc_maps_line(&io
,
402 sizeof(event
->mmap2
.filename
),
403 event
->mmap2
.filename
))
406 if ((rdclock() - t
) > timeout
) {
407 pr_warning("Reading %s/proc/%d/task/%d/maps time out. "
408 "You may want to increase "
409 "the time limit by --proc-map-timeout\n",
410 machine
->root_dir
, pid
, pid
);
415 event
->mmap2
.ino_generation
= 0;
418 * Just like the kernel, see __perf_event_mmap in kernel/perf_event.c
420 if (machine__is_host(machine
))
421 event
->header
.misc
= PERF_RECORD_MISC_USER
;
423 event
->header
.misc
= PERF_RECORD_MISC_GUEST_USER
;
425 if ((event
->mmap2
.prot
& PROT_EXEC
) == 0) {
426 if (!mmap_data
|| (event
->mmap2
.prot
& PROT_READ
) == 0)
429 event
->header
.misc
|= PERF_RECORD_MISC_MMAP_DATA
;
434 event
->header
.misc
|= PERF_RECORD_MISC_PROC_MAP_PARSE_TIMEOUT
;
436 if (!strcmp(event
->mmap2
.filename
, ""))
437 strcpy(event
->mmap2
.filename
, anonstr
);
439 if (hugetlbfs_mnt_len
&&
440 !strncmp(event
->mmap2
.filename
, hugetlbfs_mnt
,
441 hugetlbfs_mnt_len
)) {
442 strcpy(event
->mmap2
.filename
, anonstr
);
443 event
->mmap2
.flags
|= MAP_HUGETLB
;
446 size
= strlen(event
->mmap2
.filename
) + 1;
447 size
= PERF_ALIGN(size
, sizeof(u64
));
448 event
->mmap2
.len
-= event
->mmap
.start
;
449 event
->mmap2
.header
.size
= (sizeof(event
->mmap2
) -
450 (sizeof(event
->mmap2
.filename
) - size
));
451 memset(event
->mmap2
.filename
+ size
, 0, machine
->id_hdr_size
);
452 event
->mmap2
.header
.size
+= machine
->id_hdr_size
;
453 event
->mmap2
.pid
= tgid
;
454 event
->mmap2
.tid
= pid
;
456 if (perf_tool__process_synth_event(tool
, event
, machine
, process
) != 0) {
469 #ifdef HAVE_FILE_HANDLE
470 static int perf_event__synthesize_cgroup(struct perf_tool
*tool
,
471 union perf_event
*event
,
472 char *path
, size_t mount_len
,
473 perf_event__handler_t process
,
474 struct machine
*machine
)
476 size_t event_size
= sizeof(event
->cgroup
) - sizeof(event
->cgroup
.path
);
477 size_t path_len
= strlen(path
) - mount_len
+ 1;
479 struct file_handle fh
;
484 while (path_len
% sizeof(u64
))
485 path
[mount_len
+ path_len
++] = '\0';
487 memset(&event
->cgroup
, 0, event_size
);
489 event
->cgroup
.header
.type
= PERF_RECORD_CGROUP
;
490 event
->cgroup
.header
.size
= event_size
+ path_len
+ machine
->id_hdr_size
;
492 handle
.fh
.handle_bytes
= sizeof(handle
.cgroup_id
);
493 if (name_to_handle_at(AT_FDCWD
, path
, &handle
.fh
, &mount_id
, 0) < 0) {
494 pr_debug("stat failed: %s\n", path
);
498 event
->cgroup
.id
= handle
.cgroup_id
;
499 strncpy(event
->cgroup
.path
, path
+ mount_len
, path_len
);
500 memset(event
->cgroup
.path
+ path_len
, 0, machine
->id_hdr_size
);
502 if (perf_tool__process_synth_event(tool
, event
, machine
, process
) < 0) {
503 pr_debug("process synth event failed\n");
510 static int perf_event__walk_cgroup_tree(struct perf_tool
*tool
,
511 union perf_event
*event
,
512 char *path
, size_t mount_len
,
513 perf_event__handler_t process
,
514 struct machine
*machine
)
516 size_t pos
= strlen(path
);
521 if (perf_event__synthesize_cgroup(tool
, event
, path
, mount_len
,
522 process
, machine
) < 0)
527 pr_debug("failed to open directory: %s\n", path
);
531 while ((dent
= readdir(d
)) != NULL
) {
532 if (dent
->d_type
!= DT_DIR
)
534 if (!strcmp(dent
->d_name
, ".") ||
535 !strcmp(dent
->d_name
, ".."))
538 /* any sane path should be less than PATH_MAX */
539 if (strlen(path
) + strlen(dent
->d_name
) + 1 >= PATH_MAX
)
542 if (path
[pos
- 1] != '/')
544 strcat(path
, dent
->d_name
);
546 ret
= perf_event__walk_cgroup_tree(tool
, event
, path
,
547 mount_len
, process
, machine
);
558 int perf_event__synthesize_cgroups(struct perf_tool
*tool
,
559 perf_event__handler_t process
,
560 struct machine
*machine
)
562 union perf_event event
;
563 char cgrp_root
[PATH_MAX
];
564 size_t mount_len
; /* length of mount point in the path */
566 if (cgroupfs_find_mountpoint(cgrp_root
, PATH_MAX
, "perf_event") < 0) {
567 pr_debug("cannot find cgroup mount point\n");
571 mount_len
= strlen(cgrp_root
);
572 /* make sure the path starts with a slash (after mount point) */
573 strcat(cgrp_root
, "/");
575 if (perf_event__walk_cgroup_tree(tool
, &event
, cgrp_root
, mount_len
,
576 process
, machine
) < 0)
582 int perf_event__synthesize_cgroups(struct perf_tool
*tool __maybe_unused
,
583 perf_event__handler_t process __maybe_unused
,
584 struct machine
*machine __maybe_unused
)
590 int perf_event__synthesize_modules(struct perf_tool
*tool
, perf_event__handler_t process
,
591 struct machine
*machine
)
595 struct maps
*maps
= machine__kernel_maps(machine
);
596 union perf_event
*event
= zalloc((sizeof(event
->mmap
) +
597 machine
->id_hdr_size
));
599 pr_debug("Not enough memory synthesizing mmap event "
600 "for kernel modules\n");
604 event
->header
.type
= PERF_RECORD_MMAP
;
607 * kernel uses 0 for user space maps, see kernel/perf_event.c
610 if (machine__is_host(machine
))
611 event
->header
.misc
= PERF_RECORD_MISC_KERNEL
;
613 event
->header
.misc
= PERF_RECORD_MISC_GUEST_KERNEL
;
615 maps__for_each_entry(maps
, pos
) {
618 if (!__map__is_kmodule(pos
))
621 size
= PERF_ALIGN(pos
->dso
->long_name_len
+ 1, sizeof(u64
));
622 event
->mmap
.header
.type
= PERF_RECORD_MMAP
;
623 event
->mmap
.header
.size
= (sizeof(event
->mmap
) -
624 (sizeof(event
->mmap
.filename
) - size
));
625 memset(event
->mmap
.filename
+ size
, 0, machine
->id_hdr_size
);
626 event
->mmap
.header
.size
+= machine
->id_hdr_size
;
627 event
->mmap
.start
= pos
->start
;
628 event
->mmap
.len
= pos
->end
- pos
->start
;
629 event
->mmap
.pid
= machine
->pid
;
631 memcpy(event
->mmap
.filename
, pos
->dso
->long_name
,
632 pos
->dso
->long_name_len
+ 1);
633 if (perf_tool__process_synth_event(tool
, event
, machine
, process
) != 0) {
643 static int __event__synthesize_thread(union perf_event
*comm_event
,
644 union perf_event
*mmap_event
,
645 union perf_event
*fork_event
,
646 union perf_event
*namespaces_event
,
647 pid_t pid
, int full
, perf_event__handler_t process
,
648 struct perf_tool
*tool
, struct machine
*machine
, bool mmap_data
)
650 char filename
[PATH_MAX
];
652 struct dirent
*dirent
;
656 /* special case: only send one comm event using passed in pid */
658 tgid
= perf_event__synthesize_comm(tool
, comm_event
, pid
,
664 if (perf_event__synthesize_namespaces(tool
, namespaces_event
, pid
,
665 tgid
, process
, machine
) < 0)
669 * send mmap only for thread group leader
670 * see thread__init_maps()
673 perf_event__synthesize_mmap_events(tool
, mmap_event
, pid
, tgid
,
674 process
, machine
, mmap_data
))
680 if (machine__is_default_guest(machine
))
683 snprintf(filename
, sizeof(filename
), "%s/proc/%d/task",
684 machine
->root_dir
, pid
);
686 tasks
= opendir(filename
);
688 pr_debug("couldn't open %s\n", filename
);
692 while ((dirent
= readdir(tasks
)) != NULL
) {
696 _pid
= strtol(dirent
->d_name
, &end
, 10);
701 if (perf_event__prepare_comm(comm_event
, _pid
, machine
,
705 if (perf_event__synthesize_fork(tool
, fork_event
, _pid
, tgid
,
706 ppid
, process
, machine
) < 0)
709 if (perf_event__synthesize_namespaces(tool
, namespaces_event
, _pid
,
710 tgid
, process
, machine
) < 0)
714 * Send the prepared comm event
716 if (perf_tool__process_synth_event(tool
, comm_event
, machine
, process
) != 0)
721 /* process the parent's maps too */
722 rc
= perf_event__synthesize_mmap_events(tool
, mmap_event
, pid
, tgid
,
723 process
, machine
, mmap_data
);
733 int perf_event__synthesize_thread_map(struct perf_tool
*tool
,
734 struct perf_thread_map
*threads
,
735 perf_event__handler_t process
,
736 struct machine
*machine
,
739 union perf_event
*comm_event
, *mmap_event
, *fork_event
;
740 union perf_event
*namespaces_event
;
741 int err
= -1, thread
, j
;
743 comm_event
= malloc(sizeof(comm_event
->comm
) + machine
->id_hdr_size
);
744 if (comm_event
== NULL
)
747 mmap_event
= malloc(sizeof(mmap_event
->mmap2
) + machine
->id_hdr_size
);
748 if (mmap_event
== NULL
)
751 fork_event
= malloc(sizeof(fork_event
->fork
) + machine
->id_hdr_size
);
752 if (fork_event
== NULL
)
755 namespaces_event
= malloc(sizeof(namespaces_event
->namespaces
) +
756 (NR_NAMESPACES
* sizeof(struct perf_ns_link_info
)) +
757 machine
->id_hdr_size
);
758 if (namespaces_event
== NULL
)
762 for (thread
= 0; thread
< threads
->nr
; ++thread
) {
763 if (__event__synthesize_thread(comm_event
, mmap_event
,
764 fork_event
, namespaces_event
,
765 perf_thread_map__pid(threads
, thread
), 0,
766 process
, tool
, machine
,
773 * comm.pid is set to thread group id by
774 * perf_event__synthesize_comm
776 if ((int) comm_event
->comm
.pid
!= perf_thread_map__pid(threads
, thread
)) {
777 bool need_leader
= true;
779 /* is thread group leader in thread_map? */
780 for (j
= 0; j
< threads
->nr
; ++j
) {
781 if ((int) comm_event
->comm
.pid
== perf_thread_map__pid(threads
, j
)) {
787 /* if not, generate events for it */
789 __event__synthesize_thread(comm_event
, mmap_event
,
790 fork_event
, namespaces_event
,
791 comm_event
->comm
.pid
, 0,
792 process
, tool
, machine
,
799 free(namespaces_event
);
810 static int __perf_event__synthesize_threads(struct perf_tool
*tool
,
811 perf_event__handler_t process
,
812 struct machine
*machine
,
814 struct dirent
**dirent
,
818 union perf_event
*comm_event
, *mmap_event
, *fork_event
;
819 union perf_event
*namespaces_event
;
825 comm_event
= malloc(sizeof(comm_event
->comm
) + machine
->id_hdr_size
);
826 if (comm_event
== NULL
)
829 mmap_event
= malloc(sizeof(mmap_event
->mmap2
) + machine
->id_hdr_size
);
830 if (mmap_event
== NULL
)
833 fork_event
= malloc(sizeof(fork_event
->fork
) + machine
->id_hdr_size
);
834 if (fork_event
== NULL
)
837 namespaces_event
= malloc(sizeof(namespaces_event
->namespaces
) +
838 (NR_NAMESPACES
* sizeof(struct perf_ns_link_info
)) +
839 machine
->id_hdr_size
);
840 if (namespaces_event
== NULL
)
843 for (i
= start
; i
< start
+ num
; i
++) {
844 if (!isdigit(dirent
[i
]->d_name
[0]))
847 pid
= (pid_t
)strtol(dirent
[i
]->d_name
, &end
, 10);
848 /* only interested in proper numerical dirents */
852 * We may race with exiting thread, so don't stop just because
853 * one thread couldn't be synthesized.
855 __event__synthesize_thread(comm_event
, mmap_event
, fork_event
,
856 namespaces_event
, pid
, 1, process
,
857 tool
, machine
, mmap_data
);
861 free(namespaces_event
);
872 struct synthesize_threads_arg
{
873 struct perf_tool
*tool
;
874 perf_event__handler_t process
;
875 struct machine
*machine
;
877 struct dirent
**dirent
;
882 static void *synthesize_threads_worker(void *arg
)
884 struct synthesize_threads_arg
*args
= arg
;
886 __perf_event__synthesize_threads(args
->tool
, args
->process
,
887 args
->machine
, args
->mmap_data
,
889 args
->start
, args
->num
);
893 int perf_event__synthesize_threads(struct perf_tool
*tool
,
894 perf_event__handler_t process
,
895 struct machine
*machine
,
897 unsigned int nr_threads_synthesize
)
899 struct synthesize_threads_arg
*args
= NULL
;
900 pthread_t
*synthesize_threads
= NULL
;
901 char proc_path
[PATH_MAX
];
902 struct dirent
**dirent
;
910 if (machine__is_default_guest(machine
))
913 snprintf(proc_path
, sizeof(proc_path
), "%s/proc", machine
->root_dir
);
914 n
= scandir(proc_path
, &dirent
, 0, alphasort
);
918 if (nr_threads_synthesize
== UINT_MAX
)
919 thread_nr
= sysconf(_SC_NPROCESSORS_ONLN
);
921 thread_nr
= nr_threads_synthesize
;
923 if (thread_nr
<= 1) {
924 err
= __perf_event__synthesize_threads(tool
, process
,
932 synthesize_threads
= calloc(sizeof(pthread_t
), thread_nr
);
933 if (synthesize_threads
== NULL
)
936 args
= calloc(sizeof(*args
), thread_nr
);
940 num_per_thread
= n
/ thread_nr
;
942 for (i
= 0; i
< thread_nr
; i
++) {
944 args
[i
].process
= process
;
945 args
[i
].machine
= machine
;
946 args
[i
].mmap_data
= mmap_data
;
947 args
[i
].dirent
= dirent
;
949 for (i
= 0; i
< m
; i
++) {
950 args
[i
].num
= num_per_thread
+ 1;
951 args
[i
].start
= i
* args
[i
].num
;
954 base
= args
[i
-1].start
+ args
[i
-1].num
;
955 for (j
= i
; j
< thread_nr
; j
++) {
956 args
[j
].num
= num_per_thread
;
957 args
[j
].start
= base
+ (j
- i
) * args
[i
].num
;
960 for (i
= 0; i
< thread_nr
; i
++) {
961 if (pthread_create(&synthesize_threads
[i
], NULL
,
962 synthesize_threads_worker
, &args
[i
]))
967 for (i
= 0; i
< thread_nr
; i
++)
968 pthread_join(synthesize_threads
[i
], NULL
);
971 free(synthesize_threads
);
973 for (i
= 0; i
< n
; i
++)
980 int __weak
perf_event__synthesize_extra_kmaps(struct perf_tool
*tool __maybe_unused
,
981 perf_event__handler_t process __maybe_unused
,
982 struct machine
*machine __maybe_unused
)
987 static int __perf_event__synthesize_kernel_mmap(struct perf_tool
*tool
,
988 perf_event__handler_t process
,
989 struct machine
*machine
)
992 struct map
*map
= machine__kernel_map(machine
);
995 union perf_event
*event
;
1000 kmap
= map__kmap(map
);
1001 if (!kmap
->ref_reloc_sym
)
1005 * We should get this from /sys/kernel/sections/.text, but till that is
1006 * available use this, and after it is use this as a fallback for older
1009 event
= zalloc((sizeof(event
->mmap
) + machine
->id_hdr_size
));
1010 if (event
== NULL
) {
1011 pr_debug("Not enough memory synthesizing mmap event "
1012 "for kernel modules\n");
1016 if (machine__is_host(machine
)) {
1018 * kernel uses PERF_RECORD_MISC_USER for user space maps,
1019 * see kernel/perf_event.c __perf_event_mmap
1021 event
->header
.misc
= PERF_RECORD_MISC_KERNEL
;
1023 event
->header
.misc
= PERF_RECORD_MISC_GUEST_KERNEL
;
1026 size
= snprintf(event
->mmap
.filename
, sizeof(event
->mmap
.filename
),
1027 "%s%s", machine
->mmap_name
, kmap
->ref_reloc_sym
->name
) + 1;
1028 size
= PERF_ALIGN(size
, sizeof(u64
));
1029 event
->mmap
.header
.type
= PERF_RECORD_MMAP
;
1030 event
->mmap
.header
.size
= (sizeof(event
->mmap
) -
1031 (sizeof(event
->mmap
.filename
) - size
) + machine
->id_hdr_size
);
1032 event
->mmap
.pgoff
= kmap
->ref_reloc_sym
->addr
;
1033 event
->mmap
.start
= map
->start
;
1034 event
->mmap
.len
= map
->end
- event
->mmap
.start
;
1035 event
->mmap
.pid
= machine
->pid
;
1037 err
= perf_tool__process_synth_event(tool
, event
, machine
, process
);
1043 int perf_event__synthesize_kernel_mmap(struct perf_tool
*tool
,
1044 perf_event__handler_t process
,
1045 struct machine
*machine
)
1049 err
= __perf_event__synthesize_kernel_mmap(tool
, process
, machine
);
1053 return perf_event__synthesize_extra_kmaps(tool
, process
, machine
);
1056 int perf_event__synthesize_thread_map2(struct perf_tool
*tool
,
1057 struct perf_thread_map
*threads
,
1058 perf_event__handler_t process
,
1059 struct machine
*machine
)
1061 union perf_event
*event
;
1064 size
= sizeof(event
->thread_map
);
1065 size
+= threads
->nr
* sizeof(event
->thread_map
.entries
[0]);
1067 event
= zalloc(size
);
1071 event
->header
.type
= PERF_RECORD_THREAD_MAP
;
1072 event
->header
.size
= size
;
1073 event
->thread_map
.nr
= threads
->nr
;
1075 for (i
= 0; i
< threads
->nr
; i
++) {
1076 struct perf_record_thread_map_entry
*entry
= &event
->thread_map
.entries
[i
];
1077 char *comm
= perf_thread_map__comm(threads
, i
);
1082 entry
->pid
= perf_thread_map__pid(threads
, i
);
1083 strncpy((char *) &entry
->comm
, comm
, sizeof(entry
->comm
));
1086 err
= process(tool
, event
, NULL
, machine
);
1092 static void synthesize_cpus(struct cpu_map_entries
*cpus
,
1093 struct perf_cpu_map
*map
)
1099 for (i
= 0; i
< map
->nr
; i
++)
1100 cpus
->cpu
[i
] = map
->map
[i
];
1103 static void synthesize_mask(struct perf_record_record_cpu_map
*mask
,
1104 struct perf_cpu_map
*map
, int max
)
1108 mask
->nr
= BITS_TO_LONGS(max
);
1109 mask
->long_size
= sizeof(long);
1111 for (i
= 0; i
< map
->nr
; i
++)
1112 set_bit(map
->map
[i
], mask
->mask
);
1115 static size_t cpus_size(struct perf_cpu_map
*map
)
1117 return sizeof(struct cpu_map_entries
) + map
->nr
* sizeof(u16
);
1120 static size_t mask_size(struct perf_cpu_map
*map
, int *max
)
1126 for (i
= 0; i
< map
->nr
; i
++) {
1127 /* bit possition of the cpu is + 1 */
1128 int bit
= map
->map
[i
] + 1;
1134 return sizeof(struct perf_record_record_cpu_map
) + BITS_TO_LONGS(*max
) * sizeof(long);
1137 void *cpu_map_data__alloc(struct perf_cpu_map
*map
, size_t *size
, u16
*type
, int *max
)
1139 size_t size_cpus
, size_mask
;
1140 bool is_dummy
= perf_cpu_map__empty(map
);
1143 * Both array and mask data have variable size based
1144 * on the number of cpus and their actual values.
1145 * The size of the 'struct perf_record_cpu_map_data' is:
1147 * array = size of 'struct cpu_map_entries' +
1148 * number of cpus * sizeof(u64)
1150 * mask = size of 'struct perf_record_record_cpu_map' +
1151 * maximum cpu bit converted to size of longs
1153 * and finaly + the size of 'struct perf_record_cpu_map_data'.
1155 size_cpus
= cpus_size(map
);
1156 size_mask
= mask_size(map
, max
);
1158 if (is_dummy
|| (size_cpus
< size_mask
)) {
1160 *type
= PERF_CPU_MAP__CPUS
;
1163 *type
= PERF_CPU_MAP__MASK
;
1166 *size
+= sizeof(struct perf_record_cpu_map_data
);
1167 *size
= PERF_ALIGN(*size
, sizeof(u64
));
1168 return zalloc(*size
);
1171 void cpu_map_data__synthesize(struct perf_record_cpu_map_data
*data
, struct perf_cpu_map
*map
,
1177 case PERF_CPU_MAP__CPUS
:
1178 synthesize_cpus((struct cpu_map_entries
*) data
->data
, map
);
1180 case PERF_CPU_MAP__MASK
:
1181 synthesize_mask((struct perf_record_record_cpu_map
*)data
->data
, map
, max
);
1187 static struct perf_record_cpu_map
*cpu_map_event__new(struct perf_cpu_map
*map
)
1189 size_t size
= sizeof(struct perf_record_cpu_map
);
1190 struct perf_record_cpu_map
*event
;
1194 event
= cpu_map_data__alloc(map
, &size
, &type
, &max
);
1198 event
->header
.type
= PERF_RECORD_CPU_MAP
;
1199 event
->header
.size
= size
;
1200 event
->data
.type
= type
;
1202 cpu_map_data__synthesize(&event
->data
, map
, type
, max
);
1206 int perf_event__synthesize_cpu_map(struct perf_tool
*tool
,
1207 struct perf_cpu_map
*map
,
1208 perf_event__handler_t process
,
1209 struct machine
*machine
)
1211 struct perf_record_cpu_map
*event
;
1214 event
= cpu_map_event__new(map
);
1218 err
= process(tool
, (union perf_event
*) event
, NULL
, machine
);
1224 int perf_event__synthesize_stat_config(struct perf_tool
*tool
,
1225 struct perf_stat_config
*config
,
1226 perf_event__handler_t process
,
1227 struct machine
*machine
)
1229 struct perf_record_stat_config
*event
;
1230 int size
, i
= 0, err
;
1232 size
= sizeof(*event
);
1233 size
+= (PERF_STAT_CONFIG_TERM__MAX
* sizeof(event
->data
[0]));
1235 event
= zalloc(size
);
1239 event
->header
.type
= PERF_RECORD_STAT_CONFIG
;
1240 event
->header
.size
= size
;
1241 event
->nr
= PERF_STAT_CONFIG_TERM__MAX
;
1243 #define ADD(__term, __val) \
1244 event->data[i].tag = PERF_STAT_CONFIG_TERM__##__term; \
1245 event->data[i].val = __val; \
1248 ADD(AGGR_MODE
, config
->aggr_mode
)
1249 ADD(INTERVAL
, config
->interval
)
1250 ADD(SCALE
, config
->scale
)
1252 WARN_ONCE(i
!= PERF_STAT_CONFIG_TERM__MAX
,
1253 "stat config terms unbalanced\n");
1256 err
= process(tool
, (union perf_event
*) event
, NULL
, machine
);
1262 int perf_event__synthesize_stat(struct perf_tool
*tool
,
1263 u32 cpu
, u32 thread
, u64 id
,
1264 struct perf_counts_values
*count
,
1265 perf_event__handler_t process
,
1266 struct machine
*machine
)
1268 struct perf_record_stat event
;
1270 event
.header
.type
= PERF_RECORD_STAT
;
1271 event
.header
.size
= sizeof(event
);
1272 event
.header
.misc
= 0;
1276 event
.thread
= thread
;
1277 event
.val
= count
->val
;
1278 event
.ena
= count
->ena
;
1279 event
.run
= count
->run
;
1281 return process(tool
, (union perf_event
*) &event
, NULL
, machine
);
1284 int perf_event__synthesize_stat_round(struct perf_tool
*tool
,
1285 u64 evtime
, u64 type
,
1286 perf_event__handler_t process
,
1287 struct machine
*machine
)
1289 struct perf_record_stat_round event
;
1291 event
.header
.type
= PERF_RECORD_STAT_ROUND
;
1292 event
.header
.size
= sizeof(event
);
1293 event
.header
.misc
= 0;
1295 event
.time
= evtime
;
1298 return process(tool
, (union perf_event
*) &event
, NULL
, machine
);
1301 size_t perf_event__sample_event_size(const struct perf_sample
*sample
, u64 type
, u64 read_format
)
1303 size_t sz
, result
= sizeof(struct perf_record_sample
);
1305 if (type
& PERF_SAMPLE_IDENTIFIER
)
1306 result
+= sizeof(u64
);
1308 if (type
& PERF_SAMPLE_IP
)
1309 result
+= sizeof(u64
);
1311 if (type
& PERF_SAMPLE_TID
)
1312 result
+= sizeof(u64
);
1314 if (type
& PERF_SAMPLE_TIME
)
1315 result
+= sizeof(u64
);
1317 if (type
& PERF_SAMPLE_ADDR
)
1318 result
+= sizeof(u64
);
1320 if (type
& PERF_SAMPLE_ID
)
1321 result
+= sizeof(u64
);
1323 if (type
& PERF_SAMPLE_STREAM_ID
)
1324 result
+= sizeof(u64
);
1326 if (type
& PERF_SAMPLE_CPU
)
1327 result
+= sizeof(u64
);
1329 if (type
& PERF_SAMPLE_PERIOD
)
1330 result
+= sizeof(u64
);
1332 if (type
& PERF_SAMPLE_READ
) {
1333 result
+= sizeof(u64
);
1334 if (read_format
& PERF_FORMAT_TOTAL_TIME_ENABLED
)
1335 result
+= sizeof(u64
);
1336 if (read_format
& PERF_FORMAT_TOTAL_TIME_RUNNING
)
1337 result
+= sizeof(u64
);
1338 /* PERF_FORMAT_ID is forced for PERF_SAMPLE_READ */
1339 if (read_format
& PERF_FORMAT_GROUP
) {
1340 sz
= sample
->read
.group
.nr
*
1341 sizeof(struct sample_read_value
);
1344 result
+= sizeof(u64
);
1348 if (type
& PERF_SAMPLE_CALLCHAIN
) {
1349 sz
= (sample
->callchain
->nr
+ 1) * sizeof(u64
);
1353 if (type
& PERF_SAMPLE_RAW
) {
1354 result
+= sizeof(u32
);
1355 result
+= sample
->raw_size
;
1358 if (type
& PERF_SAMPLE_BRANCH_STACK
) {
1359 sz
= sample
->branch_stack
->nr
* sizeof(struct branch_entry
);
1361 sz
+= 2 * sizeof(u64
);
1365 if (type
& PERF_SAMPLE_REGS_USER
) {
1366 if (sample
->user_regs
.abi
) {
1367 result
+= sizeof(u64
);
1368 sz
= hweight64(sample
->user_regs
.mask
) * sizeof(u64
);
1371 result
+= sizeof(u64
);
1375 if (type
& PERF_SAMPLE_STACK_USER
) {
1376 sz
= sample
->user_stack
.size
;
1377 result
+= sizeof(u64
);
1380 result
+= sizeof(u64
);
1384 if (type
& PERF_SAMPLE_WEIGHT
)
1385 result
+= sizeof(u64
);
1387 if (type
& PERF_SAMPLE_DATA_SRC
)
1388 result
+= sizeof(u64
);
1390 if (type
& PERF_SAMPLE_TRANSACTION
)
1391 result
+= sizeof(u64
);
1393 if (type
& PERF_SAMPLE_REGS_INTR
) {
1394 if (sample
->intr_regs
.abi
) {
1395 result
+= sizeof(u64
);
1396 sz
= hweight64(sample
->intr_regs
.mask
) * sizeof(u64
);
1399 result
+= sizeof(u64
);
1403 if (type
& PERF_SAMPLE_PHYS_ADDR
)
1404 result
+= sizeof(u64
);
1406 if (type
& PERF_SAMPLE_CGROUP
)
1407 result
+= sizeof(u64
);
1409 if (type
& PERF_SAMPLE_AUX
) {
1410 result
+= sizeof(u64
);
1411 result
+= sample
->aux_sample
.size
;
1417 int perf_event__synthesize_sample(union perf_event
*event
, u64 type
, u64 read_format
,
1418 const struct perf_sample
*sample
)
1423 * used for cross-endian analysis. See git commit 65014ab3
1424 * for why this goofiness is needed.
1428 array
= event
->sample
.array
;
1430 if (type
& PERF_SAMPLE_IDENTIFIER
) {
1431 *array
= sample
->id
;
1435 if (type
& PERF_SAMPLE_IP
) {
1436 *array
= sample
->ip
;
1440 if (type
& PERF_SAMPLE_TID
) {
1441 u
.val32
[0] = sample
->pid
;
1442 u
.val32
[1] = sample
->tid
;
1447 if (type
& PERF_SAMPLE_TIME
) {
1448 *array
= sample
->time
;
1452 if (type
& PERF_SAMPLE_ADDR
) {
1453 *array
= sample
->addr
;
1457 if (type
& PERF_SAMPLE_ID
) {
1458 *array
= sample
->id
;
1462 if (type
& PERF_SAMPLE_STREAM_ID
) {
1463 *array
= sample
->stream_id
;
1467 if (type
& PERF_SAMPLE_CPU
) {
1468 u
.val32
[0] = sample
->cpu
;
1474 if (type
& PERF_SAMPLE_PERIOD
) {
1475 *array
= sample
->period
;
1479 if (type
& PERF_SAMPLE_READ
) {
1480 if (read_format
& PERF_FORMAT_GROUP
)
1481 *array
= sample
->read
.group
.nr
;
1483 *array
= sample
->read
.one
.value
;
1486 if (read_format
& PERF_FORMAT_TOTAL_TIME_ENABLED
) {
1487 *array
= sample
->read
.time_enabled
;
1491 if (read_format
& PERF_FORMAT_TOTAL_TIME_RUNNING
) {
1492 *array
= sample
->read
.time_running
;
1496 /* PERF_FORMAT_ID is forced for PERF_SAMPLE_READ */
1497 if (read_format
& PERF_FORMAT_GROUP
) {
1498 sz
= sample
->read
.group
.nr
*
1499 sizeof(struct sample_read_value
);
1500 memcpy(array
, sample
->read
.group
.values
, sz
);
1501 array
= (void *)array
+ sz
;
1503 *array
= sample
->read
.one
.id
;
1508 if (type
& PERF_SAMPLE_CALLCHAIN
) {
1509 sz
= (sample
->callchain
->nr
+ 1) * sizeof(u64
);
1510 memcpy(array
, sample
->callchain
, sz
);
1511 array
= (void *)array
+ sz
;
1514 if (type
& PERF_SAMPLE_RAW
) {
1515 u
.val32
[0] = sample
->raw_size
;
1517 array
= (void *)array
+ sizeof(u32
);
1519 memcpy(array
, sample
->raw_data
, sample
->raw_size
);
1520 array
= (void *)array
+ sample
->raw_size
;
1523 if (type
& PERF_SAMPLE_BRANCH_STACK
) {
1524 sz
= sample
->branch_stack
->nr
* sizeof(struct branch_entry
);
1526 sz
+= 2 * sizeof(u64
);
1527 memcpy(array
, sample
->branch_stack
, sz
);
1528 array
= (void *)array
+ sz
;
1531 if (type
& PERF_SAMPLE_REGS_USER
) {
1532 if (sample
->user_regs
.abi
) {
1533 *array
++ = sample
->user_regs
.abi
;
1534 sz
= hweight64(sample
->user_regs
.mask
) * sizeof(u64
);
1535 memcpy(array
, sample
->user_regs
.regs
, sz
);
1536 array
= (void *)array
+ sz
;
1542 if (type
& PERF_SAMPLE_STACK_USER
) {
1543 sz
= sample
->user_stack
.size
;
1546 memcpy(array
, sample
->user_stack
.data
, sz
);
1547 array
= (void *)array
+ sz
;
1552 if (type
& PERF_SAMPLE_WEIGHT
) {
1553 *array
= sample
->weight
;
1557 if (type
& PERF_SAMPLE_DATA_SRC
) {
1558 *array
= sample
->data_src
;
1562 if (type
& PERF_SAMPLE_TRANSACTION
) {
1563 *array
= sample
->transaction
;
1567 if (type
& PERF_SAMPLE_REGS_INTR
) {
1568 if (sample
->intr_regs
.abi
) {
1569 *array
++ = sample
->intr_regs
.abi
;
1570 sz
= hweight64(sample
->intr_regs
.mask
) * sizeof(u64
);
1571 memcpy(array
, sample
->intr_regs
.regs
, sz
);
1572 array
= (void *)array
+ sz
;
1578 if (type
& PERF_SAMPLE_PHYS_ADDR
) {
1579 *array
= sample
->phys_addr
;
1583 if (type
& PERF_SAMPLE_CGROUP
) {
1584 *array
= sample
->cgroup
;
1588 if (type
& PERF_SAMPLE_AUX
) {
1589 sz
= sample
->aux_sample
.size
;
1591 memcpy(array
, sample
->aux_sample
.data
, sz
);
1592 array
= (void *)array
+ sz
;
1598 int perf_event__synthesize_id_index(struct perf_tool
*tool
, perf_event__handler_t process
,
1599 struct evlist
*evlist
, struct machine
*machine
)
1601 union perf_event
*ev
;
1602 struct evsel
*evsel
;
1603 size_t nr
= 0, i
= 0, sz
, max_nr
, n
;
1606 pr_debug2("Synthesizing id index\n");
1608 max_nr
= (UINT16_MAX
- sizeof(struct perf_record_id_index
)) /
1609 sizeof(struct id_index_entry
);
1611 evlist__for_each_entry(evlist
, evsel
)
1612 nr
+= evsel
->core
.ids
;
1614 n
= nr
> max_nr
? max_nr
: nr
;
1615 sz
= sizeof(struct perf_record_id_index
) + n
* sizeof(struct id_index_entry
);
1620 ev
->id_index
.header
.type
= PERF_RECORD_ID_INDEX
;
1621 ev
->id_index
.header
.size
= sz
;
1622 ev
->id_index
.nr
= n
;
1624 evlist__for_each_entry(evlist
, evsel
) {
1627 for (j
= 0; j
< evsel
->core
.ids
; j
++) {
1628 struct id_index_entry
*e
;
1629 struct perf_sample_id
*sid
;
1632 err
= process(tool
, ev
, NULL
, machine
);
1639 e
= &ev
->id_index
.entries
[i
++];
1641 e
->id
= evsel
->core
.id
[j
];
1643 sid
= perf_evlist__id2sid(evlist
, e
->id
);
1655 sz
= sizeof(struct perf_record_id_index
) + nr
* sizeof(struct id_index_entry
);
1656 ev
->id_index
.header
.size
= sz
;
1657 ev
->id_index
.nr
= nr
;
1659 err
= process(tool
, ev
, NULL
, machine
);
1666 int __machine__synthesize_threads(struct machine
*machine
, struct perf_tool
*tool
,
1667 struct target
*target
, struct perf_thread_map
*threads
,
1668 perf_event__handler_t process
, bool data_mmap
,
1669 unsigned int nr_threads_synthesize
)
1671 if (target__has_task(target
))
1672 return perf_event__synthesize_thread_map(tool
, threads
, process
, machine
, data_mmap
);
1673 else if (target__has_cpu(target
))
1674 return perf_event__synthesize_threads(tool
, process
,
1676 nr_threads_synthesize
);
1677 /* command specified */
1681 int machine__synthesize_threads(struct machine
*machine
, struct target
*target
,
1682 struct perf_thread_map
*threads
, bool data_mmap
,
1683 unsigned int nr_threads_synthesize
)
1685 return __machine__synthesize_threads(machine
, NULL
, target
, threads
,
1686 perf_event__process
, data_mmap
,
1687 nr_threads_synthesize
);
1690 static struct perf_record_event_update
*event_update_event__new(size_t size
, u64 type
, u64 id
)
1692 struct perf_record_event_update
*ev
;
1694 size
+= sizeof(*ev
);
1695 size
= PERF_ALIGN(size
, sizeof(u64
));
1699 ev
->header
.type
= PERF_RECORD_EVENT_UPDATE
;
1700 ev
->header
.size
= (u16
)size
;
1707 int perf_event__synthesize_event_update_unit(struct perf_tool
*tool
, struct evsel
*evsel
,
1708 perf_event__handler_t process
)
1710 size_t size
= strlen(evsel
->unit
);
1711 struct perf_record_event_update
*ev
;
1714 ev
= event_update_event__new(size
+ 1, PERF_EVENT_UPDATE__UNIT
, evsel
->core
.id
[0]);
1718 strlcpy(ev
->data
, evsel
->unit
, size
+ 1);
1719 err
= process(tool
, (union perf_event
*)ev
, NULL
, NULL
);
1724 int perf_event__synthesize_event_update_scale(struct perf_tool
*tool
, struct evsel
*evsel
,
1725 perf_event__handler_t process
)
1727 struct perf_record_event_update
*ev
;
1728 struct perf_record_event_update_scale
*ev_data
;
1731 ev
= event_update_event__new(sizeof(*ev_data
), PERF_EVENT_UPDATE__SCALE
, evsel
->core
.id
[0]);
1735 ev_data
= (struct perf_record_event_update_scale
*)ev
->data
;
1736 ev_data
->scale
= evsel
->scale
;
1737 err
= process(tool
, (union perf_event
*)ev
, NULL
, NULL
);
1742 int perf_event__synthesize_event_update_name(struct perf_tool
*tool
, struct evsel
*evsel
,
1743 perf_event__handler_t process
)
1745 struct perf_record_event_update
*ev
;
1746 size_t len
= strlen(evsel
->name
);
1749 ev
= event_update_event__new(len
+ 1, PERF_EVENT_UPDATE__NAME
, evsel
->core
.id
[0]);
1753 strlcpy(ev
->data
, evsel
->name
, len
+ 1);
1754 err
= process(tool
, (union perf_event
*)ev
, NULL
, NULL
);
1759 int perf_event__synthesize_event_update_cpus(struct perf_tool
*tool
, struct evsel
*evsel
,
1760 perf_event__handler_t process
)
1762 size_t size
= sizeof(struct perf_record_event_update
);
1763 struct perf_record_event_update
*ev
;
1767 if (!evsel
->core
.own_cpus
)
1770 ev
= cpu_map_data__alloc(evsel
->core
.own_cpus
, &size
, &type
, &max
);
1774 ev
->header
.type
= PERF_RECORD_EVENT_UPDATE
;
1775 ev
->header
.size
= (u16
)size
;
1776 ev
->type
= PERF_EVENT_UPDATE__CPUS
;
1777 ev
->id
= evsel
->core
.id
[0];
1779 cpu_map_data__synthesize((struct perf_record_cpu_map_data
*)ev
->data
,
1780 evsel
->core
.own_cpus
, type
, max
);
1782 err
= process(tool
, (union perf_event
*)ev
, NULL
, NULL
);
1787 int perf_event__synthesize_attrs(struct perf_tool
*tool
, struct evlist
*evlist
,
1788 perf_event__handler_t process
)
1790 struct evsel
*evsel
;
1793 evlist__for_each_entry(evlist
, evsel
) {
1794 err
= perf_event__synthesize_attr(tool
, &evsel
->core
.attr
, evsel
->core
.ids
,
1795 evsel
->core
.id
, process
);
1797 pr_debug("failed to create perf header attribute\n");
1805 static bool has_unit(struct evsel
*evsel
)
1807 return evsel
->unit
&& *evsel
->unit
;
1810 static bool has_scale(struct evsel
*evsel
)
1812 return evsel
->scale
!= 1;
1815 int perf_event__synthesize_extra_attr(struct perf_tool
*tool
, struct evlist
*evsel_list
,
1816 perf_event__handler_t process
, bool is_pipe
)
1818 struct evsel
*evsel
;
1822 * Synthesize other events stuff not carried within
1823 * attr event - unit, scale, name
1825 evlist__for_each_entry(evsel_list
, evsel
) {
1826 if (!evsel
->supported
)
1830 * Synthesize unit and scale only if it's defined.
1832 if (has_unit(evsel
)) {
1833 err
= perf_event__synthesize_event_update_unit(tool
, evsel
, process
);
1835 pr_err("Couldn't synthesize evsel unit.\n");
1840 if (has_scale(evsel
)) {
1841 err
= perf_event__synthesize_event_update_scale(tool
, evsel
, process
);
1843 pr_err("Couldn't synthesize evsel evsel.\n");
1848 if (evsel
->core
.own_cpus
) {
1849 err
= perf_event__synthesize_event_update_cpus(tool
, evsel
, process
);
1851 pr_err("Couldn't synthesize evsel cpus.\n");
1857 * Name is needed only for pipe output,
1858 * perf.data carries event names.
1861 err
= perf_event__synthesize_event_update_name(tool
, evsel
, process
);
1863 pr_err("Couldn't synthesize evsel name.\n");
1871 int perf_event__synthesize_attr(struct perf_tool
*tool
, struct perf_event_attr
*attr
,
1872 u32 ids
, u64
*id
, perf_event__handler_t process
)
1874 union perf_event
*ev
;
1878 size
= sizeof(struct perf_event_attr
);
1879 size
= PERF_ALIGN(size
, sizeof(u64
));
1880 size
+= sizeof(struct perf_event_header
);
1881 size
+= ids
* sizeof(u64
);
1888 ev
->attr
.attr
= *attr
;
1889 memcpy(ev
->attr
.id
, id
, ids
* sizeof(u64
));
1891 ev
->attr
.header
.type
= PERF_RECORD_HEADER_ATTR
;
1892 ev
->attr
.header
.size
= (u16
)size
;
1894 if (ev
->attr
.header
.size
== size
)
1895 err
= process(tool
, ev
, NULL
, NULL
);
1904 int perf_event__synthesize_tracing_data(struct perf_tool
*tool
, int fd
, struct evlist
*evlist
,
1905 perf_event__handler_t process
)
1907 union perf_event ev
;
1908 struct tracing_data
*tdata
;
1909 ssize_t size
= 0, aligned_size
= 0, padding
;
1913 * We are going to store the size of the data followed
1914 * by the data contents. Since the fd descriptor is a pipe,
1915 * we cannot seek back to store the size of the data once
1916 * we know it. Instead we:
1918 * - write the tracing data to the temp file
1919 * - get/write the data size to pipe
1920 * - write the tracing data from the temp file
1923 tdata
= tracing_data_get(&evlist
->core
.entries
, fd
, true);
1927 memset(&ev
, 0, sizeof(ev
));
1929 ev
.tracing_data
.header
.type
= PERF_RECORD_HEADER_TRACING_DATA
;
1931 aligned_size
= PERF_ALIGN(size
, sizeof(u64
));
1932 padding
= aligned_size
- size
;
1933 ev
.tracing_data
.header
.size
= sizeof(ev
.tracing_data
);
1934 ev
.tracing_data
.size
= aligned_size
;
1936 process(tool
, &ev
, NULL
, NULL
);
1939 * The put function will copy all the tracing data
1940 * stored in temp file to the pipe.
1942 tracing_data_put(tdata
);
1944 ff
= (struct feat_fd
){ .fd
= fd
};
1945 if (write_padded(&ff
, NULL
, 0, padding
))
1948 return aligned_size
;
1951 int perf_event__synthesize_build_id(struct perf_tool
*tool
, struct dso
*pos
, u16 misc
,
1952 perf_event__handler_t process
, struct machine
*machine
)
1954 union perf_event ev
;
1960 memset(&ev
, 0, sizeof(ev
));
1962 len
= pos
->long_name_len
+ 1;
1963 len
= PERF_ALIGN(len
, NAME_ALIGN
);
1964 memcpy(&ev
.build_id
.build_id
, pos
->build_id
, sizeof(pos
->build_id
));
1965 ev
.build_id
.header
.type
= PERF_RECORD_HEADER_BUILD_ID
;
1966 ev
.build_id
.header
.misc
= misc
;
1967 ev
.build_id
.pid
= machine
->pid
;
1968 ev
.build_id
.header
.size
= sizeof(ev
.build_id
) + len
;
1969 memcpy(&ev
.build_id
.filename
, pos
->long_name
, pos
->long_name_len
);
1971 return process(tool
, &ev
, NULL
, machine
);
1974 int perf_event__synthesize_stat_events(struct perf_stat_config
*config
, struct perf_tool
*tool
,
1975 struct evlist
*evlist
, perf_event__handler_t process
, bool attrs
)
1980 err
= perf_event__synthesize_attrs(tool
, evlist
, process
);
1982 pr_err("Couldn't synthesize attrs.\n");
1987 err
= perf_event__synthesize_extra_attr(tool
, evlist
, process
, attrs
);
1988 err
= perf_event__synthesize_thread_map2(tool
, evlist
->core
.threads
, process
, NULL
);
1990 pr_err("Couldn't synthesize thread map.\n");
1994 err
= perf_event__synthesize_cpu_map(tool
, evlist
->core
.cpus
, process
, NULL
);
1996 pr_err("Couldn't synthesize thread map.\n");
2000 err
= perf_event__synthesize_stat_config(tool
, config
, process
, NULL
);
2002 pr_err("Couldn't synthesize config.\n");
2009 int __weak
perf_event__synth_time_conv(const struct perf_event_mmap_page
*pc __maybe_unused
,
2010 struct perf_tool
*tool __maybe_unused
,
2011 perf_event__handler_t process __maybe_unused
,
2012 struct machine
*machine __maybe_unused
)
2017 extern const struct perf_header_feature_ops feat_ops
[HEADER_LAST_FEATURE
];
2019 int perf_event__synthesize_features(struct perf_tool
*tool
, struct perf_session
*session
,
2020 struct evlist
*evlist
, perf_event__handler_t process
)
2022 struct perf_header
*header
= &session
->header
;
2023 struct perf_record_header_feature
*fe
;
2028 sz_hdr
= sizeof(fe
->header
);
2029 sz
= sizeof(union perf_event
);
2030 /* get a nice alignment */
2031 sz
= PERF_ALIGN(sz
, page_size
);
2033 memset(&ff
, 0, sizeof(ff
));
2035 ff
.buf
= malloc(sz
);
2039 ff
.size
= sz
- sz_hdr
;
2040 ff
.ph
= &session
->header
;
2042 for_each_set_bit(feat
, header
->adds_features
, HEADER_FEAT_BITS
) {
2043 if (!feat_ops
[feat
].synthesize
) {
2044 pr_debug("No record header feature for header :%d\n", feat
);
2048 ff
.offset
= sizeof(*fe
);
2050 ret
= feat_ops
[feat
].write(&ff
, evlist
);
2051 if (ret
|| ff
.offset
<= (ssize_t
)sizeof(*fe
)) {
2052 pr_debug("Error writing feature\n");
2055 /* ff.buf may have changed due to realloc in do_write() */
2057 memset(fe
, 0, sizeof(*fe
));
2060 fe
->header
.type
= PERF_RECORD_HEADER_FEATURE
;
2061 fe
->header
.size
= ff
.offset
;
2063 ret
= process(tool
, ff
.buf
, NULL
, NULL
);
2070 /* Send HEADER_LAST_FEATURE mark. */
2072 fe
->feat_id
= HEADER_LAST_FEATURE
;
2073 fe
->header
.type
= PERF_RECORD_HEADER_FEATURE
;
2074 fe
->header
.size
= sizeof(*fe
);
2076 ret
= process(tool
, ff
.buf
, NULL
, NULL
);