1 // SPDX-License-Identifier: GPL-2.0
10 #include <linux/bitmap.h>
13 #include <linux/ctype.h>
14 #include <linux/zalloc.h>
16 static int max_cpu_num
;
17 static int max_present_cpu_num
;
18 static int max_node_num
;
19 static int *cpunode_map
;
21 static struct perf_cpu_map
*cpu_map__from_entries(struct cpu_map_entries
*cpus
)
23 struct perf_cpu_map
*map
;
25 map
= perf_cpu_map__empty_new(cpus
->nr
);
29 for (i
= 0; i
< cpus
->nr
; i
++) {
31 * Special treatment for -1, which is not real cpu number,
32 * and we need to use (int) -1 to initialize map[i],
33 * otherwise it would become 65535.
35 if (cpus
->cpu
[i
] == (u16
) -1)
38 map
->map
[i
] = (int) cpus
->cpu
[i
];
45 static struct perf_cpu_map
*cpu_map__from_mask(struct perf_record_record_cpu_map
*mask
)
47 struct perf_cpu_map
*map
;
48 int nr
, nbits
= mask
->nr
* mask
->long_size
* BITS_PER_BYTE
;
50 nr
= bitmap_weight(mask
->mask
, nbits
);
52 map
= perf_cpu_map__empty_new(nr
);
56 for_each_set_bit(cpu
, mask
->mask
, nbits
)
63 struct perf_cpu_map
*cpu_map__new_data(struct perf_record_cpu_map_data
*data
)
65 if (data
->type
== PERF_CPU_MAP__CPUS
)
66 return cpu_map__from_entries((struct cpu_map_entries
*)data
->data
);
68 return cpu_map__from_mask((struct perf_record_record_cpu_map
*)data
->data
);
71 size_t cpu_map__fprintf(struct perf_cpu_map
*map
, FILE *fp
)
76 cpu_map__snprint(map
, buf
, sizeof(buf
));
77 return fprintf(fp
, "%s\n", buf
);
81 struct perf_cpu_map
*perf_cpu_map__empty_new(int nr
)
83 struct perf_cpu_map
*cpus
= malloc(sizeof(*cpus
) + sizeof(int) * nr
);
89 for (i
= 0; i
< nr
; i
++)
92 refcount_set(&cpus
->refcnt
, 1);
98 struct cpu_aggr_map
*cpu_aggr_map__empty_new(int nr
)
100 struct cpu_aggr_map
*cpus
= malloc(sizeof(*cpus
) + sizeof(struct aggr_cpu_id
) * nr
);
106 for (i
= 0; i
< nr
; i
++)
107 cpus
->map
[i
] = cpu_map__empty_aggr_cpu_id();
109 refcount_set(&cpus
->refcnt
, 1);
115 static int cpu__get_topology_int(int cpu
, const char *name
, int *value
)
119 snprintf(path
, PATH_MAX
,
120 "devices/system/cpu/cpu%d/topology/%s", cpu
, name
);
122 return sysfs__read_int(path
, value
);
125 int cpu_map__get_socket_id(int cpu
)
127 int value
, ret
= cpu__get_topology_int(cpu
, "physical_package_id", &value
);
131 struct aggr_cpu_id
cpu_map__get_socket(struct perf_cpu_map
*map
, int idx
,
132 void *data __maybe_unused
)
135 struct aggr_cpu_id id
= cpu_map__empty_aggr_cpu_id();
142 id
.socket
= cpu_map__get_socket_id(cpu
);
146 static int cmp_aggr_cpu_id(const void *a_pointer
, const void *b_pointer
)
148 struct aggr_cpu_id
*a
= (struct aggr_cpu_id
*)a_pointer
;
149 struct aggr_cpu_id
*b
= (struct aggr_cpu_id
*)b_pointer
;
151 if (a
->node
!= b
->node
)
152 return a
->node
- b
->node
;
153 else if (a
->socket
!= b
->socket
)
154 return a
->socket
- b
->socket
;
155 else if (a
->die
!= b
->die
)
156 return a
->die
- b
->die
;
157 else if (a
->core
!= b
->core
)
158 return a
->core
- b
->core
;
160 return a
->thread
- b
->thread
;
163 int cpu_map__build_map(struct perf_cpu_map
*cpus
, struct cpu_aggr_map
**res
,
164 struct aggr_cpu_id (*f
)(struct perf_cpu_map
*map
, int cpu
, void *data
),
168 struct cpu_aggr_map
*c
= cpu_aggr_map__empty_new(nr
);
170 struct aggr_cpu_id s1
;
175 /* Reset size as it may only be partially filled */
178 for (cpu
= 0; cpu
< nr
; cpu
++) {
179 s1
= f(cpus
, cpu
, data
);
180 for (s2
= 0; s2
< c
->nr
; s2
++) {
181 if (cpu_map__compare_aggr_cpu_id(s1
, c
->map
[s2
]))
189 /* ensure we process id in increasing order */
190 qsort(c
->map
, c
->nr
, sizeof(struct aggr_cpu_id
), cmp_aggr_cpu_id
);
196 int cpu_map__get_die_id(int cpu
)
198 int value
, ret
= cpu__get_topology_int(cpu
, "die_id", &value
);
203 struct aggr_cpu_id
cpu_map__get_die(struct perf_cpu_map
*map
, int idx
, void *data
)
206 struct aggr_cpu_id id
= cpu_map__empty_aggr_cpu_id();
213 die
= cpu_map__get_die_id(cpu
);
214 /* There is no die_id on legacy system. */
219 * die_id is relative to socket, so start
220 * with the socket ID and then add die to
223 id
= cpu_map__get_socket(map
, idx
, data
);
224 if (cpu_map__aggr_cpu_id_is_empty(id
))
231 int cpu_map__get_core_id(int cpu
)
233 int value
, ret
= cpu__get_topology_int(cpu
, "core_id", &value
);
237 int cpu_map__get_node_id(int cpu
)
239 return cpu__get_node(cpu
);
242 struct aggr_cpu_id
cpu_map__get_core(struct perf_cpu_map
*map
, int idx
, void *data
)
245 struct aggr_cpu_id id
= cpu_map__empty_aggr_cpu_id();
252 cpu
= cpu_map__get_core_id(cpu
);
254 /* cpu_map__get_die returns a struct with socket and die set*/
255 id
= cpu_map__get_die(map
, idx
, data
);
256 if (cpu_map__aggr_cpu_id_is_empty(id
))
260 * core_id is relative to socket and die, we need a global id.
261 * So we combine the result from cpu_map__get_die with the core id
267 struct aggr_cpu_id
cpu_map__get_node(struct perf_cpu_map
*map
, int idx
, void *data __maybe_unused
)
269 struct aggr_cpu_id id
= cpu_map__empty_aggr_cpu_id();
271 if (idx
< 0 || idx
>= map
->nr
)
274 id
.node
= cpu_map__get_node_id(map
->map
[idx
]);
278 int cpu_map__build_socket_map(struct perf_cpu_map
*cpus
, struct cpu_aggr_map
**sockp
)
280 return cpu_map__build_map(cpus
, sockp
, cpu_map__get_socket
, NULL
);
283 int cpu_map__build_die_map(struct perf_cpu_map
*cpus
, struct cpu_aggr_map
**diep
)
285 return cpu_map__build_map(cpus
, diep
, cpu_map__get_die
, NULL
);
288 int cpu_map__build_core_map(struct perf_cpu_map
*cpus
, struct cpu_aggr_map
**corep
)
290 return cpu_map__build_map(cpus
, corep
, cpu_map__get_core
, NULL
);
293 int cpu_map__build_node_map(struct perf_cpu_map
*cpus
, struct cpu_aggr_map
**numap
)
295 return cpu_map__build_map(cpus
, numap
, cpu_map__get_node
, NULL
);
298 /* setup simple routines to easily access node numbers given a cpu number */
299 static int get_max_num(char *path
, int *max
)
305 if (filename__read_str(path
, &buf
, &num
))
310 /* start on the right, to find highest node num */
312 if ((buf
[num
] == ',') || (buf
[num
] == '-')) {
317 if (sscanf(&buf
[num
], "%d", max
) < 1) {
322 /* convert from 0-based to 1-based */
330 /* Determine highest possible cpu in the system for sparse allocation */
331 static void set_max_cpu_num(void)
339 max_present_cpu_num
= 4096;
341 mnt
= sysfs__mountpoint();
345 /* get the highest possible cpu number for a sparse allocation */
346 ret
= snprintf(path
, PATH_MAX
, "%s/devices/system/cpu/possible", mnt
);
347 if (ret
>= PATH_MAX
) {
348 pr_err("sysfs path crossed PATH_MAX(%d) size\n", PATH_MAX
);
352 ret
= get_max_num(path
, &max_cpu_num
);
356 /* get the highest present cpu number for a sparse allocation */
357 ret
= snprintf(path
, PATH_MAX
, "%s/devices/system/cpu/present", mnt
);
358 if (ret
>= PATH_MAX
) {
359 pr_err("sysfs path crossed PATH_MAX(%d) size\n", PATH_MAX
);
363 ret
= get_max_num(path
, &max_present_cpu_num
);
367 pr_err("Failed to read max cpus, using default of %d\n", max_cpu_num
);
370 /* Determine highest possible node in the system for sparse allocation */
371 static void set_max_node_num(void)
380 mnt
= sysfs__mountpoint();
384 /* get the highest possible cpu number for a sparse allocation */
385 ret
= snprintf(path
, PATH_MAX
, "%s/devices/system/node/possible", mnt
);
386 if (ret
>= PATH_MAX
) {
387 pr_err("sysfs path crossed PATH_MAX(%d) size\n", PATH_MAX
);
391 ret
= get_max_num(path
, &max_node_num
);
395 pr_err("Failed to read max nodes, using default of %d\n", max_node_num
);
398 int cpu__max_node(void)
400 if (unlikely(!max_node_num
))
406 int cpu__max_cpu(void)
408 if (unlikely(!max_cpu_num
))
414 int cpu__max_present_cpu(void)
416 if (unlikely(!max_present_cpu_num
))
419 return max_present_cpu_num
;
423 int cpu__get_node(int cpu
)
425 if (unlikely(cpunode_map
== NULL
)) {
426 pr_debug("cpu_map not initialized\n");
430 return cpunode_map
[cpu
];
433 static int init_cpunode_map(void)
440 cpunode_map
= calloc(max_cpu_num
, sizeof(int));
442 pr_err("%s: calloc failed\n", __func__
);
446 for (i
= 0; i
< max_cpu_num
; i
++)
452 int cpu__setup_cpunode_map(void)
454 struct dirent
*dent1
, *dent2
;
456 unsigned int cpu
, mem
;
462 /* initialize globals */
463 if (init_cpunode_map())
466 mnt
= sysfs__mountpoint();
470 n
= snprintf(path
, PATH_MAX
, "%s/devices/system/node", mnt
);
472 pr_err("sysfs path crossed PATH_MAX(%d) size\n", PATH_MAX
);
476 dir1
= opendir(path
);
480 /* walk tree and setup map */
481 while ((dent1
= readdir(dir1
)) != NULL
) {
482 if (dent1
->d_type
!= DT_DIR
|| sscanf(dent1
->d_name
, "node%u", &mem
) < 1)
485 n
= snprintf(buf
, PATH_MAX
, "%s/%s", path
, dent1
->d_name
);
487 pr_err("sysfs path crossed PATH_MAX(%d) size\n", PATH_MAX
);
494 while ((dent2
= readdir(dir2
)) != NULL
) {
495 if (dent2
->d_type
!= DT_LNK
|| sscanf(dent2
->d_name
, "cpu%u", &cpu
) < 1)
497 cpunode_map
[cpu
] = mem
;
505 bool cpu_map__has(struct perf_cpu_map
*cpus
, int cpu
)
507 return perf_cpu_map__idx(cpus
, cpu
) != -1;
510 int cpu_map__cpu(struct perf_cpu_map
*cpus
, int idx
)
512 return cpus
->map
[idx
];
515 size_t cpu_map__snprint(struct perf_cpu_map
*map
, char *buf
, size_t size
)
517 int i
, cpu
, start
= -1;
521 #define COMMA first ? "" : ","
523 for (i
= 0; i
< map
->nr
+ 1; i
++) {
524 bool last
= i
== map
->nr
;
526 cpu
= last
? INT_MAX
: map
->map
[i
];
531 ret
+= snprintf(buf
+ ret
, size
- ret
,
535 } else if (((i
- start
) != (cpu
- map
->map
[start
])) || last
) {
539 ret
+= snprintf(buf
+ ret
, size
- ret
,
543 ret
+= snprintf(buf
+ ret
, size
- ret
,
545 map
->map
[start
], map
->map
[end
]);
554 pr_debug2("cpumask list: %s\n", buf
);
558 static char hex_char(unsigned char val
)
563 return val
- 10 + 'a';
567 size_t cpu_map__snprint_mask(struct perf_cpu_map
*map
, char *buf
, size_t size
)
571 unsigned char *bitmap
;
572 int last_cpu
= cpu_map__cpu(map
, map
->nr
- 1);
577 bitmap
= zalloc(last_cpu
/ 8 + 1);
578 if (bitmap
== NULL
) {
583 for (i
= 0; i
< map
->nr
; i
++) {
584 cpu
= cpu_map__cpu(map
, i
);
585 bitmap
[cpu
/ 8] |= 1 << (cpu
% 8);
588 for (cpu
= last_cpu
/ 4 * 4; cpu
>= 0; cpu
-= 4) {
589 unsigned char bits
= bitmap
[cpu
/ 8];
596 *ptr
++ = hex_char(bits
);
597 if ((cpu
% 32) == 0 && cpu
> 0)
603 buf
[size
- 1] = '\0';
607 const struct perf_cpu_map
*cpu_map__online(void) /* thread unsafe */
609 static const struct perf_cpu_map
*online
= NULL
;
612 online
= perf_cpu_map__new(NULL
); /* from /sys/devices/system/cpu/online */
617 bool cpu_map__compare_aggr_cpu_id(struct aggr_cpu_id a
, struct aggr_cpu_id b
)
619 return a
.thread
== b
.thread
&&
621 a
.socket
== b
.socket
&&
626 bool cpu_map__aggr_cpu_id_is_empty(struct aggr_cpu_id a
)
628 return a
.thread
== -1 &&
635 struct aggr_cpu_id
cpu_map__empty_aggr_cpu_id(void)
637 struct aggr_cpu_id ret
= {