1 // SPDX-License-Identifier: GPL-2.0
10 #include <linux/bitmap.h>
13 #include <linux/ctype.h>
14 #include <linux/zalloc.h>
16 static int max_cpu_num
;
17 static int max_present_cpu_num
;
18 static int max_node_num
;
19 static int *cpunode_map
;
21 static struct perf_cpu_map
*cpu_map__from_entries(struct cpu_map_entries
*cpus
)
23 struct perf_cpu_map
*map
;
25 map
= perf_cpu_map__empty_new(cpus
->nr
);
29 for (i
= 0; i
< cpus
->nr
; i
++) {
31 * Special treatment for -1, which is not real cpu number,
32 * and we need to use (int) -1 to initialize map[i],
33 * otherwise it would become 65535.
35 if (cpus
->cpu
[i
] == (u16
) -1)
38 map
->map
[i
] = (int) cpus
->cpu
[i
];
45 static struct perf_cpu_map
*cpu_map__from_mask(struct perf_record_record_cpu_map
*mask
)
47 struct perf_cpu_map
*map
;
48 int nr
, nbits
= mask
->nr
* mask
->long_size
* BITS_PER_BYTE
;
50 nr
= bitmap_weight(mask
->mask
, nbits
);
52 map
= perf_cpu_map__empty_new(nr
);
56 for_each_set_bit(cpu
, mask
->mask
, nbits
)
63 struct perf_cpu_map
*cpu_map__new_data(struct perf_record_cpu_map_data
*data
)
65 if (data
->type
== PERF_CPU_MAP__CPUS
)
66 return cpu_map__from_entries((struct cpu_map_entries
*)data
->data
);
68 return cpu_map__from_mask((struct perf_record_record_cpu_map
*)data
->data
);
71 size_t cpu_map__fprintf(struct perf_cpu_map
*map
, FILE *fp
)
76 cpu_map__snprint(map
, buf
, sizeof(buf
));
77 return fprintf(fp
, "%s\n", buf
);
81 struct perf_cpu_map
*perf_cpu_map__empty_new(int nr
)
83 struct perf_cpu_map
*cpus
= malloc(sizeof(*cpus
) + sizeof(int) * nr
);
89 for (i
= 0; i
< nr
; i
++)
92 refcount_set(&cpus
->refcnt
, 1);
98 static int cpu__get_topology_int(int cpu
, const char *name
, int *value
)
102 snprintf(path
, PATH_MAX
,
103 "devices/system/cpu/cpu%d/topology/%s", cpu
, name
);
105 return sysfs__read_int(path
, value
);
108 int cpu_map__get_socket_id(int cpu
)
110 int value
, ret
= cpu__get_topology_int(cpu
, "physical_package_id", &value
);
114 int cpu_map__get_socket(struct perf_cpu_map
*map
, int idx
, void *data __maybe_unused
)
123 return cpu_map__get_socket_id(cpu
);
126 static int cmp_ids(const void *a
, const void *b
)
128 return *(int *)a
- *(int *)b
;
131 int cpu_map__build_map(struct perf_cpu_map
*cpus
, struct perf_cpu_map
**res
,
132 int (*f
)(struct perf_cpu_map
*map
, int cpu
, void *data
),
135 struct perf_cpu_map
*c
;
139 /* allocate as much as possible */
140 c
= calloc(1, sizeof(*c
) + nr
* sizeof(int));
144 for (cpu
= 0; cpu
< nr
; cpu
++) {
145 s1
= f(cpus
, cpu
, data
);
146 for (s2
= 0; s2
< c
->nr
; s2
++) {
147 if (s1
== c
->map
[s2
])
155 /* ensure we process id in increasing order */
156 qsort(c
->map
, c
->nr
, sizeof(int), cmp_ids
);
158 refcount_set(&c
->refcnt
, 1);
163 int cpu_map__get_die_id(int cpu
)
165 int value
, ret
= cpu__get_topology_int(cpu
, "die_id", &value
);
170 int cpu_map__get_die(struct perf_cpu_map
*map
, int idx
, void *data
)
179 die_id
= cpu_map__get_die_id(cpu
);
180 /* There is no die_id on legacy system. */
184 s
= cpu_map__get_socket(map
, idx
, data
);
189 * Encode socket in bit range 15:8
190 * die_id is relative to socket, and
191 * we need a global id. So we combine
194 if (WARN_ONCE(die_id
>> 8, "The die id number is too big.\n"))
197 if (WARN_ONCE(s
>> 8, "The socket id number is too big.\n"))
200 return (s
<< 8) | (die_id
& 0xff);
203 int cpu_map__get_core_id(int cpu
)
205 int value
, ret
= cpu__get_topology_int(cpu
, "core_id", &value
);
209 int cpu_map__get_node_id(int cpu
)
211 return cpu__get_node(cpu
);
214 int cpu_map__get_core(struct perf_cpu_map
*map
, int idx
, void *data
)
223 cpu
= cpu_map__get_core_id(cpu
);
225 /* s_die is the combination of socket + die id */
226 s_die
= cpu_map__get_die(map
, idx
, data
);
231 * encode socket in bit range 31:24
232 * encode die id in bit range 23:16
233 * core_id is relative to socket and die,
234 * we need a global id. So we combine
235 * socket + die id + core id
237 if (WARN_ONCE(cpu
>> 16, "The core id number is too big.\n"))
240 return (s_die
<< 16) | (cpu
& 0xffff);
243 int cpu_map__get_node(struct perf_cpu_map
*map
, int idx
, void *data __maybe_unused
)
245 if (idx
< 0 || idx
>= map
->nr
)
248 return cpu_map__get_node_id(map
->map
[idx
]);
251 int cpu_map__build_socket_map(struct perf_cpu_map
*cpus
, struct perf_cpu_map
**sockp
)
253 return cpu_map__build_map(cpus
, sockp
, cpu_map__get_socket
, NULL
);
256 int cpu_map__build_die_map(struct perf_cpu_map
*cpus
, struct perf_cpu_map
**diep
)
258 return cpu_map__build_map(cpus
, diep
, cpu_map__get_die
, NULL
);
261 int cpu_map__build_core_map(struct perf_cpu_map
*cpus
, struct perf_cpu_map
**corep
)
263 return cpu_map__build_map(cpus
, corep
, cpu_map__get_core
, NULL
);
266 int cpu_map__build_node_map(struct perf_cpu_map
*cpus
, struct perf_cpu_map
**numap
)
268 return cpu_map__build_map(cpus
, numap
, cpu_map__get_node
, NULL
);
271 /* setup simple routines to easily access node numbers given a cpu number */
272 static int get_max_num(char *path
, int *max
)
278 if (filename__read_str(path
, &buf
, &num
))
283 /* start on the right, to find highest node num */
285 if ((buf
[num
] == ',') || (buf
[num
] == '-')) {
290 if (sscanf(&buf
[num
], "%d", max
) < 1) {
295 /* convert from 0-based to 1-based */
303 /* Determine highest possible cpu in the system for sparse allocation */
304 static void set_max_cpu_num(void)
312 max_present_cpu_num
= 4096;
314 mnt
= sysfs__mountpoint();
318 /* get the highest possible cpu number for a sparse allocation */
319 ret
= snprintf(path
, PATH_MAX
, "%s/devices/system/cpu/possible", mnt
);
320 if (ret
>= PATH_MAX
) {
321 pr_err("sysfs path crossed PATH_MAX(%d) size\n", PATH_MAX
);
325 ret
= get_max_num(path
, &max_cpu_num
);
329 /* get the highest present cpu number for a sparse allocation */
330 ret
= snprintf(path
, PATH_MAX
, "%s/devices/system/cpu/present", mnt
);
331 if (ret
>= PATH_MAX
) {
332 pr_err("sysfs path crossed PATH_MAX(%d) size\n", PATH_MAX
);
336 ret
= get_max_num(path
, &max_present_cpu_num
);
340 pr_err("Failed to read max cpus, using default of %d\n", max_cpu_num
);
343 /* Determine highest possible node in the system for sparse allocation */
344 static void set_max_node_num(void)
353 mnt
= sysfs__mountpoint();
357 /* get the highest possible cpu number for a sparse allocation */
358 ret
= snprintf(path
, PATH_MAX
, "%s/devices/system/node/possible", mnt
);
359 if (ret
>= PATH_MAX
) {
360 pr_err("sysfs path crossed PATH_MAX(%d) size\n", PATH_MAX
);
364 ret
= get_max_num(path
, &max_node_num
);
368 pr_err("Failed to read max nodes, using default of %d\n", max_node_num
);
371 int cpu__max_node(void)
373 if (unlikely(!max_node_num
))
379 int cpu__max_cpu(void)
381 if (unlikely(!max_cpu_num
))
387 int cpu__max_present_cpu(void)
389 if (unlikely(!max_present_cpu_num
))
392 return max_present_cpu_num
;
396 int cpu__get_node(int cpu
)
398 if (unlikely(cpunode_map
== NULL
)) {
399 pr_debug("cpu_map not initialized\n");
403 return cpunode_map
[cpu
];
406 static int init_cpunode_map(void)
413 cpunode_map
= calloc(max_cpu_num
, sizeof(int));
415 pr_err("%s: calloc failed\n", __func__
);
419 for (i
= 0; i
< max_cpu_num
; i
++)
425 int cpu__setup_cpunode_map(void)
427 struct dirent
*dent1
, *dent2
;
429 unsigned int cpu
, mem
;
435 /* initialize globals */
436 if (init_cpunode_map())
439 mnt
= sysfs__mountpoint();
443 n
= snprintf(path
, PATH_MAX
, "%s/devices/system/node", mnt
);
445 pr_err("sysfs path crossed PATH_MAX(%d) size\n", PATH_MAX
);
449 dir1
= opendir(path
);
453 /* walk tree and setup map */
454 while ((dent1
= readdir(dir1
)) != NULL
) {
455 if (dent1
->d_type
!= DT_DIR
|| sscanf(dent1
->d_name
, "node%u", &mem
) < 1)
458 n
= snprintf(buf
, PATH_MAX
, "%s/%s", path
, dent1
->d_name
);
460 pr_err("sysfs path crossed PATH_MAX(%d) size\n", PATH_MAX
);
467 while ((dent2
= readdir(dir2
)) != NULL
) {
468 if (dent2
->d_type
!= DT_LNK
|| sscanf(dent2
->d_name
, "cpu%u", &cpu
) < 1)
470 cpunode_map
[cpu
] = mem
;
478 bool cpu_map__has(struct perf_cpu_map
*cpus
, int cpu
)
480 return perf_cpu_map__idx(cpus
, cpu
) != -1;
483 int cpu_map__cpu(struct perf_cpu_map
*cpus
, int idx
)
485 return cpus
->map
[idx
];
488 size_t cpu_map__snprint(struct perf_cpu_map
*map
, char *buf
, size_t size
)
490 int i
, cpu
, start
= -1;
494 #define COMMA first ? "" : ","
496 for (i
= 0; i
< map
->nr
+ 1; i
++) {
497 bool last
= i
== map
->nr
;
499 cpu
= last
? INT_MAX
: map
->map
[i
];
504 ret
+= snprintf(buf
+ ret
, size
- ret
,
508 } else if (((i
- start
) != (cpu
- map
->map
[start
])) || last
) {
512 ret
+= snprintf(buf
+ ret
, size
- ret
,
516 ret
+= snprintf(buf
+ ret
, size
- ret
,
518 map
->map
[start
], map
->map
[end
]);
527 pr_debug2("cpumask list: %s\n", buf
);
531 static char hex_char(unsigned char val
)
536 return val
- 10 + 'a';
540 size_t cpu_map__snprint_mask(struct perf_cpu_map
*map
, char *buf
, size_t size
)
544 unsigned char *bitmap
;
545 int last_cpu
= cpu_map__cpu(map
, map
->nr
- 1);
550 bitmap
= zalloc(last_cpu
/ 8 + 1);
551 if (bitmap
== NULL
) {
556 for (i
= 0; i
< map
->nr
; i
++) {
557 cpu
= cpu_map__cpu(map
, i
);
558 bitmap
[cpu
/ 8] |= 1 << (cpu
% 8);
561 for (cpu
= last_cpu
/ 4 * 4; cpu
>= 0; cpu
-= 4) {
562 unsigned char bits
= bitmap
[cpu
/ 8];
569 *ptr
++ = hex_char(bits
);
570 if ((cpu
% 32) == 0 && cpu
> 0)
576 buf
[size
- 1] = '\0';
580 const struct perf_cpu_map
*cpu_map__online(void) /* thread unsafe */
582 static const struct perf_cpu_map
*online
= NULL
;
585 online
= perf_cpu_map__new(NULL
); /* from /sys/devices/system/cpu/online */