1 // SPDX-License-Identifier: GPL-2.0
10 #include <linux/bitmap.h>
13 #include "sane_ctype.h"
15 static int max_cpu_num
;
16 static int max_present_cpu_num
;
17 static int max_node_num
;
18 static int *cpunode_map
;
20 static struct cpu_map
*cpu_map__default_new(void)
25 nr_cpus
= sysconf(_SC_NPROCESSORS_ONLN
);
29 cpus
= malloc(sizeof(*cpus
) + nr_cpus
* sizeof(int));
32 for (i
= 0; i
< nr_cpus
; ++i
)
36 refcount_set(&cpus
->refcnt
, 1);
42 static struct cpu_map
*cpu_map__trim_new(int nr_cpus
, int *tmp_cpus
)
44 size_t payload_size
= nr_cpus
* sizeof(int);
45 struct cpu_map
*cpus
= malloc(sizeof(*cpus
) + payload_size
);
49 memcpy(cpus
->map
, tmp_cpus
, payload_size
);
50 refcount_set(&cpus
->refcnt
, 1);
56 struct cpu_map
*cpu_map__read(FILE *file
)
58 struct cpu_map
*cpus
= NULL
;
60 int *tmp_cpus
= NULL
, *tmp
;
68 n
= fscanf(file
, "%u%c", &cpu
, &sep
);
72 int new_max
= nr_cpus
+ cpu
- prev
- 1;
74 if (new_max
>= max_entries
) {
75 max_entries
= new_max
+ MAX_NR_CPUS
/ 2;
76 tmp
= realloc(tmp_cpus
, max_entries
* sizeof(int));
83 tmp_cpus
[nr_cpus
++] = prev
;
85 if (nr_cpus
== max_entries
) {
86 max_entries
+= MAX_NR_CPUS
;
87 tmp
= realloc(tmp_cpus
, max_entries
* sizeof(int));
93 tmp_cpus
[nr_cpus
++] = cpu
;
94 if (n
== 2 && sep
== '-')
98 if (n
== 1 || sep
== '\n')
103 cpus
= cpu_map__trim_new(nr_cpus
, tmp_cpus
);
105 cpus
= cpu_map__default_new();
111 static struct cpu_map
*cpu_map__read_all_cpu_map(void)
113 struct cpu_map
*cpus
= NULL
;
116 onlnf
= fopen("/sys/devices/system/cpu/online", "r");
118 return cpu_map__default_new();
120 cpus
= cpu_map__read(onlnf
);
125 struct cpu_map
*cpu_map__new(const char *cpu_list
)
127 struct cpu_map
*cpus
= NULL
;
128 unsigned long start_cpu
, end_cpu
= 0;
131 int *tmp_cpus
= NULL
, *tmp
;
135 return cpu_map__read_all_cpu_map();
138 * must handle the case of empty cpumap to cover
139 * TOPOLOGY header for NUMA nodes with no CPU
140 * ( e.g., because of CPU hotplug)
142 if (!isdigit(*cpu_list
) && *cpu_list
!= '\0')
145 while (isdigit(*cpu_list
)) {
147 start_cpu
= strtoul(cpu_list
, &p
, 0);
148 if (start_cpu
>= INT_MAX
149 || (*p
!= '\0' && *p
!= ',' && *p
!= '-'))
155 end_cpu
= strtoul(cpu_list
, &p
, 0);
157 if (end_cpu
>= INT_MAX
|| (*p
!= '\0' && *p
!= ','))
160 if (end_cpu
< start_cpu
)
166 for (; start_cpu
<= end_cpu
; start_cpu
++) {
167 /* check for duplicates */
168 for (i
= 0; i
< nr_cpus
; i
++)
169 if (tmp_cpus
[i
] == (int)start_cpu
)
172 if (nr_cpus
== max_entries
) {
173 max_entries
+= MAX_NR_CPUS
;
174 tmp
= realloc(tmp_cpus
, max_entries
* sizeof(int));
179 tmp_cpus
[nr_cpus
++] = (int)start_cpu
;
188 cpus
= cpu_map__trim_new(nr_cpus
, tmp_cpus
);
189 else if (*cpu_list
!= '\0')
190 cpus
= cpu_map__default_new();
192 cpus
= cpu_map__dummy_new();
199 static struct cpu_map
*cpu_map__from_entries(struct cpu_map_entries
*cpus
)
203 map
= cpu_map__empty_new(cpus
->nr
);
207 for (i
= 0; i
< cpus
->nr
; i
++) {
209 * Special treatment for -1, which is not real cpu number,
210 * and we need to use (int) -1 to initialize map[i],
211 * otherwise it would become 65535.
213 if (cpus
->cpu
[i
] == (u16
) -1)
216 map
->map
[i
] = (int) cpus
->cpu
[i
];
223 static struct cpu_map
*cpu_map__from_mask(struct cpu_map_mask
*mask
)
226 int nr
, nbits
= mask
->nr
* mask
->long_size
* BITS_PER_BYTE
;
228 nr
= bitmap_weight(mask
->mask
, nbits
);
230 map
= cpu_map__empty_new(nr
);
234 for_each_set_bit(cpu
, mask
->mask
, nbits
)
241 struct cpu_map
*cpu_map__new_data(struct cpu_map_data
*data
)
243 if (data
->type
== PERF_CPU_MAP__CPUS
)
244 return cpu_map__from_entries((struct cpu_map_entries
*)data
->data
);
246 return cpu_map__from_mask((struct cpu_map_mask
*)data
->data
);
249 size_t cpu_map__fprintf(struct cpu_map
*map
, FILE *fp
)
254 cpu_map__snprint(map
, buf
, sizeof(buf
));
255 return fprintf(fp
, "%s\n", buf
);
259 struct cpu_map
*cpu_map__dummy_new(void)
261 struct cpu_map
*cpus
= malloc(sizeof(*cpus
) + sizeof(int));
266 refcount_set(&cpus
->refcnt
, 1);
272 struct cpu_map
*cpu_map__empty_new(int nr
)
274 struct cpu_map
*cpus
= malloc(sizeof(*cpus
) + sizeof(int) * nr
);
280 for (i
= 0; i
< nr
; i
++)
283 refcount_set(&cpus
->refcnt
, 1);
289 static void cpu_map__delete(struct cpu_map
*map
)
292 WARN_ONCE(refcount_read(&map
->refcnt
) != 0,
293 "cpu_map refcnt unbalanced\n");
298 struct cpu_map
*cpu_map__get(struct cpu_map
*map
)
301 refcount_inc(&map
->refcnt
);
305 void cpu_map__put(struct cpu_map
*map
)
307 if (map
&& refcount_dec_and_test(&map
->refcnt
))
308 cpu_map__delete(map
);
311 static int cpu__get_topology_int(int cpu
, const char *name
, int *value
)
315 snprintf(path
, PATH_MAX
,
316 "devices/system/cpu/cpu%d/topology/%s", cpu
, name
);
318 return sysfs__read_int(path
, value
);
321 int cpu_map__get_socket_id(int cpu
)
323 int value
, ret
= cpu__get_topology_int(cpu
, "physical_package_id", &value
);
327 int cpu_map__get_socket(struct cpu_map
*map
, int idx
, void *data __maybe_unused
)
336 return cpu_map__get_socket_id(cpu
);
339 static int cmp_ids(const void *a
, const void *b
)
341 return *(int *)a
- *(int *)b
;
344 int cpu_map__build_map(struct cpu_map
*cpus
, struct cpu_map
**res
,
345 int (*f
)(struct cpu_map
*map
, int cpu
, void *data
),
352 /* allocate as much as possible */
353 c
= calloc(1, sizeof(*c
) + nr
* sizeof(int));
357 for (cpu
= 0; cpu
< nr
; cpu
++) {
358 s1
= f(cpus
, cpu
, data
);
359 for (s2
= 0; s2
< c
->nr
; s2
++) {
360 if (s1
== c
->map
[s2
])
368 /* ensure we process id in increasing order */
369 qsort(c
->map
, c
->nr
, sizeof(int), cmp_ids
);
371 refcount_set(&c
->refcnt
, 1);
376 int cpu_map__get_core_id(int cpu
)
378 int value
, ret
= cpu__get_topology_int(cpu
, "core_id", &value
);
382 int cpu_map__get_core(struct cpu_map
*map
, int idx
, void *data
)
391 cpu
= cpu_map__get_core_id(cpu
);
393 s
= cpu_map__get_socket(map
, idx
, data
);
398 * encode socket in upper 16 bits
399 * core_id is relative to socket, and
400 * we need a global id. So we combine
403 return (s
<< 16) | (cpu
& 0xffff);
406 int cpu_map__build_socket_map(struct cpu_map
*cpus
, struct cpu_map
**sockp
)
408 return cpu_map__build_map(cpus
, sockp
, cpu_map__get_socket
, NULL
);
411 int cpu_map__build_core_map(struct cpu_map
*cpus
, struct cpu_map
**corep
)
413 return cpu_map__build_map(cpus
, corep
, cpu_map__get_core
, NULL
);
416 /* setup simple routines to easily access node numbers given a cpu number */
417 static int get_max_num(char *path
, int *max
)
423 if (filename__read_str(path
, &buf
, &num
))
428 /* start on the right, to find highest node num */
430 if ((buf
[num
] == ',') || (buf
[num
] == '-')) {
435 if (sscanf(&buf
[num
], "%d", max
) < 1) {
440 /* convert from 0-based to 1-based */
448 /* Determine highest possible cpu in the system for sparse allocation */
449 static void set_max_cpu_num(void)
457 max_present_cpu_num
= 4096;
459 mnt
= sysfs__mountpoint();
463 /* get the highest possible cpu number for a sparse allocation */
464 ret
= snprintf(path
, PATH_MAX
, "%s/devices/system/cpu/possible", mnt
);
465 if (ret
== PATH_MAX
) {
466 pr_err("sysfs path crossed PATH_MAX(%d) size\n", PATH_MAX
);
470 ret
= get_max_num(path
, &max_cpu_num
);
474 /* get the highest present cpu number for a sparse allocation */
475 ret
= snprintf(path
, PATH_MAX
, "%s/devices/system/cpu/present", mnt
);
476 if (ret
== PATH_MAX
) {
477 pr_err("sysfs path crossed PATH_MAX(%d) size\n", PATH_MAX
);
481 ret
= get_max_num(path
, &max_present_cpu_num
);
485 pr_err("Failed to read max cpus, using default of %d\n", max_cpu_num
);
488 /* Determine highest possible node in the system for sparse allocation */
489 static void set_max_node_num(void)
498 mnt
= sysfs__mountpoint();
502 /* get the highest possible cpu number for a sparse allocation */
503 ret
= snprintf(path
, PATH_MAX
, "%s/devices/system/node/possible", mnt
);
504 if (ret
== PATH_MAX
) {
505 pr_err("sysfs path crossed PATH_MAX(%d) size\n", PATH_MAX
);
509 ret
= get_max_num(path
, &max_node_num
);
513 pr_err("Failed to read max nodes, using default of %d\n", max_node_num
);
516 int cpu__max_node(void)
518 if (unlikely(!max_node_num
))
524 int cpu__max_cpu(void)
526 if (unlikely(!max_cpu_num
))
532 int cpu__max_present_cpu(void)
534 if (unlikely(!max_present_cpu_num
))
537 return max_present_cpu_num
;
541 int cpu__get_node(int cpu
)
543 if (unlikely(cpunode_map
== NULL
)) {
544 pr_debug("cpu_map not initialized\n");
548 return cpunode_map
[cpu
];
551 static int init_cpunode_map(void)
558 cpunode_map
= calloc(max_cpu_num
, sizeof(int));
560 pr_err("%s: calloc failed\n", __func__
);
564 for (i
= 0; i
< max_cpu_num
; i
++)
570 int cpu__setup_cpunode_map(void)
572 struct dirent
*dent1
, *dent2
;
574 unsigned int cpu
, mem
;
580 /* initialize globals */
581 if (init_cpunode_map())
584 mnt
= sysfs__mountpoint();
588 n
= snprintf(path
, PATH_MAX
, "%s/devices/system/node", mnt
);
590 pr_err("sysfs path crossed PATH_MAX(%d) size\n", PATH_MAX
);
594 dir1
= opendir(path
);
598 /* walk tree and setup map */
599 while ((dent1
= readdir(dir1
)) != NULL
) {
600 if (dent1
->d_type
!= DT_DIR
|| sscanf(dent1
->d_name
, "node%u", &mem
) < 1)
603 n
= snprintf(buf
, PATH_MAX
, "%s/%s", path
, dent1
->d_name
);
605 pr_err("sysfs path crossed PATH_MAX(%d) size\n", PATH_MAX
);
612 while ((dent2
= readdir(dir2
)) != NULL
) {
613 if (dent2
->d_type
!= DT_LNK
|| sscanf(dent2
->d_name
, "cpu%u", &cpu
) < 1)
615 cpunode_map
[cpu
] = mem
;
623 bool cpu_map__has(struct cpu_map
*cpus
, int cpu
)
625 return cpu_map__idx(cpus
, cpu
) != -1;
628 int cpu_map__idx(struct cpu_map
*cpus
, int cpu
)
632 for (i
= 0; i
< cpus
->nr
; ++i
) {
633 if (cpus
->map
[i
] == cpu
)
640 int cpu_map__cpu(struct cpu_map
*cpus
, int idx
)
642 return cpus
->map
[idx
];
645 size_t cpu_map__snprint(struct cpu_map
*map
, char *buf
, size_t size
)
647 int i
, cpu
, start
= -1;
651 #define COMMA first ? "" : ","
653 for (i
= 0; i
< map
->nr
+ 1; i
++) {
654 bool last
= i
== map
->nr
;
656 cpu
= last
? INT_MAX
: map
->map
[i
];
661 ret
+= snprintf(buf
+ ret
, size
- ret
,
665 } else if (((i
- start
) != (cpu
- map
->map
[start
])) || last
) {
669 ret
+= snprintf(buf
+ ret
, size
- ret
,
673 ret
+= snprintf(buf
+ ret
, size
- ret
,
675 map
->map
[start
], map
->map
[end
]);
684 pr_debug2("cpumask list: %s\n", buf
);
688 static char hex_char(unsigned char val
)
693 return val
- 10 + 'a';
697 size_t cpu_map__snprint_mask(struct cpu_map
*map
, char *buf
, size_t size
)
701 unsigned char *bitmap
;
702 int last_cpu
= cpu_map__cpu(map
, map
->nr
- 1);
704 bitmap
= zalloc((last_cpu
+ 7) / 8);
705 if (bitmap
== NULL
) {
710 for (i
= 0; i
< map
->nr
; i
++) {
711 cpu
= cpu_map__cpu(map
, i
);
712 bitmap
[cpu
/ 8] |= 1 << (cpu
% 8);
715 for (cpu
= last_cpu
/ 4 * 4; cpu
>= 0; cpu
-= 4) {
716 unsigned char bits
= bitmap
[cpu
/ 8];
723 *ptr
++ = hex_char(bits
);
724 if ((cpu
% 32) == 0 && cpu
> 0)
730 buf
[size
- 1] = '\0';
734 const struct cpu_map
*cpu_map__online(void) /* thread unsafe */
736 static const struct cpu_map
*online
= NULL
;
739 online
= cpu_map__new(NULL
); /* from /sys/devices/system/cpu/online */