1 /* Common code for 32 and 64-bit NUMA */
2 #include <linux/topology.h>
3 #include <linux/module.h>
4 #include <linux/bootmem.h>
8 int __initdata numa_off
;
10 static __init
int numa_setup(char *opt
)
14 if (!strncmp(opt
, "off", 3))
16 #ifdef CONFIG_NUMA_EMU
17 if (!strncmp(opt
, "fake=", 5))
18 numa_emu_cmdline(opt
+ 5);
20 #ifdef CONFIG_ACPI_NUMA
21 if (!strncmp(opt
, "noacpi", 6))
26 early_param("numa", numa_setup
);
29 * apicid, cpu, node mappings
31 s16 __apicid_to_node
[MAX_LOCAL_APIC
] __cpuinitdata
= {
32 [0 ... MAX_LOCAL_APIC
-1] = NUMA_NO_NODE
35 cpumask_var_t node_to_cpumask_map
[MAX_NUMNODES
];
36 EXPORT_SYMBOL(node_to_cpumask_map
);
39 * Map cpu index to node index
41 DEFINE_EARLY_PER_CPU(int, x86_cpu_to_node_map
, NUMA_NO_NODE
);
42 EXPORT_EARLY_PER_CPU_SYMBOL(x86_cpu_to_node_map
);
44 void __cpuinit
numa_set_node(int cpu
, int node
)
46 int *cpu_to_node_map
= early_per_cpu_ptr(x86_cpu_to_node_map
);
48 /* early setting, no percpu area yet */
49 if (cpu_to_node_map
) {
50 cpu_to_node_map
[cpu
] = node
;
54 #ifdef CONFIG_DEBUG_PER_CPU_MAPS
55 if (cpu
>= nr_cpu_ids
|| !cpu_possible(cpu
)) {
56 printk(KERN_ERR
"numa_set_node: invalid cpu# (%d)\n", cpu
);
61 per_cpu(x86_cpu_to_node_map
, cpu
) = node
;
63 if (node
!= NUMA_NO_NODE
)
64 set_cpu_numa_node(cpu
, node
);
67 void __cpuinit
numa_clear_node(int cpu
)
69 numa_set_node(cpu
, NUMA_NO_NODE
);
73 * Allocate node_to_cpumask_map based on number of available nodes
74 * Requires node_possible_map to be valid.
76 * Note: node_to_cpumask() is not valid until after this is done.
77 * (Use CONFIG_DEBUG_PER_CPU_MAPS to check this.)
79 void __init
setup_node_to_cpumask_map(void)
81 unsigned int node
, num
= 0;
83 /* setup nr_node_ids if not done yet */
84 if (nr_node_ids
== MAX_NUMNODES
) {
85 for_each_node_mask(node
, node_possible_map
)
87 nr_node_ids
= num
+ 1;
90 /* allocate the map */
91 for (node
= 0; node
< nr_node_ids
; node
++)
92 alloc_bootmem_cpumask_var(&node_to_cpumask_map
[node
]);
94 /* cpumask_of_node() will now work */
95 pr_debug("Node to cpumask map for %d nodes\n", nr_node_ids
);
99 * There are unfortunately some poorly designed mainboards around that
100 * only connect memory to a single CPU. This breaks the 1:1 cpu->node
101 * mapping. To avoid this fill in the mapping for all possible CPUs,
102 * as the number of CPUs is not known yet. We round robin the existing
105 void __init
numa_init_array(void)
109 rr
= first_node(node_online_map
);
110 for (i
= 0; i
< nr_cpu_ids
; i
++) {
111 if (early_cpu_to_node(i
) != NUMA_NO_NODE
)
113 numa_set_node(i
, rr
);
114 rr
= next_node(rr
, node_online_map
);
115 if (rr
== MAX_NUMNODES
)
116 rr
= first_node(node_online_map
);
120 static __init
int find_near_online_node(int node
)
123 int min_val
= INT_MAX
;
126 for_each_online_node(n
) {
127 val
= node_distance(node
, n
);
139 * Setup early cpu_to_node.
141 * Populate cpu_to_node[] only if x86_cpu_to_apicid[],
142 * and apicid_to_node[] tables have valid entries for a CPU.
143 * This means we skip cpu_to_node[] initialisation for NUMA
144 * emulation and faking node case (when running a kernel compiled
145 * for NUMA on a non NUMA box), which is OK as cpu_to_node[]
146 * is already initialized in a round robin manner at numa_init_array,
147 * prior to this call, and this initialization is good enough
148 * for the fake NUMA cases.
150 * Called before the per_cpu areas are setup.
152 void __init
init_cpu_to_node(void)
155 u16
*cpu_to_apicid
= early_per_cpu_ptr(x86_cpu_to_apicid
);
157 BUG_ON(cpu_to_apicid
== NULL
);
159 for_each_possible_cpu(cpu
) {
160 int node
= numa_cpu_node(cpu
);
162 if (node
== NUMA_NO_NODE
)
164 if (!node_online(node
))
165 node
= find_near_online_node(node
);
166 numa_set_node(cpu
, node
);
170 #ifndef CONFIG_DEBUG_PER_CPU_MAPS
172 # ifndef CONFIG_NUMA_EMU
173 void __cpuinit
numa_add_cpu(int cpu
)
175 cpumask_set_cpu(cpu
, node_to_cpumask_map
[early_cpu_to_node(cpu
)]);
178 void __cpuinit
numa_remove_cpu(int cpu
)
180 cpumask_clear_cpu(cpu
, node_to_cpumask_map
[early_cpu_to_node(cpu
)]);
182 # endif /* !CONFIG_NUMA_EMU */
184 #else /* !CONFIG_DEBUG_PER_CPU_MAPS */
186 int __cpu_to_node(int cpu
)
188 if (early_per_cpu_ptr(x86_cpu_to_node_map
)) {
190 "cpu_to_node(%d): usage too early!\n", cpu
);
192 return early_per_cpu_ptr(x86_cpu_to_node_map
)[cpu
];
194 return per_cpu(x86_cpu_to_node_map
, cpu
);
196 EXPORT_SYMBOL(__cpu_to_node
);
199 * Same function as cpu_to_node() but used if called before the
200 * per_cpu areas are setup.
202 int early_cpu_to_node(int cpu
)
204 if (early_per_cpu_ptr(x86_cpu_to_node_map
))
205 return early_per_cpu_ptr(x86_cpu_to_node_map
)[cpu
];
207 if (!cpu_possible(cpu
)) {
209 "early_cpu_to_node(%d): no per_cpu area!\n", cpu
);
213 return per_cpu(x86_cpu_to_node_map
, cpu
);
216 struct cpumask __cpuinit
*debug_cpumask_set_cpu(int cpu
, int enable
)
218 int node
= early_cpu_to_node(cpu
);
219 struct cpumask
*mask
;
222 if (node
== NUMA_NO_NODE
) {
223 /* early_cpu_to_node() already emits a warning and trace */
226 mask
= node_to_cpumask_map
[node
];
228 pr_err("node_to_cpumask_map[%i] NULL\n", node
);
233 cpulist_scnprintf(buf
, sizeof(buf
), mask
);
234 printk(KERN_DEBUG
"%s cpu %d node %d: mask now %s\n",
235 enable
? "numa_add_cpu" : "numa_remove_cpu",
240 # ifndef CONFIG_NUMA_EMU
241 static void __cpuinit
numa_set_cpumask(int cpu
, int enable
)
243 struct cpumask
*mask
;
245 mask
= debug_cpumask_set_cpu(cpu
, enable
);
250 cpumask_set_cpu(cpu
, mask
);
252 cpumask_clear_cpu(cpu
, mask
);
255 void __cpuinit
numa_add_cpu(int cpu
)
257 numa_set_cpumask(cpu
, 1);
260 void __cpuinit
numa_remove_cpu(int cpu
)
262 numa_set_cpumask(cpu
, 0);
264 # endif /* !CONFIG_NUMA_EMU */
267 * Returns a pointer to the bitmask of CPUs on Node 'node'.
269 const struct cpumask
*cpumask_of_node(int node
)
271 if (node
>= nr_node_ids
) {
273 "cpumask_of_node(%d): node > nr_node_ids(%d)\n",
276 return cpu_none_mask
;
278 if (node_to_cpumask_map
[node
] == NULL
) {
280 "cpumask_of_node(%d): no node_to_cpumask_map!\n",
283 return cpu_online_mask
;
285 return node_to_cpumask_map
[node
];
287 EXPORT_SYMBOL(cpumask_of_node
);
289 #endif /* !CONFIG_DEBUG_PER_CPU_MAPS */