2 * This program is free software; you can distribute it and/or modify it
3 * under the terms of the GNU General Public License (Version 2) as
4 * published by the Free Software Foundation.
6 * This program is distributed in the hope it will be useful, but WITHOUT
7 * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
8 * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License
11 * You should have received a copy of the GNU General Public License along
12 * with this program; if not, write to the Free Software Foundation, Inc.,
13 * 59 Temple Place - Suite 330, Boston MA 02111-1307, USA.
15 * Copyright (C) 2004, 05, 06 MIPS Technologies, Inc.
16 * Elizabeth Clarke (beth@mips.com)
17 * Ralf Baechle (ralf@linux-mips.org)
18 * Copyright (C) 2006 Ralf Baechle (ralf@linux-mips.org)
20 #include <linux/kernel.h>
21 #include <linux/sched.h>
22 #include <linux/cpumask.h>
23 #include <linux/interrupt.h>
24 #include <linux/compiler.h>
25 #include <linux/smp.h>
27 #include <asm/atomic.h>
28 #include <asm/cacheflush.h>
30 #include <asm/processor.h>
31 #include <asm/system.h>
32 #include <asm/hardirq.h>
33 #include <asm/mmu_context.h>
35 #include <asm/mipsregs.h>
36 #include <asm/mipsmtregs.h>
37 #include <asm/mips_mt.h>
39 #define MIPS_CPU_IPI_RESCHED_IRQ 0
40 #define MIPS_CPU_IPI_CALL_IRQ 1
42 static int cpu_ipi_resched_irq
, cpu_ipi_call_irq
;
45 static void dump_mtregisters(int vpe
, int tc
)
47 printk("vpe %d tc %d\n", vpe
, tc
);
51 printk(" c0 status 0x%lx\n", read_vpe_c0_status());
52 printk(" vpecontrol 0x%lx\n", read_vpe_c0_vpecontrol());
53 printk(" vpeconf0 0x%lx\n", read_vpe_c0_vpeconf0());
54 printk(" tcstatus 0x%lx\n", read_tc_c0_tcstatus());
55 printk(" tcrestart 0x%lx\n", read_tc_c0_tcrestart());
56 printk(" tcbind 0x%lx\n", read_tc_c0_tcbind());
57 printk(" tchalt 0x%lx\n", read_tc_c0_tchalt());
61 void __init
sanitize_tlb_entries(void)
64 unsigned long mvpconf0
, ncpu
;
70 set_c0_mvpcontrol(MVPCONTROL_VPC
);
72 back_to_back_c0_hazard();
74 /* Disable TLB sharing */
75 clear_c0_mvpcontrol(MVPCONTROL_STLB
);
77 mvpconf0
= read_c0_mvpconf0();
79 printk(KERN_INFO
"MVPConf0 0x%lx TLBS %lx PTLBE %ld\n", mvpconf0
,
80 (mvpconf0
& MVPCONF0_TLBS
) >> MVPCONF0_TLBS_SHIFT
,
81 (mvpconf0
& MVPCONF0_PTLBE
) >> MVPCONF0_PTLBE_SHIFT
);
83 tlbsiz
= (mvpconf0
& MVPCONF0_PTLBE
) >> MVPCONF0_PTLBE_SHIFT
;
84 ncpu
= ((mvpconf0
& MVPCONF0_PVPE
) >> MVPCONF0_PVPE_SHIFT
) + 1;
86 printk(" tlbsiz %d ncpu %ld\n", tlbsiz
, ncpu
);
89 /* share them out across the vpe's */
92 printk(KERN_INFO
"setting Config1.MMU_size to %d\n", tlbsiz
);
94 for (i
= 0; i
< ncpu
; i
++) {
98 write_c0_config1((read_c0_config1() & ~(0x3f << 25)) | (tlbsiz
<< 25));
100 write_vpe_c0_config1((read_vpe_c0_config1() & ~(0x3f << 25)) |
105 clear_c0_mvpcontrol(MVPCONTROL_VPC
);
108 static void ipi_resched_dispatch(void)
110 do_IRQ(MIPS_CPU_IRQ_BASE
+ MIPS_CPU_IPI_RESCHED_IRQ
);
113 static void ipi_call_dispatch(void)
115 do_IRQ(MIPS_CPU_IRQ_BASE
+ MIPS_CPU_IPI_CALL_IRQ
);
118 static irqreturn_t
ipi_resched_interrupt(int irq
, void *dev_id
)
123 static irqreturn_t
ipi_call_interrupt(int irq
, void *dev_id
)
125 smp_call_function_interrupt();
130 static struct irqaction irq_resched
= {
131 .handler
= ipi_resched_interrupt
,
132 .flags
= IRQF_DISABLED
|IRQF_PERCPU
,
133 .name
= "IPI_resched"
136 static struct irqaction irq_call
= {
137 .handler
= ipi_call_interrupt
,
138 .flags
= IRQF_DISABLED
|IRQF_PERCPU
,
142 static void __init
smp_copy_vpe_config(void)
145 (read_c0_status() & ~(ST0_IM
| ST0_IE
| ST0_KSU
)) | ST0_CU0
);
147 /* set config to be the same as vpe0, particularly kseg0 coherency alg */
148 write_vpe_c0_config( read_c0_config());
150 /* make sure there are no software interrupts pending */
151 write_vpe_c0_cause(0);
153 /* Propagate Config7 */
154 write_vpe_c0_config7(read_c0_config7());
156 write_vpe_c0_count(read_c0_count());
159 static unsigned int __init
smp_vpe_init(unsigned int tc
, unsigned int mvpconf0
,
162 if (tc
> ((mvpconf0
& MVPCONF0_PVPE
) >> MVPCONF0_PVPE_SHIFT
))
165 /* Deactivate all but VPE 0 */
167 unsigned long tmp
= read_vpe_c0_vpeconf0();
169 tmp
&= ~VPECONF0_VPA
;
173 write_vpe_c0_vpeconf0(tmp
);
175 /* Record this as available CPU */
176 cpu_set(tc
, phys_cpu_present_map
);
177 __cpu_number_map
[tc
] = ++ncpu
;
178 __cpu_logical_map
[ncpu
] = tc
;
181 /* Disable multi-threading with TC's */
182 write_vpe_c0_vpecontrol(read_vpe_c0_vpecontrol() & ~VPECONTROL_TE
);
185 smp_copy_vpe_config();
190 static void __init
smp_tc_init(unsigned int tc
, unsigned int mvpconf0
)
197 /* bind a TC to each VPE, May as well put all excess TC's
199 if (tc
>= (((mvpconf0
& MVPCONF0_PVPE
) >> MVPCONF0_PVPE_SHIFT
)+1))
200 write_tc_c0_tcbind(read_tc_c0_tcbind() | ((mvpconf0
& MVPCONF0_PVPE
) >> MVPCONF0_PVPE_SHIFT
));
202 write_tc_c0_tcbind(read_tc_c0_tcbind() | tc
);
205 write_vpe_c0_vpeconf0(read_vpe_c0_vpeconf0() | (tc
<< VPECONF0_XTC_SHIFT
));
208 tmp
= read_tc_c0_tcstatus();
210 /* mark not allocated and not dynamically allocatable */
211 tmp
&= ~(TCSTATUS_A
| TCSTATUS_DA
);
212 tmp
|= TCSTATUS_IXMT
; /* interrupt exempt */
213 write_tc_c0_tcstatus(tmp
);
215 write_tc_c0_tchalt(TCHALT_H
);
218 static void vsmp_send_ipi_single(int cpu
, unsigned int action
)
224 local_irq_save(flags
);
226 vpflags
= dvpe(); /* cant access the other CPU's registers whilst MVPE enabled */
229 case SMP_CALL_FUNCTION
:
233 case SMP_RESCHEDULE_YOURSELF
:
239 /* 1:1 mapping of vpe and tc... */
241 write_vpe_c0_cause(read_vpe_c0_cause() | i
);
244 local_irq_restore(flags
);
247 static void vsmp_send_ipi_mask(cpumask_t mask
, unsigned int action
)
251 for_each_cpu_mask(i
, mask
)
252 vsmp_send_ipi_single(i
, action
);
255 static void __cpuinit
vsmp_init_secondary(void)
257 /* Enable per-cpu interrupts */
259 /* This is Malta specific: IPI,performance and timer inetrrupts */
260 write_c0_status((read_c0_status() & ~ST0_IM
) |
261 (STATUSF_IP0
| STATUSF_IP1
| STATUSF_IP6
| STATUSF_IP7
));
264 static void __cpuinit
vsmp_smp_finish(void)
266 write_c0_compare(read_c0_count() + (8* mips_hpt_frequency
/HZ
));
268 #ifdef CONFIG_MIPS_MT_FPAFF
269 /* If we have an FPU, enroll ourselves in the FPU-full mask */
271 cpu_set(smp_processor_id(), mt_fpu_cpumask
);
272 #endif /* CONFIG_MIPS_MT_FPAFF */
277 static void vsmp_cpus_done(void)
282 * Setup the PC, SP, and GP of a secondary processor and start it
284 * smp_bootstrap is the place to resume from
285 * __KSTK_TOS(idle) is apparently the stack pointer
286 * (unsigned long)idle->thread_info the gp
287 * assumes a 1:1 mapping of TC => VPE
289 static void __cpuinit
vsmp_boot_secondary(int cpu
, struct task_struct
*idle
)
291 struct thread_info
*gp
= task_thread_info(idle
);
293 set_c0_mvpcontrol(MVPCONTROL_VPC
);
298 write_tc_c0_tcrestart((unsigned long)&smp_bootstrap
);
300 /* enable the tc this vpe/cpu will be running */
301 write_tc_c0_tcstatus((read_tc_c0_tcstatus() & ~TCSTATUS_IXMT
) | TCSTATUS_A
);
303 write_tc_c0_tchalt(0);
306 write_vpe_c0_vpeconf0(read_vpe_c0_vpeconf0() | VPECONF0_VPA
);
309 write_tc_gpr_sp( __KSTK_TOS(idle
));
312 write_tc_gpr_gp((unsigned long)gp
);
314 flush_icache_range((unsigned long)gp
,
315 (unsigned long)(gp
+ sizeof(struct thread_info
)));
317 /* finally out of configuration and into chaos */
318 clear_c0_mvpcontrol(MVPCONTROL_VPC
);
324 * Common setup before any secondaries are started
325 * Make sure all CPU's are in a sensible state before we boot any of the
328 static void __init
vsmp_smp_setup(void)
330 unsigned int mvpconf0
, ntc
, tc
, ncpu
= 0;
333 #ifdef CONFIG_MIPS_MT_FPAFF
334 /* If we have an FPU, enroll ourselves in the FPU-full mask */
336 cpu_set(0, mt_fpu_cpumask
);
337 #endif /* CONFIG_MIPS_MT_FPAFF */
341 /* disable MT so we can configure */
345 /* Put MVPE's into 'configuration state' */
346 set_c0_mvpcontrol(MVPCONTROL_VPC
);
348 mvpconf0
= read_c0_mvpconf0();
349 ntc
= (mvpconf0
& MVPCONF0_PTC
) >> MVPCONF0_PTC_SHIFT
;
351 nvpe
= ((mvpconf0
& MVPCONF0_PVPE
) >> MVPCONF0_PVPE_SHIFT
) + 1;
352 smp_num_siblings
= nvpe
;
354 /* we'll always have more TC's than VPE's, so loop setting everything
355 to a sensible state */
356 for (tc
= 0; tc
<= ntc
; tc
++) {
359 smp_tc_init(tc
, mvpconf0
);
360 ncpu
= smp_vpe_init(tc
, mvpconf0
, ncpu
);
363 /* Release config state */
364 clear_c0_mvpcontrol(MVPCONTROL_VPC
);
366 /* We'll wait until starting the secondaries before starting MVPE */
368 printk(KERN_INFO
"Detected %i available secondary CPU(s)\n", ncpu
);
371 static void __init
vsmp_prepare_cpus(unsigned int max_cpus
)
373 mips_mt_set_cpuoptions();
375 /* set up ipi interrupts */
377 set_vi_handler(MIPS_CPU_IPI_RESCHED_IRQ
, ipi_resched_dispatch
);
378 set_vi_handler(MIPS_CPU_IPI_CALL_IRQ
, ipi_call_dispatch
);
381 cpu_ipi_resched_irq
= MIPS_CPU_IRQ_BASE
+ MIPS_CPU_IPI_RESCHED_IRQ
;
382 cpu_ipi_call_irq
= MIPS_CPU_IRQ_BASE
+ MIPS_CPU_IPI_CALL_IRQ
;
384 setup_irq(cpu_ipi_resched_irq
, &irq_resched
);
385 setup_irq(cpu_ipi_call_irq
, &irq_call
);
387 set_irq_handler(cpu_ipi_resched_irq
, handle_percpu_irq
);
388 set_irq_handler(cpu_ipi_call_irq
, handle_percpu_irq
);
391 struct plat_smp_ops vsmp_smp_ops
= {
392 .send_ipi_single
= vsmp_send_ipi_single
,
393 .send_ipi_mask
= vsmp_send_ipi_mask
,
394 .init_secondary
= vsmp_init_secondary
,
395 .smp_finish
= vsmp_smp_finish
,
396 .cpus_done
= vsmp_cpus_done
,
397 .boot_secondary
= vsmp_boot_secondary
,
398 .smp_setup
= vsmp_smp_setup
,
399 .prepare_cpus
= vsmp_prepare_cpus
,