2 * This file is subject to the terms and conditions of the GNU General Public
3 * License. See the file "COPYING" in the main directory of this archive
6 * Copyright (C) 2008 Ralf Baechle (ralf@linux-mips.org)
7 * Copyright (C) 2012 MIPS Technologies, Inc. All rights reserved.
9 #include <linux/bitmap.h>
10 #include <linux/init.h>
11 #include <linux/smp.h>
12 #include <linux/irq.h>
13 #include <linux/clocksource.h>
17 #include <asm/setup.h>
18 #include <asm/traps.h>
19 #include <asm/gcmpregs.h>
20 #include <linux/hardirq.h>
21 #include <asm-generic/bitops/find.h>
23 unsigned int gic_frequency
;
24 unsigned int gic_present
;
25 unsigned long _gic_base
;
26 unsigned int gic_irq_base
;
27 unsigned int gic_irq_flags
[GIC_NUM_INTRS
];
29 /* The index into this array is the vector # of the interrupt. */
30 struct gic_shared_intr_map gic_shared_intr_map
[GIC_NUM_INTRS
];
32 static struct gic_pcpu_mask pcpu_masks
[NR_CPUS
];
33 static struct gic_pending_regs pending_regs
[NR_CPUS
];
34 static struct gic_intrmask_regs intrmask_regs
[NR_CPUS
];
36 #if defined(CONFIG_CSRC_GIC) || defined(CONFIG_CEVT_GIC)
37 cycle_t
gic_read_count(void)
39 unsigned int hi
, hi2
, lo
;
42 GICREAD(GIC_REG(SHARED
, GIC_SH_COUNTER_63_32
), hi
);
43 GICREAD(GIC_REG(SHARED
, GIC_SH_COUNTER_31_00
), lo
);
44 GICREAD(GIC_REG(SHARED
, GIC_SH_COUNTER_63_32
), hi2
);
47 return (((cycle_t
) hi
) << 32) + lo
;
50 void gic_write_compare(cycle_t cnt
)
52 GICWRITE(GIC_REG(VPE_LOCAL
, GIC_VPE_COMPARE_HI
),
54 GICWRITE(GIC_REG(VPE_LOCAL
, GIC_VPE_COMPARE_LO
),
55 (int)(cnt
& 0xffffffff));
58 cycle_t
gic_read_compare(void)
62 GICREAD(GIC_REG(VPE_LOCAL
, GIC_VPE_COMPARE_HI
), hi
);
63 GICREAD(GIC_REG(VPE_LOCAL
, GIC_VPE_COMPARE_LO
), lo
);
65 return (((cycle_t
) hi
) << 32) + lo
;
69 unsigned int gic_get_timer_pending(void)
71 unsigned int vpe_pending
;
73 GICWRITE(GIC_REG(VPE_LOCAL
, GIC_VPE_OTHER_ADDR
), 0);
74 GICREAD(GIC_REG(VPE_OTHER
, GIC_VPE_PEND
), vpe_pending
);
75 return (vpe_pending
& GIC_VPE_PEND_TIMER_MSK
);
78 void gic_bind_eic_interrupt(int irq
, int set
)
80 /* Convert irq vector # to hw int # */
81 irq
-= GIC_PIN_TO_VEC_OFFSET
;
83 /* Set irq to use shadow set */
84 GICWRITE(GIC_REG_ADDR(VPE_LOCAL
, GIC_VPE_EIC_SS(irq
)), set
);
87 void gic_send_ipi(unsigned int intr
)
89 GICWRITE(GIC_REG(SHARED
, GIC_SH_WEDGE
), 0x80000000 | intr
);
92 static void gic_eic_irq_dispatch(void)
94 unsigned int cause
= read_c0_cause();
97 irq
= (cause
& ST0_IM
) >> STATUSB_IP2
;
102 do_IRQ(gic_irq_base
+ irq
);
104 spurious_interrupt();
107 static void __init
vpe_local_setup(unsigned int numvpes
)
109 unsigned long timer_intr
= GIC_INT_TMR
;
110 unsigned long perf_intr
= GIC_INT_PERFCTR
;
111 unsigned int vpe_ctl
;
116 * GIC timer interrupt -> CPU HW Int X (vector X+2) ->
117 * map to pin X+2-1 (since GIC adds 1)
119 timer_intr
+= (GIC_CPU_TO_VEC_OFFSET
- GIC_PIN_TO_VEC_OFFSET
);
121 * GIC perfcnt interrupt -> CPU HW Int X (vector X+2) ->
122 * map to pin X+2-1 (since GIC adds 1)
124 perf_intr
+= (GIC_CPU_TO_VEC_OFFSET
- GIC_PIN_TO_VEC_OFFSET
);
128 * Setup the default performance counter timer interrupts
131 for (i
= 0; i
< numvpes
; i
++) {
132 GICWRITE(GIC_REG(VPE_LOCAL
, GIC_VPE_OTHER_ADDR
), i
);
134 /* Are Interrupts locally routable? */
135 GICREAD(GIC_REG(VPE_OTHER
, GIC_VPE_CTL
), vpe_ctl
);
136 if (vpe_ctl
& GIC_VPE_CTL_TIMER_RTBL_MSK
)
137 GICWRITE(GIC_REG(VPE_OTHER
, GIC_VPE_TIMER_MAP
),
138 GIC_MAP_TO_PIN_MSK
| timer_intr
);
140 set_vi_handler(timer_intr
+ GIC_PIN_TO_VEC_OFFSET
,
141 gic_eic_irq_dispatch
);
142 gic_shared_intr_map
[timer_intr
+ GIC_PIN_TO_VEC_OFFSET
].local_intr_mask
|= GIC_VPE_RMASK_TIMER_MSK
;
145 if (vpe_ctl
& GIC_VPE_CTL_PERFCNT_RTBL_MSK
)
146 GICWRITE(GIC_REG(VPE_OTHER
, GIC_VPE_PERFCTR_MAP
),
147 GIC_MAP_TO_PIN_MSK
| perf_intr
);
149 set_vi_handler(perf_intr
+ GIC_PIN_TO_VEC_OFFSET
, gic_eic_irq_dispatch
);
150 gic_shared_intr_map
[perf_intr
+ GIC_PIN_TO_VEC_OFFSET
].local_intr_mask
|= GIC_VPE_RMASK_PERFCNT_MSK
;
155 unsigned int gic_compare_int(void)
157 unsigned int pending
;
159 GICREAD(GIC_REG(VPE_LOCAL
, GIC_VPE_PEND
), pending
);
160 if (pending
& GIC_VPE_PEND_CMP_MSK
)
166 unsigned int gic_get_int(void)
169 unsigned long *pending
, *intrmask
, *pcpu_mask
;
170 unsigned long *pending_abs
, *intrmask_abs
;
172 /* Get per-cpu bitmaps */
173 pending
= pending_regs
[smp_processor_id()].pending
;
174 intrmask
= intrmask_regs
[smp_processor_id()].intrmask
;
175 pcpu_mask
= pcpu_masks
[smp_processor_id()].pcpu_mask
;
177 pending_abs
= (unsigned long *) GIC_REG_ABS_ADDR(SHARED
,
178 GIC_SH_PEND_31_0_OFS
);
179 intrmask_abs
= (unsigned long *) GIC_REG_ABS_ADDR(SHARED
,
180 GIC_SH_MASK_31_0_OFS
);
182 for (i
= 0; i
< BITS_TO_LONGS(GIC_NUM_INTRS
); i
++) {
183 GICREAD(*pending_abs
, pending
[i
]);
184 GICREAD(*intrmask_abs
, intrmask
[i
]);
189 bitmap_and(pending
, pending
, intrmask
, GIC_NUM_INTRS
);
190 bitmap_and(pending
, pending
, pcpu_mask
, GIC_NUM_INTRS
);
192 return find_first_bit(pending
, GIC_NUM_INTRS
);
195 static void gic_mask_irq(struct irq_data
*d
)
197 GIC_CLR_INTR_MASK(d
->irq
- gic_irq_base
);
200 static void gic_unmask_irq(struct irq_data
*d
)
202 GIC_SET_INTR_MASK(d
->irq
- gic_irq_base
);
206 static DEFINE_SPINLOCK(gic_lock
);
208 static int gic_set_affinity(struct irq_data
*d
, const struct cpumask
*cpumask
,
211 unsigned int irq
= (d
->irq
- gic_irq_base
);
212 cpumask_t tmp
= CPU_MASK_NONE
;
216 cpumask_and(&tmp
, cpumask
, cpu_online_mask
);
220 /* Assumption : cpumask refers to a single CPU */
221 spin_lock_irqsave(&gic_lock
, flags
);
223 /* Re-route this IRQ */
224 GIC_SH_MAP_TO_VPE_SMASK(irq
, first_cpu(tmp
));
226 /* Update the pcpu_masks */
227 for (i
= 0; i
< NR_CPUS
; i
++)
228 clear_bit(irq
, pcpu_masks
[i
].pcpu_mask
);
229 set_bit(irq
, pcpu_masks
[first_cpu(tmp
)].pcpu_mask
);
232 cpumask_copy(d
->affinity
, cpumask
);
233 spin_unlock_irqrestore(&gic_lock
, flags
);
235 return IRQ_SET_MASK_OK_NOCOPY
;
239 static struct irq_chip gic_irq_controller
= {
241 .irq_ack
= gic_irq_ack
,
242 .irq_mask
= gic_mask_irq
,
243 .irq_mask_ack
= gic_mask_irq
,
244 .irq_unmask
= gic_unmask_irq
,
245 .irq_eoi
= gic_finish_irq
,
247 .irq_set_affinity
= gic_set_affinity
,
251 static void __init
gic_setup_intr(unsigned int intr
, unsigned int cpu
,
252 unsigned int pin
, unsigned int polarity
, unsigned int trigtype
,
255 struct gic_shared_intr_map
*map_ptr
;
257 /* Setup Intr to Pin mapping */
258 if (pin
& GIC_MAP_TO_NMI_MSK
) {
259 GICWRITE(GIC_REG_ADDR(SHARED
, GIC_SH_MAP_TO_PIN(intr
)), pin
);
260 /* FIXME: hack to route NMI to all cpu's */
261 for (cpu
= 0; cpu
< NR_CPUS
; cpu
+= 32) {
262 GICWRITE(GIC_REG_ADDR(SHARED
,
263 GIC_SH_MAP_TO_VPE_REG_OFF(intr
, cpu
)),
267 GICWRITE(GIC_REG_ADDR(SHARED
, GIC_SH_MAP_TO_PIN(intr
)),
268 GIC_MAP_TO_PIN_MSK
| pin
);
269 /* Setup Intr to CPU mapping */
270 GIC_SH_MAP_TO_VPE_SMASK(intr
, cpu
);
272 set_vi_handler(pin
+ GIC_PIN_TO_VEC_OFFSET
,
273 gic_eic_irq_dispatch
);
274 map_ptr
= &gic_shared_intr_map
[pin
+ GIC_PIN_TO_VEC_OFFSET
];
275 if (map_ptr
->num_shared_intr
>= GIC_MAX_SHARED_INTR
)
277 map_ptr
->intr_list
[map_ptr
->num_shared_intr
++] = intr
;
281 /* Setup Intr Polarity */
282 GIC_SET_POLARITY(intr
, polarity
);
284 /* Setup Intr Trigger Type */
285 GIC_SET_TRIGGER(intr
, trigtype
);
287 /* Init Intr Masks */
288 GIC_CLR_INTR_MASK(intr
);
289 /* Initialise per-cpu Interrupt software masks */
290 if (flags
& GIC_FLAG_IPI
)
291 set_bit(intr
, pcpu_masks
[cpu
].pcpu_mask
);
292 if ((flags
& GIC_FLAG_TRANSPARENT
) && (cpu_has_veic
== 0))
293 GIC_SET_INTR_MASK(intr
);
294 if (trigtype
== GIC_TRIG_EDGE
)
295 gic_irq_flags
[intr
] |= GIC_TRIG_EDGE
;
298 static void __init
gic_basic_init(int numintrs
, int numvpes
,
299 struct gic_intr_map
*intrmap
, int mapsize
)
302 unsigned int pin_offset
= 0;
304 board_bind_eic_interrupt
= &gic_bind_eic_interrupt
;
307 for (i
= 0; i
< numintrs
; i
++) {
308 GIC_SET_POLARITY(i
, GIC_POL_POS
);
309 GIC_SET_TRIGGER(i
, GIC_TRIG_LEVEL
);
310 GIC_CLR_INTR_MASK(i
);
311 if (i
< GIC_NUM_INTRS
) {
312 gic_irq_flags
[i
] = 0;
313 gic_shared_intr_map
[i
].num_shared_intr
= 0;
314 gic_shared_intr_map
[i
].local_intr_mask
= 0;
319 * In EIC mode, the HW_INT# is offset by (2-1). Need to subtract
320 * one because the GIC will add one (since 0=no intr).
323 pin_offset
= (GIC_CPU_TO_VEC_OFFSET
- GIC_PIN_TO_VEC_OFFSET
);
325 /* Setup specifics */
326 for (i
= 0; i
< mapsize
; i
++) {
327 cpu
= intrmap
[i
].cpunum
;
328 if (cpu
== GIC_UNUSED
)
330 if (cpu
== 0 && i
!= 0 && intrmap
[i
].flags
== 0)
334 intrmap
[i
].pin
+ pin_offset
,
340 vpe_local_setup(numvpes
);
343 void __init
gic_init(unsigned long gic_base_addr
,
344 unsigned long gic_addrspace_size
,
345 struct gic_intr_map
*intr_map
, unsigned int intr_map_size
,
346 unsigned int irqbase
)
348 unsigned int gicconfig
;
349 int numvpes
, numintrs
;
351 _gic_base
= (unsigned long) ioremap_nocache(gic_base_addr
,
353 gic_irq_base
= irqbase
;
355 GICREAD(GIC_REG(SHARED
, GIC_SH_CONFIG
), gicconfig
);
356 numintrs
= (gicconfig
& GIC_SH_CONFIG_NUMINTRS_MSK
) >>
357 GIC_SH_CONFIG_NUMINTRS_SHF
;
358 numintrs
= ((numintrs
+ 1) * 8);
360 numvpes
= (gicconfig
& GIC_SH_CONFIG_NUMVPES_MSK
) >>
361 GIC_SH_CONFIG_NUMVPES_SHF
;
362 numvpes
= numvpes
+ 1;
364 gic_basic_init(numintrs
, numvpes
, intr_map
, intr_map_size
);
366 gic_platform_init(numintrs
, &gic_irq_controller
);