1 // SPDX-License-Identifier: GPL-2.0
3 * ip30-irq.c: Highlevel interrupt handling for IP30 architecture.
5 #include <linux/errno.h>
6 #include <linux/init.h>
7 #include <linux/interrupt.h>
9 #include <linux/percpu.h>
10 #include <linux/spinlock.h>
11 #include <linux/tick.h>
12 #include <linux/types.h>
14 #include <asm/irq_cpu.h>
15 #include <asm/sgi/heart.h>
17 #include "ip30-common.h"
19 struct heart_irq_data
{
24 static DECLARE_BITMAP(heart_irq_map
, HEART_NUM_IRQS
);
26 static DEFINE_PER_CPU(unsigned long, irq_enable_mask
);
28 static inline int heart_alloc_int(void)
33 bit
= find_first_zero_bit(heart_irq_map
, HEART_NUM_IRQS
);
34 if (bit
>= HEART_NUM_IRQS
)
37 if (test_and_set_bit(bit
, heart_irq_map
))
43 static void ip30_error_irq(struct irq_desc
*desc
)
45 u64 pending
, mask
, cause
, error_irqs
, err_reg
;
46 int cpu
= smp_processor_id();
49 pending
= heart_read(&heart_regs
->isr
);
50 mask
= heart_read(&heart_regs
->imr
[cpu
]);
51 cause
= heart_read(&heart_regs
->cause
);
52 error_irqs
= (pending
& HEART_L4_INT_MASK
& mask
);
54 /* Bail if there's nothing to process (how did we get here, then?) */
55 if (unlikely(!error_irqs
))
58 /* Prevent any of the error IRQs from firing again. */
59 heart_write(mask
& ~(pending
), &heart_regs
->imr
[cpu
]);
61 /* Ack all error IRQs. */
62 heart_write(HEART_L4_INT_MASK
, &heart_regs
->clear_isr
);
65 * If we also have a cause value, then something happened, so loop
66 * through the error IRQs and report a "heart attack" for each one
67 * and print the value of the HEART cause register. This is really
68 * primitive right now, but it should hopefully work until a more
69 * robust error handling routine can be put together.
71 * Refer to heart.h for the HC_* macros to work out the cause
75 pr_alert("IP30: CPU%d: HEART ATTACK! ISR = 0x%.16llx, IMR = 0x%.16llx, CAUSE = 0x%.16llx\n",
76 cpu
, pending
, mask
, cause
);
78 if (cause
& HC_COR_MEM_ERR
) {
79 err_reg
= heart_read(&heart_regs
->mem_err_addr
);
80 pr_alert(" HEART_MEMERR_ADDR = 0x%.16llx\n", err_reg
);
83 /* i = 63; i >= 51; i-- */
84 for (i
= HEART_ERR_MASK_END
; i
>= HEART_ERR_MASK_START
; i
--)
85 if ((pending
>> i
) & 1)
86 pr_alert(" HEART Error IRQ #%d\n", i
);
88 /* XXX: Seems possible to loop forever here, so panic(). */
89 panic("IP30: Fatal Error !\n");
92 /* Unmask the error IRQs. */
93 heart_write(mask
, &heart_regs
->imr
[cpu
]);
96 static void ip30_normal_irq(struct irq_desc
*desc
)
98 int cpu
= smp_processor_id();
99 struct irq_domain
*domain
;
103 pend
= heart_read(&heart_regs
->isr
);
104 mask
= (heart_read(&heart_regs
->imr
[cpu
]) &
105 (HEART_L0_INT_MASK
| HEART_L1_INT_MASK
| HEART_L2_INT_MASK
));
112 if (pend
& BIT_ULL(HEART_L2_INT_RESCHED_CPU_0
)) {
113 heart_write(BIT_ULL(HEART_L2_INT_RESCHED_CPU_0
),
114 &heart_regs
->clear_isr
);
116 } else if (pend
& BIT_ULL(HEART_L2_INT_RESCHED_CPU_1
)) {
117 heart_write(BIT_ULL(HEART_L2_INT_RESCHED_CPU_1
),
118 &heart_regs
->clear_isr
);
120 } else if (pend
& BIT_ULL(HEART_L2_INT_CALL_CPU_0
)) {
121 heart_write(BIT_ULL(HEART_L2_INT_CALL_CPU_0
),
122 &heart_regs
->clear_isr
);
123 generic_smp_call_function_interrupt();
124 } else if (pend
& BIT_ULL(HEART_L2_INT_CALL_CPU_1
)) {
125 heart_write(BIT_ULL(HEART_L2_INT_CALL_CPU_1
),
126 &heart_regs
->clear_isr
);
127 generic_smp_call_function_interrupt();
131 domain
= irq_desc_get_handler_data(desc
);
132 irq
= irq_linear_revmap(domain
, __ffs(pend
));
134 generic_handle_irq(irq
);
136 spurious_interrupt();
140 static void ip30_ack_heart_irq(struct irq_data
*d
)
142 heart_write(BIT_ULL(d
->hwirq
), &heart_regs
->clear_isr
);
145 static void ip30_mask_heart_irq(struct irq_data
*d
)
147 struct heart_irq_data
*hd
= irq_data_get_irq_chip_data(d
);
148 unsigned long *mask
= &per_cpu(irq_enable_mask
, hd
->cpu
);
150 clear_bit(d
->hwirq
, mask
);
151 heart_write(*mask
, &heart_regs
->imr
[hd
->cpu
]);
154 static void ip30_mask_and_ack_heart_irq(struct irq_data
*d
)
156 struct heart_irq_data
*hd
= irq_data_get_irq_chip_data(d
);
157 unsigned long *mask
= &per_cpu(irq_enable_mask
, hd
->cpu
);
159 clear_bit(d
->hwirq
, mask
);
160 heart_write(*mask
, &heart_regs
->imr
[hd
->cpu
]);
161 heart_write(BIT_ULL(d
->hwirq
), &heart_regs
->clear_isr
);
164 static void ip30_unmask_heart_irq(struct irq_data
*d
)
166 struct heart_irq_data
*hd
= irq_data_get_irq_chip_data(d
);
167 unsigned long *mask
= &per_cpu(irq_enable_mask
, hd
->cpu
);
169 set_bit(d
->hwirq
, mask
);
170 heart_write(*mask
, &heart_regs
->imr
[hd
->cpu
]);
173 static int ip30_set_heart_irq_affinity(struct irq_data
*d
,
174 const struct cpumask
*mask
, bool force
)
176 struct heart_irq_data
*hd
= irq_data_get_irq_chip_data(d
);
181 if (irqd_is_started(d
))
182 ip30_mask_and_ack_heart_irq(d
);
184 hd
->cpu
= cpumask_first_and(mask
, cpu_online_mask
);
186 if (irqd_is_started(d
))
187 ip30_unmask_heart_irq(d
);
189 irq_data_update_effective_affinity(d
, cpumask_of(hd
->cpu
));
194 static struct irq_chip heart_irq_chip
= {
196 .irq_ack
= ip30_ack_heart_irq
,
197 .irq_mask
= ip30_mask_heart_irq
,
198 .irq_mask_ack
= ip30_mask_and_ack_heart_irq
,
199 .irq_unmask
= ip30_unmask_heart_irq
,
200 .irq_set_affinity
= ip30_set_heart_irq_affinity
,
203 static int heart_domain_alloc(struct irq_domain
*domain
, unsigned int virq
,
204 unsigned int nr_irqs
, void *arg
)
206 struct irq_alloc_info
*info
= arg
;
207 struct heart_irq_data
*hd
;
210 if (nr_irqs
> 1 || !info
)
213 hd
= kzalloc(sizeof(*hd
), GFP_KERNEL
);
217 hwirq
= heart_alloc_int();
222 irq_domain_set_info(domain
, virq
, hwirq
, &heart_irq_chip
, hd
,
223 handle_level_irq
, NULL
, NULL
);
228 static void heart_domain_free(struct irq_domain
*domain
,
229 unsigned int virq
, unsigned int nr_irqs
)
231 struct irq_data
*irqd
;
236 irqd
= irq_domain_get_irq_data(domain
, virq
);
238 clear_bit(irqd
->hwirq
, heart_irq_map
);
239 kfree(irqd
->chip_data
);
243 static const struct irq_domain_ops heart_domain_ops
= {
244 .alloc
= heart_domain_alloc
,
245 .free
= heart_domain_free
,
248 void __init
ip30_install_ipi(void)
250 int cpu
= smp_processor_id();
251 unsigned long *mask
= &per_cpu(irq_enable_mask
, cpu
);
253 set_bit(HEART_L2_INT_RESCHED_CPU_0
+ cpu
, mask
);
254 heart_write(BIT_ULL(HEART_L2_INT_RESCHED_CPU_0
+ cpu
),
255 &heart_regs
->clear_isr
);
256 set_bit(HEART_L2_INT_CALL_CPU_0
+ cpu
, mask
);
257 heart_write(BIT_ULL(HEART_L2_INT_CALL_CPU_0
+ cpu
),
258 &heart_regs
->clear_isr
);
260 heart_write(*mask
, &heart_regs
->imr
[cpu
]);
263 void __init
arch_init_irq(void)
265 struct irq_domain
*domain
;
266 struct fwnode_handle
*fn
;
273 heart_write(HEART_CLR_ALL_MASK
, &heart_regs
->imr
[0]);
274 heart_write(HEART_CLR_ALL_MASK
, &heart_regs
->imr
[1]);
275 heart_write(HEART_CLR_ALL_MASK
, &heart_regs
->imr
[2]);
276 heart_write(HEART_CLR_ALL_MASK
, &heart_regs
->imr
[3]);
278 /* Ack everything. */
279 heart_write(HEART_ACK_ALL_MASK
, &heart_regs
->clear_isr
);
281 /* Enable specific HEART error IRQs for each CPU. */
282 mask
= &per_cpu(irq_enable_mask
, 0);
283 *mask
|= HEART_CPU0_ERR_MASK
;
284 heart_write(*mask
, &heart_regs
->imr
[0]);
285 mask
= &per_cpu(irq_enable_mask
, 1);
286 *mask
|= HEART_CPU1_ERR_MASK
;
287 heart_write(*mask
, &heart_regs
->imr
[1]);
290 * Some HEART bits are reserved by hardware or by software convention.
291 * Mark these as reserved right away so they won't be accidentally
294 set_bit(HEART_L0_INT_GENERIC
, heart_irq_map
);
295 set_bit(HEART_L0_INT_FLOW_CTRL_HWTR_0
, heart_irq_map
);
296 set_bit(HEART_L0_INT_FLOW_CTRL_HWTR_1
, heart_irq_map
);
297 set_bit(HEART_L2_INT_RESCHED_CPU_0
, heart_irq_map
);
298 set_bit(HEART_L2_INT_RESCHED_CPU_1
, heart_irq_map
);
299 set_bit(HEART_L2_INT_CALL_CPU_0
, heart_irq_map
);
300 set_bit(HEART_L2_INT_CALL_CPU_1
, heart_irq_map
);
301 set_bit(HEART_L3_INT_TIMER
, heart_irq_map
);
303 /* Reserve the error interrupts (#51 to #63). */
304 for (i
= HEART_L4_INT_XWID_ERR_9
; i
<= HEART_L4_INT_HEART_EXCP
; i
++)
305 set_bit(i
, heart_irq_map
);
307 fn
= irq_domain_alloc_named_fwnode("HEART");
311 domain
= irq_domain_create_linear(fn
, HEART_NUM_IRQS
,
312 &heart_domain_ops
, NULL
);
313 WARN_ON(domain
== NULL
);
317 irq_set_default_host(domain
);
319 irq_set_percpu_devid(IP30_HEART_L0_IRQ
);
320 irq_set_chained_handler_and_data(IP30_HEART_L0_IRQ
, ip30_normal_irq
,
322 irq_set_percpu_devid(IP30_HEART_L1_IRQ
);
323 irq_set_chained_handler_and_data(IP30_HEART_L1_IRQ
, ip30_normal_irq
,
325 irq_set_percpu_devid(IP30_HEART_L2_IRQ
);
326 irq_set_chained_handler_and_data(IP30_HEART_L2_IRQ
, ip30_normal_irq
,
328 irq_set_percpu_devid(IP30_HEART_ERR_IRQ
);
329 irq_set_chained_handler_and_data(IP30_HEART_ERR_IRQ
, ip30_error_irq
,