2 * linux/kernel/irq/chip.c
4 * Copyright (C) 1992, 1998-2006 Linus Torvalds, Ingo Molnar
5 * Copyright (C) 2005-2006, Thomas Gleixner, Russell King
7 * This file contains the core interrupt handling code, for irq-chip
10 * Detailed information is available in Documentation/DocBook/genericirq
13 #include <linux/irq.h>
14 #include <linux/msi.h>
15 #include <linux/module.h>
16 #include <linux/interrupt.h>
17 #include <linux/kernel_stat.h>
19 #include "internals.h"
22 * irq_set_chip - set the irq chip for an irq
24 * @chip: pointer to irq chip description structure
26 int irq_set_chip(unsigned int irq
, struct irq_chip
*chip
)
29 struct irq_desc
*desc
= irq_get_desc_lock(irq
, &flags
, 0);
37 desc
->irq_data
.chip
= chip
;
38 irq_put_desc_unlock(desc
, flags
);
40 * For !CONFIG_SPARSE_IRQ make the irq show up in
41 * allocated_irqs. For the CONFIG_SPARSE_IRQ case, it is
42 * already marked, and this call is harmless.
47 EXPORT_SYMBOL(irq_set_chip
);
50 * irq_set_type - set the irq trigger type for an irq
52 * @type: IRQ_TYPE_{LEVEL,EDGE}_* value - see include/linux/irq.h
54 int irq_set_irq_type(unsigned int irq
, unsigned int type
)
57 struct irq_desc
*desc
= irq_get_desc_buslock(irq
, &flags
, IRQ_GET_DESC_CHECK_GLOBAL
);
63 type
&= IRQ_TYPE_SENSE_MASK
;
64 if (type
!= IRQ_TYPE_NONE
)
65 ret
= __irq_set_trigger(desc
, irq
, type
);
66 irq_put_desc_busunlock(desc
, flags
);
69 EXPORT_SYMBOL(irq_set_irq_type
);
72 * irq_set_handler_data - set irq handler data for an irq
73 * @irq: Interrupt number
74 * @data: Pointer to interrupt specific data
76 * Set the hardware irq controller data for an irq
78 int irq_set_handler_data(unsigned int irq
, void *data
)
81 struct irq_desc
*desc
= irq_get_desc_lock(irq
, &flags
, 0);
85 desc
->irq_data
.handler_data
= data
;
86 irq_put_desc_unlock(desc
, flags
);
89 EXPORT_SYMBOL(irq_set_handler_data
);
92 * irq_set_msi_desc - set MSI descriptor data for an irq
93 * @irq: Interrupt number
94 * @entry: Pointer to MSI descriptor data
96 * Set the MSI descriptor entry for an irq
98 int irq_set_msi_desc(unsigned int irq
, struct msi_desc
*entry
)
101 struct irq_desc
*desc
= irq_get_desc_lock(irq
, &flags
, IRQ_GET_DESC_CHECK_GLOBAL
);
105 desc
->irq_data
.msi_desc
= entry
;
108 irq_put_desc_unlock(desc
, flags
);
113 * irq_set_chip_data - set irq chip data for an irq
114 * @irq: Interrupt number
115 * @data: Pointer to chip specific data
117 * Set the hardware irq chip data for an irq
119 int irq_set_chip_data(unsigned int irq
, void *data
)
122 struct irq_desc
*desc
= irq_get_desc_lock(irq
, &flags
, 0);
126 desc
->irq_data
.chip_data
= data
;
127 irq_put_desc_unlock(desc
, flags
);
130 EXPORT_SYMBOL(irq_set_chip_data
);
132 struct irq_data
*irq_get_irq_data(unsigned int irq
)
134 struct irq_desc
*desc
= irq_to_desc(irq
);
136 return desc
? &desc
->irq_data
: NULL
;
138 EXPORT_SYMBOL_GPL(irq_get_irq_data
);
140 static void irq_state_clr_disabled(struct irq_desc
*desc
)
142 irqd_clear(&desc
->irq_data
, IRQD_IRQ_DISABLED
);
145 static void irq_state_set_disabled(struct irq_desc
*desc
)
147 irqd_set(&desc
->irq_data
, IRQD_IRQ_DISABLED
);
150 static void irq_state_clr_masked(struct irq_desc
*desc
)
152 irqd_clear(&desc
->irq_data
, IRQD_IRQ_MASKED
);
155 static void irq_state_set_masked(struct irq_desc
*desc
)
157 irqd_set(&desc
->irq_data
, IRQD_IRQ_MASKED
);
160 int irq_startup(struct irq_desc
*desc
)
162 irq_state_clr_disabled(desc
);
165 if (desc
->irq_data
.chip
->irq_startup
) {
166 int ret
= desc
->irq_data
.chip
->irq_startup(&desc
->irq_data
);
167 irq_state_clr_masked(desc
);
175 void irq_shutdown(struct irq_desc
*desc
)
177 irq_state_set_disabled(desc
);
179 if (desc
->irq_data
.chip
->irq_shutdown
)
180 desc
->irq_data
.chip
->irq_shutdown(&desc
->irq_data
);
181 else if (desc
->irq_data
.chip
->irq_disable
)
182 desc
->irq_data
.chip
->irq_disable(&desc
->irq_data
);
184 desc
->irq_data
.chip
->irq_mask(&desc
->irq_data
);
185 irq_state_set_masked(desc
);
188 void irq_enable(struct irq_desc
*desc
)
190 irq_state_clr_disabled(desc
);
191 if (desc
->irq_data
.chip
->irq_enable
)
192 desc
->irq_data
.chip
->irq_enable(&desc
->irq_data
);
194 desc
->irq_data
.chip
->irq_unmask(&desc
->irq_data
);
195 irq_state_clr_masked(desc
);
198 void irq_disable(struct irq_desc
*desc
)
200 irq_state_set_disabled(desc
);
201 if (desc
->irq_data
.chip
->irq_disable
) {
202 desc
->irq_data
.chip
->irq_disable(&desc
->irq_data
);
203 irq_state_set_masked(desc
);
207 void irq_percpu_enable(struct irq_desc
*desc
, unsigned int cpu
)
209 if (desc
->irq_data
.chip
->irq_enable
)
210 desc
->irq_data
.chip
->irq_enable(&desc
->irq_data
);
212 desc
->irq_data
.chip
->irq_unmask(&desc
->irq_data
);
213 cpumask_set_cpu(cpu
, desc
->percpu_enabled
);
216 void irq_percpu_disable(struct irq_desc
*desc
, unsigned int cpu
)
218 if (desc
->irq_data
.chip
->irq_disable
)
219 desc
->irq_data
.chip
->irq_disable(&desc
->irq_data
);
221 desc
->irq_data
.chip
->irq_mask(&desc
->irq_data
);
222 cpumask_clear_cpu(cpu
, desc
->percpu_enabled
);
225 static inline void mask_ack_irq(struct irq_desc
*desc
)
227 if (desc
->irq_data
.chip
->irq_mask_ack
)
228 desc
->irq_data
.chip
->irq_mask_ack(&desc
->irq_data
);
230 desc
->irq_data
.chip
->irq_mask(&desc
->irq_data
);
231 if (desc
->irq_data
.chip
->irq_ack
)
232 desc
->irq_data
.chip
->irq_ack(&desc
->irq_data
);
234 irq_state_set_masked(desc
);
237 void mask_irq(struct irq_desc
*desc
)
239 if (desc
->irq_data
.chip
->irq_mask
) {
240 desc
->irq_data
.chip
->irq_mask(&desc
->irq_data
);
241 irq_state_set_masked(desc
);
245 void unmask_irq(struct irq_desc
*desc
)
247 if (desc
->irq_data
.chip
->irq_unmask
) {
248 desc
->irq_data
.chip
->irq_unmask(&desc
->irq_data
);
249 irq_state_clr_masked(desc
);
254 * handle_nested_irq - Handle a nested irq from a irq thread
255 * @irq: the interrupt number
257 * Handle interrupts which are nested into a threaded interrupt
258 * handler. The handler function is called inside the calling
261 void handle_nested_irq(unsigned int irq
)
263 struct irq_desc
*desc
= irq_to_desc(irq
);
264 struct irqaction
*action
;
265 irqreturn_t action_ret
;
269 raw_spin_lock_irq(&desc
->lock
);
271 kstat_incr_irqs_this_cpu(irq
, desc
);
273 action
= desc
->action
;
274 if (unlikely(!action
|| irqd_irq_disabled(&desc
->irq_data
)))
277 irqd_set(&desc
->irq_data
, IRQD_IRQ_INPROGRESS
);
278 raw_spin_unlock_irq(&desc
->lock
);
280 action_ret
= action
->thread_fn(action
->irq
, action
->dev_id
);
282 note_interrupt(irq
, desc
, action_ret
);
284 raw_spin_lock_irq(&desc
->lock
);
285 irqd_clear(&desc
->irq_data
, IRQD_IRQ_INPROGRESS
);
288 raw_spin_unlock_irq(&desc
->lock
);
290 EXPORT_SYMBOL_GPL(handle_nested_irq
);
292 static bool irq_check_poll(struct irq_desc
*desc
)
294 if (!(desc
->istate
& IRQS_POLL_INPROGRESS
))
296 return irq_wait_for_poll(desc
);
300 * handle_simple_irq - Simple and software-decoded IRQs.
301 * @irq: the interrupt number
302 * @desc: the interrupt description structure for this irq
304 * Simple interrupts are either sent from a demultiplexing interrupt
305 * handler or come from hardware, where no interrupt hardware control
308 * Note: The caller is expected to handle the ack, clear, mask and
309 * unmask issues if necessary.
312 handle_simple_irq(unsigned int irq
, struct irq_desc
*desc
)
314 raw_spin_lock(&desc
->lock
);
316 if (unlikely(irqd_irq_inprogress(&desc
->irq_data
)))
317 if (!irq_check_poll(desc
))
320 desc
->istate
&= ~(IRQS_REPLAY
| IRQS_WAITING
);
321 kstat_incr_irqs_this_cpu(irq
, desc
);
323 if (unlikely(!desc
->action
|| irqd_irq_disabled(&desc
->irq_data
)))
326 handle_irq_event(desc
);
329 raw_spin_unlock(&desc
->lock
);
331 EXPORT_SYMBOL_GPL(handle_simple_irq
);
334 * handle_level_irq - Level type irq handler
335 * @irq: the interrupt number
336 * @desc: the interrupt description structure for this irq
338 * Level type interrupts are active as long as the hardware line has
339 * the active level. This may require to mask the interrupt and unmask
340 * it after the associated handler has acknowledged the device, so the
341 * interrupt line is back to inactive.
344 handle_level_irq(unsigned int irq
, struct irq_desc
*desc
)
346 raw_spin_lock(&desc
->lock
);
349 if (unlikely(irqd_irq_inprogress(&desc
->irq_data
)))
350 if (!irq_check_poll(desc
))
353 desc
->istate
&= ~(IRQS_REPLAY
| IRQS_WAITING
);
354 kstat_incr_irqs_this_cpu(irq
, desc
);
357 * If its disabled or no action available
358 * keep it masked and get out of here
360 if (unlikely(!desc
->action
|| irqd_irq_disabled(&desc
->irq_data
)))
363 handle_irq_event(desc
);
365 if (!irqd_irq_disabled(&desc
->irq_data
) && !(desc
->istate
& IRQS_ONESHOT
))
368 raw_spin_unlock(&desc
->lock
);
370 EXPORT_SYMBOL_GPL(handle_level_irq
);
372 #ifdef CONFIG_IRQ_PREFLOW_FASTEOI
373 static inline void preflow_handler(struct irq_desc
*desc
)
375 if (desc
->preflow_handler
)
376 desc
->preflow_handler(&desc
->irq_data
);
379 static inline void preflow_handler(struct irq_desc
*desc
) { }
383 * handle_fasteoi_irq - irq handler for transparent controllers
384 * @irq: the interrupt number
385 * @desc: the interrupt description structure for this irq
387 * Only a single callback will be issued to the chip: an ->eoi()
388 * call when the interrupt has been serviced. This enables support
389 * for modern forms of interrupt handlers, which handle the flow
390 * details in hardware, transparently.
393 handle_fasteoi_irq(unsigned int irq
, struct irq_desc
*desc
)
395 raw_spin_lock(&desc
->lock
);
397 if (unlikely(irqd_irq_inprogress(&desc
->irq_data
)))
398 if (!irq_check_poll(desc
))
401 desc
->istate
&= ~(IRQS_REPLAY
| IRQS_WAITING
);
402 kstat_incr_irqs_this_cpu(irq
, desc
);
405 * If its disabled or no action available
406 * then mask it and get out of here:
408 if (unlikely(!desc
->action
|| irqd_irq_disabled(&desc
->irq_data
))) {
409 desc
->istate
|= IRQS_PENDING
;
414 if (desc
->istate
& IRQS_ONESHOT
)
417 preflow_handler(desc
);
418 handle_irq_event(desc
);
421 desc
->irq_data
.chip
->irq_eoi(&desc
->irq_data
);
423 raw_spin_unlock(&desc
->lock
);
426 if (!(desc
->irq_data
.chip
->flags
& IRQCHIP_EOI_IF_HANDLED
))
432 * handle_edge_irq - edge type IRQ handler
433 * @irq: the interrupt number
434 * @desc: the interrupt description structure for this irq
436 * Interrupt occures on the falling and/or rising edge of a hardware
437 * signal. The occurrence is latched into the irq controller hardware
438 * and must be acked in order to be reenabled. After the ack another
439 * interrupt can happen on the same source even before the first one
440 * is handled by the associated event handler. If this happens it
441 * might be necessary to disable (mask) the interrupt depending on the
442 * controller hardware. This requires to reenable the interrupt inside
443 * of the loop which handles the interrupts which have arrived while
444 * the handler was running. If all pending interrupts are handled, the
448 handle_edge_irq(unsigned int irq
, struct irq_desc
*desc
)
450 raw_spin_lock(&desc
->lock
);
452 desc
->istate
&= ~(IRQS_REPLAY
| IRQS_WAITING
);
454 * If we're currently running this IRQ, or its disabled,
455 * we shouldn't process the IRQ. Mark it pending, handle
456 * the necessary masking and go out
458 if (unlikely(irqd_irq_disabled(&desc
->irq_data
) ||
459 irqd_irq_inprogress(&desc
->irq_data
) || !desc
->action
)) {
460 if (!irq_check_poll(desc
)) {
461 desc
->istate
|= IRQS_PENDING
;
466 kstat_incr_irqs_this_cpu(irq
, desc
);
468 /* Start handling the irq */
469 desc
->irq_data
.chip
->irq_ack(&desc
->irq_data
);
472 if (unlikely(!desc
->action
)) {
478 * When another irq arrived while we were handling
479 * one, we could have masked the irq.
480 * Renable it, if it was not disabled in meantime.
482 if (unlikely(desc
->istate
& IRQS_PENDING
)) {
483 if (!irqd_irq_disabled(&desc
->irq_data
) &&
484 irqd_irq_masked(&desc
->irq_data
))
488 handle_irq_event(desc
);
490 } while ((desc
->istate
& IRQS_PENDING
) &&
491 !irqd_irq_disabled(&desc
->irq_data
));
494 raw_spin_unlock(&desc
->lock
);
497 #ifdef CONFIG_IRQ_EDGE_EOI_HANDLER
499 * handle_edge_eoi_irq - edge eoi type IRQ handler
500 * @irq: the interrupt number
501 * @desc: the interrupt description structure for this irq
503 * Similar as the above handle_edge_irq, but using eoi and w/o the
506 void handle_edge_eoi_irq(unsigned int irq
, struct irq_desc
*desc
)
508 struct irq_chip
*chip
= irq_desc_get_chip(desc
);
510 raw_spin_lock(&desc
->lock
);
512 desc
->istate
&= ~(IRQS_REPLAY
| IRQS_WAITING
);
514 * If we're currently running this IRQ, or its disabled,
515 * we shouldn't process the IRQ. Mark it pending, handle
516 * the necessary masking and go out
518 if (unlikely(irqd_irq_disabled(&desc
->irq_data
) ||
519 irqd_irq_inprogress(&desc
->irq_data
) || !desc
->action
)) {
520 if (!irq_check_poll(desc
)) {
521 desc
->istate
|= IRQS_PENDING
;
525 kstat_incr_irqs_this_cpu(irq
, desc
);
528 if (unlikely(!desc
->action
))
531 handle_irq_event(desc
);
533 } while ((desc
->istate
& IRQS_PENDING
) &&
534 !irqd_irq_disabled(&desc
->irq_data
));
537 chip
->irq_eoi(&desc
->irq_data
);
538 raw_spin_unlock(&desc
->lock
);
543 * handle_percpu_irq - Per CPU local irq handler
544 * @irq: the interrupt number
545 * @desc: the interrupt description structure for this irq
547 * Per CPU interrupts on SMP machines without locking requirements
550 handle_percpu_irq(unsigned int irq
, struct irq_desc
*desc
)
552 struct irq_chip
*chip
= irq_desc_get_chip(desc
);
554 kstat_incr_irqs_this_cpu(irq
, desc
);
557 chip
->irq_ack(&desc
->irq_data
);
559 handle_irq_event_percpu(desc
, desc
->action
);
562 chip
->irq_eoi(&desc
->irq_data
);
566 * handle_percpu_devid_irq - Per CPU local irq handler with per cpu dev ids
567 * @irq: the interrupt number
568 * @desc: the interrupt description structure for this irq
570 * Per CPU interrupts on SMP machines without locking requirements. Same as
571 * handle_percpu_irq() above but with the following extras:
573 * action->percpu_dev_id is a pointer to percpu variables which
574 * contain the real device id for the cpu on which this handler is
577 void handle_percpu_devid_irq(unsigned int irq
, struct irq_desc
*desc
)
579 struct irq_chip
*chip
= irq_desc_get_chip(desc
);
580 struct irqaction
*action
= desc
->action
;
581 void *dev_id
= __this_cpu_ptr(action
->percpu_dev_id
);
584 kstat_incr_irqs_this_cpu(irq
, desc
);
587 chip
->irq_ack(&desc
->irq_data
);
589 trace_irq_handler_entry(irq
, action
);
590 res
= action
->handler(irq
, dev_id
);
591 trace_irq_handler_exit(irq
, action
, res
);
594 chip
->irq_eoi(&desc
->irq_data
);
598 __irq_set_handler(unsigned int irq
, irq_flow_handler_t handle
, int is_chained
,
602 struct irq_desc
*desc
= irq_get_desc_buslock(irq
, &flags
, 0);
608 handle
= handle_bad_irq
;
610 if (WARN_ON(desc
->irq_data
.chip
== &no_irq_chip
))
615 if (handle
== handle_bad_irq
) {
616 if (desc
->irq_data
.chip
!= &no_irq_chip
)
618 irq_state_set_disabled(desc
);
621 desc
->handle_irq
= handle
;
624 if (handle
!= handle_bad_irq
&& is_chained
) {
625 irq_settings_set_noprobe(desc
);
626 irq_settings_set_norequest(desc
);
627 irq_settings_set_nothread(desc
);
631 irq_put_desc_busunlock(desc
, flags
);
633 EXPORT_SYMBOL_GPL(__irq_set_handler
);
636 irq_set_chip_and_handler_name(unsigned int irq
, struct irq_chip
*chip
,
637 irq_flow_handler_t handle
, const char *name
)
639 irq_set_chip(irq
, chip
);
640 __irq_set_handler(irq
, handle
, 0, name
);
643 void irq_modify_status(unsigned int irq
, unsigned long clr
, unsigned long set
)
646 struct irq_desc
*desc
= irq_get_desc_lock(irq
, &flags
, 0);
650 irq_settings_clr_and_set(desc
, clr
, set
);
652 irqd_clear(&desc
->irq_data
, IRQD_NO_BALANCING
| IRQD_PER_CPU
|
653 IRQD_TRIGGER_MASK
| IRQD_LEVEL
| IRQD_MOVE_PCNTXT
);
654 if (irq_settings_has_no_balance_set(desc
))
655 irqd_set(&desc
->irq_data
, IRQD_NO_BALANCING
);
656 if (irq_settings_is_per_cpu(desc
))
657 irqd_set(&desc
->irq_data
, IRQD_PER_CPU
);
658 if (irq_settings_can_move_pcntxt(desc
))
659 irqd_set(&desc
->irq_data
, IRQD_MOVE_PCNTXT
);
660 if (irq_settings_is_level(desc
))
661 irqd_set(&desc
->irq_data
, IRQD_LEVEL
);
663 irqd_set(&desc
->irq_data
, irq_settings_get_trigger_mask(desc
));
665 irq_put_desc_unlock(desc
, flags
);
667 EXPORT_SYMBOL_GPL(irq_modify_status
);
670 * irq_cpu_online - Invoke all irq_cpu_online functions.
672 * Iterate through all irqs and invoke the chip.irq_cpu_online()
675 void irq_cpu_online(void)
677 struct irq_desc
*desc
;
678 struct irq_chip
*chip
;
682 for_each_active_irq(irq
) {
683 desc
= irq_to_desc(irq
);
687 raw_spin_lock_irqsave(&desc
->lock
, flags
);
689 chip
= irq_data_get_irq_chip(&desc
->irq_data
);
690 if (chip
&& chip
->irq_cpu_online
&&
691 (!(chip
->flags
& IRQCHIP_ONOFFLINE_ENABLED
) ||
692 !irqd_irq_disabled(&desc
->irq_data
)))
693 chip
->irq_cpu_online(&desc
->irq_data
);
695 raw_spin_unlock_irqrestore(&desc
->lock
, flags
);
700 * irq_cpu_offline - Invoke all irq_cpu_offline functions.
702 * Iterate through all irqs and invoke the chip.irq_cpu_offline()
705 void irq_cpu_offline(void)
707 struct irq_desc
*desc
;
708 struct irq_chip
*chip
;
712 for_each_active_irq(irq
) {
713 desc
= irq_to_desc(irq
);
717 raw_spin_lock_irqsave(&desc
->lock
, flags
);
719 chip
= irq_data_get_irq_chip(&desc
->irq_data
);
720 if (chip
&& chip
->irq_cpu_offline
&&
721 (!(chip
->flags
& IRQCHIP_ONOFFLINE_ENABLED
) ||
722 !irqd_irq_disabled(&desc
->irq_data
)))
723 chip
->irq_cpu_offline(&desc
->irq_data
);
725 raw_spin_unlock_irqrestore(&desc
->lock
, flags
);