1 #include <linux/linkage.h>
2 #include <linux/errno.h>
3 #include <linux/signal.h>
4 #include <linux/sched.h>
5 #include <linux/ioport.h>
6 #include <linux/interrupt.h>
7 #include <linux/timex.h>
8 #include <linux/random.h>
9 #include <linux/kprobes.h>
10 #include <linux/init.h>
11 #include <linux/kernel_stat.h>
12 #include <linux/device.h>
13 #include <linux/bitops.h>
14 #include <linux/acpi.h>
16 #include <linux/delay.h>
18 #include <linux/atomic.h>
19 #include <asm/timer.h>
20 #include <asm/hw_irq.h>
21 #include <asm/pgtable.h>
24 #include <asm/setup.h>
25 #include <asm/i8259.h>
26 #include <asm/traps.h>
30 * ISA PIC or low IO-APIC triggered (INTA-cycle or APIC) interrupts:
31 * (these are usually mapped to vectors 0x30-0x3f)
35 * The IO-APIC gives us many more interrupt sources. Most of these
36 * are unused but an SMP system is supposed to have enough memory ...
37 * sometimes (mostly wrt. hw bugs) we get corrupted vectors all
38 * across the spectrum, so we really want to be prepared to get all
39 * of these. Plus, more powerful systems might have more than 64
42 * (these are usually mapped into the 0x30-0xff vector range)
46 * IRQ2 is cascade interrupt to second interrupt controller
48 static struct irqaction irq2
= {
51 .flags
= IRQF_NO_THREAD
,
54 DEFINE_PER_CPU(vector_irq_t
, vector_irq
) = {
55 [0 ... NR_VECTORS
- 1] = VECTOR_UNUSED
,
58 int vector_used_by_percpu_irq(unsigned int vector
)
62 for_each_online_cpu(cpu
) {
63 if (!IS_ERR_OR_NULL(per_cpu(vector_irq
, cpu
)[vector
]))
70 void __init
init_ISA_irqs(void)
72 struct irq_chip
*chip
= legacy_pic
->chip
;
75 #if defined(CONFIG_X86_64) || defined(CONFIG_X86_LOCAL_APIC)
80 for (i
= 0; i
< nr_legacy_irqs(); i
++)
81 irq_set_chip_and_handler(i
, chip
, handle_level_irq
);
84 void __init
init_IRQ(void)
89 * On cpu 0, Assign ISA_IRQ_VECTOR(irq) to IRQ 0..15.
90 * If these IRQ's are handled by legacy interrupt-controllers like PIC,
91 * then this configuration will likely be static after the boot. If
92 * these IRQ's are handled by more mordern controllers like IO-APIC,
93 * then this vector space can be freed and re-used dynamically as the
96 for (i
= 0; i
< nr_legacy_irqs(); i
++)
97 per_cpu(vector_irq
, 0)[ISA_IRQ_VECTOR(i
)] = irq_to_desc(i
);
99 x86_init
.irqs
.intr_init();
102 static void __init
smp_intr_init(void)
106 * The reschedule interrupt is a CPU-to-CPU reschedule-helper
107 * IPI, driven by wakeup.
109 alloc_intr_gate(RESCHEDULE_VECTOR
, reschedule_interrupt
);
111 /* IPI for generic function call */
112 alloc_intr_gate(CALL_FUNCTION_VECTOR
, call_function_interrupt
);
114 /* IPI for generic single function call */
115 alloc_intr_gate(CALL_FUNCTION_SINGLE_VECTOR
,
116 call_function_single_interrupt
);
118 /* Low priority IPI to cleanup after moving an irq */
119 set_intr_gate(IRQ_MOVE_CLEANUP_VECTOR
, irq_move_cleanup_interrupt
);
120 set_bit(IRQ_MOVE_CLEANUP_VECTOR
, used_vectors
);
122 /* IPI used for rebooting/stopping */
123 alloc_intr_gate(REBOOT_VECTOR
, reboot_interrupt
);
124 #endif /* CONFIG_SMP */
127 static void __init
apic_intr_init(void)
131 #ifdef CONFIG_X86_THERMAL_VECTOR
132 alloc_intr_gate(THERMAL_APIC_VECTOR
, thermal_interrupt
);
134 #ifdef CONFIG_X86_MCE_THRESHOLD
135 alloc_intr_gate(THRESHOLD_APIC_VECTOR
, threshold_interrupt
);
138 #ifdef CONFIG_X86_MCE_AMD
139 alloc_intr_gate(DEFERRED_ERROR_VECTOR
, deferred_error_interrupt
);
142 #ifdef CONFIG_X86_LOCAL_APIC
143 /* self generated IPI for local APIC timer */
144 alloc_intr_gate(LOCAL_TIMER_VECTOR
, apic_timer_interrupt
);
146 /* IPI for X86 platform specific use */
147 alloc_intr_gate(X86_PLATFORM_IPI_VECTOR
, x86_platform_ipi
);
148 #ifdef CONFIG_HAVE_KVM
149 /* IPI for KVM to deliver posted interrupt */
150 alloc_intr_gate(POSTED_INTR_VECTOR
, kvm_posted_intr_ipi
);
151 /* IPI for KVM to deliver interrupt to wake up tasks */
152 alloc_intr_gate(POSTED_INTR_WAKEUP_VECTOR
, kvm_posted_intr_wakeup_ipi
);
155 /* IPI vectors for APIC spurious and error interrupts */
156 alloc_intr_gate(SPURIOUS_APIC_VECTOR
, spurious_interrupt
);
157 alloc_intr_gate(ERROR_APIC_VECTOR
, error_interrupt
);
159 /* IRQ work interrupts: */
160 # ifdef CONFIG_IRQ_WORK
161 alloc_intr_gate(IRQ_WORK_VECTOR
, irq_work_interrupt
);
167 void __init
native_init_IRQ(void)
171 /* Execute any quirks before the call gates are initialised: */
172 x86_init
.irqs
.pre_vector_init();
177 * Cover the whole vector space, no vector can escape
178 * us. (some of these will be overridden and become
179 * 'special' SMP interrupts)
181 i
= FIRST_EXTERNAL_VECTOR
;
182 #ifndef CONFIG_X86_LOCAL_APIC
183 #define first_system_vector NR_VECTORS
185 for_each_clear_bit_from(i
, used_vectors
, first_system_vector
) {
186 /* IA32_SYSCALL_VECTOR could be used in trap_init already. */
187 set_intr_gate(i
, irq_entries_start
+
188 8 * (i
- FIRST_EXTERNAL_VECTOR
));
190 #ifdef CONFIG_X86_LOCAL_APIC
191 for_each_clear_bit_from(i
, used_vectors
, NR_VECTORS
)
192 set_intr_gate(i
, spurious_interrupt
);
195 if (!acpi_ioapic
&& !of_ioapic
&& nr_legacy_irqs())
199 irq_ctx_init(smp_processor_id());