x86: prepare for tlb merge
[deliverable/linux.git] / arch / x86 / kernel / irqinit_32.c
CommitLineData
1da177e4
LT
1#include <linux/errno.h>
2#include <linux/signal.h>
3#include <linux/sched.h>
4#include <linux/ioport.h>
5#include <linux/interrupt.h>
6#include <linux/slab.h>
7#include <linux/random.h>
1da177e4
LT
8#include <linux/init.h>
9#include <linux/kernel_stat.h>
10#include <linux/sysdev.h>
11#include <linux/bitops.h>
aa09e6cd
JSR
12#include <linux/io.h>
13#include <linux/delay.h>
1da177e4 14
1da177e4
LT
15#include <asm/atomic.h>
16#include <asm/system.h>
1da177e4
LT
17#include <asm/timer.h>
18#include <asm/pgtable.h>
1da177e4
LT
19#include <asm/desc.h>
20#include <asm/apic.h>
21#include <asm/arch_hooks.h>
22#include <asm/i8259.h>
aa09e6cd 23#include <asm/traps.h>
1da177e4 24
1da177e4
LT
25
26/*
27 * Note that on a 486, we don't want to do a SIGFPE on an irq13
28 * as the irq is unreliable, and exception 16 works correctly
29 * (ie as explained in the intel literature). On a 386, you
30 * can't use exception 16 due to bad IBM design, so we have to
31 * rely on the less exact irq13.
32 *
33 * Careful.. Not only is IRQ13 unreliable, but it is also
34 * leads to races. IBM designers who came up with it should
35 * be shot.
36 */
1da177e4 37
7d12e780 38static irqreturn_t math_error_irq(int cpl, void *dev_id)
1da177e4 39{
aa09e6cd 40 outb(0, 0xF0);
1da177e4
LT
41 if (ignore_fpu_irq || !boot_cpu_data.hard_math)
42 return IRQ_NONE;
65ea5b03 43 math_error((void __user *)get_irq_regs()->ip);
1da177e4
LT
44 return IRQ_HANDLED;
45}
46
47/*
48 * New motherboards sometimes make IRQ 13 be a PCI interrupt,
49 * so allow interrupt sharing.
50 */
6a61f6a5
TG
51static struct irqaction fpu_irq = {
52 .handler = math_error_irq,
53 .mask = CPU_MASK_NONE,
54 .name = "fpu",
55};
1da177e4 56
aa09e6cd 57void __init init_ISA_irqs(void)
1da177e4
LT
58{
59 int i;
60
61#ifdef CONFIG_X86_LOCAL_APIC
62 init_bsp_APIC();
63#endif
64 init_8259A(0);
65
7c6357da
AD
66 /*
67 * 16 old-style INTA-cycle interrupts:
68 */
99d093d1 69 for (i = 0; i < NR_IRQS_LEGACY; i++) {
ee32c973 70 struct irq_desc *desc = irq_to_desc(i);
199751d7
YL
71
72 desc->status = IRQ_DISABLED;
73 desc->action = NULL;
74 desc->depth = 1;
75
7c6357da
AD
76 set_irq_chip_and_handler_name(i, &i8259A_chip,
77 handle_level_irq, "XT");
1da177e4
LT
78 }
79}
80
2ae111cd
CG
81/*
82 * IRQ2 is cascade interrupt to second interrupt controller
83 */
84static struct irqaction irq2 = {
85 .handler = no_action,
86 .mask = CPU_MASK_NONE,
87 .name = "cascade",
88};
89
497c9a19
YL
90DEFINE_PER_CPU(vector_irq_t, vector_irq) = {
91 [0 ... IRQ0_VECTOR - 1] = -1,
92 [IRQ0_VECTOR] = 0,
93 [IRQ1_VECTOR] = 1,
94 [IRQ2_VECTOR] = 2,
95 [IRQ3_VECTOR] = 3,
96 [IRQ4_VECTOR] = 4,
97 [IRQ5_VECTOR] = 5,
98 [IRQ6_VECTOR] = 6,
99 [IRQ7_VECTOR] = 7,
100 [IRQ8_VECTOR] = 8,
101 [IRQ9_VECTOR] = 9,
102 [IRQ10_VECTOR] = 10,
103 [IRQ11_VECTOR] = 11,
104 [IRQ12_VECTOR] = 12,
105 [IRQ13_VECTOR] = 13,
106 [IRQ14_VECTOR] = 14,
107 [IRQ15_VECTOR] = 15,
108 [IRQ15_VECTOR + 1 ... NR_VECTORS - 1] = -1
109};
110
b77b881f
YL
111int vector_used_by_percpu_irq(unsigned int vector)
112{
113 int cpu;
114
115 for_each_online_cpu(cpu) {
116 if (per_cpu(vector_irq, cpu)[vector] != -1)
117 return 1;
118 }
119
120 return 0;
121}
122
d3561b7f
RR
123/* Overridden in paravirt.c */
124void init_IRQ(void) __attribute__((weak, alias("native_init_IRQ")));
125
126void __init native_init_IRQ(void)
1da177e4
LT
127{
128 int i;
129
130 /* all the set up before the call gates are initialised */
131 pre_intr_init_hook();
132
133 /*
134 * Cover the whole vector space, no vector can escape
135 * us. (some of these will be overridden and become
136 * 'special' SMP interrupts)
137 */
497c9a19 138 for (i = FIRST_EXTERNAL_VECTOR; i < NR_VECTORS; i++) {
dbeb2be2 139 /* SYSCALL_VECTOR was reserved in trap_init. */
497c9a19 140 if (i != SYSCALL_VECTOR)
4687518c 141 set_intr_gate(i, interrupt[i-FIRST_EXTERNAL_VECTOR]);
1da177e4
LT
142 }
143
2ae111cd 144
497c9a19 145#if defined(CONFIG_X86_LOCAL_APIC) && defined(CONFIG_SMP)
2ae111cd
CG
146 /*
147 * The reschedule interrupt is a CPU-to-CPU reschedule-helper
148 * IPI, driven by wakeup.
149 */
150 alloc_intr_gate(RESCHEDULE_VECTOR, reschedule_interrupt);
151
152 /* IPI for invalidation */
153 alloc_intr_gate(INVALIDATE_TLB_VECTOR, invalidate_interrupt);
154
155 /* IPI for generic function call */
156 alloc_intr_gate(CALL_FUNCTION_VECTOR, call_function_interrupt);
157
158 /* IPI for single call function */
b77b881f
YL
159 alloc_intr_gate(CALL_FUNCTION_SINGLE_VECTOR,
160 call_function_single_interrupt);
497c9a19
YL
161
162 /* Low priority IPI to cleanup after moving an irq */
163 set_intr_gate(IRQ_MOVE_CLEANUP_VECTOR, irq_move_cleanup_interrupt);
b77b881f 164 set_bit(IRQ_MOVE_CLEANUP_VECTOR, used_vectors);
2ae111cd
CG
165#endif
166
167#ifdef CONFIG_X86_LOCAL_APIC
168 /* self generated IPI for local APIC timer */
169 alloc_intr_gate(LOCAL_TIMER_VECTOR, apic_timer_interrupt);
170
171 /* IPI vectors for APIC spurious and error interrupts */
172 alloc_intr_gate(SPURIOUS_APIC_VECTOR, spurious_interrupt);
173 alloc_intr_gate(ERROR_APIC_VECTOR, error_interrupt);
174#endif
175
176#if defined(CONFIG_X86_LOCAL_APIC) && defined(CONFIG_X86_MCE_P4THERMAL)
177 /* thermal monitor LVT interrupt */
178 alloc_intr_gate(THERMAL_APIC_VECTOR, thermal_interrupt);
179#endif
180
181 if (!acpi_ioapic)
182 setup_irq(2, &irq2);
183
1da177e4
LT
184 /* setup after call gates are initialised (usually add in
185 * the architecture specific gates)
186 */
187 intr_init_hook();
188
1da177e4
LT
189 /*
190 * External FPU? Set up irq13 if so, for
191 * original braindamaged IBM FERR coupling.
192 */
193 if (boot_cpu_data.hard_math && !cpu_has_fpu)
194 setup_irq(FPU_IRQ, &fpu_irq);
195
196 irq_ctx_init(smp_processor_id());
197}
This page took 0.391668 seconds and 5 git commands to generate.