]>
Commit | Line | Data |
---|---|---|
1da177e4 LT |
1 | /* |
2 | * linux/kernel/irq/manage.c | |
3 | * | |
4 | * Copyright (C) 1992, 1998-2004 Linus Torvalds, Ingo Molnar | |
5 | * | |
6 | * This file contains driver APIs to the irq subsystem. | |
7 | */ | |
8 | ||
b77d6adc | 9 | #include <linux/config.h> |
1da177e4 LT |
10 | #include <linux/irq.h> |
11 | #include <linux/module.h> | |
12 | #include <linux/random.h> | |
13 | #include <linux/interrupt.h> | |
14 | ||
15 | #include "internals.h" | |
16 | ||
17 | #ifdef CONFIG_SMP | |
18 | ||
19 | cpumask_t irq_affinity[NR_IRQS] = { [0 ... NR_IRQS-1] = CPU_MASK_ALL }; | |
20 | ||
54d5d424 AR |
21 | #if defined (CONFIG_GENERIC_PENDING_IRQ) || defined (CONFIG_IRQBALANCE) |
22 | cpumask_t __cacheline_aligned pending_irq_cpumask[NR_IRQS]; | |
23 | #endif | |
24 | ||
1da177e4 LT |
25 | /** |
26 | * synchronize_irq - wait for pending IRQ handlers (on other CPUs) | |
1e5d5331 | 27 | * @irq: interrupt number to wait for |
1da177e4 LT |
28 | * |
29 | * This function waits for any pending IRQ handlers for this interrupt | |
30 | * to complete before returning. If you use this function while | |
31 | * holding a resource the IRQ handler may need you will deadlock. | |
32 | * | |
33 | * This function may be called - with care - from IRQ context. | |
34 | */ | |
35 | void synchronize_irq(unsigned int irq) | |
36 | { | |
37 | struct irq_desc *desc = irq_desc + irq; | |
38 | ||
c2b5a251 MW |
39 | if (irq >= NR_IRQS) |
40 | return; | |
41 | ||
1da177e4 LT |
42 | while (desc->status & IRQ_INPROGRESS) |
43 | cpu_relax(); | |
44 | } | |
45 | ||
46 | EXPORT_SYMBOL(synchronize_irq); | |
47 | ||
48 | #endif | |
49 | ||
50 | /** | |
51 | * disable_irq_nosync - disable an irq without waiting | |
52 | * @irq: Interrupt to disable | |
53 | * | |
54 | * Disable the selected interrupt line. Disables and Enables are | |
55 | * nested. | |
56 | * Unlike disable_irq(), this function does not ensure existing | |
57 | * instances of the IRQ handler have completed before returning. | |
58 | * | |
59 | * This function may be called from IRQ context. | |
60 | */ | |
61 | void disable_irq_nosync(unsigned int irq) | |
62 | { | |
63 | irq_desc_t *desc = irq_desc + irq; | |
64 | unsigned long flags; | |
65 | ||
c2b5a251 MW |
66 | if (irq >= NR_IRQS) |
67 | return; | |
68 | ||
1da177e4 LT |
69 | spin_lock_irqsave(&desc->lock, flags); |
70 | if (!desc->depth++) { | |
71 | desc->status |= IRQ_DISABLED; | |
72 | desc->handler->disable(irq); | |
73 | } | |
74 | spin_unlock_irqrestore(&desc->lock, flags); | |
75 | } | |
76 | ||
77 | EXPORT_SYMBOL(disable_irq_nosync); | |
78 | ||
79 | /** | |
80 | * disable_irq - disable an irq and wait for completion | |
81 | * @irq: Interrupt to disable | |
82 | * | |
83 | * Disable the selected interrupt line. Enables and Disables are | |
84 | * nested. | |
85 | * This function waits for any pending IRQ handlers for this interrupt | |
86 | * to complete before returning. If you use this function while | |
87 | * holding a resource the IRQ handler may need you will deadlock. | |
88 | * | |
89 | * This function may be called - with care - from IRQ context. | |
90 | */ | |
91 | void disable_irq(unsigned int irq) | |
92 | { | |
93 | irq_desc_t *desc = irq_desc + irq; | |
94 | ||
c2b5a251 MW |
95 | if (irq >= NR_IRQS) |
96 | return; | |
97 | ||
1da177e4 LT |
98 | disable_irq_nosync(irq); |
99 | if (desc->action) | |
100 | synchronize_irq(irq); | |
101 | } | |
102 | ||
103 | EXPORT_SYMBOL(disable_irq); | |
104 | ||
105 | /** | |
106 | * enable_irq - enable handling of an irq | |
107 | * @irq: Interrupt to enable | |
108 | * | |
109 | * Undoes the effect of one call to disable_irq(). If this | |
110 | * matches the last disable, processing of interrupts on this | |
111 | * IRQ line is re-enabled. | |
112 | * | |
113 | * This function may be called from IRQ context. | |
114 | */ | |
115 | void enable_irq(unsigned int irq) | |
116 | { | |
117 | irq_desc_t *desc = irq_desc + irq; | |
118 | unsigned long flags; | |
119 | ||
c2b5a251 MW |
120 | if (irq >= NR_IRQS) |
121 | return; | |
122 | ||
1da177e4 LT |
123 | spin_lock_irqsave(&desc->lock, flags); |
124 | switch (desc->depth) { | |
125 | case 0: | |
126 | WARN_ON(1); | |
127 | break; | |
128 | case 1: { | |
129 | unsigned int status = desc->status & ~IRQ_DISABLED; | |
130 | ||
131 | desc->status = status; | |
132 | if ((status & (IRQ_PENDING | IRQ_REPLAY)) == IRQ_PENDING) { | |
133 | desc->status = status | IRQ_REPLAY; | |
134 | hw_resend_irq(desc->handler,irq); | |
135 | } | |
136 | desc->handler->enable(irq); | |
137 | /* fall-through */ | |
138 | } | |
139 | default: | |
140 | desc->depth--; | |
141 | } | |
142 | spin_unlock_irqrestore(&desc->lock, flags); | |
143 | } | |
144 | ||
145 | EXPORT_SYMBOL(enable_irq); | |
146 | ||
147 | /* | |
148 | * Internal function that tells the architecture code whether a | |
149 | * particular irq has been exclusively allocated or is available | |
150 | * for driver use. | |
151 | */ | |
152 | int can_request_irq(unsigned int irq, unsigned long irqflags) | |
153 | { | |
154 | struct irqaction *action; | |
155 | ||
156 | if (irq >= NR_IRQS) | |
157 | return 0; | |
158 | ||
159 | action = irq_desc[irq].action; | |
160 | if (action) | |
161 | if (irqflags & action->flags & SA_SHIRQ) | |
162 | action = NULL; | |
163 | ||
164 | return !action; | |
165 | } | |
166 | ||
167 | /* | |
168 | * Internal function to register an irqaction - typically used to | |
169 | * allocate special interrupts that are part of the architecture. | |
170 | */ | |
171 | int setup_irq(unsigned int irq, struct irqaction * new) | |
172 | { | |
173 | struct irq_desc *desc = irq_desc + irq; | |
174 | struct irqaction *old, **p; | |
175 | unsigned long flags; | |
176 | int shared = 0; | |
177 | ||
c2b5a251 MW |
178 | if (irq >= NR_IRQS) |
179 | return -EINVAL; | |
180 | ||
1da177e4 LT |
181 | if (desc->handler == &no_irq_type) |
182 | return -ENOSYS; | |
183 | /* | |
184 | * Some drivers like serial.c use request_irq() heavily, | |
185 | * so we have to be careful not to interfere with a | |
186 | * running system. | |
187 | */ | |
188 | if (new->flags & SA_SAMPLE_RANDOM) { | |
189 | /* | |
190 | * This function might sleep, we want to call it first, | |
191 | * outside of the atomic block. | |
192 | * Yes, this might clear the entropy pool if the wrong | |
193 | * driver is attempted to be loaded, without actually | |
194 | * installing a new handler, but is this really a problem, | |
195 | * only the sysadmin is able to do this. | |
196 | */ | |
197 | rand_initialize_irq(irq); | |
198 | } | |
199 | ||
200 | /* | |
201 | * The following block of code has to be executed atomically | |
202 | */ | |
203 | spin_lock_irqsave(&desc->lock,flags); | |
204 | p = &desc->action; | |
205 | if ((old = *p) != NULL) { | |
206 | /* Can't share interrupts unless both agree to */ | |
f5163427 DS |
207 | if (!(old->flags & new->flags & SA_SHIRQ)) |
208 | goto mismatch; | |
209 | ||
210 | #if defined(ARCH_HAS_IRQ_PER_CPU) && defined(SA_PERCPU_IRQ) | |
211 | /* All handlers must agree on per-cpuness */ | |
212 | if ((old->flags & IRQ_PER_CPU) != (new->flags & IRQ_PER_CPU)) | |
213 | goto mismatch; | |
214 | #endif | |
1da177e4 LT |
215 | |
216 | /* add new interrupt at end of irq queue */ | |
217 | do { | |
218 | p = &old->next; | |
219 | old = *p; | |
220 | } while (old); | |
221 | shared = 1; | |
222 | } | |
223 | ||
224 | *p = new; | |
f5163427 DS |
225 | #if defined(ARCH_HAS_IRQ_PER_CPU) && defined(SA_PERCPU_IRQ) |
226 | if (new->flags & SA_PERCPU_IRQ) | |
227 | desc->status |= IRQ_PER_CPU; | |
228 | #endif | |
1da177e4 LT |
229 | if (!shared) { |
230 | desc->depth = 0; | |
231 | desc->status &= ~(IRQ_DISABLED | IRQ_AUTODETECT | | |
232 | IRQ_WAITING | IRQ_INPROGRESS); | |
233 | if (desc->handler->startup) | |
234 | desc->handler->startup(irq); | |
235 | else | |
236 | desc->handler->enable(irq); | |
237 | } | |
238 | spin_unlock_irqrestore(&desc->lock,flags); | |
239 | ||
240 | new->irq = irq; | |
241 | register_irq_proc(irq); | |
242 | new->dir = NULL; | |
243 | register_handler_proc(irq, new); | |
244 | ||
245 | return 0; | |
f5163427 DS |
246 | |
247 | mismatch: | |
248 | spin_unlock_irqrestore(&desc->lock, flags); | |
249 | printk(KERN_ERR "%s: irq handler mismatch\n", __FUNCTION__); | |
250 | dump_stack(); | |
251 | return -EBUSY; | |
1da177e4 LT |
252 | } |
253 | ||
254 | /** | |
255 | * free_irq - free an interrupt | |
256 | * @irq: Interrupt line to free | |
257 | * @dev_id: Device identity to free | |
258 | * | |
259 | * Remove an interrupt handler. The handler is removed and if the | |
260 | * interrupt line is no longer in use by any driver it is disabled. | |
261 | * On a shared IRQ the caller must ensure the interrupt is disabled | |
262 | * on the card it drives before calling this function. The function | |
263 | * does not return until any executing interrupts for this IRQ | |
264 | * have completed. | |
265 | * | |
266 | * This function must not be called from interrupt context. | |
267 | */ | |
268 | void free_irq(unsigned int irq, void *dev_id) | |
269 | { | |
270 | struct irq_desc *desc; | |
271 | struct irqaction **p; | |
272 | unsigned long flags; | |
273 | ||
cd7b24bb | 274 | WARN_ON(in_interrupt()); |
1da177e4 LT |
275 | if (irq >= NR_IRQS) |
276 | return; | |
277 | ||
278 | desc = irq_desc + irq; | |
279 | spin_lock_irqsave(&desc->lock,flags); | |
280 | p = &desc->action; | |
281 | for (;;) { | |
282 | struct irqaction * action = *p; | |
283 | ||
284 | if (action) { | |
285 | struct irqaction **pp = p; | |
286 | ||
287 | p = &action->next; | |
288 | if (action->dev_id != dev_id) | |
289 | continue; | |
290 | ||
291 | /* Found it - now remove it from the list of entries */ | |
292 | *pp = action->next; | |
dbce706e | 293 | |
b77d6adc PBG |
294 | /* Currently used only by UML, might disappear one day.*/ |
295 | #ifdef CONFIG_IRQ_RELEASE_METHOD | |
dbce706e PBG |
296 | if (desc->handler->release) |
297 | desc->handler->release(irq, dev_id); | |
b77d6adc | 298 | #endif |
dbce706e | 299 | |
1da177e4 LT |
300 | if (!desc->action) { |
301 | desc->status |= IRQ_DISABLED; | |
302 | if (desc->handler->shutdown) | |
303 | desc->handler->shutdown(irq); | |
304 | else | |
305 | desc->handler->disable(irq); | |
306 | } | |
307 | spin_unlock_irqrestore(&desc->lock,flags); | |
308 | unregister_handler_proc(irq, action); | |
309 | ||
310 | /* Make sure it's not being used on another CPU */ | |
311 | synchronize_irq(irq); | |
312 | kfree(action); | |
313 | return; | |
314 | } | |
315 | printk(KERN_ERR "Trying to free free IRQ%d\n",irq); | |
316 | spin_unlock_irqrestore(&desc->lock,flags); | |
317 | return; | |
318 | } | |
319 | } | |
320 | ||
321 | EXPORT_SYMBOL(free_irq); | |
322 | ||
323 | /** | |
324 | * request_irq - allocate an interrupt line | |
325 | * @irq: Interrupt line to allocate | |
326 | * @handler: Function to be called when the IRQ occurs | |
327 | * @irqflags: Interrupt type flags | |
328 | * @devname: An ascii name for the claiming device | |
329 | * @dev_id: A cookie passed back to the handler function | |
330 | * | |
331 | * This call allocates interrupt resources and enables the | |
332 | * interrupt line and IRQ handling. From the point this | |
333 | * call is made your handler function may be invoked. Since | |
334 | * your handler function must clear any interrupt the board | |
335 | * raises, you must take care both to initialise your hardware | |
336 | * and to set up the interrupt handler in the right order. | |
337 | * | |
338 | * Dev_id must be globally unique. Normally the address of the | |
339 | * device data structure is used as the cookie. Since the handler | |
340 | * receives this value it makes sense to use it. | |
341 | * | |
342 | * If your interrupt is shared you must pass a non NULL dev_id | |
343 | * as this is required when freeing the interrupt. | |
344 | * | |
345 | * Flags: | |
346 | * | |
347 | * SA_SHIRQ Interrupt is shared | |
348 | * SA_INTERRUPT Disable local interrupts while processing | |
349 | * SA_SAMPLE_RANDOM The interrupt can be used for entropy | |
350 | * | |
351 | */ | |
352 | int request_irq(unsigned int irq, | |
353 | irqreturn_t (*handler)(int, void *, struct pt_regs *), | |
354 | unsigned long irqflags, const char * devname, void *dev_id) | |
355 | { | |
356 | struct irqaction * action; | |
357 | int retval; | |
358 | ||
359 | /* | |
360 | * Sanity-check: shared interrupts must pass in a real dev-ID, | |
361 | * otherwise we'll have trouble later trying to figure out | |
362 | * which interrupt is which (messes up the interrupt freeing | |
363 | * logic etc). | |
364 | */ | |
365 | if ((irqflags & SA_SHIRQ) && !dev_id) | |
366 | return -EINVAL; | |
367 | if (irq >= NR_IRQS) | |
368 | return -EINVAL; | |
369 | if (!handler) | |
370 | return -EINVAL; | |
371 | ||
372 | action = kmalloc(sizeof(struct irqaction), GFP_ATOMIC); | |
373 | if (!action) | |
374 | return -ENOMEM; | |
375 | ||
376 | action->handler = handler; | |
377 | action->flags = irqflags; | |
378 | cpus_clear(action->mask); | |
379 | action->name = devname; | |
380 | action->next = NULL; | |
381 | action->dev_id = dev_id; | |
382 | ||
eee45269 IK |
383 | select_smp_affinity(irq); |
384 | ||
1da177e4 LT |
385 | retval = setup_irq(irq, action); |
386 | if (retval) | |
387 | kfree(action); | |
388 | ||
389 | return retval; | |
390 | } | |
391 | ||
392 | EXPORT_SYMBOL(request_irq); | |
393 |