2 * Copyright (c) 1991 The Regents of the University of California.
5 * This code is derived from software contributed to Berkeley by
8 * Redistribution and use in source and binary forms, with or without
9 * modification, are permitted provided that the following conditions
11 * 1. Redistributions of source code must retain the above copyright
12 * notice, this list of conditions and the following disclaimer.
13 * 2. Redistributions in binary form must reproduce the above copyright
14 * notice, this list of conditions and the following disclaimer in the
15 * documentation and/or other materials provided with the distribution.
16 * 3. All advertising materials mentioning features or use of this software
17 * must display the following acknowledgement:
18 * This product includes software developed by the University of
19 * California, Berkeley and its contributors.
20 * 4. Neither the name of the University nor the names of its contributors
21 * may be used to endorse or promote products derived from this software
22 * without specific prior written permission.
24 * THIS SOFTWARE IS PROVIDED BY THE REGENTS AND CONTRIBUTORS ``AS IS'' AND
25 * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
26 * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
27 * ARE DISCLAIMED. IN NO EVENT SHALL THE REGENTS OR CONTRIBUTORS BE LIABLE
28 * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
29 * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
30 * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
31 * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
32 * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
33 * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
36 * from: @(#)isa.c 7.2 (Berkeley) 5/13/91
37 * $FreeBSD: src/sys/i386/isa/intr_machdep.c,v 1.29.2.5 2001/10/14 06:54:27 luigi Exp $
38 * $DragonFly: src/sys/platform/pc32/isa/intr_machdep.c,v 1.29 2005/06/03 17:14:51 dillon Exp $
41 * This file contains an aggregated module marked:
42 * Copyright (c) 1997, Stefan Esser <se@freebsd.org>
43 * All rights reserved.
44 * See the notice for details.
49 #include "opt_auto_eoi.h"
51 #include <sys/param.h>
53 #include <machine/lock.h>
55 #include <sys/systm.h>
56 #include <sys/syslog.h>
57 #include <sys/malloc.h>
58 #include <sys/errno.h>
59 #include <sys/interrupt.h>
60 #include <machine/ipl.h>
61 #include <machine/md_var.h>
62 #include <machine/segments.h>
64 #include <machine/globaldata.h>
66 #include <sys/thread2.h>
68 #include <machine/smptests.h> /** FAST_HI */
69 #include <machine/smp.h>
70 #include <bus/isa/i386/isa.h>
71 #include <i386/isa/icu.h>
74 #include <bus/isa/isavar.h>
76 #include <i386/isa/intr_machdep.h>
77 #include <bus/isa/isavar.h>
78 #include <sys/interrupt.h>
80 #include <machine/clock.h>
82 #include <machine/cpu.h>
85 #include <bus/mca/i386/mca_machdep.h>
88 /* XXX should be in suitable include files */
89 #define ICU_IMR_OFFSET 1 /* IO_ICU{1,2} + 1 */
94 * This is to accommodate "mixed-mode" programming for
95 * motherboards that don't connect the 8254 to the IO APIC.
100 #define NR_INTRNAMES (1 + ICU_LEN + 2 * ICU_LEN)
102 static inthand2_t isa_strayintr;
103 #if defined(FAST_HI) && defined(APIC_IO)
104 static inthand2_t isa_wrongintr;
106 static void init_i8259(void);
108 void *intr_unit[ICU_LEN*2];
109 u_long *intr_countp[ICU_LEN*2];
110 inthand2_t *intr_handler[ICU_LEN*2] = {
111 isa_strayintr, isa_strayintr, isa_strayintr, isa_strayintr,
112 isa_strayintr, isa_strayintr, isa_strayintr, isa_strayintr,
113 isa_strayintr, isa_strayintr, isa_strayintr, isa_strayintr,
114 isa_strayintr, isa_strayintr, isa_strayintr, isa_strayintr,
115 isa_strayintr, isa_strayintr, isa_strayintr, isa_strayintr,
116 isa_strayintr, isa_strayintr, isa_strayintr, isa_strayintr,
117 isa_strayintr, isa_strayintr, isa_strayintr, isa_strayintr,
118 isa_strayintr, isa_strayintr, isa_strayintr, isa_strayintr,
121 static struct md_intr_info {
124 int mihandler_installed;
126 } intr_info[ICU_LEN*2];
128 static inthand_t *fastintr[ICU_LEN] = {
129 &IDTVEC(fastintr0), &IDTVEC(fastintr1),
130 &IDTVEC(fastintr2), &IDTVEC(fastintr3),
131 &IDTVEC(fastintr4), &IDTVEC(fastintr5),
132 &IDTVEC(fastintr6), &IDTVEC(fastintr7),
133 &IDTVEC(fastintr8), &IDTVEC(fastintr9),
134 &IDTVEC(fastintr10), &IDTVEC(fastintr11),
135 &IDTVEC(fastintr12), &IDTVEC(fastintr13),
136 &IDTVEC(fastintr14), &IDTVEC(fastintr15),
138 &IDTVEC(fastintr16), &IDTVEC(fastintr17),
139 &IDTVEC(fastintr18), &IDTVEC(fastintr19),
140 &IDTVEC(fastintr20), &IDTVEC(fastintr21),
141 &IDTVEC(fastintr22), &IDTVEC(fastintr23),
145 unpendhand_t *fastunpend[ICU_LEN] = {
146 IDTVEC(fastunpend0), IDTVEC(fastunpend1),
147 IDTVEC(fastunpend2), IDTVEC(fastunpend3),
148 IDTVEC(fastunpend4), IDTVEC(fastunpend5),
149 IDTVEC(fastunpend6), IDTVEC(fastunpend7),
150 IDTVEC(fastunpend8), IDTVEC(fastunpend9),
151 IDTVEC(fastunpend10), IDTVEC(fastunpend11),
152 IDTVEC(fastunpend12), IDTVEC(fastunpend13),
153 IDTVEC(fastunpend14), IDTVEC(fastunpend15),
155 IDTVEC(fastunpend16), IDTVEC(fastunpend17),
156 IDTVEC(fastunpend18), IDTVEC(fastunpend19),
157 IDTVEC(fastunpend20), IDTVEC(fastunpend21),
158 IDTVEC(fastunpend22), IDTVEC(fastunpend23),
162 static inthand_t *slowintr[ICU_LEN] = {
163 &IDTVEC(intr0), &IDTVEC(intr1), &IDTVEC(intr2), &IDTVEC(intr3),
164 &IDTVEC(intr4), &IDTVEC(intr5), &IDTVEC(intr6), &IDTVEC(intr7),
165 &IDTVEC(intr8), &IDTVEC(intr9), &IDTVEC(intr10), &IDTVEC(intr11),
166 &IDTVEC(intr12), &IDTVEC(intr13), &IDTVEC(intr14), &IDTVEC(intr15),
168 &IDTVEC(intr16), &IDTVEC(intr17), &IDTVEC(intr18), &IDTVEC(intr19),
169 &IDTVEC(intr20), &IDTVEC(intr21), &IDTVEC(intr22), &IDTVEC(intr23),
173 #define NMI_PARITY (1 << 7)
174 #define NMI_IOCHAN (1 << 6)
175 #define ENMI_WATCHDOG (1 << 7)
176 #define ENMI_BUSTIMER (1 << 6)
177 #define ENMI_IOSTATUS (1 << 5)
180 * Handle a NMI, possibly a machine check.
181 * return true to panic system, false to ignore.
188 int isa_port = inb(0x61);
189 int eisa_port = inb(0x461);
191 log(LOG_CRIT, "NMI ISA %x, EISA %x\n", isa_port, eisa_port);
193 if (MCA_system && mca_bus_nmi())
197 if (isa_port & NMI_PARITY) {
198 log(LOG_CRIT, "RAM parity error, likely hardware failure.");
202 if (isa_port & NMI_IOCHAN) {
203 log(LOG_CRIT, "I/O channel check, likely hardware failure.");
208 * On a real EISA machine, this will never happen. However it can
209 * happen on ISA machines which implement XT style floating point
210 * error handling (very rare). Save them from a meaningless panic.
212 if (eisa_port == 0xff)
215 if (eisa_port & ENMI_WATCHDOG) {
216 log(LOG_CRIT, "EISA watchdog timer expired, likely hardware failure.");
220 if (eisa_port & ENMI_BUSTIMER) {
221 log(LOG_CRIT, "EISA bus timeout, likely hardware failure.");
225 if (eisa_port & ENMI_IOSTATUS) {
226 log(LOG_CRIT, "EISA I/O port status error.");
233 * ICU reinitialize when ICU configuration has lost.
241 for(i=0;i<ICU_LEN;i++)
242 if(intr_handler[i] != isa_strayintr)
248 * Fill in default interrupt table (in case of spurious interrupt
249 * during configuration of kernel, setup interrupt control unit
257 for (i = 0; i < ICU_LEN; i++)
258 icu_unset(i, isa_strayintr);
266 /* initialize 8259's */
269 outb(IO_ICU1, 0x19); /* reset; program device, four bytes */
272 outb(IO_ICU1, 0x11); /* reset; program device, four bytes */
274 outb(IO_ICU1+ICU_IMR_OFFSET, NRSVIDT); /* starting at this vector index */
275 outb(IO_ICU1+ICU_IMR_OFFSET, IRQ_SLAVE); /* slave on line 7 */
277 outb(IO_ICU1+ICU_IMR_OFFSET, 2 | 1); /* auto EOI, 8086 mode */
279 outb(IO_ICU1+ICU_IMR_OFFSET, 1); /* 8086 mode */
281 outb(IO_ICU1+ICU_IMR_OFFSET, 0xff); /* leave interrupts masked */
282 outb(IO_ICU1, 0x0a); /* default to IRR on read */
283 outb(IO_ICU1, 0xc0 | (3 - 1)); /* pri order 3-7, 0-2 (com2 first) */
287 outb(IO_ICU2, 0x19); /* reset; program device, four bytes */
290 outb(IO_ICU2, 0x11); /* reset; program device, four bytes */
292 outb(IO_ICU2+ICU_IMR_OFFSET, NRSVIDT+8); /* staring at this vector index */
293 outb(IO_ICU2+ICU_IMR_OFFSET, ICU_SLAVEID); /* my slave id is 7 */
295 outb(IO_ICU2+ICU_IMR_OFFSET, 2 | 1); /* auto EOI, 8086 mode */
297 outb(IO_ICU2+ICU_IMR_OFFSET,1); /* 8086 mode */
299 outb(IO_ICU2+ICU_IMR_OFFSET, 0xff); /* leave interrupts masked */
300 outb(IO_ICU2, 0x0a); /* default to IRR on read */
304 * Caught a stray interrupt, notify
307 isa_strayintr(void *vcookiep)
309 int intr = (void **)vcookiep - &intr_unit[0];
311 /* DON'T BOTHER FOR NOW! */
312 /* for some reason, we get bursts of intr #7, even if not enabled! */
314 * Well the reason you got bursts of intr #7 is because someone
315 * raised an interrupt line and dropped it before the 8259 could
316 * prioritize it. This is documented in the intel data book. This
317 * means you have BAD hardware! I have changed this so that only
318 * the first 5 get logged, then it quits logging them, and puts
319 * out a special message. rgrimes 3/25/1993
322 * XXX TODO print a different message for #7 if it is for a
323 * glitch. Glitches can be distinguished from real #7's by
324 * testing that the in-service bit is _not_ set. The test
325 * must be done before sending an EOI so it can't be done if
326 * we are using AUTO_EOI_1.
328 if (intrcnt[1 + intr] <= 5)
329 log(LOG_ERR, "stray irq %d\n", intr);
330 if (intrcnt[1 + intr] == 5)
332 "too many stray irq %d's; not logging any more\n", intr);
335 #if defined(FAST_HI) && defined(APIC_IO)
338 * This occurs if we mis-programmed the APIC and its vector is still
339 * pointing to the slow vector even when we thought we reprogrammed it
340 * to the high vector. This can occur when interrupts are improperly
341 * routed by the APIC. The unit data is opaque so we have to try to
342 * find it in the unit array.
345 isa_wrongintr(void *vcookiep)
349 for (intr = 0; intr < ICU_LEN*2; ++intr) {
350 if (intr_unit[intr] == vcookiep)
353 if (intr == ICU_LEN*2) {
354 log(LOG_ERR, "stray unknown irq (APIC misprogrammed)\n");
355 } else if (intrcnt[1 + intr] <= 5) {
356 log(LOG_ERR, "stray irq ~%d (APIC misprogrammed)\n", intr);
357 } else if (intrcnt[1 + intr] == 6) {
359 "too many stray irq ~%d's; not logging any more\n", intr);
367 * Return a bitmap of the current interrupt requests. This is 8259-specific
368 * and is only suitable for use at probe time.
371 isa_irq_pending(void)
378 return ((irr2 << 8) | irr1);
383 update_intr_masks(void)
388 for (intr=0; intr < ICU_LEN; intr ++) {
390 /* no 8259 SLAVE to ignore */
392 if (intr==ICU_SLAVEID) continue; /* ignore 8259 SLAVE output */
394 maskptr = intr_info[intr].maskp;
397 *maskptr |= SWI_CLOCK_MASK | (1 << intr);
399 if (mask != intr_info[intr].mask) {
401 printf ("intr_mask[%2d] old=%08x new=%08x ptr=%p.\n",
402 intr, intr_info[intr].mask, mask, maskptr);
404 intr_info[intr].mask = mask;
413 update_intrname(int intr, char *name)
417 int name_index, off, strayintr;
420 * Initialise strings for bitbucket and stray interrupt counters.
421 * These have statically allocated indices 0 and 1 through ICU_LEN.
423 if (intrnames[0] == '\0') {
424 off = sprintf(intrnames, "???") + 1;
425 for (strayintr = 0; strayintr < ICU_LEN; strayintr++)
426 off += sprintf(intrnames + off, "stray irq%d",
432 if (snprintf(buf, sizeof(buf), "%s irq%d", name, intr) >= sizeof(buf))
436 * Search for `buf' in `intrnames'. In the usual case when it is
437 * not found, append it to the end if there is enough space (the \0
438 * terminator for the previous string, if any, becomes a separator).
440 for (cp = intrnames, name_index = 0;
441 cp != eintrnames && name_index < NR_INTRNAMES;
442 cp += strlen(cp) + 1, name_index++) {
444 if (strlen(buf) >= eintrnames - cp)
449 if (strcmp(cp, buf) == 0)
454 printf("update_intrname: counting %s irq%d as %s\n", name, intr,
458 intr_countp[intr] = &intrcnt[name_index];
462 * NOTE! intr_handler[] is only used for FAST interrupts, the *vector.s
463 * code ignores it for normal interrupts.
466 icu_setup(int intr, inthand2_t *handler, void *arg, u_int *maskptr, int flags)
468 #if defined(FAST_HI) && defined(APIC_IO)
469 int select; /* the select register is 8 bits */
471 u_int32_t value; /* the window register is 32 bits */
474 u_int mask = (maskptr ? *maskptr : 0);
477 if ((u_int)intr >= ICU_LEN) /* no 8259 SLAVE to ignore */
479 if ((u_int)intr >= ICU_LEN || intr == ICU_SLAVEID)
482 if (intr_handler[intr] != isa_strayintr)
486 cpu_disable_intr(); /* YYY */
487 intr_handler[intr] = handler;
488 intr_unit[intr] = arg;
489 intr_info[intr].maskp = maskptr;
490 intr_info[intr].mask = mask | SWI_CLOCK_MASK | (1 << intr);
492 /* YYY fast ints supported and mp protected but ... */
495 #if defined(FAST_HI) && defined(APIC_IO)
496 if (flags & INTR_FAST) {
498 * Install a spurious interrupt in the low space in case
499 * the IO apic is not properly reprogrammed.
501 vector = TPR_SLOW_INTS + intr;
502 setidt(vector, isa_wrongintr,
503 SDT_SYS386IGT, SEL_KPL, GSEL(GCODE_SEL, SEL_KPL));
504 vector = TPR_FAST_INTS + intr;
505 setidt(vector, fastintr[intr],
506 SDT_SYS386IGT, SEL_KPL, GSEL(GCODE_SEL, SEL_KPL));
508 vector = TPR_SLOW_INTS + intr;
509 #ifdef APIC_INTR_REORDER
510 #ifdef APIC_INTR_HIGHPRI_CLOCK
511 /* XXX: Hack (kludge?) for more accurate clock. */
512 if (intr == apic_8254_intr || intr == 8) {
513 vector = TPR_FAST_INTS + intr;
517 setidt(vector, slowintr[intr],
518 SDT_SYS386IGT, SEL_KPL, GSEL(GCODE_SEL, SEL_KPL));
520 #ifdef APIC_INTR_REORDER
521 set_lapic_isrloc(intr, vector);
524 * Reprogram the vector in the IO APIC.
526 * XXX EOI/mask a pending (stray) interrupt on the old vector?
528 if (int_to_apicintpin[intr].ioapic >= 0) {
529 select = int_to_apicintpin[intr].redirindex;
530 value = io_apic_read(int_to_apicintpin[intr].ioapic,
531 select) & ~IOART_INTVEC;
532 io_apic_write(int_to_apicintpin[intr].ioapic,
533 select, value | vector);
536 setidt(ICU_OFFSET + intr,
537 flags & INTR_FAST ? fastintr[intr] : slowintr[intr],
538 SDT_SYS386IGT, SEL_KPL, GSEL(GCODE_SEL, SEL_KPL));
539 #endif /* FAST_HI && APIC_IO */
546 icu_unset(intr, handler)
552 if ((u_int)intr >= ICU_LEN || handler != intr_handler[intr]) {
553 printf("icu_unset: invalid handler %d %p/%p\n", intr, handler,
554 (((u_int)intr >= ICU_LEN) ? (void *)-1 : intr_handler[intr]));
560 cpu_disable_intr(); /* YYY */
561 intr_countp[intr] = &intrcnt[1 + intr];
562 intr_handler[intr] = isa_strayintr;
563 intr_info[intr].maskp = NULL;
564 intr_info[intr].mask = HWI_MASK | SWI_MASK;
565 intr_unit[intr] = &intr_unit[intr];
567 /* XXX how do I re-create dvp here? */
568 setidt(flags & INTR_FAST ? TPR_FAST_INTS + intr : TPR_SLOW_INTS + intr,
569 slowintr[intr], SDT_SYS386IGT, SEL_KPL, GSEL(GCODE_SEL, SEL_KPL));
571 #ifdef APIC_INTR_REORDER
572 set_lapic_isrloc(intr, ICU_OFFSET + intr);
574 setidt(ICU_OFFSET + intr, slowintr[intr], SDT_SYS386IGT, SEL_KPL,
575 GSEL(GCODE_SEL, SEL_KPL));
582 /* The following notice applies beyond this point in the file */
585 * Copyright (c) 1997, Stefan Esser <se@freebsd.org>
586 * All rights reserved.
588 * Redistribution and use in source and binary forms, with or without
589 * modification, are permitted provided that the following conditions
591 * 1. Redistributions of source code must retain the above copyright
592 * notice unmodified, this list of conditions, and the following
594 * 2. Redistributions in binary form must reproduce the above copyright
595 * notice, this list of conditions and the following disclaimer in the
596 * documentation and/or other materials provided with the distribution.
598 * THIS SOFTWARE IS PROVIDED BY THE AUTHOR ``AS IS'' AND ANY EXPRESS OR
599 * IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES
600 * OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED.
601 * IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR ANY DIRECT, INDIRECT,
602 * INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT
603 * NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
604 * DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
605 * THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
606 * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF
607 * THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
609 * $FreeBSD: src/sys/i386/isa/intr_machdep.c,v 1.29.2.5 2001/10/14 06:54:27 luigi Exp $
613 typedef struct intrec {
622 lwkt_serialize_t serializer;
623 volatile int in_handler;
626 static intrec *intreclist_head[ICU_LEN];
629 * The interrupt multiplexer calls each of the handlers in turn. A handler
630 * is called only if we can successfully obtain the interlock, meaning
631 * (1) we aren't recursed and (2) the handler has not been disabled via
632 * inthand_disabled().
634 * XXX the IPL is currently raised as necessary for the handler. However,
635 * IPLs are not MP safe so the IPL code will be removed when the device
636 * drivers, BIO, and VM no longer depend on it.
644 for (pp = arg; (p = *pp) != NULL; pp = &p->next) {
647 * New handler dispatch method. Only the serializer
648 * is used to interlock access. Note that this
649 * API includes a handler disablement feature.
651 lwkt_serialize_handler_call(p->serializer,
652 p->handler, p->argument);
655 * Old handlers may expect multiple interrupt
656 * sources to be masked. We must use a critical
660 p->handler(p->argument);
667 find_idesc(unsigned *maskptr, int irq)
669 intrec *p = intreclist_head[irq];
671 while (p && p->maskptr != maskptr)
678 * Both the low level handler and the shared interrupt multiplexer
679 * block out further interrupts as set in the handlers "mask", while
680 * the handler is running. In fact *maskptr should be used for this
681 * purpose, but since this requires one more pointer dereference on
682 * each interrupt, we rather bother update "mask" whenever *maskptr
683 * changes. The function "update_masks" should be called **after**
684 * all manipulation of the linked list of interrupt handlers hung
685 * off of intrdec_head[irq] is complete, since the chain of handlers
686 * will both determine the *maskptr values and the instances of mask
687 * that are fixed. This function should be called with the irq for
688 * which a new handler has been add blocked, since the masks may not
689 * yet know about the use of this irq for a device of a certain class.
693 update_mux_masks(void)
696 for (irq = 0; irq < ICU_LEN; irq++) {
697 intrec *idesc = intreclist_head[irq];
698 while (idesc != NULL) {
699 if (idesc->maskptr != NULL) {
700 /* our copy of *maskptr may be stale, refresh */
701 idesc->mask = *idesc->maskptr;
709 update_masks(intrmask_t *maskptr, int irq)
711 intrmask_t mask = 1 << irq;
716 if (find_idesc(maskptr, irq) == NULL) {
717 /* no reference to this maskptr was found in this irq's chain */
720 /* a reference to this maskptr was found in this irq's chain */
723 /* we need to update all values in the intr_mask[irq] array */
725 /* update mask in chains of the interrupt multiplex handler as well */
730 * Add an interrupt handler to the linked list hung off of intreclist_head[irq]
731 * and install a shared interrupt multiplex handler. Install an interrupt
732 * thread for each interrupt (though FAST interrupts will not use it).
733 * The preemption procedure checks the CPL. lwkt_preempt() will check
734 * relative thread priorities for us as long as we properly pass through
737 * The interrupt thread has already been put on the run queue, so if we cannot
738 * preempt we should force a reschedule.
741 cpu_intr_preempt(struct thread *td, int critpri)
743 struct md_intr_info *info = td->td_info.intdata;
745 if ((curthread->td_cpl & (1 << info->irq)) == 0)
746 lwkt_preempt(td, critpri);
748 need_lwkt_resched(); /* XXX may not be required */
752 add_intrdesc(intrec *idesc)
754 int irq = idesc->intr;
759 * There are two ways to enter intr_mux(). (1) via the scheduled
760 * interrupt thread or (2) directly. The thread mechanism is the
761 * normal mechanism used by SLOW interrupts, while the direct method
762 * is used by FAST interrupts.
764 * We need to create an interrupt thread if none exists.
766 if (intr_info[irq].mihandler_installed == 0) {
769 intr_info[irq].mihandler_installed = 1;
770 intr_info[irq].irq = irq;
771 td = register_int(irq, intr_mux, &intreclist_head[irq], idesc->name, idesc->maskptr);
772 td->td_info.intdata = &intr_info[irq];
773 td->td_preemptable = cpu_intr_preempt;
774 printf("installed MI handler for int %d\n", irq);
777 headp = &intreclist_head[irq];
784 if ((idesc->flags & INTR_EXCL) || (head->flags & INTR_EXCL)) {
785 printf("\tdevice combination doesn't support "
786 "shared irq%d\n", irq);
789 if ((idesc->flags & INTR_FAST) || (head->flags & INTR_FAST)) {
790 printf("\tdevice combination doesn't support "
791 "multiple FAST interrupts on IRQ%d\n", irq);
796 * Always install intr_mux as the hard handler so it can deal with
797 * individual enablement on handlers.
800 if (icu_setup(irq, idesc->handler, idesc->argument, idesc->maskptr, idesc->flags) != 0)
802 update_intrname(irq, idesc->name);
803 } else if (head->next == NULL) {
804 icu_unset(irq, head->handler);
805 if (icu_setup(irq, intr_mux, &intreclist_head[irq], 0, 0) != 0)
807 if (bootverbose && head->next == NULL)
808 printf("\tusing shared irq%d.\n", irq);
809 update_intrname(irq, "mux");
813 * Append to the end of the chain and update our SPL masks.
815 while (*headp != NULL)
816 headp = &(*headp)->next;
819 update_masks(idesc->maskptr, irq);
824 * Create and activate an interrupt handler descriptor data structure.
826 * The dev_instance pointer is required for resource management, and will
827 * only be passed through to resource_claim().
829 * There will be functions that derive a driver and unit name from a
830 * dev_instance variable, and those functions will be used to maintain the
831 * interrupt counter label array referenced by systat and vmstat to report
832 * device interrupt rates (->update_intrlabels).
834 * Add the interrupt handler descriptor data structure created by an
835 * earlier call of create_intr() to the linked list for its irq and
836 * adjust the interrupt masks if necessary.
838 * WARNING: This is an internal function and not to be used by device
839 * drivers. It is subject to change without notice.
843 inthand_add(const char *name, int irq, inthand2_t handler, void *arg,
844 intrmask_t *maskptr, int flags, lwkt_serialize_t serializer)
849 if (ICU_LEN > 8 * sizeof *maskptr) {
850 printf("create_intr: ICU_LEN of %d too high for %d bit intrmask\n",
851 ICU_LEN, 8 * sizeof *maskptr);
854 if ((unsigned)irq >= ICU_LEN) {
855 printf("create_intr: requested irq%d too high, limit is %d\n",
860 idesc = malloc(sizeof *idesc, M_DEVBUF, M_WAITOK | M_ZERO);
866 idesc->name = malloc(strlen(name) + 1, M_DEVBUF, M_WAITOK);
867 if (idesc->name == NULL) {
868 free(idesc, M_DEVBUF);
871 strcpy(idesc->name, name);
873 idesc->handler = handler;
874 idesc->argument = arg;
875 idesc->maskptr = maskptr;
877 idesc->flags = flags;
878 idesc->serializer = serializer;
880 /* add irq to class selected by maskptr */
882 errcode = add_intrdesc(idesc);
887 printf("\tintr_connect(irq%d) failed, result=%d\n",
889 free(idesc->name, M_DEVBUF);
890 free(idesc, M_DEVBUF);
898 * Deactivate and remove the interrupt handler descriptor data connected
899 * created by an earlier call of intr_connect() from the linked list and
900 * adjust theinterrupt masks if necessary.
902 * Return the memory held by the interrupt handler descriptor data structure
903 * to the system. Make sure, the handler is not actively used anymore, before.
906 inthand_remove(intrec *idesc)
908 intrec **hook, *head;
918 * Find and remove the interrupt descriptor.
920 hook = &intreclist_head[irq];
921 while (*hook != idesc) {
926 hook = &(*hook)->next;
931 * If the list is now empty, revert the hard vector to the spurious
934 head = intreclist_head[irq];
937 * No more interrupts on this irq
939 icu_unset(irq, idesc->handler);
940 update_intrname(irq, NULL);
941 } else if (head->next) {
943 * This irq is still shared (has at least two handlers)
944 * (the name should already be set to "mux").
948 * This irq is no longer shared
950 icu_unset(irq, intr_mux);
951 icu_setup(irq, head->handler, head->argument, head->maskptr, head->flags);
952 update_intrname(irq, head->name);
954 update_masks(idesc->maskptr, irq);
956 free(idesc, M_DEVBUF);
964 * This function is called by an interrupt thread when it has completed
965 * processing a loop. We re-enable interrupts and interlock with
968 * See kern/kern_intr.c for more information.
971 ithread_done(int irq)
973 struct mdglobaldata *gd = mdcpu;
977 td = gd->mi.gd_curthread;
979 KKASSERT(td->td_pri >= TDPRI_CRIT);
980 lwkt_deschedule_self(td);
982 if (gd->gd_ipending & mask) {
983 atomic_clear_int_nonlocked(&gd->gd_ipending, mask);
985 lwkt_schedule_self(td);
993 * forward_fast_remote()
995 * This function is called from the receiving end of an IPIQ when a
996 * remote cpu wishes to forward a fast interrupt to us. All we have to
997 * do is set the interrupt pending and let the IPI's doreti deal with it.
1000 forward_fastint_remote(void *arg)
1003 struct mdglobaldata *gd = mdcpu;
1005 atomic_set_int_nonlocked(&gd->gd_fpending, 1 << irq);
1006 atomic_set_int_nonlocked(&gd->mi.gd_reqflags, RQF_INTPEND);