1
0
mirror of https://git.FreeBSD.org/src.git synced 2024-12-24 11:29:10 +00:00
freebsd/sys/alpha/isa/isa.c
John Baldwin e0f66ef861 Reorganize the interrupt handling code a bit to make a few things cleaner
and increase flexibility to allow various different approaches to be tried
in the future.
- Split struct ithd up into two pieces.  struct intr_event holds the list
  of interrupt handlers associated with interrupt sources.
  struct intr_thread contains the data relative to an interrupt thread.
  Currently we still provide a 1:1 relationship of events to threads
  with the exception that events only have an associated thread if there
  is at least one threaded interrupt handler attached to the event.  This
  means that on x86 we no longer have 4 bazillion interrupt threads with
  no handlers.  It also means that interrupt events with only INTR_FAST
  handlers no longer have an associated thread either.
- Renamed struct intrhand to struct intr_handler to follow the struct
  intr_foo naming convention.  This did require renaming the powerpc
  MD struct intr_handler to struct ppc_intr_handler.
- INTR_FAST no longer implies INTR_EXCL on all architectures except for
  powerpc.  This means that multiple INTR_FAST handlers can attach to the
  same interrupt and that INTR_FAST and non-INTR_FAST handlers can attach
  to the same interrupt.  Sharing INTR_FAST handlers may not always be
  desirable, but having sio(4) and uhci(4) fight over an IRQ isn't fun
  either.  Drivers can always still use INTR_EXCL to ask for an interrupt
  exclusively.  The way this sharing works is that when an interrupt
  comes in, all the INTR_FAST handlers are executed first, and if any
  threaded handlers exist, the interrupt thread is scheduled afterwards.
  This type of layout also makes it possible to investigate using interrupt
  filters ala OS X where the filter determines whether or not its companion
  threaded handler should run.
- Aside from the INTR_FAST changes above, the impact on MD interrupt code
  is mostly just 's/ithread/intr_event/'.
- A new MI ddb command 'show intrs' walks the list of interrupt events
  dumping their state.  It also has a '/v' verbose switch which dumps
  info about all of the handlers attached to each event.
- We currently don't destroy an interrupt thread when the last threaded
  handler is removed because it would suck for things like ppbus(8)'s
  braindead behavior.  The code is present, though, it is just under
  #if 0 for now.
- Move the code to actually execute the threaded handlers for an interrrupt
  event into a separate function so that ithread_loop() becomes more
  readable.  Previously this code was all in the middle of ithread_loop()
  and indented halfway across the screen.
- Made struct intr_thread private to kern_intr.c and replaced td_ithd
  with a thread private flag TDP_ITHREAD.
- In statclock, check curthread against idlethread directly rather than
  curthread's proc against idlethread's proc. (Not really related to intr
  changes)

Tested on:	alpha, amd64, i386, sparc64
Tested on:	arm, ia64 (older version of patch by cognet and marcel)
2005-10-25 19:48:48 +00:00

419 lines
9.5 KiB
C

/*-
* Copyright (c) 1998 Doug Rabson
* All rights reserved.
*
* Redistribution and use in source and binary forms, with or without
* modification, are permitted provided that the following conditions
* are met:
* 1. Redistributions of source code must retain the above copyright
* notice, this list of conditions and the following disclaimer.
* 2. Redistributions in binary form must reproduce the above copyright
* notice, this list of conditions and the following disclaimer in the
* documentation and/or other materials provided with the distribution.
*
* THIS SOFTWARE IS PROVIDED BY THE AUTHOR AND CONTRIBUTORS ``AS IS'' AND
* ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
* IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
* ARE DISCLAIMED. IN NO EVENT SHALL THE AUTHOR OR CONTRIBUTORS BE LIABLE
* FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
* DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
* OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
* HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
* LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
* OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
* SUCH DAMAGE.
*/
#include <sys/cdefs.h>
__FBSDID("$FreeBSD$");
#include <sys/param.h>
#include <sys/systm.h>
#include <sys/kernel.h>
#include <sys/lock.h>
#include <sys/module.h>
#include <sys/mutex.h>
#include <sys/bus.h>
#include <machine/bus.h>
#include <sys/malloc.h>
#include <sys/proc.h>
#include <sys/rman.h>
#include <sys/interrupt.h>
#include <isa/isareg.h>
#include <isa/isavar.h>
#include <isa/isa_common.h>
#include <alpha/isa/isavar.h>
#include <machine/intr.h>
#include <machine/intrcnt.h>
#include <machine/resource.h>
#include <machine/cpuconf.h>
static struct rman isa_irq_rman;
static struct rman isa_drq_rman;
static void
isa_intr_enable(int irq)
{
if (irq < 8)
outb(IO_ICU1+1, inb(IO_ICU1+1) & ~(1 << irq));
else
outb(IO_ICU2+1, inb(IO_ICU2+1) & ~(1 << (irq - 8)));
}
static void
isa_intr_disable(int irq)
{
if (irq < 8)
outb(IO_ICU1+1, inb(IO_ICU1+1) | (1 << irq));
else
outb(IO_ICU2+1, inb(IO_ICU2+1) | (1 << (irq - 8)));
}
intrmask_t
isa_irq_pending(void)
{
u_char irr1;
u_char irr2;
irr1 = inb(IO_ICU1);
irr2 = inb(IO_ICU2);
return ((irr2 << 8) | irr1);
}
intrmask_t
isa_irq_mask(void)
{
u_char irr1;
u_char irr2;
irr1 = inb(IO_ICU1+1);
irr2 = inb(IO_ICU2+1);
return ((irr2 << 8) | irr1);
}
void
isa_init(device_t dev)
{
isa_init_intr();
}
void
isa_init_intr(void)
{
static int initted = 0;
if (initted) return;
initted = 1;
isa_irq_rman.rm_start = 0;
isa_irq_rman.rm_end = 15;
isa_irq_rman.rm_type = RMAN_ARRAY;
isa_irq_rman.rm_descr = "ISA Interrupt request lines";
if (rman_init(&isa_irq_rman)
|| rman_manage_region(&isa_irq_rman, 0, 1)
|| rman_manage_region(&isa_irq_rman, 3, 15))
panic("isa_probe isa_irq_rman");
isa_drq_rman.rm_start = 0;
isa_drq_rman.rm_end = 7;
isa_drq_rman.rm_type = RMAN_ARRAY;
isa_drq_rman.rm_descr = "ISA DMA request lines";
if (rman_init(&isa_drq_rman)
|| rman_manage_region(&isa_drq_rman, 0, 7))
panic("isa_probe isa_drq_rman");
/* mask all isa interrupts */
outb(IO_ICU1+1, 0xff);
outb(IO_ICU2+1, 0xff);
/* make sure chaining irq is enabled */
isa_intr_enable(2);
}
struct resource *
isa_alloc_intr(device_t bus, device_t child, int irq)
{
return rman_reserve_resource(&isa_irq_rman, irq, irq, 1,
0, child);
}
struct resource *
isa_alloc_intrs(device_t bus, device_t child, u_long start, u_long end)
{
return rman_reserve_resource(&isa_irq_rman, start, end,
end - start + 1, 0, child);
}
int
isa_release_intr(device_t bus, device_t child, struct resource *r)
{
return rman_release_resource(r);
}
/*
* This implementation simply passes the request up to the parent
* bus, which in our case is the pci chipset device, substituting any
* configured values if the caller defaulted. We can get away with
* this because there is no special mapping for ISA resources on this
* platform. When porting this code to another architecture, it may be
* necessary to interpose a mapping layer here.
*
* We manage our own interrupt resources since ISA interrupts go through
* the ISA PIC, not the PCI interrupt controller.
*/
struct resource *
isa_alloc_resource(device_t bus, device_t child, int type, int *rid,
u_long start, u_long end, u_long count, u_int flags)
{
/*
* Consider adding a resource definition.
*/
int passthrough = (device_get_parent(child) != bus);
int isdefault = (start == 0UL && end == ~0UL);
struct isa_device* idev = DEVTOISA(child);
struct resource_list *rl = &idev->id_resources;
struct resource_list_entry *rle;
struct resource *res;
if (!passthrough && !isdefault) {
rle = resource_list_find(rl, type, *rid);
if (!rle) {
if (*rid < 0)
return 0;
switch (type) {
case SYS_RES_IRQ:
if (*rid >= ISA_NIRQ)
return 0;
break;
case SYS_RES_DRQ:
if (*rid >= ISA_NDRQ)
return 0;
break;
case SYS_RES_MEMORY:
if (*rid >= ISA_NMEM)
return 0;
break;
case SYS_RES_IOPORT:
if (*rid >= ISA_NPORT)
return 0;
break;
default:
return 0;
}
resource_list_add(rl, type, *rid, start, end, count);
}
}
if (type != SYS_RES_IRQ && type != SYS_RES_DRQ)
return resource_list_alloc(rl, bus, child, type, rid,
start, end, count, flags);
if (!passthrough) {
rl = device_get_ivars(child);
rle = resource_list_find(rl, type, *rid);
if (!rle)
return 0;
if (rle->res)
panic("isa_alloc_resource: resource entry is busy");
if (isdefault) {
start = end = rle->start;
count = 1;
}
}
if (type == SYS_RES_IRQ)
res = rman_reserve_resource(&isa_irq_rman, start, start, 1,
0, child);
else
res = rman_reserve_resource(&isa_drq_rman, start, start, 1,
0, child);
if (res && !passthrough) {
rle = resource_list_find(rl, type, *rid);
rle->start = rman_get_start(res);
rle->end = rman_get_end(res);
rle->count = 1;
rle->res = res;
}
return res;
}
int
isa_release_resource(device_t bus, device_t child, int type, int rid,
struct resource *res)
{
int passthrough = (device_get_parent(child) != bus);
struct isa_device* idev = DEVTOISA(child);
struct resource_list *rl = &idev->id_resources;
struct resource_list_entry *rle;
int error;
if (type != SYS_RES_IRQ)
return resource_list_release(rl, bus, child, type, rid, res);
error = rman_release_resource(res);
if (!passthrough && !error) {
rle = resource_list_find(rl, SYS_RES_IRQ, rid);
if (rle)
rle->res = NULL;
else
error = ENOENT;
}
return error;
}
struct isa_intr {
void *ih;
driver_intr_t *intr;
void *arg;
int irq;
};
/*
* Wrap ISA interrupt routines so that we can feed non-specific
* EOI to the PICs.
*/
static void
isa_handle_fast_intr(void *arg)
{
struct isa_intr *ii = arg;
int irq = ii->irq;
ii->intr(ii->arg);
mtx_lock_spin(&icu_lock);
if (irq > 7)
outb(IO_ICU2, 0x20 | (irq & 7));
outb(IO_ICU1, 0x20 | (irq > 7 ? 2 : irq));
mtx_unlock_spin(&icu_lock);
}
static void
isa_handle_intr(void *arg)
{
struct isa_intr *ii = arg;
ii->intr(ii->arg);
}
/*
* Send a non-specific EIO early, then disable the source
*/
static void
isa_disable_intr(uintptr_t vector)
{
int irq;
irq = (vector - 0x800) >> 4;
mtx_lock_spin(&icu_lock);
if (irq > 7)
outb(IO_ICU2, 0x20 | (irq & 7));
outb(IO_ICU1, 0x20 | (irq > 7 ? 2 : irq));
isa_intr_disable(irq);
mtx_unlock_spin(&icu_lock);
}
static void
isa_enable_intr(uintptr_t vector)
{
int irq;
irq = (vector - 0x800) >> 4;
mtx_lock_spin(&icu_lock);
isa_intr_enable(irq);
mtx_unlock_spin(&icu_lock);
}
int
isa_setup_intr(device_t dev, device_t child,
struct resource *irq, int flags,
driver_intr_t *intr, void *arg, void **cookiep)
{
struct isa_intr *ii;
int error;
if (platform.isa_setup_intr)
return platform.isa_setup_intr(dev, child, irq, flags,
intr, arg, cookiep);
if (irq == NULL)
return ENODEV;
error = rman_activate_resource(irq);
if (error)
return error;
ii = malloc(sizeof(struct isa_intr), M_DEVBUF, M_NOWAIT);
if (!ii)
return ENOMEM;
ii->intr = intr;
ii->arg = arg;
ii->irq = rman_get_start(irq);
error = alpha_setup_intr(
device_get_nameunit(child ? child : dev),
0x800 + (ii->irq << 4),
((flags & INTR_FAST) ? isa_handle_fast_intr :
isa_handle_intr), ii, flags, &ii->ih,
&intrcnt[INTRCNT_ISA_IRQ + ii->irq],
isa_disable_intr, isa_enable_intr);
if (error) {
free(ii, M_DEVBUF);
return error;
}
mtx_lock_spin(&icu_lock);
isa_intr_enable(ii->irq);
mtx_unlock_spin(&icu_lock);
*cookiep = ii;
if (child)
device_printf(child, "interrupting at ISA irq %d\n",
(int)ii->irq);
return 0;
}
int
isa_teardown_intr(device_t dev, device_t child,
struct resource *irq, void *cookie)
{
struct isa_intr *ii = cookie;
struct intr_handler *ih, *handler = (struct intr_handler *)ii->ih;
struct intr_event *ie = handler->ih_event;
int num_handlers = 0;
mtx_lock(&ie->ie_lock);
TAILQ_FOREACH(ih, &ie->ie_handlers, ih_next)
num_handlers++;
mtx_unlock(&ie->ie_lock);
/*
* Only disable the interrupt in hardware if there are no
* other handlers sharing it.
*/
if (num_handlers == 1) {
mtx_lock_spin(&icu_lock);
isa_intr_disable(ii->irq);
mtx_unlock_spin(&icu_lock);
if (platform.isa_teardown_intr) {
platform.isa_teardown_intr(dev, child, irq, cookie);
return 0;
}
}
alpha_teardown_intr(ii->ih);
return 0;
}