2 * Copyright (c) 2000 Michael Smith
3 * Copyright (c) 2001 Scott Long
4 * Copyright (c) 2000 BSDi
5 * Copyright (c) 2001 Adaptec, Inc.
8 * Redistribution and use in source and binary forms, with or without
9 * modification, are permitted provided that the following conditions
11 * 1. Redistributions of source code must retain the above copyright
12 * notice, this list of conditions and the following disclaimer.
13 * 2. Redistributions in binary form must reproduce the above copyright
14 * notice, this list of conditions and the following disclaimer in the
15 * documentation and/or other materials provided with the distribution.
17 * THIS SOFTWARE IS PROVIDED BY THE AUTHOR AND CONTRIBUTORS ``AS IS'' AND
18 * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
19 * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
20 * ARE DISCLAIMED. IN NO EVENT SHALL THE AUTHOR OR CONTRIBUTORS BE LIABLE
21 * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
22 * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
23 * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
24 * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
25 * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
26 * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
29 * $FreeBSD: src/sys/dev/aac/aac.c,v 1.9.2.14 2003/04/08 13:22:08 scottl Exp $
30 * $DragonFly: src/sys/dev/raid/aac/aac.c,v 1.34 2008/01/20 03:40:35 pavalos Exp $
34 * Driver for the Adaptec 'FSA' family of PCI/SCSI RAID adapters.
36 #define AAC_DRIVER_VERSION 0x02000000
37 #define AAC_DRIVERNAME "aac"
41 /* #include <stddef.h> */
42 #include <sys/param.h>
43 #include <sys/systm.h>
44 #include <sys/malloc.h>
45 #include <sys/kernel.h>
46 #include <sys/kthread.h>
47 #include <sys/sysctl.h>
48 #include <sys/event.h>
52 #include <sys/devicestat.h>
54 #include <sys/signalvar.h>
56 #include <sys/eventhandler.h>
59 #include <sys/mplock2.h>
61 #include <bus/pci/pcireg.h>
62 #include <bus/pci/pcivar.h>
65 #include "aac_ioctl.h"
67 #include "aac_tables.h"
69 static void aac_startup(void *arg);
70 static void aac_add_container(struct aac_softc *sc,
71 struct aac_mntinforesp *mir, int f);
72 static void aac_get_bus_info(struct aac_softc *sc);
73 static int aac_shutdown(device_t dev);
75 /* Command Processing */
76 static void aac_timeout(void *ssc);
77 static int aac_map_command(struct aac_command *cm);
78 static void aac_complete(void *context, int pending);
79 static int aac_bio_command(struct aac_softc *sc, struct aac_command **cmp);
80 static void aac_bio_complete(struct aac_command *cm);
81 static int aac_wait_command(struct aac_command *cm);
82 static void aac_command_thread(struct aac_softc *sc);
84 /* Command Buffer Management */
85 static void aac_map_command_sg(void *arg, bus_dma_segment_t *segs,
87 static void aac_map_command_helper(void *arg, bus_dma_segment_t *segs,
89 static int aac_alloc_commands(struct aac_softc *sc);
90 static void aac_free_commands(struct aac_softc *sc);
91 static void aac_unmap_command(struct aac_command *cm);
93 /* Hardware Interface */
94 static void aac_common_map(void *arg, bus_dma_segment_t *segs, int nseg,
96 static int aac_check_firmware(struct aac_softc *sc);
97 static int aac_init(struct aac_softc *sc);
98 static int aac_sync_command(struct aac_softc *sc, u_int32_t command,
99 u_int32_t arg0, u_int32_t arg1, u_int32_t arg2,
100 u_int32_t arg3, u_int32_t *sp);
101 static int aac_enqueue_fib(struct aac_softc *sc, int queue,
102 struct aac_command *cm);
103 static int aac_dequeue_fib(struct aac_softc *sc, int queue,
104 u_int32_t *fib_size, struct aac_fib **fib_addr);
105 static int aac_enqueue_response(struct aac_softc *sc, int queue,
106 struct aac_fib *fib);
108 /* Falcon/PPC interface */
109 static int aac_fa_get_fwstatus(struct aac_softc *sc);
110 static void aac_fa_qnotify(struct aac_softc *sc, int qbit);
111 static int aac_fa_get_istatus(struct aac_softc *sc);
112 static void aac_fa_clear_istatus(struct aac_softc *sc, int mask);
113 static void aac_fa_set_mailbox(struct aac_softc *sc, u_int32_t command,
114 u_int32_t arg0, u_int32_t arg1,
115 u_int32_t arg2, u_int32_t arg3);
116 static int aac_fa_get_mailbox(struct aac_softc *sc, int mb);
117 static void aac_fa_set_interrupts(struct aac_softc *sc, int enable);
119 struct aac_interface aac_fa_interface = {
123 aac_fa_clear_istatus,
126 aac_fa_set_interrupts,
130 /* StrongARM interface */
131 static int aac_sa_get_fwstatus(struct aac_softc *sc);
132 static void aac_sa_qnotify(struct aac_softc *sc, int qbit);
133 static int aac_sa_get_istatus(struct aac_softc *sc);
134 static void aac_sa_clear_istatus(struct aac_softc *sc, int mask);
135 static void aac_sa_set_mailbox(struct aac_softc *sc, u_int32_t command,
136 u_int32_t arg0, u_int32_t arg1,
137 u_int32_t arg2, u_int32_t arg3);
138 static int aac_sa_get_mailbox(struct aac_softc *sc, int mb);
139 static void aac_sa_set_interrupts(struct aac_softc *sc, int enable);
141 struct aac_interface aac_sa_interface = {
145 aac_sa_clear_istatus,
148 aac_sa_set_interrupts,
152 /* i960Rx interface */
153 static int aac_rx_get_fwstatus(struct aac_softc *sc);
154 static void aac_rx_qnotify(struct aac_softc *sc, int qbit);
155 static int aac_rx_get_istatus(struct aac_softc *sc);
156 static void aac_rx_clear_istatus(struct aac_softc *sc, int mask);
157 static void aac_rx_set_mailbox(struct aac_softc *sc, u_int32_t command,
158 u_int32_t arg0, u_int32_t arg1,
159 u_int32_t arg2, u_int32_t arg3);
160 static int aac_rx_get_mailbox(struct aac_softc *sc, int mb);
161 static void aac_rx_set_interrupts(struct aac_softc *sc, int enable);
162 static int aac_rx_send_command(struct aac_softc *sc, struct aac_command *cm);
163 static int aac_rx_get_outb_queue(struct aac_softc *sc);
164 static void aac_rx_set_outb_queue(struct aac_softc *sc, int index);
166 struct aac_interface aac_rx_interface = {
170 aac_rx_clear_istatus,
173 aac_rx_set_interrupts,
175 aac_rx_get_outb_queue,
176 aac_rx_set_outb_queue
179 /* Rocket/MIPS interface */
180 static int aac_rkt_get_fwstatus(struct aac_softc *sc);
181 static void aac_rkt_qnotify(struct aac_softc *sc, int qbit);
182 static int aac_rkt_get_istatus(struct aac_softc *sc);
183 static void aac_rkt_clear_istatus(struct aac_softc *sc, int mask);
184 static void aac_rkt_set_mailbox(struct aac_softc *sc, u_int32_t command,
185 u_int32_t arg0, u_int32_t arg1,
186 u_int32_t arg2, u_int32_t arg3);
187 static int aac_rkt_get_mailbox(struct aac_softc *sc, int mb);
188 static void aac_rkt_set_interrupts(struct aac_softc *sc, int enable);
189 static int aac_rkt_send_command(struct aac_softc *sc, struct aac_command *cm);
190 static int aac_rkt_get_outb_queue(struct aac_softc *sc);
191 static void aac_rkt_set_outb_queue(struct aac_softc *sc, int index);
193 struct aac_interface aac_rkt_interface = {
194 aac_rkt_get_fwstatus,
197 aac_rkt_clear_istatus,
200 aac_rkt_set_interrupts,
201 aac_rkt_send_command,
202 aac_rkt_get_outb_queue,
203 aac_rkt_set_outb_queue
206 /* Debugging and Diagnostics */
207 static void aac_describe_controller(struct aac_softc *sc);
208 static char *aac_describe_code(struct aac_code_lookup *table,
211 /* Management Interface */
212 static d_open_t aac_open;
213 static d_close_t aac_close;
214 static d_ioctl_t aac_ioctl;
215 static d_kqfilter_t aac_kqfilter;
216 static void aac_filter_detach(struct knote *kn);
217 static int aac_filter(struct knote *kn, long hint);
218 static int aac_ioctl_sendfib(struct aac_softc *sc, caddr_t ufib) __unused;
219 static void aac_handle_aif(struct aac_softc *sc,
220 struct aac_fib *fib);
221 static int aac_rev_check(struct aac_softc *sc, caddr_t udata);
222 static int aac_getnext_aif(struct aac_softc *sc, caddr_t arg);
223 static int aac_return_aif(struct aac_softc *sc, caddr_t uptr);
224 static int aac_query_disk(struct aac_softc *sc, caddr_t uptr);
225 static int aac_get_pci_info(struct aac_softc *sc, caddr_t uptr);
226 static void aac_ioctl_event(struct aac_softc *sc,
227 struct aac_event *event, void *arg);
229 #define AAC_CDEV_MAJOR 150
231 static struct dev_ops aac_ops = {
232 { "aac", AAC_CDEV_MAJOR, D_KQFILTER },
234 .d_close = aac_close,
235 .d_ioctl = aac_ioctl,
236 .d_kqfilter = aac_kqfilter
239 DECLARE_DUMMY_MODULE(aac);
241 MALLOC_DEFINE(M_AACBUF, "aacbuf", "Buffers for the AAC driver");
244 SYSCTL_NODE(_hw, OID_AUTO, aac, CTLFLAG_RD, 0, "AAC driver parameters");
251 * Initialise the controller and softc
254 aac_attach(struct aac_softc *sc)
259 callout_init(&sc->aac_watchdog);
262 * Initialise per-controller queues.
267 aac_initq_complete(sc);
271 * Initialise command-completion task.
273 TASK_INIT(&sc->aac_task_complete, 0, aac_complete, sc);
275 /* mark controller as suspended until we get ourselves organised */
276 sc->aac_state |= AAC_STATE_SUSPEND;
279 * Check that the firmware on the card is supported.
281 if ((error = aac_check_firmware(sc)) != 0)
287 AAC_LOCK_INIT(&sc->aac_aifq_lock, "AAC AIF lock");
288 AAC_LOCK_INIT(&sc->aac_io_lock, "AAC I/O lock");
289 AAC_LOCK_INIT(&sc->aac_container_lock, "AAC container lock");
290 TAILQ_INIT(&sc->aac_container_tqh);
291 TAILQ_INIT(&sc->aac_ev_cmfree);
294 /* Initialize the local AIF queue pointers */
295 sc->aac_aifq_head = sc->aac_aifq_tail = AAC_AIFQ_LENGTH;
298 * Initialise the adapter.
300 if ((error = aac_init(sc)) != 0)
304 * Allocate and connect our interrupt.
307 if ((sc->aac_irq = bus_alloc_resource_any(sc->aac_dev, SYS_RES_IRQ,
310 RF_ACTIVE)) == NULL) {
311 device_printf(sc->aac_dev, "can't allocate interrupt\n");
314 if (sc->flags & AAC_FLAGS_NEW_COMM) {
315 if (bus_setup_intr(sc->aac_dev, sc->aac_irq,
317 sc, &sc->aac_intr, NULL)) {
318 device_printf(sc->aac_dev, "can't set up interrupt\n");
322 if (bus_setup_intr(sc->aac_dev, sc->aac_irq, 0,
323 aac_fast_intr, sc, &sc->aac_intr, NULL)) {
324 device_printf(sc->aac_dev,
325 "can't set up FAST interrupt\n");
326 if (bus_setup_intr(sc->aac_dev, sc->aac_irq,
328 sc, &sc->aac_intr, NULL)) {
329 device_printf(sc->aac_dev,
330 "can't set up MPSAFE interrupt\n");
337 * Print a little information about the controller.
339 aac_describe_controller(sc);
342 * Register to probe our containers later.
344 sc->aac_ich.ich_func = aac_startup;
345 sc->aac_ich.ich_arg = sc;
346 sc->aac_ich.ich_desc = "aac";
347 if (config_intrhook_establish(&sc->aac_ich) != 0) {
348 device_printf(sc->aac_dev,
349 "can't establish configuration hook\n");
354 * Make the control device.
356 unit = device_get_unit(sc->aac_dev);
357 sc->aac_dev_t = make_dev(&aac_ops, unit, UID_ROOT, GID_OPERATOR,
358 0640, "aac%d", unit);
359 sc->aac_dev_t->si_drv1 = sc;
360 reference_dev(sc->aac_dev_t);
362 /* Create the AIF thread */
363 if (kthread_create((void(*)(void *))aac_command_thread, sc,
364 &sc->aifthread, "aac%daif", unit))
365 panic("Could not create AIF thread\n");
367 /* Register the shutdown method to only be called post-dump */
368 if ((sc->eh = EVENTHANDLER_REGISTER(shutdown_post_sync, aac_shutdown,
369 sc->aac_dev, SHUTDOWN_PRI_DRIVER)) == NULL)
370 device_printf(sc->aac_dev,
371 "shutdown event registration failed\n");
373 /* Register with CAM for the non-DASD devices */
374 if ((sc->flags & AAC_FLAGS_ENABLE_CAM) != 0) {
375 TAILQ_INIT(&sc->aac_sim_tqh);
376 aac_get_bus_info(sc);
383 aac_add_event(struct aac_softc *sc, struct aac_event *event)
386 switch (event->ev_type & AAC_EVENT_MASK) {
387 case AAC_EVENT_CMFREE:
388 TAILQ_INSERT_TAIL(&sc->aac_ev_cmfree, event, ev_links);
391 device_printf(sc->aac_dev, "aac_add event: unknown event %d\n",
400 * Probe for containers, create disks.
403 aac_startup(void *arg)
405 struct aac_softc *sc;
407 struct aac_mntinfo *mi;
408 struct aac_mntinforesp *mir = NULL;
409 int count = 0, i = 0;
413 sc = (struct aac_softc *)arg;
415 /* disconnect ourselves from the intrhook chain */
416 config_intrhook_disestablish(&sc->aac_ich);
418 AAC_LOCK_ACQUIRE(&sc->aac_io_lock);
419 aac_alloc_sync_fib(sc, &fib);
420 mi = (struct aac_mntinfo *)&fib->data[0];
422 /* loop over possible containers */
424 /* request information on this container */
425 bzero(mi, sizeof(struct aac_mntinfo));
426 mi->Command = VM_NameServe;
427 mi->MntType = FT_FILESYS;
429 if (aac_sync_fib(sc, ContainerCommand, 0, fib,
430 sizeof(struct aac_mntinfo))) {
431 device_printf(sc->aac_dev,
432 "error probing container %d", i);
437 mir = (struct aac_mntinforesp *)&fib->data[0];
438 /* XXX Need to check if count changed */
439 count = mir->MntRespCount;
440 aac_add_container(sc, mir, 0);
442 } while ((i < count) && (i < AAC_MAX_CONTAINERS));
444 aac_release_sync_fib(sc);
445 AAC_LOCK_RELEASE(&sc->aac_io_lock);
447 /* poke the bus to actually attach the child devices */
448 if (bus_generic_attach(sc->aac_dev))
449 device_printf(sc->aac_dev, "bus_generic_attach failed\n");
451 /* mark the controller up */
452 sc->aac_state &= ~AAC_STATE_SUSPEND;
454 /* enable interrupts now */
455 AAC_UNMASK_INTERRUPTS(sc);
459 * Create a device to respresent a new container
462 aac_add_container(struct aac_softc *sc, struct aac_mntinforesp *mir, int f)
464 struct aac_container *co;
468 * Check container volume type for validity. Note that many of
469 * the possible types may never show up.
471 if ((mir->Status == ST_OK) && (mir->MntTable[0].VolType != CT_NONE)) {
472 co = (struct aac_container *)kmalloc(sizeof *co, M_AACBUF,
474 debug(1, "id %x name '%.16s' size %u type %d",
475 mir->MntTable[0].ObjectId,
476 mir->MntTable[0].FileSystemName,
477 mir->MntTable[0].Capacity, mir->MntTable[0].VolType);
479 if ((child = device_add_child(sc->aac_dev, "aacd", -1)) == NULL)
480 device_printf(sc->aac_dev, "device_add_child failed\n");
482 device_set_ivars(child, co);
483 device_set_desc(child, aac_describe_code(aac_container_types,
484 mir->MntTable[0].VolType));
487 bcopy(&mir->MntTable[0], &co->co_mntobj,
488 sizeof(struct aac_mntobj));
489 AAC_LOCK_ACQUIRE(&sc->aac_container_lock);
490 TAILQ_INSERT_TAIL(&sc->aac_container_tqh, co, co_link);
491 AAC_LOCK_RELEASE(&sc->aac_container_lock);
496 * Free all of the resources associated with (sc)
498 * Should not be called if the controller is active.
501 aac_free(struct aac_softc *sc)
506 /* remove the control device */
507 if (sc->aac_dev_t != NULL)
508 destroy_dev(sc->aac_dev_t);
510 /* throw away any FIB buffers, discard the FIB DMA tag */
511 aac_free_commands(sc);
512 if (sc->aac_fib_dmat)
513 bus_dma_tag_destroy(sc->aac_fib_dmat);
515 kfree(sc->aac_commands, M_AACBUF);
517 /* destroy the common area */
518 if (sc->aac_common) {
519 bus_dmamap_unload(sc->aac_common_dmat, sc->aac_common_dmamap);
520 bus_dmamem_free(sc->aac_common_dmat, sc->aac_common,
521 sc->aac_common_dmamap);
523 if (sc->aac_common_dmat)
524 bus_dma_tag_destroy(sc->aac_common_dmat);
526 /* disconnect the interrupt handler */
528 bus_teardown_intr(sc->aac_dev, sc->aac_irq, sc->aac_intr);
529 if (sc->aac_irq != NULL)
530 bus_release_resource(sc->aac_dev, SYS_RES_IRQ, sc->aac_irq_rid,
533 /* destroy data-transfer DMA tag */
534 if (sc->aac_buffer_dmat)
535 bus_dma_tag_destroy(sc->aac_buffer_dmat);
537 /* destroy the parent DMA tag */
538 if (sc->aac_parent_dmat)
539 bus_dma_tag_destroy(sc->aac_parent_dmat);
541 /* release the register window mapping */
542 if (sc->aac_regs_resource != NULL) {
543 bus_release_resource(sc->aac_dev, SYS_RES_MEMORY,
544 sc->aac_regs_rid, sc->aac_regs_resource);
546 dev_ops_remove_minor(&aac_ops, device_get_unit(sc->aac_dev));
550 * Disconnect from the controller completely, in preparation for unload.
553 aac_detach(device_t dev)
555 struct aac_softc *sc;
556 struct aac_container *co;
562 sc = device_get_softc(dev);
564 callout_stop(&sc->aac_watchdog);
566 if (sc->aac_state & AAC_STATE_OPEN)
569 /* Remove the child containers */
570 while ((co = TAILQ_FIRST(&sc->aac_container_tqh)) != NULL) {
571 error = device_delete_child(dev, co->co_disk);
574 TAILQ_REMOVE(&sc->aac_container_tqh, co, co_link);
578 /* Remove the CAM SIMs */
579 while ((sim = TAILQ_FIRST(&sc->aac_sim_tqh)) != NULL) {
580 TAILQ_REMOVE(&sc->aac_sim_tqh, sim, sim_link);
581 error = device_delete_child(dev, sim->sim_dev);
584 kfree(sim, M_AACBUF);
587 if (sc->aifflags & AAC_AIFFLAGS_RUNNING) {
588 sc->aifflags |= AAC_AIFFLAGS_EXIT;
589 wakeup(sc->aifthread);
590 tsleep(sc->aac_dev, PCATCH, "aacdch", 30 * hz);
593 if (sc->aifflags & AAC_AIFFLAGS_RUNNING)
594 panic("Cannot shutdown AIF thread\n");
596 if ((error = aac_shutdown(dev)))
599 EVENTHANDLER_DEREGISTER(shutdown_post_sync, sc->eh);
603 lockuninit(&sc->aac_aifq_lock);
604 lockuninit(&sc->aac_io_lock);
605 lockuninit(&sc->aac_container_lock);
611 * Bring the controller down to a dormant state and detach all child devices.
613 * This function is called before detach or system shutdown.
615 * Note that we can assume that the bioq on the controller is empty, as we won't
616 * allow shutdown if any device is open.
619 aac_shutdown(device_t dev)
621 struct aac_softc *sc;
623 struct aac_close_command *cc;
627 sc = device_get_softc(dev);
629 sc->aac_state |= AAC_STATE_SUSPEND;
632 * Send a Container shutdown followed by a HostShutdown FIB to the
633 * controller to convince it that we don't want to talk to it anymore.
634 * We've been closed and all I/O completed already
636 device_printf(sc->aac_dev, "shutting down controller...");
638 AAC_LOCK_ACQUIRE(&sc->aac_io_lock);
639 aac_alloc_sync_fib(sc, &fib);
640 cc = (struct aac_close_command *)&fib->data[0];
642 bzero(cc, sizeof(struct aac_close_command));
643 cc->Command = VM_CloseAll;
644 cc->ContainerId = 0xffffffff;
645 if (aac_sync_fib(sc, ContainerCommand, 0, fib,
646 sizeof(struct aac_close_command)))
647 kprintf("FAILED.\n");
654 * XXX Issuing this command to the controller makes it shut down
655 * but also keeps it from coming back up without a reset of the
656 * PCI bus. This is not desirable if you are just unloading the
657 * driver module with the intent to reload it later.
659 if (aac_sync_fib(sc, FsaHostShutdown, AAC_FIBSTATE_SHUTDOWN,
661 kprintf("FAILED.\n");
668 AAC_MASK_INTERRUPTS(sc);
669 aac_release_sync_fib(sc);
670 AAC_LOCK_RELEASE(&sc->aac_io_lock);
676 * Bring the controller to a quiescent state, ready for system suspend.
679 aac_suspend(device_t dev)
681 struct aac_softc *sc;
685 sc = device_get_softc(dev);
687 sc->aac_state |= AAC_STATE_SUSPEND;
689 AAC_MASK_INTERRUPTS(sc);
694 * Bring the controller back to a state ready for operation.
697 aac_resume(device_t dev)
699 struct aac_softc *sc;
703 sc = device_get_softc(dev);
705 sc->aac_state &= ~AAC_STATE_SUSPEND;
706 AAC_UNMASK_INTERRUPTS(sc);
711 * Interrupt handler for NEW_COMM interface.
714 aac_new_intr(void *arg)
716 struct aac_softc *sc;
717 u_int32_t index, fast;
718 struct aac_command *cm;
724 sc = (struct aac_softc *)arg;
726 AAC_LOCK_ACQUIRE(&sc->aac_io_lock);
728 index = AAC_GET_OUTB_QUEUE(sc);
729 if (index == 0xffffffff)
730 index = AAC_GET_OUTB_QUEUE(sc);
731 if (index == 0xffffffff)
734 if (index == 0xfffffffe) {
735 /* XXX This means that the controller wants
736 * more work. Ignore it for now.
741 fib = (struct aac_fib *)kmalloc(sizeof *fib, M_AACBUF,
744 for (i = 0; i < sizeof(struct aac_fib)/4; ++i)
745 ((u_int32_t *)fib)[i] = AAC_GETREG4(sc, index + i*4);
746 aac_handle_aif(sc, fib);
747 kfree(fib, M_AACBUF);
750 * AIF memory is owned by the adapter, so let it
751 * know that we are done with it.
753 AAC_SET_OUTB_QUEUE(sc, index);
754 AAC_CLEAR_ISTATUS(sc, AAC_DB_RESPONSE_READY);
757 cm = sc->aac_commands + (index >> 2);
760 fib->Header.XferState |= AAC_FIBSTATE_DONEADAP;
761 *((u_int32_t *)(fib->data)) = AAC_ERROR_NORMAL;
764 aac_unmap_command(cm);
765 cm->cm_flags |= AAC_CMD_COMPLETED;
767 /* is there a completion handler? */
768 if (cm->cm_complete != NULL) {
771 /* assume that someone is sleeping on this
776 sc->flags &= ~AAC_QUEUE_FRZN;
779 /* see if we can start some more I/O */
780 if ((sc->flags & AAC_QUEUE_FRZN) == 0)
783 AAC_LOCK_RELEASE(&sc->aac_io_lock);
787 aac_fast_intr(void *arg)
789 struct aac_softc *sc;
794 sc = (struct aac_softc *)arg;
797 * Read the status register directly. This is faster than taking the
798 * driver lock and reading the queues directly. It also saves having
799 * to turn parts of the driver lock into a spin mutex, which would be
802 reason = AAC_GET_ISTATUS(sc);
803 AAC_CLEAR_ISTATUS(sc, reason);
805 /* handle completion processing */
806 if (reason & AAC_DB_RESPONSE_READY)
807 taskqueue_enqueue(taskqueue_swi, &sc->aac_task_complete);
809 /* controller wants to talk to us */
810 if (reason & (AAC_DB_PRINTF | AAC_DB_COMMAND_READY)) {
812 * XXX Make sure that we don't get fooled by strange messages
813 * that start with a NULL.
815 if ((reason & AAC_DB_PRINTF) &&
816 (sc->aac_common->ac_printf[0] == 0))
817 sc->aac_common->ac_printf[0] = 32;
820 * This might miss doing the actual wakeup. However, the
821 * ssleep that this is waking up has a timeout, so it will
822 * wake up eventually. AIFs and printfs are low enough
823 * priority that they can handle hanging out for a few seconds
826 wakeup(sc->aifthread);
835 * Start as much queued I/O as possible on the controller
838 aac_startio(struct aac_softc *sc)
840 struct aac_command *cm;
844 if (sc->flags & AAC_QUEUE_FRZN)
849 * Try to get a command that's been put off for lack of
852 cm = aac_dequeue_ready(sc);
855 * Try to build a command off the bio queue (ignore error
859 aac_bio_command(sc, &cm);
866 * Try to give the command to the controller. Any error is
867 * catastrophic since it means that bus_dmamap_load() failed.
869 if (aac_map_command(cm) != 0)
870 panic("aac: error mapping command %p\n", cm);
875 * Deliver a command to the controller; allocate controller resources at the
876 * last moment when possible.
879 aac_map_command(struct aac_command *cm)
881 struct aac_softc *sc;
889 /* don't map more than once */
890 if (cm->cm_flags & AAC_CMD_MAPPED)
891 panic("aac: command %p already mapped", cm);
893 if (cm->cm_datalen != 0) {
894 error = bus_dmamap_load(sc->aac_buffer_dmat, cm->cm_datamap,
895 cm->cm_data, cm->cm_datalen,
896 aac_map_command_sg, cm, 0);
897 if (error == EINPROGRESS) {
898 debug(1, "freezing queue\n");
899 sc->flags |= AAC_QUEUE_FRZN;
903 aac_map_command_sg(cm, NULL, 0, 0);
909 * Handle notification of one or more FIBs coming from the controller.
912 aac_command_thread(struct aac_softc *sc)
920 AAC_LOCK_ACQUIRE(&sc->aac_io_lock);
921 sc->aifflags = AAC_AIFFLAGS_RUNNING;
923 while ((sc->aifflags & AAC_AIFFLAGS_EXIT) == 0) {
925 if ((sc->aifflags & AAC_AIFFLAGS_PENDING) == 0) {
926 tsleep_interlock(sc->aifthread, 0);
927 AAC_LOCK_RELEASE(&sc->aac_io_lock);
928 retval = tsleep(sc->aifthread, PINTERLOCKED,
929 "aifthd", AAC_PERIODIC_INTERVAL * hz);
930 AAC_LOCK_ACQUIRE(&sc->aac_io_lock);
933 * First see if any FIBs need to be allocated. This needs
934 * to be called without the driver lock because contigmalloc
935 * will grab Giant, and would result in an LOR.
937 if ((sc->aifflags & AAC_AIFFLAGS_ALLOCFIBS) != 0) {
938 AAC_LOCK_RELEASE(&sc->aac_io_lock);
939 aac_alloc_commands(sc);
940 AAC_LOCK_ACQUIRE(&sc->aac_io_lock);
941 sc->aifflags &= ~AAC_AIFFLAGS_ALLOCFIBS;
946 * While we're here, check to see if any commands are stuck.
947 * This is pretty low-priority, so it's ok if it doesn't
950 if (retval == EWOULDBLOCK)
953 /* Check the hardware printf message buffer */
954 if (sc->aac_common->ac_printf[0] != 0)
955 aac_print_printf(sc);
957 /* Also check to see if the adapter has a command for us. */
958 if (sc->flags & AAC_FLAGS_NEW_COMM)
961 if (aac_dequeue_fib(sc, AAC_HOST_NORM_CMD_QUEUE,
965 AAC_PRINT_FIB(sc, fib);
967 switch (fib->Header.Command) {
969 aac_handle_aif(sc, fib);
972 device_printf(sc->aac_dev, "unknown command "
973 "from controller\n");
977 if ((fib->Header.XferState == 0) ||
978 (fib->Header.StructType != AAC_FIBTYPE_TFIB)) {
982 /* Return the AIF to the controller. */
983 if (fib->Header.XferState & AAC_FIBSTATE_FROMADAP) {
984 fib->Header.XferState |= AAC_FIBSTATE_DONEHOST;
985 *(AAC_FSAStatus*)fib->data = ST_OK;
987 /* XXX Compute the Size field? */
988 size = fib->Header.Size;
989 if (size > sizeof(struct aac_fib)) {
990 size = sizeof(struct aac_fib);
991 fib->Header.Size = size;
994 * Since we did not generate this command, it
995 * cannot go through the normal
996 * enqueue->startio chain.
998 aac_enqueue_response(sc,
999 AAC_ADAP_NORM_RESP_QUEUE,
1004 sc->aifflags &= ~AAC_AIFFLAGS_RUNNING;
1005 AAC_LOCK_RELEASE(&sc->aac_io_lock);
1006 wakeup(sc->aac_dev);
1012 * Process completed commands.
1015 aac_complete(void *context, int pending)
1017 struct aac_softc *sc;
1018 struct aac_command *cm;
1019 struct aac_fib *fib;
1024 sc = (struct aac_softc *)context;
1026 AAC_LOCK_ACQUIRE(&sc->aac_io_lock);
1028 /* pull completed commands off the queue */
1030 /* look for completed FIBs on our queue */
1031 if (aac_dequeue_fib(sc, AAC_HOST_NORM_RESP_QUEUE, &fib_size,
1033 break; /* nothing to do */
1035 /* get the command, unmap and queue for later processing */
1036 cm = sc->aac_commands + fib->Header.SenderData;
1038 AAC_PRINT_FIB(sc, fib);
1041 aac_remove_busy(cm);
1042 aac_unmap_command(cm); /* XXX defer? */
1043 cm->cm_flags |= AAC_CMD_COMPLETED;
1045 /* is there a completion handler? */
1046 if (cm->cm_complete != NULL) {
1047 cm->cm_complete(cm);
1049 /* assume that someone is sleeping on this command */
1054 /* see if we can start some more I/O */
1055 sc->flags &= ~AAC_QUEUE_FRZN;
1058 AAC_LOCK_RELEASE(&sc->aac_io_lock);
1062 * Handle a bio submitted from a disk device.
1065 aac_submit_bio(struct aac_disk *ad, struct bio *bio)
1067 struct aac_softc *sc;
1071 bio->bio_driver_info = ad;
1072 sc = ad->ad_controller;
1074 /* queue the BIO and try to get some work done */
1075 aac_enqueue_bio(sc, bio);
1080 * Get a bio and build a command to go with it.
1083 aac_bio_command(struct aac_softc *sc, struct aac_command **cmp)
1085 struct aac_command *cm;
1086 struct aac_fib *fib;
1087 struct aac_disk *ad;
1093 /* get the resources we will need */
1096 if (aac_alloc_command(sc, &cm)) /* get a command */
1098 if ((bio = aac_dequeue_bio(sc)) == NULL)
1101 /* fill out the command */
1103 cm->cm_data = (void *)bp->b_data;
1104 cm->cm_datalen = bp->b_bcount;
1105 cm->cm_complete = aac_bio_complete;
1106 cm->cm_private = bio;
1107 cm->cm_timestamp = time_second;
1108 cm->cm_queue = AAC_ADAP_NORM_CMD_QUEUE;
1112 fib->Header.Size = sizeof(struct aac_fib_header);
1113 fib->Header.XferState =
1114 AAC_FIBSTATE_HOSTOWNED |
1115 AAC_FIBSTATE_INITIALISED |
1116 AAC_FIBSTATE_EMPTY |
1117 AAC_FIBSTATE_FROMHOST |
1118 AAC_FIBSTATE_REXPECTED |
1120 AAC_FIBSTATE_ASYNC |
1121 AAC_FIBSTATE_FAST_RESPONSE;
1123 /* build the read/write request */
1124 ad = (struct aac_disk *)bio->bio_driver_info;
1126 if (sc->flags & AAC_FLAGS_RAW_IO) {
1127 struct aac_raw_io *raw;
1128 raw = (struct aac_raw_io *)&fib->data[0];
1129 fib->Header.Command = RawIo;
1130 raw->BlockNumber = bio->bio_offset / AAC_BLOCK_SIZE;
1131 raw->ByteCount = bp->b_bcount;
1132 raw->ContainerId = ad->ad_container->co_mntobj.ObjectId;
1134 raw->BpComplete = 0;
1135 fib->Header.Size += sizeof(struct aac_raw_io);
1136 cm->cm_sgtable = (struct aac_sg_table *)&raw->SgMapRaw;
1137 if (bp->b_cmd == BUF_CMD_READ) {
1139 cm->cm_flags |= AAC_CMD_DATAIN;
1142 cm->cm_flags |= AAC_CMD_DATAOUT;
1144 } else if ((sc->flags & AAC_FLAGS_SG_64BIT) == 0) {
1145 fib->Header.Command = ContainerCommand;
1146 if (bp->b_cmd == BUF_CMD_READ) {
1147 struct aac_blockread *br;
1148 br = (struct aac_blockread *)&fib->data[0];
1149 br->Command = VM_CtBlockRead;
1150 br->ContainerId = ad->ad_container->co_mntobj.ObjectId;
1151 br->BlockNumber = bio->bio_offset / AAC_BLOCK_SIZE;
1152 br->ByteCount = bp->b_bcount;
1153 fib->Header.Size += sizeof(struct aac_blockread);
1154 cm->cm_sgtable = &br->SgMap;
1155 cm->cm_flags |= AAC_CMD_DATAIN;
1157 struct aac_blockwrite *bw;
1158 bw = (struct aac_blockwrite *)&fib->data[0];
1159 bw->Command = VM_CtBlockWrite;
1160 bw->ContainerId = ad->ad_container->co_mntobj.ObjectId;
1161 bw->BlockNumber = bio->bio_offset / AAC_BLOCK_SIZE;
1162 bw->ByteCount = bp->b_bcount;
1163 bw->Stable = CUNSTABLE;
1164 fib->Header.Size += sizeof(struct aac_blockwrite);
1165 cm->cm_flags |= AAC_CMD_DATAOUT;
1166 cm->cm_sgtable = &bw->SgMap;
1169 fib->Header.Command = ContainerCommand64;
1170 if (bp->b_cmd == BUF_CMD_READ) {
1171 struct aac_blockread64 *br;
1172 br = (struct aac_blockread64 *)&fib->data[0];
1173 br->Command = VM_CtHostRead64;
1174 br->ContainerId = ad->ad_container->co_mntobj.ObjectId;
1175 br->SectorCount = bp->b_bcount / AAC_BLOCK_SIZE;
1176 br->BlockNumber = bio->bio_offset / AAC_BLOCK_SIZE;
1179 fib->Header.Size += sizeof(struct aac_blockread64);
1180 cm->cm_flags |= AAC_CMD_DATAOUT;
1181 cm->cm_sgtable = (struct aac_sg_table *)&br->SgMap64;
1183 struct aac_blockwrite64 *bw;
1184 bw = (struct aac_blockwrite64 *)&fib->data[0];
1185 bw->Command = VM_CtHostWrite64;
1186 bw->ContainerId = ad->ad_container->co_mntobj.ObjectId;
1187 bw->SectorCount = bp->b_bcount / AAC_BLOCK_SIZE;
1188 bw->BlockNumber = bio->bio_offset / AAC_BLOCK_SIZE;
1191 fib->Header.Size += sizeof(struct aac_blockwrite64);
1192 cm->cm_flags |= AAC_CMD_DATAIN;
1193 cm->cm_sgtable = (struct aac_sg_table *)&bw->SgMap64;
1202 aac_enqueue_bio(sc, bio);
1204 aac_release_command(cm);
1209 * Handle a bio-instigated command that has been completed.
1212 aac_bio_complete(struct aac_command *cm)
1214 struct aac_blockread_response *brr;
1215 struct aac_blockwrite_response *bwr;
1219 AAC_FSAStatus status;
1221 /* fetch relevant status and then release the command */
1222 bio = (struct bio *)cm->cm_private;
1224 if (bp->b_cmd == BUF_CMD_READ) {
1225 brr = (struct aac_blockread_response *)&cm->cm_fib->data[0];
1226 status = brr->Status;
1228 bwr = (struct aac_blockwrite_response *)&cm->cm_fib->data[0];
1229 status = bwr->Status;
1231 aac_release_command(cm);
1233 /* fix up the bio based on status */
1234 if (status == ST_OK) {
1239 bp->b_flags |= B_ERROR;
1240 /* pass an error string out to the disk layer */
1241 code = aac_describe_code(aac_command_status_table, status);
1243 aac_biodone(bio, code);
1247 * Dump a block of data to the controller. If the queue is full, tell the
1248 * caller to hold off and wait for the queue to drain.
1251 aac_dump_enqueue(struct aac_disk *ad, u_int64_t lba, void *data, int dumppages)
1253 struct aac_softc *sc;
1254 struct aac_command *cm;
1255 struct aac_fib *fib;
1256 struct aac_blockwrite *bw;
1258 sc = ad->ad_controller;
1261 KKASSERT(lba <= 0x100000000ULL);
1263 if (aac_alloc_command(sc, &cm))
1266 /* fill out the command */
1268 cm->cm_datalen = dumppages * PAGE_SIZE;
1269 cm->cm_complete = NULL;
1270 cm->cm_private = NULL;
1271 cm->cm_timestamp = time_second;
1272 cm->cm_queue = AAC_ADAP_NORM_CMD_QUEUE;
1276 fib->Header.XferState =
1277 AAC_FIBSTATE_HOSTOWNED |
1278 AAC_FIBSTATE_INITIALISED |
1279 AAC_FIBSTATE_FROMHOST |
1280 AAC_FIBSTATE_REXPECTED |
1282 fib->Header.Command = ContainerCommand;
1283 fib->Header.Size = sizeof(struct aac_fib_header);
1285 bw = (struct aac_blockwrite *)&fib->data[0];
1286 bw->Command = VM_CtBlockWrite;
1287 bw->ContainerId = ad->ad_container->co_mntobj.ObjectId;
1288 bw->BlockNumber = lba;
1289 bw->ByteCount = dumppages * PAGE_SIZE;
1290 bw->Stable = CUNSTABLE; /* XXX what's appropriate here? */
1291 fib->Header.Size += sizeof(struct aac_blockwrite);
1292 cm->cm_flags |= AAC_CMD_DATAOUT;
1293 cm->cm_sgtable = &bw->SgMap;
1295 return (aac_map_command(cm));
1299 * Wait for the card's queue to drain when dumping. Also check for monitor
1303 aac_dump_complete(struct aac_softc *sc)
1305 struct aac_fib *fib;
1306 struct aac_command *cm;
1308 u_int32_t pi, ci, fib_size;
1311 reason = AAC_GET_ISTATUS(sc);
1312 if (reason & AAC_DB_RESPONSE_READY) {
1313 AAC_CLEAR_ISTATUS(sc, AAC_DB_RESPONSE_READY);
1315 if (aac_dequeue_fib(sc,
1316 AAC_HOST_NORM_RESP_QUEUE,
1319 cm = (struct aac_command *)
1320 fib->Header.SenderData;
1322 AAC_PRINT_FIB(sc, fib);
1324 aac_remove_busy(cm);
1325 aac_unmap_command(cm);
1326 aac_enqueue_complete(cm);
1327 aac_release_command(cm);
1331 if (reason & AAC_DB_PRINTF) {
1332 AAC_CLEAR_ISTATUS(sc, AAC_DB_PRINTF);
1333 aac_print_printf(sc);
1335 pi = sc->aac_queues->qt_qindex[AAC_ADAP_NORM_CMD_QUEUE][
1336 AAC_PRODUCER_INDEX];
1337 ci = sc->aac_queues->qt_qindex[AAC_ADAP_NORM_CMD_QUEUE][
1338 AAC_CONSUMER_INDEX];
1345 * Submit a command to the controller, return when it completes.
1346 * XXX This is very dangerous! If the card has gone out to lunch, we could
1347 * be stuck here forever. At the same time, signals are not caught
1348 * because there is a risk that a signal could wakeup the sleep before
1349 * the card has a chance to complete the command. Since there is no way
1350 * to cancel a command that is in progress, we can't protect against the
1351 * card completing a command late and spamming the command and data
1352 * memory. So, we are held hostage until the command completes.
1355 aac_wait_command(struct aac_command *cm)
1357 struct aac_softc *sc;
1364 /* Put the command on the ready queue and get things going */
1365 cm->cm_queue = AAC_ADAP_NORM_CMD_QUEUE;
1366 aac_enqueue_ready(cm);
1369 KKASSERT(lockstatus(&sc->aac_io_lock, curthread) != 0);
1370 tsleep_interlock(cm, 0);
1371 AAC_LOCK_RELEASE(&sc->aac_io_lock);
1372 error = tsleep(cm, PINTERLOCKED, "aacwait", 0);
1373 AAC_LOCK_ACQUIRE(&sc->aac_io_lock);
1378 *Command Buffer Management
1382 * Allocate a command.
1385 aac_alloc_command(struct aac_softc *sc, struct aac_command **cmp)
1387 struct aac_command *cm;
1391 if ((cm = aac_dequeue_free(sc)) == NULL) {
1392 if (sc->total_fibs < sc->aac_max_fibs) {
1393 sc->aifflags |= AAC_AIFFLAGS_ALLOCFIBS;
1394 wakeup(sc->aifthread);
1404 * Release a command back to the freelist.
1407 aac_release_command(struct aac_command *cm)
1409 struct aac_event *event;
1410 struct aac_softc *sc;
1414 /* (re)initialise the command/FIB */
1415 cm->cm_sgtable = NULL;
1417 cm->cm_complete = NULL;
1418 cm->cm_private = NULL;
1419 cm->cm_fib->Header.XferState = AAC_FIBSTATE_EMPTY;
1420 cm->cm_fib->Header.StructType = AAC_FIBTYPE_TFIB;
1421 cm->cm_fib->Header.Flags = 0;
1422 cm->cm_fib->Header.SenderSize = cm->cm_sc->aac_max_fib_size;
1425 * These are duplicated in aac_start to cover the case where an
1426 * intermediate stage may have destroyed them. They're left
1427 * initialised here for debugging purposes only.
1429 cm->cm_fib->Header.ReceiverFibAddress = (u_int32_t)cm->cm_fibphys;
1430 cm->cm_fib->Header.SenderData = 0;
1432 aac_enqueue_free(cm);
1435 event = TAILQ_FIRST(&sc->aac_ev_cmfree);
1436 if (event != NULL) {
1437 TAILQ_REMOVE(&sc->aac_ev_cmfree, event, ev_links);
1438 event->ev_callback(sc, event, event->ev_arg);
1443 * Map helper for command/FIB allocation.
1446 aac_map_command_helper(void *arg, bus_dma_segment_t *segs, int nseg, int error)
1450 fibphys = (uint64_t *)arg;
1454 *fibphys = segs[0].ds_addr;
1458 * Allocate and initialise commands/FIBs for this adapter.
1461 aac_alloc_commands(struct aac_softc *sc)
1463 struct aac_command *cm;
1464 struct aac_fibmap *fm;
1470 if (sc->total_fibs + sc->aac_max_fibs_alloc > sc->aac_max_fibs)
1473 fm = kmalloc(sizeof(struct aac_fibmap), M_AACBUF, M_INTWAIT | M_ZERO);
1475 /* allocate the FIBs in DMAable memory and load them */
1476 if (bus_dmamem_alloc(sc->aac_fib_dmat, (void **)&fm->aac_fibs,
1477 BUS_DMA_NOWAIT, &fm->aac_fibmap)) {
1478 device_printf(sc->aac_dev,
1479 "Not enough contiguous memory available.\n");
1480 kfree(fm, M_AACBUF);
1484 /* Ignore errors since this doesn't bounce */
1485 bus_dmamap_load(sc->aac_fib_dmat, fm->aac_fibmap, fm->aac_fibs,
1486 sc->aac_max_fibs_alloc * sc->aac_max_fib_size,
1487 aac_map_command_helper, &fibphys, 0);
1489 /* initialise constant fields in the command structure */
1490 bzero(fm->aac_fibs, sc->aac_max_fibs_alloc * sc->aac_max_fib_size);
1491 for (i = 0; i < sc->aac_max_fibs_alloc; i++) {
1492 cm = sc->aac_commands + sc->total_fibs;
1493 fm->aac_commands = cm;
1495 cm->cm_fib = (struct aac_fib *)
1496 ((u_int8_t *)fm->aac_fibs + i*sc->aac_max_fib_size);
1497 cm->cm_fibphys = fibphys + i*sc->aac_max_fib_size;
1498 cm->cm_index = sc->total_fibs;
1500 if ((error = bus_dmamap_create(sc->aac_buffer_dmat, 0,
1501 &cm->cm_datamap)) != 0)
1503 AAC_LOCK_ACQUIRE(&sc->aac_io_lock);
1504 aac_release_command(cm);
1506 AAC_LOCK_RELEASE(&sc->aac_io_lock);
1510 AAC_LOCK_ACQUIRE(&sc->aac_io_lock);
1511 TAILQ_INSERT_TAIL(&sc->aac_fibmap_tqh, fm, fm_link);
1512 debug(1, "total_fibs= %d\n", sc->total_fibs);
1513 AAC_LOCK_RELEASE(&sc->aac_io_lock);
1517 bus_dmamap_unload(sc->aac_fib_dmat, fm->aac_fibmap);
1518 bus_dmamem_free(sc->aac_fib_dmat, fm->aac_fibs, fm->aac_fibmap);
1519 kfree(fm, M_AACBUF);
1524 * Free FIBs owned by this adapter.
1527 aac_free_commands(struct aac_softc *sc)
1529 struct aac_fibmap *fm;
1530 struct aac_command *cm;
1535 while ((fm = TAILQ_FIRST(&sc->aac_fibmap_tqh)) != NULL) {
1537 TAILQ_REMOVE(&sc->aac_fibmap_tqh, fm, fm_link);
1539 * We check against total_fibs to handle partially
1542 for (i = 0; i < sc->aac_max_fibs_alloc && sc->total_fibs--; i++) {
1543 cm = fm->aac_commands + i;
1544 bus_dmamap_destroy(sc->aac_buffer_dmat, cm->cm_datamap);
1546 bus_dmamap_unload(sc->aac_fib_dmat, fm->aac_fibmap);
1547 bus_dmamem_free(sc->aac_fib_dmat, fm->aac_fibs, fm->aac_fibmap);
1548 kfree(fm, M_AACBUF);
1553 * Command-mapping helper function - populate this command's s/g table.
1556 aac_map_command_sg(void *arg, bus_dma_segment_t *segs, int nseg, int error)
1558 struct aac_softc *sc;
1559 struct aac_command *cm;
1560 struct aac_fib *fib;
1565 cm = (struct aac_command *)arg;
1569 /* copy into the FIB */
1570 if (cm->cm_sgtable != NULL) {
1571 if (fib->Header.Command == RawIo) {
1572 struct aac_sg_tableraw *sg;
1573 sg = (struct aac_sg_tableraw *)cm->cm_sgtable;
1575 for (i = 0; i < nseg; i++) {
1576 sg->SgEntryRaw[i].SgAddress = segs[i].ds_addr;
1577 sg->SgEntryRaw[i].SgByteCount = segs[i].ds_len;
1578 sg->SgEntryRaw[i].Next = 0;
1579 sg->SgEntryRaw[i].Prev = 0;
1580 sg->SgEntryRaw[i].Flags = 0;
1582 /* update the FIB size for the s/g count */
1583 fib->Header.Size += nseg*sizeof(struct aac_sg_entryraw);
1584 } else if ((cm->cm_sc->flags & AAC_FLAGS_SG_64BIT) == 0) {
1585 struct aac_sg_table *sg;
1586 sg = cm->cm_sgtable;
1588 for (i = 0; i < nseg; i++) {
1589 sg->SgEntry[i].SgAddress = segs[i].ds_addr;
1590 sg->SgEntry[i].SgByteCount = segs[i].ds_len;
1592 /* update the FIB size for the s/g count */
1593 fib->Header.Size += nseg*sizeof(struct aac_sg_entry);
1595 struct aac_sg_table64 *sg;
1596 sg = (struct aac_sg_table64 *)cm->cm_sgtable;
1598 for (i = 0; i < nseg; i++) {
1599 sg->SgEntry64[i].SgAddress = segs[i].ds_addr;
1600 sg->SgEntry64[i].SgByteCount = segs[i].ds_len;
1602 /* update the FIB size for the s/g count */
1603 fib->Header.Size += nseg*sizeof(struct aac_sg_entry64);
1607 /* Fix up the address values in the FIB. Use the command array index
1608 * instead of a pointer since these fields are only 32 bits. Shift
1609 * the SenderFibAddress over to make room for the fast response bit
1610 * and for the AIF bit
1612 cm->cm_fib->Header.SenderFibAddress = (cm->cm_index << 2);
1613 cm->cm_fib->Header.ReceiverFibAddress = (u_int32_t)cm->cm_fibphys;
1615 /* save a pointer to the command for speedy reverse-lookup */
1616 cm->cm_fib->Header.SenderData = cm->cm_index;
1618 if (cm->cm_flags & AAC_CMD_DATAIN)
1619 bus_dmamap_sync(sc->aac_buffer_dmat, cm->cm_datamap,
1620 BUS_DMASYNC_PREREAD);
1621 if (cm->cm_flags & AAC_CMD_DATAOUT)
1622 bus_dmamap_sync(sc->aac_buffer_dmat, cm->cm_datamap,
1623 BUS_DMASYNC_PREWRITE);
1624 cm->cm_flags |= AAC_CMD_MAPPED;
1626 if (sc->flags & AAC_FLAGS_NEW_COMM) {
1627 int count = 10000000L;
1628 while (AAC_SEND_COMMAND(sc, cm) != 0) {
1630 aac_unmap_command(cm);
1631 sc->flags |= AAC_QUEUE_FRZN;
1632 aac_requeue_ready(cm);
1635 DELAY(5); /* wait 5 usec. */
1638 /* Put the FIB on the outbound queue */
1639 if (aac_enqueue_fib(sc, cm->cm_queue, cm) == EBUSY) {
1640 aac_unmap_command(cm);
1641 sc->flags |= AAC_QUEUE_FRZN;
1642 aac_requeue_ready(cm);
1648 * Unmap a command from controller-visible space.
1651 aac_unmap_command(struct aac_command *cm)
1653 struct aac_softc *sc;
1659 if (!(cm->cm_flags & AAC_CMD_MAPPED))
1662 if (cm->cm_datalen != 0) {
1663 if (cm->cm_flags & AAC_CMD_DATAIN)
1664 bus_dmamap_sync(sc->aac_buffer_dmat, cm->cm_datamap,
1665 BUS_DMASYNC_POSTREAD);
1666 if (cm->cm_flags & AAC_CMD_DATAOUT)
1667 bus_dmamap_sync(sc->aac_buffer_dmat, cm->cm_datamap,
1668 BUS_DMASYNC_POSTWRITE);
1670 bus_dmamap_unload(sc->aac_buffer_dmat, cm->cm_datamap);
1672 cm->cm_flags &= ~AAC_CMD_MAPPED;
1676 * Hardware Interface
1680 * Initialise the adapter.
1683 aac_common_map(void *arg, bus_dma_segment_t *segs, int nseg, int error)
1685 struct aac_softc *sc;
1689 sc = (struct aac_softc *)arg;
1691 sc->aac_common_busaddr = segs[0].ds_addr;
1695 aac_check_firmware(struct aac_softc *sc)
1697 u_int32_t major, minor, options = 0, atu_size = 0;
1703 * Retrieve the firmware version numbers. Dell PERC2/QC cards with
1704 * firmware version 1.x are not compatible with this driver.
1706 if (sc->flags & AAC_FLAGS_PERC2QC) {
1707 if (aac_sync_command(sc, AAC_MONKER_GETKERNVER, 0, 0, 0, 0,
1709 device_printf(sc->aac_dev,
1710 "Error reading firmware version\n");
1714 /* These numbers are stored as ASCII! */
1715 major = (AAC_GET_MAILBOX(sc, 1) & 0xff) - 0x30;
1716 minor = (AAC_GET_MAILBOX(sc, 2) & 0xff) - 0x30;
1718 device_printf(sc->aac_dev,
1719 "Firmware version %d.%d is not supported.\n",
1726 * Retrieve the capabilities/supported options word so we know what
1727 * work-arounds to enable. Some firmware revs don't support this
1730 if (aac_sync_command(sc, AAC_MONKER_GETINFO, 0, 0, 0, 0, &status)) {
1731 if (status != AAC_SRB_STS_INVALID_REQUEST) {
1732 device_printf(sc->aac_dev,
1733 "RequestAdapterInfo failed\n");
1737 options = AAC_GET_MAILBOX(sc, 1);
1738 atu_size = AAC_GET_MAILBOX(sc, 2);
1739 sc->supported_options = options;
1741 if ((options & AAC_SUPPORTED_4GB_WINDOW) != 0 &&
1742 (sc->flags & AAC_FLAGS_NO4GB) == 0)
1743 sc->flags |= AAC_FLAGS_4GB_WINDOW;
1744 if (options & AAC_SUPPORTED_NONDASD)
1745 sc->flags |= AAC_FLAGS_ENABLE_CAM;
1746 if ((options & AAC_SUPPORTED_SGMAP_HOST64) != 0
1747 && (sizeof(bus_addr_t) > 4)) {
1748 device_printf(sc->aac_dev,
1749 "Enabling 64-bit address support\n");
1750 sc->flags |= AAC_FLAGS_SG_64BIT;
1752 if ((options & AAC_SUPPORTED_NEW_COMM)
1753 && sc->aac_if.aif_send_command)
1754 sc->flags |= AAC_FLAGS_NEW_COMM;
1755 if (options & AAC_SUPPORTED_64BIT_ARRAYSIZE)
1756 sc->flags |= AAC_FLAGS_ARRAY_64BIT;
1759 /* Check for broken hardware that does a lower number of commands */
1760 sc->aac_max_fibs = (sc->flags & AAC_FLAGS_256FIBS ? 256:512);
1762 /* Remap mem. resource, if required */
1763 if ((sc->flags & AAC_FLAGS_NEW_COMM) &&
1764 atu_size > rman_get_size(sc->aac_regs_resource)) {
1765 bus_release_resource(
1766 sc->aac_dev, SYS_RES_MEMORY,
1767 sc->aac_regs_rid, sc->aac_regs_resource);
1768 sc->aac_regs_resource = bus_alloc_resource(
1769 sc->aac_dev, SYS_RES_MEMORY, &sc->aac_regs_rid,
1770 0ul, ~0ul, atu_size, RF_ACTIVE);
1771 if (sc->aac_regs_resource == NULL) {
1772 sc->aac_regs_resource = bus_alloc_resource_any(
1773 sc->aac_dev, SYS_RES_MEMORY,
1774 &sc->aac_regs_rid, RF_ACTIVE);
1775 if (sc->aac_regs_resource == NULL) {
1776 device_printf(sc->aac_dev,
1777 "couldn't allocate register window\n");
1780 sc->flags &= ~AAC_FLAGS_NEW_COMM;
1782 sc->aac_btag = rman_get_bustag(sc->aac_regs_resource);
1783 sc->aac_bhandle = rman_get_bushandle(sc->aac_regs_resource);
1786 /* Read preferred settings */
1787 sc->aac_max_fib_size = sizeof(struct aac_fib);
1788 sc->aac_max_sectors = 128; /* 64KB */
1789 if (sc->flags & AAC_FLAGS_SG_64BIT)
1790 sc->aac_sg_tablesize = (AAC_FIB_DATASIZE
1791 - sizeof(struct aac_blockwrite64)
1792 + sizeof(struct aac_sg_table64))
1793 / sizeof(struct aac_sg_table64);
1795 sc->aac_sg_tablesize = (AAC_FIB_DATASIZE
1796 - sizeof(struct aac_blockwrite)
1797 + sizeof(struct aac_sg_table))
1798 / sizeof(struct aac_sg_table);
1800 if (!aac_sync_command(sc, AAC_MONKER_GETCOMMPREF, 0, 0, 0, 0, NULL)) {
1801 options = AAC_GET_MAILBOX(sc, 1);
1802 sc->aac_max_fib_size = (options & 0xFFFF);
1803 sc->aac_max_sectors = (options >> 16) << 1;
1804 options = AAC_GET_MAILBOX(sc, 2);
1805 sc->aac_sg_tablesize = (options >> 16);
1806 options = AAC_GET_MAILBOX(sc, 3);
1807 sc->aac_max_fibs = (options & 0xFFFF);
1809 if (sc->aac_max_fib_size > PAGE_SIZE)
1810 sc->aac_max_fib_size = PAGE_SIZE;
1811 sc->aac_max_fibs_alloc = PAGE_SIZE / sc->aac_max_fib_size;
1817 aac_init(struct aac_softc *sc)
1819 struct aac_adapter_init *ip;
1821 u_int32_t code, qoffset;
1827 * First wait for the adapter to come ready.
1831 code = AAC_GET_FWSTATUS(sc);
1832 if (code & AAC_SELF_TEST_FAILED) {
1833 device_printf(sc->aac_dev, "FATAL: selftest failed\n");
1836 if (code & AAC_KERNEL_PANIC) {
1837 device_printf(sc->aac_dev,
1838 "FATAL: controller kernel panic\n");
1841 if (time_second > (then + AAC_BOOT_TIMEOUT)) {
1842 device_printf(sc->aac_dev,
1843 "FATAL: controller not coming ready, "
1844 "status %x\n", code);
1847 } while (!(code & AAC_UP_AND_RUNNING));
1851 * Create DMA tag for mapping buffers into controller-addressable space.
1853 if (bus_dma_tag_create(sc->aac_parent_dmat, /* parent */
1854 1, 0, /* algnmnt, boundary */
1855 (sc->flags & AAC_FLAGS_SG_64BIT) ?
1857 BUS_SPACE_MAXADDR_32BIT, /* lowaddr */
1858 BUS_SPACE_MAXADDR, /* highaddr */
1859 NULL, NULL, /* filter, filterarg */
1860 MAXBSIZE, /* maxsize */
1861 sc->aac_sg_tablesize, /* nsegments */
1862 MAXBSIZE, /* maxsegsize */
1863 BUS_DMA_ALLOCNOW, /* flags */
1864 &sc->aac_buffer_dmat)) {
1865 device_printf(sc->aac_dev, "can't allocate buffer DMA tag\n");
1870 * Create DMA tag for mapping FIBs into controller-addressable space..
1872 if (bus_dma_tag_create(sc->aac_parent_dmat, /* parent */
1873 1, 0, /* algnmnt, boundary */
1874 (sc->flags & AAC_FLAGS_4GB_WINDOW) ?
1875 BUS_SPACE_MAXADDR_32BIT :
1876 0x7fffffff, /* lowaddr */
1877 BUS_SPACE_MAXADDR, /* highaddr */
1878 NULL, NULL, /* filter, filterarg */
1879 sc->aac_max_fibs_alloc *
1880 sc->aac_max_fib_size, /* maxsize */
1882 sc->aac_max_fibs_alloc *
1883 sc->aac_max_fib_size, /* maxsegsize */
1885 &sc->aac_fib_dmat)) {
1886 device_printf(sc->aac_dev, "can't allocate FIB DMA tag\n");
1891 * Create DMA tag for the common structure and allocate it.
1893 if (bus_dma_tag_create(sc->aac_parent_dmat, /* parent */
1894 1, 0, /* algnmnt, boundary */
1895 (sc->flags & AAC_FLAGS_4GB_WINDOW) ?
1896 BUS_SPACE_MAXADDR_32BIT :
1897 0x7fffffff, /* lowaddr */
1898 BUS_SPACE_MAXADDR, /* highaddr */
1899 NULL, NULL, /* filter, filterarg */
1900 8192 + sizeof(struct aac_common), /* maxsize */
1902 BUS_SPACE_MAXSIZE_32BIT, /* maxsegsize */
1904 &sc->aac_common_dmat)) {
1905 device_printf(sc->aac_dev,
1906 "can't allocate common structure DMA tag\n");
1909 if (bus_dmamem_alloc(sc->aac_common_dmat, (void **)&sc->aac_common,
1910 BUS_DMA_NOWAIT, &sc->aac_common_dmamap)) {
1911 device_printf(sc->aac_dev, "can't allocate common structure\n");
1915 * Work around a bug in the 2120 and 2200 that cannot DMA commands
1916 * below address 8192 in physical memory.
1917 * XXX If the padding is not needed, can it be put to use instead
1920 bus_dmamap_load(sc->aac_common_dmat, sc->aac_common_dmamap,
1921 sc->aac_common, 8192 + sizeof(*sc->aac_common),
1922 aac_common_map, sc, 0);
1924 if (sc->aac_common_busaddr < 8192) {
1926 (struct aac_common *)((uint8_t *)sc->aac_common + 8192);
1927 sc->aac_common_busaddr += 8192;
1929 bzero(sc->aac_common, sizeof(*sc->aac_common));
1931 /* Allocate some FIBs and associated command structs */
1932 TAILQ_INIT(&sc->aac_fibmap_tqh);
1933 sc->aac_commands = kmalloc(sc->aac_max_fibs * sizeof(struct aac_command),
1934 M_AACBUF, M_INTWAIT | M_ZERO);
1935 while (sc->total_fibs < AAC_PREALLOCATE_FIBS) {
1936 if (aac_alloc_commands(sc) != 0)
1939 if (sc->total_fibs == 0)
1943 * Fill in the init structure. This tells the adapter about the
1944 * physical location of various important shared data structures.
1946 ip = &sc->aac_common->ac_init;
1947 ip->InitStructRevision = AAC_INIT_STRUCT_REVISION;
1948 if (sc->aac_max_fib_size > sizeof(struct aac_fib)) {
1949 ip->InitStructRevision = AAC_INIT_STRUCT_REVISION_4;
1950 sc->flags |= AAC_FLAGS_RAW_IO;
1952 ip->MiniPortRevision = AAC_INIT_STRUCT_MINIPORT_REVISION;
1954 ip->AdapterFibsPhysicalAddress = sc->aac_common_busaddr +
1955 offsetof(struct aac_common, ac_fibs);
1956 ip->AdapterFibsVirtualAddress = 0;
1957 ip->AdapterFibsSize = AAC_ADAPTER_FIBS * sizeof(struct aac_fib);
1958 ip->AdapterFibAlign = sizeof(struct aac_fib);
1960 ip->PrintfBufferAddress = sc->aac_common_busaddr +
1961 offsetof(struct aac_common, ac_printf);
1962 ip->PrintfBufferSize = AAC_PRINTF_BUFSIZE;
1965 * The adapter assumes that pages are 4K in size, except on some
1966 * broken firmware versions that do the page->byte conversion twice,
1967 * therefore 'assuming' that this value is in 16MB units (2^24).
1968 * Round up since the granularity is so high.
1970 /* XXX why should the adapter care? */
1971 ip->HostPhysMemPages = ctob((int)Maxmem) / AAC_PAGE_SIZE;
1972 if (sc->flags & AAC_FLAGS_BROKEN_MEMMAP) {
1973 ip->HostPhysMemPages =
1974 (ip->HostPhysMemPages + AAC_PAGE_SIZE) / AAC_PAGE_SIZE;
1976 ip->HostElapsedSeconds = time_second; /* reset later if invalid */
1979 if (sc->flags & AAC_FLAGS_NEW_COMM) {
1980 ip->InitFlags = INITFLAGS_NEW_COMM_SUPPORTED;
1981 device_printf(sc->aac_dev, "New comm. interface enabled\n");
1984 ip->MaxIoCommands = sc->aac_max_fibs;
1985 ip->MaxIoSize = sc->aac_max_sectors << 9;
1986 ip->MaxFibSize = sc->aac_max_fib_size;
1989 * Initialise FIB queues. Note that it appears that the layout of the
1990 * indexes and the segmentation of the entries may be mandated by the
1991 * adapter, which is only told about the base of the queue index fields.
1993 * The initial values of the indices are assumed to inform the adapter
1994 * of the sizes of the respective queues, and theoretically it could
1995 * work out the entire layout of the queue structures from this. We
1996 * take the easy route and just lay this area out like everyone else
1999 * The Linux driver uses a much more complex scheme whereby several
2000 * header records are kept for each queue. We use a couple of generic
2001 * list manipulation functions which 'know' the size of each list by
2002 * virtue of a table.
2004 qoffset = offsetof(struct aac_common, ac_qbuf) + AAC_QUEUE_ALIGN;
2005 qoffset &= ~(AAC_QUEUE_ALIGN - 1);
2007 (struct aac_queue_table *)((uintptr_t)sc->aac_common + qoffset);
2008 ip->CommHeaderAddress = sc->aac_common_busaddr + qoffset;
2010 sc->aac_queues->qt_qindex[AAC_HOST_NORM_CMD_QUEUE][AAC_PRODUCER_INDEX] =
2011 AAC_HOST_NORM_CMD_ENTRIES;
2012 sc->aac_queues->qt_qindex[AAC_HOST_NORM_CMD_QUEUE][AAC_CONSUMER_INDEX] =
2013 AAC_HOST_NORM_CMD_ENTRIES;
2014 sc->aac_queues->qt_qindex[AAC_HOST_HIGH_CMD_QUEUE][AAC_PRODUCER_INDEX] =
2015 AAC_HOST_HIGH_CMD_ENTRIES;
2016 sc->aac_queues->qt_qindex[AAC_HOST_HIGH_CMD_QUEUE][AAC_CONSUMER_INDEX] =
2017 AAC_HOST_HIGH_CMD_ENTRIES;
2018 sc->aac_queues->qt_qindex[AAC_ADAP_NORM_CMD_QUEUE][AAC_PRODUCER_INDEX] =
2019 AAC_ADAP_NORM_CMD_ENTRIES;
2020 sc->aac_queues->qt_qindex[AAC_ADAP_NORM_CMD_QUEUE][AAC_CONSUMER_INDEX] =
2021 AAC_ADAP_NORM_CMD_ENTRIES;
2022 sc->aac_queues->qt_qindex[AAC_ADAP_HIGH_CMD_QUEUE][AAC_PRODUCER_INDEX] =
2023 AAC_ADAP_HIGH_CMD_ENTRIES;
2024 sc->aac_queues->qt_qindex[AAC_ADAP_HIGH_CMD_QUEUE][AAC_CONSUMER_INDEX] =
2025 AAC_ADAP_HIGH_CMD_ENTRIES;
2026 sc->aac_queues->qt_qindex[AAC_HOST_NORM_RESP_QUEUE][AAC_PRODUCER_INDEX]=
2027 AAC_HOST_NORM_RESP_ENTRIES;
2028 sc->aac_queues->qt_qindex[AAC_HOST_NORM_RESP_QUEUE][AAC_CONSUMER_INDEX]=
2029 AAC_HOST_NORM_RESP_ENTRIES;
2030 sc->aac_queues->qt_qindex[AAC_HOST_HIGH_RESP_QUEUE][AAC_PRODUCER_INDEX]=
2031 AAC_HOST_HIGH_RESP_ENTRIES;
2032 sc->aac_queues->qt_qindex[AAC_HOST_HIGH_RESP_QUEUE][AAC_CONSUMER_INDEX]=
2033 AAC_HOST_HIGH_RESP_ENTRIES;
2034 sc->aac_queues->qt_qindex[AAC_ADAP_NORM_RESP_QUEUE][AAC_PRODUCER_INDEX]=
2035 AAC_ADAP_NORM_RESP_ENTRIES;
2036 sc->aac_queues->qt_qindex[AAC_ADAP_NORM_RESP_QUEUE][AAC_CONSUMER_INDEX]=
2037 AAC_ADAP_NORM_RESP_ENTRIES;
2038 sc->aac_queues->qt_qindex[AAC_ADAP_HIGH_RESP_QUEUE][AAC_PRODUCER_INDEX]=
2039 AAC_ADAP_HIGH_RESP_ENTRIES;
2040 sc->aac_queues->qt_qindex[AAC_ADAP_HIGH_RESP_QUEUE][AAC_CONSUMER_INDEX]=
2041 AAC_ADAP_HIGH_RESP_ENTRIES;
2042 sc->aac_qentries[AAC_HOST_NORM_CMD_QUEUE] =
2043 &sc->aac_queues->qt_HostNormCmdQueue[0];
2044 sc->aac_qentries[AAC_HOST_HIGH_CMD_QUEUE] =
2045 &sc->aac_queues->qt_HostHighCmdQueue[0];
2046 sc->aac_qentries[AAC_ADAP_NORM_CMD_QUEUE] =
2047 &sc->aac_queues->qt_AdapNormCmdQueue[0];
2048 sc->aac_qentries[AAC_ADAP_HIGH_CMD_QUEUE] =
2049 &sc->aac_queues->qt_AdapHighCmdQueue[0];
2050 sc->aac_qentries[AAC_HOST_NORM_RESP_QUEUE] =
2051 &sc->aac_queues->qt_HostNormRespQueue[0];
2052 sc->aac_qentries[AAC_HOST_HIGH_RESP_QUEUE] =
2053 &sc->aac_queues->qt_HostHighRespQueue[0];
2054 sc->aac_qentries[AAC_ADAP_NORM_RESP_QUEUE] =
2055 &sc->aac_queues->qt_AdapNormRespQueue[0];
2056 sc->aac_qentries[AAC_ADAP_HIGH_RESP_QUEUE] =
2057 &sc->aac_queues->qt_AdapHighRespQueue[0];
2060 * Do controller-type-specific initialisation
2062 switch (sc->aac_hwif) {
2063 case AAC_HWIF_I960RX:
2064 AAC_SETREG4(sc, AAC_RX_ODBR, ~0);
2067 AAC_SETREG4(sc, AAC_RKT_ODBR, ~0);
2074 * Give the init structure to the controller.
2076 if (aac_sync_command(sc, AAC_MONKER_INITSTRUCT,
2077 sc->aac_common_busaddr +
2078 offsetof(struct aac_common, ac_init), 0, 0, 0,
2080 device_printf(sc->aac_dev,
2081 "error establishing init structure\n");
2092 * Send a synchronous command to the controller and wait for a result.
2093 * Indicate if the controller completed the command with an error status.
2096 aac_sync_command(struct aac_softc *sc, u_int32_t command,
2097 u_int32_t arg0, u_int32_t arg1, u_int32_t arg2, u_int32_t arg3,
2105 /* populate the mailbox */
2106 AAC_SET_MAILBOX(sc, command, arg0, arg1, arg2, arg3);
2108 /* ensure the sync command doorbell flag is cleared */
2109 AAC_CLEAR_ISTATUS(sc, AAC_DB_SYNC_COMMAND);
2111 /* then set it to signal the adapter */
2112 AAC_QNOTIFY(sc, AAC_DB_SYNC_COMMAND);
2114 /* spin waiting for the command to complete */
2117 if (time_second > (then + AAC_IMMEDIATE_TIMEOUT)) {
2118 debug(1, "timed out");
2121 } while (!(AAC_GET_ISTATUS(sc) & AAC_DB_SYNC_COMMAND));
2123 /* clear the completion flag */
2124 AAC_CLEAR_ISTATUS(sc, AAC_DB_SYNC_COMMAND);
2126 /* get the command status */
2127 status = AAC_GET_MAILBOX(sc, 0);
2131 if (status != AAC_SRB_STS_SUCCESS)
2137 aac_sync_fib(struct aac_softc *sc, u_int32_t command, u_int32_t xferstate,
2138 struct aac_fib *fib, u_int16_t datasize)
2141 KKASSERT(lockstatus(&sc->aac_io_lock, curthread) != 0);
2143 if (datasize > AAC_FIB_DATASIZE)
2147 * Set up the sync FIB
2149 fib->Header.XferState = AAC_FIBSTATE_HOSTOWNED |
2150 AAC_FIBSTATE_INITIALISED |
2152 fib->Header.XferState |= xferstate;
2153 fib->Header.Command = command;
2154 fib->Header.StructType = AAC_FIBTYPE_TFIB;
2155 fib->Header.Size = sizeof(struct aac_fib) + datasize;
2156 fib->Header.SenderSize = sizeof(struct aac_fib);
2157 fib->Header.SenderFibAddress = 0; /* Not needed */
2158 fib->Header.ReceiverFibAddress = sc->aac_common_busaddr +
2159 offsetof(struct aac_common,
2163 * Give the FIB to the controller, wait for a response.
2165 if (aac_sync_command(sc, AAC_MONKER_SYNCFIB,
2166 fib->Header.ReceiverFibAddress, 0, 0, 0, NULL)) {
2167 debug(2, "IO error");
2175 * Adapter-space FIB queue manipulation
2177 * Note that the queue implementation here is a little funky; neither the PI or
2178 * CI will ever be zero. This behaviour is a controller feature.
2184 {AAC_HOST_NORM_CMD_ENTRIES, AAC_DB_COMMAND_NOT_FULL},
2185 {AAC_HOST_HIGH_CMD_ENTRIES, 0},
2186 {AAC_ADAP_NORM_CMD_ENTRIES, AAC_DB_COMMAND_READY},
2187 {AAC_ADAP_HIGH_CMD_ENTRIES, 0},
2188 {AAC_HOST_NORM_RESP_ENTRIES, AAC_DB_RESPONSE_NOT_FULL},
2189 {AAC_HOST_HIGH_RESP_ENTRIES, 0},
2190 {AAC_ADAP_NORM_RESP_ENTRIES, AAC_DB_RESPONSE_READY},
2191 {AAC_ADAP_HIGH_RESP_ENTRIES, 0}
2195 * Atomically insert an entry into the nominated queue, returns 0 on success or
2196 * EBUSY if the queue is full.
2198 * Note: it would be more efficient to defer notifying the controller in
2199 * the case where we may be inserting several entries in rapid succession,
2200 * but implementing this usefully may be difficult (it would involve a
2201 * separate queue/notify interface).
2204 aac_enqueue_fib(struct aac_softc *sc, int queue, struct aac_command *cm)
2213 fib_size = cm->cm_fib->Header.Size;
2214 fib_addr = cm->cm_fib->Header.ReceiverFibAddress;
2216 /* get the producer/consumer indices */
2217 pi = sc->aac_queues->qt_qindex[queue][AAC_PRODUCER_INDEX];
2218 ci = sc->aac_queues->qt_qindex[queue][AAC_CONSUMER_INDEX];
2220 /* wrap the queue? */
2221 if (pi >= aac_qinfo[queue].size)
2224 /* check for queue full */
2225 if ((pi + 1) == ci) {
2230 * To avoid a race with its completion interrupt, place this command on
2231 * the busy queue prior to advertising it to the controller.
2233 aac_enqueue_busy(cm);
2237 /* populate queue entry */
2238 (sc->aac_qentries[queue] + pi)->aq_fib_size = fib_size;
2239 (sc->aac_qentries[queue] + pi)->aq_fib_addr = fib_addr;
2241 /* update producer index */
2242 sc->aac_queues->qt_qindex[queue][AAC_PRODUCER_INDEX] = pi + 1;
2244 /* notify the adapter if we know how */
2245 if (aac_qinfo[queue].notify != 0)
2246 AAC_QNOTIFY(sc, aac_qinfo[queue].notify);
2255 * Atomically remove one entry from the nominated queue, returns 0 on
2256 * success or ENOENT if the queue is empty.
2259 aac_dequeue_fib(struct aac_softc *sc, int queue, u_int32_t *fib_size,
2260 struct aac_fib **fib_addr)
2263 u_int32_t fib_index;
2269 /* get the producer/consumer indices */
2270 pi = sc->aac_queues->qt_qindex[queue][AAC_PRODUCER_INDEX];
2271 ci = sc->aac_queues->qt_qindex[queue][AAC_CONSUMER_INDEX];
2273 /* check for queue empty */
2279 /* wrap the pi so the following test works */
2280 if (pi >= aac_qinfo[queue].size)
2287 /* wrap the queue? */
2288 if (ci >= aac_qinfo[queue].size)
2291 /* fetch the entry */
2292 *fib_size = (sc->aac_qentries[queue] + ci)->aq_fib_size;
2295 case AAC_HOST_NORM_CMD_QUEUE:
2296 case AAC_HOST_HIGH_CMD_QUEUE:
2298 * The aq_fib_addr is only 32 bits wide so it can't be counted
2299 * on to hold an address. For AIF's, the adapter assumes
2300 * that it's giving us an address into the array of AIF fibs.
2301 * Therefore, we have to convert it to an index.
2303 fib_index = (sc->aac_qentries[queue] + ci)->aq_fib_addr /
2304 sizeof(struct aac_fib);
2305 *fib_addr = &sc->aac_common->ac_fibs[fib_index];
2308 case AAC_HOST_NORM_RESP_QUEUE:
2309 case AAC_HOST_HIGH_RESP_QUEUE:
2311 struct aac_command *cm;
2314 * As above, an index is used instead of an actual address.
2315 * Gotta shift the index to account for the fast response
2316 * bit. No other correction is needed since this value was
2317 * originally provided by the driver via the SenderFibAddress
2320 fib_index = (sc->aac_qentries[queue] + ci)->aq_fib_addr;
2321 cm = sc->aac_commands + (fib_index >> 2);
2322 *fib_addr = cm->cm_fib;
2325 * Is this a fast response? If it is, update the fib fields in
2326 * local memory since the whole fib isn't DMA'd back up.
2328 if (fib_index & 0x01) {
2329 (*fib_addr)->Header.XferState |= AAC_FIBSTATE_DONEADAP;
2330 *((u_int32_t*)((*fib_addr)->data)) = AAC_ERROR_NORMAL;
2335 panic("Invalid queue in aac_dequeue_fib()");
2339 /* update consumer index */
2340 sc->aac_queues->qt_qindex[queue][AAC_CONSUMER_INDEX] = ci + 1;
2342 /* if we have made the queue un-full, notify the adapter */
2343 if (notify && (aac_qinfo[queue].notify != 0))
2344 AAC_QNOTIFY(sc, aac_qinfo[queue].notify);
2352 * Put our response to an Adapter Initialed Fib on the response queue
2355 aac_enqueue_response(struct aac_softc *sc, int queue, struct aac_fib *fib)
2364 /* Tell the adapter where the FIB is */
2365 fib_size = fib->Header.Size;
2366 fib_addr = fib->Header.SenderFibAddress;
2367 fib->Header.ReceiverFibAddress = fib_addr;
2369 /* get the producer/consumer indices */
2370 pi = sc->aac_queues->qt_qindex[queue][AAC_PRODUCER_INDEX];
2371 ci = sc->aac_queues->qt_qindex[queue][AAC_CONSUMER_INDEX];
2373 /* wrap the queue? */
2374 if (pi >= aac_qinfo[queue].size)
2377 /* check for queue full */
2378 if ((pi + 1) == ci) {
2383 /* populate queue entry */
2384 (sc->aac_qentries[queue] + pi)->aq_fib_size = fib_size;
2385 (sc->aac_qentries[queue] + pi)->aq_fib_addr = fib_addr;
2387 /* update producer index */
2388 sc->aac_queues->qt_qindex[queue][AAC_PRODUCER_INDEX] = pi + 1;
2390 /* notify the adapter if we know how */
2391 if (aac_qinfo[queue].notify != 0)
2392 AAC_QNOTIFY(sc, aac_qinfo[queue].notify);
2401 * Check for commands that have been outstanding for a suspiciously long time,
2402 * and complain about them.
2405 aac_timeout(void *xsc)
2407 struct aac_softc *sc = xsc;
2408 struct aac_command *cm;
2412 * Traverse the busy command list, bitch about late commands once
2416 deadline = time_second - AAC_CMD_TIMEOUT;
2417 TAILQ_FOREACH(cm, &sc->aac_busy, cm_link) {
2418 if ((cm->cm_timestamp < deadline)
2419 /* && !(cm->cm_flags & AAC_CMD_TIMEDOUT) */) {
2420 cm->cm_flags |= AAC_CMD_TIMEDOUT;
2421 device_printf(sc->aac_dev,
2422 "COMMAND %p TIMEOUT AFTER %d SECONDS\n",
2423 cm, (int)(time_second-cm->cm_timestamp));
2424 AAC_PRINT_FIB(sc, cm->cm_fib);
2429 code = AAC_GET_FWSTATUS(sc);
2430 if (code != AAC_UP_AND_RUNNING) {
2431 device_printf(sc->aac_dev, "WARNING! Controller is no "
2432 "longer running! code= 0x%x\n", code);
2439 * Interface Function Vectors
2443 * Read the current firmware status word.
2446 aac_sa_get_fwstatus(struct aac_softc *sc)
2450 return(AAC_GETREG4(sc, AAC_SA_FWSTATUS));
2454 aac_rx_get_fwstatus(struct aac_softc *sc)
2458 return(AAC_GETREG4(sc, AAC_RX_FWSTATUS));
2462 aac_fa_get_fwstatus(struct aac_softc *sc)
2468 val = AAC_GETREG4(sc, AAC_FA_FWSTATUS);
2473 aac_rkt_get_fwstatus(struct aac_softc *sc)
2477 return(AAC_GETREG4(sc, AAC_RKT_FWSTATUS));
2481 * Notify the controller of a change in a given queue
2485 aac_sa_qnotify(struct aac_softc *sc, int qbit)
2489 AAC_SETREG2(sc, AAC_SA_DOORBELL1_SET, qbit);
2493 aac_rx_qnotify(struct aac_softc *sc, int qbit)
2497 AAC_SETREG4(sc, AAC_RX_IDBR, qbit);
2501 aac_fa_qnotify(struct aac_softc *sc, int qbit)
2505 AAC_SETREG2(sc, AAC_FA_DOORBELL1, qbit);
2510 aac_rkt_qnotify(struct aac_softc *sc, int qbit)
2514 AAC_SETREG4(sc, AAC_RKT_IDBR, qbit);
2518 * Get the interrupt reason bits
2521 aac_sa_get_istatus(struct aac_softc *sc)
2525 return(AAC_GETREG2(sc, AAC_SA_DOORBELL0));
2529 aac_rx_get_istatus(struct aac_softc *sc)
2533 return(AAC_GETREG4(sc, AAC_RX_ODBR));
2537 aac_fa_get_istatus(struct aac_softc *sc)
2543 val = AAC_GETREG2(sc, AAC_FA_DOORBELL0);
2548 aac_rkt_get_istatus(struct aac_softc *sc)
2552 return(AAC_GETREG4(sc, AAC_RKT_ODBR));
2556 * Clear some interrupt reason bits
2559 aac_sa_clear_istatus(struct aac_softc *sc, int mask)
2563 AAC_SETREG2(sc, AAC_SA_DOORBELL0_CLEAR, mask);
2567 aac_rx_clear_istatus(struct aac_softc *sc, int mask)
2571 AAC_SETREG4(sc, AAC_RX_ODBR, mask);
2575 aac_fa_clear_istatus(struct aac_softc *sc, int mask)
2579 AAC_SETREG2(sc, AAC_FA_DOORBELL0_CLEAR, mask);
2584 aac_rkt_clear_istatus(struct aac_softc *sc, int mask)
2588 AAC_SETREG4(sc, AAC_RKT_ODBR, mask);
2592 * Populate the mailbox and set the command word
2595 aac_sa_set_mailbox(struct aac_softc *sc, u_int32_t command,
2596 u_int32_t arg0, u_int32_t arg1, u_int32_t arg2, u_int32_t arg3)
2600 AAC_SETREG4(sc, AAC_SA_MAILBOX, command);
2601 AAC_SETREG4(sc, AAC_SA_MAILBOX + 4, arg0);
2602 AAC_SETREG4(sc, AAC_SA_MAILBOX + 8, arg1);
2603 AAC_SETREG4(sc, AAC_SA_MAILBOX + 12, arg2);
2604 AAC_SETREG4(sc, AAC_SA_MAILBOX + 16, arg3);
2608 aac_rx_set_mailbox(struct aac_softc *sc, u_int32_t command,
2609 u_int32_t arg0, u_int32_t arg1, u_int32_t arg2, u_int32_t arg3)
2613 AAC_SETREG4(sc, AAC_RX_MAILBOX, command);
2614 AAC_SETREG4(sc, AAC_RX_MAILBOX + 4, arg0);
2615 AAC_SETREG4(sc, AAC_RX_MAILBOX + 8, arg1);
2616 AAC_SETREG4(sc, AAC_RX_MAILBOX + 12, arg2);
2617 AAC_SETREG4(sc, AAC_RX_MAILBOX + 16, arg3);
2621 aac_fa_set_mailbox(struct aac_softc *sc, u_int32_t command,
2622 u_int32_t arg0, u_int32_t arg1, u_int32_t arg2, u_int32_t arg3)
2626 AAC_SETREG4(sc, AAC_FA_MAILBOX, command);
2628 AAC_SETREG4(sc, AAC_FA_MAILBOX + 4, arg0);
2630 AAC_SETREG4(sc, AAC_FA_MAILBOX + 8, arg1);
2632 AAC_SETREG4(sc, AAC_FA_MAILBOX + 12, arg2);
2634 AAC_SETREG4(sc, AAC_FA_MAILBOX + 16, arg3);
2639 aac_rkt_set_mailbox(struct aac_softc *sc, u_int32_t command, u_int32_t arg0,
2640 u_int32_t arg1, u_int32_t arg2, u_int32_t arg3)
2644 AAC_SETREG4(sc, AAC_RKT_MAILBOX, command);
2645 AAC_SETREG4(sc, AAC_RKT_MAILBOX + 4, arg0);
2646 AAC_SETREG4(sc, AAC_RKT_MAILBOX + 8, arg1);
2647 AAC_SETREG4(sc, AAC_RKT_MAILBOX + 12, arg2);
2648 AAC_SETREG4(sc, AAC_RKT_MAILBOX + 16, arg3);
2652 * Fetch the immediate command status word
2655 aac_sa_get_mailbox(struct aac_softc *sc, int mb)
2659 return(AAC_GETREG4(sc, AAC_SA_MAILBOX + (mb * 4)));
2663 aac_rx_get_mailbox(struct aac_softc *sc, int mb)
2667 return(AAC_GETREG4(sc, AAC_RX_MAILBOX + (mb * 4)));
2671 aac_fa_get_mailbox(struct aac_softc *sc, int mb)
2677 val = AAC_GETREG4(sc, AAC_FA_MAILBOX + (mb * 4));
2682 aac_rkt_get_mailbox(struct aac_softc *sc, int mb)
2686 return(AAC_GETREG4(sc, AAC_RKT_MAILBOX + (mb * 4)));
2690 * Set/clear interrupt masks
2693 aac_sa_set_interrupts(struct aac_softc *sc, int enable)
2695 debug(2, "%sable interrupts", enable ? "en" : "dis");
2698 AAC_SETREG2((sc), AAC_SA_MASK0_CLEAR, AAC_DB_INTERRUPTS);
2700 AAC_SETREG2((sc), AAC_SA_MASK0_SET, ~0);
2705 aac_rx_set_interrupts(struct aac_softc *sc, int enable)
2707 debug(2, "%sable interrupts", enable ? "en" : "dis");
2710 if (sc->flags & AAC_FLAGS_NEW_COMM)
2711 AAC_SETREG4(sc, AAC_RX_OIMR, ~AAC_DB_INT_NEW_COMM);
2713 AAC_SETREG4(sc, AAC_RX_OIMR, ~AAC_DB_INTERRUPTS);
2715 AAC_SETREG4(sc, AAC_RX_OIMR, ~0);
2720 aac_fa_set_interrupts(struct aac_softc *sc, int enable)
2722 debug(2, "%sable interrupts", enable ? "en" : "dis");
2725 AAC_SETREG2((sc), AAC_FA_MASK0_CLEAR, AAC_DB_INTERRUPTS);
2728 AAC_SETREG2((sc), AAC_FA_MASK0, ~0);
2734 aac_rkt_set_interrupts(struct aac_softc *sc, int enable)
2736 debug(2, "%sable interrupts", enable ? "en" : "dis");
2739 if (sc->flags & AAC_FLAGS_NEW_COMM)
2740 AAC_SETREG4(sc, AAC_RKT_OIMR, ~AAC_DB_INT_NEW_COMM);
2742 AAC_SETREG4(sc, AAC_RKT_OIMR, ~AAC_DB_INTERRUPTS);
2744 AAC_SETREG4(sc, AAC_RKT_OIMR, ~0);
2749 * New comm. interface: Send command functions
2752 aac_rx_send_command(struct aac_softc *sc, struct aac_command *cm)
2754 u_int32_t index, device;
2756 debug(2, "send command (new comm.)");
2758 index = AAC_GETREG4(sc, AAC_RX_IQUE);
2759 if (index == 0xffffffffL)
2760 index = AAC_GETREG4(sc, AAC_RX_IQUE);
2761 if (index == 0xffffffffL)
2763 aac_enqueue_busy(cm);
2765 AAC_SETREG4(sc, device, (u_int32_t)(cm->cm_fibphys & 0xffffffffUL));
2767 AAC_SETREG4(sc, device, (u_int32_t)(cm->cm_fibphys >> 32));
2769 AAC_SETREG4(sc, device, cm->cm_fib->Header.Size);
2770 AAC_SETREG4(sc, AAC_RX_IQUE, index);
2775 aac_rkt_send_command(struct aac_softc *sc, struct aac_command *cm)
2777 u_int32_t index, device;
2779 debug(2, "send command (new comm.)");
2781 index = AAC_GETREG4(sc, AAC_RKT_IQUE);
2782 if (index == 0xffffffffL)
2783 index = AAC_GETREG4(sc, AAC_RKT_IQUE);
2784 if (index == 0xffffffffL)
2786 aac_enqueue_busy(cm);
2788 AAC_SETREG4(sc, device, (u_int32_t)(cm->cm_fibphys & 0xffffffffUL));
2790 AAC_SETREG4(sc, device, (u_int32_t)(cm->cm_fibphys >> 32));
2792 AAC_SETREG4(sc, device, cm->cm_fib->Header.Size);
2793 AAC_SETREG4(sc, AAC_RKT_IQUE, index);
2798 * New comm. interface: get, set outbound queue index
2801 aac_rx_get_outb_queue(struct aac_softc *sc)
2805 return(AAC_GETREG4(sc, AAC_RX_OQUE));
2809 aac_rkt_get_outb_queue(struct aac_softc *sc)
2813 return(AAC_GETREG4(sc, AAC_RKT_OQUE));
2817 aac_rx_set_outb_queue(struct aac_softc *sc, int index)
2821 AAC_SETREG4(sc, AAC_RX_OQUE, index);
2825 aac_rkt_set_outb_queue(struct aac_softc *sc, int index)
2829 AAC_SETREG4(sc, AAC_RKT_OQUE, index);
2833 * Debugging and Diagnostics
2837 * Print some information about the controller.
2840 aac_describe_controller(struct aac_softc *sc)
2842 struct aac_fib *fib;
2843 struct aac_adapter_info *info;
2847 AAC_LOCK_ACQUIRE(&sc->aac_io_lock);
2848 aac_alloc_sync_fib(sc, &fib);
2851 if (aac_sync_fib(sc, RequestAdapterInfo, 0, fib, 1)) {
2852 device_printf(sc->aac_dev, "RequestAdapterInfo failed\n");
2853 aac_release_sync_fib(sc);
2854 AAC_LOCK_RELEASE(&sc->aac_io_lock);
2858 /* save the kernel revision structure for later use */
2859 info = (struct aac_adapter_info *)&fib->data[0];
2860 sc->aac_revision = info->KernelRevision;
2862 device_printf(sc->aac_dev, "Adaptec Raid Controller %d.%d.%d-%d\n",
2863 AAC_DRIVER_VERSION >> 24,
2864 (AAC_DRIVER_VERSION >> 16) & 0xFF,
2865 AAC_DRIVER_VERSION & 0xFF,
2869 device_printf(sc->aac_dev, "%s %dMHz, %dMB memory "
2870 "(%dMB cache, %dMB execution), %s\n",
2871 aac_describe_code(aac_cpu_variant, info->CpuVariant),
2872 info->ClockSpeed, info->TotalMem / (1024 * 1024),
2873 info->BufferMem / (1024 * 1024),
2874 info->ExecutionMem / (1024 * 1024),
2875 aac_describe_code(aac_battery_platform,
2876 info->batteryPlatform));
2878 device_printf(sc->aac_dev,
2879 "Kernel %d.%d-%d, Build %d, S/N %6X\n",
2880 info->KernelRevision.external.comp.major,
2881 info->KernelRevision.external.comp.minor,
2882 info->KernelRevision.external.comp.dash,
2883 info->KernelRevision.buildNumber,
2884 (u_int32_t)(info->SerialNumber & 0xffffff));
2886 device_printf(sc->aac_dev, "Supported Options=%b\n",
2887 sc->supported_options,
2909 aac_release_sync_fib(sc);
2910 AAC_LOCK_RELEASE(&sc->aac_io_lock);
2914 * Look up a text description of a numeric error code and return a pointer to
2918 aac_describe_code(struct aac_code_lookup *table, u_int32_t code)
2922 for (i = 0; table[i].string != NULL; i++)
2923 if (table[i].code == code)
2924 return(table[i].string);
2925 return(table[i + 1].string);
2929 * Management Interface
2933 aac_open(struct dev_open_args *ap)
2935 cdev_t dev = ap->a_head.a_dev;
2936 struct aac_softc *sc;
2942 /* Check to make sure the device isn't already open */
2943 if (sc->aac_state & AAC_STATE_OPEN) {
2946 sc->aac_state |= AAC_STATE_OPEN;
2952 aac_close(struct dev_close_args *ap)
2954 cdev_t dev = ap->a_head.a_dev;
2955 struct aac_softc *sc;
2961 /* Mark this unit as no longer open */
2962 sc->aac_state &= ~AAC_STATE_OPEN;
2968 aac_ioctl(struct dev_ioctl_args *ap)
2970 cdev_t dev = ap->a_head.a_dev;
2971 caddr_t arg = ap->a_data;
2972 struct aac_softc *sc = dev->si_drv1;
2978 if (ap->a_cmd == AACIO_STATS) {
2979 union aac_statrequest *as = (union aac_statrequest *)arg;
2981 switch (as->as_item) {
2987 bcopy(&sc->aac_qstat[as->as_item], &as->as_qstat,
2988 sizeof(struct aac_qstat));
2997 arg = *(caddr_t *)arg;
2999 switch (ap->a_cmd) {
3000 /* AACIO_STATS already handled above */
3001 case FSACTL_SENDFIB:
3002 debug(1, "FSACTL_SENDFIB");
3003 error = aac_ioctl_sendfib(sc, arg);
3005 case FSACTL_AIF_THREAD:
3006 debug(1, "FSACTL_AIF_THREAD");
3009 case FSACTL_OPEN_GET_ADAPTER_FIB:
3010 debug(1, "FSACTL_OPEN_GET_ADAPTER_FIB");
3012 * Pass the caller out an AdapterFibContext.
3014 * Note that because we only support one opener, we
3015 * basically ignore this. Set the caller's context to a magic
3016 * number just in case.
3018 * The Linux code hands the driver a pointer into kernel space,
3019 * and then trusts it when the caller hands it back. Aiee!
3020 * Here, we give it the proc pointer of the per-adapter aif
3021 * thread. It's only used as a sanity check in other calls.
3023 cookie = (uint32_t)(uintptr_t)sc->aifthread;
3024 error = copyout(&cookie, arg, sizeof(cookie));
3026 case FSACTL_GET_NEXT_ADAPTER_FIB:
3027 debug(1, "FSACTL_GET_NEXT_ADAPTER_FIB");
3028 error = aac_getnext_aif(sc, arg);
3030 case FSACTL_CLOSE_GET_ADAPTER_FIB:
3031 debug(1, "FSACTL_CLOSE_GET_ADAPTER_FIB");
3032 /* don't do anything here */
3034 case FSACTL_MINIPORT_REV_CHECK:
3035 debug(1, "FSACTL_MINIPORT_REV_CHECK");
3036 error = aac_rev_check(sc, arg);
3038 case FSACTL_QUERY_DISK:
3039 debug(1, "FSACTL_QUERY_DISK");
3040 error = aac_query_disk(sc, arg);
3042 case FSACTL_DELETE_DISK:
3044 * We don't trust the underland to tell us when to delete a
3045 * container, rather we rely on an AIF coming from the
3050 case FSACTL_GET_PCI_INFO:
3051 arg = *(caddr_t*)arg;
3052 case FSACTL_LNX_GET_PCI_INFO:
3053 debug(1, "FSACTL_GET_PCI_INFO");
3054 error = aac_get_pci_info(sc, arg);
3057 debug(1, "unsupported cmd 0x%lx\n", ap->a_cmd);
3064 static struct filterops aac_filterops =
3065 { FILTEROP_ISFD, NULL, aac_filter_detach, aac_filter };
3068 aac_kqfilter(struct dev_kqfilter_args *ap)
3070 cdev_t dev = ap->a_head.a_dev;
3071 struct aac_softc *sc = dev->si_drv1;
3072 struct knote *kn = ap->a_kn;
3073 struct klist *klist;
3077 switch (kn->kn_filter) {
3079 kn->kn_fop = &aac_filterops;
3080 kn->kn_hook = (caddr_t)sc;
3083 ap->a_result = EOPNOTSUPP;
3087 klist = &sc->rcv_kq.ki_note;
3088 knote_insert(klist, kn);
3094 aac_filter_detach(struct knote *kn)
3096 struct aac_softc *sc = (struct aac_softc *)kn->kn_hook;
3097 struct klist *klist;
3099 klist = &sc->rcv_kq.ki_note;
3100 knote_remove(klist, kn);
3104 aac_filter(struct knote *kn, long hint)
3106 struct aac_softc *sc = (struct aac_softc *)kn->kn_hook;
3109 AAC_LOCK_ACQUIRE(&sc->aac_aifq_lock);
3110 if (sc->aac_aifq_tail != sc->aac_aifq_head)
3112 AAC_LOCK_RELEASE(&sc->aac_aifq_lock);
3119 aac_ioctl_event(struct aac_softc *sc, struct aac_event *event, void *arg)
3122 switch (event->ev_type) {
3123 case AAC_EVENT_CMFREE:
3124 AAC_LOCK_ACQUIRE(&sc->aac_io_lock);
3125 if (aac_alloc_command(sc, (struct aac_command **)arg)) {
3126 aac_add_event(sc, event);
3127 AAC_LOCK_RELEASE(&sc->aac_io_lock);
3130 kfree(event, M_AACBUF);
3132 AAC_LOCK_RELEASE(&sc->aac_io_lock);
3140 * Send a FIB supplied from userspace
3143 aac_ioctl_sendfib(struct aac_softc *sc, caddr_t ufib)
3145 struct aac_command *cm;
3155 AAC_LOCK_ACQUIRE(&sc->aac_io_lock);
3156 if (aac_alloc_command(sc, &cm)) {
3157 struct aac_event *event;
3159 event = kmalloc(sizeof(struct aac_event), M_AACBUF,
3160 M_INTWAIT | M_ZERO);
3161 event->ev_type = AAC_EVENT_CMFREE;
3162 event->ev_callback = aac_ioctl_event;
3163 event->ev_arg = &cm;
3164 aac_add_event(sc, event);
3165 tsleep_interlock(&cm, 0);
3166 AAC_LOCK_RELEASE(&sc->aac_io_lock);
3167 tsleep(&cm, PINTERLOCKED, "sendfib", 0);
3168 AAC_LOCK_ACQUIRE(&sc->aac_io_lock);
3170 AAC_LOCK_RELEASE(&sc->aac_io_lock);
3173 * Fetch the FIB header, then re-copy to get data as well.
3175 if ((error = copyin(ufib, cm->cm_fib,
3176 sizeof(struct aac_fib_header))) != 0)
3178 size = cm->cm_fib->Header.Size + sizeof(struct aac_fib_header);
3179 if (size > sizeof(struct aac_fib)) {
3180 device_printf(sc->aac_dev, "incoming FIB oversized (%d > %zd)\n",
3181 size, sizeof(struct aac_fib));
3182 size = sizeof(struct aac_fib);
3184 if ((error = copyin(ufib, cm->cm_fib, size)) != 0)
3186 cm->cm_fib->Header.Size = size;
3187 cm->cm_timestamp = time_second;
3190 * Pass the FIB to the controller, wait for it to complete.
3192 AAC_LOCK_ACQUIRE(&sc->aac_io_lock);
3193 if ((error = aac_wait_command(cm)) != 0) {
3194 device_printf(sc->aac_dev,
3195 "aac_wait_command return %d\n", error);
3198 AAC_LOCK_RELEASE(&sc->aac_io_lock);
3201 * Copy the FIB and data back out to the caller.
3203 size = cm->cm_fib->Header.Size;
3204 if (size > sizeof(struct aac_fib)) {
3205 device_printf(sc->aac_dev, "outbound FIB oversized (%d > %zd)\n",
3206 size, sizeof(struct aac_fib));
3207 size = sizeof(struct aac_fib);
3209 error = copyout(cm->cm_fib, ufib, size);
3210 AAC_LOCK_ACQUIRE(&sc->aac_io_lock);
3214 aac_release_command(cm);
3217 AAC_LOCK_RELEASE(&sc->aac_io_lock);
3222 * Handle an AIF sent to us by the controller; queue it for later reference.
3223 * If the queue fills up, then drop the older entries.
3226 aac_handle_aif(struct aac_softc *sc, struct aac_fib *fib)
3228 struct aac_aif_command *aif;
3229 struct aac_container *co, *co_next;
3230 struct aac_mntinfo *mi;
3231 struct aac_mntinforesp *mir = NULL;
3234 int count = 0, added = 0, i = 0;
3238 aif = (struct aac_aif_command*)&fib->data[0];
3239 aac_print_aif(sc, aif);
3241 /* Is it an event that we should care about? */
3242 switch (aif->command) {
3243 case AifCmdEventNotify:
3244 switch (aif->data.EN.type) {
3245 case AifEnAddContainer:
3246 case AifEnDeleteContainer:
3248 * A container was added or deleted, but the message
3249 * doesn't tell us anything else! Re-enumerate the
3250 * containers and sort things out.
3252 aac_alloc_sync_fib(sc, &fib);
3253 mi = (struct aac_mntinfo *)&fib->data[0];
3256 * Ask the controller for its containers one at
3258 * XXX What if the controller's list changes
3259 * midway through this enumaration?
3260 * XXX This should be done async.
3262 bzero(mi, sizeof(struct aac_mntinfo));
3263 mi->Command = VM_NameServe;
3264 mi->MntType = FT_FILESYS;
3266 rsize = sizeof(mir);
3267 if (aac_sync_fib(sc, ContainerCommand, 0, fib,
3268 sizeof(struct aac_mntinfo))) {
3269 device_printf(sc->aac_dev,
3270 "Error probing container %d\n", i);
3274 mir = (struct aac_mntinforesp *)&fib->data[0];
3275 /* XXX Need to check if count changed */
3276 count = mir->MntRespCount;
3279 * Check the container against our list.
3280 * co->co_found was already set to 0 in a
3283 if ((mir->Status == ST_OK) &&
3284 (mir->MntTable[0].VolType != CT_NONE)) {
3287 &sc->aac_container_tqh,
3289 if (co->co_mntobj.ObjectId ==
3290 mir->MntTable[0].ObjectId) {
3297 * If the container matched, continue
3306 * This is a new container. Do all the
3307 * appropriate things to set it up.
3309 aac_add_container(sc, mir, 1);
3313 } while ((i < count) && (i < AAC_MAX_CONTAINERS));
3314 aac_release_sync_fib(sc);
3317 * Go through our list of containers and see which ones
3318 * were not marked 'found'. Since the controller didn't
3319 * list them they must have been deleted. Do the
3320 * appropriate steps to destroy the device. Also reset
3321 * the co->co_found field.
3323 co = TAILQ_FIRST(&sc->aac_container_tqh);
3324 while (co != NULL) {
3325 if (co->co_found == 0) {
3326 AAC_LOCK_RELEASE(&sc->aac_io_lock);
3328 device_delete_child(sc->aac_dev,
3331 AAC_LOCK_ACQUIRE(&sc->aac_io_lock);
3332 co_next = TAILQ_NEXT(co, co_link);
3333 AAC_LOCK_ACQUIRE(&sc->
3334 aac_container_lock);
3335 TAILQ_REMOVE(&sc->aac_container_tqh, co,
3337 AAC_LOCK_RELEASE(&sc->
3338 aac_container_lock);
3339 kfree(co, M_AACBUF);
3343 co = TAILQ_NEXT(co, co_link);
3347 /* Attach the newly created containers */
3349 AAC_LOCK_RELEASE(&sc->aac_io_lock);
3351 bus_generic_attach(sc->aac_dev);
3353 AAC_LOCK_ACQUIRE(&sc->aac_io_lock);
3366 /* Copy the AIF data to the AIF queue for ioctl retrieval */
3367 AAC_LOCK_ACQUIRE(&sc->aac_aifq_lock);
3368 next = (sc->aac_aifq_head + 1) % AAC_AIFQ_LENGTH;
3369 if (next != sc->aac_aifq_tail) {
3370 bcopy(aif, &sc->aac_aifq[next], sizeof(struct aac_aif_command));
3371 sc->aac_aifq_head = next;
3373 /* On the off chance that someone is sleeping for an aif... */
3374 if (sc->aac_state & AAC_STATE_AIF_SLEEPER)
3375 wakeup(sc->aac_aifq);
3376 /* token may have been lost */
3377 /* Wakeup any poll()ers */
3378 KNOTE(&sc->rcv_kq.ki_note, 0);
3379 /* token may have been lost */
3381 AAC_LOCK_RELEASE(&sc->aac_aifq_lock);
3387 * Return the Revision of the driver to userspace and check to see if the
3388 * userspace app is possibly compatible. This is extremely bogus since
3389 * our driver doesn't follow Adaptec's versioning system. Cheat by just
3390 * returning what the card reported.
3393 aac_rev_check(struct aac_softc *sc, caddr_t udata)
3395 struct aac_rev_check rev_check;
3396 struct aac_rev_check_resp rev_check_resp;
3402 * Copyin the revision struct from userspace
3404 if ((error = copyin(udata, (caddr_t)&rev_check,
3405 sizeof(struct aac_rev_check))) != 0) {
3409 debug(2, "Userland revision= %d\n",
3410 rev_check.callingRevision.buildNumber);
3413 * Doctor up the response struct.
3415 rev_check_resp.possiblyCompatible = 1;
3416 rev_check_resp.adapterSWRevision.external.ul =
3417 sc->aac_revision.external.ul;
3418 rev_check_resp.adapterSWRevision.buildNumber =
3419 sc->aac_revision.buildNumber;
3421 return(copyout((caddr_t)&rev_check_resp, udata,
3422 sizeof(struct aac_rev_check_resp)));
3426 * Pass the caller the next AIF in their queue
3429 aac_getnext_aif(struct aac_softc *sc, caddr_t arg)
3431 struct get_adapter_fib_ioctl agf;
3436 if ((error = copyin(arg, &agf, sizeof(agf))) == 0) {
3439 * Check the magic number that we gave the caller.
3441 if (agf.AdapterFibContext != (int)(uintptr_t)sc->aifthread) {
3445 error = aac_return_aif(sc, agf.AifFib);
3447 if ((error == EAGAIN) && (agf.Wait)) {
3448 sc->aac_state |= AAC_STATE_AIF_SLEEPER;
3449 while (error == EAGAIN) {
3450 error = tsleep(sc->aac_aifq,
3451 PCATCH, "aacaif", 0);
3453 error = aac_return_aif(sc,
3456 sc->aac_state &= ~AAC_STATE_AIF_SLEEPER;
3464 * Hand the next AIF off the top of the queue out to userspace.
3466 * YYY token could be lost during copyout
3469 aac_return_aif(struct aac_softc *sc, caddr_t uptr)
3475 AAC_LOCK_ACQUIRE(&sc->aac_aifq_lock);
3476 if (sc->aac_aifq_tail == sc->aac_aifq_head) {
3477 AAC_LOCK_RELEASE(&sc->aac_aifq_lock);
3481 next = (sc->aac_aifq_tail + 1) % AAC_AIFQ_LENGTH;
3482 error = copyout(&sc->aac_aifq[next], uptr,
3483 sizeof(struct aac_aif_command));
3485 device_printf(sc->aac_dev,
3486 "aac_return_aif: copyout returned %d\n", error);
3488 sc->aac_aifq_tail = next;
3490 AAC_LOCK_RELEASE(&sc->aac_aifq_lock);
3495 aac_get_pci_info(struct aac_softc *sc, caddr_t uptr)
3497 struct aac_pci_info {
3505 pciinf.bus = pci_get_bus(sc->aac_dev);
3506 pciinf.slot = pci_get_slot(sc->aac_dev);
3508 error = copyout((caddr_t)&pciinf, uptr,
3509 sizeof(struct aac_pci_info));
3515 * Give the userland some information about the container. The AAC arch
3516 * expects the driver to be a SCSI passthrough type driver, so it expects
3517 * the containers to have b:t:l numbers. Fake it.
3520 aac_query_disk(struct aac_softc *sc, caddr_t uptr)
3522 struct aac_query_disk query_disk;
3523 struct aac_container *co;
3524 struct aac_disk *disk;
3531 error = copyin(uptr, (caddr_t)&query_disk,
3532 sizeof(struct aac_query_disk));
3536 id = query_disk.ContainerNumber;
3540 AAC_LOCK_ACQUIRE(&sc->aac_container_lock);
3541 TAILQ_FOREACH(co, &sc->aac_container_tqh, co_link) {
3542 if (co->co_mntobj.ObjectId == id)
3547 query_disk.Valid = 0;
3548 query_disk.Locked = 0;
3549 query_disk.Deleted = 1; /* XXX is this right? */
3551 disk = device_get_softc(co->co_disk);
3552 query_disk.Valid = 1;
3554 (disk->ad_flags & AAC_DISK_OPEN) ? 1 : 0;
3555 query_disk.Deleted = 0;
3556 query_disk.Bus = device_get_unit(sc->aac_dev);
3557 query_disk.Target = disk->unit;
3559 query_disk.UnMapped = 0;
3560 bcopy(disk->ad_dev_t->si_name,
3561 &query_disk.diskDeviceName[0], 10);
3563 AAC_LOCK_RELEASE(&sc->aac_container_lock);
3565 error = copyout((caddr_t)&query_disk, uptr,
3566 sizeof(struct aac_query_disk));
3572 aac_get_bus_info(struct aac_softc *sc)
3574 struct aac_fib *fib;
3575 struct aac_ctcfg *c_cmd;
3576 struct aac_ctcfg_resp *c_resp;
3577 struct aac_vmioctl *vmi;
3578 struct aac_vmi_businf_resp *vmi_resp;
3579 struct aac_getbusinf businfo;
3580 struct aac_sim *caminf;
3582 int i, found, error;
3584 AAC_LOCK_ACQUIRE(&sc->aac_io_lock);
3585 aac_alloc_sync_fib(sc, &fib);
3586 c_cmd = (struct aac_ctcfg *)&fib->data[0];
3587 bzero(c_cmd, sizeof(struct aac_ctcfg));
3589 c_cmd->Command = VM_ContainerConfig;
3590 c_cmd->cmd = CT_GET_SCSI_METHOD;
3593 error = aac_sync_fib(sc, ContainerCommand, 0, fib,
3594 sizeof(struct aac_ctcfg));
3596 device_printf(sc->aac_dev, "Error %d sending "
3597 "VM_ContainerConfig command\n", error);
3598 aac_release_sync_fib(sc);
3599 AAC_LOCK_RELEASE(&sc->aac_io_lock);
3603 c_resp = (struct aac_ctcfg_resp *)&fib->data[0];
3604 if (c_resp->Status != ST_OK) {
3605 device_printf(sc->aac_dev, "VM_ContainerConfig returned 0x%x\n",
3607 aac_release_sync_fib(sc);
3608 AAC_LOCK_RELEASE(&sc->aac_io_lock);
3612 sc->scsi_method_id = c_resp->param;
3614 vmi = (struct aac_vmioctl *)&fib->data[0];
3615 bzero(vmi, sizeof(struct aac_vmioctl));
3617 vmi->Command = VM_Ioctl;
3618 vmi->ObjType = FT_DRIVE;
3619 vmi->MethId = sc->scsi_method_id;
3621 vmi->IoctlCmd = GetBusInfo;
3623 error = aac_sync_fib(sc, ContainerCommand, 0, fib,
3624 sizeof(struct aac_vmioctl));
3626 device_printf(sc->aac_dev, "Error %d sending VMIoctl command\n",
3628 aac_release_sync_fib(sc);
3629 AAC_LOCK_RELEASE(&sc->aac_io_lock);
3633 vmi_resp = (struct aac_vmi_businf_resp *)&fib->data[0];
3634 if (vmi_resp->Status != ST_OK) {
3635 debug(1, "VM_Ioctl returned %d\n", vmi_resp->Status);
3636 aac_release_sync_fib(sc);
3637 AAC_LOCK_RELEASE(&sc->aac_io_lock);
3641 bcopy(&vmi_resp->BusInf, &businfo, sizeof(struct aac_getbusinf));
3642 aac_release_sync_fib(sc);
3643 AAC_LOCK_RELEASE(&sc->aac_io_lock);
3646 for (i = 0; i < businfo.BusCount; i++) {
3647 if (businfo.BusValid[i] != AAC_BUS_VALID)
3650 caminf = (struct aac_sim *)kmalloc(sizeof(struct aac_sim),
3651 M_AACBUF, M_INTWAIT | M_ZERO);
3653 child = device_add_child(sc->aac_dev, "aacp", -1);
3654 if (child == NULL) {
3655 device_printf(sc->aac_dev,
3656 "device_add_child failed for passthrough bus %d\n",
3658 kfree(caminf, M_AACBUF);
3662 caminf->TargetsPerBus = businfo.TargetsPerBus;
3663 caminf->BusNumber = i;
3664 caminf->InitiatorBusId = businfo.InitiatorBusId[i];
3665 caminf->aac_sc = sc;
3666 caminf->sim_dev = child;
3668 device_set_ivars(child, caminf);
3669 device_set_desc(child, "SCSI Passthrough Bus");
3670 TAILQ_INSERT_TAIL(&sc->aac_sim_tqh, caminf, sim_link);
3676 bus_generic_attach(sc->aac_dev);