2 * Copyright (c) 1998 - 2006 Søren Schmidt <sos@FreeBSD.org>
5 * Redistribution and use in source and binary forms, with or without
6 * modification, are permitted provided that the following conditions
8 * 1. Redistributions of source code must retain the above copyright
9 * notice, this list of conditions and the following disclaimer,
10 * without modification, immediately at the beginning of the file.
11 * 2. Redistributions in binary form must reproduce the above copyright
12 * notice, this list of conditions and the following disclaimer in the
13 * documentation and/or other materials provided with the distribution.
15 * THIS SOFTWARE IS PROVIDED BY THE AUTHOR ``AS IS'' AND ANY EXPRESS OR
16 * IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES
17 * OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED.
18 * IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR ANY DIRECT, INDIRECT,
19 * INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT
20 * NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
21 * DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
22 * THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
23 * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF
24 * THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
26 * $FreeBSD: src/sys/dev/ata/ata-queue.c,v 1.65 2006/07/21 19:13:05 imp Exp $
27 * $DragonFly: src/sys/dev/disk/nata/ata-queue.c,v 1.4 2006/12/22 23:26:16 swildner Exp $
32 #include <sys/param.h>
34 #include <sys/callout.h>
36 #include <sys/queue.h>
37 #include <sys/spinlock2.h>
38 #include <sys/systm.h>
39 #include <sys/taskqueue.h>
45 static void ata_completed(void *, int);
46 static void ata_sort_queue(struct ata_channel *ch, struct ata_request *request);
47 static char *ata_skey2str(u_int8_t);
50 ata_queue_request(struct ata_request *request)
52 struct ata_channel *ch;
54 /* treat request as virgin (this might be an ATA_R_REQUEUE) */
55 request->result = request->status = request->error = 0;
57 /* check that that the device is still valid */
58 if (!(request->parent = device_get_parent(request->dev))) {
59 request->result = ENXIO;
60 if (request->callback)
61 (request->callback)(request);
64 ch = device_get_softc(request->parent);
65 callout_init_mp(&request->callout); /* serialization done via state_mtx */
66 if (!request->callback && !(request->flags & ATA_R_REQUEUE))
67 spin_init(&request->done);
69 /* in ATA_STALL_QUEUE state we call HW directly */
70 if ((ch->state & ATA_STALL_QUEUE) && (request->flags & ATA_R_CONTROL)) {
71 spin_lock_wr(&ch->state_mtx);
72 ch->running = request;
73 if (ch->hw.begin_transaction(request) == ATA_OP_FINISHED) {
75 if (!request->callback)
76 spin_uninit(&request->done);
77 spin_unlock_wr(&ch->state_mtx);
80 spin_unlock_wr(&ch->state_mtx);
82 /* otherwise put request on the locked queue at the specified location */
84 spin_lock_wr(&ch->queue_mtx);
85 if (request->flags & ATA_R_AT_HEAD)
86 TAILQ_INSERT_HEAD(&ch->ata_queue, request, chain);
87 else if (request->flags & ATA_R_ORDERED)
88 ata_sort_queue(ch, request);
90 TAILQ_INSERT_TAIL(&ch->ata_queue, request, chain);
91 spin_unlock_wr(&ch->queue_mtx);
92 ATA_DEBUG_RQ(request, "queued");
96 /* if this is a requeued request callback/sleep we're done */
97 if (request->flags & ATA_R_REQUEUE)
100 /* if this is not a callback wait until request is completed */
101 if (!request->callback) {
102 ATA_DEBUG_RQ(request, "wait for completion");
104 /* interlock against wakeup */
105 spin_lock_wr(&request->done);
106 /* check if the request was completed already */
107 if (!(request->flags & ATA_R_COMPLETED))
108 msleep(request, &request->done, 0, "ATA request completion "
109 "wait", request->timeout * hz * 4);
110 spin_unlock_wr(&request->done);
111 /* check if the request was completed while sleeping */
112 if (!(request->flags & ATA_R_COMPLETED)) {
114 device_printf(request->dev, "WARNING - %s taskqueue timeout - "
115 "completing request directly\n",
116 ata_cmd2str(request));
117 request->flags |= ATA_R_DANGER1;
118 ata_completed(request, 0);
121 spin_uninit(&request->done);
126 ata_controlcmd(device_t dev, u_int8_t command, u_int16_t feature,
127 u_int64_t lba, u_int16_t count)
129 struct ata_request *request = ata_alloc_request();
134 request->u.ata.command = command;
135 request->u.ata.lba = lba;
136 request->u.ata.count = count;
137 request->u.ata.feature = feature;
138 request->flags = ATA_R_CONTROL;
139 request->timeout = 1;
140 request->retries = 0;
141 ata_queue_request(request);
142 error = request->result;
143 ata_free_request(request);
149 ata_atapicmd(device_t dev, u_int8_t *ccb, caddr_t data,
150 int count, int flags, int timeout)
152 struct ata_request *request = ata_alloc_request();
153 struct ata_device *atadev = device_get_softc(dev);
158 if ((atadev->param.config & ATA_PROTO_MASK) == ATA_PROTO_ATAPI_12)
159 bcopy(ccb, request->u.atapi.ccb, 12);
161 bcopy(ccb, request->u.atapi.ccb, 16);
162 request->data = data;
163 request->bytecount = count;
164 request->transfersize = min(request->bytecount, 65534);
165 request->flags = flags | ATA_R_ATAPI;
166 request->timeout = timeout;
167 request->retries = 0;
168 ata_queue_request(request);
169 error = request->result;
170 ata_free_request(request);
176 ata_start(device_t dev)
178 struct ata_channel *ch = device_get_softc(dev);
179 struct ata_request *request;
180 struct ata_composite *cptr;
181 int dependencies = 0;
183 /* if we have a request on the queue try to get it running */
184 spin_lock_wr(&ch->queue_mtx);
185 if ((request = TAILQ_FIRST(&ch->ata_queue))) {
187 /* we need the locking function to get the lock for this channel */
188 if (ATA_LOCKING(dev, ATA_LF_LOCK) == ch->unit) {
190 /* check for composite dependencies */
191 if ((cptr = request->composite)) {
192 spin_lock_wr(&cptr->lock);
193 if ((request->flags & ATA_R_WRITE) &&
194 (cptr->wr_depend & cptr->rd_done) != cptr->wr_depend) {
197 spin_unlock_wr(&cptr->lock);
200 /* check we are in the right state and has no dependencies */
201 spin_lock_wr(&ch->state_mtx);
202 if (ch->state == ATA_IDLE && !dependencies) {
203 ATA_DEBUG_RQ(request, "starting");
204 TAILQ_REMOVE(&ch->ata_queue, request, chain);
205 ch->running = request;
206 ch->state = ATA_ACTIVE;
208 /* if we are the freezing point release it */
209 if (ch->freezepoint == request)
210 ch->freezepoint = NULL;
212 if (ch->hw.begin_transaction(request) == ATA_OP_FINISHED) {
214 ch->state = ATA_IDLE;
215 spin_unlock_wr(&ch->state_mtx);
216 spin_unlock_wr(&ch->queue_mtx);
217 ATA_LOCKING(dev, ATA_LF_UNLOCK);
222 spin_unlock_wr(&ch->state_mtx);
223 spin_unlock_wr(&ch->queue_mtx);
224 while (!ata_interrupt(ch))
229 spin_unlock_wr(&ch->state_mtx);
232 spin_unlock_wr(&ch->queue_mtx);
236 ata_finish(struct ata_request *request)
238 struct ata_channel *ch = device_get_softc(request->parent);
241 * if in ATA_STALL_QUEUE state or request has ATA_R_DIRECT flags set
242 * we need to call ata_complete() directly here (no taskqueue involvement)
245 (ch->state & ATA_STALL_QUEUE) || (request->flags & ATA_R_DIRECT)) {
246 ATA_DEBUG_RQ(request, "finish directly");
247 ata_completed(request, 0);
250 /* put request on the proper taskqueue for completition */
251 TASK_INIT(&request->task, 0, ata_completed, request);
252 ATA_DEBUG_RQ(request, "finish taskqueue_swi");
253 taskqueue_enqueue(taskqueue_swi, &request->task);
258 ata_completed(void *context, int dummy)
260 struct ata_request *request = (struct ata_request *)context;
261 struct ata_channel *ch = device_get_softc(request->parent);
262 struct ata_device *atadev = device_get_softc(request->dev);
263 struct ata_composite *composite;
265 if (request->flags & ATA_R_DANGER2) {
266 device_printf(request->dev,
267 "WARNING - %s freeing taskqueue zombie request\n",
268 ata_cmd2str(request));
269 request->flags &= ~(ATA_R_DANGER1 | ATA_R_DANGER2);
270 ata_free_request(request);
273 if (request->flags & ATA_R_DANGER1)
274 request->flags |= ATA_R_DANGER2;
276 ATA_DEBUG_RQ(request, "completed entered");
278 /* if we had a timeout, reinit channel and deal with the falldown */
279 if (request->flags & ATA_R_TIMEOUT) {
281 * if the channel is still present and
282 * reinit succeeds and
283 * the device doesn't get detached and
284 * there are retries left we reinject this request
286 if (ch && !ata_reinit(ch->dev) && !request->result &&
287 (request->retries-- > 0)) {
288 if (!(request->flags & ATA_R_QUIET)) {
289 device_printf(request->dev,
290 "TIMEOUT - %s retrying (%d retr%s left)",
291 ata_cmd2str(request), request->retries,
292 request->retries == 1 ? "y" : "ies");
293 if (!(request->flags & (ATA_R_ATAPI | ATA_R_CONTROL)))
294 kprintf(" LBA=%ju", request->u.ata.lba);
297 request->flags &= ~(ATA_R_TIMEOUT | ATA_R_DEBUG);
298 request->flags |= (ATA_R_AT_HEAD | ATA_R_REQUEUE);
299 ATA_DEBUG_RQ(request, "completed reinject");
300 ata_queue_request(request);
304 /* ran out of good intentions so finish with error */
305 if (!request->result) {
306 if (!(request->flags & ATA_R_QUIET)) {
308 device_printf(request->dev, "FAILURE - %s timed out",
309 ata_cmd2str(request));
310 if (!(request->flags & (ATA_R_ATAPI | ATA_R_CONTROL)))
311 kprintf(" LBA=%ju", request->u.ata.lba);
315 request->result = EIO;
318 else if (!(request->flags & ATA_R_ATAPI) ){
319 /* if this is a soft ECC error warn about it */
320 /* XXX SOS we could do WARF here */
321 if ((request->status & (ATA_S_CORR | ATA_S_ERROR)) == ATA_S_CORR) {
322 device_printf(request->dev,
323 "WARNING - %s soft error (ECC corrected)",
324 ata_cmd2str(request));
325 if (!(request->flags & (ATA_R_ATAPI | ATA_R_CONTROL)))
326 kprintf(" LBA=%ju", request->u.ata.lba);
330 /* if this is a UDMA CRC error we reinject if there are retries left */
331 if (request->flags & ATA_R_DMA && request->error & ATA_E_ICRC) {
332 if (request->retries-- > 0) {
333 device_printf(request->dev,
334 "WARNING - %s UDMA ICRC error (retrying request)",
335 ata_cmd2str(request));
336 if (!(request->flags & (ATA_R_ATAPI | ATA_R_CONTROL)))
337 kprintf(" LBA=%ju", request->u.ata.lba);
339 request->flags |= (ATA_R_AT_HEAD | ATA_R_REQUEUE);
340 ata_queue_request(request);
346 switch (request->flags & ATA_R_ATAPI) {
350 if (!request->result && request->status & ATA_S_ERROR) {
351 if (!(request->flags & ATA_R_QUIET)) {
352 device_printf(request->dev,
353 "FAILURE - %s status=%b error=%b",
354 ata_cmd2str(request),
355 request->status, "\20\10BUSY\7READY\6DMA_READY"
356 "\5DSC\4DRQ\3CORRECTABLE\2INDEX\1ERROR",
357 request->error, "\20\10ICRC\7UNCORRECTABLE"
358 "\6MEDIA_CHANGED\5NID_NOT_FOUND"
359 "\4MEDIA_CHANGE_REQEST"
360 "\3ABORTED\2NO_MEDIA\1ILLEGAL_LENGTH");
361 if ((request->flags & ATA_R_DMA) &&
362 (request->dmastat & ATA_BMSTAT_ERROR))
363 kprintf(" dma=0x%02x", request->dmastat);
364 if (!(request->flags & (ATA_R_ATAPI | ATA_R_CONTROL)))
365 kprintf(" LBA=%ju", request->u.ata.lba);
368 request->result = EIO;
374 /* skip if result already set */
378 /* if we have a sensekey -> request sense from device */
379 if ((request->error & ATA_E_ATAPI_SENSE_MASK) &&
380 (request->u.atapi.ccb[0] != ATAPI_REQUEST_SENSE)) {
381 static u_int8_t ccb[16] = { ATAPI_REQUEST_SENSE, 0, 0, 0,
382 sizeof(struct atapi_sense),
383 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 };
385 request->u.atapi.saved_cmd = request->u.atapi.ccb[0];
386 bcopy(ccb, request->u.atapi.ccb, 16);
387 request->data = (caddr_t)&request->u.atapi.sense;
388 request->bytecount = sizeof(struct atapi_sense);
389 request->donecount = 0;
390 request->transfersize = sizeof(struct atapi_sense);
391 request->timeout = 5;
392 request->flags &= (ATA_R_ATAPI | ATA_R_QUIET);
393 request->flags |= (ATA_R_READ | ATA_R_AT_HEAD | ATA_R_REQUEUE);
394 ATA_DEBUG_RQ(request, "autoissue request sense");
395 ata_queue_request(request);
399 switch (request->u.atapi.sense.key & ATA_SENSE_KEY_MASK) {
400 case ATA_SENSE_RECOVERED_ERROR:
401 device_printf(request->dev, "WARNING - %s recovered error\n",
402 ata_cmd2str(request));
405 case ATA_SENSE_NO_SENSE:
409 case ATA_SENSE_NOT_READY:
410 request->result = EBUSY;
413 case ATA_SENSE_UNIT_ATTENTION:
414 atadev->flags |= ATA_D_MEDIA_CHANGED;
415 request->result = EIO;
419 request->result = EIO;
420 if (request->flags & ATA_R_QUIET)
423 device_printf(request->dev,
424 "FAILURE - %s %s asc=0x%02x ascq=0x%02x ",
425 ata_cmd2str(request), ata_skey2str(
426 (request->u.atapi.sense.key & ATA_SENSE_KEY_MASK)),
427 request->u.atapi.sense.asc,
428 request->u.atapi.sense.ascq);
429 if (request->u.atapi.sense.specific & ATA_SENSE_SPEC_VALID)
430 kprintf("sks=0x%02x 0x%02x 0x%02x\n",
431 request->u.atapi.sense.specific & ATA_SENSE_SPEC_MASK,
432 request->u.atapi.sense.specific1,
433 request->u.atapi.sense.specific2);
438 if ((request->u.atapi.sense.key & ATA_SENSE_KEY_MASK ?
439 request->u.atapi.sense.key & ATA_SENSE_KEY_MASK :
441 request->result = EIO;
444 ATA_DEBUG_RQ(request, "completed callback/wakeup");
446 /* if we are part of a composite operation we need to maintain progress */
447 if ((composite = request->composite)) {
450 spin_lock_wr(&composite->lock);
452 /* update whats done */
453 if (request->flags & ATA_R_READ)
454 composite->rd_done |= (1 << request->this);
455 if (request->flags & ATA_R_WRITE)
456 composite->wr_done |= (1 << request->this);
458 /* find ready to go dependencies */
459 if (composite->wr_depend &&
460 (composite->rd_done & composite->wr_depend)==composite->wr_depend &&
461 (composite->wr_needed & (~composite->wr_done))) {
462 index = composite->wr_needed & ~composite->wr_done;
465 spin_unlock_wr(&composite->lock);
467 /* if we have any ready candidates kick them off */
471 for (bit = 0; bit < MAX_COMPOSITES; bit++) {
472 if (index & (1 << bit))
473 ata_start(device_get_parent(composite->request[bit]->dev));
478 /* get results back to the initiator for this request */
479 if (request->callback)
480 (request->callback)(request);
482 spin_lock_wr(&request->done);
483 request->flags |= ATA_R_COMPLETED;
485 spin_unlock_wr(&request->done);
488 /* only call ata_start if channel is present */
494 ata_timeout(struct ata_request *request)
496 struct ata_channel *ch = device_get_softc(request->parent);
498 /* acquire state_mtx, softclock_handler() doesn't do this for us */
499 spin_lock_wr(&ch->state_mtx);
501 /*request->flags |= ATA_R_DEBUG;*/
502 ATA_DEBUG_RQ(request, "timeout");
505 * if we have an ATA_ACTIVE request running, we flag the request
506 * ATA_R_TIMEOUT so ata_finish will handle it correctly
507 * also NULL out the running request so we wont loose
508 * the race with an eventual interrupt arriving late
510 if (ch->state == ATA_ACTIVE) {
511 request->flags |= ATA_R_TIMEOUT;
512 spin_unlock_wr(&ch->state_mtx);
513 ATA_LOCKING(ch->dev, ATA_LF_UNLOCK);
517 spin_unlock_wr(&ch->state_mtx);
522 ata_fail_requests(device_t dev)
524 struct ata_channel *ch = device_get_softc(device_get_parent(dev));
525 struct ata_request *request, *tmp;
526 TAILQ_HEAD(, ata_request) fail_requests;
527 TAILQ_INIT(&fail_requests);
529 /* grap all channel locks to avoid races */
530 spin_lock_wr(&ch->queue_mtx);
531 spin_lock_wr(&ch->state_mtx);
533 /* do we have any running request to care about ? */
534 if ((request = ch->running) && (!dev || request->dev == dev)) {
535 callout_stop(&request->callout);
537 request->result = ENXIO;
538 TAILQ_INSERT_TAIL(&fail_requests, request, chain);
541 /* fail all requests queued on this channel for device dev if !NULL */
542 TAILQ_FOREACH_MUTABLE(request, &ch->ata_queue, chain, tmp) {
543 if (!dev || request->dev == dev) {
544 TAILQ_REMOVE(&ch->ata_queue, request, chain);
545 request->result = ENXIO;
546 TAILQ_INSERT_TAIL(&fail_requests, request, chain);
550 spin_unlock_wr(&ch->state_mtx);
551 spin_unlock_wr(&ch->queue_mtx);
553 /* finish up all requests collected above */
554 TAILQ_FOREACH_MUTABLE(request, &fail_requests, chain, tmp) {
555 TAILQ_REMOVE(&fail_requests, request, chain);
561 ata_get_lba(struct ata_request *request)
563 if (request->flags & ATA_R_ATAPI) {
564 switch (request->u.atapi.ccb[0]) {
566 case ATAPI_WRITE_BIG:
568 return (request->u.atapi.ccb[5]) | (request->u.atapi.ccb[4]<<8) |
569 (request->u.atapi.ccb[3]<<16)|(request->u.atapi.ccb[2]<<24);
572 return (request->u.atapi.ccb[4]) | (request->u.atapi.ccb[3]<<8) |
573 (request->u.atapi.ccb[2]<<16);
579 return request->u.ata.lba;
583 ata_sort_queue(struct ata_channel *ch, struct ata_request *request)
585 struct ata_request *this, *next;
587 this = TAILQ_FIRST(&ch->ata_queue);
589 /* if the queue is empty just insert */
591 if (request->composite)
592 ch->freezepoint = request;
593 TAILQ_INSERT_TAIL(&ch->ata_queue, request, chain);
597 /* dont sort frozen parts of the queue */
599 this = ch->freezepoint;
601 /* if position is less than head we add after tipping point */
602 if (ata_get_lba(request) < ata_get_lba(this)) {
603 while ((next = TAILQ_NEXT(this, chain))) {
605 /* have we reached the tipping point */
606 if (ata_get_lba(next) < ata_get_lba(this)) {
608 /* sort the insert */
610 if (ata_get_lba(request) < ata_get_lba(next))
613 } while ((next = TAILQ_NEXT(this, chain)));
620 /* we are after head so sort the insert before tipping point */
622 while ((next = TAILQ_NEXT(this, chain))) {
623 if (ata_get_lba(next) < ata_get_lba(this) ||
624 ata_get_lba(request) < ata_get_lba(next))
630 if (request->composite)
631 ch->freezepoint = request;
632 TAILQ_INSERT_AFTER(&ch->ata_queue, this, request, chain);
636 ata_cmd2str(struct ata_request *request)
638 static char buffer[20];
640 if (request->flags & ATA_R_ATAPI) {
641 switch (request->u.atapi.sense.key ?
642 request->u.atapi.saved_cmd : request->u.atapi.ccb[0]) {
643 case 0x00: return ("TEST_UNIT_READY");
644 case 0x01: return ("REZERO");
645 case 0x03: return ("REQUEST_SENSE");
646 case 0x04: return ("FORMAT");
647 case 0x08: return ("READ");
648 case 0x0a: return ("WRITE");
649 case 0x10: return ("WEOF");
650 case 0x11: return ("SPACE");
651 case 0x12: return ("INQUIRY");
652 case 0x15: return ("MODE_SELECT");
653 case 0x19: return ("ERASE");
654 case 0x1a: return ("MODE_SENSE");
655 case 0x1b: return ("START_STOP");
656 case 0x1e: return ("PREVENT_ALLOW");
657 case 0x23: return ("ATAPI_READ_FORMAT_CAPACITIES");
658 case 0x25: return ("READ_CAPACITY");
659 case 0x28: return ("READ_BIG");
660 case 0x2a: return ("WRITE_BIG");
661 case 0x2b: return ("LOCATE");
662 case 0x34: return ("READ_POSITION");
663 case 0x35: return ("SYNCHRONIZE_CACHE");
664 case 0x3b: return ("WRITE_BUFFER");
665 case 0x3c: return ("READ_BUFFER");
666 case 0x42: return ("READ_SUBCHANNEL");
667 case 0x43: return ("READ_TOC");
668 case 0x45: return ("PLAY_10");
669 case 0x47: return ("PLAY_MSF");
670 case 0x48: return ("PLAY_TRACK");
671 case 0x4b: return ("PAUSE");
672 case 0x51: return ("READ_DISK_INFO");
673 case 0x52: return ("READ_TRACK_INFO");
674 case 0x53: return ("RESERVE_TRACK");
675 case 0x54: return ("SEND_OPC_INFO");
676 case 0x55: return ("MODE_SELECT_BIG");
677 case 0x58: return ("REPAIR_TRACK");
678 case 0x59: return ("READ_MASTER_CUE");
679 case 0x5a: return ("MODE_SENSE_BIG");
680 case 0x5b: return ("CLOSE_TRACK/SESSION");
681 case 0x5c: return ("READ_BUFFER_CAPACITY");
682 case 0x5d: return ("SEND_CUE_SHEET");
683 case 0x96: return ("SERVICE_ACTION_IN");
684 case 0xa1: return ("BLANK_CMD");
685 case 0xa3: return ("SEND_KEY");
686 case 0xa4: return ("REPORT_KEY");
687 case 0xa5: return ("PLAY_12");
688 case 0xa6: return ("LOAD_UNLOAD");
689 case 0xad: return ("READ_DVD_STRUCTURE");
690 case 0xb4: return ("PLAY_CD");
691 case 0xbb: return ("SET_SPEED");
692 case 0xbd: return ("MECH_STATUS");
693 case 0xbe: return ("READ_CD");
694 case 0xff: return ("POLL_DSC");
698 switch (request->u.ata.command) {
699 case 0x00: return ("NOP");
700 case 0x08: return ("DEVICE_RESET");
701 case 0x20: return ("READ");
702 case 0x24: return ("READ48");
703 case 0x25: return ("READ_DMA48");
704 case 0x26: return ("READ_DMA_QUEUED48");
705 case 0x29: return ("READ_MUL48");
706 case 0x30: return ("WRITE");
707 case 0x34: return ("WRITE48");
708 case 0x35: return ("WRITE_DMA48");
709 case 0x36: return ("WRITE_DMA_QUEUED48");
710 case 0x39: return ("WRITE_MUL48");
711 case 0x70: return ("SEEK");
712 case 0xa0: return ("PACKET_CMD");
713 case 0xa1: return ("ATAPI_IDENTIFY");
714 case 0xa2: return ("SERVICE");
715 case 0xc0: return ("CFA ERASE");
716 case 0xc4: return ("READ_MUL");
717 case 0xc5: return ("WRITE_MUL");
718 case 0xc6: return ("SET_MULTI");
719 case 0xc7: return ("READ_DMA_QUEUED");
720 case 0xc8: return ("READ_DMA");
721 case 0xca: return ("WRITE_DMA");
722 case 0xcc: return ("WRITE_DMA_QUEUED");
723 case 0xe6: return ("SLEEP");
724 case 0xe7: return ("FLUSHCACHE");
725 case 0xea: return ("FLUSHCACHE48");
726 case 0xec: return ("ATA_IDENTIFY");
728 switch (request->u.ata.feature) {
729 case 0x03: return ("SETFEATURES SET TRANSFER MODE");
730 case 0x02: return ("SETFEATURES ENABLE WCACHE");
731 case 0x82: return ("SETFEATURES DISABLE WCACHE");
732 case 0xaa: return ("SETFEATURES ENABLE RCACHE");
733 case 0x55: return ("SETFEATURES DISABLE RCACHE");
735 ksprintf(buffer, "SETFEATURES 0x%02x", request->u.ata.feature);
739 ksprintf(buffer, "unknown CMD (0x%02x)", request->u.ata.command);
744 ata_skey2str(u_int8_t skey)
747 case 0x00: return ("NO SENSE");
748 case 0x01: return ("RECOVERED ERROR");
749 case 0x02: return ("NOT READY");
750 case 0x03: return ("MEDIUM ERROR");
751 case 0x04: return ("HARDWARE ERROR");
752 case 0x05: return ("ILLEGAL REQUEST");
753 case 0x06: return ("UNIT ATTENTION");
754 case 0x07: return ("DATA PROTECT");
755 case 0x08: return ("BLANK CHECK");
756 case 0x09: return ("VENDOR SPECIFIC");
757 case 0x0a: return ("COPY ABORTED");
758 case 0x0b: return ("ABORTED COMMAND");
759 case 0x0c: return ("EQUAL");
760 case 0x0d: return ("VOLUME OVERFLOW");
761 case 0x0e: return ("MISCOMPARE");
762 case 0x0f: return ("RESERVED");
763 default: return("UNKNOWN");