aacraid_cam.c revision 315813
1/*-
2 * Copyright (c) 2002-2010 Adaptec, Inc.
3 * Copyright (c) 2010-2012 PMC-Sierra, Inc.
4 * All rights reserved.
5 *
6 * Redistribution and use in source and binary forms, with or without
7 * modification, are permitted provided that the following conditions
8 * are met:
9 * 1. Redistributions of source code must retain the above copyright
10 *    notice, this list of conditions and the following disclaimer.
11 * 2. Redistributions in binary form must reproduce the above copyright
12 *    notice, this list of conditions and the following disclaimer in the
13 *    documentation and/or other materials provided with the distribution.
14 *
15 * THIS SOFTWARE IS PROVIDED BY THE AUTHOR AND CONTRIBUTORS ``AS IS'' AND
16 * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
17 * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
18 * ARE DISCLAIMED.  IN NO EVENT SHALL THE AUTHOR OR CONTRIBUTORS BE LIABLE
19 * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
20 * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
21 * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
22 * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
23 * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
24 * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
25 * SUCH DAMAGE.
26 */
27
28#include <sys/cdefs.h>
29__FBSDID("$FreeBSD: stable/10/sys/dev/aacraid/aacraid_cam.c 315813 2017-03-23 06:41:13Z mav $");
30
31/*
32 * CAM front-end for communicating with non-DASD devices
33 */
34
35#include "opt_aacraid.h"
36
37#include <sys/param.h>
38#include <sys/systm.h>
39#include <sys/kernel.h>
40#include <sys/sysctl.h>
41#include <sys/lock.h>
42#include <sys/malloc.h>
43#include <sys/module.h>
44#include <sys/mutex.h>
45
46#include <cam/cam.h>
47#include <cam/cam_ccb.h>
48#include <cam/cam_debug.h>
49#include <cam/cam_periph.h>
50#if __FreeBSD_version < 801000
51#include <cam/cam_xpt_periph.h>
52#endif
53#include <cam/cam_sim.h>
54#include <cam/cam_xpt_sim.h>
55#include <cam/scsi/scsi_all.h>
56#include <cam/scsi/scsi_message.h>
57
58#include <sys/bus.h>
59#include <sys/conf.h>
60#include <sys/disk.h>
61
62#include <machine/md_var.h>
63#include <machine/bus.h>
64#include <sys/rman.h>
65
66#include <vm/vm.h>
67#include <vm/pmap.h>
68
69#include <dev/aacraid/aacraid_reg.h>
70#include <sys/aac_ioctl.h>
71#include <dev/aacraid/aacraid_debug.h>
72#include <dev/aacraid/aacraid_var.h>
73
74#if __FreeBSD_version >= 700025
75#ifndef	CAM_NEW_TRAN_CODE
76#define	CAM_NEW_TRAN_CODE	1
77#endif
78#endif
79
80#ifndef SVPD_SUPPORTED_PAGE_LIST
81struct scsi_vpd_supported_page_list
82{
83	u_int8_t device;
84	u_int8_t page_code;
85#define	SVPD_SUPPORTED_PAGE_LIST 0x00
86	u_int8_t reserved;
87	u_int8_t length;	/* number of VPD entries */
88#define	SVPD_SUPPORTED_PAGES_SIZE	251
89	u_int8_t list[SVPD_SUPPORTED_PAGES_SIZE];
90};
91#endif
92
93/************************** Version Compatibility *************************/
94#if	__FreeBSD_version < 700031
95#define	aac_sim_alloc(a,b,c,d,e,f,g,h,i)	cam_sim_alloc(a,b,c,d,e,g,h,i)
96#else
97#define	aac_sim_alloc				cam_sim_alloc
98#endif
99
100struct aac_cam {
101	device_t		dev;
102	struct aac_sim		*inf;
103	struct cam_sim		*sim;
104	struct cam_path		*path;
105};
106
107static int aac_cam_probe(device_t dev);
108static int aac_cam_attach(device_t dev);
109static int aac_cam_detach(device_t dev);
110static void aac_cam_action(struct cam_sim *, union ccb *);
111static void aac_cam_poll(struct cam_sim *);
112static void aac_cam_complete(struct aac_command *);
113static void aac_container_complete(struct aac_command *);
114#if __FreeBSD_version >= 700000
115static void aac_cam_rescan(struct aac_softc *sc, uint32_t channel,
116	uint32_t target_id);
117#endif
118static void aac_set_scsi_error(struct aac_softc *sc, union ccb *ccb,
119	u_int8_t status, u_int8_t key, u_int8_t asc, u_int8_t ascq);
120static int aac_load_map_command_sg(struct aac_softc *, struct aac_command *);
121static u_int64_t aac_eval_blockno(u_int8_t *);
122static void aac_container_rw_command(struct cam_sim *, union ccb *, u_int8_t *);
123static void aac_container_special_command(struct cam_sim *, union ccb *,
124	u_int8_t *);
125static void aac_passthrough_command(struct cam_sim *, union ccb *);
126
127static u_int32_t aac_cam_reset_bus(struct cam_sim *, union ccb *);
128static u_int32_t aac_cam_abort_ccb(struct cam_sim *, union ccb *);
129static u_int32_t aac_cam_term_io(struct cam_sim *, union ccb *);
130
131static devclass_t	aacraid_pass_devclass;
132
133static device_method_t	aacraid_pass_methods[] = {
134	DEVMETHOD(device_probe,		aac_cam_probe),
135	DEVMETHOD(device_attach,	aac_cam_attach),
136	DEVMETHOD(device_detach,	aac_cam_detach),
137	{ 0, 0 }
138};
139
140static driver_t	aacraid_pass_driver = {
141	"aacraidp",
142	aacraid_pass_methods,
143	sizeof(struct aac_cam)
144};
145
146DRIVER_MODULE(aacraidp, aacraid, aacraid_pass_driver, aacraid_pass_devclass, 0, 0);
147MODULE_DEPEND(aacraidp, cam, 1, 1, 1);
148
149MALLOC_DEFINE(M_AACRAIDCAM, "aacraidcam", "AACRAID CAM info");
150
151static void
152aac_set_scsi_error(struct aac_softc *sc, union ccb *ccb, u_int8_t status,
153	u_int8_t key, u_int8_t asc, u_int8_t ascq)
154{
155#if __FreeBSD_version >= 900000
156	struct scsi_sense_data_fixed *sense =
157		(struct scsi_sense_data_fixed *)&ccb->csio.sense_data;
158#else
159	struct scsi_sense_data *sense = &ccb->csio.sense_data;
160#endif
161
162	fwprintf(sc, HBA_FLAGS_DBG_FUNCTION_ENTRY_B, "Error %d!", status);
163
164	ccb->ccb_h.status = CAM_SCSI_STATUS_ERROR;
165	ccb->csio.scsi_status = status;
166	if (status == SCSI_STATUS_CHECK_COND) {
167		ccb->ccb_h.status |= CAM_AUTOSNS_VALID;
168		bzero(&ccb->csio.sense_data, ccb->csio.sense_len);
169		ccb->csio.sense_data.error_code =
170			SSD_CURRENT_ERROR | SSD_ERRCODE_VALID;
171		sense->flags = key;
172		if (ccb->csio.sense_len >= 14) {
173			sense->extra_len = 6;
174			sense->add_sense_code = asc;
175			sense->add_sense_code_qual = ascq;
176		}
177	}
178}
179
180#if __FreeBSD_version >= 700000
181static void
182aac_cam_rescan(struct aac_softc *sc, uint32_t channel, uint32_t target_id)
183{
184	union ccb *ccb;
185	struct aac_sim *sim;
186	struct aac_cam *camsc;
187
188	if (target_id == AAC_CAM_TARGET_WILDCARD)
189		target_id = CAM_TARGET_WILDCARD;
190
191	TAILQ_FOREACH(sim, &sc->aac_sim_tqh, sim_link) {
192		camsc = sim->aac_cam;
193		if (camsc == NULL || camsc->inf == NULL ||
194		    camsc->inf->BusNumber != channel)
195			continue;
196
197		ccb = xpt_alloc_ccb_nowait();
198		if (ccb == NULL) {
199			device_printf(sc->aac_dev,
200			    "Cannot allocate ccb for bus rescan.\n");
201			return;
202		}
203
204		if (xpt_create_path(&ccb->ccb_h.path, xpt_periph,
205		    cam_sim_path(camsc->sim),
206		    target_id, CAM_LUN_WILDCARD) != CAM_REQ_CMP) {
207			xpt_free_ccb(ccb);
208			device_printf(sc->aac_dev,
209			    "Cannot create path for bus rescan.\n");
210			return;
211		}
212		xpt_rescan(ccb);
213		break;
214	}
215}
216#endif
217
218static void
219aac_cam_event(struct aac_softc *sc, struct aac_event *event, void *arg)
220{
221	union ccb *ccb;
222	struct aac_cam *camsc;
223
224	switch (event->ev_type) {
225	case AAC_EVENT_CMFREE:
226		ccb = arg;
227		camsc = ccb->ccb_h.sim_priv.entries[0].ptr;
228		free(event, M_AACRAIDCAM);
229		xpt_release_simq(camsc->sim, 1);
230		ccb->ccb_h.status = CAM_REQUEUE_REQ;
231		xpt_done(ccb);
232		break;
233	default:
234		device_printf(sc->aac_dev, "unknown event %d in aac_cam\n",
235		    event->ev_type);
236		break;
237	}
238
239	return;
240}
241
242static int
243aac_cam_probe(device_t dev)
244{
245	struct aac_softc *sc;
246	struct aac_cam *camsc;
247
248	camsc = (struct aac_cam *)device_get_softc(dev);
249	if (!camsc->inf)
250		return (0);
251	sc = camsc->inf->aac_sc;
252	fwprintf(sc, HBA_FLAGS_DBG_FUNCTION_ENTRY_B, "");
253	return (0);
254}
255
256static int
257aac_cam_detach(device_t dev)
258{
259	struct aac_softc *sc;
260	struct aac_cam *camsc;
261
262	camsc = (struct aac_cam *)device_get_softc(dev);
263	if (!camsc->inf)
264		return (0);
265	sc = camsc->inf->aac_sc;
266	fwprintf(sc, HBA_FLAGS_DBG_FUNCTION_ENTRY_B, "");
267	camsc->inf->aac_cam = NULL;
268
269	mtx_lock(&sc->aac_io_lock);
270
271	xpt_async(AC_LOST_DEVICE, camsc->path, NULL);
272	xpt_free_path(camsc->path);
273	xpt_bus_deregister(cam_sim_path(camsc->sim));
274	cam_sim_free(camsc->sim, /*free_devq*/TRUE);
275
276	sc->cam_rescan_cb = NULL;
277
278	mtx_unlock(&sc->aac_io_lock);
279
280	return (0);
281}
282
283/*
284 * Register the driver as a CAM SIM
285 */
286static int
287aac_cam_attach(device_t dev)
288{
289	struct cam_devq *devq;
290	struct cam_sim *sim;
291	struct cam_path *path;
292	struct aac_cam *camsc;
293	struct aac_sim *inf;
294
295	camsc = (struct aac_cam *)device_get_softc(dev);
296	inf = (struct aac_sim *)device_get_ivars(dev);
297	if (!inf)
298		return (EIO);
299	fwprintf(inf->aac_sc, HBA_FLAGS_DBG_FUNCTION_ENTRY_B, "");
300	camsc->inf = inf;
301	camsc->inf->aac_cam = camsc;
302
303	devq = cam_simq_alloc(inf->TargetsPerBus);
304	if (devq == NULL)
305		return (EIO);
306
307	sim = aac_sim_alloc(aac_cam_action, aac_cam_poll, "aacraidp", camsc,
308	    device_get_unit(dev), &inf->aac_sc->aac_io_lock, 1, 1, devq);
309	if (sim == NULL) {
310		cam_simq_free(devq);
311		return (EIO);
312	}
313
314	/* Since every bus has it's own sim, every bus 'appears' as bus 0 */
315	mtx_lock(&inf->aac_sc->aac_io_lock);
316	if (aac_xpt_bus_register(sim, dev, 0) != CAM_SUCCESS) {
317		cam_sim_free(sim, TRUE);
318		mtx_unlock(&inf->aac_sc->aac_io_lock);
319		return (EIO);
320	}
321
322	if (xpt_create_path(&path, NULL, cam_sim_path(sim),
323	    CAM_TARGET_WILDCARD, CAM_LUN_WILDCARD) != CAM_REQ_CMP) {
324		xpt_bus_deregister(cam_sim_path(sim));
325		cam_sim_free(sim, TRUE);
326		mtx_unlock(&inf->aac_sc->aac_io_lock);
327		return (EIO);
328	}
329
330#if __FreeBSD_version >= 700000
331	inf->aac_sc->cam_rescan_cb = aac_cam_rescan;
332#endif
333	mtx_unlock(&inf->aac_sc->aac_io_lock);
334
335	camsc->sim = sim;
336	camsc->path = path;
337
338	return (0);
339}
340
341static u_int64_t
342aac_eval_blockno(u_int8_t *cmdp)
343{
344	u_int64_t blockno;
345
346	switch (cmdp[0]) {
347	case READ_6:
348	case WRITE_6:
349		blockno = scsi_3btoul(((struct scsi_rw_6 *)cmdp)->addr);
350		break;
351	case READ_10:
352	case WRITE_10:
353		blockno = scsi_4btoul(((struct scsi_rw_10 *)cmdp)->addr);
354		break;
355	case READ_12:
356	case WRITE_12:
357		blockno = scsi_4btoul(((struct scsi_rw_12 *)cmdp)->addr);
358		break;
359	case READ_16:
360	case WRITE_16:
361		blockno = scsi_8btou64(((struct scsi_rw_16 *)cmdp)->addr);
362		break;
363	default:
364		blockno = 0;
365		break;
366	}
367	return(blockno);
368}
369
370static void
371aac_container_rw_command(struct cam_sim *sim, union ccb *ccb, u_int8_t *cmdp)
372{
373	struct	aac_cam *camsc;
374	struct	aac_softc *sc;
375	struct	aac_command *cm;
376	struct	aac_fib *fib;
377	u_int64_t blockno;
378
379	camsc = (struct aac_cam *)cam_sim_softc(sim);
380	sc = camsc->inf->aac_sc;
381	mtx_assert(&sc->aac_io_lock, MA_OWNED);
382
383	if (aacraid_alloc_command(sc, &cm)) {
384		struct aac_event *event;
385
386		xpt_freeze_simq(sim, 1);
387		ccb->ccb_h.status = CAM_RESRC_UNAVAIL;
388		ccb->ccb_h.sim_priv.entries[0].ptr = camsc;
389		event = malloc(sizeof(struct aac_event), M_AACRAIDCAM,
390		    M_NOWAIT | M_ZERO);
391		if (event == NULL) {
392			device_printf(sc->aac_dev,
393			    "Warning, out of memory for event\n");
394			return;
395		}
396		event->ev_callback = aac_cam_event;
397		event->ev_arg = ccb;
398		event->ev_type = AAC_EVENT_CMFREE;
399		aacraid_add_event(sc, event);
400		return;
401	}
402
403	fib = cm->cm_fib;
404	switch (ccb->ccb_h.flags & CAM_DIR_MASK) {
405	case CAM_DIR_IN:
406		cm->cm_flags |= AAC_CMD_DATAIN;
407		break;
408	case CAM_DIR_OUT:
409		cm->cm_flags |= AAC_CMD_DATAOUT;
410		break;
411	case CAM_DIR_NONE:
412		break;
413	default:
414		cm->cm_flags |= AAC_CMD_DATAIN | AAC_CMD_DATAOUT;
415		break;
416	}
417
418	blockno = aac_eval_blockno(cmdp);
419
420	cm->cm_complete = aac_container_complete;
421	cm->cm_ccb = ccb;
422	cm->cm_timestamp = time_uptime;
423	cm->cm_data = (void *)ccb->csio.data_ptr;
424	cm->cm_datalen = ccb->csio.dxfer_len;
425
426	fib->Header.Size = sizeof(struct aac_fib_header);
427	fib->Header.XferState =
428		AAC_FIBSTATE_HOSTOWNED   |
429		AAC_FIBSTATE_INITIALISED |
430		AAC_FIBSTATE_EMPTY	 |
431		AAC_FIBSTATE_FROMHOST	 |
432		AAC_FIBSTATE_REXPECTED   |
433		AAC_FIBSTATE_NORM	 |
434		AAC_FIBSTATE_ASYNC	 |
435		AAC_FIBSTATE_FAST_RESPONSE;
436
437	if (sc->flags & AAC_FLAGS_NEW_COMM_TYPE2) {
438		struct aac_raw_io2 *raw;
439		raw = (struct aac_raw_io2 *)&fib->data[0];
440		bzero(raw, sizeof(struct aac_raw_io2));
441		fib->Header.Command = RawIo2;
442		raw->strtBlkLow = (u_int32_t)blockno;
443		raw->strtBlkHigh = (u_int32_t)(blockno >> 32);
444		raw->byteCnt = cm->cm_datalen;
445		raw->ldNum = ccb->ccb_h.target_id;
446		fib->Header.Size += sizeof(struct aac_raw_io2);
447		cm->cm_sgtable = (struct aac_sg_table *)raw->sge;
448		if (cm->cm_flags & AAC_CMD_DATAIN)
449			raw->flags = RIO2_IO_TYPE_READ | RIO2_SG_FORMAT_IEEE1212;
450		else
451			raw->flags = RIO2_IO_TYPE_WRITE | RIO2_SG_FORMAT_IEEE1212;
452	} else if (sc->flags & AAC_FLAGS_RAW_IO) {
453		struct aac_raw_io *raw;
454		raw = (struct aac_raw_io *)&fib->data[0];
455		bzero(raw, sizeof(struct aac_raw_io));
456		fib->Header.Command = RawIo;
457		raw->BlockNumber = blockno;
458		raw->ByteCount = cm->cm_datalen;
459		raw->ContainerId = ccb->ccb_h.target_id;
460		fib->Header.Size += sizeof(struct aac_raw_io);
461		cm->cm_sgtable = (struct aac_sg_table *)
462			&raw->SgMapRaw;
463		if (cm->cm_flags & AAC_CMD_DATAIN)
464			raw->Flags = 1;
465	} else if ((sc->flags & AAC_FLAGS_SG_64BIT) == 0) {
466		fib->Header.Command = ContainerCommand;
467		if (cm->cm_flags & AAC_CMD_DATAIN) {
468			struct aac_blockread *br;
469			br = (struct aac_blockread *)&fib->data[0];
470			br->Command = VM_CtBlockRead;
471			br->ContainerId = ccb->ccb_h.target_id;
472			br->BlockNumber = blockno;
473			br->ByteCount = cm->cm_datalen;
474			fib->Header.Size += sizeof(struct aac_blockread);
475			cm->cm_sgtable = &br->SgMap;
476		} else {
477			struct aac_blockwrite *bw;
478			bw = (struct aac_blockwrite *)&fib->data[0];
479			bw->Command = VM_CtBlockWrite;
480			bw->ContainerId = ccb->ccb_h.target_id;
481			bw->BlockNumber = blockno;
482			bw->ByteCount = cm->cm_datalen;
483			bw->Stable = CUNSTABLE;
484			fib->Header.Size += sizeof(struct aac_blockwrite);
485			cm->cm_sgtable = &bw->SgMap;
486		}
487	} else {
488		fib->Header.Command = ContainerCommand64;
489		if (cm->cm_flags & AAC_CMD_DATAIN) {
490			struct aac_blockread64 *br;
491			br = (struct aac_blockread64 *)&fib->data[0];
492			br->Command = VM_CtHostRead64;
493			br->ContainerId = ccb->ccb_h.target_id;
494			br->SectorCount = cm->cm_datalen/AAC_BLOCK_SIZE;
495			br->BlockNumber = blockno;
496			br->Pad = 0;
497			br->Flags = 0;
498			fib->Header.Size += sizeof(struct aac_blockread64);
499			cm->cm_sgtable = (struct aac_sg_table *)&br->SgMap64;
500		} else {
501			struct aac_blockwrite64 *bw;
502			bw = (struct aac_blockwrite64 *)&fib->data[0];
503			bw->Command = VM_CtHostWrite64;
504			bw->ContainerId = ccb->ccb_h.target_id;
505			bw->SectorCount = cm->cm_datalen/AAC_BLOCK_SIZE;
506			bw->BlockNumber = blockno;
507			bw->Pad = 0;
508			bw->Flags = 0;
509			fib->Header.Size += sizeof(struct aac_blockwrite64);
510			cm->cm_sgtable = (struct aac_sg_table *)&bw->SgMap64;
511		}
512	}
513	aac_enqueue_ready(cm);
514	aacraid_startio(cm->cm_sc);
515}
516
517static void
518aac_container_special_command(struct cam_sim *sim, union ccb *ccb,
519	u_int8_t *cmdp)
520{
521	struct	aac_cam *camsc;
522	struct	aac_softc *sc;
523	struct	aac_container *co;
524
525	camsc = (struct aac_cam *)cam_sim_softc(sim);
526	sc = camsc->inf->aac_sc;
527	mtx_assert(&sc->aac_io_lock, MA_OWNED);
528
529	TAILQ_FOREACH(co, &sc->aac_container_tqh, co_link) {
530		fwprintf(sc, HBA_FLAGS_DBG_ERROR_B, "found container %d search for %d", co->co_mntobj.ObjectId, ccb->ccb_h.target_id);
531		if (co->co_mntobj.ObjectId == ccb->ccb_h.target_id)
532			break;
533	}
534	if (co == NULL || ccb->ccb_h.target_lun != 0) {
535		fwprintf(sc, HBA_FLAGS_DBG_ERROR_B,
536			"Container not present: cmd 0x%x id %d lun %d len %d",
537			*cmdp, ccb->ccb_h.target_id, ccb->ccb_h.target_lun,
538			ccb->csio.dxfer_len);
539		ccb->ccb_h.status = CAM_DEV_NOT_THERE;
540		xpt_done(ccb);
541		return;
542	}
543
544	if (ccb->csio.dxfer_len)
545		bzero(ccb->csio.data_ptr, ccb->csio.dxfer_len);
546
547	switch (*cmdp) {
548	case INQUIRY:
549	{
550		struct scsi_inquiry *inq = (struct scsi_inquiry *)cmdp;
551
552		fwprintf(sc, HBA_FLAGS_DBG_COMM_B,
553		"Container INQUIRY id %d lun %d len %d VPD 0x%x Page 0x%x",
554			ccb->ccb_h.target_id, ccb->ccb_h.target_lun,
555			ccb->csio.dxfer_len, inq->byte2, inq->page_code);
556		if (!(inq->byte2 & SI_EVPD)) {
557			struct scsi_inquiry_data *p =
558				(struct scsi_inquiry_data *)ccb->csio.data_ptr;
559			if (inq->page_code != 0) {
560				aac_set_scsi_error(sc, ccb,
561					SCSI_STATUS_CHECK_COND,
562					SSD_KEY_ILLEGAL_REQUEST, 0x24, 0x00);
563				xpt_done(ccb);
564				return;
565			}
566			p->device = T_DIRECT;
567			p->version = SCSI_REV_SPC2;
568			p->response_format = 2;
569			if (ccb->csio.dxfer_len >= 36) {
570				p->additional_length = 31;
571				p->flags = SID_WBus16|SID_Sync|SID_CmdQue;
572				/* OEM Vendor defines */
573				strncpy(p->vendor, "Adaptec ", sizeof(p->vendor));
574				strncpy(p->product, "Array           ",
575				    sizeof(p->product));
576				strncpy(p->revision, "V1.0",
577				    sizeof(p->revision));
578			}
579		} else {
580			if (inq->page_code == SVPD_SUPPORTED_PAGE_LIST) {
581				struct scsi_vpd_supported_page_list *p =
582					(struct scsi_vpd_supported_page_list *)
583					ccb->csio.data_ptr;
584				p->device = T_DIRECT;
585				p->page_code = SVPD_SUPPORTED_PAGE_LIST;
586				p->length = 2;
587				p->list[0] = SVPD_SUPPORTED_PAGE_LIST;
588				p->list[1] = SVPD_UNIT_SERIAL_NUMBER;
589			} else if (inq->page_code == SVPD_UNIT_SERIAL_NUMBER) {
590				struct scsi_vpd_unit_serial_number *p =
591					(struct scsi_vpd_unit_serial_number *)
592					ccb->csio.data_ptr;
593				p->device = T_DIRECT;
594				p->page_code = SVPD_UNIT_SERIAL_NUMBER;
595				p->length = sprintf((char *)p->serial_num,
596					"%08X%02X", co->co_uid,
597					ccb->ccb_h.target_id);
598			} else {
599				aac_set_scsi_error(sc, ccb,
600					SCSI_STATUS_CHECK_COND,
601					SSD_KEY_ILLEGAL_REQUEST, 0x24, 0x00);
602				xpt_done(ccb);
603				return;
604			}
605		}
606		ccb->ccb_h.status = CAM_REQ_CMP;
607		break;
608	}
609
610	case REPORT_LUNS:
611		fwprintf(sc, HBA_FLAGS_DBG_COMM_B,
612		"Container REPORT_LUNS id %d lun %d len %d",
613		ccb->ccb_h.target_id, ccb->ccb_h.target_lun,
614		ccb->csio.dxfer_len);
615		ccb->ccb_h.status = CAM_REQ_CMP;
616		break;
617
618	case START_STOP:
619	{
620		struct scsi_start_stop_unit *ss =
621			(struct scsi_start_stop_unit *)cmdp;
622		fwprintf(sc, HBA_FLAGS_DBG_COMM_B,
623		"Container START_STOP id %d lun %d len %d",
624		ccb->ccb_h.target_id, ccb->ccb_h.target_lun,
625		ccb->csio.dxfer_len);
626		if (sc->aac_support_opt2 & AAC_SUPPORTED_POWER_MANAGEMENT) {
627			struct aac_command *cm;
628			struct aac_fib *fib;
629			struct aac_cnt_config *ccfg;
630
631			if (aacraid_alloc_command(sc, &cm)) {
632				struct aac_event *event;
633
634				xpt_freeze_simq(sim, 1);
635				ccb->ccb_h.status = CAM_RESRC_UNAVAIL;
636				ccb->ccb_h.sim_priv.entries[0].ptr = camsc;
637				event = malloc(sizeof(struct aac_event), M_AACRAIDCAM,
638					M_NOWAIT | M_ZERO);
639				if (event == NULL) {
640					device_printf(sc->aac_dev,
641						"Warning, out of memory for event\n");
642					return;
643				}
644				event->ev_callback = aac_cam_event;
645				event->ev_arg = ccb;
646				event->ev_type = AAC_EVENT_CMFREE;
647				aacraid_add_event(sc, event);
648				return;
649			}
650
651			fib = cm->cm_fib;
652			cm->cm_timestamp = time_uptime;
653			cm->cm_datalen = 0;
654
655			fib->Header.Size =
656				sizeof(struct aac_fib_header) + sizeof(struct aac_cnt_config);
657			fib->Header.XferState =
658				AAC_FIBSTATE_HOSTOWNED   |
659				AAC_FIBSTATE_INITIALISED |
660				AAC_FIBSTATE_EMPTY	 |
661				AAC_FIBSTATE_FROMHOST	 |
662				AAC_FIBSTATE_REXPECTED   |
663				AAC_FIBSTATE_NORM	 |
664				AAC_FIBSTATE_ASYNC	 |
665				AAC_FIBSTATE_FAST_RESPONSE;
666			fib->Header.Command = ContainerCommand;
667
668			/* Start unit */
669			ccfg = (struct aac_cnt_config *)&fib->data[0];
670			bzero(ccfg, sizeof (*ccfg) - CT_PACKET_SIZE);
671			ccfg->Command = VM_ContainerConfig;
672			ccfg->CTCommand.command = CT_PM_DRIVER_SUPPORT;
673			ccfg->CTCommand.param[0] = (ss->how & SSS_START ?
674				AAC_PM_DRIVERSUP_START_UNIT :
675				AAC_PM_DRIVERSUP_STOP_UNIT);
676			ccfg->CTCommand.param[1] = co->co_mntobj.ObjectId;
677			ccfg->CTCommand.param[2] = 0;	/* 1 - immediate */
678
679			if (aacraid_wait_command(cm) != 0 ||
680				*(u_int32_t *)&fib->data[0] != 0) {
681				printf("Power Management: Error start/stop container %d\n",
682				co->co_mntobj.ObjectId);
683			}
684			aacraid_release_command(cm);
685		}
686		ccb->ccb_h.status = CAM_REQ_CMP;
687		break;
688	}
689
690	case TEST_UNIT_READY:
691		fwprintf(sc, HBA_FLAGS_DBG_COMM_B,
692		"Container TEST_UNIT_READY id %d lun %d len %d",
693		ccb->ccb_h.target_id, ccb->ccb_h.target_lun,
694		ccb->csio.dxfer_len);
695		ccb->ccb_h.status = CAM_REQ_CMP;
696		break;
697
698	case REQUEST_SENSE:
699		fwprintf(sc, HBA_FLAGS_DBG_COMM_B,
700		"Container REQUEST_SENSE id %d lun %d len %d",
701		ccb->ccb_h.target_id, ccb->ccb_h.target_lun,
702		ccb->csio.dxfer_len);
703		ccb->ccb_h.status = CAM_REQ_CMP;
704		break;
705
706	case READ_CAPACITY:
707	{
708		struct scsi_read_capacity_data *p =
709			(struct scsi_read_capacity_data *)ccb->csio.data_ptr;
710		fwprintf(sc, HBA_FLAGS_DBG_COMM_B,
711		"Container READ_CAPACITY id %d lun %d len %d",
712		ccb->ccb_h.target_id, ccb->ccb_h.target_lun,
713		ccb->csio.dxfer_len);
714		scsi_ulto4b(co->co_mntobj.ObjExtension.BlockDevice.BlockSize, p->length);
715		/* check if greater than 2TB */
716		if (co->co_mntobj.CapacityHigh) {
717			if (sc->flags & AAC_FLAGS_LBA_64BIT)
718				scsi_ulto4b(0xffffffff, p->addr);
719		} else {
720			scsi_ulto4b(co->co_mntobj.Capacity-1, p->addr);
721		}
722		ccb->ccb_h.status = CAM_REQ_CMP;
723		break;
724	}
725
726	case SERVICE_ACTION_IN:
727	{
728		struct scsi_read_capacity_data_long *p =
729			(struct scsi_read_capacity_data_long *)
730			ccb->csio.data_ptr;
731		fwprintf(sc, HBA_FLAGS_DBG_COMM_B,
732		"Container SERVICE_ACTION_IN id %d lun %d len %d",
733		ccb->ccb_h.target_id, ccb->ccb_h.target_lun,
734		ccb->csio.dxfer_len);
735		if (((struct scsi_read_capacity_16 *)cmdp)->service_action !=
736			SRC16_SERVICE_ACTION) {
737			aac_set_scsi_error(sc, ccb, SCSI_STATUS_CHECK_COND,
738				SSD_KEY_ILLEGAL_REQUEST, 0x24, 0x00);
739			xpt_done(ccb);
740			return;
741		}
742		scsi_ulto4b(co->co_mntobj.ObjExtension.BlockDevice.BlockSize, p->length);
743		scsi_ulto4b(co->co_mntobj.CapacityHigh, p->addr);
744		scsi_ulto4b(co->co_mntobj.Capacity-1, &p->addr[4]);
745
746		if (ccb->csio.dxfer_len >= 14) {
747			u_int32_t mapping = co->co_mntobj.ObjExtension.BlockDevice.bdLgclPhysMap;
748			p->prot_lbppbe = 0;
749			while (mapping > 1) {
750				mapping >>= 1;
751				p->prot_lbppbe++;
752			}
753			p->prot_lbppbe &= 0x0f;
754		}
755
756		ccb->ccb_h.status = CAM_REQ_CMP;
757		break;
758	}
759
760	case MODE_SENSE_6:
761	{
762		struct scsi_mode_sense_6 *msp =(struct scsi_mode_sense_6 *)cmdp;
763		struct ms6_data {
764			struct scsi_mode_hdr_6 hd;
765			struct scsi_mode_block_descr bd;
766			char pages;
767		} *p = (struct ms6_data *)ccb->csio.data_ptr;
768		char *pagep;
769		int return_all_pages = FALSE;
770
771		fwprintf(sc, HBA_FLAGS_DBG_COMM_B,
772		"Container MODE_SENSE id %d lun %d len %d page %d",
773		ccb->ccb_h.target_id, ccb->ccb_h.target_lun,
774		ccb->csio.dxfer_len, msp->page);
775		p->hd.datalen = sizeof(struct scsi_mode_hdr_6) - 1;
776		if (co->co_mntobj.ContentState & AAC_FSCS_READONLY)
777			p->hd.dev_specific = 0x80;	/* WP */
778		p->hd.dev_specific |= 0x10;	/* DPOFUA */
779		if (msp->byte2 & SMS_DBD) {
780			p->hd.block_descr_len = 0;
781		} else {
782			p->hd.block_descr_len =
783				sizeof(struct scsi_mode_block_descr);
784			p->hd.datalen += p->hd.block_descr_len;
785			scsi_ulto3b(co->co_mntobj.ObjExtension.BlockDevice.BlockSize, p->bd.block_len);
786			if (co->co_mntobj.Capacity > 0xffffff ||
787				co->co_mntobj.CapacityHigh) {
788				p->bd.num_blocks[0] = 0xff;
789				p->bd.num_blocks[1] = 0xff;
790				p->bd.num_blocks[2] = 0xff;
791			} else {
792				p->bd.num_blocks[0] = (u_int8_t)
793					(co->co_mntobj.Capacity >> 16);
794				p->bd.num_blocks[1] = (u_int8_t)
795					(co->co_mntobj.Capacity >> 8);
796				p->bd.num_blocks[2] = (u_int8_t)
797					(co->co_mntobj.Capacity);
798			}
799		}
800		pagep = &p->pages;
801		switch (msp->page & SMS_PAGE_CODE) {
802		case SMS_ALL_PAGES_PAGE:
803			return_all_pages = TRUE;
804		case SMS_CONTROL_MODE_PAGE:
805		{
806			struct scsi_control_page *cp =
807				(struct scsi_control_page *)pagep;
808
809			if (ccb->csio.dxfer_len <= p->hd.datalen + 8) {
810				aac_set_scsi_error(sc, ccb,
811					SCSI_STATUS_CHECK_COND,
812					SSD_KEY_ILLEGAL_REQUEST, 0x24, 0x00);
813				xpt_done(ccb);
814				return;
815			}
816			cp->page_code = SMS_CONTROL_MODE_PAGE;
817			cp->page_length = 6;
818			p->hd.datalen += 8;
819			pagep += 8;
820			if (!return_all_pages)
821				break;
822		}
823		case SMS_VENDOR_SPECIFIC_PAGE:
824			break;
825		default:
826			aac_set_scsi_error(sc, ccb, SCSI_STATUS_CHECK_COND,
827				SSD_KEY_ILLEGAL_REQUEST, 0x24, 0x00);
828			xpt_done(ccb);
829			return;
830		}
831		ccb->ccb_h.status = CAM_REQ_CMP;
832		break;
833	}
834
835	case SYNCHRONIZE_CACHE:
836		fwprintf(sc, HBA_FLAGS_DBG_COMM_B,
837		"Container SYNCHRONIZE_CACHE id %d lun %d len %d",
838		ccb->ccb_h.target_id, ccb->ccb_h.target_lun,
839		ccb->csio.dxfer_len);
840		ccb->ccb_h.status = CAM_REQ_CMP;
841		break;
842
843	default:
844		fwprintf(sc, HBA_FLAGS_DBG_ERROR_B,
845		"Container unsupp. cmd 0x%x id %d lun %d len %d",
846		*cmdp, ccb->ccb_h.target_id, ccb->ccb_h.target_lun,
847		ccb->csio.dxfer_len);
848		ccb->ccb_h.status = CAM_REQ_CMP; /*CAM_REQ_INVALID*/
849		break;
850	}
851	xpt_done(ccb);
852}
853
854static void
855aac_passthrough_command(struct cam_sim *sim, union ccb *ccb)
856{
857	struct	aac_cam *camsc;
858	struct	aac_softc *sc;
859	struct	aac_command *cm;
860	struct	aac_fib *fib;
861	struct	aac_srb *srb;
862
863	camsc = (struct aac_cam *)cam_sim_softc(sim);
864	sc = camsc->inf->aac_sc;
865	mtx_assert(&sc->aac_io_lock, MA_OWNED);
866
867	if (aacraid_alloc_command(sc, &cm)) {
868		struct aac_event *event;
869
870		xpt_freeze_simq(sim, 1);
871		ccb->ccb_h.status = CAM_RESRC_UNAVAIL;
872		ccb->ccb_h.sim_priv.entries[0].ptr = camsc;
873		event = malloc(sizeof(struct aac_event), M_AACRAIDCAM,
874		    M_NOWAIT | M_ZERO);
875		if (event == NULL) {
876			device_printf(sc->aac_dev,
877			    "Warning, out of memory for event\n");
878			return;
879		}
880		event->ev_callback = aac_cam_event;
881		event->ev_arg = ccb;
882		event->ev_type = AAC_EVENT_CMFREE;
883		aacraid_add_event(sc, event);
884		return;
885	}
886
887	fib = cm->cm_fib;
888	switch (ccb->ccb_h.flags & CAM_DIR_MASK) {
889	case CAM_DIR_IN:
890		cm->cm_flags |= AAC_CMD_DATAIN;
891		break;
892	case CAM_DIR_OUT:
893		cm->cm_flags |= AAC_CMD_DATAOUT;
894		break;
895	case CAM_DIR_NONE:
896		break;
897	default:
898		cm->cm_flags |= AAC_CMD_DATAIN | AAC_CMD_DATAOUT;
899		break;
900	}
901
902	srb = (struct aac_srb *)&fib->data[0];
903	srb->function = AAC_SRB_FUNC_EXECUTE_SCSI;
904	if (cm->cm_flags & (AAC_CMD_DATAIN|AAC_CMD_DATAOUT))
905		srb->flags = AAC_SRB_FLAGS_UNSPECIFIED_DIRECTION;
906	if (cm->cm_flags & AAC_CMD_DATAIN)
907		srb->flags = AAC_SRB_FLAGS_DATA_IN;
908	else if (cm->cm_flags & AAC_CMD_DATAOUT)
909		srb->flags = AAC_SRB_FLAGS_DATA_OUT;
910	else
911		srb->flags = AAC_SRB_FLAGS_NO_DATA_XFER;
912
913	/*
914	 * Copy the CDB into the SRB.  It's only 6-16 bytes,
915	 * so a copy is not too expensive.
916	 */
917	srb->cdb_len = ccb->csio.cdb_len;
918	if (ccb->ccb_h.flags & CAM_CDB_POINTER)
919		bcopy(ccb->csio.cdb_io.cdb_ptr, (u_int8_t *)&srb->cdb[0],
920			srb->cdb_len);
921	else
922		bcopy(ccb->csio.cdb_io.cdb_bytes, (u_int8_t *)&srb->cdb[0],
923			srb->cdb_len);
924
925	/* Set command */
926	fib->Header.Command = (sc->flags & AAC_FLAGS_SG_64BIT) ?
927		ScsiPortCommandU64 : ScsiPortCommand;
928	fib->Header.Size = sizeof(struct aac_fib_header) +
929			sizeof(struct aac_srb);
930
931	/* Map the s/g list */
932	cm->cm_sgtable = &srb->sg_map;
933	if ((ccb->ccb_h.flags & CAM_DIR_MASK) != CAM_DIR_NONE) {
934		/*
935		 * Arrange things so that the S/G
936		 * map will get set up automagically
937		 */
938		cm->cm_data = (void *)ccb->csio.data_ptr;
939		cm->cm_datalen = ccb->csio.dxfer_len;
940		srb->data_len = ccb->csio.dxfer_len;
941	} else {
942		cm->cm_data = NULL;
943		cm->cm_datalen = 0;
944		srb->data_len = 0;
945	}
946
947	srb->bus = camsc->inf->BusNumber - 1; /* Bus no. rel. to the card */
948	srb->target = ccb->ccb_h.target_id;
949	srb->lun = ccb->ccb_h.target_lun;
950	srb->timeout = ccb->ccb_h.timeout;	/* XXX */
951	srb->retry_limit = 0;
952
953	cm->cm_complete = aac_cam_complete;
954	cm->cm_ccb = ccb;
955	cm->cm_timestamp = time_uptime;
956
957	fib->Header.XferState =
958			AAC_FIBSTATE_HOSTOWNED	|
959			AAC_FIBSTATE_INITIALISED	|
960			AAC_FIBSTATE_FROMHOST	|
961			AAC_FIBSTATE_REXPECTED	|
962			AAC_FIBSTATE_NORM	|
963			AAC_FIBSTATE_ASYNC	 |
964			AAC_FIBSTATE_FAST_RESPONSE;
965
966	aac_enqueue_ready(cm);
967	aacraid_startio(cm->cm_sc);
968}
969
970static void
971aac_cam_action(struct cam_sim *sim, union ccb *ccb)
972{
973	struct	aac_cam *camsc;
974	struct	aac_softc *sc;
975
976	camsc = (struct aac_cam *)cam_sim_softc(sim);
977	sc = camsc->inf->aac_sc;
978	fwprintf(sc, HBA_FLAGS_DBG_FUNCTION_ENTRY_B, "");
979	mtx_assert(&sc->aac_io_lock, MA_OWNED);
980
981	/* Synchronous ops, and ops that don't require communication with the
982	 * controller */
983	switch(ccb->ccb_h.func_code) {
984	case XPT_SCSI_IO:
985		/* This is handled down below */
986		break;
987	case XPT_CALC_GEOMETRY:
988	{
989		struct ccb_calc_geometry *ccg;
990		u_int32_t size_mb;
991		u_int32_t secs_per_cylinder;
992
993		ccg = &ccb->ccg;
994		size_mb = ccg->volume_size /
995		    ((1024L * 1024L) / ccg->block_size);
996		if (size_mb >= (2 * 1024)) {		/* 2GB */
997			ccg->heads = 255;
998			ccg->secs_per_track = 63;
999		} else if (size_mb >= (1 * 1024)) {	/* 1GB */
1000			ccg->heads = 128;
1001			ccg->secs_per_track = 32;
1002		} else {
1003			ccg->heads = 64;
1004			ccg->secs_per_track = 32;
1005		}
1006		secs_per_cylinder = ccg->heads * ccg->secs_per_track;
1007		ccg->cylinders = ccg->volume_size / secs_per_cylinder;
1008
1009		ccb->ccb_h.status = CAM_REQ_CMP;
1010		xpt_done(ccb);
1011		return;
1012	}
1013	case XPT_PATH_INQ:
1014	{
1015		struct ccb_pathinq *cpi = &ccb->cpi;
1016
1017		cpi->version_num = 1;
1018		cpi->target_sprt = 0;
1019		cpi->hba_eng_cnt = 0;
1020		cpi->max_target = camsc->inf->TargetsPerBus;
1021		cpi->max_lun = 8;	/* Per the controller spec */
1022		cpi->initiator_id = camsc->inf->InitiatorBusId;
1023		cpi->bus_id = camsc->inf->BusNumber;
1024#if __FreeBSD_version >= 800000
1025		cpi->maxio = sc->aac_max_sectors << 9;
1026#endif
1027
1028		/*
1029		 * Resetting via the passthrough or parallel bus scan
1030		 * causes problems.
1031		 */
1032		cpi->hba_misc = PIM_NOBUSRESET;
1033		cpi->hba_inquiry = PI_TAG_ABLE;
1034		cpi->base_transfer_speed = 300000;
1035#ifdef CAM_NEW_TRAN_CODE
1036		cpi->hba_misc |= PIM_SEQSCAN;
1037		cpi->protocol = PROTO_SCSI;
1038		cpi->transport = XPORT_SAS;
1039		cpi->transport_version = 0;
1040		cpi->protocol_version = SCSI_REV_SPC2;
1041#endif
1042		strlcpy(cpi->sim_vid, "FreeBSD", SIM_IDLEN);
1043		strlcpy(cpi->hba_vid, "PMC-Sierra", HBA_IDLEN);
1044		strlcpy(cpi->dev_name, cam_sim_name(sim), DEV_IDLEN);
1045		cpi->unit_number = cam_sim_unit(sim);
1046		ccb->ccb_h.status = CAM_REQ_CMP;
1047		xpt_done(ccb);
1048		return;
1049	}
1050	case XPT_GET_TRAN_SETTINGS:
1051	{
1052#ifdef CAM_NEW_TRAN_CODE
1053		struct ccb_trans_settings_scsi *scsi =
1054			&ccb->cts.proto_specific.scsi;
1055		struct ccb_trans_settings_spi *spi =
1056			&ccb->cts.xport_specific.spi;
1057		ccb->cts.protocol = PROTO_SCSI;
1058		ccb->cts.protocol_version = SCSI_REV_SPC2;
1059		ccb->cts.transport = XPORT_SAS;
1060		ccb->cts.transport_version = 0;
1061		scsi->valid = CTS_SCSI_VALID_TQ;
1062		scsi->flags = CTS_SCSI_FLAGS_TAG_ENB;
1063		spi->valid |= CTS_SPI_VALID_DISC;
1064		spi->flags |= CTS_SPI_FLAGS_DISC_ENB;
1065#else
1066		ccb->cts.flags = ~(CCB_TRANS_DISC_ENB | CCB_TRANS_TAG_ENB);
1067		ccb->cts.valid = CCB_TRANS_DISC_VALID | CCB_TRANS_TQ_VALID;
1068#endif
1069		ccb->ccb_h.status = CAM_REQ_CMP;
1070		xpt_done(ccb);
1071		return;
1072	}
1073	case XPT_SET_TRAN_SETTINGS:
1074		ccb->ccb_h.status = CAM_FUNC_NOTAVAIL;
1075		xpt_done(ccb);
1076		return;
1077	case XPT_RESET_BUS:
1078		if (!(sc->flags & AAC_FLAGS_CAM_NORESET) &&
1079			camsc->inf->BusType != CONTAINER_BUS) {
1080			ccb->ccb_h.status = aac_cam_reset_bus(sim, ccb);
1081		} else {
1082			ccb->ccb_h.status = CAM_REQ_CMP;
1083		}
1084		xpt_done(ccb);
1085		return;
1086	case XPT_RESET_DEV:
1087		ccb->ccb_h.status = CAM_REQ_CMP;
1088		xpt_done(ccb);
1089		return;
1090	case XPT_ABORT:
1091		ccb->ccb_h.status = aac_cam_abort_ccb(sim, ccb);
1092		xpt_done(ccb);
1093		return;
1094	case XPT_TERM_IO:
1095		ccb->ccb_h.status = aac_cam_term_io(sim, ccb);
1096		xpt_done(ccb);
1097		return;
1098	default:
1099		device_printf(sc->aac_dev, "Unsupported command 0x%x\n",
1100		    ccb->ccb_h.func_code);
1101		ccb->ccb_h.status = CAM_PROVIDE_FAIL;
1102		xpt_done(ccb);
1103		return;
1104	}
1105
1106	/* Async ops that require communcation with the controller */
1107	if (camsc->inf->BusType == CONTAINER_BUS) {
1108		u_int8_t *cmdp;
1109
1110		if (ccb->ccb_h.flags & CAM_CDB_POINTER)
1111			cmdp = ccb->csio.cdb_io.cdb_ptr;
1112		else
1113			cmdp = &ccb->csio.cdb_io.cdb_bytes[0];
1114
1115		if (*cmdp==READ_6 || *cmdp==WRITE_6 || *cmdp==READ_10 ||
1116			*cmdp==WRITE_10 || *cmdp==READ_12 || *cmdp==WRITE_12 ||
1117			*cmdp==READ_16 || *cmdp==WRITE_16)
1118			aac_container_rw_command(sim, ccb, cmdp);
1119		else
1120			aac_container_special_command(sim, ccb, cmdp);
1121	} else {
1122		aac_passthrough_command(sim, ccb);
1123	}
1124}
1125
1126static void
1127aac_cam_poll(struct cam_sim *sim)
1128{
1129	/*
1130	 * Pinging the interrupt routine isn't very safe, nor is it
1131	 * really necessary.  Do nothing.
1132	 */
1133}
1134
1135static void
1136aac_container_complete(struct aac_command *cm)
1137{
1138	struct	aac_softc *sc;
1139	union	ccb *ccb;
1140	u_int32_t status;
1141
1142	sc = cm->cm_sc;
1143	fwprintf(sc, HBA_FLAGS_DBG_FUNCTION_ENTRY_B, "");
1144	ccb = cm->cm_ccb;
1145	status = ((u_int32_t *)cm->cm_fib->data)[0];
1146
1147	if (cm->cm_flags & AAC_CMD_RESET) {
1148		ccb->ccb_h.status = CAM_SCSI_BUS_RESET;
1149	} else if (status == ST_OK) {
1150		ccb->ccb_h.status = CAM_REQ_CMP;
1151	} else if (status == ST_NOT_READY) {
1152		ccb->ccb_h.status = CAM_BUSY;
1153	} else {
1154		ccb->ccb_h.status = CAM_REQ_CMP_ERR;
1155	}
1156
1157	aacraid_release_command(cm);
1158	xpt_done(ccb);
1159}
1160
1161static void
1162aac_cam_complete(struct aac_command *cm)
1163{
1164	union	ccb *ccb;
1165	struct 	aac_srb_response *srbr;
1166	struct	aac_softc *sc;
1167
1168	sc = cm->cm_sc;
1169	fwprintf(sc, HBA_FLAGS_DBG_FUNCTION_ENTRY_B, "");
1170	ccb = cm->cm_ccb;
1171	srbr = (struct aac_srb_response *)&cm->cm_fib->data[0];
1172
1173	if (cm->cm_flags & AAC_CMD_FASTRESP) {
1174		/* fast response */
1175		srbr->srb_status = CAM_REQ_CMP;
1176		srbr->scsi_status = SCSI_STATUS_OK;
1177		srbr->sense_len = 0;
1178	}
1179
1180	if (cm->cm_flags & AAC_CMD_RESET) {
1181		ccb->ccb_h.status = CAM_SCSI_BUS_RESET;
1182	} else if (srbr->fib_status != 0) {
1183		device_printf(sc->aac_dev, "Passthru FIB failed!\n");
1184		ccb->ccb_h.status = CAM_REQ_ABORTED;
1185	} else {
1186		/*
1187		 * The SRB error codes just happen to match the CAM error
1188		 * codes.  How convienient!
1189		 */
1190		ccb->ccb_h.status = srbr->srb_status;
1191
1192		/* Take care of SCSI_IO ops. */
1193		if (ccb->ccb_h.func_code == XPT_SCSI_IO) {
1194			u_int8_t command, device;
1195
1196			ccb->csio.scsi_status = srbr->scsi_status;
1197
1198			/* Take care of autosense */
1199			if (srbr->sense_len) {
1200				int sense_len, scsi_sense_len;
1201
1202				scsi_sense_len = sizeof(struct scsi_sense_data);
1203				bzero(&ccb->csio.sense_data, scsi_sense_len);
1204				sense_len = (srbr->sense_len >
1205				    scsi_sense_len) ? scsi_sense_len :
1206				    srbr->sense_len;
1207				bcopy(&srbr->sense[0], &ccb->csio.sense_data,
1208				    srbr->sense_len);
1209				ccb->csio.sense_len = sense_len;
1210				ccb->ccb_h.status |= CAM_AUTOSNS_VALID;
1211				// scsi_sense_print(&ccb->csio);
1212			}
1213
1214			/* If this is an inquiry command, fake things out */
1215			if (ccb->ccb_h.flags & CAM_CDB_POINTER)
1216				command = ccb->csio.cdb_io.cdb_ptr[0];
1217			else
1218				command = ccb->csio.cdb_io.cdb_bytes[0];
1219
1220			if (command == INQUIRY) {
1221				if (ccb->ccb_h.status == CAM_REQ_CMP) {
1222				  device = ccb->csio.data_ptr[0] & 0x1f;
1223				  /*
1224				   * We want DASD and PROC devices to only be
1225				   * visible through the pass device.
1226				   */
1227				  if ((device == T_DIRECT &&
1228				    !(sc->aac_feature_bits & AAC_SUPPL_SUPPORTED_JBOD)) ||
1229				    (device == T_PROCESSOR))
1230				    ccb->csio.data_ptr[0] =
1231				  	((device & 0xe0) | T_NODEVICE);
1232
1233				  /* handle phys. components of a log. drive */
1234				  if (ccb->csio.data_ptr[0] & 0x20) {
1235					if (sc->hint_flags & 8) {
1236					  /* expose phys. device (daXX) */
1237					  ccb->csio.data_ptr[0] &= 0xdf;
1238					} else {
1239					  /* phys. device only visible through pass device (passXX) */
1240					  ccb->csio.data_ptr[0] |= 0x10;
1241					}
1242				  }
1243				} else if (ccb->ccb_h.status == CAM_SEL_TIMEOUT &&
1244				  ccb->ccb_h.target_lun != 0) {
1245				  /* fix for INQUIRYs on Lun>0 */
1246				  ccb->ccb_h.status = CAM_DEV_NOT_THERE;
1247				}
1248			}
1249		}
1250	}
1251
1252	aacraid_release_command(cm);
1253	xpt_done(ccb);
1254}
1255
1256static u_int32_t
1257aac_cam_reset_bus(struct cam_sim *sim, union ccb *ccb)
1258{
1259	struct aac_command *cm;
1260	struct aac_fib *fib;
1261	struct aac_softc *sc;
1262	struct aac_cam *camsc;
1263	struct aac_vmioctl *vmi;
1264	struct aac_resetbus *rbc;
1265	u_int32_t rval;
1266
1267	camsc = (struct aac_cam *)cam_sim_softc(sim);
1268	sc = camsc->inf->aac_sc;
1269
1270	if (sc == NULL) {
1271		printf("aac: Null sc?\n");
1272		return (CAM_REQ_ABORTED);
1273	}
1274
1275	if (aacraid_alloc_command(sc, &cm)) {
1276		struct aac_event *event;
1277
1278		xpt_freeze_simq(sim, 1);
1279		ccb->ccb_h.status = CAM_RESRC_UNAVAIL;
1280		ccb->ccb_h.sim_priv.entries[0].ptr = camsc;
1281		event = malloc(sizeof(struct aac_event), M_AACRAIDCAM,
1282			M_NOWAIT | M_ZERO);
1283		if (event == NULL) {
1284			device_printf(sc->aac_dev,
1285				"Warning, out of memory for event\n");
1286			return (CAM_REQ_ABORTED);
1287		}
1288		event->ev_callback = aac_cam_event;
1289		event->ev_arg = ccb;
1290		event->ev_type = AAC_EVENT_CMFREE;
1291		aacraid_add_event(sc, event);
1292		return (CAM_REQ_ABORTED);
1293	}
1294
1295	fib = cm->cm_fib;
1296	cm->cm_timestamp = time_uptime;
1297	cm->cm_datalen = 0;
1298
1299	fib->Header.Size =
1300		sizeof(struct aac_fib_header) + sizeof(struct aac_vmioctl);
1301	fib->Header.XferState =
1302		AAC_FIBSTATE_HOSTOWNED   |
1303		AAC_FIBSTATE_INITIALISED |
1304		AAC_FIBSTATE_EMPTY	 |
1305		AAC_FIBSTATE_FROMHOST	 |
1306		AAC_FIBSTATE_REXPECTED   |
1307		AAC_FIBSTATE_NORM	 |
1308		AAC_FIBSTATE_ASYNC	 |
1309		AAC_FIBSTATE_FAST_RESPONSE;
1310	fib->Header.Command = ContainerCommand;
1311
1312	vmi = (struct aac_vmioctl *)&fib->data[0];
1313	bzero(vmi, sizeof(struct aac_vmioctl));
1314
1315	vmi->Command = VM_Ioctl;
1316	vmi->ObjType = FT_DRIVE;
1317	vmi->MethId = sc->scsi_method_id;
1318	vmi->ObjId = 0;
1319	vmi->IoctlCmd = ResetBus;
1320
1321	rbc = (struct aac_resetbus *)&vmi->IoctlBuf[0];
1322	rbc->BusNumber = camsc->inf->BusNumber - 1;
1323
1324	if (aacraid_wait_command(cm) != 0) {
1325		device_printf(sc->aac_dev,"Error sending ResetBus command\n");
1326		rval = CAM_REQ_ABORTED;
1327	} else {
1328		rval = CAM_REQ_CMP;
1329	}
1330	aacraid_release_command(cm);
1331	return (rval);
1332}
1333
1334static u_int32_t
1335aac_cam_abort_ccb(struct cam_sim *sim, union ccb *ccb)
1336{
1337	return (CAM_UA_ABORT);
1338}
1339
1340static u_int32_t
1341aac_cam_term_io(struct cam_sim *sim, union ccb *ccb)
1342{
1343	return (CAM_UA_TERMIO);
1344}
1345
1346static int
1347aac_load_map_command_sg(struct aac_softc *sc, struct aac_command *cm)
1348{
1349	int error;
1350
1351	fwprintf(sc, HBA_FLAGS_DBG_FUNCTION_ENTRY_B, "");
1352	error = bus_dmamap_load(sc->aac_buffer_dmat,
1353				cm->cm_datamap, cm->cm_data, cm->cm_datalen,
1354				aacraid_map_command_sg, cm, 0);
1355	if (error == EINPROGRESS) {
1356		fwprintf(sc, HBA_FLAGS_DBG_INIT_B, "freezing queue\n");
1357		sc->flags |= AAC_QUEUE_FRZN;
1358		error = 0;
1359	} else if (error != 0) {
1360		panic("aac_load_map_command_sg: unexpected error %d from "
1361	     		"busdma", error);
1362	}
1363	return(error);
1364}
1365
1366/*
1367 * Start as much queued I/O as possible on the controller
1368 */
1369void
1370aacraid_startio(struct aac_softc *sc)
1371{
1372	struct aac_command *cm;
1373
1374	fwprintf(sc, HBA_FLAGS_DBG_FUNCTION_ENTRY_B, "");
1375
1376	for (;;) {
1377		if (sc->aac_state & AAC_STATE_RESET) {
1378			fwprintf(sc, HBA_FLAGS_DBG_ERROR_B, "AAC_STATE_RESET");
1379			break;
1380		}
1381		/*
1382		 * This flag might be set if the card is out of resources.
1383		 * Checking it here prevents an infinite loop of deferrals.
1384		 */
1385		if (sc->flags & AAC_QUEUE_FRZN) {
1386			fwprintf(sc, HBA_FLAGS_DBG_ERROR_B, "AAC_QUEUE_FRZN");
1387			break;
1388		}
1389
1390		/*
1391		 * Try to get a command that's been put off for lack of
1392		 * resources
1393		 */
1394		if (sc->flags & AAC_FLAGS_SYNC_MODE) {
1395			/* sync. transfer mode */
1396			if (sc->aac_sync_cm)
1397				break;
1398			cm = aac_dequeue_ready(sc);
1399			sc->aac_sync_cm = cm;
1400		} else {
1401			cm = aac_dequeue_ready(sc);
1402		}
1403
1404		/* nothing to do? */
1405		if (cm == NULL)
1406			break;
1407
1408		/* don't map more than once */
1409		if (cm->cm_flags & AAC_CMD_MAPPED)
1410			panic("aac: command %p already mapped", cm);
1411
1412		/*
1413		 * Set up the command to go to the controller.  If there are no
1414		 * data buffers associated with the command then it can bypass
1415		 * busdma.
1416		 */
1417		if (cm->cm_datalen)
1418			aac_load_map_command_sg(sc, cm);
1419		else
1420			aacraid_map_command_sg(cm, NULL, 0, 0);
1421	}
1422}
1423