2/* 3 * Platform (FreeBSD) dependent common attachment code for Qlogic adapters. 4 * 5 * Copyright (c) 1997, 1998, 1999, 2000 by Matthew Jacob 6 * 7 * Redistribution and use in source and binary forms, with or without 8 * modification, are permitted provided that the following conditions 9 * are met: 10 * 1. Redistributions of source code must retain the above copyright 11 * notice immediately at the beginning of the file, without modification, 12 * this list of conditions, and the following disclaimer. 13 * 2. The name of the author may not be used to endorse or promote products 14 * derived from this software without specific prior written permission. 15 * 16 * THIS SOFTWARE IS PROVIDED BY THE AUTHOR AND CONTRIBUTORS ``AS IS'' AND 17 * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE 18 * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE 19 * ARE DISCLAIMED. IN NO EVENT SHALL THE AUTHOR OR CONTRIBUTORS BE LIABLE FOR 20 * ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL 21 * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS 22 * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) 23 * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT 24 * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY 25 * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF 26 * SUCH DAMAGE. 27 */ 28#include <dev/isp/isp_freebsd.h> 29#include <machine/stdarg.h> /* for use by isp_prt below */ 30 31static void isp_intr_enable(void *); 32static void isp_cam_async(void *, u_int32_t, struct cam_path *, void *); 33static void isp_poll(struct cam_sim *); 34static void isp_relsim(void *); 35static timeout_t isp_watchdog; 36static void isp_action(struct cam_sim *, union ccb *); 37 38 39static struct ispsoftc *isplist = NULL; 40 41void 42isp_attach(struct ispsoftc *isp) 43{ 44 int primary, secondary; 45 struct ccb_setasync csa; 46 struct cam_devq *devq; 47 struct cam_sim *sim; 48 struct cam_path *path; 49 50 /* 51 * Establish (in case of 12X0) which bus is the primary. 52 */ 53 54 primary = 0; 55 secondary = 1; 56 57 /* 58 * Create the device queue for our SIM(s). 59 */ 60 devq = cam_simq_alloc(isp->isp_maxcmds); 61 if (devq == NULL) { 62 return; 63 } 64 65 /* 66 * Construct our SIM entry. 67 */ 68 sim = cam_sim_alloc(isp_action, isp_poll, "isp", isp, 69 isp->isp_unit, 1, isp->isp_maxcmds, devq); 70 if (sim == NULL) { 71 cam_simq_free(devq); 72 return; 73 } 74 75 isp->isp_osinfo.ehook.ich_func = isp_intr_enable; 76 isp->isp_osinfo.ehook.ich_arg = isp; 77 if (config_intrhook_establish(&isp->isp_osinfo.ehook) != 0) { 78 isp_prt(isp, ISP_LOGERR, 79 "could not establish interrupt enable hook"); 80 cam_sim_free(sim, TRUE); 81 return; 82 } 83 84 if (xpt_bus_register(sim, primary) != CAM_SUCCESS) { 85 cam_sim_free(sim, TRUE); 86 return; 87 } 88 89 if (xpt_create_path(&path, NULL, cam_sim_path(sim), 90 CAM_TARGET_WILDCARD, CAM_LUN_WILDCARD) != CAM_REQ_CMP) { 91 xpt_bus_deregister(cam_sim_path(sim)); 92 cam_sim_free(sim, TRUE); 93 return; 94 } 95 96 xpt_setup_ccb(&csa.ccb_h, path, 5); 97 csa.ccb_h.func_code = XPT_SASYNC_CB; 98 csa.event_enable = AC_LOST_DEVICE; 99 csa.callback = isp_cam_async; 100 csa.callback_arg = sim; 101 xpt_action((union ccb *)&csa); 102 isp->isp_sim = sim; 103 isp->isp_path = path; 104 105 /* 106 * If we have a second channel, construct SIM entry for that. 107 */ 108 if (IS_DUALBUS(isp)) { 109 sim = cam_sim_alloc(isp_action, isp_poll, "isp", isp, 110 isp->isp_unit, 1, isp->isp_maxcmds, devq); 111 if (sim == NULL) { 112 xpt_bus_deregister(cam_sim_path(isp->isp_sim)); 113 xpt_free_path(isp->isp_path); 114 cam_simq_free(devq); 115 return; 116 } 117 if (xpt_bus_register(sim, secondary) != CAM_SUCCESS) { 118 xpt_bus_deregister(cam_sim_path(isp->isp_sim)); 119 xpt_free_path(isp->isp_path); 120 cam_sim_free(sim, TRUE); 121 return; 122 } 123 124 if (xpt_create_path(&path, NULL, cam_sim_path(sim), 125 CAM_TARGET_WILDCARD, CAM_LUN_WILDCARD) != CAM_REQ_CMP) { 126 xpt_bus_deregister(cam_sim_path(isp->isp_sim)); 127 xpt_free_path(isp->isp_path); 128 xpt_bus_deregister(cam_sim_path(sim)); 129 cam_sim_free(sim, TRUE); 130 return; 131 } 132 133 xpt_setup_ccb(&csa.ccb_h, path, 5); 134 csa.ccb_h.func_code = XPT_SASYNC_CB; 135 csa.event_enable = AC_LOST_DEVICE; 136 csa.callback = isp_cam_async; 137 csa.callback_arg = sim; 138 xpt_action((union ccb *)&csa); 139 isp->isp_sim2 = sim; 140 isp->isp_path2 = path; 141 } 142 isp->isp_state = ISP_RUNSTATE; 143 ENABLE_INTS(isp); 144 if (isplist == NULL) { 145 isplist = isp; 146 } else { 147 struct ispsoftc *tmp = isplist; 148 while (tmp->isp_osinfo.next) { 149 tmp = tmp->isp_osinfo.next; 150 } 151 tmp->isp_osinfo.next = isp; 152 } 153} 154 155static void 156isp_intr_enable(void *arg) 157{ 158 struct ispsoftc *isp = arg; 159 ENABLE_INTS(isp); 160 isp->isp_osinfo.intsok = 1; 161 /* Release our hook so that the boot can continue. */ 162 config_intrhook_disestablish(&isp->isp_osinfo.ehook); 163} 164 165/* 166 * Put the target mode functions here, because some are inlines 167 */ 168 169#ifdef ISP_TARGET_MODE 170 171static __inline int is_lun_enabled(struct ispsoftc *, lun_id_t); 172static __inline int are_any_luns_enabled(struct ispsoftc *); 173static __inline tstate_t *get_lun_statep(struct ispsoftc *, lun_id_t); 174static __inline void rls_lun_statep(struct ispsoftc *, tstate_t *); 175static __inline int isp_psema_sig_rqe(struct ispsoftc *); 176static __inline int isp_cv_wait_timed_rqe(struct ispsoftc *, int); 177static __inline void isp_cv_signal_rqe(struct ispsoftc *, int); 178static __inline void isp_vsema_rqe(struct ispsoftc *); 179static cam_status 180create_lun_state(struct ispsoftc *, struct cam_path *, tstate_t **); 181static void destroy_lun_state(struct ispsoftc *, tstate_t *); 182static void isp_en_lun(struct ispsoftc *, union ccb *); 183static cam_status isp_abort_tgt_ccb(struct ispsoftc *, union ccb *); 184static cam_status isp_target_start_ctio(struct ispsoftc *, union ccb *); 185static cam_status isp_target_putback_atio(struct ispsoftc *, union ccb *); 186static timeout_t isp_refire_putback_atio; 187 188static int isp_handle_platform_atio(struct ispsoftc *, at_entry_t *); 189static int isp_handle_platform_atio2(struct ispsoftc *, at2_entry_t *); 190static int isp_handle_platform_ctio(struct ispsoftc *, void *); 191static void isp_handle_platform_ctio_part2(struct ispsoftc *, union ccb *); 192 193static __inline int 194is_lun_enabled(struct ispsoftc *isp, lun_id_t lun) 195{ 196 tstate_t *tptr; 197 ISP_LOCK(isp); 198 if ((tptr = isp->isp_osinfo.lun_hash[LUN_HASH_FUNC(lun)]) == NULL) { 199 ISP_UNLOCK(isp); 200 return (0); 201 } 202 do { 203 if (tptr->lun == (lun_id_t) lun) { 204 ISP_UNLOCK(isp); 205 return (1); 206 } 207 } while ((tptr = tptr->next) != NULL); 208 ISP_UNLOCK(isp); 209 return (0); 210} 211 212static __inline int 213are_any_luns_enabled(struct ispsoftc *isp) 214{ 215 int i; 216 for (i = 0; i < LUN_HASH_SIZE; i++) { 217 if (isp->isp_osinfo.lun_hash[i]) { 218 return (1); 219 } 220 } 221 return (0); 222} 223 224static __inline tstate_t * 225get_lun_statep(struct ispsoftc *isp, lun_id_t lun) 226{ 227 tstate_t *tptr; 228 229 ISP_LOCK(isp); 230 if (lun == CAM_LUN_WILDCARD) { 231 tptr = &isp->isp_osinfo.tsdflt; 232 tptr->hold++; 233 ISP_UNLOCK(isp); 234 return (tptr); 235 } else { 236 tptr = isp->isp_osinfo.lun_hash[LUN_HASH_FUNC(lun)]; 237 } 238 if (tptr == NULL) { 239 ISP_UNLOCK(isp); 240 return (NULL); 241 } 242 243 do { 244 if (tptr->lun == lun) { 245 tptr->hold++; 246 ISP_UNLOCK(isp); 247 return (tptr); 248 } 249 } while ((tptr = tptr->next) != NULL); 250 ISP_UNLOCK(isp); 251 return (tptr); 252} 253 254static __inline void 255rls_lun_statep(struct ispsoftc *isp, tstate_t *tptr) 256{ 257 if (tptr->hold) 258 tptr->hold--; 259} 260 261static __inline int 262isp_psema_sig_rqe(struct ispsoftc *isp) 263{ 264 ISP_LOCK(isp); 265 while (isp->isp_osinfo.tmflags & TM_BUSY) { 266 isp->isp_osinfo.tmflags |= TM_WANTED; 267 if (tsleep(&isp->isp_osinfo.tmflags, PRIBIO|PCATCH, "i0", 0)) { 268 ISP_UNLOCK(isp); 269 return (-1); 270 } 271 isp->isp_osinfo.tmflags |= TM_BUSY; 272 } 273 ISP_UNLOCK(isp); 274 return (0); 275} 276 277static __inline int 278isp_cv_wait_timed_rqe(struct ispsoftc *isp, int timo) 279{ 280 ISP_LOCK(isp); 281 if (tsleep(&isp->isp_osinfo.rstatus, PRIBIO, "qt1", timo)) { 282 ISP_UNLOCK(isp); 283 return (-1); 284 } 285 ISP_UNLOCK(isp); 286 return (0); 287} 288 289static __inline void 290isp_cv_signal_rqe(struct ispsoftc *isp, int status) 291{ 292 isp->isp_osinfo.rstatus = status; 293 wakeup(&isp->isp_osinfo.rstatus); 294} 295 296static __inline void 297isp_vsema_rqe(struct ispsoftc *isp) 298{ 299 ISP_LOCK(isp); 300 if (isp->isp_osinfo.tmflags & TM_WANTED) { 301 isp->isp_osinfo.tmflags &= ~TM_WANTED; 302 wakeup(&isp->isp_osinfo.tmflags); 303 } 304 isp->isp_osinfo.tmflags &= ~TM_BUSY; 305 ISP_UNLOCK(isp); 306} 307 308static cam_status 309create_lun_state(struct ispsoftc *isp, struct cam_path *path, tstate_t **rslt) 310{ 311 cam_status status; 312 lun_id_t lun; 313 tstate_t *tptr, *new; 314 315 lun = xpt_path_lun_id(path); 316 if (lun < 0) { 317 return (CAM_LUN_INVALID); 318 } 319 if (is_lun_enabled(isp, lun)) { 320 return (CAM_LUN_ALRDY_ENA); 321 } 322 new = (tstate_t *) malloc(sizeof (tstate_t), M_DEVBUF, M_NOWAIT|M_ZERO); 323 if (new == NULL) { 324 return (CAM_RESRC_UNAVAIL); 325 } 326 327 status = xpt_create_path(&new->owner, NULL, xpt_path_path_id(path), 328 xpt_path_target_id(path), xpt_path_lun_id(path)); 329 if (status != CAM_REQ_CMP) { 330 free(new, M_DEVBUF); 331 return (status); 332 } 333 new->lun = lun; 334 SLIST_INIT(&new->atios); 335 SLIST_INIT(&new->inots); 336 new->hold = 1; 337 338 ISP_LOCK(isp); 339 if ((tptr = isp->isp_osinfo.lun_hash[LUN_HASH_FUNC(lun)]) == NULL) { 340 isp->isp_osinfo.lun_hash[LUN_HASH_FUNC(lun)] = new; 341 } else { 342 while (tptr->next) 343 tptr = tptr->next; 344 tptr->next = new; 345 } 346 ISP_UNLOCK(isp); 347 *rslt = new; 348 return (CAM_REQ_CMP); 349} 350 351static __inline void 352destroy_lun_state(struct ispsoftc *isp, tstate_t *tptr) 353{ 354 tstate_t *lw, *pw; 355 356 ISP_LOCK(isp); 357 if (tptr->hold) { 358 ISP_UNLOCK(isp); 359 return; 360 } 361 pw = isp->isp_osinfo.lun_hash[LUN_HASH_FUNC(tptr->lun)]; 362 if (pw == NULL) { 363 ISP_UNLOCK(isp); 364 return; 365 } else if (pw->lun == tptr->lun) { 366 isp->isp_osinfo.lun_hash[LUN_HASH_FUNC(tptr->lun)] = pw->next; 367 } else { 368 lw = pw; 369 pw = lw->next; 370 while (pw) { 371 if (pw->lun == tptr->lun) { 372 lw->next = pw->next; 373 break; 374 } 375 lw = pw; 376 pw = pw->next; 377 } 378 if (pw == NULL) { 379 ISP_UNLOCK(isp); 380 return; 381 } 382 } 383 free(tptr, M_DEVBUF); 384 ISP_UNLOCK(isp); 385} 386 387static void 388isp_en_lun(struct ispsoftc *isp, union ccb *ccb) 389{ 390 const char *lfmt = "Lun now %sabled for target mode\n"; 391 struct ccb_en_lun *cel = &ccb->cel; 392 tstate_t *tptr; 393 u_int16_t rstat; 394 int bus, frozen = 0; 395 lun_id_t lun; 396 target_id_t tgt; 397 398 399 bus = XS_CHANNEL(ccb); 400 tgt = ccb->ccb_h.target_id; 401 lun = ccb->ccb_h.target_lun; 402 403 /* 404 * Do some sanity checking first. 405 */ 406 407 if (lun < 0 || lun >= (lun_id_t) isp->isp_maxluns) { 408 ccb->ccb_h.status = CAM_LUN_INVALID; 409 return; 410 } 411 if (IS_SCSI(isp)) { 412 if (tgt != CAM_TARGET_WILDCARD && 413 tgt != ((sdparam *) isp->isp_param)->isp_initiator_id) { 414 ccb->ccb_h.status = CAM_TID_INVALID; 415 return; 416 } 417 } else { 418 if (tgt != CAM_TARGET_WILDCARD && 419 tgt != ((fcparam *) isp->isp_param)->isp_loopid) { 420 ccb->ccb_h.status = CAM_TID_INVALID; 421 return; 422 } 423 } 424 425 /* 426 * If Fibre Channel, stop and drain all activity to this bus. 427 */ 428 if (IS_FC(isp)) { 429 ISP_LOCK(isp); 430 frozen = 1; 431 xpt_freeze_simq(isp->isp_sim, 1); 432 isp->isp_osinfo.drain = 1; 433 /* ISP_UNLOCK(isp); XXX NEED CV_WAIT HERE XXX */ 434 while (isp->isp_osinfo.drain) { 435 tsleep(&isp->isp_osinfo.drain, PRIBIO, "ispdrain", 0); 436 } 437 ISP_UNLOCK(isp); 438 } 439 440 /* 441 * Check to see if we're enabling on fibre channel and 442 * don't yet have a notion of who the heck we are (no 443 * loop yet). 444 */ 445 if (IS_FC(isp) && cel->enable && 446 (isp->isp_osinfo.tmflags & TM_TMODE_ENABLED) == 0) { 447 int rv= 2 * 1000000; 448 fcparam *fcp = isp->isp_param; 449 450 ISP_LOCK(isp); 451 rv = isp_control(isp, ISPCTL_FCLINK_TEST, &rv); 452 ISP_UNLOCK(isp); 453 if (rv || fcp->isp_fwstate != FW_READY) { 454 xpt_print_path(ccb->ccb_h.path); 455 printf("link status not good yet\n"); 456 ccb->ccb_h.status = CAM_REQ_CMP_ERR; 457 if (frozen) 458 xpt_release_simq(isp->isp_sim, 1); 459 return; 460 } 461 ISP_LOCK(isp); 462 rv = isp_control(isp, ISPCTL_PDB_SYNC, NULL); 463 ISP_UNLOCK(isp); 464 if (rv || fcp->isp_fwstate != FW_READY) { 465 xpt_print_path(ccb->ccb_h.path); 466 printf("could not get a good port database read\n"); 467 ccb->ccb_h.status = CAM_REQ_CMP_ERR; 468 if (frozen) 469 xpt_release_simq(isp->isp_sim, 1); 470 return; 471 } 472 } 473 474 475 /* 476 * Next check to see whether this is a target/lun wildcard action. 477 * 478 * If so, we enable/disable target mode but don't do any lun enabling. 479 */ 480 if (lun == CAM_LUN_WILDCARD && tgt == CAM_TARGET_WILDCARD) { 481 int av; 482 tptr = &isp->isp_osinfo.tsdflt; 483 if (cel->enable) { 484 if (isp->isp_osinfo.tmflags & TM_TMODE_ENABLED) { 485 ccb->ccb_h.status = CAM_LUN_ALRDY_ENA; 486 if (frozen) 487 xpt_release_simq(isp->isp_sim, 1); 488 return; 489 } 490 ccb->ccb_h.status = 491 xpt_create_path(&tptr->owner, NULL, 492 xpt_path_path_id(ccb->ccb_h.path), 493 xpt_path_target_id(ccb->ccb_h.path), 494 xpt_path_lun_id(ccb->ccb_h.path)); 495 if (ccb->ccb_h.status != CAM_REQ_CMP) { 496 if (frozen) 497 xpt_release_simq(isp->isp_sim, 1); 498 return; 499 } 500 SLIST_INIT(&tptr->atios); 501 SLIST_INIT(&tptr->inots); 502 av = 1; 503 ISP_LOCK(isp); 504 av = isp_control(isp, ISPCTL_TOGGLE_TMODE, &av); 505 if (av) { 506 ccb->ccb_h.status = CAM_FUNC_NOTAVAIL; 507 xpt_free_path(tptr->owner); 508 ISP_UNLOCK(isp); 509 if (frozen) 510 xpt_release_simq(isp->isp_sim, 1); 511 return; 512 } 513 isp->isp_osinfo.tmflags |= TM_TMODE_ENABLED; 514 ISP_UNLOCK(isp); 515 } else { 516 if ((isp->isp_osinfo.tmflags & TM_TMODE_ENABLED) == 0) { 517 ccb->ccb_h.status = CAM_LUN_INVALID; 518 if (frozen) 519 xpt_release_simq(isp->isp_sim, 1); 520 return; 521 } 522 if (are_any_luns_enabled(isp)) { 523 ccb->ccb_h.status = CAM_SCSI_BUSY; 524 if (frozen) 525 xpt_release_simq(isp->isp_sim, 1); 526 return; 527 } 528 av = 0; 529 ISP_LOCK(isp); 530 av = isp_control(isp, ISPCTL_TOGGLE_TMODE, &av); 531 if (av) { 532 ccb->ccb_h.status = CAM_FUNC_NOTAVAIL; 533 ISP_UNLOCK(isp); 534 if (frozen) 535 xpt_release_simq(isp->isp_sim, 1); 536 return; 537 } 538 isp->isp_osinfo.tmflags &= ~TM_TMODE_ENABLED; 539 ISP_UNLOCK(isp); 540 ccb->ccb_h.status = CAM_REQ_CMP; 541 } 542 xpt_print_path(ccb->ccb_h.path); 543 printf(lfmt, (cel->enable) ? "en" : "dis"); 544 if (frozen) 545 xpt_release_simq(isp->isp_sim, 1); 546 return; 547 } 548 549 /* 550 * We can move along now... 551 */ 552 553 if (frozen) 554 xpt_release_simq(isp->isp_sim, 1); 555 556 557 if (cel->enable) { 558 ccb->ccb_h.status = 559 create_lun_state(isp, ccb->ccb_h.path, &tptr); 560 if (ccb->ccb_h.status != CAM_REQ_CMP) { 561 return; 562 } 563 } else { 564 tptr = get_lun_statep(isp, lun); 565 if (tptr == NULL) { 566 ccb->ccb_h.status = CAM_LUN_INVALID; 567 return; 568 } 569 } 570 571 if (isp_psema_sig_rqe(isp)) { 572 rls_lun_statep(isp, tptr); 573 if (cel->enable) 574 destroy_lun_state(isp, tptr); 575 ccb->ccb_h.status = CAM_REQ_CMP_ERR; 576 return; 577 } 578 579 ISP_LOCK(isp); 580 if (cel->enable) { 581 u_int32_t seq = isp->isp_osinfo.rollinfo++; 582 rstat = LUN_ERR; 583 if (isp_lun_cmd(isp, RQSTYPE_ENABLE_LUN, bus, tgt, lun, seq)) { 584 xpt_print_path(ccb->ccb_h.path); 585 printf("isp_lun_cmd failed\n"); 586 goto out; 587 } 588 if (isp_cv_wait_timed_rqe(isp, 30 * hz)) { 589 xpt_print_path(ccb->ccb_h.path); 590 printf("wait for ENABLE LUN timed out\n"); 591 goto out; 592 } 593 rstat = isp->isp_osinfo.rstatus; 594 if (rstat != LUN_OK) { 595 xpt_print_path(ccb->ccb_h.path); 596 printf("ENABLE LUN returned 0x%x\n", rstat); 597 goto out; 598 } 599 } else { 600 u_int32_t seq; 601 602 seq = isp->isp_osinfo.rollinfo++; 603 rstat = LUN_ERR; 604 605 if (isp_lun_cmd(isp, -RQSTYPE_MODIFY_LUN, bus, tgt, lun, seq)) { 606 xpt_print_path(ccb->ccb_h.path); 607 printf("isp_lun_cmd failed\n"); 608 goto out; 609 } 610 if (isp_cv_wait_timed_rqe(isp, 30 * hz)) { 611 xpt_print_path(ccb->ccb_h.path); 612 printf("wait for MODIFY LUN timed out\n"); 613 goto out; 614 } 615 rstat = isp->isp_osinfo.rstatus; 616 if (rstat != LUN_OK) { 617 xpt_print_path(ccb->ccb_h.path); 618 printf("MODIFY LUN returned 0x%x\n", rstat); 619 goto out; 620 } 621 rstat = LUN_ERR; 622 seq = isp->isp_osinfo.rollinfo++; 623 624 if (isp_lun_cmd(isp, -RQSTYPE_ENABLE_LUN, bus, tgt, lun, seq)) { 625 xpt_print_path(ccb->ccb_h.path); 626 printf("isp_lun_cmd failed\n"); 627 goto out; 628 } 629 if (isp_cv_wait_timed_rqe(isp, 30 * hz)) { 630 xpt_print_path(ccb->ccb_h.path); 631 printf("wait for ENABLE LUN timed out\n"); 632 goto out; 633 } 634 rstat = isp->isp_osinfo.rstatus; 635 if (rstat != LUN_OK) { 636 xpt_print_path(ccb->ccb_h.path); 637 printf("ENABLE LUN returned 0x%x\n", rstat); 638 goto out; 639 } 640 } 641out: 642 isp_vsema_rqe(isp); 643 ISP_UNLOCK(isp); 644 645 if (rstat != LUN_OK) { 646 xpt_print_path(ccb->ccb_h.path); 647 printf("lun %sable failed\n", (cel->enable) ? "en" : "dis"); 648 ccb->ccb_h.status = CAM_REQ_CMP_ERR; 649 rls_lun_statep(isp, tptr); 650 if (cel->enable) 651 destroy_lun_state(isp, tptr); 652 } else { 653 xpt_print_path(ccb->ccb_h.path); 654 printf(lfmt, (cel->enable) ? "en" : "dis"); 655 rls_lun_statep(isp, tptr); 656 if (cel->enable == 0) { 657 destroy_lun_state(isp, tptr); 658 } 659 ccb->ccb_h.status = CAM_REQ_CMP; 660 } 661} 662 663static cam_status 664isp_abort_tgt_ccb(struct ispsoftc *isp, union ccb *ccb) 665{ 666 tstate_t *tptr; 667 struct ccb_hdr_slist *lp; 668 struct ccb_hdr *curelm; 669 int found; 670 union ccb *accb = ccb->cab.abort_ccb; 671 672 if (accb->ccb_h.target_id != CAM_TARGET_WILDCARD) { 673 if (IS_FC(isp) && (accb->ccb_h.target_id != 674 ((fcparam *) isp->isp_param)->isp_loopid)) { 675 return (CAM_PATH_INVALID); 676 } else if (IS_SCSI(isp) && (accb->ccb_h.target_id != 677 ((sdparam *) isp->isp_param)->isp_initiator_id)) { 678 return (CAM_PATH_INVALID); 679 } 680 } 681 tptr = get_lun_statep(isp, accb->ccb_h.target_lun); 682 if (tptr == NULL) { 683 return (CAM_PATH_INVALID); 684 } 685 if (accb->ccb_h.func_code == XPT_ACCEPT_TARGET_IO) { 686 lp = &tptr->atios; 687 } else if (accb->ccb_h.func_code == XPT_IMMED_NOTIFY) { 688 lp = &tptr->inots; 689 } else { 690 rls_lun_statep(isp, tptr); 691 return (CAM_UA_ABORT); 692 } 693 curelm = SLIST_FIRST(lp); 694 found = 0; 695 if (curelm == &accb->ccb_h) { 696 found = 1; 697 SLIST_REMOVE_HEAD(lp, sim_links.sle); 698 } else { 699 while(curelm != NULL) { 700 struct ccb_hdr *nextelm; 701 702 nextelm = SLIST_NEXT(curelm, sim_links.sle); 703 if (nextelm == &accb->ccb_h) { 704 found = 1; 705 SLIST_NEXT(curelm, sim_links.sle) = 706 SLIST_NEXT(nextelm, sim_links.sle); 707 break; 708 } 709 curelm = nextelm; 710 } 711 } 712 rls_lun_statep(isp, tptr); 713 if (found) { 714 accb->ccb_h.status = CAM_REQ_ABORTED; 715 return (CAM_REQ_CMP); 716 } 717 return(CAM_PATH_INVALID); 718} 719 720static cam_status 721isp_target_start_ctio(struct ispsoftc *isp, union ccb *ccb) 722{ 723 void *qe; 724 struct ccb_scsiio *cso = &ccb->csio; 725 u_int32_t *hp, save_handle; 726 u_int16_t iptr, optr; 727 728 729 if (isp_getrqentry(isp, &iptr, &optr, &qe)) { 730 xpt_print_path(ccb->ccb_h.path); 731 printf("Request Queue Overflow in isp_target_start_ctio\n"); 732 return (CAM_RESRC_UNAVAIL); 733 } 734 bzero(qe, QENTRY_LEN); 735 736 /* 737 * We're either moving data or completing a command here. 738 */ 739 740 if (IS_FC(isp)) { 741 struct ccb_accept_tio *atiop; 742 ct2_entry_t *cto = qe; 743 744 cto->ct_header.rqs_entry_type = RQSTYPE_CTIO2; 745 cto->ct_header.rqs_entry_count = 1; 746 cto->ct_iid = cso->init_id; 747 if (isp->isp_maxluns <= 16) { 748 cto->ct_lun = ccb->ccb_h.target_lun; 749 } 750 /* 751 * Start with a residual based on what the original datalength 752 * was supposed to be. Basically, we ignore what CAM has set 753 * for residuals. The data transfer routines will knock off 754 * the residual for each byte actually moved- and also will 755 * be responsible for setting the underrun flag. 756 */ 757 /* HACK! HACK! */ 758 if ((atiop = ccb->ccb_h.periph_priv.entries[1].ptr) != NULL) { 759 cto->ct_resid = atiop->ccb_h.spriv_field0; 760 } 761 762 /* 763 * We always have to use the tag_id- it has the RX_ID 764 * for this exchage. 765 */ 766 cto->ct_rxid = cso->tag_id; 767 if (cso->dxfer_len == 0) { 768 cto->ct_flags |= CT2_FLAG_MODE1 | CT2_NO_DATA; 769 if (ccb->ccb_h.flags & CAM_SEND_STATUS) { 770 cto->ct_flags |= CT2_SENDSTATUS; 771 cto->rsp.m1.ct_scsi_status = cso->scsi_status; 772 } 773 if ((ccb->ccb_h.flags & CAM_SEND_SENSE) != 0) { 774 int m = min(cso->sense_len, MAXRESPLEN); 775 bcopy(&cso->sense_data, cto->rsp.m1.ct_resp, m); 776 cto->rsp.m1.ct_senselen = m; 777 cto->rsp.m1.ct_scsi_status |= CT2_SNSLEN_VALID; 778 } 779 } else { 780 cto->ct_flags |= CT2_FLAG_MODE0; 781 if ((cso->ccb_h.flags & CAM_DIR_MASK) == CAM_DIR_IN) { 782 cto->ct_flags |= CT2_DATA_IN; 783 } else { 784 cto->ct_flags |= CT2_DATA_OUT; 785 } 786 if ((ccb->ccb_h.flags & CAM_SEND_STATUS) != 0) { 787 cto->ct_flags |= CT2_SENDSTATUS; 788 cto->rsp.m0.ct_scsi_status = cso->scsi_status; 789 } 790 /* 791 * If we're sending data and status back together, 792 * we can't also send back sense data as well. 793 */ 794 ccb->ccb_h.flags &= ~CAM_SEND_SENSE; 795 } 796 if (cto->ct_flags & CAM_SEND_STATUS) { 797 isp_prt(isp, ISP_LOGTDEBUG2, 798 "CTIO2 RX_ID 0x%x SCSI STATUS 0x%x datalength %u", 799 cto->ct_rxid, cso->scsi_status, cto->ct_resid); 800 } 801 hp = &cto->ct_reserved; 802 } else { 803 ct_entry_t *cto = qe; 804 805 cto->ct_header.rqs_entry_type = RQSTYPE_CTIO; 806 cto->ct_header.rqs_entry_count = 1; 807 cto->ct_iid = cso->init_id; 808 cto->ct_tgt = ccb->ccb_h.target_id; 809 cto->ct_lun = ccb->ccb_h.target_lun; 810 if (cso->tag_id && cso->tag_action) { 811 /* 812 * We don't specify a tag type for regular SCSI. 813 * Just the tag value and set the flag. 814 */ 815 cto->ct_tag_val = cso->tag_id; 816 cto->ct_flags |= CT_TQAE; 817 } 818 if (ccb->ccb_h.flags & CAM_DIS_DISCONNECT) { 819 cto->ct_flags |= CT_NODISC; 820 } 821 if (cso->dxfer_len == 0) { 822 cto->ct_flags |= CT_NO_DATA; 823 } else if ((cso->ccb_h.flags & CAM_DIR_MASK) == CAM_DIR_IN) { 824 cto->ct_flags |= CT_DATA_IN; 825 } else { 826 cto->ct_flags |= CT_DATA_OUT; 827 } 828 if (ccb->ccb_h.flags & CAM_SEND_STATUS) { 829 cto->ct_flags |= CT_SENDSTATUS; 830 cto->ct_scsi_status = cso->scsi_status; 831 cto->ct_resid = cso->resid; 832 } 833 if (cto->ct_flags & CAM_SEND_STATUS) { 834 isp_prt(isp, ISP_LOGTDEBUG2, 835 "CTIO SCSI STATUS 0x%x resid %d", 836 cso->scsi_status, cso->resid); 837 } 838 hp = &cto->ct_reserved; 839 ccb->ccb_h.flags &= ~CAM_SEND_SENSE; 840 } 841 842 if (isp_save_xs(isp, (XS_T *)ccb, hp)) { 843 xpt_print_path(ccb->ccb_h.path); 844 printf("No XFLIST pointers for isp_target_start_ctio\n"); 845 return (CAM_RESRC_UNAVAIL); 846 } 847 848 849 /* 850 * Call the dma setup routines for this entry (and any subsequent 851 * CTIOs) if there's data to move, and then tell the f/w it's got 852 * new things to play with. As with isp_start's usage of DMA setup, 853 * any swizzling is done in the machine dependent layer. Because 854 * of this, we put the request onto the queue area first in native 855 * format. 856 */ 857 858 save_handle = *hp; 859 switch (ISP_DMASETUP(isp, cso, qe, &iptr, optr)) { 860 case CMD_QUEUED: 861 ISP_ADD_REQUEST(isp, iptr); 862 return (CAM_REQ_INPROG); 863 864 case CMD_EAGAIN: 865 ccb->ccb_h.status = CAM_RESRC_UNAVAIL; 866 isp_destroy_handle(isp, save_handle); 867 return (CAM_RESRC_UNAVAIL); 868 869 default: 870 isp_destroy_handle(isp, save_handle); 871 return (XS_ERR(ccb)); 872 } 873} 874 875static cam_status 876isp_target_putback_atio(struct ispsoftc *isp, union ccb *ccb) 877{ 878 void *qe; 879 struct ccb_accept_tio *atiop; 880 u_int16_t iptr, optr; 881 882 if (isp_getrqentry(isp, &iptr, &optr, &qe)) { 883 xpt_print_path(ccb->ccb_h.path); 884 printf("Request Queue Overflow in isp_target_putback_atio\n"); 885 return (CAM_RESRC_UNAVAIL); 886 } 887 bzero(qe, QENTRY_LEN); 888 atiop = (struct ccb_accept_tio *) ccb; 889 if (IS_FC(isp)) { 890 at2_entry_t *at = qe; 891 at->at_header.rqs_entry_type = RQSTYPE_ATIO2; 892 at->at_header.rqs_entry_count = 1; 893 if (isp->isp_maxluns > 16) { 894 at->at_scclun = (uint16_t) atiop->ccb_h.target_lun; 895 } else { 896 at->at_lun = (uint8_t) atiop->ccb_h.target_lun; 897 } 898 at->at_status = CT_OK; 899 at->at_rxid = atiop->tag_id; 900 ISP_SWIZ_ATIO2(isp, qe, qe); 901 } else { 902 at_entry_t *at = qe; 903 at->at_header.rqs_entry_type = RQSTYPE_ATIO; 904 at->at_header.rqs_entry_count = 1; 905 at->at_iid = atiop->init_id; 906 at->at_tgt = atiop->ccb_h.target_id; 907 at->at_lun = atiop->ccb_h.target_lun; 908 at->at_status = CT_OK; 909 if (atiop->ccb_h.status & CAM_TAG_ACTION_VALID) { 910 at->at_tag_type = atiop->tag_action; 911 } 912 at->at_tag_val = atiop->tag_id; 913 ISP_SWIZ_ATIO(isp, qe, qe); 914 } 915 ISP_TDQE(isp, "isp_target_putback_atio", (int) optr, qe); 916 ISP_ADD_REQUEST(isp, iptr); 917 return (CAM_REQ_CMP); 918} 919 920static void 921isp_refire_putback_atio(void *arg) 922{ 923 union ccb *ccb = arg; 924 int s = splcam(); 925 if (isp_target_putback_atio(XS_ISP(ccb), ccb) != CAM_REQ_CMP) { 926 (void) timeout(isp_refire_putback_atio, ccb, 10); 927 } else { 928 isp_handle_platform_ctio_part2(XS_ISP(ccb), ccb); 929 } 930 splx(s); 931} 932 933/* 934 * Handle ATIO stuff that the generic code can't. 935 * This means handling CDBs. 936 */ 937 938static int 939isp_handle_platform_atio(struct ispsoftc *isp, at_entry_t *aep) 940{ 941 tstate_t *tptr; 942 int status; 943 struct ccb_accept_tio *atiop; 944 945 /* 946 * The firmware status (except for the QLTM_SVALID bit) 947 * indicates why this ATIO was sent to us. 948 * 949 * If QLTM_SVALID is set, the firware has recommended Sense Data. 950 * 951 * If the DISCONNECTS DISABLED bit is set in the flags field, 952 * we're still connected on the SCSI bus - i.e. the initiator 953 * did not set DiscPriv in the identify message. We don't care 954 * about this so it's ignored. 955 */ 956 status = aep->at_status; 957 if ((status & ~QLTM_SVALID) == AT_PHASE_ERROR) { 958 /* 959 * Bus Phase Sequence error. We should have sense data 960 * suggested by the f/w. I'm not sure quite yet what 961 * to do about this for CAM. 962 */ 963 printf("%s: PHASE ERROR\n", isp->isp_name); 964 isp_endcmd(isp, aep, SCSI_STATUS_BUSY, 0); 965 return (0); 966 } 967 if ((status & ~QLTM_SVALID) != AT_CDB) { 968 printf("%s: bogus atio (0x%x) leaked to platform\n", 969 isp->isp_name, status); 970 isp_endcmd(isp, aep, SCSI_STATUS_BUSY, 0); 971 return (0); 972 } 973 974 tptr = get_lun_statep(isp, aep->at_lun); 975 if (tptr == NULL) { 976 tptr = get_lun_statep(isp, CAM_LUN_WILDCARD); 977 } 978 979 if (tptr == NULL) { 980 /* 981 * Because we can't autofeed sense data back with 982 * a command for parallel SCSI, we can't give back 983 * a CHECK CONDITION. We'll give back a BUSY status 984 * instead. This works out okay because the only 985 * time we should, in fact, get this, is in the 986 * case that somebody configured us without the 987 * blackhole driver, so they get what they deserve. 988 */ 989 isp_endcmd(isp, aep, SCSI_STATUS_BUSY, 0); 990 return (0); 991 } 992 993 atiop = (struct ccb_accept_tio *) SLIST_FIRST(&tptr->atios); 994 if (atiop == NULL) { 995 /* 996 * Because we can't autofeed sense data back with 997 * a command for parallel SCSI, we can't give back 998 * a CHECK CONDITION. We'll give back a QUEUE FULL status 999 * instead. This works out okay because the only time we 1000 * should, in fact, get this, is in the case that we've 1001 * run out of ATIOS. 1002 */ 1003 xpt_print_path(tptr->owner); 1004 printf("no ATIOS for lun %d from initiator %d\n", 1005 aep->at_lun, aep->at_iid); 1006 rls_lun_statep(isp, tptr); 1007 if (aep->at_flags & AT_TQAE) 1008 isp_endcmd(isp, aep, SCSI_STATUS_QUEUE_FULL, 0); 1009 else 1010 isp_endcmd(isp, aep, SCSI_STATUS_BUSY, 0); 1011 return (0); 1012 } 1013 SLIST_REMOVE_HEAD(&tptr->atios, sim_links.sle); 1014 if (tptr == &isp->isp_osinfo.tsdflt) { 1015 atiop->ccb_h.target_id = aep->at_tgt; 1016 atiop->ccb_h.target_lun = aep->at_lun; 1017 } 1018 if (aep->at_flags & AT_NODISC) { 1019 atiop->ccb_h.flags = CAM_DIS_DISCONNECT; 1020 } else { 1021 atiop->ccb_h.flags = 0; 1022 } 1023 1024 if (status & QLTM_SVALID) { 1025 size_t amt = imin(QLTM_SENSELEN, sizeof (atiop->sense_data)); 1026 atiop->sense_len = amt; 1027 MEMCPY(&atiop->sense_data, aep->at_sense, amt); 1028 } else { 1029 atiop->sense_len = 0; 1030 } 1031 1032 atiop->init_id = aep->at_iid; 1033 atiop->cdb_len = aep->at_cdblen; 1034 MEMCPY(atiop->cdb_io.cdb_bytes, aep->at_cdb, aep->at_cdblen); 1035 atiop->ccb_h.status = CAM_CDB_RECVD; 1036 atiop->tag_id = aep->at_tag_val; 1037 if ((atiop->tag_action = aep->at_tag_type) != 0) { 1038 atiop->ccb_h.status |= CAM_TAG_ACTION_VALID; 1039 } 1040 xpt_done((union ccb*)atiop); 1041 isp_prt(isp, ISP_LOGTDEBUG2, 1042 "ATIO CDB=0x%x iid%d->lun%d tag 0x%x ttype 0x%x %s", 1043 aep->at_cdb[0] & 0xff, aep->at_iid, aep->at_lun, 1044 aep->at_tag_val & 0xff, aep->at_tag_type, 1045 (aep->at_flags & AT_NODISC)? "nondisc" : "disconnecting"); 1046 rls_lun_statep(isp, tptr); 1047 return (0); 1048} 1049 1050static int 1051isp_handle_platform_atio2(struct ispsoftc *isp, at2_entry_t *aep) 1052{ 1053 lun_id_t lun; 1054 tstate_t *tptr; 1055 struct ccb_accept_tio *atiop; 1056 1057 /* 1058 * The firmware status (except for the QLTM_SVALID bit) 1059 * indicates why this ATIO was sent to us. 1060 * 1061 * If QLTM_SVALID is set, the firware has recommended Sense Data. 1062 */ 1063 if ((aep->at_status & ~QLTM_SVALID) != AT_CDB) { 1064 printf("%s: bogus atio (0x%x) leaked to platform\n", 1065 isp->isp_name, aep->at_status); 1066 isp_endcmd(isp, aep, SCSI_STATUS_BUSY, 0); 1067 return (0); 1068 } 1069 1070 if (isp->isp_maxluns > 16) { 1071 lun = aep->at_scclun; 1072 } else { 1073 lun = aep->at_lun; 1074 } 1075 tptr = get_lun_statep(isp, lun); 1076 if (tptr == NULL) { 1077 tptr = get_lun_statep(isp, CAM_LUN_WILDCARD); 1078 } 1079 1080 if (tptr == NULL) { 1081 /* 1082 * What we'd like to know is whether or not we have a listener 1083 * upstream that really hasn't configured yet. If we do, then 1084 * we can give a more sensible reply here. If not, then we can 1085 * reject this out of hand. 1086 * 1087 * Choices for what to send were 1088 * 1089 * Not Ready, Unit Not Self-Configured Yet 1090 * (0x2,0x3e,0x00) 1091 * 1092 * for the former and 1093 * 1094 * Illegal Request, Logical Unit Not Supported 1095 * (0x5,0x25,0x00) 1096 * 1097 * for the latter. 1098 * 1099 * We used to decide whether there was at least one listener 1100 * based upon whether the black hole driver was configured. 1101 * However, recent config(8) changes have made this hard to do 1102 * at this time. 1103 * 1104 */ 1105 u_int32_t ccode = SCSI_STATUS_BUSY; 1106 1107 /* 1108 * Because we can't autofeed sense data back with 1109 * a command for parallel SCSI, we can't give back 1110 * a CHECK CONDITION. We'll give back a BUSY status 1111 * instead. This works out okay because the only 1112 * time we should, in fact, get this, is in the 1113 * case that somebody configured us without the 1114 * blackhole driver, so they get what they deserve. 1115 */ 1116 isp_endcmd(isp, aep, ccode, 0); 1117 return (0); 1118 } 1119 1120 atiop = (struct ccb_accept_tio *) SLIST_FIRST(&tptr->atios); 1121 if (atiop == NULL) { 1122 /* 1123 * Because we can't autofeed sense data back with 1124 * a command for parallel SCSI, we can't give back 1125 * a CHECK CONDITION. We'll give back a QUEUE FULL status 1126 * instead. This works out okay because the only time we 1127 * should, in fact, get this, is in the case that we've 1128 * run out of ATIOS. 1129 */ 1130 xpt_print_path(tptr->owner); 1131 printf("no ATIOS for lun %d from initiator %d\n", 1132 lun, aep->at_iid); 1133 rls_lun_statep(isp, tptr); 1134 if (aep->at_flags & AT_TQAE) 1135 isp_endcmd(isp, aep, SCSI_STATUS_QUEUE_FULL, 0); 1136 else 1137 isp_endcmd(isp, aep, SCSI_STATUS_BUSY, 0); 1138 return (0); 1139 } 1140 SLIST_REMOVE_HEAD(&tptr->atios, sim_links.sle); 1141 1142 if (tptr == &isp->isp_osinfo.tsdflt) { 1143 atiop->ccb_h.target_id = 1144 ((fcparam *)isp->isp_param)->isp_loopid; 1145 atiop->ccb_h.target_lun = lun; 1146 } 1147 if (aep->at_status & QLTM_SVALID) { 1148 size_t amt = imin(QLTM_SENSELEN, sizeof (atiop->sense_data)); 1149 atiop->sense_len = amt; 1150 MEMCPY(&atiop->sense_data, aep->at_sense, amt); 1151 } else { 1152 atiop->sense_len = 0; 1153 } 1154 1155 atiop->init_id = aep->at_iid; 1156 atiop->cdb_len = ATIO2_CDBLEN; 1157 MEMCPY(atiop->cdb_io.cdb_bytes, aep->at_cdb, ATIO2_CDBLEN); 1158 atiop->ccb_h.status = CAM_CDB_RECVD; 1159 atiop->tag_id = aep->at_rxid; 1160 switch (aep->at_taskflags & ATIO2_TC_ATTR_MASK) { 1161 case ATIO2_TC_ATTR_SIMPLEQ: 1162 atiop->tag_action = MSG_SIMPLE_Q_TAG; 1163 break; 1164 case ATIO2_TC_ATTR_HEADOFQ: 1165 atiop->tag_action = MSG_HEAD_OF_Q_TAG; 1166 break; 1167 case ATIO2_TC_ATTR_ORDERED: 1168 atiop->tag_action = MSG_ORDERED_Q_TAG; 1169 break; 1170 case ATIO2_TC_ATTR_ACAQ: /* ?? */ 1171 case ATIO2_TC_ATTR_UNTAGGED: 1172 default: 1173 atiop->tag_action = 0; 1174 break; 1175 } 1176 if (atiop->tag_action != 0) { 1177 atiop->ccb_h.status |= CAM_TAG_ACTION_VALID; 1178 } 1179 1180 /* 1181 * Preserve overall command datalength in private field. 1182 */ 1183 atiop->ccb_h.spriv_field0 = aep->at_datalen; 1184 1185 xpt_done((union ccb*)atiop); 1186 isp_prt(isp, ISP_LOGTDEBUG2, 1187 "ATIO2 RX_ID 0x%x CDB=0x%x iid%d->lun%d tattr 0x%x datalen %u", 1188 aep->at_rxid & 0xffff, aep->at_cdb[0] & 0xff, aep->at_iid, 1189 lun, aep->at_taskflags, aep->at_datalen); 1190 rls_lun_statep(isp, tptr); 1191 return (0); 1192} 1193 1194static int 1195isp_handle_platform_ctio(struct ispsoftc *isp, void *arg) 1196{ 1197 union ccb *ccb; 1198 int sentstatus, ok, notify_cam; 1199 1200 /* 1201 * CTIO and CTIO2 are close enough.... 1202 */ 1203 1204 ccb = (union ccb *) isp_find_xs(isp, ((ct_entry_t *)arg)->ct_reserved); 1205 KASSERT((ccb != NULL), ("null ccb in isp_handle_platform_ctio")); 1206 isp_destroy_handle(isp, ((ct_entry_t *)arg)->ct_reserved); 1207 1208 if (IS_FC(isp)) { 1209 ct2_entry_t *ct = arg; 1210 sentstatus = ct->ct_flags & CT2_SENDSTATUS; 1211 ok = (ct->ct_status & ~QLTM_SVALID) == CT_OK; 1212 if (ok && (ccb->ccb_h.flags & CAM_SEND_SENSE)) { 1213 ccb->ccb_h.status |= CAM_SENT_SENSE; 1214 } 1215 isp_prt(isp, ISP_LOGTDEBUG2, 1216 "CTIO2 RX_ID 0x%x sts 0x%x flg 0x%x sns %d FIN", 1217 ct->ct_rxid, ct->ct_status, ct->ct_flags, 1218 (ccb->ccb_h.status & CAM_SENT_SENSE) != 0); 1219 notify_cam = ct->ct_header.rqs_seqno; 1220 } else { 1221 ct_entry_t *ct = arg; 1222 sentstatus = ct->ct_flags & CT_SENDSTATUS; 1223 ok = (ct->ct_status & ~QLTM_SVALID) == CT_OK; 1224 isp_prt(isp, ISP_LOGTDEBUG2, 1225 "CTIO tag 0x%x sts 0x%x flg 0x%x FIN", 1226 ct->ct_tag_val, ct->ct_status, ct->ct_flags); 1227 notify_cam = ct->ct_header.rqs_seqno; 1228 } 1229 1230 /* 1231 * We're here either because data transfers are done (and 1232 * it's time to send a final status CTIO) or because the final 1233 * status CTIO is done. We don't get called for all intermediate 1234 * CTIOs that happen for a large data transfer. 1235 * 1236 * In any case, for this platform, the upper layers figure out 1237 * what to do next, so all we do here is collect status and 1238 * pass information along. The exception is that we clear 1239 * the notion of handling a non-disconnecting command here. 1240 */ 1241 1242 if (sentstatus) { 1243 /* 1244 * Data transfer done. See if all went okay. 1245 */ 1246 if (ok) { 1247 ccb->csio.resid = 0; 1248 } else { 1249 ccb->csio.resid = ccb->csio.dxfer_len; 1250 } 1251 } 1252 1253 if (notify_cam == 0) { 1254 isp_prt(isp, ISP_LOGTDEBUG1, "Intermediate CTIO done"); 1255 return (0); 1256 } 1257 isp_prt(isp, ISP_LOGTDEBUG1, "Final CTIO done"); 1258 if (isp_target_putback_atio(isp, ccb) != CAM_REQ_CMP) { 1259 (void) timeout(isp_refire_putback_atio, ccb, 10); 1260 } else { 1261 isp_handle_platform_ctio_part2(isp, ccb); 1262 } 1263 return (0); 1264} 1265 1266static void 1267isp_handle_platform_ctio_part2(struct ispsoftc *isp, union ccb *ccb) 1268{ 1269 if ((ccb->ccb_h.status & CAM_STATUS_MASK) == CAM_REQ_INPROG) { 1270 ccb->ccb_h.status |= CAM_REQ_CMP; 1271 } 1272 ccb->ccb_h.status &= ~CAM_SIM_QUEUED; 1273 if (isp->isp_osinfo.simqfrozen & SIMQFRZ_RESOURCE) { 1274 isp->isp_osinfo.simqfrozen &= ~SIMQFRZ_RESOURCE; 1275 if (isp->isp_osinfo.simqfrozen == 0) { 1276 if ((ccb->ccb_h.status & CAM_DEV_QFRZN) == 0) { 1277 isp_prt(isp, ISP_LOGDEBUG2, "ctio->relsimq"); 1278 ccb->ccb_h.status |= CAM_RELEASE_SIMQ; 1279 } else { 1280 isp_prt(isp, ISP_LOGDEBUG2, "ctio->devqfrozen"); 1281 } 1282 } else { 1283 isp_prt(isp, ISP_LOGDEBUG2, 1284 "ctio->simqfrozen(%x)", isp->isp_osinfo.simqfrozen); 1285 } 1286 } 1287 xpt_done(ccb); 1288} 1289#endif 1290 1291static void 1292isp_cam_async(void *cbarg, u_int32_t code, struct cam_path *path, void *arg) 1293{ 1294 struct cam_sim *sim; 1295 struct ispsoftc *isp; 1296 1297 sim = (struct cam_sim *)cbarg; 1298 isp = (struct ispsoftc *) cam_sim_softc(sim); 1299 switch (code) { 1300 case AC_LOST_DEVICE: 1301 if (IS_SCSI(isp)) { 1302 u_int16_t oflags, nflags; 1303 sdparam *sdp = isp->isp_param; 1304 int rvf, tgt; 1305 1306 tgt = xpt_path_target_id(path); 1307 rvf = ISP_FW_REVX(isp->isp_fwrev); 1308 ISP_LOCK(isp); 1309 sdp += cam_sim_bus(sim); 1310 isp->isp_update |= (1 << cam_sim_bus(sim)); 1311 nflags = DPARM_SAFE_DFLT; 1312 if (rvf >= ISP_FW_REV(7, 55, 0) || 1313 (ISP_FW_REV(4, 55, 0) <= rvf && 1314 (rvf < ISP_FW_REV(5, 0, 0)))) { 1315 nflags |= DPARM_NARROW | DPARM_ASYNC; 1316 } 1317 oflags = sdp->isp_devparam[tgt].dev_flags; 1318 sdp->isp_devparam[tgt].dev_flags = nflags; 1319 sdp->isp_devparam[tgt].dev_update = 1; 1320 (void) isp_control(isp, ISPCTL_UPDATE_PARAMS, NULL); 1321 sdp->isp_devparam[tgt].dev_flags = oflags; 1322 ISP_UNLOCK(isp); 1323 } 1324 break; 1325 default: 1326 printf("%s: isp_attach Async Code 0x%x\n", isp->isp_name, code); 1327 break; 1328 } 1329} 1330 1331static void 1332isp_poll(struct cam_sim *sim) 1333{ 1334 struct ispsoftc *isp = cam_sim_softc(sim); 1335 ISP_LOCK(isp); 1336 (void) isp_intr(isp); 1337 ISP_UNLOCK(isp); 1338} 1339 1340static void 1341isp_relsim(void *arg) 1342{ 1343 struct ispsoftc *isp = arg; 1344 ISP_LOCK(isp); 1345 if (isp->isp_osinfo.simqfrozen & SIMQFRZ_TIMED) { 1346 int wasfrozen = isp->isp_osinfo.simqfrozen & SIMQFRZ_TIMED; 1347 isp->isp_osinfo.simqfrozen &= ~SIMQFRZ_TIMED; 1348 if (wasfrozen && isp->isp_osinfo.simqfrozen == 0) { 1349 xpt_release_simq(isp->isp_sim, 1); 1350 isp_prt(isp, ISP_LOGDEBUG2, "timed relsimq"); 1351 } 1352 } 1353 ISP_UNLOCK(isp); 1354} 1355 1356static void 1357isp_watchdog(void *arg) 1358{ 1359 XS_T *xs = arg; 1360 struct ispsoftc *isp = XS_ISP(xs); 1361 u_int32_t handle; 1362 1363 /* 1364 * We've decided this command is dead. Make sure we're not trying 1365 * to kill a command that's already dead by getting it's handle and 1366 * and seeing whether it's still alive. 1367 */ 1368 ISP_LOCK(isp); 1369 handle = isp_find_handle(isp, xs); 1370 if (handle) { 1371 u_int16_t r; 1372 1373 if (XS_CMD_DONE_P(xs)) { 1374 isp_prt(isp, ISP_LOGDEBUG1, 1375 "watchdog found done cmd (handle 0x%x)", handle); 1376 ISP_UNLOCK(isp); 1377 return; 1378 } 1379 1380 if (XS_CMD_WDOG_P(xs)) { 1381 isp_prt(isp, ISP_LOGDEBUG2, 1382 "recursive watchdog (handle 0x%x)", handle); 1383 ISP_UNLOCK(isp); 1384 return; 1385 } 1386 1387 XS_CMD_S_WDOG(xs); 1388 1389 r = ISP_READ(isp, BIU_ISR); 1390 1391 if (INT_PENDING(isp, r) && isp_intr(isp) && XS_CMD_DONE_P(xs)) { 1392 isp_prt(isp, ISP_LOGDEBUG2, 1393 "watchdog cleanup (%x, %x)", handle, r); 1394 xpt_done((union ccb *) xs); 1395 } else if (XS_CMD_GRACE_P(xs)) { 1396 /* 1397 * Make sure the command is *really* dead before we 1398 * release the handle (and DMA resources) for reuse. 1399 */ 1400 (void) isp_control(isp, ISPCTL_ABORT_CMD, arg); 1401 1402 /* 1403 * After this point, the comamnd is really dead. 1404 */ 1405 if (XS_XFRLEN(xs)) { 1406 ISP_DMAFREE(isp, xs, handle); 1407 } 1408 isp_destroy_handle(isp, handle); 1409 xpt_print_path(xs->ccb_h.path); 1410 printf("%s: watchdog timeout (%x, %x)\n", 1411 isp->isp_name, handle, r); 1412 XS_SETERR(xs, CAM_CMD_TIMEOUT); 1413 XS_CMD_C_WDOG(xs); 1414 isp_done(xs); 1415 } else { 1416 u_int16_t iptr, optr; 1417 ispreq_t *mp; 1418 1419 XS_CMD_C_WDOG(xs); 1420 xs->ccb_h.timeout_ch = timeout(isp_watchdog, xs, hz); 1421 if (isp_getrqentry(isp, &iptr, &optr, (void **) &mp)) { 1422 ISP_UNLOCK(isp); 1423 return; 1424 } 1425 XS_CMD_S_GRACE(xs); 1426 MEMZERO((void *) mp, sizeof (*mp)); 1427 mp->req_header.rqs_entry_count = 1; 1428 mp->req_header.rqs_entry_type = RQSTYPE_MARKER; 1429 mp->req_modifier = SYNC_ALL; 1430 mp->req_target = XS_CHANNEL(xs) << 7; 1431 ISP_SWIZZLE_REQUEST(isp, mp); 1432 ISP_ADD_REQUEST(isp, iptr); 1433 } 1434 } else { 1435 isp_prt(isp, ISP_LOGDEBUG2, "watchdog with no command"); 1436 } 1437 ISP_UNLOCK(isp); 1438} 1439 1440static void 1441isp_action(struct cam_sim *sim, union ccb *ccb) 1442{ 1443 int bus, tgt, error; 1444 struct ispsoftc *isp; 1445 struct ccb_trans_settings *cts; 1446 1447 CAM_DEBUG(ccb->ccb_h.path, CAM_DEBUG_TRACE, ("isp_action\n")); 1448 1449 isp = (struct ispsoftc *)cam_sim_softc(sim); 1450 ccb->ccb_h.sim_priv.entries[0].field = 0; 1451 ccb->ccb_h.sim_priv.entries[1].ptr = isp; 1452 if (isp->isp_state != ISP_RUNSTATE && 1453 ccb->ccb_h.func_code == XPT_SCSI_IO) { 1454 ISP_LOCK(isp); 1455 isp_init(isp); 1456 if (isp->isp_state != ISP_INITSTATE) { 1457 ISP_UNLOCK(isp); 1458 /* 1459 * Lie. Say it was a selection timeout. 1460 */ 1461 ccb->ccb_h.status = CAM_SEL_TIMEOUT | CAM_DEV_QFRZN; 1462 xpt_freeze_devq(ccb->ccb_h.path, 1); 1463 xpt_done(ccb); 1464 return; 1465 } 1466 isp->isp_state = ISP_RUNSTATE; 1467 ISP_UNLOCK(isp); 1468 } 1469 isp_prt(isp, ISP_LOGDEBUG2, "isp_action code %x", ccb->ccb_h.func_code); 1470 1471 switch (ccb->ccb_h.func_code) { 1472 case XPT_SCSI_IO: /* Execute the requested I/O operation */ 1473 /* 1474 * Do a couple of preliminary checks... 1475 */ 1476 if ((ccb->ccb_h.flags & CAM_CDB_POINTER) != 0) { 1477 if ((ccb->ccb_h.flags & CAM_CDB_PHYS) != 0) { 1478 ccb->ccb_h.status = CAM_REQ_INVALID; 1479 xpt_done(ccb); 1480 break; 1481 } 1482 } 1483#ifdef DIAGNOSTIC 1484 if (ccb->ccb_h.target_id > (ISP_MAX_TARGETS(isp) - 1)) { 1485 ccb->ccb_h.status = CAM_PATH_INVALID; 1486 } else if (ccb->ccb_h.target_lun > (ISP_MAX_LUNS(isp) - 1)) { 1487 ccb->ccb_h.status = CAM_PATH_INVALID; 1488 } 1489 if (ccb->ccb_h.status == CAM_PATH_INVALID) { 1490 isp_prt(isp, ISP_LOGERR, 1491 "invalid tgt/lun (%d.%d) in XPT_SCSI_IO", 1492 ccb->ccb_h.target_id, ccb->ccb_h.target_lun); 1493 xpt_done(ccb); 1494 break; 1495 } 1496#endif 1497 ((struct ccb_scsiio *) ccb)->scsi_status = SCSI_STATUS_OK; 1498 ISP_LOCK(isp); 1499 error = isp_start((XS_T *) ccb); 1500 ISP_UNLOCK(isp); 1501 switch (error) { 1502 case CMD_QUEUED: 1503 ccb->ccb_h.status |= CAM_SIM_QUEUED; 1504 if (ccb->ccb_h.timeout != CAM_TIME_INFINITY) { 1505 u_int64_t ticks = (u_int64_t) hz; 1506 if (ccb->ccb_h.timeout == CAM_TIME_DEFAULT) 1507 ticks = 60 * 1000 * ticks; 1508 else 1509 ticks = ccb->ccb_h.timeout * hz; 1510 ticks = ((ticks + 999) / 1000) + hz + hz; 1511 if (ticks >= 0x80000000) { 1512 isp_prt(isp, ISP_LOGERR, 1513 "timeout overflow"); 1514 ticks = 0x80000000; 1515 } 1516 ccb->ccb_h.timeout_ch = timeout(isp_watchdog, 1517 (caddr_t)ccb, (int)ticks); 1518 } else { 1519 callout_handle_init(&ccb->ccb_h.timeout_ch); 1520 } 1521 break; 1522 case CMD_RQLATER: 1523 if (isp->isp_osinfo.simqfrozen == 0) { 1524 isp_prt(isp, ISP_LOGDEBUG2, 1525 "RQLATER freeze simq"); 1526 isp->isp_osinfo.simqfrozen |= SIMQFRZ_TIMED; 1527 timeout(isp_relsim, isp, 500); 1528 xpt_freeze_simq(sim, 1); 1529 } 1530 XS_SETERR(ccb, CAM_REQUEUE_REQ); 1531 xpt_done(ccb); 1532 break; 1533 case CMD_EAGAIN: 1534 if (isp->isp_osinfo.simqfrozen == 0) { 1535 xpt_freeze_simq(sim, 1); 1536 isp_prt(isp, ISP_LOGDEBUG2, 1537 "EAGAIN freeze simq"); 1538 } 1539 isp->isp_osinfo.simqfrozen |= SIMQFRZ_RESOURCE; 1540 XS_SETERR(ccb, CAM_REQUEUE_REQ); 1541 xpt_done(ccb); 1542 break; 1543 case CMD_COMPLETE: 1544 ISP_LOCK(isp); 1545 isp_done((struct ccb_scsiio *) ccb); 1546 ISP_UNLOCK(isp); 1547 break; 1548 default: 1549 isp_prt(isp, ISP_LOGERR, 1550 "What's this? 0x%x at %d in file %s", 1551 error, __LINE__, __FILE__); 1552 XS_SETERR(ccb, CAM_REQ_CMP_ERR); 1553 xpt_done(ccb); 1554 } 1555 break; 1556 1557#ifdef ISP_TARGET_MODE 1558 case XPT_EN_LUN: /* Enable LUN as a target */ 1559 isp_en_lun(isp, ccb); 1560 xpt_done(ccb); 1561 break; 1562 1563 case XPT_NOTIFY_ACK: /* recycle notify ack */ 1564 case XPT_IMMED_NOTIFY: /* Add Immediate Notify Resource */ 1565 case XPT_ACCEPT_TARGET_IO: /* Add Accept Target IO Resource */ 1566 { 1567 tstate_t *tptr = get_lun_statep(isp, ccb->ccb_h.target_lun); 1568 if (tptr == NULL) { 1569 ccb->ccb_h.status = CAM_LUN_INVALID; 1570 xpt_done(ccb); 1571 break; 1572 } 1573 ccb->ccb_h.sim_priv.entries[0].field = 0; 1574 ccb->ccb_h.sim_priv.entries[1].ptr = isp; 1575 ISP_LOCK(isp); 1576 if (ccb->ccb_h.func_code == XPT_ACCEPT_TARGET_IO) { 1577#if 0 1578 (void) isp_target_putback_atio(isp, ccb); 1579#endif 1580 SLIST_INSERT_HEAD(&tptr->atios, 1581 &ccb->ccb_h, sim_links.sle); 1582 } else { 1583 SLIST_INSERT_HEAD(&tptr->inots, &ccb->ccb_h, 1584 sim_links.sle); 1585 } 1586 ISP_UNLOCK(isp); 1587 rls_lun_statep(isp, tptr); 1588 ccb->ccb_h.status = CAM_REQ_INPROG; 1589 break; 1590 } 1591 case XPT_CONT_TARGET_IO: 1592 { 1593 ISP_LOCK(isp); 1594 ccb->ccb_h.status = isp_target_start_ctio(isp, ccb); 1595 if (ccb->ccb_h.status != CAM_REQ_INPROG) { 1596 if (isp->isp_osinfo.simqfrozen == 0) { 1597 xpt_freeze_simq(sim, 1); 1598 xpt_print_path(ccb->ccb_h.path); 1599 printf("XPT_CONT_TARGET_IO freeze simq\n"); 1600 } 1601 isp->isp_osinfo.simqfrozen |= SIMQFRZ_RESOURCE; 1602 XS_SETERR(ccb, CAM_REQUEUE_REQ); 1603 xpt_done(ccb); 1604 } else { 1605 ccb->ccb_h.status |= CAM_SIM_QUEUED; 1606 } 1607 ISP_UNLOCK(isp); 1608 break; 1609 } 1610#endif 1611 case XPT_RESET_DEV: /* BDR the specified SCSI device */ 1612 1613 bus = cam_sim_bus(xpt_path_sim(ccb->ccb_h.path)); 1614 tgt = ccb->ccb_h.target_id; 1615 tgt |= (bus << 16); 1616 1617 ISP_LOCK(isp); 1618 error = isp_control(isp, ISPCTL_RESET_DEV, &tgt); 1619 ISP_UNLOCK(isp); 1620 if (error) { 1621 ccb->ccb_h.status = CAM_REQ_CMP_ERR; 1622 } else { 1623 ccb->ccb_h.status = CAM_REQ_CMP; 1624 } 1625 xpt_done(ccb); 1626 break; 1627 case XPT_ABORT: /* Abort the specified CCB */ 1628 { 1629 union ccb *accb = ccb->cab.abort_ccb; 1630 switch (accb->ccb_h.func_code) { 1631#ifdef ISP_TARGET_MODE 1632 case XPT_ACCEPT_TARGET_IO: 1633 case XPT_IMMED_NOTIFY: 1634 ccb->ccb_h.status = isp_abort_tgt_ccb(isp, ccb); 1635 break; 1636 case XPT_CONT_TARGET_IO: 1637 isp_prt(isp, ISP_LOGERR, "cannot abort CTIOs yet"); 1638 ccb->ccb_h.status = CAM_UA_ABORT; 1639 break; 1640#endif 1641 case XPT_SCSI_IO: 1642 ISP_LOCK(isp); 1643 error = isp_control(isp, ISPCTL_ABORT_CMD, ccb); 1644 ISP_UNLOCK(isp); 1645 if (error) { 1646 ccb->ccb_h.status = CAM_UA_ABORT; 1647 } else { 1648 ccb->ccb_h.status = CAM_REQ_CMP; 1649 } 1650 break; 1651 default: 1652 ccb->ccb_h.status = CAM_REQ_INVALID; 1653 break; 1654 } 1655 xpt_done(ccb); 1656 break; 1657 } 1658 case XPT_SET_TRAN_SETTINGS: /* Nexus Settings */ 1659 1660 cts = &ccb->cts; 1661 tgt = cts->ccb_h.target_id; 1662 ISP_LOCK(isp); 1663 if (IS_SCSI(isp)) { 1664 sdparam *sdp = isp->isp_param; 1665 u_int16_t *dptr; 1666 1667 bus = cam_sim_bus(xpt_path_sim(cts->ccb_h.path)); 1668 1669 sdp += bus; 1670#if 0 1671 if (cts->flags & CCB_TRANS_CURRENT_SETTINGS) 1672 dptr = &sdp->isp_devparam[tgt].cur_dflags; 1673 else 1674 dptr = &sdp->isp_devparam[tgt].dev_flags; 1675#else 1676 /* 1677 * We always update (internally) from dev_flags 1678 * so any request to change settings just gets 1679 * vectored to that location. 1680 */ 1681 dptr = &sdp->isp_devparam[tgt].dev_flags; 1682#endif 1683 1684 /* 1685 * Note that these operations affect the 1686 * the goal flags (dev_flags)- not 1687 * the current state flags. Then we mark 1688 * things so that the next operation to 1689 * this HBA will cause the update to occur. 1690 */ 1691 if (cts->valid & CCB_TRANS_DISC_VALID) { 1692 if ((cts->flags & CCB_TRANS_DISC_ENB) != 0) { 1693 *dptr |= DPARM_DISC; 1694 } else { 1695 *dptr &= ~DPARM_DISC; 1696 } 1697 } 1698 if (cts->valid & CCB_TRANS_TQ_VALID) { 1699 if ((cts->flags & CCB_TRANS_TAG_ENB) != 0) { 1700 *dptr |= DPARM_TQING; 1701 } else { 1702 *dptr &= ~DPARM_TQING; 1703 } 1704 } 1705 if (cts->valid & CCB_TRANS_BUS_WIDTH_VALID) { 1706 switch (cts->bus_width) { 1707 case MSG_EXT_WDTR_BUS_16_BIT: 1708 *dptr |= DPARM_WIDE; 1709 break; 1710 default: 1711 *dptr &= ~DPARM_WIDE; 1712 } 1713 } 1714 /* 1715 * Any SYNC RATE of nonzero and SYNC_OFFSET 1716 * of nonzero will cause us to go to the 1717 * selected (from NVRAM) maximum value for 1718 * this device. At a later point, we'll 1719 * allow finer control. 1720 */ 1721 if ((cts->valid & CCB_TRANS_SYNC_RATE_VALID) && 1722 (cts->valid & CCB_TRANS_SYNC_OFFSET_VALID) && 1723 (cts->sync_offset > 0)) { 1724 *dptr |= DPARM_SYNC; 1725 } else { 1726 *dptr &= ~DPARM_SYNC; 1727 } 1728 *dptr |= DPARM_SAFE_DFLT; 1729 isp_prt(isp, ISP_LOGDEBUG0, 1730 "%d.%d set %s period 0x%x offset 0x%x flags 0x%x", 1731 bus, tgt, (cts->flags & CCB_TRANS_CURRENT_SETTINGS)? 1732 "current" : "user", 1733 sdp->isp_devparam[tgt].sync_period, 1734 sdp->isp_devparam[tgt].sync_offset, 1735 sdp->isp_devparam[tgt].dev_flags); 1736 sdp->isp_devparam[tgt].dev_update = 1; 1737 isp->isp_update |= (1 << bus); 1738 } 1739 ISP_UNLOCK(isp); 1740 ccb->ccb_h.status = CAM_REQ_CMP; 1741 xpt_done(ccb); 1742 break; 1743 1744 case XPT_GET_TRAN_SETTINGS: 1745 1746 cts = &ccb->cts; 1747 tgt = cts->ccb_h.target_id; 1748 if (IS_FC(isp)) { 1749 /* 1750 * a lot of normal SCSI things don't make sense. 1751 */ 1752 cts->flags = CCB_TRANS_TAG_ENB | CCB_TRANS_DISC_ENB; 1753 cts->valid = CCB_TRANS_DISC_VALID | CCB_TRANS_TQ_VALID; 1754 /* 1755 * How do you measure the width of a high 1756 * speed serial bus? Well, in bytes. 1757 * 1758 * Offset and period make no sense, though, so we set 1759 * (above) a 'base' transfer speed to be gigabit. 1760 */ 1761 cts->bus_width = MSG_EXT_WDTR_BUS_8_BIT; 1762 } else { 1763 sdparam *sdp = isp->isp_param; 1764 u_int16_t dval, pval, oval; 1765 int bus = cam_sim_bus(xpt_path_sim(cts->ccb_h.path)); 1766 1767 sdp += bus; 1768 if (cts->flags & CCB_TRANS_CURRENT_SETTINGS) { 1769 ISP_LOCK(isp); 1770 sdp->isp_devparam[tgt].dev_refresh = 1; 1771 isp->isp_update |= (1 << bus); 1772 (void) isp_control(isp, ISPCTL_UPDATE_PARAMS, 1773 NULL); 1774 ISP_UNLOCK(isp); 1775 dval = sdp->isp_devparam[tgt].cur_dflags; 1776 oval = sdp->isp_devparam[tgt].cur_offset; 1777 pval = sdp->isp_devparam[tgt].cur_period; 1778 } else { 1779 dval = sdp->isp_devparam[tgt].dev_flags; 1780 oval = sdp->isp_devparam[tgt].sync_offset; 1781 pval = sdp->isp_devparam[tgt].sync_period; 1782 } 1783 1784 ISP_LOCK(isp); 1785 cts->flags &= ~(CCB_TRANS_DISC_ENB|CCB_TRANS_TAG_ENB); 1786 1787 if (dval & DPARM_DISC) { 1788 cts->flags |= CCB_TRANS_DISC_ENB; 1789 } 1790 if (dval & DPARM_TQING) { 1791 cts->flags |= CCB_TRANS_TAG_ENB; 1792 } 1793 if (dval & DPARM_WIDE) { 1794 cts->bus_width = MSG_EXT_WDTR_BUS_16_BIT; 1795 } else { 1796 cts->bus_width = MSG_EXT_WDTR_BUS_8_BIT; 1797 } 1798 cts->valid = CCB_TRANS_BUS_WIDTH_VALID | 1799 CCB_TRANS_DISC_VALID | CCB_TRANS_TQ_VALID; 1800 1801 if ((dval & DPARM_SYNC) && oval != 0) { 1802 cts->sync_period = pval; 1803 cts->sync_offset = oval; 1804 cts->valid |= 1805 CCB_TRANS_SYNC_RATE_VALID | 1806 CCB_TRANS_SYNC_OFFSET_VALID; 1807 } 1808 ISP_UNLOCK(isp); 1809 isp_prt(isp, ISP_LOGDEBUG0, 1810 "%d.%d get %s period 0x%x offset 0x%x flags 0x%x", 1811 bus, tgt, (cts->flags & CCB_TRANS_CURRENT_SETTINGS)? 1812 "current" : "user", pval, oval, dval); 1813 } 1814 ccb->ccb_h.status = CAM_REQ_CMP; 1815 xpt_done(ccb); 1816 break; 1817 1818 case XPT_CALC_GEOMETRY: 1819 { 1820 struct ccb_calc_geometry *ccg; 1821 u_int32_t secs_per_cylinder; 1822 u_int32_t size_mb; 1823 1824 ccg = &ccb->ccg; 1825 if (ccg->block_size == 0) { 1826 isp_prt(isp, ISP_LOGERR, 1827 "%d.%d XPT_CALC_GEOMETRY block size 0?", 1828 ccg->ccb_h.target_id, ccg->ccb_h.target_lun); 1829 ccb->ccb_h.status = CAM_REQ_INVALID; 1830 xpt_done(ccb); 1831 break; 1832 } 1833 size_mb = ccg->volume_size /((1024L * 1024L) / ccg->block_size); 1834 if (size_mb > 1024) { 1835 ccg->heads = 255; 1836 ccg->secs_per_track = 63; 1837 } else { 1838 ccg->heads = 64; 1839 ccg->secs_per_track = 32; 1840 } 1841 secs_per_cylinder = ccg->heads * ccg->secs_per_track; 1842 ccg->cylinders = ccg->volume_size / secs_per_cylinder; 1843 ccb->ccb_h.status = CAM_REQ_CMP; 1844 xpt_done(ccb); 1845 break; 1846 } 1847 case XPT_RESET_BUS: /* Reset the specified bus */ 1848 bus = cam_sim_bus(sim); 1849 ISP_LOCK(isp); 1850 error = isp_control(isp, ISPCTL_RESET_BUS, &bus); 1851 ISP_UNLOCK(isp); 1852 if (error) 1853 ccb->ccb_h.status = CAM_REQ_CMP_ERR; 1854 else { 1855 if (cam_sim_bus(sim) && isp->isp_path2 != NULL) 1856 xpt_async(AC_BUS_RESET, isp->isp_path2, NULL); 1857 else if (isp->isp_path != NULL) 1858 xpt_async(AC_BUS_RESET, isp->isp_path, NULL); 1859 ccb->ccb_h.status = CAM_REQ_CMP; 1860 } 1861 xpt_done(ccb); 1862 break; 1863 1864 case XPT_TERM_IO: /* Terminate the I/O process */ 1865 ccb->ccb_h.status = CAM_REQ_INVALID; 1866 xpt_done(ccb); 1867 break; 1868 1869 case XPT_PATH_INQ: /* Path routing inquiry */ 1870 { 1871 struct ccb_pathinq *cpi = &ccb->cpi; 1872 1873 cpi->version_num = 1; 1874#ifdef ISP_TARGET_MODE 1875 cpi->target_sprt = PIT_PROCESSOR | PIT_DISCONNECT | PIT_TERM_IO; 1876#else 1877 cpi->target_sprt = 0; 1878#endif 1879 cpi->hba_eng_cnt = 0; 1880 cpi->max_target = ISP_MAX_TARGETS(isp) - 1; 1881 cpi->max_lun = ISP_MAX_LUNS(isp) - 1; 1882 cpi->bus_id = cam_sim_bus(sim); 1883 if (IS_FC(isp)) { 1884 cpi->hba_misc = PIM_NOBUSRESET; 1885 /* 1886 * Because our loop ID can shift from time to time, 1887 * make our initiator ID out of range of our bus. 1888 */ 1889 cpi->initiator_id = cpi->max_target + 1; 1890 1891 /* 1892 * Set base transfer capabilities for Fibre Channel. 1893 * Technically not correct because we don't know 1894 * what media we're running on top of- but we'll 1895 * look good if we always say 100MB/s. 1896 */ 1897 cpi->base_transfer_speed = 100000; 1898 cpi->hba_inquiry = PI_TAG_ABLE; 1899 } else { 1900 sdparam *sdp = isp->isp_param; 1901 sdp += cam_sim_bus(xpt_path_sim(cpi->ccb_h.path)); 1902 cpi->hba_inquiry = PI_SDTR_ABLE|PI_TAG_ABLE|PI_WIDE_16; 1903 cpi->hba_misc = 0; 1904 cpi->initiator_id = sdp->isp_initiator_id; 1905 cpi->base_transfer_speed = 3300; 1906 } 1907 strncpy(cpi->sim_vid, "FreeBSD", SIM_IDLEN); 1908 strncpy(cpi->hba_vid, "Qlogic", HBA_IDLEN); 1909 strncpy(cpi->dev_name, cam_sim_name(sim), DEV_IDLEN); 1910 cpi->unit_number = cam_sim_unit(sim); 1911 cpi->ccb_h.status = CAM_REQ_CMP; 1912 xpt_done(ccb); 1913 break; 1914 } 1915 default: 1916 ccb->ccb_h.status = CAM_REQ_INVALID; 1917 xpt_done(ccb); 1918 break; 1919 } 1920} 1921 1922#define ISPDDB (CAM_DEBUG_INFO|CAM_DEBUG_TRACE|CAM_DEBUG_CDB) 1923void 1924isp_done(struct ccb_scsiio *sccb) 1925{ 1926 struct ispsoftc *isp = XS_ISP(sccb); 1927 1928 if (XS_NOERR(sccb)) 1929 XS_SETERR(sccb, CAM_REQ_CMP); 1930 1931 if ((sccb->ccb_h.status & CAM_STATUS_MASK) == CAM_REQ_CMP && 1932 (sccb->scsi_status != SCSI_STATUS_OK)) { 1933 sccb->ccb_h.status &= ~CAM_STATUS_MASK; 1934 if ((sccb->scsi_status == SCSI_STATUS_CHECK_COND) && 1935 (sccb->ccb_h.status & CAM_AUTOSNS_VALID) == 0) { 1936 sccb->ccb_h.status |= CAM_AUTOSENSE_FAIL; 1937 } else { 1938 sccb->ccb_h.status |= CAM_SCSI_STATUS_ERROR; 1939 } 1940 } 1941 1942 sccb->ccb_h.status &= ~CAM_SIM_QUEUED; 1943 if ((sccb->ccb_h.status & CAM_STATUS_MASK) != CAM_REQ_CMP) { 1944 if ((sccb->ccb_h.status & CAM_DEV_QFRZN) == 0) { 1945 sccb->ccb_h.status |= CAM_DEV_QFRZN; 1946 xpt_freeze_devq(sccb->ccb_h.path, 1); 1947 if (sccb->scsi_status != SCSI_STATUS_OK) 1948 isp_prt(isp, ISP_LOGDEBUG2, 1949 "freeze devq %d.%d %x %x", 1950 sccb->ccb_h.target_id, 1951 sccb->ccb_h.target_lun, sccb->ccb_h.status, 1952 sccb->scsi_status); 1953 } 1954 } 1955 1956 /* 1957 * If we were frozen waiting resources, clear that we were frozen 1958 * waiting for resources. If we are no longer frozen, and the devq 1959 * isn't frozen, mark the completing CCB to have the XPT layer 1960 * release the simq. 1961 */ 1962 if (isp->isp_osinfo.simqfrozen & SIMQFRZ_RESOURCE) { 1963 isp->isp_osinfo.simqfrozen &= ~SIMQFRZ_RESOURCE; 1964 if (isp->isp_osinfo.simqfrozen == 0) { 1965 if ((sccb->ccb_h.status & CAM_DEV_QFRZN) == 0) { 1966 isp_prt(isp, ISP_LOGDEBUG2, 1967 "isp_done->relsimq"); 1968 sccb->ccb_h.status |= CAM_RELEASE_SIMQ; 1969 } else { 1970 isp_prt(isp, ISP_LOGDEBUG2, 1971 "isp_done->devq frozen"); 1972 } 1973 } else { 1974 isp_prt(isp, ISP_LOGDEBUG2, 1975 "isp_done -> simqfrozen = %x", 1976 isp->isp_osinfo.simqfrozen); 1977 } 1978 } 1979 if ((CAM_DEBUGGED(sccb->ccb_h.path, ISPDDB)) && 1980 (sccb->ccb_h.status & CAM_STATUS_MASK) != CAM_REQ_CMP) { 1981 xpt_print_path(sccb->ccb_h.path); 1982 printf("cam completion status 0x%x\n", sccb->ccb_h.status); 1983 } 1984 1985 XS_CMD_S_DONE(sccb); 1986 if (XS_CMD_WDOG_P(sccb) == 0) { 1987 untimeout(isp_watchdog, (caddr_t)sccb, sccb->ccb_h.timeout_ch); 1988 if (XS_CMD_GRACE_P(sccb)) { 1989 isp_prt(isp, ISP_LOGDEBUG2, 1990 "finished command on borrowed time"); 1991 } 1992 XS_CMD_S_CLEAR(sccb); 1993 ISP_UNLOCK(isp); 1994#ifdef ISP_SMPLOCK
| 2/* 3 * Platform (FreeBSD) dependent common attachment code for Qlogic adapters. 4 * 5 * Copyright (c) 1997, 1998, 1999, 2000 by Matthew Jacob 6 * 7 * Redistribution and use in source and binary forms, with or without 8 * modification, are permitted provided that the following conditions 9 * are met: 10 * 1. Redistributions of source code must retain the above copyright 11 * notice immediately at the beginning of the file, without modification, 12 * this list of conditions, and the following disclaimer. 13 * 2. The name of the author may not be used to endorse or promote products 14 * derived from this software without specific prior written permission. 15 * 16 * THIS SOFTWARE IS PROVIDED BY THE AUTHOR AND CONTRIBUTORS ``AS IS'' AND 17 * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE 18 * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE 19 * ARE DISCLAIMED. IN NO EVENT SHALL THE AUTHOR OR CONTRIBUTORS BE LIABLE FOR 20 * ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL 21 * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS 22 * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) 23 * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT 24 * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY 25 * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF 26 * SUCH DAMAGE. 27 */ 28#include <dev/isp/isp_freebsd.h> 29#include <machine/stdarg.h> /* for use by isp_prt below */ 30 31static void isp_intr_enable(void *); 32static void isp_cam_async(void *, u_int32_t, struct cam_path *, void *); 33static void isp_poll(struct cam_sim *); 34static void isp_relsim(void *); 35static timeout_t isp_watchdog; 36static void isp_action(struct cam_sim *, union ccb *); 37 38 39static struct ispsoftc *isplist = NULL; 40 41void 42isp_attach(struct ispsoftc *isp) 43{ 44 int primary, secondary; 45 struct ccb_setasync csa; 46 struct cam_devq *devq; 47 struct cam_sim *sim; 48 struct cam_path *path; 49 50 /* 51 * Establish (in case of 12X0) which bus is the primary. 52 */ 53 54 primary = 0; 55 secondary = 1; 56 57 /* 58 * Create the device queue for our SIM(s). 59 */ 60 devq = cam_simq_alloc(isp->isp_maxcmds); 61 if (devq == NULL) { 62 return; 63 } 64 65 /* 66 * Construct our SIM entry. 67 */ 68 sim = cam_sim_alloc(isp_action, isp_poll, "isp", isp, 69 isp->isp_unit, 1, isp->isp_maxcmds, devq); 70 if (sim == NULL) { 71 cam_simq_free(devq); 72 return; 73 } 74 75 isp->isp_osinfo.ehook.ich_func = isp_intr_enable; 76 isp->isp_osinfo.ehook.ich_arg = isp; 77 if (config_intrhook_establish(&isp->isp_osinfo.ehook) != 0) { 78 isp_prt(isp, ISP_LOGERR, 79 "could not establish interrupt enable hook"); 80 cam_sim_free(sim, TRUE); 81 return; 82 } 83 84 if (xpt_bus_register(sim, primary) != CAM_SUCCESS) { 85 cam_sim_free(sim, TRUE); 86 return; 87 } 88 89 if (xpt_create_path(&path, NULL, cam_sim_path(sim), 90 CAM_TARGET_WILDCARD, CAM_LUN_WILDCARD) != CAM_REQ_CMP) { 91 xpt_bus_deregister(cam_sim_path(sim)); 92 cam_sim_free(sim, TRUE); 93 return; 94 } 95 96 xpt_setup_ccb(&csa.ccb_h, path, 5); 97 csa.ccb_h.func_code = XPT_SASYNC_CB; 98 csa.event_enable = AC_LOST_DEVICE; 99 csa.callback = isp_cam_async; 100 csa.callback_arg = sim; 101 xpt_action((union ccb *)&csa); 102 isp->isp_sim = sim; 103 isp->isp_path = path; 104 105 /* 106 * If we have a second channel, construct SIM entry for that. 107 */ 108 if (IS_DUALBUS(isp)) { 109 sim = cam_sim_alloc(isp_action, isp_poll, "isp", isp, 110 isp->isp_unit, 1, isp->isp_maxcmds, devq); 111 if (sim == NULL) { 112 xpt_bus_deregister(cam_sim_path(isp->isp_sim)); 113 xpt_free_path(isp->isp_path); 114 cam_simq_free(devq); 115 return; 116 } 117 if (xpt_bus_register(sim, secondary) != CAM_SUCCESS) { 118 xpt_bus_deregister(cam_sim_path(isp->isp_sim)); 119 xpt_free_path(isp->isp_path); 120 cam_sim_free(sim, TRUE); 121 return; 122 } 123 124 if (xpt_create_path(&path, NULL, cam_sim_path(sim), 125 CAM_TARGET_WILDCARD, CAM_LUN_WILDCARD) != CAM_REQ_CMP) { 126 xpt_bus_deregister(cam_sim_path(isp->isp_sim)); 127 xpt_free_path(isp->isp_path); 128 xpt_bus_deregister(cam_sim_path(sim)); 129 cam_sim_free(sim, TRUE); 130 return; 131 } 132 133 xpt_setup_ccb(&csa.ccb_h, path, 5); 134 csa.ccb_h.func_code = XPT_SASYNC_CB; 135 csa.event_enable = AC_LOST_DEVICE; 136 csa.callback = isp_cam_async; 137 csa.callback_arg = sim; 138 xpt_action((union ccb *)&csa); 139 isp->isp_sim2 = sim; 140 isp->isp_path2 = path; 141 } 142 isp->isp_state = ISP_RUNSTATE; 143 ENABLE_INTS(isp); 144 if (isplist == NULL) { 145 isplist = isp; 146 } else { 147 struct ispsoftc *tmp = isplist; 148 while (tmp->isp_osinfo.next) { 149 tmp = tmp->isp_osinfo.next; 150 } 151 tmp->isp_osinfo.next = isp; 152 } 153} 154 155static void 156isp_intr_enable(void *arg) 157{ 158 struct ispsoftc *isp = arg; 159 ENABLE_INTS(isp); 160 isp->isp_osinfo.intsok = 1; 161 /* Release our hook so that the boot can continue. */ 162 config_intrhook_disestablish(&isp->isp_osinfo.ehook); 163} 164 165/* 166 * Put the target mode functions here, because some are inlines 167 */ 168 169#ifdef ISP_TARGET_MODE 170 171static __inline int is_lun_enabled(struct ispsoftc *, lun_id_t); 172static __inline int are_any_luns_enabled(struct ispsoftc *); 173static __inline tstate_t *get_lun_statep(struct ispsoftc *, lun_id_t); 174static __inline void rls_lun_statep(struct ispsoftc *, tstate_t *); 175static __inline int isp_psema_sig_rqe(struct ispsoftc *); 176static __inline int isp_cv_wait_timed_rqe(struct ispsoftc *, int); 177static __inline void isp_cv_signal_rqe(struct ispsoftc *, int); 178static __inline void isp_vsema_rqe(struct ispsoftc *); 179static cam_status 180create_lun_state(struct ispsoftc *, struct cam_path *, tstate_t **); 181static void destroy_lun_state(struct ispsoftc *, tstate_t *); 182static void isp_en_lun(struct ispsoftc *, union ccb *); 183static cam_status isp_abort_tgt_ccb(struct ispsoftc *, union ccb *); 184static cam_status isp_target_start_ctio(struct ispsoftc *, union ccb *); 185static cam_status isp_target_putback_atio(struct ispsoftc *, union ccb *); 186static timeout_t isp_refire_putback_atio; 187 188static int isp_handle_platform_atio(struct ispsoftc *, at_entry_t *); 189static int isp_handle_platform_atio2(struct ispsoftc *, at2_entry_t *); 190static int isp_handle_platform_ctio(struct ispsoftc *, void *); 191static void isp_handle_platform_ctio_part2(struct ispsoftc *, union ccb *); 192 193static __inline int 194is_lun_enabled(struct ispsoftc *isp, lun_id_t lun) 195{ 196 tstate_t *tptr; 197 ISP_LOCK(isp); 198 if ((tptr = isp->isp_osinfo.lun_hash[LUN_HASH_FUNC(lun)]) == NULL) { 199 ISP_UNLOCK(isp); 200 return (0); 201 } 202 do { 203 if (tptr->lun == (lun_id_t) lun) { 204 ISP_UNLOCK(isp); 205 return (1); 206 } 207 } while ((tptr = tptr->next) != NULL); 208 ISP_UNLOCK(isp); 209 return (0); 210} 211 212static __inline int 213are_any_luns_enabled(struct ispsoftc *isp) 214{ 215 int i; 216 for (i = 0; i < LUN_HASH_SIZE; i++) { 217 if (isp->isp_osinfo.lun_hash[i]) { 218 return (1); 219 } 220 } 221 return (0); 222} 223 224static __inline tstate_t * 225get_lun_statep(struct ispsoftc *isp, lun_id_t lun) 226{ 227 tstate_t *tptr; 228 229 ISP_LOCK(isp); 230 if (lun == CAM_LUN_WILDCARD) { 231 tptr = &isp->isp_osinfo.tsdflt; 232 tptr->hold++; 233 ISP_UNLOCK(isp); 234 return (tptr); 235 } else { 236 tptr = isp->isp_osinfo.lun_hash[LUN_HASH_FUNC(lun)]; 237 } 238 if (tptr == NULL) { 239 ISP_UNLOCK(isp); 240 return (NULL); 241 } 242 243 do { 244 if (tptr->lun == lun) { 245 tptr->hold++; 246 ISP_UNLOCK(isp); 247 return (tptr); 248 } 249 } while ((tptr = tptr->next) != NULL); 250 ISP_UNLOCK(isp); 251 return (tptr); 252} 253 254static __inline void 255rls_lun_statep(struct ispsoftc *isp, tstate_t *tptr) 256{ 257 if (tptr->hold) 258 tptr->hold--; 259} 260 261static __inline int 262isp_psema_sig_rqe(struct ispsoftc *isp) 263{ 264 ISP_LOCK(isp); 265 while (isp->isp_osinfo.tmflags & TM_BUSY) { 266 isp->isp_osinfo.tmflags |= TM_WANTED; 267 if (tsleep(&isp->isp_osinfo.tmflags, PRIBIO|PCATCH, "i0", 0)) { 268 ISP_UNLOCK(isp); 269 return (-1); 270 } 271 isp->isp_osinfo.tmflags |= TM_BUSY; 272 } 273 ISP_UNLOCK(isp); 274 return (0); 275} 276 277static __inline int 278isp_cv_wait_timed_rqe(struct ispsoftc *isp, int timo) 279{ 280 ISP_LOCK(isp); 281 if (tsleep(&isp->isp_osinfo.rstatus, PRIBIO, "qt1", timo)) { 282 ISP_UNLOCK(isp); 283 return (-1); 284 } 285 ISP_UNLOCK(isp); 286 return (0); 287} 288 289static __inline void 290isp_cv_signal_rqe(struct ispsoftc *isp, int status) 291{ 292 isp->isp_osinfo.rstatus = status; 293 wakeup(&isp->isp_osinfo.rstatus); 294} 295 296static __inline void 297isp_vsema_rqe(struct ispsoftc *isp) 298{ 299 ISP_LOCK(isp); 300 if (isp->isp_osinfo.tmflags & TM_WANTED) { 301 isp->isp_osinfo.tmflags &= ~TM_WANTED; 302 wakeup(&isp->isp_osinfo.tmflags); 303 } 304 isp->isp_osinfo.tmflags &= ~TM_BUSY; 305 ISP_UNLOCK(isp); 306} 307 308static cam_status 309create_lun_state(struct ispsoftc *isp, struct cam_path *path, tstate_t **rslt) 310{ 311 cam_status status; 312 lun_id_t lun; 313 tstate_t *tptr, *new; 314 315 lun = xpt_path_lun_id(path); 316 if (lun < 0) { 317 return (CAM_LUN_INVALID); 318 } 319 if (is_lun_enabled(isp, lun)) { 320 return (CAM_LUN_ALRDY_ENA); 321 } 322 new = (tstate_t *) malloc(sizeof (tstate_t), M_DEVBUF, M_NOWAIT|M_ZERO); 323 if (new == NULL) { 324 return (CAM_RESRC_UNAVAIL); 325 } 326 327 status = xpt_create_path(&new->owner, NULL, xpt_path_path_id(path), 328 xpt_path_target_id(path), xpt_path_lun_id(path)); 329 if (status != CAM_REQ_CMP) { 330 free(new, M_DEVBUF); 331 return (status); 332 } 333 new->lun = lun; 334 SLIST_INIT(&new->atios); 335 SLIST_INIT(&new->inots); 336 new->hold = 1; 337 338 ISP_LOCK(isp); 339 if ((tptr = isp->isp_osinfo.lun_hash[LUN_HASH_FUNC(lun)]) == NULL) { 340 isp->isp_osinfo.lun_hash[LUN_HASH_FUNC(lun)] = new; 341 } else { 342 while (tptr->next) 343 tptr = tptr->next; 344 tptr->next = new; 345 } 346 ISP_UNLOCK(isp); 347 *rslt = new; 348 return (CAM_REQ_CMP); 349} 350 351static __inline void 352destroy_lun_state(struct ispsoftc *isp, tstate_t *tptr) 353{ 354 tstate_t *lw, *pw; 355 356 ISP_LOCK(isp); 357 if (tptr->hold) { 358 ISP_UNLOCK(isp); 359 return; 360 } 361 pw = isp->isp_osinfo.lun_hash[LUN_HASH_FUNC(tptr->lun)]; 362 if (pw == NULL) { 363 ISP_UNLOCK(isp); 364 return; 365 } else if (pw->lun == tptr->lun) { 366 isp->isp_osinfo.lun_hash[LUN_HASH_FUNC(tptr->lun)] = pw->next; 367 } else { 368 lw = pw; 369 pw = lw->next; 370 while (pw) { 371 if (pw->lun == tptr->lun) { 372 lw->next = pw->next; 373 break; 374 } 375 lw = pw; 376 pw = pw->next; 377 } 378 if (pw == NULL) { 379 ISP_UNLOCK(isp); 380 return; 381 } 382 } 383 free(tptr, M_DEVBUF); 384 ISP_UNLOCK(isp); 385} 386 387static void 388isp_en_lun(struct ispsoftc *isp, union ccb *ccb) 389{ 390 const char *lfmt = "Lun now %sabled for target mode\n"; 391 struct ccb_en_lun *cel = &ccb->cel; 392 tstate_t *tptr; 393 u_int16_t rstat; 394 int bus, frozen = 0; 395 lun_id_t lun; 396 target_id_t tgt; 397 398 399 bus = XS_CHANNEL(ccb); 400 tgt = ccb->ccb_h.target_id; 401 lun = ccb->ccb_h.target_lun; 402 403 /* 404 * Do some sanity checking first. 405 */ 406 407 if (lun < 0 || lun >= (lun_id_t) isp->isp_maxluns) { 408 ccb->ccb_h.status = CAM_LUN_INVALID; 409 return; 410 } 411 if (IS_SCSI(isp)) { 412 if (tgt != CAM_TARGET_WILDCARD && 413 tgt != ((sdparam *) isp->isp_param)->isp_initiator_id) { 414 ccb->ccb_h.status = CAM_TID_INVALID; 415 return; 416 } 417 } else { 418 if (tgt != CAM_TARGET_WILDCARD && 419 tgt != ((fcparam *) isp->isp_param)->isp_loopid) { 420 ccb->ccb_h.status = CAM_TID_INVALID; 421 return; 422 } 423 } 424 425 /* 426 * If Fibre Channel, stop and drain all activity to this bus. 427 */ 428 if (IS_FC(isp)) { 429 ISP_LOCK(isp); 430 frozen = 1; 431 xpt_freeze_simq(isp->isp_sim, 1); 432 isp->isp_osinfo.drain = 1; 433 /* ISP_UNLOCK(isp); XXX NEED CV_WAIT HERE XXX */ 434 while (isp->isp_osinfo.drain) { 435 tsleep(&isp->isp_osinfo.drain, PRIBIO, "ispdrain", 0); 436 } 437 ISP_UNLOCK(isp); 438 } 439 440 /* 441 * Check to see if we're enabling on fibre channel and 442 * don't yet have a notion of who the heck we are (no 443 * loop yet). 444 */ 445 if (IS_FC(isp) && cel->enable && 446 (isp->isp_osinfo.tmflags & TM_TMODE_ENABLED) == 0) { 447 int rv= 2 * 1000000; 448 fcparam *fcp = isp->isp_param; 449 450 ISP_LOCK(isp); 451 rv = isp_control(isp, ISPCTL_FCLINK_TEST, &rv); 452 ISP_UNLOCK(isp); 453 if (rv || fcp->isp_fwstate != FW_READY) { 454 xpt_print_path(ccb->ccb_h.path); 455 printf("link status not good yet\n"); 456 ccb->ccb_h.status = CAM_REQ_CMP_ERR; 457 if (frozen) 458 xpt_release_simq(isp->isp_sim, 1); 459 return; 460 } 461 ISP_LOCK(isp); 462 rv = isp_control(isp, ISPCTL_PDB_SYNC, NULL); 463 ISP_UNLOCK(isp); 464 if (rv || fcp->isp_fwstate != FW_READY) { 465 xpt_print_path(ccb->ccb_h.path); 466 printf("could not get a good port database read\n"); 467 ccb->ccb_h.status = CAM_REQ_CMP_ERR; 468 if (frozen) 469 xpt_release_simq(isp->isp_sim, 1); 470 return; 471 } 472 } 473 474 475 /* 476 * Next check to see whether this is a target/lun wildcard action. 477 * 478 * If so, we enable/disable target mode but don't do any lun enabling. 479 */ 480 if (lun == CAM_LUN_WILDCARD && tgt == CAM_TARGET_WILDCARD) { 481 int av; 482 tptr = &isp->isp_osinfo.tsdflt; 483 if (cel->enable) { 484 if (isp->isp_osinfo.tmflags & TM_TMODE_ENABLED) { 485 ccb->ccb_h.status = CAM_LUN_ALRDY_ENA; 486 if (frozen) 487 xpt_release_simq(isp->isp_sim, 1); 488 return; 489 } 490 ccb->ccb_h.status = 491 xpt_create_path(&tptr->owner, NULL, 492 xpt_path_path_id(ccb->ccb_h.path), 493 xpt_path_target_id(ccb->ccb_h.path), 494 xpt_path_lun_id(ccb->ccb_h.path)); 495 if (ccb->ccb_h.status != CAM_REQ_CMP) { 496 if (frozen) 497 xpt_release_simq(isp->isp_sim, 1); 498 return; 499 } 500 SLIST_INIT(&tptr->atios); 501 SLIST_INIT(&tptr->inots); 502 av = 1; 503 ISP_LOCK(isp); 504 av = isp_control(isp, ISPCTL_TOGGLE_TMODE, &av); 505 if (av) { 506 ccb->ccb_h.status = CAM_FUNC_NOTAVAIL; 507 xpt_free_path(tptr->owner); 508 ISP_UNLOCK(isp); 509 if (frozen) 510 xpt_release_simq(isp->isp_sim, 1); 511 return; 512 } 513 isp->isp_osinfo.tmflags |= TM_TMODE_ENABLED; 514 ISP_UNLOCK(isp); 515 } else { 516 if ((isp->isp_osinfo.tmflags & TM_TMODE_ENABLED) == 0) { 517 ccb->ccb_h.status = CAM_LUN_INVALID; 518 if (frozen) 519 xpt_release_simq(isp->isp_sim, 1); 520 return; 521 } 522 if (are_any_luns_enabled(isp)) { 523 ccb->ccb_h.status = CAM_SCSI_BUSY; 524 if (frozen) 525 xpt_release_simq(isp->isp_sim, 1); 526 return; 527 } 528 av = 0; 529 ISP_LOCK(isp); 530 av = isp_control(isp, ISPCTL_TOGGLE_TMODE, &av); 531 if (av) { 532 ccb->ccb_h.status = CAM_FUNC_NOTAVAIL; 533 ISP_UNLOCK(isp); 534 if (frozen) 535 xpt_release_simq(isp->isp_sim, 1); 536 return; 537 } 538 isp->isp_osinfo.tmflags &= ~TM_TMODE_ENABLED; 539 ISP_UNLOCK(isp); 540 ccb->ccb_h.status = CAM_REQ_CMP; 541 } 542 xpt_print_path(ccb->ccb_h.path); 543 printf(lfmt, (cel->enable) ? "en" : "dis"); 544 if (frozen) 545 xpt_release_simq(isp->isp_sim, 1); 546 return; 547 } 548 549 /* 550 * We can move along now... 551 */ 552 553 if (frozen) 554 xpt_release_simq(isp->isp_sim, 1); 555 556 557 if (cel->enable) { 558 ccb->ccb_h.status = 559 create_lun_state(isp, ccb->ccb_h.path, &tptr); 560 if (ccb->ccb_h.status != CAM_REQ_CMP) { 561 return; 562 } 563 } else { 564 tptr = get_lun_statep(isp, lun); 565 if (tptr == NULL) { 566 ccb->ccb_h.status = CAM_LUN_INVALID; 567 return; 568 } 569 } 570 571 if (isp_psema_sig_rqe(isp)) { 572 rls_lun_statep(isp, tptr); 573 if (cel->enable) 574 destroy_lun_state(isp, tptr); 575 ccb->ccb_h.status = CAM_REQ_CMP_ERR; 576 return; 577 } 578 579 ISP_LOCK(isp); 580 if (cel->enable) { 581 u_int32_t seq = isp->isp_osinfo.rollinfo++; 582 rstat = LUN_ERR; 583 if (isp_lun_cmd(isp, RQSTYPE_ENABLE_LUN, bus, tgt, lun, seq)) { 584 xpt_print_path(ccb->ccb_h.path); 585 printf("isp_lun_cmd failed\n"); 586 goto out; 587 } 588 if (isp_cv_wait_timed_rqe(isp, 30 * hz)) { 589 xpt_print_path(ccb->ccb_h.path); 590 printf("wait for ENABLE LUN timed out\n"); 591 goto out; 592 } 593 rstat = isp->isp_osinfo.rstatus; 594 if (rstat != LUN_OK) { 595 xpt_print_path(ccb->ccb_h.path); 596 printf("ENABLE LUN returned 0x%x\n", rstat); 597 goto out; 598 } 599 } else { 600 u_int32_t seq; 601 602 seq = isp->isp_osinfo.rollinfo++; 603 rstat = LUN_ERR; 604 605 if (isp_lun_cmd(isp, -RQSTYPE_MODIFY_LUN, bus, tgt, lun, seq)) { 606 xpt_print_path(ccb->ccb_h.path); 607 printf("isp_lun_cmd failed\n"); 608 goto out; 609 } 610 if (isp_cv_wait_timed_rqe(isp, 30 * hz)) { 611 xpt_print_path(ccb->ccb_h.path); 612 printf("wait for MODIFY LUN timed out\n"); 613 goto out; 614 } 615 rstat = isp->isp_osinfo.rstatus; 616 if (rstat != LUN_OK) { 617 xpt_print_path(ccb->ccb_h.path); 618 printf("MODIFY LUN returned 0x%x\n", rstat); 619 goto out; 620 } 621 rstat = LUN_ERR; 622 seq = isp->isp_osinfo.rollinfo++; 623 624 if (isp_lun_cmd(isp, -RQSTYPE_ENABLE_LUN, bus, tgt, lun, seq)) { 625 xpt_print_path(ccb->ccb_h.path); 626 printf("isp_lun_cmd failed\n"); 627 goto out; 628 } 629 if (isp_cv_wait_timed_rqe(isp, 30 * hz)) { 630 xpt_print_path(ccb->ccb_h.path); 631 printf("wait for ENABLE LUN timed out\n"); 632 goto out; 633 } 634 rstat = isp->isp_osinfo.rstatus; 635 if (rstat != LUN_OK) { 636 xpt_print_path(ccb->ccb_h.path); 637 printf("ENABLE LUN returned 0x%x\n", rstat); 638 goto out; 639 } 640 } 641out: 642 isp_vsema_rqe(isp); 643 ISP_UNLOCK(isp); 644 645 if (rstat != LUN_OK) { 646 xpt_print_path(ccb->ccb_h.path); 647 printf("lun %sable failed\n", (cel->enable) ? "en" : "dis"); 648 ccb->ccb_h.status = CAM_REQ_CMP_ERR; 649 rls_lun_statep(isp, tptr); 650 if (cel->enable) 651 destroy_lun_state(isp, tptr); 652 } else { 653 xpt_print_path(ccb->ccb_h.path); 654 printf(lfmt, (cel->enable) ? "en" : "dis"); 655 rls_lun_statep(isp, tptr); 656 if (cel->enable == 0) { 657 destroy_lun_state(isp, tptr); 658 } 659 ccb->ccb_h.status = CAM_REQ_CMP; 660 } 661} 662 663static cam_status 664isp_abort_tgt_ccb(struct ispsoftc *isp, union ccb *ccb) 665{ 666 tstate_t *tptr; 667 struct ccb_hdr_slist *lp; 668 struct ccb_hdr *curelm; 669 int found; 670 union ccb *accb = ccb->cab.abort_ccb; 671 672 if (accb->ccb_h.target_id != CAM_TARGET_WILDCARD) { 673 if (IS_FC(isp) && (accb->ccb_h.target_id != 674 ((fcparam *) isp->isp_param)->isp_loopid)) { 675 return (CAM_PATH_INVALID); 676 } else if (IS_SCSI(isp) && (accb->ccb_h.target_id != 677 ((sdparam *) isp->isp_param)->isp_initiator_id)) { 678 return (CAM_PATH_INVALID); 679 } 680 } 681 tptr = get_lun_statep(isp, accb->ccb_h.target_lun); 682 if (tptr == NULL) { 683 return (CAM_PATH_INVALID); 684 } 685 if (accb->ccb_h.func_code == XPT_ACCEPT_TARGET_IO) { 686 lp = &tptr->atios; 687 } else if (accb->ccb_h.func_code == XPT_IMMED_NOTIFY) { 688 lp = &tptr->inots; 689 } else { 690 rls_lun_statep(isp, tptr); 691 return (CAM_UA_ABORT); 692 } 693 curelm = SLIST_FIRST(lp); 694 found = 0; 695 if (curelm == &accb->ccb_h) { 696 found = 1; 697 SLIST_REMOVE_HEAD(lp, sim_links.sle); 698 } else { 699 while(curelm != NULL) { 700 struct ccb_hdr *nextelm; 701 702 nextelm = SLIST_NEXT(curelm, sim_links.sle); 703 if (nextelm == &accb->ccb_h) { 704 found = 1; 705 SLIST_NEXT(curelm, sim_links.sle) = 706 SLIST_NEXT(nextelm, sim_links.sle); 707 break; 708 } 709 curelm = nextelm; 710 } 711 } 712 rls_lun_statep(isp, tptr); 713 if (found) { 714 accb->ccb_h.status = CAM_REQ_ABORTED; 715 return (CAM_REQ_CMP); 716 } 717 return(CAM_PATH_INVALID); 718} 719 720static cam_status 721isp_target_start_ctio(struct ispsoftc *isp, union ccb *ccb) 722{ 723 void *qe; 724 struct ccb_scsiio *cso = &ccb->csio; 725 u_int32_t *hp, save_handle; 726 u_int16_t iptr, optr; 727 728 729 if (isp_getrqentry(isp, &iptr, &optr, &qe)) { 730 xpt_print_path(ccb->ccb_h.path); 731 printf("Request Queue Overflow in isp_target_start_ctio\n"); 732 return (CAM_RESRC_UNAVAIL); 733 } 734 bzero(qe, QENTRY_LEN); 735 736 /* 737 * We're either moving data or completing a command here. 738 */ 739 740 if (IS_FC(isp)) { 741 struct ccb_accept_tio *atiop; 742 ct2_entry_t *cto = qe; 743 744 cto->ct_header.rqs_entry_type = RQSTYPE_CTIO2; 745 cto->ct_header.rqs_entry_count = 1; 746 cto->ct_iid = cso->init_id; 747 if (isp->isp_maxluns <= 16) { 748 cto->ct_lun = ccb->ccb_h.target_lun; 749 } 750 /* 751 * Start with a residual based on what the original datalength 752 * was supposed to be. Basically, we ignore what CAM has set 753 * for residuals. The data transfer routines will knock off 754 * the residual for each byte actually moved- and also will 755 * be responsible for setting the underrun flag. 756 */ 757 /* HACK! HACK! */ 758 if ((atiop = ccb->ccb_h.periph_priv.entries[1].ptr) != NULL) { 759 cto->ct_resid = atiop->ccb_h.spriv_field0; 760 } 761 762 /* 763 * We always have to use the tag_id- it has the RX_ID 764 * for this exchage. 765 */ 766 cto->ct_rxid = cso->tag_id; 767 if (cso->dxfer_len == 0) { 768 cto->ct_flags |= CT2_FLAG_MODE1 | CT2_NO_DATA; 769 if (ccb->ccb_h.flags & CAM_SEND_STATUS) { 770 cto->ct_flags |= CT2_SENDSTATUS; 771 cto->rsp.m1.ct_scsi_status = cso->scsi_status; 772 } 773 if ((ccb->ccb_h.flags & CAM_SEND_SENSE) != 0) { 774 int m = min(cso->sense_len, MAXRESPLEN); 775 bcopy(&cso->sense_data, cto->rsp.m1.ct_resp, m); 776 cto->rsp.m1.ct_senselen = m; 777 cto->rsp.m1.ct_scsi_status |= CT2_SNSLEN_VALID; 778 } 779 } else { 780 cto->ct_flags |= CT2_FLAG_MODE0; 781 if ((cso->ccb_h.flags & CAM_DIR_MASK) == CAM_DIR_IN) { 782 cto->ct_flags |= CT2_DATA_IN; 783 } else { 784 cto->ct_flags |= CT2_DATA_OUT; 785 } 786 if ((ccb->ccb_h.flags & CAM_SEND_STATUS) != 0) { 787 cto->ct_flags |= CT2_SENDSTATUS; 788 cto->rsp.m0.ct_scsi_status = cso->scsi_status; 789 } 790 /* 791 * If we're sending data and status back together, 792 * we can't also send back sense data as well. 793 */ 794 ccb->ccb_h.flags &= ~CAM_SEND_SENSE; 795 } 796 if (cto->ct_flags & CAM_SEND_STATUS) { 797 isp_prt(isp, ISP_LOGTDEBUG2, 798 "CTIO2 RX_ID 0x%x SCSI STATUS 0x%x datalength %u", 799 cto->ct_rxid, cso->scsi_status, cto->ct_resid); 800 } 801 hp = &cto->ct_reserved; 802 } else { 803 ct_entry_t *cto = qe; 804 805 cto->ct_header.rqs_entry_type = RQSTYPE_CTIO; 806 cto->ct_header.rqs_entry_count = 1; 807 cto->ct_iid = cso->init_id; 808 cto->ct_tgt = ccb->ccb_h.target_id; 809 cto->ct_lun = ccb->ccb_h.target_lun; 810 if (cso->tag_id && cso->tag_action) { 811 /* 812 * We don't specify a tag type for regular SCSI. 813 * Just the tag value and set the flag. 814 */ 815 cto->ct_tag_val = cso->tag_id; 816 cto->ct_flags |= CT_TQAE; 817 } 818 if (ccb->ccb_h.flags & CAM_DIS_DISCONNECT) { 819 cto->ct_flags |= CT_NODISC; 820 } 821 if (cso->dxfer_len == 0) { 822 cto->ct_flags |= CT_NO_DATA; 823 } else if ((cso->ccb_h.flags & CAM_DIR_MASK) == CAM_DIR_IN) { 824 cto->ct_flags |= CT_DATA_IN; 825 } else { 826 cto->ct_flags |= CT_DATA_OUT; 827 } 828 if (ccb->ccb_h.flags & CAM_SEND_STATUS) { 829 cto->ct_flags |= CT_SENDSTATUS; 830 cto->ct_scsi_status = cso->scsi_status; 831 cto->ct_resid = cso->resid; 832 } 833 if (cto->ct_flags & CAM_SEND_STATUS) { 834 isp_prt(isp, ISP_LOGTDEBUG2, 835 "CTIO SCSI STATUS 0x%x resid %d", 836 cso->scsi_status, cso->resid); 837 } 838 hp = &cto->ct_reserved; 839 ccb->ccb_h.flags &= ~CAM_SEND_SENSE; 840 } 841 842 if (isp_save_xs(isp, (XS_T *)ccb, hp)) { 843 xpt_print_path(ccb->ccb_h.path); 844 printf("No XFLIST pointers for isp_target_start_ctio\n"); 845 return (CAM_RESRC_UNAVAIL); 846 } 847 848 849 /* 850 * Call the dma setup routines for this entry (and any subsequent 851 * CTIOs) if there's data to move, and then tell the f/w it's got 852 * new things to play with. As with isp_start's usage of DMA setup, 853 * any swizzling is done in the machine dependent layer. Because 854 * of this, we put the request onto the queue area first in native 855 * format. 856 */ 857 858 save_handle = *hp; 859 switch (ISP_DMASETUP(isp, cso, qe, &iptr, optr)) { 860 case CMD_QUEUED: 861 ISP_ADD_REQUEST(isp, iptr); 862 return (CAM_REQ_INPROG); 863 864 case CMD_EAGAIN: 865 ccb->ccb_h.status = CAM_RESRC_UNAVAIL; 866 isp_destroy_handle(isp, save_handle); 867 return (CAM_RESRC_UNAVAIL); 868 869 default: 870 isp_destroy_handle(isp, save_handle); 871 return (XS_ERR(ccb)); 872 } 873} 874 875static cam_status 876isp_target_putback_atio(struct ispsoftc *isp, union ccb *ccb) 877{ 878 void *qe; 879 struct ccb_accept_tio *atiop; 880 u_int16_t iptr, optr; 881 882 if (isp_getrqentry(isp, &iptr, &optr, &qe)) { 883 xpt_print_path(ccb->ccb_h.path); 884 printf("Request Queue Overflow in isp_target_putback_atio\n"); 885 return (CAM_RESRC_UNAVAIL); 886 } 887 bzero(qe, QENTRY_LEN); 888 atiop = (struct ccb_accept_tio *) ccb; 889 if (IS_FC(isp)) { 890 at2_entry_t *at = qe; 891 at->at_header.rqs_entry_type = RQSTYPE_ATIO2; 892 at->at_header.rqs_entry_count = 1; 893 if (isp->isp_maxluns > 16) { 894 at->at_scclun = (uint16_t) atiop->ccb_h.target_lun; 895 } else { 896 at->at_lun = (uint8_t) atiop->ccb_h.target_lun; 897 } 898 at->at_status = CT_OK; 899 at->at_rxid = atiop->tag_id; 900 ISP_SWIZ_ATIO2(isp, qe, qe); 901 } else { 902 at_entry_t *at = qe; 903 at->at_header.rqs_entry_type = RQSTYPE_ATIO; 904 at->at_header.rqs_entry_count = 1; 905 at->at_iid = atiop->init_id; 906 at->at_tgt = atiop->ccb_h.target_id; 907 at->at_lun = atiop->ccb_h.target_lun; 908 at->at_status = CT_OK; 909 if (atiop->ccb_h.status & CAM_TAG_ACTION_VALID) { 910 at->at_tag_type = atiop->tag_action; 911 } 912 at->at_tag_val = atiop->tag_id; 913 ISP_SWIZ_ATIO(isp, qe, qe); 914 } 915 ISP_TDQE(isp, "isp_target_putback_atio", (int) optr, qe); 916 ISP_ADD_REQUEST(isp, iptr); 917 return (CAM_REQ_CMP); 918} 919 920static void 921isp_refire_putback_atio(void *arg) 922{ 923 union ccb *ccb = arg; 924 int s = splcam(); 925 if (isp_target_putback_atio(XS_ISP(ccb), ccb) != CAM_REQ_CMP) { 926 (void) timeout(isp_refire_putback_atio, ccb, 10); 927 } else { 928 isp_handle_platform_ctio_part2(XS_ISP(ccb), ccb); 929 } 930 splx(s); 931} 932 933/* 934 * Handle ATIO stuff that the generic code can't. 935 * This means handling CDBs. 936 */ 937 938static int 939isp_handle_platform_atio(struct ispsoftc *isp, at_entry_t *aep) 940{ 941 tstate_t *tptr; 942 int status; 943 struct ccb_accept_tio *atiop; 944 945 /* 946 * The firmware status (except for the QLTM_SVALID bit) 947 * indicates why this ATIO was sent to us. 948 * 949 * If QLTM_SVALID is set, the firware has recommended Sense Data. 950 * 951 * If the DISCONNECTS DISABLED bit is set in the flags field, 952 * we're still connected on the SCSI bus - i.e. the initiator 953 * did not set DiscPriv in the identify message. We don't care 954 * about this so it's ignored. 955 */ 956 status = aep->at_status; 957 if ((status & ~QLTM_SVALID) == AT_PHASE_ERROR) { 958 /* 959 * Bus Phase Sequence error. We should have sense data 960 * suggested by the f/w. I'm not sure quite yet what 961 * to do about this for CAM. 962 */ 963 printf("%s: PHASE ERROR\n", isp->isp_name); 964 isp_endcmd(isp, aep, SCSI_STATUS_BUSY, 0); 965 return (0); 966 } 967 if ((status & ~QLTM_SVALID) != AT_CDB) { 968 printf("%s: bogus atio (0x%x) leaked to platform\n", 969 isp->isp_name, status); 970 isp_endcmd(isp, aep, SCSI_STATUS_BUSY, 0); 971 return (0); 972 } 973 974 tptr = get_lun_statep(isp, aep->at_lun); 975 if (tptr == NULL) { 976 tptr = get_lun_statep(isp, CAM_LUN_WILDCARD); 977 } 978 979 if (tptr == NULL) { 980 /* 981 * Because we can't autofeed sense data back with 982 * a command for parallel SCSI, we can't give back 983 * a CHECK CONDITION. We'll give back a BUSY status 984 * instead. This works out okay because the only 985 * time we should, in fact, get this, is in the 986 * case that somebody configured us without the 987 * blackhole driver, so they get what they deserve. 988 */ 989 isp_endcmd(isp, aep, SCSI_STATUS_BUSY, 0); 990 return (0); 991 } 992 993 atiop = (struct ccb_accept_tio *) SLIST_FIRST(&tptr->atios); 994 if (atiop == NULL) { 995 /* 996 * Because we can't autofeed sense data back with 997 * a command for parallel SCSI, we can't give back 998 * a CHECK CONDITION. We'll give back a QUEUE FULL status 999 * instead. This works out okay because the only time we 1000 * should, in fact, get this, is in the case that we've 1001 * run out of ATIOS. 1002 */ 1003 xpt_print_path(tptr->owner); 1004 printf("no ATIOS for lun %d from initiator %d\n", 1005 aep->at_lun, aep->at_iid); 1006 rls_lun_statep(isp, tptr); 1007 if (aep->at_flags & AT_TQAE) 1008 isp_endcmd(isp, aep, SCSI_STATUS_QUEUE_FULL, 0); 1009 else 1010 isp_endcmd(isp, aep, SCSI_STATUS_BUSY, 0); 1011 return (0); 1012 } 1013 SLIST_REMOVE_HEAD(&tptr->atios, sim_links.sle); 1014 if (tptr == &isp->isp_osinfo.tsdflt) { 1015 atiop->ccb_h.target_id = aep->at_tgt; 1016 atiop->ccb_h.target_lun = aep->at_lun; 1017 } 1018 if (aep->at_flags & AT_NODISC) { 1019 atiop->ccb_h.flags = CAM_DIS_DISCONNECT; 1020 } else { 1021 atiop->ccb_h.flags = 0; 1022 } 1023 1024 if (status & QLTM_SVALID) { 1025 size_t amt = imin(QLTM_SENSELEN, sizeof (atiop->sense_data)); 1026 atiop->sense_len = amt; 1027 MEMCPY(&atiop->sense_data, aep->at_sense, amt); 1028 } else { 1029 atiop->sense_len = 0; 1030 } 1031 1032 atiop->init_id = aep->at_iid; 1033 atiop->cdb_len = aep->at_cdblen; 1034 MEMCPY(atiop->cdb_io.cdb_bytes, aep->at_cdb, aep->at_cdblen); 1035 atiop->ccb_h.status = CAM_CDB_RECVD; 1036 atiop->tag_id = aep->at_tag_val; 1037 if ((atiop->tag_action = aep->at_tag_type) != 0) { 1038 atiop->ccb_h.status |= CAM_TAG_ACTION_VALID; 1039 } 1040 xpt_done((union ccb*)atiop); 1041 isp_prt(isp, ISP_LOGTDEBUG2, 1042 "ATIO CDB=0x%x iid%d->lun%d tag 0x%x ttype 0x%x %s", 1043 aep->at_cdb[0] & 0xff, aep->at_iid, aep->at_lun, 1044 aep->at_tag_val & 0xff, aep->at_tag_type, 1045 (aep->at_flags & AT_NODISC)? "nondisc" : "disconnecting"); 1046 rls_lun_statep(isp, tptr); 1047 return (0); 1048} 1049 1050static int 1051isp_handle_platform_atio2(struct ispsoftc *isp, at2_entry_t *aep) 1052{ 1053 lun_id_t lun; 1054 tstate_t *tptr; 1055 struct ccb_accept_tio *atiop; 1056 1057 /* 1058 * The firmware status (except for the QLTM_SVALID bit) 1059 * indicates why this ATIO was sent to us. 1060 * 1061 * If QLTM_SVALID is set, the firware has recommended Sense Data. 1062 */ 1063 if ((aep->at_status & ~QLTM_SVALID) != AT_CDB) { 1064 printf("%s: bogus atio (0x%x) leaked to platform\n", 1065 isp->isp_name, aep->at_status); 1066 isp_endcmd(isp, aep, SCSI_STATUS_BUSY, 0); 1067 return (0); 1068 } 1069 1070 if (isp->isp_maxluns > 16) { 1071 lun = aep->at_scclun; 1072 } else { 1073 lun = aep->at_lun; 1074 } 1075 tptr = get_lun_statep(isp, lun); 1076 if (tptr == NULL) { 1077 tptr = get_lun_statep(isp, CAM_LUN_WILDCARD); 1078 } 1079 1080 if (tptr == NULL) { 1081 /* 1082 * What we'd like to know is whether or not we have a listener 1083 * upstream that really hasn't configured yet. If we do, then 1084 * we can give a more sensible reply here. If not, then we can 1085 * reject this out of hand. 1086 * 1087 * Choices for what to send were 1088 * 1089 * Not Ready, Unit Not Self-Configured Yet 1090 * (0x2,0x3e,0x00) 1091 * 1092 * for the former and 1093 * 1094 * Illegal Request, Logical Unit Not Supported 1095 * (0x5,0x25,0x00) 1096 * 1097 * for the latter. 1098 * 1099 * We used to decide whether there was at least one listener 1100 * based upon whether the black hole driver was configured. 1101 * However, recent config(8) changes have made this hard to do 1102 * at this time. 1103 * 1104 */ 1105 u_int32_t ccode = SCSI_STATUS_BUSY; 1106 1107 /* 1108 * Because we can't autofeed sense data back with 1109 * a command for parallel SCSI, we can't give back 1110 * a CHECK CONDITION. We'll give back a BUSY status 1111 * instead. This works out okay because the only 1112 * time we should, in fact, get this, is in the 1113 * case that somebody configured us without the 1114 * blackhole driver, so they get what they deserve. 1115 */ 1116 isp_endcmd(isp, aep, ccode, 0); 1117 return (0); 1118 } 1119 1120 atiop = (struct ccb_accept_tio *) SLIST_FIRST(&tptr->atios); 1121 if (atiop == NULL) { 1122 /* 1123 * Because we can't autofeed sense data back with 1124 * a command for parallel SCSI, we can't give back 1125 * a CHECK CONDITION. We'll give back a QUEUE FULL status 1126 * instead. This works out okay because the only time we 1127 * should, in fact, get this, is in the case that we've 1128 * run out of ATIOS. 1129 */ 1130 xpt_print_path(tptr->owner); 1131 printf("no ATIOS for lun %d from initiator %d\n", 1132 lun, aep->at_iid); 1133 rls_lun_statep(isp, tptr); 1134 if (aep->at_flags & AT_TQAE) 1135 isp_endcmd(isp, aep, SCSI_STATUS_QUEUE_FULL, 0); 1136 else 1137 isp_endcmd(isp, aep, SCSI_STATUS_BUSY, 0); 1138 return (0); 1139 } 1140 SLIST_REMOVE_HEAD(&tptr->atios, sim_links.sle); 1141 1142 if (tptr == &isp->isp_osinfo.tsdflt) { 1143 atiop->ccb_h.target_id = 1144 ((fcparam *)isp->isp_param)->isp_loopid; 1145 atiop->ccb_h.target_lun = lun; 1146 } 1147 if (aep->at_status & QLTM_SVALID) { 1148 size_t amt = imin(QLTM_SENSELEN, sizeof (atiop->sense_data)); 1149 atiop->sense_len = amt; 1150 MEMCPY(&atiop->sense_data, aep->at_sense, amt); 1151 } else { 1152 atiop->sense_len = 0; 1153 } 1154 1155 atiop->init_id = aep->at_iid; 1156 atiop->cdb_len = ATIO2_CDBLEN; 1157 MEMCPY(atiop->cdb_io.cdb_bytes, aep->at_cdb, ATIO2_CDBLEN); 1158 atiop->ccb_h.status = CAM_CDB_RECVD; 1159 atiop->tag_id = aep->at_rxid; 1160 switch (aep->at_taskflags & ATIO2_TC_ATTR_MASK) { 1161 case ATIO2_TC_ATTR_SIMPLEQ: 1162 atiop->tag_action = MSG_SIMPLE_Q_TAG; 1163 break; 1164 case ATIO2_TC_ATTR_HEADOFQ: 1165 atiop->tag_action = MSG_HEAD_OF_Q_TAG; 1166 break; 1167 case ATIO2_TC_ATTR_ORDERED: 1168 atiop->tag_action = MSG_ORDERED_Q_TAG; 1169 break; 1170 case ATIO2_TC_ATTR_ACAQ: /* ?? */ 1171 case ATIO2_TC_ATTR_UNTAGGED: 1172 default: 1173 atiop->tag_action = 0; 1174 break; 1175 } 1176 if (atiop->tag_action != 0) { 1177 atiop->ccb_h.status |= CAM_TAG_ACTION_VALID; 1178 } 1179 1180 /* 1181 * Preserve overall command datalength in private field. 1182 */ 1183 atiop->ccb_h.spriv_field0 = aep->at_datalen; 1184 1185 xpt_done((union ccb*)atiop); 1186 isp_prt(isp, ISP_LOGTDEBUG2, 1187 "ATIO2 RX_ID 0x%x CDB=0x%x iid%d->lun%d tattr 0x%x datalen %u", 1188 aep->at_rxid & 0xffff, aep->at_cdb[0] & 0xff, aep->at_iid, 1189 lun, aep->at_taskflags, aep->at_datalen); 1190 rls_lun_statep(isp, tptr); 1191 return (0); 1192} 1193 1194static int 1195isp_handle_platform_ctio(struct ispsoftc *isp, void *arg) 1196{ 1197 union ccb *ccb; 1198 int sentstatus, ok, notify_cam; 1199 1200 /* 1201 * CTIO and CTIO2 are close enough.... 1202 */ 1203 1204 ccb = (union ccb *) isp_find_xs(isp, ((ct_entry_t *)arg)->ct_reserved); 1205 KASSERT((ccb != NULL), ("null ccb in isp_handle_platform_ctio")); 1206 isp_destroy_handle(isp, ((ct_entry_t *)arg)->ct_reserved); 1207 1208 if (IS_FC(isp)) { 1209 ct2_entry_t *ct = arg; 1210 sentstatus = ct->ct_flags & CT2_SENDSTATUS; 1211 ok = (ct->ct_status & ~QLTM_SVALID) == CT_OK; 1212 if (ok && (ccb->ccb_h.flags & CAM_SEND_SENSE)) { 1213 ccb->ccb_h.status |= CAM_SENT_SENSE; 1214 } 1215 isp_prt(isp, ISP_LOGTDEBUG2, 1216 "CTIO2 RX_ID 0x%x sts 0x%x flg 0x%x sns %d FIN", 1217 ct->ct_rxid, ct->ct_status, ct->ct_flags, 1218 (ccb->ccb_h.status & CAM_SENT_SENSE) != 0); 1219 notify_cam = ct->ct_header.rqs_seqno; 1220 } else { 1221 ct_entry_t *ct = arg; 1222 sentstatus = ct->ct_flags & CT_SENDSTATUS; 1223 ok = (ct->ct_status & ~QLTM_SVALID) == CT_OK; 1224 isp_prt(isp, ISP_LOGTDEBUG2, 1225 "CTIO tag 0x%x sts 0x%x flg 0x%x FIN", 1226 ct->ct_tag_val, ct->ct_status, ct->ct_flags); 1227 notify_cam = ct->ct_header.rqs_seqno; 1228 } 1229 1230 /* 1231 * We're here either because data transfers are done (and 1232 * it's time to send a final status CTIO) or because the final 1233 * status CTIO is done. We don't get called for all intermediate 1234 * CTIOs that happen for a large data transfer. 1235 * 1236 * In any case, for this platform, the upper layers figure out 1237 * what to do next, so all we do here is collect status and 1238 * pass information along. The exception is that we clear 1239 * the notion of handling a non-disconnecting command here. 1240 */ 1241 1242 if (sentstatus) { 1243 /* 1244 * Data transfer done. See if all went okay. 1245 */ 1246 if (ok) { 1247 ccb->csio.resid = 0; 1248 } else { 1249 ccb->csio.resid = ccb->csio.dxfer_len; 1250 } 1251 } 1252 1253 if (notify_cam == 0) { 1254 isp_prt(isp, ISP_LOGTDEBUG1, "Intermediate CTIO done"); 1255 return (0); 1256 } 1257 isp_prt(isp, ISP_LOGTDEBUG1, "Final CTIO done"); 1258 if (isp_target_putback_atio(isp, ccb) != CAM_REQ_CMP) { 1259 (void) timeout(isp_refire_putback_atio, ccb, 10); 1260 } else { 1261 isp_handle_platform_ctio_part2(isp, ccb); 1262 } 1263 return (0); 1264} 1265 1266static void 1267isp_handle_platform_ctio_part2(struct ispsoftc *isp, union ccb *ccb) 1268{ 1269 if ((ccb->ccb_h.status & CAM_STATUS_MASK) == CAM_REQ_INPROG) { 1270 ccb->ccb_h.status |= CAM_REQ_CMP; 1271 } 1272 ccb->ccb_h.status &= ~CAM_SIM_QUEUED; 1273 if (isp->isp_osinfo.simqfrozen & SIMQFRZ_RESOURCE) { 1274 isp->isp_osinfo.simqfrozen &= ~SIMQFRZ_RESOURCE; 1275 if (isp->isp_osinfo.simqfrozen == 0) { 1276 if ((ccb->ccb_h.status & CAM_DEV_QFRZN) == 0) { 1277 isp_prt(isp, ISP_LOGDEBUG2, "ctio->relsimq"); 1278 ccb->ccb_h.status |= CAM_RELEASE_SIMQ; 1279 } else { 1280 isp_prt(isp, ISP_LOGDEBUG2, "ctio->devqfrozen"); 1281 } 1282 } else { 1283 isp_prt(isp, ISP_LOGDEBUG2, 1284 "ctio->simqfrozen(%x)", isp->isp_osinfo.simqfrozen); 1285 } 1286 } 1287 xpt_done(ccb); 1288} 1289#endif 1290 1291static void 1292isp_cam_async(void *cbarg, u_int32_t code, struct cam_path *path, void *arg) 1293{ 1294 struct cam_sim *sim; 1295 struct ispsoftc *isp; 1296 1297 sim = (struct cam_sim *)cbarg; 1298 isp = (struct ispsoftc *) cam_sim_softc(sim); 1299 switch (code) { 1300 case AC_LOST_DEVICE: 1301 if (IS_SCSI(isp)) { 1302 u_int16_t oflags, nflags; 1303 sdparam *sdp = isp->isp_param; 1304 int rvf, tgt; 1305 1306 tgt = xpt_path_target_id(path); 1307 rvf = ISP_FW_REVX(isp->isp_fwrev); 1308 ISP_LOCK(isp); 1309 sdp += cam_sim_bus(sim); 1310 isp->isp_update |= (1 << cam_sim_bus(sim)); 1311 nflags = DPARM_SAFE_DFLT; 1312 if (rvf >= ISP_FW_REV(7, 55, 0) || 1313 (ISP_FW_REV(4, 55, 0) <= rvf && 1314 (rvf < ISP_FW_REV(5, 0, 0)))) { 1315 nflags |= DPARM_NARROW | DPARM_ASYNC; 1316 } 1317 oflags = sdp->isp_devparam[tgt].dev_flags; 1318 sdp->isp_devparam[tgt].dev_flags = nflags; 1319 sdp->isp_devparam[tgt].dev_update = 1; 1320 (void) isp_control(isp, ISPCTL_UPDATE_PARAMS, NULL); 1321 sdp->isp_devparam[tgt].dev_flags = oflags; 1322 ISP_UNLOCK(isp); 1323 } 1324 break; 1325 default: 1326 printf("%s: isp_attach Async Code 0x%x\n", isp->isp_name, code); 1327 break; 1328 } 1329} 1330 1331static void 1332isp_poll(struct cam_sim *sim) 1333{ 1334 struct ispsoftc *isp = cam_sim_softc(sim); 1335 ISP_LOCK(isp); 1336 (void) isp_intr(isp); 1337 ISP_UNLOCK(isp); 1338} 1339 1340static void 1341isp_relsim(void *arg) 1342{ 1343 struct ispsoftc *isp = arg; 1344 ISP_LOCK(isp); 1345 if (isp->isp_osinfo.simqfrozen & SIMQFRZ_TIMED) { 1346 int wasfrozen = isp->isp_osinfo.simqfrozen & SIMQFRZ_TIMED; 1347 isp->isp_osinfo.simqfrozen &= ~SIMQFRZ_TIMED; 1348 if (wasfrozen && isp->isp_osinfo.simqfrozen == 0) { 1349 xpt_release_simq(isp->isp_sim, 1); 1350 isp_prt(isp, ISP_LOGDEBUG2, "timed relsimq"); 1351 } 1352 } 1353 ISP_UNLOCK(isp); 1354} 1355 1356static void 1357isp_watchdog(void *arg) 1358{ 1359 XS_T *xs = arg; 1360 struct ispsoftc *isp = XS_ISP(xs); 1361 u_int32_t handle; 1362 1363 /* 1364 * We've decided this command is dead. Make sure we're not trying 1365 * to kill a command that's already dead by getting it's handle and 1366 * and seeing whether it's still alive. 1367 */ 1368 ISP_LOCK(isp); 1369 handle = isp_find_handle(isp, xs); 1370 if (handle) { 1371 u_int16_t r; 1372 1373 if (XS_CMD_DONE_P(xs)) { 1374 isp_prt(isp, ISP_LOGDEBUG1, 1375 "watchdog found done cmd (handle 0x%x)", handle); 1376 ISP_UNLOCK(isp); 1377 return; 1378 } 1379 1380 if (XS_CMD_WDOG_P(xs)) { 1381 isp_prt(isp, ISP_LOGDEBUG2, 1382 "recursive watchdog (handle 0x%x)", handle); 1383 ISP_UNLOCK(isp); 1384 return; 1385 } 1386 1387 XS_CMD_S_WDOG(xs); 1388 1389 r = ISP_READ(isp, BIU_ISR); 1390 1391 if (INT_PENDING(isp, r) && isp_intr(isp) && XS_CMD_DONE_P(xs)) { 1392 isp_prt(isp, ISP_LOGDEBUG2, 1393 "watchdog cleanup (%x, %x)", handle, r); 1394 xpt_done((union ccb *) xs); 1395 } else if (XS_CMD_GRACE_P(xs)) { 1396 /* 1397 * Make sure the command is *really* dead before we 1398 * release the handle (and DMA resources) for reuse. 1399 */ 1400 (void) isp_control(isp, ISPCTL_ABORT_CMD, arg); 1401 1402 /* 1403 * After this point, the comamnd is really dead. 1404 */ 1405 if (XS_XFRLEN(xs)) { 1406 ISP_DMAFREE(isp, xs, handle); 1407 } 1408 isp_destroy_handle(isp, handle); 1409 xpt_print_path(xs->ccb_h.path); 1410 printf("%s: watchdog timeout (%x, %x)\n", 1411 isp->isp_name, handle, r); 1412 XS_SETERR(xs, CAM_CMD_TIMEOUT); 1413 XS_CMD_C_WDOG(xs); 1414 isp_done(xs); 1415 } else { 1416 u_int16_t iptr, optr; 1417 ispreq_t *mp; 1418 1419 XS_CMD_C_WDOG(xs); 1420 xs->ccb_h.timeout_ch = timeout(isp_watchdog, xs, hz); 1421 if (isp_getrqentry(isp, &iptr, &optr, (void **) &mp)) { 1422 ISP_UNLOCK(isp); 1423 return; 1424 } 1425 XS_CMD_S_GRACE(xs); 1426 MEMZERO((void *) mp, sizeof (*mp)); 1427 mp->req_header.rqs_entry_count = 1; 1428 mp->req_header.rqs_entry_type = RQSTYPE_MARKER; 1429 mp->req_modifier = SYNC_ALL; 1430 mp->req_target = XS_CHANNEL(xs) << 7; 1431 ISP_SWIZZLE_REQUEST(isp, mp); 1432 ISP_ADD_REQUEST(isp, iptr); 1433 } 1434 } else { 1435 isp_prt(isp, ISP_LOGDEBUG2, "watchdog with no command"); 1436 } 1437 ISP_UNLOCK(isp); 1438} 1439 1440static void 1441isp_action(struct cam_sim *sim, union ccb *ccb) 1442{ 1443 int bus, tgt, error; 1444 struct ispsoftc *isp; 1445 struct ccb_trans_settings *cts; 1446 1447 CAM_DEBUG(ccb->ccb_h.path, CAM_DEBUG_TRACE, ("isp_action\n")); 1448 1449 isp = (struct ispsoftc *)cam_sim_softc(sim); 1450 ccb->ccb_h.sim_priv.entries[0].field = 0; 1451 ccb->ccb_h.sim_priv.entries[1].ptr = isp; 1452 if (isp->isp_state != ISP_RUNSTATE && 1453 ccb->ccb_h.func_code == XPT_SCSI_IO) { 1454 ISP_LOCK(isp); 1455 isp_init(isp); 1456 if (isp->isp_state != ISP_INITSTATE) { 1457 ISP_UNLOCK(isp); 1458 /* 1459 * Lie. Say it was a selection timeout. 1460 */ 1461 ccb->ccb_h.status = CAM_SEL_TIMEOUT | CAM_DEV_QFRZN; 1462 xpt_freeze_devq(ccb->ccb_h.path, 1); 1463 xpt_done(ccb); 1464 return; 1465 } 1466 isp->isp_state = ISP_RUNSTATE; 1467 ISP_UNLOCK(isp); 1468 } 1469 isp_prt(isp, ISP_LOGDEBUG2, "isp_action code %x", ccb->ccb_h.func_code); 1470 1471 switch (ccb->ccb_h.func_code) { 1472 case XPT_SCSI_IO: /* Execute the requested I/O operation */ 1473 /* 1474 * Do a couple of preliminary checks... 1475 */ 1476 if ((ccb->ccb_h.flags & CAM_CDB_POINTER) != 0) { 1477 if ((ccb->ccb_h.flags & CAM_CDB_PHYS) != 0) { 1478 ccb->ccb_h.status = CAM_REQ_INVALID; 1479 xpt_done(ccb); 1480 break; 1481 } 1482 } 1483#ifdef DIAGNOSTIC 1484 if (ccb->ccb_h.target_id > (ISP_MAX_TARGETS(isp) - 1)) { 1485 ccb->ccb_h.status = CAM_PATH_INVALID; 1486 } else if (ccb->ccb_h.target_lun > (ISP_MAX_LUNS(isp) - 1)) { 1487 ccb->ccb_h.status = CAM_PATH_INVALID; 1488 } 1489 if (ccb->ccb_h.status == CAM_PATH_INVALID) { 1490 isp_prt(isp, ISP_LOGERR, 1491 "invalid tgt/lun (%d.%d) in XPT_SCSI_IO", 1492 ccb->ccb_h.target_id, ccb->ccb_h.target_lun); 1493 xpt_done(ccb); 1494 break; 1495 } 1496#endif 1497 ((struct ccb_scsiio *) ccb)->scsi_status = SCSI_STATUS_OK; 1498 ISP_LOCK(isp); 1499 error = isp_start((XS_T *) ccb); 1500 ISP_UNLOCK(isp); 1501 switch (error) { 1502 case CMD_QUEUED: 1503 ccb->ccb_h.status |= CAM_SIM_QUEUED; 1504 if (ccb->ccb_h.timeout != CAM_TIME_INFINITY) { 1505 u_int64_t ticks = (u_int64_t) hz; 1506 if (ccb->ccb_h.timeout == CAM_TIME_DEFAULT) 1507 ticks = 60 * 1000 * ticks; 1508 else 1509 ticks = ccb->ccb_h.timeout * hz; 1510 ticks = ((ticks + 999) / 1000) + hz + hz; 1511 if (ticks >= 0x80000000) { 1512 isp_prt(isp, ISP_LOGERR, 1513 "timeout overflow"); 1514 ticks = 0x80000000; 1515 } 1516 ccb->ccb_h.timeout_ch = timeout(isp_watchdog, 1517 (caddr_t)ccb, (int)ticks); 1518 } else { 1519 callout_handle_init(&ccb->ccb_h.timeout_ch); 1520 } 1521 break; 1522 case CMD_RQLATER: 1523 if (isp->isp_osinfo.simqfrozen == 0) { 1524 isp_prt(isp, ISP_LOGDEBUG2, 1525 "RQLATER freeze simq"); 1526 isp->isp_osinfo.simqfrozen |= SIMQFRZ_TIMED; 1527 timeout(isp_relsim, isp, 500); 1528 xpt_freeze_simq(sim, 1); 1529 } 1530 XS_SETERR(ccb, CAM_REQUEUE_REQ); 1531 xpt_done(ccb); 1532 break; 1533 case CMD_EAGAIN: 1534 if (isp->isp_osinfo.simqfrozen == 0) { 1535 xpt_freeze_simq(sim, 1); 1536 isp_prt(isp, ISP_LOGDEBUG2, 1537 "EAGAIN freeze simq"); 1538 } 1539 isp->isp_osinfo.simqfrozen |= SIMQFRZ_RESOURCE; 1540 XS_SETERR(ccb, CAM_REQUEUE_REQ); 1541 xpt_done(ccb); 1542 break; 1543 case CMD_COMPLETE: 1544 ISP_LOCK(isp); 1545 isp_done((struct ccb_scsiio *) ccb); 1546 ISP_UNLOCK(isp); 1547 break; 1548 default: 1549 isp_prt(isp, ISP_LOGERR, 1550 "What's this? 0x%x at %d in file %s", 1551 error, __LINE__, __FILE__); 1552 XS_SETERR(ccb, CAM_REQ_CMP_ERR); 1553 xpt_done(ccb); 1554 } 1555 break; 1556 1557#ifdef ISP_TARGET_MODE 1558 case XPT_EN_LUN: /* Enable LUN as a target */ 1559 isp_en_lun(isp, ccb); 1560 xpt_done(ccb); 1561 break; 1562 1563 case XPT_NOTIFY_ACK: /* recycle notify ack */ 1564 case XPT_IMMED_NOTIFY: /* Add Immediate Notify Resource */ 1565 case XPT_ACCEPT_TARGET_IO: /* Add Accept Target IO Resource */ 1566 { 1567 tstate_t *tptr = get_lun_statep(isp, ccb->ccb_h.target_lun); 1568 if (tptr == NULL) { 1569 ccb->ccb_h.status = CAM_LUN_INVALID; 1570 xpt_done(ccb); 1571 break; 1572 } 1573 ccb->ccb_h.sim_priv.entries[0].field = 0; 1574 ccb->ccb_h.sim_priv.entries[1].ptr = isp; 1575 ISP_LOCK(isp); 1576 if (ccb->ccb_h.func_code == XPT_ACCEPT_TARGET_IO) { 1577#if 0 1578 (void) isp_target_putback_atio(isp, ccb); 1579#endif 1580 SLIST_INSERT_HEAD(&tptr->atios, 1581 &ccb->ccb_h, sim_links.sle); 1582 } else { 1583 SLIST_INSERT_HEAD(&tptr->inots, &ccb->ccb_h, 1584 sim_links.sle); 1585 } 1586 ISP_UNLOCK(isp); 1587 rls_lun_statep(isp, tptr); 1588 ccb->ccb_h.status = CAM_REQ_INPROG; 1589 break; 1590 } 1591 case XPT_CONT_TARGET_IO: 1592 { 1593 ISP_LOCK(isp); 1594 ccb->ccb_h.status = isp_target_start_ctio(isp, ccb); 1595 if (ccb->ccb_h.status != CAM_REQ_INPROG) { 1596 if (isp->isp_osinfo.simqfrozen == 0) { 1597 xpt_freeze_simq(sim, 1); 1598 xpt_print_path(ccb->ccb_h.path); 1599 printf("XPT_CONT_TARGET_IO freeze simq\n"); 1600 } 1601 isp->isp_osinfo.simqfrozen |= SIMQFRZ_RESOURCE; 1602 XS_SETERR(ccb, CAM_REQUEUE_REQ); 1603 xpt_done(ccb); 1604 } else { 1605 ccb->ccb_h.status |= CAM_SIM_QUEUED; 1606 } 1607 ISP_UNLOCK(isp); 1608 break; 1609 } 1610#endif 1611 case XPT_RESET_DEV: /* BDR the specified SCSI device */ 1612 1613 bus = cam_sim_bus(xpt_path_sim(ccb->ccb_h.path)); 1614 tgt = ccb->ccb_h.target_id; 1615 tgt |= (bus << 16); 1616 1617 ISP_LOCK(isp); 1618 error = isp_control(isp, ISPCTL_RESET_DEV, &tgt); 1619 ISP_UNLOCK(isp); 1620 if (error) { 1621 ccb->ccb_h.status = CAM_REQ_CMP_ERR; 1622 } else { 1623 ccb->ccb_h.status = CAM_REQ_CMP; 1624 } 1625 xpt_done(ccb); 1626 break; 1627 case XPT_ABORT: /* Abort the specified CCB */ 1628 { 1629 union ccb *accb = ccb->cab.abort_ccb; 1630 switch (accb->ccb_h.func_code) { 1631#ifdef ISP_TARGET_MODE 1632 case XPT_ACCEPT_TARGET_IO: 1633 case XPT_IMMED_NOTIFY: 1634 ccb->ccb_h.status = isp_abort_tgt_ccb(isp, ccb); 1635 break; 1636 case XPT_CONT_TARGET_IO: 1637 isp_prt(isp, ISP_LOGERR, "cannot abort CTIOs yet"); 1638 ccb->ccb_h.status = CAM_UA_ABORT; 1639 break; 1640#endif 1641 case XPT_SCSI_IO: 1642 ISP_LOCK(isp); 1643 error = isp_control(isp, ISPCTL_ABORT_CMD, ccb); 1644 ISP_UNLOCK(isp); 1645 if (error) { 1646 ccb->ccb_h.status = CAM_UA_ABORT; 1647 } else { 1648 ccb->ccb_h.status = CAM_REQ_CMP; 1649 } 1650 break; 1651 default: 1652 ccb->ccb_h.status = CAM_REQ_INVALID; 1653 break; 1654 } 1655 xpt_done(ccb); 1656 break; 1657 } 1658 case XPT_SET_TRAN_SETTINGS: /* Nexus Settings */ 1659 1660 cts = &ccb->cts; 1661 tgt = cts->ccb_h.target_id; 1662 ISP_LOCK(isp); 1663 if (IS_SCSI(isp)) { 1664 sdparam *sdp = isp->isp_param; 1665 u_int16_t *dptr; 1666 1667 bus = cam_sim_bus(xpt_path_sim(cts->ccb_h.path)); 1668 1669 sdp += bus; 1670#if 0 1671 if (cts->flags & CCB_TRANS_CURRENT_SETTINGS) 1672 dptr = &sdp->isp_devparam[tgt].cur_dflags; 1673 else 1674 dptr = &sdp->isp_devparam[tgt].dev_flags; 1675#else 1676 /* 1677 * We always update (internally) from dev_flags 1678 * so any request to change settings just gets 1679 * vectored to that location. 1680 */ 1681 dptr = &sdp->isp_devparam[tgt].dev_flags; 1682#endif 1683 1684 /* 1685 * Note that these operations affect the 1686 * the goal flags (dev_flags)- not 1687 * the current state flags. Then we mark 1688 * things so that the next operation to 1689 * this HBA will cause the update to occur. 1690 */ 1691 if (cts->valid & CCB_TRANS_DISC_VALID) { 1692 if ((cts->flags & CCB_TRANS_DISC_ENB) != 0) { 1693 *dptr |= DPARM_DISC; 1694 } else { 1695 *dptr &= ~DPARM_DISC; 1696 } 1697 } 1698 if (cts->valid & CCB_TRANS_TQ_VALID) { 1699 if ((cts->flags & CCB_TRANS_TAG_ENB) != 0) { 1700 *dptr |= DPARM_TQING; 1701 } else { 1702 *dptr &= ~DPARM_TQING; 1703 } 1704 } 1705 if (cts->valid & CCB_TRANS_BUS_WIDTH_VALID) { 1706 switch (cts->bus_width) { 1707 case MSG_EXT_WDTR_BUS_16_BIT: 1708 *dptr |= DPARM_WIDE; 1709 break; 1710 default: 1711 *dptr &= ~DPARM_WIDE; 1712 } 1713 } 1714 /* 1715 * Any SYNC RATE of nonzero and SYNC_OFFSET 1716 * of nonzero will cause us to go to the 1717 * selected (from NVRAM) maximum value for 1718 * this device. At a later point, we'll 1719 * allow finer control. 1720 */ 1721 if ((cts->valid & CCB_TRANS_SYNC_RATE_VALID) && 1722 (cts->valid & CCB_TRANS_SYNC_OFFSET_VALID) && 1723 (cts->sync_offset > 0)) { 1724 *dptr |= DPARM_SYNC; 1725 } else { 1726 *dptr &= ~DPARM_SYNC; 1727 } 1728 *dptr |= DPARM_SAFE_DFLT; 1729 isp_prt(isp, ISP_LOGDEBUG0, 1730 "%d.%d set %s period 0x%x offset 0x%x flags 0x%x", 1731 bus, tgt, (cts->flags & CCB_TRANS_CURRENT_SETTINGS)? 1732 "current" : "user", 1733 sdp->isp_devparam[tgt].sync_period, 1734 sdp->isp_devparam[tgt].sync_offset, 1735 sdp->isp_devparam[tgt].dev_flags); 1736 sdp->isp_devparam[tgt].dev_update = 1; 1737 isp->isp_update |= (1 << bus); 1738 } 1739 ISP_UNLOCK(isp); 1740 ccb->ccb_h.status = CAM_REQ_CMP; 1741 xpt_done(ccb); 1742 break; 1743 1744 case XPT_GET_TRAN_SETTINGS: 1745 1746 cts = &ccb->cts; 1747 tgt = cts->ccb_h.target_id; 1748 if (IS_FC(isp)) { 1749 /* 1750 * a lot of normal SCSI things don't make sense. 1751 */ 1752 cts->flags = CCB_TRANS_TAG_ENB | CCB_TRANS_DISC_ENB; 1753 cts->valid = CCB_TRANS_DISC_VALID | CCB_TRANS_TQ_VALID; 1754 /* 1755 * How do you measure the width of a high 1756 * speed serial bus? Well, in bytes. 1757 * 1758 * Offset and period make no sense, though, so we set 1759 * (above) a 'base' transfer speed to be gigabit. 1760 */ 1761 cts->bus_width = MSG_EXT_WDTR_BUS_8_BIT; 1762 } else { 1763 sdparam *sdp = isp->isp_param; 1764 u_int16_t dval, pval, oval; 1765 int bus = cam_sim_bus(xpt_path_sim(cts->ccb_h.path)); 1766 1767 sdp += bus; 1768 if (cts->flags & CCB_TRANS_CURRENT_SETTINGS) { 1769 ISP_LOCK(isp); 1770 sdp->isp_devparam[tgt].dev_refresh = 1; 1771 isp->isp_update |= (1 << bus); 1772 (void) isp_control(isp, ISPCTL_UPDATE_PARAMS, 1773 NULL); 1774 ISP_UNLOCK(isp); 1775 dval = sdp->isp_devparam[tgt].cur_dflags; 1776 oval = sdp->isp_devparam[tgt].cur_offset; 1777 pval = sdp->isp_devparam[tgt].cur_period; 1778 } else { 1779 dval = sdp->isp_devparam[tgt].dev_flags; 1780 oval = sdp->isp_devparam[tgt].sync_offset; 1781 pval = sdp->isp_devparam[tgt].sync_period; 1782 } 1783 1784 ISP_LOCK(isp); 1785 cts->flags &= ~(CCB_TRANS_DISC_ENB|CCB_TRANS_TAG_ENB); 1786 1787 if (dval & DPARM_DISC) { 1788 cts->flags |= CCB_TRANS_DISC_ENB; 1789 } 1790 if (dval & DPARM_TQING) { 1791 cts->flags |= CCB_TRANS_TAG_ENB; 1792 } 1793 if (dval & DPARM_WIDE) { 1794 cts->bus_width = MSG_EXT_WDTR_BUS_16_BIT; 1795 } else { 1796 cts->bus_width = MSG_EXT_WDTR_BUS_8_BIT; 1797 } 1798 cts->valid = CCB_TRANS_BUS_WIDTH_VALID | 1799 CCB_TRANS_DISC_VALID | CCB_TRANS_TQ_VALID; 1800 1801 if ((dval & DPARM_SYNC) && oval != 0) { 1802 cts->sync_period = pval; 1803 cts->sync_offset = oval; 1804 cts->valid |= 1805 CCB_TRANS_SYNC_RATE_VALID | 1806 CCB_TRANS_SYNC_OFFSET_VALID; 1807 } 1808 ISP_UNLOCK(isp); 1809 isp_prt(isp, ISP_LOGDEBUG0, 1810 "%d.%d get %s period 0x%x offset 0x%x flags 0x%x", 1811 bus, tgt, (cts->flags & CCB_TRANS_CURRENT_SETTINGS)? 1812 "current" : "user", pval, oval, dval); 1813 } 1814 ccb->ccb_h.status = CAM_REQ_CMP; 1815 xpt_done(ccb); 1816 break; 1817 1818 case XPT_CALC_GEOMETRY: 1819 { 1820 struct ccb_calc_geometry *ccg; 1821 u_int32_t secs_per_cylinder; 1822 u_int32_t size_mb; 1823 1824 ccg = &ccb->ccg; 1825 if (ccg->block_size == 0) { 1826 isp_prt(isp, ISP_LOGERR, 1827 "%d.%d XPT_CALC_GEOMETRY block size 0?", 1828 ccg->ccb_h.target_id, ccg->ccb_h.target_lun); 1829 ccb->ccb_h.status = CAM_REQ_INVALID; 1830 xpt_done(ccb); 1831 break; 1832 } 1833 size_mb = ccg->volume_size /((1024L * 1024L) / ccg->block_size); 1834 if (size_mb > 1024) { 1835 ccg->heads = 255; 1836 ccg->secs_per_track = 63; 1837 } else { 1838 ccg->heads = 64; 1839 ccg->secs_per_track = 32; 1840 } 1841 secs_per_cylinder = ccg->heads * ccg->secs_per_track; 1842 ccg->cylinders = ccg->volume_size / secs_per_cylinder; 1843 ccb->ccb_h.status = CAM_REQ_CMP; 1844 xpt_done(ccb); 1845 break; 1846 } 1847 case XPT_RESET_BUS: /* Reset the specified bus */ 1848 bus = cam_sim_bus(sim); 1849 ISP_LOCK(isp); 1850 error = isp_control(isp, ISPCTL_RESET_BUS, &bus); 1851 ISP_UNLOCK(isp); 1852 if (error) 1853 ccb->ccb_h.status = CAM_REQ_CMP_ERR; 1854 else { 1855 if (cam_sim_bus(sim) && isp->isp_path2 != NULL) 1856 xpt_async(AC_BUS_RESET, isp->isp_path2, NULL); 1857 else if (isp->isp_path != NULL) 1858 xpt_async(AC_BUS_RESET, isp->isp_path, NULL); 1859 ccb->ccb_h.status = CAM_REQ_CMP; 1860 } 1861 xpt_done(ccb); 1862 break; 1863 1864 case XPT_TERM_IO: /* Terminate the I/O process */ 1865 ccb->ccb_h.status = CAM_REQ_INVALID; 1866 xpt_done(ccb); 1867 break; 1868 1869 case XPT_PATH_INQ: /* Path routing inquiry */ 1870 { 1871 struct ccb_pathinq *cpi = &ccb->cpi; 1872 1873 cpi->version_num = 1; 1874#ifdef ISP_TARGET_MODE 1875 cpi->target_sprt = PIT_PROCESSOR | PIT_DISCONNECT | PIT_TERM_IO; 1876#else 1877 cpi->target_sprt = 0; 1878#endif 1879 cpi->hba_eng_cnt = 0; 1880 cpi->max_target = ISP_MAX_TARGETS(isp) - 1; 1881 cpi->max_lun = ISP_MAX_LUNS(isp) - 1; 1882 cpi->bus_id = cam_sim_bus(sim); 1883 if (IS_FC(isp)) { 1884 cpi->hba_misc = PIM_NOBUSRESET; 1885 /* 1886 * Because our loop ID can shift from time to time, 1887 * make our initiator ID out of range of our bus. 1888 */ 1889 cpi->initiator_id = cpi->max_target + 1; 1890 1891 /* 1892 * Set base transfer capabilities for Fibre Channel. 1893 * Technically not correct because we don't know 1894 * what media we're running on top of- but we'll 1895 * look good if we always say 100MB/s. 1896 */ 1897 cpi->base_transfer_speed = 100000; 1898 cpi->hba_inquiry = PI_TAG_ABLE; 1899 } else { 1900 sdparam *sdp = isp->isp_param; 1901 sdp += cam_sim_bus(xpt_path_sim(cpi->ccb_h.path)); 1902 cpi->hba_inquiry = PI_SDTR_ABLE|PI_TAG_ABLE|PI_WIDE_16; 1903 cpi->hba_misc = 0; 1904 cpi->initiator_id = sdp->isp_initiator_id; 1905 cpi->base_transfer_speed = 3300; 1906 } 1907 strncpy(cpi->sim_vid, "FreeBSD", SIM_IDLEN); 1908 strncpy(cpi->hba_vid, "Qlogic", HBA_IDLEN); 1909 strncpy(cpi->dev_name, cam_sim_name(sim), DEV_IDLEN); 1910 cpi->unit_number = cam_sim_unit(sim); 1911 cpi->ccb_h.status = CAM_REQ_CMP; 1912 xpt_done(ccb); 1913 break; 1914 } 1915 default: 1916 ccb->ccb_h.status = CAM_REQ_INVALID; 1917 xpt_done(ccb); 1918 break; 1919 } 1920} 1921 1922#define ISPDDB (CAM_DEBUG_INFO|CAM_DEBUG_TRACE|CAM_DEBUG_CDB) 1923void 1924isp_done(struct ccb_scsiio *sccb) 1925{ 1926 struct ispsoftc *isp = XS_ISP(sccb); 1927 1928 if (XS_NOERR(sccb)) 1929 XS_SETERR(sccb, CAM_REQ_CMP); 1930 1931 if ((sccb->ccb_h.status & CAM_STATUS_MASK) == CAM_REQ_CMP && 1932 (sccb->scsi_status != SCSI_STATUS_OK)) { 1933 sccb->ccb_h.status &= ~CAM_STATUS_MASK; 1934 if ((sccb->scsi_status == SCSI_STATUS_CHECK_COND) && 1935 (sccb->ccb_h.status & CAM_AUTOSNS_VALID) == 0) { 1936 sccb->ccb_h.status |= CAM_AUTOSENSE_FAIL; 1937 } else { 1938 sccb->ccb_h.status |= CAM_SCSI_STATUS_ERROR; 1939 } 1940 } 1941 1942 sccb->ccb_h.status &= ~CAM_SIM_QUEUED; 1943 if ((sccb->ccb_h.status & CAM_STATUS_MASK) != CAM_REQ_CMP) { 1944 if ((sccb->ccb_h.status & CAM_DEV_QFRZN) == 0) { 1945 sccb->ccb_h.status |= CAM_DEV_QFRZN; 1946 xpt_freeze_devq(sccb->ccb_h.path, 1); 1947 if (sccb->scsi_status != SCSI_STATUS_OK) 1948 isp_prt(isp, ISP_LOGDEBUG2, 1949 "freeze devq %d.%d %x %x", 1950 sccb->ccb_h.target_id, 1951 sccb->ccb_h.target_lun, sccb->ccb_h.status, 1952 sccb->scsi_status); 1953 } 1954 } 1955 1956 /* 1957 * If we were frozen waiting resources, clear that we were frozen 1958 * waiting for resources. If we are no longer frozen, and the devq 1959 * isn't frozen, mark the completing CCB to have the XPT layer 1960 * release the simq. 1961 */ 1962 if (isp->isp_osinfo.simqfrozen & SIMQFRZ_RESOURCE) { 1963 isp->isp_osinfo.simqfrozen &= ~SIMQFRZ_RESOURCE; 1964 if (isp->isp_osinfo.simqfrozen == 0) { 1965 if ((sccb->ccb_h.status & CAM_DEV_QFRZN) == 0) { 1966 isp_prt(isp, ISP_LOGDEBUG2, 1967 "isp_done->relsimq"); 1968 sccb->ccb_h.status |= CAM_RELEASE_SIMQ; 1969 } else { 1970 isp_prt(isp, ISP_LOGDEBUG2, 1971 "isp_done->devq frozen"); 1972 } 1973 } else { 1974 isp_prt(isp, ISP_LOGDEBUG2, 1975 "isp_done -> simqfrozen = %x", 1976 isp->isp_osinfo.simqfrozen); 1977 } 1978 } 1979 if ((CAM_DEBUGGED(sccb->ccb_h.path, ISPDDB)) && 1980 (sccb->ccb_h.status & CAM_STATUS_MASK) != CAM_REQ_CMP) { 1981 xpt_print_path(sccb->ccb_h.path); 1982 printf("cam completion status 0x%x\n", sccb->ccb_h.status); 1983 } 1984 1985 XS_CMD_S_DONE(sccb); 1986 if (XS_CMD_WDOG_P(sccb) == 0) { 1987 untimeout(isp_watchdog, (caddr_t)sccb, sccb->ccb_h.timeout_ch); 1988 if (XS_CMD_GRACE_P(sccb)) { 1989 isp_prt(isp, ISP_LOGDEBUG2, 1990 "finished command on borrowed time"); 1991 } 1992 XS_CMD_S_CLEAR(sccb); 1993 ISP_UNLOCK(isp); 1994#ifdef ISP_SMPLOCK
|