icp.c revision 1.13
1/* $NetBSD: icp.c,v 1.13 2004/09/13 12:55:47 drochner Exp $ */ 2 3/*- 4 * Copyright (c) 2002, 2003 The NetBSD Foundation, Inc. 5 * All rights reserved. 6 * 7 * This code is derived from software contributed to The NetBSD Foundation 8 * by Andrew Doran, and by Jason R. Thorpe of Wasabi Systems, Inc. 9 * 10 * Redistribution and use in source and binary forms, with or without 11 * modification, are permitted provided that the following conditions 12 * are met: 13 * 1. Redistributions of source code must retain the above copyright 14 * notice, this list of conditions and the following disclaimer. 15 * 2. Redistributions in binary form must reproduce the above copyright 16 * notice, this list of conditions and the following disclaimer in the 17 * documentation and/or other materials provided with the distribution. 18 * 3. All advertising materials mentioning features or use of this software 19 * must display the following acknowledgement: 20 * This product includes software developed by the NetBSD 21 * Foundation, Inc. and its contributors. 22 * 4. Neither the name of The NetBSD Foundation nor the names of its 23 * contributors may be used to endorse or promote products derived 24 * from this software without specific prior written permission. 25 * 26 * THIS SOFTWARE IS PROVIDED BY THE NETBSD FOUNDATION, INC. AND CONTRIBUTORS 27 * ``AS IS'' AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED 28 * TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR 29 * PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE FOUNDATION OR CONTRIBUTORS 30 * BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR 31 * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF 32 * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS 33 * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN 34 * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) 35 * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE 36 * POSSIBILITY OF SUCH DAMAGE. 37 */ 38 39/* 40 * Copyright (c) 1999, 2000 Niklas Hallqvist. All rights reserved. 41 * 42 * Redistribution and use in source and binary forms, with or without 43 * modification, are permitted provided that the following conditions 44 * are met: 45 * 1. Redistributions of source code must retain the above copyright 46 * notice, this list of conditions and the following disclaimer. 47 * 2. Redistributions in binary form must reproduce the above copyright 48 * notice, this list of conditions and the following disclaimer in the 49 * documentation and/or other materials provided with the distribution. 50 * 3. All advertising materials mentioning features or use of this software 51 * must display the following acknowledgement: 52 * This product includes software developed by Niklas Hallqvist. 53 * 4. The name of the author may not be used to endorse or promote products 54 * derived from this software without specific prior written permission. 55 * 56 * THIS SOFTWARE IS PROVIDED BY THE AUTHOR ``AS IS'' AND ANY EXPRESS OR 57 * IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES 58 * OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED. 59 * IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR ANY DIRECT, INDIRECT, 60 * INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT 61 * NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, 62 * DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY 63 * THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT 64 * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF 65 * THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. 66 * 67 * from OpenBSD: gdt_common.c,v 1.12 2001/07/04 06:43:18 niklas Exp 68 */ 69 70/* 71 * This driver would not have written if it was not for the hardware donations 72 * from both ICP-Vortex and �ko.neT. I want to thank them for their support. 73 * 74 * Re-worked for NetBSD by Andrew Doran. Test hardware kindly supplied by 75 * Intel. 76 * 77 * Support for the ICP-Vortex management tools added by 78 * Jason R. Thorpe of Wasabi Systems, Inc., based on code 79 * provided by Achim Leubner <achim.leubner@intel.com>. 80 * 81 * Additional support for dynamic rescan of cacheservice drives by 82 * Jason R. Thorpe of Wasabi Systems, Inc. 83 */ 84 85#include <sys/cdefs.h> 86__KERNEL_RCSID(0, "$NetBSD: icp.c,v 1.13 2004/09/13 12:55:47 drochner Exp $"); 87 88#include <sys/param.h> 89#include <sys/systm.h> 90#include <sys/kernel.h> 91#include <sys/device.h> 92#include <sys/queue.h> 93#include <sys/proc.h> 94#include <sys/buf.h> 95#include <sys/endian.h> 96#include <sys/malloc.h> 97#include <sys/disk.h> 98 99#include <uvm/uvm_extern.h> 100 101#include <machine/bswap.h> 102#include <machine/bus.h> 103 104#include <dev/pci/pcireg.h> 105#include <dev/pci/pcivar.h> 106#include <dev/pci/pcidevs.h> 107 108#include <dev/ic/icpreg.h> 109#include <dev/ic/icpvar.h> 110 111#include <dev/scsipi/scsipi_all.h> 112#include <dev/scsipi/scsiconf.h> 113 114#include "locators.h" 115 116int icp_async_event(struct icp_softc *, int); 117void icp_ccb_submit(struct icp_softc *icp, struct icp_ccb *ic); 118void icp_chain(struct icp_softc *); 119int icp_print(void *, const char *); 120int icp_submatch(struct device *, struct cfdata *, 121 const locdesc_t *, void *); 122void icp_watchdog(void *); 123void icp_ucmd_intr(struct icp_ccb *); 124void icp_recompute_openings(struct icp_softc *); 125 126int icp_count; /* total # of controllers, for ioctl interface */ 127 128/* 129 * Statistics for the ioctl interface to query. 130 * 131 * XXX Global. They should probably be made per-controller 132 * XXX at some point. 133 */ 134gdt_statist_t icp_stats; 135 136int 137icp_init(struct icp_softc *icp, const char *intrstr) 138{ 139 struct icp_attach_args icpa; 140 struct icp_binfo binfo; 141 struct icp_ccb *ic; 142 u_int16_t cdev_cnt; 143 int i, j, state, feat, nsegs, rv; 144 int help[2]; 145 locdesc_t *ldesc = (void *)help; /* XXX */ 146 147 state = 0; 148 149 if (intrstr != NULL) 150 aprint_normal("%s: interrupting at %s\n", icp->icp_dv.dv_xname, 151 intrstr); 152 153 SIMPLEQ_INIT(&icp->icp_ccb_queue); 154 SIMPLEQ_INIT(&icp->icp_ccb_freelist); 155 SIMPLEQ_INIT(&icp->icp_ucmd_queue); 156 callout_init(&icp->icp_wdog_callout); 157 158 /* 159 * Allocate a scratch area. 160 */ 161 if (bus_dmamap_create(icp->icp_dmat, ICP_SCRATCH_SIZE, 1, 162 ICP_SCRATCH_SIZE, 0, BUS_DMA_NOWAIT | BUS_DMA_ALLOCNOW, 163 &icp->icp_scr_dmamap) != 0) { 164 aprint_error("%s: cannot create scratch dmamap\n", 165 icp->icp_dv.dv_xname); 166 return (1); 167 } 168 state++; 169 170 if (bus_dmamem_alloc(icp->icp_dmat, ICP_SCRATCH_SIZE, PAGE_SIZE, 0, 171 icp->icp_scr_seg, 1, &nsegs, BUS_DMA_NOWAIT) != 0) { 172 aprint_error("%s: cannot alloc scratch dmamem\n", 173 icp->icp_dv.dv_xname); 174 goto bail_out; 175 } 176 state++; 177 178 if (bus_dmamem_map(icp->icp_dmat, icp->icp_scr_seg, nsegs, 179 ICP_SCRATCH_SIZE, &icp->icp_scr, 0)) { 180 aprint_error("%s: cannot map scratch dmamem\n", 181 icp->icp_dv.dv_xname); 182 goto bail_out; 183 } 184 state++; 185 186 if (bus_dmamap_load(icp->icp_dmat, icp->icp_scr_dmamap, icp->icp_scr, 187 ICP_SCRATCH_SIZE, NULL, BUS_DMA_NOWAIT)) { 188 aprint_error("%s: cannot load scratch dmamap\n", 189 icp->icp_dv.dv_xname); 190 goto bail_out; 191 } 192 state++; 193 194 /* 195 * Allocate and initialize the command control blocks. 196 */ 197 ic = malloc(sizeof(*ic) * ICP_NCCBS, M_DEVBUF, M_NOWAIT | M_ZERO); 198 if ((icp->icp_ccbs = ic) == NULL) { 199 aprint_error("%s: malloc() failed\n", icp->icp_dv.dv_xname); 200 goto bail_out; 201 } 202 state++; 203 204 for (i = 0; i < ICP_NCCBS; i++, ic++) { 205 /* 206 * The first two command indexes have special meanings, so 207 * we can't use them. 208 */ 209 ic->ic_ident = i + 2; 210 rv = bus_dmamap_create(icp->icp_dmat, ICP_MAX_XFER, 211 ICP_MAXSG, ICP_MAX_XFER, 0, 212 BUS_DMA_NOWAIT | BUS_DMA_ALLOCNOW, 213 &ic->ic_xfer_map); 214 if (rv != 0) 215 break; 216 icp->icp_nccbs++; 217 icp_ccb_free(icp, ic); 218 } 219#ifdef DIAGNOSTIC 220 if (icp->icp_nccbs != ICP_NCCBS) 221 aprint_error("%s: %d/%d CCBs usable\n", icp->icp_dv.dv_xname, 222 icp->icp_nccbs, ICP_NCCBS); 223#endif 224 225 /* 226 * Initalize the controller. 227 */ 228 if (!icp_cmd(icp, ICP_SCREENSERVICE, ICP_INIT, 0, 0, 0)) { 229 aprint_error("%s: screen service init error %d\n", 230 icp->icp_dv.dv_xname, icp->icp_status); 231 goto bail_out; 232 } 233 234 if (!icp_cmd(icp, ICP_CACHESERVICE, ICP_INIT, ICP_LINUX_OS, 0, 0)) { 235 aprint_error("%s: cache service init error %d\n", 236 icp->icp_dv.dv_xname, icp->icp_status); 237 goto bail_out; 238 } 239 240 icp_cmd(icp, ICP_CACHESERVICE, ICP_UNFREEZE_IO, 0, 0, 0); 241 242 if (!icp_cmd(icp, ICP_CACHESERVICE, ICP_MOUNT, 0xffff, 1, 0)) { 243 aprint_error("%s: cache service mount error %d\n", 244 icp->icp_dv.dv_xname, icp->icp_status); 245 goto bail_out; 246 } 247 248 if (!icp_cmd(icp, ICP_CACHESERVICE, ICP_INIT, ICP_LINUX_OS, 0, 0)) { 249 aprint_error("%s: cache service post-mount init error %d\n", 250 icp->icp_dv.dv_xname, icp->icp_status); 251 goto bail_out; 252 } 253 cdev_cnt = (u_int16_t)icp->icp_info; 254 icp->icp_fw_vers = icp->icp_service; 255 256 if (!icp_cmd(icp, ICP_SCSIRAWSERVICE, ICP_INIT, 0, 0, 0)) { 257 aprint_error("%s: raw service init error %d\n", 258 icp->icp_dv.dv_xname, icp->icp_status); 259 goto bail_out; 260 } 261 262 /* 263 * Set/get raw service features (scatter/gather). 264 */ 265 feat = 0; 266 if (icp_cmd(icp, ICP_SCSIRAWSERVICE, ICP_SET_FEAT, ICP_SCATTER_GATHER, 267 0, 0)) 268 if (icp_cmd(icp, ICP_SCSIRAWSERVICE, ICP_GET_FEAT, 0, 0, 0)) 269 feat = icp->icp_info; 270 271 if ((feat & ICP_SCATTER_GATHER) == 0) { 272#ifdef DIAGNOSTIC 273 aprint_normal( 274 "%s: scatter/gather not supported (raw service)\n", 275 icp->icp_dv.dv_xname); 276#endif 277 } else 278 icp->icp_features |= ICP_FEAT_RAWSERVICE; 279 280 /* 281 * Set/get cache service features (scatter/gather). 282 */ 283 feat = 0; 284 if (icp_cmd(icp, ICP_CACHESERVICE, ICP_SET_FEAT, 0, 285 ICP_SCATTER_GATHER, 0)) 286 if (icp_cmd(icp, ICP_CACHESERVICE, ICP_GET_FEAT, 0, 0, 0)) 287 feat = icp->icp_info; 288 289 if ((feat & ICP_SCATTER_GATHER) == 0) { 290#ifdef DIAGNOSTIC 291 aprint_normal( 292 "%s: scatter/gather not supported (cache service)\n", 293 icp->icp_dv.dv_xname); 294#endif 295 } else 296 icp->icp_features |= ICP_FEAT_CACHESERVICE; 297 298 /* 299 * Pull some information from the board and dump. 300 */ 301 if (!icp_cmd(icp, ICP_CACHESERVICE, ICP_IOCTL, ICP_BOARD_INFO, 302 ICP_INVALID_CHANNEL, sizeof(struct icp_binfo))) { 303 aprint_error("%s: unable to retrive board info\n", 304 icp->icp_dv.dv_xname); 305 goto bail_out; 306 } 307 memcpy(&binfo, icp->icp_scr, sizeof(binfo)); 308 309 aprint_normal( 310 "%s: model <%s>, firmware <%s>, %d channel(s), %dMB memory\n", 311 icp->icp_dv.dv_xname, binfo.bi_type_string, binfo.bi_raid_string, 312 binfo.bi_chan_count, le32toh(binfo.bi_memsize) >> 20); 313 314 /* 315 * Determine the number of devices, and number of openings per 316 * device. 317 */ 318 if (icp->icp_features & ICP_FEAT_CACHESERVICE) { 319 for (j = 0; j < cdev_cnt && j < ICP_MAX_HDRIVES; j++) { 320 if (!icp_cmd(icp, ICP_CACHESERVICE, ICP_INFO, j, 0, 321 0)) 322 continue; 323 324 icp->icp_cdr[j].cd_size = icp->icp_info; 325 if (icp->icp_cdr[j].cd_size != 0) 326 icp->icp_ndevs++; 327 328 if (icp_cmd(icp, ICP_CACHESERVICE, ICP_DEVTYPE, j, 0, 329 0)) 330 icp->icp_cdr[j].cd_type = icp->icp_info; 331 } 332 } 333 334 if (icp->icp_features & ICP_FEAT_RAWSERVICE) { 335 icp->icp_nchan = binfo.bi_chan_count; 336 icp->icp_ndevs += icp->icp_nchan; 337 } 338 339 icp_recompute_openings(icp); 340 341 /* 342 * Attach SCSI channels. 343 */ 344 if (icp->icp_features & ICP_FEAT_RAWSERVICE) { 345 struct icp_ioc_version *iv; 346 struct icp_rawioc *ri; 347 struct icp_getch *gc; 348 349 iv = (struct icp_ioc_version *)icp->icp_scr; 350 iv->iv_version = htole32(ICP_IOC_NEWEST); 351 iv->iv_listents = ICP_MAXBUS; 352 iv->iv_firstchan = 0; 353 iv->iv_lastchan = ICP_MAXBUS - 1; 354 iv->iv_listoffset = htole32(sizeof(*iv)); 355 356 if (icp_cmd(icp, ICP_CACHESERVICE, ICP_IOCTL, 357 ICP_IOCHAN_RAW_DESC, ICP_INVALID_CHANNEL, 358 sizeof(*iv) + ICP_MAXBUS * sizeof(*ri))) { 359 ri = (struct icp_rawioc *)(iv + 1); 360 for (j = 0; j < binfo.bi_chan_count; j++, ri++) 361 icp->icp_bus_id[j] = ri->ri_procid; 362 } else { 363 /* 364 * Fall back to the old method. 365 */ 366 gc = (struct icp_getch *)icp->icp_scr; 367 368 for (j = 0; j < binfo.bi_chan_count; j++) { 369 if (!icp_cmd(icp, ICP_CACHESERVICE, ICP_IOCTL, 370 ICP_SCSI_CHAN_CNT | ICP_L_CTRL_PATTERN, 371 ICP_IO_CHANNEL | ICP_INVALID_CHANNEL, 372 sizeof(*gc))) { 373 aprint_error( 374 "%s: unable to get chan info", 375 icp->icp_dv.dv_xname); 376 goto bail_out; 377 } 378 icp->icp_bus_id[j] = gc->gc_scsiid; 379 } 380 } 381 382 for (j = 0; j < binfo.bi_chan_count; j++) { 383 if (icp->icp_bus_id[j] > ICP_MAXID_FC) 384 icp->icp_bus_id[j] = ICP_MAXID_FC; 385 386 icpa.icpa_unit = j + ICPA_UNIT_SCSI; 387 388 ldesc->len = 1; 389 ldesc->locs[ICPCF_UNIT] = j + ICPA_UNIT_SCSI; 390 391 icp->icp_children[icpa.icpa_unit] = 392 config_found_sm_loc(&icp->icp_dv, "icp", ldesc, 393 &icpa, icp_print, icp_submatch); 394 } 395 } 396 397 /* 398 * Attach cache devices. 399 */ 400 if (icp->icp_features & ICP_FEAT_CACHESERVICE) { 401 for (j = 0; j < cdev_cnt && j < ICP_MAX_HDRIVES; j++) { 402 if (icp->icp_cdr[j].cd_size == 0) 403 continue; 404 405 icpa.icpa_unit = j; 406 407 ldesc->len = 1; 408 ldesc->locs[ICPCF_UNIT] = j; 409 410 icp->icp_children[icpa.icpa_unit] = 411 config_found_sm_loc(&icp->icp_dv, "icp", ldesc, 412 &icpa, icp_print, icp_submatch); 413 } 414 } 415 416 /* 417 * Start the watchdog. 418 */ 419 icp_watchdog(icp); 420 421 /* 422 * Count the controller, and we're done! 423 */ 424 icp_count++; 425 426 return (0); 427 428 bail_out: 429 if (state > 4) 430 for (j = 0; j < i; j++) 431 bus_dmamap_destroy(icp->icp_dmat, 432 icp->icp_ccbs[j].ic_xfer_map); 433 if (state > 3) 434 free(icp->icp_ccbs, M_DEVBUF); 435 if (state > 2) 436 bus_dmamap_unload(icp->icp_dmat, icp->icp_scr_dmamap); 437 if (state > 1) 438 bus_dmamem_unmap(icp->icp_dmat, icp->icp_scr, 439 ICP_SCRATCH_SIZE); 440 if (state > 0) 441 bus_dmamem_free(icp->icp_dmat, icp->icp_scr_seg, nsegs); 442 bus_dmamap_destroy(icp->icp_dmat, icp->icp_scr_dmamap); 443 444 return (1); 445} 446 447void 448icp_register_servicecb(struct icp_softc *icp, int unit, 449 const struct icp_servicecb *cb) 450{ 451 452 icp->icp_servicecb[unit] = cb; 453} 454 455void 456icp_rescan(struct icp_softc *icp, int unit) 457{ 458 struct icp_attach_args icpa; 459 u_int newsize, newtype; 460 int help[2]; 461 locdesc_t *ldesc = (void *)help; /* XXX */ 462 463 /* 464 * NOTE: It is very important that the queue be frozen and not 465 * commands running when this is called. The ioctl mutex must 466 * also be held. 467 */ 468 469 KASSERT(icp->icp_qfreeze != 0); 470 KASSERT(icp->icp_running == 0); 471 KASSERT(unit < ICP_MAX_HDRIVES); 472 473 if (!icp_cmd(icp, ICP_CACHESERVICE, ICP_INFO, unit, 0, 0)) { 474#ifdef ICP_DEBUG 475 printf("%s: rescan: unit %d ICP_INFO failed -> 0x%04x\n", 476 icp->icp_dv.dv_xname, unit, icp->icp_status); 477#endif 478 goto gone; 479 } 480 if ((newsize = icp->icp_info) == 0) { 481#ifdef ICP_DEBUG 482 printf("%s: rescan: unit %d has zero size\n", 483 icp->icp_dv.dv_xname, unit); 484#endif 485 gone: 486 /* 487 * Host drive is no longer present; detach if a child 488 * is currently there. 489 */ 490 if (icp->icp_cdr[unit].cd_size != 0) 491 icp->icp_ndevs--; 492 icp->icp_cdr[unit].cd_size = 0; 493 if (icp->icp_children[unit] != NULL) { 494 (void) config_detach(icp->icp_children[unit], 495 DETACH_FORCE); 496 icp->icp_children[unit] = NULL; 497 } 498 return; 499 } 500 501 if (icp_cmd(icp, ICP_CACHESERVICE, ICP_DEVTYPE, unit, 0, 0)) 502 newtype = icp->icp_info; 503 else { 504#ifdef ICP_DEBUG 505 printf("%s: rescan: unit %d ICP_DEVTYPE failed\n", 506 icp->icp_dv.dv_xname, unit); 507#endif 508 newtype = 0; /* XXX? */ 509 } 510 511#ifdef ICP_DEBUG 512 printf("%s: rescan: unit %d old %u/%u, new %u/%u\n", 513 icp->icp_dv.dv_xname, unit, icp->icp_cdr[unit].cd_size, 514 icp->icp_cdr[unit].cd_type, newsize, newtype); 515#endif 516 517 /* 518 * If the type or size changed, detach any old child (if it exists) 519 * and attach a new one. 520 */ 521 if (icp->icp_children[unit] == NULL || 522 newsize != icp->icp_cdr[unit].cd_size || 523 newtype != icp->icp_cdr[unit].cd_type) { 524 if (icp->icp_cdr[unit].cd_size == 0) 525 icp->icp_ndevs++; 526 icp->icp_cdr[unit].cd_size = newsize; 527 icp->icp_cdr[unit].cd_type = newtype; 528 if (icp->icp_children[unit] != NULL) 529 (void) config_detach(icp->icp_children[unit], 530 DETACH_FORCE); 531 532 icpa.icpa_unit = unit; 533 534 ldesc->len = 1; 535 ldesc->locs[ICPCF_UNIT] = unit; 536 537 icp->icp_children[unit] = config_found_sm_loc(&icp->icp_dv, 538 "icp", ldesc, &icpa, icp_print, icp_submatch); 539 } 540 541 icp_recompute_openings(icp); 542} 543 544void 545icp_rescan_all(struct icp_softc *icp) 546{ 547 int unit; 548 u_int16_t cdev_cnt; 549 550 /* 551 * This is the old method of rescanning the host drives. We 552 * start by reinitializing the cache service. 553 */ 554 if (!icp_cmd(icp, ICP_CACHESERVICE, ICP_INIT, ICP_LINUX_OS, 0, 0)) { 555 printf("%s: unable to re-initialize cache service for rescan\n", 556 icp->icp_dv.dv_xname); 557 return; 558 } 559 cdev_cnt = (u_int16_t) icp->icp_info; 560 561 /* For each host drive, do the new-style rescan. */ 562 for (unit = 0; unit < cdev_cnt && unit < ICP_MAX_HDRIVES; unit++) 563 icp_rescan(icp, unit); 564 565 /* Now detach anything in the slots after cdev_cnt. */ 566 for (; unit < ICP_MAX_HDRIVES; unit++) { 567 if (icp->icp_cdr[unit].cd_size != 0) { 568#ifdef ICP_DEBUG 569 printf("%s: rescan all: unit %d < new cdev_cnt (%d)\n", 570 icp->icp_dv.dv_xname, unit, cdev_cnt); 571#endif 572 icp->icp_ndevs--; 573 icp->icp_cdr[unit].cd_size = 0; 574 if (icp->icp_children[unit] != NULL) { 575 (void) config_detach(icp->icp_children[unit], 576 DETACH_FORCE); 577 icp->icp_children[unit] = NULL; 578 } 579 } 580 } 581 582 icp_recompute_openings(icp); 583} 584 585void 586icp_recompute_openings(struct icp_softc *icp) 587{ 588 int unit, openings; 589 590 if (icp->icp_ndevs != 0) 591 openings = 592 (icp->icp_nccbs - ICP_NCCB_RESERVE) / icp->icp_ndevs; 593 else 594 openings = 0; 595 if (openings == icp->icp_openings) 596 return; 597 icp->icp_openings = openings; 598 599#ifdef ICP_DEBUG 600 printf("%s: %d device%s, %d openings per device\n", 601 icp->icp_dv.dv_xname, icp->icp_ndevs, 602 icp->icp_ndevs == 1 ? "" : "s", icp->icp_openings); 603#endif 604 605 for (unit = 0; unit < ICP_MAX_HDRIVES + ICP_MAXBUS; unit++) { 606 if (icp->icp_children[unit] != NULL) 607 (*icp->icp_servicecb[unit]->iscb_openings)( 608 icp->icp_children[unit], icp->icp_openings); 609 } 610} 611 612void 613icp_watchdog(void *cookie) 614{ 615 struct icp_softc *icp; 616 int s; 617 618 icp = cookie; 619 620 s = splbio(); 621 icp_intr(icp); 622 if (ICP_HAS_WORK(icp)) 623 icp_ccb_enqueue(icp, NULL); 624 splx(s); 625 626 callout_reset(&icp->icp_wdog_callout, hz * ICP_WATCHDOG_FREQ, 627 icp_watchdog, icp); 628} 629 630int 631icp_print(void *aux, const char *pnp) 632{ 633 struct icp_attach_args *icpa; 634 const char *str; 635 636 icpa = (struct icp_attach_args *)aux; 637 638 if (pnp != NULL) { 639 if (icpa->icpa_unit < ICPA_UNIT_SCSI) 640 str = "block device"; 641 else 642 str = "SCSI channel"; 643 aprint_normal("%s at %s", str, pnp); 644 } 645 aprint_normal(" unit %d", icpa->icpa_unit); 646 647 return (UNCONF); 648} 649 650int 651icp_submatch(struct device *parent, struct cfdata *cf, 652 const locdesc_t *ldesc, void *aux) 653{ 654 655 if (cf->cf_loc[ICPCF_UNIT] != ICPCF_UNIT_DEFAULT && 656 cf->cf_loc[ICPCF_UNIT] != ldesc->locs[ICPCF_UNIT]) 657 return (0); 658 659 return (config_match(parent, cf, aux)); 660} 661 662int 663icp_async_event(struct icp_softc *icp, int service) 664{ 665 666 if (service == ICP_SCREENSERVICE) { 667 if (icp->icp_status == ICP_S_MSG_REQUEST) { 668 /* XXX */ 669 } 670 } else { 671 if ((icp->icp_fw_vers & 0xff) >= 0x1a) { 672 icp->icp_evt.size = 0; 673 icp->icp_evt.eu.async.ionode = icp->icp_dv.dv_unit; 674 icp->icp_evt.eu.async.status = icp->icp_status; 675 /* 676 * Severity and event string are filled in by the 677 * hardware interface interrupt handler. 678 */ 679 printf("%s: %s\n", icp->icp_dv.dv_xname, 680 icp->icp_evt.event_string); 681 } else { 682 icp->icp_evt.size = sizeof(icp->icp_evt.eu.async); 683 icp->icp_evt.eu.async.ionode = icp->icp_dv.dv_unit; 684 icp->icp_evt.eu.async.service = service; 685 icp->icp_evt.eu.async.status = icp->icp_status; 686 icp->icp_evt.eu.async.info = icp->icp_info; 687 /* XXXJRT FIX THIS */ 688 *(u_int32_t *) icp->icp_evt.eu.async.scsi_coord = 689 icp->icp_info2; 690 } 691 icp_store_event(icp, GDT_ES_ASYNC, service, &icp->icp_evt); 692 } 693 694 return (0); 695} 696 697int 698icp_intr(void *cookie) 699{ 700 struct icp_softc *icp; 701 struct icp_intr_ctx ctx; 702 struct icp_ccb *ic; 703 704 icp = cookie; 705 706 ctx.istatus = (*icp->icp_get_status)(icp); 707 if (!ctx.istatus) { 708 icp->icp_status = ICP_S_NO_STATUS; 709 return (0); 710 } 711 712 (*icp->icp_intr)(icp, &ctx); 713 714 icp->icp_status = ctx.cmd_status; 715 icp->icp_service = ctx.service; 716 icp->icp_info = ctx.info; 717 icp->icp_info2 = ctx.info2; 718 719 switch (ctx.istatus) { 720 case ICP_ASYNCINDEX: 721 icp_async_event(icp, ctx.service); 722 return (1); 723 724 case ICP_SPEZINDEX: 725 printf("%s: uninitialized or unknown service (%d/%d)\n", 726 icp->icp_dv.dv_xname, ctx.info, ctx.info2); 727 icp->icp_evt.size = sizeof(icp->icp_evt.eu.driver); 728 icp->icp_evt.eu.driver.ionode = icp->icp_dv.dv_unit; 729 icp_store_event(icp, GDT_ES_DRIVER, 4, &icp->icp_evt); 730 return (1); 731 } 732 733 if ((ctx.istatus - 2) > icp->icp_nccbs) 734 panic("icp_intr: bad command index returned"); 735 736 ic = &icp->icp_ccbs[ctx.istatus - 2]; 737 ic->ic_status = icp->icp_status; 738 739 if ((ic->ic_flags & IC_ALLOCED) == 0) { 740 /* XXX ICP's "iir" driver just sends an event here. */ 741 panic("icp_intr: inactive CCB identified"); 742 } 743 744 /* 745 * Try to protect ourselves from the running command count already 746 * being 0 (e.g. if a polled command times out). 747 */ 748 KDASSERT(icp->icp_running != 0); 749 if (--icp->icp_running == 0 && 750 (icp->icp_flags & ICP_F_WAIT_FREEZE) != 0) { 751 icp->icp_flags &= ~ICP_F_WAIT_FREEZE; 752 wakeup(&icp->icp_qfreeze); 753 } 754 755 switch (icp->icp_status) { 756 case ICP_S_BSY: 757#ifdef ICP_DEBUG 758 printf("%s: ICP_S_BSY received\n", icp->icp_dv.dv_xname); 759#endif 760 if (__predict_false((ic->ic_flags & IC_UCMD) != 0)) 761 SIMPLEQ_INSERT_HEAD(&icp->icp_ucmd_queue, ic, ic_chain); 762 else 763 SIMPLEQ_INSERT_HEAD(&icp->icp_ccb_queue, ic, ic_chain); 764 break; 765 766 default: 767 ic->ic_flags |= IC_COMPLETE; 768 769 if ((ic->ic_flags & IC_WAITING) != 0) 770 wakeup(ic); 771 else if (ic->ic_intr != NULL) 772 (*ic->ic_intr)(ic); 773 774 if (ICP_HAS_WORK(icp)) 775 icp_ccb_enqueue(icp, NULL); 776 777 break; 778 } 779 780 return (1); 781} 782 783struct icp_ucmd_ctx { 784 gdt_ucmd_t *iu_ucmd; 785 u_int32_t iu_cnt; 786}; 787 788void 789icp_ucmd_intr(struct icp_ccb *ic) 790{ 791 struct icp_softc *icp = (void *) ic->ic_dv; 792 struct icp_ucmd_ctx *iu = ic->ic_context; 793 gdt_ucmd_t *ucmd = iu->iu_ucmd; 794 795 ucmd->status = icp->icp_status; 796 ucmd->info = icp->icp_info; 797 798 if (iu->iu_cnt != 0) { 799 bus_dmamap_sync(icp->icp_dmat, 800 icp->icp_scr_dmamap, 801 ICP_SCRATCH_UCMD, iu->iu_cnt, 802 BUS_DMASYNC_POSTREAD | BUS_DMASYNC_POSTWRITE); 803 memcpy(ucmd->data, 804 icp->icp_scr + ICP_SCRATCH_UCMD, iu->iu_cnt); 805 } 806 807 icp->icp_ucmd_ccb = NULL; 808 809 ic->ic_flags |= IC_COMPLETE; 810 wakeup(ic); 811} 812 813/* 814 * NOTE: We assume that it is safe to sleep here! 815 */ 816int 817icp_cmd(struct icp_softc *icp, u_int8_t service, u_int16_t opcode, 818 u_int32_t arg1, u_int32_t arg2, u_int32_t arg3) 819{ 820 struct icp_ioctlcmd *icmd; 821 struct icp_cachecmd *cc; 822 struct icp_rawcmd *rc; 823 int retries, rv; 824 struct icp_ccb *ic; 825 826 retries = ICP_RETRIES; 827 828 do { 829 ic = icp_ccb_alloc_wait(icp); 830 memset(&ic->ic_cmd, 0, sizeof(ic->ic_cmd)); 831 ic->ic_cmd.cmd_opcode = htole16(opcode); 832 833 switch (service) { 834 case ICP_CACHESERVICE: 835 if (opcode == ICP_IOCTL) { 836 icmd = &ic->ic_cmd.cmd_packet.ic; 837 icmd->ic_subfunc = htole16(arg1); 838 icmd->ic_channel = htole32(arg2); 839 icmd->ic_bufsize = htole32(arg3); 840 icmd->ic_addr = 841 htole32(icp->icp_scr_seg[0].ds_addr); 842 843 bus_dmamap_sync(icp->icp_dmat, 844 icp->icp_scr_dmamap, 0, arg3, 845 BUS_DMASYNC_PREWRITE | 846 BUS_DMASYNC_PREREAD); 847 } else { 848 cc = &ic->ic_cmd.cmd_packet.cc; 849 cc->cc_deviceno = htole16(arg1); 850 cc->cc_blockno = htole32(arg2); 851 } 852 break; 853 854 case ICP_SCSIRAWSERVICE: 855 rc = &ic->ic_cmd.cmd_packet.rc; 856 rc->rc_direction = htole32(arg1); 857 rc->rc_bus = arg2; 858 rc->rc_target = arg3; 859 rc->rc_lun = arg3 >> 8; 860 break; 861 } 862 863 ic->ic_service = service; 864 ic->ic_cmdlen = sizeof(ic->ic_cmd); 865 rv = icp_ccb_poll(icp, ic, 10000); 866 867 switch (service) { 868 case ICP_CACHESERVICE: 869 if (opcode == ICP_IOCTL) { 870 bus_dmamap_sync(icp->icp_dmat, 871 icp->icp_scr_dmamap, 0, arg3, 872 BUS_DMASYNC_POSTWRITE | 873 BUS_DMASYNC_POSTREAD); 874 } 875 break; 876 } 877 878 icp_ccb_free(icp, ic); 879 } while (rv != 0 && --retries > 0); 880 881 return (icp->icp_status == ICP_S_OK); 882} 883 884int 885icp_ucmd(struct icp_softc *icp, gdt_ucmd_t *ucmd) 886{ 887 struct icp_ccb *ic; 888 struct icp_ucmd_ctx iu; 889 u_int32_t cnt; 890 int error; 891 892 if (ucmd->service == ICP_CACHESERVICE) { 893 if (ucmd->command.cmd_opcode == ICP_IOCTL) { 894 cnt = ucmd->command.cmd_packet.ic.ic_bufsize; 895 if (cnt > GDT_SCRATCH_SZ) { 896 printf("%s: scratch buffer too small (%d/%d)\n", 897 icp->icp_dv.dv_xname, GDT_SCRATCH_SZ, cnt); 898 return (EINVAL); 899 } 900 } else { 901 cnt = ucmd->command.cmd_packet.cc.cc_blockcnt * 902 ICP_SECTOR_SIZE; 903 if (cnt > GDT_SCRATCH_SZ) { 904 printf("%s: scratch buffer too small (%d/%d)\n", 905 icp->icp_dv.dv_xname, GDT_SCRATCH_SZ, cnt); 906 return (EINVAL); 907 } 908 } 909 } else { 910 cnt = ucmd->command.cmd_packet.rc.rc_sdlen + 911 ucmd->command.cmd_packet.rc.rc_sense_len; 912 if (cnt > GDT_SCRATCH_SZ) { 913 printf("%s: scratch buffer too small (%d/%d)\n", 914 icp->icp_dv.dv_xname, GDT_SCRATCH_SZ, cnt); 915 return (EINVAL); 916 } 917 } 918 919 iu.iu_ucmd = ucmd; 920 iu.iu_cnt = cnt; 921 922 ic = icp_ccb_alloc_wait(icp); 923 memset(&ic->ic_cmd, 0, sizeof(ic->ic_cmd)); 924 ic->ic_cmd.cmd_opcode = htole16(ucmd->command.cmd_opcode); 925 926 if (ucmd->service == ICP_CACHESERVICE) { 927 if (ucmd->command.cmd_opcode == ICP_IOCTL) { 928 struct icp_ioctlcmd *icmd, *uicmd; 929 930 icmd = &ic->ic_cmd.cmd_packet.ic; 931 uicmd = &ucmd->command.cmd_packet.ic; 932 933 icmd->ic_subfunc = htole16(uicmd->ic_subfunc); 934 icmd->ic_channel = htole32(uicmd->ic_channel); 935 icmd->ic_bufsize = htole32(uicmd->ic_bufsize); 936 icmd->ic_addr = 937 htole32(icp->icp_scr_seg[0].ds_addr + 938 ICP_SCRATCH_UCMD); 939 } else { 940 struct icp_cachecmd *cc, *ucc; 941 942 cc = &ic->ic_cmd.cmd_packet.cc; 943 ucc = &ucmd->command.cmd_packet.cc; 944 945 cc->cc_deviceno = htole16(ucc->cc_deviceno); 946 cc->cc_blockno = htole32(ucc->cc_blockno); 947 cc->cc_blockcnt = htole32(ucc->cc_blockcnt); 948 cc->cc_addr = htole32(0xffffffffU); 949 cc->cc_nsgent = htole32(1); 950 cc->cc_sg[0].sg_addr = 951 htole32(icp->icp_scr_seg[0].ds_addr + 952 ICP_SCRATCH_UCMD); 953 cc->cc_sg[0].sg_len = htole32(cnt); 954 } 955 } else { 956 struct icp_rawcmd *rc, *urc; 957 958 rc = &ic->ic_cmd.cmd_packet.rc; 959 urc = &ucmd->command.cmd_packet.rc; 960 961 rc->rc_direction = htole32(urc->rc_direction); 962 rc->rc_sdata = htole32(0xffffffffU); 963 rc->rc_sdlen = htole32(urc->rc_sdlen); 964 rc->rc_clen = htole32(urc->rc_clen); 965 memcpy(rc->rc_cdb, urc->rc_cdb, sizeof(rc->rc_cdb)); 966 rc->rc_target = urc->rc_target; 967 rc->rc_lun = urc->rc_lun; 968 rc->rc_bus = urc->rc_bus; 969 rc->rc_sense_len = htole32(urc->rc_sense_len); 970 rc->rc_sense_addr = 971 htole32(icp->icp_scr_seg[0].ds_addr + 972 ICP_SCRATCH_UCMD + urc->rc_sdlen); 973 rc->rc_nsgent = htole32(1); 974 rc->rc_sg[0].sg_addr = 975 htole32(icp->icp_scr_seg[0].ds_addr + ICP_SCRATCH_UCMD); 976 rc->rc_sg[0].sg_len = htole32(cnt - urc->rc_sense_len); 977 } 978 979 ic->ic_service = ucmd->service; 980 ic->ic_cmdlen = sizeof(ic->ic_cmd); 981 ic->ic_context = &iu; 982 983 /* 984 * XXX What units are ucmd->timeout in? Until we know, we 985 * XXX just pull a number out of thin air. 986 */ 987 if (__predict_false((error = icp_ccb_wait_user(icp, ic, 30000)) != 0)) 988 printf("%s: error %d waiting for ucmd to complete\n", 989 icp->icp_dv.dv_xname, error); 990 991 /* icp_ucmd_intr() has updated ucmd. */ 992 icp_ccb_free(icp, ic); 993 994 return (error); 995} 996 997struct icp_ccb * 998icp_ccb_alloc(struct icp_softc *icp) 999{ 1000 struct icp_ccb *ic; 1001 int s; 1002 1003 s = splbio(); 1004 if (__predict_false((ic = 1005 SIMPLEQ_FIRST(&icp->icp_ccb_freelist)) == NULL)) { 1006 splx(s); 1007 return (NULL); 1008 } 1009 SIMPLEQ_REMOVE_HEAD(&icp->icp_ccb_freelist, ic_chain); 1010 splx(s); 1011 1012 ic->ic_flags = IC_ALLOCED; 1013 return (ic); 1014} 1015 1016struct icp_ccb * 1017icp_ccb_alloc_wait(struct icp_softc *icp) 1018{ 1019 struct icp_ccb *ic; 1020 int s; 1021 1022 s = splbio(); 1023 while ((ic = SIMPLEQ_FIRST(&icp->icp_ccb_freelist)) == NULL) { 1024 icp->icp_flags |= ICP_F_WAIT_CCB; 1025 (void) tsleep(&icp->icp_ccb_freelist, PRIBIO, "icpccb", 0); 1026 } 1027 SIMPLEQ_REMOVE_HEAD(&icp->icp_ccb_freelist, ic_chain); 1028 splx(s); 1029 1030 ic->ic_flags = IC_ALLOCED; 1031 return (ic); 1032} 1033 1034void 1035icp_ccb_free(struct icp_softc *icp, struct icp_ccb *ic) 1036{ 1037 int s; 1038 1039 s = splbio(); 1040 ic->ic_flags = 0; 1041 ic->ic_intr = NULL; 1042 SIMPLEQ_INSERT_HEAD(&icp->icp_ccb_freelist, ic, ic_chain); 1043 if (__predict_false((icp->icp_flags & ICP_F_WAIT_CCB) != 0)) { 1044 icp->icp_flags &= ~ICP_F_WAIT_CCB; 1045 wakeup(&icp->icp_ccb_freelist); 1046 } 1047 splx(s); 1048} 1049 1050void 1051icp_ccb_enqueue(struct icp_softc *icp, struct icp_ccb *ic) 1052{ 1053 int s; 1054 1055 s = splbio(); 1056 1057 if (ic != NULL) { 1058 if (__predict_false((ic->ic_flags & IC_UCMD) != 0)) 1059 SIMPLEQ_INSERT_TAIL(&icp->icp_ucmd_queue, ic, ic_chain); 1060 else 1061 SIMPLEQ_INSERT_TAIL(&icp->icp_ccb_queue, ic, ic_chain); 1062 } 1063 1064 for (; icp->icp_qfreeze == 0;) { 1065 if (__predict_false((ic = 1066 SIMPLEQ_FIRST(&icp->icp_ucmd_queue)) != NULL)) { 1067 struct icp_ucmd_ctx *iu = ic->ic_context; 1068 gdt_ucmd_t *ucmd = iu->iu_ucmd; 1069 1070 /* 1071 * All user-generated commands share the same 1072 * scratch space, so if one is already running, 1073 * we have to stall the command queue. 1074 */ 1075 if (icp->icp_ucmd_ccb != NULL) 1076 break; 1077 if ((*icp->icp_test_busy)(icp)) 1078 break; 1079 icp->icp_ucmd_ccb = ic; 1080 1081 if (iu->iu_cnt != 0) { 1082 memcpy(icp->icp_scr + ICP_SCRATCH_UCMD, 1083 ucmd->data, iu->iu_cnt); 1084 bus_dmamap_sync(icp->icp_dmat, 1085 icp->icp_scr_dmamap, 1086 ICP_SCRATCH_UCMD, iu->iu_cnt, 1087 BUS_DMASYNC_PREREAD | 1088 BUS_DMASYNC_PREWRITE); 1089 } 1090 } else if (__predict_true((ic = 1091 SIMPLEQ_FIRST(&icp->icp_ccb_queue)) != NULL)) { 1092 if ((*icp->icp_test_busy)(icp)) 1093 break; 1094 } else { 1095 /* no command found */ 1096 break; 1097 } 1098 icp_ccb_submit(icp, ic); 1099 if (__predict_false((ic->ic_flags & IC_UCMD) != 0)) 1100 SIMPLEQ_REMOVE_HEAD(&icp->icp_ucmd_queue, ic_chain); 1101 else 1102 SIMPLEQ_REMOVE_HEAD(&icp->icp_ccb_queue, ic_chain); 1103 } 1104 1105 splx(s); 1106} 1107 1108int 1109icp_ccb_map(struct icp_softc *icp, struct icp_ccb *ic, void *data, int size, 1110 int dir) 1111{ 1112 struct icp_sg *sg; 1113 int nsegs, i, rv; 1114 bus_dmamap_t xfer; 1115 1116 xfer = ic->ic_xfer_map; 1117 1118 rv = bus_dmamap_load(icp->icp_dmat, xfer, data, size, NULL, 1119 BUS_DMA_NOWAIT | BUS_DMA_STREAMING | 1120 ((dir & IC_XFER_IN) ? BUS_DMA_READ : BUS_DMA_WRITE)); 1121 if (rv != 0) 1122 return (rv); 1123 1124 nsegs = xfer->dm_nsegs; 1125 ic->ic_xfer_size = size; 1126 ic->ic_nsgent = nsegs; 1127 ic->ic_flags |= dir; 1128 sg = ic->ic_sg; 1129 1130 if (sg != NULL) { 1131 for (i = 0; i < nsegs; i++, sg++) { 1132 sg->sg_addr = htole32(xfer->dm_segs[i].ds_addr); 1133 sg->sg_len = htole32(xfer->dm_segs[i].ds_len); 1134 } 1135 } else if (nsegs > 1) 1136 panic("icp_ccb_map: no SG list specified, but nsegs > 1"); 1137 1138 if ((dir & IC_XFER_OUT) != 0) 1139 i = BUS_DMASYNC_PREWRITE; 1140 else /* if ((dir & IC_XFER_IN) != 0) */ 1141 i = BUS_DMASYNC_PREREAD; 1142 1143 bus_dmamap_sync(icp->icp_dmat, xfer, 0, ic->ic_xfer_size, i); 1144 return (0); 1145} 1146 1147void 1148icp_ccb_unmap(struct icp_softc *icp, struct icp_ccb *ic) 1149{ 1150 int i; 1151 1152 if ((ic->ic_flags & IC_XFER_OUT) != 0) 1153 i = BUS_DMASYNC_POSTWRITE; 1154 else /* if ((ic->ic_flags & IC_XFER_IN) != 0) */ 1155 i = BUS_DMASYNC_POSTREAD; 1156 1157 bus_dmamap_sync(icp->icp_dmat, ic->ic_xfer_map, 0, ic->ic_xfer_size, i); 1158 bus_dmamap_unload(icp->icp_dmat, ic->ic_xfer_map); 1159} 1160 1161int 1162icp_ccb_poll(struct icp_softc *icp, struct icp_ccb *ic, int timo) 1163{ 1164 int s, rv; 1165 1166 s = splbio(); 1167 1168 for (timo = ICP_BUSY_WAIT_MS * 100; timo != 0; timo--) { 1169 if (!(*icp->icp_test_busy)(icp)) 1170 break; 1171 DELAY(10); 1172 } 1173 if (timo == 0) { 1174 printf("%s: submit: busy\n", icp->icp_dv.dv_xname); 1175 return (EAGAIN); 1176 } 1177 1178 icp_ccb_submit(icp, ic); 1179 1180 if (cold) { 1181 for (timo *= 10; timo != 0; timo--) { 1182 DELAY(100); 1183 icp_intr(icp); 1184 if ((ic->ic_flags & IC_COMPLETE) != 0) 1185 break; 1186 } 1187 } else { 1188 ic->ic_flags |= IC_WAITING; 1189 while ((ic->ic_flags & IC_COMPLETE) == 0) { 1190 if ((rv = tsleep(ic, PRIBIO, "icpwccb", 1191 mstohz(timo))) != 0) { 1192 timo = 0; 1193 break; 1194 } 1195 } 1196 } 1197 1198 if (timo != 0) { 1199 if (ic->ic_status != ICP_S_OK) { 1200#ifdef ICP_DEBUG 1201 printf("%s: request failed; status=0x%04x\n", 1202 icp->icp_dv.dv_xname, ic->ic_status); 1203#endif 1204 rv = EIO; 1205 } else 1206 rv = 0; 1207 } else { 1208 printf("%s: command timed out\n", icp->icp_dv.dv_xname); 1209 rv = EIO; 1210 } 1211 1212 while ((*icp->icp_test_busy)(icp) != 0) 1213 DELAY(10); 1214 1215 splx(s); 1216 1217 return (rv); 1218} 1219 1220int 1221icp_ccb_wait(struct icp_softc *icp, struct icp_ccb *ic, int timo) 1222{ 1223 int s, rv; 1224 1225 ic->ic_flags |= IC_WAITING; 1226 1227 s = splbio(); 1228 icp_ccb_enqueue(icp, ic); 1229 while ((ic->ic_flags & IC_COMPLETE) == 0) { 1230 if ((rv = tsleep(ic, PRIBIO, "icpwccb", mstohz(timo))) != 0) { 1231 splx(s); 1232 return (rv); 1233 } 1234 } 1235 splx(s); 1236 1237 if (ic->ic_status != ICP_S_OK) { 1238 printf("%s: command failed; status=%x\n", icp->icp_dv.dv_xname, 1239 ic->ic_status); 1240 return (EIO); 1241 } 1242 1243 return (0); 1244} 1245 1246int 1247icp_ccb_wait_user(struct icp_softc *icp, struct icp_ccb *ic, int timo) 1248{ 1249 int s, rv; 1250 1251 ic->ic_dv = &icp->icp_dv; 1252 ic->ic_intr = icp_ucmd_intr; 1253 ic->ic_flags |= IC_UCMD; 1254 1255 s = splbio(); 1256 icp_ccb_enqueue(icp, ic); 1257 while ((ic->ic_flags & IC_COMPLETE) == 0) { 1258 if ((rv = tsleep(ic, PRIBIO, "icpwuccb", mstohz(timo))) != 0) { 1259 splx(s); 1260 return (rv); 1261 } 1262 } 1263 splx(s); 1264 1265 return (0); 1266} 1267 1268void 1269icp_ccb_submit(struct icp_softc *icp, struct icp_ccb *ic) 1270{ 1271 1272 ic->ic_cmdlen = (ic->ic_cmdlen + 3) & ~3; 1273 1274 (*icp->icp_set_sema0)(icp); 1275 DELAY(10); 1276 1277 ic->ic_cmd.cmd_boardnode = htole32(ICP_LOCALBOARD); 1278 ic->ic_cmd.cmd_cmdindex = htole32(ic->ic_ident); 1279 1280 icp->icp_running++; 1281 1282 (*icp->icp_copy_cmd)(icp, ic); 1283 (*icp->icp_release_event)(icp, ic); 1284} 1285 1286int 1287icp_freeze(struct icp_softc *icp) 1288{ 1289 int s, error = 0; 1290 1291 s = splbio(); 1292 if (icp->icp_qfreeze++ == 0) { 1293 while (icp->icp_running != 0) { 1294 icp->icp_flags |= ICP_F_WAIT_FREEZE; 1295 error = tsleep(&icp->icp_qfreeze, PRIBIO|PCATCH, 1296 "icpqfrz", 0); 1297 if (error != 0 && --icp->icp_qfreeze == 0 && 1298 ICP_HAS_WORK(icp)) { 1299 icp_ccb_enqueue(icp, NULL); 1300 break; 1301 } 1302 } 1303 } 1304 splx(s); 1305 1306 return (error); 1307} 1308 1309void 1310icp_unfreeze(struct icp_softc *icp) 1311{ 1312 int s; 1313 1314 s = splbio(); 1315 KDASSERT(icp->icp_qfreeze != 0); 1316 if (--icp->icp_qfreeze == 0 && ICP_HAS_WORK(icp)) 1317 icp_ccb_enqueue(icp, NULL); 1318 splx(s); 1319} 1320 1321/* XXX Global - should be per-controller? XXX */ 1322static gdt_evt_str icp_event_buffer[ICP_MAX_EVENTS]; 1323static int icp_event_oldidx; 1324static int icp_event_lastidx; 1325 1326gdt_evt_str * 1327icp_store_event(struct icp_softc *icp, u_int16_t source, u_int16_t idx, 1328 gdt_evt_data *evt) 1329{ 1330 gdt_evt_str *e; 1331 1332 /* no source == no event */ 1333 if (source == 0) 1334 return (NULL); 1335 1336 e = &icp_event_buffer[icp_event_lastidx]; 1337 if (e->event_source == source && e->event_idx == idx && 1338 ((evt->size != 0 && e->event_data.size != 0 && 1339 memcmp(&e->event_data.eu, &evt->eu, evt->size) == 0) || 1340 (evt->size == 0 && e->event_data.size == 0 && 1341 strcmp((char *) e->event_data.event_string, 1342 (char *) evt->event_string) == 0))) { 1343 e->last_stamp = time.tv_sec; 1344 e->same_count++; 1345 } else { 1346 if (icp_event_buffer[icp_event_lastidx].event_source != 0) { 1347 icp_event_lastidx++; 1348 if (icp_event_lastidx == ICP_MAX_EVENTS) 1349 icp_event_lastidx = 0; 1350 if (icp_event_lastidx == icp_event_oldidx) { 1351 icp_event_oldidx++; 1352 if (icp_event_oldidx == ICP_MAX_EVENTS) 1353 icp_event_oldidx = 0; 1354 } 1355 } 1356 e = &icp_event_buffer[icp_event_lastidx]; 1357 e->event_source = source; 1358 e->event_idx = idx; 1359 e->first_stamp = e->last_stamp = time.tv_sec; 1360 e->same_count = 1; 1361 e->event_data = *evt; 1362 e->application = 0; 1363 } 1364 return (e); 1365} 1366 1367int 1368icp_read_event(struct icp_softc *icp, int handle, gdt_evt_str *estr) 1369{ 1370 gdt_evt_str *e; 1371 int eindex, s; 1372 1373 s = splbio(); 1374 1375 if (handle == -1) 1376 eindex = icp_event_oldidx; 1377 else 1378 eindex = handle; 1379 1380 estr->event_source = 0; 1381 1382 if (eindex < 0 || eindex >= ICP_MAX_EVENTS) { 1383 splx(s); 1384 return (eindex); 1385 } 1386 1387 e = &icp_event_buffer[eindex]; 1388 if (e->event_source != 0) { 1389 if (eindex != icp_event_lastidx) { 1390 eindex++; 1391 if (eindex == ICP_MAX_EVENTS) 1392 eindex = 0; 1393 } else 1394 eindex = -1; 1395 memcpy(estr, e, sizeof(gdt_evt_str)); 1396 } 1397 1398 splx(s); 1399 1400 return (eindex); 1401} 1402 1403void 1404icp_readapp_event(struct icp_softc *icp, u_int8_t application, 1405 gdt_evt_str *estr) 1406{ 1407 gdt_evt_str *e; 1408 int found = 0, eindex, s; 1409 1410 s = splbio(); 1411 1412 eindex = icp_event_oldidx; 1413 for (;;) { 1414 e = &icp_event_buffer[eindex]; 1415 if (e->event_source == 0) 1416 break; 1417 if ((e->application & application) == 0) { 1418 e->application |= application; 1419 found = 1; 1420 break; 1421 } 1422 if (eindex == icp_event_lastidx) 1423 break; 1424 eindex++; 1425 if (eindex == ICP_MAX_EVENTS) 1426 eindex = 0; 1427 } 1428 if (found) 1429 memcpy(estr, e, sizeof(gdt_evt_str)); 1430 else 1431 estr->event_source = 0; 1432 1433 splx(s); 1434} 1435 1436void 1437icp_clear_events(struct icp_softc *icp) 1438{ 1439 int s; 1440 1441 s = splbio(); 1442 icp_event_oldidx = icp_event_lastidx = 0; 1443 memset(icp_event_buffer, 0, sizeof(icp_event_buffer)); 1444 splx(s); 1445} 1446