lsi64854.c (226381) | lsi64854.c (226947) |
---|---|
1/*- 2 * Copyright (c) 2004 Scott Long 3 * All rights reserved. 4 * 5 * Redistribution and use in source and binary forms, with or without 6 * modification, are permitted provided that the following conditions 7 * are met: 8 * 1. Redistributions of source code must retain the above copyright --- 43 unchanged lines hidden (view full) --- 52 * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS 53 * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN 54 * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) 55 * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE 56 * POSSIBILITY OF SUCH DAMAGE. 57 */ 58 59#include <sys/cdefs.h> | 1/*- 2 * Copyright (c) 2004 Scott Long 3 * All rights reserved. 4 * 5 * Redistribution and use in source and binary forms, with or without 6 * modification, are permitted provided that the following conditions 7 * are met: 8 * 1. Redistributions of source code must retain the above copyright --- 43 unchanged lines hidden (view full) --- 52 * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS 53 * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN 54 * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) 55 * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE 56 * POSSIBILITY OF SUCH DAMAGE. 57 */ 58 59#include <sys/cdefs.h> |
60__FBSDID("$FreeBSD: head/sys/sparc64/sbus/lsi64854.c 226381 2011-10-15 09:29:43Z marius $"); | 60__FBSDID("$FreeBSD: head/sys/sparc64/sbus/lsi64854.c 226947 2011-10-30 21:17:42Z marius $"); |
61 62#include <sys/param.h> 63#include <sys/systm.h> 64#include <sys/bus.h> 65#include <sys/kernel.h> 66#include <sys/lock.h> 67#include <sys/mutex.h> 68#include <sys/rman.h> --- 20 unchanged lines hidden (view full) --- 89 do { \ 90 if ((lsi64854debug & (a)) != 0) \ 91 printf x; \ 92 } while (/* CONSTCOND */0) 93#else 94#define DPRINTF(a,x) 95#endif 96 | 61 62#include <sys/param.h> 63#include <sys/systm.h> 64#include <sys/bus.h> 65#include <sys/kernel.h> 66#include <sys/lock.h> 67#include <sys/mutex.h> 68#include <sys/rman.h> --- 20 unchanged lines hidden (view full) --- 89 do { \ 90 if ((lsi64854debug & (a)) != 0) \ 91 printf x; \ 92 } while (/* CONSTCOND */0) 93#else 94#define DPRINTF(a,x) 95#endif 96 |
97#define MAX_DMA_SZ (16*1024*1024) | 97/* 98 * The rules say we cannot transfer more than the limit of this DMA chip (64k 99 * for old and 16Mb for new), and we cannot cross a 16Mb boundary. 100 */ 101#define MAX_DMA_SZ (64 * 1024) 102#define BOUNDARY (16 * 1024 * 1024) |
98 99static void lsi64854_reset(struct lsi64854_softc *); 100static void lsi64854_map_scsi(void *, bus_dma_segment_t *, int, int); 101static int lsi64854_setup(struct lsi64854_softc *, void **, size_t *, 102 int, size_t *); 103static int lsi64854_scsi_intr(void *); 104static int lsi64854_enet_intr(void *); 105static int lsi64854_setup_pp(struct lsi64854_softc *, void **, --- 14 unchanged lines hidden (view full) --- 120 bus_dma_lock_t *lockfunc; 121 struct ncr53c9x_softc *nsc; 122 void *lockfuncarg; 123 uint32_t csr; 124 int error; 125 126 lockfunc = NULL; 127 lockfuncarg = NULL; | 103 104static void lsi64854_reset(struct lsi64854_softc *); 105static void lsi64854_map_scsi(void *, bus_dma_segment_t *, int, int); 106static int lsi64854_setup(struct lsi64854_softc *, void **, size_t *, 107 int, size_t *); 108static int lsi64854_scsi_intr(void *); 109static int lsi64854_enet_intr(void *); 110static int lsi64854_setup_pp(struct lsi64854_softc *, void **, --- 14 unchanged lines hidden (view full) --- 125 bus_dma_lock_t *lockfunc; 126 struct ncr53c9x_softc *nsc; 127 void *lockfuncarg; 128 uint32_t csr; 129 int error; 130 131 lockfunc = NULL; 132 lockfuncarg = NULL; |
133 sc->sc_maxdmasize = MAX_DMA_SZ; |
|
128 129 switch (sc->sc_channel) { 130 case L64854_CHANNEL_SCSI: 131 nsc = sc->sc_client; 132 if (NCR_LOCK_INITIALIZED(nsc) == 0) { 133 device_printf(sc->sc_dev, "mutex not initialized\n"); 134 return (ENXIO); 135 } 136 lockfunc = busdma_lock_mutex; 137 lockfuncarg = &nsc->sc_lock; | 134 135 switch (sc->sc_channel) { 136 case L64854_CHANNEL_SCSI: 137 nsc = sc->sc_client; 138 if (NCR_LOCK_INITIALIZED(nsc) == 0) { 139 device_printf(sc->sc_dev, "mutex not initialized\n"); 140 return (ENXIO); 141 } 142 lockfunc = busdma_lock_mutex; 143 lockfuncarg = &nsc->sc_lock; |
144 sc->sc_maxdmasize = nsc->sc_maxxfer; |
|
138 sc->intr = lsi64854_scsi_intr; 139 sc->setup = lsi64854_setup; 140 break; 141 case L64854_CHANNEL_ENET: 142 sc->intr = lsi64854_enet_intr; 143 break; 144 case L64854_CHANNEL_PP: 145 sc->intr = lsi64854_pp_intr; 146 sc->setup = lsi64854_setup_pp; 147 break; 148 default: 149 device_printf(sc->sc_dev, "unknown channel\n"); 150 } 151 sc->reset = lsi64854_reset; 152 153 if (sc->setup != NULL) { 154 error = bus_dma_tag_create( 155 sc->sc_parent_dmat, /* parent */ | 145 sc->intr = lsi64854_scsi_intr; 146 sc->setup = lsi64854_setup; 147 break; 148 case L64854_CHANNEL_ENET: 149 sc->intr = lsi64854_enet_intr; 150 break; 151 case L64854_CHANNEL_PP: 152 sc->intr = lsi64854_pp_intr; 153 sc->setup = lsi64854_setup_pp; 154 break; 155 default: 156 device_printf(sc->sc_dev, "unknown channel\n"); 157 } 158 sc->reset = lsi64854_reset; 159 160 if (sc->setup != NULL) { 161 error = bus_dma_tag_create( 162 sc->sc_parent_dmat, /* parent */ |
156 1, 0, /* alignment, boundary */ | 163 1, BOUNDARY, /* alignment, boundary */ |
157 BUS_SPACE_MAXADDR, /* lowaddr */ 158 BUS_SPACE_MAXADDR, /* highaddr */ 159 NULL, NULL, /* filter, filterarg */ | 164 BUS_SPACE_MAXADDR, /* lowaddr */ 165 BUS_SPACE_MAXADDR, /* highaddr */ 166 NULL, NULL, /* filter, filterarg */ |
160 MAX_DMA_SZ, /* maxsize */ | 167 sc->sc_maxdmasize, /* maxsize */ |
161 1, /* nsegments */ | 168 1, /* nsegments */ |
162 MAX_DMA_SZ, /* maxsegsize */ | 169 sc->sc_maxdmasize, /* maxsegsize */ |
163 BUS_DMA_ALLOCNOW, /* flags */ 164 lockfunc, lockfuncarg, /* lockfunc, lockfuncarg */ 165 &sc->sc_buffer_dmat); 166 if (error != 0) { 167 device_printf(sc->sc_dev, 168 "cannot allocate buffer DMA tag\n"); 169 return (error); 170 } --- 74 unchanged lines hidden (view full) --- 245 uint32_t csr; \ 246 /* \ 247 * DMA rev0 & rev1: we are not allowed to touch the DMA "flush" \ 248 * and "drain" bits while it is still thinking about a \ 249 * request. \ 250 * other revs: D_ESC_R_PEND bit reads as 0 \ 251 */ \ 252 DMAWAIT(sc, L64854_GCSR(sc) & D_ESC_R_PEND, "R_PEND", dontpanic);\ | 170 BUS_DMA_ALLOCNOW, /* flags */ 171 lockfunc, lockfuncarg, /* lockfunc, lockfuncarg */ 172 &sc->sc_buffer_dmat); 173 if (error != 0) { 174 device_printf(sc->sc_dev, 175 "cannot allocate buffer DMA tag\n"); 176 return (error); 177 } --- 74 unchanged lines hidden (view full) --- 252 uint32_t csr; \ 253 /* \ 254 * DMA rev0 & rev1: we are not allowed to touch the DMA "flush" \ 255 * and "drain" bits while it is still thinking about a \ 256 * request. \ 257 * other revs: D_ESC_R_PEND bit reads as 0 \ 258 */ \ 259 DMAWAIT(sc, L64854_GCSR(sc) & D_ESC_R_PEND, "R_PEND", dontpanic);\ |
253 if (sc->sc_rev != DMAREV_HME) { \ 254 /* \ 255 * Select drain bit based on revision \ 256 * also clears errors and D_TC flag \ 257 */ \ 258 csr = L64854_GCSR(sc); \ 259 if (sc->sc_rev == DMAREV_1 || sc->sc_rev == DMAREV_0) \ 260 csr |= D_ESC_DRAIN; \ 261 else \ 262 csr |= L64854_INVALIDATE; \ | 260 if (sc->sc_rev != DMAREV_HME) { \ 261 /* \ 262 * Select drain bit based on revision \ 263 * also clears errors and D_TC flag \ 264 */ \ 265 csr = L64854_GCSR(sc); \ 266 if (sc->sc_rev == DMAREV_1 || sc->sc_rev == DMAREV_0) \ 267 csr |= D_ESC_DRAIN; \ 268 else \ 269 csr |= L64854_INVALIDATE; \ |
263 \ | 270 \ |
264 L64854_SCSR(sc,csr); \ | 271 L64854_SCSR(sc, csr); \ |
265 } \ 266 /* \ 267 * Wait for draining to finish \ 268 * rev0 & rev1 call this PACKCNT \ 269 */ \ | 272 } \ 273 /* \ 274 * Wait for draining to finish \ 275 * rev0 & rev1 call this PACKCNT \ 276 */ \ |
270 DMAWAIT(sc, L64854_GCSR(sc) & L64854_DRAINING, "DRAINING", dontpanic);\ | 277 DMAWAIT(sc, L64854_GCSR(sc) & L64854_DRAINING, "DRAINING", \ 278 dontpanic); \ |
271} while (/* CONSTCOND */0) 272 273#define DMA_FLUSH(sc, dontpanic) do { \ 274 uint32_t csr; \ 275 /* \ 276 * DMA rev0 & rev1: we are not allowed to touch the DMA "flush" \ 277 * and "drain" bits while it is still thinking about a \ 278 * request. \ 279 * other revs: D_ESC_R_PEND bit reads as 0 \ 280 */ \ 281 DMAWAIT(sc, L64854_GCSR(sc) & D_ESC_R_PEND, "R_PEND", dontpanic);\ 282 csr = L64854_GCSR(sc); \ 283 csr &= ~(L64854_WRITE|L64854_EN_DMA); /* no-ops on ENET */ \ 284 csr |= L64854_INVALIDATE; /* XXX FAS ? */ \ | 279} while (/* CONSTCOND */0) 280 281#define DMA_FLUSH(sc, dontpanic) do { \ 282 uint32_t csr; \ 283 /* \ 284 * DMA rev0 & rev1: we are not allowed to touch the DMA "flush" \ 285 * and "drain" bits while it is still thinking about a \ 286 * request. \ 287 * other revs: D_ESC_R_PEND bit reads as 0 \ 288 */ \ 289 DMAWAIT(sc, L64854_GCSR(sc) & D_ESC_R_PEND, "R_PEND", dontpanic);\ 290 csr = L64854_GCSR(sc); \ 291 csr &= ~(L64854_WRITE|L64854_EN_DMA); /* no-ops on ENET */ \ 292 csr |= L64854_INVALIDATE; /* XXX FAS ? */ \ |
285 L64854_SCSR(sc,csr); \ | 293 L64854_SCSR(sc, csr); \ |
286} while (/* CONSTCOND */0) 287 288static void 289lsi64854_reset(struct lsi64854_softc *sc) 290{ | 294} while (/* CONSTCOND */0) 295 296static void 297lsi64854_reset(struct lsi64854_softc *sc) 298{ |
299 bus_dma_tag_t dmat; 300 bus_dmamap_t dmam; |
|
291 uint32_t csr; 292 293 DMA_FLUSH(sc, 1); 294 csr = L64854_GCSR(sc); 295 296 DPRINTF(LDB_ANY, ("%s: csr 0x%x\n", __func__, csr)); 297 298 if (sc->sc_dmasize != 0) { | 301 uint32_t csr; 302 303 DMA_FLUSH(sc, 1); 304 csr = L64854_GCSR(sc); 305 306 DPRINTF(LDB_ANY, ("%s: csr 0x%x\n", __func__, csr)); 307 308 if (sc->sc_dmasize != 0) { |
299 bus_dmamap_sync(sc->sc_buffer_dmat, sc->sc_dmamap, 300 (csr & D_WRITE) != 0 ? BUS_DMASYNC_PREREAD : 301 BUS_DMASYNC_PREWRITE); 302 bus_dmamap_unload(sc->sc_buffer_dmat, sc->sc_dmamap); | 309 dmat = sc->sc_buffer_dmat; 310 dmam = sc->sc_dmamap; 311 bus_dmamap_sync(dmat, dmam, (csr & D_WRITE) != 0 ? 312 BUS_DMASYNC_PREREAD : BUS_DMASYNC_PREWRITE); 313 bus_dmamap_unload(dmat, dmam); |
303 } 304 305 if (sc->sc_rev == DMAREV_HME) 306 L64854_SCSR(sc, csr | D_HW_RESET_FAS366); 307 308 csr |= L64854_RESET; /* reset DMA */ 309 L64854_SCSR(sc, csr); 310 DELAY(200); /* > 10 Sbus clocks(?) */ --- 48 unchanged lines hidden (view full) --- 359 360static void 361lsi64854_map_scsi(void *arg, bus_dma_segment_t *segs, int nseg, int error) 362{ 363 struct lsi64854_softc *sc; 364 365 sc = (struct lsi64854_softc *)arg; 366 | 314 } 315 316 if (sc->sc_rev == DMAREV_HME) 317 L64854_SCSR(sc, csr | D_HW_RESET_FAS366); 318 319 csr |= L64854_RESET; /* reset DMA */ 320 L64854_SCSR(sc, csr); 321 DELAY(200); /* > 10 Sbus clocks(?) */ --- 48 unchanged lines hidden (view full) --- 370 371static void 372lsi64854_map_scsi(void *arg, bus_dma_segment_t *segs, int nseg, int error) 373{ 374 struct lsi64854_softc *sc; 375 376 sc = (struct lsi64854_softc *)arg; 377 |
378 if (error != 0) 379 return; |
|
367 if (nseg != 1) 368 panic("%s: cannot map %d segments\n", __func__, nseg); 369 370 bus_dmamap_sync(sc->sc_buffer_dmat, sc->sc_dmamap, | 380 if (nseg != 1) 381 panic("%s: cannot map %d segments\n", __func__, nseg); 382 383 bus_dmamap_sync(sc->sc_buffer_dmat, sc->sc_dmamap, |
371 sc->sc_datain ? BUS_DMASYNC_PREREAD : BUS_DMASYNC_PREWRITE); | 384 sc->sc_datain != 0 ? BUS_DMASYNC_PREREAD : BUS_DMASYNC_PREWRITE); |
372 bus_write_4(sc->sc_res, L64854_REG_ADDR, segs[0].ds_addr); 373} 374 | 385 bus_write_4(sc->sc_res, L64854_REG_ADDR, segs[0].ds_addr); 386} 387 |
375#define DMAMAX(a) (MAX_DMA_SZ - ((a) & (MAX_DMA_SZ - 1))) | |
376/* 377 * setup a DMA transfer 378 */ 379static int 380lsi64854_setup(struct lsi64854_softc *sc, void **addr, size_t *len, 381 int datain, size_t *dmasize) 382{ 383 long bcnt; | 388/* 389 * setup a DMA transfer 390 */ 391static int 392lsi64854_setup(struct lsi64854_softc *sc, void **addr, size_t *len, 393 int datain, size_t *dmasize) 394{ 395 long bcnt; |
396 int error; |
|
384 uint32_t csr; 385 386 DMA_FLUSH(sc, 0); 387 388#if 0 389 DMACSR(sc) &= ~D_INT_EN; 390#endif 391 sc->sc_dmaaddr = addr; 392 sc->sc_dmalen = len; 393 sc->sc_datain = datain; 394 | 397 uint32_t csr; 398 399 DMA_FLUSH(sc, 0); 400 401#if 0 402 DMACSR(sc) &= ~D_INT_EN; 403#endif 404 sc->sc_dmaaddr = addr; 405 sc->sc_dmalen = len; 406 sc->sc_datain = datain; 407 |
395 /* 396 * The rules say we cannot transfer more than the limit 397 * of this DMA chip (64k for old and 16Mb for new), 398 * and we cannot cross a 16Mb boundary. 399 */ 400 *dmasize = sc->sc_dmasize = 401 ulmin(*dmasize, DMAMAX((size_t)*sc->sc_dmaaddr)); | 408 KASSERT(*dmasize <= sc->sc_maxdmasize, 409 ("%s: transfer size %ld too large", __func__, (long)*dmasize)); |
402 | 410 |
403 DPRINTF(LDB_ANY, ("%s: dmasize=%ld\n", __func__, (long)sc->sc_dmasize)); | 411 sc->sc_dmasize = *dmasize; |
404 | 412 |
413 DPRINTF(LDB_ANY, ("%s: dmasize=%ld\n", __func__, (long)*dmasize)); 414 |
|
405 /* 406 * XXX what length? 407 */ 408 if (sc->sc_rev == DMAREV_HME) { 409 L64854_SCSR(sc, sc->sc_dmactl | L64854_RESET); 410 L64854_SCSR(sc, sc->sc_dmactl); 411 412 bus_write_4(sc->sc_res, L64854_REG_CNT, *dmasize); 413 } 414 | 415 /* 416 * XXX what length? 417 */ 418 if (sc->sc_rev == DMAREV_HME) { 419 L64854_SCSR(sc, sc->sc_dmactl | L64854_RESET); 420 L64854_SCSR(sc, sc->sc_dmactl); 421 422 bus_write_4(sc->sc_res, L64854_REG_CNT, *dmasize); 423 } 424 |
415 /* Program the DMA address */ 416 if (sc->sc_dmasize != 0) 417 if (bus_dmamap_load(sc->sc_buffer_dmat, sc->sc_dmamap, 418 *sc->sc_dmaaddr, sc->sc_dmasize, lsi64854_map_scsi, sc, 0)) 419 panic("%s: cannot allocate DVMA address", __func__); | 425 /* 426 * Load the transfer buffer and program the DMA address. 427 * Note that the NCR53C9x core can't handle EINPROGRESS so we set 428 * BUS_DMA_NOWAIT. 429 */ 430 if (*dmasize != 0) { 431 error = bus_dmamap_load(sc->sc_buffer_dmat, sc->sc_dmamap, 432 *sc->sc_dmaaddr, *dmasize, lsi64854_map_scsi, sc, 433 BUS_DMA_NOWAIT); 434 if (error != 0) 435 return (error); 436 } |
420 421 if (sc->sc_rev == DMAREV_ESC) { 422 /* DMA ESC chip bug work-around */ | 437 438 if (sc->sc_rev == DMAREV_ESC) { 439 /* DMA ESC chip bug work-around */ |
423 bcnt = sc->sc_dmasize; | 440 bcnt = *dmasize; |
424 if (((bcnt + (long)*sc->sc_dmaaddr) & PAGE_MASK_8K) != 0) 425 bcnt = roundup(bcnt, PAGE_SIZE_8K); 426 bus_write_4(sc->sc_res, L64854_REG_CNT, bcnt); 427 } 428 | 441 if (((bcnt + (long)*sc->sc_dmaaddr) & PAGE_MASK_8K) != 0) 442 bcnt = roundup(bcnt, PAGE_SIZE_8K); 443 bus_write_4(sc->sc_res, L64854_REG_CNT, bcnt); 444 } 445 |
429 /* Setup DMA control register */ | 446 /* Setup the DMA control register. */ |
430 csr = L64854_GCSR(sc); 431 | 447 csr = L64854_GCSR(sc); 448 |
432 if (datain) | 449 if (datain != 0) |
433 csr |= L64854_WRITE; 434 else 435 csr &= ~L64854_WRITE; 436 csr |= L64854_INT_EN; 437 438 if (sc->sc_rev == DMAREV_HME) 439 csr |= (D_DSBL_SCSI_DRN | D_EN_DMA); 440 --- 9 unchanged lines hidden (view full) --- 450 * 451 * return 1 if it was a DMA continue. 452 */ 453static int 454lsi64854_scsi_intr(void *arg) 455{ 456 struct lsi64854_softc *sc = arg; 457 struct ncr53c9x_softc *nsc = sc->sc_client; | 450 csr |= L64854_WRITE; 451 else 452 csr &= ~L64854_WRITE; 453 csr |= L64854_INT_EN; 454 455 if (sc->sc_rev == DMAREV_HME) 456 csr |= (D_DSBL_SCSI_DRN | D_EN_DMA); 457 --- 9 unchanged lines hidden (view full) --- 467 * 468 * return 1 if it was a DMA continue. 469 */ 470static int 471lsi64854_scsi_intr(void *arg) 472{ 473 struct lsi64854_softc *sc = arg; 474 struct ncr53c9x_softc *nsc = sc->sc_client; |
458 int trans, resid; | 475 bus_dma_tag_t dmat; 476 bus_dmamap_t dmam; 477 size_t dmasize; 478 int lxfer, resid, trans; |
459 uint32_t csr; 460 461 csr = L64854_GCSR(sc); 462 463 DPRINTF(LDB_SCSI, ("%s: addr 0x%x, csr %b\n", __func__, | 479 uint32_t csr; 480 481 csr = L64854_GCSR(sc); 482 483 DPRINTF(LDB_SCSI, ("%s: addr 0x%x, csr %b\n", __func__, |
464 bus_read_4(sc->sc_res, L64854_REG_ADDR), csr, DDMACSR_BITS)); | 484 bus_read_4(sc->sc_res, L64854_REG_ADDR), csr, DDMACSR_BITS)); |
465 | 485 |
466 if (csr & (D_ERR_PEND|D_SLAVE_ERR)) { 467 device_printf(sc->sc_dev, "error: csr=%b\n", csr, DDMACSR_BITS); 468 csr &= ~D_EN_DMA; /* Stop DMA */ | 486 if (csr & (D_ERR_PEND | D_SLAVE_ERR)) { 487 device_printf(sc->sc_dev, "error: csr=%b\n", csr, 488 DDMACSR_BITS); 489 csr &= ~D_EN_DMA; /* Stop DMA. */ |
469 /* Invalidate the queue; SLAVE_ERR bit is write-to-clear */ | 490 /* Invalidate the queue; SLAVE_ERR bit is write-to-clear */ |
470 csr |= D_INVALIDATE|D_SLAVE_ERR; | 491 csr |= D_INVALIDATE | D_SLAVE_ERR; |
471 L64854_SCSR(sc, csr); 472 return (-1); 473 } 474 475 /* This is an "assertion" :) */ 476 if (sc->sc_active == 0) 477 panic("%s: DMA wasn't active", __func__); 478 479 DMA_DRAIN(sc, 0); 480 481 /* DMA has stopped */ 482 csr &= ~D_EN_DMA; 483 L64854_SCSR(sc, csr); 484 sc->sc_active = 0; 485 | 492 L64854_SCSR(sc, csr); 493 return (-1); 494 } 495 496 /* This is an "assertion" :) */ 497 if (sc->sc_active == 0) 498 panic("%s: DMA wasn't active", __func__); 499 500 DMA_DRAIN(sc, 0); 501 502 /* DMA has stopped */ 503 csr &= ~D_EN_DMA; 504 L64854_SCSR(sc, csr); 505 sc->sc_active = 0; 506 |
486 if (sc->sc_dmasize == 0) { 487 /* A "Transfer Pad" operation completed */ 488 DPRINTF(LDB_SCSI, ("%s: discarded %d bytes (tcl=%d, tcm=%d)\n", 489 __func__, NCR_READ_REG(nsc, NCR_TCL) | | 507 dmasize = sc->sc_dmasize; 508 if (dmasize == 0) { 509 /* A "Transfer Pad" operation completed. */ 510 DPRINTF(LDB_SCSI, ("%s: discarded %d bytes (tcl=%d, " 511 "tcm=%d)\n", __func__, NCR_READ_REG(nsc, NCR_TCL) | |
490 (NCR_READ_REG(nsc, NCR_TCM) << 8), 491 NCR_READ_REG(nsc, NCR_TCL), NCR_READ_REG(nsc, NCR_TCM))); 492 return (0); 493 } 494 495 resid = 0; 496 /* 497 * If a transfer onto the SCSI bus gets interrupted by the device 498 * (e.g. for a SAVEPOINTER message), the data in the FIFO counts 499 * as residual since the NCR53C9X counter registers get decremented 500 * as bytes are clocked into the FIFO. 501 */ | 512 (NCR_READ_REG(nsc, NCR_TCM) << 8), 513 NCR_READ_REG(nsc, NCR_TCL), NCR_READ_REG(nsc, NCR_TCM))); 514 return (0); 515 } 516 517 resid = 0; 518 /* 519 * If a transfer onto the SCSI bus gets interrupted by the device 520 * (e.g. for a SAVEPOINTER message), the data in the FIFO counts 521 * as residual since the NCR53C9X counter registers get decremented 522 * as bytes are clocked into the FIFO. 523 */ |
502 if (!(csr & D_WRITE) && | 524 if ((csr & D_WRITE) == 0 && |
503 (resid = (NCR_READ_REG(nsc, NCR_FFLAG) & NCRFIFO_FF)) != 0) { 504 DPRINTF(LDB_SCSI, ("%s: empty esp FIFO of %d ", __func__, 505 resid)); 506 if (nsc->sc_rev == NCR_VARIANT_FAS366 && 507 (NCR_READ_REG(nsc, NCR_CFG3) & NCRFASCFG3_EWIDE)) 508 resid <<= 1; 509 } 510 511 if ((nsc->sc_espstat & NCRSTAT_TC) == 0) { | 525 (resid = (NCR_READ_REG(nsc, NCR_FFLAG) & NCRFIFO_FF)) != 0) { 526 DPRINTF(LDB_SCSI, ("%s: empty esp FIFO of %d ", __func__, 527 resid)); 528 if (nsc->sc_rev == NCR_VARIANT_FAS366 && 529 (NCR_READ_REG(nsc, NCR_CFG3) & NCRFASCFG3_EWIDE)) 530 resid <<= 1; 531 } 532 533 if ((nsc->sc_espstat & NCRSTAT_TC) == 0) { |
534 lxfer = nsc->sc_features & NCR_F_LARGEXFER; |
|
512 /* | 535 /* |
513 * `Terminal count' is off, so read the residue | 536 * "Terminal count" is off, so read the residue |
514 * out of the NCR53C9X counter registers. 515 */ 516 resid += (NCR_READ_REG(nsc, NCR_TCL) | 517 (NCR_READ_REG(nsc, NCR_TCM) << 8) | | 537 * out of the NCR53C9X counter registers. 538 */ 539 resid += (NCR_READ_REG(nsc, NCR_TCL) | 540 (NCR_READ_REG(nsc, NCR_TCM) << 8) | |
518 ((nsc->sc_cfg2 & NCRCFG2_FE) ? 519 (NCR_READ_REG(nsc, NCR_TCH) << 16) : 0)); | 541 (lxfer != 0 ? (NCR_READ_REG(nsc, NCR_TCH) << 16) : 0)); |
520 | 542 |
521 if (resid == 0 && sc->sc_dmasize == 65536 && 522 (nsc->sc_cfg2 & NCRCFG2_FE) == 0) 523 /* A transfer of 64K is encoded as `TCL=TCM=0' */ | 543 if (resid == 0 && dmasize == 65536 && lxfer == 0) 544 /* A transfer of 64k is encoded as TCL=TCM=0. */ |
524 resid = 65536; 525 } 526 | 545 resid = 65536; 546 } 547 |
527 trans = sc->sc_dmasize - resid; | 548 trans = dmasize - resid; |
528 if (trans < 0) { /* transferred < 0? */ 529#if 0 530 /* 531 * This situation can happen in perfectly normal operation 532 * if the ESP is reselected while using DMA to select 533 * another target. As such, don't print the warning. 534 */ 535 device_printf(sc->sc_dev, "xfer (%d) > req (%d)\n", trans, | 549 if (trans < 0) { /* transferred < 0? */ 550#if 0 551 /* 552 * This situation can happen in perfectly normal operation 553 * if the ESP is reselected while using DMA to select 554 * another target. As such, don't print the warning. 555 */ 556 device_printf(sc->sc_dev, "xfer (%d) > req (%d)\n", trans, |
536 sc->sc_dmasize); | 557 dmasize); |
537#endif | 558#endif |
538 trans = sc->sc_dmasize; | 559 trans = dmasize; |
539 } 540 541 DPRINTF(LDB_SCSI, ("%s: tcl=%d, tcm=%d, tch=%d; trans=%d, resid=%d\n", 542 __func__, NCR_READ_REG(nsc, NCR_TCL), NCR_READ_REG(nsc, NCR_TCM), | 560 } 561 562 DPRINTF(LDB_SCSI, ("%s: tcl=%d, tcm=%d, tch=%d; trans=%d, resid=%d\n", 563 __func__, NCR_READ_REG(nsc, NCR_TCL), NCR_READ_REG(nsc, NCR_TCM), |
543 (nsc->sc_cfg2 & NCRCFG2_FE) ? NCR_READ_REG(nsc, NCR_TCH) : 0, 544 trans, resid)); | 564 (nsc->sc_sc_features & NCR_F_LARGEXFER) != 0 ? 565 NCR_READ_REG(nsc, NCR_TCH) : 0, trans, resid)); |
545 | 566 |
546 if (sc->sc_dmasize != 0) { 547 bus_dmamap_sync(sc->sc_buffer_dmat, sc->sc_dmamap, 548 (csr & D_WRITE) != 0 ? BUS_DMASYNC_POSTREAD : 549 BUS_DMASYNC_POSTWRITE); 550 bus_dmamap_unload(sc->sc_buffer_dmat, sc->sc_dmamap); | 567 if (dmasize != 0) { 568 dmat = sc->sc_buffer_dmat; 569 dmam = sc->sc_dmamap; 570 bus_dmamap_sync(dmat, dmam, (csr & D_WRITE) != 0 ? 571 BUS_DMASYNC_POSTREAD : BUS_DMASYNC_POSTWRITE); 572 bus_dmamap_unload(dmat, dmam); |
551 } 552 553 *sc->sc_dmalen -= trans; 554 *sc->sc_dmaaddr = (char *)*sc->sc_dmaaddr + trans; 555 556#if 0 /* this is not normal operation just yet */ 557 if (*sc->sc_dmalen == 0 || nsc->sc_phase != nsc->sc_prevphase) 558 return (0); 559 560 /* and again */ 561 dma_start(sc, sc->sc_dmaaddr, sc->sc_dmalen, DMACSR(sc) & D_WRITE); 562 return (1); 563#endif 564 return (0); 565} 566 567/* | 573 } 574 575 *sc->sc_dmalen -= trans; 576 *sc->sc_dmaaddr = (char *)*sc->sc_dmaaddr + trans; 577 578#if 0 /* this is not normal operation just yet */ 579 if (*sc->sc_dmalen == 0 || nsc->sc_phase != nsc->sc_prevphase) 580 return (0); 581 582 /* and again */ 583 dma_start(sc, sc->sc_dmaaddr, sc->sc_dmalen, DMACSR(sc) & D_WRITE); 584 return (1); 585#endif 586 return (0); 587} 588 589/* |
568 * Pseudo (chained) interrupt to le driver to handle DMA errors. | 590 * Pseudo (chained) interrupt to le(4) driver to handle DMA errors |
569 */ 570static int 571lsi64854_enet_intr(void *arg) 572{ 573 struct lsi64854_softc *sc = arg; 574 uint32_t csr; 575 int i, rv; 576 577 csr = L64854_GCSR(sc); 578 579 /* If the DMA logic shows an interrupt, claim it */ 580 rv = ((csr & E_INT_PEND) != 0) ? 1 : 0; 581 | 591 */ 592static int 593lsi64854_enet_intr(void *arg) 594{ 595 struct lsi64854_softc *sc = arg; 596 uint32_t csr; 597 int i, rv; 598 599 csr = L64854_GCSR(sc); 600 601 /* If the DMA logic shows an interrupt, claim it */ 602 rv = ((csr & E_INT_PEND) != 0) ? 1 : 0; 603 |
582 if (csr & (E_ERR_PEND|E_SLAVE_ERR)) { 583 device_printf(sc->sc_dev, "error: csr=%b\n", csr, EDMACSR_BITS); 584 csr &= ~L64854_EN_DMA; /* Stop DMA */ | 604 if (csr & (E_ERR_PEND | E_SLAVE_ERR)) { 605 device_printf(sc->sc_dev, "error: csr=%b\n", csr, 606 EDMACSR_BITS); 607 csr &= ~L64854_EN_DMA; /* Stop DMA. */ |
585 /* Invalidate the queue; SLAVE_ERR bit is write-to-clear */ | 608 /* Invalidate the queue; SLAVE_ERR bit is write-to-clear */ |
586 csr |= E_INVALIDATE|E_SLAVE_ERR; | 609 csr |= E_INVALIDATE | E_SLAVE_ERR; |
587 L64854_SCSR(sc, csr); 588 /* Will be drained with the LE_C0_IDON interrupt. */ 589 sc->sc_dodrain = 1; 590 return (-1); 591 } 592 593 /* XXX - is this necessary with E_DSBL_WR_INVAL on? */ 594 if (sc->sc_dodrain) { --- 10 unchanged lines hidden (view full) --- 605 606static void 607lsi64854_map_pp(void *arg, bus_dma_segment_t *segs, int nsegs, int error) 608{ 609 struct lsi64854_softc *sc; 610 611 sc = (struct lsi64854_softc *)arg; 612 | 610 L64854_SCSR(sc, csr); 611 /* Will be drained with the LE_C0_IDON interrupt. */ 612 sc->sc_dodrain = 1; 613 return (-1); 614 } 615 616 /* XXX - is this necessary with E_DSBL_WR_INVAL on? */ 617 if (sc->sc_dodrain) { --- 10 unchanged lines hidden (view full) --- 628 629static void 630lsi64854_map_pp(void *arg, bus_dma_segment_t *segs, int nsegs, int error) 631{ 632 struct lsi64854_softc *sc; 633 634 sc = (struct lsi64854_softc *)arg; 635 |
636 if (error != 0) 637 return; |
|
613 if (nsegs != 1) 614 panic("%s: cannot map %d segments\n", __func__, nsegs); 615 | 638 if (nsegs != 1) 639 panic("%s: cannot map %d segments\n", __func__, nsegs); 640 |
616 bus_dmamap_sync(sc->sc_buffer_dmat, sc->sc_dmamap, sc->sc_datain ? 617 BUS_DMASYNC_PREREAD : BUS_DMASYNC_PREWRITE); | 641 bus_dmamap_sync(sc->sc_buffer_dmat, sc->sc_dmamap, 642 sc->sc_datain != 0 ? BUS_DMASYNC_PREREAD : BUS_DMASYNC_PREWRITE); |
618 bus_write_4(sc->sc_res, L64854_REG_ADDR, segs[0].ds_addr); 619 620 bus_write_4(sc->sc_res, L64854_REG_CNT, sc->sc_dmasize); 621} 622 623/* | 643 bus_write_4(sc->sc_res, L64854_REG_ADDR, segs[0].ds_addr); 644 645 bus_write_4(sc->sc_res, L64854_REG_CNT, sc->sc_dmasize); 646} 647 648/* |
624 * setup a DMA transfer | 649 * Setup a DMA transfer. |
625 */ 626static int 627lsi64854_setup_pp(struct lsi64854_softc *sc, void **addr, size_t *len, 628 int datain, size_t *dmasize) 629{ | 650 */ 651static int 652lsi64854_setup_pp(struct lsi64854_softc *sc, void **addr, size_t *len, 653 int datain, size_t *dmasize) 654{ |
655 int error; |
|
630 uint32_t csr; 631 632 DMA_FLUSH(sc, 0); 633 634 sc->sc_dmaaddr = addr; 635 sc->sc_dmalen = len; 636 sc->sc_datain = datain; 637 638 DPRINTF(LDB_PP, ("%s: pp start %ld@%p,%d\n", __func__, | 656 uint32_t csr; 657 658 DMA_FLUSH(sc, 0); 659 660 sc->sc_dmaaddr = addr; 661 sc->sc_dmalen = len; 662 sc->sc_datain = datain; 663 664 DPRINTF(LDB_PP, ("%s: pp start %ld@%p,%d\n", __func__, |
639 (long)*sc->sc_dmalen, *sc->sc_dmaaddr, datain ? 1 : 0)); | 665 (long)*sc->sc_dmalen, *sc->sc_dmaaddr, datain != 0 ? 1 : 0)); |
640 | 666 |
641 /* 642 * the rules say we cannot transfer more than the limit 643 * of this DMA chip (64k for old and 16Mb for new), 644 * and we cannot cross a 16Mb boundary. 645 */ 646 *dmasize = sc->sc_dmasize = 647 ulmin(*dmasize, DMAMAX((size_t)*sc->sc_dmaaddr)); | 667 KASSERT(*dmasize <= sc->sc_maxdmasize, 668 ("%s: transfer size %ld too large", __func__, (long)*dmasize)); |
648 | 669 |
649 DPRINTF(LDB_PP, ("%s: dmasize=%ld\n", __func__, (long)sc->sc_dmasize)); | 670 sc->sc_dmasize = *dmasize; |
650 | 671 |
651 /* Program the DMA address */ 652 if (sc->sc_dmasize != 0) 653 if (bus_dmamap_load(sc->sc_buffer_dmat, sc->sc_dmamap, 654 *sc->sc_dmaaddr, sc->sc_dmasize, lsi64854_map_pp, sc, 0)) 655 panic("%s: pp cannot allocate DVMA address", __func__); | 672 DPRINTF(LDB_PP, ("%s: dmasize=%ld\n", __func__, (long)*dmasize)); |
656 | 673 |
657 /* Setup DMA control register */ | 674 /* Load the transfer buffer and program the DMA address. */ 675 if (*dmasize != 0) { 676 error = bus_dmamap_load(sc->sc_buffer_dmat, sc->sc_dmamap, 677 *sc->sc_dmaaddr, *dmasize, lsi64854_map_pp, sc, 678 BUS_DMA_NOWAIT); 679 if (error != 0) 680 return (error); 681 } 682 683 /* Setup the DMA control register. */ |
658 csr = L64854_GCSR(sc); 659 csr &= ~L64854_BURST_SIZE; 660 if (sc->sc_burst == 32) 661 csr |= L64854_BURST_32; 662 else if (sc->sc_burst == 16) 663 csr |= L64854_BURST_16; 664 else 665 csr |= L64854_BURST_0; | 684 csr = L64854_GCSR(sc); 685 csr &= ~L64854_BURST_SIZE; 686 if (sc->sc_burst == 32) 687 csr |= L64854_BURST_32; 688 else if (sc->sc_burst == 16) 689 csr |= L64854_BURST_16; 690 else 691 csr |= L64854_BURST_0; |
666 csr |= P_EN_DMA|P_INT_EN|P_EN_CNT; | 692 csr |= P_EN_DMA | P_INT_EN | P_EN_CNT; |
667#if 0 | 693#if 0 |
668 /* This bit is read-only in PP csr register */ 669 if (datain) | 694 /* This bit is read-only in PP csr register. */ 695 if (datain != 0) |
670 csr |= P_WRITE; 671 else 672 csr &= ~P_WRITE; 673#endif 674 L64854_SCSR(sc, csr); 675 676 return (0); 677} 678 679/* | 696 csr |= P_WRITE; 697 else 698 csr &= ~P_WRITE; 699#endif 700 L64854_SCSR(sc, csr); 701 702 return (0); 703} 704 705/* |
680 * Parallel port DMA interrupt. | 706 * Parallel port DMA interrupt |
681 */ 682static int 683lsi64854_pp_intr(void *arg) 684{ 685 struct lsi64854_softc *sc = arg; | 707 */ 708static int 709lsi64854_pp_intr(void *arg) 710{ 711 struct lsi64854_softc *sc = arg; |
712 bus_dma_tag_t dmat; 713 bus_dmamap_t dmam; 714 size_t dmasize; |
|
686 int ret, trans, resid = 0; 687 uint32_t csr; 688 689 csr = L64854_GCSR(sc); 690 691 DPRINTF(LDB_PP, ("%s: addr 0x%x, csr %b\n", __func__, 692 bus_read_4(sc->sc_res, L64854_REG_ADDR), csr, PDMACSR_BITS)); 693 | 715 int ret, trans, resid = 0; 716 uint32_t csr; 717 718 csr = L64854_GCSR(sc); 719 720 DPRINTF(LDB_PP, ("%s: addr 0x%x, csr %b\n", __func__, 721 bus_read_4(sc->sc_res, L64854_REG_ADDR), csr, PDMACSR_BITS)); 722 |
694 if (csr & (P_ERR_PEND|P_SLAVE_ERR)) { | 723 if ((csr & (P_ERR_PEND | P_SLAVE_ERR)) != 0) { |
695 resid = bus_read_4(sc->sc_res, L64854_REG_CNT); 696 device_printf(sc->sc_dev, "error: resid %d csr=%b\n", resid, 697 csr, PDMACSR_BITS); | 724 resid = bus_read_4(sc->sc_res, L64854_REG_CNT); 725 device_printf(sc->sc_dev, "error: resid %d csr=%b\n", resid, 726 csr, PDMACSR_BITS); |
698 csr &= ~P_EN_DMA; /* Stop DMA */ | 727 csr &= ~P_EN_DMA; /* Stop DMA. */ |
699 /* Invalidate the queue; SLAVE_ERR bit is write-to-clear */ | 728 /* Invalidate the queue; SLAVE_ERR bit is write-to-clear */ |
700 csr |= P_INVALIDATE|P_SLAVE_ERR; | 729 csr |= P_INVALIDATE | P_SLAVE_ERR; |
701 L64854_SCSR(sc, csr); 702 return (-1); 703 } 704 705 ret = (csr & P_INT_PEND) != 0; 706 707 if (sc->sc_active != 0) { 708 DMA_DRAIN(sc, 0); 709 resid = bus_read_4(sc->sc_res, L64854_REG_CNT); 710 } 711 712 /* DMA has stopped */ 713 csr &= ~D_EN_DMA; 714 L64854_SCSR(sc, csr); 715 sc->sc_active = 0; 716 | 730 L64854_SCSR(sc, csr); 731 return (-1); 732 } 733 734 ret = (csr & P_INT_PEND) != 0; 735 736 if (sc->sc_active != 0) { 737 DMA_DRAIN(sc, 0); 738 resid = bus_read_4(sc->sc_res, L64854_REG_CNT); 739 } 740 741 /* DMA has stopped */ 742 csr &= ~D_EN_DMA; 743 L64854_SCSR(sc, csr); 744 sc->sc_active = 0; 745 |
717 trans = sc->sc_dmasize - resid; | 746 dmasize = sc->sc_dmasize; 747 trans = dmasize - resid; |
718 if (trans < 0) /* transferred < 0? */ | 748 if (trans < 0) /* transferred < 0? */ |
719 trans = sc->sc_dmasize; | 749 trans = dmasize; |
720 *sc->sc_dmalen -= trans; 721 *sc->sc_dmaaddr = (char *)*sc->sc_dmaaddr + trans; 722 | 750 *sc->sc_dmalen -= trans; 751 *sc->sc_dmaaddr = (char *)*sc->sc_dmaaddr + trans; 752 |
723 if (sc->sc_dmasize != 0) { 724 bus_dmamap_sync(sc->sc_buffer_dmat, sc->sc_dmamap, 725 (csr & D_WRITE) != 0 ? BUS_DMASYNC_POSTREAD : 726 BUS_DMASYNC_POSTWRITE); 727 bus_dmamap_unload(sc->sc_buffer_dmat, sc->sc_dmamap); | 753 if (dmasize != 0) { 754 dmat = sc->sc_buffer_dmat; 755 dmam = sc->sc_dmamap; 756 bus_dmamap_sync(dmat, dmam, (csr & D_WRITE) != 0 ? 757 BUS_DMASYNC_POSTREAD : BUS_DMASYNC_POSTWRITE); 758 bus_dmamap_unload(dmat, dmam); |
728 } 729 730 return (ret != 0); 731} | 759 } 760 761 return (ret != 0); 762} |