1/* -*- mode: c; c-basic-offset: 8 -*- */ 2 3/* NCR (or Symbios) 53c700 and 53c700-66 Driver 4 * 5 * Copyright (C) 2001 by James.Bottomley@HansenPartnership.com 6**----------------------------------------------------------------------------- 7** 8** This program is free software; you can redistribute it and/or modify 9** it under the terms of the GNU General Public License as published by 10** the Free Software Foundation; either version 2 of the License, or 11** (at your option) any later version. 12** 13** This program is distributed in the hope that it will be useful, 14** but WITHOUT ANY WARRANTY; without even the implied warranty of 15** MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the 16** GNU General Public License for more details. 17** 18** You should have received a copy of the GNU General Public License 19** along with this program; if not, write to the Free Software 20** Foundation, Inc., 675 Mass Ave, Cambridge, MA 02139, USA. 21** 22**----------------------------------------------------------------------------- 23 */ 24 25/* Notes: 26 * 27 * This driver is designed exclusively for these chips (virtually the 28 * earliest of the scripts engine chips). They need their own drivers 29 * because they are missing so many of the scripts and snazzy register 30 * features of their elder brothers (the 710, 720 and 770). 31 * 32 * The 700 is the lowliest of the line, it can only do async SCSI. 33 * The 700-66 can at least do synchronous SCSI up to 10MHz. 34 * 35 * The 700 chip has no host bus interface logic of its own. However, 36 * it is usually mapped to a location with well defined register 37 * offsets. Therefore, if you can determine the base address and the 38 * irq your board incorporating this chip uses, you can probably use 39 * this driver to run it (although you'll probably have to write a 40 * minimal wrapper for the purpose---see the NCR_D700 driver for 41 * details about how to do this). 42 * 43 * 44 * TODO List: 45 * 46 * 1. Better statistics in the proc fs 47 * 48 * 2. Implement message queue (queues SCSI messages like commands) and make 49 * the abort and device reset functions use them. 50 * */ 51 52/* CHANGELOG 53 * 54 * Version 2.8 55 * 56 * Fixed bad bug affecting tag starvation processing (previously the 57 * driver would hang the system if too many tags starved. Also fixed 58 * bad bug having to do with 10 byte command processing and REQUEST 59 * SENSE (the command would loop forever getting a transfer length 60 * mismatch in the CMD phase). 61 * 62 * Version 2.7 63 * 64 * Fixed scripts problem which caused certain devices (notably CDRWs) 65 * to hang on initial INQUIRY. Updated NCR_700_readl/writel to use 66 * __raw_readl/writel for parisc compatibility (Thomas 67 * Bogendoerfer). Added missing SCp->request_bufflen initialisation 68 * for sense requests (Ryan Bradetich). 69 * 70 * Version 2.6 71 * 72 * Following test of the 64 bit parisc kernel by Richard Hirst, 73 * several problems have now been corrected. Also adds support for 74 * consistent memory allocation. 75 * 76 * Version 2.5 77 * 78 * More Compatibility changes for 710 (now actually works). Enhanced 79 * support for odd clock speeds which constrain SDTR negotiations. 80 * correct cacheline separation for scsi messages and status for 81 * incoherent architectures. Use of the pci mapping functions on 82 * buffers to begin support for 64 bit drivers. 83 * 84 * Version 2.4 85 * 86 * Added support for the 53c710 chip (in 53c700 emulation mode only---no 87 * special 53c710 instructions or registers are used). 88 * 89 * Version 2.3 90 * 91 * More endianness/cache coherency changes. 92 * 93 * Better bad device handling (handles devices lying about tag 94 * queueing support and devices which fail to provide sense data on 95 * contingent allegiance conditions) 96 * 97 * Many thanks to Richard Hirst <rhirst@linuxcare.com> for patiently 98 * debugging this driver on the parisc architecture and suggesting 99 * many improvements and bug fixes. 100 * 101 * Thanks also go to Linuxcare Inc. for providing several PARISC 102 * machines for me to debug the driver on. 103 * 104 * Version 2.2 105 * 106 * Made the driver mem or io mapped; added endian invariance; added 107 * dma cache flushing operations for architectures which need it; 108 * added support for more varied clocking speeds. 109 * 110 * Version 2.1 111 * 112 * Initial modularisation from the D700. See NCR_D700.c for the rest of 113 * the changelog. 114 * */ 115#define NCR_700_VERSION "2.8" 116 117#include <linux/kernel.h> 118#include <linux/types.h> 119#include <linux/string.h> 120#include <linux/slab.h> 121#include <linux/ioport.h> 122#include <linux/delay.h> 123#include <linux/spinlock.h> 124#include <linux/completion.h> 125#include <linux/init.h> 126#include <linux/proc_fs.h> 127#include <linux/blkdev.h> 128#include <linux/module.h> 129#include <linux/interrupt.h> 130#include <linux/device.h> 131#include <asm/dma.h> 132#include <asm/system.h> 133#include <asm/io.h> 134#include <asm/pgtable.h> 135#include <asm/byteorder.h> 136 137#include <scsi/scsi.h> 138#include <scsi/scsi_cmnd.h> 139#include <scsi/scsi_dbg.h> 140#include <scsi/scsi_eh.h> 141#include <scsi/scsi_host.h> 142#include <scsi/scsi_tcq.h> 143#include <scsi/scsi_transport.h> 144#include <scsi/scsi_transport_spi.h> 145 146#include "53c700.h" 147 148/* NOTE: For 64 bit drivers there are points in the code where we use 149 * a non dereferenceable pointer to point to a structure in dma-able 150 * memory (which is 32 bits) so that we can use all of the structure 151 * operations but take the address at the end. This macro allows us 152 * to truncate the 64 bit pointer down to 32 bits without the compiler 153 * complaining */ 154#define to32bit(x) ((__u32)((unsigned long)(x))) 155 156#ifdef NCR_700_DEBUG 157#define STATIC 158#else 159#define STATIC static 160#endif 161 162MODULE_AUTHOR("James Bottomley"); 163MODULE_DESCRIPTION("53c700 and 53c700-66 Driver"); 164MODULE_LICENSE("GPL"); 165 166/* This is the script */ 167#include "53c700_d.h" 168 169 170STATIC int NCR_700_queuecommand(struct scsi_cmnd *, void (*done)(struct scsi_cmnd *)); 171STATIC int NCR_700_abort(struct scsi_cmnd * SCpnt); 172STATIC int NCR_700_bus_reset(struct scsi_cmnd * SCpnt); 173STATIC int NCR_700_host_reset(struct scsi_cmnd * SCpnt); 174STATIC void NCR_700_chip_setup(struct Scsi_Host *host); 175STATIC void NCR_700_chip_reset(struct Scsi_Host *host); 176STATIC int NCR_700_slave_alloc(struct scsi_device *SDpnt); 177STATIC int NCR_700_slave_configure(struct scsi_device *SDpnt); 178STATIC void NCR_700_slave_destroy(struct scsi_device *SDpnt); 179static int NCR_700_change_queue_depth(struct scsi_device *SDpnt, int depth, int reason); 180static int NCR_700_change_queue_type(struct scsi_device *SDpnt, int depth); 181 182STATIC struct device_attribute *NCR_700_dev_attrs[]; 183 184STATIC struct scsi_transport_template *NCR_700_transport_template = NULL; 185 186static char *NCR_700_phase[] = { 187 "", 188 "after selection", 189 "before command phase", 190 "after command phase", 191 "after status phase", 192 "after data in phase", 193 "after data out phase", 194 "during data phase", 195}; 196 197static char *NCR_700_condition[] = { 198 "", 199 "NOT MSG_OUT", 200 "UNEXPECTED PHASE", 201 "NOT MSG_IN", 202 "UNEXPECTED MSG", 203 "MSG_IN", 204 "SDTR_MSG RECEIVED", 205 "REJECT_MSG RECEIVED", 206 "DISCONNECT_MSG RECEIVED", 207 "MSG_OUT", 208 "DATA_IN", 209 210}; 211 212static char *NCR_700_fatal_messages[] = { 213 "unexpected message after reselection", 214 "still MSG_OUT after message injection", 215 "not MSG_IN after selection", 216 "Illegal message length received", 217}; 218 219static char *NCR_700_SBCL_bits[] = { 220 "IO ", 221 "CD ", 222 "MSG ", 223 "ATN ", 224 "SEL ", 225 "BSY ", 226 "ACK ", 227 "REQ ", 228}; 229 230static char *NCR_700_SBCL_to_phase[] = { 231 "DATA_OUT", 232 "DATA_IN", 233 "CMD_OUT", 234 "STATE", 235 "ILLEGAL PHASE", 236 "ILLEGAL PHASE", 237 "MSG OUT", 238 "MSG IN", 239}; 240 241/* This translates the SDTR message offset and period to a value 242 * which can be loaded into the SXFER_REG. 243 * 244 * NOTE: According to SCSI-2, the true transfer period (in ns) is 245 * actually four times this period value */ 246static inline __u8 247NCR_700_offset_period_to_sxfer(struct NCR_700_Host_Parameters *hostdata, 248 __u8 offset, __u8 period) 249{ 250 int XFERP; 251 252 __u8 min_xferp = (hostdata->chip710 253 ? NCR_710_MIN_XFERP : NCR_700_MIN_XFERP); 254 __u8 max_offset = (hostdata->chip710 255 ? NCR_710_MAX_OFFSET : NCR_700_MAX_OFFSET); 256 257 if(offset == 0) 258 return 0; 259 260 if(period < hostdata->min_period) { 261 printk(KERN_WARNING "53c700: Period %dns is less than this chip's minimum, setting to %d\n", period*4, NCR_700_MIN_PERIOD*4); 262 period = hostdata->min_period; 263 } 264 XFERP = (period*4 * hostdata->sync_clock)/1000 - 4; 265 if(offset > max_offset) { 266 printk(KERN_WARNING "53c700: Offset %d exceeds chip maximum, setting to %d\n", 267 offset, max_offset); 268 offset = max_offset; 269 } 270 if(XFERP < min_xferp) { 271 XFERP = min_xferp; 272 } 273 return (offset & 0x0f) | (XFERP & 0x07)<<4; 274} 275 276static inline __u8 277NCR_700_get_SXFER(struct scsi_device *SDp) 278{ 279 struct NCR_700_Host_Parameters *hostdata = 280 (struct NCR_700_Host_Parameters *)SDp->host->hostdata[0]; 281 282 return NCR_700_offset_period_to_sxfer(hostdata, 283 spi_offset(SDp->sdev_target), 284 spi_period(SDp->sdev_target)); 285} 286 287struct Scsi_Host * 288NCR_700_detect(struct scsi_host_template *tpnt, 289 struct NCR_700_Host_Parameters *hostdata, struct device *dev) 290{ 291 dma_addr_t pScript, pSlots; 292 __u8 *memory; 293 __u32 *script; 294 struct Scsi_Host *host; 295 static int banner = 0; 296 int j; 297 298 if(tpnt->sdev_attrs == NULL) 299 tpnt->sdev_attrs = NCR_700_dev_attrs; 300 301 memory = dma_alloc_noncoherent(hostdata->dev, TOTAL_MEM_SIZE, 302 &pScript, GFP_KERNEL); 303 if(memory == NULL) { 304 printk(KERN_ERR "53c700: Failed to allocate memory for driver, detatching\n"); 305 return NULL; 306 } 307 308 script = (__u32 *)memory; 309 hostdata->msgin = memory + MSGIN_OFFSET; 310 hostdata->msgout = memory + MSGOUT_OFFSET; 311 hostdata->status = memory + STATUS_OFFSET; 312 hostdata->slots = (struct NCR_700_command_slot *)(memory + SLOTS_OFFSET); 313 hostdata->dev = dev; 314 315 pSlots = pScript + SLOTS_OFFSET; 316 317 /* Fill in the missing routines from the host template */ 318 tpnt->queuecommand = NCR_700_queuecommand; 319 tpnt->eh_abort_handler = NCR_700_abort; 320 tpnt->eh_bus_reset_handler = NCR_700_bus_reset; 321 tpnt->eh_host_reset_handler = NCR_700_host_reset; 322 tpnt->can_queue = NCR_700_COMMAND_SLOTS_PER_HOST; 323 tpnt->sg_tablesize = NCR_700_SG_SEGMENTS; 324 tpnt->cmd_per_lun = NCR_700_CMD_PER_LUN; 325 tpnt->use_clustering = ENABLE_CLUSTERING; 326 tpnt->slave_configure = NCR_700_slave_configure; 327 tpnt->slave_destroy = NCR_700_slave_destroy; 328 tpnt->slave_alloc = NCR_700_slave_alloc; 329 tpnt->change_queue_depth = NCR_700_change_queue_depth; 330 tpnt->change_queue_type = NCR_700_change_queue_type; 331 332 if(tpnt->name == NULL) 333 tpnt->name = "53c700"; 334 if(tpnt->proc_name == NULL) 335 tpnt->proc_name = "53c700"; 336 337 host = scsi_host_alloc(tpnt, 4); 338 if (!host) 339 return NULL; 340 memset(hostdata->slots, 0, sizeof(struct NCR_700_command_slot) 341 * NCR_700_COMMAND_SLOTS_PER_HOST); 342 for (j = 0; j < NCR_700_COMMAND_SLOTS_PER_HOST; j++) { 343 dma_addr_t offset = (dma_addr_t)((unsigned long)&hostdata->slots[j].SG[0] 344 - (unsigned long)&hostdata->slots[0].SG[0]); 345 hostdata->slots[j].pSG = (struct NCR_700_SG_List *)((unsigned long)(pSlots + offset)); 346 if(j == 0) 347 hostdata->free_list = &hostdata->slots[j]; 348 else 349 hostdata->slots[j-1].ITL_forw = &hostdata->slots[j]; 350 hostdata->slots[j].state = NCR_700_SLOT_FREE; 351 } 352 353 for (j = 0; j < ARRAY_SIZE(SCRIPT); j++) 354 script[j] = bS_to_host(SCRIPT[j]); 355 356 /* adjust all labels to be bus physical */ 357 for (j = 0; j < PATCHES; j++) 358 script[LABELPATCHES[j]] = bS_to_host(pScript + SCRIPT[LABELPATCHES[j]]); 359 /* now patch up fixed addresses. */ 360 script_patch_32(hostdata->dev, script, MessageLocation, 361 pScript + MSGOUT_OFFSET); 362 script_patch_32(hostdata->dev, script, StatusAddress, 363 pScript + STATUS_OFFSET); 364 script_patch_32(hostdata->dev, script, ReceiveMsgAddress, 365 pScript + MSGIN_OFFSET); 366 367 hostdata->script = script; 368 hostdata->pScript = pScript; 369 dma_sync_single_for_device(hostdata->dev, pScript, sizeof(SCRIPT), DMA_TO_DEVICE); 370 hostdata->state = NCR_700_HOST_FREE; 371 hostdata->cmd = NULL; 372 host->max_id = 8; 373 host->max_lun = NCR_700_MAX_LUNS; 374 BUG_ON(NCR_700_transport_template == NULL); 375 host->transportt = NCR_700_transport_template; 376 host->unique_id = (unsigned long)hostdata->base; 377 hostdata->eh_complete = NULL; 378 host->hostdata[0] = (unsigned long)hostdata; 379 /* kick the chip */ 380 NCR_700_writeb(0xff, host, CTEST9_REG); 381 if (hostdata->chip710) 382 hostdata->rev = (NCR_700_readb(host, CTEST8_REG)>>4) & 0x0f; 383 else 384 hostdata->rev = (NCR_700_readb(host, CTEST7_REG)>>4) & 0x0f; 385 hostdata->fast = (NCR_700_readb(host, CTEST9_REG) == 0); 386 if (banner == 0) { 387 printk(KERN_NOTICE "53c700: Version " NCR_700_VERSION " By James.Bottomley@HansenPartnership.com\n"); 388 banner = 1; 389 } 390 printk(KERN_NOTICE "scsi%d: %s rev %d %s\n", host->host_no, 391 hostdata->chip710 ? "53c710" : 392 (hostdata->fast ? "53c700-66" : "53c700"), 393 hostdata->rev, hostdata->differential ? 394 "(Differential)" : ""); 395 /* reset the chip */ 396 NCR_700_chip_reset(host); 397 398 if (scsi_add_host(host, dev)) { 399 dev_printk(KERN_ERR, dev, "53c700: scsi_add_host failed\n"); 400 scsi_host_put(host); 401 return NULL; 402 } 403 404 spi_signalling(host) = hostdata->differential ? SPI_SIGNAL_HVD : 405 SPI_SIGNAL_SE; 406 407 return host; 408} 409 410int 411NCR_700_release(struct Scsi_Host *host) 412{ 413 struct NCR_700_Host_Parameters *hostdata = 414 (struct NCR_700_Host_Parameters *)host->hostdata[0]; 415 416 dma_free_noncoherent(hostdata->dev, TOTAL_MEM_SIZE, 417 hostdata->script, hostdata->pScript); 418 return 1; 419} 420 421static inline __u8 422NCR_700_identify(int can_disconnect, __u8 lun) 423{ 424 return IDENTIFY_BASE | 425 ((can_disconnect) ? 0x40 : 0) | 426 (lun & NCR_700_LUN_MASK); 427} 428 429/* 430 * Function : static int data_residual (Scsi_Host *host) 431 * 432 * Purpose : return residual data count of what's in the chip. If you 433 * really want to know what this function is doing, it's almost a 434 * direct transcription of the algorithm described in the 53c710 435 * guide, except that the DBC and DFIFO registers are only 6 bits 436 * wide on a 53c700. 437 * 438 * Inputs : host - SCSI host */ 439static inline int 440NCR_700_data_residual (struct Scsi_Host *host) { 441 struct NCR_700_Host_Parameters *hostdata = 442 (struct NCR_700_Host_Parameters *)host->hostdata[0]; 443 int count, synchronous = 0; 444 unsigned int ddir; 445 446 if(hostdata->chip710) { 447 count = ((NCR_700_readb(host, DFIFO_REG) & 0x7f) - 448 (NCR_700_readl(host, DBC_REG) & 0x7f)) & 0x7f; 449 } else { 450 count = ((NCR_700_readb(host, DFIFO_REG) & 0x3f) - 451 (NCR_700_readl(host, DBC_REG) & 0x3f)) & 0x3f; 452 } 453 454 if(hostdata->fast) 455 synchronous = NCR_700_readb(host, SXFER_REG) & 0x0f; 456 457 /* get the data direction */ 458 ddir = NCR_700_readb(host, CTEST0_REG) & 0x01; 459 460 if (ddir) { 461 /* Receive */ 462 if (synchronous) 463 count += (NCR_700_readb(host, SSTAT2_REG) & 0xf0) >> 4; 464 else 465 if (NCR_700_readb(host, SSTAT1_REG) & SIDL_REG_FULL) 466 ++count; 467 } else { 468 /* Send */ 469 __u8 sstat = NCR_700_readb(host, SSTAT1_REG); 470 if (sstat & SODL_REG_FULL) 471 ++count; 472 if (synchronous && (sstat & SODR_REG_FULL)) 473 ++count; 474 } 475#ifdef NCR_700_DEBUG 476 if(count) 477 printk("RESIDUAL IS %d (ddir %d)\n", count, ddir); 478#endif 479 return count; 480} 481 482/* print out the SCSI wires and corresponding phase from the SBCL register 483 * in the chip */ 484static inline char * 485sbcl_to_string(__u8 sbcl) 486{ 487 int i; 488 static char ret[256]; 489 490 ret[0]='\0'; 491 for(i=0; i<8; i++) { 492 if((1<<i) & sbcl) 493 strcat(ret, NCR_700_SBCL_bits[i]); 494 } 495 strcat(ret, NCR_700_SBCL_to_phase[sbcl & 0x07]); 496 return ret; 497} 498 499static inline __u8 500bitmap_to_number(__u8 bitmap) 501{ 502 __u8 i; 503 504 for(i=0; i<8 && !(bitmap &(1<<i)); i++) 505 ; 506 return i; 507} 508 509/* Pull a slot off the free list */ 510STATIC struct NCR_700_command_slot * 511find_empty_slot(struct NCR_700_Host_Parameters *hostdata) 512{ 513 struct NCR_700_command_slot *slot = hostdata->free_list; 514 515 if(slot == NULL) { 516 /* sanity check */ 517 if(hostdata->command_slot_count != NCR_700_COMMAND_SLOTS_PER_HOST) 518 printk(KERN_ERR "SLOTS FULL, but count is %d, should be %d\n", hostdata->command_slot_count, NCR_700_COMMAND_SLOTS_PER_HOST); 519 return NULL; 520 } 521 522 if(slot->state != NCR_700_SLOT_FREE) 523 /* should panic! */ 524 printk(KERN_ERR "BUSY SLOT ON FREE LIST!!!\n"); 525 526 527 hostdata->free_list = slot->ITL_forw; 528 slot->ITL_forw = NULL; 529 530 531 /* NOTE: set the state to busy here, not queued, since this 532 * indicates the slot is in use and cannot be run by the IRQ 533 * finish routine. If we cannot queue the command when it 534 * is properly build, we then change to NCR_700_SLOT_QUEUED */ 535 slot->state = NCR_700_SLOT_BUSY; 536 slot->flags = 0; 537 hostdata->command_slot_count++; 538 539 return slot; 540} 541 542STATIC void 543free_slot(struct NCR_700_command_slot *slot, 544 struct NCR_700_Host_Parameters *hostdata) 545{ 546 if((slot->state & NCR_700_SLOT_MASK) != NCR_700_SLOT_MAGIC) { 547 printk(KERN_ERR "53c700: SLOT %p is not MAGIC!!!\n", slot); 548 } 549 if(slot->state == NCR_700_SLOT_FREE) { 550 printk(KERN_ERR "53c700: SLOT %p is FREE!!!\n", slot); 551 } 552 553 slot->resume_offset = 0; 554 slot->cmnd = NULL; 555 slot->state = NCR_700_SLOT_FREE; 556 slot->ITL_forw = hostdata->free_list; 557 hostdata->free_list = slot; 558 hostdata->command_slot_count--; 559} 560 561 562/* This routine really does very little. The command is indexed on 563 the ITL and (if tagged) the ITLQ lists in _queuecommand */ 564STATIC void 565save_for_reselection(struct NCR_700_Host_Parameters *hostdata, 566 struct scsi_cmnd *SCp, __u32 dsp) 567{ 568 /* Its just possible that this gets executed twice */ 569 if(SCp != NULL) { 570 struct NCR_700_command_slot *slot = 571 (struct NCR_700_command_slot *)SCp->host_scribble; 572 573 slot->resume_offset = dsp; 574 } 575 hostdata->state = NCR_700_HOST_FREE; 576 hostdata->cmd = NULL; 577} 578 579STATIC inline void 580NCR_700_unmap(struct NCR_700_Host_Parameters *hostdata, struct scsi_cmnd *SCp, 581 struct NCR_700_command_slot *slot) 582{ 583 if(SCp->sc_data_direction != DMA_NONE && 584 SCp->sc_data_direction != DMA_BIDIRECTIONAL) 585 scsi_dma_unmap(SCp); 586} 587 588STATIC inline void 589NCR_700_scsi_done(struct NCR_700_Host_Parameters *hostdata, 590 struct scsi_cmnd *SCp, int result) 591{ 592 hostdata->state = NCR_700_HOST_FREE; 593 hostdata->cmd = NULL; 594 595 if(SCp != NULL) { 596 struct NCR_700_command_slot *slot = 597 (struct NCR_700_command_slot *)SCp->host_scribble; 598 599 dma_unmap_single(hostdata->dev, slot->pCmd, 600 MAX_COMMAND_SIZE, DMA_TO_DEVICE); 601 if (slot->flags == NCR_700_FLAG_AUTOSENSE) { 602 char *cmnd = NCR_700_get_sense_cmnd(SCp->device); 603#ifdef NCR_700_DEBUG 604 printk(" ORIGINAL CMD %p RETURNED %d, new return is %d sense is\n", 605 SCp, SCp->cmnd[7], result); 606 scsi_print_sense("53c700", SCp); 607 608#endif 609 dma_unmap_single(hostdata->dev, slot->dma_handle, 610 SCSI_SENSE_BUFFERSIZE, DMA_FROM_DEVICE); 611 /* restore the old result if the request sense was 612 * successful */ 613 if (result == 0) 614 result = cmnd[7]; 615 /* restore the original length */ 616 SCp->cmd_len = cmnd[8]; 617 } else 618 NCR_700_unmap(hostdata, SCp, slot); 619 620 free_slot(slot, hostdata); 621#ifdef NCR_700_DEBUG 622 if(NCR_700_get_depth(SCp->device) == 0 || 623 NCR_700_get_depth(SCp->device) > SCp->device->queue_depth) 624 printk(KERN_ERR "Invalid depth in NCR_700_scsi_done(): %d\n", 625 NCR_700_get_depth(SCp->device)); 626#endif /* NCR_700_DEBUG */ 627 NCR_700_set_depth(SCp->device, NCR_700_get_depth(SCp->device) - 1); 628 629 SCp->host_scribble = NULL; 630 SCp->result = result; 631 SCp->scsi_done(SCp); 632 } else { 633 printk(KERN_ERR "53c700: SCSI DONE HAS NULL SCp\n"); 634 } 635} 636 637 638STATIC void 639NCR_700_internal_bus_reset(struct Scsi_Host *host) 640{ 641 /* Bus reset */ 642 NCR_700_writeb(ASSERT_RST, host, SCNTL1_REG); 643 udelay(50); 644 NCR_700_writeb(0, host, SCNTL1_REG); 645 646} 647 648STATIC void 649NCR_700_chip_setup(struct Scsi_Host *host) 650{ 651 struct NCR_700_Host_Parameters *hostdata = 652 (struct NCR_700_Host_Parameters *)host->hostdata[0]; 653 __u8 min_period; 654 __u8 min_xferp = (hostdata->chip710 ? NCR_710_MIN_XFERP : NCR_700_MIN_XFERP); 655 656 if(hostdata->chip710) { 657 __u8 burst_disable = 0; 658 __u8 burst_length = 0; 659 660 switch (hostdata->burst_length) { 661 case 1: 662 burst_length = BURST_LENGTH_1; 663 break; 664 case 2: 665 burst_length = BURST_LENGTH_2; 666 break; 667 case 4: 668 burst_length = BURST_LENGTH_4; 669 break; 670 case 8: 671 burst_length = BURST_LENGTH_8; 672 break; 673 default: 674 burst_disable = BURST_DISABLE; 675 break; 676 } 677 hostdata->dcntl_extra |= COMPAT_700_MODE; 678 679 NCR_700_writeb(hostdata->dcntl_extra, host, DCNTL_REG); 680 NCR_700_writeb(burst_length | hostdata->dmode_extra, 681 host, DMODE_710_REG); 682 NCR_700_writeb(burst_disable | hostdata->ctest7_extra | 683 (hostdata->differential ? DIFF : 0), 684 host, CTEST7_REG); 685 NCR_700_writeb(BTB_TIMER_DISABLE, host, CTEST0_REG); 686 NCR_700_writeb(FULL_ARBITRATION | ENABLE_PARITY | PARITY 687 | AUTO_ATN, host, SCNTL0_REG); 688 } else { 689 NCR_700_writeb(BURST_LENGTH_8 | hostdata->dmode_extra, 690 host, DMODE_700_REG); 691 NCR_700_writeb(hostdata->differential ? 692 DIFF : 0, host, CTEST7_REG); 693 if(hostdata->fast) { 694 /* this is for 700-66, does nothing on 700 */ 695 NCR_700_writeb(LAST_DIS_ENBL | ENABLE_ACTIVE_NEGATION 696 | GENERATE_RECEIVE_PARITY, host, 697 CTEST8_REG); 698 } else { 699 NCR_700_writeb(FULL_ARBITRATION | ENABLE_PARITY 700 | PARITY | AUTO_ATN, host, SCNTL0_REG); 701 } 702 } 703 704 NCR_700_writeb(1 << host->this_id, host, SCID_REG); 705 NCR_700_writeb(0, host, SBCL_REG); 706 NCR_700_writeb(ASYNC_OPERATION, host, SXFER_REG); 707 708 NCR_700_writeb(PHASE_MM_INT | SEL_TIMEOUT_INT | GROSS_ERR_INT | UX_DISC_INT 709 | RST_INT | PAR_ERR_INT | SELECT_INT, host, SIEN_REG); 710 711 NCR_700_writeb(ABORT_INT | INT_INST_INT | ILGL_INST_INT, host, DIEN_REG); 712 NCR_700_writeb(ENABLE_SELECT, host, SCNTL1_REG); 713 if(hostdata->clock > 75) { 714 printk(KERN_ERR "53c700: Clock speed %dMHz is too high: 75Mhz is the maximum this chip can be driven at\n", hostdata->clock); 715 /* do the best we can, but the async clock will be out 716 * of spec: sync divider 2, async divider 3 */ 717 DEBUG(("53c700: sync 2 async 3\n")); 718 NCR_700_writeb(SYNC_DIV_2_0, host, SBCL_REG); 719 NCR_700_writeb(ASYNC_DIV_3_0 | hostdata->dcntl_extra, host, DCNTL_REG); 720 hostdata->sync_clock = hostdata->clock/2; 721 } else if(hostdata->clock > 50 && hostdata->clock <= 75) { 722 /* sync divider 1.5, async divider 3 */ 723 DEBUG(("53c700: sync 1.5 async 3\n")); 724 NCR_700_writeb(SYNC_DIV_1_5, host, SBCL_REG); 725 NCR_700_writeb(ASYNC_DIV_3_0 | hostdata->dcntl_extra, host, DCNTL_REG); 726 hostdata->sync_clock = hostdata->clock*2; 727 hostdata->sync_clock /= 3; 728 729 } else if(hostdata->clock > 37 && hostdata->clock <= 50) { 730 /* sync divider 1, async divider 2 */ 731 DEBUG(("53c700: sync 1 async 2\n")); 732 NCR_700_writeb(SYNC_DIV_1_0, host, SBCL_REG); 733 NCR_700_writeb(ASYNC_DIV_2_0 | hostdata->dcntl_extra, host, DCNTL_REG); 734 hostdata->sync_clock = hostdata->clock; 735 } else if(hostdata->clock > 25 && hostdata->clock <=37) { 736 /* sync divider 1, async divider 1.5 */ 737 DEBUG(("53c700: sync 1 async 1.5\n")); 738 NCR_700_writeb(SYNC_DIV_1_0, host, SBCL_REG); 739 NCR_700_writeb(ASYNC_DIV_1_5 | hostdata->dcntl_extra, host, DCNTL_REG); 740 hostdata->sync_clock = hostdata->clock; 741 } else { 742 DEBUG(("53c700: sync 1 async 1\n")); 743 NCR_700_writeb(SYNC_DIV_1_0, host, SBCL_REG); 744 NCR_700_writeb(ASYNC_DIV_1_0 | hostdata->dcntl_extra, host, DCNTL_REG); 745 /* sync divider 1, async divider 1 */ 746 hostdata->sync_clock = hostdata->clock; 747 } 748 /* Calculate the actual minimum period that can be supported 749 * by our synchronous clock speed. See the 710 manual for 750 * exact details of this calculation which is based on a 751 * setting of the SXFER register */ 752 min_period = 1000*(4+min_xferp)/(4*hostdata->sync_clock); 753 hostdata->min_period = NCR_700_MIN_PERIOD; 754 if(min_period > NCR_700_MIN_PERIOD) 755 hostdata->min_period = min_period; 756} 757 758STATIC void 759NCR_700_chip_reset(struct Scsi_Host *host) 760{ 761 struct NCR_700_Host_Parameters *hostdata = 762 (struct NCR_700_Host_Parameters *)host->hostdata[0]; 763 if(hostdata->chip710) { 764 NCR_700_writeb(SOFTWARE_RESET_710, host, ISTAT_REG); 765 udelay(100); 766 767 NCR_700_writeb(0, host, ISTAT_REG); 768 } else { 769 NCR_700_writeb(SOFTWARE_RESET, host, DCNTL_REG); 770 udelay(100); 771 772 NCR_700_writeb(0, host, DCNTL_REG); 773 } 774 775 mdelay(1000); 776 777 NCR_700_chip_setup(host); 778} 779 780/* The heart of the message processing engine is that the instruction 781 * immediately after the INT is the normal case (and so must be CLEAR 782 * ACK). If we want to do something else, we call that routine in 783 * scripts and set temp to be the normal case + 8 (skipping the CLEAR 784 * ACK) so that the routine returns correctly to resume its activity 785 * */ 786STATIC __u32 787process_extended_message(struct Scsi_Host *host, 788 struct NCR_700_Host_Parameters *hostdata, 789 struct scsi_cmnd *SCp, __u32 dsp, __u32 dsps) 790{ 791 __u32 resume_offset = dsp, temp = dsp + 8; 792 __u8 pun = 0xff, lun = 0xff; 793 794 if(SCp != NULL) { 795 pun = SCp->device->id; 796 lun = SCp->device->lun; 797 } 798 799 switch(hostdata->msgin[2]) { 800 case A_SDTR_MSG: 801 if(SCp != NULL && NCR_700_is_flag_set(SCp->device, NCR_700_DEV_BEGIN_SYNC_NEGOTIATION)) { 802 struct scsi_target *starget = SCp->device->sdev_target; 803 __u8 period = hostdata->msgin[3]; 804 __u8 offset = hostdata->msgin[4]; 805 806 if(offset == 0 || period == 0) { 807 offset = 0; 808 period = 0; 809 } 810 811 spi_offset(starget) = offset; 812 spi_period(starget) = period; 813 814 if(NCR_700_is_flag_set(SCp->device, NCR_700_DEV_PRINT_SYNC_NEGOTIATION)) { 815 spi_display_xfer_agreement(starget); 816 NCR_700_clear_flag(SCp->device, NCR_700_DEV_PRINT_SYNC_NEGOTIATION); 817 } 818 819 NCR_700_set_flag(SCp->device, NCR_700_DEV_NEGOTIATED_SYNC); 820 NCR_700_clear_flag(SCp->device, NCR_700_DEV_BEGIN_SYNC_NEGOTIATION); 821 822 NCR_700_writeb(NCR_700_get_SXFER(SCp->device), 823 host, SXFER_REG); 824 825 } else { 826 /* SDTR message out of the blue, reject it */ 827 shost_printk(KERN_WARNING, host, 828 "Unexpected SDTR msg\n"); 829 hostdata->msgout[0] = A_REJECT_MSG; 830 dma_cache_sync(hostdata->dev, hostdata->msgout, 1, DMA_TO_DEVICE); 831 script_patch_16(hostdata->dev, hostdata->script, 832 MessageCount, 1); 833 /* SendMsgOut returns, so set up the return 834 * address */ 835 resume_offset = hostdata->pScript + Ent_SendMessageWithATN; 836 } 837 break; 838 839 case A_WDTR_MSG: 840 printk(KERN_INFO "scsi%d: (%d:%d), Unsolicited WDTR after CMD, Rejecting\n", 841 host->host_no, pun, lun); 842 hostdata->msgout[0] = A_REJECT_MSG; 843 dma_cache_sync(hostdata->dev, hostdata->msgout, 1, DMA_TO_DEVICE); 844 script_patch_16(hostdata->dev, hostdata->script, MessageCount, 845 1); 846 resume_offset = hostdata->pScript + Ent_SendMessageWithATN; 847 848 break; 849 850 default: 851 printk(KERN_INFO "scsi%d (%d:%d): Unexpected message %s: ", 852 host->host_no, pun, lun, 853 NCR_700_phase[(dsps & 0xf00) >> 8]); 854 spi_print_msg(hostdata->msgin); 855 printk("\n"); 856 /* just reject it */ 857 hostdata->msgout[0] = A_REJECT_MSG; 858 dma_cache_sync(hostdata->dev, hostdata->msgout, 1, DMA_TO_DEVICE); 859 script_patch_16(hostdata->dev, hostdata->script, MessageCount, 860 1); 861 /* SendMsgOut returns, so set up the return 862 * address */ 863 resume_offset = hostdata->pScript + Ent_SendMessageWithATN; 864 } 865 NCR_700_writel(temp, host, TEMP_REG); 866 return resume_offset; 867} 868 869STATIC __u32 870process_message(struct Scsi_Host *host, struct NCR_700_Host_Parameters *hostdata, 871 struct scsi_cmnd *SCp, __u32 dsp, __u32 dsps) 872{ 873 /* work out where to return to */ 874 __u32 temp = dsp + 8, resume_offset = dsp; 875 __u8 pun = 0xff, lun = 0xff; 876 877 if(SCp != NULL) { 878 pun = SCp->device->id; 879 lun = SCp->device->lun; 880 } 881 882#ifdef NCR_700_DEBUG 883 printk("scsi%d (%d:%d): message %s: ", host->host_no, pun, lun, 884 NCR_700_phase[(dsps & 0xf00) >> 8]); 885 spi_print_msg(hostdata->msgin); 886 printk("\n"); 887#endif 888 889 switch(hostdata->msgin[0]) { 890 891 case A_EXTENDED_MSG: 892 resume_offset = process_extended_message(host, hostdata, SCp, 893 dsp, dsps); 894 break; 895 896 case A_REJECT_MSG: 897 if(SCp != NULL && NCR_700_is_flag_set(SCp->device, NCR_700_DEV_BEGIN_SYNC_NEGOTIATION)) { 898 /* Rejected our sync negotiation attempt */ 899 spi_period(SCp->device->sdev_target) = 900 spi_offset(SCp->device->sdev_target) = 0; 901 NCR_700_set_flag(SCp->device, NCR_700_DEV_NEGOTIATED_SYNC); 902 NCR_700_clear_flag(SCp->device, NCR_700_DEV_BEGIN_SYNC_NEGOTIATION); 903 } else if(SCp != NULL && NCR_700_get_tag_neg_state(SCp->device) == NCR_700_DURING_TAG_NEGOTIATION) { 904 /* rejected our first simple tag message */ 905 scmd_printk(KERN_WARNING, SCp, 906 "Rejected first tag queue attempt, turning off tag queueing\n"); 907 /* we're done negotiating */ 908 NCR_700_set_tag_neg_state(SCp->device, NCR_700_FINISHED_TAG_NEGOTIATION); 909 hostdata->tag_negotiated &= ~(1<<scmd_id(SCp)); 910 SCp->device->tagged_supported = 0; 911 scsi_deactivate_tcq(SCp->device, host->cmd_per_lun); 912 } else { 913 shost_printk(KERN_WARNING, host, 914 "(%d:%d) Unexpected REJECT Message %s\n", 915 pun, lun, 916 NCR_700_phase[(dsps & 0xf00) >> 8]); 917 /* however, just ignore it */ 918 } 919 break; 920 921 case A_PARITY_ERROR_MSG: 922 printk(KERN_ERR "scsi%d (%d:%d) Parity Error!\n", host->host_no, 923 pun, lun); 924 NCR_700_internal_bus_reset(host); 925 break; 926 case A_SIMPLE_TAG_MSG: 927 printk(KERN_INFO "scsi%d (%d:%d) SIMPLE TAG %d %s\n", host->host_no, 928 pun, lun, hostdata->msgin[1], 929 NCR_700_phase[(dsps & 0xf00) >> 8]); 930 /* just ignore it */ 931 break; 932 default: 933 printk(KERN_INFO "scsi%d (%d:%d): Unexpected message %s: ", 934 host->host_no, pun, lun, 935 NCR_700_phase[(dsps & 0xf00) >> 8]); 936 937 spi_print_msg(hostdata->msgin); 938 printk("\n"); 939 /* just reject it */ 940 hostdata->msgout[0] = A_REJECT_MSG; 941 dma_cache_sync(hostdata->dev, hostdata->msgout, 1, DMA_TO_DEVICE); 942 script_patch_16(hostdata->dev, hostdata->script, MessageCount, 943 1); 944 /* SendMsgOut returns, so set up the return 945 * address */ 946 resume_offset = hostdata->pScript + Ent_SendMessageWithATN; 947 948 break; 949 } 950 NCR_700_writel(temp, host, TEMP_REG); 951 /* set us up to receive another message */ 952 dma_cache_sync(hostdata->dev, hostdata->msgin, MSG_ARRAY_SIZE, DMA_FROM_DEVICE); 953 return resume_offset; 954} 955 956STATIC __u32 957process_script_interrupt(__u32 dsps, __u32 dsp, struct scsi_cmnd *SCp, 958 struct Scsi_Host *host, 959 struct NCR_700_Host_Parameters *hostdata) 960{ 961 __u32 resume_offset = 0; 962 __u8 pun = 0xff, lun=0xff; 963 964 if(SCp != NULL) { 965 pun = SCp->device->id; 966 lun = SCp->device->lun; 967 } 968 969 if(dsps == A_GOOD_STATUS_AFTER_STATUS) { 970 DEBUG((" COMMAND COMPLETE, status=%02x\n", 971 hostdata->status[0])); 972 /* OK, if TCQ still under negotiation, we now know it works */ 973 if (NCR_700_get_tag_neg_state(SCp->device) == NCR_700_DURING_TAG_NEGOTIATION) 974 NCR_700_set_tag_neg_state(SCp->device, 975 NCR_700_FINISHED_TAG_NEGOTIATION); 976 977 /* check for contingent allegiance contitions */ 978 if(status_byte(hostdata->status[0]) == CHECK_CONDITION || 979 status_byte(hostdata->status[0]) == COMMAND_TERMINATED) { 980 struct NCR_700_command_slot *slot = 981 (struct NCR_700_command_slot *)SCp->host_scribble; 982 if(slot->flags == NCR_700_FLAG_AUTOSENSE) { 983 /* OOPS: bad device, returning another 984 * contingent allegiance condition */ 985 scmd_printk(KERN_ERR, SCp, 986 "broken device is looping in contingent allegiance: ignoring\n"); 987 NCR_700_scsi_done(hostdata, SCp, hostdata->status[0]); 988 } else { 989 char *cmnd = 990 NCR_700_get_sense_cmnd(SCp->device); 991#ifdef NCR_DEBUG 992 scsi_print_command(SCp); 993 printk(" cmd %p has status %d, requesting sense\n", 994 SCp, hostdata->status[0]); 995#endif 996 /* we can destroy the command here 997 * because the contingent allegiance 998 * condition will cause a retry which 999 * will re-copy the command from the 1000 * saved data_cmnd. We also unmap any 1001 * data associated with the command 1002 * here */ 1003 NCR_700_unmap(hostdata, SCp, slot); 1004 dma_unmap_single(hostdata->dev, slot->pCmd, 1005 MAX_COMMAND_SIZE, 1006 DMA_TO_DEVICE); 1007 1008 cmnd[0] = REQUEST_SENSE; 1009 cmnd[1] = (SCp->device->lun & 0x7) << 5; 1010 cmnd[2] = 0; 1011 cmnd[3] = 0; 1012 cmnd[4] = SCSI_SENSE_BUFFERSIZE; 1013 cmnd[5] = 0; 1014 /* Here's a quiet hack: the 1015 * REQUEST_SENSE command is six bytes, 1016 * so store a flag indicating that 1017 * this was an internal sense request 1018 * and the original status at the end 1019 * of the command */ 1020 cmnd[6] = NCR_700_INTERNAL_SENSE_MAGIC; 1021 cmnd[7] = hostdata->status[0]; 1022 cmnd[8] = SCp->cmd_len; 1023 SCp->cmd_len = 6; /* command length for 1024 * REQUEST_SENSE */ 1025 slot->pCmd = dma_map_single(hostdata->dev, cmnd, MAX_COMMAND_SIZE, DMA_TO_DEVICE); 1026 slot->dma_handle = dma_map_single(hostdata->dev, SCp->sense_buffer, SCSI_SENSE_BUFFERSIZE, DMA_FROM_DEVICE); 1027 slot->SG[0].ins = bS_to_host(SCRIPT_MOVE_DATA_IN | SCSI_SENSE_BUFFERSIZE); 1028 slot->SG[0].pAddr = bS_to_host(slot->dma_handle); 1029 slot->SG[1].ins = bS_to_host(SCRIPT_RETURN); 1030 slot->SG[1].pAddr = 0; 1031 slot->resume_offset = hostdata->pScript; 1032 dma_cache_sync(hostdata->dev, slot->SG, sizeof(slot->SG[0])*2, DMA_TO_DEVICE); 1033 dma_cache_sync(hostdata->dev, SCp->sense_buffer, SCSI_SENSE_BUFFERSIZE, DMA_FROM_DEVICE); 1034 1035 /* queue the command for reissue */ 1036 slot->state = NCR_700_SLOT_QUEUED; 1037 slot->flags = NCR_700_FLAG_AUTOSENSE; 1038 hostdata->state = NCR_700_HOST_FREE; 1039 hostdata->cmd = NULL; 1040 } 1041 } else { 1042 // Currently rely on the mid layer evaluation 1043 // of the tag queuing capability 1044 // 1045 //if(status_byte(hostdata->status[0]) == GOOD && 1046 // SCp->cmnd[0] == INQUIRY && SCp->use_sg == 0) { 1047 // /* Piggy back the tag queueing support 1048 // * on this command */ 1049 // dma_sync_single_for_cpu(hostdata->dev, 1050 // slot->dma_handle, 1051 // SCp->request_bufflen, 1052 // DMA_FROM_DEVICE); 1053 // if(((char *)SCp->request_buffer)[7] & 0x02) { 1054 // scmd_printk(KERN_INFO, SCp, 1055 // "Enabling Tag Command Queuing\n"); 1056 // hostdata->tag_negotiated |= (1<<scmd_id(SCp)); 1057 // NCR_700_set_flag(SCp->device, NCR_700_DEV_BEGIN_TAG_QUEUEING); 1058 // } else { 1059 // NCR_700_clear_flag(SCp->device, NCR_700_DEV_BEGIN_TAG_QUEUEING); 1060 // hostdata->tag_negotiated &= ~(1<<scmd_id(SCp)); 1061 // } 1062 //} 1063 NCR_700_scsi_done(hostdata, SCp, hostdata->status[0]); 1064 } 1065 } else if((dsps & 0xfffff0f0) == A_UNEXPECTED_PHASE) { 1066 __u8 i = (dsps & 0xf00) >> 8; 1067 1068 scmd_printk(KERN_ERR, SCp, "UNEXPECTED PHASE %s (%s)\n", 1069 NCR_700_phase[i], 1070 sbcl_to_string(NCR_700_readb(host, SBCL_REG))); 1071 scmd_printk(KERN_ERR, SCp, " len = %d, cmd =", 1072 SCp->cmd_len); 1073 scsi_print_command(SCp); 1074 1075 NCR_700_internal_bus_reset(host); 1076 } else if((dsps & 0xfffff000) == A_FATAL) { 1077 int i = (dsps & 0xfff); 1078 1079 printk(KERN_ERR "scsi%d: (%d:%d) FATAL ERROR: %s\n", 1080 host->host_no, pun, lun, NCR_700_fatal_messages[i]); 1081 if(dsps == A_FATAL_ILLEGAL_MSG_LENGTH) { 1082 printk(KERN_ERR " msg begins %02x %02x\n", 1083 hostdata->msgin[0], hostdata->msgin[1]); 1084 } 1085 NCR_700_internal_bus_reset(host); 1086 } else if((dsps & 0xfffff0f0) == A_DISCONNECT) { 1087#ifdef NCR_700_DEBUG 1088 __u8 i = (dsps & 0xf00) >> 8; 1089 1090 printk("scsi%d: (%d:%d), DISCONNECTED (%d) %s\n", 1091 host->host_no, pun, lun, 1092 i, NCR_700_phase[i]); 1093#endif 1094 save_for_reselection(hostdata, SCp, dsp); 1095 1096 } else if(dsps == A_RESELECTION_IDENTIFIED) { 1097 __u8 lun; 1098 struct NCR_700_command_slot *slot; 1099 __u8 reselection_id = hostdata->reselection_id; 1100 struct scsi_device *SDp; 1101 1102 lun = hostdata->msgin[0] & 0x1f; 1103 1104 hostdata->reselection_id = 0xff; 1105 DEBUG(("scsi%d: (%d:%d) RESELECTED!\n", 1106 host->host_no, reselection_id, lun)); 1107 /* clear the reselection indicator */ 1108 SDp = __scsi_device_lookup(host, 0, reselection_id, lun); 1109 if(unlikely(SDp == NULL)) { 1110 printk(KERN_ERR "scsi%d: (%d:%d) HAS NO device\n", 1111 host->host_no, reselection_id, lun); 1112 BUG(); 1113 } 1114 if(hostdata->msgin[1] == A_SIMPLE_TAG_MSG) { 1115 struct scsi_cmnd *SCp = scsi_find_tag(SDp, hostdata->msgin[2]); 1116 if(unlikely(SCp == NULL)) { 1117 printk(KERN_ERR "scsi%d: (%d:%d) no saved request for tag %d\n", 1118 host->host_no, reselection_id, lun, hostdata->msgin[2]); 1119 BUG(); 1120 } 1121 1122 slot = (struct NCR_700_command_slot *)SCp->host_scribble; 1123 DDEBUG(KERN_DEBUG, SDp, 1124 "reselection is tag %d, slot %p(%d)\n", 1125 hostdata->msgin[2], slot, slot->tag); 1126 } else { 1127 struct scsi_cmnd *SCp = scsi_find_tag(SDp, SCSI_NO_TAG); 1128 if(unlikely(SCp == NULL)) { 1129 sdev_printk(KERN_ERR, SDp, 1130 "no saved request for untagged cmd\n"); 1131 BUG(); 1132 } 1133 slot = (struct NCR_700_command_slot *)SCp->host_scribble; 1134 } 1135 1136 if(slot == NULL) { 1137 printk(KERN_ERR "scsi%d: (%d:%d) RESELECTED but no saved command (MSG = %02x %02x %02x)!!\n", 1138 host->host_no, reselection_id, lun, 1139 hostdata->msgin[0], hostdata->msgin[1], 1140 hostdata->msgin[2]); 1141 } else { 1142 if(hostdata->state != NCR_700_HOST_BUSY) 1143 printk(KERN_ERR "scsi%d: FATAL, host not busy during valid reselection!\n", 1144 host->host_no); 1145 resume_offset = slot->resume_offset; 1146 hostdata->cmd = slot->cmnd; 1147 1148 /* re-patch for this command */ 1149 script_patch_32_abs(hostdata->dev, hostdata->script, 1150 CommandAddress, slot->pCmd); 1151 script_patch_16(hostdata->dev, hostdata->script, 1152 CommandCount, slot->cmnd->cmd_len); 1153 script_patch_32_abs(hostdata->dev, hostdata->script, 1154 SGScriptStartAddress, 1155 to32bit(&slot->pSG[0].ins)); 1156 1157 /* Note: setting SXFER only works if we're 1158 * still in the MESSAGE phase, so it is vital 1159 * that ACK is still asserted when we process 1160 * the reselection message. The resume offset 1161 * should therefore always clear ACK */ 1162 NCR_700_writeb(NCR_700_get_SXFER(hostdata->cmd->device), 1163 host, SXFER_REG); 1164 dma_cache_sync(hostdata->dev, hostdata->msgin, 1165 MSG_ARRAY_SIZE, DMA_FROM_DEVICE); 1166 dma_cache_sync(hostdata->dev, hostdata->msgout, 1167 MSG_ARRAY_SIZE, DMA_TO_DEVICE); 1168 /* I'm just being paranoid here, the command should 1169 * already have been flushed from the cache */ 1170 dma_cache_sync(hostdata->dev, slot->cmnd->cmnd, 1171 slot->cmnd->cmd_len, DMA_TO_DEVICE); 1172 1173 1174 1175 } 1176 } else if(dsps == A_RESELECTED_DURING_SELECTION) { 1177 1178 /* This section is full of debugging code because I've 1179 * never managed to reach it. I think what happens is 1180 * that, because the 700 runs with selection 1181 * interrupts enabled the whole time that we take a 1182 * selection interrupt before we manage to get to the 1183 * reselected script interrupt */ 1184 1185 __u8 reselection_id = NCR_700_readb(host, SFBR_REG); 1186 struct NCR_700_command_slot *slot; 1187 1188 /* Take out our own ID */ 1189 reselection_id &= ~(1<<host->this_id); 1190 1191 /* I've never seen this happen, so keep this as a printk rather 1192 * than a debug */ 1193 printk(KERN_INFO "scsi%d: (%d:%d) RESELECTION DURING SELECTION, dsp=%08x[%04x] state=%d, count=%d\n", 1194 host->host_no, reselection_id, lun, dsp, dsp - hostdata->pScript, hostdata->state, hostdata->command_slot_count); 1195 1196 { 1197 __u32 SG = (__u32)bS_to_cpu(hostdata->script[A_SGScriptStartAddress_used[0]]); 1198 int i; 1199 1200 for(i=0; i< NCR_700_COMMAND_SLOTS_PER_HOST; i++) { 1201 if(SG >= to32bit(&hostdata->slots[i].pSG[0]) 1202 && SG <= to32bit(&hostdata->slots[i].pSG[NCR_700_SG_SEGMENTS])) 1203 break; 1204 } 1205 printk(KERN_INFO "IDENTIFIED SG segment as being %08x in slot %p, cmd %p, slot->resume_offset=%08x\n", SG, &hostdata->slots[i], hostdata->slots[i].cmnd, hostdata->slots[i].resume_offset); 1206 SCp = hostdata->slots[i].cmnd; 1207 } 1208 1209 if(SCp != NULL) { 1210 slot = (struct NCR_700_command_slot *)SCp->host_scribble; 1211 /* change slot from busy to queued to redo command */ 1212 slot->state = NCR_700_SLOT_QUEUED; 1213 } 1214 hostdata->cmd = NULL; 1215 1216 if(reselection_id == 0) { 1217 if(hostdata->reselection_id == 0xff) { 1218 printk(KERN_ERR "scsi%d: Invalid reselection during selection!!\n", host->host_no); 1219 return 0; 1220 } else { 1221 printk(KERN_ERR "scsi%d: script reselected and we took a selection interrupt\n", 1222 host->host_no); 1223 reselection_id = hostdata->reselection_id; 1224 } 1225 } else { 1226 1227 /* convert to real ID */ 1228 reselection_id = bitmap_to_number(reselection_id); 1229 } 1230 hostdata->reselection_id = reselection_id; 1231 /* just in case we have a stale simple tag message, clear it */ 1232 hostdata->msgin[1] = 0; 1233 dma_cache_sync(hostdata->dev, hostdata->msgin, 1234 MSG_ARRAY_SIZE, DMA_BIDIRECTIONAL); 1235 if(hostdata->tag_negotiated & (1<<reselection_id)) { 1236 resume_offset = hostdata->pScript + Ent_GetReselectionWithTag; 1237 } else { 1238 resume_offset = hostdata->pScript + Ent_GetReselectionData; 1239 } 1240 } else if(dsps == A_COMPLETED_SELECTION_AS_TARGET) { 1241 /* we've just disconnected from the bus, do nothing since 1242 * a return here will re-run the queued command slot 1243 * that may have been interrupted by the initial selection */ 1244 DEBUG((" SELECTION COMPLETED\n")); 1245 } else if((dsps & 0xfffff0f0) == A_MSG_IN) { 1246 resume_offset = process_message(host, hostdata, SCp, 1247 dsp, dsps); 1248 } else if((dsps & 0xfffff000) == 0) { 1249 __u8 i = (dsps & 0xf0) >> 4, j = (dsps & 0xf00) >> 8; 1250 printk(KERN_ERR "scsi%d: (%d:%d), unhandled script condition %s %s at %04x\n", 1251 host->host_no, pun, lun, NCR_700_condition[i], 1252 NCR_700_phase[j], dsp - hostdata->pScript); 1253 if(SCp != NULL) { 1254 struct scatterlist *sg; 1255 1256 scsi_print_command(SCp); 1257 scsi_for_each_sg(SCp, sg, scsi_sg_count(SCp) + 1, i) { 1258 printk(KERN_INFO " SG[%d].length = %d, move_insn=%08x, addr %08x\n", i, sg->length, ((struct NCR_700_command_slot *)SCp->host_scribble)->SG[i].ins, ((struct NCR_700_command_slot *)SCp->host_scribble)->SG[i].pAddr); 1259 } 1260 } 1261 NCR_700_internal_bus_reset(host); 1262 } else if((dsps & 0xfffff000) == A_DEBUG_INTERRUPT) { 1263 printk(KERN_NOTICE "scsi%d (%d:%d) DEBUG INTERRUPT %d AT %08x[%04x], continuing\n", 1264 host->host_no, pun, lun, dsps & 0xfff, dsp, dsp - hostdata->pScript); 1265 resume_offset = dsp; 1266 } else { 1267 printk(KERN_ERR "scsi%d: (%d:%d), unidentified script interrupt 0x%x at %04x\n", 1268 host->host_no, pun, lun, dsps, dsp - hostdata->pScript); 1269 NCR_700_internal_bus_reset(host); 1270 } 1271 return resume_offset; 1272} 1273 1274/* We run the 53c700 with selection interrupts always enabled. This 1275 * means that the chip may be selected as soon as the bus frees. On a 1276 * busy bus, this can be before the scripts engine finishes its 1277 * processing. Therefore, part of the selection processing has to be 1278 * to find out what the scripts engine is doing and complete the 1279 * function if necessary (i.e. process the pending disconnect or save 1280 * the interrupted initial selection */ 1281STATIC inline __u32 1282process_selection(struct Scsi_Host *host, __u32 dsp) 1283{ 1284 __u8 id = 0; /* Squash compiler warning */ 1285 int count = 0; 1286 __u32 resume_offset = 0; 1287 struct NCR_700_Host_Parameters *hostdata = 1288 (struct NCR_700_Host_Parameters *)host->hostdata[0]; 1289 struct scsi_cmnd *SCp = hostdata->cmd; 1290 __u8 sbcl; 1291 1292 for(count = 0; count < 5; count++) { 1293 id = NCR_700_readb(host, hostdata->chip710 ? 1294 CTEST9_REG : SFBR_REG); 1295 1296 /* Take out our own ID */ 1297 id &= ~(1<<host->this_id); 1298 if(id != 0) 1299 break; 1300 udelay(5); 1301 } 1302 sbcl = NCR_700_readb(host, SBCL_REG); 1303 if((sbcl & SBCL_IO) == 0) { 1304 /* mark as having been selected rather than reselected */ 1305 id = 0xff; 1306 } else { 1307 /* convert to real ID */ 1308 hostdata->reselection_id = id = bitmap_to_number(id); 1309 DEBUG(("scsi%d: Reselected by %d\n", 1310 host->host_no, id)); 1311 } 1312 if(hostdata->state == NCR_700_HOST_BUSY && SCp != NULL) { 1313 struct NCR_700_command_slot *slot = 1314 (struct NCR_700_command_slot *)SCp->host_scribble; 1315 DEBUG((" ID %d WARNING: RESELECTION OF BUSY HOST, saving cmd %p, slot %p, addr %x [%04x], resume %x!\n", id, hostdata->cmd, slot, dsp, dsp - hostdata->pScript, resume_offset)); 1316 1317 switch(dsp - hostdata->pScript) { 1318 case Ent_Disconnect1: 1319 case Ent_Disconnect2: 1320 save_for_reselection(hostdata, SCp, Ent_Disconnect2 + hostdata->pScript); 1321 break; 1322 case Ent_Disconnect3: 1323 case Ent_Disconnect4: 1324 save_for_reselection(hostdata, SCp, Ent_Disconnect4 + hostdata->pScript); 1325 break; 1326 case Ent_Disconnect5: 1327 case Ent_Disconnect6: 1328 save_for_reselection(hostdata, SCp, Ent_Disconnect6 + hostdata->pScript); 1329 break; 1330 case Ent_Disconnect7: 1331 case Ent_Disconnect8: 1332 save_for_reselection(hostdata, SCp, Ent_Disconnect8 + hostdata->pScript); 1333 break; 1334 case Ent_Finish1: 1335 case Ent_Finish2: 1336 process_script_interrupt(A_GOOD_STATUS_AFTER_STATUS, dsp, SCp, host, hostdata); 1337 break; 1338 1339 default: 1340 slot->state = NCR_700_SLOT_QUEUED; 1341 break; 1342 } 1343 } 1344 hostdata->state = NCR_700_HOST_BUSY; 1345 hostdata->cmd = NULL; 1346 /* clear any stale simple tag message */ 1347 hostdata->msgin[1] = 0; 1348 dma_cache_sync(hostdata->dev, hostdata->msgin, MSG_ARRAY_SIZE, 1349 DMA_BIDIRECTIONAL); 1350 1351 if(id == 0xff) { 1352 /* Selected as target, Ignore */ 1353 resume_offset = hostdata->pScript + Ent_SelectedAsTarget; 1354 } else if(hostdata->tag_negotiated & (1<<id)) { 1355 resume_offset = hostdata->pScript + Ent_GetReselectionWithTag; 1356 } else { 1357 resume_offset = hostdata->pScript + Ent_GetReselectionData; 1358 } 1359 return resume_offset; 1360} 1361 1362static inline void 1363NCR_700_clear_fifo(struct Scsi_Host *host) { 1364 const struct NCR_700_Host_Parameters *hostdata 1365 = (struct NCR_700_Host_Parameters *)host->hostdata[0]; 1366 if(hostdata->chip710) { 1367 NCR_700_writeb(CLR_FIFO_710, host, CTEST8_REG); 1368 } else { 1369 NCR_700_writeb(CLR_FIFO, host, DFIFO_REG); 1370 } 1371} 1372 1373static inline void 1374NCR_700_flush_fifo(struct Scsi_Host *host) { 1375 const struct NCR_700_Host_Parameters *hostdata 1376 = (struct NCR_700_Host_Parameters *)host->hostdata[0]; 1377 if(hostdata->chip710) { 1378 NCR_700_writeb(FLUSH_DMA_FIFO_710, host, CTEST8_REG); 1379 udelay(10); 1380 NCR_700_writeb(0, host, CTEST8_REG); 1381 } else { 1382 NCR_700_writeb(FLUSH_DMA_FIFO, host, DFIFO_REG); 1383 udelay(10); 1384 NCR_700_writeb(0, host, DFIFO_REG); 1385 } 1386} 1387 1388 1389/* The queue lock with interrupts disabled must be held on entry to 1390 * this function */ 1391STATIC int 1392NCR_700_start_command(struct scsi_cmnd *SCp) 1393{ 1394 struct NCR_700_command_slot *slot = 1395 (struct NCR_700_command_slot *)SCp->host_scribble; 1396 struct NCR_700_Host_Parameters *hostdata = 1397 (struct NCR_700_Host_Parameters *)SCp->device->host->hostdata[0]; 1398 __u16 count = 1; /* for IDENTIFY message */ 1399 1400 if(hostdata->state != NCR_700_HOST_FREE) { 1401 /* keep this inside the lock to close the race window where 1402 * the running command finishes on another CPU while we don't 1403 * change the state to queued on this one */ 1404 slot->state = NCR_700_SLOT_QUEUED; 1405 1406 DEBUG(("scsi%d: host busy, queueing command %p, slot %p\n", 1407 SCp->device->host->host_no, slot->cmnd, slot)); 1408 return 0; 1409 } 1410 hostdata->state = NCR_700_HOST_BUSY; 1411 hostdata->cmd = SCp; 1412 slot->state = NCR_700_SLOT_BUSY; 1413 /* keep interrupts disabled until we have the command correctly 1414 * set up so we cannot take a selection interrupt */ 1415 1416 hostdata->msgout[0] = NCR_700_identify((SCp->cmnd[0] != REQUEST_SENSE && 1417 slot->flags != NCR_700_FLAG_AUTOSENSE), 1418 SCp->device->lun); 1419 /* for INQUIRY or REQUEST_SENSE commands, we cannot be sure 1420 * if the negotiated transfer parameters still hold, so 1421 * always renegotiate them */ 1422 if(SCp->cmnd[0] == INQUIRY || SCp->cmnd[0] == REQUEST_SENSE || 1423 slot->flags == NCR_700_FLAG_AUTOSENSE) { 1424 NCR_700_clear_flag(SCp->device, NCR_700_DEV_NEGOTIATED_SYNC); 1425 } 1426 1427 /* REQUEST_SENSE is asking for contingent I_T_L(_Q) status. 1428 * If a contingent allegiance condition exists, the device 1429 * will refuse all tags, so send the request sense as untagged 1430 * */ 1431 if((hostdata->tag_negotiated & (1<<scmd_id(SCp))) 1432 && (slot->tag != SCSI_NO_TAG && SCp->cmnd[0] != REQUEST_SENSE && 1433 slot->flags != NCR_700_FLAG_AUTOSENSE)) { 1434 count += scsi_populate_tag_msg(SCp, &hostdata->msgout[count]); 1435 } 1436 1437 if(hostdata->fast && 1438 NCR_700_is_flag_clear(SCp->device, NCR_700_DEV_NEGOTIATED_SYNC)) { 1439 count += spi_populate_sync_msg(&hostdata->msgout[count], 1440 spi_period(SCp->device->sdev_target), 1441 spi_offset(SCp->device->sdev_target)); 1442 NCR_700_set_flag(SCp->device, NCR_700_DEV_BEGIN_SYNC_NEGOTIATION); 1443 } 1444 1445 script_patch_16(hostdata->dev, hostdata->script, MessageCount, count); 1446 1447 1448 script_patch_ID(hostdata->dev, hostdata->script, 1449 Device_ID, 1<<scmd_id(SCp)); 1450 1451 script_patch_32_abs(hostdata->dev, hostdata->script, CommandAddress, 1452 slot->pCmd); 1453 script_patch_16(hostdata->dev, hostdata->script, CommandCount, 1454 SCp->cmd_len); 1455 /* finally plumb the beginning of the SG list into the script 1456 * */ 1457 script_patch_32_abs(hostdata->dev, hostdata->script, 1458 SGScriptStartAddress, to32bit(&slot->pSG[0].ins)); 1459 NCR_700_clear_fifo(SCp->device->host); 1460 1461 if(slot->resume_offset == 0) 1462 slot->resume_offset = hostdata->pScript; 1463 /* now perform all the writebacks and invalidates */ 1464 dma_cache_sync(hostdata->dev, hostdata->msgout, count, DMA_TO_DEVICE); 1465 dma_cache_sync(hostdata->dev, hostdata->msgin, MSG_ARRAY_SIZE, 1466 DMA_FROM_DEVICE); 1467 dma_cache_sync(hostdata->dev, SCp->cmnd, SCp->cmd_len, DMA_TO_DEVICE); 1468 dma_cache_sync(hostdata->dev, hostdata->status, 1, DMA_FROM_DEVICE); 1469 1470 /* set the synchronous period/offset */ 1471 NCR_700_writeb(NCR_700_get_SXFER(SCp->device), 1472 SCp->device->host, SXFER_REG); 1473 NCR_700_writel(slot->temp, SCp->device->host, TEMP_REG); 1474 NCR_700_writel(slot->resume_offset, SCp->device->host, DSP_REG); 1475 1476 return 1; 1477} 1478 1479irqreturn_t 1480NCR_700_intr(int irq, void *dev_id) 1481{ 1482 struct Scsi_Host *host = (struct Scsi_Host *)dev_id; 1483 struct NCR_700_Host_Parameters *hostdata = 1484 (struct NCR_700_Host_Parameters *)host->hostdata[0]; 1485 __u8 istat; 1486 __u32 resume_offset = 0; 1487 __u8 pun = 0xff, lun = 0xff; 1488 unsigned long flags; 1489 int handled = 0; 1490 1491 /* Use the host lock to serialise access to the 53c700 1492 * hardware. Note: In future, we may need to take the queue 1493 * lock to enter the done routines. When that happens, we 1494 * need to ensure that for this driver, the host lock and the 1495 * queue lock point to the same thing. */ 1496 spin_lock_irqsave(host->host_lock, flags); 1497 if((istat = NCR_700_readb(host, ISTAT_REG)) 1498 & (SCSI_INT_PENDING | DMA_INT_PENDING)) { 1499 __u32 dsps; 1500 __u8 sstat0 = 0, dstat = 0; 1501 __u32 dsp; 1502 struct scsi_cmnd *SCp = hostdata->cmd; 1503 enum NCR_700_Host_State state; 1504 1505 handled = 1; 1506 state = hostdata->state; 1507 SCp = hostdata->cmd; 1508 1509 if(istat & SCSI_INT_PENDING) { 1510 udelay(10); 1511 1512 sstat0 = NCR_700_readb(host, SSTAT0_REG); 1513 } 1514 1515 if(istat & DMA_INT_PENDING) { 1516 udelay(10); 1517 1518 dstat = NCR_700_readb(host, DSTAT_REG); 1519 } 1520 1521 dsps = NCR_700_readl(host, DSPS_REG); 1522 dsp = NCR_700_readl(host, DSP_REG); 1523 1524 DEBUG(("scsi%d: istat %02x sstat0 %02x dstat %02x dsp %04x[%08x] dsps 0x%x\n", 1525 host->host_no, istat, sstat0, dstat, 1526 (dsp - (__u32)(hostdata->pScript))/4, 1527 dsp, dsps)); 1528 1529 if(SCp != NULL) { 1530 pun = SCp->device->id; 1531 lun = SCp->device->lun; 1532 } 1533 1534 if(sstat0 & SCSI_RESET_DETECTED) { 1535 struct scsi_device *SDp; 1536 int i; 1537 1538 hostdata->state = NCR_700_HOST_BUSY; 1539 1540 printk(KERN_ERR "scsi%d: Bus Reset detected, executing command %p, slot %p, dsp %08x[%04x]\n", 1541 host->host_no, SCp, SCp == NULL ? NULL : SCp->host_scribble, dsp, dsp - hostdata->pScript); 1542 1543 scsi_report_bus_reset(host, 0); 1544 1545 /* clear all the negotiated parameters */ 1546 __shost_for_each_device(SDp, host) 1547 NCR_700_clear_flag(SDp, ~0); 1548 1549 /* clear all the slots and their pending commands */ 1550 for(i = 0; i < NCR_700_COMMAND_SLOTS_PER_HOST; i++) { 1551 struct scsi_cmnd *SCp; 1552 struct NCR_700_command_slot *slot = 1553 &hostdata->slots[i]; 1554 1555 if(slot->state == NCR_700_SLOT_FREE) 1556 continue; 1557 1558 SCp = slot->cmnd; 1559 printk(KERN_ERR " failing command because of reset, slot %p, cmnd %p\n", 1560 slot, SCp); 1561 free_slot(slot, hostdata); 1562 SCp->host_scribble = NULL; 1563 NCR_700_set_depth(SCp->device, 0); 1564 /* NOTE: deadlock potential here: we 1565 * rely on mid-layer guarantees that 1566 * scsi_done won't try to issue the 1567 * command again otherwise we'll 1568 * deadlock on the 1569 * hostdata->state_lock */ 1570 SCp->result = DID_RESET << 16; 1571 SCp->scsi_done(SCp); 1572 } 1573 mdelay(25); 1574 NCR_700_chip_setup(host); 1575 1576 hostdata->state = NCR_700_HOST_FREE; 1577 hostdata->cmd = NULL; 1578 /* signal back if this was an eh induced reset */ 1579 if(hostdata->eh_complete != NULL) 1580 complete(hostdata->eh_complete); 1581 goto out_unlock; 1582 } else if(sstat0 & SELECTION_TIMEOUT) { 1583 DEBUG(("scsi%d: (%d:%d) selection timeout\n", 1584 host->host_no, pun, lun)); 1585 NCR_700_scsi_done(hostdata, SCp, DID_NO_CONNECT<<16); 1586 } else if(sstat0 & PHASE_MISMATCH) { 1587 struct NCR_700_command_slot *slot = (SCp == NULL) ? NULL : 1588 (struct NCR_700_command_slot *)SCp->host_scribble; 1589 1590 if(dsp == Ent_SendMessage + 8 + hostdata->pScript) { 1591 /* It wants to reply to some part of 1592 * our message */ 1593#ifdef NCR_700_DEBUG 1594 __u32 temp = NCR_700_readl(host, TEMP_REG); 1595 int count = (hostdata->script[Ent_SendMessage/4] & 0xffffff) - ((NCR_700_readl(host, DBC_REG) & 0xffffff) + NCR_700_data_residual(host)); 1596 printk("scsi%d (%d:%d) PHASE MISMATCH IN SEND MESSAGE %d remain, return %p[%04x], phase %s\n", host->host_no, pun, lun, count, (void *)temp, temp - hostdata->pScript, sbcl_to_string(NCR_700_readb(host, SBCL_REG))); 1597#endif 1598 resume_offset = hostdata->pScript + Ent_SendMessagePhaseMismatch; 1599 } else if(dsp >= to32bit(&slot->pSG[0].ins) && 1600 dsp <= to32bit(&slot->pSG[NCR_700_SG_SEGMENTS].ins)) { 1601 int data_transfer = NCR_700_readl(host, DBC_REG) & 0xffffff; 1602 int SGcount = (dsp - to32bit(&slot->pSG[0].ins))/sizeof(struct NCR_700_SG_List); 1603 int residual = NCR_700_data_residual(host); 1604 int i; 1605#ifdef NCR_700_DEBUG 1606 __u32 naddr = NCR_700_readl(host, DNAD_REG); 1607 1608 printk("scsi%d: (%d:%d) Expected phase mismatch in slot->SG[%d], transferred 0x%x\n", 1609 host->host_no, pun, lun, 1610 SGcount, data_transfer); 1611 scsi_print_command(SCp); 1612 if(residual) { 1613 printk("scsi%d: (%d:%d) Expected phase mismatch in slot->SG[%d], transferred 0x%x, residual %d\n", 1614 host->host_no, pun, lun, 1615 SGcount, data_transfer, residual); 1616 } 1617#endif 1618 data_transfer += residual; 1619 1620 if(data_transfer != 0) { 1621 int count; 1622 __u32 pAddr; 1623 1624 SGcount--; 1625 1626 count = (bS_to_cpu(slot->SG[SGcount].ins) & 0x00ffffff); 1627 DEBUG(("DATA TRANSFER MISMATCH, count = %d, transferred %d\n", count, count-data_transfer)); 1628 slot->SG[SGcount].ins &= bS_to_host(0xff000000); 1629 slot->SG[SGcount].ins |= bS_to_host(data_transfer); 1630 pAddr = bS_to_cpu(slot->SG[SGcount].pAddr); 1631 pAddr += (count - data_transfer); 1632#ifdef NCR_700_DEBUG 1633 if(pAddr != naddr) { 1634 printk("scsi%d (%d:%d) transfer mismatch pAddr=%lx, naddr=%lx, data_transfer=%d, residual=%d\n", host->host_no, pun, lun, (unsigned long)pAddr, (unsigned long)naddr, data_transfer, residual); 1635 } 1636#endif 1637 slot->SG[SGcount].pAddr = bS_to_host(pAddr); 1638 } 1639 /* set the executed moves to nops */ 1640 for(i=0; i<SGcount; i++) { 1641 slot->SG[i].ins = bS_to_host(SCRIPT_NOP); 1642 slot->SG[i].pAddr = 0; 1643 } 1644 dma_cache_sync(hostdata->dev, slot->SG, sizeof(slot->SG), DMA_TO_DEVICE); 1645 /* and pretend we disconnected after 1646 * the command phase */ 1647 resume_offset = hostdata->pScript + Ent_MsgInDuringData; 1648 /* make sure all the data is flushed */ 1649 NCR_700_flush_fifo(host); 1650 } else { 1651 __u8 sbcl = NCR_700_readb(host, SBCL_REG); 1652 printk(KERN_ERR "scsi%d: (%d:%d) phase mismatch at %04x, phase %s\n", 1653 host->host_no, pun, lun, dsp - hostdata->pScript, sbcl_to_string(sbcl)); 1654 NCR_700_internal_bus_reset(host); 1655 } 1656 1657 } else if(sstat0 & SCSI_GROSS_ERROR) { 1658 printk(KERN_ERR "scsi%d: (%d:%d) GROSS ERROR\n", 1659 host->host_no, pun, lun); 1660 NCR_700_scsi_done(hostdata, SCp, DID_ERROR<<16); 1661 } else if(sstat0 & PARITY_ERROR) { 1662 printk(KERN_ERR "scsi%d: (%d:%d) PARITY ERROR\n", 1663 host->host_no, pun, lun); 1664 NCR_700_scsi_done(hostdata, SCp, DID_ERROR<<16); 1665 } else if(dstat & SCRIPT_INT_RECEIVED) { 1666 DEBUG(("scsi%d: (%d:%d) ====>SCRIPT INTERRUPT<====\n", 1667 host->host_no, pun, lun)); 1668 resume_offset = process_script_interrupt(dsps, dsp, SCp, host, hostdata); 1669 } else if(dstat & (ILGL_INST_DETECTED)) { 1670 printk(KERN_ERR "scsi%d: (%d:%d) Illegal Instruction detected at 0x%08x[0x%x]!!!\n" 1671 " Please email James.Bottomley@HansenPartnership.com with the details\n", 1672 host->host_no, pun, lun, 1673 dsp, dsp - hostdata->pScript); 1674 NCR_700_scsi_done(hostdata, SCp, DID_ERROR<<16); 1675 } else if(dstat & (WATCH_DOG_INTERRUPT|ABORTED)) { 1676 printk(KERN_ERR "scsi%d: (%d:%d) serious DMA problem, dstat=%02x\n", 1677 host->host_no, pun, lun, dstat); 1678 NCR_700_scsi_done(hostdata, SCp, DID_ERROR<<16); 1679 } 1680 1681 1682 /* NOTE: selection interrupt processing MUST occur 1683 * after script interrupt processing to correctly cope 1684 * with the case where we process a disconnect and 1685 * then get reselected before we process the 1686 * disconnection */ 1687 if(sstat0 & SELECTED) { 1688 1689 resume_offset = process_selection(host, dsp); 1690 1691 } 1692 1693 } 1694 1695 if(resume_offset) { 1696 if(hostdata->state != NCR_700_HOST_BUSY) { 1697 printk(KERN_ERR "scsi%d: Driver error: resume at 0x%08x [0x%04x] with non busy host!\n", 1698 host->host_no, resume_offset, resume_offset - hostdata->pScript); 1699 hostdata->state = NCR_700_HOST_BUSY; 1700 } 1701 1702 DEBUG(("Attempting to resume at %x\n", resume_offset)); 1703 NCR_700_clear_fifo(host); 1704 NCR_700_writel(resume_offset, host, DSP_REG); 1705 } 1706 /* There is probably a technical no-no about this: If we're a 1707 * shared interrupt and we got this interrupt because the 1708 * other device needs servicing not us, we're still going to 1709 * check our queued commands here---of course, there shouldn't 1710 * be any outstanding.... */ 1711 if(hostdata->state == NCR_700_HOST_FREE) { 1712 int i; 1713 1714 for(i = 0; i < NCR_700_COMMAND_SLOTS_PER_HOST; i++) { 1715 /* fairness: always run the queue from the last 1716 * position we left off */ 1717 int j = (i + hostdata->saved_slot_position) 1718 % NCR_700_COMMAND_SLOTS_PER_HOST; 1719 1720 if(hostdata->slots[j].state != NCR_700_SLOT_QUEUED) 1721 continue; 1722 if(NCR_700_start_command(hostdata->slots[j].cmnd)) { 1723 DEBUG(("scsi%d: Issuing saved command slot %p, cmd %p\t\n", 1724 host->host_no, &hostdata->slots[j], 1725 hostdata->slots[j].cmnd)); 1726 hostdata->saved_slot_position = j + 1; 1727 } 1728 1729 break; 1730 } 1731 } 1732 out_unlock: 1733 spin_unlock_irqrestore(host->host_lock, flags); 1734 return IRQ_RETVAL(handled); 1735} 1736 1737STATIC int 1738NCR_700_queuecommand(struct scsi_cmnd *SCp, void (*done)(struct scsi_cmnd *)) 1739{ 1740 struct NCR_700_Host_Parameters *hostdata = 1741 (struct NCR_700_Host_Parameters *)SCp->device->host->hostdata[0]; 1742 __u32 move_ins; 1743 enum dma_data_direction direction; 1744 struct NCR_700_command_slot *slot; 1745 1746 if(hostdata->command_slot_count >= NCR_700_COMMAND_SLOTS_PER_HOST) { 1747 /* We're over our allocation, this should never happen 1748 * since we report the max allocation to the mid layer */ 1749 printk(KERN_WARNING "scsi%d: Command depth has gone over queue depth\n", SCp->device->host->host_no); 1750 return 1; 1751 } 1752 /* check for untagged commands. We cannot have any outstanding 1753 * commands if we accept them. Commands could be untagged because: 1754 * 1755 * - The tag negotiated bitmap is clear 1756 * - The blk layer sent and untagged command 1757 */ 1758 if(NCR_700_get_depth(SCp->device) != 0 1759 && (!(hostdata->tag_negotiated & (1<<scmd_id(SCp))) 1760 || !blk_rq_tagged(SCp->request))) { 1761 CDEBUG(KERN_ERR, SCp, "has non zero depth %d\n", 1762 NCR_700_get_depth(SCp->device)); 1763 return SCSI_MLQUEUE_DEVICE_BUSY; 1764 } 1765 if(NCR_700_get_depth(SCp->device) >= SCp->device->queue_depth) { 1766 CDEBUG(KERN_ERR, SCp, "has max tag depth %d\n", 1767 NCR_700_get_depth(SCp->device)); 1768 return SCSI_MLQUEUE_DEVICE_BUSY; 1769 } 1770 NCR_700_set_depth(SCp->device, NCR_700_get_depth(SCp->device) + 1); 1771 1772 /* begin the command here */ 1773 /* no need to check for NULL, test for command_slot_count above 1774 * ensures a slot is free */ 1775 slot = find_empty_slot(hostdata); 1776 1777 slot->cmnd = SCp; 1778 1779 SCp->scsi_done = done; 1780 SCp->host_scribble = (unsigned char *)slot; 1781 SCp->SCp.ptr = NULL; 1782 SCp->SCp.buffer = NULL; 1783 1784#ifdef NCR_700_DEBUG 1785 printk("53c700: scsi%d, command ", SCp->device->host->host_no); 1786 scsi_print_command(SCp); 1787#endif 1788 if(blk_rq_tagged(SCp->request) 1789 && (hostdata->tag_negotiated &(1<<scmd_id(SCp))) == 0 1790 && NCR_700_get_tag_neg_state(SCp->device) == NCR_700_START_TAG_NEGOTIATION) { 1791 scmd_printk(KERN_ERR, SCp, "Enabling Tag Command Queuing\n"); 1792 hostdata->tag_negotiated |= (1<<scmd_id(SCp)); 1793 NCR_700_set_tag_neg_state(SCp->device, NCR_700_DURING_TAG_NEGOTIATION); 1794 } 1795 1796 if(!blk_rq_tagged(SCp->request) 1797 && (hostdata->tag_negotiated &(1<<scmd_id(SCp)))) { 1798 scmd_printk(KERN_INFO, SCp, "Disabling Tag Command Queuing\n"); 1799 hostdata->tag_negotiated &= ~(1<<scmd_id(SCp)); 1800 } 1801 1802 if((hostdata->tag_negotiated &(1<<scmd_id(SCp))) 1803 && scsi_get_tag_type(SCp->device)) { 1804 slot->tag = SCp->request->tag; 1805 CDEBUG(KERN_DEBUG, SCp, "sending out tag %d, slot %p\n", 1806 slot->tag, slot); 1807 } else { 1808 slot->tag = SCSI_NO_TAG; 1809 /* must populate current_cmnd for scsi_find_tag to work */ 1810 SCp->device->current_cmnd = SCp; 1811 } 1812 /* sanity check: some of the commands generated by the mid-layer 1813 * have an eccentric idea of their sc_data_direction */ 1814 if(!scsi_sg_count(SCp) && !scsi_bufflen(SCp) && 1815 SCp->sc_data_direction != DMA_NONE) { 1816#ifdef NCR_700_DEBUG 1817 printk("53c700: Command"); 1818 scsi_print_command(SCp); 1819 printk("Has wrong data direction %d\n", SCp->sc_data_direction); 1820#endif 1821 SCp->sc_data_direction = DMA_NONE; 1822 } 1823 1824 switch (SCp->cmnd[0]) { 1825 case REQUEST_SENSE: 1826 /* clear the internal sense magic */ 1827 SCp->cmnd[6] = 0; 1828 /* fall through */ 1829 default: 1830 /* OK, get it from the command */ 1831 switch(SCp->sc_data_direction) { 1832 case DMA_BIDIRECTIONAL: 1833 default: 1834 printk(KERN_ERR "53c700: Unknown command for data direction "); 1835 scsi_print_command(SCp); 1836 1837 move_ins = 0; 1838 break; 1839 case DMA_NONE: 1840 move_ins = 0; 1841 break; 1842 case DMA_FROM_DEVICE: 1843 move_ins = SCRIPT_MOVE_DATA_IN; 1844 break; 1845 case DMA_TO_DEVICE: 1846 move_ins = SCRIPT_MOVE_DATA_OUT; 1847 break; 1848 } 1849 } 1850 1851 /* now build the scatter gather list */ 1852 direction = SCp->sc_data_direction; 1853 if(move_ins != 0) { 1854 int i; 1855 int sg_count; 1856 dma_addr_t vPtr = 0; 1857 struct scatterlist *sg; 1858 __u32 count = 0; 1859 1860 sg_count = scsi_dma_map(SCp); 1861 BUG_ON(sg_count < 0); 1862 1863 scsi_for_each_sg(SCp, sg, sg_count, i) { 1864 vPtr = sg_dma_address(sg); 1865 count = sg_dma_len(sg); 1866 1867 slot->SG[i].ins = bS_to_host(move_ins | count); 1868 DEBUG((" scatter block %d: move %d[%08x] from 0x%lx\n", 1869 i, count, slot->SG[i].ins, (unsigned long)vPtr)); 1870 slot->SG[i].pAddr = bS_to_host(vPtr); 1871 } 1872 slot->SG[i].ins = bS_to_host(SCRIPT_RETURN); 1873 slot->SG[i].pAddr = 0; 1874 dma_cache_sync(hostdata->dev, slot->SG, sizeof(slot->SG), DMA_TO_DEVICE); 1875 DEBUG((" SETTING %08lx to %x\n", 1876 (&slot->pSG[i].ins), 1877 slot->SG[i].ins)); 1878 } 1879 slot->resume_offset = 0; 1880 slot->pCmd = dma_map_single(hostdata->dev, SCp->cmnd, 1881 MAX_COMMAND_SIZE, DMA_TO_DEVICE); 1882 NCR_700_start_command(SCp); 1883 return 0; 1884} 1885 1886STATIC int 1887NCR_700_abort(struct scsi_cmnd * SCp) 1888{ 1889 struct NCR_700_command_slot *slot; 1890 1891 scmd_printk(KERN_INFO, SCp, 1892 "New error handler wants to abort command\n\t"); 1893 scsi_print_command(SCp); 1894 1895 slot = (struct NCR_700_command_slot *)SCp->host_scribble; 1896 1897 if(slot == NULL) 1898 /* no outstanding command to abort */ 1899 return SUCCESS; 1900 if(SCp->cmnd[0] == TEST_UNIT_READY) { 1901 NCR_700_internal_bus_reset(SCp->device->host); 1902 /* still drop through and return failed */ 1903 } 1904 return FAILED; 1905 1906} 1907 1908STATIC int 1909NCR_700_bus_reset(struct scsi_cmnd * SCp) 1910{ 1911 DECLARE_COMPLETION_ONSTACK(complete); 1912 struct NCR_700_Host_Parameters *hostdata = 1913 (struct NCR_700_Host_Parameters *)SCp->device->host->hostdata[0]; 1914 1915 scmd_printk(KERN_INFO, SCp, 1916 "New error handler wants BUS reset, cmd %p\n\t", SCp); 1917 scsi_print_command(SCp); 1918 1919 /* In theory, eh_complete should always be null because the 1920 * eh is single threaded, but just in case we're handling a 1921 * reset via sg or something */ 1922 spin_lock_irq(SCp->device->host->host_lock); 1923 while (hostdata->eh_complete != NULL) { 1924 spin_unlock_irq(SCp->device->host->host_lock); 1925 msleep_interruptible(100); 1926 spin_lock_irq(SCp->device->host->host_lock); 1927 } 1928 1929 hostdata->eh_complete = &complete; 1930 NCR_700_internal_bus_reset(SCp->device->host); 1931 1932 spin_unlock_irq(SCp->device->host->host_lock); 1933 wait_for_completion(&complete); 1934 spin_lock_irq(SCp->device->host->host_lock); 1935 1936 hostdata->eh_complete = NULL; 1937 /* Revalidate the transport parameters of the failing device */ 1938 if(hostdata->fast) 1939 spi_schedule_dv_device(SCp->device); 1940 1941 spin_unlock_irq(SCp->device->host->host_lock); 1942 return SUCCESS; 1943} 1944 1945STATIC int 1946NCR_700_host_reset(struct scsi_cmnd * SCp) 1947{ 1948 scmd_printk(KERN_INFO, SCp, "New error handler wants HOST reset\n\t"); 1949 scsi_print_command(SCp); 1950 1951 spin_lock_irq(SCp->device->host->host_lock); 1952 1953 NCR_700_internal_bus_reset(SCp->device->host); 1954 NCR_700_chip_reset(SCp->device->host); 1955 1956 spin_unlock_irq(SCp->device->host->host_lock); 1957 1958 return SUCCESS; 1959} 1960 1961STATIC void 1962NCR_700_set_period(struct scsi_target *STp, int period) 1963{ 1964 struct Scsi_Host *SHp = dev_to_shost(STp->dev.parent); 1965 struct NCR_700_Host_Parameters *hostdata = 1966 (struct NCR_700_Host_Parameters *)SHp->hostdata[0]; 1967 1968 if(!hostdata->fast) 1969 return; 1970 1971 if(period < hostdata->min_period) 1972 period = hostdata->min_period; 1973 1974 spi_period(STp) = period; 1975 spi_flags(STp) &= ~(NCR_700_DEV_NEGOTIATED_SYNC | 1976 NCR_700_DEV_BEGIN_SYNC_NEGOTIATION); 1977 spi_flags(STp) |= NCR_700_DEV_PRINT_SYNC_NEGOTIATION; 1978} 1979 1980STATIC void 1981NCR_700_set_offset(struct scsi_target *STp, int offset) 1982{ 1983 struct Scsi_Host *SHp = dev_to_shost(STp->dev.parent); 1984 struct NCR_700_Host_Parameters *hostdata = 1985 (struct NCR_700_Host_Parameters *)SHp->hostdata[0]; 1986 int max_offset = hostdata->chip710 1987 ? NCR_710_MAX_OFFSET : NCR_700_MAX_OFFSET; 1988 1989 if(!hostdata->fast) 1990 return; 1991 1992 if(offset > max_offset) 1993 offset = max_offset; 1994 1995 /* if we're currently async, make sure the period is reasonable */ 1996 if(spi_offset(STp) == 0 && (spi_period(STp) < hostdata->min_period || 1997 spi_period(STp) > 0xff)) 1998 spi_period(STp) = hostdata->min_period; 1999 2000 spi_offset(STp) = offset; 2001 spi_flags(STp) &= ~(NCR_700_DEV_NEGOTIATED_SYNC | 2002 NCR_700_DEV_BEGIN_SYNC_NEGOTIATION); 2003 spi_flags(STp) |= NCR_700_DEV_PRINT_SYNC_NEGOTIATION; 2004} 2005 2006STATIC int 2007NCR_700_slave_alloc(struct scsi_device *SDp) 2008{ 2009 SDp->hostdata = kzalloc(sizeof(struct NCR_700_Device_Parameters), 2010 GFP_KERNEL); 2011 2012 if (!SDp->hostdata) 2013 return -ENOMEM; 2014 2015 return 0; 2016} 2017 2018STATIC int 2019NCR_700_slave_configure(struct scsi_device *SDp) 2020{ 2021 struct NCR_700_Host_Parameters *hostdata = 2022 (struct NCR_700_Host_Parameters *)SDp->host->hostdata[0]; 2023 2024 /* to do here: allocate memory; build a queue_full list */ 2025 if(SDp->tagged_supported) { 2026 scsi_set_tag_type(SDp, MSG_ORDERED_TAG); 2027 scsi_activate_tcq(SDp, NCR_700_DEFAULT_TAGS); 2028 NCR_700_set_tag_neg_state(SDp, NCR_700_START_TAG_NEGOTIATION); 2029 } else { 2030 /* initialise to default depth */ 2031 scsi_adjust_queue_depth(SDp, 0, SDp->host->cmd_per_lun); 2032 } 2033 if(hostdata->fast) { 2034 /* Find the correct offset and period via domain validation */ 2035 if (!spi_initial_dv(SDp->sdev_target)) 2036 spi_dv_device(SDp); 2037 } else { 2038 spi_offset(SDp->sdev_target) = 0; 2039 spi_period(SDp->sdev_target) = 0; 2040 } 2041 return 0; 2042} 2043 2044STATIC void 2045NCR_700_slave_destroy(struct scsi_device *SDp) 2046{ 2047 kfree(SDp->hostdata); 2048 SDp->hostdata = NULL; 2049} 2050 2051static int 2052NCR_700_change_queue_depth(struct scsi_device *SDp, int depth, int reason) 2053{ 2054 if (reason != SCSI_QDEPTH_DEFAULT) 2055 return -EOPNOTSUPP; 2056 2057 if (depth > NCR_700_MAX_TAGS) 2058 depth = NCR_700_MAX_TAGS; 2059 2060 scsi_adjust_queue_depth(SDp, scsi_get_tag_type(SDp), depth); 2061 return depth; 2062} 2063 2064static int NCR_700_change_queue_type(struct scsi_device *SDp, int tag_type) 2065{ 2066 int change_tag = ((tag_type ==0 && scsi_get_tag_type(SDp) != 0) 2067 || (tag_type != 0 && scsi_get_tag_type(SDp) == 0)); 2068 struct NCR_700_Host_Parameters *hostdata = 2069 (struct NCR_700_Host_Parameters *)SDp->host->hostdata[0]; 2070 2071 scsi_set_tag_type(SDp, tag_type); 2072 2073 /* We have a global (per target) flag to track whether TCQ is 2074 * enabled, so we'll be turning it off for the entire target here. 2075 * our tag algorithm will fail if we mix tagged and untagged commands, 2076 * so quiesce the device before doing this */ 2077 if (change_tag) 2078 scsi_target_quiesce(SDp->sdev_target); 2079 2080 if (!tag_type) { 2081 /* shift back to the default unqueued number of commands 2082 * (the user can still raise this) */ 2083 scsi_deactivate_tcq(SDp, SDp->host->cmd_per_lun); 2084 hostdata->tag_negotiated &= ~(1 << sdev_id(SDp)); 2085 } else { 2086 /* Here, we cleared the negotiation flag above, so this 2087 * will force the driver to renegotiate */ 2088 scsi_activate_tcq(SDp, SDp->queue_depth); 2089 if (change_tag) 2090 NCR_700_set_tag_neg_state(SDp, NCR_700_START_TAG_NEGOTIATION); 2091 } 2092 if (change_tag) 2093 scsi_target_resume(SDp->sdev_target); 2094 2095 return tag_type; 2096} 2097 2098static ssize_t 2099NCR_700_show_active_tags(struct device *dev, struct device_attribute *attr, char *buf) 2100{ 2101 struct scsi_device *SDp = to_scsi_device(dev); 2102 2103 return snprintf(buf, 20, "%d\n", NCR_700_get_depth(SDp)); 2104} 2105 2106static struct device_attribute NCR_700_active_tags_attr = { 2107 .attr = { 2108 .name = "active_tags", 2109 .mode = S_IRUGO, 2110 }, 2111 .show = NCR_700_show_active_tags, 2112}; 2113 2114STATIC struct device_attribute *NCR_700_dev_attrs[] = { 2115 &NCR_700_active_tags_attr, 2116 NULL, 2117}; 2118 2119EXPORT_SYMBOL(NCR_700_detect); 2120EXPORT_SYMBOL(NCR_700_release); 2121EXPORT_SYMBOL(NCR_700_intr); 2122 2123static struct spi_function_template NCR_700_transport_functions = { 2124 .set_period = NCR_700_set_period, 2125 .show_period = 1, 2126 .set_offset = NCR_700_set_offset, 2127 .show_offset = 1, 2128}; 2129 2130static int __init NCR_700_init(void) 2131{ 2132 NCR_700_transport_template = spi_attach_transport(&NCR_700_transport_functions); 2133 if(!NCR_700_transport_template) 2134 return -ENODEV; 2135 return 0; 2136} 2137 2138static void __exit NCR_700_exit(void) 2139{ 2140 spi_release_transport(NCR_700_transport_template); 2141} 2142 2143module_init(NCR_700_init); 2144module_exit(NCR_700_exit); 2145