1/* 2 * sata_sx4.c - Promise SATA 3 * 4 * Maintained by: Jeff Garzik <jgarzik@pobox.com> 5 * Please ALWAYS copy linux-ide@vger.kernel.org 6 * on emails. 7 * 8 * Copyright 2003-2004 Red Hat, Inc. 9 * 10 * 11 * This program is free software; you can redistribute it and/or modify 12 * it under the terms of the GNU General Public License as published by 13 * the Free Software Foundation; either version 2, or (at your option) 14 * any later version. 15 * 16 * This program is distributed in the hope that it will be useful, 17 * but WITHOUT ANY WARRANTY; without even the implied warranty of 18 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the 19 * GNU General Public License for more details. 20 * 21 * You should have received a copy of the GNU General Public License 22 * along with this program; see the file COPYING. If not, write to 23 * the Free Software Foundation, 675 Mass Ave, Cambridge, MA 02139, USA. 24 * 25 * 26 * libata documentation is available via 'make {ps|pdf}docs', 27 * as Documentation/DocBook/libata.* 28 * 29 * Hardware documentation available under NDA. 30 * 31 */ 32 33/* 34 Theory of operation 35 ------------------- 36 37 The SX4 (PDC20621) chip features a single Host DMA (HDMA) copy 38 engine, DIMM memory, and four ATA engines (one per SATA port). 39 Data is copied to/from DIMM memory by the HDMA engine, before 40 handing off to one (or more) of the ATA engines. The ATA 41 engines operate solely on DIMM memory. 42 43 The SX4 behaves like a PATA chip, with no SATA controls or 44 knowledge whatsoever, leading to the presumption that 45 PATA<->SATA bridges exist on SX4 boards, external to the 46 PDC20621 chip itself. 47 48 The chip is quite capable, supporting an XOR engine and linked 49 hardware commands (permits a string to transactions to be 50 submitted and waited-on as a single unit), and an optional 51 microprocessor. 52 53 The limiting factor is largely software. This Linux driver was 54 written to multiplex the single HDMA engine to copy disk 55 transactions into a fixed DIMM memory space, from where an ATA 56 engine takes over. As a result, each WRITE looks like this: 57 58 submit HDMA packet to hardware 59 hardware copies data from system memory to DIMM 60 hardware raises interrupt 61 62 submit ATA packet to hardware 63 hardware executes ATA WRITE command, w/ data in DIMM 64 hardware raises interrupt 65 66 and each READ looks like this: 67 68 submit ATA packet to hardware 69 hardware executes ATA READ command, w/ data in DIMM 70 hardware raises interrupt 71 72 submit HDMA packet to hardware 73 hardware copies data from DIMM to system memory 74 hardware raises interrupt 75 76 This is a very slow, lock-step way of doing things that can 77 certainly be improved by motivated kernel hackers. 78 79 */ 80 81#include <linux/kernel.h> 82#include <linux/module.h> 83#include <linux/pci.h> 84#include <linux/slab.h> 85#include <linux/init.h> 86#include <linux/blkdev.h> 87#include <linux/delay.h> 88#include <linux/interrupt.h> 89#include <linux/device.h> 90#include <scsi/scsi_host.h> 91#include <scsi/scsi_cmnd.h> 92#include <linux/libata.h> 93#include "sata_promise.h" 94 95#define DRV_NAME "sata_sx4" 96#define DRV_VERSION "0.12" 97 98 99enum { 100 PDC_MMIO_BAR = 3, 101 PDC_DIMM_BAR = 4, 102 103 PDC_PRD_TBL = 0x44, /* Direct command DMA table addr */ 104 105 PDC_PKT_SUBMIT = 0x40, /* Command packet pointer addr */ 106 PDC_HDMA_PKT_SUBMIT = 0x100, /* Host DMA packet pointer addr */ 107 PDC_INT_SEQMASK = 0x40, /* Mask of asserted SEQ INTs */ 108 PDC_HDMA_CTLSTAT = 0x12C, /* Host DMA control / status */ 109 110 PDC_CTLSTAT = 0x60, /* IDEn control / status */ 111 112 PDC_20621_SEQCTL = 0x400, 113 PDC_20621_SEQMASK = 0x480, 114 PDC_20621_GENERAL_CTL = 0x484, 115 PDC_20621_PAGE_SIZE = (32 * 1024), 116 117 /* chosen, not constant, values; we design our own DIMM mem map */ 118 PDC_20621_DIMM_WINDOW = 0x0C, /* page# for 32K DIMM window */ 119 PDC_20621_DIMM_BASE = 0x00200000, 120 PDC_20621_DIMM_DATA = (64 * 1024), 121 PDC_DIMM_DATA_STEP = (256 * 1024), 122 PDC_DIMM_WINDOW_STEP = (8 * 1024), 123 PDC_DIMM_HOST_PRD = (6 * 1024), 124 PDC_DIMM_HOST_PKT = (128 * 0), 125 PDC_DIMM_HPKT_PRD = (128 * 1), 126 PDC_DIMM_ATA_PKT = (128 * 2), 127 PDC_DIMM_APKT_PRD = (128 * 3), 128 PDC_DIMM_HEADER_SZ = PDC_DIMM_APKT_PRD + 128, 129 PDC_PAGE_WINDOW = 0x40, 130 PDC_PAGE_DATA = PDC_PAGE_WINDOW + 131 (PDC_20621_DIMM_DATA / PDC_20621_PAGE_SIZE), 132 PDC_PAGE_SET = PDC_DIMM_DATA_STEP / PDC_20621_PAGE_SIZE, 133 134 PDC_CHIP0_OFS = 0xC0000, /* offset of chip #0 */ 135 136 PDC_20621_ERR_MASK = (1<<19) | (1<<20) | (1<<21) | (1<<22) | 137 (1<<23), 138 139 board_20621 = 0, /* FastTrak S150 SX4 */ 140 141 PDC_MASK_INT = (1 << 10), /* HDMA/ATA mask int */ 142 PDC_RESET = (1 << 11), /* HDMA/ATA reset */ 143 PDC_DMA_ENABLE = (1 << 7), /* DMA start/stop */ 144 145 PDC_MAX_HDMA = 32, 146 PDC_HDMA_Q_MASK = (PDC_MAX_HDMA - 1), 147 148 PDC_DIMM0_SPD_DEV_ADDRESS = 0x50, 149 PDC_DIMM1_SPD_DEV_ADDRESS = 0x51, 150 PDC_I2C_CONTROL = 0x48, 151 PDC_I2C_ADDR_DATA = 0x4C, 152 PDC_DIMM0_CONTROL = 0x80, 153 PDC_DIMM1_CONTROL = 0x84, 154 PDC_SDRAM_CONTROL = 0x88, 155 PDC_I2C_WRITE = 0, /* master -> slave */ 156 PDC_I2C_READ = (1 << 6), /* master <- slave */ 157 PDC_I2C_START = (1 << 7), /* start I2C proto */ 158 PDC_I2C_MASK_INT = (1 << 5), /* mask I2C interrupt */ 159 PDC_I2C_COMPLETE = (1 << 16), /* I2C normal compl. */ 160 PDC_I2C_NO_ACK = (1 << 20), /* slave no-ack addr */ 161 PDC_DIMM_SPD_SUBADDRESS_START = 0x00, 162 PDC_DIMM_SPD_SUBADDRESS_END = 0x7F, 163 PDC_DIMM_SPD_ROW_NUM = 3, 164 PDC_DIMM_SPD_COLUMN_NUM = 4, 165 PDC_DIMM_SPD_MODULE_ROW = 5, 166 PDC_DIMM_SPD_TYPE = 11, 167 PDC_DIMM_SPD_FRESH_RATE = 12, 168 PDC_DIMM_SPD_BANK_NUM = 17, 169 PDC_DIMM_SPD_CAS_LATENCY = 18, 170 PDC_DIMM_SPD_ATTRIBUTE = 21, 171 PDC_DIMM_SPD_ROW_PRE_CHARGE = 27, 172 PDC_DIMM_SPD_ROW_ACTIVE_DELAY = 28, 173 PDC_DIMM_SPD_RAS_CAS_DELAY = 29, 174 PDC_DIMM_SPD_ACTIVE_PRECHARGE = 30, 175 PDC_DIMM_SPD_SYSTEM_FREQ = 126, 176 PDC_CTL_STATUS = 0x08, 177 PDC_DIMM_WINDOW_CTLR = 0x0C, 178 PDC_TIME_CONTROL = 0x3C, 179 PDC_TIME_PERIOD = 0x40, 180 PDC_TIME_COUNTER = 0x44, 181 PDC_GENERAL_CTLR = 0x484, 182 PCI_PLL_INIT = 0x8A531824, 183 PCI_X_TCOUNT = 0xEE1E5CFF, 184 185 /* PDC_TIME_CONTROL bits */ 186 PDC_TIMER_BUZZER = (1 << 10), 187 PDC_TIMER_MODE_PERIODIC = 0, /* bits 9:8 == 00 */ 188 PDC_TIMER_MODE_ONCE = (1 << 8), /* bits 9:8 == 01 */ 189 PDC_TIMER_ENABLE = (1 << 7), 190 PDC_TIMER_MASK_INT = (1 << 5), 191 PDC_TIMER_SEQ_MASK = 0x1f, /* SEQ ID for timer */ 192 PDC_TIMER_DEFAULT = PDC_TIMER_MODE_ONCE | 193 PDC_TIMER_ENABLE | 194 PDC_TIMER_MASK_INT, 195}; 196 197#define ECC_ERASE_BUF_SZ (128 * 1024) 198 199struct pdc_port_priv { 200 u8 dimm_buf[(ATA_PRD_SZ * ATA_MAX_PRD) + 512]; 201 u8 *pkt; 202 dma_addr_t pkt_dma; 203}; 204 205struct pdc_host_priv { 206 unsigned int doing_hdma; 207 unsigned int hdma_prod; 208 unsigned int hdma_cons; 209 struct { 210 struct ata_queued_cmd *qc; 211 unsigned int seq; 212 unsigned long pkt_ofs; 213 } hdma[32]; 214}; 215 216 217static int pdc_sata_init_one(struct pci_dev *pdev, const struct pci_device_id *ent); 218static void pdc_error_handler(struct ata_port *ap); 219static void pdc_freeze(struct ata_port *ap); 220static void pdc_thaw(struct ata_port *ap); 221static int pdc_port_start(struct ata_port *ap); 222static void pdc20621_qc_prep(struct ata_queued_cmd *qc); 223static void pdc_tf_load_mmio(struct ata_port *ap, const struct ata_taskfile *tf); 224static void pdc_exec_command_mmio(struct ata_port *ap, const struct ata_taskfile *tf); 225static unsigned int pdc20621_dimm_init(struct ata_host *host); 226static int pdc20621_detect_dimm(struct ata_host *host); 227static unsigned int pdc20621_i2c_read(struct ata_host *host, 228 u32 device, u32 subaddr, u32 *pdata); 229static int pdc20621_prog_dimm0(struct ata_host *host); 230static unsigned int pdc20621_prog_dimm_global(struct ata_host *host); 231#ifdef ATA_VERBOSE_DEBUG 232static void pdc20621_get_from_dimm(struct ata_host *host, 233 void *psource, u32 offset, u32 size); 234#endif 235static void pdc20621_put_to_dimm(struct ata_host *host, 236 void *psource, u32 offset, u32 size); 237static void pdc20621_irq_clear(struct ata_port *ap); 238static unsigned int pdc20621_qc_issue(struct ata_queued_cmd *qc); 239static int pdc_softreset(struct ata_link *link, unsigned int *class, 240 unsigned long deadline); 241static void pdc_post_internal_cmd(struct ata_queued_cmd *qc); 242static int pdc_check_atapi_dma(struct ata_queued_cmd *qc); 243 244 245static struct scsi_host_template pdc_sata_sht = { 246 ATA_BASE_SHT(DRV_NAME), 247 .sg_tablesize = LIBATA_MAX_PRD, 248 .dma_boundary = ATA_DMA_BOUNDARY, 249}; 250 251/* TODO: inherit from base port_ops after converting to new EH */ 252static struct ata_port_operations pdc_20621_ops = { 253 .inherits = &ata_sff_port_ops, 254 255 .check_atapi_dma = pdc_check_atapi_dma, 256 .qc_prep = pdc20621_qc_prep, 257 .qc_issue = pdc20621_qc_issue, 258 259 .freeze = pdc_freeze, 260 .thaw = pdc_thaw, 261 .softreset = pdc_softreset, 262 .error_handler = pdc_error_handler, 263 .lost_interrupt = ATA_OP_NULL, 264 .post_internal_cmd = pdc_post_internal_cmd, 265 266 .port_start = pdc_port_start, 267 268 .sff_tf_load = pdc_tf_load_mmio, 269 .sff_exec_command = pdc_exec_command_mmio, 270 .sff_irq_clear = pdc20621_irq_clear, 271}; 272 273static const struct ata_port_info pdc_port_info[] = { 274 /* board_20621 */ 275 { 276 .flags = ATA_FLAG_SATA | ATA_FLAG_NO_LEGACY | 277 ATA_FLAG_SRST | ATA_FLAG_MMIO | 278 ATA_FLAG_NO_ATAPI | ATA_FLAG_PIO_POLLING, 279 .pio_mask = ATA_PIO4, 280 .mwdma_mask = ATA_MWDMA2, 281 .udma_mask = ATA_UDMA6, 282 .port_ops = &pdc_20621_ops, 283 }, 284 285}; 286 287static const struct pci_device_id pdc_sata_pci_tbl[] = { 288 { PCI_VDEVICE(PROMISE, 0x6622), board_20621 }, 289 290 { } /* terminate list */ 291}; 292 293static struct pci_driver pdc_sata_pci_driver = { 294 .name = DRV_NAME, 295 .id_table = pdc_sata_pci_tbl, 296 .probe = pdc_sata_init_one, 297 .remove = ata_pci_remove_one, 298}; 299 300 301static int pdc_port_start(struct ata_port *ap) 302{ 303 struct device *dev = ap->host->dev; 304 struct pdc_port_priv *pp; 305 306 pp = devm_kzalloc(dev, sizeof(*pp), GFP_KERNEL); 307 if (!pp) 308 return -ENOMEM; 309 310 pp->pkt = dmam_alloc_coherent(dev, 128, &pp->pkt_dma, GFP_KERNEL); 311 if (!pp->pkt) 312 return -ENOMEM; 313 314 ap->private_data = pp; 315 316 return 0; 317} 318 319static inline void pdc20621_ata_sg(struct ata_taskfile *tf, u8 *buf, 320 unsigned int portno, 321 unsigned int total_len) 322{ 323 u32 addr; 324 unsigned int dw = PDC_DIMM_APKT_PRD >> 2; 325 __le32 *buf32 = (__le32 *) buf; 326 327 /* output ATA packet S/G table */ 328 addr = PDC_20621_DIMM_BASE + PDC_20621_DIMM_DATA + 329 (PDC_DIMM_DATA_STEP * portno); 330 VPRINTK("ATA sg addr 0x%x, %d\n", addr, addr); 331 buf32[dw] = cpu_to_le32(addr); 332 buf32[dw + 1] = cpu_to_le32(total_len | ATA_PRD_EOT); 333 334 VPRINTK("ATA PSG @ %x == (0x%x, 0x%x)\n", 335 PDC_20621_DIMM_BASE + 336 (PDC_DIMM_WINDOW_STEP * portno) + 337 PDC_DIMM_APKT_PRD, 338 buf32[dw], buf32[dw + 1]); 339} 340 341static inline void pdc20621_host_sg(struct ata_taskfile *tf, u8 *buf, 342 unsigned int portno, 343 unsigned int total_len) 344{ 345 u32 addr; 346 unsigned int dw = PDC_DIMM_HPKT_PRD >> 2; 347 __le32 *buf32 = (__le32 *) buf; 348 349 /* output Host DMA packet S/G table */ 350 addr = PDC_20621_DIMM_BASE + PDC_20621_DIMM_DATA + 351 (PDC_DIMM_DATA_STEP * portno); 352 353 buf32[dw] = cpu_to_le32(addr); 354 buf32[dw + 1] = cpu_to_le32(total_len | ATA_PRD_EOT); 355 356 VPRINTK("HOST PSG @ %x == (0x%x, 0x%x)\n", 357 PDC_20621_DIMM_BASE + 358 (PDC_DIMM_WINDOW_STEP * portno) + 359 PDC_DIMM_HPKT_PRD, 360 buf32[dw], buf32[dw + 1]); 361} 362 363static inline unsigned int pdc20621_ata_pkt(struct ata_taskfile *tf, 364 unsigned int devno, u8 *buf, 365 unsigned int portno) 366{ 367 unsigned int i, dw; 368 __le32 *buf32 = (__le32 *) buf; 369 u8 dev_reg; 370 371 unsigned int dimm_sg = PDC_20621_DIMM_BASE + 372 (PDC_DIMM_WINDOW_STEP * portno) + 373 PDC_DIMM_APKT_PRD; 374 VPRINTK("ENTER, dimm_sg == 0x%x, %d\n", dimm_sg, dimm_sg); 375 376 i = PDC_DIMM_ATA_PKT; 377 378 /* 379 * Set up ATA packet 380 */ 381 if ((tf->protocol == ATA_PROT_DMA) && (!(tf->flags & ATA_TFLAG_WRITE))) 382 buf[i++] = PDC_PKT_READ; 383 else if (tf->protocol == ATA_PROT_NODATA) 384 buf[i++] = PDC_PKT_NODATA; 385 else 386 buf[i++] = 0; 387 buf[i++] = 0; /* reserved */ 388 buf[i++] = portno + 1; /* seq. id */ 389 buf[i++] = 0xff; /* delay seq. id */ 390 391 /* dimm dma S/G, and next-pkt */ 392 dw = i >> 2; 393 if (tf->protocol == ATA_PROT_NODATA) 394 buf32[dw] = 0; 395 else 396 buf32[dw] = cpu_to_le32(dimm_sg); 397 buf32[dw + 1] = 0; 398 i += 8; 399 400 if (devno == 0) 401 dev_reg = ATA_DEVICE_OBS; 402 else 403 dev_reg = ATA_DEVICE_OBS | ATA_DEV1; 404 405 /* select device */ 406 buf[i++] = (1 << 5) | PDC_PKT_CLEAR_BSY | ATA_REG_DEVICE; 407 buf[i++] = dev_reg; 408 409 /* device control register */ 410 buf[i++] = (1 << 5) | PDC_REG_DEVCTL; 411 buf[i++] = tf->ctl; 412 413 return i; 414} 415 416static inline void pdc20621_host_pkt(struct ata_taskfile *tf, u8 *buf, 417 unsigned int portno) 418{ 419 unsigned int dw; 420 u32 tmp; 421 __le32 *buf32 = (__le32 *) buf; 422 423 unsigned int host_sg = PDC_20621_DIMM_BASE + 424 (PDC_DIMM_WINDOW_STEP * portno) + 425 PDC_DIMM_HOST_PRD; 426 unsigned int dimm_sg = PDC_20621_DIMM_BASE + 427 (PDC_DIMM_WINDOW_STEP * portno) + 428 PDC_DIMM_HPKT_PRD; 429 VPRINTK("ENTER, dimm_sg == 0x%x, %d\n", dimm_sg, dimm_sg); 430 VPRINTK("host_sg == 0x%x, %d\n", host_sg, host_sg); 431 432 dw = PDC_DIMM_HOST_PKT >> 2; 433 434 /* 435 * Set up Host DMA packet 436 */ 437 if ((tf->protocol == ATA_PROT_DMA) && (!(tf->flags & ATA_TFLAG_WRITE))) 438 tmp = PDC_PKT_READ; 439 else 440 tmp = 0; 441 tmp |= ((portno + 1 + 4) << 16); /* seq. id */ 442 tmp |= (0xff << 24); /* delay seq. id */ 443 buf32[dw + 0] = cpu_to_le32(tmp); 444 buf32[dw + 1] = cpu_to_le32(host_sg); 445 buf32[dw + 2] = cpu_to_le32(dimm_sg); 446 buf32[dw + 3] = 0; 447 448 VPRINTK("HOST PKT @ %x == (0x%x 0x%x 0x%x 0x%x)\n", 449 PDC_20621_DIMM_BASE + (PDC_DIMM_WINDOW_STEP * portno) + 450 PDC_DIMM_HOST_PKT, 451 buf32[dw + 0], 452 buf32[dw + 1], 453 buf32[dw + 2], 454 buf32[dw + 3]); 455} 456 457static void pdc20621_dma_prep(struct ata_queued_cmd *qc) 458{ 459 struct scatterlist *sg; 460 struct ata_port *ap = qc->ap; 461 struct pdc_port_priv *pp = ap->private_data; 462 void __iomem *mmio = ap->host->iomap[PDC_MMIO_BAR]; 463 void __iomem *dimm_mmio = ap->host->iomap[PDC_DIMM_BAR]; 464 unsigned int portno = ap->port_no; 465 unsigned int i, si, idx, total_len = 0, sgt_len; 466 __le32 *buf = (__le32 *) &pp->dimm_buf[PDC_DIMM_HEADER_SZ]; 467 468 WARN_ON(!(qc->flags & ATA_QCFLAG_DMAMAP)); 469 470 VPRINTK("ata%u: ENTER\n", ap->print_id); 471 472 /* hard-code chip #0 */ 473 mmio += PDC_CHIP0_OFS; 474 475 /* 476 * Build S/G table 477 */ 478 idx = 0; 479 for_each_sg(qc->sg, sg, qc->n_elem, si) { 480 buf[idx++] = cpu_to_le32(sg_dma_address(sg)); 481 buf[idx++] = cpu_to_le32(sg_dma_len(sg)); 482 total_len += sg_dma_len(sg); 483 } 484 buf[idx - 1] |= cpu_to_le32(ATA_PRD_EOT); 485 sgt_len = idx * 4; 486 487 /* 488 * Build ATA, host DMA packets 489 */ 490 pdc20621_host_sg(&qc->tf, &pp->dimm_buf[0], portno, total_len); 491 pdc20621_host_pkt(&qc->tf, &pp->dimm_buf[0], portno); 492 493 pdc20621_ata_sg(&qc->tf, &pp->dimm_buf[0], portno, total_len); 494 i = pdc20621_ata_pkt(&qc->tf, qc->dev->devno, &pp->dimm_buf[0], portno); 495 496 if (qc->tf.flags & ATA_TFLAG_LBA48) 497 i = pdc_prep_lba48(&qc->tf, &pp->dimm_buf[0], i); 498 else 499 i = pdc_prep_lba28(&qc->tf, &pp->dimm_buf[0], i); 500 501 pdc_pkt_footer(&qc->tf, &pp->dimm_buf[0], i); 502 503 /* copy three S/G tables and two packets to DIMM MMIO window */ 504 memcpy_toio(dimm_mmio + (portno * PDC_DIMM_WINDOW_STEP), 505 &pp->dimm_buf, PDC_DIMM_HEADER_SZ); 506 memcpy_toio(dimm_mmio + (portno * PDC_DIMM_WINDOW_STEP) + 507 PDC_DIMM_HOST_PRD, 508 &pp->dimm_buf[PDC_DIMM_HEADER_SZ], sgt_len); 509 510 /* force host FIFO dump */ 511 writel(0x00000001, mmio + PDC_20621_GENERAL_CTL); 512 513 readl(dimm_mmio); /* MMIO PCI posting flush */ 514 515 VPRINTK("ata pkt buf ofs %u, prd size %u, mmio copied\n", i, sgt_len); 516} 517 518static void pdc20621_nodata_prep(struct ata_queued_cmd *qc) 519{ 520 struct ata_port *ap = qc->ap; 521 struct pdc_port_priv *pp = ap->private_data; 522 void __iomem *mmio = ap->host->iomap[PDC_MMIO_BAR]; 523 void __iomem *dimm_mmio = ap->host->iomap[PDC_DIMM_BAR]; 524 unsigned int portno = ap->port_no; 525 unsigned int i; 526 527 VPRINTK("ata%u: ENTER\n", ap->print_id); 528 529 /* hard-code chip #0 */ 530 mmio += PDC_CHIP0_OFS; 531 532 i = pdc20621_ata_pkt(&qc->tf, qc->dev->devno, &pp->dimm_buf[0], portno); 533 534 if (qc->tf.flags & ATA_TFLAG_LBA48) 535 i = pdc_prep_lba48(&qc->tf, &pp->dimm_buf[0], i); 536 else 537 i = pdc_prep_lba28(&qc->tf, &pp->dimm_buf[0], i); 538 539 pdc_pkt_footer(&qc->tf, &pp->dimm_buf[0], i); 540 541 /* copy three S/G tables and two packets to DIMM MMIO window */ 542 memcpy_toio(dimm_mmio + (portno * PDC_DIMM_WINDOW_STEP), 543 &pp->dimm_buf, PDC_DIMM_HEADER_SZ); 544 545 /* force host FIFO dump */ 546 writel(0x00000001, mmio + PDC_20621_GENERAL_CTL); 547 548 readl(dimm_mmio); /* MMIO PCI posting flush */ 549 550 VPRINTK("ata pkt buf ofs %u, mmio copied\n", i); 551} 552 553static void pdc20621_qc_prep(struct ata_queued_cmd *qc) 554{ 555 switch (qc->tf.protocol) { 556 case ATA_PROT_DMA: 557 pdc20621_dma_prep(qc); 558 break; 559 case ATA_PROT_NODATA: 560 pdc20621_nodata_prep(qc); 561 break; 562 default: 563 break; 564 } 565} 566 567static void __pdc20621_push_hdma(struct ata_queued_cmd *qc, 568 unsigned int seq, 569 u32 pkt_ofs) 570{ 571 struct ata_port *ap = qc->ap; 572 struct ata_host *host = ap->host; 573 void __iomem *mmio = host->iomap[PDC_MMIO_BAR]; 574 575 /* hard-code chip #0 */ 576 mmio += PDC_CHIP0_OFS; 577 578 writel(0x00000001, mmio + PDC_20621_SEQCTL + (seq * 4)); 579 readl(mmio + PDC_20621_SEQCTL + (seq * 4)); /* flush */ 580 581 writel(pkt_ofs, mmio + PDC_HDMA_PKT_SUBMIT); 582 readl(mmio + PDC_HDMA_PKT_SUBMIT); /* flush */ 583} 584 585static void pdc20621_push_hdma(struct ata_queued_cmd *qc, 586 unsigned int seq, 587 u32 pkt_ofs) 588{ 589 struct ata_port *ap = qc->ap; 590 struct pdc_host_priv *pp = ap->host->private_data; 591 unsigned int idx = pp->hdma_prod & PDC_HDMA_Q_MASK; 592 593 if (!pp->doing_hdma) { 594 __pdc20621_push_hdma(qc, seq, pkt_ofs); 595 pp->doing_hdma = 1; 596 return; 597 } 598 599 pp->hdma[idx].qc = qc; 600 pp->hdma[idx].seq = seq; 601 pp->hdma[idx].pkt_ofs = pkt_ofs; 602 pp->hdma_prod++; 603} 604 605static void pdc20621_pop_hdma(struct ata_queued_cmd *qc) 606{ 607 struct ata_port *ap = qc->ap; 608 struct pdc_host_priv *pp = ap->host->private_data; 609 unsigned int idx = pp->hdma_cons & PDC_HDMA_Q_MASK; 610 611 /* if nothing on queue, we're done */ 612 if (pp->hdma_prod == pp->hdma_cons) { 613 pp->doing_hdma = 0; 614 return; 615 } 616 617 __pdc20621_push_hdma(pp->hdma[idx].qc, pp->hdma[idx].seq, 618 pp->hdma[idx].pkt_ofs); 619 pp->hdma_cons++; 620} 621 622#ifdef ATA_VERBOSE_DEBUG 623static void pdc20621_dump_hdma(struct ata_queued_cmd *qc) 624{ 625 struct ata_port *ap = qc->ap; 626 unsigned int port_no = ap->port_no; 627 void __iomem *dimm_mmio = ap->host->iomap[PDC_DIMM_BAR]; 628 629 dimm_mmio += (port_no * PDC_DIMM_WINDOW_STEP); 630 dimm_mmio += PDC_DIMM_HOST_PKT; 631 632 printk(KERN_ERR "HDMA[0] == 0x%08X\n", readl(dimm_mmio)); 633 printk(KERN_ERR "HDMA[1] == 0x%08X\n", readl(dimm_mmio + 4)); 634 printk(KERN_ERR "HDMA[2] == 0x%08X\n", readl(dimm_mmio + 8)); 635 printk(KERN_ERR "HDMA[3] == 0x%08X\n", readl(dimm_mmio + 12)); 636} 637#else 638static inline void pdc20621_dump_hdma(struct ata_queued_cmd *qc) { } 639#endif /* ATA_VERBOSE_DEBUG */ 640 641static void pdc20621_packet_start(struct ata_queued_cmd *qc) 642{ 643 struct ata_port *ap = qc->ap; 644 struct ata_host *host = ap->host; 645 unsigned int port_no = ap->port_no; 646 void __iomem *mmio = host->iomap[PDC_MMIO_BAR]; 647 unsigned int rw = (qc->tf.flags & ATA_TFLAG_WRITE); 648 u8 seq = (u8) (port_no + 1); 649 unsigned int port_ofs; 650 651 /* hard-code chip #0 */ 652 mmio += PDC_CHIP0_OFS; 653 654 VPRINTK("ata%u: ENTER\n", ap->print_id); 655 656 wmb(); /* flush PRD, pkt writes */ 657 658 port_ofs = PDC_20621_DIMM_BASE + (PDC_DIMM_WINDOW_STEP * port_no); 659 660 /* if writing, we (1) DMA to DIMM, then (2) do ATA command */ 661 if (rw && qc->tf.protocol == ATA_PROT_DMA) { 662 seq += 4; 663 664 pdc20621_dump_hdma(qc); 665 pdc20621_push_hdma(qc, seq, port_ofs + PDC_DIMM_HOST_PKT); 666 VPRINTK("queued ofs 0x%x (%u), seq %u\n", 667 port_ofs + PDC_DIMM_HOST_PKT, 668 port_ofs + PDC_DIMM_HOST_PKT, 669 seq); 670 } else { 671 writel(0x00000001, mmio + PDC_20621_SEQCTL + (seq * 4)); 672 readl(mmio + PDC_20621_SEQCTL + (seq * 4)); /* flush */ 673 674 writel(port_ofs + PDC_DIMM_ATA_PKT, 675 ap->ioaddr.cmd_addr + PDC_PKT_SUBMIT); 676 readl(ap->ioaddr.cmd_addr + PDC_PKT_SUBMIT); 677 VPRINTK("submitted ofs 0x%x (%u), seq %u\n", 678 port_ofs + PDC_DIMM_ATA_PKT, 679 port_ofs + PDC_DIMM_ATA_PKT, 680 seq); 681 } 682} 683 684static unsigned int pdc20621_qc_issue(struct ata_queued_cmd *qc) 685{ 686 switch (qc->tf.protocol) { 687 case ATA_PROT_NODATA: 688 if (qc->tf.flags & ATA_TFLAG_POLLING) 689 break; 690 /*FALLTHROUGH*/ 691 case ATA_PROT_DMA: 692 pdc20621_packet_start(qc); 693 return 0; 694 695 case ATAPI_PROT_DMA: 696 BUG(); 697 break; 698 699 default: 700 break; 701 } 702 703 return ata_sff_qc_issue(qc); 704} 705 706static inline unsigned int pdc20621_host_intr(struct ata_port *ap, 707 struct ata_queued_cmd *qc, 708 unsigned int doing_hdma, 709 void __iomem *mmio) 710{ 711 unsigned int port_no = ap->port_no; 712 unsigned int port_ofs = 713 PDC_20621_DIMM_BASE + (PDC_DIMM_WINDOW_STEP * port_no); 714 u8 status; 715 unsigned int handled = 0; 716 717 VPRINTK("ENTER\n"); 718 719 if ((qc->tf.protocol == ATA_PROT_DMA) && /* read */ 720 (!(qc->tf.flags & ATA_TFLAG_WRITE))) { 721 722 /* step two - DMA from DIMM to host */ 723 if (doing_hdma) { 724 VPRINTK("ata%u: read hdma, 0x%x 0x%x\n", ap->print_id, 725 readl(mmio + 0x104), readl(mmio + PDC_HDMA_CTLSTAT)); 726 /* get drive status; clear intr; complete txn */ 727 qc->err_mask |= ac_err_mask(ata_wait_idle(ap)); 728 ata_qc_complete(qc); 729 pdc20621_pop_hdma(qc); 730 } 731 732 /* step one - exec ATA command */ 733 else { 734 u8 seq = (u8) (port_no + 1 + 4); 735 VPRINTK("ata%u: read ata, 0x%x 0x%x\n", ap->print_id, 736 readl(mmio + 0x104), readl(mmio + PDC_HDMA_CTLSTAT)); 737 738 /* submit hdma pkt */ 739 pdc20621_dump_hdma(qc); 740 pdc20621_push_hdma(qc, seq, 741 port_ofs + PDC_DIMM_HOST_PKT); 742 } 743 handled = 1; 744 745 } else if (qc->tf.protocol == ATA_PROT_DMA) { /* write */ 746 747 /* step one - DMA from host to DIMM */ 748 if (doing_hdma) { 749 u8 seq = (u8) (port_no + 1); 750 VPRINTK("ata%u: write hdma, 0x%x 0x%x\n", ap->print_id, 751 readl(mmio + 0x104), readl(mmio + PDC_HDMA_CTLSTAT)); 752 753 /* submit ata pkt */ 754 writel(0x00000001, mmio + PDC_20621_SEQCTL + (seq * 4)); 755 readl(mmio + PDC_20621_SEQCTL + (seq * 4)); 756 writel(port_ofs + PDC_DIMM_ATA_PKT, 757 ap->ioaddr.cmd_addr + PDC_PKT_SUBMIT); 758 readl(ap->ioaddr.cmd_addr + PDC_PKT_SUBMIT); 759 } 760 761 /* step two - execute ATA command */ 762 else { 763 VPRINTK("ata%u: write ata, 0x%x 0x%x\n", ap->print_id, 764 readl(mmio + 0x104), readl(mmio + PDC_HDMA_CTLSTAT)); 765 /* get drive status; clear intr; complete txn */ 766 qc->err_mask |= ac_err_mask(ata_wait_idle(ap)); 767 ata_qc_complete(qc); 768 pdc20621_pop_hdma(qc); 769 } 770 handled = 1; 771 772 /* command completion, but no data xfer */ 773 } else if (qc->tf.protocol == ATA_PROT_NODATA) { 774 775 status = ata_sff_busy_wait(ap, ATA_BUSY | ATA_DRQ, 1000); 776 DPRINTK("BUS_NODATA (drv_stat 0x%X)\n", status); 777 qc->err_mask |= ac_err_mask(status); 778 ata_qc_complete(qc); 779 handled = 1; 780 781 } else { 782 ap->stats.idle_irq++; 783 } 784 785 return handled; 786} 787 788static void pdc20621_irq_clear(struct ata_port *ap) 789{ 790 ioread8(ap->ioaddr.status_addr); 791} 792 793static irqreturn_t pdc20621_interrupt(int irq, void *dev_instance) 794{ 795 struct ata_host *host = dev_instance; 796 struct ata_port *ap; 797 u32 mask = 0; 798 unsigned int i, tmp, port_no; 799 unsigned int handled = 0; 800 void __iomem *mmio_base; 801 802 VPRINTK("ENTER\n"); 803 804 if (!host || !host->iomap[PDC_MMIO_BAR]) { 805 VPRINTK("QUICK EXIT\n"); 806 return IRQ_NONE; 807 } 808 809 mmio_base = host->iomap[PDC_MMIO_BAR]; 810 811 /* reading should also clear interrupts */ 812 mmio_base += PDC_CHIP0_OFS; 813 mask = readl(mmio_base + PDC_20621_SEQMASK); 814 VPRINTK("mask == 0x%x\n", mask); 815 816 if (mask == 0xffffffff) { 817 VPRINTK("QUICK EXIT 2\n"); 818 return IRQ_NONE; 819 } 820 mask &= 0xffff; /* only 16 tags possible */ 821 if (!mask) { 822 VPRINTK("QUICK EXIT 3\n"); 823 return IRQ_NONE; 824 } 825 826 spin_lock(&host->lock); 827 828 for (i = 1; i < 9; i++) { 829 port_no = i - 1; 830 if (port_no > 3) 831 port_no -= 4; 832 if (port_no >= host->n_ports) 833 ap = NULL; 834 else 835 ap = host->ports[port_no]; 836 tmp = mask & (1 << i); 837 VPRINTK("seq %u, port_no %u, ap %p, tmp %x\n", i, port_no, ap, tmp); 838 if (tmp && ap) { 839 struct ata_queued_cmd *qc; 840 841 qc = ata_qc_from_tag(ap, ap->link.active_tag); 842 if (qc && (!(qc->tf.flags & ATA_TFLAG_POLLING))) 843 handled += pdc20621_host_intr(ap, qc, (i > 4), 844 mmio_base); 845 } 846 } 847 848 spin_unlock(&host->lock); 849 850 VPRINTK("mask == 0x%x\n", mask); 851 852 VPRINTK("EXIT\n"); 853 854 return IRQ_RETVAL(handled); 855} 856 857static void pdc_freeze(struct ata_port *ap) 858{ 859 void __iomem *mmio = ap->ioaddr.cmd_addr; 860 u32 tmp; 861 862 863 tmp = readl(mmio + PDC_CTLSTAT); 864 tmp |= PDC_MASK_INT; 865 tmp &= ~PDC_DMA_ENABLE; 866 writel(tmp, mmio + PDC_CTLSTAT); 867 readl(mmio + PDC_CTLSTAT); /* flush */ 868} 869 870static void pdc_thaw(struct ata_port *ap) 871{ 872 void __iomem *mmio = ap->ioaddr.cmd_addr; 873 u32 tmp; 874 875 876 /* clear IRQ */ 877 ioread8(ap->ioaddr.status_addr); 878 879 /* turn IRQ back on */ 880 tmp = readl(mmio + PDC_CTLSTAT); 881 tmp &= ~PDC_MASK_INT; 882 writel(tmp, mmio + PDC_CTLSTAT); 883 readl(mmio + PDC_CTLSTAT); /* flush */ 884} 885 886static void pdc_reset_port(struct ata_port *ap) 887{ 888 void __iomem *mmio = ap->ioaddr.cmd_addr + PDC_CTLSTAT; 889 unsigned int i; 890 u32 tmp; 891 892 893 for (i = 11; i > 0; i--) { 894 tmp = readl(mmio); 895 if (tmp & PDC_RESET) 896 break; 897 898 udelay(100); 899 900 tmp |= PDC_RESET; 901 writel(tmp, mmio); 902 } 903 904 tmp &= ~PDC_RESET; 905 writel(tmp, mmio); 906 readl(mmio); /* flush */ 907} 908 909static int pdc_softreset(struct ata_link *link, unsigned int *class, 910 unsigned long deadline) 911{ 912 pdc_reset_port(link->ap); 913 return ata_sff_softreset(link, class, deadline); 914} 915 916static void pdc_error_handler(struct ata_port *ap) 917{ 918 if (!(ap->pflags & ATA_PFLAG_FROZEN)) 919 pdc_reset_port(ap); 920 921 ata_sff_error_handler(ap); 922} 923 924static void pdc_post_internal_cmd(struct ata_queued_cmd *qc) 925{ 926 struct ata_port *ap = qc->ap; 927 928 /* make DMA engine forget about the failed command */ 929 if (qc->flags & ATA_QCFLAG_FAILED) 930 pdc_reset_port(ap); 931} 932 933static int pdc_check_atapi_dma(struct ata_queued_cmd *qc) 934{ 935 u8 *scsicmd = qc->scsicmd->cmnd; 936 int pio = 1; /* atapi dma off by default */ 937 938 /* Whitelist commands that may use DMA. */ 939 switch (scsicmd[0]) { 940 case WRITE_12: 941 case WRITE_10: 942 case WRITE_6: 943 case READ_12: 944 case READ_10: 945 case READ_6: 946 case 0xad: /* READ_DVD_STRUCTURE */ 947 case 0xbe: /* READ_CD */ 948 pio = 0; 949 } 950 /* -45150 (FFFF4FA2) to -1 (FFFFFFFF) shall use PIO mode */ 951 if (scsicmd[0] == WRITE_10) { 952 unsigned int lba = 953 (scsicmd[2] << 24) | 954 (scsicmd[3] << 16) | 955 (scsicmd[4] << 8) | 956 scsicmd[5]; 957 if (lba >= 0xFFFF4FA2) 958 pio = 1; 959 } 960 return pio; 961} 962 963static void pdc_tf_load_mmio(struct ata_port *ap, const struct ata_taskfile *tf) 964{ 965 WARN_ON(tf->protocol == ATA_PROT_DMA || 966 tf->protocol == ATAPI_PROT_DMA); 967 ata_sff_tf_load(ap, tf); 968} 969 970 971static void pdc_exec_command_mmio(struct ata_port *ap, const struct ata_taskfile *tf) 972{ 973 WARN_ON(tf->protocol == ATA_PROT_DMA || 974 tf->protocol == ATAPI_PROT_DMA); 975 ata_sff_exec_command(ap, tf); 976} 977 978 979static void pdc_sata_setup_port(struct ata_ioports *port, void __iomem *base) 980{ 981 port->cmd_addr = base; 982 port->data_addr = base; 983 port->feature_addr = 984 port->error_addr = base + 0x4; 985 port->nsect_addr = base + 0x8; 986 port->lbal_addr = base + 0xc; 987 port->lbam_addr = base + 0x10; 988 port->lbah_addr = base + 0x14; 989 port->device_addr = base + 0x18; 990 port->command_addr = 991 port->status_addr = base + 0x1c; 992 port->altstatus_addr = 993 port->ctl_addr = base + 0x38; 994} 995 996 997#ifdef ATA_VERBOSE_DEBUG 998static void pdc20621_get_from_dimm(struct ata_host *host, void *psource, 999 u32 offset, u32 size) 1000{ 1001 u32 window_size; 1002 u16 idx; 1003 u8 page_mask; 1004 long dist; 1005 void __iomem *mmio = host->iomap[PDC_MMIO_BAR]; 1006 void __iomem *dimm_mmio = host->iomap[PDC_DIMM_BAR]; 1007 1008 /* hard-code chip #0 */ 1009 mmio += PDC_CHIP0_OFS; 1010 1011 page_mask = 0x00; 1012 window_size = 0x2000 * 4; /* 32K byte uchar size */ 1013 idx = (u16) (offset / window_size); 1014 1015 writel(0x01, mmio + PDC_GENERAL_CTLR); 1016 readl(mmio + PDC_GENERAL_CTLR); 1017 writel(((idx) << page_mask), mmio + PDC_DIMM_WINDOW_CTLR); 1018 readl(mmio + PDC_DIMM_WINDOW_CTLR); 1019 1020 offset -= (idx * window_size); 1021 idx++; 1022 dist = ((long) (window_size - (offset + size))) >= 0 ? size : 1023 (long) (window_size - offset); 1024 memcpy_fromio((char *) psource, (char *) (dimm_mmio + offset / 4), 1025 dist); 1026 1027 psource += dist; 1028 size -= dist; 1029 for (; (long) size >= (long) window_size ;) { 1030 writel(0x01, mmio + PDC_GENERAL_CTLR); 1031 readl(mmio + PDC_GENERAL_CTLR); 1032 writel(((idx) << page_mask), mmio + PDC_DIMM_WINDOW_CTLR); 1033 readl(mmio + PDC_DIMM_WINDOW_CTLR); 1034 memcpy_fromio((char *) psource, (char *) (dimm_mmio), 1035 window_size / 4); 1036 psource += window_size; 1037 size -= window_size; 1038 idx++; 1039 } 1040 1041 if (size) { 1042 writel(0x01, mmio + PDC_GENERAL_CTLR); 1043 readl(mmio + PDC_GENERAL_CTLR); 1044 writel(((idx) << page_mask), mmio + PDC_DIMM_WINDOW_CTLR); 1045 readl(mmio + PDC_DIMM_WINDOW_CTLR); 1046 memcpy_fromio((char *) psource, (char *) (dimm_mmio), 1047 size / 4); 1048 } 1049} 1050#endif 1051 1052 1053static void pdc20621_put_to_dimm(struct ata_host *host, void *psource, 1054 u32 offset, u32 size) 1055{ 1056 u32 window_size; 1057 u16 idx; 1058 u8 page_mask; 1059 long dist; 1060 void __iomem *mmio = host->iomap[PDC_MMIO_BAR]; 1061 void __iomem *dimm_mmio = host->iomap[PDC_DIMM_BAR]; 1062 1063 /* hard-code chip #0 */ 1064 mmio += PDC_CHIP0_OFS; 1065 1066 page_mask = 0x00; 1067 window_size = 0x2000 * 4; /* 32K byte uchar size */ 1068 idx = (u16) (offset / window_size); 1069 1070 writel(((idx) << page_mask), mmio + PDC_DIMM_WINDOW_CTLR); 1071 readl(mmio + PDC_DIMM_WINDOW_CTLR); 1072 offset -= (idx * window_size); 1073 idx++; 1074 dist = ((long)(s32)(window_size - (offset + size))) >= 0 ? size : 1075 (long) (window_size - offset); 1076 memcpy_toio(dimm_mmio + offset / 4, psource, dist); 1077 writel(0x01, mmio + PDC_GENERAL_CTLR); 1078 readl(mmio + PDC_GENERAL_CTLR); 1079 1080 psource += dist; 1081 size -= dist; 1082 for (; (long) size >= (long) window_size ;) { 1083 writel(((idx) << page_mask), mmio + PDC_DIMM_WINDOW_CTLR); 1084 readl(mmio + PDC_DIMM_WINDOW_CTLR); 1085 memcpy_toio(dimm_mmio, psource, window_size / 4); 1086 writel(0x01, mmio + PDC_GENERAL_CTLR); 1087 readl(mmio + PDC_GENERAL_CTLR); 1088 psource += window_size; 1089 size -= window_size; 1090 idx++; 1091 } 1092 1093 if (size) { 1094 writel(((idx) << page_mask), mmio + PDC_DIMM_WINDOW_CTLR); 1095 readl(mmio + PDC_DIMM_WINDOW_CTLR); 1096 memcpy_toio(dimm_mmio, psource, size / 4); 1097 writel(0x01, mmio + PDC_GENERAL_CTLR); 1098 readl(mmio + PDC_GENERAL_CTLR); 1099 } 1100} 1101 1102 1103static unsigned int pdc20621_i2c_read(struct ata_host *host, u32 device, 1104 u32 subaddr, u32 *pdata) 1105{ 1106 void __iomem *mmio = host->iomap[PDC_MMIO_BAR]; 1107 u32 i2creg = 0; 1108 u32 status; 1109 u32 count = 0; 1110 1111 /* hard-code chip #0 */ 1112 mmio += PDC_CHIP0_OFS; 1113 1114 i2creg |= device << 24; 1115 i2creg |= subaddr << 16; 1116 1117 /* Set the device and subaddress */ 1118 writel(i2creg, mmio + PDC_I2C_ADDR_DATA); 1119 readl(mmio + PDC_I2C_ADDR_DATA); 1120 1121 /* Write Control to perform read operation, mask int */ 1122 writel(PDC_I2C_READ | PDC_I2C_START | PDC_I2C_MASK_INT, 1123 mmio + PDC_I2C_CONTROL); 1124 1125 for (count = 0; count <= 1000; count ++) { 1126 status = readl(mmio + PDC_I2C_CONTROL); 1127 if (status & PDC_I2C_COMPLETE) { 1128 status = readl(mmio + PDC_I2C_ADDR_DATA); 1129 break; 1130 } else if (count == 1000) 1131 return 0; 1132 } 1133 1134 *pdata = (status >> 8) & 0x000000ff; 1135 return 1; 1136} 1137 1138 1139static int pdc20621_detect_dimm(struct ata_host *host) 1140{ 1141 u32 data = 0; 1142 if (pdc20621_i2c_read(host, PDC_DIMM0_SPD_DEV_ADDRESS, 1143 PDC_DIMM_SPD_SYSTEM_FREQ, &data)) { 1144 if (data == 100) 1145 return 100; 1146 } else 1147 return 0; 1148 1149 if (pdc20621_i2c_read(host, PDC_DIMM0_SPD_DEV_ADDRESS, 9, &data)) { 1150 if (data <= 0x75) 1151 return 133; 1152 } else 1153 return 0; 1154 1155 return 0; 1156} 1157 1158 1159static int pdc20621_prog_dimm0(struct ata_host *host) 1160{ 1161 u32 spd0[50]; 1162 u32 data = 0; 1163 int size, i; 1164 u8 bdimmsize; 1165 void __iomem *mmio = host->iomap[PDC_MMIO_BAR]; 1166 static const struct { 1167 unsigned int reg; 1168 unsigned int ofs; 1169 } pdc_i2c_read_data [] = { 1170 { PDC_DIMM_SPD_TYPE, 11 }, 1171 { PDC_DIMM_SPD_FRESH_RATE, 12 }, 1172 { PDC_DIMM_SPD_COLUMN_NUM, 4 }, 1173 { PDC_DIMM_SPD_ATTRIBUTE, 21 }, 1174 { PDC_DIMM_SPD_ROW_NUM, 3 }, 1175 { PDC_DIMM_SPD_BANK_NUM, 17 }, 1176 { PDC_DIMM_SPD_MODULE_ROW, 5 }, 1177 { PDC_DIMM_SPD_ROW_PRE_CHARGE, 27 }, 1178 { PDC_DIMM_SPD_ROW_ACTIVE_DELAY, 28 }, 1179 { PDC_DIMM_SPD_RAS_CAS_DELAY, 29 }, 1180 { PDC_DIMM_SPD_ACTIVE_PRECHARGE, 30 }, 1181 { PDC_DIMM_SPD_CAS_LATENCY, 18 }, 1182 }; 1183 1184 /* hard-code chip #0 */ 1185 mmio += PDC_CHIP0_OFS; 1186 1187 for (i = 0; i < ARRAY_SIZE(pdc_i2c_read_data); i++) 1188 pdc20621_i2c_read(host, PDC_DIMM0_SPD_DEV_ADDRESS, 1189 pdc_i2c_read_data[i].reg, 1190 &spd0[pdc_i2c_read_data[i].ofs]); 1191 1192 data |= (spd0[4] - 8) | ((spd0[21] != 0) << 3) | ((spd0[3]-11) << 4); 1193 data |= ((spd0[17] / 4) << 6) | ((spd0[5] / 2) << 7) | 1194 ((((spd0[27] + 9) / 10) - 1) << 8) ; 1195 data |= (((((spd0[29] > spd0[28]) 1196 ? spd0[29] : spd0[28]) + 9) / 10) - 1) << 10; 1197 data |= ((spd0[30] - spd0[29] + 9) / 10 - 2) << 12; 1198 1199 if (spd0[18] & 0x08) 1200 data |= ((0x03) << 14); 1201 else if (spd0[18] & 0x04) 1202 data |= ((0x02) << 14); 1203 else if (spd0[18] & 0x01) 1204 data |= ((0x01) << 14); 1205 else 1206 data |= (0 << 14); 1207 1208 /* 1209 Calculate the size of bDIMMSize (power of 2) and 1210 merge the DIMM size by program start/end address. 1211 */ 1212 1213 bdimmsize = spd0[4] + (spd0[5] / 2) + spd0[3] + (spd0[17] / 2) + 3; 1214 size = (1 << bdimmsize) >> 20; 1215 data |= (((size / 16) - 1) << 16); 1216 data |= (0 << 23); 1217 data |= 8; 1218 writel(data, mmio + PDC_DIMM0_CONTROL); 1219 readl(mmio + PDC_DIMM0_CONTROL); 1220 return size; 1221} 1222 1223 1224static unsigned int pdc20621_prog_dimm_global(struct ata_host *host) 1225{ 1226 u32 data, spd0; 1227 int error, i; 1228 void __iomem *mmio = host->iomap[PDC_MMIO_BAR]; 1229 1230 /* hard-code chip #0 */ 1231 mmio += PDC_CHIP0_OFS; 1232 1233 /* 1234 Set To Default : DIMM Module Global Control Register (0x022259F1) 1235 DIMM Arbitration Disable (bit 20) 1236 DIMM Data/Control Output Driving Selection (bit12 - bit15) 1237 Refresh Enable (bit 17) 1238 */ 1239 1240 data = 0x022259F1; 1241 writel(data, mmio + PDC_SDRAM_CONTROL); 1242 readl(mmio + PDC_SDRAM_CONTROL); 1243 1244 /* Turn on for ECC */ 1245 pdc20621_i2c_read(host, PDC_DIMM0_SPD_DEV_ADDRESS, 1246 PDC_DIMM_SPD_TYPE, &spd0); 1247 if (spd0 == 0x02) { 1248 data |= (0x01 << 16); 1249 writel(data, mmio + PDC_SDRAM_CONTROL); 1250 readl(mmio + PDC_SDRAM_CONTROL); 1251 printk(KERN_ERR "Local DIMM ECC Enabled\n"); 1252 } 1253 1254 /* DIMM Initialization Select/Enable (bit 18/19) */ 1255 data &= (~(1<<18)); 1256 data |= (1<<19); 1257 writel(data, mmio + PDC_SDRAM_CONTROL); 1258 1259 error = 1; 1260 for (i = 1; i <= 10; i++) { /* polling ~5 secs */ 1261 data = readl(mmio + PDC_SDRAM_CONTROL); 1262 if (!(data & (1<<19))) { 1263 error = 0; 1264 break; 1265 } 1266 msleep(i*100); 1267 } 1268 return error; 1269} 1270 1271 1272static unsigned int pdc20621_dimm_init(struct ata_host *host) 1273{ 1274 int speed, size, length; 1275 u32 addr, spd0, pci_status; 1276 u32 time_period = 0; 1277 u32 tcount = 0; 1278 u32 ticks = 0; 1279 u32 clock = 0; 1280 u32 fparam = 0; 1281 void __iomem *mmio = host->iomap[PDC_MMIO_BAR]; 1282 1283 /* hard-code chip #0 */ 1284 mmio += PDC_CHIP0_OFS; 1285 1286 /* Initialize PLL based upon PCI Bus Frequency */ 1287 1288 /* Initialize Time Period Register */ 1289 writel(0xffffffff, mmio + PDC_TIME_PERIOD); 1290 time_period = readl(mmio + PDC_TIME_PERIOD); 1291 VPRINTK("Time Period Register (0x40): 0x%x\n", time_period); 1292 1293 /* Enable timer */ 1294 writel(PDC_TIMER_DEFAULT, mmio + PDC_TIME_CONTROL); 1295 readl(mmio + PDC_TIME_CONTROL); 1296 1297 /* Wait 3 seconds */ 1298 msleep(3000); 1299 1300 /* 1301 When timer is enabled, counter is decreased every internal 1302 clock cycle. 1303 */ 1304 1305 tcount = readl(mmio + PDC_TIME_COUNTER); 1306 VPRINTK("Time Counter Register (0x44): 0x%x\n", tcount); 1307 1308 /* 1309 If SX4 is on PCI-X bus, after 3 seconds, the timer counter 1310 register should be >= (0xffffffff - 3x10^8). 1311 */ 1312 if (tcount >= PCI_X_TCOUNT) { 1313 ticks = (time_period - tcount); 1314 VPRINTK("Num counters 0x%x (%d)\n", ticks, ticks); 1315 1316 clock = (ticks / 300000); 1317 VPRINTK("10 * Internal clk = 0x%x (%d)\n", clock, clock); 1318 1319 clock = (clock * 33); 1320 VPRINTK("10 * Internal clk * 33 = 0x%x (%d)\n", clock, clock); 1321 1322 /* PLL F Param (bit 22:16) */ 1323 fparam = (1400000 / clock) - 2; 1324 VPRINTK("PLL F Param: 0x%x (%d)\n", fparam, fparam); 1325 1326 /* OD param = 0x2 (bit 31:30), R param = 0x5 (bit 29:25) */ 1327 pci_status = (0x8a001824 | (fparam << 16)); 1328 } else 1329 pci_status = PCI_PLL_INIT; 1330 1331 /* Initialize PLL. */ 1332 VPRINTK("pci_status: 0x%x\n", pci_status); 1333 writel(pci_status, mmio + PDC_CTL_STATUS); 1334 readl(mmio + PDC_CTL_STATUS); 1335 1336 /* 1337 Read SPD of DIMM by I2C interface, 1338 and program the DIMM Module Controller. 1339 */ 1340 if (!(speed = pdc20621_detect_dimm(host))) { 1341 printk(KERN_ERR "Detect Local DIMM Fail\n"); 1342 return 1; /* DIMM error */ 1343 } 1344 VPRINTK("Local DIMM Speed = %d\n", speed); 1345 1346 /* Programming DIMM0 Module Control Register (index_CID0:80h) */ 1347 size = pdc20621_prog_dimm0(host); 1348 VPRINTK("Local DIMM Size = %dMB\n", size); 1349 1350 /* Programming DIMM Module Global Control Register (index_CID0:88h) */ 1351 if (pdc20621_prog_dimm_global(host)) { 1352 printk(KERN_ERR "Programming DIMM Module Global Control Register Fail\n"); 1353 return 1; 1354 } 1355 1356#ifdef ATA_VERBOSE_DEBUG 1357 { 1358 u8 test_parttern1[40] = 1359 {0x55,0xAA,'P','r','o','m','i','s','e',' ', 1360 'N','o','t',' ','Y','e','t',' ', 1361 'D','e','f','i','n','e','d',' ', 1362 '1','.','1','0', 1363 '9','8','0','3','1','6','1','2',0,0}; 1364 u8 test_parttern2[40] = {0}; 1365 1366 pdc20621_put_to_dimm(host, test_parttern2, 0x10040, 40); 1367 pdc20621_put_to_dimm(host, test_parttern2, 0x40, 40); 1368 1369 pdc20621_put_to_dimm(host, test_parttern1, 0x10040, 40); 1370 pdc20621_get_from_dimm(host, test_parttern2, 0x40, 40); 1371 printk(KERN_ERR "%x, %x, %s\n", test_parttern2[0], 1372 test_parttern2[1], &(test_parttern2[2])); 1373 pdc20621_get_from_dimm(host, test_parttern2, 0x10040, 1374 40); 1375 printk(KERN_ERR "%x, %x, %s\n", test_parttern2[0], 1376 test_parttern2[1], &(test_parttern2[2])); 1377 1378 pdc20621_put_to_dimm(host, test_parttern1, 0x40, 40); 1379 pdc20621_get_from_dimm(host, test_parttern2, 0x40, 40); 1380 printk(KERN_ERR "%x, %x, %s\n", test_parttern2[0], 1381 test_parttern2[1], &(test_parttern2[2])); 1382 } 1383#endif 1384 1385 /* ECC initiliazation. */ 1386 1387 pdc20621_i2c_read(host, PDC_DIMM0_SPD_DEV_ADDRESS, 1388 PDC_DIMM_SPD_TYPE, &spd0); 1389 if (spd0 == 0x02) { 1390 void *buf; 1391 VPRINTK("Start ECC initialization\n"); 1392 addr = 0; 1393 length = size * 1024 * 1024; 1394 buf = kzalloc(ECC_ERASE_BUF_SZ, GFP_KERNEL); 1395 while (addr < length) { 1396 pdc20621_put_to_dimm(host, buf, addr, 1397 ECC_ERASE_BUF_SZ); 1398 addr += ECC_ERASE_BUF_SZ; 1399 } 1400 kfree(buf); 1401 VPRINTK("Finish ECC initialization\n"); 1402 } 1403 return 0; 1404} 1405 1406 1407static void pdc_20621_init(struct ata_host *host) 1408{ 1409 u32 tmp; 1410 void __iomem *mmio = host->iomap[PDC_MMIO_BAR]; 1411 1412 /* hard-code chip #0 */ 1413 mmio += PDC_CHIP0_OFS; 1414 1415 /* 1416 * Select page 0x40 for our 32k DIMM window 1417 */ 1418 tmp = readl(mmio + PDC_20621_DIMM_WINDOW) & 0xffff0000; 1419 tmp |= PDC_PAGE_WINDOW; /* page 40h; arbitrarily selected */ 1420 writel(tmp, mmio + PDC_20621_DIMM_WINDOW); 1421 1422 /* 1423 * Reset Host DMA 1424 */ 1425 tmp = readl(mmio + PDC_HDMA_CTLSTAT); 1426 tmp |= PDC_RESET; 1427 writel(tmp, mmio + PDC_HDMA_CTLSTAT); 1428 readl(mmio + PDC_HDMA_CTLSTAT); /* flush */ 1429 1430 udelay(10); 1431 1432 tmp = readl(mmio + PDC_HDMA_CTLSTAT); 1433 tmp &= ~PDC_RESET; 1434 writel(tmp, mmio + PDC_HDMA_CTLSTAT); 1435 readl(mmio + PDC_HDMA_CTLSTAT); /* flush */ 1436} 1437 1438static int pdc_sata_init_one(struct pci_dev *pdev, 1439 const struct pci_device_id *ent) 1440{ 1441 static int printed_version; 1442 const struct ata_port_info *ppi[] = 1443 { &pdc_port_info[ent->driver_data], NULL }; 1444 struct ata_host *host; 1445 struct pdc_host_priv *hpriv; 1446 int i, rc; 1447 1448 if (!printed_version++) 1449 dev_printk(KERN_DEBUG, &pdev->dev, "version " DRV_VERSION "\n"); 1450 1451 /* allocate host */ 1452 host = ata_host_alloc_pinfo(&pdev->dev, ppi, 4); 1453 hpriv = devm_kzalloc(&pdev->dev, sizeof(*hpriv), GFP_KERNEL); 1454 if (!host || !hpriv) 1455 return -ENOMEM; 1456 1457 host->private_data = hpriv; 1458 1459 /* acquire resources and fill host */ 1460 rc = pcim_enable_device(pdev); 1461 if (rc) 1462 return rc; 1463 1464 rc = pcim_iomap_regions(pdev, (1 << PDC_MMIO_BAR) | (1 << PDC_DIMM_BAR), 1465 DRV_NAME); 1466 if (rc == -EBUSY) 1467 pcim_pin_device(pdev); 1468 if (rc) 1469 return rc; 1470 host->iomap = pcim_iomap_table(pdev); 1471 1472 for (i = 0; i < 4; i++) { 1473 struct ata_port *ap = host->ports[i]; 1474 void __iomem *base = host->iomap[PDC_MMIO_BAR] + PDC_CHIP0_OFS; 1475 unsigned int offset = 0x200 + i * 0x80; 1476 1477 pdc_sata_setup_port(&ap->ioaddr, base + offset); 1478 1479 ata_port_pbar_desc(ap, PDC_MMIO_BAR, -1, "mmio"); 1480 ata_port_pbar_desc(ap, PDC_DIMM_BAR, -1, "dimm"); 1481 ata_port_pbar_desc(ap, PDC_MMIO_BAR, offset, "port"); 1482 } 1483 1484 /* configure and activate */ 1485 rc = pci_set_dma_mask(pdev, ATA_DMA_MASK); 1486 if (rc) 1487 return rc; 1488 rc = pci_set_consistent_dma_mask(pdev, ATA_DMA_MASK); 1489 if (rc) 1490 return rc; 1491 1492 if (pdc20621_dimm_init(host)) 1493 return -ENOMEM; 1494 pdc_20621_init(host); 1495 1496 pci_set_master(pdev); 1497 return ata_host_activate(host, pdev->irq, pdc20621_interrupt, 1498 IRQF_SHARED, &pdc_sata_sht); 1499} 1500 1501 1502static int __init pdc_sata_init(void) 1503{ 1504 return pci_register_driver(&pdc_sata_pci_driver); 1505} 1506 1507 1508static void __exit pdc_sata_exit(void) 1509{ 1510 pci_unregister_driver(&pdc_sata_pci_driver); 1511} 1512 1513 1514MODULE_AUTHOR("Jeff Garzik"); 1515MODULE_DESCRIPTION("Promise SATA low-level driver"); 1516MODULE_LICENSE("GPL"); 1517MODULE_DEVICE_TABLE(pci, pdc_sata_pci_tbl); 1518MODULE_VERSION(DRV_VERSION); 1519 1520module_init(pdc_sata_init); 1521module_exit(pdc_sata_exit); 1522