1/* 2 * linux/drivers/net/irda/pxaficp_ir.c 3 * 4 * Based on sa1100_ir.c by Russell King 5 * 6 * Changes copyright (C) 2003-2005 MontaVista Software, Inc. 7 * 8 * This program is free software; you can redistribute it and/or modify 9 * it under the terms of the GNU General Public License version 2 as 10 * published by the Free Software Foundation. 11 * 12 * Infra-red driver (SIR/FIR) for the PXA2xx embedded microprocessor 13 * 14 */ 15#include <linux/module.h> 16#include <linux/types.h> 17#include <linux/init.h> 18#include <linux/errno.h> 19#include <linux/netdevice.h> 20#include <linux/slab.h> 21#include <linux/rtnetlink.h> 22#include <linux/interrupt.h> 23#include <linux/dma-mapping.h> 24#include <linux/platform_device.h> 25#include <linux/pm.h> 26 27#include <net/irda/irda.h> 28#include <net/irda/irmod.h> 29#include <net/irda/wrapper.h> 30#include <net/irda/irda_device.h> 31 32#include <asm/irq.h> 33#include <asm/dma.h> 34#include <asm/delay.h> 35#include <asm/hardware.h> 36#include <asm/arch/irda.h> 37#include <asm/arch/pxa-regs.h> 38 39#ifdef CONFIG_MACH_MAINSTONE 40#include <asm/arch/mainstone.h> 41#endif 42 43#define IrSR_RXPL_NEG_IS_ZERO (1<<4) 44#define IrSR_RXPL_POS_IS_ZERO 0x0 45#define IrSR_TXPL_NEG_IS_ZERO (1<<3) 46#define IrSR_TXPL_POS_IS_ZERO 0x0 47#define IrSR_XMODE_PULSE_1_6 (1<<2) 48#define IrSR_XMODE_PULSE_3_16 0x0 49#define IrSR_RCVEIR_IR_MODE (1<<1) 50#define IrSR_RCVEIR_UART_MODE 0x0 51#define IrSR_XMITIR_IR_MODE (1<<0) 52#define IrSR_XMITIR_UART_MODE 0x0 53 54#define IrSR_IR_RECEIVE_ON (\ 55 IrSR_RXPL_NEG_IS_ZERO | \ 56 IrSR_TXPL_POS_IS_ZERO | \ 57 IrSR_XMODE_PULSE_3_16 | \ 58 IrSR_RCVEIR_IR_MODE | \ 59 IrSR_XMITIR_UART_MODE) 60 61#define IrSR_IR_TRANSMIT_ON (\ 62 IrSR_RXPL_NEG_IS_ZERO | \ 63 IrSR_TXPL_POS_IS_ZERO | \ 64 IrSR_XMODE_PULSE_3_16 | \ 65 IrSR_RCVEIR_UART_MODE | \ 66 IrSR_XMITIR_IR_MODE) 67 68struct pxa_irda { 69 int speed; 70 int newspeed; 71 unsigned long last_oscr; 72 73 unsigned char *dma_rx_buff; 74 unsigned char *dma_tx_buff; 75 dma_addr_t dma_rx_buff_phy; 76 dma_addr_t dma_tx_buff_phy; 77 unsigned int dma_tx_buff_len; 78 int txdma; 79 int rxdma; 80 81 struct net_device_stats stats; 82 struct irlap_cb *irlap; 83 struct qos_info qos; 84 85 iobuff_t tx_buff; 86 iobuff_t rx_buff; 87 88 struct device *dev; 89 struct pxaficp_platform_data *pdata; 90}; 91 92 93#define IS_FIR(si) ((si)->speed >= 4000000) 94#define IRDA_FRAME_SIZE_LIMIT 2047 95 96inline static void pxa_irda_fir_dma_rx_start(struct pxa_irda *si) 97{ 98 DCSR(si->rxdma) = DCSR_NODESC; 99 DSADR(si->rxdma) = __PREG(ICDR); 100 DTADR(si->rxdma) = si->dma_rx_buff_phy; 101 DCMD(si->rxdma) = DCMD_INCTRGADDR | DCMD_FLOWSRC | DCMD_WIDTH1 | DCMD_BURST32 | IRDA_FRAME_SIZE_LIMIT; 102 DCSR(si->rxdma) |= DCSR_RUN; 103} 104 105inline static void pxa_irda_fir_dma_tx_start(struct pxa_irda *si) 106{ 107 DCSR(si->txdma) = DCSR_NODESC; 108 DSADR(si->txdma) = si->dma_tx_buff_phy; 109 DTADR(si->txdma) = __PREG(ICDR); 110 DCMD(si->txdma) = DCMD_INCSRCADDR | DCMD_FLOWTRG | DCMD_ENDIRQEN | DCMD_WIDTH1 | DCMD_BURST32 | si->dma_tx_buff_len; 111 DCSR(si->txdma) |= DCSR_RUN; 112} 113 114/* 115 * Set the IrDA communications speed. 116 */ 117static int pxa_irda_set_speed(struct pxa_irda *si, int speed) 118{ 119 unsigned long flags; 120 unsigned int divisor; 121 122 switch (speed) { 123 case 9600: case 19200: case 38400: 124 case 57600: case 115200: 125 126 /* refer to PXA250/210 Developer's Manual 10-7 */ 127 /* BaudRate = 14.7456 MHz / (16*Divisor) */ 128 divisor = 14745600 / (16 * speed); 129 130 local_irq_save(flags); 131 132 if (IS_FIR(si)) { 133 /* stop RX DMA */ 134 DCSR(si->rxdma) &= ~DCSR_RUN; 135 /* disable FICP */ 136 ICCR0 = 0; 137 pxa_set_cken(CKEN_FICP, 0); 138 139 /* set board transceiver to SIR mode */ 140 si->pdata->transceiver_mode(si->dev, IR_SIRMODE); 141 142 /* configure GPIO46/47 */ 143 pxa_gpio_mode(GPIO46_STRXD_MD); 144 pxa_gpio_mode(GPIO47_STTXD_MD); 145 146 /* enable the STUART clock */ 147 pxa_set_cken(CKEN_STUART, 1); 148 } 149 150 /* disable STUART first */ 151 STIER = 0; 152 153 /* access DLL & DLH */ 154 STLCR |= LCR_DLAB; 155 STDLL = divisor & 0xff; 156 STDLH = divisor >> 8; 157 STLCR &= ~LCR_DLAB; 158 159 si->speed = speed; 160 STISR = IrSR_IR_RECEIVE_ON | IrSR_XMODE_PULSE_1_6; 161 STIER = IER_UUE | IER_RLSE | IER_RAVIE | IER_RTIOE; 162 163 local_irq_restore(flags); 164 break; 165 166 case 4000000: 167 local_irq_save(flags); 168 169 /* disable STUART */ 170 STIER = 0; 171 STISR = 0; 172 pxa_set_cken(CKEN_STUART, 0); 173 174 /* disable FICP first */ 175 ICCR0 = 0; 176 177 /* set board transceiver to FIR mode */ 178 si->pdata->transceiver_mode(si->dev, IR_FIRMODE); 179 180 /* configure GPIO46/47 */ 181 pxa_gpio_mode(GPIO46_ICPRXD_MD); 182 pxa_gpio_mode(GPIO47_ICPTXD_MD); 183 184 /* enable the FICP clock */ 185 pxa_set_cken(CKEN_FICP, 1); 186 187 si->speed = speed; 188 pxa_irda_fir_dma_rx_start(si); 189 ICCR0 = ICCR0_ITR | ICCR0_RXE; 190 191 local_irq_restore(flags); 192 break; 193 194 default: 195 return -EINVAL; 196 } 197 198 return 0; 199} 200 201/* SIR interrupt service routine. */ 202static irqreturn_t pxa_irda_sir_irq(int irq, void *dev_id) 203{ 204 struct net_device *dev = dev_id; 205 struct pxa_irda *si = netdev_priv(dev); 206 int iir, lsr, data; 207 208 iir = STIIR; 209 210 switch (iir & 0x0F) { 211 case 0x06: /* Receiver Line Status */ 212 lsr = STLSR; 213 while (lsr & LSR_FIFOE) { 214 data = STRBR; 215 if (lsr & (LSR_OE | LSR_PE | LSR_FE | LSR_BI)) { 216 printk(KERN_DEBUG "pxa_ir: sir receiving error\n"); 217 si->stats.rx_errors++; 218 if (lsr & LSR_FE) 219 si->stats.rx_frame_errors++; 220 if (lsr & LSR_OE) 221 si->stats.rx_fifo_errors++; 222 } else { 223 si->stats.rx_bytes++; 224 async_unwrap_char(dev, &si->stats, &si->rx_buff, data); 225 } 226 lsr = STLSR; 227 } 228 dev->last_rx = jiffies; 229 si->last_oscr = OSCR; 230 break; 231 232 case 0x04: /* Received Data Available */ 233 /* forth through */ 234 235 case 0x0C: /* Character Timeout Indication */ 236 do { 237 si->stats.rx_bytes++; 238 async_unwrap_char(dev, &si->stats, &si->rx_buff, STRBR); 239 } while (STLSR & LSR_DR); 240 dev->last_rx = jiffies; 241 si->last_oscr = OSCR; 242 break; 243 244 case 0x02: /* Transmit FIFO Data Request */ 245 while ((si->tx_buff.len) && (STLSR & LSR_TDRQ)) { 246 STTHR = *si->tx_buff.data++; 247 si->tx_buff.len -= 1; 248 } 249 250 if (si->tx_buff.len == 0) { 251 si->stats.tx_packets++; 252 si->stats.tx_bytes += si->tx_buff.data - 253 si->tx_buff.head; 254 255 /* We need to ensure that the transmitter has finished. */ 256 while ((STLSR & LSR_TEMT) == 0) 257 cpu_relax(); 258 si->last_oscr = OSCR; 259 260 /* 261 * Ok, we've finished transmitting. Now enable 262 * the receiver. Sometimes we get a receive IRQ 263 * immediately after a transmit... 264 */ 265 if (si->newspeed) { 266 pxa_irda_set_speed(si, si->newspeed); 267 si->newspeed = 0; 268 } else { 269 /* enable IR Receiver, disable IR Transmitter */ 270 STISR = IrSR_IR_RECEIVE_ON | IrSR_XMODE_PULSE_1_6; 271 /* enable STUART and receive interrupts */ 272 STIER = IER_UUE | IER_RLSE | IER_RAVIE | IER_RTIOE; 273 } 274 /* I'm hungry! */ 275 netif_wake_queue(dev); 276 } 277 break; 278 } 279 280 return IRQ_HANDLED; 281} 282 283/* FIR Receive DMA interrupt handler */ 284static void pxa_irda_fir_dma_rx_irq(int channel, void *data) 285{ 286 int dcsr = DCSR(channel); 287 288 DCSR(channel) = dcsr & ~DCSR_RUN; 289 290 printk(KERN_DEBUG "pxa_ir: fir rx dma bus error %#x\n", dcsr); 291} 292 293/* FIR Transmit DMA interrupt handler */ 294static void pxa_irda_fir_dma_tx_irq(int channel, void *data) 295{ 296 struct net_device *dev = data; 297 struct pxa_irda *si = netdev_priv(dev); 298 int dcsr; 299 300 dcsr = DCSR(channel); 301 DCSR(channel) = dcsr & ~DCSR_RUN; 302 303 if (dcsr & DCSR_ENDINTR) { 304 si->stats.tx_packets++; 305 si->stats.tx_bytes += si->dma_tx_buff_len; 306 } else { 307 si->stats.tx_errors++; 308 } 309 310 while (ICSR1 & ICSR1_TBY) 311 cpu_relax(); 312 si->last_oscr = OSCR; 313 314 /* 315 * HACK: It looks like the TBY bit is dropped too soon. 316 * Without this delay things break. 317 */ 318 udelay(120); 319 320 if (si->newspeed) { 321 pxa_irda_set_speed(si, si->newspeed); 322 si->newspeed = 0; 323 } else { 324 int i = 64; 325 326 ICCR0 = 0; 327 pxa_irda_fir_dma_rx_start(si); 328 while ((ICSR1 & ICSR1_RNE) && i--) 329 (void)ICDR; 330 ICCR0 = ICCR0_ITR | ICCR0_RXE; 331 332 if (i < 0) 333 printk(KERN_ERR "pxa_ir: cannot clear Rx FIFO!\n"); 334 } 335 netif_wake_queue(dev); 336} 337 338/* EIF(Error in FIFO/End in Frame) handler for FIR */ 339static void pxa_irda_fir_irq_eif(struct pxa_irda *si, struct net_device *dev, int icsr0) 340{ 341 unsigned int len, stat, data; 342 343 /* Get the current data position. */ 344 len = DTADR(si->rxdma) - si->dma_rx_buff_phy; 345 346 do { 347 /* Read Status, and then Data. */ 348 stat = ICSR1; 349 rmb(); 350 data = ICDR; 351 352 if (stat & (ICSR1_CRE | ICSR1_ROR)) { 353 si->stats.rx_errors++; 354 if (stat & ICSR1_CRE) { 355 printk(KERN_DEBUG "pxa_ir: fir receive CRC error\n"); 356 si->stats.rx_crc_errors++; 357 } 358 if (stat & ICSR1_ROR) { 359 printk(KERN_DEBUG "pxa_ir: fir receive overrun\n"); 360 si->stats.rx_over_errors++; 361 } 362 } else { 363 si->dma_rx_buff[len++] = data; 364 } 365 /* If we hit the end of frame, there's no point in continuing. */ 366 if (stat & ICSR1_EOF) 367 break; 368 } while (ICSR0 & ICSR0_EIF); 369 370 if (stat & ICSR1_EOF) { 371 /* end of frame. */ 372 struct sk_buff *skb; 373 374 if (icsr0 & ICSR0_FRE) { 375 printk(KERN_ERR "pxa_ir: dropping erroneous frame\n"); 376 si->stats.rx_dropped++; 377 return; 378 } 379 380 skb = alloc_skb(len+1,GFP_ATOMIC); 381 if (!skb) { 382 printk(KERN_ERR "pxa_ir: fir out of memory for receive skb\n"); 383 si->stats.rx_dropped++; 384 return; 385 } 386 387 /* Align IP header to 20 bytes */ 388 skb_reserve(skb, 1); 389 skb_copy_to_linear_data(skb, si->dma_rx_buff, len); 390 skb_put(skb, len); 391 392 /* Feed it to IrLAP */ 393 skb->dev = dev; 394 skb_reset_mac_header(skb); 395 skb->protocol = htons(ETH_P_IRDA); 396 netif_rx(skb); 397 398 si->stats.rx_packets++; 399 si->stats.rx_bytes += len; 400 401 dev->last_rx = jiffies; 402 } 403} 404 405/* FIR interrupt handler */ 406static irqreturn_t pxa_irda_fir_irq(int irq, void *dev_id) 407{ 408 struct net_device *dev = dev_id; 409 struct pxa_irda *si = netdev_priv(dev); 410 int icsr0, i = 64; 411 412 /* stop RX DMA */ 413 DCSR(si->rxdma) &= ~DCSR_RUN; 414 si->last_oscr = OSCR; 415 icsr0 = ICSR0; 416 417 if (icsr0 & (ICSR0_FRE | ICSR0_RAB)) { 418 if (icsr0 & ICSR0_FRE) { 419 printk(KERN_DEBUG "pxa_ir: fir receive frame error\n"); 420 si->stats.rx_frame_errors++; 421 } else { 422 printk(KERN_DEBUG "pxa_ir: fir receive abort\n"); 423 si->stats.rx_errors++; 424 } 425 ICSR0 = icsr0 & (ICSR0_FRE | ICSR0_RAB); 426 } 427 428 if (icsr0 & ICSR0_EIF) { 429 /* An error in FIFO occured, or there is a end of frame */ 430 pxa_irda_fir_irq_eif(si, dev, icsr0); 431 } 432 433 ICCR0 = 0; 434 pxa_irda_fir_dma_rx_start(si); 435 while ((ICSR1 & ICSR1_RNE) && i--) 436 (void)ICDR; 437 ICCR0 = ICCR0_ITR | ICCR0_RXE; 438 439 if (i < 0) 440 printk(KERN_ERR "pxa_ir: cannot clear Rx FIFO!\n"); 441 442 return IRQ_HANDLED; 443} 444 445/* hard_xmit interface of irda device */ 446static int pxa_irda_hard_xmit(struct sk_buff *skb, struct net_device *dev) 447{ 448 struct pxa_irda *si = netdev_priv(dev); 449 int speed = irda_get_next_speed(skb); 450 451 /* 452 * Does this packet contain a request to change the interface 453 * speed? If so, remember it until we complete the transmission 454 * of this frame. 455 */ 456 if (speed != si->speed && speed != -1) 457 si->newspeed = speed; 458 459 /* 460 * If this is an empty frame, we can bypass a lot. 461 */ 462 if (skb->len == 0) { 463 if (si->newspeed) { 464 si->newspeed = 0; 465 pxa_irda_set_speed(si, speed); 466 } 467 dev_kfree_skb(skb); 468 return 0; 469 } 470 471 netif_stop_queue(dev); 472 473 if (!IS_FIR(si)) { 474 si->tx_buff.data = si->tx_buff.head; 475 si->tx_buff.len = async_wrap_skb(skb, si->tx_buff.data, si->tx_buff.truesize); 476 477 /* Disable STUART interrupts and switch to transmit mode. */ 478 STIER = 0; 479 STISR = IrSR_IR_TRANSMIT_ON | IrSR_XMODE_PULSE_1_6; 480 481 /* enable STUART and transmit interrupts */ 482 STIER = IER_UUE | IER_TIE; 483 } else { 484 unsigned long mtt = irda_get_mtt(skb); 485 486 si->dma_tx_buff_len = skb->len; 487 skb_copy_from_linear_data(skb, si->dma_tx_buff, skb->len); 488 489 if (mtt) 490 while ((unsigned)(OSCR - si->last_oscr)/4 < mtt) 491 cpu_relax(); 492 493 /* stop RX DMA, disable FICP */ 494 DCSR(si->rxdma) &= ~DCSR_RUN; 495 ICCR0 = 0; 496 497 pxa_irda_fir_dma_tx_start(si); 498 ICCR0 = ICCR0_ITR | ICCR0_TXE; 499 } 500 501 dev_kfree_skb(skb); 502 dev->trans_start = jiffies; 503 return 0; 504} 505 506static int pxa_irda_ioctl(struct net_device *dev, struct ifreq *ifreq, int cmd) 507{ 508 struct if_irda_req *rq = (struct if_irda_req *)ifreq; 509 struct pxa_irda *si = netdev_priv(dev); 510 int ret; 511 512 switch (cmd) { 513 case SIOCSBANDWIDTH: 514 ret = -EPERM; 515 if (capable(CAP_NET_ADMIN)) { 516 /* 517 * We are unable to set the speed if the 518 * device is not running. 519 */ 520 if (netif_running(dev)) { 521 ret = pxa_irda_set_speed(si, 522 rq->ifr_baudrate); 523 } else { 524 printk(KERN_INFO "pxa_ir: SIOCSBANDWIDTH: !netif_running\n"); 525 ret = 0; 526 } 527 } 528 break; 529 530 case SIOCSMEDIABUSY: 531 ret = -EPERM; 532 if (capable(CAP_NET_ADMIN)) { 533 irda_device_set_media_busy(dev, TRUE); 534 ret = 0; 535 } 536 break; 537 538 case SIOCGRECEIVING: 539 ret = 0; 540 rq->ifr_receiving = IS_FIR(si) ? 0 541 : si->rx_buff.state != OUTSIDE_FRAME; 542 break; 543 544 default: 545 ret = -EOPNOTSUPP; 546 break; 547 } 548 549 return ret; 550} 551 552static struct net_device_stats *pxa_irda_stats(struct net_device *dev) 553{ 554 struct pxa_irda *si = netdev_priv(dev); 555 return &si->stats; 556} 557 558static void pxa_irda_startup(struct pxa_irda *si) 559{ 560 /* Disable STUART interrupts */ 561 STIER = 0; 562 /* enable STUART interrupt to the processor */ 563 STMCR = MCR_OUT2; 564 /* configure SIR frame format: StartBit - Data 7 ... Data 0 - Stop Bit */ 565 STLCR = LCR_WLS0 | LCR_WLS1; 566 /* enable FIFO, we use FIFO to improve performance */ 567 STFCR = FCR_TRFIFOE | FCR_ITL_32; 568 569 /* disable FICP */ 570 ICCR0 = 0; 571 /* configure FICP ICCR2 */ 572 ICCR2 = ICCR2_TXP | ICCR2_TRIG_32; 573 574 /* configure DMAC */ 575 DRCMR17 = si->rxdma | DRCMR_MAPVLD; 576 DRCMR18 = si->txdma | DRCMR_MAPVLD; 577 578 /* force SIR reinitialization */ 579 si->speed = 4000000; 580 pxa_irda_set_speed(si, 9600); 581 582 printk(KERN_DEBUG "pxa_ir: irda startup\n"); 583} 584 585static void pxa_irda_shutdown(struct pxa_irda *si) 586{ 587 unsigned long flags; 588 589 local_irq_save(flags); 590 591 /* disable STUART and interrupt */ 592 STIER = 0; 593 /* disable STUART SIR mode */ 594 STISR = 0; 595 /* disable the STUART clock */ 596 pxa_set_cken(CKEN_STUART, 0); 597 598 /* disable DMA */ 599 DCSR(si->txdma) &= ~DCSR_RUN; 600 DCSR(si->rxdma) &= ~DCSR_RUN; 601 /* disable FICP */ 602 ICCR0 = 0; 603 /* disable the FICP clock */ 604 pxa_set_cken(CKEN_FICP, 0); 605 606 DRCMR17 = 0; 607 DRCMR18 = 0; 608 609 local_irq_restore(flags); 610 611 /* power off board transceiver */ 612 si->pdata->transceiver_mode(si->dev, IR_OFF); 613 614 printk(KERN_DEBUG "pxa_ir: irda shutdown\n"); 615} 616 617static int pxa_irda_start(struct net_device *dev) 618{ 619 struct pxa_irda *si = netdev_priv(dev); 620 int err; 621 622 si->speed = 9600; 623 624 err = request_irq(IRQ_STUART, pxa_irda_sir_irq, 0, dev->name, dev); 625 if (err) 626 goto err_irq1; 627 628 err = request_irq(IRQ_ICP, pxa_irda_fir_irq, 0, dev->name, dev); 629 if (err) 630 goto err_irq2; 631 632 /* 633 * The interrupt must remain disabled for now. 634 */ 635 disable_irq(IRQ_STUART); 636 disable_irq(IRQ_ICP); 637 638 err = -EBUSY; 639 si->rxdma = pxa_request_dma("FICP_RX",DMA_PRIO_LOW, pxa_irda_fir_dma_rx_irq, dev); 640 if (si->rxdma < 0) 641 goto err_rx_dma; 642 643 si->txdma = pxa_request_dma("FICP_TX",DMA_PRIO_LOW, pxa_irda_fir_dma_tx_irq, dev); 644 if (si->txdma < 0) 645 goto err_tx_dma; 646 647 err = -ENOMEM; 648 si->dma_rx_buff = dma_alloc_coherent(si->dev, IRDA_FRAME_SIZE_LIMIT, 649 &si->dma_rx_buff_phy, GFP_KERNEL ); 650 if (!si->dma_rx_buff) 651 goto err_dma_rx_buff; 652 653 si->dma_tx_buff = dma_alloc_coherent(si->dev, IRDA_FRAME_SIZE_LIMIT, 654 &si->dma_tx_buff_phy, GFP_KERNEL ); 655 if (!si->dma_tx_buff) 656 goto err_dma_tx_buff; 657 658 /* Setup the serial port for the initial speed. */ 659 pxa_irda_startup(si); 660 661 /* 662 * Open a new IrLAP layer instance. 663 */ 664 si->irlap = irlap_open(dev, &si->qos, "pxa"); 665 err = -ENOMEM; 666 if (!si->irlap) 667 goto err_irlap; 668 669 /* 670 * Now enable the interrupt and start the queue 671 */ 672 enable_irq(IRQ_STUART); 673 enable_irq(IRQ_ICP); 674 netif_start_queue(dev); 675 676 printk(KERN_DEBUG "pxa_ir: irda driver opened\n"); 677 678 return 0; 679 680err_irlap: 681 pxa_irda_shutdown(si); 682 dma_free_coherent(si->dev, IRDA_FRAME_SIZE_LIMIT, si->dma_tx_buff, si->dma_tx_buff_phy); 683err_dma_tx_buff: 684 dma_free_coherent(si->dev, IRDA_FRAME_SIZE_LIMIT, si->dma_rx_buff, si->dma_rx_buff_phy); 685err_dma_rx_buff: 686 pxa_free_dma(si->txdma); 687err_tx_dma: 688 pxa_free_dma(si->rxdma); 689err_rx_dma: 690 free_irq(IRQ_ICP, dev); 691err_irq2: 692 free_irq(IRQ_STUART, dev); 693err_irq1: 694 695 return err; 696} 697 698static int pxa_irda_stop(struct net_device *dev) 699{ 700 struct pxa_irda *si = netdev_priv(dev); 701 702 netif_stop_queue(dev); 703 704 pxa_irda_shutdown(si); 705 706 /* Stop IrLAP */ 707 if (si->irlap) { 708 irlap_close(si->irlap); 709 si->irlap = NULL; 710 } 711 712 free_irq(IRQ_STUART, dev); 713 free_irq(IRQ_ICP, dev); 714 715 pxa_free_dma(si->rxdma); 716 pxa_free_dma(si->txdma); 717 718 if (si->dma_rx_buff) 719 dma_free_coherent(si->dev, IRDA_FRAME_SIZE_LIMIT, si->dma_tx_buff, si->dma_tx_buff_phy); 720 if (si->dma_tx_buff) 721 dma_free_coherent(si->dev, IRDA_FRAME_SIZE_LIMIT, si->dma_rx_buff, si->dma_rx_buff_phy); 722 723 printk(KERN_DEBUG "pxa_ir: irda driver closed\n"); 724 return 0; 725} 726 727static int pxa_irda_suspend(struct platform_device *_dev, pm_message_t state) 728{ 729 struct net_device *dev = platform_get_drvdata(_dev); 730 struct pxa_irda *si; 731 732 if (dev && netif_running(dev)) { 733 si = netdev_priv(dev); 734 netif_device_detach(dev); 735 pxa_irda_shutdown(si); 736 } 737 738 return 0; 739} 740 741static int pxa_irda_resume(struct platform_device *_dev) 742{ 743 struct net_device *dev = platform_get_drvdata(_dev); 744 struct pxa_irda *si; 745 746 if (dev && netif_running(dev)) { 747 si = netdev_priv(dev); 748 pxa_irda_startup(si); 749 netif_device_attach(dev); 750 netif_wake_queue(dev); 751 } 752 753 return 0; 754} 755 756 757static int pxa_irda_init_iobuf(iobuff_t *io, int size) 758{ 759 io->head = kmalloc(size, GFP_KERNEL | GFP_DMA); 760 if (io->head != NULL) { 761 io->truesize = size; 762 io->in_frame = FALSE; 763 io->state = OUTSIDE_FRAME; 764 io->data = io->head; 765 } 766 return io->head ? 0 : -ENOMEM; 767} 768 769static int pxa_irda_probe(struct platform_device *pdev) 770{ 771 struct net_device *dev; 772 struct pxa_irda *si; 773 unsigned int baudrate_mask; 774 int err; 775 776 if (!pdev->dev.platform_data) 777 return -ENODEV; 778 779 err = request_mem_region(__PREG(STUART), 0x24, "IrDA") ? 0 : -EBUSY; 780 if (err) 781 goto err_mem_1; 782 783 err = request_mem_region(__PREG(FICP), 0x1c, "IrDA") ? 0 : -EBUSY; 784 if (err) 785 goto err_mem_2; 786 787 dev = alloc_irdadev(sizeof(struct pxa_irda)); 788 if (!dev) 789 goto err_mem_3; 790 791 si = netdev_priv(dev); 792 si->dev = &pdev->dev; 793 si->pdata = pdev->dev.platform_data; 794 795 /* 796 * Initialise the SIR buffers 797 */ 798 err = pxa_irda_init_iobuf(&si->rx_buff, 14384); 799 if (err) 800 goto err_mem_4; 801 err = pxa_irda_init_iobuf(&si->tx_buff, 4000); 802 if (err) 803 goto err_mem_5; 804 805 dev->hard_start_xmit = pxa_irda_hard_xmit; 806 dev->open = pxa_irda_start; 807 dev->stop = pxa_irda_stop; 808 dev->do_ioctl = pxa_irda_ioctl; 809 dev->get_stats = pxa_irda_stats; 810 811 irda_init_max_qos_capabilies(&si->qos); 812 813 baudrate_mask = 0; 814 if (si->pdata->transceiver_cap & IR_SIRMODE) 815 baudrate_mask |= IR_9600|IR_19200|IR_38400|IR_57600|IR_115200; 816 if (si->pdata->transceiver_cap & IR_FIRMODE) 817 baudrate_mask |= IR_4000000 << 8; 818 819 si->qos.baud_rate.bits &= baudrate_mask; 820 si->qos.min_turn_time.bits = 7; /* 1ms or more */ 821 822 irda_qos_bits_to_value(&si->qos); 823 824 err = register_netdev(dev); 825 826 if (err == 0) 827 dev_set_drvdata(&pdev->dev, dev); 828 829 if (err) { 830 kfree(si->tx_buff.head); 831err_mem_5: 832 kfree(si->rx_buff.head); 833err_mem_4: 834 free_netdev(dev); 835err_mem_3: 836 release_mem_region(__PREG(FICP), 0x1c); 837err_mem_2: 838 release_mem_region(__PREG(STUART), 0x24); 839 } 840err_mem_1: 841 return err; 842} 843 844static int pxa_irda_remove(struct platform_device *_dev) 845{ 846 struct net_device *dev = platform_get_drvdata(_dev); 847 848 if (dev) { 849 struct pxa_irda *si = netdev_priv(dev); 850 unregister_netdev(dev); 851 kfree(si->tx_buff.head); 852 kfree(si->rx_buff.head); 853 free_netdev(dev); 854 } 855 856 release_mem_region(__PREG(STUART), 0x24); 857 release_mem_region(__PREG(FICP), 0x1c); 858 859 return 0; 860} 861 862static struct platform_driver pxa_ir_driver = { 863 .driver = { 864 .name = "pxa2xx-ir", 865 }, 866 .probe = pxa_irda_probe, 867 .remove = pxa_irda_remove, 868 .suspend = pxa_irda_suspend, 869 .resume = pxa_irda_resume, 870}; 871 872static int __init pxa_irda_init(void) 873{ 874 return platform_driver_register(&pxa_ir_driver); 875} 876 877static void __exit pxa_irda_exit(void) 878{ 879 platform_driver_unregister(&pxa_ir_driver); 880} 881 882module_init(pxa_irda_init); 883module_exit(pxa_irda_exit); 884 885MODULE_LICENSE("GPL"); 886