1/* 2 * Network device driver for Cell Processor-Based Blade and Celleb platform 3 * 4 * (C) Copyright IBM Corp. 2005 5 * (C) Copyright 2006 TOSHIBA CORPORATION 6 * 7 * Authors : Utz Bacher <utz.bacher@de.ibm.com> 8 * Jens Osterkamp <Jens.Osterkamp@de.ibm.com> 9 * 10 * This program is free software; you can redistribute it and/or modify 11 * it under the terms of the GNU General Public License as published by 12 * the Free Software Foundation; either version 2, or (at your option) 13 * any later version. 14 * 15 * This program is distributed in the hope that it will be useful, 16 * but WITHOUT ANY WARRANTY; without even the implied warranty of 17 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the 18 * GNU General Public License for more details. 19 * 20 * You should have received a copy of the GNU General Public License 21 * along with this program; if not, write to the Free Software 22 * Foundation, Inc., 675 Mass Ave, Cambridge, MA 02139, USA. 23 */ 24 25#include <linux/compiler.h> 26#include <linux/crc32.h> 27#include <linux/delay.h> 28#include <linux/etherdevice.h> 29#include <linux/ethtool.h> 30#include <linux/firmware.h> 31#include <linux/if_vlan.h> 32#include <linux/in.h> 33#include <linux/init.h> 34#include <linux/gfp.h> 35#include <linux/ioport.h> 36#include <linux/ip.h> 37#include <linux/kernel.h> 38#include <linux/mii.h> 39#include <linux/module.h> 40#include <linux/netdevice.h> 41#include <linux/device.h> 42#include <linux/pci.h> 43#include <linux/skbuff.h> 44#include <linux/tcp.h> 45#include <linux/types.h> 46#include <linux/vmalloc.h> 47#include <linux/wait.h> 48#include <linux/workqueue.h> 49#include <linux/bitops.h> 50#include <asm/pci-bridge.h> 51#include <net/checksum.h> 52 53#include "spider_net.h" 54 55MODULE_AUTHOR("Utz Bacher <utz.bacher@de.ibm.com> and Jens Osterkamp " \ 56 "<Jens.Osterkamp@de.ibm.com>"); 57MODULE_DESCRIPTION("Spider Southbridge Gigabit Ethernet driver"); 58MODULE_LICENSE("GPL"); 59MODULE_VERSION(VERSION); 60MODULE_FIRMWARE(SPIDER_NET_FIRMWARE_NAME); 61 62static int rx_descriptors = SPIDER_NET_RX_DESCRIPTORS_DEFAULT; 63static int tx_descriptors = SPIDER_NET_TX_DESCRIPTORS_DEFAULT; 64 65module_param(rx_descriptors, int, 0444); 66module_param(tx_descriptors, int, 0444); 67 68MODULE_PARM_DESC(rx_descriptors, "number of descriptors used " \ 69 "in rx chains"); 70MODULE_PARM_DESC(tx_descriptors, "number of descriptors used " \ 71 "in tx chain"); 72 73char spider_net_driver_name[] = "spidernet"; 74 75static DEFINE_PCI_DEVICE_TABLE(spider_net_pci_tbl) = { 76 { PCI_VENDOR_ID_TOSHIBA_2, PCI_DEVICE_ID_TOSHIBA_SPIDER_NET, 77 PCI_ANY_ID, PCI_ANY_ID, 0, 0, 0UL }, 78 { 0, } 79}; 80 81MODULE_DEVICE_TABLE(pci, spider_net_pci_tbl); 82 83/** 84 * spider_net_read_reg - reads an SMMIO register of a card 85 * @card: device structure 86 * @reg: register to read from 87 * 88 * returns the content of the specified SMMIO register. 89 */ 90static inline u32 91spider_net_read_reg(struct spider_net_card *card, u32 reg) 92{ 93 /* We use the powerpc specific variants instead of readl_be() because 94 * we know spidernet is not a real PCI device and we can thus avoid the 95 * performance hit caused by the PCI workarounds. 96 */ 97 return in_be32(card->regs + reg); 98} 99 100/** 101 * spider_net_write_reg - writes to an SMMIO register of a card 102 * @card: device structure 103 * @reg: register to write to 104 * @value: value to write into the specified SMMIO register 105 */ 106static inline void 107spider_net_write_reg(struct spider_net_card *card, u32 reg, u32 value) 108{ 109 /* We use the powerpc specific variants instead of writel_be() because 110 * we know spidernet is not a real PCI device and we can thus avoid the 111 * performance hit caused by the PCI workarounds. 112 */ 113 out_be32(card->regs + reg, value); 114} 115 116/** spider_net_write_phy - write to phy register 117 * @netdev: adapter to be written to 118 * @mii_id: id of MII 119 * @reg: PHY register 120 * @val: value to be written to phy register 121 * 122 * spider_net_write_phy_register writes to an arbitrary PHY 123 * register via the spider GPCWOPCMD register. We assume the queue does 124 * not run full (not more than 15 commands outstanding). 125 **/ 126static void 127spider_net_write_phy(struct net_device *netdev, int mii_id, 128 int reg, int val) 129{ 130 struct spider_net_card *card = netdev_priv(netdev); 131 u32 writevalue; 132 133 writevalue = ((u32)mii_id << 21) | 134 ((u32)reg << 16) | ((u32)val); 135 136 spider_net_write_reg(card, SPIDER_NET_GPCWOPCMD, writevalue); 137} 138 139/** spider_net_read_phy - read from phy register 140 * @netdev: network device to be read from 141 * @mii_id: id of MII 142 * @reg: PHY register 143 * 144 * Returns value read from PHY register 145 * 146 * spider_net_write_phy reads from an arbitrary PHY 147 * register via the spider GPCROPCMD register 148 **/ 149static int 150spider_net_read_phy(struct net_device *netdev, int mii_id, int reg) 151{ 152 struct spider_net_card *card = netdev_priv(netdev); 153 u32 readvalue; 154 155 readvalue = ((u32)mii_id << 21) | ((u32)reg << 16); 156 spider_net_write_reg(card, SPIDER_NET_GPCROPCMD, readvalue); 157 158 /* we don't use semaphores to wait for an SPIDER_NET_GPROPCMPINT 159 * interrupt, as we poll for the completion of the read operation 160 * in spider_net_read_phy. Should take about 50 us */ 161 do { 162 readvalue = spider_net_read_reg(card, SPIDER_NET_GPCROPCMD); 163 } while (readvalue & SPIDER_NET_GPREXEC); 164 165 readvalue &= SPIDER_NET_GPRDAT_MASK; 166 167 return readvalue; 168} 169 170/** 171 * spider_net_setup_aneg - initial auto-negotiation setup 172 * @card: device structure 173 **/ 174static void 175spider_net_setup_aneg(struct spider_net_card *card) 176{ 177 struct mii_phy *phy = &card->phy; 178 u32 advertise = 0; 179 u16 bmsr, estat; 180 181 bmsr = spider_net_read_phy(card->netdev, phy->mii_id, MII_BMSR); 182 estat = spider_net_read_phy(card->netdev, phy->mii_id, MII_ESTATUS); 183 184 if (bmsr & BMSR_10HALF) 185 advertise |= ADVERTISED_10baseT_Half; 186 if (bmsr & BMSR_10FULL) 187 advertise |= ADVERTISED_10baseT_Full; 188 if (bmsr & BMSR_100HALF) 189 advertise |= ADVERTISED_100baseT_Half; 190 if (bmsr & BMSR_100FULL) 191 advertise |= ADVERTISED_100baseT_Full; 192 193 if ((bmsr & BMSR_ESTATEN) && (estat & ESTATUS_1000_TFULL)) 194 advertise |= SUPPORTED_1000baseT_Full; 195 if ((bmsr & BMSR_ESTATEN) && (estat & ESTATUS_1000_THALF)) 196 advertise |= SUPPORTED_1000baseT_Half; 197 198 mii_phy_probe(phy, phy->mii_id); 199 phy->def->ops->setup_aneg(phy, advertise); 200 201} 202 203/** 204 * spider_net_rx_irq_off - switch off rx irq on this spider card 205 * @card: device structure 206 * 207 * switches off rx irq by masking them out in the GHIINTnMSK register 208 */ 209static void 210spider_net_rx_irq_off(struct spider_net_card *card) 211{ 212 u32 regvalue; 213 214 regvalue = SPIDER_NET_INT0_MASK_VALUE & (~SPIDER_NET_RXINT); 215 spider_net_write_reg(card, SPIDER_NET_GHIINT0MSK, regvalue); 216} 217 218/** 219 * spider_net_rx_irq_on - switch on rx irq on this spider card 220 * @card: device structure 221 * 222 * switches on rx irq by enabling them in the GHIINTnMSK register 223 */ 224static void 225spider_net_rx_irq_on(struct spider_net_card *card) 226{ 227 u32 regvalue; 228 229 regvalue = SPIDER_NET_INT0_MASK_VALUE | SPIDER_NET_RXINT; 230 spider_net_write_reg(card, SPIDER_NET_GHIINT0MSK, regvalue); 231} 232 233/** 234 * spider_net_set_promisc - sets the unicast address or the promiscuous mode 235 * @card: card structure 236 * 237 * spider_net_set_promisc sets the unicast destination address filter and 238 * thus either allows for non-promisc mode or promisc mode 239 */ 240static void 241spider_net_set_promisc(struct spider_net_card *card) 242{ 243 u32 macu, macl; 244 struct net_device *netdev = card->netdev; 245 246 if (netdev->flags & IFF_PROMISC) { 247 /* clear destination entry 0 */ 248 spider_net_write_reg(card, SPIDER_NET_GMRUAFILnR, 0); 249 spider_net_write_reg(card, SPIDER_NET_GMRUAFILnR + 0x04, 0); 250 spider_net_write_reg(card, SPIDER_NET_GMRUA0FIL15R, 251 SPIDER_NET_PROMISC_VALUE); 252 } else { 253 macu = netdev->dev_addr[0]; 254 macu <<= 8; 255 macu |= netdev->dev_addr[1]; 256 memcpy(&macl, &netdev->dev_addr[2], sizeof(macl)); 257 258 macu |= SPIDER_NET_UA_DESCR_VALUE; 259 spider_net_write_reg(card, SPIDER_NET_GMRUAFILnR, macu); 260 spider_net_write_reg(card, SPIDER_NET_GMRUAFILnR + 0x04, macl); 261 spider_net_write_reg(card, SPIDER_NET_GMRUA0FIL15R, 262 SPIDER_NET_NONPROMISC_VALUE); 263 } 264} 265 266/** 267 * spider_net_get_mac_address - read mac address from spider card 268 * @card: device structure 269 * 270 * reads MAC address from GMACUNIMACU and GMACUNIMACL registers 271 */ 272static int 273spider_net_get_mac_address(struct net_device *netdev) 274{ 275 struct spider_net_card *card = netdev_priv(netdev); 276 u32 macl, macu; 277 278 macl = spider_net_read_reg(card, SPIDER_NET_GMACUNIMACL); 279 macu = spider_net_read_reg(card, SPIDER_NET_GMACUNIMACU); 280 281 netdev->dev_addr[0] = (macu >> 24) & 0xff; 282 netdev->dev_addr[1] = (macu >> 16) & 0xff; 283 netdev->dev_addr[2] = (macu >> 8) & 0xff; 284 netdev->dev_addr[3] = macu & 0xff; 285 netdev->dev_addr[4] = (macl >> 8) & 0xff; 286 netdev->dev_addr[5] = macl & 0xff; 287 288 if (!is_valid_ether_addr(&netdev->dev_addr[0])) 289 return -EINVAL; 290 291 return 0; 292} 293 294/** 295 * spider_net_get_descr_status -- returns the status of a descriptor 296 * @descr: descriptor to look at 297 * 298 * returns the status as in the dmac_cmd_status field of the descriptor 299 */ 300static inline int 301spider_net_get_descr_status(struct spider_net_hw_descr *hwdescr) 302{ 303 return hwdescr->dmac_cmd_status & SPIDER_NET_DESCR_IND_PROC_MASK; 304} 305 306/** 307 * spider_net_free_chain - free descriptor chain 308 * @card: card structure 309 * @chain: address of chain 310 * 311 */ 312static void 313spider_net_free_chain(struct spider_net_card *card, 314 struct spider_net_descr_chain *chain) 315{ 316 struct spider_net_descr *descr; 317 318 descr = chain->ring; 319 do { 320 descr->bus_addr = 0; 321 descr->hwdescr->next_descr_addr = 0; 322 descr = descr->next; 323 } while (descr != chain->ring); 324 325 dma_free_coherent(&card->pdev->dev, chain->num_desc, 326 chain->hwring, chain->dma_addr); 327} 328 329/** 330 * spider_net_init_chain - alloc and link descriptor chain 331 * @card: card structure 332 * @chain: address of chain 333 * 334 * We manage a circular list that mirrors the hardware structure, 335 * except that the hardware uses bus addresses. 336 * 337 * Returns 0 on success, <0 on failure 338 */ 339static int 340spider_net_init_chain(struct spider_net_card *card, 341 struct spider_net_descr_chain *chain) 342{ 343 int i; 344 struct spider_net_descr *descr; 345 struct spider_net_hw_descr *hwdescr; 346 dma_addr_t buf; 347 size_t alloc_size; 348 349 alloc_size = chain->num_desc * sizeof(struct spider_net_hw_descr); 350 351 chain->hwring = dma_alloc_coherent(&card->pdev->dev, alloc_size, 352 &chain->dma_addr, GFP_KERNEL); 353 354 if (!chain->hwring) 355 return -ENOMEM; 356 357 memset(chain->ring, 0, chain->num_desc * sizeof(struct spider_net_descr)); 358 359 /* Set up the hardware pointers in each descriptor */ 360 descr = chain->ring; 361 hwdescr = chain->hwring; 362 buf = chain->dma_addr; 363 for (i=0; i < chain->num_desc; i++, descr++, hwdescr++) { 364 hwdescr->dmac_cmd_status = SPIDER_NET_DESCR_NOT_IN_USE; 365 hwdescr->next_descr_addr = 0; 366 367 descr->hwdescr = hwdescr; 368 descr->bus_addr = buf; 369 descr->next = descr + 1; 370 descr->prev = descr - 1; 371 372 buf += sizeof(struct spider_net_hw_descr); 373 } 374 /* do actual circular list */ 375 (descr-1)->next = chain->ring; 376 chain->ring->prev = descr-1; 377 378 spin_lock_init(&chain->lock); 379 chain->head = chain->ring; 380 chain->tail = chain->ring; 381 return 0; 382} 383 384/** 385 * spider_net_free_rx_chain_contents - frees descr contents in rx chain 386 * @card: card structure 387 * 388 * returns 0 on success, <0 on failure 389 */ 390static void 391spider_net_free_rx_chain_contents(struct spider_net_card *card) 392{ 393 struct spider_net_descr *descr; 394 395 descr = card->rx_chain.head; 396 do { 397 if (descr->skb) { 398 pci_unmap_single(card->pdev, descr->hwdescr->buf_addr, 399 SPIDER_NET_MAX_FRAME, 400 PCI_DMA_BIDIRECTIONAL); 401 dev_kfree_skb(descr->skb); 402 descr->skb = NULL; 403 } 404 descr = descr->next; 405 } while (descr != card->rx_chain.head); 406} 407 408/** 409 * spider_net_prepare_rx_descr - Reinitialize RX descriptor 410 * @card: card structure 411 * @descr: descriptor to re-init 412 * 413 * Return 0 on success, <0 on failure. 414 * 415 * Allocates a new rx skb, iommu-maps it and attaches it to the 416 * descriptor. Mark the descriptor as activated, ready-to-use. 417 */ 418static int 419spider_net_prepare_rx_descr(struct spider_net_card *card, 420 struct spider_net_descr *descr) 421{ 422 struct spider_net_hw_descr *hwdescr = descr->hwdescr; 423 dma_addr_t buf; 424 int offset; 425 int bufsize; 426 427 /* we need to round up the buffer size to a multiple of 128 */ 428 bufsize = (SPIDER_NET_MAX_FRAME + SPIDER_NET_RXBUF_ALIGN - 1) & 429 (~(SPIDER_NET_RXBUF_ALIGN - 1)); 430 431 /* and we need to have it 128 byte aligned, therefore we allocate a 432 * bit more */ 433 /* allocate an skb */ 434 descr->skb = netdev_alloc_skb(card->netdev, 435 bufsize + SPIDER_NET_RXBUF_ALIGN - 1); 436 if (!descr->skb) { 437 if (netif_msg_rx_err(card) && net_ratelimit()) 438 dev_err(&card->netdev->dev, 439 "Not enough memory to allocate rx buffer\n"); 440 card->spider_stats.alloc_rx_skb_error++; 441 return -ENOMEM; 442 } 443 hwdescr->buf_size = bufsize; 444 hwdescr->result_size = 0; 445 hwdescr->valid_size = 0; 446 hwdescr->data_status = 0; 447 hwdescr->data_error = 0; 448 449 offset = ((unsigned long)descr->skb->data) & 450 (SPIDER_NET_RXBUF_ALIGN - 1); 451 if (offset) 452 skb_reserve(descr->skb, SPIDER_NET_RXBUF_ALIGN - offset); 453 /* iommu-map the skb */ 454 buf = pci_map_single(card->pdev, descr->skb->data, 455 SPIDER_NET_MAX_FRAME, PCI_DMA_FROMDEVICE); 456 if (pci_dma_mapping_error(card->pdev, buf)) { 457 dev_kfree_skb_any(descr->skb); 458 descr->skb = NULL; 459 if (netif_msg_rx_err(card) && net_ratelimit()) 460 dev_err(&card->netdev->dev, "Could not iommu-map rx buffer\n"); 461 card->spider_stats.rx_iommu_map_error++; 462 hwdescr->dmac_cmd_status = SPIDER_NET_DESCR_NOT_IN_USE; 463 } else { 464 hwdescr->buf_addr = buf; 465 wmb(); 466 hwdescr->dmac_cmd_status = SPIDER_NET_DESCR_CARDOWNED | 467 SPIDER_NET_DMAC_NOINTR_COMPLETE; 468 } 469 470 return 0; 471} 472 473/** 474 * spider_net_enable_rxchtails - sets RX dmac chain tail addresses 475 * @card: card structure 476 * 477 * spider_net_enable_rxchtails sets the RX DMAC chain tail addresses in the 478 * chip by writing to the appropriate register. DMA is enabled in 479 * spider_net_enable_rxdmac. 480 */ 481static inline void 482spider_net_enable_rxchtails(struct spider_net_card *card) 483{ 484 /* assume chain is aligned correctly */ 485 spider_net_write_reg(card, SPIDER_NET_GDADCHA , 486 card->rx_chain.tail->bus_addr); 487} 488 489/** 490 * spider_net_enable_rxdmac - enables a receive DMA controller 491 * @card: card structure 492 * 493 * spider_net_enable_rxdmac enables the DMA controller by setting RX_DMA_EN 494 * in the GDADMACCNTR register 495 */ 496static inline void 497spider_net_enable_rxdmac(struct spider_net_card *card) 498{ 499 wmb(); 500 spider_net_write_reg(card, SPIDER_NET_GDADMACCNTR, 501 SPIDER_NET_DMA_RX_VALUE); 502} 503 504/** 505 * spider_net_disable_rxdmac - disables the receive DMA controller 506 * @card: card structure 507 * 508 * spider_net_disable_rxdmac terminates processing on the DMA controller 509 * by turing off the DMA controller, with the force-end flag set. 510 */ 511static inline void 512spider_net_disable_rxdmac(struct spider_net_card *card) 513{ 514 spider_net_write_reg(card, SPIDER_NET_GDADMACCNTR, 515 SPIDER_NET_DMA_RX_FEND_VALUE); 516} 517 518/** 519 * spider_net_refill_rx_chain - refills descriptors/skbs in the rx chains 520 * @card: card structure 521 * 522 * refills descriptors in the rx chain: allocates skbs and iommu-maps them. 523 */ 524static void 525spider_net_refill_rx_chain(struct spider_net_card *card) 526{ 527 struct spider_net_descr_chain *chain = &card->rx_chain; 528 unsigned long flags; 529 530 /* one context doing the refill (and a second context seeing that 531 * and omitting it) is ok. If called by NAPI, we'll be called again 532 * as spider_net_decode_one_descr is called several times. If some 533 * interrupt calls us, the NAPI is about to clean up anyway. */ 534 if (!spin_trylock_irqsave(&chain->lock, flags)) 535 return; 536 537 while (spider_net_get_descr_status(chain->head->hwdescr) == 538 SPIDER_NET_DESCR_NOT_IN_USE) { 539 if (spider_net_prepare_rx_descr(card, chain->head)) 540 break; 541 chain->head = chain->head->next; 542 } 543 544 spin_unlock_irqrestore(&chain->lock, flags); 545} 546 547/** 548 * spider_net_alloc_rx_skbs - Allocates rx skbs in rx descriptor chains 549 * @card: card structure 550 * 551 * Returns 0 on success, <0 on failure. 552 */ 553static int 554spider_net_alloc_rx_skbs(struct spider_net_card *card) 555{ 556 struct spider_net_descr_chain *chain = &card->rx_chain; 557 struct spider_net_descr *start = chain->tail; 558 struct spider_net_descr *descr = start; 559 560 /* Link up the hardware chain pointers */ 561 do { 562 descr->prev->hwdescr->next_descr_addr = descr->bus_addr; 563 descr = descr->next; 564 } while (descr != start); 565 566 /* Put at least one buffer into the chain. if this fails, 567 * we've got a problem. If not, spider_net_refill_rx_chain 568 * will do the rest at the end of this function. */ 569 if (spider_net_prepare_rx_descr(card, chain->head)) 570 goto error; 571 else 572 chain->head = chain->head->next; 573 574 /* This will allocate the rest of the rx buffers; 575 * if not, it's business as usual later on. */ 576 spider_net_refill_rx_chain(card); 577 spider_net_enable_rxdmac(card); 578 return 0; 579 580error: 581 spider_net_free_rx_chain_contents(card); 582 return -ENOMEM; 583} 584 585/** 586 * spider_net_get_multicast_hash - generates hash for multicast filter table 587 * @addr: multicast address 588 * 589 * returns the hash value. 590 * 591 * spider_net_get_multicast_hash calculates a hash value for a given multicast 592 * address, that is used to set the multicast filter tables 593 */ 594static u8 595spider_net_get_multicast_hash(struct net_device *netdev, __u8 *addr) 596{ 597 u32 crc; 598 u8 hash; 599 char addr_for_crc[ETH_ALEN] = { 0, }; 600 int i, bit; 601 602 for (i = 0; i < ETH_ALEN * 8; i++) { 603 bit = (addr[i / 8] >> (i % 8)) & 1; 604 addr_for_crc[ETH_ALEN - 1 - i / 8] += bit << (7 - (i % 8)); 605 } 606 607 crc = crc32_be(~0, addr_for_crc, netdev->addr_len); 608 609 hash = (crc >> 27); 610 hash <<= 3; 611 hash |= crc & 7; 612 hash &= 0xff; 613 614 return hash; 615} 616 617/** 618 * spider_net_set_multi - sets multicast addresses and promisc flags 619 * @netdev: interface device structure 620 * 621 * spider_net_set_multi configures multicast addresses as needed for the 622 * netdev interface. It also sets up multicast, allmulti and promisc 623 * flags appropriately 624 */ 625static void 626spider_net_set_multi(struct net_device *netdev) 627{ 628 struct netdev_hw_addr *ha; 629 u8 hash; 630 int i; 631 u32 reg; 632 struct spider_net_card *card = netdev_priv(netdev); 633 unsigned long bitmask[SPIDER_NET_MULTICAST_HASHES / BITS_PER_LONG] = 634 {0, }; 635 636 spider_net_set_promisc(card); 637 638 if (netdev->flags & IFF_ALLMULTI) { 639 for (i = 0; i < SPIDER_NET_MULTICAST_HASHES; i++) { 640 set_bit(i, bitmask); 641 } 642 goto write_hash; 643 } 644 645 /* well, we know, what the broadcast hash value is: it's xfd 646 hash = spider_net_get_multicast_hash(netdev, netdev->broadcast); */ 647 set_bit(0xfd, bitmask); 648 649 netdev_for_each_mc_addr(ha, netdev) { 650 hash = spider_net_get_multicast_hash(netdev, ha->addr); 651 set_bit(hash, bitmask); 652 } 653 654write_hash: 655 for (i = 0; i < SPIDER_NET_MULTICAST_HASHES / 4; i++) { 656 reg = 0; 657 if (test_bit(i * 4, bitmask)) 658 reg += 0x08; 659 reg <<= 8; 660 if (test_bit(i * 4 + 1, bitmask)) 661 reg += 0x08; 662 reg <<= 8; 663 if (test_bit(i * 4 + 2, bitmask)) 664 reg += 0x08; 665 reg <<= 8; 666 if (test_bit(i * 4 + 3, bitmask)) 667 reg += 0x08; 668 669 spider_net_write_reg(card, SPIDER_NET_GMRMHFILnR + i * 4, reg); 670 } 671} 672 673/** 674 * spider_net_prepare_tx_descr - fill tx descriptor with skb data 675 * @card: card structure 676 * @skb: packet to use 677 * 678 * returns 0 on success, <0 on failure. 679 * 680 * fills out the descriptor structure with skb data and len. Copies data, 681 * if needed (32bit DMA!) 682 */ 683static int 684spider_net_prepare_tx_descr(struct spider_net_card *card, 685 struct sk_buff *skb) 686{ 687 struct spider_net_descr_chain *chain = &card->tx_chain; 688 struct spider_net_descr *descr; 689 struct spider_net_hw_descr *hwdescr; 690 dma_addr_t buf; 691 unsigned long flags; 692 693 buf = pci_map_single(card->pdev, skb->data, skb->len, PCI_DMA_TODEVICE); 694 if (pci_dma_mapping_error(card->pdev, buf)) { 695 if (netif_msg_tx_err(card) && net_ratelimit()) 696 dev_err(&card->netdev->dev, "could not iommu-map packet (%p, %i). " 697 "Dropping packet\n", skb->data, skb->len); 698 card->spider_stats.tx_iommu_map_error++; 699 return -ENOMEM; 700 } 701 702 spin_lock_irqsave(&chain->lock, flags); 703 descr = card->tx_chain.head; 704 if (descr->next == chain->tail->prev) { 705 spin_unlock_irqrestore(&chain->lock, flags); 706 pci_unmap_single(card->pdev, buf, skb->len, PCI_DMA_TODEVICE); 707 return -ENOMEM; 708 } 709 hwdescr = descr->hwdescr; 710 chain->head = descr->next; 711 712 descr->skb = skb; 713 hwdescr->buf_addr = buf; 714 hwdescr->buf_size = skb->len; 715 hwdescr->next_descr_addr = 0; 716 hwdescr->data_status = 0; 717 718 hwdescr->dmac_cmd_status = 719 SPIDER_NET_DESCR_CARDOWNED | SPIDER_NET_DMAC_TXFRMTL; 720 spin_unlock_irqrestore(&chain->lock, flags); 721 722 if (skb->ip_summed == CHECKSUM_PARTIAL) 723 switch (ip_hdr(skb)->protocol) { 724 case IPPROTO_TCP: 725 hwdescr->dmac_cmd_status |= SPIDER_NET_DMAC_TCP; 726 break; 727 case IPPROTO_UDP: 728 hwdescr->dmac_cmd_status |= SPIDER_NET_DMAC_UDP; 729 break; 730 } 731 732 /* Chain the bus address, so that the DMA engine finds this descr. */ 733 wmb(); 734 descr->prev->hwdescr->next_descr_addr = descr->bus_addr; 735 736 card->netdev->trans_start = jiffies; /* set netdev watchdog timer */ 737 return 0; 738} 739 740static int 741spider_net_set_low_watermark(struct spider_net_card *card) 742{ 743 struct spider_net_descr *descr = card->tx_chain.tail; 744 struct spider_net_hw_descr *hwdescr; 745 unsigned long flags; 746 int status; 747 int cnt=0; 748 int i; 749 750 /* Measure the length of the queue. Measurement does not 751 * need to be precise -- does not need a lock. */ 752 while (descr != card->tx_chain.head) { 753 status = descr->hwdescr->dmac_cmd_status & SPIDER_NET_DESCR_NOT_IN_USE; 754 if (status == SPIDER_NET_DESCR_NOT_IN_USE) 755 break; 756 descr = descr->next; 757 cnt++; 758 } 759 760 /* If TX queue is short, don't even bother with interrupts */ 761 if (cnt < card->tx_chain.num_desc/4) 762 return cnt; 763 764 /* Set low-watermark 3/4th's of the way into the queue. */ 765 descr = card->tx_chain.tail; 766 cnt = (cnt*3)/4; 767 for (i=0;i<cnt; i++) 768 descr = descr->next; 769 770 /* Set the new watermark, clear the old watermark */ 771 spin_lock_irqsave(&card->tx_chain.lock, flags); 772 descr->hwdescr->dmac_cmd_status |= SPIDER_NET_DESCR_TXDESFLG; 773 if (card->low_watermark && card->low_watermark != descr) { 774 hwdescr = card->low_watermark->hwdescr; 775 hwdescr->dmac_cmd_status = 776 hwdescr->dmac_cmd_status & ~SPIDER_NET_DESCR_TXDESFLG; 777 } 778 card->low_watermark = descr; 779 spin_unlock_irqrestore(&card->tx_chain.lock, flags); 780 return cnt; 781} 782 783/** 784 * spider_net_release_tx_chain - processes sent tx descriptors 785 * @card: adapter structure 786 * @brutal: if set, don't care about whether descriptor seems to be in use 787 * 788 * returns 0 if the tx ring is empty, otherwise 1. 789 * 790 * spider_net_release_tx_chain releases the tx descriptors that spider has 791 * finished with (if non-brutal) or simply release tx descriptors (if brutal). 792 * If some other context is calling this function, we return 1 so that we're 793 * scheduled again (if we were scheduled) and will not lose initiative. 794 */ 795static int 796spider_net_release_tx_chain(struct spider_net_card *card, int brutal) 797{ 798 struct net_device *dev = card->netdev; 799 struct spider_net_descr_chain *chain = &card->tx_chain; 800 struct spider_net_descr *descr; 801 struct spider_net_hw_descr *hwdescr; 802 struct sk_buff *skb; 803 u32 buf_addr; 804 unsigned long flags; 805 int status; 806 807 while (1) { 808 spin_lock_irqsave(&chain->lock, flags); 809 if (chain->tail == chain->head) { 810 spin_unlock_irqrestore(&chain->lock, flags); 811 return 0; 812 } 813 descr = chain->tail; 814 hwdescr = descr->hwdescr; 815 816 status = spider_net_get_descr_status(hwdescr); 817 switch (status) { 818 case SPIDER_NET_DESCR_COMPLETE: 819 dev->stats.tx_packets++; 820 dev->stats.tx_bytes += descr->skb->len; 821 break; 822 823 case SPIDER_NET_DESCR_CARDOWNED: 824 if (!brutal) { 825 spin_unlock_irqrestore(&chain->lock, flags); 826 return 1; 827 } 828 829 /* fallthrough, if we release the descriptors 830 * brutally (then we don't care about 831 * SPIDER_NET_DESCR_CARDOWNED) */ 832 833 case SPIDER_NET_DESCR_RESPONSE_ERROR: 834 case SPIDER_NET_DESCR_PROTECTION_ERROR: 835 case SPIDER_NET_DESCR_FORCE_END: 836 if (netif_msg_tx_err(card)) 837 dev_err(&card->netdev->dev, "forcing end of tx descriptor " 838 "with status x%02x\n", status); 839 dev->stats.tx_errors++; 840 break; 841 842 default: 843 dev->stats.tx_dropped++; 844 if (!brutal) { 845 spin_unlock_irqrestore(&chain->lock, flags); 846 return 1; 847 } 848 } 849 850 chain->tail = descr->next; 851 hwdescr->dmac_cmd_status |= SPIDER_NET_DESCR_NOT_IN_USE; 852 skb = descr->skb; 853 descr->skb = NULL; 854 buf_addr = hwdescr->buf_addr; 855 spin_unlock_irqrestore(&chain->lock, flags); 856 857 /* unmap the skb */ 858 if (skb) { 859 pci_unmap_single(card->pdev, buf_addr, skb->len, 860 PCI_DMA_TODEVICE); 861 dev_kfree_skb(skb); 862 } 863 } 864 return 0; 865} 866 867/** 868 * spider_net_kick_tx_dma - enables TX DMA processing 869 * @card: card structure 870 * 871 * This routine will start the transmit DMA running if 872 * it is not already running. This routine ned only be 873 * called when queueing a new packet to an empty tx queue. 874 * Writes the current tx chain head as start address 875 * of the tx descriptor chain and enables the transmission 876 * DMA engine. 877 */ 878static inline void 879spider_net_kick_tx_dma(struct spider_net_card *card) 880{ 881 struct spider_net_descr *descr; 882 883 if (spider_net_read_reg(card, SPIDER_NET_GDTDMACCNTR) & 884 SPIDER_NET_TX_DMA_EN) 885 goto out; 886 887 descr = card->tx_chain.tail; 888 for (;;) { 889 if (spider_net_get_descr_status(descr->hwdescr) == 890 SPIDER_NET_DESCR_CARDOWNED) { 891 spider_net_write_reg(card, SPIDER_NET_GDTDCHA, 892 descr->bus_addr); 893 spider_net_write_reg(card, SPIDER_NET_GDTDMACCNTR, 894 SPIDER_NET_DMA_TX_VALUE); 895 break; 896 } 897 if (descr == card->tx_chain.head) 898 break; 899 descr = descr->next; 900 } 901 902out: 903 mod_timer(&card->tx_timer, jiffies + SPIDER_NET_TX_TIMER); 904} 905 906/** 907 * spider_net_xmit - transmits a frame over the device 908 * @skb: packet to send out 909 * @netdev: interface device structure 910 * 911 * returns 0 on success, !0 on failure 912 */ 913static int 914spider_net_xmit(struct sk_buff *skb, struct net_device *netdev) 915{ 916 int cnt; 917 struct spider_net_card *card = netdev_priv(netdev); 918 919 spider_net_release_tx_chain(card, 0); 920 921 if (spider_net_prepare_tx_descr(card, skb) != 0) { 922 netdev->stats.tx_dropped++; 923 netif_stop_queue(netdev); 924 return NETDEV_TX_BUSY; 925 } 926 927 cnt = spider_net_set_low_watermark(card); 928 if (cnt < 5) 929 spider_net_kick_tx_dma(card); 930 return NETDEV_TX_OK; 931} 932 933/** 934 * spider_net_cleanup_tx_ring - cleans up the TX ring 935 * @card: card structure 936 * 937 * spider_net_cleanup_tx_ring is called by either the tx_timer 938 * or from the NAPI polling routine. 939 * This routine releases resources associted with transmitted 940 * packets, including updating the queue tail pointer. 941 */ 942static void 943spider_net_cleanup_tx_ring(struct spider_net_card *card) 944{ 945 if ((spider_net_release_tx_chain(card, 0) != 0) && 946 (card->netdev->flags & IFF_UP)) { 947 spider_net_kick_tx_dma(card); 948 netif_wake_queue(card->netdev); 949 } 950} 951 952/** 953 * spider_net_do_ioctl - called for device ioctls 954 * @netdev: interface device structure 955 * @ifr: request parameter structure for ioctl 956 * @cmd: command code for ioctl 957 * 958 * returns 0 on success, <0 on failure. Currently, we have no special ioctls. 959 * -EOPNOTSUPP is returned, if an unknown ioctl was requested 960 */ 961static int 962spider_net_do_ioctl(struct net_device *netdev, struct ifreq *ifr, int cmd) 963{ 964 switch (cmd) { 965 default: 966 return -EOPNOTSUPP; 967 } 968} 969 970/** 971 * spider_net_pass_skb_up - takes an skb from a descriptor and passes it on 972 * @descr: descriptor to process 973 * @card: card structure 974 * 975 * Fills out skb structure and passes the data to the stack. 976 * The descriptor state is not changed. 977 */ 978static void 979spider_net_pass_skb_up(struct spider_net_descr *descr, 980 struct spider_net_card *card) 981{ 982 struct spider_net_hw_descr *hwdescr = descr->hwdescr; 983 struct sk_buff *skb = descr->skb; 984 struct net_device *netdev = card->netdev; 985 u32 data_status = hwdescr->data_status; 986 u32 data_error = hwdescr->data_error; 987 988 skb_put(skb, hwdescr->valid_size); 989 990 /* the card seems to add 2 bytes of junk in front 991 * of the ethernet frame */ 992#define SPIDER_MISALIGN 2 993 skb_pull(skb, SPIDER_MISALIGN); 994 skb->protocol = eth_type_trans(skb, netdev); 995 996 /* checksum offload */ 997 if (card->options.rx_csum) { 998 if ( ( (data_status & SPIDER_NET_DATA_STATUS_CKSUM_MASK) == 999 SPIDER_NET_DATA_STATUS_CKSUM_MASK) && 1000 !(data_error & SPIDER_NET_DATA_ERR_CKSUM_MASK)) 1001 skb->ip_summed = CHECKSUM_UNNECESSARY; 1002 else 1003 skb->ip_summed = CHECKSUM_NONE; 1004 } else 1005 skb->ip_summed = CHECKSUM_NONE; 1006 1007 if (data_status & SPIDER_NET_VLAN_PACKET) { 1008 /* further enhancements: HW-accel VLAN 1009 * vlan_hwaccel_receive_skb 1010 */ 1011 } 1012 1013 /* update netdevice statistics */ 1014 netdev->stats.rx_packets++; 1015 netdev->stats.rx_bytes += skb->len; 1016 1017 /* pass skb up to stack */ 1018 netif_receive_skb(skb); 1019} 1020 1021static void show_rx_chain(struct spider_net_card *card) 1022{ 1023 struct spider_net_descr_chain *chain = &card->rx_chain; 1024 struct spider_net_descr *start= chain->tail; 1025 struct spider_net_descr *descr= start; 1026 struct spider_net_hw_descr *hwd = start->hwdescr; 1027 struct device *dev = &card->netdev->dev; 1028 u32 curr_desc, next_desc; 1029 int status; 1030 1031 int tot = 0; 1032 int cnt = 0; 1033 int off = start - chain->ring; 1034 int cstat = hwd->dmac_cmd_status; 1035 1036 dev_info(dev, "Total number of descrs=%d\n", 1037 chain->num_desc); 1038 dev_info(dev, "Chain tail located at descr=%d, status=0x%x\n", 1039 off, cstat); 1040 1041 curr_desc = spider_net_read_reg(card, SPIDER_NET_GDACTDPA); 1042 next_desc = spider_net_read_reg(card, SPIDER_NET_GDACNEXTDA); 1043 1044 status = cstat; 1045 do 1046 { 1047 hwd = descr->hwdescr; 1048 off = descr - chain->ring; 1049 status = hwd->dmac_cmd_status; 1050 1051 if (descr == chain->head) 1052 dev_info(dev, "Chain head is at %d, head status=0x%x\n", 1053 off, status); 1054 1055 if (curr_desc == descr->bus_addr) 1056 dev_info(dev, "HW curr desc (GDACTDPA) is at %d, status=0x%x\n", 1057 off, status); 1058 1059 if (next_desc == descr->bus_addr) 1060 dev_info(dev, "HW next desc (GDACNEXTDA) is at %d, status=0x%x\n", 1061 off, status); 1062 1063 if (hwd->next_descr_addr == 0) 1064 dev_info(dev, "chain is cut at %d\n", off); 1065 1066 if (cstat != status) { 1067 int from = (chain->num_desc + off - cnt) % chain->num_desc; 1068 int to = (chain->num_desc + off - 1) % chain->num_desc; 1069 dev_info(dev, "Have %d (from %d to %d) descrs " 1070 "with stat=0x%08x\n", cnt, from, to, cstat); 1071 cstat = status; 1072 cnt = 0; 1073 } 1074 1075 cnt ++; 1076 tot ++; 1077 descr = descr->next; 1078 } while (descr != start); 1079 1080 dev_info(dev, "Last %d descrs with stat=0x%08x " 1081 "for a total of %d descrs\n", cnt, cstat, tot); 1082 1083#ifdef DEBUG 1084 /* Now dump the whole ring */ 1085 descr = start; 1086 do 1087 { 1088 struct spider_net_hw_descr *hwd = descr->hwdescr; 1089 status = spider_net_get_descr_status(hwd); 1090 cnt = descr - chain->ring; 1091 dev_info(dev, "Descr %d stat=0x%08x skb=%p\n", 1092 cnt, status, descr->skb); 1093 dev_info(dev, "bus addr=%08x buf addr=%08x sz=%d\n", 1094 descr->bus_addr, hwd->buf_addr, hwd->buf_size); 1095 dev_info(dev, "next=%08x result sz=%d valid sz=%d\n", 1096 hwd->next_descr_addr, hwd->result_size, 1097 hwd->valid_size); 1098 dev_info(dev, "dmac=%08x data stat=%08x data err=%08x\n", 1099 hwd->dmac_cmd_status, hwd->data_status, 1100 hwd->data_error); 1101 dev_info(dev, "\n"); 1102 1103 descr = descr->next; 1104 } while (descr != start); 1105#endif 1106 1107} 1108 1109/** 1110 * spider_net_resync_head_ptr - Advance head ptr past empty descrs 1111 * 1112 * If the driver fails to keep up and empty the queue, then the 1113 * hardware wil run out of room to put incoming packets. This 1114 * will cause the hardware to skip descrs that are full (instead 1115 * of halting/retrying). Thus, once the driver runs, it wil need 1116 * to "catch up" to where the hardware chain pointer is at. 1117 */ 1118static void spider_net_resync_head_ptr(struct spider_net_card *card) 1119{ 1120 unsigned long flags; 1121 struct spider_net_descr_chain *chain = &card->rx_chain; 1122 struct spider_net_descr *descr; 1123 int i, status; 1124 1125 /* Advance head pointer past any empty descrs */ 1126 descr = chain->head; 1127 status = spider_net_get_descr_status(descr->hwdescr); 1128 1129 if (status == SPIDER_NET_DESCR_NOT_IN_USE) 1130 return; 1131 1132 spin_lock_irqsave(&chain->lock, flags); 1133 1134 descr = chain->head; 1135 status = spider_net_get_descr_status(descr->hwdescr); 1136 for (i=0; i<chain->num_desc; i++) { 1137 if (status != SPIDER_NET_DESCR_CARDOWNED) break; 1138 descr = descr->next; 1139 status = spider_net_get_descr_status(descr->hwdescr); 1140 } 1141 chain->head = descr; 1142 1143 spin_unlock_irqrestore(&chain->lock, flags); 1144} 1145 1146static int spider_net_resync_tail_ptr(struct spider_net_card *card) 1147{ 1148 struct spider_net_descr_chain *chain = &card->rx_chain; 1149 struct spider_net_descr *descr; 1150 int i, status; 1151 1152 /* Advance tail pointer past any empty and reaped descrs */ 1153 descr = chain->tail; 1154 status = spider_net_get_descr_status(descr->hwdescr); 1155 1156 for (i=0; i<chain->num_desc; i++) { 1157 if ((status != SPIDER_NET_DESCR_CARDOWNED) && 1158 (status != SPIDER_NET_DESCR_NOT_IN_USE)) break; 1159 descr = descr->next; 1160 status = spider_net_get_descr_status(descr->hwdescr); 1161 } 1162 chain->tail = descr; 1163 1164 if ((i == chain->num_desc) || (i == 0)) 1165 return 1; 1166 return 0; 1167} 1168 1169/** 1170 * spider_net_decode_one_descr - processes an RX descriptor 1171 * @card: card structure 1172 * 1173 * Returns 1 if a packet has been sent to the stack, otherwise 0. 1174 * 1175 * Processes an RX descriptor by iommu-unmapping the data buffer 1176 * and passing the packet up to the stack. This function is called 1177 * in softirq context, e.g. either bottom half from interrupt or 1178 * NAPI polling context. 1179 */ 1180static int 1181spider_net_decode_one_descr(struct spider_net_card *card) 1182{ 1183 struct net_device *dev = card->netdev; 1184 struct spider_net_descr_chain *chain = &card->rx_chain; 1185 struct spider_net_descr *descr = chain->tail; 1186 struct spider_net_hw_descr *hwdescr = descr->hwdescr; 1187 u32 hw_buf_addr; 1188 int status; 1189 1190 status = spider_net_get_descr_status(hwdescr); 1191 1192 /* Nothing in the descriptor, or ring must be empty */ 1193 if ((status == SPIDER_NET_DESCR_CARDOWNED) || 1194 (status == SPIDER_NET_DESCR_NOT_IN_USE)) 1195 return 0; 1196 1197 /* descriptor definitively used -- move on tail */ 1198 chain->tail = descr->next; 1199 1200 /* unmap descriptor */ 1201 hw_buf_addr = hwdescr->buf_addr; 1202 hwdescr->buf_addr = 0xffffffff; 1203 pci_unmap_single(card->pdev, hw_buf_addr, 1204 SPIDER_NET_MAX_FRAME, PCI_DMA_FROMDEVICE); 1205 1206 if ( (status == SPIDER_NET_DESCR_RESPONSE_ERROR) || 1207 (status == SPIDER_NET_DESCR_PROTECTION_ERROR) || 1208 (status == SPIDER_NET_DESCR_FORCE_END) ) { 1209 if (netif_msg_rx_err(card)) 1210 dev_err(&dev->dev, 1211 "dropping RX descriptor with state %d\n", status); 1212 dev->stats.rx_dropped++; 1213 goto bad_desc; 1214 } 1215 1216 if ( (status != SPIDER_NET_DESCR_COMPLETE) && 1217 (status != SPIDER_NET_DESCR_FRAME_END) ) { 1218 if (netif_msg_rx_err(card)) 1219 dev_err(&card->netdev->dev, 1220 "RX descriptor with unknown state %d\n", status); 1221 card->spider_stats.rx_desc_unk_state++; 1222 goto bad_desc; 1223 } 1224 1225 /* The cases we'll throw away the packet immediately */ 1226 if (hwdescr->data_error & SPIDER_NET_DESTROY_RX_FLAGS) { 1227 if (netif_msg_rx_err(card)) 1228 dev_err(&card->netdev->dev, 1229 "error in received descriptor found, " 1230 "data_status=x%08x, data_error=x%08x\n", 1231 hwdescr->data_status, hwdescr->data_error); 1232 goto bad_desc; 1233 } 1234 1235 if (hwdescr->dmac_cmd_status & SPIDER_NET_DESCR_BAD_STATUS) { 1236 dev_err(&card->netdev->dev, "bad status, cmd_status=x%08x\n", 1237 hwdescr->dmac_cmd_status); 1238 pr_err("buf_addr=x%08x\n", hw_buf_addr); 1239 pr_err("buf_size=x%08x\n", hwdescr->buf_size); 1240 pr_err("next_descr_addr=x%08x\n", hwdescr->next_descr_addr); 1241 pr_err("result_size=x%08x\n", hwdescr->result_size); 1242 pr_err("valid_size=x%08x\n", hwdescr->valid_size); 1243 pr_err("data_status=x%08x\n", hwdescr->data_status); 1244 pr_err("data_error=x%08x\n", hwdescr->data_error); 1245 pr_err("which=%ld\n", descr - card->rx_chain.ring); 1246 1247 card->spider_stats.rx_desc_error++; 1248 goto bad_desc; 1249 } 1250 1251 /* Ok, we've got a packet in descr */ 1252 spider_net_pass_skb_up(descr, card); 1253 descr->skb = NULL; 1254 hwdescr->dmac_cmd_status = SPIDER_NET_DESCR_NOT_IN_USE; 1255 return 1; 1256 1257bad_desc: 1258 if (netif_msg_rx_err(card)) 1259 show_rx_chain(card); 1260 dev_kfree_skb_irq(descr->skb); 1261 descr->skb = NULL; 1262 hwdescr->dmac_cmd_status = SPIDER_NET_DESCR_NOT_IN_USE; 1263 return 0; 1264} 1265 1266/** 1267 * spider_net_poll - NAPI poll function called by the stack to return packets 1268 * @netdev: interface device structure 1269 * @budget: number of packets we can pass to the stack at most 1270 * 1271 * returns 0 if no more packets available to the driver/stack. Returns 1, 1272 * if the quota is exceeded, but the driver has still packets. 1273 * 1274 * spider_net_poll returns all packets from the rx descriptors to the stack 1275 * (using netif_receive_skb). If all/enough packets are up, the driver 1276 * reenables interrupts and returns 0. If not, 1 is returned. 1277 */ 1278static int spider_net_poll(struct napi_struct *napi, int budget) 1279{ 1280 struct spider_net_card *card = container_of(napi, struct spider_net_card, napi); 1281 int packets_done = 0; 1282 1283 while (packets_done < budget) { 1284 if (!spider_net_decode_one_descr(card)) 1285 break; 1286 1287 packets_done++; 1288 } 1289 1290 if ((packets_done == 0) && (card->num_rx_ints != 0)) { 1291 if (!spider_net_resync_tail_ptr(card)) 1292 packets_done = budget; 1293 spider_net_resync_head_ptr(card); 1294 } 1295 card->num_rx_ints = 0; 1296 1297 spider_net_refill_rx_chain(card); 1298 spider_net_enable_rxdmac(card); 1299 1300 spider_net_cleanup_tx_ring(card); 1301 1302 /* if all packets are in the stack, enable interrupts and return 0 */ 1303 /* if not, return 1 */ 1304 if (packets_done < budget) { 1305 napi_complete(napi); 1306 spider_net_rx_irq_on(card); 1307 card->ignore_rx_ramfull = 0; 1308 } 1309 1310 return packets_done; 1311} 1312 1313/** 1314 * spider_net_change_mtu - changes the MTU of an interface 1315 * @netdev: interface device structure 1316 * @new_mtu: new MTU value 1317 * 1318 * returns 0 on success, <0 on failure 1319 */ 1320static int 1321spider_net_change_mtu(struct net_device *netdev, int new_mtu) 1322{ 1323 /* no need to re-alloc skbs or so -- the max mtu is about 2.3k 1324 * and mtu is outbound only anyway */ 1325 if ( (new_mtu < SPIDER_NET_MIN_MTU ) || 1326 (new_mtu > SPIDER_NET_MAX_MTU) ) 1327 return -EINVAL; 1328 netdev->mtu = new_mtu; 1329 return 0; 1330} 1331 1332/** 1333 * spider_net_set_mac - sets the MAC of an interface 1334 * @netdev: interface device structure 1335 * @ptr: pointer to new MAC address 1336 * 1337 * Returns 0 on success, <0 on failure. Currently, we don't support this 1338 * and will always return EOPNOTSUPP. 1339 */ 1340static int 1341spider_net_set_mac(struct net_device *netdev, void *p) 1342{ 1343 struct spider_net_card *card = netdev_priv(netdev); 1344 u32 macl, macu, regvalue; 1345 struct sockaddr *addr = p; 1346 1347 if (!is_valid_ether_addr(addr->sa_data)) 1348 return -EADDRNOTAVAIL; 1349 1350 /* switch off GMACTPE and GMACRPE */ 1351 regvalue = spider_net_read_reg(card, SPIDER_NET_GMACOPEMD); 1352 regvalue &= ~((1 << 5) | (1 << 6)); 1353 spider_net_write_reg(card, SPIDER_NET_GMACOPEMD, regvalue); 1354 1355 /* write mac */ 1356 macu = (addr->sa_data[0]<<24) + (addr->sa_data[1]<<16) + 1357 (addr->sa_data[2]<<8) + (addr->sa_data[3]); 1358 macl = (addr->sa_data[4]<<8) + (addr->sa_data[5]); 1359 spider_net_write_reg(card, SPIDER_NET_GMACUNIMACU, macu); 1360 spider_net_write_reg(card, SPIDER_NET_GMACUNIMACL, macl); 1361 1362 /* switch GMACTPE and GMACRPE back on */ 1363 regvalue = spider_net_read_reg(card, SPIDER_NET_GMACOPEMD); 1364 regvalue |= ((1 << 5) | (1 << 6)); 1365 spider_net_write_reg(card, SPIDER_NET_GMACOPEMD, regvalue); 1366 1367 spider_net_set_promisc(card); 1368 1369 /* look up, whether we have been successful */ 1370 if (spider_net_get_mac_address(netdev)) 1371 return -EADDRNOTAVAIL; 1372 if (memcmp(netdev->dev_addr,addr->sa_data,netdev->addr_len)) 1373 return -EADDRNOTAVAIL; 1374 1375 return 0; 1376} 1377 1378/** 1379 * spider_net_link_reset 1380 * @netdev: net device structure 1381 * 1382 * This is called when the PHY_LINK signal is asserted. For the blade this is 1383 * not connected so we should never get here. 1384 * 1385 */ 1386static void 1387spider_net_link_reset(struct net_device *netdev) 1388{ 1389 1390 struct spider_net_card *card = netdev_priv(netdev); 1391 1392 del_timer_sync(&card->aneg_timer); 1393 1394 /* clear interrupt, block further interrupts */ 1395 spider_net_write_reg(card, SPIDER_NET_GMACST, 1396 spider_net_read_reg(card, SPIDER_NET_GMACST)); 1397 spider_net_write_reg(card, SPIDER_NET_GMACINTEN, 0); 1398 1399 /* reset phy and setup aneg */ 1400 card->aneg_count = 0; 1401 card->medium = BCM54XX_COPPER; 1402 spider_net_setup_aneg(card); 1403 mod_timer(&card->aneg_timer, jiffies + SPIDER_NET_ANEG_TIMER); 1404 1405} 1406 1407/** 1408 * spider_net_handle_error_irq - handles errors raised by an interrupt 1409 * @card: card structure 1410 * @status_reg: interrupt status register 0 (GHIINT0STS) 1411 * 1412 * spider_net_handle_error_irq treats or ignores all error conditions 1413 * found when an interrupt is presented 1414 */ 1415static void 1416spider_net_handle_error_irq(struct spider_net_card *card, u32 status_reg, 1417 u32 error_reg1, u32 error_reg2) 1418{ 1419 u32 i; 1420 int show_error = 1; 1421 1422 /* check GHIINT0STS ************************************/ 1423 if (status_reg) 1424 for (i = 0; i < 32; i++) 1425 if (status_reg & (1<<i)) 1426 switch (i) 1427 { 1428 /* let error_reg1 and error_reg2 evaluation decide, what to do 1429 case SPIDER_NET_PHYINT: 1430 case SPIDER_NET_GMAC2INT: 1431 case SPIDER_NET_GMAC1INT: 1432 case SPIDER_NET_GFIFOINT: 1433 case SPIDER_NET_DMACINT: 1434 case SPIDER_NET_GSYSINT: 1435 break; */ 1436 1437 case SPIDER_NET_GIPSINT: 1438 show_error = 0; 1439 break; 1440 1441 case SPIDER_NET_GPWOPCMPINT: 1442 /* PHY write operation completed */ 1443 show_error = 0; 1444 break; 1445 case SPIDER_NET_GPROPCMPINT: 1446 /* PHY read operation completed */ 1447 /* we don't use semaphores, as we poll for the completion 1448 * of the read operation in spider_net_read_phy. Should take 1449 * about 50 us */ 1450 show_error = 0; 1451 break; 1452 case SPIDER_NET_GPWFFINT: 1453 /* PHY command queue full */ 1454 if (netif_msg_intr(card)) 1455 dev_err(&card->netdev->dev, "PHY write queue full\n"); 1456 show_error = 0; 1457 break; 1458 1459 /* case SPIDER_NET_GRMDADRINT: not used. print a message */ 1460 /* case SPIDER_NET_GRMARPINT: not used. print a message */ 1461 /* case SPIDER_NET_GRMMPINT: not used. print a message */ 1462 1463 case SPIDER_NET_GDTDEN0INT: 1464 /* someone has set TX_DMA_EN to 0 */ 1465 show_error = 0; 1466 break; 1467 1468 case SPIDER_NET_GDDDEN0INT: /* fallthrough */ 1469 case SPIDER_NET_GDCDEN0INT: /* fallthrough */ 1470 case SPIDER_NET_GDBDEN0INT: /* fallthrough */ 1471 case SPIDER_NET_GDADEN0INT: 1472 /* someone has set RX_DMA_EN to 0 */ 1473 show_error = 0; 1474 break; 1475 1476 /* RX interrupts */ 1477 case SPIDER_NET_GDDFDCINT: 1478 case SPIDER_NET_GDCFDCINT: 1479 case SPIDER_NET_GDBFDCINT: 1480 case SPIDER_NET_GDAFDCINT: 1481 /* case SPIDER_NET_GDNMINT: not used. print a message */ 1482 /* case SPIDER_NET_GCNMINT: not used. print a message */ 1483 /* case SPIDER_NET_GBNMINT: not used. print a message */ 1484 /* case SPIDER_NET_GANMINT: not used. print a message */ 1485 /* case SPIDER_NET_GRFNMINT: not used. print a message */ 1486 show_error = 0; 1487 break; 1488 1489 /* TX interrupts */ 1490 case SPIDER_NET_GDTFDCINT: 1491 show_error = 0; 1492 break; 1493 case SPIDER_NET_GTTEDINT: 1494 show_error = 0; 1495 break; 1496 case SPIDER_NET_GDTDCEINT: 1497 /* chain end. If a descriptor should be sent, kick off 1498 * tx dma 1499 if (card->tx_chain.tail != card->tx_chain.head) 1500 spider_net_kick_tx_dma(card); 1501 */ 1502 show_error = 0; 1503 break; 1504 1505 /* case SPIDER_NET_G1TMCNTINT: not used. print a message */ 1506 /* case SPIDER_NET_GFREECNTINT: not used. print a message */ 1507 } 1508 1509 /* check GHIINT1STS ************************************/ 1510 if (error_reg1) 1511 for (i = 0; i < 32; i++) 1512 if (error_reg1 & (1<<i)) 1513 switch (i) 1514 { 1515 case SPIDER_NET_GTMFLLINT: 1516 /* TX RAM full may happen on a usual case. 1517 * Logging is not needed. */ 1518 show_error = 0; 1519 break; 1520 case SPIDER_NET_GRFDFLLINT: /* fallthrough */ 1521 case SPIDER_NET_GRFCFLLINT: /* fallthrough */ 1522 case SPIDER_NET_GRFBFLLINT: /* fallthrough */ 1523 case SPIDER_NET_GRFAFLLINT: /* fallthrough */ 1524 case SPIDER_NET_GRMFLLINT: 1525 /* Could happen when rx chain is full */ 1526 if (card->ignore_rx_ramfull == 0) { 1527 card->ignore_rx_ramfull = 1; 1528 spider_net_resync_head_ptr(card); 1529 spider_net_refill_rx_chain(card); 1530 spider_net_enable_rxdmac(card); 1531 card->num_rx_ints ++; 1532 napi_schedule(&card->napi); 1533 } 1534 show_error = 0; 1535 break; 1536 1537 /* case SPIDER_NET_GTMSHTINT: problem, print a message */ 1538 case SPIDER_NET_GDTINVDINT: 1539 /* allrighty. tx from previous descr ok */ 1540 show_error = 0; 1541 break; 1542 1543 /* chain end */ 1544 case SPIDER_NET_GDDDCEINT: /* fallthrough */ 1545 case SPIDER_NET_GDCDCEINT: /* fallthrough */ 1546 case SPIDER_NET_GDBDCEINT: /* fallthrough */ 1547 case SPIDER_NET_GDADCEINT: 1548 spider_net_resync_head_ptr(card); 1549 spider_net_refill_rx_chain(card); 1550 spider_net_enable_rxdmac(card); 1551 card->num_rx_ints ++; 1552 napi_schedule(&card->napi); 1553 show_error = 0; 1554 break; 1555 1556 /* invalid descriptor */ 1557 case SPIDER_NET_GDDINVDINT: /* fallthrough */ 1558 case SPIDER_NET_GDCINVDINT: /* fallthrough */ 1559 case SPIDER_NET_GDBINVDINT: /* fallthrough */ 1560 case SPIDER_NET_GDAINVDINT: 1561 /* Could happen when rx chain is full */ 1562 spider_net_resync_head_ptr(card); 1563 spider_net_refill_rx_chain(card); 1564 spider_net_enable_rxdmac(card); 1565 card->num_rx_ints ++; 1566 napi_schedule(&card->napi); 1567 show_error = 0; 1568 break; 1569 1570 /* case SPIDER_NET_GDTRSERINT: problem, print a message */ 1571 /* case SPIDER_NET_GDDRSERINT: problem, print a message */ 1572 /* case SPIDER_NET_GDCRSERINT: problem, print a message */ 1573 /* case SPIDER_NET_GDBRSERINT: problem, print a message */ 1574 /* case SPIDER_NET_GDARSERINT: problem, print a message */ 1575 /* case SPIDER_NET_GDSERINT: problem, print a message */ 1576 /* case SPIDER_NET_GDTPTERINT: problem, print a message */ 1577 /* case SPIDER_NET_GDDPTERINT: problem, print a message */ 1578 /* case SPIDER_NET_GDCPTERINT: problem, print a message */ 1579 /* case SPIDER_NET_GDBPTERINT: problem, print a message */ 1580 /* case SPIDER_NET_GDAPTERINT: problem, print a message */ 1581 default: 1582 show_error = 1; 1583 break; 1584 } 1585 1586 /* check GHIINT2STS ************************************/ 1587 if (error_reg2) 1588 for (i = 0; i < 32; i++) 1589 if (error_reg2 & (1<<i)) 1590 switch (i) 1591 { 1592 /* there is nothing we can (want to) do at this time. Log a 1593 * message, we can switch on and off the specific values later on 1594 case SPIDER_NET_GPROPERINT: 1595 case SPIDER_NET_GMCTCRSNGINT: 1596 case SPIDER_NET_GMCTLCOLINT: 1597 case SPIDER_NET_GMCTTMOTINT: 1598 case SPIDER_NET_GMCRCAERINT: 1599 case SPIDER_NET_GMCRCALERINT: 1600 case SPIDER_NET_GMCRALNERINT: 1601 case SPIDER_NET_GMCROVRINT: 1602 case SPIDER_NET_GMCRRNTINT: 1603 case SPIDER_NET_GMCRRXERINT: 1604 case SPIDER_NET_GTITCSERINT: 1605 case SPIDER_NET_GTIFMTERINT: 1606 case SPIDER_NET_GTIPKTRVKINT: 1607 case SPIDER_NET_GTISPINGINT: 1608 case SPIDER_NET_GTISADNGINT: 1609 case SPIDER_NET_GTISPDNGINT: 1610 case SPIDER_NET_GRIFMTERINT: 1611 case SPIDER_NET_GRIPKTRVKINT: 1612 case SPIDER_NET_GRISPINGINT: 1613 case SPIDER_NET_GRISADNGINT: 1614 case SPIDER_NET_GRISPDNGINT: 1615 break; 1616 */ 1617 default: 1618 break; 1619 } 1620 1621 if ((show_error) && (netif_msg_intr(card)) && net_ratelimit()) 1622 dev_err(&card->netdev->dev, "Error interrupt, GHIINT0STS = 0x%08x, " 1623 "GHIINT1STS = 0x%08x, GHIINT2STS = 0x%08x\n", 1624 status_reg, error_reg1, error_reg2); 1625 1626 /* clear interrupt sources */ 1627 spider_net_write_reg(card, SPIDER_NET_GHIINT1STS, error_reg1); 1628 spider_net_write_reg(card, SPIDER_NET_GHIINT2STS, error_reg2); 1629} 1630 1631/** 1632 * spider_net_interrupt - interrupt handler for spider_net 1633 * @irq: interrupt number 1634 * @ptr: pointer to net_device 1635 * 1636 * returns IRQ_HANDLED, if interrupt was for driver, or IRQ_NONE, if no 1637 * interrupt found raised by card. 1638 * 1639 * This is the interrupt handler, that turns off 1640 * interrupts for this device and makes the stack poll the driver 1641 */ 1642static irqreturn_t 1643spider_net_interrupt(int irq, void *ptr) 1644{ 1645 struct net_device *netdev = ptr; 1646 struct spider_net_card *card = netdev_priv(netdev); 1647 u32 status_reg, error_reg1, error_reg2; 1648 1649 status_reg = spider_net_read_reg(card, SPIDER_NET_GHIINT0STS); 1650 error_reg1 = spider_net_read_reg(card, SPIDER_NET_GHIINT1STS); 1651 error_reg2 = spider_net_read_reg(card, SPIDER_NET_GHIINT2STS); 1652 1653 if (!(status_reg & SPIDER_NET_INT0_MASK_VALUE) && 1654 !(error_reg1 & SPIDER_NET_INT1_MASK_VALUE) && 1655 !(error_reg2 & SPIDER_NET_INT2_MASK_VALUE)) 1656 return IRQ_NONE; 1657 1658 if (status_reg & SPIDER_NET_RXINT ) { 1659 spider_net_rx_irq_off(card); 1660 napi_schedule(&card->napi); 1661 card->num_rx_ints ++; 1662 } 1663 if (status_reg & SPIDER_NET_TXINT) 1664 napi_schedule(&card->napi); 1665 1666 if (status_reg & SPIDER_NET_LINKINT) 1667 spider_net_link_reset(netdev); 1668 1669 if (status_reg & SPIDER_NET_ERRINT ) 1670 spider_net_handle_error_irq(card, status_reg, 1671 error_reg1, error_reg2); 1672 1673 /* clear interrupt sources */ 1674 spider_net_write_reg(card, SPIDER_NET_GHIINT0STS, status_reg); 1675 1676 return IRQ_HANDLED; 1677} 1678 1679#ifdef CONFIG_NET_POLL_CONTROLLER 1680/** 1681 * spider_net_poll_controller - artificial interrupt for netconsole etc. 1682 * @netdev: interface device structure 1683 * 1684 * see Documentation/networking/netconsole.txt 1685 */ 1686static void 1687spider_net_poll_controller(struct net_device *netdev) 1688{ 1689 disable_irq(netdev->irq); 1690 spider_net_interrupt(netdev->irq, netdev); 1691 enable_irq(netdev->irq); 1692} 1693#endif /* CONFIG_NET_POLL_CONTROLLER */ 1694 1695/** 1696 * spider_net_enable_interrupts - enable interrupts 1697 * @card: card structure 1698 * 1699 * spider_net_enable_interrupt enables several interrupts 1700 */ 1701static void 1702spider_net_enable_interrupts(struct spider_net_card *card) 1703{ 1704 spider_net_write_reg(card, SPIDER_NET_GHIINT0MSK, 1705 SPIDER_NET_INT0_MASK_VALUE); 1706 spider_net_write_reg(card, SPIDER_NET_GHIINT1MSK, 1707 SPIDER_NET_INT1_MASK_VALUE); 1708 spider_net_write_reg(card, SPIDER_NET_GHIINT2MSK, 1709 SPIDER_NET_INT2_MASK_VALUE); 1710} 1711 1712/** 1713 * spider_net_disable_interrupts - disable interrupts 1714 * @card: card structure 1715 * 1716 * spider_net_disable_interrupts disables all the interrupts 1717 */ 1718static void 1719spider_net_disable_interrupts(struct spider_net_card *card) 1720{ 1721 spider_net_write_reg(card, SPIDER_NET_GHIINT0MSK, 0); 1722 spider_net_write_reg(card, SPIDER_NET_GHIINT1MSK, 0); 1723 spider_net_write_reg(card, SPIDER_NET_GHIINT2MSK, 0); 1724 spider_net_write_reg(card, SPIDER_NET_GMACINTEN, 0); 1725} 1726 1727/** 1728 * spider_net_init_card - initializes the card 1729 * @card: card structure 1730 * 1731 * spider_net_init_card initializes the card so that other registers can 1732 * be used 1733 */ 1734static void 1735spider_net_init_card(struct spider_net_card *card) 1736{ 1737 spider_net_write_reg(card, SPIDER_NET_CKRCTRL, 1738 SPIDER_NET_CKRCTRL_STOP_VALUE); 1739 1740 spider_net_write_reg(card, SPIDER_NET_CKRCTRL, 1741 SPIDER_NET_CKRCTRL_RUN_VALUE); 1742 1743 /* trigger ETOMOD signal */ 1744 spider_net_write_reg(card, SPIDER_NET_GMACOPEMD, 1745 spider_net_read_reg(card, SPIDER_NET_GMACOPEMD) | 0x4); 1746 1747 spider_net_disable_interrupts(card); 1748} 1749 1750/** 1751 * spider_net_enable_card - enables the card by setting all kinds of regs 1752 * @card: card structure 1753 * 1754 * spider_net_enable_card sets a lot of SMMIO registers to enable the device 1755 */ 1756static void 1757spider_net_enable_card(struct spider_net_card *card) 1758{ 1759 int i; 1760 /* the following array consists of (register),(value) pairs 1761 * that are set in this function. A register of 0 ends the list */ 1762 u32 regs[][2] = { 1763 { SPIDER_NET_GRESUMINTNUM, 0 }, 1764 { SPIDER_NET_GREINTNUM, 0 }, 1765 1766 /* set interrupt frame number registers */ 1767 /* clear the single DMA engine registers first */ 1768 { SPIDER_NET_GFAFRMNUM, SPIDER_NET_GFXFRAMES_VALUE }, 1769 { SPIDER_NET_GFBFRMNUM, SPIDER_NET_GFXFRAMES_VALUE }, 1770 { SPIDER_NET_GFCFRMNUM, SPIDER_NET_GFXFRAMES_VALUE }, 1771 { SPIDER_NET_GFDFRMNUM, SPIDER_NET_GFXFRAMES_VALUE }, 1772 /* then set, what we really need */ 1773 { SPIDER_NET_GFFRMNUM, SPIDER_NET_FRAMENUM_VALUE }, 1774 1775 /* timer counter registers and stuff */ 1776 { SPIDER_NET_GFREECNNUM, 0 }, 1777 { SPIDER_NET_GONETIMENUM, 0 }, 1778 { SPIDER_NET_GTOUTFRMNUM, 0 }, 1779 1780 /* RX mode setting */ 1781 { SPIDER_NET_GRXMDSET, SPIDER_NET_RXMODE_VALUE }, 1782 /* TX mode setting */ 1783 { SPIDER_NET_GTXMDSET, SPIDER_NET_TXMODE_VALUE }, 1784 /* IPSEC mode setting */ 1785 { SPIDER_NET_GIPSECINIT, SPIDER_NET_IPSECINIT_VALUE }, 1786 1787 { SPIDER_NET_GFTRESTRT, SPIDER_NET_RESTART_VALUE }, 1788 1789 { SPIDER_NET_GMRWOLCTRL, 0 }, 1790 { SPIDER_NET_GTESTMD, 0x10000000 }, 1791 { SPIDER_NET_GTTQMSK, 0x00400040 }, 1792 1793 { SPIDER_NET_GMACINTEN, 0 }, 1794 1795 /* flow control stuff */ 1796 { SPIDER_NET_GMACAPAUSE, SPIDER_NET_MACAPAUSE_VALUE }, 1797 { SPIDER_NET_GMACTXPAUSE, SPIDER_NET_TXPAUSE_VALUE }, 1798 1799 { SPIDER_NET_GMACBSTLMT, SPIDER_NET_BURSTLMT_VALUE }, 1800 { 0, 0} 1801 }; 1802 1803 i = 0; 1804 while (regs[i][0]) { 1805 spider_net_write_reg(card, regs[i][0], regs[i][1]); 1806 i++; 1807 } 1808 1809 /* clear unicast filter table entries 1 to 14 */ 1810 for (i = 1; i <= 14; i++) { 1811 spider_net_write_reg(card, 1812 SPIDER_NET_GMRUAFILnR + i * 8, 1813 0x00080000); 1814 spider_net_write_reg(card, 1815 SPIDER_NET_GMRUAFILnR + i * 8 + 4, 1816 0x00000000); 1817 } 1818 1819 spider_net_write_reg(card, SPIDER_NET_GMRUA0FIL15R, 0x08080000); 1820 1821 spider_net_write_reg(card, SPIDER_NET_ECMODE, SPIDER_NET_ECMODE_VALUE); 1822 1823 /* set chain tail address for RX chains and 1824 * enable DMA */ 1825 spider_net_enable_rxchtails(card); 1826 spider_net_enable_rxdmac(card); 1827 1828 spider_net_write_reg(card, SPIDER_NET_GRXDMAEN, SPIDER_NET_WOL_VALUE); 1829 1830 spider_net_write_reg(card, SPIDER_NET_GMACLENLMT, 1831 SPIDER_NET_LENLMT_VALUE); 1832 spider_net_write_reg(card, SPIDER_NET_GMACOPEMD, 1833 SPIDER_NET_OPMODE_VALUE); 1834 1835 spider_net_write_reg(card, SPIDER_NET_GDTDMACCNTR, 1836 SPIDER_NET_GDTBSTA); 1837} 1838 1839/** 1840 * spider_net_download_firmware - loads firmware into the adapter 1841 * @card: card structure 1842 * @firmware_ptr: pointer to firmware data 1843 * 1844 * spider_net_download_firmware loads the firmware data into the 1845 * adapter. It assumes the length etc. to be allright. 1846 */ 1847static int 1848spider_net_download_firmware(struct spider_net_card *card, 1849 const void *firmware_ptr) 1850{ 1851 int sequencer, i; 1852 const u32 *fw_ptr = firmware_ptr; 1853 1854 /* stop sequencers */ 1855 spider_net_write_reg(card, SPIDER_NET_GSINIT, 1856 SPIDER_NET_STOP_SEQ_VALUE); 1857 1858 for (sequencer = 0; sequencer < SPIDER_NET_FIRMWARE_SEQS; 1859 sequencer++) { 1860 spider_net_write_reg(card, 1861 SPIDER_NET_GSnPRGADR + sequencer * 8, 0); 1862 for (i = 0; i < SPIDER_NET_FIRMWARE_SEQWORDS; i++) { 1863 spider_net_write_reg(card, SPIDER_NET_GSnPRGDAT + 1864 sequencer * 8, *fw_ptr); 1865 fw_ptr++; 1866 } 1867 } 1868 1869 if (spider_net_read_reg(card, SPIDER_NET_GSINIT)) 1870 return -EIO; 1871 1872 spider_net_write_reg(card, SPIDER_NET_GSINIT, 1873 SPIDER_NET_RUN_SEQ_VALUE); 1874 1875 return 0; 1876} 1877 1878/** 1879 * spider_net_init_firmware - reads in firmware parts 1880 * @card: card structure 1881 * 1882 * Returns 0 on success, <0 on failure 1883 * 1884 * spider_net_init_firmware opens the sequencer firmware and does some basic 1885 * checks. This function opens and releases the firmware structure. A call 1886 * to download the firmware is performed before the release. 1887 * 1888 * Firmware format 1889 * =============== 1890 * spider_fw.bin is expected to be a file containing 6*1024*4 bytes, 4k being 1891 * the program for each sequencer. Use the command 1892 * tail -q -n +2 Seq_code1_0x088.txt Seq_code2_0x090.txt \ 1893 * Seq_code3_0x098.txt Seq_code4_0x0A0.txt Seq_code5_0x0A8.txt \ 1894 * Seq_code6_0x0B0.txt | xxd -r -p -c4 > spider_fw.bin 1895 * 1896 * to generate spider_fw.bin, if you have sequencer programs with something 1897 * like the following contents for each sequencer: 1898 * <ONE LINE COMMENT> 1899 * <FIRST 4-BYTES-WORD FOR SEQUENCER> 1900 * <SECOND 4-BYTES-WORD FOR SEQUENCER> 1901 * ... 1902 * <1024th 4-BYTES-WORD FOR SEQUENCER> 1903 */ 1904static int 1905spider_net_init_firmware(struct spider_net_card *card) 1906{ 1907 struct firmware *firmware = NULL; 1908 struct device_node *dn; 1909 const u8 *fw_prop = NULL; 1910 int err = -ENOENT; 1911 int fw_size; 1912 1913 if (request_firmware((const struct firmware **)&firmware, 1914 SPIDER_NET_FIRMWARE_NAME, &card->pdev->dev) == 0) { 1915 if ( (firmware->size != SPIDER_NET_FIRMWARE_LEN) && 1916 netif_msg_probe(card) ) { 1917 dev_err(&card->netdev->dev, 1918 "Incorrect size of spidernet firmware in " \ 1919 "filesystem. Looking in host firmware...\n"); 1920 goto try_host_fw; 1921 } 1922 err = spider_net_download_firmware(card, firmware->data); 1923 1924 release_firmware(firmware); 1925 if (err) 1926 goto try_host_fw; 1927 1928 goto done; 1929 } 1930 1931try_host_fw: 1932 dn = pci_device_to_OF_node(card->pdev); 1933 if (!dn) 1934 goto out_err; 1935 1936 fw_prop = of_get_property(dn, "firmware", &fw_size); 1937 if (!fw_prop) 1938 goto out_err; 1939 1940 if ( (fw_size != SPIDER_NET_FIRMWARE_LEN) && 1941 netif_msg_probe(card) ) { 1942 dev_err(&card->netdev->dev, 1943 "Incorrect size of spidernet firmware in host firmware\n"); 1944 goto done; 1945 } 1946 1947 err = spider_net_download_firmware(card, fw_prop); 1948 1949done: 1950 return err; 1951out_err: 1952 if (netif_msg_probe(card)) 1953 dev_err(&card->netdev->dev, 1954 "Couldn't find spidernet firmware in filesystem " \ 1955 "or host firmware\n"); 1956 return err; 1957} 1958 1959/** 1960 * spider_net_open - called upon ifonfig up 1961 * @netdev: interface device structure 1962 * 1963 * returns 0 on success, <0 on failure 1964 * 1965 * spider_net_open allocates all the descriptors and memory needed for 1966 * operation, sets up multicast list and enables interrupts 1967 */ 1968int 1969spider_net_open(struct net_device *netdev) 1970{ 1971 struct spider_net_card *card = netdev_priv(netdev); 1972 int result; 1973 1974 result = spider_net_init_firmware(card); 1975 if (result) 1976 goto init_firmware_failed; 1977 1978 /* start probing with copper */ 1979 card->aneg_count = 0; 1980 card->medium = BCM54XX_COPPER; 1981 spider_net_setup_aneg(card); 1982 if (card->phy.def->phy_id) 1983 mod_timer(&card->aneg_timer, jiffies + SPIDER_NET_ANEG_TIMER); 1984 1985 result = spider_net_init_chain(card, &card->tx_chain); 1986 if (result) 1987 goto alloc_tx_failed; 1988 card->low_watermark = NULL; 1989 1990 result = spider_net_init_chain(card, &card->rx_chain); 1991 if (result) 1992 goto alloc_rx_failed; 1993 1994 /* Allocate rx skbs */ 1995 if (spider_net_alloc_rx_skbs(card)) 1996 goto alloc_skbs_failed; 1997 1998 spider_net_set_multi(netdev); 1999 2000 /* further enhancement: setup hw vlan, if needed */ 2001 2002 result = -EBUSY; 2003 if (request_irq(netdev->irq, spider_net_interrupt, 2004 IRQF_SHARED, netdev->name, netdev)) 2005 goto register_int_failed; 2006 2007 spider_net_enable_card(card); 2008 2009 netif_start_queue(netdev); 2010 netif_carrier_on(netdev); 2011 napi_enable(&card->napi); 2012 2013 spider_net_enable_interrupts(card); 2014 2015 return 0; 2016 2017register_int_failed: 2018 spider_net_free_rx_chain_contents(card); 2019alloc_skbs_failed: 2020 spider_net_free_chain(card, &card->rx_chain); 2021alloc_rx_failed: 2022 spider_net_free_chain(card, &card->tx_chain); 2023alloc_tx_failed: 2024 del_timer_sync(&card->aneg_timer); 2025init_firmware_failed: 2026 return result; 2027} 2028 2029/** 2030 * spider_net_link_phy 2031 * @data: used for pointer to card structure 2032 * 2033 */ 2034static void spider_net_link_phy(unsigned long data) 2035{ 2036 struct spider_net_card *card = (struct spider_net_card *)data; 2037 struct mii_phy *phy = &card->phy; 2038 2039 /* if link didn't come up after SPIDER_NET_ANEG_TIMEOUT tries, setup phy again */ 2040 if (card->aneg_count > SPIDER_NET_ANEG_TIMEOUT) { 2041 2042 pr_debug("%s: link is down trying to bring it up\n", 2043 card->netdev->name); 2044 2045 switch (card->medium) { 2046 case BCM54XX_COPPER: 2047 /* enable fiber with autonegotiation first */ 2048 if (phy->def->ops->enable_fiber) 2049 phy->def->ops->enable_fiber(phy, 1); 2050 card->medium = BCM54XX_FIBER; 2051 break; 2052 2053 case BCM54XX_FIBER: 2054 /* fiber didn't come up, try to disable fiber autoneg */ 2055 if (phy->def->ops->enable_fiber) 2056 phy->def->ops->enable_fiber(phy, 0); 2057 card->medium = BCM54XX_UNKNOWN; 2058 break; 2059 2060 case BCM54XX_UNKNOWN: 2061 /* copper, fiber with and without failed, 2062 * retry from beginning */ 2063 spider_net_setup_aneg(card); 2064 card->medium = BCM54XX_COPPER; 2065 break; 2066 } 2067 2068 card->aneg_count = 0; 2069 mod_timer(&card->aneg_timer, jiffies + SPIDER_NET_ANEG_TIMER); 2070 return; 2071 } 2072 2073 /* link still not up, try again later */ 2074 if (!(phy->def->ops->poll_link(phy))) { 2075 card->aneg_count++; 2076 mod_timer(&card->aneg_timer, jiffies + SPIDER_NET_ANEG_TIMER); 2077 return; 2078 } 2079 2080 /* link came up, get abilities */ 2081 phy->def->ops->read_link(phy); 2082 2083 spider_net_write_reg(card, SPIDER_NET_GMACST, 2084 spider_net_read_reg(card, SPIDER_NET_GMACST)); 2085 spider_net_write_reg(card, SPIDER_NET_GMACINTEN, 0x4); 2086 2087 if (phy->speed == 1000) 2088 spider_net_write_reg(card, SPIDER_NET_GMACMODE, 0x00000001); 2089 else 2090 spider_net_write_reg(card, SPIDER_NET_GMACMODE, 0); 2091 2092 card->aneg_count = 0; 2093 2094 pr_info("%s: link up, %i Mbps, %s-duplex %sautoneg.\n", 2095 card->netdev->name, phy->speed, 2096 phy->duplex == 1 ? "Full" : "Half", 2097 phy->autoneg == 1 ? "" : "no "); 2098} 2099 2100/** 2101 * spider_net_setup_phy - setup PHY 2102 * @card: card structure 2103 * 2104 * returns 0 on success, <0 on failure 2105 * 2106 * spider_net_setup_phy is used as part of spider_net_probe. 2107 **/ 2108static int 2109spider_net_setup_phy(struct spider_net_card *card) 2110{ 2111 struct mii_phy *phy = &card->phy; 2112 2113 spider_net_write_reg(card, SPIDER_NET_GDTDMASEL, 2114 SPIDER_NET_DMASEL_VALUE); 2115 spider_net_write_reg(card, SPIDER_NET_GPCCTRL, 2116 SPIDER_NET_PHY_CTRL_VALUE); 2117 2118 phy->dev = card->netdev; 2119 phy->mdio_read = spider_net_read_phy; 2120 phy->mdio_write = spider_net_write_phy; 2121 2122 for (phy->mii_id = 1; phy->mii_id <= 31; phy->mii_id++) { 2123 unsigned short id; 2124 id = spider_net_read_phy(card->netdev, phy->mii_id, MII_BMSR); 2125 if (id != 0x0000 && id != 0xffff) { 2126 if (!mii_phy_probe(phy, phy->mii_id)) { 2127 pr_info("Found %s.\n", phy->def->name); 2128 break; 2129 } 2130 } 2131 } 2132 2133 return 0; 2134} 2135 2136static void 2137spider_net_workaround_rxramfull(struct spider_net_card *card) 2138{ 2139 int i, sequencer = 0; 2140 2141 /* cancel reset */ 2142 spider_net_write_reg(card, SPIDER_NET_CKRCTRL, 2143 SPIDER_NET_CKRCTRL_RUN_VALUE); 2144 2145 /* empty sequencer data */ 2146 for (sequencer = 0; sequencer < SPIDER_NET_FIRMWARE_SEQS; 2147 sequencer++) { 2148 spider_net_write_reg(card, SPIDER_NET_GSnPRGADR + 2149 sequencer * 8, 0x0); 2150 for (i = 0; i < SPIDER_NET_FIRMWARE_SEQWORDS; i++) { 2151 spider_net_write_reg(card, SPIDER_NET_GSnPRGDAT + 2152 sequencer * 8, 0x0); 2153 } 2154 } 2155 2156 /* set sequencer operation */ 2157 spider_net_write_reg(card, SPIDER_NET_GSINIT, 0x000000fe); 2158 2159 /* reset */ 2160 spider_net_write_reg(card, SPIDER_NET_CKRCTRL, 2161 SPIDER_NET_CKRCTRL_STOP_VALUE); 2162} 2163 2164/** 2165 * spider_net_stop - called upon ifconfig down 2166 * @netdev: interface device structure 2167 * 2168 * always returns 0 2169 */ 2170int 2171spider_net_stop(struct net_device *netdev) 2172{ 2173 struct spider_net_card *card = netdev_priv(netdev); 2174 2175 napi_disable(&card->napi); 2176 netif_carrier_off(netdev); 2177 netif_stop_queue(netdev); 2178 del_timer_sync(&card->tx_timer); 2179 del_timer_sync(&card->aneg_timer); 2180 2181 spider_net_disable_interrupts(card); 2182 2183 free_irq(netdev->irq, netdev); 2184 2185 spider_net_write_reg(card, SPIDER_NET_GDTDMACCNTR, 2186 SPIDER_NET_DMA_TX_FEND_VALUE); 2187 2188 /* turn off DMA, force end */ 2189 spider_net_disable_rxdmac(card); 2190 2191 /* release chains */ 2192 spider_net_release_tx_chain(card, 1); 2193 spider_net_free_rx_chain_contents(card); 2194 2195 spider_net_free_chain(card, &card->tx_chain); 2196 spider_net_free_chain(card, &card->rx_chain); 2197 2198 return 0; 2199} 2200 2201/** 2202 * spider_net_tx_timeout_task - task scheduled by the watchdog timeout 2203 * function (to be called not under interrupt status) 2204 * @data: data, is interface device structure 2205 * 2206 * called as task when tx hangs, resets interface (if interface is up) 2207 */ 2208static void 2209spider_net_tx_timeout_task(struct work_struct *work) 2210{ 2211 struct spider_net_card *card = 2212 container_of(work, struct spider_net_card, tx_timeout_task); 2213 struct net_device *netdev = card->netdev; 2214 2215 if (!(netdev->flags & IFF_UP)) 2216 goto out; 2217 2218 netif_device_detach(netdev); 2219 spider_net_stop(netdev); 2220 2221 spider_net_workaround_rxramfull(card); 2222 spider_net_init_card(card); 2223 2224 if (spider_net_setup_phy(card)) 2225 goto out; 2226 2227 spider_net_open(netdev); 2228 spider_net_kick_tx_dma(card); 2229 netif_device_attach(netdev); 2230 2231out: 2232 atomic_dec(&card->tx_timeout_task_counter); 2233} 2234 2235/** 2236 * spider_net_tx_timeout - called when the tx timeout watchdog kicks in. 2237 * @netdev: interface device structure 2238 * 2239 * called, if tx hangs. Schedules a task that resets the interface 2240 */ 2241static void 2242spider_net_tx_timeout(struct net_device *netdev) 2243{ 2244 struct spider_net_card *card; 2245 2246 card = netdev_priv(netdev); 2247 atomic_inc(&card->tx_timeout_task_counter); 2248 if (netdev->flags & IFF_UP) 2249 schedule_work(&card->tx_timeout_task); 2250 else 2251 atomic_dec(&card->tx_timeout_task_counter); 2252 card->spider_stats.tx_timeouts++; 2253} 2254 2255static const struct net_device_ops spider_net_ops = { 2256 .ndo_open = spider_net_open, 2257 .ndo_stop = spider_net_stop, 2258 .ndo_start_xmit = spider_net_xmit, 2259 .ndo_set_multicast_list = spider_net_set_multi, 2260 .ndo_set_mac_address = spider_net_set_mac, 2261 .ndo_change_mtu = spider_net_change_mtu, 2262 .ndo_do_ioctl = spider_net_do_ioctl, 2263 .ndo_tx_timeout = spider_net_tx_timeout, 2264 .ndo_validate_addr = eth_validate_addr, 2265 /* HW VLAN */ 2266#ifdef CONFIG_NET_POLL_CONTROLLER 2267 /* poll controller */ 2268 .ndo_poll_controller = spider_net_poll_controller, 2269#endif /* CONFIG_NET_POLL_CONTROLLER */ 2270}; 2271 2272/** 2273 * spider_net_setup_netdev_ops - initialization of net_device operations 2274 * @netdev: net_device structure 2275 * 2276 * fills out function pointers in the net_device structure 2277 */ 2278static void 2279spider_net_setup_netdev_ops(struct net_device *netdev) 2280{ 2281 netdev->netdev_ops = &spider_net_ops; 2282 netdev->watchdog_timeo = SPIDER_NET_WATCHDOG_TIMEOUT; 2283 /* ethtool ops */ 2284 netdev->ethtool_ops = &spider_net_ethtool_ops; 2285} 2286 2287/** 2288 * spider_net_setup_netdev - initialization of net_device 2289 * @card: card structure 2290 * 2291 * Returns 0 on success or <0 on failure 2292 * 2293 * spider_net_setup_netdev initializes the net_device structure 2294 **/ 2295static int 2296spider_net_setup_netdev(struct spider_net_card *card) 2297{ 2298 int result; 2299 struct net_device *netdev = card->netdev; 2300 struct device_node *dn; 2301 struct sockaddr addr; 2302 const u8 *mac; 2303 2304 SET_NETDEV_DEV(netdev, &card->pdev->dev); 2305 2306 pci_set_drvdata(card->pdev, netdev); 2307 2308 init_timer(&card->tx_timer); 2309 card->tx_timer.function = 2310 (void (*)(unsigned long)) spider_net_cleanup_tx_ring; 2311 card->tx_timer.data = (unsigned long) card; 2312 netdev->irq = card->pdev->irq; 2313 2314 card->aneg_count = 0; 2315 init_timer(&card->aneg_timer); 2316 card->aneg_timer.function = spider_net_link_phy; 2317 card->aneg_timer.data = (unsigned long) card; 2318 2319 card->options.rx_csum = SPIDER_NET_RX_CSUM_DEFAULT; 2320 2321 netif_napi_add(netdev, &card->napi, 2322 spider_net_poll, SPIDER_NET_NAPI_WEIGHT); 2323 2324 spider_net_setup_netdev_ops(netdev); 2325 2326 netdev->features = NETIF_F_IP_CSUM | NETIF_F_LLTX; 2327 /* some time: NETIF_F_HW_VLAN_TX | NETIF_F_HW_VLAN_RX | 2328 * NETIF_F_HW_VLAN_FILTER */ 2329 2330 netdev->irq = card->pdev->irq; 2331 card->num_rx_ints = 0; 2332 card->ignore_rx_ramfull = 0; 2333 2334 dn = pci_device_to_OF_node(card->pdev); 2335 if (!dn) 2336 return -EIO; 2337 2338 mac = of_get_property(dn, "local-mac-address", NULL); 2339 if (!mac) 2340 return -EIO; 2341 memcpy(addr.sa_data, mac, ETH_ALEN); 2342 2343 result = spider_net_set_mac(netdev, &addr); 2344 if ((result) && (netif_msg_probe(card))) 2345 dev_err(&card->netdev->dev, 2346 "Failed to set MAC address: %i\n", result); 2347 2348 result = register_netdev(netdev); 2349 if (result) { 2350 if (netif_msg_probe(card)) 2351 dev_err(&card->netdev->dev, 2352 "Couldn't register net_device: %i\n", result); 2353 return result; 2354 } 2355 2356 if (netif_msg_probe(card)) 2357 pr_info("Initialized device %s.\n", netdev->name); 2358 2359 return 0; 2360} 2361 2362/** 2363 * spider_net_alloc_card - allocates net_device and card structure 2364 * 2365 * returns the card structure or NULL in case of errors 2366 * 2367 * the card and net_device structures are linked to each other 2368 */ 2369static struct spider_net_card * 2370spider_net_alloc_card(void) 2371{ 2372 struct net_device *netdev; 2373 struct spider_net_card *card; 2374 size_t alloc_size; 2375 2376 alloc_size = sizeof(struct spider_net_card) + 2377 (tx_descriptors + rx_descriptors) * sizeof(struct spider_net_descr); 2378 netdev = alloc_etherdev(alloc_size); 2379 if (!netdev) 2380 return NULL; 2381 2382 card = netdev_priv(netdev); 2383 card->netdev = netdev; 2384 card->msg_enable = SPIDER_NET_DEFAULT_MSG; 2385 INIT_WORK(&card->tx_timeout_task, spider_net_tx_timeout_task); 2386 init_waitqueue_head(&card->waitq); 2387 atomic_set(&card->tx_timeout_task_counter, 0); 2388 2389 card->rx_chain.num_desc = rx_descriptors; 2390 card->rx_chain.ring = card->darray; 2391 card->tx_chain.num_desc = tx_descriptors; 2392 card->tx_chain.ring = card->darray + rx_descriptors; 2393 2394 return card; 2395} 2396 2397/** 2398 * spider_net_undo_pci_setup - releases PCI ressources 2399 * @card: card structure 2400 * 2401 * spider_net_undo_pci_setup releases the mapped regions 2402 */ 2403static void 2404spider_net_undo_pci_setup(struct spider_net_card *card) 2405{ 2406 iounmap(card->regs); 2407 pci_release_regions(card->pdev); 2408} 2409 2410/** 2411 * spider_net_setup_pci_dev - sets up the device in terms of PCI operations 2412 * @pdev: PCI device 2413 * 2414 * Returns the card structure or NULL if any errors occur 2415 * 2416 * spider_net_setup_pci_dev initializes pdev and together with the 2417 * functions called in spider_net_open configures the device so that 2418 * data can be transferred over it 2419 * The net_device structure is attached to the card structure, if the 2420 * function returns without error. 2421 **/ 2422static struct spider_net_card * 2423spider_net_setup_pci_dev(struct pci_dev *pdev) 2424{ 2425 struct spider_net_card *card; 2426 unsigned long mmio_start, mmio_len; 2427 2428 if (pci_enable_device(pdev)) { 2429 dev_err(&pdev->dev, "Couldn't enable PCI device\n"); 2430 return NULL; 2431 } 2432 2433 if (!(pci_resource_flags(pdev, 0) & IORESOURCE_MEM)) { 2434 dev_err(&pdev->dev, 2435 "Couldn't find proper PCI device base address.\n"); 2436 goto out_disable_dev; 2437 } 2438 2439 if (pci_request_regions(pdev, spider_net_driver_name)) { 2440 dev_err(&pdev->dev, 2441 "Couldn't obtain PCI resources, aborting.\n"); 2442 goto out_disable_dev; 2443 } 2444 2445 pci_set_master(pdev); 2446 2447 card = spider_net_alloc_card(); 2448 if (!card) { 2449 dev_err(&pdev->dev, 2450 "Couldn't allocate net_device structure, aborting.\n"); 2451 goto out_release_regions; 2452 } 2453 card->pdev = pdev; 2454 2455 /* fetch base address and length of first resource */ 2456 mmio_start = pci_resource_start(pdev, 0); 2457 mmio_len = pci_resource_len(pdev, 0); 2458 2459 card->netdev->mem_start = mmio_start; 2460 card->netdev->mem_end = mmio_start + mmio_len; 2461 card->regs = ioremap(mmio_start, mmio_len); 2462 2463 if (!card->regs) { 2464 dev_err(&pdev->dev, 2465 "Couldn't obtain PCI resources, aborting.\n"); 2466 goto out_release_regions; 2467 } 2468 2469 return card; 2470 2471out_release_regions: 2472 pci_release_regions(pdev); 2473out_disable_dev: 2474 pci_disable_device(pdev); 2475 pci_set_drvdata(pdev, NULL); 2476 return NULL; 2477} 2478 2479/** 2480 * spider_net_probe - initialization of a device 2481 * @pdev: PCI device 2482 * @ent: entry in the device id list 2483 * 2484 * Returns 0 on success, <0 on failure 2485 * 2486 * spider_net_probe initializes pdev and registers a net_device 2487 * structure for it. After that, the device can be ifconfig'ed up 2488 **/ 2489static int __devinit 2490spider_net_probe(struct pci_dev *pdev, const struct pci_device_id *ent) 2491{ 2492 int err = -EIO; 2493 struct spider_net_card *card; 2494 2495 card = spider_net_setup_pci_dev(pdev); 2496 if (!card) 2497 goto out; 2498 2499 spider_net_workaround_rxramfull(card); 2500 spider_net_init_card(card); 2501 2502 err = spider_net_setup_phy(card); 2503 if (err) 2504 goto out_undo_pci; 2505 2506 err = spider_net_setup_netdev(card); 2507 if (err) 2508 goto out_undo_pci; 2509 2510 return 0; 2511 2512out_undo_pci: 2513 spider_net_undo_pci_setup(card); 2514 free_netdev(card->netdev); 2515out: 2516 return err; 2517} 2518 2519/** 2520 * spider_net_remove - removal of a device 2521 * @pdev: PCI device 2522 * 2523 * Returns 0 on success, <0 on failure 2524 * 2525 * spider_net_remove is called to remove the device and unregisters the 2526 * net_device 2527 **/ 2528static void __devexit 2529spider_net_remove(struct pci_dev *pdev) 2530{ 2531 struct net_device *netdev; 2532 struct spider_net_card *card; 2533 2534 netdev = pci_get_drvdata(pdev); 2535 card = netdev_priv(netdev); 2536 2537 wait_event(card->waitq, 2538 atomic_read(&card->tx_timeout_task_counter) == 0); 2539 2540 unregister_netdev(netdev); 2541 2542 /* switch off card */ 2543 spider_net_write_reg(card, SPIDER_NET_CKRCTRL, 2544 SPIDER_NET_CKRCTRL_STOP_VALUE); 2545 spider_net_write_reg(card, SPIDER_NET_CKRCTRL, 2546 SPIDER_NET_CKRCTRL_RUN_VALUE); 2547 2548 spider_net_undo_pci_setup(card); 2549 free_netdev(netdev); 2550} 2551 2552static struct pci_driver spider_net_driver = { 2553 .name = spider_net_driver_name, 2554 .id_table = spider_net_pci_tbl, 2555 .probe = spider_net_probe, 2556 .remove = __devexit_p(spider_net_remove) 2557}; 2558 2559/** 2560 * spider_net_init - init function when the driver is loaded 2561 * 2562 * spider_net_init registers the device driver 2563 */ 2564static int __init spider_net_init(void) 2565{ 2566 printk(KERN_INFO "Spidernet version %s.\n", VERSION); 2567 2568 if (rx_descriptors < SPIDER_NET_RX_DESCRIPTORS_MIN) { 2569 rx_descriptors = SPIDER_NET_RX_DESCRIPTORS_MIN; 2570 pr_info("adjusting rx descriptors to %i.\n", rx_descriptors); 2571 } 2572 if (rx_descriptors > SPIDER_NET_RX_DESCRIPTORS_MAX) { 2573 rx_descriptors = SPIDER_NET_RX_DESCRIPTORS_MAX; 2574 pr_info("adjusting rx descriptors to %i.\n", rx_descriptors); 2575 } 2576 if (tx_descriptors < SPIDER_NET_TX_DESCRIPTORS_MIN) { 2577 tx_descriptors = SPIDER_NET_TX_DESCRIPTORS_MIN; 2578 pr_info("adjusting tx descriptors to %i.\n", tx_descriptors); 2579 } 2580 if (tx_descriptors > SPIDER_NET_TX_DESCRIPTORS_MAX) { 2581 tx_descriptors = SPIDER_NET_TX_DESCRIPTORS_MAX; 2582 pr_info("adjusting tx descriptors to %i.\n", tx_descriptors); 2583 } 2584 2585 return pci_register_driver(&spider_net_driver); 2586} 2587 2588/** 2589 * spider_net_cleanup - exit function when driver is unloaded 2590 * 2591 * spider_net_cleanup unregisters the device driver 2592 */ 2593static void __exit spider_net_cleanup(void) 2594{ 2595 pci_unregister_driver(&spider_net_driver); 2596} 2597 2598module_init(spider_net_init); 2599module_exit(spider_net_cleanup); 2600