1/*- 2 * Copyright (c) 2000 Berkeley Software Design, Inc. 3 * Copyright (c) 1997, 1998, 1999, 2000 4 * Bill Paul <wpaul@osd.bsdi.com>. All rights reserved. 5 * 6 * Redistribution and use in source and binary forms, with or without 7 * modification, are permitted provided that the following conditions 8 * are met: 9 * 1. Redistributions of source code must retain the above copyright 10 * notice, this list of conditions and the following disclaimer. 11 * 2. Redistributions in binary form must reproduce the above copyright 12 * notice, this list of conditions and the following disclaimer in the 13 * documentation and/or other materials provided with the distribution. 14 * 3. All advertising materials mentioning features or use of this software 15 * must display the following acknowledgement: 16 * This product includes software developed by Bill Paul. 17 * 4. Neither the name of the author nor the names of any co-contributors 18 * may be used to endorse or promote products derived from this software 19 * without specific prior written permission. 20 * 21 * THIS SOFTWARE IS PROVIDED BY Bill Paul AND CONTRIBUTORS ``AS IS'' AND 22 * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE 23 * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE 24 * ARE DISCLAIMED. IN NO EVENT SHALL Bill Paul OR THE VOICES IN HIS HEAD 25 * BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR 26 * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF 27 * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS 28 * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN 29 * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) 30 * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF 31 * THE POSSIBILITY OF SUCH DAMAGE. 32 */ 33 34#include <sys/cdefs.h> 35__FBSDID("$FreeBSD$"); 36 37/* 38 * AMD Am79c972 fast ethernet PCI NIC driver. Datasheets are available 39 * from http://www.amd.com. 40 * 41 * The AMD PCnet/PCI controllers are more advanced and functional 42 * versions of the venerable 7990 LANCE. The PCnet/PCI chips retain 43 * backwards compatibility with the LANCE and thus can be made 44 * to work with older LANCE drivers. This is in fact how the 45 * PCnet/PCI chips were supported in FreeBSD originally. The trouble 46 * is that the PCnet/PCI devices offer several performance enhancements 47 * which can't be exploited in LANCE compatibility mode. Chief among 48 * these enhancements is the ability to perform PCI DMA operations 49 * using 32-bit addressing (which eliminates the need for ISA 50 * bounce-buffering), and special receive buffer alignment (which 51 * allows the receive handler to pass packets to the upper protocol 52 * layers without copying on both the x86 and alpha platforms). 53 */ 54 55#include <sys/param.h> 56#include <sys/systm.h> 57#include <sys/sockio.h> 58#include <sys/mbuf.h> 59#include <sys/malloc.h> 60#include <sys/kernel.h> 61#include <sys/module.h> 62#include <sys/socket.h> 63 64#include <net/if.h> 65#include <net/if_arp.h> 66#include <net/ethernet.h> 67#include <net/if_dl.h> 68#include <net/if_media.h> 69#include <net/if_types.h> 70 71#include <net/bpf.h> 72 73#include <vm/vm.h> /* for vtophys */ 74#include <vm/pmap.h> /* for vtophys */ 75#include <machine/bus.h> 76#include <machine/resource.h> 77#include <sys/bus.h> 78#include <sys/rman.h> 79 80#include <dev/mii/mii.h> 81#include <dev/mii/miivar.h> 82 83#include <dev/pci/pcireg.h> 84#include <dev/pci/pcivar.h> 85 86#define PCN_USEIOSPACE 87 88#include <dev/pcn/if_pcnreg.h> 89 90MODULE_DEPEND(pcn, pci, 1, 1, 1); 91MODULE_DEPEND(pcn, ether, 1, 1, 1); 92MODULE_DEPEND(pcn, miibus, 1, 1, 1); 93 94/* "device miibus" required. See GENERIC if you get errors here. */ 95#include "miibus_if.h" 96 97/* 98 * Various supported device vendors/types and their names. 99 */ 100static const struct pcn_type pcn_devs[] = { 101 { PCN_VENDORID, PCN_DEVICEID_PCNET, "AMD PCnet/PCI 10/100BaseTX" }, 102 { PCN_VENDORID, PCN_DEVICEID_HOME, "AMD PCnet/Home HomePNA" }, 103 { 0, 0, NULL } 104}; 105 106static const struct pcn_chipid { 107 u_int32_t id; 108 const char *name; 109} pcn_chipid[] = { 110 { Am79C971, "Am79C971" }, 111 { Am79C972, "Am79C972" }, 112 { Am79C973, "Am79C973" }, 113 { Am79C978, "Am79C978" }, 114 { Am79C975, "Am79C975" }, 115 { Am79C976, "Am79C976" }, 116 { 0, NULL }, 117}; 118 119static const char *pcn_chipid_name(u_int32_t); 120static u_int32_t pcn_chip_id(device_t); 121static const struct pcn_type *pcn_match(u_int16_t, u_int16_t); 122 123static u_int32_t pcn_csr_read(struct pcn_softc *, int); 124static u_int16_t pcn_csr_read16(struct pcn_softc *, int); 125static u_int16_t pcn_bcr_read16(struct pcn_softc *, int); 126static void pcn_csr_write(struct pcn_softc *, int, int); 127static u_int32_t pcn_bcr_read(struct pcn_softc *, int); 128static void pcn_bcr_write(struct pcn_softc *, int, int); 129 130static int pcn_probe(device_t); 131static int pcn_attach(device_t); 132static int pcn_detach(device_t); 133 134static int pcn_newbuf(struct pcn_softc *, int, struct mbuf *); 135static int pcn_encap(struct pcn_softc *, struct mbuf *, u_int32_t *); 136static void pcn_rxeof(struct pcn_softc *); 137static void pcn_txeof(struct pcn_softc *); 138static void pcn_intr(void *); 139static void pcn_tick(void *); 140static void pcn_start(struct ifnet *); 141static void pcn_start_locked(struct ifnet *); 142static int pcn_ioctl(struct ifnet *, u_long, caddr_t); 143static void pcn_init(void *); 144static void pcn_init_locked(struct pcn_softc *); 145static void pcn_stop(struct pcn_softc *); 146static void pcn_watchdog(struct pcn_softc *); 147static int pcn_shutdown(device_t); 148static int pcn_ifmedia_upd(struct ifnet *); 149static void pcn_ifmedia_sts(struct ifnet *, struct ifmediareq *); 150 151static int pcn_miibus_readreg(device_t, int, int); 152static int pcn_miibus_writereg(device_t, int, int, int); 153static void pcn_miibus_statchg(device_t); 154 155static void pcn_setfilt(struct ifnet *); 156static void pcn_setmulti(struct pcn_softc *); 157static void pcn_reset(struct pcn_softc *); 158static int pcn_list_rx_init(struct pcn_softc *); 159static int pcn_list_tx_init(struct pcn_softc *); 160 161#ifdef PCN_USEIOSPACE 162#define PCN_RES SYS_RES_IOPORT 163#define PCN_RID PCN_PCI_LOIO 164#else 165#define PCN_RES SYS_RES_MEMORY 166#define PCN_RID PCN_PCI_LOMEM 167#endif 168 169static device_method_t pcn_methods[] = { 170 /* Device interface */ 171 DEVMETHOD(device_probe, pcn_probe), 172 DEVMETHOD(device_attach, pcn_attach), 173 DEVMETHOD(device_detach, pcn_detach), 174 DEVMETHOD(device_shutdown, pcn_shutdown), 175 176 /* MII interface */ 177 DEVMETHOD(miibus_readreg, pcn_miibus_readreg), 178 DEVMETHOD(miibus_writereg, pcn_miibus_writereg), 179 DEVMETHOD(miibus_statchg, pcn_miibus_statchg), 180 181 DEVMETHOD_END 182}; 183 184static driver_t pcn_driver = { 185 "pcn", 186 pcn_methods, 187 sizeof(struct pcn_softc) 188}; 189 190static devclass_t pcn_devclass; 191 192DRIVER_MODULE(pcn, pci, pcn_driver, pcn_devclass, 0, 0); 193DRIVER_MODULE(miibus, pcn, miibus_driver, miibus_devclass, 0, 0); 194 195#define PCN_CSR_SETBIT(sc, reg, x) \ 196 pcn_csr_write(sc, reg, pcn_csr_read(sc, reg) | (x)) 197 198#define PCN_CSR_CLRBIT(sc, reg, x) \ 199 pcn_csr_write(sc, reg, pcn_csr_read(sc, reg) & ~(x)) 200 201#define PCN_BCR_SETBIT(sc, reg, x) \ 202 pcn_bcr_write(sc, reg, pcn_bcr_read(sc, reg) | (x)) 203 204#define PCN_BCR_CLRBIT(sc, reg, x) \ 205 pcn_bcr_write(sc, reg, pcn_bcr_read(sc, reg) & ~(x)) 206 207static u_int32_t 208pcn_csr_read(sc, reg) 209 struct pcn_softc *sc; 210 int reg; 211{ 212 CSR_WRITE_4(sc, PCN_IO32_RAP, reg); 213 return(CSR_READ_4(sc, PCN_IO32_RDP)); 214} 215 216static u_int16_t 217pcn_csr_read16(sc, reg) 218 struct pcn_softc *sc; 219 int reg; 220{ 221 CSR_WRITE_2(sc, PCN_IO16_RAP, reg); 222 return(CSR_READ_2(sc, PCN_IO16_RDP)); 223} 224 225static void 226pcn_csr_write(sc, reg, val) 227 struct pcn_softc *sc; 228 int reg; 229 int val; 230{ 231 CSR_WRITE_4(sc, PCN_IO32_RAP, reg); 232 CSR_WRITE_4(sc, PCN_IO32_RDP, val); 233 return; 234} 235 236static u_int32_t 237pcn_bcr_read(sc, reg) 238 struct pcn_softc *sc; 239 int reg; 240{ 241 CSR_WRITE_4(sc, PCN_IO32_RAP, reg); 242 return(CSR_READ_4(sc, PCN_IO32_BDP)); 243} 244 245static u_int16_t 246pcn_bcr_read16(sc, reg) 247 struct pcn_softc *sc; 248 int reg; 249{ 250 CSR_WRITE_2(sc, PCN_IO16_RAP, reg); 251 return(CSR_READ_2(sc, PCN_IO16_BDP)); 252} 253 254static void 255pcn_bcr_write(sc, reg, val) 256 struct pcn_softc *sc; 257 int reg; 258 int val; 259{ 260 CSR_WRITE_4(sc, PCN_IO32_RAP, reg); 261 CSR_WRITE_4(sc, PCN_IO32_BDP, val); 262 return; 263} 264 265static int 266pcn_miibus_readreg(dev, phy, reg) 267 device_t dev; 268 int phy, reg; 269{ 270 struct pcn_softc *sc; 271 int val; 272 273 sc = device_get_softc(dev); 274 275 /* 276 * At least Am79C971 with DP83840A wedge when isolating the 277 * external PHY so we can't allow multiple external PHYs. 278 * There are cards that use Am79C971 with both the internal 279 * and an external PHY though. 280 * For internal PHYs it doesn't really matter whether we can 281 * isolate the remaining internal and the external ones in 282 * the PHY drivers as the internal PHYs have to be enabled 283 * individually in PCN_BCR_PHYSEL, PCN_CSR_MODE, etc. 284 * With Am79C97{3,5,8} we don't support switching beetween 285 * the internal and external PHYs, yet, so we can't allow 286 * multiple PHYs with these either. 287 * Am79C97{2,6} actually only support external PHYs (not 288 * connectable internal ones respond at the usual addresses, 289 * which don't hurt if we let them show up on the bus) and 290 * isolating them works. 291 */ 292 if (((sc->pcn_type == Am79C971 && phy != PCN_PHYAD_10BT) || 293 sc->pcn_type == Am79C973 || sc->pcn_type == Am79C975 || 294 sc->pcn_type == Am79C978) && sc->pcn_extphyaddr != -1 && 295 phy != sc->pcn_extphyaddr) 296 return(0); 297 298 pcn_bcr_write(sc, PCN_BCR_MIIADDR, reg | (phy << 5)); 299 val = pcn_bcr_read(sc, PCN_BCR_MIIDATA) & 0xFFFF; 300 if (val == 0xFFFF) 301 return(0); 302 303 if (((sc->pcn_type == Am79C971 && phy != PCN_PHYAD_10BT) || 304 sc->pcn_type == Am79C973 || sc->pcn_type == Am79C975 || 305 sc->pcn_type == Am79C978) && sc->pcn_extphyaddr == -1) 306 sc->pcn_extphyaddr = phy; 307 308 return(val); 309} 310 311static int 312pcn_miibus_writereg(dev, phy, reg, data) 313 device_t dev; 314 int phy, reg, data; 315{ 316 struct pcn_softc *sc; 317 318 sc = device_get_softc(dev); 319 320 pcn_bcr_write(sc, PCN_BCR_MIIADDR, reg | (phy << 5)); 321 pcn_bcr_write(sc, PCN_BCR_MIIDATA, data); 322 323 return(0); 324} 325 326static void 327pcn_miibus_statchg(dev) 328 device_t dev; 329{ 330 struct pcn_softc *sc; 331 struct mii_data *mii; 332 333 sc = device_get_softc(dev); 334 mii = device_get_softc(sc->pcn_miibus); 335 336 if ((mii->mii_media_active & IFM_GMASK) == IFM_FDX) { 337 PCN_BCR_SETBIT(sc, PCN_BCR_DUPLEX, PCN_DUPLEX_FDEN); 338 } else { 339 PCN_BCR_CLRBIT(sc, PCN_BCR_DUPLEX, PCN_DUPLEX_FDEN); 340 } 341 342 return; 343} 344 345static void 346pcn_setmulti(sc) 347 struct pcn_softc *sc; 348{ 349 struct ifnet *ifp; 350 struct ifmultiaddr *ifma; 351 u_int32_t h, i; 352 u_int16_t hashes[4] = { 0, 0, 0, 0 }; 353 354 ifp = sc->pcn_ifp; 355 356 PCN_CSR_SETBIT(sc, PCN_CSR_EXTCTL1, PCN_EXTCTL1_SPND); 357 358 if (ifp->if_flags & IFF_ALLMULTI || ifp->if_flags & IFF_PROMISC) { 359 for (i = 0; i < 4; i++) 360 pcn_csr_write(sc, PCN_CSR_MAR0 + i, 0xFFFF); 361 PCN_CSR_CLRBIT(sc, PCN_CSR_EXTCTL1, PCN_EXTCTL1_SPND); 362 return; 363 } 364 365 /* first, zot all the existing hash bits */ 366 for (i = 0; i < 4; i++) 367 pcn_csr_write(sc, PCN_CSR_MAR0 + i, 0); 368 369 /* now program new ones */ 370 if_maddr_rlock(ifp); 371 TAILQ_FOREACH(ifma, &ifp->if_multiaddrs, ifma_link) { 372 if (ifma->ifma_addr->sa_family != AF_LINK) 373 continue; 374 h = ether_crc32_le(LLADDR((struct sockaddr_dl *) 375 ifma->ifma_addr), ETHER_ADDR_LEN) >> 26; 376 hashes[h >> 4] |= 1 << (h & 0xF); 377 } 378 if_maddr_runlock(ifp); 379 380 for (i = 0; i < 4; i++) 381 pcn_csr_write(sc, PCN_CSR_MAR0 + i, hashes[i]); 382 383 PCN_CSR_CLRBIT(sc, PCN_CSR_EXTCTL1, PCN_EXTCTL1_SPND); 384 385 return; 386} 387 388static void 389pcn_reset(sc) 390 struct pcn_softc *sc; 391{ 392 /* 393 * Issue a reset by reading from the RESET register. 394 * Note that we don't know if the chip is operating in 395 * 16-bit or 32-bit mode at this point, so we attempt 396 * to reset the chip both ways. If one fails, the other 397 * will succeed. 398 */ 399 CSR_READ_2(sc, PCN_IO16_RESET); 400 CSR_READ_4(sc, PCN_IO32_RESET); 401 402 /* Wait a little while for the chip to get its brains in order. */ 403 DELAY(1000); 404 405 /* Select 32-bit (DWIO) mode */ 406 CSR_WRITE_4(sc, PCN_IO32_RDP, 0); 407 408 /* Select software style 3. */ 409 pcn_bcr_write(sc, PCN_BCR_SSTYLE, PCN_SWSTYLE_PCNETPCI_BURST); 410 411 return; 412} 413 414static const char * 415pcn_chipid_name(u_int32_t id) 416{ 417 const struct pcn_chipid *p; 418 419 p = pcn_chipid; 420 while (p->name) { 421 if (id == p->id) 422 return (p->name); 423 p++; 424 } 425 return ("Unknown"); 426} 427 428static u_int32_t 429pcn_chip_id(device_t dev) 430{ 431 struct pcn_softc *sc; 432 u_int32_t chip_id; 433 434 sc = device_get_softc(dev); 435 /* 436 * Note: we can *NOT* put the chip into 437 * 32-bit mode yet. The le(4) driver will only 438 * work in 16-bit mode, and once the chip 439 * goes into 32-bit mode, the only way to 440 * get it out again is with a hardware reset. 441 * So if pcn_probe() is called before the 442 * le(4) driver's probe routine, the chip will 443 * be locked into 32-bit operation and the 444 * le(4) driver will be unable to attach to it. 445 * Note II: if the chip happens to already 446 * be in 32-bit mode, we still need to check 447 * the chip ID, but first we have to detect 448 * 32-bit mode using only 16-bit operations. 449 * The safest way to do this is to read the 450 * PCI subsystem ID from BCR23/24 and compare 451 * that with the value read from PCI config 452 * space. 453 */ 454 chip_id = pcn_bcr_read16(sc, PCN_BCR_PCISUBSYSID); 455 chip_id <<= 16; 456 chip_id |= pcn_bcr_read16(sc, PCN_BCR_PCISUBVENID); 457 /* 458 * Note III: the test for 0x10001000 is a hack to 459 * pacify VMware, who's pseudo-PCnet interface is 460 * broken. Reading the subsystem register from PCI 461 * config space yields 0x00000000 while reading the 462 * same value from I/O space yields 0x10001000. It's 463 * not supposed to be that way. 464 */ 465 if (chip_id == pci_read_config(dev, 466 PCIR_SUBVEND_0, 4) || chip_id == 0x10001000) { 467 /* We're in 16-bit mode. */ 468 chip_id = pcn_csr_read16(sc, PCN_CSR_CHIPID1); 469 chip_id <<= 16; 470 chip_id |= pcn_csr_read16(sc, PCN_CSR_CHIPID0); 471 } else { 472 /* We're in 32-bit mode. */ 473 chip_id = pcn_csr_read(sc, PCN_CSR_CHIPID1); 474 chip_id <<= 16; 475 chip_id |= pcn_csr_read(sc, PCN_CSR_CHIPID0); 476 } 477 478 return (chip_id); 479} 480 481static const struct pcn_type * 482pcn_match(u_int16_t vid, u_int16_t did) 483{ 484 const struct pcn_type *t; 485 486 t = pcn_devs; 487 while (t->pcn_name != NULL) { 488 if ((vid == t->pcn_vid) && (did == t->pcn_did)) 489 return (t); 490 t++; 491 } 492 return (NULL); 493} 494 495/* 496 * Probe for an AMD chip. Check the PCI vendor and device 497 * IDs against our list and return a device name if we find a match. 498 */ 499static int 500pcn_probe(dev) 501 device_t dev; 502{ 503 const struct pcn_type *t; 504 struct pcn_softc *sc; 505 int rid; 506 u_int32_t chip_id; 507 508 t = pcn_match(pci_get_vendor(dev), pci_get_device(dev)); 509 if (t == NULL) 510 return (ENXIO); 511 sc = device_get_softc(dev); 512 513 /* 514 * Temporarily map the I/O space so we can read the chip ID register. 515 */ 516 rid = PCN_RID; 517 sc->pcn_res = bus_alloc_resource_any(dev, PCN_RES, &rid, RF_ACTIVE); 518 if (sc->pcn_res == NULL) { 519 device_printf(dev, "couldn't map ports/memory\n"); 520 return(ENXIO); 521 } 522 sc->pcn_btag = rman_get_bustag(sc->pcn_res); 523 sc->pcn_bhandle = rman_get_bushandle(sc->pcn_res); 524 525 chip_id = pcn_chip_id(dev); 526 527 bus_release_resource(dev, PCN_RES, PCN_RID, sc->pcn_res); 528 529 switch((chip_id >> 12) & PART_MASK) { 530 case Am79C971: 531 case Am79C972: 532 case Am79C973: 533 case Am79C975: 534 case Am79C976: 535 case Am79C978: 536 break; 537 default: 538 return(ENXIO); 539 } 540 device_set_desc(dev, t->pcn_name); 541 return(BUS_PROBE_DEFAULT); 542} 543 544/* 545 * Attach the interface. Allocate softc structures, do ifmedia 546 * setup and ethernet/BPF attach. 547 */ 548static int 549pcn_attach(dev) 550 device_t dev; 551{ 552 u_int32_t eaddr[2]; 553 struct pcn_softc *sc; 554 struct mii_data *mii; 555 struct mii_softc *miisc; 556 struct ifnet *ifp; 557 int error = 0, rid; 558 559 sc = device_get_softc(dev); 560 561 /* Initialize our mutex. */ 562 mtx_init(&sc->pcn_mtx, device_get_nameunit(dev), MTX_NETWORK_LOCK, 563 MTX_DEF); 564 /* 565 * Map control/status registers. 566 */ 567 pci_enable_busmaster(dev); 568 569 /* Retrieve the chip ID */ 570 sc->pcn_type = (pcn_chip_id(dev) >> 12) & PART_MASK; 571 device_printf(dev, "Chip ID %04x (%s)\n", 572 sc->pcn_type, pcn_chipid_name(sc->pcn_type)); 573 574 rid = PCN_RID; 575 sc->pcn_res = bus_alloc_resource_any(dev, PCN_RES, &rid, RF_ACTIVE); 576 577 if (sc->pcn_res == NULL) { 578 device_printf(dev, "couldn't map ports/memory\n"); 579 error = ENXIO; 580 goto fail; 581 } 582 583 sc->pcn_btag = rman_get_bustag(sc->pcn_res); 584 sc->pcn_bhandle = rman_get_bushandle(sc->pcn_res); 585 586 /* Allocate interrupt */ 587 rid = 0; 588 sc->pcn_irq = bus_alloc_resource_any(dev, SYS_RES_IRQ, &rid, 589 RF_SHAREABLE | RF_ACTIVE); 590 591 if (sc->pcn_irq == NULL) { 592 device_printf(dev, "couldn't map interrupt\n"); 593 error = ENXIO; 594 goto fail; 595 } 596 597 /* Reset the adapter. */ 598 pcn_reset(sc); 599 600 /* 601 * Get station address from the EEPROM. 602 */ 603 eaddr[0] = CSR_READ_4(sc, PCN_IO32_APROM00); 604 eaddr[1] = CSR_READ_4(sc, PCN_IO32_APROM01); 605 606 callout_init_mtx(&sc->pcn_stat_callout, &sc->pcn_mtx, 0); 607 608 sc->pcn_ldata = contigmalloc(sizeof(struct pcn_list_data), M_DEVBUF, 609 M_NOWAIT, 0, 0xffffffff, PAGE_SIZE, 0); 610 611 if (sc->pcn_ldata == NULL) { 612 device_printf(dev, "no memory for list buffers!\n"); 613 error = ENXIO; 614 goto fail; 615 } 616 bzero(sc->pcn_ldata, sizeof(struct pcn_list_data)); 617 618 ifp = sc->pcn_ifp = if_alloc(IFT_ETHER); 619 if (ifp == NULL) { 620 device_printf(dev, "can not if_alloc()\n"); 621 error = ENOSPC; 622 goto fail; 623 } 624 ifp->if_softc = sc; 625 if_initname(ifp, device_get_name(dev), device_get_unit(dev)); 626 ifp->if_flags = IFF_BROADCAST | IFF_SIMPLEX | IFF_MULTICAST; 627 ifp->if_ioctl = pcn_ioctl; 628 ifp->if_start = pcn_start; 629 ifp->if_init = pcn_init; 630 ifp->if_snd.ifq_maxlen = PCN_TX_LIST_CNT - 1; 631 632 /* 633 * Do MII setup. 634 * See the comment in pcn_miibus_readreg() for why we can't 635 * universally pass MIIF_NOISOLATE here. 636 */ 637 sc->pcn_extphyaddr = -1; 638 error = mii_attach(dev, &sc->pcn_miibus, ifp, pcn_ifmedia_upd, 639 pcn_ifmedia_sts, BMSR_DEFCAPMASK, MII_PHY_ANY, MII_OFFSET_ANY, 0); 640 if (error != 0) { 641 device_printf(dev, "attaching PHYs failed\n"); 642 goto fail; 643 } 644 /* 645 * Record the media instances of internal PHYs, which map the 646 * built-in interfaces to the MII, so we can set the active 647 * PHY/port based on the currently selected media. 648 */ 649 sc->pcn_inst_10bt = -1; 650 mii = device_get_softc(sc->pcn_miibus); 651 LIST_FOREACH(miisc, &mii->mii_phys, mii_list) { 652 switch (miisc->mii_phy) { 653 case PCN_PHYAD_10BT: 654 sc->pcn_inst_10bt = miisc->mii_inst; 655 break; 656 /* 657 * XXX deal with the Am79C97{3,5} internal 100baseT 658 * and the Am79C978 internal HomePNA PHYs. 659 */ 660 } 661 } 662 663 /* 664 * Call MI attach routine. 665 */ 666 ether_ifattach(ifp, (u_int8_t *) eaddr); 667 668 /* Hook interrupt last to avoid having to lock softc */ 669 error = bus_setup_intr(dev, sc->pcn_irq, INTR_TYPE_NET | INTR_MPSAFE, 670 NULL, pcn_intr, sc, &sc->pcn_intrhand); 671 672 if (error) { 673 device_printf(dev, "couldn't set up irq\n"); 674 ether_ifdetach(ifp); 675 goto fail; 676 } 677 678fail: 679 if (error) 680 pcn_detach(dev); 681 682 return(error); 683} 684 685/* 686 * Shutdown hardware and free up resources. This can be called any 687 * time after the mutex has been initialized. It is called in both 688 * the error case in attach and the normal detach case so it needs 689 * to be careful about only freeing resources that have actually been 690 * allocated. 691 */ 692static int 693pcn_detach(dev) 694 device_t dev; 695{ 696 struct pcn_softc *sc; 697 struct ifnet *ifp; 698 699 sc = device_get_softc(dev); 700 ifp = sc->pcn_ifp; 701 702 KASSERT(mtx_initialized(&sc->pcn_mtx), ("pcn mutex not initialized")); 703 704 /* These should only be active if attach succeeded */ 705 if (device_is_attached(dev)) { 706 PCN_LOCK(sc); 707 pcn_reset(sc); 708 pcn_stop(sc); 709 PCN_UNLOCK(sc); 710 callout_drain(&sc->pcn_stat_callout); 711 ether_ifdetach(ifp); 712 } 713 if (sc->pcn_miibus) 714 device_delete_child(dev, sc->pcn_miibus); 715 bus_generic_detach(dev); 716 717 if (sc->pcn_intrhand) 718 bus_teardown_intr(dev, sc->pcn_irq, sc->pcn_intrhand); 719 if (sc->pcn_irq) 720 bus_release_resource(dev, SYS_RES_IRQ, 0, sc->pcn_irq); 721 if (sc->pcn_res) 722 bus_release_resource(dev, PCN_RES, PCN_RID, sc->pcn_res); 723 724 if (ifp) 725 if_free(ifp); 726 727 if (sc->pcn_ldata) { 728 contigfree(sc->pcn_ldata, sizeof(struct pcn_list_data), 729 M_DEVBUF); 730 } 731 732 mtx_destroy(&sc->pcn_mtx); 733 734 return(0); 735} 736 737/* 738 * Initialize the transmit descriptors. 739 */ 740static int 741pcn_list_tx_init(sc) 742 struct pcn_softc *sc; 743{ 744 struct pcn_list_data *ld; 745 struct pcn_ring_data *cd; 746 int i; 747 748 cd = &sc->pcn_cdata; 749 ld = sc->pcn_ldata; 750 751 for (i = 0; i < PCN_TX_LIST_CNT; i++) { 752 cd->pcn_tx_chain[i] = NULL; 753 ld->pcn_tx_list[i].pcn_tbaddr = 0; 754 ld->pcn_tx_list[i].pcn_txctl = 0; 755 ld->pcn_tx_list[i].pcn_txstat = 0; 756 } 757 758 cd->pcn_tx_prod = cd->pcn_tx_cons = cd->pcn_tx_cnt = 0; 759 760 return(0); 761} 762 763 764/* 765 * Initialize the RX descriptors and allocate mbufs for them. 766 */ 767static int 768pcn_list_rx_init(sc) 769 struct pcn_softc *sc; 770{ 771 struct pcn_ring_data *cd; 772 int i; 773 774 cd = &sc->pcn_cdata; 775 776 for (i = 0; i < PCN_RX_LIST_CNT; i++) { 777 if (pcn_newbuf(sc, i, NULL) == ENOBUFS) 778 return(ENOBUFS); 779 } 780 781 cd->pcn_rx_prod = 0; 782 783 return(0); 784} 785 786/* 787 * Initialize an RX descriptor and attach an MBUF cluster. 788 */ 789static int 790pcn_newbuf(sc, idx, m) 791 struct pcn_softc *sc; 792 int idx; 793 struct mbuf *m; 794{ 795 struct mbuf *m_new = NULL; 796 struct pcn_rx_desc *c; 797 798 c = &sc->pcn_ldata->pcn_rx_list[idx]; 799 800 if (m == NULL) { 801 MGETHDR(m_new, M_NOWAIT, MT_DATA); 802 if (m_new == NULL) 803 return(ENOBUFS); 804 805 MCLGET(m_new, M_NOWAIT); 806 if (!(m_new->m_flags & M_EXT)) { 807 m_freem(m_new); 808 return(ENOBUFS); 809 } 810 m_new->m_len = m_new->m_pkthdr.len = MCLBYTES; 811 } else { 812 m_new = m; 813 m_new->m_len = m_new->m_pkthdr.len = MCLBYTES; 814 m_new->m_data = m_new->m_ext.ext_buf; 815 } 816 817 m_adj(m_new, ETHER_ALIGN); 818 819 sc->pcn_cdata.pcn_rx_chain[idx] = m_new; 820 c->pcn_rbaddr = vtophys(mtod(m_new, caddr_t)); 821 c->pcn_bufsz = (~(PCN_RXLEN) + 1) & PCN_RXLEN_BUFSZ; 822 c->pcn_bufsz |= PCN_RXLEN_MBO; 823 c->pcn_rxstat = PCN_RXSTAT_STP|PCN_RXSTAT_ENP|PCN_RXSTAT_OWN; 824 825 return(0); 826} 827 828/* 829 * A frame has been uploaded: pass the resulting mbuf chain up to 830 * the higher level protocols. 831 */ 832static void 833pcn_rxeof(sc) 834 struct pcn_softc *sc; 835{ 836 struct mbuf *m; 837 struct ifnet *ifp; 838 struct pcn_rx_desc *cur_rx; 839 int i; 840 841 PCN_LOCK_ASSERT(sc); 842 843 ifp = sc->pcn_ifp; 844 i = sc->pcn_cdata.pcn_rx_prod; 845 846 while(PCN_OWN_RXDESC(&sc->pcn_ldata->pcn_rx_list[i])) { 847 cur_rx = &sc->pcn_ldata->pcn_rx_list[i]; 848 m = sc->pcn_cdata.pcn_rx_chain[i]; 849 sc->pcn_cdata.pcn_rx_chain[i] = NULL; 850 851 /* 852 * If an error occurs, update stats, clear the 853 * status word and leave the mbuf cluster in place: 854 * it should simply get re-used next time this descriptor 855 * comes up in the ring. 856 */ 857 if (cur_rx->pcn_rxstat & PCN_RXSTAT_ERR) { 858 ifp->if_ierrors++; 859 pcn_newbuf(sc, i, m); 860 PCN_INC(i, PCN_RX_LIST_CNT); 861 continue; 862 } 863 864 if (pcn_newbuf(sc, i, NULL)) { 865 /* Ran out of mbufs; recycle this one. */ 866 pcn_newbuf(sc, i, m); 867 ifp->if_ierrors++; 868 PCN_INC(i, PCN_RX_LIST_CNT); 869 continue; 870 } 871 872 PCN_INC(i, PCN_RX_LIST_CNT); 873 874 /* No errors; receive the packet. */ 875 ifp->if_ipackets++; 876 m->m_len = m->m_pkthdr.len = 877 cur_rx->pcn_rxlen - ETHER_CRC_LEN; 878 m->m_pkthdr.rcvif = ifp; 879 880 PCN_UNLOCK(sc); 881 (*ifp->if_input)(ifp, m); 882 PCN_LOCK(sc); 883 } 884 885 sc->pcn_cdata.pcn_rx_prod = i; 886 887 return; 888} 889 890/* 891 * A frame was downloaded to the chip. It's safe for us to clean up 892 * the list buffers. 893 */ 894 895static void 896pcn_txeof(sc) 897 struct pcn_softc *sc; 898{ 899 struct pcn_tx_desc *cur_tx = NULL; 900 struct ifnet *ifp; 901 u_int32_t idx; 902 903 ifp = sc->pcn_ifp; 904 905 /* 906 * Go through our tx list and free mbufs for those 907 * frames that have been transmitted. 908 */ 909 idx = sc->pcn_cdata.pcn_tx_cons; 910 while (idx != sc->pcn_cdata.pcn_tx_prod) { 911 cur_tx = &sc->pcn_ldata->pcn_tx_list[idx]; 912 913 if (!PCN_OWN_TXDESC(cur_tx)) 914 break; 915 916 if (!(cur_tx->pcn_txctl & PCN_TXCTL_ENP)) { 917 sc->pcn_cdata.pcn_tx_cnt--; 918 PCN_INC(idx, PCN_TX_LIST_CNT); 919 continue; 920 } 921 922 if (cur_tx->pcn_txctl & PCN_TXCTL_ERR) { 923 ifp->if_oerrors++; 924 if (cur_tx->pcn_txstat & PCN_TXSTAT_EXDEF) 925 ifp->if_collisions++; 926 if (cur_tx->pcn_txstat & PCN_TXSTAT_RTRY) 927 ifp->if_collisions++; 928 } 929 930 ifp->if_collisions += 931 cur_tx->pcn_txstat & PCN_TXSTAT_TRC; 932 933 ifp->if_opackets++; 934 if (sc->pcn_cdata.pcn_tx_chain[idx] != NULL) { 935 m_freem(sc->pcn_cdata.pcn_tx_chain[idx]); 936 sc->pcn_cdata.pcn_tx_chain[idx] = NULL; 937 } 938 939 sc->pcn_cdata.pcn_tx_cnt--; 940 PCN_INC(idx, PCN_TX_LIST_CNT); 941 } 942 943 if (idx != sc->pcn_cdata.pcn_tx_cons) { 944 /* Some buffers have been freed. */ 945 sc->pcn_cdata.pcn_tx_cons = idx; 946 ifp->if_drv_flags &= ~IFF_DRV_OACTIVE; 947 } 948 sc->pcn_timer = (sc->pcn_cdata.pcn_tx_cnt == 0) ? 0 : 5; 949 950 return; 951} 952 953static void 954pcn_tick(xsc) 955 void *xsc; 956{ 957 struct pcn_softc *sc; 958 struct mii_data *mii; 959 struct ifnet *ifp; 960 961 sc = xsc; 962 ifp = sc->pcn_ifp; 963 PCN_LOCK_ASSERT(sc); 964 965 mii = device_get_softc(sc->pcn_miibus); 966 mii_tick(mii); 967 968 /* link just died */ 969 if (sc->pcn_link & !(mii->mii_media_status & IFM_ACTIVE)) 970 sc->pcn_link = 0; 971 972 /* link just came up, restart */ 973 if (!sc->pcn_link && mii->mii_media_status & IFM_ACTIVE && 974 IFM_SUBTYPE(mii->mii_media_active) != IFM_NONE) { 975 sc->pcn_link++; 976 if (ifp->if_snd.ifq_head != NULL) 977 pcn_start_locked(ifp); 978 } 979 980 if (sc->pcn_timer > 0 && --sc->pcn_timer == 0) 981 pcn_watchdog(sc); 982 callout_reset(&sc->pcn_stat_callout, hz, pcn_tick, sc); 983 984 return; 985} 986 987static void 988pcn_intr(arg) 989 void *arg; 990{ 991 struct pcn_softc *sc; 992 struct ifnet *ifp; 993 u_int32_t status; 994 995 sc = arg; 996 ifp = sc->pcn_ifp; 997 998 PCN_LOCK(sc); 999 1000 /* Suppress unwanted interrupts */ 1001 if (!(ifp->if_flags & IFF_UP)) { 1002 pcn_stop(sc); 1003 PCN_UNLOCK(sc); 1004 return; 1005 } 1006 1007 CSR_WRITE_4(sc, PCN_IO32_RAP, PCN_CSR_CSR); 1008 1009 while ((status = CSR_READ_4(sc, PCN_IO32_RDP)) & PCN_CSR_INTR) { 1010 CSR_WRITE_4(sc, PCN_IO32_RDP, status); 1011 1012 if (status & PCN_CSR_RINT) 1013 pcn_rxeof(sc); 1014 1015 if (status & PCN_CSR_TINT) 1016 pcn_txeof(sc); 1017 1018 if (status & PCN_CSR_ERR) { 1019 pcn_init_locked(sc); 1020 break; 1021 } 1022 } 1023 1024 if (ifp->if_snd.ifq_head != NULL) 1025 pcn_start_locked(ifp); 1026 1027 PCN_UNLOCK(sc); 1028 return; 1029} 1030 1031/* 1032 * Encapsulate an mbuf chain in a descriptor by coupling the mbuf data 1033 * pointers to the fragment pointers. 1034 */ 1035static int 1036pcn_encap(sc, m_head, txidx) 1037 struct pcn_softc *sc; 1038 struct mbuf *m_head; 1039 u_int32_t *txidx; 1040{ 1041 struct pcn_tx_desc *f = NULL; 1042 struct mbuf *m; 1043 int frag, cur, cnt = 0; 1044 1045 /* 1046 * Start packing the mbufs in this chain into 1047 * the fragment pointers. Stop when we run out 1048 * of fragments or hit the end of the mbuf chain. 1049 */ 1050 m = m_head; 1051 cur = frag = *txidx; 1052 1053 for (m = m_head; m != NULL; m = m->m_next) { 1054 if (m->m_len == 0) 1055 continue; 1056 1057 if ((PCN_TX_LIST_CNT - (sc->pcn_cdata.pcn_tx_cnt + cnt)) < 2) 1058 return(ENOBUFS); 1059 f = &sc->pcn_ldata->pcn_tx_list[frag]; 1060 f->pcn_txctl = (~(m->m_len) + 1) & PCN_TXCTL_BUFSZ; 1061 f->pcn_txctl |= PCN_TXCTL_MBO; 1062 f->pcn_tbaddr = vtophys(mtod(m, vm_offset_t)); 1063 if (cnt == 0) 1064 f->pcn_txctl |= PCN_TXCTL_STP; 1065 else 1066 f->pcn_txctl |= PCN_TXCTL_OWN; 1067 cur = frag; 1068 PCN_INC(frag, PCN_TX_LIST_CNT); 1069 cnt++; 1070 } 1071 1072 if (m != NULL) 1073 return(ENOBUFS); 1074 1075 sc->pcn_cdata.pcn_tx_chain[cur] = m_head; 1076 sc->pcn_ldata->pcn_tx_list[cur].pcn_txctl |= 1077 PCN_TXCTL_ENP|PCN_TXCTL_ADD_FCS|PCN_TXCTL_MORE_LTINT; 1078 sc->pcn_ldata->pcn_tx_list[*txidx].pcn_txctl |= PCN_TXCTL_OWN; 1079 sc->pcn_cdata.pcn_tx_cnt += cnt; 1080 *txidx = frag; 1081 1082 return(0); 1083} 1084 1085/* 1086 * Main transmit routine. To avoid having to do mbuf copies, we put pointers 1087 * to the mbuf data regions directly in the transmit lists. We also save a 1088 * copy of the pointers since the transmit list fragment pointers are 1089 * physical addresses. 1090 */ 1091static void 1092pcn_start(ifp) 1093 struct ifnet *ifp; 1094{ 1095 struct pcn_softc *sc; 1096 1097 sc = ifp->if_softc; 1098 PCN_LOCK(sc); 1099 pcn_start_locked(ifp); 1100 PCN_UNLOCK(sc); 1101} 1102 1103static void 1104pcn_start_locked(ifp) 1105 struct ifnet *ifp; 1106{ 1107 struct pcn_softc *sc; 1108 struct mbuf *m_head = NULL; 1109 u_int32_t idx; 1110 1111 sc = ifp->if_softc; 1112 1113 PCN_LOCK_ASSERT(sc); 1114 1115 if (!sc->pcn_link) 1116 return; 1117 1118 idx = sc->pcn_cdata.pcn_tx_prod; 1119 1120 if (ifp->if_drv_flags & IFF_DRV_OACTIVE) 1121 return; 1122 1123 while(sc->pcn_cdata.pcn_tx_chain[idx] == NULL) { 1124 IF_DEQUEUE(&ifp->if_snd, m_head); 1125 if (m_head == NULL) 1126 break; 1127 1128 if (pcn_encap(sc, m_head, &idx)) { 1129 IF_PREPEND(&ifp->if_snd, m_head); 1130 ifp->if_drv_flags |= IFF_DRV_OACTIVE; 1131 break; 1132 } 1133 1134 /* 1135 * If there's a BPF listener, bounce a copy of this frame 1136 * to him. 1137 */ 1138 BPF_MTAP(ifp, m_head); 1139 1140 } 1141 1142 /* Transmit */ 1143 sc->pcn_cdata.pcn_tx_prod = idx; 1144 pcn_csr_write(sc, PCN_CSR_CSR, PCN_CSR_TX|PCN_CSR_INTEN); 1145 1146 /* 1147 * Set a timeout in case the chip goes out to lunch. 1148 */ 1149 sc->pcn_timer = 5; 1150 1151 return; 1152} 1153 1154static void 1155pcn_setfilt(ifp) 1156 struct ifnet *ifp; 1157{ 1158 struct pcn_softc *sc; 1159 1160 sc = ifp->if_softc; 1161 1162 /* If we want promiscuous mode, set the allframes bit. */ 1163 if (ifp->if_flags & IFF_PROMISC) { 1164 PCN_CSR_SETBIT(sc, PCN_CSR_MODE, PCN_MODE_PROMISC); 1165 } else { 1166 PCN_CSR_CLRBIT(sc, PCN_CSR_MODE, PCN_MODE_PROMISC); 1167 } 1168 1169 /* Set the capture broadcast bit to capture broadcast frames. */ 1170 if (ifp->if_flags & IFF_BROADCAST) { 1171 PCN_CSR_CLRBIT(sc, PCN_CSR_MODE, PCN_MODE_RXNOBROAD); 1172 } else { 1173 PCN_CSR_SETBIT(sc, PCN_CSR_MODE, PCN_MODE_RXNOBROAD); 1174 } 1175 1176 return; 1177} 1178 1179static void 1180pcn_init(xsc) 1181 void *xsc; 1182{ 1183 struct pcn_softc *sc = xsc; 1184 1185 PCN_LOCK(sc); 1186 pcn_init_locked(sc); 1187 PCN_UNLOCK(sc); 1188} 1189 1190static void 1191pcn_init_locked(sc) 1192 struct pcn_softc *sc; 1193{ 1194 struct ifnet *ifp = sc->pcn_ifp; 1195 struct mii_data *mii = NULL; 1196 struct ifmedia_entry *ife; 1197 1198 PCN_LOCK_ASSERT(sc); 1199 1200 /* 1201 * Cancel pending I/O and free all RX/TX buffers. 1202 */ 1203 pcn_stop(sc); 1204 pcn_reset(sc); 1205 1206 mii = device_get_softc(sc->pcn_miibus); 1207 ife = mii->mii_media.ifm_cur; 1208 1209 /* Set MAC address */ 1210 pcn_csr_write(sc, PCN_CSR_PAR0, 1211 ((u_int16_t *)IF_LLADDR(sc->pcn_ifp))[0]); 1212 pcn_csr_write(sc, PCN_CSR_PAR1, 1213 ((u_int16_t *)IF_LLADDR(sc->pcn_ifp))[1]); 1214 pcn_csr_write(sc, PCN_CSR_PAR2, 1215 ((u_int16_t *)IF_LLADDR(sc->pcn_ifp))[2]); 1216 1217 /* Init circular RX list. */ 1218 if (pcn_list_rx_init(sc) == ENOBUFS) { 1219 if_printf(ifp, "initialization failed: no " 1220 "memory for rx buffers\n"); 1221 pcn_stop(sc); 1222 return; 1223 } 1224 1225 /* 1226 * Init tx descriptors. 1227 */ 1228 pcn_list_tx_init(sc); 1229 1230 /* Clear PCN_MISC_ASEL so we can set the port via PCN_CSR_MODE. */ 1231 PCN_BCR_CLRBIT(sc, PCN_BCR_MISCCFG, PCN_MISC_ASEL); 1232 1233 /* 1234 * Set up the port based on the currently selected media. 1235 * For Am79C978 we've to unconditionally set PCN_PORT_MII and 1236 * set the PHY in PCN_BCR_PHYSEL instead. 1237 */ 1238 if (sc->pcn_type != Am79C978 && 1239 IFM_INST(ife->ifm_media) == sc->pcn_inst_10bt) 1240 pcn_csr_write(sc, PCN_CSR_MODE, PCN_PORT_10BASET); 1241 else 1242 pcn_csr_write(sc, PCN_CSR_MODE, PCN_PORT_MII); 1243 1244 /* Set up RX filter. */ 1245 pcn_setfilt(ifp); 1246 1247 /* 1248 * Load the multicast filter. 1249 */ 1250 pcn_setmulti(sc); 1251 1252 /* 1253 * Load the addresses of the RX and TX lists. 1254 */ 1255 pcn_csr_write(sc, PCN_CSR_RXADDR0, 1256 vtophys(&sc->pcn_ldata->pcn_rx_list[0]) & 0xFFFF); 1257 pcn_csr_write(sc, PCN_CSR_RXADDR1, 1258 (vtophys(&sc->pcn_ldata->pcn_rx_list[0]) >> 16) & 0xFFFF); 1259 pcn_csr_write(sc, PCN_CSR_TXADDR0, 1260 vtophys(&sc->pcn_ldata->pcn_tx_list[0]) & 0xFFFF); 1261 pcn_csr_write(sc, PCN_CSR_TXADDR1, 1262 (vtophys(&sc->pcn_ldata->pcn_tx_list[0]) >> 16) & 0xFFFF); 1263 1264 /* Set the RX and TX ring sizes. */ 1265 pcn_csr_write(sc, PCN_CSR_RXRINGLEN, (~PCN_RX_LIST_CNT) + 1); 1266 pcn_csr_write(sc, PCN_CSR_TXRINGLEN, (~PCN_TX_LIST_CNT) + 1); 1267 1268 /* We're not using the initialization block. */ 1269 pcn_csr_write(sc, PCN_CSR_IAB1, 0); 1270 1271 /* Enable fast suspend mode. */ 1272 PCN_CSR_SETBIT(sc, PCN_CSR_EXTCTL2, PCN_EXTCTL2_FASTSPNDE); 1273 1274 /* 1275 * Enable burst read and write. Also set the no underflow 1276 * bit. This will avoid transmit underruns in certain 1277 * conditions while still providing decent performance. 1278 */ 1279 PCN_BCR_SETBIT(sc, PCN_BCR_BUSCTL, PCN_BUSCTL_NOUFLOW| 1280 PCN_BUSCTL_BREAD|PCN_BUSCTL_BWRITE); 1281 1282 /* Enable graceful recovery from underflow. */ 1283 PCN_CSR_SETBIT(sc, PCN_CSR_IMR, PCN_IMR_DXSUFLO); 1284 1285 /* Enable auto-padding of short TX frames. */ 1286 PCN_CSR_SETBIT(sc, PCN_CSR_TFEAT, PCN_TFEAT_PAD_TX); 1287 1288 /* Disable MII autoneg (we handle this ourselves). */ 1289 PCN_BCR_SETBIT(sc, PCN_BCR_MIICTL, PCN_MIICTL_DANAS); 1290 1291 if (sc->pcn_type == Am79C978) 1292 /* XXX support other PHYs? */ 1293 pcn_bcr_write(sc, PCN_BCR_PHYSEL, 1294 PCN_PHYSEL_PCNET|PCN_PHY_HOMEPNA); 1295 1296 /* Enable interrupts and start the controller running. */ 1297 pcn_csr_write(sc, PCN_CSR_CSR, PCN_CSR_INTEN|PCN_CSR_START); 1298 1299 mii_mediachg(mii); 1300 1301 ifp->if_drv_flags |= IFF_DRV_RUNNING; 1302 ifp->if_drv_flags &= ~IFF_DRV_OACTIVE; 1303 1304 callout_reset(&sc->pcn_stat_callout, hz, pcn_tick, sc); 1305 1306 return; 1307} 1308 1309/* 1310 * Set media options. 1311 */ 1312static int 1313pcn_ifmedia_upd(ifp) 1314 struct ifnet *ifp; 1315{ 1316 struct pcn_softc *sc; 1317 1318 sc = ifp->if_softc; 1319 1320 PCN_LOCK(sc); 1321 1322 /* 1323 * At least Am79C971 with DP83840A can wedge when switching 1324 * from the internal 10baseT PHY to the external PHY without 1325 * issuing pcn_reset(). For setting the port in PCN_CSR_MODE 1326 * the PCnet chip has to be powered down or stopped anyway 1327 * and although documented otherwise it doesn't take effect 1328 * until the next initialization. 1329 */ 1330 sc->pcn_link = 0; 1331 pcn_stop(sc); 1332 pcn_reset(sc); 1333 pcn_init_locked(sc); 1334 if (ifp->if_snd.ifq_head != NULL) 1335 pcn_start_locked(ifp); 1336 1337 PCN_UNLOCK(sc); 1338 1339 return(0); 1340} 1341 1342/* 1343 * Report current media status. 1344 */ 1345static void 1346pcn_ifmedia_sts(ifp, ifmr) 1347 struct ifnet *ifp; 1348 struct ifmediareq *ifmr; 1349{ 1350 struct pcn_softc *sc; 1351 struct mii_data *mii; 1352 1353 sc = ifp->if_softc; 1354 1355 mii = device_get_softc(sc->pcn_miibus); 1356 PCN_LOCK(sc); 1357 mii_pollstat(mii); 1358 ifmr->ifm_active = mii->mii_media_active; 1359 ifmr->ifm_status = mii->mii_media_status; 1360 PCN_UNLOCK(sc); 1361 1362 return; 1363} 1364 1365static int 1366pcn_ioctl(ifp, command, data) 1367 struct ifnet *ifp; 1368 u_long command; 1369 caddr_t data; 1370{ 1371 struct pcn_softc *sc = ifp->if_softc; 1372 struct ifreq *ifr = (struct ifreq *) data; 1373 struct mii_data *mii = NULL; 1374 int error = 0; 1375 1376 switch(command) { 1377 case SIOCSIFFLAGS: 1378 PCN_LOCK(sc); 1379 if (ifp->if_flags & IFF_UP) { 1380 if (ifp->if_drv_flags & IFF_DRV_RUNNING && 1381 ifp->if_flags & IFF_PROMISC && 1382 !(sc->pcn_if_flags & IFF_PROMISC)) { 1383 PCN_CSR_SETBIT(sc, PCN_CSR_EXTCTL1, 1384 PCN_EXTCTL1_SPND); 1385 pcn_setfilt(ifp); 1386 PCN_CSR_CLRBIT(sc, PCN_CSR_EXTCTL1, 1387 PCN_EXTCTL1_SPND); 1388 pcn_csr_write(sc, PCN_CSR_CSR, 1389 PCN_CSR_INTEN|PCN_CSR_START); 1390 } else if (ifp->if_drv_flags & IFF_DRV_RUNNING && 1391 !(ifp->if_flags & IFF_PROMISC) && 1392 sc->pcn_if_flags & IFF_PROMISC) { 1393 PCN_CSR_SETBIT(sc, PCN_CSR_EXTCTL1, 1394 PCN_EXTCTL1_SPND); 1395 pcn_setfilt(ifp); 1396 PCN_CSR_CLRBIT(sc, PCN_CSR_EXTCTL1, 1397 PCN_EXTCTL1_SPND); 1398 pcn_csr_write(sc, PCN_CSR_CSR, 1399 PCN_CSR_INTEN|PCN_CSR_START); 1400 } else if (!(ifp->if_drv_flags & IFF_DRV_RUNNING)) 1401 pcn_init_locked(sc); 1402 } else { 1403 if (ifp->if_drv_flags & IFF_DRV_RUNNING) 1404 pcn_stop(sc); 1405 } 1406 sc->pcn_if_flags = ifp->if_flags; 1407 PCN_UNLOCK(sc); 1408 error = 0; 1409 break; 1410 case SIOCADDMULTI: 1411 case SIOCDELMULTI: 1412 PCN_LOCK(sc); 1413 pcn_setmulti(sc); 1414 PCN_UNLOCK(sc); 1415 error = 0; 1416 break; 1417 case SIOCGIFMEDIA: 1418 case SIOCSIFMEDIA: 1419 mii = device_get_softc(sc->pcn_miibus); 1420 error = ifmedia_ioctl(ifp, ifr, &mii->mii_media, command); 1421 break; 1422 default: 1423 error = ether_ioctl(ifp, command, data); 1424 break; 1425 } 1426 1427 return(error); 1428} 1429 1430static void 1431pcn_watchdog(struct pcn_softc *sc) 1432{ 1433 struct ifnet *ifp; 1434 1435 PCN_LOCK_ASSERT(sc); 1436 ifp = sc->pcn_ifp; 1437 1438 ifp->if_oerrors++; 1439 if_printf(ifp, "watchdog timeout\n"); 1440 1441 pcn_stop(sc); 1442 pcn_reset(sc); 1443 pcn_init_locked(sc); 1444 1445 if (ifp->if_snd.ifq_head != NULL) 1446 pcn_start_locked(ifp); 1447} 1448 1449/* 1450 * Stop the adapter and free any mbufs allocated to the 1451 * RX and TX lists. 1452 */ 1453static void 1454pcn_stop(struct pcn_softc *sc) 1455{ 1456 register int i; 1457 struct ifnet *ifp; 1458 1459 PCN_LOCK_ASSERT(sc); 1460 ifp = sc->pcn_ifp; 1461 sc->pcn_timer = 0; 1462 1463 callout_stop(&sc->pcn_stat_callout); 1464 1465 /* Turn off interrupts */ 1466 PCN_CSR_CLRBIT(sc, PCN_CSR_CSR, PCN_CSR_INTEN); 1467 /* Stop adapter */ 1468 PCN_CSR_SETBIT(sc, PCN_CSR_CSR, PCN_CSR_STOP); 1469 sc->pcn_link = 0; 1470 1471 /* 1472 * Free data in the RX lists. 1473 */ 1474 for (i = 0; i < PCN_RX_LIST_CNT; i++) { 1475 if (sc->pcn_cdata.pcn_rx_chain[i] != NULL) { 1476 m_freem(sc->pcn_cdata.pcn_rx_chain[i]); 1477 sc->pcn_cdata.pcn_rx_chain[i] = NULL; 1478 } 1479 } 1480 bzero((char *)&sc->pcn_ldata->pcn_rx_list, 1481 sizeof(sc->pcn_ldata->pcn_rx_list)); 1482 1483 /* 1484 * Free the TX list buffers. 1485 */ 1486 for (i = 0; i < PCN_TX_LIST_CNT; i++) { 1487 if (sc->pcn_cdata.pcn_tx_chain[i] != NULL) { 1488 m_freem(sc->pcn_cdata.pcn_tx_chain[i]); 1489 sc->pcn_cdata.pcn_tx_chain[i] = NULL; 1490 } 1491 } 1492 1493 bzero((char *)&sc->pcn_ldata->pcn_tx_list, 1494 sizeof(sc->pcn_ldata->pcn_tx_list)); 1495 1496 ifp->if_drv_flags &= ~(IFF_DRV_RUNNING | IFF_DRV_OACTIVE); 1497 1498 return; 1499} 1500 1501/* 1502 * Stop all chip I/O so that the kernel's probe routines don't 1503 * get confused by errant DMAs when rebooting. 1504 */ 1505static int 1506pcn_shutdown(device_t dev) 1507{ 1508 struct pcn_softc *sc; 1509 1510 sc = device_get_softc(dev); 1511 1512 PCN_LOCK(sc); 1513 pcn_reset(sc); 1514 pcn_stop(sc); 1515 PCN_UNLOCK(sc); 1516 1517 return 0; 1518} 1519