112#if EFSYS_OPT_MCDI_LOGGING 113#define SFXGE_PARAM_MCDI_LOGGING SFXGE_PARAM(mcdi_logging) 114static int sfxge_mcdi_logging = 0; 115TUNABLE_INT(SFXGE_PARAM_MCDI_LOGGING, &sfxge_mcdi_logging); 116#endif 117 118static void 119sfxge_reset(void *arg, int npending); 120 121static int 122sfxge_estimate_rsrc_limits(struct sfxge_softc *sc) 123{ 124 efx_drv_limits_t limits; 125 int rc; 126 unsigned int evq_max; 127 uint32_t evq_allocated; 128 uint32_t rxq_allocated; 129 uint32_t txq_allocated; 130 131 /* 132 * Limit the number of event queues to: 133 * - number of CPUs 134 * - hardwire maximum RSS channels 135 * - administratively specified maximum RSS channels 136 */ 137 evq_max = MIN(mp_ncpus, EFX_MAXRSS); 138 if (sc->max_rss_channels > 0) 139 evq_max = MIN(evq_max, sc->max_rss_channels); 140 141 memset(&limits, 0, sizeof(limits)); 142 143 limits.edl_min_evq_count = 1; 144 limits.edl_max_evq_count = evq_max; 145 limits.edl_min_txq_count = SFXGE_TXQ_NTYPES; 146 limits.edl_max_txq_count = evq_max + SFXGE_TXQ_NTYPES - 1; 147 limits.edl_min_rxq_count = 1; 148 limits.edl_max_rxq_count = evq_max; 149 150 efx_nic_set_drv_limits(sc->enp, &limits); 151 152 if ((rc = efx_nic_init(sc->enp)) != 0) 153 return (rc); 154 155 rc = efx_nic_get_vi_pool(sc->enp, &evq_allocated, &rxq_allocated, 156 &txq_allocated); 157 if (rc != 0) { 158 efx_nic_fini(sc->enp); 159 return (rc); 160 } 161 162 KASSERT(txq_allocated >= SFXGE_TXQ_NTYPES, 163 ("txq_allocated < SFXGE_TXQ_NTYPES")); 164 165 sc->evq_max = MIN(evq_allocated, evq_max); 166 sc->evq_max = MIN(rxq_allocated, sc->evq_max); 167 sc->evq_max = MIN(txq_allocated - (SFXGE_TXQ_NTYPES - 1), 168 sc->evq_max); 169 170 KASSERT(sc->evq_max <= evq_max, 171 ("allocated more than maximum requested")); 172 173 /* 174 * NIC is kept initialized in the case of success to be able to 175 * initialize port to find out media types. 176 */ 177 return (0); 178} 179 180static int 181sfxge_set_drv_limits(struct sfxge_softc *sc) 182{ 183 efx_drv_limits_t limits; 184 185 memset(&limits, 0, sizeof(limits)); 186 187 /* Limits are strict since take into account initial estimation */ 188 limits.edl_min_evq_count = limits.edl_max_evq_count = 189 sc->intr.n_alloc; 190 limits.edl_min_txq_count = limits.edl_max_txq_count = 191 sc->intr.n_alloc + SFXGE_TXQ_NTYPES - 1; 192 limits.edl_min_rxq_count = limits.edl_max_rxq_count = 193 sc->intr.n_alloc; 194 195 return (efx_nic_set_drv_limits(sc->enp, &limits)); 196} 197 198static int 199sfxge_start(struct sfxge_softc *sc) 200{ 201 int rc; 202 203 SFXGE_ADAPTER_LOCK_ASSERT_OWNED(sc); 204 205 if (sc->init_state == SFXGE_STARTED) 206 return (0); 207 208 if (sc->init_state != SFXGE_REGISTERED) { 209 rc = EINVAL; 210 goto fail; 211 } 212 213 /* Set required resource limits */ 214 if ((rc = sfxge_set_drv_limits(sc)) != 0) 215 goto fail; 216 217 if ((rc = efx_nic_init(sc->enp)) != 0) 218 goto fail; 219 220 /* Start processing interrupts. */ 221 if ((rc = sfxge_intr_start(sc)) != 0) 222 goto fail2; 223 224 /* Start processing events. */ 225 if ((rc = sfxge_ev_start(sc)) != 0) 226 goto fail3; 227 228 /* Fire up the port. */ 229 if ((rc = sfxge_port_start(sc)) != 0) 230 goto fail4; 231 232 /* Start the receiver side. */ 233 if ((rc = sfxge_rx_start(sc)) != 0) 234 goto fail5; 235 236 /* Start the transmitter side. */ 237 if ((rc = sfxge_tx_start(sc)) != 0) 238 goto fail6; 239 240 sc->init_state = SFXGE_STARTED; 241 242 /* Tell the stack we're running. */ 243 sc->ifnet->if_drv_flags |= IFF_DRV_RUNNING; 244 sc->ifnet->if_drv_flags &= ~IFF_DRV_OACTIVE; 245 246 return (0); 247 248fail6: 249 sfxge_rx_stop(sc); 250 251fail5: 252 sfxge_port_stop(sc); 253 254fail4: 255 sfxge_ev_stop(sc); 256 257fail3: 258 sfxge_intr_stop(sc); 259 260fail2: 261 efx_nic_fini(sc->enp); 262 263fail: 264 device_printf(sc->dev, "sfxge_start: %d\n", rc); 265 266 return (rc); 267} 268 269static void 270sfxge_if_init(void *arg) 271{ 272 struct sfxge_softc *sc; 273 274 sc = (struct sfxge_softc *)arg; 275 276 SFXGE_ADAPTER_LOCK(sc); 277 (void)sfxge_start(sc); 278 SFXGE_ADAPTER_UNLOCK(sc); 279} 280 281static void 282sfxge_stop(struct sfxge_softc *sc) 283{ 284 SFXGE_ADAPTER_LOCK_ASSERT_OWNED(sc); 285 286 if (sc->init_state != SFXGE_STARTED) 287 return; 288 289 sc->init_state = SFXGE_REGISTERED; 290 291 /* Stop the transmitter. */ 292 sfxge_tx_stop(sc); 293 294 /* Stop the receiver. */ 295 sfxge_rx_stop(sc); 296 297 /* Stop the port. */ 298 sfxge_port_stop(sc); 299 300 /* Stop processing events. */ 301 sfxge_ev_stop(sc); 302 303 /* Stop processing interrupts. */ 304 sfxge_intr_stop(sc); 305 306 efx_nic_fini(sc->enp); 307 308 sc->ifnet->if_drv_flags &= ~IFF_DRV_RUNNING; 309} 310 311 312static int 313sfxge_vpd_ioctl(struct sfxge_softc *sc, sfxge_ioc_t *ioc) 314{ 315 efx_vpd_value_t value; 316 int rc = 0; 317 318 switch (ioc->u.vpd.op) { 319 case SFXGE_VPD_OP_GET_KEYWORD: 320 value.evv_tag = ioc->u.vpd.tag; 321 value.evv_keyword = ioc->u.vpd.keyword; 322 rc = efx_vpd_get(sc->enp, sc->vpd_data, sc->vpd_size, &value); 323 if (rc != 0) 324 break; 325 ioc->u.vpd.len = MIN(ioc->u.vpd.len, value.evv_length); 326 if (ioc->u.vpd.payload != 0) { 327 rc = copyout(value.evv_value, ioc->u.vpd.payload, 328 ioc->u.vpd.len); 329 } 330 break; 331 case SFXGE_VPD_OP_SET_KEYWORD: 332 if (ioc->u.vpd.len > sizeof(value.evv_value)) 333 return (EINVAL); 334 value.evv_tag = ioc->u.vpd.tag; 335 value.evv_keyword = ioc->u.vpd.keyword; 336 value.evv_length = ioc->u.vpd.len; 337 rc = copyin(ioc->u.vpd.payload, value.evv_value, value.evv_length); 338 if (rc != 0) 339 break; 340 rc = efx_vpd_set(sc->enp, sc->vpd_data, sc->vpd_size, &value); 341 if (rc != 0) 342 break; 343 rc = efx_vpd_verify(sc->enp, sc->vpd_data, sc->vpd_size); 344 if (rc != 0) 345 break; 346 rc = efx_vpd_write(sc->enp, sc->vpd_data, sc->vpd_size); 347 break; 348 default: 349 rc = EOPNOTSUPP; 350 break; 351 } 352 353 return (rc); 354} 355 356static int 357sfxge_private_ioctl(struct sfxge_softc *sc, sfxge_ioc_t *ioc) 358{ 359 switch (ioc->op) { 360 case SFXGE_MCDI_IOC: 361 return (sfxge_mcdi_ioctl(sc, ioc)); 362 case SFXGE_NVRAM_IOC: 363 return (sfxge_nvram_ioctl(sc, ioc)); 364 case SFXGE_VPD_IOC: 365 return (sfxge_vpd_ioctl(sc, ioc)); 366 default: 367 return (EOPNOTSUPP); 368 } 369} 370 371 372static int 373sfxge_if_ioctl(struct ifnet *ifp, unsigned long command, caddr_t data) 374{ 375 struct sfxge_softc *sc; 376 struct ifreq *ifr; 377 sfxge_ioc_t ioc; 378 int error; 379 380 ifr = (struct ifreq *)data; 381 sc = ifp->if_softc; 382 error = 0; 383 384 switch (command) { 385 case SIOCSIFFLAGS: 386 SFXGE_ADAPTER_LOCK(sc); 387 if (ifp->if_flags & IFF_UP) { 388 if (ifp->if_drv_flags & IFF_DRV_RUNNING) { 389 if ((ifp->if_flags ^ sc->if_flags) & 390 (IFF_PROMISC | IFF_ALLMULTI)) { 391 sfxge_mac_filter_set(sc); 392 } 393 } else 394 sfxge_start(sc); 395 } else 396 if (ifp->if_drv_flags & IFF_DRV_RUNNING) 397 sfxge_stop(sc); 398 sc->if_flags = ifp->if_flags; 399 SFXGE_ADAPTER_UNLOCK(sc); 400 break; 401 case SIOCSIFMTU: 402 if (ifr->ifr_mtu == ifp->if_mtu) { 403 /* Nothing to do */ 404 error = 0; 405 } else if (ifr->ifr_mtu > SFXGE_MAX_MTU) { 406 error = EINVAL; 407 } else if (!(ifp->if_drv_flags & IFF_DRV_RUNNING)) { 408 ifp->if_mtu = ifr->ifr_mtu; 409 error = 0; 410 } else { 411 /* Restart required */ 412 SFXGE_ADAPTER_LOCK(sc); 413 sfxge_stop(sc); 414 ifp->if_mtu = ifr->ifr_mtu; 415 error = sfxge_start(sc); 416 SFXGE_ADAPTER_UNLOCK(sc); 417 if (error != 0) { 418 ifp->if_flags &= ~IFF_UP; 419 ifp->if_drv_flags &= ~IFF_DRV_RUNNING; 420 if_down(ifp); 421 } 422 } 423 break; 424 case SIOCADDMULTI: 425 case SIOCDELMULTI: 426 if (ifp->if_drv_flags & IFF_DRV_RUNNING) 427 sfxge_mac_filter_set(sc); 428 break; 429 case SIOCSIFCAP: 430 { 431 int reqcap = ifr->ifr_reqcap; 432 int capchg_mask; 433 434 SFXGE_ADAPTER_LOCK(sc); 435 436 /* Capabilities to be changed in accordance with request */ 437 capchg_mask = ifp->if_capenable ^ reqcap; 438 439 /* 440 * The networking core already rejects attempts to 441 * enable capabilities we don't have. We still have 442 * to reject attempts to disable capabilities that we 443 * can't (yet) disable. 444 */ 445 KASSERT((reqcap & ~ifp->if_capabilities) == 0, 446 ("Unsupported capabilities 0x%x requested 0x%x vs " 447 "supported 0x%x", 448 reqcap & ~ifp->if_capabilities, 449 reqcap , ifp->if_capabilities)); 450 if (capchg_mask & SFXGE_CAP_FIXED) { 451 error = EINVAL; 452 SFXGE_ADAPTER_UNLOCK(sc); 453 break; 454 } 455 456 /* Check request before any changes */ 457 if ((capchg_mask & IFCAP_TSO4) && 458 (reqcap & (IFCAP_TSO4 | IFCAP_TXCSUM)) == IFCAP_TSO4) { 459 error = EAGAIN; 460 SFXGE_ADAPTER_UNLOCK(sc); 461 if_printf(ifp, "enable txcsum before tso4\n"); 462 break; 463 } 464 if ((capchg_mask & IFCAP_TSO6) && 465 (reqcap & (IFCAP_TSO6 | IFCAP_TXCSUM_IPV6)) == IFCAP_TSO6) { 466 error = EAGAIN; 467 SFXGE_ADAPTER_UNLOCK(sc); 468 if_printf(ifp, "enable txcsum6 before tso6\n"); 469 break; 470 } 471 472 if (reqcap & IFCAP_TXCSUM) { 473 ifp->if_hwassist |= (CSUM_IP | CSUM_TCP | CSUM_UDP); 474 } else { 475 ifp->if_hwassist &= ~(CSUM_IP | CSUM_TCP | CSUM_UDP); 476 if (reqcap & IFCAP_TSO4) { 477 reqcap &= ~IFCAP_TSO4; 478 if_printf(ifp, 479 "tso4 disabled due to -txcsum\n"); 480 } 481 } 482 if (reqcap & IFCAP_TXCSUM_IPV6) { 483 ifp->if_hwassist |= (CSUM_TCP_IPV6 | CSUM_UDP_IPV6); 484 } else { 485 ifp->if_hwassist &= ~(CSUM_TCP_IPV6 | CSUM_UDP_IPV6); 486 if (reqcap & IFCAP_TSO6) { 487 reqcap &= ~IFCAP_TSO6; 488 if_printf(ifp, 489 "tso6 disabled due to -txcsum6\n"); 490 } 491 } 492 493 /* 494 * The kernel takes both IFCAP_TSOx and CSUM_TSO into 495 * account before using TSO. So, we do not touch 496 * checksum flags when IFCAP_TSOx is modified. 497 * Note that CSUM_TSO is (CSUM_IP_TSO|CSUM_IP6_TSO), 498 * but both bits are set in IPv4 and IPv6 mbufs. 499 */ 500 501 ifp->if_capenable = reqcap; 502 503 SFXGE_ADAPTER_UNLOCK(sc); 504 break; 505 } 506 case SIOCSIFMEDIA: 507 case SIOCGIFMEDIA: 508 error = ifmedia_ioctl(ifp, ifr, &sc->media, command); 509 break; 510 case SIOCGPRIVATE_0: 511 error = priv_check(curthread, PRIV_DRIVER); 512 if (error != 0) 513 break; 514 error = copyin(ifr->ifr_data, &ioc, sizeof(ioc)); 515 if (error != 0) 516 return (error); 517 error = sfxge_private_ioctl(sc, &ioc); 518 if (error == 0) { 519 error = copyout(&ioc, ifr->ifr_data, sizeof(ioc)); 520 } 521 break; 522 default: 523 error = ether_ioctl(ifp, command, data); 524 } 525 526 return (error); 527} 528 529static void 530sfxge_tick(void *arg) 531{ 532 struct sfxge_softc *sc = arg; 533 534 sfxge_port_update_stats(sc); 535 sfxge_tx_update_stats(sc); 536 537 callout_reset(&sc->tick_callout, sfxge_stats_update_period, 538 sfxge_tick, sc); 539} 540 541static void 542sfxge_ifnet_fini(struct ifnet *ifp) 543{ 544 struct sfxge_softc *sc = ifp->if_softc; 545 546 callout_drain(&sc->tick_callout); 547 548 SFXGE_ADAPTER_LOCK(sc); 549 sfxge_stop(sc); 550 SFXGE_ADAPTER_UNLOCK(sc); 551 552 ifmedia_removeall(&sc->media); 553 ether_ifdetach(ifp); 554 if_free(ifp); 555} 556 557static int 558sfxge_ifnet_init(struct ifnet *ifp, struct sfxge_softc *sc) 559{ 560 const efx_nic_cfg_t *encp = efx_nic_cfg_get(sc->enp); 561 device_t dev; 562 int rc; 563 564 dev = sc->dev; 565 sc->ifnet = ifp; 566 567 if_initname(ifp, device_get_name(dev), device_get_unit(dev)); 568 ifp->if_init = sfxge_if_init; 569 ifp->if_softc = sc; 570 ifp->if_flags = IFF_BROADCAST | IFF_SIMPLEX | IFF_MULTICAST; 571 ifp->if_ioctl = sfxge_if_ioctl; 572 573 ifp->if_capabilities = SFXGE_CAP; 574 ifp->if_capenable = SFXGE_CAP_ENABLE; 575 ifp->if_hw_tsomax = SFXGE_TSO_MAX_SIZE; 576 ifp->if_hw_tsomaxsegcount = SFXGE_TX_MAPPING_MAX_SEG; 577 ifp->if_hw_tsomaxsegsize = PAGE_SIZE; 578 579#ifdef SFXGE_LRO 580 ifp->if_capabilities |= IFCAP_LRO; 581 ifp->if_capenable |= IFCAP_LRO; 582#endif 583 584 if (encp->enc_hw_tx_insert_vlan_enabled) { 585 ifp->if_capabilities |= IFCAP_VLAN_HWTAGGING; 586 ifp->if_capenable |= IFCAP_VLAN_HWTAGGING; 587 } 588 ifp->if_hwassist = CSUM_TCP | CSUM_UDP | CSUM_IP | CSUM_TSO | 589 CSUM_TCP_IPV6 | CSUM_UDP_IPV6; 590 591 ether_ifattach(ifp, encp->enc_mac_addr); 592 593 ifp->if_transmit = sfxge_if_transmit; 594 ifp->if_qflush = sfxge_if_qflush; 595 596 callout_init(&sc->tick_callout, B_TRUE); 597 598 DBGPRINT(sc->dev, "ifmedia_init"); 599 if ((rc = sfxge_port_ifmedia_init(sc)) != 0) 600 goto fail; 601 602 callout_reset(&sc->tick_callout, sfxge_stats_update_period, 603 sfxge_tick, sc); 604 605 return (0); 606 607fail: 608 ether_ifdetach(sc->ifnet); 609 return (rc); 610} 611 612void 613sfxge_sram_buf_tbl_alloc(struct sfxge_softc *sc, size_t n, uint32_t *idp) 614{ 615 KASSERT(sc->buffer_table_next + n <= 616 efx_nic_cfg_get(sc->enp)->enc_buftbl_limit, 617 ("buffer table full")); 618 619 *idp = sc->buffer_table_next; 620 sc->buffer_table_next += n; 621} 622 623static int 624sfxge_bar_init(struct sfxge_softc *sc) 625{ 626 efsys_bar_t *esbp = &sc->bar; 627 628 esbp->esb_rid = PCIR_BAR(EFX_MEM_BAR); 629 if ((esbp->esb_res = bus_alloc_resource_any(sc->dev, SYS_RES_MEMORY, 630 &esbp->esb_rid, RF_ACTIVE)) == NULL) { 631 device_printf(sc->dev, "Cannot allocate BAR region %d\n", 632 EFX_MEM_BAR); 633 return (ENXIO); 634 } 635 esbp->esb_tag = rman_get_bustag(esbp->esb_res); 636 esbp->esb_handle = rman_get_bushandle(esbp->esb_res); 637 638 SFXGE_BAR_LOCK_INIT(esbp, device_get_nameunit(sc->dev)); 639 640 return (0); 641} 642 643static void 644sfxge_bar_fini(struct sfxge_softc *sc) 645{ 646 efsys_bar_t *esbp = &sc->bar; 647 648 bus_release_resource(sc->dev, SYS_RES_MEMORY, esbp->esb_rid, 649 esbp->esb_res); 650 SFXGE_BAR_LOCK_DESTROY(esbp); 651} 652 653static int 654sfxge_create(struct sfxge_softc *sc) 655{ 656 device_t dev; 657 efx_nic_t *enp; 658 int error; 659 char rss_param_name[sizeof(SFXGE_PARAM(%d.max_rss_channels))]; 660#if EFSYS_OPT_MCDI_LOGGING 661 char mcdi_log_param_name[sizeof(SFXGE_PARAM(%d.mcdi_logging))]; 662#endif 663 664 dev = sc->dev; 665 666 SFXGE_ADAPTER_LOCK_INIT(sc, device_get_nameunit(sc->dev)); 667 668 sc->max_rss_channels = 0; 669 snprintf(rss_param_name, sizeof(rss_param_name), 670 SFXGE_PARAM(%d.max_rss_channels), 671 (int)device_get_unit(dev)); 672 TUNABLE_INT_FETCH(rss_param_name, &sc->max_rss_channels); 673#if EFSYS_OPT_MCDI_LOGGING 674 sc->mcdi_logging = sfxge_mcdi_logging; 675 snprintf(mcdi_log_param_name, sizeof(mcdi_log_param_name), 676 SFXGE_PARAM(%d.mcdi_logging), 677 (int)device_get_unit(dev)); 678 TUNABLE_INT_FETCH(mcdi_log_param_name, &sc->mcdi_logging); 679#endif 680 681 sc->stats_node = SYSCTL_ADD_NODE( 682 device_get_sysctl_ctx(dev), 683 SYSCTL_CHILDREN(device_get_sysctl_tree(dev)), 684 OID_AUTO, "stats", CTLFLAG_RD, NULL, "Statistics"); 685 if (sc->stats_node == NULL) { 686 error = ENOMEM; 687 goto fail; 688 } 689 690 TASK_INIT(&sc->task_reset, 0, sfxge_reset, sc); 691 692 (void) pci_enable_busmaster(dev); 693 694 /* Initialize DMA mappings. */ 695 DBGPRINT(sc->dev, "dma_init..."); 696 if ((error = sfxge_dma_init(sc)) != 0) 697 goto fail; 698 699 /* Map the device registers. */ 700 DBGPRINT(sc->dev, "bar_init..."); 701 if ((error = sfxge_bar_init(sc)) != 0) 702 goto fail; 703 704 error = efx_family(pci_get_vendor(dev), pci_get_device(dev), 705 &sc->family); 706 KASSERT(error == 0, ("Family should be filtered by sfxge_probe()")); 707 708 DBGPRINT(sc->dev, "nic_create..."); 709 710 /* Create the common code nic object. */ 711 SFXGE_EFSYS_LOCK_INIT(&sc->enp_lock, 712 device_get_nameunit(sc->dev), "nic"); 713 if ((error = efx_nic_create(sc->family, (efsys_identifier_t *)sc, 714 &sc->bar, &sc->enp_lock, &enp)) != 0) 715 goto fail3; 716 sc->enp = enp; 717 718 if (!ISP2(sfxge_rx_ring_entries) || 719 (sfxge_rx_ring_entries < EFX_RXQ_MINNDESCS) || 720 (sfxge_rx_ring_entries > EFX_RXQ_MAXNDESCS)) { 721 log(LOG_ERR, "%s=%d must be power of 2 from %u to %u", 722 SFXGE_PARAM_RX_RING, sfxge_rx_ring_entries, 723 EFX_RXQ_MINNDESCS, EFX_RXQ_MAXNDESCS); 724 error = EINVAL; 725 goto fail_rx_ring_entries; 726 } 727 sc->rxq_entries = sfxge_rx_ring_entries; 728 729 if (!ISP2(sfxge_tx_ring_entries) || 730 (sfxge_tx_ring_entries < EFX_TXQ_MINNDESCS) || 731 (sfxge_tx_ring_entries > EFX_TXQ_MAXNDESCS(efx_nic_cfg_get(enp)))) { 732 log(LOG_ERR, "%s=%d must be power of 2 from %u to %u", 733 SFXGE_PARAM_TX_RING, sfxge_tx_ring_entries, 734 EFX_TXQ_MINNDESCS, EFX_TXQ_MAXNDESCS(efx_nic_cfg_get(enp))); 735 error = EINVAL; 736 goto fail_tx_ring_entries; 737 } 738 sc->txq_entries = sfxge_tx_ring_entries; 739 740 /* Initialize MCDI to talk to the microcontroller. */ 741 DBGPRINT(sc->dev, "mcdi_init..."); 742 if ((error = sfxge_mcdi_init(sc)) != 0) 743 goto fail4; 744 745 /* Probe the NIC and build the configuration data area. */ 746 DBGPRINT(sc->dev, "nic_probe..."); 747 if ((error = efx_nic_probe(enp)) != 0) 748 goto fail5; 749 750 SYSCTL_ADD_STRING(device_get_sysctl_ctx(dev), 751 SYSCTL_CHILDREN(device_get_sysctl_tree(dev)), 752 OID_AUTO, "version", CTLFLAG_RD, 753 SFXGE_VERSION_STRING, 0, 754 "Driver version"); 755 756 SYSCTL_ADD_UINT(device_get_sysctl_ctx(dev), 757 SYSCTL_CHILDREN(device_get_sysctl_tree(dev)), 758 OID_AUTO, "phy_type", CTLFLAG_RD, 759 NULL, efx_nic_cfg_get(enp)->enc_phy_type, 760 "PHY type"); 761 762 /* Initialize the NVRAM. */ 763 DBGPRINT(sc->dev, "nvram_init..."); 764 if ((error = efx_nvram_init(enp)) != 0) 765 goto fail6; 766 767 /* Initialize the VPD. */ 768 DBGPRINT(sc->dev, "vpd_init..."); 769 if ((error = efx_vpd_init(enp)) != 0) 770 goto fail7; 771 772 efx_mcdi_new_epoch(enp); 773 774 /* Reset the NIC. */ 775 DBGPRINT(sc->dev, "nic_reset..."); 776 if ((error = efx_nic_reset(enp)) != 0) 777 goto fail8; 778 779 /* Initialize buffer table allocation. */ 780 sc->buffer_table_next = 0; 781 782 /* 783 * Guarantee minimum and estimate maximum number of event queues 784 * to take it into account when MSI-X interrupts are allocated. 785 * It initializes NIC and keeps it initialized on success. 786 */ 787 if ((error = sfxge_estimate_rsrc_limits(sc)) != 0) 788 goto fail8; 789 790 /* Set up interrupts. */ 791 DBGPRINT(sc->dev, "intr_init..."); 792 if ((error = sfxge_intr_init(sc)) != 0) 793 goto fail9; 794 795 /* Initialize event processing state. */ 796 DBGPRINT(sc->dev, "ev_init..."); 797 if ((error = sfxge_ev_init(sc)) != 0) 798 goto fail11; 799 800 /* Initialize port state. */ 801 DBGPRINT(sc->dev, "port_init..."); 802 if ((error = sfxge_port_init(sc)) != 0) 803 goto fail12; 804 805 /* Initialize receive state. */ 806 DBGPRINT(sc->dev, "rx_init..."); 807 if ((error = sfxge_rx_init(sc)) != 0) 808 goto fail13; 809 810 /* Initialize transmit state. */ 811 DBGPRINT(sc->dev, "tx_init..."); 812 if ((error = sfxge_tx_init(sc)) != 0) 813 goto fail14; 814 815 sc->init_state = SFXGE_INITIALIZED; 816 817 DBGPRINT(sc->dev, "success"); 818 return (0); 819 820fail14: 821 sfxge_rx_fini(sc); 822 823fail13: 824 sfxge_port_fini(sc); 825 826fail12: 827 sfxge_ev_fini(sc); 828 829fail11: 830 sfxge_intr_fini(sc); 831 832fail9: 833 efx_nic_fini(sc->enp); 834 835fail8: 836 efx_vpd_fini(enp); 837 838fail7: 839 efx_nvram_fini(enp); 840 841fail6: 842 efx_nic_unprobe(enp); 843 844fail5: 845 sfxge_mcdi_fini(sc); 846 847fail4: 848fail_tx_ring_entries: 849fail_rx_ring_entries: 850 sc->enp = NULL; 851 efx_nic_destroy(enp); 852 SFXGE_EFSYS_LOCK_DESTROY(&sc->enp_lock); 853 854fail3: 855 sfxge_bar_fini(sc); 856 (void) pci_disable_busmaster(sc->dev); 857 858fail: 859 DBGPRINT(sc->dev, "failed %d", error); 860 sc->dev = NULL; 861 SFXGE_ADAPTER_LOCK_DESTROY(sc); 862 return (error); 863} 864 865static void 866sfxge_destroy(struct sfxge_softc *sc) 867{ 868 efx_nic_t *enp; 869 870 /* Clean up transmit state. */ 871 sfxge_tx_fini(sc); 872 873 /* Clean up receive state. */ 874 sfxge_rx_fini(sc); 875 876 /* Clean up port state. */ 877 sfxge_port_fini(sc); 878 879 /* Clean up event processing state. */ 880 sfxge_ev_fini(sc); 881 882 /* Clean up interrupts. */ 883 sfxge_intr_fini(sc); 884 885 /* Tear down common code subsystems. */ 886 efx_nic_reset(sc->enp); 887 efx_vpd_fini(sc->enp); 888 efx_nvram_fini(sc->enp); 889 efx_nic_unprobe(sc->enp); 890 891 /* Tear down MCDI. */ 892 sfxge_mcdi_fini(sc); 893 894 /* Destroy common code context. */ 895 enp = sc->enp; 896 sc->enp = NULL; 897 efx_nic_destroy(enp); 898 899 /* Free DMA memory. */ 900 sfxge_dma_fini(sc); 901 902 /* Free mapped BARs. */ 903 sfxge_bar_fini(sc); 904 905 (void) pci_disable_busmaster(sc->dev); 906 907 taskqueue_drain(taskqueue_thread, &sc->task_reset); 908 909 /* Destroy the softc lock. */ 910 SFXGE_ADAPTER_LOCK_DESTROY(sc); 911} 912 913static int 914sfxge_vpd_handler(SYSCTL_HANDLER_ARGS) 915{ 916 struct sfxge_softc *sc = arg1; 917 efx_vpd_value_t value; 918 int rc; 919 920 value.evv_tag = arg2 >> 16; 921 value.evv_keyword = arg2 & 0xffff; 922 if ((rc = efx_vpd_get(sc->enp, sc->vpd_data, sc->vpd_size, &value)) 923 != 0) 924 return (rc); 925 926 return (SYSCTL_OUT(req, value.evv_value, value.evv_length)); 927} 928 929static void 930sfxge_vpd_try_add(struct sfxge_softc *sc, struct sysctl_oid_list *list, 931 efx_vpd_tag_t tag, const char *keyword) 932{ 933 struct sysctl_ctx_list *ctx = device_get_sysctl_ctx(sc->dev); 934 efx_vpd_value_t value; 935 936 /* Check whether VPD tag/keyword is present */ 937 value.evv_tag = tag; 938 value.evv_keyword = EFX_VPD_KEYWORD(keyword[0], keyword[1]); 939 if (efx_vpd_get(sc->enp, sc->vpd_data, sc->vpd_size, &value) != 0) 940 return; 941 942 SYSCTL_ADD_PROC( 943 ctx, list, OID_AUTO, keyword, CTLTYPE_STRING|CTLFLAG_RD, 944 sc, tag << 16 | EFX_VPD_KEYWORD(keyword[0], keyword[1]), 945 sfxge_vpd_handler, "A", ""); 946} 947 948static int 949sfxge_vpd_init(struct sfxge_softc *sc) 950{ 951 struct sysctl_ctx_list *ctx = device_get_sysctl_ctx(sc->dev); 952 struct sysctl_oid *vpd_node; 953 struct sysctl_oid_list *vpd_list; 954 char keyword[3]; 955 efx_vpd_value_t value; 956 int rc; 957 958 if ((rc = efx_vpd_size(sc->enp, &sc->vpd_size)) != 0) { 959 /* 960 * Unpriviledged functions deny VPD access. 961 * Simply skip VPD in this case. 962 */ 963 if (rc == EACCES) 964 goto done; 965 goto fail; 966 } 967 sc->vpd_data = malloc(sc->vpd_size, M_SFXGE, M_WAITOK); 968 if ((rc = efx_vpd_read(sc->enp, sc->vpd_data, sc->vpd_size)) != 0) 969 goto fail2; 970 971 /* Copy ID (product name) into device description, and log it. */ 972 value.evv_tag = EFX_VPD_ID; 973 if (efx_vpd_get(sc->enp, sc->vpd_data, sc->vpd_size, &value) == 0) { 974 value.evv_value[value.evv_length] = 0; 975 device_set_desc_copy(sc->dev, value.evv_value); 976 device_printf(sc->dev, "%s\n", value.evv_value); 977 } 978 979 vpd_node = SYSCTL_ADD_NODE( 980 ctx, SYSCTL_CHILDREN(device_get_sysctl_tree(sc->dev)), 981 OID_AUTO, "vpd", CTLFLAG_RD, NULL, "Vital Product Data"); 982 vpd_list = SYSCTL_CHILDREN(vpd_node); 983 984 /* Add sysctls for all expected and any vendor-defined keywords. */ 985 sfxge_vpd_try_add(sc, vpd_list, EFX_VPD_RO, "PN"); 986 sfxge_vpd_try_add(sc, vpd_list, EFX_VPD_RO, "EC"); 987 sfxge_vpd_try_add(sc, vpd_list, EFX_VPD_RO, "SN"); 988 keyword[0] = 'V'; 989 keyword[2] = 0; 990 for (keyword[1] = '0'; keyword[1] <= '9'; keyword[1]++) 991 sfxge_vpd_try_add(sc, vpd_list, EFX_VPD_RO, keyword); 992 for (keyword[1] = 'A'; keyword[1] <= 'Z'; keyword[1]++) 993 sfxge_vpd_try_add(sc, vpd_list, EFX_VPD_RO, keyword); 994 995done: 996 return (0); 997 998fail2: 999 free(sc->vpd_data, M_SFXGE); 1000fail: 1001 return (rc); 1002} 1003 1004static void 1005sfxge_vpd_fini(struct sfxge_softc *sc) 1006{ 1007 free(sc->vpd_data, M_SFXGE); 1008} 1009 1010static void 1011sfxge_reset(void *arg, int npending) 1012{ 1013 struct sfxge_softc *sc; 1014 int rc; 1015 unsigned attempt; 1016 1017 (void)npending; 1018 1019 sc = (struct sfxge_softc *)arg; 1020 1021 SFXGE_ADAPTER_LOCK(sc); 1022 1023 if (sc->init_state != SFXGE_STARTED) 1024 goto done; 1025 1026 sfxge_stop(sc); 1027 efx_nic_reset(sc->enp);
|