Lines Matching refs:nic

80 #include "nic.h"
88 #define NICVF_CORE_LOCK_INIT(nic) \
89 sx_init(&(nic)->core_sx, device_get_nameunit((nic)->dev))
91 #define NICVF_CORE_LOCK_DESTROY(nic) \
92 sx_destroy(&(nic)->core_sx)
94 #define NICVF_CORE_LOCK(nic) sx_xlock(&(nic)->core_sx)
95 #define NICVF_CORE_UNLOCK(nic) sx_xunlock(&(nic)->core_sx)
97 #define NICVF_CORE_LOCK_ASSERT(nic) sx_assert(&(nic)->core_sx, SA_XLOCKED)
184 struct nicvf *nic;
190 nic = device_get_softc(dev);
191 nic->dev = dev;
192 nic->pnicvf = nic;
194 NICVF_CORE_LOCK_INIT(nic);
197 nic->hw_tso = TRUE;
200 nic->reg_base = bus_alloc_resource_any(dev, SYS_RES_MEMORY, &rid,
202 if (nic->reg_base == NULL) {
208 nic->max_queues = qcount;
210 err = nicvf_set_qset_resources(nic);
215 err = nicvf_allocate_misc_interrupt(nic);
219 NICVF_CORE_LOCK(nic);
220 err = nicvf_enable_misc_interrupt(nic);
221 NICVF_CORE_UNLOCK(nic);
225 err = nicvf_allocate_net_interrupts(nic);
233 if (memcmp(nic->hwaddr, zeromac, ETHER_ADDR_LEN) == 0) {
235 memcpy(nic->hwaddr, hwaddr, ETHER_ADDR_LEN);
236 NICVF_CORE_LOCK(nic);
237 nicvf_hw_set_mac_addr(nic, hwaddr);
238 NICVF_CORE_UNLOCK(nic);
242 nic->cpi_alg = CPI_ALG_NONE;
243 NICVF_CORE_LOCK(nic);
244 nicvf_config_cpi(nic);
246 if (nic->qs->rq_cnt > 1)
247 nicvf_rss_init(nic);
248 NICVF_CORE_UNLOCK(nic);
250 err = nicvf_setup_ifnet(nic);
256 err = nicvf_setup_ifmedia(nic);
262 mtx_init(&nic->stats_mtx, "VNIC stats", NULL, MTX_DEF);
263 callout_init_mtx(&nic->stats_callout, &nic->stats_mtx, 0);
265 ether_ifattach(nic->ifp, nic->hwaddr);
270 if_free(nic->ifp);
272 nicvf_release_all_interrupts(nic);
274 bus_release_resource(dev, SYS_RES_MEMORY, rman_get_rid(nic->reg_base),
275 nic->reg_base);
283 struct nicvf *nic;
285 nic = device_get_softc(dev);
287 NICVF_CORE_LOCK(nic);
289 nicvf_stop_locked(nic);
291 mtx_destroy(&nic->stats_mtx);
293 nicvf_release_all_interrupts(nic);
295 if (nic->reg_base != NULL) {
297 rman_get_rid(nic->reg_base), nic->reg_base);
301 ifmedia_removeall(&nic->if_media);
303 if_free(nic->ifp);
304 NICVF_CORE_UNLOCK(nic);
306 NICVF_CORE_LOCK_DESTROY(nic);
333 nicvf_setup_ifnet(struct nicvf *nic)
339 device_printf(nic->dev, "Could not allocate ifnet structure\n");
343 nic->ifp = ifp;
345 if_setsoftc(ifp, nic);
346 if_initname(ifp, device_get_name(nic->dev), device_get_unit(nic->dev));
363 if (nic->hw_tso) {
379 if (nic->hw_tso)
387 nicvf_setup_ifmedia(struct nicvf *nic)
390 ifmedia_init(&nic->if_media, IFM_IMASK, nicvf_media_change,
398 ifmedia_add(&nic->if_media, (IFM_ETHER | IFM_10_T | IFM_FDX),
400 ifmedia_add(&nic->if_media, (IFM_ETHER | IFM_100_TX | IFM_FDX),
402 ifmedia_add(&nic->if_media, (IFM_ETHER | IFM_1000_T | IFM_FDX),
404 ifmedia_add(&nic->if_media, (IFM_ETHER | IFM_10G_SR | IFM_FDX),
406 ifmedia_add(&nic->if_media, (IFM_ETHER | IFM_40G_CR4 | IFM_FDX),
408 ifmedia_add(&nic->if_media, (IFM_ETHER | IFM_AUTO | IFM_FDX),
411 ifmedia_set(&nic->if_media, (IFM_ETHER | IFM_AUTO | IFM_FDX));
419 struct nicvf *nic;
430 nic = if_getsoftc(ifp);
452 nicvf_if_init(nic);
468 NICVF_CORE_LOCK(nic);
469 err = nicvf_update_hw_max_frs(nic, ifr->ifr_mtu);
472 NICVF_CORE_UNLOCK(nic);
476 NICVF_CORE_LOCK(nic);
480 if ((flags ^ nic->if_flags) & IFF_PROMISC) {
483 nicvf_set_promiscous(nic);
487 if ((flags ^ nic->if_flags) & IFF_ALLMULTI) {
490 nicvf_set_multicast(nic);
494 nicvf_if_init_locked(nic);
497 nicvf_stop_locked(nic);
499 nic->if_flags = flags;
500 NICVF_CORE_UNLOCK(nic);
507 NICVF_CORE_LOCK(nic);
509 nicvf_set_multicast(nic);
510 NICVF_CORE_UNLOCK(nic);
517 err = ifmedia_ioctl(ifp, ifr, &nic->if_media, cmd);
530 if ((mask & IFCAP_TSO4) && nic->hw_tso)
537 NICVF_CORE_LOCK(nic);
539 if ((if_getdrvflags(nic->ifp) & IFF_DRV_RUNNING) != 0) {
548 rq_idx < nic->qs->rq_cnt; rq_idx++) {
549 rq = &nic->qs->rq[rq_idx];
553 NICVF_CORE_UNLOCK(nic);
567 nicvf_if_init_locked(struct nicvf *nic)
569 struct queue_set *qs = nic->qs;
575 NICVF_CORE_LOCK_ASSERT(nic);
576 ifp = nic->ifp;
579 nicvf_stop_locked(nic);
581 err = nicvf_enable_misc_interrupt(nic);
590 if (memcmp(nic->hwaddr, if_addr, ETHER_ADDR_LEN) != 0) {
591 memcpy(nic->hwaddr, if_addr, ETHER_ADDR_LEN);
592 nicvf_hw_set_mac_addr(nic, if_addr);
596 err = nicvf_init_resources(nic);
603 nicvf_reg_write(nic, NIC_VF_INT, ~0UL);
605 nicvf_enable_intr(nic, NICVF_INTR_QS_ERR, 0);
609 nicvf_enable_intr(nic, NICVF_INTR_CQ, qidx);
613 nicvf_enable_intr(nic, NICVF_INTR_RBDR, qidx);
615 nic->drv_stats.txq_stop = 0;
616 nic->drv_stats.txq_wake = 0;
622 callout_reset(&nic->stats_callout, hz, nicvf_tick_stats, nic);
634 struct nicvf *nic = if_softc;
636 NICVF_CORE_LOCK(nic);
637 nicvf_if_init_locked(nic);
638 NICVF_CORE_UNLOCK(nic);
644 struct nicvf *nic = if_getsoftc(ifp);
645 struct queue_set *qs = nic->qs;
653 device_get_nameunit(nic->dev));
678 IFF_DRV_RUNNING) || !nic->link_up || (err != 0)) {
700 struct nicvf *nic;
706 nic = if_getsoftc(ifp);
707 qs = nic->qs;
722 struct nicvf *nic;
726 nic = if_getsoftc(ifp);
727 hw_stats = &nic->hw_stats;
728 drv_stats = &nic->drv_stats;
756 struct nicvf *nic = if_getsoftc(ifp);
758 NICVF_CORE_LOCK(nic);
763 if (nic->link_up) {
768 switch (nic->speed) {
789 if (nic->duplex)
794 NICVF_CORE_UNLOCK(nic);
806 nicvf_reg_write(struct nicvf *nic, bus_space_handle_t offset, uint64_t val)
809 bus_write_8(nic->reg_base, offset, val);
813 nicvf_reg_read(struct nicvf *nic, uint64_t offset)
816 return (bus_read_8(nic->reg_base, offset));
820 nicvf_queue_reg_write(struct nicvf *nic, bus_space_handle_t offset,
824 bus_write_8(nic->reg_base, offset + (qidx << NIC_Q_NUM_SHIFT), val);
828 nicvf_queue_reg_read(struct nicvf *nic, bus_space_handle_t offset,
832 return (bus_read_8(nic->reg_base, offset + (qidx << NIC_Q_NUM_SHIFT)));
837 nicvf_write_to_mbx(struct nicvf *nic, union nic_mbx *mbx)
841 nicvf_reg_write(nic, NIC_VF_PF_MAILBOX_0_1 + 0, msg[0]);
842 nicvf_reg_write(nic, NIC_VF_PF_MAILBOX_0_1 + 8, msg[1]);
846 nicvf_send_msg_to_pf(struct nicvf *nic, union nic_mbx *mbx)
851 NICVF_CORE_LOCK_ASSERT(nic);
853 nic->pf_acked = FALSE;
854 nic->pf_nacked = FALSE;
856 nicvf_write_to_mbx(nic, mbx);
859 while (!nic->pf_acked) {
860 if (nic->pf_nacked)
865 if (nic->pf_acked)
869 device_printf(nic->dev,
871 (mbx->msg.msg & 0xFF), nic->vf_id);
884 nicvf_check_pf_ready(struct nicvf *nic)
889 if (nicvf_send_msg_to_pf(nic, &mbx)) {
890 device_printf(nic->dev,
899 nicvf_read_bgx_stats(struct nicvf *nic, struct bgx_stats_msg *bgx)
903 nic->bgx_stats.rx_stats[bgx->idx] = bgx->stats;
905 nic->bgx_stats.tx_stats[bgx->idx] = bgx->stats;
909 nicvf_handle_mbx_intr(struct nicvf *nic)
920 *mbx_data = nicvf_reg_read(nic, mbx_addr);
927 nic->pf_acked = TRUE;
928 nic->vf_id = mbx.nic_cfg.vf_id & 0x7F;
929 nic->tns_mode = mbx.nic_cfg.tns_mode & 0x7F;
930 nic->node = mbx.nic_cfg.node_id;
931 memcpy(nic->hwaddr, mbx.nic_cfg.mac_addr, ETHER_ADDR_LEN);
932 nic->loopback_supported = mbx.nic_cfg.loopback_supported;
933 nic->link_up = FALSE;
934 nic->duplex = 0;
935 nic->speed = 0;
938 nic->pf_acked = TRUE;
941 nic->pf_nacked = TRUE;
944 nic->rss_info.rss_size = mbx.rss_size.ind_tbl_size;
945 nic->pf_acked = TRUE;
948 nicvf_read_bgx_stats(nic, &mbx.bgx_stats);
949 nic->pf_acked = TRUE;
952 nic->pf_acked = TRUE;
953 nic->link_up = mbx.link_status.link_up;
954 nic->duplex = mbx.link_status.duplex;
955 nic->speed = mbx.link_status.speed;
956 if (nic->link_up) {
957 if_setbaudrate(nic->ifp, nic->speed * 1000000);
958 if_link_state_change(nic->ifp, LINK_STATE_UP);
960 if_setbaudrate(nic->ifp, 0);
961 if_link_state_change(nic->ifp, LINK_STATE_DOWN);
965 device_printf(nic->dev,
969 nicvf_clear_intr(nic, NICVF_INTR_MBOX, 0);
973 nicvf_update_hw_max_frs(struct nicvf *nic, int mtu)
979 mbx.frs.vf_id = nic->vf_id;
981 return nicvf_send_msg_to_pf(nic, &mbx);
985 nicvf_hw_set_mac_addr(struct nicvf *nic, uint8_t *hwaddr)
990 mbx.mac.vf_id = nic->vf_id;
993 return (nicvf_send_msg_to_pf(nic, &mbx));
997 nicvf_config_cpi(struct nicvf *nic)
1002 mbx.cpi_cfg.vf_id = nic->vf_id;
1003 mbx.cpi_cfg.cpi_alg = nic->cpi_alg;
1004 mbx.cpi_cfg.rq_cnt = nic->qs->rq_cnt;
1006 nicvf_send_msg_to_pf(nic, &mbx);
1010 nicvf_get_rss_size(struct nicvf *nic)
1015 mbx.rss_size.vf_id = nic->vf_id;
1016 nicvf_send_msg_to_pf(nic, &mbx);
1020 nicvf_config_rss(struct nicvf *nic)
1027 rss = &nic->rss_info;
1031 mbx.rss_cfg.vf_id = nic->vf_id;
1043 nicvf_send_msg_to_pf(nic, &mbx);
1050 nicvf_set_rss_key(struct nicvf *nic)
1056 rss = &nic->rss_info;
1060 nicvf_reg_write(nic, key_addr, rss->key[idx]);
1066 nicvf_rss_init(struct nicvf *nic)
1071 nicvf_get_rss_size(nic);
1073 rss = &nic->rss_info;
1074 if (nic->cpi_alg != CPI_ALG_NONE) {
1089 nicvf_set_rss_key(nic);
1092 nicvf_reg_write(nic, NIC_VNIC_RSS_CFG, rss->cfg);
1096 rss->ind_tbl[idx] = idx % nic->rx_queues;
1098 nicvf_config_rss(nic);
1104 nicvf_init_resources(struct nicvf *nic)
1112 nicvf_qset_config(nic, TRUE);
1115 err = nicvf_config_data_transfer(nic, TRUE);
1117 device_printf(nic->dev,
1123 nicvf_write_to_mbx(nic, &mbx);
1131 struct nicvf *nic = (struct nicvf *)arg;
1134 intr = nicvf_reg_read(nic, NIC_VF_INT);
1139 nicvf_handle_mbx_intr(nic);
1145 struct nicvf *nic;
1150 nic = cq->nic;
1154 nicvf_disable_intr(nic, NICVF_INTR_CQ, qidx);
1159 nicvf_clear_intr(nic, NICVF_INTR_CQ, qidx);
1167 struct nicvf *nic;
1172 nic = (struct nicvf *)arg;
1175 for (qidx = 0; qidx < nic->qs->rbdr_cnt; qidx++) {
1176 if (!nicvf_is_intr_enabled(nic, NICVF_INTR_RBDR, qidx))
1178 nicvf_disable_intr(nic, NICVF_INTR_RBDR, qidx);
1180 qs = nic->qs;
1184 nicvf_clear_intr(nic, NICVF_INTR_RBDR, qidx);
1191 struct nicvf *nic = (struct nicvf *)arg;
1192 struct queue_set *qs = nic->qs;
1195 nicvf_disable_intr(nic, NICVF_INTR_QS_ERR, 0);
1197 nicvf_clear_intr(nic, NICVF_INTR_QS_ERR, 0);
1202 nicvf_enable_msix(struct nicvf *nic)
1208 dinfo = device_get_ivars(nic->dev);
1210 nic->msix_table_res =
1211 bus_alloc_resource_any(nic->dev, SYS_RES_MEMORY, &rid, RF_ACTIVE);
1212 if (nic->msix_table_res == NULL) {
1213 device_printf(nic->dev,
1218 count = nic->num_vec = NIC_VF_MSIX_VECTORS;
1220 ret = pci_alloc_msix(nic->dev, &count);
1221 if ((ret != 0) || (count != nic->num_vec)) {
1222 device_printf(nic->dev,
1224 nic->num_vec, ret);
1228 nic->msix_enabled = 1;
1233 nicvf_disable_msix(struct nicvf *nic)
1236 if (nic->msix_enabled) {
1237 pci_release_msi(nic->dev);
1238 nic->msix_enabled = 0;
1239 nic->num_vec = 0;
1244 nicvf_release_all_interrupts(struct nicvf *nic)
1251 for (irq = 0; irq < nic->num_vec; irq++) {
1252 res = nic->msix_entries[irq].irq_res;
1256 if (nic->msix_entries[irq].handle != NULL) {
1257 err = bus_teardown_intr(nic->dev,
1258 nic->msix_entries[irq].irq_res,
1259 nic->msix_entries[irq].handle);
1262 nic->msix_entries[irq].handle = NULL;
1265 bus_release_resource(nic->dev, SYS_RES_IRQ,
1266 rman_get_rid(res), nic->msix_entries[irq].irq_res);
1267 nic->msix_entries[irq].irq_res = NULL;
1270 nicvf_disable_msix(nic);
1278 nicvf_allocate_misc_interrupt(struct nicvf *nic)
1285 if (nic->msix_enabled)
1289 if (nicvf_enable_msix(nic) != 0)
1294 nic->msix_entries[irq].irq_res = bus_alloc_resource_any(nic->dev,
1296 if (nic->msix_entries[irq].irq_res == NULL) {
1297 device_printf(nic->dev,
1299 device_get_unit(nic->dev));
1303 ret = bus_setup_intr(nic->dev, nic->msix_entries[irq].irq_res,
1304 (INTR_MPSAFE | INTR_TYPE_MISC), NULL, nicvf_misc_intr_handler, nic,
1305 &nic->msix_entries[irq].handle);
1307 res = nic->msix_entries[irq].irq_res;
1308 bus_release_resource(nic->dev, SYS_RES_IRQ,
1310 nic->msix_entries[irq].irq_res = NULL;
1318 nicvf_enable_misc_interrupt(struct nicvf *nic)
1322 nicvf_enable_intr(nic, NICVF_INTR_MBOX, 0);
1325 if (!nicvf_check_pf_ready(nic)) {
1326 nicvf_disable_intr(nic, NICVF_INTR_MBOX, 0);
1334 nicvf_release_net_interrupts(struct nicvf *nic)
1341 res = nic->msix_entries[irq].irq_res;
1345 if (nic->msix_entries[irq].handle != NULL) {
1346 err = bus_teardown_intr(nic->dev,
1347 nic->msix_entries[irq].irq_res,
1348 nic->msix_entries[irq].handle);
1357 bus_release_resource(nic->dev, SYS_RES_IRQ, rman_get_rid(res),
1359 nic->msix_entries[irq].irq_res = NULL;
1363 res = nic->msix_entries[irq].irq_res;
1367 if (nic->msix_entries[irq].handle != NULL) {
1368 err = bus_teardown_intr(nic->dev,
1369 nic->msix_entries[irq].irq_res,
1370 nic->msix_entries[irq].handle);
1379 bus_release_resource(nic->dev, SYS_RES_IRQ, rman_get_rid(res),
1381 nic->msix_entries[irq].irq_res = NULL;
1385 res = nic->msix_entries[irq].irq_res;
1388 if (nic->msix_entries[irq].handle != NULL) {
1389 err = bus_teardown_intr(nic->dev,
1390 nic->msix_entries[irq].irq_res,
1391 nic->msix_entries[irq].handle);
1400 bus_release_resource(nic->dev, SYS_RES_IRQ, rman_get_rid(res),
1402 nic->msix_entries[irq].irq_res = NULL;
1407 nicvf_allocate_net_interrupts(struct nicvf *nic)
1415 if (!nic->msix_enabled) {
1416 device_printf(nic->dev, "Cannot alloacte queue interrups. "
1423 if (irq >= (NICVF_INTR_ID_CQ + nic->qs->cq_cnt))
1428 nic->msix_entries[irq].irq_res = bus_alloc_resource_any(nic->dev,
1430 if (nic->msix_entries[irq].irq_res == NULL) {
1431 device_printf(nic->dev,
1433 (irq - NICVF_INTR_ID_CQ), device_get_unit(nic->dev));
1437 ret = bus_setup_intr(nic->dev, nic->msix_entries[irq].irq_res,
1439 NULL, &nic->qs->cq[qidx], &nic->msix_entries[irq].handle);
1441 device_printf(nic->dev,
1443 (irq - NICVF_INTR_ID_CQ), device_get_unit(nic->dev));
1446 cpuid = (device_get_unit(nic->dev) * CMP_QUEUE_CNT) + qidx;
1453 nic->qs->cq[qidx].cmp_cpuid = cpuid;
1455 device_printf(nic->dev, "bind CQ%d IRQ to CPU%d\n",
1459 bus_bind_intr(nic->dev, nic->msix_entries[irq].irq_res, cpuid);
1464 if (irq >= (NICVF_INTR_ID_RBDR + nic->qs->rbdr_cnt))
1468 nic->msix_entries[irq].irq_res = bus_alloc_resource_any(nic->dev,
1470 if (nic->msix_entries[irq].irq_res == NULL) {
1471 device_printf(nic->dev,
1474 device_get_unit(nic->dev));
1478 ret = bus_setup_intr(nic->dev, nic->msix_entries[irq].irq_res,
1480 nicvf_rbdr_intr_handler, nic,
1481 &nic->msix_entries[irq].handle);
1483 device_printf(nic->dev,
1486 device_get_unit(nic->dev));
1494 nic->msix_entries[irq].irq_res = bus_alloc_resource_any(nic->dev,
1496 if (nic->msix_entries[irq].irq_res == NULL) {
1497 device_printf(nic->dev,
1499 device_get_unit(nic->dev));
1503 ret = bus_setup_intr(nic->dev, nic->msix_entries[irq].irq_res,
1505 nic, &nic->msix_entries[irq].handle);
1507 device_printf(nic->dev,
1509 device_get_unit(nic->dev));
1515 nicvf_release_net_interrupts(nic);
1520 nicvf_stop_locked(struct nicvf *nic)
1524 struct queue_set *qs = nic->qs;
1527 NICVF_CORE_LOCK_ASSERT(nic);
1529 callout_drain(&nic->stats_callout);
1531 ifp = nic->ifp;
1534 nicvf_send_msg_to_pf(nic, &mbx);
1538 nicvf_disable_intr(nic, NICVF_INTR_RBDR, qidx);
1539 nicvf_clear_intr(nic, NICVF_INTR_RBDR, qidx);
1541 nicvf_disable_intr(nic, NICVF_INTR_QS_ERR, 0);
1542 nicvf_clear_intr(nic, NICVF_INTR_QS_ERR, 0);
1548 nicvf_config_data_transfer(nic, FALSE);
1551 nicvf_qset_config(nic, FALSE);
1554 nicvf_disable_intr(nic, NICVF_INTR_MBOX, 0);
1560 nicvf_update_stats(struct nicvf *nic)
1563 struct nicvf_hw_stats *stats = &nic->hw_stats;
1564 struct nicvf_drv_stats *drv_stats = &nic->drv_stats;
1565 struct queue_set *qs = nic->qs;
1568 nicvf_reg_read(nic, NIC_VNIC_RX_STAT_0_13 | ((reg) << 3))
1570 nicvf_reg_read(nic, NIC_VNIC_TX_STAT_0_4 | ((reg) << 3))
1600 nicvf_update_rq_stats(nic, qidx);
1602 nicvf_update_sq_stats(nic, qidx);
1608 struct nicvf *nic;
1610 nic = (struct nicvf *)arg;
1613 nicvf_update_stats(nic);
1615 callout_reset(&nic->stats_callout, hz, nicvf_tick_stats, nic);