if_cnmac.c revision 1.9
1/* $NetBSD: if_cnmac.c,v 1.9 2018/06/26 06:47:58 msaitoh Exp $ */ 2 3#include <sys/cdefs.h> 4#if 0 5__KERNEL_RCSID(0, "$NetBSD: if_cnmac.c,v 1.9 2018/06/26 06:47:58 msaitoh Exp $"); 6#endif 7 8#include "opt_octeon.h" 9 10#ifdef OCTEON_ETH_DEBUG 11 12#ifndef DIAGNOSTIC 13#define DIAGNOSTIC 14#endif 15 16#ifndef DEBUG 17#define DEBUG 18#endif 19 20#endif 21 22/* 23 * If no free send buffer is available, free all the sent buffer and bail out. 24 */ 25#define OCTEON_ETH_SEND_QUEUE_CHECK 26 27/* XXX XXX XXX XXX XXX XXX */ 28 29#include <sys/param.h> 30#include <sys/systm.h> 31#include <sys/pool.h> 32#include <sys/mbuf.h> 33#include <sys/malloc.h> 34#include <sys/kernel.h> 35#include <sys/socket.h> 36#include <sys/ioctl.h> 37#include <sys/errno.h> 38#include <sys/device.h> 39#include <sys/queue.h> 40#include <sys/conf.h> 41#include <sys/sysctl.h> 42#include <sys/syslog.h> 43 44#include <net/if.h> 45#include <net/if_dl.h> 46#include <net/if_media.h> 47#include <net/if_ether.h> 48#include <net/route.h> 49 50#include <net/bpf.h> 51 52#include <netinet/in.h> 53#include <netinet/in_systm.h> 54#include <netinet/in_var.h> 55#include <netinet/ip.h> 56 57#include <sys/bus.h> 58#include <machine/intr.h> 59#include <machine/endian.h> 60#include <machine/locore.h> 61 62#include <dev/mii/mii.h> 63#include <dev/mii/miivar.h> 64 65#include <mips/cpuregs.h> 66 67#include <mips/cavium/dev/octeon_asxreg.h> 68#include <mips/cavium/dev/octeon_ciureg.h> 69#include <mips/cavium/dev/octeon_npireg.h> 70#include <mips/cavium/dev/octeon_gmxreg.h> 71#include <mips/cavium/dev/octeon_ipdreg.h> 72#include <mips/cavium/dev/octeon_pipreg.h> 73#include <mips/cavium/dev/octeon_powreg.h> 74#include <mips/cavium/dev/octeon_faureg.h> 75#include <mips/cavium/dev/octeon_fpareg.h> 76#include <mips/cavium/dev/octeon_bootbusreg.h> 77#include <mips/cavium/include/iobusvar.h> 78#include <mips/cavium/octeonvar.h> 79#include <mips/cavium/dev/octeon_fpavar.h> 80#include <mips/cavium/dev/octeon_gmxvar.h> 81#include <mips/cavium/dev/octeon_fauvar.h> 82#include <mips/cavium/dev/octeon_powvar.h> 83#include <mips/cavium/dev/octeon_ipdvar.h> 84#include <mips/cavium/dev/octeon_pipvar.h> 85#include <mips/cavium/dev/octeon_pkovar.h> 86#include <mips/cavium/dev/octeon_asxvar.h> 87#include <mips/cavium/dev/octeon_smivar.h> 88#include <mips/cavium/dev/if_cnmacvar.h> 89 90#ifdef OCTEON_ETH_DEBUG 91#define OCTEON_ETH_KASSERT(x) KASSERT(x) 92#define OCTEON_ETH_KDASSERT(x) KDASSERT(x) 93#else 94#define OCTEON_ETH_KASSERT(x) 95#define OCTEON_ETH_KDASSERT(x) 96#endif 97 98/* 99 * Set the PKO to think command buffers are an odd length. This makes it so we 100 * never have to divide a comamnd across two buffers. 101 */ 102#define OCTEON_POOL_NWORDS_CMD \ 103 (((uint32_t)OCTEON_POOL_SIZE_CMD / sizeof(uint64_t)) - 1) 104#define FPA_COMMAND_BUFFER_POOL_NWORDS OCTEON_POOL_NWORDS_CMD /* XXX */ 105 106static void octeon_eth_buf_init(struct octeon_eth_softc *); 107 108static int octeon_eth_match(device_t, struct cfdata *, void *); 109static void octeon_eth_attach(device_t, device_t, void *); 110static void octeon_eth_pip_init(struct octeon_eth_softc *); 111static void octeon_eth_ipd_init(struct octeon_eth_softc *); 112static void octeon_eth_pko_init(struct octeon_eth_softc *); 113static void octeon_eth_asx_init(struct octeon_eth_softc *); 114static void octeon_eth_smi_init(struct octeon_eth_softc *); 115 116static void octeon_eth_board_mac_addr(uint8_t *, size_t, struct octeon_eth_softc *); 117 118static int octeon_eth_mii_readreg(device_t, int, int); 119static void octeon_eth_mii_writereg(device_t, int, int, int); 120static void octeon_eth_mii_statchg(struct ifnet *); 121 122static int octeon_eth_mediainit(struct octeon_eth_softc *); 123static void octeon_eth_mediastatus(struct ifnet *, struct ifmediareq *); 124static int octeon_eth_mediachange(struct ifnet *); 125 126static inline void octeon_eth_send_queue_flush_prefetch(struct octeon_eth_softc *); 127static inline void octeon_eth_send_queue_flush_fetch(struct octeon_eth_softc *); 128static inline void octeon_eth_send_queue_flush(struct octeon_eth_softc *); 129static inline void octeon_eth_send_queue_flush_sync(struct octeon_eth_softc *); 130static inline int octeon_eth_send_queue_is_full(struct octeon_eth_softc *); 131static inline void octeon_eth_send_queue_add(struct octeon_eth_softc *, 132 struct mbuf *, uint64_t *); 133static inline void octeon_eth_send_queue_del(struct octeon_eth_softc *, 134 struct mbuf **, uint64_t **); 135static inline int octeon_eth_buf_free_work(struct octeon_eth_softc *, 136 uint64_t *, uint64_t); 137static inline void octeon_eth_buf_ext_free_m(struct mbuf *, void *, size_t, void *); 138static inline void octeon_eth_buf_ext_free_ext(struct mbuf *, void *, size_t, void *); 139 140static int octeon_eth_ioctl(struct ifnet *, u_long, void *); 141static void octeon_eth_watchdog(struct ifnet *); 142static int octeon_eth_init(struct ifnet *); 143static void octeon_eth_stop(struct ifnet *, int); 144static void octeon_eth_start(struct ifnet *); 145 146static inline int octeon_eth_send_cmd(struct octeon_eth_softc *, uint64_t, 147 uint64_t, int *); 148static inline uint64_t octeon_eth_send_makecmd_w1(int, paddr_t); 149static inline uint64_t octeon_eth_send_makecmd_w0(uint64_t, uint64_t, size_t, 150 int); 151static inline int octeon_eth_send_makecmd_gbuf(struct octeon_eth_softc *, 152 struct mbuf *, uint64_t *, int *); 153static inline int octeon_eth_send_makecmd(struct octeon_eth_softc *, 154 struct mbuf *, uint64_t *, uint64_t *, uint64_t *); 155static inline int octeon_eth_send_buf(struct octeon_eth_softc *, 156 struct mbuf *, uint64_t *, int *); 157static inline int octeon_eth_send(struct octeon_eth_softc *, 158 struct mbuf *, int *); 159 160static int octeon_eth_reset(struct octeon_eth_softc *); 161static int octeon_eth_configure(struct octeon_eth_softc *); 162static int octeon_eth_configure_common(struct octeon_eth_softc *); 163 164static void octeon_eth_tick_free(void *arg); 165static void octeon_eth_tick_misc(void *); 166 167static inline int octeon_eth_recv_mbuf(struct octeon_eth_softc *, 168 uint64_t *, struct mbuf **); 169static inline int octeon_eth_recv_check_code(struct octeon_eth_softc *, 170 uint64_t); 171static inline int octeon_eth_recv_check_jumbo(struct octeon_eth_softc *, 172 uint64_t); 173static inline int octeon_eth_recv_check_link(struct octeon_eth_softc *, 174 uint64_t); 175static inline int octeon_eth_recv_check(struct octeon_eth_softc *, 176 uint64_t); 177static inline int octeon_eth_recv(struct octeon_eth_softc *, uint64_t *); 178static void octeon_eth_recv_redir(struct ifnet *, struct mbuf *); 179static inline void octeon_eth_recv_intr(void *, uint64_t *); 180 181/* device driver context */ 182static struct octeon_eth_softc *octeon_eth_gsc[GMX_PORT_NUNITS]; 183static void *octeon_eth_pow_recv_ih; 184 185/* sysctl'able parameters */ 186int octeon_eth_param_pko_cmd_w0_n2 = 1; 187int octeon_eth_param_pip_dyn_rs = 1; 188int octeon_eth_param_redir = 0; 189int octeon_eth_param_pktbuf = 0; 190int octeon_eth_param_rate = 0; 191int octeon_eth_param_intr = 0; 192 193CFATTACH_DECL_NEW(cnmac, sizeof(struct octeon_eth_softc), 194 octeon_eth_match, octeon_eth_attach, NULL, NULL); 195 196#ifdef OCTEON_ETH_DEBUG 197 198static const struct octeon_evcnt_entry octeon_evcnt_entries[] = { 199#define _ENTRY(name, type, parent, descr) \ 200 OCTEON_EVCNT_ENTRY(struct octeon_eth_softc, name, type, parent, descr) 201 _ENTRY(rx, MISC, NULL, "rx"), 202 _ENTRY(rxint, INTR, NULL, "rx intr"), 203 _ENTRY(rxrs, MISC, NULL, "rx dynamic short"), 204 _ENTRY(rxbufpkalloc, MISC, NULL, "rx buf pkt alloc"), 205 _ENTRY(rxbufpkput, MISC, NULL, "rx buf pkt put"), 206 _ENTRY(rxbufwqalloc, MISC, NULL, "rx buf wqe alloc"), 207 _ENTRY(rxbufwqput, MISC, NULL, "rx buf wqe put"), 208 _ENTRY(rxerrcode, MISC, NULL, "rx code error"), 209 _ENTRY(rxerrfix, MISC, NULL, "rx fixup error"), 210 _ENTRY(rxerrjmb, MISC, NULL, "rx jmb error"), 211 _ENTRY(rxerrlink, MISC, NULL, "rx link error"), 212 _ENTRY(rxerroff, MISC, NULL, "rx offload error"), 213 _ENTRY(rxonperrshort, MISC, NULL, "rx onp fixup short error"), 214 _ENTRY(rxonperrpreamble, MISC, NULL, "rx onp fixup preamble error"), 215 _ENTRY(rxonperrcrc, MISC, NULL, "rx onp fixup crc error"), 216 _ENTRY(rxonperraddress, MISC, NULL, "rx onp fixup address error"), 217 _ENTRY(rxonponp, MISC, NULL, "rx onp fixup onp packets"), 218 _ENTRY(rxonpok, MISC, NULL, "rx onp fixup success packets"), 219 _ENTRY(tx, MISC, NULL, "tx"), 220 _ENTRY(txadd, MISC, NULL, "tx add"), 221 _ENTRY(txbufcballoc, MISC, NULL, "tx buf cb alloc"), 222 _ENTRY(txbufcbget, MISC, NULL, "tx buf cb get"), 223 _ENTRY(txbufgballoc, MISC, NULL, "tx buf gb alloc"), 224 _ENTRY(txbufgbget, MISC, NULL, "tx buf gb get"), 225 _ENTRY(txbufgbput, MISC, NULL, "tx buf gb put"), 226 _ENTRY(txdel, MISC, NULL, "tx del"), 227 _ENTRY(txerr, MISC, NULL, "tx error"), 228 _ENTRY(txerrcmd, MISC, NULL, "tx cmd error"), 229 _ENTRY(txerrgbuf, MISC, NULL, "tx gbuf error"), 230 _ENTRY(txerrlink, MISC, NULL, "tx link error"), 231 _ENTRY(txerrmkcmd, MISC, NULL, "tx makecmd error"), 232#undef _ENTRY 233}; 234#endif 235 236/* ---- buffer management */ 237 238static const struct octeon_eth_pool_param { 239 int poolno; 240 size_t size; 241 size_t nelems; 242} octeon_eth_pool_params[] = { 243#define _ENTRY(x) { OCTEON_POOL_NO_##x, OCTEON_POOL_SIZE_##x, OCTEON_POOL_NELEMS_##x } 244 _ENTRY(PKT), 245 _ENTRY(WQE), 246 _ENTRY(CMD), 247 _ENTRY(SG) 248#undef _ENTRY 249}; 250struct octeon_fpa_buf *octeon_eth_pools[8/* XXX */]; 251#define octeon_eth_fb_pkt octeon_eth_pools[OCTEON_POOL_NO_PKT] 252#define octeon_eth_fb_wqe octeon_eth_pools[OCTEON_POOL_NO_WQE] 253#define octeon_eth_fb_cmd octeon_eth_pools[OCTEON_POOL_NO_CMD] 254#define octeon_eth_fb_sg octeon_eth_pools[OCTEON_POOL_NO_SG] 255 256static void 257octeon_eth_buf_init(struct octeon_eth_softc *sc) 258{ 259 static int once; 260 int i; 261 const struct octeon_eth_pool_param *pp; 262 struct octeon_fpa_buf *fb; 263 264 if (once == 1) 265 return; 266 once = 1; 267 268 for (i = 0; i < (int)__arraycount(octeon_eth_pool_params); i++) { 269 pp = &octeon_eth_pool_params[i]; 270 octeon_fpa_buf_init(pp->poolno, pp->size, pp->nelems, &fb); 271 octeon_eth_pools[i] = fb; 272 } 273} 274 275/* ---- autoconf */ 276 277static int 278octeon_eth_match(device_t parent, struct cfdata *match, void *aux) 279{ 280 struct octeon_gmx_attach_args *ga = aux; 281 282 if (strcmp(match->cf_name, ga->ga_name) != 0) { 283 return 0; 284 } 285 return 1; 286} 287 288static void 289octeon_eth_attach(device_t parent, device_t self, void *aux) 290{ 291 struct octeon_eth_softc *sc = device_private(self); 292 struct octeon_gmx_attach_args *ga = aux; 293 struct ifnet *ifp = &sc->sc_ethercom.ec_if; 294 uint8_t enaddr[ETHER_ADDR_LEN]; 295 296 sc->sc_dev = self; 297 sc->sc_regt = ga->ga_regt; 298 sc->sc_port = ga->ga_portno; 299 sc->sc_port_type = ga->ga_port_type; 300 sc->sc_gmx = ga->ga_gmx; 301 sc->sc_gmx_port = ga->ga_gmx_port; 302 303 sc->sc_init_flag = 0; 304 /* 305 * XXXUEBAYASI 306 * Setting PIP_IP_OFFSET[OFFSET] to 8 causes panic ... why??? 307 */ 308 sc->sc_ip_offset = 0/* XXX */; 309 310 if (MIPS_PRID_IMPL(mips_options.mips_cpu_id) <= MIPS_CN30XX) { 311 SET(sc->sc_quirks, OCTEON_ETH_QUIRKS_NO_PRE_ALIGN); 312 SET(sc->sc_quirks, OCTEON_ETH_QUIRKS_NO_RX_INBND); 313 } 314 315 octeon_eth_board_mac_addr(enaddr, sizeof(enaddr), sc); 316 printf("%s: Ethernet address %s\n", device_xname(sc->sc_dev), 317 ether_sprintf(enaddr)); 318 319 octeon_eth_gsc[sc->sc_port] = sc; 320 321 SIMPLEQ_INIT(&sc->sc_sendq); 322 sc->sc_soft_req_thresh = 15/* XXX */; 323 sc->sc_ext_callback_cnt = 0; 324 325 octeon_gmx_stats_init(sc->sc_gmx_port); 326 327 callout_init(&sc->sc_tick_misc_ch, 0); 328 callout_init(&sc->sc_tick_free_ch, 0); 329 330 octeon_fau_op_init(&sc->sc_fau_done, 331 OCTEON_CVMSEG_ETHER_OFFSET(sc->sc_port, csm_ether_fau_done), 332 OCT_FAU_REG_ADDR_END - (8 * (sc->sc_port + 1))/* XXX */); 333 octeon_fau_op_set_8(&sc->sc_fau_done, 0); 334 335 octeon_eth_pip_init(sc); 336 octeon_eth_ipd_init(sc); 337 octeon_eth_pko_init(sc); 338 octeon_eth_asx_init(sc); 339 octeon_eth_smi_init(sc); 340 341 sc->sc_gmx_port->sc_ipd = sc->sc_ipd; 342 sc->sc_gmx_port->sc_port_asx = sc->sc_asx; 343 sc->sc_gmx_port->sc_port_mii = &sc->sc_mii; 344 sc->sc_gmx_port->sc_port_ec = &sc->sc_ethercom; 345 /* XXX */ 346 sc->sc_gmx_port->sc_quirks = sc->sc_quirks; 347 348 /* XXX */ 349 sc->sc_pow = &octeon_pow_softc; 350 351 octeon_eth_mediainit(sc); 352 353 strncpy(ifp->if_xname, device_xname(sc->sc_dev), sizeof(ifp->if_xname)); 354 ifp->if_softc = sc; 355 ifp->if_flags = IFF_BROADCAST | IFF_SIMPLEX | IFF_MULTICAST; 356 ifp->if_ioctl = octeon_eth_ioctl; 357 ifp->if_start = octeon_eth_start; 358 ifp->if_watchdog = octeon_eth_watchdog; 359 ifp->if_init = octeon_eth_init; 360 ifp->if_stop = octeon_eth_stop; 361 IFQ_SET_MAXLEN(&ifp->if_snd, max(GATHER_QUEUE_SIZE, IFQ_MAXLEN)); 362 IFQ_SET_READY(&ifp->if_snd); 363 364 /* XXX: not yet tx checksum */ 365 ifp->if_capabilities = 366 IFCAP_CSUM_IPv4_Rx | IFCAP_CSUM_TCPv4_Rx | IFCAP_CSUM_UDPv4_Rx | 367 IFCAP_CSUM_TCPv6_Rx | IFCAP_CSUM_UDPv6_Rx; 368 369 /* 802.1Q VLAN-sized frames are supported */ 370 sc->sc_ethercom.ec_capabilities |= ETHERCAP_VLAN_MTU; 371 372 octeon_gmx_set_mac_addr(sc->sc_gmx_port, enaddr); 373 octeon_gmx_set_filter(sc->sc_gmx_port); 374 375 if_attach(ifp); 376 ether_ifattach(ifp, enaddr); 377 378 /* XXX */ 379 sc->sc_rate_recv_check_link_cap.tv_sec = 1; 380 sc->sc_rate_recv_check_jumbo_cap.tv_sec = 1; 381 sc->sc_rate_recv_check_code_cap.tv_sec = 1; 382 sc->sc_rate_recv_fixup_odd_nibble_short_cap.tv_sec = 1; 383 sc->sc_rate_recv_fixup_odd_nibble_preamble_cap.tv_sec = 1; 384 sc->sc_rate_recv_fixup_odd_nibble_crc_cap.tv_sec = 1; 385#ifdef OCTEON_ETH_DEBUG 386 sc->sc_rate_recv_fixup_odd_nibble_addr_cap.tv_sec = 1; 387#endif 388 /* XXX */ 389 390#if 1 391 octeon_eth_buf_init(sc); 392#endif 393 394 if (octeon_eth_pow_recv_ih == NULL) 395 octeon_eth_pow_recv_ih = octeon_pow_intr_establish(OCTEON_POW_GROUP_PIP, 396 IPL_NET, octeon_eth_recv_intr, NULL, NULL); 397 398 OCTEON_EVCNT_ATTACH_EVCNTS(sc, octeon_evcnt_entries, 399 device_xname(sc->sc_dev)); 400} 401 402/* ---- submodules */ 403 404/* XXX */ 405static void 406octeon_eth_pip_init(struct octeon_eth_softc *sc) 407{ 408 struct octeon_pip_attach_args pip_aa; 409 410 pip_aa.aa_port = sc->sc_port; 411 pip_aa.aa_regt = sc->sc_regt; 412 pip_aa.aa_tag_type = POW_TAG_TYPE_ORDERED/* XXX */; 413 pip_aa.aa_receive_group = OCTEON_POW_GROUP_PIP; 414 pip_aa.aa_ip_offset = sc->sc_ip_offset; 415 octeon_pip_init(&pip_aa, &sc->sc_pip); 416} 417 418/* XXX */ 419static void 420octeon_eth_ipd_init(struct octeon_eth_softc *sc) 421{ 422 struct octeon_ipd_attach_args ipd_aa; 423 424 ipd_aa.aa_port = sc->sc_port; 425 ipd_aa.aa_regt = sc->sc_regt; 426 ipd_aa.aa_first_mbuff_skip = 184/* XXX */; 427 ipd_aa.aa_not_first_mbuff_skip = 0/* XXX */; 428 octeon_ipd_init(&ipd_aa, &sc->sc_ipd); 429} 430 431/* XXX */ 432static void 433octeon_eth_pko_init(struct octeon_eth_softc *sc) 434{ 435 struct octeon_pko_attach_args pko_aa; 436 437 pko_aa.aa_port = sc->sc_port; 438 pko_aa.aa_regt = sc->sc_regt; 439 pko_aa.aa_cmdptr = &sc->sc_cmdptr; 440 pko_aa.aa_cmd_buf_pool = OCTEON_POOL_NO_CMD; 441 pko_aa.aa_cmd_buf_size = OCTEON_POOL_NWORDS_CMD; 442 octeon_pko_init(&pko_aa, &sc->sc_pko); 443} 444 445/* XXX */ 446static void 447octeon_eth_asx_init(struct octeon_eth_softc *sc) 448{ 449 struct octeon_asx_attach_args asx_aa; 450 451 asx_aa.aa_port = sc->sc_port; 452 asx_aa.aa_regt = sc->sc_regt; 453 octeon_asx_init(&asx_aa, &sc->sc_asx); 454} 455 456static void 457octeon_eth_smi_init(struct octeon_eth_softc *sc) 458{ 459 struct octeon_smi_attach_args smi_aa; 460 461 smi_aa.aa_port = sc->sc_port; 462 smi_aa.aa_regt = sc->sc_regt; 463 octeon_smi_init(&smi_aa, &sc->sc_smi); 464 octeon_smi_set_clock(sc->sc_smi, 0x1464ULL); /* XXX */ 465} 466 467/* ---- XXX */ 468 469#define ADDR2UINT64(u, a) \ 470 do { \ 471 u = \ 472 (((uint64_t)a[0] << 40) | ((uint64_t)a[1] << 32) | \ 473 ((uint64_t)a[2] << 24) | ((uint64_t)a[3] << 16) | \ 474 ((uint64_t)a[4] << 8) | ((uint64_t)a[5] << 0)); \ 475 } while (0) 476#define UINT642ADDR(a, u) \ 477 do { \ 478 a[0] = (uint8_t)((u) >> 40); a[1] = (uint8_t)((u) >> 32); \ 479 a[2] = (uint8_t)((u) >> 24); a[3] = (uint8_t)((u) >> 16); \ 480 a[4] = (uint8_t)((u) >> 8); a[5] = (uint8_t)((u) >> 0); \ 481 } while (0) 482 483static void 484octeon_eth_board_mac_addr(uint8_t *enaddr, size_t size, struct octeon_eth_softc *sc) 485{ 486 prop_dictionary_t dict; 487 prop_data_t ea; 488 489 dict = device_properties(sc->sc_dev); 490 KASSERT(dict != NULL); 491 ea = prop_dictionary_get(dict, "mac-address"); 492 KASSERT(ea != NULL); 493 memcpy(enaddr, prop_data_data_nocopy(ea), size); 494} 495 496/* ---- media */ 497 498static int 499octeon_eth_mii_readreg(device_t self, int phy_addr, int reg) 500{ 501 struct octeon_eth_softc *sc = device_private(self); 502 503 return octeon_smi_read(sc->sc_smi, phy_addr, reg); 504} 505 506static void 507octeon_eth_mii_writereg(device_t self, int phy_addr, int reg, int value) 508{ 509 struct octeon_eth_softc *sc = device_private(self); 510 511 octeon_smi_write(sc->sc_smi, phy_addr, reg, value); 512} 513 514static void 515octeon_eth_mii_statchg(struct ifnet *ifp) 516{ 517 struct octeon_eth_softc *sc = ifp->if_softc; 518 519 octeon_pko_port_enable(sc->sc_pko, 0); 520 octeon_gmx_port_enable(sc->sc_gmx_port, 0); 521 522 octeon_eth_reset(sc); 523 524 if (ISSET(ifp->if_flags, IFF_RUNNING)) 525 octeon_gmx_set_filter(sc->sc_gmx_port); 526 527 octeon_pko_port_enable(sc->sc_pko, 1); 528 octeon_gmx_port_enable(sc->sc_gmx_port, 1); 529} 530 531static int 532octeon_eth_mediainit(struct octeon_eth_softc *sc) 533{ 534 struct ifnet *ifp = &sc->sc_ethercom.ec_if; 535 prop_object_t phy; 536 537 sc->sc_mii.mii_ifp = ifp; 538 sc->sc_mii.mii_readreg = octeon_eth_mii_readreg; 539 sc->sc_mii.mii_writereg = octeon_eth_mii_writereg; 540 sc->sc_mii.mii_statchg = octeon_eth_mii_statchg; 541 ifmedia_init(&sc->sc_mii.mii_media, 0, octeon_eth_mediachange, 542 octeon_eth_mediastatus); 543 544 phy = prop_dictionary_get(device_properties(sc->sc_dev), "phy-addr"); 545 KASSERT(phy != NULL); 546 547 mii_attach(sc->sc_dev, &sc->sc_mii, 548 0xffffffff, prop_number_integer_value(phy), 549 MII_OFFSET_ANY, MIIF_DOPAUSE); 550 551 /* XXX XXX XXX */ 552 if (LIST_FIRST(&sc->sc_mii.mii_phys) != NULL) { 553 /* XXX XXX XXX */ 554 ifmedia_set(&sc->sc_mii.mii_media, IFM_ETHER | IFM_AUTO); 555 /* XXX XXX XXX */ 556 } else { 557 /* XXX XXX XXX */ 558 ifmedia_add(&sc->sc_mii.mii_media, IFM_ETHER | IFM_NONE, 559 MII_MEDIA_NONE, NULL); 560 /* XXX XXX XXX */ 561 ifmedia_set(&sc->sc_mii.mii_media, IFM_ETHER | IFM_NONE); 562 /* XXX XXX XXX */ 563 } 564 /* XXX XXX XXX */ 565 566 return 0; 567} 568 569static void 570octeon_eth_mediastatus(struct ifnet *ifp, struct ifmediareq *ifmr) 571{ 572 struct octeon_eth_softc *sc = ifp->if_softc; 573 574 mii_pollstat(&sc->sc_mii); 575 576 ifmr->ifm_status = sc->sc_mii.mii_media_status; 577 ifmr->ifm_active = sc->sc_mii.mii_media_active; 578 ifmr->ifm_active = (sc->sc_mii.mii_media_active & ~IFM_ETH_FMASK) | 579 sc->sc_gmx_port->sc_port_flowflags; 580} 581 582static int 583octeon_eth_mediachange(struct ifnet *ifp) 584{ 585 struct octeon_eth_softc *sc = ifp->if_softc; 586 587 mii_mediachg(&sc->sc_mii); 588 589 return 0; 590} 591 592/* ---- send buffer garbage collection */ 593 594static inline void 595octeon_eth_send_queue_flush_prefetch(struct octeon_eth_softc *sc) 596{ 597 OCTEON_ETH_KASSERT(sc->sc_prefetch == 0); 598 octeon_fau_op_inc_fetch_8(&sc->sc_fau_done, 0); 599 sc->sc_prefetch = 1; 600} 601 602static inline void 603octeon_eth_send_queue_flush_fetch(struct octeon_eth_softc *sc) 604{ 605#ifndef OCTEON_ETH_DEBUG 606 if (!sc->sc_prefetch) 607 return; 608#endif 609 OCTEON_ETH_KASSERT(sc->sc_prefetch == 1); 610 sc->sc_hard_done_cnt = octeon_fau_op_inc_read_8(&sc->sc_fau_done); 611 OCTEON_ETH_KASSERT(sc->sc_hard_done_cnt <= 0); 612 sc->sc_prefetch = 0; 613} 614 615static inline void 616octeon_eth_send_queue_flush(struct octeon_eth_softc *sc) 617{ 618 struct ifnet *ifp = &sc->sc_ethercom.ec_if; 619 const int64_t sent_count = sc->sc_hard_done_cnt; 620 int i; 621 622 OCTEON_ETH_KASSERT(sc->sc_flush == 0); 623 OCTEON_ETH_KASSERT(sent_count <= 0); 624 625 for (i = 0; i < 0 - sent_count; i++) { 626 struct mbuf *m; 627 uint64_t *gbuf; 628 629 octeon_eth_send_queue_del(sc, &m, &gbuf); 630 631 octeon_fpa_buf_put(octeon_eth_fb_sg, gbuf); 632 OCTEON_EVCNT_INC(sc, txbufgbput); 633 634 m_freem(m); 635 636 CLR(ifp->if_flags, IFF_OACTIVE); 637 } 638 639 octeon_fau_op_inc_fetch_8(&sc->sc_fau_done, i); 640 sc->sc_flush = i; 641} 642 643static inline void 644octeon_eth_send_queue_flush_sync(struct octeon_eth_softc *sc) 645{ 646 if (sc->sc_flush == 0) 647 return; 648 649 OCTEON_ETH_KASSERT(sc->sc_flush > 0); 650 651 /* XXX XXX XXX */ 652 octeon_fau_op_inc_read_8(&sc->sc_fau_done); 653 sc->sc_soft_req_cnt -= sc->sc_flush; 654 OCTEON_ETH_KASSERT(sc->sc_soft_req_cnt >= 0); 655 /* XXX XXX XXX */ 656 657 sc->sc_flush = 0; 658} 659 660static inline int 661octeon_eth_send_queue_is_full(struct octeon_eth_softc *sc) 662{ 663#ifdef OCTEON_ETH_SEND_QUEUE_CHECK 664 int64_t nofree_cnt; 665 666 nofree_cnt = sc->sc_soft_req_cnt + sc->sc_hard_done_cnt; 667 668 if (__predict_false(nofree_cnt == GATHER_QUEUE_SIZE - 1)) { 669 octeon_eth_send_queue_flush(sc); 670 OCTEON_EVCNT_INC(sc, txerrgbuf); 671 octeon_eth_send_queue_flush_sync(sc); 672 return 1; 673 } 674 675#endif 676 return 0; 677} 678 679/* 680 * (Ab)use m_nextpkt and m_paddr to maintain mbuf chain and pointer to gather 681 * buffer. Other mbuf members may be used by m_freem(), so don't touch them! 682 */ 683 684struct _send_queue_entry { 685 union { 686 struct mbuf _sqe_s_mbuf; 687 struct { 688 char _sqe_s_entry_pad[offsetof(struct mbuf, m_nextpkt)]; 689 SIMPLEQ_ENTRY(_send_queue_entry) _sqe_s_entry_entry; 690 } _sqe_s_entry; 691 struct { 692 char _sqe_s_gbuf_pad[offsetof(struct mbuf, m_paddr)]; 693 uint64_t *_sqe_s_gbuf_gbuf; 694 } _sqe_s_gbuf; 695 } _sqe_u; 696#define _sqe_entry _sqe_u._sqe_s_entry._sqe_s_entry_entry 697#define _sqe_gbuf _sqe_u._sqe_s_gbuf._sqe_s_gbuf_gbuf 698}; 699 700static inline void 701octeon_eth_send_queue_add(struct octeon_eth_softc *sc, struct mbuf *m, 702 uint64_t *gbuf) 703{ 704 struct _send_queue_entry *sqe = (struct _send_queue_entry *)m; 705 706 sqe->_sqe_gbuf = gbuf; 707 SIMPLEQ_INSERT_TAIL(&sc->sc_sendq, sqe, _sqe_entry); 708 709 if ((m->m_flags & M_EXT) && m->m_ext.ext_free != NULL) 710 sc->sc_ext_callback_cnt++; 711 712 OCTEON_EVCNT_INC(sc, txadd); 713} 714 715static inline void 716octeon_eth_send_queue_del(struct octeon_eth_softc *sc, struct mbuf **rm, 717 uint64_t **rgbuf) 718{ 719 struct _send_queue_entry *sqe; 720 721 sqe = SIMPLEQ_FIRST(&sc->sc_sendq); 722 OCTEON_ETH_KASSERT(sqe != NULL); 723 SIMPLEQ_REMOVE_HEAD(&sc->sc_sendq, _sqe_entry); 724 725 *rm = (void *)sqe; 726 *rgbuf = sqe->_sqe_gbuf; 727 728 if (((*rm)->m_flags & M_EXT) && (*rm)->m_ext.ext_free != NULL) { 729 sc->sc_ext_callback_cnt--; 730 OCTEON_ETH_KASSERT(sc->sc_ext_callback_cnt >= 0); 731 } 732 733 OCTEON_EVCNT_INC(sc, txdel); 734} 735 736static inline int 737octeon_eth_buf_free_work(struct octeon_eth_softc *sc, uint64_t *work, 738 uint64_t word2) 739{ 740 /* XXX when jumbo frame */ 741 if (ISSET(word2, PIP_WQE_WORD2_IP_BUFS)) { 742 paddr_t addr; 743 paddr_t start_buffer; 744 745 addr = work[3] & PIP_WQE_WORD3_ADDR; 746 start_buffer = addr & ~(2048 - 1); 747 748 octeon_fpa_buf_put_paddr(octeon_eth_fb_pkt, start_buffer); 749 OCTEON_EVCNT_INC(sc, rxbufpkput); 750 } 751 752 octeon_fpa_buf_put(octeon_eth_fb_wqe, work); 753 OCTEON_EVCNT_INC(sc, rxbufwqput); 754 755 return 0; 756} 757 758static inline void 759octeon_eth_buf_ext_free_m(struct mbuf *m, void *buf, size_t size, void *arg) 760{ 761 uint64_t *work = (void *)arg; 762#ifdef OCTEON_ETH_DEBUG 763 struct octeon_eth_softc *sc = (void *)(uintptr_t)work[0]; 764#endif 765 int s = splnet(); 766 767 OCTEON_EVCNT_INC(sc, rxrs); 768 769 octeon_fpa_buf_put(octeon_eth_fb_wqe, work); 770 OCTEON_EVCNT_INC(sc, rxbufwqput); 771 772 OCTEON_ETH_KASSERT(m != NULL); 773 774 pool_cache_put(mb_cache, m); 775 776 splx(s); 777} 778 779static inline void 780octeon_eth_buf_ext_free_ext(struct mbuf *m, void *buf, size_t size, 781 void *arg) 782{ 783 uint64_t *work = (void *)arg; 784#ifdef OCTEON_ETH_DEBUG 785 struct octeon_eth_softc *sc = (void *)(uintptr_t)work[0]; 786#endif 787 int s = splnet(); 788 789 octeon_fpa_buf_put(octeon_eth_fb_wqe, work); 790 OCTEON_EVCNT_INC(sc, rxbufwqput); 791 792 octeon_fpa_buf_put(octeon_eth_fb_pkt, buf); 793 OCTEON_EVCNT_INC(sc, rxbufpkput); 794 795 OCTEON_ETH_KASSERT(m != NULL); 796 797 pool_cache_put(mb_cache, m); 798 799 splx(s); 800} 801 802/* ---- ifnet interfaces */ 803 804static int 805octeon_eth_ioctl(struct ifnet *ifp, u_long cmd, void *data) 806{ 807 struct octeon_eth_softc *sc = ifp->if_softc; 808 struct ifreq *ifr = (struct ifreq *)data; 809 int s, error; 810 811 s = splnet(); 812 switch (cmd) { 813 case SIOCSIFMEDIA: 814 /* Flow control requires full-duplex mode. */ 815 if (IFM_SUBTYPE(ifr->ifr_media) == IFM_AUTO || 816 (ifr->ifr_media & IFM_FDX) == 0) { 817 ifr->ifr_media &= ~IFM_ETH_FMASK; 818 } 819 if (IFM_SUBTYPE(ifr->ifr_media) != IFM_AUTO) { 820 if ((ifr->ifr_media & IFM_ETH_FMASK) == IFM_FLOW) { 821 ifr->ifr_media |= 822 IFM_ETH_TXPAUSE | IFM_ETH_RXPAUSE; 823 } 824 sc->sc_gmx_port->sc_port_flowflags = 825 ifr->ifr_media & IFM_ETH_FMASK; 826 } 827 /* FALLTHROUGH */ 828 case SIOCGIFMEDIA: 829 /* XXX: Flow contorol */ 830 error = ifmedia_ioctl(ifp, ifr, &sc->sc_mii.mii_media, cmd); 831 break; 832 default: 833 error = ether_ioctl(ifp, cmd, data); 834 if (error == ENETRESET) { 835 /* 836 * Multicast list has changed; set the hardware filter 837 * accordingly. 838 */ 839 if (ISSET(ifp->if_flags, IFF_RUNNING)) 840 octeon_gmx_set_filter(sc->sc_gmx_port); 841 error = 0; 842 } 843 break; 844 } 845 octeon_eth_start(ifp); 846 splx(s); 847 848 return (error); 849} 850 851/* ---- send (output) */ 852 853static inline uint64_t 854octeon_eth_send_makecmd_w0(uint64_t fau0, uint64_t fau1, size_t len, int segs) 855{ 856 return octeon_pko_cmd_word0( 857 OCT_FAU_OP_SIZE_64, /* sz1 */ 858 OCT_FAU_OP_SIZE_64, /* sz0 */ 859 1, fau1, 1, fau0, /* s1, reg1, s0, reg0 */ 860 0, /* le */ 861 octeon_eth_param_pko_cmd_w0_n2, /* n2 */ 862 1, 0, /* q, r */ 863 (segs == 1) ? 0 : 1, /* g */ 864 0, 0, 1, /* ipoffp1, ii, df */ 865 segs, (int)len); /* segs, totalbytes */ 866} 867 868static inline uint64_t 869octeon_eth_send_makecmd_w1(int size, paddr_t addr) 870{ 871 return octeon_pko_cmd_word1( 872 0, 0, /* i, back */ 873 FPA_GATHER_BUFFER_POOL, /* pool */ 874 size, addr); /* size, addr */ 875} 876 877static inline int 878octeon_eth_send_makecmd_gbuf(struct octeon_eth_softc *sc, struct mbuf *m0, 879 uint64_t *gbuf, int *rsegs) 880{ 881 struct mbuf *m; 882 int segs = 0; 883 uintptr_t laddr, rlen, nlen; 884 885 for (m = m0; m != NULL; m = m->m_next) { 886 887 if (__predict_false(m->m_len == 0)) 888 continue; 889 890#if 0 891 OCTEON_ETH_KASSERT(((uint32_t)m->m_data & (PAGE_SIZE - 1)) 892 == (kvtophys((vaddr_t)m->m_data) & (PAGE_SIZE - 1))); 893#endif 894 895 /* 896 * aligned 4k 897 */ 898 laddr = (uintptr_t)m->m_data & (PAGE_SIZE - 1); 899 900 if (laddr + m->m_len > PAGE_SIZE) { 901 /* XXX XXX XXX */ 902 rlen = PAGE_SIZE - laddr; 903 nlen = m->m_len - rlen; 904 *(gbuf + segs) = octeon_eth_send_makecmd_w1(rlen, 905 kvtophys((vaddr_t)m->m_data)); 906 segs++; 907 if (segs > 63) { 908 return 1; 909 } 910 /* XXX XXX XXX */ 911 } else { 912 rlen = 0; 913 nlen = m->m_len; 914 } 915 916 *(gbuf + segs) = octeon_eth_send_makecmd_w1(nlen, 917 kvtophys((vaddr_t)(m->m_data + rlen))); 918 segs++; 919 if (segs > 63) { 920 return 1; 921 } 922 } 923 924 OCTEON_ETH_KASSERT(m == NULL); 925 926 *rsegs = segs; 927 928 return 0; 929} 930 931static inline int 932octeon_eth_send_makecmd(struct octeon_eth_softc *sc, struct mbuf *m, 933 uint64_t *gbuf, uint64_t *rpko_cmd_w0, uint64_t *rpko_cmd_w1) 934{ 935 uint64_t pko_cmd_w0, pko_cmd_w1; 936 int segs; 937 int result = 0; 938 939 if (octeon_eth_send_makecmd_gbuf(sc, m, gbuf, &segs)) { 940 log(LOG_WARNING, "%s: there are a lot of number of segments" 941 " of transmission data", device_xname(sc->sc_dev)); 942 result = 1; 943 goto done; 944 } 945 946 /* 947 * segs == 1 -> link mode (single continuous buffer) 948 * WORD1[size] is number of bytes pointed by segment 949 * 950 * segs > 1 -> gather mode (scatter-gather buffer) 951 * WORD1[size] is number of segments 952 */ 953 pko_cmd_w0 = octeon_eth_send_makecmd_w0(sc->sc_fau_done.fd_regno, 954 0, m->m_pkthdr.len, segs); 955 if (segs == 1) { 956 pko_cmd_w1 = octeon_eth_send_makecmd_w1( 957 m->m_pkthdr.len, kvtophys((vaddr_t)m->m_data)); 958 } else { 959#ifdef __mips_n32 960 KASSERT(MIPS_KSEG0_P(gbuf)); 961 pko_cmd_w1 = octeon_eth_send_makecmd_w1(segs, 962 MIPS_KSEG0_TO_PHYS(gbuf)); 963#else 964 pko_cmd_w1 = octeon_eth_send_makecmd_w1(segs, 965 MIPS_XKPHYS_TO_PHYS(gbuf)); 966#endif 967 } 968 969 *rpko_cmd_w0 = pko_cmd_w0; 970 *rpko_cmd_w1 = pko_cmd_w1; 971 972done: 973 return result; 974} 975 976static inline int 977octeon_eth_send_cmd(struct octeon_eth_softc *sc, uint64_t pko_cmd_w0, 978 uint64_t pko_cmd_w1, int *pwdc) 979{ 980 uint64_t *cmdptr; 981 int result = 0; 982 983#ifdef __mips_n32 984 KASSERT((sc->sc_cmdptr.cmdptr & ~MIPS_PHYS_MASK) == 0); 985 cmdptr = (uint64_t *)MIPS_PHYS_TO_KSEG0(sc->sc_cmdptr.cmdptr); 986#else 987 cmdptr = (uint64_t *)MIPS_PHYS_TO_XKPHYS_CACHED(sc->sc_cmdptr.cmdptr); 988#endif 989 cmdptr += sc->sc_cmdptr.cmdptr_idx; 990 991 OCTEON_ETH_KASSERT(cmdptr != NULL); 992 993 *cmdptr++ = pko_cmd_w0; 994 *cmdptr++ = pko_cmd_w1; 995 996 OCTEON_ETH_KASSERT(sc->sc_cmdptr.cmdptr_idx + 2 <= FPA_COMMAND_BUFFER_POOL_NWORDS - 1); 997 998 if (sc->sc_cmdptr.cmdptr_idx + 2 == FPA_COMMAND_BUFFER_POOL_NWORDS - 1) { 999 paddr_t buf; 1000 1001 buf = octeon_fpa_buf_get_paddr(octeon_eth_fb_cmd); 1002 if (buf == 0) { 1003 log(LOG_WARNING, 1004 "%s: can not allocate command buffer from free pool allocator\n", 1005 device_xname(sc->sc_dev)); 1006 result = 1; 1007 goto done; 1008 } 1009 OCTEON_EVCNT_INC(sc, txbufcbget); 1010 *cmdptr++ = buf; 1011 sc->sc_cmdptr.cmdptr = (uint64_t)buf; 1012 sc->sc_cmdptr.cmdptr_idx = 0; 1013 } else { 1014 sc->sc_cmdptr.cmdptr_idx += 2; 1015 } 1016 1017 *pwdc += 2; 1018 1019done: 1020 return result; 1021} 1022 1023static inline int 1024octeon_eth_send_buf(struct octeon_eth_softc *sc, struct mbuf *m, 1025 uint64_t *gbuf, int *pwdc) 1026{ 1027 int result = 0, error; 1028 uint64_t pko_cmd_w0, pko_cmd_w1; 1029 1030 error = octeon_eth_send_makecmd(sc, m, gbuf, &pko_cmd_w0, &pko_cmd_w1); 1031 if (error != 0) { 1032 /* already logging */ 1033 OCTEON_EVCNT_INC(sc, txerrmkcmd); 1034 result = error; 1035 goto done; 1036 } 1037 1038 error = octeon_eth_send_cmd(sc, pko_cmd_w0, pko_cmd_w1, pwdc); 1039 if (error != 0) { 1040 /* already logging */ 1041 OCTEON_EVCNT_INC(sc, txerrcmd); 1042 result = error; 1043 } 1044 1045done: 1046 return result; 1047} 1048 1049static inline int 1050octeon_eth_send(struct octeon_eth_softc *sc, struct mbuf *m, int *pwdc) 1051{ 1052 paddr_t gaddr = 0; 1053 uint64_t *gbuf = NULL; 1054 int result = 0, error; 1055 1056 OCTEON_EVCNT_INC(sc, tx); 1057 1058 gaddr = octeon_fpa_buf_get_paddr(octeon_eth_fb_sg); 1059 if (gaddr == 0) { 1060 log(LOG_WARNING, 1061 "%s: can not allocate gather buffer from free pool allocator\n", 1062 device_xname(sc->sc_dev)); 1063 OCTEON_EVCNT_INC(sc, txerrgbuf); 1064 result = 1; 1065 goto done; 1066 } 1067 OCTEON_EVCNT_INC(sc, txbufgbget); 1068 1069#ifdef __mips_n32 1070 KASSERT((gaddr & ~MIPS_PHYS_MASK) == 0); 1071 gbuf = (uint64_t *)(uintptr_t)MIPS_PHYS_TO_KSEG0(gaddr); 1072#else 1073 gbuf = (uint64_t *)(uintptr_t)MIPS_PHYS_TO_XKPHYS_CACHED(gaddr); 1074#endif 1075 1076 OCTEON_ETH_KASSERT(gbuf != NULL); 1077 1078 error = octeon_eth_send_buf(sc, m, gbuf, pwdc); 1079 if (error != 0) { 1080 /* already logging */ 1081 octeon_fpa_buf_put_paddr(octeon_eth_fb_sg, gaddr); 1082 OCTEON_EVCNT_INC(sc, txbufgbput); 1083 result = error; 1084 goto done; 1085 } 1086 1087 octeon_eth_send_queue_add(sc, m, gbuf); 1088 1089done: 1090 return result; 1091} 1092 1093static void 1094octeon_eth_start(struct ifnet *ifp) 1095{ 1096 struct octeon_eth_softc *sc = ifp->if_softc; 1097 struct mbuf *m; 1098 int wdc = 0; 1099 1100 /* 1101 * performance tuning 1102 * presend iobdma request 1103 */ 1104 octeon_eth_send_queue_flush_prefetch(sc); 1105 1106 if ((ifp->if_flags & (IFF_RUNNING | IFF_OACTIVE)) != IFF_RUNNING) 1107 goto last; 1108 1109 /* XXX assume that OCTEON doesn't buffer packets */ 1110 if (__predict_false(!octeon_gmx_link_status(sc->sc_gmx_port))) { 1111 /* dequeue and drop them */ 1112 while (1) { 1113 IFQ_DEQUEUE(&ifp->if_snd, m); 1114 if (m == NULL) 1115 break; 1116 1117 m_freem(m); 1118 IF_DROP(&ifp->if_snd); 1119 OCTEON_EVCNT_INC(sc, txerrlink); 1120 } 1121 goto last; 1122 } 1123 1124 for (;;) { 1125 IFQ_POLL(&ifp->if_snd, m); 1126 if (__predict_false(m == NULL)) 1127 break; 1128 1129 /* XXX XXX XXX */ 1130 octeon_eth_send_queue_flush_fetch(sc); 1131 1132 /* 1133 * If no free send buffer is available, free all the sent buffer 1134 * and bail out. 1135 */ 1136 if (octeon_eth_send_queue_is_full(sc)) { 1137 SET(ifp->if_flags, IFF_OACTIVE); 1138 if (wdc > 0) 1139 octeon_pko_op_doorbell_write(sc->sc_port, 1140 sc->sc_port, wdc); 1141 return; 1142 } 1143 /* XXX XXX XXX */ 1144 1145 IFQ_DEQUEUE(&ifp->if_snd, m); 1146 1147 bpf_mtap(ifp, m, BPF_D_OUT); 1148 1149 /* XXX XXX XXX */ 1150 if (sc->sc_soft_req_cnt > sc->sc_soft_req_thresh) 1151 octeon_eth_send_queue_flush(sc); 1152 if (octeon_eth_send(sc, m, &wdc)) { 1153 IF_DROP(&ifp->if_snd); 1154 m_freem(m); 1155 log(LOG_WARNING, 1156 "%s: failed in the transmission of the packet\n", 1157 device_xname(sc->sc_dev)); 1158 OCTEON_EVCNT_INC(sc, txerr); 1159 } else { 1160 sc->sc_soft_req_cnt++; 1161 } 1162 if (sc->sc_flush) 1163 octeon_eth_send_queue_flush_sync(sc); 1164 /* XXX XXX XXX */ 1165 1166 /* 1167 * send next iobdma request 1168 */ 1169 octeon_eth_send_queue_flush_prefetch(sc); 1170 } 1171 1172 if (wdc > 0) 1173 octeon_pko_op_doorbell_write(sc->sc_port, sc->sc_port, wdc); 1174 1175/* 1176 * Don't schedule send-buffer-free callout every time - those buffers are freed 1177 * by "free tick". This makes some packets like NFS slower. 1178 */ 1179#ifdef OCTEON_ETH_USENFS 1180 if (__predict_false(sc->sc_ext_callback_cnt > 0)) { 1181 int timo; 1182 1183 /* ??? */ 1184 timo = hz - (100 * sc->sc_ext_callback_cnt); 1185 if (timo < 10) 1186 timo = 10; 1187 callout_schedule(&sc->sc_tick_free_ch, timo); 1188 } 1189#endif 1190 1191last: 1192 octeon_eth_send_queue_flush_fetch(sc); 1193} 1194 1195static void 1196octeon_eth_watchdog(struct ifnet *ifp) 1197{ 1198 struct octeon_eth_softc *sc = ifp->if_softc; 1199 1200 printf("%s: device timeout\n", device_xname(sc->sc_dev)); 1201 1202 octeon_eth_configure(sc); 1203 1204 SET(ifp->if_flags, IFF_RUNNING); 1205 CLR(ifp->if_flags, IFF_OACTIVE); 1206 ifp->if_timer = 0; 1207 1208 octeon_eth_start(ifp); 1209} 1210 1211static int 1212octeon_eth_init(struct ifnet *ifp) 1213{ 1214 struct octeon_eth_softc *sc = ifp->if_softc; 1215 1216 /* XXX don't disable commonly used parts!!! XXX */ 1217 if (sc->sc_init_flag == 0) { 1218 /* Cancel any pending I/O. */ 1219 octeon_eth_stop(ifp, 0); 1220 1221 /* Initialize the device */ 1222 octeon_eth_configure(sc); 1223 1224 octeon_pko_enable(sc->sc_pko); 1225 octeon_ipd_enable(sc->sc_ipd); 1226 1227 sc->sc_init_flag = 1; 1228 } else { 1229 octeon_gmx_port_enable(sc->sc_gmx_port, 1); 1230 } 1231 octeon_eth_mediachange(ifp); 1232 1233 octeon_gmx_set_filter(sc->sc_gmx_port); 1234 1235 callout_reset(&sc->sc_tick_misc_ch, hz, octeon_eth_tick_misc, sc); 1236 callout_reset(&sc->sc_tick_free_ch, hz, octeon_eth_tick_free, sc); 1237 1238 SET(ifp->if_flags, IFF_RUNNING); 1239 CLR(ifp->if_flags, IFF_OACTIVE); 1240 1241 return 0; 1242} 1243 1244static void 1245octeon_eth_stop(struct ifnet *ifp, int disable) 1246{ 1247 struct octeon_eth_softc *sc = ifp->if_softc; 1248 1249 callout_stop(&sc->sc_tick_misc_ch); 1250 callout_stop(&sc->sc_tick_free_ch); 1251 1252 mii_down(&sc->sc_mii); 1253 1254 octeon_gmx_port_enable(sc->sc_gmx_port, 0); 1255 1256 /* Mark the interface as down and cancel the watchdog timer. */ 1257 CLR(ifp->if_flags, IFF_RUNNING | IFF_OACTIVE); 1258 ifp->if_timer = 0; 1259} 1260 1261/* ---- misc */ 1262 1263#define PKO_INDEX_MASK ((1ULL << 12/* XXX */) - 1) 1264 1265static int 1266octeon_eth_reset(struct octeon_eth_softc *sc) 1267{ 1268 octeon_gmx_reset_speed(sc->sc_gmx_port); 1269 octeon_gmx_reset_flowctl(sc->sc_gmx_port); 1270 octeon_gmx_reset_timing(sc->sc_gmx_port); 1271 1272 return 0; 1273} 1274 1275static int 1276octeon_eth_configure(struct octeon_eth_softc *sc) 1277{ 1278 octeon_gmx_port_enable(sc->sc_gmx_port, 0); 1279 1280 octeon_eth_reset(sc); 1281 1282 octeon_eth_configure_common(sc); 1283 1284 octeon_pko_port_config(sc->sc_pko); 1285 octeon_pko_port_enable(sc->sc_pko, 1); 1286 octeon_pip_port_config(sc->sc_pip); 1287 1288 octeon_gmx_tx_stats_rd_clr(sc->sc_gmx_port, 1); 1289 octeon_gmx_rx_stats_rd_clr(sc->sc_gmx_port, 1); 1290 1291 octeon_gmx_port_enable(sc->sc_gmx_port, 1); 1292 1293 return 0; 1294} 1295 1296static int 1297octeon_eth_configure_common(struct octeon_eth_softc *sc) 1298{ 1299 static int once; 1300 1301 if (once == 1) 1302 return 0; 1303 once = 1; 1304 1305 octeon_ipd_config(sc->sc_ipd); 1306#ifdef OCTEON_ETH_IPD_RED 1307 octeon_ipd_red(sc->sc_ipd, RECV_QUEUE_SIZE >> 2, RECV_QUEUE_SIZE >> 3); 1308#endif 1309 octeon_pko_config(sc->sc_pko); 1310 1311 octeon_pow_config(sc->sc_pow, OCTEON_POW_GROUP_PIP); 1312 1313 return 0; 1314} 1315 1316/* ---- receive (input) */ 1317 1318static inline int 1319octeon_eth_recv_mbuf(struct octeon_eth_softc *sc, uint64_t *work, 1320 struct mbuf **rm) 1321{ 1322 struct mbuf *m; 1323 void (*ext_free)(struct mbuf *, void *, size_t, void *); 1324 void *ext_buf; 1325 size_t ext_size; 1326 void *data; 1327 uint64_t word1 = work[1]; 1328 uint64_t word2 = work[2]; 1329 uint64_t word3 = work[3]; 1330 1331 MGETHDR(m, M_NOWAIT, MT_DATA); 1332 if (m == NULL) 1333 return 1; 1334 OCTEON_ETH_KASSERT(m != NULL); 1335 1336 if ((word2 & PIP_WQE_WORD2_IP_BUFS) == 0) { 1337 /* Dynamic short */ 1338 ext_free = octeon_eth_buf_ext_free_m; 1339 ext_buf = &work[4]; 1340 ext_size = 96; 1341 1342 data = &work[4 + sc->sc_ip_offset / sizeof(uint64_t)]; 1343 } else { 1344 vaddr_t addr; 1345 vaddr_t start_buffer; 1346 1347#ifdef __mips_n32 1348 KASSERT((word3 & ~MIPS_PHYS_MASK) == 0); 1349 addr = MIPS_PHYS_TO_KSEG0(word3 & PIP_WQE_WORD3_ADDR); 1350#else 1351 addr = MIPS_PHYS_TO_XKPHYS_CACHED(word3 & PIP_WQE_WORD3_ADDR); 1352#endif 1353 start_buffer = addr & ~(2048 - 1); 1354 1355 ext_free = octeon_eth_buf_ext_free_ext; 1356 ext_buf = (void *)start_buffer; 1357 ext_size = 2048; 1358 1359 data = (void *)addr; 1360 } 1361 1362 /* embed sc pointer into work[0] for _ext_free evcnt */ 1363 work[0] = (uintptr_t)sc; 1364 1365 MEXTADD(m, ext_buf, ext_size, 0, ext_free, work); 1366 OCTEON_ETH_KASSERT(ISSET(m->m_flags, M_EXT)); 1367 1368 m->m_data = data; 1369 m->m_len = m->m_pkthdr.len = (word1 & PIP_WQE_WORD1_LEN) >> 48; 1370 m_set_rcvif(m, &sc->sc_ethercom.ec_if); 1371 /* 1372 * not readonly buffer 1373 */ 1374 m->m_flags |= M_EXT_RW; 1375 1376 *rm = m; 1377 1378 OCTEON_ETH_KASSERT(*rm != NULL); 1379 1380 return 0; 1381} 1382 1383static inline int 1384octeon_eth_recv_check_code(struct octeon_eth_softc *sc, uint64_t word2) 1385{ 1386 uint64_t opecode = word2 & PIP_WQE_WORD2_NOIP_OPECODE; 1387 1388 if (__predict_true(!ISSET(word2, PIP_WQE_WORD2_NOIP_RE))) 1389 return 0; 1390 1391 /* this error is harmless */ 1392 if (opecode == PIP_OVER_ERR) 1393 return 0; 1394 1395 return 1; 1396} 1397 1398static inline int 1399octeon_eth_recv_check_jumbo(struct octeon_eth_softc *sc, uint64_t word2) 1400{ 1401 if (__predict_false((word2 & PIP_WQE_WORD2_IP_BUFS) > (1ULL << 56))) 1402 return 1; 1403 return 0; 1404} 1405 1406static inline int 1407octeon_eth_recv_check_link(struct octeon_eth_softc *sc, uint64_t word2) 1408{ 1409 if (__predict_false(!octeon_gmx_link_status(sc->sc_gmx_port))) 1410 return 1; 1411 return 0; 1412} 1413 1414static inline int 1415octeon_eth_recv_check(struct octeon_eth_softc *sc, uint64_t word2) 1416{ 1417 if (__predict_false(octeon_eth_recv_check_link(sc, word2)) != 0) { 1418 if (ratecheck(&sc->sc_rate_recv_check_link_last, 1419 &sc->sc_rate_recv_check_link_cap)) 1420 log(LOG_DEBUG, 1421 "%s: link is not up, the packet was dropped\n", 1422 device_xname(sc->sc_dev)); 1423 OCTEON_EVCNT_INC(sc, rxerrlink); 1424 return 1; 1425 } 1426 1427#if 0 /* XXX Performance tunig (Jumbo-frame is not supported yet!) */ 1428 if (__predict_false(octeon_eth_recv_check_jumbo(sc, word2)) != 0) { 1429 /* XXX jumbo frame */ 1430 if (ratecheck(&sc->sc_rate_recv_check_jumbo_last, 1431 &sc->sc_rate_recv_check_jumbo_cap)) 1432 log(LOG_DEBUG, 1433 "jumbo frame was received\n"); 1434 OCTEON_EVCNT_INC(sc, rxerrjmb); 1435 return 1; 1436 } 1437#endif 1438 1439 if (__predict_false(octeon_eth_recv_check_code(sc, word2)) != 0) { 1440 1441 if ((word2 & PIP_WQE_WORD2_NOIP_OPECODE) == 1442 PIP_WQE_WORD2_RE_OPCODE_LENGTH) { 1443 /* no logging */ 1444 /* XXX inclement special error count */ 1445 } else if ((word2 & PIP_WQE_WORD2_NOIP_OPECODE) == 1446 PIP_WQE_WORD2_RE_OPCODE_PARTIAL) { 1447 /* not an erorr. it's because of overload */ 1448 } else { 1449 1450 if (ratecheck(&sc->sc_rate_recv_check_code_last, 1451 &sc->sc_rate_recv_check_code_cap)) 1452 log(LOG_WARNING, 1453 "%s: reception error, packet dropped " 1454 "(error code = %" PRId64 ")\n", 1455 device_xname(sc->sc_dev), word2 & PIP_WQE_WORD2_NOIP_OPECODE); 1456 } 1457 OCTEON_EVCNT_INC(sc, rxerrcode); 1458 return 1; 1459 } 1460 1461 return 0; 1462} 1463 1464static inline int 1465octeon_eth_recv(struct octeon_eth_softc *sc, uint64_t *work) 1466{ 1467 int result = 0; 1468 struct ifnet *ifp; 1469 struct mbuf *m; 1470 uint64_t word2; 1471 1472 /* XXX XXX XXX */ 1473 /* 1474 * performance tuning 1475 * presend iobdma request 1476 */ 1477 if (sc->sc_soft_req_cnt > sc->sc_soft_req_thresh) { 1478 octeon_eth_send_queue_flush_prefetch(sc); 1479 } 1480 /* XXX XXX XXX */ 1481 1482 OCTEON_ETH_KASSERT(sc != NULL); 1483 OCTEON_ETH_KASSERT(work != NULL); 1484 1485 OCTEON_EVCNT_INC(sc, rx); 1486 1487 word2 = work[2]; 1488 ifp = &sc->sc_ethercom.ec_if; 1489 1490 OCTEON_ETH_KASSERT(ifp != NULL); 1491 1492 if (__predict_false(octeon_eth_recv_check(sc, word2) != 0)) { 1493 ifp->if_ierrors++; 1494 result = 1; 1495 octeon_eth_buf_free_work(sc, work, word2); 1496 goto drop; 1497 } 1498 1499 if (__predict_false(octeon_eth_recv_mbuf(sc, work, &m) != 0)) { 1500 ifp->if_ierrors++; 1501 result = 1; 1502 octeon_eth_buf_free_work(sc, work, word2); 1503 goto drop; 1504 } 1505 1506 /* work[0] .. work[3] may not be valid any more */ 1507 1508 OCTEON_ETH_KASSERT(m != NULL); 1509 1510 octeon_ipd_offload(word2, m->m_data, &m->m_pkthdr.csum_flags); 1511 1512 /* XXX XXX XXX */ 1513 if (sc->sc_soft_req_cnt > sc->sc_soft_req_thresh) { 1514 octeon_eth_send_queue_flush_fetch(sc); 1515 octeon_eth_send_queue_flush(sc); 1516 } 1517 1518 /* XXX XXX XXX */ 1519 if (sc->sc_flush) 1520 octeon_eth_send_queue_flush_sync(sc); 1521 /* XXX XXX XXX */ 1522 1523 if_percpuq_enqueue(ifp->if_percpuq, m); 1524 1525 return 0; 1526 1527drop: 1528 /* XXX XXX XXX */ 1529 if (sc->sc_soft_req_cnt > sc->sc_soft_req_thresh) { 1530 octeon_eth_send_queue_flush_fetch(sc); 1531 } 1532 /* XXX XXX XXX */ 1533 1534 return result; 1535} 1536 1537static void 1538octeon_eth_recv_redir(struct ifnet *ifp, struct mbuf *m) 1539{ 1540 struct octeon_eth_softc *rsc = ifp->if_softc; 1541 struct octeon_eth_softc *sc = NULL; 1542 int i, wdc = 0; 1543 1544 for (i = 0; i < 3 /* XXX */; i++) { 1545 if (rsc->sc_redir & (1 << i)) 1546 sc = octeon_eth_gsc[i]; 1547 } 1548 1549 if (sc == NULL) { 1550 m_freem(m); 1551 return; 1552 } 1553 octeon_eth_send_queue_flush_prefetch(sc); 1554 1555 octeon_eth_send_queue_flush_fetch(sc); 1556 1557 if (octeon_eth_send_queue_is_full(sc)) { 1558 m_freem(m); 1559 return; 1560 } 1561 if (sc->sc_soft_req_cnt > sc->sc_soft_req_thresh) 1562 octeon_eth_send_queue_flush(sc); 1563 1564 if (octeon_eth_send(sc, m, &wdc)) { 1565 IF_DROP(&ifp->if_snd); 1566 m_freem(m); 1567 } else { 1568 octeon_pko_op_doorbell_write(sc->sc_port, sc->sc_port, wdc); 1569 sc->sc_soft_req_cnt++; 1570 } 1571 1572 if (sc->sc_flush) 1573 octeon_eth_send_queue_flush_sync(sc); 1574} 1575 1576static inline void 1577octeon_eth_recv_intr(void *data, uint64_t *work) 1578{ 1579 struct octeon_eth_softc *sc; 1580 int port; 1581 1582 OCTEON_ETH_KASSERT(work != NULL); 1583 1584 port = (work[1] & PIP_WQE_WORD1_IPRT) >> 42; 1585 1586 OCTEON_ETH_KASSERT(port < GMX_PORT_NUNITS); 1587 1588 sc = octeon_eth_gsc[port]; 1589 1590 OCTEON_ETH_KASSERT(sc != NULL); 1591 OCTEON_ETH_KASSERT(port == sc->sc_port); 1592 1593 /* XXX process all work queue entries anyway */ 1594 1595 (void)octeon_eth_recv(sc, work); 1596} 1597 1598/* ---- tick */ 1599 1600/* 1601 * octeon_eth_tick_free 1602 * 1603 * => garbage collect send gather buffer / mbuf 1604 * => called at softclock 1605 */ 1606static void 1607octeon_eth_tick_free(void *arg) 1608{ 1609 struct octeon_eth_softc *sc = arg; 1610 int timo; 1611 int s; 1612 1613 s = splnet(); 1614 /* XXX XXX XXX */ 1615 if (sc->sc_soft_req_cnt > 0) { 1616 octeon_eth_send_queue_flush_prefetch(sc); 1617 octeon_eth_send_queue_flush_fetch(sc); 1618 octeon_eth_send_queue_flush(sc); 1619 octeon_eth_send_queue_flush_sync(sc); 1620 } 1621 /* XXX XXX XXX */ 1622 1623 /* XXX XXX XXX */ 1624 /* ??? */ 1625 timo = hz - (100 * sc->sc_ext_callback_cnt); 1626 if (timo < 10) 1627 timo = 10; 1628 callout_schedule(&sc->sc_tick_free_ch, timo); 1629 /* XXX XXX XXX */ 1630 splx(s); 1631} 1632 1633/* 1634 * octeon_eth_tick_misc 1635 * 1636 * => collect statistics 1637 * => check link status 1638 * => called at softclock 1639 */ 1640static void 1641octeon_eth_tick_misc(void *arg) 1642{ 1643 struct octeon_eth_softc *sc = arg; 1644 struct ifnet *ifp; 1645 int s; 1646 1647 s = splnet(); 1648 1649 ifp = &sc->sc_ethercom.ec_if; 1650 1651 octeon_gmx_stats(sc->sc_gmx_port); 1652 octeon_pip_stats(sc->sc_pip, ifp, sc->sc_port); 1653 mii_tick(&sc->sc_mii); 1654 1655 splx(s); 1656 1657 callout_schedule(&sc->sc_tick_misc_ch, hz); 1658} 1659 1660/* ---- odd nibble preamble workaround (software CRC processing) */ 1661 1662/* ---- sysctl */ 1663 1664static int octeon_eth_sysctl_verify(SYSCTLFN_ARGS); 1665static int octeon_eth_sysctl_pool(SYSCTLFN_ARGS); 1666static int octeon_eth_sysctl_rd(SYSCTLFN_ARGS); 1667 1668static int octeon_eth_sysctl_pkocmdw0n2_num; 1669static int octeon_eth_sysctl_pipdynrs_num; 1670static int octeon_eth_sysctl_redir_num; 1671static int octeon_eth_sysctl_pkt_pool_num; 1672static int octeon_eth_sysctl_wqe_pool_num; 1673static int octeon_eth_sysctl_cmd_pool_num; 1674static int octeon_eth_sysctl_sg_pool_num; 1675static int octeon_eth_sysctl_pktbuf_num; 1676 1677/* 1678 * Set up sysctl(3) MIB, hw.cnmac.*. 1679 */ 1680SYSCTL_SETUP(sysctl_octeon_eth, "sysctl cnmac subtree setup") 1681{ 1682 int rc; 1683 int octeon_eth_sysctl_root_num; 1684 const struct sysctlnode *node; 1685 1686 if ((rc = sysctl_createv(clog, 0, NULL, NULL, 1687 0, CTLTYPE_NODE, "hw", NULL, 1688 NULL, 0, NULL, 0, CTL_HW, CTL_EOL)) != 0) { 1689 goto err; 1690 } 1691 1692 if ((rc = sysctl_createv(clog, 0, NULL, &node, 1693 0, CTLTYPE_NODE, "cnmac", 1694 SYSCTL_DESCR("cnmac interface controls"), 1695 NULL, 0, NULL, 0, CTL_HW, CTL_CREATE, CTL_EOL)) != 0) { 1696 goto err; 1697 } 1698 1699 octeon_eth_sysctl_root_num = node->sysctl_num; 1700 1701 if ((rc = sysctl_createv(clog, 0, NULL, &node, 1702 CTLFLAG_PERMANENT | CTLFLAG_READWRITE, 1703 CTLTYPE_INT, "pko_cmd_w0_n2", 1704 SYSCTL_DESCR("PKO command WORD0 N2 bit"), 1705 octeon_eth_sysctl_verify, 0, 1706 &octeon_eth_param_pko_cmd_w0_n2, 1707 0, CTL_HW, octeon_eth_sysctl_root_num, CTL_CREATE, 1708 CTL_EOL)) != 0) { 1709 goto err; 1710 } 1711 1712 octeon_eth_sysctl_pkocmdw0n2_num = node->sysctl_num; 1713 1714 if ((rc = sysctl_createv(clog, 0, NULL, &node, 1715 CTLFLAG_PERMANENT | CTLFLAG_READWRITE, 1716 CTLTYPE_INT, "pip_dyn_rs", 1717 SYSCTL_DESCR("PIP dynamic short in WQE"), 1718 octeon_eth_sysctl_verify, 0, 1719 &octeon_eth_param_pip_dyn_rs, 1720 0, CTL_HW, octeon_eth_sysctl_root_num, CTL_CREATE, 1721 CTL_EOL)) != 0) { 1722 goto err; 1723 } 1724 1725 octeon_eth_sysctl_pipdynrs_num = node->sysctl_num; 1726 1727 if ((rc = sysctl_createv(clog, 0, NULL, &node, 1728 CTLFLAG_PERMANENT | CTLFLAG_READWRITE, 1729 CTLTYPE_INT, "redir", 1730 SYSCTL_DESCR("input port redirection"), 1731 octeon_eth_sysctl_verify, 0, 1732 &octeon_eth_param_redir, 1733 0, CTL_HW, octeon_eth_sysctl_root_num, CTL_CREATE, 1734 CTL_EOL)) != 0) { 1735 goto err; 1736 } 1737 1738 octeon_eth_sysctl_redir_num = node->sysctl_num; 1739 1740 if ((rc = sysctl_createv(clog, 0, NULL, &node, 1741 CTLFLAG_PERMANENT, 1742 CTLTYPE_INT, "pkt_pool", 1743 SYSCTL_DESCR("packet pool available"), 1744 octeon_eth_sysctl_pool, 0, NULL, 1745 0, CTL_HW, octeon_eth_sysctl_root_num, CTL_CREATE, 1746 CTL_EOL)) != 0) { 1747 goto err; 1748 } 1749 1750 octeon_eth_sysctl_pkt_pool_num = node->sysctl_num; 1751 1752 if ((rc = sysctl_createv(clog, 0, NULL, &node, 1753 CTLFLAG_PERMANENT, 1754 CTLTYPE_INT, "wqe_pool", 1755 SYSCTL_DESCR("wqe pool available"), 1756 octeon_eth_sysctl_pool, 0, NULL, 1757 0, CTL_HW, octeon_eth_sysctl_root_num, CTL_CREATE, 1758 CTL_EOL)) != 0) { 1759 goto err; 1760 } 1761 1762 octeon_eth_sysctl_wqe_pool_num = node->sysctl_num; 1763 1764 if ((rc = sysctl_createv(clog, 0, NULL, &node, 1765 CTLFLAG_PERMANENT, 1766 CTLTYPE_INT, "cmd_pool", 1767 SYSCTL_DESCR("cmd pool available"), 1768 octeon_eth_sysctl_pool, 0, NULL, 1769 0, CTL_HW, octeon_eth_sysctl_root_num, CTL_CREATE, 1770 CTL_EOL)) != 0) { 1771 goto err; 1772 } 1773 1774 octeon_eth_sysctl_cmd_pool_num = node->sysctl_num; 1775 1776 if ((rc = sysctl_createv(clog, 0, NULL, &node, 1777 CTLFLAG_PERMANENT, 1778 CTLTYPE_INT, "sg_pool", 1779 SYSCTL_DESCR("sg pool available"), 1780 octeon_eth_sysctl_pool, 0, NULL, 1781 0, CTL_HW, octeon_eth_sysctl_root_num, CTL_CREATE, 1782 CTL_EOL)) != 0) { 1783 goto err; 1784 } 1785 1786 octeon_eth_sysctl_sg_pool_num = node->sysctl_num; 1787 1788 if ((rc = sysctl_createv(clog, 0, NULL, &node, 1789 CTLFLAG_PERMANENT | CTLFLAG_READONLY, 1790 CTLTYPE_INT, "pktbuf", 1791 SYSCTL_DESCR("input packet buffer size on POW"), 1792 octeon_eth_sysctl_rd, 0, 1793 &octeon_eth_param_pktbuf, 1794 0, CTL_HW, octeon_eth_sysctl_root_num, CTL_CREATE, 1795 CTL_EOL)) != 0) { 1796 goto err; 1797 } 1798 1799 octeon_eth_sysctl_pktbuf_num = node->sysctl_num; 1800 1801 return; 1802 1803err: 1804 aprint_error("%s: syctl_createv failed (rc = %d)\n", __func__, rc); 1805} 1806 1807static int 1808octeon_eth_sysctl_verify(SYSCTLFN_ARGS) 1809{ 1810 int error, v; 1811 struct sysctlnode node; 1812 struct octeon_eth_softc *sc; 1813 int i; 1814 int s; 1815 1816 node = *rnode; 1817 v = *(int *)rnode->sysctl_data; 1818 node.sysctl_data = &v; 1819 error = sysctl_lookup(SYSCTLFN_CALL(&node)); 1820 if (error || newp == NULL) 1821 return error; 1822 1823 if (node.sysctl_num == octeon_eth_sysctl_pkocmdw0n2_num) { 1824 if (v < 0 || v > 1) 1825 return EINVAL; 1826 *(int *)rnode->sysctl_data = v; 1827 return 0; 1828 } 1829 1830 if (node.sysctl_num == octeon_eth_sysctl_pipdynrs_num) { 1831 if (v < 0 || v > 1) 1832 return EINVAL; 1833 *(int *)rnode->sysctl_data = v; 1834 s = splnet(); 1835 for (i = 0; i < 3/* XXX */; i++) { 1836 sc = octeon_eth_gsc[i]; /* XXX */ 1837 octeon_pip_prt_cfg_enable(sc->sc_pip, PIP_PRT_CFGN_DYN_RS, v); 1838 } 1839 splx(s); 1840 return 0; 1841 } 1842 1843 if (node.sysctl_num == octeon_eth_sysctl_redir_num) { 1844 if (v & ~((0x7 << (4 * 0)) | (0x7 << (4 * 1)) | (0x7 << (4 * 2)))) 1845 return EINVAL; 1846 *(int *)rnode->sysctl_data = v; 1847 s = splnet(); 1848 for (i = 0; i < 3/* XXX */; i++) { 1849 struct ifnet *ifp; 1850 1851 sc = octeon_eth_gsc[i]; /* XXX */ 1852 ifp = &sc->sc_ethercom.ec_if; 1853 1854 sc->sc_redir = (octeon_eth_param_redir >> (4 * i)) & 0x7; 1855 if (sc->sc_redir == 0) { 1856 if (ISSET(ifp->if_flags, IFF_PROMISC)) { 1857 CLR(ifp->if_flags, IFF_PROMISC); 1858 octeon_eth_mii_statchg(ifp); 1859 /* octeon_gmx_set_filter(sc->sc_gmx_port); */ 1860 } 1861 ifp->_if_input = ether_input; 1862 } 1863 else { 1864 if (!ISSET(ifp->if_flags, IFF_PROMISC)) { 1865 SET(ifp->if_flags, IFF_PROMISC); 1866 octeon_eth_mii_statchg(ifp); 1867 /* octeon_gmx_set_filter(sc->sc_gmx_port); */ 1868 } 1869 ifp->_if_input = octeon_eth_recv_redir; 1870 } 1871 } 1872 splx(s); 1873 return 0; 1874 } 1875 1876 return EINVAL; 1877} 1878 1879static int 1880octeon_eth_sysctl_pool(SYSCTLFN_ARGS) 1881{ 1882 int error, newval = 0; 1883 struct sysctlnode node; 1884 int s; 1885 1886 node = *rnode; 1887 node.sysctl_data = &newval; 1888 s = splnet(); 1889 if (node.sysctl_num == octeon_eth_sysctl_pkt_pool_num) { 1890 error = octeon_fpa_available_fpa_pool(&newval, OCTEON_POOL_NO_PKT); 1891 } else if (node.sysctl_num == octeon_eth_sysctl_wqe_pool_num) { 1892 error = octeon_fpa_available_fpa_pool(&newval, OCTEON_POOL_NO_WQE); 1893 } else if (node.sysctl_num == octeon_eth_sysctl_cmd_pool_num) { 1894 error = octeon_fpa_available_fpa_pool(&newval, OCTEON_POOL_NO_CMD); 1895 } else if (node.sysctl_num == octeon_eth_sysctl_sg_pool_num) { 1896 error = octeon_fpa_available_fpa_pool(&newval, OCTEON_POOL_NO_SG); 1897 } else { 1898 splx(s); 1899 return EINVAL; 1900 } 1901 splx(s); 1902 if (error) 1903 return error; 1904 error = sysctl_lookup(SYSCTLFN_CALL(&node)); 1905 if (error || newp == NULL) 1906 return error; 1907 1908 return 0; 1909} 1910 1911static int 1912octeon_eth_sysctl_rd(SYSCTLFN_ARGS) 1913{ 1914 int error, v; 1915 struct sysctlnode node; 1916 int s; 1917 1918 node = *rnode; 1919 v = *(int *)rnode->sysctl_data; 1920 node.sysctl_data = &v; 1921 error = sysctl_lookup(SYSCTLFN_CALL(&node)); 1922 if (error || newp != NULL) 1923 return error; 1924 1925 if (node.sysctl_num == octeon_eth_sysctl_pktbuf_num) { 1926 uint64_t tmp; 1927 int n; 1928 1929 s = splnet(); 1930 tmp = octeon_fpa_query(0); 1931 n = (int)tmp; 1932 splx(s); 1933 *(int *)rnode->sysctl_data = n; 1934 octeon_eth_param_pktbuf = n; 1935 *(int *)oldp = n; 1936 return 0; 1937 } 1938 1939 return EINVAL; 1940} 1941