if_cnmac.c revision 1.17
1/*	$NetBSD: if_cnmac.c,v 1.17 2020/02/18 15:00:42 thorpej Exp $	*/
2
3#include <sys/cdefs.h>
4#if 0
5__KERNEL_RCSID(0, "$NetBSD: if_cnmac.c,v 1.17 2020/02/18 15:00:42 thorpej Exp $");
6#endif
7
8#include "opt_octeon.h"
9
10#ifdef	OCTEON_ETH_DEBUG
11
12#ifndef DIAGNOSTIC
13#define	DIAGNOSTIC
14#endif
15
16#ifndef DEBUG
17#define	DEBUG
18#endif
19
20#endif
21
22/*
23 * If no free send buffer is available, free all the sent buffers and bail out.
24 */
25#define OCTEON_ETH_SEND_QUEUE_CHECK
26
27/* XXX XXX XXX XXX XXX XXX */
28
29#include <sys/param.h>
30#include <sys/systm.h>
31#include <sys/pool.h>
32#include <sys/mbuf.h>
33#include <sys/malloc.h>
34#include <sys/kernel.h>
35#include <sys/socket.h>
36#include <sys/ioctl.h>
37#include <sys/errno.h>
38#include <sys/device.h>
39#include <sys/queue.h>
40#include <sys/conf.h>
41#include <sys/sysctl.h>
42#include <sys/syslog.h>
43
44#include <net/if.h>
45#include <net/if_dl.h>
46#include <net/if_media.h>
47#include <net/if_ether.h>
48#include <net/route.h>
49#include <net/bpf.h>
50
51#include <netinet/in.h>
52#include <netinet/in_systm.h>
53#include <netinet/in_var.h>
54#include <netinet/ip.h>
55
56#include <sys/bus.h>
57#include <machine/intr.h>
58#include <machine/endian.h>
59#include <machine/locore.h>
60
61#include <dev/mii/mii.h>
62#include <dev/mii/miivar.h>
63
64#include <mips/cpuregs.h>
65
66#include <mips/cavium/dev/octeon_asxreg.h>
67#include <mips/cavium/dev/octeon_ciureg.h>
68#include <mips/cavium/dev/octeon_npireg.h>
69#include <mips/cavium/dev/octeon_gmxreg.h>
70#include <mips/cavium/dev/octeon_ipdreg.h>
71#include <mips/cavium/dev/octeon_pipreg.h>
72#include <mips/cavium/dev/octeon_powreg.h>
73#include <mips/cavium/dev/octeon_faureg.h>
74#include <mips/cavium/dev/octeon_fpareg.h>
75#include <mips/cavium/dev/octeon_bootbusreg.h>
76#include <mips/cavium/include/iobusvar.h>
77#include <mips/cavium/octeonvar.h>
78#include <mips/cavium/dev/octeon_fpavar.h>
79#include <mips/cavium/dev/octeon_gmxvar.h>
80#include <mips/cavium/dev/octeon_fauvar.h>
81#include <mips/cavium/dev/octeon_powvar.h>
82#include <mips/cavium/dev/octeon_ipdvar.h>
83#include <mips/cavium/dev/octeon_pipvar.h>
84#include <mips/cavium/dev/octeon_pkovar.h>
85#include <mips/cavium/dev/octeon_asxvar.h>
86#include <mips/cavium/dev/octeon_smivar.h>
87#include <mips/cavium/dev/if_cnmacvar.h>
88
89#ifdef OCTEON_ETH_DEBUG
90#define	OCTEON_ETH_KASSERT(x)	KASSERT(x)
91#define	OCTEON_ETH_KDASSERT(x)	KDASSERT(x)
92#else
93#define	OCTEON_ETH_KASSERT(x)
94#define	OCTEON_ETH_KDASSERT(x)
95#endif
96
97/*
98 * Set the PKO to think command buffers are an odd length.  This makes it so we
99 * never have to divide a comamnd across two buffers.
100 */
101#define OCTEON_POOL_NWORDS_CMD	\
102	    (((uint32_t)OCTEON_POOL_SIZE_CMD / sizeof(uint64_t)) - 1)
103#define FPA_COMMAND_BUFFER_POOL_NWORDS	OCTEON_POOL_NWORDS_CMD	/* XXX */
104
105static void	octeon_eth_buf_init(struct octeon_eth_softc *);
106
107static int	octeon_eth_match(device_t, struct cfdata *, void *);
108static void	octeon_eth_attach(device_t, device_t, void *);
109static void	octeon_eth_pip_init(struct octeon_eth_softc *);
110static void	octeon_eth_ipd_init(struct octeon_eth_softc *);
111static void	octeon_eth_pko_init(struct octeon_eth_softc *);
112static void	octeon_eth_asx_init(struct octeon_eth_softc *);
113static void	octeon_eth_smi_init(struct octeon_eth_softc *);
114
115static void	octeon_eth_board_mac_addr(uint8_t *, size_t,
116    struct octeon_eth_softc *);
117
118static int	octeon_eth_mii_readreg(device_t, int, int, uint16_t *);
119static int	octeon_eth_mii_writereg(device_t, int, int, uint16_t);
120static void	octeon_eth_mii_statchg(struct ifnet *);
121
122static int	octeon_eth_mediainit(struct octeon_eth_softc *);
123static void	octeon_eth_mediastatus(struct ifnet *, struct ifmediareq *);
124
125static inline void octeon_eth_send_queue_flush_prefetch(struct octeon_eth_softc *);
126static inline void octeon_eth_send_queue_flush_fetch(struct octeon_eth_softc *);
127static inline void octeon_eth_send_queue_flush(struct octeon_eth_softc *);
128static inline void octeon_eth_send_queue_flush_sync(struct octeon_eth_softc *);
129static inline int octeon_eth_send_queue_is_full(struct octeon_eth_softc *);
130static inline void octeon_eth_send_queue_add(struct octeon_eth_softc *,
131    struct mbuf *, uint64_t *);
132static inline void octeon_eth_send_queue_del(struct octeon_eth_softc *,
133    struct mbuf **, uint64_t **);
134static inline int octeon_eth_buf_free_work(struct octeon_eth_softc *,
135    uint64_t *, uint64_t);
136static inline void octeon_eth_buf_ext_free_m(struct mbuf *, void *, size_t,
137    void *);
138static inline void octeon_eth_buf_ext_free_ext(struct mbuf *, void *, size_t,
139    void *);
140
141static int	octeon_eth_ioctl(struct ifnet *, u_long, void *);
142static void	octeon_eth_watchdog(struct ifnet *);
143static int	octeon_eth_init(struct ifnet *);
144static void	octeon_eth_stop(struct ifnet *, int);
145static void	octeon_eth_start(struct ifnet *);
146
147static inline int octeon_eth_send_cmd(struct octeon_eth_softc *, uint64_t,
148    uint64_t, int *);
149static inline uint64_t	octeon_eth_send_makecmd_w1(int, paddr_t);
150static inline uint64_t	octeon_eth_send_makecmd_w0(uint64_t, uint64_t, size_t,
151    int);
152static inline int octeon_eth_send_makecmd_gbuf(struct octeon_eth_softc *,
153    struct mbuf *, uint64_t *, int *);
154static inline int octeon_eth_send_makecmd(struct octeon_eth_softc *,
155    struct mbuf *, uint64_t *, uint64_t *, uint64_t *);
156static inline int octeon_eth_send_buf(struct octeon_eth_softc *,
157    struct mbuf *, uint64_t *, int *);
158static inline int octeon_eth_send(struct octeon_eth_softc *,
159    struct mbuf *, int *);
160
161static int	octeon_eth_reset(struct octeon_eth_softc *);
162static int	octeon_eth_configure(struct octeon_eth_softc *);
163static int	octeon_eth_configure_common(struct octeon_eth_softc *);
164
165static void	octeon_eth_tick_free(void *);
166static void	octeon_eth_tick_misc(void *);
167
168static inline int octeon_eth_recv_mbuf(struct octeon_eth_softc *,
169    uint64_t *, struct mbuf **);
170static inline int octeon_eth_recv_check_code(struct octeon_eth_softc *,
171    uint64_t);
172static inline int octeon_eth_recv_check_jumbo(struct octeon_eth_softc *,
173    uint64_t);
174static inline int octeon_eth_recv_check_link(struct octeon_eth_softc *,
175    uint64_t);
176static inline int octeon_eth_recv_check(struct octeon_eth_softc *,
177    uint64_t);
178static inline int octeon_eth_recv(struct octeon_eth_softc *, uint64_t *);
179static void	octeon_eth_recv_redir(struct ifnet *, struct mbuf *);
180static inline void octeon_eth_recv_intr(void *, uint64_t *);
181
182/* Device driver context */
183static struct	octeon_eth_softc *octeon_eth_gsc[GMX_PORT_NUNITS];
184static void	*octeon_eth_pow_recv_ih;
185
186/* sysctl'able parameters */
187int		octeon_eth_param_pko_cmd_w0_n2 = 1;
188int		octeon_eth_param_pip_dyn_rs = 1;
189int		octeon_eth_param_redir = 0;
190int		octeon_eth_param_pktbuf = 0;
191int		octeon_eth_param_rate = 0;
192int		octeon_eth_param_intr = 0;
193
194CFATTACH_DECL_NEW(cnmac, sizeof(struct octeon_eth_softc),
195    octeon_eth_match, octeon_eth_attach, NULL, NULL);
196
197#ifdef OCTEON_ETH_DEBUG
198
199static const struct octeon_evcnt_entry octeon_evcnt_entries[] = {
200#define	_ENTRY(name, type, parent, descr) \
201	OCTEON_EVCNT_ENTRY(struct octeon_eth_softc, name, type, parent, descr)
202	_ENTRY(rx,			MISC, NULL, "rx"),
203	_ENTRY(rxint,			INTR, NULL, "rx intr"),
204	_ENTRY(rxrs,			MISC, NULL, "rx dynamic short"),
205	_ENTRY(rxbufpkalloc,		MISC, NULL, "rx buf pkt alloc"),
206	_ENTRY(rxbufpkput,		MISC, NULL, "rx buf pkt put"),
207	_ENTRY(rxbufwqalloc,		MISC, NULL, "rx buf wqe alloc"),
208	_ENTRY(rxbufwqput,		MISC, NULL, "rx buf wqe put"),
209	_ENTRY(rxerrcode,		MISC, NULL, "rx code error"),
210	_ENTRY(rxerrfix,		MISC, NULL, "rx fixup error"),
211	_ENTRY(rxerrjmb,		MISC, NULL, "rx jmb error"),
212	_ENTRY(rxerrlink,		MISC, NULL, "rx link error"),
213	_ENTRY(rxerroff,		MISC, NULL, "rx offload error"),
214	_ENTRY(rxonperrshort,		MISC, NULL, "rx onp fixup short error"),
215	_ENTRY(rxonperrpreamble,	MISC, NULL, "rx onp fixup preamble error"),
216	_ENTRY(rxonperrcrc,		MISC, NULL, "rx onp fixup crc error"),
217	_ENTRY(rxonperraddress,		MISC, NULL, "rx onp fixup address error"),
218	_ENTRY(rxonponp,		MISC, NULL, "rx onp fixup onp packets"),
219	_ENTRY(rxonpok,			MISC, NULL, "rx onp fixup success packets"),
220	_ENTRY(tx,			MISC, NULL, "tx"),
221	_ENTRY(txadd,			MISC, NULL, "tx add"),
222	_ENTRY(txbufcballoc,		MISC, NULL, "tx buf cb alloc"),
223	_ENTRY(txbufcbget,		MISC, NULL, "tx buf cb get"),
224	_ENTRY(txbufgballoc,		MISC, NULL, "tx buf gb alloc"),
225	_ENTRY(txbufgbget,		MISC, NULL, "tx buf gb get"),
226	_ENTRY(txbufgbput,		MISC, NULL, "tx buf gb put"),
227	_ENTRY(txdel,			MISC, NULL, "tx del"),
228	_ENTRY(txerr,			MISC, NULL, "tx error"),
229	_ENTRY(txerrcmd,		MISC, NULL, "tx cmd error"),
230	_ENTRY(txerrgbuf,		MISC, NULL, "tx gbuf error"),
231	_ENTRY(txerrlink,		MISC, NULL, "tx link error"),
232	_ENTRY(txerrmkcmd,		MISC, NULL, "tx makecmd error"),
233#undef	_ENTRY
234};
235#endif
236
237/* ---- buffer management */
238
239static const struct octeon_eth_pool_param {
240	int			poolno;
241	size_t			size;
242	size_t			nelems;
243} octeon_eth_pool_params[] = {
244#define	_ENTRY(x)	{ OCTEON_POOL_NO_##x, OCTEON_POOL_SIZE_##x, OCTEON_POOL_NELEMS_##x }
245	_ENTRY(PKT),
246	_ENTRY(WQE),
247	_ENTRY(CMD),
248	_ENTRY(SG)
249#undef	_ENTRY
250};
251struct octeon_fpa_buf	*octeon_eth_pools[8/* XXX */];
252#define	octeon_eth_fb_pkt	octeon_eth_pools[OCTEON_POOL_NO_PKT]
253#define	octeon_eth_fb_wqe	octeon_eth_pools[OCTEON_POOL_NO_WQE]
254#define	octeon_eth_fb_cmd	octeon_eth_pools[OCTEON_POOL_NO_CMD]
255#define	octeon_eth_fb_sg	octeon_eth_pools[OCTEON_POOL_NO_SG]
256
257static void
258octeon_eth_buf_init(struct octeon_eth_softc *sc)
259{
260	static int once;
261	int i;
262	const struct octeon_eth_pool_param *pp;
263	struct octeon_fpa_buf *fb;
264
265	if (once == 1)
266		return;
267	once = 1;
268
269	for (i = 0; i < (int)__arraycount(octeon_eth_pool_params); i++) {
270		pp = &octeon_eth_pool_params[i];
271		octeon_fpa_buf_init(pp->poolno, pp->size, pp->nelems, &fb);
272		octeon_eth_pools[i] = fb;
273	}
274}
275
276/* ---- autoconf */
277
278static int
279octeon_eth_match(device_t parent, struct cfdata *match, void *aux)
280{
281	struct octeon_gmx_attach_args *ga = aux;
282
283	if (strcmp(match->cf_name, ga->ga_name) != 0) {
284		return 0;
285	}
286	return 1;
287}
288
289static void
290octeon_eth_attach(device_t parent, device_t self, void *aux)
291{
292	struct octeon_eth_softc *sc = device_private(self);
293	struct octeon_gmx_attach_args *ga = aux;
294	struct ifnet *ifp = &sc->sc_ethercom.ec_if;
295	uint8_t enaddr[ETHER_ADDR_LEN];
296
297	sc->sc_dev = self;
298	sc->sc_regt = ga->ga_regt;
299	sc->sc_port = ga->ga_portno;
300	sc->sc_port_type = ga->ga_port_type;
301	sc->sc_gmx = ga->ga_gmx;
302	sc->sc_gmx_port = ga->ga_gmx_port;
303
304	sc->sc_init_flag = 0;
305	/*
306	 * XXXUEBAYASI
307	 * Setting PIP_IP_OFFSET[OFFSET] to 8 causes panic ... why???
308	 */
309	sc->sc_ip_offset = 0/* XXX */;
310
311	if (MIPS_PRID_IMPL(mips_options.mips_cpu_id) <= MIPS_CN30XX) {
312		SET(sc->sc_quirks, OCTEON_ETH_QUIRKS_NO_PRE_ALIGN);
313		SET(sc->sc_quirks, OCTEON_ETH_QUIRKS_NO_RX_INBND);
314	}
315
316	octeon_eth_board_mac_addr(enaddr, sizeof(enaddr), sc);
317	printf("%s: Ethernet address %s\n", device_xname(sc->sc_dev),
318	    ether_sprintf(enaddr));
319
320	octeon_eth_gsc[sc->sc_port] = sc;
321
322	SIMPLEQ_INIT(&sc->sc_sendq);
323	sc->sc_soft_req_thresh = 15/* XXX */;
324	sc->sc_ext_callback_cnt = 0;
325
326	octeon_gmx_stats_init(sc->sc_gmx_port);
327
328	callout_init(&sc->sc_tick_misc_ch, 0);
329	callout_init(&sc->sc_tick_free_ch, 0);
330
331	octeon_fau_op_init(&sc->sc_fau_done,
332	    OCTEON_CVMSEG_ETHER_OFFSET(sc->sc_port, csm_ether_fau_done),
333	    OCT_FAU_REG_ADDR_END - (8 * (sc->sc_port + 1))/* XXX */);
334	octeon_fau_op_set_8(&sc->sc_fau_done, 0);
335
336	octeon_eth_pip_init(sc);
337	octeon_eth_ipd_init(sc);
338	octeon_eth_pko_init(sc);
339	octeon_eth_asx_init(sc);
340	octeon_eth_smi_init(sc);
341
342	sc->sc_gmx_port->sc_ipd = sc->sc_ipd;
343	sc->sc_gmx_port->sc_port_asx = sc->sc_asx;
344	sc->sc_gmx_port->sc_port_mii = &sc->sc_mii;
345	sc->sc_gmx_port->sc_port_ec = &sc->sc_ethercom;
346	/* XXX */
347	sc->sc_gmx_port->sc_quirks = sc->sc_quirks;
348
349	/* XXX */
350	sc->sc_pow = &octeon_pow_softc;
351
352	octeon_eth_mediainit(sc);
353
354	strncpy(ifp->if_xname, device_xname(sc->sc_dev), sizeof(ifp->if_xname));
355	ifp->if_softc = sc;
356	ifp->if_flags = IFF_BROADCAST | IFF_SIMPLEX | IFF_MULTICAST;
357	ifp->if_ioctl = octeon_eth_ioctl;
358	ifp->if_start = octeon_eth_start;
359	ifp->if_watchdog = octeon_eth_watchdog;
360	ifp->if_init = octeon_eth_init;
361	ifp->if_stop = octeon_eth_stop;
362	IFQ_SET_MAXLEN(&ifp->if_snd, uimax(GATHER_QUEUE_SIZE, IFQ_MAXLEN));
363	IFQ_SET_READY(&ifp->if_snd);
364
365	/* XXX: not yet tx checksum */
366	ifp->if_capabilities =
367	    IFCAP_CSUM_IPv4_Rx | IFCAP_CSUM_TCPv4_Rx | IFCAP_CSUM_UDPv4_Rx |
368	    IFCAP_CSUM_TCPv6_Rx | IFCAP_CSUM_UDPv6_Rx;
369
370	/* 802.1Q VLAN-sized frames are supported */
371	sc->sc_ethercom.ec_capabilities |= ETHERCAP_VLAN_MTU;
372
373	octeon_gmx_set_mac_addr(sc->sc_gmx_port, enaddr);
374
375	if_attach(ifp);
376	ether_ifattach(ifp, enaddr);
377	octeon_gmx_set_filter(sc->sc_gmx_port);
378
379	/* XXX */
380	sc->sc_rate_recv_check_link_cap.tv_sec = 1;
381	sc->sc_rate_recv_check_jumbo_cap.tv_sec = 1;
382	sc->sc_rate_recv_check_code_cap.tv_sec = 1;
383	sc->sc_rate_recv_fixup_odd_nibble_short_cap.tv_sec = 1;
384	sc->sc_rate_recv_fixup_odd_nibble_preamble_cap.tv_sec = 1;
385	sc->sc_rate_recv_fixup_odd_nibble_crc_cap.tv_sec = 1;
386#ifdef OCTEON_ETH_DEBUG
387	sc->sc_rate_recv_fixup_odd_nibble_addr_cap.tv_sec = 1;
388#endif
389	/* XXX */
390
391#if 1
392	octeon_eth_buf_init(sc);
393#endif
394
395	if (octeon_eth_pow_recv_ih == NULL)
396		octeon_eth_pow_recv_ih
397		    = octeon_pow_intr_establish(OCTEON_POW_GROUP_PIP,
398			IPL_NET, octeon_eth_recv_intr, NULL, NULL);
399
400	OCTEON_EVCNT_ATTACH_EVCNTS(sc, octeon_evcnt_entries,
401	    device_xname(sc->sc_dev));
402}
403
404/* ---- submodules */
405
406/* XXX */
407static void
408octeon_eth_pip_init(struct octeon_eth_softc *sc)
409{
410	struct octeon_pip_attach_args pip_aa;
411
412	pip_aa.aa_port = sc->sc_port;
413	pip_aa.aa_regt = sc->sc_regt;
414	pip_aa.aa_tag_type = POW_TAG_TYPE_ORDERED/* XXX */;
415	pip_aa.aa_receive_group = OCTEON_POW_GROUP_PIP;
416	pip_aa.aa_ip_offset = sc->sc_ip_offset;
417	octeon_pip_init(&pip_aa, &sc->sc_pip);
418}
419
420/* XXX */
421static void
422octeon_eth_ipd_init(struct octeon_eth_softc *sc)
423{
424	struct octeon_ipd_attach_args ipd_aa;
425
426	ipd_aa.aa_port = sc->sc_port;
427	ipd_aa.aa_regt = sc->sc_regt;
428	ipd_aa.aa_first_mbuff_skip = 184/* XXX */;
429	ipd_aa.aa_not_first_mbuff_skip = 0/* XXX */;
430	octeon_ipd_init(&ipd_aa, &sc->sc_ipd);
431}
432
433/* XXX */
434static void
435octeon_eth_pko_init(struct octeon_eth_softc *sc)
436{
437	struct octeon_pko_attach_args pko_aa;
438
439	pko_aa.aa_port = sc->sc_port;
440	pko_aa.aa_regt = sc->sc_regt;
441	pko_aa.aa_cmdptr = &sc->sc_cmdptr;
442	pko_aa.aa_cmd_buf_pool = OCTEON_POOL_NO_CMD;
443	pko_aa.aa_cmd_buf_size = OCTEON_POOL_NWORDS_CMD;
444	octeon_pko_init(&pko_aa, &sc->sc_pko);
445}
446
447/* XXX */
448static void
449octeon_eth_asx_init(struct octeon_eth_softc *sc)
450{
451	struct octeon_asx_attach_args asx_aa;
452
453	asx_aa.aa_port = sc->sc_port;
454	asx_aa.aa_regt = sc->sc_regt;
455	octeon_asx_init(&asx_aa, &sc->sc_asx);
456}
457
458static void
459octeon_eth_smi_init(struct octeon_eth_softc *sc)
460{
461	struct octeon_smi_attach_args smi_aa;
462
463	smi_aa.aa_port = sc->sc_port;
464	smi_aa.aa_regt = sc->sc_regt;
465	octeon_smi_init(&smi_aa, &sc->sc_smi);
466	octeon_smi_set_clock(sc->sc_smi, 0x1464ULL); /* XXX */
467}
468
469/* ---- XXX */
470
471#define	ADDR2UINT64(u, a) \
472	do { \
473		u = \
474		    (((uint64_t)a[0] << 40) | ((uint64_t)a[1] << 32) | \
475		     ((uint64_t)a[2] << 24) | ((uint64_t)a[3] << 16) | \
476		     ((uint64_t)a[4] <<	 8) | ((uint64_t)a[5] <<  0)); \
477	} while (0)
478#define	UINT642ADDR(a, u) \
479	do { \
480		a[0] = (uint8_t)((u) >> 40); a[1] = (uint8_t)((u) >> 32); \
481		a[2] = (uint8_t)((u) >> 24); a[3] = (uint8_t)((u) >> 16); \
482		a[4] = (uint8_t)((u) >>	 8); a[5] = (uint8_t)((u) >>  0); \
483	} while (0)
484
485static void
486octeon_eth_board_mac_addr(uint8_t *enaddr, size_t size,
487    struct octeon_eth_softc *sc)
488{
489	prop_dictionary_t dict;
490	prop_data_t ea;
491
492	dict = device_properties(sc->sc_dev);
493	KASSERT(dict != NULL);
494	ea = prop_dictionary_get(dict, "mac-address");
495	KASSERT(ea != NULL);
496	memcpy(enaddr, prop_data_data_nocopy(ea), size);
497}
498
499/* ---- media */
500
501static int
502octeon_eth_mii_readreg(device_t self, int phy_addr, int reg, uint16_t *val)
503{
504	struct octeon_eth_softc *sc = device_private(self);
505
506	return octeon_smi_read(sc->sc_smi, phy_addr, reg, val);
507}
508
509static int
510octeon_eth_mii_writereg(device_t self, int phy_addr, int reg, uint16_t val)
511{
512	struct octeon_eth_softc *sc = device_private(self);
513
514	return octeon_smi_write(sc->sc_smi, phy_addr, reg, val);
515}
516
517static void
518octeon_eth_mii_statchg(struct ifnet *ifp)
519{
520	struct octeon_eth_softc *sc = ifp->if_softc;
521
522	octeon_pko_port_enable(sc->sc_pko, 0);
523	octeon_gmx_port_enable(sc->sc_gmx_port, 0);
524
525	octeon_eth_reset(sc);
526
527	if (ISSET(ifp->if_flags, IFF_RUNNING))
528		octeon_gmx_set_filter(sc->sc_gmx_port);
529
530	octeon_pko_port_enable(sc->sc_pko, 1);
531	octeon_gmx_port_enable(sc->sc_gmx_port, 1);
532}
533
534static int
535octeon_eth_mediainit(struct octeon_eth_softc *sc)
536{
537	struct ifnet *ifp = &sc->sc_ethercom.ec_if;
538	struct mii_data *mii = &sc->sc_mii;
539	prop_object_t phy;
540
541	mii->mii_ifp = ifp;
542	mii->mii_readreg = octeon_eth_mii_readreg;
543	mii->mii_writereg = octeon_eth_mii_writereg;
544	mii->mii_statchg = octeon_eth_mii_statchg;
545	sc->sc_ethercom.ec_mii = mii;
546
547	/* Initialize ifmedia structures. */
548	ifmedia_init(&mii->mii_media, 0, ether_mediachange,
549	    octeon_eth_mediastatus);
550
551	phy = prop_dictionary_get(device_properties(sc->sc_dev), "phy-addr");
552	KASSERT(phy != NULL);
553
554	mii_attach(sc->sc_dev, mii, 0xffffffff, prop_number_integer_value(phy),
555	    MII_OFFSET_ANY, MIIF_DOPAUSE);
556
557	/* XXX XXX XXX */
558	if (LIST_FIRST(&mii->mii_phys) != NULL) {
559		/* XXX XXX XXX */
560		ifmedia_set(&mii->mii_media, IFM_ETHER | IFM_AUTO);
561		/* XXX XXX XXX */
562	} else {
563		/* XXX XXX XXX */
564		ifmedia_add(&mii->mii_media, IFM_ETHER | IFM_NONE,
565		    MII_MEDIA_NONE, NULL);
566		/* XXX XXX XXX */
567		ifmedia_set(&mii->mii_media, IFM_ETHER | IFM_NONE);
568		/* XXX XXX XXX */
569	}
570	/* XXX XXX XXX */
571
572	return 0;
573}
574
575static void
576octeon_eth_mediastatus(struct ifnet *ifp, struct ifmediareq *ifmr)
577{
578	struct octeon_eth_softc *sc = ifp->if_softc;
579
580	mii_pollstat(&sc->sc_mii);
581
582	ifmr->ifm_status = sc->sc_mii.mii_media_status;
583	ifmr->ifm_active = sc->sc_mii.mii_media_active;
584	ifmr->ifm_active = (sc->sc_mii.mii_media_active & ~IFM_ETH_FMASK) |
585	    sc->sc_gmx_port->sc_port_flowflags;
586}
587
588/* ---- send buffer garbage collection */
589
590static inline void
591octeon_eth_send_queue_flush_prefetch(struct octeon_eth_softc *sc)
592{
593	OCTEON_ETH_KASSERT(sc->sc_prefetch == 0);
594	octeon_fau_op_inc_fetch_8(&sc->sc_fau_done, 0);
595	sc->sc_prefetch = 1;
596}
597
598static inline void
599octeon_eth_send_queue_flush_fetch(struct octeon_eth_softc *sc)
600{
601#ifndef	 OCTEON_ETH_DEBUG
602	if (!sc->sc_prefetch)
603		return;
604#endif
605	OCTEON_ETH_KASSERT(sc->sc_prefetch == 1);
606	sc->sc_hard_done_cnt = octeon_fau_op_inc_read_8(&sc->sc_fau_done);
607	OCTEON_ETH_KASSERT(sc->sc_hard_done_cnt <= 0);
608	sc->sc_prefetch = 0;
609}
610
611static inline void
612octeon_eth_send_queue_flush(struct octeon_eth_softc *sc)
613{
614	struct ifnet *ifp = &sc->sc_ethercom.ec_if;
615	const int64_t sent_count = sc->sc_hard_done_cnt;
616	int i;
617
618	OCTEON_ETH_KASSERT(sc->sc_flush == 0);
619	OCTEON_ETH_KASSERT(sent_count <= 0);
620
621	for (i = 0; i < 0 - sent_count; i++) {
622		struct mbuf *m;
623		uint64_t *gbuf;
624
625		octeon_eth_send_queue_del(sc, &m, &gbuf);
626
627		octeon_fpa_buf_put(octeon_eth_fb_sg, gbuf);
628		OCTEON_EVCNT_INC(sc, txbufgbput);
629
630		m_freem(m);
631
632		CLR(ifp->if_flags, IFF_OACTIVE);
633	}
634
635	octeon_fau_op_inc_fetch_8(&sc->sc_fau_done, i);
636	sc->sc_flush = i;
637}
638
639static inline void
640octeon_eth_send_queue_flush_sync(struct octeon_eth_softc *sc)
641{
642	if (sc->sc_flush == 0)
643		return;
644
645	OCTEON_ETH_KASSERT(sc->sc_flush > 0);
646
647	/* XXX XXX XXX */
648	octeon_fau_op_inc_read_8(&sc->sc_fau_done);
649	sc->sc_soft_req_cnt -= sc->sc_flush;
650	OCTEON_ETH_KASSERT(sc->sc_soft_req_cnt >= 0);
651	/* XXX XXX XXX */
652
653	sc->sc_flush = 0;
654}
655
656static inline int
657octeon_eth_send_queue_is_full(struct octeon_eth_softc *sc)
658{
659#ifdef OCTEON_ETH_SEND_QUEUE_CHECK
660	int64_t nofree_cnt;
661
662	nofree_cnt = sc->sc_soft_req_cnt + sc->sc_hard_done_cnt;
663
664	if (__predict_false(nofree_cnt == GATHER_QUEUE_SIZE - 1)) {
665		octeon_eth_send_queue_flush(sc);
666		OCTEON_EVCNT_INC(sc, txerrgbuf);
667		octeon_eth_send_queue_flush_sync(sc);
668		return 1;
669	}
670
671#endif
672	return 0;
673}
674
675/*
676 * (Ab)use m_nextpkt and m_paddr to maintain mbuf chain and pointer to gather
677 * buffer.  Other mbuf members may be used by m_freem(), so don't touch them!
678 */
679
680struct _send_queue_entry {
681	union {
682		struct mbuf _sqe_s_mbuf;
683		struct {
684			char _sqe_s_entry_pad[offsetof(struct mbuf, m_nextpkt)];
685			SIMPLEQ_ENTRY(_send_queue_entry) _sqe_s_entry_entry;
686		} _sqe_s_entry;
687		struct {
688			char _sqe_s_gbuf_pad[offsetof(struct mbuf, m_paddr)];
689			uint64_t *_sqe_s_gbuf_gbuf;
690		} _sqe_s_gbuf;
691	} _sqe_u;
692#define	_sqe_entry	_sqe_u._sqe_s_entry._sqe_s_entry_entry
693#define	_sqe_gbuf	_sqe_u._sqe_s_gbuf._sqe_s_gbuf_gbuf
694};
695
696static inline void
697octeon_eth_send_queue_add(struct octeon_eth_softc *sc, struct mbuf *m,
698    uint64_t *gbuf)
699{
700	struct _send_queue_entry *sqe = (struct _send_queue_entry *)m;
701
702	sqe->_sqe_gbuf = gbuf;
703	SIMPLEQ_INSERT_TAIL(&sc->sc_sendq, sqe, _sqe_entry);
704
705	if ((m->m_flags & M_EXT) && m->m_ext.ext_free != NULL)
706		sc->sc_ext_callback_cnt++;
707
708	OCTEON_EVCNT_INC(sc, txadd);
709}
710
711static inline void
712octeon_eth_send_queue_del(struct octeon_eth_softc *sc, struct mbuf **rm,
713    uint64_t **rgbuf)
714{
715	struct _send_queue_entry *sqe;
716
717	sqe = SIMPLEQ_FIRST(&sc->sc_sendq);
718	OCTEON_ETH_KASSERT(sqe != NULL);
719	SIMPLEQ_REMOVE_HEAD(&sc->sc_sendq, _sqe_entry);
720
721	*rm = (void *)sqe;
722	*rgbuf = sqe->_sqe_gbuf;
723
724	if (((*rm)->m_flags & M_EXT) && (*rm)->m_ext.ext_free != NULL) {
725		sc->sc_ext_callback_cnt--;
726		OCTEON_ETH_KASSERT(sc->sc_ext_callback_cnt >= 0);
727	}
728
729	OCTEON_EVCNT_INC(sc, txdel);
730}
731
732static inline int
733octeon_eth_buf_free_work(struct octeon_eth_softc *sc, uint64_t *work,
734    uint64_t word2)
735{
736	/* XXX when jumbo frame */
737	if (ISSET(word2, PIP_WQE_WORD2_IP_BUFS)) {
738		paddr_t addr;
739		paddr_t start_buffer;
740
741		addr = work[3] & PIP_WQE_WORD3_ADDR;
742		start_buffer = addr & ~(2048 - 1);
743
744		octeon_fpa_buf_put_paddr(octeon_eth_fb_pkt, start_buffer);
745		OCTEON_EVCNT_INC(sc, rxbufpkput);
746	}
747
748	octeon_fpa_buf_put(octeon_eth_fb_wqe, work);
749	OCTEON_EVCNT_INC(sc, rxbufwqput);
750
751	return 0;
752}
753
754static inline void
755octeon_eth_buf_ext_free_m(struct mbuf *m, void *buf, size_t size, void *arg)
756{
757	uint64_t *work = (void *)arg;
758#ifdef OCTEON_ETH_DEBUG
759	struct octeon_eth_softc *sc = (void *)(uintptr_t)work[0];
760#endif
761	int s = splnet();
762
763	OCTEON_EVCNT_INC(sc, rxrs);
764
765	octeon_fpa_buf_put(octeon_eth_fb_wqe, work);
766	OCTEON_EVCNT_INC(sc, rxbufwqput);
767
768	OCTEON_ETH_KASSERT(m != NULL);
769
770	pool_cache_put(mb_cache, m);
771
772	splx(s);
773}
774
775static inline void
776octeon_eth_buf_ext_free_ext(struct mbuf *m, void *buf, size_t size, void *arg)
777{
778	uint64_t *work = (void *)arg;
779#ifdef OCTEON_ETH_DEBUG
780	struct octeon_eth_softc *sc = (void *)(uintptr_t)work[0];
781#endif
782	int s = splnet();
783
784	octeon_fpa_buf_put(octeon_eth_fb_wqe, work);
785	OCTEON_EVCNT_INC(sc, rxbufwqput);
786
787	octeon_fpa_buf_put(octeon_eth_fb_pkt, buf);
788	OCTEON_EVCNT_INC(sc, rxbufpkput);
789
790	OCTEON_ETH_KASSERT(m != NULL);
791
792	pool_cache_put(mb_cache, m);
793
794	splx(s);
795}
796
797/* ---- ifnet interfaces */
798
799static int
800octeon_eth_ioctl(struct ifnet *ifp, u_long cmd, void *data)
801{
802	struct octeon_eth_softc *sc = ifp->if_softc;
803	struct ifreq *ifr = (struct ifreq *)data;
804	int s, error;
805
806	s = splnet();
807	switch (cmd) {
808	case SIOCSIFMEDIA:
809		/* Flow control requires full-duplex mode. */
810		if (IFM_SUBTYPE(ifr->ifr_media) == IFM_AUTO ||
811		    (ifr->ifr_media & IFM_FDX) == 0) {
812			ifr->ifr_media &= ~IFM_ETH_FMASK;
813		}
814		if (IFM_SUBTYPE(ifr->ifr_media) != IFM_AUTO) {
815			if ((ifr->ifr_media & IFM_ETH_FMASK) == IFM_FLOW) {
816				ifr->ifr_media |=
817				    IFM_ETH_TXPAUSE | IFM_ETH_RXPAUSE;
818			}
819			sc->sc_gmx_port->sc_port_flowflags =
820				ifr->ifr_media & IFM_ETH_FMASK;
821		}
822		error = ifmedia_ioctl(ifp, ifr, &sc->sc_mii.mii_media, cmd);
823		break;
824	default:
825		error = ether_ioctl(ifp, cmd, data);
826		if (error == ENETRESET) {
827			/*
828			 * Multicast list has changed; set the hardware filter
829			 * accordingly.
830			 */
831			if (ISSET(ifp->if_flags, IFF_RUNNING))
832				octeon_gmx_set_filter(sc->sc_gmx_port);
833			error = 0;
834		}
835		break;
836	}
837	octeon_eth_start(ifp);
838	splx(s);
839
840	return error;
841}
842
843/* ---- send (output) */
844
845static inline uint64_t
846octeon_eth_send_makecmd_w0(uint64_t fau0, uint64_t fau1, size_t len, int segs)
847{
848	return octeon_pko_cmd_word0(
849		OCT_FAU_OP_SIZE_64,		/* sz1 */
850		OCT_FAU_OP_SIZE_64,		/* sz0 */
851		1, fau1, 1, fau0,		/* s1, reg1, s0, reg0 */
852		0,				/* le */
853		octeon_eth_param_pko_cmd_w0_n2,	/* n2 */
854		1, 0,				/* q, r */
855		(segs == 1) ? 0 : 1,		/* g */
856		0, 0, 1,			/* ipoffp1, ii, df */
857		segs, (int)len);		/* segs, totalbytes */
858}
859
860static inline uint64_t
861octeon_eth_send_makecmd_w1(int size, paddr_t addr)
862{
863	return octeon_pko_cmd_word1(
864		0, 0,				/* i, back */
865		FPA_GATHER_BUFFER_POOL,		/* pool */
866		size, addr);			/* size, addr */
867}
868
869static inline int
870octeon_eth_send_makecmd_gbuf(struct octeon_eth_softc *sc, struct mbuf *m0,
871    uint64_t *gbuf, int *rsegs)
872{
873	struct mbuf *m;
874	int segs = 0;
875	uintptr_t laddr, rlen, nlen;
876
877	for (m = m0; m != NULL; m = m->m_next) {
878
879		if (__predict_false(m->m_len == 0))
880			continue;
881
882#if 0
883		OCTEON_ETH_KASSERT(((uint32_t)m->m_data & (PAGE_SIZE - 1))
884		   == (kvtophys((vaddr_t)m->m_data) & (PAGE_SIZE - 1)));
885#endif
886
887		/* Aligned 4k */
888		laddr = (uintptr_t)m->m_data & (PAGE_SIZE - 1);
889
890		if (laddr + m->m_len > PAGE_SIZE) {
891			/* XXX XXX XXX */
892			rlen = PAGE_SIZE - laddr;
893			nlen = m->m_len - rlen;
894			*(gbuf + segs) = octeon_eth_send_makecmd_w1(rlen,
895			    kvtophys((vaddr_t)m->m_data));
896			segs++;
897			if (segs > 63) {
898				return 1;
899			}
900			/* XXX XXX XXX */
901		} else {
902			rlen = 0;
903			nlen = m->m_len;
904		}
905
906		*(gbuf + segs) = octeon_eth_send_makecmd_w1(nlen,
907		    kvtophys((vaddr_t)(m->m_data + rlen)));
908		segs++;
909		if (segs > 63) {
910			return 1;
911		}
912	}
913
914	OCTEON_ETH_KASSERT(m == NULL);
915
916	*rsegs = segs;
917
918	return 0;
919}
920
921static inline int
922octeon_eth_send_makecmd(struct octeon_eth_softc *sc, struct mbuf *m,
923    uint64_t *gbuf, uint64_t *rpko_cmd_w0, uint64_t *rpko_cmd_w1)
924{
925	uint64_t pko_cmd_w0, pko_cmd_w1;
926	int segs;
927	int result = 0;
928
929	if (octeon_eth_send_makecmd_gbuf(sc, m, gbuf, &segs)) {
930		log(LOG_WARNING, "%s: there are a lot of number of segments"
931		    " of transmission data", device_xname(sc->sc_dev));
932		result = 1;
933		goto done;
934	}
935
936	/*
937	 * segs == 1	-> link mode (single continuous buffer)
938	 *		   WORD1[size] is number of bytes pointed by segment
939	 *
940	 * segs > 1	-> gather mode (scatter-gather buffer)
941	 *		   WORD1[size] is number of segments
942	 */
943	pko_cmd_w0 = octeon_eth_send_makecmd_w0(sc->sc_fau_done.fd_regno,
944	    0, m->m_pkthdr.len, segs);
945	if (segs == 1) {
946		pko_cmd_w1 = octeon_eth_send_makecmd_w1(
947		    m->m_pkthdr.len, kvtophys((vaddr_t)m->m_data));
948	} else {
949#ifdef __mips_n32
950		KASSERT(MIPS_KSEG0_P(gbuf));
951		pko_cmd_w1 = octeon_eth_send_makecmd_w1(segs,
952		    MIPS_KSEG0_TO_PHYS(gbuf));
953#else
954		pko_cmd_w1 = octeon_eth_send_makecmd_w1(segs,
955		    MIPS_XKPHYS_TO_PHYS(gbuf));
956#endif
957	}
958
959	*rpko_cmd_w0 = pko_cmd_w0;
960	*rpko_cmd_w1 = pko_cmd_w1;
961
962done:
963	return result;
964}
965
966static inline int
967octeon_eth_send_cmd(struct octeon_eth_softc *sc, uint64_t pko_cmd_w0,
968    uint64_t pko_cmd_w1, int *pwdc)
969{
970	uint64_t *cmdptr;
971	int result = 0;
972
973#ifdef __mips_n32
974	KASSERT((sc->sc_cmdptr.cmdptr & ~MIPS_PHYS_MASK) == 0);
975	cmdptr = (uint64_t *)MIPS_PHYS_TO_KSEG0(sc->sc_cmdptr.cmdptr);
976#else
977	cmdptr = (uint64_t *)MIPS_PHYS_TO_XKPHYS_CACHED(sc->sc_cmdptr.cmdptr);
978#endif
979	cmdptr += sc->sc_cmdptr.cmdptr_idx;
980
981	OCTEON_ETH_KASSERT(cmdptr != NULL);
982
983	*cmdptr++ = pko_cmd_w0;
984	*cmdptr++ = pko_cmd_w1;
985
986	OCTEON_ETH_KASSERT(sc->sc_cmdptr.cmdptr_idx + 2 <= FPA_COMMAND_BUFFER_POOL_NWORDS - 1);
987
988	if (sc->sc_cmdptr.cmdptr_idx + 2 == FPA_COMMAND_BUFFER_POOL_NWORDS - 1) {
989		paddr_t buf;
990
991		buf = octeon_fpa_buf_get_paddr(octeon_eth_fb_cmd);
992		if (buf == 0) {
993			log(LOG_WARNING,
994			    "%s: can not allocate command buffer from free pool allocator\n",
995			    device_xname(sc->sc_dev));
996			result = 1;
997			goto done;
998		}
999		OCTEON_EVCNT_INC(sc, txbufcbget);
1000		*cmdptr++ = buf;
1001		sc->sc_cmdptr.cmdptr = (uint64_t)buf;
1002		sc->sc_cmdptr.cmdptr_idx = 0;
1003	} else {
1004		sc->sc_cmdptr.cmdptr_idx += 2;
1005	}
1006
1007	*pwdc += 2;
1008
1009done:
1010	return result;
1011}
1012
1013static inline int
1014octeon_eth_send_buf(struct octeon_eth_softc *sc, struct mbuf *m,
1015    uint64_t *gbuf, int *pwdc)
1016{
1017	int result = 0, error;
1018	uint64_t pko_cmd_w0, pko_cmd_w1;
1019
1020	error = octeon_eth_send_makecmd(sc, m, gbuf, &pko_cmd_w0, &pko_cmd_w1);
1021	if (error != 0) {
1022		/* Already logging */
1023		OCTEON_EVCNT_INC(sc, txerrmkcmd);
1024		result = error;
1025		goto done;
1026	}
1027
1028	error = octeon_eth_send_cmd(sc, pko_cmd_w0, pko_cmd_w1, pwdc);
1029	if (error != 0) {
1030		/* Already logging */
1031		OCTEON_EVCNT_INC(sc, txerrcmd);
1032		result = error;
1033	}
1034
1035done:
1036	return result;
1037}
1038
1039static inline int
1040octeon_eth_send(struct octeon_eth_softc *sc, struct mbuf *m, int *pwdc)
1041{
1042	paddr_t gaddr = 0;
1043	uint64_t *gbuf = NULL;
1044	int result = 0, error;
1045
1046	OCTEON_EVCNT_INC(sc, tx);
1047
1048	gaddr = octeon_fpa_buf_get_paddr(octeon_eth_fb_sg);
1049	if (gaddr == 0) {
1050		log(LOG_WARNING, "%s: can not allocate gather buffer from "
1051		    "free pool allocator\n", device_xname(sc->sc_dev));
1052		OCTEON_EVCNT_INC(sc, txerrgbuf);
1053		result = 1;
1054		goto done;
1055	}
1056	OCTEON_EVCNT_INC(sc, txbufgbget);
1057
1058#ifdef __mips_n32
1059	KASSERT((gaddr & ~MIPS_PHYS_MASK) == 0);
1060	gbuf = (uint64_t *)(uintptr_t)MIPS_PHYS_TO_KSEG0(gaddr);
1061#else
1062	gbuf = (uint64_t *)(uintptr_t)MIPS_PHYS_TO_XKPHYS_CACHED(gaddr);
1063#endif
1064
1065	OCTEON_ETH_KASSERT(gbuf != NULL);
1066
1067	error = octeon_eth_send_buf(sc, m, gbuf, pwdc);
1068	if (error != 0) {
1069		/* Already logging */
1070		octeon_fpa_buf_put_paddr(octeon_eth_fb_sg, gaddr);
1071		OCTEON_EVCNT_INC(sc, txbufgbput);
1072		result = error;
1073		goto done;
1074	}
1075
1076	octeon_eth_send_queue_add(sc, m, gbuf);
1077
1078done:
1079	return result;
1080}
1081
1082static void
1083octeon_eth_start(struct ifnet *ifp)
1084{
1085	struct octeon_eth_softc *sc = ifp->if_softc;
1086	struct mbuf *m;
1087	int wdc = 0;
1088
1089	/*
1090	 * Performance tuning
1091	 * pre-send iobdma request
1092	 */
1093	octeon_eth_send_queue_flush_prefetch(sc);
1094
1095	if ((ifp->if_flags & (IFF_RUNNING | IFF_OACTIVE)) != IFF_RUNNING)
1096		goto last;
1097
1098	/* XXX assume that OCTEON doesn't buffer packets */
1099	if (__predict_false(!octeon_gmx_link_status(sc->sc_gmx_port))) {
1100		/* Dequeue and drop them */
1101		while (1) {
1102			IFQ_DEQUEUE(&ifp->if_snd, m);
1103			if (m == NULL)
1104				break;
1105
1106			m_freem(m);
1107			IF_DROP(&ifp->if_snd);
1108			OCTEON_EVCNT_INC(sc, txerrlink);
1109		}
1110		goto last;
1111	}
1112
1113	for (;;) {
1114		IFQ_POLL(&ifp->if_snd, m);
1115		if (__predict_false(m == NULL))
1116			break;
1117
1118		/* XXX XXX XXX */
1119		octeon_eth_send_queue_flush_fetch(sc);
1120
1121		/*
1122		 * If no free send buffer is available, free all the sent
1123		 * buffers and bail out.
1124		 */
1125		if (octeon_eth_send_queue_is_full(sc)) {
1126			SET(ifp->if_flags, IFF_OACTIVE);
1127			if (wdc > 0)
1128				octeon_pko_op_doorbell_write(sc->sc_port,
1129				    sc->sc_port, wdc);
1130			return;
1131		}
1132		/* XXX XXX XXX */
1133
1134		IFQ_DEQUEUE(&ifp->if_snd, m);
1135
1136		bpf_mtap(ifp, m, BPF_D_OUT);
1137
1138		/* XXX XXX XXX */
1139		if (sc->sc_soft_req_cnt > sc->sc_soft_req_thresh)
1140			octeon_eth_send_queue_flush(sc);
1141		if (octeon_eth_send(sc, m, &wdc)) {
1142			IF_DROP(&ifp->if_snd);
1143			m_freem(m);
1144			log(LOG_WARNING,
1145			  "%s: failed in the transmission of the packet\n",
1146			  device_xname(sc->sc_dev));
1147			OCTEON_EVCNT_INC(sc, txerr);
1148		} else
1149			sc->sc_soft_req_cnt++;
1150
1151		if (sc->sc_flush)
1152			octeon_eth_send_queue_flush_sync(sc);
1153		/* XXX XXX XXX */
1154
1155		/* Send next iobdma request */
1156		octeon_eth_send_queue_flush_prefetch(sc);
1157	}
1158
1159	if (wdc > 0)
1160		octeon_pko_op_doorbell_write(sc->sc_port, sc->sc_port, wdc);
1161
1162/*
1163 * Don't schedule send-buffer-free callout every time - those buffers are freed
1164 * by "free tick".  This makes some packets like NFS slower.
1165 */
1166#ifdef OCTEON_ETH_USENFS
1167	if (__predict_false(sc->sc_ext_callback_cnt > 0)) {
1168		int timo;
1169
1170		/* ??? */
1171		timo = hz - (100 * sc->sc_ext_callback_cnt);
1172		if (timo < 10)
1173			timo = 10;
1174		callout_schedule(&sc->sc_tick_free_ch, timo);
1175	}
1176#endif
1177
1178last:
1179	octeon_eth_send_queue_flush_fetch(sc);
1180}
1181
1182static void
1183octeon_eth_watchdog(struct ifnet *ifp)
1184{
1185	struct octeon_eth_softc *sc = ifp->if_softc;
1186
1187	printf("%s: device timeout\n", device_xname(sc->sc_dev));
1188
1189	octeon_eth_configure(sc);
1190
1191	SET(ifp->if_flags, IFF_RUNNING);
1192	CLR(ifp->if_flags, IFF_OACTIVE);
1193	ifp->if_timer = 0;
1194
1195	octeon_eth_start(ifp);
1196}
1197
1198static int
1199octeon_eth_init(struct ifnet *ifp)
1200{
1201	struct octeon_eth_softc *sc = ifp->if_softc;
1202
1203	/* XXX don't disable commonly used parts!!! XXX */
1204	if (sc->sc_init_flag == 0) {
1205		/* Cancel any pending I/O. */
1206		octeon_eth_stop(ifp, 0);
1207
1208		/* Initialize the device */
1209		octeon_eth_configure(sc);
1210
1211		octeon_pko_enable(sc->sc_pko);
1212		octeon_ipd_enable(sc->sc_ipd);
1213
1214		sc->sc_init_flag = 1;
1215	} else {
1216		octeon_gmx_port_enable(sc->sc_gmx_port, 1);
1217	}
1218	mii_ifmedia_change(&sc->sc_mii);
1219
1220	octeon_gmx_set_filter(sc->sc_gmx_port);
1221
1222	callout_reset(&sc->sc_tick_misc_ch, hz, octeon_eth_tick_misc, sc);
1223	callout_reset(&sc->sc_tick_free_ch, hz, octeon_eth_tick_free, sc);
1224
1225	SET(ifp->if_flags, IFF_RUNNING);
1226	CLR(ifp->if_flags, IFF_OACTIVE);
1227
1228	return 0;
1229}
1230
1231static void
1232octeon_eth_stop(struct ifnet *ifp, int disable)
1233{
1234	struct octeon_eth_softc *sc = ifp->if_softc;
1235
1236	callout_stop(&sc->sc_tick_misc_ch);
1237	callout_stop(&sc->sc_tick_free_ch);
1238
1239	mii_down(&sc->sc_mii);
1240
1241	octeon_gmx_port_enable(sc->sc_gmx_port, 0);
1242
1243	/* Mark the interface as down and cancel the watchdog timer. */
1244	CLR(ifp->if_flags, IFF_RUNNING | IFF_OACTIVE);
1245	ifp->if_timer = 0;
1246}
1247
1248/* ---- misc */
1249
1250#define PKO_INDEX_MASK	((1ULL << 12/* XXX */) - 1)
1251
1252static int
1253octeon_eth_reset(struct octeon_eth_softc *sc)
1254{
1255	octeon_gmx_reset_speed(sc->sc_gmx_port);
1256	octeon_gmx_reset_flowctl(sc->sc_gmx_port);
1257	octeon_gmx_reset_timing(sc->sc_gmx_port);
1258
1259	return 0;
1260}
1261
1262static int
1263octeon_eth_configure(struct octeon_eth_softc *sc)
1264{
1265	octeon_gmx_port_enable(sc->sc_gmx_port, 0);
1266
1267	octeon_eth_reset(sc);
1268
1269	octeon_eth_configure_common(sc);
1270
1271	octeon_pko_port_config(sc->sc_pko);
1272	octeon_pko_port_enable(sc->sc_pko, 1);
1273	octeon_pip_port_config(sc->sc_pip);
1274
1275	octeon_gmx_tx_stats_rd_clr(sc->sc_gmx_port, 1);
1276	octeon_gmx_rx_stats_rd_clr(sc->sc_gmx_port, 1);
1277
1278	octeon_gmx_port_enable(sc->sc_gmx_port, 1);
1279
1280	return 0;
1281}
1282
1283static int
1284octeon_eth_configure_common(struct octeon_eth_softc *sc)
1285{
1286	static int once;
1287
1288	if (once == 1)
1289		return 0;
1290	once = 1;
1291
1292	octeon_ipd_config(sc->sc_ipd);
1293#ifdef OCTEON_ETH_IPD_RED
1294	octeon_ipd_red(sc->sc_ipd, RECV_QUEUE_SIZE >> 2, RECV_QUEUE_SIZE >> 3);
1295#endif
1296	octeon_pko_config(sc->sc_pko);
1297
1298	octeon_pow_config(sc->sc_pow, OCTEON_POW_GROUP_PIP);
1299
1300	return 0;
1301}
1302
1303/* ---- receive (input) */
1304
1305static inline int
1306octeon_eth_recv_mbuf(struct octeon_eth_softc *sc, uint64_t *work,
1307    struct mbuf **rm)
1308{
1309	struct mbuf *m;
1310	void (*ext_free)(struct mbuf *, void *, size_t, void *);
1311	void *ext_buf;
1312	size_t ext_size;
1313	void *data;
1314	uint64_t word1 = work[1];
1315	uint64_t word2 = work[2];
1316	uint64_t word3 = work[3];
1317
1318	MGETHDR(m, M_NOWAIT, MT_DATA);
1319	if (m == NULL)
1320		return 1;
1321	OCTEON_ETH_KASSERT(m != NULL);
1322
1323	if ((word2 & PIP_WQE_WORD2_IP_BUFS) == 0) {
1324		/* Dynamic short */
1325		ext_free = octeon_eth_buf_ext_free_m;
1326		ext_buf = &work[4];
1327		ext_size = 96;
1328
1329		data = &work[4 + sc->sc_ip_offset / sizeof(uint64_t)];
1330	} else {
1331		vaddr_t addr;
1332		vaddr_t start_buffer;
1333
1334#ifdef __mips_n32
1335		KASSERT((word3 & ~MIPS_PHYS_MASK) == 0);
1336		addr = MIPS_PHYS_TO_KSEG0(word3 & PIP_WQE_WORD3_ADDR);
1337#else
1338		addr = MIPS_PHYS_TO_XKPHYS_CACHED(word3 & PIP_WQE_WORD3_ADDR);
1339#endif
1340		start_buffer = addr & ~(2048 - 1);
1341
1342		ext_free = octeon_eth_buf_ext_free_ext;
1343		ext_buf = (void *)start_buffer;
1344		ext_size = 2048;
1345
1346		data = (void *)addr;
1347	}
1348
1349	/* Embed sc pointer into work[0] for _ext_free evcnt */
1350	work[0] = (uintptr_t)sc;
1351
1352	MEXTADD(m, ext_buf, ext_size, 0, ext_free, work);
1353	OCTEON_ETH_KASSERT(ISSET(m->m_flags, M_EXT));
1354
1355	m->m_data = data;
1356	m->m_len = m->m_pkthdr.len = (word1 & PIP_WQE_WORD1_LEN) >> 48;
1357	m_set_rcvif(m, &sc->sc_ethercom.ec_if);
1358
1359	/* Not readonly buffer */
1360	m->m_flags |= M_EXT_RW;
1361
1362	*rm = m;
1363
1364	OCTEON_ETH_KASSERT(*rm != NULL);
1365
1366	return 0;
1367}
1368
1369static inline int
1370octeon_eth_recv_check_code(struct octeon_eth_softc *sc, uint64_t word2)
1371{
1372	uint64_t opecode = word2 & PIP_WQE_WORD2_NOIP_OPECODE;
1373
1374	if (__predict_true(!ISSET(word2, PIP_WQE_WORD2_NOIP_RE)))
1375		return 0;
1376
1377	/* This error is harmless */
1378	if (opecode == PIP_OVER_ERR)
1379		return 0;
1380
1381	return 1;
1382}
1383
1384static inline int
1385octeon_eth_recv_check_jumbo(struct octeon_eth_softc *sc, uint64_t word2)
1386{
1387	if (__predict_false((word2 & PIP_WQE_WORD2_IP_BUFS) > (1ULL << 56)))
1388		return 1;
1389	return 0;
1390}
1391
1392static inline int
1393octeon_eth_recv_check_link(struct octeon_eth_softc *sc, uint64_t word2)
1394{
1395	if (__predict_false(!octeon_gmx_link_status(sc->sc_gmx_port)))
1396		return 1;
1397	return 0;
1398}
1399
1400static inline int
1401octeon_eth_recv_check(struct octeon_eth_softc *sc, uint64_t word2)
1402{
1403	if (__predict_false(octeon_eth_recv_check_link(sc, word2)) != 0) {
1404		if (ratecheck(&sc->sc_rate_recv_check_link_last,
1405		    &sc->sc_rate_recv_check_link_cap))
1406			log(LOG_DEBUG,
1407			    "%s: link is not up, the packet was dropped\n",
1408			    device_xname(sc->sc_dev));
1409		OCTEON_EVCNT_INC(sc, rxerrlink);
1410		return 1;
1411	}
1412
1413#if 0 /* XXX Performance tuning (Jumbo-frame is not supported yet!) */
1414	if (__predict_false(octeon_eth_recv_check_jumbo(sc, word2)) != 0) {
1415		/* XXX jumbo frame */
1416		if (ratecheck(&sc->sc_rate_recv_check_jumbo_last,
1417		    &sc->sc_rate_recv_check_jumbo_cap))
1418			log(LOG_DEBUG,
1419			    "jumbo frame was received\n");
1420		OCTEON_EVCNT_INC(sc, rxerrjmb);
1421		return 1;
1422	}
1423#endif
1424
1425	if (__predict_false(octeon_eth_recv_check_code(sc, word2)) != 0) {
1426
1427		if ((word2 & PIP_WQE_WORD2_NOIP_OPECODE) ==
1428				PIP_WQE_WORD2_RE_OPCODE_LENGTH) {
1429			/* No logging */
1430			/* XXX increment special error count */
1431		} else if ((word2 & PIP_WQE_WORD2_NOIP_OPECODE) ==
1432				PIP_WQE_WORD2_RE_OPCODE_PARTIAL) {
1433			/* Not an error, it's because of overload */
1434		} else {
1435
1436			if (ratecheck(&sc->sc_rate_recv_check_code_last,
1437			    &sc->sc_rate_recv_check_code_cap))
1438				log(LOG_WARNING,
1439				    "%s: reception error, packet dropped "
1440				    "(error code = %" PRId64 ")\n",
1441				    device_xname(sc->sc_dev), word2 & PIP_WQE_WORD2_NOIP_OPECODE);
1442		}
1443		OCTEON_EVCNT_INC(sc, rxerrcode);
1444		return 1;
1445	}
1446
1447	return 0;
1448}
1449
1450static inline int
1451octeon_eth_recv(struct octeon_eth_softc *sc, uint64_t *work)
1452{
1453	int result = 0;
1454	struct ifnet *ifp;
1455	struct mbuf *m;
1456	uint64_t word2;
1457
1458	/* XXX XXX XXX */
1459	/*
1460	 * Performance tuning
1461	 * pre-send iobdma request
1462	 */
1463	if (sc->sc_soft_req_cnt > sc->sc_soft_req_thresh) {
1464		octeon_eth_send_queue_flush_prefetch(sc);
1465	}
1466	/* XXX XXX XXX */
1467
1468	OCTEON_ETH_KASSERT(sc != NULL);
1469	OCTEON_ETH_KASSERT(work != NULL);
1470
1471	OCTEON_EVCNT_INC(sc, rx);
1472
1473	word2 = work[2];
1474	ifp = &sc->sc_ethercom.ec_if;
1475
1476	OCTEON_ETH_KASSERT(ifp != NULL);
1477
1478	if (__predict_false(octeon_eth_recv_check(sc, word2) != 0)) {
1479		if_statinc(ifp, if_ierrors);
1480		result = 1;
1481		octeon_eth_buf_free_work(sc, work, word2);
1482		goto drop;
1483	}
1484
1485	if (__predict_false(octeon_eth_recv_mbuf(sc, work, &m) != 0)) {
1486		if_statinc(ifp, if_ierrors);
1487		result = 1;
1488		octeon_eth_buf_free_work(sc, work, word2);
1489		goto drop;
1490	}
1491
1492	/* work[0] .. work[3] may not be valid any more */
1493
1494	OCTEON_ETH_KASSERT(m != NULL);
1495
1496	octeon_ipd_offload(word2, m->m_data, &m->m_pkthdr.csum_flags);
1497
1498	/* XXX XXX XXX */
1499	if (sc->sc_soft_req_cnt > sc->sc_soft_req_thresh) {
1500		octeon_eth_send_queue_flush_fetch(sc);
1501		octeon_eth_send_queue_flush(sc);
1502	}
1503
1504	/* XXX XXX XXX */
1505	if (sc->sc_flush)
1506		octeon_eth_send_queue_flush_sync(sc);
1507	/* XXX XXX XXX */
1508
1509	if_percpuq_enqueue(ifp->if_percpuq, m);
1510
1511	return 0;
1512
1513drop:
1514	/* XXX XXX XXX */
1515	if (sc->sc_soft_req_cnt > sc->sc_soft_req_thresh) {
1516		octeon_eth_send_queue_flush_fetch(sc);
1517	}
1518	/* XXX XXX XXX */
1519
1520	return result;
1521}
1522
1523static void
1524octeon_eth_recv_redir(struct ifnet *ifp, struct mbuf *m)
1525{
1526	struct octeon_eth_softc *rsc = ifp->if_softc;
1527	struct octeon_eth_softc *sc = NULL;
1528	int i, wdc = 0;
1529
1530	for (i = 0; i < 3 /* XXX */; i++) {
1531		if (rsc->sc_redir & (1 << i))
1532			sc = octeon_eth_gsc[i];
1533	}
1534
1535	if (sc == NULL) {
1536		m_freem(m);
1537		return;
1538	}
1539	octeon_eth_send_queue_flush_prefetch(sc);
1540
1541	octeon_eth_send_queue_flush_fetch(sc);
1542
1543	if (octeon_eth_send_queue_is_full(sc)) {
1544		m_freem(m);
1545		return;
1546	}
1547	if (sc->sc_soft_req_cnt > sc->sc_soft_req_thresh)
1548		octeon_eth_send_queue_flush(sc);
1549
1550	if (octeon_eth_send(sc, m, &wdc)) {
1551		IF_DROP(&ifp->if_snd);
1552		m_freem(m);
1553	} else {
1554		octeon_pko_op_doorbell_write(sc->sc_port, sc->sc_port, wdc);
1555		sc->sc_soft_req_cnt++;
1556	}
1557
1558	if (sc->sc_flush)
1559		octeon_eth_send_queue_flush_sync(sc);
1560}
1561
1562static inline void
1563octeon_eth_recv_intr(void *data, uint64_t *work)
1564{
1565	struct octeon_eth_softc *sc;
1566	int port;
1567
1568	OCTEON_ETH_KASSERT(work != NULL);
1569
1570	port = (work[1] & PIP_WQE_WORD1_IPRT) >> 42;
1571
1572	OCTEON_ETH_KASSERT(port < GMX_PORT_NUNITS);
1573
1574	sc = octeon_eth_gsc[port];
1575
1576	OCTEON_ETH_KASSERT(sc != NULL);
1577	OCTEON_ETH_KASSERT(port == sc->sc_port);
1578
1579	/* XXX process all work queue entries anyway */
1580
1581	(void)octeon_eth_recv(sc, work);
1582}
1583
1584/* ---- tick */
1585
1586/*
1587 * octeon_eth_tick_free
1588 *
1589 * => garbage collect send gather buffer / mbuf
1590 * => called at softclock
1591 */
1592static void
1593octeon_eth_tick_free(void *arg)
1594{
1595	struct octeon_eth_softc *sc = arg;
1596	int timo;
1597	int s;
1598
1599	s = splnet();
1600	/* XXX XXX XXX */
1601	if (sc->sc_soft_req_cnt > 0) {
1602		octeon_eth_send_queue_flush_prefetch(sc);
1603		octeon_eth_send_queue_flush_fetch(sc);
1604		octeon_eth_send_queue_flush(sc);
1605		octeon_eth_send_queue_flush_sync(sc);
1606	}
1607	/* XXX XXX XXX */
1608
1609	/* XXX XXX XXX */
1610	/* ??? */
1611	timo = hz - (100 * sc->sc_ext_callback_cnt);
1612	if (timo < 10)
1613		 timo = 10;
1614	callout_schedule(&sc->sc_tick_free_ch, timo);
1615	/* XXX XXX XXX */
1616	splx(s);
1617}
1618
1619/*
1620 * octeon_eth_tick_misc
1621 *
1622 * => collect statistics
1623 * => check link status
1624 * => called at softclock
1625 */
1626static void
1627octeon_eth_tick_misc(void *arg)
1628{
1629	struct octeon_eth_softc *sc = arg;
1630	struct ifnet *ifp;
1631	int s;
1632
1633	s = splnet();
1634
1635	ifp = &sc->sc_ethercom.ec_if;
1636
1637	octeon_gmx_stats(sc->sc_gmx_port);
1638	octeon_pip_stats(sc->sc_pip, ifp, sc->sc_port);
1639	mii_tick(&sc->sc_mii);
1640
1641	splx(s);
1642
1643	callout_schedule(&sc->sc_tick_misc_ch, hz);
1644}
1645
1646/* ---- Odd nibble preamble workaround (software CRC processing) */
1647
1648/* ---- sysctl */
1649
1650static int	octeon_eth_sysctl_verify(SYSCTLFN_ARGS);
1651static int	octeon_eth_sysctl_pool(SYSCTLFN_ARGS);
1652static int	octeon_eth_sysctl_rd(SYSCTLFN_ARGS);
1653
1654static int	octeon_eth_sysctl_pkocmdw0n2_num;
1655static int	octeon_eth_sysctl_pipdynrs_num;
1656static int	octeon_eth_sysctl_redir_num;
1657static int	octeon_eth_sysctl_pkt_pool_num;
1658static int	octeon_eth_sysctl_wqe_pool_num;
1659static int	octeon_eth_sysctl_cmd_pool_num;
1660static int	octeon_eth_sysctl_sg_pool_num;
1661static int	octeon_eth_sysctl_pktbuf_num;
1662
1663/*
1664 * Set up sysctl(3) MIB, hw.cnmac.*.
1665 */
1666SYSCTL_SETUP(sysctl_octeon_eth, "sysctl cnmac subtree setup")
1667{
1668	int rc;
1669	int octeon_eth_sysctl_root_num;
1670	const struct sysctlnode *node;
1671
1672	if ((rc = sysctl_createv(clog, 0, NULL, NULL,
1673		    0, CTLTYPE_NODE, "hw", NULL,
1674		    NULL, 0, NULL, 0, CTL_HW, CTL_EOL)) != 0) {
1675		goto err;
1676	}
1677
1678	if ((rc = sysctl_createv(clog, 0, NULL, &node,
1679		    0, CTLTYPE_NODE, "cnmac",
1680		    SYSCTL_DESCR("cnmac interface controls"),
1681		    NULL, 0, NULL, 0, CTL_HW, CTL_CREATE, CTL_EOL)) != 0) {
1682		goto err;
1683	}
1684
1685	octeon_eth_sysctl_root_num = node->sysctl_num;
1686
1687	if ((rc = sysctl_createv(clog, 0, NULL, &node,
1688		    CTLFLAG_PERMANENT | CTLFLAG_READWRITE,
1689		    CTLTYPE_INT, "pko_cmd_w0_n2",
1690		    SYSCTL_DESCR("PKO command WORD0 N2 bit"),
1691		    octeon_eth_sysctl_verify, 0,
1692		    &octeon_eth_param_pko_cmd_w0_n2,
1693		    0, CTL_HW, octeon_eth_sysctl_root_num, CTL_CREATE,
1694		    CTL_EOL)) != 0) {
1695		goto err;
1696	}
1697
1698	octeon_eth_sysctl_pkocmdw0n2_num = node->sysctl_num;
1699
1700	if ((rc = sysctl_createv(clog, 0, NULL, &node,
1701		    CTLFLAG_PERMANENT | CTLFLAG_READWRITE,
1702		    CTLTYPE_INT, "pip_dyn_rs",
1703		    SYSCTL_DESCR("PIP dynamic short in WQE"),
1704		    octeon_eth_sysctl_verify, 0,
1705		    &octeon_eth_param_pip_dyn_rs,
1706		    0, CTL_HW, octeon_eth_sysctl_root_num, CTL_CREATE,
1707		    CTL_EOL)) != 0) {
1708		goto err;
1709	}
1710
1711	octeon_eth_sysctl_pipdynrs_num = node->sysctl_num;
1712
1713	if ((rc = sysctl_createv(clog, 0, NULL, &node,
1714		    CTLFLAG_PERMANENT | CTLFLAG_READWRITE,
1715		    CTLTYPE_INT, "redir",
1716		    SYSCTL_DESCR("input port redirection"),
1717		    octeon_eth_sysctl_verify, 0,
1718		    &octeon_eth_param_redir,
1719		    0, CTL_HW, octeon_eth_sysctl_root_num, CTL_CREATE,
1720		    CTL_EOL)) != 0) {
1721		goto err;
1722	}
1723
1724	octeon_eth_sysctl_redir_num = node->sysctl_num;
1725
1726	if ((rc = sysctl_createv(clog, 0, NULL, &node,
1727		    CTLFLAG_PERMANENT,
1728		    CTLTYPE_INT, "pkt_pool",
1729		    SYSCTL_DESCR("packet pool available"),
1730		    octeon_eth_sysctl_pool, 0, NULL,
1731		    0, CTL_HW, octeon_eth_sysctl_root_num, CTL_CREATE,
1732		    CTL_EOL)) != 0) {
1733		goto err;
1734	}
1735
1736	octeon_eth_sysctl_pkt_pool_num = node->sysctl_num;
1737
1738	if ((rc = sysctl_createv(clog, 0, NULL, &node,
1739		    CTLFLAG_PERMANENT,
1740		    CTLTYPE_INT, "wqe_pool",
1741		    SYSCTL_DESCR("wqe pool available"),
1742		    octeon_eth_sysctl_pool, 0, NULL,
1743		    0, CTL_HW, octeon_eth_sysctl_root_num, CTL_CREATE,
1744		    CTL_EOL)) != 0) {
1745		goto err;
1746	}
1747
1748	octeon_eth_sysctl_wqe_pool_num = node->sysctl_num;
1749
1750	if ((rc = sysctl_createv(clog, 0, NULL, &node,
1751		    CTLFLAG_PERMANENT,
1752		    CTLTYPE_INT, "cmd_pool",
1753		    SYSCTL_DESCR("cmd pool available"),
1754		    octeon_eth_sysctl_pool, 0, NULL,
1755		    0, CTL_HW, octeon_eth_sysctl_root_num, CTL_CREATE,
1756		    CTL_EOL)) != 0) {
1757		goto err;
1758	}
1759
1760	octeon_eth_sysctl_cmd_pool_num = node->sysctl_num;
1761
1762	if ((rc = sysctl_createv(clog, 0, NULL, &node,
1763		    CTLFLAG_PERMANENT,
1764		    CTLTYPE_INT, "sg_pool",
1765		    SYSCTL_DESCR("sg pool available"),
1766		    octeon_eth_sysctl_pool, 0, NULL,
1767		    0, CTL_HW, octeon_eth_sysctl_root_num, CTL_CREATE,
1768		    CTL_EOL)) != 0) {
1769		goto err;
1770	}
1771
1772	octeon_eth_sysctl_sg_pool_num = node->sysctl_num;
1773
1774	if ((rc = sysctl_createv(clog, 0, NULL, &node,
1775		    CTLFLAG_PERMANENT | CTLFLAG_READONLY,
1776		    CTLTYPE_INT, "pktbuf",
1777		    SYSCTL_DESCR("input packet buffer size on POW"),
1778		    octeon_eth_sysctl_rd, 0,
1779		    &octeon_eth_param_pktbuf,
1780		    0, CTL_HW, octeon_eth_sysctl_root_num, CTL_CREATE,
1781		    CTL_EOL)) != 0) {
1782		goto err;
1783	}
1784
1785	octeon_eth_sysctl_pktbuf_num = node->sysctl_num;
1786
1787	return;
1788
1789err:
1790	aprint_error("%s: syctl_createv failed (rc = %d)\n", __func__, rc);
1791}
1792
1793static int
1794octeon_eth_sysctl_verify(SYSCTLFN_ARGS)
1795{
1796	int error, v;
1797	struct sysctlnode node;
1798	struct octeon_eth_softc *sc;
1799	int i;
1800	int s;
1801
1802	node = *rnode;
1803	v = *(int *)rnode->sysctl_data;
1804	node.sysctl_data = &v;
1805	error = sysctl_lookup(SYSCTLFN_CALL(&node));
1806	if (error || newp == NULL)
1807		return error;
1808
1809	if (node.sysctl_num == octeon_eth_sysctl_pkocmdw0n2_num) {
1810		if (v < 0 || v > 1)
1811			return EINVAL;
1812		*(int *)rnode->sysctl_data = v;
1813		return 0;
1814	}
1815
1816	if (node.sysctl_num == octeon_eth_sysctl_pipdynrs_num) {
1817		if (v < 0 || v > 1)
1818			return EINVAL;
1819		*(int *)rnode->sysctl_data = v;
1820		s = splnet();
1821		for (i = 0; i < 3/* XXX */; i++) {
1822			sc = octeon_eth_gsc[i];	/* XXX */
1823			octeon_pip_prt_cfg_enable(sc->sc_pip,
1824			    PIP_PRT_CFGN_DYN_RS, v);
1825		}
1826		splx(s);
1827		return 0;
1828	}
1829
1830	if (node.sysctl_num == octeon_eth_sysctl_redir_num) {
1831		if (v & ~((0x7 << (4 * 0)) | (0x7 << (4 * 1)) | (0x7 << (4 * 2))))
1832			return EINVAL;
1833		*(int *)rnode->sysctl_data = v;
1834		s = splnet();
1835		for (i = 0; i < 3/* XXX */; i++) {
1836			struct ifnet *ifp;
1837
1838			sc = octeon_eth_gsc[i];	/* XXX */
1839			ifp = &sc->sc_ethercom.ec_if;
1840
1841			sc->sc_redir
1842			    = (octeon_eth_param_redir >> (4 * i)) & 0x7;
1843			if (sc->sc_redir == 0) {
1844				if (ISSET(ifp->if_flags, IFF_PROMISC)) {
1845					CLR(ifp->if_flags, IFF_PROMISC);
1846					octeon_eth_mii_statchg(ifp);
1847					/* octeon_gmx_set_filter(sc->sc_gmx_port); */
1848				}
1849				ifp->_if_input = ether_input;
1850			}
1851			else {
1852				if (!ISSET(ifp->if_flags, IFF_PROMISC)) {
1853					SET(ifp->if_flags, IFF_PROMISC);
1854					octeon_eth_mii_statchg(ifp);
1855					/* octeon_gmx_set_filter(sc->sc_gmx_port); */
1856				}
1857				ifp->_if_input = octeon_eth_recv_redir;
1858			}
1859		}
1860		splx(s);
1861		return 0;
1862	}
1863
1864	return EINVAL;
1865}
1866
1867static int
1868octeon_eth_sysctl_pool(SYSCTLFN_ARGS)
1869{
1870	int error, newval = 0;
1871	struct sysctlnode node;
1872	int s;
1873
1874	node = *rnode;
1875	node.sysctl_data = &newval;
1876	s = splnet();
1877	if (node.sysctl_num == octeon_eth_sysctl_pkt_pool_num) {
1878		error = octeon_fpa_available_fpa_pool(&newval,
1879		    OCTEON_POOL_NO_PKT);
1880	} else if (node.sysctl_num == octeon_eth_sysctl_wqe_pool_num) {
1881		error = octeon_fpa_available_fpa_pool(&newval,
1882		    OCTEON_POOL_NO_WQE);
1883	} else if (node.sysctl_num == octeon_eth_sysctl_cmd_pool_num) {
1884		error = octeon_fpa_available_fpa_pool(&newval,
1885		    OCTEON_POOL_NO_CMD);
1886	} else if (node.sysctl_num == octeon_eth_sysctl_sg_pool_num) {
1887		error = octeon_fpa_available_fpa_pool(&newval,
1888		    OCTEON_POOL_NO_SG);
1889	} else {
1890		splx(s);
1891		return EINVAL;
1892	}
1893	splx(s);
1894	if (error)
1895		return error;
1896	error = sysctl_lookup(SYSCTLFN_CALL(&node));
1897	if (error || newp == NULL)
1898		return error;
1899
1900	return 0;
1901}
1902
1903static int
1904octeon_eth_sysctl_rd(SYSCTLFN_ARGS)
1905{
1906	int error, v;
1907	struct sysctlnode node;
1908	int s;
1909
1910	node = *rnode;
1911	v = *(int *)rnode->sysctl_data;
1912	node.sysctl_data = &v;
1913	error = sysctl_lookup(SYSCTLFN_CALL(&node));
1914	if (error || newp != NULL)
1915		return error;
1916
1917	if (node.sysctl_num == octeon_eth_sysctl_pktbuf_num) {
1918		uint64_t tmp;
1919		int n;
1920
1921		s = splnet();
1922		tmp = octeon_fpa_query(0);
1923		n = (int)tmp;
1924		splx(s);
1925		*(int *)rnode->sysctl_data = n;
1926		octeon_eth_param_pktbuf = n;
1927		*(int *)oldp = n;
1928		return 0;
1929	}
1930
1931	return EINVAL;
1932}
1933