if_re.c revision 119976
1/*
2 * Copyright (c) 1997, 1998-2003
3 *	Bill Paul <wpaul@windriver.com>.  All rights reserved.
4 *
5 * Redistribution and use in source and binary forms, with or without
6 * modification, are permitted provided that the following conditions
7 * are met:
8 * 1. Redistributions of source code must retain the above copyright
9 *    notice, this list of conditions and the following disclaimer.
10 * 2. Redistributions in binary form must reproduce the above copyright
11 *    notice, this list of conditions and the following disclaimer in the
12 *    documentation and/or other materials provided with the distribution.
13 * 3. All advertising materials mentioning features or use of this software
14 *    must display the following acknowledgement:
15 *	This product includes software developed by Bill Paul.
16 * 4. Neither the name of the author nor the names of any co-contributors
17 *    may be used to endorse or promote products derived from this software
18 *    without specific prior written permission.
19 *
20 * THIS SOFTWARE IS PROVIDED BY Bill Paul AND CONTRIBUTORS ``AS IS'' AND
21 * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
22 * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
23 * ARE DISCLAIMED.  IN NO EVENT SHALL Bill Paul OR THE VOICES IN HIS HEAD
24 * BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
25 * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
26 * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
27 * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
28 * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
29 * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF
30 * THE POSSIBILITY OF SUCH DAMAGE.
31 */
32
33/*
34 * RealTek 8139C+/8169/8169S/8110S PCI NIC driver
35 *
36 * Written by Bill Paul <wpaul@windriver.com>
37 * Senior Networking Software Engineer
38 * Wind River Systems
39 */
40
41/*
42 * This driver is designed to support RealTek's next generation of
43 * 10/100 and 10/100/1000 PCI ethernet controllers. There are currently
44 * four devices in this family: the RTL8139C+, the RTL8169, the RTL8169S
45 * and the RTL8110S.
46 *
47 * The 8139C+ is a 10/100 ethernet chip. It is backwards compatible
48 * with the older 8139 family, however it also supports a special
49 * C+ mode of operation that provides several new performance enhancing
50 * features. These include:
51 *
52 *	o Descriptor based DMA mechanism. Each descriptor represents
53 *	  a single packet fragment. Data buffers may be aligned on
54 *	  any byte boundary.
55 *
56 *	o 64-bit DMA
57 *
58 *	o TCP/IP checksum offload for both RX and TX
59 *
60 *	o High and normal priority transmit DMA rings
61 *
62 *	o VLAN tag insertion and extraction
63 *
64 *	o TCP large send (segmentation offload)
65 *
66 * Like the 8139, the 8139C+ also has a built-in 10/100 PHY. The C+
67 * programming API is fairly straightforward. The RX filtering, EEPROM
68 * access and PHY access is the same as it is on the older 8139 series
69 * chips.
70 *
71 * The 8169 is a 64-bit 10/100/1000 gigabit ethernet MAC. It has almost the
72 * same programming API and feature set as the 8139C+ with the following
73 * differences and additions:
74 *
75 *	o 1000Mbps mode
76 *
77 *	o Jumbo frames
78 *
79 * 	o GMII and TBI ports/registers for interfacing with copper
80 *	  or fiber PHYs
81 *
82 *      o RX and TX DMA rings can have up to 1024 descriptors
83 *        (the 8139C+ allows a maximum of 64)
84 *
85 *	o Slight differences in register layout from the 8139C+
86 *
87 * The TX start and timer interrupt registers are at different locations
88 * on the 8169 than they are on the 8139C+. Also, the status word in the
89 * RX descriptor has a slightly different bit layout. The 8169 does not
90 * have a built-in PHY. Most reference boards use a Marvell 88E1000 'Alaska'
91 * copper gigE PHY.
92 *
93 * The 8169S/8110S 10/100/1000 devices have built-in copper gigE PHYs
94 * (the 'S' stands for 'single-chip'). These devices have the same
95 * programming API as the older 8169, but also have some vendor-specific
96 * registers for the on-board PHY. The 8110S is a LAN-on-motherboard
97 * part designed to be pin-compatible with the RealTek 8100 10/100 chip.
98 *
99 * This driver takes advantage of the RX and TX checksum offload and
100 * VLAN tag insertion/extraction features. It also implements TX
101 * interrupt moderation using the timer interrupt registers, which
102 * significantly reduces TX interrupt load. There is also support
103 * for jumbo frames, however the 8169/8169S/8110S can not transmit
104 * jumbo frames larger than 7.5K, so the max MTU possible with this
105 * driver is 7500 bytes.
106 */
107
108#include <sys/cdefs.h>
109__FBSDID("$FreeBSD: head/sys/dev/re/if_re.c 119976 2003-09-11 03:53:46Z wpaul $");
110
111#include <sys/param.h>
112#include <sys/endian.h>
113#include <sys/systm.h>
114#include <sys/sockio.h>
115#include <sys/mbuf.h>
116#include <sys/malloc.h>
117#include <sys/kernel.h>
118#include <sys/socket.h>
119
120#include <net/if.h>
121#include <net/if_arp.h>
122#include <net/ethernet.h>
123#include <net/if_dl.h>
124#include <net/if_media.h>
125#include <net/if_vlan_var.h>
126
127#include <net/bpf.h>
128
129#include <machine/bus_pio.h>
130#include <machine/bus_memio.h>
131#include <machine/bus.h>
132#include <machine/resource.h>
133#include <sys/bus.h>
134#include <sys/rman.h>
135
136#include <dev/mii/mii.h>
137#include <dev/mii/miivar.h>
138
139#include <dev/pci/pcireg.h>
140#include <dev/pci/pcivar.h>
141
142MODULE_DEPEND(re, pci, 1, 1, 1);
143MODULE_DEPEND(re, ether, 1, 1, 1);
144MODULE_DEPEND(re, miibus, 1, 1, 1);
145
146/* "controller miibus0" required.  See GENERIC if you get errors here. */
147#include "miibus_if.h"
148
149/*
150 * Default to using PIO access for this driver.
151 */
152#define RE_USEIOSPACE
153
154#include <pci/if_rlreg.h>
155
156#define RE_CSUM_FEATURES    (CSUM_IP | CSUM_TCP | CSUM_UDP)
157
158/*
159 * Various supported device vendors/types and their names.
160 */
161static struct rl_type re_devs[] = {
162	{ RT_VENDORID, RT_DEVICEID_8139, RL_HWREV_8139CPLUS,
163		"RealTek 8139C+ 10/100BaseTX" },
164	{ RT_VENDORID, RT_DEVICEID_8169, RL_HWREV_8169,
165		"RealTek 8169 Gigabit Ethernet" },
166	{ RT_VENDORID, RT_DEVICEID_8169, RL_HWREV_8169S,
167		"RealTek 8169S Single-chip Gigabit Ethernet" },
168	{ RT_VENDORID, RT_DEVICEID_8169, RL_HWREV_8110S,
169		"RealTek 8110S Single-chip Gigabit Ethernet" },
170	{ 0, 0, 0, NULL }
171};
172
173static struct rl_hwrev re_hwrevs[] = {
174	{ RL_HWREV_8139, RL_8139,  "" },
175	{ RL_HWREV_8139A, RL_8139, "A" },
176	{ RL_HWREV_8139AG, RL_8139, "A-G" },
177	{ RL_HWREV_8139B, RL_8139, "B" },
178	{ RL_HWREV_8130, RL_8139, "8130" },
179	{ RL_HWREV_8139C, RL_8139, "C" },
180	{ RL_HWREV_8139D, RL_8139, "8139D/8100B/8100C" },
181	{ RL_HWREV_8139CPLUS, RL_8139CPLUS, "C+"},
182	{ RL_HWREV_8169, RL_8169, "8169"},
183	{ RL_HWREV_8169S, RL_8169, "8169S"},
184	{ RL_HWREV_8110S, RL_8169, "8110S"},
185	{ RL_HWREV_8100, RL_8139, "8100"},
186	{ RL_HWREV_8101, RL_8139, "8101"},
187	{ 0, 0, NULL }
188};
189
190static int re_probe		(device_t);
191static int re_attach		(device_t);
192static int re_detach		(device_t);
193
194static int re_encap		(struct rl_softc *, struct mbuf *, int *);
195
196static void re_dma_map_addr	(void *, bus_dma_segment_t *, int, int);
197static void re_dma_map_desc	(void *, bus_dma_segment_t *, int,
198				    bus_size_t, int);
199static int re_allocmem		(device_t, struct rl_softc *);
200static int re_newbuf		(struct rl_softc *, int, struct mbuf *);
201static int re_rx_list_init	(struct rl_softc *);
202static int re_tx_list_init	(struct rl_softc *);
203static void re_rxeof		(struct rl_softc *);
204static void re_txeof		(struct rl_softc *);
205static void re_intr		(void *);
206static void re_tick		(void *);
207static void re_start		(struct ifnet *);
208static int re_ioctl		(struct ifnet *, u_long, caddr_t);
209static void re_init		(void *);
210static void re_stop		(struct rl_softc *);
211static void re_watchdog		(struct ifnet *);
212static int re_suspend		(device_t);
213static int re_resume		(device_t);
214static void re_shutdown		(device_t);
215static int re_ifmedia_upd	(struct ifnet *);
216static void re_ifmedia_sts	(struct ifnet *, struct ifmediareq *);
217
218static void re_eeprom_putbyte	(struct rl_softc *, int);
219static void re_eeprom_getword	(struct rl_softc *, int, u_int16_t *);
220static void re_read_eeprom	(struct rl_softc *, caddr_t, int, int, int);
221static int re_gmii_readreg	(device_t, int, int);
222static int re_gmii_writereg	(device_t, int, int, int);
223
224static int re_miibus_readreg	(device_t, int, int);
225static int re_miibus_writereg	(device_t, int, int, int);
226static void re_miibus_statchg	(device_t);
227
228static u_int8_t re_calchash	(caddr_t);
229static void re_setmulti		(struct rl_softc *);
230static void re_reset		(struct rl_softc *);
231
232static int re_diag		(struct rl_softc *);
233
234#ifdef RE_USEIOSPACE
235#define RL_RES			SYS_RES_IOPORT
236#define RL_RID			RL_PCI_LOIO
237#else
238#define RL_RES			SYS_RES_MEMORY
239#define RL_RID			RL_PCI_LOMEM
240#endif
241
242static device_method_t re_methods[] = {
243	/* Device interface */
244	DEVMETHOD(device_probe,		re_probe),
245	DEVMETHOD(device_attach,	re_attach),
246	DEVMETHOD(device_detach,	re_detach),
247	DEVMETHOD(device_suspend,	re_suspend),
248	DEVMETHOD(device_resume,	re_resume),
249	DEVMETHOD(device_shutdown,	re_shutdown),
250
251	/* bus interface */
252	DEVMETHOD(bus_print_child,	bus_generic_print_child),
253	DEVMETHOD(bus_driver_added,	bus_generic_driver_added),
254
255	/* MII interface */
256	DEVMETHOD(miibus_readreg,	re_miibus_readreg),
257	DEVMETHOD(miibus_writereg,	re_miibus_writereg),
258	DEVMETHOD(miibus_statchg,	re_miibus_statchg),
259
260	{ 0, 0 }
261};
262
263static driver_t re_driver = {
264	"re",
265	re_methods,
266	sizeof(struct rl_softc)
267};
268
269static devclass_t re_devclass;
270
271DRIVER_MODULE(re, pci, re_driver, re_devclass, 0, 0);
272DRIVER_MODULE(re, cardbus, re_driver, re_devclass, 0, 0);
273DRIVER_MODULE(miibus, re, miibus_driver, miibus_devclass, 0, 0);
274
275#define EE_SET(x)					\
276	CSR_WRITE_1(sc, RL_EECMD,			\
277		CSR_READ_1(sc, RL_EECMD) | x)
278
279#define EE_CLR(x)					\
280	CSR_WRITE_1(sc, RL_EECMD,			\
281		CSR_READ_1(sc, RL_EECMD) & ~x)
282
283/*
284 * Send a read command and address to the EEPROM, check for ACK.
285 */
286static void
287re_eeprom_putbyte(sc, addr)
288	struct rl_softc		*sc;
289	int			addr;
290{
291	register int		d, i;
292
293	d = addr | sc->rl_eecmd_read;
294
295	/*
296	 * Feed in each bit and strobe the clock.
297	 */
298	for (i = 0x400; i; i >>= 1) {
299		if (d & i) {
300			EE_SET(RL_EE_DATAIN);
301		} else {
302			EE_CLR(RL_EE_DATAIN);
303		}
304		DELAY(100);
305		EE_SET(RL_EE_CLK);
306		DELAY(150);
307		EE_CLR(RL_EE_CLK);
308		DELAY(100);
309	}
310
311	return;
312}
313
314/*
315 * Read a word of data stored in the EEPROM at address 'addr.'
316 */
317static void
318re_eeprom_getword(sc, addr, dest)
319	struct rl_softc		*sc;
320	int			addr;
321	u_int16_t		*dest;
322{
323	register int		i;
324	u_int16_t		word = 0;
325
326	/* Enter EEPROM access mode. */
327	CSR_WRITE_1(sc, RL_EECMD, RL_EEMODE_PROGRAM|RL_EE_SEL);
328
329	/*
330	 * Send address of word we want to read.
331	 */
332	re_eeprom_putbyte(sc, addr);
333
334	CSR_WRITE_1(sc, RL_EECMD, RL_EEMODE_PROGRAM|RL_EE_SEL);
335
336	/*
337	 * Start reading bits from EEPROM.
338	 */
339	for (i = 0x8000; i; i >>= 1) {
340		EE_SET(RL_EE_CLK);
341		DELAY(100);
342		if (CSR_READ_1(sc, RL_EECMD) & RL_EE_DATAOUT)
343			word |= i;
344		EE_CLR(RL_EE_CLK);
345		DELAY(100);
346	}
347
348	/* Turn off EEPROM access mode. */
349	CSR_WRITE_1(sc, RL_EECMD, RL_EEMODE_OFF);
350
351	*dest = word;
352
353	return;
354}
355
356/*
357 * Read a sequence of words from the EEPROM.
358 */
359static void
360re_read_eeprom(sc, dest, off, cnt, swap)
361	struct rl_softc		*sc;
362	caddr_t			dest;
363	int			off;
364	int			cnt;
365	int			swap;
366{
367	int			i;
368	u_int16_t		word = 0, *ptr;
369
370	for (i = 0; i < cnt; i++) {
371		re_eeprom_getword(sc, off + i, &word);
372		ptr = (u_int16_t *)(dest + (i * 2));
373		if (swap)
374			*ptr = ntohs(word);
375		else
376			*ptr = word;
377	}
378
379	return;
380}
381
382static int
383re_gmii_readreg(dev, phy, reg)
384	device_t		dev;
385	int			phy, reg;
386{
387	struct rl_softc		*sc;
388	u_int32_t		rval;
389	int			i;
390
391	if (phy != 1)
392		return(0);
393
394	sc = device_get_softc(dev);
395
396	/* Let the rgephy driver read the GMEDIASTAT register */
397
398	if (reg == RL_GMEDIASTAT) {
399		rval = CSR_READ_1(sc, RL_GMEDIASTAT);
400		return(rval);
401	}
402
403	CSR_WRITE_4(sc, RL_PHYAR, reg << 16);
404	DELAY(1000);
405
406	for (i = 0; i < RL_TIMEOUT; i++) {
407		rval = CSR_READ_4(sc, RL_PHYAR);
408		if (rval & RL_PHYAR_BUSY)
409			break;
410		DELAY(100);
411	}
412
413	if (i == RL_TIMEOUT) {
414		printf ("re%d: PHY read failed\n", sc->rl_unit);
415		return (0);
416	}
417
418	return (rval & RL_PHYAR_PHYDATA);
419}
420
421static int
422re_gmii_writereg(dev, phy, reg, data)
423	device_t		dev;
424	int			phy, reg, data;
425{
426	struct rl_softc		*sc;
427	u_int32_t		rval;
428	int			i;
429
430	sc = device_get_softc(dev);
431
432	CSR_WRITE_4(sc, RL_PHYAR, (reg << 16) |
433	    (data & RL_PHYAR_PHYDATA) | RL_PHYAR_BUSY);
434	DELAY(1000);
435
436	for (i = 0; i < RL_TIMEOUT; i++) {
437		rval = CSR_READ_4(sc, RL_PHYAR);
438		if (!(rval & RL_PHYAR_BUSY))
439			break;
440		DELAY(100);
441	}
442
443	if (i == RL_TIMEOUT) {
444		printf ("re%d: PHY write failed\n", sc->rl_unit);
445		return (0);
446	}
447
448	return (0);
449}
450
451static int
452re_miibus_readreg(dev, phy, reg)
453	device_t		dev;
454	int			phy, reg;
455{
456	struct rl_softc		*sc;
457	u_int16_t		rval = 0;
458	u_int16_t		re8139_reg = 0;
459
460	sc = device_get_softc(dev);
461	RL_LOCK(sc);
462
463	if (sc->rl_type == RL_8169) {
464		rval = re_gmii_readreg(dev, phy, reg);
465		RL_UNLOCK(sc);
466		return (rval);
467	}
468
469	/* Pretend the internal PHY is only at address 0 */
470	if (phy) {
471		RL_UNLOCK(sc);
472		return(0);
473	}
474	switch(reg) {
475	case MII_BMCR:
476		re8139_reg = RL_BMCR;
477		break;
478	case MII_BMSR:
479		re8139_reg = RL_BMSR;
480		break;
481	case MII_ANAR:
482		re8139_reg = RL_ANAR;
483		break;
484	case MII_ANER:
485		re8139_reg = RL_ANER;
486		break;
487	case MII_ANLPAR:
488		re8139_reg = RL_LPAR;
489		break;
490	case MII_PHYIDR1:
491	case MII_PHYIDR2:
492		RL_UNLOCK(sc);
493		return(0);
494	/*
495	 * Allow the rlphy driver to read the media status
496	 * register. If we have a link partner which does not
497	 * support NWAY, this is the register which will tell
498	 * us the results of parallel detection.
499	 */
500	case RL_MEDIASTAT:
501		rval = CSR_READ_1(sc, RL_MEDIASTAT);
502		RL_UNLOCK(sc);
503		return(rval);
504	default:
505		printf("re%d: bad phy register\n", sc->rl_unit);
506		RL_UNLOCK(sc);
507		return(0);
508	}
509	rval = CSR_READ_2(sc, re8139_reg);
510	RL_UNLOCK(sc);
511	return(rval);
512}
513
514static int
515re_miibus_writereg(dev, phy, reg, data)
516	device_t		dev;
517	int			phy, reg, data;
518{
519	struct rl_softc		*sc;
520	u_int16_t		re8139_reg = 0;
521	int			rval = 0;
522
523	sc = device_get_softc(dev);
524	RL_LOCK(sc);
525
526	if (sc->rl_type == RL_8169) {
527		rval = re_gmii_writereg(dev, phy, reg, data);
528		RL_UNLOCK(sc);
529		return (rval);
530	}
531
532	/* Pretend the internal PHY is only at address 0 */
533	if (phy) {
534		RL_UNLOCK(sc);
535		return(0);
536	}
537	switch(reg) {
538	case MII_BMCR:
539		re8139_reg = RL_BMCR;
540		break;
541	case MII_BMSR:
542		re8139_reg = RL_BMSR;
543		break;
544	case MII_ANAR:
545		re8139_reg = RL_ANAR;
546		break;
547	case MII_ANER:
548		re8139_reg = RL_ANER;
549		break;
550	case MII_ANLPAR:
551		re8139_reg = RL_LPAR;
552		break;
553	case MII_PHYIDR1:
554	case MII_PHYIDR2:
555		RL_UNLOCK(sc);
556		return(0);
557		break;
558	default:
559		printf("re%d: bad phy register\n", sc->rl_unit);
560		RL_UNLOCK(sc);
561		return(0);
562	}
563	CSR_WRITE_2(sc, re8139_reg, data);
564	RL_UNLOCK(sc);
565	return(0);
566}
567
568static void
569re_miibus_statchg(dev)
570	device_t		dev;
571{
572	return;
573}
574
575/*
576 * Calculate CRC of a multicast group address, return the upper 6 bits.
577 */
578static u_int8_t
579re_calchash(addr)
580	caddr_t			addr;
581{
582	u_int32_t		crc, carry;
583	int			i, j;
584	u_int8_t		c;
585
586	/* Compute CRC for the address value. */
587	crc = 0xFFFFFFFF; /* initial value */
588
589	for (i = 0; i < 6; i++) {
590		c = *(addr + i);
591		for (j = 0; j < 8; j++) {
592			carry = ((crc & 0x80000000) ? 1 : 0) ^ (c & 0x01);
593			crc <<= 1;
594			c >>= 1;
595			if (carry)
596				crc = (crc ^ 0x04c11db6) | carry;
597		}
598	}
599
600	/* return the filter bit position */
601	return(crc >> 26);
602}
603
604/*
605 * Program the 64-bit multicast hash filter.
606 */
607static void
608re_setmulti(sc)
609	struct rl_softc		*sc;
610{
611	struct ifnet		*ifp;
612	int			h = 0;
613	u_int32_t		hashes[2] = { 0, 0 };
614	struct ifmultiaddr	*ifma;
615	u_int32_t		rxfilt;
616	int			mcnt = 0;
617
618	ifp = &sc->arpcom.ac_if;
619
620	rxfilt = CSR_READ_4(sc, RL_RXCFG);
621
622	if (ifp->if_flags & IFF_ALLMULTI || ifp->if_flags & IFF_PROMISC) {
623		rxfilt |= RL_RXCFG_RX_MULTI;
624		CSR_WRITE_4(sc, RL_RXCFG, rxfilt);
625		CSR_WRITE_4(sc, RL_MAR0, 0xFFFFFFFF);
626		CSR_WRITE_4(sc, RL_MAR4, 0xFFFFFFFF);
627		return;
628	}
629
630	/* first, zot all the existing hash bits */
631	CSR_WRITE_4(sc, RL_MAR0, 0);
632	CSR_WRITE_4(sc, RL_MAR4, 0);
633
634	/* now program new ones */
635	TAILQ_FOREACH(ifma, &ifp->if_multiaddrs, ifma_link) {
636		if (ifma->ifma_addr->sa_family != AF_LINK)
637			continue;
638		h = re_calchash(LLADDR((struct sockaddr_dl *)ifma->ifma_addr));
639		if (h < 32)
640			hashes[0] |= (1 << h);
641		else
642			hashes[1] |= (1 << (h - 32));
643		mcnt++;
644	}
645
646	if (mcnt)
647		rxfilt |= RL_RXCFG_RX_MULTI;
648	else
649		rxfilt &= ~RL_RXCFG_RX_MULTI;
650
651	CSR_WRITE_4(sc, RL_RXCFG, rxfilt);
652	CSR_WRITE_4(sc, RL_MAR0, hashes[0]);
653	CSR_WRITE_4(sc, RL_MAR4, hashes[1]);
654
655	return;
656}
657
658static void
659re_reset(sc)
660	struct rl_softc		*sc;
661{
662	register int		i;
663
664	CSR_WRITE_1(sc, RL_COMMAND, RL_CMD_RESET);
665
666	for (i = 0; i < RL_TIMEOUT; i++) {
667		DELAY(10);
668		if (!(CSR_READ_1(sc, RL_COMMAND) & RL_CMD_RESET))
669			break;
670	}
671	if (i == RL_TIMEOUT)
672		printf("re%d: reset never completed!\n", sc->rl_unit);
673
674	CSR_WRITE_1(sc, 0x82, 1);
675
676	return;
677}
678
679/*
680 * The following routine is designed to test for a defect on some
681 * 32-bit 8169 cards. Some of these NICs have the REQ64# and ACK64#
682 * lines connected to the bus, however for a 32-bit only card, they
683 * should be pulled high. The result of this defect is that the
684 * NIC will not work right if you plug it into a 64-bit slot: DMA
685 * operations will be done with 64-bit transfers, which will fail
686 * because the 64-bit data lines aren't connected.
687 *
688 * There's no way to work around this (short of talking a soldering
689 * iron to the board), however we can detect it. The method we use
690 * here is to put the NIC into digital loopback mode, set the receiver
691 * to promiscuous mode, and then try to send a frame. We then compare
692 * the frame data we sent to what was received. If the data matches,
693 * then the NIC is working correctly, otherwise we know the user has
694 * a defective NIC which has been mistakenly plugged into a 64-bit PCI
695 * slot. In the latter case, there's no way the NIC can work correctly,
696 * so we print out a message on the console and abort the device attach.
697 */
698
699static int
700re_diag(sc)
701	struct rl_softc		*sc;
702{
703	struct ifnet		*ifp = &sc->arpcom.ac_if;
704	struct mbuf		*m0;
705	struct ether_header	*eh;
706	struct rl_desc		*cur_rx;
707	u_int16_t		status;
708	u_int32_t		rxstat;
709	int			total_len, i, error = 0;
710	u_int8_t		dst[] = { 0x00, 'h', 'e', 'l', 'l', 'o' };
711	u_int8_t		src[] = { 0x00, 'w', 'o', 'r', 'l', 'd' };
712
713	/* Allocate a single mbuf */
714
715	MGETHDR(m0, M_DONTWAIT, MT_DATA);
716	if (m0 == NULL)
717		return(ENOBUFS);
718
719	/*
720	 * Initialize the NIC in test mode. This sets the chip up
721	 * so that it can send and receive frames, but performs the
722	 * following special functions:
723	 * - Puts receiver in promiscuous mode
724	 * - Enables digital loopback mode
725	 * - Leaves interrupts turned off
726	 */
727
728	ifp->if_flags |= IFF_PROMISC;
729	sc->rl_testmode = 1;
730	re_init(sc);
731
732	/* Put some data in the mbuf */
733
734	eh = mtod(m0, struct ether_header *);
735	bcopy ((char *)&dst, eh->ether_dhost, ETHER_ADDR_LEN);
736	bcopy ((char *)&src, eh->ether_shost, ETHER_ADDR_LEN);
737	eh->ether_type = htons(ETHERTYPE_IP);
738	m0->m_pkthdr.len = m0->m_len = ETHER_MIN_LEN - ETHER_CRC_LEN;
739
740	/* Queue the packet, start transmission */
741
742	IF_HANDOFF(&ifp->if_snd, m0, ifp);
743	re_start(ifp);
744	m0 = NULL;
745
746	/* Wait for it to propagate through the chip */
747
748	for (i = 0; i < RL_TIMEOUT; i++) {
749		status = CSR_READ_2(sc, RL_ISR);
750		if (status & RL_ISR_RX_OK)
751			break;
752		DELAY(10);
753	}
754
755	if (i == RL_TIMEOUT) {
756		printf("re%d: diagnostic failed, failed to receive packet "
757		    "in loopback mode\n", sc->rl_unit);
758		error = EIO;
759		goto done;
760	}
761
762	/*
763	 * The packet should have been dumped into the first
764	 * entry in the RX DMA ring. Grab it from there.
765	 */
766
767	bus_dmamap_sync(sc->rl_ldata.rl_rx_list_tag,
768	    sc->rl_ldata.rl_rx_list_map,
769	    BUS_DMASYNC_POSTREAD);
770	bus_dmamap_sync(sc->rl_ldata.rl_mtag,
771	    sc->rl_ldata.rl_rx_dmamap[0],
772	    BUS_DMASYNC_POSTWRITE);
773	bus_dmamap_unload(sc->rl_ldata.rl_mtag,
774	    sc->rl_ldata.rl_rx_dmamap[0]);
775
776	m0 = sc->rl_ldata.rl_rx_mbuf[0];
777	sc->rl_ldata.rl_rx_mbuf[0] = NULL;
778	eh = mtod(m0, struct ether_header *);
779
780	cur_rx = &sc->rl_ldata.rl_rx_list[0];
781	total_len = RL_RXBYTES(cur_rx);
782	rxstat = le32toh(cur_rx->rl_cmdstat);
783
784	if (total_len != ETHER_MIN_LEN) {
785		printf("re%d: diagnostic failed, received short packet\n",
786		    sc->rl_unit);
787		error = EIO;
788		goto done;
789	}
790
791	/* Test that the received packet data matches what we sent. */
792
793	if (bcmp((char *)&eh->ether_dhost, (char *)&dst, ETHER_ADDR_LEN) ||
794	    bcmp((char *)&eh->ether_shost, (char *)&src, ETHER_ADDR_LEN) ||
795	    ntohs(eh->ether_type) != ETHERTYPE_IP) {
796		printf("re%d: WARNING, DMA FAILURE!\n", sc->rl_unit);
797		printf("re%d: expected TX data: %6D/%6D/0x%x\n", sc->rl_unit,
798		    dst, ":", src, ":", ETHERTYPE_IP);
799		printf("re%d: received RX data: %6D/%6D/0x%x\n", sc->rl_unit,
800		    eh->ether_dhost, ":",  eh->ether_shost, ":",
801		    ntohs(eh->ether_type));
802		printf("re%d: You may have a defective 32-bit NIC plugged "
803		    "into a 64-bit PCI slot.\n", sc->rl_unit);
804		printf("re%d: Please re-install the NIC in a 32-bit slot "
805		    "for proper operation.\n", sc->rl_unit);
806		printf("re%d: Read the re(4) man page for more details.\n",
807		    sc->rl_unit);
808		error = EIO;
809	}
810
811done:
812	/* Turn interface off, release resources */
813
814	sc->rl_testmode = 0;
815	ifp->if_flags &= ~IFF_PROMISC;
816	re_stop(sc);
817	if (m0 != NULL)
818		m_freem(m0);
819
820	return (error);
821}
822
823/*
824 * Probe for a RealTek 8139C+/8169/8110 chip. Check the PCI vendor and device
825 * IDs against our list and return a device name if we find a match.
826 */
827static int
828re_probe(dev)
829	device_t		dev;
830{
831	struct rl_type		*t;
832	struct rl_softc		*sc;
833	int			rid;
834	u_int32_t		hwrev;
835
836	t = re_devs;
837	sc = device_get_softc(dev);
838
839	while(t->rl_name != NULL) {
840		if ((pci_get_vendor(dev) == t->rl_vid) &&
841		    (pci_get_device(dev) == t->rl_did)) {
842
843			/*
844			 * Temporarily map the I/O space
845			 * so we can read the chip ID register.
846			 */
847			rid = RL_RID;
848			sc->rl_res = bus_alloc_resource(dev, RL_RES, &rid,
849			    0, ~0, 1, RF_ACTIVE);
850			if (sc->rl_res == NULL) {
851				device_printf(dev,
852				    "couldn't map ports/memory\n");
853				return(ENXIO);
854			}
855			sc->rl_btag = rman_get_bustag(sc->rl_res);
856			sc->rl_bhandle = rman_get_bushandle(sc->rl_res);
857			mtx_init(&sc->rl_mtx,
858			    device_get_nameunit(dev),
859			    MTX_NETWORK_LOCK, MTX_DEF);
860			RL_LOCK(sc);
861			hwrev = CSR_READ_4(sc, RL_TXCFG) & RL_TXCFG_HWREV;
862			bus_release_resource(dev, RL_RES,
863			    RL_RID, sc->rl_res);
864			RL_UNLOCK(sc);
865			mtx_destroy(&sc->rl_mtx);
866			if (t->rl_basetype == hwrev) {
867				device_set_desc(dev, t->rl_name);
868				return(0);
869			}
870		}
871		t++;
872	}
873
874	return(ENXIO);
875}
876
877/*
878 * This routine takes the segment list provided as the result of
879 * a bus_dma_map_load() operation and assigns the addresses/lengths
880 * to RealTek DMA descriptors. This can be called either by the RX
881 * code or the TX code. In the RX case, we'll probably wind up mapping
882 * at most one segment. For the TX case, there could be any number of
883 * segments since TX packets may span multiple mbufs. In either case,
884 * if the number of segments is larger than the rl_maxsegs limit
885 * specified by the caller, we abort the mapping operation. Sadly,
886 * whoever designed the buffer mapping API did not provide a way to
887 * return an error from here, so we have to fake it a bit.
888 */
889
890static void
891re_dma_map_desc(arg, segs, nseg, mapsize, error)
892	void			*arg;
893	bus_dma_segment_t	*segs;
894	int			nseg;
895	bus_size_t		mapsize;
896	int			error;
897{
898	struct rl_dmaload_arg	*ctx;
899	struct rl_desc		*d = NULL;
900	int			i = 0, idx;
901
902	if (error)
903		return;
904
905	ctx = arg;
906
907	/* Signal error to caller if there's too many segments */
908	if (nseg > ctx->rl_maxsegs) {
909		ctx->rl_maxsegs = 0;
910		return;
911	}
912
913	/*
914	 * Map the segment array into descriptors. Note that we set the
915	 * start-of-frame and end-of-frame markers for either TX or RX, but
916	 * they really only have meaning in the TX case. (In the RX case,
917	 * it's the chip that tells us where packets begin and end.)
918	 * We also keep track of the end of the ring and set the
919	 * end-of-ring bits as needed, and we set the ownership bits
920	 * in all except the very first descriptor. (The caller will
921	 * set this descriptor later when it start transmission or
922	 * reception.)
923	 */
924	idx = ctx->rl_idx;
925	while(1) {
926		u_int32_t		cmdstat;
927		d = &ctx->rl_ring[idx];
928		if (le32toh(d->rl_cmdstat) & RL_RDESC_STAT_OWN) {
929			ctx->rl_maxsegs = 0;
930			return;
931		}
932		cmdstat = segs[i].ds_len;
933		d->rl_bufaddr_lo = htole32(RL_ADDR_LO(segs[i].ds_addr));
934		d->rl_bufaddr_hi = htole32(RL_ADDR_HI(segs[i].ds_addr));
935		if (i == 0)
936			cmdstat |= RL_TDESC_CMD_SOF;
937		else
938			cmdstat |= RL_TDESC_CMD_OWN;
939		if (idx == (RL_RX_DESC_CNT - 1))
940			cmdstat |= RL_TDESC_CMD_EOR;
941		d->rl_cmdstat = htole32(cmdstat | ctx->rl_flags);
942		i++;
943		if (i == nseg)
944			break;
945		RL_DESC_INC(idx);
946	}
947
948	d->rl_cmdstat |= htole32(RL_TDESC_CMD_EOF);
949	ctx->rl_maxsegs = nseg;
950	ctx->rl_idx = idx;
951
952	return;
953}
954
955/*
956 * Map a single buffer address.
957 */
958
959static void
960re_dma_map_addr(arg, segs, nseg, error)
961	void			*arg;
962	bus_dma_segment_t	*segs;
963	int			nseg;
964	int			error;
965{
966	u_int32_t		*addr;
967
968	if (error)
969		return;
970
971	KASSERT(nseg == 1, ("too many DMA segments, %d should be 1", nseg));
972	addr = arg;
973	*addr = segs->ds_addr;
974
975	return;
976}
977
978static int
979re_allocmem(dev, sc)
980	device_t		dev;
981	struct rl_softc		*sc;
982{
983	int			error;
984	int			nseg;
985	int			i;
986
987	/*
988	 * Allocate map for RX mbufs.
989	 */
990	nseg = 32;
991	error = bus_dma_tag_create(sc->rl_parent_tag, ETHER_ALIGN, 0,
992	    BUS_SPACE_MAXADDR_32BIT, BUS_SPACE_MAXADDR, NULL,
993	    NULL, MCLBYTES * nseg, nseg, RL_JLEN, BUS_DMA_ALLOCNOW,
994	    NULL, NULL, &sc->rl_ldata.rl_mtag);
995	if (error) {
996		device_printf(dev, "could not allocate dma tag\n");
997		return (ENOMEM);
998	}
999
1000	/*
1001	 * Allocate map for TX descriptor list.
1002	 */
1003	error = bus_dma_tag_create(sc->rl_parent_tag, RL_RING_ALIGN,
1004	    0, BUS_SPACE_MAXADDR_32BIT, BUS_SPACE_MAXADDR, NULL,
1005            NULL, RL_TX_LIST_SZ, 1, RL_TX_LIST_SZ, BUS_DMA_ALLOCNOW,
1006	    NULL, NULL, &sc->rl_ldata.rl_tx_list_tag);
1007	if (error) {
1008		device_printf(dev, "could not allocate dma tag\n");
1009		return (ENOMEM);
1010	}
1011
1012	/* Allocate DMA'able memory for the TX ring */
1013
1014        error = bus_dmamem_alloc(sc->rl_ldata.rl_tx_list_tag,
1015	    (void **)&sc->rl_ldata.rl_tx_list, BUS_DMA_NOWAIT | BUS_DMA_ZERO,
1016            &sc->rl_ldata.rl_tx_list_map);
1017        if (error)
1018                return (ENOMEM);
1019
1020	/* Load the map for the TX ring. */
1021
1022	error = bus_dmamap_load(sc->rl_ldata.rl_tx_list_tag,
1023	     sc->rl_ldata.rl_tx_list_map, sc->rl_ldata.rl_tx_list,
1024	     RL_TX_LIST_SZ, re_dma_map_addr,
1025	     &sc->rl_ldata.rl_tx_list_addr, BUS_DMA_NOWAIT);
1026
1027	/* Create DMA maps for TX buffers */
1028
1029	for (i = 0; i < RL_TX_DESC_CNT; i++) {
1030		error = bus_dmamap_create(sc->rl_ldata.rl_mtag, 0,
1031			    &sc->rl_ldata.rl_tx_dmamap[i]);
1032		if (error) {
1033			device_printf(dev, "can't create DMA map for TX\n");
1034			return(ENOMEM);
1035		}
1036	}
1037
1038	/*
1039	 * Allocate map for RX descriptor list.
1040	 */
1041	error = bus_dma_tag_create(sc->rl_parent_tag, RL_RING_ALIGN,
1042	    0, BUS_SPACE_MAXADDR_32BIT, BUS_SPACE_MAXADDR, NULL,
1043            NULL, RL_TX_LIST_SZ, 1, RL_TX_LIST_SZ, BUS_DMA_ALLOCNOW,
1044	    NULL, NULL, &sc->rl_ldata.rl_rx_list_tag);
1045	if (error) {
1046		device_printf(dev, "could not allocate dma tag\n");
1047		return (ENOMEM);
1048	}
1049
1050	/* Allocate DMA'able memory for the RX ring */
1051
1052        error = bus_dmamem_alloc(sc->rl_ldata.rl_rx_list_tag,
1053	    (void **)&sc->rl_ldata.rl_rx_list, BUS_DMA_NOWAIT | BUS_DMA_ZERO,
1054            &sc->rl_ldata.rl_rx_list_map);
1055        if (error)
1056                return (ENOMEM);
1057
1058	/* Load the map for the RX ring. */
1059
1060	error = bus_dmamap_load(sc->rl_ldata.rl_rx_list_tag,
1061	     sc->rl_ldata.rl_rx_list_map, sc->rl_ldata.rl_rx_list,
1062	     RL_TX_LIST_SZ, re_dma_map_addr,
1063	     &sc->rl_ldata.rl_rx_list_addr, BUS_DMA_NOWAIT);
1064
1065	/* Create DMA maps for RX buffers */
1066
1067	for (i = 0; i < RL_RX_DESC_CNT; i++) {
1068		error = bus_dmamap_create(sc->rl_ldata.rl_mtag, 0,
1069			    &sc->rl_ldata.rl_rx_dmamap[i]);
1070		if (error) {
1071			device_printf(dev, "can't create DMA map for RX\n");
1072			return(ENOMEM);
1073		}
1074	}
1075
1076	return(0);
1077}
1078
1079/*
1080 * Attach the interface. Allocate softc structures, do ifmedia
1081 * setup and ethernet/BPF attach.
1082 */
1083static int
1084re_attach(dev)
1085	device_t		dev;
1086{
1087	u_char			eaddr[ETHER_ADDR_LEN];
1088	u_int16_t		as[3];
1089	struct rl_softc		*sc;
1090	struct ifnet		*ifp;
1091	struct rl_hwrev		*hw_rev;
1092	int			hwrev;
1093	u_int16_t		re_did = 0;
1094	int			unit, error = 0, rid, i;
1095
1096	sc = device_get_softc(dev);
1097	unit = device_get_unit(dev);
1098
1099	mtx_init(&sc->rl_mtx, device_get_nameunit(dev), MTX_NETWORK_LOCK,
1100	    MTX_DEF | MTX_RECURSE);
1101#ifndef BURN_BRIDGES
1102	/*
1103	 * Handle power management nonsense.
1104	 */
1105
1106	if (pci_get_powerstate(dev) != PCI_POWERSTATE_D0) {
1107		u_int32_t		iobase, membase, irq;
1108
1109		/* Save important PCI config data. */
1110		iobase = pci_read_config(dev, RL_PCI_LOIO, 4);
1111		membase = pci_read_config(dev, RL_PCI_LOMEM, 4);
1112		irq = pci_read_config(dev, RL_PCI_INTLINE, 4);
1113
1114		/* Reset the power state. */
1115		printf("re%d: chip is is in D%d power mode "
1116		    "-- setting to D0\n", unit,
1117		    pci_get_powerstate(dev));
1118
1119		pci_set_powerstate(dev, PCI_POWERSTATE_D0);
1120
1121		/* Restore PCI config data. */
1122		pci_write_config(dev, RL_PCI_LOIO, iobase, 4);
1123		pci_write_config(dev, RL_PCI_LOMEM, membase, 4);
1124		pci_write_config(dev, RL_PCI_INTLINE, irq, 4);
1125	}
1126#endif
1127	/*
1128	 * Map control/status registers.
1129	 */
1130	pci_enable_busmaster(dev);
1131
1132	rid = RL_RID;
1133	sc->rl_res = bus_alloc_resource(dev, RL_RES, &rid,
1134	    0, ~0, 1, RF_ACTIVE);
1135
1136	if (sc->rl_res == NULL) {
1137		printf ("re%d: couldn't map ports/memory\n", unit);
1138		error = ENXIO;
1139		goto fail;
1140	}
1141
1142	sc->rl_btag = rman_get_bustag(sc->rl_res);
1143	sc->rl_bhandle = rman_get_bushandle(sc->rl_res);
1144
1145	/* Allocate interrupt */
1146	rid = 0;
1147	sc->rl_irq = bus_alloc_resource(dev, SYS_RES_IRQ, &rid, 0, ~0, 1,
1148	    RF_SHAREABLE | RF_ACTIVE);
1149
1150	if (sc->rl_irq == NULL) {
1151		printf("re%d: couldn't map interrupt\n", unit);
1152		error = ENXIO;
1153		goto fail;
1154	}
1155
1156	/* Reset the adapter. */
1157	re_reset(sc);
1158	CSR_WRITE_1(sc, RL_EECMD, RL_EEMODE_PROGRAM);
1159	sc->rl_eecmd_read = RL_EECMD_READ_6BIT;
1160	re_read_eeprom(sc, (caddr_t)&re_did, 0, 1, 0);
1161	if (re_did != 0x8129)
1162		sc->rl_eecmd_read = RL_EECMD_READ_8BIT;
1163
1164	/*
1165	 * Get station address from the EEPROM.
1166	 */
1167	re_read_eeprom(sc, (caddr_t)as, RL_EE_EADDR, 3, 0);
1168	for (i = 0; i < 3; i++) {
1169		eaddr[(i * 2) + 0] = as[i] & 0xff;
1170		eaddr[(i * 2) + 1] = as[i] >> 8;
1171	}
1172
1173	CSR_WRITE_1(sc, RL_EECMD, RL_EEMODE_OFF);
1174
1175	/*
1176	 * A RealTek chip was detected. Inform the world.
1177	 */
1178	printf("re%d: Ethernet address: %6D\n", unit, eaddr, ":");
1179
1180	sc->rl_unit = unit;
1181	bcopy(eaddr, (char *)&sc->arpcom.ac_enaddr, ETHER_ADDR_LEN);
1182
1183	hw_rev = re_hwrevs;
1184	hwrev = CSR_READ_4(sc, RL_TXCFG) & RL_TXCFG_HWREV;
1185	while (hw_rev->rl_desc != NULL) {
1186		if (hw_rev->rl_rev == hwrev) {
1187			sc->rl_type = hw_rev->rl_type;
1188			break;
1189		}
1190		hw_rev++;
1191	}
1192
1193	/*
1194	 * Allocate the parent bus DMA tag appropriate for PCI.
1195	 */
1196#define RL_NSEG_NEW 32
1197	error = bus_dma_tag_create(NULL,	/* parent */
1198			1, 0,			/* alignment, boundary */
1199			BUS_SPACE_MAXADDR_32BIT,/* lowaddr */
1200			BUS_SPACE_MAXADDR,	/* highaddr */
1201			NULL, NULL,		/* filter, filterarg */
1202			MAXBSIZE, RL_NSEG_NEW,	/* maxsize, nsegments */
1203			BUS_SPACE_MAXSIZE_32BIT,/* maxsegsize */
1204			BUS_DMA_ALLOCNOW,	/* flags */
1205			NULL, NULL,		/* lockfunc, lockarg */
1206			&sc->rl_parent_tag);
1207	if (error)
1208		goto fail;
1209
1210	error = re_allocmem(dev, sc);
1211
1212	if (error)
1213		goto fail;
1214
1215	/* Do MII setup */
1216	if (mii_phy_probe(dev, &sc->rl_miibus,
1217	    re_ifmedia_upd, re_ifmedia_sts)) {
1218		printf("re%d: MII without any phy!\n", sc->rl_unit);
1219		error = ENXIO;
1220		goto fail;
1221	}
1222
1223	ifp = &sc->arpcom.ac_if;
1224	ifp->if_softc = sc;
1225	ifp->if_unit = unit;
1226	ifp->if_name = "re";
1227	ifp->if_mtu = ETHERMTU;
1228	ifp->if_flags = IFF_BROADCAST | IFF_SIMPLEX | IFF_MULTICAST;
1229	ifp->if_ioctl = re_ioctl;
1230	ifp->if_output = ether_output;
1231	ifp->if_capabilities = IFCAP_VLAN_MTU;
1232	ifp->if_start = re_start;
1233	ifp->if_hwassist = RE_CSUM_FEATURES;
1234	ifp->if_capabilities |= IFCAP_HWCSUM|IFCAP_VLAN_HWTAGGING;
1235	ifp->if_watchdog = re_watchdog;
1236	ifp->if_init = re_init;
1237	if (sc->rl_type == RL_8169)
1238		ifp->if_baudrate = 1000000000;
1239	else
1240		ifp->if_baudrate = 100000000;
1241	ifp->if_snd.ifq_maxlen = RL_IFQ_MAXLEN;
1242	ifp->if_capenable = ifp->if_capabilities;
1243
1244	callout_handle_init(&sc->rl_stat_ch);
1245
1246	/*
1247	 * Call MI attach routine.
1248	 */
1249	ether_ifattach(ifp, eaddr);
1250
1251	/* Perform hardware diagnostic. */
1252	error = re_diag(sc);
1253
1254	if (error) {
1255		printf("re%d: attach aborted due to hardware diag failure\n",
1256		    unit);
1257		ether_ifdetach(ifp);
1258		goto fail;
1259	}
1260
1261	/* Hook interrupt last to avoid having to lock softc */
1262	error = bus_setup_intr(dev, sc->rl_irq, INTR_TYPE_NET,
1263	    re_intr, sc, &sc->rl_intrhand);
1264
1265	if (error) {
1266		printf("re%d: couldn't set up irq\n", unit);
1267		ether_ifdetach(ifp);
1268		goto fail;
1269	}
1270
1271fail:
1272	if (error)
1273		re_detach(dev);
1274
1275	return (error);
1276}
1277
1278/*
1279 * Shutdown hardware and free up resources. This can be called any
1280 * time after the mutex has been initialized. It is called in both
1281 * the error case in attach and the normal detach case so it needs
1282 * to be careful about only freeing resources that have actually been
1283 * allocated.
1284 */
1285static int
1286re_detach(dev)
1287	device_t		dev;
1288{
1289	struct rl_softc		*sc;
1290	struct ifnet		*ifp;
1291	int			i;
1292
1293	sc = device_get_softc(dev);
1294	KASSERT(mtx_initialized(&sc->rl_mtx), ("rl mutex not initialized"));
1295	RL_LOCK(sc);
1296	ifp = &sc->arpcom.ac_if;
1297
1298	/* These should only be active if attach succeeded */
1299	if (device_is_attached(dev)) {
1300		re_stop(sc);
1301		/*
1302		 * Force off the IFF_UP flag here, in case someone
1303		 * still had a BPF descriptor attached to this
1304		 * interface. If they do, ether_ifattach() will cause
1305		 * the BPF code to try and clear the promisc mode
1306		 * flag, which will bubble down to re_ioctl(),
1307		 * which will try to call re_init() again. This will
1308		 * turn the NIC back on and restart the MII ticker,
1309		 * which will panic the system when the kernel tries
1310		 * to invoke the re_tick() function that isn't there
1311		 * anymore.
1312		 */
1313		ifp->if_flags &= ~IFF_UP;
1314		ether_ifdetach(ifp);
1315	}
1316	if (sc->rl_miibus)
1317		device_delete_child(dev, sc->rl_miibus);
1318	bus_generic_detach(dev);
1319
1320	if (sc->rl_intrhand)
1321		bus_teardown_intr(dev, sc->rl_irq, sc->rl_intrhand);
1322	if (sc->rl_irq)
1323		bus_release_resource(dev, SYS_RES_IRQ, 0, sc->rl_irq);
1324	if (sc->rl_res)
1325		bus_release_resource(dev, RL_RES, RL_RID, sc->rl_res);
1326
1327
1328	/* Unload and free the RX DMA ring memory and map */
1329
1330	if (sc->rl_ldata.rl_rx_list_tag) {
1331		bus_dmamap_unload(sc->rl_ldata.rl_rx_list_tag,
1332		    sc->rl_ldata.rl_rx_list_map);
1333		bus_dmamem_free(sc->rl_ldata.rl_rx_list_tag,
1334		    sc->rl_ldata.rl_rx_list,
1335		    sc->rl_ldata.rl_rx_list_map);
1336		bus_dma_tag_destroy(sc->rl_ldata.rl_rx_list_tag);
1337	}
1338
1339	/* Unload and free the TX DMA ring memory and map */
1340
1341	if (sc->rl_ldata.rl_tx_list_tag) {
1342		bus_dmamap_unload(sc->rl_ldata.rl_tx_list_tag,
1343		    sc->rl_ldata.rl_tx_list_map);
1344		bus_dmamem_free(sc->rl_ldata.rl_tx_list_tag,
1345		    sc->rl_ldata.rl_tx_list,
1346		    sc->rl_ldata.rl_tx_list_map);
1347		bus_dma_tag_destroy(sc->rl_ldata.rl_tx_list_tag);
1348	}
1349
1350	/* Destroy all the RX and TX buffer maps */
1351
1352	if (sc->rl_ldata.rl_mtag) {
1353		for (i = 0; i < RL_TX_DESC_CNT; i++)
1354			bus_dmamap_destroy(sc->rl_ldata.rl_mtag,
1355			    sc->rl_ldata.rl_tx_dmamap[i]);
1356		for (i = 0; i < RL_RX_DESC_CNT; i++)
1357			bus_dmamap_destroy(sc->rl_ldata.rl_mtag,
1358			    sc->rl_ldata.rl_rx_dmamap[i]);
1359		bus_dma_tag_destroy(sc->rl_ldata.rl_mtag);
1360	}
1361
1362	/* Unload and free the stats buffer and map */
1363
1364	if (sc->rl_ldata.rl_stag) {
1365		bus_dmamap_unload(sc->rl_ldata.rl_stag,
1366		    sc->rl_ldata.rl_rx_list_map);
1367		bus_dmamem_free(sc->rl_ldata.rl_stag,
1368		    sc->rl_ldata.rl_stats,
1369		    sc->rl_ldata.rl_smap);
1370		bus_dma_tag_destroy(sc->rl_ldata.rl_stag);
1371	}
1372
1373	if (sc->rl_parent_tag)
1374		bus_dma_tag_destroy(sc->rl_parent_tag);
1375
1376	RL_UNLOCK(sc);
1377	mtx_destroy(&sc->rl_mtx);
1378
1379	return(0);
1380}
1381
1382static int
1383re_newbuf(sc, idx, m)
1384	struct rl_softc		*sc;
1385	int			idx;
1386	struct mbuf		*m;
1387{
1388	struct rl_dmaload_arg	arg;
1389	struct mbuf		*n = NULL;
1390	int			error;
1391
1392	if (m == NULL) {
1393		n = m_getcl(M_DONTWAIT, MT_DATA, M_PKTHDR);
1394		if (n == NULL)
1395			return(ENOBUFS);
1396		m = n;
1397	} else
1398		m->m_data = m->m_ext.ext_buf;
1399
1400	/*
1401	 * Initialize mbuf length fields and fixup
1402	 * alignment so that the frame payload is
1403	 * longword aligned.
1404	 */
1405	m->m_len = m->m_pkthdr.len = MCLBYTES;
1406	m_adj(m, ETHER_ALIGN);
1407
1408	arg.sc = sc;
1409	arg.rl_idx = idx;
1410	arg.rl_maxsegs = 1;
1411	arg.rl_flags = 0;
1412	arg.rl_ring = sc->rl_ldata.rl_rx_list;
1413
1414        error = bus_dmamap_load_mbuf(sc->rl_ldata.rl_mtag,
1415	    sc->rl_ldata.rl_rx_dmamap[idx], m, re_dma_map_desc,
1416	    &arg, BUS_DMA_NOWAIT);
1417	if (error || arg.rl_maxsegs != 1) {
1418		if (n != NULL)
1419			m_freem(n);
1420		return (ENOMEM);
1421	}
1422
1423	sc->rl_ldata.rl_rx_list[idx].rl_cmdstat |= htole32(RL_RDESC_CMD_OWN);
1424	sc->rl_ldata.rl_rx_mbuf[idx] = m;
1425
1426        bus_dmamap_sync(sc->rl_ldata.rl_mtag,
1427	    sc->rl_ldata.rl_rx_dmamap[idx],
1428	    BUS_DMASYNC_PREREAD);
1429
1430	return(0);
1431}
1432
1433static int
1434re_tx_list_init(sc)
1435	struct rl_softc		*sc;
1436{
1437	bzero ((char *)sc->rl_ldata.rl_tx_list, RL_TX_LIST_SZ);
1438	bzero ((char *)&sc->rl_ldata.rl_tx_mbuf,
1439	    (RL_TX_DESC_CNT * sizeof(struct mbuf *)));
1440
1441	bus_dmamap_sync(sc->rl_ldata.rl_tx_list_tag,
1442	    sc->rl_ldata.rl_tx_list_map, BUS_DMASYNC_PREWRITE);
1443	sc->rl_ldata.rl_tx_prodidx = 0;
1444	sc->rl_ldata.rl_tx_considx = 0;
1445	sc->rl_ldata.rl_tx_free = RL_TX_DESC_CNT;
1446
1447	return(0);
1448}
1449
1450static int
1451re_rx_list_init(sc)
1452	struct rl_softc		*sc;
1453{
1454	int			i;
1455
1456	bzero ((char *)sc->rl_ldata.rl_rx_list, RL_RX_LIST_SZ);
1457	bzero ((char *)&sc->rl_ldata.rl_rx_mbuf,
1458	    (RL_RX_DESC_CNT * sizeof(struct mbuf *)));
1459
1460	for (i = 0; i < RL_RX_DESC_CNT; i++) {
1461		if (re_newbuf(sc, i, NULL) == ENOBUFS)
1462			return(ENOBUFS);
1463	}
1464
1465	/* Flush the RX descriptors */
1466
1467	bus_dmamap_sync(sc->rl_ldata.rl_rx_list_tag,
1468	    sc->rl_ldata.rl_rx_list_map,
1469	    BUS_DMASYNC_PREWRITE|BUS_DMASYNC_PREREAD);
1470
1471	sc->rl_ldata.rl_rx_prodidx = 0;
1472	sc->rl_head = sc->rl_tail = NULL;
1473
1474	return(0);
1475}
1476
1477/*
1478 * RX handler for C+ and 8169. For the gigE chips, we support
1479 * the reception of jumbo frames that have been fragmented
1480 * across multiple 2K mbuf cluster buffers.
1481 */
1482static void
1483re_rxeof(sc)
1484	struct rl_softc		*sc;
1485{
1486	struct mbuf		*m;
1487	struct ifnet		*ifp;
1488	int			i, total_len;
1489	struct rl_desc		*cur_rx;
1490	u_int32_t		rxstat, rxvlan;
1491
1492	ifp = &sc->arpcom.ac_if;
1493	i = sc->rl_ldata.rl_rx_prodidx;
1494
1495	/* Invalidate the descriptor memory */
1496
1497	bus_dmamap_sync(sc->rl_ldata.rl_rx_list_tag,
1498	    sc->rl_ldata.rl_rx_list_map,
1499	    BUS_DMASYNC_POSTREAD);
1500
1501	while (!RL_OWN(&sc->rl_ldata.rl_rx_list[i])) {
1502
1503		cur_rx = &sc->rl_ldata.rl_rx_list[i];
1504		m = sc->rl_ldata.rl_rx_mbuf[i];
1505		total_len = RL_RXBYTES(cur_rx);
1506		rxstat = le32toh(cur_rx->rl_cmdstat);
1507		rxvlan = le32toh(cur_rx->rl_vlanctl);
1508
1509		/* Invalidate the RX mbuf and unload its map */
1510
1511		bus_dmamap_sync(sc->rl_ldata.rl_mtag,
1512		    sc->rl_ldata.rl_rx_dmamap[i],
1513		    BUS_DMASYNC_POSTWRITE);
1514		bus_dmamap_unload(sc->rl_ldata.rl_mtag,
1515		    sc->rl_ldata.rl_rx_dmamap[i]);
1516
1517		if (!(rxstat & RL_RDESC_STAT_EOF)) {
1518			m->m_len = MCLBYTES - ETHER_ALIGN;
1519			if (sc->rl_head == NULL)
1520				sc->rl_head = sc->rl_tail = m;
1521			else {
1522				m->m_flags &= ~M_PKTHDR;
1523				sc->rl_tail->m_next = m;
1524				sc->rl_tail = m;
1525			}
1526			re_newbuf(sc, i, NULL);
1527			RL_DESC_INC(i);
1528			continue;
1529		}
1530
1531		/*
1532		 * NOTE: for the 8139C+, the frame length field
1533		 * is always 12 bits in size, but for the gigE chips,
1534		 * it is 13 bits (since the max RX frame length is 16K).
1535		 * Unfortunately, all 32 bits in the status word
1536		 * were already used, so to make room for the extra
1537		 * length bit, RealTek took out the 'frame alignment
1538		 * error' bit and shifted the other status bits
1539		 * over one slot. The OWN, EOR, FS and LS bits are
1540		 * still in the same places. We have already extracted
1541		 * the frame length and checked the OWN bit, so rather
1542		 * than using an alternate bit mapping, we shift the
1543		 * status bits one space to the right so we can evaluate
1544		 * them using the 8169 status as though it was in the
1545		 * same format as that of the 8139C+.
1546		 */
1547		if (sc->rl_type == RL_8169)
1548			rxstat >>= 1;
1549
1550		if (rxstat & RL_RDESC_STAT_RXERRSUM) {
1551			ifp->if_ierrors++;
1552			/*
1553			 * If this is part of a multi-fragment packet,
1554			 * discard all the pieces.
1555			 */
1556			if (sc->rl_head != NULL) {
1557				m_freem(sc->rl_head);
1558				sc->rl_head = sc->rl_tail = NULL;
1559			}
1560			re_newbuf(sc, i, m);
1561			RL_DESC_INC(i);
1562			continue;
1563		}
1564
1565		/*
1566		 * If allocating a replacement mbuf fails,
1567		 * reload the current one.
1568		 */
1569
1570		if (re_newbuf(sc, i, NULL)) {
1571			ifp->if_ierrors++;
1572			if (sc->rl_head != NULL) {
1573				m_freem(sc->rl_head);
1574				sc->rl_head = sc->rl_tail = NULL;
1575			}
1576			re_newbuf(sc, i, m);
1577			RL_DESC_INC(i);
1578			continue;
1579		}
1580
1581		RL_DESC_INC(i);
1582
1583		if (sc->rl_head != NULL) {
1584			m->m_len = total_len % (MCLBYTES - ETHER_ALIGN);
1585			/*
1586			 * Special case: if there's 4 bytes or less
1587			 * in this buffer, the mbuf can be discarded:
1588			 * the last 4 bytes is the CRC, which we don't
1589			 * care about anyway.
1590			 */
1591			if (m->m_len <= ETHER_CRC_LEN) {
1592				sc->rl_tail->m_len -=
1593				    (ETHER_CRC_LEN - m->m_len);
1594				m_freem(m);
1595			} else {
1596				m->m_len -= ETHER_CRC_LEN;
1597				m->m_flags &= ~M_PKTHDR;
1598				sc->rl_tail->m_next = m;
1599			}
1600			m = sc->rl_head;
1601			sc->rl_head = sc->rl_tail = NULL;
1602			m->m_pkthdr.len = total_len - ETHER_CRC_LEN;
1603		} else
1604			m->m_pkthdr.len = m->m_len =
1605			    (total_len - ETHER_CRC_LEN);
1606
1607		ifp->if_ipackets++;
1608		m->m_pkthdr.rcvif = ifp;
1609
1610		/* Do RX checksumming if enabled */
1611
1612		if (ifp->if_capenable & IFCAP_RXCSUM) {
1613
1614			/* Check IP header checksum */
1615			if (rxstat & RL_RDESC_STAT_PROTOID)
1616				m->m_pkthdr.csum_flags |= CSUM_IP_CHECKED;
1617			if (!(rxstat & RL_RDESC_STAT_IPSUMBAD))
1618				m->m_pkthdr.csum_flags |= CSUM_IP_VALID;
1619
1620			/* Check TCP/UDP checksum */
1621			if ((RL_TCPPKT(rxstat) &&
1622			    !(rxstat & RL_RDESC_STAT_TCPSUMBAD)) ||
1623			    (RL_UDPPKT(rxstat) &&
1624			    !(rxstat & RL_RDESC_STAT_UDPSUMBAD))) {
1625				m->m_pkthdr.csum_flags |=
1626				    CSUM_DATA_VALID|CSUM_PSEUDO_HDR;
1627				m->m_pkthdr.csum_data = 0xffff;
1628			}
1629		}
1630
1631		if (rxvlan & RL_RDESC_VLANCTL_TAG)
1632			VLAN_INPUT_TAG(ifp, m,
1633			    ntohs((rxvlan & RL_RDESC_VLANCTL_DATA)), continue);
1634		(*ifp->if_input)(ifp, m);
1635	}
1636
1637	/* Flush the RX DMA ring */
1638
1639	bus_dmamap_sync(sc->rl_ldata.rl_rx_list_tag,
1640	    sc->rl_ldata.rl_rx_list_map,
1641	    BUS_DMASYNC_PREWRITE|BUS_DMASYNC_PREREAD);
1642
1643	sc->rl_ldata.rl_rx_prodidx = i;
1644
1645	return;
1646}
1647
1648static void
1649re_txeof(sc)
1650	struct rl_softc		*sc;
1651{
1652	struct ifnet		*ifp;
1653	u_int32_t		txstat;
1654	int			idx;
1655
1656	ifp = &sc->arpcom.ac_if;
1657	idx = sc->rl_ldata.rl_tx_considx;
1658
1659	/* Invalidate the TX descriptor list */
1660
1661	bus_dmamap_sync(sc->rl_ldata.rl_tx_list_tag,
1662	    sc->rl_ldata.rl_tx_list_map,
1663	    BUS_DMASYNC_POSTREAD);
1664
1665	while (idx != sc->rl_ldata.rl_tx_prodidx) {
1666
1667		txstat = le32toh(sc->rl_ldata.rl_tx_list[idx].rl_cmdstat);
1668		if (txstat & RL_TDESC_CMD_OWN)
1669			break;
1670
1671		/*
1672		 * We only stash mbufs in the last descriptor
1673		 * in a fragment chain, which also happens to
1674		 * be the only place where the TX status bits
1675		 * are valid.
1676		 */
1677
1678		if (txstat & RL_TDESC_CMD_EOF) {
1679			m_freem(sc->rl_ldata.rl_tx_mbuf[idx]);
1680			sc->rl_ldata.rl_tx_mbuf[idx] = NULL;
1681			bus_dmamap_unload(sc->rl_ldata.rl_mtag,
1682			    sc->rl_ldata.rl_tx_dmamap[idx]);
1683			if (txstat & (RL_TDESC_STAT_EXCESSCOL|
1684			    RL_TDESC_STAT_COLCNT))
1685				ifp->if_collisions++;
1686			if (txstat & RL_TDESC_STAT_TXERRSUM)
1687				ifp->if_oerrors++;
1688			else
1689				ifp->if_opackets++;
1690		}
1691		sc->rl_ldata.rl_tx_free++;
1692		RL_DESC_INC(idx);
1693	}
1694
1695	/* No changes made to the TX ring, so no flush needed */
1696
1697	if (idx != sc->rl_ldata.rl_tx_considx) {
1698		sc->rl_ldata.rl_tx_considx = idx;
1699		ifp->if_flags &= ~IFF_OACTIVE;
1700		ifp->if_timer = 0;
1701	}
1702
1703	/*
1704	 * If not all descriptors have been released reaped yet,
1705	 * reload the timer so that we will eventually get another
1706	 * interrupt that will cause us to re-enter this routine.
1707	 * This is done in case the transmitter has gone idle.
1708	 */
1709	if (sc->rl_ldata.rl_tx_free != RL_TX_DESC_CNT)
1710                CSR_WRITE_4(sc, RL_TIMERCNT, 1);
1711
1712	return;
1713}
1714
1715static void
1716re_tick(xsc)
1717	void			*xsc;
1718{
1719	struct rl_softc		*sc;
1720	struct mii_data		*mii;
1721
1722	sc = xsc;
1723	RL_LOCK(sc);
1724	mii = device_get_softc(sc->rl_miibus);
1725
1726	mii_tick(mii);
1727
1728	sc->rl_stat_ch = timeout(re_tick, sc, hz);
1729	RL_UNLOCK(sc);
1730
1731	return;
1732}
1733
1734#ifdef DEVICE_POLLING
1735static void
1736re_poll (struct ifnet *ifp, enum poll_cmd cmd, int count)
1737{
1738	struct rl_softc *sc = ifp->if_softc;
1739
1740	RL_LOCK(sc);
1741	if (cmd == POLL_DEREGISTER) { /* final call, enable interrupts */
1742		CSR_WRITE_2(sc, RL_IMR, RL_INTRS_CPLUS);
1743		goto done;
1744	}
1745
1746	sc->rxcycles = count;
1747	re_rxeof(sc);
1748	re_txeof(sc);
1749
1750	if (ifp->if_snd.ifq_head != NULL)
1751		(*ifp->if_start)(ifp);
1752
1753	if (cmd == POLL_AND_CHECK_STATUS) { /* also check status register */
1754		u_int16_t       status;
1755
1756		status = CSR_READ_2(sc, RL_ISR);
1757		if (status == 0xffff)
1758			goto done;
1759		if (status)
1760			CSR_WRITE_2(sc, RL_ISR, status);
1761
1762		/*
1763		 * XXX check behaviour on receiver stalls.
1764		 */
1765
1766		if (status & RL_ISR_SYSTEM_ERR) {
1767			re_reset(sc);
1768			re_init(sc);
1769		}
1770	}
1771done:
1772	RL_UNLOCK(sc);
1773}
1774#endif /* DEVICE_POLLING */
1775
1776static void
1777re_intr(arg)
1778	void			*arg;
1779{
1780	struct rl_softc		*sc;
1781	struct ifnet		*ifp;
1782	u_int16_t		status;
1783
1784	sc = arg;
1785
1786	if (sc->suspended) {
1787		return;
1788	}
1789
1790	RL_LOCK(sc);
1791	ifp = &sc->arpcom.ac_if;
1792
1793	if (!(ifp->if_flags & IFF_UP)) {
1794		RL_UNLOCK(sc);
1795		return;
1796	}
1797
1798#ifdef DEVICE_POLLING
1799	if  (ifp->if_flags & IFF_POLLING)
1800		goto done;
1801	if (ether_poll_register(re_poll, ifp)) { /* ok, disable interrupts */
1802		CSR_WRITE_2(sc, RL_IMR, 0x0000);
1803		re_poll(ifp, 0, 1);
1804		goto done;
1805	}
1806#endif /* DEVICE_POLLING */
1807
1808	for (;;) {
1809
1810		status = CSR_READ_2(sc, RL_ISR);
1811		/* If the card has gone away the read returns 0xffff. */
1812		if (status == 0xffff)
1813			break;
1814		if (status)
1815			CSR_WRITE_2(sc, RL_ISR, status);
1816
1817		if ((status & RL_INTRS_CPLUS) == 0)
1818			break;
1819
1820		if (status & RL_ISR_RX_OK)
1821			re_rxeof(sc);
1822
1823		if (status & RL_ISR_RX_ERR)
1824			re_rxeof(sc);
1825
1826		if ((status & RL_ISR_TIMEOUT_EXPIRED) ||
1827		    (status & RL_ISR_TX_ERR) ||
1828		    (status & RL_ISR_TX_DESC_UNAVAIL))
1829			re_txeof(sc);
1830
1831		if (status & RL_ISR_SYSTEM_ERR) {
1832			re_reset(sc);
1833			re_init(sc);
1834		}
1835
1836		if (status & RL_ISR_LINKCHG) {
1837			untimeout(re_tick, sc, sc->rl_stat_ch);
1838			re_tick(sc);
1839		}
1840	}
1841
1842	if (ifp->if_snd.ifq_head != NULL)
1843		(*ifp->if_start)(ifp);
1844
1845#ifdef DEVICE_POLLING
1846done:
1847#endif
1848	RL_UNLOCK(sc);
1849
1850	return;
1851}
1852
1853static int
1854re_encap(sc, m_head, idx)
1855	struct rl_softc		*sc;
1856	struct mbuf		*m_head;
1857	int			*idx;
1858{
1859	struct mbuf		*m_new = NULL;
1860	struct rl_dmaload_arg	arg;
1861	bus_dmamap_t		map;
1862	int			error;
1863	struct m_tag		*mtag;
1864
1865	if (sc->rl_ldata.rl_tx_free < 4)
1866		return(EFBIG);
1867
1868	/*
1869	 * Set up checksum offload. Note: checksum offload bits must
1870	 * appear in all descriptors of a multi-descriptor transmit
1871	 * attempt. (This is according to testing done with an 8169
1872	 * chip. I'm not sure if this is a requirement or a bug.)
1873	 */
1874
1875	arg.rl_flags = 0;
1876
1877	if (m_head->m_pkthdr.csum_flags & CSUM_IP)
1878		arg.rl_flags |= RL_TDESC_CMD_IPCSUM;
1879	if (m_head->m_pkthdr.csum_flags & CSUM_TCP)
1880		arg.rl_flags |= RL_TDESC_CMD_TCPCSUM;
1881	if (m_head->m_pkthdr.csum_flags & CSUM_UDP)
1882		arg.rl_flags |= RL_TDESC_CMD_UDPCSUM;
1883
1884	arg.sc = sc;
1885	arg.rl_idx = *idx;
1886	arg.rl_maxsegs = sc->rl_ldata.rl_tx_free;
1887	arg.rl_ring = sc->rl_ldata.rl_tx_list;
1888
1889	map = sc->rl_ldata.rl_tx_dmamap[*idx];
1890	error = bus_dmamap_load_mbuf(sc->rl_ldata.rl_mtag, map,
1891	    m_head, re_dma_map_desc, &arg, BUS_DMA_NOWAIT);
1892
1893	if (error && error != EFBIG) {
1894		printf("re%d: can't map mbuf (error %d)\n", sc->rl_unit, error);
1895		return(ENOBUFS);
1896	}
1897
1898	/* Too many segments to map, coalesce into a single mbuf */
1899
1900	if (error || arg.rl_maxsegs == 0) {
1901		m_new = m_defrag(m_head, M_DONTWAIT);
1902		if (m_new == NULL)
1903			return(1);
1904		else
1905			m_head = m_new;
1906
1907		arg.sc = sc;
1908		arg.rl_idx = *idx;
1909		arg.rl_maxsegs = sc->rl_ldata.rl_tx_free;
1910		arg.rl_ring = sc->rl_ldata.rl_tx_list;
1911
1912		error = bus_dmamap_load_mbuf(sc->rl_ldata.rl_mtag, map,
1913		    m_head, re_dma_map_desc, &arg, BUS_DMA_NOWAIT);
1914		if (error) {
1915			printf("re%d: can't map mbuf (error %d)\n",
1916			    sc->rl_unit, error);
1917			return(EFBIG);
1918		}
1919	}
1920
1921	/*
1922	 * Insure that the map for this transmission
1923	 * is placed at the array index of the last descriptor
1924	 * in this chain.
1925	 */
1926	sc->rl_ldata.rl_tx_dmamap[*idx] =
1927	    sc->rl_ldata.rl_tx_dmamap[arg.rl_idx];
1928	sc->rl_ldata.rl_tx_dmamap[arg.rl_idx] = map;
1929
1930	sc->rl_ldata.rl_tx_mbuf[arg.rl_idx] = m_head;
1931	sc->rl_ldata.rl_tx_free -= arg.rl_maxsegs;
1932
1933	/*
1934	 * Set up hardware VLAN tagging. Note: vlan tag info must
1935	 * appear in the first descriptor of a multi-descriptor
1936	 * transmission attempt.
1937	 */
1938
1939	mtag = VLAN_OUTPUT_TAG(&sc->arpcom.ac_if, m_head);
1940	if (mtag != NULL)
1941		sc->rl_ldata.rl_tx_list[*idx].rl_vlanctl =
1942		    htole32(htons(VLAN_TAG_VALUE(mtag)) | RL_TDESC_VLANCTL_TAG);
1943
1944	/* Transfer ownership of packet to the chip. */
1945
1946	sc->rl_ldata.rl_tx_list[arg.rl_idx].rl_cmdstat |=
1947	    htole32(RL_TDESC_CMD_OWN);
1948	if (*idx != arg.rl_idx)
1949		sc->rl_ldata.rl_tx_list[*idx].rl_cmdstat |=
1950		    htole32(RL_TDESC_CMD_OWN);
1951
1952	RL_DESC_INC(arg.rl_idx);
1953	*idx = arg.rl_idx;
1954
1955	return(0);
1956}
1957
1958/*
1959 * Main transmit routine for C+ and gigE NICs.
1960 */
1961
1962static void
1963re_start(ifp)
1964	struct ifnet		*ifp;
1965{
1966	struct rl_softc		*sc;
1967	struct mbuf		*m_head = NULL;
1968	int			idx;
1969
1970	sc = ifp->if_softc;
1971	RL_LOCK(sc);
1972
1973	idx = sc->rl_ldata.rl_tx_prodidx;
1974
1975	while (sc->rl_ldata.rl_tx_mbuf[idx] == NULL) {
1976		IF_DEQUEUE(&ifp->if_snd, m_head);
1977		if (m_head == NULL)
1978			break;
1979
1980		if (re_encap(sc, m_head, &idx)) {
1981			IF_PREPEND(&ifp->if_snd, m_head);
1982			ifp->if_flags |= IFF_OACTIVE;
1983			break;
1984		}
1985
1986		/*
1987		 * If there's a BPF listener, bounce a copy of this frame
1988		 * to him.
1989		 */
1990		BPF_MTAP(ifp, m_head);
1991	}
1992
1993	/* Flush the TX descriptors */
1994
1995	bus_dmamap_sync(sc->rl_ldata.rl_tx_list_tag,
1996	    sc->rl_ldata.rl_tx_list_map,
1997	    BUS_DMASYNC_PREWRITE|BUS_DMASYNC_PREREAD);
1998
1999	sc->rl_ldata.rl_tx_prodidx = idx;
2000
2001	/*
2002	 * RealTek put the TX poll request register in a different
2003	 * location on the 8169 gigE chip. I don't know why.
2004	 */
2005
2006	if (sc->rl_type == RL_8169)
2007		CSR_WRITE_2(sc, RL_GTXSTART, RL_TXSTART_START);
2008	else
2009		CSR_WRITE_2(sc, RL_TXSTART, RL_TXSTART_START);
2010
2011	/*
2012	 * Use the countdown timer for interrupt moderation.
2013	 * 'TX done' interrupts are disabled. Instead, we reset the
2014	 * countdown timer, which will begin counting until it hits
2015	 * the value in the TIMERINT register, and then trigger an
2016	 * interrupt. Each time we write to the TIMERCNT register,
2017	 * the timer count is reset to 0.
2018	 */
2019	CSR_WRITE_4(sc, RL_TIMERCNT, 1);
2020
2021	RL_UNLOCK(sc);
2022
2023	/*
2024	 * Set a timeout in case the chip goes out to lunch.
2025	 */
2026	ifp->if_timer = 5;
2027
2028	return;
2029}
2030
2031static void
2032re_init(xsc)
2033	void			*xsc;
2034{
2035	struct rl_softc		*sc = xsc;
2036	struct ifnet		*ifp = &sc->arpcom.ac_if;
2037	struct mii_data		*mii;
2038	u_int32_t		rxcfg = 0;
2039
2040	RL_LOCK(sc);
2041	mii = device_get_softc(sc->rl_miibus);
2042
2043	/*
2044	 * Cancel pending I/O and free all RX/TX buffers.
2045	 */
2046	re_stop(sc);
2047
2048	/*
2049	 * Init our MAC address.  Even though the chipset
2050	 * documentation doesn't mention it, we need to enter "Config
2051	 * register write enable" mode to modify the ID registers.
2052	 */
2053	CSR_WRITE_1(sc, RL_EECMD, RL_EEMODE_WRITECFG);
2054	CSR_WRITE_STREAM_4(sc, RL_IDR0,
2055	    *(u_int32_t *)(&sc->arpcom.ac_enaddr[0]));
2056	CSR_WRITE_STREAM_4(sc, RL_IDR4,
2057	    *(u_int32_t *)(&sc->arpcom.ac_enaddr[4]));
2058	CSR_WRITE_1(sc, RL_EECMD, RL_EEMODE_OFF);
2059
2060	/*
2061	 * For C+ mode, initialize the RX descriptors and mbufs.
2062	 */
2063	re_rx_list_init(sc);
2064	re_tx_list_init(sc);
2065
2066	/*
2067	 * Enable transmit and receive.
2068	 */
2069	CSR_WRITE_1(sc, RL_COMMAND, RL_CMD_TX_ENB|RL_CMD_RX_ENB);
2070
2071	/*
2072	 * Set the initial TX and RX configuration.
2073	 */
2074	if (sc->rl_testmode)
2075		CSR_WRITE_4(sc, RL_TXCFG, RL_TXCFG_CONFIG|RL_LOOPTEST_ON);
2076	else
2077		CSR_WRITE_4(sc, RL_TXCFG, RL_TXCFG_CONFIG);
2078	CSR_WRITE_4(sc, RL_RXCFG, RL_RXCFG_CONFIG);
2079
2080	/* Set the individual bit to receive frames for this host only. */
2081	rxcfg = CSR_READ_4(sc, RL_RXCFG);
2082	rxcfg |= RL_RXCFG_RX_INDIV;
2083
2084	/* If we want promiscuous mode, set the allframes bit. */
2085	if (ifp->if_flags & IFF_PROMISC) {
2086		rxcfg |= RL_RXCFG_RX_ALLPHYS;
2087		CSR_WRITE_4(sc, RL_RXCFG, rxcfg);
2088	} else {
2089		rxcfg &= ~RL_RXCFG_RX_ALLPHYS;
2090		CSR_WRITE_4(sc, RL_RXCFG, rxcfg);
2091	}
2092
2093	/*
2094	 * Set capture broadcast bit to capture broadcast frames.
2095	 */
2096	if (ifp->if_flags & IFF_BROADCAST) {
2097		rxcfg |= RL_RXCFG_RX_BROAD;
2098		CSR_WRITE_4(sc, RL_RXCFG, rxcfg);
2099	} else {
2100		rxcfg &= ~RL_RXCFG_RX_BROAD;
2101		CSR_WRITE_4(sc, RL_RXCFG, rxcfg);
2102	}
2103
2104	/*
2105	 * Program the multicast filter, if necessary.
2106	 */
2107	re_setmulti(sc);
2108
2109#ifdef DEVICE_POLLING
2110	/*
2111	 * Disable interrupts if we are polling.
2112	 */
2113	if (ifp->if_flags & IFF_POLLING)
2114		CSR_WRITE_2(sc, RL_IMR, 0);
2115	else	/* otherwise ... */
2116#endif /* DEVICE_POLLING */
2117	/*
2118	 * Enable interrupts.
2119	 */
2120	if (sc->rl_testmode)
2121		CSR_WRITE_2(sc, RL_IMR, 0);
2122	else
2123		CSR_WRITE_2(sc, RL_IMR, RL_INTRS_CPLUS);
2124
2125	/* Set initial TX threshold */
2126	sc->rl_txthresh = RL_TX_THRESH_INIT;
2127
2128	/* Start RX/TX process. */
2129	CSR_WRITE_4(sc, RL_MISSEDPKT, 0);
2130#ifdef notdef
2131	/* Enable receiver and transmitter. */
2132	CSR_WRITE_1(sc, RL_COMMAND, RL_CMD_TX_ENB|RL_CMD_RX_ENB);
2133#endif
2134	/*
2135	 * If this is a C+ capable chip, enable C+ RX and TX mode,
2136	 * and load the addresses of the RX and TX lists into the chip.
2137	 */
2138	CSR_WRITE_2(sc, RL_CPLUS_CMD, RL_CPLUSCMD_RXENB|
2139	    RL_CPLUSCMD_TXENB|RL_CPLUSCMD_PCI_MRW|
2140	    RL_CPLUSCMD_VLANSTRIP|
2141	    (ifp->if_capenable & IFCAP_RXCSUM ?
2142	    RL_CPLUSCMD_RXCSUM_ENB : 0));
2143
2144	CSR_WRITE_4(sc, RL_RXLIST_ADDR_HI,
2145	    RL_ADDR_HI(sc->rl_ldata.rl_rx_list_addr));
2146	CSR_WRITE_4(sc, RL_RXLIST_ADDR_LO,
2147	    RL_ADDR_LO(sc->rl_ldata.rl_rx_list_addr));
2148
2149	CSR_WRITE_4(sc, RL_TXLIST_ADDR_HI,
2150	    RL_ADDR_HI(sc->rl_ldata.rl_tx_list_addr));
2151	CSR_WRITE_4(sc, RL_TXLIST_ADDR_LO,
2152	    RL_ADDR_LO(sc->rl_ldata.rl_tx_list_addr));
2153
2154	CSR_WRITE_1(sc, RL_EARLY_TX_THRESH, 16);
2155
2156	/*
2157	 * Initialize the timer interrupt register so that
2158	 * a timer interrupt will be generated once the timer
2159	 * reaches a certain number of ticks. The timer is
2160	 * reloaded on each transmit. This gives us TX interrupt
2161	 * moderation, which dramatically improves TX frame rate.
2162	 */
2163
2164	if (sc->rl_type == RL_8169)
2165		CSR_WRITE_4(sc, RL_TIMERINT_8169, 0x800);
2166	else
2167		CSR_WRITE_4(sc, RL_TIMERINT, 0x400);
2168
2169	/*
2170	 * For 8169 gigE NICs, set the max allowed RX packet
2171	 * size so we can receive jumbo frames.
2172	 */
2173	if (sc->rl_type == RL_8169)
2174		CSR_WRITE_2(sc, RL_MAXRXPKTLEN, 16383);
2175
2176	if (sc->rl_testmode) {
2177		RL_UNLOCK(sc);
2178		return;
2179	}
2180
2181	mii_mediachg(mii);
2182
2183	CSR_WRITE_1(sc, RL_CFG1, RL_CFG1_DRVLOAD|RL_CFG1_FULLDUPLEX);
2184
2185	ifp->if_flags |= IFF_RUNNING;
2186	ifp->if_flags &= ~IFF_OACTIVE;
2187
2188	sc->rl_stat_ch = timeout(re_tick, sc, hz);
2189	RL_UNLOCK(sc);
2190
2191	return;
2192}
2193
2194/*
2195 * Set media options.
2196 */
2197static int
2198re_ifmedia_upd(ifp)
2199	struct ifnet		*ifp;
2200{
2201	struct rl_softc		*sc;
2202	struct mii_data		*mii;
2203
2204	sc = ifp->if_softc;
2205	mii = device_get_softc(sc->rl_miibus);
2206	mii_mediachg(mii);
2207
2208	return(0);
2209}
2210
2211/*
2212 * Report current media status.
2213 */
2214static void
2215re_ifmedia_sts(ifp, ifmr)
2216	struct ifnet		*ifp;
2217	struct ifmediareq	*ifmr;
2218{
2219	struct rl_softc		*sc;
2220	struct mii_data		*mii;
2221
2222	sc = ifp->if_softc;
2223	mii = device_get_softc(sc->rl_miibus);
2224
2225	mii_pollstat(mii);
2226	ifmr->ifm_active = mii->mii_media_active;
2227	ifmr->ifm_status = mii->mii_media_status;
2228
2229	return;
2230}
2231
2232static int
2233re_ioctl(ifp, command, data)
2234	struct ifnet		*ifp;
2235	u_long			command;
2236	caddr_t			data;
2237{
2238	struct rl_softc		*sc = ifp->if_softc;
2239	struct ifreq		*ifr = (struct ifreq *) data;
2240	struct mii_data		*mii;
2241	int			error = 0;
2242
2243	RL_LOCK(sc);
2244
2245	switch(command) {
2246	case SIOCSIFMTU:
2247		if (ifr->ifr_mtu > RL_JUMBO_MTU)
2248			error = EINVAL;
2249		ifp->if_mtu = ifr->ifr_mtu;
2250		break;
2251	case SIOCSIFFLAGS:
2252		if (ifp->if_flags & IFF_UP) {
2253			re_init(sc);
2254		} else {
2255			if (ifp->if_flags & IFF_RUNNING)
2256				re_stop(sc);
2257		}
2258		error = 0;
2259		break;
2260	case SIOCADDMULTI:
2261	case SIOCDELMULTI:
2262		re_setmulti(sc);
2263		error = 0;
2264		break;
2265	case SIOCGIFMEDIA:
2266	case SIOCSIFMEDIA:
2267		mii = device_get_softc(sc->rl_miibus);
2268		error = ifmedia_ioctl(ifp, ifr, &mii->mii_media, command);
2269		break;
2270	case SIOCSIFCAP:
2271		ifp->if_capenable = ifr->ifr_reqcap;
2272		if (ifp->if_capenable & IFCAP_TXCSUM)
2273			ifp->if_hwassist = RE_CSUM_FEATURES;
2274		else
2275			ifp->if_hwassist = 0;
2276		if (ifp->if_flags & IFF_RUNNING)
2277			re_init(sc);
2278		break;
2279	default:
2280		error = ether_ioctl(ifp, command, data);
2281		break;
2282	}
2283
2284	RL_UNLOCK(sc);
2285
2286	return(error);
2287}
2288
2289static void
2290re_watchdog(ifp)
2291	struct ifnet		*ifp;
2292{
2293	struct rl_softc		*sc;
2294
2295	sc = ifp->if_softc;
2296	RL_LOCK(sc);
2297	printf("re%d: watchdog timeout\n", sc->rl_unit);
2298	ifp->if_oerrors++;
2299
2300	re_txeof(sc);
2301	re_rxeof(sc);
2302
2303	re_init(sc);
2304
2305	RL_UNLOCK(sc);
2306
2307	return;
2308}
2309
2310/*
2311 * Stop the adapter and free any mbufs allocated to the
2312 * RX and TX lists.
2313 */
2314static void
2315re_stop(sc)
2316	struct rl_softc		*sc;
2317{
2318	register int		i;
2319	struct ifnet		*ifp;
2320
2321	RL_LOCK(sc);
2322	ifp = &sc->arpcom.ac_if;
2323	ifp->if_timer = 0;
2324
2325	untimeout(re_tick, sc, sc->rl_stat_ch);
2326	ifp->if_flags &= ~(IFF_RUNNING | IFF_OACTIVE);
2327#ifdef DEVICE_POLLING
2328	ether_poll_deregister(ifp);
2329#endif /* DEVICE_POLLING */
2330
2331	CSR_WRITE_1(sc, RL_COMMAND, 0x00);
2332	CSR_WRITE_2(sc, RL_IMR, 0x0000);
2333
2334	if (sc->rl_head != NULL) {
2335		m_freem(sc->rl_head);
2336		sc->rl_head = sc->rl_tail = NULL;
2337	}
2338
2339	/* Free the TX list buffers. */
2340
2341	for (i = 0; i < RL_TX_DESC_CNT; i++) {
2342		if (sc->rl_ldata.rl_tx_mbuf[i] != NULL) {
2343			bus_dmamap_unload(sc->rl_ldata.rl_mtag,
2344			    sc->rl_ldata.rl_tx_dmamap[i]);
2345			m_freem(sc->rl_ldata.rl_tx_mbuf[i]);
2346			sc->rl_ldata.rl_tx_mbuf[i] = NULL;
2347		}
2348	}
2349
2350	/* Free the RX list buffers. */
2351
2352	for (i = 0; i < RL_RX_DESC_CNT; i++) {
2353		if (sc->rl_ldata.rl_rx_mbuf[i] != NULL) {
2354			bus_dmamap_unload(sc->rl_ldata.rl_mtag,
2355			    sc->rl_ldata.rl_rx_dmamap[i]);
2356			m_freem(sc->rl_ldata.rl_rx_mbuf[i]);
2357			sc->rl_ldata.rl_rx_mbuf[i] = NULL;
2358		}
2359	}
2360
2361	RL_UNLOCK(sc);
2362	return;
2363}
2364
2365/*
2366 * Device suspend routine.  Stop the interface and save some PCI
2367 * settings in case the BIOS doesn't restore them properly on
2368 * resume.
2369 */
2370static int
2371re_suspend(dev)
2372	device_t		dev;
2373{
2374	register int		i;
2375	struct rl_softc		*sc;
2376
2377	sc = device_get_softc(dev);
2378
2379	re_stop(sc);
2380
2381	for (i = 0; i < 5; i++)
2382		sc->saved_maps[i] = pci_read_config(dev, PCIR_MAPS + i * 4, 4);
2383	sc->saved_biosaddr = pci_read_config(dev, PCIR_BIOS, 4);
2384	sc->saved_intline = pci_read_config(dev, PCIR_INTLINE, 1);
2385	sc->saved_cachelnsz = pci_read_config(dev, PCIR_CACHELNSZ, 1);
2386	sc->saved_lattimer = pci_read_config(dev, PCIR_LATTIMER, 1);
2387
2388	sc->suspended = 1;
2389
2390	return (0);
2391}
2392
2393/*
2394 * Device resume routine.  Restore some PCI settings in case the BIOS
2395 * doesn't, re-enable busmastering, and restart the interface if
2396 * appropriate.
2397 */
2398static int
2399re_resume(dev)
2400	device_t		dev;
2401{
2402	register int		i;
2403	struct rl_softc		*sc;
2404	struct ifnet		*ifp;
2405
2406	sc = device_get_softc(dev);
2407	ifp = &sc->arpcom.ac_if;
2408
2409	/* better way to do this? */
2410	for (i = 0; i < 5; i++)
2411		pci_write_config(dev, PCIR_MAPS + i * 4, sc->saved_maps[i], 4);
2412	pci_write_config(dev, PCIR_BIOS, sc->saved_biosaddr, 4);
2413	pci_write_config(dev, PCIR_INTLINE, sc->saved_intline, 1);
2414	pci_write_config(dev, PCIR_CACHELNSZ, sc->saved_cachelnsz, 1);
2415	pci_write_config(dev, PCIR_LATTIMER, sc->saved_lattimer, 1);
2416
2417	/* reenable busmastering */
2418	pci_enable_busmaster(dev);
2419	pci_enable_io(dev, RL_RES);
2420
2421	/* reinitialize interface if necessary */
2422	if (ifp->if_flags & IFF_UP)
2423		re_init(sc);
2424
2425	sc->suspended = 0;
2426
2427	return (0);
2428}
2429
2430/*
2431 * Stop all chip I/O so that the kernel's probe routines don't
2432 * get confused by errant DMAs when rebooting.
2433 */
2434static void
2435re_shutdown(dev)
2436	device_t		dev;
2437{
2438	struct rl_softc		*sc;
2439
2440	sc = device_get_softc(dev);
2441
2442	re_stop(sc);
2443
2444	return;
2445}
2446