if_hatmvar.h revision 121677
1/*
2 * Copyright (c) 2001-2003
3 *	Fraunhofer Institute for Open Communication Systems (FhG Fokus).
4 * 	All rights reserved.
5 *
6 * Redistribution and use in source and binary forms, with or without
7 * modification, are permitted provided that the following conditions
8 * are met:
9 * 1. Redistributions of source code must retain the above copyright
10 *    notice, this list of conditions and the following disclaimer.
11 * 2. Redistributions in binary form must reproduce the above copyright
12 *    notice, this list of conditions and the following disclaimer in the
13 *    documentation and/or other materials provided with the distribution.
14 *
15 * THIS SOFTWARE IS PROVIDED BY THE AUTHOR AND CONTRIBUTORS ``AS IS'' AND
16 * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
17 * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
18 * ARE DISCLAIMED.  IN NO EVENT SHALL THE AUTHOR OR CONTRIBUTORS BE LIABLE
19 * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
20 * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
21 * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
22 * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
23 * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
24 * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
25 * SUCH DAMAGE.
26 *
27 * Author: Hartmut Brandt <harti@freebsd.org>
28 *
29 * $FreeBSD: head/sys/dev/hatm/if_hatmvar.h 121677 2003-10-29 13:21:38Z harti $
30 *
31 * Fore HE driver for NATM
32 */
33
34/*
35 * Debug statistics of the HE driver
36 */
37struct istats {
38	uint32_t	tdprq_full;
39	uint32_t	hbuf_error;
40	uint32_t	crc_error;
41	uint32_t	len_error;
42	uint32_t	flow_closed;
43	uint32_t	flow_drop;
44	uint32_t	tpd_no_mem;
45	uint32_t	rx_seg;
46	uint32_t	empty_hbuf;
47	uint32_t	short_aal5;
48	uint32_t	badlen_aal5;
49	uint32_t	bug_bad_isw;
50	uint32_t	bug_no_irq_upd;
51	uint32_t	itype_tbrq;
52	uint32_t	itype_tpd;
53	uint32_t	itype_rbps;
54	uint32_t	itype_rbpl;
55	uint32_t	itype_rbrq;
56	uint32_t	itype_rbrqt;
57	uint32_t	itype_unknown;
58	uint32_t	itype_phys;
59	uint32_t	itype_err;
60	uint32_t	defrag;
61	uint32_t	mcc;
62	uint32_t	oec;
63	uint32_t	dcc;
64	uint32_t	cec;
65	uint32_t	no_rcv_mbuf;
66};
67
68/* Card memory layout parameters */
69#define HE_CONFIG_MEM_LAYOUT {						\
70	{			/* 155 */				\
71	  20,			/* cells_per_row */			\
72	  1024,			/* bytes_per_row */			\
73	  512,			/* r0_numrows */			\
74	  1018,			/* tx_numrows */			\
75	  512,			/* r1_numrows */			\
76	  6,			/* r0_startrow */			\
77	  2			/* cells_per_lbuf */			\
78	}, {			/* 622 */				\
79	  40,			/* cells_per_row */			\
80	  2048,			/* bytes_per_row */			\
81	  256,			/* r0_numrows */			\
82	  512,			/* tx_numrows */			\
83	  256,			/* r1_numrows */			\
84	  0,			/* r0_startrow */			\
85	  4			/* cells_per_lbuf */			\
86	}								\
87}
88
89/*********************************************************************/
90struct hatm_softc;
91
92/*
93 * A chunk of DMA-able memory
94 */
95struct dmamem {
96	u_int		size;		/* in bytes */
97	u_int		align;		/* alignement */
98	bus_dma_tag_t	tag;		/* DMA tag */
99	void		*base;		/* the memory */
100	bus_addr_t	paddr;		/* physical address */
101	bus_dmamap_t	map;		/* the MAP */
102};
103
104/*
105 * RBP (Receive Buffer Pool) queue entry and queue.
106 */
107struct herbp {
108	u_int		size;		/* RBP number of entries (power of two) */
109	u_int		thresh;		/* interrupt treshold */
110	uint32_t	bsize;		/* buffer size in bytes */
111	u_int		offset;		/* free space at start for small bufs */
112	uint32_t	mask;		/* mask for index */
113	struct dmamem	mem;		/* the queue area */
114	struct he_rbpen	*rbp;
115	uint32_t	head, tail;	/* head and tail */
116};
117
118/*
119 * RBRQ (Receive Buffer Return Queue) entry and queue.
120 */
121struct herbrq {
122	u_int		size;		/* number of entries */
123	u_int		thresh;		/* interrupt threshold */
124	u_int		tout;		/* timeout value */
125	u_int		pcnt;		/* packet count threshold */
126	struct dmamem	mem;		/* memory */
127	struct he_rbrqen *rbrq;
128	uint32_t	head;		/* driver end */
129};
130
131/*
132 * TPDRQ (Transmit Packet Descriptor Ready Queue) entry and queue
133 */
134struct hetpdrq {
135	u_int		size;		/* number of entries */
136	struct dmamem	mem;		/* memory */
137	struct he_tpdrqen *tpdrq;
138	u_int		head;		/* head (copy of adapter) */
139	u_int		tail;		/* written back to adapter */
140};
141
142/*
143 * TBRQ (Transmit Buffer Return Queue) entry and queue
144 */
145struct hetbrq {
146	u_int		size;		/* number of entries */
147	u_int		thresh;		/* interrupt threshold */
148	struct dmamem	mem;		/* memory */
149	struct he_tbrqen *tbrq;
150	u_int		head;		/* adapter end */
151};
152
153/*==================================================================*/
154
155/*
156 * TPDs are 32 byte and must be aligned on 64 byte boundaries. That means,
157 * that half of the space is free. We use this space to plug in a link for
158 * the list of free TPDs. Note, that the m_act member of the mbufs contain
159 * a pointer to the dmamap.
160 *
161 * The maximum number of TDPs is the size of the common transmit packet
162 * descriptor ready queue plus the sizes of the transmit buffer return queues
163 * (currently only queue 0). We allocate and map these TPD when initializing
164 * the card. We also allocate on DMA map for each TPD. Only the map in the
165 * last TPD of a packets is used when a packet is transmitted.
166 * This is signalled by having the mbuf member of this TPD non-zero and
167 * pointing to the mbuf.
168 */
169#define HE_TPD_SIZE		64
170struct tpd {
171	struct he_tpd		tpd;	/* at beginning */
172	SLIST_ENTRY(tpd)	link;	/* free cid list link */
173	struct mbuf		*mbuf;	/* the buf chain */
174	bus_dmamap_t		map;	/* map */
175	uint32_t		cid;	/* CID */
176	uint16_t		no;	/* number of this tpd */
177};
178SLIST_HEAD(tpd_list, tpd);
179
180#define TPD_SET_USED(SC, I) do {				\
181	(SC)->tpd_used[(I) / 8] |= (1 << ((I) % 8));		\
182    } while (0)
183
184#define TPD_CLR_USED(SC, I) do {				\
185	(SC)->tpd_used[(I) / 8] &= ~(1 << ((I) % 8));		\
186    } while (0)
187
188#define TPD_TST_USED(SC, I) ((SC)->tpd_used[(I) / 8] & (1 << ((I) % 8)))
189
190#define TPD_ADDR(SC, I) ((struct tpd *)((char *)sc->tpds.base +	\
191    (I) * HE_TPD_SIZE))
192
193/*==================================================================*/
194
195/*
196 * External MBUFs. The card needs a lot of mbufs in the pools for high
197 * performance. The problem with using mbufs directly is that we would need
198 * a dmamap for each of the mbufs. This can exhaust iommu space on the sparc
199 * and it eats also a lot of processing time. So we use external mbufs
200 * for the small buffers and clusters for the large buffers.
201 * For receive group 0 we use 5 ATM cells, for group 1 one (52 byte) ATM
202 * cell. The mbuf storage is allocated pagewise and one dmamap is used per
203 * page.
204 *
205 * The handle we give to the card for the small buffers is a word combined
206 * of the page number and the number of the chunk in the page. This restricts
207 * the number of chunks per page to 256 (8 bit) and the number of pages to
208 * 65536 (16 bits).
209 *
210 * A chunk may be in one of three states: free, on the card and floating around
211 * in the system. If it is free, it is on one of the two free lists and
212 * start with a struct mbufx_free. Each page has a bitmap that tracks where
213 * its chunks are.
214 *
215 * For large buffers we use mbuf clusters. Here we have two problems: we need
216 * to track the buffers on the card (in the case we want to stop it) and
217 * we need to map the 64bit mbuf address to a 26bit handle for 64-bit machines.
218 * The card uses the buffers in the order we give it to the card. Therefor
219 * we can use a private array holding pointers to the mbufs as a circular
220 * queue for both tasks. This is done with the lbufs member of softc. The
221 * handle for these buffer is the lbufs index ored with a flag.
222 */
223#define MBUF0_SIZE	(5 * 48)	/* 240 */
224#define MBUF1_SIZE	(52)
225
226#define MBUF0_CHUNK	256		/* 16 free bytes */
227#define MBUF1_CHUNK	96		/* 44 free bytes */
228#ifdef XXX
229#define MBUF0_OFFSET	(MBUF0_CHUNK - sizeof(struct mbuf_chunk_hdr) \
230    - MBUF0_SIZE)
231#else
232#define MBUF0_OFFSET	0
233#endif
234#define MBUF1_OFFSET	(MBUF1_CHUNK - sizeof(struct mbuf_chunk_hdr) \
235    - MBUF1_SIZE)
236#define MBUFL_OFFSET	16		/* two pointers for HARP */
237
238#define MBUF_ALLOC_SIZE	(PAGE_SIZE)
239
240/* each allocated page has one of these structures at its very end. */
241struct mbuf_page_hdr {
242	uint8_t		card[32];	/* bitmap for on-card */
243	uint16_t	nchunks;	/* chunks on this page */
244	bus_dmamap_t	map;		/* the DMA MAP */
245	uint32_t	phys;		/* physical base address */
246	uint32_t	hdroff;		/* chunk header offset */
247	uint32_t	chunksize;	/* chunk size */
248};
249struct mbuf_page {
250	char	storage[MBUF_ALLOC_SIZE - sizeof(struct mbuf_page_hdr)];
251	struct mbuf_page_hdr	hdr;
252};
253
254/* numbers per page */
255#define MBUF0_PER_PAGE	((MBUF_ALLOC_SIZE - sizeof(struct mbuf_page_hdr)) / \
256    MBUF0_CHUNK)
257#define MBUF1_PER_PAGE	((MBUF_ALLOC_SIZE - sizeof(struct mbuf_page_hdr)) / \
258    MBUF1_CHUNK)
259
260#define MBUF_CLR_BIT(ARRAY, BIT) ((ARRAY)[(BIT) / 8] &= ~(1 << ((BIT) % 8)))
261#define MBUF_SET_BIT(ARRAY, BIT) ((ARRAY)[(BIT) / 8] |= (1 << ((BIT) % 8)))
262#define MBUF_TST_BIT(ARRAY, BIT) ((ARRAY)[(BIT) / 8] & (1 << ((BIT) % 8)))
263
264#define MBUF_MAKE_HANDLE(PAGENO, CHUNKNO) \
265	(((PAGENO) << 10) | (CHUNKNO))
266
267#define MBUF_PARSE_HANDLE(HANDLE, PAGENO, CHUNKNO) do {	\
268	(CHUNKNO) = (HANDLE) & 0x3ff;			\
269	(PAGENO) = ((HANDLE) >> 10) & 0x3ff;		\
270    } while (0)
271
272#define MBUF_LARGE_FLAG	(1 << 20)
273
274/* chunks have the following structure at the end (4 byte) */
275struct mbuf_chunk_hdr {
276	uint16_t		pageno;
277	uint16_t		chunkno;
278	u_int			ref_cnt;
279};
280
281#define MBUFX_STORAGE_SIZE(X) (MBUF##X##_CHUNK	\
282    - sizeof(struct mbuf_chunk_hdr))
283
284struct mbuf0_chunk {
285	char			storage[MBUFX_STORAGE_SIZE(0)];
286	struct mbuf_chunk_hdr	hdr;
287};
288
289struct mbuf1_chunk {
290	char			storage[MBUFX_STORAGE_SIZE(1)];
291	struct mbuf_chunk_hdr	hdr;
292};
293
294struct mbufx_free {
295	struct mbufx_free	*link;
296};
297
298/*==================================================================*/
299
300/*
301 * Interrupt queue
302 */
303struct heirq {
304	u_int		size;	/* number of entries */
305	u_int		thresh;	/* re-interrupt threshold */
306	u_int		line;	/* interrupt line to use */
307	struct dmamem	mem;	/* interrupt queues */
308	uint32_t *	irq;	/* interrupt queue */
309	uint32_t 	head;	/* head index */
310	uint32_t *	tailp;	/* pointer to tail */
311	struct hatm_softc *sc;	/* back pointer */
312	u_int		group;	/* interrupt group */
313};
314
315/*
316 * This structure describes all information for a VCC open on the card.
317 * The array of these structures is indexed by the compressed connection ID
318 * (CID). This structure must begin with the atmio_vcc.
319 */
320struct hevcc {
321	struct atmio_vcc param;		/* traffic parameters */
322	void *		rxhand;		/* NATM protocol block */
323	u_int		vflags;		/* private flags */
324	uint32_t	ipackets;
325	uint32_t	opackets;
326	uint32_t	ibytes;
327	uint32_t	obytes;
328
329	u_int		rc;		/* rate control group for CBR */
330	struct mbuf *	chain;		/* partial received PDU */
331	struct mbuf *	last;		/* last mbuf in chain */
332	u_int		ntpds;		/* number of active TPDs */
333};
334#define HE_VCC_OPEN		0x000f0000
335#define HE_VCC_RX_OPEN		0x00010000
336#define HE_VCC_RX_CLOSING	0x00020000
337#define HE_VCC_TX_OPEN		0x00040000
338#define HE_VCC_TX_CLOSING	0x00080000
339#define HE_VCC_FLOW_CTRL	0x00100000
340
341/*
342 * CBR rate groups
343 */
344struct herg {
345	u_int	refcnt;		/* how many connections reference this group */
346	u_int	rate;		/* the value */
347};
348
349/*
350 * Softc
351 */
352struct hatm_softc {
353	struct ifatm		ifatm;		/* common ATM stuff */
354	struct mtx		mtx;		/* lock */
355	struct ifmedia		media;		/* media */
356	device_t		dev;		/* device */
357	int			memid;		/* resoure id for memory */
358	struct resource *	memres;		/* memory resource */
359	bus_space_handle_t	memh;		/* handle */
360	bus_space_tag_t		memt;		/* ... and tag */
361	bus_dma_tag_t		parent_tag;	/* global restriction */
362	struct cv		vcc_cv;		/* condition variable */
363	int			irqid;		/* resource id */
364	struct resource *	irqres;		/* resource */
365	void *			ih;		/* interrupt handle */
366	struct utopia		utopia;		/* utopia state */
367
368	/* rest has to be reset by stop */
369	int			he622;		/* this is a HE622 */
370	int			pci64;		/* 64bit bus */
371	char			prod_id[HE_EEPROM_PROD_ID_LEN + 1];
372	char			rev[HE_EEPROM_REV_LEN + 1];
373	struct heirq		irq_0;		/* interrupt queues 0 */
374
375	/* generic network controller state */
376	u_int			cells_per_row;
377	u_int			bytes_per_row;
378	u_int			r0_numrows;
379	u_int			tx_numrows;
380	u_int			r1_numrows;
381	u_int			r0_startrow;
382	u_int			tx_startrow;
383	u_int			r1_startrow;
384	u_int			cells_per_lbuf;
385	u_int			r0_numbuffs;
386	u_int			r1_numbuffs;
387	u_int			tx_numbuffs;
388
389	/* HSP */
390	struct he_hsp		*hsp;
391	struct dmamem		hsp_mem;
392
393	/*** TX ***/
394	struct hetbrq		tbrq;		/* TBRQ 0 */
395	struct hetpdrq		tpdrq;		/* TPDRQ */
396	struct tpd_list		tpd_free;	/* Free TPDs */
397	u_int			tpd_nfree;	/* number of free TPDs */
398	u_int			tpd_total;	/* total TPDs */
399	uint8_t			*tpd_used;	/* bitmap of used TPDs */
400	struct dmamem		tpds;		/* TPD memory */
401	bus_dma_tag_t		tx_tag;		/* DMA tag for all tx mbufs */
402
403	/*** RX ***/
404	/* receive/transmit groups */
405	struct herbp		rbp_s0;		/* RBPS0 */
406	struct herbp		rbp_l0;		/* RBPL0 */
407	struct herbp		rbp_s1;		/* RBPS1 */
408	struct herbrq		rbrq_0;		/* RBRQ0 */
409	struct herbrq		rbrq_1;		/* RBRQ1 */
410
411	/* list of external mbuf storage */
412	bus_dma_tag_t		mbuf_tag;
413	struct mbuf_page	**mbuf_pages;
414	u_int			mbuf_npages;
415	struct mbufx_free	*mbuf_list[2];
416
417	/* mbuf cluster tracking and mapping for group 0 */
418	struct mbuf		**lbufs;	/* mbufs */
419	bus_dmamap_t		*rmaps;		/* DMA maps */
420	u_int			lbufs_size;
421	u_int			lbufs_next;
422
423	/* VCCs */
424	struct hevcc		*vccs[HE_MAX_VCCS];
425	u_int			cbr_bw;		/* BW allocated to CBR */
426	u_int			max_tpd;	/* per VCC */
427	u_int			open_vccs;
428	uma_zone_t		vcc_zone;
429
430	/* rate groups */
431	struct herg		rate_ctrl[HE_REGN_CS_STPER];
432
433	/* memory offsets */
434	u_int			tsrb, tsrc, tsrd;
435	u_int			rsrb;
436
437	struct cv		cv_rcclose;	/* condition variable */
438	uint32_t		rate_grid[16][16]; /* our copy */
439
440	/* sysctl support */
441	struct sysctl_ctx_list	sysctl_ctx;
442	struct sysctl_oid	*sysctl_tree;
443
444	/* internal statistics */
445	struct istats		istats;
446
447#ifdef HATM_DEBUG
448	/* debugging */
449	u_int			debug;
450#endif
451};
452
453#define READ4(SC,OFF)	bus_space_read_4(SC->memt, SC->memh, (OFF))
454#define READ2(SC,OFF)	bus_space_read_2(SC->memt, SC->memh, (OFF))
455#define READ1(SC,OFF)	bus_space_read_1(SC->memt, SC->memh, (OFF))
456
457#define WRITE4(SC,OFF,VAL) bus_space_write_4(SC->memt, SC->memh, (OFF), (VAL))
458#define WRITE2(SC,OFF,VAL) bus_space_write_2(SC->memt, SC->memh, (OFF), (VAL))
459#define WRITE1(SC,OFF,VAL) bus_space_write_1(SC->memt, SC->memh, (OFF), (VAL))
460
461#define BARRIER_R(SC) bus_space_barrier(SC->memt, SC->memh, 0, HE_REGO_END, \
462	BUS_SPACE_BARRIER_READ)
463#define BARRIER_W(SC) bus_space_barrier(SC->memt, SC->memh, 0, HE_REGO_END, \
464	BUS_SPACE_BARRIER_WRITE)
465#define BARRIER_RW(SC) bus_space_barrier(SC->memt, SC->memh, 0, HE_REGO_END, \
466	BUS_SPACE_BARRIER_READ | BUS_SPACE_BARRIER_WRITE)
467
468#define READ_SUNI(SC,OFF) READ4(SC, HE_REGO_SUNI + 4 * (OFF))
469#define WRITE_SUNI(SC,OFF,VAL) WRITE4(SC, HE_REGO_SUNI + 4 * (OFF), (VAL))
470
471#define READ_LB4(SC,OFF)						\
472    ({									\
473	WRITE4(SC, HE_REGO_LB_MEM_ADDR, (OFF));				\
474	WRITE4(SC, HE_REGO_LB_MEM_ACCESS,				\
475	    (HE_REGM_LB_MEM_HNDSHK | HE_REGM_LB_MEM_READ));		\
476	while((READ4(SC, HE_REGO_LB_MEM_ACCESS) & HE_REGM_LB_MEM_HNDSHK))\
477		;							\
478	READ4(SC, HE_REGO_LB_MEM_DATA);					\
479    })
480#define WRITE_LB4(SC,OFF,VAL)						\
481    do {								\
482	WRITE4(SC, HE_REGO_LB_MEM_ADDR, (OFF));				\
483	WRITE4(SC, HE_REGO_LB_MEM_DATA, (VAL));				\
484	WRITE4(SC, HE_REGO_LB_MEM_ACCESS,				\
485	    (HE_REGM_LB_MEM_HNDSHK | HE_REGM_LB_MEM_WRITE));		\
486	while((READ4(SC, HE_REGO_LB_MEM_ACCESS) & HE_REGM_LB_MEM_HNDSHK))\
487		;							\
488    } while(0)
489
490#define WRITE_MEM4(SC,OFF,VAL,SPACE)					\
491    do {								\
492	WRITE4(SC, HE_REGO_CON_DAT, (VAL));				\
493	WRITE4(SC, HE_REGO_CON_CTL,					\
494	    (SPACE | HE_REGM_CON_WE | HE_REGM_CON_STATUS | (OFF)));	\
495	while((READ4(SC, HE_REGO_CON_CTL) & HE_REGM_CON_STATUS) != 0)	\
496		;							\
497    } while(0)
498
499#define READ_MEM4(SC,OFF,SPACE)					\
500    ({									\
501	WRITE4(SC, HE_REGO_CON_CTL,					\
502	    (SPACE | HE_REGM_CON_STATUS | (OFF)));			\
503	while((READ4(SC, HE_REGO_CON_CTL) & HE_REGM_CON_STATUS) != 0)	\
504		;							\
505	READ4(SC, HE_REGO_CON_DAT);					\
506    })
507
508#define WRITE_TCM4(SC,OFF,VAL) WRITE_MEM4(SC,(OFF),(VAL),HE_REGM_CON_TCM)
509#define WRITE_RCM4(SC,OFF,VAL) WRITE_MEM4(SC,(OFF),(VAL),HE_REGM_CON_RCM)
510#define WRITE_MBOX4(SC,OFF,VAL) WRITE_MEM4(SC,(OFF),(VAL),HE_REGM_CON_MBOX)
511
512#define READ_TCM4(SC,OFF) READ_MEM4(SC,(OFF),HE_REGM_CON_TCM)
513#define READ_RCM4(SC,OFF) READ_MEM4(SC,(OFF),HE_REGM_CON_RCM)
514#define READ_MBOX4(SC,OFF) READ_MEM4(SC,(OFF),HE_REGM_CON_MBOX)
515
516#define WRITE_TCM(SC,OFF,BYTES,VAL) 					\
517	WRITE_MEM4(SC,(OFF) | ((~(BYTES) & 0xf) << HE_REGS_CON_DIS),	\
518	    (VAL), HE_REGM_CON_TCM)
519#define WRITE_RCM(SC,OFF,BYTES,VAL) 					\
520	WRITE_MEM4(SC,(OFF) | ((~(BYTES) & 0xf) << HE_REGS_CON_DIS),	\
521	    (VAL), HE_REGM_CON_RCM)
522
523#define READ_TSR(SC,CID,NR)						\
524    ({									\
525	uint32_t _v;							\
526	if((NR) <= 7) {							\
527		_v = READ_TCM4(SC, HE_REGO_TSRA(0,CID,NR));		\
528	} else if((NR) <= 11) {						\
529		_v = READ_TCM4(SC, HE_REGO_TSRB((SC)->tsrb,CID,(NR-8)));\
530	} else if((NR) <= 13) {						\
531		_v = READ_TCM4(SC, HE_REGO_TSRC((SC)->tsrc,CID,(NR-12)));\
532	} else {							\
533		_v = READ_TCM4(SC, HE_REGO_TSRD((SC)->tsrd,CID));	\
534	}								\
535	_v;								\
536    })
537
538#define WRITE_TSR(SC,CID,NR,BEN,VAL)					\
539    do {								\
540	if((NR) <= 7) {							\
541		WRITE_TCM(SC, HE_REGO_TSRA(0,CID,NR),BEN,VAL);		\
542	} else if((NR) <= 11) {						\
543		WRITE_TCM(SC, HE_REGO_TSRB((SC)->tsrb,CID,(NR-8)),BEN,VAL);\
544	} else if((NR) <= 13) {						\
545		WRITE_TCM(SC, HE_REGO_TSRC((SC)->tsrc,CID,(NR-12)),BEN,VAL);\
546	} else {							\
547		WRITE_TCM(SC, HE_REGO_TSRD((SC)->tsrd,CID),BEN,VAL);	\
548	}								\
549    } while(0)
550
551#define READ_RSR(SC,CID,NR)						\
552    ({									\
553	uint32_t _v;							\
554	if((NR) <= 7) {							\
555		_v = READ_RCM4(SC, HE_REGO_RSRA(0,CID,NR));		\
556	} else {							\
557		_v = READ_RCM4(SC, HE_REGO_RSRB((SC)->rsrb,CID,(NR-8)));\
558	}								\
559	_v;								\
560    })
561
562#define WRITE_RSR(SC,CID,NR,BEN,VAL)					\
563    do {								\
564	if((NR) <= 7) {							\
565		WRITE_RCM(SC, HE_REGO_RSRA(0,CID,NR),BEN,VAL);		\
566	} else {							\
567		WRITE_RCM(SC, HE_REGO_RSRB((SC)->rsrb,CID,(NR-8)),BEN,VAL);\
568	}								\
569    } while(0)
570
571#ifdef HATM_DEBUG
572#define DBG(SC, FL, PRINT) do {						\
573	if((SC)->debug & DBG_##FL) { 					\
574		if_printf(&(SC)->ifatm.ifnet, "%s: ", __func__);	\
575		printf PRINT;						\
576		printf("\n");						\
577	}								\
578    } while (0)
579
580enum {
581	DBG_RX		= 0x0001,
582	DBG_TX		= 0x0002,
583	DBG_VCC		= 0x0004,
584	DBG_IOCTL	= 0x0008,
585	DBG_ATTACH	= 0x0010,
586	DBG_INTR	= 0x0020,
587	DBG_DMA		= 0x0040,
588	DBG_DMAH	= 0x0080,
589
590	DBG_ALL		= 0x00ff
591};
592
593#else
594#define DBG(SC, FL, PRINT)
595#endif
596
597u_int hatm_cps2atmf(uint32_t);
598u_int hatm_atmf2cps(uint32_t);
599
600void hatm_intr(void *);
601int hatm_ioctl(struct ifnet *, u_long, caddr_t);
602void hatm_initialize(struct hatm_softc *);
603void hatm_stop(struct hatm_softc *sc);
604void hatm_start(struct ifnet *);
605
606void hatm_rx(struct hatm_softc *sc, u_int cid, u_int flags, struct mbuf *m,
607    u_int len);
608void hatm_tx_complete(struct hatm_softc *sc, struct tpd *tpd, uint32_t);
609
610int hatm_tx_vcc_can_open(struct hatm_softc *sc, u_int cid, struct hevcc *);
611void hatm_tx_vcc_open(struct hatm_softc *sc, u_int cid);
612void hatm_rx_vcc_open(struct hatm_softc *sc, u_int cid);
613void hatm_tx_vcc_close(struct hatm_softc *sc, u_int cid);
614void hatm_rx_vcc_close(struct hatm_softc *sc, u_int cid);
615void hatm_tx_vcc_closed(struct hatm_softc *sc, u_int cid);
616void hatm_vcc_closed(struct hatm_softc *sc, u_int cid);
617void hatm_load_vc(struct hatm_softc *sc, u_int cid, int reopen);
618