if_mwl.c revision 289165
1/*-
2 * Copyright (c) 2007-2009 Sam Leffler, Errno Consulting
3 * Copyright (c) 2007-2008 Marvell Semiconductor, Inc.
4 * All rights reserved.
5 *
6 * Redistribution and use in source and binary forms, with or without
7 * modification, are permitted provided that the following conditions
8 * are met:
9 * 1. Redistributions of source code must retain the above copyright
10 *    notice, this list of conditions and the following disclaimer,
11 *    without modification.
12 * 2. Redistributions in binary form must reproduce at minimum a disclaimer
13 *    similar to the "NO WARRANTY" disclaimer below ("Disclaimer") and any
14 *    redistribution must be conditioned upon including a substantially
15 *    similar Disclaimer requirement for further binary redistribution.
16 *
17 * NO WARRANTY
18 * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
19 * ``AS IS'' AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
20 * LIMITED TO, THE IMPLIED WARRANTIES OF NONINFRINGEMENT, MERCHANTIBILITY
21 * AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL
22 * THE COPYRIGHT HOLDERS OR CONTRIBUTORS BE LIABLE FOR SPECIAL, EXEMPLARY,
23 * OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
24 * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
25 * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER
26 * IN CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
27 * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF
28 * THE POSSIBILITY OF SUCH DAMAGES.
29 */
30
31#include <sys/cdefs.h>
32__FBSDID("$FreeBSD: head/sys/dev/mwl/if_mwl.c 289165 2015-10-12 04:55:20Z adrian $");
33
34/*
35 * Driver for the Marvell 88W8363 Wireless LAN controller.
36 */
37
38#include "opt_inet.h"
39#include "opt_mwl.h"
40#include "opt_wlan.h"
41
42#include <sys/param.h>
43#include <sys/systm.h>
44#include <sys/sysctl.h>
45#include <sys/mbuf.h>
46#include <sys/malloc.h>
47#include <sys/lock.h>
48#include <sys/mutex.h>
49#include <sys/kernel.h>
50#include <sys/socket.h>
51#include <sys/sockio.h>
52#include <sys/errno.h>
53#include <sys/callout.h>
54#include <sys/bus.h>
55#include <sys/endian.h>
56#include <sys/kthread.h>
57#include <sys/taskqueue.h>
58
59#include <machine/bus.h>
60
61#include <net/if.h>
62#include <net/if_var.h>
63#include <net/if_dl.h>
64#include <net/if_media.h>
65#include <net/if_types.h>
66#include <net/if_arp.h>
67#include <net/ethernet.h>
68#include <net/if_llc.h>
69
70#include <net/bpf.h>
71
72#include <net80211/ieee80211_var.h>
73#include <net80211/ieee80211_input.h>
74#include <net80211/ieee80211_regdomain.h>
75
76#ifdef INET
77#include <netinet/in.h>
78#include <netinet/if_ether.h>
79#endif /* INET */
80
81#include <dev/mwl/if_mwlvar.h>
82#include <dev/mwl/mwldiag.h>
83
84/* idiomatic shorthands: MS = mask+shift, SM = shift+mask */
85#define	MS(v,x)	(((v) & x) >> x##_S)
86#define	SM(v,x)	(((v) << x##_S) & x)
87
88static struct ieee80211vap *mwl_vap_create(struct ieee80211com *,
89		    const char [IFNAMSIZ], int, enum ieee80211_opmode, int,
90		    const uint8_t [IEEE80211_ADDR_LEN],
91		    const uint8_t [IEEE80211_ADDR_LEN]);
92static void	mwl_vap_delete(struct ieee80211vap *);
93static int	mwl_setupdma(struct mwl_softc *);
94static int	mwl_hal_reset(struct mwl_softc *sc);
95static int	mwl_init(struct mwl_softc *);
96static void	mwl_parent(struct ieee80211com *);
97static int	mwl_reset(struct ieee80211vap *, u_long);
98static void	mwl_stop(struct mwl_softc *);
99static void	mwl_start(struct mwl_softc *);
100static int	mwl_transmit(struct ieee80211com *, struct mbuf *);
101static int	mwl_raw_xmit(struct ieee80211_node *, struct mbuf *,
102			const struct ieee80211_bpf_params *);
103static int	mwl_media_change(struct ifnet *);
104static void	mwl_watchdog(void *);
105static int	mwl_ioctl(struct ieee80211com *, u_long, void *);
106static void	mwl_radar_proc(void *, int);
107static void	mwl_chanswitch_proc(void *, int);
108static void	mwl_bawatchdog_proc(void *, int);
109static int	mwl_key_alloc(struct ieee80211vap *,
110			struct ieee80211_key *,
111			ieee80211_keyix *, ieee80211_keyix *);
112static int	mwl_key_delete(struct ieee80211vap *,
113			const struct ieee80211_key *);
114static int	mwl_key_set(struct ieee80211vap *,
115			const struct ieee80211_key *);
116static int	_mwl_key_set(struct ieee80211vap *,
117			const struct ieee80211_key *,
118			const uint8_t mac[IEEE80211_ADDR_LEN]);
119static int	mwl_mode_init(struct mwl_softc *);
120static void	mwl_update_mcast(struct ieee80211com *);
121static void	mwl_update_promisc(struct ieee80211com *);
122static void	mwl_updateslot(struct ieee80211com *);
123static int	mwl_beacon_setup(struct ieee80211vap *);
124static void	mwl_beacon_update(struct ieee80211vap *, int);
125#ifdef MWL_HOST_PS_SUPPORT
126static void	mwl_update_ps(struct ieee80211vap *, int);
127static int	mwl_set_tim(struct ieee80211_node *, int);
128#endif
129static int	mwl_dma_setup(struct mwl_softc *);
130static void	mwl_dma_cleanup(struct mwl_softc *);
131static struct ieee80211_node *mwl_node_alloc(struct ieee80211vap *,
132		    const uint8_t [IEEE80211_ADDR_LEN]);
133static void	mwl_node_cleanup(struct ieee80211_node *);
134static void	mwl_node_drain(struct ieee80211_node *);
135static void	mwl_node_getsignal(const struct ieee80211_node *,
136			int8_t *, int8_t *);
137static void	mwl_node_getmimoinfo(const struct ieee80211_node *,
138			struct ieee80211_mimo_info *);
139static int	mwl_rxbuf_init(struct mwl_softc *, struct mwl_rxbuf *);
140static void	mwl_rx_proc(void *, int);
141static void	mwl_txq_init(struct mwl_softc *sc, struct mwl_txq *, int);
142static int	mwl_tx_setup(struct mwl_softc *, int, int);
143static int	mwl_wme_update(struct ieee80211com *);
144static void	mwl_tx_cleanupq(struct mwl_softc *, struct mwl_txq *);
145static void	mwl_tx_cleanup(struct mwl_softc *);
146static uint16_t	mwl_calcformat(uint8_t rate, const struct ieee80211_node *);
147static int	mwl_tx_start(struct mwl_softc *, struct ieee80211_node *,
148			     struct mwl_txbuf *, struct mbuf *);
149static void	mwl_tx_proc(void *, int);
150static int	mwl_chan_set(struct mwl_softc *, struct ieee80211_channel *);
151static void	mwl_draintxq(struct mwl_softc *);
152static void	mwl_cleartxq(struct mwl_softc *, struct ieee80211vap *);
153static int	mwl_recv_action(struct ieee80211_node *,
154			const struct ieee80211_frame *,
155			const uint8_t *, const uint8_t *);
156static int	mwl_addba_request(struct ieee80211_node *,
157			struct ieee80211_tx_ampdu *, int dialogtoken,
158			int baparamset, int batimeout);
159static int	mwl_addba_response(struct ieee80211_node *,
160			struct ieee80211_tx_ampdu *, int status,
161			int baparamset, int batimeout);
162static void	mwl_addba_stop(struct ieee80211_node *,
163			struct ieee80211_tx_ampdu *);
164static int	mwl_startrecv(struct mwl_softc *);
165static MWL_HAL_APMODE mwl_getapmode(const struct ieee80211vap *,
166			struct ieee80211_channel *);
167static int	mwl_setapmode(struct ieee80211vap *, struct ieee80211_channel*);
168static void	mwl_scan_start(struct ieee80211com *);
169static void	mwl_scan_end(struct ieee80211com *);
170static void	mwl_set_channel(struct ieee80211com *);
171static int	mwl_peerstadb(struct ieee80211_node *,
172			int aid, int staid, MWL_HAL_PEERINFO *pi);
173static int	mwl_localstadb(struct ieee80211vap *);
174static int	mwl_newstate(struct ieee80211vap *, enum ieee80211_state, int);
175static int	allocstaid(struct mwl_softc *sc, int aid);
176static void	delstaid(struct mwl_softc *sc, int staid);
177static void	mwl_newassoc(struct ieee80211_node *, int);
178static void	mwl_agestations(void *);
179static int	mwl_setregdomain(struct ieee80211com *,
180			struct ieee80211_regdomain *, int,
181			struct ieee80211_channel []);
182static void	mwl_getradiocaps(struct ieee80211com *, int, int *,
183			struct ieee80211_channel []);
184static int	mwl_getchannels(struct mwl_softc *);
185
186static void	mwl_sysctlattach(struct mwl_softc *);
187static void	mwl_announce(struct mwl_softc *);
188
189SYSCTL_NODE(_hw, OID_AUTO, mwl, CTLFLAG_RD, 0, "Marvell driver parameters");
190
191static	int mwl_rxdesc = MWL_RXDESC;		/* # rx desc's to allocate */
192SYSCTL_INT(_hw_mwl, OID_AUTO, rxdesc, CTLFLAG_RW, &mwl_rxdesc,
193	    0, "rx descriptors allocated");
194static	int mwl_rxbuf = MWL_RXBUF;		/* # rx buffers to allocate */
195SYSCTL_INT(_hw_mwl, OID_AUTO, rxbuf, CTLFLAG_RWTUN, &mwl_rxbuf,
196	    0, "rx buffers allocated");
197static	int mwl_txbuf = MWL_TXBUF;		/* # tx buffers to allocate */
198SYSCTL_INT(_hw_mwl, OID_AUTO, txbuf, CTLFLAG_RWTUN, &mwl_txbuf,
199	    0, "tx buffers allocated");
200static	int mwl_txcoalesce = 8;		/* # tx packets to q before poking f/w*/
201SYSCTL_INT(_hw_mwl, OID_AUTO, txcoalesce, CTLFLAG_RWTUN, &mwl_txcoalesce,
202	    0, "tx buffers to send at once");
203static	int mwl_rxquota = MWL_RXBUF;		/* # max buffers to process */
204SYSCTL_INT(_hw_mwl, OID_AUTO, rxquota, CTLFLAG_RWTUN, &mwl_rxquota,
205	    0, "max rx buffers to process per interrupt");
206static	int mwl_rxdmalow = 3;			/* # min buffers for wakeup */
207SYSCTL_INT(_hw_mwl, OID_AUTO, rxdmalow, CTLFLAG_RWTUN, &mwl_rxdmalow,
208	    0, "min free rx buffers before restarting traffic");
209
210#ifdef MWL_DEBUG
211static	int mwl_debug = 0;
212SYSCTL_INT(_hw_mwl, OID_AUTO, debug, CTLFLAG_RWTUN, &mwl_debug,
213	    0, "control debugging printfs");
214enum {
215	MWL_DEBUG_XMIT		= 0x00000001,	/* basic xmit operation */
216	MWL_DEBUG_XMIT_DESC	= 0x00000002,	/* xmit descriptors */
217	MWL_DEBUG_RECV		= 0x00000004,	/* basic recv operation */
218	MWL_DEBUG_RECV_DESC	= 0x00000008,	/* recv descriptors */
219	MWL_DEBUG_RESET		= 0x00000010,	/* reset processing */
220	MWL_DEBUG_BEACON 	= 0x00000020,	/* beacon handling */
221	MWL_DEBUG_INTR		= 0x00000040,	/* ISR */
222	MWL_DEBUG_TX_PROC	= 0x00000080,	/* tx ISR proc */
223	MWL_DEBUG_RX_PROC	= 0x00000100,	/* rx ISR proc */
224	MWL_DEBUG_KEYCACHE	= 0x00000200,	/* key cache management */
225	MWL_DEBUG_STATE		= 0x00000400,	/* 802.11 state transitions */
226	MWL_DEBUG_NODE		= 0x00000800,	/* node management */
227	MWL_DEBUG_RECV_ALL	= 0x00001000,	/* trace all frames (beacons) */
228	MWL_DEBUG_TSO		= 0x00002000,	/* TSO processing */
229	MWL_DEBUG_AMPDU		= 0x00004000,	/* BA stream handling */
230	MWL_DEBUG_ANY		= 0xffffffff
231};
232#define	IS_BEACON(wh) \
233    ((wh->i_fc[0] & (IEEE80211_FC0_TYPE_MASK|IEEE80211_FC0_SUBTYPE_MASK)) == \
234	 (IEEE80211_FC0_TYPE_MGT|IEEE80211_FC0_SUBTYPE_BEACON))
235#define	IFF_DUMPPKTS_RECV(sc, wh) \
236    ((sc->sc_debug & MWL_DEBUG_RECV) && \
237      ((sc->sc_debug & MWL_DEBUG_RECV_ALL) || !IS_BEACON(wh)))
238#define	IFF_DUMPPKTS_XMIT(sc) \
239	(sc->sc_debug & MWL_DEBUG_XMIT)
240
241#define	DPRINTF(sc, m, fmt, ...) do {				\
242	if (sc->sc_debug & (m))					\
243		printf(fmt, __VA_ARGS__);			\
244} while (0)
245#define	KEYPRINTF(sc, hk, mac) do {				\
246	if (sc->sc_debug & MWL_DEBUG_KEYCACHE)			\
247		mwl_keyprint(sc, __func__, hk, mac);		\
248} while (0)
249static	void mwl_printrxbuf(const struct mwl_rxbuf *bf, u_int ix);
250static	void mwl_printtxbuf(const struct mwl_txbuf *bf, u_int qnum, u_int ix);
251#else
252#define	IFF_DUMPPKTS_RECV(sc, wh)	0
253#define	IFF_DUMPPKTS_XMIT(sc)		0
254#define	DPRINTF(sc, m, fmt, ...)	do { (void )sc; } while (0)
255#define	KEYPRINTF(sc, k, mac)		do { (void )sc; } while (0)
256#endif
257
258static MALLOC_DEFINE(M_MWLDEV, "mwldev", "mwl driver dma buffers");
259
260/*
261 * Each packet has fixed front matter: a 2-byte length
262 * of the payload, followed by a 4-address 802.11 header
263 * (regardless of the actual header and always w/o any
264 * QoS header).  The payload then follows.
265 */
266struct mwltxrec {
267	uint16_t fwlen;
268	struct ieee80211_frame_addr4 wh;
269} __packed;
270
271/*
272 * Read/Write shorthands for accesses to BAR 0.  Note
273 * that all BAR 1 operations are done in the "hal" and
274 * there should be no reference to them here.
275 */
276#ifdef MWL_DEBUG
277static __inline uint32_t
278RD4(struct mwl_softc *sc, bus_size_t off)
279{
280	return bus_space_read_4(sc->sc_io0t, sc->sc_io0h, off);
281}
282#endif
283
284static __inline void
285WR4(struct mwl_softc *sc, bus_size_t off, uint32_t val)
286{
287	bus_space_write_4(sc->sc_io0t, sc->sc_io0h, off, val);
288}
289
290int
291mwl_attach(uint16_t devid, struct mwl_softc *sc)
292{
293	struct ieee80211com *ic = &sc->sc_ic;
294	struct mwl_hal *mh;
295	int error = 0;
296
297	DPRINTF(sc, MWL_DEBUG_ANY, "%s: devid 0x%x\n", __func__, devid);
298
299	/*
300	 * Setup the RX free list lock early, so it can be consistently
301	 * removed.
302	 */
303	MWL_RXFREE_INIT(sc);
304
305	mh = mwl_hal_attach(sc->sc_dev, devid,
306	    sc->sc_io1h, sc->sc_io1t, sc->sc_dmat);
307	if (mh == NULL) {
308		device_printf(sc->sc_dev, "unable to attach HAL\n");
309		error = EIO;
310		goto bad;
311	}
312	sc->sc_mh = mh;
313	/*
314	 * Load firmware so we can get setup.  We arbitrarily
315	 * pick station firmware; we'll re-load firmware as
316	 * needed so setting up the wrong mode isn't a big deal.
317	 */
318	if (mwl_hal_fwload(mh, NULL) != 0) {
319		device_printf(sc->sc_dev, "unable to setup builtin firmware\n");
320		error = EIO;
321		goto bad1;
322	}
323	if (mwl_hal_gethwspecs(mh, &sc->sc_hwspecs) != 0) {
324		device_printf(sc->sc_dev, "unable to fetch h/w specs\n");
325		error = EIO;
326		goto bad1;
327	}
328	error = mwl_getchannels(sc);
329	if (error != 0)
330		goto bad1;
331
332	sc->sc_txantenna = 0;		/* h/w default */
333	sc->sc_rxantenna = 0;		/* h/w default */
334	sc->sc_invalid = 0;		/* ready to go, enable int handling */
335	sc->sc_ageinterval = MWL_AGEINTERVAL;
336
337	/*
338	 * Allocate tx+rx descriptors and populate the lists.
339	 * We immediately push the information to the firmware
340	 * as otherwise it gets upset.
341	 */
342	error = mwl_dma_setup(sc);
343	if (error != 0) {
344		device_printf(sc->sc_dev, "failed to setup descriptors: %d\n",
345		    error);
346		goto bad1;
347	}
348	error = mwl_setupdma(sc);	/* push to firmware */
349	if (error != 0)			/* NB: mwl_setupdma prints msg */
350		goto bad1;
351
352	callout_init(&sc->sc_timer, 1);
353	callout_init_mtx(&sc->sc_watchdog, &sc->sc_mtx, 0);
354	mbufq_init(&sc->sc_snd, ifqmaxlen);
355
356	sc->sc_tq = taskqueue_create("mwl_taskq", M_NOWAIT,
357		taskqueue_thread_enqueue, &sc->sc_tq);
358	taskqueue_start_threads(&sc->sc_tq, 1, PI_NET,
359		"%s taskq", device_get_nameunit(sc->sc_dev));
360
361	TASK_INIT(&sc->sc_rxtask, 0, mwl_rx_proc, sc);
362	TASK_INIT(&sc->sc_radartask, 0, mwl_radar_proc, sc);
363	TASK_INIT(&sc->sc_chanswitchtask, 0, mwl_chanswitch_proc, sc);
364	TASK_INIT(&sc->sc_bawatchdogtask, 0, mwl_bawatchdog_proc, sc);
365
366	/* NB: insure BK queue is the lowest priority h/w queue */
367	if (!mwl_tx_setup(sc, WME_AC_BK, MWL_WME_AC_BK)) {
368		device_printf(sc->sc_dev,
369		    "unable to setup xmit queue for %s traffic!\n",
370		     ieee80211_wme_acnames[WME_AC_BK]);
371		error = EIO;
372		goto bad2;
373	}
374	if (!mwl_tx_setup(sc, WME_AC_BE, MWL_WME_AC_BE) ||
375	    !mwl_tx_setup(sc, WME_AC_VI, MWL_WME_AC_VI) ||
376	    !mwl_tx_setup(sc, WME_AC_VO, MWL_WME_AC_VO)) {
377		/*
378		 * Not enough hardware tx queues to properly do WME;
379		 * just punt and assign them all to the same h/w queue.
380		 * We could do a better job of this if, for example,
381		 * we allocate queues when we switch from station to
382		 * AP mode.
383		 */
384		if (sc->sc_ac2q[WME_AC_VI] != NULL)
385			mwl_tx_cleanupq(sc, sc->sc_ac2q[WME_AC_VI]);
386		if (sc->sc_ac2q[WME_AC_BE] != NULL)
387			mwl_tx_cleanupq(sc, sc->sc_ac2q[WME_AC_BE]);
388		sc->sc_ac2q[WME_AC_BE] = sc->sc_ac2q[WME_AC_BK];
389		sc->sc_ac2q[WME_AC_VI] = sc->sc_ac2q[WME_AC_BK];
390		sc->sc_ac2q[WME_AC_VO] = sc->sc_ac2q[WME_AC_BK];
391	}
392	TASK_INIT(&sc->sc_txtask, 0, mwl_tx_proc, sc);
393
394	ic->ic_softc = sc;
395	ic->ic_name = device_get_nameunit(sc->sc_dev);
396	/* XXX not right but it's not used anywhere important */
397	ic->ic_phytype = IEEE80211_T_OFDM;
398	ic->ic_opmode = IEEE80211_M_STA;
399	ic->ic_caps =
400		  IEEE80211_C_STA		/* station mode supported */
401		| IEEE80211_C_HOSTAP		/* hostap mode */
402		| IEEE80211_C_MONITOR		/* monitor mode */
403#if 0
404		| IEEE80211_C_IBSS		/* ibss, nee adhoc, mode */
405		| IEEE80211_C_AHDEMO		/* adhoc demo mode */
406#endif
407		| IEEE80211_C_MBSS		/* mesh point link mode */
408		| IEEE80211_C_WDS		/* WDS supported */
409		| IEEE80211_C_SHPREAMBLE	/* short preamble supported */
410		| IEEE80211_C_SHSLOT		/* short slot time supported */
411		| IEEE80211_C_WME		/* WME/WMM supported */
412		| IEEE80211_C_BURST		/* xmit bursting supported */
413		| IEEE80211_C_WPA		/* capable of WPA1+WPA2 */
414		| IEEE80211_C_BGSCAN		/* capable of bg scanning */
415		| IEEE80211_C_TXFRAG		/* handle tx frags */
416		| IEEE80211_C_TXPMGT		/* capable of txpow mgt */
417		| IEEE80211_C_DFS		/* DFS supported */
418		;
419
420	ic->ic_htcaps =
421		  IEEE80211_HTCAP_SMPS_ENA	/* SM PS mode enabled */
422		| IEEE80211_HTCAP_CHWIDTH40	/* 40MHz channel width */
423		| IEEE80211_HTCAP_SHORTGI20	/* short GI in 20MHz */
424		| IEEE80211_HTCAP_SHORTGI40	/* short GI in 40MHz */
425		| IEEE80211_HTCAP_RXSTBC_2STREAM/* 1-2 spatial streams */
426#if MWL_AGGR_SIZE == 7935
427		| IEEE80211_HTCAP_MAXAMSDU_7935	/* max A-MSDU length */
428#else
429		| IEEE80211_HTCAP_MAXAMSDU_3839	/* max A-MSDU length */
430#endif
431#if 0
432		| IEEE80211_HTCAP_PSMP		/* PSMP supported */
433		| IEEE80211_HTCAP_40INTOLERANT	/* 40MHz intolerant */
434#endif
435		/* s/w capabilities */
436		| IEEE80211_HTC_HT		/* HT operation */
437		| IEEE80211_HTC_AMPDU		/* tx A-MPDU */
438		| IEEE80211_HTC_AMSDU		/* tx A-MSDU */
439		| IEEE80211_HTC_SMPS		/* SMPS available */
440		;
441
442	/*
443	 * Mark h/w crypto support.
444	 * XXX no way to query h/w support.
445	 */
446	ic->ic_cryptocaps |= IEEE80211_CRYPTO_WEP
447			  |  IEEE80211_CRYPTO_AES_CCM
448			  |  IEEE80211_CRYPTO_TKIP
449			  |  IEEE80211_CRYPTO_TKIPMIC
450			  ;
451	/*
452	 * Transmit requires space in the packet for a special
453	 * format transmit record and optional padding between
454	 * this record and the payload.  Ask the net80211 layer
455	 * to arrange this when encapsulating packets so we can
456	 * add it efficiently.
457	 */
458	ic->ic_headroom = sizeof(struct mwltxrec) -
459		sizeof(struct ieee80211_frame);
460
461	IEEE80211_ADDR_COPY(ic->ic_macaddr, sc->sc_hwspecs.macAddr);
462
463	/* call MI attach routine. */
464	ieee80211_ifattach(ic);
465	ic->ic_setregdomain = mwl_setregdomain;
466	ic->ic_getradiocaps = mwl_getradiocaps;
467	/* override default methods */
468	ic->ic_raw_xmit = mwl_raw_xmit;
469	ic->ic_newassoc = mwl_newassoc;
470	ic->ic_updateslot = mwl_updateslot;
471	ic->ic_update_mcast = mwl_update_mcast;
472	ic->ic_update_promisc = mwl_update_promisc;
473	ic->ic_wme.wme_update = mwl_wme_update;
474	ic->ic_transmit = mwl_transmit;
475	ic->ic_ioctl = mwl_ioctl;
476	ic->ic_parent = mwl_parent;
477
478	ic->ic_node_alloc = mwl_node_alloc;
479	sc->sc_node_cleanup = ic->ic_node_cleanup;
480	ic->ic_node_cleanup = mwl_node_cleanup;
481	sc->sc_node_drain = ic->ic_node_drain;
482	ic->ic_node_drain = mwl_node_drain;
483	ic->ic_node_getsignal = mwl_node_getsignal;
484	ic->ic_node_getmimoinfo = mwl_node_getmimoinfo;
485
486	ic->ic_scan_start = mwl_scan_start;
487	ic->ic_scan_end = mwl_scan_end;
488	ic->ic_set_channel = mwl_set_channel;
489
490	sc->sc_recv_action = ic->ic_recv_action;
491	ic->ic_recv_action = mwl_recv_action;
492	sc->sc_addba_request = ic->ic_addba_request;
493	ic->ic_addba_request = mwl_addba_request;
494	sc->sc_addba_response = ic->ic_addba_response;
495	ic->ic_addba_response = mwl_addba_response;
496	sc->sc_addba_stop = ic->ic_addba_stop;
497	ic->ic_addba_stop = mwl_addba_stop;
498
499	ic->ic_vap_create = mwl_vap_create;
500	ic->ic_vap_delete = mwl_vap_delete;
501
502	ieee80211_radiotap_attach(ic,
503	    &sc->sc_tx_th.wt_ihdr, sizeof(sc->sc_tx_th),
504		MWL_TX_RADIOTAP_PRESENT,
505	    &sc->sc_rx_th.wr_ihdr, sizeof(sc->sc_rx_th),
506		MWL_RX_RADIOTAP_PRESENT);
507	/*
508	 * Setup dynamic sysctl's now that country code and
509	 * regdomain are available from the hal.
510	 */
511	mwl_sysctlattach(sc);
512
513	if (bootverbose)
514		ieee80211_announce(ic);
515	mwl_announce(sc);
516	return 0;
517bad2:
518	mwl_dma_cleanup(sc);
519bad1:
520	mwl_hal_detach(mh);
521bad:
522	MWL_RXFREE_DESTROY(sc);
523	sc->sc_invalid = 1;
524	return error;
525}
526
527int
528mwl_detach(struct mwl_softc *sc)
529{
530	struct ieee80211com *ic = &sc->sc_ic;
531
532	MWL_LOCK(sc);
533	mwl_stop(sc);
534	MWL_UNLOCK(sc);
535	/*
536	 * NB: the order of these is important:
537	 * o call the 802.11 layer before detaching the hal to
538	 *   insure callbacks into the driver to delete global
539	 *   key cache entries can be handled
540	 * o reclaim the tx queue data structures after calling
541	 *   the 802.11 layer as we'll get called back to reclaim
542	 *   node state and potentially want to use them
543	 * o to cleanup the tx queues the hal is called, so detach
544	 *   it last
545	 * Other than that, it's straightforward...
546	 */
547	ieee80211_ifdetach(ic);
548	callout_drain(&sc->sc_watchdog);
549	mwl_dma_cleanup(sc);
550	MWL_RXFREE_DESTROY(sc);
551	mwl_tx_cleanup(sc);
552	mwl_hal_detach(sc->sc_mh);
553	mbufq_drain(&sc->sc_snd);
554
555	return 0;
556}
557
558/*
559 * MAC address handling for multiple BSS on the same radio.
560 * The first vap uses the MAC address from the EEPROM.  For
561 * subsequent vap's we set the U/L bit (bit 1) in the MAC
562 * address and use the next six bits as an index.
563 */
564static void
565assign_address(struct mwl_softc *sc, uint8_t mac[IEEE80211_ADDR_LEN], int clone)
566{
567	int i;
568
569	if (clone && mwl_hal_ismbsscapable(sc->sc_mh)) {
570		/* NB: we only do this if h/w supports multiple bssid */
571		for (i = 0; i < 32; i++)
572			if ((sc->sc_bssidmask & (1<<i)) == 0)
573				break;
574		if (i != 0)
575			mac[0] |= (i << 2)|0x2;
576	} else
577		i = 0;
578	sc->sc_bssidmask |= 1<<i;
579	if (i == 0)
580		sc->sc_nbssid0++;
581}
582
583static void
584reclaim_address(struct mwl_softc *sc, const uint8_t mac[IEEE80211_ADDR_LEN])
585{
586	int i = mac[0] >> 2;
587	if (i != 0 || --sc->sc_nbssid0 == 0)
588		sc->sc_bssidmask &= ~(1<<i);
589}
590
591static struct ieee80211vap *
592mwl_vap_create(struct ieee80211com *ic, const char name[IFNAMSIZ], int unit,
593    enum ieee80211_opmode opmode, int flags,
594    const uint8_t bssid[IEEE80211_ADDR_LEN],
595    const uint8_t mac0[IEEE80211_ADDR_LEN])
596{
597	struct mwl_softc *sc = ic->ic_softc;
598	struct mwl_hal *mh = sc->sc_mh;
599	struct ieee80211vap *vap, *apvap;
600	struct mwl_hal_vap *hvap;
601	struct mwl_vap *mvp;
602	uint8_t mac[IEEE80211_ADDR_LEN];
603
604	IEEE80211_ADDR_COPY(mac, mac0);
605	switch (opmode) {
606	case IEEE80211_M_HOSTAP:
607	case IEEE80211_M_MBSS:
608		if ((flags & IEEE80211_CLONE_MACADDR) == 0)
609			assign_address(sc, mac, flags & IEEE80211_CLONE_BSSID);
610		hvap = mwl_hal_newvap(mh, MWL_HAL_AP, mac);
611		if (hvap == NULL) {
612			if ((flags & IEEE80211_CLONE_MACADDR) == 0)
613				reclaim_address(sc, mac);
614			return NULL;
615		}
616		break;
617	case IEEE80211_M_STA:
618		if ((flags & IEEE80211_CLONE_MACADDR) == 0)
619			assign_address(sc, mac, flags & IEEE80211_CLONE_BSSID);
620		hvap = mwl_hal_newvap(mh, MWL_HAL_STA, mac);
621		if (hvap == NULL) {
622			if ((flags & IEEE80211_CLONE_MACADDR) == 0)
623				reclaim_address(sc, mac);
624			return NULL;
625		}
626		/* no h/w beacon miss support; always use s/w */
627		flags |= IEEE80211_CLONE_NOBEACONS;
628		break;
629	case IEEE80211_M_WDS:
630		hvap = NULL;		/* NB: we use associated AP vap */
631		if (sc->sc_napvaps == 0)
632			return NULL;	/* no existing AP vap */
633		break;
634	case IEEE80211_M_MONITOR:
635		hvap = NULL;
636		break;
637	case IEEE80211_M_IBSS:
638	case IEEE80211_M_AHDEMO:
639	default:
640		return NULL;
641	}
642
643	mvp = malloc(sizeof(struct mwl_vap), M_80211_VAP, M_WAITOK | M_ZERO);
644	mvp->mv_hvap = hvap;
645	if (opmode == IEEE80211_M_WDS) {
646		/*
647		 * WDS vaps must have an associated AP vap; find one.
648		 * XXX not right.
649		 */
650		TAILQ_FOREACH(apvap, &ic->ic_vaps, iv_next)
651			if (apvap->iv_opmode == IEEE80211_M_HOSTAP) {
652				mvp->mv_ap_hvap = MWL_VAP(apvap)->mv_hvap;
653				break;
654			}
655		KASSERT(mvp->mv_ap_hvap != NULL, ("no ap vap"));
656	}
657	vap = &mvp->mv_vap;
658	ieee80211_vap_setup(ic, vap, name, unit, opmode, flags, bssid);
659	/* override with driver methods */
660	mvp->mv_newstate = vap->iv_newstate;
661	vap->iv_newstate = mwl_newstate;
662	vap->iv_max_keyix = 0;	/* XXX */
663	vap->iv_key_alloc = mwl_key_alloc;
664	vap->iv_key_delete = mwl_key_delete;
665	vap->iv_key_set = mwl_key_set;
666#ifdef MWL_HOST_PS_SUPPORT
667	if (opmode == IEEE80211_M_HOSTAP || opmode == IEEE80211_M_MBSS) {
668		vap->iv_update_ps = mwl_update_ps;
669		mvp->mv_set_tim = vap->iv_set_tim;
670		vap->iv_set_tim = mwl_set_tim;
671	}
672#endif
673	vap->iv_reset = mwl_reset;
674	vap->iv_update_beacon = mwl_beacon_update;
675
676	/* override max aid so sta's cannot assoc when we're out of sta id's */
677	vap->iv_max_aid = MWL_MAXSTAID;
678	/* override default A-MPDU rx parameters */
679	vap->iv_ampdu_rxmax = IEEE80211_HTCAP_MAXRXAMPDU_64K;
680	vap->iv_ampdu_density = IEEE80211_HTCAP_MPDUDENSITY_4;
681
682	/* complete setup */
683	ieee80211_vap_attach(vap, mwl_media_change, ieee80211_media_status,
684	    mac);
685
686	switch (vap->iv_opmode) {
687	case IEEE80211_M_HOSTAP:
688	case IEEE80211_M_MBSS:
689	case IEEE80211_M_STA:
690		/*
691		 * Setup sta db entry for local address.
692		 */
693		mwl_localstadb(vap);
694		if (vap->iv_opmode == IEEE80211_M_HOSTAP ||
695		    vap->iv_opmode == IEEE80211_M_MBSS)
696			sc->sc_napvaps++;
697		else
698			sc->sc_nstavaps++;
699		break;
700	case IEEE80211_M_WDS:
701		sc->sc_nwdsvaps++;
702		break;
703	default:
704		break;
705	}
706	/*
707	 * Setup overall operating mode.
708	 */
709	if (sc->sc_napvaps)
710		ic->ic_opmode = IEEE80211_M_HOSTAP;
711	else if (sc->sc_nstavaps)
712		ic->ic_opmode = IEEE80211_M_STA;
713	else
714		ic->ic_opmode = opmode;
715
716	return vap;
717}
718
719static void
720mwl_vap_delete(struct ieee80211vap *vap)
721{
722	struct mwl_vap *mvp = MWL_VAP(vap);
723	struct mwl_softc *sc = vap->iv_ic->ic_softc;
724	struct mwl_hal *mh = sc->sc_mh;
725	struct mwl_hal_vap *hvap = mvp->mv_hvap;
726	enum ieee80211_opmode opmode = vap->iv_opmode;
727
728	/* XXX disallow ap vap delete if WDS still present */
729	if (sc->sc_running) {
730		/* quiesce h/w while we remove the vap */
731		mwl_hal_intrset(mh, 0);		/* disable interrupts */
732	}
733	ieee80211_vap_detach(vap);
734	switch (opmode) {
735	case IEEE80211_M_HOSTAP:
736	case IEEE80211_M_MBSS:
737	case IEEE80211_M_STA:
738		KASSERT(hvap != NULL, ("no hal vap handle"));
739		(void) mwl_hal_delstation(hvap, vap->iv_myaddr);
740		mwl_hal_delvap(hvap);
741		if (opmode == IEEE80211_M_HOSTAP || opmode == IEEE80211_M_MBSS)
742			sc->sc_napvaps--;
743		else
744			sc->sc_nstavaps--;
745		/* XXX don't do it for IEEE80211_CLONE_MACADDR */
746		reclaim_address(sc, vap->iv_myaddr);
747		break;
748	case IEEE80211_M_WDS:
749		sc->sc_nwdsvaps--;
750		break;
751	default:
752		break;
753	}
754	mwl_cleartxq(sc, vap);
755	free(mvp, M_80211_VAP);
756	if (sc->sc_running)
757		mwl_hal_intrset(mh, sc->sc_imask);
758}
759
760void
761mwl_suspend(struct mwl_softc *sc)
762{
763
764	MWL_LOCK(sc);
765	mwl_stop(sc);
766	MWL_UNLOCK(sc);
767}
768
769void
770mwl_resume(struct mwl_softc *sc)
771{
772	int error = EDOOFUS;
773
774	MWL_LOCK(sc);
775	if (sc->sc_ic.ic_nrunning > 0)
776		error = mwl_init(sc);
777	MWL_UNLOCK(sc);
778
779	if (error == 0)
780		ieee80211_start_all(&sc->sc_ic);	/* start all vap's */
781}
782
783void
784mwl_shutdown(void *arg)
785{
786	struct mwl_softc *sc = arg;
787
788	MWL_LOCK(sc);
789	mwl_stop(sc);
790	MWL_UNLOCK(sc);
791}
792
793/*
794 * Interrupt handler.  Most of the actual processing is deferred.
795 */
796void
797mwl_intr(void *arg)
798{
799	struct mwl_softc *sc = arg;
800	struct mwl_hal *mh = sc->sc_mh;
801	uint32_t status;
802
803	if (sc->sc_invalid) {
804		/*
805		 * The hardware is not ready/present, don't touch anything.
806		 * Note this can happen early on if the IRQ is shared.
807		 */
808		DPRINTF(sc, MWL_DEBUG_ANY, "%s: invalid; ignored\n", __func__);
809		return;
810	}
811	/*
812	 * Figure out the reason(s) for the interrupt.
813	 */
814	mwl_hal_getisr(mh, &status);		/* NB: clears ISR too */
815	if (status == 0)			/* must be a shared irq */
816		return;
817
818	DPRINTF(sc, MWL_DEBUG_INTR, "%s: status 0x%x imask 0x%x\n",
819	    __func__, status, sc->sc_imask);
820	if (status & MACREG_A2HRIC_BIT_RX_RDY)
821		taskqueue_enqueue(sc->sc_tq, &sc->sc_rxtask);
822	if (status & MACREG_A2HRIC_BIT_TX_DONE)
823		taskqueue_enqueue(sc->sc_tq, &sc->sc_txtask);
824	if (status & MACREG_A2HRIC_BIT_BA_WATCHDOG)
825		taskqueue_enqueue(sc->sc_tq, &sc->sc_bawatchdogtask);
826	if (status & MACREG_A2HRIC_BIT_OPC_DONE)
827		mwl_hal_cmddone(mh);
828	if (status & MACREG_A2HRIC_BIT_MAC_EVENT) {
829		;
830	}
831	if (status & MACREG_A2HRIC_BIT_ICV_ERROR) {
832		/* TKIP ICV error */
833		sc->sc_stats.mst_rx_badtkipicv++;
834	}
835	if (status & MACREG_A2HRIC_BIT_QUEUE_EMPTY) {
836		/* 11n aggregation queue is empty, re-fill */
837		;
838	}
839	if (status & MACREG_A2HRIC_BIT_QUEUE_FULL) {
840		;
841	}
842	if (status & MACREG_A2HRIC_BIT_RADAR_DETECT) {
843		/* radar detected, process event */
844		taskqueue_enqueue(sc->sc_tq, &sc->sc_radartask);
845	}
846	if (status & MACREG_A2HRIC_BIT_CHAN_SWITCH) {
847		/* DFS channel switch */
848		taskqueue_enqueue(sc->sc_tq, &sc->sc_chanswitchtask);
849	}
850}
851
852static void
853mwl_radar_proc(void *arg, int pending)
854{
855	struct mwl_softc *sc = arg;
856	struct ieee80211com *ic = &sc->sc_ic;
857
858	DPRINTF(sc, MWL_DEBUG_ANY, "%s: radar detected, pending %u\n",
859	    __func__, pending);
860
861	sc->sc_stats.mst_radardetect++;
862	/* XXX stop h/w BA streams? */
863
864	IEEE80211_LOCK(ic);
865	ieee80211_dfs_notify_radar(ic, ic->ic_curchan);
866	IEEE80211_UNLOCK(ic);
867}
868
869static void
870mwl_chanswitch_proc(void *arg, int pending)
871{
872	struct mwl_softc *sc = arg;
873	struct ieee80211com *ic = &sc->sc_ic;
874
875	DPRINTF(sc, MWL_DEBUG_ANY, "%s: channel switch notice, pending %u\n",
876	    __func__, pending);
877
878	IEEE80211_LOCK(ic);
879	sc->sc_csapending = 0;
880	ieee80211_csa_completeswitch(ic);
881	IEEE80211_UNLOCK(ic);
882}
883
884static void
885mwl_bawatchdog(const MWL_HAL_BASTREAM *sp)
886{
887	struct ieee80211_node *ni = sp->data[0];
888
889	/* send DELBA and drop the stream */
890	ieee80211_ampdu_stop(ni, sp->data[1], IEEE80211_REASON_UNSPECIFIED);
891}
892
893static void
894mwl_bawatchdog_proc(void *arg, int pending)
895{
896	struct mwl_softc *sc = arg;
897	struct mwl_hal *mh = sc->sc_mh;
898	const MWL_HAL_BASTREAM *sp;
899	uint8_t bitmap, n;
900
901	sc->sc_stats.mst_bawatchdog++;
902
903	if (mwl_hal_getwatchdogbitmap(mh, &bitmap) != 0) {
904		DPRINTF(sc, MWL_DEBUG_AMPDU,
905		    "%s: could not get bitmap\n", __func__);
906		sc->sc_stats.mst_bawatchdog_failed++;
907		return;
908	}
909	DPRINTF(sc, MWL_DEBUG_AMPDU, "%s: bitmap 0x%x\n", __func__, bitmap);
910	if (bitmap == 0xff) {
911		n = 0;
912		/* disable all ba streams */
913		for (bitmap = 0; bitmap < 8; bitmap++) {
914			sp = mwl_hal_bastream_lookup(mh, bitmap);
915			if (sp != NULL) {
916				mwl_bawatchdog(sp);
917				n++;
918			}
919		}
920		if (n == 0) {
921			DPRINTF(sc, MWL_DEBUG_AMPDU,
922			    "%s: no BA streams found\n", __func__);
923			sc->sc_stats.mst_bawatchdog_empty++;
924		}
925	} else if (bitmap != 0xaa) {
926		/* disable a single ba stream */
927		sp = mwl_hal_bastream_lookup(mh, bitmap);
928		if (sp != NULL) {
929			mwl_bawatchdog(sp);
930		} else {
931			DPRINTF(sc, MWL_DEBUG_AMPDU,
932			    "%s: no BA stream %d\n", __func__, bitmap);
933			sc->sc_stats.mst_bawatchdog_notfound++;
934		}
935	}
936}
937
938/*
939 * Convert net80211 channel to a HAL channel.
940 */
941static void
942mwl_mapchan(MWL_HAL_CHANNEL *hc, const struct ieee80211_channel *chan)
943{
944	hc->channel = chan->ic_ieee;
945
946	*(uint32_t *)&hc->channelFlags = 0;
947	if (IEEE80211_IS_CHAN_2GHZ(chan))
948		hc->channelFlags.FreqBand = MWL_FREQ_BAND_2DOT4GHZ;
949	else if (IEEE80211_IS_CHAN_5GHZ(chan))
950		hc->channelFlags.FreqBand = MWL_FREQ_BAND_5GHZ;
951	if (IEEE80211_IS_CHAN_HT40(chan)) {
952		hc->channelFlags.ChnlWidth = MWL_CH_40_MHz_WIDTH;
953		if (IEEE80211_IS_CHAN_HT40U(chan))
954			hc->channelFlags.ExtChnlOffset = MWL_EXT_CH_ABOVE_CTRL_CH;
955		else
956			hc->channelFlags.ExtChnlOffset = MWL_EXT_CH_BELOW_CTRL_CH;
957	} else
958		hc->channelFlags.ChnlWidth = MWL_CH_20_MHz_WIDTH;
959	/* XXX 10MHz channels */
960}
961
962/*
963 * Inform firmware of our tx/rx dma setup.  The BAR 0
964 * writes below are for compatibility with older firmware.
965 * For current firmware we send this information with a
966 * cmd block via mwl_hal_sethwdma.
967 */
968static int
969mwl_setupdma(struct mwl_softc *sc)
970{
971	int error, i;
972
973	sc->sc_hwdma.rxDescRead = sc->sc_rxdma.dd_desc_paddr;
974	WR4(sc, sc->sc_hwspecs.rxDescRead, sc->sc_hwdma.rxDescRead);
975	WR4(sc, sc->sc_hwspecs.rxDescWrite, sc->sc_hwdma.rxDescRead);
976
977	for (i = 0; i < MWL_NUM_TX_QUEUES-MWL_NUM_ACK_QUEUES; i++) {
978		struct mwl_txq *txq = &sc->sc_txq[i];
979		sc->sc_hwdma.wcbBase[i] = txq->dma.dd_desc_paddr;
980		WR4(sc, sc->sc_hwspecs.wcbBase[i], sc->sc_hwdma.wcbBase[i]);
981	}
982	sc->sc_hwdma.maxNumTxWcb = mwl_txbuf;
983	sc->sc_hwdma.maxNumWCB = MWL_NUM_TX_QUEUES-MWL_NUM_ACK_QUEUES;
984
985	error = mwl_hal_sethwdma(sc->sc_mh, &sc->sc_hwdma);
986	if (error != 0) {
987		device_printf(sc->sc_dev,
988		    "unable to setup tx/rx dma; hal status %u\n", error);
989		/* XXX */
990	}
991	return error;
992}
993
994/*
995 * Inform firmware of tx rate parameters.
996 * Called after a channel change.
997 */
998static int
999mwl_setcurchanrates(struct mwl_softc *sc)
1000{
1001	struct ieee80211com *ic = &sc->sc_ic;
1002	const struct ieee80211_rateset *rs;
1003	MWL_HAL_TXRATE rates;
1004
1005	memset(&rates, 0, sizeof(rates));
1006	rs = ieee80211_get_suprates(ic, ic->ic_curchan);
1007	/* rate used to send management frames */
1008	rates.MgtRate = rs->rs_rates[0] & IEEE80211_RATE_VAL;
1009	/* rate used to send multicast frames */
1010	rates.McastRate = rates.MgtRate;
1011
1012	return mwl_hal_settxrate_auto(sc->sc_mh, &rates);
1013}
1014
1015/*
1016 * Inform firmware of tx rate parameters.  Called whenever
1017 * user-settable params change and after a channel change.
1018 */
1019static int
1020mwl_setrates(struct ieee80211vap *vap)
1021{
1022	struct mwl_vap *mvp = MWL_VAP(vap);
1023	struct ieee80211_node *ni = vap->iv_bss;
1024	const struct ieee80211_txparam *tp = ni->ni_txparms;
1025	MWL_HAL_TXRATE rates;
1026
1027	KASSERT(vap->iv_state == IEEE80211_S_RUN, ("state %d", vap->iv_state));
1028
1029	/*
1030	 * Update the h/w rate map.
1031	 * NB: 0x80 for MCS is passed through unchanged
1032	 */
1033	memset(&rates, 0, sizeof(rates));
1034	/* rate used to send management frames */
1035	rates.MgtRate = tp->mgmtrate;
1036	/* rate used to send multicast frames */
1037	rates.McastRate = tp->mcastrate;
1038
1039	/* while here calculate EAPOL fixed rate cookie */
1040	mvp->mv_eapolformat = htole16(mwl_calcformat(rates.MgtRate, ni));
1041
1042	return mwl_hal_settxrate(mvp->mv_hvap,
1043	    tp->ucastrate != IEEE80211_FIXED_RATE_NONE ?
1044		RATE_FIXED : RATE_AUTO, &rates);
1045}
1046
1047/*
1048 * Setup a fixed xmit rate cookie for EAPOL frames.
1049 */
1050static void
1051mwl_seteapolformat(struct ieee80211vap *vap)
1052{
1053	struct mwl_vap *mvp = MWL_VAP(vap);
1054	struct ieee80211_node *ni = vap->iv_bss;
1055	enum ieee80211_phymode mode;
1056	uint8_t rate;
1057
1058	KASSERT(vap->iv_state == IEEE80211_S_RUN, ("state %d", vap->iv_state));
1059
1060	mode = ieee80211_chan2mode(ni->ni_chan);
1061	/*
1062	 * Use legacy rates when operating a mixed HT+non-HT bss.
1063	 * NB: this may violate POLA for sta and wds vap's.
1064	 */
1065	if (mode == IEEE80211_MODE_11NA &&
1066	    (vap->iv_flags_ht & IEEE80211_FHT_PUREN) == 0)
1067		rate = vap->iv_txparms[IEEE80211_MODE_11A].mgmtrate;
1068	else if (mode == IEEE80211_MODE_11NG &&
1069	    (vap->iv_flags_ht & IEEE80211_FHT_PUREN) == 0)
1070		rate = vap->iv_txparms[IEEE80211_MODE_11G].mgmtrate;
1071	else
1072		rate = vap->iv_txparms[mode].mgmtrate;
1073
1074	mvp->mv_eapolformat = htole16(mwl_calcformat(rate, ni));
1075}
1076
1077/*
1078 * Map SKU+country code to region code for radar bin'ing.
1079 */
1080static int
1081mwl_map2regioncode(const struct ieee80211_regdomain *rd)
1082{
1083	switch (rd->regdomain) {
1084	case SKU_FCC:
1085	case SKU_FCC3:
1086		return DOMAIN_CODE_FCC;
1087	case SKU_CA:
1088		return DOMAIN_CODE_IC;
1089	case SKU_ETSI:
1090	case SKU_ETSI2:
1091	case SKU_ETSI3:
1092		if (rd->country == CTRY_SPAIN)
1093			return DOMAIN_CODE_SPAIN;
1094		if (rd->country == CTRY_FRANCE || rd->country == CTRY_FRANCE2)
1095			return DOMAIN_CODE_FRANCE;
1096		/* XXX force 1.3.1 radar type */
1097		return DOMAIN_CODE_ETSI_131;
1098	case SKU_JAPAN:
1099		return DOMAIN_CODE_MKK;
1100	case SKU_ROW:
1101		return DOMAIN_CODE_DGT;	/* Taiwan */
1102	case SKU_APAC:
1103	case SKU_APAC2:
1104	case SKU_APAC3:
1105		return DOMAIN_CODE_AUS;	/* Australia */
1106	}
1107	/* XXX KOREA? */
1108	return DOMAIN_CODE_FCC;			/* XXX? */
1109}
1110
1111static int
1112mwl_hal_reset(struct mwl_softc *sc)
1113{
1114	struct ieee80211com *ic = &sc->sc_ic;
1115	struct mwl_hal *mh = sc->sc_mh;
1116
1117	mwl_hal_setantenna(mh, WL_ANTENNATYPE_RX, sc->sc_rxantenna);
1118	mwl_hal_setantenna(mh, WL_ANTENNATYPE_TX, sc->sc_txantenna);
1119	mwl_hal_setradio(mh, 1, WL_AUTO_PREAMBLE);
1120	mwl_hal_setwmm(sc->sc_mh, (ic->ic_flags & IEEE80211_F_WME) != 0);
1121	mwl_chan_set(sc, ic->ic_curchan);
1122	/* NB: RF/RA performance tuned for indoor mode */
1123	mwl_hal_setrateadaptmode(mh, 0);
1124	mwl_hal_setoptimizationlevel(mh,
1125	    (ic->ic_flags & IEEE80211_F_BURST) != 0);
1126
1127	mwl_hal_setregioncode(mh, mwl_map2regioncode(&ic->ic_regdomain));
1128
1129	mwl_hal_setaggampduratemode(mh, 1, 80);		/* XXX */
1130	mwl_hal_setcfend(mh, 0);			/* XXX */
1131
1132	return 1;
1133}
1134
1135static int
1136mwl_init(struct mwl_softc *sc)
1137{
1138	struct mwl_hal *mh = sc->sc_mh;
1139	int error = 0;
1140
1141	MWL_LOCK_ASSERT(sc);
1142
1143	/*
1144	 * Stop anything previously setup.  This is safe
1145	 * whether this is the first time through or not.
1146	 */
1147	mwl_stop(sc);
1148
1149	/*
1150	 * Push vap-independent state to the firmware.
1151	 */
1152	if (!mwl_hal_reset(sc)) {
1153		device_printf(sc->sc_dev, "unable to reset hardware\n");
1154		return EIO;
1155	}
1156
1157	/*
1158	 * Setup recv (once); transmit is already good to go.
1159	 */
1160	error = mwl_startrecv(sc);
1161	if (error != 0) {
1162		device_printf(sc->sc_dev, "unable to start recv logic\n");
1163		return error;
1164	}
1165
1166	/*
1167	 * Enable interrupts.
1168	 */
1169	sc->sc_imask = MACREG_A2HRIC_BIT_RX_RDY
1170		     | MACREG_A2HRIC_BIT_TX_DONE
1171		     | MACREG_A2HRIC_BIT_OPC_DONE
1172#if 0
1173		     | MACREG_A2HRIC_BIT_MAC_EVENT
1174#endif
1175		     | MACREG_A2HRIC_BIT_ICV_ERROR
1176		     | MACREG_A2HRIC_BIT_RADAR_DETECT
1177		     | MACREG_A2HRIC_BIT_CHAN_SWITCH
1178#if 0
1179		     | MACREG_A2HRIC_BIT_QUEUE_EMPTY
1180#endif
1181		     | MACREG_A2HRIC_BIT_BA_WATCHDOG
1182		     | MACREQ_A2HRIC_BIT_TX_ACK
1183		     ;
1184
1185	sc->sc_running = 1;
1186	mwl_hal_intrset(mh, sc->sc_imask);
1187	callout_reset(&sc->sc_watchdog, hz, mwl_watchdog, sc);
1188
1189	return 0;
1190}
1191
1192static void
1193mwl_stop(struct mwl_softc *sc)
1194{
1195
1196	MWL_LOCK_ASSERT(sc);
1197	if (sc->sc_running) {
1198		/*
1199		 * Shutdown the hardware and driver.
1200		 */
1201		sc->sc_running = 0;
1202		callout_stop(&sc->sc_watchdog);
1203		sc->sc_tx_timer = 0;
1204		mwl_draintxq(sc);
1205	}
1206}
1207
1208static int
1209mwl_reset_vap(struct ieee80211vap *vap, int state)
1210{
1211	struct mwl_hal_vap *hvap = MWL_VAP(vap)->mv_hvap;
1212	struct ieee80211com *ic = vap->iv_ic;
1213
1214	if (state == IEEE80211_S_RUN)
1215		mwl_setrates(vap);
1216	/* XXX off by 1? */
1217	mwl_hal_setrtsthreshold(hvap, vap->iv_rtsthreshold);
1218	/* XXX auto? 20/40 split? */
1219	mwl_hal_sethtgi(hvap, (vap->iv_flags_ht &
1220	    (IEEE80211_FHT_SHORTGI20|IEEE80211_FHT_SHORTGI40)) ? 1 : 0);
1221	mwl_hal_setnprot(hvap, ic->ic_htprotmode == IEEE80211_PROT_NONE ?
1222	    HTPROTECT_NONE : HTPROTECT_AUTO);
1223	/* XXX txpower cap */
1224
1225	/* re-setup beacons */
1226	if (state == IEEE80211_S_RUN &&
1227	    (vap->iv_opmode == IEEE80211_M_HOSTAP ||
1228	     vap->iv_opmode == IEEE80211_M_MBSS ||
1229	     vap->iv_opmode == IEEE80211_M_IBSS)) {
1230		mwl_setapmode(vap, vap->iv_bss->ni_chan);
1231		mwl_hal_setnprotmode(hvap,
1232		    MS(ic->ic_curhtprotmode, IEEE80211_HTINFO_OPMODE));
1233		return mwl_beacon_setup(vap);
1234	}
1235	return 0;
1236}
1237
1238/*
1239 * Reset the hardware w/o losing operational state.
1240 * Used to to reset or reload hardware state for a vap.
1241 */
1242static int
1243mwl_reset(struct ieee80211vap *vap, u_long cmd)
1244{
1245	struct mwl_hal_vap *hvap = MWL_VAP(vap)->mv_hvap;
1246	int error = 0;
1247
1248	if (hvap != NULL) {			/* WDS, MONITOR, etc. */
1249		struct ieee80211com *ic = vap->iv_ic;
1250		struct mwl_softc *sc = ic->ic_softc;
1251		struct mwl_hal *mh = sc->sc_mh;
1252
1253		/* XXX handle DWDS sta vap change */
1254		/* XXX do we need to disable interrupts? */
1255		mwl_hal_intrset(mh, 0);		/* disable interrupts */
1256		error = mwl_reset_vap(vap, vap->iv_state);
1257		mwl_hal_intrset(mh, sc->sc_imask);
1258	}
1259	return error;
1260}
1261
1262/*
1263 * Allocate a tx buffer for sending a frame.  The
1264 * packet is assumed to have the WME AC stored so
1265 * we can use it to select the appropriate h/w queue.
1266 */
1267static struct mwl_txbuf *
1268mwl_gettxbuf(struct mwl_softc *sc, struct mwl_txq *txq)
1269{
1270	struct mwl_txbuf *bf;
1271
1272	/*
1273	 * Grab a TX buffer and associated resources.
1274	 */
1275	MWL_TXQ_LOCK(txq);
1276	bf = STAILQ_FIRST(&txq->free);
1277	if (bf != NULL) {
1278		STAILQ_REMOVE_HEAD(&txq->free, bf_list);
1279		txq->nfree--;
1280	}
1281	MWL_TXQ_UNLOCK(txq);
1282	if (bf == NULL)
1283		DPRINTF(sc, MWL_DEBUG_XMIT,
1284		    "%s: out of xmit buffers on q %d\n", __func__, txq->qnum);
1285	return bf;
1286}
1287
1288/*
1289 * Return a tx buffer to the queue it came from.  Note there
1290 * are two cases because we must preserve the order of buffers
1291 * as it reflects the fixed order of descriptors in memory
1292 * (the firmware pre-fetches descriptors so we cannot reorder).
1293 */
1294static void
1295mwl_puttxbuf_head(struct mwl_txq *txq, struct mwl_txbuf *bf)
1296{
1297	bf->bf_m = NULL;
1298	bf->bf_node = NULL;
1299	MWL_TXQ_LOCK(txq);
1300	STAILQ_INSERT_HEAD(&txq->free, bf, bf_list);
1301	txq->nfree++;
1302	MWL_TXQ_UNLOCK(txq);
1303}
1304
1305static void
1306mwl_puttxbuf_tail(struct mwl_txq *txq, struct mwl_txbuf *bf)
1307{
1308	bf->bf_m = NULL;
1309	bf->bf_node = NULL;
1310	MWL_TXQ_LOCK(txq);
1311	STAILQ_INSERT_TAIL(&txq->free, bf, bf_list);
1312	txq->nfree++;
1313	MWL_TXQ_UNLOCK(txq);
1314}
1315
1316static int
1317mwl_transmit(struct ieee80211com *ic, struct mbuf *m)
1318{
1319	struct mwl_softc *sc = ic->ic_softc;
1320	int error;
1321
1322	MWL_LOCK(sc);
1323	if (!sc->sc_running) {
1324		MWL_UNLOCK(sc);
1325		return (ENXIO);
1326	}
1327	error = mbufq_enqueue(&sc->sc_snd, m);
1328	if (error) {
1329		MWL_UNLOCK(sc);
1330		return (error);
1331	}
1332	mwl_start(sc);
1333	MWL_UNLOCK(sc);
1334	return (0);
1335}
1336
1337static void
1338mwl_start(struct mwl_softc *sc)
1339{
1340	struct ieee80211_node *ni;
1341	struct mwl_txbuf *bf;
1342	struct mbuf *m;
1343	struct mwl_txq *txq = NULL;	/* XXX silence gcc */
1344	int nqueued;
1345
1346	MWL_LOCK_ASSERT(sc);
1347	if (!sc->sc_running || sc->sc_invalid)
1348		return;
1349	nqueued = 0;
1350	while ((m = mbufq_dequeue(&sc->sc_snd)) != NULL) {
1351		/*
1352		 * Grab the node for the destination.
1353		 */
1354		ni = (struct ieee80211_node *) m->m_pkthdr.rcvif;
1355		KASSERT(ni != NULL, ("no node"));
1356		m->m_pkthdr.rcvif = NULL;	/* committed, clear ref */
1357		/*
1358		 * Grab a TX buffer and associated resources.
1359		 * We honor the classification by the 802.11 layer.
1360		 */
1361		txq = sc->sc_ac2q[M_WME_GETAC(m)];
1362		bf = mwl_gettxbuf(sc, txq);
1363		if (bf == NULL) {
1364			m_freem(m);
1365			ieee80211_free_node(ni);
1366#ifdef MWL_TX_NODROP
1367			sc->sc_stats.mst_tx_qstop++;
1368			break;
1369#else
1370			DPRINTF(sc, MWL_DEBUG_XMIT,
1371			    "%s: tail drop on q %d\n", __func__, txq->qnum);
1372			sc->sc_stats.mst_tx_qdrop++;
1373			continue;
1374#endif /* MWL_TX_NODROP */
1375		}
1376
1377		/*
1378		 * Pass the frame to the h/w for transmission.
1379		 */
1380		if (mwl_tx_start(sc, ni, bf, m)) {
1381			if_inc_counter(ni->ni_vap->iv_ifp,
1382			    IFCOUNTER_OERRORS, 1);
1383			mwl_puttxbuf_head(txq, bf);
1384			ieee80211_free_node(ni);
1385			continue;
1386		}
1387		nqueued++;
1388		if (nqueued >= mwl_txcoalesce) {
1389			/*
1390			 * Poke the firmware to process queued frames;
1391			 * see below about (lack of) locking.
1392			 */
1393			nqueued = 0;
1394			mwl_hal_txstart(sc->sc_mh, 0/*XXX*/);
1395		}
1396	}
1397	if (nqueued) {
1398		/*
1399		 * NB: We don't need to lock against tx done because
1400		 * this just prods the firmware to check the transmit
1401		 * descriptors.  The firmware will also start fetching
1402		 * descriptors by itself if it notices new ones are
1403		 * present when it goes to deliver a tx done interrupt
1404		 * to the host. So if we race with tx done processing
1405		 * it's ok.  Delivering the kick here rather than in
1406		 * mwl_tx_start is an optimization to avoid poking the
1407		 * firmware for each packet.
1408		 *
1409		 * NB: the queue id isn't used so 0 is ok.
1410		 */
1411		mwl_hal_txstart(sc->sc_mh, 0/*XXX*/);
1412	}
1413}
1414
1415static int
1416mwl_raw_xmit(struct ieee80211_node *ni, struct mbuf *m,
1417	const struct ieee80211_bpf_params *params)
1418{
1419	struct ieee80211com *ic = ni->ni_ic;
1420	struct mwl_softc *sc = ic->ic_softc;
1421	struct mwl_txbuf *bf;
1422	struct mwl_txq *txq;
1423
1424	if (!sc->sc_running || sc->sc_invalid) {
1425		m_freem(m);
1426		return ENETDOWN;
1427	}
1428	/*
1429	 * Grab a TX buffer and associated resources.
1430	 * Note that we depend on the classification
1431	 * by the 802.11 layer to get to the right h/w
1432	 * queue.  Management frames must ALWAYS go on
1433	 * queue 1 but we cannot just force that here
1434	 * because we may receive non-mgt frames.
1435	 */
1436	txq = sc->sc_ac2q[M_WME_GETAC(m)];
1437	bf = mwl_gettxbuf(sc, txq);
1438	if (bf == NULL) {
1439		sc->sc_stats.mst_tx_qstop++;
1440		m_freem(m);
1441		return ENOBUFS;
1442	}
1443	/*
1444	 * Pass the frame to the h/w for transmission.
1445	 */
1446	if (mwl_tx_start(sc, ni, bf, m)) {
1447		mwl_puttxbuf_head(txq, bf);
1448
1449		return EIO;		/* XXX */
1450	}
1451	/*
1452	 * NB: We don't need to lock against tx done because
1453	 * this just prods the firmware to check the transmit
1454	 * descriptors.  The firmware will also start fetching
1455	 * descriptors by itself if it notices new ones are
1456	 * present when it goes to deliver a tx done interrupt
1457	 * to the host. So if we race with tx done processing
1458	 * it's ok.  Delivering the kick here rather than in
1459	 * mwl_tx_start is an optimization to avoid poking the
1460	 * firmware for each packet.
1461	 *
1462	 * NB: the queue id isn't used so 0 is ok.
1463	 */
1464	mwl_hal_txstart(sc->sc_mh, 0/*XXX*/);
1465	return 0;
1466}
1467
1468static int
1469mwl_media_change(struct ifnet *ifp)
1470{
1471	struct ieee80211vap *vap = ifp->if_softc;
1472	int error;
1473
1474	error = ieee80211_media_change(ifp);
1475	/* NB: only the fixed rate can change and that doesn't need a reset */
1476	if (error == ENETRESET) {
1477		mwl_setrates(vap);
1478		error = 0;
1479	}
1480	return error;
1481}
1482
1483#ifdef MWL_DEBUG
1484static void
1485mwl_keyprint(struct mwl_softc *sc, const char *tag,
1486	const MWL_HAL_KEYVAL *hk, const uint8_t mac[IEEE80211_ADDR_LEN])
1487{
1488	static const char *ciphers[] = {
1489		"WEP",
1490		"TKIP",
1491		"AES-CCM",
1492	};
1493	int i, n;
1494
1495	printf("%s: [%u] %-7s", tag, hk->keyIndex, ciphers[hk->keyTypeId]);
1496	for (i = 0, n = hk->keyLen; i < n; i++)
1497		printf(" %02x", hk->key.aes[i]);
1498	printf(" mac %s", ether_sprintf(mac));
1499	if (hk->keyTypeId == KEY_TYPE_ID_TKIP) {
1500		printf(" %s", "rxmic");
1501		for (i = 0; i < sizeof(hk->key.tkip.rxMic); i++)
1502			printf(" %02x", hk->key.tkip.rxMic[i]);
1503		printf(" txmic");
1504		for (i = 0; i < sizeof(hk->key.tkip.txMic); i++)
1505			printf(" %02x", hk->key.tkip.txMic[i]);
1506	}
1507	printf(" flags 0x%x\n", hk->keyFlags);
1508}
1509#endif
1510
1511/*
1512 * Allocate a key cache slot for a unicast key.  The
1513 * firmware handles key allocation and every station is
1514 * guaranteed key space so we are always successful.
1515 */
1516static int
1517mwl_key_alloc(struct ieee80211vap *vap, struct ieee80211_key *k,
1518	ieee80211_keyix *keyix, ieee80211_keyix *rxkeyix)
1519{
1520	struct mwl_softc *sc = vap->iv_ic->ic_softc;
1521
1522	if (k->wk_keyix != IEEE80211_KEYIX_NONE ||
1523	    (k->wk_flags & IEEE80211_KEY_GROUP)) {
1524		if (!(&vap->iv_nw_keys[0] <= k &&
1525		      k < &vap->iv_nw_keys[IEEE80211_WEP_NKID])) {
1526			/* should not happen */
1527			DPRINTF(sc, MWL_DEBUG_KEYCACHE,
1528				"%s: bogus group key\n", __func__);
1529			return 0;
1530		}
1531		/* give the caller what they requested */
1532		*keyix = *rxkeyix = k - vap->iv_nw_keys;
1533	} else {
1534		/*
1535		 * Firmware handles key allocation.
1536		 */
1537		*keyix = *rxkeyix = 0;
1538	}
1539	return 1;
1540}
1541
1542/*
1543 * Delete a key entry allocated by mwl_key_alloc.
1544 */
1545static int
1546mwl_key_delete(struct ieee80211vap *vap, const struct ieee80211_key *k)
1547{
1548	struct mwl_softc *sc = vap->iv_ic->ic_softc;
1549	struct mwl_hal_vap *hvap = MWL_VAP(vap)->mv_hvap;
1550	MWL_HAL_KEYVAL hk;
1551	const uint8_t bcastaddr[IEEE80211_ADDR_LEN] =
1552	    { 0xff, 0xff, 0xff, 0xff, 0xff, 0xff };
1553
1554	if (hvap == NULL) {
1555		if (vap->iv_opmode != IEEE80211_M_WDS) {
1556			/* XXX monitor mode? */
1557			DPRINTF(sc, MWL_DEBUG_KEYCACHE,
1558			    "%s: no hvap for opmode %d\n", __func__,
1559			    vap->iv_opmode);
1560			return 0;
1561		}
1562		hvap = MWL_VAP(vap)->mv_ap_hvap;
1563	}
1564
1565	DPRINTF(sc, MWL_DEBUG_KEYCACHE, "%s: delete key %u\n",
1566	    __func__, k->wk_keyix);
1567
1568	memset(&hk, 0, sizeof(hk));
1569	hk.keyIndex = k->wk_keyix;
1570	switch (k->wk_cipher->ic_cipher) {
1571	case IEEE80211_CIPHER_WEP:
1572		hk.keyTypeId = KEY_TYPE_ID_WEP;
1573		break;
1574	case IEEE80211_CIPHER_TKIP:
1575		hk.keyTypeId = KEY_TYPE_ID_TKIP;
1576		break;
1577	case IEEE80211_CIPHER_AES_CCM:
1578		hk.keyTypeId = KEY_TYPE_ID_AES;
1579		break;
1580	default:
1581		/* XXX should not happen */
1582		DPRINTF(sc, MWL_DEBUG_KEYCACHE, "%s: unknown cipher %d\n",
1583		    __func__, k->wk_cipher->ic_cipher);
1584		return 0;
1585	}
1586	return (mwl_hal_keyreset(hvap, &hk, bcastaddr) == 0);	/*XXX*/
1587}
1588
1589static __inline int
1590addgroupflags(MWL_HAL_KEYVAL *hk, const struct ieee80211_key *k)
1591{
1592	if (k->wk_flags & IEEE80211_KEY_GROUP) {
1593		if (k->wk_flags & IEEE80211_KEY_XMIT)
1594			hk->keyFlags |= KEY_FLAG_TXGROUPKEY;
1595		if (k->wk_flags & IEEE80211_KEY_RECV)
1596			hk->keyFlags |= KEY_FLAG_RXGROUPKEY;
1597		return 1;
1598	} else
1599		return 0;
1600}
1601
1602/*
1603 * Set the key cache contents for the specified key.  Key cache
1604 * slot(s) must already have been allocated by mwl_key_alloc.
1605 */
1606static int
1607mwl_key_set(struct ieee80211vap *vap, const struct ieee80211_key *k)
1608{
1609	return (_mwl_key_set(vap, k, k->wk_macaddr));
1610}
1611
1612static int
1613_mwl_key_set(struct ieee80211vap *vap, const struct ieee80211_key *k,
1614	const uint8_t mac[IEEE80211_ADDR_LEN])
1615{
1616#define	GRPXMIT	(IEEE80211_KEY_XMIT | IEEE80211_KEY_GROUP)
1617/* NB: static wep keys are marked GROUP+tx/rx; GTK will be tx or rx */
1618#define	IEEE80211_IS_STATICKEY(k) \
1619	(((k)->wk_flags & (GRPXMIT|IEEE80211_KEY_RECV)) == \
1620	 (GRPXMIT|IEEE80211_KEY_RECV))
1621	struct mwl_softc *sc = vap->iv_ic->ic_softc;
1622	struct mwl_hal_vap *hvap = MWL_VAP(vap)->mv_hvap;
1623	const struct ieee80211_cipher *cip = k->wk_cipher;
1624	const uint8_t *macaddr;
1625	MWL_HAL_KEYVAL hk;
1626
1627	KASSERT((k->wk_flags & IEEE80211_KEY_SWCRYPT) == 0,
1628		("s/w crypto set?"));
1629
1630	if (hvap == NULL) {
1631		if (vap->iv_opmode != IEEE80211_M_WDS) {
1632			/* XXX monitor mode? */
1633			DPRINTF(sc, MWL_DEBUG_KEYCACHE,
1634			    "%s: no hvap for opmode %d\n", __func__,
1635			    vap->iv_opmode);
1636			return 0;
1637		}
1638		hvap = MWL_VAP(vap)->mv_ap_hvap;
1639	}
1640	memset(&hk, 0, sizeof(hk));
1641	hk.keyIndex = k->wk_keyix;
1642	switch (cip->ic_cipher) {
1643	case IEEE80211_CIPHER_WEP:
1644		hk.keyTypeId = KEY_TYPE_ID_WEP;
1645		hk.keyLen = k->wk_keylen;
1646		if (k->wk_keyix == vap->iv_def_txkey)
1647			hk.keyFlags = KEY_FLAG_WEP_TXKEY;
1648		if (!IEEE80211_IS_STATICKEY(k)) {
1649			/* NB: WEP is never used for the PTK */
1650			(void) addgroupflags(&hk, k);
1651		}
1652		break;
1653	case IEEE80211_CIPHER_TKIP:
1654		hk.keyTypeId = KEY_TYPE_ID_TKIP;
1655		hk.key.tkip.tsc.high = (uint32_t)(k->wk_keytsc >> 16);
1656		hk.key.tkip.tsc.low = (uint16_t)k->wk_keytsc;
1657		hk.keyFlags = KEY_FLAG_TSC_VALID | KEY_FLAG_MICKEY_VALID;
1658		hk.keyLen = k->wk_keylen + IEEE80211_MICBUF_SIZE;
1659		if (!addgroupflags(&hk, k))
1660			hk.keyFlags |= KEY_FLAG_PAIRWISE;
1661		break;
1662	case IEEE80211_CIPHER_AES_CCM:
1663		hk.keyTypeId = KEY_TYPE_ID_AES;
1664		hk.keyLen = k->wk_keylen;
1665		if (!addgroupflags(&hk, k))
1666			hk.keyFlags |= KEY_FLAG_PAIRWISE;
1667		break;
1668	default:
1669		/* XXX should not happen */
1670		DPRINTF(sc, MWL_DEBUG_KEYCACHE, "%s: unknown cipher %d\n",
1671		    __func__, k->wk_cipher->ic_cipher);
1672		return 0;
1673	}
1674	/*
1675	 * NB: tkip mic keys get copied here too; the layout
1676	 *     just happens to match that in ieee80211_key.
1677	 */
1678	memcpy(hk.key.aes, k->wk_key, hk.keyLen);
1679
1680	/*
1681	 * Locate address of sta db entry for writing key;
1682	 * the convention unfortunately is somewhat different
1683	 * than how net80211, hostapd, and wpa_supplicant think.
1684	 */
1685	if (vap->iv_opmode == IEEE80211_M_STA) {
1686		/*
1687		 * NB: keys plumbed before the sta reaches AUTH state
1688		 * will be discarded or written to the wrong sta db
1689		 * entry because iv_bss is meaningless.  This is ok
1690		 * (right now) because we handle deferred plumbing of
1691		 * WEP keys when the sta reaches AUTH state.
1692		 */
1693		macaddr = vap->iv_bss->ni_bssid;
1694		if ((k->wk_flags & IEEE80211_KEY_GROUP) == 0) {
1695			/* XXX plumb to local sta db too for static key wep */
1696			mwl_hal_keyset(hvap, &hk, vap->iv_myaddr);
1697		}
1698	} else if (vap->iv_opmode == IEEE80211_M_WDS &&
1699	    vap->iv_state != IEEE80211_S_RUN) {
1700		/*
1701		 * Prior to RUN state a WDS vap will not it's BSS node
1702		 * setup so we will plumb the key to the wrong mac
1703		 * address (it'll be our local address).  Workaround
1704		 * this for the moment by grabbing the correct address.
1705		 */
1706		macaddr = vap->iv_des_bssid;
1707	} else if ((k->wk_flags & GRPXMIT) == GRPXMIT)
1708		macaddr = vap->iv_myaddr;
1709	else
1710		macaddr = mac;
1711	KEYPRINTF(sc, &hk, macaddr);
1712	return (mwl_hal_keyset(hvap, &hk, macaddr) == 0);
1713#undef IEEE80211_IS_STATICKEY
1714#undef GRPXMIT
1715}
1716
1717/*
1718 * Set the multicast filter contents into the hardware.
1719 * XXX f/w has no support; just defer to the os.
1720 */
1721static void
1722mwl_setmcastfilter(struct mwl_softc *sc)
1723{
1724#if 0
1725	struct ether_multi *enm;
1726	struct ether_multistep estep;
1727	uint8_t macs[IEEE80211_ADDR_LEN*MWL_HAL_MCAST_MAX];/* XXX stack use */
1728	uint8_t *mp;
1729	int nmc;
1730
1731	mp = macs;
1732	nmc = 0;
1733	ETHER_FIRST_MULTI(estep, &sc->sc_ec, enm);
1734	while (enm != NULL) {
1735		/* XXX Punt on ranges. */
1736		if (nmc == MWL_HAL_MCAST_MAX ||
1737		    !IEEE80211_ADDR_EQ(enm->enm_addrlo, enm->enm_addrhi)) {
1738			ifp->if_flags |= IFF_ALLMULTI;
1739			return;
1740		}
1741		IEEE80211_ADDR_COPY(mp, enm->enm_addrlo);
1742		mp += IEEE80211_ADDR_LEN, nmc++;
1743		ETHER_NEXT_MULTI(estep, enm);
1744	}
1745	ifp->if_flags &= ~IFF_ALLMULTI;
1746	mwl_hal_setmcast(sc->sc_mh, nmc, macs);
1747#endif
1748}
1749
1750static int
1751mwl_mode_init(struct mwl_softc *sc)
1752{
1753	struct ieee80211com *ic = &sc->sc_ic;
1754	struct mwl_hal *mh = sc->sc_mh;
1755
1756	/*
1757	 * NB: Ignore promisc in hostap mode; it's set by the
1758	 * bridge.  This is wrong but we have no way to
1759	 * identify internal requests (from the bridge)
1760	 * versus external requests such as for tcpdump.
1761	 */
1762	mwl_hal_setpromisc(mh, ic->ic_promisc > 0 &&
1763	    ic->ic_opmode != IEEE80211_M_HOSTAP);
1764	mwl_setmcastfilter(sc);
1765
1766	return 0;
1767}
1768
1769/*
1770 * Callback from the 802.11 layer after a multicast state change.
1771 */
1772static void
1773mwl_update_mcast(struct ieee80211com *ic)
1774{
1775	struct mwl_softc *sc = ic->ic_softc;
1776
1777	mwl_setmcastfilter(sc);
1778}
1779
1780/*
1781 * Callback from the 802.11 layer after a promiscuous mode change.
1782 * Note this interface does not check the operating mode as this
1783 * is an internal callback and we are expected to honor the current
1784 * state (e.g. this is used for setting the interface in promiscuous
1785 * mode when operating in hostap mode to do ACS).
1786 */
1787static void
1788mwl_update_promisc(struct ieee80211com *ic)
1789{
1790	struct mwl_softc *sc = ic->ic_softc;
1791
1792	mwl_hal_setpromisc(sc->sc_mh, ic->ic_promisc > 0);
1793}
1794
1795/*
1796 * Callback from the 802.11 layer to update the slot time
1797 * based on the current setting.  We use it to notify the
1798 * firmware of ERP changes and the f/w takes care of things
1799 * like slot time and preamble.
1800 */
1801static void
1802mwl_updateslot(struct ieee80211com *ic)
1803{
1804	struct mwl_softc *sc = ic->ic_softc;
1805	struct mwl_hal *mh = sc->sc_mh;
1806	int prot;
1807
1808	/* NB: can be called early; suppress needless cmds */
1809	if (!sc->sc_running)
1810		return;
1811
1812	/*
1813	 * Calculate the ERP flags.  The firwmare will use
1814	 * this to carry out the appropriate measures.
1815	 */
1816	prot = 0;
1817	if (IEEE80211_IS_CHAN_ANYG(ic->ic_curchan)) {
1818		if ((ic->ic_flags & IEEE80211_F_SHSLOT) == 0)
1819			prot |= IEEE80211_ERP_NON_ERP_PRESENT;
1820		if (ic->ic_flags & IEEE80211_F_USEPROT)
1821			prot |= IEEE80211_ERP_USE_PROTECTION;
1822		if (ic->ic_flags & IEEE80211_F_USEBARKER)
1823			prot |= IEEE80211_ERP_LONG_PREAMBLE;
1824	}
1825
1826	DPRINTF(sc, MWL_DEBUG_RESET,
1827	    "%s: chan %u MHz/flags 0x%x %s slot, (prot 0x%x ic_flags 0x%x)\n",
1828	    __func__, ic->ic_curchan->ic_freq, ic->ic_curchan->ic_flags,
1829	    ic->ic_flags & IEEE80211_F_SHSLOT ? "short" : "long", prot,
1830	    ic->ic_flags);
1831
1832	mwl_hal_setgprot(mh, prot);
1833}
1834
1835/*
1836 * Setup the beacon frame.
1837 */
1838static int
1839mwl_beacon_setup(struct ieee80211vap *vap)
1840{
1841	struct mwl_hal_vap *hvap = MWL_VAP(vap)->mv_hvap;
1842	struct ieee80211_node *ni = vap->iv_bss;
1843	struct mbuf *m;
1844
1845	m = ieee80211_beacon_alloc(ni);
1846	if (m == NULL)
1847		return ENOBUFS;
1848	mwl_hal_setbeacon(hvap, mtod(m, const void *), m->m_len);
1849	m_free(m);
1850
1851	return 0;
1852}
1853
1854/*
1855 * Update the beacon frame in response to a change.
1856 */
1857static void
1858mwl_beacon_update(struct ieee80211vap *vap, int item)
1859{
1860	struct mwl_hal_vap *hvap = MWL_VAP(vap)->mv_hvap;
1861	struct ieee80211com *ic = vap->iv_ic;
1862
1863	KASSERT(hvap != NULL, ("no beacon"));
1864	switch (item) {
1865	case IEEE80211_BEACON_ERP:
1866		mwl_updateslot(ic);
1867		break;
1868	case IEEE80211_BEACON_HTINFO:
1869		mwl_hal_setnprotmode(hvap,
1870		    MS(ic->ic_curhtprotmode, IEEE80211_HTINFO_OPMODE));
1871		break;
1872	case IEEE80211_BEACON_CAPS:
1873	case IEEE80211_BEACON_WME:
1874	case IEEE80211_BEACON_APPIE:
1875	case IEEE80211_BEACON_CSA:
1876		break;
1877	case IEEE80211_BEACON_TIM:
1878		/* NB: firmware always forms TIM */
1879		return;
1880	}
1881	/* XXX retain beacon frame and update */
1882	mwl_beacon_setup(vap);
1883}
1884
1885static void
1886mwl_load_cb(void *arg, bus_dma_segment_t *segs, int nsegs, int error)
1887{
1888	bus_addr_t *paddr = (bus_addr_t*) arg;
1889	KASSERT(error == 0, ("error %u on bus_dma callback", error));
1890	*paddr = segs->ds_addr;
1891}
1892
1893#ifdef MWL_HOST_PS_SUPPORT
1894/*
1895 * Handle power save station occupancy changes.
1896 */
1897static void
1898mwl_update_ps(struct ieee80211vap *vap, int nsta)
1899{
1900	struct mwl_vap *mvp = MWL_VAP(vap);
1901
1902	if (nsta == 0 || mvp->mv_last_ps_sta == 0)
1903		mwl_hal_setpowersave_bss(mvp->mv_hvap, nsta);
1904	mvp->mv_last_ps_sta = nsta;
1905}
1906
1907/*
1908 * Handle associated station power save state changes.
1909 */
1910static int
1911mwl_set_tim(struct ieee80211_node *ni, int set)
1912{
1913	struct ieee80211vap *vap = ni->ni_vap;
1914	struct mwl_vap *mvp = MWL_VAP(vap);
1915
1916	if (mvp->mv_set_tim(ni, set)) {		/* NB: state change */
1917		mwl_hal_setpowersave_sta(mvp->mv_hvap,
1918		    IEEE80211_AID(ni->ni_associd), set);
1919		return 1;
1920	} else
1921		return 0;
1922}
1923#endif /* MWL_HOST_PS_SUPPORT */
1924
1925static int
1926mwl_desc_setup(struct mwl_softc *sc, const char *name,
1927	struct mwl_descdma *dd,
1928	int nbuf, size_t bufsize, int ndesc, size_t descsize)
1929{
1930	uint8_t *ds;
1931	int error;
1932
1933	DPRINTF(sc, MWL_DEBUG_RESET,
1934	    "%s: %s DMA: %u bufs (%ju) %u desc/buf (%ju)\n",
1935	    __func__, name, nbuf, (uintmax_t) bufsize,
1936	    ndesc, (uintmax_t) descsize);
1937
1938	dd->dd_name = name;
1939	dd->dd_desc_len = nbuf * ndesc * descsize;
1940
1941	/*
1942	 * Setup DMA descriptor area.
1943	 */
1944	error = bus_dma_tag_create(bus_get_dma_tag(sc->sc_dev),	/* parent */
1945		       PAGE_SIZE, 0,		/* alignment, bounds */
1946		       BUS_SPACE_MAXADDR_32BIT,	/* lowaddr */
1947		       BUS_SPACE_MAXADDR,	/* highaddr */
1948		       NULL, NULL,		/* filter, filterarg */
1949		       dd->dd_desc_len,		/* maxsize */
1950		       1,			/* nsegments */
1951		       dd->dd_desc_len,		/* maxsegsize */
1952		       BUS_DMA_ALLOCNOW,	/* flags */
1953		       NULL,			/* lockfunc */
1954		       NULL,			/* lockarg */
1955		       &dd->dd_dmat);
1956	if (error != 0) {
1957		device_printf(sc->sc_dev, "cannot allocate %s DMA tag\n", dd->dd_name);
1958		return error;
1959	}
1960
1961	/* allocate descriptors */
1962	error = bus_dmamem_alloc(dd->dd_dmat, (void**) &dd->dd_desc,
1963				 BUS_DMA_NOWAIT | BUS_DMA_COHERENT,
1964				 &dd->dd_dmamap);
1965	if (error != 0) {
1966		device_printf(sc->sc_dev, "unable to alloc memory for %u %s descriptors, "
1967			"error %u\n", nbuf * ndesc, dd->dd_name, error);
1968		goto fail1;
1969	}
1970
1971	error = bus_dmamap_load(dd->dd_dmat, dd->dd_dmamap,
1972				dd->dd_desc, dd->dd_desc_len,
1973				mwl_load_cb, &dd->dd_desc_paddr,
1974				BUS_DMA_NOWAIT);
1975	if (error != 0) {
1976		device_printf(sc->sc_dev, "unable to map %s descriptors, error %u\n",
1977			dd->dd_name, error);
1978		goto fail2;
1979	}
1980
1981	ds = dd->dd_desc;
1982	memset(ds, 0, dd->dd_desc_len);
1983	DPRINTF(sc, MWL_DEBUG_RESET,
1984	    "%s: %s DMA map: %p (%lu) -> 0x%jx (%lu)\n",
1985	    __func__, dd->dd_name, ds, (u_long) dd->dd_desc_len,
1986	    (uintmax_t) dd->dd_desc_paddr, /*XXX*/ (u_long) dd->dd_desc_len);
1987
1988	return 0;
1989fail2:
1990	bus_dmamem_free(dd->dd_dmat, dd->dd_desc, dd->dd_dmamap);
1991fail1:
1992	bus_dma_tag_destroy(dd->dd_dmat);
1993	memset(dd, 0, sizeof(*dd));
1994	return error;
1995#undef DS2PHYS
1996}
1997
1998static void
1999mwl_desc_cleanup(struct mwl_softc *sc, struct mwl_descdma *dd)
2000{
2001	bus_dmamap_unload(dd->dd_dmat, dd->dd_dmamap);
2002	bus_dmamem_free(dd->dd_dmat, dd->dd_desc, dd->dd_dmamap);
2003	bus_dma_tag_destroy(dd->dd_dmat);
2004
2005	memset(dd, 0, sizeof(*dd));
2006}
2007
2008/*
2009 * Construct a tx q's free list.  The order of entries on
2010 * the list must reflect the physical layout of tx descriptors
2011 * because the firmware pre-fetches descriptors.
2012 *
2013 * XXX might be better to use indices into the buffer array.
2014 */
2015static void
2016mwl_txq_reset(struct mwl_softc *sc, struct mwl_txq *txq)
2017{
2018	struct mwl_txbuf *bf;
2019	int i;
2020
2021	bf = txq->dma.dd_bufptr;
2022	STAILQ_INIT(&txq->free);
2023	for (i = 0; i < mwl_txbuf; i++, bf++)
2024		STAILQ_INSERT_TAIL(&txq->free, bf, bf_list);
2025	txq->nfree = i;
2026}
2027
2028#define	DS2PHYS(_dd, _ds) \
2029	((_dd)->dd_desc_paddr + ((caddr_t)(_ds) - (caddr_t)(_dd)->dd_desc))
2030
2031static int
2032mwl_txdma_setup(struct mwl_softc *sc, struct mwl_txq *txq)
2033{
2034	int error, bsize, i;
2035	struct mwl_txbuf *bf;
2036	struct mwl_txdesc *ds;
2037
2038	error = mwl_desc_setup(sc, "tx", &txq->dma,
2039			mwl_txbuf, sizeof(struct mwl_txbuf),
2040			MWL_TXDESC, sizeof(struct mwl_txdesc));
2041	if (error != 0)
2042		return error;
2043
2044	/* allocate and setup tx buffers */
2045	bsize = mwl_txbuf * sizeof(struct mwl_txbuf);
2046	bf = malloc(bsize, M_MWLDEV, M_NOWAIT | M_ZERO);
2047	if (bf == NULL) {
2048		device_printf(sc->sc_dev, "malloc of %u tx buffers failed\n",
2049			mwl_txbuf);
2050		return ENOMEM;
2051	}
2052	txq->dma.dd_bufptr = bf;
2053
2054	ds = txq->dma.dd_desc;
2055	for (i = 0; i < mwl_txbuf; i++, bf++, ds += MWL_TXDESC) {
2056		bf->bf_desc = ds;
2057		bf->bf_daddr = DS2PHYS(&txq->dma, ds);
2058		error = bus_dmamap_create(sc->sc_dmat, BUS_DMA_NOWAIT,
2059				&bf->bf_dmamap);
2060		if (error != 0) {
2061			device_printf(sc->sc_dev, "unable to create dmamap for tx "
2062				"buffer %u, error %u\n", i, error);
2063			return error;
2064		}
2065	}
2066	mwl_txq_reset(sc, txq);
2067	return 0;
2068}
2069
2070static void
2071mwl_txdma_cleanup(struct mwl_softc *sc, struct mwl_txq *txq)
2072{
2073	struct mwl_txbuf *bf;
2074	int i;
2075
2076	bf = txq->dma.dd_bufptr;
2077	for (i = 0; i < mwl_txbuf; i++, bf++) {
2078		KASSERT(bf->bf_m == NULL, ("mbuf on free list"));
2079		KASSERT(bf->bf_node == NULL, ("node on free list"));
2080		if (bf->bf_dmamap != NULL)
2081			bus_dmamap_destroy(sc->sc_dmat, bf->bf_dmamap);
2082	}
2083	STAILQ_INIT(&txq->free);
2084	txq->nfree = 0;
2085	if (txq->dma.dd_bufptr != NULL) {
2086		free(txq->dma.dd_bufptr, M_MWLDEV);
2087		txq->dma.dd_bufptr = NULL;
2088	}
2089	if (txq->dma.dd_desc_len != 0)
2090		mwl_desc_cleanup(sc, &txq->dma);
2091}
2092
2093static int
2094mwl_rxdma_setup(struct mwl_softc *sc)
2095{
2096	int error, jumbosize, bsize, i;
2097	struct mwl_rxbuf *bf;
2098	struct mwl_jumbo *rbuf;
2099	struct mwl_rxdesc *ds;
2100	caddr_t data;
2101
2102	error = mwl_desc_setup(sc, "rx", &sc->sc_rxdma,
2103			mwl_rxdesc, sizeof(struct mwl_rxbuf),
2104			1, sizeof(struct mwl_rxdesc));
2105	if (error != 0)
2106		return error;
2107
2108	/*
2109	 * Receive is done to a private pool of jumbo buffers.
2110	 * This allows us to attach to mbuf's and avoid re-mapping
2111	 * memory on each rx we post.  We allocate a large chunk
2112	 * of memory and manage it in the driver.  The mbuf free
2113	 * callback method is used to reclaim frames after sending
2114	 * them up the stack.  By default we allocate 2x the number of
2115	 * rx descriptors configured so we have some slop to hold
2116	 * us while frames are processed.
2117	 */
2118	if (mwl_rxbuf < 2*mwl_rxdesc) {
2119		device_printf(sc->sc_dev,
2120		    "too few rx dma buffers (%d); increasing to %d\n",
2121		    mwl_rxbuf, 2*mwl_rxdesc);
2122		mwl_rxbuf = 2*mwl_rxdesc;
2123	}
2124	jumbosize = roundup(MWL_AGGR_SIZE, PAGE_SIZE);
2125	sc->sc_rxmemsize = mwl_rxbuf*jumbosize;
2126
2127	error = bus_dma_tag_create(sc->sc_dmat,	/* parent */
2128		       PAGE_SIZE, 0,		/* alignment, bounds */
2129		       BUS_SPACE_MAXADDR_32BIT,	/* lowaddr */
2130		       BUS_SPACE_MAXADDR,	/* highaddr */
2131		       NULL, NULL,		/* filter, filterarg */
2132		       sc->sc_rxmemsize,	/* maxsize */
2133		       1,			/* nsegments */
2134		       sc->sc_rxmemsize,	/* maxsegsize */
2135		       BUS_DMA_ALLOCNOW,	/* flags */
2136		       NULL,			/* lockfunc */
2137		       NULL,			/* lockarg */
2138		       &sc->sc_rxdmat);
2139	if (error != 0) {
2140		device_printf(sc->sc_dev, "could not create rx DMA tag\n");
2141		return error;
2142	}
2143
2144	error = bus_dmamem_alloc(sc->sc_rxdmat, (void**) &sc->sc_rxmem,
2145				 BUS_DMA_NOWAIT | BUS_DMA_COHERENT,
2146				 &sc->sc_rxmap);
2147	if (error != 0) {
2148		device_printf(sc->sc_dev, "could not alloc %ju bytes of rx DMA memory\n",
2149		    (uintmax_t) sc->sc_rxmemsize);
2150		return error;
2151	}
2152
2153	error = bus_dmamap_load(sc->sc_rxdmat, sc->sc_rxmap,
2154				sc->sc_rxmem, sc->sc_rxmemsize,
2155				mwl_load_cb, &sc->sc_rxmem_paddr,
2156				BUS_DMA_NOWAIT);
2157	if (error != 0) {
2158		device_printf(sc->sc_dev, "could not load rx DMA map\n");
2159		return error;
2160	}
2161
2162	/*
2163	 * Allocate rx buffers and set them up.
2164	 */
2165	bsize = mwl_rxdesc * sizeof(struct mwl_rxbuf);
2166	bf = malloc(bsize, M_MWLDEV, M_NOWAIT | M_ZERO);
2167	if (bf == NULL) {
2168		device_printf(sc->sc_dev, "malloc of %u rx buffers failed\n", bsize);
2169		return error;
2170	}
2171	sc->sc_rxdma.dd_bufptr = bf;
2172
2173	STAILQ_INIT(&sc->sc_rxbuf);
2174	ds = sc->sc_rxdma.dd_desc;
2175	for (i = 0; i < mwl_rxdesc; i++, bf++, ds++) {
2176		bf->bf_desc = ds;
2177		bf->bf_daddr = DS2PHYS(&sc->sc_rxdma, ds);
2178		/* pre-assign dma buffer */
2179		bf->bf_data = ((uint8_t *)sc->sc_rxmem) + (i*jumbosize);
2180		/* NB: tail is intentional to preserve descriptor order */
2181		STAILQ_INSERT_TAIL(&sc->sc_rxbuf, bf, bf_list);
2182	}
2183
2184	/*
2185	 * Place remainder of dma memory buffers on the free list.
2186	 */
2187	SLIST_INIT(&sc->sc_rxfree);
2188	for (; i < mwl_rxbuf; i++) {
2189		data = ((uint8_t *)sc->sc_rxmem) + (i*jumbosize);
2190		rbuf = MWL_JUMBO_DATA2BUF(data);
2191		SLIST_INSERT_HEAD(&sc->sc_rxfree, rbuf, next);
2192		sc->sc_nrxfree++;
2193	}
2194	return 0;
2195}
2196#undef DS2PHYS
2197
2198static void
2199mwl_rxdma_cleanup(struct mwl_softc *sc)
2200{
2201	if (sc->sc_rxmem_paddr != 0) {
2202		bus_dmamap_unload(sc->sc_rxdmat, sc->sc_rxmap);
2203		sc->sc_rxmem_paddr = 0;
2204	}
2205	if (sc->sc_rxmem != NULL) {
2206		bus_dmamem_free(sc->sc_rxdmat, sc->sc_rxmem, sc->sc_rxmap);
2207		sc->sc_rxmem = NULL;
2208	}
2209	if (sc->sc_rxdma.dd_bufptr != NULL) {
2210		free(sc->sc_rxdma.dd_bufptr, M_MWLDEV);
2211		sc->sc_rxdma.dd_bufptr = NULL;
2212	}
2213	if (sc->sc_rxdma.dd_desc_len != 0)
2214		mwl_desc_cleanup(sc, &sc->sc_rxdma);
2215}
2216
2217static int
2218mwl_dma_setup(struct mwl_softc *sc)
2219{
2220	int error, i;
2221
2222	error = mwl_rxdma_setup(sc);
2223	if (error != 0) {
2224		mwl_rxdma_cleanup(sc);
2225		return error;
2226	}
2227
2228	for (i = 0; i < MWL_NUM_TX_QUEUES; i++) {
2229		error = mwl_txdma_setup(sc, &sc->sc_txq[i]);
2230		if (error != 0) {
2231			mwl_dma_cleanup(sc);
2232			return error;
2233		}
2234	}
2235	return 0;
2236}
2237
2238static void
2239mwl_dma_cleanup(struct mwl_softc *sc)
2240{
2241	int i;
2242
2243	for (i = 0; i < MWL_NUM_TX_QUEUES; i++)
2244		mwl_txdma_cleanup(sc, &sc->sc_txq[i]);
2245	mwl_rxdma_cleanup(sc);
2246}
2247
2248static struct ieee80211_node *
2249mwl_node_alloc(struct ieee80211vap *vap, const uint8_t mac[IEEE80211_ADDR_LEN])
2250{
2251	struct ieee80211com *ic = vap->iv_ic;
2252	struct mwl_softc *sc = ic->ic_softc;
2253	const size_t space = sizeof(struct mwl_node);
2254	struct mwl_node *mn;
2255
2256	mn = malloc(space, M_80211_NODE, M_NOWAIT|M_ZERO);
2257	if (mn == NULL) {
2258		/* XXX stat+msg */
2259		return NULL;
2260	}
2261	DPRINTF(sc, MWL_DEBUG_NODE, "%s: mn %p\n", __func__, mn);
2262	return &mn->mn_node;
2263}
2264
2265static void
2266mwl_node_cleanup(struct ieee80211_node *ni)
2267{
2268	struct ieee80211com *ic = ni->ni_ic;
2269        struct mwl_softc *sc = ic->ic_softc;
2270	struct mwl_node *mn = MWL_NODE(ni);
2271
2272	DPRINTF(sc, MWL_DEBUG_NODE, "%s: ni %p ic %p staid %d\n",
2273	    __func__, ni, ni->ni_ic, mn->mn_staid);
2274
2275	if (mn->mn_staid != 0) {
2276		struct ieee80211vap *vap = ni->ni_vap;
2277
2278		if (mn->mn_hvap != NULL) {
2279			if (vap->iv_opmode == IEEE80211_M_STA)
2280				mwl_hal_delstation(mn->mn_hvap, vap->iv_myaddr);
2281			else
2282				mwl_hal_delstation(mn->mn_hvap, ni->ni_macaddr);
2283		}
2284		/*
2285		 * NB: legacy WDS peer sta db entry is installed using
2286		 * the associate ap's hvap; use it again to delete it.
2287		 * XXX can vap be NULL?
2288		 */
2289		else if (vap->iv_opmode == IEEE80211_M_WDS &&
2290		    MWL_VAP(vap)->mv_ap_hvap != NULL)
2291			mwl_hal_delstation(MWL_VAP(vap)->mv_ap_hvap,
2292			    ni->ni_macaddr);
2293		delstaid(sc, mn->mn_staid);
2294		mn->mn_staid = 0;
2295	}
2296	sc->sc_node_cleanup(ni);
2297}
2298
2299/*
2300 * Reclaim rx dma buffers from packets sitting on the ampdu
2301 * reorder queue for a station.  We replace buffers with a
2302 * system cluster (if available).
2303 */
2304static void
2305mwl_ampdu_rxdma_reclaim(struct ieee80211_rx_ampdu *rap)
2306{
2307#if 0
2308	int i, n, off;
2309	struct mbuf *m;
2310	void *cl;
2311
2312	n = rap->rxa_qframes;
2313	for (i = 0; i < rap->rxa_wnd && n > 0; i++) {
2314		m = rap->rxa_m[i];
2315		if (m == NULL)
2316			continue;
2317		n--;
2318		/* our dma buffers have a well-known free routine */
2319		if ((m->m_flags & M_EXT) == 0 ||
2320		    m->m_ext.ext_free != mwl_ext_free)
2321			continue;
2322		/*
2323		 * Try to allocate a cluster and move the data.
2324		 */
2325		off = m->m_data - m->m_ext.ext_buf;
2326		if (off + m->m_pkthdr.len > MCLBYTES) {
2327			/* XXX no AMSDU for now */
2328			continue;
2329		}
2330		cl = pool_cache_get_paddr(&mclpool_cache, 0,
2331		    &m->m_ext.ext_paddr);
2332		if (cl != NULL) {
2333			/*
2334			 * Copy the existing data to the cluster, remove
2335			 * the rx dma buffer, and attach the cluster in
2336			 * its place.  Note we preserve the offset to the
2337			 * data so frames being bridged can still prepend
2338			 * their headers without adding another mbuf.
2339			 */
2340			memcpy((caddr_t) cl + off, m->m_data, m->m_pkthdr.len);
2341			MEXTREMOVE(m);
2342			MEXTADD(m, cl, MCLBYTES, 0, NULL, &mclpool_cache);
2343			/* setup mbuf like _MCLGET does */
2344			m->m_flags |= M_CLUSTER | M_EXT_RW;
2345			_MOWNERREF(m, M_EXT | M_CLUSTER);
2346			/* NB: m_data is clobbered by MEXTADDR, adjust */
2347			m->m_data += off;
2348		}
2349	}
2350#endif
2351}
2352
2353/*
2354 * Callback to reclaim resources.  We first let the
2355 * net80211 layer do it's thing, then if we are still
2356 * blocked by a lack of rx dma buffers we walk the ampdu
2357 * reorder q's to reclaim buffers by copying to a system
2358 * cluster.
2359 */
2360static void
2361mwl_node_drain(struct ieee80211_node *ni)
2362{
2363	struct ieee80211com *ic = ni->ni_ic;
2364        struct mwl_softc *sc = ic->ic_softc;
2365	struct mwl_node *mn = MWL_NODE(ni);
2366
2367	DPRINTF(sc, MWL_DEBUG_NODE, "%s: ni %p vap %p staid %d\n",
2368	    __func__, ni, ni->ni_vap, mn->mn_staid);
2369
2370	/* NB: call up first to age out ampdu q's */
2371	sc->sc_node_drain(ni);
2372
2373	/* XXX better to not check low water mark? */
2374	if (sc->sc_rxblocked && mn->mn_staid != 0 &&
2375	    (ni->ni_flags & IEEE80211_NODE_HT)) {
2376		uint8_t tid;
2377		/*
2378		 * Walk the reorder q and reclaim rx dma buffers by copying
2379		 * the packet contents into clusters.
2380		 */
2381		for (tid = 0; tid < WME_NUM_TID; tid++) {
2382			struct ieee80211_rx_ampdu *rap;
2383
2384			rap = &ni->ni_rx_ampdu[tid];
2385			if ((rap->rxa_flags & IEEE80211_AGGR_XCHGPEND) == 0)
2386				continue;
2387			if (rap->rxa_qframes)
2388				mwl_ampdu_rxdma_reclaim(rap);
2389		}
2390	}
2391}
2392
2393static void
2394mwl_node_getsignal(const struct ieee80211_node *ni, int8_t *rssi, int8_t *noise)
2395{
2396	*rssi = ni->ni_ic->ic_node_getrssi(ni);
2397#ifdef MWL_ANT_INFO_SUPPORT
2398#if 0
2399	/* XXX need to smooth data */
2400	*noise = -MWL_NODE_CONST(ni)->mn_ai.nf;
2401#else
2402	*noise = -95;		/* XXX */
2403#endif
2404#else
2405	*noise = -95;		/* XXX */
2406#endif
2407}
2408
2409/*
2410 * Convert Hardware per-antenna rssi info to common format:
2411 * Let a1, a2, a3 represent the amplitudes per chain
2412 * Let amax represent max[a1, a2, a3]
2413 * Rssi1_dBm = RSSI_dBm + 20*log10(a1/amax)
2414 * Rssi1_dBm = RSSI_dBm + 20*log10(a1) - 20*log10(amax)
2415 * We store a table that is 4*20*log10(idx) - the extra 4 is to store or
2416 * maintain some extra precision.
2417 *
2418 * Values are stored in .5 db format capped at 127.
2419 */
2420static void
2421mwl_node_getmimoinfo(const struct ieee80211_node *ni,
2422	struct ieee80211_mimo_info *mi)
2423{
2424#define	CVT(_dst, _src) do {						\
2425	(_dst) = rssi + ((logdbtbl[_src] - logdbtbl[rssi_max]) >> 2);	\
2426	(_dst) = (_dst) > 64 ? 127 : ((_dst) << 1);			\
2427} while (0)
2428	static const int8_t logdbtbl[32] = {
2429	       0,   0,  24,  38,  48,  56,  62,  68,
2430	      72,  76,  80,  83,  86,  89,  92,  94,
2431	      96,  98, 100, 102, 104, 106, 107, 109,
2432	     110, 112, 113, 115, 116, 117, 118, 119
2433	};
2434	const struct mwl_node *mn = MWL_NODE_CONST(ni);
2435	uint8_t rssi = mn->mn_ai.rsvd1/2;		/* XXX */
2436	uint32_t rssi_max;
2437
2438	rssi_max = mn->mn_ai.rssi_a;
2439	if (mn->mn_ai.rssi_b > rssi_max)
2440		rssi_max = mn->mn_ai.rssi_b;
2441	if (mn->mn_ai.rssi_c > rssi_max)
2442		rssi_max = mn->mn_ai.rssi_c;
2443
2444	CVT(mi->rssi[0], mn->mn_ai.rssi_a);
2445	CVT(mi->rssi[1], mn->mn_ai.rssi_b);
2446	CVT(mi->rssi[2], mn->mn_ai.rssi_c);
2447
2448	mi->noise[0] = mn->mn_ai.nf_a;
2449	mi->noise[1] = mn->mn_ai.nf_b;
2450	mi->noise[2] = mn->mn_ai.nf_c;
2451#undef CVT
2452}
2453
2454static __inline void *
2455mwl_getrxdma(struct mwl_softc *sc)
2456{
2457	struct mwl_jumbo *buf;
2458	void *data;
2459
2460	/*
2461	 * Allocate from jumbo pool.
2462	 */
2463	MWL_RXFREE_LOCK(sc);
2464	buf = SLIST_FIRST(&sc->sc_rxfree);
2465	if (buf == NULL) {
2466		DPRINTF(sc, MWL_DEBUG_ANY,
2467		    "%s: out of rx dma buffers\n", __func__);
2468		sc->sc_stats.mst_rx_nodmabuf++;
2469		data = NULL;
2470	} else {
2471		SLIST_REMOVE_HEAD(&sc->sc_rxfree, next);
2472		sc->sc_nrxfree--;
2473		data = MWL_JUMBO_BUF2DATA(buf);
2474	}
2475	MWL_RXFREE_UNLOCK(sc);
2476	return data;
2477}
2478
2479static __inline void
2480mwl_putrxdma(struct mwl_softc *sc, void *data)
2481{
2482	struct mwl_jumbo *buf;
2483
2484	/* XXX bounds check data */
2485	MWL_RXFREE_LOCK(sc);
2486	buf = MWL_JUMBO_DATA2BUF(data);
2487	SLIST_INSERT_HEAD(&sc->sc_rxfree, buf, next);
2488	sc->sc_nrxfree++;
2489	MWL_RXFREE_UNLOCK(sc);
2490}
2491
2492static int
2493mwl_rxbuf_init(struct mwl_softc *sc, struct mwl_rxbuf *bf)
2494{
2495	struct mwl_rxdesc *ds;
2496
2497	ds = bf->bf_desc;
2498	if (bf->bf_data == NULL) {
2499		bf->bf_data = mwl_getrxdma(sc);
2500		if (bf->bf_data == NULL) {
2501			/* mark descriptor to be skipped */
2502			ds->RxControl = EAGLE_RXD_CTRL_OS_OWN;
2503			/* NB: don't need PREREAD */
2504			MWL_RXDESC_SYNC(sc, ds, BUS_DMASYNC_PREWRITE);
2505			sc->sc_stats.mst_rxbuf_failed++;
2506			return ENOMEM;
2507		}
2508	}
2509	/*
2510	 * NB: DMA buffer contents is known to be unmodified
2511	 *     so there's no need to flush the data cache.
2512	 */
2513
2514	/*
2515	 * Setup descriptor.
2516	 */
2517	ds->QosCtrl = 0;
2518	ds->RSSI = 0;
2519	ds->Status = EAGLE_RXD_STATUS_IDLE;
2520	ds->Channel = 0;
2521	ds->PktLen = htole16(MWL_AGGR_SIZE);
2522	ds->SQ2 = 0;
2523	ds->pPhysBuffData = htole32(MWL_JUMBO_DMA_ADDR(sc, bf->bf_data));
2524	/* NB: don't touch pPhysNext, set once */
2525	ds->RxControl = EAGLE_RXD_CTRL_DRIVER_OWN;
2526	MWL_RXDESC_SYNC(sc, ds, BUS_DMASYNC_PREREAD | BUS_DMASYNC_PREWRITE);
2527
2528	return 0;
2529}
2530
2531static void
2532mwl_ext_free(struct mbuf *m, void *data, void *arg)
2533{
2534	struct mwl_softc *sc = arg;
2535
2536	/* XXX bounds check data */
2537	mwl_putrxdma(sc, data);
2538	/*
2539	 * If we were previously blocked by a lack of rx dma buffers
2540	 * check if we now have enough to restart rx interrupt handling.
2541	 * NB: we know we are called at splvm which is above splnet.
2542	 */
2543	if (sc->sc_rxblocked && sc->sc_nrxfree > mwl_rxdmalow) {
2544		sc->sc_rxblocked = 0;
2545		mwl_hal_intrset(sc->sc_mh, sc->sc_imask);
2546	}
2547}
2548
2549struct mwl_frame_bar {
2550	u_int8_t	i_fc[2];
2551	u_int8_t	i_dur[2];
2552	u_int8_t	i_ra[IEEE80211_ADDR_LEN];
2553	u_int8_t	i_ta[IEEE80211_ADDR_LEN];
2554	/* ctl, seq, FCS */
2555} __packed;
2556
2557/*
2558 * Like ieee80211_anyhdrsize, but handles BAR frames
2559 * specially so the logic below to piece the 802.11
2560 * header together works.
2561 */
2562static __inline int
2563mwl_anyhdrsize(const void *data)
2564{
2565	const struct ieee80211_frame *wh = data;
2566
2567	if ((wh->i_fc[0]&IEEE80211_FC0_TYPE_MASK) == IEEE80211_FC0_TYPE_CTL) {
2568		switch (wh->i_fc[0] & IEEE80211_FC0_SUBTYPE_MASK) {
2569		case IEEE80211_FC0_SUBTYPE_CTS:
2570		case IEEE80211_FC0_SUBTYPE_ACK:
2571			return sizeof(struct ieee80211_frame_ack);
2572		case IEEE80211_FC0_SUBTYPE_BAR:
2573			return sizeof(struct mwl_frame_bar);
2574		}
2575		return sizeof(struct ieee80211_frame_min);
2576	} else
2577		return ieee80211_hdrsize(data);
2578}
2579
2580static void
2581mwl_handlemicerror(struct ieee80211com *ic, const uint8_t *data)
2582{
2583	const struct ieee80211_frame *wh;
2584	struct ieee80211_node *ni;
2585
2586	wh = (const struct ieee80211_frame *)(data + sizeof(uint16_t));
2587	ni = ieee80211_find_rxnode(ic, (const struct ieee80211_frame_min *) wh);
2588	if (ni != NULL) {
2589		ieee80211_notify_michael_failure(ni->ni_vap, wh, 0);
2590		ieee80211_free_node(ni);
2591	}
2592}
2593
2594/*
2595 * Convert hardware signal strength to rssi.  The value
2596 * provided by the device has the noise floor added in;
2597 * we need to compensate for this but we don't have that
2598 * so we use a fixed value.
2599 *
2600 * The offset of 8 is good for both 2.4 and 5GHz.  The LNA
2601 * offset is already set as part of the initial gain.  This
2602 * will give at least +/- 3dB for 2.4GHz and +/- 5dB for 5GHz.
2603 */
2604static __inline int
2605cvtrssi(uint8_t ssi)
2606{
2607	int rssi = (int) ssi + 8;
2608	/* XXX hack guess until we have a real noise floor */
2609	rssi = 2*(87 - rssi);	/* NB: .5 dBm units */
2610	return (rssi < 0 ? 0 : rssi > 127 ? 127 : rssi);
2611}
2612
2613static void
2614mwl_rx_proc(void *arg, int npending)
2615{
2616	struct mwl_softc *sc = arg;
2617	struct ieee80211com *ic = &sc->sc_ic;
2618	struct mwl_rxbuf *bf;
2619	struct mwl_rxdesc *ds;
2620	struct mbuf *m;
2621	struct ieee80211_qosframe *wh;
2622	struct ieee80211_qosframe_addr4 *wh4;
2623	struct ieee80211_node *ni;
2624	struct mwl_node *mn;
2625	int off, len, hdrlen, pktlen, rssi, ntodo;
2626	uint8_t *data, status;
2627	void *newdata;
2628	int16_t nf;
2629
2630	DPRINTF(sc, MWL_DEBUG_RX_PROC, "%s: pending %u rdptr 0x%x wrptr 0x%x\n",
2631	    __func__, npending, RD4(sc, sc->sc_hwspecs.rxDescRead),
2632	    RD4(sc, sc->sc_hwspecs.rxDescWrite));
2633	nf = -96;			/* XXX */
2634	bf = sc->sc_rxnext;
2635	for (ntodo = mwl_rxquota; ntodo > 0; ntodo--) {
2636		if (bf == NULL)
2637			bf = STAILQ_FIRST(&sc->sc_rxbuf);
2638		ds = bf->bf_desc;
2639		data = bf->bf_data;
2640		if (data == NULL) {
2641			/*
2642			 * If data allocation failed previously there
2643			 * will be no buffer; try again to re-populate it.
2644			 * Note the firmware will not advance to the next
2645			 * descriptor with a dma buffer so we must mimic
2646			 * this or we'll get out of sync.
2647			 */
2648			DPRINTF(sc, MWL_DEBUG_ANY,
2649			    "%s: rx buf w/o dma memory\n", __func__);
2650			(void) mwl_rxbuf_init(sc, bf);
2651			sc->sc_stats.mst_rx_dmabufmissing++;
2652			break;
2653		}
2654		MWL_RXDESC_SYNC(sc, ds,
2655		    BUS_DMASYNC_POSTREAD | BUS_DMASYNC_POSTWRITE);
2656		if (ds->RxControl != EAGLE_RXD_CTRL_DMA_OWN)
2657			break;
2658#ifdef MWL_DEBUG
2659		if (sc->sc_debug & MWL_DEBUG_RECV_DESC)
2660			mwl_printrxbuf(bf, 0);
2661#endif
2662		status = ds->Status;
2663		if (status & EAGLE_RXD_STATUS_DECRYPT_ERR_MASK) {
2664			counter_u64_add(ic->ic_ierrors, 1);
2665			sc->sc_stats.mst_rx_crypto++;
2666			/*
2667			 * NB: Check EAGLE_RXD_STATUS_GENERAL_DECRYPT_ERR
2668			 *     for backwards compatibility.
2669			 */
2670			if (status != EAGLE_RXD_STATUS_GENERAL_DECRYPT_ERR &&
2671			    (status & EAGLE_RXD_STATUS_TKIP_MIC_DECRYPT_ERR)) {
2672				/*
2673				 * MIC error, notify upper layers.
2674				 */
2675				bus_dmamap_sync(sc->sc_rxdmat, sc->sc_rxmap,
2676				    BUS_DMASYNC_POSTREAD);
2677				mwl_handlemicerror(ic, data);
2678				sc->sc_stats.mst_rx_tkipmic++;
2679			}
2680			/* XXX too painful to tap packets */
2681			goto rx_next;
2682		}
2683		/*
2684		 * Sync the data buffer.
2685		 */
2686		len = le16toh(ds->PktLen);
2687		bus_dmamap_sync(sc->sc_rxdmat, sc->sc_rxmap, BUS_DMASYNC_POSTREAD);
2688		/*
2689		 * The 802.11 header is provided all or in part at the front;
2690		 * use it to calculate the true size of the header that we'll
2691		 * construct below.  We use this to figure out where to copy
2692		 * payload prior to constructing the header.
2693		 */
2694		hdrlen = mwl_anyhdrsize(data + sizeof(uint16_t));
2695		off = sizeof(uint16_t) + sizeof(struct ieee80211_frame_addr4);
2696
2697		/* calculate rssi early so we can re-use for each aggregate */
2698		rssi = cvtrssi(ds->RSSI);
2699
2700		pktlen = hdrlen + (len - off);
2701		/*
2702		 * NB: we know our frame is at least as large as
2703		 * IEEE80211_MIN_LEN because there is a 4-address
2704		 * frame at the front.  Hence there's no need to
2705		 * vet the packet length.  If the frame in fact
2706		 * is too small it should be discarded at the
2707		 * net80211 layer.
2708		 */
2709
2710		/*
2711		 * Attach dma buffer to an mbuf.  We tried
2712		 * doing this based on the packet size (i.e.
2713		 * copying small packets) but it turns out to
2714		 * be a net loss.  The tradeoff might be system
2715		 * dependent (cache architecture is important).
2716		 */
2717		MGETHDR(m, M_NOWAIT, MT_DATA);
2718		if (m == NULL) {
2719			DPRINTF(sc, MWL_DEBUG_ANY,
2720			    "%s: no rx mbuf\n", __func__);
2721			sc->sc_stats.mst_rx_nombuf++;
2722			goto rx_next;
2723		}
2724		/*
2725		 * Acquire the replacement dma buffer before
2726		 * processing the frame.  If we're out of dma
2727		 * buffers we disable rx interrupts and wait
2728		 * for the free pool to reach mlw_rxdmalow buffers
2729		 * before starting to do work again.  If the firmware
2730		 * runs out of descriptors then it will toss frames
2731		 * which is better than our doing it as that can
2732		 * starve our processing.  It is also important that
2733		 * we always process rx'd frames in case they are
2734		 * A-MPDU as otherwise the host's view of the BA
2735		 * window may get out of sync with the firmware.
2736		 */
2737		newdata = mwl_getrxdma(sc);
2738		if (newdata == NULL) {
2739			/* NB: stat+msg in mwl_getrxdma */
2740			m_free(m);
2741			/* disable RX interrupt and mark state */
2742			mwl_hal_intrset(sc->sc_mh,
2743			    sc->sc_imask &~ MACREG_A2HRIC_BIT_RX_RDY);
2744			sc->sc_rxblocked = 1;
2745			ieee80211_drain(ic);
2746			/* XXX check rxblocked and immediately start again? */
2747			goto rx_stop;
2748		}
2749		bf->bf_data = newdata;
2750		/*
2751		 * Attach the dma buffer to the mbuf;
2752		 * mwl_rxbuf_init will re-setup the rx
2753		 * descriptor using the replacement dma
2754		 * buffer we just installed above.
2755		 */
2756		MEXTADD(m, data, MWL_AGGR_SIZE, mwl_ext_free,
2757		    data, sc, 0, EXT_NET_DRV);
2758		m->m_data += off - hdrlen;
2759		m->m_pkthdr.len = m->m_len = pktlen;
2760		/* NB: dma buffer assumed read-only */
2761
2762		/*
2763		 * Piece 802.11 header together.
2764		 */
2765		wh = mtod(m, struct ieee80211_qosframe *);
2766		/* NB: don't need to do this sometimes but ... */
2767		/* XXX special case so we can memcpy after m_devget? */
2768		ovbcopy(data + sizeof(uint16_t), wh, hdrlen);
2769		if (IEEE80211_QOS_HAS_SEQ(wh)) {
2770			if (IEEE80211_IS_DSTODS(wh)) {
2771				wh4 = mtod(m,
2772				    struct ieee80211_qosframe_addr4*);
2773				*(uint16_t *)wh4->i_qos = ds->QosCtrl;
2774			} else {
2775				*(uint16_t *)wh->i_qos = ds->QosCtrl;
2776			}
2777		}
2778		/*
2779		 * The f/w strips WEP header but doesn't clear
2780		 * the WEP bit; mark the packet with M_WEP so
2781		 * net80211 will treat the data as decrypted.
2782		 * While here also clear the PWR_MGT bit since
2783		 * power save is handled by the firmware and
2784		 * passing this up will potentially cause the
2785		 * upper layer to put a station in power save
2786		 * (except when configured with MWL_HOST_PS_SUPPORT).
2787		 */
2788		if (wh->i_fc[1] & IEEE80211_FC1_PROTECTED)
2789			m->m_flags |= M_WEP;
2790#ifdef MWL_HOST_PS_SUPPORT
2791		wh->i_fc[1] &= ~IEEE80211_FC1_PROTECTED;
2792#else
2793		wh->i_fc[1] &= ~(IEEE80211_FC1_PROTECTED |
2794		    IEEE80211_FC1_PWR_MGT);
2795#endif
2796
2797		if (ieee80211_radiotap_active(ic)) {
2798			struct mwl_rx_radiotap_header *tap = &sc->sc_rx_th;
2799
2800			tap->wr_flags = 0;
2801			tap->wr_rate = ds->Rate;
2802			tap->wr_antsignal = rssi + nf;
2803			tap->wr_antnoise = nf;
2804		}
2805		if (IFF_DUMPPKTS_RECV(sc, wh)) {
2806			ieee80211_dump_pkt(ic, mtod(m, caddr_t),
2807			    len, ds->Rate, rssi);
2808		}
2809		/* dispatch */
2810		ni = ieee80211_find_rxnode(ic,
2811		    (const struct ieee80211_frame_min *) wh);
2812		if (ni != NULL) {
2813			mn = MWL_NODE(ni);
2814#ifdef MWL_ANT_INFO_SUPPORT
2815			mn->mn_ai.rssi_a = ds->ai.rssi_a;
2816			mn->mn_ai.rssi_b = ds->ai.rssi_b;
2817			mn->mn_ai.rssi_c = ds->ai.rssi_c;
2818			mn->mn_ai.rsvd1 = rssi;
2819#endif
2820			/* tag AMPDU aggregates for reorder processing */
2821			if (ni->ni_flags & IEEE80211_NODE_HT)
2822				m->m_flags |= M_AMPDU;
2823			(void) ieee80211_input(ni, m, rssi, nf);
2824			ieee80211_free_node(ni);
2825		} else
2826			(void) ieee80211_input_all(ic, m, rssi, nf);
2827rx_next:
2828		/* NB: ignore ENOMEM so we process more descriptors */
2829		(void) mwl_rxbuf_init(sc, bf);
2830		bf = STAILQ_NEXT(bf, bf_list);
2831	}
2832rx_stop:
2833	sc->sc_rxnext = bf;
2834
2835	if (mbufq_first(&sc->sc_snd) != NULL) {
2836		/* NB: kick fw; the tx thread may have been preempted */
2837		mwl_hal_txstart(sc->sc_mh, 0);
2838		mwl_start(sc);
2839	}
2840}
2841
2842static void
2843mwl_txq_init(struct mwl_softc *sc, struct mwl_txq *txq, int qnum)
2844{
2845	struct mwl_txbuf *bf, *bn;
2846	struct mwl_txdesc *ds;
2847
2848	MWL_TXQ_LOCK_INIT(sc, txq);
2849	txq->qnum = qnum;
2850	txq->txpri = 0;	/* XXX */
2851#if 0
2852	/* NB: q setup by mwl_txdma_setup XXX */
2853	STAILQ_INIT(&txq->free);
2854#endif
2855	STAILQ_FOREACH(bf, &txq->free, bf_list) {
2856		bf->bf_txq = txq;
2857
2858		ds = bf->bf_desc;
2859		bn = STAILQ_NEXT(bf, bf_list);
2860		if (bn == NULL)
2861			bn = STAILQ_FIRST(&txq->free);
2862		ds->pPhysNext = htole32(bn->bf_daddr);
2863	}
2864	STAILQ_INIT(&txq->active);
2865}
2866
2867/*
2868 * Setup a hardware data transmit queue for the specified
2869 * access control.  We record the mapping from ac's
2870 * to h/w queues for use by mwl_tx_start.
2871 */
2872static int
2873mwl_tx_setup(struct mwl_softc *sc, int ac, int mvtype)
2874{
2875	struct mwl_txq *txq;
2876
2877	if (ac >= nitems(sc->sc_ac2q)) {
2878		device_printf(sc->sc_dev, "AC %u out of range, max %zu!\n",
2879			ac, nitems(sc->sc_ac2q));
2880		return 0;
2881	}
2882	if (mvtype >= MWL_NUM_TX_QUEUES) {
2883		device_printf(sc->sc_dev, "mvtype %u out of range, max %u!\n",
2884			mvtype, MWL_NUM_TX_QUEUES);
2885		return 0;
2886	}
2887	txq = &sc->sc_txq[mvtype];
2888	mwl_txq_init(sc, txq, mvtype);
2889	sc->sc_ac2q[ac] = txq;
2890	return 1;
2891}
2892
2893/*
2894 * Update WME parameters for a transmit queue.
2895 */
2896static int
2897mwl_txq_update(struct mwl_softc *sc, int ac)
2898{
2899#define	MWL_EXPONENT_TO_VALUE(v)	((1<<v)-1)
2900	struct ieee80211com *ic = &sc->sc_ic;
2901	struct mwl_txq *txq = sc->sc_ac2q[ac];
2902	struct wmeParams *wmep = &ic->ic_wme.wme_chanParams.cap_wmeParams[ac];
2903	struct mwl_hal *mh = sc->sc_mh;
2904	int aifs, cwmin, cwmax, txoplim;
2905
2906	aifs = wmep->wmep_aifsn;
2907	/* XXX in sta mode need to pass log values for cwmin/max */
2908	cwmin = MWL_EXPONENT_TO_VALUE(wmep->wmep_logcwmin);
2909	cwmax = MWL_EXPONENT_TO_VALUE(wmep->wmep_logcwmax);
2910	txoplim = wmep->wmep_txopLimit;		/* NB: units of 32us */
2911
2912	if (mwl_hal_setedcaparams(mh, txq->qnum, cwmin, cwmax, aifs, txoplim)) {
2913		device_printf(sc->sc_dev, "unable to update hardware queue "
2914			"parameters for %s traffic!\n",
2915			ieee80211_wme_acnames[ac]);
2916		return 0;
2917	}
2918	return 1;
2919#undef MWL_EXPONENT_TO_VALUE
2920}
2921
2922/*
2923 * Callback from the 802.11 layer to update WME parameters.
2924 */
2925static int
2926mwl_wme_update(struct ieee80211com *ic)
2927{
2928	struct mwl_softc *sc = ic->ic_softc;
2929
2930	return !mwl_txq_update(sc, WME_AC_BE) ||
2931	    !mwl_txq_update(sc, WME_AC_BK) ||
2932	    !mwl_txq_update(sc, WME_AC_VI) ||
2933	    !mwl_txq_update(sc, WME_AC_VO) ? EIO : 0;
2934}
2935
2936/*
2937 * Reclaim resources for a setup queue.
2938 */
2939static void
2940mwl_tx_cleanupq(struct mwl_softc *sc, struct mwl_txq *txq)
2941{
2942	/* XXX hal work? */
2943	MWL_TXQ_LOCK_DESTROY(txq);
2944}
2945
2946/*
2947 * Reclaim all tx queue resources.
2948 */
2949static void
2950mwl_tx_cleanup(struct mwl_softc *sc)
2951{
2952	int i;
2953
2954	for (i = 0; i < MWL_NUM_TX_QUEUES; i++)
2955		mwl_tx_cleanupq(sc, &sc->sc_txq[i]);
2956}
2957
2958static int
2959mwl_tx_dmasetup(struct mwl_softc *sc, struct mwl_txbuf *bf, struct mbuf *m0)
2960{
2961	struct mbuf *m;
2962	int error;
2963
2964	/*
2965	 * Load the DMA map so any coalescing is done.  This
2966	 * also calculates the number of descriptors we need.
2967	 */
2968	error = bus_dmamap_load_mbuf_sg(sc->sc_dmat, bf->bf_dmamap, m0,
2969				     bf->bf_segs, &bf->bf_nseg,
2970				     BUS_DMA_NOWAIT);
2971	if (error == EFBIG) {
2972		/* XXX packet requires too many descriptors */
2973		bf->bf_nseg = MWL_TXDESC+1;
2974	} else if (error != 0) {
2975		sc->sc_stats.mst_tx_busdma++;
2976		m_freem(m0);
2977		return error;
2978	}
2979	/*
2980	 * Discard null packets and check for packets that
2981	 * require too many TX descriptors.  We try to convert
2982	 * the latter to a cluster.
2983	 */
2984	if (error == EFBIG) {		/* too many desc's, linearize */
2985		sc->sc_stats.mst_tx_linear++;
2986#if MWL_TXDESC > 1
2987		m = m_collapse(m0, M_NOWAIT, MWL_TXDESC);
2988#else
2989		m = m_defrag(m0, M_NOWAIT);
2990#endif
2991		if (m == NULL) {
2992			m_freem(m0);
2993			sc->sc_stats.mst_tx_nombuf++;
2994			return ENOMEM;
2995		}
2996		m0 = m;
2997		error = bus_dmamap_load_mbuf_sg(sc->sc_dmat, bf->bf_dmamap, m0,
2998					     bf->bf_segs, &bf->bf_nseg,
2999					     BUS_DMA_NOWAIT);
3000		if (error != 0) {
3001			sc->sc_stats.mst_tx_busdma++;
3002			m_freem(m0);
3003			return error;
3004		}
3005		KASSERT(bf->bf_nseg <= MWL_TXDESC,
3006		    ("too many segments after defrag; nseg %u", bf->bf_nseg));
3007	} else if (bf->bf_nseg == 0) {		/* null packet, discard */
3008		sc->sc_stats.mst_tx_nodata++;
3009		m_freem(m0);
3010		return EIO;
3011	}
3012	DPRINTF(sc, MWL_DEBUG_XMIT, "%s: m %p len %u\n",
3013		__func__, m0, m0->m_pkthdr.len);
3014	bus_dmamap_sync(sc->sc_dmat, bf->bf_dmamap, BUS_DMASYNC_PREWRITE);
3015	bf->bf_m = m0;
3016
3017	return 0;
3018}
3019
3020static __inline int
3021mwl_cvtlegacyrate(int rate)
3022{
3023	switch (rate) {
3024	case 2:	 return 0;
3025	case 4:	 return 1;
3026	case 11: return 2;
3027	case 22: return 3;
3028	case 44: return 4;
3029	case 12: return 5;
3030	case 18: return 6;
3031	case 24: return 7;
3032	case 36: return 8;
3033	case 48: return 9;
3034	case 72: return 10;
3035	case 96: return 11;
3036	case 108:return 12;
3037	}
3038	return 0;
3039}
3040
3041/*
3042 * Calculate fixed tx rate information per client state;
3043 * this value is suitable for writing to the Format field
3044 * of a tx descriptor.
3045 */
3046static uint16_t
3047mwl_calcformat(uint8_t rate, const struct ieee80211_node *ni)
3048{
3049	uint16_t fmt;
3050
3051	fmt = SM(3, EAGLE_TXD_ANTENNA)
3052	    | (IEEE80211_IS_CHAN_HT40D(ni->ni_chan) ?
3053		EAGLE_TXD_EXTCHAN_LO : EAGLE_TXD_EXTCHAN_HI);
3054	if (rate & IEEE80211_RATE_MCS) {	/* HT MCS */
3055		fmt |= EAGLE_TXD_FORMAT_HT
3056		    /* NB: 0x80 implicitly stripped from ucastrate */
3057		    | SM(rate, EAGLE_TXD_RATE);
3058		/* XXX short/long GI may be wrong; re-check */
3059		if (IEEE80211_IS_CHAN_HT40(ni->ni_chan)) {
3060			fmt |= EAGLE_TXD_CHW_40
3061			    | (ni->ni_htcap & IEEE80211_HTCAP_SHORTGI40 ?
3062			        EAGLE_TXD_GI_SHORT : EAGLE_TXD_GI_LONG);
3063		} else {
3064			fmt |= EAGLE_TXD_CHW_20
3065			    | (ni->ni_htcap & IEEE80211_HTCAP_SHORTGI20 ?
3066			        EAGLE_TXD_GI_SHORT : EAGLE_TXD_GI_LONG);
3067		}
3068	} else {			/* legacy rate */
3069		fmt |= EAGLE_TXD_FORMAT_LEGACY
3070		    | SM(mwl_cvtlegacyrate(rate), EAGLE_TXD_RATE)
3071		    | EAGLE_TXD_CHW_20
3072		    /* XXX iv_flags & IEEE80211_F_SHPREAMBLE? */
3073		    | (ni->ni_capinfo & IEEE80211_CAPINFO_SHORT_PREAMBLE ?
3074			EAGLE_TXD_PREAMBLE_SHORT : EAGLE_TXD_PREAMBLE_LONG);
3075	}
3076	return fmt;
3077}
3078
3079static int
3080mwl_tx_start(struct mwl_softc *sc, struct ieee80211_node *ni, struct mwl_txbuf *bf,
3081    struct mbuf *m0)
3082{
3083	struct ieee80211com *ic = &sc->sc_ic;
3084	struct ieee80211vap *vap = ni->ni_vap;
3085	int error, iswep, ismcast;
3086	int hdrlen, copyhdrlen, pktlen;
3087	struct mwl_txdesc *ds;
3088	struct mwl_txq *txq;
3089	struct ieee80211_frame *wh;
3090	struct mwltxrec *tr;
3091	struct mwl_node *mn;
3092	uint16_t qos;
3093#if MWL_TXDESC > 1
3094	int i;
3095#endif
3096
3097	wh = mtod(m0, struct ieee80211_frame *);
3098	iswep = wh->i_fc[1] & IEEE80211_FC1_PROTECTED;
3099	ismcast = IEEE80211_IS_MULTICAST(wh->i_addr1);
3100	hdrlen = ieee80211_anyhdrsize(wh);
3101	copyhdrlen = hdrlen;
3102	pktlen = m0->m_pkthdr.len;
3103	if (IEEE80211_QOS_HAS_SEQ(wh)) {
3104		if (IEEE80211_IS_DSTODS(wh)) {
3105			qos = *(uint16_t *)
3106			    (((struct ieee80211_qosframe_addr4 *) wh)->i_qos);
3107			copyhdrlen -= sizeof(qos);
3108		} else
3109			qos = *(uint16_t *)
3110			    (((struct ieee80211_qosframe *) wh)->i_qos);
3111	} else
3112		qos = 0;
3113
3114	if (iswep) {
3115		const struct ieee80211_cipher *cip;
3116		struct ieee80211_key *k;
3117
3118		/*
3119		 * Construct the 802.11 header+trailer for an encrypted
3120		 * frame. The only reason this can fail is because of an
3121		 * unknown or unsupported cipher/key type.
3122		 *
3123		 * NB: we do this even though the firmware will ignore
3124		 *     what we've done for WEP and TKIP as we need the
3125		 *     ExtIV filled in for CCMP and this also adjusts
3126		 *     the headers which simplifies our work below.
3127		 */
3128		k = ieee80211_crypto_encap(ni, m0);
3129		if (k == NULL) {
3130			/*
3131			 * This can happen when the key is yanked after the
3132			 * frame was queued.  Just discard the frame; the
3133			 * 802.11 layer counts failures and provides
3134			 * debugging/diagnostics.
3135			 */
3136			m_freem(m0);
3137			return EIO;
3138		}
3139		/*
3140		 * Adjust the packet length for the crypto additions
3141		 * done during encap and any other bits that the f/w
3142		 * will add later on.
3143		 */
3144		cip = k->wk_cipher;
3145		pktlen += cip->ic_header + cip->ic_miclen + cip->ic_trailer;
3146
3147		/* packet header may have moved, reset our local pointer */
3148		wh = mtod(m0, struct ieee80211_frame *);
3149	}
3150
3151	if (ieee80211_radiotap_active_vap(vap)) {
3152		sc->sc_tx_th.wt_flags = 0;	/* XXX */
3153		if (iswep)
3154			sc->sc_tx_th.wt_flags |= IEEE80211_RADIOTAP_F_WEP;
3155#if 0
3156		sc->sc_tx_th.wt_rate = ds->DataRate;
3157#endif
3158		sc->sc_tx_th.wt_txpower = ni->ni_txpower;
3159		sc->sc_tx_th.wt_antenna = sc->sc_txantenna;
3160
3161		ieee80211_radiotap_tx(vap, m0);
3162	}
3163	/*
3164	 * Copy up/down the 802.11 header; the firmware requires
3165	 * we present a 2-byte payload length followed by a
3166	 * 4-address header (w/o QoS), followed (optionally) by
3167	 * any WEP/ExtIV header (but only filled in for CCMP).
3168	 * We are assured the mbuf has sufficient headroom to
3169	 * prepend in-place by the setup of ic_headroom in
3170	 * mwl_attach.
3171	 */
3172	if (hdrlen < sizeof(struct mwltxrec)) {
3173		const int space = sizeof(struct mwltxrec) - hdrlen;
3174		if (M_LEADINGSPACE(m0) < space) {
3175			/* NB: should never happen */
3176			device_printf(sc->sc_dev,
3177			    "not enough headroom, need %d found %zd, "
3178			    "m_flags 0x%x m_len %d\n",
3179			    space, M_LEADINGSPACE(m0), m0->m_flags, m0->m_len);
3180			ieee80211_dump_pkt(ic,
3181			    mtod(m0, const uint8_t *), m0->m_len, 0, -1);
3182			m_freem(m0);
3183			sc->sc_stats.mst_tx_noheadroom++;
3184			return EIO;
3185		}
3186		M_PREPEND(m0, space, M_NOWAIT);
3187	}
3188	tr = mtod(m0, struct mwltxrec *);
3189	if (wh != (struct ieee80211_frame *) &tr->wh)
3190		ovbcopy(wh, &tr->wh, hdrlen);
3191	/*
3192	 * Note: the "firmware length" is actually the length
3193	 * of the fully formed "802.11 payload".  That is, it's
3194	 * everything except for the 802.11 header.  In particular
3195	 * this includes all crypto material including the MIC!
3196	 */
3197	tr->fwlen = htole16(pktlen - hdrlen);
3198
3199	/*
3200	 * Load the DMA map so any coalescing is done.  This
3201	 * also calculates the number of descriptors we need.
3202	 */
3203	error = mwl_tx_dmasetup(sc, bf, m0);
3204	if (error != 0) {
3205		/* NB: stat collected in mwl_tx_dmasetup */
3206		DPRINTF(sc, MWL_DEBUG_XMIT,
3207		    "%s: unable to setup dma\n", __func__);
3208		return error;
3209	}
3210	bf->bf_node = ni;			/* NB: held reference */
3211	m0 = bf->bf_m;				/* NB: may have changed */
3212	tr = mtod(m0, struct mwltxrec *);
3213	wh = (struct ieee80211_frame *)&tr->wh;
3214
3215	/*
3216	 * Formulate tx descriptor.
3217	 */
3218	ds = bf->bf_desc;
3219	txq = bf->bf_txq;
3220
3221	ds->QosCtrl = qos;			/* NB: already little-endian */
3222#if MWL_TXDESC == 1
3223	/*
3224	 * NB: multiframes should be zero because the descriptors
3225	 *     are initialized to zero.  This should handle the case
3226	 *     where the driver is built with MWL_TXDESC=1 but we are
3227	 *     using firmware with multi-segment support.
3228	 */
3229	ds->PktPtr = htole32(bf->bf_segs[0].ds_addr);
3230	ds->PktLen = htole16(bf->bf_segs[0].ds_len);
3231#else
3232	ds->multiframes = htole32(bf->bf_nseg);
3233	ds->PktLen = htole16(m0->m_pkthdr.len);
3234	for (i = 0; i < bf->bf_nseg; i++) {
3235		ds->PktPtrArray[i] = htole32(bf->bf_segs[i].ds_addr);
3236		ds->PktLenArray[i] = htole16(bf->bf_segs[i].ds_len);
3237	}
3238#endif
3239	/* NB: pPhysNext, DataRate, and SapPktInfo setup once, don't touch */
3240	ds->Format = 0;
3241	ds->pad = 0;
3242	ds->ack_wcb_addr = 0;
3243
3244	mn = MWL_NODE(ni);
3245	/*
3246	 * Select transmit rate.
3247	 */
3248	switch (wh->i_fc[0] & IEEE80211_FC0_TYPE_MASK) {
3249	case IEEE80211_FC0_TYPE_MGT:
3250		sc->sc_stats.mst_tx_mgmt++;
3251		/* fall thru... */
3252	case IEEE80211_FC0_TYPE_CTL:
3253		/* NB: assign to BE q to avoid bursting */
3254		ds->TxPriority = MWL_WME_AC_BE;
3255		break;
3256	case IEEE80211_FC0_TYPE_DATA:
3257		if (!ismcast) {
3258			const struct ieee80211_txparam *tp = ni->ni_txparms;
3259			/*
3260			 * EAPOL frames get forced to a fixed rate and w/o
3261			 * aggregation; otherwise check for any fixed rate
3262			 * for the client (may depend on association state).
3263			 */
3264			if (m0->m_flags & M_EAPOL) {
3265				const struct mwl_vap *mvp = MWL_VAP_CONST(vap);
3266				ds->Format = mvp->mv_eapolformat;
3267				ds->pad = htole16(
3268				    EAGLE_TXD_FIXED_RATE | EAGLE_TXD_DONT_AGGR);
3269			} else if (tp->ucastrate != IEEE80211_FIXED_RATE_NONE) {
3270				/* XXX pre-calculate per node */
3271				ds->Format = htole16(
3272				    mwl_calcformat(tp->ucastrate, ni));
3273				ds->pad = htole16(EAGLE_TXD_FIXED_RATE);
3274			}
3275			/* NB: EAPOL frames will never have qos set */
3276			if (qos == 0)
3277				ds->TxPriority = txq->qnum;
3278#if MWL_MAXBA > 3
3279			else if (mwl_bastream_match(&mn->mn_ba[3], qos))
3280				ds->TxPriority = mn->mn_ba[3].txq;
3281#endif
3282#if MWL_MAXBA > 2
3283			else if (mwl_bastream_match(&mn->mn_ba[2], qos))
3284				ds->TxPriority = mn->mn_ba[2].txq;
3285#endif
3286#if MWL_MAXBA > 1
3287			else if (mwl_bastream_match(&mn->mn_ba[1], qos))
3288				ds->TxPriority = mn->mn_ba[1].txq;
3289#endif
3290#if MWL_MAXBA > 0
3291			else if (mwl_bastream_match(&mn->mn_ba[0], qos))
3292				ds->TxPriority = mn->mn_ba[0].txq;
3293#endif
3294			else
3295				ds->TxPriority = txq->qnum;
3296		} else
3297			ds->TxPriority = txq->qnum;
3298		break;
3299	default:
3300		device_printf(sc->sc_dev, "bogus frame type 0x%x (%s)\n",
3301			wh->i_fc[0] & IEEE80211_FC0_TYPE_MASK, __func__);
3302		sc->sc_stats.mst_tx_badframetype++;
3303		m_freem(m0);
3304		return EIO;
3305	}
3306
3307	if (IFF_DUMPPKTS_XMIT(sc))
3308		ieee80211_dump_pkt(ic,
3309		    mtod(m0, const uint8_t *)+sizeof(uint16_t),
3310		    m0->m_len - sizeof(uint16_t), ds->DataRate, -1);
3311
3312	MWL_TXQ_LOCK(txq);
3313	ds->Status = htole32(EAGLE_TXD_STATUS_FW_OWNED);
3314	STAILQ_INSERT_TAIL(&txq->active, bf, bf_list);
3315	MWL_TXDESC_SYNC(txq, ds, BUS_DMASYNC_PREREAD | BUS_DMASYNC_PREWRITE);
3316
3317	sc->sc_tx_timer = 5;
3318	MWL_TXQ_UNLOCK(txq);
3319
3320	return 0;
3321}
3322
3323static __inline int
3324mwl_cvtlegacyrix(int rix)
3325{
3326	static const int ieeerates[] =
3327	    { 2, 4, 11, 22, 44, 12, 18, 24, 36, 48, 72, 96, 108 };
3328	return (rix < nitems(ieeerates) ? ieeerates[rix] : 0);
3329}
3330
3331/*
3332 * Process completed xmit descriptors from the specified queue.
3333 */
3334static int
3335mwl_tx_processq(struct mwl_softc *sc, struct mwl_txq *txq)
3336{
3337#define	EAGLE_TXD_STATUS_MCAST \
3338	(EAGLE_TXD_STATUS_MULTICAST_TX | EAGLE_TXD_STATUS_BROADCAST_TX)
3339	struct ieee80211com *ic = &sc->sc_ic;
3340	struct mwl_txbuf *bf;
3341	struct mwl_txdesc *ds;
3342	struct ieee80211_node *ni;
3343	struct mwl_node *an;
3344	int nreaped;
3345	uint32_t status;
3346
3347	DPRINTF(sc, MWL_DEBUG_TX_PROC, "%s: tx queue %u\n", __func__, txq->qnum);
3348	for (nreaped = 0;; nreaped++) {
3349		MWL_TXQ_LOCK(txq);
3350		bf = STAILQ_FIRST(&txq->active);
3351		if (bf == NULL) {
3352			MWL_TXQ_UNLOCK(txq);
3353			break;
3354		}
3355		ds = bf->bf_desc;
3356		MWL_TXDESC_SYNC(txq, ds,
3357		    BUS_DMASYNC_POSTREAD | BUS_DMASYNC_POSTWRITE);
3358		if (ds->Status & htole32(EAGLE_TXD_STATUS_FW_OWNED)) {
3359			MWL_TXQ_UNLOCK(txq);
3360			break;
3361		}
3362		STAILQ_REMOVE_HEAD(&txq->active, bf_list);
3363		MWL_TXQ_UNLOCK(txq);
3364
3365#ifdef MWL_DEBUG
3366		if (sc->sc_debug & MWL_DEBUG_XMIT_DESC)
3367			mwl_printtxbuf(bf, txq->qnum, nreaped);
3368#endif
3369		ni = bf->bf_node;
3370		if (ni != NULL) {
3371			an = MWL_NODE(ni);
3372			status = le32toh(ds->Status);
3373			if (status & EAGLE_TXD_STATUS_OK) {
3374				uint16_t Format = le16toh(ds->Format);
3375				uint8_t txant = MS(Format, EAGLE_TXD_ANTENNA);
3376
3377				sc->sc_stats.mst_ant_tx[txant]++;
3378				if (status & EAGLE_TXD_STATUS_OK_RETRY)
3379					sc->sc_stats.mst_tx_retries++;
3380				if (status & EAGLE_TXD_STATUS_OK_MORE_RETRY)
3381					sc->sc_stats.mst_tx_mretries++;
3382				if (txq->qnum >= MWL_WME_AC_VO)
3383					ic->ic_wme.wme_hipri_traffic++;
3384				ni->ni_txrate = MS(Format, EAGLE_TXD_RATE);
3385				if ((Format & EAGLE_TXD_FORMAT_HT) == 0) {
3386					ni->ni_txrate = mwl_cvtlegacyrix(
3387					    ni->ni_txrate);
3388				} else
3389					ni->ni_txrate |= IEEE80211_RATE_MCS;
3390				sc->sc_stats.mst_tx_rate = ni->ni_txrate;
3391			} else {
3392				if (status & EAGLE_TXD_STATUS_FAILED_LINK_ERROR)
3393					sc->sc_stats.mst_tx_linkerror++;
3394				if (status & EAGLE_TXD_STATUS_FAILED_XRETRY)
3395					sc->sc_stats.mst_tx_xretries++;
3396				if (status & EAGLE_TXD_STATUS_FAILED_AGING)
3397					sc->sc_stats.mst_tx_aging++;
3398				if (bf->bf_m->m_flags & M_FF)
3399					sc->sc_stats.mst_ff_txerr++;
3400			}
3401			if (bf->bf_m->m_flags & M_TXCB)
3402				/* XXX strip fw len in case header inspected */
3403				m_adj(bf->bf_m, sizeof(uint16_t));
3404			ieee80211_tx_complete(ni, bf->bf_m,
3405			    (status & EAGLE_TXD_STATUS_OK) == 0);
3406		} else
3407			m_freem(bf->bf_m);
3408		ds->Status = htole32(EAGLE_TXD_STATUS_IDLE);
3409
3410		bus_dmamap_sync(sc->sc_dmat, bf->bf_dmamap,
3411		    BUS_DMASYNC_POSTWRITE);
3412		bus_dmamap_unload(sc->sc_dmat, bf->bf_dmamap);
3413
3414		mwl_puttxbuf_tail(txq, bf);
3415	}
3416	return nreaped;
3417#undef EAGLE_TXD_STATUS_MCAST
3418}
3419
3420/*
3421 * Deferred processing of transmit interrupt; special-cased
3422 * for four hardware queues, 0-3.
3423 */
3424static void
3425mwl_tx_proc(void *arg, int npending)
3426{
3427	struct mwl_softc *sc = arg;
3428	int nreaped;
3429
3430	/*
3431	 * Process each active queue.
3432	 */
3433	nreaped = 0;
3434	if (!STAILQ_EMPTY(&sc->sc_txq[0].active))
3435		nreaped += mwl_tx_processq(sc, &sc->sc_txq[0]);
3436	if (!STAILQ_EMPTY(&sc->sc_txq[1].active))
3437		nreaped += mwl_tx_processq(sc, &sc->sc_txq[1]);
3438	if (!STAILQ_EMPTY(&sc->sc_txq[2].active))
3439		nreaped += mwl_tx_processq(sc, &sc->sc_txq[2]);
3440	if (!STAILQ_EMPTY(&sc->sc_txq[3].active))
3441		nreaped += mwl_tx_processq(sc, &sc->sc_txq[3]);
3442
3443	if (nreaped != 0) {
3444		sc->sc_tx_timer = 0;
3445		if (mbufq_first(&sc->sc_snd) != NULL) {
3446			/* NB: kick fw; the tx thread may have been preempted */
3447			mwl_hal_txstart(sc->sc_mh, 0);
3448			mwl_start(sc);
3449		}
3450	}
3451}
3452
3453static void
3454mwl_tx_draintxq(struct mwl_softc *sc, struct mwl_txq *txq)
3455{
3456	struct ieee80211_node *ni;
3457	struct mwl_txbuf *bf;
3458	u_int ix;
3459
3460	/*
3461	 * NB: this assumes output has been stopped and
3462	 *     we do not need to block mwl_tx_tasklet
3463	 */
3464	for (ix = 0;; ix++) {
3465		MWL_TXQ_LOCK(txq);
3466		bf = STAILQ_FIRST(&txq->active);
3467		if (bf == NULL) {
3468			MWL_TXQ_UNLOCK(txq);
3469			break;
3470		}
3471		STAILQ_REMOVE_HEAD(&txq->active, bf_list);
3472		MWL_TXQ_UNLOCK(txq);
3473#ifdef MWL_DEBUG
3474		if (sc->sc_debug & MWL_DEBUG_RESET) {
3475			struct ieee80211com *ic = &sc->sc_ic;
3476			const struct mwltxrec *tr =
3477			    mtod(bf->bf_m, const struct mwltxrec *);
3478			mwl_printtxbuf(bf, txq->qnum, ix);
3479			ieee80211_dump_pkt(ic, (const uint8_t *)&tr->wh,
3480				bf->bf_m->m_len - sizeof(tr->fwlen), 0, -1);
3481		}
3482#endif /* MWL_DEBUG */
3483		bus_dmamap_unload(sc->sc_dmat, bf->bf_dmamap);
3484		ni = bf->bf_node;
3485		if (ni != NULL) {
3486			/*
3487			 * Reclaim node reference.
3488			 */
3489			ieee80211_free_node(ni);
3490		}
3491		m_freem(bf->bf_m);
3492
3493		mwl_puttxbuf_tail(txq, bf);
3494	}
3495}
3496
3497/*
3498 * Drain the transmit queues and reclaim resources.
3499 */
3500static void
3501mwl_draintxq(struct mwl_softc *sc)
3502{
3503	int i;
3504
3505	for (i = 0; i < MWL_NUM_TX_QUEUES; i++)
3506		mwl_tx_draintxq(sc, &sc->sc_txq[i]);
3507	sc->sc_tx_timer = 0;
3508}
3509
3510#ifdef MWL_DIAGAPI
3511/*
3512 * Reset the transmit queues to a pristine state after a fw download.
3513 */
3514static void
3515mwl_resettxq(struct mwl_softc *sc)
3516{
3517	int i;
3518
3519	for (i = 0; i < MWL_NUM_TX_QUEUES; i++)
3520		mwl_txq_reset(sc, &sc->sc_txq[i]);
3521}
3522#endif /* MWL_DIAGAPI */
3523
3524/*
3525 * Clear the transmit queues of any frames submitted for the
3526 * specified vap.  This is done when the vap is deleted so we
3527 * don't potentially reference the vap after it is gone.
3528 * Note we cannot remove the frames; we only reclaim the node
3529 * reference.
3530 */
3531static void
3532mwl_cleartxq(struct mwl_softc *sc, struct ieee80211vap *vap)
3533{
3534	struct mwl_txq *txq;
3535	struct mwl_txbuf *bf;
3536	int i;
3537
3538	for (i = 0; i < MWL_NUM_TX_QUEUES; i++) {
3539		txq = &sc->sc_txq[i];
3540		MWL_TXQ_LOCK(txq);
3541		STAILQ_FOREACH(bf, &txq->active, bf_list) {
3542			struct ieee80211_node *ni = bf->bf_node;
3543			if (ni != NULL && ni->ni_vap == vap) {
3544				bf->bf_node = NULL;
3545				ieee80211_free_node(ni);
3546			}
3547		}
3548		MWL_TXQ_UNLOCK(txq);
3549	}
3550}
3551
3552static int
3553mwl_recv_action(struct ieee80211_node *ni, const struct ieee80211_frame *wh,
3554	const uint8_t *frm, const uint8_t *efrm)
3555{
3556	struct mwl_softc *sc = ni->ni_ic->ic_softc;
3557	const struct ieee80211_action *ia;
3558
3559	ia = (const struct ieee80211_action *) frm;
3560	if (ia->ia_category == IEEE80211_ACTION_CAT_HT &&
3561	    ia->ia_action == IEEE80211_ACTION_HT_MIMOPWRSAVE) {
3562		const struct ieee80211_action_ht_mimopowersave *mps =
3563		    (const struct ieee80211_action_ht_mimopowersave *) ia;
3564
3565		mwl_hal_setmimops(sc->sc_mh, ni->ni_macaddr,
3566		    mps->am_control & IEEE80211_A_HT_MIMOPWRSAVE_ENA,
3567		    MS(mps->am_control, IEEE80211_A_HT_MIMOPWRSAVE_MODE));
3568		return 0;
3569	} else
3570		return sc->sc_recv_action(ni, wh, frm, efrm);
3571}
3572
3573static int
3574mwl_addba_request(struct ieee80211_node *ni, struct ieee80211_tx_ampdu *tap,
3575	int dialogtoken, int baparamset, int batimeout)
3576{
3577	struct mwl_softc *sc = ni->ni_ic->ic_softc;
3578	struct ieee80211vap *vap = ni->ni_vap;
3579	struct mwl_node *mn = MWL_NODE(ni);
3580	struct mwl_bastate *bas;
3581
3582	bas = tap->txa_private;
3583	if (bas == NULL) {
3584		const MWL_HAL_BASTREAM *sp;
3585		/*
3586		 * Check for a free BA stream slot.
3587		 */
3588#if MWL_MAXBA > 3
3589		if (mn->mn_ba[3].bastream == NULL)
3590			bas = &mn->mn_ba[3];
3591		else
3592#endif
3593#if MWL_MAXBA > 2
3594		if (mn->mn_ba[2].bastream == NULL)
3595			bas = &mn->mn_ba[2];
3596		else
3597#endif
3598#if MWL_MAXBA > 1
3599		if (mn->mn_ba[1].bastream == NULL)
3600			bas = &mn->mn_ba[1];
3601		else
3602#endif
3603#if MWL_MAXBA > 0
3604		if (mn->mn_ba[0].bastream == NULL)
3605			bas = &mn->mn_ba[0];
3606		else
3607#endif
3608		{
3609			/* sta already has max BA streams */
3610			/* XXX assign BA stream to highest priority tid */
3611			DPRINTF(sc, MWL_DEBUG_AMPDU,
3612			    "%s: already has max bastreams\n", __func__);
3613			sc->sc_stats.mst_ampdu_reject++;
3614			return 0;
3615		}
3616		/* NB: no held reference to ni */
3617		sp = mwl_hal_bastream_alloc(MWL_VAP(vap)->mv_hvap,
3618		    (baparamset & IEEE80211_BAPS_POLICY_IMMEDIATE) != 0,
3619		    ni->ni_macaddr, tap->txa_tid, ni->ni_htparam,
3620		    ni, tap);
3621		if (sp == NULL) {
3622			/*
3623			 * No available stream, return 0 so no
3624			 * a-mpdu aggregation will be done.
3625			 */
3626			DPRINTF(sc, MWL_DEBUG_AMPDU,
3627			    "%s: no bastream available\n", __func__);
3628			sc->sc_stats.mst_ampdu_nostream++;
3629			return 0;
3630		}
3631		DPRINTF(sc, MWL_DEBUG_AMPDU, "%s: alloc bastream %p\n",
3632		    __func__, sp);
3633		/* NB: qos is left zero so we won't match in mwl_tx_start */
3634		bas->bastream = sp;
3635		tap->txa_private = bas;
3636	}
3637	/* fetch current seq# from the firmware; if available */
3638	if (mwl_hal_bastream_get_seqno(sc->sc_mh, bas->bastream,
3639	    vap->iv_opmode == IEEE80211_M_STA ? vap->iv_myaddr : ni->ni_macaddr,
3640	    &tap->txa_start) != 0)
3641		tap->txa_start = 0;
3642	return sc->sc_addba_request(ni, tap, dialogtoken, baparamset, batimeout);
3643}
3644
3645static int
3646mwl_addba_response(struct ieee80211_node *ni, struct ieee80211_tx_ampdu *tap,
3647	int code, int baparamset, int batimeout)
3648{
3649	struct mwl_softc *sc = ni->ni_ic->ic_softc;
3650	struct mwl_bastate *bas;
3651
3652	bas = tap->txa_private;
3653	if (bas == NULL) {
3654		/* XXX should not happen */
3655		DPRINTF(sc, MWL_DEBUG_AMPDU,
3656		    "%s: no BA stream allocated, TID %d\n",
3657		    __func__, tap->txa_tid);
3658		sc->sc_stats.mst_addba_nostream++;
3659		return 0;
3660	}
3661	if (code == IEEE80211_STATUS_SUCCESS) {
3662		struct ieee80211vap *vap = ni->ni_vap;
3663		int bufsiz, error;
3664
3665		/*
3666		 * Tell the firmware to setup the BA stream;
3667		 * we know resources are available because we
3668		 * pre-allocated one before forming the request.
3669		 */
3670		bufsiz = MS(baparamset, IEEE80211_BAPS_BUFSIZ);
3671		if (bufsiz == 0)
3672			bufsiz = IEEE80211_AGGR_BAWMAX;
3673		error = mwl_hal_bastream_create(MWL_VAP(vap)->mv_hvap,
3674		    bas->bastream, bufsiz, bufsiz, tap->txa_start);
3675		if (error != 0) {
3676			/*
3677			 * Setup failed, return immediately so no a-mpdu
3678			 * aggregation will be done.
3679			 */
3680			mwl_hal_bastream_destroy(sc->sc_mh, bas->bastream);
3681			mwl_bastream_free(bas);
3682			tap->txa_private = NULL;
3683
3684			DPRINTF(sc, MWL_DEBUG_AMPDU,
3685			    "%s: create failed, error %d, bufsiz %d TID %d "
3686			    "htparam 0x%x\n", __func__, error, bufsiz,
3687			    tap->txa_tid, ni->ni_htparam);
3688			sc->sc_stats.mst_bacreate_failed++;
3689			return 0;
3690		}
3691		/* NB: cache txq to avoid ptr indirect */
3692		mwl_bastream_setup(bas, tap->txa_tid, bas->bastream->txq);
3693		DPRINTF(sc, MWL_DEBUG_AMPDU,
3694		    "%s: bastream %p assigned to txq %d TID %d bufsiz %d "
3695		    "htparam 0x%x\n", __func__, bas->bastream,
3696		    bas->txq, tap->txa_tid, bufsiz, ni->ni_htparam);
3697	} else {
3698		/*
3699		 * Other side NAK'd us; return the resources.
3700		 */
3701		DPRINTF(sc, MWL_DEBUG_AMPDU,
3702		    "%s: request failed with code %d, destroy bastream %p\n",
3703		    __func__, code, bas->bastream);
3704		mwl_hal_bastream_destroy(sc->sc_mh, bas->bastream);
3705		mwl_bastream_free(bas);
3706		tap->txa_private = NULL;
3707	}
3708	/* NB: firmware sends BAR so we don't need to */
3709	return sc->sc_addba_response(ni, tap, code, baparamset, batimeout);
3710}
3711
3712static void
3713mwl_addba_stop(struct ieee80211_node *ni, struct ieee80211_tx_ampdu *tap)
3714{
3715	struct mwl_softc *sc = ni->ni_ic->ic_softc;
3716	struct mwl_bastate *bas;
3717
3718	bas = tap->txa_private;
3719	if (bas != NULL) {
3720		DPRINTF(sc, MWL_DEBUG_AMPDU, "%s: destroy bastream %p\n",
3721		    __func__, bas->bastream);
3722		mwl_hal_bastream_destroy(sc->sc_mh, bas->bastream);
3723		mwl_bastream_free(bas);
3724		tap->txa_private = NULL;
3725	}
3726	sc->sc_addba_stop(ni, tap);
3727}
3728
3729/*
3730 * Setup the rx data structures.  This should only be
3731 * done once or we may get out of sync with the firmware.
3732 */
3733static int
3734mwl_startrecv(struct mwl_softc *sc)
3735{
3736	if (!sc->sc_recvsetup) {
3737		struct mwl_rxbuf *bf, *prev;
3738		struct mwl_rxdesc *ds;
3739
3740		prev = NULL;
3741		STAILQ_FOREACH(bf, &sc->sc_rxbuf, bf_list) {
3742			int error = mwl_rxbuf_init(sc, bf);
3743			if (error != 0) {
3744				DPRINTF(sc, MWL_DEBUG_RECV,
3745					"%s: mwl_rxbuf_init failed %d\n",
3746					__func__, error);
3747				return error;
3748			}
3749			if (prev != NULL) {
3750				ds = prev->bf_desc;
3751				ds->pPhysNext = htole32(bf->bf_daddr);
3752			}
3753			prev = bf;
3754		}
3755		if (prev != NULL) {
3756			ds = prev->bf_desc;
3757			ds->pPhysNext =
3758			    htole32(STAILQ_FIRST(&sc->sc_rxbuf)->bf_daddr);
3759		}
3760		sc->sc_recvsetup = 1;
3761	}
3762	mwl_mode_init(sc);		/* set filters, etc. */
3763	return 0;
3764}
3765
3766static MWL_HAL_APMODE
3767mwl_getapmode(const struct ieee80211vap *vap, struct ieee80211_channel *chan)
3768{
3769	MWL_HAL_APMODE mode;
3770
3771	if (IEEE80211_IS_CHAN_HT(chan)) {
3772		if (vap->iv_flags_ht & IEEE80211_FHT_PUREN)
3773			mode = AP_MODE_N_ONLY;
3774		else if (IEEE80211_IS_CHAN_5GHZ(chan))
3775			mode = AP_MODE_AandN;
3776		else if (vap->iv_flags & IEEE80211_F_PUREG)
3777			mode = AP_MODE_GandN;
3778		else
3779			mode = AP_MODE_BandGandN;
3780	} else if (IEEE80211_IS_CHAN_ANYG(chan)) {
3781		if (vap->iv_flags & IEEE80211_F_PUREG)
3782			mode = AP_MODE_G_ONLY;
3783		else
3784			mode = AP_MODE_MIXED;
3785	} else if (IEEE80211_IS_CHAN_B(chan))
3786		mode = AP_MODE_B_ONLY;
3787	else if (IEEE80211_IS_CHAN_A(chan))
3788		mode = AP_MODE_A_ONLY;
3789	else
3790		mode = AP_MODE_MIXED;		/* XXX should not happen? */
3791	return mode;
3792}
3793
3794static int
3795mwl_setapmode(struct ieee80211vap *vap, struct ieee80211_channel *chan)
3796{
3797	struct mwl_hal_vap *hvap = MWL_VAP(vap)->mv_hvap;
3798	return mwl_hal_setapmode(hvap, mwl_getapmode(vap, chan));
3799}
3800
3801/*
3802 * Set/change channels.
3803 */
3804static int
3805mwl_chan_set(struct mwl_softc *sc, struct ieee80211_channel *chan)
3806{
3807	struct mwl_hal *mh = sc->sc_mh;
3808	struct ieee80211com *ic = &sc->sc_ic;
3809	MWL_HAL_CHANNEL hchan;
3810	int maxtxpow;
3811
3812	DPRINTF(sc, MWL_DEBUG_RESET, "%s: chan %u MHz/flags 0x%x\n",
3813	    __func__, chan->ic_freq, chan->ic_flags);
3814
3815	/*
3816	 * Convert to a HAL channel description with
3817	 * the flags constrained to reflect the current
3818	 * operating mode.
3819	 */
3820	mwl_mapchan(&hchan, chan);
3821	mwl_hal_intrset(mh, 0);		/* disable interrupts */
3822#if 0
3823	mwl_draintxq(sc);		/* clear pending tx frames */
3824#endif
3825	mwl_hal_setchannel(mh, &hchan);
3826	/*
3827	 * Tx power is cap'd by the regulatory setting and
3828	 * possibly a user-set limit.  We pass the min of
3829	 * these to the hal to apply them to the cal data
3830	 * for this channel.
3831	 * XXX min bound?
3832	 */
3833	maxtxpow = 2*chan->ic_maxregpower;
3834	if (maxtxpow > ic->ic_txpowlimit)
3835		maxtxpow = ic->ic_txpowlimit;
3836	mwl_hal_settxpower(mh, &hchan, maxtxpow / 2);
3837	/* NB: potentially change mcast/mgt rates */
3838	mwl_setcurchanrates(sc);
3839
3840	/*
3841	 * Update internal state.
3842	 */
3843	sc->sc_tx_th.wt_chan_freq = htole16(chan->ic_freq);
3844	sc->sc_rx_th.wr_chan_freq = htole16(chan->ic_freq);
3845	if (IEEE80211_IS_CHAN_A(chan)) {
3846		sc->sc_tx_th.wt_chan_flags = htole16(IEEE80211_CHAN_A);
3847		sc->sc_rx_th.wr_chan_flags = htole16(IEEE80211_CHAN_A);
3848	} else if (IEEE80211_IS_CHAN_ANYG(chan)) {
3849		sc->sc_tx_th.wt_chan_flags = htole16(IEEE80211_CHAN_G);
3850		sc->sc_rx_th.wr_chan_flags = htole16(IEEE80211_CHAN_G);
3851	} else {
3852		sc->sc_tx_th.wt_chan_flags = htole16(IEEE80211_CHAN_B);
3853		sc->sc_rx_th.wr_chan_flags = htole16(IEEE80211_CHAN_B);
3854	}
3855	sc->sc_curchan = hchan;
3856	mwl_hal_intrset(mh, sc->sc_imask);
3857
3858	return 0;
3859}
3860
3861static void
3862mwl_scan_start(struct ieee80211com *ic)
3863{
3864	struct mwl_softc *sc = ic->ic_softc;
3865
3866	DPRINTF(sc, MWL_DEBUG_STATE, "%s\n", __func__);
3867}
3868
3869static void
3870mwl_scan_end(struct ieee80211com *ic)
3871{
3872	struct mwl_softc *sc = ic->ic_softc;
3873
3874	DPRINTF(sc, MWL_DEBUG_STATE, "%s\n", __func__);
3875}
3876
3877static void
3878mwl_set_channel(struct ieee80211com *ic)
3879{
3880	struct mwl_softc *sc = ic->ic_softc;
3881
3882	(void) mwl_chan_set(sc, ic->ic_curchan);
3883}
3884
3885/*
3886 * Handle a channel switch request.  We inform the firmware
3887 * and mark the global state to suppress various actions.
3888 * NB: we issue only one request to the fw; we may be called
3889 * multiple times if there are multiple vap's.
3890 */
3891static void
3892mwl_startcsa(struct ieee80211vap *vap)
3893{
3894	struct ieee80211com *ic = vap->iv_ic;
3895	struct mwl_softc *sc = ic->ic_softc;
3896	MWL_HAL_CHANNEL hchan;
3897
3898	if (sc->sc_csapending)
3899		return;
3900
3901	mwl_mapchan(&hchan, ic->ic_csa_newchan);
3902	/* 1 =>'s quiet channel */
3903	mwl_hal_setchannelswitchie(sc->sc_mh, &hchan, 1, ic->ic_csa_count);
3904	sc->sc_csapending = 1;
3905}
3906
3907/*
3908 * Plumb any static WEP key for the station.  This is
3909 * necessary as we must propagate the key from the
3910 * global key table of the vap to each sta db entry.
3911 */
3912static void
3913mwl_setanywepkey(struct ieee80211vap *vap, const uint8_t mac[IEEE80211_ADDR_LEN])
3914{
3915	if ((vap->iv_flags & (IEEE80211_F_PRIVACY|IEEE80211_F_WPA)) ==
3916		IEEE80211_F_PRIVACY &&
3917	    vap->iv_def_txkey != IEEE80211_KEYIX_NONE &&
3918	    vap->iv_nw_keys[vap->iv_def_txkey].wk_keyix != IEEE80211_KEYIX_NONE)
3919		(void) _mwl_key_set(vap, &vap->iv_nw_keys[vap->iv_def_txkey],
3920				    mac);
3921}
3922
3923static int
3924mwl_peerstadb(struct ieee80211_node *ni, int aid, int staid, MWL_HAL_PEERINFO *pi)
3925{
3926#define	WME(ie) ((const struct ieee80211_wme_info *) ie)
3927	struct ieee80211vap *vap = ni->ni_vap;
3928	struct mwl_hal_vap *hvap;
3929	int error;
3930
3931	if (vap->iv_opmode == IEEE80211_M_WDS) {
3932		/*
3933		 * WDS vap's do not have a f/w vap; instead they piggyback
3934		 * on an AP vap and we must install the sta db entry and
3935		 * crypto state using that AP's handle (the WDS vap has none).
3936		 */
3937		hvap = MWL_VAP(vap)->mv_ap_hvap;
3938	} else
3939		hvap = MWL_VAP(vap)->mv_hvap;
3940	error = mwl_hal_newstation(hvap, ni->ni_macaddr,
3941	    aid, staid, pi,
3942	    ni->ni_flags & (IEEE80211_NODE_QOS | IEEE80211_NODE_HT),
3943	    ni->ni_ies.wme_ie != NULL ? WME(ni->ni_ies.wme_ie)->wme_info : 0);
3944	if (error == 0) {
3945		/*
3946		 * Setup security for this station.  For sta mode this is
3947		 * needed even though do the same thing on transition to
3948		 * AUTH state because the call to mwl_hal_newstation
3949		 * clobbers the crypto state we setup.
3950		 */
3951		mwl_setanywepkey(vap, ni->ni_macaddr);
3952	}
3953	return error;
3954#undef WME
3955}
3956
3957static void
3958mwl_setglobalkeys(struct ieee80211vap *vap)
3959{
3960	struct ieee80211_key *wk;
3961
3962	wk = &vap->iv_nw_keys[0];
3963	for (; wk < &vap->iv_nw_keys[IEEE80211_WEP_NKID]; wk++)
3964		if (wk->wk_keyix != IEEE80211_KEYIX_NONE)
3965			(void) _mwl_key_set(vap, wk, vap->iv_myaddr);
3966}
3967
3968/*
3969 * Convert a legacy rate set to a firmware bitmask.
3970 */
3971static uint32_t
3972get_rate_bitmap(const struct ieee80211_rateset *rs)
3973{
3974	uint32_t rates;
3975	int i;
3976
3977	rates = 0;
3978	for (i = 0; i < rs->rs_nrates; i++)
3979		switch (rs->rs_rates[i] & IEEE80211_RATE_VAL) {
3980		case 2:	  rates |= 0x001; break;
3981		case 4:	  rates |= 0x002; break;
3982		case 11:  rates |= 0x004; break;
3983		case 22:  rates |= 0x008; break;
3984		case 44:  rates |= 0x010; break;
3985		case 12:  rates |= 0x020; break;
3986		case 18:  rates |= 0x040; break;
3987		case 24:  rates |= 0x080; break;
3988		case 36:  rates |= 0x100; break;
3989		case 48:  rates |= 0x200; break;
3990		case 72:  rates |= 0x400; break;
3991		case 96:  rates |= 0x800; break;
3992		case 108: rates |= 0x1000; break;
3993		}
3994	return rates;
3995}
3996
3997/*
3998 * Construct an HT firmware bitmask from an HT rate set.
3999 */
4000static uint32_t
4001get_htrate_bitmap(const struct ieee80211_htrateset *rs)
4002{
4003	uint32_t rates;
4004	int i;
4005
4006	rates = 0;
4007	for (i = 0; i < rs->rs_nrates; i++) {
4008		if (rs->rs_rates[i] < 16)
4009			rates |= 1<<rs->rs_rates[i];
4010	}
4011	return rates;
4012}
4013
4014/*
4015 * Craft station database entry for station.
4016 * NB: use host byte order here, the hal handles byte swapping.
4017 */
4018static MWL_HAL_PEERINFO *
4019mkpeerinfo(MWL_HAL_PEERINFO *pi, const struct ieee80211_node *ni)
4020{
4021	const struct ieee80211vap *vap = ni->ni_vap;
4022
4023	memset(pi, 0, sizeof(*pi));
4024	pi->LegacyRateBitMap = get_rate_bitmap(&ni->ni_rates);
4025	pi->CapInfo = ni->ni_capinfo;
4026	if (ni->ni_flags & IEEE80211_NODE_HT) {
4027		/* HT capabilities, etc */
4028		pi->HTCapabilitiesInfo = ni->ni_htcap;
4029		/* XXX pi.HTCapabilitiesInfo */
4030	        pi->MacHTParamInfo = ni->ni_htparam;
4031		pi->HTRateBitMap = get_htrate_bitmap(&ni->ni_htrates);
4032		pi->AddHtInfo.ControlChan = ni->ni_htctlchan;
4033		pi->AddHtInfo.AddChan = ni->ni_ht2ndchan;
4034		pi->AddHtInfo.OpMode = ni->ni_htopmode;
4035		pi->AddHtInfo.stbc = ni->ni_htstbc;
4036
4037		/* constrain according to local configuration */
4038		if ((vap->iv_flags_ht & IEEE80211_FHT_SHORTGI40) == 0)
4039			pi->HTCapabilitiesInfo &= ~IEEE80211_HTCAP_SHORTGI40;
4040		if ((vap->iv_flags_ht & IEEE80211_FHT_SHORTGI20) == 0)
4041			pi->HTCapabilitiesInfo &= ~IEEE80211_HTCAP_SHORTGI20;
4042		if (ni->ni_chw != 40)
4043			pi->HTCapabilitiesInfo &= ~IEEE80211_HTCAP_CHWIDTH40;
4044	}
4045	return pi;
4046}
4047
4048/*
4049 * Re-create the local sta db entry for a vap to ensure
4050 * up to date WME state is pushed to the firmware.  Because
4051 * this resets crypto state this must be followed by a
4052 * reload of any keys in the global key table.
4053 */
4054static int
4055mwl_localstadb(struct ieee80211vap *vap)
4056{
4057#define	WME(ie) ((const struct ieee80211_wme_info *) ie)
4058	struct mwl_hal_vap *hvap = MWL_VAP(vap)->mv_hvap;
4059	struct ieee80211_node *bss;
4060	MWL_HAL_PEERINFO pi;
4061	int error;
4062
4063	switch (vap->iv_opmode) {
4064	case IEEE80211_M_STA:
4065		bss = vap->iv_bss;
4066		error = mwl_hal_newstation(hvap, vap->iv_myaddr, 0, 0,
4067		    vap->iv_state == IEEE80211_S_RUN ?
4068			mkpeerinfo(&pi, bss) : NULL,
4069		    (bss->ni_flags & (IEEE80211_NODE_QOS | IEEE80211_NODE_HT)),
4070		    bss->ni_ies.wme_ie != NULL ?
4071			WME(bss->ni_ies.wme_ie)->wme_info : 0);
4072		if (error == 0)
4073			mwl_setglobalkeys(vap);
4074		break;
4075	case IEEE80211_M_HOSTAP:
4076	case IEEE80211_M_MBSS:
4077		error = mwl_hal_newstation(hvap, vap->iv_myaddr,
4078		    0, 0, NULL, vap->iv_flags & IEEE80211_F_WME, 0);
4079		if (error == 0)
4080			mwl_setglobalkeys(vap);
4081		break;
4082	default:
4083		error = 0;
4084		break;
4085	}
4086	return error;
4087#undef WME
4088}
4089
4090static int
4091mwl_newstate(struct ieee80211vap *vap, enum ieee80211_state nstate, int arg)
4092{
4093	struct mwl_vap *mvp = MWL_VAP(vap);
4094	struct mwl_hal_vap *hvap = mvp->mv_hvap;
4095	struct ieee80211com *ic = vap->iv_ic;
4096	struct ieee80211_node *ni = NULL;
4097	struct mwl_softc *sc = ic->ic_softc;
4098	struct mwl_hal *mh = sc->sc_mh;
4099	enum ieee80211_state ostate = vap->iv_state;
4100	int error;
4101
4102	DPRINTF(sc, MWL_DEBUG_STATE, "%s: %s: %s -> %s\n",
4103	    vap->iv_ifp->if_xname, __func__,
4104	    ieee80211_state_name[ostate], ieee80211_state_name[nstate]);
4105
4106	callout_stop(&sc->sc_timer);
4107	/*
4108	 * Clear current radar detection state.
4109	 */
4110	if (ostate == IEEE80211_S_CAC) {
4111		/* stop quiet mode radar detection */
4112		mwl_hal_setradardetection(mh, DR_CHK_CHANNEL_AVAILABLE_STOP);
4113	} else if (sc->sc_radarena) {
4114		/* stop in-service radar detection */
4115		mwl_hal_setradardetection(mh, DR_DFS_DISABLE);
4116		sc->sc_radarena = 0;
4117	}
4118	/*
4119	 * Carry out per-state actions before doing net80211 work.
4120	 */
4121	if (nstate == IEEE80211_S_INIT) {
4122		/* NB: only ap+sta vap's have a fw entity */
4123		if (hvap != NULL)
4124			mwl_hal_stop(hvap);
4125	} else if (nstate == IEEE80211_S_SCAN) {
4126		mwl_hal_start(hvap);
4127		/* NB: this disables beacon frames */
4128		mwl_hal_setinframode(hvap);
4129	} else if (nstate == IEEE80211_S_AUTH) {
4130		/*
4131		 * Must create a sta db entry in case a WEP key needs to
4132		 * be plumbed.  This entry will be overwritten if we
4133		 * associate; otherwise it will be reclaimed on node free.
4134		 */
4135		ni = vap->iv_bss;
4136		MWL_NODE(ni)->mn_hvap = hvap;
4137		(void) mwl_peerstadb(ni, 0, 0, NULL);
4138	} else if (nstate == IEEE80211_S_CSA) {
4139		/* XXX move to below? */
4140		if (vap->iv_opmode == IEEE80211_M_HOSTAP ||
4141		    vap->iv_opmode == IEEE80211_M_MBSS)
4142			mwl_startcsa(vap);
4143	} else if (nstate == IEEE80211_S_CAC) {
4144		/* XXX move to below? */
4145		/* stop ap xmit and enable quiet mode radar detection */
4146		mwl_hal_setradardetection(mh, DR_CHK_CHANNEL_AVAILABLE_START);
4147	}
4148
4149	/*
4150	 * Invoke the parent method to do net80211 work.
4151	 */
4152	error = mvp->mv_newstate(vap, nstate, arg);
4153
4154	/*
4155	 * Carry out work that must be done after net80211 runs;
4156	 * this work requires up to date state (e.g. iv_bss).
4157	 */
4158	if (error == 0 && nstate == IEEE80211_S_RUN) {
4159		/* NB: collect bss node again, it may have changed */
4160		ni = vap->iv_bss;
4161
4162		DPRINTF(sc, MWL_DEBUG_STATE,
4163		    "%s: %s(RUN): iv_flags 0x%08x bintvl %d bssid %s "
4164		    "capinfo 0x%04x chan %d\n",
4165		    vap->iv_ifp->if_xname, __func__, vap->iv_flags,
4166		    ni->ni_intval, ether_sprintf(ni->ni_bssid), ni->ni_capinfo,
4167		    ieee80211_chan2ieee(ic, ic->ic_curchan));
4168
4169		/*
4170		 * Recreate local sta db entry to update WME/HT state.
4171		 */
4172		mwl_localstadb(vap);
4173		switch (vap->iv_opmode) {
4174		case IEEE80211_M_HOSTAP:
4175		case IEEE80211_M_MBSS:
4176			if (ostate == IEEE80211_S_CAC) {
4177				/* enable in-service radar detection */
4178				mwl_hal_setradardetection(mh,
4179				    DR_IN_SERVICE_MONITOR_START);
4180				sc->sc_radarena = 1;
4181			}
4182			/*
4183			 * Allocate and setup the beacon frame
4184			 * (and related state).
4185			 */
4186			error = mwl_reset_vap(vap, IEEE80211_S_RUN);
4187			if (error != 0) {
4188				DPRINTF(sc, MWL_DEBUG_STATE,
4189				    "%s: beacon setup failed, error %d\n",
4190				    __func__, error);
4191				goto bad;
4192			}
4193			/* NB: must be after setting up beacon */
4194			mwl_hal_start(hvap);
4195			break;
4196		case IEEE80211_M_STA:
4197			DPRINTF(sc, MWL_DEBUG_STATE, "%s: %s: aid 0x%x\n",
4198			    vap->iv_ifp->if_xname, __func__, ni->ni_associd);
4199			/*
4200			 * Set state now that we're associated.
4201			 */
4202			mwl_hal_setassocid(hvap, ni->ni_bssid, ni->ni_associd);
4203			mwl_setrates(vap);
4204			mwl_hal_setrtsthreshold(hvap, vap->iv_rtsthreshold);
4205			if ((vap->iv_flags & IEEE80211_F_DWDS) &&
4206			    sc->sc_ndwdsvaps++ == 0)
4207				mwl_hal_setdwds(mh, 1);
4208			break;
4209		case IEEE80211_M_WDS:
4210			DPRINTF(sc, MWL_DEBUG_STATE, "%s: %s: bssid %s\n",
4211			    vap->iv_ifp->if_xname, __func__,
4212			    ether_sprintf(ni->ni_bssid));
4213			mwl_seteapolformat(vap);
4214			break;
4215		default:
4216			break;
4217		}
4218		/*
4219		 * Set CS mode according to operating channel;
4220		 * this mostly an optimization for 5GHz.
4221		 *
4222		 * NB: must follow mwl_hal_start which resets csmode
4223		 */
4224		if (IEEE80211_IS_CHAN_5GHZ(ic->ic_bsschan))
4225			mwl_hal_setcsmode(mh, CSMODE_AGGRESSIVE);
4226		else
4227			mwl_hal_setcsmode(mh, CSMODE_AUTO_ENA);
4228		/*
4229		 * Start timer to prod firmware.
4230		 */
4231		if (sc->sc_ageinterval != 0)
4232			callout_reset(&sc->sc_timer, sc->sc_ageinterval*hz,
4233			    mwl_agestations, sc);
4234	} else if (nstate == IEEE80211_S_SLEEP) {
4235		/* XXX set chip in power save */
4236	} else if ((vap->iv_flags & IEEE80211_F_DWDS) &&
4237	    --sc->sc_ndwdsvaps == 0)
4238		mwl_hal_setdwds(mh, 0);
4239bad:
4240	return error;
4241}
4242
4243/*
4244 * Manage station id's; these are separate from AID's
4245 * as AID's may have values out of the range of possible
4246 * station id's acceptable to the firmware.
4247 */
4248static int
4249allocstaid(struct mwl_softc *sc, int aid)
4250{
4251	int staid;
4252
4253	if (!(0 < aid && aid < MWL_MAXSTAID) || isset(sc->sc_staid, aid)) {
4254		/* NB: don't use 0 */
4255		for (staid = 1; staid < MWL_MAXSTAID; staid++)
4256			if (isclr(sc->sc_staid, staid))
4257				break;
4258	} else
4259		staid = aid;
4260	setbit(sc->sc_staid, staid);
4261	return staid;
4262}
4263
4264static void
4265delstaid(struct mwl_softc *sc, int staid)
4266{
4267	clrbit(sc->sc_staid, staid);
4268}
4269
4270/*
4271 * Setup driver-specific state for a newly associated node.
4272 * Note that we're called also on a re-associate, the isnew
4273 * param tells us if this is the first time or not.
4274 */
4275static void
4276mwl_newassoc(struct ieee80211_node *ni, int isnew)
4277{
4278	struct ieee80211vap *vap = ni->ni_vap;
4279        struct mwl_softc *sc = vap->iv_ic->ic_softc;
4280	struct mwl_node *mn = MWL_NODE(ni);
4281	MWL_HAL_PEERINFO pi;
4282	uint16_t aid;
4283	int error;
4284
4285	aid = IEEE80211_AID(ni->ni_associd);
4286	if (isnew) {
4287		mn->mn_staid = allocstaid(sc, aid);
4288		mn->mn_hvap = MWL_VAP(vap)->mv_hvap;
4289	} else {
4290		mn = MWL_NODE(ni);
4291		/* XXX reset BA stream? */
4292	}
4293	DPRINTF(sc, MWL_DEBUG_NODE, "%s: mac %s isnew %d aid %d staid %d\n",
4294	    __func__, ether_sprintf(ni->ni_macaddr), isnew, aid, mn->mn_staid);
4295	error = mwl_peerstadb(ni, aid, mn->mn_staid, mkpeerinfo(&pi, ni));
4296	if (error != 0) {
4297		DPRINTF(sc, MWL_DEBUG_NODE,
4298		    "%s: error %d creating sta db entry\n",
4299		    __func__, error);
4300		/* XXX how to deal with error? */
4301	}
4302}
4303
4304/*
4305 * Periodically poke the firmware to age out station state
4306 * (power save queues, pending tx aggregates).
4307 */
4308static void
4309mwl_agestations(void *arg)
4310{
4311	struct mwl_softc *sc = arg;
4312
4313	mwl_hal_setkeepalive(sc->sc_mh);
4314	if (sc->sc_ageinterval != 0)		/* NB: catch dynamic changes */
4315		callout_schedule(&sc->sc_timer, sc->sc_ageinterval*hz);
4316}
4317
4318static const struct mwl_hal_channel *
4319findhalchannel(const MWL_HAL_CHANNELINFO *ci, int ieee)
4320{
4321	int i;
4322
4323	for (i = 0; i < ci->nchannels; i++) {
4324		const struct mwl_hal_channel *hc = &ci->channels[i];
4325		if (hc->ieee == ieee)
4326			return hc;
4327	}
4328	return NULL;
4329}
4330
4331static int
4332mwl_setregdomain(struct ieee80211com *ic, struct ieee80211_regdomain *rd,
4333	int nchan, struct ieee80211_channel chans[])
4334{
4335	struct mwl_softc *sc = ic->ic_softc;
4336	struct mwl_hal *mh = sc->sc_mh;
4337	const MWL_HAL_CHANNELINFO *ci;
4338	int i;
4339
4340	for (i = 0; i < nchan; i++) {
4341		struct ieee80211_channel *c = &chans[i];
4342		const struct mwl_hal_channel *hc;
4343
4344		if (IEEE80211_IS_CHAN_2GHZ(c)) {
4345			mwl_hal_getchannelinfo(mh, MWL_FREQ_BAND_2DOT4GHZ,
4346			    IEEE80211_IS_CHAN_HT40(c) ?
4347				MWL_CH_40_MHz_WIDTH : MWL_CH_20_MHz_WIDTH, &ci);
4348		} else if (IEEE80211_IS_CHAN_5GHZ(c)) {
4349			mwl_hal_getchannelinfo(mh, MWL_FREQ_BAND_5GHZ,
4350			    IEEE80211_IS_CHAN_HT40(c) ?
4351				MWL_CH_40_MHz_WIDTH : MWL_CH_20_MHz_WIDTH, &ci);
4352		} else {
4353			device_printf(sc->sc_dev,
4354			    "%s: channel %u freq %u/0x%x not 2.4/5GHz\n",
4355			    __func__, c->ic_ieee, c->ic_freq, c->ic_flags);
4356			return EINVAL;
4357		}
4358		/*
4359		 * Verify channel has cal data and cap tx power.
4360		 */
4361		hc = findhalchannel(ci, c->ic_ieee);
4362		if (hc != NULL) {
4363			if (c->ic_maxpower > 2*hc->maxTxPow)
4364				c->ic_maxpower = 2*hc->maxTxPow;
4365			goto next;
4366		}
4367		if (IEEE80211_IS_CHAN_HT40(c)) {
4368			/*
4369			 * Look for the extension channel since the
4370			 * hal table only has the primary channel.
4371			 */
4372			hc = findhalchannel(ci, c->ic_extieee);
4373			if (hc != NULL) {
4374				if (c->ic_maxpower > 2*hc->maxTxPow)
4375					c->ic_maxpower = 2*hc->maxTxPow;
4376				goto next;
4377			}
4378		}
4379		device_printf(sc->sc_dev,
4380		    "%s: no cal data for channel %u ext %u freq %u/0x%x\n",
4381		    __func__, c->ic_ieee, c->ic_extieee,
4382		    c->ic_freq, c->ic_flags);
4383		return EINVAL;
4384	next:
4385		;
4386	}
4387	return 0;
4388}
4389
4390#define	IEEE80211_CHAN_HTG	(IEEE80211_CHAN_HT|IEEE80211_CHAN_G)
4391#define	IEEE80211_CHAN_HTA	(IEEE80211_CHAN_HT|IEEE80211_CHAN_A)
4392
4393static void
4394addchan(struct ieee80211_channel *c, int freq, int flags, int ieee, int txpow)
4395{
4396	c->ic_freq = freq;
4397	c->ic_flags = flags;
4398	c->ic_ieee = ieee;
4399	c->ic_minpower = 0;
4400	c->ic_maxpower = 2*txpow;
4401	c->ic_maxregpower = txpow;
4402}
4403
4404static const struct ieee80211_channel *
4405findchannel(const struct ieee80211_channel chans[], int nchans,
4406	int freq, int flags)
4407{
4408	const struct ieee80211_channel *c;
4409	int i;
4410
4411	for (i = 0; i < nchans; i++) {
4412		c = &chans[i];
4413		if (c->ic_freq == freq && c->ic_flags == flags)
4414			return c;
4415	}
4416	return NULL;
4417}
4418
4419static void
4420addht40channels(struct ieee80211_channel chans[], int maxchans, int *nchans,
4421	const MWL_HAL_CHANNELINFO *ci, int flags)
4422{
4423	struct ieee80211_channel *c;
4424	const struct ieee80211_channel *extc;
4425	const struct mwl_hal_channel *hc;
4426	int i;
4427
4428	c = &chans[*nchans];
4429
4430	flags &= ~IEEE80211_CHAN_HT;
4431	for (i = 0; i < ci->nchannels; i++) {
4432		/*
4433		 * Each entry defines an HT40 channel pair; find the
4434		 * extension channel above and the insert the pair.
4435		 */
4436		hc = &ci->channels[i];
4437		extc = findchannel(chans, *nchans, hc->freq+20,
4438		    flags | IEEE80211_CHAN_HT20);
4439		if (extc != NULL) {
4440			if (*nchans >= maxchans)
4441				break;
4442			addchan(c, hc->freq, flags | IEEE80211_CHAN_HT40U,
4443			    hc->ieee, hc->maxTxPow);
4444			c->ic_extieee = extc->ic_ieee;
4445			c++, (*nchans)++;
4446			if (*nchans >= maxchans)
4447				break;
4448			addchan(c, extc->ic_freq, flags | IEEE80211_CHAN_HT40D,
4449			    extc->ic_ieee, hc->maxTxPow);
4450			c->ic_extieee = hc->ieee;
4451			c++, (*nchans)++;
4452		}
4453	}
4454}
4455
4456static void
4457addchannels(struct ieee80211_channel chans[], int maxchans, int *nchans,
4458	const MWL_HAL_CHANNELINFO *ci, int flags)
4459{
4460	struct ieee80211_channel *c;
4461	int i;
4462
4463	c = &chans[*nchans];
4464
4465	for (i = 0; i < ci->nchannels; i++) {
4466		const struct mwl_hal_channel *hc;
4467
4468		hc = &ci->channels[i];
4469		if (*nchans >= maxchans)
4470			break;
4471		addchan(c, hc->freq, flags, hc->ieee, hc->maxTxPow);
4472		c++, (*nchans)++;
4473		if (flags == IEEE80211_CHAN_G || flags == IEEE80211_CHAN_HTG) {
4474			/* g channel have a separate b-only entry */
4475			if (*nchans >= maxchans)
4476				break;
4477			c[0] = c[-1];
4478			c[-1].ic_flags = IEEE80211_CHAN_B;
4479			c++, (*nchans)++;
4480		}
4481		if (flags == IEEE80211_CHAN_HTG) {
4482			/* HT g channel have a separate g-only entry */
4483			if (*nchans >= maxchans)
4484				break;
4485			c[-1].ic_flags = IEEE80211_CHAN_G;
4486			c[0] = c[-1];
4487			c[0].ic_flags &= ~IEEE80211_CHAN_HT;
4488			c[0].ic_flags |= IEEE80211_CHAN_HT20;	/* HT20 */
4489			c++, (*nchans)++;
4490		}
4491		if (flags == IEEE80211_CHAN_HTA) {
4492			/* HT a channel have a separate a-only entry */
4493			if (*nchans >= maxchans)
4494				break;
4495			c[-1].ic_flags = IEEE80211_CHAN_A;
4496			c[0] = c[-1];
4497			c[0].ic_flags &= ~IEEE80211_CHAN_HT;
4498			c[0].ic_flags |= IEEE80211_CHAN_HT20;	/* HT20 */
4499			c++, (*nchans)++;
4500		}
4501	}
4502}
4503
4504static void
4505getchannels(struct mwl_softc *sc, int maxchans, int *nchans,
4506	struct ieee80211_channel chans[])
4507{
4508	const MWL_HAL_CHANNELINFO *ci;
4509
4510	/*
4511	 * Use the channel info from the hal to craft the
4512	 * channel list.  Note that we pass back an unsorted
4513	 * list; the caller is required to sort it for us
4514	 * (if desired).
4515	 */
4516	*nchans = 0;
4517	if (mwl_hal_getchannelinfo(sc->sc_mh,
4518	    MWL_FREQ_BAND_2DOT4GHZ, MWL_CH_20_MHz_WIDTH, &ci) == 0)
4519		addchannels(chans, maxchans, nchans, ci, IEEE80211_CHAN_HTG);
4520	if (mwl_hal_getchannelinfo(sc->sc_mh,
4521	    MWL_FREQ_BAND_5GHZ, MWL_CH_20_MHz_WIDTH, &ci) == 0)
4522		addchannels(chans, maxchans, nchans, ci, IEEE80211_CHAN_HTA);
4523	if (mwl_hal_getchannelinfo(sc->sc_mh,
4524	    MWL_FREQ_BAND_2DOT4GHZ, MWL_CH_40_MHz_WIDTH, &ci) == 0)
4525		addht40channels(chans, maxchans, nchans, ci, IEEE80211_CHAN_HTG);
4526	if (mwl_hal_getchannelinfo(sc->sc_mh,
4527	    MWL_FREQ_BAND_5GHZ, MWL_CH_40_MHz_WIDTH, &ci) == 0)
4528		addht40channels(chans, maxchans, nchans, ci, IEEE80211_CHAN_HTA);
4529}
4530
4531static void
4532mwl_getradiocaps(struct ieee80211com *ic,
4533	int maxchans, int *nchans, struct ieee80211_channel chans[])
4534{
4535	struct mwl_softc *sc = ic->ic_softc;
4536
4537	getchannels(sc, maxchans, nchans, chans);
4538}
4539
4540static int
4541mwl_getchannels(struct mwl_softc *sc)
4542{
4543	struct ieee80211com *ic = &sc->sc_ic;
4544
4545	/*
4546	 * Use the channel info from the hal to craft the
4547	 * channel list for net80211.  Note that we pass up
4548	 * an unsorted list; net80211 will sort it for us.
4549	 */
4550	memset(ic->ic_channels, 0, sizeof(ic->ic_channels));
4551	ic->ic_nchans = 0;
4552	getchannels(sc, IEEE80211_CHAN_MAX, &ic->ic_nchans, ic->ic_channels);
4553
4554	ic->ic_regdomain.regdomain = SKU_DEBUG;
4555	ic->ic_regdomain.country = CTRY_DEFAULT;
4556	ic->ic_regdomain.location = 'I';
4557	ic->ic_regdomain.isocc[0] = ' ';	/* XXX? */
4558	ic->ic_regdomain.isocc[1] = ' ';
4559	return (ic->ic_nchans == 0 ? EIO : 0);
4560}
4561#undef IEEE80211_CHAN_HTA
4562#undef IEEE80211_CHAN_HTG
4563
4564#ifdef MWL_DEBUG
4565static void
4566mwl_printrxbuf(const struct mwl_rxbuf *bf, u_int ix)
4567{
4568	const struct mwl_rxdesc *ds = bf->bf_desc;
4569	uint32_t status = le32toh(ds->Status);
4570
4571	printf("R[%2u] (DS.V:%p DS.P:0x%jx) NEXT:%08x DATA:%08x RC:%02x%s\n"
4572	       "      STAT:%02x LEN:%04x RSSI:%02x CHAN:%02x RATE:%02x QOS:%04x HT:%04x\n",
4573	    ix, ds, (uintmax_t)bf->bf_daddr, le32toh(ds->pPhysNext),
4574	    le32toh(ds->pPhysBuffData), ds->RxControl,
4575	    ds->RxControl != EAGLE_RXD_CTRL_DRIVER_OWN ?
4576	        "" : (status & EAGLE_RXD_STATUS_OK) ? " *" : " !",
4577	    ds->Status, le16toh(ds->PktLen), ds->RSSI, ds->Channel,
4578	    ds->Rate, le16toh(ds->QosCtrl), le16toh(ds->HtSig2));
4579}
4580
4581static void
4582mwl_printtxbuf(const struct mwl_txbuf *bf, u_int qnum, u_int ix)
4583{
4584	const struct mwl_txdesc *ds = bf->bf_desc;
4585	uint32_t status = le32toh(ds->Status);
4586
4587	printf("Q%u[%3u]", qnum, ix);
4588	printf(" (DS.V:%p DS.P:0x%jx)\n", ds, (uintmax_t)bf->bf_daddr);
4589	printf("    NEXT:%08x DATA:%08x LEN:%04x STAT:%08x%s\n",
4590	    le32toh(ds->pPhysNext),
4591	    le32toh(ds->PktPtr), le16toh(ds->PktLen), status,
4592	    status & EAGLE_TXD_STATUS_USED ?
4593		"" : (status & 3) != 0 ? " *" : " !");
4594	printf("    RATE:%02x PRI:%x QOS:%04x SAP:%08x FORMAT:%04x\n",
4595	    ds->DataRate, ds->TxPriority, le16toh(ds->QosCtrl),
4596	    le32toh(ds->SapPktInfo), le16toh(ds->Format));
4597#if MWL_TXDESC > 1
4598	printf("    MULTIFRAMES:%u LEN:%04x %04x %04x %04x %04x %04x\n"
4599	    , le32toh(ds->multiframes)
4600	    , le16toh(ds->PktLenArray[0]), le16toh(ds->PktLenArray[1])
4601	    , le16toh(ds->PktLenArray[2]), le16toh(ds->PktLenArray[3])
4602	    , le16toh(ds->PktLenArray[4]), le16toh(ds->PktLenArray[5])
4603	);
4604	printf("    DATA:%08x %08x %08x %08x %08x %08x\n"
4605	    , le32toh(ds->PktPtrArray[0]), le32toh(ds->PktPtrArray[1])
4606	    , le32toh(ds->PktPtrArray[2]), le32toh(ds->PktPtrArray[3])
4607	    , le32toh(ds->PktPtrArray[4]), le32toh(ds->PktPtrArray[5])
4608	);
4609#endif
4610#if 0
4611{ const uint8_t *cp = (const uint8_t *) ds;
4612  int i;
4613  for (i = 0; i < sizeof(struct mwl_txdesc); i++) {
4614	printf("%02x ", cp[i]);
4615	if (((i+1) % 16) == 0)
4616		printf("\n");
4617  }
4618  printf("\n");
4619}
4620#endif
4621}
4622#endif /* MWL_DEBUG */
4623
4624#if 0
4625static void
4626mwl_txq_dump(struct mwl_txq *txq)
4627{
4628	struct mwl_txbuf *bf;
4629	int i = 0;
4630
4631	MWL_TXQ_LOCK(txq);
4632	STAILQ_FOREACH(bf, &txq->active, bf_list) {
4633		struct mwl_txdesc *ds = bf->bf_desc;
4634		MWL_TXDESC_SYNC(txq, ds,
4635		    BUS_DMASYNC_POSTREAD | BUS_DMASYNC_POSTWRITE);
4636#ifdef MWL_DEBUG
4637		mwl_printtxbuf(bf, txq->qnum, i);
4638#endif
4639		i++;
4640	}
4641	MWL_TXQ_UNLOCK(txq);
4642}
4643#endif
4644
4645static void
4646mwl_watchdog(void *arg)
4647{
4648	struct mwl_softc *sc = arg;
4649
4650	callout_reset(&sc->sc_watchdog, hz, mwl_watchdog, sc);
4651	if (sc->sc_tx_timer == 0 || --sc->sc_tx_timer > 0)
4652		return;
4653
4654	if (sc->sc_running && !sc->sc_invalid) {
4655		if (mwl_hal_setkeepalive(sc->sc_mh))
4656			device_printf(sc->sc_dev,
4657			    "transmit timeout (firmware hung?)\n");
4658		else
4659			device_printf(sc->sc_dev,
4660			    "transmit timeout\n");
4661#if 0
4662		mwl_reset(sc);
4663mwl_txq_dump(&sc->sc_txq[0]);/*XXX*/
4664#endif
4665		counter_u64_add(sc->sc_ic.ic_oerrors, 1);
4666		sc->sc_stats.mst_watchdog++;
4667	}
4668}
4669
4670#ifdef MWL_DIAGAPI
4671/*
4672 * Diagnostic interface to the HAL.  This is used by various
4673 * tools to do things like retrieve register contents for
4674 * debugging.  The mechanism is intentionally opaque so that
4675 * it can change frequently w/o concern for compatiblity.
4676 */
4677static int
4678mwl_ioctl_diag(struct mwl_softc *sc, struct mwl_diag *md)
4679{
4680	struct mwl_hal *mh = sc->sc_mh;
4681	u_int id = md->md_id & MWL_DIAG_ID;
4682	void *indata = NULL;
4683	void *outdata = NULL;
4684	u_int32_t insize = md->md_in_size;
4685	u_int32_t outsize = md->md_out_size;
4686	int error = 0;
4687
4688	if (md->md_id & MWL_DIAG_IN) {
4689		/*
4690		 * Copy in data.
4691		 */
4692		indata = malloc(insize, M_TEMP, M_NOWAIT);
4693		if (indata == NULL) {
4694			error = ENOMEM;
4695			goto bad;
4696		}
4697		error = copyin(md->md_in_data, indata, insize);
4698		if (error)
4699			goto bad;
4700	}
4701	if (md->md_id & MWL_DIAG_DYN) {
4702		/*
4703		 * Allocate a buffer for the results (otherwise the HAL
4704		 * returns a pointer to a buffer where we can read the
4705		 * results).  Note that we depend on the HAL leaving this
4706		 * pointer for us to use below in reclaiming the buffer;
4707		 * may want to be more defensive.
4708		 */
4709		outdata = malloc(outsize, M_TEMP, M_NOWAIT);
4710		if (outdata == NULL) {
4711			error = ENOMEM;
4712			goto bad;
4713		}
4714	}
4715	if (mwl_hal_getdiagstate(mh, id, indata, insize, &outdata, &outsize)) {
4716		if (outsize < md->md_out_size)
4717			md->md_out_size = outsize;
4718		if (outdata != NULL)
4719			error = copyout(outdata, md->md_out_data,
4720					md->md_out_size);
4721	} else {
4722		error = EINVAL;
4723	}
4724bad:
4725	if ((md->md_id & MWL_DIAG_IN) && indata != NULL)
4726		free(indata, M_TEMP);
4727	if ((md->md_id & MWL_DIAG_DYN) && outdata != NULL)
4728		free(outdata, M_TEMP);
4729	return error;
4730}
4731
4732static int
4733mwl_ioctl_reset(struct mwl_softc *sc, struct mwl_diag *md)
4734{
4735	struct mwl_hal *mh = sc->sc_mh;
4736	int error;
4737
4738	MWL_LOCK_ASSERT(sc);
4739
4740	if (md->md_id == 0 && mwl_hal_fwload(mh, NULL) != 0) {
4741		device_printf(sc->sc_dev, "unable to load firmware\n");
4742		return EIO;
4743	}
4744	if (mwl_hal_gethwspecs(mh, &sc->sc_hwspecs) != 0) {
4745		device_printf(sc->sc_dev, "unable to fetch h/w specs\n");
4746		return EIO;
4747	}
4748	error = mwl_setupdma(sc);
4749	if (error != 0) {
4750		/* NB: mwl_setupdma prints a msg */
4751		return error;
4752	}
4753	/*
4754	 * Reset tx/rx data structures; after reload we must
4755	 * re-start the driver's notion of the next xmit/recv.
4756	 */
4757	mwl_draintxq(sc);		/* clear pending frames */
4758	mwl_resettxq(sc);		/* rebuild tx q lists */
4759	sc->sc_rxnext = NULL;		/* force rx to start at the list head */
4760	return 0;
4761}
4762#endif /* MWL_DIAGAPI */
4763
4764static void
4765mwl_parent(struct ieee80211com *ic)
4766{
4767	struct mwl_softc *sc = ic->ic_softc;
4768	int startall = 0;
4769
4770	MWL_LOCK(sc);
4771	if (ic->ic_nrunning > 0) {
4772		if (sc->sc_running) {
4773			/*
4774			 * To avoid rescanning another access point,
4775			 * do not call mwl_init() here.  Instead,
4776			 * only reflect promisc mode settings.
4777			 */
4778			mwl_mode_init(sc);
4779		} else {
4780			/*
4781			 * Beware of being called during attach/detach
4782			 * to reset promiscuous mode.  In that case we
4783			 * will still be marked UP but not RUNNING.
4784			 * However trying to re-init the interface
4785			 * is the wrong thing to do as we've already
4786			 * torn down much of our state.  There's
4787			 * probably a better way to deal with this.
4788			 */
4789			if (!sc->sc_invalid) {
4790				mwl_init(sc);	/* XXX lose error */
4791				startall = 1;
4792			}
4793		}
4794	} else
4795		mwl_stop(sc);
4796	MWL_UNLOCK(sc);
4797	if (startall)
4798		ieee80211_start_all(ic);
4799}
4800
4801static int
4802mwl_ioctl(struct ieee80211com *ic, u_long cmd, void *data)
4803{
4804	struct mwl_softc *sc = ic->ic_softc;
4805	struct ifreq *ifr = data;
4806	int error = 0;
4807
4808	switch (cmd) {
4809	case SIOCGMVSTATS:
4810		mwl_hal_gethwstats(sc->sc_mh, &sc->sc_stats.hw_stats);
4811#if 0
4812		/* NB: embed these numbers to get a consistent view */
4813		sc->sc_stats.mst_tx_packets =
4814		    ifp->if_get_counter(ifp, IFCOUNTER_OPACKETS);
4815		sc->sc_stats.mst_rx_packets =
4816		    ifp->if_get_counter(ifp, IFCOUNTER_IPACKETS);
4817#endif
4818		/*
4819		 * NB: Drop the softc lock in case of a page fault;
4820		 * we'll accept any potential inconsisentcy in the
4821		 * statistics.  The alternative is to copy the data
4822		 * to a local structure.
4823		 */
4824		return (copyout(&sc->sc_stats,
4825				ifr->ifr_data, sizeof (sc->sc_stats)));
4826#ifdef MWL_DIAGAPI
4827	case SIOCGMVDIAG:
4828		/* XXX check privs */
4829		return mwl_ioctl_diag(sc, (struct mwl_diag *) ifr);
4830	case SIOCGMVRESET:
4831		/* XXX check privs */
4832		MWL_LOCK(sc);
4833		error = mwl_ioctl_reset(sc,(struct mwl_diag *) ifr);
4834		MWL_UNLOCK(sc);
4835		break;
4836#endif /* MWL_DIAGAPI */
4837	default:
4838		error = ENOTTY;
4839		break;
4840	}
4841	return (error);
4842}
4843
4844#ifdef	MWL_DEBUG
4845static int
4846mwl_sysctl_debug(SYSCTL_HANDLER_ARGS)
4847{
4848	struct mwl_softc *sc = arg1;
4849	int debug, error;
4850
4851	debug = sc->sc_debug | (mwl_hal_getdebug(sc->sc_mh) << 24);
4852	error = sysctl_handle_int(oidp, &debug, 0, req);
4853	if (error || !req->newptr)
4854		return error;
4855	mwl_hal_setdebug(sc->sc_mh, debug >> 24);
4856	sc->sc_debug = debug & 0x00ffffff;
4857	return 0;
4858}
4859#endif /* MWL_DEBUG */
4860
4861static void
4862mwl_sysctlattach(struct mwl_softc *sc)
4863{
4864#ifdef	MWL_DEBUG
4865	struct sysctl_ctx_list *ctx = device_get_sysctl_ctx(sc->sc_dev);
4866	struct sysctl_oid *tree = device_get_sysctl_tree(sc->sc_dev);
4867
4868	sc->sc_debug = mwl_debug;
4869	SYSCTL_ADD_PROC(ctx, SYSCTL_CHILDREN(tree), OID_AUTO,
4870		"debug", CTLTYPE_INT | CTLFLAG_RW, sc, 0,
4871		mwl_sysctl_debug, "I", "control debugging printfs");
4872#endif
4873}
4874
4875/*
4876 * Announce various information on device/driver attach.
4877 */
4878static void
4879mwl_announce(struct mwl_softc *sc)
4880{
4881
4882	device_printf(sc->sc_dev, "Rev A%d hardware, v%d.%d.%d.%d firmware (regioncode %d)\n",
4883		sc->sc_hwspecs.hwVersion,
4884		(sc->sc_hwspecs.fwReleaseNumber>>24) & 0xff,
4885		(sc->sc_hwspecs.fwReleaseNumber>>16) & 0xff,
4886		(sc->sc_hwspecs.fwReleaseNumber>>8) & 0xff,
4887		(sc->sc_hwspecs.fwReleaseNumber>>0) & 0xff,
4888		sc->sc_hwspecs.regionCode);
4889	sc->sc_fwrelease = sc->sc_hwspecs.fwReleaseNumber;
4890
4891	if (bootverbose) {
4892		int i;
4893		for (i = 0; i <= WME_AC_VO; i++) {
4894			struct mwl_txq *txq = sc->sc_ac2q[i];
4895			device_printf(sc->sc_dev, "Use hw queue %u for %s traffic\n",
4896				txq->qnum, ieee80211_wme_acnames[i]);
4897		}
4898	}
4899	if (bootverbose || mwl_rxdesc != MWL_RXDESC)
4900		device_printf(sc->sc_dev, "using %u rx descriptors\n", mwl_rxdesc);
4901	if (bootverbose || mwl_rxbuf != MWL_RXBUF)
4902		device_printf(sc->sc_dev, "using %u rx buffers\n", mwl_rxbuf);
4903	if (bootverbose || mwl_txbuf != MWL_TXBUF)
4904		device_printf(sc->sc_dev, "using %u tx buffers\n", mwl_txbuf);
4905	if (bootverbose && mwl_hal_ismbsscapable(sc->sc_mh))
4906		device_printf(sc->sc_dev, "multi-bss support\n");
4907#ifdef MWL_TX_NODROP
4908	if (bootverbose)
4909		device_printf(sc->sc_dev, "no tx drop\n");
4910#endif
4911}
4912