if_mwl.c revision 288087
1/*-
2 * Copyright (c) 2007-2009 Sam Leffler, Errno Consulting
3 * Copyright (c) 2007-2008 Marvell Semiconductor, Inc.
4 * All rights reserved.
5 *
6 * Redistribution and use in source and binary forms, with or without
7 * modification, are permitted provided that the following conditions
8 * are met:
9 * 1. Redistributions of source code must retain the above copyright
10 *    notice, this list of conditions and the following disclaimer,
11 *    without modification.
12 * 2. Redistributions in binary form must reproduce at minimum a disclaimer
13 *    similar to the "NO WARRANTY" disclaimer below ("Disclaimer") and any
14 *    redistribution must be conditioned upon including a substantially
15 *    similar Disclaimer requirement for further binary redistribution.
16 *
17 * NO WARRANTY
18 * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
19 * ``AS IS'' AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
20 * LIMITED TO, THE IMPLIED WARRANTIES OF NONINFRINGEMENT, MERCHANTIBILITY
21 * AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL
22 * THE COPYRIGHT HOLDERS OR CONTRIBUTORS BE LIABLE FOR SPECIAL, EXEMPLARY,
23 * OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
24 * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
25 * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER
26 * IN CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
27 * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF
28 * THE POSSIBILITY OF SUCH DAMAGES.
29 */
30
31#include <sys/cdefs.h>
32__FBSDID("$FreeBSD: head/sys/dev/mwl/if_mwl.c 288087 2015-09-22 02:44:59Z adrian $");
33
34/*
35 * Driver for the Marvell 88W8363 Wireless LAN controller.
36 */
37
38#include "opt_inet.h"
39#include "opt_mwl.h"
40#include "opt_wlan.h"
41
42#include <sys/param.h>
43#include <sys/systm.h>
44#include <sys/sysctl.h>
45#include <sys/mbuf.h>
46#include <sys/malloc.h>
47#include <sys/lock.h>
48#include <sys/mutex.h>
49#include <sys/kernel.h>
50#include <sys/socket.h>
51#include <sys/sockio.h>
52#include <sys/errno.h>
53#include <sys/callout.h>
54#include <sys/bus.h>
55#include <sys/endian.h>
56#include <sys/kthread.h>
57#include <sys/taskqueue.h>
58
59#include <machine/bus.h>
60
61#include <net/if.h>
62#include <net/if_var.h>
63#include <net/if_dl.h>
64#include <net/if_media.h>
65#include <net/if_types.h>
66#include <net/if_arp.h>
67#include <net/ethernet.h>
68#include <net/if_llc.h>
69
70#include <net/bpf.h>
71
72#include <net80211/ieee80211_var.h>
73#include <net80211/ieee80211_input.h>
74#include <net80211/ieee80211_regdomain.h>
75
76#ifdef INET
77#include <netinet/in.h>
78#include <netinet/if_ether.h>
79#endif /* INET */
80
81#include <dev/mwl/if_mwlvar.h>
82#include <dev/mwl/mwldiag.h>
83
84/* idiomatic shorthands: MS = mask+shift, SM = shift+mask */
85#define	MS(v,x)	(((v) & x) >> x##_S)
86#define	SM(v,x)	(((v) << x##_S) & x)
87
88static struct ieee80211vap *mwl_vap_create(struct ieee80211com *,
89		    const char [IFNAMSIZ], int, enum ieee80211_opmode, int,
90		    const uint8_t [IEEE80211_ADDR_LEN],
91		    const uint8_t [IEEE80211_ADDR_LEN]);
92static void	mwl_vap_delete(struct ieee80211vap *);
93static int	mwl_setupdma(struct mwl_softc *);
94static int	mwl_hal_reset(struct mwl_softc *sc);
95static int	mwl_init(struct mwl_softc *);
96static void	mwl_parent(struct ieee80211com *);
97static int	mwl_reset(struct ieee80211vap *, u_long);
98static void	mwl_stop(struct mwl_softc *);
99static void	mwl_start(struct mwl_softc *);
100static int	mwl_transmit(struct ieee80211com *, struct mbuf *);
101static int	mwl_raw_xmit(struct ieee80211_node *, struct mbuf *,
102			const struct ieee80211_bpf_params *);
103static int	mwl_media_change(struct ifnet *);
104static void	mwl_watchdog(void *);
105static int	mwl_ioctl(struct ieee80211com *, u_long, void *);
106static void	mwl_radar_proc(void *, int);
107static void	mwl_chanswitch_proc(void *, int);
108static void	mwl_bawatchdog_proc(void *, int);
109static int	mwl_key_alloc(struct ieee80211vap *,
110			struct ieee80211_key *,
111			ieee80211_keyix *, ieee80211_keyix *);
112static int	mwl_key_delete(struct ieee80211vap *,
113			const struct ieee80211_key *);
114static int	mwl_key_set(struct ieee80211vap *, const struct ieee80211_key *,
115			const uint8_t mac[IEEE80211_ADDR_LEN]);
116static int	mwl_mode_init(struct mwl_softc *);
117static void	mwl_update_mcast(struct ieee80211com *);
118static void	mwl_update_promisc(struct ieee80211com *);
119static void	mwl_updateslot(struct ieee80211com *);
120static int	mwl_beacon_setup(struct ieee80211vap *);
121static void	mwl_beacon_update(struct ieee80211vap *, int);
122#ifdef MWL_HOST_PS_SUPPORT
123static void	mwl_update_ps(struct ieee80211vap *, int);
124static int	mwl_set_tim(struct ieee80211_node *, int);
125#endif
126static int	mwl_dma_setup(struct mwl_softc *);
127static void	mwl_dma_cleanup(struct mwl_softc *);
128static struct ieee80211_node *mwl_node_alloc(struct ieee80211vap *,
129		    const uint8_t [IEEE80211_ADDR_LEN]);
130static void	mwl_node_cleanup(struct ieee80211_node *);
131static void	mwl_node_drain(struct ieee80211_node *);
132static void	mwl_node_getsignal(const struct ieee80211_node *,
133			int8_t *, int8_t *);
134static void	mwl_node_getmimoinfo(const struct ieee80211_node *,
135			struct ieee80211_mimo_info *);
136static int	mwl_rxbuf_init(struct mwl_softc *, struct mwl_rxbuf *);
137static void	mwl_rx_proc(void *, int);
138static void	mwl_txq_init(struct mwl_softc *sc, struct mwl_txq *, int);
139static int	mwl_tx_setup(struct mwl_softc *, int, int);
140static int	mwl_wme_update(struct ieee80211com *);
141static void	mwl_tx_cleanupq(struct mwl_softc *, struct mwl_txq *);
142static void	mwl_tx_cleanup(struct mwl_softc *);
143static uint16_t	mwl_calcformat(uint8_t rate, const struct ieee80211_node *);
144static int	mwl_tx_start(struct mwl_softc *, struct ieee80211_node *,
145			     struct mwl_txbuf *, struct mbuf *);
146static void	mwl_tx_proc(void *, int);
147static int	mwl_chan_set(struct mwl_softc *, struct ieee80211_channel *);
148static void	mwl_draintxq(struct mwl_softc *);
149static void	mwl_cleartxq(struct mwl_softc *, struct ieee80211vap *);
150static int	mwl_recv_action(struct ieee80211_node *,
151			const struct ieee80211_frame *,
152			const uint8_t *, const uint8_t *);
153static int	mwl_addba_request(struct ieee80211_node *,
154			struct ieee80211_tx_ampdu *, int dialogtoken,
155			int baparamset, int batimeout);
156static int	mwl_addba_response(struct ieee80211_node *,
157			struct ieee80211_tx_ampdu *, int status,
158			int baparamset, int batimeout);
159static void	mwl_addba_stop(struct ieee80211_node *,
160			struct ieee80211_tx_ampdu *);
161static int	mwl_startrecv(struct mwl_softc *);
162static MWL_HAL_APMODE mwl_getapmode(const struct ieee80211vap *,
163			struct ieee80211_channel *);
164static int	mwl_setapmode(struct ieee80211vap *, struct ieee80211_channel*);
165static void	mwl_scan_start(struct ieee80211com *);
166static void	mwl_scan_end(struct ieee80211com *);
167static void	mwl_set_channel(struct ieee80211com *);
168static int	mwl_peerstadb(struct ieee80211_node *,
169			int aid, int staid, MWL_HAL_PEERINFO *pi);
170static int	mwl_localstadb(struct ieee80211vap *);
171static int	mwl_newstate(struct ieee80211vap *, enum ieee80211_state, int);
172static int	allocstaid(struct mwl_softc *sc, int aid);
173static void	delstaid(struct mwl_softc *sc, int staid);
174static void	mwl_newassoc(struct ieee80211_node *, int);
175static void	mwl_agestations(void *);
176static int	mwl_setregdomain(struct ieee80211com *,
177			struct ieee80211_regdomain *, int,
178			struct ieee80211_channel []);
179static void	mwl_getradiocaps(struct ieee80211com *, int, int *,
180			struct ieee80211_channel []);
181static int	mwl_getchannels(struct mwl_softc *);
182
183static void	mwl_sysctlattach(struct mwl_softc *);
184static void	mwl_announce(struct mwl_softc *);
185
186SYSCTL_NODE(_hw, OID_AUTO, mwl, CTLFLAG_RD, 0, "Marvell driver parameters");
187
188static	int mwl_rxdesc = MWL_RXDESC;		/* # rx desc's to allocate */
189SYSCTL_INT(_hw_mwl, OID_AUTO, rxdesc, CTLFLAG_RW, &mwl_rxdesc,
190	    0, "rx descriptors allocated");
191static	int mwl_rxbuf = MWL_RXBUF;		/* # rx buffers to allocate */
192SYSCTL_INT(_hw_mwl, OID_AUTO, rxbuf, CTLFLAG_RWTUN, &mwl_rxbuf,
193	    0, "rx buffers allocated");
194static	int mwl_txbuf = MWL_TXBUF;		/* # tx buffers to allocate */
195SYSCTL_INT(_hw_mwl, OID_AUTO, txbuf, CTLFLAG_RWTUN, &mwl_txbuf,
196	    0, "tx buffers allocated");
197static	int mwl_txcoalesce = 8;		/* # tx packets to q before poking f/w*/
198SYSCTL_INT(_hw_mwl, OID_AUTO, txcoalesce, CTLFLAG_RWTUN, &mwl_txcoalesce,
199	    0, "tx buffers to send at once");
200static	int mwl_rxquota = MWL_RXBUF;		/* # max buffers to process */
201SYSCTL_INT(_hw_mwl, OID_AUTO, rxquota, CTLFLAG_RWTUN, &mwl_rxquota,
202	    0, "max rx buffers to process per interrupt");
203static	int mwl_rxdmalow = 3;			/* # min buffers for wakeup */
204SYSCTL_INT(_hw_mwl, OID_AUTO, rxdmalow, CTLFLAG_RWTUN, &mwl_rxdmalow,
205	    0, "min free rx buffers before restarting traffic");
206
207#ifdef MWL_DEBUG
208static	int mwl_debug = 0;
209SYSCTL_INT(_hw_mwl, OID_AUTO, debug, CTLFLAG_RWTUN, &mwl_debug,
210	    0, "control debugging printfs");
211enum {
212	MWL_DEBUG_XMIT		= 0x00000001,	/* basic xmit operation */
213	MWL_DEBUG_XMIT_DESC	= 0x00000002,	/* xmit descriptors */
214	MWL_DEBUG_RECV		= 0x00000004,	/* basic recv operation */
215	MWL_DEBUG_RECV_DESC	= 0x00000008,	/* recv descriptors */
216	MWL_DEBUG_RESET		= 0x00000010,	/* reset processing */
217	MWL_DEBUG_BEACON 	= 0x00000020,	/* beacon handling */
218	MWL_DEBUG_INTR		= 0x00000040,	/* ISR */
219	MWL_DEBUG_TX_PROC	= 0x00000080,	/* tx ISR proc */
220	MWL_DEBUG_RX_PROC	= 0x00000100,	/* rx ISR proc */
221	MWL_DEBUG_KEYCACHE	= 0x00000200,	/* key cache management */
222	MWL_DEBUG_STATE		= 0x00000400,	/* 802.11 state transitions */
223	MWL_DEBUG_NODE		= 0x00000800,	/* node management */
224	MWL_DEBUG_RECV_ALL	= 0x00001000,	/* trace all frames (beacons) */
225	MWL_DEBUG_TSO		= 0x00002000,	/* TSO processing */
226	MWL_DEBUG_AMPDU		= 0x00004000,	/* BA stream handling */
227	MWL_DEBUG_ANY		= 0xffffffff
228};
229#define	IS_BEACON(wh) \
230    ((wh->i_fc[0] & (IEEE80211_FC0_TYPE_MASK|IEEE80211_FC0_SUBTYPE_MASK)) == \
231	 (IEEE80211_FC0_TYPE_MGT|IEEE80211_FC0_SUBTYPE_BEACON))
232#define	IFF_DUMPPKTS_RECV(sc, wh) \
233    ((sc->sc_debug & MWL_DEBUG_RECV) && \
234      ((sc->sc_debug & MWL_DEBUG_RECV_ALL) || !IS_BEACON(wh)))
235#define	IFF_DUMPPKTS_XMIT(sc) \
236	(sc->sc_debug & MWL_DEBUG_XMIT)
237
238#define	DPRINTF(sc, m, fmt, ...) do {				\
239	if (sc->sc_debug & (m))					\
240		printf(fmt, __VA_ARGS__);			\
241} while (0)
242#define	KEYPRINTF(sc, hk, mac) do {				\
243	if (sc->sc_debug & MWL_DEBUG_KEYCACHE)			\
244		mwl_keyprint(sc, __func__, hk, mac);		\
245} while (0)
246static	void mwl_printrxbuf(const struct mwl_rxbuf *bf, u_int ix);
247static	void mwl_printtxbuf(const struct mwl_txbuf *bf, u_int qnum, u_int ix);
248#else
249#define	IFF_DUMPPKTS_RECV(sc, wh)	0
250#define	IFF_DUMPPKTS_XMIT(sc)		0
251#define	DPRINTF(sc, m, fmt, ...)	do { (void )sc; } while (0)
252#define	KEYPRINTF(sc, k, mac)		do { (void )sc; } while (0)
253#endif
254
255static MALLOC_DEFINE(M_MWLDEV, "mwldev", "mwl driver dma buffers");
256
257/*
258 * Each packet has fixed front matter: a 2-byte length
259 * of the payload, followed by a 4-address 802.11 header
260 * (regardless of the actual header and always w/o any
261 * QoS header).  The payload then follows.
262 */
263struct mwltxrec {
264	uint16_t fwlen;
265	struct ieee80211_frame_addr4 wh;
266} __packed;
267
268/*
269 * Read/Write shorthands for accesses to BAR 0.  Note
270 * that all BAR 1 operations are done in the "hal" and
271 * there should be no reference to them here.
272 */
273#ifdef MWL_DEBUG
274static __inline uint32_t
275RD4(struct mwl_softc *sc, bus_size_t off)
276{
277	return bus_space_read_4(sc->sc_io0t, sc->sc_io0h, off);
278}
279#endif
280
281static __inline void
282WR4(struct mwl_softc *sc, bus_size_t off, uint32_t val)
283{
284	bus_space_write_4(sc->sc_io0t, sc->sc_io0h, off, val);
285}
286
287int
288mwl_attach(uint16_t devid, struct mwl_softc *sc)
289{
290	struct ieee80211com *ic = &sc->sc_ic;
291	struct mwl_hal *mh;
292	int error = 0;
293
294	DPRINTF(sc, MWL_DEBUG_ANY, "%s: devid 0x%x\n", __func__, devid);
295
296	/*
297	 * Setup the RX free list lock early, so it can be consistently
298	 * removed.
299	 */
300	MWL_RXFREE_INIT(sc);
301
302	mh = mwl_hal_attach(sc->sc_dev, devid,
303	    sc->sc_io1h, sc->sc_io1t, sc->sc_dmat);
304	if (mh == NULL) {
305		device_printf(sc->sc_dev, "unable to attach HAL\n");
306		error = EIO;
307		goto bad;
308	}
309	sc->sc_mh = mh;
310	/*
311	 * Load firmware so we can get setup.  We arbitrarily
312	 * pick station firmware; we'll re-load firmware as
313	 * needed so setting up the wrong mode isn't a big deal.
314	 */
315	if (mwl_hal_fwload(mh, NULL) != 0) {
316		device_printf(sc->sc_dev, "unable to setup builtin firmware\n");
317		error = EIO;
318		goto bad1;
319	}
320	if (mwl_hal_gethwspecs(mh, &sc->sc_hwspecs) != 0) {
321		device_printf(sc->sc_dev, "unable to fetch h/w specs\n");
322		error = EIO;
323		goto bad1;
324	}
325	error = mwl_getchannels(sc);
326	if (error != 0)
327		goto bad1;
328
329	sc->sc_txantenna = 0;		/* h/w default */
330	sc->sc_rxantenna = 0;		/* h/w default */
331	sc->sc_invalid = 0;		/* ready to go, enable int handling */
332	sc->sc_ageinterval = MWL_AGEINTERVAL;
333
334	/*
335	 * Allocate tx+rx descriptors and populate the lists.
336	 * We immediately push the information to the firmware
337	 * as otherwise it gets upset.
338	 */
339	error = mwl_dma_setup(sc);
340	if (error != 0) {
341		device_printf(sc->sc_dev, "failed to setup descriptors: %d\n",
342		    error);
343		goto bad1;
344	}
345	error = mwl_setupdma(sc);	/* push to firmware */
346	if (error != 0)			/* NB: mwl_setupdma prints msg */
347		goto bad1;
348
349	callout_init(&sc->sc_timer, 1);
350	callout_init_mtx(&sc->sc_watchdog, &sc->sc_mtx, 0);
351	mbufq_init(&sc->sc_snd, ifqmaxlen);
352
353	sc->sc_tq = taskqueue_create("mwl_taskq", M_NOWAIT,
354		taskqueue_thread_enqueue, &sc->sc_tq);
355	taskqueue_start_threads(&sc->sc_tq, 1, PI_NET,
356		"%s taskq", device_get_nameunit(sc->sc_dev));
357
358	TASK_INIT(&sc->sc_rxtask, 0, mwl_rx_proc, sc);
359	TASK_INIT(&sc->sc_radartask, 0, mwl_radar_proc, sc);
360	TASK_INIT(&sc->sc_chanswitchtask, 0, mwl_chanswitch_proc, sc);
361	TASK_INIT(&sc->sc_bawatchdogtask, 0, mwl_bawatchdog_proc, sc);
362
363	/* NB: insure BK queue is the lowest priority h/w queue */
364	if (!mwl_tx_setup(sc, WME_AC_BK, MWL_WME_AC_BK)) {
365		device_printf(sc->sc_dev,
366		    "unable to setup xmit queue for %s traffic!\n",
367		     ieee80211_wme_acnames[WME_AC_BK]);
368		error = EIO;
369		goto bad2;
370	}
371	if (!mwl_tx_setup(sc, WME_AC_BE, MWL_WME_AC_BE) ||
372	    !mwl_tx_setup(sc, WME_AC_VI, MWL_WME_AC_VI) ||
373	    !mwl_tx_setup(sc, WME_AC_VO, MWL_WME_AC_VO)) {
374		/*
375		 * Not enough hardware tx queues to properly do WME;
376		 * just punt and assign them all to the same h/w queue.
377		 * We could do a better job of this if, for example,
378		 * we allocate queues when we switch from station to
379		 * AP mode.
380		 */
381		if (sc->sc_ac2q[WME_AC_VI] != NULL)
382			mwl_tx_cleanupq(sc, sc->sc_ac2q[WME_AC_VI]);
383		if (sc->sc_ac2q[WME_AC_BE] != NULL)
384			mwl_tx_cleanupq(sc, sc->sc_ac2q[WME_AC_BE]);
385		sc->sc_ac2q[WME_AC_BE] = sc->sc_ac2q[WME_AC_BK];
386		sc->sc_ac2q[WME_AC_VI] = sc->sc_ac2q[WME_AC_BK];
387		sc->sc_ac2q[WME_AC_VO] = sc->sc_ac2q[WME_AC_BK];
388	}
389	TASK_INIT(&sc->sc_txtask, 0, mwl_tx_proc, sc);
390
391	ic->ic_softc = sc;
392	ic->ic_name = device_get_nameunit(sc->sc_dev);
393	/* XXX not right but it's not used anywhere important */
394	ic->ic_phytype = IEEE80211_T_OFDM;
395	ic->ic_opmode = IEEE80211_M_STA;
396	ic->ic_caps =
397		  IEEE80211_C_STA		/* station mode supported */
398		| IEEE80211_C_HOSTAP		/* hostap mode */
399		| IEEE80211_C_MONITOR		/* monitor mode */
400#if 0
401		| IEEE80211_C_IBSS		/* ibss, nee adhoc, mode */
402		| IEEE80211_C_AHDEMO		/* adhoc demo mode */
403#endif
404		| IEEE80211_C_MBSS		/* mesh point link mode */
405		| IEEE80211_C_WDS		/* WDS supported */
406		| IEEE80211_C_SHPREAMBLE	/* short preamble supported */
407		| IEEE80211_C_SHSLOT		/* short slot time supported */
408		| IEEE80211_C_WME		/* WME/WMM supported */
409		| IEEE80211_C_BURST		/* xmit bursting supported */
410		| IEEE80211_C_WPA		/* capable of WPA1+WPA2 */
411		| IEEE80211_C_BGSCAN		/* capable of bg scanning */
412		| IEEE80211_C_TXFRAG		/* handle tx frags */
413		| IEEE80211_C_TXPMGT		/* capable of txpow mgt */
414		| IEEE80211_C_DFS		/* DFS supported */
415		;
416
417	ic->ic_htcaps =
418		  IEEE80211_HTCAP_SMPS_ENA	/* SM PS mode enabled */
419		| IEEE80211_HTCAP_CHWIDTH40	/* 40MHz channel width */
420		| IEEE80211_HTCAP_SHORTGI20	/* short GI in 20MHz */
421		| IEEE80211_HTCAP_SHORTGI40	/* short GI in 40MHz */
422		| IEEE80211_HTCAP_RXSTBC_2STREAM/* 1-2 spatial streams */
423#if MWL_AGGR_SIZE == 7935
424		| IEEE80211_HTCAP_MAXAMSDU_7935	/* max A-MSDU length */
425#else
426		| IEEE80211_HTCAP_MAXAMSDU_3839	/* max A-MSDU length */
427#endif
428#if 0
429		| IEEE80211_HTCAP_PSMP		/* PSMP supported */
430		| IEEE80211_HTCAP_40INTOLERANT	/* 40MHz intolerant */
431#endif
432		/* s/w capabilities */
433		| IEEE80211_HTC_HT		/* HT operation */
434		| IEEE80211_HTC_AMPDU		/* tx A-MPDU */
435		| IEEE80211_HTC_AMSDU		/* tx A-MSDU */
436		| IEEE80211_HTC_SMPS		/* SMPS available */
437		;
438
439	/*
440	 * Mark h/w crypto support.
441	 * XXX no way to query h/w support.
442	 */
443	ic->ic_cryptocaps |= IEEE80211_CRYPTO_WEP
444			  |  IEEE80211_CRYPTO_AES_CCM
445			  |  IEEE80211_CRYPTO_TKIP
446			  |  IEEE80211_CRYPTO_TKIPMIC
447			  ;
448	/*
449	 * Transmit requires space in the packet for a special
450	 * format transmit record and optional padding between
451	 * this record and the payload.  Ask the net80211 layer
452	 * to arrange this when encapsulating packets so we can
453	 * add it efficiently.
454	 */
455	ic->ic_headroom = sizeof(struct mwltxrec) -
456		sizeof(struct ieee80211_frame);
457
458	IEEE80211_ADDR_COPY(ic->ic_macaddr, sc->sc_hwspecs.macAddr);
459
460	/* call MI attach routine. */
461	ieee80211_ifattach(ic);
462	ic->ic_setregdomain = mwl_setregdomain;
463	ic->ic_getradiocaps = mwl_getradiocaps;
464	/* override default methods */
465	ic->ic_raw_xmit = mwl_raw_xmit;
466	ic->ic_newassoc = mwl_newassoc;
467	ic->ic_updateslot = mwl_updateslot;
468	ic->ic_update_mcast = mwl_update_mcast;
469	ic->ic_update_promisc = mwl_update_promisc;
470	ic->ic_wme.wme_update = mwl_wme_update;
471	ic->ic_transmit = mwl_transmit;
472	ic->ic_ioctl = mwl_ioctl;
473	ic->ic_parent = mwl_parent;
474
475	ic->ic_node_alloc = mwl_node_alloc;
476	sc->sc_node_cleanup = ic->ic_node_cleanup;
477	ic->ic_node_cleanup = mwl_node_cleanup;
478	sc->sc_node_drain = ic->ic_node_drain;
479	ic->ic_node_drain = mwl_node_drain;
480	ic->ic_node_getsignal = mwl_node_getsignal;
481	ic->ic_node_getmimoinfo = mwl_node_getmimoinfo;
482
483	ic->ic_scan_start = mwl_scan_start;
484	ic->ic_scan_end = mwl_scan_end;
485	ic->ic_set_channel = mwl_set_channel;
486
487	sc->sc_recv_action = ic->ic_recv_action;
488	ic->ic_recv_action = mwl_recv_action;
489	sc->sc_addba_request = ic->ic_addba_request;
490	ic->ic_addba_request = mwl_addba_request;
491	sc->sc_addba_response = ic->ic_addba_response;
492	ic->ic_addba_response = mwl_addba_response;
493	sc->sc_addba_stop = ic->ic_addba_stop;
494	ic->ic_addba_stop = mwl_addba_stop;
495
496	ic->ic_vap_create = mwl_vap_create;
497	ic->ic_vap_delete = mwl_vap_delete;
498
499	ieee80211_radiotap_attach(ic,
500	    &sc->sc_tx_th.wt_ihdr, sizeof(sc->sc_tx_th),
501		MWL_TX_RADIOTAP_PRESENT,
502	    &sc->sc_rx_th.wr_ihdr, sizeof(sc->sc_rx_th),
503		MWL_RX_RADIOTAP_PRESENT);
504	/*
505	 * Setup dynamic sysctl's now that country code and
506	 * regdomain are available from the hal.
507	 */
508	mwl_sysctlattach(sc);
509
510	if (bootverbose)
511		ieee80211_announce(ic);
512	mwl_announce(sc);
513	return 0;
514bad2:
515	mwl_dma_cleanup(sc);
516bad1:
517	mwl_hal_detach(mh);
518bad:
519	MWL_RXFREE_DESTROY(sc);
520	sc->sc_invalid = 1;
521	return error;
522}
523
524int
525mwl_detach(struct mwl_softc *sc)
526{
527	struct ieee80211com *ic = &sc->sc_ic;
528
529	MWL_LOCK(sc);
530	mwl_stop(sc);
531	MWL_UNLOCK(sc);
532	/*
533	 * NB: the order of these is important:
534	 * o call the 802.11 layer before detaching the hal to
535	 *   insure callbacks into the driver to delete global
536	 *   key cache entries can be handled
537	 * o reclaim the tx queue data structures after calling
538	 *   the 802.11 layer as we'll get called back to reclaim
539	 *   node state and potentially want to use them
540	 * o to cleanup the tx queues the hal is called, so detach
541	 *   it last
542	 * Other than that, it's straightforward...
543	 */
544	ieee80211_ifdetach(ic);
545	callout_drain(&sc->sc_watchdog);
546	mwl_dma_cleanup(sc);
547	MWL_RXFREE_DESTROY(sc);
548	mwl_tx_cleanup(sc);
549	mwl_hal_detach(sc->sc_mh);
550	mbufq_drain(&sc->sc_snd);
551
552	return 0;
553}
554
555/*
556 * MAC address handling for multiple BSS on the same radio.
557 * The first vap uses the MAC address from the EEPROM.  For
558 * subsequent vap's we set the U/L bit (bit 1) in the MAC
559 * address and use the next six bits as an index.
560 */
561static void
562assign_address(struct mwl_softc *sc, uint8_t mac[IEEE80211_ADDR_LEN], int clone)
563{
564	int i;
565
566	if (clone && mwl_hal_ismbsscapable(sc->sc_mh)) {
567		/* NB: we only do this if h/w supports multiple bssid */
568		for (i = 0; i < 32; i++)
569			if ((sc->sc_bssidmask & (1<<i)) == 0)
570				break;
571		if (i != 0)
572			mac[0] |= (i << 2)|0x2;
573	} else
574		i = 0;
575	sc->sc_bssidmask |= 1<<i;
576	if (i == 0)
577		sc->sc_nbssid0++;
578}
579
580static void
581reclaim_address(struct mwl_softc *sc, const uint8_t mac[IEEE80211_ADDR_LEN])
582{
583	int i = mac[0] >> 2;
584	if (i != 0 || --sc->sc_nbssid0 == 0)
585		sc->sc_bssidmask &= ~(1<<i);
586}
587
588static struct ieee80211vap *
589mwl_vap_create(struct ieee80211com *ic, const char name[IFNAMSIZ], int unit,
590    enum ieee80211_opmode opmode, int flags,
591    const uint8_t bssid[IEEE80211_ADDR_LEN],
592    const uint8_t mac0[IEEE80211_ADDR_LEN])
593{
594	struct mwl_softc *sc = ic->ic_softc;
595	struct mwl_hal *mh = sc->sc_mh;
596	struct ieee80211vap *vap, *apvap;
597	struct mwl_hal_vap *hvap;
598	struct mwl_vap *mvp;
599	uint8_t mac[IEEE80211_ADDR_LEN];
600
601	IEEE80211_ADDR_COPY(mac, mac0);
602	switch (opmode) {
603	case IEEE80211_M_HOSTAP:
604	case IEEE80211_M_MBSS:
605		if ((flags & IEEE80211_CLONE_MACADDR) == 0)
606			assign_address(sc, mac, flags & IEEE80211_CLONE_BSSID);
607		hvap = mwl_hal_newvap(mh, MWL_HAL_AP, mac);
608		if (hvap == NULL) {
609			if ((flags & IEEE80211_CLONE_MACADDR) == 0)
610				reclaim_address(sc, mac);
611			return NULL;
612		}
613		break;
614	case IEEE80211_M_STA:
615		if ((flags & IEEE80211_CLONE_MACADDR) == 0)
616			assign_address(sc, mac, flags & IEEE80211_CLONE_BSSID);
617		hvap = mwl_hal_newvap(mh, MWL_HAL_STA, mac);
618		if (hvap == NULL) {
619			if ((flags & IEEE80211_CLONE_MACADDR) == 0)
620				reclaim_address(sc, mac);
621			return NULL;
622		}
623		/* no h/w beacon miss support; always use s/w */
624		flags |= IEEE80211_CLONE_NOBEACONS;
625		break;
626	case IEEE80211_M_WDS:
627		hvap = NULL;		/* NB: we use associated AP vap */
628		if (sc->sc_napvaps == 0)
629			return NULL;	/* no existing AP vap */
630		break;
631	case IEEE80211_M_MONITOR:
632		hvap = NULL;
633		break;
634	case IEEE80211_M_IBSS:
635	case IEEE80211_M_AHDEMO:
636	default:
637		return NULL;
638	}
639
640	mvp = malloc(sizeof(struct mwl_vap), M_80211_VAP, M_WAITOK | M_ZERO);
641	mvp->mv_hvap = hvap;
642	if (opmode == IEEE80211_M_WDS) {
643		/*
644		 * WDS vaps must have an associated AP vap; find one.
645		 * XXX not right.
646		 */
647		TAILQ_FOREACH(apvap, &ic->ic_vaps, iv_next)
648			if (apvap->iv_opmode == IEEE80211_M_HOSTAP) {
649				mvp->mv_ap_hvap = MWL_VAP(apvap)->mv_hvap;
650				break;
651			}
652		KASSERT(mvp->mv_ap_hvap != NULL, ("no ap vap"));
653	}
654	vap = &mvp->mv_vap;
655	ieee80211_vap_setup(ic, vap, name, unit, opmode, flags, bssid);
656	/* override with driver methods */
657	mvp->mv_newstate = vap->iv_newstate;
658	vap->iv_newstate = mwl_newstate;
659	vap->iv_max_keyix = 0;	/* XXX */
660	vap->iv_key_alloc = mwl_key_alloc;
661	vap->iv_key_delete = mwl_key_delete;
662	vap->iv_key_set = mwl_key_set;
663#ifdef MWL_HOST_PS_SUPPORT
664	if (opmode == IEEE80211_M_HOSTAP || opmode == IEEE80211_M_MBSS) {
665		vap->iv_update_ps = mwl_update_ps;
666		mvp->mv_set_tim = vap->iv_set_tim;
667		vap->iv_set_tim = mwl_set_tim;
668	}
669#endif
670	vap->iv_reset = mwl_reset;
671	vap->iv_update_beacon = mwl_beacon_update;
672
673	/* override max aid so sta's cannot assoc when we're out of sta id's */
674	vap->iv_max_aid = MWL_MAXSTAID;
675	/* override default A-MPDU rx parameters */
676	vap->iv_ampdu_rxmax = IEEE80211_HTCAP_MAXRXAMPDU_64K;
677	vap->iv_ampdu_density = IEEE80211_HTCAP_MPDUDENSITY_4;
678
679	/* complete setup */
680	ieee80211_vap_attach(vap, mwl_media_change, ieee80211_media_status,
681	    mac);
682
683	switch (vap->iv_opmode) {
684	case IEEE80211_M_HOSTAP:
685	case IEEE80211_M_MBSS:
686	case IEEE80211_M_STA:
687		/*
688		 * Setup sta db entry for local address.
689		 */
690		mwl_localstadb(vap);
691		if (vap->iv_opmode == IEEE80211_M_HOSTAP ||
692		    vap->iv_opmode == IEEE80211_M_MBSS)
693			sc->sc_napvaps++;
694		else
695			sc->sc_nstavaps++;
696		break;
697	case IEEE80211_M_WDS:
698		sc->sc_nwdsvaps++;
699		break;
700	default:
701		break;
702	}
703	/*
704	 * Setup overall operating mode.
705	 */
706	if (sc->sc_napvaps)
707		ic->ic_opmode = IEEE80211_M_HOSTAP;
708	else if (sc->sc_nstavaps)
709		ic->ic_opmode = IEEE80211_M_STA;
710	else
711		ic->ic_opmode = opmode;
712
713	return vap;
714}
715
716static void
717mwl_vap_delete(struct ieee80211vap *vap)
718{
719	struct mwl_vap *mvp = MWL_VAP(vap);
720	struct mwl_softc *sc = vap->iv_ic->ic_softc;
721	struct mwl_hal *mh = sc->sc_mh;
722	struct mwl_hal_vap *hvap = mvp->mv_hvap;
723	enum ieee80211_opmode opmode = vap->iv_opmode;
724
725	/* XXX disallow ap vap delete if WDS still present */
726	if (sc->sc_running) {
727		/* quiesce h/w while we remove the vap */
728		mwl_hal_intrset(mh, 0);		/* disable interrupts */
729	}
730	ieee80211_vap_detach(vap);
731	switch (opmode) {
732	case IEEE80211_M_HOSTAP:
733	case IEEE80211_M_MBSS:
734	case IEEE80211_M_STA:
735		KASSERT(hvap != NULL, ("no hal vap handle"));
736		(void) mwl_hal_delstation(hvap, vap->iv_myaddr);
737		mwl_hal_delvap(hvap);
738		if (opmode == IEEE80211_M_HOSTAP || opmode == IEEE80211_M_MBSS)
739			sc->sc_napvaps--;
740		else
741			sc->sc_nstavaps--;
742		/* XXX don't do it for IEEE80211_CLONE_MACADDR */
743		reclaim_address(sc, vap->iv_myaddr);
744		break;
745	case IEEE80211_M_WDS:
746		sc->sc_nwdsvaps--;
747		break;
748	default:
749		break;
750	}
751	mwl_cleartxq(sc, vap);
752	free(mvp, M_80211_VAP);
753	if (sc->sc_running)
754		mwl_hal_intrset(mh, sc->sc_imask);
755}
756
757void
758mwl_suspend(struct mwl_softc *sc)
759{
760
761	MWL_LOCK(sc);
762	mwl_stop(sc);
763	MWL_UNLOCK(sc);
764}
765
766void
767mwl_resume(struct mwl_softc *sc)
768{
769	int error = EDOOFUS;
770
771	MWL_LOCK(sc);
772	if (sc->sc_ic.ic_nrunning > 0)
773		error = mwl_init(sc);
774	MWL_UNLOCK(sc);
775
776	if (error == 0)
777		ieee80211_start_all(&sc->sc_ic);	/* start all vap's */
778}
779
780void
781mwl_shutdown(void *arg)
782{
783	struct mwl_softc *sc = arg;
784
785	MWL_LOCK(sc);
786	mwl_stop(sc);
787	MWL_UNLOCK(sc);
788}
789
790/*
791 * Interrupt handler.  Most of the actual processing is deferred.
792 */
793void
794mwl_intr(void *arg)
795{
796	struct mwl_softc *sc = arg;
797	struct mwl_hal *mh = sc->sc_mh;
798	uint32_t status;
799
800	if (sc->sc_invalid) {
801		/*
802		 * The hardware is not ready/present, don't touch anything.
803		 * Note this can happen early on if the IRQ is shared.
804		 */
805		DPRINTF(sc, MWL_DEBUG_ANY, "%s: invalid; ignored\n", __func__);
806		return;
807	}
808	/*
809	 * Figure out the reason(s) for the interrupt.
810	 */
811	mwl_hal_getisr(mh, &status);		/* NB: clears ISR too */
812	if (status == 0)			/* must be a shared irq */
813		return;
814
815	DPRINTF(sc, MWL_DEBUG_INTR, "%s: status 0x%x imask 0x%x\n",
816	    __func__, status, sc->sc_imask);
817	if (status & MACREG_A2HRIC_BIT_RX_RDY)
818		taskqueue_enqueue(sc->sc_tq, &sc->sc_rxtask);
819	if (status & MACREG_A2HRIC_BIT_TX_DONE)
820		taskqueue_enqueue(sc->sc_tq, &sc->sc_txtask);
821	if (status & MACREG_A2HRIC_BIT_BA_WATCHDOG)
822		taskqueue_enqueue(sc->sc_tq, &sc->sc_bawatchdogtask);
823	if (status & MACREG_A2HRIC_BIT_OPC_DONE)
824		mwl_hal_cmddone(mh);
825	if (status & MACREG_A2HRIC_BIT_MAC_EVENT) {
826		;
827	}
828	if (status & MACREG_A2HRIC_BIT_ICV_ERROR) {
829		/* TKIP ICV error */
830		sc->sc_stats.mst_rx_badtkipicv++;
831	}
832	if (status & MACREG_A2HRIC_BIT_QUEUE_EMPTY) {
833		/* 11n aggregation queue is empty, re-fill */
834		;
835	}
836	if (status & MACREG_A2HRIC_BIT_QUEUE_FULL) {
837		;
838	}
839	if (status & MACREG_A2HRIC_BIT_RADAR_DETECT) {
840		/* radar detected, process event */
841		taskqueue_enqueue(sc->sc_tq, &sc->sc_radartask);
842	}
843	if (status & MACREG_A2HRIC_BIT_CHAN_SWITCH) {
844		/* DFS channel switch */
845		taskqueue_enqueue(sc->sc_tq, &sc->sc_chanswitchtask);
846	}
847}
848
849static void
850mwl_radar_proc(void *arg, int pending)
851{
852	struct mwl_softc *sc = arg;
853	struct ieee80211com *ic = &sc->sc_ic;
854
855	DPRINTF(sc, MWL_DEBUG_ANY, "%s: radar detected, pending %u\n",
856	    __func__, pending);
857
858	sc->sc_stats.mst_radardetect++;
859	/* XXX stop h/w BA streams? */
860
861	IEEE80211_LOCK(ic);
862	ieee80211_dfs_notify_radar(ic, ic->ic_curchan);
863	IEEE80211_UNLOCK(ic);
864}
865
866static void
867mwl_chanswitch_proc(void *arg, int pending)
868{
869	struct mwl_softc *sc = arg;
870	struct ieee80211com *ic = &sc->sc_ic;
871
872	DPRINTF(sc, MWL_DEBUG_ANY, "%s: channel switch notice, pending %u\n",
873	    __func__, pending);
874
875	IEEE80211_LOCK(ic);
876	sc->sc_csapending = 0;
877	ieee80211_csa_completeswitch(ic);
878	IEEE80211_UNLOCK(ic);
879}
880
881static void
882mwl_bawatchdog(const MWL_HAL_BASTREAM *sp)
883{
884	struct ieee80211_node *ni = sp->data[0];
885
886	/* send DELBA and drop the stream */
887	ieee80211_ampdu_stop(ni, sp->data[1], IEEE80211_REASON_UNSPECIFIED);
888}
889
890static void
891mwl_bawatchdog_proc(void *arg, int pending)
892{
893	struct mwl_softc *sc = arg;
894	struct mwl_hal *mh = sc->sc_mh;
895	const MWL_HAL_BASTREAM *sp;
896	uint8_t bitmap, n;
897
898	sc->sc_stats.mst_bawatchdog++;
899
900	if (mwl_hal_getwatchdogbitmap(mh, &bitmap) != 0) {
901		DPRINTF(sc, MWL_DEBUG_AMPDU,
902		    "%s: could not get bitmap\n", __func__);
903		sc->sc_stats.mst_bawatchdog_failed++;
904		return;
905	}
906	DPRINTF(sc, MWL_DEBUG_AMPDU, "%s: bitmap 0x%x\n", __func__, bitmap);
907	if (bitmap == 0xff) {
908		n = 0;
909		/* disable all ba streams */
910		for (bitmap = 0; bitmap < 8; bitmap++) {
911			sp = mwl_hal_bastream_lookup(mh, bitmap);
912			if (sp != NULL) {
913				mwl_bawatchdog(sp);
914				n++;
915			}
916		}
917		if (n == 0) {
918			DPRINTF(sc, MWL_DEBUG_AMPDU,
919			    "%s: no BA streams found\n", __func__);
920			sc->sc_stats.mst_bawatchdog_empty++;
921		}
922	} else if (bitmap != 0xaa) {
923		/* disable a single ba stream */
924		sp = mwl_hal_bastream_lookup(mh, bitmap);
925		if (sp != NULL) {
926			mwl_bawatchdog(sp);
927		} else {
928			DPRINTF(sc, MWL_DEBUG_AMPDU,
929			    "%s: no BA stream %d\n", __func__, bitmap);
930			sc->sc_stats.mst_bawatchdog_notfound++;
931		}
932	}
933}
934
935/*
936 * Convert net80211 channel to a HAL channel.
937 */
938static void
939mwl_mapchan(MWL_HAL_CHANNEL *hc, const struct ieee80211_channel *chan)
940{
941	hc->channel = chan->ic_ieee;
942
943	*(uint32_t *)&hc->channelFlags = 0;
944	if (IEEE80211_IS_CHAN_2GHZ(chan))
945		hc->channelFlags.FreqBand = MWL_FREQ_BAND_2DOT4GHZ;
946	else if (IEEE80211_IS_CHAN_5GHZ(chan))
947		hc->channelFlags.FreqBand = MWL_FREQ_BAND_5GHZ;
948	if (IEEE80211_IS_CHAN_HT40(chan)) {
949		hc->channelFlags.ChnlWidth = MWL_CH_40_MHz_WIDTH;
950		if (IEEE80211_IS_CHAN_HT40U(chan))
951			hc->channelFlags.ExtChnlOffset = MWL_EXT_CH_ABOVE_CTRL_CH;
952		else
953			hc->channelFlags.ExtChnlOffset = MWL_EXT_CH_BELOW_CTRL_CH;
954	} else
955		hc->channelFlags.ChnlWidth = MWL_CH_20_MHz_WIDTH;
956	/* XXX 10MHz channels */
957}
958
959/*
960 * Inform firmware of our tx/rx dma setup.  The BAR 0
961 * writes below are for compatibility with older firmware.
962 * For current firmware we send this information with a
963 * cmd block via mwl_hal_sethwdma.
964 */
965static int
966mwl_setupdma(struct mwl_softc *sc)
967{
968	int error, i;
969
970	sc->sc_hwdma.rxDescRead = sc->sc_rxdma.dd_desc_paddr;
971	WR4(sc, sc->sc_hwspecs.rxDescRead, sc->sc_hwdma.rxDescRead);
972	WR4(sc, sc->sc_hwspecs.rxDescWrite, sc->sc_hwdma.rxDescRead);
973
974	for (i = 0; i < MWL_NUM_TX_QUEUES-MWL_NUM_ACK_QUEUES; i++) {
975		struct mwl_txq *txq = &sc->sc_txq[i];
976		sc->sc_hwdma.wcbBase[i] = txq->dma.dd_desc_paddr;
977		WR4(sc, sc->sc_hwspecs.wcbBase[i], sc->sc_hwdma.wcbBase[i]);
978	}
979	sc->sc_hwdma.maxNumTxWcb = mwl_txbuf;
980	sc->sc_hwdma.maxNumWCB = MWL_NUM_TX_QUEUES-MWL_NUM_ACK_QUEUES;
981
982	error = mwl_hal_sethwdma(sc->sc_mh, &sc->sc_hwdma);
983	if (error != 0) {
984		device_printf(sc->sc_dev,
985		    "unable to setup tx/rx dma; hal status %u\n", error);
986		/* XXX */
987	}
988	return error;
989}
990
991/*
992 * Inform firmware of tx rate parameters.
993 * Called after a channel change.
994 */
995static int
996mwl_setcurchanrates(struct mwl_softc *sc)
997{
998	struct ieee80211com *ic = &sc->sc_ic;
999	const struct ieee80211_rateset *rs;
1000	MWL_HAL_TXRATE rates;
1001
1002	memset(&rates, 0, sizeof(rates));
1003	rs = ieee80211_get_suprates(ic, ic->ic_curchan);
1004	/* rate used to send management frames */
1005	rates.MgtRate = rs->rs_rates[0] & IEEE80211_RATE_VAL;
1006	/* rate used to send multicast frames */
1007	rates.McastRate = rates.MgtRate;
1008
1009	return mwl_hal_settxrate_auto(sc->sc_mh, &rates);
1010}
1011
1012/*
1013 * Inform firmware of tx rate parameters.  Called whenever
1014 * user-settable params change and after a channel change.
1015 */
1016static int
1017mwl_setrates(struct ieee80211vap *vap)
1018{
1019	struct mwl_vap *mvp = MWL_VAP(vap);
1020	struct ieee80211_node *ni = vap->iv_bss;
1021	const struct ieee80211_txparam *tp = ni->ni_txparms;
1022	MWL_HAL_TXRATE rates;
1023
1024	KASSERT(vap->iv_state == IEEE80211_S_RUN, ("state %d", vap->iv_state));
1025
1026	/*
1027	 * Update the h/w rate map.
1028	 * NB: 0x80 for MCS is passed through unchanged
1029	 */
1030	memset(&rates, 0, sizeof(rates));
1031	/* rate used to send management frames */
1032	rates.MgtRate = tp->mgmtrate;
1033	/* rate used to send multicast frames */
1034	rates.McastRate = tp->mcastrate;
1035
1036	/* while here calculate EAPOL fixed rate cookie */
1037	mvp->mv_eapolformat = htole16(mwl_calcformat(rates.MgtRate, ni));
1038
1039	return mwl_hal_settxrate(mvp->mv_hvap,
1040	    tp->ucastrate != IEEE80211_FIXED_RATE_NONE ?
1041		RATE_FIXED : RATE_AUTO, &rates);
1042}
1043
1044/*
1045 * Setup a fixed xmit rate cookie for EAPOL frames.
1046 */
1047static void
1048mwl_seteapolformat(struct ieee80211vap *vap)
1049{
1050	struct mwl_vap *mvp = MWL_VAP(vap);
1051	struct ieee80211_node *ni = vap->iv_bss;
1052	enum ieee80211_phymode mode;
1053	uint8_t rate;
1054
1055	KASSERT(vap->iv_state == IEEE80211_S_RUN, ("state %d", vap->iv_state));
1056
1057	mode = ieee80211_chan2mode(ni->ni_chan);
1058	/*
1059	 * Use legacy rates when operating a mixed HT+non-HT bss.
1060	 * NB: this may violate POLA for sta and wds vap's.
1061	 */
1062	if (mode == IEEE80211_MODE_11NA &&
1063	    (vap->iv_flags_ht & IEEE80211_FHT_PUREN) == 0)
1064		rate = vap->iv_txparms[IEEE80211_MODE_11A].mgmtrate;
1065	else if (mode == IEEE80211_MODE_11NG &&
1066	    (vap->iv_flags_ht & IEEE80211_FHT_PUREN) == 0)
1067		rate = vap->iv_txparms[IEEE80211_MODE_11G].mgmtrate;
1068	else
1069		rate = vap->iv_txparms[mode].mgmtrate;
1070
1071	mvp->mv_eapolformat = htole16(mwl_calcformat(rate, ni));
1072}
1073
1074/*
1075 * Map SKU+country code to region code for radar bin'ing.
1076 */
1077static int
1078mwl_map2regioncode(const struct ieee80211_regdomain *rd)
1079{
1080	switch (rd->regdomain) {
1081	case SKU_FCC:
1082	case SKU_FCC3:
1083		return DOMAIN_CODE_FCC;
1084	case SKU_CA:
1085		return DOMAIN_CODE_IC;
1086	case SKU_ETSI:
1087	case SKU_ETSI2:
1088	case SKU_ETSI3:
1089		if (rd->country == CTRY_SPAIN)
1090			return DOMAIN_CODE_SPAIN;
1091		if (rd->country == CTRY_FRANCE || rd->country == CTRY_FRANCE2)
1092			return DOMAIN_CODE_FRANCE;
1093		/* XXX force 1.3.1 radar type */
1094		return DOMAIN_CODE_ETSI_131;
1095	case SKU_JAPAN:
1096		return DOMAIN_CODE_MKK;
1097	case SKU_ROW:
1098		return DOMAIN_CODE_DGT;	/* Taiwan */
1099	case SKU_APAC:
1100	case SKU_APAC2:
1101	case SKU_APAC3:
1102		return DOMAIN_CODE_AUS;	/* Australia */
1103	}
1104	/* XXX KOREA? */
1105	return DOMAIN_CODE_FCC;			/* XXX? */
1106}
1107
1108static int
1109mwl_hal_reset(struct mwl_softc *sc)
1110{
1111	struct ieee80211com *ic = &sc->sc_ic;
1112	struct mwl_hal *mh = sc->sc_mh;
1113
1114	mwl_hal_setantenna(mh, WL_ANTENNATYPE_RX, sc->sc_rxantenna);
1115	mwl_hal_setantenna(mh, WL_ANTENNATYPE_TX, sc->sc_txantenna);
1116	mwl_hal_setradio(mh, 1, WL_AUTO_PREAMBLE);
1117	mwl_hal_setwmm(sc->sc_mh, (ic->ic_flags & IEEE80211_F_WME) != 0);
1118	mwl_chan_set(sc, ic->ic_curchan);
1119	/* NB: RF/RA performance tuned for indoor mode */
1120	mwl_hal_setrateadaptmode(mh, 0);
1121	mwl_hal_setoptimizationlevel(mh,
1122	    (ic->ic_flags & IEEE80211_F_BURST) != 0);
1123
1124	mwl_hal_setregioncode(mh, mwl_map2regioncode(&ic->ic_regdomain));
1125
1126	mwl_hal_setaggampduratemode(mh, 1, 80);		/* XXX */
1127	mwl_hal_setcfend(mh, 0);			/* XXX */
1128
1129	return 1;
1130}
1131
1132static int
1133mwl_init(struct mwl_softc *sc)
1134{
1135	struct mwl_hal *mh = sc->sc_mh;
1136	int error = 0;
1137
1138	MWL_LOCK_ASSERT(sc);
1139
1140	/*
1141	 * Stop anything previously setup.  This is safe
1142	 * whether this is the first time through or not.
1143	 */
1144	mwl_stop(sc);
1145
1146	/*
1147	 * Push vap-independent state to the firmware.
1148	 */
1149	if (!mwl_hal_reset(sc)) {
1150		device_printf(sc->sc_dev, "unable to reset hardware\n");
1151		return EIO;
1152	}
1153
1154	/*
1155	 * Setup recv (once); transmit is already good to go.
1156	 */
1157	error = mwl_startrecv(sc);
1158	if (error != 0) {
1159		device_printf(sc->sc_dev, "unable to start recv logic\n");
1160		return error;
1161	}
1162
1163	/*
1164	 * Enable interrupts.
1165	 */
1166	sc->sc_imask = MACREG_A2HRIC_BIT_RX_RDY
1167		     | MACREG_A2HRIC_BIT_TX_DONE
1168		     | MACREG_A2HRIC_BIT_OPC_DONE
1169#if 0
1170		     | MACREG_A2HRIC_BIT_MAC_EVENT
1171#endif
1172		     | MACREG_A2HRIC_BIT_ICV_ERROR
1173		     | MACREG_A2HRIC_BIT_RADAR_DETECT
1174		     | MACREG_A2HRIC_BIT_CHAN_SWITCH
1175#if 0
1176		     | MACREG_A2HRIC_BIT_QUEUE_EMPTY
1177#endif
1178		     | MACREG_A2HRIC_BIT_BA_WATCHDOG
1179		     | MACREQ_A2HRIC_BIT_TX_ACK
1180		     ;
1181
1182	sc->sc_running = 1;
1183	mwl_hal_intrset(mh, sc->sc_imask);
1184	callout_reset(&sc->sc_watchdog, hz, mwl_watchdog, sc);
1185
1186	return 0;
1187}
1188
1189static void
1190mwl_stop(struct mwl_softc *sc)
1191{
1192
1193	MWL_LOCK_ASSERT(sc);
1194	if (sc->sc_running) {
1195		/*
1196		 * Shutdown the hardware and driver.
1197		 */
1198		sc->sc_running = 0;
1199		callout_stop(&sc->sc_watchdog);
1200		sc->sc_tx_timer = 0;
1201		mwl_draintxq(sc);
1202	}
1203}
1204
1205static int
1206mwl_reset_vap(struct ieee80211vap *vap, int state)
1207{
1208	struct mwl_hal_vap *hvap = MWL_VAP(vap)->mv_hvap;
1209	struct ieee80211com *ic = vap->iv_ic;
1210
1211	if (state == IEEE80211_S_RUN)
1212		mwl_setrates(vap);
1213	/* XXX off by 1? */
1214	mwl_hal_setrtsthreshold(hvap, vap->iv_rtsthreshold);
1215	/* XXX auto? 20/40 split? */
1216	mwl_hal_sethtgi(hvap, (vap->iv_flags_ht &
1217	    (IEEE80211_FHT_SHORTGI20|IEEE80211_FHT_SHORTGI40)) ? 1 : 0);
1218	mwl_hal_setnprot(hvap, ic->ic_htprotmode == IEEE80211_PROT_NONE ?
1219	    HTPROTECT_NONE : HTPROTECT_AUTO);
1220	/* XXX txpower cap */
1221
1222	/* re-setup beacons */
1223	if (state == IEEE80211_S_RUN &&
1224	    (vap->iv_opmode == IEEE80211_M_HOSTAP ||
1225	     vap->iv_opmode == IEEE80211_M_MBSS ||
1226	     vap->iv_opmode == IEEE80211_M_IBSS)) {
1227		mwl_setapmode(vap, vap->iv_bss->ni_chan);
1228		mwl_hal_setnprotmode(hvap,
1229		    MS(ic->ic_curhtprotmode, IEEE80211_HTINFO_OPMODE));
1230		return mwl_beacon_setup(vap);
1231	}
1232	return 0;
1233}
1234
1235/*
1236 * Reset the hardware w/o losing operational state.
1237 * Used to to reset or reload hardware state for a vap.
1238 */
1239static int
1240mwl_reset(struct ieee80211vap *vap, u_long cmd)
1241{
1242	struct mwl_hal_vap *hvap = MWL_VAP(vap)->mv_hvap;
1243	int error = 0;
1244
1245	if (hvap != NULL) {			/* WDS, MONITOR, etc. */
1246		struct ieee80211com *ic = vap->iv_ic;
1247		struct mwl_softc *sc = ic->ic_softc;
1248		struct mwl_hal *mh = sc->sc_mh;
1249
1250		/* XXX handle DWDS sta vap change */
1251		/* XXX do we need to disable interrupts? */
1252		mwl_hal_intrset(mh, 0);		/* disable interrupts */
1253		error = mwl_reset_vap(vap, vap->iv_state);
1254		mwl_hal_intrset(mh, sc->sc_imask);
1255	}
1256	return error;
1257}
1258
1259/*
1260 * Allocate a tx buffer for sending a frame.  The
1261 * packet is assumed to have the WME AC stored so
1262 * we can use it to select the appropriate h/w queue.
1263 */
1264static struct mwl_txbuf *
1265mwl_gettxbuf(struct mwl_softc *sc, struct mwl_txq *txq)
1266{
1267	struct mwl_txbuf *bf;
1268
1269	/*
1270	 * Grab a TX buffer and associated resources.
1271	 */
1272	MWL_TXQ_LOCK(txq);
1273	bf = STAILQ_FIRST(&txq->free);
1274	if (bf != NULL) {
1275		STAILQ_REMOVE_HEAD(&txq->free, bf_list);
1276		txq->nfree--;
1277	}
1278	MWL_TXQ_UNLOCK(txq);
1279	if (bf == NULL)
1280		DPRINTF(sc, MWL_DEBUG_XMIT,
1281		    "%s: out of xmit buffers on q %d\n", __func__, txq->qnum);
1282	return bf;
1283}
1284
1285/*
1286 * Return a tx buffer to the queue it came from.  Note there
1287 * are two cases because we must preserve the order of buffers
1288 * as it reflects the fixed order of descriptors in memory
1289 * (the firmware pre-fetches descriptors so we cannot reorder).
1290 */
1291static void
1292mwl_puttxbuf_head(struct mwl_txq *txq, struct mwl_txbuf *bf)
1293{
1294	bf->bf_m = NULL;
1295	bf->bf_node = NULL;
1296	MWL_TXQ_LOCK(txq);
1297	STAILQ_INSERT_HEAD(&txq->free, bf, bf_list);
1298	txq->nfree++;
1299	MWL_TXQ_UNLOCK(txq);
1300}
1301
1302static void
1303mwl_puttxbuf_tail(struct mwl_txq *txq, struct mwl_txbuf *bf)
1304{
1305	bf->bf_m = NULL;
1306	bf->bf_node = NULL;
1307	MWL_TXQ_LOCK(txq);
1308	STAILQ_INSERT_TAIL(&txq->free, bf, bf_list);
1309	txq->nfree++;
1310	MWL_TXQ_UNLOCK(txq);
1311}
1312
1313static int
1314mwl_transmit(struct ieee80211com *ic, struct mbuf *m)
1315{
1316	struct mwl_softc *sc = ic->ic_softc;
1317	int error;
1318
1319	MWL_LOCK(sc);
1320	if (!sc->sc_running) {
1321		MWL_UNLOCK(sc);
1322		return (ENXIO);
1323	}
1324	error = mbufq_enqueue(&sc->sc_snd, m);
1325	if (error) {
1326		MWL_UNLOCK(sc);
1327		return (error);
1328	}
1329	mwl_start(sc);
1330	MWL_UNLOCK(sc);
1331	return (0);
1332}
1333
1334static void
1335mwl_start(struct mwl_softc *sc)
1336{
1337	struct ieee80211_node *ni;
1338	struct mwl_txbuf *bf;
1339	struct mbuf *m;
1340	struct mwl_txq *txq = NULL;	/* XXX silence gcc */
1341	int nqueued;
1342
1343	MWL_LOCK_ASSERT(sc);
1344	if (!sc->sc_running || sc->sc_invalid)
1345		return;
1346	nqueued = 0;
1347	while ((m = mbufq_dequeue(&sc->sc_snd)) != NULL) {
1348		/*
1349		 * Grab the node for the destination.
1350		 */
1351		ni = (struct ieee80211_node *) m->m_pkthdr.rcvif;
1352		KASSERT(ni != NULL, ("no node"));
1353		m->m_pkthdr.rcvif = NULL;	/* committed, clear ref */
1354		/*
1355		 * Grab a TX buffer and associated resources.
1356		 * We honor the classification by the 802.11 layer.
1357		 */
1358		txq = sc->sc_ac2q[M_WME_GETAC(m)];
1359		bf = mwl_gettxbuf(sc, txq);
1360		if (bf == NULL) {
1361			m_freem(m);
1362			ieee80211_free_node(ni);
1363#ifdef MWL_TX_NODROP
1364			sc->sc_stats.mst_tx_qstop++;
1365			break;
1366#else
1367			DPRINTF(sc, MWL_DEBUG_XMIT,
1368			    "%s: tail drop on q %d\n", __func__, txq->qnum);
1369			sc->sc_stats.mst_tx_qdrop++;
1370			continue;
1371#endif /* MWL_TX_NODROP */
1372		}
1373
1374		/*
1375		 * Pass the frame to the h/w for transmission.
1376		 */
1377		if (mwl_tx_start(sc, ni, bf, m)) {
1378			if_inc_counter(ni->ni_vap->iv_ifp,
1379			    IFCOUNTER_OERRORS, 1);
1380			mwl_puttxbuf_head(txq, bf);
1381			ieee80211_free_node(ni);
1382			continue;
1383		}
1384		nqueued++;
1385		if (nqueued >= mwl_txcoalesce) {
1386			/*
1387			 * Poke the firmware to process queued frames;
1388			 * see below about (lack of) locking.
1389			 */
1390			nqueued = 0;
1391			mwl_hal_txstart(sc->sc_mh, 0/*XXX*/);
1392		}
1393	}
1394	if (nqueued) {
1395		/*
1396		 * NB: We don't need to lock against tx done because
1397		 * this just prods the firmware to check the transmit
1398		 * descriptors.  The firmware will also start fetching
1399		 * descriptors by itself if it notices new ones are
1400		 * present when it goes to deliver a tx done interrupt
1401		 * to the host. So if we race with tx done processing
1402		 * it's ok.  Delivering the kick here rather than in
1403		 * mwl_tx_start is an optimization to avoid poking the
1404		 * firmware for each packet.
1405		 *
1406		 * NB: the queue id isn't used so 0 is ok.
1407		 */
1408		mwl_hal_txstart(sc->sc_mh, 0/*XXX*/);
1409	}
1410}
1411
1412static int
1413mwl_raw_xmit(struct ieee80211_node *ni, struct mbuf *m,
1414	const struct ieee80211_bpf_params *params)
1415{
1416	struct ieee80211com *ic = ni->ni_ic;
1417	struct mwl_softc *sc = ic->ic_softc;
1418	struct mwl_txbuf *bf;
1419	struct mwl_txq *txq;
1420
1421	if (!sc->sc_running || sc->sc_invalid) {
1422		ieee80211_free_node(ni);
1423		m_freem(m);
1424		return ENETDOWN;
1425	}
1426	/*
1427	 * Grab a TX buffer and associated resources.
1428	 * Note that we depend on the classification
1429	 * by the 802.11 layer to get to the right h/w
1430	 * queue.  Management frames must ALWAYS go on
1431	 * queue 1 but we cannot just force that here
1432	 * because we may receive non-mgt frames.
1433	 */
1434	txq = sc->sc_ac2q[M_WME_GETAC(m)];
1435	bf = mwl_gettxbuf(sc, txq);
1436	if (bf == NULL) {
1437		sc->sc_stats.mst_tx_qstop++;
1438		ieee80211_free_node(ni);
1439		m_freem(m);
1440		return ENOBUFS;
1441	}
1442	/*
1443	 * Pass the frame to the h/w for transmission.
1444	 */
1445	if (mwl_tx_start(sc, ni, bf, m)) {
1446		mwl_puttxbuf_head(txq, bf);
1447
1448		ieee80211_free_node(ni);
1449		return EIO;		/* XXX */
1450	}
1451	/*
1452	 * NB: We don't need to lock against tx done because
1453	 * this just prods the firmware to check the transmit
1454	 * descriptors.  The firmware will also start fetching
1455	 * descriptors by itself if it notices new ones are
1456	 * present when it goes to deliver a tx done interrupt
1457	 * to the host. So if we race with tx done processing
1458	 * it's ok.  Delivering the kick here rather than in
1459	 * mwl_tx_start is an optimization to avoid poking the
1460	 * firmware for each packet.
1461	 *
1462	 * NB: the queue id isn't used so 0 is ok.
1463	 */
1464	mwl_hal_txstart(sc->sc_mh, 0/*XXX*/);
1465	return 0;
1466}
1467
1468static int
1469mwl_media_change(struct ifnet *ifp)
1470{
1471	struct ieee80211vap *vap = ifp->if_softc;
1472	int error;
1473
1474	error = ieee80211_media_change(ifp);
1475	/* NB: only the fixed rate can change and that doesn't need a reset */
1476	if (error == ENETRESET) {
1477		mwl_setrates(vap);
1478		error = 0;
1479	}
1480	return error;
1481}
1482
1483#ifdef MWL_DEBUG
1484static void
1485mwl_keyprint(struct mwl_softc *sc, const char *tag,
1486	const MWL_HAL_KEYVAL *hk, const uint8_t mac[IEEE80211_ADDR_LEN])
1487{
1488	static const char *ciphers[] = {
1489		"WEP",
1490		"TKIP",
1491		"AES-CCM",
1492	};
1493	int i, n;
1494
1495	printf("%s: [%u] %-7s", tag, hk->keyIndex, ciphers[hk->keyTypeId]);
1496	for (i = 0, n = hk->keyLen; i < n; i++)
1497		printf(" %02x", hk->key.aes[i]);
1498	printf(" mac %s", ether_sprintf(mac));
1499	if (hk->keyTypeId == KEY_TYPE_ID_TKIP) {
1500		printf(" %s", "rxmic");
1501		for (i = 0; i < sizeof(hk->key.tkip.rxMic); i++)
1502			printf(" %02x", hk->key.tkip.rxMic[i]);
1503		printf(" txmic");
1504		for (i = 0; i < sizeof(hk->key.tkip.txMic); i++)
1505			printf(" %02x", hk->key.tkip.txMic[i]);
1506	}
1507	printf(" flags 0x%x\n", hk->keyFlags);
1508}
1509#endif
1510
1511/*
1512 * Allocate a key cache slot for a unicast key.  The
1513 * firmware handles key allocation and every station is
1514 * guaranteed key space so we are always successful.
1515 */
1516static int
1517mwl_key_alloc(struct ieee80211vap *vap, struct ieee80211_key *k,
1518	ieee80211_keyix *keyix, ieee80211_keyix *rxkeyix)
1519{
1520	struct mwl_softc *sc = vap->iv_ic->ic_softc;
1521
1522	if (k->wk_keyix != IEEE80211_KEYIX_NONE ||
1523	    (k->wk_flags & IEEE80211_KEY_GROUP)) {
1524		if (!(&vap->iv_nw_keys[0] <= k &&
1525		      k < &vap->iv_nw_keys[IEEE80211_WEP_NKID])) {
1526			/* should not happen */
1527			DPRINTF(sc, MWL_DEBUG_KEYCACHE,
1528				"%s: bogus group key\n", __func__);
1529			return 0;
1530		}
1531		/* give the caller what they requested */
1532		*keyix = *rxkeyix = k - vap->iv_nw_keys;
1533	} else {
1534		/*
1535		 * Firmware handles key allocation.
1536		 */
1537		*keyix = *rxkeyix = 0;
1538	}
1539	return 1;
1540}
1541
1542/*
1543 * Delete a key entry allocated by mwl_key_alloc.
1544 */
1545static int
1546mwl_key_delete(struct ieee80211vap *vap, const struct ieee80211_key *k)
1547{
1548	struct mwl_softc *sc = vap->iv_ic->ic_softc;
1549	struct mwl_hal_vap *hvap = MWL_VAP(vap)->mv_hvap;
1550	MWL_HAL_KEYVAL hk;
1551	const uint8_t bcastaddr[IEEE80211_ADDR_LEN] =
1552	    { 0xff, 0xff, 0xff, 0xff, 0xff, 0xff };
1553
1554	if (hvap == NULL) {
1555		if (vap->iv_opmode != IEEE80211_M_WDS) {
1556			/* XXX monitor mode? */
1557			DPRINTF(sc, MWL_DEBUG_KEYCACHE,
1558			    "%s: no hvap for opmode %d\n", __func__,
1559			    vap->iv_opmode);
1560			return 0;
1561		}
1562		hvap = MWL_VAP(vap)->mv_ap_hvap;
1563	}
1564
1565	DPRINTF(sc, MWL_DEBUG_KEYCACHE, "%s: delete key %u\n",
1566	    __func__, k->wk_keyix);
1567
1568	memset(&hk, 0, sizeof(hk));
1569	hk.keyIndex = k->wk_keyix;
1570	switch (k->wk_cipher->ic_cipher) {
1571	case IEEE80211_CIPHER_WEP:
1572		hk.keyTypeId = KEY_TYPE_ID_WEP;
1573		break;
1574	case IEEE80211_CIPHER_TKIP:
1575		hk.keyTypeId = KEY_TYPE_ID_TKIP;
1576		break;
1577	case IEEE80211_CIPHER_AES_CCM:
1578		hk.keyTypeId = KEY_TYPE_ID_AES;
1579		break;
1580	default:
1581		/* XXX should not happen */
1582		DPRINTF(sc, MWL_DEBUG_KEYCACHE, "%s: unknown cipher %d\n",
1583		    __func__, k->wk_cipher->ic_cipher);
1584		return 0;
1585	}
1586	return (mwl_hal_keyreset(hvap, &hk, bcastaddr) == 0);	/*XXX*/
1587}
1588
1589static __inline int
1590addgroupflags(MWL_HAL_KEYVAL *hk, const struct ieee80211_key *k)
1591{
1592	if (k->wk_flags & IEEE80211_KEY_GROUP) {
1593		if (k->wk_flags & IEEE80211_KEY_XMIT)
1594			hk->keyFlags |= KEY_FLAG_TXGROUPKEY;
1595		if (k->wk_flags & IEEE80211_KEY_RECV)
1596			hk->keyFlags |= KEY_FLAG_RXGROUPKEY;
1597		return 1;
1598	} else
1599		return 0;
1600}
1601
1602/*
1603 * Set the key cache contents for the specified key.  Key cache
1604 * slot(s) must already have been allocated by mwl_key_alloc.
1605 */
1606static int
1607mwl_key_set(struct ieee80211vap *vap, const struct ieee80211_key *k,
1608	const uint8_t mac[IEEE80211_ADDR_LEN])
1609{
1610#define	GRPXMIT	(IEEE80211_KEY_XMIT | IEEE80211_KEY_GROUP)
1611/* NB: static wep keys are marked GROUP+tx/rx; GTK will be tx or rx */
1612#define	IEEE80211_IS_STATICKEY(k) \
1613	(((k)->wk_flags & (GRPXMIT|IEEE80211_KEY_RECV)) == \
1614	 (GRPXMIT|IEEE80211_KEY_RECV))
1615	struct mwl_softc *sc = vap->iv_ic->ic_softc;
1616	struct mwl_hal_vap *hvap = MWL_VAP(vap)->mv_hvap;
1617	const struct ieee80211_cipher *cip = k->wk_cipher;
1618	const uint8_t *macaddr;
1619	MWL_HAL_KEYVAL hk;
1620
1621	KASSERT((k->wk_flags & IEEE80211_KEY_SWCRYPT) == 0,
1622		("s/w crypto set?"));
1623
1624	if (hvap == NULL) {
1625		if (vap->iv_opmode != IEEE80211_M_WDS) {
1626			/* XXX monitor mode? */
1627			DPRINTF(sc, MWL_DEBUG_KEYCACHE,
1628			    "%s: no hvap for opmode %d\n", __func__,
1629			    vap->iv_opmode);
1630			return 0;
1631		}
1632		hvap = MWL_VAP(vap)->mv_ap_hvap;
1633	}
1634	memset(&hk, 0, sizeof(hk));
1635	hk.keyIndex = k->wk_keyix;
1636	switch (cip->ic_cipher) {
1637	case IEEE80211_CIPHER_WEP:
1638		hk.keyTypeId = KEY_TYPE_ID_WEP;
1639		hk.keyLen = k->wk_keylen;
1640		if (k->wk_keyix == vap->iv_def_txkey)
1641			hk.keyFlags = KEY_FLAG_WEP_TXKEY;
1642		if (!IEEE80211_IS_STATICKEY(k)) {
1643			/* NB: WEP is never used for the PTK */
1644			(void) addgroupflags(&hk, k);
1645		}
1646		break;
1647	case IEEE80211_CIPHER_TKIP:
1648		hk.keyTypeId = KEY_TYPE_ID_TKIP;
1649		hk.key.tkip.tsc.high = (uint32_t)(k->wk_keytsc >> 16);
1650		hk.key.tkip.tsc.low = (uint16_t)k->wk_keytsc;
1651		hk.keyFlags = KEY_FLAG_TSC_VALID | KEY_FLAG_MICKEY_VALID;
1652		hk.keyLen = k->wk_keylen + IEEE80211_MICBUF_SIZE;
1653		if (!addgroupflags(&hk, k))
1654			hk.keyFlags |= KEY_FLAG_PAIRWISE;
1655		break;
1656	case IEEE80211_CIPHER_AES_CCM:
1657		hk.keyTypeId = KEY_TYPE_ID_AES;
1658		hk.keyLen = k->wk_keylen;
1659		if (!addgroupflags(&hk, k))
1660			hk.keyFlags |= KEY_FLAG_PAIRWISE;
1661		break;
1662	default:
1663		/* XXX should not happen */
1664		DPRINTF(sc, MWL_DEBUG_KEYCACHE, "%s: unknown cipher %d\n",
1665		    __func__, k->wk_cipher->ic_cipher);
1666		return 0;
1667	}
1668	/*
1669	 * NB: tkip mic keys get copied here too; the layout
1670	 *     just happens to match that in ieee80211_key.
1671	 */
1672	memcpy(hk.key.aes, k->wk_key, hk.keyLen);
1673
1674	/*
1675	 * Locate address of sta db entry for writing key;
1676	 * the convention unfortunately is somewhat different
1677	 * than how net80211, hostapd, and wpa_supplicant think.
1678	 */
1679	if (vap->iv_opmode == IEEE80211_M_STA) {
1680		/*
1681		 * NB: keys plumbed before the sta reaches AUTH state
1682		 * will be discarded or written to the wrong sta db
1683		 * entry because iv_bss is meaningless.  This is ok
1684		 * (right now) because we handle deferred plumbing of
1685		 * WEP keys when the sta reaches AUTH state.
1686		 */
1687		macaddr = vap->iv_bss->ni_bssid;
1688		if ((k->wk_flags & IEEE80211_KEY_GROUP) == 0) {
1689			/* XXX plumb to local sta db too for static key wep */
1690			mwl_hal_keyset(hvap, &hk, vap->iv_myaddr);
1691		}
1692	} else if (vap->iv_opmode == IEEE80211_M_WDS &&
1693	    vap->iv_state != IEEE80211_S_RUN) {
1694		/*
1695		 * Prior to RUN state a WDS vap will not it's BSS node
1696		 * setup so we will plumb the key to the wrong mac
1697		 * address (it'll be our local address).  Workaround
1698		 * this for the moment by grabbing the correct address.
1699		 */
1700		macaddr = vap->iv_des_bssid;
1701	} else if ((k->wk_flags & GRPXMIT) == GRPXMIT)
1702		macaddr = vap->iv_myaddr;
1703	else
1704		macaddr = mac;
1705	KEYPRINTF(sc, &hk, macaddr);
1706	return (mwl_hal_keyset(hvap, &hk, macaddr) == 0);
1707#undef IEEE80211_IS_STATICKEY
1708#undef GRPXMIT
1709}
1710
1711/*
1712 * Set the multicast filter contents into the hardware.
1713 * XXX f/w has no support; just defer to the os.
1714 */
1715static void
1716mwl_setmcastfilter(struct mwl_softc *sc)
1717{
1718#if 0
1719	struct ether_multi *enm;
1720	struct ether_multistep estep;
1721	uint8_t macs[IEEE80211_ADDR_LEN*MWL_HAL_MCAST_MAX];/* XXX stack use */
1722	uint8_t *mp;
1723	int nmc;
1724
1725	mp = macs;
1726	nmc = 0;
1727	ETHER_FIRST_MULTI(estep, &sc->sc_ec, enm);
1728	while (enm != NULL) {
1729		/* XXX Punt on ranges. */
1730		if (nmc == MWL_HAL_MCAST_MAX ||
1731		    !IEEE80211_ADDR_EQ(enm->enm_addrlo, enm->enm_addrhi)) {
1732			ifp->if_flags |= IFF_ALLMULTI;
1733			return;
1734		}
1735		IEEE80211_ADDR_COPY(mp, enm->enm_addrlo);
1736		mp += IEEE80211_ADDR_LEN, nmc++;
1737		ETHER_NEXT_MULTI(estep, enm);
1738	}
1739	ifp->if_flags &= ~IFF_ALLMULTI;
1740	mwl_hal_setmcast(sc->sc_mh, nmc, macs);
1741#endif
1742}
1743
1744static int
1745mwl_mode_init(struct mwl_softc *sc)
1746{
1747	struct ieee80211com *ic = &sc->sc_ic;
1748	struct mwl_hal *mh = sc->sc_mh;
1749
1750	/*
1751	 * NB: Ignore promisc in hostap mode; it's set by the
1752	 * bridge.  This is wrong but we have no way to
1753	 * identify internal requests (from the bridge)
1754	 * versus external requests such as for tcpdump.
1755	 */
1756	mwl_hal_setpromisc(mh, ic->ic_promisc > 0 &&
1757	    ic->ic_opmode != IEEE80211_M_HOSTAP);
1758	mwl_setmcastfilter(sc);
1759
1760	return 0;
1761}
1762
1763/*
1764 * Callback from the 802.11 layer after a multicast state change.
1765 */
1766static void
1767mwl_update_mcast(struct ieee80211com *ic)
1768{
1769	struct mwl_softc *sc = ic->ic_softc;
1770
1771	mwl_setmcastfilter(sc);
1772}
1773
1774/*
1775 * Callback from the 802.11 layer after a promiscuous mode change.
1776 * Note this interface does not check the operating mode as this
1777 * is an internal callback and we are expected to honor the current
1778 * state (e.g. this is used for setting the interface in promiscuous
1779 * mode when operating in hostap mode to do ACS).
1780 */
1781static void
1782mwl_update_promisc(struct ieee80211com *ic)
1783{
1784	struct mwl_softc *sc = ic->ic_softc;
1785
1786	mwl_hal_setpromisc(sc->sc_mh, ic->ic_promisc > 0);
1787}
1788
1789/*
1790 * Callback from the 802.11 layer to update the slot time
1791 * based on the current setting.  We use it to notify the
1792 * firmware of ERP changes and the f/w takes care of things
1793 * like slot time and preamble.
1794 */
1795static void
1796mwl_updateslot(struct ieee80211com *ic)
1797{
1798	struct mwl_softc *sc = ic->ic_softc;
1799	struct mwl_hal *mh = sc->sc_mh;
1800	int prot;
1801
1802	/* NB: can be called early; suppress needless cmds */
1803	if (!sc->sc_running)
1804		return;
1805
1806	/*
1807	 * Calculate the ERP flags.  The firwmare will use
1808	 * this to carry out the appropriate measures.
1809	 */
1810	prot = 0;
1811	if (IEEE80211_IS_CHAN_ANYG(ic->ic_curchan)) {
1812		if ((ic->ic_flags & IEEE80211_F_SHSLOT) == 0)
1813			prot |= IEEE80211_ERP_NON_ERP_PRESENT;
1814		if (ic->ic_flags & IEEE80211_F_USEPROT)
1815			prot |= IEEE80211_ERP_USE_PROTECTION;
1816		if (ic->ic_flags & IEEE80211_F_USEBARKER)
1817			prot |= IEEE80211_ERP_LONG_PREAMBLE;
1818	}
1819
1820	DPRINTF(sc, MWL_DEBUG_RESET,
1821	    "%s: chan %u MHz/flags 0x%x %s slot, (prot 0x%x ic_flags 0x%x)\n",
1822	    __func__, ic->ic_curchan->ic_freq, ic->ic_curchan->ic_flags,
1823	    ic->ic_flags & IEEE80211_F_SHSLOT ? "short" : "long", prot,
1824	    ic->ic_flags);
1825
1826	mwl_hal_setgprot(mh, prot);
1827}
1828
1829/*
1830 * Setup the beacon frame.
1831 */
1832static int
1833mwl_beacon_setup(struct ieee80211vap *vap)
1834{
1835	struct mwl_hal_vap *hvap = MWL_VAP(vap)->mv_hvap;
1836	struct ieee80211_node *ni = vap->iv_bss;
1837	struct ieee80211_beacon_offsets bo;
1838	struct mbuf *m;
1839
1840	m = ieee80211_beacon_alloc(ni, &bo);
1841	if (m == NULL)
1842		return ENOBUFS;
1843	mwl_hal_setbeacon(hvap, mtod(m, const void *), m->m_len);
1844	m_free(m);
1845
1846	return 0;
1847}
1848
1849/*
1850 * Update the beacon frame in response to a change.
1851 */
1852static void
1853mwl_beacon_update(struct ieee80211vap *vap, int item)
1854{
1855	struct mwl_hal_vap *hvap = MWL_VAP(vap)->mv_hvap;
1856	struct ieee80211com *ic = vap->iv_ic;
1857
1858	KASSERT(hvap != NULL, ("no beacon"));
1859	switch (item) {
1860	case IEEE80211_BEACON_ERP:
1861		mwl_updateslot(ic);
1862		break;
1863	case IEEE80211_BEACON_HTINFO:
1864		mwl_hal_setnprotmode(hvap,
1865		    MS(ic->ic_curhtprotmode, IEEE80211_HTINFO_OPMODE));
1866		break;
1867	case IEEE80211_BEACON_CAPS:
1868	case IEEE80211_BEACON_WME:
1869	case IEEE80211_BEACON_APPIE:
1870	case IEEE80211_BEACON_CSA:
1871		break;
1872	case IEEE80211_BEACON_TIM:
1873		/* NB: firmware always forms TIM */
1874		return;
1875	}
1876	/* XXX retain beacon frame and update */
1877	mwl_beacon_setup(vap);
1878}
1879
1880static void
1881mwl_load_cb(void *arg, bus_dma_segment_t *segs, int nsegs, int error)
1882{
1883	bus_addr_t *paddr = (bus_addr_t*) arg;
1884	KASSERT(error == 0, ("error %u on bus_dma callback", error));
1885	*paddr = segs->ds_addr;
1886}
1887
1888#ifdef MWL_HOST_PS_SUPPORT
1889/*
1890 * Handle power save station occupancy changes.
1891 */
1892static void
1893mwl_update_ps(struct ieee80211vap *vap, int nsta)
1894{
1895	struct mwl_vap *mvp = MWL_VAP(vap);
1896
1897	if (nsta == 0 || mvp->mv_last_ps_sta == 0)
1898		mwl_hal_setpowersave_bss(mvp->mv_hvap, nsta);
1899	mvp->mv_last_ps_sta = nsta;
1900}
1901
1902/*
1903 * Handle associated station power save state changes.
1904 */
1905static int
1906mwl_set_tim(struct ieee80211_node *ni, int set)
1907{
1908	struct ieee80211vap *vap = ni->ni_vap;
1909	struct mwl_vap *mvp = MWL_VAP(vap);
1910
1911	if (mvp->mv_set_tim(ni, set)) {		/* NB: state change */
1912		mwl_hal_setpowersave_sta(mvp->mv_hvap,
1913		    IEEE80211_AID(ni->ni_associd), set);
1914		return 1;
1915	} else
1916		return 0;
1917}
1918#endif /* MWL_HOST_PS_SUPPORT */
1919
1920static int
1921mwl_desc_setup(struct mwl_softc *sc, const char *name,
1922	struct mwl_descdma *dd,
1923	int nbuf, size_t bufsize, int ndesc, size_t descsize)
1924{
1925	uint8_t *ds;
1926	int error;
1927
1928	DPRINTF(sc, MWL_DEBUG_RESET,
1929	    "%s: %s DMA: %u bufs (%ju) %u desc/buf (%ju)\n",
1930	    __func__, name, nbuf, (uintmax_t) bufsize,
1931	    ndesc, (uintmax_t) descsize);
1932
1933	dd->dd_name = name;
1934	dd->dd_desc_len = nbuf * ndesc * descsize;
1935
1936	/*
1937	 * Setup DMA descriptor area.
1938	 */
1939	error = bus_dma_tag_create(bus_get_dma_tag(sc->sc_dev),	/* parent */
1940		       PAGE_SIZE, 0,		/* alignment, bounds */
1941		       BUS_SPACE_MAXADDR_32BIT,	/* lowaddr */
1942		       BUS_SPACE_MAXADDR,	/* highaddr */
1943		       NULL, NULL,		/* filter, filterarg */
1944		       dd->dd_desc_len,		/* maxsize */
1945		       1,			/* nsegments */
1946		       dd->dd_desc_len,		/* maxsegsize */
1947		       BUS_DMA_ALLOCNOW,	/* flags */
1948		       NULL,			/* lockfunc */
1949		       NULL,			/* lockarg */
1950		       &dd->dd_dmat);
1951	if (error != 0) {
1952		device_printf(sc->sc_dev, "cannot allocate %s DMA tag\n", dd->dd_name);
1953		return error;
1954	}
1955
1956	/* allocate descriptors */
1957	error = bus_dmamem_alloc(dd->dd_dmat, (void**) &dd->dd_desc,
1958				 BUS_DMA_NOWAIT | BUS_DMA_COHERENT,
1959				 &dd->dd_dmamap);
1960	if (error != 0) {
1961		device_printf(sc->sc_dev, "unable to alloc memory for %u %s descriptors, "
1962			"error %u\n", nbuf * ndesc, dd->dd_name, error);
1963		goto fail1;
1964	}
1965
1966	error = bus_dmamap_load(dd->dd_dmat, dd->dd_dmamap,
1967				dd->dd_desc, dd->dd_desc_len,
1968				mwl_load_cb, &dd->dd_desc_paddr,
1969				BUS_DMA_NOWAIT);
1970	if (error != 0) {
1971		device_printf(sc->sc_dev, "unable to map %s descriptors, error %u\n",
1972			dd->dd_name, error);
1973		goto fail2;
1974	}
1975
1976	ds = dd->dd_desc;
1977	memset(ds, 0, dd->dd_desc_len);
1978	DPRINTF(sc, MWL_DEBUG_RESET,
1979	    "%s: %s DMA map: %p (%lu) -> 0x%jx (%lu)\n",
1980	    __func__, dd->dd_name, ds, (u_long) dd->dd_desc_len,
1981	    (uintmax_t) dd->dd_desc_paddr, /*XXX*/ (u_long) dd->dd_desc_len);
1982
1983	return 0;
1984fail2:
1985	bus_dmamem_free(dd->dd_dmat, dd->dd_desc, dd->dd_dmamap);
1986fail1:
1987	bus_dma_tag_destroy(dd->dd_dmat);
1988	memset(dd, 0, sizeof(*dd));
1989	return error;
1990#undef DS2PHYS
1991}
1992
1993static void
1994mwl_desc_cleanup(struct mwl_softc *sc, struct mwl_descdma *dd)
1995{
1996	bus_dmamap_unload(dd->dd_dmat, dd->dd_dmamap);
1997	bus_dmamem_free(dd->dd_dmat, dd->dd_desc, dd->dd_dmamap);
1998	bus_dma_tag_destroy(dd->dd_dmat);
1999
2000	memset(dd, 0, sizeof(*dd));
2001}
2002
2003/*
2004 * Construct a tx q's free list.  The order of entries on
2005 * the list must reflect the physical layout of tx descriptors
2006 * because the firmware pre-fetches descriptors.
2007 *
2008 * XXX might be better to use indices into the buffer array.
2009 */
2010static void
2011mwl_txq_reset(struct mwl_softc *sc, struct mwl_txq *txq)
2012{
2013	struct mwl_txbuf *bf;
2014	int i;
2015
2016	bf = txq->dma.dd_bufptr;
2017	STAILQ_INIT(&txq->free);
2018	for (i = 0; i < mwl_txbuf; i++, bf++)
2019		STAILQ_INSERT_TAIL(&txq->free, bf, bf_list);
2020	txq->nfree = i;
2021}
2022
2023#define	DS2PHYS(_dd, _ds) \
2024	((_dd)->dd_desc_paddr + ((caddr_t)(_ds) - (caddr_t)(_dd)->dd_desc))
2025
2026static int
2027mwl_txdma_setup(struct mwl_softc *sc, struct mwl_txq *txq)
2028{
2029	int error, bsize, i;
2030	struct mwl_txbuf *bf;
2031	struct mwl_txdesc *ds;
2032
2033	error = mwl_desc_setup(sc, "tx", &txq->dma,
2034			mwl_txbuf, sizeof(struct mwl_txbuf),
2035			MWL_TXDESC, sizeof(struct mwl_txdesc));
2036	if (error != 0)
2037		return error;
2038
2039	/* allocate and setup tx buffers */
2040	bsize = mwl_txbuf * sizeof(struct mwl_txbuf);
2041	bf = malloc(bsize, M_MWLDEV, M_NOWAIT | M_ZERO);
2042	if (bf == NULL) {
2043		device_printf(sc->sc_dev, "malloc of %u tx buffers failed\n",
2044			mwl_txbuf);
2045		return ENOMEM;
2046	}
2047	txq->dma.dd_bufptr = bf;
2048
2049	ds = txq->dma.dd_desc;
2050	for (i = 0; i < mwl_txbuf; i++, bf++, ds += MWL_TXDESC) {
2051		bf->bf_desc = ds;
2052		bf->bf_daddr = DS2PHYS(&txq->dma, ds);
2053		error = bus_dmamap_create(sc->sc_dmat, BUS_DMA_NOWAIT,
2054				&bf->bf_dmamap);
2055		if (error != 0) {
2056			device_printf(sc->sc_dev, "unable to create dmamap for tx "
2057				"buffer %u, error %u\n", i, error);
2058			return error;
2059		}
2060	}
2061	mwl_txq_reset(sc, txq);
2062	return 0;
2063}
2064
2065static void
2066mwl_txdma_cleanup(struct mwl_softc *sc, struct mwl_txq *txq)
2067{
2068	struct mwl_txbuf *bf;
2069	int i;
2070
2071	bf = txq->dma.dd_bufptr;
2072	for (i = 0; i < mwl_txbuf; i++, bf++) {
2073		KASSERT(bf->bf_m == NULL, ("mbuf on free list"));
2074		KASSERT(bf->bf_node == NULL, ("node on free list"));
2075		if (bf->bf_dmamap != NULL)
2076			bus_dmamap_destroy(sc->sc_dmat, bf->bf_dmamap);
2077	}
2078	STAILQ_INIT(&txq->free);
2079	txq->nfree = 0;
2080	if (txq->dma.dd_bufptr != NULL) {
2081		free(txq->dma.dd_bufptr, M_MWLDEV);
2082		txq->dma.dd_bufptr = NULL;
2083	}
2084	if (txq->dma.dd_desc_len != 0)
2085		mwl_desc_cleanup(sc, &txq->dma);
2086}
2087
2088static int
2089mwl_rxdma_setup(struct mwl_softc *sc)
2090{
2091	int error, jumbosize, bsize, i;
2092	struct mwl_rxbuf *bf;
2093	struct mwl_jumbo *rbuf;
2094	struct mwl_rxdesc *ds;
2095	caddr_t data;
2096
2097	error = mwl_desc_setup(sc, "rx", &sc->sc_rxdma,
2098			mwl_rxdesc, sizeof(struct mwl_rxbuf),
2099			1, sizeof(struct mwl_rxdesc));
2100	if (error != 0)
2101		return error;
2102
2103	/*
2104	 * Receive is done to a private pool of jumbo buffers.
2105	 * This allows us to attach to mbuf's and avoid re-mapping
2106	 * memory on each rx we post.  We allocate a large chunk
2107	 * of memory and manage it in the driver.  The mbuf free
2108	 * callback method is used to reclaim frames after sending
2109	 * them up the stack.  By default we allocate 2x the number of
2110	 * rx descriptors configured so we have some slop to hold
2111	 * us while frames are processed.
2112	 */
2113	if (mwl_rxbuf < 2*mwl_rxdesc) {
2114		device_printf(sc->sc_dev,
2115		    "too few rx dma buffers (%d); increasing to %d\n",
2116		    mwl_rxbuf, 2*mwl_rxdesc);
2117		mwl_rxbuf = 2*mwl_rxdesc;
2118	}
2119	jumbosize = roundup(MWL_AGGR_SIZE, PAGE_SIZE);
2120	sc->sc_rxmemsize = mwl_rxbuf*jumbosize;
2121
2122	error = bus_dma_tag_create(sc->sc_dmat,	/* parent */
2123		       PAGE_SIZE, 0,		/* alignment, bounds */
2124		       BUS_SPACE_MAXADDR_32BIT,	/* lowaddr */
2125		       BUS_SPACE_MAXADDR,	/* highaddr */
2126		       NULL, NULL,		/* filter, filterarg */
2127		       sc->sc_rxmemsize,	/* maxsize */
2128		       1,			/* nsegments */
2129		       sc->sc_rxmemsize,	/* maxsegsize */
2130		       BUS_DMA_ALLOCNOW,	/* flags */
2131		       NULL,			/* lockfunc */
2132		       NULL,			/* lockarg */
2133		       &sc->sc_rxdmat);
2134	if (error != 0) {
2135		device_printf(sc->sc_dev, "could not create rx DMA tag\n");
2136		return error;
2137	}
2138
2139	error = bus_dmamem_alloc(sc->sc_rxdmat, (void**) &sc->sc_rxmem,
2140				 BUS_DMA_NOWAIT | BUS_DMA_COHERENT,
2141				 &sc->sc_rxmap);
2142	if (error != 0) {
2143		device_printf(sc->sc_dev, "could not alloc %ju bytes of rx DMA memory\n",
2144		    (uintmax_t) sc->sc_rxmemsize);
2145		return error;
2146	}
2147
2148	error = bus_dmamap_load(sc->sc_rxdmat, sc->sc_rxmap,
2149				sc->sc_rxmem, sc->sc_rxmemsize,
2150				mwl_load_cb, &sc->sc_rxmem_paddr,
2151				BUS_DMA_NOWAIT);
2152	if (error != 0) {
2153		device_printf(sc->sc_dev, "could not load rx DMA map\n");
2154		return error;
2155	}
2156
2157	/*
2158	 * Allocate rx buffers and set them up.
2159	 */
2160	bsize = mwl_rxdesc * sizeof(struct mwl_rxbuf);
2161	bf = malloc(bsize, M_MWLDEV, M_NOWAIT | M_ZERO);
2162	if (bf == NULL) {
2163		device_printf(sc->sc_dev, "malloc of %u rx buffers failed\n", bsize);
2164		return error;
2165	}
2166	sc->sc_rxdma.dd_bufptr = bf;
2167
2168	STAILQ_INIT(&sc->sc_rxbuf);
2169	ds = sc->sc_rxdma.dd_desc;
2170	for (i = 0; i < mwl_rxdesc; i++, bf++, ds++) {
2171		bf->bf_desc = ds;
2172		bf->bf_daddr = DS2PHYS(&sc->sc_rxdma, ds);
2173		/* pre-assign dma buffer */
2174		bf->bf_data = ((uint8_t *)sc->sc_rxmem) + (i*jumbosize);
2175		/* NB: tail is intentional to preserve descriptor order */
2176		STAILQ_INSERT_TAIL(&sc->sc_rxbuf, bf, bf_list);
2177	}
2178
2179	/*
2180	 * Place remainder of dma memory buffers on the free list.
2181	 */
2182	SLIST_INIT(&sc->sc_rxfree);
2183	for (; i < mwl_rxbuf; i++) {
2184		data = ((uint8_t *)sc->sc_rxmem) + (i*jumbosize);
2185		rbuf = MWL_JUMBO_DATA2BUF(data);
2186		SLIST_INSERT_HEAD(&sc->sc_rxfree, rbuf, next);
2187		sc->sc_nrxfree++;
2188	}
2189	return 0;
2190}
2191#undef DS2PHYS
2192
2193static void
2194mwl_rxdma_cleanup(struct mwl_softc *sc)
2195{
2196	if (sc->sc_rxmem_paddr != 0) {
2197		bus_dmamap_unload(sc->sc_rxdmat, sc->sc_rxmap);
2198		sc->sc_rxmem_paddr = 0;
2199	}
2200	if (sc->sc_rxmem != NULL) {
2201		bus_dmamem_free(sc->sc_rxdmat, sc->sc_rxmem, sc->sc_rxmap);
2202		sc->sc_rxmem = NULL;
2203	}
2204	if (sc->sc_rxdma.dd_bufptr != NULL) {
2205		free(sc->sc_rxdma.dd_bufptr, M_MWLDEV);
2206		sc->sc_rxdma.dd_bufptr = NULL;
2207	}
2208	if (sc->sc_rxdma.dd_desc_len != 0)
2209		mwl_desc_cleanup(sc, &sc->sc_rxdma);
2210}
2211
2212static int
2213mwl_dma_setup(struct mwl_softc *sc)
2214{
2215	int error, i;
2216
2217	error = mwl_rxdma_setup(sc);
2218	if (error != 0) {
2219		mwl_rxdma_cleanup(sc);
2220		return error;
2221	}
2222
2223	for (i = 0; i < MWL_NUM_TX_QUEUES; i++) {
2224		error = mwl_txdma_setup(sc, &sc->sc_txq[i]);
2225		if (error != 0) {
2226			mwl_dma_cleanup(sc);
2227			return error;
2228		}
2229	}
2230	return 0;
2231}
2232
2233static void
2234mwl_dma_cleanup(struct mwl_softc *sc)
2235{
2236	int i;
2237
2238	for (i = 0; i < MWL_NUM_TX_QUEUES; i++)
2239		mwl_txdma_cleanup(sc, &sc->sc_txq[i]);
2240	mwl_rxdma_cleanup(sc);
2241}
2242
2243static struct ieee80211_node *
2244mwl_node_alloc(struct ieee80211vap *vap, const uint8_t mac[IEEE80211_ADDR_LEN])
2245{
2246	struct ieee80211com *ic = vap->iv_ic;
2247	struct mwl_softc *sc = ic->ic_softc;
2248	const size_t space = sizeof(struct mwl_node);
2249	struct mwl_node *mn;
2250
2251	mn = malloc(space, M_80211_NODE, M_NOWAIT|M_ZERO);
2252	if (mn == NULL) {
2253		/* XXX stat+msg */
2254		return NULL;
2255	}
2256	DPRINTF(sc, MWL_DEBUG_NODE, "%s: mn %p\n", __func__, mn);
2257	return &mn->mn_node;
2258}
2259
2260static void
2261mwl_node_cleanup(struct ieee80211_node *ni)
2262{
2263	struct ieee80211com *ic = ni->ni_ic;
2264        struct mwl_softc *sc = ic->ic_softc;
2265	struct mwl_node *mn = MWL_NODE(ni);
2266
2267	DPRINTF(sc, MWL_DEBUG_NODE, "%s: ni %p ic %p staid %d\n",
2268	    __func__, ni, ni->ni_ic, mn->mn_staid);
2269
2270	if (mn->mn_staid != 0) {
2271		struct ieee80211vap *vap = ni->ni_vap;
2272
2273		if (mn->mn_hvap != NULL) {
2274			if (vap->iv_opmode == IEEE80211_M_STA)
2275				mwl_hal_delstation(mn->mn_hvap, vap->iv_myaddr);
2276			else
2277				mwl_hal_delstation(mn->mn_hvap, ni->ni_macaddr);
2278		}
2279		/*
2280		 * NB: legacy WDS peer sta db entry is installed using
2281		 * the associate ap's hvap; use it again to delete it.
2282		 * XXX can vap be NULL?
2283		 */
2284		else if (vap->iv_opmode == IEEE80211_M_WDS &&
2285		    MWL_VAP(vap)->mv_ap_hvap != NULL)
2286			mwl_hal_delstation(MWL_VAP(vap)->mv_ap_hvap,
2287			    ni->ni_macaddr);
2288		delstaid(sc, mn->mn_staid);
2289		mn->mn_staid = 0;
2290	}
2291	sc->sc_node_cleanup(ni);
2292}
2293
2294/*
2295 * Reclaim rx dma buffers from packets sitting on the ampdu
2296 * reorder queue for a station.  We replace buffers with a
2297 * system cluster (if available).
2298 */
2299static void
2300mwl_ampdu_rxdma_reclaim(struct ieee80211_rx_ampdu *rap)
2301{
2302#if 0
2303	int i, n, off;
2304	struct mbuf *m;
2305	void *cl;
2306
2307	n = rap->rxa_qframes;
2308	for (i = 0; i < rap->rxa_wnd && n > 0; i++) {
2309		m = rap->rxa_m[i];
2310		if (m == NULL)
2311			continue;
2312		n--;
2313		/* our dma buffers have a well-known free routine */
2314		if ((m->m_flags & M_EXT) == 0 ||
2315		    m->m_ext.ext_free != mwl_ext_free)
2316			continue;
2317		/*
2318		 * Try to allocate a cluster and move the data.
2319		 */
2320		off = m->m_data - m->m_ext.ext_buf;
2321		if (off + m->m_pkthdr.len > MCLBYTES) {
2322			/* XXX no AMSDU for now */
2323			continue;
2324		}
2325		cl = pool_cache_get_paddr(&mclpool_cache, 0,
2326		    &m->m_ext.ext_paddr);
2327		if (cl != NULL) {
2328			/*
2329			 * Copy the existing data to the cluster, remove
2330			 * the rx dma buffer, and attach the cluster in
2331			 * its place.  Note we preserve the offset to the
2332			 * data so frames being bridged can still prepend
2333			 * their headers without adding another mbuf.
2334			 */
2335			memcpy((caddr_t) cl + off, m->m_data, m->m_pkthdr.len);
2336			MEXTREMOVE(m);
2337			MEXTADD(m, cl, MCLBYTES, 0, NULL, &mclpool_cache);
2338			/* setup mbuf like _MCLGET does */
2339			m->m_flags |= M_CLUSTER | M_EXT_RW;
2340			_MOWNERREF(m, M_EXT | M_CLUSTER);
2341			/* NB: m_data is clobbered by MEXTADDR, adjust */
2342			m->m_data += off;
2343		}
2344	}
2345#endif
2346}
2347
2348/*
2349 * Callback to reclaim resources.  We first let the
2350 * net80211 layer do it's thing, then if we are still
2351 * blocked by a lack of rx dma buffers we walk the ampdu
2352 * reorder q's to reclaim buffers by copying to a system
2353 * cluster.
2354 */
2355static void
2356mwl_node_drain(struct ieee80211_node *ni)
2357{
2358	struct ieee80211com *ic = ni->ni_ic;
2359        struct mwl_softc *sc = ic->ic_softc;
2360	struct mwl_node *mn = MWL_NODE(ni);
2361
2362	DPRINTF(sc, MWL_DEBUG_NODE, "%s: ni %p vap %p staid %d\n",
2363	    __func__, ni, ni->ni_vap, mn->mn_staid);
2364
2365	/* NB: call up first to age out ampdu q's */
2366	sc->sc_node_drain(ni);
2367
2368	/* XXX better to not check low water mark? */
2369	if (sc->sc_rxblocked && mn->mn_staid != 0 &&
2370	    (ni->ni_flags & IEEE80211_NODE_HT)) {
2371		uint8_t tid;
2372		/*
2373		 * Walk the reorder q and reclaim rx dma buffers by copying
2374		 * the packet contents into clusters.
2375		 */
2376		for (tid = 0; tid < WME_NUM_TID; tid++) {
2377			struct ieee80211_rx_ampdu *rap;
2378
2379			rap = &ni->ni_rx_ampdu[tid];
2380			if ((rap->rxa_flags & IEEE80211_AGGR_XCHGPEND) == 0)
2381				continue;
2382			if (rap->rxa_qframes)
2383				mwl_ampdu_rxdma_reclaim(rap);
2384		}
2385	}
2386}
2387
2388static void
2389mwl_node_getsignal(const struct ieee80211_node *ni, int8_t *rssi, int8_t *noise)
2390{
2391	*rssi = ni->ni_ic->ic_node_getrssi(ni);
2392#ifdef MWL_ANT_INFO_SUPPORT
2393#if 0
2394	/* XXX need to smooth data */
2395	*noise = -MWL_NODE_CONST(ni)->mn_ai.nf;
2396#else
2397	*noise = -95;		/* XXX */
2398#endif
2399#else
2400	*noise = -95;		/* XXX */
2401#endif
2402}
2403
2404/*
2405 * Convert Hardware per-antenna rssi info to common format:
2406 * Let a1, a2, a3 represent the amplitudes per chain
2407 * Let amax represent max[a1, a2, a3]
2408 * Rssi1_dBm = RSSI_dBm + 20*log10(a1/amax)
2409 * Rssi1_dBm = RSSI_dBm + 20*log10(a1) - 20*log10(amax)
2410 * We store a table that is 4*20*log10(idx) - the extra 4 is to store or
2411 * maintain some extra precision.
2412 *
2413 * Values are stored in .5 db format capped at 127.
2414 */
2415static void
2416mwl_node_getmimoinfo(const struct ieee80211_node *ni,
2417	struct ieee80211_mimo_info *mi)
2418{
2419#define	CVT(_dst, _src) do {						\
2420	(_dst) = rssi + ((logdbtbl[_src] - logdbtbl[rssi_max]) >> 2);	\
2421	(_dst) = (_dst) > 64 ? 127 : ((_dst) << 1);			\
2422} while (0)
2423	static const int8_t logdbtbl[32] = {
2424	       0,   0,  24,  38,  48,  56,  62,  68,
2425	      72,  76,  80,  83,  86,  89,  92,  94,
2426	      96,  98, 100, 102, 104, 106, 107, 109,
2427	     110, 112, 113, 115, 116, 117, 118, 119
2428	};
2429	const struct mwl_node *mn = MWL_NODE_CONST(ni);
2430	uint8_t rssi = mn->mn_ai.rsvd1/2;		/* XXX */
2431	uint32_t rssi_max;
2432
2433	rssi_max = mn->mn_ai.rssi_a;
2434	if (mn->mn_ai.rssi_b > rssi_max)
2435		rssi_max = mn->mn_ai.rssi_b;
2436	if (mn->mn_ai.rssi_c > rssi_max)
2437		rssi_max = mn->mn_ai.rssi_c;
2438
2439	CVT(mi->rssi[0], mn->mn_ai.rssi_a);
2440	CVT(mi->rssi[1], mn->mn_ai.rssi_b);
2441	CVT(mi->rssi[2], mn->mn_ai.rssi_c);
2442
2443	mi->noise[0] = mn->mn_ai.nf_a;
2444	mi->noise[1] = mn->mn_ai.nf_b;
2445	mi->noise[2] = mn->mn_ai.nf_c;
2446#undef CVT
2447}
2448
2449static __inline void *
2450mwl_getrxdma(struct mwl_softc *sc)
2451{
2452	struct mwl_jumbo *buf;
2453	void *data;
2454
2455	/*
2456	 * Allocate from jumbo pool.
2457	 */
2458	MWL_RXFREE_LOCK(sc);
2459	buf = SLIST_FIRST(&sc->sc_rxfree);
2460	if (buf == NULL) {
2461		DPRINTF(sc, MWL_DEBUG_ANY,
2462		    "%s: out of rx dma buffers\n", __func__);
2463		sc->sc_stats.mst_rx_nodmabuf++;
2464		data = NULL;
2465	} else {
2466		SLIST_REMOVE_HEAD(&sc->sc_rxfree, next);
2467		sc->sc_nrxfree--;
2468		data = MWL_JUMBO_BUF2DATA(buf);
2469	}
2470	MWL_RXFREE_UNLOCK(sc);
2471	return data;
2472}
2473
2474static __inline void
2475mwl_putrxdma(struct mwl_softc *sc, void *data)
2476{
2477	struct mwl_jumbo *buf;
2478
2479	/* XXX bounds check data */
2480	MWL_RXFREE_LOCK(sc);
2481	buf = MWL_JUMBO_DATA2BUF(data);
2482	SLIST_INSERT_HEAD(&sc->sc_rxfree, buf, next);
2483	sc->sc_nrxfree++;
2484	MWL_RXFREE_UNLOCK(sc);
2485}
2486
2487static int
2488mwl_rxbuf_init(struct mwl_softc *sc, struct mwl_rxbuf *bf)
2489{
2490	struct mwl_rxdesc *ds;
2491
2492	ds = bf->bf_desc;
2493	if (bf->bf_data == NULL) {
2494		bf->bf_data = mwl_getrxdma(sc);
2495		if (bf->bf_data == NULL) {
2496			/* mark descriptor to be skipped */
2497			ds->RxControl = EAGLE_RXD_CTRL_OS_OWN;
2498			/* NB: don't need PREREAD */
2499			MWL_RXDESC_SYNC(sc, ds, BUS_DMASYNC_PREWRITE);
2500			sc->sc_stats.mst_rxbuf_failed++;
2501			return ENOMEM;
2502		}
2503	}
2504	/*
2505	 * NB: DMA buffer contents is known to be unmodified
2506	 *     so there's no need to flush the data cache.
2507	 */
2508
2509	/*
2510	 * Setup descriptor.
2511	 */
2512	ds->QosCtrl = 0;
2513	ds->RSSI = 0;
2514	ds->Status = EAGLE_RXD_STATUS_IDLE;
2515	ds->Channel = 0;
2516	ds->PktLen = htole16(MWL_AGGR_SIZE);
2517	ds->SQ2 = 0;
2518	ds->pPhysBuffData = htole32(MWL_JUMBO_DMA_ADDR(sc, bf->bf_data));
2519	/* NB: don't touch pPhysNext, set once */
2520	ds->RxControl = EAGLE_RXD_CTRL_DRIVER_OWN;
2521	MWL_RXDESC_SYNC(sc, ds, BUS_DMASYNC_PREREAD | BUS_DMASYNC_PREWRITE);
2522
2523	return 0;
2524}
2525
2526static void
2527mwl_ext_free(struct mbuf *m, void *data, void *arg)
2528{
2529	struct mwl_softc *sc = arg;
2530
2531	/* XXX bounds check data */
2532	mwl_putrxdma(sc, data);
2533	/*
2534	 * If we were previously blocked by a lack of rx dma buffers
2535	 * check if we now have enough to restart rx interrupt handling.
2536	 * NB: we know we are called at splvm which is above splnet.
2537	 */
2538	if (sc->sc_rxblocked && sc->sc_nrxfree > mwl_rxdmalow) {
2539		sc->sc_rxblocked = 0;
2540		mwl_hal_intrset(sc->sc_mh, sc->sc_imask);
2541	}
2542}
2543
2544struct mwl_frame_bar {
2545	u_int8_t	i_fc[2];
2546	u_int8_t	i_dur[2];
2547	u_int8_t	i_ra[IEEE80211_ADDR_LEN];
2548	u_int8_t	i_ta[IEEE80211_ADDR_LEN];
2549	/* ctl, seq, FCS */
2550} __packed;
2551
2552/*
2553 * Like ieee80211_anyhdrsize, but handles BAR frames
2554 * specially so the logic below to piece the 802.11
2555 * header together works.
2556 */
2557static __inline int
2558mwl_anyhdrsize(const void *data)
2559{
2560	const struct ieee80211_frame *wh = data;
2561
2562	if ((wh->i_fc[0]&IEEE80211_FC0_TYPE_MASK) == IEEE80211_FC0_TYPE_CTL) {
2563		switch (wh->i_fc[0] & IEEE80211_FC0_SUBTYPE_MASK) {
2564		case IEEE80211_FC0_SUBTYPE_CTS:
2565		case IEEE80211_FC0_SUBTYPE_ACK:
2566			return sizeof(struct ieee80211_frame_ack);
2567		case IEEE80211_FC0_SUBTYPE_BAR:
2568			return sizeof(struct mwl_frame_bar);
2569		}
2570		return sizeof(struct ieee80211_frame_min);
2571	} else
2572		return ieee80211_hdrsize(data);
2573}
2574
2575static void
2576mwl_handlemicerror(struct ieee80211com *ic, const uint8_t *data)
2577{
2578	const struct ieee80211_frame *wh;
2579	struct ieee80211_node *ni;
2580
2581	wh = (const struct ieee80211_frame *)(data + sizeof(uint16_t));
2582	ni = ieee80211_find_rxnode(ic, (const struct ieee80211_frame_min *) wh);
2583	if (ni != NULL) {
2584		ieee80211_notify_michael_failure(ni->ni_vap, wh, 0);
2585		ieee80211_free_node(ni);
2586	}
2587}
2588
2589/*
2590 * Convert hardware signal strength to rssi.  The value
2591 * provided by the device has the noise floor added in;
2592 * we need to compensate for this but we don't have that
2593 * so we use a fixed value.
2594 *
2595 * The offset of 8 is good for both 2.4 and 5GHz.  The LNA
2596 * offset is already set as part of the initial gain.  This
2597 * will give at least +/- 3dB for 2.4GHz and +/- 5dB for 5GHz.
2598 */
2599static __inline int
2600cvtrssi(uint8_t ssi)
2601{
2602	int rssi = (int) ssi + 8;
2603	/* XXX hack guess until we have a real noise floor */
2604	rssi = 2*(87 - rssi);	/* NB: .5 dBm units */
2605	return (rssi < 0 ? 0 : rssi > 127 ? 127 : rssi);
2606}
2607
2608static void
2609mwl_rx_proc(void *arg, int npending)
2610{
2611	struct mwl_softc *sc = arg;
2612	struct ieee80211com *ic = &sc->sc_ic;
2613	struct mwl_rxbuf *bf;
2614	struct mwl_rxdesc *ds;
2615	struct mbuf *m;
2616	struct ieee80211_qosframe *wh;
2617	struct ieee80211_qosframe_addr4 *wh4;
2618	struct ieee80211_node *ni;
2619	struct mwl_node *mn;
2620	int off, len, hdrlen, pktlen, rssi, ntodo;
2621	uint8_t *data, status;
2622	void *newdata;
2623	int16_t nf;
2624
2625	DPRINTF(sc, MWL_DEBUG_RX_PROC, "%s: pending %u rdptr 0x%x wrptr 0x%x\n",
2626	    __func__, npending, RD4(sc, sc->sc_hwspecs.rxDescRead),
2627	    RD4(sc, sc->sc_hwspecs.rxDescWrite));
2628	nf = -96;			/* XXX */
2629	bf = sc->sc_rxnext;
2630	for (ntodo = mwl_rxquota; ntodo > 0; ntodo--) {
2631		if (bf == NULL)
2632			bf = STAILQ_FIRST(&sc->sc_rxbuf);
2633		ds = bf->bf_desc;
2634		data = bf->bf_data;
2635		if (data == NULL) {
2636			/*
2637			 * If data allocation failed previously there
2638			 * will be no buffer; try again to re-populate it.
2639			 * Note the firmware will not advance to the next
2640			 * descriptor with a dma buffer so we must mimic
2641			 * this or we'll get out of sync.
2642			 */
2643			DPRINTF(sc, MWL_DEBUG_ANY,
2644			    "%s: rx buf w/o dma memory\n", __func__);
2645			(void) mwl_rxbuf_init(sc, bf);
2646			sc->sc_stats.mst_rx_dmabufmissing++;
2647			break;
2648		}
2649		MWL_RXDESC_SYNC(sc, ds,
2650		    BUS_DMASYNC_POSTREAD | BUS_DMASYNC_POSTWRITE);
2651		if (ds->RxControl != EAGLE_RXD_CTRL_DMA_OWN)
2652			break;
2653#ifdef MWL_DEBUG
2654		if (sc->sc_debug & MWL_DEBUG_RECV_DESC)
2655			mwl_printrxbuf(bf, 0);
2656#endif
2657		status = ds->Status;
2658		if (status & EAGLE_RXD_STATUS_DECRYPT_ERR_MASK) {
2659			counter_u64_add(ic->ic_ierrors, 1);
2660			sc->sc_stats.mst_rx_crypto++;
2661			/*
2662			 * NB: Check EAGLE_RXD_STATUS_GENERAL_DECRYPT_ERR
2663			 *     for backwards compatibility.
2664			 */
2665			if (status != EAGLE_RXD_STATUS_GENERAL_DECRYPT_ERR &&
2666			    (status & EAGLE_RXD_STATUS_TKIP_MIC_DECRYPT_ERR)) {
2667				/*
2668				 * MIC error, notify upper layers.
2669				 */
2670				bus_dmamap_sync(sc->sc_rxdmat, sc->sc_rxmap,
2671				    BUS_DMASYNC_POSTREAD);
2672				mwl_handlemicerror(ic, data);
2673				sc->sc_stats.mst_rx_tkipmic++;
2674			}
2675			/* XXX too painful to tap packets */
2676			goto rx_next;
2677		}
2678		/*
2679		 * Sync the data buffer.
2680		 */
2681		len = le16toh(ds->PktLen);
2682		bus_dmamap_sync(sc->sc_rxdmat, sc->sc_rxmap, BUS_DMASYNC_POSTREAD);
2683		/*
2684		 * The 802.11 header is provided all or in part at the front;
2685		 * use it to calculate the true size of the header that we'll
2686		 * construct below.  We use this to figure out where to copy
2687		 * payload prior to constructing the header.
2688		 */
2689		hdrlen = mwl_anyhdrsize(data + sizeof(uint16_t));
2690		off = sizeof(uint16_t) + sizeof(struct ieee80211_frame_addr4);
2691
2692		/* calculate rssi early so we can re-use for each aggregate */
2693		rssi = cvtrssi(ds->RSSI);
2694
2695		pktlen = hdrlen + (len - off);
2696		/*
2697		 * NB: we know our frame is at least as large as
2698		 * IEEE80211_MIN_LEN because there is a 4-address
2699		 * frame at the front.  Hence there's no need to
2700		 * vet the packet length.  If the frame in fact
2701		 * is too small it should be discarded at the
2702		 * net80211 layer.
2703		 */
2704
2705		/*
2706		 * Attach dma buffer to an mbuf.  We tried
2707		 * doing this based on the packet size (i.e.
2708		 * copying small packets) but it turns out to
2709		 * be a net loss.  The tradeoff might be system
2710		 * dependent (cache architecture is important).
2711		 */
2712		MGETHDR(m, M_NOWAIT, MT_DATA);
2713		if (m == NULL) {
2714			DPRINTF(sc, MWL_DEBUG_ANY,
2715			    "%s: no rx mbuf\n", __func__);
2716			sc->sc_stats.mst_rx_nombuf++;
2717			goto rx_next;
2718		}
2719		/*
2720		 * Acquire the replacement dma buffer before
2721		 * processing the frame.  If we're out of dma
2722		 * buffers we disable rx interrupts and wait
2723		 * for the free pool to reach mlw_rxdmalow buffers
2724		 * before starting to do work again.  If the firmware
2725		 * runs out of descriptors then it will toss frames
2726		 * which is better than our doing it as that can
2727		 * starve our processing.  It is also important that
2728		 * we always process rx'd frames in case they are
2729		 * A-MPDU as otherwise the host's view of the BA
2730		 * window may get out of sync with the firmware.
2731		 */
2732		newdata = mwl_getrxdma(sc);
2733		if (newdata == NULL) {
2734			/* NB: stat+msg in mwl_getrxdma */
2735			m_free(m);
2736			/* disable RX interrupt and mark state */
2737			mwl_hal_intrset(sc->sc_mh,
2738			    sc->sc_imask &~ MACREG_A2HRIC_BIT_RX_RDY);
2739			sc->sc_rxblocked = 1;
2740			ieee80211_drain(ic);
2741			/* XXX check rxblocked and immediately start again? */
2742			goto rx_stop;
2743		}
2744		bf->bf_data = newdata;
2745		/*
2746		 * Attach the dma buffer to the mbuf;
2747		 * mwl_rxbuf_init will re-setup the rx
2748		 * descriptor using the replacement dma
2749		 * buffer we just installed above.
2750		 */
2751		MEXTADD(m, data, MWL_AGGR_SIZE, mwl_ext_free,
2752		    data, sc, 0, EXT_NET_DRV);
2753		m->m_data += off - hdrlen;
2754		m->m_pkthdr.len = m->m_len = pktlen;
2755		/* NB: dma buffer assumed read-only */
2756
2757		/*
2758		 * Piece 802.11 header together.
2759		 */
2760		wh = mtod(m, struct ieee80211_qosframe *);
2761		/* NB: don't need to do this sometimes but ... */
2762		/* XXX special case so we can memcpy after m_devget? */
2763		ovbcopy(data + sizeof(uint16_t), wh, hdrlen);
2764		if (IEEE80211_QOS_HAS_SEQ(wh)) {
2765			if (IEEE80211_IS_DSTODS(wh)) {
2766				wh4 = mtod(m,
2767				    struct ieee80211_qosframe_addr4*);
2768				*(uint16_t *)wh4->i_qos = ds->QosCtrl;
2769			} else {
2770				*(uint16_t *)wh->i_qos = ds->QosCtrl;
2771			}
2772		}
2773		/*
2774		 * The f/w strips WEP header but doesn't clear
2775		 * the WEP bit; mark the packet with M_WEP so
2776		 * net80211 will treat the data as decrypted.
2777		 * While here also clear the PWR_MGT bit since
2778		 * power save is handled by the firmware and
2779		 * passing this up will potentially cause the
2780		 * upper layer to put a station in power save
2781		 * (except when configured with MWL_HOST_PS_SUPPORT).
2782		 */
2783		if (wh->i_fc[1] & IEEE80211_FC1_PROTECTED)
2784			m->m_flags |= M_WEP;
2785#ifdef MWL_HOST_PS_SUPPORT
2786		wh->i_fc[1] &= ~IEEE80211_FC1_PROTECTED;
2787#else
2788		wh->i_fc[1] &= ~(IEEE80211_FC1_PROTECTED |
2789		    IEEE80211_FC1_PWR_MGT);
2790#endif
2791
2792		if (ieee80211_radiotap_active(ic)) {
2793			struct mwl_rx_radiotap_header *tap = &sc->sc_rx_th;
2794
2795			tap->wr_flags = 0;
2796			tap->wr_rate = ds->Rate;
2797			tap->wr_antsignal = rssi + nf;
2798			tap->wr_antnoise = nf;
2799		}
2800		if (IFF_DUMPPKTS_RECV(sc, wh)) {
2801			ieee80211_dump_pkt(ic, mtod(m, caddr_t),
2802			    len, ds->Rate, rssi);
2803		}
2804		/* dispatch */
2805		ni = ieee80211_find_rxnode(ic,
2806		    (const struct ieee80211_frame_min *) wh);
2807		if (ni != NULL) {
2808			mn = MWL_NODE(ni);
2809#ifdef MWL_ANT_INFO_SUPPORT
2810			mn->mn_ai.rssi_a = ds->ai.rssi_a;
2811			mn->mn_ai.rssi_b = ds->ai.rssi_b;
2812			mn->mn_ai.rssi_c = ds->ai.rssi_c;
2813			mn->mn_ai.rsvd1 = rssi;
2814#endif
2815			/* tag AMPDU aggregates for reorder processing */
2816			if (ni->ni_flags & IEEE80211_NODE_HT)
2817				m->m_flags |= M_AMPDU;
2818			(void) ieee80211_input(ni, m, rssi, nf);
2819			ieee80211_free_node(ni);
2820		} else
2821			(void) ieee80211_input_all(ic, m, rssi, nf);
2822rx_next:
2823		/* NB: ignore ENOMEM so we process more descriptors */
2824		(void) mwl_rxbuf_init(sc, bf);
2825		bf = STAILQ_NEXT(bf, bf_list);
2826	}
2827rx_stop:
2828	sc->sc_rxnext = bf;
2829
2830	if (mbufq_first(&sc->sc_snd) != NULL) {
2831		/* NB: kick fw; the tx thread may have been preempted */
2832		mwl_hal_txstart(sc->sc_mh, 0);
2833		mwl_start(sc);
2834	}
2835}
2836
2837static void
2838mwl_txq_init(struct mwl_softc *sc, struct mwl_txq *txq, int qnum)
2839{
2840	struct mwl_txbuf *bf, *bn;
2841	struct mwl_txdesc *ds;
2842
2843	MWL_TXQ_LOCK_INIT(sc, txq);
2844	txq->qnum = qnum;
2845	txq->txpri = 0;	/* XXX */
2846#if 0
2847	/* NB: q setup by mwl_txdma_setup XXX */
2848	STAILQ_INIT(&txq->free);
2849#endif
2850	STAILQ_FOREACH(bf, &txq->free, bf_list) {
2851		bf->bf_txq = txq;
2852
2853		ds = bf->bf_desc;
2854		bn = STAILQ_NEXT(bf, bf_list);
2855		if (bn == NULL)
2856			bn = STAILQ_FIRST(&txq->free);
2857		ds->pPhysNext = htole32(bn->bf_daddr);
2858	}
2859	STAILQ_INIT(&txq->active);
2860}
2861
2862/*
2863 * Setup a hardware data transmit queue for the specified
2864 * access control.  We record the mapping from ac's
2865 * to h/w queues for use by mwl_tx_start.
2866 */
2867static int
2868mwl_tx_setup(struct mwl_softc *sc, int ac, int mvtype)
2869{
2870	struct mwl_txq *txq;
2871
2872	if (ac >= nitems(sc->sc_ac2q)) {
2873		device_printf(sc->sc_dev, "AC %u out of range, max %zu!\n",
2874			ac, nitems(sc->sc_ac2q));
2875		return 0;
2876	}
2877	if (mvtype >= MWL_NUM_TX_QUEUES) {
2878		device_printf(sc->sc_dev, "mvtype %u out of range, max %u!\n",
2879			mvtype, MWL_NUM_TX_QUEUES);
2880		return 0;
2881	}
2882	txq = &sc->sc_txq[mvtype];
2883	mwl_txq_init(sc, txq, mvtype);
2884	sc->sc_ac2q[ac] = txq;
2885	return 1;
2886}
2887
2888/*
2889 * Update WME parameters for a transmit queue.
2890 */
2891static int
2892mwl_txq_update(struct mwl_softc *sc, int ac)
2893{
2894#define	MWL_EXPONENT_TO_VALUE(v)	((1<<v)-1)
2895	struct ieee80211com *ic = &sc->sc_ic;
2896	struct mwl_txq *txq = sc->sc_ac2q[ac];
2897	struct wmeParams *wmep = &ic->ic_wme.wme_chanParams.cap_wmeParams[ac];
2898	struct mwl_hal *mh = sc->sc_mh;
2899	int aifs, cwmin, cwmax, txoplim;
2900
2901	aifs = wmep->wmep_aifsn;
2902	/* XXX in sta mode need to pass log values for cwmin/max */
2903	cwmin = MWL_EXPONENT_TO_VALUE(wmep->wmep_logcwmin);
2904	cwmax = MWL_EXPONENT_TO_VALUE(wmep->wmep_logcwmax);
2905	txoplim = wmep->wmep_txopLimit;		/* NB: units of 32us */
2906
2907	if (mwl_hal_setedcaparams(mh, txq->qnum, cwmin, cwmax, aifs, txoplim)) {
2908		device_printf(sc->sc_dev, "unable to update hardware queue "
2909			"parameters for %s traffic!\n",
2910			ieee80211_wme_acnames[ac]);
2911		return 0;
2912	}
2913	return 1;
2914#undef MWL_EXPONENT_TO_VALUE
2915}
2916
2917/*
2918 * Callback from the 802.11 layer to update WME parameters.
2919 */
2920static int
2921mwl_wme_update(struct ieee80211com *ic)
2922{
2923	struct mwl_softc *sc = ic->ic_softc;
2924
2925	return !mwl_txq_update(sc, WME_AC_BE) ||
2926	    !mwl_txq_update(sc, WME_AC_BK) ||
2927	    !mwl_txq_update(sc, WME_AC_VI) ||
2928	    !mwl_txq_update(sc, WME_AC_VO) ? EIO : 0;
2929}
2930
2931/*
2932 * Reclaim resources for a setup queue.
2933 */
2934static void
2935mwl_tx_cleanupq(struct mwl_softc *sc, struct mwl_txq *txq)
2936{
2937	/* XXX hal work? */
2938	MWL_TXQ_LOCK_DESTROY(txq);
2939}
2940
2941/*
2942 * Reclaim all tx queue resources.
2943 */
2944static void
2945mwl_tx_cleanup(struct mwl_softc *sc)
2946{
2947	int i;
2948
2949	for (i = 0; i < MWL_NUM_TX_QUEUES; i++)
2950		mwl_tx_cleanupq(sc, &sc->sc_txq[i]);
2951}
2952
2953static int
2954mwl_tx_dmasetup(struct mwl_softc *sc, struct mwl_txbuf *bf, struct mbuf *m0)
2955{
2956	struct mbuf *m;
2957	int error;
2958
2959	/*
2960	 * Load the DMA map so any coalescing is done.  This
2961	 * also calculates the number of descriptors we need.
2962	 */
2963	error = bus_dmamap_load_mbuf_sg(sc->sc_dmat, bf->bf_dmamap, m0,
2964				     bf->bf_segs, &bf->bf_nseg,
2965				     BUS_DMA_NOWAIT);
2966	if (error == EFBIG) {
2967		/* XXX packet requires too many descriptors */
2968		bf->bf_nseg = MWL_TXDESC+1;
2969	} else if (error != 0) {
2970		sc->sc_stats.mst_tx_busdma++;
2971		m_freem(m0);
2972		return error;
2973	}
2974	/*
2975	 * Discard null packets and check for packets that
2976	 * require too many TX descriptors.  We try to convert
2977	 * the latter to a cluster.
2978	 */
2979	if (error == EFBIG) {		/* too many desc's, linearize */
2980		sc->sc_stats.mst_tx_linear++;
2981#if MWL_TXDESC > 1
2982		m = m_collapse(m0, M_NOWAIT, MWL_TXDESC);
2983#else
2984		m = m_defrag(m0, M_NOWAIT);
2985#endif
2986		if (m == NULL) {
2987			m_freem(m0);
2988			sc->sc_stats.mst_tx_nombuf++;
2989			return ENOMEM;
2990		}
2991		m0 = m;
2992		error = bus_dmamap_load_mbuf_sg(sc->sc_dmat, bf->bf_dmamap, m0,
2993					     bf->bf_segs, &bf->bf_nseg,
2994					     BUS_DMA_NOWAIT);
2995		if (error != 0) {
2996			sc->sc_stats.mst_tx_busdma++;
2997			m_freem(m0);
2998			return error;
2999		}
3000		KASSERT(bf->bf_nseg <= MWL_TXDESC,
3001		    ("too many segments after defrag; nseg %u", bf->bf_nseg));
3002	} else if (bf->bf_nseg == 0) {		/* null packet, discard */
3003		sc->sc_stats.mst_tx_nodata++;
3004		m_freem(m0);
3005		return EIO;
3006	}
3007	DPRINTF(sc, MWL_DEBUG_XMIT, "%s: m %p len %u\n",
3008		__func__, m0, m0->m_pkthdr.len);
3009	bus_dmamap_sync(sc->sc_dmat, bf->bf_dmamap, BUS_DMASYNC_PREWRITE);
3010	bf->bf_m = m0;
3011
3012	return 0;
3013}
3014
3015static __inline int
3016mwl_cvtlegacyrate(int rate)
3017{
3018	switch (rate) {
3019	case 2:	 return 0;
3020	case 4:	 return 1;
3021	case 11: return 2;
3022	case 22: return 3;
3023	case 44: return 4;
3024	case 12: return 5;
3025	case 18: return 6;
3026	case 24: return 7;
3027	case 36: return 8;
3028	case 48: return 9;
3029	case 72: return 10;
3030	case 96: return 11;
3031	case 108:return 12;
3032	}
3033	return 0;
3034}
3035
3036/*
3037 * Calculate fixed tx rate information per client state;
3038 * this value is suitable for writing to the Format field
3039 * of a tx descriptor.
3040 */
3041static uint16_t
3042mwl_calcformat(uint8_t rate, const struct ieee80211_node *ni)
3043{
3044	uint16_t fmt;
3045
3046	fmt = SM(3, EAGLE_TXD_ANTENNA)
3047	    | (IEEE80211_IS_CHAN_HT40D(ni->ni_chan) ?
3048		EAGLE_TXD_EXTCHAN_LO : EAGLE_TXD_EXTCHAN_HI);
3049	if (rate & IEEE80211_RATE_MCS) {	/* HT MCS */
3050		fmt |= EAGLE_TXD_FORMAT_HT
3051		    /* NB: 0x80 implicitly stripped from ucastrate */
3052		    | SM(rate, EAGLE_TXD_RATE);
3053		/* XXX short/long GI may be wrong; re-check */
3054		if (IEEE80211_IS_CHAN_HT40(ni->ni_chan)) {
3055			fmt |= EAGLE_TXD_CHW_40
3056			    | (ni->ni_htcap & IEEE80211_HTCAP_SHORTGI40 ?
3057			        EAGLE_TXD_GI_SHORT : EAGLE_TXD_GI_LONG);
3058		} else {
3059			fmt |= EAGLE_TXD_CHW_20
3060			    | (ni->ni_htcap & IEEE80211_HTCAP_SHORTGI20 ?
3061			        EAGLE_TXD_GI_SHORT : EAGLE_TXD_GI_LONG);
3062		}
3063	} else {			/* legacy rate */
3064		fmt |= EAGLE_TXD_FORMAT_LEGACY
3065		    | SM(mwl_cvtlegacyrate(rate), EAGLE_TXD_RATE)
3066		    | EAGLE_TXD_CHW_20
3067		    /* XXX iv_flags & IEEE80211_F_SHPREAMBLE? */
3068		    | (ni->ni_capinfo & IEEE80211_CAPINFO_SHORT_PREAMBLE ?
3069			EAGLE_TXD_PREAMBLE_SHORT : EAGLE_TXD_PREAMBLE_LONG);
3070	}
3071	return fmt;
3072}
3073
3074static int
3075mwl_tx_start(struct mwl_softc *sc, struct ieee80211_node *ni, struct mwl_txbuf *bf,
3076    struct mbuf *m0)
3077{
3078	struct ieee80211com *ic = &sc->sc_ic;
3079	struct ieee80211vap *vap = ni->ni_vap;
3080	int error, iswep, ismcast;
3081	int hdrlen, copyhdrlen, pktlen;
3082	struct mwl_txdesc *ds;
3083	struct mwl_txq *txq;
3084	struct ieee80211_frame *wh;
3085	struct mwltxrec *tr;
3086	struct mwl_node *mn;
3087	uint16_t qos;
3088#if MWL_TXDESC > 1
3089	int i;
3090#endif
3091
3092	wh = mtod(m0, struct ieee80211_frame *);
3093	iswep = wh->i_fc[1] & IEEE80211_FC1_PROTECTED;
3094	ismcast = IEEE80211_IS_MULTICAST(wh->i_addr1);
3095	hdrlen = ieee80211_anyhdrsize(wh);
3096	copyhdrlen = hdrlen;
3097	pktlen = m0->m_pkthdr.len;
3098	if (IEEE80211_QOS_HAS_SEQ(wh)) {
3099		if (IEEE80211_IS_DSTODS(wh)) {
3100			qos = *(uint16_t *)
3101			    (((struct ieee80211_qosframe_addr4 *) wh)->i_qos);
3102			copyhdrlen -= sizeof(qos);
3103		} else
3104			qos = *(uint16_t *)
3105			    (((struct ieee80211_qosframe *) wh)->i_qos);
3106	} else
3107		qos = 0;
3108
3109	if (iswep) {
3110		const struct ieee80211_cipher *cip;
3111		struct ieee80211_key *k;
3112
3113		/*
3114		 * Construct the 802.11 header+trailer for an encrypted
3115		 * frame. The only reason this can fail is because of an
3116		 * unknown or unsupported cipher/key type.
3117		 *
3118		 * NB: we do this even though the firmware will ignore
3119		 *     what we've done for WEP and TKIP as we need the
3120		 *     ExtIV filled in for CCMP and this also adjusts
3121		 *     the headers which simplifies our work below.
3122		 */
3123		k = ieee80211_crypto_encap(ni, m0);
3124		if (k == NULL) {
3125			/*
3126			 * This can happen when the key is yanked after the
3127			 * frame was queued.  Just discard the frame; the
3128			 * 802.11 layer counts failures and provides
3129			 * debugging/diagnostics.
3130			 */
3131			m_freem(m0);
3132			return EIO;
3133		}
3134		/*
3135		 * Adjust the packet length for the crypto additions
3136		 * done during encap and any other bits that the f/w
3137		 * will add later on.
3138		 */
3139		cip = k->wk_cipher;
3140		pktlen += cip->ic_header + cip->ic_miclen + cip->ic_trailer;
3141
3142		/* packet header may have moved, reset our local pointer */
3143		wh = mtod(m0, struct ieee80211_frame *);
3144	}
3145
3146	if (ieee80211_radiotap_active_vap(vap)) {
3147		sc->sc_tx_th.wt_flags = 0;	/* XXX */
3148		if (iswep)
3149			sc->sc_tx_th.wt_flags |= IEEE80211_RADIOTAP_F_WEP;
3150#if 0
3151		sc->sc_tx_th.wt_rate = ds->DataRate;
3152#endif
3153		sc->sc_tx_th.wt_txpower = ni->ni_txpower;
3154		sc->sc_tx_th.wt_antenna = sc->sc_txantenna;
3155
3156		ieee80211_radiotap_tx(vap, m0);
3157	}
3158	/*
3159	 * Copy up/down the 802.11 header; the firmware requires
3160	 * we present a 2-byte payload length followed by a
3161	 * 4-address header (w/o QoS), followed (optionally) by
3162	 * any WEP/ExtIV header (but only filled in for CCMP).
3163	 * We are assured the mbuf has sufficient headroom to
3164	 * prepend in-place by the setup of ic_headroom in
3165	 * mwl_attach.
3166	 */
3167	if (hdrlen < sizeof(struct mwltxrec)) {
3168		const int space = sizeof(struct mwltxrec) - hdrlen;
3169		if (M_LEADINGSPACE(m0) < space) {
3170			/* NB: should never happen */
3171			device_printf(sc->sc_dev,
3172			    "not enough headroom, need %d found %zd, "
3173			    "m_flags 0x%x m_len %d\n",
3174			    space, M_LEADINGSPACE(m0), m0->m_flags, m0->m_len);
3175			ieee80211_dump_pkt(ic,
3176			    mtod(m0, const uint8_t *), m0->m_len, 0, -1);
3177			m_freem(m0);
3178			sc->sc_stats.mst_tx_noheadroom++;
3179			return EIO;
3180		}
3181		M_PREPEND(m0, space, M_NOWAIT);
3182	}
3183	tr = mtod(m0, struct mwltxrec *);
3184	if (wh != (struct ieee80211_frame *) &tr->wh)
3185		ovbcopy(wh, &tr->wh, hdrlen);
3186	/*
3187	 * Note: the "firmware length" is actually the length
3188	 * of the fully formed "802.11 payload".  That is, it's
3189	 * everything except for the 802.11 header.  In particular
3190	 * this includes all crypto material including the MIC!
3191	 */
3192	tr->fwlen = htole16(pktlen - hdrlen);
3193
3194	/*
3195	 * Load the DMA map so any coalescing is done.  This
3196	 * also calculates the number of descriptors we need.
3197	 */
3198	error = mwl_tx_dmasetup(sc, bf, m0);
3199	if (error != 0) {
3200		/* NB: stat collected in mwl_tx_dmasetup */
3201		DPRINTF(sc, MWL_DEBUG_XMIT,
3202		    "%s: unable to setup dma\n", __func__);
3203		return error;
3204	}
3205	bf->bf_node = ni;			/* NB: held reference */
3206	m0 = bf->bf_m;				/* NB: may have changed */
3207	tr = mtod(m0, struct mwltxrec *);
3208	wh = (struct ieee80211_frame *)&tr->wh;
3209
3210	/*
3211	 * Formulate tx descriptor.
3212	 */
3213	ds = bf->bf_desc;
3214	txq = bf->bf_txq;
3215
3216	ds->QosCtrl = qos;			/* NB: already little-endian */
3217#if MWL_TXDESC == 1
3218	/*
3219	 * NB: multiframes should be zero because the descriptors
3220	 *     are initialized to zero.  This should handle the case
3221	 *     where the driver is built with MWL_TXDESC=1 but we are
3222	 *     using firmware with multi-segment support.
3223	 */
3224	ds->PktPtr = htole32(bf->bf_segs[0].ds_addr);
3225	ds->PktLen = htole16(bf->bf_segs[0].ds_len);
3226#else
3227	ds->multiframes = htole32(bf->bf_nseg);
3228	ds->PktLen = htole16(m0->m_pkthdr.len);
3229	for (i = 0; i < bf->bf_nseg; i++) {
3230		ds->PktPtrArray[i] = htole32(bf->bf_segs[i].ds_addr);
3231		ds->PktLenArray[i] = htole16(bf->bf_segs[i].ds_len);
3232	}
3233#endif
3234	/* NB: pPhysNext, DataRate, and SapPktInfo setup once, don't touch */
3235	ds->Format = 0;
3236	ds->pad = 0;
3237	ds->ack_wcb_addr = 0;
3238
3239	mn = MWL_NODE(ni);
3240	/*
3241	 * Select transmit rate.
3242	 */
3243	switch (wh->i_fc[0] & IEEE80211_FC0_TYPE_MASK) {
3244	case IEEE80211_FC0_TYPE_MGT:
3245		sc->sc_stats.mst_tx_mgmt++;
3246		/* fall thru... */
3247	case IEEE80211_FC0_TYPE_CTL:
3248		/* NB: assign to BE q to avoid bursting */
3249		ds->TxPriority = MWL_WME_AC_BE;
3250		break;
3251	case IEEE80211_FC0_TYPE_DATA:
3252		if (!ismcast) {
3253			const struct ieee80211_txparam *tp = ni->ni_txparms;
3254			/*
3255			 * EAPOL frames get forced to a fixed rate and w/o
3256			 * aggregation; otherwise check for any fixed rate
3257			 * for the client (may depend on association state).
3258			 */
3259			if (m0->m_flags & M_EAPOL) {
3260				const struct mwl_vap *mvp = MWL_VAP_CONST(vap);
3261				ds->Format = mvp->mv_eapolformat;
3262				ds->pad = htole16(
3263				    EAGLE_TXD_FIXED_RATE | EAGLE_TXD_DONT_AGGR);
3264			} else if (tp->ucastrate != IEEE80211_FIXED_RATE_NONE) {
3265				/* XXX pre-calculate per node */
3266				ds->Format = htole16(
3267				    mwl_calcformat(tp->ucastrate, ni));
3268				ds->pad = htole16(EAGLE_TXD_FIXED_RATE);
3269			}
3270			/* NB: EAPOL frames will never have qos set */
3271			if (qos == 0)
3272				ds->TxPriority = txq->qnum;
3273#if MWL_MAXBA > 3
3274			else if (mwl_bastream_match(&mn->mn_ba[3], qos))
3275				ds->TxPriority = mn->mn_ba[3].txq;
3276#endif
3277#if MWL_MAXBA > 2
3278			else if (mwl_bastream_match(&mn->mn_ba[2], qos))
3279				ds->TxPriority = mn->mn_ba[2].txq;
3280#endif
3281#if MWL_MAXBA > 1
3282			else if (mwl_bastream_match(&mn->mn_ba[1], qos))
3283				ds->TxPriority = mn->mn_ba[1].txq;
3284#endif
3285#if MWL_MAXBA > 0
3286			else if (mwl_bastream_match(&mn->mn_ba[0], qos))
3287				ds->TxPriority = mn->mn_ba[0].txq;
3288#endif
3289			else
3290				ds->TxPriority = txq->qnum;
3291		} else
3292			ds->TxPriority = txq->qnum;
3293		break;
3294	default:
3295		device_printf(sc->sc_dev, "bogus frame type 0x%x (%s)\n",
3296			wh->i_fc[0] & IEEE80211_FC0_TYPE_MASK, __func__);
3297		sc->sc_stats.mst_tx_badframetype++;
3298		m_freem(m0);
3299		return EIO;
3300	}
3301
3302	if (IFF_DUMPPKTS_XMIT(sc))
3303		ieee80211_dump_pkt(ic,
3304		    mtod(m0, const uint8_t *)+sizeof(uint16_t),
3305		    m0->m_len - sizeof(uint16_t), ds->DataRate, -1);
3306
3307	MWL_TXQ_LOCK(txq);
3308	ds->Status = htole32(EAGLE_TXD_STATUS_FW_OWNED);
3309	STAILQ_INSERT_TAIL(&txq->active, bf, bf_list);
3310	MWL_TXDESC_SYNC(txq, ds, BUS_DMASYNC_PREREAD | BUS_DMASYNC_PREWRITE);
3311
3312	sc->sc_tx_timer = 5;
3313	MWL_TXQ_UNLOCK(txq);
3314
3315	return 0;
3316}
3317
3318static __inline int
3319mwl_cvtlegacyrix(int rix)
3320{
3321	static const int ieeerates[] =
3322	    { 2, 4, 11, 22, 44, 12, 18, 24, 36, 48, 72, 96, 108 };
3323	return (rix < nitems(ieeerates) ? ieeerates[rix] : 0);
3324}
3325
3326/*
3327 * Process completed xmit descriptors from the specified queue.
3328 */
3329static int
3330mwl_tx_processq(struct mwl_softc *sc, struct mwl_txq *txq)
3331{
3332#define	EAGLE_TXD_STATUS_MCAST \
3333	(EAGLE_TXD_STATUS_MULTICAST_TX | EAGLE_TXD_STATUS_BROADCAST_TX)
3334	struct ieee80211com *ic = &sc->sc_ic;
3335	struct mwl_txbuf *bf;
3336	struct mwl_txdesc *ds;
3337	struct ieee80211_node *ni;
3338	struct mwl_node *an;
3339	int nreaped;
3340	uint32_t status;
3341
3342	DPRINTF(sc, MWL_DEBUG_TX_PROC, "%s: tx queue %u\n", __func__, txq->qnum);
3343	for (nreaped = 0;; nreaped++) {
3344		MWL_TXQ_LOCK(txq);
3345		bf = STAILQ_FIRST(&txq->active);
3346		if (bf == NULL) {
3347			MWL_TXQ_UNLOCK(txq);
3348			break;
3349		}
3350		ds = bf->bf_desc;
3351		MWL_TXDESC_SYNC(txq, ds,
3352		    BUS_DMASYNC_POSTREAD | BUS_DMASYNC_POSTWRITE);
3353		if (ds->Status & htole32(EAGLE_TXD_STATUS_FW_OWNED)) {
3354			MWL_TXQ_UNLOCK(txq);
3355			break;
3356		}
3357		STAILQ_REMOVE_HEAD(&txq->active, bf_list);
3358		MWL_TXQ_UNLOCK(txq);
3359
3360#ifdef MWL_DEBUG
3361		if (sc->sc_debug & MWL_DEBUG_XMIT_DESC)
3362			mwl_printtxbuf(bf, txq->qnum, nreaped);
3363#endif
3364		ni = bf->bf_node;
3365		if (ni != NULL) {
3366			an = MWL_NODE(ni);
3367			status = le32toh(ds->Status);
3368			if (status & EAGLE_TXD_STATUS_OK) {
3369				uint16_t Format = le16toh(ds->Format);
3370				uint8_t txant = MS(Format, EAGLE_TXD_ANTENNA);
3371
3372				sc->sc_stats.mst_ant_tx[txant]++;
3373				if (status & EAGLE_TXD_STATUS_OK_RETRY)
3374					sc->sc_stats.mst_tx_retries++;
3375				if (status & EAGLE_TXD_STATUS_OK_MORE_RETRY)
3376					sc->sc_stats.mst_tx_mretries++;
3377				if (txq->qnum >= MWL_WME_AC_VO)
3378					ic->ic_wme.wme_hipri_traffic++;
3379				ni->ni_txrate = MS(Format, EAGLE_TXD_RATE);
3380				if ((Format & EAGLE_TXD_FORMAT_HT) == 0) {
3381					ni->ni_txrate = mwl_cvtlegacyrix(
3382					    ni->ni_txrate);
3383				} else
3384					ni->ni_txrate |= IEEE80211_RATE_MCS;
3385				sc->sc_stats.mst_tx_rate = ni->ni_txrate;
3386			} else {
3387				if (status & EAGLE_TXD_STATUS_FAILED_LINK_ERROR)
3388					sc->sc_stats.mst_tx_linkerror++;
3389				if (status & EAGLE_TXD_STATUS_FAILED_XRETRY)
3390					sc->sc_stats.mst_tx_xretries++;
3391				if (status & EAGLE_TXD_STATUS_FAILED_AGING)
3392					sc->sc_stats.mst_tx_aging++;
3393				if (bf->bf_m->m_flags & M_FF)
3394					sc->sc_stats.mst_ff_txerr++;
3395			}
3396			if (bf->bf_m->m_flags & M_TXCB)
3397				/* XXX strip fw len in case header inspected */
3398				m_adj(bf->bf_m, sizeof(uint16_t));
3399			ieee80211_tx_complete(ni, bf->bf_m,
3400			    (status & EAGLE_TXD_STATUS_OK) == 0);
3401		} else
3402			m_freem(bf->bf_m);
3403		ds->Status = htole32(EAGLE_TXD_STATUS_IDLE);
3404
3405		bus_dmamap_sync(sc->sc_dmat, bf->bf_dmamap,
3406		    BUS_DMASYNC_POSTWRITE);
3407		bus_dmamap_unload(sc->sc_dmat, bf->bf_dmamap);
3408
3409		mwl_puttxbuf_tail(txq, bf);
3410	}
3411	return nreaped;
3412#undef EAGLE_TXD_STATUS_MCAST
3413}
3414
3415/*
3416 * Deferred processing of transmit interrupt; special-cased
3417 * for four hardware queues, 0-3.
3418 */
3419static void
3420mwl_tx_proc(void *arg, int npending)
3421{
3422	struct mwl_softc *sc = arg;
3423	int nreaped;
3424
3425	/*
3426	 * Process each active queue.
3427	 */
3428	nreaped = 0;
3429	if (!STAILQ_EMPTY(&sc->sc_txq[0].active))
3430		nreaped += mwl_tx_processq(sc, &sc->sc_txq[0]);
3431	if (!STAILQ_EMPTY(&sc->sc_txq[1].active))
3432		nreaped += mwl_tx_processq(sc, &sc->sc_txq[1]);
3433	if (!STAILQ_EMPTY(&sc->sc_txq[2].active))
3434		nreaped += mwl_tx_processq(sc, &sc->sc_txq[2]);
3435	if (!STAILQ_EMPTY(&sc->sc_txq[3].active))
3436		nreaped += mwl_tx_processq(sc, &sc->sc_txq[3]);
3437
3438	if (nreaped != 0) {
3439		sc->sc_tx_timer = 0;
3440		if (mbufq_first(&sc->sc_snd) != NULL) {
3441			/* NB: kick fw; the tx thread may have been preempted */
3442			mwl_hal_txstart(sc->sc_mh, 0);
3443			mwl_start(sc);
3444		}
3445	}
3446}
3447
3448static void
3449mwl_tx_draintxq(struct mwl_softc *sc, struct mwl_txq *txq)
3450{
3451	struct ieee80211_node *ni;
3452	struct mwl_txbuf *bf;
3453	u_int ix;
3454
3455	/*
3456	 * NB: this assumes output has been stopped and
3457	 *     we do not need to block mwl_tx_tasklet
3458	 */
3459	for (ix = 0;; ix++) {
3460		MWL_TXQ_LOCK(txq);
3461		bf = STAILQ_FIRST(&txq->active);
3462		if (bf == NULL) {
3463			MWL_TXQ_UNLOCK(txq);
3464			break;
3465		}
3466		STAILQ_REMOVE_HEAD(&txq->active, bf_list);
3467		MWL_TXQ_UNLOCK(txq);
3468#ifdef MWL_DEBUG
3469		if (sc->sc_debug & MWL_DEBUG_RESET) {
3470			struct ieee80211com *ic = &sc->sc_ic;
3471			const struct mwltxrec *tr =
3472			    mtod(bf->bf_m, const struct mwltxrec *);
3473			mwl_printtxbuf(bf, txq->qnum, ix);
3474			ieee80211_dump_pkt(ic, (const uint8_t *)&tr->wh,
3475				bf->bf_m->m_len - sizeof(tr->fwlen), 0, -1);
3476		}
3477#endif /* MWL_DEBUG */
3478		bus_dmamap_unload(sc->sc_dmat, bf->bf_dmamap);
3479		ni = bf->bf_node;
3480		if (ni != NULL) {
3481			/*
3482			 * Reclaim node reference.
3483			 */
3484			ieee80211_free_node(ni);
3485		}
3486		m_freem(bf->bf_m);
3487
3488		mwl_puttxbuf_tail(txq, bf);
3489	}
3490}
3491
3492/*
3493 * Drain the transmit queues and reclaim resources.
3494 */
3495static void
3496mwl_draintxq(struct mwl_softc *sc)
3497{
3498	int i;
3499
3500	for (i = 0; i < MWL_NUM_TX_QUEUES; i++)
3501		mwl_tx_draintxq(sc, &sc->sc_txq[i]);
3502	sc->sc_tx_timer = 0;
3503}
3504
3505#ifdef MWL_DIAGAPI
3506/*
3507 * Reset the transmit queues to a pristine state after a fw download.
3508 */
3509static void
3510mwl_resettxq(struct mwl_softc *sc)
3511{
3512	int i;
3513
3514	for (i = 0; i < MWL_NUM_TX_QUEUES; i++)
3515		mwl_txq_reset(sc, &sc->sc_txq[i]);
3516}
3517#endif /* MWL_DIAGAPI */
3518
3519/*
3520 * Clear the transmit queues of any frames submitted for the
3521 * specified vap.  This is done when the vap is deleted so we
3522 * don't potentially reference the vap after it is gone.
3523 * Note we cannot remove the frames; we only reclaim the node
3524 * reference.
3525 */
3526static void
3527mwl_cleartxq(struct mwl_softc *sc, struct ieee80211vap *vap)
3528{
3529	struct mwl_txq *txq;
3530	struct mwl_txbuf *bf;
3531	int i;
3532
3533	for (i = 0; i < MWL_NUM_TX_QUEUES; i++) {
3534		txq = &sc->sc_txq[i];
3535		MWL_TXQ_LOCK(txq);
3536		STAILQ_FOREACH(bf, &txq->active, bf_list) {
3537			struct ieee80211_node *ni = bf->bf_node;
3538			if (ni != NULL && ni->ni_vap == vap) {
3539				bf->bf_node = NULL;
3540				ieee80211_free_node(ni);
3541			}
3542		}
3543		MWL_TXQ_UNLOCK(txq);
3544	}
3545}
3546
3547static int
3548mwl_recv_action(struct ieee80211_node *ni, const struct ieee80211_frame *wh,
3549	const uint8_t *frm, const uint8_t *efrm)
3550{
3551	struct mwl_softc *sc = ni->ni_ic->ic_softc;
3552	const struct ieee80211_action *ia;
3553
3554	ia = (const struct ieee80211_action *) frm;
3555	if (ia->ia_category == IEEE80211_ACTION_CAT_HT &&
3556	    ia->ia_action == IEEE80211_ACTION_HT_MIMOPWRSAVE) {
3557		const struct ieee80211_action_ht_mimopowersave *mps =
3558		    (const struct ieee80211_action_ht_mimopowersave *) ia;
3559
3560		mwl_hal_setmimops(sc->sc_mh, ni->ni_macaddr,
3561		    mps->am_control & IEEE80211_A_HT_MIMOPWRSAVE_ENA,
3562		    MS(mps->am_control, IEEE80211_A_HT_MIMOPWRSAVE_MODE));
3563		return 0;
3564	} else
3565		return sc->sc_recv_action(ni, wh, frm, efrm);
3566}
3567
3568static int
3569mwl_addba_request(struct ieee80211_node *ni, struct ieee80211_tx_ampdu *tap,
3570	int dialogtoken, int baparamset, int batimeout)
3571{
3572	struct mwl_softc *sc = ni->ni_ic->ic_softc;
3573	struct ieee80211vap *vap = ni->ni_vap;
3574	struct mwl_node *mn = MWL_NODE(ni);
3575	struct mwl_bastate *bas;
3576
3577	bas = tap->txa_private;
3578	if (bas == NULL) {
3579		const MWL_HAL_BASTREAM *sp;
3580		/*
3581		 * Check for a free BA stream slot.
3582		 */
3583#if MWL_MAXBA > 3
3584		if (mn->mn_ba[3].bastream == NULL)
3585			bas = &mn->mn_ba[3];
3586		else
3587#endif
3588#if MWL_MAXBA > 2
3589		if (mn->mn_ba[2].bastream == NULL)
3590			bas = &mn->mn_ba[2];
3591		else
3592#endif
3593#if MWL_MAXBA > 1
3594		if (mn->mn_ba[1].bastream == NULL)
3595			bas = &mn->mn_ba[1];
3596		else
3597#endif
3598#if MWL_MAXBA > 0
3599		if (mn->mn_ba[0].bastream == NULL)
3600			bas = &mn->mn_ba[0];
3601		else
3602#endif
3603		{
3604			/* sta already has max BA streams */
3605			/* XXX assign BA stream to highest priority tid */
3606			DPRINTF(sc, MWL_DEBUG_AMPDU,
3607			    "%s: already has max bastreams\n", __func__);
3608			sc->sc_stats.mst_ampdu_reject++;
3609			return 0;
3610		}
3611		/* NB: no held reference to ni */
3612		sp = mwl_hal_bastream_alloc(MWL_VAP(vap)->mv_hvap,
3613		    (baparamset & IEEE80211_BAPS_POLICY_IMMEDIATE) != 0,
3614		    ni->ni_macaddr, tap->txa_tid, ni->ni_htparam,
3615		    ni, tap);
3616		if (sp == NULL) {
3617			/*
3618			 * No available stream, return 0 so no
3619			 * a-mpdu aggregation will be done.
3620			 */
3621			DPRINTF(sc, MWL_DEBUG_AMPDU,
3622			    "%s: no bastream available\n", __func__);
3623			sc->sc_stats.mst_ampdu_nostream++;
3624			return 0;
3625		}
3626		DPRINTF(sc, MWL_DEBUG_AMPDU, "%s: alloc bastream %p\n",
3627		    __func__, sp);
3628		/* NB: qos is left zero so we won't match in mwl_tx_start */
3629		bas->bastream = sp;
3630		tap->txa_private = bas;
3631	}
3632	/* fetch current seq# from the firmware; if available */
3633	if (mwl_hal_bastream_get_seqno(sc->sc_mh, bas->bastream,
3634	    vap->iv_opmode == IEEE80211_M_STA ? vap->iv_myaddr : ni->ni_macaddr,
3635	    &tap->txa_start) != 0)
3636		tap->txa_start = 0;
3637	return sc->sc_addba_request(ni, tap, dialogtoken, baparamset, batimeout);
3638}
3639
3640static int
3641mwl_addba_response(struct ieee80211_node *ni, struct ieee80211_tx_ampdu *tap,
3642	int code, int baparamset, int batimeout)
3643{
3644	struct mwl_softc *sc = ni->ni_ic->ic_softc;
3645	struct mwl_bastate *bas;
3646
3647	bas = tap->txa_private;
3648	if (bas == NULL) {
3649		/* XXX should not happen */
3650		DPRINTF(sc, MWL_DEBUG_AMPDU,
3651		    "%s: no BA stream allocated, TID %d\n",
3652		    __func__, tap->txa_tid);
3653		sc->sc_stats.mst_addba_nostream++;
3654		return 0;
3655	}
3656	if (code == IEEE80211_STATUS_SUCCESS) {
3657		struct ieee80211vap *vap = ni->ni_vap;
3658		int bufsiz, error;
3659
3660		/*
3661		 * Tell the firmware to setup the BA stream;
3662		 * we know resources are available because we
3663		 * pre-allocated one before forming the request.
3664		 */
3665		bufsiz = MS(baparamset, IEEE80211_BAPS_BUFSIZ);
3666		if (bufsiz == 0)
3667			bufsiz = IEEE80211_AGGR_BAWMAX;
3668		error = mwl_hal_bastream_create(MWL_VAP(vap)->mv_hvap,
3669		    bas->bastream, bufsiz, bufsiz, tap->txa_start);
3670		if (error != 0) {
3671			/*
3672			 * Setup failed, return immediately so no a-mpdu
3673			 * aggregation will be done.
3674			 */
3675			mwl_hal_bastream_destroy(sc->sc_mh, bas->bastream);
3676			mwl_bastream_free(bas);
3677			tap->txa_private = NULL;
3678
3679			DPRINTF(sc, MWL_DEBUG_AMPDU,
3680			    "%s: create failed, error %d, bufsiz %d TID %d "
3681			    "htparam 0x%x\n", __func__, error, bufsiz,
3682			    tap->txa_tid, ni->ni_htparam);
3683			sc->sc_stats.mst_bacreate_failed++;
3684			return 0;
3685		}
3686		/* NB: cache txq to avoid ptr indirect */
3687		mwl_bastream_setup(bas, tap->txa_tid, bas->bastream->txq);
3688		DPRINTF(sc, MWL_DEBUG_AMPDU,
3689		    "%s: bastream %p assigned to txq %d TID %d bufsiz %d "
3690		    "htparam 0x%x\n", __func__, bas->bastream,
3691		    bas->txq, tap->txa_tid, bufsiz, ni->ni_htparam);
3692	} else {
3693		/*
3694		 * Other side NAK'd us; return the resources.
3695		 */
3696		DPRINTF(sc, MWL_DEBUG_AMPDU,
3697		    "%s: request failed with code %d, destroy bastream %p\n",
3698		    __func__, code, bas->bastream);
3699		mwl_hal_bastream_destroy(sc->sc_mh, bas->bastream);
3700		mwl_bastream_free(bas);
3701		tap->txa_private = NULL;
3702	}
3703	/* NB: firmware sends BAR so we don't need to */
3704	return sc->sc_addba_response(ni, tap, code, baparamset, batimeout);
3705}
3706
3707static void
3708mwl_addba_stop(struct ieee80211_node *ni, struct ieee80211_tx_ampdu *tap)
3709{
3710	struct mwl_softc *sc = ni->ni_ic->ic_softc;
3711	struct mwl_bastate *bas;
3712
3713	bas = tap->txa_private;
3714	if (bas != NULL) {
3715		DPRINTF(sc, MWL_DEBUG_AMPDU, "%s: destroy bastream %p\n",
3716		    __func__, bas->bastream);
3717		mwl_hal_bastream_destroy(sc->sc_mh, bas->bastream);
3718		mwl_bastream_free(bas);
3719		tap->txa_private = NULL;
3720	}
3721	sc->sc_addba_stop(ni, tap);
3722}
3723
3724/*
3725 * Setup the rx data structures.  This should only be
3726 * done once or we may get out of sync with the firmware.
3727 */
3728static int
3729mwl_startrecv(struct mwl_softc *sc)
3730{
3731	if (!sc->sc_recvsetup) {
3732		struct mwl_rxbuf *bf, *prev;
3733		struct mwl_rxdesc *ds;
3734
3735		prev = NULL;
3736		STAILQ_FOREACH(bf, &sc->sc_rxbuf, bf_list) {
3737			int error = mwl_rxbuf_init(sc, bf);
3738			if (error != 0) {
3739				DPRINTF(sc, MWL_DEBUG_RECV,
3740					"%s: mwl_rxbuf_init failed %d\n",
3741					__func__, error);
3742				return error;
3743			}
3744			if (prev != NULL) {
3745				ds = prev->bf_desc;
3746				ds->pPhysNext = htole32(bf->bf_daddr);
3747			}
3748			prev = bf;
3749		}
3750		if (prev != NULL) {
3751			ds = prev->bf_desc;
3752			ds->pPhysNext =
3753			    htole32(STAILQ_FIRST(&sc->sc_rxbuf)->bf_daddr);
3754		}
3755		sc->sc_recvsetup = 1;
3756	}
3757	mwl_mode_init(sc);		/* set filters, etc. */
3758	return 0;
3759}
3760
3761static MWL_HAL_APMODE
3762mwl_getapmode(const struct ieee80211vap *vap, struct ieee80211_channel *chan)
3763{
3764	MWL_HAL_APMODE mode;
3765
3766	if (IEEE80211_IS_CHAN_HT(chan)) {
3767		if (vap->iv_flags_ht & IEEE80211_FHT_PUREN)
3768			mode = AP_MODE_N_ONLY;
3769		else if (IEEE80211_IS_CHAN_5GHZ(chan))
3770			mode = AP_MODE_AandN;
3771		else if (vap->iv_flags & IEEE80211_F_PUREG)
3772			mode = AP_MODE_GandN;
3773		else
3774			mode = AP_MODE_BandGandN;
3775	} else if (IEEE80211_IS_CHAN_ANYG(chan)) {
3776		if (vap->iv_flags & IEEE80211_F_PUREG)
3777			mode = AP_MODE_G_ONLY;
3778		else
3779			mode = AP_MODE_MIXED;
3780	} else if (IEEE80211_IS_CHAN_B(chan))
3781		mode = AP_MODE_B_ONLY;
3782	else if (IEEE80211_IS_CHAN_A(chan))
3783		mode = AP_MODE_A_ONLY;
3784	else
3785		mode = AP_MODE_MIXED;		/* XXX should not happen? */
3786	return mode;
3787}
3788
3789static int
3790mwl_setapmode(struct ieee80211vap *vap, struct ieee80211_channel *chan)
3791{
3792	struct mwl_hal_vap *hvap = MWL_VAP(vap)->mv_hvap;
3793	return mwl_hal_setapmode(hvap, mwl_getapmode(vap, chan));
3794}
3795
3796/*
3797 * Set/change channels.
3798 */
3799static int
3800mwl_chan_set(struct mwl_softc *sc, struct ieee80211_channel *chan)
3801{
3802	struct mwl_hal *mh = sc->sc_mh;
3803	struct ieee80211com *ic = &sc->sc_ic;
3804	MWL_HAL_CHANNEL hchan;
3805	int maxtxpow;
3806
3807	DPRINTF(sc, MWL_DEBUG_RESET, "%s: chan %u MHz/flags 0x%x\n",
3808	    __func__, chan->ic_freq, chan->ic_flags);
3809
3810	/*
3811	 * Convert to a HAL channel description with
3812	 * the flags constrained to reflect the current
3813	 * operating mode.
3814	 */
3815	mwl_mapchan(&hchan, chan);
3816	mwl_hal_intrset(mh, 0);		/* disable interrupts */
3817#if 0
3818	mwl_draintxq(sc);		/* clear pending tx frames */
3819#endif
3820	mwl_hal_setchannel(mh, &hchan);
3821	/*
3822	 * Tx power is cap'd by the regulatory setting and
3823	 * possibly a user-set limit.  We pass the min of
3824	 * these to the hal to apply them to the cal data
3825	 * for this channel.
3826	 * XXX min bound?
3827	 */
3828	maxtxpow = 2*chan->ic_maxregpower;
3829	if (maxtxpow > ic->ic_txpowlimit)
3830		maxtxpow = ic->ic_txpowlimit;
3831	mwl_hal_settxpower(mh, &hchan, maxtxpow / 2);
3832	/* NB: potentially change mcast/mgt rates */
3833	mwl_setcurchanrates(sc);
3834
3835	/*
3836	 * Update internal state.
3837	 */
3838	sc->sc_tx_th.wt_chan_freq = htole16(chan->ic_freq);
3839	sc->sc_rx_th.wr_chan_freq = htole16(chan->ic_freq);
3840	if (IEEE80211_IS_CHAN_A(chan)) {
3841		sc->sc_tx_th.wt_chan_flags = htole16(IEEE80211_CHAN_A);
3842		sc->sc_rx_th.wr_chan_flags = htole16(IEEE80211_CHAN_A);
3843	} else if (IEEE80211_IS_CHAN_ANYG(chan)) {
3844		sc->sc_tx_th.wt_chan_flags = htole16(IEEE80211_CHAN_G);
3845		sc->sc_rx_th.wr_chan_flags = htole16(IEEE80211_CHAN_G);
3846	} else {
3847		sc->sc_tx_th.wt_chan_flags = htole16(IEEE80211_CHAN_B);
3848		sc->sc_rx_th.wr_chan_flags = htole16(IEEE80211_CHAN_B);
3849	}
3850	sc->sc_curchan = hchan;
3851	mwl_hal_intrset(mh, sc->sc_imask);
3852
3853	return 0;
3854}
3855
3856static void
3857mwl_scan_start(struct ieee80211com *ic)
3858{
3859	struct mwl_softc *sc = ic->ic_softc;
3860
3861	DPRINTF(sc, MWL_DEBUG_STATE, "%s\n", __func__);
3862}
3863
3864static void
3865mwl_scan_end(struct ieee80211com *ic)
3866{
3867	struct mwl_softc *sc = ic->ic_softc;
3868
3869	DPRINTF(sc, MWL_DEBUG_STATE, "%s\n", __func__);
3870}
3871
3872static void
3873mwl_set_channel(struct ieee80211com *ic)
3874{
3875	struct mwl_softc *sc = ic->ic_softc;
3876
3877	(void) mwl_chan_set(sc, ic->ic_curchan);
3878}
3879
3880/*
3881 * Handle a channel switch request.  We inform the firmware
3882 * and mark the global state to suppress various actions.
3883 * NB: we issue only one request to the fw; we may be called
3884 * multiple times if there are multiple vap's.
3885 */
3886static void
3887mwl_startcsa(struct ieee80211vap *vap)
3888{
3889	struct ieee80211com *ic = vap->iv_ic;
3890	struct mwl_softc *sc = ic->ic_softc;
3891	MWL_HAL_CHANNEL hchan;
3892
3893	if (sc->sc_csapending)
3894		return;
3895
3896	mwl_mapchan(&hchan, ic->ic_csa_newchan);
3897	/* 1 =>'s quiet channel */
3898	mwl_hal_setchannelswitchie(sc->sc_mh, &hchan, 1, ic->ic_csa_count);
3899	sc->sc_csapending = 1;
3900}
3901
3902/*
3903 * Plumb any static WEP key for the station.  This is
3904 * necessary as we must propagate the key from the
3905 * global key table of the vap to each sta db entry.
3906 */
3907static void
3908mwl_setanywepkey(struct ieee80211vap *vap, const uint8_t mac[IEEE80211_ADDR_LEN])
3909{
3910	if ((vap->iv_flags & (IEEE80211_F_PRIVACY|IEEE80211_F_WPA)) ==
3911		IEEE80211_F_PRIVACY &&
3912	    vap->iv_def_txkey != IEEE80211_KEYIX_NONE &&
3913	    vap->iv_nw_keys[vap->iv_def_txkey].wk_keyix != IEEE80211_KEYIX_NONE)
3914		(void) mwl_key_set(vap, &vap->iv_nw_keys[vap->iv_def_txkey], mac);
3915}
3916
3917static int
3918mwl_peerstadb(struct ieee80211_node *ni, int aid, int staid, MWL_HAL_PEERINFO *pi)
3919{
3920#define	WME(ie) ((const struct ieee80211_wme_info *) ie)
3921	struct ieee80211vap *vap = ni->ni_vap;
3922	struct mwl_hal_vap *hvap;
3923	int error;
3924
3925	if (vap->iv_opmode == IEEE80211_M_WDS) {
3926		/*
3927		 * WDS vap's do not have a f/w vap; instead they piggyback
3928		 * on an AP vap and we must install the sta db entry and
3929		 * crypto state using that AP's handle (the WDS vap has none).
3930		 */
3931		hvap = MWL_VAP(vap)->mv_ap_hvap;
3932	} else
3933		hvap = MWL_VAP(vap)->mv_hvap;
3934	error = mwl_hal_newstation(hvap, ni->ni_macaddr,
3935	    aid, staid, pi,
3936	    ni->ni_flags & (IEEE80211_NODE_QOS | IEEE80211_NODE_HT),
3937	    ni->ni_ies.wme_ie != NULL ? WME(ni->ni_ies.wme_ie)->wme_info : 0);
3938	if (error == 0) {
3939		/*
3940		 * Setup security for this station.  For sta mode this is
3941		 * needed even though do the same thing on transition to
3942		 * AUTH state because the call to mwl_hal_newstation
3943		 * clobbers the crypto state we setup.
3944		 */
3945		mwl_setanywepkey(vap, ni->ni_macaddr);
3946	}
3947	return error;
3948#undef WME
3949}
3950
3951static void
3952mwl_setglobalkeys(struct ieee80211vap *vap)
3953{
3954	struct ieee80211_key *wk;
3955
3956	wk = &vap->iv_nw_keys[0];
3957	for (; wk < &vap->iv_nw_keys[IEEE80211_WEP_NKID]; wk++)
3958		if (wk->wk_keyix != IEEE80211_KEYIX_NONE)
3959			(void) mwl_key_set(vap, wk, vap->iv_myaddr);
3960}
3961
3962/*
3963 * Convert a legacy rate set to a firmware bitmask.
3964 */
3965static uint32_t
3966get_rate_bitmap(const struct ieee80211_rateset *rs)
3967{
3968	uint32_t rates;
3969	int i;
3970
3971	rates = 0;
3972	for (i = 0; i < rs->rs_nrates; i++)
3973		switch (rs->rs_rates[i] & IEEE80211_RATE_VAL) {
3974		case 2:	  rates |= 0x001; break;
3975		case 4:	  rates |= 0x002; break;
3976		case 11:  rates |= 0x004; break;
3977		case 22:  rates |= 0x008; break;
3978		case 44:  rates |= 0x010; break;
3979		case 12:  rates |= 0x020; break;
3980		case 18:  rates |= 0x040; break;
3981		case 24:  rates |= 0x080; break;
3982		case 36:  rates |= 0x100; break;
3983		case 48:  rates |= 0x200; break;
3984		case 72:  rates |= 0x400; break;
3985		case 96:  rates |= 0x800; break;
3986		case 108: rates |= 0x1000; break;
3987		}
3988	return rates;
3989}
3990
3991/*
3992 * Construct an HT firmware bitmask from an HT rate set.
3993 */
3994static uint32_t
3995get_htrate_bitmap(const struct ieee80211_htrateset *rs)
3996{
3997	uint32_t rates;
3998	int i;
3999
4000	rates = 0;
4001	for (i = 0; i < rs->rs_nrates; i++) {
4002		if (rs->rs_rates[i] < 16)
4003			rates |= 1<<rs->rs_rates[i];
4004	}
4005	return rates;
4006}
4007
4008/*
4009 * Craft station database entry for station.
4010 * NB: use host byte order here, the hal handles byte swapping.
4011 */
4012static MWL_HAL_PEERINFO *
4013mkpeerinfo(MWL_HAL_PEERINFO *pi, const struct ieee80211_node *ni)
4014{
4015	const struct ieee80211vap *vap = ni->ni_vap;
4016
4017	memset(pi, 0, sizeof(*pi));
4018	pi->LegacyRateBitMap = get_rate_bitmap(&ni->ni_rates);
4019	pi->CapInfo = ni->ni_capinfo;
4020	if (ni->ni_flags & IEEE80211_NODE_HT) {
4021		/* HT capabilities, etc */
4022		pi->HTCapabilitiesInfo = ni->ni_htcap;
4023		/* XXX pi.HTCapabilitiesInfo */
4024	        pi->MacHTParamInfo = ni->ni_htparam;
4025		pi->HTRateBitMap = get_htrate_bitmap(&ni->ni_htrates);
4026		pi->AddHtInfo.ControlChan = ni->ni_htctlchan;
4027		pi->AddHtInfo.AddChan = ni->ni_ht2ndchan;
4028		pi->AddHtInfo.OpMode = ni->ni_htopmode;
4029		pi->AddHtInfo.stbc = ni->ni_htstbc;
4030
4031		/* constrain according to local configuration */
4032		if ((vap->iv_flags_ht & IEEE80211_FHT_SHORTGI40) == 0)
4033			pi->HTCapabilitiesInfo &= ~IEEE80211_HTCAP_SHORTGI40;
4034		if ((vap->iv_flags_ht & IEEE80211_FHT_SHORTGI20) == 0)
4035			pi->HTCapabilitiesInfo &= ~IEEE80211_HTCAP_SHORTGI20;
4036		if (ni->ni_chw != 40)
4037			pi->HTCapabilitiesInfo &= ~IEEE80211_HTCAP_CHWIDTH40;
4038	}
4039	return pi;
4040}
4041
4042/*
4043 * Re-create the local sta db entry for a vap to ensure
4044 * up to date WME state is pushed to the firmware.  Because
4045 * this resets crypto state this must be followed by a
4046 * reload of any keys in the global key table.
4047 */
4048static int
4049mwl_localstadb(struct ieee80211vap *vap)
4050{
4051#define	WME(ie) ((const struct ieee80211_wme_info *) ie)
4052	struct mwl_hal_vap *hvap = MWL_VAP(vap)->mv_hvap;
4053	struct ieee80211_node *bss;
4054	MWL_HAL_PEERINFO pi;
4055	int error;
4056
4057	switch (vap->iv_opmode) {
4058	case IEEE80211_M_STA:
4059		bss = vap->iv_bss;
4060		error = mwl_hal_newstation(hvap, vap->iv_myaddr, 0, 0,
4061		    vap->iv_state == IEEE80211_S_RUN ?
4062			mkpeerinfo(&pi, bss) : NULL,
4063		    (bss->ni_flags & (IEEE80211_NODE_QOS | IEEE80211_NODE_HT)),
4064		    bss->ni_ies.wme_ie != NULL ?
4065			WME(bss->ni_ies.wme_ie)->wme_info : 0);
4066		if (error == 0)
4067			mwl_setglobalkeys(vap);
4068		break;
4069	case IEEE80211_M_HOSTAP:
4070	case IEEE80211_M_MBSS:
4071		error = mwl_hal_newstation(hvap, vap->iv_myaddr,
4072		    0, 0, NULL, vap->iv_flags & IEEE80211_F_WME, 0);
4073		if (error == 0)
4074			mwl_setglobalkeys(vap);
4075		break;
4076	default:
4077		error = 0;
4078		break;
4079	}
4080	return error;
4081#undef WME
4082}
4083
4084static int
4085mwl_newstate(struct ieee80211vap *vap, enum ieee80211_state nstate, int arg)
4086{
4087	struct mwl_vap *mvp = MWL_VAP(vap);
4088	struct mwl_hal_vap *hvap = mvp->mv_hvap;
4089	struct ieee80211com *ic = vap->iv_ic;
4090	struct ieee80211_node *ni = NULL;
4091	struct mwl_softc *sc = ic->ic_softc;
4092	struct mwl_hal *mh = sc->sc_mh;
4093	enum ieee80211_state ostate = vap->iv_state;
4094	int error;
4095
4096	DPRINTF(sc, MWL_DEBUG_STATE, "%s: %s: %s -> %s\n",
4097	    vap->iv_ifp->if_xname, __func__,
4098	    ieee80211_state_name[ostate], ieee80211_state_name[nstate]);
4099
4100	callout_stop(&sc->sc_timer);
4101	/*
4102	 * Clear current radar detection state.
4103	 */
4104	if (ostate == IEEE80211_S_CAC) {
4105		/* stop quiet mode radar detection */
4106		mwl_hal_setradardetection(mh, DR_CHK_CHANNEL_AVAILABLE_STOP);
4107	} else if (sc->sc_radarena) {
4108		/* stop in-service radar detection */
4109		mwl_hal_setradardetection(mh, DR_DFS_DISABLE);
4110		sc->sc_radarena = 0;
4111	}
4112	/*
4113	 * Carry out per-state actions before doing net80211 work.
4114	 */
4115	if (nstate == IEEE80211_S_INIT) {
4116		/* NB: only ap+sta vap's have a fw entity */
4117		if (hvap != NULL)
4118			mwl_hal_stop(hvap);
4119	} else if (nstate == IEEE80211_S_SCAN) {
4120		mwl_hal_start(hvap);
4121		/* NB: this disables beacon frames */
4122		mwl_hal_setinframode(hvap);
4123	} else if (nstate == IEEE80211_S_AUTH) {
4124		/*
4125		 * Must create a sta db entry in case a WEP key needs to
4126		 * be plumbed.  This entry will be overwritten if we
4127		 * associate; otherwise it will be reclaimed on node free.
4128		 */
4129		ni = vap->iv_bss;
4130		MWL_NODE(ni)->mn_hvap = hvap;
4131		(void) mwl_peerstadb(ni, 0, 0, NULL);
4132	} else if (nstate == IEEE80211_S_CSA) {
4133		/* XXX move to below? */
4134		if (vap->iv_opmode == IEEE80211_M_HOSTAP ||
4135		    vap->iv_opmode == IEEE80211_M_MBSS)
4136			mwl_startcsa(vap);
4137	} else if (nstate == IEEE80211_S_CAC) {
4138		/* XXX move to below? */
4139		/* stop ap xmit and enable quiet mode radar detection */
4140		mwl_hal_setradardetection(mh, DR_CHK_CHANNEL_AVAILABLE_START);
4141	}
4142
4143	/*
4144	 * Invoke the parent method to do net80211 work.
4145	 */
4146	error = mvp->mv_newstate(vap, nstate, arg);
4147
4148	/*
4149	 * Carry out work that must be done after net80211 runs;
4150	 * this work requires up to date state (e.g. iv_bss).
4151	 */
4152	if (error == 0 && nstate == IEEE80211_S_RUN) {
4153		/* NB: collect bss node again, it may have changed */
4154		ni = vap->iv_bss;
4155
4156		DPRINTF(sc, MWL_DEBUG_STATE,
4157		    "%s: %s(RUN): iv_flags 0x%08x bintvl %d bssid %s "
4158		    "capinfo 0x%04x chan %d\n",
4159		    vap->iv_ifp->if_xname, __func__, vap->iv_flags,
4160		    ni->ni_intval, ether_sprintf(ni->ni_bssid), ni->ni_capinfo,
4161		    ieee80211_chan2ieee(ic, ic->ic_curchan));
4162
4163		/*
4164		 * Recreate local sta db entry to update WME/HT state.
4165		 */
4166		mwl_localstadb(vap);
4167		switch (vap->iv_opmode) {
4168		case IEEE80211_M_HOSTAP:
4169		case IEEE80211_M_MBSS:
4170			if (ostate == IEEE80211_S_CAC) {
4171				/* enable in-service radar detection */
4172				mwl_hal_setradardetection(mh,
4173				    DR_IN_SERVICE_MONITOR_START);
4174				sc->sc_radarena = 1;
4175			}
4176			/*
4177			 * Allocate and setup the beacon frame
4178			 * (and related state).
4179			 */
4180			error = mwl_reset_vap(vap, IEEE80211_S_RUN);
4181			if (error != 0) {
4182				DPRINTF(sc, MWL_DEBUG_STATE,
4183				    "%s: beacon setup failed, error %d\n",
4184				    __func__, error);
4185				goto bad;
4186			}
4187			/* NB: must be after setting up beacon */
4188			mwl_hal_start(hvap);
4189			break;
4190		case IEEE80211_M_STA:
4191			DPRINTF(sc, MWL_DEBUG_STATE, "%s: %s: aid 0x%x\n",
4192			    vap->iv_ifp->if_xname, __func__, ni->ni_associd);
4193			/*
4194			 * Set state now that we're associated.
4195			 */
4196			mwl_hal_setassocid(hvap, ni->ni_bssid, ni->ni_associd);
4197			mwl_setrates(vap);
4198			mwl_hal_setrtsthreshold(hvap, vap->iv_rtsthreshold);
4199			if ((vap->iv_flags & IEEE80211_F_DWDS) &&
4200			    sc->sc_ndwdsvaps++ == 0)
4201				mwl_hal_setdwds(mh, 1);
4202			break;
4203		case IEEE80211_M_WDS:
4204			DPRINTF(sc, MWL_DEBUG_STATE, "%s: %s: bssid %s\n",
4205			    vap->iv_ifp->if_xname, __func__,
4206			    ether_sprintf(ni->ni_bssid));
4207			mwl_seteapolformat(vap);
4208			break;
4209		default:
4210			break;
4211		}
4212		/*
4213		 * Set CS mode according to operating channel;
4214		 * this mostly an optimization for 5GHz.
4215		 *
4216		 * NB: must follow mwl_hal_start which resets csmode
4217		 */
4218		if (IEEE80211_IS_CHAN_5GHZ(ic->ic_bsschan))
4219			mwl_hal_setcsmode(mh, CSMODE_AGGRESSIVE);
4220		else
4221			mwl_hal_setcsmode(mh, CSMODE_AUTO_ENA);
4222		/*
4223		 * Start timer to prod firmware.
4224		 */
4225		if (sc->sc_ageinterval != 0)
4226			callout_reset(&sc->sc_timer, sc->sc_ageinterval*hz,
4227			    mwl_agestations, sc);
4228	} else if (nstate == IEEE80211_S_SLEEP) {
4229		/* XXX set chip in power save */
4230	} else if ((vap->iv_flags & IEEE80211_F_DWDS) &&
4231	    --sc->sc_ndwdsvaps == 0)
4232		mwl_hal_setdwds(mh, 0);
4233bad:
4234	return error;
4235}
4236
4237/*
4238 * Manage station id's; these are separate from AID's
4239 * as AID's may have values out of the range of possible
4240 * station id's acceptable to the firmware.
4241 */
4242static int
4243allocstaid(struct mwl_softc *sc, int aid)
4244{
4245	int staid;
4246
4247	if (!(0 < aid && aid < MWL_MAXSTAID) || isset(sc->sc_staid, aid)) {
4248		/* NB: don't use 0 */
4249		for (staid = 1; staid < MWL_MAXSTAID; staid++)
4250			if (isclr(sc->sc_staid, staid))
4251				break;
4252	} else
4253		staid = aid;
4254	setbit(sc->sc_staid, staid);
4255	return staid;
4256}
4257
4258static void
4259delstaid(struct mwl_softc *sc, int staid)
4260{
4261	clrbit(sc->sc_staid, staid);
4262}
4263
4264/*
4265 * Setup driver-specific state for a newly associated node.
4266 * Note that we're called also on a re-associate, the isnew
4267 * param tells us if this is the first time or not.
4268 */
4269static void
4270mwl_newassoc(struct ieee80211_node *ni, int isnew)
4271{
4272	struct ieee80211vap *vap = ni->ni_vap;
4273        struct mwl_softc *sc = vap->iv_ic->ic_softc;
4274	struct mwl_node *mn = MWL_NODE(ni);
4275	MWL_HAL_PEERINFO pi;
4276	uint16_t aid;
4277	int error;
4278
4279	aid = IEEE80211_AID(ni->ni_associd);
4280	if (isnew) {
4281		mn->mn_staid = allocstaid(sc, aid);
4282		mn->mn_hvap = MWL_VAP(vap)->mv_hvap;
4283	} else {
4284		mn = MWL_NODE(ni);
4285		/* XXX reset BA stream? */
4286	}
4287	DPRINTF(sc, MWL_DEBUG_NODE, "%s: mac %s isnew %d aid %d staid %d\n",
4288	    __func__, ether_sprintf(ni->ni_macaddr), isnew, aid, mn->mn_staid);
4289	error = mwl_peerstadb(ni, aid, mn->mn_staid, mkpeerinfo(&pi, ni));
4290	if (error != 0) {
4291		DPRINTF(sc, MWL_DEBUG_NODE,
4292		    "%s: error %d creating sta db entry\n",
4293		    __func__, error);
4294		/* XXX how to deal with error? */
4295	}
4296}
4297
4298/*
4299 * Periodically poke the firmware to age out station state
4300 * (power save queues, pending tx aggregates).
4301 */
4302static void
4303mwl_agestations(void *arg)
4304{
4305	struct mwl_softc *sc = arg;
4306
4307	mwl_hal_setkeepalive(sc->sc_mh);
4308	if (sc->sc_ageinterval != 0)		/* NB: catch dynamic changes */
4309		callout_schedule(&sc->sc_timer, sc->sc_ageinterval*hz);
4310}
4311
4312static const struct mwl_hal_channel *
4313findhalchannel(const MWL_HAL_CHANNELINFO *ci, int ieee)
4314{
4315	int i;
4316
4317	for (i = 0; i < ci->nchannels; i++) {
4318		const struct mwl_hal_channel *hc = &ci->channels[i];
4319		if (hc->ieee == ieee)
4320			return hc;
4321	}
4322	return NULL;
4323}
4324
4325static int
4326mwl_setregdomain(struct ieee80211com *ic, struct ieee80211_regdomain *rd,
4327	int nchan, struct ieee80211_channel chans[])
4328{
4329	struct mwl_softc *sc = ic->ic_softc;
4330	struct mwl_hal *mh = sc->sc_mh;
4331	const MWL_HAL_CHANNELINFO *ci;
4332	int i;
4333
4334	for (i = 0; i < nchan; i++) {
4335		struct ieee80211_channel *c = &chans[i];
4336		const struct mwl_hal_channel *hc;
4337
4338		if (IEEE80211_IS_CHAN_2GHZ(c)) {
4339			mwl_hal_getchannelinfo(mh, MWL_FREQ_BAND_2DOT4GHZ,
4340			    IEEE80211_IS_CHAN_HT40(c) ?
4341				MWL_CH_40_MHz_WIDTH : MWL_CH_20_MHz_WIDTH, &ci);
4342		} else if (IEEE80211_IS_CHAN_5GHZ(c)) {
4343			mwl_hal_getchannelinfo(mh, MWL_FREQ_BAND_5GHZ,
4344			    IEEE80211_IS_CHAN_HT40(c) ?
4345				MWL_CH_40_MHz_WIDTH : MWL_CH_20_MHz_WIDTH, &ci);
4346		} else {
4347			device_printf(sc->sc_dev,
4348			    "%s: channel %u freq %u/0x%x not 2.4/5GHz\n",
4349			    __func__, c->ic_ieee, c->ic_freq, c->ic_flags);
4350			return EINVAL;
4351		}
4352		/*
4353		 * Verify channel has cal data and cap tx power.
4354		 */
4355		hc = findhalchannel(ci, c->ic_ieee);
4356		if (hc != NULL) {
4357			if (c->ic_maxpower > 2*hc->maxTxPow)
4358				c->ic_maxpower = 2*hc->maxTxPow;
4359			goto next;
4360		}
4361		if (IEEE80211_IS_CHAN_HT40(c)) {
4362			/*
4363			 * Look for the extension channel since the
4364			 * hal table only has the primary channel.
4365			 */
4366			hc = findhalchannel(ci, c->ic_extieee);
4367			if (hc != NULL) {
4368				if (c->ic_maxpower > 2*hc->maxTxPow)
4369					c->ic_maxpower = 2*hc->maxTxPow;
4370				goto next;
4371			}
4372		}
4373		device_printf(sc->sc_dev,
4374		    "%s: no cal data for channel %u ext %u freq %u/0x%x\n",
4375		    __func__, c->ic_ieee, c->ic_extieee,
4376		    c->ic_freq, c->ic_flags);
4377		return EINVAL;
4378	next:
4379		;
4380	}
4381	return 0;
4382}
4383
4384#define	IEEE80211_CHAN_HTG	(IEEE80211_CHAN_HT|IEEE80211_CHAN_G)
4385#define	IEEE80211_CHAN_HTA	(IEEE80211_CHAN_HT|IEEE80211_CHAN_A)
4386
4387static void
4388addchan(struct ieee80211_channel *c, int freq, int flags, int ieee, int txpow)
4389{
4390	c->ic_freq = freq;
4391	c->ic_flags = flags;
4392	c->ic_ieee = ieee;
4393	c->ic_minpower = 0;
4394	c->ic_maxpower = 2*txpow;
4395	c->ic_maxregpower = txpow;
4396}
4397
4398static const struct ieee80211_channel *
4399findchannel(const struct ieee80211_channel chans[], int nchans,
4400	int freq, int flags)
4401{
4402	const struct ieee80211_channel *c;
4403	int i;
4404
4405	for (i = 0; i < nchans; i++) {
4406		c = &chans[i];
4407		if (c->ic_freq == freq && c->ic_flags == flags)
4408			return c;
4409	}
4410	return NULL;
4411}
4412
4413static void
4414addht40channels(struct ieee80211_channel chans[], int maxchans, int *nchans,
4415	const MWL_HAL_CHANNELINFO *ci, int flags)
4416{
4417	struct ieee80211_channel *c;
4418	const struct ieee80211_channel *extc;
4419	const struct mwl_hal_channel *hc;
4420	int i;
4421
4422	c = &chans[*nchans];
4423
4424	flags &= ~IEEE80211_CHAN_HT;
4425	for (i = 0; i < ci->nchannels; i++) {
4426		/*
4427		 * Each entry defines an HT40 channel pair; find the
4428		 * extension channel above and the insert the pair.
4429		 */
4430		hc = &ci->channels[i];
4431		extc = findchannel(chans, *nchans, hc->freq+20,
4432		    flags | IEEE80211_CHAN_HT20);
4433		if (extc != NULL) {
4434			if (*nchans >= maxchans)
4435				break;
4436			addchan(c, hc->freq, flags | IEEE80211_CHAN_HT40U,
4437			    hc->ieee, hc->maxTxPow);
4438			c->ic_extieee = extc->ic_ieee;
4439			c++, (*nchans)++;
4440			if (*nchans >= maxchans)
4441				break;
4442			addchan(c, extc->ic_freq, flags | IEEE80211_CHAN_HT40D,
4443			    extc->ic_ieee, hc->maxTxPow);
4444			c->ic_extieee = hc->ieee;
4445			c++, (*nchans)++;
4446		}
4447	}
4448}
4449
4450static void
4451addchannels(struct ieee80211_channel chans[], int maxchans, int *nchans,
4452	const MWL_HAL_CHANNELINFO *ci, int flags)
4453{
4454	struct ieee80211_channel *c;
4455	int i;
4456
4457	c = &chans[*nchans];
4458
4459	for (i = 0; i < ci->nchannels; i++) {
4460		const struct mwl_hal_channel *hc;
4461
4462		hc = &ci->channels[i];
4463		if (*nchans >= maxchans)
4464			break;
4465		addchan(c, hc->freq, flags, hc->ieee, hc->maxTxPow);
4466		c++, (*nchans)++;
4467		if (flags == IEEE80211_CHAN_G || flags == IEEE80211_CHAN_HTG) {
4468			/* g channel have a separate b-only entry */
4469			if (*nchans >= maxchans)
4470				break;
4471			c[0] = c[-1];
4472			c[-1].ic_flags = IEEE80211_CHAN_B;
4473			c++, (*nchans)++;
4474		}
4475		if (flags == IEEE80211_CHAN_HTG) {
4476			/* HT g channel have a separate g-only entry */
4477			if (*nchans >= maxchans)
4478				break;
4479			c[-1].ic_flags = IEEE80211_CHAN_G;
4480			c[0] = c[-1];
4481			c[0].ic_flags &= ~IEEE80211_CHAN_HT;
4482			c[0].ic_flags |= IEEE80211_CHAN_HT20;	/* HT20 */
4483			c++, (*nchans)++;
4484		}
4485		if (flags == IEEE80211_CHAN_HTA) {
4486			/* HT a channel have a separate a-only entry */
4487			if (*nchans >= maxchans)
4488				break;
4489			c[-1].ic_flags = IEEE80211_CHAN_A;
4490			c[0] = c[-1];
4491			c[0].ic_flags &= ~IEEE80211_CHAN_HT;
4492			c[0].ic_flags |= IEEE80211_CHAN_HT20;	/* HT20 */
4493			c++, (*nchans)++;
4494		}
4495	}
4496}
4497
4498static void
4499getchannels(struct mwl_softc *sc, int maxchans, int *nchans,
4500	struct ieee80211_channel chans[])
4501{
4502	const MWL_HAL_CHANNELINFO *ci;
4503
4504	/*
4505	 * Use the channel info from the hal to craft the
4506	 * channel list.  Note that we pass back an unsorted
4507	 * list; the caller is required to sort it for us
4508	 * (if desired).
4509	 */
4510	*nchans = 0;
4511	if (mwl_hal_getchannelinfo(sc->sc_mh,
4512	    MWL_FREQ_BAND_2DOT4GHZ, MWL_CH_20_MHz_WIDTH, &ci) == 0)
4513		addchannels(chans, maxchans, nchans, ci, IEEE80211_CHAN_HTG);
4514	if (mwl_hal_getchannelinfo(sc->sc_mh,
4515	    MWL_FREQ_BAND_5GHZ, MWL_CH_20_MHz_WIDTH, &ci) == 0)
4516		addchannels(chans, maxchans, nchans, ci, IEEE80211_CHAN_HTA);
4517	if (mwl_hal_getchannelinfo(sc->sc_mh,
4518	    MWL_FREQ_BAND_2DOT4GHZ, MWL_CH_40_MHz_WIDTH, &ci) == 0)
4519		addht40channels(chans, maxchans, nchans, ci, IEEE80211_CHAN_HTG);
4520	if (mwl_hal_getchannelinfo(sc->sc_mh,
4521	    MWL_FREQ_BAND_5GHZ, MWL_CH_40_MHz_WIDTH, &ci) == 0)
4522		addht40channels(chans, maxchans, nchans, ci, IEEE80211_CHAN_HTA);
4523}
4524
4525static void
4526mwl_getradiocaps(struct ieee80211com *ic,
4527	int maxchans, int *nchans, struct ieee80211_channel chans[])
4528{
4529	struct mwl_softc *sc = ic->ic_softc;
4530
4531	getchannels(sc, maxchans, nchans, chans);
4532}
4533
4534static int
4535mwl_getchannels(struct mwl_softc *sc)
4536{
4537	struct ieee80211com *ic = &sc->sc_ic;
4538
4539	/*
4540	 * Use the channel info from the hal to craft the
4541	 * channel list for net80211.  Note that we pass up
4542	 * an unsorted list; net80211 will sort it for us.
4543	 */
4544	memset(ic->ic_channels, 0, sizeof(ic->ic_channels));
4545	ic->ic_nchans = 0;
4546	getchannels(sc, IEEE80211_CHAN_MAX, &ic->ic_nchans, ic->ic_channels);
4547
4548	ic->ic_regdomain.regdomain = SKU_DEBUG;
4549	ic->ic_regdomain.country = CTRY_DEFAULT;
4550	ic->ic_regdomain.location = 'I';
4551	ic->ic_regdomain.isocc[0] = ' ';	/* XXX? */
4552	ic->ic_regdomain.isocc[1] = ' ';
4553	return (ic->ic_nchans == 0 ? EIO : 0);
4554}
4555#undef IEEE80211_CHAN_HTA
4556#undef IEEE80211_CHAN_HTG
4557
4558#ifdef MWL_DEBUG
4559static void
4560mwl_printrxbuf(const struct mwl_rxbuf *bf, u_int ix)
4561{
4562	const struct mwl_rxdesc *ds = bf->bf_desc;
4563	uint32_t status = le32toh(ds->Status);
4564
4565	printf("R[%2u] (DS.V:%p DS.P:0x%jx) NEXT:%08x DATA:%08x RC:%02x%s\n"
4566	       "      STAT:%02x LEN:%04x RSSI:%02x CHAN:%02x RATE:%02x QOS:%04x HT:%04x\n",
4567	    ix, ds, (uintmax_t)bf->bf_daddr, le32toh(ds->pPhysNext),
4568	    le32toh(ds->pPhysBuffData), ds->RxControl,
4569	    ds->RxControl != EAGLE_RXD_CTRL_DRIVER_OWN ?
4570	        "" : (status & EAGLE_RXD_STATUS_OK) ? " *" : " !",
4571	    ds->Status, le16toh(ds->PktLen), ds->RSSI, ds->Channel,
4572	    ds->Rate, le16toh(ds->QosCtrl), le16toh(ds->HtSig2));
4573}
4574
4575static void
4576mwl_printtxbuf(const struct mwl_txbuf *bf, u_int qnum, u_int ix)
4577{
4578	const struct mwl_txdesc *ds = bf->bf_desc;
4579	uint32_t status = le32toh(ds->Status);
4580
4581	printf("Q%u[%3u]", qnum, ix);
4582	printf(" (DS.V:%p DS.P:0x%jx)\n", ds, (uintmax_t)bf->bf_daddr);
4583	printf("    NEXT:%08x DATA:%08x LEN:%04x STAT:%08x%s\n",
4584	    le32toh(ds->pPhysNext),
4585	    le32toh(ds->PktPtr), le16toh(ds->PktLen), status,
4586	    status & EAGLE_TXD_STATUS_USED ?
4587		"" : (status & 3) != 0 ? " *" : " !");
4588	printf("    RATE:%02x PRI:%x QOS:%04x SAP:%08x FORMAT:%04x\n",
4589	    ds->DataRate, ds->TxPriority, le16toh(ds->QosCtrl),
4590	    le32toh(ds->SapPktInfo), le16toh(ds->Format));
4591#if MWL_TXDESC > 1
4592	printf("    MULTIFRAMES:%u LEN:%04x %04x %04x %04x %04x %04x\n"
4593	    , le32toh(ds->multiframes)
4594	    , le16toh(ds->PktLenArray[0]), le16toh(ds->PktLenArray[1])
4595	    , le16toh(ds->PktLenArray[2]), le16toh(ds->PktLenArray[3])
4596	    , le16toh(ds->PktLenArray[4]), le16toh(ds->PktLenArray[5])
4597	);
4598	printf("    DATA:%08x %08x %08x %08x %08x %08x\n"
4599	    , le32toh(ds->PktPtrArray[0]), le32toh(ds->PktPtrArray[1])
4600	    , le32toh(ds->PktPtrArray[2]), le32toh(ds->PktPtrArray[3])
4601	    , le32toh(ds->PktPtrArray[4]), le32toh(ds->PktPtrArray[5])
4602	);
4603#endif
4604#if 0
4605{ const uint8_t *cp = (const uint8_t *) ds;
4606  int i;
4607  for (i = 0; i < sizeof(struct mwl_txdesc); i++) {
4608	printf("%02x ", cp[i]);
4609	if (((i+1) % 16) == 0)
4610		printf("\n");
4611  }
4612  printf("\n");
4613}
4614#endif
4615}
4616#endif /* MWL_DEBUG */
4617
4618#if 0
4619static void
4620mwl_txq_dump(struct mwl_txq *txq)
4621{
4622	struct mwl_txbuf *bf;
4623	int i = 0;
4624
4625	MWL_TXQ_LOCK(txq);
4626	STAILQ_FOREACH(bf, &txq->active, bf_list) {
4627		struct mwl_txdesc *ds = bf->bf_desc;
4628		MWL_TXDESC_SYNC(txq, ds,
4629		    BUS_DMASYNC_POSTREAD | BUS_DMASYNC_POSTWRITE);
4630#ifdef MWL_DEBUG
4631		mwl_printtxbuf(bf, txq->qnum, i);
4632#endif
4633		i++;
4634	}
4635	MWL_TXQ_UNLOCK(txq);
4636}
4637#endif
4638
4639static void
4640mwl_watchdog(void *arg)
4641{
4642	struct mwl_softc *sc = arg;
4643
4644	callout_reset(&sc->sc_watchdog, hz, mwl_watchdog, sc);
4645	if (sc->sc_tx_timer == 0 || --sc->sc_tx_timer > 0)
4646		return;
4647
4648	if (sc->sc_running && !sc->sc_invalid) {
4649		if (mwl_hal_setkeepalive(sc->sc_mh))
4650			device_printf(sc->sc_dev,
4651			    "transmit timeout (firmware hung?)\n");
4652		else
4653			device_printf(sc->sc_dev,
4654			    "transmit timeout\n");
4655#if 0
4656		mwl_reset(sc);
4657mwl_txq_dump(&sc->sc_txq[0]);/*XXX*/
4658#endif
4659		counter_u64_add(sc->sc_ic.ic_oerrors, 1);
4660		sc->sc_stats.mst_watchdog++;
4661	}
4662}
4663
4664#ifdef MWL_DIAGAPI
4665/*
4666 * Diagnostic interface to the HAL.  This is used by various
4667 * tools to do things like retrieve register contents for
4668 * debugging.  The mechanism is intentionally opaque so that
4669 * it can change frequently w/o concern for compatiblity.
4670 */
4671static int
4672mwl_ioctl_diag(struct mwl_softc *sc, struct mwl_diag *md)
4673{
4674	struct mwl_hal *mh = sc->sc_mh;
4675	u_int id = md->md_id & MWL_DIAG_ID;
4676	void *indata = NULL;
4677	void *outdata = NULL;
4678	u_int32_t insize = md->md_in_size;
4679	u_int32_t outsize = md->md_out_size;
4680	int error = 0;
4681
4682	if (md->md_id & MWL_DIAG_IN) {
4683		/*
4684		 * Copy in data.
4685		 */
4686		indata = malloc(insize, M_TEMP, M_NOWAIT);
4687		if (indata == NULL) {
4688			error = ENOMEM;
4689			goto bad;
4690		}
4691		error = copyin(md->md_in_data, indata, insize);
4692		if (error)
4693			goto bad;
4694	}
4695	if (md->md_id & MWL_DIAG_DYN) {
4696		/*
4697		 * Allocate a buffer for the results (otherwise the HAL
4698		 * returns a pointer to a buffer where we can read the
4699		 * results).  Note that we depend on the HAL leaving this
4700		 * pointer for us to use below in reclaiming the buffer;
4701		 * may want to be more defensive.
4702		 */
4703		outdata = malloc(outsize, M_TEMP, M_NOWAIT);
4704		if (outdata == NULL) {
4705			error = ENOMEM;
4706			goto bad;
4707		}
4708	}
4709	if (mwl_hal_getdiagstate(mh, id, indata, insize, &outdata, &outsize)) {
4710		if (outsize < md->md_out_size)
4711			md->md_out_size = outsize;
4712		if (outdata != NULL)
4713			error = copyout(outdata, md->md_out_data,
4714					md->md_out_size);
4715	} else {
4716		error = EINVAL;
4717	}
4718bad:
4719	if ((md->md_id & MWL_DIAG_IN) && indata != NULL)
4720		free(indata, M_TEMP);
4721	if ((md->md_id & MWL_DIAG_DYN) && outdata != NULL)
4722		free(outdata, M_TEMP);
4723	return error;
4724}
4725
4726static int
4727mwl_ioctl_reset(struct mwl_softc *sc, struct mwl_diag *md)
4728{
4729	struct mwl_hal *mh = sc->sc_mh;
4730	int error;
4731
4732	MWL_LOCK_ASSERT(sc);
4733
4734	if (md->md_id == 0 && mwl_hal_fwload(mh, NULL) != 0) {
4735		device_printf(sc->sc_dev, "unable to load firmware\n");
4736		return EIO;
4737	}
4738	if (mwl_hal_gethwspecs(mh, &sc->sc_hwspecs) != 0) {
4739		device_printf(sc->sc_dev, "unable to fetch h/w specs\n");
4740		return EIO;
4741	}
4742	error = mwl_setupdma(sc);
4743	if (error != 0) {
4744		/* NB: mwl_setupdma prints a msg */
4745		return error;
4746	}
4747	/*
4748	 * Reset tx/rx data structures; after reload we must
4749	 * re-start the driver's notion of the next xmit/recv.
4750	 */
4751	mwl_draintxq(sc);		/* clear pending frames */
4752	mwl_resettxq(sc);		/* rebuild tx q lists */
4753	sc->sc_rxnext = NULL;		/* force rx to start at the list head */
4754	return 0;
4755}
4756#endif /* MWL_DIAGAPI */
4757
4758static void
4759mwl_parent(struct ieee80211com *ic)
4760{
4761	struct mwl_softc *sc = ic->ic_softc;
4762	int startall = 0;
4763
4764	MWL_LOCK(sc);
4765	if (ic->ic_nrunning > 0) {
4766		if (sc->sc_running) {
4767			/*
4768			 * To avoid rescanning another access point,
4769			 * do not call mwl_init() here.  Instead,
4770			 * only reflect promisc mode settings.
4771			 */
4772			mwl_mode_init(sc);
4773		} else {
4774			/*
4775			 * Beware of being called during attach/detach
4776			 * to reset promiscuous mode.  In that case we
4777			 * will still be marked UP but not RUNNING.
4778			 * However trying to re-init the interface
4779			 * is the wrong thing to do as we've already
4780			 * torn down much of our state.  There's
4781			 * probably a better way to deal with this.
4782			 */
4783			if (!sc->sc_invalid) {
4784				mwl_init(sc);	/* XXX lose error */
4785				startall = 1;
4786			}
4787		}
4788	} else
4789		mwl_stop(sc);
4790	MWL_UNLOCK(sc);
4791	if (startall)
4792		ieee80211_start_all(ic);
4793}
4794
4795static int
4796mwl_ioctl(struct ieee80211com *ic, u_long cmd, void *data)
4797{
4798	struct mwl_softc *sc = ic->ic_softc;
4799	struct ifreq *ifr = data;
4800	int error = 0;
4801
4802	switch (cmd) {
4803	case SIOCGMVSTATS:
4804		mwl_hal_gethwstats(sc->sc_mh, &sc->sc_stats.hw_stats);
4805#if 0
4806		/* NB: embed these numbers to get a consistent view */
4807		sc->sc_stats.mst_tx_packets =
4808		    ifp->if_get_counter(ifp, IFCOUNTER_OPACKETS);
4809		sc->sc_stats.mst_rx_packets =
4810		    ifp->if_get_counter(ifp, IFCOUNTER_IPACKETS);
4811#endif
4812		/*
4813		 * NB: Drop the softc lock in case of a page fault;
4814		 * we'll accept any potential inconsisentcy in the
4815		 * statistics.  The alternative is to copy the data
4816		 * to a local structure.
4817		 */
4818		return (copyout(&sc->sc_stats,
4819				ifr->ifr_data, sizeof (sc->sc_stats)));
4820#ifdef MWL_DIAGAPI
4821	case SIOCGMVDIAG:
4822		/* XXX check privs */
4823		return mwl_ioctl_diag(sc, (struct mwl_diag *) ifr);
4824	case SIOCGMVRESET:
4825		/* XXX check privs */
4826		MWL_LOCK(sc);
4827		error = mwl_ioctl_reset(sc,(struct mwl_diag *) ifr);
4828		MWL_UNLOCK(sc);
4829		break;
4830#endif /* MWL_DIAGAPI */
4831	default:
4832		error = ENOTTY;
4833		break;
4834	}
4835	return (error);
4836}
4837
4838#ifdef	MWL_DEBUG
4839static int
4840mwl_sysctl_debug(SYSCTL_HANDLER_ARGS)
4841{
4842	struct mwl_softc *sc = arg1;
4843	int debug, error;
4844
4845	debug = sc->sc_debug | (mwl_hal_getdebug(sc->sc_mh) << 24);
4846	error = sysctl_handle_int(oidp, &debug, 0, req);
4847	if (error || !req->newptr)
4848		return error;
4849	mwl_hal_setdebug(sc->sc_mh, debug >> 24);
4850	sc->sc_debug = debug & 0x00ffffff;
4851	return 0;
4852}
4853#endif /* MWL_DEBUG */
4854
4855static void
4856mwl_sysctlattach(struct mwl_softc *sc)
4857{
4858#ifdef	MWL_DEBUG
4859	struct sysctl_ctx_list *ctx = device_get_sysctl_ctx(sc->sc_dev);
4860	struct sysctl_oid *tree = device_get_sysctl_tree(sc->sc_dev);
4861
4862	sc->sc_debug = mwl_debug;
4863	SYSCTL_ADD_PROC(ctx, SYSCTL_CHILDREN(tree), OID_AUTO,
4864		"debug", CTLTYPE_INT | CTLFLAG_RW, sc, 0,
4865		mwl_sysctl_debug, "I", "control debugging printfs");
4866#endif
4867}
4868
4869/*
4870 * Announce various information on device/driver attach.
4871 */
4872static void
4873mwl_announce(struct mwl_softc *sc)
4874{
4875
4876	device_printf(sc->sc_dev, "Rev A%d hardware, v%d.%d.%d.%d firmware (regioncode %d)\n",
4877		sc->sc_hwspecs.hwVersion,
4878		(sc->sc_hwspecs.fwReleaseNumber>>24) & 0xff,
4879		(sc->sc_hwspecs.fwReleaseNumber>>16) & 0xff,
4880		(sc->sc_hwspecs.fwReleaseNumber>>8) & 0xff,
4881		(sc->sc_hwspecs.fwReleaseNumber>>0) & 0xff,
4882		sc->sc_hwspecs.regionCode);
4883	sc->sc_fwrelease = sc->sc_hwspecs.fwReleaseNumber;
4884
4885	if (bootverbose) {
4886		int i;
4887		for (i = 0; i <= WME_AC_VO; i++) {
4888			struct mwl_txq *txq = sc->sc_ac2q[i];
4889			device_printf(sc->sc_dev, "Use hw queue %u for %s traffic\n",
4890				txq->qnum, ieee80211_wme_acnames[i]);
4891		}
4892	}
4893	if (bootverbose || mwl_rxdesc != MWL_RXDESC)
4894		device_printf(sc->sc_dev, "using %u rx descriptors\n", mwl_rxdesc);
4895	if (bootverbose || mwl_rxbuf != MWL_RXBUF)
4896		device_printf(sc->sc_dev, "using %u rx buffers\n", mwl_rxbuf);
4897	if (bootverbose || mwl_txbuf != MWL_TXBUF)
4898		device_printf(sc->sc_dev, "using %u tx buffers\n", mwl_txbuf);
4899	if (bootverbose && mwl_hal_ismbsscapable(sc->sc_mh))
4900		device_printf(sc->sc_dev, "multi-bss support\n");
4901#ifdef MWL_TX_NODROP
4902	if (bootverbose)
4903		device_printf(sc->sc_dev, "no tx drop\n");
4904#endif
4905}
4906