1/* $OpenBSD: if_rtwn.c,v 1.6 2015/08/28 00:03:53 deraadt Exp $ */ 2 3/*- 4 * Copyright (c) 2010 Damien Bergamini <damien.bergamini@free.fr> 5 * Copyright (c) 2015 Stefan Sperling <stsp@openbsd.org> 6 * Copyright (c) 2016 Andriy Voskoboinyk <avos@FreeBSD.org> 7 * 8 * Permission to use, copy, modify, and distribute this software for any 9 * purpose with or without fee is hereby granted, provided that the above 10 * copyright notice and this permission notice appear in all copies. 11 * 12 * THE SOFTWARE IS PROVIDED "AS IS" AND THE AUTHOR DISCLAIMS ALL WARRANTIES 13 * WITH REGARD TO THIS SOFTWARE INCLUDING ALL IMPLIED WARRANTIES OF 14 * MERCHANTABILITY AND FITNESS. IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR 15 * ANY SPECIAL, DIRECT, INDIRECT, OR CONSEQUENTIAL DAMAGES OR ANY DAMAGES 16 * WHATSOEVER RESULTING FROM LOSS OF USE, DATA OR PROFITS, WHETHER IN AN 17 * ACTION OF CONTRACT, NEGLIGENCE OR OTHER TORTIOUS ACTION, ARISING OUT OF 18 * OR IN CONNECTION WITH THE USE OR PERFORMANCE OF THIS SOFTWARE. 19 */ 20 21#include <sys/cdefs.h> 22__FBSDID("$FreeBSD$"); 23 24#include "opt_wlan.h" 25 26#include <sys/param.h> 27#include <sys/lock.h> 28#include <sys/mutex.h> 29#include <sys/mbuf.h> 30#include <sys/kernel.h> 31#include <sys/socket.h> 32#include <sys/systm.h> 33#include <sys/malloc.h> 34#include <sys/queue.h> 35#include <sys/taskqueue.h> 36#include <sys/bus.h> 37#include <sys/endian.h> 38 39#include <machine/bus.h> 40#include <machine/resource.h> 41#include <sys/rman.h> 42 43#include <net/if.h> 44#include <net/ethernet.h> 45#include <net/if_media.h> 46 47#include <net80211/ieee80211_var.h> 48 49#include <dev/rtwn/if_rtwnreg.h> 50#include <dev/rtwn/if_rtwnvar.h> 51#include <dev/rtwn/if_rtwn_debug.h> 52#include <dev/rtwn/if_rtwn_rx.h> 53#include <dev/rtwn/if_rtwn_task.h> 54#include <dev/rtwn/if_rtwn_tx.h> 55 56#include <dev/rtwn/pci/rtwn_pci_var.h> 57#include <dev/rtwn/pci/rtwn_pci_rx.h> 58 59void 60rtwn_pci_dma_map_addr(void *arg, bus_dma_segment_t *segs, int nsegs, 61 int error) 62{ 63 64 if (error != 0) 65 return; 66 KASSERT(nsegs == 1, ("too many DMA segments, %d should be 1", nsegs)); 67 *(bus_addr_t *)arg = segs[0].ds_addr; 68} 69 70void 71rtwn_pci_setup_rx_desc(struct rtwn_pci_softc *pc, 72 struct rtwn_rx_stat_pci *desc, bus_addr_t addr, size_t len, int idx) 73{ 74 75 memset(desc, 0, sizeof(*desc)); 76 desc->rxdw0 = htole32(SM(RTWN_RXDW0_PKTLEN, len) | 77 ((idx == RTWN_PCI_RX_LIST_COUNT - 1) ? RTWN_RXDW0_EOR : 0)); 78 desc->rxbufaddr = htole32(addr); 79 bus_space_barrier(pc->pc_st, pc->pc_sh, 0, pc->pc_mapsize, 80 BUS_SPACE_BARRIER_WRITE); 81 desc->rxdw0 |= htole32(RTWN_RXDW0_OWN); 82} 83 84static void 85rtwn_pci_rx_frame(struct rtwn_pci_softc *pc) 86{ 87 struct rtwn_softc *sc = &pc->pc_sc; 88 struct rtwn_rx_ring *ring = &pc->rx_ring; 89 struct rtwn_rx_stat_pci *rx_desc = &ring->desc[ring->cur]; 90 struct rtwn_rx_data *rx_data = &ring->rx_data[ring->cur]; 91 struct ieee80211com *ic = &sc->sc_ic; 92 struct ieee80211_node *ni; 93 uint32_t rxdw0; 94 struct mbuf *m, *m1; 95 int infosz, pktlen, shift, error; 96 97 /* Dump Rx descriptor. */ 98 RTWN_DPRINTF(sc, RTWN_DEBUG_RECV_DESC, 99 "%s: dw: 0 %08X, 1 %08X, 2 %08X, 3 %08X, 4 %08X, tsfl %08X, " 100 "addr: %08X (64: %08X)\n", 101 __func__, le32toh(rx_desc->rxdw0), le32toh(rx_desc->rxdw1), 102 le32toh(rx_desc->rxdw2), le32toh(rx_desc->rxdw3), 103 le32toh(rx_desc->rxdw4), le32toh(rx_desc->tsf_low), 104 le32toh(rx_desc->rxbufaddr), le32toh(rx_desc->rxbufaddr64)); 105 106 rxdw0 = le32toh(rx_desc->rxdw0); 107 if (__predict_false(rxdw0 & (RTWN_RXDW0_CRCERR | RTWN_RXDW0_ICVERR))) { 108 /* 109 * This should not happen since we setup our Rx filter 110 * to not receive these frames. 111 */ 112 RTWN_DPRINTF(sc, RTWN_DEBUG_RECV, 113 "%s: RX flags error (%s)\n", __func__, 114 rxdw0 & RTWN_RXDW0_CRCERR ? "CRC" : "ICV"); 115 goto fail; 116 } 117 118 pktlen = MS(rxdw0, RTWN_RXDW0_PKTLEN); 119 if (__predict_false(pktlen < sizeof(struct ieee80211_frame_ack) || 120 pktlen > MJUMPAGESIZE)) { 121 RTWN_DPRINTF(sc, RTWN_DEBUG_RECV, 122 "%s: frame is too short/long: %d\n", __func__, pktlen); 123 goto fail; 124 } 125 126 infosz = MS(rxdw0, RTWN_RXDW0_INFOSZ) * 8; 127 shift = MS(rxdw0, RTWN_RXDW0_SHIFT); 128 129 m1 = m_getjcl(M_NOWAIT, MT_DATA, M_PKTHDR, MJUMPAGESIZE); 130 if (__predict_false(m1 == NULL)) { 131 device_printf(sc->sc_dev, "%s: could not allocate RX mbuf\n", 132 __func__); 133 goto fail; 134 } 135 bus_dmamap_sync(ring->data_dmat, rx_data->map, BUS_DMASYNC_POSTREAD); 136 bus_dmamap_unload(ring->data_dmat, rx_data->map); 137 138 error = bus_dmamap_load(ring->data_dmat, rx_data->map, mtod(m1, void *), 139 MJUMPAGESIZE, rtwn_pci_dma_map_addr, &rx_data->paddr, 0); 140 if (error != 0) { 141 m_freem(m1); 142 143 error = bus_dmamap_load(ring->data_dmat, rx_data->map, 144 mtod(rx_data->m, void *), MJUMPAGESIZE, 145 rtwn_pci_dma_map_addr, &rx_data->paddr, BUS_DMA_NOWAIT); 146 if (error != 0) 147 panic("%s: could not load old RX mbuf", 148 device_get_name(sc->sc_dev)); 149 150 goto fail; 151 } 152 153 /* Finalize mbuf. */ 154 m = rx_data->m; 155 rx_data->m = m1; 156 m->m_pkthdr.len = m->m_len = pktlen + infosz + shift; 157 158 ni = rtwn_rx_common(sc, m, rx_desc); 159 160 RTWN_DPRINTF(sc, RTWN_DEBUG_RECV, 161 "%s: Rx frame len %d, infosz %d, shift %d\n", 162 __func__, pktlen, infosz, shift); 163 164 /* Send the frame to the 802.11 layer. */ 165 RTWN_UNLOCK(sc); 166 if (ni != NULL) { 167 (void)ieee80211_input_mimo(ni, m); 168 /* Node is no longer needed. */ 169 ieee80211_free_node(ni); 170 } else 171 (void)ieee80211_input_mimo_all(ic, m); 172 173 RTWN_LOCK(sc); 174 175 return; 176 177fail: 178 counter_u64_add(ic->ic_ierrors, 1); 179} 180 181static int 182rtwn_pci_rx_buf_copy(struct rtwn_pci_softc *pc) 183{ 184 struct rtwn_rx_ring *ring = &pc->rx_ring; 185 struct rtwn_rx_stat_pci *rx_desc = &ring->desc[ring->cur]; 186 struct rtwn_rx_data *rx_data = &ring->rx_data[ring->cur]; 187 uint32_t rxdw0; 188 int desc_size, pktlen; 189 190 /* 191 * NB: tx_report() / c2h_report() expects to see USB Rx 192 * descriptor - same as for PCIe, but without rxbufaddr* fields. 193 */ 194 desc_size = sizeof(struct rtwn_rx_stat_common); 195 KASSERT(sizeof(pc->pc_rx_buf) >= desc_size, 196 ("adjust size for PCIe Rx buffer!")); 197 198 memcpy(pc->pc_rx_buf, rx_desc, desc_size); 199 200 rxdw0 = le32toh(rx_desc->rxdw0); 201 pktlen = MS(rxdw0, RTWN_RXDW0_PKTLEN); 202 203 if (pktlen > sizeof(pc->pc_rx_buf) - desc_size) 204 { 205 /* Looks like an ordinary Rx frame. */ 206 return (desc_size); 207 } 208 209 bus_dmamap_sync(ring->data_dmat, rx_data->map, BUS_DMASYNC_POSTREAD); 210 memcpy(pc->pc_rx_buf + desc_size, mtod(rx_data->m, void *), pktlen); 211 212 return (desc_size + pktlen); 213} 214 215static void 216rtwn_pci_tx_report(struct rtwn_pci_softc *pc, int len) 217{ 218 struct rtwn_softc *sc = &pc->pc_sc; 219 220 if (sc->sc_ratectl != RTWN_RATECTL_NET80211) { 221 /* shouldn't happen */ 222 device_printf(sc->sc_dev, 223 "%s called while ratectl = %d!\n", 224 __func__, sc->sc_ratectl); 225 return; 226 } 227 228 RTWN_NT_LOCK(sc); 229 rtwn_handle_tx_report(sc, pc->pc_rx_buf, len); 230 RTWN_NT_UNLOCK(sc); 231 232#ifdef IEEE80211_SUPPORT_SUPERG 233 /* 234 * NB: this will executed only when 'report' bit is set. 235 */ 236 if (sc->sc_tx_n_active > 0 && --sc->sc_tx_n_active <= 1) 237 rtwn_cmd_sleepable(sc, NULL, 0, rtwn_ff_flush_all); 238#endif 239} 240 241static void 242rtwn_pci_c2h_report(struct rtwn_pci_softc *pc, int len) 243{ 244 rtwn_handle_c2h_report(&pc->pc_sc, pc->pc_rx_buf, len); 245} 246 247static void 248rtwn_pci_tx_done(struct rtwn_softc *sc, int qid) 249{ 250 struct rtwn_pci_softc *pc = RTWN_PCI_SOFTC(sc); 251 struct rtwn_tx_ring *ring = &pc->tx_ring[qid]; 252 struct rtwn_tx_desc_common *desc; 253 struct rtwn_tx_data *data; 254 255 RTWN_DPRINTF(sc, RTWN_DEBUG_INTR, "%s: qid %d, last %d, cur %d\n", 256 __func__, qid, ring->last, ring->cur); 257 258 bus_dmamap_sync(ring->desc_dmat, ring->desc_map, 259 BUS_DMASYNC_POSTREAD | BUS_DMASYNC_POSTWRITE); 260 261 while(ring->last != ring->cur) { 262 data = &ring->tx_data[ring->last]; 263 desc = (struct rtwn_tx_desc_common *) 264 ((uint8_t *)ring->desc + sc->txdesc_len * ring->last); 265 266 KASSERT(data->m != NULL, ("no mbuf")); 267 268 if (desc->flags0 & RTWN_FLAGS0_OWN) 269 break; 270 271 /* Unmap and free mbuf. */ 272 bus_dmamap_sync(ring->data_dmat, data->map, 273 BUS_DMASYNC_POSTWRITE); 274 bus_dmamap_unload(ring->data_dmat, data->map); 275 276 if (data->ni != NULL) { /* not a beacon frame */ 277 ieee80211_tx_complete(data->ni, data->m, 0); 278 279 data->ni = NULL; 280 ring->queued--; 281 KASSERT(ring->queued >= 0, 282 ("ring->queued (qid %d) underflow!\n", qid)); 283 } else 284 m_freem(data->m); 285 286 data->m = NULL; 287 ring->last = (ring->last + 1) % RTWN_PCI_TX_LIST_COUNT; 288#ifndef D4054 289 if (ring->queued > 0) 290 sc->sc_tx_timer = 5; 291 else 292 sc->sc_tx_timer = 0; 293#endif 294 } 295 296 if ((sc->qfullmsk & (1 << qid)) != 0 && 297 ring->queued < (RTWN_PCI_TX_LIST_COUNT - 1)) { 298 sc->qfullmsk &= ~(1 << qid); 299 rtwn_start(sc); 300 } 301 302#ifdef IEEE80211_SUPPORT_SUPERG 303 /* 304 * If the TX active queue drops below a certain 305 * threshold, ensure we age fast-frames out so they're 306 * transmitted. 307 */ 308 if (sc->sc_ratectl != RTWN_RATECTL_NET80211 && ring->queued <= 1) { 309 /* 310 * XXX TODO: just make this a callout timer schedule 311 * so we can flush the FF staging queue if we're 312 * approaching idle. 313 */ 314 rtwn_cmd_sleepable(sc, NULL, 0, rtwn_ff_flush_all); 315 } 316#endif 317} 318 319static void 320rtwn_pci_rx_done(struct rtwn_softc *sc) 321{ 322 struct rtwn_pci_softc *pc = RTWN_PCI_SOFTC(sc); 323 struct rtwn_rx_ring *ring = &pc->rx_ring; 324 struct rtwn_rx_stat_pci *rx_desc; 325 struct rtwn_rx_data *rx_data; 326 int len; 327 328 bus_dmamap_sync(ring->desc_dmat, ring->desc_map, BUS_DMASYNC_POSTREAD); 329 330 for (;;) { 331 rx_desc = &ring->desc[ring->cur]; 332 rx_data = &ring->rx_data[ring->cur]; 333 334 if (le32toh(rx_desc->rxdw0) & RTWN_RXDW0_OWN) 335 break; 336 337 len = rtwn_pci_rx_buf_copy(pc); 338 339 switch (rtwn_classify_intr(sc, pc->pc_rx_buf, len)) { 340 case RTWN_RX_DATA: 341 rtwn_pci_rx_frame(pc); 342 break; 343 case RTWN_RX_TX_REPORT: 344 rtwn_pci_tx_report(pc, len); 345 break; 346 case RTWN_RX_OTHER: 347 rtwn_pci_c2h_report(pc, len); 348 break; 349 default: 350 /* NOTREACHED */ 351 KASSERT(0, ("unknown Rx classification code")); 352 break; 353 } 354 355 /* Update / reset RX descriptor (and set OWN bit). */ 356 rtwn_pci_setup_rx_desc(pc, rx_desc, rx_data->paddr, 357 MJUMPAGESIZE, ring->cur); 358 359 if (!(sc->sc_flags & RTWN_RUNNING)) 360 return; 361 362 /* NB: device can reuse current descriptor. */ 363 bus_dmamap_sync(ring->desc_dmat, ring->desc_map, 364 BUS_DMASYNC_POSTREAD); 365 366 if (le32toh(rx_desc->rxdw0) & RTWN_RXDW0_OWN) 367 ring->cur = (ring->cur + 1) % RTWN_PCI_RX_LIST_COUNT; 368 } 369} 370 371void 372rtwn_pci_intr(void *arg) 373{ 374 struct rtwn_softc *sc = arg; 375 struct rtwn_pci_softc *pc = RTWN_PCI_SOFTC(sc); 376 int i, status, tx_rings; 377 378 RTWN_LOCK(sc); 379 status = rtwn_pci_get_intr_status(pc, &tx_rings); 380 RTWN_DPRINTF(sc, RTWN_DEBUG_INTR, "%s: status %08X, tx_rings %08X\n", 381 __func__, status, tx_rings); 382 if (status == 0 && tx_rings == 0) 383 goto unlock; 384 385 if (status & (RTWN_PCI_INTR_RX | RTWN_PCI_INTR_TX_REPORT)) { 386 rtwn_pci_rx_done(sc); 387 if (!(sc->sc_flags & RTWN_RUNNING)) 388 goto unlock; 389 } 390 391 if (tx_rings != 0) 392 for (i = 0; i < RTWN_PCI_NTXQUEUES; i++) 393 if (tx_rings & (1 << i)) 394 rtwn_pci_tx_done(sc, i); 395 396 if (sc->sc_flags & RTWN_RUNNING) 397 rtwn_pci_enable_intr(pc); 398unlock: 399 RTWN_UNLOCK(sc); 400} 401