Lines Matching refs:tid

126     int tid);
128 int tid);
132 struct ieee80211_node *ni, struct mbuf *m0, int *tid);
135 struct ath_tid *tid, struct ath_buf *bf);
1379 ath_tx_do_ratelookup(struct ath_softc *sc, struct ath_buf *bf, int tid,
1395 pktlen, tid, is_aggr, &rix, &try0, &rate, &maxdur, &maxpktlen);
1418 ath_tx_update_clrdmask(struct ath_softc *sc, struct ath_tid *tid,
1509 struct ath_tid *tid = &an->an_tid[bf->bf_state.bfs_tid];
1525 ath_tx_do_ratelookup(sc, bf, tid->tid, bf->bf_state.bfs_pktlen, false);
1533 tid->hwq_depth++;
1911 int tid;
1948 tid = ath_tx_gettid(sc, m0);
1997 is_ampdu_tx = ath_tx_ampdu_running(sc, ATH_NODE(ni), tid);
1998 is_ampdu_pending = ath_tx_ampdu_pending(sc, ATH_NODE(ni), tid);
2001 DPRINTF(sc, ATH_DEBUG_SW_TX, "%s: tid=%d, ac=%d, is_ampdu=%d\n",
2002 __func__, tid, pri, is_ampdu);
2005 bf->bf_state.bfs_tid = tid;
2070 "%s: tid %d: ampdu pending, seqno %d\n",
2071 __func__, tid, M_SEQNO_GET(m0));
2131 ath_tx_leak_count_update(sc, tid, bf);
2193 "%s: overriding tid %d pri %d -> %d\n",
2408 ath_tx_leak_count_update(sc, tid, bf);
2609 struct mbuf *m0, int *tid)
2639 *tid = (int) _IEEE80211_MASKSHIFT(baparamset, IEEE80211_BAPS_TID);
2658 struct ath_tid *tid, struct ath_buf *bf)
2668 tap = ath_tx_get_tx_tid(an, tid->tid);
2679 "%s: re-added? tid=%d, seqno %d; window %d:%d; "
2681 __func__, tid->tid, SEQNO(bf->bf_state.bfs_seqno),
2682 tap->txa_start, tap->txa_wnd, tid->baw_head,
2683 tid->baw_tail);
2692 "%s: bf=%p: outside of BAW?? tid=%d, seqno %d; window %d:%d; "
2694 __func__, bf, tid->tid, SEQNO(bf->bf_state.bfs_seqno),
2695 tap->txa_start, tap->txa_wnd, tid->baw_head,
2696 tid->baw_tail);
2704 cindex = (tid->baw_head + index) & (ATH_TID_MAX_BUFS - 1);
2706 "%s: tid=%d, seqno %d; window %d:%d; index=%d cindex=%d "
2708 __func__, tid->tid, SEQNO(bf->bf_state.bfs_seqno),
2709 tap->txa_start, tap->txa_wnd, index, cindex, tid->baw_head,
2710 tid->baw_tail);
2713 assert(tid->tx_buf[cindex] == NULL);
2715 if (tid->tx_buf[cindex] != NULL) {
2719 __func__, index, cindex, tid->baw_head, tid->baw_tail);
2723 tid->tx_buf[cindex],
2724 SEQNO(tid->tx_buf[cindex]->bf_state.bfs_seqno),
2729 tid->tx_buf[cindex] = bf;
2731 if (index >= ((tid->baw_tail - tid->baw_head) &
2733 tid->baw_tail = cindex;
2734 INCR(tid->baw_tail, ATH_TID_MAX_BUFS);
2749 struct ath_tid *tid, struct ath_buf *old_bf, struct ath_buf *new_bf)
2757 tap = ath_tx_get_tx_tid(an, tid->tid);
2759 cindex = (tid->baw_head + index) & (ATH_TID_MAX_BUFS - 1);
2776 if (tid->tx_buf[cindex] != old_bf) {
2784 tid->tx_buf[cindex] = new_bf;
2796 struct ath_tid *tid, const struct ath_buf *bf)
2804 tap = ath_tx_get_tx_tid(an, tid->tid);
2806 cindex = (tid->baw_head + index) & (ATH_TID_MAX_BUFS - 1);
2809 "%s: tid=%d, baw=%d:%d, seqno=%d, index=%d, cindex=%d, "
2811 __func__, tid->tid, tap->txa_start, tap->txa_wnd, seqno, index,
2812 cindex, tid->baw_head, tid->baw_tail);
2824 if (tid->tx_buf[cindex] != bf) {
2828 tid->tx_buf[cindex],
2829 (tid->tx_buf[cindex] != NULL) ?
2830 SEQNO(tid->tx_buf[cindex]->bf_state.bfs_seqno) : -1);
2833 tid->tx_buf[cindex] = NULL;
2835 while (tid->baw_head != tid->baw_tail &&
2836 !tid->tx_buf[tid->baw_head]) {
2838 INCR(tid->baw_head, ATH_TID_MAX_BUFS);
2841 "%s: tid=%d: baw is now %d:%d, baw head=%d\n",
2842 __func__, tid->tid, tap->txa_start, tap->txa_wnd, tid->baw_head);
2846 ath_tx_leak_count_update(struct ath_softc *sc, struct ath_tid *tid,
2853 if (tid->an->an_leak_count > 0) {
2859 if ((tid->an->an_stack_psq > 0)
2860 || (tid->an->an_swq_depth > 0))
2868 tid->an->an_node.ni_macaddr,
2870 tid->an->an_leak_count,
2871 tid->an->an_stack_psq,
2872 tid->an->an_swq_depth,
2881 tid->an->an_leak_count --;
2886 ath_tx_tid_can_tx_or_sched(struct ath_softc *sc, struct ath_tid *tid)
2891 if (tid->an->an_leak_count > 0) {
2894 if (tid->paused)
2908 ath_tx_tid_sched(struct ath_softc *sc, struct ath_tid *tid)
2910 struct ath_txq *txq = sc->sc_ac2q[tid->ac];
2919 if (! ath_tx_tid_can_tx_or_sched(sc, tid))
2922 if (tid->sched)
2925 tid->sched = 1;
2932 if (tid->an->an_leak_count) {
2933 TAILQ_INSERT_HEAD(&txq->axq_tidq, tid, axq_qelem);
2935 TAILQ_INSERT_TAIL(&txq->axq_tidq, tid, axq_qelem);
2951 TAILQ_INSERT_TAIL(&txq->axq_tidq, tid, axq_qelem);
2961 ath_tx_tid_unsched(struct ath_softc *sc, struct ath_tid *tid)
2963 struct ath_txq *txq = sc->sc_ac2q[tid->ac];
2967 if (tid->sched == 0)
2970 tid->sched = 0;
2971 TAILQ_REMOVE(&txq->axq_tidq, tid, axq_qelem);
2988 int tid;
2993 tid = ieee80211_gettid(wh);
2995 DPRINTF(sc, ATH_DEBUG_SW_TX, "%s: tid=%d, qos has seq=%d\n",
2996 __func__, tid, IEEE80211_QOS_HAS_SEQ(wh));
3029 seqno = ni->ni_txseqs[tid];
3030 INCR(ni->ni_txseqs[tid], IEEE80211_SEQ_RANGE);
3037 "%s: -> subtype=0x%x, tid=%d, seqno=%d\n",
3038 __func__, subtype, tid, seqno);
3051 struct ath_tid *tid = &an->an_tid[bf->bf_state.bfs_tid];
3056 tap = ath_tx_get_tx_tid(an, tid->tid);
3059 if (! ath_tx_tid_can_tx_or_sched(sc, tid)) {
3060 ATH_TID_INSERT_HEAD(tid, bf, bf_list);
3069 ATH_TID_INSERT_HEAD(tid, bf, bf_list);
3070 ath_tx_tid_sched(sc, tid);
3093 ath_tx_update_clrdmask(sc, tid, bf);
3096 ath_tx_do_ratelookup(sc, bf, tid->tid, bf->bf_state.bfs_pktlen,
3108 tid->hwq_depth++;
3112 ath_tx_addto_baw(sc, an, tid, bf);
3124 ath_tx_leak_count_update(sc, tid, bf);
3143 int pri, tid;
3151 tid = ath_tx_gettid(sc, m0);
3152 atid = &an->an_tid[tid];
3154 DPRINTF(sc, ATH_DEBUG_SW_TX, "%s: bf=%p, pri=%d, tid=%d, qos=%d\n",
3155 __func__, bf, pri, tid, IEEE80211_QOS_HAS_SEQ(wh));
3159 bf->bf_state.bfs_tid = tid;
3183 } else if (ath_tx_ampdu_pending(sc, an, tid)) {
3188 } else if (ath_tx_ampdu_running(sc, an, tid)) {
3331 atid->tid = i;
3356 ath_tx_tid_pause(struct ath_softc *sc, struct ath_tid *tid)
3360 tid->paused++;
3361 DPRINTF(sc, ATH_DEBUG_SW_TX_CTRL, "%s: [%6D]: tid=%d, paused = %d\n",
3363 tid->an->an_node.ni_macaddr, ":",
3364 tid->tid,
3365 tid->paused);
3372 ath_tx_tid_resume(struct ath_softc *sc, struct ath_tid *tid)
3381 if (tid->paused == 0) {
3383 "%s: [%6D]: tid=%d, paused=0?\n",
3385 tid->an->an_node.ni_macaddr, ":",
3386 tid->tid);
3388 tid->paused--;
3392 "%s: [%6D]: tid=%d, unpaused = %d\n",
3394 tid->an->an_node.ni_macaddr, ":",
3395 tid->tid,
3396 tid->paused);
3398 if (tid->paused)
3405 ath_tx_set_clrdmask(sc, tid->an);
3407 if (tid->axq_depth == 0)
3411 if (tid->isfiltered == 1) {
3417 ath_tx_tid_sched(sc, tid);
3430 ath_tx_tid_filt_addbuf(struct ath_softc *sc, struct ath_tid *tid,
3436 if (!tid->isfiltered)
3446 ATH_TID_FILT_INSERT_TAIL(tid, bf, bf_list);
3455 ath_tx_tid_filt_comp_buf(struct ath_softc *sc, struct ath_tid *tid,
3461 if (! tid->isfiltered) {
3462 DPRINTF(sc, ATH_DEBUG_SW_TX_FILT, "%s: tid=%d; filter transition\n",
3463 __func__, tid->tid);
3464 tid->isfiltered = 1;
3465 ath_tx_tid_pause(sc, tid);
3469 ath_tx_tid_filt_addbuf(sc, tid, bf);
3480 ath_tx_tid_filt_comp_complete(struct ath_softc *sc, struct ath_tid *tid)
3487 if (tid->hwq_depth != 0)
3490 DPRINTF(sc, ATH_DEBUG_SW_TX_FILT, "%s: tid=%d, hwq=0, transition back\n",
3491 __func__, tid->tid);
3492 if (tid->isfiltered == 1) {
3493 tid->isfiltered = 0;
3498 ath_tx_set_clrdmask(sc, tid->an);
3501 while ((bf = ATH_TID_FILT_LAST(tid, ath_bufhead_s)) != NULL) {
3502 ATH_TID_FILT_REMOVE(tid, bf, bf_list);
3503 ATH_TID_INSERT_HEAD(tid, bf, bf_list);
3508 ath_tx_tid_resume(sc, tid);
3523 ath_tx_tid_filt_comp_single(struct ath_softc *sc, struct ath_tid *tid,
3550 nbf = ath_tx_retry_clone(sc, tid->an, tid, bf);
3564 ath_tx_tid_filt_comp_buf(sc, tid, nbf);
3568 ath_tx_tid_filt_comp_complete(sc, tid);
3574 ath_tx_tid_filt_comp_aggr(struct ath_softc *sc, struct ath_tid *tid,
3592 "%s: tid=%d, bf=%p, seqno=%d, exceeded retries\n",
3594 tid->tid,
3602 nbf = ath_tx_retry_clone(sc, tid->an, tid, bf);
3604 "%s: tid=%d, busy buffer cloned: %p -> %p, seqno=%d\n",
3605 __func__, tid->tid, bf, nbf, SEQNO(bf->bf_state.bfs_seqno));
3616 "%s: tid=%d, buffer couldn't be cloned! (%p) seqno=%d\n",
3617 __func__, tid->tid, bf, SEQNO(bf->bf_state.bfs_seqno));
3620 ath_tx_tid_filt_comp_buf(sc, tid, nbf);
3626 ath_tx_tid_filt_comp_complete(sc, tid);
3633 ath_tx_tid_bar_suspend(struct ath_softc *sc, struct ath_tid *tid)
3639 "%s: tid=%d, bar_wait=%d, bar_tx=%d, called\n",
3641 tid->tid,
3642 tid->bar_wait,
3643 tid->bar_tx);
3646 if (tid->bar_tx) {
3652 if (tid->bar_wait)
3656 tid->bar_wait = 1;
3659 ath_tx_tid_pause(sc, tid);
3667 ath_tx_tid_bar_unsuspend(struct ath_softc *sc, struct ath_tid *tid)
3675 tid->an->an_node.ni_macaddr,
3677 tid->tid);
3679 if (tid->bar_tx == 0 || tid->bar_wait == 0) {
3682 __func__, tid->an->an_node.ni_macaddr, ":",
3683 tid->tid, tid->bar_tx, tid->bar_wait);
3686 tid->bar_tx = tid->bar_wait = 0;
3687 ath_tx_tid_resume(sc, tid);
3696 ath_tx_tid_bar_tx_ready(struct ath_softc *sc, struct ath_tid *tid)
3701 if (tid->bar_wait == 0 || tid->hwq_depth > 0)
3707 tid->an->an_node.ni_macaddr,
3709 tid->tid);
3727 ath_tx_tid_bar_tx(struct ath_softc *sc, struct ath_tid *tid)
3736 tid->an->an_node.ni_macaddr,
3738 tid->tid);
3740 tap = ath_tx_get_tx_tid(tid->an, tid->tid);
3745 if (tid->bar_wait == 0 || tid->bar_tx == 1) {
3748 __func__, tid->an->an_node.ni_macaddr, ":",
3749 tid->tid, tid->bar_tx, tid->bar_wait);
3754 if (tid->hwq_depth > 0) {
3758 tid->an->an_node.ni_macaddr,
3760 tid->tid,
3761 tid->hwq_depth);
3766 tid->bar_tx = 1;
3772 ath_tx_set_clrdmask(sc, tid->an);
3783 tid->an->an_node.ni_macaddr,
3785 tid->tid,
3792 if (ieee80211_send_bar(&tid->an->an_node, tap, tap->txa_start) == 0) {
3802 __func__, tid->an->an_node.ni_macaddr, ":",
3803 tid->tid);
3804 ath_tx_tid_bar_unsuspend(sc, tid);
3809 struct ath_tid *tid, ath_bufhead *bf_cq, struct ath_buf *bf)
3818 if (ath_tx_ampdu_running(sc, an, tid->tid) &&
3826 ath_tx_update_baw(sc, an, tid, bf);
3849 const char *pfx, struct ath_tid *tid, struct ath_buf *bf)
3855 txq = sc->sc_ac2q[tid->ac];
3856 tap = ath_tx_get_tx_tid(an, tid->tid);
3881 "%s: %s: %6D: bf=%p: tid txq_depth=%d hwq_depth=%d, bar_wait=%d, "
3888 tid->axq_depth,
3889 tid->hwq_depth,
3890 tid->bar_wait,
3891 tid->isfiltered);
3893 "%s: %s: %6D: tid %d: "
3901 tid->tid,
3902 tid->sched, tid->paused,
3903 tid->incomp, tid->baw_head,
3904 tid->baw_tail, tap == NULL ? -1 : tap->txa_start,
3905 ni->ni_txseqs[tid->tid]);
3931 struct ath_tid *tid, ath_bufhead *bf_cq)
3938 tap = ath_tx_get_tx_tid(an, tid->tid);
3945 bf = ATH_TID_FIRST(tid);
3951 ath_tx_tid_drain_print(sc, an, "norm", tid, bf);
3955 ATH_TID_REMOVE(tid, bf, bf_list);
3956 ath_tx_tid_drain_pkt(sc, an, tid, bf_cq, bf);
3962 bf = ATH_TID_FILT_FIRST(tid);
3967 ath_tx_tid_drain_print(sc, an, "filt", tid, bf);
3971 ATH_TID_FILT_REMOVE(tid, bf, bf_list);
3972 ath_tx_tid_drain_pkt(sc, an, tid, bf_cq, bf);
3982 ath_tx_set_clrdmask(sc, tid->an);
4006 tid->tid,
4009 ni->ni_txseqs[tid->tid] = tap->txa_start;
4010 tid->baw_tail = tid->baw_head;
4020 ath_tx_tid_reset(struct ath_softc *sc, struct ath_tid *tid)
4024 tid->bar_wait = tid->bar_tx = tid->isfiltered = 0;
4025 tid->paused = tid->sched = tid->addba_tx_pending = 0;
4026 tid->incomp = tid->cleanup_inprogress = 0;
4038 if (tid->bar_wait) {
4039 if (tid->paused > 0) {
4040 tid->paused --;
4053 if (tid->isfiltered) {
4054 if (tid->paused > 0) {
4055 tid->paused --;
4064 tid->bar_wait = 0;
4065 tid->bar_tx = 0;
4066 tid->isfiltered = 0;
4067 tid->sched = 0;
4068 tid->addba_tx_pending = 0;
4091 int tid;
4114 for (tid = 0; tid < IEEE80211_TID_SIZE; tid++) {
4115 struct ath_tid *atid = &an->an_tid[tid];
4120 /* Remove this tid from the list of active tids */
4146 struct ath_tid *tid;
4158 tid = TAILQ_FIRST(&txq->axq_tidq);
4159 ath_tx_tid_drain(sc, tid->an, tid, &bf_cq);
4160 ath_tx_tid_unsched(sc, tid);
4192 int tid = bf->bf_state.bfs_tid;
4193 struct ath_tid *atid = &an->an_tid[tid];
4232 __func__, tid);
4279 int tid = bf->bf_state.bfs_tid;
4280 struct ath_tid *atid = &an->an_tid[tid];
4283 __func__, tid, atid->incomp);
4300 __func__, tid);
4316 int tid, struct ath_buf *bf_head, ath_bufhead *bf_cq)
4318 struct ath_tid *atid = &an->an_tid[tid];
4380 ath_tx_tid_cleanup(struct ath_softc *sc, struct ath_node *an, int tid,
4383 struct ath_tid *atid = &an->an_tid[tid];
4389 "%s: TID %d: called; inprogress=%d\n", __func__, tid,
4420 ath_tx_tid_cleanup_frame(sc, an, tid, bf, bf_cq);
4445 __func__, tid, atid->incomp);
4452 struct ath_tid *tid, struct ath_buf *bf)
4498 ath_tx_switch_baw_buf(sc, an, tid, bf, nbf);
4521 int tid = bf->bf_state.bfs_tid;
4522 struct ath_tid *atid = &an->an_tid[tid];
4527 tap = ath_tx_get_tx_tid(an, tid);
4612 int tid = bf->bf_state.bfs_tid;
4613 struct ath_tid *atid = &an->an_tid[tid];
4674 struct ath_tid *tid)
4697 tap = ath_tx_get_tx_tid(an, tid->tid);
4717 ATH_TID_INSERT_HEAD(tid, bf, bf_list);
4723 ath_tx_tid_sched(sc, tid);
4734 ath_tx_tid_bar_suspend(sc, tid);
4740 if (ath_tx_tid_bar_tx_ready(sc, tid))
4741 ath_tx_tid_bar_tx(sc, tid);
4764 int tid = bf_first->bf_state.bfs_tid;
4765 struct ath_tid *atid = &an->an_tid[tid];
4789 __func__, tid);
4828 int tid = bf_first->bf_state.bfs_tid;
4829 struct ath_tid *atid = &an->an_tid[tid];
4956 tap = ath_tx_get_tx_tid(an, tid);
4988 * out tid 1 - the aggregate frames are all marked as TID 1,
4994 if (tid != ts.ts_tid) {
4995 DPRINTF(sc, ATH_DEBUG_SW_TX_AGGR, "%s: tid %d != hw tid %d\n",
4996 __func__, tid, ts.ts_tid);
5169 int tid = bf->bf_state.bfs_tid;
5170 struct ath_tid *atid = &an->an_tid[tid];
5200 if (tid == IEEE80211_NONQOS_TID)
5204 "%s: bf=%p: tid=%d, hwq_depth=%d, seqno=%d\n",
5318 __func__, tid, SEQNO(bf->bf_state.bfs_seqno));
5372 struct ath_tid *tid)
5380 tap = ath_tx_get_tx_tid(an, tid->tid);
5387 TAILQ_FOREACH(bf, &tid->tid_q, bf_list) {
5428 struct ath_tid *tid)
5431 struct ath_txq *txq = sc->sc_ac2q[tid->ac];
5437 DPRINTF(sc, ATH_DEBUG_SW_TX, "%s: tid=%d\n", __func__, tid->tid);
5445 tap = ath_tx_get_tx_tid(an, tid->tid);
5447 if (tid->tid == IEEE80211_NONQOS_TID)
5462 if (! ath_tx_tid_can_tx_or_sched(sc, tid))
5465 bf = ATH_TID_FIRST(tid);
5478 ATH_TID_REMOVE(tid, bf, bf_list);
5497 ath_tx_update_clrdmask(sc, tid, bf);
5499 ath_tx_do_ratelookup(sc, bf, tid->tid,
5521 swq_pktbytes = ath_tx_tid_swq_depth_bytes(sc, an, tid);
5522 ath_tx_do_ratelookup(sc, bf, tid->tid, swq_pktbytes, true);
5535 status = ath_tx_form_aggr(sc, an, tid, &bf_q);
5565 ath_tx_update_clrdmask(sc, tid, bf);
5586 ath_tx_update_clrdmask(sc, tid, bf);
5621 ath_tx_leak_count_update(sc, tid, bf);
5628 tid->hwq_depth++;
5659 struct ath_tid *tid)
5662 struct ath_txq *txq = sc->sc_ac2q[tid->ac];
5665 __func__, an, tid->tid);
5670 if (ath_tx_ampdu_pending(sc, an, tid->tid))
5671 DPRINTF(sc, ATH_DEBUG_SW_TX, "%s: tid=%d, ampdu pending?\n",
5672 __func__, tid->tid);
5673 if (ath_tx_ampdu_running(sc, an, tid->tid))
5674 DPRINTF(sc, ATH_DEBUG_SW_TX, "%s: tid=%d, ampdu running?\n",
5675 __func__, tid->tid);
5685 if (! ath_tx_tid_can_tx_or_sched(sc, tid))
5688 bf = ATH_TID_FIRST(tid);
5693 ATH_TID_REMOVE(tid, bf, bf_list);
5696 if (tid->tid != bf->bf_state.bfs_tid) {
5698 " tid %d\n", __func__, bf->bf_state.bfs_tid,
5699 tid->tid);
5711 ath_tx_update_clrdmask(sc, tid, bf);
5714 ath_tx_do_ratelookup(sc, bf, tid->tid,
5727 ath_tx_leak_count_update(sc, tid, bf);
5731 tid->hwq_depth++;
5752 struct ath_tid *tid, *next, *last;
5797 TAILQ_FOREACH_SAFE(tid, &txq->axq_tidq, axq_qelem, next) {
5802 DPRINTF(sc, ATH_DEBUG_SW_TX, "%s: tid=%d, paused=%d\n",
5803 __func__, tid->tid, tid->paused);
5804 ath_tx_tid_unsched(sc, tid);
5809 if (! ath_tx_tid_can_tx_or_sched(sc, tid)) {
5812 if (ath_tx_ampdu_running(sc, tid->an, tid->tid))
5813 ath_tx_tid_hw_queue_aggr(sc, tid->an, tid);
5815 ath_tx_tid_hw_queue_norm(sc, tid->an, tid);
5818 if (tid->axq_depth != 0)
5819 ath_tx_tid_sched(sc, tid);
5844 if (tid == last)
5857 ath_tx_get_tx_tid(struct ath_node *an, int tid)
5862 if (tid == IEEE80211_NONQOS_TID)
5865 tap = &ni->ni_tx_ampdu[tid];
5873 ath_tx_ampdu_running(struct ath_softc *sc, struct ath_node *an, int tid)
5877 if (tid == IEEE80211_NONQOS_TID)
5880 tap = ath_tx_get_tx_tid(an, tid);
5891 ath_tx_ampdu_pending(struct ath_softc *sc, struct ath_node *an, int tid)
5895 if (tid == IEEE80211_NONQOS_TID)
5898 tap = ath_tx_get_tx_tid(an, tid);
5922 int tid = tap->txa_tid;
5924 struct ath_tid *atid = &an->an_tid[tid];
5969 __func__, tap->txa_start, ni->ni_txseqs[tid]);
6000 int tid = tap->txa_tid;
6002 struct ath_tid *atid = &an->an_tid[tid];
6013 __func__, tap->txa_start, ni->ni_txseqs[tid]);
6030 tap->txa_start = ni->ni_txseqs[tid];
6046 int tid = tap->txa_tid;
6048 struct ath_tid *atid = &an->an_tid[tid];
6097 ath_tx_tid_cleanup(sc, an, tid, &bf_cq);
6123 struct ath_tid *tid;
6134 tid = &an->an_tid[i];
6135 if (tid->hwq_depth == 0)
6147 if (! tid->cleanup_inprogress) {
6148 ath_tx_tid_pause(sc, tid);
6153 if (! tid->cleanup_inprogress)
6154 ath_tx_tid_resume(sc, tid);
6181 int tid = tap->txa_tid;
6183 struct ath_tid *atid = &an->an_tid[tid];
6188 "%s: %6D: called; txa_tid=%d, atid->tid=%d, status=%d, attempts=%d, txa_start=%d, txa_seqpending=%d\n",
6193 atid->tid,
6214 device_printf(sc->sc_dev, "%s: tid=%d; txa_start=%d, old=%d, adjusting\n",
6216 tid,
6254 int tid = tap->txa_tid;
6256 struct ath_tid *atid = &an->an_tid[tid];
6263 tid);
6306 * 'active tid' list and this requires both locks to be held.
6313 int tid;
6328 for (tid = 0; tid < IEEE80211_TID_SIZE; tid++) {
6329 atid = &an->an_tid[tid];
6350 int tid;
6372 for (tid = 0; tid < IEEE80211_TID_SIZE; tid++) {
6373 atid = &an->an_tid[tid];