Lines Matching refs:tid

124     int tid);
126 int tid);
130 struct ieee80211_node *ni, struct mbuf *m0, int *tid);
133 struct ath_tid *tid, struct ath_buf *bf);
1377 ath_tx_do_ratelookup(struct ath_softc *sc, struct ath_buf *bf, int tid,
1393 pktlen, tid, is_aggr, &rix, &try0, &rate, &maxdur, &maxpktlen);
1416 ath_tx_update_clrdmask(struct ath_softc *sc, struct ath_tid *tid,
1507 struct ath_tid *tid = &an->an_tid[bf->bf_state.bfs_tid];
1523 ath_tx_do_ratelookup(sc, bf, tid->tid, bf->bf_state.bfs_pktlen, false);
1531 tid->hwq_depth++;
1909 int tid;
1946 tid = ath_tx_gettid(sc, m0);
1995 is_ampdu_tx = ath_tx_ampdu_running(sc, ATH_NODE(ni), tid);
1996 is_ampdu_pending = ath_tx_ampdu_pending(sc, ATH_NODE(ni), tid);
1999 DPRINTF(sc, ATH_DEBUG_SW_TX, "%s: tid=%d, ac=%d, is_ampdu=%d\n",
2000 __func__, tid, pri, is_ampdu);
2003 bf->bf_state.bfs_tid = tid;
2068 "%s: tid %d: ampdu pending, seqno %d\n",
2069 __func__, tid, M_SEQNO_GET(m0));
2129 ath_tx_leak_count_update(sc, tid, bf);
2191 "%s: overriding tid %d pri %d -> %d\n",
2406 ath_tx_leak_count_update(sc, tid, bf);
2607 struct mbuf *m0, int *tid)
2637 *tid = (int) _IEEE80211_MASKSHIFT(baparamset, IEEE80211_BAPS_TID);
2656 struct ath_tid *tid, struct ath_buf *bf)
2666 tap = ath_tx_get_tx_tid(an, tid->tid);
2677 "%s: re-added? tid=%d, seqno %d; window %d:%d; "
2679 __func__, tid->tid, SEQNO(bf->bf_state.bfs_seqno),
2680 tap->txa_start, tap->txa_wnd, tid->baw_head,
2681 tid->baw_tail);
2690 "%s: bf=%p: outside of BAW?? tid=%d, seqno %d; window %d:%d; "
2692 __func__, bf, tid->tid, SEQNO(bf->bf_state.bfs_seqno),
2693 tap->txa_start, tap->txa_wnd, tid->baw_head,
2694 tid->baw_tail);
2702 cindex = (tid->baw_head + index) & (ATH_TID_MAX_BUFS - 1);
2704 "%s: tid=%d, seqno %d; window %d:%d; index=%d cindex=%d "
2706 __func__, tid->tid, SEQNO(bf->bf_state.bfs_seqno),
2707 tap->txa_start, tap->txa_wnd, index, cindex, tid->baw_head,
2708 tid->baw_tail);
2711 assert(tid->tx_buf[cindex] == NULL);
2713 if (tid->tx_buf[cindex] != NULL) {
2717 __func__, index, cindex, tid->baw_head, tid->baw_tail);
2721 tid->tx_buf[cindex],
2722 SEQNO(tid->tx_buf[cindex]->bf_state.bfs_seqno),
2727 tid->tx_buf[cindex] = bf;
2729 if (index >= ((tid->baw_tail - tid->baw_head) &
2731 tid->baw_tail = cindex;
2732 INCR(tid->baw_tail, ATH_TID_MAX_BUFS);
2747 struct ath_tid *tid, struct ath_buf *old_bf, struct ath_buf *new_bf)
2755 tap = ath_tx_get_tx_tid(an, tid->tid);
2757 cindex = (tid->baw_head + index) & (ATH_TID_MAX_BUFS - 1);
2774 if (tid->tx_buf[cindex] != old_bf) {
2782 tid->tx_buf[cindex] = new_bf;
2794 struct ath_tid *tid, const struct ath_buf *bf)
2802 tap = ath_tx_get_tx_tid(an, tid->tid);
2804 cindex = (tid->baw_head + index) & (ATH_TID_MAX_BUFS - 1);
2807 "%s: tid=%d, baw=%d:%d, seqno=%d, index=%d, cindex=%d, "
2809 __func__, tid->tid, tap->txa_start, tap->txa_wnd, seqno, index,
2810 cindex, tid->baw_head, tid->baw_tail);
2822 if (tid->tx_buf[cindex] != bf) {
2826 tid->tx_buf[cindex],
2827 (tid->tx_buf[cindex] != NULL) ?
2828 SEQNO(tid->tx_buf[cindex]->bf_state.bfs_seqno) : -1);
2831 tid->tx_buf[cindex] = NULL;
2833 while (tid->baw_head != tid->baw_tail &&
2834 !tid->tx_buf[tid->baw_head]) {
2836 INCR(tid->baw_head, ATH_TID_MAX_BUFS);
2839 "%s: tid=%d: baw is now %d:%d, baw head=%d\n",
2840 __func__, tid->tid, tap->txa_start, tap->txa_wnd, tid->baw_head);
2844 ath_tx_leak_count_update(struct ath_softc *sc, struct ath_tid *tid,
2851 if (tid->an->an_leak_count > 0) {
2857 if ((tid->an->an_stack_psq > 0)
2858 || (tid->an->an_swq_depth > 0))
2866 tid->an->an_node.ni_macaddr,
2868 tid->an->an_leak_count,
2869 tid->an->an_stack_psq,
2870 tid->an->an_swq_depth,
2879 tid->an->an_leak_count --;
2884 ath_tx_tid_can_tx_or_sched(struct ath_softc *sc, struct ath_tid *tid)
2889 if (tid->an->an_leak_count > 0) {
2892 if (tid->paused)
2906 ath_tx_tid_sched(struct ath_softc *sc, struct ath_tid *tid)
2908 struct ath_txq *txq = sc->sc_ac2q[tid->ac];
2917 if (! ath_tx_tid_can_tx_or_sched(sc, tid))
2920 if (tid->sched)
2923 tid->sched = 1;
2930 if (tid->an->an_leak_count) {
2931 TAILQ_INSERT_HEAD(&txq->axq_tidq, tid, axq_qelem);
2933 TAILQ_INSERT_TAIL(&txq->axq_tidq, tid, axq_qelem);
2949 TAILQ_INSERT_TAIL(&txq->axq_tidq, tid, axq_qelem);
2959 ath_tx_tid_unsched(struct ath_softc *sc, struct ath_tid *tid)
2961 struct ath_txq *txq = sc->sc_ac2q[tid->ac];
2965 if (tid->sched == 0)
2968 tid->sched = 0;
2969 TAILQ_REMOVE(&txq->axq_tidq, tid, axq_qelem);
2986 int tid;
2991 tid = ieee80211_gettid(wh);
2993 DPRINTF(sc, ATH_DEBUG_SW_TX, "%s: tid=%d, qos has seq=%d\n",
2994 __func__, tid, IEEE80211_QOS_HAS_SEQ(wh));
3027 seqno = ni->ni_txseqs[tid];
3028 INCR(ni->ni_txseqs[tid], IEEE80211_SEQ_RANGE);
3035 "%s: -> subtype=0x%x, tid=%d, seqno=%d\n",
3036 __func__, subtype, tid, seqno);
3049 struct ath_tid *tid = &an->an_tid[bf->bf_state.bfs_tid];
3054 tap = ath_tx_get_tx_tid(an, tid->tid);
3057 if (! ath_tx_tid_can_tx_or_sched(sc, tid)) {
3058 ATH_TID_INSERT_HEAD(tid, bf, bf_list);
3067 ATH_TID_INSERT_HEAD(tid, bf, bf_list);
3068 ath_tx_tid_sched(sc, tid);
3091 ath_tx_update_clrdmask(sc, tid, bf);
3094 ath_tx_do_ratelookup(sc, bf, tid->tid, bf->bf_state.bfs_pktlen,
3106 tid->hwq_depth++;
3110 ath_tx_addto_baw(sc, an, tid, bf);
3122 ath_tx_leak_count_update(sc, tid, bf);
3141 int pri, tid;
3149 tid = ath_tx_gettid(sc, m0);
3150 atid = &an->an_tid[tid];
3152 DPRINTF(sc, ATH_DEBUG_SW_TX, "%s: bf=%p, pri=%d, tid=%d, qos=%d\n",
3153 __func__, bf, pri, tid, IEEE80211_QOS_HAS_SEQ(wh));
3157 bf->bf_state.bfs_tid = tid;
3181 } else if (ath_tx_ampdu_pending(sc, an, tid)) {
3186 } else if (ath_tx_ampdu_running(sc, an, tid)) {
3329 atid->tid = i;
3354 ath_tx_tid_pause(struct ath_softc *sc, struct ath_tid *tid)
3358 tid->paused++;
3359 DPRINTF(sc, ATH_DEBUG_SW_TX_CTRL, "%s: [%6D]: tid=%d, paused = %d\n",
3361 tid->an->an_node.ni_macaddr, ":",
3362 tid->tid,
3363 tid->paused);
3370 ath_tx_tid_resume(struct ath_softc *sc, struct ath_tid *tid)
3379 if (tid->paused == 0) {
3381 "%s: [%6D]: tid=%d, paused=0?\n",
3383 tid->an->an_node.ni_macaddr, ":",
3384 tid->tid);
3386 tid->paused--;
3390 "%s: [%6D]: tid=%d, unpaused = %d\n",
3392 tid->an->an_node.ni_macaddr, ":",
3393 tid->tid,
3394 tid->paused);
3396 if (tid->paused)
3403 ath_tx_set_clrdmask(sc, tid->an);
3405 if (tid->axq_depth == 0)
3409 if (tid->isfiltered == 1) {
3415 ath_tx_tid_sched(sc, tid);
3428 ath_tx_tid_filt_addbuf(struct ath_softc *sc, struct ath_tid *tid,
3434 if (!tid->isfiltered)
3444 ATH_TID_FILT_INSERT_TAIL(tid, bf, bf_list);
3453 ath_tx_tid_filt_comp_buf(struct ath_softc *sc, struct ath_tid *tid,
3459 if (! tid->isfiltered) {
3460 DPRINTF(sc, ATH_DEBUG_SW_TX_FILT, "%s: tid=%d; filter transition\n",
3461 __func__, tid->tid);
3462 tid->isfiltered = 1;
3463 ath_tx_tid_pause(sc, tid);
3467 ath_tx_tid_filt_addbuf(sc, tid, bf);
3478 ath_tx_tid_filt_comp_complete(struct ath_softc *sc, struct ath_tid *tid)
3485 if (tid->hwq_depth != 0)
3488 DPRINTF(sc, ATH_DEBUG_SW_TX_FILT, "%s: tid=%d, hwq=0, transition back\n",
3489 __func__, tid->tid);
3490 if (tid->isfiltered == 1) {
3491 tid->isfiltered = 0;
3496 ath_tx_set_clrdmask(sc, tid->an);
3499 while ((bf = ATH_TID_FILT_LAST(tid, ath_bufhead_s)) != NULL) {
3500 ATH_TID_FILT_REMOVE(tid, bf, bf_list);
3501 ATH_TID_INSERT_HEAD(tid, bf, bf_list);
3506 ath_tx_tid_resume(sc, tid);
3521 ath_tx_tid_filt_comp_single(struct ath_softc *sc, struct ath_tid *tid,
3548 nbf = ath_tx_retry_clone(sc, tid->an, tid, bf);
3562 ath_tx_tid_filt_comp_buf(sc, tid, nbf);
3566 ath_tx_tid_filt_comp_complete(sc, tid);
3572 ath_tx_tid_filt_comp_aggr(struct ath_softc *sc, struct ath_tid *tid,
3590 "%s: tid=%d, bf=%p, seqno=%d, exceeded retries\n",
3592 tid->tid,
3600 nbf = ath_tx_retry_clone(sc, tid->an, tid, bf);
3602 "%s: tid=%d, busy buffer cloned: %p -> %p, seqno=%d\n",
3603 __func__, tid->tid, bf, nbf, SEQNO(bf->bf_state.bfs_seqno));
3614 "%s: tid=%d, buffer couldn't be cloned! (%p) seqno=%d\n",
3615 __func__, tid->tid, bf, SEQNO(bf->bf_state.bfs_seqno));
3618 ath_tx_tid_filt_comp_buf(sc, tid, nbf);
3624 ath_tx_tid_filt_comp_complete(sc, tid);
3631 ath_tx_tid_bar_suspend(struct ath_softc *sc, struct ath_tid *tid)
3637 "%s: tid=%d, bar_wait=%d, bar_tx=%d, called\n",
3639 tid->tid,
3640 tid->bar_wait,
3641 tid->bar_tx);
3644 if (tid->bar_tx) {
3650 if (tid->bar_wait)
3654 tid->bar_wait = 1;
3657 ath_tx_tid_pause(sc, tid);
3665 ath_tx_tid_bar_unsuspend(struct ath_softc *sc, struct ath_tid *tid)
3673 tid->an->an_node.ni_macaddr,
3675 tid->tid);
3677 if (tid->bar_tx == 0 || tid->bar_wait == 0) {
3680 __func__, tid->an->an_node.ni_macaddr, ":",
3681 tid->tid, tid->bar_tx, tid->bar_wait);
3684 tid->bar_tx = tid->bar_wait = 0;
3685 ath_tx_tid_resume(sc, tid);
3694 ath_tx_tid_bar_tx_ready(struct ath_softc *sc, struct ath_tid *tid)
3699 if (tid->bar_wait == 0 || tid->hwq_depth > 0)
3705 tid->an->an_node.ni_macaddr,
3707 tid->tid);
3725 ath_tx_tid_bar_tx(struct ath_softc *sc, struct ath_tid *tid)
3734 tid->an->an_node.ni_macaddr,
3736 tid->tid);
3738 tap = ath_tx_get_tx_tid(tid->an, tid->tid);
3743 if (tid->bar_wait == 0 || tid->bar_tx == 1) {
3746 __func__, tid->an->an_node.ni_macaddr, ":",
3747 tid->tid, tid->bar_tx, tid->bar_wait);
3752 if (tid->hwq_depth > 0) {
3756 tid->an->an_node.ni_macaddr,
3758 tid->tid,
3759 tid->hwq_depth);
3764 tid->bar_tx = 1;
3770 ath_tx_set_clrdmask(sc, tid->an);
3781 tid->an->an_node.ni_macaddr,
3783 tid->tid,
3790 if (ieee80211_send_bar(&tid->an->an_node, tap, tap->txa_start) == 0) {
3800 __func__, tid->an->an_node.ni_macaddr, ":",
3801 tid->tid);
3802 ath_tx_tid_bar_unsuspend(sc, tid);
3807 struct ath_tid *tid, ath_bufhead *bf_cq, struct ath_buf *bf)
3816 if (ath_tx_ampdu_running(sc, an, tid->tid) &&
3824 ath_tx_update_baw(sc, an, tid, bf);
3847 const char *pfx, struct ath_tid *tid, struct ath_buf *bf)
3853 txq = sc->sc_ac2q[tid->ac];
3854 tap = ath_tx_get_tx_tid(an, tid->tid);
3879 "%s: %s: %6D: bf=%p: tid txq_depth=%d hwq_depth=%d, bar_wait=%d, "
3886 tid->axq_depth,
3887 tid->hwq_depth,
3888 tid->bar_wait,
3889 tid->isfiltered);
3891 "%s: %s: %6D: tid %d: "
3899 tid->tid,
3900 tid->sched, tid->paused,
3901 tid->incomp, tid->baw_head,
3902 tid->baw_tail, tap == NULL ? -1 : tap->txa_start,
3903 ni->ni_txseqs[tid->tid]);
3929 struct ath_tid *tid, ath_bufhead *bf_cq)
3936 tap = ath_tx_get_tx_tid(an, tid->tid);
3943 bf = ATH_TID_FIRST(tid);
3949 ath_tx_tid_drain_print(sc, an, "norm", tid, bf);
3953 ATH_TID_REMOVE(tid, bf, bf_list);
3954 ath_tx_tid_drain_pkt(sc, an, tid, bf_cq, bf);
3960 bf = ATH_TID_FILT_FIRST(tid);
3965 ath_tx_tid_drain_print(sc, an, "filt", tid, bf);
3969 ATH_TID_FILT_REMOVE(tid, bf, bf_list);
3970 ath_tx_tid_drain_pkt(sc, an, tid, bf_cq, bf);
3980 ath_tx_set_clrdmask(sc, tid->an);
4004 tid->tid,
4007 ni->ni_txseqs[tid->tid] = tap->txa_start;
4008 tid->baw_tail = tid->baw_head;
4018 ath_tx_tid_reset(struct ath_softc *sc, struct ath_tid *tid)
4022 tid->bar_wait = tid->bar_tx = tid->isfiltered = 0;
4023 tid->paused = tid->sched = tid->addba_tx_pending = 0;
4024 tid->incomp = tid->cleanup_inprogress = 0;
4036 if (tid->bar_wait) {
4037 if (tid->paused > 0) {
4038 tid->paused --;
4051 if (tid->isfiltered) {
4052 if (tid->paused > 0) {
4053 tid->paused --;
4062 tid->bar_wait = 0;
4063 tid->bar_tx = 0;
4064 tid->isfiltered = 0;
4065 tid->sched = 0;
4066 tid->addba_tx_pending = 0;
4089 int tid;
4112 for (tid = 0; tid < IEEE80211_TID_SIZE; tid++) {
4113 struct ath_tid *atid = &an->an_tid[tid];
4118 /* Remove this tid from the list of active tids */
4144 struct ath_tid *tid;
4156 tid = TAILQ_FIRST(&txq->axq_tidq);
4157 ath_tx_tid_drain(sc, tid->an, tid, &bf_cq);
4158 ath_tx_tid_unsched(sc, tid);
4190 int tid = bf->bf_state.bfs_tid;
4191 struct ath_tid *atid = &an->an_tid[tid];
4230 __func__, tid);
4277 int tid = bf->bf_state.bfs_tid;
4278 struct ath_tid *atid = &an->an_tid[tid];
4281 __func__, tid, atid->incomp);
4298 __func__, tid);
4314 int tid, struct ath_buf *bf_head, ath_bufhead *bf_cq)
4316 struct ath_tid *atid = &an->an_tid[tid];
4378 ath_tx_tid_cleanup(struct ath_softc *sc, struct ath_node *an, int tid,
4381 struct ath_tid *atid = &an->an_tid[tid];
4387 "%s: TID %d: called; inprogress=%d\n", __func__, tid,
4418 ath_tx_tid_cleanup_frame(sc, an, tid, bf, bf_cq);
4443 __func__, tid, atid->incomp);
4450 struct ath_tid *tid, struct ath_buf *bf)
4496 ath_tx_switch_baw_buf(sc, an, tid, bf, nbf);
4519 int tid = bf->bf_state.bfs_tid;
4520 struct ath_tid *atid = &an->an_tid[tid];
4525 tap = ath_tx_get_tx_tid(an, tid);
4610 int tid = bf->bf_state.bfs_tid;
4611 struct ath_tid *atid = &an->an_tid[tid];
4672 struct ath_tid *tid)
4695 tap = ath_tx_get_tx_tid(an, tid->tid);
4715 ATH_TID_INSERT_HEAD(tid, bf, bf_list);
4721 ath_tx_tid_sched(sc, tid);
4732 ath_tx_tid_bar_suspend(sc, tid);
4738 if (ath_tx_tid_bar_tx_ready(sc, tid))
4739 ath_tx_tid_bar_tx(sc, tid);
4762 int tid = bf_first->bf_state.bfs_tid;
4763 struct ath_tid *atid = &an->an_tid[tid];
4787 __func__, tid);
4826 int tid = bf_first->bf_state.bfs_tid;
4827 struct ath_tid *atid = &an->an_tid[tid];
4954 tap = ath_tx_get_tx_tid(an, tid);
4986 * out tid 1 - the aggregate frames are all marked as TID 1,
4992 if (tid != ts.ts_tid) {
4993 DPRINTF(sc, ATH_DEBUG_SW_TX_AGGR, "%s: tid %d != hw tid %d\n",
4994 __func__, tid, ts.ts_tid);
5167 int tid = bf->bf_state.bfs_tid;
5168 struct ath_tid *atid = &an->an_tid[tid];
5198 if (tid == IEEE80211_NONQOS_TID)
5202 "%s: bf=%p: tid=%d, hwq_depth=%d, seqno=%d\n",
5316 __func__, tid, SEQNO(bf->bf_state.bfs_seqno));
5370 struct ath_tid *tid)
5378 tap = ath_tx_get_tx_tid(an, tid->tid);
5385 TAILQ_FOREACH(bf, &tid->tid_q, bf_list) {
5426 struct ath_tid *tid)
5429 struct ath_txq *txq = sc->sc_ac2q[tid->ac];
5435 DPRINTF(sc, ATH_DEBUG_SW_TX, "%s: tid=%d\n", __func__, tid->tid);
5443 tap = ath_tx_get_tx_tid(an, tid->tid);
5445 if (tid->tid == IEEE80211_NONQOS_TID)
5460 if (! ath_tx_tid_can_tx_or_sched(sc, tid))
5463 bf = ATH_TID_FIRST(tid);
5476 ATH_TID_REMOVE(tid, bf, bf_list);
5495 ath_tx_update_clrdmask(sc, tid, bf);
5497 ath_tx_do_ratelookup(sc, bf, tid->tid,
5519 swq_pktbytes = ath_tx_tid_swq_depth_bytes(sc, an, tid);
5520 ath_tx_do_ratelookup(sc, bf, tid->tid, swq_pktbytes, true);
5533 status = ath_tx_form_aggr(sc, an, tid, &bf_q);
5563 ath_tx_update_clrdmask(sc, tid, bf);
5584 ath_tx_update_clrdmask(sc, tid, bf);
5619 ath_tx_leak_count_update(sc, tid, bf);
5626 tid->hwq_depth++;
5657 struct ath_tid *tid)
5660 struct ath_txq *txq = sc->sc_ac2q[tid->ac];
5663 __func__, an, tid->tid);
5668 if (ath_tx_ampdu_pending(sc, an, tid->tid))
5669 DPRINTF(sc, ATH_DEBUG_SW_TX, "%s: tid=%d, ampdu pending?\n",
5670 __func__, tid->tid);
5671 if (ath_tx_ampdu_running(sc, an, tid->tid))
5672 DPRINTF(sc, ATH_DEBUG_SW_TX, "%s: tid=%d, ampdu running?\n",
5673 __func__, tid->tid);
5683 if (! ath_tx_tid_can_tx_or_sched(sc, tid))
5686 bf = ATH_TID_FIRST(tid);
5691 ATH_TID_REMOVE(tid, bf, bf_list);
5694 if (tid->tid != bf->bf_state.bfs_tid) {
5696 " tid %d\n", __func__, bf->bf_state.bfs_tid,
5697 tid->tid);
5709 ath_tx_update_clrdmask(sc, tid, bf);
5712 ath_tx_do_ratelookup(sc, bf, tid->tid,
5725 ath_tx_leak_count_update(sc, tid, bf);
5729 tid->hwq_depth++;
5750 struct ath_tid *tid, *next, *last;
5795 TAILQ_FOREACH_SAFE(tid, &txq->axq_tidq, axq_qelem, next) {
5800 DPRINTF(sc, ATH_DEBUG_SW_TX, "%s: tid=%d, paused=%d\n",
5801 __func__, tid->tid, tid->paused);
5802 ath_tx_tid_unsched(sc, tid);
5807 if (! ath_tx_tid_can_tx_or_sched(sc, tid)) {
5810 if (ath_tx_ampdu_running(sc, tid->an, tid->tid))
5811 ath_tx_tid_hw_queue_aggr(sc, tid->an, tid);
5813 ath_tx_tid_hw_queue_norm(sc, tid->an, tid);
5816 if (tid->axq_depth != 0)
5817 ath_tx_tid_sched(sc, tid);
5842 if (tid == last)
5855 ath_tx_get_tx_tid(struct ath_node *an, int tid)
5860 if (tid == IEEE80211_NONQOS_TID)
5863 tap = &ni->ni_tx_ampdu[tid];
5871 ath_tx_ampdu_running(struct ath_softc *sc, struct ath_node *an, int tid)
5875 if (tid == IEEE80211_NONQOS_TID)
5878 tap = ath_tx_get_tx_tid(an, tid);
5889 ath_tx_ampdu_pending(struct ath_softc *sc, struct ath_node *an, int tid)
5893 if (tid == IEEE80211_NONQOS_TID)
5896 tap = ath_tx_get_tx_tid(an, tid);
5920 int tid = tap->txa_tid;
5922 struct ath_tid *atid = &an->an_tid[tid];
5967 __func__, tap->txa_start, ni->ni_txseqs[tid]);
5998 int tid = tap->txa_tid;
6000 struct ath_tid *atid = &an->an_tid[tid];
6011 __func__, tap->txa_start, ni->ni_txseqs[tid]);
6028 tap->txa_start = ni->ni_txseqs[tid];
6044 int tid = tap->txa_tid;
6046 struct ath_tid *atid = &an->an_tid[tid];
6095 ath_tx_tid_cleanup(sc, an, tid, &bf_cq);
6121 struct ath_tid *tid;
6132 tid = &an->an_tid[i];
6133 if (tid->hwq_depth == 0)
6145 if (! tid->cleanup_inprogress) {
6146 ath_tx_tid_pause(sc, tid);
6151 if (! tid->cleanup_inprogress)
6152 ath_tx_tid_resume(sc, tid);
6179 int tid = tap->txa_tid;
6181 struct ath_tid *atid = &an->an_tid[tid];
6186 "%s: %6D: called; txa_tid=%d, atid->tid=%d, status=%d, attempts=%d, txa_start=%d, txa_seqpending=%d\n",
6191 atid->tid,
6212 device_printf(sc->sc_dev, "%s: tid=%d; txa_start=%d, old=%d, adjusting\n",
6214 tid,
6252 int tid = tap->txa_tid;
6254 struct ath_tid *atid = &an->an_tid[tid];
6261 tid);
6304 * 'active tid' list and this requires both locks to be held.
6311 int tid;
6326 for (tid = 0; tid < IEEE80211_TID_SIZE; tid++) {
6327 atid = &an->an_tid[tid];
6348 int tid;
6370 for (tid = 0; tid < IEEE80211_TID_SIZE; tid++) {
6371 atid = &an->an_tid[tid];