Lines Matching defs:ra_list

100 	struct mwifiex_ra_list_tbl *ra_list;
102 ra_list = kzalloc(sizeof(struct mwifiex_ra_list_tbl), GFP_ATOMIC);
103 if (!ra_list)
106 INIT_LIST_HEAD(&ra_list->list);
107 skb_queue_head_init(&ra_list->skb_head);
109 memcpy(ra_list->ra, ra, ETH_ALEN);
111 ra_list->total_pkt_count = 0;
113 mwifiex_dbg(adapter, INFO, "info: allocated ra_list %p\n", ra_list);
115 return ra_list;
141 struct mwifiex_ra_list_tbl *ra_list;
147 ra_list = mwifiex_wmm_allocate_ralist_node(adapter, ra);
149 "info: created ra_list %p\n", ra_list);
151 if (!ra_list)
154 ra_list->is_11n_enabled = 0;
155 ra_list->tdls_link = false;
156 ra_list->ba_status = BA_SETUP_NONE;
157 ra_list->amsdu_in_ampdu = false;
161 ra_list->tdls_link = true;
162 ra_list->is_11n_enabled =
165 ra_list->is_11n_enabled = IS_11N_ENABLED(priv);
171 ra_list->tx_paused = node->tx_pause;
172 ra_list->is_11n_enabled =
174 if (ra_list->is_11n_enabled)
175 ra_list->max_amsdu = node->max_amsdu;
180 ra_list, ra_list->is_11n_enabled);
182 if (ra_list->is_11n_enabled) {
183 ra_list->ba_pkt_count = 0;
184 ra_list->ba_packet_thr =
187 list_add_tail(&ra_list->list,
188 &priv->wmm.tid_tbl_ptr[i].ra_list);
504 struct mwifiex_ra_list_tbl *ra_list)
509 skb_queue_walk_safe(&ra_list->skb_head, skb, tmp) {
510 skb_unlink(skb, &ra_list->skb_head);
525 struct mwifiex_ra_list_tbl *ra_list;
527 list_for_each_entry(ra_list, ra_list_head, list)
528 mwifiex_wmm_del_pkts_in_ralist_node(priv, ra_list);
540 ra_list);
551 struct mwifiex_ra_list_tbl *ra_list, *tmp_node;
556 "info: ra_list: freeing buf for tid %d\n", i);
557 list_for_each_entry_safe(ra_list, tmp_node,
558 &priv->wmm.tid_tbl_ptr[i].ra_list,
560 list_del(&ra_list->list);
561 kfree(ra_list);
564 INIT_LIST_HEAD(&priv->wmm.tid_tbl_ptr[i].ra_list);
630 struct mwifiex_ra_list_tbl *ra_list;
632 list_for_each_entry(ra_list, &priv->wmm.tid_tbl_ptr[tid].ra_list,
634 if (!memcmp(ra_list->ra, ra_addr, ETH_ALEN))
635 return ra_list;
644 struct mwifiex_ra_list_tbl *ra_list;
651 ra_list = mwifiex_wmm_get_ralist_node(priv, i, mac);
652 if (ra_list && ra_list->tx_paused != tx_pause) {
653 pkt_cnt += ra_list->total_pkt_count;
654 ra_list->tx_paused = tx_pause;
657 ra_list->total_pkt_count;
660 ra_list->total_pkt_count;
683 struct mwifiex_ra_list_tbl *ra_list;
690 list_for_each_entry(ra_list, &priv->wmm.tid_tbl_ptr[i].ra_list,
692 if (!memcmp(ra_list->ra, mac, ETH_ALEN))
695 if (ra_list->tx_paused != tx_pause) {
696 pkt_cnt += ra_list->total_pkt_count;
697 ra_list->tx_paused = tx_pause;
700 ra_list->total_pkt_count;
703 ra_list->total_pkt_count;
732 struct mwifiex_ra_list_tbl *ra_list;
734 ra_list = mwifiex_wmm_get_ralist_node(priv, tid, ra_addr);
735 if (ra_list)
736 return ra_list;
749 struct mwifiex_ra_list_tbl *ra_list;
755 ra_list = mwifiex_wmm_get_ralist_node(priv, i, ra_addr);
757 if (!ra_list)
759 mwifiex_wmm_del_pkts_in_ralist_node(priv, ra_list);
760 if (ra_list->tx_paused)
761 priv->wmm.pkts_paused[i] -= ra_list->total_pkt_count;
763 atomic_sub(ra_list->total_pkt_count,
765 list_del(&ra_list->list);
766 kfree(ra_list);
777 struct mwifiex_ra_list_tbl *ra_list, int ptr_index)
781 list_for_each_entry(rlist, &priv->wmm.tid_tbl_ptr[ptr_index].ra_list,
783 if (rlist == ra_list)
817 struct mwifiex_ra_list_tbl *ra_list;
858 ra_list = mwifiex_wmm_get_queue_raptr(priv, tid_down,
867 list_head = priv->wmm.tid_tbl_ptr[tid_down].ra_list;
868 ra_list = list_first_entry_or_null(&list_head,
876 ra_list = mwifiex_wmm_get_queue_raptr(priv, tid_down, ra);
879 if (!ra_list) {
885 skb_queue_tail(&ra_list->skb_head, skb);
887 ra_list->ba_pkt_count++;
888 ra_list->total_pkt_count++;
895 if (ra_list->tx_paused)
1126 list_for_each_entry(ptr, &tid_ptr->ra_list,
1195 list_move(&tid_ptr->ra_list, &ra->list);