Lines Matching refs:cache
2150 void ieee80211_init_frag_cache(struct ieee80211_fragment_cache *cache)
2154 for (i = 0; i < ARRAY_SIZE(cache->entries); i++)
2155 skb_queue_head_init(&cache->entries[i].skb_list);
2158 void ieee80211_destroy_frag_cache(struct ieee80211_fragment_cache *cache)
2162 for (i = 0; i < ARRAY_SIZE(cache->entries); i++)
2163 __skb_queue_purge(&cache->entries[i].skb_list);
2167 ieee80211_reassemble_add(struct ieee80211_fragment_cache *cache,
2173 entry = &cache->entries[cache->next++];
2174 if (cache->next >= IEEE80211_FRAGMENT_MAX)
2175 cache->next = 0;
2192 ieee80211_reassemble_find(struct ieee80211_fragment_cache *cache,
2199 idx = cache->next;
2208 entry = &cache->entries[idx];
2249 struct ieee80211_fragment_cache *cache = &rx->sdata->frags;
2268 cache = &rx->sta->frags;
2291 entry = ieee80211_reassemble_add(cache, frag, seq,
2323 * fragment cache. Add this fragment to the end of the pending entry.
2325 entry = ieee80211_reassemble_find(cache, frag, seq,
2344 /* Prevent mixed key and fragment cache attacks */
2364 /* Drop this as a mixed key or fragment cache attack, even
2902 /* flush fast xmit cache if the address path changed */