/linux-master/net/unix/ |
H A D | garbage.c | 163 struct sk_buff *next; local 166 skb_queue_walk_safe(&x->sk_receive_queue, skb, next) { 203 struct sk_buff *next; local 211 skb_queue_walk_safe(&x->sk_receive_queue, skb, next) { 223 u = list_entry(embryos.next, struct unix_sock, link); 257 struct unix_sock *u, *next; local 288 list_for_each_entry_safe(u, next, &gc_inflight_list, link) { 322 while (cursor.next != &gc_candidates) { 323 u = list_entry(cursor.next, struct unix_sock, link); 356 u = list_entry(not_cycle_list.next, struc [all...] |
/linux-master/net/tls/ |
H A D | tls_strp.c | 171 goto next; 183 next: 184 iter = iter->next; 305 last->next = nskb; 442 skb = skb->next;
|
/linux-master/net/openvswitch/ |
H A D | conntrack.c | 1596 struct hlist_node *next; local 1598 hlist_for_each_entry_safe(ct_limit, next, head, hlist_node)
|
/linux-master/net/mac80211/ |
H A D | tx.c | 86 * transmit next fragment plus 2 x ACK plus 3 x SIFS 187 * transmit next fragment plus ACK and 2 x SIFS. */ 189 /* next fragment */ 986 * chain them (using skb as the first fragment) to skb->next. 1090 struct sk_buff *next = skb_queue_next(&tx->skbs, skb); local 1091 next_len = next->len; 1338 struct sk_buff *next; local 1341 skb_list_walk_safe(skb, skb, next) 3478 frag_tail = &(*frag_tail)->next; 3504 * padding to the next on 3764 struct sk_buff *next; local 4277 struct sk_buff *next; local 4584 struct sk_buff *next; local 4606 struct sk_buff *seg, *next; local [all...] |
H A D | rx.c | 1178 /* release the buffer until next missing frame */ 2173 entry = &cache->entries[cache->next++]; 2174 if (cache->next >= IEEE80211_FRAGMENT_MAX) 2175 cache->next = 0; 2199 idx = cache->next; 2218 * Check ftype and addresses are equal, else check next fragment 2297 * next fragment has a sequential PN value. 2985 /* unable to resolve next hop */ 3821 /* will return in the next handlers */ 3847 /* skip known-bad action frames and return them in the next handle [all...] |
H A D | mlme.c | 2256 goto next; 2268 next: 2658 /* downgrade next lower non-ACM AC */ 3192 /* on the next assoc, re-program HT/VHT parameters */ 4534 * next beacon and update then.
|
/linux-master/net/ipv4/ |
H A D | tcp_ao.c | 1071 struct hlist_node *next; local 1094 hlist_for_each_entry_safe(key, next, &ao_info->head, node) { 1759 * Free the key with next RCU grace period (in case it was
|
H A D | route.c | 226 .next = rt_cache_seq_next, 304 .next = rt_cpu_seq_next, 3043 goto next; 3046 goto next; 3050 goto next; 3056 goto next; 3064 next:
|
/linux-master/net/bluetooth/ |
H A D | l2cap_core.c | 906 /* Get next available identificator. 2208 frag = &(*frag)->next; 2969 frag = &(*frag)->next; 5615 new_frag->next = NULL; 5617 (*last_frag)->next = new_frag; 7174 /* Find the next fixed channel in BT_LISTEN state, continue iteration 7188 c = list_entry(chan_list.next, typeof(*c), global_l); 7244 struct l2cap_chan *chan, *next; local 7248 goto next; 7262 next [all...] |
H A D | hci_sync.c | 475 /* Get the next instance to advertise BEFORE we remove 937 * address be updated at the next cycle. 1949 struct adv_info *next = NULL; local 1956 /* Get the next instance to advertise BEFORE we remove 1961 next = hci_get_next_instance(hdev, instance); 1972 if (next && next->instance == instance) 1973 next = NULL; 1984 if (next && !ext_adv_capable(hdev)) 1985 hci_schedule_adv_instance_sync(hdev, next [all...] |
/linux-master/net/ax25/ |
H A D | af_ax25.c | 114 * list meanwhile and thus the next pointer is 1994 .next = ax25_info_next,
|
/linux-master/mm/ |
H A D | hugetlb.c | 1465 * next node from which to allocate, handling wrap at end of node 1484 * next node id whether or not we find a free huge page to free so 1485 * that the next attempt to free addresses the next node. 1828 node = node->next; 2357 * Remove huge page from pool from next node to free. Attempt to keep 2377 folio = list_entry(h->hugepage_freelists[node].next, 3301 /* allocate from next node when distributing huge pages */ 3706 struct folio *folio, *next; local 3708 list_for_each_entry_safe(folio, next, free [all...] |
/linux-master/kernel/ |
H A D | vmcore_info.c | 191 VMCOREINFO_OFFSET(list_head, next);
|
/linux-master/include/net/ |
H A D | sock.h | 143 * @skc_tw_rcv_nxt: (aka tw_rcv_nxt) TCP window next expected seq number 724 return hlist_entry_safe(sk->sk_node.next, struct sock, sk_node); 729 return (!is_a_nulls(sk->sk_nulls_node.next)) ? 730 hlist_nulls_entry(sk->sk_nulls_node.next, 1058 sk->sk_backlog.tail->next = skb; 1061 skb->next = NULL; 1186 * caches using SLAB_TYPESAFE_BY_RCU should let .next pointer from nulls nodes 1191 if (offsetof(struct sock, sk_node.next) != 0) 1192 memset(sk, 0, offsetof(struct sock, sk_node.next)); 2728 * @tskey: filled in with next sk_tske [all...] |
H A D | macsec.h | 121 * @next_pn: packet number expected for the next packet 160 struct macsec_rx_sc __rcu *next; member in struct:macsec_rx_sc 172 * @next_pn: packet number to use for the next packet 257 * @update_pn: when updating the SA, update the next PN
|
/linux-master/include/linux/ |
H A D | mm.h | 221 /* to align the pointer to the (next) page boundary */ 2118 * folio_next - Move to the next physical folio. 2123 * folio to the next. Do not use it if the memory is only virtually 2129 * Return: The next struct folio. 2271 * lru.next has bit 1 set if the page is allocated from the 2275 return (uintptr_t)page->lru.next & BIT(1); 2286 * lru.next has bit 1 set if the page is allocated from the 2290 return (uintptr_t)folio->lru.next & BIT(1); 2299 page->lru.next = (void *)BIT(1); 2304 page->lru.next [all...] |
/linux-master/fs/ntfs3/ |
H A D | file.c | 134 goto next; 156 next:
|
/linux-master/fs/netfs/ |
H A D | buffered_write.c | 569 pgoff_t last = (start + len - 1) / PAGE_SIZE, next; local 578 next = index + 1; 582 next = folio_next_index(folio); 594 } while (index = next, index <= last); 607 pgoff_t last = (start + len - 1) / PAGE_SIZE, next; local 616 next = index + 1; 620 next = folio_next_index(folio); 627 } while (index = next, index <= last);
|
/linux-master/drivers/video/fbdev/core/ |
H A D | fb_defio.c | 241 struct fb_deferred_io_pageref *pageref, *next; local 257 list_for_each_entry_safe(pageref, next, &fbdefio->pagereflist, list)
|
/linux-master/drivers/net/ |
H A D | macsec.c | 56 sc = rcu_dereference_bh(sc->next)) 60 sc = rtnl_dereference(sc->next)) 1431 rx_scp = &rx_sc->next, rx_sc = rtnl_dereference(*rx_scp)) { 1435 rcu_assign_pointer(*rx_scp, rx_sc->next); 1472 rcu_assign_pointer(rx_sc->next, secy->rx_sc); 3344 goto next; 3347 goto next; 3352 next: 3991 rcu_assign_pointer(macsec->secy.rx_sc, rx_sc->next);
|
H A D | gtp.c | 456 /* seq, npdu and next should be counted to the length of the GTP packet 1101 struct hlist_node *next; local 1106 hlist_for_each_entry_safe(pctx, next, >p->tid_hash[i], hlist_tid)
|
/linux-master/drivers/net/ethernet/wangxun/libwx/ |
H A D | wx_lib.c | 296 /* move offset up to the next cache line */ 375 /* update next to alloc since we have filled the ring */ 402 * This function updates next to clean. If the buffer is an EOP buffer 404 * sk_buff in the next buffer to be chained and return true indicating 413 /* fetch, update, and store next to clean */ 775 /* move us one more past the eop_desc for start of next pkt */ 785 /* issue prefetch for next Tx descriptor */ 884 /* For the next check */ 1467 * otherwise try next time 1713 ring->next [all...] |
/linux-master/drivers/net/ethernet/ti/ |
H A D | am65-cpts.c | 244 struct list_head *this, *next; local 248 list_for_each_safe(this, next, &cpts->events) { 827 struct list_head *this, *next; local 836 list_for_each_safe(this, next, &events) {
|
/linux-master/drivers/net/ethernet/mellanox/mlxsw/ |
H A D | spectrum_acl_tcam.c | 218 /* Check if the next entry in the list has the same vregion. */ 219 if (region->list.next != &group->region_list && 400 /* If the next region is defined, place the new one 401 * before it. The next one is a sibling. 500 * under some of the next vregions. 502 if (pos->next != &vgroup->vregion_list) { /* not last */ 503 vregion2 = list_entry(pos->next, typeof(*vregion2), 524 * use this vregion. There is still some hope that the next
|
/linux-master/drivers/net/ethernet/marvell/octeontx2/af/ |
H A D | rvu_npc.c | 1676 /* Profile image offsets are rounded up to next 8 multiple.*/ 2377 u16 index, next, end; local 2387 next = find_next_bit(map, end, index); 2388 if (*max_area < (next - index)) { 2389 *max_area = next - index; 2393 if (next < end) { 2394 start = next + 1; 2406 u16 index, next; local 2417 next = find_next_bit(map, end, index); 2418 if (next < [all...] |