Lines Matching refs:next
1465 * next node from which to allocate, handling wrap at end of node
1484 * next node id whether or not we find a free huge page to free so
1485 * that the next attempt to free addresses the next node.
1828 node = node->next;
2357 * Remove huge page from pool from next node to free. Attempt to keep
2377 folio = list_entry(h->hugepage_freelists[node].next,
3301 /* allocate from next node when distributing huge pages */
3706 struct folio *folio, *next;
3708 list_for_each_entry_safe(folio, next, freel, lru) {
4753 /* Go to parse next node*/
6621 * here we just wait to defer the next page fault to avoid busy loop and