Lines Matching refs:page
28 * the page allocator becomes available.
97 struct page *shadow, *origin;
114 bool kmsan_memblock_free_pages(struct page *page, unsigned int order)
116 struct page *shadow, *origin;
119 held_back[order].shadow = page;
123 held_back[order].origin = page;
128 kmsan_setup_meta(page, shadow, origin, order);
137 struct page *items[MAX_BLOCKS];
147 static void smallstack_push(struct smallstack *stack, struct page *pages)
155 static struct page *smallstack_pop(struct smallstack *stack)
157 struct page *ret;
168 struct page *page, *shadow, *origin;
171 page = smallstack_pop(&collect);
174 kmsan_setup_meta(page, shadow, origin, collect.order);
175 __free_pages_core(page, collect.order);
185 struct page *page;
190 page = smallstack_pop(&collect);
191 smallstack_push(&tmp, &page[0]);
192 smallstack_push(&tmp, &page[1 << tmp.order]);
198 * Memblock is about to go away. Split the page blocks left over in held_back[]