Lines Matching refs:reuse
232 * How many struct page structs need to be reset. When we reuse the head
280 * @reuse: reuse address.
285 unsigned long reuse)
293 BUG_ON(start - reuse != PAGE_SIZE);
295 return vmemmap_remap_range(reuse, end, &walk);
300 * to the page which @reuse is mapped to, then free vmemmap
306 * @reuse: reuse address.
314 unsigned long reuse,
321 .reuse_addr = reuse,
325 int nid = page_to_nid((struct page *)reuse);
348 * - The range [@start, @end) and the range [@reuse, @reuse + PAGE_SIZE)
350 * - The @reuse address is part of the range [@reuse, @end) that we are
352 * - The @reuse address is the first in the complete range.
354 * So we need to make sure that @start and @reuse meet the above rules.
356 BUG_ON(start - reuse != PAGE_SIZE);
358 ret = vmemmap_remap_range(reuse, end, &walk);
360 end = reuse + walk.nr_walked * PAGE_SIZE;
369 .reuse_addr = reuse,
374 vmemmap_remap_range(reuse, end, &walk);
410 * @reuse: reuse address.
416 unsigned long reuse, unsigned long flags)
421 .reuse_addr = reuse,
427 BUG_ON(start - reuse != PAGE_SIZE);
432 return vmemmap_remap_range(reuse, end, &walk);