Lines Matching defs:page

282 void __reset_page_owner(struct page *page, unsigned short order)
290 page_ext = page_ext_get(page);
313 noinline void __set_page_owner(struct page *page, unsigned short order,
322 page_ext = page_ext_get(page);
332 void __set_page_owner_migrate_reason(struct page *page, int reason)
334 struct page_ext *page_ext = page_ext_get(page);
345 void __split_page_owner(struct page *page, int old_order, int new_order)
348 struct page_ext *page_ext = page_ext_get(page);
371 old_ext = page_ext_get(&old->page);
375 new_ext = page_ext_get(&newfolio->page);
416 struct page *page;
429 * Walk the zone in pageblock_nr_pages steps. If a page block spans
434 page = pfn_to_online_page(pfn);
435 if (!page) {
443 pageblock_mt = get_pageblock_migratetype(page);
447 page = pfn_to_page(pfn);
449 if (page_zone(page) != zone)
452 if (PageBuddy(page)) {
455 freepage_order = buddy_order_unsafe(page);
461 if (PageReserved(page))
464 page_ext = page_ext_get(page);
500 struct page *page)
509 memcg_data = READ_ONCE(page->memcg_data);
515 "Slab cache page\n");
517 memcg = page_memcg_check(page);
525 PageMemcgKmem(page) ? "(via objcg) " : "",
537 struct page *page, struct page_owner *page_owner,
556 pageblock_mt = get_pageblock_migratetype(page);
564 &page->flags);
576 ret = print_page_owner_memcg(kbuf, count, ret, page);
593 void __dump_page_owner(const struct page *page)
595 struct page_ext *page_ext = page_ext_get((void *)page);
602 pr_alert("There is not page extension available.\n");
617 pr_alert("page_owner tracks the page as allocated\n");
619 pr_alert("page_owner tracks the page as freed\n");
621 pr_alert("page last allocated via order %u, migratetype %s, gfp_mask %#x(%pGg), pid %d, tgid %d (%s), ts %llu, free_ts %llu\n",
636 pr_alert("page last free pid %d tgid %d stack trace:\n",
642 pr_alert("page has been migrated, last migrate reason: %s\n",
651 struct page *page;
659 page = NULL;
668 /* Find an allocated page */
673 * the rcu lock and copying the page owner information to
679 * If the new page is in a new MAX_ORDER_NR_PAGES area,
687 page = pfn_to_page(pfn);
688 if (PageBuddy(page)) {
689 unsigned long freepage_order = buddy_order_unsafe(page);
696 page_ext = page_ext_get(page);
736 return print_page_owner(buf, count, pfn, page,
767 * Walk the zone in pageblock_nr_pages steps. If a page block spans
783 struct page *page = pfn_to_page(pfn);
786 if (page_zone(page) != zone)
791 * careful when reading buddy page order. The only
796 if (PageBuddy(page)) {
797 unsigned long order = buddy_order_unsafe(page);
804 if (PageReserved(page))
807 page_ext = page_ext_get(page);
815 /* Found early allocated page */
826 pr_info("Node %d, zone %8s: page owner found early allocated %lu pages\n",