Lines Matching refs:hpage
1062 static int alloc_charge_hpage(struct page **hpage, struct mm_struct *mm,
1071 *hpage = NULL;
1077 *hpage = NULL;
1083 *hpage = folio_page(folio, 0);
1096 struct page *hpage;
1112 result = alloc_charge_hpage(&hpage, mm, cc);
1211 result = __collapse_huge_page_copy(pte, hpage, pmd, _pmd,
1218 folio = page_folio(hpage);
1227 _pmd = mk_huge_pmd(hpage, vma->vm_page_prot);
1239 hpage = NULL;
1245 if (hpage)
1246 put_page(hpage);
1447 /* hpage must be locked, and mmap_lock must be held */
1449 pmd_t *pmdp, struct page *hpage)
1458 VM_BUG_ON(!PageTransHuge(hpage));
1461 if (do_set_pmd(&vmf, hpage))
1464 get_page(hpage);
1571 * page table, but the new page will not be a subpage of hpage.
1800 struct page *hpage;
1814 result = alloc_charge_hpage(&hpage, mm, cc);
1818 __SetPageLocked(hpage);
1820 __SetPageSwapBacked(hpage);
1821 hpage->index = start;
1822 hpage->mapping = mapping;
2057 clear_highpage(hpage + (index % HPAGE_PMD_NR));
2060 if (copy_mc_highpage(hpage + (page->index % HPAGE_PMD_NR), page) > 0) {
2067 clear_highpage(hpage + (index % HPAGE_PMD_NR));
2130 folio = page_folio(hpage);
2144 * Mark hpage as uptodate before inserting it into the page cache so
2212 hpage->mapping = NULL;
2214 unlock_page(hpage);
2215 put_page(hpage);
2218 trace_mm_khugepaged_collapse_file(mm, hpage, index, is_shmem, addr, file, nr, result);