/linux-master/mm/ |
H A D | hwpoison-inject.c | 18 struct page *hpage; local 28 hpage = compound_head(p); 33 shake_page(hpage); 37 if (!PageLRU(hpage) && !PageHuge(p) && !is_free_buddy_page(p)) 45 err = hwpoison_filter(hpage);
|
H A D | khugepaged.c | 1062 static int alloc_charge_hpage(struct page **hpage, struct mm_struct *mm, argument 1071 *hpage = NULL; 1077 *hpage = NULL; 1083 *hpage = folio_page(folio, 0); 1096 struct page *hpage; local 1112 result = alloc_charge_hpage(&hpage, mm, cc); 1211 result = __collapse_huge_page_copy(pte, hpage, pmd, _pmd, 1218 folio = page_folio(hpage); 1227 _pmd = mk_huge_pmd(hpage, vma->vm_page_prot); 1239 hpage 1448 set_huge_pmd(struct vm_area_struct *vma, unsigned long addr, pmd_t *pmdp, struct page *hpage) argument 1800 struct page *hpage; local [all...] |
H A D | memory-failure.c | 1571 int flags, struct page *hpage) 1573 struct folio *folio = page_folio(hpage); 1579 bool mlocked = PageMlocked(hpage); 1587 if (!(PageLRU(hpage) || PageHuge(p))) 1608 mapping = page_mapping(hpage); 1609 if (!(flags & MF_MUST_KILL) && !PageDirty(hpage) && mapping && 1611 if (page_mkclean(hpage)) { 1612 SetPageDirty(hpage); 1627 if (PageHuge(hpage) && !PageAnon(hpage)) { 1570 hwpoison_user_mappings(struct page *p, unsigned long pfn, int flags, struct page *hpage) argument 2200 struct page *hpage; local [all...] |
H A D | hugetlb.c | 2185 struct address_space *hugetlb_page_mapping_lock_write(struct page *hpage) argument 2187 struct address_space *mapping = page_mapping(hpage);
|
/linux-master/fs/verity/ |
H A D | verify.c | 17 * @hpage, has already been verified. 19 static bool is_hash_block_verified(struct fsverity_info *vi, struct page *hpage, argument 35 return PageChecked(hpage); 60 if (PageChecked(hpage)) { 77 SetPageChecked(hpage); 149 struct page *hpage; local 172 hpage = inode->i_sb->s_vop->read_merkle_tree_page(inode, 175 if (IS_ERR(hpage)) { 178 PTR_ERR(hpage), hpage_idx); 181 haddr = kmap_local_page(hpage) 200 struct page *hpage = hblocks[level - 1].page; local [all...] |
/linux-master/include/trace/events/ |
H A D | huge_memory.h | 210 TP_PROTO(struct mm_struct *mm, struct page *hpage, pgoff_t index, 213 TP_ARGS(mm, hpage, index, addr, is_shmem, file, nr, result), 227 __entry->hpfn = hpage ? page_to_pfn(hpage) : -1;
|
/linux-master/io_uring/ |
H A D | rsrc.c | 816 int nr_pages, struct page *hpage) 824 if (compound_head(pages[i]) == hpage) 835 if (compound_head(imu->bvec[j].bv_page) == hpage) 854 struct page *hpage; local 856 hpage = compound_head(pages[i]); 857 if (hpage == *last_hpage) 859 *last_hpage = hpage; 860 if (headpage_already_acct(ctx, pages, i, hpage)) 862 imu->acct_pages += page_size(hpage) >> PAGE_SHIFT; 815 headpage_already_acct(struct io_ring_ctx *ctx, struct page **pages, int nr_pages, struct page *hpage) argument
|
/linux-master/arch/powerpc/kvm/ |
H A D | book3s_pr.c | 648 struct page *hpage; local 653 hpage = gfn_to_page(vcpu->kvm, pte->raddr >> PAGE_SHIFT); 654 if (is_error_page(hpage)) 661 get_page(hpage); 662 page = kmap_atomic(hpage); 670 put_page(hpage);
|
/linux-master/include/linux/ |
H A D | hugetlb.h | 178 struct address_space *hugetlb_page_mapping_lock_write(struct page *hpage); 302 struct page *hpage) 301 hugetlb_page_mapping_lock_write( struct page *hpage) argument
|
/linux-master/arch/s390/kvm/ |
H A D | kvm-s390.c | 200 static int hpage; variable 201 module_param(hpage, int, 0444); 202 MODULE_PARM_DESC(hpage, "1m huge page backing support"); 590 if (hpage && !kvm_is_ucontrol(kvm)) 844 else if (!hpage || kvm->arch.use_cmma || kvm_is_ucontrol(kvm)) 5844 if (nested && hpage) {
|
/linux-master/kernel/trace/ |
H A D | ring_buffer.c | 1943 struct buffer_page *hpage = rb_set_head_page(cpu_buffer); local 1945 if (!hpage) 1947 head_page = &hpage->list;
|