/linux-master/arch/arc/include/asm/ |
H A D | pgtable.h | 25 #define ZERO_PAGE(vaddr) (virt_to_page(empty_zero_page))
|
H A D | pgtable-levels.h | 112 #define p4d_page(p4d) virt_to_page(p4d_pgtable(p4d)) 134 #define pud_page(pud) virt_to_page(pud_pgtable(pud)) 162 #define pmd_page(pmd) virt_to_page((void *)pmd_page_vaddr(pmd))
|
/linux-master/arch/parisc/mm/ |
H A D | ioremap.c | 37 for (page = virt_to_page(t_addr); 38 page <= virt_to_page(t_end); page++) {
|
/linux-master/arch/arm/include/asm/vdso/ |
H A D | vsyscall.h | 27 flush_dcache_page(virt_to_page(vdata));
|
/linux-master/arch/s390/include/asm/ |
H A D | kfence.h | 38 __kernel_map_pages(virt_to_page((void *)addr), 1, !protect);
|
/linux-master/lib/ |
H A D | test_free_pages.c | 20 struct page *page = virt_to_page((void *)addr);
|
/linux-master/arch/powerpc/include/asm/ |
H A D | kfence.h | 26 struct page *page = virt_to_page((void *)addr);
|
/linux-master/arch/m68k/include/asm/ |
H A D | pgtable_no.h | 38 #define ZERO_PAGE(vaddr) (virt_to_page(empty_zero_page))
|
H A D | page_no.h | 32 #define virt_to_page(addr) (mem_map + (((unsigned long)(addr)-PAGE_OFFSET) >> PAGE_SHIFT)) macro
|
/linux-master/mm/ |
H A D | bootmem_info.c | 60 page = virt_to_page(memmap); 69 page = virt_to_page(usage); 93 page = virt_to_page(usage); 109 page = virt_to_page(pgdat);
|
H A D | ptdump.c | 37 if (pgd_page(val) == virt_to_page(lm_alias(kasan_early_shadow_p4d))) 60 if (p4d_page(val) == virt_to_page(lm_alias(kasan_early_shadow_pud))) 83 if (pud_page(val) == virt_to_page(lm_alias(kasan_early_shadow_pmd))) 105 if (pmd_page(val) == virt_to_page(lm_alias(kasan_early_shadow_pte)))
|
H A D | percpu-km.c | 104 return virt_to_page(addr);
|
/linux-master/arch/powerpc/mm/ |
H A D | pgtable_64.c | 108 return virt_to_page(p4d_pgtable(p4d)); 119 return virt_to_page(pud_pgtable(pud)); 138 return virt_to_page(pmd_page_vaddr(pmd));
|
/linux-master/arch/loongarch/include/asm/ |
H A D | page.h | 90 #define virt_to_page(kaddr) phys_to_page(__pa(kaddr)) macro 102 #define virt_to_page(kaddr) \ macro 111 #define virt_to_pfn(kaddr) page_to_pfn(virt_to_page(kaddr))
|
H A D | kfence.h | 42 set_page_address(virt_to_page(kaddr), vaddr);
|
/linux-master/arch/xtensa/mm/ |
H A D | kasan_init.c | 26 mk_pte(virt_to_page(kasan_early_shadow_page), 93 mk_pte(virt_to_page(kasan_early_shadow_page),
|
/linux-master/fs/ubifs/ |
H A D | crypto.c | 53 err = fscrypt_encrypt_block_inplace(inode, virt_to_page(p), pad_len, 78 err = fscrypt_decrypt_block_inplace(inode, virt_to_page(&dn->data),
|
/linux-master/arch/csky/kernel/ |
H A D | vdso.c | 36 pg = virt_to_page(vdso_start + (i << PAGE_SHIFT)); 39 vdso_pagelist[i] = virt_to_page(vdso_data);
|
/linux-master/arch/powerpc/mm/kasan/ |
H A D | init_book3e_64.c | 19 return p4d_page(p4d) == virt_to_page(lm_alias(kasan_early_shadow_pud)); 24 return pud_page(pud) == virt_to_page(lm_alias(kasan_early_shadow_pmd)); 29 return pmd_page(pmd) == virt_to_page(lm_alias(kasan_early_shadow_pte));
|
/linux-master/arch/sh/kernel/vsyscall/ |
H A D | vsyscall.c | 43 syscall_pages[0] = virt_to_page(syscall_page);
|
/linux-master/arch/x86/include/asm/ |
H A D | page.h | 65 * virt_to_page(kaddr) returns a valid pointer if and only if 68 #define virt_to_page(kaddr) pfn_to_page(__pa(kaddr) >> PAGE_SHIFT) macro
|
/linux-master/arch/openrisc/include/asm/ |
H A D | page.h | 80 #define virt_to_page(addr) \ macro
|
/linux-master/arch/nios2/include/asm/ |
H A D | page.h | 89 # define virt_to_page(vaddr) pfn_to_page(PFN_DOWN(virt_to_phys(vaddr))) macro
|
/linux-master/arch/nios2/mm/ |
H A D | ioremap.c | 139 for (page = virt_to_page(t_addr); 140 page <= virt_to_page(t_end); page++)
|
/linux-master/include/asm-generic/ |
H A D | page.h | 88 #define virt_to_page(addr) pfn_to_page(virt_to_pfn(addr)) macro
|