/linux-master/include/linux/ |
H A D | kcore.h | 39 extern int __init register_mem_pfn_is_ram(int (*fn)(unsigned long pfn));
|
H A D | memory_hotplug.h | 68 struct page *pfn_to_online_page(unsigned long pfn); 173 extern int mhp_init_memmap_on_memory(unsigned long pfn, unsigned long nr_pages, 175 extern void mhp_deinit_memmap_on_memory(unsigned long pfn, unsigned long nr_pages); 176 extern int online_pages(unsigned long pfn, unsigned long nr_pages, 246 #define pfn_to_online_page(pfn) \ 249 if (pfn_valid(pfn)) \ 250 ___page = pfn_to_page(pfn); \ 368 extern int sparse_add_section(int nid, unsigned long pfn, 371 extern void sparse_remove_section(unsigned long pfn, unsigned long nr_pages,
|
/linux-master/include/xen/interface/hvm/ |
H A D | hvm_op.h | 37 /* Return hvmmem_type_t for the specified pfn. */ 45 uint64_t pfn; member in struct:xen_hvm_get_mem_type
|
/linux-master/arch/x86/include/asm/ |
H A D | memtype.h | 24 extern bool pat_pfn_immune_to_uc_mtrr(unsigned long pfn);
|
H A D | page.h | 72 static __always_inline void *pfn_to_kaddr(unsigned long pfn) argument 74 return __va(pfn << PAGE_SHIFT);
|
/linux-master/include/trace/events/ |
H A D | huge_memory.h | 65 __field(unsigned long, pfn) 75 __entry->pfn = page ? page_to_pfn(page) : -1; 85 __entry->pfn, 125 __field(unsigned long, pfn) 133 __entry->pfn = page ? page_to_pfn(page) : -1; 141 __entry->pfn, 184 __field(unsigned long, pfn) 193 __entry->pfn = page ? page_to_pfn(page) : -1; 202 __entry->pfn,
|
/linux-master/arch/mips/include/asm/ |
H A D | pgtable-32.h | 158 pfn_pte(unsigned long pfn, pgprot_t prot) argument 162 pte.pte_low = (pfn >> _PAGE_PRESENT_SHIFT) | 164 pte.pte_high = (pfn << PFN_PTE_SHIFT) | 174 static inline pte_t pfn_pte(unsigned long pfn, pgprot_t prot) argument 178 pte.pte_high = (pfn << 6) | (pgprot_val(prot) & 0x3f); 188 #define pfn_pte(pfn, prot) __pte(((unsigned long long)(pfn) << PFN_PTE_SHIFT) | pgprot_val(prot)) 189 #define pfn_pmd(pfn, prot) __pmd(((unsigned long long)(pfn) << PFN_PTE_SHIFT) | pgprot_val(prot))
|
/linux-master/arch/microblaze/mm/ |
H A D | init.c | 18 #include <linux/pfn.h> 58 unsigned long pfn; local 60 for (pfn = max_low_pfn; pfn < max_pfn; ++pfn) { 61 struct page *page = pfn_to_page(pfn); 64 if (!memblock_is_reserved(pfn << PAGE_SHIFT)) 138 int page_is_ram(unsigned long pfn) argument 140 return pfn < max_low_pfn;
|
/linux-master/arch/arm64/kvm/hyp/include/nvhe/ |
H A D | gfp.h | 32 int hyp_pool_init(struct hyp_pool *pool, u64 pfn, unsigned int nr_pages,
|
H A D | memory.h | 31 #define hyp_pfn_to_phys(pfn) ((phys_addr_t)((pfn) << PAGE_SHIFT))
|
/linux-master/arch/mips/mm/ |
H A D | ioremap.c | 50 unsigned long offset, pfn, last_pfn; local 83 pfn = PFN_DOWN(phys_addr); 85 if (walk_system_ram_range(pfn, last_pfn - pfn + 1, NULL,
|
/linux-master/arch/x86/xen/ |
H A D | setup.c | 125 unsigned long __ref xen_chk_extra_mem(unsigned long pfn) argument 130 if (pfn >= xen_extra_mem[i].start_pfn && 131 pfn < xen_extra_mem[i].start_pfn + xen_extra_mem[i].n_pfns) 135 return IDENTITY_FRAME(pfn); 143 unsigned long pfn, pfn_s, pfn_e; local 151 for (pfn = pfn_s; pfn < pfn_e; pfn++) 152 set_phys_to_machine(pfn, INVALID_P2M_ENTRY); 157 * Finds the next RAM pfn availabl 218 unsigned long pfn, end; local 249 xen_update_mem_tables(unsigned long pfn, unsigned long mfn) argument 348 unsigned long pfn; local 459 unsigned long mfn_save, pfn; local [all...] |
/linux-master/arch/x86/mm/pat/ |
H A D | set_memory.c | 14 #include <linux/pfn.h> 50 unsigned long pfn; member in struct:cpa_data 252 static bool __cpa_pfn_in_highmap(unsigned long pfn) argument 258 return within_inclusive(pfn, highmap_start_pfn(), highmap_end_pfn()); 263 static bool __cpa_pfn_in_highmap(unsigned long pfn) argument 464 * The .rodata section needs to be read-only. Using the pfn catches all 551 unsigned long pfn, const char *txt) 563 lvltxt[warnlvl], txt, start, end, pfn, (unsigned long long)pgprot_val(prot), 574 unsigned long pfn, unsigned long npg, 591 check_conflict(warnlvl, prot, res, start, end, pfn, "Tex 549 check_conflict(int warnlvl, pgprot_t prot, pgprotval_t val, unsigned long start, unsigned long end, unsigned long pfn, const char *txt) argument 573 static_protections(pgprot_t prot, unsigned long start, unsigned long pfn, unsigned long npg, unsigned long lpsize, int warnlvl) argument 621 verify_rwx(pgprot_t old, pgprot_t new, unsigned long start, unsigned long pfn, unsigned long npg) argument 848 unsigned long numpages, pmask, psize, lpaddr, pfn, old_pfn; local 1005 split_set_pte(struct cpa_data *cpa, pte_t *pte, unsigned long pfn, pgprot_t ref_prot, unsigned long address, unsigned long size) argument 1045 unsigned long lpaddr, lpinc, ref_pfn, pfn, pfninc = 1; local 1107 unsigned long pfn = PFN_DOWN(__pa(address)); local 1612 unsigned long pfn = pte_pfn(old_pte); local 2020 set_mce_nospec(unsigned long pfn) argument 2049 clear_mce_nospec(unsigned long pfn) argument 2416 kernel_map_pages_in_pgd(pgd_t *pgd, u64 pfn, unsigned long address, unsigned numpages, unsigned long page_flags) argument [all...] |
/linux-master/arch/sparc/include/asm/ |
H A D | pgtable_32.h | 258 #define pfn_pte(pfn, prot) mk_pte(pfn_to_page(pfn), prot) 396 * For sparc32&64, the pfn in io_remap_pfn_range() carries <iospace> in 399 #define MK_IOSPACE_PFN(space, pfn) (pfn | (space << (BITS_PER_LONG - 4))) 400 #define GET_IOSPACE(pfn) (pfn >> (BITS_PER_LONG - 4)) 401 #define GET_PFN(pfn) (pfn & 0x0fffffffUL) 407 unsigned long from, unsigned long pfn, 406 io_remap_pfn_range(struct vm_area_struct *vma, unsigned long from, unsigned long pfn, unsigned long size, pgprot_t prot) argument [all...] |
/linux-master/arch/x86/kvm/mmu/ |
H A D | spte.c | 88 static bool kvm_is_mmio_pfn(kvm_pfn_t pfn) argument 90 if (pfn_valid(pfn)) 91 return !is_zero_pfn(pfn) && PageReserved(pfn_to_page(pfn)) && 102 (!pat_enabled() || pat_pfn_immune_to_uc_mtrr(pfn)); 104 return !e820__mapped_raw_any(pfn_to_hpa(pfn), 105 pfn_to_hpa(pfn + 1) - 1, 139 unsigned int pte_access, gfn_t gfn, kvm_pfn_t pfn, 194 kvm_is_mmio_pfn(pfn)); 200 if (shadow_me_value && !kvm_is_mmio_pfn(pfn)) 137 make_spte(struct kvm_vcpu *vcpu, struct kvm_mmu_page *sp, const struct kvm_memory_slot *slot, unsigned int pte_access, gfn_t gfn, kvm_pfn_t pfn, u64 old_spte, bool prefetch, bool can_unsync, bool host_writable, u64 *new_spte) argument [all...] |
/linux-master/arch/powerpc/kvm/ |
H A D | trace_pr.h | 69 __field( u64, pfn ) 78 __entry->pfn = pte->pfn; 87 TP_printk("Map: hvpn=%llx pfn=%llx ea=%lx vp=%llx ra=%lx [%x]", 88 __entry->host_vpn, __entry->pfn, __entry->eaddr, 98 __field( u64, pfn ) 107 __entry->pfn = pte->pfn; 116 TP_printk("Flush: hva=%llx pfn=%llx ea=%lx vp=%llx ra=%lx [%x]", 117 __entry->host_vpn, __entry->pfn, __entr [all...] |
/linux-master/arch/xtensa/kernel/ |
H A D | pci-dma.c | 28 unsigned long pfn = PFN_DOWN(paddr); local 29 struct page *page = pfn_to_page(pfn);
|
/linux-master/arch/arm/include/asm/mach/ |
H A D | pci.h | 68 extern void pci_map_io_early(unsigned long pfn); 70 static inline void pci_map_io_early(unsigned long pfn) {} argument
|
/linux-master/arch/x86/mm/ |
H A D | iomap_32.c | 47 void __iomem *__iomap_local_pfn_prot(unsigned long pfn, pgprot_t prot) argument 63 return (void __force __iomem *)__kmap_local_pfn_prot(pfn, prot);
|
/linux-master/arch/powerpc/include/asm/nohash/64/ |
H A D | pgtable-4k.h | 90 #define remap_4k_pfn(vma, addr, pfn, prot) \ 91 remap_pfn_range((vma), (addr), (pfn), PAGE_SIZE, (prot))
|
/linux-master/arch/arm/mach-spear/ |
H A D | spear3xx.c | 54 .pfn = __phys_to_pfn(SPEAR_ICM1_2_BASE), 59 .pfn = __phys_to_pfn(SPEAR_ICM3_SMI_CTRL_BASE),
|
/linux-master/arch/mips/alchemy/common/ |
H A D | setup.c | 98 unsigned long pfn, unsigned long size, pgprot_t prot) 100 phys_addr_t phys_addr = fixup_bigphys_addr(pfn << PAGE_SHIFT, size); 97 io_remap_pfn_range(struct vm_area_struct *vma, unsigned long vaddr, unsigned long pfn, unsigned long size, pgprot_t prot) argument
|
/linux-master/arch/riscv/kernel/ |
H A D | vdso.c | 64 unsigned long pfn; local 81 pfn = sym_to_pfn(vdso_info->vdso_code_start); 84 vdso_pagelist[i] = pfn_to_page(pfn + i); 128 unsigned long pfn; local 133 pfn = page_to_pfn(timens_page); 135 pfn = sym_to_pfn(vdso_data); 148 pfn = sym_to_pfn(vdso_data); 155 return vmf_insert_pfn(vma, vmf->address, pfn);
|
/linux-master/arch/arm64/include/asm/ |
H A D | kexec.h | 84 extern bool crash_is_nosave(unsigned long pfn); 91 static inline bool crash_is_nosave(unsigned long pfn) {return false; } argument
|
/linux-master/drivers/dax/ |
H A D | device.c | 76 static void dax_set_mapping(struct vm_fault *vmf, pfn_t pfn, argument 92 struct page *page = pfn_to_page(pfn_t_to_pfn(pfn) + i); 108 pfn_t pfn; local 129 pfn = phys_to_pfn_t(phys, PFN_DEV|PFN_MAP); 131 dax_set_mapping(vmf, pfn, fault_size); 133 return vmf_insert_mixed(vmf->vma, vmf->address, pfn); 143 pfn_t pfn; local 172 pfn = phys_to_pfn_t(phys, PFN_DEV|PFN_MAP); 174 dax_set_mapping(vmf, pfn, fault_size); 176 return vmf_insert_pfn_pmd(vmf, pfn, vm 187 pfn_t pfn; local [all...] |