/linux-master/include/asm-generic/ |
H A D | memory_model.h | 5 #include <linux/pfn.h> 18 #define __pfn_to_page(pfn) (mem_map + ((pfn) - ARCH_PFN_OFFSET)) 23 static inline int pfn_valid(unsigned long pfn) argument 29 return pfn >= pfn_offset && (pfn - pfn_offset) < max_mapnr; 37 #define __pfn_to_page(pfn) (vmemmap + (pfn)) 51 #define __pfn_to_page(pfn) \ 52 ({ unsigned long __pfn = (pfn); \ [all...] |
/linux-master/lib/ |
H A D | devmem_is_allowed.c | 21 int devmem_is_allowed(unsigned long pfn) argument 23 if (iomem_is_exclusive(PFN_PHYS(pfn))) 25 if (!page_is_ram(pfn))
|
/linux-master/include/linux/ |
H A D | pfn_t.h | 8 * PFN_SG_CHAIN - pfn is a pointer to the next scatterlist entry 9 * PFN_SG_LAST - pfn references a page and is the last scatterlist entry 10 * PFN_DEV - pfn is not covered by system memmap by default 11 * PFN_MAP - pfn has a dynamic page mapping established by a device driver 29 static inline pfn_t __pfn_to_pfn_t(unsigned long pfn, u64 flags) argument 31 pfn_t pfn_t = { .val = pfn | (flags & PFN_FLAGS_MASK), }; 36 /* a default pfn to pfn_t conversion assumes that @pfn is pfn_valid() */ 37 static inline pfn_t pfn_to_pfn_t(unsigned long pfn) argument 39 return __pfn_to_pfn_t(pfn, 47 pfn_t_has_page(pfn_t pfn) argument 52 pfn_t_to_pfn(pfn_t pfn) argument 57 pfn_t_to_page(pfn_t pfn) argument 64 pfn_t_to_phys(pfn_t pfn) argument 74 pfn_t_valid(pfn_t pfn) argument 80 pfn_t_pte(pfn_t pfn, pgprot_t pgprot) argument 87 pfn_t_pmd(pfn_t pfn, pgprot_t pgprot) argument 93 pfn_t_pud(pfn_t pfn, pgprot_t pgprot) argument 101 pfn_t_devmap(pfn_t pfn) argument 108 pfn_t_devmap(pfn_t pfn) argument 121 pfn_t_special(pfn_t pfn) argument 126 pfn_t_special(pfn_t pfn) argument [all...] |
H A D | pageblock-flags.h | 56 #define pageblock_align(pfn) ALIGN((pfn), pageblock_nr_pages) 57 #define pageblock_aligned(pfn) IS_ALIGNED((pfn), pageblock_nr_pages) 58 #define pageblock_start_pfn(pfn) ALIGN_DOWN((pfn), pageblock_nr_pages) 59 #define pageblock_end_pfn(pfn) ALIGN((pfn) + 1, pageblock_nr_pages) 65 unsigned long pfn, 70 unsigned long pfn, [all...] |
/linux-master/arch/x86/kernel/ |
H A D | crash_dump_32.c | 15 static inline bool is_crashed_pfn_valid(unsigned long pfn) argument 22 * test checks high 12 bits for being zero (pfn will be shifted left 25 return pte_pfn(pfn_pte(pfn, __pgprot(0))) == pfn; 31 ssize_t copy_oldmem_page(struct iov_iter *iter, unsigned long pfn, size_t csize, argument 39 if (!is_crashed_pfn_valid(pfn)) 42 vaddr = kmap_local_pfn(pfn);
|
H A D | crash_dump_64.c | 15 static ssize_t __copy_oldmem_page(struct iov_iter *iter, unsigned long pfn, argument 25 vaddr = (__force void *)ioremap_encrypted(pfn << PAGE_SHIFT, PAGE_SIZE); 27 vaddr = (__force void *)ioremap_cache(pfn << PAGE_SHIFT, PAGE_SIZE); 38 ssize_t copy_oldmem_page(struct iov_iter *iter, unsigned long pfn, size_t csize, argument 41 return __copy_oldmem_page(iter, pfn, csize, offset, false); 49 ssize_t copy_oldmem_page_encrypted(struct iov_iter *iter, unsigned long pfn, argument 52 return __copy_oldmem_page(iter, pfn, csize, offset, true);
|
/linux-master/arch/powerpc/kernel/ |
H A D | suspend.c | 15 * pfn_is_nosave - check if given pfn is in the 'nosave' section 18 int pfn_is_nosave(unsigned long pfn) argument 22 return (pfn >= nosave_begin_pfn) && (pfn < nosave_end_pfn);
|
/linux-master/arch/xtensa/kernel/ |
H A D | hibernate.c | 7 int pfn_is_nosave(unsigned long pfn) argument 12 return (pfn >= nosave_begin_pfn) && (pfn < nosave_end_pfn);
|
/linux-master/tools/testing/scatterlist/ |
H A D | main.c | 11 unsigned *pfn; member in struct:test 28 #define pfn(...) (unsigned []){ __VA_ARGS__ } macro 42 printf(" %x", test->pfn[i]); 56 { -EINVAL, 1, pfn(0), NULL, PAGE_SIZE, 0, 1 }, 57 { 0, 1, pfn(0), NULL, PAGE_SIZE, PAGE_SIZE + 1, 1 }, 58 { 0, 1, pfn(0), NULL, PAGE_SIZE, sgmax, 1 }, 59 { 0, 1, pfn(0), NULL, 1, sgmax, 1 }, 60 { 0, 2, pfn(0, 1), NULL, 2 * PAGE_SIZE, sgmax, 1 }, 61 { 0, 2, pfn(1, 0), NULL, 2 * PAGE_SIZE, sgmax, 2 }, 62 { 0, 3, pfn( [all...] |
/linux-master/arch/sh/include/asm/ |
H A D | mmzone.h | 11 static inline int pfn_to_nid(unsigned long pfn) argument 16 if (pfn >= node_start_pfn(nid) && pfn <= node_end_pfn(nid)) 22 static inline struct pglist_data *pfn_to_pgdat(unsigned long pfn) argument 24 return NODE_DATA(pfn_to_nid(pfn));
|
/linux-master/include/trace/events/ |
H A D | cma.h | 13 TP_PROTO(const char *name, unsigned long pfn, const struct page *page, 16 TP_ARGS(name, pfn, page, count), 20 __field(unsigned long, pfn) 27 __entry->pfn = pfn; 32 TP_printk("name=%s pfn=0x%lx page=%p count=%lu", 34 __entry->pfn, 65 TP_PROTO(const char *name, unsigned long pfn, const struct page *page, 68 TP_ARGS(name, pfn, page, count, align, errorno), 72 __field(unsigned long, pfn) [all...] |
H A D | ksm.h | 123 * @pfn: page frame number of ksm page 132 TP_PROTO(unsigned long pfn, void *rmap_item, void *mm, int err), 134 TP_ARGS(pfn, rmap_item, mm, err), 137 __field(unsigned long, pfn) 144 __entry->pfn = pfn; 150 TP_printk("ksm pfn %lu rmap_item %p mm %p error %d", 151 __entry->pfn, __entry->rmap_item, __entry->mm, __entry->err) 158 * @pfn: page frame number of ksm page 167 TP_PROTO(void *ksm_page, unsigned long pfn, voi [all...] |
/linux-master/arch/x86/include/asm/xen/ |
H A D | page.h | 8 #include <linux/pfn.h> 56 extern int xen_alloc_p2m_entry(unsigned long pfn); 58 extern unsigned long get_phys_to_machine(unsigned long pfn); 59 extern bool set_phys_to_machine(unsigned long pfn, unsigned long mfn); 60 extern bool __set_phys_to_machine(unsigned long pfn, unsigned long mfn); 133 static inline unsigned long __pfn_to_mfn(unsigned long pfn) argument 137 if (pfn < xen_p2m_size) 138 mfn = xen_p2m_addr[pfn]; 139 else if (unlikely(pfn < xen_max_p2m_pfn)) 140 return get_phys_to_machine(pfn); 150 __pfn_to_mfn(unsigned long pfn) argument 156 pfn_to_mfn(unsigned long pfn) argument 176 phys_to_machine_mapping_valid(unsigned long pfn) argument 186 unsigned long pfn; local 206 unsigned long pfn; local 243 pfn_to_gfn(unsigned long pfn) argument 285 unsigned long pfn; local [all...] |
/linux-master/mm/ |
H A D | page_isolation.c | 38 unsigned long pfn; local 55 for (pfn = start_pfn; pfn < end_pfn; pfn++) { 56 page = pfn_to_page(pfn); 93 pfn += skip_pages - 1; 105 pfn += (1 << buddy_order(page)) - 1; 267 __first_valid_page(unsigned long pfn, unsigned long nr_pages) argument 274 page = pfn_to_online_page(pfn + i); 285 * @boundary_pfn: pageblock-aligned pfn tha 312 unsigned long pfn; local 532 unsigned long pfn; local 586 unsigned long pfn; local 608 __test_page_isolated_in_pageblock(unsigned long pfn, unsigned long end_pfn, int flags) argument 658 unsigned long pfn, flags; local [all...] |
H A D | page_ext.c | 157 unsigned long pfn = page_to_pfn(page); local 171 index = pfn - round_down(node_start_pfn(page_to_nid(page)), 236 unsigned long pfn = page_to_pfn(page); local 237 struct mem_section *section = __pfn_to_section(pfn); 249 return get_entry(page_ext, pfn); 268 static int __meminit init_section_page_ext(unsigned long pfn, int nid) argument 274 section = __pfn_to_section(pfn); 283 * The value stored in section->page_ext is (base - pfn) 295 * The passed "pfn" may not be aligned to SECTION. For the calculation 298 pfn 320 __free_page_ext(unsigned long pfn) argument 342 __invalidate_page_ext(unsigned long pfn) argument 358 unsigned long start, end, pfn; local 390 unsigned long start, end, pfn; local 444 unsigned long pfn; local [all...] |
H A D | page_idle.c | 32 * This function tries to get a user memory page by pfn as described above. 34 static struct folio *page_idle_get_folio(unsigned long pfn) argument 36 struct page *page = pfn_to_online_page(pfn); 120 unsigned long pfn, end_pfn; local 126 pfn = pos * BITS_PER_BYTE; 127 if (pfn >= max_pfn) 130 end_pfn = pfn + count * BITS_PER_BYTE; 134 for (; pfn < end_pfn; pfn++) { 135 bit = pfn 165 unsigned long pfn, end_pfn; local [all...] |
/linux-master/arch/sh/kernel/ |
H A D | swsusp.c | 19 int pfn_is_nosave(unsigned long pfn) argument 24 return (pfn >= begin_pfn) && (pfn < end_pfn);
|
/linux-master/arch/sparc/power/ |
H A D | hibernate.c | 19 * pfn_is_nosave - check if given pfn is in the 'nosave' section 22 int pfn_is_nosave(unsigned long pfn) argument 27 return (pfn >= nosave_begin_pfn) && (pfn < nosave_end_pfn);
|
/linux-master/include/xen/arm/ |
H A D | page.h | 7 #include <linux/pfn.h> 15 #define phys_to_machine_mapping_valid(pfn) (1) 33 * The pseudo-physical frame (pfn) used in all the helpers is always based 43 unsigned long __pfn_to_mfn(unsigned long pfn); 47 static inline unsigned long pfn_to_gfn(unsigned long pfn) argument 49 return pfn; 58 static inline unsigned long pfn_to_bfn(unsigned long pfn) argument 63 mfn = __pfn_to_mfn(pfn); 68 return pfn; 103 bool __set_phys_to_machine(unsigned long pfn, unsigne 107 set_phys_to_machine(unsigned long pfn, unsigned long mfn) argument [all...] |
/linux-master/tools/testing/memblock/linux/ |
H A D | mmzone.h | 25 #define pageblock_align(pfn) ALIGN((pfn), pageblock_nr_pages) 26 #define pageblock_start_pfn(pfn) ALIGN_DOWN((pfn), pageblock_nr_pages)
|
/linux-master/arch/csky/mm/ |
H A D | ioremap.c | 8 pgprot_t phys_mem_access_prot(struct file *file, unsigned long pfn, argument 11 if (!pfn_valid(pfn)) {
|
/linux-master/arch/mips/kernel/ |
H A D | crash_dump.c | 6 ssize_t copy_oldmem_page(struct iov_iter *iter, unsigned long pfn, argument 14 vaddr = kmap_local_pfn(pfn);
|
/linux-master/arch/x86/xen/ |
H A D | p2m.c | 59 * PFN is that we could find ourselves where pfn_to_mfn(pfn)==pfn for a 60 * non-identity pfn. To protect ourselves against we elect to set (and get) the 126 static inline unsigned p2m_top_index(unsigned long pfn) argument 128 BUG_ON(pfn >= MAX_P2M_PFN); 129 return pfn / (P2M_MID_PER_PAGE * P2M_PER_PAGE); 132 static inline unsigned p2m_mid_index(unsigned long pfn) argument 134 return (pfn / P2M_PER_PAGE) % P2M_MID_PER_PAGE; 169 static void p2m_init_identity(unsigned long *p2m, unsigned long pfn) argument 174 p2m[i] = IDENTITY_FRAME(pfn 214 unsigned long pfn, mfn; local 292 unsigned long pfn; local 308 xen_p2m_elem_type(unsigned long pfn) argument 329 unsigned long pfn; local 430 get_phys_to_machine(unsigned long pfn) argument 526 xen_alloc_p2m_entry(unsigned long pfn) argument 628 unsigned long pfn; local 645 __set_phys_to_machine(unsigned long pfn, unsigned long mfn) argument 673 set_phys_to_machine(unsigned long pfn, unsigned long mfn) argument 706 unsigned long mfn, pfn; local 783 unsigned long pfn = page_to_pfn(pages[i]); local 807 unsigned long pfn, first_pfn; local [all...] |
/linux-master/arch/x86/virt/svm/ |
H A D | sev.c | 237 static struct rmpentry *get_rmpentry(u64 pfn) argument 239 if (WARN_ON_ONCE(pfn > rmptable_max_pfn)) 242 return &rmptable[pfn]; 245 static struct rmpentry *__snp_lookup_rmpentry(u64 pfn, int *level) argument 252 entry = get_rmpentry(pfn); 261 large_entry = get_rmpentry(pfn & PFN_PMD_MASK); 270 int snp_lookup_rmpentry(u64 pfn, bool *assigned, int *level) argument 274 e = __snp_lookup_rmpentry(pfn, level); 289 static void dump_rmpentry(u64 pfn) argument 295 e = __snp_lookup_rmpentry(pfn, 361 psmash(u64 pfn) argument 408 adjust_direct_map(u64 pfn, int rmp_level) argument 470 rmpupdate(u64 pfn, struct rmp_state *state) argument 503 rmp_make_private(u64 pfn, u64 gpa, enum pg_level level, u32 asid, bool immutable) argument 519 rmp_make_shared(u64 pfn, enum pg_level level) argument 530 snp_leak_pages(u64 pfn, unsigned int npages) argument [all...] |
/linux-master/arch/mips/power/ |
H A D | cpu.c | 37 int pfn_is_nosave(unsigned long pfn) argument 42 return (pfn >= nosave_begin_pfn) && (pfn < nosave_end_pfn);
|