/linux-master/arch/arm/include/asm/mach/ |
H A D | map.h | 16 unsigned long pfn; member in struct:map_desc
|
/linux-master/arch/arm/mach-spear/ |
H A D | spear1310.c | 43 .pfn = __phys_to_pfn(SPEAR1310_RAS_GRP1_BASE),
|
/linux-master/arch/arm/mach-gemini/ |
H A D | board-dt.c | 19 .pfn = __phys_to_pfn(CONFIG_DEBUG_UART_PHYS),
|
/linux-master/arch/arm64/include/asm/ |
H A D | page.h | 44 int pfn_is_map_memory(unsigned long pfn);
|
/linux-master/arch/sh/mm/ |
H A D | cache-sh7705.c | 143 unsigned long pfn = folio_pfn(folio); local 147 __flush_dcache_page((pfn + i) * PAGE_SIZE); 171 unsigned long pfn = data->addr2; local 173 __flush_dcache_page(pfn << PAGE_SHIFT);
|
/linux-master/drivers/gpu/drm/nouveau/nvkm/subdev/mmu/ |
H A D | vmmgp100.c | 82 for (; ptes; ptes--, map->pfn++) { 85 if (!(*map->pfn & NVKM_VMM_PFN_V)) 88 if (!(*map->pfn & NVKM_VMM_PFN_W)) 91 if (!(*map->pfn & NVKM_VMM_PFN_A)) 94 if (!(*map->pfn & NVKM_VMM_PFN_VRAM)) { 95 addr = *map->pfn >> NVKM_VMM_PFN_ADDR_SHIFT; 105 data |= (*map->pfn & NVKM_VMM_PFN_ADDR) >> 4; 176 .pfn = gp100_vmm_pgt_pfn, 319 for (; ptes; ptes--, map->pfn++) { 322 if (!(*map->pfn [all...] |
/linux-master/arch/arc/include/asm/ |
H A D | pgtable-levels.h | 144 #define pfn_pmd(pfn,prot) __pmd(((pfn) << PAGE_SHIFT) | pgprot_val(prot)) 179 #define pfn_pte(pfn, prot) __pte(__pfn_to_phys(pfn) | pgprot_val(prot))
|
H A D | hugepage.h | 47 #define pfn_pmd(pfn, prot) (__pmd(((pfn) << PAGE_SHIFT) | pgprot_val(prot)))
|
/linux-master/arch/s390/kernel/ |
H A D | crash_dump.c | 166 ssize_t copy_oldmem_page(struct iov_iter *iter, unsigned long pfn, size_t csize, argument 171 src = pfn_to_phys(pfn) + offset; 182 unsigned long from, unsigned long pfn, 188 if (pfn < oldmem_data.size >> PAGE_SHIFT) { 189 size_old = min(size, oldmem_data.size - (pfn << PAGE_SHIFT)); 191 pfn + (oldmem_data.start >> PAGE_SHIFT), 197 pfn += size_old >> PAGE_SHIFT; 199 return remap_pfn_range(vma, from, pfn, size, prot); 210 unsigned long pfn, 216 if (pfn < hsa_en 181 remap_oldmem_pfn_range_kdump(struct vm_area_struct *vma, unsigned long from, unsigned long pfn, unsigned long size, pgprot_t prot) argument 208 remap_oldmem_pfn_range_zfcpdump(struct vm_area_struct *vma, unsigned long from, unsigned long pfn, unsigned long size, pgprot_t prot) argument 230 remap_oldmem_pfn_range(struct vm_area_struct *vma, unsigned long from, unsigned long pfn, unsigned long size, pgprot_t prot) argument [all...] |
H A D | vdso.c | 73 unsigned long addr, pfn; local 78 pfn = virt_to_pfn(vdso_data); 85 err = vmf_insert_pfn(vma, addr, pfn); 88 pfn = page_to_pfn(timens_page); 102 pfn = virt_to_pfn(vdso_data); 108 return vmf_insert_pfn(vma, vmf->address, pfn);
|
/linux-master/tools/testing/selftests/mm/ |
H A D | transhuge-stress.c | 99 int64_t pfn; local 101 pfn = allocate_transhuge(p, pagemap_fd); 103 if (pfn < 0) { 106 size_t idx = pfn >> (HPAGE_SHIFT - pshift());
|
/linux-master/arch/x86/xen/ |
H A D | enlighten_pvh.c | 33 u64 pfn; local 40 pfn = __pa(hypercall_page); 41 wrmsr_safe(msr, (u32)pfn, (u32)(pfn >> 32));
|
/linux-master/drivers/virtio/ |
H A D | virtio_mem.c | 274 static void virtio_mem_fake_offline_going_offline(unsigned long pfn, 276 static void virtio_mem_fake_offline_cancel_offline(unsigned long pfn, 924 unsigned long pfn; local 930 pfn = PFN_DOWN(virtio_mem_mb_id_to_phys(mb_id) + 932 virtio_mem_fake_offline_going_offline(pfn, nr_pages); 940 unsigned long pfn; local 946 pfn = PFN_DOWN(virtio_mem_mb_id_to_phys(mb_id) + 948 virtio_mem_fake_offline_cancel_offline(pfn, nr_pages); 954 unsigned long pfn, 964 virtio_mem_fake_offline_going_offline(pfn, nr_page 952 virtio_mem_bbm_notify_going_offline(struct virtio_mem *vm, unsigned long bb_id, unsigned long pfn, unsigned long nr_pages) argument 967 virtio_mem_bbm_notify_cancel_offline(struct virtio_mem *vm, unsigned long bb_id, unsigned long pfn, unsigned long nr_pages) argument 1118 virtio_mem_set_fake_offline(unsigned long pfn, unsigned long nr_pages, bool onlined) argument 1139 virtio_mem_clear_fake_offline(unsigned long pfn, unsigned long nr_pages, bool onlined) argument 1155 virtio_mem_fake_online(unsigned long pfn, unsigned long nr_pages) argument 1192 virtio_mem_fake_offline(struct virtio_mem *vm, unsigned long pfn, unsigned long nr_pages) argument 1236 virtio_mem_fake_offline_going_offline(unsigned long pfn, unsigned long nr_pages) argument 1260 virtio_mem_fake_offline_cancel_offline(unsigned long pfn, unsigned long nr_pages) argument 1698 unsigned long pfn, nr_pages; local 2139 unsigned long pfn; local 2204 unsigned long pfn; local 2224 unsigned long pfn; local 2661 virtio_mem_vmcore_pfn_is_ram(struct vmcore_cb *cb, unsigned long pfn) argument [all...] |
/linux-master/include/asm-generic/ |
H A D | page.h | 82 static inline void *pfn_to_virt(unsigned long pfn) argument 84 return __va(pfn) << PAGE_SHIFT;
|
/linux-master/arch/arm64/kernel/ |
H A D | vdso.c | 87 unsigned long pfn; local 106 pfn = sym_to_pfn(vdso_info[abi].vdso_code_start); 109 vdso_pagelist[i] = pfn_to_page(pfn + i); 155 unsigned long pfn; local 160 pfn = page_to_pfn(timens_page); 162 pfn = sym_to_pfn(vdso_data); 175 pfn = sym_to_pfn(vdso_data); 182 return vmf_insert_pfn(vma, vmf->address, pfn);
|
/linux-master/arch/arm64/mm/ |
H A D | mmap.c | 60 int valid_mmap_phys_addr_range(unsigned long pfn, size_t size) argument 62 return !(((pfn << PAGE_SHIFT) + size) & ~PHYS_MASK);
|
/linux-master/arch/x86/mm/pat/ |
H A D | cpa-test.c | 139 unsigned long pfn = get_random_u32_below(max_pfn_mapped); local 141 addr[i] = (unsigned long)__va(pfn << PAGE_SHIFT); 143 len[i] = min_t(unsigned long, len[i], max_pfn_mapped - pfn - 1); 167 if (test_bit(pfn + k, bm)) { 171 __set_bit(pfn + k, bm); 173 pages[k] = pfn_to_page(pfn + k);
|
/linux-master/mm/ |
H A D | page_table_check.c | 61 static void page_table_check_clear(unsigned long pfn, unsigned long pgcnt) argument 68 if (!pfn_valid(pfn)) 71 page = pfn_to_page(pfn); 97 static void page_table_check_set(unsigned long pfn, unsigned long pgcnt, argument 105 if (!pfn_valid(pfn)) 108 page = pfn_to_page(pfn);
|
H A D | page_alloc.c | 346 unsigned long pfn) 349 return section_to_usemap(__pfn_to_section(pfn)); 355 static inline int pfn_to_bitidx(const struct page *page, unsigned long pfn) argument 358 pfn &= (PAGES_PER_SECTION-1); 360 pfn = pfn - pageblock_start_pfn(page_zone(page)->zone_start_pfn); 362 return (pfn >> pageblock_order) * NR_PAGEBLOCK_BITS; 368 * @pfn: The target page frame number 374 unsigned long pfn, unsigned long mask) 380 bitmap = get_pageblock_bitmap(page, pfn); 345 get_pageblock_bitmap(const struct page *page, unsigned long pfn) argument 373 get_pfnblock_flags_mask(const struct page *page, unsigned long pfn, unsigned long mask) argument 393 get_pfnblock_migratetype(const struct page *page, unsigned long pfn) argument 406 set_pfnblock_flags_mask(struct page *page, unsigned long flags, unsigned long pfn, unsigned long mask) argument 447 unsigned long pfn = page_to_pfn(page); local 725 buddy_merge_likely(unsigned long pfn, unsigned long buddy_pfn, struct page *page, unsigned int order) argument 765 __free_one_page(struct page *page, unsigned long pfn, struct zone *zone, unsigned int order, int migratetype, fpi_t fpi_flags) argument 864 unsigned long pfn; local 892 __free_one_page(pfn_to_page(pfn), pfn, zone, free_page_order, local 1247 free_one_page(struct zone *zone, struct page *page, unsigned long pfn, unsigned int order, int migratetype, fpi_t fpi_flags) argument 1267 unsigned long pfn = page_to_pfn(page); local 1622 unsigned long pfn; local 1657 unsigned long start_pfn, end_pfn, pfn; local 2342 free_unref_page_prepare(struct page *page, unsigned long pfn, unsigned int order) argument 2484 unsigned long pfn = page_to_pfn(page); local 2500 free_one_page(page_zone(page), page, pfn, order, migratetype, FPI_NONE); local 2531 unsigned long pfn = folio_pfn(folio); local 2546 free_one_page(folio_zone(folio), &folio->page, pfn, local 6248 unsigned long pfn = start; local 6529 unsigned long ret, pfn, flags; local 6564 free_contig_range(unsigned long pfn, unsigned long nr_pages) argument 6626 unsigned long pfn = start_pfn; local 6671 unsigned long pfn = page_to_pfn(page); local 6723 unsigned long pfn = page_to_pfn(page); local 6760 unsigned long pfn = page_to_pfn(page); local [all...] |
/linux-master/arch/mips/mm/ |
H A D | cache.c | 36 unsigned long pfn); 149 unsigned long pfn, addr; local 153 pfn = pte_pfn(pte); 154 if (unlikely(!pfn_valid(pfn))) 157 folio = page_folio(pfn_to_page(pfn)); 159 address -= offset_in_folio(folio, pfn << PAGE_SHIFT);
|
/linux-master/arch/arm/include/asm/ |
H A D | cacheflush.h | 235 unsigned long user_addr, unsigned long pfn, unsigned int nr) 251 #define flush_cache_pages(vma, addr, pfn, nr) \ 252 vivt_flush_cache_pages(vma, addr, pfn, nr) 257 unsigned long pfn, unsigned int nr); 261 #define flush_cache_page(vma, addr, pfn) flush_cache_pages(vma, addr, pfn, 1) 234 vivt_flush_cache_pages(struct vm_area_struct *vma, unsigned long user_addr, unsigned long pfn, unsigned int nr) argument
|
/linux-master/fs/proc/ |
H A D | kcore.c | 61 static int (*mem_pfn_is_ram)(unsigned long pfn); 63 int __init register_mem_pfn_is_ram(int (*fn)(unsigned long pfn)) argument 71 static int pfn_is_ram(unsigned long pfn) argument 74 return mem_pfn_is_ram(pfn); 142 /* calculate vmemmap's address from given system ram pfn and register it */ 146 unsigned long pfn = __pa(ent->addr) >> PAGE_SHIFT; local 152 start = ((unsigned long)pfn_to_page(pfn)) & PAGE_MASK; 153 end = ((unsigned long)pfn_to_page(pfn + nr_pages)) - 1; 185 kclist_add_private(unsigned long pfn, unsigned long nr_pages, void *arg) argument 191 if (!pfn_valid(pfn)) 473 unsigned long pfn; local [all...] |
/linux-master/fs/ |
H A D | dax.c | 74 static void *dax_make_entry(pfn_t pfn, unsigned long flags) argument 76 return xa_mk_value(flags | (pfn_t_to_pfn(pfn) << DAX_SHIFT)); 319 #define for_each_mapped_pfn(entry, pfn) \ 320 for (pfn = dax_to_pfn(entry); \ 321 pfn < dax_end_pfn(entry); pfn++) 359 unsigned long size = dax_entry_size(entry), pfn, index; local 366 for_each_mapped_pfn(entry, pfn) { 367 struct page *page = pfn_to_page(pfn); 382 unsigned long pfn; local 404 unsigned long pfn; local 869 dax_insert_entry(struct xa_state *xas, struct vm_fault *vmf, const struct iomap_iter *iter, void *entry, pfn_t pfn, unsigned long flags) argument 929 unsigned long pfn, index, count, end; local 1191 pfn_t pfn = pfn_to_pfn_t(my_zero_pfn(vaddr)); local 1213 pfn_t pfn; local 1595 dax_fault_synchronous_pfnp(pfn_t *pfnp, pfn_t pfn) argument 1653 pfn_t pfn; local 1929 dax_insert_pfn_mkwrite(struct vm_fault *vmf, pfn_t pfn, unsigned int order) argument 1973 dax_finish_sync_fault(struct vm_fault *vmf, unsigned int order, pfn_t pfn) argument [all...] |
/linux-master/arch/powerpc/kvm/ |
H A D | book3s_64_mmu_host.c | 90 unsigned long pfn; local 97 pfn = kvmppc_gpa_to_pfn(vcpu, orig_pte->raddr, iswrite, &writable); 98 if (is_error_noslot_pfn(pfn)) { 104 hpaddr = pfn << PAGE_SHIFT; 124 kvm_set_pfn_accessed(pfn); 129 kvm_set_pfn_dirty(pfn); 135 kvmppc_mmu_flush_icache(pfn); 197 cpte->pfn = pfn; 206 kvm_release_pfn_clean(pfn); [all...] |
/linux-master/arch/powerpc/lib/ |
H A D | code-patching.c | 239 unsigned long pfn = get_patch_pfn(addr); local 241 return map_kernel_page(text_poke_addr, (pfn << PAGE_SHIFT), PAGE_KERNEL); 285 unsigned long pfn = get_patch_pfn(addr); local 298 __set_pte_at(patching_mm, text_poke_addr, pte, pfn_pte(pfn, PAGE_KERNEL), 0); 331 unsigned long pfn = get_patch_pfn(addr); local 337 __set_pte_at(&init_mm, text_poke_addr, pte, pfn_pte(pfn, PAGE_KERNEL), 0); 408 unsigned long pfn = get_patch_pfn(addr); local 423 __set_pte_at(patching_mm, text_poke_addr, pte, pfn_pte(pfn, PAGE_KERNEL), 0); 456 unsigned long pfn = get_patch_pfn(addr); local 466 __set_pte_at(&init_mm, text_poke_addr, pte, pfn_pte(pfn, PAGE_KERNE [all...] |