/linux-master/arch/arm/mach-footbridge/ |
H A D | common.c | 221 .pfn = __phys_to_pfn(DC21285_ARMCSR_BASE), 227 .pfn = __phys_to_pfn(DC21285_PCI_MEM), 232 .pfn = __phys_to_pfn(DC21285_PCI_TYPE_0_CONFIG), 237 .pfn = __phys_to_pfn(DC21285_PCI_TYPE_1_CONFIG), 242 .pfn = __phys_to_pfn(DC21285_PCI_IACK),
|
/linux-master/drivers/gpu/drm/i915/ |
H A D | i915_scatterlist.h | 10 #include <linux/pfn.h> 26 unsigned long pfn; member in union:sgt_iter::__anon698 42 s.pfn = page_to_pfn(sg_page(s.sgp)); 113 ((__pp) = (__iter).pfn == 0 ? NULL : \ 114 pfn_to_page((__iter).pfn + ((__iter).curr >> PAGE_SHIFT))); \
|
/linux-master/arch/powerpc/include/asm/nohash/32/ |
H A D | mmu-8xx.h | 226 static inline bool arch_vmap_try_size(unsigned long addr, unsigned long end, u64 pfn, argument 238 if (!IS_ALIGNED(PFN_PHYS(pfn), size)) 245 u64 pfn, unsigned int max_page_shift) 247 if (arch_vmap_try_size(addr, end, pfn, max_page_shift, SZ_512K)) 251 if (arch_vmap_try_size(addr, end, pfn, max_page_shift, SZ_16K)) 244 arch_vmap_pte_range_map_size(unsigned long addr, unsigned long end, u64 pfn, unsigned int max_page_shift) argument
|
/linux-master/arch/riscv/include/asm/ |
H A D | page.h | 12 #include <linux/pfn.h> 176 #define pfn_to_phys(pfn) (PFN_PHYS(pfn)) 179 #define pfn_to_virt(pfn) (__va(pfn_to_phys(pfn)))
|
/linux-master/arch/x86/mm/ |
H A D | mmap.c | 224 int valid_mmap_phys_addr_range(unsigned long pfn, size_t count) argument 226 phys_addr_t addr = (phys_addr_t)pfn << PAGE_SHIFT; 238 bool pfn_modify_allowed(unsigned long pfn, pgprot_t prot) argument 245 if (pfn_valid(pfn)) 247 if (pfn >= l1tf_pfn_limit() && !capable(CAP_SYS_ADMIN))
|
H A D | init.c | 127 unsigned long pfn; local 152 pfn = ret >> PAGE_SHIFT; 154 pfn = pgt_buf_end; 161 adr = __va((pfn + i) << PAGE_SHIFT); 165 return __va(pfn << PAGE_SHIFT); 404 unsigned long pfn; local 410 pfn = start_pfn = PFN_DOWN(start); 418 if (pfn == 0) 421 end_pfn = round_up(pfn, PFN_DOWN(PMD_SIZE)); 423 end_pfn = round_up(pfn, PFN_DOW [all...] |
/linux-master/arch/x86/mm/pat/ |
H A D | memtype.c | 456 u64 pfn; local 470 for (pfn = (start >> PAGE_SHIFT); pfn < (end >> PAGE_SHIFT); ++pfn) { 473 page = pfn_to_page(pfn); 488 for (pfn = (start >> PAGE_SHIFT); pfn < (end >> PAGE_SHIFT); ++pfn) { 489 page = pfn_to_page(pfn); 498 u64 pfn; local 707 pat_pfn_immune_to_uc_mtrr(unsigned long pfn) argument 783 phys_mem_access_prot(struct file *file, unsigned long pfn, unsigned long size, pgprot_t vma_prot) argument 794 range_is_allowed(unsigned long pfn, unsigned long size) argument 800 range_is_allowed(unsigned long pfn, unsigned long size) argument 819 phys_mem_access_prot_allowed(struct file *file, unsigned long pfn, unsigned long size, pgprot_t *vma_prot) argument 1035 track_pfn_remap(struct vm_area_struct *vma, pgprot_t *prot, unsigned long pfn, unsigned long addr, unsigned long size) argument 1075 track_pfn_insert(struct vm_area_struct *vma, pgprot_t *prot, pfn_t pfn) argument 1093 untrack_pfn(struct vm_area_struct *vma, unsigned long pfn, unsigned long size, bool mm_wr_locked) argument [all...] |
/linux-master/arch/arm/mach-s3c/ |
H A D | s3c64xx.c | 101 .pfn = __phys_to_pfn(S3C64XX_PA_SYSCON), 106 .pfn = __phys_to_pfn(S3C64XX_PA_SROM), 111 .pfn = __phys_to_pfn(S3C_PA_UART), 116 .pfn = __phys_to_pfn(S3C64XX_PA_VIC0), 121 .pfn = __phys_to_pfn(S3C64XX_PA_VIC1), 126 .pfn = __phys_to_pfn(S3C_PA_TIMER), 131 .pfn = __phys_to_pfn(S3C64XX_PA_GPIO), 136 .pfn = __phys_to_pfn(S3C64XX_PA_MODEM), 141 .pfn = __phys_to_pfn(S3C64XX_PA_WATCHDOG), 146 .pfn [all...] |
/linux-master/arch/x86/entry/vdso/ |
H A D | vma.c | 132 unsigned long pfn; local 154 pfn = __pa_symbol(&__vvar_page) >> PAGE_SHIFT; 174 err = vmf_insert_pfn(vma, addr, pfn); 178 pfn = page_to_pfn(timens_page); 181 return vmf_insert_pfn(vma, vmf->address, pfn); 191 pfn = hv_get_tsc_pfn(); 193 if (pfn && vclock_was_used(VDSO_CLOCKMODE_HVCLOCK)) 194 return vmf_insert_pfn(vma, vmf->address, pfn); 201 pfn = __pa_symbol(&__vvar_page) >> PAGE_SHIFT; 202 return vmf_insert_pfn(vma, vmf->address, pfn); [all...] |
/linux-master/lib/ |
H A D | test_hmm.c | 210 unsigned long pfn; local 212 for (pfn = (range->start >> PAGE_SHIFT); 213 pfn < (range->end >> PAGE_SHIFT); 214 pfn++, pfns++) { 233 entry = xa_store(&dmirror->pt, pfn, entry, GFP_ATOMIC); 244 unsigned long pfn; local 252 xa_for_each_range(&dmirror->pt, pfn, entry, start >> PAGE_SHIFT, 254 xa_erase(&dmirror->pt, pfn); 363 unsigned long pfn; local 368 for (pfn 429 unsigned long pfn; local 498 unsigned long pfn; local 696 unsigned long pfn; local 712 unsigned long pfn, mapped = 0; local 746 unsigned long pfn; local [all...] |
/linux-master/arch/arm64/kvm/ |
H A D | mmu.c | 183 static bool kvm_is_device_pfn(unsigned long pfn) argument 185 return !pfn_is_map_memory(pfn); 431 u64 pfn; member in struct:hyp_shared_pfn 439 static struct hyp_shared_pfn *find_shared_pfn(u64 pfn, struct rb_node ***node, argument 449 if (this->pfn < pfn) 451 else if (this->pfn > pfn) 460 static int share_pfn_hyp(u64 pfn) argument 467 this = find_shared_pfn(pfn, 490 unshare_pfn_hyp(u64 pfn) argument 519 u64 pfn; local 551 u64 pfn; local 1288 kvm_pfn_t pfn = *pfnp; local 1355 sanitise_mte_tags(struct kvm *kvm, kvm_pfn_t pfn, unsigned long size) argument 1391 kvm_pfn_t pfn; local [all...] |
/linux-master/mm/ |
H A D | page_alloc.c | 329 unsigned long pfn) 332 return section_to_usemap(__pfn_to_section(pfn)); 338 static inline int pfn_to_bitidx(const struct page *page, unsigned long pfn) argument 341 pfn &= (PAGES_PER_SECTION-1); 343 pfn = pfn - pageblock_start_pfn(page_zone(page)->zone_start_pfn); 345 return (pfn >> pageblock_order) * NR_PAGEBLOCK_BITS; 351 * @pfn: The target page frame number 357 unsigned long pfn, unsigned long mask) 363 bitmap = get_pageblock_bitmap(page, pfn); 328 get_pageblock_bitmap(const struct page *page, unsigned long pfn) argument 356 get_pfnblock_flags_mask(const struct page *page, unsigned long pfn, unsigned long mask) argument 376 get_pfnblock_migratetype(const struct page *page, unsigned long pfn) argument 389 set_pfnblock_flags_mask(struct page *page, unsigned long flags, unsigned long pfn, unsigned long mask) argument 430 unsigned long pfn = page_to_pfn(page); local 717 buddy_merge_likely(unsigned long pfn, unsigned long buddy_pfn, struct page *page, unsigned int order) argument 757 __free_one_page(struct page *page, unsigned long pfn, struct zone *zone, unsigned int order, int migratetype, fpi_t fpi_flags) argument 1169 unsigned long pfn; local 1189 free_one_page(struct zone *zone, struct page *page, unsigned long pfn, unsigned int order, fpi_t fpi_flags) argument 1205 unsigned long pfn = page_to_pfn(page); local 1554 unsigned long pfn, end_pfn; local 1589 unsigned long pfn, start, end; local 1652 unsigned long pfn = start_pfn; local 1672 split_large_buddy(struct zone *zone, struct page *page, unsigned long pfn, int order) argument 1714 unsigned long start_pfn, pfn; local 2575 unsigned long pfn = page_to_pfn(page); local 2596 free_one_page(page_zone(page), page, pfn, order, FPI_NONE); local 2627 unsigned long pfn = folio_pfn(folio); local 2640 pfn, order, FPI_NONE); local 2653 unsigned long pfn = folio_pfn(folio); local 6331 unsigned long pfn = start; local 6594 unsigned long ret, pfn, flags; local 6629 free_contig_range(unsigned long pfn, unsigned long nr_pages) argument 6691 unsigned long pfn = start_pfn; local 6737 unsigned long pfn = page_to_pfn(page); local 6797 unsigned long pfn = page_to_pfn(page); local 6838 unsigned long pfn = page_to_pfn(page); local [all...] |
H A D | compaction.c | 66 #define block_start_pfn(pfn, order) round_down(pfn, 1UL << (order)) 67 #define block_end_pfn(pfn, order) ALIGN((pfn) + 1, 1UL << (order)) 118 unsigned long pfn = page_to_pfn(page); local 127 if (pfn > high_pfn) 128 high_pfn = pfn; 327 __reset_isolation_pfn(struct zone *zone, unsigned long pfn, bool check_source, argument 330 struct page *page = pfn_to_online_page(pfn); 358 block_pfn = pageblock_start_pfn(pfn); 485 update_cached_migrate(struct compact_control *cc, unsigned long pfn) argument 507 update_pageblock_skip(struct compact_control *cc, struct page *page, unsigned long pfn) argument 532 update_pageblock_skip(struct compact_control *cc, struct page *page, unsigned long pfn) argument 537 update_cached_migrate(struct compact_control *cc, unsigned long pfn) argument 748 unsigned long isolated, pfn, block_start_pfn, block_end_pfn; local 1380 unsigned long pfn, block_start_pfn, block_end_pfn; local 1515 fast_isolate_around(struct compact_control *cc, unsigned long pfn) argument 1618 unsigned long pfn; local 1951 update_fast_start_pfn(struct compact_control *cc, unsigned long pfn) argument 1984 unsigned long pfn = cc->migrate_pfn; local [all...] |
/linux-master/arch/x86/include/asm/ |
H A D | paravirt.h | 104 static inline void notify_page_enc_status_changed(unsigned long pfn, argument 107 PVOP_VCALL3(mmu.notify_page_enc_status_changed, pfn, npages, enc); 354 static inline void paravirt_alloc_pte(struct mm_struct *mm, unsigned long pfn) argument 356 PVOP_VCALL2(mmu.alloc_pte, mm, pfn); 358 static inline void paravirt_release_pte(unsigned long pfn) argument 360 PVOP_VCALL1(mmu.release_pte, pfn); 363 static inline void paravirt_alloc_pmd(struct mm_struct *mm, unsigned long pfn) argument 365 PVOP_VCALL2(mmu.alloc_pmd, mm, pfn); 368 static inline void paravirt_release_pmd(unsigned long pfn) argument 370 PVOP_VCALL1(mmu.release_pmd, pfn); 373 paravirt_alloc_pud(struct mm_struct *mm, unsigned long pfn) argument 377 paravirt_release_pud(unsigned long pfn) argument 382 paravirt_alloc_p4d(struct mm_struct *mm, unsigned long pfn) argument 387 paravirt_release_p4d(unsigned long pfn) argument [all...] |
/linux-master/arch/s390/include/asm/ |
H A D | page.h | 168 static inline int devmem_is_allowed(unsigned long pfn) argument 247 #define pfn_to_phys(pfn) ((pfn) << PAGE_SHIFT) 252 static inline void *pfn_to_virt(unsigned long pfn) argument 254 return __va(pfn_to_phys(pfn)); 262 #define pfn_to_kaddr(pfn) pfn_to_virt(pfn)
|
/linux-master/arch/sh/mm/ |
H A D | cache-sh4.c | 121 unsigned long pfn = folio_pfn(folio); local 128 pfn * PAGE_SIZE); 130 pfn++; 219 unsigned long address, pfn, phys; local 227 pfn = data->addr2; 228 phys = pfn << PAGE_SHIFT; 229 page = pfn_to_page(pfn);
|
H A D | numa.c | 14 #include <linux/pfn.h>
|
/linux-master/kernel/ |
H A D | iomem.c | 30 unsigned long pfn = PHYS_PFN(offset); local 33 if (pfn_valid(pfn) && !PageHighMem(pfn_to_page(pfn)) &&
|
/linux-master/arch/arm/mm/ |
H A D | mmu.c | 713 pgprot_t phys_mem_access_prot(struct file *file, unsigned long pfn, argument 716 if (!pfn_valid(pfn)) 767 unsigned long end, unsigned long pfn, 774 set_pte_ext(pte, pfn_pte(pfn, __pgprot(type->prot_pte)), 776 pfn++; 880 phys = __pfn_to_phys(md->pfn); 885 (long long)__pfn_to_phys((u64)md->pfn), addr); 897 (long long)__pfn_to_phys((u64)md->pfn), addr); 901 if ((addr | length | __pfn_to_phys(md->pfn)) & ~SUPERSECTION_MASK) { 903 (long long)__pfn_to_phys((u64)md->pfn), add 766 alloc_init_pte(pmd_t *pmd, unsigned long addr, unsigned long end, unsigned long pfn, const struct mem_type *type, void *(*alloc)(unsigned long sz), bool ng) argument [all...] |
/linux-master/arch/x86/xen/ |
H A D | mmu.c | 3 #include <linux/pfn.h>
|
/linux-master/drivers/infiniband/hw/mthca/ |
H A D | mthca_uar.c | 44 uar->pfn = (pci_resource_start(dev->pdev, 2) >> PAGE_SHIFT) + uar->index;
|
/linux-master/arch/nios2/mm/ |
H A D | ioremap.c | 27 unsigned long pfn; local 37 pfn = PFN_DOWN(phys_addr); 43 set_pte(pte, pfn_pte(pfn, pgprot)); 45 pfn++;
|
/linux-master/arch/arm/mach-imx/ |
H A D | hardware.h | 104 .pfn = __phys_to_pfn(soc ## _ ## name ## _BASE_ADDR), \
|
/linux-master/arch/powerpc/kernel/ |
H A D | mce_power.c | 33 unsigned long pfn, flags; local 44 pfn = ULONG_MAX; 50 pfn = ULONG_MAX; 55 pfn = pte_pfn(pte); 58 pfn = pte_pfn(__pte(pte_val(pte) | (addr & rpnmask))); 62 return pfn; 459 unsigned long pfn, instr_addr; local 463 pfn = addr_to_pfn(regs, regs->nip); 464 if (pfn != ULONG_MAX) { 465 instr_addr = (pfn << PAGE_SHIF 550 unsigned long pfn; local [all...] |
/linux-master/drivers/gpu/drm/gma500/ |
H A D | gtt.h | 36 uint32_t psb_gtt_mask_pte(uint32_t pfn, int type);
|