/linux-master/arch/arm64/kvm/ |
H A D | mmu.c | 183 static bool kvm_is_device_pfn(unsigned long pfn) argument 185 return !pfn_is_map_memory(pfn); 431 u64 pfn; member in struct:hyp_shared_pfn 439 static struct hyp_shared_pfn *find_shared_pfn(u64 pfn, struct rb_node ***node, argument 449 if (this->pfn < pfn) 451 else if (this->pfn > pfn) 460 static int share_pfn_hyp(u64 pfn) argument 467 this = find_shared_pfn(pfn, 490 unshare_pfn_hyp(u64 pfn) argument 519 u64 pfn; local 551 u64 pfn; local 1288 kvm_pfn_t pfn = *pfnp; local 1355 sanitise_mte_tags(struct kvm *kvm, kvm_pfn_t pfn, unsigned long size) argument 1391 kvm_pfn_t pfn; local 1773 kvm_pfn_t pfn = pte_pfn(range->arg.pte); local [all...] |
/linux-master/arch/x86/platform/efi/ |
H A D | efi_64.c | 180 unsigned long pfn, text, pf, rodata, tramp; local 191 pfn = pa_memmap >> PAGE_SHIFT; 193 if (kernel_map_pages_in_pgd(pgd, pfn, pa_memmap, num_pages, pf)) { 250 pfn = rodata >> PAGE_SHIFT; 253 if (kernel_map_pages_in_pgd(pgd, pfn, rodata, npages, pf)) { 259 pfn = tramp >> PAGE_SHIFT; 262 if (kernel_map_pages_in_pgd(pgd, pfn, tramp, 1, pf)) { 273 unsigned long pfn; local 299 pfn = md->phys_addr >> PAGE_SHIFT; 300 if (kernel_map_pages_in_pgd(pgd, pfn, v 371 unsigned long pfn; local [all...] |
/linux-master/drivers/gpu/drm/ttm/ |
H A D | ttm_bo_vm.c | 190 unsigned long pfn; local 247 pfn = ttm_bo_io_mem_pfn(bo, page_offset); 255 pfn = page_to_pfn(page); 266 ret = vmf_insert_pfn_prot(vma, address, pfn, prot); 298 unsigned long pfn; local 310 pfn = page_to_pfn(page); 315 ret = vmf_insert_pfn_prot(vma, address, pfn, prot);
|
/linux-master/arch/mips/kvm/ |
H A D | mmu.c | 517 kvm_pfn_t pfn = 0; /* silence bogus GCC warning */ local 533 pfn = pte_pfn(*ptep); 545 pfn = pte_pfn(*ptep); 547 kvm_set_pfn_dirty(pfn); 558 kvm_set_pfn_accessed(pfn); 594 kvm_pfn_t pfn; local 614 * Used to check for invalidations in progress, of the pfn that is 632 pfn = gfn_to_pfn_prot(kvm, gfn, write_fault, &writeable); 633 if (is_error_noslot_pfn(pfn)) { 639 /* Check if an invalidation has taken place since we got pfn */ [all...] |
/linux-master/drivers/char/ |
H A D | mem.c | 27 #include <linux/pfn.h> 53 static inline int valid_mmap_phys_addr_range(unsigned long pfn, size_t size) argument 60 static inline int page_is_allowed(unsigned long pfn) argument 62 return devmem_is_allowed(pfn); 64 static inline int range_is_allowed(unsigned long pfn, unsigned long size) argument 66 u64 from = ((u64)pfn) << PAGE_SHIFT; 71 if (!devmem_is_allowed(pfn)) 74 pfn++; 79 static inline int page_is_allowed(unsigned long pfn) argument 83 static inline int range_is_allowed(unsigned long pfn, unsigne argument 260 phys_mem_access_prot_allowed(struct file *file, unsigned long pfn, unsigned long size, pgprot_t *vma_prot) argument 287 phys_mem_access_prot(struct file *file, unsigned long pfn, unsigned long size, pgprot_t vma_prot) argument [all...] |
/linux-master/drivers/vdpa/vdpa_user/ |
H A D | iova_domain.c | 139 unsigned long pfn = PFN_DOWN(orig); local 147 page = pfn_to_page(pfn); 154 pfn++; 233 unsigned long pfn, bounce_pfns; local 237 for (pfn = 0; pfn < bounce_pfns; pfn++) { 238 map = &domain->bounce_maps[pfn]; 578 unsigned long pfn, bounce_pfns; local 600 for (pfn [all...] |
/linux-master/arch/x86/include/asm/ |
H A D | paravirt_types.h | 143 void (*notify_page_enc_status_changed)(unsigned long pfn, int npages, bool enc); 163 void (*alloc_pte)(struct mm_struct *mm, unsigned long pfn); 164 void (*alloc_pmd)(struct mm_struct *mm, unsigned long pfn); 165 void (*alloc_pud)(struct mm_struct *mm, unsigned long pfn); 166 void (*alloc_p4d)(struct mm_struct *mm, unsigned long pfn); 167 void (*release_pte)(unsigned long pfn); 168 void (*release_pmd)(unsigned long pfn); 169 void (*release_pud)(unsigned long pfn); 170 void (*release_p4d)(unsigned long pfn); 208 /* Sometimes the physical address is a pfn, an [all...] |
H A D | pgtable.h | 225 phys_addr_t pfn = pte_val(pte); local 226 pfn ^= protnone_mask(pfn); 227 return (pfn & PTE_PFN_MASK) >> PAGE_SHIFT; 232 phys_addr_t pfn = pmd_val(pmd); local 233 pfn ^= protnone_mask(pfn); 234 return (pfn & pmd_pfn_mask(pmd)) >> PAGE_SHIFT; 239 phys_addr_t pfn = pud_val(pud); local 240 pfn 765 phys_addr_t pfn = (phys_addr_t)page_nr << PAGE_SHIFT; local 773 phys_addr_t pfn = (phys_addr_t)page_nr << PAGE_SHIFT; local 781 phys_addr_t pfn = (phys_addr_t)page_nr << PAGE_SHIFT; local [all...] |
/linux-master/mm/ |
H A D | memremap.c | 76 unsigned long pfn = PHYS_PFN(range->start); local 79 return pfn; 80 return pfn + vmem_altmap_offset(pgmap_altmap(pgmap)); 83 bool pgmap_pfn_valid(struct dev_pagemap *pgmap, unsigned long pfn) argument 90 if (pfn >= PHYS_PFN(range->start) && 91 pfn <= PHYS_PFN(range->end)) 92 return pfn >= pfn_first(pgmap, i); 427 * get_dev_pagemap() - take a new live reference on the dev_pagemap for @pfn 428 * @pfn: page frame number to lookup page_map 431 * If @pgmap is non-NULL and covers @pfn i 434 get_dev_pagemap(unsigned long pfn, struct dev_pagemap *pgmap) argument [all...] |
H A D | debug.c | 55 unsigned long pfn, unsigned long idx) 71 pr_warn("page: refcount:%d mapcount:%d mapping:%p index:%#lx pfn:%#lx\n", 73 folio->index + idx, pfn); 101 is_migrate_cma_folio(folio, pfn) ? " CMA" : ""); 117 unsigned long pfn = page_to_pfn(page); local 149 __dump_folio(foliop, &precise, pfn, idx); 54 __dump_folio(struct folio *folio, struct page *page, unsigned long pfn, unsigned long idx) argument
|
H A D | hmm.c | 96 * multiple page faults into one request and sets flags per pfn for 99 * waste to have the user pre-fill the pfn arrays with a default 193 unsigned long pfn, npages, i; local 204 pfn = pmd_pfn(pmd) + ((addr & ~PMD_MASK) >> PAGE_SHIFT); 205 for (i = 0; addr < end; addr += PAGE_SIZE, i++, pfn++) 206 hmm_pfns[i] = pfn | cpu_flags; 295 * Bypass devmap pte such as DAX page when all pfn requested 361 * huge or device mapping one and compute corresponding pfn 433 unsigned long i, npages, pfn; local 455 pfn 477 unsigned long addr = start, i, pfn; local [all...] |
/linux-master/arch/nios2/mm/ |
H A D | cacheflush.c | 153 unsigned long pfn) 213 unsigned long pfn = pte_pfn(pte); local 219 if (!pfn_valid(pfn)) 226 if (is_zero_pfn(pfn)) 229 folio = page_folio(pfn_to_page(pfn)); 152 flush_cache_page(struct vm_area_struct *vma, unsigned long vmaddr, unsigned long pfn) argument
|
/linux-master/arch/xtensa/mm/ |
H A D | cache.c | 198 unsigned long pfn) 202 unsigned long phys = page_to_phys(pfn_to_page(pfn)); 217 unsigned long pfn = pte_pfn(*ptep); local 221 if (!pfn_valid(pfn)) 224 folio = page_folio(pfn_to_page(pfn)); 197 local_flush_cache_page(struct vm_area_struct *vma, unsigned long address, unsigned long pfn) argument
|
/linux-master/arch/csky/include/asm/ |
H A D | pgtable.h | 39 #define pfn_pte(pfn, prot) __pte(((unsigned long long)(pfn) << PAGE_SHIFT) \ 222 extern pgprot_t phys_mem_access_prot(struct file *file, unsigned long pfn, 271 #define io_remap_pfn_range(vma, vaddr, pfn, size, prot) \ 272 remap_pfn_range(vma, vaddr, pfn, size, prot)
|
/linux-master/arch/powerpc/kvm/ |
H A D | book3s_hv_uvmem.c | 173 * when the GFN is paged-in. Its pfn[] has KVMPPC_GFN_UVMEM_PFN flag 176 * the GFN is pagedout. Its pfn[] has KVMPPC_GFN_MEM_PFN flag set. 179 * Shared GFN is associated with a normal-PFN. Its pfn[] has 183 * Normal GFN is associated with normal-PFN. Its pfn[] has 307 /* mark the GFN as secure-GFN associated with @uvmem pfn device-PFN. */ 522 unsigned long pfn; local 558 pfn = page_to_pfn(dpage); 568 ret = uv_page_out(kvm->arch.lpid, pfn << page_shift, 572 *mig.dst = migrate_pfn(pfn); 749 unsigned long pfn; local 884 unsigned long pfn; local 1020 unsigned long pfn = page_to_pfn(page) - local 1086 unsigned long pfn; local [all...] |
/linux-master/arch/arm64/mm/ |
H A D | trans_pgd.c | 53 * Before marking this entry valid, check the pfn should 236 unsigned long pfn = __phys_to_pfn(dst_addr); local 243 prev_level_entry = pte_val(pfn_pte(pfn, PAGE_KERNEL_ROX)); 257 pfn = virt_to_pfn(levels[this_level]); 258 prev_level_entry = pte_val(pfn_pte(pfn, 265 *trans_ttbr0 = phys_to_ttbr(__pfn_to_phys(pfn));
|
/linux-master/arch/sparc/include/asm/ |
H A D | leon.h | 254 #define _pfn_valid(pfn) ((pfn < last_valid_pfn) && (pfn >= PFN(phys_base)))
|
/linux-master/arch/arm/mach-rpc/ |
H A D | riscpc.c | 70 .pfn = __phys_to_pfn(SCREEN_START), 75 .pfn = __phys_to_pfn(IO_START), 80 .pfn = __phys_to_pfn(EASI_START),
|
/linux-master/include/trace/events/ |
H A D | filemap.h | 23 __field(unsigned long, pfn) 31 __entry->pfn = folio_pfn(folio); 41 TP_printk("dev %d:%d ino %lx pfn=0x%lx ofs=%lu order=%u", 44 __entry->pfn,
|
H A D | xen.h | 287 TP_PROTO(struct mm_struct *mm, unsigned long pfn, unsigned level, bool pinned), 288 TP_ARGS(mm, pfn, level, pinned), 291 __field(unsigned long, pfn) 296 __entry->pfn = pfn; 299 TP_printk("mm %p pfn %lx level %d %spinned", 300 __entry->mm, __entry->pfn, __entry->level, 305 TP_PROTO(unsigned long pfn, unsigned level, bool pinned), 306 TP_ARGS(pfn, level, pinned), 308 __field(unsigned long, pfn) [all...] |
/linux-master/arch/arm/mm/ |
H A D | nommu.c | 204 void __iomem *__arm_ioremap_pfn(unsigned long pfn, unsigned long offset, argument 207 if (pfn >= (0x100000000ULL >> PAGE_SHIFT)) 209 return (void __iomem *) (offset + (pfn << PAGE_SHIFT));
|
/linux-master/arch/mips/include/asm/ |
H A D | page.h | 59 #include <linux/pfn.h> 211 #define pfn_to_kaddr(pfn) __va((pfn) << PAGE_SHIFT)
|
/linux-master/arch/powerpc/include/asm/ |
H A D | page.h | 233 static inline const void *pfn_to_kaddr(unsigned long pfn) argument 235 return __va(pfn << PAGE_SHIFT); 316 extern int devmem_is_allowed(unsigned long pfn);
|
H A D | pgtable.h | 66 * Select all bits except the pfn 121 pgprot_t __phys_mem_access_prot(unsigned long pfn, unsigned long size, 125 static inline pgprot_t phys_mem_access_prot(struct file *file, unsigned long pfn, argument 128 return __phys_mem_access_prot(pfn, size, vma_prot);
|
/linux-master/arch/um/include/asm/ |
H A D | page.h | 21 #include <linux/pfn.h> 109 #define pfn_to_phys(pfn) PFN_PHYS(pfn)
|