Searched refs:pfn (Results 201 - 225 of 620) sorted by relevance

1234567891011>>

/linux-master/arch/arm64/kvm/
H A Dmmu.c183 static bool kvm_is_device_pfn(unsigned long pfn) argument
185 return !pfn_is_map_memory(pfn);
431 u64 pfn; member in struct:hyp_shared_pfn
439 static struct hyp_shared_pfn *find_shared_pfn(u64 pfn, struct rb_node ***node, argument
449 if (this->pfn < pfn)
451 else if (this->pfn > pfn)
460 static int share_pfn_hyp(u64 pfn) argument
467 this = find_shared_pfn(pfn,
490 unshare_pfn_hyp(u64 pfn) argument
519 u64 pfn; local
551 u64 pfn; local
1288 kvm_pfn_t pfn = *pfnp; local
1355 sanitise_mte_tags(struct kvm *kvm, kvm_pfn_t pfn, unsigned long size) argument
1391 kvm_pfn_t pfn; local
1773 kvm_pfn_t pfn = pte_pfn(range->arg.pte); local
[all...]
/linux-master/arch/x86/platform/efi/
H A Defi_64.c180 unsigned long pfn, text, pf, rodata, tramp; local
191 pfn = pa_memmap >> PAGE_SHIFT;
193 if (kernel_map_pages_in_pgd(pgd, pfn, pa_memmap, num_pages, pf)) {
250 pfn = rodata >> PAGE_SHIFT;
253 if (kernel_map_pages_in_pgd(pgd, pfn, rodata, npages, pf)) {
259 pfn = tramp >> PAGE_SHIFT;
262 if (kernel_map_pages_in_pgd(pgd, pfn, tramp, 1, pf)) {
273 unsigned long pfn; local
299 pfn = md->phys_addr >> PAGE_SHIFT;
300 if (kernel_map_pages_in_pgd(pgd, pfn, v
371 unsigned long pfn; local
[all...]
/linux-master/drivers/gpu/drm/ttm/
H A Dttm_bo_vm.c190 unsigned long pfn; local
247 pfn = ttm_bo_io_mem_pfn(bo, page_offset);
255 pfn = page_to_pfn(page);
266 ret = vmf_insert_pfn_prot(vma, address, pfn, prot);
298 unsigned long pfn; local
310 pfn = page_to_pfn(page);
315 ret = vmf_insert_pfn_prot(vma, address, pfn, prot);
/linux-master/arch/mips/kvm/
H A Dmmu.c517 kvm_pfn_t pfn = 0; /* silence bogus GCC warning */ local
533 pfn = pte_pfn(*ptep);
545 pfn = pte_pfn(*ptep);
547 kvm_set_pfn_dirty(pfn);
558 kvm_set_pfn_accessed(pfn);
594 kvm_pfn_t pfn; local
614 * Used to check for invalidations in progress, of the pfn that is
632 pfn = gfn_to_pfn_prot(kvm, gfn, write_fault, &writeable);
633 if (is_error_noslot_pfn(pfn)) {
639 /* Check if an invalidation has taken place since we got pfn */
[all...]
/linux-master/drivers/char/
H A Dmem.c27 #include <linux/pfn.h>
53 static inline int valid_mmap_phys_addr_range(unsigned long pfn, size_t size) argument
60 static inline int page_is_allowed(unsigned long pfn) argument
62 return devmem_is_allowed(pfn);
64 static inline int range_is_allowed(unsigned long pfn, unsigned long size) argument
66 u64 from = ((u64)pfn) << PAGE_SHIFT;
71 if (!devmem_is_allowed(pfn))
74 pfn++;
79 static inline int page_is_allowed(unsigned long pfn) argument
83 static inline int range_is_allowed(unsigned long pfn, unsigne argument
260 phys_mem_access_prot_allowed(struct file *file, unsigned long pfn, unsigned long size, pgprot_t *vma_prot) argument
287 phys_mem_access_prot(struct file *file, unsigned long pfn, unsigned long size, pgprot_t vma_prot) argument
[all...]
/linux-master/drivers/vdpa/vdpa_user/
H A Diova_domain.c139 unsigned long pfn = PFN_DOWN(orig); local
147 page = pfn_to_page(pfn);
154 pfn++;
233 unsigned long pfn, bounce_pfns; local
237 for (pfn = 0; pfn < bounce_pfns; pfn++) {
238 map = &domain->bounce_maps[pfn];
578 unsigned long pfn, bounce_pfns; local
600 for (pfn
[all...]
/linux-master/arch/x86/include/asm/
H A Dparavirt_types.h143 void (*notify_page_enc_status_changed)(unsigned long pfn, int npages, bool enc);
163 void (*alloc_pte)(struct mm_struct *mm, unsigned long pfn);
164 void (*alloc_pmd)(struct mm_struct *mm, unsigned long pfn);
165 void (*alloc_pud)(struct mm_struct *mm, unsigned long pfn);
166 void (*alloc_p4d)(struct mm_struct *mm, unsigned long pfn);
167 void (*release_pte)(unsigned long pfn);
168 void (*release_pmd)(unsigned long pfn);
169 void (*release_pud)(unsigned long pfn);
170 void (*release_p4d)(unsigned long pfn);
208 /* Sometimes the physical address is a pfn, an
[all...]
H A Dpgtable.h225 phys_addr_t pfn = pte_val(pte); local
226 pfn ^= protnone_mask(pfn);
227 return (pfn & PTE_PFN_MASK) >> PAGE_SHIFT;
232 phys_addr_t pfn = pmd_val(pmd); local
233 pfn ^= protnone_mask(pfn);
234 return (pfn & pmd_pfn_mask(pmd)) >> PAGE_SHIFT;
239 phys_addr_t pfn = pud_val(pud); local
240 pfn
765 phys_addr_t pfn = (phys_addr_t)page_nr << PAGE_SHIFT; local
773 phys_addr_t pfn = (phys_addr_t)page_nr << PAGE_SHIFT; local
781 phys_addr_t pfn = (phys_addr_t)page_nr << PAGE_SHIFT; local
[all...]
/linux-master/mm/
H A Dmemremap.c76 unsigned long pfn = PHYS_PFN(range->start); local
79 return pfn;
80 return pfn + vmem_altmap_offset(pgmap_altmap(pgmap));
83 bool pgmap_pfn_valid(struct dev_pagemap *pgmap, unsigned long pfn) argument
90 if (pfn >= PHYS_PFN(range->start) &&
91 pfn <= PHYS_PFN(range->end))
92 return pfn >= pfn_first(pgmap, i);
427 * get_dev_pagemap() - take a new live reference on the dev_pagemap for @pfn
428 * @pfn: page frame number to lookup page_map
431 * If @pgmap is non-NULL and covers @pfn i
434 get_dev_pagemap(unsigned long pfn, struct dev_pagemap *pgmap) argument
[all...]
H A Ddebug.c55 unsigned long pfn, unsigned long idx)
71 pr_warn("page: refcount:%d mapcount:%d mapping:%p index:%#lx pfn:%#lx\n",
73 folio->index + idx, pfn);
101 is_migrate_cma_folio(folio, pfn) ? " CMA" : "");
117 unsigned long pfn = page_to_pfn(page); local
149 __dump_folio(foliop, &precise, pfn, idx);
54 __dump_folio(struct folio *folio, struct page *page, unsigned long pfn, unsigned long idx) argument
H A Dhmm.c96 * multiple page faults into one request and sets flags per pfn for
99 * waste to have the user pre-fill the pfn arrays with a default
193 unsigned long pfn, npages, i; local
204 pfn = pmd_pfn(pmd) + ((addr & ~PMD_MASK) >> PAGE_SHIFT);
205 for (i = 0; addr < end; addr += PAGE_SIZE, i++, pfn++)
206 hmm_pfns[i] = pfn | cpu_flags;
295 * Bypass devmap pte such as DAX page when all pfn requested
361 * huge or device mapping one and compute corresponding pfn
433 unsigned long i, npages, pfn; local
455 pfn
477 unsigned long addr = start, i, pfn; local
[all...]
/linux-master/arch/nios2/mm/
H A Dcacheflush.c153 unsigned long pfn)
213 unsigned long pfn = pte_pfn(pte); local
219 if (!pfn_valid(pfn))
226 if (is_zero_pfn(pfn))
229 folio = page_folio(pfn_to_page(pfn));
152 flush_cache_page(struct vm_area_struct *vma, unsigned long vmaddr, unsigned long pfn) argument
/linux-master/arch/xtensa/mm/
H A Dcache.c198 unsigned long pfn)
202 unsigned long phys = page_to_phys(pfn_to_page(pfn));
217 unsigned long pfn = pte_pfn(*ptep); local
221 if (!pfn_valid(pfn))
224 folio = page_folio(pfn_to_page(pfn));
197 local_flush_cache_page(struct vm_area_struct *vma, unsigned long address, unsigned long pfn) argument
/linux-master/arch/csky/include/asm/
H A Dpgtable.h39 #define pfn_pte(pfn, prot) __pte(((unsigned long long)(pfn) << PAGE_SHIFT) \
222 extern pgprot_t phys_mem_access_prot(struct file *file, unsigned long pfn,
271 #define io_remap_pfn_range(vma, vaddr, pfn, size, prot) \
272 remap_pfn_range(vma, vaddr, pfn, size, prot)
/linux-master/arch/powerpc/kvm/
H A Dbook3s_hv_uvmem.c173 * when the GFN is paged-in. Its pfn[] has KVMPPC_GFN_UVMEM_PFN flag
176 * the GFN is pagedout. Its pfn[] has KVMPPC_GFN_MEM_PFN flag set.
179 * Shared GFN is associated with a normal-PFN. Its pfn[] has
183 * Normal GFN is associated with normal-PFN. Its pfn[] has
307 /* mark the GFN as secure-GFN associated with @uvmem pfn device-PFN. */
522 unsigned long pfn; local
558 pfn = page_to_pfn(dpage);
568 ret = uv_page_out(kvm->arch.lpid, pfn << page_shift,
572 *mig.dst = migrate_pfn(pfn);
749 unsigned long pfn; local
884 unsigned long pfn; local
1020 unsigned long pfn = page_to_pfn(page) - local
1086 unsigned long pfn; local
[all...]
/linux-master/arch/arm64/mm/
H A Dtrans_pgd.c53 * Before marking this entry valid, check the pfn should
236 unsigned long pfn = __phys_to_pfn(dst_addr); local
243 prev_level_entry = pte_val(pfn_pte(pfn, PAGE_KERNEL_ROX));
257 pfn = virt_to_pfn(levels[this_level]);
258 prev_level_entry = pte_val(pfn_pte(pfn,
265 *trans_ttbr0 = phys_to_ttbr(__pfn_to_phys(pfn));
/linux-master/arch/sparc/include/asm/
H A Dleon.h254 #define _pfn_valid(pfn) ((pfn < last_valid_pfn) && (pfn >= PFN(phys_base)))
/linux-master/arch/arm/mach-rpc/
H A Driscpc.c70 .pfn = __phys_to_pfn(SCREEN_START),
75 .pfn = __phys_to_pfn(IO_START),
80 .pfn = __phys_to_pfn(EASI_START),
/linux-master/include/trace/events/
H A Dfilemap.h23 __field(unsigned long, pfn)
31 __entry->pfn = folio_pfn(folio);
41 TP_printk("dev %d:%d ino %lx pfn=0x%lx ofs=%lu order=%u",
44 __entry->pfn,
H A Dxen.h287 TP_PROTO(struct mm_struct *mm, unsigned long pfn, unsigned level, bool pinned),
288 TP_ARGS(mm, pfn, level, pinned),
291 __field(unsigned long, pfn)
296 __entry->pfn = pfn;
299 TP_printk("mm %p pfn %lx level %d %spinned",
300 __entry->mm, __entry->pfn, __entry->level,
305 TP_PROTO(unsigned long pfn, unsigned level, bool pinned),
306 TP_ARGS(pfn, level, pinned),
308 __field(unsigned long, pfn)
[all...]
/linux-master/arch/arm/mm/
H A Dnommu.c204 void __iomem *__arm_ioremap_pfn(unsigned long pfn, unsigned long offset, argument
207 if (pfn >= (0x100000000ULL >> PAGE_SHIFT))
209 return (void __iomem *) (offset + (pfn << PAGE_SHIFT));
/linux-master/arch/mips/include/asm/
H A Dpage.h59 #include <linux/pfn.h>
211 #define pfn_to_kaddr(pfn) __va((pfn) << PAGE_SHIFT)
/linux-master/arch/powerpc/include/asm/
H A Dpage.h233 static inline const void *pfn_to_kaddr(unsigned long pfn) argument
235 return __va(pfn << PAGE_SHIFT);
316 extern int devmem_is_allowed(unsigned long pfn);
H A Dpgtable.h66 * Select all bits except the pfn
121 pgprot_t __phys_mem_access_prot(unsigned long pfn, unsigned long size,
125 static inline pgprot_t phys_mem_access_prot(struct file *file, unsigned long pfn, argument
128 return __phys_mem_access_prot(pfn, size, vma_prot);
/linux-master/arch/um/include/asm/
H A Dpage.h21 #include <linux/pfn.h>
109 #define pfn_to_phys(pfn) PFN_PHYS(pfn)

Completed in 178 milliseconds

1234567891011>>