/linux-master/include/linux/ |
H A D | mm.h | 17 #include <linux/pfn.h> 1161 extern int page_is_ram(unsigned long pfn); 1924 static inline struct folio *pfn_folio(unsigned long pfn) argument 1926 return page_folio(pfn_to_page(pfn)); 2057 unsigned long node, unsigned long pfn) 2062 set_page_section(page, pfn_to_section_nr(pfn)); 2399 unsigned long *pfn); 3206 static inline int early_pfn_to_nid(unsigned long pfn) argument 3212 extern int __meminit early_pfn_to_nid(unsigned long pfn); 3565 unsigned long pfn, unsigne 2056 set_page_links(struct page *page, enum zone_type zone, unsigned long node, unsigned long pfn) argument 3599 io_remap_pfn_range(struct vm_area_struct *vma, unsigned long addr, unsigned long pfn, unsigned long size, pgprot_t prot) argument 3989 memory_failure_queue(unsigned long pfn, int flags) argument 3993 __get_huge_page_for_hwpoison(unsigned long pfn, int flags, bool *migratable_cleared) argument 3999 num_poisoned_pages_inc(unsigned long pfn) argument 4003 num_poisoned_pages_sub(unsigned long pfn, long i) argument 4018 memblk_nr_poison_inc(unsigned long pfn) argument 4022 memblk_nr_poison_sub(unsigned long pfn, long i) argument 4028 arch_memory_failure(unsigned long pfn, int flags) argument 4200 pfn_is_unaccepted_memory(unsigned long pfn) argument [all...] |
H A D | pgtable.h | 5 #include <linux/pfn.h> 1398 * memory type of pfn mappings specified by the remap_pfn_range, 1403 * track_pfn_remap is called when a _new_ pfn mapping is being established 1404 * by remap_pfn_range() for physical range indicated by pfn and size. 1407 unsigned long pfn, unsigned long addr, 1414 * track_pfn_insert is called when a _new_ single pfn is established 1418 pfn_t pfn) 1433 * untrack can be called for a specific region indicated by pfn and size or 1434 * can be for the entire vma (in which case pfn, size are zero). 1437 unsigned long pfn, unsigne 1406 track_pfn_remap(struct vm_area_struct *vma, pgprot_t *prot, unsigned long pfn, unsigned long addr, unsigned long size) argument 1417 track_pfn_insert(struct vm_area_struct *vma, pgprot_t *prot, pfn_t pfn) argument 1436 untrack_pfn(struct vm_area_struct *vma, unsigned long pfn, unsigned long size, bool mm_wr_locked) argument 1463 is_zero_pfn(unsigned long pfn) argument 1473 is_zero_pfn(unsigned long pfn) argument 1486 is_zero_pfn(unsigned long pfn) argument 1672 pfn_modify_allowed(unsigned long pfn, pgprot_t prot) argument [all...] |
H A D | io-mapping.h | 32 #include <linux/pfn.h> 229 unsigned long addr, unsigned long pfn, unsigned long size);
|
H A D | migrate.h | 174 static inline unsigned long migrate_pfn(unsigned long pfn) argument 176 return (pfn << MIGRATE_PFN_SHIFT) | MIGRATE_PFN_VALID;
|
/linux-master/drivers/base/ |
H A D | memory.c | 58 static inline unsigned long pfn_to_block_id(unsigned long pfn) argument 60 return memory_block_id(pfn_to_section_nr(pfn)); 596 u64 pfn; local 599 if (kstrtoull(buf, 0, &pfn) < 0) 601 pfn >>= PAGE_SHIFT; 602 ret = soft_offline_page(pfn, 0); 612 u64 pfn; local 615 if (kstrtoull(buf, 0, &pfn) < 0) 617 pfn >>= PAGE_SHIFT; 618 ret = memory_failure(pfn, MF_SW_SIMULATE 1230 memblk_nr_poison_inc(unsigned long pfn) argument 1239 memblk_nr_poison_sub(unsigned long pfn, long i) argument [all...] |
/linux-master/arch/sparc/mm/ |
H A D | iommu.c | 53 #define MKIOPTE(pfn, perm) (((((pfn)<<8) & IOPTE_PAGE) | (perm)) & ~IOPTE_WAZ) 191 unsigned long pfn = __phys_to_pfn(paddr); local 213 /* page color = pfn of page */ 214 ioptex = bit_map_string_get(&iommu->usemap, npages, pfn); 223 iopte_val(*iopte) = MKIOPTE(pfn, IOPERM); 227 pfn++;
|
/linux-master/arch/riscv/mm/ |
H A D | pageattr.c | 113 unsigned long pfn = _pmd_pfn(pmdp_get(pmdp)); local 124 set_pte(ptep_new, pfn_pte(pfn + i, prot)); 153 unsigned long pfn = _pud_pfn(pudp_get(pudp)); local 165 pfn_pmd(pfn + ((i * PMD_SIZE) >> PAGE_SHIFT), prot)); 202 unsigned long pfn = _p4d_pfn(p4dp_get(p4dp)); local 218 pfn_pud(pfn + ((i * PUD_SIZE) >> PAGE_SHIFT), prot));
|
/linux-master/include/ras/ |
H A D | ras_event.h | 344 * unsigned long pfn - Page Frame Number of the corrupted page 400 TP_PROTO(unsigned long pfn, 404 TP_ARGS(pfn, type, result), 407 __field(unsigned long, pfn) 413 __entry->pfn = pfn; 418 TP_printk("pfn %#lx: recovery action for %s: %s", 419 __entry->pfn,
|
/linux-master/mm/ |
H A D | memory.c | 566 * to pfn. To get around this restriction, we allow arbitrary mappings so long 585 unsigned long pfn = pte_pfn(pte); local 594 if (is_zero_pfn(pfn)) 615 if (!pfn_valid(pfn)) 621 if (pfn == vma->vm_pgoff + off) 628 if (is_zero_pfn(pfn)) 632 if (unlikely(pfn > highest_memmap_pfn)) { 642 return pfn_to_page(pfn); 659 unsigned long pfn = pmd_pfn(pmd); local 668 if (!pfn_valid(pfn)) 2263 insert_pfn(struct vm_area_struct *vma, unsigned long addr, pfn_t pfn, pgprot_t prot, bool mkwrite) argument 2350 vmf_insert_pfn_prot(struct vm_area_struct *vma, unsigned long addr, unsigned long pfn, pgprot_t pgprot) argument 2398 vmf_insert_pfn(struct vm_area_struct *vma, unsigned long addr, unsigned long pfn) argument 2405 vm_mixed_ok(struct vm_area_struct *vma, pfn_t pfn) argument 2419 __vm_insert_mixed(struct vm_area_struct *vma, unsigned long addr, pfn_t pfn, bool mkwrite) argument 2465 vmf_insert_mixed(struct vm_area_struct *vma, unsigned long addr, pfn_t pfn) argument 2477 vmf_insert_mixed_mkwrite(struct vm_area_struct *vma, unsigned long addr, pfn_t pfn) argument 2489 remap_pte_range(struct mm_struct *mm, pmd_t *pmd, unsigned long addr, unsigned long end, unsigned long pfn, pgprot_t prot) argument 2515 remap_pmd_range(struct mm_struct *mm, pud_t *pud, unsigned long addr, unsigned long end, unsigned long pfn, pgprot_t prot) argument 2538 remap_pud_range(struct mm_struct *mm, p4d_t *p4d, unsigned long addr, unsigned long end, unsigned long pfn, pgprot_t prot) argument 2560 remap_p4d_range(struct mm_struct *mm, pgd_t *pgd, unsigned long addr, unsigned long end, unsigned long pfn, pgprot_t prot) argument 2586 remap_pfn_range_notrack(struct vm_area_struct *vma, unsigned long addr, unsigned long pfn, unsigned long size, pgprot_t prot) argument 2651 remap_pfn_range(struct vm_area_struct *vma, unsigned long addr, unsigned long pfn, unsigned long size, pgprot_t prot) argument 2684 unsigned long vm_len, pfn, pages; local 5941 follow_pfn(struct vm_area_struct *vma, unsigned long address, unsigned long *pfn) argument [all...] |
/linux-master/drivers/hv/ |
H A D | hv_balloon.c | 593 unsigned long pfn) 598 if ((pfn < has->covered_start_pfn) || (pfn >= has->covered_end_pfn)) 603 if ((pfn >= gap->start_pfn) && (pfn < gap->end_pfn)) 613 unsigned long pfn = start_pfn, count = 0; local 617 while (pfn < start_pfn + nr_pages) { 619 * Search for HAS which covers the pfn and when we find one 624 while ((pfn >= has->start_pfn) && 625 (pfn < ha 592 has_pfn_is_backed(struct hv_hotadd_state *has, unsigned long pfn) argument 785 unsigned long pfn = page_to_pfn(pg); local [all...] |
/linux-master/drivers/nvdimm/ |
H A D | pmem.c | 32 #include "pfn.h" 68 unsigned long pfn_start, pfn_end, pfn; local 76 for (pfn = pfn_start; pfn < pfn_end; pfn++) { 77 struct page *page = pfn_to_page(pfn); 85 clear_mce_nospec(pfn); 245 pfn_t *pfn) 256 if (pfn) 257 *pfn 243 __pmem_direct_access(struct pmem_device *pmem, pgoff_t pgoff, long nr_pages, enum dax_access_mode mode, void **kaddr, pfn_t *pfn) argument 304 pmem_dax_direct_access(struct dax_device *dax_dev, pgoff_t pgoff, long nr_pages, enum dax_access_mode mode, void **kaddr, pfn_t *pfn) argument 434 pmem_pagemap_memory_failure(struct dev_pagemap *pgmap, unsigned long pfn, unsigned long nr_pages, int mf_flags) argument [all...] |
/linux-master/arch/arm/mm/ |
H A D | mmap.c | 164 int valid_mmap_phys_addr_range(unsigned long pfn, size_t size) argument 166 return (pfn + (size >> PAGE_SHIFT)) <= (1 + (PHYS_MASK >> PAGE_SHIFT));
|
/linux-master/tools/testing/scatterlist/linux/ |
H A D | mm.h | 53 #define pfn_to_page(pfn) (void *)((pfn) * PAGE_SIZE)
|
/linux-master/drivers/xen/ |
H A D | unpopulated-alloc.c | 109 xen_pfn_t pfn = PFN_DOWN(res->start); local 112 if (!set_phys_to_machine(pfn + i, INVALID_P2M_ENTRY)) {
|
H A D | grant-dma-ops.c | 14 #include <linux/pfn.h> 84 unsigned long pfn; local 99 pfn = virt_to_pfn(ret); 108 pfn_to_gfn(pfn + i), 0);
|
/linux-master/include/xen/ |
H A D | xen-ops.h | 48 xen_pfn_t *pfn, int nr, int *err_ptr, pgprot_t prot, 52 xen_pfn_t *pfn, int nr, int *err_ptr, 51 xen_remap_pfn(struct vm_area_struct *vma, unsigned long addr, xen_pfn_t *pfn, int nr, int *err_ptr, pgprot_t prot, unsigned int domid, bool no_translate) argument
|
/linux-master/arch/arm/mach-pxa/ |
H A D | pxa25x.c | 163 .pfn = __phys_to_pfn(PXA2XX_SMEMC_BASE), 168 .pfn = __phys_to_pfn(0x00000000),
|
/linux-master/drivers/misc/sgi-gru/ |
H A D | gruhandles.c | 157 tfh->pfn = paddr >> GRU_PADDR_SHIFT; 173 tfh->pfn = paddr >> GRU_PADDR_SHIFT;
|
/linux-master/drivers/firmware/efi/ |
H A D | unaccepted_memory.c | 208 unsigned long pfn) 210 return !pfn_is_unaccepted_memory(pfn); 207 unaccepted_memory_vmcore_pfn_is_ram(struct vmcore_cb *cb, unsigned long pfn) argument
|
/linux-master/arch/sparc/include/asm/ |
H A D | page_64.h | 149 #define pfn_to_kaddr(pfn) __va((pfn) << PAGE_SHIFT)
|
/linux-master/arch/sh/include/asm/ |
H A D | page.h | 149 #define pfn_to_kaddr(pfn) __va((pfn) << PAGE_SHIFT)
|
/linux-master/kernel/debug/kdb/ |
H A D | kdb_support.c | 319 unsigned long pfn; local 323 pfn = (addr >> PAGE_SHIFT); 324 if (!pfn_valid(pfn)) 326 page = pfn_to_page(pfn);
|
/linux-master/drivers/gpu/drm/gma500/ |
H A D | gem.c | 261 unsigned long pfn; local 294 pfn = (dev_priv->stolen_base + pobj->offset) >> PAGE_SHIFT; 296 pfn = page_to_pfn(pobj->pages[page_offset]); 297 ret = vmf_insert_pfn(vma, vmf->address, pfn);
|
/linux-master/drivers/net/ethernet/mellanox/mlx5/core/ |
H A D | uar.c | 99 phys_addr_t pfn; local 135 pfn = uar2pfn(mdev, up->index); 137 up->map = ioremap_wc(pfn << PAGE_SHIFT, PAGE_SIZE); 143 up->map = ioremap(pfn << PAGE_SHIFT, PAGE_SIZE);
|
/linux-master/arch/arm/mach-sa1100/ |
H A D | generic.c | 364 .pfn = __phys_to_pfn(0x80000000), 369 .pfn = __phys_to_pfn(0x90000000), 374 .pfn = __phys_to_pfn(0xa0000000), 379 .pfn = __phys_to_pfn(0xb0000000),
|