/linux-master/arch/x86/power/ |
H A D | hibernate.c | 44 * pfn_is_nosave - check if given pfn is in the 'nosave' section 46 int pfn_is_nosave(unsigned long pfn) argument 54 return pfn >= nosave_begin_pfn && pfn < nosave_end_pfn;
|
/linux-master/arch/alpha/include/asm/ |
H A D | pgtable.h | 179 #define PHYS_TWIDDLE(pfn) \ 180 ((((pfn) & KSEG_PFN) == (0x40000000000UL >> PAGE_SHIFT)) \ 181 ? ((pfn) ^= KSEG_PFN) : (pfn)) 183 #define PHYS_TWIDDLE(pfn) (pfn)
|
/linux-master/drivers/edac/ |
H A D | cell_edac.c | 38 unsigned long address, pfn, offset, syndrome; local 47 pfn = address >> PAGE_SHIFT; 53 csrow->first_page + pfn, offset, syndrome, 61 unsigned long address, pfn, offset; local 70 pfn = address >> PAGE_SHIFT; 75 csrow->first_page + pfn, offset, 0,
|
/linux-master/arch/powerpc/mm/ |
H A D | pgtable.c | 63 unsigned long pfn = pte_pfn(pte); local 66 if (unlikely(!pfn_valid(pfn))) 68 page = pfn_to_page(pfn); 362 unsigned long pfn = vmalloc_to_pfn(va); local 364 BUG_ON(!pfn); 365 return __pa(pfn_to_kaddr(pfn)) + offset_in_page(va);
|
H A D | cacheflush.c | 170 unsigned long pfn = folio_pfn(folio); local 172 flush_dcache_icache_phys((pfn + i) * PAGE_SIZE);
|
/linux-master/arch/riscv/include/asm/ |
H A D | pgtable-64.h | 213 static inline pud_t pfn_pud(unsigned long pfn, pgprot_t prot) argument 215 return __pud((pfn << _PAGE_PFN_SHIFT) | pgprot_val(prot)); 253 static inline pmd_t pfn_pmd(unsigned long pfn, pgprot_t prot) argument 259 return __pmd((pfn << _PAGE_PFN_SHIFT) | prot_val); 316 static inline p4d_t pfn_p4d(unsigned long pfn, pgprot_t prot) argument 318 return __p4d((pfn << _PAGE_PFN_SHIFT) | pgprot_val(prot));
|
/linux-master/drivers/net/ethernet/ibm/ehea/ |
H A D | ehea_qmr.c | 608 static int ehea_update_busmap(unsigned long pfn, unsigned long nr_pages, int add) argument 621 start_section = (pfn * PAGE_SIZE) / EHEA_SECTSIZE; 651 int ehea_add_sect_bmap(unsigned long pfn, unsigned long nr_pages) argument 656 ret = ehea_update_busmap(pfn, nr_pages, EHEA_BUSMAP_ADD_SECT); 661 int ehea_rem_sect_bmap(unsigned long pfn, unsigned long nr_pages) argument 666 ret = ehea_update_busmap(pfn, nr_pages, EHEA_BUSMAP_REM_SECT); 671 static int ehea_is_hugepage(unsigned long pfn) argument 673 if (pfn & EHEA_HUGEPAGE_PFN_MASK) 676 if (page_shift(pfn_to_page(pfn)) != EHEA_HUGEPAGESHIFT) 686 unsigned long pfn, start_pf local [all...] |
/linux-master/arch/arm/mach-versatile/ |
H A D | integrator_ap.c | 47 .pfn = __phys_to_pfn(INTEGRATOR_IC_BASE), 52 .pfn = __phys_to_pfn(INTEGRATOR_UART0_BASE),
|
/linux-master/mm/ |
H A D | shuffle.c | 37 unsigned long pfn, int order) 39 struct page *page = pfn_to_online_page(pfn); 108 * a swap target. If the selected pfn is a hole, retry 109 * up to SHUFFLE_RETRY attempts find a random valid pfn 36 shuffle_valid_page(struct zone *zone, unsigned long pfn, int order) argument
|
H A D | migrate_device.c | 109 unsigned long mpfn = 0, pfn; local 146 pfn = pte_pfn(pte); 147 if (is_zero_pfn(pfn) && 161 mpfn = migrate_pfn(pfn) | MIGRATE_PFN_MIGRATE; 471 * corresponding entry in the dst array with the pfn value of the destination 643 unsigned long pfn = pte_pfn(orig_pte); local 645 if (!is_zero_pfn(pfn)) 882 * @start: starting pfn in the range to migrate. 901 unsigned long i, pfn; local 903 for (pfn [all...] |
/linux-master/arch/powerpc/include/asm/book3s/64/ |
H A D | hash-4k.h | 74 #define remap_4k_pfn(vma, addr, pfn, prot) \ 75 remap_pfn_range((vma), (addr), (pfn), PAGE_SIZE, (prot))
|
/linux-master/arch/loongarch/include/asm/ |
H A D | kvm_mmu.h | 21 #define kvm_pfn_pte(pfn, prot) (((pfn) << PFN_PTE_SHIFT) | pgprot_val(prot))
|
/linux-master/mm/damon/ |
H A D | ops-common.c | 16 * Get an online page for a pfn if it's in the LRU list. Otherwise, returns 22 struct folio *damon_get_folio(unsigned long pfn) argument 24 struct page *page = pfn_to_online_page(pfn);
|
/linux-master/arch/powerpc/kernel/ |
H A D | crash_dump.c | 72 ssize_t copy_oldmem_page(struct iov_iter *iter, unsigned long pfn, argument 82 paddr = pfn << PAGE_SHIFT;
|
/linux-master/arch/x86/include/asm/xen/ |
H A D | interface_64.h | 133 #define xen_pfn_to_cr3(pfn) ((unsigned long)(pfn) << 12)
|
/linux-master/drivers/iommu/ |
H A D | iova.c | 22 unsigned long pfn, 275 private_find_iova(struct iova_domain *iovad, unsigned long pfn) argument 284 if (pfn < iova->pfn_lo) 286 else if (pfn > iova->pfn_hi) 289 return iova; /* pfn falls within iova's range */ 303 * find_iova - finds an iova for a given pfn 305 * @pfn: - page frame number 307 * given domain which matches the given pfn. 309 struct iova *find_iova(struct iova_domain *iovad, unsigned long pfn) argument 316 iova = private_find_iova(iovad, pfn); 348 free_iova(struct iova_domain *iovad, unsigned long pfn) argument 424 free_iova_fast(struct iova_domain *iovad, unsigned long pfn, unsigned long size) argument 653 unsigned long pfn; local 667 iova_magazine_push(struct iova_magazine *mag, unsigned long pfn) argument 801 iova_rcache_insert(struct iova_domain *iovad, unsigned long pfn, unsigned long size) argument [all...] |
/linux-master/drivers/xen/ |
H A D | balloon.c | 301 unsigned long pfn, i; local 303 pfn = PFN_DOWN(resource->start); 305 if (!set_phys_to_machine(pfn + i, INVALID_P2M_ENTRY)) { 345 pr_debug("Online %lu pages starting at pfn 0x%lx\n", size, start_pfn); 676 unsigned long pfn, extra_pfn_end; local 693 for (pfn = start_pfn; pfn < extra_pfn_end; pfn++) 694 balloon_append(pfn_to_page(pfn));
|
/linux-master/arch/loongarch/mm/ |
H A D | mmap.c | 147 int valid_mmap_phys_addr_range(unsigned long pfn, size_t size) argument 149 return !(((pfn << PAGE_SHIFT) + size) & ~(GENMASK_ULL(cpu_pabits, 0)));
|
/linux-master/drivers/net/wireless/broadcom/brcm80211/brcmfmac/ |
H A D | pno.c | 201 struct brcmf_pno_net_param_le pfn; local 204 pfn.auth = cpu_to_le32(WLAN_AUTH_OPEN); 205 pfn.wpa_auth = cpu_to_le32(BRCMF_PNO_WPA_AUTH_ANY); 206 pfn.wsec = cpu_to_le32(0); 207 pfn.infra = cpu_to_le32(1); 208 pfn.flags = 0; 210 pfn.flags = cpu_to_le32(1 << BRCMF_PNO_HIDDEN_BIT); 211 pfn.ssid.SSID_len = cpu_to_le32(ssid->ssid_len); 212 memcpy(pfn.ssid.SSID, ssid->ssid, ssid->ssid_len); 215 err = brcmf_fil_iovar_data_set(ifp, "pfn_add", &pfn, sizeo [all...] |
/linux-master/drivers/vfio/ |
H A D | vfio_iommu_type1.c | 98 struct rb_root pfn_list; /* Ex-user pinned pfn list */ 130 unsigned long pfn; /* Host pfn */ member in struct:vfio_pfn 154 static int put_pfn(unsigned long pfn, int prot); 321 * Helper Functions for host iova-pfn list 368 unsigned long pfn) 377 vpfn->pfn = pfn; 406 ret = put_pfn(vpfn->pfn, dma->prot); 450 * pfn conversio 367 vfio_add_to_pfn_list(struct vfio_dma *dma, dma_addr_t iova, unsigned long pfn) argument 454 is_invalid_reserved_pfn(unsigned long pfn) argument 462 put_pfn(unsigned long pfn, int prot) argument 498 unsigned long pfn = page_to_pfn(batch->pages[batch->offset]); local 512 follow_fault_pfn(struct vm_area_struct *vma, struct mm_struct *mm, unsigned long vaddr, unsigned long *pfn, bool write_fault) argument 555 vaddr_get_pfns(struct mm_struct *mm, unsigned long vaddr, long npages, int prot, unsigned long *pfn, struct page **pages) argument 605 unsigned long pfn; local 712 vfio_unpin_pages_remote(struct vfio_dma *dma, dma_addr_t iova, unsigned long pfn, long npage, bool do_accounting) argument 1418 vfio_iommu_map(struct vfio_iommu *iommu, dma_addr_t iova, unsigned long pfn, long npage, int prot) argument 1453 unsigned long pfn, limit = rlimit(RLIMIT_MEMLOCK) >> PAGE_SHIFT; local 1720 unsigned long pfn; local [all...] |
/linux-master/arch/arm64/mm/ |
H A D | hugetlbpage.c | 243 unsigned long pfn, dpfn; local 259 pfn = pte_pfn(pte); 265 for (i = 0; i < ncontig; i++, ptep++, addr += pgsize, pfn += dpfn) 266 __set_ptes(mm, addr, ptep, pfn_pte(pfn, hugeprot), 1); 452 unsigned long pfn = pte_pfn(pte), dpfn; local 476 for (i = 0; i < ncontig; i++, ptep++, addr += pgsize, pfn += dpfn) 477 __set_ptes(mm, addr, ptep, pfn_pte(pfn, hugeprot), 1); 485 unsigned long pfn, dpfn; local 503 pfn = pte_pfn(pte); 505 for (i = 0; i < ncontig; i++, ptep++, addr += pgsize, pfn [all...] |
/linux-master/arch/sparc/include/asm/ |
H A D | pgtable_64.h | 221 static inline pte_t pfn_pte(unsigned long pfn, pgprot_t prot) argument 223 unsigned long paddr = pfn << PAGE_SHIFT; 840 unsigned long pfn; local 842 pfn = pte_pfn(pte); 844 return ((unsigned long) __va(pfn << PAGE_SHIFT)); 850 unsigned long pfn; local 852 pfn = pte_pfn(pte); 854 return ((pmd_t *) __va(pfn << PAGE_SHIFT)); 1046 * For sparc32&64, the pfn in io_remap_pfn_range() carries <iospace> in 1049 #define MK_IOSPACE_PFN(space, pfn) (pf 1089 io_remap_pfn_range(struct vm_area_struct *vma, unsigned long from, unsigned long pfn, unsigned long size, pgprot_t prot) argument [all...] |
/linux-master/arch/loongarch/kvm/ |
H A D | mmu.c | 501 kvm_pfn_t pfn = pte_pfn(range->arg.pte); local 524 kvm_set_pte(ptep, kvm_pfn_pte(pfn, __pgprot(prot_bits))); 570 kvm_pfn_t pfn = 0; local 616 pfn = kvm_pte_pfn(new); 621 * Fixme: pfn may be freed after mmu_lock 622 * kvm_try_get_pfn(pfn)/kvm_release_pfn pair to prevent this? 625 kvm_set_pfn_accessed(pfn); 629 kvm_set_pfn_dirty(pfn); 805 kvm_pfn_t pfn; local 832 * Used to check for invalidations in progress, of the pfn tha [all...] |
/linux-master/drivers/scsi/csiostor/ |
H A D | csio_mb.c | 104 * @mpfn: Master pfn 281 (FW_LDST_CMD_LC_F | FW_LDST_CMD_FN_V(hw->pfn)); 450 FW_IQ_CMD_PFN_V(iq_params->pfn) | 507 FW_IQ_CMD_PFN_V(iq_params->pfn) | 632 FW_IQ_CMD_PFN_V(iq_params->pfn) | 667 FW_EQ_OFLD_CMD_PFN_V(eq_ofld_params->pfn) | 715 FW_EQ_OFLD_CMD_PFN_V(eq_ofld_params->pfn) | 818 FW_EQ_OFLD_CMD_PFN_V(eq_ofld_params->pfn) | 1162 uint32_t ctl_reg = PF_REG(hw->pfn, CIM_PF_MAILBOX_CTRL_A); 1163 uint32_t data_reg = PF_REG(hw->pfn, CIM_PF_MAILBOX_DATA_ [all...] |
/linux-master/fs/proc/ |
H A D | vmcore.c | 104 static bool pfn_is_ram(unsigned long pfn) argument 113 ret = cb->pfn_is_ram(cb, pfn); 134 unsigned long pfn, offset; local 143 pfn = (unsigned long)(*ppos / PAGE_SIZE); 152 /* If pfn is not ram, return zeros for sparse dump files */ 153 if (!pfn_is_ram(pfn)) { 157 tmp = copy_oldmem_page_encrypted(iter, pfn, 161 tmp = copy_oldmem_page(iter, pfn, nr_bytes, 172 ++pfn; 225 unsigned long from, unsigned long pfn, 224 remap_oldmem_pfn_range(struct vm_area_struct *vma, unsigned long from, unsigned long pfn, unsigned long size, pgprot_t prot) argument 235 copy_oldmem_page_encrypted(struct iov_iter *iter, unsigned long pfn, size_t csize, unsigned long offset) argument 501 remap_oldmem_pfn_checked(struct vm_area_struct *vma, unsigned long from, unsigned long pfn, unsigned long size, pgprot_t prot) argument 551 vmcore_remap_oldmem_pfn(struct vm_area_struct *vma, unsigned long from, unsigned long pfn, unsigned long size, pgprot_t prot) argument 591 u64 pfn; local [all...] |