Searched refs:pfn (Results 276 - 300 of 620) sorted by relevance

<<11121314151617181920>>

/linux-master/include/linux/
H A Dmm.h17 #include <linux/pfn.h>
1161 extern int page_is_ram(unsigned long pfn);
1924 static inline struct folio *pfn_folio(unsigned long pfn) argument
1926 return page_folio(pfn_to_page(pfn));
2057 unsigned long node, unsigned long pfn)
2062 set_page_section(page, pfn_to_section_nr(pfn));
2399 unsigned long *pfn);
3206 static inline int early_pfn_to_nid(unsigned long pfn) argument
3212 extern int __meminit early_pfn_to_nid(unsigned long pfn);
3565 unsigned long pfn, unsigne
2056 set_page_links(struct page *page, enum zone_type zone, unsigned long node, unsigned long pfn) argument
3599 io_remap_pfn_range(struct vm_area_struct *vma, unsigned long addr, unsigned long pfn, unsigned long size, pgprot_t prot) argument
3989 memory_failure_queue(unsigned long pfn, int flags) argument
3993 __get_huge_page_for_hwpoison(unsigned long pfn, int flags, bool *migratable_cleared) argument
3999 num_poisoned_pages_inc(unsigned long pfn) argument
4003 num_poisoned_pages_sub(unsigned long pfn, long i) argument
4018 memblk_nr_poison_inc(unsigned long pfn) argument
4022 memblk_nr_poison_sub(unsigned long pfn, long i) argument
4028 arch_memory_failure(unsigned long pfn, int flags) argument
4200 pfn_is_unaccepted_memory(unsigned long pfn) argument
[all...]
H A Dpgtable.h5 #include <linux/pfn.h>
1398 * memory type of pfn mappings specified by the remap_pfn_range,
1403 * track_pfn_remap is called when a _new_ pfn mapping is being established
1404 * by remap_pfn_range() for physical range indicated by pfn and size.
1407 unsigned long pfn, unsigned long addr,
1414 * track_pfn_insert is called when a _new_ single pfn is established
1418 pfn_t pfn)
1433 * untrack can be called for a specific region indicated by pfn and size or
1434 * can be for the entire vma (in which case pfn, size are zero).
1437 unsigned long pfn, unsigne
1406 track_pfn_remap(struct vm_area_struct *vma, pgprot_t *prot, unsigned long pfn, unsigned long addr, unsigned long size) argument
1417 track_pfn_insert(struct vm_area_struct *vma, pgprot_t *prot, pfn_t pfn) argument
1436 untrack_pfn(struct vm_area_struct *vma, unsigned long pfn, unsigned long size, bool mm_wr_locked) argument
1463 is_zero_pfn(unsigned long pfn) argument
1473 is_zero_pfn(unsigned long pfn) argument
1486 is_zero_pfn(unsigned long pfn) argument
1672 pfn_modify_allowed(unsigned long pfn, pgprot_t prot) argument
[all...]
H A Dio-mapping.h32 #include <linux/pfn.h>
229 unsigned long addr, unsigned long pfn, unsigned long size);
H A Dmigrate.h174 static inline unsigned long migrate_pfn(unsigned long pfn) argument
176 return (pfn << MIGRATE_PFN_SHIFT) | MIGRATE_PFN_VALID;
/linux-master/drivers/base/
H A Dmemory.c58 static inline unsigned long pfn_to_block_id(unsigned long pfn) argument
60 return memory_block_id(pfn_to_section_nr(pfn));
596 u64 pfn; local
599 if (kstrtoull(buf, 0, &pfn) < 0)
601 pfn >>= PAGE_SHIFT;
602 ret = soft_offline_page(pfn, 0);
612 u64 pfn; local
615 if (kstrtoull(buf, 0, &pfn) < 0)
617 pfn >>= PAGE_SHIFT;
618 ret = memory_failure(pfn, MF_SW_SIMULATE
1230 memblk_nr_poison_inc(unsigned long pfn) argument
1239 memblk_nr_poison_sub(unsigned long pfn, long i) argument
[all...]
/linux-master/arch/sparc/mm/
H A Diommu.c53 #define MKIOPTE(pfn, perm) (((((pfn)<<8) & IOPTE_PAGE) | (perm)) & ~IOPTE_WAZ)
191 unsigned long pfn = __phys_to_pfn(paddr); local
213 /* page color = pfn of page */
214 ioptex = bit_map_string_get(&iommu->usemap, npages, pfn);
223 iopte_val(*iopte) = MKIOPTE(pfn, IOPERM);
227 pfn++;
/linux-master/arch/riscv/mm/
H A Dpageattr.c113 unsigned long pfn = _pmd_pfn(pmdp_get(pmdp)); local
124 set_pte(ptep_new, pfn_pte(pfn + i, prot));
153 unsigned long pfn = _pud_pfn(pudp_get(pudp)); local
165 pfn_pmd(pfn + ((i * PMD_SIZE) >> PAGE_SHIFT), prot));
202 unsigned long pfn = _p4d_pfn(p4dp_get(p4dp)); local
218 pfn_pud(pfn + ((i * PUD_SIZE) >> PAGE_SHIFT), prot));
/linux-master/include/ras/
H A Dras_event.h344 * unsigned long pfn - Page Frame Number of the corrupted page
400 TP_PROTO(unsigned long pfn,
404 TP_ARGS(pfn, type, result),
407 __field(unsigned long, pfn)
413 __entry->pfn = pfn;
418 TP_printk("pfn %#lx: recovery action for %s: %s",
419 __entry->pfn,
/linux-master/mm/
H A Dmemory.c566 * to pfn. To get around this restriction, we allow arbitrary mappings so long
585 unsigned long pfn = pte_pfn(pte); local
594 if (is_zero_pfn(pfn))
615 if (!pfn_valid(pfn))
621 if (pfn == vma->vm_pgoff + off)
628 if (is_zero_pfn(pfn))
632 if (unlikely(pfn > highest_memmap_pfn)) {
642 return pfn_to_page(pfn);
659 unsigned long pfn = pmd_pfn(pmd); local
668 if (!pfn_valid(pfn))
2263 insert_pfn(struct vm_area_struct *vma, unsigned long addr, pfn_t pfn, pgprot_t prot, bool mkwrite) argument
2350 vmf_insert_pfn_prot(struct vm_area_struct *vma, unsigned long addr, unsigned long pfn, pgprot_t pgprot) argument
2398 vmf_insert_pfn(struct vm_area_struct *vma, unsigned long addr, unsigned long pfn) argument
2405 vm_mixed_ok(struct vm_area_struct *vma, pfn_t pfn) argument
2419 __vm_insert_mixed(struct vm_area_struct *vma, unsigned long addr, pfn_t pfn, bool mkwrite) argument
2465 vmf_insert_mixed(struct vm_area_struct *vma, unsigned long addr, pfn_t pfn) argument
2477 vmf_insert_mixed_mkwrite(struct vm_area_struct *vma, unsigned long addr, pfn_t pfn) argument
2489 remap_pte_range(struct mm_struct *mm, pmd_t *pmd, unsigned long addr, unsigned long end, unsigned long pfn, pgprot_t prot) argument
2515 remap_pmd_range(struct mm_struct *mm, pud_t *pud, unsigned long addr, unsigned long end, unsigned long pfn, pgprot_t prot) argument
2538 remap_pud_range(struct mm_struct *mm, p4d_t *p4d, unsigned long addr, unsigned long end, unsigned long pfn, pgprot_t prot) argument
2560 remap_p4d_range(struct mm_struct *mm, pgd_t *pgd, unsigned long addr, unsigned long end, unsigned long pfn, pgprot_t prot) argument
2586 remap_pfn_range_notrack(struct vm_area_struct *vma, unsigned long addr, unsigned long pfn, unsigned long size, pgprot_t prot) argument
2651 remap_pfn_range(struct vm_area_struct *vma, unsigned long addr, unsigned long pfn, unsigned long size, pgprot_t prot) argument
2684 unsigned long vm_len, pfn, pages; local
5941 follow_pfn(struct vm_area_struct *vma, unsigned long address, unsigned long *pfn) argument
[all...]
/linux-master/drivers/hv/
H A Dhv_balloon.c593 unsigned long pfn)
598 if ((pfn < has->covered_start_pfn) || (pfn >= has->covered_end_pfn))
603 if ((pfn >= gap->start_pfn) && (pfn < gap->end_pfn))
613 unsigned long pfn = start_pfn, count = 0; local
617 while (pfn < start_pfn + nr_pages) {
619 * Search for HAS which covers the pfn and when we find one
624 while ((pfn >= has->start_pfn) &&
625 (pfn < ha
592 has_pfn_is_backed(struct hv_hotadd_state *has, unsigned long pfn) argument
785 unsigned long pfn = page_to_pfn(pg); local
[all...]
/linux-master/drivers/nvdimm/
H A Dpmem.c32 #include "pfn.h"
68 unsigned long pfn_start, pfn_end, pfn; local
76 for (pfn = pfn_start; pfn < pfn_end; pfn++) {
77 struct page *page = pfn_to_page(pfn);
85 clear_mce_nospec(pfn);
245 pfn_t *pfn)
256 if (pfn)
257 *pfn
243 __pmem_direct_access(struct pmem_device *pmem, pgoff_t pgoff, long nr_pages, enum dax_access_mode mode, void **kaddr, pfn_t *pfn) argument
304 pmem_dax_direct_access(struct dax_device *dax_dev, pgoff_t pgoff, long nr_pages, enum dax_access_mode mode, void **kaddr, pfn_t *pfn) argument
434 pmem_pagemap_memory_failure(struct dev_pagemap *pgmap, unsigned long pfn, unsigned long nr_pages, int mf_flags) argument
[all...]
/linux-master/arch/arm/mm/
H A Dmmap.c164 int valid_mmap_phys_addr_range(unsigned long pfn, size_t size) argument
166 return (pfn + (size >> PAGE_SHIFT)) <= (1 + (PHYS_MASK >> PAGE_SHIFT));
/linux-master/tools/testing/scatterlist/linux/
H A Dmm.h53 #define pfn_to_page(pfn) (void *)((pfn) * PAGE_SIZE)
/linux-master/drivers/xen/
H A Dunpopulated-alloc.c109 xen_pfn_t pfn = PFN_DOWN(res->start); local
112 if (!set_phys_to_machine(pfn + i, INVALID_P2M_ENTRY)) {
H A Dgrant-dma-ops.c14 #include <linux/pfn.h>
84 unsigned long pfn; local
99 pfn = virt_to_pfn(ret);
108 pfn_to_gfn(pfn + i), 0);
/linux-master/include/xen/
H A Dxen-ops.h48 xen_pfn_t *pfn, int nr, int *err_ptr, pgprot_t prot,
52 xen_pfn_t *pfn, int nr, int *err_ptr,
51 xen_remap_pfn(struct vm_area_struct *vma, unsigned long addr, xen_pfn_t *pfn, int nr, int *err_ptr, pgprot_t prot, unsigned int domid, bool no_translate) argument
/linux-master/arch/arm/mach-pxa/
H A Dpxa25x.c163 .pfn = __phys_to_pfn(PXA2XX_SMEMC_BASE),
168 .pfn = __phys_to_pfn(0x00000000),
/linux-master/drivers/misc/sgi-gru/
H A Dgruhandles.c157 tfh->pfn = paddr >> GRU_PADDR_SHIFT;
173 tfh->pfn = paddr >> GRU_PADDR_SHIFT;
/linux-master/drivers/firmware/efi/
H A Dunaccepted_memory.c208 unsigned long pfn)
210 return !pfn_is_unaccepted_memory(pfn);
207 unaccepted_memory_vmcore_pfn_is_ram(struct vmcore_cb *cb, unsigned long pfn) argument
/linux-master/arch/sparc/include/asm/
H A Dpage_64.h149 #define pfn_to_kaddr(pfn) __va((pfn) << PAGE_SHIFT)
/linux-master/arch/sh/include/asm/
H A Dpage.h149 #define pfn_to_kaddr(pfn) __va((pfn) << PAGE_SHIFT)
/linux-master/kernel/debug/kdb/
H A Dkdb_support.c319 unsigned long pfn; local
323 pfn = (addr >> PAGE_SHIFT);
324 if (!pfn_valid(pfn))
326 page = pfn_to_page(pfn);
/linux-master/drivers/gpu/drm/gma500/
H A Dgem.c261 unsigned long pfn; local
294 pfn = (dev_priv->stolen_base + pobj->offset) >> PAGE_SHIFT;
296 pfn = page_to_pfn(pobj->pages[page_offset]);
297 ret = vmf_insert_pfn(vma, vmf->address, pfn);
/linux-master/drivers/net/ethernet/mellanox/mlx5/core/
H A Duar.c99 phys_addr_t pfn; local
135 pfn = uar2pfn(mdev, up->index);
137 up->map = ioremap_wc(pfn << PAGE_SHIFT, PAGE_SIZE);
143 up->map = ioremap(pfn << PAGE_SHIFT, PAGE_SIZE);
/linux-master/arch/arm/mach-sa1100/
H A Dgeneric.c364 .pfn = __phys_to_pfn(0x80000000),
369 .pfn = __phys_to_pfn(0x90000000),
374 .pfn = __phys_to_pfn(0xa0000000),
379 .pfn = __phys_to_pfn(0xb0000000),

Completed in 241 milliseconds

<<11121314151617181920>>