/linux-master/include/linux/ |
H A D | iova.h | 34 unsigned long start_pfn; /* Lower limit for this domain */ member in struct:iova_domain 96 unsigned long start_pfn); 149 unsigned long start_pfn) 147 init_iova_domain(struct iova_domain *iovad, unsigned long granule, unsigned long start_pfn) argument
|
H A D | memory.h | 88 int arch_get_memory_phys_device(unsigned long start_pfn); 110 unsigned long start_pfn; member in struct:memory_notify
|
/linux-master/arch/x86/platform/efi/ |
H A D | efi_32.c | 38 u64 start_pfn, end_pfn, end; local 42 start_pfn = PFN_DOWN(md->phys_addr); 47 if (pfn_range_is_mapped(start_pfn, end_pfn)) {
|
/linux-master/arch/mips/mm/ |
H A D | ioremap.c | 25 static int __ioremap_check_ram(unsigned long start_pfn, unsigned long nr_pages, argument 31 if (pfn_valid(start_pfn + i) && 32 !PageReserved(pfn_to_page(start_pfn + i)))
|
/linux-master/include/xen/ |
H A D | page.h | 37 unsigned long start_pfn; member in struct:xen_memory_region
|
/linux-master/arch/x86/xen/ |
H A D | enlighten.c | 358 void __init xen_add_extra_mem(unsigned long start_pfn, unsigned long n_pfns) argument 369 xen_extra_mem[i].start_pfn = start_pfn; 374 if (xen_extra_mem[i].start_pfn + xen_extra_mem[i].n_pfns == 375 start_pfn) { 383 memblock_reserve(PFN_PHYS(start_pfn), PFN_PHYS(n_pfns));
|
/linux-master/drivers/base/ |
H A D | arch_numa.c | 219 static void __init setup_node_data(int nid, u64 start_pfn, u64 end_pfn) argument 226 if (start_pfn >= end_pfn) 246 NODE_DATA(nid)->node_start_pfn = start_pfn; 247 NODE_DATA(nid)->node_spanned_pages = end_pfn - start_pfn; 363 unsigned long start_pfn, end_pfn; local 365 get_pfn_range_for_nid(nid, &start_pfn, &end_pfn); 366 setup_node_data(nid, start_pfn, end_pfn);
|
/linux-master/mm/ |
H A D | sparse.c | 129 static void __meminit mminit_validate_memmodel_limits(unsigned long *start_pfn, argument 138 if (*start_pfn > max_sparsemem_pfn) { 141 *start_pfn, *end_pfn, max_sparsemem_pfn); 143 *start_pfn = max_sparsemem_pfn; 148 *start_pfn, *end_pfn, max_sparsemem_pfn); 592 void online_mem_sections(unsigned long start_pfn, unsigned long end_pfn) argument 596 for (pfn = start_pfn; pfn < end_pfn; pfn += PAGES_PER_SECTION) { 610 void offline_mem_sections(unsigned long start_pfn, unsigned long end_pfn) argument 614 for (pfn = start_pfn; pfn < end_pfn; pfn += PAGES_PER_SECTION) { 874 * @start_pfn 890 sparse_add_section(int nid, unsigned long start_pfn, unsigned long nr_pages, struct vmem_altmap *altmap, struct dev_pagemap *pgmap) argument [all...] |
H A D | compaction.c | 262 static unsigned long skip_offline_sections(unsigned long start_pfn) argument 264 unsigned long start_nr = pfn_to_section_nr(start_pfn); 282 static unsigned long skip_offline_sections_reverse(unsigned long start_pfn) argument 284 unsigned long start_nr = pfn_to_section_nr(start_pfn); 297 static unsigned long skip_offline_sections(unsigned long start_pfn) argument 302 static unsigned long skip_offline_sections_reverse(unsigned long start_pfn) argument 608 unsigned long *start_pfn, 618 unsigned long blockpfn = *start_pfn; 710 trace_mm_compaction_isolate_freepages(*start_pfn, blockpfn, 714 *start_pfn 607 isolate_freepages_block(struct compact_control *cc, unsigned long *start_pfn, unsigned long end_pfn, struct list_head *freelist, unsigned int stride, bool strict) argument 745 isolate_freepages_range(struct compact_control *cc, unsigned long start_pfn, unsigned long end_pfn) argument 902 unsigned long start_pfn = low_pfn; local 1377 isolate_migratepages_range(struct compact_control *cc, unsigned long start_pfn, unsigned long end_pfn) argument 1517 unsigned long start_pfn, end_pfn; local 2528 unsigned long start_pfn = cc->zone->zone_start_pfn; local [all...] |
/linux-master/drivers/xen/ |
H A D | balloon.c | 342 unsigned long start_pfn = page_to_pfn(page); local 345 pr_debug("Online %lu pages starting at pfn 0x%lx\n", size, start_pfn); 348 p = pfn_to_page(start_pfn + i); 675 unsigned long start_pfn, pages; local 684 start_pfn = xen_extra_mem[i].start_pfn; 691 extra_pfn_end = min(max_pfn, start_pfn + pages); 693 for (pfn = start_pfn; pfn < extra_pfn_end; pfn++) 696 balloon_stats.total_pages += extra_pfn_end - start_pfn;
|
/linux-master/arch/loongarch/mm/ |
H A D | init.c | 95 unsigned long start_pfn = start >> PAGE_SHIFT; local 99 ret = __add_pages(nid, start_pfn, nr_pages, params); 110 unsigned long start_pfn = start >> PAGE_SHIFT; local 112 struct page *page = pfn_to_page(start_pfn); 117 __remove_pages(start_pfn, nr_pages, altmap);
|
/linux-master/arch/s390/mm/ |
H A D | init.c | 256 mem_data.start = arg->start_pfn << PAGE_SHIFT; 278 unsigned long start_pfn = PFN_DOWN(start); local 290 rc = __add_pages(nid, start_pfn, size_pages, params); 298 unsigned long start_pfn = start >> PAGE_SHIFT; local 301 __remove_pages(start_pfn, nr_pages, altmap);
|
/linux-master/arch/sh/kernel/ |
H A D | setup.c | 193 void __init __add_active_range(unsigned int nid, unsigned long start_pfn, argument 201 start = start_pfn << PAGE_SHIFT; 211 start_pfn, end_pfn); 235 memblock_set_node(PFN_PHYS(start_pfn), PFN_PHYS(end_pfn - start_pfn),
|
/linux-master/drivers/gpu/drm/gma500/ |
H A D | mmu.h | 71 uint32_t start_pfn,
|
/linux-master/arch/x86/virt/vmx/tdx/ |
H A D | tdx.c | 185 static int add_tdx_memblock(struct list_head *tmb_list, unsigned long start_pfn, argument 195 tmb->start_pfn = start_pfn; 224 unsigned long start_pfn, end_pfn; local 227 for_each_mem_pfn_range(i, MAX_NUMNODES, &start_pfn, &end_pfn, &nid) { 234 start_pfn = max(start_pfn, PHYS_PFN(SZ_1M)); 235 if (start_pfn >= end_pfn) 243 ret = add_tdx_memblock(tmb_list, start_pfn, end_pfn, nid); 426 start = TDMR_ALIGN_DOWN(PFN_PHYS(tmb->start_pfn)); 1372 is_tdx_memory(unsigned long start_pfn, unsigned long end_pfn) argument [all...] |
H A D | tdx.h | 97 unsigned long start_pfn; member in struct:tdx_memblock
|
/linux-master/arch/x86/include/asm/ |
H A D | page_types.h | 66 bool pfn_range_is_mapped(unsigned long start_pfn, unsigned long end_pfn);
|
/linux-master/arch/loongarch/kernel/ |
H A D | numa.c | 213 unsigned long start_pfn, end_pfn; local 220 get_pfn_range_for_nid(node, &start_pfn, &end_pfn); 221 pr_info("Node%d: start_pfn=0x%lx, end_pfn=0x%lx\n", 222 node, start_pfn, end_pfn); 236 pr_info(" start_pfn:0x%llx, end_pfn:0x%llx, num_physpages:0x%lx\n", 397 unsigned long start_pfn, end_pfn; local 399 get_pfn_range_for_nid(node, &start_pfn, &end_pfn);
|
/linux-master/arch/powerpc/mm/ |
H A D | init_64.c | 76 unsigned long start_pfn; local 80 start_pfn = (offset / sizeof(struct page)) & PAGE_SUBSECTION_MASK; 81 return pfn_to_page(start_pfn); 191 unsigned long start_pfn = page_to_pfn((struct page *)start); local 193 if ((start_pfn + nr_pfn - 1) > altmap->end_pfn) 196 if (start_pfn < altmap->base_pfn)
|
H A D | numa.c | 1031 unsigned long start_pfn, end_pfn; local 1038 for_each_mem_pfn_range(i, MAX_NUMNODES, &start_pfn, &end_pfn, NULL) { 1040 memblock_set_node(PFN_PHYS(start_pfn), 1041 PFN_PHYS(end_pfn - start_pfn), 1083 static void __init setup_node_data(int nid, u64 start_pfn, u64 end_pfn) argument 1085 u64 spanned_pages = end_pfn - start_pfn; 1108 NODE_DATA(nid)->node_start_pfn = start_pfn; 1218 unsigned long start_pfn, end_pfn; local 1220 get_pfn_range_for_nid(nid, &start_pfn, &end_pfn); 1221 setup_node_data(nid, start_pfn, end_pf [all...] |
/linux-master/arch/mips/sgi-ip27/ |
H A D | ip27-memory.c | 355 unsigned long start_pfn, end_pfn; local 357 get_pfn_range_for_nid(node, &start_pfn, &end_pfn); 365 NODE_DATA(node)->node_start_pfn = start_pfn; 366 NODE_DATA(node)->node_spanned_pages = end_pfn - start_pfn;
|
/linux-master/arch/s390/include/asm/ |
H A D | diag.h | 57 static inline void diag10_range(unsigned long start_pfn, unsigned long num_pfn) argument 61 start_addr = pfn_to_phys(start_pfn); 62 end_addr = pfn_to_phys(start_pfn + num_pfn - 1);
|
/linux-master/drivers/iommu/intel/ |
H A D | iommu.c | 933 unsigned long start_pfn, 940 WARN_ON(start_pfn > last_pfn)) 946 first_pte = pte = dma_pfn_level_pte(domain, start_pfn, 1, &large_page); 948 start_pfn = align_to_level(start_pfn + 1, large_page + 1); 953 start_pfn += lvl_to_nr_pages(large_page); 955 } while (start_pfn <= last_pfn && !first_pte_in_page(pte)); 960 } while (start_pfn && start_pfn <= last_pfn); 965 unsigned long pfn, unsigned long start_pfn, 932 dma_pte_clear_range(struct dmar_domain *domain, unsigned long start_pfn, unsigned long last_pfn) argument 963 dma_pte_free_level(struct dmar_domain *domain, int level, int retain_level, struct dma_pte *pte, unsigned long pfn, unsigned long start_pfn, unsigned long last_pfn) argument 1006 dma_pte_free_pagetable(struct dmar_domain *domain, unsigned long start_pfn, unsigned long last_pfn, int retain_level) argument 1050 dma_pte_clear_level(struct dmar_domain *domain, int level, struct dma_pte *pte, unsigned long pfn, unsigned long start_pfn, unsigned long last_pfn, struct list_head *freelist) argument 1097 domain_unmap(struct dmar_domain *domain, unsigned long start_pfn, unsigned long last_pfn, struct list_head *freelist) argument 2029 switch_to_super_page(struct dmar_domain *domain, unsigned long start_pfn, unsigned long end_pfn, int level) argument 2282 unsigned long start_pfn, end_pfn; local 4081 unsigned long start_pfn, last_pfn; local 4129 unsigned long start_pfn; local [all...] |
/linux-master/kernel/power/ |
H A D | snapshot.c | 396 unsigned long start_pfn; /* Zone start page frame */ member in struct:mem_zone_bm_rtree 555 zone->start_pfn = start; 757 if (pfn >= zone->start_pfn && pfn < zone->end_pfn) 764 if (pfn >= curr->start_pfn && pfn < curr->end_pfn) { 786 ((pfn - zone->start_pfn) & ~BM_BLOCK_MASK) == bm->cur.node_pfn) 790 block_nr = (pfn - zone->start_pfn) >> BM_BLOCK_SHIFT; 805 bm->cur.node_pfn = (pfn - zone->start_pfn) & ~BM_BLOCK_MASK; 810 *bit_nr = (pfn - zone->start_pfn) & BM_BLOCK_MASK; 935 pages = bm->cur.zone->end_pfn - bm->cur.zone->start_pfn; 940 pfn = bm->cur.zone->start_pfn 957 unsigned long start_pfn; member in struct:nosave_region 997 register_nosave_region(unsigned long start_pfn, unsigned long end_pfn) argument [all...] |
/linux-master/arch/x86/mm/ |
H A D | ioremap.c | 72 unsigned long start_pfn, stop_pfn; local 78 start_pfn = (res->start + PAGE_SIZE - 1) >> PAGE_SHIFT; 80 if (stop_pfn > start_pfn) { 81 for (i = 0; i < (stop_pfn - start_pfn); ++i) 82 if (pfn_valid(start_pfn + i) && 83 !PageReserved(pfn_to_page(start_pfn + i)))
|