/linux-master/include/trace/events/ |
H A D | page_isolation.h | 14 unsigned long end_pfn, 17 TP_ARGS(start_pfn, end_pfn, fin_pfn), 21 __field(unsigned long, end_pfn) 27 __entry->end_pfn = end_pfn; 31 TP_printk("start_pfn=0x%lx end_pfn=0x%lx fin_pfn=0x%lx ret=%s", 32 __entry->start_pfn, __entry->end_pfn, __entry->fin_pfn, 33 __entry->end_pfn <= __entry->fin_pfn ? "success" : "fail")
|
H A D | compaction.h | 18 unsigned long end_pfn, 22 TP_ARGS(start_pfn, end_pfn, nr_scanned, nr_taken), 26 __field(unsigned long, end_pfn) 33 __entry->end_pfn = end_pfn; 40 __entry->end_pfn, 49 unsigned long end_pfn, 53 TP_ARGS(start_pfn, end_pfn, nr_scanned, nr_taken) 60 unsigned long end_pfn, 64 TP_ARGS(start_pfn, end_pfn, nr_scanne [all...] |
/linux-master/include/linux/ |
H A D | page-isolation.h | 40 int start_isolate_page_range(unsigned long start_pfn, unsigned long end_pfn, 43 void undo_isolate_page_range(unsigned long start_pfn, unsigned long end_pfn, 46 int test_pages_isolated(unsigned long start_pfn, unsigned long end_pfn,
|
H A D | node.h | 109 unsigned long end_pfn, 113 unsigned long end_pfn, 133 unsigned long end_pfn = start_pfn + pgdat->node_spanned_pages; local 138 register_memory_blocks_under_node(nid, start_pfn, end_pfn, 112 register_memory_blocks_under_node(int nid, unsigned long start_pfn, unsigned long end_pfn, enum meminit_context context) argument
|
/linux-master/arch/sh/mm/ |
H A D | numa.c | 28 unsigned long start_pfn, end_pfn; local 34 end_pfn = PFN_DOWN(end); 41 __add_active_range(nid, start_pfn, end_pfn); 52 NODE_DATA(nid)->node_spanned_pages = end_pfn - start_pfn;
|
H A D | init.c | 210 unsigned long start_pfn, end_pfn; local 212 get_pfn_range_for_nid(nid, &start_pfn, &end_pfn); 224 NODE_DATA(nid)->node_spanned_pages = end_pfn - start_pfn; 229 unsigned long start_pfn, end_pfn; local 233 for_each_mem_pfn_range(i, MAX_NUMNODES, &start_pfn, &end_pfn, NULL) 234 __add_active_range(0, start_pfn, end_pfn);
|
/linux-master/arch/sh/kernel/ |
H A D | swsusp.c | 22 unsigned long end_pfn = PAGE_ALIGN(__pa(&__nosave_end)) >> PAGE_SHIFT; local 24 return (pfn >= begin_pfn) && (pfn < end_pfn);
|
/linux-master/arch/x86/mm/ |
H A D | init.c | 328 unsigned long start_pfn, unsigned long end_pfn, 331 if (start_pfn < end_pfn) { 335 mr[nr_range].end = end_pfn<<PAGE_SHIFT; 406 unsigned long start_pfn, end_pfn, limit_pfn; local 422 end_pfn = PFN_DOWN(PMD_SIZE); 424 end_pfn = round_up(pfn, PFN_DOWN(PMD_SIZE)); 426 end_pfn = round_up(pfn, PFN_DOWN(PMD_SIZE)); 428 if (end_pfn > limit_pfn) 429 end_pfn = limit_pfn; 430 if (start_pfn < end_pfn) { 327 save_mr(struct map_range *mr, int nr_range, unsigned long start_pfn, unsigned long end_pfn, unsigned long page_size_mask) argument 505 add_pfn_range_mapped(unsigned long start_pfn, unsigned long end_pfn) argument 518 pfn_range_is_mapped(unsigned long start_pfn, unsigned long end_pfn) argument 575 unsigned long start_pfn, end_pfn; local 1005 unsigned long start_pfn, end_pfn; local [all...] |
/linux-master/mm/ |
H A D | page_idle.c | 120 unsigned long pfn, end_pfn; local 130 end_pfn = pfn + count * BITS_PER_BYTE; 131 if (end_pfn > max_pfn) 132 end_pfn = max_pfn; 134 for (; pfn < end_pfn; pfn++) { 165 unsigned long pfn, end_pfn; local 175 end_pfn = pfn + count * BITS_PER_BYTE; 176 if (end_pfn > max_pfn) 177 end_pfn = max_pfn; 179 for (; pfn < end_pfn; pf [all...] |
H A D | page_isolation.c | 19 * This function checks whether the range [start_pfn, end_pfn) includes 33 static struct page *has_unmovable_pages(unsigned long start_pfn, unsigned long end_pfn, argument 41 pageblock_start_pfn(end_pfn - 1)); 55 for (pfn = start_pfn; pfn < end_pfn; pfn++) { 144 * present in [start_pfn, end_pfn). The pageblock must intersect with 145 * [start_pfn, end_pfn). 148 unsigned long start_pfn, unsigned long end_pfn) 171 * Pass the intersection of [start_pfn, end_pfn) and the page's pageblock 176 end_pfn); 487 * @end_pfn 147 set_migratetype_isolate(struct page *page, int migratetype, int isol_flags, unsigned long start_pfn, unsigned long end_pfn) argument 529 start_isolate_page_range(unsigned long start_pfn, unsigned long end_pfn, int migratetype, int flags, gfp_t gfp_flags) argument 583 undo_isolate_page_range(unsigned long start_pfn, unsigned long end_pfn, int migratetype) argument 608 __test_page_isolated_in_pageblock(unsigned long pfn, unsigned long end_pfn, int flags) argument 655 test_pages_isolated(unsigned long start_pfn, unsigned long end_pfn, int isol_flags) argument [all...] |
H A D | mm_init.c | 300 unsigned long start_pfn, end_pfn; local 303 for_each_mem_pfn_range(i, MAX_NUMNODES, &start_pfn, &end_pfn, &nid) { 304 unsigned long pages = end_pfn - start_pfn; 462 unsigned long start_pfn, end_pfn; local 480 for_each_mem_pfn_range(i, nid, &start_pfn, &end_pfn, NULL) { 484 if (start_pfn >= end_pfn) 490 kernel_pages = min(end_pfn, usable_startpfn) 499 if (end_pfn <= usable_startpfn) { 507 zone_movable_pfn[nid] = end_pfn; 515 * start_pfn->end_pfn 549 unsigned long start_pfn, end_pfn; local 602 unsigned long start_pfn, end_pfn; local 672 defer_init(int nid, unsigned long pfn, unsigned long end_pfn) argument 732 defer_init(int nid, unsigned long pfn, unsigned long end_pfn) argument 752 unsigned long end_pfn = PFN_UP(end); local 854 unsigned long pfn, end_pfn = start_pfn + size; local 910 memmap_init_zone_range(struct zone *zone, unsigned long start_pfn, unsigned long end_pfn, unsigned long *hole_pfn) argument 936 unsigned long start_pfn, end_pfn; local 1043 unsigned long pfn, end_pfn = head_pfn + nr_pages; local 1069 unsigned long pfn, end_pfn = start_pfn + nr_pages; local 1152 unsigned long start_pfn, end_pfn; local 1170 absent_pages_in_range(unsigned long start_pfn, unsigned long end_pfn) argument 1196 unsigned long start_pfn, end_pfn; local 1685 get_pfn_range_for_nid(unsigned int nid, unsigned long *start_pfn, unsigned long *end_pfn) argument 1707 unsigned long end_pfn = 0; local 1791 unsigned long start_pfn, end_pfn; local 2014 deferred_free_pages(unsigned long pfn, unsigned long end_pfn) argument 2039 deferred_init_pages(struct zone *zone, unsigned long pfn, unsigned long end_pfn) argument 2104 deferred_init_maxorder(u64 *i, struct zone *zone, unsigned long *start_pfn, unsigned long *end_pfn) argument 2148 deferred_init_memmap_chunk(unsigned long start_pfn, unsigned long end_pfn, void *arg) argument [all...] |
H A D | bootmem_info.c | 104 unsigned long i, pfn, end_pfn, nr_pages; local 115 end_pfn = pgdat_end_pfn(pgdat); 118 for (; pfn < end_pfn; pfn += PAGES_PER_SECTION) {
|
H A D | memory_hotplug.c | 372 const unsigned long end_pfn = pfn + nr_pages; local 399 for (; pfn < end_pfn; pfn += cur_nr_pages) { 401 cur_nr_pages = min(end_pfn - pfn, 413 /* find the smallest valid pfn in the range [start_pfn, end_pfn) */ 416 unsigned long end_pfn) 418 for (; start_pfn < end_pfn; start_pfn += PAGES_PER_SUBSECTION) { 434 /* find the biggest valid pfn in the range [start_pfn, end_pfn). */ 437 unsigned long end_pfn) 442 pfn = end_pfn - 1; 460 unsigned long end_pfn) 414 find_smallest_section_pfn(int nid, struct zone *zone, unsigned long start_pfn, unsigned long end_pfn) argument 435 find_biggest_section_pfn(int nid, struct zone *zone, unsigned long start_pfn, unsigned long end_pfn) argument 459 shrink_zone_span(struct zone *zone, unsigned long start_pfn, unsigned long end_pfn) argument 506 unsigned long end_pfn = zone_end_pfn(zone); local 531 const unsigned long end_pfn = start_pfn + nr_pages; local 576 const unsigned long end_pfn = pfn + nr_pages; local 646 const unsigned long end_pfn = start_pfn + nr_pages; local 968 unsigned long online_pages = 0, max_pages, end_pfn; local 1092 unsigned long end_pfn = pfn + nr_pages; local 1126 unsigned long end_pfn = pfn + nr_pages; local 1775 do_migrate_range(unsigned long start_pfn, unsigned long end_pfn) argument 1946 const unsigned long end_pfn = start_pfn + nr_pages; local [all...] |
H A D | shuffle.c | 71 * the span, end_pfn - start_pfn. 84 unsigned long end_pfn = zone_end_pfn(z); local 90 for (i = start_pfn; i < end_pfn; i += order_pages) { 97 * (@start_pfn to @end_pfn), to more likely be valid compared to
|
/linux-master/arch/sh/include/asm/ |
H A D | mmzone.h | 41 unsigned long end_pfn);
|
/linux-master/arch/x86/include/asm/ |
H A D | highmem.h | 73 unsigned long end_pfn);
|
H A D | mtrr.h | 73 extern int mtrr_trim_uncached_memory(unsigned long end_pfn); 115 static inline int mtrr_trim_uncached_memory(unsigned long end_pfn) argument
|
/linux-master/arch/mips/loongson64/ |
H A D | numa.c | 88 unsigned long start_pfn, end_pfn; local 97 get_pfn_range_for_nid(node, &start_pfn, &end_pfn); 98 pr_info("Node%d: start_pfn=0x%lx, end_pfn=0x%lx\n", 99 node, start_pfn, end_pfn); 112 NODE_DATA(node)->node_spanned_pages = end_pfn - start_pfn; 122 max_low_pfn = end_pfn;
|
/linux-master/arch/sparc/mm/ |
H A D | init_32.c | 65 unsigned long end_pfn = (sp_banks[i].base_addr + sp_banks[i].num_bytes) >> PAGE_SHIFT; local 67 if (end_pfn <= max_low_pfn) 73 nr += end_pfn - start_pfn; 235 static void map_high_region(unsigned long start_pfn, unsigned long end_pfn) argument 240 printk("mapping high region %08lx - %08lx\n", start_pfn, end_pfn); 243 for (tmp = start_pfn; tmp < end_pfn; tmp++) 284 unsigned long end_pfn = (sp_banks[i].base_addr + sp_banks[i].num_bytes) >> PAGE_SHIFT; local 286 if (end_pfn <= highstart_pfn) 292 map_high_region(start_pfn, end_pfn);
|
/linux-master/arch/x86/platform/efi/ |
H A D | efi_32.c | 38 u64 start_pfn, end_pfn, end; local 45 end_pfn = PFN_UP(end); 47 if (pfn_range_is_mapped(start_pfn, end_pfn)) {
|
/linux-master/arch/loongarch/kernel/ |
H A D | numa.c | 213 unsigned long start_pfn, end_pfn; local 220 get_pfn_range_for_nid(node, &start_pfn, &end_pfn); 221 pr_info("Node%d: start_pfn=0x%lx, end_pfn=0x%lx\n", 222 node, start_pfn, end_pfn); 236 pr_info(" start_pfn:0x%llx, end_pfn:0x%llx, num_physpages:0x%lx\n", 397 unsigned long start_pfn, end_pfn; local 399 get_pfn_range_for_nid(node, &start_pfn, &end_pfn); 401 if (end_pfn > max_low_pfn) 402 max_low_pfn = end_pfn;
|
/linux-master/drivers/gpu/drm/imagination/ |
H A D | pvr_vm_mips.c | 156 s32 end_pfn; local 172 end_pfn = (end & fw_dev->fw_heap_info.offset_mask) >> ROGUE_MIPSFW_LOG2_PAGE_SIZE_4K; 181 for (pfn = start_pfn; pfn <= end_pfn; pfn++) { 229 const u32 end_pfn = (end & fw_dev->fw_heap_info.offset_mask) >> local 232 for (u32 pfn = start_pfn; pfn < end_pfn; pfn++)
|
/linux-master/drivers/base/ |
H A D | arch_numa.c | 219 static void __init setup_node_data(int nid, u64 start_pfn, u64 end_pfn) argument 226 if (start_pfn >= end_pfn) 247 NODE_DATA(nid)->node_spanned_pages = end_pfn - start_pfn; 363 unsigned long start_pfn, end_pfn; local 365 get_pfn_range_for_nid(nid, &start_pfn, &end_pfn); 366 setup_node_data(nid, start_pfn, end_pfn);
|
/linux-master/arch/powerpc/mm/ |
H A D | mem.c | 117 unsigned long end_pfn = PFN_UP(start + size); local 119 if (end_pfn > max_pfn) { 120 max_pfn = end_pfn; 121 max_low_pfn = end_pfn;
|
/linux-master/arch/x86/virt/vmx/tdx/ |
H A D | tdx.h | 98 unsigned long end_pfn; member in struct:tdx_memblock
|