/linux-master/include/trace/events/ |
H A D | page_isolation.h | 13 unsigned long start_pfn, 17 TP_ARGS(start_pfn, end_pfn, fin_pfn), 20 __field(unsigned long, start_pfn) 26 __entry->start_pfn = start_pfn; 31 TP_printk("start_pfn=0x%lx end_pfn=0x%lx fin_pfn=0x%lx ret=%s", 32 __entry->start_pfn, __entry->end_pfn, __entry->fin_pfn,
|
H A D | compaction.h | 17 unsigned long start_pfn, 22 TP_ARGS(start_pfn, end_pfn, nr_scanned, nr_taken), 25 __field(unsigned long, start_pfn) 32 __entry->start_pfn = start_pfn; 39 __entry->start_pfn, 48 unsigned long start_pfn, 53 TP_ARGS(start_pfn, end_pfn, nr_scanned, nr_taken) 59 unsigned long start_pfn, 64 TP_ARGS(start_pfn, end_pf [all...] |
/linux-master/arch/sh/mm/ |
H A D | numa.c | 28 unsigned long start_pfn, end_pfn; local 33 start_pfn = PFN_DOWN(start); 41 __add_active_range(nid, start_pfn, end_pfn); 51 NODE_DATA(nid)->node_start_pfn = start_pfn; 52 NODE_DATA(nid)->node_spanned_pages = end_pfn - start_pfn;
|
H A D | init.c | 210 unsigned long start_pfn, end_pfn; local 212 get_pfn_range_for_nid(nid, &start_pfn, &end_pfn); 223 NODE_DATA(nid)->node_start_pfn = start_pfn; 224 NODE_DATA(nid)->node_spanned_pages = end_pfn - start_pfn; 229 unsigned long start_pfn, end_pfn; local 233 for_each_mem_pfn_range(i, MAX_NUMNODES, &start_pfn, &end_pfn, NULL) 234 __add_active_range(0, start_pfn, end_pfn); 247 unsigned long start_pfn; local 255 start_pfn = PFN_UP(__pa(_end)); 263 memblock_reserve(start, (PFN_PHYS(start_pfn) 403 unsigned long start_pfn = PFN_DOWN(start); local 420 unsigned long start_pfn = PFN_DOWN(start); local [all...] |
/linux-master/include/linux/ |
H A D | page-isolation.h | 40 int start_isolate_page_range(unsigned long start_pfn, unsigned long end_pfn, 43 void undo_isolate_page_range(unsigned long start_pfn, unsigned long end_pfn, 46 int test_pages_isolated(unsigned long start_pfn, unsigned long end_pfn,
|
H A D | node.h | 108 void register_memory_blocks_under_node(int nid, unsigned long start_pfn, 112 static inline void register_memory_blocks_under_node(int nid, unsigned long start_pfn, argument 132 unsigned long start_pfn = pgdat->node_start_pfn; local 133 unsigned long end_pfn = start_pfn + pgdat->node_spanned_pages; 138 register_memory_blocks_under_node(nid, start_pfn, end_pfn,
|
H A D | memory_hotplug.h | 178 extern void __offline_isolated_pages(unsigned long start_pfn, 205 extern void __remove_pages(unsigned long start_pfn, unsigned long nr_pages, 209 extern int __add_pages(int nid, unsigned long start_pfn, unsigned long nr_pages, 213 static inline int add_pages(int nid, unsigned long start_pfn, argument 216 return __add_pages(nid, start_pfn, nr_pages, params); 219 int add_pages(int nid, unsigned long start_pfn, unsigned long nr_pages, 330 extern int offline_pages(unsigned long start_pfn, unsigned long nr_pages, 339 static inline int offline_pages(unsigned long start_pfn, unsigned long nr_pages, argument 362 extern void move_pfn_range_to_zone(struct zone *zone, unsigned long start_pfn, 366 unsigned long start_pfn, [all...] |
/linux-master/arch/powerpc/platforms/powernv/ |
H A D | memtrace.c | 91 static void memtrace_clear_range(unsigned long start_pfn, argument 97 for (pfn = start_pfn; pfn < start_pfn + nr_pages; pfn++) { 106 flush_dcache_range_chunked((unsigned long)pfn_to_kaddr(start_pfn), 107 (unsigned long)pfn_to_kaddr(start_pfn + nr_pages), 114 unsigned long pfn, start_pfn; local 125 start_pfn = page_to_pfn(page); 132 memtrace_clear_range(start_pfn, nr_pages); 138 for (pfn = start_pfn; pfn < start_pfn 217 const unsigned long start_pfn = PHYS_PFN(start); local [all...] |
/linux-master/mm/ |
H A D | page_isolation.c | 19 * This function checks whether the range [start_pfn, end_pfn) includes 33 static struct page *has_unmovable_pages(unsigned long start_pfn, unsigned long end_pfn, argument 36 struct page *page = pfn_to_page(start_pfn); 40 VM_BUG_ON(pageblock_start_pfn(start_pfn) != 55 for (pfn = start_pfn; pfn < end_pfn; pfn++) { 144 * present in [start_pfn, end_pfn). The pageblock must intersect with 145 * [start_pfn, end_pfn). 148 unsigned long start_pfn, unsigned long end_pfn) 171 * Pass the intersection of [start_pfn, end_pfn) and the page's pageblock 174 check_unmovable_start = max(page_to_pfn(page), start_pfn); 147 set_migratetype_isolate(struct page *page, int migratetype, int isol_flags, unsigned long start_pfn, unsigned long end_pfn) argument 310 unsigned long start_pfn; local 529 start_isolate_page_range(unsigned long start_pfn, unsigned long end_pfn, int migratetype, int flags, gfp_t gfp_flags) argument 583 undo_isolate_page_range(unsigned long start_pfn, unsigned long end_pfn, int migratetype) argument 655 test_pages_isolated(unsigned long start_pfn, unsigned long end_pfn, int isol_flags) argument [all...] |
H A D | mm_init.c | 300 unsigned long start_pfn, end_pfn; local 303 for_each_mem_pfn_range(i, MAX_NUMNODES, &start_pfn, &end_pfn, &nid) { 304 unsigned long pages = end_pfn - start_pfn; 462 unsigned long start_pfn, end_pfn; local 480 for_each_mem_pfn_range(i, nid, &start_pfn, &end_pfn, NULL) { 483 start_pfn = max(start_pfn, zone_movable_pfn[nid]); 484 if (start_pfn >= end_pfn) 488 if (start_pfn < usable_startpfn) { 491 - start_pfn; 549 unsigned long start_pfn, end_pfn; local 602 unsigned long start_pfn, end_pfn; local 751 unsigned long start_pfn = PFN_DOWN(start); local 849 memmap_init_range(unsigned long size, int nid, unsigned long zone, unsigned long start_pfn, unsigned long zone_end_pfn, enum meminit_context context, struct vmem_altmap *altmap, int migratetype) argument 910 memmap_init_zone_range(struct zone *zone, unsigned long start_pfn, unsigned long end_pfn, unsigned long *hole_pfn) argument 936 unsigned long start_pfn, end_pfn; local 1064 memmap_init_zone_device(struct zone *zone, unsigned long start_pfn, unsigned long nr_pages, struct dev_pagemap *pgmap) argument 1152 unsigned long start_pfn, end_pfn; local 1170 absent_pages_in_range(unsigned long start_pfn, unsigned long end_pfn) argument 1196 unsigned long start_pfn, end_pfn; local 1685 get_pfn_range_for_nid(unsigned int nid, unsigned long *start_pfn, unsigned long *end_pfn) argument 1706 unsigned long start_pfn = 0; local 1791 unsigned long start_pfn, end_pfn; local 2104 deferred_init_maxorder(u64 *i, struct zone *zone, unsigned long *start_pfn, unsigned long *end_pfn) argument 2148 deferred_init_memmap_chunk(unsigned long start_pfn, unsigned long end_pfn, void *arg) argument [all...] |
H A D | page_ext.c | 354 static int __meminit online_page_ext(unsigned long start_pfn, argument 361 start = SECTION_ALIGN_DOWN(start_pfn); 362 end = SECTION_ALIGN_UP(start_pfn + nr_pages); 367 * "start_pfn" passed to us is a pfn which is an arg for 368 * online__pages(), and start_pfn should exist. 370 nid = pfn_to_nid(start_pfn); 387 static void __meminit offline_page_ext(unsigned long start_pfn, argument 392 start = SECTION_ALIGN_DOWN(start_pfn); 393 end = SECTION_ALIGN_UP(start_pfn + nr_pages); 421 ret = online_page_ext(mn->start_pfn, 451 unsigned long start_pfn, end_pfn; local [all...] |
H A D | memory_hotplug.c | 413 /* find the smallest valid pfn in the range [start_pfn, end_pfn) */ 415 unsigned long start_pfn, 418 for (; start_pfn < end_pfn; start_pfn += PAGES_PER_SUBSECTION) { 419 if (unlikely(!pfn_to_online_page(start_pfn))) 422 if (unlikely(pfn_to_nid(start_pfn) != nid)) 425 if (zone != page_zone(pfn_to_page(start_pfn))) 428 return start_pfn; 434 /* find the biggest valid pfn in the range [start_pfn, end_pfn). */ 436 unsigned long start_pfn, 414 find_smallest_section_pfn(int nid, struct zone *zone, unsigned long start_pfn, unsigned long end_pfn) argument 435 find_biggest_section_pfn(int nid, struct zone *zone, unsigned long start_pfn, unsigned long end_pfn) argument 459 shrink_zone_span(struct zone *zone, unsigned long start_pfn, unsigned long end_pfn) argument 527 remove_pfn_range_from_zone(struct zone *zone, unsigned long start_pfn, unsigned long nr_pages) argument 644 online_pages_range(unsigned long start_pfn, unsigned long nr_pages) argument 705 resize_zone_range(struct zone *zone, unsigned long start_pfn, unsigned long nr_pages) argument 716 resize_pgdat_range(struct pglist_data *pgdat, unsigned long start_pfn, unsigned long nr_pages) argument 750 move_pfn_range_to_zone(struct zone *zone, unsigned long start_pfn, unsigned long nr_pages, struct vmem_altmap *altmap, int migratetype) argument 897 default_kernel_zone_for_pfn(int nid, unsigned long start_pfn, unsigned long nr_pages) argument 1024 default_zone_for_pfn(int nid, unsigned long start_pfn, unsigned long nr_pages) argument 1048 zone_for_pfn_range(int online_type, int nid, struct memory_group *group, unsigned long start_pfn, unsigned long nr_pages) argument 1775 do_migrate_range(unsigned long start_pfn, unsigned long end_pfn) argument 1931 count_system_ram_pages_cb(unsigned long start_pfn, unsigned long nr_pages, void *data) argument 1943 offline_pages(unsigned long start_pfn, unsigned long nr_pages, struct zone *zone, struct memory_group *group) argument [all...] |
H A D | shuffle.c | 71 * the span, end_pfn - start_pfn. 83 unsigned long start_pfn = z->zone_start_pfn; local 89 start_pfn = ALIGN(start_pfn, order_pages); 90 for (i = start_pfn; i < end_pfn; i += order_pages) { 97 * (@start_pfn to @end_pfn), to more likely be valid compared to
|
H A D | bootmem_info.c | 43 static void __init register_page_bootmem_info_section(unsigned long start_pfn) argument 50 section_nr = pfn_to_section_nr(start_pfn); 78 static void __init register_page_bootmem_info_section(unsigned long start_pfn) argument 85 section_nr = pfn_to_section_nr(start_pfn);
|
/linux-master/arch/x86/mm/ |
H A D | init.c | 328 unsigned long start_pfn, unsigned long end_pfn, 331 if (start_pfn < end_pfn) { 334 mr[nr_range].start = start_pfn<<PAGE_SHIFT; 406 unsigned long start_pfn, end_pfn, limit_pfn; local 413 pfn = start_pfn = PFN_DOWN(start); 430 if (start_pfn < end_pfn) { 431 nr_range = save_mr(mr, nr_range, start_pfn, end_pfn, 0); 436 start_pfn = round_up(pfn, PFN_DOWN(PMD_SIZE)); 445 if (start_pfn < end_pfn) { 446 nr_range = save_mr(mr, nr_range, start_pfn, end_pf 327 save_mr(struct map_range *mr, int nr_range, unsigned long start_pfn, unsigned long end_pfn, unsigned long page_size_mask) argument 505 add_pfn_range_mapped(unsigned long start_pfn, unsigned long end_pfn) argument 518 pfn_range_is_mapped(unsigned long start_pfn, unsigned long end_pfn) argument 575 unsigned long start_pfn, end_pfn; local 1005 unsigned long start_pfn, end_pfn; local [all...] |
/linux-master/arch/sparc/mm/ |
H A D | init_32.c | 64 unsigned long start_pfn = sp_banks[i].base_addr >> PAGE_SHIFT; local 70 if (start_pfn < max_low_pfn) 71 start_pfn = max_low_pfn; 73 nr += end_pfn - start_pfn; 133 unsigned long start_pfn, bytes_avail, size; local 171 start_pfn = (unsigned long)__pa(PAGE_ALIGN((unsigned long) &_end)); 174 start_pfn >>= PAGE_SHIFT; 192 size = (start_pfn << PAGE_SHIFT) - phys_base; 235 static void map_high_region(unsigned long start_pfn, unsigned long end_pfn) argument 240 printk("mapping high region %08lx - %08lx\n", start_pfn, end_pf 283 unsigned long start_pfn = sp_banks[i].base_addr >> PAGE_SHIFT; local [all...] |
/linux-master/arch/mips/loongson64/ |
H A D | numa.c | 88 unsigned long start_pfn, end_pfn; local 97 get_pfn_range_for_nid(node, &start_pfn, &end_pfn); 98 pr_info("Node%d: start_pfn=0x%lx, end_pfn=0x%lx\n", 99 node, start_pfn, end_pfn); 111 NODE_DATA(node)->node_start_pfn = start_pfn; 112 NODE_DATA(node)->node_spanned_pages = end_pfn - start_pfn; 134 memblock_reserve(0, PAGE_SIZE * start_pfn);
|
/linux-master/arch/sh/include/asm/ |
H A D | mmzone.h | 40 void __init __add_active_range(unsigned int nid, unsigned long start_pfn,
|
/linux-master/arch/x86/include/asm/ |
H A D | highmem.h | 72 extern void add_highpages_with_active_regions(int nid, unsigned long start_pfn,
|
/linux-master/arch/x86/xen/ |
H A D | setup.c | 84 static void __init xen_del_extra_mem(unsigned long start_pfn, argument 91 start_r = xen_extra_mem[i].start_pfn; 95 if (start_r == start_pfn) { 97 xen_extra_mem[i].start_pfn += n_pfns; 102 if (start_r + size_r == start_pfn + n_pfns) { 108 if (start_pfn > start_r && start_pfn < start_r + size_r) { 109 BUG_ON(start_pfn + n_pfns > start_r + size_r); 110 xen_extra_mem[i].n_pfns = start_pfn - start_r; 112 xen_add_extra_mem(start_pfn 215 xen_set_identity_and_release_chunk(unsigned long start_pfn, unsigned long end_pfn, unsigned long nr_pages) argument 290 xen_do_set_identity_and_remap_chunk( unsigned long start_pfn, unsigned long size, unsigned long remap_pfn) argument 344 xen_set_identity_and_remap_chunk( unsigned long start_pfn, unsigned long end_pfn, unsigned long nr_pages, unsigned long remap_pfn) argument 400 xen_count_remap_pages( unsigned long start_pfn, unsigned long end_pfn, unsigned long nr_pages, unsigned long remap_pages) argument 410 xen_foreach_remap_area(unsigned long nr_pages, unsigned long (*func)(unsigned long start_pfn, unsigned long end_pfn, unsigned long nr_pages, unsigned long last_val)) argument 433 unsigned long start_pfn = PFN_DOWN(start); local [all...] |
/linux-master/arch/parisc/mm/ |
H A D | init.c | 132 if (pmem_ranges[j-1].start_pfn < 133 pmem_ranges[j].start_pfn) { 148 if (pmem_ranges[i].start_pfn - 149 (pmem_ranges[i-1].start_pfn + 154 pmem_ranges[i].start_pfn - 155 (pmem_ranges[i-1].start_pfn + 171 start = (pmem_ranges[i].start_pfn << PAGE_SHIFT); 225 end_pfn = pmem_ranges[0].start_pfn + pmem_ranges[0].pages; 228 hole_pages = pmem_ranges[i].start_pfn - end_pfn; 230 pmem_holes[npmem_holes].start_pfn 248 unsigned long start_pfn; local [all...] |
/linux-master/drivers/hv/ |
H A D | hv_balloon.c | 425 * The range start_pfn : end_pfn specifies the range 427 * start_pfn : ha_end_pfn specifies the range that we have 437 unsigned long start_pfn; member in struct:hv_hotadd_state 450 unsigned long start_pfn; member in struct:hv_hotadd_gap 603 if ((pfn >= gap->start_pfn) && (pfn < gap->end_pfn)) 610 static unsigned long hv_page_offline_check(unsigned long start_pfn, argument 613 unsigned long pfn = start_pfn, count = 0; 617 while (pfn < start_pfn + nr_pages) { 624 while ((pfn >= has->start_pfn) && 626 (pfn < start_pfn 707 hv_bring_pgs_online(struct hv_hotadd_state *has, unsigned long start_pfn, unsigned long size) argument 723 unsigned long start_pfn; local 799 pfn_covered(unsigned long start_pfn, unsigned long pfn_cnt) argument 860 unsigned long start_pfn = pg_start; local [all...] |
/linux-master/arch/powerpc/mm/ |
H A D | mem.c | 126 int __ref add_pages(int nid, unsigned long start_pfn, unsigned long nr_pages, argument 131 ret = __add_pages(nid, start_pfn, nr_pages, params); 136 update_end_of_memory_vars(start_pfn << PAGE_SHIFT, 145 unsigned long start_pfn = start >> PAGE_SHIFT; local 152 rc = add_pages(nid, start_pfn, nr_pages, params); 160 unsigned long start_pfn = start >> PAGE_SHIFT; local 163 __remove_pages(start_pfn, nr_pages, altmap);
|
/linux-master/drivers/gpu/drm/imagination/ |
H A D | pvr_vm_mips.c | 155 s32 start_pfn; local 171 start_pfn = (start & fw_dev->fw_heap_info.offset_mask) >> ROGUE_MIPSFW_LOG2_PAGE_SIZE_4K; 181 for (pfn = start_pfn; pfn <= end_pfn; pfn++) { 186 (pfn - start_pfn) << 204 while (--pfn >= start_pfn) 227 const u32 start_pfn = (start & fw_dev->fw_heap_info.offset_mask) >> local 232 for (u32 pfn = start_pfn; pfn < end_pfn; pfn++)
|
/linux-master/drivers/base/ |
H A D | memory.c | 188 unsigned long start_pfn = section_nr_to_pfn(mem->start_section_nr); local 199 start_pfn, nr_pages); 211 arg.altmap_start_pfn = start_pfn; 213 arg.start_pfn = start_pfn + nr_vmemmap_pages; 222 ret = mhp_init_memmap_on_memory(start_pfn, nr_vmemmap_pages, 228 ret = online_pages(start_pfn + nr_vmemmap_pages, 232 mhp_deinit_memmap_on_memory(start_pfn, nr_vmemmap_pages); 241 adjust_present_page_count(pfn_to_page(start_pfn), mem->group, 259 unsigned long start_pfn local 428 unsigned long start_pfn = section_nr_to_pfn(mem->start_section_nr); local 435 print_allowed_zone(char *buf, int len, int nid, struct memory_group *group, unsigned long start_pfn, unsigned long nr_pages, int online_type, struct zone *default_zone) argument 453 unsigned long start_pfn = section_nr_to_pfn(mem->start_section_nr); local 629 arch_get_memory_phys_device(unsigned long start_pfn) argument 705 const unsigned long start_pfn = section_nr_to_pfn(mem->start_section_nr); local [all...] |