Searched refs:start_pfn (Results 1 - 25 of 101) sorted by relevance

12345

/linux-master/include/trace/events/
H A Dpage_isolation.h13 unsigned long start_pfn,
17 TP_ARGS(start_pfn, end_pfn, fin_pfn),
20 __field(unsigned long, start_pfn)
26 __entry->start_pfn = start_pfn;
31 TP_printk("start_pfn=0x%lx end_pfn=0x%lx fin_pfn=0x%lx ret=%s",
32 __entry->start_pfn, __entry->end_pfn, __entry->fin_pfn,
H A Dcompaction.h17 unsigned long start_pfn,
22 TP_ARGS(start_pfn, end_pfn, nr_scanned, nr_taken),
25 __field(unsigned long, start_pfn)
32 __entry->start_pfn = start_pfn;
39 __entry->start_pfn,
48 unsigned long start_pfn,
53 TP_ARGS(start_pfn, end_pfn, nr_scanned, nr_taken)
59 unsigned long start_pfn,
64 TP_ARGS(start_pfn, end_pf
[all...]
/linux-master/arch/sh/mm/
H A Dnuma.c28 unsigned long start_pfn, end_pfn; local
33 start_pfn = PFN_DOWN(start);
41 __add_active_range(nid, start_pfn, end_pfn);
51 NODE_DATA(nid)->node_start_pfn = start_pfn;
52 NODE_DATA(nid)->node_spanned_pages = end_pfn - start_pfn;
H A Dinit.c210 unsigned long start_pfn, end_pfn; local
212 get_pfn_range_for_nid(nid, &start_pfn, &end_pfn);
223 NODE_DATA(nid)->node_start_pfn = start_pfn;
224 NODE_DATA(nid)->node_spanned_pages = end_pfn - start_pfn;
229 unsigned long start_pfn, end_pfn; local
233 for_each_mem_pfn_range(i, MAX_NUMNODES, &start_pfn, &end_pfn, NULL)
234 __add_active_range(0, start_pfn, end_pfn);
247 unsigned long start_pfn; local
255 start_pfn = PFN_UP(__pa(_end));
263 memblock_reserve(start, (PFN_PHYS(start_pfn)
403 unsigned long start_pfn = PFN_DOWN(start); local
420 unsigned long start_pfn = PFN_DOWN(start); local
[all...]
/linux-master/include/linux/
H A Dpage-isolation.h40 int start_isolate_page_range(unsigned long start_pfn, unsigned long end_pfn,
43 void undo_isolate_page_range(unsigned long start_pfn, unsigned long end_pfn,
46 int test_pages_isolated(unsigned long start_pfn, unsigned long end_pfn,
H A Dnode.h108 void register_memory_blocks_under_node(int nid, unsigned long start_pfn,
112 static inline void register_memory_blocks_under_node(int nid, unsigned long start_pfn, argument
132 unsigned long start_pfn = pgdat->node_start_pfn; local
133 unsigned long end_pfn = start_pfn + pgdat->node_spanned_pages;
138 register_memory_blocks_under_node(nid, start_pfn, end_pfn,
H A Dmemory_hotplug.h178 extern void __offline_isolated_pages(unsigned long start_pfn,
205 extern void __remove_pages(unsigned long start_pfn, unsigned long nr_pages,
209 extern int __add_pages(int nid, unsigned long start_pfn, unsigned long nr_pages,
213 static inline int add_pages(int nid, unsigned long start_pfn, argument
216 return __add_pages(nid, start_pfn, nr_pages, params);
219 int add_pages(int nid, unsigned long start_pfn, unsigned long nr_pages,
330 extern int offline_pages(unsigned long start_pfn, unsigned long nr_pages,
339 static inline int offline_pages(unsigned long start_pfn, unsigned long nr_pages, argument
362 extern void move_pfn_range_to_zone(struct zone *zone, unsigned long start_pfn,
366 unsigned long start_pfn,
[all...]
/linux-master/arch/powerpc/platforms/powernv/
H A Dmemtrace.c91 static void memtrace_clear_range(unsigned long start_pfn, argument
97 for (pfn = start_pfn; pfn < start_pfn + nr_pages; pfn++) {
106 flush_dcache_range_chunked((unsigned long)pfn_to_kaddr(start_pfn),
107 (unsigned long)pfn_to_kaddr(start_pfn + nr_pages),
114 unsigned long pfn, start_pfn; local
125 start_pfn = page_to_pfn(page);
132 memtrace_clear_range(start_pfn, nr_pages);
138 for (pfn = start_pfn; pfn < start_pfn
217 const unsigned long start_pfn = PHYS_PFN(start); local
[all...]
/linux-master/mm/
H A Dpage_isolation.c19 * This function checks whether the range [start_pfn, end_pfn) includes
33 static struct page *has_unmovable_pages(unsigned long start_pfn, unsigned long end_pfn, argument
36 struct page *page = pfn_to_page(start_pfn);
40 VM_BUG_ON(pageblock_start_pfn(start_pfn) !=
55 for (pfn = start_pfn; pfn < end_pfn; pfn++) {
144 * present in [start_pfn, end_pfn). The pageblock must intersect with
145 * [start_pfn, end_pfn).
148 unsigned long start_pfn, unsigned long end_pfn)
171 * Pass the intersection of [start_pfn, end_pfn) and the page's pageblock
174 check_unmovable_start = max(page_to_pfn(page), start_pfn);
147 set_migratetype_isolate(struct page *page, int migratetype, int isol_flags, unsigned long start_pfn, unsigned long end_pfn) argument
310 unsigned long start_pfn; local
529 start_isolate_page_range(unsigned long start_pfn, unsigned long end_pfn, int migratetype, int flags, gfp_t gfp_flags) argument
583 undo_isolate_page_range(unsigned long start_pfn, unsigned long end_pfn, int migratetype) argument
655 test_pages_isolated(unsigned long start_pfn, unsigned long end_pfn, int isol_flags) argument
[all...]
H A Dmm_init.c300 unsigned long start_pfn, end_pfn; local
303 for_each_mem_pfn_range(i, MAX_NUMNODES, &start_pfn, &end_pfn, &nid) {
304 unsigned long pages = end_pfn - start_pfn;
462 unsigned long start_pfn, end_pfn; local
480 for_each_mem_pfn_range(i, nid, &start_pfn, &end_pfn, NULL) {
483 start_pfn = max(start_pfn, zone_movable_pfn[nid]);
484 if (start_pfn >= end_pfn)
488 if (start_pfn < usable_startpfn) {
491 - start_pfn;
549 unsigned long start_pfn, end_pfn; local
602 unsigned long start_pfn, end_pfn; local
751 unsigned long start_pfn = PFN_DOWN(start); local
849 memmap_init_range(unsigned long size, int nid, unsigned long zone, unsigned long start_pfn, unsigned long zone_end_pfn, enum meminit_context context, struct vmem_altmap *altmap, int migratetype) argument
910 memmap_init_zone_range(struct zone *zone, unsigned long start_pfn, unsigned long end_pfn, unsigned long *hole_pfn) argument
936 unsigned long start_pfn, end_pfn; local
1064 memmap_init_zone_device(struct zone *zone, unsigned long start_pfn, unsigned long nr_pages, struct dev_pagemap *pgmap) argument
1152 unsigned long start_pfn, end_pfn; local
1170 absent_pages_in_range(unsigned long start_pfn, unsigned long end_pfn) argument
1196 unsigned long start_pfn, end_pfn; local
1685 get_pfn_range_for_nid(unsigned int nid, unsigned long *start_pfn, unsigned long *end_pfn) argument
1706 unsigned long start_pfn = 0; local
1791 unsigned long start_pfn, end_pfn; local
2104 deferred_init_maxorder(u64 *i, struct zone *zone, unsigned long *start_pfn, unsigned long *end_pfn) argument
2148 deferred_init_memmap_chunk(unsigned long start_pfn, unsigned long end_pfn, void *arg) argument
[all...]
H A Dpage_ext.c354 static int __meminit online_page_ext(unsigned long start_pfn, argument
361 start = SECTION_ALIGN_DOWN(start_pfn);
362 end = SECTION_ALIGN_UP(start_pfn + nr_pages);
367 * "start_pfn" passed to us is a pfn which is an arg for
368 * online__pages(), and start_pfn should exist.
370 nid = pfn_to_nid(start_pfn);
387 static void __meminit offline_page_ext(unsigned long start_pfn, argument
392 start = SECTION_ALIGN_DOWN(start_pfn);
393 end = SECTION_ALIGN_UP(start_pfn + nr_pages);
421 ret = online_page_ext(mn->start_pfn,
451 unsigned long start_pfn, end_pfn; local
[all...]
H A Dmemory_hotplug.c413 /* find the smallest valid pfn in the range [start_pfn, end_pfn) */
415 unsigned long start_pfn,
418 for (; start_pfn < end_pfn; start_pfn += PAGES_PER_SUBSECTION) {
419 if (unlikely(!pfn_to_online_page(start_pfn)))
422 if (unlikely(pfn_to_nid(start_pfn) != nid))
425 if (zone != page_zone(pfn_to_page(start_pfn)))
428 return start_pfn;
434 /* find the biggest valid pfn in the range [start_pfn, end_pfn). */
436 unsigned long start_pfn,
414 find_smallest_section_pfn(int nid, struct zone *zone, unsigned long start_pfn, unsigned long end_pfn) argument
435 find_biggest_section_pfn(int nid, struct zone *zone, unsigned long start_pfn, unsigned long end_pfn) argument
459 shrink_zone_span(struct zone *zone, unsigned long start_pfn, unsigned long end_pfn) argument
527 remove_pfn_range_from_zone(struct zone *zone, unsigned long start_pfn, unsigned long nr_pages) argument
644 online_pages_range(unsigned long start_pfn, unsigned long nr_pages) argument
705 resize_zone_range(struct zone *zone, unsigned long start_pfn, unsigned long nr_pages) argument
716 resize_pgdat_range(struct pglist_data *pgdat, unsigned long start_pfn, unsigned long nr_pages) argument
750 move_pfn_range_to_zone(struct zone *zone, unsigned long start_pfn, unsigned long nr_pages, struct vmem_altmap *altmap, int migratetype) argument
897 default_kernel_zone_for_pfn(int nid, unsigned long start_pfn, unsigned long nr_pages) argument
1024 default_zone_for_pfn(int nid, unsigned long start_pfn, unsigned long nr_pages) argument
1048 zone_for_pfn_range(int online_type, int nid, struct memory_group *group, unsigned long start_pfn, unsigned long nr_pages) argument
1775 do_migrate_range(unsigned long start_pfn, unsigned long end_pfn) argument
1931 count_system_ram_pages_cb(unsigned long start_pfn, unsigned long nr_pages, void *data) argument
1943 offline_pages(unsigned long start_pfn, unsigned long nr_pages, struct zone *zone, struct memory_group *group) argument
[all...]
H A Dshuffle.c71 * the span, end_pfn - start_pfn.
83 unsigned long start_pfn = z->zone_start_pfn; local
89 start_pfn = ALIGN(start_pfn, order_pages);
90 for (i = start_pfn; i < end_pfn; i += order_pages) {
97 * (@start_pfn to @end_pfn), to more likely be valid compared to
H A Dbootmem_info.c43 static void __init register_page_bootmem_info_section(unsigned long start_pfn) argument
50 section_nr = pfn_to_section_nr(start_pfn);
78 static void __init register_page_bootmem_info_section(unsigned long start_pfn) argument
85 section_nr = pfn_to_section_nr(start_pfn);
/linux-master/arch/x86/mm/
H A Dinit.c328 unsigned long start_pfn, unsigned long end_pfn,
331 if (start_pfn < end_pfn) {
334 mr[nr_range].start = start_pfn<<PAGE_SHIFT;
406 unsigned long start_pfn, end_pfn, limit_pfn; local
413 pfn = start_pfn = PFN_DOWN(start);
430 if (start_pfn < end_pfn) {
431 nr_range = save_mr(mr, nr_range, start_pfn, end_pfn, 0);
436 start_pfn = round_up(pfn, PFN_DOWN(PMD_SIZE));
445 if (start_pfn < end_pfn) {
446 nr_range = save_mr(mr, nr_range, start_pfn, end_pf
327 save_mr(struct map_range *mr, int nr_range, unsigned long start_pfn, unsigned long end_pfn, unsigned long page_size_mask) argument
505 add_pfn_range_mapped(unsigned long start_pfn, unsigned long end_pfn) argument
518 pfn_range_is_mapped(unsigned long start_pfn, unsigned long end_pfn) argument
575 unsigned long start_pfn, end_pfn; local
1005 unsigned long start_pfn, end_pfn; local
[all...]
/linux-master/arch/sparc/mm/
H A Dinit_32.c64 unsigned long start_pfn = sp_banks[i].base_addr >> PAGE_SHIFT; local
70 if (start_pfn < max_low_pfn)
71 start_pfn = max_low_pfn;
73 nr += end_pfn - start_pfn;
133 unsigned long start_pfn, bytes_avail, size; local
171 start_pfn = (unsigned long)__pa(PAGE_ALIGN((unsigned long) &_end));
174 start_pfn >>= PAGE_SHIFT;
192 size = (start_pfn << PAGE_SHIFT) - phys_base;
235 static void map_high_region(unsigned long start_pfn, unsigned long end_pfn) argument
240 printk("mapping high region %08lx - %08lx\n", start_pfn, end_pf
283 unsigned long start_pfn = sp_banks[i].base_addr >> PAGE_SHIFT; local
[all...]
/linux-master/arch/mips/loongson64/
H A Dnuma.c88 unsigned long start_pfn, end_pfn; local
97 get_pfn_range_for_nid(node, &start_pfn, &end_pfn);
98 pr_info("Node%d: start_pfn=0x%lx, end_pfn=0x%lx\n",
99 node, start_pfn, end_pfn);
111 NODE_DATA(node)->node_start_pfn = start_pfn;
112 NODE_DATA(node)->node_spanned_pages = end_pfn - start_pfn;
134 memblock_reserve(0, PAGE_SIZE * start_pfn);
/linux-master/arch/sh/include/asm/
H A Dmmzone.h40 void __init __add_active_range(unsigned int nid, unsigned long start_pfn,
/linux-master/arch/x86/include/asm/
H A Dhighmem.h72 extern void add_highpages_with_active_regions(int nid, unsigned long start_pfn,
/linux-master/arch/x86/xen/
H A Dsetup.c84 static void __init xen_del_extra_mem(unsigned long start_pfn, argument
91 start_r = xen_extra_mem[i].start_pfn;
95 if (start_r == start_pfn) {
97 xen_extra_mem[i].start_pfn += n_pfns;
102 if (start_r + size_r == start_pfn + n_pfns) {
108 if (start_pfn > start_r && start_pfn < start_r + size_r) {
109 BUG_ON(start_pfn + n_pfns > start_r + size_r);
110 xen_extra_mem[i].n_pfns = start_pfn - start_r;
112 xen_add_extra_mem(start_pfn
215 xen_set_identity_and_release_chunk(unsigned long start_pfn, unsigned long end_pfn, unsigned long nr_pages) argument
290 xen_do_set_identity_and_remap_chunk( unsigned long start_pfn, unsigned long size, unsigned long remap_pfn) argument
344 xen_set_identity_and_remap_chunk( unsigned long start_pfn, unsigned long end_pfn, unsigned long nr_pages, unsigned long remap_pfn) argument
400 xen_count_remap_pages( unsigned long start_pfn, unsigned long end_pfn, unsigned long nr_pages, unsigned long remap_pages) argument
410 xen_foreach_remap_area(unsigned long nr_pages, unsigned long (*func)(unsigned long start_pfn, unsigned long end_pfn, unsigned long nr_pages, unsigned long last_val)) argument
433 unsigned long start_pfn = PFN_DOWN(start); local
[all...]
/linux-master/arch/parisc/mm/
H A Dinit.c132 if (pmem_ranges[j-1].start_pfn <
133 pmem_ranges[j].start_pfn) {
148 if (pmem_ranges[i].start_pfn -
149 (pmem_ranges[i-1].start_pfn +
154 pmem_ranges[i].start_pfn -
155 (pmem_ranges[i-1].start_pfn +
171 start = (pmem_ranges[i].start_pfn << PAGE_SHIFT);
225 end_pfn = pmem_ranges[0].start_pfn + pmem_ranges[0].pages;
228 hole_pages = pmem_ranges[i].start_pfn - end_pfn;
230 pmem_holes[npmem_holes].start_pfn
248 unsigned long start_pfn; local
[all...]
/linux-master/drivers/hv/
H A Dhv_balloon.c425 * The range start_pfn : end_pfn specifies the range
427 * start_pfn : ha_end_pfn specifies the range that we have
437 unsigned long start_pfn; member in struct:hv_hotadd_state
450 unsigned long start_pfn; member in struct:hv_hotadd_gap
603 if ((pfn >= gap->start_pfn) && (pfn < gap->end_pfn))
610 static unsigned long hv_page_offline_check(unsigned long start_pfn, argument
613 unsigned long pfn = start_pfn, count = 0;
617 while (pfn < start_pfn + nr_pages) {
624 while ((pfn >= has->start_pfn) &&
626 (pfn < start_pfn
707 hv_bring_pgs_online(struct hv_hotadd_state *has, unsigned long start_pfn, unsigned long size) argument
723 unsigned long start_pfn; local
799 pfn_covered(unsigned long start_pfn, unsigned long pfn_cnt) argument
860 unsigned long start_pfn = pg_start; local
[all...]
/linux-master/arch/powerpc/mm/
H A Dmem.c126 int __ref add_pages(int nid, unsigned long start_pfn, unsigned long nr_pages, argument
131 ret = __add_pages(nid, start_pfn, nr_pages, params);
136 update_end_of_memory_vars(start_pfn << PAGE_SHIFT,
145 unsigned long start_pfn = start >> PAGE_SHIFT; local
152 rc = add_pages(nid, start_pfn, nr_pages, params);
160 unsigned long start_pfn = start >> PAGE_SHIFT; local
163 __remove_pages(start_pfn, nr_pages, altmap);
/linux-master/drivers/gpu/drm/imagination/
H A Dpvr_vm_mips.c155 s32 start_pfn; local
171 start_pfn = (start & fw_dev->fw_heap_info.offset_mask) >> ROGUE_MIPSFW_LOG2_PAGE_SIZE_4K;
181 for (pfn = start_pfn; pfn <= end_pfn; pfn++) {
186 (pfn - start_pfn) <<
204 while (--pfn >= start_pfn)
227 const u32 start_pfn = (start & fw_dev->fw_heap_info.offset_mask) >> local
232 for (u32 pfn = start_pfn; pfn < end_pfn; pfn++)
/linux-master/drivers/base/
H A Dmemory.c188 unsigned long start_pfn = section_nr_to_pfn(mem->start_section_nr); local
199 start_pfn, nr_pages);
211 arg.altmap_start_pfn = start_pfn;
213 arg.start_pfn = start_pfn + nr_vmemmap_pages;
222 ret = mhp_init_memmap_on_memory(start_pfn, nr_vmemmap_pages,
228 ret = online_pages(start_pfn + nr_vmemmap_pages,
232 mhp_deinit_memmap_on_memory(start_pfn, nr_vmemmap_pages);
241 adjust_present_page_count(pfn_to_page(start_pfn), mem->group,
259 unsigned long start_pfn local
428 unsigned long start_pfn = section_nr_to_pfn(mem->start_section_nr); local
435 print_allowed_zone(char *buf, int len, int nid, struct memory_group *group, unsigned long start_pfn, unsigned long nr_pages, int online_type, struct zone *default_zone) argument
453 unsigned long start_pfn = section_nr_to_pfn(mem->start_section_nr); local
629 arch_get_memory_phys_device(unsigned long start_pfn) argument
705 const unsigned long start_pfn = section_nr_to_pfn(mem->start_section_nr); local
[all...]

Completed in 252 milliseconds

12345