/linux-master/arch/x86/mm/ |
H A D | numa_32.c | 38 highstart_pfn = highend_pfn = max_pfn; 39 if (max_pfn > max_low_pfn)
|
H A D | init_32.c | 557 max_low_pfn = max_pfn; 562 if (highmem_pages >= max_pfn) { 564 pages_to_mb(highmem_pages), pages_to_mb(max_pfn)); 595 highmem_pages = max_pfn - MAXMEM_PFN; 597 if (highmem_pages + MAXMEM_PFN < max_pfn) 598 max_pfn = MAXMEM_PFN + highmem_pages; 600 if (highmem_pages + MAXMEM_PFN > max_pfn) { 602 pages_to_mb(max_pfn - MAXMEM_PFN), 609 if (max_pfn > MAX_NONPAE_PFN) 613 max_pfn [all...] |
H A D | kaslr.c | 96 memory_tb = DIV_ROUND_UP(max_pfn << PAGE_SHIFT, 1UL << TB_SHIFT) +
|
H A D | numa_emulation.c | 203 unsigned long max_pfn = PHYS_PFN(max_addr); local 207 return PFN_PHYS((max_pfn - base_pfn - hole_pfns) / nr_nodes); 375 const u64 max_addr = PFN_PHYS(max_pfn);
|
/linux-master/arch/powerpc/include/asm/ |
H A D | mmzone.h | 36 extern unsigned long max_pfn;
|
/linux-master/arch/csky/kernel/ |
H A D | setup.c | 30 max_low_pfn = max_pfn = PFN_DOWN(memblock_end_of_DRAM()); 32 size = max_pfn - min_low_pfn; 48 max_zone_pfn[ZONE_HIGHMEM] = max_pfn; 51 highend_pfn = max_pfn;
|
/linux-master/mm/ |
H A D | page_idle.c | 127 if (pfn >= max_pfn) 131 if (end_pfn > max_pfn) 132 end_pfn = max_pfn; 172 if (pfn >= max_pfn) 176 if (end_pfn > max_pfn) 177 end_pfn = max_pfn;
|
/linux-master/arch/xtensa/mm/ |
H A D | init.c | 57 max_pfn = PFN_DOWN(memblock_end_of_DRAM()); 58 max_low_pfn = min(max_pfn, MAX_LOW_PFN); 76 [ZONE_HIGHMEM] = max_pfn, 117 max_mapnr = max_pfn - ARCH_PFN_OFFSET;
|
/linux-master/arch/microblaze/mm/ |
H A D | init.c | 60 for (pfn = max_low_pfn; pfn < max_pfn; ++pfn) { 89 zones_size[ZONE_HIGHMEM] = max_pfn; 91 zones_size[ZONE_DMA] = max_pfn; 115 max_pfn = ((u64)memory_start + (u64)memory_size) >> PAGE_SHIFT; 120 pr_info("%s: max_pfn: %#lx\n", __func__, max_pfn);
|
/linux-master/arch/powerpc/mm/ |
H A D | mem.c | 112 * After memory hotplug the variables max_pfn, max_low_pfn and high_memory need 119 if (end_pfn > max_pfn) { 120 max_pfn = end_pfn; 122 high_memory = (void *)__va(max_pfn * PAGE_SIZE - 1) + 1; 135 /* update max_pfn, max_low_pfn and high_memory */ 171 max_low_pfn = max_pfn = memblock_end_of_DRAM() >> PAGE_SHIFT; 263 max_zone_pfns[ZONE_HIGHMEM] = max_pfn;
|
/linux-master/arch/x86/include/asm/ |
H A D | page_64.h | 14 extern unsigned long max_pfn;
|
/linux-master/arch/csky/include/asm/ |
H A D | pgalloc.h | 71 extern void mmu_init(unsigned long min_pfn, unsigned long max_pfn);
|
/linux-master/scripts/gdb/linux/ |
H A D | page_owner.py | 34 max_pfn = None variable in class:DumpPageOwner 65 self.max_pfn = int(gdb.parse_and_eval("max_pfn")) 102 if pfn < self.min_pfn or pfn > self.max_pfn or (not self.p_ops.pfn_valid(pfn)): 151 while pfn < self.max_pfn:
|
/linux-master/arch/x86/kernel/ |
H A D | setup.c | 942 max_pfn = e820__end_of_ram_pfn(); 946 if (mtrr_trim_uncached_memory(max_pfn)) 947 max_pfn = e820__end_of_ram_pfn(); 949 max_possible_pfn = max_pfn; 952 * Define random base addresses for memory sections after max_pfn is 965 if (max_pfn > (1UL<<(32 - PAGE_SHIFT))) 968 max_low_pfn = max_pfn; 970 high_memory = (void *)__va(max_pfn * PAGE_SIZE - 1) + 1; 1157 e820__register_nosave_regions(max_pfn);
|
/linux-master/arch/sparc/mm/ |
H A D | init_32.c | 176 max_pfn = end_of_phys_memory >> PAGE_SHIFT; 178 max_low_pfn = max_pfn; 179 highstart_pfn = highend_pfn = max_pfn; 202 return max_pfn;
|
/linux-master/arch/arm/mm/ |
H A D | init.c | 209 find_limits(&min_low_pfn, &max_low_pfn, &max_pfn); 225 zone_sizes_init(min_low_pfn, max_low_pfn, max_pfn); 274 swiotlb_init(max_pfn > arm_dma_pfn_limit, SWIOTLB_VERBOSE); 277 set_max_mapnr(pfn_to_page(max_pfn) - mem_map);
|
/linux-master/arch/arm64/mm/ |
H A D | init.c | 155 max_zone_pfns[ZONE_NORMAL] = max_pfn; 327 max_pfn = max_low_pfn = max; 371 bool swiotlb = max_pfn > PFN_DOWN(arm64_dma_phys_limit);
|
/linux-master/arch/x86/xen/ |
H A D | setup.c | 687 unsigned long max_pfn, pfn_s, n_pfns; local 698 max_pfn = xen_get_pages_limit(); 699 max_pfn = min(max_pfn, xen_start_info->nr_pages); 700 mem_end = PFN_PHYS(max_pfn); 753 max_pages += xen_foreach_remap_area(max_pfn, xen_count_remap_pages); 755 if (max_pages > max_pfn) 756 extra_pages += max_pages - max_pfn; 765 extra_pages = min3(EXTRA_MEM_RATIO * min(max_pfn, PFN_DOWN(MAXMEM)), 766 extra_pages, max_pages - max_pfn); [all...] |
H A D | xen-ops.h | 37 void xen_setup_kernel_pagetable(pgd_t *pgd, unsigned long max_pfn);
|
/linux-master/arch/arc/mm/ |
H A D | init.c | 94 max_low_pfn = max_pfn = PFN_DOWN(low_mem_start + low_mem_sz);
|
/linux-master/arch/nios2/kernel/ |
H A D | setup.c | 158 find_limits(&min_low_pfn, &max_low_pfn, &max_pfn);
|
/linux-master/arch/microblaze/include/asm/ |
H A D | page.h | 89 extern unsigned long max_pfn;
|
/linux-master/arch/alpha/mm/ |
H A D | init.c | 247 max_pfn = max_low_pfn; 250 max_zone_pfn[ZONE_NORMAL] = max_pfn;
|
/linux-master/arch/csky/mm/ |
H A D | init.c | 128 void __init mmu_init(unsigned long min_pfn, unsigned long max_pfn) argument 142 for (i = min_pfn; i < max_pfn; i++)
|
/linux-master/arch/parisc/mm/ |
H A D | init.c | 246 max_pfn = 0; 262 if ((start_pfn + npages) > max_pfn) 263 max_pfn = start_pfn + npages; 278 max_low_pfn = max_pfn; 562 high_memory = __va((max_pfn << PAGE_SHIFT));
|