/linux-master/arch/x86/mm/ |
H A D | numa_32.c | 39 if (max_pfn > max_low_pfn) 40 highstart_pfn = max_low_pfn; 45 high_memory = (void *) __va(max_low_pfn * PAGE_SIZE - 1) + 1; 48 pages_to_mb(max_low_pfn)); 49 printk(KERN_DEBUG "max_low_pfn = %lx, highstart_pfn = %lx\n", 50 max_low_pfn, highstart_pfn); 53 (ulong) pfn_to_kaddr(max_low_pfn));
|
H A D | physaddr.c | 78 /* max_low_pfn is set early, but not _that_ early */ 79 if (max_low_pfn) { 80 VIRTUAL_BUG_ON((phys_addr >> PAGE_SHIFT) > max_low_pfn);
|
H A D | init_32.c | 245 * of max_low_pfn pages, by creating page tables starting from address 449 * definition. And max_low_pfn is set to VMALLOC_END physical 451 * should have pte used near max_low_pfn or one pmd is not present. 453 for (pfn = max_low_pfn; pfn < 1<<(32-PAGE_SHIFT); pfn++) { 467 pr_warn("try to clear pte for ram above max_low_pfn: pfn: %lx pmd: %p pmd phys: %lx, but pmd is big page and is not using pte !\n", 476 printk(KERN_DEBUG "clearing pte for ram above max_low_pfn: pfn: %lx pmd: %p pmd phys: %lx pte: %p pte phys: %lx\n", 556 /* max_low_pfn is 0, we already have early_res support */ 557 max_low_pfn = max_pfn; 568 if (max_low_pfn - highmem_pages < 64*1024*1024/PAGE_SIZE) { 573 max_low_pfn [all...] |
H A D | init.c | 359 if ((end >> PAGE_SHIFT) > max_low_pfn) 561 * create direct mappings for all pfns from [0 to max_low_pfn) and 765 end = max_low_pfn << PAGE_SHIFT; 795 if (max_pfn > max_low_pfn) { 796 /* can we preserve max_low_pfn ?*/ 797 max_low_pfn = max_pfn; 1047 max_zone_pfns[ZONE_DMA] = min(MAX_DMA_PFN, max_low_pfn); 1050 max_zone_pfns[ZONE_DMA32] = min(MAX_DMA32_PFN, max_low_pfn); 1052 max_zone_pfns[ZONE_NORMAL] = max_low_pfn;
|
/linux-master/arch/csky/kernel/ |
H A D | setup.c | 30 max_low_pfn = max_pfn = PFN_DOWN(memblock_end_of_DRAM()); 35 max_low_pfn = min_low_pfn + lowmem_size; 40 max_low_pfn = min_low_pfn + sseg_size; 43 max_zone_pfn[ZONE_NORMAL] = max_low_pfn; 45 mmu_init(min_low_pfn, max_low_pfn); 50 highstart_pfn = max_low_pfn; 53 memblock_set_current_limit(PFN_PHYS(max_low_pfn));
|
/linux-master/arch/xtensa/mm/ |
H A D | init.c | 58 max_low_pfn = min(max_pfn, MAX_LOW_PFN); 61 (phys_addr_t)max_low_pfn << PAGE_SHIFT); 63 memblock_set_current_limit(PFN_PHYS(max_low_pfn)); 64 dma_contiguous_reserve(PFN_PHYS(max_low_pfn)); 74 [ZONE_NORMAL] = max_low_pfn, 85 unsigned long max_low = max_low_pfn; 118 high_memory = (void *)__va(max_low_pfn << PAGE_SHIFT); 153 (max_low_pfn - min_low_pfn) * PAGE_SIZE, 155 min_low_pfn * PAGE_SIZE, max_low_pfn * PAGE_SIZE, 157 ((max_low_pfn [all...] |
/linux-master/arch/loongarch/kernel/ |
H A D | mem.c | 34 if (max_low_pfn < (mem_end >> PAGE_SHIFT)) 35 max_low_pfn = mem_end >> PAGE_SHIFT; 52 memblock_set_current_limit(PFN_PHYS(max_low_pfn));
|
H A D | numa.c | 380 max_low_pfn = PHYS_PFN(memblock_end_of_DRAM()); 401 if (end_pfn > max_low_pfn) 402 max_low_pfn = end_pfn; 407 zones_size[ZONE_NORMAL] = max_low_pfn; 413 high_memory = (void *) __va(max_low_pfn << PAGE_SHIFT);
|
/linux-master/arch/hexagon/kernel/ |
H A D | dma.c | 36 * Our max_low_pfn should have been backed off by 16MB in mm/init.c to create 41 return dma_init_global_coherent(PFN_PHYS(max_low_pfn),
|
/linux-master/arch/arc/mm/ |
H A D | init.c | 94 max_low_pfn = max_pfn = PFN_DOWN(low_mem_start + low_mem_sz); 126 max_zone_pfn[ZONE_NORMAL] = max_low_pfn; 160 max_mapnr = max_low_pfn - min_low_pfn; 199 (pfn >= min_low_pfn && pfn <= max_low_pfn);
|
/linux-master/arch/hexagon/mm/ |
H A D | init.c | 102 max_zone_pfn[ZONE_NORMAL] = max_low_pfn; 169 max_low_pfn = bootmem_lastpg - PFN_DOWN(DMA_RESERVED_BYTES); 171 memblock_reserve(PFN_PHYS(max_low_pfn), DMA_RESERVED_BYTES); 176 printk(KERN_INFO "max_low_pfn: 0x%08lx\n", max_low_pfn);
|
/linux-master/arch/sparc/mm/ |
H A D | init_32.c | 67 if (end_pfn <= max_low_pfn) 70 if (start_pfn < max_low_pfn) 71 start_pfn = max_low_pfn; 178 max_low_pfn = max_pfn; 181 if (max_low_pfn > pfn_base + (SRMMU_MAXMEM >> PAGE_SHIFT)) { 183 max_low_pfn = calc_max_low_pfn(); 200 memblock_set_current_limit(max_low_pfn << PAGE_SHIFT); 279 high_memory = __va(max_low_pfn << PAGE_SHIFT);
|
/linux-master/arch/microblaze/mm/ |
H A D | init.c | 46 EXPORT_SYMBOL(max_low_pfn); variable 60 for (pfn = max_low_pfn; pfn < max_pfn; ++pfn) { 88 zones_size[ZONE_DMA] = max_low_pfn; 106 * max_low_pfn 114 max_low_pfn = ((u64)memory_start + (u64)lowmem_size) >> PAGE_SHIFT; 119 pr_info("%s: max_low_pfn: %#lx\n", __func__, max_low_pfn); 140 return pfn < max_low_pfn;
|
/linux-master/arch/powerpc/mm/ |
H A D | mem.c | 112 * After memory hotplug the variables max_pfn, max_low_pfn and high_memory need 121 max_low_pfn = end_pfn; 135 /* update max_pfn, max_low_pfn and high_memory */ 171 max_low_pfn = max_pfn = memblock_end_of_DRAM() >> PAGE_SHIFT; 174 max_low_pfn = lowmem_end_addr >> PAGE_SHIFT; 258 max_zone_pfns[ZONE_DMA] = min(max_low_pfn, 261 max_zone_pfns[ZONE_NORMAL] = max_low_pfn; 291 high_memory = (void *) __va(max_low_pfn * PAGE_SIZE);
|
/linux-master/arch/x86/power/ |
H A D | hibernate_32.c | 78 * of max_low_pfn pages, by creating page tables starting from address 98 if (pfn >= max_low_pfn) 102 if (pfn >= max_low_pfn) 121 if (pfn >= max_low_pfn)
|
/linux-master/arch/openrisc/mm/ |
H A D | init.c | 48 max_zone_pfn[ZONE_NORMAL] = max_low_pfn; 196 max_mapnr = max_low_pfn; 197 high_memory = (void *)__va(max_low_pfn * PAGE_SIZE);
|
/linux-master/arch/mips/loongson64/ |
H A D | numa.c | 122 max_low_pfn = end_pfn; 154 max_low_pfn = PHYS_PFN(memblock_end_of_DRAM()); 177 zones_size[ZONE_NORMAL] = max_low_pfn;
|
/linux-master/arch/nios2/kernel/ |
H A D | setup.c | 158 find_limits(&min_low_pfn, &max_low_pfn, &max_pfn); 159 max_mapnr = max_low_pfn;
|
/linux-master/arch/mips/mm/ |
H A D | init.c | 416 max_zone_pfns[ZONE_NORMAL] = max_low_pfn; 420 if (cpu_has_dc_aliases && max_low_pfn != highend_pfn) { 423 (highend_pfn - max_low_pfn) << (PAGE_SHIFT - 10)); 424 max_zone_pfns[ZONE_HIGHMEM] = max_low_pfn; 426 max_mapnr = max_low_pfn; 430 max_mapnr = max_low_pfn; 433 max_mapnr = max_low_pfn; 435 high_memory = (void *) __va(max_low_pfn << PAGE_SHIFT);
|
/linux-master/arch/mips/kernel/ |
H A D | setup.c | 221 if (__pa(initrd_end) > PFN_PHYS(max_low_pfn)) { 286 /* max_low_pfn is not a number of pages but the end pfn of low mem */ 308 * Skip highmem here so we get an accurate max_low_pfn if low 317 if (end > max_low_pfn) 318 max_low_pfn = end; 321 if (min_low_pfn >= max_low_pfn) 325 max_low_pfn = PFN_DOWN(HIGHMEM_START); 327 highstart_pfn = max_low_pfn; 330 max_pfn = max_low_pfn; 656 * This cannot be done before max_low_pfn i [all...] |
/linux-master/arch/arm/mm/ |
H A D | init.c | 209 find_limits(&min_low_pfn, &max_low_pfn, &max_pfn); 212 (phys_addr_t)max_low_pfn << PAGE_SHIFT); 225 zone_sizes_init(min_low_pfn, max_low_pfn, max_pfn); 242 unsigned long max_low = max_low_pfn;
|
/linux-master/arch/alpha/mm/ |
H A D | init.c | 22 #include <linux/memblock.h> /* max_low_pfn */ 247 max_pfn = max_low_pfn; 279 set_max_mapnr(max_low_pfn); 280 high_memory = (void *) __va(max_low_pfn * PAGE_SIZE);
|
/linux-master/arch/csky/mm/ |
H A D | init.c | 55 if (__pa(initrd_end) > PFN_PHYS(max_low_pfn)) { 91 set_max_mapnr(max_low_pfn - ARCH_PFN_OFFSET); 93 high_memory = (void *) __va(max_low_pfn << PAGE_SHIFT);
|
/linux-master/arch/loongarch/mm/ |
H A D | init.c | 73 max_zone_pfns[ZONE_NORMAL] = max_low_pfn; 80 max_mapnr = max_low_pfn; 81 high_memory = (void *) __va(max_low_pfn << PAGE_SHIFT);
|
/linux-master/arch/s390/mm/ |
H A D | init.c | 101 max_zone_pfns[ZONE_NORMAL] = max_low_pfn; 160 set_max_mapnr(max_low_pfn); 161 high_memory = (void *) __va(max_low_pfn * PAGE_SIZE);
|