Searched refs:PAGE_END (Results 1 - 8 of 8) sorted by relevance
/linux-master/arch/arm64/include/asm/ |
H A D | memory.h | 29 * PAGE_OFFSET all the way to PAGE_END for 48-bit. This allows us to 39 * PAGE_END - the end of the linear map, where all other kernel mappings begin. 106 #define PAGE_END KASAN_SHADOW_START macro 110 #define PAGE_END (_PAGE_END(VA_BITS_MIN)) macro 328 * lives in the [PAGE_OFFSET, PAGE_END) interval at the bottom of the 331 #define __is_lm_address(addr) (((u64)(addr) - PAGE_OFFSET) < (PAGE_END - PAGE_OFFSET))
|
/linux-master/scripts/gdb/linux/ |
H A D | mm.py | 114 self.PAGE_END = self.KASAN_SHADOW_END - (1 << (self.vabits_actual - self.KASAN_SHADOW_SCALE_SHIFT)) 116 self.PAGE_END = self._PAGE_END(self.VA_BITS_MIN) 190 if (addr - self.PAGE_OFFSET) < (self.PAGE_END - self.PAGE_OFFSET):
|
/linux-master/arch/arm64/kernel/ |
H A D | hibernate.c | 419 PAGE_END);
|
H A D | machine_kexec.c | 135 rc = trans_pgd_create_copy(&info, &trans_pgd, PAGE_OFFSET, PAGE_END);
|
/linux-master/arch/arm64/mm/ |
H A D | init.c | 191 s64 linear_region_size = PAGE_END - _PAGE_OFFSET(vabits_actual);
|
H A D | ptdump.c | 364 { PAGE_END, "Linear Mapping end" },
|
H A D | mmu.c | 1285 WARN_ON((start < PAGE_OFFSET) || (end > PAGE_END)); 1288 free_empty_tables(start, end, PAGE_OFFSET, PAGE_END); 1295 u64 end_linear_pa = __pa(PAGE_END - 1); 1312 * Linear mapping region is the range [PAGE_OFFSET..(PAGE_END - 1)] 1313 * accommodating both its ends but excluding PAGE_END. Max physical
|
H A D | kasan_init.c | 345 kasan_populate_early_shadow(kasan_mem_to_shadow((void *)PAGE_END),
|
Completed in 182 milliseconds