/linux-master/arch/xtensa/mm/ |
H A D | tlb.c | 153 if (end > start && start >= TASK_SIZE && end <= PAGE_OFFSET &&
|
/linux-master/arch/x86/kernel/ |
H A D | setup.c | 269 initrd_start = relocated_ramdisk + PAGE_OFFSET; 315 initrd_start = ramdisk_image + PAGE_OFFSET;
|
/linux-master/arch/x86/include/asm/ |
H A D | pgtable.h | 1180 #define KERNEL_PGD_BOUNDARY pgd_index(PAGE_OFFSET)
|
H A D | io.h | 233 * to PAGE_OFFSET is pure coincidence - it does not mean ISA values 238 #define __ISA_IO_base ((char __iomem *)(PAGE_OFFSET))
|
/linux-master/arch/riscv/include/asm/ |
H A D | pgtable.h | 15 #define KERNEL_LINK_ADDR PAGE_OFFSET 25 #define KERNEL_LINK_ADDR PAGE_OFFSET 40 #define VMALLOC_END PAGE_OFFSET 41 #define VMALLOC_START (PAGE_OFFSET - VMALLOC_SIZE) 48 #define BPF_JIT_REGION_START (PAGE_OFFSET - BPF_JIT_REGION_SIZE)
|
/linux-master/arch/powerpc/mm/ |
H A D | mem.c | 432 if (MODULES_VADDR < PAGE_OFFSET && MODULES_END > limit) {
|
H A D | cacheflush.c | 20 icbi((void *)PAGE_OFFSET);
|
/linux-master/arch/parisc/include/asm/ |
H A D | page.h | 133 #define PAGE_OFFSET ((unsigned long)__PAGE_OFFSET) macro 148 #define __pa(x) ((unsigned long)(x)-PAGE_OFFSET) 149 #define __va(x) ((void *)((unsigned long)(x)+PAGE_OFFSET))
|
/linux-master/arch/mips/mm/ |
H A D | mmap.c | 123 if ((vaddr < PAGE_OFFSET) || (vaddr >= MAP_BASE))
|
/linux-master/arch/loongarch/mm/ |
H A D | mmap.c | 117 if ((vaddr < PAGE_OFFSET) || (vaddr >= vm_map_base))
|
/linux-master/arch/arm64/include/asm/ |
H A D | pgtable.h | 27 #define VMEMMAP_UNUSED_NPAGES ((_PAGE_OFFSET(vabits_actual) - PAGE_OFFSET) >> PAGE_SHIFT)
|
/linux-master/arch/arm/kernel/ |
H A D | traps.c | 126 if (sp < PAGE_OFFSET ||
|
/linux-master/arch/arm/include/asm/ |
H A D | pgtable-3level.h | 52 #define USER_PTRS_PER_PGD (PAGE_OFFSET / PGDIR_SIZE)
|
H A D | pgtable-3level-hwdef.h | 80 * TTBR0/TTBR1 split (PAGE_OFFSET): 85 * Only use this feature if PHYS_OFFSET <= PAGE_OFFSET, otherwise 97 #define TTBR1_SIZE (((PAGE_OFFSET >> 30) - 1) << 16)
|
/linux-master/include/linux/ |
H A D | kvm_host.h | 136 * architectures with KVM_HVA_ERR_BAD other than PAGE_OFFSET (e.g. s390) 141 #define KVM_HVA_ERR_BAD (PAGE_OFFSET) 142 #define KVM_HVA_ERR_RO_BAD (PAGE_OFFSET + PAGE_SIZE) 146 return addr >= PAGE_OFFSET;
|
/linux-master/arch/sparc/include/asm/ |
H A D | pgtable_32.h | 50 #define USER_PTRS_PER_PGD PAGE_OFFSET / PGDIR_SIZE
|
/linux-master/arch/riscv/mm/ |
H A D | init.c | 148 print_ml("lowmem", (unsigned long)PAGE_OFFSET, 236 kernel_map.va_pa_offset = PAGE_OFFSET - phys_ram_base; 1103 0UL : PAGE_OFFSET - kernel_map.phys_addr; 1115 BUG_ON((PAGE_OFFSET % PGDIR_SIZE) != 0); 1271 if (start <= __pa(PAGE_OFFSET) && 1272 __pa(PAGE_OFFSET) < end) 1273 start = __pa(PAGE_OFFSET); 1274 if (end >= __pa(PAGE_OFFSET) + memory_limit) 1275 end = __pa(PAGE_OFFSET) + memory_limit;
|
/linux-master/arch/powerpc/net/ |
H A D | bpf_jit_comp64.c | 948 PPC_LI64(tmp2_reg, PAGE_OFFSET);
|
/linux-master/arch/powerpc/mm/ptdump/ |
H A D | hashpagetable.c | 332 if (ea < PAGE_OFFSET) 479 for (addr = PAGE_OFFSET; addr < PAGE_OFFSET + 503 address_markers[0].start_address = PAGE_OFFSET; 519 .start_address = PAGE_OFFSET,
|
/linux-master/arch/powerpc/mm/book3s32/ |
H A D | mmu.c | 151 setbat(idx, PAGE_OFFSET + base, base, size, PAGE_KERNEL_X); 161 unsigned long border = (unsigned long)__srwx_boundary - PAGE_OFFSET; 164 size = roundup_pow_of_two((unsigned long)_einittext - PAGE_OFFSET); 165 setibat(0, PAGE_OFFSET, 0, size, PAGE_KERNEL_X); 200 unsigned long base = (unsigned long)_stext - PAGE_OFFSET; 201 unsigned long top = ALIGN((unsigned long)_etext - PAGE_OFFSET, SZ_128K); 202 unsigned long border = (unsigned long)__init_begin - PAGE_OFFSET; 207 setibat(i++, PAGE_OFFSET + base, base, size, PAGE_KERNEL_TEXT); 218 setibat(i++, PAGE_OFFSET + base, base, size, PAGE_KERNEL_TEXT); 397 unsigned int hash = (unsigned int)Hash - PAGE_OFFSET; [all...] |
/linux-master/arch/powerpc/kernel/ |
H A D | head_book3s_32.S | 432 lis r2, (swapper_pg_dir - PAGE_OFFSET)@ha /* if kernel address, use */ 434 addi r2, r2, (swapper_pg_dir - PAGE_OFFSET)@l /* kernel page table */ 498 lis r2, (swapper_pg_dir - PAGE_OFFSET)@ha /* if kernel address, use */ 500 addi r2, r2, (swapper_pg_dir - PAGE_OFFSET)@l /* kernel page table */ 578 lis r2, (swapper_pg_dir - PAGE_OFFSET)@ha /* if kernel address, use */ 580 addi r2, r2, (swapper_pg_dir - PAGE_OFFSET)@l /* kernel page table */ 863 lis r4, (swapper_pg_dir - PAGE_OFFSET)@h 864 ori r4, r4, (swapper_pg_dir - PAGE_OFFSET)@l 893 lis r6, early_hash - PAGE_OFFSET@h 967 lis r4, (swapper_pg_dir - PAGE_OFFSET) [all...] |
H A D | head_8xx.S | 49 cmpli cr0, \scratch, PAGE_OFFSET@h 210 oris r10, r10, (swapper_pg_dir - PAGE_OFFSET)@ha 214 lwz r11, (swapper_pg_dir-PAGE_OFFSET)@l(r10) /* Get level 1 entry */ 240 0: lwz r10, (itlb_miss_counter - PAGE_OFFSET)@l(0) 242 stw r10, (itlb_miss_counter - PAGE_OFFSET)@l(0) 261 oris r10, r10, (swapper_pg_dir - PAGE_OFFSET)@ha 264 lwz r11, (swapper_pg_dir-PAGE_OFFSET)@l(r10) /* Get level 1 entry */ 299 0: lwz r10, (dtlb_miss_counter - PAGE_OFFSET)@l(0) 301 stw r10, (dtlb_miss_counter - PAGE_OFFSET)@l(0) 359 cmplwi cr1, r11, (.Ldtlbie - PAGE_OFFSET) [all...] |
/linux-master/arch/powerpc/include/asm/ |
H A D | kasan.h | 23 #define KASAN_KERN_START ALIGN_DOWN(PAGE_OFFSET - SZ_256M, SZ_256M) 25 #define KASAN_KERN_START PAGE_OFFSET
|
/linux-master/arch/arm64/kvm/ |
H A D | mmu.c | 1871 kern_hyp_va(PAGE_OFFSET), 1874 if (hyp_idmap_start >= kern_hyp_va(PAGE_OFFSET) &&
|
/linux-master/arch/arm/mm/ |
H A D | proc-v7-3level.S | 119 cmp \ttbr1, \tmp, lsr #12 @ PHYS_OFFSET > PAGE_OFFSET? 126 * Only use split TTBRs if PHYS_OFFSET <= PAGE_OFFSET (cmp above),
|