/linux-master/arch/x86/mm/ |
H A D | fault.c | 240 if (!(address >= VMALLOC_START && address < VMALLOC_END)) 271 addr >= TASK_SIZE_MAX && addr < VMALLOC_END;
|
H A D | kasan_init_64.c | 403 kasan_mem_to_shadow((void *)VMALLOC_END)); 407 kasan_mem_to_shadow((void *)VMALLOC_END)); 410 kasan_mem_to_shadow((void *)VMALLOC_END + 1),
|
H A D | init_32.c | 448 * from VMALLOC_END to pkmap or fixmap according to VMALLOC_END 449 * definition. And max_low_pfn is set to VMALLOC_END physical 748 BUILD_BUG_ON(VMALLOC_END > PKMAP_BASE); 751 BUILD_BUG_ON(VMALLOC_START >= VMALLOC_END); 757 BUG_ON(VMALLOC_END > PKMAP_BASE); 759 BUG_ON(VMALLOC_START >= VMALLOC_END);
|
H A D | dump_pagetables.c | 467 address_markers[VMALLOC_END_NR].start_address = VMALLOC_END;
|
/linux-master/kernel/bpf/ |
H A D | core.c | 1020 return VMALLOC_END - VMALLOC_START;
|
H A D | syscall.c | 307 return __vmalloc_node_range(size, align, VMALLOC_START, VMALLOC_END,
|
/linux-master/arch/arm64/net/ |
H A D | bpf_jit_comp.c | 1793 return VMALLOC_END - VMALLOC_START;
|
/linux-master/arch/riscv/mm/ |
H A D | init.c | 142 (unsigned long)VMALLOC_END); 1479 preallocate_pgd_pages_range(VMALLOC_START, VMALLOC_END, "vmalloc");
|
/linux-master/arch/riscv/include/asm/ |
H A D | pgtable.h | 40 #define VMALLOC_END PAGE_OFFSET macro 49 #define BPF_JIT_REGION_END (VMALLOC_END) 901 #define VMALLOC_END TASK_SIZE macro
|
/linux-master/kernel/ |
H A D | fork.c | 310 VMALLOC_START, VMALLOC_END,
|
/linux-master/mm/ |
H A D | vmalloc.c | 82 return addr >= VMALLOC_START && addr < VMALLOC_END; 1915 if (vstart != VMALLOC_START || vend != VMALLOC_END || 2423 * If we had a constant VMALLOC_START and VMALLOC_END, we'd like to be able 2424 * to #define VMALLOC_SPACE (VMALLOC_END-VMALLOC_START). Guess 2575 VMALLOC_START, VMALLOC_END, 2884 BUG_ON(addr > VMALLOC_END); 2933 VMALLOC_START, VMALLOC_END, 3030 BUG_ON(addr > VMALLOC_END - vm->size); 3138 VMALLOC_START, VMALLOC_END, 3147 VMALLOC_START, VMALLOC_END, [all...] |
H A D | util.c | 659 return __vmalloc_node_range(size, 1, VMALLOC_START, VMALLOC_END,
|
/linux-master/arch/x86/hyperv/ |
H A D | hv_init.c | 535 VMALLOC_END, GFP_KERNEL, PAGE_KERNEL_ROX,
|
/linux-master/arch/loongarch/include/asm/ |
H A D | pgtable.h | 94 #define VMALLOC_END \ macro 98 #define VMALLOC_END \ macro 103 #define vmemmap ((struct page *)((VMALLOC_END + PMD_SIZE) & PMD_MASK))
|
/linux-master/arch/arm/kernel/ |
H A D | traps.c | 964 if (start < VMALLOC_END && end > VMALLOC_START)
|
/linux-master/arch/powerpc/mm/book3s64/ |
H A D | hash_utils.c | 1615 } else if (ea < VMALLOC_END) {
|
/linux-master/arch/powerpc/include/asm/book3s/64/ |
H A D | pgtable.h | 246 #define VMALLOC_END __vmalloc_end macro
|
/linux-master/scripts/gdb/linux/ |
H A D | vmalloc.py | 20 return addr >= pg_ops.VMALLOC_START and addr < pg_ops.VMALLOC_END
|
/linux-master/kernel/module/ |
H A D | main.c | 1615 return __vmalloc_node_range(size, 1, VMALLOC_START, VMALLOC_END,
|
/linux-master/include/linux/ |
H A D | vmalloc.h | 288 #define VMALLOC_TOTAL (VMALLOC_END - VMALLOC_START)
|
/linux-master/fs/proc/ |
H A D | kcore.c | 678 if (MODULES_VADDR != VMALLOC_START && MODULES_END != VMALLOC_END) { 700 VMALLOC_END - VMALLOC_START, KCORE_VMALLOC);
|
/linux-master/arch/sparc/mm/ |
H A D | init_64.c | 1955 unsigned long VMALLOC_END = 0x0000010000000000UL; variable 1956 EXPORT_SYMBOL(VMALLOC_END); variable 2026 VMALLOC_END = ((sparc64_va_hole_bottom >> 1) + 2032 VMALLOC_START, VMALLOC_END);
|
/linux-master/arch/sparc/include/asm/ |
H A D | pgtable_64.h | 46 #define VMEMMAP_BASE VMALLOC_END 84 extern unsigned long VMALLOC_END;
|
/linux-master/arch/s390/mm/ |
H A D | dump_pagetables.c | 305 address_markers[VMALLOC_END_NR].start_address = VMALLOC_END;
|
/linux-master/arch/s390/include/asm/ |
H A D | pgtable.h | 89 extern unsigned long __bootdata_preserved(VMALLOC_END);
|