/linux-master/arch/mips/include/asm/ |
H A D | pgtable-64.h | 39 * fault address - VMALLOC_START. 141 #define VMALLOC_START (MAP_BASE + (2 * PAGE_SIZE)) macro 148 VMALLOC_START != CKSSEG
|
H A D | pgtable-32.h | 96 #define VMALLOC_START MAP_BASE macro
|
/linux-master/arch/um/include/asm/ |
H A D | pgtable.h | 49 #define VMALLOC_START ((end_iomem + VMALLOC_OFFSET) & ~(VMALLOC_OFFSET-1)) macro 52 #define MODULES_VADDR VMALLOC_START
|
/linux-master/arch/arm/mm/ |
H A D | ioremap.c | 124 memcpy(pgd_offset(mm, VMALLOC_START), 125 pgd_offset_k(VMALLOC_START), 127 pgd_index(VMALLOC_START)));
|
H A D | dump.c | 441 address_markers[4].start_address = VMALLOC_START; 443 address_markers[2].start_address = VMALLOC_START;
|
H A D | kasan_init.c | 245 kasan_populate_early_shadow(kasan_mem_to_shadow((void *)VMALLOC_START),
|
/linux-master/arch/x86/mm/ |
H A D | kasan_init_64.c | 393 kasan_mem_to_shadow((void *)VMALLOC_START)); 402 kasan_mem_to_shadow((void *)VMALLOC_START), 406 kasan_mem_to_shadow((void *)VMALLOC_START),
|
H A D | init_32.c | 751 BUILD_BUG_ON(VMALLOC_START >= VMALLOC_END); 759 BUG_ON(VMALLOC_START >= VMALLOC_END); 760 BUG_ON((unsigned long)high_memory > VMALLOC_START);
|
H A D | dump_pagetables.c | 455 address_markers[VMALLOC_START_NR].start_address = VMALLOC_START; 466 address_markers[VMALLOC_START_NR].start_address = VMALLOC_START;
|
/linux-master/arch/powerpc/kernel/ |
H A D | module.c | 125 return __module_alloc(size, VMALLOC_START, VMALLOC_END, false);
|
/linux-master/arch/powerpc/mm/kasan/ |
H A D | init_book3e_64.c | 118 kasan_remove_zero_shadow((void *)VMALLOC_START, VMALLOC_SIZE);
|
H A D | init_32.c | 114 unsigned long k_start = (unsigned long)kasan_mem_to_shadow((void *)VMALLOC_START);
|
/linux-master/arch/nios2/mm/ |
H A D | fault.c | 67 if (unlikely(address >= VMALLOC_START && address <= VMALLOC_END)) {
|
/linux-master/arch/openrisc/mm/ |
H A D | fault.c | 77 if (address >= VMALLOC_START &&
|
/linux-master/arch/arc/mm/ |
H A D | fault.c | 89 if (address >= VMALLOC_START && !user_mode(regs)) {
|
/linux-master/arch/sh/mm/ |
H A D | init.c | 376 (unsigned long)VMALLOC_START, VMALLOC_END, 377 (VMALLOC_END - VMALLOC_START) >> 20,
|
/linux-master/arch/alpha/include/asm/ |
H A D | pgtable.h | 53 #define VMALLOC_START 0xfffffe0000000000 macro 55 #define VMALLOC_START (-2*PGDIR_SIZE) macro
|
/linux-master/arch/riscv/mm/ |
H A D | kasan_init.c | 495 (void *)kasan_mem_to_shadow((void *)VMALLOC_START)); 499 (void *)kasan_mem_to_shadow((void *)VMALLOC_START), 506 kasan_populate_early_shadow((void *)kasan_mem_to_shadow((void *)VMALLOC_START),
|
/linux-master/arch/powerpc/include/asm/book3s/32/ |
H A D | pgtable.h | 168 #define IOREMAP_START VMALLOC_START 183 * from VMALLOC_START). For this reason we have ioremap_bot to check when 190 #define VMALLOC_START ((((long)high_memory + VMALLOC_OFFSET) & ~(VMALLOC_OFFSET-1))) macro
|
/linux-master/arch/s390/boot/ |
H A D | vmem.c | 118 untracked_end = VMALLOC_START; 120 kasan_populate(VMALLOC_START, MODULES_END, POPULATE_KASAN_SHALLOW);
|
/linux-master/arch/riscv/kernel/probes/ |
H A D | kprobes.c | 110 return __vmalloc_node_range(PAGE_SIZE, 1, VMALLOC_START, VMALLOC_END,
|
/linux-master/arch/csky/mm/ |
H A D | fault.c | 214 if (unlikely((addr >= VMALLOC_START) && (addr <= VMALLOC_END))) {
|
/linux-master/arch/m68k/mm/ |
H A D | mcfmmu.c | 25 #define KMAPAREA(x) ((x >= VMALLOC_START) && (x < KMAP_END))
|
/linux-master/arch/powerpc/mm/ |
H A D | mem.c | 335 VMALLOC_START, VMALLOC_END);
|
/linux-master/arch/mips/mm/ |
H A D | fault.c | 84 if (unlikely(address >= VMALLOC_START && address <= VMALLOC_END))
|