/linux-master/arch/arm/mm/ |
H A D | kasan_init.c | 245 kasan_populate_early_shadow(kasan_mem_to_shadow((void *)VMALLOC_START),
|
H A D | mmu.c | 992 (md->virtual < VMALLOC_START || md->virtual >= VMALLOC_END)) { 1338 addr < VMALLOC_START; addr += PMD_SIZE) 1372 * mappings above VMALLOC_START, except early fixmap, we might remove debug 1393 for (addr = VMALLOC_START; addr < (FIXADDR_TOP & PMD_MASK); addr += PMD_SIZE)
|
/linux-master/arch/s390/mm/ |
H A D | dump_pagetables.c | 304 address_markers[VMALLOC_NR].start_address = VMALLOC_START;
|
/linux-master/arch/arc/include/asm/ |
H A D | entry-compact.h | 151 brlo sp, VMALLOC_START, 88f variable
|
/linux-master/arch/mips/mm/ |
H A D | fault.c | 84 if (unlikely(address >= VMALLOC_START && address <= VMALLOC_END))
|
/linux-master/arch/nios2/include/asm/ |
H A D | pgtable.h | 27 #define VMALLOC_START CONFIG_NIOS2_KERNEL_MMU_REGION_BASE macro
|
/linux-master/arch/mips/include/asm/ |
H A D | pgtable-32.h | 96 #define VMALLOC_START MAP_BASE macro
|
/linux-master/include/linux/ |
H A D | vmalloc.h | 318 #define VMALLOC_TOTAL (VMALLOC_END - VMALLOC_START)
|
/linux-master/kernel/ |
H A D | vmcore_info.c | 149 vmcoreinfo_append_str("NUMBER(VMALLOC_START)=0x%lx\n", (unsigned long) VMALLOC_START);
|
/linux-master/arch/riscv/mm/ |
H A D | fault.c | 251 unlikely(addr >= VMALLOC_START && addr < VMALLOC_END)) {
|
H A D | ptdump.c | 392 address_markers[VMALLOC_START_NR].start_address = VMALLOC_START;
|
H A D | init.c | 142 print_ml("vmalloc", (unsigned long)VMALLOC_START, 1497 preallocate_pgd_pages_range(VMALLOC_START, VMALLOC_END, "vmalloc"); 1518 .start = VMALLOC_START,
|
/linux-master/mm/ |
H A D | vmalloc.c | 83 return addr >= VMALLOC_START && addr < VMALLOC_END; 1583 vstart = VMALLOC_START + rnd; 1923 if (vstart != VMALLOC_START || vend != VMALLOC_END || 2447 * If we had a constant VMALLOC_START and VMALLOC_END, we'd like to be able 2448 * to #define VMALLOC_SPACE (VMALLOC_END-VMALLOC_START). Guess 2559 addr -= VMALLOC_START & ~(VMAP_BLOCK_SIZE-1); 2599 VMALLOC_START, VMALLOC_END, 2907 BUG_ON(addr < VMALLOC_START); 2957 VMALLOC_START, VMALLOC_END, 3044 unsigned long addr = ALIGN(VMALLOC_START, alig [all...] |
/linux-master/arch/alpha/kernel/ |
H A D | core_titan.c | 557 if (addr >= VMALLOC_START) 566 if (addr >= VMALLOC_START)
|
/linux-master/arch/sh/mm/ |
H A D | fault.c | 183 if (!(address >= VMALLOC_START && address < __FAULT_ADDR_LIMIT))
|
/linux-master/arch/openrisc/include/asm/ |
H A D | pgtable.h | 91 #define VMALLOC_START (PAGE_OFFSET-0x04000000UL) macro
|
/linux-master/mm/kmsan/ |
H A D | shadow.c | 61 off = addr64 - VMALLOC_START;
|
/linux-master/arch/um/kernel/ |
H A D | um_arch.c | 389 start_vm = VMALLOC_START;
|
/linux-master/arch/microblaze/include/asm/ |
H A D | pgtable.h | 39 #define VMALLOC_START (CONFIG_KERNEL_START + CONFIG_LOWMEM_SIZE) macro
|
/linux-master/arch/loongarch/mm/ |
H A D | kasan_init.c | 262 kasan_populate_early_shadow(kasan_mem_to_shadow((void *)VMALLOC_START),
|
/linux-master/arch/arm64/mm/ |
H A D | ptdump.c | 371 { VMALLOC_START, "vmalloc() area" },
|
H A D | kasan_init.c | 348 BUILD_BUG_ON(VMALLOC_START != MODULES_END);
|
/linux-master/arch/arm/include/asm/ |
H A D | pgtable.h | 51 #define VMALLOC_START (((unsigned long)high_memory + VMALLOC_OFFSET) & ~(VMALLOC_OFFSET-1)) macro
|
/linux-master/arch/powerpc/mm/ptdump/ |
H A D | ptdump.c | 270 address_markers[i++].start_address = VMALLOC_START;
|
/linux-master/arch/x86/hyperv/ |
H A D | hv_init.c | 534 hv_hypercall_pg = __vmalloc_node_range(PAGE_SIZE, 1, VMALLOC_START,
|