/linux-master/arch/arm/mm/ |
H A D | fault-armv.c | 54 (pfn << PAGE_SHIFT) + PAGE_SIZE);
|
H A D | fault.c | 164 if (addr < PAGE_SIZE) {
|
H A D | flush.c | 48 : "r" (to), "r" (to + PAGE_SIZE - 1), "r" (zero) 55 unsigned long offset = vaddr & (PAGE_SIZE - 1); 214 i * PAGE_SIZE); 215 __cpuc_flush_dcache_area(addr, PAGE_SIZE); 222 __cpuc_flush_dcache_area(addr, PAGE_SIZE); 274 start += offset * PAGE_SIZE; 276 if (start + nr * PAGE_SIZE > vma->vm_end) 277 nr = (vma->vm_end - start) / PAGE_SIZE; 410 __cpuc_flush_dcache_area(page_address(page), PAGE_SIZE); local
|
H A D | init.c | 128 unsigned long pageblock_size = PAGE_SIZE * pageblock_nr_pages; 299 BUILD_BUG_ON(PKMAP_BASE + LAST_PKMAP * PAGE_SIZE > PAGE_OFFSET); 300 BUG_ON(PKMAP_BASE + LAST_PKMAP * PAGE_SIZE > PAGE_OFFSET); 481 start = round_down(start, PAGE_SIZE); 483 end = round_up(end, PAGE_SIZE);
|
H A D | ioremap.c | 113 return vmap_page_range(virt, virt + PAGE_SIZE, phys, 411 set_memory_ro((unsigned long)ptr, PAGE_ALIGN(size) / PAGE_SIZE);
|
H A D | kasan_init.c | 53 next = addr + PAGE_SIZE; 59 p = kasan_alloc_block_raw(PAGE_SIZE); 65 memset(p, KASAN_SHADOW_INIT, PAGE_SIZE); 107 kasan_alloc_block(PAGE_SIZE); 139 void *p = kasan_alloc_block(PAGE_SIZE); 293 set_pte_at(&init_mm, KASAN_SHADOW_START + i*PAGE_SIZE, 302 memset(kasan_early_shadow_page, 0, PAGE_SIZE);
|
H A D | mmu.c | 416 local_flush_tlb_kernel_range(vaddr, vaddr + PAGE_SIZE); 777 } while (pte++, addr += PAGE_SIZE, addr != end); 1158 map.length = PAGE_SIZE; 1344 #define SWAPPER_PG_DIR_SIZE (PAGE_SIZE + \ 1386 vectors = early_alloc(PAGE_SIZE * 2); 1442 map.length = PAGE_SIZE; 1452 map.length = PAGE_SIZE * 2; 1459 map.virtual = 0xffff0000 + PAGE_SIZE; 1460 map.length = PAGE_SIZE; 1752 map.length = PAGE_SIZE; [all...] |
H A D | nommu.c | 98 memblock_reserve(vectors_base, 2 * PAGE_SIZE); 165 zero_page = (void *)memblock_alloc(PAGE_SIZE, PAGE_SIZE); 168 __func__, PAGE_SIZE, PAGE_SIZE); 191 __cpuc_flush_dcache_area(page_address(page), PAGE_SIZE); local
|
H A D | pageattr.c | 36 * This function assumes that the range is mapped with PAGE_SIZE pages. 58 unsigned long end = PAGE_ALIGN(addr) + numpages * PAGE_SIZE; 104 return __change_memory_common(addr, PAGE_SIZE * numpages, 108 return __change_memory_common(addr, PAGE_SIZE * numpages,
|
H A D | pmsa-v7.c | 464 err |= mpu_setup_region(region++, vectors_base, ilog2(2 * PAGE_SIZE),
|
H A D | pmsa-v8.c | 266 subtract_range(mem, ARRAY_SIZE(mem), vectors_base, vectors_base + 2 * PAGE_SIZE); 267 subtract_range(io, ARRAY_SIZE(io), vectors_base, vectors_base + 2 * PAGE_SIZE); 301 err |= pmsav8_setup_vector(region++, vectors_base, vectors_base + 2 * PAGE_SIZE);
|
/linux-master/arch/arm64/include/asm/ |
H A D | assembler.h | 514 tst \src, #(PAGE_SIZE - 1)
|
H A D | elf.h | 123 #define ELF_EXEC_PAGESIZE PAGE_SIZE
|
H A D | fixmap.h | 46 FIX_FDT = FIX_FDT_END + DIV_ROUND_UP(MAX_FDT_SIZE, PAGE_SIZE) + 1, 76 #define NR_FIX_BTMAPS (SZ_256K / PAGE_SIZE)
|
H A D | hugetlb.h | 76 __flush_tlb_range(vma, start, end, PAGE_SIZE, false, 0);
|
H A D | kernel-pgtable.h | 59 #define INIT_DIR_SIZE (PAGE_SIZE * (EARLY_PAGES(SWAPPER_PGTABLE_LEVELS, KIMAGE_VADDR, _end, EXTRA_PAGE) \ 63 #define INIT_IDMAP_DIR_SIZE ((INIT_IDMAP_DIR_PAGES + EARLY_IDMAP_EXTRA_PAGES) * PAGE_SIZE) 66 #define INIT_IDMAP_FDT_SIZE ((INIT_IDMAP_FDT_PAGES + EARLY_IDMAP_EXTRA_FDT_PAGES) * PAGE_SIZE)
|
H A D | kvm_pkvm.h | 34 start = ALIGN_DOWN(start, PAGE_SIZE); 35 end = ALIGN(end, PAGE_SIZE); 128 return (2 * KVM_FFA_MBOX_NR_PAGES) + DIV_ROUND_UP(desc_max, PAGE_SIZE);
|
H A D | memory.h | 152 #define NVHE_STACKTRACE_SIZE ((OVERFLOW_STACK_SIZE + PAGE_SIZE) / 2 + sizeof(long)) 201 #define RESERVED_SWAPPER_OFFSET (PAGE_SIZE) 207 #define TRAMP_SWAPPER_OFFSET (2 * PAGE_SIZE) 412 u64 __addr = PAGE_OFFSET + (__idx * PAGE_SIZE); \ 417 u64 __idx = (__tag_reset((u64)x) - PAGE_OFFSET) / PAGE_SIZE; \
|
H A D | mte-def.h | 10 #define MTE_GRANULES_PER_PAGE (PAGE_SIZE / MTE_GRANULE_SIZE)
|
H A D | page-def.h | 15 #define PAGE_SIZE (_AC(1, UL) << PAGE_SHIFT) macro 16 #define PAGE_MASK (~(PAGE_SIZE-1))
|
H A D | pgalloc.h | 101 BUG_ON((unsigned long)p4d & (PAGE_SIZE-1));
|
H A D | pgtable-hwdef.h | 86 #define CONT_PTE_SIZE (CONT_PTES * PAGE_SIZE)
|
H A D | pgtable.h | 70 extern unsigned long empty_zero_page[PAGE_SIZE / sizeof(unsigned long)]; 627 #define pte_leaf_size(pte) (pte_cont(pte) ? CONT_PTE_SIZE : PAGE_SIZE) 826 return (pud_t *)PTR_ALIGN_DOWN(p4dp, PAGE_SIZE) + pud_index(addr); 950 return (p4d_t *)PTR_ALIGN_DOWN(pgdp, PAGE_SIZE) + p4d_index(addr); 1161 addr += PAGE_SIZE; 1174 addr += PAGE_SIZE; 1226 for (i = 0; i < nr; i++, address += PAGE_SIZE, ptep++) 1352 return PAGE_SIZE == SZ_4K;
|
H A D | processor.h | 65 #define TASK_SIZE_32 (UL(0x100000000) - PAGE_SIZE)
|
H A D | tlbflush.h | 68 * PAGE_SIZE. Used by TTL. 79 switch (PAGE_SIZE) { 288 (uaddr & PAGE_MASK) + PAGE_SIZE); 483 __flush_tlb_range(vma, start, end, PAGE_SIZE, false, TLBI_TTL_UNKNOWN); 490 if ((end - start) > (MAX_DVM_OPS * PAGE_SIZE)) {
|