/linux-master/arch/arm64/mm/ |
H A D | kasan_init.c | 26 static pgd_t tmp_pg_dir[PTRS_PER_PTE] __initdata __aligned(PAGE_SIZE); 37 void *p = memblock_alloc_try_nid(PAGE_SIZE, PAGE_SIZE, 42 __func__, PAGE_SIZE, PAGE_SIZE, node, 50 void *p = memblock_alloc_try_nid_raw(PAGE_SIZE, PAGE_SIZE, 56 __func__, PAGE_SIZE, PAGE_SIZE, node, 126 memset(__va(page_phys), KASAN_SHADOW_INIT, PAGE_SIZE); local [all...] |
H A D | mmu.c | 65 unsigned long empty_zero_page[PAGE_SIZE / sizeof(unsigned long)] __page_aligned_bss; 114 phys = memblock_phys_alloc_range(PAGE_SIZE, PAGE_SIZE, 0, 126 memset(ptr, 0, PAGE_SIZE); 193 phys += PAGE_SIZE; 194 } while (ptep++, addr += PAGE_SIZE, addr != end); 576 kfence_pool = memblock_phys_alloc(KFENCE_POOL_SIZE, PAGE_SIZE); 698 size += PAGE_SIZE; 735 for (i = 0; i < DIV_ROUND_UP(entry_tramp_text_size(), PAGE_SIZE); i++) 737 pa_start + i * PAGE_SIZE, pro 856 PAGE_SIZE, altmap); local [all...] |
H A D | pageattr.c | 49 * This function assumes that the range is mapped with PAGE_SIZE pages. 71 unsigned long size = PAGE_SIZE * numpages; 112 PAGE_SIZE, set_mask, clear_mask); 156 return __change_memory_common(addr, PAGE_SIZE * numpages, 160 return __change_memory_common(addr, PAGE_SIZE * numpages, 177 PAGE_SIZE, change_page_range, &data); 192 PAGE_SIZE, change_page_range, &data);
|
H A D | pgd.c | 22 if (PGD_SIZE == PAGE_SIZE)
|
H A D | proc.S | 239 add pte, temp_pte, #PAGE_SIZE * (\level + 2) 246 add cur_\type\()p, temp_pte, #PAGE_SIZE * (\level + 2)
|
H A D | ptdump.c | 206 st->uxn_pages += (addr - st->start_address) / PAGE_SIZE; 221 st->wx_pages += (addr - st->start_address) / PAGE_SIZE;
|
H A D | trans_pgd.c | 78 } while (dst_ptep++, src_ptep++, addr += PAGE_SIZE, addr != end);
|
/linux-master/arch/csky/abiv2/ |
H A D | cacheflush.c | 27 icache_inv_range(address, address + nr*PAGE_SIZE); 30 i * PAGE_SIZE); 32 dcache_wb_range(addr, addr + PAGE_SIZE); 34 icache_inv_range(addr, addr + PAGE_SIZE);
|
/linux-master/arch/csky/include/asm/ |
H A D | memory.h | 13 #define VMALLOC_START (PAGE_OFFSET + LOWMEM_LIMIT + (PAGE_SIZE * 8)) 14 #define VMALLOC_END (PKMAP_BASE - (PAGE_SIZE * 2)) 22 #define FIXADDR_TCM _AC(FIXADDR_TOP - (TCM_NR_PAGES * PAGE_SIZE), UL)
|
H A D | page.h | 14 #define PAGE_SIZE (_AC(1, UL) << PAGE_SHIFT) macro 15 #define PAGE_MASK (~(PAGE_SIZE - 1)) 16 #define THREAD_SIZE (PAGE_SIZE * 2) 43 #define clear_page(page) memset((page), 0, PAGE_SIZE) 44 #define copy_page(to, from) memcpy((to), (from), PAGE_SIZE)
|
H A D | pgalloc.h | 36 for (i = 0; i < PAGE_SIZE/sizeof(pte_t); i++)
|
H A D | pgtable.h | 22 #define PTRS_PER_PGD (PAGE_SIZE / sizeof(pgd_t)) 24 #define PTRS_PER_PTE (PAGE_SIZE / sizeof(pte_t)) 79 extern unsigned long empty_zero_page[PAGE_SIZE / sizeof(unsigned long)];
|
H A D | processor.h | 29 #define TASK_SIZE (PAGE_OFFSET - (PAGE_SIZE * 8))
|
H A D | shmparam.h | 6 #define SHMLBA (4 * PAGE_SIZE)
|
/linux-master/arch/csky/kernel/ |
H A D | ptrace.c | 480 (int) (((unsigned long) current) + 2 * PAGE_SIZE));
|
H A D | vdso.c | 79 ret = install_special_mapping(mm, vdso_base, PAGE_SIZE, 94 (long)vma->vm_mm->context.vdso + PAGE_SIZE))
|
H A D | vmlinux.lds.S | 27 . = ALIGN(PAGE_SIZE); 45 . = ALIGN(PAGE_SIZE); 47 INIT_TEXT_SECTION(PAGE_SIZE) 48 INIT_DATA_SECTION(PAGE_SIZE) 50 . = ALIGN(PAGE_SIZE); 54 RO_DATA(PAGE_SIZE) 55 RW_DATA(L1_CACHE_BYTES, PAGE_SIZE, THREAD_SIZE) 60 . = ALIGN(PAGE_SIZE); 80 #define ITCM_SIZE CONFIG_ITCM_NR_PAGES * PAGE_SIZE 101 . = ALIGN(PAGE_SIZE); [all...] |
/linux-master/arch/csky/kernel/vdso/ |
H A D | vdso.S | 10 .balign PAGE_SIZE 13 .balign PAGE_SIZE
|
H A D | vdso.lds.S | 9 PROVIDE(_vdso_data = . + PAGE_SIZE);
|
/linux-master/arch/csky/mm/ |
H A D | cachev1.c | 53 if (unlikely((end - start) >= PAGE_SIZE) ||
|
H A D | dma-mapping.c | 25 if (offset + len > PAGE_SIZE) 26 len = PAGE_SIZE - offset; 42 start += PAGE_SIZE;
|
H A D | highmem.c | 29 fixrange_init(vaddr, vaddr + PAGE_SIZE*LAST_PKMAP, swapper_pg_dir);
|
H A D | init.c | 41 unsigned long empty_zero_page[PAGE_SIZE / sizeof(unsigned long)] 176 pte = (pte_t *) memblock_alloc_low(PAGE_SIZE, PAGE_SIZE); 179 __func__, PAGE_SIZE, 180 PAGE_SIZE);
|
H A D | tcm.c | 51 paddr = paddr + PAGE_SIZE; 70 paddr = paddr + PAGE_SIZE; 133 u32 pool_size = (u32) (TCM_NR_PAGES * PAGE_SIZE) 139 u32 pool_size = (u32) (CONFIG_DTCM_NR_PAGES * PAGE_SIZE)
|
H A D | tlb.c | 16 #define TLB_ENTRY_SIZE (PAGE_SIZE * 2) 70 start += 2*PAGE_SIZE; 83 start += 2*PAGE_SIZE; 110 start += 2*PAGE_SIZE; 123 start += 2*PAGE_SIZE;
|