/linux-master/arch/arc/kernel/ |
H A D | vmlinux.lds.S | 33 . = ALIGN(PAGE_SIZE); 59 . = ALIGN(PAGE_SIZE); 80 . = ALIGN(PAGE_SIZE); 99 RO_DATA(PAGE_SIZE) 105 RW_DATA(L1_CACHE_BYTES, PAGE_SIZE, THREAD_SIZE) 112 . = ALIGN(PAGE_SIZE); 122 . = ALIGN(PAGE_SIZE);
|
/linux-master/drivers/crypto/intel/qat/qat_common/ |
H A D | adf_gen4_pm_debugfs.c | 150 static_assert(sizeof(struct icp_qat_fw_init_admin_pm_info) < PAGE_SIZE); 166 pm_info = kmalloc(PAGE_SIZE, GFP_KERNEL); 170 pm_kv = kmalloc(PAGE_SIZE, GFP_KERNEL); 176 p_state_addr = dma_map_single(&GET_DEV(accel_dev), pm_info, PAGE_SIZE, 183 ret = adf_get_pm_info(accel_dev, p_state_addr, PAGE_SIZE); 184 dma_unmap_single(&GET_DEV(accel_dev), p_state_addr, PAGE_SIZE, local 192 len += scnprintf(&pm_kv[len], PAGE_SIZE - len, 195 pm_info_regs, PAGE_SIZE - len, 197 len += scnprintf(&pm_kv[len], PAGE_SIZE - len, "max_pwrreq: %#x\n", 199 len += scnprintf(&pm_kv[len], PAGE_SIZE [all...] |
/linux-master/arch/powerpc/kernel/ |
H A D | proc_powerpc.c | 21 return fixed_size_llseek(file, off, whence, PAGE_SIZE); 28 pde_data(file_inode(file)), PAGE_SIZE); 33 if ((vma->vm_end - vma->vm_start) > PAGE_SIZE) 38 PAGE_SIZE, vma->vm_page_prot); 57 proc_set_size(pde, PAGE_SIZE);
|
/linux-master/drivers/gpu/drm/nouveau/ |
H A D | nouveau_bo85b5.c | 62 0x031c, PAGE_SIZE, 63 0x0320, PAGE_SIZE, 64 0x0324, PAGE_SIZE, 69 src_offset += (PAGE_SIZE * line_count); 70 dst_offset += (PAGE_SIZE * line_count);
|
H A D | nouveau_bo90b5.c | 55 0x031c, PAGE_SIZE, 56 0x0320, PAGE_SIZE, 57 0x0324, PAGE_SIZE, 62 src_offset += (PAGE_SIZE * line_count); 63 dst_offset += (PAGE_SIZE * line_count);
|
/linux-master/drivers/gpu/drm/i915/gem/selftests/ |
H A D | i915_gem_object.c | 21 obj = i915_gem_object_create_shmem(i915, PAGE_SIZE); 45 nreal * PAGE_SIZE, 46 to_gt(i915)->ggtt->vm.total + PAGE_SIZE); 53 nreal, obj->base.size / PAGE_SIZE, err); 57 for (n = 0; n < obj->base.size / PAGE_SIZE; n++) {
|
/linux-master/arch/x86/include/asm/ |
H A D | page_64_types.h | 16 #define THREAD_SIZE (PAGE_SIZE << THREAD_SIZE_ORDER) 19 #define EXCEPTION_STKSZ (PAGE_SIZE << EXCEPTION_STACK_ORDER) 22 #define IRQ_STACK_SIZE (PAGE_SIZE << IRQ_STACK_ORDER) 61 #define task_size_max() ((_AC(1,UL) << __VIRTUAL_MASK_SHIFT) - PAGE_SIZE) 65 #define DEFAULT_MAP_WINDOW ((1UL << 47) - PAGE_SIZE)
|
/linux-master/arch/powerpc/include/asm/ |
H A D | fixmap.h | 49 FIX_EARLY_DEBUG_BASE = FIX_EARLY_DEBUG_TOP+(ALIGN(SZ_128K, PAGE_SIZE)/PAGE_SIZE)-1, 56 #define FIX_IMMR_SIZE (512 * 1024 / PAGE_SIZE) 63 #define FIX_IMMR_SIZE (SZ_2M / PAGE_SIZE) 72 #define NR_FIX_BTMAPS (SZ_256K / PAGE_SIZE)
|
/linux-master/arch/openrisc/include/asm/ |
H A D | page.h | 23 #define PAGE_SIZE (1 << PAGE_SHIFT) macro 25 #define PAGE_SIZE (1UL << PAGE_SHIFT) macro 27 #define PAGE_MASK (~(PAGE_SIZE-1)) 39 #define clear_page(page) memset((page), 0, PAGE_SIZE) 40 #define copy_page(to, from) memcpy((to), (from), PAGE_SIZE)
|
/linux-master/arch/x86/kernel/ |
H A D | vmlinux.lds.S | 85 . = ALIGN(PAGE_SIZE); \ 94 . = ALIGN(PAGE_SIZE); \ 161 . = ALIGN(PAGE_SIZE); 164 RO_DATA(PAGE_SIZE) 180 PAGE_ALIGNED_DATA(PAGE_SIZE) 198 . = ALIGN(PAGE_SIZE); 216 . = __vvar_beginning_hack + PAGE_SIZE; 219 . = ALIGN(__vvar_page + PAGE_SIZE, PAGE_SIZE); 222 . = ALIGN(PAGE_SIZE); [all...] |
/linux-master/arch/arm64/kernel/ |
H A D | vmlinux.lds.S | 18 . = ALIGN(PAGE_SIZE); \ 22 . = ALIGN(PAGE_SIZE); \ 27 . = ALIGN(PAGE_SIZE); \ 42 . = ALIGN(PAGE_SIZE); \ 51 #define SBSS_ALIGN PAGE_SIZE 78 . = ALIGN(PAGE_SIZE); \ 85 . = ALIGN(PAGE_SIZE); \ 116 . = ALIGN(PAGE_SIZE); \ 119 . = ALIGN(PAGE_SIZE); \ 190 RO_DATA(PAGE_SIZE) [all...] |
/linux-master/tools/testing/selftests/proc/ |
H A D | thread-self.c | 49 const int PAGE_SIZE = sysconf(_SC_PAGESIZE); local 56 stack = mmap(NULL, 2 * PAGE_SIZE, PROT_READ|PROT_WRITE, MAP_PRIVATE|MAP_ANONYMOUS, -1, 0); 59 pid = clone(f, stack + PAGE_SIZE, CLONE_THREAD|CLONE_SIGHAND|CLONE_VM, (void *)1);
|
/linux-master/arch/hexagon/include/asm/ |
H A D | mem-layout.h | 65 #define VMALLOC_OFFSET PAGE_SIZE 97 #define PKMAP_BASE (FIXADDR_START-PAGE_SIZE*LAST_PKMAP) 103 #define VMALLOC_END (PKMAP_BASE-PAGE_SIZE*2)
|
/linux-master/arch/alpha/kernel/ |
H A D | vmlinux.lds.S | 40 __init_begin = ALIGN(PAGE_SIZE); 41 INIT_TEXT_SECTION(PAGE_SIZE) 44 /* Align to THREAD_SIZE rather than PAGE_SIZE here so any padding page 52 RW_DATA(L1_CACHE_BYTES, PAGE_SIZE, THREAD_SIZE)
|
/linux-master/lib/ |
H A D | test_user_copy.c | 52 if (test(size < 2 * PAGE_SIZE, "buffer too small")) 62 start = PAGE_SIZE - (size / 2); 194 kmem = kmalloc(PAGE_SIZE * 2, GFP_KERNEL); 198 user_addr = vm_mmap(NULL, 0, PAGE_SIZE * 2, 213 memset(kmem, 0x3a, PAGE_SIZE * 2); 214 ret |= test(copy_to_user(usermem, kmem, PAGE_SIZE), 216 memset(kmem, 0x0, PAGE_SIZE); 217 ret |= test(copy_from_user(kmem, usermem, PAGE_SIZE), 219 ret |= test(memcmp(kmem, kmem + PAGE_SIZE, PAGE_SIZE), [all...] |
/linux-master/arch/powerpc/mm/kasan/ |
H A D | init_book3s_64.c | 32 k_start = ALIGN_DOWN((unsigned long)kasan_mem_to_shadow(start), PAGE_SIZE); 33 k_end = ALIGN((unsigned long)kasan_mem_to_shadow(end), PAGE_SIZE); 35 va = memblock_alloc(k_end - k_start, PAGE_SIZE); 36 for (k_cur = k_start; k_cur < k_end; k_cur += PAGE_SIZE, va += PAGE_SIZE) 93 memset(kasan_early_shadow_page, 0, PAGE_SIZE);
|
/linux-master/arch/sparc/kernel/ |
H A D | vmlinux.lds.S | 61 RO_DATA(PAGE_SIZE) 81 . = ALIGN(PAGE_SIZE); 82 __init_begin = ALIGN(PAGE_SIZE); 83 INIT_TEXT_SECTION(PAGE_SIZE) 173 . = ALIGN(PAGE_SIZE); 182 . = ALIGN(PAGE_SIZE);
|
/linux-master/arch/xtensa/mm/ |
H A D | kasan_init.c | 37 unsigned long n_pages = (end - start) / PAGE_SIZE; 42 pte_t *pte = memblock_alloc(n_pages * sizeof(pte_t), PAGE_SIZE); 46 __func__, n_pages * sizeof(pte_t), PAGE_SIZE); 55 memblock_phys_alloc_range(PAGE_SIZE, PAGE_SIZE, 97 memset(kasan_early_shadow_page, 0, PAGE_SIZE);
|
/linux-master/arch/arm/include/asm/ |
H A D | page-nommu.h | 11 #define clear_page(page) memset((page), 0, PAGE_SIZE) 12 #define copy_page(to,from) memcpy((to), (from), PAGE_SIZE)
|
/linux-master/arch/x86/realmode/rm/ |
H A D | realmode.lds.S | 36 . = ALIGN(PAGE_SIZE); 54 . = ALIGN(PAGE_SIZE);
|
/linux-master/arch/um/kernel/ |
H A D | mem.c | 62 memset(empty_zero_page, 0, PAGE_SIZE); 86 pte_t *pte = (pte_t *) memblock_alloc_low(PAGE_SIZE, 87 PAGE_SIZE); 90 __func__, PAGE_SIZE, PAGE_SIZE); 101 pmd_t *pmd_table = (pmd_t *) memblock_alloc_low(PAGE_SIZE, PAGE_SIZE); 104 __func__, PAGE_SIZE, PAGE_SIZE); 152 v = (unsigned long) memblock_alloc_low(size, PAGE_SIZE); [all...] |
/linux-master/arch/m68k/mm/ |
H A D | sun3kmap.c | 55 phys += PAGE_SIZE; 56 virt += PAGE_SIZE; 72 offset = phys & (PAGE_SIZE-1); 73 phys &= ~(PAGE_SIZE-1); 85 pages = size / PAGE_SIZE; 92 seg_pages = (SUN3_PMEG_SIZE - (virt & SUN3_PMEG_MASK)) / PAGE_SIZE; 99 phys += seg_pages * PAGE_SIZE; 100 virt += seg_pages * PAGE_SIZE;
|
/linux-master/fs/freevxfs/ |
H A D | vxfs_immed.c | 38 memcpy_to_page(folio_page(folio, i), 0, src, PAGE_SIZE); local 39 src += PAGE_SIZE;
|
/linux-master/arch/x86/mm/ |
H A D | numa_32.c | 43 high_memory = (void *) __va(highstart_pfn * PAGE_SIZE - 1) + 1; 45 high_memory = (void *) __va(max_low_pfn * PAGE_SIZE - 1) + 1;
|
/linux-master/drivers/misc/lkdtm/ |
H A D | usercopy.c | 81 user_addr = vm_mmap(NULL, 0, PAGE_SIZE, 127 vm_munmap(user_addr, PAGE_SIZE); 149 user_addr = vm_mmap(NULL, 0, PAGE_SIZE, 192 vm_munmap(user_addr, PAGE_SIZE); 225 user_alloc = vm_mmap(NULL, 0, PAGE_SIZE, 269 vm_munmap(user_alloc, PAGE_SIZE); 315 user_addr = vm_mmap(NULL, 0, PAGE_SIZE, 334 unconst + PAGE_SIZE)) { 342 vm_munmap(user_addr, PAGE_SIZE); 346 * This expects "kaddr" to point to a PAGE_SIZE allocatio [all...] |