/linux-master/arch/powerpc/include/asm/ |
H A D | page_32.h | 46 for (i = 0; i < PAGE_SIZE / L1_CACHE_BYTES; i++, addr += L1_CACHE_BYTES)
|
H A D | kfence.h | 39 flush_tlb_kernel_range(addr, addr + PAGE_SIZE);
|
/linux-master/arch/loongarch/vdso/ |
H A D | vgetcpu.c | 24 return (struct vdso_pcpu_data *)(get_vdso_data() + VVAR_LOONGARCH_PAGES_START * PAGE_SIZE);
|
/linux-master/arch/arm/kernel/ |
H A D | vmlinux-xip.lds.S | 61 RO_DATA(PAGE_SIZE) 116 RW_DATA(L1_CACHE_BYTES, PAGE_SIZE, THREAD_SIZE) 122 . = ALIGN(PAGE_SIZE); 146 . = ALIGN(PAGE_SIZE);
|
/linux-master/arch/m68k/atari/ |
H A D | stram.c | 98 PAGE_SIZE); 101 __func__, pool_size, PAGE_SIZE); 125 stram_pool.start = PAGE_SIZE; 170 PAGE_SIZE, NULL, NULL);
|
/linux-master/arch/m68k/mm/ |
H A D | mcfmmu.c | 45 empty_zero_page = memblock_alloc(PAGE_SIZE, PAGE_SIZE); 48 __func__, PAGE_SIZE, PAGE_SIZE); 54 size = (size + PAGE_SIZE) & ~(PAGE_SIZE-1); 55 next_pgtable = (unsigned long) memblock_alloc(size, PAGE_SIZE); 58 __func__, size, PAGE_SIZE); 77 address += PAGE_SIZE;
|
/linux-master/arch/powerpc/kernel/ |
H A D | crash_dump.c | 81 csize = min_t(size_t, csize, PAGE_SIZE); 88 vaddr = ioremap_cache(paddr, PAGE_SIZE); 126 for (addr = begin; addr < end; addr += PAGE_SIZE) { 128 if (addr <= rtas_end && ((addr + PAGE_SIZE) > rtas_start))
|
H A D | vmlinux.lds.S | 36 #if STRICT_ALIGN_SIZE < PAGE_SIZE 130 . = ALIGN(PAGE_SIZE); 135 RO_DATA(PAGE_SIZE) 275 . = ALIGN(PAGE_SIZE); 289 . = ALIGN(PAGE_SIZE); 358 . = ALIGN(PAGE_SIZE); 365 . = ALIGN(PAGE_SIZE); 380 PAGE_ALIGNED_DATA(PAGE_SIZE) 391 . = ALIGN(PAGE_SIZE); 398 . = ALIGN(PAGE_SIZE); [all...] |
/linux-master/include/linux/ |
H A D | mm_types_task.h | 40 #if (BITS_PER_LONG > 32) || (PAGE_SIZE >= 65536)
|
/linux-master/mm/ |
H A D | process_vm_access.c | 36 size_t copy = PAGE_SIZE - offset; 59 #define PVM_MAX_USER_PAGES (PVM_MAX_KMALLOC_PAGES * PAGE_SIZE / sizeof(struct page *)) 90 nr_pages = (addr + len - 1) / PAGE_SIZE - addr / PAGE_SIZE + 1; 114 bytes = pinned_pages * PAGE_SIZE - start_offset; 124 pa += pinned_pages * PAGE_SIZE; 176 / PAGE_SIZE - (unsigned long)rvec[i].iov_base 177 / PAGE_SIZE + 1; 188 process_pages = kmalloc(min_t(size_t, PVM_MAX_KMALLOC_PAGES * PAGE_SIZE,
|
/linux-master/security/integrity/ima/ |
H A D | ima_kexec.c | 84 struct kexec_buf kbuf = { .image = image, .buf_align = PAGE_SIZE, 100 if (binary_runtime_size >= ULONG_MAX - PAGE_SIZE) 104 PAGE_SIZE / 2, PAGE_SIZE);
|
/linux-master/arch/riscv/kernel/ |
H A D | vmlinux.lds.S | 36 . = ALIGN(PAGE_SIZE); 132 RW_DATA(L1_CACHE_BYTES, PAGE_SIZE, THREAD_ALIGN) 159 BSS_SECTION(PAGE_SIZE, PAGE_SIZE, 0)
|
/linux-master/arch/x86/kvm/ |
H A D | kvm_onhyperv.h | 26 *p_hv_pa_pg = kzalloc(PAGE_SIZE, GFP_KERNEL_ACCOUNT);
|
/linux-master/arch/xtensa/include/asm/ |
H A D | cacheflush.h | 69 #if defined(CONFIG_MMU) && (DCACHE_WAY_SIZE > PAGE_SIZE) 78 #if defined(CONFIG_MMU) && (ICACHE_WAY_SIZE > PAGE_SIZE) 95 ((DCACHE_WAY_SIZE > PAGE_SIZE) || defined(CONFIG_SMP)) 168 #if defined(CONFIG_MMU) && (DCACHE_WAY_SIZE > PAGE_SIZE)
|
/linux-master/arch/x86/boot/compressed/ |
H A D | pgtable_64.c | 65 bios_start = round_down(bios_start, PAGE_SIZE); 86 new = round_down(new, PAGE_SIZE); 190 memcpy(trampoline_32bit, (void *)src, PAGE_SIZE); 198 memcpy(pgtable, trampoline_32bit, PAGE_SIZE);
|
/linux-master/arch/sparc/include/asm/ |
H A D | page_32.h | 15 #define PAGE_SIZE (_AC(1, UL) << PAGE_SHIFT) macro 16 #define PAGE_MASK (~(PAGE_SIZE-1)) 20 #define clear_page(page) memset((void *)(page), 0, PAGE_SIZE) 21 #define copy_page(to,from) memcpy((void *)(to), (void *)(from), PAGE_SIZE)
|
/linux-master/fs/xfs/ |
H A D | xfs_buf_mem.h | 9 #define XMBUF_BLOCKSIZE (PAGE_SIZE)
|
/linux-master/lib/ |
H A D | test_vmalloc.c | 90 * Maximum 1024 pages, if PAGE_SIZE is 4096. 97 size = ((rnd % 10) + 1) * PAGE_SIZE; 122 ptr = __vmalloc_node(PAGE_SIZE, align, GFP_KERNEL|__GFP_ZERO, 0, 139 ptr = __vmalloc_node(5 * PAGE_SIZE, THREAD_ALIGN << 1, 159 p = vmalloc(n * PAGE_SIZE); 183 ptr[i] = vmalloc(1 * PAGE_SIZE); 186 ptr_1 = vmalloc(100 * PAGE_SIZE); 190 ptr_2 = vmalloc(1 * PAGE_SIZE); 222 junk_length *= (32 * 1024 * 1024 / PAGE_SIZE); 235 ptr[i] = vmalloc(1 * PAGE_SIZE); [all...] |
/linux-master/fs/xfs/scrub/ |
H A D | xfile.h | 22 #define XFILE_MAX_FOLIO_SIZE (PAGE_SIZE << MAX_PAGECACHE_ORDER)
|
/linux-master/arch/arm/mm/ |
H A D | flush.c | 48 : "r" (to), "r" (to + PAGE_SIZE - 1), "r" (zero) 55 unsigned long offset = vaddr & (PAGE_SIZE - 1); 214 i * PAGE_SIZE); 215 __cpuc_flush_dcache_area(addr, PAGE_SIZE); 222 __cpuc_flush_dcache_area(addr, PAGE_SIZE); 274 start += offset * PAGE_SIZE; 276 if (start + nr * PAGE_SIZE > vma->vm_end) 277 nr = (vma->vm_end - start) / PAGE_SIZE; 410 __cpuc_flush_dcache_area(page_address(page), PAGE_SIZE); local
|
/linux-master/fs/ecryptfs/ |
H A D | mmap.c | 113 loff_t num_extents_per_page = (PAGE_SIZE 129 memset(page_virt, 0, PAGE_SIZE); 190 PAGE_SIZE, 207 page, page->index, 0, PAGE_SIZE, 242 if ((i_size_read(inode) / PAGE_SIZE) != page->index) 244 end_byte_in_page = i_size_read(inode) % PAGE_SIZE; 247 zero_user_segment(page, end_byte_in_page, PAGE_SIZE); 287 page, index, 0, PAGE_SIZE, mapping->host); 313 page, index, 0, PAGE_SIZE, 327 zero_user(page, 0, PAGE_SIZE); [all...] |
/linux-master/fs/btrfs/ |
H A D | zlib.c | 24 #define ZLIB_DFLTCC_BUF_SIZE (4 * PAGE_SIZE) 80 workspace->buf = kmalloc(PAGE_SIZE, GFP_KERNEL); 81 workspace->buf_size = PAGE_SIZE; 109 const unsigned long max_out = nr_dest_pages * PAGE_SIZE; 136 workspace->strm.avail_out = PAGE_SIZE; 145 in_buf_pages = min(DIV_ROUND_UP(bytes_left, PAGE_SIZE), 146 workspace->buf_size / PAGE_SIZE); 158 copy_page(workspace->buf + i * PAGE_SIZE, 160 start += PAGE_SIZE; 171 start += PAGE_SIZE; [all...] |
/linux-master/arch/sh/mm/ |
H A D | init.c | 140 pmd = memblock_alloc(PAGE_SIZE, PAGE_SIZE); 143 __func__, PAGE_SIZE, PAGE_SIZE); 156 pte = memblock_alloc(PAGE_SIZE, PAGE_SIZE); 159 __func__, PAGE_SIZE, PAGE_SIZE); 263 memblock_reserve(start, (PFN_PHYS(start_pfn) + PAGE_SIZE - 1) - start); 358 memset(empty_zero_page, 0, PAGE_SIZE); [all...] |
/linux-master/drivers/gpu/drm/i915/gt/ |
H A D | selftest_reset.c | 45 tmp = kmalloc(PAGE_SIZE, GFP_KERNEL); 96 PAGE_SIZE); 101 memset_io(s, STACK_MAGIC, PAGE_SIZE); 104 if (i915_memcpy_from_wc(tmp, in, PAGE_SIZE)) 106 crc[page] = crc32_le(0, in, PAGE_SIZE); 111 ggtt->vm.clear_range(&ggtt->vm, ggtt->error_capture.start, PAGE_SIZE); 139 PAGE_SIZE); 142 if (i915_memcpy_from_wc(tmp, in, PAGE_SIZE)) 144 x = crc32_le(0, in, PAGE_SIZE); 153 igt_hexdump(in, PAGE_SIZE); [all...] |
/linux-master/include/asm-generic/ |
H A D | tlb.h | 75 * freeing. __tlb_remove_page() assumes PAGE_SIZE. Both will return a 208 ((PAGE_SIZE - sizeof(struct mmu_table_batch)) / sizeof(void *)) 259 ((PAGE_SIZE - sizeof(struct mmu_gather_batch)) / sizeof(void *)) 479 return __tlb_remove_page_size(tlb, page, delay_rmap, PAGE_SIZE); 488 return tlb_remove_page_size(tlb, page, PAGE_SIZE); 617 tlb_flush_pte_range(tlb, address, PAGE_SIZE); \ 631 tlb_flush_pte_range(tlb, address, PAGE_SIZE * nr); 637 address += PAGE_SIZE; 698 * For now w.r.t page table cache, mark the range_size as PAGE_SIZE 704 tlb_flush_pmd_range(tlb, address, PAGE_SIZE); \ [all...] |