/linux-master/arch/powerpc/kernel/ |
H A D | vmlinux.lds.S | 36 #if STRICT_ALIGN_SIZE < PAGE_SIZE 130 . = ALIGN(PAGE_SIZE); 135 RO_DATA(PAGE_SIZE) 275 . = ALIGN(PAGE_SIZE); 289 . = ALIGN(PAGE_SIZE); 358 . = ALIGN(PAGE_SIZE); 365 . = ALIGN(PAGE_SIZE); 380 PAGE_ALIGNED_DATA(PAGE_SIZE) 391 . = ALIGN(PAGE_SIZE); 398 . = ALIGN(PAGE_SIZE); [all...] |
/linux-master/include/asm-generic/ |
H A D | tlb.h | 75 * freeing. __tlb_remove_page() assumes PAGE_SIZE. Both will return a 208 ((PAGE_SIZE - sizeof(struct mmu_table_batch)) / sizeof(void *)) 259 ((PAGE_SIZE - sizeof(struct mmu_gather_batch)) / sizeof(void *)) 479 return __tlb_remove_page_size(tlb, page, delay_rmap, PAGE_SIZE); 488 return tlb_remove_page_size(tlb, page, PAGE_SIZE); 617 tlb_flush_pte_range(tlb, address, PAGE_SIZE); \ 631 tlb_flush_pte_range(tlb, address, PAGE_SIZE * nr); 637 address += PAGE_SIZE; 698 * For now w.r.t page table cache, mark the range_size as PAGE_SIZE 704 tlb_flush_pmd_range(tlb, address, PAGE_SIZE); \ [all...] |
/linux-master/mm/kmsan/ |
H A D | hooks.c | 163 nr = (end - start) / PAGE_SIZE; 165 for (int i = 0; i < nr; i++, off += PAGE_SIZE, clean = i) { 174 vmalloc_shadow(start + off + PAGE_SIZE), prot, &shadow, 183 vmalloc_origin(start + off + PAGE_SIZE), prot, &origin, 188 vmalloc_shadow(start + off + PAGE_SIZE)); 210 vmalloc_shadow(start + clean * PAGE_SIZE)); 213 vmalloc_origin(start + clean * PAGE_SIZE)); 230 nr = (end - start) / PAGE_SIZE; 235 i++, v_shadow += PAGE_SIZE, v_origin += PAGE_SIZE) { [all...] |
/linux-master/arch/arm64/kvm/hyp/nvhe/ |
H A D | mm.c | 56 /* The allocated size is always a multiple of PAGE_SIZE */ 125 for (virt_addr = start; virt_addr < end; virt_addr += PAGE_SIZE) { 129 err = kvm_pgtable_hyp_map(&pkvm_pgtable, virt_addr, PAGE_SIZE, 156 start = ALIGN_DOWN((u64)hyp_phys_to_page(start), PAGE_SIZE); 302 return kvm_pgtable_walk(&pkvm_pgtable, addr, PAGE_SIZE, &walker); 311 ret = pkvm_alloc_private_va_range(PAGE_SIZE, &addr); 315 ret = kvm_pgtable_hyp_map(&pkvm_pgtable, addr, PAGE_SIZE, 333 start = ALIGN_DOWN(start, PAGE_SIZE); 336 end = ALIGN(end, PAGE_SIZE); 366 size = PAGE_SIZE * [all...] |
/linux-master/drivers/gpu/drm/tests/ |
H A D | drm_buddy_test.c | 332 mm_size = PAGE_SIZE << max_order; 333 KUNIT_ASSERT_FALSE_MSG(test, drm_buddy_init(&mm, mm_size, PAGE_SIZE), 347 size = get_size(order, PAGE_SIZE); 361 size = get_size(0, PAGE_SIZE); 371 size = get_size(top, PAGE_SIZE); 382 size = get_size(order, PAGE_SIZE); 411 mm_size = PAGE_SIZE << max_order; 412 KUNIT_ASSERT_FALSE_MSG(test, drm_buddy_init(&mm, mm_size, PAGE_SIZE), 418 size = get_size(order, PAGE_SIZE); 431 size = get_size(0, PAGE_SIZE); [all...] |
/linux-master/drivers/spi/ |
H A D | spi-test.h | 13 #define SPI_TEST_MAX_SIZE (32 * PAGE_SIZE) 121 1021, 1024, 1031, 4093, PAGE_SIZE, 4099, 65536, 65537
|
/linux-master/tools/testing/selftests/powerpc/benchmarks/ |
H A D | mmap_bench.c | 17 #define PAGE_SIZE (1UL << 16) macro 18 #define CHUNK_COUNT (MEMSIZE/PAGE_SIZE)
|
/linux-master/net/core/ |
H A D | hwbm.c | 15 if (likely(bm_pool->frag_size <= PAGE_SIZE)) 28 if (likely(frag_size <= PAGE_SIZE))
|
/linux-master/sound/firewire/ |
H A D | packets-buffer.c | 37 packets_per_page = PAGE_SIZE / packet_size; 54 b->packets[i].offset = page_index * PAGE_SIZE + offset_in_page;
|
/linux-master/arch/sh/kernel/vsyscall/ |
H A D | vsyscall.c | 67 addr = get_unmapped_area(NULL, 0, PAGE_SIZE, 0, 0); 73 ret = install_special_mapping(mm, addr, PAGE_SIZE,
|
/linux-master/arch/sh/mm/ |
H A D | pgtable.c | 24 PAGE_SIZE, SLAB_PANIC, pgd_ctor); 28 PAGE_SIZE, SLAB_PANIC, NULL);
|
/linux-master/arch/xtensa/kernel/ |
H A D | pci-dma.c | 27 unsigned long off = paddr & (PAGE_SIZE - 1); 35 size_t sz = min_t(size_t, size, PAGE_SIZE - off);
|
/linux-master/drivers/gpu/drm/i915/gem/selftests/ |
H A D | i915_gem_phys.c | 21 obj = i915_gem_object_create_shmem(i915, PAGE_SIZE); 36 err = i915_gem_object_attach_phys(obj, PAGE_SIZE);
|
/linux-master/arch/um/include/shared/ |
H A D | common-offsets.h | 7 DEFINE(UM_KERN_PAGE_SIZE, PAGE_SIZE);
|
/linux-master/arch/powerpc/mm/kasan/ |
H A D | book3s_32.c | 38 phys = memblock_phys_alloc_range(k_end - k_nobat, PAGE_SIZE, 0, 50 for (k_cur = k_nobat; k_cur < k_end; k_cur += PAGE_SIZE) {
|
/linux-master/arch/m68k/mm/ |
H A D | kmap.c | 35 #define IO_SIZE PAGE_SIZE 101 virtaddr += PAGE_SIZE; 102 size -= PAGE_SIZE; 282 virtaddr += PAGE_SIZE; 283 physaddr += PAGE_SIZE; 284 size -= PAGE_SIZE; 394 virtaddr += PAGE_SIZE; 395 size -= PAGE_SIZE;
|
/linux-master/arch/arm64/kernel/vdso32/ |
H A D | vdso.lds.S | 21 PROVIDE_HIDDEN(_vdso_data = . - __VVAR_PAGES * PAGE_SIZE); 23 PROVIDE_HIDDEN(_timens_data = _vdso_data + PAGE_SIZE);
|
/linux-master/fs/xfs/ |
H A D | xfs_bio_io.c | 9 return bio_max_segs(howmany(count, PAGE_SIZE)); 36 unsigned int len = min_t(unsigned, left, PAGE_SIZE - off);
|
/linux-master/arch/parisc/include/asm/ |
H A D | pgalloc.h | 27 memset(pgd, 0, PAGE_SIZE << PGD_TABLE_ORDER); 53 memset ((void *)pmd, 0, PAGE_SIZE << PMD_TABLE_ORDER);
|
/linux-master/arch/x86/kernel/ |
H A D | crash_dump_64.c | 25 vaddr = (__force void *)ioremap_encrypted(pfn << PAGE_SHIFT, PAGE_SIZE); 27 vaddr = (__force void *)ioremap_cache(pfn << PAGE_SHIFT, PAGE_SIZE);
|
/linux-master/arch/alpha/boot/ |
H A D | bootpz.c | 86 for (vaddr = vstart; vaddr <= vend; vaddr += PAGE_SIZE) 183 #define NEXT_PAGE(a) (((a) | (PAGE_SIZE - 1)) + 1) 255 NEXT_PAGE(K_COPY_IMAGE_START + KERNEL_SIZE + PAGE_SIZE) 387 uncompressed_image_start += PAGE_SIZE; 388 uncompressed_image_end += PAGE_SIZE; 389 initrd_image_start += PAGE_SIZE; 441 move_stack(initrd_image_start - PAGE_SIZE); 451 memset((char*)ZERO_PGE, 0, PAGE_SIZE);
|
/linux-master/arch/x86/kernel/cpu/sgx/ |
H A D | sgx.h | 73 return section->phys_addr + index * PAGE_SIZE; 83 return section->virt_addr + index * PAGE_SIZE;
|
/linux-master/tools/power/acpi/os_specific/service_layers/ |
H A D | osunixmap.c | 43 #ifdef PAGE_SIZE 44 return PAGE_SIZE;
|
/linux-master/arch/s390/include/asm/ |
H A D | thread_info.h | 24 #define BOOT_STACK_SIZE (PAGE_SIZE << 2) 25 #define THREAD_SIZE (PAGE_SIZE << THREAD_SIZE_ORDER)
|
/linux-master/fs/romfs/ |
H A D | mmap-nommu.c | 33 lpages = (len + PAGE_SIZE - 1) >> PAGE_SHIFT; 37 maxpages = (isize + PAGE_SIZE - 1) >> PAGE_SHIFT;
|