/linux-master/tools/testing/selftests/powerpc/mm/ |
H A D | large_vm_gpr_corruption.c | 106 unsigned long page_size, addr, *p; local 118 page_size = sysconf(_SC_PAGESIZE); 123 p = mmap((void *)addr, page_size, PROT_READ | PROT_WRITE,
|
/linux-master/tools/testing/selftests/bpf/progs/ |
H A D | sockopt_inherit.c | 12 __s32 page_size = 0; variable 78 if (ctx->optlen > page_size) 107 if (ctx->optlen > page_size)
|
/linux-master/drivers/nvdimm/ |
H A D | pfn.h | 45 __le32 page_size; member in struct:nd_pfn_sb
|
/linux-master/drivers/gpu/drm/i915/gem/ |
H A D | i915_gem_lmem.h | 31 resource_size_t page_size,
|
H A D | i915_gem_region.h | 57 resource_size_t page_size,
|
/linux-master/tools/testing/selftests/kvm/lib/aarch64/ |
H A D | ucall.c | 13 vm_vaddr_t mmio_gva = vm_vaddr_unused_gap(vm, vm->page_size, KVM_UTIL_MIN_VADDR);
|
/linux-master/drivers/accel/habanalabs/common/mmu/ |
H A D | mmu.c | 32 return hl_mem_area_inside_range(virt_addr, prop->dmmu.page_size, 163 * @page_size: page size 175 u32 page_size, u32 *real_page_size, bool is_dram_addr) 181 if ((page_size % mmu_prop->page_size) == 0) { 182 *real_page_size = mmu_prop->page_size; 187 page_size, mmu_prop->page_size >> 10); 192 static struct hl_mmu_properties *hl_mmu_get_prop(struct hl_device *hdev, u32 page_size, argument 199 else if ((page_size 174 hl_mmu_get_real_page_size(struct hl_device *hdev, struct hl_mmu_properties *mmu_prop, u32 page_size, u32 *real_page_size, bool is_dram_addr) argument 226 hl_mmu_unmap_page(struct hl_ctx *ctx, u64 virt_addr, u32 page_size, bool flush_pte) argument 292 hl_mmu_map_page(struct hl_ctx *ctx, u64 virt_addr, u64 phys_addr, u32 page_size, bool flush_pte) argument 387 u32 page_size; local 447 u32 page_size; local [all...] |
/linux-master/drivers/accel/habanalabs/common/ |
H A D | memory.c | 29 static int set_alloc_page_size(struct hl_device *hdev, struct hl_mem_in *args, u32 *page_size) argument 38 if (prop->supports_user_set_page_size && args->alloc.page_size) { 39 psize = args->alloc.page_size; 49 *page_size = psize; 94 u32 num_curr_pgs, page_size; local 100 rc = set_alloc_page_size(hdev, args, &page_size); 104 num_pgs = DIV_ROUND_UP_ULL(args->alloc.mem_size, page_size); 105 total_size = num_pgs * page_size; 115 if (is_power_of_2(page_size)) 117 total_size, NULL, page_size); 843 u32 npages, page_size = PAGE_SIZE, local 947 u32 page_size = phys_pg_pack->page_size; local 1011 u32 page_size; local 1074 u32 page_size = hdev->asic_prop.pmmu.page_size, local 1534 alloc_sgt_from_device_pages(struct hl_device *hdev, u64 *pages, u64 npages, u64 page_size, u64 exported_size, u64 offset, struct device *dev, enum dma_data_direction dir) argument 1727 u64 *pages, npages, page_size, exported_size, offset; local 2518 va_range_init(struct hl_device *hdev, struct hl_va_range **va_ranges, enum hl_va_range_type range_type, u64 start, u64 end, u32 page_size) argument [all...] |
/linux-master/tools/testing/selftests/bpf/prog_tests/ |
H A D | user_ringbuf.c | 101 int page_size = getpagesize(); local 111 mmap_ptr = mmap(NULL, page_size, PROT_READ, MAP_SHARED, rb_fd, 0); 113 ASSERT_ERR(mprotect(mmap_ptr, page_size, PROT_WRITE), "write_cons_pos_protect"); 114 ASSERT_ERR(mprotect(mmap_ptr, page_size, PROT_EXEC), "exec_cons_pos_protect"); 115 ASSERT_ERR_PTR(mremap(mmap_ptr, 0, 4 * page_size, MREMAP_MAYMOVE), "wr_prod_pos"); 118 ASSERT_OK(munmap(mmap_ptr, page_size), "unmap_ro_cons"); 121 mmap_ptr = mmap(NULL, page_size, PROT_READ | PROT_WRITE, MAP_SHARED, 122 rb_fd, page_size); 124 ASSERT_ERR(mprotect(mmap_ptr, page_size, PROT_EXEC), "exec_prod_pos_protect"); 127 ASSERT_OK(munmap(mmap_ptr, page_size), "unmap_rw_pro 206 int rb_fd, page_size = getpagesize(); local [all...] |
/linux-master/drivers/misc/lkdtm/ |
H A D | powerpc.c | 9 static void insert_slb_entry(unsigned long p, int ssize, int page_size) argument 13 flags = SLB_VSID_KERNEL | mmu_psize_defs[page_size].sllp;
|
/linux-master/arch/riscv/include/asm/ |
H A D | tlbflush.h | 39 unsigned long end, unsigned int page_size); 77 #define flush_tlb_mm_range(mm, start, end, page_size) flush_tlb_all()
|
/linux-master/arch/powerpc/platforms/ps3/ |
H A D | mm.c | 353 DBG("%s:%d: page_size %u\n", func, line, r->page_size); 389 DBG("%s:%d: r.page_size %u\n", func, line, c->region->page_size); 401 unsigned long aligned_bus = ALIGN_DOWN(bus_addr, 1 << r->page_size); 403 1 << r->page_size); 430 unsigned long aligned_lpar = ALIGN_DOWN(lpar_addr, 1 << r->page_size); 432 1 << r->page_size); 479 for (iopage = 0; iopage < (c->len >> r->page_size); iopage++) { 480 offset = (1 << r->page_size) * iopag 1118 ps3_dma_region_init(struct ps3_system_bus_device *dev, struct ps3_dma_region *r, enum ps3_dma_page_size page_size, enum ps3_dma_region_type region_type, void *addr, unsigned long len) argument [all...] |
/linux-master/include/asm-generic/ |
H A D | tlb.h | 144 * changes the size and provides mmu_gather::page_size to tlb_flush(). 270 bool delay_rmap, int page_size); 355 unsigned int page_size; member in struct:mmu_gather 470 struct page *page, int page_size) 472 if (__tlb_remove_page_size(tlb, page, false, page_size)) 503 unsigned int page_size) 506 if (tlb->page_size && tlb->page_size != page_size) { 511 tlb->page_size 469 tlb_remove_page_size(struct mmu_gather *tlb, struct page *page, int page_size) argument 502 tlb_change_page_size(struct mmu_gather *tlb, unsigned int page_size) argument [all...] |
/linux-master/tools/perf/arch/x86/util/ |
H A D | intel-bts.c | 26 #include <internal/lib.h> // page_size 155 opts->auxtrace_mmap_pages = MiB(4) / page_size; 157 opts->auxtrace_mmap_pages = KiB(128) / page_size; 159 opts->mmap_pages = KiB(256) / page_size; 163 opts->mmap_pages = KiB(256) / page_size; 167 opts->auxtrace_mmap_pages * (size_t)page_size; 171 sz = round_up(sz, page_size) / page_size; 175 opts->auxtrace_mmap_pages * (size_t)page_size) { 178 opts->auxtrace_mmap_pages * (size_t)page_size); [all...] |
/linux-master/arch/powerpc/mm/book3s64/ |
H A D | radix_tlb.c | 432 unsigned long pid, unsigned long page_size, 438 for (addr = start; addr < end; addr += page_size) 453 unsigned long pid, unsigned long page_size, 459 __tlbiel_va_range(start, end, pid, page_size, psize); 464 unsigned long pid, unsigned long page_size, 470 for (addr = start; addr < end; addr += page_size) 473 fixup_tlbie_va_range(addr - page_size, pid, ap); 521 unsigned long page_size; member in struct:tlbiel_va_range 530 _tlbiel_va_range(t->start, t->end, t->pid, t->page_size, 546 unsigned long pid, unsigned long page_size, 431 __tlbiel_va_range(unsigned long start, unsigned long end, unsigned long pid, unsigned long page_size, unsigned long psize) argument 452 _tlbiel_va_range(unsigned long start, unsigned long end, unsigned long pid, unsigned long page_size, unsigned long psize, bool also_pwc) argument 463 __tlbie_va_range(unsigned long start, unsigned long end, unsigned long pid, unsigned long page_size, unsigned long psize) argument 545 _tlbie_va_range(unsigned long start, unsigned long end, unsigned long pid, unsigned long page_size, unsigned long psize, bool also_pwc) argument 556 _tlbiel_va_range_multicast(struct mm_struct *mm, unsigned long start, unsigned long end, unsigned long pid, unsigned long page_size, unsigned long psize, bool also_pwc) argument 1023 unsigned long page_size = 1UL << page_shift; local 1130 radix_get_mmu_psize(int page_size) argument 1148 radix__flush_tlb_lpid_page(unsigned int lpid, unsigned long addr, unsigned long page_size) argument 1188 int page_size = tlb->page_size; local 1241 unsigned long page_size = 1UL << page_shift; local 1497 __tlbie_va_range_lpid(unsigned long start, unsigned long end, unsigned long pid, unsigned long lpid, unsigned long page_size, unsigned long psize) argument 1511 _tlbie_va_range_lpid(unsigned long start, unsigned long end, unsigned long pid, unsigned long lpid, unsigned long page_size, unsigned long psize, bool also_pwc) argument [all...] |
/linux-master/tools/testing/selftests/mm/ |
H A D | uffd-common.c | 13 unsigned long nr_cpus, nr_pages, nr_pages_per_cpu, page_size; variable 45 if (madvise(rel_area, nr_pages * page_size, MADV_DONTNEED)) 51 *alloc_area = mmap(NULL, nr_pages * page_size, PROT_READ | PROT_WRITE, 67 if (madvise(rel_area, nr_pages * page_size, MADV_DONTNEED)) 70 if (madvise(rel_area, nr_pages * page_size, MADV_REMOVE)) 77 off_t size = nr_pages * page_size; 121 if (madvise(rel_area, nr_pages * page_size, MADV_REMOVE)) 128 size_t bytes = nr_pages * page_size, hpage_size = read_pmd_pagesize(); 261 if (munmap(*area, nr_pages * page_size)) 445 wp_range(uffd, msg->arg.pagefault.address, page_size, fals [all...] |
H A D | uffd-common.h | 60 ((pthread_mutex_t *) ((___area) + (___nr)*page_size)) 67 ((___area) + (___nr)*page_size + \ 100 extern unsigned long nr_cpus, nr_pages, nr_pages_per_cpu, page_size;
|
H A D | hmm-tests.c | 71 unsigned int page_size; local 94 unsigned int page_size; local 136 self->page_size = sysconf(_SC_PAGE_SIZE); 137 self->page_shift = ffs(self->page_size) - 1; 147 self->page_size = sysconf(_SC_PAGE_SIZE); 148 self->page_shift = ffs(self->page_size) - 1; 302 npages = ALIGN(HMM_BUFFER_SIZE, self->page_size) >> self->page_shift; 324 i = 2 * self->page_size / sizeof(*ptr); 333 val = *(int *)(buffer->ptr + self->page_size); 344 for (i = 0; i < 2 * self->page_size / sizeo [all...] |
/linux-master/drivers/misc/cxl/ |
H A D | cxllib.c | 204 unsigned long *page_size) 216 *page_size = vma_kernel_pagesize(vma); 228 unsigned long page_size; local 239 rc = get_vma_info(mm, addr, &vma_start, &vma_end, &page_size); 243 for (dar = (addr & ~(page_size - 1)); dar < (addr + size); 244 dar += page_size) { 260 &page_size); 202 get_vma_info(struct mm_struct *mm, u64 addr, u64 *vma_start, u64 *vma_end, unsigned long *page_size) argument
|
/linux-master/arch/powerpc/mm/nohash/ |
H A D | book3e_pgtable.c | 24 unsigned long page_size, 41 for (i = 0; i < page_size; i += PAGE_SIZE) 49 unsigned long page_size) 23 vmemmap_create_mapping(unsigned long start, unsigned long page_size, unsigned long phys) argument 48 vmemmap_remove_mapping(unsigned long start, unsigned long page_size) argument
|
/linux-master/drivers/iommu/amd/ |
H A D | io_pgtable.c | 50 static u64 *first_pte_l7(u64 *pte, unsigned long *page_size, argument 61 if (page_size) 62 *page_size = pg_size; 197 unsigned long page_size, 205 BUG_ON(!is_power_of_2(page_size)); 219 address = PAGE_SIZE_ALIGN(address, page_size); 220 end_lvl = PAGE_SIZE_LEVEL(page_size); 292 unsigned long *page_size) 297 *page_size = 0; 304 *page_size 195 alloc_pte(struct protection_domain *domain, unsigned long address, unsigned long page_size, u64 **pte_page, gfp_t gfp, bool *updated) argument 290 fetch_pte(struct amd_io_pgtable *pgtable, unsigned long address, unsigned long *page_size) argument [all...] |
/linux-master/tools/testing/selftests/kvm/ |
H A D | memslot_perf_test.c | 190 uint32_t guest_page_size = data->vm->page_size; 221 uint32_t guest_page_size = data->vm->page_size; 260 uint32_t guest_page_size = data->vm->page_size; 291 guest_page_size = vm_guest_mode_params[VM_MODE_DEFAULT].page_size; 295 TEST_ASSERT(data->vm->page_size == guest_page_size, "Invalid VM page size"); 350 sync->guest_page_size = data->vm->page_size; 447 uint32_t page_size = (typeof(page_size))READ_ONCE(sync->guest_page_size); local 458 ptr += page_size) 476 uint32_t page_size local 543 uint32_t page_size = (typeof(page_size))READ_ONCE(sync->guest_page_size); local [all...] |
H A D | memslot_modification_stress_test.c | 68 uint64_t pages = max_t(int, vm->page_size, getpagesize()) / vm->page_size; 76 gpa = memstress_args.gpa - pages * vm->page_size;
|
/linux-master/arch/riscv/kvm/ |
H A D | mmu.c | 60 static int gstage_page_size_to_level(unsigned long page_size, u32 *out_level) argument 66 if (page_size == (psz << (i * gstage_index_bits))) { 180 unsigned long page_size, 188 ret = gstage_page_size_to_level(page_size, &level); 234 unsigned long next_page_size, page_size; local 236 ret = gstage_level_to_page_size(ptep_level, &page_size); 240 BUG_ON(addr & (page_size - 1)); 276 unsigned long page_size; local 282 ret = gstage_level_to_page_size(ptep_level, &page_size); 289 if (!(addr & (page_size 177 gstage_map_page(struct kvm *kvm, struct kvm_mmu_memory_cache *pcache, gpa_t gpa, phys_addr_t hpa, unsigned long page_size, bool page_rdonly, bool page_exec) argument 312 unsigned long page_size; local [all...] |
/linux-master/tools/testing/selftests/kvm/lib/x86_64/ |
H A D | vmx.c | 30 uint64_t page_size:1; member in struct:eptPageTableEntry 378 pte->page_size = (current_level == target_level); 379 if (pte->page_size) 392 TEST_ASSERT(!pte->page_size, 402 const uint64_t page_size = PG_LEVEL_SIZE(target_level); local 412 TEST_ASSERT((nested_paddr % page_size) == 0, 414 " nested_paddr: 0x%lx page_size: 0x%lx", 415 nested_paddr, page_size); 418 " nested_paddr: 0x%lx vm->max_gfn: 0x%lx vm->page_size: 0x%x", 419 paddr, vm->max_gfn, vm->page_size); 477 size_t page_size = PG_LEVEL_SIZE(level); local [all...] |