Searched refs:page_size (Results 51 - 75 of 411) sorted by relevance

1234567891011>>

/linux-master/tools/testing/selftests/powerpc/mm/
H A Dlarge_vm_gpr_corruption.c106 unsigned long page_size, addr, *p; local
118 page_size = sysconf(_SC_PAGESIZE);
123 p = mmap((void *)addr, page_size, PROT_READ | PROT_WRITE,
/linux-master/tools/testing/selftests/bpf/progs/
H A Dsockopt_inherit.c12 __s32 page_size = 0; variable
78 if (ctx->optlen > page_size)
107 if (ctx->optlen > page_size)
/linux-master/drivers/nvdimm/
H A Dpfn.h45 __le32 page_size; member in struct:nd_pfn_sb
/linux-master/drivers/gpu/drm/i915/gem/
H A Di915_gem_lmem.h31 resource_size_t page_size,
H A Di915_gem_region.h57 resource_size_t page_size,
/linux-master/tools/testing/selftests/kvm/lib/aarch64/
H A Ducall.c13 vm_vaddr_t mmio_gva = vm_vaddr_unused_gap(vm, vm->page_size, KVM_UTIL_MIN_VADDR);
/linux-master/drivers/accel/habanalabs/common/mmu/
H A Dmmu.c32 return hl_mem_area_inside_range(virt_addr, prop->dmmu.page_size,
163 * @page_size: page size
175 u32 page_size, u32 *real_page_size, bool is_dram_addr)
181 if ((page_size % mmu_prop->page_size) == 0) {
182 *real_page_size = mmu_prop->page_size;
187 page_size, mmu_prop->page_size >> 10);
192 static struct hl_mmu_properties *hl_mmu_get_prop(struct hl_device *hdev, u32 page_size, argument
199 else if ((page_size
174 hl_mmu_get_real_page_size(struct hl_device *hdev, struct hl_mmu_properties *mmu_prop, u32 page_size, u32 *real_page_size, bool is_dram_addr) argument
226 hl_mmu_unmap_page(struct hl_ctx *ctx, u64 virt_addr, u32 page_size, bool flush_pte) argument
292 hl_mmu_map_page(struct hl_ctx *ctx, u64 virt_addr, u64 phys_addr, u32 page_size, bool flush_pte) argument
387 u32 page_size; local
447 u32 page_size; local
[all...]
/linux-master/drivers/accel/habanalabs/common/
H A Dmemory.c29 static int set_alloc_page_size(struct hl_device *hdev, struct hl_mem_in *args, u32 *page_size) argument
38 if (prop->supports_user_set_page_size && args->alloc.page_size) {
39 psize = args->alloc.page_size;
49 *page_size = psize;
94 u32 num_curr_pgs, page_size; local
100 rc = set_alloc_page_size(hdev, args, &page_size);
104 num_pgs = DIV_ROUND_UP_ULL(args->alloc.mem_size, page_size);
105 total_size = num_pgs * page_size;
115 if (is_power_of_2(page_size))
117 total_size, NULL, page_size);
843 u32 npages, page_size = PAGE_SIZE, local
947 u32 page_size = phys_pg_pack->page_size; local
1011 u32 page_size; local
1074 u32 page_size = hdev->asic_prop.pmmu.page_size, local
1534 alloc_sgt_from_device_pages(struct hl_device *hdev, u64 *pages, u64 npages, u64 page_size, u64 exported_size, u64 offset, struct device *dev, enum dma_data_direction dir) argument
1727 u64 *pages, npages, page_size, exported_size, offset; local
2518 va_range_init(struct hl_device *hdev, struct hl_va_range **va_ranges, enum hl_va_range_type range_type, u64 start, u64 end, u32 page_size) argument
[all...]
/linux-master/tools/testing/selftests/bpf/prog_tests/
H A Duser_ringbuf.c101 int page_size = getpagesize(); local
111 mmap_ptr = mmap(NULL, page_size, PROT_READ, MAP_SHARED, rb_fd, 0);
113 ASSERT_ERR(mprotect(mmap_ptr, page_size, PROT_WRITE), "write_cons_pos_protect");
114 ASSERT_ERR(mprotect(mmap_ptr, page_size, PROT_EXEC), "exec_cons_pos_protect");
115 ASSERT_ERR_PTR(mremap(mmap_ptr, 0, 4 * page_size, MREMAP_MAYMOVE), "wr_prod_pos");
118 ASSERT_OK(munmap(mmap_ptr, page_size), "unmap_ro_cons");
121 mmap_ptr = mmap(NULL, page_size, PROT_READ | PROT_WRITE, MAP_SHARED,
122 rb_fd, page_size);
124 ASSERT_ERR(mprotect(mmap_ptr, page_size, PROT_EXEC), "exec_prod_pos_protect");
127 ASSERT_OK(munmap(mmap_ptr, page_size), "unmap_rw_pro
206 int rb_fd, page_size = getpagesize(); local
[all...]
/linux-master/drivers/misc/lkdtm/
H A Dpowerpc.c9 static void insert_slb_entry(unsigned long p, int ssize, int page_size) argument
13 flags = SLB_VSID_KERNEL | mmu_psize_defs[page_size].sllp;
/linux-master/arch/riscv/include/asm/
H A Dtlbflush.h39 unsigned long end, unsigned int page_size);
77 #define flush_tlb_mm_range(mm, start, end, page_size) flush_tlb_all()
/linux-master/arch/powerpc/platforms/ps3/
H A Dmm.c353 DBG("%s:%d: page_size %u\n", func, line, r->page_size);
389 DBG("%s:%d: r.page_size %u\n", func, line, c->region->page_size);
401 unsigned long aligned_bus = ALIGN_DOWN(bus_addr, 1 << r->page_size);
403 1 << r->page_size);
430 unsigned long aligned_lpar = ALIGN_DOWN(lpar_addr, 1 << r->page_size);
432 1 << r->page_size);
479 for (iopage = 0; iopage < (c->len >> r->page_size); iopage++) {
480 offset = (1 << r->page_size) * iopag
1118 ps3_dma_region_init(struct ps3_system_bus_device *dev, struct ps3_dma_region *r, enum ps3_dma_page_size page_size, enum ps3_dma_region_type region_type, void *addr, unsigned long len) argument
[all...]
/linux-master/include/asm-generic/
H A Dtlb.h144 * changes the size and provides mmu_gather::page_size to tlb_flush().
270 bool delay_rmap, int page_size);
355 unsigned int page_size; member in struct:mmu_gather
470 struct page *page, int page_size)
472 if (__tlb_remove_page_size(tlb, page, false, page_size))
503 unsigned int page_size)
506 if (tlb->page_size && tlb->page_size != page_size) {
511 tlb->page_size
469 tlb_remove_page_size(struct mmu_gather *tlb, struct page *page, int page_size) argument
502 tlb_change_page_size(struct mmu_gather *tlb, unsigned int page_size) argument
[all...]
/linux-master/tools/perf/arch/x86/util/
H A Dintel-bts.c26 #include <internal/lib.h> // page_size
155 opts->auxtrace_mmap_pages = MiB(4) / page_size;
157 opts->auxtrace_mmap_pages = KiB(128) / page_size;
159 opts->mmap_pages = KiB(256) / page_size;
163 opts->mmap_pages = KiB(256) / page_size;
167 opts->auxtrace_mmap_pages * (size_t)page_size;
171 sz = round_up(sz, page_size) / page_size;
175 opts->auxtrace_mmap_pages * (size_t)page_size) {
178 opts->auxtrace_mmap_pages * (size_t)page_size);
[all...]
/linux-master/arch/powerpc/mm/book3s64/
H A Dradix_tlb.c432 unsigned long pid, unsigned long page_size,
438 for (addr = start; addr < end; addr += page_size)
453 unsigned long pid, unsigned long page_size,
459 __tlbiel_va_range(start, end, pid, page_size, psize);
464 unsigned long pid, unsigned long page_size,
470 for (addr = start; addr < end; addr += page_size)
473 fixup_tlbie_va_range(addr - page_size, pid, ap);
521 unsigned long page_size; member in struct:tlbiel_va_range
530 _tlbiel_va_range(t->start, t->end, t->pid, t->page_size,
546 unsigned long pid, unsigned long page_size,
431 __tlbiel_va_range(unsigned long start, unsigned long end, unsigned long pid, unsigned long page_size, unsigned long psize) argument
452 _tlbiel_va_range(unsigned long start, unsigned long end, unsigned long pid, unsigned long page_size, unsigned long psize, bool also_pwc) argument
463 __tlbie_va_range(unsigned long start, unsigned long end, unsigned long pid, unsigned long page_size, unsigned long psize) argument
545 _tlbie_va_range(unsigned long start, unsigned long end, unsigned long pid, unsigned long page_size, unsigned long psize, bool also_pwc) argument
556 _tlbiel_va_range_multicast(struct mm_struct *mm, unsigned long start, unsigned long end, unsigned long pid, unsigned long page_size, unsigned long psize, bool also_pwc) argument
1023 unsigned long page_size = 1UL << page_shift; local
1130 radix_get_mmu_psize(int page_size) argument
1148 radix__flush_tlb_lpid_page(unsigned int lpid, unsigned long addr, unsigned long page_size) argument
1188 int page_size = tlb->page_size; local
1241 unsigned long page_size = 1UL << page_shift; local
1497 __tlbie_va_range_lpid(unsigned long start, unsigned long end, unsigned long pid, unsigned long lpid, unsigned long page_size, unsigned long psize) argument
1511 _tlbie_va_range_lpid(unsigned long start, unsigned long end, unsigned long pid, unsigned long lpid, unsigned long page_size, unsigned long psize, bool also_pwc) argument
[all...]
/linux-master/tools/testing/selftests/mm/
H A Duffd-common.c13 unsigned long nr_cpus, nr_pages, nr_pages_per_cpu, page_size; variable
45 if (madvise(rel_area, nr_pages * page_size, MADV_DONTNEED))
51 *alloc_area = mmap(NULL, nr_pages * page_size, PROT_READ | PROT_WRITE,
67 if (madvise(rel_area, nr_pages * page_size, MADV_DONTNEED))
70 if (madvise(rel_area, nr_pages * page_size, MADV_REMOVE))
77 off_t size = nr_pages * page_size;
121 if (madvise(rel_area, nr_pages * page_size, MADV_REMOVE))
128 size_t bytes = nr_pages * page_size, hpage_size = read_pmd_pagesize();
261 if (munmap(*area, nr_pages * page_size))
445 wp_range(uffd, msg->arg.pagefault.address, page_size, fals
[all...]
H A Duffd-common.h60 ((pthread_mutex_t *) ((___area) + (___nr)*page_size))
67 ((___area) + (___nr)*page_size + \
100 extern unsigned long nr_cpus, nr_pages, nr_pages_per_cpu, page_size;
H A Dhmm-tests.c71 unsigned int page_size; local
94 unsigned int page_size; local
136 self->page_size = sysconf(_SC_PAGE_SIZE);
137 self->page_shift = ffs(self->page_size) - 1;
147 self->page_size = sysconf(_SC_PAGE_SIZE);
148 self->page_shift = ffs(self->page_size) - 1;
302 npages = ALIGN(HMM_BUFFER_SIZE, self->page_size) >> self->page_shift;
324 i = 2 * self->page_size / sizeof(*ptr);
333 val = *(int *)(buffer->ptr + self->page_size);
344 for (i = 0; i < 2 * self->page_size / sizeo
[all...]
/linux-master/drivers/misc/cxl/
H A Dcxllib.c204 unsigned long *page_size)
216 *page_size = vma_kernel_pagesize(vma);
228 unsigned long page_size; local
239 rc = get_vma_info(mm, addr, &vma_start, &vma_end, &page_size);
243 for (dar = (addr & ~(page_size - 1)); dar < (addr + size);
244 dar += page_size) {
260 &page_size);
202 get_vma_info(struct mm_struct *mm, u64 addr, u64 *vma_start, u64 *vma_end, unsigned long *page_size) argument
/linux-master/arch/powerpc/mm/nohash/
H A Dbook3e_pgtable.c24 unsigned long page_size,
41 for (i = 0; i < page_size; i += PAGE_SIZE)
49 unsigned long page_size)
23 vmemmap_create_mapping(unsigned long start, unsigned long page_size, unsigned long phys) argument
48 vmemmap_remove_mapping(unsigned long start, unsigned long page_size) argument
/linux-master/drivers/iommu/amd/
H A Dio_pgtable.c50 static u64 *first_pte_l7(u64 *pte, unsigned long *page_size, argument
61 if (page_size)
62 *page_size = pg_size;
197 unsigned long page_size,
205 BUG_ON(!is_power_of_2(page_size));
219 address = PAGE_SIZE_ALIGN(address, page_size);
220 end_lvl = PAGE_SIZE_LEVEL(page_size);
292 unsigned long *page_size)
297 *page_size = 0;
304 *page_size
195 alloc_pte(struct protection_domain *domain, unsigned long address, unsigned long page_size, u64 **pte_page, gfp_t gfp, bool *updated) argument
290 fetch_pte(struct amd_io_pgtable *pgtable, unsigned long address, unsigned long *page_size) argument
[all...]
/linux-master/tools/testing/selftests/kvm/
H A Dmemslot_perf_test.c190 uint32_t guest_page_size = data->vm->page_size;
221 uint32_t guest_page_size = data->vm->page_size;
260 uint32_t guest_page_size = data->vm->page_size;
291 guest_page_size = vm_guest_mode_params[VM_MODE_DEFAULT].page_size;
295 TEST_ASSERT(data->vm->page_size == guest_page_size, "Invalid VM page size");
350 sync->guest_page_size = data->vm->page_size;
447 uint32_t page_size = (typeof(page_size))READ_ONCE(sync->guest_page_size); local
458 ptr += page_size)
476 uint32_t page_size local
543 uint32_t page_size = (typeof(page_size))READ_ONCE(sync->guest_page_size); local
[all...]
H A Dmemslot_modification_stress_test.c68 uint64_t pages = max_t(int, vm->page_size, getpagesize()) / vm->page_size;
76 gpa = memstress_args.gpa - pages * vm->page_size;
/linux-master/arch/riscv/kvm/
H A Dmmu.c60 static int gstage_page_size_to_level(unsigned long page_size, u32 *out_level) argument
66 if (page_size == (psz << (i * gstage_index_bits))) {
180 unsigned long page_size,
188 ret = gstage_page_size_to_level(page_size, &level);
234 unsigned long next_page_size, page_size; local
236 ret = gstage_level_to_page_size(ptep_level, &page_size);
240 BUG_ON(addr & (page_size - 1));
276 unsigned long page_size; local
282 ret = gstage_level_to_page_size(ptep_level, &page_size);
289 if (!(addr & (page_size
177 gstage_map_page(struct kvm *kvm, struct kvm_mmu_memory_cache *pcache, gpa_t gpa, phys_addr_t hpa, unsigned long page_size, bool page_rdonly, bool page_exec) argument
312 unsigned long page_size; local
[all...]
/linux-master/tools/testing/selftests/kvm/lib/x86_64/
H A Dvmx.c30 uint64_t page_size:1; member in struct:eptPageTableEntry
378 pte->page_size = (current_level == target_level);
379 if (pte->page_size)
392 TEST_ASSERT(!pte->page_size,
402 const uint64_t page_size = PG_LEVEL_SIZE(target_level); local
412 TEST_ASSERT((nested_paddr % page_size) == 0,
414 " nested_paddr: 0x%lx page_size: 0x%lx",
415 nested_paddr, page_size);
418 " nested_paddr: 0x%lx vm->max_gfn: 0x%lx vm->page_size: 0x%x",
419 paddr, vm->max_gfn, vm->page_size);
477 size_t page_size = PG_LEVEL_SIZE(level); local
[all...]

Completed in 262 milliseconds

1234567891011>>