/linux-master/tools/testing/selftests/powerpc/copyloops/ |
H A D | exc_validate.c | 81 int page_size; local 85 page_size = getpagesize(); 86 p = mmap(NULL, page_size * 2, PROT_READ|PROT_WRITE, 94 memset(p, 0, page_size); 98 if (mprotect(p + page_size, page_size, PROT_NONE)) { 103 q = p + page_size - MAX_LEN;
|
/linux-master/tools/testing/selftests/powerpc/mm/ |
H A D | bad_accesses.c | 69 unsigned long i, j, addr, region_shift, page_shift, page_size; local 82 page_size = sysconf(_SC_PAGESIZE); 83 if (page_size == (64 * 1024)) 88 if (page_size == (64 * 1024) || !hash_mmu) { 93 } else if (page_size == (4 * 1024) && hash_mmu) {
|
H A D | stack_expansion_ldst.c | 157 static void test_one_type(enum access_type type, unsigned long page_size, unsigned long rlim_cur) argument 162 for (delta = page_size; delta <= rlim_cur; delta += page_size) 173 unsigned long page_size; local 176 page_size = getpagesize(); 181 test_one_type(LOAD, page_size, rlimit.rlim_cur); 183 test_one_type(STORE, page_size, rlimit.rlim_cur);
|
/linux-master/tools/testing/selftests/mm/ |
H A D | mremap_test.c | 57 #define PTE page_size 213 static void mremap_expand_merge(FILE *maps_fp, unsigned long page_size) argument 219 start = mmap(NULL, 3 * page_size, PROT_READ | PROT_WRITE, 227 munmap(start + page_size, page_size); 228 remap = mremap(start, page_size, 2 * page_size, 0); 231 munmap(start, page_size); 232 munmap(start + 2 * page_size, page_size); 251 mremap_expand_merge_offset(FILE *maps_fp, unsigned long page_size) argument 650 int page_size; local [all...] |
H A D | memfd_secret.c | 37 static unsigned long page_size; variable 65 if (len % page_size != 0) 66 len = (len/page_size) * page_size; 184 mem = mmap(NULL, page_size, prot, mode, fd, 0); 190 ftruncate(fd, page_size); 191 memset(mem, PATTERN, page_size); 236 page_size = sysconf(_SC_PAGE_SIZE); 237 if (!page_size) 248 printf("page_size [all...] |
H A D | pagemap_ioctl.c | 37 int page_size; variable 41 #define LEN(region) ((region.end - region.start)/page_size) 190 mem_size = num_pages * page_size; 210 vec_size = mem_size/page_size; 245 mem_size = num_pages * page_size; 318 for (i = 0; i < mem_size; i += 2 * page_size) 326 ksft_test_result(ret == mem_size/(page_size * 2), 352 for (i = 0; i < mem_size; i += 2 * page_size) 354 mem[(mem_size/page_size - 1) * page_size] 1310 get_dirty_pages_reset(char *mem, unsigned int count, int reset, int page_size) argument 1384 transact_test(int page_size) argument [all...] |
H A D | uffd-stress.c | 158 copy_page_retry(uffd, page_nr * page_size); 166 wp_range(uffd, (unsigned long)area_dst + start_nr * page_size, local 167 nr_pages_per_cpu * page_size, true); local 174 copy_page_retry(uffd, page_nr * page_size); 249 uint64_t mem_size = nr_pages * page_size; 256 if (posix_memalign(&area, page_size, page_size)) 259 bzero(zeropage, page_size); 334 nr_pages * page_size, false); local 396 page_size [all...] |
H A D | ksm_tests.c | 348 long page_count, int timeout, size_t page_size) 359 map_ptr = allocate_memory(NULL, prot, mapping, '*', page_size * page_count); 363 if (ksm_merge_pages(merge_type, map_ptr, page_size * page_count, start_time, timeout)) 369 munmap(map_ptr, page_size * page_count); 377 munmap(map_ptr, page_size * page_count); 381 static int check_ksm_unmerge(int merge_type, int mapping, int prot, int timeout, size_t page_size) argument 393 map_ptr = allocate_memory(NULL, prot, mapping, '*', page_size * page_count); 397 if (ksm_merge_pages(merge_type, map_ptr, page_size * page_count, start_time, timeout)) 402 memset(map_ptr + page_size, '+', 1); 411 munmap(map_ptr, page_size * page_coun 347 check_ksm_merge(int merge_type, int mapping, int prot, long page_count, int timeout, size_t page_size) argument 421 check_ksm_zero_page_merge(int merge_type, int mapping, int prot, long page_count, int timeout, bool use_zero_pages, size_t page_size) argument 486 check_ksm_numa_merge(int merge_type, int mapping, int prot, int timeout, bool merge_across_nodes, size_t page_size) argument 706 ksm_cow_time(int merge_type, int mapping, int prot, int timeout, size_t page_size) argument 784 size_t page_size = sysconf(_SC_PAGESIZE); local [all...] |
H A D | khugepaged.c | 29 static unsigned long page_size; variable 276 for (i = start / page_size; i < end / page_size; i++) 277 p[i * page_size / sizeof(*p)] = i + 0xdead0000; 338 for (i = start / page_size; i < end / page_size; i++) { 339 if (p[i * page_size / sizeof(*p)] != i + 0xdead0000) { 341 i, p[i * page_size / sizeof(*p)]); 634 madvise(p, page_size, MADV_DONTNEED); 671 ops->fault(p, 0, page_size); [all...] |
/linux-master/tools/testing/selftests/kvm/lib/s390x/ |
H A D | processor.c | 17 TEST_ASSERT(vm->page_size == 4096, "Unsupported page size: 0x%x", 18 vm->page_size); 26 memset(addr_gpa2hva(vm, paddr), 0xff, PAGES_PER_REGION * vm->page_size); 43 memset(addr_gpa2hva(vm, taddr), 0xff, PAGES_PER_REGION * vm->page_size); 55 TEST_ASSERT((gva % vm->page_size) == 0, 57 " vaddr: 0x%lx vm->page_size: 0x%x", 58 gva, vm->page_size); 63 TEST_ASSERT((gpa % vm->page_size) == 0, 65 " paddr: 0x%lx vm->page_size: 0x%x", 66 gva, vm->page_size); [all...] |
/linux-master/tools/testing/selftests/bpf/prog_tests/ |
H A D | ringbuf_multi.c | 48 int page_size = getpagesize(); local 56 ASSERT_EQ(bpf_map__max_entries(skel->maps.ringbuf1), page_size, "rb1_size_before"); 57 ASSERT_OK(bpf_map__set_max_entries(skel->maps.ringbuf1, page_size + 1), "rb1_resize"); 58 ASSERT_EQ(bpf_map__max_entries(skel->maps.ringbuf1), 2 * page_size, "rb1_size_after"); 59 ASSERT_OK(bpf_map__set_max_entries(skel->maps.ringbuf1, page_size), "rb1_reset"); 60 ASSERT_EQ(bpf_map__max_entries(skel->maps.ringbuf1), page_size, "rb1_size_final"); 62 proto_fd = bpf_map_create(BPF_MAP_TYPE_RINGBUF, NULL, 0, 0, page_size, NULL); 78 if (!ASSERT_ERR(bpf_map__set_max_entries(skel->maps.ringbuf1, 3 * page_size), "rb1_resize_after_load"))
|
/linux-master/include/linux/ |
H A D | iova_bitmap.h | 20 unsigned long page_size, 30 unsigned long page_size, 28 iova_bitmap_alloc(unsigned long iova, size_t length, unsigned long page_size, u64 __user *data) argument
|
/linux-master/tools/testing/selftests/bpf/progs/ |
H A D | sockopt_qos_to_cc.c | 12 __s32 page_size = 0; variable 44 if (ctx->optlen > page_size)
|
H A D | sockopt_sk.c | 10 int page_size = 0; /* userspace should set it */ variable 104 if (optval_end - optval != page_size) 134 if (ctx->optlen > page_size) 195 if (ctx->optlen != page_size * 2) 209 if (optval_end - optval != page_size) 235 if (ctx->optlen > page_size)
|
/linux-master/drivers/vfio/pci/pds/ |
H A D | dirty.h | 13 u64 page_size; member in struct:pds_vfio_region 39 u64 *page_size);
|
/linux-master/drivers/infiniband/hw/mlx5/ |
H A D | mem.c | 41 void mlx5_ib_populate_pas(struct ib_umem *umem, size_t page_size, __be64 *pas, argument 46 rdma_umem_for_each_dma_block (umem, &biter, page_size) { 64 unsigned long page_size; local 67 page_size = ib_umem_find_best_pgoff(umem, pgsz_bitmap, pgoff_bitmask); 68 if (!page_size) 74 * Reduce the page_size, and thus the page_offset and quanta, until the 75 * page_offset fits into the mailbox field. Once page_size < scale this 78 page_offset = ib_umem_dma_offset(umem, page_size); 79 while (page_offset & ~(u64)(page_offset_mask * (page_size / scale))) { 80 page_size / [all...] |
/linux-master/drivers/net/ethernet/qlogic/qed/ |
H A D | qed_chain.c | 22 params->page_size); 24 params->page_size, 33 chain->page_size = params->page_size; 88 dma_free_coherent(dev, chain->page_size, virt, phys); 101 dma_free_coherent(&cdev->pdev->dev, chain->page_size, 119 dma_free_coherent(dev, chain->page_size, entry->virt_addr, 164 chain_size = ELEMS_PER_PAGE(params->elem_size, params->page_size); 207 virt = dma_alloc_coherent(dev, chain->page_size, &phys, 238 virt = dma_alloc_coherent(&cdev->pdev->dev, chain->page_size, [all...] |
/linux-master/drivers/staging/rtl8723bs/hal/ |
H A D | hal_sdio.c | 62 u32 page_size; local 65 rtw_hal_get_def_var(padapter, HAL_DEF_TX_PAGE_SIZE, &page_size); 67 lenHQ = ((numHQ + numPubQ) >> 1) * page_size; 68 lenNQ = ((numNQ + numPubQ) >> 1) * page_size; 69 lenLQ = ((numLQ + numPubQ) >> 1) * page_size;
|
/linux-master/arch/powerpc/mm/ |
H A D | init_64.c | 188 unsigned long page_size) 190 unsigned long nr_pfn = page_size / sizeof(struct page); 206 unsigned long page_size = 1 << mmu_psize_defs[mmu_vmemmap_psize].shift; local 209 start = ALIGN_DOWN(start, page_size); 213 for (; start < end; start += page_size) { 223 if (vmemmap_populated(start, page_size)) 231 if (altmap && !altmap_cross_boundary(altmap, start, page_size)) { 232 p = vmemmap_alloc_block_buf(page_size, node, altmap); 239 p = vmemmap_alloc_block_buf(page_size, node, NULL); 252 int nr_pfns = page_size >> PAGE_SHIF 187 altmap_cross_boundary(struct vmem_altmap *altmap, unsigned long start, unsigned long page_size) argument 322 unsigned long page_size = 1 << mmu_psize_defs[mmu_vmemmap_psize].shift; local [all...] |
/linux-master/drivers/misc/ |
H A D | vmw_balloon.c | 245 enum vmballoon_page_size_type page_size; member in struct:vmballoon_ctl 560 * @page_size: the size of the page. 565 unsigned int vmballoon_page_order(enum vmballoon_page_size_type page_size) argument 567 return page_size == VMW_BALLOON_2M_PAGE ? VMW_BALLOON_2M_ORDER : 0; 572 * @page_size: the size of the page. 577 vmballoon_page_in_frames(enum vmballoon_page_size_type page_size) argument 579 return 1 << vmballoon_page_order(page_size); 585 * @page_size: the size of the page. 589 enum vmballoon_page_size_type page_size) 593 for (i = 0; i < vmballoon_page_in_frames(page_size); 588 vmballoon_mark_page_offline(struct page *page, enum vmballoon_page_size_type page_size) argument 603 vmballoon_mark_page_online(struct page *page, enum vmballoon_page_size_type page_size) argument 703 vmballoon_handle_one_result(struct vmballoon *b, struct page *page, enum vmballoon_page_size_type page_size, unsigned long status) argument 771 vmballoon_lock_op(struct vmballoon *b, unsigned int num_pages, enum vmballoon_page_size_type page_size, enum vmballoon_op op) argument 909 vmballoon_release_page_list(struct list_head *page_list, int *n_pages, enum vmballoon_page_size_type page_size) argument 985 vmballoon_enqueue_page_list(struct vmballoon *b, struct list_head *pages, unsigned int *n_pages, enum vmballoon_page_size_type page_size) argument 1028 vmballoon_dequeue_page_list(struct vmballoon *b, struct list_head *pages, unsigned int *n_pages, enum vmballoon_page_size_type page_size, unsigned int n_req_pages) argument [all...] |
/linux-master/include/trace/events/ |
H A D | habanalabs.h | 17 TP_PROTO(struct device *dev, u64 virt_addr, u64 phys_addr, u32 page_size, bool flush_pte), 19 TP_ARGS(dev, virt_addr, phys_addr, page_size, flush_pte), 25 __field(u32, page_size) 33 __entry->page_size = page_size; 41 __entry->page_size, 46 TP_PROTO(struct device *dev, u64 virt_addr, u64 phys_addr, u32 page_size, bool flush_pte), 47 TP_ARGS(dev, virt_addr, phys_addr, page_size, flush_pte)); 50 TP_PROTO(struct device *dev, u64 virt_addr, u64 phys_addr, u32 page_size, bool flush_pte), 51 TP_ARGS(dev, virt_addr, phys_addr, page_size, flush_pt [all...] |
/linux-master/tools/perf/arch/arm64/util/ |
H A D | arm-spe.c | 20 #include <internal/lib.h> // page_size 76 * The default auxtrace mmap size is 4MiB/page_size for privileged users, 128KiB for 78 * will be reduced from the default 512KiB/page_size to 256KiB/page_size, otherwise the 87 opts->auxtrace_mmap_pages = MiB(4) / page_size; 89 opts->auxtrace_mmap_pages = KiB(128) / page_size; 91 opts->mmap_pages = KiB(256) / page_size; 94 opts->mmap_pages = KiB(256) / page_size; 102 opts->auxtrace_snapshot_size = opts->auxtrace_mmap_pages * (size_t)page_size; 111 sz = round_up(sz, page_size) / page_siz [all...] |
/linux-master/arch/arm/mach-rpc/include/mach/ |
H A D | uncompress.h | 19 unsigned long page_size; member in struct:param_struct 116 unsigned int nr_pages = 0, page_size = PAGE_SIZE; local 129 page_size = PAGE_SIZE; 135 page_size = params->page_size; 179 if (nr_pages * page_size < 4096*1024) error("<4M of mem\n");
|
/linux-master/drivers/gpu/drm/i915/gem/ |
H A D | i915_gem_region.c | 36 resource_size_t page_size, 60 if (page_size) 61 default_page_size = page_size; 94 err = mem->ops->init_object(mem, obj, offset, size, page_size, flags); 109 resource_size_t page_size, 113 size, page_size, flags); 33 __i915_gem_object_create_region(struct intel_memory_region *mem, resource_size_t offset, resource_size_t size, resource_size_t page_size, unsigned int flags) argument 107 i915_gem_object_create_region(struct intel_memory_region *mem, resource_size_t size, resource_size_t page_size, unsigned int flags) argument
|
/linux-master/tools/testing/selftests/futex/functional/ |
H A D | futex_wait_uninitialized_heap.c | 69 long page_size; local 89 page_size = sysconf(_SC_PAGESIZE); 91 buf = mmap(NULL, page_size, PROT_READ|PROT_WRITE,
|