/linux-master/tools/lib/perf/ |
H A D | core.c | 36 page_size = sysconf(_SC_PAGE_SIZE);
|
/linux-master/drivers/infiniband/hw/hns/ |
H A D | hns_roce_alloc.c | 69 u32 trunk_size, page_size, alloced_size; local 85 page_size = 1 << buf->page_shift; 92 buf->trunk_shift = order_base_2(ALIGN(page_size, PAGE_SIZE)); 126 buf->npages = DIV_ROUND_UP(alloced_size, page_size);
|
/linux-master/tools/perf/arch/arm64/util/ |
H A D | hisi-ptt.c | 14 #include <internal/lib.h> // page_size 73 opts->auxtrace_mmap_pages = MiB(16) / page_size; 75 opts->auxtrace_mmap_pages = KiB(128) / page_size; 77 opts->mmap_pages = KiB(256) / page_size; 83 size_t sz = opts->auxtrace_mmap_pages * (size_t)page_size;
|
/linux-master/tools/lib/bpf/ |
H A D | ringbuf.c | 38 size_t page_size; member in struct:ring_buffer 49 size_t page_size; member in struct:user_ring_buffer 63 munmap(r->consumer_pos, rb->page_size); 67 munmap(r->producer_pos, rb->page_size + 2 * (r->mask + 1)); 123 tmp = mmap(NULL, rb->page_size, PROT_READ | PROT_WRITE, MAP_SHARED, map_fd, 0); 136 mmap_sz = rb->page_size + 2 * (__u64)info.max_entries; 142 tmp = mmap(NULL, (size_t)mmap_sz, PROT_READ, MAP_SHARED, map_fd, rb->page_size); 150 r->data = tmp + rb->page_size; 203 rb->page_size = getpagesize(); 388 munmap(rb->consumer_pos, rb->page_size); [all...] |
/linux-master/tools/testing/selftests/mm/ |
H A D | mlock-random-test.c | 138 int page_size = 0; local 174 page_size = get_proc_page_size((unsigned long)p); 176 if (locked_vm_size > PAGE_ALIGN(alloc_size, page_size) + page_size)
|
/linux-master/drivers/net/ethernet/huawei/hinic/ |
H A D | hinic_hw_io.h | 101 u8 page_size; member in struct:hinic_wq_page_size 107 u32 page_size);
|
/linux-master/drivers/gpu/drm/i915/gem/ |
H A D | i915_gem_lmem.c | 83 * up depending on the @page_size. The final object size can be fished out from 85 * @page_size: The requested minimum page size in bytes for this object. This is 93 * @page_size. If this is smaller than the regions min_page_size then it can 102 resource_size_t page_size, 106 size, page_size, flags); 100 __i915_gem_object_create_lmem_with_ps(struct drm_i915_private *i915, resource_size_t size, resource_size_t page_size, unsigned int flags) argument
|
/linux-master/arch/powerpc/include/asm/book3s/64/ |
H A D | tlbflush-radix.h | 37 unsigned long page_size); 45 unsigned long page_size) 43 radix__flush_tlb_lpid_page(unsigned int lpid, unsigned long addr, unsigned long page_size) argument
|
/linux-master/tools/testing/selftests/kvm/x86_64/ |
H A D | hyperv_extended_hypercalls.c | 61 memset(addr_gva2hva(vm, hcall_in_page), 0x0, vm->page_size); 65 memset(addr_gva2hva(vm, hcall_out_page), 0x0, vm->page_size);
|
/linux-master/tools/testing/selftests/arm64/mte/ |
H A D | check_child_memory.c | 23 static size_t page_size; variable 26 /* page size - 1*/ 0, /* page_size */ 0, /* page size + 1 */ 0 149 page_size = getpagesize(); 150 if (!page_size) { 154 sizes[item - 3] = page_size - 1; 155 sizes[item - 2] = page_size; 156 sizes[item - 1] = page_size + 1;
|
/linux-master/tools/testing/selftests/user_events/ |
H A D | perf_test.c | 138 int page_size = sysconf(_SC_PAGESIZE); local 170 perf_page = mmap(NULL, page_size * 2, PROT_READ, MAP_SHARED, fd, 0); 191 munmap(perf_page, page_size * 2); 202 int page_size = sysconf(_SC_PAGESIZE); local 232 perf_page = mmap(NULL, page_size * 2, PROT_READ, MAP_SHARED, fd, 0); 244 munmap(perf_page, page_size * 2);
|
/linux-master/tools/testing/selftests/kvm/lib/aarch64/ |
H A D | processor.c | 23 return (v + vm->page_size) & ~(vm->page_size - 1); 64 return (vm->page_size == SZ_4K || vm->page_size == SZ_16K) && 115 size_t nr_pages = page_align(vm, ptrs_per_pgd(vm) * 8) / vm->page_size; 132 TEST_ASSERT((vaddr % vm->page_size) == 0, 134 " vaddr: 0x%lx vm->page_size: 0x%x", vaddr, vm->page_size); 138 TEST_ASSERT((paddr % vm->page_size) == 0, 140 " paddr: 0x%lx vm->page_size [all...] |
/linux-master/tools/testing/selftests/bpf/prog_tests/ |
H A D | cgroup_getset_retval.c | 28 obj->bss->page_size = sysconf(_SC_PAGESIZE); 64 obj->bss->page_size = sysconf(_SC_PAGESIZE); 107 obj->bss->page_size = sysconf(_SC_PAGESIZE); 143 obj->bss->page_size = sysconf(_SC_PAGESIZE); 188 obj->bss->page_size = sysconf(_SC_PAGESIZE); 237 obj->bss->page_size = sysconf(_SC_PAGESIZE); 283 obj->bss->page_size = sysconf(_SC_PAGESIZE); 336 obj->bss->page_size = sysconf(_SC_PAGESIZE); 378 obj->bss->page_size = sysconf(_SC_PAGESIZE); 417 obj->bss->page_size [all...] |
/linux-master/drivers/gpu/drm/i915/gem/selftests/ |
H A D | huge_pages.c | 56 unsigned int page_size = page_sizes[i]; local 58 if (HAS_PAGE_SIZES(i915, page_size) && rem >= page_size) 59 return page_size; 112 unsigned int page_size = BIT(bit); local 113 int order = get_order(page_size); 123 sg_set_page(sg, page, page_size, 0); 124 sg_page_sizes |= page_size; 127 rem -= page_size; 134 } while ((rem - ((page_size 237 unsigned int page_size = get_largest_page_size(i915, rem); local 271 unsigned int page_size; local 516 unsigned int page_size = BIT(bit); local 596 unsigned int page_size = BIT(bit); local 794 unsigned int page_size = page_sizes[i]; local [all...] |
/linux-master/tools/testing/selftests/kvm/lib/ |
H A D | kvm_util.c | 283 vm->page_size = vm_guest_mode_params[vm->mode].page_size; 372 uint64_t page_size = vm_guest_mode_params[mode].page_size; local 402 nr_pages += ucall_nr_pages_required(page_size); 834 amt = vm->page_size - (ptr1 % vm->page_size); 836 amt = vm->page_size - (ptr2 % vm->page_size); 970 size_t mem_size = npages * vm->page_size; 1565 size_t page_size = vm->page_size; local 1755 uint32_t page_size = getpagesize(); local [all...] |
/linux-master/drivers/memstick/core/ |
H A D | mspro_block.c | 74 __be16 page_size; member in struct:mspro_sys_info 143 unsigned short page_size; member in struct:mspro_block_data 298 rc += sysfs_emit_at(buffer, rc, "page size: %x\n", be16_to_cpu(x_sys->page_size)); 536 / msb->page_size)) { 560 t_offset += msb->current_page * msb->page_size; 565 msb->page_size, offset_in_page(t_offset)); 604 .data_count = cpu_to_be16((uint16_t)(length / msb->page_size)), 610 do_div(offset, msb->page_size); 686 / msb->page_size; 691 t_len *= msb->page_size; [all...] |
H A D | ms_block.h | 79 unsigned short page_size; member in struct:ms_boot_attr_info 112 unsigned short page_size; member in struct:ms_cis_idi 169 unsigned short page_size; member in struct:msb_data
|
/linux-master/tools/testing/selftests/landlock/ |
H A D | base_test.c | 26 const long page_size = sysconf(_SC_PAGESIZE); local 27 char *const buf = malloc(page_size + 1); 49 ASSERT_EQ(-1, landlock_create_ruleset(ruleset_attr, page_size + 1, 0)); 59 ASSERT_EQ(-1, landlock_create_ruleset(ruleset_attr, page_size, 0)); 63 buf[page_size - 2] = '.'; 64 ASSERT_EQ(-1, landlock_create_ruleset(ruleset_attr, page_size, 0)); 67 ASSERT_EQ(-1, landlock_create_ruleset(ruleset_attr, page_size + 1, 0));
|
/linux-master/mm/ |
H A D | mmu_gather.c | 166 int page_size) 174 VM_WARN_ON(tlb->page_size != page_size); 175 VM_WARN_ON_ONCE(nr_pages != 1 && page_size != PAGE_SIZE); 213 bool delay_rmap, int page_size) 215 return __tlb_remove_folio_pages_size(tlb, page, 1, delay_rmap, page_size); 394 tlb->page_size = 0; 164 __tlb_remove_folio_pages_size(struct mmu_gather *tlb, struct page *page, unsigned int nr_pages, bool delay_rmap, int page_size) argument 212 __tlb_remove_page_size(struct mmu_gather *tlb, struct page *page, bool delay_rmap, int page_size) argument
|
/linux-master/drivers/vfio/pci/pds/ |
H A D | dirty.c | 222 region->page_size = region_page_size; 245 u64 *page_size) 251 u64 region_page_size = *page_size; 309 "region_info[%d]: region_start 0x%llx region_end 0x%lx region_size 0x%llx page_count %u page_size %llu\n", 473 u64 page_size = region->page_size; local 494 u64 addr = abs_bit_i * page_size + region_start; 496 iova_bitmap_set(dirty_bitmap, addr, page_size); 545 pages = DIV_ROUND_UP(length, region->page_size); 550 "vf%u: iova 0x%lx length %lu page_size 243 pds_vfio_dirty_enable(struct pds_vfio_pci_device *pds_vfio, struct rb_root_cached *ranges, u32 nnodes, u64 *page_size) argument 615 pds_vfio_dma_logging_start(struct vfio_device *vdev, struct rb_root_cached *ranges, u32 nnodes, u64 *page_size) argument [all...] |
/linux-master/tools/testing/selftests/kvm/lib/riscv/ |
H A D | processor.c | 30 return (v + vm->page_size) & ~(vm->page_size - 1); 70 size_t nr_pages = page_align(vm, ptrs_per_pte(vm) * 8) / vm->page_size; 86 TEST_ASSERT((vaddr % vm->page_size) == 0, 88 " vaddr: 0x%lx vm->page_size: 0x%x", vaddr, vm->page_size); 92 TEST_ASSERT((paddr % vm->page_size) == 0, 94 " paddr: 0x%lx vm->page_size: 0x%x", paddr, vm->page_size); 97 " paddr: 0x%lx vm->max_gfn: 0x%lx vm->page_size [all...] |
/linux-master/include/linux/qed/ |
H A D | qed_chain.h | 124 u32 page_size; member in struct:qed_chain 153 u32 page_size; member in struct:qed_chain_init_params 163 #define ELEMS_PER_PAGE(elem_size, page_size) \ 164 ((page_size) / (elem_size)) 171 #define USABLE_ELEMS_PER_PAGE(elem_size, page_size, mode) \ 172 ((u32)(ELEMS_PER_PAGE((elem_size), (page_size)) - \ 175 #define QED_CHAIN_PAGE_CNT(elem_cnt, elem_size, page_size, mode) \ 177 USABLE_ELEMS_PER_PAGE((elem_size), (page_size), (mode))) 635 p_chain->page_size);
|
/linux-master/drivers/accel/habanalabs/common/mmu/ |
H A D | mmu_v2_hr.c | 190 static int hl_mmu_v2_get_last_hop(struct hl_mmu_properties *mmu_prop, u32 page_size) argument 198 if (page_size <= (1 << mmu_prop->hop_shifts[hop])) 207 u32 page_size, bool is_dram_addr) 227 else if (page_size == prop->pmmu_huge.page_size) 232 hop_last = hl_mmu_v2_get_last_hop(mmu_prop, page_size); 341 is_dram_addr = hl_mem_area_inside_range(virt_addr, prop->dmmu.page_size, 344 is_pmmu_addr = hl_mem_area_inside_range(virt_addr, prop->pmmu.page_size, 348 prop->pmmu_huge.page_size, 205 _hl_mmu_v2_hr_map(struct hl_ctx *ctx, u64 virt_addr, u64 phys_addr, u32 page_size, bool is_dram_addr) argument
|
/linux-master/drivers/mtd/devices/ |
H A D | spear_smi.c | 190 * @page_size: Page size of NOR-flash. 202 u32 page_size; member in struct:spear_snor_flash 680 u32 page_offset, page_size; local 695 page_offset = (u32)to % flash->page_size; 698 if (page_offset + len <= flash->page_size) { 706 page_size = flash->page_size - page_offset; 709 page_size); 713 *retlen += page_size; 716 for (i = page_size; [all...] |
/linux-master/tools/perf/arch/x86/util/ |
H A D | intel-pt.c | 33 #include <internal/lib.h> // page_size 665 opts->auxtrace_mmap_pages = MiB(4) / page_size; 667 opts->auxtrace_mmap_pages = KiB(128) / page_size; 669 opts->mmap_pages = KiB(256) / page_size; 673 opts->mmap_pages = KiB(256) / page_size; 677 opts->auxtrace_mmap_pages * (size_t)page_size; 681 sz = round_up(sz, page_size) / page_size; 685 opts->auxtrace_mmap_pages * (size_t)page_size) { 688 opts->auxtrace_mmap_pages * (size_t)page_size); [all...] |