/linux-master/arch/x86/kernel/ |
H A D | sys_x86_64.c | 117 unsigned long len, unsigned long pgoff, unsigned long flags) 145 info.align_offset = pgoff << PAGE_SHIFT; 155 const unsigned long len, const unsigned long pgoff, 207 info.align_offset = pgoff << PAGE_SHIFT; 224 return arch_get_unmapped_area(filp, addr0, len, pgoff, flags); 116 arch_get_unmapped_area(struct file *filp, unsigned long addr, unsigned long len, unsigned long pgoff, unsigned long flags) argument 154 arch_get_unmapped_area_topdown(struct file *filp, const unsigned long addr0, const unsigned long len, const unsigned long pgoff, const unsigned long flags) argument
|
/linux-master/include/linux/ |
H A D | dax.h | 42 size_t (*recovery_write)(struct dax_device *dax_dev, pgoff_t pgoff, 69 size_t dax_recovery_write(struct dax_device *dax_dev, pgoff_t pgoff, 125 pgoff_t pgoff, void *addr, size_t bytes, struct iov_iter *i) 231 long dax_direct_access(struct dax_device *dax_dev, pgoff_t pgoff, long nr_pages, 233 size_t dax_copy_from_iter(struct dax_device *dax_dev, pgoff_t pgoff, void *addr, 235 size_t dax_copy_to_iter(struct dax_device *dax_dev, pgoff_t pgoff, void *addr, 237 int dax_zero_page_range(struct dax_device *dax_dev, pgoff_t pgoff, 124 dax_recovery_write(struct dax_device *dax_dev, pgoff_t pgoff, void *addr, size_t bytes, struct iov_iter *i) argument
|
H A D | coredump.h | 15 unsigned long pgoff; member in struct:core_vma_metadata
|
H A D | futex.h | 35 unsigned long pgoff; member in struct:futex_key::__anon11
|
/linux-master/drivers/md/ |
H A D | dm-stripe.c | 301 static struct dax_device *stripe_dax_pgoff(struct dm_target *ti, pgoff_t *pgoff) argument 308 stripe_map_sector(sc, *pgoff * PAGE_SECTORS, &stripe, &dev_sector); 312 *pgoff = (get_start_sect(bdev) + dev_sector) >> PAGE_SECTORS_SHIFT; 316 static long stripe_dax_direct_access(struct dm_target *ti, pgoff_t pgoff, argument 320 struct dax_device *dax_dev = stripe_dax_pgoff(ti, &pgoff); 322 return dax_direct_access(dax_dev, pgoff, nr_pages, mode, kaddr, pfn); 325 static int stripe_dax_zero_page_range(struct dm_target *ti, pgoff_t pgoff, argument 328 struct dax_device *dax_dev = stripe_dax_pgoff(ti, &pgoff); 330 return dax_zero_page_range(dax_dev, pgoff, nr_pages); 333 static size_t stripe_dax_recovery_write(struct dm_target *ti, pgoff_t pgoff, argument [all...] |
/linux-master/kernel/bpf/ |
H A D | arena.c | 18 * In this example 7f7d26200000 is the address of the first page (pgoff=0). 22 * ffffc90001e69000 + 26200000 == ffffc90028069000 is "pgoff=0" within 4Gb 266 ret = mtree_insert(&arena->mt, vmf->pgoff, MT_ENTRY, GFP_KERNEL); 273 mtree_erase(&arena->mt, vmf->pgoff); 279 mtree_erase(&arena->mt, vmf->pgoff); 295 unsigned long len, unsigned long pgoff, 302 if (pgoff) 412 long pgoff = 0; local 422 pgoff = compute_pgoff(arena, uaddr); 423 if (pgoff > page_cnt_ma 294 arena_get_unmapped_area(struct file *filp, unsigned long addr, unsigned long len, unsigned long pgoff, unsigned long flags) argument 490 long kaddr, pgoff, i; local [all...] |
/linux-master/tools/perf/util/ |
H A D | copyfile.c | 48 loff_t pgoff; local 50 pgoff = off_in & ~(page_size - 1); 51 off_in -= pgoff; 53 ptr = mmap(NULL, off_in + size, PROT_READ, MAP_PRIVATE, ifd, pgoff);
|
H A D | map.h | 29 u64 pgoff; local 63 return RC_CHK_ACCESS(map)->pgoff; 167 u64 start, u64 end, u64 pgoff, struct dso *dso); 173 u64 pgoff, struct dso_id *id, u32 prot, u32 flags, 270 static inline void map__set_pgoff(struct map *map, u64 pgoff) argument 272 RC_CHK_ACCESS(map)->pgoff = pgoff; 277 RC_CHK_ACCESS(map)->pgoff += inc;
|
/linux-master/drivers/nvdimm/ |
H A D | pmem.h | 32 long __pmem_direct_access(struct pmem_device *pmem, pgoff_t pgoff,
|
/linux-master/virt/kvm/ |
H A D | guest_memfd.c | 60 pgoff_t pgoff = slot->gmem.pgoff; local 63 .start = slot->base_gfn + max(pgoff, start) - pgoff, 64 .end = slot->base_gfn + min(pgoff + slot->npages, end) - pgoff, 407 BUILD_BUG_ON(sizeof(gfn_t) != sizeof(slot->gmem.pgoff)); 443 slot->gmem.pgoff = start; 461 unsigned long start = slot->gmem.pgoff; 488 pgoff_t index = gfn - slot->base_gfn + slot->gmem.pgoff; [all...] |
/linux-master/arch/sparc/kernel/ |
H A D | sys_sparc_64.c | 82 unsigned long pgoff) 85 unsigned long off = (pgoff<<PAGE_SHIFT) & (SHMLBA-1); 90 unsigned long arch_get_unmapped_area(struct file *filp, unsigned long addr, unsigned long len, unsigned long pgoff, unsigned long flags) argument 103 ((addr - (pgoff << PAGE_SHIFT)) & (SHMLBA - 1))) 119 addr = COLOR_ALIGN(addr, pgoff); 134 info.align_offset = pgoff << PAGE_SHIFT; 149 const unsigned long len, const unsigned long pgoff, 167 ((addr - (pgoff << PAGE_SHIFT)) & (SHMLBA - 1))) 182 addr = COLOR_ALIGN(addr, pgoff); 197 info.align_offset = pgoff << PAGE_SHIF 81 COLOR_ALIGN(unsigned long addr, unsigned long pgoff) argument 148 arch_get_unmapped_area_topdown(struct file *filp, const unsigned long addr0, const unsigned long len, const unsigned long pgoff, const unsigned long flags) argument 218 get_fb_unmapped_area(struct file *filp, unsigned long orig_addr, unsigned long len, unsigned long pgoff, unsigned long flags) argument [all...] |
/linux-master/kernel/events/ |
H A D | ring_buffer.c | 672 pgoff_t pgoff, int nr_pages, long watermark, int flags) 763 rb->aux_pgoff = pgoff; 783 __perf_mmap_to_page(struct perf_buffer *rb, unsigned long pgoff) argument 785 if (pgoff > rb->nr_pages) 788 if (pgoff == 0) 791 return virt_to_page(rb->data_pages[pgoff - 1]); 873 __perf_mmap_to_page(struct perf_buffer *rb, unsigned long pgoff) argument 876 if (pgoff > data_page_nr(rb)) 879 return vmalloc_to_page((void *)rb->user_page + pgoff * PAGE_SIZE); 954 perf_mmap_to_page(struct perf_buffer *rb, unsigned long pgoff) argument 671 rb_alloc_aux(struct perf_buffer *rb, struct perf_event *event, pgoff_t pgoff, int nr_pages, long watermark, int flags) argument [all...] |
/linux-master/drivers/video/fbdev/core/ |
H A D | fb_defio.c | 45 unsigned long pgoff = offset >> PAGE_SHIFT; local 48 if (WARN_ON_ONCE(pgoff >= info->npagerefs)) 52 pageref = &info->pagerefs[pgoff]; 64 pageref->offset = pgoff << PAGE_SHIFT; 100 offset = vmf->pgoff << PAGE_SHIFT; 116 page->index = vmf->pgoff; /* for page_mkclean() */ 199 unsigned long offset = vmf->pgoff << PAGE_SHIFT;
|
/linux-master/drivers/infiniband/core/ |
H A D | ib_core_uverbs.c | 105 * @pgoff: The mmap offset >> PAGE_SHIFT 118 unsigned long pgoff) 122 if (pgoff > U32_MAX) 127 entry = xa_load(&ucontext->mmap_xa, pgoff); 134 if (!entry || entry->start_pgoff != pgoff || entry->driver_removed || 140 ibdev_dbg(ucontext->device, "mmap: pgoff[%#lx] npages[%#zx] returned\n", 141 pgoff, entry->npages); 195 ibdev_dbg(ucontext->device, "mmap: pgoff[%#lx] npages[%#zx] removed\n", 249 * @min_pgoff: minimum pgoff to be returned 250 * @max_pgoff: maximum pgoff t 117 rdma_user_mmap_entry_get_pgoff(struct ib_ucontext *ucontext, unsigned long pgoff) argument [all...] |
H A D | umem.c | 84 unsigned long va, pgoff; local 107 pgoff = umem->address & ~PAGE_MASK; 113 mask |= (sg_dma_address(sg) + pgoff) ^ va; 114 va += sg_dma_len(sg) - pgoff; 121 pgoff = 0;
|
/linux-master/arch/arm/mm/ |
H A D | fault-armv.c | 141 pgoff_t pgoff; local 144 pgoff = vma->vm_pgoff + ((addr - vma->vm_start) >> PAGE_SHIFT); 152 vma_interval_tree_foreach(mpnt, &mapping->i_mmap, pgoff, pgoff) { 162 offset = (pgoff - mpnt->vm_pgoff) << PAGE_SHIFT;
|
H A D | flush.c | 242 pgoff_t pgoff, pgoff_end; local 250 pgoff = folio->index; 251 pgoff_end = pgoff + folio_nr_pages(folio) - 1; 254 vma_interval_tree_foreach(vma, &mapping->i_mmap, pgoff, pgoff_end) { 269 offset = pgoff - vma->vm_pgoff;
|
/linux-master/arch/nios2/mm/ |
H A D | cacheflush.c | 79 pgoff_t pgoff; local 82 pgoff = folio->index; 85 vma_interval_tree_foreach(vma, &mapping->i_mmap, pgoff, pgoff + nr - 1) { 93 start = vma->vm_start + ((pgoff - vma->vm_pgoff) << PAGE_SHIFT);
|
/linux-master/fs/xfs/ |
H A D | xfs_notify_failure.c | 78 pgoff_t pgoff; local 103 pgoff = xfs_failure_pgoff(mp, rec, notify); 108 error = mf_dax_kill_procs(mapping, pgoff, pgcnt, 113 invalidate_inode_pages2_range(mapping, pgoff, 114 pgoff + pgcnt - 1);
|
/linux-master/fs/ramfs/ |
H A D | file-nommu.c | 29 unsigned long pgoff, 204 unsigned long pgoff, unsigned long flags) 217 if (pgoff >= maxpages) 220 if (maxpages - pgoff < lpages) 227 nr_folios = filemap_get_folios_contig(inode->i_mapping, &pgoff, 202 ramfs_nommu_get_unmapped_area(struct file *file, unsigned long addr, unsigned long len, unsigned long pgoff, unsigned long flags) argument
|
/linux-master/drivers/iommu/iommufd/ |
H A D | iova_bitmap.c | 42 unsigned long pgoff; member in struct:iova_bitmap_map 208 mapped->pgoff = offset_in_page(addr); 307 bytes = (bitmap->mapped.npages << PAGE_SHIFT) - bitmap->mapped.pgoff; 448 mapped->pgshift) + mapped->pgoff * BITS_PER_BYTE; 450 mapped->pgshift) + mapped->pgoff * BITS_PER_BYTE;
|
/linux-master/arch/mips/kernel/ |
H A D | syscall.c | 74 unsigned long, pgoff) 76 if (pgoff & (~PAGE_MASK >> 12)) 80 pgoff >> (PAGE_SHIFT - 12));
|
/linux-master/arch/powerpc/kernel/ |
H A D | syscalls.c | 54 unsigned long, fd, unsigned long, pgoff) 56 return do_mmap2(addr, len, prot, flags, fd, pgoff, PAGE_SHIFT-12);
|
/linux-master/drivers/dax/ |
H A D | super.c | 139 * dax_direct_access() - translate a device pgoff to an absolute pfn 141 * @pgoff: offset in pages from the start of the device to translate 145 * @pfn: output parameter that returns an absolute pfn translation of @pgoff 148 * pages accessible at the device relative @pgoff. 150 long dax_direct_access(struct dax_device *dax_dev, pgoff_t pgoff, long nr_pages, argument 164 avail = dax_dev->ops->direct_access(dax_dev, pgoff, nr_pages, 172 size_t dax_copy_from_iter(struct dax_device *dax_dev, pgoff_t pgoff, void *addr, argument 188 size_t dax_copy_to_iter(struct dax_device *dax_dev, pgoff_t pgoff, void *addr, argument 204 int dax_zero_page_range(struct dax_device *dax_dev, pgoff_t pgoff, argument 219 ret = dax_dev->ops->zero_page_range(dax_dev, pgoff, nr_page 224 dax_recovery_write(struct dax_device *dax_dev, pgoff_t pgoff, void *addr, size_t bytes, struct iov_iter *iter) argument [all...] |
/linux-master/drivers/usb/gadget/function/ |
H A D | uvc_queue.h | 88 unsigned long pgoff);
|