/linux-master/arch/s390/mm/ |
H A D | hugetlbpage.c | 258 unsigned long pgoff, unsigned long flags) 274 unsigned long pgoff, unsigned long flags) 306 unsigned long len, unsigned long pgoff, unsigned long flags) 333 pgoff, flags); 336 pgoff, flags); 256 hugetlb_get_unmapped_area_bottomup(struct file *file, unsigned long addr, unsigned long len, unsigned long pgoff, unsigned long flags) argument 272 hugetlb_get_unmapped_area_topdown(struct file *file, unsigned long addr0, unsigned long len, unsigned long pgoff, unsigned long flags) argument 305 hugetlb_get_unmapped_area(struct file *file, unsigned long addr, unsigned long len, unsigned long pgoff, unsigned long flags) argument
|
/linux-master/mm/ |
H A D | hugetlb.c | 5876 pgoff_t pgoff; local 5883 pgoff = ((address - vma->vm_start) >> PAGE_SHIFT) + 5893 vma_interval_tree_foreach(iter_vma, &mapping->i_mmap, pgoff, pgoff) { 6174 hash = hugetlb_fault_mutex_hash(mapping, vmf->pgoff); 6450 .pgoff = vma_hugecache_offset(h, vma, haddr), 6465 hash = hugetlb_fault_mutex_hash(mapping, vmf.pgoff); 6499 return hugetlb_no_page(mm, vma, mapping, vmf.pgoff, address, 6547 vmf.pgoff);
|
H A D | shmem.c | 2205 vmf->pgoff >= shmem_falloc->start && 2206 vmf->pgoff < shmem_falloc->next) { 2255 err = shmem_get_folio_gfp(inode, vmf->pgoff, &folio, SGP_CACHE, 2260 vmf->page = folio_file_page(folio, vmf->pgoff); 2268 unsigned long pgoff, unsigned long flags) 2282 addr = get_area(file, uaddr, len, pgoff, flags); 2327 offset = (pgoff << PAGE_SHIFT) & (HPAGE_PMD_SIZE-1); 2616 pgoff_t pgoff = linear_page_index(dst_vma, dst_addr); local 2637 folio = shmem_alloc_folio(gfp, info, pgoff); 2691 if (unlikely(pgoff > 2266 shmem_get_unmapped_area(struct file *file, unsigned long uaddr, unsigned long len, unsigned long pgoff, unsigned long flags) argument 4800 shmem_get_unmapped_area(struct file *file, unsigned long addr, unsigned long len, unsigned long pgoff, unsigned long flags) argument [all...] |
H A D | memory-failure.c | 616 pgoff_t pgoff; local 622 pgoff = page_to_pgoff(page); 631 pgoff, pgoff) { 653 pgoff_t pgoff; local 657 pgoff = page_to_pgoff(page); 663 vma_interval_tree_foreach(vma, &mapping->i_mmap, pgoff, 664 pgoff) { 683 struct list_head *to_kill, pgoff_t pgoff) 685 __add_to_kill(tsk, p, vma, to_kill, 0, pgoff); 681 add_to_kill_fsdax(struct task_struct *tsk, struct page *p, struct vm_area_struct *vma, struct list_head *to_kill, pgoff_t pgoff) argument 691 collect_procs_fsdax(struct page *page, struct address_space *mapping, pgoff_t pgoff, struct list_head *to_kill, bool pre_remove) argument [all...] |
H A D | internal.h | 713 pgoff_t pgoff, addr; local 726 pgoff = folio_pgoff(folio); 729 if (!in_range(pgoff, vma->vm_pgoff, vma_pglen)) 732 addr = vma->vm_start + ((pgoff - vma->vm_pgoff) << PAGE_SHIFT); 797 vma_pgoff_address(pgoff_t pgoff, unsigned long nr_pages, argument 802 if (pgoff >= vma->vm_pgoff) { 804 ((pgoff - vma->vm_pgoff) << PAGE_SHIFT); 808 } else if (pgoff + nr_pages - 1 >= vma->vm_pgoff) { 836 pgoff_t pgoff; local 839 /* Common case, plus ->pgoff i 1215 vma_set_range(struct vm_area_struct *vma, unsigned long start, unsigned long end, pgoff_t pgoff) argument [all...] |
H A D | huge_memory.c | 855 unsigned long len, unsigned long pgoff, unsigned long flags) 858 loff_t off = (loff_t)pgoff << PAGE_SHIFT; 864 return current->mm->get_unmapped_area(filp, addr, len, pgoff, flags); 854 thp_get_unmapped_area(struct file *filp, unsigned long addr, unsigned long len, unsigned long pgoff, unsigned long flags) argument
|
H A D | vmalloc.c | 4346 * @pgoff: offset from @kaddr to start at 4359 void *kaddr, unsigned long pgoff, 4366 if (check_shl_overflow(pgoff, PAGE_SHIFT, &off)) 4408 * @pgoff: number of pages into addr before first page to map 4419 unsigned long pgoff) 4422 addr, pgoff, 4358 remap_vmalloc_range_partial(struct vm_area_struct *vma, unsigned long uaddr, void *kaddr, unsigned long pgoff, unsigned long size) argument 4418 remap_vmalloc_range(struct vm_area_struct *vma, void *addr, unsigned long pgoff) argument
|
H A D | memory.c | 4833 pgoff_t vma_off = vmf->pgoff - vmf->vma->vm_pgoff; 4853 vmf->pgoff + from_pte - pte_off, 4854 vmf->pgoff + to_pte - pte_off); 5358 .pgoff = linear_page_index(vma, address),
|
H A D | userfaultfd.c | 357 pgoff_t pgoff = linear_page_index(dst_vma, dst_addr); local 362 ret = shmem_get_folio(inode, pgoff, &folio, SGP_NOALLOC); 373 page = folio_file_page(folio, pgoff);
|
H A D | filemap.c | 3096 DEFINE_READAHEAD(ractl, file, ra, mapping, vmf->pgoff); 3147 ra->start = max_t(long, 0, vmf->pgoff - ra->ra_pages / 2); 3165 DEFINE_READAHEAD(ractl, file, ra, file->f_mapping, vmf->pgoff); 3256 pgoff_t max_idx, index = vmf->pgoff;
|
/linux-master/include/linux/ |
H A D | mm.h | 518 * pgoff should be used in favour of virtual_address, if possible. 524 pgoff_t pgoff; /* Logical page offset based on vma */ member in struct:vm_fault::__anon82 3277 unsigned long start, unsigned long end, pgoff_t pgoff, 3280 unsigned long start, unsigned long end, pgoff_t pgoff); 3285 unsigned long addr, unsigned long len, pgoff_t pgoff, 3391 unsigned long len, vm_flags_t vm_flags, unsigned long pgoff, 3395 vm_flags_t vm_flags, unsigned long pgoff, unsigned long *populate,
|
H A D | shmem_fs.h | 97 unsigned long len, unsigned long pgoff, unsigned long flags);
|
H A D | bpf.h | 144 unsigned long len, unsigned long pgoff,
|
/linux-master/drivers/video/fbdev/core/ |
H A D | fb_defio.c | 45 unsigned long pgoff = offset >> PAGE_SHIFT; local 48 if (WARN_ON_ONCE(pgoff >= info->npagerefs)) 52 pageref = &info->pagerefs[pgoff]; 64 pageref->offset = pgoff << PAGE_SHIFT; 100 offset = vmf->pgoff << PAGE_SHIFT; 116 page->index = vmf->pgoff; /* for page_mkclean() */ 199 unsigned long offset = vmf->pgoff << PAGE_SHIFT;
|
/linux-master/drivers/md/ |
H A D | dm.c | 1232 static long dm_dax_direct_access(struct dax_device *dax_dev, pgoff_t pgoff, argument 1237 sector_t sector = pgoff * PAGE_SECTORS; 1252 ret = ti->type->direct_access(ti, pgoff, nr_pages, mode, kaddr, pfn); 1260 static int dm_dax_zero_page_range(struct dax_device *dax_dev, pgoff_t pgoff, argument 1264 sector_t sector = pgoff * PAGE_SECTORS; 1280 ret = ti->type->dax_zero_page_range(ti, pgoff, nr_pages); 1287 static size_t dm_dax_recovery_write(struct dax_device *dax_dev, pgoff_t pgoff, argument 1291 sector_t sector = pgoff * PAGE_SECTORS; 1300 ret = ti->type->dax_recovery_write(ti, pgoff, addr, bytes, i);
|
/linux-master/fs/btrfs/ |
H A D | scrub.c | 695 unsigned int pgoff = scrub_stripe_get_page_offset(stripe, sector_nr); local 740 ret = btrfs_check_sector_csum(fs_info, page, pgoff, csum_buf, sector->csum); 830 int pgoff; local 834 pgoff = scrub_stripe_get_page_offset(stripe, i); 854 ret = bio_add_page(&bbio->bio, page, fs_info->sectorsize, pgoff); 1203 unsigned int pgoff = scrub_stripe_get_page_offset(stripe, sector_nr); local 1221 ret = bio_add_page(&bbio->bio, page, fs_info->sectorsize, pgoff); 1667 unsigned int pgoff = scrub_stripe_get_page_offset(stripe, i); local 1707 __bio_add_page(&bbio->bio, page, fs_info->sectorsize, pgoff); 1749 unsigned int pgoff local [all...] |
H A D | inode.c | 3293 u32 pgoff, u8 *csum, const u8 * const csum_expected) 3298 ASSERT(pgoff + fs_info->sectorsize <= PAGE_SIZE); 3302 kaddr = kmap_local_page(page) + pgoff; 3292 btrfs_check_sector_csum(struct btrfs_fs_info *fs_info, struct page *page, u32 pgoff, u8 *csum, const u8 * const csum_expected) argument
|
/linux-master/virt/kvm/ |
H A D | kvm_main.c | 4128 static bool kvm_page_in_dirty_ring(struct kvm *kvm, unsigned long pgoff) argument 4131 return (pgoff >= KVM_DIRTY_LOG_PAGE_OFFSET) && 4132 (pgoff < KVM_DIRTY_LOG_PAGE_OFFSET + 4144 if (vmf->pgoff == 0) 4147 else if (vmf->pgoff == KVM_PIO_PAGE_OFFSET) 4151 else if (vmf->pgoff == KVM_COALESCED_MMIO_PAGE_OFFSET) 4154 else if (kvm_page_in_dirty_ring(vcpu->kvm, vmf->pgoff)) 4157 vmf->pgoff - KVM_DIRTY_LOG_PAGE_OFFSET);
|
/linux-master/io_uring/ |
H A D | io_uring.c | 3428 loff_t pgoff, size_t sz) 3431 loff_t offset = pgoff << PAGE_SHIFT; 3490 unsigned long pgoff, unsigned long flags) 3502 ptr = io_uring_validate_mmap_request(filp, pgoff, len); 3514 * - use the same pgoff which the get_unmapped_area() uses to 3521 pgoff = 0; /* has been translated to ptr above */ 3524 pgoff = addr >> PAGE_SHIFT; 3528 return current->mm->get_unmapped_area(filp, addr, len, pgoff, flags); 3545 unsigned long pgoff, unsigned long flags) 3549 ptr = io_uring_validate_mmap_request(file, pgoff, le 3427 io_uring_validate_mmap_request(struct file *file, loff_t pgoff, size_t sz) argument 3488 io_uring_mmu_get_unmapped_area(struct file *filp, unsigned long addr, unsigned long len, unsigned long pgoff, unsigned long flags) argument 3543 io_uring_nommu_get_unmapped_area(struct file *file, unsigned long addr, unsigned long len, unsigned long pgoff, unsigned long flags) argument [all...] |
/linux-master/fs/ceph/ |
H A D | addr.c | 1598 loff_t off = (loff_t)vmf->pgoff << PAGE_SHIFT;
|
/linux-master/drivers/scsi/ |
H A D | sg.c | 1233 offset = vmf->pgoff << PAGE_SHIFT;
|
/linux-master/arch/loongarch/mm/ |
H A D | mmap.c | 14 #define COLOUR_ALIGN(addr, pgoff) \ 16 + (((pgoff) << PAGE_SHIFT) & SHM_ALIGN_MASK)) 21 unsigned long addr0, unsigned long len, unsigned long pgoff, 43 ((addr - (pgoff << PAGE_SHIFT)) & SHM_ALIGN_MASK)) 55 addr = COLOUR_ALIGN(addr, pgoff); 67 info.align_offset = pgoff << PAGE_SHIFT; 93 unsigned long len, unsigned long pgoff, unsigned long flags) 96 addr0, len, pgoff, flags, UP); 104 unsigned long addr0, unsigned long len, unsigned long pgoff, 108 addr0, len, pgoff, flag 20 arch_get_unmapped_area_common(struct file *filp, unsigned long addr0, unsigned long len, unsigned long pgoff, unsigned long flags, enum mmap_allocation_direction dir) argument 92 arch_get_unmapped_area(struct file *filp, unsigned long addr0, unsigned long len, unsigned long pgoff, unsigned long flags) argument 103 arch_get_unmapped_area_topdown(struct file *filp, unsigned long addr0, unsigned long len, unsigned long pgoff, unsigned long flags) argument [all...] |
/linux-master/kernel/bpf/ |
H A D | syscall.c | 975 unsigned long len, unsigned long pgoff, 981 return map->ops->map_get_unmapped_area(filp, addr, len, pgoff, flags); 983 return current->mm->get_unmapped_area(filp, addr, len, pgoff, flags); 974 bpf_get_unmapped_area(struct file *filp, unsigned long addr, unsigned long len, unsigned long pgoff, unsigned long flags) argument
|
/linux-master/fs/cramfs/ |
H A D | inode.c | 288 static u32 cramfs_get_block_range(struct inode *inode, u32 pgoff, u32 *pages) argument 299 blockptrs = (u32 *)(sbi->linear_virt_addr + OFFSET(inode) + pgoff * 4); 309 pgoff+i, pgoff + *pages - 1, 351 unsigned long address, pgoff = vma->vm_pgoff; local 371 if (pgoff >= max_pages) 373 pages = min(vma_pages(vma), max_pages - pgoff); 375 offset = cramfs_get_block_range(inode, pgoff, &pages); 385 if (pgoff + pages == max_pages && cramfs_last_page_is_shared(inode)) { 425 pgoff, addres 443 cramfs_physmem_get_unmapped_area(struct file *file, unsigned long addr, unsigned long len, unsigned long pgoff, unsigned long flags) argument [all...] |
/linux-master/security/selinux/ |
H A D | selinuxfs.c | 451 offset = vmf->pgoff << PAGE_SHIFT;
|