/linux-master/mm/ |
H A D | memory.c | 107 static vm_fault_t do_fault(struct vm_fault *vmf); 108 static vm_fault_t do_anonymous_page(struct vm_fault *vmf); 109 static bool vmf_pte_changed(struct vm_fault *vmf); 115 static bool vmf_orig_pte_uffd_wp(struct vm_fault *vmf) argument 117 if (!(vmf->flags & FAULT_FLAG_ORIG_PTE_VALID)) 120 return pte_marker_uffd_wp(vmf->orig_pte); 2939 static inline int pte_unmap_same(struct vm_fault *vmf) argument 2944 spin_lock(vmf->ptl); 2945 same = pte_same(ptep_get(vmf->pte), vmf 2960 __wp_page_copy_user(struct page *dst, struct page *src, struct vm_fault *vmf) argument 3080 do_page_mkwrite(struct vm_fault *vmf, struct folio *folio) argument 3113 fault_dirty_shared_page(struct vm_fault *vmf) argument 3199 vmf_can_call_fault(const struct vm_fault *vmf) argument 3209 vmf_anon_prepare(struct vm_fault *vmf) argument 3241 wp_page_copy(struct vm_fault *vmf) argument 3414 finish_mkwrite_fault(struct vm_fault *vmf, struct folio *folio) argument 3438 wp_pfn_shared(struct vm_fault *vmf) argument 3794 remove_device_exclusive_entry(struct vm_fault *vmf) argument 3855 pte_marker_clear(struct vm_fault *vmf) argument 3875 do_pte_missing(struct vm_fault *vmf) argument 3887 pte_marker_handle_uffd_wp(struct vm_fault *vmf) argument 3899 handle_pte_marker(struct vm_fault *vmf) argument 3930 do_swap_page(struct vm_fault *vmf) argument 4306 alloc_anon_folio(struct vm_fault *vmf) argument 4383 do_anonymous_page(struct vm_fault *vmf) argument 4504 __do_fault(struct vm_fault *vmf) argument 4561 deposit_prealloc_pte(struct vm_fault *vmf) argument 4574 do_set_pmd(struct vm_fault *vmf, struct page *page) argument 4639 do_set_pmd(struct vm_fault *vmf, struct page *page) argument 4653 set_pte_range(struct vm_fault *vmf, struct folio *folio, struct page *page, unsigned int nr, unsigned long addr) argument 4690 vmf_pte_changed(struct vm_fault *vmf) argument 4713 finish_fault(struct vm_fault *vmf) argument 4828 do_fault_around(struct vm_fault *vmf) argument 4861 should_fault_around(struct vm_fault *vmf) argument 4874 do_read_fault(struct vm_fault *vmf) argument 4906 do_cow_fault(struct vm_fault *vmf) argument 4944 do_shared_fault(struct vm_fault *vmf) argument 4994 do_fault(struct vm_fault *vmf) argument 5055 do_numa_page(struct vm_fault *vmf) argument 5171 create_huge_pmd(struct vm_fault *vmf) argument 5182 wp_huge_pmd(struct vm_fault *vmf) argument 5213 create_huge_pud(struct vm_fault *vmf) argument 5227 wp_huge_pud(struct vm_fault *vmf, pud_t orig_pud) argument 5266 handle_pte_fault(struct vm_fault *vmf) argument 5353 struct vm_fault vmf = { local [all...] |
H A D | secretmem.c | 50 static vm_fault_t secretmem_fault(struct vm_fault *vmf) argument 52 struct address_space *mapping = vmf->vma->vm_file->f_mapping; 53 struct inode *inode = file_inode(vmf->vma->vm_file); 54 pgoff_t offset = vmf->pgoff; 55 gfp_t gfp = vmf->gfp_mask; 62 if (((loff_t)vmf->pgoff << PAGE_SHIFT) >= i_size_read(inode)) 105 vmf->page = page;
|
H A D | huge_memory.c | 868 static vm_fault_t __do_huge_pmd_anonymous_page(struct vm_fault *vmf, argument 871 struct vm_area_struct *vma = vmf->vma; 874 unsigned long haddr = vmf->address & HPAGE_PMD_MASK; 893 clear_huge_page(page, vmf->address, HPAGE_PMD_NR); 901 vmf->ptl = pmd_lock(vma->vm_mm, vmf->pmd); 902 if (unlikely(!pmd_none(*vmf->pmd))) { 913 spin_unlock(vmf->ptl); 916 ret = handle_userfault(vmf, VM_UFFD_MISSING); 925 pgtable_trans_huge_deposit(vma->vm_mm, vmf 996 do_huge_pmd_anonymous_page(struct vm_fault *vmf) argument 1114 vmf_insert_pfn_pmd(struct vm_fault *vmf, pfn_t pfn, bool write) argument 1203 vmf_insert_pfn_pud(struct vm_fault *vmf, pfn_t pfn, bool write) argument 1485 huge_pud_set_accessed(struct vm_fault *vmf, pud_t orig_pud) argument 1499 huge_pmd_set_accessed(struct vm_fault *vmf) argument 1513 do_huge_pmd_wp_page(struct vm_fault *vmf) argument 1713 do_huge_pmd_numa_page(struct vm_fault *vmf) argument [all...] |
H A D | filemap.c | 1670 vm_fault_t __folio_lock_or_retry(struct folio *folio, struct vm_fault *vmf) argument 1672 unsigned int flags = vmf->flags; 1682 release_fault_lock(vmf); 1694 release_fault_lock(vmf); 3040 * @vmf - the vm_fault for this fault. 3050 static int lock_folio_maybe_drop_mmap(struct vm_fault *vmf, struct folio *folio, argument 3061 if (vmf->flags & FAULT_FLAG_RETRY_NOWAIT) 3064 *fpin = maybe_unlock_mmap_for_io(vmf, *fpin); 3065 if (vmf->flags & FAULT_FLAG_KILLABLE) { 3075 release_fault_lock(vmf); 3091 do_sync_mmap_readahead(struct vm_fault *vmf) argument 3160 do_async_mmap_readahead(struct vm_fault *vmf, struct folio *folio) argument 3184 filemap_fault_recheck_pte_none(struct vm_fault *vmf) argument 3249 filemap_fault(struct vm_fault *vmf) argument 3408 filemap_map_pmd(struct vm_fault *vmf, struct folio *folio, pgoff_t start) argument 3481 filemap_map_folio_range(struct vm_fault *vmf, struct folio *folio, unsigned long start, unsigned long addr, unsigned int nr_pages, unsigned int *mmap_miss) argument 3534 filemap_map_order0_folio(struct vm_fault *vmf, struct folio *folio, unsigned long addr, unsigned int *mmap_miss) argument 3563 filemap_map_pages(struct vm_fault *vmf, pgoff_t start_pgoff, pgoff_t end_pgoff) argument 3627 filemap_page_mkwrite(struct vm_fault *vmf) argument 3682 filemap_page_mkwrite(struct vm_fault *vmf) argument [all...] |
/linux-master/drivers/dax/ |
H A D | device.c | 76 static void dax_set_mapping(struct vm_fault *vmf, pfn_t pfn, argument 80 struct file *filp = vmf->vma->vm_file; 88 pgoff = linear_page_index(vmf->vma, 89 ALIGN(vmf->address, fault_size)); 104 struct vm_fault *vmf) 111 if (check_vma(dev_dax, vmf->vma, __func__)) 123 phys = dax_pgoff_to_phys(dev_dax, vmf->pgoff, PAGE_SIZE); 125 dev_dbg(dev, "pgoff_to_phys(%#lx) failed\n", vmf->pgoff); 131 dax_set_mapping(vmf, pfn, fault_size); 133 return vmf_insert_mixed(vmf 103 __dev_dax_pte_fault(struct dev_dax *dev_dax, struct vm_fault *vmf) argument 136 __dev_dax_pmd_fault(struct dev_dax *dev_dax, struct vm_fault *vmf) argument 180 __dev_dax_pud_fault(struct dev_dax *dev_dax, struct vm_fault *vmf) argument 224 __dev_dax_pud_fault(struct dev_dax *dev_dax, struct vm_fault *vmf) argument 231 dev_dax_huge_fault(struct vm_fault *vmf, unsigned int order) argument 257 dev_dax_fault(struct vm_fault *vmf) argument [all...] |
/linux-master/include/trace/events/ |
H A D | fs_dax.h | 11 TP_PROTO(struct inode *inode, struct vm_fault *vmf, 13 TP_ARGS(inode, vmf, max_pgoff, result), 29 __entry->vm_start = vmf->vma->vm_start; 30 __entry->vm_end = vmf->vma->vm_end; 31 __entry->vm_flags = vmf->vma->vm_flags; 32 __entry->address = vmf->address; 33 __entry->flags = vmf->flags; 34 __entry->pgoff = vmf->pgoff; 56 TP_PROTO(struct inode *inode, struct vm_fault *vmf, \ 58 TP_ARGS(inode, vmf, max_pgof [all...] |
/linux-master/drivers/gpu/drm/ttm/ |
H A D | ttm_bo_vm.c | 42 struct vm_fault *vmf) 57 if (fault_flag_allow_retry_first(vmf->flags)) { 58 if (vmf->flags & FAULT_FLAG_RETRY_NOWAIT) 62 mmap_read_unlock(vmf->vma->vm_mm); 98 * @vmf: The fault structure handed to the callback 117 struct vm_fault *vmf) 131 if (fault_flag_allow_retry_first(vmf->flags)) { 132 if (!(vmf->flags & FAULT_FLAG_RETRY_NOWAIT)) { 134 mmap_read_unlock(vmf->vma->vm_mm); 165 * @vmf 41 ttm_bo_vm_fault_idle(struct ttm_buffer_object *bo, struct vm_fault *vmf) argument 116 ttm_bo_vm_reserve(struct ttm_buffer_object *bo, struct vm_fault *vmf) argument 181 ttm_bo_vm_fault_reserved(struct vm_fault *vmf, pgprot_t prot, pgoff_t num_prefault) argument 291 ttm_bo_vm_dummy_page(struct vm_fault *vmf, pgprot_t prot) argument 321 ttm_bo_vm_fault(struct vm_fault *vmf) argument [all...] |
/linux-master/fs/ocfs2/ |
H A D | mmap.c | 31 static vm_fault_t ocfs2_fault(struct vm_fault *vmf) argument 33 struct vm_area_struct *vma = vmf->vma; 38 ret = filemap_fault(vmf); 42 vma, vmf->page, vmf->pgoff); 113 static vm_fault_t ocfs2_page_mkwrite(struct vm_fault *vmf) argument 115 struct page *page = vmf->page; 116 struct inode *inode = file_inode(vmf->vma->vm_file); 144 ret = __ocfs2_page_mkwrite(vmf->vma->vm_file, di_bh, page);
|
/linux-master/fs/ |
H A D | dax.c | 830 static int copy_cow_page_dax(struct vm_fault *vmf, const struct iomap_iter *iter) argument 844 vto = kmap_atomic(vmf->cow_page); 845 copy_user_page(vto, kaddr, vmf->address, vmf->cow_page); 869 static void *dax_insert_entry(struct xa_state *xas, struct vm_fault *vmf, argument 873 struct address_space *mapping = vmf->vma->vm_file->f_mapping; 876 bool dirty = write && !dax_fault_is_synchronous(iter, vmf->vma); 898 dax_associate_entry(new_entry, mapping, vmf->vma, vmf->address, 1186 static vm_fault_t dax_load_hole(struct xa_state *xas, struct vm_fault *vmf, argument 1202 dax_pmd_load_hole(struct xa_state *xas, struct vm_fault *vmf, const struct iomap_iter *iter, void **entry) argument 1254 dax_pmd_load_hole(struct xa_state *xas, struct vm_fault *vmf, const struct iomap_iter *iter, void **entry) argument 1603 dax_fault_cow_page(struct vm_fault *vmf, const struct iomap_iter *iter) argument 1642 dax_fault_iter(struct vm_fault *vmf, const struct iomap_iter *iter, pfn_t *pfnp, struct xa_state *xas, void **entry, bool pmd) argument 1697 dax_iomap_pte_fault(struct vm_fault *vmf, pfn_t *pfnp, int *iomap_errp, const struct iomap_ops *ops) argument 1774 dax_fault_check_fallback(struct vm_fault *vmf, struct xa_state *xas, pgoff_t max_pgoff) argument 1807 dax_iomap_pmd_fault(struct vm_fault *vmf, pfn_t *pfnp, const struct iomap_ops *ops) argument 1887 dax_iomap_pmd_fault(struct vm_fault *vmf, pfn_t *pfnp, const struct iomap_ops *ops) argument 1907 dax_iomap_fault(struct vm_fault *vmf, unsigned int order, pfn_t *pfnp, int *iomap_errp, const struct iomap_ops *ops) argument 1929 dax_insert_pfn_mkwrite(struct vm_fault *vmf, pfn_t pfn, unsigned int order) argument 1973 dax_finish_sync_fault(struct vm_fault *vmf, unsigned int order, pfn_t pfn) argument [all...] |
/linux-master/drivers/video/fbdev/core/ |
H A D | fb_defio.c | 94 static vm_fault_t fb_deferred_io_fault(struct vm_fault *vmf) argument 98 struct fb_info *info = vmf->vma->vm_private_data; 100 offset = vmf->pgoff << PAGE_SHIFT; 110 if (vmf->vma->vm_file) 111 page->mapping = vmf->vma->vm_file->f_mapping; 116 page->index = vmf->pgoff; /* for page_mkclean() */ 118 vmf->page = page; 186 * @vmf: The VM fault 197 static vm_fault_t fb_deferred_io_page_mkwrite(struct fb_info *info, struct vm_fault *vmf) argument 199 unsigned long offset = vmf 208 fb_deferred_io_mkwrite(struct vm_fault *vmf) argument [all...] |
/linux-master/drivers/xen/ |
H A D | privcmd-buf.c | 117 static vm_fault_t privcmd_buf_vma_fault(struct vm_fault *vmf) argument 120 vmf->vma, vmf->vma->vm_start, vmf->vma->vm_end, 121 vmf->pgoff, (void *)vmf->address);
|
/linux-master/arch/x86/entry/vdso/ |
H A D | vma.c | 63 struct vm_area_struct *vma, struct vm_fault *vmf) 67 if (!image || (vmf->pgoff << PAGE_SHIFT) >= image->size) 70 vmf->page = virt_to_page(image->data + (vmf->pgoff << PAGE_SHIFT)); 71 get_page(vmf->page); 129 struct vm_area_struct *vma, struct vm_fault *vmf) 138 sym_offset = (long)(vmf->pgoff << PAGE_SHIFT) + 173 addr = vmf->address + (image->sym_timens_page - sym_offset); 181 return vmf_insert_pfn(vma, vmf->address, pfn); 186 return vmf_insert_pfn_prot(vma, vmf 62 vdso_fault(const struct vm_special_mapping *sm, struct vm_area_struct *vma, struct vm_fault *vmf) argument 128 vvar_fault(const struct vm_special_mapping *sm, struct vm_area_struct *vma, struct vm_fault *vmf) argument [all...] |
/linux-master/fs/nilfs2/ |
H A D | file.c | 45 static vm_fault_t nilfs_page_mkwrite(struct vm_fault *vmf) argument 47 struct vm_area_struct *vma = vmf->vma; 48 struct folio *folio = page_folio(vmf->page); 101 ret = block_page_mkwrite(vma, vmf, nilfs_get_block);
|
/linux-master/drivers/gpu/drm/vmwgfx/ |
H A D | vmwgfx_page_dirty.c | 375 vm_fault_t vmw_bo_vm_mkwrite(struct vm_fault *vmf) argument 377 struct vm_area_struct *vma = vmf->vma; 389 save_flags = vmf->flags; 390 vmf->flags &= ~FAULT_FLAG_ALLOW_RETRY; 391 ret = ttm_bo_vm_reserve(bo, vmf); 392 vmf->flags = save_flags; 396 page_offset = vmf->pgoff - drm_vma_node_start(&bo->base.vma_node); 416 vm_fault_t vmw_bo_vm_fault(struct vm_fault *vmf) argument 418 struct vm_area_struct *vma = vmf->vma; 426 ret = ttm_bo_vm_reserve(bo, vmf); [all...] |
/linux-master/fs/xfs/ |
H A D | xfs_file.c | 1326 struct vm_fault *vmf, 1331 return dax_iomap_fault(vmf, order, pfn, NULL, 1332 (write_fault && !vmf->cow_page) ? 1339 struct vm_fault *vmf, 1361 struct vm_fault *vmf, 1365 struct inode *inode = file_inode(vmf->vma->vm_file); 1374 file_update_time(vmf->vma->vm_file); 1383 ret = xfs_dax_fault(vmf, order, write_fault, &pfn); 1385 ret = dax_finish_sync_fault(vmf, order, pfn); 1387 ret = iomap_page_mkwrite(vmf, 1325 xfs_dax_fault( struct vm_fault *vmf, unsigned int order, bool write_fault, pfn_t *pfn) argument 1338 xfs_dax_fault( struct vm_fault *vmf, unsigned int order, bool write_fault, pfn_t *pfn) argument 1360 __xfs_filemap_fault( struct vm_fault *vmf, unsigned int order, bool write_fault) argument 1401 xfs_is_write_fault( struct vm_fault *vmf) argument 1409 xfs_filemap_fault( struct vm_fault *vmf) argument 1419 xfs_filemap_huge_fault( struct vm_fault *vmf, unsigned int order) argument 1432 xfs_filemap_page_mkwrite( struct vm_fault *vmf) argument 1444 xfs_filemap_pfn_mkwrite( struct vm_fault *vmf) argument [all...] |
/linux-master/arch/hexagon/include/asm/ |
H A D | cacheflush.h | 61 static inline void update_mmu_cache_range(struct vm_fault *vmf, argument
|
/linux-master/arch/microblaze/include/asm/ |
H A D | tlbflush.h | 36 #define update_mmu_cache_range(vmf, vma, addr, ptep, nr) do { } while (0)
|
/linux-master/arch/loongarch/kernel/ |
H A D | vdso.c | 49 struct vm_area_struct *vma, struct vm_fault *vmf) 54 switch (vmf->pgoff) { 76 pfn = sym_to_pfn(&loongarch_vdso_data) + vmf->pgoff - VVAR_LOONGARCH_PAGES_START; 82 return vmf_insert_pfn(vma, vmf->address, pfn); 48 vvar_fault(const struct vm_special_mapping *sm, struct vm_area_struct *vma, struct vm_fault *vmf) argument
|
/linux-master/sound/usb/usx2y/ |
H A D | usX2Yhwdep.c | 21 static vm_fault_t snd_us428ctls_vm_fault(struct vm_fault *vmf) argument 28 vmf->vma->vm_start, 29 vmf->pgoff); 31 offset = vmf->pgoff << PAGE_SHIFT; 32 vaddr = (char *)((struct usx2ydev *)vmf->vma->vm_private_data)->us428ctls_sharedmem + offset; 35 vmf->page = page;
|
/linux-master/include/linux/ |
H A D | huge_mm.h | 10 vm_fault_t do_huge_pmd_anonymous_page(struct vm_fault *vmf); 14 void huge_pmd_set_accessed(struct vm_fault *vmf); 20 void huge_pud_set_accessed(struct vm_fault *vmf, pud_t orig_pud); 22 static inline void huge_pud_set_accessed(struct vm_fault *vmf, pud_t orig_pud) argument 27 vm_fault_t do_huge_pmd_wp_page(struct vm_fault *vmf); 40 vm_fault_t vmf_insert_pfn_pmd(struct vm_fault *vmf, pfn_t pfn, bool write); 41 vm_fault_t vmf_insert_pfn_pud(struct vm_fault *vmf, pfn_t pfn, bool write); 350 vm_fault_t do_huge_pmd_numa_page(struct vm_fault *vmf); 481 static inline vm_fault_t do_huge_pmd_numa_page(struct vm_fault *vmf) argument
|
/linux-master/drivers/misc/ocxl/ |
H A D | sysfs.c | 109 static vm_fault_t global_mmio_fault(struct vm_fault *vmf) argument 111 struct vm_area_struct *vma = vmf->vma; 115 if (vmf->pgoff >= (afu->config.global_mmio_size >> PAGE_SHIFT)) 118 offset = vmf->pgoff; 120 return vmf_insert_pfn(vma, vmf->address, offset);
|
H A D | context.c | 139 static vm_fault_t ocxl_mmap_fault(struct vm_fault *vmf) argument 141 struct vm_area_struct *vma = vmf->vma; 146 offset = vmf->pgoff << PAGE_SHIFT; 148 ctx->pasid, vmf->address, offset); 151 ret = map_pp_mmio(vma, vmf->address, offset, ctx); 153 ret = map_afu_irq(vma, vmf->address, offset, ctx);
|
/linux-master/drivers/misc/cxl/ |
H A D | context.c | 126 static vm_fault_t cxl_mmap_fault(struct vm_fault *vmf) argument 128 struct vm_area_struct *vma = vmf->vma; 133 offset = vmf->pgoff << PAGE_SHIFT; 136 __func__, ctx->pe, vmf->address, offset); 161 vmf->page = ctx->ff_page; 168 ret = vmf_insert_pfn(vma, vmf->address, (area + offset) >> PAGE_SHIFT);
|
/linux-master/fs/ext2/ |
H A D | file.c | 93 static vm_fault_t ext2_dax_fault(struct vm_fault *vmf) argument 95 struct inode *inode = file_inode(vmf->vma->vm_file); 97 bool write = (vmf->flags & FAULT_FLAG_WRITE) && 98 (vmf->vma->vm_flags & VM_SHARED); 102 file_update_time(vmf->vma->vm_file); 106 ret = dax_iomap_fault(vmf, 0, NULL, NULL, &ext2_iomap_ops);
|
/linux-master/drivers/char/agp/ |
H A D | alpha-agp.c | 14 static vm_fault_t alpha_core_agp_vm_fault(struct vm_fault *vmf) argument 21 dma_addr = vmf->address - vmf->vma->vm_start + agp->aperture.bus_base; 32 vmf->page = page;
|