Searched refs:vmf (Results 1 - 25 of 170) sorted by relevance

1234567

/linux-master/mm/
H A Dmemory.c107 static vm_fault_t do_fault(struct vm_fault *vmf);
108 static vm_fault_t do_anonymous_page(struct vm_fault *vmf);
109 static bool vmf_pte_changed(struct vm_fault *vmf);
115 static __always_inline bool vmf_orig_pte_uffd_wp(struct vm_fault *vmf) argument
117 if (!userfaultfd_wp(vmf->vma))
119 if (!(vmf->flags & FAULT_FLAG_ORIG_PTE_VALID))
122 return pte_marker_uffd_wp(vmf->orig_pte);
2947 static inline int pte_unmap_same(struct vm_fault *vmf) argument
2952 spin_lock(vmf->ptl);
2953 same = pte_same(ptep_get(vmf
2968 __wp_page_copy_user(struct page *dst, struct page *src, struct vm_fault *vmf) argument
3088 do_page_mkwrite(struct vm_fault *vmf, struct folio *folio) argument
3121 fault_dirty_shared_page(struct vm_fault *vmf) argument
3207 vmf_can_call_fault(const struct vm_fault *vmf) argument
3232 vmf_anon_prepare(struct vm_fault *vmf) argument
3269 wp_page_copy(struct vm_fault *vmf) argument
3437 finish_mkwrite_fault(struct vm_fault *vmf, struct folio *folio) argument
3461 wp_pfn_shared(struct vm_fault *vmf) argument
3817 remove_device_exclusive_entry(struct vm_fault *vmf) argument
3878 pte_marker_clear(struct vm_fault *vmf) argument
3898 do_pte_missing(struct vm_fault *vmf) argument
3910 pte_marker_handle_uffd_wp(struct vm_fault *vmf) argument
3922 handle_pte_marker(struct vm_fault *vmf) argument
3953 do_swap_page(struct vm_fault *vmf) argument
4329 alloc_anon_folio(struct vm_fault *vmf) argument
4411 do_anonymous_page(struct vm_fault *vmf) argument
4535 __do_fault(struct vm_fault *vmf) argument
4592 deposit_prealloc_pte(struct vm_fault *vmf) argument
4605 do_set_pmd(struct vm_fault *vmf, struct page *page) argument
4670 do_set_pmd(struct vm_fault *vmf, struct page *page) argument
4684 set_pte_range(struct vm_fault *vmf, struct folio *folio, struct page *page, unsigned int nr, unsigned long addr) argument
4718 vmf_pte_changed(struct vm_fault *vmf) argument
4741 finish_fault(struct vm_fault *vmf) argument
4860 do_fault_around(struct vm_fault *vmf) argument
4893 should_fault_around(struct vm_fault *vmf) argument
4906 do_read_fault(struct vm_fault *vmf) argument
4938 do_cow_fault(struct vm_fault *vmf) argument
4976 do_shared_fault(struct vm_fault *vmf) argument
5026 do_fault(struct vm_fault *vmf) argument
5070 numa_migrate_prep(struct folio *folio, struct vm_fault *vmf, unsigned long addr, int page_nid, int *flags) argument
5089 numa_rebuild_single_mapping(struct vm_fault *vmf, struct vm_area_struct *vma, unsigned long fault_addr, pte_t *fault_pte, bool writable) argument
5104 numa_rebuild_large_mapping(struct vm_fault *vmf, struct vm_area_struct *vma, struct folio *folio, pte_t fault_pte, bool ignore_writable, bool pte_write_upgrade) argument
5137 do_numa_page(struct vm_fault *vmf) argument
5251 create_huge_pmd(struct vm_fault *vmf) argument
5262 wp_huge_pmd(struct vm_fault *vmf) argument
5293 create_huge_pud(struct vm_fault *vmf) argument
5307 wp_huge_pud(struct vm_fault *vmf, pud_t orig_pud) argument
5346 handle_pte_fault(struct vm_fault *vmf) argument
5433 struct vm_fault vmf = { local
[all...]
H A Dsecretmem.c50 static vm_fault_t secretmem_fault(struct vm_fault *vmf) argument
52 struct address_space *mapping = vmf->vma->vm_file->f_mapping;
53 struct inode *inode = file_inode(vmf->vma->vm_file);
54 pgoff_t offset = vmf->pgoff;
55 gfp_t gfp = vmf->gfp_mask;
62 if (((loff_t)vmf->pgoff << PAGE_SHIFT) >= i_size_read(inode))
105 vmf->page = page;
H A Dhuge_memory.c918 static vm_fault_t __do_huge_pmd_anonymous_page(struct vm_fault *vmf, argument
921 struct vm_area_struct *vma = vmf->vma;
924 unsigned long haddr = vmf->address & HPAGE_PMD_MASK;
945 clear_huge_page(page, vmf->address, HPAGE_PMD_NR);
953 vmf->ptl = pmd_lock(vma->vm_mm, vmf->pmd);
954 if (unlikely(!pmd_none(*vmf->pmd))) {
965 spin_unlock(vmf->ptl);
968 ret = handle_userfault(vmf, VM_UFFD_MISSING);
977 pgtable_trans_huge_deposit(vma->vm_mm, vmf
1049 do_huge_pmd_anonymous_page(struct vm_fault *vmf) argument
1171 vmf_insert_pfn_pmd(struct vm_fault *vmf, pfn_t pfn, bool write) argument
1260 vmf_insert_pfn_pud(struct vm_fault *vmf, pfn_t pfn, bool write) argument
1499 huge_pud_set_accessed(struct vm_fault *vmf, pud_t orig_pud) argument
1513 huge_pmd_set_accessed(struct vm_fault *vmf) argument
1527 do_huge_pmd_wp_page(struct vm_fault *vmf) argument
1645 do_huge_pmd_numa_page(struct vm_fault *vmf) argument
[all...]
H A Dhugetlb.c5273 static vm_fault_t hugetlb_vm_op_fault(struct vm_fault *vmf) argument
5921 struct vm_fault *vmf)
5923 struct vm_area_struct *vma = vmf->vma;
5925 const bool unshare = vmf->flags & FAULT_FLAG_UNSHARE;
5926 pte_t pte = huge_ptep_get(vmf->pte);
5954 set_huge_ptep_writable(vma, vmf->address, vmf->pte);
5980 set_huge_ptep_writable(vma, vmf->address, vmf->pte);
6007 spin_unlock(vmf
5920 hugetlb_wp(struct folio *pagecache_folio, struct vm_fault *vmf) argument
6165 hugetlb_handle_userfault(struct vm_fault *vmf, struct address_space *mapping, unsigned long reason) argument
6199 hugetlb_no_page(struct address_space *mapping, struct vm_fault *vmf) argument
6444 struct vm_fault vmf = { local
[all...]
H A Dfilemap.c1694 vm_fault_t __folio_lock_or_retry(struct folio *folio, struct vm_fault *vmf) argument
1696 unsigned int flags = vmf->flags;
1706 release_fault_lock(vmf);
1718 release_fault_lock(vmf);
3064 * @vmf - the vm_fault for this fault.
3074 static int lock_folio_maybe_drop_mmap(struct vm_fault *vmf, struct folio *folio, argument
3085 if (vmf->flags & FAULT_FLAG_RETRY_NOWAIT)
3088 *fpin = maybe_unlock_mmap_for_io(vmf, *fpin);
3089 if (vmf->flags & FAULT_FLAG_KILLABLE) {
3099 release_fault_lock(vmf);
3115 do_sync_mmap_readahead(struct vm_fault *vmf) argument
3184 do_async_mmap_readahead(struct vm_fault *vmf, struct folio *folio) argument
3208 filemap_fault_recheck_pte_none(struct vm_fault *vmf) argument
3273 filemap_fault(struct vm_fault *vmf) argument
3432 filemap_map_pmd(struct vm_fault *vmf, struct folio *folio, pgoff_t start) argument
3505 filemap_map_folio_range(struct vm_fault *vmf, struct folio *folio, unsigned long start, unsigned long addr, unsigned int nr_pages, unsigned long *rss, unsigned int *mmap_miss) argument
3568 filemap_map_order0_folio(struct vm_fault *vmf, struct folio *folio, unsigned long addr, unsigned long *rss, unsigned int *mmap_miss) argument
3600 filemap_map_pages(struct vm_fault *vmf, pgoff_t start_pgoff, pgoff_t end_pgoff) argument
3668 filemap_page_mkwrite(struct vm_fault *vmf) argument
3723 filemap_page_mkwrite(struct vm_fault *vmf) argument
[all...]
/linux-master/drivers/dax/
H A Ddevice.c76 static void dax_set_mapping(struct vm_fault *vmf, pfn_t pfn, argument
80 struct file *filp = vmf->vma->vm_file;
88 pgoff = linear_page_index(vmf->vma,
89 ALIGN(vmf->address, fault_size));
104 struct vm_fault *vmf)
111 if (check_vma(dev_dax, vmf->vma, __func__))
123 phys = dax_pgoff_to_phys(dev_dax, vmf->pgoff, PAGE_SIZE);
125 dev_dbg(dev, "pgoff_to_phys(%#lx) failed\n", vmf->pgoff);
131 dax_set_mapping(vmf, pfn, fault_size);
133 return vmf_insert_mixed(vmf
103 __dev_dax_pte_fault(struct dev_dax *dev_dax, struct vm_fault *vmf) argument
136 __dev_dax_pmd_fault(struct dev_dax *dev_dax, struct vm_fault *vmf) argument
180 __dev_dax_pud_fault(struct dev_dax *dev_dax, struct vm_fault *vmf) argument
224 __dev_dax_pud_fault(struct dev_dax *dev_dax, struct vm_fault *vmf) argument
231 dev_dax_huge_fault(struct vm_fault *vmf, unsigned int order) argument
257 dev_dax_fault(struct vm_fault *vmf) argument
[all...]
/linux-master/include/trace/events/
H A Dfs_dax.h11 TP_PROTO(struct inode *inode, struct vm_fault *vmf,
13 TP_ARGS(inode, vmf, max_pgoff, result),
29 __entry->vm_start = vmf->vma->vm_start;
30 __entry->vm_end = vmf->vma->vm_end;
31 __entry->vm_flags = vmf->vma->vm_flags;
32 __entry->address = vmf->address;
33 __entry->flags = vmf->flags;
34 __entry->pgoff = vmf->pgoff;
56 TP_PROTO(struct inode *inode, struct vm_fault *vmf, \
58 TP_ARGS(inode, vmf, max_pgof
[all...]
/linux-master/drivers/gpu/drm/ttm/
H A Dttm_bo_vm.c42 struct vm_fault *vmf)
57 if (fault_flag_allow_retry_first(vmf->flags)) {
58 if (vmf->flags & FAULT_FLAG_RETRY_NOWAIT)
62 mmap_read_unlock(vmf->vma->vm_mm);
98 * @vmf: The fault structure handed to the callback
117 struct vm_fault *vmf)
131 if (fault_flag_allow_retry_first(vmf->flags)) {
132 if (!(vmf->flags & FAULT_FLAG_RETRY_NOWAIT)) {
134 mmap_read_unlock(vmf->vma->vm_mm);
165 * @vmf
41 ttm_bo_vm_fault_idle(struct ttm_buffer_object *bo, struct vm_fault *vmf) argument
116 ttm_bo_vm_reserve(struct ttm_buffer_object *bo, struct vm_fault *vmf) argument
181 ttm_bo_vm_fault_reserved(struct vm_fault *vmf, pgprot_t prot, pgoff_t num_prefault) argument
291 ttm_bo_vm_dummy_page(struct vm_fault *vmf, pgprot_t prot) argument
321 ttm_bo_vm_fault(struct vm_fault *vmf) argument
[all...]
/linux-master/fs/ocfs2/
H A Dmmap.c31 static vm_fault_t ocfs2_fault(struct vm_fault *vmf) argument
33 struct vm_area_struct *vma = vmf->vma;
38 ret = filemap_fault(vmf);
42 vma, vmf->page, vmf->pgoff);
113 static vm_fault_t ocfs2_page_mkwrite(struct vm_fault *vmf) argument
115 struct page *page = vmf->page;
116 struct inode *inode = file_inode(vmf->vma->vm_file);
144 ret = __ocfs2_page_mkwrite(vmf->vma->vm_file, di_bh, page);
/linux-master/fs/
H A Ddax.c830 static int copy_cow_page_dax(struct vm_fault *vmf, const struct iomap_iter *iter) argument
844 vto = kmap_atomic(vmf->cow_page);
845 copy_user_page(vto, kaddr, vmf->address, vmf->cow_page);
869 static void *dax_insert_entry(struct xa_state *xas, struct vm_fault *vmf, argument
873 struct address_space *mapping = vmf->vma->vm_file->f_mapping;
876 bool dirty = write && !dax_fault_is_synchronous(iter, vmf->vma);
898 dax_associate_entry(new_entry, mapping, vmf->vma, vmf->address,
1186 static vm_fault_t dax_load_hole(struct xa_state *xas, struct vm_fault *vmf, argument
1202 dax_pmd_load_hole(struct xa_state *xas, struct vm_fault *vmf, const struct iomap_iter *iter, void **entry) argument
1254 dax_pmd_load_hole(struct xa_state *xas, struct vm_fault *vmf, const struct iomap_iter *iter, void **entry) argument
1603 dax_fault_cow_page(struct vm_fault *vmf, const struct iomap_iter *iter) argument
1642 dax_fault_iter(struct vm_fault *vmf, const struct iomap_iter *iter, pfn_t *pfnp, struct xa_state *xas, void **entry, bool pmd) argument
1697 dax_iomap_pte_fault(struct vm_fault *vmf, pfn_t *pfnp, int *iomap_errp, const struct iomap_ops *ops) argument
1774 dax_fault_check_fallback(struct vm_fault *vmf, struct xa_state *xas, pgoff_t max_pgoff) argument
1807 dax_iomap_pmd_fault(struct vm_fault *vmf, pfn_t *pfnp, const struct iomap_ops *ops) argument
1887 dax_iomap_pmd_fault(struct vm_fault *vmf, pfn_t *pfnp, const struct iomap_ops *ops) argument
1907 dax_iomap_fault(struct vm_fault *vmf, unsigned int order, pfn_t *pfnp, int *iomap_errp, const struct iomap_ops *ops) argument
1929 dax_insert_pfn_mkwrite(struct vm_fault *vmf, pfn_t pfn, unsigned int order) argument
1973 dax_finish_sync_fault(struct vm_fault *vmf, unsigned int order, pfn_t pfn) argument
[all...]
/linux-master/drivers/video/fbdev/core/
H A Dfb_defio.c94 static vm_fault_t fb_deferred_io_fault(struct vm_fault *vmf) argument
98 struct fb_info *info = vmf->vma->vm_private_data;
100 offset = vmf->pgoff << PAGE_SHIFT;
110 if (vmf->vma->vm_file)
111 page->mapping = vmf->vma->vm_file->f_mapping;
116 page->index = vmf->pgoff; /* for page_mkclean() */
118 vmf->page = page;
186 * @vmf: The VM fault
197 static vm_fault_t fb_deferred_io_page_mkwrite(struct fb_info *info, struct vm_fault *vmf) argument
199 unsigned long offset = vmf
208 fb_deferred_io_mkwrite(struct vm_fault *vmf) argument
[all...]
/linux-master/drivers/xen/
H A Dprivcmd-buf.c117 static vm_fault_t privcmd_buf_vma_fault(struct vm_fault *vmf) argument
120 vmf->vma, vmf->vma->vm_start, vmf->vma->vm_end,
121 vmf->pgoff, (void *)vmf->address);
/linux-master/arch/x86/entry/vdso/
H A Dvma.c63 struct vm_area_struct *vma, struct vm_fault *vmf)
67 if (!image || (vmf->pgoff << PAGE_SHIFT) >= image->size)
70 vmf->page = virt_to_page(image->data + (vmf->pgoff << PAGE_SHIFT));
71 get_page(vmf->page);
129 struct vm_area_struct *vma, struct vm_fault *vmf)
138 sym_offset = (long)(vmf->pgoff << PAGE_SHIFT) +
173 addr = vmf->address + (image->sym_timens_page - sym_offset);
181 return vmf_insert_pfn(vma, vmf->address, pfn);
186 return vmf_insert_pfn_prot(vma, vmf
62 vdso_fault(const struct vm_special_mapping *sm, struct vm_area_struct *vma, struct vm_fault *vmf) argument
128 vvar_fault(const struct vm_special_mapping *sm, struct vm_area_struct *vma, struct vm_fault *vmf) argument
[all...]
/linux-master/fs/nilfs2/
H A Dfile.c45 static vm_fault_t nilfs_page_mkwrite(struct vm_fault *vmf) argument
47 struct vm_area_struct *vma = vmf->vma;
48 struct folio *folio = page_folio(vmf->page);
101 ret = block_page_mkwrite(vma, vmf, nilfs_get_block);
/linux-master/drivers/gpu/drm/vmwgfx/
H A Dvmwgfx_page_dirty.c375 vm_fault_t vmw_bo_vm_mkwrite(struct vm_fault *vmf) argument
377 struct vm_area_struct *vma = vmf->vma;
389 save_flags = vmf->flags;
390 vmf->flags &= ~FAULT_FLAG_ALLOW_RETRY;
391 ret = ttm_bo_vm_reserve(bo, vmf);
392 vmf->flags = save_flags;
396 page_offset = vmf->pgoff - drm_vma_node_start(&bo->base.vma_node);
416 vm_fault_t vmw_bo_vm_fault(struct vm_fault *vmf) argument
418 struct vm_area_struct *vma = vmf->vma;
426 ret = ttm_bo_vm_reserve(bo, vmf);
[all...]
/linux-master/fs/xfs/
H A Dxfs_file.c1253 struct vm_fault *vmf,
1258 return dax_iomap_fault(vmf, order, pfn, NULL,
1259 (write_fault && !vmf->cow_page) ?
1266 struct vm_fault *vmf,
1288 struct vm_fault *vmf,
1292 struct inode *inode = file_inode(vmf->vma->vm_file);
1301 file_update_time(vmf->vma->vm_file);
1310 ret = xfs_dax_fault(vmf, order, write_fault, &pfn);
1312 ret = dax_finish_sync_fault(vmf, order, pfn);
1314 ret = iomap_page_mkwrite(vmf,
1252 xfs_dax_fault( struct vm_fault *vmf, unsigned int order, bool write_fault, pfn_t *pfn) argument
1265 xfs_dax_fault( struct vm_fault *vmf, unsigned int order, bool write_fault, pfn_t *pfn) argument
1287 __xfs_filemap_fault( struct vm_fault *vmf, unsigned int order, bool write_fault) argument
1328 xfs_is_write_fault( struct vm_fault *vmf) argument
1336 xfs_filemap_fault( struct vm_fault *vmf) argument
1346 xfs_filemap_huge_fault( struct vm_fault *vmf, unsigned int order) argument
1359 xfs_filemap_page_mkwrite( struct vm_fault *vmf) argument
1371 xfs_filemap_pfn_mkwrite( struct vm_fault *vmf) argument
[all...]
/linux-master/arch/hexagon/include/asm/
H A Dcacheflush.h61 static inline void update_mmu_cache_range(struct vm_fault *vmf, argument
/linux-master/arch/microblaze/include/asm/
H A Dtlbflush.h36 #define update_mmu_cache_range(vmf, vma, addr, ptep, nr) do { } while (0)
/linux-master/arch/loongarch/kernel/
H A Dvdso.c49 struct vm_area_struct *vma, struct vm_fault *vmf)
54 switch (vmf->pgoff) {
76 pfn = sym_to_pfn(&loongarch_vdso_data) + vmf->pgoff - VVAR_LOONGARCH_PAGES_START;
82 return vmf_insert_pfn(vma, vmf->address, pfn);
48 vvar_fault(const struct vm_special_mapping *sm, struct vm_area_struct *vma, struct vm_fault *vmf) argument
/linux-master/sound/usb/usx2y/
H A DusX2Yhwdep.c21 static vm_fault_t snd_us428ctls_vm_fault(struct vm_fault *vmf) argument
28 vmf->vma->vm_start,
29 vmf->pgoff);
31 offset = vmf->pgoff << PAGE_SHIFT;
32 vaddr = (char *)((struct usx2ydev *)vmf->vma->vm_private_data)->us428ctls_sharedmem + offset;
35 vmf->page = page;
/linux-master/include/linux/
H A Dhuge_mm.h10 vm_fault_t do_huge_pmd_anonymous_page(struct vm_fault *vmf);
14 void huge_pmd_set_accessed(struct vm_fault *vmf);
20 void huge_pud_set_accessed(struct vm_fault *vmf, pud_t orig_pud);
22 static inline void huge_pud_set_accessed(struct vm_fault *vmf, pud_t orig_pud) argument
27 vm_fault_t do_huge_pmd_wp_page(struct vm_fault *vmf);
40 vm_fault_t vmf_insert_pfn_pmd(struct vm_fault *vmf, pfn_t pfn, bool write);
41 vm_fault_t vmf_insert_pfn_pud(struct vm_fault *vmf, pfn_t pfn, bool write);
382 vm_fault_t do_huge_pmd_numa_page(struct vm_fault *vmf);
512 static inline vm_fault_t do_huge_pmd_numa_page(struct vm_fault *vmf) argument
/linux-master/drivers/misc/ocxl/
H A Dsysfs.c109 static vm_fault_t global_mmio_fault(struct vm_fault *vmf) argument
111 struct vm_area_struct *vma = vmf->vma;
115 if (vmf->pgoff >= (afu->config.global_mmio_size >> PAGE_SHIFT))
118 offset = vmf->pgoff;
120 return vmf_insert_pfn(vma, vmf->address, offset);
H A Dcontext.c139 static vm_fault_t ocxl_mmap_fault(struct vm_fault *vmf) argument
141 struct vm_area_struct *vma = vmf->vma;
146 offset = vmf->pgoff << PAGE_SHIFT;
148 ctx->pasid, vmf->address, offset);
151 ret = map_pp_mmio(vma, vmf->address, offset, ctx);
153 ret = map_afu_irq(vma, vmf->address, offset, ctx);
/linux-master/drivers/misc/cxl/
H A Dcontext.c126 static vm_fault_t cxl_mmap_fault(struct vm_fault *vmf) argument
128 struct vm_area_struct *vma = vmf->vma;
133 offset = vmf->pgoff << PAGE_SHIFT;
136 __func__, ctx->pe, vmf->address, offset);
161 vmf->page = ctx->ff_page;
168 ret = vmf_insert_pfn(vma, vmf->address, (area + offset) >> PAGE_SHIFT);
/linux-master/fs/ext2/
H A Dfile.c93 static vm_fault_t ext2_dax_fault(struct vm_fault *vmf) argument
95 struct inode *inode = file_inode(vmf->vma->vm_file);
97 bool write = (vmf->flags & FAULT_FLAG_WRITE) &&
98 (vmf->vma->vm_flags & VM_SHARED);
102 file_update_time(vmf->vma->vm_file);
106 ret = dax_iomap_fault(vmf, 0, NULL, NULL, &ext2_iomap_ops);

Completed in 251 milliseconds

1234567