Searched refs:vmf (Results 1 - 25 of 168) sorted by relevance

1234567

/linux-master/mm/
H A Dmemory.c107 static vm_fault_t do_fault(struct vm_fault *vmf);
108 static vm_fault_t do_anonymous_page(struct vm_fault *vmf);
109 static bool vmf_pte_changed(struct vm_fault *vmf);
115 static bool vmf_orig_pte_uffd_wp(struct vm_fault *vmf) argument
117 if (!(vmf->flags & FAULT_FLAG_ORIG_PTE_VALID))
120 return pte_marker_uffd_wp(vmf->orig_pte);
2939 static inline int pte_unmap_same(struct vm_fault *vmf) argument
2944 spin_lock(vmf->ptl);
2945 same = pte_same(ptep_get(vmf->pte), vmf
2960 __wp_page_copy_user(struct page *dst, struct page *src, struct vm_fault *vmf) argument
3080 do_page_mkwrite(struct vm_fault *vmf, struct folio *folio) argument
3113 fault_dirty_shared_page(struct vm_fault *vmf) argument
3199 vmf_can_call_fault(const struct vm_fault *vmf) argument
3209 vmf_anon_prepare(struct vm_fault *vmf) argument
3241 wp_page_copy(struct vm_fault *vmf) argument
3414 finish_mkwrite_fault(struct vm_fault *vmf, struct folio *folio) argument
3438 wp_pfn_shared(struct vm_fault *vmf) argument
3794 remove_device_exclusive_entry(struct vm_fault *vmf) argument
3855 pte_marker_clear(struct vm_fault *vmf) argument
3875 do_pte_missing(struct vm_fault *vmf) argument
3887 pte_marker_handle_uffd_wp(struct vm_fault *vmf) argument
3899 handle_pte_marker(struct vm_fault *vmf) argument
3930 do_swap_page(struct vm_fault *vmf) argument
4306 alloc_anon_folio(struct vm_fault *vmf) argument
4383 do_anonymous_page(struct vm_fault *vmf) argument
4504 __do_fault(struct vm_fault *vmf) argument
4561 deposit_prealloc_pte(struct vm_fault *vmf) argument
4574 do_set_pmd(struct vm_fault *vmf, struct page *page) argument
4639 do_set_pmd(struct vm_fault *vmf, struct page *page) argument
4653 set_pte_range(struct vm_fault *vmf, struct folio *folio, struct page *page, unsigned int nr, unsigned long addr) argument
4690 vmf_pte_changed(struct vm_fault *vmf) argument
4713 finish_fault(struct vm_fault *vmf) argument
4828 do_fault_around(struct vm_fault *vmf) argument
4861 should_fault_around(struct vm_fault *vmf) argument
4874 do_read_fault(struct vm_fault *vmf) argument
4906 do_cow_fault(struct vm_fault *vmf) argument
4944 do_shared_fault(struct vm_fault *vmf) argument
4994 do_fault(struct vm_fault *vmf) argument
5055 do_numa_page(struct vm_fault *vmf) argument
5171 create_huge_pmd(struct vm_fault *vmf) argument
5182 wp_huge_pmd(struct vm_fault *vmf) argument
5213 create_huge_pud(struct vm_fault *vmf) argument
5227 wp_huge_pud(struct vm_fault *vmf, pud_t orig_pud) argument
5266 handle_pte_fault(struct vm_fault *vmf) argument
5353 struct vm_fault vmf = { local
[all...]
H A Dsecretmem.c50 static vm_fault_t secretmem_fault(struct vm_fault *vmf) argument
52 struct address_space *mapping = vmf->vma->vm_file->f_mapping;
53 struct inode *inode = file_inode(vmf->vma->vm_file);
54 pgoff_t offset = vmf->pgoff;
55 gfp_t gfp = vmf->gfp_mask;
62 if (((loff_t)vmf->pgoff << PAGE_SHIFT) >= i_size_read(inode))
105 vmf->page = page;
H A Dhuge_memory.c868 static vm_fault_t __do_huge_pmd_anonymous_page(struct vm_fault *vmf, argument
871 struct vm_area_struct *vma = vmf->vma;
874 unsigned long haddr = vmf->address & HPAGE_PMD_MASK;
893 clear_huge_page(page, vmf->address, HPAGE_PMD_NR);
901 vmf->ptl = pmd_lock(vma->vm_mm, vmf->pmd);
902 if (unlikely(!pmd_none(*vmf->pmd))) {
913 spin_unlock(vmf->ptl);
916 ret = handle_userfault(vmf, VM_UFFD_MISSING);
925 pgtable_trans_huge_deposit(vma->vm_mm, vmf
996 do_huge_pmd_anonymous_page(struct vm_fault *vmf) argument
1114 vmf_insert_pfn_pmd(struct vm_fault *vmf, pfn_t pfn, bool write) argument
1203 vmf_insert_pfn_pud(struct vm_fault *vmf, pfn_t pfn, bool write) argument
1485 huge_pud_set_accessed(struct vm_fault *vmf, pud_t orig_pud) argument
1499 huge_pmd_set_accessed(struct vm_fault *vmf) argument
1513 do_huge_pmd_wp_page(struct vm_fault *vmf) argument
1713 do_huge_pmd_numa_page(struct vm_fault *vmf) argument
[all...]
H A Dfilemap.c1670 vm_fault_t __folio_lock_or_retry(struct folio *folio, struct vm_fault *vmf) argument
1672 unsigned int flags = vmf->flags;
1682 release_fault_lock(vmf);
1694 release_fault_lock(vmf);
3040 * @vmf - the vm_fault for this fault.
3050 static int lock_folio_maybe_drop_mmap(struct vm_fault *vmf, struct folio *folio, argument
3061 if (vmf->flags & FAULT_FLAG_RETRY_NOWAIT)
3064 *fpin = maybe_unlock_mmap_for_io(vmf, *fpin);
3065 if (vmf->flags & FAULT_FLAG_KILLABLE) {
3075 release_fault_lock(vmf);
3091 do_sync_mmap_readahead(struct vm_fault *vmf) argument
3160 do_async_mmap_readahead(struct vm_fault *vmf, struct folio *folio) argument
3184 filemap_fault_recheck_pte_none(struct vm_fault *vmf) argument
3249 filemap_fault(struct vm_fault *vmf) argument
3408 filemap_map_pmd(struct vm_fault *vmf, struct folio *folio, pgoff_t start) argument
3481 filemap_map_folio_range(struct vm_fault *vmf, struct folio *folio, unsigned long start, unsigned long addr, unsigned int nr_pages, unsigned int *mmap_miss) argument
3534 filemap_map_order0_folio(struct vm_fault *vmf, struct folio *folio, unsigned long addr, unsigned int *mmap_miss) argument
3563 filemap_map_pages(struct vm_fault *vmf, pgoff_t start_pgoff, pgoff_t end_pgoff) argument
3627 filemap_page_mkwrite(struct vm_fault *vmf) argument
3682 filemap_page_mkwrite(struct vm_fault *vmf) argument
[all...]
/linux-master/drivers/dax/
H A Ddevice.c76 static void dax_set_mapping(struct vm_fault *vmf, pfn_t pfn, argument
80 struct file *filp = vmf->vma->vm_file;
88 pgoff = linear_page_index(vmf->vma,
89 ALIGN(vmf->address, fault_size));
104 struct vm_fault *vmf)
111 if (check_vma(dev_dax, vmf->vma, __func__))
123 phys = dax_pgoff_to_phys(dev_dax, vmf->pgoff, PAGE_SIZE);
125 dev_dbg(dev, "pgoff_to_phys(%#lx) failed\n", vmf->pgoff);
131 dax_set_mapping(vmf, pfn, fault_size);
133 return vmf_insert_mixed(vmf
103 __dev_dax_pte_fault(struct dev_dax *dev_dax, struct vm_fault *vmf) argument
136 __dev_dax_pmd_fault(struct dev_dax *dev_dax, struct vm_fault *vmf) argument
180 __dev_dax_pud_fault(struct dev_dax *dev_dax, struct vm_fault *vmf) argument
224 __dev_dax_pud_fault(struct dev_dax *dev_dax, struct vm_fault *vmf) argument
231 dev_dax_huge_fault(struct vm_fault *vmf, unsigned int order) argument
257 dev_dax_fault(struct vm_fault *vmf) argument
[all...]
/linux-master/include/trace/events/
H A Dfs_dax.h11 TP_PROTO(struct inode *inode, struct vm_fault *vmf,
13 TP_ARGS(inode, vmf, max_pgoff, result),
29 __entry->vm_start = vmf->vma->vm_start;
30 __entry->vm_end = vmf->vma->vm_end;
31 __entry->vm_flags = vmf->vma->vm_flags;
32 __entry->address = vmf->address;
33 __entry->flags = vmf->flags;
34 __entry->pgoff = vmf->pgoff;
56 TP_PROTO(struct inode *inode, struct vm_fault *vmf, \
58 TP_ARGS(inode, vmf, max_pgof
[all...]
/linux-master/drivers/gpu/drm/ttm/
H A Dttm_bo_vm.c42 struct vm_fault *vmf)
57 if (fault_flag_allow_retry_first(vmf->flags)) {
58 if (vmf->flags & FAULT_FLAG_RETRY_NOWAIT)
62 mmap_read_unlock(vmf->vma->vm_mm);
98 * @vmf: The fault structure handed to the callback
117 struct vm_fault *vmf)
131 if (fault_flag_allow_retry_first(vmf->flags)) {
132 if (!(vmf->flags & FAULT_FLAG_RETRY_NOWAIT)) {
134 mmap_read_unlock(vmf->vma->vm_mm);
165 * @vmf
41 ttm_bo_vm_fault_idle(struct ttm_buffer_object *bo, struct vm_fault *vmf) argument
116 ttm_bo_vm_reserve(struct ttm_buffer_object *bo, struct vm_fault *vmf) argument
181 ttm_bo_vm_fault_reserved(struct vm_fault *vmf, pgprot_t prot, pgoff_t num_prefault) argument
291 ttm_bo_vm_dummy_page(struct vm_fault *vmf, pgprot_t prot) argument
321 ttm_bo_vm_fault(struct vm_fault *vmf) argument
[all...]
/linux-master/fs/ocfs2/
H A Dmmap.c31 static vm_fault_t ocfs2_fault(struct vm_fault *vmf) argument
33 struct vm_area_struct *vma = vmf->vma;
38 ret = filemap_fault(vmf);
42 vma, vmf->page, vmf->pgoff);
113 static vm_fault_t ocfs2_page_mkwrite(struct vm_fault *vmf) argument
115 struct page *page = vmf->page;
116 struct inode *inode = file_inode(vmf->vma->vm_file);
144 ret = __ocfs2_page_mkwrite(vmf->vma->vm_file, di_bh, page);
/linux-master/fs/
H A Ddax.c830 static int copy_cow_page_dax(struct vm_fault *vmf, const struct iomap_iter *iter) argument
844 vto = kmap_atomic(vmf->cow_page);
845 copy_user_page(vto, kaddr, vmf->address, vmf->cow_page);
869 static void *dax_insert_entry(struct xa_state *xas, struct vm_fault *vmf, argument
873 struct address_space *mapping = vmf->vma->vm_file->f_mapping;
876 bool dirty = write && !dax_fault_is_synchronous(iter, vmf->vma);
898 dax_associate_entry(new_entry, mapping, vmf->vma, vmf->address,
1186 static vm_fault_t dax_load_hole(struct xa_state *xas, struct vm_fault *vmf, argument
1202 dax_pmd_load_hole(struct xa_state *xas, struct vm_fault *vmf, const struct iomap_iter *iter, void **entry) argument
1254 dax_pmd_load_hole(struct xa_state *xas, struct vm_fault *vmf, const struct iomap_iter *iter, void **entry) argument
1603 dax_fault_cow_page(struct vm_fault *vmf, const struct iomap_iter *iter) argument
1642 dax_fault_iter(struct vm_fault *vmf, const struct iomap_iter *iter, pfn_t *pfnp, struct xa_state *xas, void **entry, bool pmd) argument
1697 dax_iomap_pte_fault(struct vm_fault *vmf, pfn_t *pfnp, int *iomap_errp, const struct iomap_ops *ops) argument
1774 dax_fault_check_fallback(struct vm_fault *vmf, struct xa_state *xas, pgoff_t max_pgoff) argument
1807 dax_iomap_pmd_fault(struct vm_fault *vmf, pfn_t *pfnp, const struct iomap_ops *ops) argument
1887 dax_iomap_pmd_fault(struct vm_fault *vmf, pfn_t *pfnp, const struct iomap_ops *ops) argument
1907 dax_iomap_fault(struct vm_fault *vmf, unsigned int order, pfn_t *pfnp, int *iomap_errp, const struct iomap_ops *ops) argument
1929 dax_insert_pfn_mkwrite(struct vm_fault *vmf, pfn_t pfn, unsigned int order) argument
1973 dax_finish_sync_fault(struct vm_fault *vmf, unsigned int order, pfn_t pfn) argument
[all...]
/linux-master/drivers/video/fbdev/core/
H A Dfb_defio.c94 static vm_fault_t fb_deferred_io_fault(struct vm_fault *vmf) argument
98 struct fb_info *info = vmf->vma->vm_private_data;
100 offset = vmf->pgoff << PAGE_SHIFT;
110 if (vmf->vma->vm_file)
111 page->mapping = vmf->vma->vm_file->f_mapping;
116 page->index = vmf->pgoff; /* for page_mkclean() */
118 vmf->page = page;
186 * @vmf: The VM fault
197 static vm_fault_t fb_deferred_io_page_mkwrite(struct fb_info *info, struct vm_fault *vmf) argument
199 unsigned long offset = vmf
208 fb_deferred_io_mkwrite(struct vm_fault *vmf) argument
[all...]
/linux-master/drivers/xen/
H A Dprivcmd-buf.c117 static vm_fault_t privcmd_buf_vma_fault(struct vm_fault *vmf) argument
120 vmf->vma, vmf->vma->vm_start, vmf->vma->vm_end,
121 vmf->pgoff, (void *)vmf->address);
/linux-master/arch/x86/entry/vdso/
H A Dvma.c63 struct vm_area_struct *vma, struct vm_fault *vmf)
67 if (!image || (vmf->pgoff << PAGE_SHIFT) >= image->size)
70 vmf->page = virt_to_page(image->data + (vmf->pgoff << PAGE_SHIFT));
71 get_page(vmf->page);
129 struct vm_area_struct *vma, struct vm_fault *vmf)
138 sym_offset = (long)(vmf->pgoff << PAGE_SHIFT) +
173 addr = vmf->address + (image->sym_timens_page - sym_offset);
181 return vmf_insert_pfn(vma, vmf->address, pfn);
186 return vmf_insert_pfn_prot(vma, vmf
62 vdso_fault(const struct vm_special_mapping *sm, struct vm_area_struct *vma, struct vm_fault *vmf) argument
128 vvar_fault(const struct vm_special_mapping *sm, struct vm_area_struct *vma, struct vm_fault *vmf) argument
[all...]
/linux-master/fs/nilfs2/
H A Dfile.c45 static vm_fault_t nilfs_page_mkwrite(struct vm_fault *vmf) argument
47 struct vm_area_struct *vma = vmf->vma;
48 struct folio *folio = page_folio(vmf->page);
101 ret = block_page_mkwrite(vma, vmf, nilfs_get_block);
/linux-master/drivers/gpu/drm/vmwgfx/
H A Dvmwgfx_page_dirty.c375 vm_fault_t vmw_bo_vm_mkwrite(struct vm_fault *vmf) argument
377 struct vm_area_struct *vma = vmf->vma;
389 save_flags = vmf->flags;
390 vmf->flags &= ~FAULT_FLAG_ALLOW_RETRY;
391 ret = ttm_bo_vm_reserve(bo, vmf);
392 vmf->flags = save_flags;
396 page_offset = vmf->pgoff - drm_vma_node_start(&bo->base.vma_node);
416 vm_fault_t vmw_bo_vm_fault(struct vm_fault *vmf) argument
418 struct vm_area_struct *vma = vmf->vma;
426 ret = ttm_bo_vm_reserve(bo, vmf);
[all...]
/linux-master/fs/xfs/
H A Dxfs_file.c1326 struct vm_fault *vmf,
1331 return dax_iomap_fault(vmf, order, pfn, NULL,
1332 (write_fault && !vmf->cow_page) ?
1339 struct vm_fault *vmf,
1361 struct vm_fault *vmf,
1365 struct inode *inode = file_inode(vmf->vma->vm_file);
1374 file_update_time(vmf->vma->vm_file);
1383 ret = xfs_dax_fault(vmf, order, write_fault, &pfn);
1385 ret = dax_finish_sync_fault(vmf, order, pfn);
1387 ret = iomap_page_mkwrite(vmf,
1325 xfs_dax_fault( struct vm_fault *vmf, unsigned int order, bool write_fault, pfn_t *pfn) argument
1338 xfs_dax_fault( struct vm_fault *vmf, unsigned int order, bool write_fault, pfn_t *pfn) argument
1360 __xfs_filemap_fault( struct vm_fault *vmf, unsigned int order, bool write_fault) argument
1401 xfs_is_write_fault( struct vm_fault *vmf) argument
1409 xfs_filemap_fault( struct vm_fault *vmf) argument
1419 xfs_filemap_huge_fault( struct vm_fault *vmf, unsigned int order) argument
1432 xfs_filemap_page_mkwrite( struct vm_fault *vmf) argument
1444 xfs_filemap_pfn_mkwrite( struct vm_fault *vmf) argument
[all...]
/linux-master/arch/hexagon/include/asm/
H A Dcacheflush.h61 static inline void update_mmu_cache_range(struct vm_fault *vmf, argument
/linux-master/arch/microblaze/include/asm/
H A Dtlbflush.h36 #define update_mmu_cache_range(vmf, vma, addr, ptep, nr) do { } while (0)
/linux-master/arch/loongarch/kernel/
H A Dvdso.c49 struct vm_area_struct *vma, struct vm_fault *vmf)
54 switch (vmf->pgoff) {
76 pfn = sym_to_pfn(&loongarch_vdso_data) + vmf->pgoff - VVAR_LOONGARCH_PAGES_START;
82 return vmf_insert_pfn(vma, vmf->address, pfn);
48 vvar_fault(const struct vm_special_mapping *sm, struct vm_area_struct *vma, struct vm_fault *vmf) argument
/linux-master/sound/usb/usx2y/
H A DusX2Yhwdep.c21 static vm_fault_t snd_us428ctls_vm_fault(struct vm_fault *vmf) argument
28 vmf->vma->vm_start,
29 vmf->pgoff);
31 offset = vmf->pgoff << PAGE_SHIFT;
32 vaddr = (char *)((struct usx2ydev *)vmf->vma->vm_private_data)->us428ctls_sharedmem + offset;
35 vmf->page = page;
/linux-master/include/linux/
H A Dhuge_mm.h10 vm_fault_t do_huge_pmd_anonymous_page(struct vm_fault *vmf);
14 void huge_pmd_set_accessed(struct vm_fault *vmf);
20 void huge_pud_set_accessed(struct vm_fault *vmf, pud_t orig_pud);
22 static inline void huge_pud_set_accessed(struct vm_fault *vmf, pud_t orig_pud) argument
27 vm_fault_t do_huge_pmd_wp_page(struct vm_fault *vmf);
40 vm_fault_t vmf_insert_pfn_pmd(struct vm_fault *vmf, pfn_t pfn, bool write);
41 vm_fault_t vmf_insert_pfn_pud(struct vm_fault *vmf, pfn_t pfn, bool write);
350 vm_fault_t do_huge_pmd_numa_page(struct vm_fault *vmf);
481 static inline vm_fault_t do_huge_pmd_numa_page(struct vm_fault *vmf) argument
/linux-master/drivers/misc/ocxl/
H A Dsysfs.c109 static vm_fault_t global_mmio_fault(struct vm_fault *vmf) argument
111 struct vm_area_struct *vma = vmf->vma;
115 if (vmf->pgoff >= (afu->config.global_mmio_size >> PAGE_SHIFT))
118 offset = vmf->pgoff;
120 return vmf_insert_pfn(vma, vmf->address, offset);
H A Dcontext.c139 static vm_fault_t ocxl_mmap_fault(struct vm_fault *vmf) argument
141 struct vm_area_struct *vma = vmf->vma;
146 offset = vmf->pgoff << PAGE_SHIFT;
148 ctx->pasid, vmf->address, offset);
151 ret = map_pp_mmio(vma, vmf->address, offset, ctx);
153 ret = map_afu_irq(vma, vmf->address, offset, ctx);
/linux-master/drivers/misc/cxl/
H A Dcontext.c126 static vm_fault_t cxl_mmap_fault(struct vm_fault *vmf) argument
128 struct vm_area_struct *vma = vmf->vma;
133 offset = vmf->pgoff << PAGE_SHIFT;
136 __func__, ctx->pe, vmf->address, offset);
161 vmf->page = ctx->ff_page;
168 ret = vmf_insert_pfn(vma, vmf->address, (area + offset) >> PAGE_SHIFT);
/linux-master/fs/ext2/
H A Dfile.c93 static vm_fault_t ext2_dax_fault(struct vm_fault *vmf) argument
95 struct inode *inode = file_inode(vmf->vma->vm_file);
97 bool write = (vmf->flags & FAULT_FLAG_WRITE) &&
98 (vmf->vma->vm_flags & VM_SHARED);
102 file_update_time(vmf->vma->vm_file);
106 ret = dax_iomap_fault(vmf, 0, NULL, NULL, &ext2_iomap_ops);
/linux-master/drivers/char/agp/
H A Dalpha-agp.c14 static vm_fault_t alpha_core_agp_vm_fault(struct vm_fault *vmf) argument
21 dma_addr = vmf->address - vmf->vma->vm_start + agp->aperture.bus_base;
32 vmf->page = page;

Completed in 292 milliseconds

1234567