/linux-master/arch/sh/include/asm/ |
H A D | pgtable.h | 105 static inline void update_mmu_cache_range(struct vm_fault *vmf, argument
|
/linux-master/include/linux/ |
H A D | memremap.h | 90 vm_fault_t (*migrate_to_ram)(struct vm_fault *vmf);
|
H A D | mm.h | 589 vm_fault_t (*fault)(struct vm_fault *vmf); 590 vm_fault_t (*huge_fault)(struct vm_fault *vmf, unsigned int order); 591 vm_fault_t (*map_pages)(struct vm_fault *vmf, 597 vm_fault_t (*page_mkwrite)(struct vm_fault *vmf); 600 vm_fault_t (*pfn_mkwrite)(struct vm_fault *vmf); 763 static inline void release_fault_lock(struct vm_fault *vmf) argument 765 if (vmf->flags & FAULT_FLAG_VMA_LOCK) 766 vma_end_read(vmf->vma); 768 mmap_read_unlock(vmf->vma->vm_mm); 771 static inline void assert_fault_locked(struct vm_fault *vmf) argument 804 release_fault_lock(struct vm_fault *vmf) argument 809 assert_fault_locked(struct vm_fault *vmf) argument [all...] |
H A D | dax.h | 245 vm_fault_t dax_iomap_fault(struct vm_fault *vmf, unsigned int order, 247 vm_fault_t dax_finish_sync_fault(struct vm_fault *vmf,
|
H A D | userfaultfd_k.h | 85 extern vm_fault_t handle_userfault(struct vm_fault *vmf, unsigned long reason); 270 static inline vm_fault_t handle_userfault(struct vm_fault *vmf, argument
|
H A D | mempolicy.h | 170 int mpol_misplaced(struct folio *folio, struct vm_fault *vmf, 286 struct vm_fault *vmf, 285 mpol_misplaced(struct folio *folio, struct vm_fault *vmf, unsigned long address) argument
|
/linux-master/arch/arc/include/asm/ |
H A D | pgtable-bits-arcv2.h | 104 void update_mmu_cache_range(struct vm_fault *vmf, struct vm_area_struct *vma,
|
/linux-master/fs/erofs/ |
H A D | data.c | 411 static vm_fault_t erofs_dax_huge_fault(struct vm_fault *vmf, argument 414 return dax_iomap_fault(vmf, order, NULL, NULL, &erofs_iomap_ops); 417 static vm_fault_t erofs_dax_fault(struct vm_fault *vmf) argument 419 return erofs_dax_huge_fault(vmf, 0);
|
/linux-master/drivers/infiniband/hw/hfi1/ |
H A D | file_ops.c | 73 static vm_fault_t vma_fault(struct vm_fault *vmf); 307 static inline void mmap_cdbg(u16 ctxt, u8 subctxt, u8 type, u8 mapio, u8 vmf, argument 313 ctxt, subctxt, type, mapio, vmf, !!memdma, 328 u8 subctxt, mapio = 0, vmf = 0, type; local 441 mmap_cdbg(ctxt, subctxt, type, mapio, vmf, memaddr, 487 vmf = 1; 520 vmf = 1; 526 vmf = 1; 533 vmf = 1; 545 vmf 595 vma_fault(struct vm_fault *vmf) argument [all...] |
/linux-master/samples/vfio-mdev/ |
H A D | mbochs.c | 777 static vm_fault_t mbochs_region_vm_fault(struct vm_fault *vmf) argument 779 struct vm_area_struct *vma = vmf->vma; 781 pgoff_t page_offset = (vmf->address - vma->vm_start) >> PAGE_SHIFT; 786 vmf->page = mbochs_get_page(mdev_state, page_offset); 787 if (!vmf->page) 816 static vm_fault_t mbochs_dmabuf_vm_fault(struct vm_fault *vmf) argument 818 struct vm_area_struct *vma = vmf->vma; 821 if (WARN_ON(vmf->pgoff >= dmabuf->pagecount)) 824 vmf->page = dmabuf->pages[vmf [all...] |
/linux-master/arch/powerpc/kvm/ |
H A D | book3s_xive_native.c | 228 static vm_fault_t xive_native_esb_fault(struct vm_fault *vmf) argument 230 struct vm_area_struct *vma = vmf->vma; 246 page_offset = vmf->pgoff - vma->vm_pgoff; 280 vmf_insert_pfn(vma, vmf->address, page >> PAGE_SHIFT); 288 static vm_fault_t xive_native_tima_fault(struct vm_fault *vmf) argument 290 struct vm_area_struct *vma = vmf->vma; 292 switch (vmf->pgoff - vma->vm_pgoff) { 297 vmf_insert_pfn(vma, vmf->address, xive_tima_os >> PAGE_SHIFT);
|
/linux-master/drivers/gpu/drm/i915/gem/ |
H A D | i915_gem_ttm.c | 1034 static vm_fault_t vm_fault_ttm(struct vm_fault *vmf) argument 1036 struct vm_area_struct *area = vmf->vma; 1049 ret = ttm_bo_vm_reserve(bo, vmf); 1110 ret = ttm_bo_vm_fault_reserved(vmf, vmf->vma->vm_page_prot, 1114 ret = ttm_bo_vm_dummy_page(vmf, vmf->vma->vm_page_prot); 1117 if (ret == VM_FAULT_RETRY && !(vmf->flags & FAULT_FLAG_RETRY_NOWAIT))
|
/linux-master/drivers/gpu/drm/amd/amdgpu/ |
H A D | amdgpu_gem.c | 48 static vm_fault_t amdgpu_gem_fault(struct vm_fault *vmf) argument 50 struct ttm_buffer_object *bo = vmf->vma->vm_private_data; 55 ret = ttm_bo_vm_reserve(bo, vmf); 66 ret = ttm_bo_vm_fault_reserved(vmf, vmf->vma->vm_page_prot, 71 ret = ttm_bo_vm_dummy_page(vmf, vmf->vma->vm_page_prot); 73 if (ret == VM_FAULT_RETRY && !(vmf->flags & FAULT_FLAG_RETRY_NOWAIT))
|
/linux-master/drivers/gpu/drm/nouveau/ |
H A D | nouveau_dmem.c | 161 static vm_fault_t nouveau_dmem_migrate_to_ram(struct vm_fault *vmf) argument 163 struct nouveau_drm *drm = page_to_drm(vmf->page); 172 .vma = vmf->vma, 173 .start = vmf->address, 174 .end = vmf->address + PAGE_SIZE, 178 .fault_page = vmf->page, 196 dpage = alloc_page_vma(GFP_HIGHUSER, vmf->vma, vmf->address);
|
/linux-master/drivers/gpu/drm/etnaviv/ |
H A D | etnaviv_gem.c | 162 static vm_fault_t etnaviv_gem_fault(struct vm_fault *vmf) argument 164 struct vm_area_struct *vma = vmf->vma; 189 /* We don't use vmf->pgoff since that has the fake offset: */ 190 pgoff = (vmf->address - vma->vm_start) >> PAGE_SHIFT; 194 VERB("Inserting %p pfn %lx, pa %lx", (void *)vmf->address, 197 return vmf_insert_pfn(vma, vmf->address, pfn);
|
/linux-master/drivers/infiniband/core/ |
H A D | uverbs_main.c | 778 static vm_fault_t rdma_umap_fault(struct vm_fault *vmf) argument 780 struct ib_uverbs_file *ufile = vmf->vma->vm_file->private_data; 781 struct rdma_umap_priv *priv = vmf->vma->vm_private_data; 788 if (!(vmf->vma->vm_flags & (VM_WRITE | VM_MAYWRITE))) { 789 vmf->page = ZERO_PAGE(vmf->address); 790 get_page(vmf->page); 797 alloc_pages(vmf->gfp_mask | __GFP_ZERO, 0); 804 vmf->page = ufile->disassociate_page; 805 get_page(vmf [all...] |
/linux-master/mm/ |
H A D | internal.h | 296 vm_fault_t vmf_anon_prepare(struct vm_fault *vmf); 297 vm_fault_t do_swap_page(struct vm_fault *vmf); 941 static inline struct file *maybe_unlock_mmap_for_io(struct vm_fault *vmf, 944 int flags = vmf->flags; 956 fpin = get_file(vmf->vma->vm_file); 957 release_fault_lock(vmf); 1177 int numa_migrate_prep(struct folio *folio, struct vm_fault *vmf,
|
/linux-master/fs/orangefs/ |
H A D | file.c | 376 static vm_fault_t orangefs_fault(struct vm_fault *vmf) argument 378 struct file *file = vmf->vma->vm_file; 389 return filemap_fault(vmf);
|
/linux-master/arch/powerpc/platforms/book3s/ |
H A D | vas-api.c | 395 static vm_fault_t vas_mmap_fault(struct vm_fault *vmf) argument 397 struct vm_area_struct *vma = vmf->vma; 423 if (txwin->task_ref.vma != vmf->vma) {
|
/linux-master/fs/netfs/ |
H A D | buffered_write.c | 521 vm_fault_t netfs_page_mkwrite(struct vm_fault *vmf, struct netfs_group *netfs_group) argument 524 struct folio *folio = page_folio(vmf->page); 525 struct file *file = vmf->vma->vm_file;
|
/linux-master/sound/usb/usx2y/ |
H A D | us122l.c | 103 static vm_fault_t usb_stream_hwdep_vm_fault(struct vm_fault *vmf) argument 108 struct us122l *us122l = vmf->vma->vm_private_data; 116 offset = vmf->pgoff << PAGE_SHIFT; 131 vmf->page = page;
|
/linux-master/fs/zonefs/ |
H A D | file.c | 282 static vm_fault_t zonefs_filemap_page_mkwrite(struct vm_fault *vmf) argument 284 struct inode *inode = file_inode(vmf->vma->vm_file); 298 file_update_time(vmf->vma->vm_file); 302 ret = iomap_page_mkwrite(vmf, &zonefs_write_iomap_ops);
|
/linux-master/arch/powerpc/include/asm/ |
H A D | pgtable.h | 143 static inline void update_mmu_cache_range(struct vm_fault *vmf, argument
|
/linux-master/drivers/vdpa/vdpa_user/ |
H A D | iova_domain.c | 506 static vm_fault_t vduse_domain_mmap_fault(struct vm_fault *vmf) argument 508 struct vduse_iova_domain *domain = vmf->vma->vm_private_data; 509 unsigned long iova = vmf->pgoff << PAGE_SHIFT; 523 vmf->page = page;
|
/linux-master/fs/afs/ |
H A D | file.c | 30 static vm_fault_t afs_vm_map_pages(struct vm_fault *vmf, pgoff_t start_pgoff, pgoff_t end_pgoff); 466 static vm_fault_t afs_vm_map_pages(struct vm_fault *vmf, pgoff_t start_pgoff, pgoff_t end_pgoff) argument 468 struct afs_vnode *vnode = AFS_FS_I(file_inode(vmf->vma->vm_file)); 471 return filemap_map_pages(vmf, start_pgoff, end_pgoff);
|