Lines Matching refs:vmf
44 struct vm_fault *vmf)
59 if (fault_flag_allow_retry_first(vmf->flags)) {
60 if (vmf->flags & FAULT_FLAG_RETRY_NOWAIT)
64 mmap_read_unlock(vmf->vma->vm_mm);
100 * @vmf: The fault structure handed to the callback
119 struct vm_fault *vmf)
133 if (fault_flag_allow_retry_first(vmf->flags)) {
134 if (!(vmf->flags & FAULT_FLAG_RETRY_NOWAIT)) {
136 mmap_read_unlock(vmf->vma->vm_mm);
167 * @vmf: The struct vm_fault given as argument to the fault callback
183 vm_fault_t ttm_bo_vm_fault_reserved(struct vm_fault *vmf,
187 struct vm_area_struct *vma = vmf->vma;
198 unsigned long address = vmf->address;
204 ret = ttm_bo_vm_fault_idle(bo, vmf);
293 vm_fault_t ttm_bo_vm_dummy_page(struct vm_fault *vmf, pgprot_t prot)
295 struct vm_area_struct *vma = vmf->vma;
323 vm_fault_t ttm_bo_vm_fault(struct vm_fault *vmf)
325 struct vm_area_struct *vma = vmf->vma;
332 ret = ttm_bo_vm_reserve(bo, vmf);
338 ret = ttm_bo_vm_fault_reserved(vmf, prot, TTM_BO_VM_NUM_PREFAULT);
341 ret = ttm_bo_vm_dummy_page(vmf, prot);
343 if (ret == VM_FAULT_RETRY && !(vmf->flags & FAULT_FLAG_RETRY_NOWAIT))
371 if (fault_flag_allow_retry_first(vmf->flags)) {
372 if (vmf->flags & FAULT_FLAG_RETRY_NOWAIT)
376 mmap_read_unlock(vmf->vma->vm_mm);
414 * @vmf: The fault structure handed to the callback
447 if (fault_flag_allow_retry_first(vmf->flags)) {
448 if (!(vmf->flags & FAULT_FLAG_RETRY_NOWAIT)) {
450 mmap_read_unlock(vmf->vma->vm_mm);
546 return ttm_bo_vm_insert_huge(vmf, bo, page_offset,