Lines Matching refs:ret

596 	int ret;
601 ret = gup_hugepte(vma, ptep, sz, addr, end, flags, pages, nr);
602 if (ret != 1)
603 return ret;
617 int nr = 0, ret;
627 ret = gup_hugepd(vma, hugepd, addr, pdshift, addr + PAGE_SIZE,
631 if (ret == 1) {
638 /* ret can be either 0 (translates to NULL) or negative */
639 return ERR_PTR(ret);
694 int ret;
732 ret = try_grab_page(page, flags);
733 if (ret)
734 page = ERR_PTR(ret);
787 int ret;
809 ret = try_grab_page(page, flags);
810 if (ret)
811 return ERR_PTR(ret);
908 int ret;
957 ret = follow_pfn_pte(vma, address, ptep, flags);
958 page = ERR_PTR(ret);
972 ret = try_grab_page(page, flags);
973 if (unlikely(ret)) {
974 page = ERR_PTR(ret);
984 ret = arch_make_page_accessible(page);
985 if (ret) {
987 page = ERR_PTR(ret);
1201 int ret = -EFAULT;
1236 ret = try_grab_page(*page, gup_flags);
1237 if (unlikely(ret))
1240 ret = 0;
1243 return ret;
1256 vm_fault_t ret;
1290 ret = handle_mm_fault(vma, address, fault_flags, NULL);
1292 if (ret & VM_FAULT_COMPLETED) {
1310 if (ret & VM_FAULT_ERROR) {
1311 int err = vm_fault_to_errno(ret, *flags);
1318 if (ret & VM_FAULT_RETRY) {
1523 long ret = 0, i = 0;
1548 ret = -ENOMEM;
1552 ret = -EINVAL;
1559 ret = get_gate_page(mm, start & PAGE_MASK,
1562 if (ret)
1569 ret = -EFAULT;
1572 ret = check_vma_flags(vma, gup_flags);
1573 if (ret)
1582 ret = -EINTR;
1589 ret = faultin_page(vma, start, &foll_flags,
1591 switch (ret) {
1596 ret = 0;
1612 ret = PTR_ERR(page);
1616 ret = PTR_ERR(page);
1654 ret = -EFAULT;
1674 return i ? i : ret;
1734 vm_fault_t ret;
1753 ret = handle_mm_fault(vma, address, fault_flags, NULL);
1755 if (ret & VM_FAULT_COMPLETED) {
1766 if (ret & VM_FAULT_ERROR) {
1767 int err = vm_fault_to_errno(ret, 0);
1774 if (ret & VM_FAULT_RETRY) {
1821 long ret, pages_done;
1857 ret = __get_user_pages(mm, start, nr_pages, flags, pages,
1861 pages_done = ret;
1867 BUG_ON(ret < 0);
1868 BUG_ON(ret >= nr_pages);
1871 if (ret > 0) {
1872 nr_pages -= ret;
1873 pages_done += ret;
1883 pages_done = ret;
1891 pages += ret;
1892 start += ret << PAGE_SHIFT;
1912 ret = mmap_read_lock_killable(mm);
1913 if (ret) {
1914 BUG_ON(ret > 0);
1916 pages_done = ret;
1921 ret = __get_user_pages(mm, start, 1, flags | FOLL_TRIED,
1925 BUG_ON(ret != 0);
1928 if (ret != 1) {
1929 BUG_ON(ret > 1);
1931 pages_done = ret;
1989 long ret;
2029 ret = __get_user_pages(mm, start, nr_pages, gup_flags,
2032 return ret;
2061 long ret;
2081 ret = __get_user_pages_locked(mm, start, nr_pages, NULL, locked,
2084 return ret;
2100 long ret = 0;
2132 ret = populate_vma_page_range(vma, nstart, nend, &locked);
2133 if (ret < 0) {
2135 ret = 0;
2140 nend = nstart + ret * PAGE_SIZE;
2141 ret = 0;
2145 return ret; /* 0 or negative error code */
2373 int ret;
2375 ret = __get_user_pages_locked(current->mm, addr, 1, &page, &locked,
2377 return (ret == 1) ? page : NULL;
2441 int ret;
2457 ret = -EBUSY;
2485 ret = -ENOMEM;
2500 return ret;
2926 int nr_start = *nr, ret = 0;
2992 ret = arch_make_page_accessible(page);
2993 if (ret) {
3003 ret = 1;
3009 return ret;
3443 int ret;
3473 ret = __gup_longterm_locked(current->mm, start, nr_pages - nr_pinned,
3476 if (ret < 0) {
3483 return ret;
3485 return ret + nr_pinned;