Lines Matching refs:ret

176 	long ret = delta;
179 return ret;
187 ret = -ENOMEM;
199 ret = delta - spool->rsv_hpages;
202 ret = 0; /* reserves already accounted for */
209 return ret;
221 long ret = delta;
235 ret = 0;
237 ret = spool->rsv_hpages + delta - spool->min_hpages;
250 return ret;
1917 long ret;
1927 ret = hugetlb_vmemmap_restore_folios(h, folio_list, &non_hvo_folios);
1928 if (ret < 0) {
1934 * At this point, list should be empty, ret should be >= 0 and there
1937 * Without HVO enabled, ret will be 0 and there is no need to call
1941 VM_WARN_ON(ret < 0);
1942 if (!list_empty(&non_hvo_folios) && ret) {
2630 int ret;
2644 ret = -ENOMEM;
2686 ret = 0;
2706 return ret;
2803 long ret;
2813 ret = region_chg(resv, idx, idx + 1, &dummy_out_regions_needed);
2821 ret = region_add(resv, idx, idx + 1, 1, NULL, NULL);
2823 VM_BUG_ON(ret < 0);
2827 ret = 0;
2831 ret = region_add(resv, idx, idx + 1, 1, NULL, NULL);
2833 VM_BUG_ON(ret < 0);
2836 ret = region_del(resv, idx, idx + 1);
2842 ret = region_del(resv, idx, idx + 1);
2844 ret = region_add(resv, idx, idx + 1, 1, NULL, NULL);
2846 VM_BUG_ON(ret < 0);
2854 return ret;
2870 if (ret > 0)
2872 if (ret == 0)
2874 return ret;
3014 int ret = 0;
3032 ret = isolated ? 0 : -EBUSY;
3078 return ret;
3088 return ret;
3095 int ret = -EBUSY;
3120 ret = 0;
3122 ret = alloc_and_dissolve_hugetlb_folio(h, folio, list);
3124 return ret;
3135 int memcg_charge_ret, ret, idx;
3190 ret = hugetlb_cgroup_charge_cgroup_rsvd(
3192 if (ret)
3196 ret = hugetlb_cgroup_charge_cgroup(idx, pages_per_huge_page(h), &h_cg);
3197 if (ret)
3338 int ret;
3345 ret = page_ref_freeze(page, 1);
3346 VM_BUG_ON(!ret);
3354 int ret;
3359 ret = folio_ref_freeze(folio, 1);
3360 VM_BUG_ON(!ret);
4936 int ret;
4941 ret = proc_hugetlb_doulongvec_minmax(table, write, buffer, length, ppos,
4943 if (ret)
4947 ret = __nr_hugepages_store_common(obey_mempolicy, h,
4950 return ret;
4975 int ret;
4985 ret = proc_hugetlb_doulongvec_minmax(table, write, buffer, length, ppos,
4987 if (ret)
4996 return ret;
5119 int ret = -ENOMEM;
5158 ret = 0;
5164 return ret;
5377 int ret = 0;
5406 ret = -ENOMEM;
5490 ret = PTR_ERR(new_folio);
5493 ret = copy_user_large_folio(new_folio,
5497 if (ret) {
5550 return ret;
5931 vm_fault_t ret = 0;
6057 ret = vmf_error(PTR_ERR(new_folio));
6065 ret = vmf_anon_prepare(vmf);
6066 if (unlikely(ret))
6070 ret = VM_FAULT_HWPOISON_LARGE | VM_FAULT_SET_HINDEX(hstate_index(h));
6116 return ret;
6205 vm_fault_t ret = VM_FAULT_SIGBUS;
6255 ret = 0;
6264 ret = vmf_anon_prepare(vmf);
6265 if (unlikely(ret))
6284 ret = vmf_error(PTR_ERR(folio));
6286 ret = 0;
6308 ret = VM_FAULT_SIGBUS;
6323 ret = VM_FAULT_HWPOISON_LARGE |
6334 ret = 0;
6350 ret = VM_FAULT_OOM;
6358 ret = 0;
6380 ret = hugetlb_wp(folio, vmf);
6397 return ret;
6437 vm_fault_t ret;
6488 ret = VM_FAULT_HWPOISON_LARGE |
6503 ret = 0;
6525 ret = VM_FAULT_HWPOISON_LARGE |
6540 ret = VM_FAULT_OOM;
6594 ret = hugetlb_wp(pagecache_folio, &vmf);
6627 return ret;
6678 int ret = -ENOMEM;
6704 ret = -EFAULT;
6715 ret = -EEXIST;
6721 ret = -ENOMEM;
6725 ret = copy_folio_from_user(folio, (const void __user *) src_addr,
6729 if (unlikely(ret)) {
6730 ret = -ENOENT;
6742 ret = -ENOMEM;
6756 ret = -EEXIST;
6764 ret = -ENOMEM;
6768 ret = copy_user_large_folio(folio, *foliop, dst_addr, dst_vma);
6771 if (ret) {
6796 ret = -EFAULT;
6806 ret = hugetlb_add_to_page_cache(folio, mapping, idx);
6807 if (ret)
6814 ret = -EIO;
6823 ret = -EEXIST;
6866 ret = 0;
6868 return ret;
7528 bool ret = true;
7534 ret = false;
7541 return ret;
7546 int ret = 0;
7553 ret = 0;
7555 ret = folio_try_get(folio);
7557 ret = -EBUSY;
7560 return ret;
7566 int ret;
7569 ret = __get_huge_page_for_hwpoison(pfn, flags, migratable_cleared);
7571 return ret;