Lines Matching refs:gfp

150 			struct folio **foliop, enum sgp_type sgp, gfp_t gfp,
763 pgoff_t index, void *expected, gfp_t gfp)
777 gfp &= GFP_RECLAIM_MASK;
778 folio_throttle_swaprate(folio, gfp);
800 } while (xas_nomem(&xas, gfp));
1568 static struct folio *shmem_swapin_cluster(swp_entry_t swap, gfp_t gfp,
1576 folio = swap_cluster_readahead(swap, gfp, mpol, ilx);
1597 * Minimize the result gfp by taking the union with the deny flags,
1606 static struct folio *shmem_alloc_hugefolio(gfp_t gfp,
1614 page = alloc_pages_mpol(gfp, HPAGE_PMD_ORDER, mpol, ilx, numa_node_id());
1620 static struct folio *shmem_alloc_folio(gfp_t gfp,
1628 page = alloc_pages_mpol(gfp, 0, mpol, ilx, numa_node_id());
1634 static struct folio *shmem_alloc_and_add_folio(gfp_t gfp,
1663 folio = shmem_alloc_hugefolio(gfp, info, index);
1668 folio = shmem_alloc_folio(gfp, info, index);
1676 gfp &= GFP_RECLAIM_MASK;
1677 error = mem_cgroup_charge(folio, fault_mm, gfp);
1689 error = shmem_add_to_page_cache(folio, mapping, index, NULL, gfp);
1743 static bool shmem_should_replace_folio(struct folio *folio, gfp_t gfp)
1745 return folio_zonenum(folio) > gfp_zone(gfp);
1748 static int shmem_replace_folio(struct folio **foliop, gfp_t gfp,
1766 gfp &= ~GFP_CONSTRAINT_MASK;
1768 new = shmem_alloc_folio(gfp, info, index);
1850 gfp_t gfp, struct mm_struct *fault_mm,
1885 folio = shmem_swapin_cluster(swap, gfp, info, index);
1912 if (shmem_should_replace_folio(folio, gfp)) {
1913 error = shmem_replace_folio(&folio, gfp, info, index);
1919 swp_to_radix_entry(swap), gfp);
1960 struct folio **foliop, enum sgp_type sgp, gfp_t gfp,
1992 sgp, gfp, fault_mm, fault_type);
2044 huge_gfp = limit_gfp_mask(huge_gfp, gfp);
2055 folio = shmem_alloc_and_add_folio(gfp, inode, index, fault_mm, false);
2239 gfp_t gfp = mapping_gfp_mask(inode->i_mapping);
2256 gfp, vmf, &ret);
2615 gfp_t gfp = mapping_gfp_mask(mapping);
2637 folio = shmem_alloc_folio(gfp, info, pgoff);
2694 ret = mem_cgroup_charge(folio, dst_vma->vm_mm, gfp);
2697 ret = shmem_add_to_page_cache(folio, mapping, pgoff, NULL, gfp);
4941 * @gfp: the page allocator flags to use if allocating
4943 * This behaves as a tmpfs "read_cache_page_gfp(mapping, index, gfp)",
4953 pgoff_t index, gfp_t gfp)
4961 gfp, NULL, NULL);
4971 return mapping_read_folio_gfp(mapping, index, gfp);
4977 pgoff_t index, gfp_t gfp)
4979 struct folio *folio = shmem_read_folio_gfp(mapping, index, gfp);