Lines Matching refs:gfp

178 struct page *__alloc_pages(gfp_t gfp, unsigned int order, int preferred_nid,
180 struct folio *__folio_alloc(gfp_t gfp, unsigned int order, int preferred_nid,
183 unsigned long __alloc_pages_bulk(gfp_t gfp, int preferred_nid,
188 unsigned long alloc_pages_bulk_array_mempolicy(gfp_t gfp,
194 alloc_pages_bulk_list(gfp_t gfp, unsigned long nr_pages, struct list_head *list)
196 return __alloc_pages_bulk(gfp, numa_mem_id(), NULL, nr_pages, list, NULL);
200 alloc_pages_bulk_array(gfp_t gfp, unsigned long nr_pages, struct page **page_array)
202 return __alloc_pages_bulk(gfp, numa_mem_id(), NULL, nr_pages, NULL, page_array);
206 alloc_pages_bulk_array_node(gfp_t gfp, int nid, unsigned long nr_pages, struct page **page_array)
211 return __alloc_pages_bulk(gfp, nid, NULL, nr_pages, NULL, page_array);
242 struct folio *__folio_alloc_node(gfp_t gfp, unsigned int order, int nid)
245 warn_if_node_offline(nid, gfp);
247 return __folio_alloc(gfp, order, nid, NULL);
265 struct page *alloc_pages(gfp_t gfp, unsigned int order);
266 struct page *alloc_pages_mpol(gfp_t gfp, unsigned int order,
268 struct folio *folio_alloc(gfp_t gfp, unsigned int order);
269 struct folio *vma_alloc_folio(gfp_t gfp, int order, struct vm_area_struct *vma,
276 static inline struct page *alloc_pages_mpol(gfp_t gfp, unsigned int order,
279 return alloc_pages(gfp, order);
281 static inline struct folio *folio_alloc(gfp_t gfp, unsigned int order)
283 return __folio_alloc_node(gfp, order, numa_node_id());
285 #define vma_alloc_folio(gfp, order, vma, addr, hugepage) \
286 folio_alloc(gfp, order)
289 static inline struct page *alloc_page_vma(gfp_t gfp,
292 struct folio *folio = vma_alloc_folio(gfp, 0, vma, addr, false);
359 static inline bool gfp_has_io_fs(gfp_t gfp)
361 return (gfp & (__GFP_IO | __GFP_FS)) == (__GFP_IO | __GFP_FS);
365 * Check if the gfp flags allow compaction - GFP_NOIO is a really