Lines Matching refs:gfp_mask
197 gfp_t gfp_mask, int node_id)
207 gfp_mask, node_id);
217 element = pool->alloc(gfp_mask, pool->pool_data);
278 gfp_t gfp_mask, int node_id)
282 pool = kzalloc_node(sizeof(*pool), gfp_mask, node_id);
287 gfp_mask, node_id)) {
380 * @gfp_mask: the usual allocation bitmask.
390 void *mempool_alloc(mempool_t *pool, gfp_t gfp_mask)
397 VM_WARN_ON_ONCE(gfp_mask & __GFP_ZERO);
398 might_alloc(gfp_mask);
400 gfp_mask |= __GFP_NOMEMALLOC; /* don't allocate emergency reserves */
401 gfp_mask |= __GFP_NORETRY; /* don't loop in __alloc_pages */
402 gfp_mask |= __GFP_NOWARN; /* failures are OK */
404 gfp_temp = gfp_mask & ~(__GFP_DIRECT_RECLAIM|__GFP_IO);
430 if (gfp_temp != gfp_mask) {
432 gfp_temp = gfp_mask;
437 if (!(gfp_mask & __GFP_DIRECT_RECLAIM)) {
561 void *mempool_alloc_slab(gfp_t gfp_mask, void *pool_data)
565 return kmem_cache_alloc(mem, gfp_mask);
580 void *mempool_kmalloc(gfp_t gfp_mask, void *pool_data)
583 return kmalloc(size, gfp_mask);
593 void *mempool_kvmalloc(gfp_t gfp_mask, void *pool_data)
596 return kvmalloc(size, gfp_mask);
610 void *mempool_alloc_pages(gfp_t gfp_mask, void *pool_data)
613 return alloc_pages(gfp_mask, order);