Lines Matching refs:gfp

283  * @gfp: Memory allocation flags.
298 bool xas_nomem(struct xa_state *xas, gfp_t gfp)
305 gfp |= __GFP_ACCOUNT;
306 xas->xa_alloc = kmem_cache_alloc_lru(radix_tree_node_cachep, xas->xa_lru, gfp);
319 * @gfp: Memory allocation flags.
325 static bool __xas_nomem(struct xa_state *xas, gfp_t gfp)
335 gfp |= __GFP_ACCOUNT;
336 if (gfpflags_allow_blocking(gfp)) {
338 xas->xa_alloc = kmem_cache_alloc_lru(radix_tree_node_cachep, xas->xa_lru, gfp);
341 xas->xa_alloc = kmem_cache_alloc_lru(radix_tree_node_cachep, xas->xa_lru, gfp);
370 gfp_t gfp = GFP_NOWAIT | __GFP_NOWARN;
373 gfp |= __GFP_ACCOUNT;
375 node = kmem_cache_alloc_lru(radix_tree_node_cachep, xas->xa_lru, gfp);
994 * @gfp: Memory allocation flags.
1001 * Context: May sleep if @gfp flags permit.
1004 gfp_t gfp)
1020 node = kmem_cache_alloc_lru(radix_tree_node_cachep, xas->xa_lru, gfp);
1528 * @gfp: Memory allocation flags.
1535 * release and reacquire xa_lock if @gfp flags permit.
1538 void *__xa_store(struct xarray *xa, unsigned long index, void *entry, gfp_t gfp)
1552 } while (__xas_nomem(&xas, gfp));
1563 * @gfp: Memory allocation flags.
1570 * May sleep if the @gfp flags permit.
1575 void *xa_store(struct xarray *xa, unsigned long index, void *entry, gfp_t gfp)
1580 curr = __xa_store(xa, index, entry, gfp);
1593 * @gfp: Memory allocation flags.
1600 * release and reacquire xa_lock if @gfp flags permit.
1604 void *old, void *entry, gfp_t gfp)
1619 } while (__xas_nomem(&xas, gfp));
1630 * @gfp: Memory allocation flags.
1637 * release and reacquire xa_lock if @gfp flags permit.
1641 int __xa_insert(struct xarray *xa, unsigned long index, void *entry, gfp_t gfp)
1660 } while (__xas_nomem(&xas, gfp));
1704 * @gfp: Memory allocation flags.
1712 * if the @gfp flags permit.
1717 unsigned long last, void *entry, gfp_t gfp)
1746 } while (xas_nomem(&xas, gfp));
1799 * @gfp: Memory allocation flags.
1809 * release and reacquire xa_lock if @gfp flags permit.
1814 struct xa_limit limit, gfp_t gfp)
1835 } while (__xas_nomem(&xas, gfp));
1848 * @gfp: Memory allocation flags.
1860 * release and reacquire xa_lock if @gfp flags permit.
1866 struct xa_limit limit, u32 *next, gfp_t gfp)
1872 ret = __xa_alloc(xa, id, entry, limit, gfp);
1880 ret = __xa_alloc(xa, id, entry, limit, gfp);