• Home
  • History
  • Annotate
  • Raw
  • Download
  • only in /netgear-R7000-V1.0.7.12_1.2.5/components/opensource/linux/linux-2.6.36/mm/

Lines Matching defs:gfpflags

1557 slab_out_of_memory(struct kmem_cache *s, gfp_t gfpflags, int nid)
1563 nid, gfpflags);
1609 static void * BCMFASTPATH_HOST __slab_alloc(struct kmem_cache *s, gfp_t gfpflags, int node,
1616 gfpflags &= ~__GFP_ZERO;
1647 new = get_partial(s, gfpflags, node);
1654 if (gfpflags & __GFP_WAIT)
1657 new = new_slab(s, gfpflags, node);
1659 if (gfpflags & __GFP_WAIT)
1672 if (!(gfpflags & __GFP_NOWARN) && printk_ratelimit())
1673 slab_out_of_memory(s, gfpflags, node);
1696 gfp_t gfpflags, int node, unsigned long addr)
1702 gfpflags &= gfp_allowed_mask;
1704 lockdep_trace_alloc(gfpflags);
1705 might_sleep_if(gfpflags & __GFP_WAIT);
1707 if (should_failslab(s->objsize, gfpflags, s->flags))
1715 object = __slab_alloc(s, gfpflags, node, addr, c);
1723 if (unlikely(gfpflags & __GFP_ZERO) && object)
1726 kmemcheck_slab_alloc(s, gfpflags, object, s->objsize);
1727 kmemleak_alloc_recursive(object, s->objsize, 1, s->flags, gfpflags);
1732 void * BCMFASTPATH_HOST kmem_cache_alloc(struct kmem_cache *s, gfp_t gfpflags)
1734 void *ret = slab_alloc(s, gfpflags, NUMA_NO_NODE, _RET_IP_);
1736 trace_kmem_cache_alloc(_RET_IP_, ret, s->objsize, s->size, gfpflags);
1743 void *kmem_cache_alloc_notrace(struct kmem_cache *s, gfp_t gfpflags)
1745 return slab_alloc(s, gfpflags, NUMA_NO_NODE, _RET_IP_);
1751 void *kmem_cache_alloc_node(struct kmem_cache *s, gfp_t gfpflags, int node)
1753 void *ret = slab_alloc(s, gfpflags, node, _RET_IP_);
1756 s->objsize, s->size, gfpflags, node);
1765 gfp_t gfpflags,
1768 return slab_alloc(s, gfpflags, node, _RET_IP_);
2098 static void early_kmem_cache_node_alloc(gfp_t gfpflags, int node)
2106 page = new_slab(kmalloc_caches, gfpflags, node);
2150 static int init_kmem_cache_nodes(struct kmem_cache *s, gfp_t gfpflags)
2158 early_kmem_cache_node_alloc(gfpflags, node);
2162 gfpflags, node);
2179 static int init_kmem_cache_nodes(struct kmem_cache *s, gfp_t gfpflags)
2319 static int kmem_cache_open(struct kmem_cache *s, gfp_t gfpflags,
2355 if (!init_kmem_cache_nodes(s, gfpflags & ~SLUB_DMA))
2358 if (alloc_kmem_cache_cpus(s, gfpflags & ~SLUB_DMA))
3299 void * BCMFASTPATH_HOST __kmalloc_track_caller(size_t size, gfp_t gfpflags, unsigned long caller)
3305 return kmalloc_large(size, gfpflags);
3307 s = get_slab(size, gfpflags);
3312 ret = slab_alloc(s, gfpflags, NUMA_NO_NODE, caller);
3315 trace_kmalloc(caller, ret, size, s->size, gfpflags);
3320 void *__kmalloc_node_track_caller(size_t size, gfp_t gfpflags,
3327 ret = kmalloc_large_node(size, gfpflags, node);
3331 gfpflags, node);
3336 s = get_slab(size, gfpflags);
3341 ret = slab_alloc(s, gfpflags, node, caller);
3344 trace_kmalloc_node(caller, ret, size, s->size, gfpflags, node);