• Home
  • History
  • Annotate
  • Raw
  • Download
  • only in /netgear-WNDR4500v2-V1.0.0.60_1.0.38/src/linux/linux-2.6/mm/

Lines Matching defs:kmem_cache

58  *  Several members in struct kmem_cache and struct slab never change, they
248 struct kmem_cache *cachep;
314 static int drain_freelist(struct kmem_cache *cache,
316 static void free_block(struct kmem_cache *cachep, void **objpp, int len,
318 static int enable_cpucache(struct kmem_cache *cachep);
377 * struct kmem_cache
382 struct kmem_cache {
406 struct kmem_cache *slabp_cache;
411 void (*ctor) (void *, struct kmem_cache *, unsigned long);
445 * We put nodelists[] at the end of kmem_cache, because we want to size
528 static int obj_offset(struct kmem_cache *cachep)
533 static int obj_size(struct kmem_cache *cachep)
538 static unsigned long long *dbg_redzone1(struct kmem_cache *cachep, void *objp)
545 static unsigned long long *dbg_redzone2(struct kmem_cache *cachep, void *objp)
556 static void **dbg_userword(struct kmem_cache *cachep, void *objp)
584 static inline void page_set_cache(struct page *page, struct kmem_cache *cache)
589 static inline struct kmem_cache *page_get_cache(struct page *page)
593 return (struct kmem_cache *)page->lru.next;
607 static inline struct kmem_cache *virt_to_cache(const void *obj)
619 static inline void *index_to_obj(struct kmem_cache *cache, struct slab *slab,
631 static inline unsigned int obj_to_index(const struct kmem_cache *cache,
668 static struct kmem_cache cache_cache = {
672 .buffer_size = sizeof(struct kmem_cache),
673 .name = "kmem_cache",
754 static inline struct array_cache *cpu_cache_get(struct kmem_cache *cachep)
759 static inline struct kmem_cache *__find_general_cachep(size_t size,
786 static struct kmem_cache *kmem_find_general_cachep(size_t size, gfp_t gfpflags)
859 static void __slab_error(const char *function, struct kmem_cache *cachep,
997 static inline int cache_free_alien(struct kmem_cache *cachep, void *objp)
1002 static inline void *alternate_node_alloc(struct kmem_cache *cachep,
1008 static inline void *____cache_alloc_node(struct kmem_cache *cachep,
1016 static void *____cache_alloc_node(struct kmem_cache *, gfp_t, int);
1017 static void *alternate_node_alloc(struct kmem_cache *, gfp_t);
1057 static void __drain_alien_cache(struct kmem_cache *cachep,
1081 static void reap_alien(struct kmem_cache *cachep, struct kmem_list3 *l3)
1095 static void drain_alien_cache(struct kmem_cache *cachep,
1112 static inline int cache_free_alien(struct kmem_cache *cachep, void *objp)
1153 struct kmem_cache *cachep;
1359 static void init_list(struct kmem_cache *cachep, struct kmem_list3 *list,
1411 * kmem_cache structures of all caches, except cache_cache itself:
1417 * The struct kmem_cache for the new cache is allocated normally.
1438 * struct kmem_cache size depends on nr_node_ids, which
1441 cache_cache.buffer_size = offsetof(struct kmem_cache, nodelists) +
1571 struct kmem_cache *cachep;
1618 static void *kmem_getpages(struct kmem_cache *cachep, gfp_t flags, int nodeid)
1653 static void kmem_freepages(struct kmem_cache *cachep, void *addr)
1678 struct kmem_cache *cachep = slab_rcu->cachep;
1688 static void store_stackinfo(struct kmem_cache *cachep, unsigned long *addr,
1721 static void poison_obj(struct kmem_cache *cachep, void *addr, unsigned char val)
1764 static void print_objinfo(struct kmem_cache *cachep, void *objp, int lines)
1793 static void check_poison_obj(struct kmem_cache *cachep, void *objp)
1864 static void slab_destroy_objs(struct kmem_cache *cachep, struct slab *slabp)
1893 static void slab_destroy_objs(struct kmem_cache *cachep, struct slab *slabp)
1907 static void slab_destroy(struct kmem_cache *cachep, struct slab *slabp)
1930 static void __init set_up_list3s(struct kmem_cache *cachep, int index)
1942 static void __kmem_cache_destroy(struct kmem_cache *cachep)
1976 static size_t calculate_slab_order(struct kmem_cache *cachep,
2033 static int __init_refok setup_cpu_cache(struct kmem_cache *cachep)
2116 struct kmem_cache *
2119 void (*ctor)(void*, struct kmem_cache *, unsigned long),
2120 void (*dtor)(void*, struct kmem_cache *, unsigned long))
2123 struct kmem_cache *cachep = NULL, *pc;
2380 static void check_spinlock_acquired(struct kmem_cache *cachep)
2388 static void check_spinlock_acquired_node(struct kmem_cache *cachep, int node)
2403 static void drain_array(struct kmem_cache *cachep, struct kmem_list3 *l3,
2409 struct kmem_cache *cachep = arg;
2421 static void drain_cpu_caches(struct kmem_cache *cachep)
2447 static int drain_freelist(struct kmem_cache *cache,
2483 static int __cache_shrink(struct kmem_cache *cachep)
2511 int kmem_cache_shrink(struct kmem_cache *cachep)
2527 * Remove a &struct kmem_cache object from the slab cache.
2539 void kmem_cache_destroy(struct kmem_cache *cachep)
2575 static struct slab *alloc_slabmgmt(struct kmem_cache *cachep, void *objp,
2603 static void cache_init_objs(struct kmem_cache *cachep,
2652 static void kmem_flagcheck(struct kmem_cache *cachep, gfp_t flags)
2662 static void *slab_get_obj(struct kmem_cache *cachep, struct slab *slabp,
2679 static void slab_put_obj(struct kmem_cache *cachep, struct slab *slabp,
2704 static void slab_map_pages(struct kmem_cache *cache, struct slab *slab,
2727 static int cache_grow(struct kmem_cache *cachep,
2822 static inline void verify_redzone_free(struct kmem_cache *cache, void *obj)
2844 static void *cache_free_debugcheck(struct kmem_cache *cachep, void *objp,
2889 static void check_slabp(struct kmem_cache *cachep, struct slab *slabp)
2922 static void *cache_alloc_refill(struct kmem_cache *cachep, gfp_t flags)
3014 static inline void cache_alloc_debugcheck_before(struct kmem_cache *cachep,
3024 static void *cache_alloc_debugcheck_after(struct kmem_cache *cachep,
3104 static int should_failslab(struct kmem_cache *cachep, gfp_t flags)
3148 static inline int should_failslab(struct kmem_cache *cachep, gfp_t flags)
3155 static inline void *____cache_alloc(struct kmem_cache *cachep, gfp_t flags)
3181 static void *alternate_node_alloc(struct kmem_cache *cachep, gfp_t flags)
3205 static void *fallback_alloc(struct kmem_cache *cache, gfp_t flags)
3275 static void *____cache_alloc_node(struct kmem_cache *cachep, gfp_t flags,
3347 __cache_alloc_node(struct kmem_cache *cachep, gfp_t flags, int nodeid,
3389 __do_cache_alloc(struct kmem_cache *cache, gfp_t flags)
3413 __do_cache_alloc(struct kmem_cache *cachep, gfp_t flags)
3421 __cache_alloc(struct kmem_cache *cachep, gfp_t flags, void *caller)
3442 static void free_block(struct kmem_cache *cachep, void **objpp, int nr_objects,
3486 static void cache_flusharray(struct kmem_cache *cachep, struct array_cache *ac)
3541 static inline void __cache_free(struct kmem_cache *cachep, void *objp)
3570 void *kmem_cache_alloc(struct kmem_cache *cachep, gfp_t flags)
3584 void *kmem_cache_zalloc(struct kmem_cache *cache, gfp_t flags)
3607 int kmem_ptr_validate(struct kmem_cache *cachep, const void *ptr)
3636 void *kmem_cache_alloc_node(struct kmem_cache *cachep, gfp_t flags, int nodeid)
3646 struct kmem_cache *cachep;
3686 struct kmem_cache *cachep;
3734 struct kmem_cache *cache, *new_cache;
3775 void kmem_cache_free(struct kmem_cache *cachep, void *objp)
3799 struct kmem_cache *c;
3813 unsigned int kmem_cache_size(struct kmem_cache *cachep)
3819 const char *kmem_cache_name(struct kmem_cache *cachep)
3828 static int alloc_kmemlist(struct kmem_cache *cachep)
3914 struct kmem_cache *cachep;
3931 static int do_tune_cpucache(struct kmem_cache *cachep, int limit,
3974 static int enable_cpucache(struct kmem_cache *cachep)
4032 void drain_array(struct kmem_cache *cachep, struct kmem_list3 *l3,
4070 struct kmem_cache *searchp;
4164 return list_entry(p, struct kmem_cache, next);
4169 struct kmem_cache *cachep = p;
4172 NULL : list_entry(cachep->next.next, struct kmem_cache, next);
4182 struct kmem_cache *cachep = p;
4310 struct kmem_cache *cachep;
4361 return list_entry(p, struct kmem_cache, next);
4394 static void handle_slab(unsigned long *n, struct kmem_cache *c, struct slab *s)
4426 struct kmem_cache *cachep = p;