/linux-master/tools/testing/radix-tree/ |
H A D | linux.c | 31 void kmem_cache_set_non_kernel(struct kmem_cache *cachep, unsigned int val) argument 33 cachep->non_kernel = val; 36 unsigned long kmem_cache_get_alloc(struct kmem_cache *cachep) argument 38 return cachep->size * cachep->nr_allocated; 41 unsigned long kmem_cache_nr_allocated(struct kmem_cache *cachep) argument 43 return cachep->nr_allocated; 46 unsigned long kmem_cache_nr_tallocated(struct kmem_cache *cachep) argument 48 return cachep->nr_tallocated; 51 void kmem_cache_zero_nr_tallocated(struct kmem_cache *cachep) argument 56 kmem_cache_alloc_lru(struct kmem_cache *cachep, struct list_lru *lru, int gfp) argument 96 __kmem_cache_free_locked(struct kmem_cache *cachep, void *objp) argument 110 kmem_cache_free_locked(struct kmem_cache *cachep, void *objp) argument 119 kmem_cache_free(struct kmem_cache *cachep, void *objp) argument 126 kmem_cache_free_bulk(struct kmem_cache *cachep, size_t size, void **list) argument 137 kmem_cache_shrink(struct kmem_cache *cachep) argument 141 kmem_cache_alloc_bulk(struct kmem_cache *cachep, gfp_t gfp, size_t size, void **p) argument [all...] |
/linux-master/tools/include/linux/ |
H A D | slab.h | 32 void *kmem_cache_alloc_lru(struct kmem_cache *cachep, struct list_lru *, int flags); 33 static inline void *kmem_cache_alloc(struct kmem_cache *cachep, int flags) argument 35 return kmem_cache_alloc_lru(cachep, NULL, flags); 37 void kmem_cache_free(struct kmem_cache *cachep, void *objp); 43 void kmem_cache_free_bulk(struct kmem_cache *cachep, size_t size, void **list); 44 int kmem_cache_alloc_bulk(struct kmem_cache *cachep, gfp_t gfp, size_t size,
|
/linux-master/drivers/scsi/snic/ |
H A D | snic_main.c | 819 struct kmem_cache *cachep; local 852 cachep = kmem_cache_create("snic_req_dfltsgl", len, SNIC_SG_DESC_ALIGN, 854 if (!cachep) { 860 snic_glob->req_cache[SNIC_REQ_CACHE_DFLT_SGL] = cachep; 865 cachep = kmem_cache_create("snic_req_maxsgl", len, SNIC_SG_DESC_ALIGN, 867 if (!cachep) { 873 snic_glob->req_cache[SNIC_REQ_CACHE_MAX_SGL] = cachep; 876 cachep = kmem_cache_create("snic_req_maxsgl", len, SNIC_SG_DESC_ALIGN, 878 if (!cachep) { 884 snic_glob->req_cache[SNIC_REQ_TM_CACHE] = cachep; [all...] |
/linux-master/arch/powerpc/mm/ |
H A D | hugetlbpage.c | 49 struct kmem_cache *cachep; local 55 cachep = PGT_CACHE(PTE_T_ORDER); 58 cachep = PGT_CACHE(pdshift - pshift); 62 if (!cachep) { 67 new = kmem_cache_alloc(cachep, pgtable_gfp_flags(mm, GFP_KERNEL)); 98 kmem_cache_free(cachep, new);
|
/linux-master/mm/ |
H A D | slab_common.c | 509 * @cachep: The cache to shrink. 516 int kmem_cache_shrink(struct kmem_cache *cachep) argument 518 kasan_cache_shrink(cachep); 520 return __kmem_cache_shrink(cachep); 1011 int cache_random_seq_create(struct kmem_cache *cachep, unsigned int count, argument 1015 if (count < 2 || cachep->random_seq) 1018 cachep->random_seq = kcalloc(count, sizeof(unsigned int), gfp); 1019 if (!cachep->random_seq) 1022 freelist_randomize(cachep->random_seq, count); 1027 void cache_random_seq_destroy(struct kmem_cache *cachep) argument [all...] |
H A D | slab.h | 614 int cache_random_seq_create(struct kmem_cache *cachep, unsigned int count, 616 void cache_random_seq_destroy(struct kmem_cache *cachep); 618 static inline int cache_random_seq_create(struct kmem_cache *cachep, argument 623 static inline void cache_random_seq_destroy(struct kmem_cache *cachep) { } argument
|
H A D | slub.c | 4316 struct kmem_cache *cachep; local 4322 cachep = virt_to_cache(x); 4323 if (WARN(cachep && cachep != s, 4325 __func__, s->name, cachep->name)) 4326 print_tracking(cachep, x); 4327 return cachep;
|
/linux-master/include/net/ |
H A D | inet_hashtables.h | 226 inet_bind_bucket_create(struct kmem_cache *cachep, struct net *net, 229 void inet_bind_bucket_destroy(struct kmem_cache *cachep, 237 inet_bind2_bucket_create(struct kmem_cache *cachep, struct net *net, 242 void inet_bind2_bucket_destroy(struct kmem_cache *cachep,
|
/linux-master/net/ipv4/ |
H A D | inet_hashtables.c | 64 struct inet_bind_bucket *inet_bind_bucket_create(struct kmem_cache *cachep, argument 70 struct inet_bind_bucket *tb = kmem_cache_alloc(cachep, GFP_ATOMIC); 87 void inet_bind_bucket_destroy(struct kmem_cache *cachep, struct inet_bind_bucket *tb) argument 91 kmem_cache_free(cachep, tb); 128 struct inet_bind2_bucket *inet_bind2_bucket_create(struct kmem_cache *cachep, argument 134 struct inet_bind2_bucket *tb2 = kmem_cache_alloc(cachep, GFP_ATOMIC); 143 void inet_bind2_bucket_destroy(struct kmem_cache *cachep, struct inet_bind2_bucket *tb) argument 148 kmem_cache_free(cachep, tb);
|
/linux-master/include/linux/ |
H A D | slab.h | 520 * @cachep: The cache to allocate from. 528 void *kmem_cache_alloc(struct kmem_cache *cachep, gfp_t flags) __assume_slab_alignment __malloc;
|
/linux-master/fs/ext4/ |
H A D | mballoc.c | 3268 struct kmem_cache *cachep = ext4_groupinfo_caches[cache_index]; local 3270 BUG_ON(!cachep); 3271 return cachep; 3319 struct kmem_cache *cachep = get_groupinfo_cache(sb->s_blocksize_bits); local 3343 meta_group_info[i] = kmem_cache_zalloc(cachep, GFP_NOFS); 3398 struct kmem_cache *cachep; local 3457 cachep = get_groupinfo_cache(sb->s_blocksize_bits); 3462 kmem_cache_free(cachep, grp); 3494 struct kmem_cache *cachep; local 3511 cachep 3763 struct kmem_cache *cachep = get_groupinfo_cache(sb->s_blocksize_bits); local [all...] |
/linux-master/fs/f2fs/ |
H A D | f2fs.h | 2812 static inline void *f2fs_kmem_cache_alloc_nofail(struct kmem_cache *cachep, argument 2817 entry = kmem_cache_alloc(cachep, flags); 2819 entry = kmem_cache_alloc(cachep, flags | __GFP_NOFAIL); 2823 static inline void *f2fs_kmem_cache_alloc(struct kmem_cache *cachep, argument 2827 return f2fs_kmem_cache_alloc_nofail(cachep, flags); 2832 return kmem_cache_alloc(cachep, flags);
|