Lines Matching defs:kmem_cache_cpu
121 * This locks protect slowpath manipulation of all kmem_cache_cpu fields
148 * local cpu doesn't change in the process and e.g. the kmem_cache_cpu pointer
384 struct kmem_cache_cpu {
2774 struct kmem_cache_cpu *c;
2786 * Assumes the slab has been already safely taken away from kmem_cache_cpu
2929 struct kmem_cache_cpu *c)
2990 struct kmem_cache_cpu *c) { }
2994 static inline void flush_slab(struct kmem_cache *s, struct kmem_cache_cpu *c)
3019 struct kmem_cache_cpu *c = per_cpu_ptr(s->cpu_slab, cpu);
3049 struct kmem_cache_cpu *c;
3065 struct kmem_cache_cpu *c = per_cpu_ptr(s->cpu_slab, cpu);
3372 unsigned long addr, struct kmem_cache_cpu *c, unsigned int orig_size)
3597 unsigned long addr, struct kmem_cache_cpu *c, unsigned int orig_size)
3620 struct kmem_cache_cpu *c;
3632 * We must guarantee that tid and kmem_cache_cpu are retrieved on the
3633 * same cpu. We read first the kmem_cache_cpu pointer and use it to read
4215 struct kmem_cache_cpu *c;
4518 struct kmem_cache_cpu *c;
4797 sizeof(struct kmem_cache_cpu));
4803 s->cpu_slab = __alloc_percpu(sizeof(struct kmem_cache_cpu),
6010 struct kmem_cache_cpu *c = per_cpu_ptr(s->cpu_slab,