• Home
  • History
  • Annotate
  • Raw
  • Download
  • only in /netgear-R7000-V1.0.7.12_1.2.5/components/opensource/linux/linux-2.6.36/mm/

Lines Matching defs:ac

983 				struct array_cache *ac, int node)
987 if (ac->avail) {
995 transfer_objects(rl3->shared, ac, ac->limit);
997 free_block(cachep, ac->entry, ac->avail, node);
998 ac->avail = 0;
1011 struct array_cache *ac = l3->alien[node];
1013 if (ac && ac->avail && spin_trylock_irq(&ac->lock)) {
1014 __drain_alien_cache(cachep, ac, node);
1015 spin_unlock_irq(&ac->lock);
1024 struct array_cache *ac;
1028 ac = alien[i];
1029 if (ac) {
1030 spin_lock_irqsave(&ac->lock, flags);
1031 __drain_alien_cache(cachep, ac, i);
1032 spin_unlock_irqrestore(&ac->lock, flags);
2467 struct array_cache *ac,
2473 struct array_cache *ac;
2477 ac = cpu_cache_get(cachep);
2479 free_block(cachep, ac->entry, ac->avail, node);
2481 ac->avail = 0;
3002 struct array_cache *ac;
3008 ac = cpu_cache_get(cachep);
3009 batchcount = ac->batchcount;
3010 if (!ac->touched && batchcount > BATCHREFILL_LIMIT) {
3020 BUG_ON(ac->avail > 0 || !l3);
3024 if (l3->shared && transfer_objects(ac, l3->shared, batchcount)) {
3057 ac->entry[ac->avail++] = slab_get_obj(cachep, slabp,
3071 l3->free_objects -= ac->avail;
3075 if (unlikely(!ac->avail)) {
3079 /* cache_grow can reenable interrupts, then ac could change. */
3080 ac = cpu_cache_get(cachep);
3081 if (!x && ac->avail == 0) /* no objects in sight? abort */
3084 if (!ac->avail) /* objects refilled by interrupt? */
3087 ac->touched = 1;
3088 return ac->entry[--ac->avail];
3170 struct array_cache *ac;
3174 ac = cpu_cache_get(cachep);
3175 if (likely(ac->avail)) {
3177 ac->touched = 1;
3178 objp = ac->entry[--ac->avail];
3183 * the 'ac' may be updated by cache_alloc_refill(),
3186 ac = cpu_cache_get(cachep);
3194 kmemleak_erase(&ac->entry[ac->avail]);
3543 static void cache_flusharray(struct kmem_cache *cachep, struct array_cache *ac)
3549 batchcount = ac->batchcount;
3551 BUG_ON(!batchcount || batchcount > ac->avail);
3563 ac->entry, sizeof(void *) * batchcount);
3569 free_block(cachep, ac->entry, batchcount, node);
3590 ac->avail -= batchcount;
3591 memmove(ac->entry, &(ac->entry[batchcount]), sizeof(void *)*ac->avail);
3600 struct array_cache *ac = cpu_cache_get(cachep);
3618 if (likely(ac->avail < ac->limit)) {
3620 ac->entry[ac->avail++] = objp;
3624 cache_flusharray(cachep, ac);
3625 ac->entry[ac->avail++] = objp;
4072 struct array_cache *ac, int force, int node)
4076 if (!ac || !ac->avail)
4078 if (ac->touched && !force) {
4079 ac->touched = 0;
4082 if (ac->avail) {
4083 tofree = force ? ac->avail : (ac->limit + 4) / 5;
4084 if (tofree > ac->avail)
4085 tofree = (ac->avail + 1) / 2;
4086 free_block(cachep, ac->entry, tofree, node);
4087 ac->avail -= tofree;
4088 memmove(ac->entry, &(ac->entry[tofree]),
4089 sizeof(void *) * ac->avail);