Lines Matching refs:arc_p
337 /* shift of arc_c for calculating both min and max arc_p */
969 #define arc_p ARCSTAT(arcstat_p) /* target size of MRU */
1207 arc_p = (arc_c >> 1);
3993 * evict so much that we drop the MRU below arc_p, though. If
3994 * we're over the meta limit more than we're over arc_p, we
3999 refcount_count(&arc_mru->arcs_size) - arc_p));
4006 * space allotted to the MFU (which is defined as arc_c - arc_p).
4010 (arc_c - arc_p)));
4116 * arc_p. So, if we're over our target cache size more than
4117 * the MRU is over arc_p, we'll evict enough to get back to
4118 * arc_p here, and then evict more from the MFU below.
4122 refcount_count(&arc_mru->arcs_size) + ameta - arc_p));
4169 * size back to arc_p, if we're still above the target cache
4279 arc_c_min, uint64_t, arc_p, uint64_t, to_free);
4285 atomic_add_64(&arc_p, -(arc_p >> arc_shrink_shift));
4288 if (arc_p > arc_c)
4289 arc_p = (arc_c >> 1);
4292 arc_p);
4295 ASSERT((int64_t)arc_p >= 0);
4783 mult = MIN(mult, 10); /* avoid wild arc_p adjustment */
4785 arc_p = MIN(arc_c - arc_p_min, arc_p + bytes * mult);
4792 delta = MIN(bytes * mult, arc_p);
4793 arc_p = MAX(arc_p_min, arc_p - delta);
4795 ASSERT((int64_t)arc_p >= 0);
4822 atomic_add_64(&arc_p, (int64_t)bytes);
4823 if (arc_p > arc_c)
4824 arc_p = arc_c;
4826 ASSERT((int64_t)arc_p >= 0);
4964 * data, and we have outgrown arc_p, update arc_p
4969 refcount_count(&arc_mru->arcs_size) > arc_p))
4970 arc_p = MIN(arc_c, arc_p + size);
6798 arc_p = (arc_c >> 1);