Lines Matching refs:sa
102 accumulate_sum(u64 delta, struct sched_avg *sa,
108 delta += sa->period_contrib;
115 sa->load_sum = decay_load(sa->load_sum, periods);
116 sa->runnable_sum =
117 decay_load(sa->runnable_sum, periods);
118 sa->util_sum = decay_load((u64)(sa->util_sum), periods);
136 1024 - sa->period_contrib, delta);
139 sa->period_contrib = delta;
142 sa->load_sum += load * contrib;
144 sa->runnable_sum += runnable * contrib << SCHED_CAPACITY_SHIFT;
146 sa->util_sum += contrib << SCHED_CAPACITY_SHIFT;
180 ___update_load_sum(u64 now, struct sched_avg *sa,
185 delta = now - sa->last_update_time;
191 sa->last_update_time = now;
203 sa->last_update_time += delta << 10;
226 if (!accumulate_sum(delta, sa, load, runnable, running))
241 * LOAD_AVG_MAX*y + sa->period_contrib
245 * LOAD_AVG_MAX - 1024 + sa->period_contrib
257 ___update_load_avg(struct sched_avg *sa, unsigned long load)
259 u32 divider = get_pelt_divider(sa);
264 sa->load_avg = div_u64(load * sa->load_sum, divider);
265 sa->runnable_avg = div_u64(sa->runnable_sum, divider);
266 WRITE_ONCE(sa->util_avg, sa->util_sum / divider);