Lines Matching refs:gcfs_rq
3967 struct cfs_rq *gcfs_rq = group_cfs_rq(se);
3970 if (!gcfs_rq)
3973 if (throttled_hierarchy(gcfs_rq))
3977 shares = READ_ONCE(gcfs_rq->tg->shares);
3979 shares = calc_group_shares(gcfs_rq);
4271 update_tg_cfs_util(struct cfs_rq *cfs_rq, struct sched_entity *se, struct cfs_rq *gcfs_rq)
4273 long delta_sum, delta_avg = gcfs_rq->avg.util_avg - se->avg.util_avg;
4288 se->avg.util_avg = gcfs_rq->avg.util_avg;
4303 update_tg_cfs_runnable(struct cfs_rq *cfs_rq, struct sched_entity *se, struct cfs_rq *gcfs_rq)
4305 long delta_sum, delta_avg = gcfs_rq->avg.runnable_avg - se->avg.runnable_avg;
4319 se->avg.runnable_avg = gcfs_rq->avg.runnable_avg;
4333 update_tg_cfs_load(struct cfs_rq *cfs_rq, struct sched_entity *se, struct cfs_rq *gcfs_rq)
4335 long delta_avg, running_sum, runnable_sum = gcfs_rq->prop_runnable_sum;
4344 gcfs_rq->prop_runnable_sum = 0;
4361 * Estimate the new unweighted runnable_sum of the gcfs_rq by
4364 if (scale_load_down(gcfs_rq->load.weight)) {
4365 load_sum = div_u64(gcfs_rq->avg.load_sum,
4366 scale_load_down(gcfs_rq->load.weight));
4409 struct cfs_rq *cfs_rq, *gcfs_rq;
4414 gcfs_rq = group_cfs_rq(se);
4415 if (!gcfs_rq->propagate)
4418 gcfs_rq->propagate = 0;
4422 add_tg_cfs_propagate(cfs_rq, gcfs_rq->prop_runnable_sum);
4424 update_tg_cfs_util(cfs_rq, se, gcfs_rq);
4425 update_tg_cfs_runnable(cfs_rq, se, gcfs_rq);
4426 update_tg_cfs_load(cfs_rq, se, gcfs_rq);
4440 struct cfs_rq *gcfs_rq = group_cfs_rq(se);
4453 if (gcfs_rq->propagate)