Lines Matching refs:last_update_time
1075 se->avg.last_update_time = cfs_rq_clock_pelt(cfs_rq);
4040 return u64_u32_load_copy(cfs_rq->avg.last_update_time,
4193 if (!(se->avg.last_update_time && prev))
4200 se->avg.last_update_time = n_last_update_time;
4506 * Estimated "now" is: last_update_time + cfs_idle_lag + rq_idle_lag, where:
4508 * last_update_time (the cfs_rq's last_update_time)
4549 * cfs_rq->avg.last_update_time is more recent than our
4635 u64_u32_store_copy(sa->last_update_time,
4637 sa->last_update_time);
4647 * cfs_rq->avg.last_update_time being current.
4664 se->avg.last_update_time = cfs_rq->avg.last_update_time;
4702 * cfs_rq->avg.last_update_time being current.
4744 if (se->avg.last_update_time && !(flags & SKIP_AGE_LOAD))
4750 if (!se->avg.last_update_time && (flags & DO_ATTACH)) {
4754 * !last_update_time means we've passed through
4784 u64 last_update_time;
4786 last_update_time = cfs_rq_last_update_time(cfs_rq);
4787 __update_load_avg_blocked_se(last_update_time, se);
6917 if (cpu_of(rq) != task_cpu(p) || !READ_ONCE(p->se.avg.last_update_time))
6940 if (cpu_of(rq) != task_cpu(p) || !READ_ONCE(p->se.avg.last_update_time))
7179 * prev_cpu's last_update_time.
7762 if (cpu != task_cpu(p) || !READ_ONCE(p->se.avg.last_update_time))
8238 * Estimate the missing time from the cfs_rq last_update_time
8246 se->avg.last_update_time = 0;
10156 if (cpu != task_cpu(p) || !READ_ONCE(p->se.avg.last_update_time))
12716 if (!se->avg.last_update_time)
12825 p->se.avg.last_update_time = 0;