Lines Matching defs:last_update_time
1081 se->avg.last_update_time = cfs_rq_clock_pelt(cfs_rq);
4046 return u64_u32_load_copy(cfs_rq->avg.last_update_time,
4199 if (!(se->avg.last_update_time && prev))
4206 se->avg.last_update_time = n_last_update_time;
4512 * Estimated "now" is: last_update_time + cfs_idle_lag + rq_idle_lag, where:
4514 * last_update_time (the cfs_rq's last_update_time)
4555 * cfs_rq->avg.last_update_time is more recent than our
4641 u64_u32_store_copy(sa->last_update_time,
4643 sa->last_update_time);
4653 * cfs_rq->avg.last_update_time being current.
4670 se->avg.last_update_time = cfs_rq->avg.last_update_time;
4708 * cfs_rq->avg.last_update_time being current.
4750 if (se->avg.last_update_time && !(flags & SKIP_AGE_LOAD))
4756 if (!se->avg.last_update_time && (flags & DO_ATTACH)) {
4760 * !last_update_time means we've passed through
4790 u64 last_update_time;
4792 last_update_time = cfs_rq_last_update_time(cfs_rq);
4793 __update_load_avg_blocked_se(last_update_time, se);
6923 if (cpu_of(rq) != task_cpu(p) || !READ_ONCE(p->se.avg.last_update_time))
6946 if (cpu_of(rq) != task_cpu(p) || !READ_ONCE(p->se.avg.last_update_time))
7185 * prev_cpu's last_update_time.
7768 if (cpu != task_cpu(p) || !READ_ONCE(p->se.avg.last_update_time))
8244 * Estimate the missing time from the cfs_rq last_update_time
8252 se->avg.last_update_time = 0;
10162 if (cpu != task_cpu(p) || !READ_ONCE(p->se.avg.last_update_time))
12722 if (!se->avg.last_update_time)
12831 p->se.avg.last_update_time = 0;