Lines Matching defs:last_update_time
1185 se->avg.last_update_time = cfs_rq_clock_pelt(cfs_rq);
4059 return u64_u32_load_copy(cfs_rq->avg.last_update_time,
4154 if (!(se->avg.last_update_time && prev))
4161 se->avg.last_update_time = n_last_update_time;
4465 * Estimated "now" is: last_update_time + cfs_idle_lag + rq_idle_lag, where:
4467 * last_update_time (the cfs_rq's last_update_time)
4508 * cfs_rq->avg.last_update_time is more recent than our
4594 u64_u32_store_copy(sa->last_update_time,
4596 sa->last_update_time);
4606 * cfs_rq->avg.last_update_time being current.
4623 se->avg.last_update_time = cfs_rq->avg.last_update_time;
4661 * cfs_rq->avg.last_update_time being current.
4703 if (se->avg.last_update_time && !(flags & SKIP_AGE_LOAD))
4709 if (!se->avg.last_update_time && (flags & DO_ATTACH)) {
4713 * !last_update_time means we've passed through
4743 u64 last_update_time;
4745 last_update_time = cfs_rq_last_update_time(cfs_rq);
4746 __update_load_avg_blocked_se(last_update_time, se);
6986 if (cpu_of(rq) != task_cpu(p) || !READ_ONCE(p->se.avg.last_update_time))
7009 if (cpu_of(rq) != task_cpu(p) || !READ_ONCE(p->se.avg.last_update_time))
7251 * prev_cpu's last_update_time.
7868 if (cpu != task_cpu(p) || !READ_ONCE(p->se.avg.last_update_time))
8354 * Estimate the missing time from the cfs_rq last_update_time
8362 se->avg.last_update_time = 0;
10430 if (cpu != task_cpu(p) || !READ_ONCE(p->se.avg.last_update_time))
13185 if (!se->avg.last_update_time)
13294 p->se.avg.last_update_time = 0;