Lines Matching refs:vruntime
550 static inline u64 max_vruntime(u64 max_vruntime, u64 vruntime)
552 s64 delta = (s64)(vruntime - max_vruntime);
554 max_vruntime = vruntime;
560 static inline u64 min_vruntime(u64 min_vruntime, u64 vruntime)
562 s64 delta = (s64)(vruntime - min_vruntime);
564 min_vruntime = vruntime;
572 return (s64)(a->vruntime - b->vruntime) < 0;
580 u64 vruntime = cfs_rq->min_vruntime;
584 vruntime = curr->vruntime;
595 vruntime = se->vruntime;
597 vruntime = min_vruntime(vruntime, se->vruntime);
602 cfs_rq->min_vruntime = max_vruntime(cfs_rq->min_vruntime, vruntime);
775 * We calculate the vruntime slice of a to-be-inserted task.
916 curr->vruntime += calc_delta_fair(delta_exec, curr);
922 trace_sched_stat_runtime(curtask, delta_exec, curr->vruntime);
4319 s64 d = se->vruntime - cfs_rq->min_vruntime;
4333 u64 vruntime = cfs_rq->min_vruntime;
4342 vruntime += sched_vslice(cfs_rq, se);
4357 vruntime -= thresh;
4361 se->vruntime = max_vruntime(se->vruntime, vruntime);
4392 * vruntime -= min_vruntime
4397 * vruntime += min_vruntime
4399 * this way the vruntime transition between RQs is done when both
4405 * vruntime -= min_vruntime
4410 * vruntime += min_vruntime
4426 se->vruntime += cfs_rq->min_vruntime;
4438 se->vruntime += cfs_rq->min_vruntime;
4568 se->vruntime -= cfs_rq->min_vruntime;
4618 delta = curr->vruntime - se->vruntime;
7195 * As blocked tasks retain absolute vruntime the migration needs to
7217 se->vruntime -= min_vruntime;
7316 s64 gran, vdiff = curr->vruntime - se->vruntime;
7490 * entity, update_curr() will update its vruntime, otherwise
11581 se->vruntime = curr->vruntime;
11590 swap(curr->vruntime, se->vruntime);
11594 se->vruntime -= cfs_rq->min_vruntime;
11633 * vruntime.
11640 * When !on_rq, vruntime of the task has usually NOT been normalized.
11725 * Fix up our vruntime so that the current sleep doesn't
11729 se->vruntime -= cfs_rq->min_vruntime;
11743 se->vruntime += cfs_rq->min_vruntime;