Lines Matching defs:delta
376 * delta /= w
378 static inline u64 calc_delta_fair(u64 delta, struct sched_entity *se)
381 delta = __calc_delta(delta, NICE_0_LOAD, &se->load);
383 return delta;
617 s64 delta = (s64)(vruntime - max_vruntime);
618 if (delta > 0)
626 s64 delta = (s64)(vruntime - min_vruntime);
627 if (delta < 0)
726 void avg_vruntime_update(struct cfs_rq *cfs_rq, s64 delta)
731 cfs_rq->avg_vruntime -= cfs_rq->avg_load * delta;
827 s64 delta = (s64)(vruntime - min_vruntime);
828 if (delta > 0) {
829 avg_vruntime_update(cfs_rq, delta);
1296 * will be 0.So it will let the delta wrong. We need to avoid this
2764 u64 runtime, delta, now;
2770 delta = runtime - p->last_sum_exec_runtime;
2777 delta = p->se.avg.load_sum;
2784 return delta;
4119 long delta = cfs_rq->avg.load_avg - cfs_rq->tg_load_avg_contrib;
4127 if (abs(delta) > cfs_rq->tg_load_avg_contrib / 64) {
4128 atomic_long_add(delta, &cfs_rq->tg->load_avg);
4472 * cfs_idle_lag (delta between rq's update and cfs_rq's update)
4475 * rq_idle_lag (delta between now and rq's update)
5752 u64 delta = rq_clock(rq) - cfs_rq->throttled_clock_self;
5756 if (SCHED_WARN_ON((s64)delta < 0))
5757 delta = 0;
5759 cfs_rq->throttled_clock_self_time += delta;
6665 s64 delta = slice - ran;
6667 if (delta < 0) {
6672 hrtick_start(rq, delta);
8404 * Clamp the delta to stay in the scheduler period range
9021 s64 delta;
9055 delta = rq_clock_task(env->src_rq) - p->se.exec_start;
9057 return delta < (s64)sysctl_sched_migration_cost;
9759 * average uses the actual delta max capacity(load).
13010 s64 delta;
13040 * Find delta after normalizing se's vruntime with its cfs_rq's
13044 delta = (s64)(sea->vruntime - seb->vruntime) +
13047 return delta > 0;