Lines Matching refs:delta

552     s64 delta = (s64)(vruntime - max_vruntime);
553 if (delta > 0) {
562 s64 delta = (s64)(vruntime - min_vruntime);
563 if (delta < 0) {
706 * delta /= w
708 static inline u64 calc_delta_fair(u64 delta, struct sched_entity *se)
711 delta = fair_calc_delta(delta, NICE_0_LOAD, &se->load);
714 return delta;
955 u64 delta;
961 delta = rq_clock(rq_of(cfs_rq)) - schedstat_val(se->statistics.wait_start);
971 __schedstat_set(se->statistics.wait_start, delta);
974 trace_sched_stat_wait(p, delta);
977 __schedstat_set(se->statistics.wait_max, max(schedstat_val(se->statistics.wait_max), delta));
979 __schedstat_add(se->statistics.wait_sum, delta);
1000 u64 delta = rq_clock(rq_of(cfs_rq)) - sleep_start;
1001 if ((s64)delta < 0) {
1002 delta = 0;
1005 if (unlikely(delta > schedstat_val(se->statistics.sleep_max))) {
1006 __schedstat_set(se->statistics.sleep_max, delta);
1010 __schedstat_add(se->statistics.sum_sleep_runtime, delta);
1013 account_scheduler_latency(tsk, delta >> 10, 1);
1014 trace_sched_stat_sleep(tsk, delta);
1018 u64 delta = rq_clock(rq_of(cfs_rq)) - block_start;
1019 if ((s64)delta < 0) {
1020 delta = 0;
1023 if (unlikely(delta > schedstat_val(se->statistics.block_max))) {
1024 __schedstat_set(se->statistics.block_max, delta);
1028 __schedstat_add(se->statistics.sum_sleep_runtime, delta);
1032 __schedstat_add(se->statistics.iowait_sum, delta);
1034 trace_sched_stat_iowait(tsk, delta);
1037 trace_sched_stat_blocked(tsk, delta);
1045 profile_hits(SLEEP_PROFILING, (void *)get_wchan(tsk), delta >> FAIR_TWENTY);
1047 account_scheduler_latency(tsk, delta >> 10, 0);
2329 u64 runtime, delta, now;
2335 delta = runtime - p->last_sum_exec_runtime;
2343 delta = p->se.avg.load_sum;
2350 return delta;
3432 long delta = cfs_rq->avg.load_avg - cfs_rq->tg_load_avg_contrib;
3441 if (abs(delta) > cfs_rq->tg_load_avg_contrib / 0x40) {
3442 atomic_long_add(delta, &cfs_rq->tg->load_avg);
3565 long delta = gcfs_rq->avg.util_avg - se->avg.util_avg;
3569 if (!delta) {
3584 add_positive(&cfs_rq->avg.util_avg, delta);
3590 long delta = gcfs_rq->avg.runnable_avg - se->avg.runnable_avg;
3594 if (!delta) {
3609 add_positive(&cfs_rq->avg.runnable_avg, delta);
3615 long delta, running_sum, runnable_sum = gcfs_rq->prop_runnable_sum;
3664 delta = load_avg - se->avg.load_avg;
3669 add_positive(&cfs_rq->avg.load_avg, delta);
4594 s64 delta;
4618 delta = curr->vruntime - se->vruntime;
4620 if (delta < 0) {
4624 if (delta > ideal_runtime) {
5699 s64 delta = slice - ran;
5701 if (delta < 0) {
5707 hrtick_start(rq, delta);
7888 s64 delta;
7920 delta = rq_clock_task(env->src_rq) - p->se.exec_start;
7922 return delta < (s64)sysctl_sched_migration_cost;
8657 * average uses the actual delta max capacity(load).