Lines Matching refs:runnable
142 static __always_inline u32 accumulate_sum(u64 delta, struct sched_avg *sa, unsigned long load, unsigned long runnable,
168 * runnable = running = 0;
182 if (runnable) {
183 sa->runnable_sum += runnable * contrib << SCHED_CAPACITY_SHIFT;
193 * We can represent the historical contribution to runnable average as the
194 * coefficients of a geometric series. To do this we sub-divide our runnable
202 * Let u_i denote the fraction of p_i that the entity was runnable.
220 static __always_inline int ___update_load_sum(u64 now, struct sched_avg *sa, unsigned long load, unsigned long runnable,
247 * running is a subset of runnable (weight) so running can't be set if
248 * runnable is clear. But there are some corner cases where the current
258 runnable = running = 0;
268 if (!accumulate_sum(delta, sa, load, runnable, running)) {
322 * runnable_sum = se_runnable() * runnable = grq->runnable_sum
325 * load_sum := runnable