Lines Matching refs:load
142 static __always_inline u32 accumulate_sum(u64 delta, struct sched_avg *sa, unsigned long load, unsigned long runnable,
163 if (load) {
167 * if (!load)
179 if (load) {
180 sa->load_sum += load * contrib;
205 * following representation of historical load:
211 * This means that the contribution to load ~32ms ago (u_32) will be weighted
212 * approximately half as much as the contribution to load within the last ms
220 static __always_inline int ___update_load_sum(u64 now, struct sched_avg *sa, unsigned long load, unsigned long runnable,
257 if (!load) {
268 if (!accumulate_sum(delta, sa, load, runnable, running)) {
299 static __always_inline void ___update_load_avg(struct sched_avg *sa, unsigned long load)
306 sa->load_avg = div_u64(load * sa->load_sum, divider);
315 * se_weight() = se->load.weight
362 if (___update_load_sum(now, &cfs_rq->avg, scale_load_down(cfs_rq->load.weight), cfs_rq->h_nr_running,