Lines Matching defs:rq
294 * and the cfs rq, to which they are attached, have the same position in the
383 int update_rt_rq_load_avg(u64 now, struct rq *rq, int running)
385 if (___update_load_sum(now, &rq->avg_rt, running, running, running)) {
386 ___update_load_avg(&rq->avg_rt, 1);
387 trace_pelt_rt_tp(rq);
405 int update_dl_rq_load_avg(u64 now, struct rq *rq, int running)
407 if (___update_load_sum(now, &rq->avg_dl, running, running, running)) {
408 ___update_load_avg(&rq->avg_dl, 1);
409 trace_pelt_dl_tp(rq);
432 int update_thermal_load_avg(u64 now, struct rq *rq, u64 capacity)
434 if (___update_load_sum(now, &rq->avg_thermal, capacity, capacity, capacity)) {
435 ___update_load_avg(&rq->avg_thermal, 1);
436 trace_pelt_thermal_tp(rq);
456 int update_irq_load_avg(struct rq *rq, u64 running)
465 running = cap_scale(running, arch_scale_freq_capacity(cpu_of(rq)));
466 running = cap_scale(running, arch_scale_cpu_capacity(cpu_of(rq)));
473 * of rq clock during which the metric is updated.
476 * We can safely remove running from rq->clock because
477 * rq->clock += delta with delta >= running
479 ret = ___update_load_sum(rq->clock - running, &rq->avg_irq, 0, 0, 0);
480 ret += ___update_load_sum(rq->clock, &rq->avg_irq, 1, 1, 1);
482 ___update_load_avg(&rq->avg_irq, 1);
483 trace_pelt_irq_tp(rq);