Home
last modified time | relevance | path

Searched refs:cpu_of (Results 1 - 5 of 5) sorted by relevance

/device/soc/rockchip/common/sdk_linux/kernel/sched/
H A Dpelt.c465 running = cap_scale(running, arch_scale_freq_capacity(cpu_of(rq))); in update_irq_load_avg()
466 running = cap_scale(running, arch_scale_cpu_capacity(cpu_of(rq))); in update_irq_load_avg()
H A Dfair.c344 int cpu = cpu_of(rq); in list_add_leaf_cfs_rq()
843 long cpu_scale = arch_scale_cpu_capacity(cpu_of(rq_of(cfs_rq))); in post_init_entity_util_avg()
4184 if (task_util(p) > capacity_orig_of(cpu_of(rq_of(cfs_rq)))) { in util_est_update()
4237 int cpu = cpu_of(rq); in update_misfit_status()
4255 task_fits = task_fits_capacity(p, capacity_of(cpu_of(rq))); in update_misfit_status()
4258 task_fits = task_fits_capacity(p, capacity_of(cpu_of(rq))); in update_misfit_status()
4952 struct cfs_rq *cfs_rq = tg->cfs_rq[cpu_of(rq)]; in tg_unthrottle_up()
4970 struct cfs_rq *cfs_rq = tg->cfs_rq[cpu_of(rq)]; in tg_throttle_down()
5010 se = cfs_rq->tg->se[cpu_of(rq_of(cfs_rq))]; in throttle_cfs_rq()
5064 se = cfs_rq->tg->se[cpu_of(r in unthrottle_cfs_rq()
[all...]
H A Drt.c289 return rq->rt.highest_prio.curr > prev->prio && !cpu_isolated(cpu_of(rq)); in need_pull_rt_task()
543 (iter = next_task_group(iter)) && (rt_rq = iter->rt_rq[cpu_of(rq)]);) \
562 int cpu = cpu_of(rq); in sched_rt_rq_enqueue()
582 int cpu = cpu_of(rq_of_rt_rq(rt_rq)); in sched_rt_rq_dequeue()
2568 if (!task_on_rq_queued(p) || rq->rt.rt_nr_running || cpu_isolated(cpu_of(rq))) { in switched_from_rt()
2613 if (p->prio < rq->curr->prio && cpu_online(cpu_of(rq))) { in switched_to_rt()
2754 if (capacity_orig_of(cpu_of(lowest_rq)) <= capacity_orig_of(task_cpu(next_task))) { in rt_active_load_balance_cpu_stop()
H A Dcore.c267 irq_delta = irq_time_read(cpu_of(rq)) - rq->prev_irq_time; in update_rq_clock_task()
292 steal = paravirt_steal_clock(cpu_of(rq)); in update_rq_clock_task()
331 delta = sched_clock_cpu(cpu_of(rq)) - rq->clock; in update_rq_clock()
368 WARN_ON_ONCE(cpu_of(rq) != smp_processor_id()); in hrtick()
421 smp_call_function_single_async(cpu_of(rq), &rq->hrtick_csd); in hrtick_start()
642 cpu = cpu_of(rq); in resched_curr()
795 int cpu = cpu_of(rq); in nohz_csd_func()
2141 stop_one_cpu(cpu_of(rq), migration_cpu_stop, &arg); in __set_cpus_allowed_ptr()
2862 if (WARN_ON_ONCE(task_cpu(p) != cpu_of(rq))) { in sched_ttwu_pending()
2863 set_task_cpu(p, cpu_of(r in sched_ttwu_pending()
[all...]
H A Dsched.h1186 static inline int cpu_of(struct rq *rq) in cpu_of() function
2071 int cpu = cpu_of(rq); in sched_update_tick_dependency()
2143 if (!cpu_active(cpu_of(rq))) { in hrtick_enabled()
2509 data = rcu_dereference_sched(*per_cpu_ptr(&cpufreq_update_util_data, cpu_of(rq))); in cpufreq_update_util()

Completed in 26 milliseconds