/kernel/linux/linux-5.10/kernel/sched/ |
H A D | rt.c | 87 void init_rt_rq(struct rt_rq *rt_rq) in init_rt_rq() argument 92 array = &rt_rq->active; in init_rt_rq() 101 rt_rq->highest_prio.curr = MAX_RT_PRIO; in init_rt_rq() 102 rt_rq->highest_prio.next = MAX_RT_PRIO; in init_rt_rq() 103 rt_rq->rt_nr_migratory = 0; in init_rt_rq() 104 rt_rq->overloaded = 0; in init_rt_rq() 105 plist_head_init(&rt_rq->pushable_tasks); in init_rt_rq() 108 rt_rq->rt_queued = 0; in init_rt_rq() 110 rt_rq in init_rt_rq() 132 rq_of_rt_rq(struct rt_rq *rt_rq) rq_of_rt_rq() argument 144 struct rt_rq *rt_rq = rt_se->rt_rq; rq_of_rt_se() local 167 init_tg_rt_entry(struct task_group *tg, struct rt_rq *rt_rq, struct sched_rt_entity *rt_se, int cpu, struct sched_rt_entity *parent) init_tg_rt_entry() argument 196 struct rt_rq *rt_rq; alloc_rt_sched_group() local 243 rq_of_rt_rq(struct rt_rq *rt_rq) rq_of_rt_rq() argument 318 update_rt_migration(struct rt_rq *rt_rq) update_rt_migration() argument 331 inc_rt_migration(struct sched_rt_entity *rt_se, struct rt_rq *rt_rq) inc_rt_migration() argument 348 dec_rt_migration(struct sched_rt_entity *rt_se, struct rt_rq *rt_rq) dec_rt_migration() argument 424 inc_rt_migration(struct sched_rt_entity *rt_se, struct rt_rq *rt_rq) inc_rt_migration() argument 429 dec_rt_migration(struct sched_rt_entity *rt_se, struct rt_rq *rt_rq) dec_rt_migration() argument 496 sched_rt_runtime(struct rt_rq *rt_rq) sched_rt_runtime() argument 504 sched_rt_period(struct rt_rq *rt_rq) sched_rt_period() argument 540 sched_rt_rq_enqueue(struct rt_rq *rt_rq) sched_rt_rq_enqueue() argument 561 sched_rt_rq_dequeue(struct rt_rq *rt_rq) sched_rt_rq_dequeue() argument 577 rt_rq_throttled(struct rt_rq *rt_rq) rt_rq_throttled() argument 584 struct rt_rq *rt_rq = group_rt_rq(rt_se); rt_se_boosted() local 612 sched_rt_bandwidth(struct rt_rq *rt_rq) sched_rt_bandwidth() argument 619 sched_rt_runtime(struct rt_rq *rt_rq) sched_rt_runtime() argument 624 sched_rt_period(struct rt_rq *rt_rq) sched_rt_period() argument 642 sched_rt_rq_enqueue(struct rt_rq *rt_rq) sched_rt_rq_enqueue() argument 653 sched_rt_rq_dequeue(struct rt_rq *rt_rq) sched_rt_rq_dequeue() argument 658 rt_rq_throttled(struct rt_rq *rt_rq) rt_rq_throttled() argument 674 sched_rt_bandwidth(struct rt_rq *rt_rq) sched_rt_bandwidth() argument 681 sched_rt_bandwidth_account(struct rt_rq *rt_rq) sched_rt_bandwidth_account() argument 693 do_balance_runtime(struct rt_rq *rt_rq) do_balance_runtime() argument 749 struct rt_rq *rt_rq; __disable_runtime() local 830 struct rt_rq *rt_rq; __enable_runtime() local 851 balance_runtime(struct rt_rq *rt_rq) balance_runtime() argument 863 balance_runtime(struct rt_rq *rt_rq) balance_runtime() argument 885 struct rt_rq *rt_rq = sched_rt_period_rt_rq(rt_b, i); do_sched_rt_period_timer() local 951 struct rt_rq *rt_rq = group_rt_rq(rt_se); rt_se_prio() local 972 sched_rt_runtime_exceeded(struct rt_rq *rt_rq) sched_rt_runtime_exceeded() argument 1047 struct rt_rq *rt_rq = rt_rq_of_se(rt_se); update_curr_rt() local 1064 dequeue_top_rt_rq(struct rt_rq *rt_rq, unsigned int count) dequeue_top_rt_rq() argument 1081 enqueue_top_rt_rq(struct rt_rq *rt_rq) enqueue_top_rt_rq() argument 1105 inc_rt_prio_smp(struct rt_rq *rt_rq, int prio, int prev_prio) inc_rt_prio_smp() argument 1121 dec_rt_prio_smp(struct rt_rq *rt_rq, int prio, int prev_prio) dec_rt_prio_smp() argument 1139 inc_rt_prio_smp(struct rt_rq *rt_rq, int prio, int prev_prio) inc_rt_prio_smp() argument 1141 dec_rt_prio_smp(struct rt_rq *rt_rq, int prio, int prev_prio) dec_rt_prio_smp() argument 1147 inc_rt_prio(struct rt_rq *rt_rq, int prio) inc_rt_prio() argument 1158 dec_rt_prio(struct rt_rq *rt_rq, int prio) dec_rt_prio() argument 1185 inc_rt_prio(struct rt_rq *rt_rq, int prio) inc_rt_prio() argument 1186 dec_rt_prio(struct rt_rq *rt_rq, int prio) dec_rt_prio() argument 1193 inc_rt_group(struct sched_rt_entity *rt_se, struct rt_rq *rt_rq) inc_rt_group() argument 1203 dec_rt_group(struct sched_rt_entity *rt_se, struct rt_rq *rt_rq) dec_rt_group() argument 1214 inc_rt_group(struct sched_rt_entity *rt_se, struct rt_rq *rt_rq) inc_rt_group() argument 1220 dec_rt_group(struct sched_rt_entity *rt_se, struct rt_rq *rt_rq) dec_rt_group() argument 1250 inc_rt_tasks(struct sched_rt_entity *rt_se, struct rt_rq *rt_rq) inc_rt_tasks() argument 1264 dec_rt_tasks(struct sched_rt_entity *rt_se, struct rt_rq *rt_rq) dec_rt_tasks() argument 1301 struct rt_rq *rt_rq = rt_rq_of_se(rt_se); __enqueue_rt_entity() local 1335 struct rt_rq *rt_rq = rt_rq_of_se(rt_se); __dequeue_rt_entity() local 1388 struct rt_rq *rt_rq = group_rt_rq(rt_se); dequeue_rt_entity() local 1430 requeue_rt_entity(struct rt_rq *rt_rq, struct sched_rt_entity *rt_se, int head) requeue_rt_entity() argument 1446 struct rt_rq *rt_rq; requeue_task_rt() local 1634 pick_next_rt_entity(struct rt_rq *rt_rq) pick_next_rt_entity() argument 1655 struct rt_rq *rt_rq = &rq->rt; _pick_next_task_rt() local 2916 struct rt_rq *rt_rq = tg->rt_rq[i]; tg_set_rt_bandwidth() local 3005 struct rt_rq *rt_rq = &cpu_rq(i)->rt; sched_rt_global_constraints() local 3106 struct rt_rq *rt_rq; print_rt_stats() local [all...] |
H A D | autogroup.c | 32 ag->tg->rt_rq = NULL; in autogroup_destroy() 89 tg->rt_rq = root_task_group.rt_rq; in autogroup_create()
|
H A D | sched.h | 441 struct rt_rq; 490 struct rt_rq **rt_rq; member 577 extern void init_tg_rt_entry(struct task_group *tg, struct rt_rq *rt_rq, 720 struct rt_rq { struct 755 static inline bool rt_rq_is_runnable(struct rt_rq *rt_rq) in rt_rq_is_runnable() argument 757 return rt_rq->rt_queued && rt_rq in rt_rq_is_runnable() [all...] |
H A D | debug.c | 632 void print_rt_rq(struct seq_file *m, int cpu, struct rt_rq *rt_rq) in print_rt_rq() argument 636 SEQ_printf_task_group_path(m, rt_rq->tg, "rt_rq[%d]:%s\n", cpu); in print_rt_rq() 639 SEQ_printf(m, "rt_rq[%d]:\n", cpu); in print_rt_rq() 643 SEQ_printf(m, " .%-30s: %Ld\n", #x, (long long)(rt_rq->x)) in print_rt_rq() 645 SEQ_printf(m, " .%-30s: %lu\n", #x, (unsigned long)(rt_rq->x)) in print_rt_rq() 647 SEQ_printf(m, " .%-30s: %Ld.%06ld\n", #x, SPLIT_NS(rt_rq->x)) in print_rt_rq()
|
H A D | deadline.c | 1194 extern bool sched_rt_bandwidth_account(struct rt_rq *rt_rq); 1330 struct rt_rq *rt_rq = &rq->rt; in update_curr_dl() local 1332 raw_spin_lock(&rt_rq->rt_runtime_lock); in update_curr_dl() 1338 if (sched_rt_bandwidth_account(rt_rq)) in update_curr_dl() 1339 rt_rq->rt_time += delta_exec; in update_curr_dl() 1340 raw_spin_unlock(&rt_rq->rt_runtime_lock); in update_curr_dl()
|
H A D | core.c | 7690 root_task_group.rt_rq = (struct rt_rq **)ptr; in sched_init()
|
/kernel/linux/linux-6.6/kernel/sched/ |
H A D | rt.c | 144 void init_rt_rq(struct rt_rq *rt_rq) in init_rt_rq() argument 149 array = &rt_rq->active; in init_rt_rq() 158 rt_rq->highest_prio.curr = MAX_RT_PRIO-1; in init_rt_rq() 159 rt_rq->highest_prio.next = MAX_RT_PRIO-1; in init_rt_rq() 160 rt_rq->rt_nr_migratory = 0; in init_rt_rq() 161 rt_rq->overloaded = 0; in init_rt_rq() 162 plist_head_init(&rt_rq->pushable_tasks); in init_rt_rq() 165 rt_rq->rt_queued = 0; in init_rt_rq() 167 rt_rq in init_rt_rq() 189 rq_of_rt_rq(struct rt_rq *rt_rq) rq_of_rt_rq() argument 201 struct rt_rq *rt_rq = rt_se->rt_rq; rq_of_rt_se() local 228 init_tg_rt_entry(struct task_group *tg, struct rt_rq *rt_rq, struct sched_rt_entity *rt_se, int cpu, struct sched_rt_entity *parent) init_tg_rt_entry() argument 257 struct rt_rq *rt_rq; alloc_rt_sched_group() local 304 rq_of_rt_rq(struct rt_rq *rt_rq) rq_of_rt_rq() argument 376 update_rt_migration(struct rt_rq *rt_rq) update_rt_migration() argument 389 inc_rt_migration(struct sched_rt_entity *rt_se, struct rt_rq *rt_rq) inc_rt_migration() argument 406 dec_rt_migration(struct sched_rt_entity *rt_se, struct rt_rq *rt_rq) dec_rt_migration() argument 483 inc_rt_migration(struct sched_rt_entity *rt_se, struct rt_rq *rt_rq) inc_rt_migration() argument 488 dec_rt_migration(struct sched_rt_entity *rt_se, struct rt_rq *rt_rq) dec_rt_migration() argument 546 sched_rt_runtime(struct rt_rq *rt_rq) sched_rt_runtime() argument 554 sched_rt_period(struct rt_rq *rt_rq) sched_rt_period() argument 590 sched_rt_rq_enqueue(struct rt_rq *rt_rq) sched_rt_rq_enqueue() argument 611 sched_rt_rq_dequeue(struct rt_rq *rt_rq) sched_rt_rq_dequeue() argument 627 rt_rq_throttled(struct rt_rq *rt_rq) rt_rq_throttled() argument 634 struct rt_rq *rt_rq = group_rt_rq(rt_se); rt_se_boosted() local 662 sched_rt_bandwidth(struct rt_rq *rt_rq) sched_rt_bandwidth() argument 669 sched_rt_runtime(struct rt_rq *rt_rq) sched_rt_runtime() argument 674 sched_rt_period(struct rt_rq *rt_rq) sched_rt_period() argument 692 sched_rt_rq_enqueue(struct rt_rq *rt_rq) sched_rt_rq_enqueue() argument 703 sched_rt_rq_dequeue(struct rt_rq *rt_rq) sched_rt_rq_dequeue() argument 708 rt_rq_throttled(struct rt_rq *rt_rq) rt_rq_throttled() argument 724 sched_rt_bandwidth(struct rt_rq *rt_rq) sched_rt_bandwidth() argument 731 sched_rt_bandwidth_account(struct rt_rq *rt_rq) sched_rt_bandwidth_account() argument 743 do_balance_runtime(struct rt_rq *rt_rq) do_balance_runtime() argument 799 struct rt_rq *rt_rq; __disable_runtime() local 880 struct rt_rq *rt_rq; __enable_runtime() local 901 balance_runtime(struct rt_rq *rt_rq) balance_runtime() argument 913 balance_runtime(struct rt_rq *rt_rq) balance_runtime() argument 937 struct rt_rq *rt_rq = sched_rt_period_rt_rq(rt_b, i); do_sched_rt_period_timer() local 1004 struct rt_rq *rt_rq = group_rt_rq(rt_se); rt_se_prio() local 1013 sched_rt_runtime_exceeded(struct rt_rq *rt_rq) sched_rt_runtime_exceeded() argument 1086 struct rt_rq *rt_rq = rt_rq_of_se(rt_se); update_curr_rt() local 1103 dequeue_top_rt_rq(struct rt_rq *rt_rq, unsigned int count) dequeue_top_rt_rq() argument 1120 enqueue_top_rt_rq(struct rt_rq *rt_rq) enqueue_top_rt_rq() argument 1144 inc_rt_prio_smp(struct rt_rq *rt_rq, int prio, int prev_prio) inc_rt_prio_smp() argument 1160 dec_rt_prio_smp(struct rt_rq *rt_rq, int prio, int prev_prio) dec_rt_prio_smp() argument 1178 inc_rt_prio_smp(struct rt_rq *rt_rq, int prio, int prev_prio) inc_rt_prio_smp() argument 1180 dec_rt_prio_smp(struct rt_rq *rt_rq, int prio, int prev_prio) dec_rt_prio_smp() argument 1186 inc_rt_prio(struct rt_rq *rt_rq, int prio) inc_rt_prio() argument 1197 dec_rt_prio(struct rt_rq *rt_rq, int prio) dec_rt_prio() argument 1225 inc_rt_prio(struct rt_rq *rt_rq, int prio) inc_rt_prio() argument 1226 dec_rt_prio(struct rt_rq *rt_rq, int prio) dec_rt_prio() argument 1233 inc_rt_group(struct sched_rt_entity *rt_se, struct rt_rq *rt_rq) inc_rt_group() argument 1243 dec_rt_group(struct sched_rt_entity *rt_se, struct rt_rq *rt_rq) dec_rt_group() argument 1254 inc_rt_group(struct sched_rt_entity *rt_se, struct rt_rq *rt_rq) inc_rt_group() argument 1260 dec_rt_group(struct sched_rt_entity *rt_se, struct rt_rq *rt_rq) dec_rt_group() argument 1290 inc_rt_tasks(struct sched_rt_entity *rt_se, struct rt_rq *rt_rq) inc_rt_tasks() argument 1304 dec_rt_tasks(struct sched_rt_entity *rt_se, struct rt_rq *rt_rq) dec_rt_tasks() argument 1352 update_stats_wait_start_rt(struct rt_rq *rt_rq, struct sched_rt_entity *rt_se) update_stats_wait_start_rt() argument 1371 update_stats_enqueue_sleeper_rt(struct rt_rq *rt_rq, struct sched_rt_entity *rt_se) update_stats_enqueue_sleeper_rt() argument 1390 update_stats_enqueue_rt(struct rt_rq *rt_rq, struct sched_rt_entity *rt_se, int flags) update_stats_enqueue_rt() argument 1401 update_stats_wait_end_rt(struct rt_rq *rt_rq, struct sched_rt_entity *rt_se) update_stats_wait_end_rt() argument 1420 update_stats_dequeue_rt(struct rt_rq *rt_rq, struct sched_rt_entity *rt_se, int flags) update_stats_dequeue_rt() argument 1447 struct rt_rq *rt_rq = rt_rq_of_se(rt_se); __enqueue_rt_entity() local 1481 struct rt_rq *rt_rq = rt_rq_of_se(rt_se); __dequeue_rt_entity() local 1538 struct rt_rq *rt_rq = group_rt_rq(rt_se); dequeue_rt_entity() local 1583 requeue_rt_entity(struct rt_rq *rt_rq, struct sched_rt_entity *rt_se, int head) requeue_rt_entity() argument 1599 struct rt_rq *rt_rq; requeue_task_rt() local 1769 struct rt_rq *rt_rq = &rq->rt; set_next_task_rt() local 1792 pick_next_rt_entity(struct rt_rq *rt_rq) pick_next_rt_entity() argument 1813 struct rt_rq *rt_rq = &rq->rt; _pick_next_task_rt() local 1850 struct rt_rq *rt_rq = &rq->rt; put_prev_task_rt() local 2975 struct rt_rq *rt_rq; task_is_throttled_rt() local 3173 struct rt_rq *rt_rq = tg->rt_rq[i]; tg_set_rt_bandwidth() local 3266 struct rt_rq *rt_rq = &cpu_rq(i)->rt; sched_rt_global_constraints() local 3370 struct rt_rq *rt_rq; print_rt_stats() local [all...] |
H A D | autogroup.c | 54 ag->tg->rt_rq = NULL; in autogroup_destroy() 111 tg->rt_rq = root_task_group.rt_rq; in autogroup_create()
|
H A D | sched.h | 417 struct rt_rq; 473 struct rt_rq **rt_rq; member 559 extern void init_tg_rt_entry(struct task_group *tg, struct rt_rq *rt_rq, 766 struct rt_rq { struct 801 static inline bool rt_rq_is_runnable(struct rt_rq *rt_rq) in rt_rq_is_runnable() argument 803 return rt_rq->rt_queued && rt_rq in rt_rq_is_runnable() [all...] |
H A D | debug.c | 710 void print_rt_rq(struct seq_file *m, int cpu, struct rt_rq *rt_rq) in print_rt_rq() argument 714 SEQ_printf_task_group_path(m, rt_rq->tg, "rt_rq[%d]:%s\n", cpu); in print_rt_rq() 717 SEQ_printf(m, "rt_rq[%d]:\n", cpu); in print_rt_rq() 721 SEQ_printf(m, " .%-30s: %Ld\n", #x, (long long)(rt_rq->x)) in print_rt_rq() 723 SEQ_printf(m, " .%-30s: %lu\n", #x, (unsigned long)(rt_rq->x)) in print_rt_rq() 725 SEQ_printf(m, " .%-30s: %Ld.%06ld\n", #x, SPLIT_NS(rt_rq->x)) in print_rt_rq()
|
H A D | deadline.c | 1386 struct rt_rq *rt_rq = &rq->rt; in update_curr_dl() local 1388 raw_spin_lock(&rt_rq->rt_runtime_lock); in update_curr_dl() 1394 if (sched_rt_bandwidth_account(rt_rq)) in update_curr_dl() 1395 rt_rq->rt_time += delta_exec; in update_curr_dl() 1396 raw_spin_unlock(&rt_rq->rt_runtime_lock); in update_curr_dl()
|
H A D | core.c | 10634 root_task_group.rt_rq = (struct rt_rq **)ptr; in sched_init()
|
/kernel/linux/linux-5.10/include/linux/ |
H A D | sched.h | 615 struct rt_rq *rt_rq; member 617 struct rt_rq *my_q;
|
/kernel/linux/linux-6.6/include/linux/ |
H A D | sched.h | 701 struct rt_rq *rt_rq; member 703 struct rt_rq *my_q;
|