/kernel/linux/linux-5.10/kernel/sched/ |
H A D | walt.c | 223 p->ravg.demand_scaled; in fixup_walt_sched_stats_common() 417 if (p->ravg.curr_window_cpu[i]) { in update_cluster_load_subtractions() 420 p->ravg.curr_window_cpu[i], new_task); in update_cluster_load_subtractions() 421 p->ravg.curr_window_cpu[i] = 0; in update_cluster_load_subtractions() 424 if (p->ravg.prev_window_cpu[i]) { in update_cluster_load_subtractions() 427 p->ravg.prev_window_cpu[i], new_task); in update_cluster_load_subtractions() 428 p->ravg.prev_window_cpu[i] = 0; in update_cluster_load_subtractions() 444 p->ravg.curr_window_cpu[new_cpu] = p->ravg.curr_window; in inter_cluster_migration_fixup() 445 p->ravg in inter_cluster_migration_fixup() [all...] |
H A D | walt.h | 40 return (p->ravg.sum_history[0] == EXITING_TASK_MARKER); in exiting_task() 64 return p->ravg.active_windows < SCHED_NEW_TASK_WINDOWS; in is_new_task() 117 fixup_cumulative_runnable_avg(&rq->walt_stats, p->ravg.demand_scaled); in walt_inc_cumulative_runnable_avg() 127 walt_fixup_cum_window_demand(rq, p->ravg.demand_scaled); in walt_inc_cumulative_runnable_avg() 137 -(s64)p->ravg.demand_scaled); in walt_dec_cumulative_runnable_avg() 145 walt_fixup_cum_window_demand(rq, -(s64)p->ravg.demand_scaled); in walt_dec_cumulative_runnable_avg()
|
H A D | debug.c | 994 P(ravg.demand); in proc_sched_show_task()
|
H A D | fair.c | 3944 return p->ravg.demand_scaled; in task_util() 3960 return p->ravg.demand_scaled; in task_util_est() 12117 p->ravg.demand_scaled); in walt_inc_cfs_rq_stats() 12123 -(s64)p->ravg.demand_scaled); in walt_dec_cfs_rq_stats() 12164 p->ravg.demand_scaled; in walt_fixup_sched_stats_fair()
|
H A D | core.c | 9170 * sched_exit() - Set EXITING_TASK_MARKER in task's ravg.demand field 9204 p->ravg.mark_start = wallclock; in sched_exit() 9205 p->ravg.sum_history[0] = EXITING_TASK_MARKER; in sched_exit()
|
/kernel/linux/linux-6.6/kernel/sched/ |
H A D | walt.c | 224 p->ravg.demand_scaled; in fixup_walt_sched_stats_common() 418 if (p->ravg.curr_window_cpu[i]) { in update_cluster_load_subtractions() 421 p->ravg.curr_window_cpu[i], new_task); in update_cluster_load_subtractions() 422 p->ravg.curr_window_cpu[i] = 0; in update_cluster_load_subtractions() 425 if (p->ravg.prev_window_cpu[i]) { in update_cluster_load_subtractions() 428 p->ravg.prev_window_cpu[i], new_task); in update_cluster_load_subtractions() 429 p->ravg.prev_window_cpu[i] = 0; in update_cluster_load_subtractions() 445 p->ravg.curr_window_cpu[new_cpu] = p->ravg.curr_window; in inter_cluster_migration_fixup() 446 p->ravg in inter_cluster_migration_fixup() [all...] |
H A D | walt.h | 41 return (p->ravg.sum_history[0] == EXITING_TASK_MARKER); in exiting_task() 65 return p->ravg.active_windows < SCHED_NEW_TASK_WINDOWS; in is_new_task() 118 fixup_cumulative_runnable_avg(&rq->walt_stats, p->ravg.demand_scaled); in walt_inc_cumulative_runnable_avg() 128 walt_fixup_cum_window_demand(rq, p->ravg.demand_scaled); in walt_inc_cumulative_runnable_avg() 138 -(s64)p->ravg.demand_scaled); in walt_dec_cumulative_runnable_avg() 146 walt_fixup_cum_window_demand(rq, -(s64)p->ravg.demand_scaled); in walt_dec_cumulative_runnable_avg()
|
H A D | debug.c | 1061 P(ravg.demand); in proc_sched_show_task()
|
H A D | fair.c | 4790 return p->ravg.demand_scaled; in task_util() 4806 return p->ravg.demand_scaled; in task_util_est() 13724 p->ravg.demand_scaled); in walt_inc_cfs_rq_stats() 13730 -(s64)p->ravg.demand_scaled); in walt_dec_cfs_rq_stats() 13771 p->ravg.demand_scaled; in walt_fixup_sched_stats_fair()
|
H A D | core.c | 12874 * sched_exit() - Set EXITING_TASK_MARKER in task's ravg.demand field 12908 p->ravg.mark_start = wallclock; in sched_exit() 12909 p->ravg.sum_history[0] = EXITING_TASK_MARKER; in sched_exit()
|
/kernel/linux/linux-5.10/kernel/sched/rtg/ |
H A D | rtg.c | 137 *src_curr_runnable_sum -= p->ravg.curr_window_cpu[cpu]; in transfer_busy_time() 138 *src_prev_runnable_sum -= p->ravg.prev_window_cpu[cpu]; in transfer_busy_time() 141 p->ravg.curr_window_cpu[cpu]; in transfer_busy_time() 143 p->ravg.prev_window_cpu[cpu]; in transfer_busy_time() 162 *src_curr_runnable_sum -= p->ravg.curr_window; in transfer_busy_time() 163 *src_prev_runnable_sum -= p->ravg.prev_window; in transfer_busy_time() 165 *src_nt_curr_runnable_sum -= p->ravg.curr_window; in transfer_busy_time() 166 *src_nt_prev_runnable_sum -= p->ravg.prev_window; in transfer_busy_time() 176 p->ravg.curr_window_cpu[i] = 0; in transfer_busy_time() 177 p->ravg in transfer_busy_time() 422 group_time_rollover(struct group_ravg *ravg) group_time_rollover() argument [all...] |
H A D | frame_rtg.h | 105 return &frame_info_rtg(frame_info)->ravg; in frame_info_rtg_load()
|
/kernel/linux/linux-6.6/kernel/sched/rtg/ |
H A D | rtg.c | 138 *src_curr_runnable_sum -= p->ravg.curr_window_cpu[cpu]; in transfer_busy_time() 139 *src_prev_runnable_sum -= p->ravg.prev_window_cpu[cpu]; in transfer_busy_time() 142 p->ravg.curr_window_cpu[cpu]; in transfer_busy_time() 144 p->ravg.prev_window_cpu[cpu]; in transfer_busy_time() 163 *src_curr_runnable_sum -= p->ravg.curr_window; in transfer_busy_time() 164 *src_prev_runnable_sum -= p->ravg.prev_window; in transfer_busy_time() 166 *src_nt_curr_runnable_sum -= p->ravg.curr_window; in transfer_busy_time() 167 *src_nt_prev_runnable_sum -= p->ravg.prev_window; in transfer_busy_time() 177 p->ravg.curr_window_cpu[i] = 0; in transfer_busy_time() 178 p->ravg in transfer_busy_time() 423 group_time_rollover(struct group_ravg *ravg) group_time_rollover() argument [all...] |
H A D | frame_rtg.h | 106 return &frame_info_rtg(frame_info)->ravg; in frame_info_rtg_load()
|
/kernel/linux/linux-5.10/include/trace/events/ |
H A D | walt.h | 113 __entry->demand = p->ravg.demand; 114 memcpy(__entry->hist, p->ravg.sum_history, 171 __entry->mark_start = p->ravg.mark_start; 172 __entry->delta_m = (wallclock - p->ravg.mark_start); 173 __entry->demand = p->ravg.demand; 174 __entry->sum = p->ravg.sum; 178 __entry->curr_window = p->ravg.curr_window; 179 __entry->prev_window = p->ravg.prev_window; 180 __window_data(__get_dynamic_array(curr_sum), p->ravg.curr_window_cpu); 181 __window_data(__get_dynamic_array(prev_sum), p->ravg [all...] |
/kernel/linux/linux-6.6/include/trace/events/ |
H A D | walt.h | 113 __entry->demand = p->ravg.demand; 114 memcpy(__entry->hist, p->ravg.sum_history, 171 __entry->mark_start = p->ravg.mark_start; 172 __entry->delta_m = (wallclock - p->ravg.mark_start); 173 __entry->demand = p->ravg.demand; 174 __entry->sum = p->ravg.sum; 178 __entry->curr_window = p->ravg.curr_window; 179 __entry->prev_window = p->ravg.prev_window; 180 __window_data(__get_dynamic_array(curr_sum), p->ravg.curr_window_cpu); 181 __window_data(__get_dynamic_array(prev_sum), p->ravg [all...] |
/kernel/linux/linux-5.10/include/linux/sched/ |
H A D | rtg.h | 36 struct group_ravg ravg; member
|
/kernel/linux/linux-6.6/include/linux/sched/ |
H A D | rtg.h | 36 struct group_ravg ravg; member
|
/kernel/linux/linux-5.10/include/linux/ |
H A D | sched.h | 562 struct ravg { struct 822 struct ravg ravg; member
|
/kernel/linux/linux-6.6/include/linux/ |
H A D | sched.h | 648 struct ravg { struct 900 struct ravg ravg; member
|