Searched refs:scale_load (Results 1 - 8 of 8) sorted by relevance
/kernel/linux/linux-5.10/kernel/sched/ |
H A D | autogroup.c | 233 shares = scale_load(sched_prio_to_weight[idx]); in proc_sched_autogroup_set_nice()
|
H A D | core.c | 865 load->weight = scale_load(WEIGHT_IDLEPRIO); in set_load_weight() 877 load->weight = scale_load(sched_prio_to_weight[prio]); in set_load_weight() 8530 return sched_group_set_shares(css_tg(css), scale_load(shareval)); in cpu_shares_write_u64() 8949 return sched_group_set_shares(css_tg(css), scale_load(weight)); in cpu_weight_write_u64() 8983 return sched_group_set_shares(css_tg(css), scale_load(weight)); in cpu_weight_nice_write_s64()
|
H A D | sched.h | 206 # define scale_load(w) ((w) << SCHED_FIXEDPOINT_SHIFT) macro 216 # define scale_load(w) (w) macro 223 * scale_load() and scale_load_down(w) to convert between them. The 226 * scale_load(sched_prio_to_weight[USER_PRIO(NICE_TO_PRIO(0))]) == NICE_0_LOAD
|
H A D | fair.c | 3154 unsigned long weight = scale_load(sched_prio_to_weight[prio]); in reweight_task() 3260 * E.g. on 64-bit for a group with tg->shares of scale_load(15)=15*1024 in calc_group_shares() 3262 * to be 15*1024*1/8=1920 instead of scale_load(MIN_SHARES)=2*1024. In in calc_group_shares() 11943 shares = clamp(shares, scale_load(MIN_SHARES), scale_load(MAX_SHARES)); in sched_group_set_shares()
|
/kernel/linux/linux-6.6/kernel/sched/ |
H A D | autogroup.c | 256 shares = scale_load(sched_prio_to_weight[idx]); in proc_sched_autogroup_set_nice()
|
H A D | sched.h | 229 # define scale_load(w) ((w) << SCHED_FIXEDPOINT_SHIFT) macro 239 # define scale_load(w) (w) macro 246 * scale_load() and scale_load_down(w) to convert between them. The 249 * scale_load(sched_prio_to_weight[NICE_TO_PRIO(0)-MAX_RT_PRIO]) == NICE_0_LOAD
|
H A D | core.c | 1320 load->weight = scale_load(WEIGHT_IDLEPRIO); in set_load_weight() 1332 load->weight = scale_load(sched_prio_to_weight[prio]); in set_load_weight() 11551 return sched_group_set_shares(css_tg(css), scale_load(shareval)); in cpu_shares_write_u64() 12096 return sched_group_set_shares(css_tg(css), scale_load(weight)); in cpu_weight_write_u64() 12130 return sched_group_set_shares(css_tg(css), scale_load(weight)); in cpu_weight_nice_write_s64()
|
H A D | fair.c | 3862 unsigned long weight = scale_load(sched_prio_to_weight[prio]); in reweight_task() 3970 * E.g. on 64-bit for a group with tg->shares of scale_load(15)=15*1024 in calc_group_shares() 3972 * to be 15*1024*1/8=1920 instead of scale_load(MIN_SHARES)=2*1024. In in calc_group_shares() 13446 shares = clamp(shares, scale_load(MIN_SHARES), scale_load(MAX_SHARES)); in __sched_group_set_shares() 13549 __sched_group_set_shares(tg, scale_load(WEIGHT_IDLEPRIO)); in sched_group_set_idle()
|
Completed in 64 milliseconds