Searched refs:NICE_0_LOAD (Results 1 - 4 of 4) sorted by relevance
/kernel/linux/linux-5.10/kernel/sched/ |
H A D | sched.h | 226 * scale_load(sched_prio_to_weight[USER_PRIO(NICE_TO_PRIO(0))]) == NICE_0_LOAD 229 #define NICE_0_LOAD (1L << NICE_0_LOAD_SHIFT) macro 530 #define ROOT_TASK_GROUP_LOAD NICE_0_LOAD
|
H A D | fair.c | 251 * Either weight := NICE_0_LOAD and lw \e sched_prio_to_wmult[], in which case 253 * fit 32 bits, and NICE_0_LOAD gives another 10 bits; therefore shift >= 22. 702 if (unlikely(se->load.weight != NICE_0_LOAD)) in calc_delta_fair() 703 delta = __calc_delta(delta, NICE_0_LOAD, &se->load); in calc_delta_fair() 4378 if (sleep_time > ((1ULL << 63) / scale_load_down(NICE_0_LOAD))) in entity_is_long_sleeper() 4422 * minimal weight: scale_load_down(NICE_0_LOAD) / MIN_SHARES. in place_entity() 4428 * 2^63 / scale_load_down(NICE_0_LOAD) ~ 104 days in place_entity() 9385 imbalance = scale_load_down(NICE_0_LOAD) * in find_idlest_group() 11829 tg->shares = NICE_0_LOAD; in alloc_fair_sched_group() 11927 update_load_set(&se->load, NICE_0_LOAD); in init_tg_cfs_entry() [all...] |
/kernel/linux/linux-6.6/kernel/sched/ |
H A D | fair.c | 341 * Either weight := NICE_0_LOAD and lw \e sched_prio_to_wmult[], in which case 343 * fit 32 bits, and NICE_0_LOAD gives another 10 bits; therefore shift >= 22. 380 if (unlikely(se->load.weight != NICE_0_LOAD)) in calc_delta_fair() 381 delta = __calc_delta(delta, NICE_0_LOAD, &se->load); in calc_delta_fair() 10668 imbalance = scale_load_down(NICE_0_LOAD) * in find_idlest_group() 13328 tg->shares = NICE_0_LOAD; in alloc_fair_sched_group() 13428 update_load_set(&se->load, NICE_0_LOAD); in init_tg_cfs_entry() 13551 __sched_group_set_shares(tg, NICE_0_LOAD); in sched_group_set_idle()
|
H A D | sched.h | 249 * scale_load(sched_prio_to_weight[NICE_TO_PRIO(0)-MAX_RT_PRIO]) == NICE_0_LOAD 252 #define NICE_0_LOAD (1L << NICE_0_LOAD_SHIFT) macro 513 #define ROOT_TASK_GROUP_LOAD NICE_0_LOAD
|
Completed in 48 milliseconds