Home
last modified time | relevance | path

Searched refs:NICE_0_LOAD (Results 1 - 4 of 4) sorted by relevance

/kernel/linux/linux-5.10/kernel/sched/
H A Dsched.h226 * scale_load(sched_prio_to_weight[USER_PRIO(NICE_TO_PRIO(0))]) == NICE_0_LOAD
229 #define NICE_0_LOAD (1L << NICE_0_LOAD_SHIFT) macro
530 #define ROOT_TASK_GROUP_LOAD NICE_0_LOAD
H A Dfair.c251 * Either weight := NICE_0_LOAD and lw \e sched_prio_to_wmult[], in which case
253 * fit 32 bits, and NICE_0_LOAD gives another 10 bits; therefore shift >= 22.
702 if (unlikely(se->load.weight != NICE_0_LOAD)) in calc_delta_fair()
703 delta = __calc_delta(delta, NICE_0_LOAD, &se->load); in calc_delta_fair()
4378 if (sleep_time > ((1ULL << 63) / scale_load_down(NICE_0_LOAD))) in entity_is_long_sleeper()
4422 * minimal weight: scale_load_down(NICE_0_LOAD) / MIN_SHARES. in place_entity()
4428 * 2^63 / scale_load_down(NICE_0_LOAD) ~ 104 days in place_entity()
9385 imbalance = scale_load_down(NICE_0_LOAD) * in find_idlest_group()
11829 tg->shares = NICE_0_LOAD; in alloc_fair_sched_group()
11927 update_load_set(&se->load, NICE_0_LOAD); in init_tg_cfs_entry()
[all...]
/kernel/linux/linux-6.6/kernel/sched/
H A Dfair.c341 * Either weight := NICE_0_LOAD and lw \e sched_prio_to_wmult[], in which case
343 * fit 32 bits, and NICE_0_LOAD gives another 10 bits; therefore shift >= 22.
380 if (unlikely(se->load.weight != NICE_0_LOAD)) in calc_delta_fair()
381 delta = __calc_delta(delta, NICE_0_LOAD, &se->load); in calc_delta_fair()
10668 imbalance = scale_load_down(NICE_0_LOAD) * in find_idlest_group()
13328 tg->shares = NICE_0_LOAD; in alloc_fair_sched_group()
13428 update_load_set(&se->load, NICE_0_LOAD); in init_tg_cfs_entry()
13551 __sched_group_set_shares(tg, NICE_0_LOAD); in sched_group_set_idle()
H A Dsched.h249 * scale_load(sched_prio_to_weight[NICE_TO_PRIO(0)-MAX_RT_PRIO]) == NICE_0_LOAD
252 #define NICE_0_LOAD (1L << NICE_0_LOAD_SHIFT) macro
513 #define ROOT_TASK_GROUP_LOAD NICE_0_LOAD

Completed in 48 milliseconds