/kernel/linux/linux-5.10/arch/arm/kernel/ |
H A D | topology.c | 150 >> (SCHED_CAPACITY_SHIFT+1); in parse_dt_topology() 153 >> (SCHED_CAPACITY_SHIFT-1)) + 1; in parse_dt_topology()
|
/kernel/linux/linux-6.6/arch/arm/kernel/ |
H A D | topology.c | 150 >> (SCHED_CAPACITY_SHIFT+1); in parse_dt_topology() 153 >> (SCHED_CAPACITY_SHIFT-1)) + 1; in parse_dt_topology()
|
/kernel/linux/linux-5.10/arch/arm64/kernel/ |
H A D | topology.c | 135 * We use a factor of 2 * SCHED_CAPACITY_SHIFT -> SCHED_CAPACITY_SCALE² in validate_cpu_freq_invariance_counters() 140 ratio = (u64)arch_timer_get_rate() << (2 * SCHED_CAPACITY_SHIFT); in validate_cpu_freq_invariance_counters() 258 * arch_max_freq_scale and the use of SCHED_CAPACITY_SHIFT. in topology_scale_freq_tick() 262 scale = div64_u64(scale >> SCHED_CAPACITY_SHIFT, in topology_scale_freq_tick()
|
/kernel/linux/linux-6.6/arch/arm64/kernel/ |
H A D | topology.c | 133 * We use a factor of 2 * SCHED_CAPACITY_SHIFT -> SCHED_CAPACITY_SCALE² in freq_inv_set_max_ratio() 138 ratio = ref_rate << (2 * SCHED_CAPACITY_SHIFT); in freq_inv_set_max_ratio() 173 * arch_max_freq_scale and the use of SCHED_CAPACITY_SHIFT. in amu_scale_freq_tick() 177 scale = div64_u64(scale >> SCHED_CAPACITY_SHIFT, in amu_scale_freq_tick()
|
/kernel/linux/linux-5.10/kernel/sched/ |
H A D | pelt.c | 148 sa->runnable_sum += runnable * contrib << SCHED_CAPACITY_SHIFT; in accumulate_sum() 150 sa->util_sum += contrib << SCHED_CAPACITY_SHIFT; in accumulate_sum()
|
H A D | pelt.h | 110 * (LOAD_AVG_MAX - 1024 + rq->cfs.avg.period_contrib) << SCHED_CAPACITY_SHIFT; 117 u32 divider = ((LOAD_AVG_MAX - 1024) << SCHED_CAPACITY_SHIFT) - LOAD_AVG_MAX; in update_idle_rq_clock_pelt()
|
H A D | sched_avg.c | 173 sched_ravg_window >> SCHED_CAPACITY_SHIFT); in sched_get_cpu_util()
|
H A D | walt.h | 57 delta = (delta * capcurr) >> SCHED_CAPACITY_SHIFT; in scale_exec_time()
|
H A D | cpufreq_schedutil.c | 421 boost = (sg_cpu->iowait_boost * max) >> SCHED_CAPACITY_SHIFT; in sugov_iowait_apply()
|
/kernel/linux/linux-6.6/kernel/sched/ |
H A D | pelt.c | 144 sa->runnable_sum += runnable * contrib << SCHED_CAPACITY_SHIFT; in accumulate_sum() 146 sa->util_sum += contrib << SCHED_CAPACITY_SHIFT; in accumulate_sum()
|
H A D | pelt.h | 128 * (LOAD_AVG_MAX - 1024 + rq->cfs.avg.period_contrib) << SCHED_CAPACITY_SHIFT; 135 u32 divider = ((LOAD_AVG_MAX - 1024) << SCHED_CAPACITY_SHIFT) - LOAD_AVG_MAX; in update_idle_rq_clock_pelt()
|
H A D | sched_avg.c | 173 sched_ravg_window >> SCHED_CAPACITY_SHIFT); in sched_get_cpu_util()
|
H A D | walt.h | 58 delta = (delta * capcurr) >> SCHED_CAPACITY_SHIFT; in scale_exec_time()
|
H A D | cpufreq_schedutil.c | 294 boost = (sg_cpu->iowait_boost * max_cap) >> SCHED_CAPACITY_SHIFT; in sugov_iowait_apply()
|
H A D | sched.h | 305 #define cap_scale(v, s) ((v)*(s) >> SCHED_CAPACITY_SHIFT) 3194 * (BW_SHIFT - SCHED_CAPACITY_SHIFT) and false otherwise. 3200 return cap >= p->dl.dl_density >> (BW_SHIFT - SCHED_CAPACITY_SHIFT); in dl_task_fits_capacity() 3946 util = rq->prev_runnable_sum << SCHED_CAPACITY_SHIFT; in cpu_util_freq_walt()
|
/kernel/linux/linux-5.10/drivers/base/ |
H A D | arch_topology.c | 53 scale = (cur_freq << SCHED_CAPACITY_SHIFT) / max_freq; in topology_set_freq_scale() 158 capacity = div64_u64(capacity << SCHED_CAPACITY_SHIFT, in topology_normalize_cpu_scale()
|
/kernel/linux/linux-6.6/drivers/cpufreq/ |
H A D | amd-pstate.c | 619 boost_ratio = div_u64(max_perf << SCHED_CAPACITY_SHIFT, in amd_get_max_freq() 622 max_freq = nominal_freq * boost_ratio >> SCHED_CAPACITY_SHIFT; in amd_get_max_freq() 656 lowest_nonlinear_ratio = div_u64(lowest_nonlinear_perf << SCHED_CAPACITY_SHIFT, in amd_get_lowest_nonlinear_freq() 659 lowest_nonlinear_freq = nominal_freq * lowest_nonlinear_ratio >> SCHED_CAPACITY_SHIFT; in amd_get_lowest_nonlinear_freq()
|
H A D | acpi-cpufreq.c | 664 return div_u64(highest_perf << SCHED_CAPACITY_SHIFT, nominal_perf); in get_max_boost_ratio() 839 policy->cpuinfo.max_freq = freq * max_boost_ratio >> SCHED_CAPACITY_SHIFT; in acpi_cpufreq_cpu_init()
|
/kernel/linux/linux-6.6/drivers/base/ |
H A D | arch_topology.c | 147 scale = (cur_freq << SCHED_CAPACITY_SHIFT) / max_freq; in topology_set_freq_scale() 289 capacity = div64_u64(capacity << SCHED_CAPACITY_SHIFT, in topology_normalize_cpu_scale()
|
/kernel/linux/linux-6.6/arch/x86/kernel/cpu/ |
H A D | aperfmperf.c | 357 if (check_shl_overflow(acnt, 2*SCHED_CAPACITY_SHIFT, &acnt)) in scale_freq_tick()
|
/kernel/linux/linux-5.10/kernel/sched/rtg/ |
H A D | frame_rtg.c | 671 frame_util = div_u64((prev_frame_load << SCHED_CAPACITY_SHIFT), in calc_prev_fake_load_util() 689 frame_util = div_u64((prev_frame_load << SCHED_CAPACITY_SHIFT), in calc_prev_frame_load_util() 760 SCHED_CAPACITY_SHIFT), frame_info->frame_time); in calc_frame_exec()
|
/kernel/linux/linux-6.6/kernel/sched/rtg/ |
H A D | frame_rtg.c | 671 frame_util = div_u64((prev_frame_load << SCHED_CAPACITY_SHIFT), in calc_prev_fake_load_util() 689 frame_util = div_u64((prev_frame_load << SCHED_CAPACITY_SHIFT), in calc_prev_frame_load_util() 760 SCHED_CAPACITY_SHIFT), frame_info->frame_time); in calc_frame_exec()
|
/kernel/linux/linux-5.10/drivers/cpufreq/ |
H A D | acpi-cpufreq.c | 662 return div_u64(highest_perf << SCHED_CAPACITY_SHIFT, nominal_perf); in get_max_boost_ratio() 837 policy->cpuinfo.max_freq = freq * max_boost_ratio >> SCHED_CAPACITY_SHIFT; in acpi_cpufreq_cpu_init()
|
/kernel/linux/linux-5.10/include/linux/ |
H A D | sched.h | 377 # define SCHED_CAPACITY_SHIFT SCHED_FIXEDPOINT_SHIFT macro 378 # define SCHED_CAPACITY_SCALE (1L << SCHED_CAPACITY_SHIFT)
|
/kernel/linux/linux-6.6/include/linux/ |
H A D | sched.h | 454 # define SCHED_CAPACITY_SHIFT SCHED_FIXEDPOINT_SHIFT macro 455 # define SCHED_CAPACITY_SCALE (1L << SCHED_CAPACITY_SHIFT)
|