/kernel/linux/linux-5.10/arch/h8300/lib/ |
H A D | mulsi3.S | 9 ; 16b * 16b = 372 states (worst case) 10 ; 32b * 32b = 724 states (worst case) 14 ; 16b * 16b = 48 states 15 ; 16b * 32b = 72 states 16 ; 32b * 32b = 92 states 21 mov.w r1,r2 ; ( 2 states) b * d 22 mulxu r0,er2 ; (22 states) 24 mov.w e0,r3 ; ( 2 states) a * d 25 beq L_skip1 ; ( 4 states) 26 mulxu r1,er3 ; (22 states) [all...] |
/kernel/linux/linux-5.10/drivers/cpuidle/governors/ |
H A D | teo.c | 11 * wakeups from idle states. Moreover, information about what happened in the 15 * which of the shallower idle states to choose. 98 * @states: Idle states data corresponding to this CPU. 105 struct teo_idle_state states[CPUIDLE_STATE_MAX]; member 131 u64 lat_ns = drv->states[dev->last_state_idx].exit_latency_ns; in teo_update() 154 * Decay the "early hits" metric for all of the states and find the in teo_update() 155 * states matching the sleep length and the measured idle duration. in teo_update() 158 unsigned int early_hits = cpu_data->states[i].early_hits; in teo_update() 160 cpu_data->states[ in teo_update() [all...] |
H A D | ladder.c | 40 struct ladder_device_state states[CPUIDLE_STATE_MAX]; member 55 ldev->states[old_idx].stats.promotion_count = 0; in ladder_do_selection() 56 ldev->states[old_idx].stats.demotion_count = 0; in ladder_do_selection() 72 int first_idx = drv->states[0].flags & CPUIDLE_FLAG_POLLING ? 1 : 0; in ladder_select_state() 82 last_state = &ldev->states[last_idx]; in ladder_select_state() 84 last_residency = dev->last_residency_ns - drv->states[last_idx].exit_latency_ns; in ladder_select_state() 90 drv->states[last_idx + 1].exit_latency_ns <= latency_req) { in ladder_select_state() 102 drv->states[last_idx].exit_latency_ns > latency_req)) { in ladder_select_state() 106 if (drv->states[i].exit_latency_ns <= latency_req) in ladder_select_state() 136 int first_idx = drv->states[ in ladder_enable_device() [all...] |
H A D | menu.c | 80 * C states, especially those with large exit latencies, can have a real 86 * The busier the system, the less impact of C states is acceptable 289 ((data->next_timer_ns < drv->states[1].target_residency_ns || in menu_select() 290 latency_req < drv->states[1].exit_latency_ns) && in menu_select() 297 *stop_tick = !(drv->states[0].flags & CPUIDLE_FLAG_POLLING); in menu_select() 339 struct cpuidle_state *s = &drv->states[i]; in menu_select() 352 if ((drv->states[idx].flags & CPUIDLE_FLAG_POLLING) && in menu_select() 369 predicted_ns = drv->states[idx].target_residency_ns; in menu_select() 379 if (drv->states[idx].target_residency_ns < TICK_NSEC && in menu_select() 392 idx = 0; /* No states enable in menu_select() [all...] |
/kernel/linux/linux-6.6/drivers/cpuidle/ |
H A D | dt_idle_genpd.c | 26 struct genpd_power_state *states, int state_count) in pd_parse_state_nodes() 32 ret = parse_state(to_of_node(states[i].fwnode), &state); in pd_parse_state_nodes() 42 states[i].data = state_buf; in pd_parse_state_nodes() 50 kfree(states[i].data); in pd_parse_state_nodes() 56 struct genpd_power_state **states, in pd_parse_states() 61 /* Parse the domain idle states. */ in pd_parse_states() 62 ret = of_genpd_parse_idle_states(np, states, state_count); in pd_parse_states() 67 ret = pd_parse_state_nodes(parse_state, *states, *state_count); in pd_parse_states() 69 kfree(*states); in pd_parse_states() 74 static void pd_free_states(struct genpd_power_state *states, in pd_free_states() argument 24 pd_parse_state_nodes( int (*parse_state)(struct device_node *, u32 *), struct genpd_power_state *states, int state_count) pd_parse_state_nodes() argument 54 pd_parse_states(struct device_node *np, int (*parse_state)(struct device_node *, u32 *), struct genpd_power_state **states, int *state_count) pd_parse_states() argument 95 struct genpd_power_state *states = NULL; dt_idle_pd_alloc() local [all...] |
H A D | cpuidle-riscv-sbi.c | 32 u32 *states; member 99 u32 *states = __this_cpu_read(sbi_cpuidle_data.states); in sbi_cpuidle_enter_state() local 100 u32 state = states[idx]; in sbi_cpuidle_enter_state() 114 u32 *states = data->states; in __sbi_enter_domain_idle_state() local 134 state = states[idx]; in __sbi_enter_domain_idle_state() 251 * of a shared state for the domain, assumes the domain states are all in sbi_dt_cpu_init_topology() 252 * deeper states. in sbi_dt_cpu_init_topology() 254 drv->states[state_coun in sbi_dt_cpu_init_topology() 271 u32 *states; sbi_cpuidle_dt_init_states() local [all...] |
H A D | cpuidle-mvebu-v7.c | 36 if (drv->states[index].flags & MVEBU_V7_FLAG_DEEP_IDLE) in mvebu_v7_enter_idle() 53 .states[0] = ARM_CPUIDLE_WFI_STATE, 54 .states[1] = { 63 .states[2] = { 77 .states[0] = ARM_CPUIDLE_WFI_STATE, 78 .states[1] = { 92 .states[0] = ARM_CPUIDLE_WFI_STATE, 93 .states[1] = {
|
H A D | cpuidle-psci.c | 56 u32 *states = data->psci_states; in __psci_enter_domain_idle_state() local 73 state = states[idx]; in __psci_enter_domain_idle_state() 233 * of a shared state for the domain, assumes the domain states are all in psci_dt_cpu_init_topology() 234 * deeper states. in psci_dt_cpu_init_topology() 236 drv->states[state_count - 1].flags |= CPUIDLE_FLAG_RCU_IDLE; in psci_dt_cpu_init_topology() 237 drv->states[state_count - 1].enter = psci_enter_domain_idle_state; in psci_dt_cpu_init_topology() 238 drv->states[state_count - 1].enter_s2idle = psci_enter_s2idle_domain_idle_state; in psci_dt_cpu_init_topology() 281 /* Idle states parsed correctly, store them in the per-cpu struct. */ in psci_dt_cpu_init_idle() 294 * idle states must not be enabled, so bail out in psci_cpu_init_idle() 350 * PSCI idle states relie in psci_idle_init_cpu() [all...] |
/kernel/linux/linux-5.10/drivers/cpuidle/ |
H A D | cpuidle-psci-domain.c | 34 struct genpd_power_state *state = &pd->states[pd->state_idx]; in psci_pd_power_off() 50 static int psci_pd_parse_state_nodes(struct genpd_power_state *states, in psci_pd_parse_state_nodes() argument 57 ret = psci_dt_parse_state_node(to_of_node(states[i].fwnode), in psci_pd_parse_state_nodes() 68 states[i].data = psci_state_buf; in psci_pd_parse_state_nodes() 76 kfree(states[i].data); in psci_pd_parse_state_nodes() 81 struct genpd_power_state **states, int *state_count) in psci_pd_parse_states() 85 /* Parse the domain idle states. */ in psci_pd_parse_states() 86 ret = of_genpd_parse_idle_states(np, states, state_count); in psci_pd_parse_states() 91 ret = psci_pd_parse_state_nodes(*states, *state_count); in psci_pd_parse_states() 93 kfree(*states); in psci_pd_parse_states() 80 psci_pd_parse_states(struct device_node *np, struct genpd_power_state **states, int *state_count) psci_pd_parse_states() argument 98 psci_pd_free_states(struct genpd_power_state *states, unsigned int state_count) psci_pd_free_states() argument 113 struct genpd_power_state *states = NULL; psci_pd_init() local [all...] |
H A D | cpuidle-mvebu-v7.c | 36 if (drv->states[index].flags & MVEBU_V7_FLAG_DEEP_IDLE) in mvebu_v7_enter_idle() 50 .states[0] = ARM_CPUIDLE_WFI_STATE, 51 .states[1] = { 59 .states[2] = { 73 .states[0] = ARM_CPUIDLE_WFI_STATE, 74 .states[1] = { 88 .states[0] = ARM_CPUIDLE_WFI_STATE, 89 .states[1] = {
|
/kernel/linux/linux-5.10/drivers/regulator/ |
H A D | gpio-regulator.c | 39 struct gpio_regulator_state *states; member 51 if (data->states[ptr].gpios == data->state) in gpio_regulator_get_value() 52 return data->states[ptr].value; in gpio_regulator_get_value() 65 if (data->states[ptr].value < best_val && in gpio_regulator_set_voltage() 66 data->states[ptr].value >= min_uV && in gpio_regulator_set_voltage() 67 data->states[ptr].value <= max_uV) { in gpio_regulator_set_voltage() 68 target = data->states[ptr].gpios; in gpio_regulator_set_voltage() 69 best_val = data->states[ptr].value; in gpio_regulator_set_voltage() 94 return data->states[selector].value; in gpio_regulator_list_voltage() 104 if (data->states[pt in gpio_regulator_set_current_limit() [all...] |
/kernel/linux/linux-6.6/drivers/cpuidle/governors/ |
H A D | ladder.c | 40 struct ladder_device_state states[CPUIDLE_STATE_MAX]; member 55 ldev->states[old_idx].stats.promotion_count = 0; in ladder_do_selection() 56 ldev->states[old_idx].stats.demotion_count = 0; in ladder_do_selection() 72 int first_idx = drv->states[0].flags & CPUIDLE_FLAG_POLLING ? 1 : 0; in ladder_select_state() 82 last_state = &ldev->states[last_idx]; in ladder_select_state() 84 last_residency = dev->last_residency_ns - drv->states[last_idx].exit_latency_ns; in ladder_select_state() 90 drv->states[last_idx + 1].exit_latency_ns <= latency_req) { in ladder_select_state() 102 drv->states[last_idx].exit_latency_ns > latency_req)) { in ladder_select_state() 106 if (drv->states[i].exit_latency_ns <= latency_req) in ladder_select_state() 136 int first_idx = drv->states[ in ladder_enable_device() [all...] |
H A D | teo.c | 20 * wakeups from idle states. Moreover, information about what happened in the 25 * states to choose instead of it. 40 * idle states provided by the %CPUIdle driver in the ascending order. That is, 75 * and all of the deeper idle states (it represents the cases in which the 79 * - The sum of the "intercepts" metrics for all of the idle states shallower 84 * - The sum of the numbers of recent intercepts for all of the idle states 91 * - Traverse the idle states shallower than the candidate one in the 95 * of the numbers of recent intercepts over all of the idle states between 243 u64 lat_ns = drv->states[dev->last_state_idx].exit_latency_ns; in teo_update() 280 target_residency_ns = drv->states[ in teo_update() [all...] |
H A D | menu.c | 82 * C states, especially those with large exit latencies, can have a real 88 * The busier the system, the less impact of C states is acceptable 315 ((data->next_timer_ns < drv->states[1].target_residency_ns || in menu_select() 316 latency_req < drv->states[1].exit_latency_ns) && in menu_select() 323 *stop_tick = !(drv->states[0].flags & CPUIDLE_FLAG_POLLING); in menu_select() 355 struct cpuidle_state *s = &drv->states[i]; in menu_select() 368 if ((drv->states[idx].flags & CPUIDLE_FLAG_POLLING) && in menu_select() 385 predicted_ns = drv->states[idx].target_residency_ns; in menu_select() 395 if (drv->states[idx].target_residency_ns < TICK_NSEC && in menu_select() 408 idx = 0; /* No states enable in menu_select() [all...] |
/kernel/linux/linux-6.6/tools/verification/dot2/ |
H A D | automata.py | 26 self.states, self.initial_state, self.final_states = self.__get_state_variables() 80 states = [] 96 states.append(state) 107 states = sorted(set(states)) 108 states.remove(initial_state) 110 # Insert the initial state at the bein og the states 111 states.insert(0, initial_state) 116 return states, initial_state, final_states 145 states [all...] |
H A D | dot2c.py | 20 enum_states_def = "states" 41 for state in self.states: 89 if self.states.__len__() > 255: 92 if self.states.__len__() > 65535: 95 if self.states.__len__() > 1000000: 96 raise Exception("Too many states: %d" % self.states.__len__()) 134 return self.__get_string_vector_per_line_content(self.states) 152 max_state_name = max(self.states, key = len).__len__() 160 nr_states = self.states [all...] |
/kernel/linux/linux-5.10/sound/usb/misc/ |
H A D | ua101.c | 66 /* bits in struct ua101::states */ 91 unsigned long states; member 159 if (test_and_clear_bit(USB_CAPTURE_RUNNING, &ua->states)) { in abort_usb_capture() 167 if (test_and_clear_bit(USB_PLAYBACK_RUNNING, &ua->states)) in abort_usb_playback() 186 if (test_bit(USB_PLAYBACK_RUNNING, &ua->states)) { in playback_urb_complete() 206 set_bit(PLAYBACK_URB_COMPLETED, &ua->states); in first_playback_urb_complete() 259 if (unlikely(!test_bit(USB_PLAYBACK_RUNNING, &ua->states))) in playback_work() 289 if (test_bit(ALSA_PLAYBACK_RUNNING, &ua->states)) in playback_work() 370 if (frames > 0 && test_bit(ALSA_CAPTURE_RUNNING, &ua->states)) in capture_urb_complete() 375 if (test_bit(USB_CAPTURE_RUNNING, &ua->states)) { in capture_urb_complete() [all...] |
/kernel/linux/linux-6.6/sound/usb/misc/ |
H A D | ua101.c | 65 /* bits in struct ua101::states */ 90 unsigned long states; member 158 if (test_and_clear_bit(USB_CAPTURE_RUNNING, &ua->states)) { in abort_usb_capture() 166 if (test_and_clear_bit(USB_PLAYBACK_RUNNING, &ua->states)) in abort_usb_playback() 185 if (test_bit(USB_PLAYBACK_RUNNING, &ua->states)) { in playback_urb_complete() 205 set_bit(PLAYBACK_URB_COMPLETED, &ua->states); in first_playback_urb_complete() 258 if (unlikely(!test_bit(USB_PLAYBACK_RUNNING, &ua->states))) in playback_work() 288 if (test_bit(ALSA_PLAYBACK_RUNNING, &ua->states)) in playback_work() 369 if (frames > 0 && test_bit(ALSA_CAPTURE_RUNNING, &ua->states)) in capture_urb_complete() 374 if (test_bit(USB_CAPTURE_RUNNING, &ua->states)) { in capture_urb_complete() [all...] |
/kernel/linux/linux-6.6/drivers/regulator/ |
H A D | irq_helpers.c | 107 stat = &rid->states[i]; in regulator_notifier_isr_work() 132 stat = &rid->states[i]; in regulator_notifier_isr_work() 214 rdev = rid->states[i].rdev; in regulator_notifier_isr() 240 stat = &rid->states[i]; in regulator_notifier_isr() 286 h->rdata.states = devm_kzalloc(dev, sizeof(*h->rdata.states) * in init_rdev_state() 288 if (!h->rdata.states) in init_rdev_state() 295 h->rdata.states[i].possible_errs = common_err; in init_rdev_state() 297 h->rdata.states[i].possible_errs |= *rdev_err++; in init_rdev_state() 298 h->rdata.states[ in init_rdev_state() [all...] |
H A D | gpio-regulator.c | 39 struct gpio_regulator_state *states; member 51 if (data->states[ptr].gpios == data->state) in gpio_regulator_get_value() 52 return data->states[ptr].value; in gpio_regulator_get_value() 65 if (data->states[ptr].value < best_val && in gpio_regulator_set_voltage() 66 data->states[ptr].value >= min_uV && in gpio_regulator_set_voltage() 67 data->states[ptr].value <= max_uV) { in gpio_regulator_set_voltage() 68 target = data->states[ptr].gpios; in gpio_regulator_set_voltage() 69 best_val = data->states[ptr].value; in gpio_regulator_set_voltage() 94 return data->states[selector].value; in gpio_regulator_list_voltage() 104 if (data->states[pt in gpio_regulator_set_current_limit() [all...] |
/kernel/linux/linux-5.10/drivers/cpufreq/ |
H A D | ia64-acpi-cpufreq.c | 97 if (value == data->acpi_data.states[i].status) in extract_clock() 98 return data->acpi_data.states[i].core_frequency; in extract_clock() 100 return data->acpi_data.states[i-1].core_frequency; in extract_clock() 159 value = (u32) data->acpi_data.states[state].control; in processor_set_freq() 254 if ((data->acpi_data.states[i].transition_latency * 1000) > in acpi_cpufreq_cpu_init() 257 data->acpi_data.states[i].transition_latency * 1000; in acpi_cpufreq_cpu_init() 266 data->acpi_data.states[i].core_frequency * 1000; in acpi_cpufreq_cpu_init() 282 (u32) data->acpi_data.states[i].core_frequency, in acpi_cpufreq_cpu_init() 283 (u32) data->acpi_data.states[i].power, in acpi_cpufreq_cpu_init() 284 (u32) data->acpi_data.states[ in acpi_cpufreq_cpu_init() [all...] |
/kernel/linux/linux-6.6/drivers/cpufreq/ |
H A D | ia64-acpi-cpufreq.c | 97 if (value == data->acpi_data.states[i].status) in extract_clock() 98 return data->acpi_data.states[i].core_frequency; in extract_clock() 100 return data->acpi_data.states[i-1].core_frequency; in extract_clock() 159 value = (u32) data->acpi_data.states[state].control; in processor_set_freq() 254 if ((data->acpi_data.states[i].transition_latency * 1000) > in acpi_cpufreq_cpu_init() 257 data->acpi_data.states[i].transition_latency * 1000; in acpi_cpufreq_cpu_init() 266 data->acpi_data.states[i].core_frequency * 1000; in acpi_cpufreq_cpu_init() 282 (u32) data->acpi_data.states[i].core_frequency, in acpi_cpufreq_cpu_init() 283 (u32) data->acpi_data.states[i].power, in acpi_cpufreq_cpu_init() 284 (u32) data->acpi_data.states[ in acpi_cpufreq_cpu_init() [all...] |
/kernel/linux/linux-5.10/drivers/gpu/drm/ |
H A D | drm_blend.c | 441 struct drm_plane_state **states; in drm_atomic_helper_crtc_normalize_zpos() local 449 states = kmalloc_array(total_planes, sizeof(*states), GFP_KERNEL); in drm_atomic_helper_crtc_normalize_zpos() 450 if (!states) in drm_atomic_helper_crtc_normalize_zpos() 454 * Normalization process might create new states for planes which in drm_atomic_helper_crtc_normalize_zpos() 464 states[n++] = plane_state; in drm_atomic_helper_crtc_normalize_zpos() 470 sort(states, n, sizeof(*states), drm_atomic_state_zpos_cmp, NULL); in drm_atomic_helper_crtc_normalize_zpos() 473 plane = states[i]->plane; in drm_atomic_helper_crtc_normalize_zpos() 475 states[ in drm_atomic_helper_crtc_normalize_zpos() [all...] |
/kernel/linux/linux-6.6/drivers/gpu/drm/ |
H A D | drm_blend.c | 448 struct drm_plane_state **states; in drm_atomic_helper_crtc_normalize_zpos() local 456 states = kmalloc_array(total_planes, sizeof(*states), GFP_KERNEL); in drm_atomic_helper_crtc_normalize_zpos() 457 if (!states) in drm_atomic_helper_crtc_normalize_zpos() 461 * Normalization process might create new states for planes which in drm_atomic_helper_crtc_normalize_zpos() 471 states[n++] = plane_state; in drm_atomic_helper_crtc_normalize_zpos() 476 sort(states, n, sizeof(*states), drm_atomic_state_zpos_cmp, NULL); in drm_atomic_helper_crtc_normalize_zpos() 479 plane = states[i]->plane; in drm_atomic_helper_crtc_normalize_zpos() 481 states[ in drm_atomic_helper_crtc_normalize_zpos() [all...] |
/kernel/linux/linux-5.10/arch/x86/kernel/acpi/ |
H A D | cstate.c | 5 * - Added _PDC for SMP C-states on Intel CPUs 106 } states[ACPI_PROCESSOR_MAX_POWER]; member 132 /* If the HW does not support any sub-states in this C-state */ in acpi_processor_ffh_cstate_probe_cpu() 174 percpu_entry->states[cx->index].eax = 0; in acpi_processor_ffh_cstate_probe() 175 percpu_entry->states[cx->index].ecx = 0; in acpi_processor_ffh_cstate_probe() 183 percpu_entry->states[cx->index].eax = cx->address; in acpi_processor_ffh_cstate_probe() 184 percpu_entry->states[cx->index].ecx = MWAIT_ECX_INTERRUPT_BREAK; in acpi_processor_ffh_cstate_probe() 205 mwait_idle_with_hints(percpu_entry->states[cx->index].eax, in acpi_processor_ffh_cstate_enter() 206 percpu_entry->states[cx->index].ecx); in acpi_processor_ffh_cstate_enter()
|