Home
last modified time | relevance | path

Searched refs:states (Results 1 - 25 of 344) sorted by relevance

12345678910>>...14

/kernel/linux/linux-5.10/arch/h8300/lib/
H A Dmulsi3.S9 ; 16b * 16b = 372 states (worst case)
10 ; 32b * 32b = 724 states (worst case)
14 ; 16b * 16b = 48 states
15 ; 16b * 32b = 72 states
16 ; 32b * 32b = 92 states
21 mov.w r1,r2 ; ( 2 states) b * d
22 mulxu r0,er2 ; (22 states)
24 mov.w e0,r3 ; ( 2 states) a * d
25 beq L_skip1 ; ( 4 states)
26 mulxu r1,er3 ; (22 states)
[all...]
/kernel/linux/linux-5.10/drivers/cpuidle/governors/
H A Dteo.c11 * wakeups from idle states. Moreover, information about what happened in the
15 * which of the shallower idle states to choose.
98 * @states: Idle states data corresponding to this CPU.
105 struct teo_idle_state states[CPUIDLE_STATE_MAX]; member
131 u64 lat_ns = drv->states[dev->last_state_idx].exit_latency_ns; in teo_update()
154 * Decay the "early hits" metric for all of the states and find the in teo_update()
155 * states matching the sleep length and the measured idle duration. in teo_update()
158 unsigned int early_hits = cpu_data->states[i].early_hits; in teo_update()
160 cpu_data->states[ in teo_update()
[all...]
H A Dladder.c40 struct ladder_device_state states[CPUIDLE_STATE_MAX]; member
55 ldev->states[old_idx].stats.promotion_count = 0; in ladder_do_selection()
56 ldev->states[old_idx].stats.demotion_count = 0; in ladder_do_selection()
72 int first_idx = drv->states[0].flags & CPUIDLE_FLAG_POLLING ? 1 : 0; in ladder_select_state()
82 last_state = &ldev->states[last_idx]; in ladder_select_state()
84 last_residency = dev->last_residency_ns - drv->states[last_idx].exit_latency_ns; in ladder_select_state()
90 drv->states[last_idx + 1].exit_latency_ns <= latency_req) { in ladder_select_state()
102 drv->states[last_idx].exit_latency_ns > latency_req)) { in ladder_select_state()
106 if (drv->states[i].exit_latency_ns <= latency_req) in ladder_select_state()
136 int first_idx = drv->states[ in ladder_enable_device()
[all...]
H A Dmenu.c80 * C states, especially those with large exit latencies, can have a real
86 * The busier the system, the less impact of C states is acceptable
289 ((data->next_timer_ns < drv->states[1].target_residency_ns || in menu_select()
290 latency_req < drv->states[1].exit_latency_ns) && in menu_select()
297 *stop_tick = !(drv->states[0].flags & CPUIDLE_FLAG_POLLING); in menu_select()
339 struct cpuidle_state *s = &drv->states[i]; in menu_select()
352 if ((drv->states[idx].flags & CPUIDLE_FLAG_POLLING) && in menu_select()
369 predicted_ns = drv->states[idx].target_residency_ns; in menu_select()
379 if (drv->states[idx].target_residency_ns < TICK_NSEC && in menu_select()
392 idx = 0; /* No states enable in menu_select()
[all...]
/kernel/linux/linux-6.6/drivers/cpuidle/
H A Ddt_idle_genpd.c26 struct genpd_power_state *states, int state_count) in pd_parse_state_nodes()
32 ret = parse_state(to_of_node(states[i].fwnode), &state); in pd_parse_state_nodes()
42 states[i].data = state_buf; in pd_parse_state_nodes()
50 kfree(states[i].data); in pd_parse_state_nodes()
56 struct genpd_power_state **states, in pd_parse_states()
61 /* Parse the domain idle states. */ in pd_parse_states()
62 ret = of_genpd_parse_idle_states(np, states, state_count); in pd_parse_states()
67 ret = pd_parse_state_nodes(parse_state, *states, *state_count); in pd_parse_states()
69 kfree(*states); in pd_parse_states()
74 static void pd_free_states(struct genpd_power_state *states, in pd_free_states() argument
24 pd_parse_state_nodes( int (*parse_state)(struct device_node *, u32 *), struct genpd_power_state *states, int state_count) pd_parse_state_nodes() argument
54 pd_parse_states(struct device_node *np, int (*parse_state)(struct device_node *, u32 *), struct genpd_power_state **states, int *state_count) pd_parse_states() argument
95 struct genpd_power_state *states = NULL; dt_idle_pd_alloc() local
[all...]
H A Dcpuidle-riscv-sbi.c32 u32 *states; member
99 u32 *states = __this_cpu_read(sbi_cpuidle_data.states); in sbi_cpuidle_enter_state() local
100 u32 state = states[idx]; in sbi_cpuidle_enter_state()
114 u32 *states = data->states; in __sbi_enter_domain_idle_state() local
134 state = states[idx]; in __sbi_enter_domain_idle_state()
251 * of a shared state for the domain, assumes the domain states are all in sbi_dt_cpu_init_topology()
252 * deeper states. in sbi_dt_cpu_init_topology()
254 drv->states[state_coun in sbi_dt_cpu_init_topology()
271 u32 *states; sbi_cpuidle_dt_init_states() local
[all...]
H A Dcpuidle-mvebu-v7.c36 if (drv->states[index].flags & MVEBU_V7_FLAG_DEEP_IDLE) in mvebu_v7_enter_idle()
53 .states[0] = ARM_CPUIDLE_WFI_STATE,
54 .states[1] = {
63 .states[2] = {
77 .states[0] = ARM_CPUIDLE_WFI_STATE,
78 .states[1] = {
92 .states[0] = ARM_CPUIDLE_WFI_STATE,
93 .states[1] = {
H A Dcpuidle-psci.c56 u32 *states = data->psci_states; in __psci_enter_domain_idle_state() local
73 state = states[idx]; in __psci_enter_domain_idle_state()
233 * of a shared state for the domain, assumes the domain states are all in psci_dt_cpu_init_topology()
234 * deeper states. in psci_dt_cpu_init_topology()
236 drv->states[state_count - 1].flags |= CPUIDLE_FLAG_RCU_IDLE; in psci_dt_cpu_init_topology()
237 drv->states[state_count - 1].enter = psci_enter_domain_idle_state; in psci_dt_cpu_init_topology()
238 drv->states[state_count - 1].enter_s2idle = psci_enter_s2idle_domain_idle_state; in psci_dt_cpu_init_topology()
281 /* Idle states parsed correctly, store them in the per-cpu struct. */ in psci_dt_cpu_init_idle()
294 * idle states must not be enabled, so bail out in psci_cpu_init_idle()
350 * PSCI idle states relie in psci_idle_init_cpu()
[all...]
/kernel/linux/linux-5.10/drivers/cpuidle/
H A Dcpuidle-psci-domain.c34 struct genpd_power_state *state = &pd->states[pd->state_idx]; in psci_pd_power_off()
50 static int psci_pd_parse_state_nodes(struct genpd_power_state *states, in psci_pd_parse_state_nodes() argument
57 ret = psci_dt_parse_state_node(to_of_node(states[i].fwnode), in psci_pd_parse_state_nodes()
68 states[i].data = psci_state_buf; in psci_pd_parse_state_nodes()
76 kfree(states[i].data); in psci_pd_parse_state_nodes()
81 struct genpd_power_state **states, int *state_count) in psci_pd_parse_states()
85 /* Parse the domain idle states. */ in psci_pd_parse_states()
86 ret = of_genpd_parse_idle_states(np, states, state_count); in psci_pd_parse_states()
91 ret = psci_pd_parse_state_nodes(*states, *state_count); in psci_pd_parse_states()
93 kfree(*states); in psci_pd_parse_states()
80 psci_pd_parse_states(struct device_node *np, struct genpd_power_state **states, int *state_count) psci_pd_parse_states() argument
98 psci_pd_free_states(struct genpd_power_state *states, unsigned int state_count) psci_pd_free_states() argument
113 struct genpd_power_state *states = NULL; psci_pd_init() local
[all...]
H A Dcpuidle-mvebu-v7.c36 if (drv->states[index].flags & MVEBU_V7_FLAG_DEEP_IDLE) in mvebu_v7_enter_idle()
50 .states[0] = ARM_CPUIDLE_WFI_STATE,
51 .states[1] = {
59 .states[2] = {
73 .states[0] = ARM_CPUIDLE_WFI_STATE,
74 .states[1] = {
88 .states[0] = ARM_CPUIDLE_WFI_STATE,
89 .states[1] = {
/kernel/linux/linux-5.10/drivers/regulator/
H A Dgpio-regulator.c39 struct gpio_regulator_state *states; member
51 if (data->states[ptr].gpios == data->state) in gpio_regulator_get_value()
52 return data->states[ptr].value; in gpio_regulator_get_value()
65 if (data->states[ptr].value < best_val && in gpio_regulator_set_voltage()
66 data->states[ptr].value >= min_uV && in gpio_regulator_set_voltage()
67 data->states[ptr].value <= max_uV) { in gpio_regulator_set_voltage()
68 target = data->states[ptr].gpios; in gpio_regulator_set_voltage()
69 best_val = data->states[ptr].value; in gpio_regulator_set_voltage()
94 return data->states[selector].value; in gpio_regulator_list_voltage()
104 if (data->states[pt in gpio_regulator_set_current_limit()
[all...]
/kernel/linux/linux-6.6/drivers/cpuidle/governors/
H A Dladder.c40 struct ladder_device_state states[CPUIDLE_STATE_MAX]; member
55 ldev->states[old_idx].stats.promotion_count = 0; in ladder_do_selection()
56 ldev->states[old_idx].stats.demotion_count = 0; in ladder_do_selection()
72 int first_idx = drv->states[0].flags & CPUIDLE_FLAG_POLLING ? 1 : 0; in ladder_select_state()
82 last_state = &ldev->states[last_idx]; in ladder_select_state()
84 last_residency = dev->last_residency_ns - drv->states[last_idx].exit_latency_ns; in ladder_select_state()
90 drv->states[last_idx + 1].exit_latency_ns <= latency_req) { in ladder_select_state()
102 drv->states[last_idx].exit_latency_ns > latency_req)) { in ladder_select_state()
106 if (drv->states[i].exit_latency_ns <= latency_req) in ladder_select_state()
136 int first_idx = drv->states[ in ladder_enable_device()
[all...]
H A Dteo.c20 * wakeups from idle states. Moreover, information about what happened in the
25 * states to choose instead of it.
40 * idle states provided by the %CPUIdle driver in the ascending order. That is,
75 * and all of the deeper idle states (it represents the cases in which the
79 * - The sum of the "intercepts" metrics for all of the idle states shallower
84 * - The sum of the numbers of recent intercepts for all of the idle states
91 * - Traverse the idle states shallower than the candidate one in the
95 * of the numbers of recent intercepts over all of the idle states between
243 u64 lat_ns = drv->states[dev->last_state_idx].exit_latency_ns; in teo_update()
280 target_residency_ns = drv->states[ in teo_update()
[all...]
H A Dmenu.c82 * C states, especially those with large exit latencies, can have a real
88 * The busier the system, the less impact of C states is acceptable
315 ((data->next_timer_ns < drv->states[1].target_residency_ns || in menu_select()
316 latency_req < drv->states[1].exit_latency_ns) && in menu_select()
323 *stop_tick = !(drv->states[0].flags & CPUIDLE_FLAG_POLLING); in menu_select()
355 struct cpuidle_state *s = &drv->states[i]; in menu_select()
368 if ((drv->states[idx].flags & CPUIDLE_FLAG_POLLING) && in menu_select()
385 predicted_ns = drv->states[idx].target_residency_ns; in menu_select()
395 if (drv->states[idx].target_residency_ns < TICK_NSEC && in menu_select()
408 idx = 0; /* No states enable in menu_select()
[all...]
/kernel/linux/linux-6.6/tools/verification/dot2/
H A Dautomata.py26 self.states, self.initial_state, self.final_states = self.__get_state_variables()
80 states = []
96 states.append(state)
107 states = sorted(set(states))
108 states.remove(initial_state)
110 # Insert the initial state at the bein og the states
111 states.insert(0, initial_state)
116 return states, initial_state, final_states
145 states
[all...]
H A Ddot2c.py20 enum_states_def = "states"
41 for state in self.states:
89 if self.states.__len__() > 255:
92 if self.states.__len__() > 65535:
95 if self.states.__len__() > 1000000:
96 raise Exception("Too many states: %d" % self.states.__len__())
134 return self.__get_string_vector_per_line_content(self.states)
152 max_state_name = max(self.states, key = len).__len__()
160 nr_states = self.states
[all...]
/kernel/linux/linux-5.10/sound/usb/misc/
H A Dua101.c66 /* bits in struct ua101::states */
91 unsigned long states; member
159 if (test_and_clear_bit(USB_CAPTURE_RUNNING, &ua->states)) { in abort_usb_capture()
167 if (test_and_clear_bit(USB_PLAYBACK_RUNNING, &ua->states)) in abort_usb_playback()
186 if (test_bit(USB_PLAYBACK_RUNNING, &ua->states)) { in playback_urb_complete()
206 set_bit(PLAYBACK_URB_COMPLETED, &ua->states); in first_playback_urb_complete()
259 if (unlikely(!test_bit(USB_PLAYBACK_RUNNING, &ua->states))) in playback_work()
289 if (test_bit(ALSA_PLAYBACK_RUNNING, &ua->states)) in playback_work()
370 if (frames > 0 && test_bit(ALSA_CAPTURE_RUNNING, &ua->states)) in capture_urb_complete()
375 if (test_bit(USB_CAPTURE_RUNNING, &ua->states)) { in capture_urb_complete()
[all...]
/kernel/linux/linux-6.6/sound/usb/misc/
H A Dua101.c65 /* bits in struct ua101::states */
90 unsigned long states; member
158 if (test_and_clear_bit(USB_CAPTURE_RUNNING, &ua->states)) { in abort_usb_capture()
166 if (test_and_clear_bit(USB_PLAYBACK_RUNNING, &ua->states)) in abort_usb_playback()
185 if (test_bit(USB_PLAYBACK_RUNNING, &ua->states)) { in playback_urb_complete()
205 set_bit(PLAYBACK_URB_COMPLETED, &ua->states); in first_playback_urb_complete()
258 if (unlikely(!test_bit(USB_PLAYBACK_RUNNING, &ua->states))) in playback_work()
288 if (test_bit(ALSA_PLAYBACK_RUNNING, &ua->states)) in playback_work()
369 if (frames > 0 && test_bit(ALSA_CAPTURE_RUNNING, &ua->states)) in capture_urb_complete()
374 if (test_bit(USB_CAPTURE_RUNNING, &ua->states)) { in capture_urb_complete()
[all...]
/kernel/linux/linux-6.6/drivers/regulator/
H A Dirq_helpers.c107 stat = &rid->states[i]; in regulator_notifier_isr_work()
132 stat = &rid->states[i]; in regulator_notifier_isr_work()
214 rdev = rid->states[i].rdev; in regulator_notifier_isr()
240 stat = &rid->states[i]; in regulator_notifier_isr()
286 h->rdata.states = devm_kzalloc(dev, sizeof(*h->rdata.states) * in init_rdev_state()
288 if (!h->rdata.states) in init_rdev_state()
295 h->rdata.states[i].possible_errs = common_err; in init_rdev_state()
297 h->rdata.states[i].possible_errs |= *rdev_err++; in init_rdev_state()
298 h->rdata.states[ in init_rdev_state()
[all...]
H A Dgpio-regulator.c39 struct gpio_regulator_state *states; member
51 if (data->states[ptr].gpios == data->state) in gpio_regulator_get_value()
52 return data->states[ptr].value; in gpio_regulator_get_value()
65 if (data->states[ptr].value < best_val && in gpio_regulator_set_voltage()
66 data->states[ptr].value >= min_uV && in gpio_regulator_set_voltage()
67 data->states[ptr].value <= max_uV) { in gpio_regulator_set_voltage()
68 target = data->states[ptr].gpios; in gpio_regulator_set_voltage()
69 best_val = data->states[ptr].value; in gpio_regulator_set_voltage()
94 return data->states[selector].value; in gpio_regulator_list_voltage()
104 if (data->states[pt in gpio_regulator_set_current_limit()
[all...]
/kernel/linux/linux-5.10/drivers/cpufreq/
H A Dia64-acpi-cpufreq.c97 if (value == data->acpi_data.states[i].status) in extract_clock()
98 return data->acpi_data.states[i].core_frequency; in extract_clock()
100 return data->acpi_data.states[i-1].core_frequency; in extract_clock()
159 value = (u32) data->acpi_data.states[state].control; in processor_set_freq()
254 if ((data->acpi_data.states[i].transition_latency * 1000) > in acpi_cpufreq_cpu_init()
257 data->acpi_data.states[i].transition_latency * 1000; in acpi_cpufreq_cpu_init()
266 data->acpi_data.states[i].core_frequency * 1000; in acpi_cpufreq_cpu_init()
282 (u32) data->acpi_data.states[i].core_frequency, in acpi_cpufreq_cpu_init()
283 (u32) data->acpi_data.states[i].power, in acpi_cpufreq_cpu_init()
284 (u32) data->acpi_data.states[ in acpi_cpufreq_cpu_init()
[all...]
/kernel/linux/linux-6.6/drivers/cpufreq/
H A Dia64-acpi-cpufreq.c97 if (value == data->acpi_data.states[i].status) in extract_clock()
98 return data->acpi_data.states[i].core_frequency; in extract_clock()
100 return data->acpi_data.states[i-1].core_frequency; in extract_clock()
159 value = (u32) data->acpi_data.states[state].control; in processor_set_freq()
254 if ((data->acpi_data.states[i].transition_latency * 1000) > in acpi_cpufreq_cpu_init()
257 data->acpi_data.states[i].transition_latency * 1000; in acpi_cpufreq_cpu_init()
266 data->acpi_data.states[i].core_frequency * 1000; in acpi_cpufreq_cpu_init()
282 (u32) data->acpi_data.states[i].core_frequency, in acpi_cpufreq_cpu_init()
283 (u32) data->acpi_data.states[i].power, in acpi_cpufreq_cpu_init()
284 (u32) data->acpi_data.states[ in acpi_cpufreq_cpu_init()
[all...]
/kernel/linux/linux-5.10/drivers/gpu/drm/
H A Ddrm_blend.c441 struct drm_plane_state **states; in drm_atomic_helper_crtc_normalize_zpos() local
449 states = kmalloc_array(total_planes, sizeof(*states), GFP_KERNEL); in drm_atomic_helper_crtc_normalize_zpos()
450 if (!states) in drm_atomic_helper_crtc_normalize_zpos()
454 * Normalization process might create new states for planes which in drm_atomic_helper_crtc_normalize_zpos()
464 states[n++] = plane_state; in drm_atomic_helper_crtc_normalize_zpos()
470 sort(states, n, sizeof(*states), drm_atomic_state_zpos_cmp, NULL); in drm_atomic_helper_crtc_normalize_zpos()
473 plane = states[i]->plane; in drm_atomic_helper_crtc_normalize_zpos()
475 states[ in drm_atomic_helper_crtc_normalize_zpos()
[all...]
/kernel/linux/linux-6.6/drivers/gpu/drm/
H A Ddrm_blend.c448 struct drm_plane_state **states; in drm_atomic_helper_crtc_normalize_zpos() local
456 states = kmalloc_array(total_planes, sizeof(*states), GFP_KERNEL); in drm_atomic_helper_crtc_normalize_zpos()
457 if (!states) in drm_atomic_helper_crtc_normalize_zpos()
461 * Normalization process might create new states for planes which in drm_atomic_helper_crtc_normalize_zpos()
471 states[n++] = plane_state; in drm_atomic_helper_crtc_normalize_zpos()
476 sort(states, n, sizeof(*states), drm_atomic_state_zpos_cmp, NULL); in drm_atomic_helper_crtc_normalize_zpos()
479 plane = states[i]->plane; in drm_atomic_helper_crtc_normalize_zpos()
481 states[ in drm_atomic_helper_crtc_normalize_zpos()
[all...]
/kernel/linux/linux-5.10/arch/x86/kernel/acpi/
H A Dcstate.c5 * - Added _PDC for SMP C-states on Intel CPUs
106 } states[ACPI_PROCESSOR_MAX_POWER]; member
132 /* If the HW does not support any sub-states in this C-state */ in acpi_processor_ffh_cstate_probe_cpu()
174 percpu_entry->states[cx->index].eax = 0; in acpi_processor_ffh_cstate_probe()
175 percpu_entry->states[cx->index].ecx = 0; in acpi_processor_ffh_cstate_probe()
183 percpu_entry->states[cx->index].eax = cx->address; in acpi_processor_ffh_cstate_probe()
184 percpu_entry->states[cx->index].ecx = MWAIT_ECX_INTERRUPT_BREAK; in acpi_processor_ffh_cstate_probe()
205 mwait_idle_with_hints(percpu_entry->states[cx->index].eax, in acpi_processor_ffh_cstate_enter()
206 percpu_entry->states[cx->index].ecx); in acpi_processor_ffh_cstate_enter()

Completed in 14 milliseconds

12345678910>>...14