1// SPDX-License-Identifier: (GPL-2.0-only OR BSD-3-Clause)
2//
3// This file is provided under a dual BSD/GPLv2 license.  When using or
4// redistributing this file, you may do so under either license.
5//
6// Copyright(c) 2018 Intel Corporation. All rights reserved.
7//
8// Authors: Liam Girdwood <liam.r.girdwood@linux.intel.com>
9//	    Ranjani Sridharan <ranjani.sridharan@linux.intel.com>
10//	    Rander Wang <rander.wang@intel.com>
11//          Keyon Jie <yang.jie@linux.intel.com>
12//
13
14/*
15 * Hardware interface for generic Intel audio DSP HDA IP
16 */
17
18#include <linux/module.h>
19#include <sound/hdaudio_ext.h>
20#include <sound/hda_register.h>
21#include <sound/hda-mlink.h>
22#include <trace/events/sof_intel.h>
23#include "../sof-audio.h"
24#include "../ops.h"
25#include "hda.h"
26#include "hda-ipc.h"
27
28static bool hda_enable_trace_D0I3_S0;
29#if IS_ENABLED(CONFIG_SND_SOC_SOF_DEBUG)
30module_param_named(enable_trace_D0I3_S0, hda_enable_trace_D0I3_S0, bool, 0444);
31MODULE_PARM_DESC(enable_trace_D0I3_S0,
32		 "SOF HDA enable trace when the DSP is in D0I3 in S0");
33#endif
34
35/*
36 * DSP Core control.
37 */
38
39static int hda_dsp_core_reset_enter(struct snd_sof_dev *sdev, unsigned int core_mask)
40{
41	u32 adspcs;
42	u32 reset;
43	int ret;
44
45	/* set reset bits for cores */
46	reset = HDA_DSP_ADSPCS_CRST_MASK(core_mask);
47	snd_sof_dsp_update_bits_unlocked(sdev, HDA_DSP_BAR,
48					 HDA_DSP_REG_ADSPCS,
49					 reset, reset);
50
51	/* poll with timeout to check if operation successful */
52	ret = snd_sof_dsp_read_poll_timeout(sdev, HDA_DSP_BAR,
53					HDA_DSP_REG_ADSPCS, adspcs,
54					((adspcs & reset) == reset),
55					HDA_DSP_REG_POLL_INTERVAL_US,
56					HDA_DSP_RESET_TIMEOUT_US);
57	if (ret < 0) {
58		dev_err(sdev->dev,
59			"error: %s: timeout on HDA_DSP_REG_ADSPCS read\n",
60			__func__);
61		return ret;
62	}
63
64	/* has core entered reset ? */
65	adspcs = snd_sof_dsp_read(sdev, HDA_DSP_BAR,
66				  HDA_DSP_REG_ADSPCS);
67	if ((adspcs & HDA_DSP_ADSPCS_CRST_MASK(core_mask)) !=
68		HDA_DSP_ADSPCS_CRST_MASK(core_mask)) {
69		dev_err(sdev->dev,
70			"error: reset enter failed: core_mask %x adspcs 0x%x\n",
71			core_mask, adspcs);
72		ret = -EIO;
73	}
74
75	return ret;
76}
77
78static int hda_dsp_core_reset_leave(struct snd_sof_dev *sdev, unsigned int core_mask)
79{
80	unsigned int crst;
81	u32 adspcs;
82	int ret;
83
84	/* clear reset bits for cores */
85	snd_sof_dsp_update_bits_unlocked(sdev, HDA_DSP_BAR,
86					 HDA_DSP_REG_ADSPCS,
87					 HDA_DSP_ADSPCS_CRST_MASK(core_mask),
88					 0);
89
90	/* poll with timeout to check if operation successful */
91	crst = HDA_DSP_ADSPCS_CRST_MASK(core_mask);
92	ret = snd_sof_dsp_read_poll_timeout(sdev, HDA_DSP_BAR,
93					    HDA_DSP_REG_ADSPCS, adspcs,
94					    !(adspcs & crst),
95					    HDA_DSP_REG_POLL_INTERVAL_US,
96					    HDA_DSP_RESET_TIMEOUT_US);
97
98	if (ret < 0) {
99		dev_err(sdev->dev,
100			"error: %s: timeout on HDA_DSP_REG_ADSPCS read\n",
101			__func__);
102		return ret;
103	}
104
105	/* has core left reset ? */
106	adspcs = snd_sof_dsp_read(sdev, HDA_DSP_BAR,
107				  HDA_DSP_REG_ADSPCS);
108	if ((adspcs & HDA_DSP_ADSPCS_CRST_MASK(core_mask)) != 0) {
109		dev_err(sdev->dev,
110			"error: reset leave failed: core_mask %x adspcs 0x%x\n",
111			core_mask, adspcs);
112		ret = -EIO;
113	}
114
115	return ret;
116}
117
118int hda_dsp_core_stall_reset(struct snd_sof_dev *sdev, unsigned int core_mask)
119{
120	/* stall core */
121	snd_sof_dsp_update_bits_unlocked(sdev, HDA_DSP_BAR,
122					 HDA_DSP_REG_ADSPCS,
123					 HDA_DSP_ADSPCS_CSTALL_MASK(core_mask),
124					 HDA_DSP_ADSPCS_CSTALL_MASK(core_mask));
125
126	/* set reset state */
127	return hda_dsp_core_reset_enter(sdev, core_mask);
128}
129
130bool hda_dsp_core_is_enabled(struct snd_sof_dev *sdev, unsigned int core_mask)
131{
132	int val;
133	bool is_enable;
134
135	val = snd_sof_dsp_read(sdev, HDA_DSP_BAR, HDA_DSP_REG_ADSPCS);
136
137#define MASK_IS_EQUAL(v, m, field) ({	\
138	u32 _m = field(m);		\
139	((v) & _m) == _m;		\
140})
141
142	is_enable = MASK_IS_EQUAL(val, core_mask, HDA_DSP_ADSPCS_CPA_MASK) &&
143		MASK_IS_EQUAL(val, core_mask, HDA_DSP_ADSPCS_SPA_MASK) &&
144		!(val & HDA_DSP_ADSPCS_CRST_MASK(core_mask)) &&
145		!(val & HDA_DSP_ADSPCS_CSTALL_MASK(core_mask));
146
147#undef MASK_IS_EQUAL
148
149	dev_dbg(sdev->dev, "DSP core(s) enabled? %d : core_mask %x\n",
150		is_enable, core_mask);
151
152	return is_enable;
153}
154
155int hda_dsp_core_run(struct snd_sof_dev *sdev, unsigned int core_mask)
156{
157	int ret;
158
159	/* leave reset state */
160	ret = hda_dsp_core_reset_leave(sdev, core_mask);
161	if (ret < 0)
162		return ret;
163
164	/* run core */
165	dev_dbg(sdev->dev, "unstall/run core: core_mask = %x\n", core_mask);
166	snd_sof_dsp_update_bits_unlocked(sdev, HDA_DSP_BAR,
167					 HDA_DSP_REG_ADSPCS,
168					 HDA_DSP_ADSPCS_CSTALL_MASK(core_mask),
169					 0);
170
171	/* is core now running ? */
172	if (!hda_dsp_core_is_enabled(sdev, core_mask)) {
173		hda_dsp_core_stall_reset(sdev, core_mask);
174		dev_err(sdev->dev, "error: DSP start core failed: core_mask %x\n",
175			core_mask);
176		ret = -EIO;
177	}
178
179	return ret;
180}
181
182/*
183 * Power Management.
184 */
185
186int hda_dsp_core_power_up(struct snd_sof_dev *sdev, unsigned int core_mask)
187{
188	struct sof_intel_hda_dev *hda = sdev->pdata->hw_pdata;
189	const struct sof_intel_dsp_desc *chip = hda->desc;
190	unsigned int cpa;
191	u32 adspcs;
192	int ret;
193
194	/* restrict core_mask to host managed cores mask */
195	core_mask &= chip->host_managed_cores_mask;
196	/* return if core_mask is not valid */
197	if (!core_mask)
198		return 0;
199
200	/* update bits */
201	snd_sof_dsp_update_bits(sdev, HDA_DSP_BAR, HDA_DSP_REG_ADSPCS,
202				HDA_DSP_ADSPCS_SPA_MASK(core_mask),
203				HDA_DSP_ADSPCS_SPA_MASK(core_mask));
204
205	/* poll with timeout to check if operation successful */
206	cpa = HDA_DSP_ADSPCS_CPA_MASK(core_mask);
207	ret = snd_sof_dsp_read_poll_timeout(sdev, HDA_DSP_BAR,
208					    HDA_DSP_REG_ADSPCS, adspcs,
209					    (adspcs & cpa) == cpa,
210					    HDA_DSP_REG_POLL_INTERVAL_US,
211					    HDA_DSP_RESET_TIMEOUT_US);
212	if (ret < 0) {
213		dev_err(sdev->dev,
214			"error: %s: timeout on HDA_DSP_REG_ADSPCS read\n",
215			__func__);
216		return ret;
217	}
218
219	/* did core power up ? */
220	adspcs = snd_sof_dsp_read(sdev, HDA_DSP_BAR,
221				  HDA_DSP_REG_ADSPCS);
222	if ((adspcs & HDA_DSP_ADSPCS_CPA_MASK(core_mask)) !=
223		HDA_DSP_ADSPCS_CPA_MASK(core_mask)) {
224		dev_err(sdev->dev,
225			"error: power up core failed core_mask %xadspcs 0x%x\n",
226			core_mask, adspcs);
227		ret = -EIO;
228	}
229
230	return ret;
231}
232
233static int hda_dsp_core_power_down(struct snd_sof_dev *sdev, unsigned int core_mask)
234{
235	u32 adspcs;
236	int ret;
237
238	/* update bits */
239	snd_sof_dsp_update_bits_unlocked(sdev, HDA_DSP_BAR,
240					 HDA_DSP_REG_ADSPCS,
241					 HDA_DSP_ADSPCS_SPA_MASK(core_mask), 0);
242
243	ret = snd_sof_dsp_read_poll_timeout(sdev, HDA_DSP_BAR,
244				HDA_DSP_REG_ADSPCS, adspcs,
245				!(adspcs & HDA_DSP_ADSPCS_CPA_MASK(core_mask)),
246				HDA_DSP_REG_POLL_INTERVAL_US,
247				HDA_DSP_PD_TIMEOUT * USEC_PER_MSEC);
248	if (ret < 0)
249		dev_err(sdev->dev,
250			"error: %s: timeout on HDA_DSP_REG_ADSPCS read\n",
251			__func__);
252
253	return ret;
254}
255
256int hda_dsp_enable_core(struct snd_sof_dev *sdev, unsigned int core_mask)
257{
258	struct sof_intel_hda_dev *hda = sdev->pdata->hw_pdata;
259	const struct sof_intel_dsp_desc *chip = hda->desc;
260	int ret;
261
262	/* restrict core_mask to host managed cores mask */
263	core_mask &= chip->host_managed_cores_mask;
264
265	/* return if core_mask is not valid or cores are already enabled */
266	if (!core_mask || hda_dsp_core_is_enabled(sdev, core_mask))
267		return 0;
268
269	/* power up */
270	ret = hda_dsp_core_power_up(sdev, core_mask);
271	if (ret < 0) {
272		dev_err(sdev->dev, "error: dsp core power up failed: core_mask %x\n",
273			core_mask);
274		return ret;
275	}
276
277	return hda_dsp_core_run(sdev, core_mask);
278}
279
280int hda_dsp_core_reset_power_down(struct snd_sof_dev *sdev,
281				  unsigned int core_mask)
282{
283	struct sof_intel_hda_dev *hda = sdev->pdata->hw_pdata;
284	const struct sof_intel_dsp_desc *chip = hda->desc;
285	int ret;
286
287	/* restrict core_mask to host managed cores mask */
288	core_mask &= chip->host_managed_cores_mask;
289
290	/* return if core_mask is not valid */
291	if (!core_mask)
292		return 0;
293
294	/* place core in reset prior to power down */
295	ret = hda_dsp_core_stall_reset(sdev, core_mask);
296	if (ret < 0) {
297		dev_err(sdev->dev, "error: dsp core reset failed: core_mask %x\n",
298			core_mask);
299		return ret;
300	}
301
302	/* power down core */
303	ret = hda_dsp_core_power_down(sdev, core_mask);
304	if (ret < 0) {
305		dev_err(sdev->dev, "error: dsp core power down fail mask %x: %d\n",
306			core_mask, ret);
307		return ret;
308	}
309
310	/* make sure we are in OFF state */
311	if (hda_dsp_core_is_enabled(sdev, core_mask)) {
312		dev_err(sdev->dev, "error: dsp core disable fail mask %x: %d\n",
313			core_mask, ret);
314		ret = -EIO;
315	}
316
317	return ret;
318}
319
320void hda_dsp_ipc_int_enable(struct snd_sof_dev *sdev)
321{
322	struct sof_intel_hda_dev *hda = sdev->pdata->hw_pdata;
323	const struct sof_intel_dsp_desc *chip = hda->desc;
324
325	if (sdev->dspless_mode_selected)
326		return;
327
328	/* enable IPC DONE and BUSY interrupts */
329	snd_sof_dsp_update_bits(sdev, HDA_DSP_BAR, chip->ipc_ctl,
330			HDA_DSP_REG_HIPCCTL_DONE | HDA_DSP_REG_HIPCCTL_BUSY,
331			HDA_DSP_REG_HIPCCTL_DONE | HDA_DSP_REG_HIPCCTL_BUSY);
332
333	/* enable IPC interrupt */
334	snd_sof_dsp_update_bits(sdev, HDA_DSP_BAR, HDA_DSP_REG_ADSPIC,
335				HDA_DSP_ADSPIC_IPC, HDA_DSP_ADSPIC_IPC);
336}
337
338void hda_dsp_ipc_int_disable(struct snd_sof_dev *sdev)
339{
340	struct sof_intel_hda_dev *hda = sdev->pdata->hw_pdata;
341	const struct sof_intel_dsp_desc *chip = hda->desc;
342
343	if (sdev->dspless_mode_selected)
344		return;
345
346	/* disable IPC interrupt */
347	snd_sof_dsp_update_bits(sdev, HDA_DSP_BAR, HDA_DSP_REG_ADSPIC,
348				HDA_DSP_ADSPIC_IPC, 0);
349
350	/* disable IPC BUSY and DONE interrupt */
351	snd_sof_dsp_update_bits(sdev, HDA_DSP_BAR, chip->ipc_ctl,
352			HDA_DSP_REG_HIPCCTL_BUSY | HDA_DSP_REG_HIPCCTL_DONE, 0);
353}
354
355static int hda_dsp_wait_d0i3c_done(struct snd_sof_dev *sdev)
356{
357	int retry = HDA_DSP_REG_POLL_RETRY_COUNT;
358	struct snd_sof_pdata *pdata = sdev->pdata;
359	const struct sof_intel_dsp_desc *chip;
360
361	chip = get_chip_info(pdata);
362	while (snd_sof_dsp_read8(sdev, HDA_DSP_HDA_BAR, chip->d0i3_offset) &
363		SOF_HDA_VS_D0I3C_CIP) {
364		if (!retry--)
365			return -ETIMEDOUT;
366		usleep_range(10, 15);
367	}
368
369	return 0;
370}
371
372static int hda_dsp_send_pm_gate_ipc(struct snd_sof_dev *sdev, u32 flags)
373{
374	const struct sof_ipc_pm_ops *pm_ops = sof_ipc_get_ops(sdev, pm);
375
376	if (pm_ops && pm_ops->set_pm_gate)
377		return pm_ops->set_pm_gate(sdev, flags);
378
379	return 0;
380}
381
382static int hda_dsp_update_d0i3c_register(struct snd_sof_dev *sdev, u8 value)
383{
384	struct snd_sof_pdata *pdata = sdev->pdata;
385	const struct sof_intel_dsp_desc *chip;
386	int ret;
387	u8 reg;
388
389	chip = get_chip_info(pdata);
390
391	/* Write to D0I3C after Command-In-Progress bit is cleared */
392	ret = hda_dsp_wait_d0i3c_done(sdev);
393	if (ret < 0) {
394		dev_err(sdev->dev, "CIP timeout before D0I3C update!\n");
395		return ret;
396	}
397
398	/* Update D0I3C register */
399	snd_sof_dsp_update8(sdev, HDA_DSP_HDA_BAR, chip->d0i3_offset,
400			    SOF_HDA_VS_D0I3C_I3, value);
401
402	/*
403	 * The value written to the D0I3C::I3 bit may not be taken into account immediately.
404	 * A delay is recommended before checking if D0I3C::CIP is cleared
405	 */
406	usleep_range(30, 40);
407
408	/* Wait for cmd in progress to be cleared before exiting the function */
409	ret = hda_dsp_wait_d0i3c_done(sdev);
410	if (ret < 0) {
411		dev_err(sdev->dev, "CIP timeout after D0I3C update!\n");
412		return ret;
413	}
414
415	reg = snd_sof_dsp_read8(sdev, HDA_DSP_HDA_BAR, chip->d0i3_offset);
416	/* Confirm d0i3 state changed with paranoia check */
417	if ((reg ^ value) & SOF_HDA_VS_D0I3C_I3) {
418		dev_err(sdev->dev, "failed to update D0I3C!\n");
419		return -EIO;
420	}
421
422	trace_sof_intel_D0I3C_updated(sdev, reg);
423
424	return 0;
425}
426
427/*
428 * d0i3 streaming is enabled if all the active streams can
429 * work in d0i3 state and playback is enabled
430 */
431static bool hda_dsp_d0i3_streaming_applicable(struct snd_sof_dev *sdev)
432{
433	struct snd_pcm_substream *substream;
434	struct snd_sof_pcm *spcm;
435	bool playback_active = false;
436	int dir;
437
438	list_for_each_entry(spcm, &sdev->pcm_list, list) {
439		for_each_pcm_streams(dir) {
440			substream = spcm->stream[dir].substream;
441			if (!substream || !substream->runtime)
442				continue;
443
444			if (!spcm->stream[dir].d0i3_compatible)
445				return false;
446
447			if (dir == SNDRV_PCM_STREAM_PLAYBACK)
448				playback_active = true;
449		}
450	}
451
452	return playback_active;
453}
454
455static int hda_dsp_set_D0_state(struct snd_sof_dev *sdev,
456				const struct sof_dsp_power_state *target_state)
457{
458	u32 flags = 0;
459	int ret;
460	u8 value = 0;
461
462	/*
463	 * Sanity check for illegal state transitions
464	 * The only allowed transitions are:
465	 * 1. D3 -> D0I0
466	 * 2. D0I0 -> D0I3
467	 * 3. D0I3 -> D0I0
468	 */
469	switch (sdev->dsp_power_state.state) {
470	case SOF_DSP_PM_D0:
471		/* Follow the sequence below for D0 substate transitions */
472		break;
473	case SOF_DSP_PM_D3:
474		/* Follow regular flow for D3 -> D0 transition */
475		return 0;
476	default:
477		dev_err(sdev->dev, "error: transition from %d to %d not allowed\n",
478			sdev->dsp_power_state.state, target_state->state);
479		return -EINVAL;
480	}
481
482	/* Set flags and register value for D0 target substate */
483	if (target_state->substate == SOF_HDA_DSP_PM_D0I3) {
484		value = SOF_HDA_VS_D0I3C_I3;
485
486		/*
487		 * Trace DMA need to be disabled when the DSP enters
488		 * D0I3 for S0Ix suspend, but it can be kept enabled
489		 * when the DSP enters D0I3 while the system is in S0
490		 * for debug purpose.
491		 */
492		if (!sdev->fw_trace_is_supported ||
493		    !hda_enable_trace_D0I3_S0 ||
494		    sdev->system_suspend_target != SOF_SUSPEND_NONE)
495			flags = HDA_PM_NO_DMA_TRACE;
496
497		if (hda_dsp_d0i3_streaming_applicable(sdev))
498			flags |= HDA_PM_PG_STREAMING;
499	} else {
500		/* prevent power gating in D0I0 */
501		flags = HDA_PM_PPG;
502	}
503
504	/* update D0I3C register */
505	ret = hda_dsp_update_d0i3c_register(sdev, value);
506	if (ret < 0)
507		return ret;
508
509	/*
510	 * Notify the DSP of the state change.
511	 * If this IPC fails, revert the D0I3C register update in order
512	 * to prevent partial state change.
513	 */
514	ret = hda_dsp_send_pm_gate_ipc(sdev, flags);
515	if (ret < 0) {
516		dev_err(sdev->dev,
517			"error: PM_GATE ipc error %d\n", ret);
518		goto revert;
519	}
520
521	return ret;
522
523revert:
524	/* fallback to the previous register value */
525	value = value ? 0 : SOF_HDA_VS_D0I3C_I3;
526
527	/*
528	 * This can fail but return the IPC error to signal that
529	 * the state change failed.
530	 */
531	hda_dsp_update_d0i3c_register(sdev, value);
532
533	return ret;
534}
535
536/* helper to log DSP state */
537static void hda_dsp_state_log(struct snd_sof_dev *sdev)
538{
539	switch (sdev->dsp_power_state.state) {
540	case SOF_DSP_PM_D0:
541		switch (sdev->dsp_power_state.substate) {
542		case SOF_HDA_DSP_PM_D0I0:
543			dev_dbg(sdev->dev, "Current DSP power state: D0I0\n");
544			break;
545		case SOF_HDA_DSP_PM_D0I3:
546			dev_dbg(sdev->dev, "Current DSP power state: D0I3\n");
547			break;
548		default:
549			dev_dbg(sdev->dev, "Unknown DSP D0 substate: %d\n",
550				sdev->dsp_power_state.substate);
551			break;
552		}
553		break;
554	case SOF_DSP_PM_D1:
555		dev_dbg(sdev->dev, "Current DSP power state: D1\n");
556		break;
557	case SOF_DSP_PM_D2:
558		dev_dbg(sdev->dev, "Current DSP power state: D2\n");
559		break;
560	case SOF_DSP_PM_D3:
561		dev_dbg(sdev->dev, "Current DSP power state: D3\n");
562		break;
563	default:
564		dev_dbg(sdev->dev, "Unknown DSP power state: %d\n",
565			sdev->dsp_power_state.state);
566		break;
567	}
568}
569
570/*
571 * All DSP power state transitions are initiated by the driver.
572 * If the requested state change fails, the error is simply returned.
573 * Further state transitions are attempted only when the set_power_save() op
574 * is called again either because of a new IPC sent to the DSP or
575 * during system suspend/resume.
576 */
577static int hda_dsp_set_power_state(struct snd_sof_dev *sdev,
578				   const struct sof_dsp_power_state *target_state)
579{
580	int ret = 0;
581
582	switch (target_state->state) {
583	case SOF_DSP_PM_D0:
584		ret = hda_dsp_set_D0_state(sdev, target_state);
585		break;
586	case SOF_DSP_PM_D3:
587		/* The only allowed transition is: D0I0 -> D3 */
588		if (sdev->dsp_power_state.state == SOF_DSP_PM_D0 &&
589		    sdev->dsp_power_state.substate == SOF_HDA_DSP_PM_D0I0)
590			break;
591
592		dev_err(sdev->dev,
593			"error: transition from %d to %d not allowed\n",
594			sdev->dsp_power_state.state, target_state->state);
595		return -EINVAL;
596	default:
597		dev_err(sdev->dev, "error: target state unsupported %d\n",
598			target_state->state);
599		return -EINVAL;
600	}
601	if (ret < 0) {
602		dev_err(sdev->dev,
603			"failed to set requested target DSP state %d substate %d\n",
604			target_state->state, target_state->substate);
605		return ret;
606	}
607
608	sdev->dsp_power_state = *target_state;
609	hda_dsp_state_log(sdev);
610	return ret;
611}
612
613int hda_dsp_set_power_state_ipc3(struct snd_sof_dev *sdev,
614				 const struct sof_dsp_power_state *target_state)
615{
616	/*
617	 * When the DSP is already in D0I3 and the target state is D0I3,
618	 * it could be the case that the DSP is in D0I3 during S0
619	 * and the system is suspending to S0Ix. Therefore,
620	 * hda_dsp_set_D0_state() must be called to disable trace DMA
621	 * by sending the PM_GATE IPC to the FW.
622	 */
623	if (target_state->substate == SOF_HDA_DSP_PM_D0I3 &&
624	    sdev->system_suspend_target == SOF_SUSPEND_S0IX)
625		return hda_dsp_set_power_state(sdev, target_state);
626
627	/*
628	 * For all other cases, return without doing anything if
629	 * the DSP is already in the target state.
630	 */
631	if (target_state->state == sdev->dsp_power_state.state &&
632	    target_state->substate == sdev->dsp_power_state.substate)
633		return 0;
634
635	return hda_dsp_set_power_state(sdev, target_state);
636}
637
638int hda_dsp_set_power_state_ipc4(struct snd_sof_dev *sdev,
639				 const struct sof_dsp_power_state *target_state)
640{
641	/* Return without doing anything if the DSP is already in the target state */
642	if (target_state->state == sdev->dsp_power_state.state &&
643	    target_state->substate == sdev->dsp_power_state.substate)
644		return 0;
645
646	return hda_dsp_set_power_state(sdev, target_state);
647}
648
649/*
650 * Audio DSP states may transform as below:-
651 *
652 *                                         Opportunistic D0I3 in S0
653 *     Runtime    +---------------------+  Delayed D0i3 work timeout
654 *     suspend    |                     +--------------------+
655 *   +------------+       D0I0(active)  |                    |
656 *   |            |                     <---------------+    |
657 *   |   +-------->                     |    New IPC	|    |
658 *   |   |Runtime +--^--+---------^--+--+ (via mailbox)	|    |
659 *   |   |resume     |  |         |  |			|    |
660 *   |   |           |  |         |  |			|    |
661 *   |   |     System|  |         |  |			|    |
662 *   |   |     resume|  | S3/S0IX |  |                  |    |
663 *   |   |	     |  | suspend |  | S0IX             |    |
664 *   |   |           |  |         |  |suspend           |    |
665 *   |   |           |  |         |  |                  |    |
666 *   |   |           |  |         |  |                  |    |
667 * +-v---+-----------+--v-------+ |  |           +------+----v----+
668 * |                            | |  +----------->                |
669 * |       D3 (suspended)       | |              |      D0I3      |
670 * |                            | +--------------+                |
671 * |                            |  System resume |                |
672 * +----------------------------+		 +----------------+
673 *
674 * S0IX suspend: The DSP is in D0I3 if any D0I3-compatible streams
675 *		 ignored the suspend trigger. Otherwise the DSP
676 *		 is in D3.
677 */
678
679static int hda_suspend(struct snd_sof_dev *sdev, bool runtime_suspend)
680{
681	struct sof_intel_hda_dev *hda = sdev->pdata->hw_pdata;
682	const struct sof_intel_dsp_desc *chip = hda->desc;
683	struct hdac_bus *bus = sof_to_bus(sdev);
684	int ret, j;
685
686	/*
687	 * The memory used for IMR boot loses its content in deeper than S3 state
688	 * We must not try IMR boot on next power up (as it will fail).
689	 *
690	 * In case of firmware crash or boot failure set the skip_imr_boot to true
691	 * as well in order to try to re-load the firmware to do a 'cold' boot.
692	 */
693	if (sdev->system_suspend_target > SOF_SUSPEND_S3 ||
694	    sdev->fw_state == SOF_FW_CRASHED ||
695	    sdev->fw_state == SOF_FW_BOOT_FAILED)
696		hda->skip_imr_boot = true;
697
698	ret = chip->disable_interrupts(sdev);
699	if (ret < 0)
700		return ret;
701
702	hda_codec_jack_wake_enable(sdev, runtime_suspend);
703
704	/* power down all hda links */
705	hda_bus_ml_suspend(bus);
706
707	if (sdev->dspless_mode_selected)
708		goto skip_dsp;
709
710	ret = chip->power_down_dsp(sdev);
711	if (ret < 0) {
712		dev_err(sdev->dev, "failed to power down DSP during suspend\n");
713		return ret;
714	}
715
716	/* reset ref counts for all cores */
717	for (j = 0; j < chip->cores_num; j++)
718		sdev->dsp_core_ref_count[j] = 0;
719
720	/* disable ppcap interrupt */
721	hda_dsp_ctrl_ppcap_enable(sdev, false);
722	hda_dsp_ctrl_ppcap_int_enable(sdev, false);
723skip_dsp:
724
725	/* disable hda bus irq and streams */
726	hda_dsp_ctrl_stop_chip(sdev);
727
728	/* disable LP retention mode */
729	snd_sof_pci_update_bits(sdev, PCI_PGCTL,
730				PCI_PGCTL_LSRMD_MASK, PCI_PGCTL_LSRMD_MASK);
731
732	/* reset controller */
733	ret = hda_dsp_ctrl_link_reset(sdev, true);
734	if (ret < 0) {
735		dev_err(sdev->dev,
736			"error: failed to reset controller during suspend\n");
737		return ret;
738	}
739
740	/* display codec can powered off after link reset */
741	hda_codec_i915_display_power(sdev, false);
742
743	return 0;
744}
745
746static int hda_resume(struct snd_sof_dev *sdev, bool runtime_resume)
747{
748	int ret;
749
750	/* display codec must be powered before link reset */
751	hda_codec_i915_display_power(sdev, true);
752
753	/*
754	 * clear TCSEL to clear playback on some HD Audio
755	 * codecs. PCI TCSEL is defined in the Intel manuals.
756	 */
757	snd_sof_pci_update_bits(sdev, PCI_TCSEL, 0x07, 0);
758
759	/* reset and start hda controller */
760	ret = hda_dsp_ctrl_init_chip(sdev);
761	if (ret < 0) {
762		dev_err(sdev->dev,
763			"error: failed to start controller after resume\n");
764		goto cleanup;
765	}
766
767	/* check jack status */
768	if (runtime_resume) {
769		hda_codec_jack_wake_enable(sdev, false);
770		if (sdev->system_suspend_target == SOF_SUSPEND_NONE)
771			hda_codec_jack_check(sdev);
772	}
773
774	if (!sdev->dspless_mode_selected) {
775		/* enable ppcap interrupt */
776		hda_dsp_ctrl_ppcap_enable(sdev, true);
777		hda_dsp_ctrl_ppcap_int_enable(sdev, true);
778	}
779
780cleanup:
781	/* display codec can powered off after controller init */
782	hda_codec_i915_display_power(sdev, false);
783
784	return 0;
785}
786
787int hda_dsp_resume(struct snd_sof_dev *sdev)
788{
789	struct sof_intel_hda_dev *hda = sdev->pdata->hw_pdata;
790	struct hdac_bus *bus = sof_to_bus(sdev);
791	struct pci_dev *pci = to_pci_dev(sdev->dev);
792	const struct sof_dsp_power_state target_state = {
793		.state = SOF_DSP_PM_D0,
794		.substate = SOF_HDA_DSP_PM_D0I0,
795	};
796	int ret;
797
798	/* resume from D0I3 */
799	if (sdev->dsp_power_state.state == SOF_DSP_PM_D0) {
800		ret = hda_bus_ml_resume(bus);
801		if (ret < 0) {
802			dev_err(sdev->dev,
803				"error %d in %s: failed to power up links",
804				ret, __func__);
805			return ret;
806		}
807
808		/* set up CORB/RIRB buffers if was on before suspend */
809		hda_codec_resume_cmd_io(sdev);
810
811		/* Set DSP power state */
812		ret = snd_sof_dsp_set_power_state(sdev, &target_state);
813		if (ret < 0) {
814			dev_err(sdev->dev, "error: setting dsp state %d substate %d\n",
815				target_state.state, target_state.substate);
816			return ret;
817		}
818
819		/* restore L1SEN bit */
820		if (hda->l1_disabled)
821			snd_sof_dsp_update_bits(sdev, HDA_DSP_HDA_BAR,
822						HDA_VS_INTEL_EM2,
823						HDA_VS_INTEL_EM2_L1SEN, 0);
824
825		/* restore and disable the system wakeup */
826		pci_restore_state(pci);
827		disable_irq_wake(pci->irq);
828		return 0;
829	}
830
831	/* init hda controller. DSP cores will be powered up during fw boot */
832	ret = hda_resume(sdev, false);
833	if (ret < 0)
834		return ret;
835
836	return snd_sof_dsp_set_power_state(sdev, &target_state);
837}
838
839int hda_dsp_runtime_resume(struct snd_sof_dev *sdev)
840{
841	const struct sof_dsp_power_state target_state = {
842		.state = SOF_DSP_PM_D0,
843	};
844	int ret;
845
846	/* init hda controller. DSP cores will be powered up during fw boot */
847	ret = hda_resume(sdev, true);
848	if (ret < 0)
849		return ret;
850
851	return snd_sof_dsp_set_power_state(sdev, &target_state);
852}
853
854int hda_dsp_runtime_idle(struct snd_sof_dev *sdev)
855{
856	struct hdac_bus *hbus = sof_to_bus(sdev);
857
858	if (hbus->codec_powered) {
859		dev_dbg(sdev->dev, "some codecs still powered (%08X), not idle\n",
860			(unsigned int)hbus->codec_powered);
861		return -EBUSY;
862	}
863
864	return 0;
865}
866
867int hda_dsp_runtime_suspend(struct snd_sof_dev *sdev)
868{
869	struct sof_intel_hda_dev *hda = sdev->pdata->hw_pdata;
870	const struct sof_dsp_power_state target_state = {
871		.state = SOF_DSP_PM_D3,
872	};
873	int ret;
874
875	if (!sdev->dspless_mode_selected) {
876		/* cancel any attempt for DSP D0I3 */
877		cancel_delayed_work_sync(&hda->d0i3_work);
878	}
879
880	/* stop hda controller and power dsp off */
881	ret = hda_suspend(sdev, true);
882	if (ret < 0)
883		return ret;
884
885	return snd_sof_dsp_set_power_state(sdev, &target_state);
886}
887
888int hda_dsp_suspend(struct snd_sof_dev *sdev, u32 target_state)
889{
890	struct sof_intel_hda_dev *hda = sdev->pdata->hw_pdata;
891	struct hdac_bus *bus = sof_to_bus(sdev);
892	struct pci_dev *pci = to_pci_dev(sdev->dev);
893	const struct sof_dsp_power_state target_dsp_state = {
894		.state = target_state,
895		.substate = target_state == SOF_DSP_PM_D0 ?
896				SOF_HDA_DSP_PM_D0I3 : 0,
897	};
898	int ret;
899
900	if (!sdev->dspless_mode_selected) {
901		/* cancel any attempt for DSP D0I3 */
902		cancel_delayed_work_sync(&hda->d0i3_work);
903	}
904
905	if (target_state == SOF_DSP_PM_D0) {
906		/* Set DSP power state */
907		ret = snd_sof_dsp_set_power_state(sdev, &target_dsp_state);
908		if (ret < 0) {
909			dev_err(sdev->dev, "error: setting dsp state %d substate %d\n",
910				target_dsp_state.state,
911				target_dsp_state.substate);
912			return ret;
913		}
914
915		/* enable L1SEN to make sure the system can enter S0Ix */
916		if (hda->l1_disabled)
917			snd_sof_dsp_update_bits(sdev, HDA_DSP_HDA_BAR, HDA_VS_INTEL_EM2,
918						HDA_VS_INTEL_EM2_L1SEN, HDA_VS_INTEL_EM2_L1SEN);
919
920		/* stop the CORB/RIRB DMA if it is On */
921		hda_codec_suspend_cmd_io(sdev);
922
923		/* no link can be powered in s0ix state */
924		ret = hda_bus_ml_suspend(bus);
925		if (ret < 0) {
926			dev_err(sdev->dev,
927				"error %d in %s: failed to power down links",
928				ret, __func__);
929			return ret;
930		}
931
932		/* enable the system waking up via IPC IRQ */
933		enable_irq_wake(pci->irq);
934		pci_save_state(pci);
935		return 0;
936	}
937
938	/* stop hda controller and power dsp off */
939	ret = hda_suspend(sdev, false);
940	if (ret < 0) {
941		dev_err(bus->dev, "error: suspending dsp\n");
942		return ret;
943	}
944
945	return snd_sof_dsp_set_power_state(sdev, &target_dsp_state);
946}
947
948static unsigned int hda_dsp_check_for_dma_streams(struct snd_sof_dev *sdev)
949{
950	struct hdac_bus *bus = sof_to_bus(sdev);
951	struct hdac_stream *s;
952	unsigned int active_streams = 0;
953	int sd_offset;
954	u32 val;
955
956	list_for_each_entry(s, &bus->stream_list, list) {
957		sd_offset = SOF_STREAM_SD_OFFSET(s);
958		val = snd_sof_dsp_read(sdev, HDA_DSP_HDA_BAR,
959				       sd_offset);
960		if (val & SOF_HDA_SD_CTL_DMA_START)
961			active_streams |= BIT(s->index);
962	}
963
964	return active_streams;
965}
966
967static int hda_dsp_s5_quirk(struct snd_sof_dev *sdev)
968{
969	int ret;
970
971	/*
972	 * Do not assume a certain timing between the prior
973	 * suspend flow, and running of this quirk function.
974	 * This is needed if the controller was just put
975	 * to reset before calling this function.
976	 */
977	usleep_range(500, 1000);
978
979	/*
980	 * Take controller out of reset to flush DMA
981	 * transactions.
982	 */
983	ret = hda_dsp_ctrl_link_reset(sdev, false);
984	if (ret < 0)
985		return ret;
986
987	usleep_range(500, 1000);
988
989	/* Restore state for shutdown, back to reset */
990	ret = hda_dsp_ctrl_link_reset(sdev, true);
991	if (ret < 0)
992		return ret;
993
994	return ret;
995}
996
997int hda_dsp_shutdown_dma_flush(struct snd_sof_dev *sdev)
998{
999	unsigned int active_streams;
1000	int ret, ret2;
1001
1002	/* check if DMA cleanup has been successful */
1003	active_streams = hda_dsp_check_for_dma_streams(sdev);
1004
1005	sdev->system_suspend_target = SOF_SUSPEND_S3;
1006	ret = snd_sof_suspend(sdev->dev);
1007
1008	if (active_streams) {
1009		dev_warn(sdev->dev,
1010			 "There were active DSP streams (%#x) at shutdown, trying to recover\n",
1011			 active_streams);
1012		ret2 = hda_dsp_s5_quirk(sdev);
1013		if (ret2 < 0)
1014			dev_err(sdev->dev, "shutdown recovery failed (%d)\n", ret2);
1015	}
1016
1017	return ret;
1018}
1019
1020int hda_dsp_shutdown(struct snd_sof_dev *sdev)
1021{
1022	sdev->system_suspend_target = SOF_SUSPEND_S3;
1023	return snd_sof_suspend(sdev->dev);
1024}
1025
1026int hda_dsp_set_hw_params_upon_resume(struct snd_sof_dev *sdev)
1027{
1028	int ret;
1029
1030	/* make sure all DAI resources are freed */
1031	ret = hda_dsp_dais_suspend(sdev);
1032	if (ret < 0)
1033		dev_warn(sdev->dev, "%s: failure in hda_dsp_dais_suspend\n", __func__);
1034
1035	return ret;
1036}
1037
1038void hda_dsp_d0i3_work(struct work_struct *work)
1039{
1040	struct sof_intel_hda_dev *hdev = container_of(work,
1041						      struct sof_intel_hda_dev,
1042						      d0i3_work.work);
1043	struct hdac_bus *bus = &hdev->hbus.core;
1044	struct snd_sof_dev *sdev = dev_get_drvdata(bus->dev);
1045	struct sof_dsp_power_state target_state = {
1046		.state = SOF_DSP_PM_D0,
1047		.substate = SOF_HDA_DSP_PM_D0I3,
1048	};
1049	int ret;
1050
1051	/* DSP can enter D0I3 iff only D0I3-compatible streams are active */
1052	if (!snd_sof_dsp_only_d0i3_compatible_stream_active(sdev))
1053		/* remain in D0I0 */
1054		return;
1055
1056	/* This can fail but error cannot be propagated */
1057	ret = snd_sof_dsp_set_power_state(sdev, &target_state);
1058	if (ret < 0)
1059		dev_err_ratelimited(sdev->dev,
1060				    "error: failed to set DSP state %d substate %d\n",
1061				    target_state.state, target_state.substate);
1062}
1063
1064int hda_dsp_core_get(struct snd_sof_dev *sdev, int core)
1065{
1066	const struct sof_ipc_pm_ops *pm_ops = sdev->ipc->ops->pm;
1067	int ret, ret1;
1068
1069	/* power up core */
1070	ret = hda_dsp_enable_core(sdev, BIT(core));
1071	if (ret < 0) {
1072		dev_err(sdev->dev, "failed to power up core %d with err: %d\n",
1073			core, ret);
1074		return ret;
1075	}
1076
1077	/* No need to send IPC for primary core or if FW boot is not complete */
1078	if (sdev->fw_state != SOF_FW_BOOT_COMPLETE || core == SOF_DSP_PRIMARY_CORE)
1079		return 0;
1080
1081	/* No need to continue the set_core_state ops is not available */
1082	if (!pm_ops->set_core_state)
1083		return 0;
1084
1085	/* Now notify DSP for secondary cores */
1086	ret = pm_ops->set_core_state(sdev, core, true);
1087	if (ret < 0) {
1088		dev_err(sdev->dev, "failed to enable secondary core '%d' failed with %d\n",
1089			core, ret);
1090		goto power_down;
1091	}
1092
1093	return ret;
1094
1095power_down:
1096	/* power down core if it is host managed and return the original error if this fails too */
1097	ret1 = hda_dsp_core_reset_power_down(sdev, BIT(core));
1098	if (ret1 < 0)
1099		dev_err(sdev->dev, "failed to power down core: %d with err: %d\n", core, ret1);
1100
1101	return ret;
1102}
1103
1104int hda_dsp_disable_interrupts(struct snd_sof_dev *sdev)
1105{
1106	hda_sdw_int_enable(sdev, false);
1107	hda_dsp_ipc_int_disable(sdev);
1108
1109	return 0;
1110}
1111