1 // SPDX-License-Identifier: GPL-2.0-only
2 /*
3  *  skl-topology.c - Implements Platform component ALSA controls/widget
4  *  handlers.
5  *
6  *  Copyright (C) 2014-2015 Intel Corp
7  *  Author: Jeeja KP <jeeja.kp@intel.com>
8  *  ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
9  */
10 
11 #include <linux/slab.h>
12 #include <linux/types.h>
13 #include <linux/firmware.h>
14 #include <linux/uuid.h>
15 #include <sound/intel-nhlt.h>
16 #include <sound/soc.h>
17 #include <sound/soc-acpi.h>
18 #include <sound/soc-topology.h>
19 #include <uapi/sound/snd_sst_tokens.h>
20 #include <uapi/sound/skl-tplg-interface.h>
21 #include "skl-sst-dsp.h"
22 #include "skl-sst-ipc.h"
23 #include "skl-topology.h"
24 #include "skl.h"
25 #include "../common/sst-dsp.h"
26 #include "../common/sst-dsp-priv.h"
27 
28 #define SKL_CH_FIXUP_MASK		(1 << 0)
29 #define SKL_RATE_FIXUP_MASK		(1 << 1)
30 #define SKL_FMT_FIXUP_MASK		(1 << 2)
31 #define SKL_IN_DIR_BIT_MASK		BIT(0)
32 #define SKL_PIN_COUNT_MASK		GENMASK(7, 4)
33 
34 static const int mic_mono_list[] = {
35 0, 1, 2, 3,
36 };
37 static const int mic_stereo_list[][SKL_CH_STEREO] = {
38 {0, 1}, {0, 2}, {0, 3}, {1, 2}, {1, 3}, {2, 3},
39 };
40 static const int mic_trio_list[][SKL_CH_TRIO] = {
41 {0, 1, 2}, {0, 1, 3}, {0, 2, 3}, {1, 2, 3},
42 };
43 static const int mic_quatro_list[][SKL_CH_QUATRO] = {
44 {0, 1, 2, 3},
45 };
46 
47 #define CHECK_HW_PARAMS(ch, freq, bps, prm_ch, prm_freq, prm_bps) \
48 	((ch == prm_ch) && (bps == prm_bps) && (freq == prm_freq))
49 
skl_tplg_d0i3_get(struct skl_dev * skl,enum d0i3_capability caps)50 void skl_tplg_d0i3_get(struct skl_dev *skl, enum d0i3_capability caps)
51 {
52 	struct skl_d0i3_data *d0i3 =  &skl->d0i3;
53 
54 	switch (caps) {
55 	case SKL_D0I3_NONE:
56 		d0i3->non_d0i3++;
57 		break;
58 
59 	case SKL_D0I3_STREAMING:
60 		d0i3->streaming++;
61 		break;
62 
63 	case SKL_D0I3_NON_STREAMING:
64 		d0i3->non_streaming++;
65 		break;
66 	}
67 }
68 
skl_tplg_d0i3_put(struct skl_dev * skl,enum d0i3_capability caps)69 void skl_tplg_d0i3_put(struct skl_dev *skl, enum d0i3_capability caps)
70 {
71 	struct skl_d0i3_data *d0i3 =  &skl->d0i3;
72 
73 	switch (caps) {
74 	case SKL_D0I3_NONE:
75 		d0i3->non_d0i3--;
76 		break;
77 
78 	case SKL_D0I3_STREAMING:
79 		d0i3->streaming--;
80 		break;
81 
82 	case SKL_D0I3_NON_STREAMING:
83 		d0i3->non_streaming--;
84 		break;
85 	}
86 }
87 
88 /*
89  * SKL DSP driver modelling uses only few DAPM widgets so for rest we will
90  * ignore. This helpers checks if the SKL driver handles this widget type
91  */
is_skl_dsp_widget_type(struct snd_soc_dapm_widget * w,struct device * dev)92 static int is_skl_dsp_widget_type(struct snd_soc_dapm_widget *w,
93 				  struct device *dev)
94 {
95 	if (w->dapm->dev != dev)
96 		return false;
97 
98 	switch (w->id) {
99 	case snd_soc_dapm_dai_link:
100 	case snd_soc_dapm_dai_in:
101 	case snd_soc_dapm_aif_in:
102 	case snd_soc_dapm_aif_out:
103 	case snd_soc_dapm_dai_out:
104 	case snd_soc_dapm_switch:
105 	case snd_soc_dapm_output:
106 	case snd_soc_dapm_mux:
107 
108 		return false;
109 	default:
110 		return true;
111 	}
112 }
113 
skl_dump_mconfig(struct skl_dev * skl,struct skl_module_cfg * mcfg)114 static void skl_dump_mconfig(struct skl_dev *skl, struct skl_module_cfg *mcfg)
115 {
116 	struct skl_module_iface *iface = &mcfg->module->formats[mcfg->fmt_idx];
117 
118 	dev_dbg(skl->dev, "Dumping config\n");
119 	dev_dbg(skl->dev, "Input Format:\n");
120 	dev_dbg(skl->dev, "channels = %d\n", iface->inputs[0].fmt.channels);
121 	dev_dbg(skl->dev, "s_freq = %d\n", iface->inputs[0].fmt.s_freq);
122 	dev_dbg(skl->dev, "ch_cfg = %d\n", iface->inputs[0].fmt.ch_cfg);
123 	dev_dbg(skl->dev, "valid bit depth = %d\n",
124 				iface->inputs[0].fmt.valid_bit_depth);
125 	dev_dbg(skl->dev, "Output Format:\n");
126 	dev_dbg(skl->dev, "channels = %d\n", iface->outputs[0].fmt.channels);
127 	dev_dbg(skl->dev, "s_freq = %d\n", iface->outputs[0].fmt.s_freq);
128 	dev_dbg(skl->dev, "valid bit depth = %d\n",
129 				iface->outputs[0].fmt.valid_bit_depth);
130 	dev_dbg(skl->dev, "ch_cfg = %d\n", iface->outputs[0].fmt.ch_cfg);
131 }
132 
skl_tplg_update_chmap(struct skl_module_fmt * fmt,int chs)133 static void skl_tplg_update_chmap(struct skl_module_fmt *fmt, int chs)
134 {
135 	int slot_map = 0xFFFFFFFF;
136 	int start_slot = 0;
137 	int i;
138 
139 	for (i = 0; i < chs; i++) {
140 		/*
141 		 * For 2 channels with starting slot as 0, slot map will
142 		 * look like 0xFFFFFF10.
143 		 */
144 		slot_map &= (~(0xF << (4 * i)) | (start_slot << (4 * i)));
145 		start_slot++;
146 	}
147 	fmt->ch_map = slot_map;
148 }
149 
skl_tplg_update_params(struct skl_module_fmt * fmt,struct skl_pipe_params * params,int fixup)150 static void skl_tplg_update_params(struct skl_module_fmt *fmt,
151 			struct skl_pipe_params *params, int fixup)
152 {
153 	if (fixup & SKL_RATE_FIXUP_MASK)
154 		fmt->s_freq = params->s_freq;
155 	if (fixup & SKL_CH_FIXUP_MASK) {
156 		fmt->channels = params->ch;
157 		skl_tplg_update_chmap(fmt, fmt->channels);
158 	}
159 	if (fixup & SKL_FMT_FIXUP_MASK) {
160 		fmt->valid_bit_depth = skl_get_bit_depth(params->s_fmt);
161 
162 		/*
163 		 * 16 bit is 16 bit container whereas 24 bit is in 32 bit
164 		 * container so update bit depth accordingly
165 		 */
166 		switch (fmt->valid_bit_depth) {
167 		case SKL_DEPTH_16BIT:
168 			fmt->bit_depth = fmt->valid_bit_depth;
169 			break;
170 
171 		default:
172 			fmt->bit_depth = SKL_DEPTH_32BIT;
173 			break;
174 		}
175 	}
176 
177 }
178 
179 /*
180  * A pipeline may have modules which impact the pcm parameters, like SRC,
181  * channel converter, format converter.
182  * We need to calculate the output params by applying the 'fixup'
183  * Topology will tell driver which type of fixup is to be applied by
184  * supplying the fixup mask, so based on that we calculate the output
185  *
186  * Now In FE the pcm hw_params is source/target format. Same is applicable
187  * for BE with its hw_params invoked.
188  * here based on FE, BE pipeline and direction we calculate the input and
189  * outfix and then apply that for a module
190  */
skl_tplg_update_params_fixup(struct skl_module_cfg * m_cfg,struct skl_pipe_params * params,bool is_fe)191 static void skl_tplg_update_params_fixup(struct skl_module_cfg *m_cfg,
192 		struct skl_pipe_params *params, bool is_fe)
193 {
194 	int in_fixup, out_fixup;
195 	struct skl_module_fmt *in_fmt, *out_fmt;
196 
197 	/* Fixups will be applied to pin 0 only */
198 	in_fmt = &m_cfg->module->formats[m_cfg->fmt_idx].inputs[0].fmt;
199 	out_fmt = &m_cfg->module->formats[m_cfg->fmt_idx].outputs[0].fmt;
200 
201 	if (params->stream == SNDRV_PCM_STREAM_PLAYBACK) {
202 		if (is_fe) {
203 			in_fixup = m_cfg->params_fixup;
204 			out_fixup = (~m_cfg->converter) &
205 					m_cfg->params_fixup;
206 		} else {
207 			out_fixup = m_cfg->params_fixup;
208 			in_fixup = (~m_cfg->converter) &
209 					m_cfg->params_fixup;
210 		}
211 	} else {
212 		if (is_fe) {
213 			out_fixup = m_cfg->params_fixup;
214 			in_fixup = (~m_cfg->converter) &
215 					m_cfg->params_fixup;
216 		} else {
217 			in_fixup = m_cfg->params_fixup;
218 			out_fixup = (~m_cfg->converter) &
219 					m_cfg->params_fixup;
220 		}
221 	}
222 
223 	skl_tplg_update_params(in_fmt, params, in_fixup);
224 	skl_tplg_update_params(out_fmt, params, out_fixup);
225 }
226 
227 /*
228  * A module needs input and output buffers, which are dependent upon pcm
229  * params, so once we have calculate params, we need buffer calculation as
230  * well.
231  */
skl_tplg_update_buffer_size(struct skl_dev * skl,struct skl_module_cfg * mcfg)232 static void skl_tplg_update_buffer_size(struct skl_dev *skl,
233 				struct skl_module_cfg *mcfg)
234 {
235 	int multiplier = 1;
236 	struct skl_module_fmt *in_fmt, *out_fmt;
237 	struct skl_module_res *res;
238 
239 	/* Since fixups is applied to pin 0 only, ibs, obs needs
240 	 * change for pin 0 only
241 	 */
242 	res = &mcfg->module->resources[mcfg->res_idx];
243 	in_fmt = &mcfg->module->formats[mcfg->fmt_idx].inputs[0].fmt;
244 	out_fmt = &mcfg->module->formats[mcfg->fmt_idx].outputs[0].fmt;
245 
246 	if (mcfg->m_type == SKL_MODULE_TYPE_SRCINT)
247 		multiplier = 5;
248 
249 	res->ibs = DIV_ROUND_UP(in_fmt->s_freq, 1000) *
250 			in_fmt->channels * (in_fmt->bit_depth >> 3) *
251 			multiplier;
252 
253 	res->obs = DIV_ROUND_UP(out_fmt->s_freq, 1000) *
254 			out_fmt->channels * (out_fmt->bit_depth >> 3) *
255 			multiplier;
256 }
257 
skl_tplg_be_dev_type(int dev_type)258 static u8 skl_tplg_be_dev_type(int dev_type)
259 {
260 	int ret;
261 
262 	switch (dev_type) {
263 	case SKL_DEVICE_BT:
264 		ret = NHLT_DEVICE_BT;
265 		break;
266 
267 	case SKL_DEVICE_DMIC:
268 		ret = NHLT_DEVICE_DMIC;
269 		break;
270 
271 	case SKL_DEVICE_I2S:
272 		ret = NHLT_DEVICE_I2S;
273 		break;
274 
275 	default:
276 		ret = NHLT_DEVICE_INVALID;
277 		break;
278 	}
279 
280 	return ret;
281 }
282 
skl_tplg_update_be_blob(struct snd_soc_dapm_widget * w,struct skl_dev * skl)283 static int skl_tplg_update_be_blob(struct snd_soc_dapm_widget *w,
284 						struct skl_dev *skl)
285 {
286 	struct skl_module_cfg *m_cfg = w->priv;
287 	int link_type, dir;
288 	u32 ch, s_freq, s_fmt, s_cont;
289 	struct nhlt_specific_cfg *cfg;
290 	u8 dev_type = skl_tplg_be_dev_type(m_cfg->dev_type);
291 	int fmt_idx = m_cfg->fmt_idx;
292 	struct skl_module_iface *m_iface = &m_cfg->module->formats[fmt_idx];
293 
294 	/* check if we already have blob */
295 	if (m_cfg->formats_config[SKL_PARAM_INIT].caps_size > 0)
296 		return 0;
297 
298 	dev_dbg(skl->dev, "Applying default cfg blob\n");
299 	switch (m_cfg->dev_type) {
300 	case SKL_DEVICE_DMIC:
301 		link_type = NHLT_LINK_DMIC;
302 		dir = SNDRV_PCM_STREAM_CAPTURE;
303 		s_freq = m_iface->inputs[0].fmt.s_freq;
304 		s_fmt = m_iface->inputs[0].fmt.valid_bit_depth;
305 		s_cont = m_iface->inputs[0].fmt.bit_depth;
306 		ch = m_iface->inputs[0].fmt.channels;
307 		break;
308 
309 	case SKL_DEVICE_I2S:
310 		link_type = NHLT_LINK_SSP;
311 		if (m_cfg->hw_conn_type == SKL_CONN_SOURCE) {
312 			dir = SNDRV_PCM_STREAM_PLAYBACK;
313 			s_freq = m_iface->outputs[0].fmt.s_freq;
314 			s_fmt = m_iface->outputs[0].fmt.valid_bit_depth;
315 			s_cont = m_iface->outputs[0].fmt.bit_depth;
316 			ch = m_iface->outputs[0].fmt.channels;
317 		} else {
318 			dir = SNDRV_PCM_STREAM_CAPTURE;
319 			s_freq = m_iface->inputs[0].fmt.s_freq;
320 			s_fmt = m_iface->inputs[0].fmt.valid_bit_depth;
321 			s_cont = m_iface->inputs[0].fmt.bit_depth;
322 			ch = m_iface->inputs[0].fmt.channels;
323 		}
324 		break;
325 
326 	default:
327 		return -EINVAL;
328 	}
329 
330 	/* update the blob based on virtual bus_id and default params */
331 	cfg = intel_nhlt_get_endpoint_blob(skl->dev, skl->nhlt, m_cfg->vbus_id,
332 					   link_type, s_fmt, s_cont, ch,
333 					   s_freq, dir, dev_type);
334 	if (cfg) {
335 		m_cfg->formats_config[SKL_PARAM_INIT].caps_size = cfg->size;
336 		m_cfg->formats_config[SKL_PARAM_INIT].caps = (u32 *)&cfg->caps;
337 	} else {
338 		dev_err(skl->dev, "Blob NULL for id %x type %d dirn %d\n",
339 					m_cfg->vbus_id, link_type, dir);
340 		dev_err(skl->dev, "PCM: ch %d, freq %d, fmt %d/%d\n",
341 					ch, s_freq, s_fmt, s_cont);
342 		return -EIO;
343 	}
344 
345 	return 0;
346 }
347 
skl_tplg_update_module_params(struct snd_soc_dapm_widget * w,struct skl_dev * skl)348 static void skl_tplg_update_module_params(struct snd_soc_dapm_widget *w,
349 							struct skl_dev *skl)
350 {
351 	struct skl_module_cfg *m_cfg = w->priv;
352 	struct skl_pipe_params *params = m_cfg->pipe->p_params;
353 	int p_conn_type = m_cfg->pipe->conn_type;
354 	bool is_fe;
355 
356 	if (!m_cfg->params_fixup)
357 		return;
358 
359 	dev_dbg(skl->dev, "Mconfig for widget=%s BEFORE updation\n",
360 				w->name);
361 
362 	skl_dump_mconfig(skl, m_cfg);
363 
364 	if (p_conn_type == SKL_PIPE_CONN_TYPE_FE)
365 		is_fe = true;
366 	else
367 		is_fe = false;
368 
369 	skl_tplg_update_params_fixup(m_cfg, params, is_fe);
370 	skl_tplg_update_buffer_size(skl, m_cfg);
371 
372 	dev_dbg(skl->dev, "Mconfig for widget=%s AFTER updation\n",
373 				w->name);
374 
375 	skl_dump_mconfig(skl, m_cfg);
376 }
377 
378 /*
379  * some modules can have multiple params set from user control and
380  * need to be set after module is initialized. If set_param flag is
381  * set module params will be done after module is initialised.
382  */
skl_tplg_set_module_params(struct snd_soc_dapm_widget * w,struct skl_dev * skl)383 static int skl_tplg_set_module_params(struct snd_soc_dapm_widget *w,
384 						struct skl_dev *skl)
385 {
386 	int i, ret;
387 	struct skl_module_cfg *mconfig = w->priv;
388 	const struct snd_kcontrol_new *k;
389 	struct soc_bytes_ext *sb;
390 	struct skl_algo_data *bc;
391 	struct skl_specific_cfg *sp_cfg;
392 
393 	if (mconfig->formats_config[SKL_PARAM_SET].caps_size > 0 &&
394 	    mconfig->formats_config[SKL_PARAM_SET].set_params == SKL_PARAM_SET) {
395 		sp_cfg = &mconfig->formats_config[SKL_PARAM_SET];
396 		ret = skl_set_module_params(skl, sp_cfg->caps,
397 					sp_cfg->caps_size,
398 					sp_cfg->param_id, mconfig);
399 		if (ret < 0)
400 			return ret;
401 	}
402 
403 	for (i = 0; i < w->num_kcontrols; i++) {
404 		k = &w->kcontrol_news[i];
405 		if (k->access & SNDRV_CTL_ELEM_ACCESS_TLV_CALLBACK) {
406 			sb = (void *) k->private_value;
407 			bc = (struct skl_algo_data *)sb->dobj.private;
408 
409 			if (bc->set_params == SKL_PARAM_SET) {
410 				ret = skl_set_module_params(skl,
411 						(u32 *)bc->params, bc->size,
412 						bc->param_id, mconfig);
413 				if (ret < 0)
414 					return ret;
415 			}
416 		}
417 	}
418 
419 	return 0;
420 }
421 
422 /*
423  * some module param can set from user control and this is required as
424  * when module is initailzed. if module param is required in init it is
425  * identifed by set_param flag. if set_param flag is not set, then this
426  * parameter needs to set as part of module init.
427  */
skl_tplg_set_module_init_data(struct snd_soc_dapm_widget * w)428 static int skl_tplg_set_module_init_data(struct snd_soc_dapm_widget *w)
429 {
430 	const struct snd_kcontrol_new *k;
431 	struct soc_bytes_ext *sb;
432 	struct skl_algo_data *bc;
433 	struct skl_module_cfg *mconfig = w->priv;
434 	int i;
435 
436 	for (i = 0; i < w->num_kcontrols; i++) {
437 		k = &w->kcontrol_news[i];
438 		if (k->access & SNDRV_CTL_ELEM_ACCESS_TLV_CALLBACK) {
439 			sb = (struct soc_bytes_ext *)k->private_value;
440 			bc = (struct skl_algo_data *)sb->dobj.private;
441 
442 			if (bc->set_params != SKL_PARAM_INIT)
443 				continue;
444 
445 			mconfig->formats_config[SKL_PARAM_INIT].caps =
446 							(u32 *)bc->params;
447 			mconfig->formats_config[SKL_PARAM_INIT].caps_size =
448 								bc->size;
449 
450 			break;
451 		}
452 	}
453 
454 	return 0;
455 }
456 
skl_tplg_module_prepare(struct skl_dev * skl,struct skl_pipe * pipe,struct snd_soc_dapm_widget * w,struct skl_module_cfg * mcfg)457 static int skl_tplg_module_prepare(struct skl_dev *skl, struct skl_pipe *pipe,
458 		struct snd_soc_dapm_widget *w, struct skl_module_cfg *mcfg)
459 {
460 	switch (mcfg->dev_type) {
461 	case SKL_DEVICE_HDAHOST:
462 		return skl_pcm_host_dma_prepare(skl->dev, pipe->p_params);
463 
464 	case SKL_DEVICE_HDALINK:
465 		return skl_pcm_link_dma_prepare(skl->dev, pipe->p_params);
466 	}
467 
468 	return 0;
469 }
470 
471 /*
472  * Inside a pipe instance, we can have various modules. These modules need
473  * to instantiated in DSP by invoking INIT_MODULE IPC, which is achieved by
474  * skl_init_module() routine, so invoke that for all modules in a pipeline
475  */
476 static int
skl_tplg_init_pipe_modules(struct skl_dev * skl,struct skl_pipe * pipe)477 skl_tplg_init_pipe_modules(struct skl_dev *skl, struct skl_pipe *pipe)
478 {
479 	struct skl_pipe_module *w_module;
480 	struct snd_soc_dapm_widget *w;
481 	struct skl_module_cfg *mconfig;
482 	u8 cfg_idx;
483 	int ret = 0;
484 
485 	list_for_each_entry(w_module, &pipe->w_list, node) {
486 		guid_t *uuid_mod;
487 		w = w_module->w;
488 		mconfig = w->priv;
489 
490 		/* check if module ids are populated */
491 		if (mconfig->id.module_id < 0) {
492 			dev_err(skl->dev,
493 					"module %pUL id not populated\n",
494 					(guid_t *)mconfig->guid);
495 			return -EIO;
496 		}
497 
498 		cfg_idx = mconfig->pipe->cur_config_idx;
499 		mconfig->fmt_idx = mconfig->mod_cfg[cfg_idx].fmt_idx;
500 		mconfig->res_idx = mconfig->mod_cfg[cfg_idx].res_idx;
501 
502 		if (mconfig->module->loadable && skl->dsp->fw_ops.load_mod) {
503 			ret = skl->dsp->fw_ops.load_mod(skl->dsp,
504 				mconfig->id.module_id, mconfig->guid);
505 			if (ret < 0)
506 				return ret;
507 		}
508 
509 		/* prepare the DMA if the module is gateway cpr */
510 		ret = skl_tplg_module_prepare(skl, pipe, w, mconfig);
511 		if (ret < 0)
512 			return ret;
513 
514 		/* update blob if blob is null for be with default value */
515 		skl_tplg_update_be_blob(w, skl);
516 
517 		/*
518 		 * apply fix/conversion to module params based on
519 		 * FE/BE params
520 		 */
521 		skl_tplg_update_module_params(w, skl);
522 		uuid_mod = (guid_t *)mconfig->guid;
523 		mconfig->id.pvt_id = skl_get_pvt_id(skl, uuid_mod,
524 						mconfig->id.instance_id);
525 		if (mconfig->id.pvt_id < 0)
526 			return ret;
527 		skl_tplg_set_module_init_data(w);
528 
529 		ret = skl_dsp_get_core(skl->dsp, mconfig->core_id);
530 		if (ret < 0) {
531 			dev_err(skl->dev, "Failed to wake up core %d ret=%d\n",
532 						mconfig->core_id, ret);
533 			return ret;
534 		}
535 
536 		ret = skl_init_module(skl, mconfig);
537 		if (ret < 0) {
538 			skl_put_pvt_id(skl, uuid_mod, &mconfig->id.pvt_id);
539 			goto err;
540 		}
541 
542 		ret = skl_tplg_set_module_params(w, skl);
543 		if (ret < 0)
544 			goto err;
545 	}
546 
547 	return 0;
548 err:
549 	skl_dsp_put_core(skl->dsp, mconfig->core_id);
550 	return ret;
551 }
552 
skl_tplg_unload_pipe_modules(struct skl_dev * skl,struct skl_pipe * pipe)553 static int skl_tplg_unload_pipe_modules(struct skl_dev *skl,
554 	 struct skl_pipe *pipe)
555 {
556 	int ret = 0;
557 	struct skl_pipe_module *w_module;
558 	struct skl_module_cfg *mconfig;
559 
560 	list_for_each_entry(w_module, &pipe->w_list, node) {
561 		guid_t *uuid_mod;
562 		mconfig  = w_module->w->priv;
563 		uuid_mod = (guid_t *)mconfig->guid;
564 
565 		if (mconfig->module->loadable && skl->dsp->fw_ops.unload_mod) {
566 			ret = skl->dsp->fw_ops.unload_mod(skl->dsp,
567 						mconfig->id.module_id);
568 			if (ret < 0)
569 				return -EIO;
570 		}
571 		skl_put_pvt_id(skl, uuid_mod, &mconfig->id.pvt_id);
572 
573 		ret = skl_dsp_put_core(skl->dsp, mconfig->core_id);
574 		if (ret < 0) {
575 			/* don't return; continue with other modules */
576 			dev_err(skl->dev, "Failed to sleep core %d ret=%d\n",
577 				mconfig->core_id, ret);
578 		}
579 	}
580 
581 	/* no modules to unload in this path, so return */
582 	return ret;
583 }
584 
skl_tplg_is_multi_fmt(struct skl_dev * skl,struct skl_pipe * pipe)585 static bool skl_tplg_is_multi_fmt(struct skl_dev *skl, struct skl_pipe *pipe)
586 {
587 	struct skl_pipe_fmt *cur_fmt;
588 	struct skl_pipe_fmt *next_fmt;
589 	int i;
590 
591 	if (pipe->nr_cfgs <= 1)
592 		return false;
593 
594 	if (pipe->conn_type != SKL_PIPE_CONN_TYPE_FE)
595 		return true;
596 
597 	for (i = 0; i < pipe->nr_cfgs - 1; i++) {
598 		if (pipe->direction == SNDRV_PCM_STREAM_PLAYBACK) {
599 			cur_fmt = &pipe->configs[i].out_fmt;
600 			next_fmt = &pipe->configs[i + 1].out_fmt;
601 		} else {
602 			cur_fmt = &pipe->configs[i].in_fmt;
603 			next_fmt = &pipe->configs[i + 1].in_fmt;
604 		}
605 
606 		if (!CHECK_HW_PARAMS(cur_fmt->channels, cur_fmt->freq,
607 				     cur_fmt->bps,
608 				     next_fmt->channels,
609 				     next_fmt->freq,
610 				     next_fmt->bps))
611 			return true;
612 	}
613 
614 	return false;
615 }
616 
617 /*
618  * Here, we select pipe format based on the pipe type and pipe
619  * direction to determine the current config index for the pipeline.
620  * The config index is then used to select proper module resources.
621  * Intermediate pipes currently have a fixed format hence we select the
622  * 0th configuratation by default for such pipes.
623  */
624 static int
skl_tplg_get_pipe_config(struct skl_dev * skl,struct skl_module_cfg * mconfig)625 skl_tplg_get_pipe_config(struct skl_dev *skl, struct skl_module_cfg *mconfig)
626 {
627 	struct skl_pipe *pipe = mconfig->pipe;
628 	struct skl_pipe_params *params = pipe->p_params;
629 	struct skl_path_config *pconfig = &pipe->configs[0];
630 	struct skl_pipe_fmt *fmt = NULL;
631 	bool in_fmt = false;
632 	int i;
633 
634 	if (pipe->nr_cfgs == 0) {
635 		pipe->cur_config_idx = 0;
636 		return 0;
637 	}
638 
639 	if (skl_tplg_is_multi_fmt(skl, pipe)) {
640 		pipe->cur_config_idx = pipe->pipe_config_idx;
641 		pipe->memory_pages = pconfig->mem_pages;
642 		dev_dbg(skl->dev, "found pipe config idx:%d\n",
643 			pipe->cur_config_idx);
644 		return 0;
645 	}
646 
647 	if (pipe->conn_type == SKL_PIPE_CONN_TYPE_NONE || pipe->nr_cfgs == 1) {
648 		dev_dbg(skl->dev, "No conn_type or just 1 pathcfg, taking 0th for %d\n",
649 			pipe->ppl_id);
650 		pipe->cur_config_idx = 0;
651 		pipe->memory_pages = pconfig->mem_pages;
652 
653 		return 0;
654 	}
655 
656 	if ((pipe->conn_type == SKL_PIPE_CONN_TYPE_FE &&
657 	     pipe->direction == SNDRV_PCM_STREAM_PLAYBACK) ||
658 	     (pipe->conn_type == SKL_PIPE_CONN_TYPE_BE &&
659 	     pipe->direction == SNDRV_PCM_STREAM_CAPTURE))
660 		in_fmt = true;
661 
662 	for (i = 0; i < pipe->nr_cfgs; i++) {
663 		pconfig = &pipe->configs[i];
664 		if (in_fmt)
665 			fmt = &pconfig->in_fmt;
666 		else
667 			fmt = &pconfig->out_fmt;
668 
669 		if (CHECK_HW_PARAMS(params->ch, params->s_freq, params->s_fmt,
670 				    fmt->channels, fmt->freq, fmt->bps)) {
671 			pipe->cur_config_idx = i;
672 			pipe->memory_pages = pconfig->mem_pages;
673 			dev_dbg(skl->dev, "Using pipe config: %d\n", i);
674 
675 			return 0;
676 		}
677 	}
678 
679 	dev_err(skl->dev, "Invalid pipe config: %d %d %d for pipe: %d\n",
680 		params->ch, params->s_freq, params->s_fmt, pipe->ppl_id);
681 	return -EINVAL;
682 }
683 
684 /*
685  * Mixer module represents a pipeline. So in the Pre-PMU event of mixer we
686  * need create the pipeline. So we do following:
687  *   - Create the pipeline
688  *   - Initialize the modules in pipeline
689  *   - finally bind all modules together
690  */
skl_tplg_mixer_dapm_pre_pmu_event(struct snd_soc_dapm_widget * w,struct skl_dev * skl)691 static int skl_tplg_mixer_dapm_pre_pmu_event(struct snd_soc_dapm_widget *w,
692 							struct skl_dev *skl)
693 {
694 	int ret;
695 	struct skl_module_cfg *mconfig = w->priv;
696 	struct skl_pipe_module *w_module;
697 	struct skl_pipe *s_pipe = mconfig->pipe;
698 	struct skl_module_cfg *src_module = NULL, *dst_module, *module;
699 	struct skl_module_deferred_bind *modules;
700 
701 	ret = skl_tplg_get_pipe_config(skl, mconfig);
702 	if (ret < 0)
703 		return ret;
704 
705 	/*
706 	 * Create a list of modules for pipe.
707 	 * This list contains modules from source to sink
708 	 */
709 	ret = skl_create_pipeline(skl, mconfig->pipe);
710 	if (ret < 0)
711 		return ret;
712 
713 	/* Init all pipe modules from source to sink */
714 	ret = skl_tplg_init_pipe_modules(skl, s_pipe);
715 	if (ret < 0)
716 		return ret;
717 
718 	/* Bind modules from source to sink */
719 	list_for_each_entry(w_module, &s_pipe->w_list, node) {
720 		dst_module = w_module->w->priv;
721 
722 		if (src_module == NULL) {
723 			src_module = dst_module;
724 			continue;
725 		}
726 
727 		ret = skl_bind_modules(skl, src_module, dst_module);
728 		if (ret < 0)
729 			return ret;
730 
731 		src_module = dst_module;
732 	}
733 
734 	/*
735 	 * When the destination module is initialized, check for these modules
736 	 * in deferred bind list. If found, bind them.
737 	 */
738 	list_for_each_entry(w_module, &s_pipe->w_list, node) {
739 		if (list_empty(&skl->bind_list))
740 			break;
741 
742 		list_for_each_entry(modules, &skl->bind_list, node) {
743 			module = w_module->w->priv;
744 			if (modules->dst == module)
745 				skl_bind_modules(skl, modules->src,
746 							modules->dst);
747 		}
748 	}
749 
750 	return 0;
751 }
752 
skl_fill_sink_instance_id(struct skl_dev * skl,u32 * params,int size,struct skl_module_cfg * mcfg)753 static int skl_fill_sink_instance_id(struct skl_dev *skl, u32 *params,
754 				int size, struct skl_module_cfg *mcfg)
755 {
756 	int i, pvt_id;
757 
758 	if (mcfg->m_type == SKL_MODULE_TYPE_KPB) {
759 		struct skl_kpb_params *kpb_params =
760 				(struct skl_kpb_params *)params;
761 		struct skl_mod_inst_map *inst = kpb_params->u.map;
762 
763 		for (i = 0; i < kpb_params->num_modules; i++) {
764 			pvt_id = skl_get_pvt_instance_id_map(skl, inst->mod_id,
765 								inst->inst_id);
766 			if (pvt_id < 0)
767 				return -EINVAL;
768 
769 			inst->inst_id = pvt_id;
770 			inst++;
771 		}
772 	}
773 
774 	return 0;
775 }
776 /*
777  * Some modules require params to be set after the module is bound to
778  * all pins connected.
779  *
780  * The module provider initializes set_param flag for such modules and we
781  * send params after binding
782  */
skl_tplg_set_module_bind_params(struct snd_soc_dapm_widget * w,struct skl_module_cfg * mcfg,struct skl_dev * skl)783 static int skl_tplg_set_module_bind_params(struct snd_soc_dapm_widget *w,
784 			struct skl_module_cfg *mcfg, struct skl_dev *skl)
785 {
786 	int i, ret;
787 	struct skl_module_cfg *mconfig = w->priv;
788 	const struct snd_kcontrol_new *k;
789 	struct soc_bytes_ext *sb;
790 	struct skl_algo_data *bc;
791 	struct skl_specific_cfg *sp_cfg;
792 	u32 *params;
793 
794 	/*
795 	 * check all out/in pins are in bind state.
796 	 * if so set the module param
797 	 */
798 	for (i = 0; i < mcfg->module->max_output_pins; i++) {
799 		if (mcfg->m_out_pin[i].pin_state != SKL_PIN_BIND_DONE)
800 			return 0;
801 	}
802 
803 	for (i = 0; i < mcfg->module->max_input_pins; i++) {
804 		if (mcfg->m_in_pin[i].pin_state != SKL_PIN_BIND_DONE)
805 			return 0;
806 	}
807 
808 	if (mconfig->formats_config[SKL_PARAM_BIND].caps_size > 0 &&
809 	    mconfig->formats_config[SKL_PARAM_BIND].set_params ==
810 								SKL_PARAM_BIND) {
811 		sp_cfg = &mconfig->formats_config[SKL_PARAM_BIND];
812 		ret = skl_set_module_params(skl, sp_cfg->caps,
813 					sp_cfg->caps_size,
814 					sp_cfg->param_id, mconfig);
815 		if (ret < 0)
816 			return ret;
817 	}
818 
819 	for (i = 0; i < w->num_kcontrols; i++) {
820 		k = &w->kcontrol_news[i];
821 		if (k->access & SNDRV_CTL_ELEM_ACCESS_TLV_CALLBACK) {
822 			sb = (void *) k->private_value;
823 			bc = (struct skl_algo_data *)sb->dobj.private;
824 
825 			if (bc->set_params == SKL_PARAM_BIND) {
826 				params = kmemdup(bc->params, bc->max, GFP_KERNEL);
827 				if (!params)
828 					return -ENOMEM;
829 
830 				skl_fill_sink_instance_id(skl, params, bc->max,
831 								mconfig);
832 
833 				ret = skl_set_module_params(skl, params,
834 						bc->max, bc->param_id, mconfig);
835 				kfree(params);
836 
837 				if (ret < 0)
838 					return ret;
839 			}
840 		}
841 	}
842 
843 	return 0;
844 }
845 
skl_get_module_id(struct skl_dev * skl,guid_t * uuid)846 static int skl_get_module_id(struct skl_dev *skl, guid_t *uuid)
847 {
848 	struct uuid_module *module;
849 
850 	list_for_each_entry(module, &skl->uuid_list, list) {
851 		if (guid_equal(uuid, &module->uuid))
852 			return module->id;
853 	}
854 
855 	return -EINVAL;
856 }
857 
skl_tplg_find_moduleid_from_uuid(struct skl_dev * skl,const struct snd_kcontrol_new * k)858 static int skl_tplg_find_moduleid_from_uuid(struct skl_dev *skl,
859 					const struct snd_kcontrol_new *k)
860 {
861 	struct soc_bytes_ext *sb = (void *) k->private_value;
862 	struct skl_algo_data *bc = (struct skl_algo_data *)sb->dobj.private;
863 	struct skl_kpb_params *uuid_params, *params;
864 	struct hdac_bus *bus = skl_to_bus(skl);
865 	int i, size, module_id;
866 
867 	if (bc->set_params == SKL_PARAM_BIND && bc->max) {
868 		uuid_params = (struct skl_kpb_params *)bc->params;
869 		size = struct_size(params, u.map, uuid_params->num_modules);
870 
871 		params = devm_kzalloc(bus->dev, size, GFP_KERNEL);
872 		if (!params)
873 			return -ENOMEM;
874 
875 		params->num_modules = uuid_params->num_modules;
876 
877 		for (i = 0; i < uuid_params->num_modules; i++) {
878 			module_id = skl_get_module_id(skl,
879 				&uuid_params->u.map_uuid[i].mod_uuid);
880 			if (module_id < 0) {
881 				devm_kfree(bus->dev, params);
882 				return -EINVAL;
883 			}
884 
885 			params->u.map[i].mod_id = module_id;
886 			params->u.map[i].inst_id =
887 				uuid_params->u.map_uuid[i].inst_id;
888 		}
889 
890 		devm_kfree(bus->dev, bc->params);
891 		bc->params = (char *)params;
892 		bc->max = size;
893 	}
894 
895 	return 0;
896 }
897 
898 /*
899  * Retrieve the module id from UUID mentioned in the
900  * post bind params
901  */
skl_tplg_add_moduleid_in_bind_params(struct skl_dev * skl,struct snd_soc_dapm_widget * w)902 void skl_tplg_add_moduleid_in_bind_params(struct skl_dev *skl,
903 				struct snd_soc_dapm_widget *w)
904 {
905 	struct skl_module_cfg *mconfig = w->priv;
906 	int i;
907 
908 	/*
909 	 * Post bind params are used for only for KPB
910 	 * to set copier instances to drain the data
911 	 * in fast mode
912 	 */
913 	if (mconfig->m_type != SKL_MODULE_TYPE_KPB)
914 		return;
915 
916 	for (i = 0; i < w->num_kcontrols; i++)
917 		if ((w->kcontrol_news[i].access &
918 			SNDRV_CTL_ELEM_ACCESS_TLV_CALLBACK) &&
919 			(skl_tplg_find_moduleid_from_uuid(skl,
920 			&w->kcontrol_news[i]) < 0))
921 			dev_err(skl->dev,
922 				"%s: invalid kpb post bind params\n",
923 				__func__);
924 }
925 
skl_tplg_module_add_deferred_bind(struct skl_dev * skl,struct skl_module_cfg * src,struct skl_module_cfg * dst)926 static int skl_tplg_module_add_deferred_bind(struct skl_dev *skl,
927 	struct skl_module_cfg *src, struct skl_module_cfg *dst)
928 {
929 	struct skl_module_deferred_bind *m_list, *modules;
930 	int i;
931 
932 	/* only supported for module with static pin connection */
933 	for (i = 0; i < dst->module->max_input_pins; i++) {
934 		struct skl_module_pin *pin = &dst->m_in_pin[i];
935 
936 		if (pin->is_dynamic)
937 			continue;
938 
939 		if ((pin->id.module_id  == src->id.module_id) &&
940 			(pin->id.instance_id  == src->id.instance_id)) {
941 
942 			if (!list_empty(&skl->bind_list)) {
943 				list_for_each_entry(modules, &skl->bind_list, node) {
944 					if (modules->src == src && modules->dst == dst)
945 						return 0;
946 				}
947 			}
948 
949 			m_list = kzalloc(sizeof(*m_list), GFP_KERNEL);
950 			if (!m_list)
951 				return -ENOMEM;
952 
953 			m_list->src = src;
954 			m_list->dst = dst;
955 
956 			list_add(&m_list->node, &skl->bind_list);
957 		}
958 	}
959 
960 	return 0;
961 }
962 
skl_tplg_bind_sinks(struct snd_soc_dapm_widget * w,struct skl_dev * skl,struct snd_soc_dapm_widget * src_w,struct skl_module_cfg * src_mconfig)963 static int skl_tplg_bind_sinks(struct snd_soc_dapm_widget *w,
964 				struct skl_dev *skl,
965 				struct snd_soc_dapm_widget *src_w,
966 				struct skl_module_cfg *src_mconfig)
967 {
968 	struct snd_soc_dapm_path *p;
969 	struct snd_soc_dapm_widget *sink = NULL, *next_sink = NULL;
970 	struct skl_module_cfg *sink_mconfig;
971 	int ret;
972 
973 	snd_soc_dapm_widget_for_each_sink_path(w, p) {
974 		if (!p->connect)
975 			continue;
976 
977 		dev_dbg(skl->dev,
978 			"%s: src widget=%s\n", __func__, w->name);
979 		dev_dbg(skl->dev,
980 			"%s: sink widget=%s\n", __func__, p->sink->name);
981 
982 		next_sink = p->sink;
983 
984 		if (!is_skl_dsp_widget_type(p->sink, skl->dev))
985 			return skl_tplg_bind_sinks(p->sink, skl, src_w, src_mconfig);
986 
987 		/*
988 		 * here we will check widgets in sink pipelines, so that
989 		 * can be any widgets type and we are only interested if
990 		 * they are ones used for SKL so check that first
991 		 */
992 		if ((p->sink->priv != NULL) &&
993 				is_skl_dsp_widget_type(p->sink, skl->dev)) {
994 
995 			sink = p->sink;
996 			sink_mconfig = sink->priv;
997 
998 			/*
999 			 * Modules other than PGA leaf can be connected
1000 			 * directly or via switch to a module in another
1001 			 * pipeline. EX: reference path
1002 			 * when the path is enabled, the dst module that needs
1003 			 * to be bound may not be initialized. if the module is
1004 			 * not initialized, add these modules in the deferred
1005 			 * bind list and when the dst module is initialised,
1006 			 * bind this module to the dst_module in deferred list.
1007 			 */
1008 			if (((src_mconfig->m_state == SKL_MODULE_INIT_DONE)
1009 				&& (sink_mconfig->m_state == SKL_MODULE_UNINIT))) {
1010 
1011 				ret = skl_tplg_module_add_deferred_bind(skl,
1012 						src_mconfig, sink_mconfig);
1013 
1014 				if (ret < 0)
1015 					return ret;
1016 
1017 			}
1018 
1019 
1020 			if (src_mconfig->m_state == SKL_MODULE_UNINIT ||
1021 				sink_mconfig->m_state == SKL_MODULE_UNINIT)
1022 				continue;
1023 
1024 			/* Bind source to sink, mixin is always source */
1025 			ret = skl_bind_modules(skl, src_mconfig, sink_mconfig);
1026 			if (ret)
1027 				return ret;
1028 
1029 			/* set module params after bind */
1030 			skl_tplg_set_module_bind_params(src_w,
1031 					src_mconfig, skl);
1032 			skl_tplg_set_module_bind_params(sink,
1033 					sink_mconfig, skl);
1034 
1035 			/* Start sinks pipe first */
1036 			if (sink_mconfig->pipe->state != SKL_PIPE_STARTED) {
1037 				if (sink_mconfig->pipe->conn_type !=
1038 							SKL_PIPE_CONN_TYPE_FE)
1039 					ret = skl_run_pipe(skl,
1040 							sink_mconfig->pipe);
1041 				if (ret)
1042 					return ret;
1043 			}
1044 		}
1045 	}
1046 
1047 	if (!sink && next_sink)
1048 		return skl_tplg_bind_sinks(next_sink, skl, src_w, src_mconfig);
1049 
1050 	return 0;
1051 }
1052 
1053 /*
1054  * A PGA represents a module in a pipeline. So in the Pre-PMU event of PGA
1055  * we need to do following:
1056  *   - Bind to sink pipeline
1057  *      Since the sink pipes can be running and we don't get mixer event on
1058  *      connect for already running mixer, we need to find the sink pipes
1059  *      here and bind to them. This way dynamic connect works.
1060  *   - Start sink pipeline, if not running
1061  *   - Then run current pipe
1062  */
skl_tplg_pga_dapm_pre_pmu_event(struct snd_soc_dapm_widget * w,struct skl_dev * skl)1063 static int skl_tplg_pga_dapm_pre_pmu_event(struct snd_soc_dapm_widget *w,
1064 							struct skl_dev *skl)
1065 {
1066 	struct skl_module_cfg *src_mconfig;
1067 	int ret = 0;
1068 
1069 	src_mconfig = w->priv;
1070 
1071 	/*
1072 	 * find which sink it is connected to, bind with the sink,
1073 	 * if sink is not started, start sink pipe first, then start
1074 	 * this pipe
1075 	 */
1076 	ret = skl_tplg_bind_sinks(w, skl, w, src_mconfig);
1077 	if (ret)
1078 		return ret;
1079 
1080 	/* Start source pipe last after starting all sinks */
1081 	if (src_mconfig->pipe->conn_type != SKL_PIPE_CONN_TYPE_FE)
1082 		return skl_run_pipe(skl, src_mconfig->pipe);
1083 
1084 	return 0;
1085 }
1086 
skl_get_src_dsp_widget(struct snd_soc_dapm_widget * w,struct skl_dev * skl)1087 static struct snd_soc_dapm_widget *skl_get_src_dsp_widget(
1088 		struct snd_soc_dapm_widget *w, struct skl_dev *skl)
1089 {
1090 	struct snd_soc_dapm_path *p;
1091 	struct snd_soc_dapm_widget *src_w = NULL;
1092 
1093 	snd_soc_dapm_widget_for_each_source_path(w, p) {
1094 		src_w = p->source;
1095 		if (!p->connect)
1096 			continue;
1097 
1098 		dev_dbg(skl->dev, "sink widget=%s\n", w->name);
1099 		dev_dbg(skl->dev, "src widget=%s\n", p->source->name);
1100 
1101 		/*
1102 		 * here we will check widgets in sink pipelines, so that can
1103 		 * be any widgets type and we are only interested if they are
1104 		 * ones used for SKL so check that first
1105 		 */
1106 		if ((p->source->priv != NULL) &&
1107 				is_skl_dsp_widget_type(p->source, skl->dev)) {
1108 			return p->source;
1109 		}
1110 	}
1111 
1112 	if (src_w != NULL)
1113 		return skl_get_src_dsp_widget(src_w, skl);
1114 
1115 	return NULL;
1116 }
1117 
1118 /*
1119  * in the Post-PMU event of mixer we need to do following:
1120  *   - Check if this pipe is running
1121  *   - if not, then
1122  *	- bind this pipeline to its source pipeline
1123  *	  if source pipe is already running, this means it is a dynamic
1124  *	  connection and we need to bind only to that pipe
1125  *	- start this pipeline
1126  */
skl_tplg_mixer_dapm_post_pmu_event(struct snd_soc_dapm_widget * w,struct skl_dev * skl)1127 static int skl_tplg_mixer_dapm_post_pmu_event(struct snd_soc_dapm_widget *w,
1128 							struct skl_dev *skl)
1129 {
1130 	int ret = 0;
1131 	struct snd_soc_dapm_widget *source, *sink;
1132 	struct skl_module_cfg *src_mconfig, *sink_mconfig;
1133 	int src_pipe_started = 0;
1134 
1135 	sink = w;
1136 	sink_mconfig = sink->priv;
1137 
1138 	/*
1139 	 * If source pipe is already started, that means source is driving
1140 	 * one more sink before this sink got connected, Since source is
1141 	 * started, bind this sink to source and start this pipe.
1142 	 */
1143 	source = skl_get_src_dsp_widget(w, skl);
1144 	if (source != NULL) {
1145 		src_mconfig = source->priv;
1146 		sink_mconfig = sink->priv;
1147 		src_pipe_started = 1;
1148 
1149 		/*
1150 		 * check pipe state, then no need to bind or start the
1151 		 * pipe
1152 		 */
1153 		if (src_mconfig->pipe->state != SKL_PIPE_STARTED)
1154 			src_pipe_started = 0;
1155 	}
1156 
1157 	if (src_pipe_started) {
1158 		ret = skl_bind_modules(skl, src_mconfig, sink_mconfig);
1159 		if (ret)
1160 			return ret;
1161 
1162 		/* set module params after bind */
1163 		skl_tplg_set_module_bind_params(source, src_mconfig, skl);
1164 		skl_tplg_set_module_bind_params(sink, sink_mconfig, skl);
1165 
1166 		if (sink_mconfig->pipe->conn_type != SKL_PIPE_CONN_TYPE_FE)
1167 			ret = skl_run_pipe(skl, sink_mconfig->pipe);
1168 	}
1169 
1170 	return ret;
1171 }
1172 
1173 /*
1174  * in the Pre-PMD event of mixer we need to do following:
1175  *   - Stop the pipe
1176  *   - find the source connections and remove that from dapm_path_list
1177  *   - unbind with source pipelines if still connected
1178  */
skl_tplg_mixer_dapm_pre_pmd_event(struct snd_soc_dapm_widget * w,struct skl_dev * skl)1179 static int skl_tplg_mixer_dapm_pre_pmd_event(struct snd_soc_dapm_widget *w,
1180 							struct skl_dev *skl)
1181 {
1182 	struct skl_module_cfg *src_mconfig, *sink_mconfig;
1183 	int ret = 0, i;
1184 
1185 	sink_mconfig = w->priv;
1186 
1187 	/* Stop the pipe */
1188 	ret = skl_stop_pipe(skl, sink_mconfig->pipe);
1189 	if (ret)
1190 		return ret;
1191 
1192 	for (i = 0; i < sink_mconfig->module->max_input_pins; i++) {
1193 		if (sink_mconfig->m_in_pin[i].pin_state == SKL_PIN_BIND_DONE) {
1194 			src_mconfig = sink_mconfig->m_in_pin[i].tgt_mcfg;
1195 			if (!src_mconfig)
1196 				continue;
1197 
1198 			ret = skl_unbind_modules(skl,
1199 						src_mconfig, sink_mconfig);
1200 		}
1201 	}
1202 
1203 	return ret;
1204 }
1205 
1206 /*
1207  * in the Post-PMD event of mixer we need to do following:
1208  *   - Unbind the modules within the pipeline
1209  *   - Delete the pipeline (modules are not required to be explicitly
1210  *     deleted, pipeline delete is enough here
1211  */
skl_tplg_mixer_dapm_post_pmd_event(struct snd_soc_dapm_widget * w,struct skl_dev * skl)1212 static int skl_tplg_mixer_dapm_post_pmd_event(struct snd_soc_dapm_widget *w,
1213 							struct skl_dev *skl)
1214 {
1215 	struct skl_module_cfg *mconfig = w->priv;
1216 	struct skl_pipe_module *w_module;
1217 	struct skl_module_cfg *src_module = NULL, *dst_module;
1218 	struct skl_pipe *s_pipe = mconfig->pipe;
1219 	struct skl_module_deferred_bind *modules, *tmp;
1220 
1221 	if (s_pipe->state == SKL_PIPE_INVALID)
1222 		return -EINVAL;
1223 
1224 	list_for_each_entry(w_module, &s_pipe->w_list, node) {
1225 		if (list_empty(&skl->bind_list))
1226 			break;
1227 
1228 		src_module = w_module->w->priv;
1229 
1230 		list_for_each_entry_safe(modules, tmp, &skl->bind_list, node) {
1231 			/*
1232 			 * When the destination module is deleted, Unbind the
1233 			 * modules from deferred bind list.
1234 			 */
1235 			if (modules->dst == src_module) {
1236 				skl_unbind_modules(skl, modules->src,
1237 						modules->dst);
1238 			}
1239 
1240 			/*
1241 			 * When the source module is deleted, remove this entry
1242 			 * from the deferred bind list.
1243 			 */
1244 			if (modules->src == src_module) {
1245 				list_del(&modules->node);
1246 				modules->src = NULL;
1247 				modules->dst = NULL;
1248 				kfree(modules);
1249 			}
1250 		}
1251 	}
1252 
1253 	list_for_each_entry(w_module, &s_pipe->w_list, node) {
1254 		dst_module = w_module->w->priv;
1255 
1256 		if (src_module == NULL) {
1257 			src_module = dst_module;
1258 			continue;
1259 		}
1260 
1261 		skl_unbind_modules(skl, src_module, dst_module);
1262 		src_module = dst_module;
1263 	}
1264 
1265 	skl_delete_pipe(skl, mconfig->pipe);
1266 
1267 	list_for_each_entry(w_module, &s_pipe->w_list, node) {
1268 		src_module = w_module->w->priv;
1269 		src_module->m_state = SKL_MODULE_UNINIT;
1270 	}
1271 
1272 	return skl_tplg_unload_pipe_modules(skl, s_pipe);
1273 }
1274 
1275 /*
1276  * in the Post-PMD event of PGA we need to do following:
1277  *   - Stop the pipeline
1278  *   - In source pipe is connected, unbind with source pipelines
1279  */
skl_tplg_pga_dapm_post_pmd_event(struct snd_soc_dapm_widget * w,struct skl_dev * skl)1280 static int skl_tplg_pga_dapm_post_pmd_event(struct snd_soc_dapm_widget *w,
1281 							struct skl_dev *skl)
1282 {
1283 	struct skl_module_cfg *src_mconfig, *sink_mconfig;
1284 	int ret = 0, i;
1285 
1286 	src_mconfig = w->priv;
1287 
1288 	/* Stop the pipe since this is a mixin module */
1289 	ret = skl_stop_pipe(skl, src_mconfig->pipe);
1290 	if (ret)
1291 		return ret;
1292 
1293 	for (i = 0; i < src_mconfig->module->max_output_pins; i++) {
1294 		if (src_mconfig->m_out_pin[i].pin_state == SKL_PIN_BIND_DONE) {
1295 			sink_mconfig = src_mconfig->m_out_pin[i].tgt_mcfg;
1296 			if (!sink_mconfig)
1297 				continue;
1298 			/*
1299 			 * This is a connecter and if path is found that means
1300 			 * unbind between source and sink has not happened yet
1301 			 */
1302 			ret = skl_unbind_modules(skl, src_mconfig,
1303 							sink_mconfig);
1304 		}
1305 	}
1306 
1307 	return ret;
1308 }
1309 
1310 /*
1311  * In modelling, we assume there will be ONLY one mixer in a pipeline. If a
1312  * second one is required that is created as another pipe entity.
1313  * The mixer is responsible for pipe management and represent a pipeline
1314  * instance
1315  */
skl_tplg_mixer_event(struct snd_soc_dapm_widget * w,struct snd_kcontrol * k,int event)1316 static int skl_tplg_mixer_event(struct snd_soc_dapm_widget *w,
1317 				struct snd_kcontrol *k, int event)
1318 {
1319 	struct snd_soc_dapm_context *dapm = w->dapm;
1320 	struct skl_dev *skl = get_skl_ctx(dapm->dev);
1321 
1322 	switch (event) {
1323 	case SND_SOC_DAPM_PRE_PMU:
1324 		return skl_tplg_mixer_dapm_pre_pmu_event(w, skl);
1325 
1326 	case SND_SOC_DAPM_POST_PMU:
1327 		return skl_tplg_mixer_dapm_post_pmu_event(w, skl);
1328 
1329 	case SND_SOC_DAPM_PRE_PMD:
1330 		return skl_tplg_mixer_dapm_pre_pmd_event(w, skl);
1331 
1332 	case SND_SOC_DAPM_POST_PMD:
1333 		return skl_tplg_mixer_dapm_post_pmd_event(w, skl);
1334 	}
1335 
1336 	return 0;
1337 }
1338 
1339 /*
1340  * In modelling, we assumed rest of the modules in pipeline are PGA. But we
1341  * are interested in last PGA (leaf PGA) in a pipeline to disconnect with
1342  * the sink when it is running (two FE to one BE or one FE to two BE)
1343  * scenarios
1344  */
skl_tplg_pga_event(struct snd_soc_dapm_widget * w,struct snd_kcontrol * k,int event)1345 static int skl_tplg_pga_event(struct snd_soc_dapm_widget *w,
1346 			struct snd_kcontrol *k, int event)
1347 
1348 {
1349 	struct snd_soc_dapm_context *dapm = w->dapm;
1350 	struct skl_dev *skl = get_skl_ctx(dapm->dev);
1351 
1352 	switch (event) {
1353 	case SND_SOC_DAPM_PRE_PMU:
1354 		return skl_tplg_pga_dapm_pre_pmu_event(w, skl);
1355 
1356 	case SND_SOC_DAPM_POST_PMD:
1357 		return skl_tplg_pga_dapm_post_pmd_event(w, skl);
1358 	}
1359 
1360 	return 0;
1361 }
1362 
skl_tplg_multi_config_set_get(struct snd_kcontrol * kcontrol,struct snd_ctl_elem_value * ucontrol,bool is_set)1363 static int skl_tplg_multi_config_set_get(struct snd_kcontrol *kcontrol,
1364 					 struct snd_ctl_elem_value *ucontrol,
1365 					 bool is_set)
1366 {
1367 	struct snd_soc_component *component =
1368 		snd_soc_kcontrol_component(kcontrol);
1369 	struct hdac_bus *bus = snd_soc_component_get_drvdata(component);
1370 	struct skl_dev *skl = bus_to_skl(bus);
1371 	struct skl_pipeline *ppl;
1372 	struct skl_pipe *pipe = NULL;
1373 	struct soc_enum *ec = (struct soc_enum *)kcontrol->private_value;
1374 	u32 *pipe_id;
1375 
1376 	if (!ec)
1377 		return -EINVAL;
1378 
1379 	if (is_set && ucontrol->value.enumerated.item[0] > ec->items)
1380 		return -EINVAL;
1381 
1382 	pipe_id = ec->dobj.private;
1383 
1384 	list_for_each_entry(ppl, &skl->ppl_list, node) {
1385 		if (ppl->pipe->ppl_id == *pipe_id) {
1386 			pipe = ppl->pipe;
1387 			break;
1388 		}
1389 	}
1390 	if (!pipe)
1391 		return -EIO;
1392 
1393 	if (is_set)
1394 		pipe->pipe_config_idx = ucontrol->value.enumerated.item[0];
1395 	else
1396 		ucontrol->value.enumerated.item[0]  =  pipe->pipe_config_idx;
1397 
1398 	return 0;
1399 }
1400 
skl_tplg_multi_config_get(struct snd_kcontrol * kcontrol,struct snd_ctl_elem_value * ucontrol)1401 static int skl_tplg_multi_config_get(struct snd_kcontrol *kcontrol,
1402 				     struct snd_ctl_elem_value *ucontrol)
1403 {
1404 	return skl_tplg_multi_config_set_get(kcontrol, ucontrol, false);
1405 }
1406 
skl_tplg_multi_config_set(struct snd_kcontrol * kcontrol,struct snd_ctl_elem_value * ucontrol)1407 static int skl_tplg_multi_config_set(struct snd_kcontrol *kcontrol,
1408 				     struct snd_ctl_elem_value *ucontrol)
1409 {
1410 	return skl_tplg_multi_config_set_get(kcontrol, ucontrol, true);
1411 }
1412 
skl_tplg_multi_config_get_dmic(struct snd_kcontrol * kcontrol,struct snd_ctl_elem_value * ucontrol)1413 static int skl_tplg_multi_config_get_dmic(struct snd_kcontrol *kcontrol,
1414 					  struct snd_ctl_elem_value *ucontrol)
1415 {
1416 	return skl_tplg_multi_config_set_get(kcontrol, ucontrol, false);
1417 }
1418 
skl_tplg_multi_config_set_dmic(struct snd_kcontrol * kcontrol,struct snd_ctl_elem_value * ucontrol)1419 static int skl_tplg_multi_config_set_dmic(struct snd_kcontrol *kcontrol,
1420 					  struct snd_ctl_elem_value *ucontrol)
1421 {
1422 	return skl_tplg_multi_config_set_get(kcontrol, ucontrol, true);
1423 }
1424 
skl_tplg_tlv_control_get(struct snd_kcontrol * kcontrol,unsigned int __user * data,unsigned int size)1425 static int skl_tplg_tlv_control_get(struct snd_kcontrol *kcontrol,
1426 			unsigned int __user *data, unsigned int size)
1427 {
1428 	struct soc_bytes_ext *sb =
1429 			(struct soc_bytes_ext *)kcontrol->private_value;
1430 	struct skl_algo_data *bc = (struct skl_algo_data *)sb->dobj.private;
1431 	struct snd_soc_dapm_widget *w = snd_soc_dapm_kcontrol_widget(kcontrol);
1432 	struct skl_module_cfg *mconfig = w->priv;
1433 	struct skl_dev *skl = get_skl_ctx(w->dapm->dev);
1434 
1435 	if (w->power)
1436 		skl_get_module_params(skl, (u32 *)bc->params,
1437 				      bc->size, bc->param_id, mconfig);
1438 
1439 	/* decrement size for TLV header */
1440 	size -= 2 * sizeof(u32);
1441 
1442 	/* check size as we don't want to send kernel data */
1443 	if (size > bc->max)
1444 		size = bc->max;
1445 
1446 	if (bc->params) {
1447 		if (copy_to_user(data, &bc->param_id, sizeof(u32)))
1448 			return -EFAULT;
1449 		if (copy_to_user(data + 1, &size, sizeof(u32)))
1450 			return -EFAULT;
1451 		if (copy_to_user(data + 2, bc->params, size))
1452 			return -EFAULT;
1453 	}
1454 
1455 	return 0;
1456 }
1457 
1458 #define SKL_PARAM_VENDOR_ID 0xff
1459 
skl_tplg_tlv_control_set(struct snd_kcontrol * kcontrol,const unsigned int __user * data,unsigned int size)1460 static int skl_tplg_tlv_control_set(struct snd_kcontrol *kcontrol,
1461 			const unsigned int __user *data, unsigned int size)
1462 {
1463 	struct snd_soc_dapm_widget *w = snd_soc_dapm_kcontrol_widget(kcontrol);
1464 	struct skl_module_cfg *mconfig = w->priv;
1465 	struct soc_bytes_ext *sb =
1466 			(struct soc_bytes_ext *)kcontrol->private_value;
1467 	struct skl_algo_data *ac = (struct skl_algo_data *)sb->dobj.private;
1468 	struct skl_dev *skl = get_skl_ctx(w->dapm->dev);
1469 
1470 	if (ac->params) {
1471 		if (size > ac->max)
1472 			return -EINVAL;
1473 		ac->size = size;
1474 
1475 		if (copy_from_user(ac->params, data, size))
1476 			return -EFAULT;
1477 
1478 		if (w->power)
1479 			return skl_set_module_params(skl,
1480 						(u32 *)ac->params, ac->size,
1481 						ac->param_id, mconfig);
1482 	}
1483 
1484 	return 0;
1485 }
1486 
skl_tplg_mic_control_get(struct snd_kcontrol * kcontrol,struct snd_ctl_elem_value * ucontrol)1487 static int skl_tplg_mic_control_get(struct snd_kcontrol *kcontrol,
1488 		struct snd_ctl_elem_value *ucontrol)
1489 {
1490 	struct snd_soc_dapm_widget *w = snd_soc_dapm_kcontrol_widget(kcontrol);
1491 	struct skl_module_cfg *mconfig = w->priv;
1492 	struct soc_enum *ec = (struct soc_enum *)kcontrol->private_value;
1493 	u32 ch_type = *((u32 *)ec->dobj.private);
1494 
1495 	if (mconfig->dmic_ch_type == ch_type)
1496 		ucontrol->value.enumerated.item[0] =
1497 					mconfig->dmic_ch_combo_index;
1498 	else
1499 		ucontrol->value.enumerated.item[0] = 0;
1500 
1501 	return 0;
1502 }
1503 
skl_fill_mic_sel_params(struct skl_module_cfg * mconfig,struct skl_mic_sel_config * mic_cfg,struct device * dev)1504 static int skl_fill_mic_sel_params(struct skl_module_cfg *mconfig,
1505 	struct skl_mic_sel_config *mic_cfg, struct device *dev)
1506 {
1507 	struct skl_specific_cfg *sp_cfg =
1508 				&mconfig->formats_config[SKL_PARAM_INIT];
1509 
1510 	sp_cfg->caps_size = sizeof(struct skl_mic_sel_config);
1511 	sp_cfg->set_params = SKL_PARAM_SET;
1512 	sp_cfg->param_id = 0x00;
1513 	if (!sp_cfg->caps) {
1514 		sp_cfg->caps = devm_kzalloc(dev, sp_cfg->caps_size, GFP_KERNEL);
1515 		if (!sp_cfg->caps)
1516 			return -ENOMEM;
1517 	}
1518 
1519 	mic_cfg->mic_switch = SKL_MIC_SEL_SWITCH;
1520 	mic_cfg->flags = 0;
1521 	memcpy(sp_cfg->caps, mic_cfg, sp_cfg->caps_size);
1522 
1523 	return 0;
1524 }
1525 
skl_tplg_mic_control_set(struct snd_kcontrol * kcontrol,struct snd_ctl_elem_value * ucontrol)1526 static int skl_tplg_mic_control_set(struct snd_kcontrol *kcontrol,
1527 			struct snd_ctl_elem_value *ucontrol)
1528 {
1529 	struct snd_soc_dapm_widget *w = snd_soc_dapm_kcontrol_widget(kcontrol);
1530 	struct skl_module_cfg *mconfig = w->priv;
1531 	struct skl_mic_sel_config mic_cfg = {0};
1532 	struct soc_enum *ec = (struct soc_enum *)kcontrol->private_value;
1533 	u32 ch_type = *((u32 *)ec->dobj.private);
1534 	const int *list;
1535 	u8 in_ch, out_ch, index;
1536 
1537 	mconfig->dmic_ch_type = ch_type;
1538 	mconfig->dmic_ch_combo_index = ucontrol->value.enumerated.item[0];
1539 
1540 	/* enum control index 0 is INVALID, so no channels to be set */
1541 	if (mconfig->dmic_ch_combo_index == 0)
1542 		return 0;
1543 
1544 	/* No valid channel selection map for index 0, so offset by 1 */
1545 	index = mconfig->dmic_ch_combo_index - 1;
1546 
1547 	switch (ch_type) {
1548 	case SKL_CH_MONO:
1549 		if (mconfig->dmic_ch_combo_index > ARRAY_SIZE(mic_mono_list))
1550 			return -EINVAL;
1551 
1552 		list = &mic_mono_list[index];
1553 		break;
1554 
1555 	case SKL_CH_STEREO:
1556 		if (mconfig->dmic_ch_combo_index > ARRAY_SIZE(mic_stereo_list))
1557 			return -EINVAL;
1558 
1559 		list = mic_stereo_list[index];
1560 		break;
1561 
1562 	case SKL_CH_TRIO:
1563 		if (mconfig->dmic_ch_combo_index > ARRAY_SIZE(mic_trio_list))
1564 			return -EINVAL;
1565 
1566 		list = mic_trio_list[index];
1567 		break;
1568 
1569 	case SKL_CH_QUATRO:
1570 		if (mconfig->dmic_ch_combo_index > ARRAY_SIZE(mic_quatro_list))
1571 			return -EINVAL;
1572 
1573 		list = mic_quatro_list[index];
1574 		break;
1575 
1576 	default:
1577 		dev_err(w->dapm->dev,
1578 				"Invalid channel %d for mic_select module\n",
1579 				ch_type);
1580 		return -EINVAL;
1581 
1582 	}
1583 
1584 	/* channel type enum map to number of chanels for that type */
1585 	for (out_ch = 0; out_ch < ch_type; out_ch++) {
1586 		in_ch = list[out_ch];
1587 		mic_cfg.blob[out_ch][in_ch] = SKL_DEFAULT_MIC_SEL_GAIN;
1588 	}
1589 
1590 	return skl_fill_mic_sel_params(mconfig, &mic_cfg, w->dapm->dev);
1591 }
1592 
1593 /*
1594  * Fill the dma id for host and link. In case of passthrough
1595  * pipeline, this will both host and link in the same
1596  * pipeline, so need to copy the link and host based on dev_type
1597  */
skl_tplg_fill_dma_id(struct skl_module_cfg * mcfg,struct skl_pipe_params * params)1598 static void skl_tplg_fill_dma_id(struct skl_module_cfg *mcfg,
1599 				struct skl_pipe_params *params)
1600 {
1601 	struct skl_pipe *pipe = mcfg->pipe;
1602 
1603 	if (pipe->passthru) {
1604 		switch (mcfg->dev_type) {
1605 		case SKL_DEVICE_HDALINK:
1606 			pipe->p_params->link_dma_id = params->link_dma_id;
1607 			pipe->p_params->link_index = params->link_index;
1608 			pipe->p_params->link_bps = params->link_bps;
1609 			break;
1610 
1611 		case SKL_DEVICE_HDAHOST:
1612 			pipe->p_params->host_dma_id = params->host_dma_id;
1613 			pipe->p_params->host_bps = params->host_bps;
1614 			break;
1615 
1616 		default:
1617 			break;
1618 		}
1619 		pipe->p_params->s_fmt = params->s_fmt;
1620 		pipe->p_params->ch = params->ch;
1621 		pipe->p_params->s_freq = params->s_freq;
1622 		pipe->p_params->stream = params->stream;
1623 		pipe->p_params->format = params->format;
1624 
1625 	} else {
1626 		memcpy(pipe->p_params, params, sizeof(*params));
1627 	}
1628 }
1629 
1630 /*
1631  * The FE params are passed by hw_params of the DAI.
1632  * On hw_params, the params are stored in Gateway module of the FE and we
1633  * need to calculate the format in DSP module configuration, that
1634  * conversion is done here
1635  */
skl_tplg_update_pipe_params(struct device * dev,struct skl_module_cfg * mconfig,struct skl_pipe_params * params)1636 int skl_tplg_update_pipe_params(struct device *dev,
1637 			struct skl_module_cfg *mconfig,
1638 			struct skl_pipe_params *params)
1639 {
1640 	struct skl_module_res *res;
1641 	struct skl_dev *skl = get_skl_ctx(dev);
1642 	struct skl_module_fmt *format = NULL;
1643 	u8 cfg_idx = mconfig->pipe->cur_config_idx;
1644 
1645 	res = &mconfig->module->resources[mconfig->res_idx];
1646 	skl_tplg_fill_dma_id(mconfig, params);
1647 	mconfig->fmt_idx = mconfig->mod_cfg[cfg_idx].fmt_idx;
1648 	mconfig->res_idx = mconfig->mod_cfg[cfg_idx].res_idx;
1649 
1650 	if (skl->nr_modules)
1651 		return 0;
1652 
1653 	if (params->stream == SNDRV_PCM_STREAM_PLAYBACK)
1654 		format = &mconfig->module->formats[mconfig->fmt_idx].inputs[0].fmt;
1655 	else
1656 		format = &mconfig->module->formats[mconfig->fmt_idx].outputs[0].fmt;
1657 
1658 	/* set the hw_params */
1659 	format->s_freq = params->s_freq;
1660 	format->channels = params->ch;
1661 	format->valid_bit_depth = skl_get_bit_depth(params->s_fmt);
1662 
1663 	/*
1664 	 * 16 bit is 16 bit container whereas 24 bit is in 32 bit
1665 	 * container so update bit depth accordingly
1666 	 */
1667 	switch (format->valid_bit_depth) {
1668 	case SKL_DEPTH_16BIT:
1669 		format->bit_depth = format->valid_bit_depth;
1670 		break;
1671 
1672 	case SKL_DEPTH_24BIT:
1673 	case SKL_DEPTH_32BIT:
1674 		format->bit_depth = SKL_DEPTH_32BIT;
1675 		break;
1676 
1677 	default:
1678 		dev_err(dev, "Invalid bit depth %x for pipe\n",
1679 				format->valid_bit_depth);
1680 		return -EINVAL;
1681 	}
1682 
1683 	if (params->stream == SNDRV_PCM_STREAM_PLAYBACK) {
1684 		res->ibs = (format->s_freq / 1000) *
1685 				(format->channels) *
1686 				(format->bit_depth >> 3);
1687 	} else {
1688 		res->obs = (format->s_freq / 1000) *
1689 				(format->channels) *
1690 				(format->bit_depth >> 3);
1691 	}
1692 
1693 	return 0;
1694 }
1695 
1696 /*
1697  * Query the module config for the FE DAI
1698  * This is used to find the hw_params set for that DAI and apply to FE
1699  * pipeline
1700  */
1701 struct skl_module_cfg *
skl_tplg_fe_get_cpr_module(struct snd_soc_dai * dai,int stream)1702 skl_tplg_fe_get_cpr_module(struct snd_soc_dai *dai, int stream)
1703 {
1704 	struct snd_soc_dapm_widget *w;
1705 	struct snd_soc_dapm_path *p = NULL;
1706 
1707 	if (stream == SNDRV_PCM_STREAM_PLAYBACK) {
1708 		w = dai->playback_widget;
1709 		snd_soc_dapm_widget_for_each_sink_path(w, p) {
1710 			if (p->connect && p->sink->power &&
1711 				!is_skl_dsp_widget_type(p->sink, dai->dev))
1712 				continue;
1713 
1714 			if (p->sink->priv) {
1715 				dev_dbg(dai->dev, "set params for %s\n",
1716 						p->sink->name);
1717 				return p->sink->priv;
1718 			}
1719 		}
1720 	} else {
1721 		w = dai->capture_widget;
1722 		snd_soc_dapm_widget_for_each_source_path(w, p) {
1723 			if (p->connect && p->source->power &&
1724 				!is_skl_dsp_widget_type(p->source, dai->dev))
1725 				continue;
1726 
1727 			if (p->source->priv) {
1728 				dev_dbg(dai->dev, "set params for %s\n",
1729 						p->source->name);
1730 				return p->source->priv;
1731 			}
1732 		}
1733 	}
1734 
1735 	return NULL;
1736 }
1737 
skl_get_mconfig_pb_cpr(struct snd_soc_dai * dai,struct snd_soc_dapm_widget * w)1738 static struct skl_module_cfg *skl_get_mconfig_pb_cpr(
1739 		struct snd_soc_dai *dai, struct snd_soc_dapm_widget *w)
1740 {
1741 	struct snd_soc_dapm_path *p;
1742 	struct skl_module_cfg *mconfig = NULL;
1743 
1744 	snd_soc_dapm_widget_for_each_source_path(w, p) {
1745 		if (w->endpoints[SND_SOC_DAPM_DIR_OUT] > 0) {
1746 			if (p->connect &&
1747 				    (p->sink->id == snd_soc_dapm_aif_out) &&
1748 				    p->source->priv) {
1749 				mconfig = p->source->priv;
1750 				return mconfig;
1751 			}
1752 			mconfig = skl_get_mconfig_pb_cpr(dai, p->source);
1753 			if (mconfig)
1754 				return mconfig;
1755 		}
1756 	}
1757 	return mconfig;
1758 }
1759 
skl_get_mconfig_cap_cpr(struct snd_soc_dai * dai,struct snd_soc_dapm_widget * w)1760 static struct skl_module_cfg *skl_get_mconfig_cap_cpr(
1761 		struct snd_soc_dai *dai, struct snd_soc_dapm_widget *w)
1762 {
1763 	struct snd_soc_dapm_path *p;
1764 	struct skl_module_cfg *mconfig = NULL;
1765 
1766 	snd_soc_dapm_widget_for_each_sink_path(w, p) {
1767 		if (w->endpoints[SND_SOC_DAPM_DIR_IN] > 0) {
1768 			if (p->connect &&
1769 				    (p->source->id == snd_soc_dapm_aif_in) &&
1770 				    p->sink->priv) {
1771 				mconfig = p->sink->priv;
1772 				return mconfig;
1773 			}
1774 			mconfig = skl_get_mconfig_cap_cpr(dai, p->sink);
1775 			if (mconfig)
1776 				return mconfig;
1777 		}
1778 	}
1779 	return mconfig;
1780 }
1781 
1782 struct skl_module_cfg *
skl_tplg_be_get_cpr_module(struct snd_soc_dai * dai,int stream)1783 skl_tplg_be_get_cpr_module(struct snd_soc_dai *dai, int stream)
1784 {
1785 	struct snd_soc_dapm_widget *w;
1786 	struct skl_module_cfg *mconfig;
1787 
1788 	if (stream == SNDRV_PCM_STREAM_PLAYBACK) {
1789 		w = dai->playback_widget;
1790 		mconfig = skl_get_mconfig_pb_cpr(dai, w);
1791 	} else {
1792 		w = dai->capture_widget;
1793 		mconfig = skl_get_mconfig_cap_cpr(dai, w);
1794 	}
1795 	return mconfig;
1796 }
1797 
skl_tplg_be_link_type(int dev_type)1798 static u8 skl_tplg_be_link_type(int dev_type)
1799 {
1800 	int ret;
1801 
1802 	switch (dev_type) {
1803 	case SKL_DEVICE_BT:
1804 		ret = NHLT_LINK_SSP;
1805 		break;
1806 
1807 	case SKL_DEVICE_DMIC:
1808 		ret = NHLT_LINK_DMIC;
1809 		break;
1810 
1811 	case SKL_DEVICE_I2S:
1812 		ret = NHLT_LINK_SSP;
1813 		break;
1814 
1815 	case SKL_DEVICE_HDALINK:
1816 		ret = NHLT_LINK_HDA;
1817 		break;
1818 
1819 	default:
1820 		ret = NHLT_LINK_INVALID;
1821 		break;
1822 	}
1823 
1824 	return ret;
1825 }
1826 
1827 /*
1828  * Fill the BE gateway parameters
1829  * The BE gateway expects a blob of parameters which are kept in the ACPI
1830  * NHLT blob, so query the blob for interface type (i2s/pdm) and instance.
1831  * The port can have multiple settings so pick based on the pipeline
1832  * parameters
1833  */
skl_tplg_be_fill_pipe_params(struct snd_soc_dai * dai,struct skl_module_cfg * mconfig,struct skl_pipe_params * params)1834 static int skl_tplg_be_fill_pipe_params(struct snd_soc_dai *dai,
1835 				struct skl_module_cfg *mconfig,
1836 				struct skl_pipe_params *params)
1837 {
1838 	struct nhlt_specific_cfg *cfg;
1839 	struct skl_pipe *pipe = mconfig->pipe;
1840 	struct skl_pipe_fmt *pipe_fmt;
1841 	struct skl_dev *skl = get_skl_ctx(dai->dev);
1842 	int link_type = skl_tplg_be_link_type(mconfig->dev_type);
1843 	u8 dev_type = skl_tplg_be_dev_type(mconfig->dev_type);
1844 
1845 	skl_tplg_fill_dma_id(mconfig, params);
1846 
1847 	if (link_type == NHLT_LINK_HDA)
1848 		return 0;
1849 
1850 	if (pipe->direction == SNDRV_PCM_STREAM_PLAYBACK)
1851 		pipe_fmt = &pipe->configs[pipe->pipe_config_idx].out_fmt;
1852 	else
1853 		pipe_fmt = &pipe->configs[pipe->pipe_config_idx].in_fmt;
1854 
1855 	/* update the blob based on virtual bus_id*/
1856 	cfg = intel_nhlt_get_endpoint_blob(dai->dev, skl->nhlt,
1857 					mconfig->vbus_id, link_type,
1858 					pipe_fmt->bps, params->s_cont,
1859 					pipe_fmt->channels, pipe_fmt->freq,
1860 					pipe->direction, dev_type);
1861 	if (cfg) {
1862 		mconfig->formats_config[SKL_PARAM_INIT].caps_size = cfg->size;
1863 		mconfig->formats_config[SKL_PARAM_INIT].caps = (u32 *)&cfg->caps;
1864 	} else {
1865 		dev_err(dai->dev, "Blob NULL for id:%d type:%d dirn:%d ch:%d, freq:%d, fmt:%d\n",
1866 			mconfig->vbus_id, link_type, params->stream,
1867 			params->ch, params->s_freq, params->s_fmt);
1868 		return -EINVAL;
1869 	}
1870 
1871 	return 0;
1872 }
1873 
skl_tplg_be_set_src_pipe_params(struct snd_soc_dai * dai,struct snd_soc_dapm_widget * w,struct skl_pipe_params * params)1874 static int skl_tplg_be_set_src_pipe_params(struct snd_soc_dai *dai,
1875 				struct snd_soc_dapm_widget *w,
1876 				struct skl_pipe_params *params)
1877 {
1878 	struct snd_soc_dapm_path *p;
1879 	int ret = -EIO;
1880 
1881 	snd_soc_dapm_widget_for_each_source_path(w, p) {
1882 		if (p->connect && is_skl_dsp_widget_type(p->source, dai->dev) &&
1883 						p->source->priv) {
1884 
1885 			ret = skl_tplg_be_fill_pipe_params(dai,
1886 						p->source->priv, params);
1887 			if (ret < 0)
1888 				return ret;
1889 		} else {
1890 			ret = skl_tplg_be_set_src_pipe_params(dai,
1891 						p->source, params);
1892 			if (ret < 0)
1893 				return ret;
1894 		}
1895 	}
1896 
1897 	return ret;
1898 }
1899 
skl_tplg_be_set_sink_pipe_params(struct snd_soc_dai * dai,struct snd_soc_dapm_widget * w,struct skl_pipe_params * params)1900 static int skl_tplg_be_set_sink_pipe_params(struct snd_soc_dai *dai,
1901 	struct snd_soc_dapm_widget *w, struct skl_pipe_params *params)
1902 {
1903 	struct snd_soc_dapm_path *p;
1904 	int ret = -EIO;
1905 
1906 	snd_soc_dapm_widget_for_each_sink_path(w, p) {
1907 		if (p->connect && is_skl_dsp_widget_type(p->sink, dai->dev) &&
1908 						p->sink->priv) {
1909 
1910 			ret = skl_tplg_be_fill_pipe_params(dai,
1911 						p->sink->priv, params);
1912 			if (ret < 0)
1913 				return ret;
1914 		} else {
1915 			ret = skl_tplg_be_set_sink_pipe_params(
1916 						dai, p->sink, params);
1917 			if (ret < 0)
1918 				return ret;
1919 		}
1920 	}
1921 
1922 	return ret;
1923 }
1924 
1925 /*
1926  * BE hw_params can be a source parameters (capture) or sink parameters
1927  * (playback). Based on sink and source we need to either find the source
1928  * list or the sink list and set the pipeline parameters
1929  */
skl_tplg_be_update_params(struct snd_soc_dai * dai,struct skl_pipe_params * params)1930 int skl_tplg_be_update_params(struct snd_soc_dai *dai,
1931 				struct skl_pipe_params *params)
1932 {
1933 	struct snd_soc_dapm_widget *w;
1934 
1935 	if (params->stream == SNDRV_PCM_STREAM_PLAYBACK) {
1936 		w = dai->playback_widget;
1937 
1938 		return skl_tplg_be_set_src_pipe_params(dai, w, params);
1939 
1940 	} else {
1941 		w = dai->capture_widget;
1942 
1943 		return skl_tplg_be_set_sink_pipe_params(dai, w, params);
1944 	}
1945 
1946 	return 0;
1947 }
1948 
1949 static const struct snd_soc_tplg_widget_events skl_tplg_widget_ops[] = {
1950 	{SKL_MIXER_EVENT, skl_tplg_mixer_event},
1951 	{SKL_VMIXER_EVENT, skl_tplg_mixer_event},
1952 	{SKL_PGA_EVENT, skl_tplg_pga_event},
1953 };
1954 
1955 static const struct snd_soc_tplg_bytes_ext_ops skl_tlv_ops[] = {
1956 	{SKL_CONTROL_TYPE_BYTE_TLV, skl_tplg_tlv_control_get,
1957 					skl_tplg_tlv_control_set},
1958 };
1959 
1960 static const struct snd_soc_tplg_kcontrol_ops skl_tplg_kcontrol_ops[] = {
1961 	{
1962 		.id = SKL_CONTROL_TYPE_MIC_SELECT,
1963 		.get = skl_tplg_mic_control_get,
1964 		.put = skl_tplg_mic_control_set,
1965 	},
1966 	{
1967 		.id = SKL_CONTROL_TYPE_MULTI_IO_SELECT,
1968 		.get = skl_tplg_multi_config_get,
1969 		.put = skl_tplg_multi_config_set,
1970 	},
1971 	{
1972 		.id = SKL_CONTROL_TYPE_MULTI_IO_SELECT_DMIC,
1973 		.get = skl_tplg_multi_config_get_dmic,
1974 		.put = skl_tplg_multi_config_set_dmic,
1975 	}
1976 };
1977 
skl_tplg_fill_pipe_cfg(struct device * dev,struct skl_pipe * pipe,u32 tkn,u32 tkn_val,int conf_idx,int dir)1978 static int skl_tplg_fill_pipe_cfg(struct device *dev,
1979 			struct skl_pipe *pipe, u32 tkn,
1980 			u32 tkn_val, int conf_idx, int dir)
1981 {
1982 	struct skl_pipe_fmt *fmt;
1983 	struct skl_path_config *config;
1984 
1985 	switch (dir) {
1986 	case SKL_DIR_IN:
1987 		fmt = &pipe->configs[conf_idx].in_fmt;
1988 		break;
1989 
1990 	case SKL_DIR_OUT:
1991 		fmt = &pipe->configs[conf_idx].out_fmt;
1992 		break;
1993 
1994 	default:
1995 		dev_err(dev, "Invalid direction: %d\n", dir);
1996 		return -EINVAL;
1997 	}
1998 
1999 	config = &pipe->configs[conf_idx];
2000 
2001 	switch (tkn) {
2002 	case SKL_TKN_U32_CFG_FREQ:
2003 		fmt->freq = tkn_val;
2004 		break;
2005 
2006 	case SKL_TKN_U8_CFG_CHAN:
2007 		fmt->channels = tkn_val;
2008 		break;
2009 
2010 	case SKL_TKN_U8_CFG_BPS:
2011 		fmt->bps = tkn_val;
2012 		break;
2013 
2014 	case SKL_TKN_U32_PATH_MEM_PGS:
2015 		config->mem_pages = tkn_val;
2016 		break;
2017 
2018 	default:
2019 		dev_err(dev, "Invalid token config: %d\n", tkn);
2020 		return -EINVAL;
2021 	}
2022 
2023 	return 0;
2024 }
2025 
skl_tplg_fill_pipe_tkn(struct device * dev,struct skl_pipe * pipe,u32 tkn,u32 tkn_val)2026 static int skl_tplg_fill_pipe_tkn(struct device *dev,
2027 			struct skl_pipe *pipe, u32 tkn,
2028 			u32 tkn_val)
2029 {
2030 
2031 	switch (tkn) {
2032 	case SKL_TKN_U32_PIPE_CONN_TYPE:
2033 		pipe->conn_type = tkn_val;
2034 		break;
2035 
2036 	case SKL_TKN_U32_PIPE_PRIORITY:
2037 		pipe->pipe_priority = tkn_val;
2038 		break;
2039 
2040 	case SKL_TKN_U32_PIPE_MEM_PGS:
2041 		pipe->memory_pages = tkn_val;
2042 		break;
2043 
2044 	case SKL_TKN_U32_PMODE:
2045 		pipe->lp_mode = tkn_val;
2046 		break;
2047 
2048 	case SKL_TKN_U32_PIPE_DIRECTION:
2049 		pipe->direction = tkn_val;
2050 		break;
2051 
2052 	case SKL_TKN_U32_NUM_CONFIGS:
2053 		pipe->nr_cfgs = tkn_val;
2054 		break;
2055 
2056 	default:
2057 		dev_err(dev, "Token not handled %d\n", tkn);
2058 		return -EINVAL;
2059 	}
2060 
2061 	return 0;
2062 }
2063 
2064 /*
2065  * Add pipeline by parsing the relevant tokens
2066  * Return an existing pipe if the pipe already exists.
2067  */
skl_tplg_add_pipe(struct device * dev,struct skl_module_cfg * mconfig,struct skl_dev * skl,struct snd_soc_tplg_vendor_value_elem * tkn_elem)2068 static int skl_tplg_add_pipe(struct device *dev,
2069 		struct skl_module_cfg *mconfig, struct skl_dev *skl,
2070 		struct snd_soc_tplg_vendor_value_elem *tkn_elem)
2071 {
2072 	struct skl_pipeline *ppl;
2073 	struct skl_pipe *pipe;
2074 	struct skl_pipe_params *params;
2075 
2076 	list_for_each_entry(ppl, &skl->ppl_list, node) {
2077 		if (ppl->pipe->ppl_id == tkn_elem->value) {
2078 			mconfig->pipe = ppl->pipe;
2079 			return -EEXIST;
2080 		}
2081 	}
2082 
2083 	ppl = devm_kzalloc(dev, sizeof(*ppl), GFP_KERNEL);
2084 	if (!ppl)
2085 		return -ENOMEM;
2086 
2087 	pipe = devm_kzalloc(dev, sizeof(*pipe), GFP_KERNEL);
2088 	if (!pipe)
2089 		return -ENOMEM;
2090 
2091 	params = devm_kzalloc(dev, sizeof(*params), GFP_KERNEL);
2092 	if (!params)
2093 		return -ENOMEM;
2094 
2095 	pipe->p_params = params;
2096 	pipe->ppl_id = tkn_elem->value;
2097 	INIT_LIST_HEAD(&pipe->w_list);
2098 
2099 	ppl->pipe = pipe;
2100 	list_add(&ppl->node, &skl->ppl_list);
2101 
2102 	mconfig->pipe = pipe;
2103 	mconfig->pipe->state = SKL_PIPE_INVALID;
2104 
2105 	return 0;
2106 }
2107 
skl_tplg_get_uuid(struct device * dev,guid_t * guid,struct snd_soc_tplg_vendor_uuid_elem * uuid_tkn)2108 static int skl_tplg_get_uuid(struct device *dev, guid_t *guid,
2109 	      struct snd_soc_tplg_vendor_uuid_elem *uuid_tkn)
2110 {
2111 	if (uuid_tkn->token == SKL_TKN_UUID) {
2112 		import_guid(guid, uuid_tkn->uuid);
2113 		return 0;
2114 	}
2115 
2116 	dev_err(dev, "Not an UUID token %d\n", uuid_tkn->token);
2117 
2118 	return -EINVAL;
2119 }
2120 
skl_tplg_fill_pin(struct device * dev,struct snd_soc_tplg_vendor_value_elem * tkn_elem,struct skl_module_pin * m_pin,int pin_index)2121 static int skl_tplg_fill_pin(struct device *dev,
2122 			struct snd_soc_tplg_vendor_value_elem *tkn_elem,
2123 			struct skl_module_pin *m_pin,
2124 			int pin_index)
2125 {
2126 	int ret;
2127 
2128 	switch (tkn_elem->token) {
2129 	case SKL_TKN_U32_PIN_MOD_ID:
2130 		m_pin[pin_index].id.module_id = tkn_elem->value;
2131 		break;
2132 
2133 	case SKL_TKN_U32_PIN_INST_ID:
2134 		m_pin[pin_index].id.instance_id = tkn_elem->value;
2135 		break;
2136 
2137 	case SKL_TKN_UUID:
2138 		ret = skl_tplg_get_uuid(dev, &m_pin[pin_index].id.mod_uuid,
2139 			(struct snd_soc_tplg_vendor_uuid_elem *)tkn_elem);
2140 		if (ret < 0)
2141 			return ret;
2142 
2143 		break;
2144 
2145 	default:
2146 		dev_err(dev, "%d Not a pin token\n", tkn_elem->token);
2147 		return -EINVAL;
2148 	}
2149 
2150 	return 0;
2151 }
2152 
2153 /*
2154  * Parse for pin config specific tokens to fill up the
2155  * module private data
2156  */
skl_tplg_fill_pins_info(struct device * dev,struct skl_module_cfg * mconfig,struct snd_soc_tplg_vendor_value_elem * tkn_elem,int dir,int pin_count)2157 static int skl_tplg_fill_pins_info(struct device *dev,
2158 		struct skl_module_cfg *mconfig,
2159 		struct snd_soc_tplg_vendor_value_elem *tkn_elem,
2160 		int dir, int pin_count)
2161 {
2162 	int ret;
2163 	struct skl_module_pin *m_pin;
2164 
2165 	switch (dir) {
2166 	case SKL_DIR_IN:
2167 		m_pin = mconfig->m_in_pin;
2168 		break;
2169 
2170 	case SKL_DIR_OUT:
2171 		m_pin = mconfig->m_out_pin;
2172 		break;
2173 
2174 	default:
2175 		dev_err(dev, "Invalid direction value\n");
2176 		return -EINVAL;
2177 	}
2178 
2179 	ret = skl_tplg_fill_pin(dev, tkn_elem, m_pin, pin_count);
2180 	if (ret < 0)
2181 		return ret;
2182 
2183 	m_pin[pin_count].in_use = false;
2184 	m_pin[pin_count].pin_state = SKL_PIN_UNBIND;
2185 
2186 	return 0;
2187 }
2188 
2189 /*
2190  * Fill up input/output module config format based
2191  * on the direction
2192  */
skl_tplg_fill_fmt(struct device * dev,struct skl_module_fmt * dst_fmt,u32 tkn,u32 value)2193 static int skl_tplg_fill_fmt(struct device *dev,
2194 		struct skl_module_fmt *dst_fmt,
2195 		u32 tkn, u32 value)
2196 {
2197 	switch (tkn) {
2198 	case SKL_TKN_U32_FMT_CH:
2199 		dst_fmt->channels  = value;
2200 		break;
2201 
2202 	case SKL_TKN_U32_FMT_FREQ:
2203 		dst_fmt->s_freq = value;
2204 		break;
2205 
2206 	case SKL_TKN_U32_FMT_BIT_DEPTH:
2207 		dst_fmt->bit_depth = value;
2208 		break;
2209 
2210 	case SKL_TKN_U32_FMT_SAMPLE_SIZE:
2211 		dst_fmt->valid_bit_depth = value;
2212 		break;
2213 
2214 	case SKL_TKN_U32_FMT_CH_CONFIG:
2215 		dst_fmt->ch_cfg = value;
2216 		break;
2217 
2218 	case SKL_TKN_U32_FMT_INTERLEAVE:
2219 		dst_fmt->interleaving_style = value;
2220 		break;
2221 
2222 	case SKL_TKN_U32_FMT_SAMPLE_TYPE:
2223 		dst_fmt->sample_type = value;
2224 		break;
2225 
2226 	case SKL_TKN_U32_FMT_CH_MAP:
2227 		dst_fmt->ch_map = value;
2228 		break;
2229 
2230 	default:
2231 		dev_err(dev, "Invalid token %d\n", tkn);
2232 		return -EINVAL;
2233 	}
2234 
2235 	return 0;
2236 }
2237 
skl_tplg_widget_fill_fmt(struct device * dev,struct skl_module_iface * fmt,u32 tkn,u32 val,u32 dir,int fmt_idx)2238 static int skl_tplg_widget_fill_fmt(struct device *dev,
2239 		struct skl_module_iface *fmt,
2240 		u32 tkn, u32 val, u32 dir, int fmt_idx)
2241 {
2242 	struct skl_module_fmt *dst_fmt;
2243 
2244 	if (!fmt)
2245 		return -EINVAL;
2246 
2247 	switch (dir) {
2248 	case SKL_DIR_IN:
2249 		dst_fmt = &fmt->inputs[fmt_idx].fmt;
2250 		break;
2251 
2252 	case SKL_DIR_OUT:
2253 		dst_fmt = &fmt->outputs[fmt_idx].fmt;
2254 		break;
2255 
2256 	default:
2257 		dev_err(dev, "Invalid direction: %d\n", dir);
2258 		return -EINVAL;
2259 	}
2260 
2261 	return skl_tplg_fill_fmt(dev, dst_fmt, tkn, val);
2262 }
2263 
skl_tplg_fill_pin_dynamic_val(struct skl_module_pin * mpin,u32 pin_count,u32 value)2264 static void skl_tplg_fill_pin_dynamic_val(
2265 		struct skl_module_pin *mpin, u32 pin_count, u32 value)
2266 {
2267 	int i;
2268 
2269 	for (i = 0; i < pin_count; i++)
2270 		mpin[i].is_dynamic = value;
2271 }
2272 
2273 /*
2274  * Resource table in the manifest has pin specific resources
2275  * like pin and pin buffer size
2276  */
skl_tplg_manifest_pin_res_tkn(struct device * dev,struct snd_soc_tplg_vendor_value_elem * tkn_elem,struct skl_module_res * res,int pin_idx,int dir)2277 static int skl_tplg_manifest_pin_res_tkn(struct device *dev,
2278 		struct snd_soc_tplg_vendor_value_elem *tkn_elem,
2279 		struct skl_module_res *res, int pin_idx, int dir)
2280 {
2281 	struct skl_module_pin_resources *m_pin;
2282 
2283 	switch (dir) {
2284 	case SKL_DIR_IN:
2285 		m_pin = &res->input[pin_idx];
2286 		break;
2287 
2288 	case SKL_DIR_OUT:
2289 		m_pin = &res->output[pin_idx];
2290 		break;
2291 
2292 	default:
2293 		dev_err(dev, "Invalid pin direction: %d\n", dir);
2294 		return -EINVAL;
2295 	}
2296 
2297 	switch (tkn_elem->token) {
2298 	case SKL_TKN_MM_U32_RES_PIN_ID:
2299 		m_pin->pin_index = tkn_elem->value;
2300 		break;
2301 
2302 	case SKL_TKN_MM_U32_PIN_BUF:
2303 		m_pin->buf_size = tkn_elem->value;
2304 		break;
2305 
2306 	default:
2307 		dev_err(dev, "Invalid token: %d\n", tkn_elem->token);
2308 		return -EINVAL;
2309 	}
2310 
2311 	return 0;
2312 }
2313 
2314 /*
2315  * Fill module specific resources from the manifest's resource
2316  * table like CPS, DMA size, mem_pages.
2317  */
skl_tplg_fill_res_tkn(struct device * dev,struct snd_soc_tplg_vendor_value_elem * tkn_elem,struct skl_module_res * res,int pin_idx,int dir)2318 static int skl_tplg_fill_res_tkn(struct device *dev,
2319 		struct snd_soc_tplg_vendor_value_elem *tkn_elem,
2320 		struct skl_module_res *res,
2321 		int pin_idx, int dir)
2322 {
2323 	int ret, tkn_count = 0;
2324 
2325 	if (!res)
2326 		return -EINVAL;
2327 
2328 	switch (tkn_elem->token) {
2329 	case SKL_TKN_MM_U32_DMA_SIZE:
2330 		res->dma_buffer_size = tkn_elem->value;
2331 		break;
2332 
2333 	case SKL_TKN_MM_U32_CPC:
2334 		res->cpc = tkn_elem->value;
2335 		break;
2336 
2337 	case SKL_TKN_U32_MEM_PAGES:
2338 		res->is_pages = tkn_elem->value;
2339 		break;
2340 
2341 	case SKL_TKN_U32_OBS:
2342 		res->obs = tkn_elem->value;
2343 		break;
2344 
2345 	case SKL_TKN_U32_IBS:
2346 		res->ibs = tkn_elem->value;
2347 		break;
2348 
2349 	case SKL_TKN_MM_U32_RES_PIN_ID:
2350 	case SKL_TKN_MM_U32_PIN_BUF:
2351 		ret = skl_tplg_manifest_pin_res_tkn(dev, tkn_elem, res,
2352 						    pin_idx, dir);
2353 		if (ret < 0)
2354 			return ret;
2355 		break;
2356 
2357 	case SKL_TKN_MM_U32_CPS:
2358 	case SKL_TKN_U32_MAX_MCPS:
2359 		/* ignore unused tokens */
2360 		break;
2361 
2362 	default:
2363 		dev_err(dev, "Not a res type token: %d", tkn_elem->token);
2364 		return -EINVAL;
2365 
2366 	}
2367 	tkn_count++;
2368 
2369 	return tkn_count;
2370 }
2371 
2372 /*
2373  * Parse tokens to fill up the module private data
2374  */
skl_tplg_get_token(struct device * dev,struct snd_soc_tplg_vendor_value_elem * tkn_elem,struct skl_dev * skl,struct skl_module_cfg * mconfig)2375 static int skl_tplg_get_token(struct device *dev,
2376 		struct snd_soc_tplg_vendor_value_elem *tkn_elem,
2377 		struct skl_dev *skl, struct skl_module_cfg *mconfig)
2378 {
2379 	int tkn_count = 0;
2380 	int ret;
2381 	static int is_pipe_exists;
2382 	static int pin_index, dir, conf_idx;
2383 	struct skl_module_iface *iface = NULL;
2384 	struct skl_module_res *res = NULL;
2385 	int res_idx = mconfig->res_idx;
2386 	int fmt_idx = mconfig->fmt_idx;
2387 
2388 	/*
2389 	 * If the manifest structure contains no modules, fill all
2390 	 * the module data to 0th index.
2391 	 * res_idx and fmt_idx are default set to 0.
2392 	 */
2393 	if (skl->nr_modules == 0) {
2394 		res = &mconfig->module->resources[res_idx];
2395 		iface = &mconfig->module->formats[fmt_idx];
2396 	}
2397 
2398 	if (tkn_elem->token > SKL_TKN_MAX)
2399 		return -EINVAL;
2400 
2401 	switch (tkn_elem->token) {
2402 	case SKL_TKN_U8_IN_QUEUE_COUNT:
2403 		mconfig->module->max_input_pins = tkn_elem->value;
2404 		break;
2405 
2406 	case SKL_TKN_U8_OUT_QUEUE_COUNT:
2407 		mconfig->module->max_output_pins = tkn_elem->value;
2408 		break;
2409 
2410 	case SKL_TKN_U8_DYN_IN_PIN:
2411 		if (!mconfig->m_in_pin)
2412 			mconfig->m_in_pin =
2413 				devm_kcalloc(dev, MAX_IN_QUEUE,
2414 					     sizeof(*mconfig->m_in_pin),
2415 					     GFP_KERNEL);
2416 		if (!mconfig->m_in_pin)
2417 			return -ENOMEM;
2418 
2419 		skl_tplg_fill_pin_dynamic_val(mconfig->m_in_pin, MAX_IN_QUEUE,
2420 					      tkn_elem->value);
2421 		break;
2422 
2423 	case SKL_TKN_U8_DYN_OUT_PIN:
2424 		if (!mconfig->m_out_pin)
2425 			mconfig->m_out_pin =
2426 				devm_kcalloc(dev, MAX_IN_QUEUE,
2427 					     sizeof(*mconfig->m_in_pin),
2428 					     GFP_KERNEL);
2429 		if (!mconfig->m_out_pin)
2430 			return -ENOMEM;
2431 
2432 		skl_tplg_fill_pin_dynamic_val(mconfig->m_out_pin, MAX_OUT_QUEUE,
2433 					      tkn_elem->value);
2434 		break;
2435 
2436 	case SKL_TKN_U8_TIME_SLOT:
2437 		mconfig->time_slot = tkn_elem->value;
2438 		break;
2439 
2440 	case SKL_TKN_U8_CORE_ID:
2441 		mconfig->core_id = tkn_elem->value;
2442 		break;
2443 
2444 	case SKL_TKN_U8_MOD_TYPE:
2445 		mconfig->m_type = tkn_elem->value;
2446 		break;
2447 
2448 	case SKL_TKN_U8_DEV_TYPE:
2449 		mconfig->dev_type = tkn_elem->value;
2450 		break;
2451 
2452 	case SKL_TKN_U8_HW_CONN_TYPE:
2453 		mconfig->hw_conn_type = tkn_elem->value;
2454 		break;
2455 
2456 	case SKL_TKN_U16_MOD_INST_ID:
2457 		mconfig->id.instance_id =
2458 		tkn_elem->value;
2459 		break;
2460 
2461 	case SKL_TKN_U32_MEM_PAGES:
2462 	case SKL_TKN_U32_MAX_MCPS:
2463 	case SKL_TKN_U32_OBS:
2464 	case SKL_TKN_U32_IBS:
2465 		ret = skl_tplg_fill_res_tkn(dev, tkn_elem, res, pin_index, dir);
2466 		if (ret < 0)
2467 			return ret;
2468 
2469 		break;
2470 
2471 	case SKL_TKN_U32_VBUS_ID:
2472 		mconfig->vbus_id = tkn_elem->value;
2473 		break;
2474 
2475 	case SKL_TKN_U32_PARAMS_FIXUP:
2476 		mconfig->params_fixup = tkn_elem->value;
2477 		break;
2478 
2479 	case SKL_TKN_U32_CONVERTER:
2480 		mconfig->converter = tkn_elem->value;
2481 		break;
2482 
2483 	case SKL_TKN_U32_D0I3_CAPS:
2484 		mconfig->d0i3_caps = tkn_elem->value;
2485 		break;
2486 
2487 	case SKL_TKN_U32_PIPE_ID:
2488 		ret = skl_tplg_add_pipe(dev,
2489 				mconfig, skl, tkn_elem);
2490 
2491 		if (ret < 0) {
2492 			if (ret == -EEXIST) {
2493 				is_pipe_exists = 1;
2494 				break;
2495 			}
2496 			return is_pipe_exists;
2497 		}
2498 
2499 		break;
2500 
2501 	case SKL_TKN_U32_PIPE_CONFIG_ID:
2502 		conf_idx = tkn_elem->value;
2503 		break;
2504 
2505 	case SKL_TKN_U32_PIPE_CONN_TYPE:
2506 	case SKL_TKN_U32_PIPE_PRIORITY:
2507 	case SKL_TKN_U32_PIPE_MEM_PGS:
2508 	case SKL_TKN_U32_PMODE:
2509 	case SKL_TKN_U32_PIPE_DIRECTION:
2510 	case SKL_TKN_U32_NUM_CONFIGS:
2511 		if (is_pipe_exists) {
2512 			ret = skl_tplg_fill_pipe_tkn(dev, mconfig->pipe,
2513 					tkn_elem->token, tkn_elem->value);
2514 			if (ret < 0)
2515 				return ret;
2516 		}
2517 
2518 		break;
2519 
2520 	case SKL_TKN_U32_PATH_MEM_PGS:
2521 	case SKL_TKN_U32_CFG_FREQ:
2522 	case SKL_TKN_U8_CFG_CHAN:
2523 	case SKL_TKN_U8_CFG_BPS:
2524 		if (mconfig->pipe->nr_cfgs) {
2525 			ret = skl_tplg_fill_pipe_cfg(dev, mconfig->pipe,
2526 					tkn_elem->token, tkn_elem->value,
2527 					conf_idx, dir);
2528 			if (ret < 0)
2529 				return ret;
2530 		}
2531 		break;
2532 
2533 	case SKL_TKN_CFG_MOD_RES_ID:
2534 		mconfig->mod_cfg[conf_idx].res_idx = tkn_elem->value;
2535 		break;
2536 
2537 	case SKL_TKN_CFG_MOD_FMT_ID:
2538 		mconfig->mod_cfg[conf_idx].fmt_idx = tkn_elem->value;
2539 		break;
2540 
2541 	/*
2542 	 * SKL_TKN_U32_DIR_PIN_COUNT token has the value for both
2543 	 * direction and the pin count. The first four bits represent
2544 	 * direction and next four the pin count.
2545 	 */
2546 	case SKL_TKN_U32_DIR_PIN_COUNT:
2547 		dir = tkn_elem->value & SKL_IN_DIR_BIT_MASK;
2548 		pin_index = (tkn_elem->value &
2549 			SKL_PIN_COUNT_MASK) >> 4;
2550 
2551 		break;
2552 
2553 	case SKL_TKN_U32_FMT_CH:
2554 	case SKL_TKN_U32_FMT_FREQ:
2555 	case SKL_TKN_U32_FMT_BIT_DEPTH:
2556 	case SKL_TKN_U32_FMT_SAMPLE_SIZE:
2557 	case SKL_TKN_U32_FMT_CH_CONFIG:
2558 	case SKL_TKN_U32_FMT_INTERLEAVE:
2559 	case SKL_TKN_U32_FMT_SAMPLE_TYPE:
2560 	case SKL_TKN_U32_FMT_CH_MAP:
2561 		ret = skl_tplg_widget_fill_fmt(dev, iface, tkn_elem->token,
2562 				tkn_elem->value, dir, pin_index);
2563 
2564 		if (ret < 0)
2565 			return ret;
2566 
2567 		break;
2568 
2569 	case SKL_TKN_U32_PIN_MOD_ID:
2570 	case SKL_TKN_U32_PIN_INST_ID:
2571 	case SKL_TKN_UUID:
2572 		ret = skl_tplg_fill_pins_info(dev,
2573 				mconfig, tkn_elem, dir,
2574 				pin_index);
2575 		if (ret < 0)
2576 			return ret;
2577 
2578 		break;
2579 
2580 	case SKL_TKN_U32_FMT_CFG_IDX:
2581 		if (tkn_elem->value > SKL_MAX_PARAMS_TYPES)
2582 			return -EINVAL;
2583 
2584 		mconfig->fmt_cfg_idx = tkn_elem->value;
2585 		break;
2586 
2587 	case SKL_TKN_U32_CAPS_SIZE:
2588 		mconfig->formats_config[mconfig->fmt_cfg_idx].caps_size =
2589 			tkn_elem->value;
2590 
2591 		break;
2592 
2593 	case SKL_TKN_U32_CAPS_SET_PARAMS:
2594 		mconfig->formats_config[mconfig->fmt_cfg_idx].set_params =
2595 				tkn_elem->value;
2596 		break;
2597 
2598 	case SKL_TKN_U32_CAPS_PARAMS_ID:
2599 		mconfig->formats_config[mconfig->fmt_cfg_idx].param_id =
2600 				tkn_elem->value;
2601 		break;
2602 
2603 	case SKL_TKN_U32_PROC_DOMAIN:
2604 		mconfig->domain =
2605 			tkn_elem->value;
2606 
2607 		break;
2608 
2609 	case SKL_TKN_U32_DMA_BUF_SIZE:
2610 		mconfig->dma_buffer_size = tkn_elem->value;
2611 		break;
2612 
2613 	case SKL_TKN_U8_IN_PIN_TYPE:
2614 	case SKL_TKN_U8_OUT_PIN_TYPE:
2615 	case SKL_TKN_U8_CONN_TYPE:
2616 		break;
2617 
2618 	default:
2619 		dev_err(dev, "Token %d not handled\n",
2620 				tkn_elem->token);
2621 		return -EINVAL;
2622 	}
2623 
2624 	tkn_count++;
2625 
2626 	return tkn_count;
2627 }
2628 
2629 /*
2630  * Parse the vendor array for specific tokens to construct
2631  * module private data
2632  */
skl_tplg_get_tokens(struct device * dev,char * pvt_data,struct skl_dev * skl,struct skl_module_cfg * mconfig,int block_size)2633 static int skl_tplg_get_tokens(struct device *dev,
2634 		char *pvt_data,	struct skl_dev *skl,
2635 		struct skl_module_cfg *mconfig, int block_size)
2636 {
2637 	struct snd_soc_tplg_vendor_array *array;
2638 	struct snd_soc_tplg_vendor_value_elem *tkn_elem;
2639 	int tkn_count = 0, ret;
2640 	int off = 0, tuple_size = 0;
2641 	bool is_module_guid = true;
2642 
2643 	if (block_size <= 0)
2644 		return -EINVAL;
2645 
2646 	while (tuple_size < block_size) {
2647 		array = (struct snd_soc_tplg_vendor_array *)(pvt_data + off);
2648 
2649 		off += array->size;
2650 
2651 		switch (array->type) {
2652 		case SND_SOC_TPLG_TUPLE_TYPE_STRING:
2653 			dev_warn(dev, "no string tokens expected for skl tplg\n");
2654 			continue;
2655 
2656 		case SND_SOC_TPLG_TUPLE_TYPE_UUID:
2657 			if (is_module_guid) {
2658 				ret = skl_tplg_get_uuid(dev, (guid_t *)mconfig->guid,
2659 							array->uuid);
2660 				is_module_guid = false;
2661 			} else {
2662 				ret = skl_tplg_get_token(dev, array->value, skl,
2663 							 mconfig);
2664 			}
2665 
2666 			if (ret < 0)
2667 				return ret;
2668 
2669 			tuple_size += sizeof(*array->uuid);
2670 
2671 			continue;
2672 
2673 		default:
2674 			tkn_elem = array->value;
2675 			tkn_count = 0;
2676 			break;
2677 		}
2678 
2679 		while (tkn_count <= (array->num_elems - 1)) {
2680 			ret = skl_tplg_get_token(dev, tkn_elem,
2681 					skl, mconfig);
2682 
2683 			if (ret < 0)
2684 				return ret;
2685 
2686 			tkn_count = tkn_count + ret;
2687 			tkn_elem++;
2688 		}
2689 
2690 		tuple_size += tkn_count * sizeof(*tkn_elem);
2691 	}
2692 
2693 	return off;
2694 }
2695 
2696 /*
2697  * Every data block is preceded by a descriptor to read the number
2698  * of data blocks, they type of the block and it's size
2699  */
skl_tplg_get_desc_blocks(struct device * dev,struct snd_soc_tplg_vendor_array * array)2700 static int skl_tplg_get_desc_blocks(struct device *dev,
2701 		struct snd_soc_tplg_vendor_array *array)
2702 {
2703 	struct snd_soc_tplg_vendor_value_elem *tkn_elem;
2704 
2705 	tkn_elem = array->value;
2706 
2707 	switch (tkn_elem->token) {
2708 	case SKL_TKN_U8_NUM_BLOCKS:
2709 	case SKL_TKN_U8_BLOCK_TYPE:
2710 	case SKL_TKN_U16_BLOCK_SIZE:
2711 		return tkn_elem->value;
2712 
2713 	default:
2714 		dev_err(dev, "Invalid descriptor token %d\n", tkn_elem->token);
2715 		break;
2716 	}
2717 
2718 	return -EINVAL;
2719 }
2720 
2721 /* Functions to parse private data from configuration file format v4 */
2722 
2723 /*
2724  * Add pipeline from topology binary into driver pipeline list
2725  *
2726  * If already added we return that instance
2727  * Otherwise we create a new instance and add into driver list
2728  */
skl_tplg_add_pipe_v4(struct device * dev,struct skl_module_cfg * mconfig,struct skl_dev * skl,struct skl_dfw_v4_pipe * dfw_pipe)2729 static int skl_tplg_add_pipe_v4(struct device *dev,
2730 			struct skl_module_cfg *mconfig, struct skl_dev *skl,
2731 			struct skl_dfw_v4_pipe *dfw_pipe)
2732 {
2733 	struct skl_pipeline *ppl;
2734 	struct skl_pipe *pipe;
2735 	struct skl_pipe_params *params;
2736 
2737 	list_for_each_entry(ppl, &skl->ppl_list, node) {
2738 		if (ppl->pipe->ppl_id == dfw_pipe->pipe_id) {
2739 			mconfig->pipe = ppl->pipe;
2740 			return 0;
2741 		}
2742 	}
2743 
2744 	ppl = devm_kzalloc(dev, sizeof(*ppl), GFP_KERNEL);
2745 	if (!ppl)
2746 		return -ENOMEM;
2747 
2748 	pipe = devm_kzalloc(dev, sizeof(*pipe), GFP_KERNEL);
2749 	if (!pipe)
2750 		return -ENOMEM;
2751 
2752 	params = devm_kzalloc(dev, sizeof(*params), GFP_KERNEL);
2753 	if (!params)
2754 		return -ENOMEM;
2755 
2756 	pipe->ppl_id = dfw_pipe->pipe_id;
2757 	pipe->memory_pages = dfw_pipe->memory_pages;
2758 	pipe->pipe_priority = dfw_pipe->pipe_priority;
2759 	pipe->conn_type = dfw_pipe->conn_type;
2760 	pipe->state = SKL_PIPE_INVALID;
2761 	pipe->p_params = params;
2762 	INIT_LIST_HEAD(&pipe->w_list);
2763 
2764 	ppl->pipe = pipe;
2765 	list_add(&ppl->node, &skl->ppl_list);
2766 
2767 	mconfig->pipe = pipe;
2768 
2769 	return 0;
2770 }
2771 
skl_fill_module_pin_info_v4(struct skl_dfw_v4_module_pin * dfw_pin,struct skl_module_pin * m_pin,bool is_dynamic,int max_pin)2772 static void skl_fill_module_pin_info_v4(struct skl_dfw_v4_module_pin *dfw_pin,
2773 					struct skl_module_pin *m_pin,
2774 					bool is_dynamic, int max_pin)
2775 {
2776 	int i;
2777 
2778 	for (i = 0; i < max_pin; i++) {
2779 		m_pin[i].id.module_id = dfw_pin[i].module_id;
2780 		m_pin[i].id.instance_id = dfw_pin[i].instance_id;
2781 		m_pin[i].in_use = false;
2782 		m_pin[i].is_dynamic = is_dynamic;
2783 		m_pin[i].pin_state = SKL_PIN_UNBIND;
2784 	}
2785 }
2786 
skl_tplg_fill_fmt_v4(struct skl_module_pin_fmt * dst_fmt,struct skl_dfw_v4_module_fmt * src_fmt,int pins)2787 static void skl_tplg_fill_fmt_v4(struct skl_module_pin_fmt *dst_fmt,
2788 				 struct skl_dfw_v4_module_fmt *src_fmt,
2789 				 int pins)
2790 {
2791 	int i;
2792 
2793 	for (i = 0; i < pins; i++) {
2794 		dst_fmt[i].fmt.channels  = src_fmt[i].channels;
2795 		dst_fmt[i].fmt.s_freq = src_fmt[i].freq;
2796 		dst_fmt[i].fmt.bit_depth = src_fmt[i].bit_depth;
2797 		dst_fmt[i].fmt.valid_bit_depth = src_fmt[i].valid_bit_depth;
2798 		dst_fmt[i].fmt.ch_cfg = src_fmt[i].ch_cfg;
2799 		dst_fmt[i].fmt.ch_map = src_fmt[i].ch_map;
2800 		dst_fmt[i].fmt.interleaving_style =
2801 						src_fmt[i].interleaving_style;
2802 		dst_fmt[i].fmt.sample_type = src_fmt[i].sample_type;
2803 	}
2804 }
2805 
skl_tplg_get_pvt_data_v4(struct snd_soc_tplg_dapm_widget * tplg_w,struct skl_dev * skl,struct device * dev,struct skl_module_cfg * mconfig)2806 static int skl_tplg_get_pvt_data_v4(struct snd_soc_tplg_dapm_widget *tplg_w,
2807 				    struct skl_dev *skl, struct device *dev,
2808 				    struct skl_module_cfg *mconfig)
2809 {
2810 	struct skl_dfw_v4_module *dfw =
2811 				(struct skl_dfw_v4_module *)tplg_w->priv.data;
2812 	int ret;
2813 	int idx = mconfig->fmt_cfg_idx;
2814 
2815 	dev_dbg(dev, "Parsing Skylake v4 widget topology data\n");
2816 
2817 	ret = guid_parse(dfw->uuid, (guid_t *)mconfig->guid);
2818 	if (ret)
2819 		return ret;
2820 	mconfig->id.module_id = -1;
2821 	mconfig->id.instance_id = dfw->instance_id;
2822 	mconfig->module->resources[0].cpc = dfw->max_mcps / 1000;
2823 	mconfig->module->resources[0].ibs = dfw->ibs;
2824 	mconfig->module->resources[0].obs = dfw->obs;
2825 	mconfig->core_id = dfw->core_id;
2826 	mconfig->module->max_input_pins = dfw->max_in_queue;
2827 	mconfig->module->max_output_pins = dfw->max_out_queue;
2828 	mconfig->module->loadable = dfw->is_loadable;
2829 	skl_tplg_fill_fmt_v4(mconfig->module->formats[0].inputs, dfw->in_fmt,
2830 			     MAX_IN_QUEUE);
2831 	skl_tplg_fill_fmt_v4(mconfig->module->formats[0].outputs, dfw->out_fmt,
2832 			     MAX_OUT_QUEUE);
2833 
2834 	mconfig->params_fixup = dfw->params_fixup;
2835 	mconfig->converter = dfw->converter;
2836 	mconfig->m_type = dfw->module_type;
2837 	mconfig->vbus_id = dfw->vbus_id;
2838 	mconfig->module->resources[0].is_pages = dfw->mem_pages;
2839 
2840 	ret = skl_tplg_add_pipe_v4(dev, mconfig, skl, &dfw->pipe);
2841 	if (ret)
2842 		return ret;
2843 
2844 	mconfig->dev_type = dfw->dev_type;
2845 	mconfig->hw_conn_type = dfw->hw_conn_type;
2846 	mconfig->time_slot = dfw->time_slot;
2847 	mconfig->formats_config[idx].caps_size = dfw->caps.caps_size;
2848 
2849 	mconfig->m_in_pin = devm_kcalloc(dev,
2850 				MAX_IN_QUEUE, sizeof(*mconfig->m_in_pin),
2851 				GFP_KERNEL);
2852 	if (!mconfig->m_in_pin)
2853 		return -ENOMEM;
2854 
2855 	mconfig->m_out_pin = devm_kcalloc(dev,
2856 				MAX_OUT_QUEUE, sizeof(*mconfig->m_out_pin),
2857 				GFP_KERNEL);
2858 	if (!mconfig->m_out_pin)
2859 		return -ENOMEM;
2860 
2861 	skl_fill_module_pin_info_v4(dfw->in_pin, mconfig->m_in_pin,
2862 				    dfw->is_dynamic_in_pin,
2863 				    mconfig->module->max_input_pins);
2864 	skl_fill_module_pin_info_v4(dfw->out_pin, mconfig->m_out_pin,
2865 				    dfw->is_dynamic_out_pin,
2866 				    mconfig->module->max_output_pins);
2867 
2868 	if (mconfig->formats_config[idx].caps_size) {
2869 		mconfig->formats_config[idx].set_params = dfw->caps.set_params;
2870 		mconfig->formats_config[idx].param_id = dfw->caps.param_id;
2871 		mconfig->formats_config[idx].caps =
2872 		devm_kzalloc(dev, mconfig->formats_config[idx].caps_size,
2873 			     GFP_KERNEL);
2874 		if (!mconfig->formats_config[idx].caps)
2875 			return -ENOMEM;
2876 		memcpy(mconfig->formats_config[idx].caps, dfw->caps.caps,
2877 		       dfw->caps.caps_size);
2878 	}
2879 
2880 	return 0;
2881 }
2882 
skl_tplg_get_caps_data(struct device * dev,char * data,struct skl_module_cfg * mconfig)2883 static int skl_tplg_get_caps_data(struct device *dev, char *data,
2884 				  struct skl_module_cfg *mconfig)
2885 {
2886 	int idx = mconfig->fmt_cfg_idx;
2887 
2888 	if (mconfig->formats_config[idx].caps_size > 0) {
2889 		mconfig->formats_config[idx].caps =
2890 			devm_kzalloc(dev, mconfig->formats_config[idx].caps_size,
2891 				     GFP_KERNEL);
2892 		if (!mconfig->formats_config[idx].caps)
2893 			return -ENOMEM;
2894 		memcpy(mconfig->formats_config[idx].caps, data,
2895 		       mconfig->formats_config[idx].caps_size);
2896 	}
2897 
2898 	return mconfig->formats_config[idx].caps_size;
2899 }
2900 
2901 /*
2902  * Parse the private data for the token and corresponding value.
2903  * The private data can have multiple data blocks. So, a data block
2904  * is preceded by a descriptor for number of blocks and a descriptor
2905  * for the type and size of the suceeding data block.
2906  */
skl_tplg_get_pvt_data(struct snd_soc_tplg_dapm_widget * tplg_w,struct skl_dev * skl,struct device * dev,struct skl_module_cfg * mconfig)2907 static int skl_tplg_get_pvt_data(struct snd_soc_tplg_dapm_widget *tplg_w,
2908 				struct skl_dev *skl, struct device *dev,
2909 				struct skl_module_cfg *mconfig)
2910 {
2911 	struct snd_soc_tplg_vendor_array *array;
2912 	int num_blocks, block_size, block_type, off = 0;
2913 	char *data;
2914 	int ret;
2915 
2916 	/*
2917 	 * v4 configuration files have a valid UUID at the start of
2918 	 * the widget's private data.
2919 	 */
2920 	if (uuid_is_valid((char *)tplg_w->priv.data))
2921 		return skl_tplg_get_pvt_data_v4(tplg_w, skl, dev, mconfig);
2922 
2923 	/* Read the NUM_DATA_BLOCKS descriptor */
2924 	array = (struct snd_soc_tplg_vendor_array *)tplg_w->priv.data;
2925 	ret = skl_tplg_get_desc_blocks(dev, array);
2926 	if (ret < 0)
2927 		return ret;
2928 	num_blocks = ret;
2929 
2930 	off += array->size;
2931 	/* Read the BLOCK_TYPE and BLOCK_SIZE descriptor */
2932 	while (num_blocks > 0) {
2933 		array = (struct snd_soc_tplg_vendor_array *)
2934 				(tplg_w->priv.data + off);
2935 
2936 		ret = skl_tplg_get_desc_blocks(dev, array);
2937 
2938 		if (ret < 0)
2939 			return ret;
2940 		block_type = ret;
2941 		off += array->size;
2942 
2943 		array = (struct snd_soc_tplg_vendor_array *)
2944 			(tplg_w->priv.data + off);
2945 
2946 		ret = skl_tplg_get_desc_blocks(dev, array);
2947 
2948 		if (ret < 0)
2949 			return ret;
2950 		block_size = ret;
2951 		off += array->size;
2952 
2953 		array = (struct snd_soc_tplg_vendor_array *)
2954 			(tplg_w->priv.data + off);
2955 
2956 		data = (tplg_w->priv.data + off);
2957 
2958 		if (block_type == SKL_TYPE_TUPLE) {
2959 			ret = skl_tplg_get_tokens(dev, data,
2960 					skl, mconfig, block_size);
2961 		} else {
2962 			ret = skl_tplg_get_caps_data(dev, data, mconfig);
2963 		}
2964 
2965 		if (ret < 0)
2966 			return ret;
2967 
2968 		--num_blocks;
2969 		off += ret;
2970 	}
2971 
2972 	return 0;
2973 }
2974 
skl_clear_pin_config(struct snd_soc_component * component,struct snd_soc_dapm_widget * w)2975 static void skl_clear_pin_config(struct snd_soc_component *component,
2976 				struct snd_soc_dapm_widget *w)
2977 {
2978 	int i;
2979 	struct skl_module_cfg *mconfig;
2980 	struct skl_pipe *pipe;
2981 
2982 	if (!strncmp(w->dapm->component->name, component->name,
2983 					strlen(component->name))) {
2984 		mconfig = w->priv;
2985 		pipe = mconfig->pipe;
2986 		for (i = 0; i < mconfig->module->max_input_pins; i++) {
2987 			mconfig->m_in_pin[i].in_use = false;
2988 			mconfig->m_in_pin[i].pin_state = SKL_PIN_UNBIND;
2989 		}
2990 		for (i = 0; i < mconfig->module->max_output_pins; i++) {
2991 			mconfig->m_out_pin[i].in_use = false;
2992 			mconfig->m_out_pin[i].pin_state = SKL_PIN_UNBIND;
2993 		}
2994 		pipe->state = SKL_PIPE_INVALID;
2995 		mconfig->m_state = SKL_MODULE_UNINIT;
2996 	}
2997 }
2998 
skl_cleanup_resources(struct skl_dev * skl)2999 void skl_cleanup_resources(struct skl_dev *skl)
3000 {
3001 	struct snd_soc_component *soc_component = skl->component;
3002 	struct snd_soc_dapm_widget *w;
3003 	struct snd_soc_card *card;
3004 
3005 	if (soc_component == NULL)
3006 		return;
3007 
3008 	card = soc_component->card;
3009 	if (!card || !card->instantiated)
3010 		return;
3011 
3012 	list_for_each_entry(w, &card->widgets, list) {
3013 		if (is_skl_dsp_widget_type(w, skl->dev) && w->priv != NULL)
3014 			skl_clear_pin_config(soc_component, w);
3015 	}
3016 
3017 	skl_clear_module_cnt(skl->dsp);
3018 }
3019 
3020 /*
3021  * Topology core widget load callback
3022  *
3023  * This is used to save the private data for each widget which gives
3024  * information to the driver about module and pipeline parameters which DSP
3025  * FW expects like ids, resource values, formats etc
3026  */
skl_tplg_widget_load(struct snd_soc_component * cmpnt,int index,struct snd_soc_dapm_widget * w,struct snd_soc_tplg_dapm_widget * tplg_w)3027 static int skl_tplg_widget_load(struct snd_soc_component *cmpnt, int index,
3028 				struct snd_soc_dapm_widget *w,
3029 				struct snd_soc_tplg_dapm_widget *tplg_w)
3030 {
3031 	int ret;
3032 	struct hdac_bus *bus = snd_soc_component_get_drvdata(cmpnt);
3033 	struct skl_dev *skl = bus_to_skl(bus);
3034 	struct skl_module_cfg *mconfig;
3035 
3036 	if (!tplg_w->priv.size)
3037 		goto bind_event;
3038 
3039 	mconfig = devm_kzalloc(bus->dev, sizeof(*mconfig), GFP_KERNEL);
3040 
3041 	if (!mconfig)
3042 		return -ENOMEM;
3043 
3044 	if (skl->nr_modules == 0) {
3045 		mconfig->module = devm_kzalloc(bus->dev,
3046 				sizeof(*mconfig->module), GFP_KERNEL);
3047 		if (!mconfig->module)
3048 			return -ENOMEM;
3049 	}
3050 
3051 	w->priv = mconfig;
3052 
3053 	/*
3054 	 * module binary can be loaded later, so set it to query when
3055 	 * module is load for a use case
3056 	 */
3057 	mconfig->id.module_id = -1;
3058 
3059 	/* To provide backward compatibility, set default as SKL_PARAM_INIT */
3060 	mconfig->fmt_cfg_idx = SKL_PARAM_INIT;
3061 
3062 	/* Parse private data for tuples */
3063 	ret = skl_tplg_get_pvt_data(tplg_w, skl, bus->dev, mconfig);
3064 	if (ret < 0)
3065 		return ret;
3066 
3067 	skl_debug_init_module(skl->debugfs, w, mconfig);
3068 
3069 bind_event:
3070 	if (tplg_w->event_type == 0) {
3071 		dev_dbg(bus->dev, "ASoC: No event handler required\n");
3072 		return 0;
3073 	}
3074 
3075 	ret = snd_soc_tplg_widget_bind_event(w, skl_tplg_widget_ops,
3076 					ARRAY_SIZE(skl_tplg_widget_ops),
3077 					tplg_w->event_type);
3078 
3079 	if (ret) {
3080 		dev_err(bus->dev, "%s: No matching event handlers found for %d\n",
3081 					__func__, tplg_w->event_type);
3082 		return -EINVAL;
3083 	}
3084 
3085 	return 0;
3086 }
3087 
skl_init_algo_data(struct device * dev,struct soc_bytes_ext * be,struct snd_soc_tplg_bytes_control * bc)3088 static int skl_init_algo_data(struct device *dev, struct soc_bytes_ext *be,
3089 					struct snd_soc_tplg_bytes_control *bc)
3090 {
3091 	struct skl_algo_data *ac;
3092 	struct skl_dfw_algo_data *dfw_ac =
3093 				(struct skl_dfw_algo_data *)bc->priv.data;
3094 
3095 	ac = devm_kzalloc(dev, sizeof(*ac), GFP_KERNEL);
3096 	if (!ac)
3097 		return -ENOMEM;
3098 
3099 	/* Fill private data */
3100 	ac->max = dfw_ac->max;
3101 	ac->param_id = dfw_ac->param_id;
3102 	ac->set_params = dfw_ac->set_params;
3103 	ac->size = dfw_ac->max;
3104 
3105 	if (ac->max) {
3106 		ac->params = devm_kzalloc(dev, ac->max, GFP_KERNEL);
3107 		if (!ac->params)
3108 			return -ENOMEM;
3109 
3110 		memcpy(ac->params, dfw_ac->params, ac->max);
3111 	}
3112 
3113 	be->dobj.private  = ac;
3114 	return 0;
3115 }
3116 
skl_init_enum_data(struct device * dev,struct soc_enum * se,struct snd_soc_tplg_enum_control * ec)3117 static int skl_init_enum_data(struct device *dev, struct soc_enum *se,
3118 				struct snd_soc_tplg_enum_control *ec)
3119 {
3120 
3121 	void *data;
3122 
3123 	if (ec->priv.size) {
3124 		data = devm_kzalloc(dev, sizeof(ec->priv.size), GFP_KERNEL);
3125 		if (!data)
3126 			return -ENOMEM;
3127 		memcpy(data, ec->priv.data, ec->priv.size);
3128 		se->dobj.private = data;
3129 	}
3130 
3131 	return 0;
3132 
3133 }
3134 
skl_tplg_control_load(struct snd_soc_component * cmpnt,int index,struct snd_kcontrol_new * kctl,struct snd_soc_tplg_ctl_hdr * hdr)3135 static int skl_tplg_control_load(struct snd_soc_component *cmpnt,
3136 				int index,
3137 				struct snd_kcontrol_new *kctl,
3138 				struct snd_soc_tplg_ctl_hdr *hdr)
3139 {
3140 	struct soc_bytes_ext *sb;
3141 	struct snd_soc_tplg_bytes_control *tplg_bc;
3142 	struct snd_soc_tplg_enum_control *tplg_ec;
3143 	struct hdac_bus *bus  = snd_soc_component_get_drvdata(cmpnt);
3144 	struct soc_enum *se;
3145 
3146 	switch (hdr->ops.info) {
3147 	case SND_SOC_TPLG_CTL_BYTES:
3148 		tplg_bc = container_of(hdr,
3149 				struct snd_soc_tplg_bytes_control, hdr);
3150 		if (kctl->access & SNDRV_CTL_ELEM_ACCESS_TLV_CALLBACK) {
3151 			sb = (struct soc_bytes_ext *)kctl->private_value;
3152 			if (tplg_bc->priv.size)
3153 				return skl_init_algo_data(
3154 						bus->dev, sb, tplg_bc);
3155 		}
3156 		break;
3157 
3158 	case SND_SOC_TPLG_CTL_ENUM:
3159 		tplg_ec = container_of(hdr,
3160 				struct snd_soc_tplg_enum_control, hdr);
3161 		if (kctl->access & SNDRV_CTL_ELEM_ACCESS_READ) {
3162 			se = (struct soc_enum *)kctl->private_value;
3163 			if (tplg_ec->priv.size)
3164 				skl_init_enum_data(bus->dev, se, tplg_ec);
3165 		}
3166 
3167 		/*
3168 		 * now that the control initializations are done, remove
3169 		 * write permission for the DMIC configuration enums to
3170 		 * avoid conflicts between NHLT settings and user interaction
3171 		 */
3172 
3173 		if (hdr->ops.get == SKL_CONTROL_TYPE_MULTI_IO_SELECT_DMIC)
3174 			kctl->access = SNDRV_CTL_ELEM_ACCESS_READ;
3175 
3176 		break;
3177 
3178 	default:
3179 		dev_dbg(bus->dev, "Control load not supported %d:%d:%d\n",
3180 			hdr->ops.get, hdr->ops.put, hdr->ops.info);
3181 		break;
3182 	}
3183 
3184 	return 0;
3185 }
3186 
skl_tplg_fill_str_mfest_tkn(struct device * dev,struct snd_soc_tplg_vendor_string_elem * str_elem,struct skl_dev * skl)3187 static int skl_tplg_fill_str_mfest_tkn(struct device *dev,
3188 		struct snd_soc_tplg_vendor_string_elem *str_elem,
3189 		struct skl_dev *skl)
3190 {
3191 	int tkn_count = 0;
3192 	static int ref_count;
3193 
3194 	switch (str_elem->token) {
3195 	case SKL_TKN_STR_LIB_NAME:
3196 		if (ref_count > skl->lib_count - 1) {
3197 			ref_count = 0;
3198 			return -EINVAL;
3199 		}
3200 
3201 		strncpy(skl->lib_info[ref_count].name,
3202 			str_elem->string,
3203 			ARRAY_SIZE(skl->lib_info[ref_count].name));
3204 		ref_count++;
3205 		break;
3206 
3207 	default:
3208 		dev_err(dev, "Not a string token %d\n", str_elem->token);
3209 		break;
3210 	}
3211 	tkn_count++;
3212 
3213 	return tkn_count;
3214 }
3215 
skl_tplg_get_str_tkn(struct device * dev,struct snd_soc_tplg_vendor_array * array,struct skl_dev * skl)3216 static int skl_tplg_get_str_tkn(struct device *dev,
3217 		struct snd_soc_tplg_vendor_array *array,
3218 		struct skl_dev *skl)
3219 {
3220 	int tkn_count = 0, ret;
3221 	struct snd_soc_tplg_vendor_string_elem *str_elem;
3222 
3223 	str_elem = (struct snd_soc_tplg_vendor_string_elem *)array->value;
3224 	while (tkn_count < array->num_elems) {
3225 		ret = skl_tplg_fill_str_mfest_tkn(dev, str_elem, skl);
3226 		str_elem++;
3227 
3228 		if (ret < 0)
3229 			return ret;
3230 
3231 		tkn_count = tkn_count + ret;
3232 	}
3233 
3234 	return tkn_count;
3235 }
3236 
skl_tplg_manifest_fill_fmt(struct device * dev,struct skl_module_iface * fmt,struct snd_soc_tplg_vendor_value_elem * tkn_elem,u32 dir,int fmt_idx)3237 static int skl_tplg_manifest_fill_fmt(struct device *dev,
3238 		struct skl_module_iface *fmt,
3239 		struct snd_soc_tplg_vendor_value_elem *tkn_elem,
3240 		u32 dir, int fmt_idx)
3241 {
3242 	struct skl_module_pin_fmt *dst_fmt;
3243 	struct skl_module_fmt *mod_fmt;
3244 	int ret;
3245 
3246 	if (!fmt)
3247 		return -EINVAL;
3248 
3249 	switch (dir) {
3250 	case SKL_DIR_IN:
3251 		dst_fmt = &fmt->inputs[fmt_idx];
3252 		break;
3253 
3254 	case SKL_DIR_OUT:
3255 		dst_fmt = &fmt->outputs[fmt_idx];
3256 		break;
3257 
3258 	default:
3259 		dev_err(dev, "Invalid direction: %d\n", dir);
3260 		return -EINVAL;
3261 	}
3262 
3263 	mod_fmt = &dst_fmt->fmt;
3264 
3265 	switch (tkn_elem->token) {
3266 	case SKL_TKN_MM_U32_INTF_PIN_ID:
3267 		dst_fmt->id = tkn_elem->value;
3268 		break;
3269 
3270 	default:
3271 		ret = skl_tplg_fill_fmt(dev, mod_fmt, tkn_elem->token,
3272 					tkn_elem->value);
3273 		if (ret < 0)
3274 			return ret;
3275 		break;
3276 	}
3277 
3278 	return 0;
3279 }
3280 
skl_tplg_fill_mod_info(struct device * dev,struct snd_soc_tplg_vendor_value_elem * tkn_elem,struct skl_module * mod)3281 static int skl_tplg_fill_mod_info(struct device *dev,
3282 		struct snd_soc_tplg_vendor_value_elem *tkn_elem,
3283 		struct skl_module *mod)
3284 {
3285 
3286 	if (!mod)
3287 		return -EINVAL;
3288 
3289 	switch (tkn_elem->token) {
3290 	case SKL_TKN_U8_IN_PIN_TYPE:
3291 		mod->input_pin_type = tkn_elem->value;
3292 		break;
3293 
3294 	case SKL_TKN_U8_OUT_PIN_TYPE:
3295 		mod->output_pin_type = tkn_elem->value;
3296 		break;
3297 
3298 	case SKL_TKN_U8_IN_QUEUE_COUNT:
3299 		mod->max_input_pins = tkn_elem->value;
3300 		break;
3301 
3302 	case SKL_TKN_U8_OUT_QUEUE_COUNT:
3303 		mod->max_output_pins = tkn_elem->value;
3304 		break;
3305 
3306 	case SKL_TKN_MM_U8_NUM_RES:
3307 		mod->nr_resources = tkn_elem->value;
3308 		break;
3309 
3310 	case SKL_TKN_MM_U8_NUM_INTF:
3311 		mod->nr_interfaces = tkn_elem->value;
3312 		break;
3313 
3314 	default:
3315 		dev_err(dev, "Invalid mod info token %d", tkn_elem->token);
3316 		return -EINVAL;
3317 	}
3318 
3319 	return 0;
3320 }
3321 
3322 
skl_tplg_get_int_tkn(struct device * dev,struct snd_soc_tplg_vendor_value_elem * tkn_elem,struct skl_dev * skl)3323 static int skl_tplg_get_int_tkn(struct device *dev,
3324 		struct snd_soc_tplg_vendor_value_elem *tkn_elem,
3325 		struct skl_dev *skl)
3326 {
3327 	int tkn_count = 0, ret;
3328 	static int mod_idx, res_val_idx, intf_val_idx, dir, pin_idx;
3329 	struct skl_module_res *res = NULL;
3330 	struct skl_module_iface *fmt = NULL;
3331 	struct skl_module *mod = NULL;
3332 	static struct skl_astate_param *astate_table;
3333 	static int astate_cfg_idx, count;
3334 	int i;
3335 	size_t size;
3336 
3337 	if (skl->modules) {
3338 		mod = skl->modules[mod_idx];
3339 		res = &mod->resources[res_val_idx];
3340 		fmt = &mod->formats[intf_val_idx];
3341 	}
3342 
3343 	switch (tkn_elem->token) {
3344 	case SKL_TKN_U32_LIB_COUNT:
3345 		skl->lib_count = tkn_elem->value;
3346 		break;
3347 
3348 	case SKL_TKN_U8_NUM_MOD:
3349 		skl->nr_modules = tkn_elem->value;
3350 		skl->modules = devm_kcalloc(dev, skl->nr_modules,
3351 				sizeof(*skl->modules), GFP_KERNEL);
3352 		if (!skl->modules)
3353 			return -ENOMEM;
3354 
3355 		for (i = 0; i < skl->nr_modules; i++) {
3356 			skl->modules[i] = devm_kzalloc(dev,
3357 					sizeof(struct skl_module), GFP_KERNEL);
3358 			if (!skl->modules[i])
3359 				return -ENOMEM;
3360 		}
3361 		break;
3362 
3363 	case SKL_TKN_MM_U8_MOD_IDX:
3364 		mod_idx = tkn_elem->value;
3365 		break;
3366 
3367 	case SKL_TKN_U32_ASTATE_COUNT:
3368 		if (astate_table != NULL) {
3369 			dev_err(dev, "More than one entry for A-State count");
3370 			return -EINVAL;
3371 		}
3372 
3373 		if (tkn_elem->value > SKL_MAX_ASTATE_CFG) {
3374 			dev_err(dev, "Invalid A-State count %d\n",
3375 				tkn_elem->value);
3376 			return -EINVAL;
3377 		}
3378 
3379 		size = struct_size(skl->cfg.astate_cfg, astate_table,
3380 				   tkn_elem->value);
3381 		skl->cfg.astate_cfg = devm_kzalloc(dev, size, GFP_KERNEL);
3382 		if (!skl->cfg.astate_cfg)
3383 			return -ENOMEM;
3384 
3385 		astate_table = skl->cfg.astate_cfg->astate_table;
3386 		count = skl->cfg.astate_cfg->count = tkn_elem->value;
3387 		break;
3388 
3389 	case SKL_TKN_U32_ASTATE_IDX:
3390 		if (tkn_elem->value >= count) {
3391 			dev_err(dev, "Invalid A-State index %d\n",
3392 				tkn_elem->value);
3393 			return -EINVAL;
3394 		}
3395 
3396 		astate_cfg_idx = tkn_elem->value;
3397 		break;
3398 
3399 	case SKL_TKN_U32_ASTATE_KCPS:
3400 		astate_table[astate_cfg_idx].kcps = tkn_elem->value;
3401 		break;
3402 
3403 	case SKL_TKN_U32_ASTATE_CLK_SRC:
3404 		astate_table[astate_cfg_idx].clk_src = tkn_elem->value;
3405 		break;
3406 
3407 	case SKL_TKN_U8_IN_PIN_TYPE:
3408 	case SKL_TKN_U8_OUT_PIN_TYPE:
3409 	case SKL_TKN_U8_IN_QUEUE_COUNT:
3410 	case SKL_TKN_U8_OUT_QUEUE_COUNT:
3411 	case SKL_TKN_MM_U8_NUM_RES:
3412 	case SKL_TKN_MM_U8_NUM_INTF:
3413 		ret = skl_tplg_fill_mod_info(dev, tkn_elem, mod);
3414 		if (ret < 0)
3415 			return ret;
3416 		break;
3417 
3418 	case SKL_TKN_U32_DIR_PIN_COUNT:
3419 		dir = tkn_elem->value & SKL_IN_DIR_BIT_MASK;
3420 		pin_idx = (tkn_elem->value & SKL_PIN_COUNT_MASK) >> 4;
3421 		break;
3422 
3423 	case SKL_TKN_MM_U32_RES_ID:
3424 		if (!res)
3425 			return -EINVAL;
3426 
3427 		res->id = tkn_elem->value;
3428 		res_val_idx = tkn_elem->value;
3429 		break;
3430 
3431 	case SKL_TKN_MM_U32_FMT_ID:
3432 		if (!fmt)
3433 			return -EINVAL;
3434 
3435 		fmt->fmt_idx = tkn_elem->value;
3436 		intf_val_idx = tkn_elem->value;
3437 		break;
3438 
3439 	case SKL_TKN_MM_U32_CPS:
3440 	case SKL_TKN_MM_U32_DMA_SIZE:
3441 	case SKL_TKN_MM_U32_CPC:
3442 	case SKL_TKN_U32_MEM_PAGES:
3443 	case SKL_TKN_U32_OBS:
3444 	case SKL_TKN_U32_IBS:
3445 	case SKL_TKN_MM_U32_RES_PIN_ID:
3446 	case SKL_TKN_MM_U32_PIN_BUF:
3447 		ret = skl_tplg_fill_res_tkn(dev, tkn_elem, res, pin_idx, dir);
3448 		if (ret < 0)
3449 			return ret;
3450 
3451 		break;
3452 
3453 	case SKL_TKN_MM_U32_NUM_IN_FMT:
3454 		if (!fmt)
3455 			return -EINVAL;
3456 
3457 		res->nr_input_pins = tkn_elem->value;
3458 		break;
3459 
3460 	case SKL_TKN_MM_U32_NUM_OUT_FMT:
3461 		if (!fmt)
3462 			return -EINVAL;
3463 
3464 		res->nr_output_pins = tkn_elem->value;
3465 		break;
3466 
3467 	case SKL_TKN_U32_FMT_CH:
3468 	case SKL_TKN_U32_FMT_FREQ:
3469 	case SKL_TKN_U32_FMT_BIT_DEPTH:
3470 	case SKL_TKN_U32_FMT_SAMPLE_SIZE:
3471 	case SKL_TKN_U32_FMT_CH_CONFIG:
3472 	case SKL_TKN_U32_FMT_INTERLEAVE:
3473 	case SKL_TKN_U32_FMT_SAMPLE_TYPE:
3474 	case SKL_TKN_U32_FMT_CH_MAP:
3475 	case SKL_TKN_MM_U32_INTF_PIN_ID:
3476 		ret = skl_tplg_manifest_fill_fmt(dev, fmt, tkn_elem,
3477 						 dir, pin_idx);
3478 		if (ret < 0)
3479 			return ret;
3480 		break;
3481 
3482 	default:
3483 		dev_err(dev, "Not a manifest token %d\n", tkn_elem->token);
3484 		return -EINVAL;
3485 	}
3486 	tkn_count++;
3487 
3488 	return tkn_count;
3489 }
3490 
3491 /*
3492  * Fill the manifest structure by parsing the tokens based on the
3493  * type.
3494  */
skl_tplg_get_manifest_tkn(struct device * dev,char * pvt_data,struct skl_dev * skl,int block_size)3495 static int skl_tplg_get_manifest_tkn(struct device *dev,
3496 		char *pvt_data, struct skl_dev *skl,
3497 		int block_size)
3498 {
3499 	int tkn_count = 0, ret;
3500 	int off = 0, tuple_size = 0;
3501 	u8 uuid_index = 0;
3502 	struct snd_soc_tplg_vendor_array *array;
3503 	struct snd_soc_tplg_vendor_value_elem *tkn_elem;
3504 
3505 	if (block_size <= 0)
3506 		return -EINVAL;
3507 
3508 	while (tuple_size < block_size) {
3509 		array = (struct snd_soc_tplg_vendor_array *)(pvt_data + off);
3510 		off += array->size;
3511 		switch (array->type) {
3512 		case SND_SOC_TPLG_TUPLE_TYPE_STRING:
3513 			ret = skl_tplg_get_str_tkn(dev, array, skl);
3514 
3515 			if (ret < 0)
3516 				return ret;
3517 			tkn_count = ret;
3518 
3519 			tuple_size += tkn_count *
3520 				sizeof(struct snd_soc_tplg_vendor_string_elem);
3521 			continue;
3522 
3523 		case SND_SOC_TPLG_TUPLE_TYPE_UUID:
3524 			if (array->uuid->token != SKL_TKN_UUID) {
3525 				dev_err(dev, "Not an UUID token: %d\n",
3526 					array->uuid->token);
3527 				return -EINVAL;
3528 			}
3529 			if (uuid_index >= skl->nr_modules) {
3530 				dev_err(dev, "Too many UUID tokens\n");
3531 				return -EINVAL;
3532 			}
3533 			import_guid(&skl->modules[uuid_index++]->uuid,
3534 				    array->uuid->uuid);
3535 
3536 			tuple_size += sizeof(*array->uuid);
3537 			continue;
3538 
3539 		default:
3540 			tkn_elem = array->value;
3541 			tkn_count = 0;
3542 			break;
3543 		}
3544 
3545 		while (tkn_count <= array->num_elems - 1) {
3546 			ret = skl_tplg_get_int_tkn(dev,
3547 					tkn_elem, skl);
3548 			if (ret < 0)
3549 				return ret;
3550 
3551 			tkn_count = tkn_count + ret;
3552 			tkn_elem++;
3553 		}
3554 		tuple_size += (tkn_count * sizeof(*tkn_elem));
3555 		tkn_count = 0;
3556 	}
3557 
3558 	return off;
3559 }
3560 
3561 /*
3562  * Parse manifest private data for tokens. The private data block is
3563  * preceded by descriptors for type and size of data block.
3564  */
skl_tplg_get_manifest_data(struct snd_soc_tplg_manifest * manifest,struct device * dev,struct skl_dev * skl)3565 static int skl_tplg_get_manifest_data(struct snd_soc_tplg_manifest *manifest,
3566 			struct device *dev, struct skl_dev *skl)
3567 {
3568 	struct snd_soc_tplg_vendor_array *array;
3569 	int num_blocks, block_size = 0, block_type, off = 0;
3570 	char *data;
3571 	int ret;
3572 
3573 	/* Read the NUM_DATA_BLOCKS descriptor */
3574 	array = (struct snd_soc_tplg_vendor_array *)manifest->priv.data;
3575 	ret = skl_tplg_get_desc_blocks(dev, array);
3576 	if (ret < 0)
3577 		return ret;
3578 	num_blocks = ret;
3579 
3580 	off += array->size;
3581 	/* Read the BLOCK_TYPE and BLOCK_SIZE descriptor */
3582 	while (num_blocks > 0) {
3583 		array = (struct snd_soc_tplg_vendor_array *)
3584 				(manifest->priv.data + off);
3585 		ret = skl_tplg_get_desc_blocks(dev, array);
3586 
3587 		if (ret < 0)
3588 			return ret;
3589 		block_type = ret;
3590 		off += array->size;
3591 
3592 		array = (struct snd_soc_tplg_vendor_array *)
3593 			(manifest->priv.data + off);
3594 
3595 		ret = skl_tplg_get_desc_blocks(dev, array);
3596 
3597 		if (ret < 0)
3598 			return ret;
3599 		block_size = ret;
3600 		off += array->size;
3601 
3602 		array = (struct snd_soc_tplg_vendor_array *)
3603 			(manifest->priv.data + off);
3604 
3605 		data = (manifest->priv.data + off);
3606 
3607 		if (block_type == SKL_TYPE_TUPLE) {
3608 			ret = skl_tplg_get_manifest_tkn(dev, data, skl,
3609 					block_size);
3610 
3611 			if (ret < 0)
3612 				return ret;
3613 
3614 			--num_blocks;
3615 		} else {
3616 			return -EINVAL;
3617 		}
3618 		off += ret;
3619 	}
3620 
3621 	return 0;
3622 }
3623 
skl_manifest_load(struct snd_soc_component * cmpnt,int index,struct snd_soc_tplg_manifest * manifest)3624 static int skl_manifest_load(struct snd_soc_component *cmpnt, int index,
3625 				struct snd_soc_tplg_manifest *manifest)
3626 {
3627 	struct hdac_bus *bus = snd_soc_component_get_drvdata(cmpnt);
3628 	struct skl_dev *skl = bus_to_skl(bus);
3629 
3630 	/* proceed only if we have private data defined */
3631 	if (manifest->priv.size == 0)
3632 		return 0;
3633 
3634 	skl_tplg_get_manifest_data(manifest, bus->dev, skl);
3635 
3636 	if (skl->lib_count > SKL_MAX_LIB) {
3637 		dev_err(bus->dev, "Exceeding max Library count. Got:%d\n",
3638 					skl->lib_count);
3639 		return  -EINVAL;
3640 	}
3641 
3642 	return 0;
3643 }
3644 
skl_tplg_complete(struct snd_soc_component * component)3645 static int skl_tplg_complete(struct snd_soc_component *component)
3646 {
3647 	struct snd_soc_dobj *dobj;
3648 	struct snd_soc_acpi_mach *mach;
3649 	struct snd_ctl_elem_value *val;
3650 	int i;
3651 
3652 	val = kmalloc(sizeof(*val), GFP_KERNEL);
3653 	if (!val)
3654 		return -ENOMEM;
3655 
3656 	mach = dev_get_platdata(component->card->dev);
3657 	list_for_each_entry(dobj, &component->dobj_list, list) {
3658 		struct snd_kcontrol *kcontrol = dobj->control.kcontrol;
3659 		struct soc_enum *se;
3660 		char **texts;
3661 		char chan_text[4];
3662 
3663 		if (dobj->type != SND_SOC_DOBJ_ENUM || !kcontrol ||
3664 		    kcontrol->put != skl_tplg_multi_config_set_dmic)
3665 			continue;
3666 
3667 		se = (struct soc_enum *)kcontrol->private_value;
3668 		texts = dobj->control.dtexts;
3669 		sprintf(chan_text, "c%d", mach->mach_params.dmic_num);
3670 
3671 		for (i = 0; i < se->items; i++) {
3672 			if (strstr(texts[i], chan_text)) {
3673 				memset(val, 0, sizeof(*val));
3674 				val->value.enumerated.item[0] = i;
3675 				kcontrol->put(kcontrol, val);
3676 			}
3677 		}
3678 	}
3679 
3680 	kfree(val);
3681 	return 0;
3682 }
3683 
3684 static struct snd_soc_tplg_ops skl_tplg_ops  = {
3685 	.widget_load = skl_tplg_widget_load,
3686 	.control_load = skl_tplg_control_load,
3687 	.bytes_ext_ops = skl_tlv_ops,
3688 	.bytes_ext_ops_count = ARRAY_SIZE(skl_tlv_ops),
3689 	.io_ops = skl_tplg_kcontrol_ops,
3690 	.io_ops_count = ARRAY_SIZE(skl_tplg_kcontrol_ops),
3691 	.manifest = skl_manifest_load,
3692 	.dai_load = skl_dai_load,
3693 	.complete = skl_tplg_complete,
3694 };
3695 
3696 /*
3697  * A pipe can have multiple modules, each of them will be a DAPM widget as
3698  * well. While managing a pipeline we need to get the list of all the
3699  * widgets in a pipelines, so this helper - skl_tplg_create_pipe_widget_list()
3700  * helps to get the SKL type widgets in that pipeline
3701  */
skl_tplg_create_pipe_widget_list(struct snd_soc_component * component)3702 static int skl_tplg_create_pipe_widget_list(struct snd_soc_component *component)
3703 {
3704 	struct snd_soc_dapm_widget *w;
3705 	struct skl_module_cfg *mcfg = NULL;
3706 	struct skl_pipe_module *p_module = NULL;
3707 	struct skl_pipe *pipe;
3708 
3709 	list_for_each_entry(w, &component->card->widgets, list) {
3710 		if (is_skl_dsp_widget_type(w, component->dev) && w->priv) {
3711 			mcfg = w->priv;
3712 			pipe = mcfg->pipe;
3713 
3714 			p_module = devm_kzalloc(component->dev,
3715 						sizeof(*p_module), GFP_KERNEL);
3716 			if (!p_module)
3717 				return -ENOMEM;
3718 
3719 			p_module->w = w;
3720 			list_add_tail(&p_module->node, &pipe->w_list);
3721 		}
3722 	}
3723 
3724 	return 0;
3725 }
3726 
skl_tplg_set_pipe_type(struct skl_dev * skl,struct skl_pipe * pipe)3727 static void skl_tplg_set_pipe_type(struct skl_dev *skl, struct skl_pipe *pipe)
3728 {
3729 	struct skl_pipe_module *w_module;
3730 	struct snd_soc_dapm_widget *w;
3731 	struct skl_module_cfg *mconfig;
3732 	bool host_found = false, link_found = false;
3733 
3734 	list_for_each_entry(w_module, &pipe->w_list, node) {
3735 		w = w_module->w;
3736 		mconfig = w->priv;
3737 
3738 		if (mconfig->dev_type == SKL_DEVICE_HDAHOST)
3739 			host_found = true;
3740 		else if (mconfig->dev_type != SKL_DEVICE_NONE)
3741 			link_found = true;
3742 	}
3743 
3744 	if (host_found && link_found)
3745 		pipe->passthru = true;
3746 	else
3747 		pipe->passthru = false;
3748 }
3749 
3750 /*
3751  * SKL topology init routine
3752  */
skl_tplg_init(struct snd_soc_component * component,struct hdac_bus * bus)3753 int skl_tplg_init(struct snd_soc_component *component, struct hdac_bus *bus)
3754 {
3755 	int ret;
3756 	const struct firmware *fw;
3757 	struct skl_dev *skl = bus_to_skl(bus);
3758 	struct skl_pipeline *ppl;
3759 
3760 	ret = request_firmware(&fw, skl->tplg_name, bus->dev);
3761 	if (ret < 0) {
3762 		char alt_tplg_name[64];
3763 
3764 		snprintf(alt_tplg_name, sizeof(alt_tplg_name), "%s-tplg.bin",
3765 			 skl->mach->drv_name);
3766 		dev_info(bus->dev, "tplg fw %s load failed with %d, trying alternative tplg name %s",
3767 			 skl->tplg_name, ret, alt_tplg_name);
3768 
3769 		ret = request_firmware(&fw, alt_tplg_name, bus->dev);
3770 		if (!ret)
3771 			goto component_load;
3772 
3773 		dev_info(bus->dev, "tplg %s failed with %d, falling back to dfw_sst.bin",
3774 			 alt_tplg_name, ret);
3775 
3776 		ret = request_firmware(&fw, "dfw_sst.bin", bus->dev);
3777 		if (ret < 0) {
3778 			dev_err(bus->dev, "Fallback tplg fw %s load failed with %d\n",
3779 					"dfw_sst.bin", ret);
3780 			return ret;
3781 		}
3782 	}
3783 
3784 component_load:
3785 	ret = snd_soc_tplg_component_load(component, &skl_tplg_ops, fw);
3786 	if (ret < 0) {
3787 		dev_err(bus->dev, "tplg component load failed%d\n", ret);
3788 		goto err;
3789 	}
3790 
3791 	ret = skl_tplg_create_pipe_widget_list(component);
3792 	if (ret < 0) {
3793 		dev_err(bus->dev, "tplg create pipe widget list failed%d\n",
3794 				ret);
3795 		goto err;
3796 	}
3797 
3798 	list_for_each_entry(ppl, &skl->ppl_list, node)
3799 		skl_tplg_set_pipe_type(skl, ppl->pipe);
3800 
3801 err:
3802 	release_firmware(fw);
3803 	return ret;
3804 }
3805 
skl_tplg_exit(struct snd_soc_component * component,struct hdac_bus * bus)3806 void skl_tplg_exit(struct snd_soc_component *component, struct hdac_bus *bus)
3807 {
3808 	struct skl_dev *skl = bus_to_skl(bus);
3809 	struct skl_pipeline *ppl, *tmp;
3810 
3811 	list_for_each_entry_safe(ppl, tmp, &skl->ppl_list, node)
3812 		list_del(&ppl->node);
3813 
3814 	/* clean up topology */
3815 	snd_soc_tplg_component_remove(component);
3816 }
3817