1 // SPDX-License-Identifier: GPL-2.0-only
2 /*
3 * skl-topology.c - Implements Platform component ALSA controls/widget
4 * handlers.
5 *
6 * Copyright (C) 2014-2015 Intel Corp
7 * Author: Jeeja KP <jeeja.kp@intel.com>
8 * ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
9 */
10
11 #include <linux/slab.h>
12 #include <linux/types.h>
13 #include <linux/firmware.h>
14 #include <linux/uuid.h>
15 #include <sound/intel-nhlt.h>
16 #include <sound/soc.h>
17 #include <sound/soc-acpi.h>
18 #include <sound/soc-topology.h>
19 #include <uapi/sound/snd_sst_tokens.h>
20 #include <uapi/sound/skl-tplg-interface.h>
21 #include "skl-sst-dsp.h"
22 #include "skl-sst-ipc.h"
23 #include "skl-topology.h"
24 #include "skl.h"
25 #include "../common/sst-dsp.h"
26 #include "../common/sst-dsp-priv.h"
27
28 #define SKL_CH_FIXUP_MASK (1 << 0)
29 #define SKL_RATE_FIXUP_MASK (1 << 1)
30 #define SKL_FMT_FIXUP_MASK (1 << 2)
31 #define SKL_IN_DIR_BIT_MASK BIT(0)
32 #define SKL_PIN_COUNT_MASK GENMASK(7, 4)
33
34 static const int mic_mono_list[] = {
35 0, 1, 2, 3,
36 };
37 static const int mic_stereo_list[][SKL_CH_STEREO] = {
38 {0, 1}, {0, 2}, {0, 3}, {1, 2}, {1, 3}, {2, 3},
39 };
40 static const int mic_trio_list[][SKL_CH_TRIO] = {
41 {0, 1, 2}, {0, 1, 3}, {0, 2, 3}, {1, 2, 3},
42 };
43 static const int mic_quatro_list[][SKL_CH_QUATRO] = {
44 {0, 1, 2, 3},
45 };
46
47 #define CHECK_HW_PARAMS(ch, freq, bps, prm_ch, prm_freq, prm_bps) \
48 ((ch == prm_ch) && (bps == prm_bps) && (freq == prm_freq))
49
skl_tplg_d0i3_get(struct skl_dev * skl,enum d0i3_capability caps)50 void skl_tplg_d0i3_get(struct skl_dev *skl, enum d0i3_capability caps)
51 {
52 struct skl_d0i3_data *d0i3 = &skl->d0i3;
53
54 switch (caps) {
55 case SKL_D0I3_NONE:
56 d0i3->non_d0i3++;
57 break;
58
59 case SKL_D0I3_STREAMING:
60 d0i3->streaming++;
61 break;
62
63 case SKL_D0I3_NON_STREAMING:
64 d0i3->non_streaming++;
65 break;
66 }
67 }
68
skl_tplg_d0i3_put(struct skl_dev * skl,enum d0i3_capability caps)69 void skl_tplg_d0i3_put(struct skl_dev *skl, enum d0i3_capability caps)
70 {
71 struct skl_d0i3_data *d0i3 = &skl->d0i3;
72
73 switch (caps) {
74 case SKL_D0I3_NONE:
75 d0i3->non_d0i3--;
76 break;
77
78 case SKL_D0I3_STREAMING:
79 d0i3->streaming--;
80 break;
81
82 case SKL_D0I3_NON_STREAMING:
83 d0i3->non_streaming--;
84 break;
85 }
86 }
87
88 /*
89 * SKL DSP driver modelling uses only few DAPM widgets so for rest we will
90 * ignore. This helpers checks if the SKL driver handles this widget type
91 */
is_skl_dsp_widget_type(struct snd_soc_dapm_widget * w,struct device * dev)92 static int is_skl_dsp_widget_type(struct snd_soc_dapm_widget *w,
93 struct device *dev)
94 {
95 if (w->dapm->dev != dev)
96 return false;
97
98 switch (w->id) {
99 case snd_soc_dapm_dai_link:
100 case snd_soc_dapm_dai_in:
101 case snd_soc_dapm_aif_in:
102 case snd_soc_dapm_aif_out:
103 case snd_soc_dapm_dai_out:
104 case snd_soc_dapm_switch:
105 case snd_soc_dapm_output:
106 case snd_soc_dapm_mux:
107
108 return false;
109 default:
110 return true;
111 }
112 }
113
skl_dump_mconfig(struct skl_dev * skl,struct skl_module_cfg * mcfg)114 static void skl_dump_mconfig(struct skl_dev *skl, struct skl_module_cfg *mcfg)
115 {
116 struct skl_module_iface *iface = &mcfg->module->formats[mcfg->fmt_idx];
117
118 dev_dbg(skl->dev, "Dumping config\n");
119 dev_dbg(skl->dev, "Input Format:\n");
120 dev_dbg(skl->dev, "channels = %d\n", iface->inputs[0].fmt.channels);
121 dev_dbg(skl->dev, "s_freq = %d\n", iface->inputs[0].fmt.s_freq);
122 dev_dbg(skl->dev, "ch_cfg = %d\n", iface->inputs[0].fmt.ch_cfg);
123 dev_dbg(skl->dev, "valid bit depth = %d\n",
124 iface->inputs[0].fmt.valid_bit_depth);
125 dev_dbg(skl->dev, "Output Format:\n");
126 dev_dbg(skl->dev, "channels = %d\n", iface->outputs[0].fmt.channels);
127 dev_dbg(skl->dev, "s_freq = %d\n", iface->outputs[0].fmt.s_freq);
128 dev_dbg(skl->dev, "valid bit depth = %d\n",
129 iface->outputs[0].fmt.valid_bit_depth);
130 dev_dbg(skl->dev, "ch_cfg = %d\n", iface->outputs[0].fmt.ch_cfg);
131 }
132
skl_tplg_update_chmap(struct skl_module_fmt * fmt,int chs)133 static void skl_tplg_update_chmap(struct skl_module_fmt *fmt, int chs)
134 {
135 int slot_map = 0xFFFFFFFF;
136 int start_slot = 0;
137 int i;
138
139 for (i = 0; i < chs; i++) {
140 /*
141 * For 2 channels with starting slot as 0, slot map will
142 * look like 0xFFFFFF10.
143 */
144 slot_map &= (~(0xF << (4 * i)) | (start_slot << (4 * i)));
145 start_slot++;
146 }
147 fmt->ch_map = slot_map;
148 }
149
skl_tplg_update_params(struct skl_module_fmt * fmt,struct skl_pipe_params * params,int fixup)150 static void skl_tplg_update_params(struct skl_module_fmt *fmt,
151 struct skl_pipe_params *params, int fixup)
152 {
153 if (fixup & SKL_RATE_FIXUP_MASK)
154 fmt->s_freq = params->s_freq;
155 if (fixup & SKL_CH_FIXUP_MASK) {
156 fmt->channels = params->ch;
157 skl_tplg_update_chmap(fmt, fmt->channels);
158 }
159 if (fixup & SKL_FMT_FIXUP_MASK) {
160 fmt->valid_bit_depth = skl_get_bit_depth(params->s_fmt);
161
162 /*
163 * 16 bit is 16 bit container whereas 24 bit is in 32 bit
164 * container so update bit depth accordingly
165 */
166 switch (fmt->valid_bit_depth) {
167 case SKL_DEPTH_16BIT:
168 fmt->bit_depth = fmt->valid_bit_depth;
169 break;
170
171 default:
172 fmt->bit_depth = SKL_DEPTH_32BIT;
173 break;
174 }
175 }
176
177 }
178
179 /*
180 * A pipeline may have modules which impact the pcm parameters, like SRC,
181 * channel converter, format converter.
182 * We need to calculate the output params by applying the 'fixup'
183 * Topology will tell driver which type of fixup is to be applied by
184 * supplying the fixup mask, so based on that we calculate the output
185 *
186 * Now In FE the pcm hw_params is source/target format. Same is applicable
187 * for BE with its hw_params invoked.
188 * here based on FE, BE pipeline and direction we calculate the input and
189 * outfix and then apply that for a module
190 */
skl_tplg_update_params_fixup(struct skl_module_cfg * m_cfg,struct skl_pipe_params * params,bool is_fe)191 static void skl_tplg_update_params_fixup(struct skl_module_cfg *m_cfg,
192 struct skl_pipe_params *params, bool is_fe)
193 {
194 int in_fixup, out_fixup;
195 struct skl_module_fmt *in_fmt, *out_fmt;
196
197 /* Fixups will be applied to pin 0 only */
198 in_fmt = &m_cfg->module->formats[m_cfg->fmt_idx].inputs[0].fmt;
199 out_fmt = &m_cfg->module->formats[m_cfg->fmt_idx].outputs[0].fmt;
200
201 if (params->stream == SNDRV_PCM_STREAM_PLAYBACK) {
202 if (is_fe) {
203 in_fixup = m_cfg->params_fixup;
204 out_fixup = (~m_cfg->converter) &
205 m_cfg->params_fixup;
206 } else {
207 out_fixup = m_cfg->params_fixup;
208 in_fixup = (~m_cfg->converter) &
209 m_cfg->params_fixup;
210 }
211 } else {
212 if (is_fe) {
213 out_fixup = m_cfg->params_fixup;
214 in_fixup = (~m_cfg->converter) &
215 m_cfg->params_fixup;
216 } else {
217 in_fixup = m_cfg->params_fixup;
218 out_fixup = (~m_cfg->converter) &
219 m_cfg->params_fixup;
220 }
221 }
222
223 skl_tplg_update_params(in_fmt, params, in_fixup);
224 skl_tplg_update_params(out_fmt, params, out_fixup);
225 }
226
227 /*
228 * A module needs input and output buffers, which are dependent upon pcm
229 * params, so once we have calculate params, we need buffer calculation as
230 * well.
231 */
skl_tplg_update_buffer_size(struct skl_dev * skl,struct skl_module_cfg * mcfg)232 static void skl_tplg_update_buffer_size(struct skl_dev *skl,
233 struct skl_module_cfg *mcfg)
234 {
235 int multiplier = 1;
236 struct skl_module_fmt *in_fmt, *out_fmt;
237 struct skl_module_res *res;
238
239 /* Since fixups is applied to pin 0 only, ibs, obs needs
240 * change for pin 0 only
241 */
242 res = &mcfg->module->resources[mcfg->res_idx];
243 in_fmt = &mcfg->module->formats[mcfg->fmt_idx].inputs[0].fmt;
244 out_fmt = &mcfg->module->formats[mcfg->fmt_idx].outputs[0].fmt;
245
246 if (mcfg->m_type == SKL_MODULE_TYPE_SRCINT)
247 multiplier = 5;
248
249 res->ibs = DIV_ROUND_UP(in_fmt->s_freq, 1000) *
250 in_fmt->channels * (in_fmt->bit_depth >> 3) *
251 multiplier;
252
253 res->obs = DIV_ROUND_UP(out_fmt->s_freq, 1000) *
254 out_fmt->channels * (out_fmt->bit_depth >> 3) *
255 multiplier;
256 }
257
skl_tplg_be_dev_type(int dev_type)258 static u8 skl_tplg_be_dev_type(int dev_type)
259 {
260 int ret;
261
262 switch (dev_type) {
263 case SKL_DEVICE_BT:
264 ret = NHLT_DEVICE_BT;
265 break;
266
267 case SKL_DEVICE_DMIC:
268 ret = NHLT_DEVICE_DMIC;
269 break;
270
271 case SKL_DEVICE_I2S:
272 ret = NHLT_DEVICE_I2S;
273 break;
274
275 default:
276 ret = NHLT_DEVICE_INVALID;
277 break;
278 }
279
280 return ret;
281 }
282
skl_tplg_update_be_blob(struct snd_soc_dapm_widget * w,struct skl_dev * skl)283 static int skl_tplg_update_be_blob(struct snd_soc_dapm_widget *w,
284 struct skl_dev *skl)
285 {
286 struct skl_module_cfg *m_cfg = w->priv;
287 int link_type, dir;
288 u32 ch, s_freq, s_fmt, s_cont;
289 struct nhlt_specific_cfg *cfg;
290 u8 dev_type = skl_tplg_be_dev_type(m_cfg->dev_type);
291 int fmt_idx = m_cfg->fmt_idx;
292 struct skl_module_iface *m_iface = &m_cfg->module->formats[fmt_idx];
293
294 /* check if we already have blob */
295 if (m_cfg->formats_config[SKL_PARAM_INIT].caps_size > 0)
296 return 0;
297
298 dev_dbg(skl->dev, "Applying default cfg blob\n");
299 switch (m_cfg->dev_type) {
300 case SKL_DEVICE_DMIC:
301 link_type = NHLT_LINK_DMIC;
302 dir = SNDRV_PCM_STREAM_CAPTURE;
303 s_freq = m_iface->inputs[0].fmt.s_freq;
304 s_fmt = m_iface->inputs[0].fmt.valid_bit_depth;
305 s_cont = m_iface->inputs[0].fmt.bit_depth;
306 ch = m_iface->inputs[0].fmt.channels;
307 break;
308
309 case SKL_DEVICE_I2S:
310 link_type = NHLT_LINK_SSP;
311 if (m_cfg->hw_conn_type == SKL_CONN_SOURCE) {
312 dir = SNDRV_PCM_STREAM_PLAYBACK;
313 s_freq = m_iface->outputs[0].fmt.s_freq;
314 s_fmt = m_iface->outputs[0].fmt.valid_bit_depth;
315 s_cont = m_iface->outputs[0].fmt.bit_depth;
316 ch = m_iface->outputs[0].fmt.channels;
317 } else {
318 dir = SNDRV_PCM_STREAM_CAPTURE;
319 s_freq = m_iface->inputs[0].fmt.s_freq;
320 s_fmt = m_iface->inputs[0].fmt.valid_bit_depth;
321 s_cont = m_iface->inputs[0].fmt.bit_depth;
322 ch = m_iface->inputs[0].fmt.channels;
323 }
324 break;
325
326 default:
327 return -EINVAL;
328 }
329
330 /* update the blob based on virtual bus_id and default params */
331 cfg = intel_nhlt_get_endpoint_blob(skl->dev, skl->nhlt, m_cfg->vbus_id,
332 link_type, s_fmt, s_cont, ch,
333 s_freq, dir, dev_type);
334 if (cfg) {
335 m_cfg->formats_config[SKL_PARAM_INIT].caps_size = cfg->size;
336 m_cfg->formats_config[SKL_PARAM_INIT].caps = (u32 *)&cfg->caps;
337 } else {
338 dev_err(skl->dev, "Blob NULL for id %x type %d dirn %d\n",
339 m_cfg->vbus_id, link_type, dir);
340 dev_err(skl->dev, "PCM: ch %d, freq %d, fmt %d/%d\n",
341 ch, s_freq, s_fmt, s_cont);
342 return -EIO;
343 }
344
345 return 0;
346 }
347
skl_tplg_update_module_params(struct snd_soc_dapm_widget * w,struct skl_dev * skl)348 static void skl_tplg_update_module_params(struct snd_soc_dapm_widget *w,
349 struct skl_dev *skl)
350 {
351 struct skl_module_cfg *m_cfg = w->priv;
352 struct skl_pipe_params *params = m_cfg->pipe->p_params;
353 int p_conn_type = m_cfg->pipe->conn_type;
354 bool is_fe;
355
356 if (!m_cfg->params_fixup)
357 return;
358
359 dev_dbg(skl->dev, "Mconfig for widget=%s BEFORE updation\n",
360 w->name);
361
362 skl_dump_mconfig(skl, m_cfg);
363
364 if (p_conn_type == SKL_PIPE_CONN_TYPE_FE)
365 is_fe = true;
366 else
367 is_fe = false;
368
369 skl_tplg_update_params_fixup(m_cfg, params, is_fe);
370 skl_tplg_update_buffer_size(skl, m_cfg);
371
372 dev_dbg(skl->dev, "Mconfig for widget=%s AFTER updation\n",
373 w->name);
374
375 skl_dump_mconfig(skl, m_cfg);
376 }
377
378 /*
379 * some modules can have multiple params set from user control and
380 * need to be set after module is initialized. If set_param flag is
381 * set module params will be done after module is initialised.
382 */
skl_tplg_set_module_params(struct snd_soc_dapm_widget * w,struct skl_dev * skl)383 static int skl_tplg_set_module_params(struct snd_soc_dapm_widget *w,
384 struct skl_dev *skl)
385 {
386 int i, ret;
387 struct skl_module_cfg *mconfig = w->priv;
388 const struct snd_kcontrol_new *k;
389 struct soc_bytes_ext *sb;
390 struct skl_algo_data *bc;
391 struct skl_specific_cfg *sp_cfg;
392
393 if (mconfig->formats_config[SKL_PARAM_SET].caps_size > 0 &&
394 mconfig->formats_config[SKL_PARAM_SET].set_params == SKL_PARAM_SET) {
395 sp_cfg = &mconfig->formats_config[SKL_PARAM_SET];
396 ret = skl_set_module_params(skl, sp_cfg->caps,
397 sp_cfg->caps_size,
398 sp_cfg->param_id, mconfig);
399 if (ret < 0)
400 return ret;
401 }
402
403 for (i = 0; i < w->num_kcontrols; i++) {
404 k = &w->kcontrol_news[i];
405 if (k->access & SNDRV_CTL_ELEM_ACCESS_TLV_CALLBACK) {
406 sb = (void *) k->private_value;
407 bc = (struct skl_algo_data *)sb->dobj.private;
408
409 if (bc->set_params == SKL_PARAM_SET) {
410 ret = skl_set_module_params(skl,
411 (u32 *)bc->params, bc->size,
412 bc->param_id, mconfig);
413 if (ret < 0)
414 return ret;
415 }
416 }
417 }
418
419 return 0;
420 }
421
422 /*
423 * some module param can set from user control and this is required as
424 * when module is initailzed. if module param is required in init it is
425 * identifed by set_param flag. if set_param flag is not set, then this
426 * parameter needs to set as part of module init.
427 */
skl_tplg_set_module_init_data(struct snd_soc_dapm_widget * w)428 static int skl_tplg_set_module_init_data(struct snd_soc_dapm_widget *w)
429 {
430 const struct snd_kcontrol_new *k;
431 struct soc_bytes_ext *sb;
432 struct skl_algo_data *bc;
433 struct skl_module_cfg *mconfig = w->priv;
434 int i;
435
436 for (i = 0; i < w->num_kcontrols; i++) {
437 k = &w->kcontrol_news[i];
438 if (k->access & SNDRV_CTL_ELEM_ACCESS_TLV_CALLBACK) {
439 sb = (struct soc_bytes_ext *)k->private_value;
440 bc = (struct skl_algo_data *)sb->dobj.private;
441
442 if (bc->set_params != SKL_PARAM_INIT)
443 continue;
444
445 mconfig->formats_config[SKL_PARAM_INIT].caps =
446 (u32 *)bc->params;
447 mconfig->formats_config[SKL_PARAM_INIT].caps_size =
448 bc->size;
449
450 break;
451 }
452 }
453
454 return 0;
455 }
456
skl_tplg_module_prepare(struct skl_dev * skl,struct skl_pipe * pipe,struct snd_soc_dapm_widget * w,struct skl_module_cfg * mcfg)457 static int skl_tplg_module_prepare(struct skl_dev *skl, struct skl_pipe *pipe,
458 struct snd_soc_dapm_widget *w, struct skl_module_cfg *mcfg)
459 {
460 switch (mcfg->dev_type) {
461 case SKL_DEVICE_HDAHOST:
462 return skl_pcm_host_dma_prepare(skl->dev, pipe->p_params);
463
464 case SKL_DEVICE_HDALINK:
465 return skl_pcm_link_dma_prepare(skl->dev, pipe->p_params);
466 }
467
468 return 0;
469 }
470
471 /*
472 * Inside a pipe instance, we can have various modules. These modules need
473 * to instantiated in DSP by invoking INIT_MODULE IPC, which is achieved by
474 * skl_init_module() routine, so invoke that for all modules in a pipeline
475 */
476 static int
skl_tplg_init_pipe_modules(struct skl_dev * skl,struct skl_pipe * pipe)477 skl_tplg_init_pipe_modules(struct skl_dev *skl, struct skl_pipe *pipe)
478 {
479 struct skl_pipe_module *w_module;
480 struct snd_soc_dapm_widget *w;
481 struct skl_module_cfg *mconfig;
482 u8 cfg_idx;
483 int ret = 0;
484
485 list_for_each_entry(w_module, &pipe->w_list, node) {
486 guid_t *uuid_mod;
487 w = w_module->w;
488 mconfig = w->priv;
489
490 /* check if module ids are populated */
491 if (mconfig->id.module_id < 0) {
492 dev_err(skl->dev,
493 "module %pUL id not populated\n",
494 (guid_t *)mconfig->guid);
495 return -EIO;
496 }
497
498 cfg_idx = mconfig->pipe->cur_config_idx;
499 mconfig->fmt_idx = mconfig->mod_cfg[cfg_idx].fmt_idx;
500 mconfig->res_idx = mconfig->mod_cfg[cfg_idx].res_idx;
501
502 if (mconfig->module->loadable && skl->dsp->fw_ops.load_mod) {
503 ret = skl->dsp->fw_ops.load_mod(skl->dsp,
504 mconfig->id.module_id, mconfig->guid);
505 if (ret < 0)
506 return ret;
507 }
508
509 /* prepare the DMA if the module is gateway cpr */
510 ret = skl_tplg_module_prepare(skl, pipe, w, mconfig);
511 if (ret < 0)
512 return ret;
513
514 /* update blob if blob is null for be with default value */
515 skl_tplg_update_be_blob(w, skl);
516
517 /*
518 * apply fix/conversion to module params based on
519 * FE/BE params
520 */
521 skl_tplg_update_module_params(w, skl);
522 uuid_mod = (guid_t *)mconfig->guid;
523 mconfig->id.pvt_id = skl_get_pvt_id(skl, uuid_mod,
524 mconfig->id.instance_id);
525 if (mconfig->id.pvt_id < 0)
526 return ret;
527 skl_tplg_set_module_init_data(w);
528
529 ret = skl_dsp_get_core(skl->dsp, mconfig->core_id);
530 if (ret < 0) {
531 dev_err(skl->dev, "Failed to wake up core %d ret=%d\n",
532 mconfig->core_id, ret);
533 return ret;
534 }
535
536 ret = skl_init_module(skl, mconfig);
537 if (ret < 0) {
538 skl_put_pvt_id(skl, uuid_mod, &mconfig->id.pvt_id);
539 goto err;
540 }
541
542 ret = skl_tplg_set_module_params(w, skl);
543 if (ret < 0)
544 goto err;
545 }
546
547 return 0;
548 err:
549 skl_dsp_put_core(skl->dsp, mconfig->core_id);
550 return ret;
551 }
552
skl_tplg_unload_pipe_modules(struct skl_dev * skl,struct skl_pipe * pipe)553 static int skl_tplg_unload_pipe_modules(struct skl_dev *skl,
554 struct skl_pipe *pipe)
555 {
556 int ret = 0;
557 struct skl_pipe_module *w_module;
558 struct skl_module_cfg *mconfig;
559
560 list_for_each_entry(w_module, &pipe->w_list, node) {
561 guid_t *uuid_mod;
562 mconfig = w_module->w->priv;
563 uuid_mod = (guid_t *)mconfig->guid;
564
565 if (mconfig->module->loadable && skl->dsp->fw_ops.unload_mod) {
566 ret = skl->dsp->fw_ops.unload_mod(skl->dsp,
567 mconfig->id.module_id);
568 if (ret < 0)
569 return -EIO;
570 }
571 skl_put_pvt_id(skl, uuid_mod, &mconfig->id.pvt_id);
572
573 ret = skl_dsp_put_core(skl->dsp, mconfig->core_id);
574 if (ret < 0) {
575 /* don't return; continue with other modules */
576 dev_err(skl->dev, "Failed to sleep core %d ret=%d\n",
577 mconfig->core_id, ret);
578 }
579 }
580
581 /* no modules to unload in this path, so return */
582 return ret;
583 }
584
skl_tplg_is_multi_fmt(struct skl_dev * skl,struct skl_pipe * pipe)585 static bool skl_tplg_is_multi_fmt(struct skl_dev *skl, struct skl_pipe *pipe)
586 {
587 struct skl_pipe_fmt *cur_fmt;
588 struct skl_pipe_fmt *next_fmt;
589 int i;
590
591 if (pipe->nr_cfgs <= 1)
592 return false;
593
594 if (pipe->conn_type != SKL_PIPE_CONN_TYPE_FE)
595 return true;
596
597 for (i = 0; i < pipe->nr_cfgs - 1; i++) {
598 if (pipe->direction == SNDRV_PCM_STREAM_PLAYBACK) {
599 cur_fmt = &pipe->configs[i].out_fmt;
600 next_fmt = &pipe->configs[i + 1].out_fmt;
601 } else {
602 cur_fmt = &pipe->configs[i].in_fmt;
603 next_fmt = &pipe->configs[i + 1].in_fmt;
604 }
605
606 if (!CHECK_HW_PARAMS(cur_fmt->channels, cur_fmt->freq,
607 cur_fmt->bps,
608 next_fmt->channels,
609 next_fmt->freq,
610 next_fmt->bps))
611 return true;
612 }
613
614 return false;
615 }
616
617 /*
618 * Here, we select pipe format based on the pipe type and pipe
619 * direction to determine the current config index for the pipeline.
620 * The config index is then used to select proper module resources.
621 * Intermediate pipes currently have a fixed format hence we select the
622 * 0th configuratation by default for such pipes.
623 */
624 static int
skl_tplg_get_pipe_config(struct skl_dev * skl,struct skl_module_cfg * mconfig)625 skl_tplg_get_pipe_config(struct skl_dev *skl, struct skl_module_cfg *mconfig)
626 {
627 struct skl_pipe *pipe = mconfig->pipe;
628 struct skl_pipe_params *params = pipe->p_params;
629 struct skl_path_config *pconfig = &pipe->configs[0];
630 struct skl_pipe_fmt *fmt = NULL;
631 bool in_fmt = false;
632 int i;
633
634 if (pipe->nr_cfgs == 0) {
635 pipe->cur_config_idx = 0;
636 return 0;
637 }
638
639 if (skl_tplg_is_multi_fmt(skl, pipe)) {
640 pipe->cur_config_idx = pipe->pipe_config_idx;
641 pipe->memory_pages = pconfig->mem_pages;
642 dev_dbg(skl->dev, "found pipe config idx:%d\n",
643 pipe->cur_config_idx);
644 return 0;
645 }
646
647 if (pipe->conn_type == SKL_PIPE_CONN_TYPE_NONE || pipe->nr_cfgs == 1) {
648 dev_dbg(skl->dev, "No conn_type or just 1 pathcfg, taking 0th for %d\n",
649 pipe->ppl_id);
650 pipe->cur_config_idx = 0;
651 pipe->memory_pages = pconfig->mem_pages;
652
653 return 0;
654 }
655
656 if ((pipe->conn_type == SKL_PIPE_CONN_TYPE_FE &&
657 pipe->direction == SNDRV_PCM_STREAM_PLAYBACK) ||
658 (pipe->conn_type == SKL_PIPE_CONN_TYPE_BE &&
659 pipe->direction == SNDRV_PCM_STREAM_CAPTURE))
660 in_fmt = true;
661
662 for (i = 0; i < pipe->nr_cfgs; i++) {
663 pconfig = &pipe->configs[i];
664 if (in_fmt)
665 fmt = &pconfig->in_fmt;
666 else
667 fmt = &pconfig->out_fmt;
668
669 if (CHECK_HW_PARAMS(params->ch, params->s_freq, params->s_fmt,
670 fmt->channels, fmt->freq, fmt->bps)) {
671 pipe->cur_config_idx = i;
672 pipe->memory_pages = pconfig->mem_pages;
673 dev_dbg(skl->dev, "Using pipe config: %d\n", i);
674
675 return 0;
676 }
677 }
678
679 dev_err(skl->dev, "Invalid pipe config: %d %d %d for pipe: %d\n",
680 params->ch, params->s_freq, params->s_fmt, pipe->ppl_id);
681 return -EINVAL;
682 }
683
684 /*
685 * Mixer module represents a pipeline. So in the Pre-PMU event of mixer we
686 * need create the pipeline. So we do following:
687 * - Create the pipeline
688 * - Initialize the modules in pipeline
689 * - finally bind all modules together
690 */
skl_tplg_mixer_dapm_pre_pmu_event(struct snd_soc_dapm_widget * w,struct skl_dev * skl)691 static int skl_tplg_mixer_dapm_pre_pmu_event(struct snd_soc_dapm_widget *w,
692 struct skl_dev *skl)
693 {
694 int ret;
695 struct skl_module_cfg *mconfig = w->priv;
696 struct skl_pipe_module *w_module;
697 struct skl_pipe *s_pipe = mconfig->pipe;
698 struct skl_module_cfg *src_module = NULL, *dst_module, *module;
699 struct skl_module_deferred_bind *modules;
700
701 ret = skl_tplg_get_pipe_config(skl, mconfig);
702 if (ret < 0)
703 return ret;
704
705 /*
706 * Create a list of modules for pipe.
707 * This list contains modules from source to sink
708 */
709 ret = skl_create_pipeline(skl, mconfig->pipe);
710 if (ret < 0)
711 return ret;
712
713 /* Init all pipe modules from source to sink */
714 ret = skl_tplg_init_pipe_modules(skl, s_pipe);
715 if (ret < 0)
716 return ret;
717
718 /* Bind modules from source to sink */
719 list_for_each_entry(w_module, &s_pipe->w_list, node) {
720 dst_module = w_module->w->priv;
721
722 if (src_module == NULL) {
723 src_module = dst_module;
724 continue;
725 }
726
727 ret = skl_bind_modules(skl, src_module, dst_module);
728 if (ret < 0)
729 return ret;
730
731 src_module = dst_module;
732 }
733
734 /*
735 * When the destination module is initialized, check for these modules
736 * in deferred bind list. If found, bind them.
737 */
738 list_for_each_entry(w_module, &s_pipe->w_list, node) {
739 if (list_empty(&skl->bind_list))
740 break;
741
742 list_for_each_entry(modules, &skl->bind_list, node) {
743 module = w_module->w->priv;
744 if (modules->dst == module)
745 skl_bind_modules(skl, modules->src,
746 modules->dst);
747 }
748 }
749
750 return 0;
751 }
752
skl_fill_sink_instance_id(struct skl_dev * skl,u32 * params,int size,struct skl_module_cfg * mcfg)753 static int skl_fill_sink_instance_id(struct skl_dev *skl, u32 *params,
754 int size, struct skl_module_cfg *mcfg)
755 {
756 int i, pvt_id;
757
758 if (mcfg->m_type == SKL_MODULE_TYPE_KPB) {
759 struct skl_kpb_params *kpb_params =
760 (struct skl_kpb_params *)params;
761 struct skl_mod_inst_map *inst = kpb_params->u.map;
762
763 for (i = 0; i < kpb_params->num_modules; i++) {
764 pvt_id = skl_get_pvt_instance_id_map(skl, inst->mod_id,
765 inst->inst_id);
766 if (pvt_id < 0)
767 return -EINVAL;
768
769 inst->inst_id = pvt_id;
770 inst++;
771 }
772 }
773
774 return 0;
775 }
776 /*
777 * Some modules require params to be set after the module is bound to
778 * all pins connected.
779 *
780 * The module provider initializes set_param flag for such modules and we
781 * send params after binding
782 */
skl_tplg_set_module_bind_params(struct snd_soc_dapm_widget * w,struct skl_module_cfg * mcfg,struct skl_dev * skl)783 static int skl_tplg_set_module_bind_params(struct snd_soc_dapm_widget *w,
784 struct skl_module_cfg *mcfg, struct skl_dev *skl)
785 {
786 int i, ret;
787 struct skl_module_cfg *mconfig = w->priv;
788 const struct snd_kcontrol_new *k;
789 struct soc_bytes_ext *sb;
790 struct skl_algo_data *bc;
791 struct skl_specific_cfg *sp_cfg;
792 u32 *params;
793
794 /*
795 * check all out/in pins are in bind state.
796 * if so set the module param
797 */
798 for (i = 0; i < mcfg->module->max_output_pins; i++) {
799 if (mcfg->m_out_pin[i].pin_state != SKL_PIN_BIND_DONE)
800 return 0;
801 }
802
803 for (i = 0; i < mcfg->module->max_input_pins; i++) {
804 if (mcfg->m_in_pin[i].pin_state != SKL_PIN_BIND_DONE)
805 return 0;
806 }
807
808 if (mconfig->formats_config[SKL_PARAM_BIND].caps_size > 0 &&
809 mconfig->formats_config[SKL_PARAM_BIND].set_params ==
810 SKL_PARAM_BIND) {
811 sp_cfg = &mconfig->formats_config[SKL_PARAM_BIND];
812 ret = skl_set_module_params(skl, sp_cfg->caps,
813 sp_cfg->caps_size,
814 sp_cfg->param_id, mconfig);
815 if (ret < 0)
816 return ret;
817 }
818
819 for (i = 0; i < w->num_kcontrols; i++) {
820 k = &w->kcontrol_news[i];
821 if (k->access & SNDRV_CTL_ELEM_ACCESS_TLV_CALLBACK) {
822 sb = (void *) k->private_value;
823 bc = (struct skl_algo_data *)sb->dobj.private;
824
825 if (bc->set_params == SKL_PARAM_BIND) {
826 params = kmemdup(bc->params, bc->max, GFP_KERNEL);
827 if (!params)
828 return -ENOMEM;
829
830 skl_fill_sink_instance_id(skl, params, bc->max,
831 mconfig);
832
833 ret = skl_set_module_params(skl, params,
834 bc->max, bc->param_id, mconfig);
835 kfree(params);
836
837 if (ret < 0)
838 return ret;
839 }
840 }
841 }
842
843 return 0;
844 }
845
skl_get_module_id(struct skl_dev * skl,guid_t * uuid)846 static int skl_get_module_id(struct skl_dev *skl, guid_t *uuid)
847 {
848 struct uuid_module *module;
849
850 list_for_each_entry(module, &skl->uuid_list, list) {
851 if (guid_equal(uuid, &module->uuid))
852 return module->id;
853 }
854
855 return -EINVAL;
856 }
857
skl_tplg_find_moduleid_from_uuid(struct skl_dev * skl,const struct snd_kcontrol_new * k)858 static int skl_tplg_find_moduleid_from_uuid(struct skl_dev *skl,
859 const struct snd_kcontrol_new *k)
860 {
861 struct soc_bytes_ext *sb = (void *) k->private_value;
862 struct skl_algo_data *bc = (struct skl_algo_data *)sb->dobj.private;
863 struct skl_kpb_params *uuid_params, *params;
864 struct hdac_bus *bus = skl_to_bus(skl);
865 int i, size, module_id;
866
867 if (bc->set_params == SKL_PARAM_BIND && bc->max) {
868 uuid_params = (struct skl_kpb_params *)bc->params;
869 size = struct_size(params, u.map, uuid_params->num_modules);
870
871 params = devm_kzalloc(bus->dev, size, GFP_KERNEL);
872 if (!params)
873 return -ENOMEM;
874
875 params->num_modules = uuid_params->num_modules;
876
877 for (i = 0; i < uuid_params->num_modules; i++) {
878 module_id = skl_get_module_id(skl,
879 &uuid_params->u.map_uuid[i].mod_uuid);
880 if (module_id < 0) {
881 devm_kfree(bus->dev, params);
882 return -EINVAL;
883 }
884
885 params->u.map[i].mod_id = module_id;
886 params->u.map[i].inst_id =
887 uuid_params->u.map_uuid[i].inst_id;
888 }
889
890 devm_kfree(bus->dev, bc->params);
891 bc->params = (char *)params;
892 bc->max = size;
893 }
894
895 return 0;
896 }
897
898 /*
899 * Retrieve the module id from UUID mentioned in the
900 * post bind params
901 */
skl_tplg_add_moduleid_in_bind_params(struct skl_dev * skl,struct snd_soc_dapm_widget * w)902 void skl_tplg_add_moduleid_in_bind_params(struct skl_dev *skl,
903 struct snd_soc_dapm_widget *w)
904 {
905 struct skl_module_cfg *mconfig = w->priv;
906 int i;
907
908 /*
909 * Post bind params are used for only for KPB
910 * to set copier instances to drain the data
911 * in fast mode
912 */
913 if (mconfig->m_type != SKL_MODULE_TYPE_KPB)
914 return;
915
916 for (i = 0; i < w->num_kcontrols; i++)
917 if ((w->kcontrol_news[i].access &
918 SNDRV_CTL_ELEM_ACCESS_TLV_CALLBACK) &&
919 (skl_tplg_find_moduleid_from_uuid(skl,
920 &w->kcontrol_news[i]) < 0))
921 dev_err(skl->dev,
922 "%s: invalid kpb post bind params\n",
923 __func__);
924 }
925
skl_tplg_module_add_deferred_bind(struct skl_dev * skl,struct skl_module_cfg * src,struct skl_module_cfg * dst)926 static int skl_tplg_module_add_deferred_bind(struct skl_dev *skl,
927 struct skl_module_cfg *src, struct skl_module_cfg *dst)
928 {
929 struct skl_module_deferred_bind *m_list, *modules;
930 int i;
931
932 /* only supported for module with static pin connection */
933 for (i = 0; i < dst->module->max_input_pins; i++) {
934 struct skl_module_pin *pin = &dst->m_in_pin[i];
935
936 if (pin->is_dynamic)
937 continue;
938
939 if ((pin->id.module_id == src->id.module_id) &&
940 (pin->id.instance_id == src->id.instance_id)) {
941
942 if (!list_empty(&skl->bind_list)) {
943 list_for_each_entry(modules, &skl->bind_list, node) {
944 if (modules->src == src && modules->dst == dst)
945 return 0;
946 }
947 }
948
949 m_list = kzalloc(sizeof(*m_list), GFP_KERNEL);
950 if (!m_list)
951 return -ENOMEM;
952
953 m_list->src = src;
954 m_list->dst = dst;
955
956 list_add(&m_list->node, &skl->bind_list);
957 }
958 }
959
960 return 0;
961 }
962
skl_tplg_bind_sinks(struct snd_soc_dapm_widget * w,struct skl_dev * skl,struct snd_soc_dapm_widget * src_w,struct skl_module_cfg * src_mconfig)963 static int skl_tplg_bind_sinks(struct snd_soc_dapm_widget *w,
964 struct skl_dev *skl,
965 struct snd_soc_dapm_widget *src_w,
966 struct skl_module_cfg *src_mconfig)
967 {
968 struct snd_soc_dapm_path *p;
969 struct snd_soc_dapm_widget *sink = NULL, *next_sink = NULL;
970 struct skl_module_cfg *sink_mconfig;
971 int ret;
972
973 snd_soc_dapm_widget_for_each_sink_path(w, p) {
974 if (!p->connect)
975 continue;
976
977 dev_dbg(skl->dev,
978 "%s: src widget=%s\n", __func__, w->name);
979 dev_dbg(skl->dev,
980 "%s: sink widget=%s\n", __func__, p->sink->name);
981
982 next_sink = p->sink;
983
984 if (!is_skl_dsp_widget_type(p->sink, skl->dev))
985 return skl_tplg_bind_sinks(p->sink, skl, src_w, src_mconfig);
986
987 /*
988 * here we will check widgets in sink pipelines, so that
989 * can be any widgets type and we are only interested if
990 * they are ones used for SKL so check that first
991 */
992 if ((p->sink->priv != NULL) &&
993 is_skl_dsp_widget_type(p->sink, skl->dev)) {
994
995 sink = p->sink;
996 sink_mconfig = sink->priv;
997
998 /*
999 * Modules other than PGA leaf can be connected
1000 * directly or via switch to a module in another
1001 * pipeline. EX: reference path
1002 * when the path is enabled, the dst module that needs
1003 * to be bound may not be initialized. if the module is
1004 * not initialized, add these modules in the deferred
1005 * bind list and when the dst module is initialised,
1006 * bind this module to the dst_module in deferred list.
1007 */
1008 if (((src_mconfig->m_state == SKL_MODULE_INIT_DONE)
1009 && (sink_mconfig->m_state == SKL_MODULE_UNINIT))) {
1010
1011 ret = skl_tplg_module_add_deferred_bind(skl,
1012 src_mconfig, sink_mconfig);
1013
1014 if (ret < 0)
1015 return ret;
1016
1017 }
1018
1019
1020 if (src_mconfig->m_state == SKL_MODULE_UNINIT ||
1021 sink_mconfig->m_state == SKL_MODULE_UNINIT)
1022 continue;
1023
1024 /* Bind source to sink, mixin is always source */
1025 ret = skl_bind_modules(skl, src_mconfig, sink_mconfig);
1026 if (ret)
1027 return ret;
1028
1029 /* set module params after bind */
1030 skl_tplg_set_module_bind_params(src_w,
1031 src_mconfig, skl);
1032 skl_tplg_set_module_bind_params(sink,
1033 sink_mconfig, skl);
1034
1035 /* Start sinks pipe first */
1036 if (sink_mconfig->pipe->state != SKL_PIPE_STARTED) {
1037 if (sink_mconfig->pipe->conn_type !=
1038 SKL_PIPE_CONN_TYPE_FE)
1039 ret = skl_run_pipe(skl,
1040 sink_mconfig->pipe);
1041 if (ret)
1042 return ret;
1043 }
1044 }
1045 }
1046
1047 if (!sink && next_sink)
1048 return skl_tplg_bind_sinks(next_sink, skl, src_w, src_mconfig);
1049
1050 return 0;
1051 }
1052
1053 /*
1054 * A PGA represents a module in a pipeline. So in the Pre-PMU event of PGA
1055 * we need to do following:
1056 * - Bind to sink pipeline
1057 * Since the sink pipes can be running and we don't get mixer event on
1058 * connect for already running mixer, we need to find the sink pipes
1059 * here and bind to them. This way dynamic connect works.
1060 * - Start sink pipeline, if not running
1061 * - Then run current pipe
1062 */
skl_tplg_pga_dapm_pre_pmu_event(struct snd_soc_dapm_widget * w,struct skl_dev * skl)1063 static int skl_tplg_pga_dapm_pre_pmu_event(struct snd_soc_dapm_widget *w,
1064 struct skl_dev *skl)
1065 {
1066 struct skl_module_cfg *src_mconfig;
1067 int ret = 0;
1068
1069 src_mconfig = w->priv;
1070
1071 /*
1072 * find which sink it is connected to, bind with the sink,
1073 * if sink is not started, start sink pipe first, then start
1074 * this pipe
1075 */
1076 ret = skl_tplg_bind_sinks(w, skl, w, src_mconfig);
1077 if (ret)
1078 return ret;
1079
1080 /* Start source pipe last after starting all sinks */
1081 if (src_mconfig->pipe->conn_type != SKL_PIPE_CONN_TYPE_FE)
1082 return skl_run_pipe(skl, src_mconfig->pipe);
1083
1084 return 0;
1085 }
1086
skl_get_src_dsp_widget(struct snd_soc_dapm_widget * w,struct skl_dev * skl)1087 static struct snd_soc_dapm_widget *skl_get_src_dsp_widget(
1088 struct snd_soc_dapm_widget *w, struct skl_dev *skl)
1089 {
1090 struct snd_soc_dapm_path *p;
1091 struct snd_soc_dapm_widget *src_w = NULL;
1092
1093 snd_soc_dapm_widget_for_each_source_path(w, p) {
1094 src_w = p->source;
1095 if (!p->connect)
1096 continue;
1097
1098 dev_dbg(skl->dev, "sink widget=%s\n", w->name);
1099 dev_dbg(skl->dev, "src widget=%s\n", p->source->name);
1100
1101 /*
1102 * here we will check widgets in sink pipelines, so that can
1103 * be any widgets type and we are only interested if they are
1104 * ones used for SKL so check that first
1105 */
1106 if ((p->source->priv != NULL) &&
1107 is_skl_dsp_widget_type(p->source, skl->dev)) {
1108 return p->source;
1109 }
1110 }
1111
1112 if (src_w != NULL)
1113 return skl_get_src_dsp_widget(src_w, skl);
1114
1115 return NULL;
1116 }
1117
1118 /*
1119 * in the Post-PMU event of mixer we need to do following:
1120 * - Check if this pipe is running
1121 * - if not, then
1122 * - bind this pipeline to its source pipeline
1123 * if source pipe is already running, this means it is a dynamic
1124 * connection and we need to bind only to that pipe
1125 * - start this pipeline
1126 */
skl_tplg_mixer_dapm_post_pmu_event(struct snd_soc_dapm_widget * w,struct skl_dev * skl)1127 static int skl_tplg_mixer_dapm_post_pmu_event(struct snd_soc_dapm_widget *w,
1128 struct skl_dev *skl)
1129 {
1130 int ret = 0;
1131 struct snd_soc_dapm_widget *source, *sink;
1132 struct skl_module_cfg *src_mconfig, *sink_mconfig;
1133 int src_pipe_started = 0;
1134
1135 sink = w;
1136 sink_mconfig = sink->priv;
1137
1138 /*
1139 * If source pipe is already started, that means source is driving
1140 * one more sink before this sink got connected, Since source is
1141 * started, bind this sink to source and start this pipe.
1142 */
1143 source = skl_get_src_dsp_widget(w, skl);
1144 if (source != NULL) {
1145 src_mconfig = source->priv;
1146 sink_mconfig = sink->priv;
1147 src_pipe_started = 1;
1148
1149 /*
1150 * check pipe state, then no need to bind or start the
1151 * pipe
1152 */
1153 if (src_mconfig->pipe->state != SKL_PIPE_STARTED)
1154 src_pipe_started = 0;
1155 }
1156
1157 if (src_pipe_started) {
1158 ret = skl_bind_modules(skl, src_mconfig, sink_mconfig);
1159 if (ret)
1160 return ret;
1161
1162 /* set module params after bind */
1163 skl_tplg_set_module_bind_params(source, src_mconfig, skl);
1164 skl_tplg_set_module_bind_params(sink, sink_mconfig, skl);
1165
1166 if (sink_mconfig->pipe->conn_type != SKL_PIPE_CONN_TYPE_FE)
1167 ret = skl_run_pipe(skl, sink_mconfig->pipe);
1168 }
1169
1170 return ret;
1171 }
1172
1173 /*
1174 * in the Pre-PMD event of mixer we need to do following:
1175 * - Stop the pipe
1176 * - find the source connections and remove that from dapm_path_list
1177 * - unbind with source pipelines if still connected
1178 */
skl_tplg_mixer_dapm_pre_pmd_event(struct snd_soc_dapm_widget * w,struct skl_dev * skl)1179 static int skl_tplg_mixer_dapm_pre_pmd_event(struct snd_soc_dapm_widget *w,
1180 struct skl_dev *skl)
1181 {
1182 struct skl_module_cfg *src_mconfig, *sink_mconfig;
1183 int ret = 0, i;
1184
1185 sink_mconfig = w->priv;
1186
1187 /* Stop the pipe */
1188 ret = skl_stop_pipe(skl, sink_mconfig->pipe);
1189 if (ret)
1190 return ret;
1191
1192 for (i = 0; i < sink_mconfig->module->max_input_pins; i++) {
1193 if (sink_mconfig->m_in_pin[i].pin_state == SKL_PIN_BIND_DONE) {
1194 src_mconfig = sink_mconfig->m_in_pin[i].tgt_mcfg;
1195 if (!src_mconfig)
1196 continue;
1197
1198 ret = skl_unbind_modules(skl,
1199 src_mconfig, sink_mconfig);
1200 }
1201 }
1202
1203 return ret;
1204 }
1205
1206 /*
1207 * in the Post-PMD event of mixer we need to do following:
1208 * - Unbind the modules within the pipeline
1209 * - Delete the pipeline (modules are not required to be explicitly
1210 * deleted, pipeline delete is enough here
1211 */
skl_tplg_mixer_dapm_post_pmd_event(struct snd_soc_dapm_widget * w,struct skl_dev * skl)1212 static int skl_tplg_mixer_dapm_post_pmd_event(struct snd_soc_dapm_widget *w,
1213 struct skl_dev *skl)
1214 {
1215 struct skl_module_cfg *mconfig = w->priv;
1216 struct skl_pipe_module *w_module;
1217 struct skl_module_cfg *src_module = NULL, *dst_module;
1218 struct skl_pipe *s_pipe = mconfig->pipe;
1219 struct skl_module_deferred_bind *modules, *tmp;
1220
1221 if (s_pipe->state == SKL_PIPE_INVALID)
1222 return -EINVAL;
1223
1224 list_for_each_entry(w_module, &s_pipe->w_list, node) {
1225 if (list_empty(&skl->bind_list))
1226 break;
1227
1228 src_module = w_module->w->priv;
1229
1230 list_for_each_entry_safe(modules, tmp, &skl->bind_list, node) {
1231 /*
1232 * When the destination module is deleted, Unbind the
1233 * modules from deferred bind list.
1234 */
1235 if (modules->dst == src_module) {
1236 skl_unbind_modules(skl, modules->src,
1237 modules->dst);
1238 }
1239
1240 /*
1241 * When the source module is deleted, remove this entry
1242 * from the deferred bind list.
1243 */
1244 if (modules->src == src_module) {
1245 list_del(&modules->node);
1246 modules->src = NULL;
1247 modules->dst = NULL;
1248 kfree(modules);
1249 }
1250 }
1251 }
1252
1253 list_for_each_entry(w_module, &s_pipe->w_list, node) {
1254 dst_module = w_module->w->priv;
1255
1256 if (src_module == NULL) {
1257 src_module = dst_module;
1258 continue;
1259 }
1260
1261 skl_unbind_modules(skl, src_module, dst_module);
1262 src_module = dst_module;
1263 }
1264
1265 skl_delete_pipe(skl, mconfig->pipe);
1266
1267 list_for_each_entry(w_module, &s_pipe->w_list, node) {
1268 src_module = w_module->w->priv;
1269 src_module->m_state = SKL_MODULE_UNINIT;
1270 }
1271
1272 return skl_tplg_unload_pipe_modules(skl, s_pipe);
1273 }
1274
1275 /*
1276 * in the Post-PMD event of PGA we need to do following:
1277 * - Stop the pipeline
1278 * - In source pipe is connected, unbind with source pipelines
1279 */
skl_tplg_pga_dapm_post_pmd_event(struct snd_soc_dapm_widget * w,struct skl_dev * skl)1280 static int skl_tplg_pga_dapm_post_pmd_event(struct snd_soc_dapm_widget *w,
1281 struct skl_dev *skl)
1282 {
1283 struct skl_module_cfg *src_mconfig, *sink_mconfig;
1284 int ret = 0, i;
1285
1286 src_mconfig = w->priv;
1287
1288 /* Stop the pipe since this is a mixin module */
1289 ret = skl_stop_pipe(skl, src_mconfig->pipe);
1290 if (ret)
1291 return ret;
1292
1293 for (i = 0; i < src_mconfig->module->max_output_pins; i++) {
1294 if (src_mconfig->m_out_pin[i].pin_state == SKL_PIN_BIND_DONE) {
1295 sink_mconfig = src_mconfig->m_out_pin[i].tgt_mcfg;
1296 if (!sink_mconfig)
1297 continue;
1298 /*
1299 * This is a connecter and if path is found that means
1300 * unbind between source and sink has not happened yet
1301 */
1302 ret = skl_unbind_modules(skl, src_mconfig,
1303 sink_mconfig);
1304 }
1305 }
1306
1307 return ret;
1308 }
1309
1310 /*
1311 * In modelling, we assume there will be ONLY one mixer in a pipeline. If a
1312 * second one is required that is created as another pipe entity.
1313 * The mixer is responsible for pipe management and represent a pipeline
1314 * instance
1315 */
skl_tplg_mixer_event(struct snd_soc_dapm_widget * w,struct snd_kcontrol * k,int event)1316 static int skl_tplg_mixer_event(struct snd_soc_dapm_widget *w,
1317 struct snd_kcontrol *k, int event)
1318 {
1319 struct snd_soc_dapm_context *dapm = w->dapm;
1320 struct skl_dev *skl = get_skl_ctx(dapm->dev);
1321
1322 switch (event) {
1323 case SND_SOC_DAPM_PRE_PMU:
1324 return skl_tplg_mixer_dapm_pre_pmu_event(w, skl);
1325
1326 case SND_SOC_DAPM_POST_PMU:
1327 return skl_tplg_mixer_dapm_post_pmu_event(w, skl);
1328
1329 case SND_SOC_DAPM_PRE_PMD:
1330 return skl_tplg_mixer_dapm_pre_pmd_event(w, skl);
1331
1332 case SND_SOC_DAPM_POST_PMD:
1333 return skl_tplg_mixer_dapm_post_pmd_event(w, skl);
1334 }
1335
1336 return 0;
1337 }
1338
1339 /*
1340 * In modelling, we assumed rest of the modules in pipeline are PGA. But we
1341 * are interested in last PGA (leaf PGA) in a pipeline to disconnect with
1342 * the sink when it is running (two FE to one BE or one FE to two BE)
1343 * scenarios
1344 */
skl_tplg_pga_event(struct snd_soc_dapm_widget * w,struct snd_kcontrol * k,int event)1345 static int skl_tplg_pga_event(struct snd_soc_dapm_widget *w,
1346 struct snd_kcontrol *k, int event)
1347
1348 {
1349 struct snd_soc_dapm_context *dapm = w->dapm;
1350 struct skl_dev *skl = get_skl_ctx(dapm->dev);
1351
1352 switch (event) {
1353 case SND_SOC_DAPM_PRE_PMU:
1354 return skl_tplg_pga_dapm_pre_pmu_event(w, skl);
1355
1356 case SND_SOC_DAPM_POST_PMD:
1357 return skl_tplg_pga_dapm_post_pmd_event(w, skl);
1358 }
1359
1360 return 0;
1361 }
1362
skl_tplg_multi_config_set_get(struct snd_kcontrol * kcontrol,struct snd_ctl_elem_value * ucontrol,bool is_set)1363 static int skl_tplg_multi_config_set_get(struct snd_kcontrol *kcontrol,
1364 struct snd_ctl_elem_value *ucontrol,
1365 bool is_set)
1366 {
1367 struct snd_soc_component *component =
1368 snd_soc_kcontrol_component(kcontrol);
1369 struct hdac_bus *bus = snd_soc_component_get_drvdata(component);
1370 struct skl_dev *skl = bus_to_skl(bus);
1371 struct skl_pipeline *ppl;
1372 struct skl_pipe *pipe = NULL;
1373 struct soc_enum *ec = (struct soc_enum *)kcontrol->private_value;
1374 u32 *pipe_id;
1375
1376 if (!ec)
1377 return -EINVAL;
1378
1379 if (is_set && ucontrol->value.enumerated.item[0] > ec->items)
1380 return -EINVAL;
1381
1382 pipe_id = ec->dobj.private;
1383
1384 list_for_each_entry(ppl, &skl->ppl_list, node) {
1385 if (ppl->pipe->ppl_id == *pipe_id) {
1386 pipe = ppl->pipe;
1387 break;
1388 }
1389 }
1390 if (!pipe)
1391 return -EIO;
1392
1393 if (is_set)
1394 pipe->pipe_config_idx = ucontrol->value.enumerated.item[0];
1395 else
1396 ucontrol->value.enumerated.item[0] = pipe->pipe_config_idx;
1397
1398 return 0;
1399 }
1400
skl_tplg_multi_config_get(struct snd_kcontrol * kcontrol,struct snd_ctl_elem_value * ucontrol)1401 static int skl_tplg_multi_config_get(struct snd_kcontrol *kcontrol,
1402 struct snd_ctl_elem_value *ucontrol)
1403 {
1404 return skl_tplg_multi_config_set_get(kcontrol, ucontrol, false);
1405 }
1406
skl_tplg_multi_config_set(struct snd_kcontrol * kcontrol,struct snd_ctl_elem_value * ucontrol)1407 static int skl_tplg_multi_config_set(struct snd_kcontrol *kcontrol,
1408 struct snd_ctl_elem_value *ucontrol)
1409 {
1410 return skl_tplg_multi_config_set_get(kcontrol, ucontrol, true);
1411 }
1412
skl_tplg_multi_config_get_dmic(struct snd_kcontrol * kcontrol,struct snd_ctl_elem_value * ucontrol)1413 static int skl_tplg_multi_config_get_dmic(struct snd_kcontrol *kcontrol,
1414 struct snd_ctl_elem_value *ucontrol)
1415 {
1416 return skl_tplg_multi_config_set_get(kcontrol, ucontrol, false);
1417 }
1418
skl_tplg_multi_config_set_dmic(struct snd_kcontrol * kcontrol,struct snd_ctl_elem_value * ucontrol)1419 static int skl_tplg_multi_config_set_dmic(struct snd_kcontrol *kcontrol,
1420 struct snd_ctl_elem_value *ucontrol)
1421 {
1422 return skl_tplg_multi_config_set_get(kcontrol, ucontrol, true);
1423 }
1424
skl_tplg_tlv_control_get(struct snd_kcontrol * kcontrol,unsigned int __user * data,unsigned int size)1425 static int skl_tplg_tlv_control_get(struct snd_kcontrol *kcontrol,
1426 unsigned int __user *data, unsigned int size)
1427 {
1428 struct soc_bytes_ext *sb =
1429 (struct soc_bytes_ext *)kcontrol->private_value;
1430 struct skl_algo_data *bc = (struct skl_algo_data *)sb->dobj.private;
1431 struct snd_soc_dapm_widget *w = snd_soc_dapm_kcontrol_widget(kcontrol);
1432 struct skl_module_cfg *mconfig = w->priv;
1433 struct skl_dev *skl = get_skl_ctx(w->dapm->dev);
1434
1435 if (w->power)
1436 skl_get_module_params(skl, (u32 *)bc->params,
1437 bc->size, bc->param_id, mconfig);
1438
1439 /* decrement size for TLV header */
1440 size -= 2 * sizeof(u32);
1441
1442 /* check size as we don't want to send kernel data */
1443 if (size > bc->max)
1444 size = bc->max;
1445
1446 if (bc->params) {
1447 if (copy_to_user(data, &bc->param_id, sizeof(u32)))
1448 return -EFAULT;
1449 if (copy_to_user(data + 1, &size, sizeof(u32)))
1450 return -EFAULT;
1451 if (copy_to_user(data + 2, bc->params, size))
1452 return -EFAULT;
1453 }
1454
1455 return 0;
1456 }
1457
1458 #define SKL_PARAM_VENDOR_ID 0xff
1459
skl_tplg_tlv_control_set(struct snd_kcontrol * kcontrol,const unsigned int __user * data,unsigned int size)1460 static int skl_tplg_tlv_control_set(struct snd_kcontrol *kcontrol,
1461 const unsigned int __user *data, unsigned int size)
1462 {
1463 struct snd_soc_dapm_widget *w = snd_soc_dapm_kcontrol_widget(kcontrol);
1464 struct skl_module_cfg *mconfig = w->priv;
1465 struct soc_bytes_ext *sb =
1466 (struct soc_bytes_ext *)kcontrol->private_value;
1467 struct skl_algo_data *ac = (struct skl_algo_data *)sb->dobj.private;
1468 struct skl_dev *skl = get_skl_ctx(w->dapm->dev);
1469
1470 if (ac->params) {
1471 if (size > ac->max)
1472 return -EINVAL;
1473 ac->size = size;
1474
1475 if (copy_from_user(ac->params, data, size))
1476 return -EFAULT;
1477
1478 if (w->power)
1479 return skl_set_module_params(skl,
1480 (u32 *)ac->params, ac->size,
1481 ac->param_id, mconfig);
1482 }
1483
1484 return 0;
1485 }
1486
skl_tplg_mic_control_get(struct snd_kcontrol * kcontrol,struct snd_ctl_elem_value * ucontrol)1487 static int skl_tplg_mic_control_get(struct snd_kcontrol *kcontrol,
1488 struct snd_ctl_elem_value *ucontrol)
1489 {
1490 struct snd_soc_dapm_widget *w = snd_soc_dapm_kcontrol_widget(kcontrol);
1491 struct skl_module_cfg *mconfig = w->priv;
1492 struct soc_enum *ec = (struct soc_enum *)kcontrol->private_value;
1493 u32 ch_type = *((u32 *)ec->dobj.private);
1494
1495 if (mconfig->dmic_ch_type == ch_type)
1496 ucontrol->value.enumerated.item[0] =
1497 mconfig->dmic_ch_combo_index;
1498 else
1499 ucontrol->value.enumerated.item[0] = 0;
1500
1501 return 0;
1502 }
1503
skl_fill_mic_sel_params(struct skl_module_cfg * mconfig,struct skl_mic_sel_config * mic_cfg,struct device * dev)1504 static int skl_fill_mic_sel_params(struct skl_module_cfg *mconfig,
1505 struct skl_mic_sel_config *mic_cfg, struct device *dev)
1506 {
1507 struct skl_specific_cfg *sp_cfg =
1508 &mconfig->formats_config[SKL_PARAM_INIT];
1509
1510 sp_cfg->caps_size = sizeof(struct skl_mic_sel_config);
1511 sp_cfg->set_params = SKL_PARAM_SET;
1512 sp_cfg->param_id = 0x00;
1513 if (!sp_cfg->caps) {
1514 sp_cfg->caps = devm_kzalloc(dev, sp_cfg->caps_size, GFP_KERNEL);
1515 if (!sp_cfg->caps)
1516 return -ENOMEM;
1517 }
1518
1519 mic_cfg->mic_switch = SKL_MIC_SEL_SWITCH;
1520 mic_cfg->flags = 0;
1521 memcpy(sp_cfg->caps, mic_cfg, sp_cfg->caps_size);
1522
1523 return 0;
1524 }
1525
skl_tplg_mic_control_set(struct snd_kcontrol * kcontrol,struct snd_ctl_elem_value * ucontrol)1526 static int skl_tplg_mic_control_set(struct snd_kcontrol *kcontrol,
1527 struct snd_ctl_elem_value *ucontrol)
1528 {
1529 struct snd_soc_dapm_widget *w = snd_soc_dapm_kcontrol_widget(kcontrol);
1530 struct skl_module_cfg *mconfig = w->priv;
1531 struct skl_mic_sel_config mic_cfg = {0};
1532 struct soc_enum *ec = (struct soc_enum *)kcontrol->private_value;
1533 u32 ch_type = *((u32 *)ec->dobj.private);
1534 const int *list;
1535 u8 in_ch, out_ch, index;
1536
1537 mconfig->dmic_ch_type = ch_type;
1538 mconfig->dmic_ch_combo_index = ucontrol->value.enumerated.item[0];
1539
1540 /* enum control index 0 is INVALID, so no channels to be set */
1541 if (mconfig->dmic_ch_combo_index == 0)
1542 return 0;
1543
1544 /* No valid channel selection map for index 0, so offset by 1 */
1545 index = mconfig->dmic_ch_combo_index - 1;
1546
1547 switch (ch_type) {
1548 case SKL_CH_MONO:
1549 if (mconfig->dmic_ch_combo_index > ARRAY_SIZE(mic_mono_list))
1550 return -EINVAL;
1551
1552 list = &mic_mono_list[index];
1553 break;
1554
1555 case SKL_CH_STEREO:
1556 if (mconfig->dmic_ch_combo_index > ARRAY_SIZE(mic_stereo_list))
1557 return -EINVAL;
1558
1559 list = mic_stereo_list[index];
1560 break;
1561
1562 case SKL_CH_TRIO:
1563 if (mconfig->dmic_ch_combo_index > ARRAY_SIZE(mic_trio_list))
1564 return -EINVAL;
1565
1566 list = mic_trio_list[index];
1567 break;
1568
1569 case SKL_CH_QUATRO:
1570 if (mconfig->dmic_ch_combo_index > ARRAY_SIZE(mic_quatro_list))
1571 return -EINVAL;
1572
1573 list = mic_quatro_list[index];
1574 break;
1575
1576 default:
1577 dev_err(w->dapm->dev,
1578 "Invalid channel %d for mic_select module\n",
1579 ch_type);
1580 return -EINVAL;
1581
1582 }
1583
1584 /* channel type enum map to number of chanels for that type */
1585 for (out_ch = 0; out_ch < ch_type; out_ch++) {
1586 in_ch = list[out_ch];
1587 mic_cfg.blob[out_ch][in_ch] = SKL_DEFAULT_MIC_SEL_GAIN;
1588 }
1589
1590 return skl_fill_mic_sel_params(mconfig, &mic_cfg, w->dapm->dev);
1591 }
1592
1593 /*
1594 * Fill the dma id for host and link. In case of passthrough
1595 * pipeline, this will both host and link in the same
1596 * pipeline, so need to copy the link and host based on dev_type
1597 */
skl_tplg_fill_dma_id(struct skl_module_cfg * mcfg,struct skl_pipe_params * params)1598 static void skl_tplg_fill_dma_id(struct skl_module_cfg *mcfg,
1599 struct skl_pipe_params *params)
1600 {
1601 struct skl_pipe *pipe = mcfg->pipe;
1602
1603 if (pipe->passthru) {
1604 switch (mcfg->dev_type) {
1605 case SKL_DEVICE_HDALINK:
1606 pipe->p_params->link_dma_id = params->link_dma_id;
1607 pipe->p_params->link_index = params->link_index;
1608 pipe->p_params->link_bps = params->link_bps;
1609 break;
1610
1611 case SKL_DEVICE_HDAHOST:
1612 pipe->p_params->host_dma_id = params->host_dma_id;
1613 pipe->p_params->host_bps = params->host_bps;
1614 break;
1615
1616 default:
1617 break;
1618 }
1619 pipe->p_params->s_fmt = params->s_fmt;
1620 pipe->p_params->ch = params->ch;
1621 pipe->p_params->s_freq = params->s_freq;
1622 pipe->p_params->stream = params->stream;
1623 pipe->p_params->format = params->format;
1624
1625 } else {
1626 memcpy(pipe->p_params, params, sizeof(*params));
1627 }
1628 }
1629
1630 /*
1631 * The FE params are passed by hw_params of the DAI.
1632 * On hw_params, the params are stored in Gateway module of the FE and we
1633 * need to calculate the format in DSP module configuration, that
1634 * conversion is done here
1635 */
skl_tplg_update_pipe_params(struct device * dev,struct skl_module_cfg * mconfig,struct skl_pipe_params * params)1636 int skl_tplg_update_pipe_params(struct device *dev,
1637 struct skl_module_cfg *mconfig,
1638 struct skl_pipe_params *params)
1639 {
1640 struct skl_module_res *res;
1641 struct skl_dev *skl = get_skl_ctx(dev);
1642 struct skl_module_fmt *format = NULL;
1643 u8 cfg_idx = mconfig->pipe->cur_config_idx;
1644
1645 res = &mconfig->module->resources[mconfig->res_idx];
1646 skl_tplg_fill_dma_id(mconfig, params);
1647 mconfig->fmt_idx = mconfig->mod_cfg[cfg_idx].fmt_idx;
1648 mconfig->res_idx = mconfig->mod_cfg[cfg_idx].res_idx;
1649
1650 if (skl->nr_modules)
1651 return 0;
1652
1653 if (params->stream == SNDRV_PCM_STREAM_PLAYBACK)
1654 format = &mconfig->module->formats[mconfig->fmt_idx].inputs[0].fmt;
1655 else
1656 format = &mconfig->module->formats[mconfig->fmt_idx].outputs[0].fmt;
1657
1658 /* set the hw_params */
1659 format->s_freq = params->s_freq;
1660 format->channels = params->ch;
1661 format->valid_bit_depth = skl_get_bit_depth(params->s_fmt);
1662
1663 /*
1664 * 16 bit is 16 bit container whereas 24 bit is in 32 bit
1665 * container so update bit depth accordingly
1666 */
1667 switch (format->valid_bit_depth) {
1668 case SKL_DEPTH_16BIT:
1669 format->bit_depth = format->valid_bit_depth;
1670 break;
1671
1672 case SKL_DEPTH_24BIT:
1673 case SKL_DEPTH_32BIT:
1674 format->bit_depth = SKL_DEPTH_32BIT;
1675 break;
1676
1677 default:
1678 dev_err(dev, "Invalid bit depth %x for pipe\n",
1679 format->valid_bit_depth);
1680 return -EINVAL;
1681 }
1682
1683 if (params->stream == SNDRV_PCM_STREAM_PLAYBACK) {
1684 res->ibs = (format->s_freq / 1000) *
1685 (format->channels) *
1686 (format->bit_depth >> 3);
1687 } else {
1688 res->obs = (format->s_freq / 1000) *
1689 (format->channels) *
1690 (format->bit_depth >> 3);
1691 }
1692
1693 return 0;
1694 }
1695
1696 /*
1697 * Query the module config for the FE DAI
1698 * This is used to find the hw_params set for that DAI and apply to FE
1699 * pipeline
1700 */
1701 struct skl_module_cfg *
skl_tplg_fe_get_cpr_module(struct snd_soc_dai * dai,int stream)1702 skl_tplg_fe_get_cpr_module(struct snd_soc_dai *dai, int stream)
1703 {
1704 struct snd_soc_dapm_widget *w;
1705 struct snd_soc_dapm_path *p = NULL;
1706
1707 if (stream == SNDRV_PCM_STREAM_PLAYBACK) {
1708 w = dai->playback_widget;
1709 snd_soc_dapm_widget_for_each_sink_path(w, p) {
1710 if (p->connect && p->sink->power &&
1711 !is_skl_dsp_widget_type(p->sink, dai->dev))
1712 continue;
1713
1714 if (p->sink->priv) {
1715 dev_dbg(dai->dev, "set params for %s\n",
1716 p->sink->name);
1717 return p->sink->priv;
1718 }
1719 }
1720 } else {
1721 w = dai->capture_widget;
1722 snd_soc_dapm_widget_for_each_source_path(w, p) {
1723 if (p->connect && p->source->power &&
1724 !is_skl_dsp_widget_type(p->source, dai->dev))
1725 continue;
1726
1727 if (p->source->priv) {
1728 dev_dbg(dai->dev, "set params for %s\n",
1729 p->source->name);
1730 return p->source->priv;
1731 }
1732 }
1733 }
1734
1735 return NULL;
1736 }
1737
skl_get_mconfig_pb_cpr(struct snd_soc_dai * dai,struct snd_soc_dapm_widget * w)1738 static struct skl_module_cfg *skl_get_mconfig_pb_cpr(
1739 struct snd_soc_dai *dai, struct snd_soc_dapm_widget *w)
1740 {
1741 struct snd_soc_dapm_path *p;
1742 struct skl_module_cfg *mconfig = NULL;
1743
1744 snd_soc_dapm_widget_for_each_source_path(w, p) {
1745 if (w->endpoints[SND_SOC_DAPM_DIR_OUT] > 0) {
1746 if (p->connect &&
1747 (p->sink->id == snd_soc_dapm_aif_out) &&
1748 p->source->priv) {
1749 mconfig = p->source->priv;
1750 return mconfig;
1751 }
1752 mconfig = skl_get_mconfig_pb_cpr(dai, p->source);
1753 if (mconfig)
1754 return mconfig;
1755 }
1756 }
1757 return mconfig;
1758 }
1759
skl_get_mconfig_cap_cpr(struct snd_soc_dai * dai,struct snd_soc_dapm_widget * w)1760 static struct skl_module_cfg *skl_get_mconfig_cap_cpr(
1761 struct snd_soc_dai *dai, struct snd_soc_dapm_widget *w)
1762 {
1763 struct snd_soc_dapm_path *p;
1764 struct skl_module_cfg *mconfig = NULL;
1765
1766 snd_soc_dapm_widget_for_each_sink_path(w, p) {
1767 if (w->endpoints[SND_SOC_DAPM_DIR_IN] > 0) {
1768 if (p->connect &&
1769 (p->source->id == snd_soc_dapm_aif_in) &&
1770 p->sink->priv) {
1771 mconfig = p->sink->priv;
1772 return mconfig;
1773 }
1774 mconfig = skl_get_mconfig_cap_cpr(dai, p->sink);
1775 if (mconfig)
1776 return mconfig;
1777 }
1778 }
1779 return mconfig;
1780 }
1781
1782 struct skl_module_cfg *
skl_tplg_be_get_cpr_module(struct snd_soc_dai * dai,int stream)1783 skl_tplg_be_get_cpr_module(struct snd_soc_dai *dai, int stream)
1784 {
1785 struct snd_soc_dapm_widget *w;
1786 struct skl_module_cfg *mconfig;
1787
1788 if (stream == SNDRV_PCM_STREAM_PLAYBACK) {
1789 w = dai->playback_widget;
1790 mconfig = skl_get_mconfig_pb_cpr(dai, w);
1791 } else {
1792 w = dai->capture_widget;
1793 mconfig = skl_get_mconfig_cap_cpr(dai, w);
1794 }
1795 return mconfig;
1796 }
1797
skl_tplg_be_link_type(int dev_type)1798 static u8 skl_tplg_be_link_type(int dev_type)
1799 {
1800 int ret;
1801
1802 switch (dev_type) {
1803 case SKL_DEVICE_BT:
1804 ret = NHLT_LINK_SSP;
1805 break;
1806
1807 case SKL_DEVICE_DMIC:
1808 ret = NHLT_LINK_DMIC;
1809 break;
1810
1811 case SKL_DEVICE_I2S:
1812 ret = NHLT_LINK_SSP;
1813 break;
1814
1815 case SKL_DEVICE_HDALINK:
1816 ret = NHLT_LINK_HDA;
1817 break;
1818
1819 default:
1820 ret = NHLT_LINK_INVALID;
1821 break;
1822 }
1823
1824 return ret;
1825 }
1826
1827 /*
1828 * Fill the BE gateway parameters
1829 * The BE gateway expects a blob of parameters which are kept in the ACPI
1830 * NHLT blob, so query the blob for interface type (i2s/pdm) and instance.
1831 * The port can have multiple settings so pick based on the pipeline
1832 * parameters
1833 */
skl_tplg_be_fill_pipe_params(struct snd_soc_dai * dai,struct skl_module_cfg * mconfig,struct skl_pipe_params * params)1834 static int skl_tplg_be_fill_pipe_params(struct snd_soc_dai *dai,
1835 struct skl_module_cfg *mconfig,
1836 struct skl_pipe_params *params)
1837 {
1838 struct nhlt_specific_cfg *cfg;
1839 struct skl_pipe *pipe = mconfig->pipe;
1840 struct skl_pipe_fmt *pipe_fmt;
1841 struct skl_dev *skl = get_skl_ctx(dai->dev);
1842 int link_type = skl_tplg_be_link_type(mconfig->dev_type);
1843 u8 dev_type = skl_tplg_be_dev_type(mconfig->dev_type);
1844
1845 skl_tplg_fill_dma_id(mconfig, params);
1846
1847 if (link_type == NHLT_LINK_HDA)
1848 return 0;
1849
1850 if (pipe->direction == SNDRV_PCM_STREAM_PLAYBACK)
1851 pipe_fmt = &pipe->configs[pipe->pipe_config_idx].out_fmt;
1852 else
1853 pipe_fmt = &pipe->configs[pipe->pipe_config_idx].in_fmt;
1854
1855 /* update the blob based on virtual bus_id*/
1856 cfg = intel_nhlt_get_endpoint_blob(dai->dev, skl->nhlt,
1857 mconfig->vbus_id, link_type,
1858 pipe_fmt->bps, params->s_cont,
1859 pipe_fmt->channels, pipe_fmt->freq,
1860 pipe->direction, dev_type);
1861 if (cfg) {
1862 mconfig->formats_config[SKL_PARAM_INIT].caps_size = cfg->size;
1863 mconfig->formats_config[SKL_PARAM_INIT].caps = (u32 *)&cfg->caps;
1864 } else {
1865 dev_err(dai->dev, "Blob NULL for id:%d type:%d dirn:%d ch:%d, freq:%d, fmt:%d\n",
1866 mconfig->vbus_id, link_type, params->stream,
1867 params->ch, params->s_freq, params->s_fmt);
1868 return -EINVAL;
1869 }
1870
1871 return 0;
1872 }
1873
skl_tplg_be_set_src_pipe_params(struct snd_soc_dai * dai,struct snd_soc_dapm_widget * w,struct skl_pipe_params * params)1874 static int skl_tplg_be_set_src_pipe_params(struct snd_soc_dai *dai,
1875 struct snd_soc_dapm_widget *w,
1876 struct skl_pipe_params *params)
1877 {
1878 struct snd_soc_dapm_path *p;
1879 int ret = -EIO;
1880
1881 snd_soc_dapm_widget_for_each_source_path(w, p) {
1882 if (p->connect && is_skl_dsp_widget_type(p->source, dai->dev) &&
1883 p->source->priv) {
1884
1885 ret = skl_tplg_be_fill_pipe_params(dai,
1886 p->source->priv, params);
1887 if (ret < 0)
1888 return ret;
1889 } else {
1890 ret = skl_tplg_be_set_src_pipe_params(dai,
1891 p->source, params);
1892 if (ret < 0)
1893 return ret;
1894 }
1895 }
1896
1897 return ret;
1898 }
1899
skl_tplg_be_set_sink_pipe_params(struct snd_soc_dai * dai,struct snd_soc_dapm_widget * w,struct skl_pipe_params * params)1900 static int skl_tplg_be_set_sink_pipe_params(struct snd_soc_dai *dai,
1901 struct snd_soc_dapm_widget *w, struct skl_pipe_params *params)
1902 {
1903 struct snd_soc_dapm_path *p;
1904 int ret = -EIO;
1905
1906 snd_soc_dapm_widget_for_each_sink_path(w, p) {
1907 if (p->connect && is_skl_dsp_widget_type(p->sink, dai->dev) &&
1908 p->sink->priv) {
1909
1910 ret = skl_tplg_be_fill_pipe_params(dai,
1911 p->sink->priv, params);
1912 if (ret < 0)
1913 return ret;
1914 } else {
1915 ret = skl_tplg_be_set_sink_pipe_params(
1916 dai, p->sink, params);
1917 if (ret < 0)
1918 return ret;
1919 }
1920 }
1921
1922 return ret;
1923 }
1924
1925 /*
1926 * BE hw_params can be a source parameters (capture) or sink parameters
1927 * (playback). Based on sink and source we need to either find the source
1928 * list or the sink list and set the pipeline parameters
1929 */
skl_tplg_be_update_params(struct snd_soc_dai * dai,struct skl_pipe_params * params)1930 int skl_tplg_be_update_params(struct snd_soc_dai *dai,
1931 struct skl_pipe_params *params)
1932 {
1933 struct snd_soc_dapm_widget *w;
1934
1935 if (params->stream == SNDRV_PCM_STREAM_PLAYBACK) {
1936 w = dai->playback_widget;
1937
1938 return skl_tplg_be_set_src_pipe_params(dai, w, params);
1939
1940 } else {
1941 w = dai->capture_widget;
1942
1943 return skl_tplg_be_set_sink_pipe_params(dai, w, params);
1944 }
1945
1946 return 0;
1947 }
1948
1949 static const struct snd_soc_tplg_widget_events skl_tplg_widget_ops[] = {
1950 {SKL_MIXER_EVENT, skl_tplg_mixer_event},
1951 {SKL_VMIXER_EVENT, skl_tplg_mixer_event},
1952 {SKL_PGA_EVENT, skl_tplg_pga_event},
1953 };
1954
1955 static const struct snd_soc_tplg_bytes_ext_ops skl_tlv_ops[] = {
1956 {SKL_CONTROL_TYPE_BYTE_TLV, skl_tplg_tlv_control_get,
1957 skl_tplg_tlv_control_set},
1958 };
1959
1960 static const struct snd_soc_tplg_kcontrol_ops skl_tplg_kcontrol_ops[] = {
1961 {
1962 .id = SKL_CONTROL_TYPE_MIC_SELECT,
1963 .get = skl_tplg_mic_control_get,
1964 .put = skl_tplg_mic_control_set,
1965 },
1966 {
1967 .id = SKL_CONTROL_TYPE_MULTI_IO_SELECT,
1968 .get = skl_tplg_multi_config_get,
1969 .put = skl_tplg_multi_config_set,
1970 },
1971 {
1972 .id = SKL_CONTROL_TYPE_MULTI_IO_SELECT_DMIC,
1973 .get = skl_tplg_multi_config_get_dmic,
1974 .put = skl_tplg_multi_config_set_dmic,
1975 }
1976 };
1977
skl_tplg_fill_pipe_cfg(struct device * dev,struct skl_pipe * pipe,u32 tkn,u32 tkn_val,int conf_idx,int dir)1978 static int skl_tplg_fill_pipe_cfg(struct device *dev,
1979 struct skl_pipe *pipe, u32 tkn,
1980 u32 tkn_val, int conf_idx, int dir)
1981 {
1982 struct skl_pipe_fmt *fmt;
1983 struct skl_path_config *config;
1984
1985 switch (dir) {
1986 case SKL_DIR_IN:
1987 fmt = &pipe->configs[conf_idx].in_fmt;
1988 break;
1989
1990 case SKL_DIR_OUT:
1991 fmt = &pipe->configs[conf_idx].out_fmt;
1992 break;
1993
1994 default:
1995 dev_err(dev, "Invalid direction: %d\n", dir);
1996 return -EINVAL;
1997 }
1998
1999 config = &pipe->configs[conf_idx];
2000
2001 switch (tkn) {
2002 case SKL_TKN_U32_CFG_FREQ:
2003 fmt->freq = tkn_val;
2004 break;
2005
2006 case SKL_TKN_U8_CFG_CHAN:
2007 fmt->channels = tkn_val;
2008 break;
2009
2010 case SKL_TKN_U8_CFG_BPS:
2011 fmt->bps = tkn_val;
2012 break;
2013
2014 case SKL_TKN_U32_PATH_MEM_PGS:
2015 config->mem_pages = tkn_val;
2016 break;
2017
2018 default:
2019 dev_err(dev, "Invalid token config: %d\n", tkn);
2020 return -EINVAL;
2021 }
2022
2023 return 0;
2024 }
2025
skl_tplg_fill_pipe_tkn(struct device * dev,struct skl_pipe * pipe,u32 tkn,u32 tkn_val)2026 static int skl_tplg_fill_pipe_tkn(struct device *dev,
2027 struct skl_pipe *pipe, u32 tkn,
2028 u32 tkn_val)
2029 {
2030
2031 switch (tkn) {
2032 case SKL_TKN_U32_PIPE_CONN_TYPE:
2033 pipe->conn_type = tkn_val;
2034 break;
2035
2036 case SKL_TKN_U32_PIPE_PRIORITY:
2037 pipe->pipe_priority = tkn_val;
2038 break;
2039
2040 case SKL_TKN_U32_PIPE_MEM_PGS:
2041 pipe->memory_pages = tkn_val;
2042 break;
2043
2044 case SKL_TKN_U32_PMODE:
2045 pipe->lp_mode = tkn_val;
2046 break;
2047
2048 case SKL_TKN_U32_PIPE_DIRECTION:
2049 pipe->direction = tkn_val;
2050 break;
2051
2052 case SKL_TKN_U32_NUM_CONFIGS:
2053 pipe->nr_cfgs = tkn_val;
2054 break;
2055
2056 default:
2057 dev_err(dev, "Token not handled %d\n", tkn);
2058 return -EINVAL;
2059 }
2060
2061 return 0;
2062 }
2063
2064 /*
2065 * Add pipeline by parsing the relevant tokens
2066 * Return an existing pipe if the pipe already exists.
2067 */
skl_tplg_add_pipe(struct device * dev,struct skl_module_cfg * mconfig,struct skl_dev * skl,struct snd_soc_tplg_vendor_value_elem * tkn_elem)2068 static int skl_tplg_add_pipe(struct device *dev,
2069 struct skl_module_cfg *mconfig, struct skl_dev *skl,
2070 struct snd_soc_tplg_vendor_value_elem *tkn_elem)
2071 {
2072 struct skl_pipeline *ppl;
2073 struct skl_pipe *pipe;
2074 struct skl_pipe_params *params;
2075
2076 list_for_each_entry(ppl, &skl->ppl_list, node) {
2077 if (ppl->pipe->ppl_id == tkn_elem->value) {
2078 mconfig->pipe = ppl->pipe;
2079 return -EEXIST;
2080 }
2081 }
2082
2083 ppl = devm_kzalloc(dev, sizeof(*ppl), GFP_KERNEL);
2084 if (!ppl)
2085 return -ENOMEM;
2086
2087 pipe = devm_kzalloc(dev, sizeof(*pipe), GFP_KERNEL);
2088 if (!pipe)
2089 return -ENOMEM;
2090
2091 params = devm_kzalloc(dev, sizeof(*params), GFP_KERNEL);
2092 if (!params)
2093 return -ENOMEM;
2094
2095 pipe->p_params = params;
2096 pipe->ppl_id = tkn_elem->value;
2097 INIT_LIST_HEAD(&pipe->w_list);
2098
2099 ppl->pipe = pipe;
2100 list_add(&ppl->node, &skl->ppl_list);
2101
2102 mconfig->pipe = pipe;
2103 mconfig->pipe->state = SKL_PIPE_INVALID;
2104
2105 return 0;
2106 }
2107
skl_tplg_get_uuid(struct device * dev,guid_t * guid,struct snd_soc_tplg_vendor_uuid_elem * uuid_tkn)2108 static int skl_tplg_get_uuid(struct device *dev, guid_t *guid,
2109 struct snd_soc_tplg_vendor_uuid_elem *uuid_tkn)
2110 {
2111 if (uuid_tkn->token == SKL_TKN_UUID) {
2112 import_guid(guid, uuid_tkn->uuid);
2113 return 0;
2114 }
2115
2116 dev_err(dev, "Not an UUID token %d\n", uuid_tkn->token);
2117
2118 return -EINVAL;
2119 }
2120
skl_tplg_fill_pin(struct device * dev,struct snd_soc_tplg_vendor_value_elem * tkn_elem,struct skl_module_pin * m_pin,int pin_index)2121 static int skl_tplg_fill_pin(struct device *dev,
2122 struct snd_soc_tplg_vendor_value_elem *tkn_elem,
2123 struct skl_module_pin *m_pin,
2124 int pin_index)
2125 {
2126 int ret;
2127
2128 switch (tkn_elem->token) {
2129 case SKL_TKN_U32_PIN_MOD_ID:
2130 m_pin[pin_index].id.module_id = tkn_elem->value;
2131 break;
2132
2133 case SKL_TKN_U32_PIN_INST_ID:
2134 m_pin[pin_index].id.instance_id = tkn_elem->value;
2135 break;
2136
2137 case SKL_TKN_UUID:
2138 ret = skl_tplg_get_uuid(dev, &m_pin[pin_index].id.mod_uuid,
2139 (struct snd_soc_tplg_vendor_uuid_elem *)tkn_elem);
2140 if (ret < 0)
2141 return ret;
2142
2143 break;
2144
2145 default:
2146 dev_err(dev, "%d Not a pin token\n", tkn_elem->token);
2147 return -EINVAL;
2148 }
2149
2150 return 0;
2151 }
2152
2153 /*
2154 * Parse for pin config specific tokens to fill up the
2155 * module private data
2156 */
skl_tplg_fill_pins_info(struct device * dev,struct skl_module_cfg * mconfig,struct snd_soc_tplg_vendor_value_elem * tkn_elem,int dir,int pin_count)2157 static int skl_tplg_fill_pins_info(struct device *dev,
2158 struct skl_module_cfg *mconfig,
2159 struct snd_soc_tplg_vendor_value_elem *tkn_elem,
2160 int dir, int pin_count)
2161 {
2162 int ret;
2163 struct skl_module_pin *m_pin;
2164
2165 switch (dir) {
2166 case SKL_DIR_IN:
2167 m_pin = mconfig->m_in_pin;
2168 break;
2169
2170 case SKL_DIR_OUT:
2171 m_pin = mconfig->m_out_pin;
2172 break;
2173
2174 default:
2175 dev_err(dev, "Invalid direction value\n");
2176 return -EINVAL;
2177 }
2178
2179 ret = skl_tplg_fill_pin(dev, tkn_elem, m_pin, pin_count);
2180 if (ret < 0)
2181 return ret;
2182
2183 m_pin[pin_count].in_use = false;
2184 m_pin[pin_count].pin_state = SKL_PIN_UNBIND;
2185
2186 return 0;
2187 }
2188
2189 /*
2190 * Fill up input/output module config format based
2191 * on the direction
2192 */
skl_tplg_fill_fmt(struct device * dev,struct skl_module_fmt * dst_fmt,u32 tkn,u32 value)2193 static int skl_tplg_fill_fmt(struct device *dev,
2194 struct skl_module_fmt *dst_fmt,
2195 u32 tkn, u32 value)
2196 {
2197 switch (tkn) {
2198 case SKL_TKN_U32_FMT_CH:
2199 dst_fmt->channels = value;
2200 break;
2201
2202 case SKL_TKN_U32_FMT_FREQ:
2203 dst_fmt->s_freq = value;
2204 break;
2205
2206 case SKL_TKN_U32_FMT_BIT_DEPTH:
2207 dst_fmt->bit_depth = value;
2208 break;
2209
2210 case SKL_TKN_U32_FMT_SAMPLE_SIZE:
2211 dst_fmt->valid_bit_depth = value;
2212 break;
2213
2214 case SKL_TKN_U32_FMT_CH_CONFIG:
2215 dst_fmt->ch_cfg = value;
2216 break;
2217
2218 case SKL_TKN_U32_FMT_INTERLEAVE:
2219 dst_fmt->interleaving_style = value;
2220 break;
2221
2222 case SKL_TKN_U32_FMT_SAMPLE_TYPE:
2223 dst_fmt->sample_type = value;
2224 break;
2225
2226 case SKL_TKN_U32_FMT_CH_MAP:
2227 dst_fmt->ch_map = value;
2228 break;
2229
2230 default:
2231 dev_err(dev, "Invalid token %d\n", tkn);
2232 return -EINVAL;
2233 }
2234
2235 return 0;
2236 }
2237
skl_tplg_widget_fill_fmt(struct device * dev,struct skl_module_iface * fmt,u32 tkn,u32 val,u32 dir,int fmt_idx)2238 static int skl_tplg_widget_fill_fmt(struct device *dev,
2239 struct skl_module_iface *fmt,
2240 u32 tkn, u32 val, u32 dir, int fmt_idx)
2241 {
2242 struct skl_module_fmt *dst_fmt;
2243
2244 if (!fmt)
2245 return -EINVAL;
2246
2247 switch (dir) {
2248 case SKL_DIR_IN:
2249 dst_fmt = &fmt->inputs[fmt_idx].fmt;
2250 break;
2251
2252 case SKL_DIR_OUT:
2253 dst_fmt = &fmt->outputs[fmt_idx].fmt;
2254 break;
2255
2256 default:
2257 dev_err(dev, "Invalid direction: %d\n", dir);
2258 return -EINVAL;
2259 }
2260
2261 return skl_tplg_fill_fmt(dev, dst_fmt, tkn, val);
2262 }
2263
skl_tplg_fill_pin_dynamic_val(struct skl_module_pin * mpin,u32 pin_count,u32 value)2264 static void skl_tplg_fill_pin_dynamic_val(
2265 struct skl_module_pin *mpin, u32 pin_count, u32 value)
2266 {
2267 int i;
2268
2269 for (i = 0; i < pin_count; i++)
2270 mpin[i].is_dynamic = value;
2271 }
2272
2273 /*
2274 * Resource table in the manifest has pin specific resources
2275 * like pin and pin buffer size
2276 */
skl_tplg_manifest_pin_res_tkn(struct device * dev,struct snd_soc_tplg_vendor_value_elem * tkn_elem,struct skl_module_res * res,int pin_idx,int dir)2277 static int skl_tplg_manifest_pin_res_tkn(struct device *dev,
2278 struct snd_soc_tplg_vendor_value_elem *tkn_elem,
2279 struct skl_module_res *res, int pin_idx, int dir)
2280 {
2281 struct skl_module_pin_resources *m_pin;
2282
2283 switch (dir) {
2284 case SKL_DIR_IN:
2285 m_pin = &res->input[pin_idx];
2286 break;
2287
2288 case SKL_DIR_OUT:
2289 m_pin = &res->output[pin_idx];
2290 break;
2291
2292 default:
2293 dev_err(dev, "Invalid pin direction: %d\n", dir);
2294 return -EINVAL;
2295 }
2296
2297 switch (tkn_elem->token) {
2298 case SKL_TKN_MM_U32_RES_PIN_ID:
2299 m_pin->pin_index = tkn_elem->value;
2300 break;
2301
2302 case SKL_TKN_MM_U32_PIN_BUF:
2303 m_pin->buf_size = tkn_elem->value;
2304 break;
2305
2306 default:
2307 dev_err(dev, "Invalid token: %d\n", tkn_elem->token);
2308 return -EINVAL;
2309 }
2310
2311 return 0;
2312 }
2313
2314 /*
2315 * Fill module specific resources from the manifest's resource
2316 * table like CPS, DMA size, mem_pages.
2317 */
skl_tplg_fill_res_tkn(struct device * dev,struct snd_soc_tplg_vendor_value_elem * tkn_elem,struct skl_module_res * res,int pin_idx,int dir)2318 static int skl_tplg_fill_res_tkn(struct device *dev,
2319 struct snd_soc_tplg_vendor_value_elem *tkn_elem,
2320 struct skl_module_res *res,
2321 int pin_idx, int dir)
2322 {
2323 int ret, tkn_count = 0;
2324
2325 if (!res)
2326 return -EINVAL;
2327
2328 switch (tkn_elem->token) {
2329 case SKL_TKN_MM_U32_DMA_SIZE:
2330 res->dma_buffer_size = tkn_elem->value;
2331 break;
2332
2333 case SKL_TKN_MM_U32_CPC:
2334 res->cpc = tkn_elem->value;
2335 break;
2336
2337 case SKL_TKN_U32_MEM_PAGES:
2338 res->is_pages = tkn_elem->value;
2339 break;
2340
2341 case SKL_TKN_U32_OBS:
2342 res->obs = tkn_elem->value;
2343 break;
2344
2345 case SKL_TKN_U32_IBS:
2346 res->ibs = tkn_elem->value;
2347 break;
2348
2349 case SKL_TKN_MM_U32_RES_PIN_ID:
2350 case SKL_TKN_MM_U32_PIN_BUF:
2351 ret = skl_tplg_manifest_pin_res_tkn(dev, tkn_elem, res,
2352 pin_idx, dir);
2353 if (ret < 0)
2354 return ret;
2355 break;
2356
2357 case SKL_TKN_MM_U32_CPS:
2358 case SKL_TKN_U32_MAX_MCPS:
2359 /* ignore unused tokens */
2360 break;
2361
2362 default:
2363 dev_err(dev, "Not a res type token: %d", tkn_elem->token);
2364 return -EINVAL;
2365
2366 }
2367 tkn_count++;
2368
2369 return tkn_count;
2370 }
2371
2372 /*
2373 * Parse tokens to fill up the module private data
2374 */
skl_tplg_get_token(struct device * dev,struct snd_soc_tplg_vendor_value_elem * tkn_elem,struct skl_dev * skl,struct skl_module_cfg * mconfig)2375 static int skl_tplg_get_token(struct device *dev,
2376 struct snd_soc_tplg_vendor_value_elem *tkn_elem,
2377 struct skl_dev *skl, struct skl_module_cfg *mconfig)
2378 {
2379 int tkn_count = 0;
2380 int ret;
2381 static int is_pipe_exists;
2382 static int pin_index, dir, conf_idx;
2383 struct skl_module_iface *iface = NULL;
2384 struct skl_module_res *res = NULL;
2385 int res_idx = mconfig->res_idx;
2386 int fmt_idx = mconfig->fmt_idx;
2387
2388 /*
2389 * If the manifest structure contains no modules, fill all
2390 * the module data to 0th index.
2391 * res_idx and fmt_idx are default set to 0.
2392 */
2393 if (skl->nr_modules == 0) {
2394 res = &mconfig->module->resources[res_idx];
2395 iface = &mconfig->module->formats[fmt_idx];
2396 }
2397
2398 if (tkn_elem->token > SKL_TKN_MAX)
2399 return -EINVAL;
2400
2401 switch (tkn_elem->token) {
2402 case SKL_TKN_U8_IN_QUEUE_COUNT:
2403 mconfig->module->max_input_pins = tkn_elem->value;
2404 break;
2405
2406 case SKL_TKN_U8_OUT_QUEUE_COUNT:
2407 mconfig->module->max_output_pins = tkn_elem->value;
2408 break;
2409
2410 case SKL_TKN_U8_DYN_IN_PIN:
2411 if (!mconfig->m_in_pin)
2412 mconfig->m_in_pin =
2413 devm_kcalloc(dev, MAX_IN_QUEUE,
2414 sizeof(*mconfig->m_in_pin),
2415 GFP_KERNEL);
2416 if (!mconfig->m_in_pin)
2417 return -ENOMEM;
2418
2419 skl_tplg_fill_pin_dynamic_val(mconfig->m_in_pin, MAX_IN_QUEUE,
2420 tkn_elem->value);
2421 break;
2422
2423 case SKL_TKN_U8_DYN_OUT_PIN:
2424 if (!mconfig->m_out_pin)
2425 mconfig->m_out_pin =
2426 devm_kcalloc(dev, MAX_IN_QUEUE,
2427 sizeof(*mconfig->m_in_pin),
2428 GFP_KERNEL);
2429 if (!mconfig->m_out_pin)
2430 return -ENOMEM;
2431
2432 skl_tplg_fill_pin_dynamic_val(mconfig->m_out_pin, MAX_OUT_QUEUE,
2433 tkn_elem->value);
2434 break;
2435
2436 case SKL_TKN_U8_TIME_SLOT:
2437 mconfig->time_slot = tkn_elem->value;
2438 break;
2439
2440 case SKL_TKN_U8_CORE_ID:
2441 mconfig->core_id = tkn_elem->value;
2442 break;
2443
2444 case SKL_TKN_U8_MOD_TYPE:
2445 mconfig->m_type = tkn_elem->value;
2446 break;
2447
2448 case SKL_TKN_U8_DEV_TYPE:
2449 mconfig->dev_type = tkn_elem->value;
2450 break;
2451
2452 case SKL_TKN_U8_HW_CONN_TYPE:
2453 mconfig->hw_conn_type = tkn_elem->value;
2454 break;
2455
2456 case SKL_TKN_U16_MOD_INST_ID:
2457 mconfig->id.instance_id =
2458 tkn_elem->value;
2459 break;
2460
2461 case SKL_TKN_U32_MEM_PAGES:
2462 case SKL_TKN_U32_MAX_MCPS:
2463 case SKL_TKN_U32_OBS:
2464 case SKL_TKN_U32_IBS:
2465 ret = skl_tplg_fill_res_tkn(dev, tkn_elem, res, pin_index, dir);
2466 if (ret < 0)
2467 return ret;
2468
2469 break;
2470
2471 case SKL_TKN_U32_VBUS_ID:
2472 mconfig->vbus_id = tkn_elem->value;
2473 break;
2474
2475 case SKL_TKN_U32_PARAMS_FIXUP:
2476 mconfig->params_fixup = tkn_elem->value;
2477 break;
2478
2479 case SKL_TKN_U32_CONVERTER:
2480 mconfig->converter = tkn_elem->value;
2481 break;
2482
2483 case SKL_TKN_U32_D0I3_CAPS:
2484 mconfig->d0i3_caps = tkn_elem->value;
2485 break;
2486
2487 case SKL_TKN_U32_PIPE_ID:
2488 ret = skl_tplg_add_pipe(dev,
2489 mconfig, skl, tkn_elem);
2490
2491 if (ret < 0) {
2492 if (ret == -EEXIST) {
2493 is_pipe_exists = 1;
2494 break;
2495 }
2496 return is_pipe_exists;
2497 }
2498
2499 break;
2500
2501 case SKL_TKN_U32_PIPE_CONFIG_ID:
2502 conf_idx = tkn_elem->value;
2503 break;
2504
2505 case SKL_TKN_U32_PIPE_CONN_TYPE:
2506 case SKL_TKN_U32_PIPE_PRIORITY:
2507 case SKL_TKN_U32_PIPE_MEM_PGS:
2508 case SKL_TKN_U32_PMODE:
2509 case SKL_TKN_U32_PIPE_DIRECTION:
2510 case SKL_TKN_U32_NUM_CONFIGS:
2511 if (is_pipe_exists) {
2512 ret = skl_tplg_fill_pipe_tkn(dev, mconfig->pipe,
2513 tkn_elem->token, tkn_elem->value);
2514 if (ret < 0)
2515 return ret;
2516 }
2517
2518 break;
2519
2520 case SKL_TKN_U32_PATH_MEM_PGS:
2521 case SKL_TKN_U32_CFG_FREQ:
2522 case SKL_TKN_U8_CFG_CHAN:
2523 case SKL_TKN_U8_CFG_BPS:
2524 if (mconfig->pipe->nr_cfgs) {
2525 ret = skl_tplg_fill_pipe_cfg(dev, mconfig->pipe,
2526 tkn_elem->token, tkn_elem->value,
2527 conf_idx, dir);
2528 if (ret < 0)
2529 return ret;
2530 }
2531 break;
2532
2533 case SKL_TKN_CFG_MOD_RES_ID:
2534 mconfig->mod_cfg[conf_idx].res_idx = tkn_elem->value;
2535 break;
2536
2537 case SKL_TKN_CFG_MOD_FMT_ID:
2538 mconfig->mod_cfg[conf_idx].fmt_idx = tkn_elem->value;
2539 break;
2540
2541 /*
2542 * SKL_TKN_U32_DIR_PIN_COUNT token has the value for both
2543 * direction and the pin count. The first four bits represent
2544 * direction and next four the pin count.
2545 */
2546 case SKL_TKN_U32_DIR_PIN_COUNT:
2547 dir = tkn_elem->value & SKL_IN_DIR_BIT_MASK;
2548 pin_index = (tkn_elem->value &
2549 SKL_PIN_COUNT_MASK) >> 4;
2550
2551 break;
2552
2553 case SKL_TKN_U32_FMT_CH:
2554 case SKL_TKN_U32_FMT_FREQ:
2555 case SKL_TKN_U32_FMT_BIT_DEPTH:
2556 case SKL_TKN_U32_FMT_SAMPLE_SIZE:
2557 case SKL_TKN_U32_FMT_CH_CONFIG:
2558 case SKL_TKN_U32_FMT_INTERLEAVE:
2559 case SKL_TKN_U32_FMT_SAMPLE_TYPE:
2560 case SKL_TKN_U32_FMT_CH_MAP:
2561 ret = skl_tplg_widget_fill_fmt(dev, iface, tkn_elem->token,
2562 tkn_elem->value, dir, pin_index);
2563
2564 if (ret < 0)
2565 return ret;
2566
2567 break;
2568
2569 case SKL_TKN_U32_PIN_MOD_ID:
2570 case SKL_TKN_U32_PIN_INST_ID:
2571 case SKL_TKN_UUID:
2572 ret = skl_tplg_fill_pins_info(dev,
2573 mconfig, tkn_elem, dir,
2574 pin_index);
2575 if (ret < 0)
2576 return ret;
2577
2578 break;
2579
2580 case SKL_TKN_U32_FMT_CFG_IDX:
2581 if (tkn_elem->value > SKL_MAX_PARAMS_TYPES)
2582 return -EINVAL;
2583
2584 mconfig->fmt_cfg_idx = tkn_elem->value;
2585 break;
2586
2587 case SKL_TKN_U32_CAPS_SIZE:
2588 mconfig->formats_config[mconfig->fmt_cfg_idx].caps_size =
2589 tkn_elem->value;
2590
2591 break;
2592
2593 case SKL_TKN_U32_CAPS_SET_PARAMS:
2594 mconfig->formats_config[mconfig->fmt_cfg_idx].set_params =
2595 tkn_elem->value;
2596 break;
2597
2598 case SKL_TKN_U32_CAPS_PARAMS_ID:
2599 mconfig->formats_config[mconfig->fmt_cfg_idx].param_id =
2600 tkn_elem->value;
2601 break;
2602
2603 case SKL_TKN_U32_PROC_DOMAIN:
2604 mconfig->domain =
2605 tkn_elem->value;
2606
2607 break;
2608
2609 case SKL_TKN_U32_DMA_BUF_SIZE:
2610 mconfig->dma_buffer_size = tkn_elem->value;
2611 break;
2612
2613 case SKL_TKN_U8_IN_PIN_TYPE:
2614 case SKL_TKN_U8_OUT_PIN_TYPE:
2615 case SKL_TKN_U8_CONN_TYPE:
2616 break;
2617
2618 default:
2619 dev_err(dev, "Token %d not handled\n",
2620 tkn_elem->token);
2621 return -EINVAL;
2622 }
2623
2624 tkn_count++;
2625
2626 return tkn_count;
2627 }
2628
2629 /*
2630 * Parse the vendor array for specific tokens to construct
2631 * module private data
2632 */
skl_tplg_get_tokens(struct device * dev,char * pvt_data,struct skl_dev * skl,struct skl_module_cfg * mconfig,int block_size)2633 static int skl_tplg_get_tokens(struct device *dev,
2634 char *pvt_data, struct skl_dev *skl,
2635 struct skl_module_cfg *mconfig, int block_size)
2636 {
2637 struct snd_soc_tplg_vendor_array *array;
2638 struct snd_soc_tplg_vendor_value_elem *tkn_elem;
2639 int tkn_count = 0, ret;
2640 int off = 0, tuple_size = 0;
2641 bool is_module_guid = true;
2642
2643 if (block_size <= 0)
2644 return -EINVAL;
2645
2646 while (tuple_size < block_size) {
2647 array = (struct snd_soc_tplg_vendor_array *)(pvt_data + off);
2648
2649 off += array->size;
2650
2651 switch (array->type) {
2652 case SND_SOC_TPLG_TUPLE_TYPE_STRING:
2653 dev_warn(dev, "no string tokens expected for skl tplg\n");
2654 continue;
2655
2656 case SND_SOC_TPLG_TUPLE_TYPE_UUID:
2657 if (is_module_guid) {
2658 ret = skl_tplg_get_uuid(dev, (guid_t *)mconfig->guid,
2659 array->uuid);
2660 is_module_guid = false;
2661 } else {
2662 ret = skl_tplg_get_token(dev, array->value, skl,
2663 mconfig);
2664 }
2665
2666 if (ret < 0)
2667 return ret;
2668
2669 tuple_size += sizeof(*array->uuid);
2670
2671 continue;
2672
2673 default:
2674 tkn_elem = array->value;
2675 tkn_count = 0;
2676 break;
2677 }
2678
2679 while (tkn_count <= (array->num_elems - 1)) {
2680 ret = skl_tplg_get_token(dev, tkn_elem,
2681 skl, mconfig);
2682
2683 if (ret < 0)
2684 return ret;
2685
2686 tkn_count = tkn_count + ret;
2687 tkn_elem++;
2688 }
2689
2690 tuple_size += tkn_count * sizeof(*tkn_elem);
2691 }
2692
2693 return off;
2694 }
2695
2696 /*
2697 * Every data block is preceded by a descriptor to read the number
2698 * of data blocks, they type of the block and it's size
2699 */
skl_tplg_get_desc_blocks(struct device * dev,struct snd_soc_tplg_vendor_array * array)2700 static int skl_tplg_get_desc_blocks(struct device *dev,
2701 struct snd_soc_tplg_vendor_array *array)
2702 {
2703 struct snd_soc_tplg_vendor_value_elem *tkn_elem;
2704
2705 tkn_elem = array->value;
2706
2707 switch (tkn_elem->token) {
2708 case SKL_TKN_U8_NUM_BLOCKS:
2709 case SKL_TKN_U8_BLOCK_TYPE:
2710 case SKL_TKN_U16_BLOCK_SIZE:
2711 return tkn_elem->value;
2712
2713 default:
2714 dev_err(dev, "Invalid descriptor token %d\n", tkn_elem->token);
2715 break;
2716 }
2717
2718 return -EINVAL;
2719 }
2720
2721 /* Functions to parse private data from configuration file format v4 */
2722
2723 /*
2724 * Add pipeline from topology binary into driver pipeline list
2725 *
2726 * If already added we return that instance
2727 * Otherwise we create a new instance and add into driver list
2728 */
skl_tplg_add_pipe_v4(struct device * dev,struct skl_module_cfg * mconfig,struct skl_dev * skl,struct skl_dfw_v4_pipe * dfw_pipe)2729 static int skl_tplg_add_pipe_v4(struct device *dev,
2730 struct skl_module_cfg *mconfig, struct skl_dev *skl,
2731 struct skl_dfw_v4_pipe *dfw_pipe)
2732 {
2733 struct skl_pipeline *ppl;
2734 struct skl_pipe *pipe;
2735 struct skl_pipe_params *params;
2736
2737 list_for_each_entry(ppl, &skl->ppl_list, node) {
2738 if (ppl->pipe->ppl_id == dfw_pipe->pipe_id) {
2739 mconfig->pipe = ppl->pipe;
2740 return 0;
2741 }
2742 }
2743
2744 ppl = devm_kzalloc(dev, sizeof(*ppl), GFP_KERNEL);
2745 if (!ppl)
2746 return -ENOMEM;
2747
2748 pipe = devm_kzalloc(dev, sizeof(*pipe), GFP_KERNEL);
2749 if (!pipe)
2750 return -ENOMEM;
2751
2752 params = devm_kzalloc(dev, sizeof(*params), GFP_KERNEL);
2753 if (!params)
2754 return -ENOMEM;
2755
2756 pipe->ppl_id = dfw_pipe->pipe_id;
2757 pipe->memory_pages = dfw_pipe->memory_pages;
2758 pipe->pipe_priority = dfw_pipe->pipe_priority;
2759 pipe->conn_type = dfw_pipe->conn_type;
2760 pipe->state = SKL_PIPE_INVALID;
2761 pipe->p_params = params;
2762 INIT_LIST_HEAD(&pipe->w_list);
2763
2764 ppl->pipe = pipe;
2765 list_add(&ppl->node, &skl->ppl_list);
2766
2767 mconfig->pipe = pipe;
2768
2769 return 0;
2770 }
2771
skl_fill_module_pin_info_v4(struct skl_dfw_v4_module_pin * dfw_pin,struct skl_module_pin * m_pin,bool is_dynamic,int max_pin)2772 static void skl_fill_module_pin_info_v4(struct skl_dfw_v4_module_pin *dfw_pin,
2773 struct skl_module_pin *m_pin,
2774 bool is_dynamic, int max_pin)
2775 {
2776 int i;
2777
2778 for (i = 0; i < max_pin; i++) {
2779 m_pin[i].id.module_id = dfw_pin[i].module_id;
2780 m_pin[i].id.instance_id = dfw_pin[i].instance_id;
2781 m_pin[i].in_use = false;
2782 m_pin[i].is_dynamic = is_dynamic;
2783 m_pin[i].pin_state = SKL_PIN_UNBIND;
2784 }
2785 }
2786
skl_tplg_fill_fmt_v4(struct skl_module_pin_fmt * dst_fmt,struct skl_dfw_v4_module_fmt * src_fmt,int pins)2787 static void skl_tplg_fill_fmt_v4(struct skl_module_pin_fmt *dst_fmt,
2788 struct skl_dfw_v4_module_fmt *src_fmt,
2789 int pins)
2790 {
2791 int i;
2792
2793 for (i = 0; i < pins; i++) {
2794 dst_fmt[i].fmt.channels = src_fmt[i].channels;
2795 dst_fmt[i].fmt.s_freq = src_fmt[i].freq;
2796 dst_fmt[i].fmt.bit_depth = src_fmt[i].bit_depth;
2797 dst_fmt[i].fmt.valid_bit_depth = src_fmt[i].valid_bit_depth;
2798 dst_fmt[i].fmt.ch_cfg = src_fmt[i].ch_cfg;
2799 dst_fmt[i].fmt.ch_map = src_fmt[i].ch_map;
2800 dst_fmt[i].fmt.interleaving_style =
2801 src_fmt[i].interleaving_style;
2802 dst_fmt[i].fmt.sample_type = src_fmt[i].sample_type;
2803 }
2804 }
2805
skl_tplg_get_pvt_data_v4(struct snd_soc_tplg_dapm_widget * tplg_w,struct skl_dev * skl,struct device * dev,struct skl_module_cfg * mconfig)2806 static int skl_tplg_get_pvt_data_v4(struct snd_soc_tplg_dapm_widget *tplg_w,
2807 struct skl_dev *skl, struct device *dev,
2808 struct skl_module_cfg *mconfig)
2809 {
2810 struct skl_dfw_v4_module *dfw =
2811 (struct skl_dfw_v4_module *)tplg_w->priv.data;
2812 int ret;
2813 int idx = mconfig->fmt_cfg_idx;
2814
2815 dev_dbg(dev, "Parsing Skylake v4 widget topology data\n");
2816
2817 ret = guid_parse(dfw->uuid, (guid_t *)mconfig->guid);
2818 if (ret)
2819 return ret;
2820 mconfig->id.module_id = -1;
2821 mconfig->id.instance_id = dfw->instance_id;
2822 mconfig->module->resources[0].cpc = dfw->max_mcps / 1000;
2823 mconfig->module->resources[0].ibs = dfw->ibs;
2824 mconfig->module->resources[0].obs = dfw->obs;
2825 mconfig->core_id = dfw->core_id;
2826 mconfig->module->max_input_pins = dfw->max_in_queue;
2827 mconfig->module->max_output_pins = dfw->max_out_queue;
2828 mconfig->module->loadable = dfw->is_loadable;
2829 skl_tplg_fill_fmt_v4(mconfig->module->formats[0].inputs, dfw->in_fmt,
2830 MAX_IN_QUEUE);
2831 skl_tplg_fill_fmt_v4(mconfig->module->formats[0].outputs, dfw->out_fmt,
2832 MAX_OUT_QUEUE);
2833
2834 mconfig->params_fixup = dfw->params_fixup;
2835 mconfig->converter = dfw->converter;
2836 mconfig->m_type = dfw->module_type;
2837 mconfig->vbus_id = dfw->vbus_id;
2838 mconfig->module->resources[0].is_pages = dfw->mem_pages;
2839
2840 ret = skl_tplg_add_pipe_v4(dev, mconfig, skl, &dfw->pipe);
2841 if (ret)
2842 return ret;
2843
2844 mconfig->dev_type = dfw->dev_type;
2845 mconfig->hw_conn_type = dfw->hw_conn_type;
2846 mconfig->time_slot = dfw->time_slot;
2847 mconfig->formats_config[idx].caps_size = dfw->caps.caps_size;
2848
2849 mconfig->m_in_pin = devm_kcalloc(dev,
2850 MAX_IN_QUEUE, sizeof(*mconfig->m_in_pin),
2851 GFP_KERNEL);
2852 if (!mconfig->m_in_pin)
2853 return -ENOMEM;
2854
2855 mconfig->m_out_pin = devm_kcalloc(dev,
2856 MAX_OUT_QUEUE, sizeof(*mconfig->m_out_pin),
2857 GFP_KERNEL);
2858 if (!mconfig->m_out_pin)
2859 return -ENOMEM;
2860
2861 skl_fill_module_pin_info_v4(dfw->in_pin, mconfig->m_in_pin,
2862 dfw->is_dynamic_in_pin,
2863 mconfig->module->max_input_pins);
2864 skl_fill_module_pin_info_v4(dfw->out_pin, mconfig->m_out_pin,
2865 dfw->is_dynamic_out_pin,
2866 mconfig->module->max_output_pins);
2867
2868 if (mconfig->formats_config[idx].caps_size) {
2869 mconfig->formats_config[idx].set_params = dfw->caps.set_params;
2870 mconfig->formats_config[idx].param_id = dfw->caps.param_id;
2871 mconfig->formats_config[idx].caps =
2872 devm_kzalloc(dev, mconfig->formats_config[idx].caps_size,
2873 GFP_KERNEL);
2874 if (!mconfig->formats_config[idx].caps)
2875 return -ENOMEM;
2876 memcpy(mconfig->formats_config[idx].caps, dfw->caps.caps,
2877 dfw->caps.caps_size);
2878 }
2879
2880 return 0;
2881 }
2882
skl_tplg_get_caps_data(struct device * dev,char * data,struct skl_module_cfg * mconfig)2883 static int skl_tplg_get_caps_data(struct device *dev, char *data,
2884 struct skl_module_cfg *mconfig)
2885 {
2886 int idx = mconfig->fmt_cfg_idx;
2887
2888 if (mconfig->formats_config[idx].caps_size > 0) {
2889 mconfig->formats_config[idx].caps =
2890 devm_kzalloc(dev, mconfig->formats_config[idx].caps_size,
2891 GFP_KERNEL);
2892 if (!mconfig->formats_config[idx].caps)
2893 return -ENOMEM;
2894 memcpy(mconfig->formats_config[idx].caps, data,
2895 mconfig->formats_config[idx].caps_size);
2896 }
2897
2898 return mconfig->formats_config[idx].caps_size;
2899 }
2900
2901 /*
2902 * Parse the private data for the token and corresponding value.
2903 * The private data can have multiple data blocks. So, a data block
2904 * is preceded by a descriptor for number of blocks and a descriptor
2905 * for the type and size of the suceeding data block.
2906 */
skl_tplg_get_pvt_data(struct snd_soc_tplg_dapm_widget * tplg_w,struct skl_dev * skl,struct device * dev,struct skl_module_cfg * mconfig)2907 static int skl_tplg_get_pvt_data(struct snd_soc_tplg_dapm_widget *tplg_w,
2908 struct skl_dev *skl, struct device *dev,
2909 struct skl_module_cfg *mconfig)
2910 {
2911 struct snd_soc_tplg_vendor_array *array;
2912 int num_blocks, block_size, block_type, off = 0;
2913 char *data;
2914 int ret;
2915
2916 /*
2917 * v4 configuration files have a valid UUID at the start of
2918 * the widget's private data.
2919 */
2920 if (uuid_is_valid((char *)tplg_w->priv.data))
2921 return skl_tplg_get_pvt_data_v4(tplg_w, skl, dev, mconfig);
2922
2923 /* Read the NUM_DATA_BLOCKS descriptor */
2924 array = (struct snd_soc_tplg_vendor_array *)tplg_w->priv.data;
2925 ret = skl_tplg_get_desc_blocks(dev, array);
2926 if (ret < 0)
2927 return ret;
2928 num_blocks = ret;
2929
2930 off += array->size;
2931 /* Read the BLOCK_TYPE and BLOCK_SIZE descriptor */
2932 while (num_blocks > 0) {
2933 array = (struct snd_soc_tplg_vendor_array *)
2934 (tplg_w->priv.data + off);
2935
2936 ret = skl_tplg_get_desc_blocks(dev, array);
2937
2938 if (ret < 0)
2939 return ret;
2940 block_type = ret;
2941 off += array->size;
2942
2943 array = (struct snd_soc_tplg_vendor_array *)
2944 (tplg_w->priv.data + off);
2945
2946 ret = skl_tplg_get_desc_blocks(dev, array);
2947
2948 if (ret < 0)
2949 return ret;
2950 block_size = ret;
2951 off += array->size;
2952
2953 array = (struct snd_soc_tplg_vendor_array *)
2954 (tplg_w->priv.data + off);
2955
2956 data = (tplg_w->priv.data + off);
2957
2958 if (block_type == SKL_TYPE_TUPLE) {
2959 ret = skl_tplg_get_tokens(dev, data,
2960 skl, mconfig, block_size);
2961 } else {
2962 ret = skl_tplg_get_caps_data(dev, data, mconfig);
2963 }
2964
2965 if (ret < 0)
2966 return ret;
2967
2968 --num_blocks;
2969 off += ret;
2970 }
2971
2972 return 0;
2973 }
2974
skl_clear_pin_config(struct snd_soc_component * component,struct snd_soc_dapm_widget * w)2975 static void skl_clear_pin_config(struct snd_soc_component *component,
2976 struct snd_soc_dapm_widget *w)
2977 {
2978 int i;
2979 struct skl_module_cfg *mconfig;
2980 struct skl_pipe *pipe;
2981
2982 if (!strncmp(w->dapm->component->name, component->name,
2983 strlen(component->name))) {
2984 mconfig = w->priv;
2985 pipe = mconfig->pipe;
2986 for (i = 0; i < mconfig->module->max_input_pins; i++) {
2987 mconfig->m_in_pin[i].in_use = false;
2988 mconfig->m_in_pin[i].pin_state = SKL_PIN_UNBIND;
2989 }
2990 for (i = 0; i < mconfig->module->max_output_pins; i++) {
2991 mconfig->m_out_pin[i].in_use = false;
2992 mconfig->m_out_pin[i].pin_state = SKL_PIN_UNBIND;
2993 }
2994 pipe->state = SKL_PIPE_INVALID;
2995 mconfig->m_state = SKL_MODULE_UNINIT;
2996 }
2997 }
2998
skl_cleanup_resources(struct skl_dev * skl)2999 void skl_cleanup_resources(struct skl_dev *skl)
3000 {
3001 struct snd_soc_component *soc_component = skl->component;
3002 struct snd_soc_dapm_widget *w;
3003 struct snd_soc_card *card;
3004
3005 if (soc_component == NULL)
3006 return;
3007
3008 card = soc_component->card;
3009 if (!card || !card->instantiated)
3010 return;
3011
3012 list_for_each_entry(w, &card->widgets, list) {
3013 if (is_skl_dsp_widget_type(w, skl->dev) && w->priv != NULL)
3014 skl_clear_pin_config(soc_component, w);
3015 }
3016
3017 skl_clear_module_cnt(skl->dsp);
3018 }
3019
3020 /*
3021 * Topology core widget load callback
3022 *
3023 * This is used to save the private data for each widget which gives
3024 * information to the driver about module and pipeline parameters which DSP
3025 * FW expects like ids, resource values, formats etc
3026 */
skl_tplg_widget_load(struct snd_soc_component * cmpnt,int index,struct snd_soc_dapm_widget * w,struct snd_soc_tplg_dapm_widget * tplg_w)3027 static int skl_tplg_widget_load(struct snd_soc_component *cmpnt, int index,
3028 struct snd_soc_dapm_widget *w,
3029 struct snd_soc_tplg_dapm_widget *tplg_w)
3030 {
3031 int ret;
3032 struct hdac_bus *bus = snd_soc_component_get_drvdata(cmpnt);
3033 struct skl_dev *skl = bus_to_skl(bus);
3034 struct skl_module_cfg *mconfig;
3035
3036 if (!tplg_w->priv.size)
3037 goto bind_event;
3038
3039 mconfig = devm_kzalloc(bus->dev, sizeof(*mconfig), GFP_KERNEL);
3040
3041 if (!mconfig)
3042 return -ENOMEM;
3043
3044 if (skl->nr_modules == 0) {
3045 mconfig->module = devm_kzalloc(bus->dev,
3046 sizeof(*mconfig->module), GFP_KERNEL);
3047 if (!mconfig->module)
3048 return -ENOMEM;
3049 }
3050
3051 w->priv = mconfig;
3052
3053 /*
3054 * module binary can be loaded later, so set it to query when
3055 * module is load for a use case
3056 */
3057 mconfig->id.module_id = -1;
3058
3059 /* To provide backward compatibility, set default as SKL_PARAM_INIT */
3060 mconfig->fmt_cfg_idx = SKL_PARAM_INIT;
3061
3062 /* Parse private data for tuples */
3063 ret = skl_tplg_get_pvt_data(tplg_w, skl, bus->dev, mconfig);
3064 if (ret < 0)
3065 return ret;
3066
3067 skl_debug_init_module(skl->debugfs, w, mconfig);
3068
3069 bind_event:
3070 if (tplg_w->event_type == 0) {
3071 dev_dbg(bus->dev, "ASoC: No event handler required\n");
3072 return 0;
3073 }
3074
3075 ret = snd_soc_tplg_widget_bind_event(w, skl_tplg_widget_ops,
3076 ARRAY_SIZE(skl_tplg_widget_ops),
3077 tplg_w->event_type);
3078
3079 if (ret) {
3080 dev_err(bus->dev, "%s: No matching event handlers found for %d\n",
3081 __func__, tplg_w->event_type);
3082 return -EINVAL;
3083 }
3084
3085 return 0;
3086 }
3087
skl_init_algo_data(struct device * dev,struct soc_bytes_ext * be,struct snd_soc_tplg_bytes_control * bc)3088 static int skl_init_algo_data(struct device *dev, struct soc_bytes_ext *be,
3089 struct snd_soc_tplg_bytes_control *bc)
3090 {
3091 struct skl_algo_data *ac;
3092 struct skl_dfw_algo_data *dfw_ac =
3093 (struct skl_dfw_algo_data *)bc->priv.data;
3094
3095 ac = devm_kzalloc(dev, sizeof(*ac), GFP_KERNEL);
3096 if (!ac)
3097 return -ENOMEM;
3098
3099 /* Fill private data */
3100 ac->max = dfw_ac->max;
3101 ac->param_id = dfw_ac->param_id;
3102 ac->set_params = dfw_ac->set_params;
3103 ac->size = dfw_ac->max;
3104
3105 if (ac->max) {
3106 ac->params = devm_kzalloc(dev, ac->max, GFP_KERNEL);
3107 if (!ac->params)
3108 return -ENOMEM;
3109
3110 memcpy(ac->params, dfw_ac->params, ac->max);
3111 }
3112
3113 be->dobj.private = ac;
3114 return 0;
3115 }
3116
skl_init_enum_data(struct device * dev,struct soc_enum * se,struct snd_soc_tplg_enum_control * ec)3117 static int skl_init_enum_data(struct device *dev, struct soc_enum *se,
3118 struct snd_soc_tplg_enum_control *ec)
3119 {
3120
3121 void *data;
3122
3123 if (ec->priv.size) {
3124 data = devm_kzalloc(dev, sizeof(ec->priv.size), GFP_KERNEL);
3125 if (!data)
3126 return -ENOMEM;
3127 memcpy(data, ec->priv.data, ec->priv.size);
3128 se->dobj.private = data;
3129 }
3130
3131 return 0;
3132
3133 }
3134
skl_tplg_control_load(struct snd_soc_component * cmpnt,int index,struct snd_kcontrol_new * kctl,struct snd_soc_tplg_ctl_hdr * hdr)3135 static int skl_tplg_control_load(struct snd_soc_component *cmpnt,
3136 int index,
3137 struct snd_kcontrol_new *kctl,
3138 struct snd_soc_tplg_ctl_hdr *hdr)
3139 {
3140 struct soc_bytes_ext *sb;
3141 struct snd_soc_tplg_bytes_control *tplg_bc;
3142 struct snd_soc_tplg_enum_control *tplg_ec;
3143 struct hdac_bus *bus = snd_soc_component_get_drvdata(cmpnt);
3144 struct soc_enum *se;
3145
3146 switch (hdr->ops.info) {
3147 case SND_SOC_TPLG_CTL_BYTES:
3148 tplg_bc = container_of(hdr,
3149 struct snd_soc_tplg_bytes_control, hdr);
3150 if (kctl->access & SNDRV_CTL_ELEM_ACCESS_TLV_CALLBACK) {
3151 sb = (struct soc_bytes_ext *)kctl->private_value;
3152 if (tplg_bc->priv.size)
3153 return skl_init_algo_data(
3154 bus->dev, sb, tplg_bc);
3155 }
3156 break;
3157
3158 case SND_SOC_TPLG_CTL_ENUM:
3159 tplg_ec = container_of(hdr,
3160 struct snd_soc_tplg_enum_control, hdr);
3161 if (kctl->access & SNDRV_CTL_ELEM_ACCESS_READ) {
3162 se = (struct soc_enum *)kctl->private_value;
3163 if (tplg_ec->priv.size)
3164 skl_init_enum_data(bus->dev, se, tplg_ec);
3165 }
3166
3167 /*
3168 * now that the control initializations are done, remove
3169 * write permission for the DMIC configuration enums to
3170 * avoid conflicts between NHLT settings and user interaction
3171 */
3172
3173 if (hdr->ops.get == SKL_CONTROL_TYPE_MULTI_IO_SELECT_DMIC)
3174 kctl->access = SNDRV_CTL_ELEM_ACCESS_READ;
3175
3176 break;
3177
3178 default:
3179 dev_dbg(bus->dev, "Control load not supported %d:%d:%d\n",
3180 hdr->ops.get, hdr->ops.put, hdr->ops.info);
3181 break;
3182 }
3183
3184 return 0;
3185 }
3186
skl_tplg_fill_str_mfest_tkn(struct device * dev,struct snd_soc_tplg_vendor_string_elem * str_elem,struct skl_dev * skl)3187 static int skl_tplg_fill_str_mfest_tkn(struct device *dev,
3188 struct snd_soc_tplg_vendor_string_elem *str_elem,
3189 struct skl_dev *skl)
3190 {
3191 int tkn_count = 0;
3192 static int ref_count;
3193
3194 switch (str_elem->token) {
3195 case SKL_TKN_STR_LIB_NAME:
3196 if (ref_count > skl->lib_count - 1) {
3197 ref_count = 0;
3198 return -EINVAL;
3199 }
3200
3201 strncpy(skl->lib_info[ref_count].name,
3202 str_elem->string,
3203 ARRAY_SIZE(skl->lib_info[ref_count].name));
3204 ref_count++;
3205 break;
3206
3207 default:
3208 dev_err(dev, "Not a string token %d\n", str_elem->token);
3209 break;
3210 }
3211 tkn_count++;
3212
3213 return tkn_count;
3214 }
3215
skl_tplg_get_str_tkn(struct device * dev,struct snd_soc_tplg_vendor_array * array,struct skl_dev * skl)3216 static int skl_tplg_get_str_tkn(struct device *dev,
3217 struct snd_soc_tplg_vendor_array *array,
3218 struct skl_dev *skl)
3219 {
3220 int tkn_count = 0, ret;
3221 struct snd_soc_tplg_vendor_string_elem *str_elem;
3222
3223 str_elem = (struct snd_soc_tplg_vendor_string_elem *)array->value;
3224 while (tkn_count < array->num_elems) {
3225 ret = skl_tplg_fill_str_mfest_tkn(dev, str_elem, skl);
3226 str_elem++;
3227
3228 if (ret < 0)
3229 return ret;
3230
3231 tkn_count = tkn_count + ret;
3232 }
3233
3234 return tkn_count;
3235 }
3236
skl_tplg_manifest_fill_fmt(struct device * dev,struct skl_module_iface * fmt,struct snd_soc_tplg_vendor_value_elem * tkn_elem,u32 dir,int fmt_idx)3237 static int skl_tplg_manifest_fill_fmt(struct device *dev,
3238 struct skl_module_iface *fmt,
3239 struct snd_soc_tplg_vendor_value_elem *tkn_elem,
3240 u32 dir, int fmt_idx)
3241 {
3242 struct skl_module_pin_fmt *dst_fmt;
3243 struct skl_module_fmt *mod_fmt;
3244 int ret;
3245
3246 if (!fmt)
3247 return -EINVAL;
3248
3249 switch (dir) {
3250 case SKL_DIR_IN:
3251 dst_fmt = &fmt->inputs[fmt_idx];
3252 break;
3253
3254 case SKL_DIR_OUT:
3255 dst_fmt = &fmt->outputs[fmt_idx];
3256 break;
3257
3258 default:
3259 dev_err(dev, "Invalid direction: %d\n", dir);
3260 return -EINVAL;
3261 }
3262
3263 mod_fmt = &dst_fmt->fmt;
3264
3265 switch (tkn_elem->token) {
3266 case SKL_TKN_MM_U32_INTF_PIN_ID:
3267 dst_fmt->id = tkn_elem->value;
3268 break;
3269
3270 default:
3271 ret = skl_tplg_fill_fmt(dev, mod_fmt, tkn_elem->token,
3272 tkn_elem->value);
3273 if (ret < 0)
3274 return ret;
3275 break;
3276 }
3277
3278 return 0;
3279 }
3280
skl_tplg_fill_mod_info(struct device * dev,struct snd_soc_tplg_vendor_value_elem * tkn_elem,struct skl_module * mod)3281 static int skl_tplg_fill_mod_info(struct device *dev,
3282 struct snd_soc_tplg_vendor_value_elem *tkn_elem,
3283 struct skl_module *mod)
3284 {
3285
3286 if (!mod)
3287 return -EINVAL;
3288
3289 switch (tkn_elem->token) {
3290 case SKL_TKN_U8_IN_PIN_TYPE:
3291 mod->input_pin_type = tkn_elem->value;
3292 break;
3293
3294 case SKL_TKN_U8_OUT_PIN_TYPE:
3295 mod->output_pin_type = tkn_elem->value;
3296 break;
3297
3298 case SKL_TKN_U8_IN_QUEUE_COUNT:
3299 mod->max_input_pins = tkn_elem->value;
3300 break;
3301
3302 case SKL_TKN_U8_OUT_QUEUE_COUNT:
3303 mod->max_output_pins = tkn_elem->value;
3304 break;
3305
3306 case SKL_TKN_MM_U8_NUM_RES:
3307 mod->nr_resources = tkn_elem->value;
3308 break;
3309
3310 case SKL_TKN_MM_U8_NUM_INTF:
3311 mod->nr_interfaces = tkn_elem->value;
3312 break;
3313
3314 default:
3315 dev_err(dev, "Invalid mod info token %d", tkn_elem->token);
3316 return -EINVAL;
3317 }
3318
3319 return 0;
3320 }
3321
3322
skl_tplg_get_int_tkn(struct device * dev,struct snd_soc_tplg_vendor_value_elem * tkn_elem,struct skl_dev * skl)3323 static int skl_tplg_get_int_tkn(struct device *dev,
3324 struct snd_soc_tplg_vendor_value_elem *tkn_elem,
3325 struct skl_dev *skl)
3326 {
3327 int tkn_count = 0, ret;
3328 static int mod_idx, res_val_idx, intf_val_idx, dir, pin_idx;
3329 struct skl_module_res *res = NULL;
3330 struct skl_module_iface *fmt = NULL;
3331 struct skl_module *mod = NULL;
3332 static struct skl_astate_param *astate_table;
3333 static int astate_cfg_idx, count;
3334 int i;
3335 size_t size;
3336
3337 if (skl->modules) {
3338 mod = skl->modules[mod_idx];
3339 res = &mod->resources[res_val_idx];
3340 fmt = &mod->formats[intf_val_idx];
3341 }
3342
3343 switch (tkn_elem->token) {
3344 case SKL_TKN_U32_LIB_COUNT:
3345 skl->lib_count = tkn_elem->value;
3346 break;
3347
3348 case SKL_TKN_U8_NUM_MOD:
3349 skl->nr_modules = tkn_elem->value;
3350 skl->modules = devm_kcalloc(dev, skl->nr_modules,
3351 sizeof(*skl->modules), GFP_KERNEL);
3352 if (!skl->modules)
3353 return -ENOMEM;
3354
3355 for (i = 0; i < skl->nr_modules; i++) {
3356 skl->modules[i] = devm_kzalloc(dev,
3357 sizeof(struct skl_module), GFP_KERNEL);
3358 if (!skl->modules[i])
3359 return -ENOMEM;
3360 }
3361 break;
3362
3363 case SKL_TKN_MM_U8_MOD_IDX:
3364 mod_idx = tkn_elem->value;
3365 break;
3366
3367 case SKL_TKN_U32_ASTATE_COUNT:
3368 if (astate_table != NULL) {
3369 dev_err(dev, "More than one entry for A-State count");
3370 return -EINVAL;
3371 }
3372
3373 if (tkn_elem->value > SKL_MAX_ASTATE_CFG) {
3374 dev_err(dev, "Invalid A-State count %d\n",
3375 tkn_elem->value);
3376 return -EINVAL;
3377 }
3378
3379 size = struct_size(skl->cfg.astate_cfg, astate_table,
3380 tkn_elem->value);
3381 skl->cfg.astate_cfg = devm_kzalloc(dev, size, GFP_KERNEL);
3382 if (!skl->cfg.astate_cfg)
3383 return -ENOMEM;
3384
3385 astate_table = skl->cfg.astate_cfg->astate_table;
3386 count = skl->cfg.astate_cfg->count = tkn_elem->value;
3387 break;
3388
3389 case SKL_TKN_U32_ASTATE_IDX:
3390 if (tkn_elem->value >= count) {
3391 dev_err(dev, "Invalid A-State index %d\n",
3392 tkn_elem->value);
3393 return -EINVAL;
3394 }
3395
3396 astate_cfg_idx = tkn_elem->value;
3397 break;
3398
3399 case SKL_TKN_U32_ASTATE_KCPS:
3400 astate_table[astate_cfg_idx].kcps = tkn_elem->value;
3401 break;
3402
3403 case SKL_TKN_U32_ASTATE_CLK_SRC:
3404 astate_table[astate_cfg_idx].clk_src = tkn_elem->value;
3405 break;
3406
3407 case SKL_TKN_U8_IN_PIN_TYPE:
3408 case SKL_TKN_U8_OUT_PIN_TYPE:
3409 case SKL_TKN_U8_IN_QUEUE_COUNT:
3410 case SKL_TKN_U8_OUT_QUEUE_COUNT:
3411 case SKL_TKN_MM_U8_NUM_RES:
3412 case SKL_TKN_MM_U8_NUM_INTF:
3413 ret = skl_tplg_fill_mod_info(dev, tkn_elem, mod);
3414 if (ret < 0)
3415 return ret;
3416 break;
3417
3418 case SKL_TKN_U32_DIR_PIN_COUNT:
3419 dir = tkn_elem->value & SKL_IN_DIR_BIT_MASK;
3420 pin_idx = (tkn_elem->value & SKL_PIN_COUNT_MASK) >> 4;
3421 break;
3422
3423 case SKL_TKN_MM_U32_RES_ID:
3424 if (!res)
3425 return -EINVAL;
3426
3427 res->id = tkn_elem->value;
3428 res_val_idx = tkn_elem->value;
3429 break;
3430
3431 case SKL_TKN_MM_U32_FMT_ID:
3432 if (!fmt)
3433 return -EINVAL;
3434
3435 fmt->fmt_idx = tkn_elem->value;
3436 intf_val_idx = tkn_elem->value;
3437 break;
3438
3439 case SKL_TKN_MM_U32_CPS:
3440 case SKL_TKN_MM_U32_DMA_SIZE:
3441 case SKL_TKN_MM_U32_CPC:
3442 case SKL_TKN_U32_MEM_PAGES:
3443 case SKL_TKN_U32_OBS:
3444 case SKL_TKN_U32_IBS:
3445 case SKL_TKN_MM_U32_RES_PIN_ID:
3446 case SKL_TKN_MM_U32_PIN_BUF:
3447 ret = skl_tplg_fill_res_tkn(dev, tkn_elem, res, pin_idx, dir);
3448 if (ret < 0)
3449 return ret;
3450
3451 break;
3452
3453 case SKL_TKN_MM_U32_NUM_IN_FMT:
3454 if (!fmt)
3455 return -EINVAL;
3456
3457 res->nr_input_pins = tkn_elem->value;
3458 break;
3459
3460 case SKL_TKN_MM_U32_NUM_OUT_FMT:
3461 if (!fmt)
3462 return -EINVAL;
3463
3464 res->nr_output_pins = tkn_elem->value;
3465 break;
3466
3467 case SKL_TKN_U32_FMT_CH:
3468 case SKL_TKN_U32_FMT_FREQ:
3469 case SKL_TKN_U32_FMT_BIT_DEPTH:
3470 case SKL_TKN_U32_FMT_SAMPLE_SIZE:
3471 case SKL_TKN_U32_FMT_CH_CONFIG:
3472 case SKL_TKN_U32_FMT_INTERLEAVE:
3473 case SKL_TKN_U32_FMT_SAMPLE_TYPE:
3474 case SKL_TKN_U32_FMT_CH_MAP:
3475 case SKL_TKN_MM_U32_INTF_PIN_ID:
3476 ret = skl_tplg_manifest_fill_fmt(dev, fmt, tkn_elem,
3477 dir, pin_idx);
3478 if (ret < 0)
3479 return ret;
3480 break;
3481
3482 default:
3483 dev_err(dev, "Not a manifest token %d\n", tkn_elem->token);
3484 return -EINVAL;
3485 }
3486 tkn_count++;
3487
3488 return tkn_count;
3489 }
3490
3491 /*
3492 * Fill the manifest structure by parsing the tokens based on the
3493 * type.
3494 */
skl_tplg_get_manifest_tkn(struct device * dev,char * pvt_data,struct skl_dev * skl,int block_size)3495 static int skl_tplg_get_manifest_tkn(struct device *dev,
3496 char *pvt_data, struct skl_dev *skl,
3497 int block_size)
3498 {
3499 int tkn_count = 0, ret;
3500 int off = 0, tuple_size = 0;
3501 u8 uuid_index = 0;
3502 struct snd_soc_tplg_vendor_array *array;
3503 struct snd_soc_tplg_vendor_value_elem *tkn_elem;
3504
3505 if (block_size <= 0)
3506 return -EINVAL;
3507
3508 while (tuple_size < block_size) {
3509 array = (struct snd_soc_tplg_vendor_array *)(pvt_data + off);
3510 off += array->size;
3511 switch (array->type) {
3512 case SND_SOC_TPLG_TUPLE_TYPE_STRING:
3513 ret = skl_tplg_get_str_tkn(dev, array, skl);
3514
3515 if (ret < 0)
3516 return ret;
3517 tkn_count = ret;
3518
3519 tuple_size += tkn_count *
3520 sizeof(struct snd_soc_tplg_vendor_string_elem);
3521 continue;
3522
3523 case SND_SOC_TPLG_TUPLE_TYPE_UUID:
3524 if (array->uuid->token != SKL_TKN_UUID) {
3525 dev_err(dev, "Not an UUID token: %d\n",
3526 array->uuid->token);
3527 return -EINVAL;
3528 }
3529 if (uuid_index >= skl->nr_modules) {
3530 dev_err(dev, "Too many UUID tokens\n");
3531 return -EINVAL;
3532 }
3533 import_guid(&skl->modules[uuid_index++]->uuid,
3534 array->uuid->uuid);
3535
3536 tuple_size += sizeof(*array->uuid);
3537 continue;
3538
3539 default:
3540 tkn_elem = array->value;
3541 tkn_count = 0;
3542 break;
3543 }
3544
3545 while (tkn_count <= array->num_elems - 1) {
3546 ret = skl_tplg_get_int_tkn(dev,
3547 tkn_elem, skl);
3548 if (ret < 0)
3549 return ret;
3550
3551 tkn_count = tkn_count + ret;
3552 tkn_elem++;
3553 }
3554 tuple_size += (tkn_count * sizeof(*tkn_elem));
3555 tkn_count = 0;
3556 }
3557
3558 return off;
3559 }
3560
3561 /*
3562 * Parse manifest private data for tokens. The private data block is
3563 * preceded by descriptors for type and size of data block.
3564 */
skl_tplg_get_manifest_data(struct snd_soc_tplg_manifest * manifest,struct device * dev,struct skl_dev * skl)3565 static int skl_tplg_get_manifest_data(struct snd_soc_tplg_manifest *manifest,
3566 struct device *dev, struct skl_dev *skl)
3567 {
3568 struct snd_soc_tplg_vendor_array *array;
3569 int num_blocks, block_size = 0, block_type, off = 0;
3570 char *data;
3571 int ret;
3572
3573 /* Read the NUM_DATA_BLOCKS descriptor */
3574 array = (struct snd_soc_tplg_vendor_array *)manifest->priv.data;
3575 ret = skl_tplg_get_desc_blocks(dev, array);
3576 if (ret < 0)
3577 return ret;
3578 num_blocks = ret;
3579
3580 off += array->size;
3581 /* Read the BLOCK_TYPE and BLOCK_SIZE descriptor */
3582 while (num_blocks > 0) {
3583 array = (struct snd_soc_tplg_vendor_array *)
3584 (manifest->priv.data + off);
3585 ret = skl_tplg_get_desc_blocks(dev, array);
3586
3587 if (ret < 0)
3588 return ret;
3589 block_type = ret;
3590 off += array->size;
3591
3592 array = (struct snd_soc_tplg_vendor_array *)
3593 (manifest->priv.data + off);
3594
3595 ret = skl_tplg_get_desc_blocks(dev, array);
3596
3597 if (ret < 0)
3598 return ret;
3599 block_size = ret;
3600 off += array->size;
3601
3602 array = (struct snd_soc_tplg_vendor_array *)
3603 (manifest->priv.data + off);
3604
3605 data = (manifest->priv.data + off);
3606
3607 if (block_type == SKL_TYPE_TUPLE) {
3608 ret = skl_tplg_get_manifest_tkn(dev, data, skl,
3609 block_size);
3610
3611 if (ret < 0)
3612 return ret;
3613
3614 --num_blocks;
3615 } else {
3616 return -EINVAL;
3617 }
3618 off += ret;
3619 }
3620
3621 return 0;
3622 }
3623
skl_manifest_load(struct snd_soc_component * cmpnt,int index,struct snd_soc_tplg_manifest * manifest)3624 static int skl_manifest_load(struct snd_soc_component *cmpnt, int index,
3625 struct snd_soc_tplg_manifest *manifest)
3626 {
3627 struct hdac_bus *bus = snd_soc_component_get_drvdata(cmpnt);
3628 struct skl_dev *skl = bus_to_skl(bus);
3629
3630 /* proceed only if we have private data defined */
3631 if (manifest->priv.size == 0)
3632 return 0;
3633
3634 skl_tplg_get_manifest_data(manifest, bus->dev, skl);
3635
3636 if (skl->lib_count > SKL_MAX_LIB) {
3637 dev_err(bus->dev, "Exceeding max Library count. Got:%d\n",
3638 skl->lib_count);
3639 return -EINVAL;
3640 }
3641
3642 return 0;
3643 }
3644
skl_tplg_complete(struct snd_soc_component * component)3645 static int skl_tplg_complete(struct snd_soc_component *component)
3646 {
3647 struct snd_soc_dobj *dobj;
3648 struct snd_soc_acpi_mach *mach;
3649 struct snd_ctl_elem_value *val;
3650 int i;
3651
3652 val = kmalloc(sizeof(*val), GFP_KERNEL);
3653 if (!val)
3654 return -ENOMEM;
3655
3656 mach = dev_get_platdata(component->card->dev);
3657 list_for_each_entry(dobj, &component->dobj_list, list) {
3658 struct snd_kcontrol *kcontrol = dobj->control.kcontrol;
3659 struct soc_enum *se;
3660 char **texts;
3661 char chan_text[4];
3662
3663 if (dobj->type != SND_SOC_DOBJ_ENUM || !kcontrol ||
3664 kcontrol->put != skl_tplg_multi_config_set_dmic)
3665 continue;
3666
3667 se = (struct soc_enum *)kcontrol->private_value;
3668 texts = dobj->control.dtexts;
3669 sprintf(chan_text, "c%d", mach->mach_params.dmic_num);
3670
3671 for (i = 0; i < se->items; i++) {
3672 if (strstr(texts[i], chan_text)) {
3673 memset(val, 0, sizeof(*val));
3674 val->value.enumerated.item[0] = i;
3675 kcontrol->put(kcontrol, val);
3676 }
3677 }
3678 }
3679
3680 kfree(val);
3681 return 0;
3682 }
3683
3684 static struct snd_soc_tplg_ops skl_tplg_ops = {
3685 .widget_load = skl_tplg_widget_load,
3686 .control_load = skl_tplg_control_load,
3687 .bytes_ext_ops = skl_tlv_ops,
3688 .bytes_ext_ops_count = ARRAY_SIZE(skl_tlv_ops),
3689 .io_ops = skl_tplg_kcontrol_ops,
3690 .io_ops_count = ARRAY_SIZE(skl_tplg_kcontrol_ops),
3691 .manifest = skl_manifest_load,
3692 .dai_load = skl_dai_load,
3693 .complete = skl_tplg_complete,
3694 };
3695
3696 /*
3697 * A pipe can have multiple modules, each of them will be a DAPM widget as
3698 * well. While managing a pipeline we need to get the list of all the
3699 * widgets in a pipelines, so this helper - skl_tplg_create_pipe_widget_list()
3700 * helps to get the SKL type widgets in that pipeline
3701 */
skl_tplg_create_pipe_widget_list(struct snd_soc_component * component)3702 static int skl_tplg_create_pipe_widget_list(struct snd_soc_component *component)
3703 {
3704 struct snd_soc_dapm_widget *w;
3705 struct skl_module_cfg *mcfg = NULL;
3706 struct skl_pipe_module *p_module = NULL;
3707 struct skl_pipe *pipe;
3708
3709 list_for_each_entry(w, &component->card->widgets, list) {
3710 if (is_skl_dsp_widget_type(w, component->dev) && w->priv) {
3711 mcfg = w->priv;
3712 pipe = mcfg->pipe;
3713
3714 p_module = devm_kzalloc(component->dev,
3715 sizeof(*p_module), GFP_KERNEL);
3716 if (!p_module)
3717 return -ENOMEM;
3718
3719 p_module->w = w;
3720 list_add_tail(&p_module->node, &pipe->w_list);
3721 }
3722 }
3723
3724 return 0;
3725 }
3726
skl_tplg_set_pipe_type(struct skl_dev * skl,struct skl_pipe * pipe)3727 static void skl_tplg_set_pipe_type(struct skl_dev *skl, struct skl_pipe *pipe)
3728 {
3729 struct skl_pipe_module *w_module;
3730 struct snd_soc_dapm_widget *w;
3731 struct skl_module_cfg *mconfig;
3732 bool host_found = false, link_found = false;
3733
3734 list_for_each_entry(w_module, &pipe->w_list, node) {
3735 w = w_module->w;
3736 mconfig = w->priv;
3737
3738 if (mconfig->dev_type == SKL_DEVICE_HDAHOST)
3739 host_found = true;
3740 else if (mconfig->dev_type != SKL_DEVICE_NONE)
3741 link_found = true;
3742 }
3743
3744 if (host_found && link_found)
3745 pipe->passthru = true;
3746 else
3747 pipe->passthru = false;
3748 }
3749
3750 /*
3751 * SKL topology init routine
3752 */
skl_tplg_init(struct snd_soc_component * component,struct hdac_bus * bus)3753 int skl_tplg_init(struct snd_soc_component *component, struct hdac_bus *bus)
3754 {
3755 int ret;
3756 const struct firmware *fw;
3757 struct skl_dev *skl = bus_to_skl(bus);
3758 struct skl_pipeline *ppl;
3759
3760 ret = request_firmware(&fw, skl->tplg_name, bus->dev);
3761 if (ret < 0) {
3762 char alt_tplg_name[64];
3763
3764 snprintf(alt_tplg_name, sizeof(alt_tplg_name), "%s-tplg.bin",
3765 skl->mach->drv_name);
3766 dev_info(bus->dev, "tplg fw %s load failed with %d, trying alternative tplg name %s",
3767 skl->tplg_name, ret, alt_tplg_name);
3768
3769 ret = request_firmware(&fw, alt_tplg_name, bus->dev);
3770 if (!ret)
3771 goto component_load;
3772
3773 dev_info(bus->dev, "tplg %s failed with %d, falling back to dfw_sst.bin",
3774 alt_tplg_name, ret);
3775
3776 ret = request_firmware(&fw, "dfw_sst.bin", bus->dev);
3777 if (ret < 0) {
3778 dev_err(bus->dev, "Fallback tplg fw %s load failed with %d\n",
3779 "dfw_sst.bin", ret);
3780 return ret;
3781 }
3782 }
3783
3784 component_load:
3785 ret = snd_soc_tplg_component_load(component, &skl_tplg_ops, fw);
3786 if (ret < 0) {
3787 dev_err(bus->dev, "tplg component load failed%d\n", ret);
3788 goto err;
3789 }
3790
3791 ret = skl_tplg_create_pipe_widget_list(component);
3792 if (ret < 0) {
3793 dev_err(bus->dev, "tplg create pipe widget list failed%d\n",
3794 ret);
3795 goto err;
3796 }
3797
3798 list_for_each_entry(ppl, &skl->ppl_list, node)
3799 skl_tplg_set_pipe_type(skl, ppl->pipe);
3800
3801 err:
3802 release_firmware(fw);
3803 return ret;
3804 }
3805
skl_tplg_exit(struct snd_soc_component * component,struct hdac_bus * bus)3806 void skl_tplg_exit(struct snd_soc_component *component, struct hdac_bus *bus)
3807 {
3808 struct skl_dev *skl = bus_to_skl(bus);
3809 struct skl_pipeline *ppl, *tmp;
3810
3811 list_for_each_entry_safe(ppl, tmp, &skl->ppl_list, node)
3812 list_del(&ppl->node);
3813
3814 /* clean up topology */
3815 snd_soc_tplg_component_remove(component);
3816 }
3817