1 /*
2 * Copyright 2015 Advanced Micro Devices, Inc.
3 *
4 * Permission is hereby granted, free of charge, to any person obtaining a
5 * copy of this software and associated documentation files (the "Software"),
6 * to deal in the Software without restriction, including without limitation
7 * the rights to use, copy, modify, merge, publish, distribute, sublicense,
8 * and/or sell copies of the Software, and to permit persons to whom the
9 * Software is furnished to do so, subject to the following conditions:
10 *
11 * The above copyright notice and this permission notice shall be included in
12 * all copies or substantial portions of the Software.
13 *
14 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
15 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
16 * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
17 * THE COPYRIGHT HOLDER(S) OR AUTHOR(S) BE LIABLE FOR ANY CLAIM, DAMAGES OR
18 * OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE,
19 * ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR
20 * OTHER DEALINGS IN THE SOFTWARE.
21 *
22 */
23
24 #include <linux/delay.h>
25 #include <linux/gfp.h>
26 #include <linux/kernel.h>
27 #include <linux/ktime.h>
28 #include <linux/slab.h>
29 #include <linux/types.h>
30
31 #include "cgs_common.h"
32 #include "smu/smu_8_0_d.h"
33 #include "smu/smu_8_0_sh_mask.h"
34 #include "smu8.h"
35 #include "smu8_fusion.h"
36 #include "smu8_smumgr.h"
37 #include "cz_ppsmc.h"
38 #include "smu_ucode_xfer_cz.h"
39 #include "gca/gfx_8_0_d.h"
40 #include "gca/gfx_8_0_sh_mask.h"
41 #include "smumgr.h"
42
43 #define SIZE_ALIGN_32(x) (((x) + 31) / 32 * 32)
44
45 static const enum smu8_scratch_entry firmware_list[] = {
46 SMU8_SCRATCH_ENTRY_UCODE_ID_SDMA0,
47 SMU8_SCRATCH_ENTRY_UCODE_ID_SDMA1,
48 SMU8_SCRATCH_ENTRY_UCODE_ID_CP_CE,
49 SMU8_SCRATCH_ENTRY_UCODE_ID_CP_PFP,
50 SMU8_SCRATCH_ENTRY_UCODE_ID_CP_ME,
51 SMU8_SCRATCH_ENTRY_UCODE_ID_CP_MEC_JT1,
52 SMU8_SCRATCH_ENTRY_UCODE_ID_CP_MEC_JT2,
53 SMU8_SCRATCH_ENTRY_UCODE_ID_RLC_G,
54 };
55
smu8_get_argument(struct pp_hwmgr * hwmgr)56 static uint32_t smu8_get_argument(struct pp_hwmgr *hwmgr)
57 {
58 if (hwmgr == NULL || hwmgr->device == NULL)
59 return 0;
60
61 return cgs_read_register(hwmgr->device,
62 mmSMU_MP1_SRBM2P_ARG_0);
63 }
64
65 /* Send a message to the SMC, and wait for its response.*/
smu8_send_msg_to_smc_with_parameter(struct pp_hwmgr * hwmgr,uint16_t msg,uint32_t parameter)66 static int smu8_send_msg_to_smc_with_parameter(struct pp_hwmgr *hwmgr,
67 uint16_t msg, uint32_t parameter)
68 {
69 int result = 0;
70 ktime_t t_start;
71 s64 elapsed_us;
72
73 if (hwmgr == NULL || hwmgr->device == NULL)
74 return -EINVAL;
75
76 result = PHM_WAIT_FIELD_UNEQUAL(hwmgr,
77 SMU_MP1_SRBM2P_RESP_0, CONTENT, 0);
78 if (result != 0) {
79 /* Read the last message to SMU, to report actual cause */
80 uint32_t val = cgs_read_register(hwmgr->device,
81 mmSMU_MP1_SRBM2P_MSG_0);
82 pr_err("%s(0x%04x) aborted; SMU still servicing msg (0x%04x)\n",
83 __func__, msg, val);
84 return result;
85 }
86 t_start = ktime_get();
87
88 cgs_write_register(hwmgr->device, mmSMU_MP1_SRBM2P_ARG_0, parameter);
89
90 cgs_write_register(hwmgr->device, mmSMU_MP1_SRBM2P_RESP_0, 0);
91 cgs_write_register(hwmgr->device, mmSMU_MP1_SRBM2P_MSG_0, msg);
92
93 result = PHM_WAIT_FIELD_UNEQUAL(hwmgr,
94 SMU_MP1_SRBM2P_RESP_0, CONTENT, 0);
95
96 elapsed_us = ktime_us_delta(ktime_get(), t_start);
97
98 WARN(result, "%s(0x%04x, %#x) timed out after %lld us\n",
99 __func__, msg, parameter, elapsed_us);
100
101 return result;
102 }
103
smu8_send_msg_to_smc(struct pp_hwmgr * hwmgr,uint16_t msg)104 static int smu8_send_msg_to_smc(struct pp_hwmgr *hwmgr, uint16_t msg)
105 {
106 return smu8_send_msg_to_smc_with_parameter(hwmgr, msg, 0);
107 }
108
smu8_set_smc_sram_address(struct pp_hwmgr * hwmgr,uint32_t smc_address,uint32_t limit)109 static int smu8_set_smc_sram_address(struct pp_hwmgr *hwmgr,
110 uint32_t smc_address, uint32_t limit)
111 {
112 if (hwmgr == NULL || hwmgr->device == NULL)
113 return -EINVAL;
114
115 if (0 != (3 & smc_address)) {
116 pr_err("SMC address must be 4 byte aligned\n");
117 return -EINVAL;
118 }
119
120 if (limit <= (smc_address + 3)) {
121 pr_err("SMC address beyond the SMC RAM area\n");
122 return -EINVAL;
123 }
124
125 cgs_write_register(hwmgr->device, mmMP0PUB_IND_INDEX_0,
126 SMN_MP1_SRAM_START_ADDR + smc_address);
127
128 return 0;
129 }
130
smu8_write_smc_sram_dword(struct pp_hwmgr * hwmgr,uint32_t smc_address,uint32_t value,uint32_t limit)131 static int smu8_write_smc_sram_dword(struct pp_hwmgr *hwmgr,
132 uint32_t smc_address, uint32_t value, uint32_t limit)
133 {
134 int result;
135
136 if (hwmgr == NULL || hwmgr->device == NULL)
137 return -EINVAL;
138
139 result = smu8_set_smc_sram_address(hwmgr, smc_address, limit);
140 if (!result)
141 cgs_write_register(hwmgr->device, mmMP0PUB_IND_DATA_0, value);
142
143 return result;
144 }
145
smu8_check_fw_load_finish(struct pp_hwmgr * hwmgr,uint32_t firmware)146 static int smu8_check_fw_load_finish(struct pp_hwmgr *hwmgr,
147 uint32_t firmware)
148 {
149 int i;
150 uint32_t index = SMN_MP1_SRAM_START_ADDR +
151 SMU8_FIRMWARE_HEADER_LOCATION +
152 offsetof(struct SMU8_Firmware_Header, UcodeLoadStatus);
153
154 if (hwmgr == NULL || hwmgr->device == NULL)
155 return -EINVAL;
156
157 cgs_write_register(hwmgr->device, mmMP0PUB_IND_INDEX, index);
158
159 for (i = 0; i < hwmgr->usec_timeout; i++) {
160 if (firmware ==
161 (cgs_read_register(hwmgr->device, mmMP0PUB_IND_DATA) & firmware))
162 break;
163 udelay(1);
164 }
165
166 if (i >= hwmgr->usec_timeout) {
167 pr_err("SMU check loaded firmware failed.\n");
168 return -EINVAL;
169 }
170
171 return 0;
172 }
173
smu8_load_mec_firmware(struct pp_hwmgr * hwmgr)174 static int smu8_load_mec_firmware(struct pp_hwmgr *hwmgr)
175 {
176 uint32_t reg_data;
177 uint32_t tmp;
178 int ret = 0;
179 struct cgs_firmware_info info = {0};
180
181 if (hwmgr == NULL || hwmgr->device == NULL)
182 return -EINVAL;
183
184 ret = cgs_get_firmware_info(hwmgr->device,
185 CGS_UCODE_ID_CP_MEC, &info);
186
187 if (ret)
188 return -EINVAL;
189
190 /* Disable MEC parsing/prefetching */
191 tmp = cgs_read_register(hwmgr->device,
192 mmCP_MEC_CNTL);
193 tmp = PHM_SET_FIELD(tmp, CP_MEC_CNTL, MEC_ME1_HALT, 1);
194 tmp = PHM_SET_FIELD(tmp, CP_MEC_CNTL, MEC_ME2_HALT, 1);
195 cgs_write_register(hwmgr->device, mmCP_MEC_CNTL, tmp);
196
197 tmp = cgs_read_register(hwmgr->device,
198 mmCP_CPC_IC_BASE_CNTL);
199
200 tmp = PHM_SET_FIELD(tmp, CP_CPC_IC_BASE_CNTL, VMID, 0);
201 tmp = PHM_SET_FIELD(tmp, CP_CPC_IC_BASE_CNTL, ATC, 0);
202 tmp = PHM_SET_FIELD(tmp, CP_CPC_IC_BASE_CNTL, CACHE_POLICY, 0);
203 tmp = PHM_SET_FIELD(tmp, CP_CPC_IC_BASE_CNTL, MTYPE, 1);
204 cgs_write_register(hwmgr->device, mmCP_CPC_IC_BASE_CNTL, tmp);
205
206 reg_data = lower_32_bits(info.mc_addr) &
207 PHM_FIELD_MASK(CP_CPC_IC_BASE_LO, IC_BASE_LO);
208 cgs_write_register(hwmgr->device, mmCP_CPC_IC_BASE_LO, reg_data);
209
210 reg_data = upper_32_bits(info.mc_addr) &
211 PHM_FIELD_MASK(CP_CPC_IC_BASE_HI, IC_BASE_HI);
212 cgs_write_register(hwmgr->device, mmCP_CPC_IC_BASE_HI, reg_data);
213
214 return 0;
215 }
216
smu8_translate_firmware_enum_to_arg(struct pp_hwmgr * hwmgr,enum smu8_scratch_entry firmware_enum)217 static uint8_t smu8_translate_firmware_enum_to_arg(struct pp_hwmgr *hwmgr,
218 enum smu8_scratch_entry firmware_enum)
219 {
220 uint8_t ret = 0;
221
222 switch (firmware_enum) {
223 case SMU8_SCRATCH_ENTRY_UCODE_ID_SDMA0:
224 ret = UCODE_ID_SDMA0;
225 break;
226 case SMU8_SCRATCH_ENTRY_UCODE_ID_SDMA1:
227 if (hwmgr->chip_id == CHIP_STONEY)
228 ret = UCODE_ID_SDMA0;
229 else
230 ret = UCODE_ID_SDMA1;
231 break;
232 case SMU8_SCRATCH_ENTRY_UCODE_ID_CP_CE:
233 ret = UCODE_ID_CP_CE;
234 break;
235 case SMU8_SCRATCH_ENTRY_UCODE_ID_CP_PFP:
236 ret = UCODE_ID_CP_PFP;
237 break;
238 case SMU8_SCRATCH_ENTRY_UCODE_ID_CP_ME:
239 ret = UCODE_ID_CP_ME;
240 break;
241 case SMU8_SCRATCH_ENTRY_UCODE_ID_CP_MEC_JT1:
242 ret = UCODE_ID_CP_MEC_JT1;
243 break;
244 case SMU8_SCRATCH_ENTRY_UCODE_ID_CP_MEC_JT2:
245 if (hwmgr->chip_id == CHIP_STONEY)
246 ret = UCODE_ID_CP_MEC_JT1;
247 else
248 ret = UCODE_ID_CP_MEC_JT2;
249 break;
250 case SMU8_SCRATCH_ENTRY_UCODE_ID_GMCON_RENG:
251 ret = UCODE_ID_GMCON_RENG;
252 break;
253 case SMU8_SCRATCH_ENTRY_UCODE_ID_RLC_G:
254 ret = UCODE_ID_RLC_G;
255 break;
256 case SMU8_SCRATCH_ENTRY_UCODE_ID_RLC_SCRATCH:
257 ret = UCODE_ID_RLC_SCRATCH;
258 break;
259 case SMU8_SCRATCH_ENTRY_UCODE_ID_RLC_SRM_ARAM:
260 ret = UCODE_ID_RLC_SRM_ARAM;
261 break;
262 case SMU8_SCRATCH_ENTRY_UCODE_ID_RLC_SRM_DRAM:
263 ret = UCODE_ID_RLC_SRM_DRAM;
264 break;
265 case SMU8_SCRATCH_ENTRY_UCODE_ID_DMCU_ERAM:
266 ret = UCODE_ID_DMCU_ERAM;
267 break;
268 case SMU8_SCRATCH_ENTRY_UCODE_ID_DMCU_IRAM:
269 ret = UCODE_ID_DMCU_IRAM;
270 break;
271 case SMU8_SCRATCH_ENTRY_UCODE_ID_POWER_PROFILING:
272 ret = TASK_ARG_INIT_MM_PWR_LOG;
273 break;
274 case SMU8_SCRATCH_ENTRY_DATA_ID_SDMA_HALT:
275 case SMU8_SCRATCH_ENTRY_DATA_ID_SYS_CLOCKGATING:
276 case SMU8_SCRATCH_ENTRY_DATA_ID_SDMA_RING_REGS:
277 case SMU8_SCRATCH_ENTRY_DATA_ID_NONGFX_REINIT:
278 case SMU8_SCRATCH_ENTRY_DATA_ID_SDMA_START:
279 case SMU8_SCRATCH_ENTRY_DATA_ID_IH_REGISTERS:
280 ret = TASK_ARG_REG_MMIO;
281 break;
282 case SMU8_SCRATCH_ENTRY_SMU8_FUSION_CLKTABLE:
283 ret = TASK_ARG_INIT_CLK_TABLE;
284 break;
285 }
286
287 return ret;
288 }
289
smu8_convert_fw_type_to_cgs(uint32_t fw_type)290 static enum cgs_ucode_id smu8_convert_fw_type_to_cgs(uint32_t fw_type)
291 {
292 enum cgs_ucode_id result = CGS_UCODE_ID_MAXIMUM;
293
294 switch (fw_type) {
295 case UCODE_ID_SDMA0:
296 result = CGS_UCODE_ID_SDMA0;
297 break;
298 case UCODE_ID_SDMA1:
299 result = CGS_UCODE_ID_SDMA1;
300 break;
301 case UCODE_ID_CP_CE:
302 result = CGS_UCODE_ID_CP_CE;
303 break;
304 case UCODE_ID_CP_PFP:
305 result = CGS_UCODE_ID_CP_PFP;
306 break;
307 case UCODE_ID_CP_ME:
308 result = CGS_UCODE_ID_CP_ME;
309 break;
310 case UCODE_ID_CP_MEC_JT1:
311 result = CGS_UCODE_ID_CP_MEC_JT1;
312 break;
313 case UCODE_ID_CP_MEC_JT2:
314 result = CGS_UCODE_ID_CP_MEC_JT2;
315 break;
316 case UCODE_ID_RLC_G:
317 result = CGS_UCODE_ID_RLC_G;
318 break;
319 default:
320 break;
321 }
322
323 return result;
324 }
325
smu8_smu_populate_single_scratch_task(struct pp_hwmgr * hwmgr,enum smu8_scratch_entry fw_enum,uint8_t type,bool is_last)326 static int smu8_smu_populate_single_scratch_task(
327 struct pp_hwmgr *hwmgr,
328 enum smu8_scratch_entry fw_enum,
329 uint8_t type, bool is_last)
330 {
331 uint8_t i;
332 struct smu8_smumgr *smu8_smu = hwmgr->smu_backend;
333 struct TOC *toc = (struct TOC *)smu8_smu->toc_buffer.kaddr;
334 struct SMU_Task *task = &toc->tasks[smu8_smu->toc_entry_used_count++];
335
336 task->type = type;
337 task->arg = smu8_translate_firmware_enum_to_arg(hwmgr, fw_enum);
338 task->next = is_last ? END_OF_TASK_LIST : smu8_smu->toc_entry_used_count;
339
340 for (i = 0; i < smu8_smu->scratch_buffer_length; i++)
341 if (smu8_smu->scratch_buffer[i].firmware_ID == fw_enum)
342 break;
343
344 if (i >= smu8_smu->scratch_buffer_length) {
345 pr_err("Invalid Firmware Type\n");
346 return -EINVAL;
347 }
348
349 task->addr.low = lower_32_bits(smu8_smu->scratch_buffer[i].mc_addr);
350 task->addr.high = upper_32_bits(smu8_smu->scratch_buffer[i].mc_addr);
351 task->size_bytes = smu8_smu->scratch_buffer[i].data_size;
352
353 if (SMU8_SCRATCH_ENTRY_DATA_ID_IH_REGISTERS == fw_enum) {
354 struct smu8_ih_meta_data *pIHReg_restore =
355 (struct smu8_ih_meta_data *)smu8_smu->scratch_buffer[i].kaddr;
356 pIHReg_restore->command =
357 METADATA_CMD_MODE0 | METADATA_PERFORM_ON_LOAD;
358 }
359
360 return 0;
361 }
362
smu8_smu_populate_single_ucode_load_task(struct pp_hwmgr * hwmgr,enum smu8_scratch_entry fw_enum,bool is_last)363 static int smu8_smu_populate_single_ucode_load_task(
364 struct pp_hwmgr *hwmgr,
365 enum smu8_scratch_entry fw_enum,
366 bool is_last)
367 {
368 uint8_t i;
369 struct smu8_smumgr *smu8_smu = hwmgr->smu_backend;
370 struct TOC *toc = (struct TOC *)smu8_smu->toc_buffer.kaddr;
371 struct SMU_Task *task = &toc->tasks[smu8_smu->toc_entry_used_count++];
372
373 task->type = TASK_TYPE_UCODE_LOAD;
374 task->arg = smu8_translate_firmware_enum_to_arg(hwmgr, fw_enum);
375 task->next = is_last ? END_OF_TASK_LIST : smu8_smu->toc_entry_used_count;
376
377 for (i = 0; i < smu8_smu->driver_buffer_length; i++)
378 if (smu8_smu->driver_buffer[i].firmware_ID == fw_enum)
379 break;
380
381 if (i >= smu8_smu->driver_buffer_length) {
382 pr_err("Invalid Firmware Type\n");
383 return -EINVAL;
384 }
385
386 task->addr.low = lower_32_bits(smu8_smu->driver_buffer[i].mc_addr);
387 task->addr.high = upper_32_bits(smu8_smu->driver_buffer[i].mc_addr);
388 task->size_bytes = smu8_smu->driver_buffer[i].data_size;
389
390 return 0;
391 }
392
smu8_smu_construct_toc_for_rlc_aram_save(struct pp_hwmgr * hwmgr)393 static int smu8_smu_construct_toc_for_rlc_aram_save(struct pp_hwmgr *hwmgr)
394 {
395 struct smu8_smumgr *smu8_smu = hwmgr->smu_backend;
396
397 smu8_smu->toc_entry_aram = smu8_smu->toc_entry_used_count;
398 smu8_smu_populate_single_scratch_task(hwmgr,
399 SMU8_SCRATCH_ENTRY_UCODE_ID_RLC_SRM_ARAM,
400 TASK_TYPE_UCODE_SAVE, true);
401
402 return 0;
403 }
404
smu8_smu_initialize_toc_empty_job_list(struct pp_hwmgr * hwmgr)405 static int smu8_smu_initialize_toc_empty_job_list(struct pp_hwmgr *hwmgr)
406 {
407 int i;
408 struct smu8_smumgr *smu8_smu = hwmgr->smu_backend;
409 struct TOC *toc = (struct TOC *)smu8_smu->toc_buffer.kaddr;
410
411 for (i = 0; i < NUM_JOBLIST_ENTRIES; i++)
412 toc->JobList[i] = (uint8_t)IGNORE_JOB;
413
414 return 0;
415 }
416
smu8_smu_construct_toc_for_vddgfx_enter(struct pp_hwmgr * hwmgr)417 static int smu8_smu_construct_toc_for_vddgfx_enter(struct pp_hwmgr *hwmgr)
418 {
419 struct smu8_smumgr *smu8_smu = hwmgr->smu_backend;
420 struct TOC *toc = (struct TOC *)smu8_smu->toc_buffer.kaddr;
421
422 toc->JobList[JOB_GFX_SAVE] = (uint8_t)smu8_smu->toc_entry_used_count;
423 smu8_smu_populate_single_scratch_task(hwmgr,
424 SMU8_SCRATCH_ENTRY_UCODE_ID_RLC_SCRATCH,
425 TASK_TYPE_UCODE_SAVE, false);
426
427 smu8_smu_populate_single_scratch_task(hwmgr,
428 SMU8_SCRATCH_ENTRY_UCODE_ID_RLC_SRM_DRAM,
429 TASK_TYPE_UCODE_SAVE, true);
430
431 return 0;
432 }
433
434
smu8_smu_construct_toc_for_vddgfx_exit(struct pp_hwmgr * hwmgr)435 static int smu8_smu_construct_toc_for_vddgfx_exit(struct pp_hwmgr *hwmgr)
436 {
437 struct smu8_smumgr *smu8_smu = hwmgr->smu_backend;
438 struct TOC *toc = (struct TOC *)smu8_smu->toc_buffer.kaddr;
439
440 toc->JobList[JOB_GFX_RESTORE] = (uint8_t)smu8_smu->toc_entry_used_count;
441
442 smu8_smu_populate_single_ucode_load_task(hwmgr,
443 SMU8_SCRATCH_ENTRY_UCODE_ID_CP_CE, false);
444 smu8_smu_populate_single_ucode_load_task(hwmgr,
445 SMU8_SCRATCH_ENTRY_UCODE_ID_CP_PFP, false);
446 smu8_smu_populate_single_ucode_load_task(hwmgr,
447 SMU8_SCRATCH_ENTRY_UCODE_ID_CP_ME, false);
448 smu8_smu_populate_single_ucode_load_task(hwmgr,
449 SMU8_SCRATCH_ENTRY_UCODE_ID_CP_MEC_JT1, false);
450
451 if (hwmgr->chip_id == CHIP_STONEY)
452 smu8_smu_populate_single_ucode_load_task(hwmgr,
453 SMU8_SCRATCH_ENTRY_UCODE_ID_CP_MEC_JT1, false);
454 else
455 smu8_smu_populate_single_ucode_load_task(hwmgr,
456 SMU8_SCRATCH_ENTRY_UCODE_ID_CP_MEC_JT2, false);
457
458 smu8_smu_populate_single_ucode_load_task(hwmgr,
459 SMU8_SCRATCH_ENTRY_UCODE_ID_RLC_G, false);
460
461 /* populate scratch */
462 smu8_smu_populate_single_scratch_task(hwmgr,
463 SMU8_SCRATCH_ENTRY_UCODE_ID_RLC_SCRATCH,
464 TASK_TYPE_UCODE_LOAD, false);
465
466 smu8_smu_populate_single_scratch_task(hwmgr,
467 SMU8_SCRATCH_ENTRY_UCODE_ID_RLC_SRM_ARAM,
468 TASK_TYPE_UCODE_LOAD, false);
469
470 smu8_smu_populate_single_scratch_task(hwmgr,
471 SMU8_SCRATCH_ENTRY_UCODE_ID_RLC_SRM_DRAM,
472 TASK_TYPE_UCODE_LOAD, true);
473
474 return 0;
475 }
476
smu8_smu_construct_toc_for_power_profiling(struct pp_hwmgr * hwmgr)477 static int smu8_smu_construct_toc_for_power_profiling(struct pp_hwmgr *hwmgr)
478 {
479 struct smu8_smumgr *smu8_smu = hwmgr->smu_backend;
480
481 smu8_smu->toc_entry_power_profiling_index = smu8_smu->toc_entry_used_count;
482
483 smu8_smu_populate_single_scratch_task(hwmgr,
484 SMU8_SCRATCH_ENTRY_UCODE_ID_POWER_PROFILING,
485 TASK_TYPE_INITIALIZE, true);
486 return 0;
487 }
488
smu8_smu_construct_toc_for_bootup(struct pp_hwmgr * hwmgr)489 static int smu8_smu_construct_toc_for_bootup(struct pp_hwmgr *hwmgr)
490 {
491 struct smu8_smumgr *smu8_smu = hwmgr->smu_backend;
492
493 smu8_smu->toc_entry_initialize_index = smu8_smu->toc_entry_used_count;
494
495 smu8_smu_populate_single_ucode_load_task(hwmgr,
496 SMU8_SCRATCH_ENTRY_UCODE_ID_SDMA0, false);
497 if (hwmgr->chip_id != CHIP_STONEY)
498 smu8_smu_populate_single_ucode_load_task(hwmgr,
499 SMU8_SCRATCH_ENTRY_UCODE_ID_SDMA1, false);
500 smu8_smu_populate_single_ucode_load_task(hwmgr,
501 SMU8_SCRATCH_ENTRY_UCODE_ID_CP_CE, false);
502 smu8_smu_populate_single_ucode_load_task(hwmgr,
503 SMU8_SCRATCH_ENTRY_UCODE_ID_CP_PFP, false);
504 smu8_smu_populate_single_ucode_load_task(hwmgr,
505 SMU8_SCRATCH_ENTRY_UCODE_ID_CP_ME, false);
506 smu8_smu_populate_single_ucode_load_task(hwmgr,
507 SMU8_SCRATCH_ENTRY_UCODE_ID_CP_MEC_JT1, false);
508 if (hwmgr->chip_id != CHIP_STONEY)
509 smu8_smu_populate_single_ucode_load_task(hwmgr,
510 SMU8_SCRATCH_ENTRY_UCODE_ID_CP_MEC_JT2, false);
511 smu8_smu_populate_single_ucode_load_task(hwmgr,
512 SMU8_SCRATCH_ENTRY_UCODE_ID_RLC_G, true);
513
514 return 0;
515 }
516
smu8_smu_construct_toc_for_clock_table(struct pp_hwmgr * hwmgr)517 static int smu8_smu_construct_toc_for_clock_table(struct pp_hwmgr *hwmgr)
518 {
519 struct smu8_smumgr *smu8_smu = hwmgr->smu_backend;
520
521 smu8_smu->toc_entry_clock_table = smu8_smu->toc_entry_used_count;
522
523 smu8_smu_populate_single_scratch_task(hwmgr,
524 SMU8_SCRATCH_ENTRY_SMU8_FUSION_CLKTABLE,
525 TASK_TYPE_INITIALIZE, true);
526
527 return 0;
528 }
529
smu8_smu_construct_toc(struct pp_hwmgr * hwmgr)530 static int smu8_smu_construct_toc(struct pp_hwmgr *hwmgr)
531 {
532 struct smu8_smumgr *smu8_smu = hwmgr->smu_backend;
533
534 smu8_smu->toc_entry_used_count = 0;
535 smu8_smu_initialize_toc_empty_job_list(hwmgr);
536 smu8_smu_construct_toc_for_rlc_aram_save(hwmgr);
537 smu8_smu_construct_toc_for_vddgfx_enter(hwmgr);
538 smu8_smu_construct_toc_for_vddgfx_exit(hwmgr);
539 smu8_smu_construct_toc_for_power_profiling(hwmgr);
540 smu8_smu_construct_toc_for_bootup(hwmgr);
541 smu8_smu_construct_toc_for_clock_table(hwmgr);
542
543 return 0;
544 }
545
smu8_smu_populate_firmware_entries(struct pp_hwmgr * hwmgr)546 static int smu8_smu_populate_firmware_entries(struct pp_hwmgr *hwmgr)
547 {
548 struct smu8_smumgr *smu8_smu = hwmgr->smu_backend;
549 uint32_t firmware_type;
550 uint32_t i;
551 int ret;
552 enum cgs_ucode_id ucode_id;
553 struct cgs_firmware_info info = {0};
554
555 smu8_smu->driver_buffer_length = 0;
556
557 for (i = 0; i < ARRAY_SIZE(firmware_list); i++) {
558
559 firmware_type = smu8_translate_firmware_enum_to_arg(hwmgr,
560 firmware_list[i]);
561
562 ucode_id = smu8_convert_fw_type_to_cgs(firmware_type);
563
564 ret = cgs_get_firmware_info(hwmgr->device,
565 ucode_id, &info);
566
567 if (ret == 0) {
568 smu8_smu->driver_buffer[i].mc_addr = info.mc_addr;
569
570 smu8_smu->driver_buffer[i].data_size = info.image_size;
571
572 smu8_smu->driver_buffer[i].firmware_ID = firmware_list[i];
573 smu8_smu->driver_buffer_length++;
574 }
575 }
576
577 return 0;
578 }
579
smu8_smu_populate_single_scratch_entry(struct pp_hwmgr * hwmgr,enum smu8_scratch_entry scratch_type,uint32_t ulsize_byte,struct smu8_buffer_entry * entry)580 static int smu8_smu_populate_single_scratch_entry(
581 struct pp_hwmgr *hwmgr,
582 enum smu8_scratch_entry scratch_type,
583 uint32_t ulsize_byte,
584 struct smu8_buffer_entry *entry)
585 {
586 struct smu8_smumgr *smu8_smu = hwmgr->smu_backend;
587 uint32_t ulsize_aligned = SIZE_ALIGN_32(ulsize_byte);
588
589 entry->data_size = ulsize_byte;
590 entry->kaddr = (char *) smu8_smu->smu_buffer.kaddr +
591 smu8_smu->smu_buffer_used_bytes;
592 entry->mc_addr = smu8_smu->smu_buffer.mc_addr + smu8_smu->smu_buffer_used_bytes;
593 entry->firmware_ID = scratch_type;
594
595 smu8_smu->smu_buffer_used_bytes += ulsize_aligned;
596
597 return 0;
598 }
599
smu8_download_pptable_settings(struct pp_hwmgr * hwmgr,void ** table)600 static int smu8_download_pptable_settings(struct pp_hwmgr *hwmgr, void **table)
601 {
602 struct smu8_smumgr *smu8_smu = hwmgr->smu_backend;
603 unsigned long i;
604
605 for (i = 0; i < smu8_smu->scratch_buffer_length; i++) {
606 if (smu8_smu->scratch_buffer[i].firmware_ID
607 == SMU8_SCRATCH_ENTRY_SMU8_FUSION_CLKTABLE)
608 break;
609 }
610
611 *table = (struct SMU8_Fusion_ClkTable *)smu8_smu->scratch_buffer[i].kaddr;
612
613 smum_send_msg_to_smc_with_parameter(hwmgr,
614 PPSMC_MSG_SetClkTableAddrHi,
615 upper_32_bits(smu8_smu->scratch_buffer[i].mc_addr),
616 NULL);
617
618 smum_send_msg_to_smc_with_parameter(hwmgr,
619 PPSMC_MSG_SetClkTableAddrLo,
620 lower_32_bits(smu8_smu->scratch_buffer[i].mc_addr),
621 NULL);
622
623 smum_send_msg_to_smc_with_parameter(hwmgr, PPSMC_MSG_ExecuteJob,
624 smu8_smu->toc_entry_clock_table,
625 NULL);
626
627 smum_send_msg_to_smc(hwmgr, PPSMC_MSG_ClkTableXferToDram, NULL);
628
629 return 0;
630 }
631
smu8_upload_pptable_settings(struct pp_hwmgr * hwmgr)632 static int smu8_upload_pptable_settings(struct pp_hwmgr *hwmgr)
633 {
634 struct smu8_smumgr *smu8_smu = hwmgr->smu_backend;
635 unsigned long i;
636
637 for (i = 0; i < smu8_smu->scratch_buffer_length; i++) {
638 if (smu8_smu->scratch_buffer[i].firmware_ID
639 == SMU8_SCRATCH_ENTRY_SMU8_FUSION_CLKTABLE)
640 break;
641 }
642
643 smum_send_msg_to_smc_with_parameter(hwmgr,
644 PPSMC_MSG_SetClkTableAddrHi,
645 upper_32_bits(smu8_smu->scratch_buffer[i].mc_addr),
646 NULL);
647
648 smum_send_msg_to_smc_with_parameter(hwmgr,
649 PPSMC_MSG_SetClkTableAddrLo,
650 lower_32_bits(smu8_smu->scratch_buffer[i].mc_addr),
651 NULL);
652
653 smum_send_msg_to_smc_with_parameter(hwmgr, PPSMC_MSG_ExecuteJob,
654 smu8_smu->toc_entry_clock_table,
655 NULL);
656
657 smum_send_msg_to_smc(hwmgr, PPSMC_MSG_ClkTableXferToSmu, NULL);
658
659 return 0;
660 }
661
smu8_request_smu_load_fw(struct pp_hwmgr * hwmgr)662 static int smu8_request_smu_load_fw(struct pp_hwmgr *hwmgr)
663 {
664 struct smu8_smumgr *smu8_smu = hwmgr->smu_backend;
665 uint32_t smc_address;
666 uint32_t fw_to_check = 0;
667 int ret;
668
669 amdgpu_ucode_init_bo(hwmgr->adev);
670
671 smu8_smu_populate_firmware_entries(hwmgr);
672
673 smu8_smu_construct_toc(hwmgr);
674
675 smc_address = SMU8_FIRMWARE_HEADER_LOCATION +
676 offsetof(struct SMU8_Firmware_Header, UcodeLoadStatus);
677
678 smu8_write_smc_sram_dword(hwmgr, smc_address, 0, smc_address+4);
679
680 smum_send_msg_to_smc_with_parameter(hwmgr,
681 PPSMC_MSG_DriverDramAddrHi,
682 upper_32_bits(smu8_smu->toc_buffer.mc_addr),
683 NULL);
684
685 smum_send_msg_to_smc_with_parameter(hwmgr,
686 PPSMC_MSG_DriverDramAddrLo,
687 lower_32_bits(smu8_smu->toc_buffer.mc_addr),
688 NULL);
689
690 smum_send_msg_to_smc(hwmgr, PPSMC_MSG_InitJobs, NULL);
691
692 smum_send_msg_to_smc_with_parameter(hwmgr,
693 PPSMC_MSG_ExecuteJob,
694 smu8_smu->toc_entry_aram,
695 NULL);
696 smum_send_msg_to_smc_with_parameter(hwmgr, PPSMC_MSG_ExecuteJob,
697 smu8_smu->toc_entry_power_profiling_index,
698 NULL);
699
700 smum_send_msg_to_smc_with_parameter(hwmgr,
701 PPSMC_MSG_ExecuteJob,
702 smu8_smu->toc_entry_initialize_index,
703 NULL);
704
705 fw_to_check = UCODE_ID_RLC_G_MASK |
706 UCODE_ID_SDMA0_MASK |
707 UCODE_ID_SDMA1_MASK |
708 UCODE_ID_CP_CE_MASK |
709 UCODE_ID_CP_ME_MASK |
710 UCODE_ID_CP_PFP_MASK |
711 UCODE_ID_CP_MEC_JT1_MASK |
712 UCODE_ID_CP_MEC_JT2_MASK;
713
714 if (hwmgr->chip_id == CHIP_STONEY)
715 fw_to_check &= ~(UCODE_ID_SDMA1_MASK | UCODE_ID_CP_MEC_JT2_MASK);
716
717 ret = smu8_check_fw_load_finish(hwmgr, fw_to_check);
718 if (ret) {
719 pr_err("SMU firmware load failed\n");
720 return ret;
721 }
722
723 ret = smu8_load_mec_firmware(hwmgr);
724 if (ret) {
725 pr_err("Mec Firmware load failed\n");
726 return ret;
727 }
728
729 return 0;
730 }
731
smu8_start_smu(struct pp_hwmgr * hwmgr)732 static int smu8_start_smu(struct pp_hwmgr *hwmgr)
733 {
734 struct amdgpu_device *adev;
735
736 uint32_t index = SMN_MP1_SRAM_START_ADDR +
737 SMU8_FIRMWARE_HEADER_LOCATION +
738 offsetof(struct SMU8_Firmware_Header, Version);
739
740 if (hwmgr == NULL || hwmgr->device == NULL)
741 return -EINVAL;
742
743 adev = hwmgr->adev;
744
745 cgs_write_register(hwmgr->device, mmMP0PUB_IND_INDEX, index);
746 hwmgr->smu_version = cgs_read_register(hwmgr->device, mmMP0PUB_IND_DATA);
747 pr_info("smu version %02d.%02d.%02d\n",
748 ((hwmgr->smu_version >> 16) & 0xFF),
749 ((hwmgr->smu_version >> 8) & 0xFF),
750 (hwmgr->smu_version & 0xFF));
751 adev->pm.fw_version = hwmgr->smu_version >> 8;
752
753 return smu8_request_smu_load_fw(hwmgr);
754 }
755
smu8_smu_init(struct pp_hwmgr * hwmgr)756 static int smu8_smu_init(struct pp_hwmgr *hwmgr)
757 {
758 int ret = 0;
759 struct smu8_smumgr *smu8_smu;
760
761 smu8_smu = kzalloc(sizeof(struct smu8_smumgr), GFP_KERNEL);
762 if (smu8_smu == NULL)
763 return -ENOMEM;
764
765 hwmgr->smu_backend = smu8_smu;
766
767 smu8_smu->toc_buffer.data_size = 4096;
768 smu8_smu->smu_buffer.data_size =
769 ALIGN(UCODE_ID_RLC_SCRATCH_SIZE_BYTE, 32) +
770 ALIGN(UCODE_ID_RLC_SRM_ARAM_SIZE_BYTE, 32) +
771 ALIGN(UCODE_ID_RLC_SRM_DRAM_SIZE_BYTE, 32) +
772 ALIGN(sizeof(struct SMU8_MultimediaPowerLogData), 32) +
773 ALIGN(sizeof(struct SMU8_Fusion_ClkTable), 32);
774
775 ret = amdgpu_bo_create_kernel((struct amdgpu_device *)hwmgr->adev,
776 smu8_smu->toc_buffer.data_size,
777 PAGE_SIZE,
778 AMDGPU_GEM_DOMAIN_VRAM,
779 &smu8_smu->toc_buffer.handle,
780 &smu8_smu->toc_buffer.mc_addr,
781 &smu8_smu->toc_buffer.kaddr);
782 if (ret)
783 goto err2;
784
785 ret = amdgpu_bo_create_kernel((struct amdgpu_device *)hwmgr->adev,
786 smu8_smu->smu_buffer.data_size,
787 PAGE_SIZE,
788 AMDGPU_GEM_DOMAIN_VRAM,
789 &smu8_smu->smu_buffer.handle,
790 &smu8_smu->smu_buffer.mc_addr,
791 &smu8_smu->smu_buffer.kaddr);
792 if (ret)
793 goto err1;
794
795 if (0 != smu8_smu_populate_single_scratch_entry(hwmgr,
796 SMU8_SCRATCH_ENTRY_UCODE_ID_RLC_SCRATCH,
797 UCODE_ID_RLC_SCRATCH_SIZE_BYTE,
798 &smu8_smu->scratch_buffer[smu8_smu->scratch_buffer_length++])) {
799 pr_err("Error when Populate Firmware Entry.\n");
800 goto err0;
801 }
802
803 if (0 != smu8_smu_populate_single_scratch_entry(hwmgr,
804 SMU8_SCRATCH_ENTRY_UCODE_ID_RLC_SRM_ARAM,
805 UCODE_ID_RLC_SRM_ARAM_SIZE_BYTE,
806 &smu8_smu->scratch_buffer[smu8_smu->scratch_buffer_length++])) {
807 pr_err("Error when Populate Firmware Entry.\n");
808 goto err0;
809 }
810 if (0 != smu8_smu_populate_single_scratch_entry(hwmgr,
811 SMU8_SCRATCH_ENTRY_UCODE_ID_RLC_SRM_DRAM,
812 UCODE_ID_RLC_SRM_DRAM_SIZE_BYTE,
813 &smu8_smu->scratch_buffer[smu8_smu->scratch_buffer_length++])) {
814 pr_err("Error when Populate Firmware Entry.\n");
815 goto err0;
816 }
817
818 if (0 != smu8_smu_populate_single_scratch_entry(hwmgr,
819 SMU8_SCRATCH_ENTRY_UCODE_ID_POWER_PROFILING,
820 sizeof(struct SMU8_MultimediaPowerLogData),
821 &smu8_smu->scratch_buffer[smu8_smu->scratch_buffer_length++])) {
822 pr_err("Error when Populate Firmware Entry.\n");
823 goto err0;
824 }
825
826 if (0 != smu8_smu_populate_single_scratch_entry(hwmgr,
827 SMU8_SCRATCH_ENTRY_SMU8_FUSION_CLKTABLE,
828 sizeof(struct SMU8_Fusion_ClkTable),
829 &smu8_smu->scratch_buffer[smu8_smu->scratch_buffer_length++])) {
830 pr_err("Error when Populate Firmware Entry.\n");
831 goto err0;
832 }
833
834 return 0;
835
836 err0:
837 amdgpu_bo_free_kernel(&smu8_smu->smu_buffer.handle,
838 &smu8_smu->smu_buffer.mc_addr,
839 &smu8_smu->smu_buffer.kaddr);
840 err1:
841 amdgpu_bo_free_kernel(&smu8_smu->toc_buffer.handle,
842 &smu8_smu->toc_buffer.mc_addr,
843 &smu8_smu->toc_buffer.kaddr);
844 err2:
845 kfree(smu8_smu);
846 return -EINVAL;
847 }
848
smu8_smu_fini(struct pp_hwmgr * hwmgr)849 static int smu8_smu_fini(struct pp_hwmgr *hwmgr)
850 {
851 struct smu8_smumgr *smu8_smu;
852
853 if (hwmgr == NULL || hwmgr->device == NULL)
854 return -EINVAL;
855
856 smu8_smu = hwmgr->smu_backend;
857 if (smu8_smu) {
858 amdgpu_bo_free_kernel(&smu8_smu->toc_buffer.handle,
859 &smu8_smu->toc_buffer.mc_addr,
860 &smu8_smu->toc_buffer.kaddr);
861 amdgpu_bo_free_kernel(&smu8_smu->smu_buffer.handle,
862 &smu8_smu->smu_buffer.mc_addr,
863 &smu8_smu->smu_buffer.kaddr);
864 kfree(smu8_smu);
865 }
866
867 return 0;
868 }
869
smu8_dpm_check_smu_features(struct pp_hwmgr * hwmgr,unsigned long check_feature)870 static bool smu8_dpm_check_smu_features(struct pp_hwmgr *hwmgr,
871 unsigned long check_feature)
872 {
873 int result;
874 uint32_t features;
875
876 result = smum_send_msg_to_smc_with_parameter(hwmgr,
877 PPSMC_MSG_GetFeatureStatus,
878 0,
879 &features);
880 if (result == 0) {
881 if (features & check_feature)
882 return true;
883 }
884
885 return false;
886 }
887
smu8_is_dpm_running(struct pp_hwmgr * hwmgr)888 static bool smu8_is_dpm_running(struct pp_hwmgr *hwmgr)
889 {
890 if (smu8_dpm_check_smu_features(hwmgr, SMU_EnabledFeatureScoreboard_SclkDpmOn))
891 return true;
892 return false;
893 }
894
895 const struct pp_smumgr_func smu8_smu_funcs = {
896 .name = "smu8_smu",
897 .smu_init = smu8_smu_init,
898 .smu_fini = smu8_smu_fini,
899 .start_smu = smu8_start_smu,
900 .check_fw_load_finish = smu8_check_fw_load_finish,
901 .request_smu_load_fw = NULL,
902 .request_smu_load_specific_fw = NULL,
903 .get_argument = smu8_get_argument,
904 .send_msg_to_smc = smu8_send_msg_to_smc,
905 .send_msg_to_smc_with_parameter = smu8_send_msg_to_smc_with_parameter,
906 .download_pptable_settings = smu8_download_pptable_settings,
907 .upload_pptable_settings = smu8_upload_pptable_settings,
908 .is_dpm_running = smu8_is_dpm_running,
909 };
910
911