1 // SPDX-License-Identifier: GPL-2.0
2 /*
3 * Copyright (C) STMicroelectronics SA 2014
4 * Authors: Fabien Dessenne <fabien.dessenne@st.com> for STMicroelectronics.
5 */
6
7 #include <linux/component.h>
8 #include <linux/delay.h>
9 #include <linux/dma-mapping.h>
10 #include <linux/firmware.h>
11 #include <linux/io.h>
12 #include <linux/module.h>
13 #include <linux/reset.h>
14 #include <linux/seq_file.h>
15
16 #include <drm/drm_atomic.h>
17 #include <drm/drm_device.h>
18 #include <drm/drm_fb_cma_helper.h>
19 #include <drm/drm_fourcc.h>
20 #include <drm/drm_gem_cma_helper.h>
21
22 #include "sti_compositor.h"
23 #include "sti_drv.h"
24 #include "sti_hqvdp_lut.h"
25 #include "sti_plane.h"
26 #include "sti_vtg.h"
27
28 /* Firmware name */
29 #define HQVDP_FMW_NAME "hqvdp-stih407.bin"
30
31 /* Regs address */
32 #define HQVDP_DMEM 0x00000000 /* 0x00000000 */
33 #define HQVDP_PMEM 0x00040000 /* 0x00040000 */
34 #define HQVDP_RD_PLUG 0x000E0000 /* 0x000E0000 */
35 #define HQVDP_RD_PLUG_CONTROL (HQVDP_RD_PLUG + 0x1000) /* 0x000E1000 */
36 #define HQVDP_RD_PLUG_PAGE_SIZE (HQVDP_RD_PLUG + 0x1004) /* 0x000E1004 */
37 #define HQVDP_RD_PLUG_MIN_OPC (HQVDP_RD_PLUG + 0x1008) /* 0x000E1008 */
38 #define HQVDP_RD_PLUG_MAX_OPC (HQVDP_RD_PLUG + 0x100C) /* 0x000E100C */
39 #define HQVDP_RD_PLUG_MAX_CHK (HQVDP_RD_PLUG + 0x1010) /* 0x000E1010 */
40 #define HQVDP_RD_PLUG_MAX_MSG (HQVDP_RD_PLUG + 0x1014) /* 0x000E1014 */
41 #define HQVDP_RD_PLUG_MIN_SPACE (HQVDP_RD_PLUG + 0x1018) /* 0x000E1018 */
42 #define HQVDP_WR_PLUG 0x000E2000 /* 0x000E2000 */
43 #define HQVDP_WR_PLUG_CONTROL (HQVDP_WR_PLUG + 0x1000) /* 0x000E3000 */
44 #define HQVDP_WR_PLUG_PAGE_SIZE (HQVDP_WR_PLUG + 0x1004) /* 0x000E3004 */
45 #define HQVDP_WR_PLUG_MIN_OPC (HQVDP_WR_PLUG + 0x1008) /* 0x000E3008 */
46 #define HQVDP_WR_PLUG_MAX_OPC (HQVDP_WR_PLUG + 0x100C) /* 0x000E300C */
47 #define HQVDP_WR_PLUG_MAX_CHK (HQVDP_WR_PLUG + 0x1010) /* 0x000E3010 */
48 #define HQVDP_WR_PLUG_MAX_MSG (HQVDP_WR_PLUG + 0x1014) /* 0x000E3014 */
49 #define HQVDP_WR_PLUG_MIN_SPACE (HQVDP_WR_PLUG + 0x1018) /* 0x000E3018 */
50 #define HQVDP_MBX 0x000E4000 /* 0x000E4000 */
51 #define HQVDP_MBX_IRQ_TO_XP70 (HQVDP_MBX + 0x0000) /* 0x000E4000 */
52 #define HQVDP_MBX_INFO_HOST (HQVDP_MBX + 0x0004) /* 0x000E4004 */
53 #define HQVDP_MBX_IRQ_TO_HOST (HQVDP_MBX + 0x0008) /* 0x000E4008 */
54 #define HQVDP_MBX_INFO_XP70 (HQVDP_MBX + 0x000C) /* 0x000E400C */
55 #define HQVDP_MBX_SW_RESET_CTRL (HQVDP_MBX + 0x0010) /* 0x000E4010 */
56 #define HQVDP_MBX_STARTUP_CTRL1 (HQVDP_MBX + 0x0014) /* 0x000E4014 */
57 #define HQVDP_MBX_STARTUP_CTRL2 (HQVDP_MBX + 0x0018) /* 0x000E4018 */
58 #define HQVDP_MBX_GP_STATUS (HQVDP_MBX + 0x001C) /* 0x000E401C */
59 #define HQVDP_MBX_NEXT_CMD (HQVDP_MBX + 0x0020) /* 0x000E4020 */
60 #define HQVDP_MBX_CURRENT_CMD (HQVDP_MBX + 0x0024) /* 0x000E4024 */
61 #define HQVDP_MBX_SOFT_VSYNC (HQVDP_MBX + 0x0028) /* 0x000E4028 */
62
63 /* Plugs config */
64 #define PLUG_CONTROL_ENABLE 0x00000001
65 #define PLUG_PAGE_SIZE_256 0x00000002
66 #define PLUG_MIN_OPC_8 0x00000003
67 #define PLUG_MAX_OPC_64 0x00000006
68 #define PLUG_MAX_CHK_2X 0x00000001
69 #define PLUG_MAX_MSG_1X 0x00000000
70 #define PLUG_MIN_SPACE_1 0x00000000
71
72 /* SW reset CTRL */
73 #define SW_RESET_CTRL_FULL BIT(0)
74 #define SW_RESET_CTRL_CORE BIT(1)
75
76 /* Startup ctrl 1 */
77 #define STARTUP_CTRL1_RST_DONE BIT(0)
78 #define STARTUP_CTRL1_AUTH_IDLE BIT(2)
79
80 /* Startup ctrl 2 */
81 #define STARTUP_CTRL2_FETCH_EN BIT(1)
82
83 /* Info xP70 */
84 #define INFO_XP70_FW_READY BIT(15)
85 #define INFO_XP70_FW_PROCESSING BIT(14)
86 #define INFO_XP70_FW_INITQUEUES BIT(13)
87
88 /* SOFT_VSYNC */
89 #define SOFT_VSYNC_HW 0x00000000
90 #define SOFT_VSYNC_SW_CMD 0x00000001
91 #define SOFT_VSYNC_SW_CTRL_IRQ 0x00000003
92
93 /* Reset & boot poll config */
94 #define POLL_MAX_ATTEMPT 50
95 #define POLL_DELAY_MS 20
96
97 #define SCALE_FACTOR 8192
98 #define SCALE_MAX_FOR_LEG_LUT_F 4096
99 #define SCALE_MAX_FOR_LEG_LUT_E 4915
100 #define SCALE_MAX_FOR_LEG_LUT_D 6654
101 #define SCALE_MAX_FOR_LEG_LUT_C 8192
102
103 enum sti_hvsrc_orient {
104 HVSRC_HORI,
105 HVSRC_VERT
106 };
107
108 /* Command structures */
109 struct sti_hqvdp_top {
110 u32 config;
111 u32 mem_format;
112 u32 current_luma;
113 u32 current_enh_luma;
114 u32 current_right_luma;
115 u32 current_enh_right_luma;
116 u32 current_chroma;
117 u32 current_enh_chroma;
118 u32 current_right_chroma;
119 u32 current_enh_right_chroma;
120 u32 output_luma;
121 u32 output_chroma;
122 u32 luma_src_pitch;
123 u32 luma_enh_src_pitch;
124 u32 luma_right_src_pitch;
125 u32 luma_enh_right_src_pitch;
126 u32 chroma_src_pitch;
127 u32 chroma_enh_src_pitch;
128 u32 chroma_right_src_pitch;
129 u32 chroma_enh_right_src_pitch;
130 u32 luma_processed_pitch;
131 u32 chroma_processed_pitch;
132 u32 input_frame_size;
133 u32 input_viewport_ori;
134 u32 input_viewport_ori_right;
135 u32 input_viewport_size;
136 u32 left_view_border_width;
137 u32 right_view_border_width;
138 u32 left_view_3d_offset_width;
139 u32 right_view_3d_offset_width;
140 u32 side_stripe_color;
141 u32 crc_reset_ctrl;
142 };
143
144 /* Configs for interlaced : no IT, no pass thru, 3 fields */
145 #define TOP_CONFIG_INTER_BTM 0x00000000
146 #define TOP_CONFIG_INTER_TOP 0x00000002
147
148 /* Config for progressive : no IT, no pass thru, 3 fields */
149 #define TOP_CONFIG_PROGRESSIVE 0x00000001
150
151 /* Default MemFormat: in=420_raster_dual out=444_raster;opaque Mem2Tv mode */
152 #define TOP_MEM_FORMAT_DFLT 0x00018060
153
154 /* Min/Max size */
155 #define MAX_WIDTH 0x1FFF
156 #define MAX_HEIGHT 0x0FFF
157 #define MIN_WIDTH 0x0030
158 #define MIN_HEIGHT 0x0010
159
160 struct sti_hqvdp_vc1re {
161 u32 ctrl_prv_csdi;
162 u32 ctrl_cur_csdi;
163 u32 ctrl_nxt_csdi;
164 u32 ctrl_cur_fmd;
165 u32 ctrl_nxt_fmd;
166 };
167
168 struct sti_hqvdp_fmd {
169 u32 config;
170 u32 viewport_ori;
171 u32 viewport_size;
172 u32 next_next_luma;
173 u32 next_next_right_luma;
174 u32 next_next_next_luma;
175 u32 next_next_next_right_luma;
176 u32 threshold_scd;
177 u32 threshold_rfd;
178 u32 threshold_move;
179 u32 threshold_cfd;
180 };
181
182 struct sti_hqvdp_csdi {
183 u32 config;
184 u32 config2;
185 u32 dcdi_config;
186 u32 prev_luma;
187 u32 prev_enh_luma;
188 u32 prev_right_luma;
189 u32 prev_enh_right_luma;
190 u32 next_luma;
191 u32 next_enh_luma;
192 u32 next_right_luma;
193 u32 next_enh_right_luma;
194 u32 prev_chroma;
195 u32 prev_enh_chroma;
196 u32 prev_right_chroma;
197 u32 prev_enh_right_chroma;
198 u32 next_chroma;
199 u32 next_enh_chroma;
200 u32 next_right_chroma;
201 u32 next_enh_right_chroma;
202 u32 prev_motion;
203 u32 prev_right_motion;
204 u32 cur_motion;
205 u32 cur_right_motion;
206 u32 next_motion;
207 u32 next_right_motion;
208 };
209
210 /* Config for progressive: by pass */
211 #define CSDI_CONFIG_PROG 0x00000000
212 /* Config for directional deinterlacing without motion */
213 #define CSDI_CONFIG_INTER_DIR 0x00000016
214 /* Additional configs for fader, blender, motion,... deinterlace algorithms */
215 #define CSDI_CONFIG2_DFLT 0x000001B3
216 #define CSDI_DCDI_CONFIG_DFLT 0x00203803
217
218 struct sti_hqvdp_hvsrc {
219 u32 hor_panoramic_ctrl;
220 u32 output_picture_size;
221 u32 init_horizontal;
222 u32 init_vertical;
223 u32 param_ctrl;
224 u32 yh_coef[NB_COEF];
225 u32 ch_coef[NB_COEF];
226 u32 yv_coef[NB_COEF];
227 u32 cv_coef[NB_COEF];
228 u32 hori_shift;
229 u32 vert_shift;
230 };
231
232 /* Default ParamCtrl: all controls enabled */
233 #define HVSRC_PARAM_CTRL_DFLT 0xFFFFFFFF
234
235 struct sti_hqvdp_iqi {
236 u32 config;
237 u32 demo_wind_size;
238 u32 pk_config;
239 u32 coeff0_coeff1;
240 u32 coeff2_coeff3;
241 u32 coeff4;
242 u32 pk_lut;
243 u32 pk_gain;
244 u32 pk_coring_level;
245 u32 cti_config;
246 u32 le_config;
247 u32 le_lut[64];
248 u32 con_bri;
249 u32 sat_gain;
250 u32 pxf_conf;
251 u32 default_color;
252 };
253
254 /* Default Config : IQI bypassed */
255 #define IQI_CONFIG_DFLT 0x00000001
256 /* Default Contrast & Brightness gain = 256 */
257 #define IQI_CON_BRI_DFLT 0x00000100
258 /* Default Saturation gain = 256 */
259 #define IQI_SAT_GAIN_DFLT 0x00000100
260 /* Default PxfConf : P2I bypassed */
261 #define IQI_PXF_CONF_DFLT 0x00000001
262
263 struct sti_hqvdp_top_status {
264 u32 processing_time;
265 u32 input_y_crc;
266 u32 input_uv_crc;
267 };
268
269 struct sti_hqvdp_fmd_status {
270 u32 fmd_repeat_move_status;
271 u32 fmd_scene_count_status;
272 u32 cfd_sum;
273 u32 field_sum;
274 u32 next_y_fmd_crc;
275 u32 next_next_y_fmd_crc;
276 u32 next_next_next_y_fmd_crc;
277 };
278
279 struct sti_hqvdp_csdi_status {
280 u32 prev_y_csdi_crc;
281 u32 cur_y_csdi_crc;
282 u32 next_y_csdi_crc;
283 u32 prev_uv_csdi_crc;
284 u32 cur_uv_csdi_crc;
285 u32 next_uv_csdi_crc;
286 u32 y_csdi_crc;
287 u32 uv_csdi_crc;
288 u32 uv_cup_crc;
289 u32 mot_csdi_crc;
290 u32 mot_cur_csdi_crc;
291 u32 mot_prev_csdi_crc;
292 };
293
294 struct sti_hqvdp_hvsrc_status {
295 u32 y_hvsrc_crc;
296 u32 u_hvsrc_crc;
297 u32 v_hvsrc_crc;
298 };
299
300 struct sti_hqvdp_iqi_status {
301 u32 pxf_it_status;
302 u32 y_iqi_crc;
303 u32 u_iqi_crc;
304 u32 v_iqi_crc;
305 };
306
307 /* Main commands. We use 2 commands one being processed by the firmware, one
308 * ready to be fetched upon next Vsync*/
309 #define NB_VDP_CMD 2
310
311 struct sti_hqvdp_cmd {
312 struct sti_hqvdp_top top;
313 struct sti_hqvdp_vc1re vc1re;
314 struct sti_hqvdp_fmd fmd;
315 struct sti_hqvdp_csdi csdi;
316 struct sti_hqvdp_hvsrc hvsrc;
317 struct sti_hqvdp_iqi iqi;
318 struct sti_hqvdp_top_status top_status;
319 struct sti_hqvdp_fmd_status fmd_status;
320 struct sti_hqvdp_csdi_status csdi_status;
321 struct sti_hqvdp_hvsrc_status hvsrc_status;
322 struct sti_hqvdp_iqi_status iqi_status;
323 };
324
325 /*
326 * STI HQVDP structure
327 *
328 * @dev: driver device
329 * @drm_dev: the drm device
330 * @regs: registers
331 * @plane: plane structure for hqvdp it self
332 * @clk: IP clock
333 * @clk_pix_main: pix main clock
334 * @reset: reset control
335 * @vtg_nb: notifier to handle VTG Vsync
336 * @btm_field_pending: is there any bottom field (interlaced frame) to display
337 * @hqvdp_cmd: buffer of commands
338 * @hqvdp_cmd_paddr: physical address of hqvdp_cmd
339 * @vtg: vtg for main data path
340 * @xp70_initialized: true if xp70 is already initialized
341 * @vtg_registered: true if registered to VTG
342 */
343 struct sti_hqvdp {
344 struct device *dev;
345 struct drm_device *drm_dev;
346 void __iomem *regs;
347 struct sti_plane plane;
348 struct clk *clk;
349 struct clk *clk_pix_main;
350 struct reset_control *reset;
351 struct notifier_block vtg_nb;
352 bool btm_field_pending;
353 void *hqvdp_cmd;
354 u32 hqvdp_cmd_paddr;
355 struct sti_vtg *vtg;
356 bool xp70_initialized;
357 bool vtg_registered;
358 };
359
360 #define to_sti_hqvdp(x) container_of(x, struct sti_hqvdp, plane)
361
362 static const uint32_t hqvdp_supported_formats[] = {
363 DRM_FORMAT_NV12,
364 };
365
366 /**
367 * sti_hqvdp_get_free_cmd
368 * @hqvdp: hqvdp structure
369 *
370 * Look for a hqvdp_cmd that is not being used (or about to be used) by the FW.
371 *
372 * RETURNS:
373 * the offset of the command to be used.
374 * -1 in error cases
375 */
sti_hqvdp_get_free_cmd(struct sti_hqvdp * hqvdp)376 static int sti_hqvdp_get_free_cmd(struct sti_hqvdp *hqvdp)
377 {
378 u32 curr_cmd, next_cmd;
379 u32 cmd = hqvdp->hqvdp_cmd_paddr;
380 int i;
381
382 curr_cmd = readl(hqvdp->regs + HQVDP_MBX_CURRENT_CMD);
383 next_cmd = readl(hqvdp->regs + HQVDP_MBX_NEXT_CMD);
384
385 for (i = 0; i < NB_VDP_CMD; i++) {
386 if ((cmd != curr_cmd) && (cmd != next_cmd))
387 return i * sizeof(struct sti_hqvdp_cmd);
388 cmd += sizeof(struct sti_hqvdp_cmd);
389 }
390
391 return -1;
392 }
393
394 /**
395 * sti_hqvdp_get_curr_cmd
396 * @hqvdp: hqvdp structure
397 *
398 * Look for the hqvdp_cmd that is being used by the FW.
399 *
400 * RETURNS:
401 * the offset of the command to be used.
402 * -1 in error cases
403 */
sti_hqvdp_get_curr_cmd(struct sti_hqvdp * hqvdp)404 static int sti_hqvdp_get_curr_cmd(struct sti_hqvdp *hqvdp)
405 {
406 u32 curr_cmd;
407 u32 cmd = hqvdp->hqvdp_cmd_paddr;
408 unsigned int i;
409
410 curr_cmd = readl(hqvdp->regs + HQVDP_MBX_CURRENT_CMD);
411
412 for (i = 0; i < NB_VDP_CMD; i++) {
413 if (cmd == curr_cmd)
414 return i * sizeof(struct sti_hqvdp_cmd);
415
416 cmd += sizeof(struct sti_hqvdp_cmd);
417 }
418
419 return -1;
420 }
421
422 /**
423 * sti_hqvdp_get_next_cmd
424 * @hqvdp: hqvdp structure
425 *
426 * Look for the next hqvdp_cmd that will be used by the FW.
427 *
428 * RETURNS:
429 * the offset of the next command that will be used.
430 * -1 in error cases
431 */
sti_hqvdp_get_next_cmd(struct sti_hqvdp * hqvdp)432 static int sti_hqvdp_get_next_cmd(struct sti_hqvdp *hqvdp)
433 {
434 int next_cmd;
435 dma_addr_t cmd = hqvdp->hqvdp_cmd_paddr;
436 unsigned int i;
437
438 next_cmd = readl(hqvdp->regs + HQVDP_MBX_NEXT_CMD);
439
440 for (i = 0; i < NB_VDP_CMD; i++) {
441 if (cmd == next_cmd)
442 return i * sizeof(struct sti_hqvdp_cmd);
443
444 cmd += sizeof(struct sti_hqvdp_cmd);
445 }
446
447 return -1;
448 }
449
450 #define DBGFS_DUMP(reg) seq_printf(s, "\n %-25s 0x%08X", #reg, \
451 readl(hqvdp->regs + reg))
452
hqvdp_dbg_get_lut(u32 * coef)453 static const char *hqvdp_dbg_get_lut(u32 *coef)
454 {
455 if (!memcmp(coef, coef_lut_a_legacy, 16))
456 return "LUT A";
457 if (!memcmp(coef, coef_lut_b, 16))
458 return "LUT B";
459 if (!memcmp(coef, coef_lut_c_y_legacy, 16))
460 return "LUT C Y";
461 if (!memcmp(coef, coef_lut_c_c_legacy, 16))
462 return "LUT C C";
463 if (!memcmp(coef, coef_lut_d_y_legacy, 16))
464 return "LUT D Y";
465 if (!memcmp(coef, coef_lut_d_c_legacy, 16))
466 return "LUT D C";
467 if (!memcmp(coef, coef_lut_e_y_legacy, 16))
468 return "LUT E Y";
469 if (!memcmp(coef, coef_lut_e_c_legacy, 16))
470 return "LUT E C";
471 if (!memcmp(coef, coef_lut_f_y_legacy, 16))
472 return "LUT F Y";
473 if (!memcmp(coef, coef_lut_f_c_legacy, 16))
474 return "LUT F C";
475 return "<UNKNOWN>";
476 }
477
hqvdp_dbg_dump_cmd(struct seq_file * s,struct sti_hqvdp_cmd * c)478 static void hqvdp_dbg_dump_cmd(struct seq_file *s, struct sti_hqvdp_cmd *c)
479 {
480 int src_w, src_h, dst_w, dst_h;
481
482 seq_puts(s, "\n\tTOP:");
483 seq_printf(s, "\n\t %-20s 0x%08X", "Config", c->top.config);
484 switch (c->top.config) {
485 case TOP_CONFIG_PROGRESSIVE:
486 seq_puts(s, "\tProgressive");
487 break;
488 case TOP_CONFIG_INTER_TOP:
489 seq_puts(s, "\tInterlaced, top field");
490 break;
491 case TOP_CONFIG_INTER_BTM:
492 seq_puts(s, "\tInterlaced, bottom field");
493 break;
494 default:
495 seq_puts(s, "\t<UNKNOWN>");
496 break;
497 }
498
499 seq_printf(s, "\n\t %-20s 0x%08X", "MemFormat", c->top.mem_format);
500 seq_printf(s, "\n\t %-20s 0x%08X", "CurrentY", c->top.current_luma);
501 seq_printf(s, "\n\t %-20s 0x%08X", "CurrentC", c->top.current_chroma);
502 seq_printf(s, "\n\t %-20s 0x%08X", "YSrcPitch", c->top.luma_src_pitch);
503 seq_printf(s, "\n\t %-20s 0x%08X", "CSrcPitch",
504 c->top.chroma_src_pitch);
505 seq_printf(s, "\n\t %-20s 0x%08X", "InputFrameSize",
506 c->top.input_frame_size);
507 seq_printf(s, "\t%dx%d",
508 c->top.input_frame_size & 0x0000FFFF,
509 c->top.input_frame_size >> 16);
510 seq_printf(s, "\n\t %-20s 0x%08X", "InputViewportSize",
511 c->top.input_viewport_size);
512 src_w = c->top.input_viewport_size & 0x0000FFFF;
513 src_h = c->top.input_viewport_size >> 16;
514 seq_printf(s, "\t%dx%d", src_w, src_h);
515
516 seq_puts(s, "\n\tHVSRC:");
517 seq_printf(s, "\n\t %-20s 0x%08X", "OutputPictureSize",
518 c->hvsrc.output_picture_size);
519 dst_w = c->hvsrc.output_picture_size & 0x0000FFFF;
520 dst_h = c->hvsrc.output_picture_size >> 16;
521 seq_printf(s, "\t%dx%d", dst_w, dst_h);
522 seq_printf(s, "\n\t %-20s 0x%08X", "ParamCtrl", c->hvsrc.param_ctrl);
523
524 seq_printf(s, "\n\t %-20s %s", "yh_coef",
525 hqvdp_dbg_get_lut(c->hvsrc.yh_coef));
526 seq_printf(s, "\n\t %-20s %s", "ch_coef",
527 hqvdp_dbg_get_lut(c->hvsrc.ch_coef));
528 seq_printf(s, "\n\t %-20s %s", "yv_coef",
529 hqvdp_dbg_get_lut(c->hvsrc.yv_coef));
530 seq_printf(s, "\n\t %-20s %s", "cv_coef",
531 hqvdp_dbg_get_lut(c->hvsrc.cv_coef));
532
533 seq_printf(s, "\n\t %-20s", "ScaleH");
534 if (dst_w > src_w)
535 seq_printf(s, " %d/1", dst_w / src_w);
536 else
537 seq_printf(s, " 1/%d", src_w / dst_w);
538
539 seq_printf(s, "\n\t %-20s", "tScaleV");
540 if (dst_h > src_h)
541 seq_printf(s, " %d/1", dst_h / src_h);
542 else
543 seq_printf(s, " 1/%d", src_h / dst_h);
544
545 seq_puts(s, "\n\tCSDI:");
546 seq_printf(s, "\n\t %-20s 0x%08X\t", "Config", c->csdi.config);
547 switch (c->csdi.config) {
548 case CSDI_CONFIG_PROG:
549 seq_puts(s, "Bypass");
550 break;
551 case CSDI_CONFIG_INTER_DIR:
552 seq_puts(s, "Deinterlace, directional");
553 break;
554 default:
555 seq_puts(s, "<UNKNOWN>");
556 break;
557 }
558
559 seq_printf(s, "\n\t %-20s 0x%08X", "Config2", c->csdi.config2);
560 seq_printf(s, "\n\t %-20s 0x%08X", "DcdiConfig", c->csdi.dcdi_config);
561 }
562
hqvdp_dbg_show(struct seq_file * s,void * data)563 static int hqvdp_dbg_show(struct seq_file *s, void *data)
564 {
565 struct drm_info_node *node = s->private;
566 struct sti_hqvdp *hqvdp = (struct sti_hqvdp *)node->info_ent->data;
567 int cmd, cmd_offset, infoxp70;
568 void *virt;
569
570 seq_printf(s, "%s: (vaddr = 0x%p)",
571 sti_plane_to_str(&hqvdp->plane), hqvdp->regs);
572
573 DBGFS_DUMP(HQVDP_MBX_IRQ_TO_XP70);
574 DBGFS_DUMP(HQVDP_MBX_INFO_HOST);
575 DBGFS_DUMP(HQVDP_MBX_IRQ_TO_HOST);
576 DBGFS_DUMP(HQVDP_MBX_INFO_XP70);
577 infoxp70 = readl(hqvdp->regs + HQVDP_MBX_INFO_XP70);
578 seq_puts(s, "\tFirmware state: ");
579 if (infoxp70 & INFO_XP70_FW_READY)
580 seq_puts(s, "idle and ready");
581 else if (infoxp70 & INFO_XP70_FW_PROCESSING)
582 seq_puts(s, "processing a picture");
583 else if (infoxp70 & INFO_XP70_FW_INITQUEUES)
584 seq_puts(s, "programming queues");
585 else
586 seq_puts(s, "NOT READY");
587
588 DBGFS_DUMP(HQVDP_MBX_SW_RESET_CTRL);
589 DBGFS_DUMP(HQVDP_MBX_STARTUP_CTRL1);
590 if (readl(hqvdp->regs + HQVDP_MBX_STARTUP_CTRL1)
591 & STARTUP_CTRL1_RST_DONE)
592 seq_puts(s, "\tReset is done");
593 else
594 seq_puts(s, "\tReset is NOT done");
595 DBGFS_DUMP(HQVDP_MBX_STARTUP_CTRL2);
596 if (readl(hqvdp->regs + HQVDP_MBX_STARTUP_CTRL2)
597 & STARTUP_CTRL2_FETCH_EN)
598 seq_puts(s, "\tFetch is enabled");
599 else
600 seq_puts(s, "\tFetch is NOT enabled");
601 DBGFS_DUMP(HQVDP_MBX_GP_STATUS);
602 DBGFS_DUMP(HQVDP_MBX_NEXT_CMD);
603 DBGFS_DUMP(HQVDP_MBX_CURRENT_CMD);
604 DBGFS_DUMP(HQVDP_MBX_SOFT_VSYNC);
605 if (!(readl(hqvdp->regs + HQVDP_MBX_SOFT_VSYNC) & 3))
606 seq_puts(s, "\tHW Vsync");
607 else
608 seq_puts(s, "\tSW Vsync ?!?!");
609
610 /* Last command */
611 cmd = readl(hqvdp->regs + HQVDP_MBX_CURRENT_CMD);
612 cmd_offset = sti_hqvdp_get_curr_cmd(hqvdp);
613 if (cmd_offset == -1) {
614 seq_puts(s, "\n\n Last command: unknown");
615 } else {
616 virt = hqvdp->hqvdp_cmd + cmd_offset;
617 seq_printf(s, "\n\n Last command: address @ 0x%x (0x%p)",
618 cmd, virt);
619 hqvdp_dbg_dump_cmd(s, (struct sti_hqvdp_cmd *)virt);
620 }
621
622 /* Next command */
623 cmd = readl(hqvdp->regs + HQVDP_MBX_NEXT_CMD);
624 cmd_offset = sti_hqvdp_get_next_cmd(hqvdp);
625 if (cmd_offset == -1) {
626 seq_puts(s, "\n\n Next command: unknown");
627 } else {
628 virt = hqvdp->hqvdp_cmd + cmd_offset;
629 seq_printf(s, "\n\n Next command address: @ 0x%x (0x%p)",
630 cmd, virt);
631 hqvdp_dbg_dump_cmd(s, (struct sti_hqvdp_cmd *)virt);
632 }
633
634 seq_putc(s, '\n');
635 return 0;
636 }
637
638 static struct drm_info_list hqvdp_debugfs_files[] = {
639 { "hqvdp", hqvdp_dbg_show, 0, NULL },
640 };
641
hqvdp_debugfs_init(struct sti_hqvdp * hqvdp,struct drm_minor * minor)642 static void hqvdp_debugfs_init(struct sti_hqvdp *hqvdp, struct drm_minor *minor)
643 {
644 unsigned int i;
645
646 for (i = 0; i < ARRAY_SIZE(hqvdp_debugfs_files); i++)
647 hqvdp_debugfs_files[i].data = hqvdp;
648
649 drm_debugfs_create_files(hqvdp_debugfs_files,
650 ARRAY_SIZE(hqvdp_debugfs_files),
651 minor->debugfs_root, minor);
652 }
653
654 /**
655 * sti_hqvdp_update_hvsrc
656 * @orient: horizontal or vertical
657 * @scale: scaling/zoom factor
658 * @hvsrc: the structure containing the LUT coef
659 *
660 * Update the Y and C Lut coef, as well as the shift param
661 *
662 * RETURNS:
663 * None.
664 */
sti_hqvdp_update_hvsrc(enum sti_hvsrc_orient orient,int scale,struct sti_hqvdp_hvsrc * hvsrc)665 static void sti_hqvdp_update_hvsrc(enum sti_hvsrc_orient orient, int scale,
666 struct sti_hqvdp_hvsrc *hvsrc)
667 {
668 const int *coef_c, *coef_y;
669 int shift_c, shift_y;
670
671 /* Get the appropriate coef tables */
672 if (scale < SCALE_MAX_FOR_LEG_LUT_F) {
673 coef_y = coef_lut_f_y_legacy;
674 coef_c = coef_lut_f_c_legacy;
675 shift_y = SHIFT_LUT_F_Y_LEGACY;
676 shift_c = SHIFT_LUT_F_C_LEGACY;
677 } else if (scale < SCALE_MAX_FOR_LEG_LUT_E) {
678 coef_y = coef_lut_e_y_legacy;
679 coef_c = coef_lut_e_c_legacy;
680 shift_y = SHIFT_LUT_E_Y_LEGACY;
681 shift_c = SHIFT_LUT_E_C_LEGACY;
682 } else if (scale < SCALE_MAX_FOR_LEG_LUT_D) {
683 coef_y = coef_lut_d_y_legacy;
684 coef_c = coef_lut_d_c_legacy;
685 shift_y = SHIFT_LUT_D_Y_LEGACY;
686 shift_c = SHIFT_LUT_D_C_LEGACY;
687 } else if (scale < SCALE_MAX_FOR_LEG_LUT_C) {
688 coef_y = coef_lut_c_y_legacy;
689 coef_c = coef_lut_c_c_legacy;
690 shift_y = SHIFT_LUT_C_Y_LEGACY;
691 shift_c = SHIFT_LUT_C_C_LEGACY;
692 } else if (scale == SCALE_MAX_FOR_LEG_LUT_C) {
693 coef_y = coef_c = coef_lut_b;
694 shift_y = shift_c = SHIFT_LUT_B;
695 } else {
696 coef_y = coef_c = coef_lut_a_legacy;
697 shift_y = shift_c = SHIFT_LUT_A_LEGACY;
698 }
699
700 if (orient == HVSRC_HORI) {
701 hvsrc->hori_shift = (shift_c << 16) | shift_y;
702 memcpy(hvsrc->yh_coef, coef_y, sizeof(hvsrc->yh_coef));
703 memcpy(hvsrc->ch_coef, coef_c, sizeof(hvsrc->ch_coef));
704 } else {
705 hvsrc->vert_shift = (shift_c << 16) | shift_y;
706 memcpy(hvsrc->yv_coef, coef_y, sizeof(hvsrc->yv_coef));
707 memcpy(hvsrc->cv_coef, coef_c, sizeof(hvsrc->cv_coef));
708 }
709 }
710
711 /**
712 * sti_hqvdp_check_hw_scaling
713 * @hqvdp: hqvdp pointer
714 * @mode: display mode with timing constraints
715 * @src_w: source width
716 * @src_h: source height
717 * @dst_w: destination width
718 * @dst_h: destination height
719 *
720 * Check if the HW is able to perform the scaling request
721 * The firmware scaling limitation is "CEIL(1/Zy) <= FLOOR(LFW)" where:
722 * Zy = OutputHeight / InputHeight
723 * LFW = (Tx * IPClock) / (MaxNbCycles * Cp)
724 * Tx : Total video mode horizontal resolution
725 * IPClock : HQVDP IP clock (Mhz)
726 * MaxNbCycles: max(InputWidth, OutputWidth)
727 * Cp: Video mode pixel clock (Mhz)
728 *
729 * RETURNS:
730 * True if the HW can scale.
731 */
sti_hqvdp_check_hw_scaling(struct sti_hqvdp * hqvdp,struct drm_display_mode * mode,int src_w,int src_h,int dst_w,int dst_h)732 static bool sti_hqvdp_check_hw_scaling(struct sti_hqvdp *hqvdp,
733 struct drm_display_mode *mode,
734 int src_w, int src_h,
735 int dst_w, int dst_h)
736 {
737 unsigned long lfw;
738 unsigned int inv_zy;
739
740 lfw = mode->htotal * (clk_get_rate(hqvdp->clk) / 1000000);
741 lfw /= max(src_w, dst_w) * mode->clock / 1000;
742
743 inv_zy = DIV_ROUND_UP(src_h, dst_h);
744
745 return (inv_zy <= lfw) ? true : false;
746 }
747
748 /**
749 * sti_hqvdp_disable
750 * @hqvdp: hqvdp pointer
751 *
752 * Disables the HQVDP plane
753 */
sti_hqvdp_disable(struct sti_hqvdp * hqvdp)754 static void sti_hqvdp_disable(struct sti_hqvdp *hqvdp)
755 {
756 int i;
757
758 DRM_DEBUG_DRIVER("%s\n", sti_plane_to_str(&hqvdp->plane));
759
760 /* Unregister VTG Vsync callback */
761 if (sti_vtg_unregister_client(hqvdp->vtg, &hqvdp->vtg_nb))
762 DRM_DEBUG_DRIVER("Warning: cannot unregister VTG notifier\n");
763
764 /* Set next cmd to NULL */
765 writel(0, hqvdp->regs + HQVDP_MBX_NEXT_CMD);
766
767 for (i = 0; i < POLL_MAX_ATTEMPT; i++) {
768 if (readl(hqvdp->regs + HQVDP_MBX_INFO_XP70)
769 & INFO_XP70_FW_READY)
770 break;
771 msleep(POLL_DELAY_MS);
772 }
773
774 /* VTG can stop now */
775 clk_disable_unprepare(hqvdp->clk_pix_main);
776
777 if (i == POLL_MAX_ATTEMPT)
778 DRM_ERROR("XP70 could not revert to idle\n");
779
780 hqvdp->plane.status = STI_PLANE_DISABLED;
781 hqvdp->vtg_registered = false;
782 }
783
784 /**
785 * sti_hqvdp_vtg_cb
786 * @nb: notifier block
787 * @evt: event message
788 * @data: private data
789 *
790 * Handle VTG Vsync event, display pending bottom field
791 *
792 * RETURNS:
793 * 0 on success.
794 */
sti_hqvdp_vtg_cb(struct notifier_block * nb,unsigned long evt,void * data)795 static int sti_hqvdp_vtg_cb(struct notifier_block *nb, unsigned long evt, void *data)
796 {
797 struct sti_hqvdp *hqvdp = container_of(nb, struct sti_hqvdp, vtg_nb);
798 int btm_cmd_offset, top_cmd_offest;
799 struct sti_hqvdp_cmd *btm_cmd, *top_cmd;
800
801 if ((evt != VTG_TOP_FIELD_EVENT) && (evt != VTG_BOTTOM_FIELD_EVENT)) {
802 DRM_DEBUG_DRIVER("Unknown event\n");
803 return 0;
804 }
805
806 if (hqvdp->plane.status == STI_PLANE_FLUSHING) {
807 /* disable need to be synchronize on vsync event */
808 DRM_DEBUG_DRIVER("Vsync event received => disable %s\n",
809 sti_plane_to_str(&hqvdp->plane));
810
811 sti_hqvdp_disable(hqvdp);
812 }
813
814 if (hqvdp->btm_field_pending) {
815 /* Create the btm field command from the current one */
816 btm_cmd_offset = sti_hqvdp_get_free_cmd(hqvdp);
817 top_cmd_offest = sti_hqvdp_get_curr_cmd(hqvdp);
818 if ((btm_cmd_offset == -1) || (top_cmd_offest == -1)) {
819 DRM_DEBUG_DRIVER("Warning: no cmd, will skip field\n");
820 return -EBUSY;
821 }
822
823 btm_cmd = hqvdp->hqvdp_cmd + btm_cmd_offset;
824 top_cmd = hqvdp->hqvdp_cmd + top_cmd_offest;
825
826 memcpy(btm_cmd, top_cmd, sizeof(*btm_cmd));
827
828 btm_cmd->top.config = TOP_CONFIG_INTER_BTM;
829 btm_cmd->top.current_luma +=
830 btm_cmd->top.luma_src_pitch / 2;
831 btm_cmd->top.current_chroma +=
832 btm_cmd->top.chroma_src_pitch / 2;
833
834 /* Post the command to mailbox */
835 writel(hqvdp->hqvdp_cmd_paddr + btm_cmd_offset,
836 hqvdp->regs + HQVDP_MBX_NEXT_CMD);
837
838 hqvdp->btm_field_pending = false;
839
840 dev_dbg(hqvdp->dev, "%s Posted command:0x%x\n",
841 __func__, hqvdp->hqvdp_cmd_paddr);
842
843 sti_plane_update_fps(&hqvdp->plane, false, true);
844 }
845
846 return 0;
847 }
848
sti_hqvdp_init(struct sti_hqvdp * hqvdp)849 static void sti_hqvdp_init(struct sti_hqvdp *hqvdp)
850 {
851 int size;
852 dma_addr_t dma_addr;
853
854 hqvdp->vtg_nb.notifier_call = sti_hqvdp_vtg_cb;
855
856 /* Allocate memory for the VDP commands */
857 size = NB_VDP_CMD * sizeof(struct sti_hqvdp_cmd);
858 hqvdp->hqvdp_cmd = dma_alloc_wc(hqvdp->dev, size,
859 &dma_addr,
860 GFP_KERNEL | GFP_DMA);
861 if (!hqvdp->hqvdp_cmd) {
862 DRM_ERROR("Failed to allocate memory for VDP cmd\n");
863 return;
864 }
865
866 hqvdp->hqvdp_cmd_paddr = (u32)dma_addr;
867 memset(hqvdp->hqvdp_cmd, 0, size);
868 }
869
sti_hqvdp_init_plugs(struct sti_hqvdp * hqvdp)870 static void sti_hqvdp_init_plugs(struct sti_hqvdp *hqvdp)
871 {
872 /* Configure Plugs (same for RD & WR) */
873 writel(PLUG_PAGE_SIZE_256, hqvdp->regs + HQVDP_RD_PLUG_PAGE_SIZE);
874 writel(PLUG_MIN_OPC_8, hqvdp->regs + HQVDP_RD_PLUG_MIN_OPC);
875 writel(PLUG_MAX_OPC_64, hqvdp->regs + HQVDP_RD_PLUG_MAX_OPC);
876 writel(PLUG_MAX_CHK_2X, hqvdp->regs + HQVDP_RD_PLUG_MAX_CHK);
877 writel(PLUG_MAX_MSG_1X, hqvdp->regs + HQVDP_RD_PLUG_MAX_MSG);
878 writel(PLUG_MIN_SPACE_1, hqvdp->regs + HQVDP_RD_PLUG_MIN_SPACE);
879 writel(PLUG_CONTROL_ENABLE, hqvdp->regs + HQVDP_RD_PLUG_CONTROL);
880
881 writel(PLUG_PAGE_SIZE_256, hqvdp->regs + HQVDP_WR_PLUG_PAGE_SIZE);
882 writel(PLUG_MIN_OPC_8, hqvdp->regs + HQVDP_WR_PLUG_MIN_OPC);
883 writel(PLUG_MAX_OPC_64, hqvdp->regs + HQVDP_WR_PLUG_MAX_OPC);
884 writel(PLUG_MAX_CHK_2X, hqvdp->regs + HQVDP_WR_PLUG_MAX_CHK);
885 writel(PLUG_MAX_MSG_1X, hqvdp->regs + HQVDP_WR_PLUG_MAX_MSG);
886 writel(PLUG_MIN_SPACE_1, hqvdp->regs + HQVDP_WR_PLUG_MIN_SPACE);
887 writel(PLUG_CONTROL_ENABLE, hqvdp->regs + HQVDP_WR_PLUG_CONTROL);
888 }
889
890 /**
891 * sti_hqvdp_start_xp70
892 * @hqvdp: hqvdp pointer
893 *
894 * Run the xP70 initialization sequence
895 */
sti_hqvdp_start_xp70(struct sti_hqvdp * hqvdp)896 static void sti_hqvdp_start_xp70(struct sti_hqvdp *hqvdp)
897 {
898 const struct firmware *firmware;
899 u32 *fw_rd_plug, *fw_wr_plug, *fw_pmem, *fw_dmem;
900 u8 *data;
901 int i;
902 struct fw_header {
903 int rd_size;
904 int wr_size;
905 int pmem_size;
906 int dmem_size;
907 } *header;
908
909 DRM_DEBUG_DRIVER("\n");
910
911 if (hqvdp->xp70_initialized) {
912 DRM_DEBUG_DRIVER("HQVDP XP70 already initialized\n");
913 return;
914 }
915
916 /* Request firmware */
917 if (request_firmware(&firmware, HQVDP_FMW_NAME, hqvdp->dev)) {
918 DRM_ERROR("Can't get HQVDP firmware\n");
919 return;
920 }
921
922 /* Check firmware parts */
923 if (!firmware) {
924 DRM_ERROR("Firmware not available\n");
925 return;
926 }
927
928 header = (struct fw_header *)firmware->data;
929 if (firmware->size < sizeof(*header)) {
930 DRM_ERROR("Invalid firmware size (%zu)\n", firmware->size);
931 goto out;
932 }
933 if ((sizeof(*header) + header->rd_size + header->wr_size +
934 header->pmem_size + header->dmem_size) != firmware->size) {
935 DRM_ERROR("Invalid fmw structure (%zu+%d+%d+%d+%d != %zu)\n",
936 sizeof(*header), header->rd_size, header->wr_size,
937 header->pmem_size, header->dmem_size,
938 firmware->size);
939 goto out;
940 }
941
942 data = (u8 *)firmware->data;
943 data += sizeof(*header);
944 fw_rd_plug = (void *)data;
945 data += header->rd_size;
946 fw_wr_plug = (void *)data;
947 data += header->wr_size;
948 fw_pmem = (void *)data;
949 data += header->pmem_size;
950 fw_dmem = (void *)data;
951
952 /* Enable clock */
953 if (clk_prepare_enable(hqvdp->clk))
954 DRM_ERROR("Failed to prepare/enable HQVDP clk\n");
955
956 /* Reset */
957 writel(SW_RESET_CTRL_FULL, hqvdp->regs + HQVDP_MBX_SW_RESET_CTRL);
958
959 for (i = 0; i < POLL_MAX_ATTEMPT; i++) {
960 if (readl(hqvdp->regs + HQVDP_MBX_STARTUP_CTRL1)
961 & STARTUP_CTRL1_RST_DONE)
962 break;
963 msleep(POLL_DELAY_MS);
964 }
965 if (i == POLL_MAX_ATTEMPT) {
966 DRM_ERROR("Could not reset\n");
967 clk_disable_unprepare(hqvdp->clk);
968 goto out;
969 }
970
971 /* Init Read & Write plugs */
972 for (i = 0; i < header->rd_size / 4; i++)
973 writel(fw_rd_plug[i], hqvdp->regs + HQVDP_RD_PLUG + i * 4);
974 for (i = 0; i < header->wr_size / 4; i++)
975 writel(fw_wr_plug[i], hqvdp->regs + HQVDP_WR_PLUG + i * 4);
976
977 sti_hqvdp_init_plugs(hqvdp);
978
979 /* Authorize Idle Mode */
980 writel(STARTUP_CTRL1_AUTH_IDLE, hqvdp->regs + HQVDP_MBX_STARTUP_CTRL1);
981
982 /* Prevent VTG interruption during the boot */
983 writel(SOFT_VSYNC_SW_CTRL_IRQ, hqvdp->regs + HQVDP_MBX_SOFT_VSYNC);
984 writel(0, hqvdp->regs + HQVDP_MBX_NEXT_CMD);
985
986 /* Download PMEM & DMEM */
987 for (i = 0; i < header->pmem_size / 4; i++)
988 writel(fw_pmem[i], hqvdp->regs + HQVDP_PMEM + i * 4);
989 for (i = 0; i < header->dmem_size / 4; i++)
990 writel(fw_dmem[i], hqvdp->regs + HQVDP_DMEM + i * 4);
991
992 /* Enable fetch */
993 writel(STARTUP_CTRL2_FETCH_EN, hqvdp->regs + HQVDP_MBX_STARTUP_CTRL2);
994
995 /* Wait end of boot */
996 for (i = 0; i < POLL_MAX_ATTEMPT; i++) {
997 if (readl(hqvdp->regs + HQVDP_MBX_INFO_XP70)
998 & INFO_XP70_FW_READY)
999 break;
1000 msleep(POLL_DELAY_MS);
1001 }
1002 if (i == POLL_MAX_ATTEMPT) {
1003 DRM_ERROR("Could not boot\n");
1004 clk_disable_unprepare(hqvdp->clk);
1005 goto out;
1006 }
1007
1008 /* Launch Vsync */
1009 writel(SOFT_VSYNC_HW, hqvdp->regs + HQVDP_MBX_SOFT_VSYNC);
1010
1011 DRM_INFO("HQVDP XP70 initialized\n");
1012
1013 hqvdp->xp70_initialized = true;
1014
1015 out:
1016 release_firmware(firmware);
1017 }
1018
sti_hqvdp_atomic_check(struct drm_plane * drm_plane,struct drm_atomic_state * state)1019 static int sti_hqvdp_atomic_check(struct drm_plane *drm_plane,
1020 struct drm_atomic_state *state)
1021 {
1022 struct drm_plane_state *new_plane_state = drm_atomic_get_new_plane_state(state,
1023 drm_plane);
1024 struct sti_plane *plane = to_sti_plane(drm_plane);
1025 struct sti_hqvdp *hqvdp = to_sti_hqvdp(plane);
1026 struct drm_crtc *crtc = new_plane_state->crtc;
1027 struct drm_framebuffer *fb = new_plane_state->fb;
1028 struct drm_crtc_state *crtc_state;
1029 struct drm_display_mode *mode;
1030 int dst_x, dst_y, dst_w, dst_h;
1031 int src_x, src_y, src_w, src_h;
1032
1033 /* no need for further checks if the plane is being disabled */
1034 if (!crtc || !fb)
1035 return 0;
1036
1037 crtc_state = drm_atomic_get_crtc_state(state, crtc);
1038 mode = &crtc_state->mode;
1039 dst_x = new_plane_state->crtc_x;
1040 dst_y = new_plane_state->crtc_y;
1041 dst_w = clamp_val(new_plane_state->crtc_w, 0, mode->hdisplay - dst_x);
1042 dst_h = clamp_val(new_plane_state->crtc_h, 0, mode->vdisplay - dst_y);
1043 /* src_x are in 16.16 format */
1044 src_x = new_plane_state->src_x >> 16;
1045 src_y = new_plane_state->src_y >> 16;
1046 src_w = new_plane_state->src_w >> 16;
1047 src_h = new_plane_state->src_h >> 16;
1048
1049 if (mode->clock && !sti_hqvdp_check_hw_scaling(hqvdp, mode,
1050 src_w, src_h,
1051 dst_w, dst_h)) {
1052 DRM_ERROR("Scaling beyond HW capabilities\n");
1053 return -EINVAL;
1054 }
1055
1056 if (!drm_fb_cma_get_gem_obj(fb, 0)) {
1057 DRM_ERROR("Can't get CMA GEM object for fb\n");
1058 return -EINVAL;
1059 }
1060
1061 /*
1062 * Input / output size
1063 * Align to upper even value
1064 */
1065 dst_w = ALIGN(dst_w, 2);
1066 dst_h = ALIGN(dst_h, 2);
1067
1068 if ((src_w > MAX_WIDTH) || (src_w < MIN_WIDTH) ||
1069 (src_h > MAX_HEIGHT) || (src_h < MIN_HEIGHT) ||
1070 (dst_w > MAX_WIDTH) || (dst_w < MIN_WIDTH) ||
1071 (dst_h > MAX_HEIGHT) || (dst_h < MIN_HEIGHT)) {
1072 DRM_ERROR("Invalid in/out size %dx%d -> %dx%d\n",
1073 src_w, src_h,
1074 dst_w, dst_h);
1075 return -EINVAL;
1076 }
1077
1078 if (!hqvdp->xp70_initialized)
1079 /* Start HQVDP XP70 coprocessor */
1080 sti_hqvdp_start_xp70(hqvdp);
1081
1082 if (!hqvdp->vtg_registered) {
1083 /* Prevent VTG shutdown */
1084 if (clk_prepare_enable(hqvdp->clk_pix_main)) {
1085 DRM_ERROR("Failed to prepare/enable pix main clk\n");
1086 return -EINVAL;
1087 }
1088
1089 /* Register VTG Vsync callback to handle bottom fields */
1090 if (sti_vtg_register_client(hqvdp->vtg,
1091 &hqvdp->vtg_nb,
1092 crtc)) {
1093 DRM_ERROR("Cannot register VTG notifier\n");
1094 clk_disable_unprepare(hqvdp->clk_pix_main);
1095 return -EINVAL;
1096 }
1097 hqvdp->vtg_registered = true;
1098 }
1099
1100 DRM_DEBUG_KMS("CRTC:%d (%s) drm plane:%d (%s)\n",
1101 crtc->base.id, sti_mixer_to_str(to_sti_mixer(crtc)),
1102 drm_plane->base.id, sti_plane_to_str(plane));
1103 DRM_DEBUG_KMS("%s dst=(%dx%d)@(%d,%d) - src=(%dx%d)@(%d,%d)\n",
1104 sti_plane_to_str(plane),
1105 dst_w, dst_h, dst_x, dst_y,
1106 src_w, src_h, src_x, src_y);
1107
1108 return 0;
1109 }
1110
sti_hqvdp_atomic_update(struct drm_plane * drm_plane,struct drm_atomic_state * state)1111 static void sti_hqvdp_atomic_update(struct drm_plane *drm_plane,
1112 struct drm_atomic_state *state)
1113 {
1114 struct drm_plane_state *oldstate = drm_atomic_get_old_plane_state(state,
1115 drm_plane);
1116 struct drm_plane_state *newstate = drm_atomic_get_new_plane_state(state,
1117 drm_plane);
1118 struct sti_plane *plane = to_sti_plane(drm_plane);
1119 struct sti_hqvdp *hqvdp = to_sti_hqvdp(plane);
1120 struct drm_crtc *crtc = newstate->crtc;
1121 struct drm_framebuffer *fb = newstate->fb;
1122 struct drm_display_mode *mode;
1123 int dst_x, dst_y, dst_w, dst_h;
1124 int src_x, src_y, src_w, src_h;
1125 struct drm_gem_cma_object *cma_obj;
1126 struct sti_hqvdp_cmd *cmd;
1127 int scale_h, scale_v;
1128 int cmd_offset;
1129
1130 if (!crtc || !fb)
1131 return;
1132
1133 if ((oldstate->fb == newstate->fb) &&
1134 (oldstate->crtc_x == newstate->crtc_x) &&
1135 (oldstate->crtc_y == newstate->crtc_y) &&
1136 (oldstate->crtc_w == newstate->crtc_w) &&
1137 (oldstate->crtc_h == newstate->crtc_h) &&
1138 (oldstate->src_x == newstate->src_x) &&
1139 (oldstate->src_y == newstate->src_y) &&
1140 (oldstate->src_w == newstate->src_w) &&
1141 (oldstate->src_h == newstate->src_h)) {
1142 /* No change since last update, do not post cmd */
1143 DRM_DEBUG_DRIVER("No change, not posting cmd\n");
1144 plane->status = STI_PLANE_UPDATED;
1145 return;
1146 }
1147
1148 mode = &crtc->mode;
1149 dst_x = newstate->crtc_x;
1150 dst_y = newstate->crtc_y;
1151 dst_w = clamp_val(newstate->crtc_w, 0, mode->hdisplay - dst_x);
1152 dst_h = clamp_val(newstate->crtc_h, 0, mode->vdisplay - dst_y);
1153 /* src_x are in 16.16 format */
1154 src_x = newstate->src_x >> 16;
1155 src_y = newstate->src_y >> 16;
1156 src_w = newstate->src_w >> 16;
1157 src_h = newstate->src_h >> 16;
1158
1159 cmd_offset = sti_hqvdp_get_free_cmd(hqvdp);
1160 if (cmd_offset == -1) {
1161 DRM_DEBUG_DRIVER("Warning: no cmd, will skip frame\n");
1162 return;
1163 }
1164 cmd = hqvdp->hqvdp_cmd + cmd_offset;
1165
1166 /* Static parameters, defaulting to progressive mode */
1167 cmd->top.config = TOP_CONFIG_PROGRESSIVE;
1168 cmd->top.mem_format = TOP_MEM_FORMAT_DFLT;
1169 cmd->hvsrc.param_ctrl = HVSRC_PARAM_CTRL_DFLT;
1170 cmd->csdi.config = CSDI_CONFIG_PROG;
1171
1172 /* VC1RE, FMD bypassed : keep everything set to 0
1173 * IQI/P2I bypassed */
1174 cmd->iqi.config = IQI_CONFIG_DFLT;
1175 cmd->iqi.con_bri = IQI_CON_BRI_DFLT;
1176 cmd->iqi.sat_gain = IQI_SAT_GAIN_DFLT;
1177 cmd->iqi.pxf_conf = IQI_PXF_CONF_DFLT;
1178
1179 cma_obj = drm_fb_cma_get_gem_obj(fb, 0);
1180
1181 DRM_DEBUG_DRIVER("drm FB:%d format:%.4s phys@:0x%lx\n", fb->base.id,
1182 (char *)&fb->format->format,
1183 (unsigned long)cma_obj->paddr);
1184
1185 /* Buffer planes address */
1186 cmd->top.current_luma = (u32)cma_obj->paddr + fb->offsets[0];
1187 cmd->top.current_chroma = (u32)cma_obj->paddr + fb->offsets[1];
1188
1189 /* Pitches */
1190 cmd->top.luma_processed_pitch = fb->pitches[0];
1191 cmd->top.luma_src_pitch = fb->pitches[0];
1192 cmd->top.chroma_processed_pitch = fb->pitches[1];
1193 cmd->top.chroma_src_pitch = fb->pitches[1];
1194
1195 /* Input / output size
1196 * Align to upper even value */
1197 dst_w = ALIGN(dst_w, 2);
1198 dst_h = ALIGN(dst_h, 2);
1199
1200 cmd->top.input_viewport_size = src_h << 16 | src_w;
1201 cmd->top.input_frame_size = src_h << 16 | src_w;
1202 cmd->hvsrc.output_picture_size = dst_h << 16 | dst_w;
1203 cmd->top.input_viewport_ori = src_y << 16 | src_x;
1204
1205 /* Handle interlaced */
1206 if (fb->flags & DRM_MODE_FB_INTERLACED) {
1207 /* Top field to display */
1208 cmd->top.config = TOP_CONFIG_INTER_TOP;
1209
1210 /* Update pitches and vert size */
1211 cmd->top.input_frame_size = (src_h / 2) << 16 | src_w;
1212 cmd->top.luma_processed_pitch *= 2;
1213 cmd->top.luma_src_pitch *= 2;
1214 cmd->top.chroma_processed_pitch *= 2;
1215 cmd->top.chroma_src_pitch *= 2;
1216
1217 /* Enable directional deinterlacing processing */
1218 cmd->csdi.config = CSDI_CONFIG_INTER_DIR;
1219 cmd->csdi.config2 = CSDI_CONFIG2_DFLT;
1220 cmd->csdi.dcdi_config = CSDI_DCDI_CONFIG_DFLT;
1221 }
1222
1223 /* Update hvsrc lut coef */
1224 scale_h = SCALE_FACTOR * dst_w / src_w;
1225 sti_hqvdp_update_hvsrc(HVSRC_HORI, scale_h, &cmd->hvsrc);
1226
1227 scale_v = SCALE_FACTOR * dst_h / src_h;
1228 sti_hqvdp_update_hvsrc(HVSRC_VERT, scale_v, &cmd->hvsrc);
1229
1230 writel(hqvdp->hqvdp_cmd_paddr + cmd_offset,
1231 hqvdp->regs + HQVDP_MBX_NEXT_CMD);
1232
1233 /* Interlaced : get ready to display the bottom field at next Vsync */
1234 if (fb->flags & DRM_MODE_FB_INTERLACED)
1235 hqvdp->btm_field_pending = true;
1236
1237 dev_dbg(hqvdp->dev, "%s Posted command:0x%x\n",
1238 __func__, hqvdp->hqvdp_cmd_paddr + cmd_offset);
1239
1240 sti_plane_update_fps(plane, true, true);
1241
1242 plane->status = STI_PLANE_UPDATED;
1243 }
1244
sti_hqvdp_atomic_disable(struct drm_plane * drm_plane,struct drm_atomic_state * state)1245 static void sti_hqvdp_atomic_disable(struct drm_plane *drm_plane,
1246 struct drm_atomic_state *state)
1247 {
1248 struct drm_plane_state *oldstate = drm_atomic_get_old_plane_state(state,
1249 drm_plane);
1250 struct sti_plane *plane = to_sti_plane(drm_plane);
1251
1252 if (!oldstate->crtc) {
1253 DRM_DEBUG_DRIVER("drm plane:%d not enabled\n",
1254 drm_plane->base.id);
1255 return;
1256 }
1257
1258 DRM_DEBUG_DRIVER("CRTC:%d (%s) drm plane:%d (%s)\n",
1259 oldstate->crtc->base.id,
1260 sti_mixer_to_str(to_sti_mixer(oldstate->crtc)),
1261 drm_plane->base.id, sti_plane_to_str(plane));
1262
1263 plane->status = STI_PLANE_DISABLING;
1264 }
1265
1266 static const struct drm_plane_helper_funcs sti_hqvdp_helpers_funcs = {
1267 .atomic_check = sti_hqvdp_atomic_check,
1268 .atomic_update = sti_hqvdp_atomic_update,
1269 .atomic_disable = sti_hqvdp_atomic_disable,
1270 };
1271
sti_hqvdp_late_register(struct drm_plane * drm_plane)1272 static int sti_hqvdp_late_register(struct drm_plane *drm_plane)
1273 {
1274 struct sti_plane *plane = to_sti_plane(drm_plane);
1275 struct sti_hqvdp *hqvdp = to_sti_hqvdp(plane);
1276
1277 hqvdp_debugfs_init(hqvdp, drm_plane->dev->primary);
1278
1279 return 0;
1280 }
1281
1282 static const struct drm_plane_funcs sti_hqvdp_plane_helpers_funcs = {
1283 .update_plane = drm_atomic_helper_update_plane,
1284 .disable_plane = drm_atomic_helper_disable_plane,
1285 .destroy = drm_plane_cleanup,
1286 .reset = drm_atomic_helper_plane_reset,
1287 .atomic_duplicate_state = drm_atomic_helper_plane_duplicate_state,
1288 .atomic_destroy_state = drm_atomic_helper_plane_destroy_state,
1289 .late_register = sti_hqvdp_late_register,
1290 };
1291
sti_hqvdp_create(struct drm_device * drm_dev,struct device * dev,int desc)1292 static struct drm_plane *sti_hqvdp_create(struct drm_device *drm_dev,
1293 struct device *dev, int desc)
1294 {
1295 struct sti_hqvdp *hqvdp = dev_get_drvdata(dev);
1296 int res;
1297
1298 hqvdp->plane.desc = desc;
1299 hqvdp->plane.status = STI_PLANE_DISABLED;
1300
1301 sti_hqvdp_init(hqvdp);
1302
1303 res = drm_universal_plane_init(drm_dev, &hqvdp->plane.drm_plane, 1,
1304 &sti_hqvdp_plane_helpers_funcs,
1305 hqvdp_supported_formats,
1306 ARRAY_SIZE(hqvdp_supported_formats),
1307 NULL, DRM_PLANE_TYPE_OVERLAY, NULL);
1308 if (res) {
1309 DRM_ERROR("Failed to initialize universal plane\n");
1310 return NULL;
1311 }
1312
1313 drm_plane_helper_add(&hqvdp->plane.drm_plane, &sti_hqvdp_helpers_funcs);
1314
1315 sti_plane_init_property(&hqvdp->plane, DRM_PLANE_TYPE_OVERLAY);
1316
1317 return &hqvdp->plane.drm_plane;
1318 }
1319
sti_hqvdp_bind(struct device * dev,struct device * master,void * data)1320 static int sti_hqvdp_bind(struct device *dev, struct device *master, void *data)
1321 {
1322 struct sti_hqvdp *hqvdp = dev_get_drvdata(dev);
1323 struct drm_device *drm_dev = data;
1324 struct drm_plane *plane;
1325
1326 DRM_DEBUG_DRIVER("\n");
1327
1328 hqvdp->drm_dev = drm_dev;
1329
1330 /* Create HQVDP plane once xp70 is initialized */
1331 plane = sti_hqvdp_create(drm_dev, hqvdp->dev, STI_HQVDP_0);
1332 if (!plane)
1333 DRM_ERROR("Can't create HQVDP plane\n");
1334
1335 return 0;
1336 }
1337
sti_hqvdp_unbind(struct device * dev,struct device * master,void * data)1338 static void sti_hqvdp_unbind(struct device *dev,
1339 struct device *master, void *data)
1340 {
1341 /* do nothing */
1342 }
1343
1344 static const struct component_ops sti_hqvdp_ops = {
1345 .bind = sti_hqvdp_bind,
1346 .unbind = sti_hqvdp_unbind,
1347 };
1348
sti_hqvdp_probe(struct platform_device * pdev)1349 static int sti_hqvdp_probe(struct platform_device *pdev)
1350 {
1351 struct device *dev = &pdev->dev;
1352 struct device_node *vtg_np;
1353 struct sti_hqvdp *hqvdp;
1354 struct resource *res;
1355
1356 DRM_DEBUG_DRIVER("\n");
1357
1358 hqvdp = devm_kzalloc(dev, sizeof(*hqvdp), GFP_KERNEL);
1359 if (!hqvdp) {
1360 DRM_ERROR("Failed to allocate HQVDP context\n");
1361 return -ENOMEM;
1362 }
1363
1364 hqvdp->dev = dev;
1365
1366 /* Get Memory resources */
1367 res = platform_get_resource(pdev, IORESOURCE_MEM, 0);
1368 if (!res) {
1369 DRM_ERROR("Get memory resource failed\n");
1370 return -ENXIO;
1371 }
1372 hqvdp->regs = devm_ioremap(dev, res->start, resource_size(res));
1373 if (!hqvdp->regs) {
1374 DRM_ERROR("Register mapping failed\n");
1375 return -ENXIO;
1376 }
1377
1378 /* Get clock resources */
1379 hqvdp->clk = devm_clk_get(dev, "hqvdp");
1380 hqvdp->clk_pix_main = devm_clk_get(dev, "pix_main");
1381 if (IS_ERR(hqvdp->clk) || IS_ERR(hqvdp->clk_pix_main)) {
1382 DRM_ERROR("Cannot get clocks\n");
1383 return -ENXIO;
1384 }
1385
1386 /* Get reset resources */
1387 hqvdp->reset = devm_reset_control_get(dev, "hqvdp");
1388 if (!IS_ERR(hqvdp->reset))
1389 reset_control_deassert(hqvdp->reset);
1390
1391 vtg_np = of_parse_phandle(pdev->dev.of_node, "st,vtg", 0);
1392 if (vtg_np)
1393 hqvdp->vtg = of_vtg_find(vtg_np);
1394 of_node_put(vtg_np);
1395
1396 platform_set_drvdata(pdev, hqvdp);
1397
1398 return component_add(&pdev->dev, &sti_hqvdp_ops);
1399 }
1400
sti_hqvdp_remove(struct platform_device * pdev)1401 static int sti_hqvdp_remove(struct platform_device *pdev)
1402 {
1403 component_del(&pdev->dev, &sti_hqvdp_ops);
1404 return 0;
1405 }
1406
1407 static const struct of_device_id hqvdp_of_match[] = {
1408 { .compatible = "st,stih407-hqvdp", },
1409 { /* end node */ }
1410 };
1411 MODULE_DEVICE_TABLE(of, hqvdp_of_match);
1412
1413 struct platform_driver sti_hqvdp_driver = {
1414 .driver = {
1415 .name = "sti-hqvdp",
1416 .owner = THIS_MODULE,
1417 .of_match_table = hqvdp_of_match,
1418 },
1419 .probe = sti_hqvdp_probe,
1420 .remove = sti_hqvdp_remove,
1421 };
1422
1423 MODULE_AUTHOR("Benjamin Gaignard <benjamin.gaignard@st.com>");
1424 MODULE_DESCRIPTION("STMicroelectronics SoC DRM driver");
1425 MODULE_LICENSE("GPL");
1426