1// SPDX-License-Identifier: (GPL-2.0 OR BSD-3-Clause)
2/*
3 * Wave5 series multi-standard codec IP - wave5 backend logic
4 *
5 * Copyright (C) 2021-2023 CHIPS&MEDIA INC
6 */
7
8#include <linux/iopoll.h>
9#include <linux/bitfield.h>
10#include "wave5-vpu.h"
11#include "wave5.h"
12#include "wave5-regdefine.h"
13
14#define FIO_TIMEOUT			10000000
15#define FIO_CTRL_READY			BIT(31)
16#define FIO_CTRL_WRITE			BIT(16)
17#define VPU_BUSY_CHECK_TIMEOUT		10000000
18#define QUEUE_REPORT_MASK		0xffff
19
20/* Encoder support fields */
21#define FEATURE_HEVC10BIT_ENC		BIT(3)
22#define FEATURE_AVC10BIT_ENC		BIT(11)
23#define FEATURE_AVC_ENCODER		BIT(1)
24#define FEATURE_HEVC_ENCODER		BIT(0)
25
26/* Decoder support fields */
27#define FEATURE_AVC_DECODER		BIT(3)
28#define FEATURE_HEVC_DECODER		BIT(2)
29
30#define FEATURE_BACKBONE		BIT(16)
31#define FEATURE_VCORE_BACKBONE		BIT(22)
32#define FEATURE_VCPU_BACKBONE		BIT(28)
33
34#define REMAP_CTRL_MAX_SIZE_BITS	((W5_REMAP_MAX_SIZE >> 12) & 0x1ff)
35#define REMAP_CTRL_REGISTER_VALUE(index)	(			\
36	(BIT(31) | (index << 12) | BIT(11) | REMAP_CTRL_MAX_SIZE_BITS)	\
37)
38
39#define FASTIO_ADDRESS_MASK		GENMASK(15, 0)
40#define SEQ_PARAM_PROFILE_MASK		GENMASK(30, 24)
41
42static void _wave5_print_reg_err(struct vpu_device *vpu_dev, u32 reg_fail_reason,
43				 const char *func);
44#define PRINT_REG_ERR(dev, reason)	_wave5_print_reg_err((dev), (reason), __func__)
45
46static inline const char *cmd_to_str(int cmd, bool is_dec)
47{
48	switch (cmd) {
49	case W5_INIT_VPU:
50		return "W5_INIT_VPU";
51	case W5_WAKEUP_VPU:
52		return "W5_WAKEUP_VPU";
53	case W5_SLEEP_VPU:
54		return "W5_SLEEP_VPU";
55	case W5_CREATE_INSTANCE:
56		return "W5_CREATE_INSTANCE";
57	case W5_FLUSH_INSTANCE:
58		return "W5_FLUSH_INSTANCE";
59	case W5_DESTROY_INSTANCE:
60		return "W5_DESTROY_INSTANCE";
61	case W5_INIT_SEQ:
62		return "W5_INIT_SEQ";
63	case W5_SET_FB:
64		return "W5_SET_FB";
65	case W5_DEC_ENC_PIC:
66		if (is_dec)
67			return "W5_DEC_PIC";
68		return "W5_ENC_PIC";
69	case W5_ENC_SET_PARAM:
70		return "W5_ENC_SET_PARAM";
71	case W5_QUERY:
72		return "W5_QUERY";
73	case W5_UPDATE_BS:
74		return "W5_UPDATE_BS";
75	case W5_MAX_VPU_COMD:
76		return "W5_MAX_VPU_COMD";
77	default:
78		return "UNKNOWN";
79	}
80}
81
82static void _wave5_print_reg_err(struct vpu_device *vpu_dev, u32 reg_fail_reason,
83				 const char *func)
84{
85	struct device *dev = vpu_dev->dev;
86	u32 reg_val;
87
88	switch (reg_fail_reason) {
89	case WAVE5_SYSERR_QUEUEING_FAIL:
90		reg_val = vpu_read_reg(vpu_dev, W5_RET_QUEUE_FAIL_REASON);
91		dev_dbg(dev, "%s: queueing failure: 0x%x\n", func, reg_val);
92		break;
93	case WAVE5_SYSERR_RESULT_NOT_READY:
94		dev_err(dev, "%s: result not ready: 0x%x\n", func, reg_fail_reason);
95		break;
96	case WAVE5_SYSERR_ACCESS_VIOLATION_HW:
97		dev_err(dev, "%s: access violation: 0x%x\n", func, reg_fail_reason);
98		break;
99	case WAVE5_SYSERR_WATCHDOG_TIMEOUT:
100		dev_err(dev, "%s: watchdog timeout: 0x%x\n", func, reg_fail_reason);
101		break;
102	case WAVE5_SYSERR_BUS_ERROR:
103		dev_err(dev, "%s: bus error: 0x%x\n", func, reg_fail_reason);
104		break;
105	case WAVE5_SYSERR_DOUBLE_FAULT:
106		dev_err(dev, "%s: double fault: 0x%x\n", func, reg_fail_reason);
107		break;
108	case WAVE5_SYSERR_VPU_STILL_RUNNING:
109		dev_err(dev, "%s: still running: 0x%x\n", func, reg_fail_reason);
110		break;
111	case WAVE5_SYSERR_VLC_BUF_FULL:
112		dev_err(dev, "%s: vlc buf full: 0x%x\n", func, reg_fail_reason);
113		break;
114	default:
115		dev_err(dev, "%s: failure:: 0x%x\n", func, reg_fail_reason);
116		break;
117	}
118}
119
120static int wave5_wait_fio_readl(struct vpu_device *vpu_dev, u32 addr, u32 val)
121{
122	u32 ctrl;
123	int ret;
124
125	ctrl = addr & 0xffff;
126	wave5_vdi_write_register(vpu_dev, W5_VPU_FIO_CTRL_ADDR, ctrl);
127	ret = read_poll_timeout(wave5_vdi_read_register, ctrl, ctrl & FIO_CTRL_READY,
128				0, FIO_TIMEOUT, false, vpu_dev, W5_VPU_FIO_CTRL_ADDR);
129	if (ret)
130		return ret;
131
132	if (wave5_vdi_read_register(vpu_dev, W5_VPU_FIO_DATA) != val)
133		return -ETIMEDOUT;
134
135	return 0;
136}
137
138static void wave5_fio_writel(struct vpu_device *vpu_dev, unsigned int addr, unsigned int data)
139{
140	int ret;
141	unsigned int ctrl;
142
143	wave5_vdi_write_register(vpu_dev, W5_VPU_FIO_DATA, data);
144	ctrl = FIELD_GET(FASTIO_ADDRESS_MASK, addr);
145	ctrl |= FIO_CTRL_WRITE;
146	wave5_vdi_write_register(vpu_dev, W5_VPU_FIO_CTRL_ADDR, ctrl);
147	ret = read_poll_timeout(wave5_vdi_read_register, ctrl, ctrl & FIO_CTRL_READY, 0,
148				FIO_TIMEOUT, false, vpu_dev, W5_VPU_FIO_CTRL_ADDR);
149	if (ret)
150		dev_dbg_ratelimited(vpu_dev->dev, "FIO write timeout: addr=0x%x data=%x\n",
151				    ctrl, data);
152}
153
154static int wave5_wait_bus_busy(struct vpu_device *vpu_dev, unsigned int addr)
155{
156	u32 gdi_status_check_value = 0x3f;
157
158	if (vpu_dev->product_code == WAVE521C_CODE ||
159	    vpu_dev->product_code == WAVE521_CODE ||
160	    vpu_dev->product_code == WAVE521E1_CODE)
161		gdi_status_check_value = 0x00ff1f3f;
162
163	return wave5_wait_fio_readl(vpu_dev, addr, gdi_status_check_value);
164}
165
166static int wave5_wait_vpu_busy(struct vpu_device *vpu_dev, unsigned int addr)
167{
168	u32 data;
169
170	return read_poll_timeout(wave5_vdi_read_register, data, data == 0,
171				 0, VPU_BUSY_CHECK_TIMEOUT, false, vpu_dev, addr);
172}
173
174static int wave5_wait_vcpu_bus_busy(struct vpu_device *vpu_dev, unsigned int addr)
175{
176	return wave5_wait_fio_readl(vpu_dev, addr, 0);
177}
178
179bool wave5_vpu_is_init(struct vpu_device *vpu_dev)
180{
181	return vpu_read_reg(vpu_dev, W5_VCPU_CUR_PC) != 0;
182}
183
184unsigned int wave5_vpu_get_product_id(struct vpu_device *vpu_dev)
185{
186	u32 val = vpu_read_reg(vpu_dev, W5_PRODUCT_NUMBER);
187
188	switch (val) {
189	case WAVE521C_CODE:
190		return PRODUCT_ID_521;
191	case WAVE521_CODE:
192	case WAVE521C_DUAL_CODE:
193	case WAVE521E1_CODE:
194	case WAVE511_CODE:
195	case WAVE517_CODE:
196	case WAVE537_CODE:
197		dev_err(vpu_dev->dev, "Unsupported product id (%x)\n", val);
198		break;
199	default:
200		dev_err(vpu_dev->dev, "Invalid product id (%x)\n", val);
201		break;
202	}
203
204	return PRODUCT_ID_NONE;
205}
206
207static void wave5_bit_issue_command(struct vpu_device *vpu_dev, struct vpu_instance *inst, u32 cmd)
208{
209	u32 instance_index;
210	u32 codec_mode;
211
212	if (inst) {
213		instance_index = inst->id;
214		codec_mode = inst->std;
215
216		vpu_write_reg(vpu_dev, W5_CMD_INSTANCE_INFO, (codec_mode << 16) |
217			      (instance_index & 0xffff));
218		vpu_write_reg(vpu_dev, W5_VPU_BUSY_STATUS, 1);
219	}
220
221	vpu_write_reg(vpu_dev, W5_COMMAND, cmd);
222
223	if (inst) {
224		dev_dbg(vpu_dev->dev, "%s: cmd=0x%x (%s)\n", __func__, cmd,
225			cmd_to_str(cmd, inst->type == VPU_INST_TYPE_DEC));
226	} else {
227		dev_dbg(vpu_dev->dev, "%s: cmd=0x%x\n", __func__, cmd);
228	}
229
230	vpu_write_reg(vpu_dev, W5_VPU_HOST_INT_REQ, 1);
231}
232
233static int wave5_vpu_firmware_command_queue_error_check(struct vpu_device *dev, u32 *fail_res)
234{
235	u32 reason = 0;
236
237	/* Check if we were able to add a command into the VCPU QUEUE */
238	if (!vpu_read_reg(dev, W5_RET_SUCCESS)) {
239		reason = vpu_read_reg(dev, W5_RET_FAIL_REASON);
240		PRINT_REG_ERR(dev, reason);
241
242		/*
243		 * The fail_res argument will be either NULL or 0.
244		 * If the fail_res argument is NULL, then just return -EIO.
245		 * Otherwise, assign the reason to fail_res, so that the
246		 * calling function can use it.
247		 */
248		if (fail_res)
249			*fail_res = reason;
250		else
251			return -EIO;
252
253		if (reason == WAVE5_SYSERR_VPU_STILL_RUNNING)
254			return -EBUSY;
255	}
256	return 0;
257}
258
259static int send_firmware_command(struct vpu_instance *inst, u32 cmd, bool check_success,
260				 u32 *queue_status, u32 *fail_result)
261{
262	int ret;
263
264	wave5_bit_issue_command(inst->dev, inst, cmd);
265	ret = wave5_wait_vpu_busy(inst->dev, W5_VPU_BUSY_STATUS);
266	if (ret) {
267		dev_warn(inst->dev->dev, "%s: command: '%s', timed out\n", __func__,
268			 cmd_to_str(cmd, inst->type == VPU_INST_TYPE_DEC));
269		return -ETIMEDOUT;
270	}
271
272	if (queue_status)
273		*queue_status = vpu_read_reg(inst->dev, W5_RET_QUEUE_STATUS);
274
275	/* In some cases we want to send multiple commands before checking
276	 * whether they are queued properly
277	 */
278	if (!check_success)
279		return 0;
280
281	return wave5_vpu_firmware_command_queue_error_check(inst->dev, fail_result);
282}
283
284static int wave5_send_query(struct vpu_device *vpu_dev, struct vpu_instance *inst,
285			    enum query_opt query_opt)
286{
287	int ret;
288
289	vpu_write_reg(vpu_dev, W5_QUERY_OPTION, query_opt);
290	vpu_write_reg(vpu_dev, W5_VPU_BUSY_STATUS, 1);
291	wave5_bit_issue_command(vpu_dev, inst, W5_QUERY);
292
293	ret = wave5_wait_vpu_busy(vpu_dev, W5_VPU_BUSY_STATUS);
294	if (ret) {
295		dev_warn(vpu_dev->dev, "command: 'W5_QUERY', timed out opt=0x%x\n", query_opt);
296		return ret;
297	}
298
299	return wave5_vpu_firmware_command_queue_error_check(vpu_dev, NULL);
300}
301
302static int setup_wave5_properties(struct device *dev)
303{
304	struct vpu_device *vpu_dev = dev_get_drvdata(dev);
305	struct vpu_attr *p_attr = &vpu_dev->attr;
306	u32 reg_val;
307	u8 *str;
308	int ret;
309	u32 hw_config_def0, hw_config_def1, hw_config_feature;
310
311	ret = wave5_send_query(vpu_dev, NULL, GET_VPU_INFO);
312	if (ret)
313		return ret;
314
315	reg_val = vpu_read_reg(vpu_dev, W5_RET_PRODUCT_NAME);
316	str = (u8 *)&reg_val;
317	p_attr->product_name[0] = str[3];
318	p_attr->product_name[1] = str[2];
319	p_attr->product_name[2] = str[1];
320	p_attr->product_name[3] = str[0];
321	p_attr->product_name[4] = 0;
322
323	p_attr->product_id = wave5_vpu_get_product_id(vpu_dev);
324	p_attr->product_version = vpu_read_reg(vpu_dev, W5_RET_PRODUCT_VERSION);
325	p_attr->fw_version = vpu_read_reg(vpu_dev, W5_RET_FW_VERSION);
326	p_attr->customer_id = vpu_read_reg(vpu_dev, W5_RET_CUSTOMER_ID);
327	hw_config_def0 = vpu_read_reg(vpu_dev, W5_RET_STD_DEF0);
328	hw_config_def1 = vpu_read_reg(vpu_dev, W5_RET_STD_DEF1);
329	hw_config_feature = vpu_read_reg(vpu_dev, W5_RET_CONF_FEATURE);
330
331	p_attr->support_hevc10bit_enc = FIELD_GET(FEATURE_HEVC10BIT_ENC, hw_config_feature);
332	p_attr->support_avc10bit_enc = FIELD_GET(FEATURE_AVC10BIT_ENC, hw_config_feature);
333
334	p_attr->support_decoders = FIELD_GET(FEATURE_AVC_DECODER, hw_config_def1) << STD_AVC;
335	p_attr->support_decoders |= FIELD_GET(FEATURE_HEVC_DECODER, hw_config_def1) << STD_HEVC;
336	p_attr->support_encoders = FIELD_GET(FEATURE_AVC_ENCODER, hw_config_def1) << STD_AVC;
337	p_attr->support_encoders |= FIELD_GET(FEATURE_HEVC_ENCODER, hw_config_def1) << STD_HEVC;
338
339	p_attr->support_backbone = FIELD_GET(FEATURE_BACKBONE, hw_config_def0);
340	p_attr->support_vcpu_backbone = FIELD_GET(FEATURE_VCPU_BACKBONE, hw_config_def0);
341	p_attr->support_vcore_backbone = FIELD_GET(FEATURE_VCORE_BACKBONE, hw_config_def0);
342
343	return 0;
344}
345
346int wave5_vpu_get_version(struct vpu_device *vpu_dev, u32 *revision)
347{
348	u32 reg_val;
349	int ret;
350
351	ret = wave5_send_query(vpu_dev, NULL, GET_VPU_INFO);
352	if (ret)
353		return ret;
354
355	reg_val = vpu_read_reg(vpu_dev, W5_RET_FW_VERSION);
356	if (revision) {
357		*revision = reg_val;
358		return 0;
359	}
360
361	return -EINVAL;
362}
363
364static void remap_page(struct vpu_device *vpu_dev, dma_addr_t code_base, u32 index)
365{
366	vpu_write_reg(vpu_dev, W5_VPU_REMAP_CTRL, REMAP_CTRL_REGISTER_VALUE(index));
367	vpu_write_reg(vpu_dev, W5_VPU_REMAP_VADDR, index * W5_REMAP_MAX_SIZE);
368	vpu_write_reg(vpu_dev, W5_VPU_REMAP_PADDR, code_base + index * W5_REMAP_MAX_SIZE);
369}
370
371int wave5_vpu_init(struct device *dev, u8 *fw, size_t size)
372{
373	struct vpu_buf *common_vb;
374	dma_addr_t code_base, temp_base;
375	u32 code_size, temp_size;
376	u32 i, reg_val, reason_code;
377	int ret;
378	struct vpu_device *vpu_dev = dev_get_drvdata(dev);
379
380	common_vb = &vpu_dev->common_mem;
381
382	code_base = common_vb->daddr;
383	/* ALIGN TO 4KB */
384	code_size = (WAVE5_MAX_CODE_BUF_SIZE & ~0xfff);
385	if (code_size < size * 2)
386		return -EINVAL;
387
388	temp_base = common_vb->daddr + WAVE5_TEMPBUF_OFFSET;
389	temp_size = WAVE5_TEMPBUF_SIZE;
390
391	ret = wave5_vdi_write_memory(vpu_dev, common_vb, 0, fw, size);
392	if (ret < 0) {
393		dev_err(vpu_dev->dev, "VPU init, Writing firmware to common buffer, fail: %d\n",
394			ret);
395		return ret;
396	}
397
398	vpu_write_reg(vpu_dev, W5_PO_CONF, 0);
399
400	/* clear registers */
401
402	for (i = W5_CMD_REG_BASE; i < W5_CMD_REG_END; i += 4)
403		vpu_write_reg(vpu_dev, i, 0x00);
404
405	remap_page(vpu_dev, code_base, W5_REMAP_INDEX0);
406	remap_page(vpu_dev, code_base, W5_REMAP_INDEX1);
407
408	vpu_write_reg(vpu_dev, W5_ADDR_CODE_BASE, code_base);
409	vpu_write_reg(vpu_dev, W5_CODE_SIZE, code_size);
410	vpu_write_reg(vpu_dev, W5_CODE_PARAM, (WAVE5_UPPER_PROC_AXI_ID << 4) | 0);
411	vpu_write_reg(vpu_dev, W5_ADDR_TEMP_BASE, temp_base);
412	vpu_write_reg(vpu_dev, W5_TEMP_SIZE, temp_size);
413
414	/* These register must be reset explicitly */
415	vpu_write_reg(vpu_dev, W5_HW_OPTION, 0);
416	wave5_fio_writel(vpu_dev, W5_BACKBONE_PROC_EXT_ADDR, 0);
417	wave5_fio_writel(vpu_dev, W5_BACKBONE_AXI_PARAM, 0);
418	vpu_write_reg(vpu_dev, W5_SEC_AXI_PARAM, 0);
419
420	/* Encoder interrupt */
421	reg_val = BIT(INT_WAVE5_ENC_SET_PARAM);
422	reg_val |= BIT(INT_WAVE5_ENC_PIC);
423	reg_val |= BIT(INT_WAVE5_BSBUF_FULL);
424	/* Decoder interrupt */
425	reg_val |= BIT(INT_WAVE5_INIT_SEQ);
426	reg_val |= BIT(INT_WAVE5_DEC_PIC);
427	reg_val |= BIT(INT_WAVE5_BSBUF_EMPTY);
428	vpu_write_reg(vpu_dev, W5_VPU_VINT_ENABLE, reg_val);
429
430	reg_val = vpu_read_reg(vpu_dev, W5_VPU_RET_VPU_CONFIG0);
431	if (FIELD_GET(FEATURE_BACKBONE, reg_val)) {
432		reg_val = ((WAVE5_PROC_AXI_ID << 28) |
433			   (WAVE5_PRP_AXI_ID << 24) |
434			   (WAVE5_FBD_Y_AXI_ID << 20) |
435			   (WAVE5_FBC_Y_AXI_ID << 16) |
436			   (WAVE5_FBD_C_AXI_ID << 12) |
437			   (WAVE5_FBC_C_AXI_ID << 8) |
438			   (WAVE5_PRI_AXI_ID << 4) |
439			   WAVE5_SEC_AXI_ID);
440		wave5_fio_writel(vpu_dev, W5_BACKBONE_PROG_AXI_ID, reg_val);
441	}
442
443	vpu_write_reg(vpu_dev, W5_VPU_BUSY_STATUS, 1);
444	vpu_write_reg(vpu_dev, W5_COMMAND, W5_INIT_VPU);
445	vpu_write_reg(vpu_dev, W5_VPU_REMAP_CORE_START, 1);
446	ret = wave5_wait_vpu_busy(vpu_dev, W5_VPU_BUSY_STATUS);
447	if (ret) {
448		dev_err(vpu_dev->dev, "VPU init(W5_VPU_REMAP_CORE_START) timeout\n");
449		return ret;
450	}
451
452	ret = wave5_vpu_firmware_command_queue_error_check(vpu_dev, &reason_code);
453	if (ret)
454		return ret;
455
456	return setup_wave5_properties(dev);
457}
458
459int wave5_vpu_build_up_dec_param(struct vpu_instance *inst,
460				 struct dec_open_param *param)
461{
462	int ret;
463	struct dec_info *p_dec_info = &inst->codec_info->dec_info;
464	struct vpu_device *vpu_dev = inst->dev;
465
466	p_dec_info->cycle_per_tick = 256;
467	if (vpu_dev->sram_buf.size) {
468		p_dec_info->sec_axi_info.use_bit_enable = 1;
469		p_dec_info->sec_axi_info.use_ip_enable = 1;
470		p_dec_info->sec_axi_info.use_lf_row_enable = 1;
471	}
472	switch (inst->std) {
473	case W_HEVC_DEC:
474		p_dec_info->seq_change_mask = SEQ_CHANGE_ENABLE_ALL_HEVC;
475		break;
476	case W_AVC_DEC:
477		p_dec_info->seq_change_mask = SEQ_CHANGE_ENABLE_ALL_AVC;
478		break;
479	default:
480		return -EINVAL;
481	}
482
483	p_dec_info->vb_work.size = WAVE521DEC_WORKBUF_SIZE;
484	ret = wave5_vdi_allocate_dma_memory(inst->dev, &p_dec_info->vb_work);
485	if (ret)
486		return ret;
487
488	vpu_write_reg(inst->dev, W5_CMD_DEC_VCORE_INFO, 1);
489
490	wave5_vdi_clear_memory(inst->dev, &p_dec_info->vb_work);
491
492	vpu_write_reg(inst->dev, W5_ADDR_WORK_BASE, p_dec_info->vb_work.daddr);
493	vpu_write_reg(inst->dev, W5_WORK_SIZE, p_dec_info->vb_work.size);
494
495	vpu_write_reg(inst->dev, W5_CMD_ADDR_SEC_AXI, vpu_dev->sram_buf.daddr);
496	vpu_write_reg(inst->dev, W5_CMD_SEC_AXI_SIZE, vpu_dev->sram_buf.size);
497
498	vpu_write_reg(inst->dev, W5_CMD_DEC_BS_START_ADDR, p_dec_info->stream_buf_start_addr);
499	vpu_write_reg(inst->dev, W5_CMD_DEC_BS_SIZE, p_dec_info->stream_buf_size);
500
501	/* NOTE: SDMA reads MSB first */
502	vpu_write_reg(inst->dev, W5_CMD_BS_PARAM, BITSTREAM_ENDIANNESS_BIG_ENDIAN);
503	/* This register must be reset explicitly */
504	vpu_write_reg(inst->dev, W5_CMD_EXT_ADDR, 0);
505	vpu_write_reg(inst->dev, W5_CMD_NUM_CQ_DEPTH_M1, (COMMAND_QUEUE_DEPTH - 1));
506
507	ret = send_firmware_command(inst, W5_CREATE_INSTANCE, true, NULL, NULL);
508	if (ret) {
509		wave5_vdi_free_dma_memory(vpu_dev, &p_dec_info->vb_work);
510		return ret;
511	}
512
513	p_dec_info->product_code = vpu_read_reg(inst->dev, W5_PRODUCT_NUMBER);
514
515	return 0;
516}
517
518int wave5_vpu_hw_flush_instance(struct vpu_instance *inst)
519{
520	struct dec_info *p_dec_info = &inst->codec_info->dec_info;
521	u32 instance_queue_count, report_queue_count;
522	u32 reg_val = 0;
523	u32 fail_res = 0;
524	int ret;
525
526	ret = send_firmware_command(inst, W5_FLUSH_INSTANCE, true, &reg_val, &fail_res);
527	if (ret)
528		return ret;
529
530	instance_queue_count = (reg_val >> 16) & 0xff;
531	report_queue_count = (reg_val & QUEUE_REPORT_MASK);
532	if (instance_queue_count != 0 || report_queue_count != 0) {
533		dev_warn(inst->dev->dev,
534			 "FLUSH_INSTANCE cmd didn't reset the amount of queued commands & reports");
535	}
536
537	/* reset our local copy of the counts */
538	p_dec_info->instance_queue_count = 0;
539	p_dec_info->report_queue_count = 0;
540
541	return 0;
542}
543
544static u32 get_bitstream_options(struct dec_info *info)
545{
546	u32 bs_option = BSOPTION_ENABLE_EXPLICIT_END;
547
548	if (info->stream_endflag)
549		bs_option |= BSOPTION_HIGHLIGHT_STREAM_END;
550	return bs_option;
551}
552
553int wave5_vpu_dec_init_seq(struct vpu_instance *inst)
554{
555	struct dec_info *p_dec_info = &inst->codec_info->dec_info;
556	u32 cmd_option = INIT_SEQ_NORMAL;
557	u32 reg_val, fail_res;
558	int ret;
559
560	if (!inst->codec_info)
561		return -EINVAL;
562
563	vpu_write_reg(inst->dev, W5_BS_RD_PTR, p_dec_info->stream_rd_ptr);
564	vpu_write_reg(inst->dev, W5_BS_WR_PTR, p_dec_info->stream_wr_ptr);
565
566	vpu_write_reg(inst->dev, W5_BS_OPTION, get_bitstream_options(p_dec_info));
567
568	vpu_write_reg(inst->dev, W5_COMMAND_OPTION, cmd_option);
569	vpu_write_reg(inst->dev, W5_CMD_DEC_USER_MASK, p_dec_info->user_data_enable);
570
571	ret = send_firmware_command(inst, W5_INIT_SEQ, true, &reg_val, &fail_res);
572	if (ret)
573		return ret;
574
575	p_dec_info->instance_queue_count = (reg_val >> 16) & 0xff;
576	p_dec_info->report_queue_count = (reg_val & QUEUE_REPORT_MASK);
577
578	dev_dbg(inst->dev->dev, "%s: init seq sent (queue %u : %u)\n", __func__,
579		p_dec_info->instance_queue_count, p_dec_info->report_queue_count);
580
581	return 0;
582}
583
584static void wave5_get_dec_seq_result(struct vpu_instance *inst, struct dec_initial_info *info)
585{
586	u32 reg_val;
587	u32 profile_compatibility_flag;
588	struct dec_info *p_dec_info = &inst->codec_info->dec_info;
589
590	p_dec_info->stream_rd_ptr = wave5_dec_get_rd_ptr(inst);
591	info->rd_ptr = p_dec_info->stream_rd_ptr;
592
593	p_dec_info->frame_display_flag = vpu_read_reg(inst->dev, W5_RET_DEC_DISP_IDC);
594
595	reg_val = vpu_read_reg(inst->dev, W5_RET_DEC_PIC_SIZE);
596	info->pic_width = ((reg_val >> 16) & 0xffff);
597	info->pic_height = (reg_val & 0xffff);
598	info->min_frame_buffer_count = vpu_read_reg(inst->dev, W5_RET_DEC_NUM_REQUIRED_FB);
599
600	reg_val = vpu_read_reg(inst->dev, W5_RET_DEC_CROP_LEFT_RIGHT);
601	info->pic_crop_rect.left = (reg_val >> 16) & 0xffff;
602	info->pic_crop_rect.right = reg_val & 0xffff;
603	reg_val = vpu_read_reg(inst->dev, W5_RET_DEC_CROP_TOP_BOTTOM);
604	info->pic_crop_rect.top = (reg_val >> 16) & 0xffff;
605	info->pic_crop_rect.bottom = reg_val & 0xffff;
606
607	reg_val = vpu_read_reg(inst->dev, W5_RET_DEC_COLOR_SAMPLE_INFO);
608	info->luma_bitdepth = reg_val & 0xf;
609	info->chroma_bitdepth = (reg_val >> 4) & 0xf;
610
611	reg_val = vpu_read_reg(inst->dev, W5_RET_DEC_SEQ_PARAM);
612	profile_compatibility_flag = (reg_val >> 12) & 0xff;
613	info->profile = (reg_val >> 24) & 0x1f;
614
615	if (inst->std == W_HEVC_DEC) {
616		/* guessing profile */
617		if (!info->profile) {
618			if ((profile_compatibility_flag & 0x06) == 0x06)
619				info->profile = HEVC_PROFILE_MAIN; /* main profile */
620			else if (profile_compatibility_flag & 0x04)
621				info->profile = HEVC_PROFILE_MAIN10; /* main10 profile */
622			else if (profile_compatibility_flag & 0x08)
623				/* main still picture profile */
624				info->profile = HEVC_PROFILE_STILLPICTURE;
625			else
626				info->profile = HEVC_PROFILE_MAIN; /* for old version HM */
627		}
628	} else if (inst->std == W_AVC_DEC) {
629		info->profile = FIELD_GET(SEQ_PARAM_PROFILE_MASK, reg_val);
630	}
631
632	info->vlc_buf_size = vpu_read_reg(inst->dev, W5_RET_VLC_BUF_SIZE);
633	info->param_buf_size = vpu_read_reg(inst->dev, W5_RET_PARAM_BUF_SIZE);
634	p_dec_info->vlc_buf_size = info->vlc_buf_size;
635	p_dec_info->param_buf_size = info->param_buf_size;
636}
637
638int wave5_vpu_dec_get_seq_info(struct vpu_instance *inst, struct dec_initial_info *info)
639{
640	int ret;
641	u32 reg_val;
642	struct dec_info *p_dec_info = &inst->codec_info->dec_info;
643
644	vpu_write_reg(inst->dev, W5_CMD_DEC_ADDR_REPORT_BASE, p_dec_info->user_data_buf_addr);
645	vpu_write_reg(inst->dev, W5_CMD_DEC_REPORT_SIZE, p_dec_info->user_data_buf_size);
646	vpu_write_reg(inst->dev, W5_CMD_DEC_REPORT_PARAM, REPORT_PARAM_ENDIANNESS_BIG_ENDIAN);
647
648	/* send QUERY cmd */
649	ret = wave5_send_query(inst->dev, inst, GET_RESULT);
650	if (ret)
651		return ret;
652
653	reg_val = vpu_read_reg(inst->dev, W5_RET_QUEUE_STATUS);
654
655	p_dec_info->instance_queue_count = (reg_val >> 16) & 0xff;
656	p_dec_info->report_queue_count = (reg_val & QUEUE_REPORT_MASK);
657
658	dev_dbg(inst->dev->dev, "%s: init seq complete (queue %u : %u)\n", __func__,
659		p_dec_info->instance_queue_count, p_dec_info->report_queue_count);
660
661	/* this is not a fatal error, set ret to -EIO but don't return immediately */
662	if (vpu_read_reg(inst->dev, W5_RET_DEC_DECODING_SUCCESS) != 1) {
663		info->seq_init_err_reason = vpu_read_reg(inst->dev, W5_RET_DEC_ERR_INFO);
664		ret = -EIO;
665	}
666
667	wave5_get_dec_seq_result(inst, info);
668
669	return ret;
670}
671
672int wave5_vpu_dec_register_framebuffer(struct vpu_instance *inst, struct frame_buffer *fb_arr,
673				       enum tiled_map_type map_type, unsigned int count)
674{
675	int ret;
676	struct dec_info *p_dec_info = &inst->codec_info->dec_info;
677	struct dec_initial_info *init_info = &p_dec_info->initial_info;
678	size_t remain, idx, j, i, cnt_8_chunk, size;
679	u32 start_no, end_no;
680	u32 reg_val, cbcr_interleave, nv21, pic_size;
681	u32 addr_y, addr_cb, addr_cr;
682	u32 mv_col_size, frame_width, frame_height, fbc_y_tbl_size, fbc_c_tbl_size;
683	struct vpu_buf vb_buf;
684	bool justified = WTL_RIGHT_JUSTIFIED;
685	u32 format_no = WTL_PIXEL_8BIT;
686	u32 color_format = 0;
687	u32 pixel_order = 1;
688	u32 bwb_flag = (map_type == LINEAR_FRAME_MAP) ? 1 : 0;
689
690	cbcr_interleave = inst->cbcr_interleave;
691	nv21 = inst->nv21;
692	mv_col_size = 0;
693	fbc_y_tbl_size = 0;
694	fbc_c_tbl_size = 0;
695
696	if (map_type >= COMPRESSED_FRAME_MAP) {
697		cbcr_interleave = 0;
698		nv21 = 0;
699
700		switch (inst->std) {
701		case W_HEVC_DEC:
702			mv_col_size = WAVE5_DEC_HEVC_BUF_SIZE(init_info->pic_width,
703							      init_info->pic_height);
704			break;
705		case W_AVC_DEC:
706			mv_col_size = WAVE5_DEC_AVC_BUF_SIZE(init_info->pic_width,
707							     init_info->pic_height);
708			break;
709		default:
710			return -EINVAL;
711		}
712
713		if (inst->std == W_HEVC_DEC || inst->std == W_AVC_DEC) {
714			size = ALIGN(ALIGN(mv_col_size, 16), BUFFER_MARGIN) + BUFFER_MARGIN;
715			ret = wave5_vdi_allocate_array(inst->dev, p_dec_info->vb_mv, count, size);
716			if (ret)
717				goto free_mv_buffers;
718		}
719
720		frame_width = init_info->pic_width;
721		frame_height = init_info->pic_height;
722		fbc_y_tbl_size = ALIGN(WAVE5_FBC_LUMA_TABLE_SIZE(frame_width, frame_height), 16);
723		fbc_c_tbl_size = ALIGN(WAVE5_FBC_CHROMA_TABLE_SIZE(frame_width, frame_height), 16);
724
725		size = ALIGN(fbc_y_tbl_size, BUFFER_MARGIN) + BUFFER_MARGIN;
726		ret = wave5_vdi_allocate_array(inst->dev, p_dec_info->vb_fbc_y_tbl, count, size);
727		if (ret)
728			goto free_fbc_y_tbl_buffers;
729
730		size = ALIGN(fbc_c_tbl_size, BUFFER_MARGIN) + BUFFER_MARGIN;
731		ret = wave5_vdi_allocate_array(inst->dev, p_dec_info->vb_fbc_c_tbl, count, size);
732		if (ret)
733			goto free_fbc_c_tbl_buffers;
734
735		pic_size = (init_info->pic_width << 16) | (init_info->pic_height);
736
737		vb_buf.size = (p_dec_info->vlc_buf_size * VLC_BUF_NUM) +
738				(p_dec_info->param_buf_size * COMMAND_QUEUE_DEPTH);
739		vb_buf.daddr = 0;
740
741		if (vb_buf.size != p_dec_info->vb_task.size) {
742			wave5_vdi_free_dma_memory(inst->dev, &p_dec_info->vb_task);
743			ret = wave5_vdi_allocate_dma_memory(inst->dev, &vb_buf);
744			if (ret)
745				goto free_fbc_c_tbl_buffers;
746
747			p_dec_info->vb_task = vb_buf;
748		}
749
750		vpu_write_reg(inst->dev, W5_CMD_SET_FB_ADDR_TASK_BUF,
751			      p_dec_info->vb_task.daddr);
752		vpu_write_reg(inst->dev, W5_CMD_SET_FB_TASK_BUF_SIZE, vb_buf.size);
753	} else {
754		pic_size = (init_info->pic_width << 16) | (init_info->pic_height);
755
756		if (inst->output_format == FORMAT_422)
757			color_format = 1;
758	}
759	vpu_write_reg(inst->dev, W5_PIC_SIZE, pic_size);
760
761	reg_val = (bwb_flag << 28) |
762		  (pixel_order << 23) |
763		  (justified << 22) |
764		  (format_no << 20) |
765		  (color_format << 19) |
766		  (nv21 << 17) |
767		  (cbcr_interleave << 16) |
768		  (fb_arr[0].stride);
769	vpu_write_reg(inst->dev, W5_COMMON_PIC_INFO, reg_val);
770
771	remain = count;
772	cnt_8_chunk = DIV_ROUND_UP(count, 8);
773	idx = 0;
774	for (j = 0; j < cnt_8_chunk; j++) {
775		reg_val = (j == cnt_8_chunk - 1) << 4 | ((j == 0) << 3);
776		vpu_write_reg(inst->dev, W5_SFB_OPTION, reg_val);
777		start_no = j * 8;
778		end_no = start_no + ((remain >= 8) ? 8 : remain) - 1;
779
780		vpu_write_reg(inst->dev, W5_SET_FB_NUM, (start_no << 8) | end_no);
781
782		for (i = 0; i < 8 && i < remain; i++) {
783			addr_y = fb_arr[i + start_no].buf_y;
784			addr_cb = fb_arr[i + start_no].buf_cb;
785			addr_cr = fb_arr[i + start_no].buf_cr;
786			vpu_write_reg(inst->dev, W5_ADDR_LUMA_BASE0 + (i << 4), addr_y);
787			vpu_write_reg(inst->dev, W5_ADDR_CB_BASE0 + (i << 4), addr_cb);
788			if (map_type >= COMPRESSED_FRAME_MAP) {
789				/* luma FBC offset table */
790				vpu_write_reg(inst->dev, W5_ADDR_FBC_Y_OFFSET0 + (i << 4),
791					      p_dec_info->vb_fbc_y_tbl[idx].daddr);
792				/* chroma FBC offset table */
793				vpu_write_reg(inst->dev, W5_ADDR_FBC_C_OFFSET0 + (i << 4),
794					      p_dec_info->vb_fbc_c_tbl[idx].daddr);
795				vpu_write_reg(inst->dev, W5_ADDR_MV_COL0 + (i << 2),
796					      p_dec_info->vb_mv[idx].daddr);
797			} else {
798				vpu_write_reg(inst->dev, W5_ADDR_CR_BASE0 + (i << 4), addr_cr);
799				vpu_write_reg(inst->dev, W5_ADDR_FBC_C_OFFSET0 + (i << 4), 0);
800				vpu_write_reg(inst->dev, W5_ADDR_MV_COL0 + (i << 2), 0);
801			}
802			idx++;
803		}
804		remain -= i;
805
806		ret = send_firmware_command(inst, W5_SET_FB, false, NULL, NULL);
807		if (ret)
808			goto free_buffers;
809	}
810
811	reg_val = vpu_read_reg(inst->dev, W5_RET_SUCCESS);
812	if (!reg_val) {
813		ret = -EIO;
814		goto free_buffers;
815	}
816
817	return 0;
818
819free_buffers:
820	wave5_vdi_free_dma_memory(inst->dev, &p_dec_info->vb_task);
821free_fbc_c_tbl_buffers:
822	for (i = 0; i < count; i++)
823		wave5_vdi_free_dma_memory(inst->dev, &p_dec_info->vb_fbc_c_tbl[i]);
824free_fbc_y_tbl_buffers:
825	for (i = 0; i < count; i++)
826		wave5_vdi_free_dma_memory(inst->dev, &p_dec_info->vb_fbc_y_tbl[i]);
827free_mv_buffers:
828	for (i = 0; i < count; i++)
829		wave5_vdi_free_dma_memory(inst->dev, &p_dec_info->vb_mv[i]);
830	return ret;
831}
832
833int wave5_vpu_decode(struct vpu_instance *inst, u32 *fail_res)
834{
835	u32 reg_val;
836	struct dec_info *p_dec_info = &inst->codec_info->dec_info;
837	int ret;
838
839	vpu_write_reg(inst->dev, W5_BS_RD_PTR, p_dec_info->stream_rd_ptr);
840	vpu_write_reg(inst->dev, W5_BS_WR_PTR, p_dec_info->stream_wr_ptr);
841
842	vpu_write_reg(inst->dev, W5_BS_OPTION, get_bitstream_options(p_dec_info));
843
844	/* secondary AXI */
845	reg_val = p_dec_info->sec_axi_info.use_bit_enable |
846		(p_dec_info->sec_axi_info.use_ip_enable << 9) |
847		(p_dec_info->sec_axi_info.use_lf_row_enable << 15);
848	vpu_write_reg(inst->dev, W5_USE_SEC_AXI, reg_val);
849
850	/* set attributes of user buffer */
851	vpu_write_reg(inst->dev, W5_CMD_DEC_USER_MASK, p_dec_info->user_data_enable);
852
853	vpu_write_reg(inst->dev, W5_COMMAND_OPTION, DEC_PIC_NORMAL);
854	vpu_write_reg(inst->dev, W5_CMD_DEC_TEMPORAL_ID_PLUS1,
855		      (p_dec_info->target_spatial_id << 9) |
856		      (p_dec_info->temp_id_select_mode << 8) | p_dec_info->target_temp_id);
857	vpu_write_reg(inst->dev, W5_CMD_SEQ_CHANGE_ENABLE_FLAG, p_dec_info->seq_change_mask);
858	/* When reordering is disabled we force the latency of the framebuffers */
859	vpu_write_reg(inst->dev, W5_CMD_DEC_FORCE_FB_LATENCY_PLUS1, !p_dec_info->reorder_enable);
860
861	ret = send_firmware_command(inst, W5_DEC_ENC_PIC, true, &reg_val, fail_res);
862	if (ret == -ETIMEDOUT)
863		return ret;
864
865	p_dec_info->instance_queue_count = (reg_val >> 16) & 0xff;
866	p_dec_info->report_queue_count = (reg_val & QUEUE_REPORT_MASK);
867
868	dev_dbg(inst->dev->dev, "%s: dec pic sent (queue %u : %u)\n", __func__,
869		p_dec_info->instance_queue_count, p_dec_info->report_queue_count);
870
871	if (ret)
872		return ret;
873
874	return 0;
875}
876
877int wave5_vpu_dec_get_result(struct vpu_instance *inst, struct dec_output_info *result)
878{
879	int ret;
880	u32 index, nal_unit_type, reg_val, sub_layer_info;
881	struct dec_info *p_dec_info = &inst->codec_info->dec_info;
882	struct vpu_device *vpu_dev = inst->dev;
883
884	vpu_write_reg(inst->dev, W5_CMD_DEC_ADDR_REPORT_BASE, p_dec_info->user_data_buf_addr);
885	vpu_write_reg(inst->dev, W5_CMD_DEC_REPORT_SIZE, p_dec_info->user_data_buf_size);
886	vpu_write_reg(inst->dev, W5_CMD_DEC_REPORT_PARAM, REPORT_PARAM_ENDIANNESS_BIG_ENDIAN);
887
888	/* send QUERY cmd */
889	ret = wave5_send_query(vpu_dev, inst, GET_RESULT);
890	if (ret)
891		return ret;
892
893	reg_val = vpu_read_reg(inst->dev, W5_RET_QUEUE_STATUS);
894
895	p_dec_info->instance_queue_count = (reg_val >> 16) & 0xff;
896	p_dec_info->report_queue_count = (reg_val & QUEUE_REPORT_MASK);
897
898	dev_dbg(inst->dev->dev, "%s: dec pic complete (queue %u : %u)\n", __func__,
899		p_dec_info->instance_queue_count, p_dec_info->report_queue_count);
900
901	reg_val = vpu_read_reg(inst->dev, W5_RET_DEC_PIC_TYPE);
902
903	nal_unit_type = (reg_val >> 4) & 0x3f;
904
905	if (inst->std == W_HEVC_DEC) {
906		if (reg_val & 0x04)
907			result->pic_type = PIC_TYPE_B;
908		else if (reg_val & 0x02)
909			result->pic_type = PIC_TYPE_P;
910		else if (reg_val & 0x01)
911			result->pic_type = PIC_TYPE_I;
912		else
913			result->pic_type = PIC_TYPE_MAX;
914		if ((nal_unit_type == 19 || nal_unit_type == 20) && result->pic_type == PIC_TYPE_I)
915			/* IDR_W_RADL, IDR_N_LP */
916			result->pic_type = PIC_TYPE_IDR;
917	} else if (inst->std == W_AVC_DEC) {
918		if (reg_val & 0x04)
919			result->pic_type = PIC_TYPE_B;
920		else if (reg_val & 0x02)
921			result->pic_type = PIC_TYPE_P;
922		else if (reg_val & 0x01)
923			result->pic_type = PIC_TYPE_I;
924		else
925			result->pic_type = PIC_TYPE_MAX;
926		if (nal_unit_type == 5 && result->pic_type == PIC_TYPE_I)
927			result->pic_type = PIC_TYPE_IDR;
928	}
929	index = vpu_read_reg(inst->dev, W5_RET_DEC_DISPLAY_INDEX);
930	result->index_frame_display = index;
931	index = vpu_read_reg(inst->dev, W5_RET_DEC_DECODED_INDEX);
932	result->index_frame_decoded = index;
933	result->index_frame_decoded_for_tiled = index;
934
935	sub_layer_info = vpu_read_reg(inst->dev, W5_RET_DEC_SUB_LAYER_INFO);
936	result->temporal_id = sub_layer_info & 0x7;
937
938	if (inst->std == W_HEVC_DEC || inst->std == W_AVC_DEC) {
939		result->decoded_poc = -1;
940		if (result->index_frame_decoded >= 0 ||
941		    result->index_frame_decoded == DECODED_IDX_FLAG_SKIP)
942			result->decoded_poc = vpu_read_reg(inst->dev, W5_RET_DEC_PIC_POC);
943	}
944
945	result->sequence_changed = vpu_read_reg(inst->dev, W5_RET_DEC_NOTIFICATION);
946	reg_val = vpu_read_reg(inst->dev, W5_RET_DEC_PIC_SIZE);
947	result->dec_pic_width = reg_val >> 16;
948	result->dec_pic_height = reg_val & 0xffff;
949
950	if (result->sequence_changed) {
951		memcpy((void *)&p_dec_info->new_seq_info, (void *)&p_dec_info->initial_info,
952		       sizeof(struct dec_initial_info));
953		wave5_get_dec_seq_result(inst, &p_dec_info->new_seq_info);
954	}
955
956	result->dec_host_cmd_tick = vpu_read_reg(inst->dev, W5_RET_DEC_HOST_CMD_TICK);
957	result->dec_decode_end_tick = vpu_read_reg(inst->dev, W5_RET_DEC_DECODING_ENC_TICK);
958
959	if (!p_dec_info->first_cycle_check) {
960		result->frame_cycle =
961			(result->dec_decode_end_tick - result->dec_host_cmd_tick) *
962			p_dec_info->cycle_per_tick;
963		vpu_dev->last_performance_cycles = result->dec_decode_end_tick;
964		p_dec_info->first_cycle_check = true;
965	} else if (result->index_frame_decoded_for_tiled != -1) {
966		result->frame_cycle =
967			(result->dec_decode_end_tick - vpu_dev->last_performance_cycles) *
968			p_dec_info->cycle_per_tick;
969		vpu_dev->last_performance_cycles = result->dec_decode_end_tick;
970		if (vpu_dev->last_performance_cycles < result->dec_host_cmd_tick)
971			result->frame_cycle =
972				(result->dec_decode_end_tick - result->dec_host_cmd_tick) *
973				p_dec_info->cycle_per_tick;
974	}
975
976	/* no remaining command. reset frame cycle. */
977	if (p_dec_info->instance_queue_count == 0 && p_dec_info->report_queue_count == 0)
978		p_dec_info->first_cycle_check = false;
979
980	return 0;
981}
982
983int wave5_vpu_re_init(struct device *dev, u8 *fw, size_t size)
984{
985	struct vpu_buf *common_vb;
986	dma_addr_t code_base, temp_base;
987	dma_addr_t old_code_base, temp_size;
988	u32 code_size, reason_code;
989	u32 reg_val;
990	struct vpu_device *vpu_dev = dev_get_drvdata(dev);
991
992	common_vb = &vpu_dev->common_mem;
993
994	code_base = common_vb->daddr;
995	/* ALIGN TO 4KB */
996	code_size = (WAVE5_MAX_CODE_BUF_SIZE & ~0xfff);
997	if (code_size < size * 2)
998		return -EINVAL;
999	temp_base = common_vb->daddr + WAVE5_TEMPBUF_OFFSET;
1000	temp_size = WAVE5_TEMPBUF_SIZE;
1001
1002	old_code_base = vpu_read_reg(vpu_dev, W5_VPU_REMAP_PADDR);
1003
1004	if (old_code_base != code_base + W5_REMAP_INDEX1 * W5_REMAP_MAX_SIZE) {
1005		int ret;
1006
1007		ret = wave5_vdi_write_memory(vpu_dev, common_vb, 0, fw, size);
1008		if (ret < 0) {
1009			dev_err(vpu_dev->dev,
1010				"VPU init, Writing firmware to common buffer, fail: %d\n", ret);
1011			return ret;
1012		}
1013
1014		vpu_write_reg(vpu_dev, W5_PO_CONF, 0);
1015
1016		ret = wave5_vpu_reset(dev, SW_RESET_ON_BOOT);
1017		if (ret < 0) {
1018			dev_err(vpu_dev->dev, "VPU init, Resetting the VPU, fail: %d\n", ret);
1019			return ret;
1020		}
1021
1022		remap_page(vpu_dev, code_base, W5_REMAP_INDEX0);
1023		remap_page(vpu_dev, code_base, W5_REMAP_INDEX1);
1024
1025		vpu_write_reg(vpu_dev, W5_ADDR_CODE_BASE, code_base);
1026		vpu_write_reg(vpu_dev, W5_CODE_SIZE, code_size);
1027		vpu_write_reg(vpu_dev, W5_CODE_PARAM, (WAVE5_UPPER_PROC_AXI_ID << 4) | 0);
1028		vpu_write_reg(vpu_dev, W5_ADDR_TEMP_BASE, temp_base);
1029		vpu_write_reg(vpu_dev, W5_TEMP_SIZE, temp_size);
1030
1031		/* These register must be reset explicitly */
1032		vpu_write_reg(vpu_dev, W5_HW_OPTION, 0);
1033		wave5_fio_writel(vpu_dev, W5_BACKBONE_PROC_EXT_ADDR, 0);
1034		wave5_fio_writel(vpu_dev, W5_BACKBONE_AXI_PARAM, 0);
1035		vpu_write_reg(vpu_dev, W5_SEC_AXI_PARAM, 0);
1036
1037		/* Encoder interrupt */
1038		reg_val = BIT(INT_WAVE5_ENC_SET_PARAM);
1039		reg_val |= BIT(INT_WAVE5_ENC_PIC);
1040		reg_val |= BIT(INT_WAVE5_BSBUF_FULL);
1041		/* Decoder interrupt */
1042		reg_val |= BIT(INT_WAVE5_INIT_SEQ);
1043		reg_val |= BIT(INT_WAVE5_DEC_PIC);
1044		reg_val |= BIT(INT_WAVE5_BSBUF_EMPTY);
1045		vpu_write_reg(vpu_dev, W5_VPU_VINT_ENABLE, reg_val);
1046
1047		reg_val = vpu_read_reg(vpu_dev, W5_VPU_RET_VPU_CONFIG0);
1048		if (FIELD_GET(FEATURE_BACKBONE, reg_val)) {
1049			reg_val = ((WAVE5_PROC_AXI_ID << 28) |
1050					(WAVE5_PRP_AXI_ID << 24) |
1051					(WAVE5_FBD_Y_AXI_ID << 20) |
1052					(WAVE5_FBC_Y_AXI_ID << 16) |
1053					(WAVE5_FBD_C_AXI_ID << 12) |
1054					(WAVE5_FBC_C_AXI_ID << 8) |
1055					(WAVE5_PRI_AXI_ID << 4) |
1056					WAVE5_SEC_AXI_ID);
1057			wave5_fio_writel(vpu_dev, W5_BACKBONE_PROG_AXI_ID, reg_val);
1058		}
1059
1060		vpu_write_reg(vpu_dev, W5_VPU_BUSY_STATUS, 1);
1061		vpu_write_reg(vpu_dev, W5_COMMAND, W5_INIT_VPU);
1062		vpu_write_reg(vpu_dev, W5_VPU_REMAP_CORE_START, 1);
1063
1064		ret = wave5_wait_vpu_busy(vpu_dev, W5_VPU_BUSY_STATUS);
1065		if (ret) {
1066			dev_err(vpu_dev->dev, "VPU reinit(W5_VPU_REMAP_CORE_START) timeout\n");
1067			return ret;
1068		}
1069
1070		ret = wave5_vpu_firmware_command_queue_error_check(vpu_dev, &reason_code);
1071		if (ret)
1072			return ret;
1073	}
1074
1075	return setup_wave5_properties(dev);
1076}
1077
1078static int wave5_vpu_sleep_wake(struct device *dev, bool i_sleep_wake, const uint16_t *code,
1079				size_t size)
1080{
1081	u32 reg_val;
1082	struct vpu_buf *common_vb;
1083	dma_addr_t code_base;
1084	u32 code_size, reason_code;
1085	struct vpu_device *vpu_dev = dev_get_drvdata(dev);
1086	int ret;
1087
1088	if (i_sleep_wake) {
1089		ret = wave5_wait_vpu_busy(vpu_dev, W5_VPU_BUSY_STATUS);
1090		if (ret)
1091			return ret;
1092
1093		/*
1094		 * Declare who has ownership for the host interface access
1095		 * 1 = VPU
1096		 * 0 = Host processor
1097		 */
1098		vpu_write_reg(vpu_dev, W5_VPU_BUSY_STATUS, 1);
1099		vpu_write_reg(vpu_dev, W5_COMMAND, W5_SLEEP_VPU);
1100		/* Send an interrupt named HOST to the VPU */
1101		vpu_write_reg(vpu_dev, W5_VPU_HOST_INT_REQ, 1);
1102
1103		ret = wave5_wait_vpu_busy(vpu_dev, W5_VPU_BUSY_STATUS);
1104		if (ret)
1105			return ret;
1106
1107		ret = wave5_vpu_firmware_command_queue_error_check(vpu_dev, &reason_code);
1108		if (ret)
1109			return ret;
1110	} else { /* restore */
1111		common_vb = &vpu_dev->common_mem;
1112
1113		code_base = common_vb->daddr;
1114		/* ALIGN TO 4KB */
1115		code_size = (WAVE5_MAX_CODE_BUF_SIZE & ~0xfff);
1116		if (code_size < size * 2) {
1117			dev_err(dev, "size too small\n");
1118			return -EINVAL;
1119		}
1120
1121		/* Power on without DEBUG mode */
1122		vpu_write_reg(vpu_dev, W5_PO_CONF, 0);
1123
1124		remap_page(vpu_dev, code_base, W5_REMAP_INDEX0);
1125		remap_page(vpu_dev, code_base, W5_REMAP_INDEX1);
1126
1127		vpu_write_reg(vpu_dev, W5_ADDR_CODE_BASE, code_base);
1128		vpu_write_reg(vpu_dev, W5_CODE_SIZE, code_size);
1129		vpu_write_reg(vpu_dev, W5_CODE_PARAM, (WAVE5_UPPER_PROC_AXI_ID << 4) | 0);
1130
1131		/* These register must be reset explicitly */
1132		vpu_write_reg(vpu_dev, W5_HW_OPTION, 0);
1133		wave5_fio_writel(vpu_dev, W5_BACKBONE_PROC_EXT_ADDR, 0);
1134		wave5_fio_writel(vpu_dev, W5_BACKBONE_AXI_PARAM, 0);
1135		vpu_write_reg(vpu_dev, W5_SEC_AXI_PARAM, 0);
1136
1137		/* Encoder interrupt */
1138		reg_val = BIT(INT_WAVE5_ENC_SET_PARAM);
1139		reg_val |= BIT(INT_WAVE5_ENC_PIC);
1140		reg_val |= BIT(INT_WAVE5_BSBUF_FULL);
1141		/* Decoder interrupt */
1142		reg_val |= BIT(INT_WAVE5_INIT_SEQ);
1143		reg_val |= BIT(INT_WAVE5_DEC_PIC);
1144		reg_val |= BIT(INT_WAVE5_BSBUF_EMPTY);
1145		vpu_write_reg(vpu_dev, W5_VPU_VINT_ENABLE, reg_val);
1146
1147		reg_val = vpu_read_reg(vpu_dev, W5_VPU_RET_VPU_CONFIG0);
1148		if (FIELD_GET(FEATURE_BACKBONE, reg_val)) {
1149			reg_val = ((WAVE5_PROC_AXI_ID << 28) |
1150					(WAVE5_PRP_AXI_ID << 24) |
1151					(WAVE5_FBD_Y_AXI_ID << 20) |
1152					(WAVE5_FBC_Y_AXI_ID << 16) |
1153					(WAVE5_FBD_C_AXI_ID << 12) |
1154					(WAVE5_FBC_C_AXI_ID << 8) |
1155					(WAVE5_PRI_AXI_ID << 4) |
1156					WAVE5_SEC_AXI_ID);
1157			wave5_fio_writel(vpu_dev, W5_BACKBONE_PROG_AXI_ID, reg_val);
1158		}
1159
1160		vpu_write_reg(vpu_dev, W5_VPU_BUSY_STATUS, 1);
1161		vpu_write_reg(vpu_dev, W5_COMMAND, W5_WAKEUP_VPU);
1162		/* Start VPU after settings */
1163		vpu_write_reg(vpu_dev, W5_VPU_REMAP_CORE_START, 1);
1164
1165		ret = wave5_wait_vpu_busy(vpu_dev, W5_VPU_BUSY_STATUS);
1166		if (ret) {
1167			dev_err(vpu_dev->dev, "VPU wakeup(W5_VPU_REMAP_CORE_START) timeout\n");
1168			return ret;
1169		}
1170
1171		return wave5_vpu_firmware_command_queue_error_check(vpu_dev, &reason_code);
1172	}
1173
1174	return 0;
1175}
1176
1177int wave5_vpu_reset(struct device *dev, enum sw_reset_mode reset_mode)
1178{
1179	u32 val = 0;
1180	int ret = 0;
1181	struct vpu_device *vpu_dev = dev_get_drvdata(dev);
1182	struct vpu_attr *p_attr = &vpu_dev->attr;
1183	/* VPU doesn't send response. force to set BUSY flag to 0. */
1184	vpu_write_reg(vpu_dev, W5_VPU_BUSY_STATUS, 0);
1185
1186	if (reset_mode == SW_RESET_SAFETY) {
1187		ret = wave5_vpu_sleep_wake(dev, true, NULL, 0);
1188		if (ret)
1189			return ret;
1190	}
1191
1192	val = vpu_read_reg(vpu_dev, W5_VPU_RET_VPU_CONFIG0);
1193	if ((val >> 16) & 0x1)
1194		p_attr->support_backbone = true;
1195	if ((val >> 22) & 0x1)
1196		p_attr->support_vcore_backbone = true;
1197	if ((val >> 28) & 0x1)
1198		p_attr->support_vcpu_backbone = true;
1199
1200	/* waiting for completion of bus transaction */
1201	if (p_attr->support_backbone) {
1202		dev_dbg(dev, "%s: backbone supported\n", __func__);
1203
1204		if (p_attr->support_vcore_backbone) {
1205			if (p_attr->support_vcpu_backbone) {
1206				/* step1 : disable request */
1207				wave5_fio_writel(vpu_dev, W5_BACKBONE_BUS_CTRL_VCPU, 0xFF);
1208
1209				/* step2 : waiting for completion of bus transaction */
1210				ret = wave5_wait_vcpu_bus_busy(vpu_dev,
1211							       W5_BACKBONE_BUS_STATUS_VCPU);
1212				if (ret) {
1213					wave5_fio_writel(vpu_dev, W5_BACKBONE_BUS_CTRL_VCPU, 0x00);
1214					return ret;
1215				}
1216			}
1217			/* step1 : disable request */
1218			wave5_fio_writel(vpu_dev, W5_BACKBONE_BUS_CTRL_VCORE0, 0x7);
1219
1220			/* step2 : waiting for completion of bus transaction */
1221			if (wave5_wait_bus_busy(vpu_dev, W5_BACKBONE_BUS_STATUS_VCORE0)) {
1222				wave5_fio_writel(vpu_dev, W5_BACKBONE_BUS_CTRL_VCORE0, 0x00);
1223				return -EBUSY;
1224			}
1225		} else {
1226			/* step1 : disable request */
1227			wave5_fio_writel(vpu_dev, W5_COMBINED_BACKBONE_BUS_CTRL, 0x7);
1228
1229			/* step2 : waiting for completion of bus transaction */
1230			if (wave5_wait_bus_busy(vpu_dev, W5_COMBINED_BACKBONE_BUS_STATUS)) {
1231				wave5_fio_writel(vpu_dev, W5_COMBINED_BACKBONE_BUS_CTRL, 0x00);
1232				return -EBUSY;
1233			}
1234		}
1235	} else {
1236		dev_dbg(dev, "%s: backbone NOT supported\n", __func__);
1237		/* step1 : disable request */
1238		wave5_fio_writel(vpu_dev, W5_GDI_BUS_CTRL, 0x100);
1239
1240		/* step2 : waiting for completion of bus transaction */
1241		ret = wave5_wait_bus_busy(vpu_dev, W5_GDI_BUS_STATUS);
1242		if (ret) {
1243			wave5_fio_writel(vpu_dev, W5_GDI_BUS_CTRL, 0x00);
1244			return ret;
1245		}
1246	}
1247
1248	switch (reset_mode) {
1249	case SW_RESET_ON_BOOT:
1250	case SW_RESET_FORCE:
1251	case SW_RESET_SAFETY:
1252		val = W5_RST_BLOCK_ALL;
1253		break;
1254	default:
1255		return -EINVAL;
1256	}
1257
1258	if (val) {
1259		vpu_write_reg(vpu_dev, W5_VPU_RESET_REQ, val);
1260
1261		ret = wave5_wait_vpu_busy(vpu_dev, W5_VPU_RESET_STATUS);
1262		if (ret) {
1263			vpu_write_reg(vpu_dev, W5_VPU_RESET_REQ, 0);
1264			return ret;
1265		}
1266		vpu_write_reg(vpu_dev, W5_VPU_RESET_REQ, 0);
1267	}
1268	/* step3 : must clear GDI_BUS_CTRL after done SW_RESET */
1269	if (p_attr->support_backbone) {
1270		if (p_attr->support_vcore_backbone) {
1271			if (p_attr->support_vcpu_backbone)
1272				wave5_fio_writel(vpu_dev, W5_BACKBONE_BUS_CTRL_VCPU, 0x00);
1273			wave5_fio_writel(vpu_dev, W5_BACKBONE_BUS_CTRL_VCORE0, 0x00);
1274		} else {
1275			wave5_fio_writel(vpu_dev, W5_COMBINED_BACKBONE_BUS_CTRL, 0x00);
1276		}
1277	} else {
1278		wave5_fio_writel(vpu_dev, W5_GDI_BUS_CTRL, 0x00);
1279	}
1280	if (reset_mode == SW_RESET_SAFETY || reset_mode == SW_RESET_FORCE)
1281		ret = wave5_vpu_sleep_wake(dev, false, NULL, 0);
1282
1283	return ret;
1284}
1285
1286int wave5_vpu_dec_finish_seq(struct vpu_instance *inst, u32 *fail_res)
1287{
1288	return send_firmware_command(inst, W5_DESTROY_INSTANCE, true, NULL, fail_res);
1289}
1290
1291int wave5_vpu_dec_set_bitstream_flag(struct vpu_instance *inst, bool eos)
1292{
1293	struct dec_info *p_dec_info = &inst->codec_info->dec_info;
1294
1295	p_dec_info->stream_endflag = eos ? 1 : 0;
1296	vpu_write_reg(inst->dev, W5_BS_OPTION, get_bitstream_options(p_dec_info));
1297	vpu_write_reg(inst->dev, W5_BS_WR_PTR, p_dec_info->stream_wr_ptr);
1298
1299	return send_firmware_command(inst, W5_UPDATE_BS, true, NULL, NULL);
1300}
1301
1302int wave5_dec_clr_disp_flag(struct vpu_instance *inst, unsigned int index)
1303{
1304	struct dec_info *p_dec_info = &inst->codec_info->dec_info;
1305	int ret;
1306
1307	vpu_write_reg(inst->dev, W5_CMD_DEC_CLR_DISP_IDC, BIT(index));
1308	vpu_write_reg(inst->dev, W5_CMD_DEC_SET_DISP_IDC, 0);
1309
1310	ret = wave5_send_query(inst->dev, inst, UPDATE_DISP_FLAG);
1311	if (ret)
1312		return ret;
1313
1314	p_dec_info->frame_display_flag = vpu_read_reg(inst->dev, W5_RET_DEC_DISP_IDC);
1315
1316	return 0;
1317}
1318
1319int wave5_dec_set_disp_flag(struct vpu_instance *inst, unsigned int index)
1320{
1321	int ret;
1322
1323	vpu_write_reg(inst->dev, W5_CMD_DEC_CLR_DISP_IDC, 0);
1324	vpu_write_reg(inst->dev, W5_CMD_DEC_SET_DISP_IDC, BIT(index));
1325
1326	ret = wave5_send_query(inst->dev, inst, UPDATE_DISP_FLAG);
1327	if (ret)
1328		return ret;
1329
1330	return 0;
1331}
1332
1333int wave5_vpu_clear_interrupt(struct vpu_instance *inst, u32 flags)
1334{
1335	u32 interrupt_reason;
1336
1337	interrupt_reason = vpu_read_reg(inst->dev, W5_VPU_VINT_REASON_USR);
1338	interrupt_reason &= ~flags;
1339	vpu_write_reg(inst->dev, W5_VPU_VINT_REASON_USR, interrupt_reason);
1340
1341	return 0;
1342}
1343
1344dma_addr_t wave5_dec_get_rd_ptr(struct vpu_instance *inst)
1345{
1346	int ret;
1347
1348	ret = wave5_send_query(inst->dev, inst, GET_BS_RD_PTR);
1349	if (ret)
1350		return inst->codec_info->dec_info.stream_rd_ptr;
1351
1352	return vpu_read_reg(inst->dev, W5_RET_QUERY_DEC_BS_RD_PTR);
1353}
1354
1355int wave5_dec_set_rd_ptr(struct vpu_instance *inst, dma_addr_t addr)
1356{
1357	int ret;
1358
1359	vpu_write_reg(inst->dev, W5_RET_QUERY_DEC_SET_BS_RD_PTR, addr);
1360
1361	ret = wave5_send_query(inst->dev, inst, SET_BS_RD_PTR);
1362
1363	return ret;
1364}
1365
1366/************************************************************************/
1367/* ENCODER functions */
1368/************************************************************************/
1369
1370int wave5_vpu_build_up_enc_param(struct device *dev, struct vpu_instance *inst,
1371				 struct enc_open_param *open_param)
1372{
1373	int ret;
1374	struct enc_info *p_enc_info = &inst->codec_info->enc_info;
1375	u32 reg_val;
1376	struct vpu_device *vpu_dev = dev_get_drvdata(dev);
1377	dma_addr_t buffer_addr;
1378	size_t buffer_size;
1379
1380	p_enc_info->cycle_per_tick = 256;
1381	if (vpu_dev->sram_buf.size) {
1382		p_enc_info->sec_axi_info.use_enc_rdo_enable = 1;
1383		p_enc_info->sec_axi_info.use_enc_lf_enable = 1;
1384	}
1385
1386	p_enc_info->vb_work.size = WAVE521ENC_WORKBUF_SIZE;
1387	ret = wave5_vdi_allocate_dma_memory(vpu_dev, &p_enc_info->vb_work);
1388	if (ret) {
1389		memset(&p_enc_info->vb_work, 0, sizeof(p_enc_info->vb_work));
1390		return ret;
1391	}
1392
1393	wave5_vdi_clear_memory(vpu_dev, &p_enc_info->vb_work);
1394
1395	vpu_write_reg(inst->dev, W5_ADDR_WORK_BASE, p_enc_info->vb_work.daddr);
1396	vpu_write_reg(inst->dev, W5_WORK_SIZE, p_enc_info->vb_work.size);
1397
1398	vpu_write_reg(inst->dev, W5_CMD_ADDR_SEC_AXI, vpu_dev->sram_buf.daddr);
1399	vpu_write_reg(inst->dev, W5_CMD_SEC_AXI_SIZE, vpu_dev->sram_buf.size);
1400
1401	reg_val = (open_param->line_buf_int_en << 6) | BITSTREAM_ENDIANNESS_BIG_ENDIAN;
1402	vpu_write_reg(inst->dev, W5_CMD_BS_PARAM, reg_val);
1403	vpu_write_reg(inst->dev, W5_CMD_EXT_ADDR, 0);
1404	vpu_write_reg(inst->dev, W5_CMD_NUM_CQ_DEPTH_M1, (COMMAND_QUEUE_DEPTH - 1));
1405
1406	/* This register must be reset explicitly */
1407	vpu_write_reg(inst->dev, W5_CMD_ENC_SRC_OPTIONS, 0);
1408	vpu_write_reg(inst->dev, W5_CMD_ENC_VCORE_INFO, 1);
1409
1410	ret = send_firmware_command(inst, W5_CREATE_INSTANCE, true, NULL, NULL);
1411	if (ret)
1412		goto free_vb_work;
1413
1414	buffer_addr = open_param->bitstream_buffer;
1415	buffer_size = open_param->bitstream_buffer_size;
1416	p_enc_info->stream_rd_ptr = buffer_addr;
1417	p_enc_info->stream_wr_ptr = buffer_addr;
1418	p_enc_info->line_buf_int_en = open_param->line_buf_int_en;
1419	p_enc_info->stream_buf_start_addr = buffer_addr;
1420	p_enc_info->stream_buf_size = buffer_size;
1421	p_enc_info->stream_buf_end_addr = buffer_addr + buffer_size;
1422	p_enc_info->stride = 0;
1423	p_enc_info->initial_info_obtained = false;
1424	p_enc_info->product_code = vpu_read_reg(inst->dev, W5_PRODUCT_NUMBER);
1425
1426	return 0;
1427free_vb_work:
1428	if (wave5_vdi_free_dma_memory(vpu_dev, &p_enc_info->vb_work))
1429		memset(&p_enc_info->vb_work, 0, sizeof(p_enc_info->vb_work));
1430	return ret;
1431}
1432
1433static void wave5_set_enc_crop_info(u32 codec, struct enc_wave_param *param, int rot_mode,
1434				    int src_width, int src_height)
1435{
1436	int aligned_width = (codec == W_HEVC_ENC) ? ALIGN(src_width, 32) : ALIGN(src_width, 16);
1437	int aligned_height = (codec == W_HEVC_ENC) ? ALIGN(src_height, 32) : ALIGN(src_height, 16);
1438	int pad_right, pad_bot;
1439	int crop_right, crop_left, crop_top, crop_bot;
1440	int prp_mode = rot_mode >> 1; /* remove prp_enable bit */
1441
1442	if (codec == W_HEVC_ENC &&
1443	    (!rot_mode || prp_mode == 14)) /* prp_mode 14 : hor_mir && ver_mir && rot_180 */
1444		return;
1445
1446	pad_right = aligned_width - src_width;
1447	pad_bot = aligned_height - src_height;
1448
1449	if (param->conf_win_right > 0)
1450		crop_right = param->conf_win_right + pad_right;
1451	else
1452		crop_right = pad_right;
1453
1454	if (param->conf_win_bot > 0)
1455		crop_bot = param->conf_win_bot + pad_bot;
1456	else
1457		crop_bot = pad_bot;
1458
1459	crop_top = param->conf_win_top;
1460	crop_left = param->conf_win_left;
1461
1462	param->conf_win_top = crop_top;
1463	param->conf_win_left = crop_left;
1464	param->conf_win_bot = crop_bot;
1465	param->conf_win_right = crop_right;
1466
1467	switch (prp_mode) {
1468	case 0:
1469		return;
1470	case 1:
1471	case 15:
1472		param->conf_win_top = crop_right;
1473		param->conf_win_left = crop_top;
1474		param->conf_win_bot = crop_left;
1475		param->conf_win_right = crop_bot;
1476		break;
1477	case 2:
1478	case 12:
1479		param->conf_win_top = crop_bot;
1480		param->conf_win_left = crop_right;
1481		param->conf_win_bot = crop_top;
1482		param->conf_win_right = crop_left;
1483		break;
1484	case 3:
1485	case 13:
1486		param->conf_win_top = crop_left;
1487		param->conf_win_left = crop_bot;
1488		param->conf_win_bot = crop_right;
1489		param->conf_win_right = crop_top;
1490		break;
1491	case 4:
1492	case 10:
1493		param->conf_win_top = crop_bot;
1494		param->conf_win_bot = crop_top;
1495		break;
1496	case 8:
1497	case 6:
1498		param->conf_win_left = crop_right;
1499		param->conf_win_right = crop_left;
1500		break;
1501	case 5:
1502	case 11:
1503		param->conf_win_top = crop_left;
1504		param->conf_win_left = crop_top;
1505		param->conf_win_bot = crop_right;
1506		param->conf_win_right = crop_bot;
1507		break;
1508	case 7:
1509	case 9:
1510		param->conf_win_top = crop_right;
1511		param->conf_win_left = crop_bot;
1512		param->conf_win_bot = crop_left;
1513		param->conf_win_right = crop_top;
1514		break;
1515	default:
1516		WARN(1, "Invalid prp_mode: %d, must be in range of 1 - 15\n", prp_mode);
1517	}
1518}
1519
1520int wave5_vpu_enc_init_seq(struct vpu_instance *inst)
1521{
1522	u32 reg_val = 0, rot_mir_mode, fixed_cu_size_mode = 0x7;
1523	struct enc_info *p_enc_info = &inst->codec_info->enc_info;
1524	struct enc_open_param *p_open_param = &p_enc_info->open_param;
1525	struct enc_wave_param *p_param = &p_open_param->wave_param;
1526
1527	/*
1528	 * OPT_COMMON:
1529	 *	the last SET_PARAM command should be called with OPT_COMMON
1530	 */
1531	rot_mir_mode = 0;
1532	if (p_enc_info->rotation_enable) {
1533		switch (p_enc_info->rotation_angle) {
1534		case 0:
1535			rot_mir_mode |= NONE_ROTATE;
1536			break;
1537		case 90:
1538			rot_mir_mode |= ROT_CLOCKWISE_90;
1539			break;
1540		case 180:
1541			rot_mir_mode |= ROT_CLOCKWISE_180;
1542			break;
1543		case 270:
1544			rot_mir_mode |= ROT_CLOCKWISE_270;
1545			break;
1546		}
1547	}
1548
1549	if (p_enc_info->mirror_enable) {
1550		switch (p_enc_info->mirror_direction) {
1551		case MIRDIR_NONE:
1552			rot_mir_mode |= NONE_ROTATE;
1553			break;
1554		case MIRDIR_VER:
1555			rot_mir_mode |= MIR_VER_FLIP;
1556			break;
1557		case MIRDIR_HOR:
1558			rot_mir_mode |= MIR_HOR_FLIP;
1559			break;
1560		case MIRDIR_HOR_VER:
1561			rot_mir_mode |= MIR_HOR_VER_FLIP;
1562			break;
1563		}
1564	}
1565
1566	wave5_set_enc_crop_info(inst->std, p_param, rot_mir_mode, p_open_param->pic_width,
1567				p_open_param->pic_height);
1568
1569	/* SET_PARAM + COMMON */
1570	vpu_write_reg(inst->dev, W5_CMD_ENC_SEQ_SET_PARAM_OPTION, OPT_COMMON);
1571	vpu_write_reg(inst->dev, W5_CMD_ENC_SEQ_SRC_SIZE, p_open_param->pic_height << 16
1572			| p_open_param->pic_width);
1573	vpu_write_reg(inst->dev, W5_CMD_ENC_SEQ_CUSTOM_MAP_ENDIAN, VDI_LITTLE_ENDIAN);
1574
1575	reg_val = p_param->profile |
1576		(p_param->level << 3) |
1577		(p_param->internal_bit_depth << 14);
1578	if (inst->std == W_HEVC_ENC)
1579		reg_val |= (p_param->tier << 12) |
1580			(p_param->tmvp_enable << 23) |
1581			(p_param->sao_enable << 24) |
1582			(p_param->skip_intra_trans << 25) |
1583			(p_param->strong_intra_smooth_enable << 27) |
1584			(p_param->en_still_picture << 30);
1585	vpu_write_reg(inst->dev, W5_CMD_ENC_SEQ_SPS_PARAM, reg_val);
1586
1587	reg_val = (p_param->lossless_enable) |
1588		(p_param->const_intra_pred_flag << 1) |
1589		(p_param->lf_cross_slice_boundary_enable << 2) |
1590		(p_param->wpp_enable << 4) |
1591		(p_param->disable_deblk << 5) |
1592		((p_param->beta_offset_div2 & 0xF) << 6) |
1593		((p_param->tc_offset_div2 & 0xF) << 10) |
1594		((p_param->chroma_cb_qp_offset & 0x1F) << 14) |
1595		((p_param->chroma_cr_qp_offset & 0x1F) << 19) |
1596		(p_param->transform8x8_enable << 29) |
1597		(p_param->entropy_coding_mode << 30);
1598	vpu_write_reg(inst->dev, W5_CMD_ENC_SEQ_PPS_PARAM, reg_val);
1599
1600	vpu_write_reg(inst->dev, W5_CMD_ENC_SEQ_GOP_PARAM, p_param->gop_preset_idx);
1601
1602	if (inst->std == W_AVC_ENC)
1603		vpu_write_reg(inst->dev, W5_CMD_ENC_SEQ_INTRA_PARAM, p_param->intra_qp |
1604				((p_param->intra_period & 0x7ff) << 6) |
1605				((p_param->avc_idr_period & 0x7ff) << 17));
1606	else if (inst->std == W_HEVC_ENC)
1607		vpu_write_reg(inst->dev, W5_CMD_ENC_SEQ_INTRA_PARAM,
1608			      p_param->decoding_refresh_type | (p_param->intra_qp << 3) |
1609				(p_param->intra_period << 16));
1610
1611	reg_val = (p_param->rdo_skip << 2) |
1612		(p_param->lambda_scaling_enable << 3) |
1613		(fixed_cu_size_mode << 5) |
1614		(p_param->intra_nx_n_enable << 8) |
1615		(p_param->max_num_merge << 18);
1616
1617	vpu_write_reg(inst->dev, W5_CMD_ENC_SEQ_RDO_PARAM, reg_val);
1618
1619	if (inst->std == W_AVC_ENC)
1620		vpu_write_reg(inst->dev, W5_CMD_ENC_SEQ_INTRA_REFRESH,
1621			      p_param->intra_mb_refresh_arg << 16 | p_param->intra_mb_refresh_mode);
1622	else if (inst->std == W_HEVC_ENC)
1623		vpu_write_reg(inst->dev, W5_CMD_ENC_SEQ_INTRA_REFRESH,
1624			      p_param->intra_refresh_arg << 16 | p_param->intra_refresh_mode);
1625
1626	vpu_write_reg(inst->dev, W5_CMD_ENC_SEQ_RC_FRAME_RATE, p_open_param->frame_rate_info);
1627	vpu_write_reg(inst->dev, W5_CMD_ENC_SEQ_RC_TARGET_RATE, p_open_param->bit_rate);
1628
1629	reg_val = p_open_param->rc_enable |
1630		(p_param->hvs_qp_enable << 2) |
1631		(p_param->hvs_qp_scale << 4) |
1632		((p_param->initial_rc_qp & 0x3F) << 14) |
1633		(p_open_param->vbv_buffer_size << 20);
1634	if (inst->std == W_AVC_ENC)
1635		reg_val |= (p_param->mb_level_rc_enable << 1);
1636	else if (inst->std == W_HEVC_ENC)
1637		reg_val |= (p_param->cu_level_rc_enable << 1);
1638	vpu_write_reg(inst->dev, W5_CMD_ENC_SEQ_RC_PARAM, reg_val);
1639
1640	vpu_write_reg(inst->dev, W5_CMD_ENC_SEQ_RC_WEIGHT_PARAM,
1641		      p_param->rc_weight_buf << 8 | p_param->rc_weight_param);
1642
1643	vpu_write_reg(inst->dev, W5_CMD_ENC_SEQ_RC_MIN_MAX_QP, p_param->min_qp_i |
1644		      (p_param->max_qp_i << 6) | (p_param->hvs_max_delta_qp << 12));
1645
1646	vpu_write_reg(inst->dev, W5_CMD_ENC_SEQ_RC_INTER_MIN_MAX_QP, p_param->min_qp_p |
1647		      (p_param->max_qp_p << 6) | (p_param->min_qp_b << 12) |
1648		      (p_param->max_qp_b << 18));
1649
1650	vpu_write_reg(inst->dev, W5_CMD_ENC_SEQ_RC_BIT_RATIO_LAYER_0_3, 0);
1651	vpu_write_reg(inst->dev, W5_CMD_ENC_SEQ_RC_BIT_RATIO_LAYER_4_7, 0);
1652	vpu_write_reg(inst->dev, W5_CMD_ENC_SEQ_ROT_PARAM, rot_mir_mode);
1653
1654	vpu_write_reg(inst->dev, W5_CMD_ENC_SEQ_BG_PARAM, 0);
1655	vpu_write_reg(inst->dev, W5_CMD_ENC_SEQ_CUSTOM_LAMBDA_ADDR, 0);
1656	vpu_write_reg(inst->dev, W5_CMD_ENC_SEQ_CONF_WIN_TOP_BOT,
1657		      p_param->conf_win_bot << 16 | p_param->conf_win_top);
1658	vpu_write_reg(inst->dev, W5_CMD_ENC_SEQ_CONF_WIN_LEFT_RIGHT,
1659		      p_param->conf_win_right << 16 | p_param->conf_win_left);
1660
1661	if (inst->std == W_AVC_ENC)
1662		vpu_write_reg(inst->dev, W5_CMD_ENC_SEQ_INDEPENDENT_SLICE,
1663			      p_param->avc_slice_arg << 16 | p_param->avc_slice_mode);
1664	else if (inst->std == W_HEVC_ENC)
1665		vpu_write_reg(inst->dev, W5_CMD_ENC_SEQ_INDEPENDENT_SLICE,
1666			      p_param->independ_slice_mode_arg << 16 |
1667			      p_param->independ_slice_mode);
1668
1669	vpu_write_reg(inst->dev, W5_CMD_ENC_SEQ_USER_SCALING_LIST_ADDR, 0);
1670	vpu_write_reg(inst->dev, W5_CMD_ENC_SEQ_NUM_UNITS_IN_TICK, 0);
1671	vpu_write_reg(inst->dev, W5_CMD_ENC_SEQ_TIME_SCALE, 0);
1672	vpu_write_reg(inst->dev, W5_CMD_ENC_SEQ_NUM_TICKS_POC_DIFF_ONE, 0);
1673
1674	if (inst->std == W_HEVC_ENC) {
1675		vpu_write_reg(inst->dev, W5_CMD_ENC_SEQ_CUSTOM_MD_PU04, 0);
1676		vpu_write_reg(inst->dev, W5_CMD_ENC_SEQ_CUSTOM_MD_PU08, 0);
1677		vpu_write_reg(inst->dev, W5_CMD_ENC_SEQ_CUSTOM_MD_PU16, 0);
1678		vpu_write_reg(inst->dev, W5_CMD_ENC_SEQ_CUSTOM_MD_PU32, 0);
1679		vpu_write_reg(inst->dev, W5_CMD_ENC_SEQ_CUSTOM_MD_CU08, 0);
1680		vpu_write_reg(inst->dev, W5_CMD_ENC_SEQ_CUSTOM_MD_CU16, 0);
1681		vpu_write_reg(inst->dev, W5_CMD_ENC_SEQ_CUSTOM_MD_CU32, 0);
1682		vpu_write_reg(inst->dev, W5_CMD_ENC_SEQ_DEPENDENT_SLICE,
1683			      p_param->depend_slice_mode_arg << 16 | p_param->depend_slice_mode);
1684
1685		vpu_write_reg(inst->dev, W5_CMD_ENC_SEQ_NR_PARAM, 0);
1686
1687		vpu_write_reg(inst->dev, W5_CMD_ENC_SEQ_NR_WEIGHT,
1688			      p_param->nr_intra_weight_y |
1689			      (p_param->nr_intra_weight_cb << 5) |
1690			      (p_param->nr_intra_weight_cr << 10) |
1691			      (p_param->nr_inter_weight_y << 15) |
1692			      (p_param->nr_inter_weight_cb << 20) |
1693			      (p_param->nr_inter_weight_cr << 25));
1694	}
1695	vpu_write_reg(inst->dev, W5_CMD_ENC_SEQ_VUI_HRD_PARAM, 0);
1696
1697	return send_firmware_command(inst, W5_ENC_SET_PARAM, true, NULL, NULL);
1698}
1699
1700int wave5_vpu_enc_get_seq_info(struct vpu_instance *inst, struct enc_initial_info *info)
1701{
1702	int ret;
1703	u32 reg_val;
1704	struct enc_info *p_enc_info = &inst->codec_info->enc_info;
1705
1706	/* send QUERY cmd */
1707	ret = wave5_send_query(inst->dev, inst, GET_RESULT);
1708	if (ret)
1709		return ret;
1710
1711	dev_dbg(inst->dev->dev, "%s: init seq\n", __func__);
1712
1713	reg_val = vpu_read_reg(inst->dev, W5_RET_QUEUE_STATUS);
1714
1715	p_enc_info->instance_queue_count = (reg_val >> 16) & 0xff;
1716	p_enc_info->report_queue_count = (reg_val & QUEUE_REPORT_MASK);
1717
1718	if (vpu_read_reg(inst->dev, W5_RET_ENC_ENCODING_SUCCESS) != 1) {
1719		info->seq_init_err_reason = vpu_read_reg(inst->dev, W5_RET_ENC_ERR_INFO);
1720		ret = -EIO;
1721	} else {
1722		info->warn_info = vpu_read_reg(inst->dev, W5_RET_ENC_WARN_INFO);
1723	}
1724
1725	info->min_frame_buffer_count = vpu_read_reg(inst->dev, W5_RET_ENC_NUM_REQUIRED_FB);
1726	info->min_src_frame_count = vpu_read_reg(inst->dev, W5_RET_ENC_MIN_SRC_BUF_NUM);
1727	info->vlc_buf_size = vpu_read_reg(inst->dev, W5_RET_VLC_BUF_SIZE);
1728	info->param_buf_size = vpu_read_reg(inst->dev, W5_RET_PARAM_BUF_SIZE);
1729	p_enc_info->vlc_buf_size = info->vlc_buf_size;
1730	p_enc_info->param_buf_size = info->param_buf_size;
1731
1732	return ret;
1733}
1734
1735static u32 calculate_luma_stride(u32 width, u32 bit_depth)
1736{
1737	return ALIGN(ALIGN(width, 16) * ((bit_depth > 8) ? 5 : 4), 32);
1738}
1739
1740static u32 calculate_chroma_stride(u32 width, u32 bit_depth)
1741{
1742	return ALIGN(ALIGN(width / 2, 16) * ((bit_depth > 8) ? 5 : 4), 32);
1743}
1744
1745int wave5_vpu_enc_register_framebuffer(struct device *dev, struct vpu_instance *inst,
1746				       struct frame_buffer *fb_arr, enum tiled_map_type map_type,
1747				       unsigned int count)
1748{
1749	struct vpu_device *vpu_dev = dev_get_drvdata(dev);
1750	int ret = 0;
1751	u32 stride;
1752	u32 start_no, end_no;
1753	size_t remain, idx, j, i, cnt_8_chunk;
1754	u32 reg_val = 0, pic_size = 0, mv_col_size, fbc_y_tbl_size, fbc_c_tbl_size;
1755	u32 sub_sampled_size = 0;
1756	u32 luma_stride, chroma_stride;
1757	u32 buf_height = 0, buf_width = 0;
1758	u32 bit_depth;
1759	bool avc_encoding = (inst->std == W_AVC_ENC);
1760	struct vpu_buf vb_mv = {0};
1761	struct vpu_buf vb_fbc_y_tbl = {0};
1762	struct vpu_buf vb_fbc_c_tbl = {0};
1763	struct vpu_buf vb_sub_sam_buf = {0};
1764	struct vpu_buf vb_task = {0};
1765	struct enc_open_param *p_open_param;
1766	struct enc_info *p_enc_info = &inst->codec_info->enc_info;
1767
1768	p_open_param = &p_enc_info->open_param;
1769	mv_col_size = 0;
1770	fbc_y_tbl_size = 0;
1771	fbc_c_tbl_size = 0;
1772	stride = p_enc_info->stride;
1773	bit_depth = p_open_param->wave_param.internal_bit_depth;
1774
1775	if (avc_encoding) {
1776		buf_width = ALIGN(p_open_param->pic_width, 16);
1777		buf_height = ALIGN(p_open_param->pic_height, 16);
1778
1779		if ((p_enc_info->rotation_angle || p_enc_info->mirror_direction) &&
1780		    !(p_enc_info->rotation_angle == 180 &&
1781					p_enc_info->mirror_direction == MIRDIR_HOR_VER)) {
1782			buf_width = ALIGN(p_open_param->pic_width, 16);
1783			buf_height = ALIGN(p_open_param->pic_height, 16);
1784		}
1785
1786		if (p_enc_info->rotation_angle == 90 || p_enc_info->rotation_angle == 270) {
1787			buf_width = ALIGN(p_open_param->pic_height, 16);
1788			buf_height = ALIGN(p_open_param->pic_width, 16);
1789		}
1790	} else {
1791		buf_width = ALIGN(p_open_param->pic_width, 8);
1792		buf_height = ALIGN(p_open_param->pic_height, 8);
1793
1794		if ((p_enc_info->rotation_angle || p_enc_info->mirror_direction) &&
1795		    !(p_enc_info->rotation_angle == 180 &&
1796					p_enc_info->mirror_direction == MIRDIR_HOR_VER)) {
1797			buf_width = ALIGN(p_open_param->pic_width, 32);
1798			buf_height = ALIGN(p_open_param->pic_height, 32);
1799		}
1800
1801		if (p_enc_info->rotation_angle == 90 || p_enc_info->rotation_angle == 270) {
1802			buf_width = ALIGN(p_open_param->pic_height, 32);
1803			buf_height = ALIGN(p_open_param->pic_width, 32);
1804		}
1805	}
1806
1807	pic_size = (buf_width << 16) | buf_height;
1808
1809	if (avc_encoding) {
1810		mv_col_size = WAVE5_ENC_AVC_BUF_SIZE(buf_width, buf_height);
1811		vb_mv.daddr = 0;
1812		vb_mv.size = ALIGN(mv_col_size * count, BUFFER_MARGIN) + BUFFER_MARGIN;
1813	} else {
1814		mv_col_size = WAVE5_ENC_HEVC_BUF_SIZE(buf_width, buf_height);
1815		mv_col_size = ALIGN(mv_col_size, 16);
1816		vb_mv.daddr = 0;
1817		vb_mv.size = ALIGN(mv_col_size * count, BUFFER_MARGIN) + BUFFER_MARGIN;
1818	}
1819
1820	ret = wave5_vdi_allocate_dma_memory(vpu_dev, &vb_mv);
1821	if (ret)
1822		return ret;
1823
1824	p_enc_info->vb_mv = vb_mv;
1825
1826	fbc_y_tbl_size = ALIGN(WAVE5_FBC_LUMA_TABLE_SIZE(buf_width, buf_height), 16);
1827	fbc_c_tbl_size = ALIGN(WAVE5_FBC_CHROMA_TABLE_SIZE(buf_width, buf_height), 16);
1828
1829	vb_fbc_y_tbl.daddr = 0;
1830	vb_fbc_y_tbl.size = ALIGN(fbc_y_tbl_size * count, BUFFER_MARGIN) + BUFFER_MARGIN;
1831	ret = wave5_vdi_allocate_dma_memory(vpu_dev, &vb_fbc_y_tbl);
1832	if (ret)
1833		goto free_vb_fbc_y_tbl;
1834
1835	p_enc_info->vb_fbc_y_tbl = vb_fbc_y_tbl;
1836
1837	vb_fbc_c_tbl.daddr = 0;
1838	vb_fbc_c_tbl.size = ALIGN(fbc_c_tbl_size * count, BUFFER_MARGIN) + BUFFER_MARGIN;
1839	ret = wave5_vdi_allocate_dma_memory(vpu_dev, &vb_fbc_c_tbl);
1840	if (ret)
1841		goto free_vb_fbc_c_tbl;
1842
1843	p_enc_info->vb_fbc_c_tbl = vb_fbc_c_tbl;
1844
1845	if (avc_encoding)
1846		sub_sampled_size = WAVE5_SUBSAMPLED_ONE_SIZE_AVC(buf_width, buf_height);
1847	else
1848		sub_sampled_size = WAVE5_SUBSAMPLED_ONE_SIZE(buf_width, buf_height);
1849	vb_sub_sam_buf.size = ALIGN(sub_sampled_size * count, BUFFER_MARGIN) + BUFFER_MARGIN;
1850	vb_sub_sam_buf.daddr = 0;
1851	ret = wave5_vdi_allocate_dma_memory(vpu_dev, &vb_sub_sam_buf);
1852	if (ret)
1853		goto free_vb_sam_buf;
1854
1855	p_enc_info->vb_sub_sam_buf = vb_sub_sam_buf;
1856
1857	vb_task.size = (p_enc_info->vlc_buf_size * VLC_BUF_NUM) +
1858			(p_enc_info->param_buf_size * COMMAND_QUEUE_DEPTH);
1859	vb_task.daddr = 0;
1860	if (p_enc_info->vb_task.size == 0) {
1861		ret = wave5_vdi_allocate_dma_memory(vpu_dev, &vb_task);
1862		if (ret)
1863			goto free_vb_task;
1864
1865		p_enc_info->vb_task = vb_task;
1866
1867		vpu_write_reg(inst->dev, W5_CMD_SET_FB_ADDR_TASK_BUF,
1868			      p_enc_info->vb_task.daddr);
1869		vpu_write_reg(inst->dev, W5_CMD_SET_FB_TASK_BUF_SIZE, vb_task.size);
1870	}
1871
1872	/* set sub-sampled buffer base addr */
1873	vpu_write_reg(inst->dev, W5_ADDR_SUB_SAMPLED_FB_BASE, vb_sub_sam_buf.daddr);
1874	/* set sub-sampled buffer size for one frame */
1875	vpu_write_reg(inst->dev, W5_SUB_SAMPLED_ONE_FB_SIZE, sub_sampled_size);
1876
1877	vpu_write_reg(inst->dev, W5_PIC_SIZE, pic_size);
1878
1879	/* set stride of luma/chroma for compressed buffer */
1880	if ((p_enc_info->rotation_angle || p_enc_info->mirror_direction) &&
1881	    !(p_enc_info->rotation_angle == 180 &&
1882	    p_enc_info->mirror_direction == MIRDIR_HOR_VER)) {
1883		luma_stride = calculate_luma_stride(buf_width, bit_depth);
1884		chroma_stride = calculate_chroma_stride(buf_width / 2, bit_depth);
1885	} else {
1886		luma_stride = calculate_luma_stride(p_open_param->pic_width, bit_depth);
1887		chroma_stride = calculate_chroma_stride(p_open_param->pic_width / 2, bit_depth);
1888	}
1889
1890	vpu_write_reg(inst->dev, W5_FBC_STRIDE, luma_stride << 16 | chroma_stride);
1891	vpu_write_reg(inst->dev, W5_COMMON_PIC_INFO, stride);
1892
1893	remain = count;
1894	cnt_8_chunk = DIV_ROUND_UP(count, 8);
1895	idx = 0;
1896	for (j = 0; j < cnt_8_chunk; j++) {
1897		reg_val = (j == cnt_8_chunk - 1) << 4 | ((j == 0) << 3);
1898		vpu_write_reg(inst->dev, W5_SFB_OPTION, reg_val);
1899		start_no = j * 8;
1900		end_no = start_no + ((remain >= 8) ? 8 : remain) - 1;
1901
1902		vpu_write_reg(inst->dev, W5_SET_FB_NUM, (start_no << 8) | end_no);
1903
1904		for (i = 0; i < 8 && i < remain; i++) {
1905			vpu_write_reg(inst->dev, W5_ADDR_LUMA_BASE0 + (i << 4), fb_arr[i +
1906					start_no].buf_y);
1907			vpu_write_reg(inst->dev, W5_ADDR_CB_BASE0 + (i << 4),
1908				      fb_arr[i + start_no].buf_cb);
1909			/* luma FBC offset table */
1910			vpu_write_reg(inst->dev, W5_ADDR_FBC_Y_OFFSET0 + (i << 4),
1911				      vb_fbc_y_tbl.daddr + idx * fbc_y_tbl_size);
1912			/* chroma FBC offset table */
1913			vpu_write_reg(inst->dev, W5_ADDR_FBC_C_OFFSET0 + (i << 4),
1914				      vb_fbc_c_tbl.daddr + idx * fbc_c_tbl_size);
1915
1916			vpu_write_reg(inst->dev, W5_ADDR_MV_COL0 + (i << 2),
1917				      vb_mv.daddr + idx * mv_col_size);
1918			idx++;
1919		}
1920		remain -= i;
1921
1922		ret = send_firmware_command(inst, W5_SET_FB, false, NULL, NULL);
1923		if (ret)
1924			goto free_vb_mem;
1925	}
1926
1927	ret = wave5_vpu_firmware_command_queue_error_check(vpu_dev, NULL);
1928	if (ret)
1929		goto free_vb_mem;
1930
1931	return ret;
1932
1933free_vb_mem:
1934	wave5_vdi_free_dma_memory(vpu_dev, &vb_task);
1935free_vb_task:
1936	wave5_vdi_free_dma_memory(vpu_dev, &vb_sub_sam_buf);
1937free_vb_sam_buf:
1938	wave5_vdi_free_dma_memory(vpu_dev, &vb_fbc_c_tbl);
1939free_vb_fbc_c_tbl:
1940	wave5_vdi_free_dma_memory(vpu_dev, &vb_fbc_y_tbl);
1941free_vb_fbc_y_tbl:
1942	wave5_vdi_free_dma_memory(vpu_dev, &vb_mv);
1943	return ret;
1944}
1945
1946int wave5_vpu_encode(struct vpu_instance *inst, struct enc_param *option, u32 *fail_res)
1947{
1948	u32 src_frame_format;
1949	u32 reg_val = 0;
1950	u32 src_stride_c = 0;
1951	struct enc_info *p_enc_info = &inst->codec_info->enc_info;
1952	struct frame_buffer *p_src_frame = option->source_frame;
1953	struct enc_open_param *p_open_param = &p_enc_info->open_param;
1954	bool justified = WTL_RIGHT_JUSTIFIED;
1955	u32 format_no = WTL_PIXEL_8BIT;
1956	int ret;
1957
1958	vpu_write_reg(inst->dev, W5_CMD_ENC_BS_START_ADDR, option->pic_stream_buffer_addr);
1959	vpu_write_reg(inst->dev, W5_CMD_ENC_BS_SIZE, option->pic_stream_buffer_size);
1960	p_enc_info->stream_buf_start_addr = option->pic_stream_buffer_addr;
1961	p_enc_info->stream_buf_size = option->pic_stream_buffer_size;
1962	p_enc_info->stream_buf_end_addr =
1963		option->pic_stream_buffer_addr + option->pic_stream_buffer_size;
1964
1965	vpu_write_reg(inst->dev, W5_CMD_ENC_PIC_SRC_AXI_SEL, DEFAULT_SRC_AXI);
1966	/* secondary AXI */
1967	reg_val = (p_enc_info->sec_axi_info.use_enc_rdo_enable << 11) |
1968		(p_enc_info->sec_axi_info.use_enc_lf_enable << 15);
1969	vpu_write_reg(inst->dev, W5_CMD_ENC_PIC_USE_SEC_AXI, reg_val);
1970
1971	vpu_write_reg(inst->dev, W5_CMD_ENC_PIC_REPORT_PARAM, 0);
1972
1973	/*
1974	 * CODEOPT_ENC_VCL is used to implicitly encode header/headers to generate bitstream.
1975	 * (use ENC_PUT_VIDEO_HEADER for give_command to encode only a header)
1976	 */
1977	if (option->code_option.implicit_header_encode)
1978		vpu_write_reg(inst->dev, W5_CMD_ENC_PIC_CODE_OPTION,
1979			      CODEOPT_ENC_HEADER_IMPLICIT | CODEOPT_ENC_VCL |
1980			      (option->code_option.encode_aud << 5) |
1981			      (option->code_option.encode_eos << 6) |
1982			      (option->code_option.encode_eob << 7));
1983	else
1984		vpu_write_reg(inst->dev, W5_CMD_ENC_PIC_CODE_OPTION,
1985			      option->code_option.implicit_header_encode |
1986			      (option->code_option.encode_vcl << 1) |
1987			      (option->code_option.encode_vps << 2) |
1988			      (option->code_option.encode_sps << 3) |
1989			      (option->code_option.encode_pps << 4) |
1990			      (option->code_option.encode_aud << 5) |
1991			      (option->code_option.encode_eos << 6) |
1992			      (option->code_option.encode_eob << 7));
1993
1994	vpu_write_reg(inst->dev, W5_CMD_ENC_PIC_PIC_PARAM, 0);
1995
1996	if (option->src_end_flag)
1997		/* no more source images. */
1998		vpu_write_reg(inst->dev, W5_CMD_ENC_PIC_SRC_PIC_IDX, 0xFFFFFFFF);
1999	else
2000		vpu_write_reg(inst->dev, W5_CMD_ENC_PIC_SRC_PIC_IDX, option->src_idx);
2001
2002	vpu_write_reg(inst->dev, W5_CMD_ENC_PIC_SRC_ADDR_Y, p_src_frame->buf_y);
2003	vpu_write_reg(inst->dev, W5_CMD_ENC_PIC_SRC_ADDR_U, p_src_frame->buf_cb);
2004	vpu_write_reg(inst->dev, W5_CMD_ENC_PIC_SRC_ADDR_V, p_src_frame->buf_cr);
2005
2006	switch (p_open_param->src_format) {
2007	case FORMAT_420:
2008	case FORMAT_422:
2009	case FORMAT_YUYV:
2010	case FORMAT_YVYU:
2011	case FORMAT_UYVY:
2012	case FORMAT_VYUY:
2013		justified = WTL_LEFT_JUSTIFIED;
2014		format_no = WTL_PIXEL_8BIT;
2015		src_stride_c = inst->cbcr_interleave ? p_src_frame->stride :
2016			(p_src_frame->stride / 2);
2017		src_stride_c = (p_open_param->src_format == FORMAT_422) ? src_stride_c * 2 :
2018			src_stride_c;
2019		break;
2020	case FORMAT_420_P10_16BIT_MSB:
2021	case FORMAT_422_P10_16BIT_MSB:
2022	case FORMAT_YUYV_P10_16BIT_MSB:
2023	case FORMAT_YVYU_P10_16BIT_MSB:
2024	case FORMAT_UYVY_P10_16BIT_MSB:
2025	case FORMAT_VYUY_P10_16BIT_MSB:
2026		justified = WTL_RIGHT_JUSTIFIED;
2027		format_no = WTL_PIXEL_16BIT;
2028		src_stride_c = inst->cbcr_interleave ? p_src_frame->stride :
2029			(p_src_frame->stride / 2);
2030		src_stride_c = (p_open_param->src_format ==
2031				FORMAT_422_P10_16BIT_MSB) ? src_stride_c * 2 : src_stride_c;
2032		break;
2033	case FORMAT_420_P10_16BIT_LSB:
2034	case FORMAT_422_P10_16BIT_LSB:
2035	case FORMAT_YUYV_P10_16BIT_LSB:
2036	case FORMAT_YVYU_P10_16BIT_LSB:
2037	case FORMAT_UYVY_P10_16BIT_LSB:
2038	case FORMAT_VYUY_P10_16BIT_LSB:
2039		justified = WTL_LEFT_JUSTIFIED;
2040		format_no = WTL_PIXEL_16BIT;
2041		src_stride_c = inst->cbcr_interleave ? p_src_frame->stride :
2042			(p_src_frame->stride / 2);
2043		src_stride_c = (p_open_param->src_format ==
2044				FORMAT_422_P10_16BIT_LSB) ? src_stride_c * 2 : src_stride_c;
2045		break;
2046	case FORMAT_420_P10_32BIT_MSB:
2047	case FORMAT_422_P10_32BIT_MSB:
2048	case FORMAT_YUYV_P10_32BIT_MSB:
2049	case FORMAT_YVYU_P10_32BIT_MSB:
2050	case FORMAT_UYVY_P10_32BIT_MSB:
2051	case FORMAT_VYUY_P10_32BIT_MSB:
2052		justified = WTL_RIGHT_JUSTIFIED;
2053		format_no = WTL_PIXEL_32BIT;
2054		src_stride_c = inst->cbcr_interleave ? p_src_frame->stride :
2055			ALIGN(p_src_frame->stride / 2, 16) * BIT(inst->cbcr_interleave);
2056		src_stride_c = (p_open_param->src_format ==
2057				FORMAT_422_P10_32BIT_MSB) ? src_stride_c * 2 : src_stride_c;
2058		break;
2059	case FORMAT_420_P10_32BIT_LSB:
2060	case FORMAT_422_P10_32BIT_LSB:
2061	case FORMAT_YUYV_P10_32BIT_LSB:
2062	case FORMAT_YVYU_P10_32BIT_LSB:
2063	case FORMAT_UYVY_P10_32BIT_LSB:
2064	case FORMAT_VYUY_P10_32BIT_LSB:
2065		justified = WTL_LEFT_JUSTIFIED;
2066		format_no = WTL_PIXEL_32BIT;
2067		src_stride_c = inst->cbcr_interleave ? p_src_frame->stride :
2068			ALIGN(p_src_frame->stride / 2, 16) * BIT(inst->cbcr_interleave);
2069		src_stride_c = (p_open_param->src_format ==
2070				FORMAT_422_P10_32BIT_LSB) ? src_stride_c * 2 : src_stride_c;
2071		break;
2072	default:
2073		return -EINVAL;
2074	}
2075
2076	src_frame_format = (inst->cbcr_interleave << 1) | (inst->nv21);
2077	switch (p_open_param->packed_format) {
2078	case PACKED_YUYV:
2079		src_frame_format = 4;
2080		break;
2081	case PACKED_YVYU:
2082		src_frame_format = 5;
2083		break;
2084	case PACKED_UYVY:
2085		src_frame_format = 6;
2086		break;
2087	case PACKED_VYUY:
2088		src_frame_format = 7;
2089		break;
2090	default:
2091		break;
2092	}
2093
2094	vpu_write_reg(inst->dev, W5_CMD_ENC_PIC_SRC_STRIDE,
2095		      (p_src_frame->stride << 16) | src_stride_c);
2096	vpu_write_reg(inst->dev, W5_CMD_ENC_PIC_SRC_FORMAT, src_frame_format |
2097		      (format_no << 3) | (justified << 5) | (PIC_SRC_ENDIANNESS_BIG_ENDIAN << 6));
2098
2099	vpu_write_reg(inst->dev, W5_CMD_ENC_PIC_CUSTOM_MAP_OPTION_ADDR, 0);
2100	vpu_write_reg(inst->dev, W5_CMD_ENC_PIC_CUSTOM_MAP_OPTION_PARAM, 0);
2101	vpu_write_reg(inst->dev, W5_CMD_ENC_PIC_LONGTERM_PIC, 0);
2102	vpu_write_reg(inst->dev, W5_CMD_ENC_PIC_WP_PIXEL_SIGMA_Y, 0);
2103	vpu_write_reg(inst->dev, W5_CMD_ENC_PIC_WP_PIXEL_SIGMA_C, 0);
2104	vpu_write_reg(inst->dev, W5_CMD_ENC_PIC_WP_PIXEL_MEAN_Y, 0);
2105	vpu_write_reg(inst->dev, W5_CMD_ENC_PIC_WP_PIXEL_MEAN_C, 0);
2106	vpu_write_reg(inst->dev, W5_CMD_ENC_PIC_PREFIX_SEI_INFO, 0);
2107	vpu_write_reg(inst->dev, W5_CMD_ENC_PIC_PREFIX_SEI_NAL_ADDR, 0);
2108	vpu_write_reg(inst->dev, W5_CMD_ENC_PIC_SUFFIX_SEI_INFO, 0);
2109	vpu_write_reg(inst->dev, W5_CMD_ENC_PIC_SUFFIX_SEI_NAL_ADDR, 0);
2110
2111	ret = send_firmware_command(inst, W5_DEC_ENC_PIC, true, &reg_val, fail_res);
2112	if (ret == -ETIMEDOUT)
2113		return ret;
2114
2115	p_enc_info->instance_queue_count = (reg_val >> 16) & 0xff;
2116	p_enc_info->report_queue_count = (reg_val & QUEUE_REPORT_MASK);
2117
2118	if (ret)
2119		return ret;
2120
2121	return 0;
2122}
2123
2124int wave5_vpu_enc_get_result(struct vpu_instance *inst, struct enc_output_info *result)
2125{
2126	int ret;
2127	u32 encoding_success;
2128	u32 reg_val;
2129	struct enc_info *p_enc_info = &inst->codec_info->enc_info;
2130	struct vpu_device *vpu_dev = inst->dev;
2131
2132	ret = wave5_send_query(inst->dev, inst, GET_RESULT);
2133	if (ret)
2134		return ret;
2135
2136	dev_dbg(inst->dev->dev, "%s: enc pic complete\n", __func__);
2137
2138	reg_val = vpu_read_reg(inst->dev, W5_RET_QUEUE_STATUS);
2139
2140	p_enc_info->instance_queue_count = (reg_val >> 16) & 0xff;
2141	p_enc_info->report_queue_count = (reg_val & QUEUE_REPORT_MASK);
2142
2143	encoding_success = vpu_read_reg(inst->dev, W5_RET_ENC_ENCODING_SUCCESS);
2144	if (!encoding_success) {
2145		result->error_reason = vpu_read_reg(inst->dev, W5_RET_ENC_ERR_INFO);
2146		return -EIO;
2147	}
2148
2149	result->warn_info = vpu_read_reg(inst->dev, W5_RET_ENC_WARN_INFO);
2150
2151	reg_val = vpu_read_reg(inst->dev, W5_RET_ENC_PIC_TYPE);
2152	result->pic_type = reg_val & 0xFFFF;
2153
2154	result->enc_vcl_nut = vpu_read_reg(inst->dev, W5_RET_ENC_VCL_NUT);
2155	/*
2156	 * To get the reconstructed frame use the following index on
2157	 * inst->frame_buf
2158	 */
2159	result->recon_frame_index = vpu_read_reg(inst->dev, W5_RET_ENC_PIC_IDX);
2160	result->enc_pic_byte = vpu_read_reg(inst->dev, W5_RET_ENC_PIC_BYTE);
2161	result->enc_src_idx = vpu_read_reg(inst->dev, W5_RET_ENC_USED_SRC_IDX);
2162	p_enc_info->stream_wr_ptr = vpu_read_reg(inst->dev, W5_RET_ENC_WR_PTR);
2163	p_enc_info->stream_rd_ptr = vpu_read_reg(inst->dev, W5_RET_ENC_RD_PTR);
2164
2165	result->bitstream_buffer = vpu_read_reg(inst->dev, W5_RET_ENC_RD_PTR);
2166	result->rd_ptr = p_enc_info->stream_rd_ptr;
2167	result->wr_ptr = p_enc_info->stream_wr_ptr;
2168
2169	/*result for header only(no vcl) encoding */
2170	if (result->recon_frame_index == RECON_IDX_FLAG_HEADER_ONLY)
2171		result->bitstream_size = result->enc_pic_byte;
2172	else if (result->recon_frame_index < 0)
2173		result->bitstream_size = 0;
2174	else
2175		result->bitstream_size = result->enc_pic_byte;
2176
2177	result->enc_host_cmd_tick = vpu_read_reg(inst->dev, W5_RET_ENC_HOST_CMD_TICK);
2178	result->enc_encode_end_tick = vpu_read_reg(inst->dev, W5_RET_ENC_ENCODING_END_TICK);
2179
2180	if (!p_enc_info->first_cycle_check) {
2181		result->frame_cycle = (result->enc_encode_end_tick - result->enc_host_cmd_tick) *
2182			p_enc_info->cycle_per_tick;
2183		p_enc_info->first_cycle_check = true;
2184	} else {
2185		result->frame_cycle =
2186			(result->enc_encode_end_tick - vpu_dev->last_performance_cycles) *
2187			p_enc_info->cycle_per_tick;
2188		if (vpu_dev->last_performance_cycles < result->enc_host_cmd_tick)
2189			result->frame_cycle = (result->enc_encode_end_tick -
2190					result->enc_host_cmd_tick) * p_enc_info->cycle_per_tick;
2191	}
2192	vpu_dev->last_performance_cycles = result->enc_encode_end_tick;
2193
2194	return 0;
2195}
2196
2197int wave5_vpu_enc_finish_seq(struct vpu_instance *inst, u32 *fail_res)
2198{
2199	return send_firmware_command(inst, W5_DESTROY_INSTANCE, true, NULL, fail_res);
2200}
2201
2202static bool wave5_vpu_enc_check_common_param_valid(struct vpu_instance *inst,
2203						   struct enc_open_param *open_param)
2204{
2205	bool low_delay = true;
2206	struct enc_wave_param *param = &open_param->wave_param;
2207	struct vpu_device *vpu_dev = inst->dev;
2208	struct device *dev = vpu_dev->dev;
2209	u32 num_ctu_row = (open_param->pic_height + 64 - 1) / 64;
2210	u32 num_ctu_col = (open_param->pic_width + 64 - 1) / 64;
2211	u32 ctu_sz = num_ctu_col * num_ctu_row;
2212
2213	if (inst->std == W_HEVC_ENC && low_delay &&
2214	    param->decoding_refresh_type == DEC_REFRESH_TYPE_CRA) {
2215		dev_warn(dev,
2216			 "dec_refresh_type(CRA) shouldn't be used together with low delay GOP\n");
2217		dev_warn(dev, "Suggested configuration parameter: decoding refresh type (IDR)\n");
2218		param->decoding_refresh_type = 2;
2219	}
2220
2221	if (param->wpp_enable && param->independ_slice_mode) {
2222		unsigned int num_ctb_in_width = ALIGN(open_param->pic_width, 64) >> 6;
2223
2224		if (param->independ_slice_mode_arg % num_ctb_in_width) {
2225			dev_err(dev, "independ_slice_mode_arg %u must be a multiple of %u\n",
2226				param->independ_slice_mode_arg, num_ctb_in_width);
2227			return false;
2228		}
2229	}
2230
2231	/* multi-slice & wpp */
2232	if (param->wpp_enable && param->depend_slice_mode) {
2233		dev_err(dev, "wpp_enable && depend_slice_mode cannot be used simultaneously\n");
2234		return false;
2235	}
2236
2237	if (!param->independ_slice_mode && param->depend_slice_mode) {
2238		dev_err(dev, "depend_slice_mode requires independ_slice_mode\n");
2239		return false;
2240	} else if (param->independ_slice_mode &&
2241		   param->depend_slice_mode == DEPEND_SLICE_MODE_RECOMMENDED &&
2242		   param->independ_slice_mode_arg < param->depend_slice_mode_arg) {
2243		dev_err(dev, "independ_slice_mode_arg: %u must be smaller than %u\n",
2244			param->independ_slice_mode_arg, param->depend_slice_mode_arg);
2245		return false;
2246	}
2247
2248	if (param->independ_slice_mode && param->independ_slice_mode_arg > 65535) {
2249		dev_err(dev, "independ_slice_mode_arg: %u must be smaller than 65535\n",
2250			param->independ_slice_mode_arg);
2251		return false;
2252	}
2253
2254	if (param->depend_slice_mode && param->depend_slice_mode_arg > 65535) {
2255		dev_err(dev, "depend_slice_mode_arg: %u must be smaller than 65535\n",
2256			param->depend_slice_mode_arg);
2257		return false;
2258	}
2259
2260	if (param->conf_win_top % 2) {
2261		dev_err(dev, "conf_win_top: %u, must be a multiple of 2\n", param->conf_win_top);
2262		return false;
2263	}
2264
2265	if (param->conf_win_bot % 2) {
2266		dev_err(dev, "conf_win_bot: %u, must be a multiple of 2\n", param->conf_win_bot);
2267		return false;
2268	}
2269
2270	if (param->conf_win_left % 2) {
2271		dev_err(dev, "conf_win_left: %u, must be a multiple of 2\n", param->conf_win_left);
2272		return false;
2273	}
2274
2275	if (param->conf_win_right % 2) {
2276		dev_err(dev, "conf_win_right: %u, Must be a multiple of 2\n",
2277			param->conf_win_right);
2278		return false;
2279	}
2280
2281	if (param->lossless_enable && open_param->rc_enable) {
2282		dev_err(dev, "option rate_control cannot be used with lossless_coding\n");
2283		return false;
2284	}
2285
2286	if (param->lossless_enable && !param->skip_intra_trans) {
2287		dev_err(dev, "option intra_trans_skip must be enabled with lossless_coding\n");
2288		return false;
2289	}
2290
2291	/* intra refresh */
2292	if (param->intra_refresh_mode && param->intra_refresh_arg == 0) {
2293		dev_err(dev, "Invalid refresh argument, mode: %u, refresh: %u must be > 0\n",
2294			param->intra_refresh_mode, param->intra_refresh_arg);
2295		return false;
2296	}
2297	switch (param->intra_refresh_mode) {
2298	case REFRESH_MODE_CTU_ROWS:
2299		if (param->intra_mb_refresh_arg > num_ctu_row)
2300			goto invalid_refresh_argument;
2301		break;
2302	case REFRESH_MODE_CTU_COLUMNS:
2303		if (param->intra_refresh_arg > num_ctu_col)
2304			goto invalid_refresh_argument;
2305		break;
2306	case REFRESH_MODE_CTU_STEP_SIZE:
2307		if (param->intra_refresh_arg > ctu_sz)
2308			goto invalid_refresh_argument;
2309		break;
2310	case REFRESH_MODE_CTUS:
2311		if (param->intra_refresh_arg > ctu_sz)
2312			goto invalid_refresh_argument;
2313		if (param->lossless_enable) {
2314			dev_err(dev, "mode: %u cannot be used lossless_enable",
2315				param->intra_refresh_mode);
2316			return false;
2317		}
2318	}
2319	return true;
2320
2321invalid_refresh_argument:
2322	dev_err(dev, "Invalid refresh argument, mode: %u, refresh: %u > W(%u)xH(%u)\n",
2323		param->intra_refresh_mode, param->intra_refresh_arg,
2324		num_ctu_row, num_ctu_col);
2325	return false;
2326}
2327
2328static bool wave5_vpu_enc_check_param_valid(struct vpu_device *vpu_dev,
2329					    struct enc_open_param *open_param)
2330{
2331	struct enc_wave_param *param = &open_param->wave_param;
2332
2333	if (open_param->rc_enable) {
2334		if (param->min_qp_i > param->max_qp_i || param->min_qp_p > param->max_qp_p ||
2335		    param->min_qp_b > param->max_qp_b) {
2336			dev_err(vpu_dev->dev, "Configuration failed because min_qp is greater than max_qp\n");
2337			dev_err(vpu_dev->dev, "Suggested configuration parameters: min_qp = max_qp\n");
2338			return false;
2339		}
2340
2341		if (open_param->bit_rate <= (int)open_param->frame_rate_info) {
2342			dev_err(vpu_dev->dev,
2343				"enc_bit_rate: %u must be greater than the frame_rate: %u\n",
2344				open_param->bit_rate, (int)open_param->frame_rate_info);
2345			return false;
2346		}
2347	}
2348
2349	return true;
2350}
2351
2352int wave5_vpu_enc_check_open_param(struct vpu_instance *inst, struct enc_open_param *open_param)
2353{
2354	u32 pic_width;
2355	u32 pic_height;
2356	s32 product_id = inst->dev->product;
2357	struct vpu_attr *p_attr = &inst->dev->attr;
2358	struct enc_wave_param *param;
2359
2360	if (!open_param)
2361		return -EINVAL;
2362
2363	param = &open_param->wave_param;
2364	pic_width = open_param->pic_width;
2365	pic_height = open_param->pic_height;
2366
2367	if (inst->id >= MAX_NUM_INSTANCE) {
2368		dev_err(inst->dev->dev, "Too many simultaneous instances: %d (max: %u)\n",
2369			inst->id, MAX_NUM_INSTANCE);
2370		return -EOPNOTSUPP;
2371	}
2372
2373	if (inst->std != W_HEVC_ENC &&
2374	    !(inst->std == W_AVC_ENC && product_id == PRODUCT_ID_521)) {
2375		dev_err(inst->dev->dev, "Unsupported encoder-codec & product combination\n");
2376		return -EOPNOTSUPP;
2377	}
2378
2379	if (param->internal_bit_depth == 10) {
2380		if (inst->std == W_HEVC_ENC && !p_attr->support_hevc10bit_enc) {
2381			dev_err(inst->dev->dev,
2382				"Flag support_hevc10bit_enc must be set to encode 10bit HEVC\n");
2383			return -EOPNOTSUPP;
2384		} else if (inst->std == W_AVC_ENC && !p_attr->support_avc10bit_enc) {
2385			dev_err(inst->dev->dev,
2386				"Flag support_avc10bit_enc must be set to encode 10bit AVC\n");
2387			return -EOPNOTSUPP;
2388		}
2389	}
2390
2391	if (!open_param->frame_rate_info) {
2392		dev_err(inst->dev->dev, "No frame rate information.\n");
2393		return -EINVAL;
2394	}
2395
2396	if (open_param->bit_rate > MAX_BIT_RATE) {
2397		dev_err(inst->dev->dev, "Invalid encoding bit-rate: %u (valid: 0-%u)\n",
2398			open_param->bit_rate, MAX_BIT_RATE);
2399		return -EINVAL;
2400	}
2401
2402	if (pic_width < W5_MIN_ENC_PIC_WIDTH || pic_width > W5_MAX_ENC_PIC_WIDTH ||
2403	    pic_height < W5_MIN_ENC_PIC_HEIGHT || pic_height > W5_MAX_ENC_PIC_HEIGHT) {
2404		dev_err(inst->dev->dev, "Invalid encoding dimension: %ux%u\n",
2405			pic_width, pic_height);
2406		return -EINVAL;
2407	}
2408
2409	if (param->profile) {
2410		if (inst->std == W_HEVC_ENC) {
2411			if ((param->profile != HEVC_PROFILE_MAIN ||
2412			     (param->profile == HEVC_PROFILE_MAIN &&
2413			      param->internal_bit_depth > 8)) &&
2414			    (param->profile != HEVC_PROFILE_MAIN10 ||
2415			     (param->profile == HEVC_PROFILE_MAIN10 &&
2416			      param->internal_bit_depth < 10)) &&
2417			    param->profile != HEVC_PROFILE_STILLPICTURE) {
2418				dev_err(inst->dev->dev,
2419					"Invalid HEVC encoding profile: %u (bit-depth: %u)\n",
2420					param->profile, param->internal_bit_depth);
2421				return -EINVAL;
2422			}
2423		} else if (inst->std == W_AVC_ENC) {
2424			if ((param->internal_bit_depth > 8 &&
2425			     param->profile != H264_PROFILE_HIGH10)) {
2426				dev_err(inst->dev->dev,
2427					"Invalid AVC encoding profile: %u (bit-depth: %u)\n",
2428					param->profile, param->internal_bit_depth);
2429				return -EINVAL;
2430			}
2431		}
2432	}
2433
2434	if (param->decoding_refresh_type > DEC_REFRESH_TYPE_IDR) {
2435		dev_err(inst->dev->dev, "Invalid decoding refresh type: %u (valid: 0-2)\n",
2436			param->decoding_refresh_type);
2437		return -EINVAL;
2438	}
2439
2440	if (param->intra_refresh_mode > REFRESH_MODE_CTUS) {
2441		dev_err(inst->dev->dev, "Invalid intra refresh mode: %d (valid: 0-4)\n",
2442			param->intra_refresh_mode);
2443		return -EINVAL;
2444	}
2445
2446	if (inst->std == W_HEVC_ENC && param->independ_slice_mode &&
2447	    param->depend_slice_mode > DEPEND_SLICE_MODE_BOOST) {
2448		dev_err(inst->dev->dev,
2449			"Can't combine slice modes: independent and fast dependent for HEVC\n");
2450		return -EINVAL;
2451	}
2452
2453	if (!param->disable_deblk) {
2454		if (param->beta_offset_div2 < -6 || param->beta_offset_div2 > 6) {
2455			dev_err(inst->dev->dev, "Invalid beta offset: %d (valid: -6-6)\n",
2456				param->beta_offset_div2);
2457			return -EINVAL;
2458		}
2459
2460		if (param->tc_offset_div2 < -6 || param->tc_offset_div2 > 6) {
2461			dev_err(inst->dev->dev, "Invalid tc offset: %d (valid: -6-6)\n",
2462				param->tc_offset_div2);
2463			return -EINVAL;
2464		}
2465	}
2466
2467	if (param->intra_qp > MAX_INTRA_QP) {
2468		dev_err(inst->dev->dev,
2469			"Invalid intra quantization parameter: %u (valid: 0-%u)\n",
2470			param->intra_qp, MAX_INTRA_QP);
2471		return -EINVAL;
2472	}
2473
2474	if (open_param->rc_enable) {
2475		if (param->min_qp_i > MAX_INTRA_QP || param->max_qp_i > MAX_INTRA_QP ||
2476		    param->min_qp_p > MAX_INTRA_QP || param->max_qp_p > MAX_INTRA_QP ||
2477		    param->min_qp_b > MAX_INTRA_QP || param->max_qp_b > MAX_INTRA_QP) {
2478			dev_err(inst->dev->dev,
2479				"Invalid quantization parameter min/max values: "
2480				"I: %u-%u, P: %u-%u, B: %u-%u (valid for each: 0-%u)\n",
2481				param->min_qp_i, param->max_qp_i, param->min_qp_p, param->max_qp_p,
2482				param->min_qp_b, param->max_qp_b, MAX_INTRA_QP);
2483			return -EINVAL;
2484		}
2485
2486		if (param->hvs_qp_enable && param->hvs_max_delta_qp > MAX_HVS_MAX_DELTA_QP) {
2487			dev_err(inst->dev->dev,
2488				"Invalid HVS max delta quantization parameter: %u (valid: 0-%u)\n",
2489				param->hvs_max_delta_qp, MAX_HVS_MAX_DELTA_QP);
2490			return -EINVAL;
2491		}
2492
2493		if (open_param->vbv_buffer_size < MIN_VBV_BUFFER_SIZE ||
2494		    open_param->vbv_buffer_size > MAX_VBV_BUFFER_SIZE) {
2495			dev_err(inst->dev->dev, "VBV buffer size: %u (valid: %u-%u)\n",
2496				open_param->vbv_buffer_size, MIN_VBV_BUFFER_SIZE,
2497				MAX_VBV_BUFFER_SIZE);
2498			return -EINVAL;
2499		}
2500	}
2501
2502	if (!wave5_vpu_enc_check_common_param_valid(inst, open_param))
2503		return -EINVAL;
2504
2505	if (!wave5_vpu_enc_check_param_valid(inst->dev, open_param))
2506		return -EINVAL;
2507
2508	if (param->chroma_cb_qp_offset < -12 || param->chroma_cb_qp_offset > 12) {
2509		dev_err(inst->dev->dev,
2510			"Invalid chroma Cb quantization parameter offset: %d (valid: -12-12)\n",
2511			param->chroma_cb_qp_offset);
2512		return -EINVAL;
2513	}
2514
2515	if (param->chroma_cr_qp_offset < -12 || param->chroma_cr_qp_offset > 12) {
2516		dev_err(inst->dev->dev,
2517			"Invalid chroma Cr quantization parameter offset: %d (valid: -12-12)\n",
2518			param->chroma_cr_qp_offset);
2519		return -EINVAL;
2520	}
2521
2522	if (param->intra_refresh_mode == REFRESH_MODE_CTU_STEP_SIZE && !param->intra_refresh_arg) {
2523		dev_err(inst->dev->dev,
2524			"Intra refresh mode CTU step-size requires an argument\n");
2525		return -EINVAL;
2526	}
2527
2528	if (inst->std == W_HEVC_ENC) {
2529		if (param->nr_intra_weight_y > MAX_INTRA_WEIGHT ||
2530		    param->nr_intra_weight_cb > MAX_INTRA_WEIGHT ||
2531		    param->nr_intra_weight_cr > MAX_INTRA_WEIGHT) {
2532			dev_err(inst->dev->dev,
2533				"Invalid intra weight Y(%u) Cb(%u) Cr(%u) (valid: %u)\n",
2534				param->nr_intra_weight_y, param->nr_intra_weight_cb,
2535				param->nr_intra_weight_cr, MAX_INTRA_WEIGHT);
2536			return -EINVAL;
2537		}
2538
2539		if (param->nr_inter_weight_y > MAX_INTER_WEIGHT ||
2540		    param->nr_inter_weight_cb > MAX_INTER_WEIGHT ||
2541		    param->nr_inter_weight_cr > MAX_INTER_WEIGHT) {
2542			dev_err(inst->dev->dev,
2543				"Invalid inter weight Y(%u) Cb(%u) Cr(%u) (valid: %u)\n",
2544				param->nr_inter_weight_y, param->nr_inter_weight_cb,
2545				param->nr_inter_weight_cr, MAX_INTER_WEIGHT);
2546			return -EINVAL;
2547		}
2548	}
2549
2550	return 0;
2551}
2552