1// SPDX-License-Identifier: GPL-2.0 2/* 3 * Rockchip Video Decoder VP9 backend 4 * 5 * Copyright (C) 2019 Collabora, Ltd. 6 * Boris Brezillon <boris.brezillon@collabora.com> 7 * Copyright (C) 2021 Collabora, Ltd. 8 * Andrzej Pietrasiewicz <andrzej.p@collabora.com> 9 * 10 * Copyright (C) 2016 Rockchip Electronics Co., Ltd. 11 * Alpha Lin <Alpha.Lin@rock-chips.com> 12 */ 13 14/* 15 * For following the vp9 spec please start reading this driver 16 * code from rkvdec_vp9_run() followed by rkvdec_vp9_done(). 17 */ 18 19#include <linux/kernel.h> 20#include <linux/vmalloc.h> 21#include <media/v4l2-mem2mem.h> 22#include <media/v4l2-vp9.h> 23 24#include "rkvdec.h" 25#include "rkvdec-regs.h" 26 27#define RKVDEC_VP9_PROBE_SIZE 4864 28#define RKVDEC_VP9_COUNT_SIZE 13232 29#define RKVDEC_VP9_MAX_SEGMAP_SIZE 73728 30 31struct rkvdec_vp9_intra_mode_probs { 32 u8 y_mode[105]; 33 u8 uv_mode[23]; 34}; 35 36struct rkvdec_vp9_intra_only_frame_probs { 37 u8 coef_intra[4][2][128]; 38 struct rkvdec_vp9_intra_mode_probs intra_mode[10]; 39}; 40 41struct rkvdec_vp9_inter_frame_probs { 42 u8 y_mode[4][9]; 43 u8 comp_mode[5]; 44 u8 comp_ref[5]; 45 u8 single_ref[5][2]; 46 u8 inter_mode[7][3]; 47 u8 interp_filter[4][2]; 48 u8 padding0[11]; 49 u8 coef[2][4][2][128]; 50 u8 uv_mode_0_2[3][9]; 51 u8 padding1[5]; 52 u8 uv_mode_3_5[3][9]; 53 u8 padding2[5]; 54 u8 uv_mode_6_8[3][9]; 55 u8 padding3[5]; 56 u8 uv_mode_9[9]; 57 u8 padding4[7]; 58 u8 padding5[16]; 59 struct { 60 u8 joint[3]; 61 u8 sign[2]; 62 u8 classes[2][10]; 63 u8 class0_bit[2]; 64 u8 bits[2][10]; 65 u8 class0_fr[2][2][3]; 66 u8 fr[2][3]; 67 u8 class0_hp[2]; 68 u8 hp[2]; 69 } mv; 70}; 71 72struct rkvdec_vp9_probs { 73 u8 partition[16][3]; 74 u8 pred[3]; 75 u8 tree[7]; 76 u8 skip[3]; 77 u8 tx32[2][3]; 78 u8 tx16[2][2]; 79 u8 tx8[2][1]; 80 u8 is_inter[4]; 81 /* 128 bit alignment */ 82 u8 padding0[3]; 83 union { 84 struct rkvdec_vp9_inter_frame_probs inter; 85 struct rkvdec_vp9_intra_only_frame_probs intra_only; 86 }; 87 /* 128 bit alignment */ 88 u8 padding1[11]; 89}; 90 91/* Data structure describing auxiliary buffer format. */ 92struct rkvdec_vp9_priv_tbl { 93 struct rkvdec_vp9_probs probs; 94 u8 segmap[2][RKVDEC_VP9_MAX_SEGMAP_SIZE]; 95}; 96 97struct rkvdec_vp9_refs_counts { 98 u32 eob[2]; 99 u32 coeff[3]; 100}; 101 102struct rkvdec_vp9_inter_frame_symbol_counts { 103 u32 partition[16][4]; 104 u32 skip[3][2]; 105 u32 inter[4][2]; 106 u32 tx32p[2][4]; 107 u32 tx16p[2][4]; 108 u32 tx8p[2][2]; 109 u32 y_mode[4][10]; 110 u32 uv_mode[10][10]; 111 u32 comp[5][2]; 112 u32 comp_ref[5][2]; 113 u32 single_ref[5][2][2]; 114 u32 mv_mode[7][4]; 115 u32 filter[4][3]; 116 u32 mv_joint[4]; 117 u32 sign[2][2]; 118 /* add 1 element for align */ 119 u32 classes[2][11 + 1]; 120 u32 class0[2][2]; 121 u32 bits[2][10][2]; 122 u32 class0_fp[2][2][4]; 123 u32 fp[2][4]; 124 u32 class0_hp[2][2]; 125 u32 hp[2][2]; 126 struct rkvdec_vp9_refs_counts ref_cnt[2][4][2][6][6]; 127}; 128 129struct rkvdec_vp9_intra_frame_symbol_counts { 130 u32 partition[4][4][4]; 131 u32 skip[3][2]; 132 u32 intra[4][2]; 133 u32 tx32p[2][4]; 134 u32 tx16p[2][4]; 135 u32 tx8p[2][2]; 136 struct rkvdec_vp9_refs_counts ref_cnt[2][4][2][6][6]; 137}; 138 139struct rkvdec_vp9_run { 140 struct rkvdec_run base; 141 const struct v4l2_ctrl_vp9_frame *decode_params; 142}; 143 144struct rkvdec_vp9_frame_info { 145 u32 valid : 1; 146 u32 segmapid : 1; 147 u32 frame_context_idx : 2; 148 u32 reference_mode : 2; 149 u32 tx_mode : 3; 150 u32 interpolation_filter : 3; 151 u32 flags; 152 u64 timestamp; 153 struct v4l2_vp9_segmentation seg; 154 struct v4l2_vp9_loop_filter lf; 155}; 156 157struct rkvdec_vp9_ctx { 158 struct rkvdec_aux_buf priv_tbl; 159 struct rkvdec_aux_buf count_tbl; 160 struct v4l2_vp9_frame_symbol_counts inter_cnts; 161 struct v4l2_vp9_frame_symbol_counts intra_cnts; 162 struct v4l2_vp9_frame_context probability_tables; 163 struct v4l2_vp9_frame_context frame_context[4]; 164 struct rkvdec_vp9_frame_info cur; 165 struct rkvdec_vp9_frame_info last; 166}; 167 168static void write_coeff_plane(const u8 coef[6][6][3], u8 *coeff_plane) 169{ 170 unsigned int idx = 0, byte_count = 0; 171 int k, m, n; 172 u8 p; 173 174 for (k = 0; k < 6; k++) { 175 for (m = 0; m < 6; m++) { 176 for (n = 0; n < 3; n++) { 177 p = coef[k][m][n]; 178 coeff_plane[idx++] = p; 179 byte_count++; 180 if (byte_count == 27) { 181 idx += 5; 182 byte_count = 0; 183 } 184 } 185 } 186 } 187} 188 189static void init_intra_only_probs(struct rkvdec_ctx *ctx, 190 const struct rkvdec_vp9_run *run) 191{ 192 struct rkvdec_vp9_ctx *vp9_ctx = ctx->priv; 193 struct rkvdec_vp9_priv_tbl *tbl = vp9_ctx->priv_tbl.cpu; 194 struct rkvdec_vp9_intra_only_frame_probs *rkprobs; 195 const struct v4l2_vp9_frame_context *probs; 196 unsigned int i, j, k; 197 198 rkprobs = &tbl->probs.intra_only; 199 probs = &vp9_ctx->probability_tables; 200 201 /* 202 * intra only 149 x 128 bits ,aligned to 152 x 128 bits coeff related 203 * prob 64 x 128 bits 204 */ 205 for (i = 0; i < ARRAY_SIZE(probs->coef); i++) { 206 for (j = 0; j < ARRAY_SIZE(probs->coef[0]); j++) 207 write_coeff_plane(probs->coef[i][j][0], 208 rkprobs->coef_intra[i][j]); 209 } 210 211 /* intra mode prob 80 x 128 bits */ 212 for (i = 0; i < ARRAY_SIZE(v4l2_vp9_kf_y_mode_prob); i++) { 213 unsigned int byte_count = 0; 214 int idx = 0; 215 216 /* vp9_kf_y_mode_prob */ 217 for (j = 0; j < ARRAY_SIZE(v4l2_vp9_kf_y_mode_prob[0]); j++) { 218 for (k = 0; k < ARRAY_SIZE(v4l2_vp9_kf_y_mode_prob[0][0]); 219 k++) { 220 u8 val = v4l2_vp9_kf_y_mode_prob[i][j][k]; 221 222 rkprobs->intra_mode[i].y_mode[idx++] = val; 223 byte_count++; 224 if (byte_count == 27) { 225 byte_count = 0; 226 idx += 5; 227 } 228 } 229 } 230 } 231 232 for (i = 0; i < sizeof(v4l2_vp9_kf_uv_mode_prob); ++i) { 233 const u8 *ptr = (const u8 *)v4l2_vp9_kf_uv_mode_prob; 234 235 rkprobs->intra_mode[i / 23].uv_mode[i % 23] = ptr[i]; 236 } 237} 238 239static void init_inter_probs(struct rkvdec_ctx *ctx, 240 const struct rkvdec_vp9_run *run) 241{ 242 struct rkvdec_vp9_ctx *vp9_ctx = ctx->priv; 243 struct rkvdec_vp9_priv_tbl *tbl = vp9_ctx->priv_tbl.cpu; 244 struct rkvdec_vp9_inter_frame_probs *rkprobs; 245 const struct v4l2_vp9_frame_context *probs; 246 unsigned int i, j, k; 247 248 rkprobs = &tbl->probs.inter; 249 probs = &vp9_ctx->probability_tables; 250 251 /* 252 * inter probs 253 * 151 x 128 bits, aligned to 152 x 128 bits 254 * inter only 255 * intra_y_mode & inter_block info 6 x 128 bits 256 */ 257 258 memcpy(rkprobs->y_mode, probs->y_mode, sizeof(rkprobs->y_mode)); 259 memcpy(rkprobs->comp_mode, probs->comp_mode, 260 sizeof(rkprobs->comp_mode)); 261 memcpy(rkprobs->comp_ref, probs->comp_ref, 262 sizeof(rkprobs->comp_ref)); 263 memcpy(rkprobs->single_ref, probs->single_ref, 264 sizeof(rkprobs->single_ref)); 265 memcpy(rkprobs->inter_mode, probs->inter_mode, 266 sizeof(rkprobs->inter_mode)); 267 memcpy(rkprobs->interp_filter, probs->interp_filter, 268 sizeof(rkprobs->interp_filter)); 269 270 /* 128 x 128 bits coeff related */ 271 for (i = 0; i < ARRAY_SIZE(probs->coef); i++) { 272 for (j = 0; j < ARRAY_SIZE(probs->coef[0]); j++) { 273 for (k = 0; k < ARRAY_SIZE(probs->coef[0][0]); k++) 274 write_coeff_plane(probs->coef[i][j][k], 275 rkprobs->coef[k][i][j]); 276 } 277 } 278 279 /* intra uv mode 6 x 128 */ 280 memcpy(rkprobs->uv_mode_0_2, &probs->uv_mode[0], 281 sizeof(rkprobs->uv_mode_0_2)); 282 memcpy(rkprobs->uv_mode_3_5, &probs->uv_mode[3], 283 sizeof(rkprobs->uv_mode_3_5)); 284 memcpy(rkprobs->uv_mode_6_8, &probs->uv_mode[6], 285 sizeof(rkprobs->uv_mode_6_8)); 286 memcpy(rkprobs->uv_mode_9, &probs->uv_mode[9], 287 sizeof(rkprobs->uv_mode_9)); 288 289 /* mv related 6 x 128 */ 290 memcpy(rkprobs->mv.joint, probs->mv.joint, 291 sizeof(rkprobs->mv.joint)); 292 memcpy(rkprobs->mv.sign, probs->mv.sign, 293 sizeof(rkprobs->mv.sign)); 294 memcpy(rkprobs->mv.classes, probs->mv.classes, 295 sizeof(rkprobs->mv.classes)); 296 memcpy(rkprobs->mv.class0_bit, probs->mv.class0_bit, 297 sizeof(rkprobs->mv.class0_bit)); 298 memcpy(rkprobs->mv.bits, probs->mv.bits, 299 sizeof(rkprobs->mv.bits)); 300 memcpy(rkprobs->mv.class0_fr, probs->mv.class0_fr, 301 sizeof(rkprobs->mv.class0_fr)); 302 memcpy(rkprobs->mv.fr, probs->mv.fr, 303 sizeof(rkprobs->mv.fr)); 304 memcpy(rkprobs->mv.class0_hp, probs->mv.class0_hp, 305 sizeof(rkprobs->mv.class0_hp)); 306 memcpy(rkprobs->mv.hp, probs->mv.hp, 307 sizeof(rkprobs->mv.hp)); 308} 309 310static void init_probs(struct rkvdec_ctx *ctx, 311 const struct rkvdec_vp9_run *run) 312{ 313 const struct v4l2_ctrl_vp9_frame *dec_params; 314 struct rkvdec_vp9_ctx *vp9_ctx = ctx->priv; 315 struct rkvdec_vp9_priv_tbl *tbl = vp9_ctx->priv_tbl.cpu; 316 struct rkvdec_vp9_probs *rkprobs = &tbl->probs; 317 const struct v4l2_vp9_segmentation *seg; 318 const struct v4l2_vp9_frame_context *probs; 319 bool intra_only; 320 321 dec_params = run->decode_params; 322 probs = &vp9_ctx->probability_tables; 323 seg = &dec_params->seg; 324 325 memset(rkprobs, 0, sizeof(*rkprobs)); 326 327 intra_only = !!(dec_params->flags & 328 (V4L2_VP9_FRAME_FLAG_KEY_FRAME | 329 V4L2_VP9_FRAME_FLAG_INTRA_ONLY)); 330 331 /* sb info 5 x 128 bit */ 332 memcpy(rkprobs->partition, 333 intra_only ? v4l2_vp9_kf_partition_probs : probs->partition, 334 sizeof(rkprobs->partition)); 335 336 memcpy(rkprobs->pred, seg->pred_probs, sizeof(rkprobs->pred)); 337 memcpy(rkprobs->tree, seg->tree_probs, sizeof(rkprobs->tree)); 338 memcpy(rkprobs->skip, probs->skip, sizeof(rkprobs->skip)); 339 memcpy(rkprobs->tx32, probs->tx32, sizeof(rkprobs->tx32)); 340 memcpy(rkprobs->tx16, probs->tx16, sizeof(rkprobs->tx16)); 341 memcpy(rkprobs->tx8, probs->tx8, sizeof(rkprobs->tx8)); 342 memcpy(rkprobs->is_inter, probs->is_inter, sizeof(rkprobs->is_inter)); 343 344 if (intra_only) 345 init_intra_only_probs(ctx, run); 346 else 347 init_inter_probs(ctx, run); 348} 349 350struct rkvdec_vp9_ref_reg { 351 u32 reg_frm_size; 352 u32 reg_hor_stride; 353 u32 reg_y_stride; 354 u32 reg_yuv_stride; 355 u32 reg_ref_base; 356}; 357 358static struct rkvdec_vp9_ref_reg ref_regs[] = { 359 { 360 .reg_frm_size = RKVDEC_REG_VP9_FRAME_SIZE(0), 361 .reg_hor_stride = RKVDEC_VP9_HOR_VIRSTRIDE(0), 362 .reg_y_stride = RKVDEC_VP9_LAST_FRAME_YSTRIDE, 363 .reg_yuv_stride = RKVDEC_VP9_LAST_FRAME_YUVSTRIDE, 364 .reg_ref_base = RKVDEC_REG_VP9_LAST_FRAME_BASE, 365 }, 366 { 367 .reg_frm_size = RKVDEC_REG_VP9_FRAME_SIZE(1), 368 .reg_hor_stride = RKVDEC_VP9_HOR_VIRSTRIDE(1), 369 .reg_y_stride = RKVDEC_VP9_GOLDEN_FRAME_YSTRIDE, 370 .reg_yuv_stride = 0, 371 .reg_ref_base = RKVDEC_REG_VP9_GOLDEN_FRAME_BASE, 372 }, 373 { 374 .reg_frm_size = RKVDEC_REG_VP9_FRAME_SIZE(2), 375 .reg_hor_stride = RKVDEC_VP9_HOR_VIRSTRIDE(2), 376 .reg_y_stride = RKVDEC_VP9_ALTREF_FRAME_YSTRIDE, 377 .reg_yuv_stride = 0, 378 .reg_ref_base = RKVDEC_REG_VP9_ALTREF_FRAME_BASE, 379 } 380}; 381 382static struct rkvdec_decoded_buffer * 383get_ref_buf(struct rkvdec_ctx *ctx, struct vb2_v4l2_buffer *dst, u64 timestamp) 384{ 385 struct v4l2_m2m_ctx *m2m_ctx = ctx->fh.m2m_ctx; 386 struct vb2_queue *cap_q = &m2m_ctx->cap_q_ctx.q; 387 struct vb2_buffer *buf; 388 389 /* 390 * If a ref is unused or invalid, address of current destination 391 * buffer is returned. 392 */ 393 buf = vb2_find_buffer(cap_q, timestamp); 394 if (!buf) 395 buf = &dst->vb2_buf; 396 397 return vb2_to_rkvdec_decoded_buf(buf); 398} 399 400static dma_addr_t get_mv_base_addr(struct rkvdec_decoded_buffer *buf) 401{ 402 unsigned int aligned_pitch, aligned_height, yuv_len; 403 404 aligned_height = round_up(buf->vp9.height, 64); 405 aligned_pitch = round_up(buf->vp9.width * buf->vp9.bit_depth, 512) / 8; 406 yuv_len = (aligned_height * aligned_pitch * 3) / 2; 407 408 return vb2_dma_contig_plane_dma_addr(&buf->base.vb.vb2_buf, 0) + 409 yuv_len; 410} 411 412static void config_ref_registers(struct rkvdec_ctx *ctx, 413 const struct rkvdec_vp9_run *run, 414 struct rkvdec_decoded_buffer *ref_buf, 415 struct rkvdec_vp9_ref_reg *ref_reg) 416{ 417 unsigned int aligned_pitch, aligned_height, y_len, yuv_len; 418 struct rkvdec_dev *rkvdec = ctx->dev; 419 420 aligned_height = round_up(ref_buf->vp9.height, 64); 421 writel_relaxed(RKVDEC_VP9_FRAMEWIDTH(ref_buf->vp9.width) | 422 RKVDEC_VP9_FRAMEHEIGHT(ref_buf->vp9.height), 423 rkvdec->regs + ref_reg->reg_frm_size); 424 425 writel_relaxed(vb2_dma_contig_plane_dma_addr(&ref_buf->base.vb.vb2_buf, 0), 426 rkvdec->regs + ref_reg->reg_ref_base); 427 428 if (&ref_buf->base.vb == run->base.bufs.dst) 429 return; 430 431 aligned_pitch = round_up(ref_buf->vp9.width * ref_buf->vp9.bit_depth, 512) / 8; 432 y_len = aligned_height * aligned_pitch; 433 yuv_len = (y_len * 3) / 2; 434 435 writel_relaxed(RKVDEC_HOR_Y_VIRSTRIDE(aligned_pitch / 16) | 436 RKVDEC_HOR_UV_VIRSTRIDE(aligned_pitch / 16), 437 rkvdec->regs + ref_reg->reg_hor_stride); 438 writel_relaxed(RKVDEC_VP9_REF_YSTRIDE(y_len / 16), 439 rkvdec->regs + ref_reg->reg_y_stride); 440 441 if (!ref_reg->reg_yuv_stride) 442 return; 443 444 writel_relaxed(RKVDEC_VP9_REF_YUVSTRIDE(yuv_len / 16), 445 rkvdec->regs + ref_reg->reg_yuv_stride); 446} 447 448static void config_seg_registers(struct rkvdec_ctx *ctx, unsigned int segid) 449{ 450 struct rkvdec_vp9_ctx *vp9_ctx = ctx->priv; 451 const struct v4l2_vp9_segmentation *seg; 452 struct rkvdec_dev *rkvdec = ctx->dev; 453 s16 feature_val; 454 int feature_id; 455 u32 val = 0; 456 457 seg = vp9_ctx->last.valid ? &vp9_ctx->last.seg : &vp9_ctx->cur.seg; 458 feature_id = V4L2_VP9_SEG_LVL_ALT_Q; 459 if (v4l2_vp9_seg_feat_enabled(seg->feature_enabled, feature_id, segid)) { 460 feature_val = seg->feature_data[segid][feature_id]; 461 val |= RKVDEC_SEGID_FRAME_QP_DELTA_EN(1) | 462 RKVDEC_SEGID_FRAME_QP_DELTA(feature_val); 463 } 464 465 feature_id = V4L2_VP9_SEG_LVL_ALT_L; 466 if (v4l2_vp9_seg_feat_enabled(seg->feature_enabled, feature_id, segid)) { 467 feature_val = seg->feature_data[segid][feature_id]; 468 val |= RKVDEC_SEGID_FRAME_LOOPFILTER_VALUE_EN(1) | 469 RKVDEC_SEGID_FRAME_LOOPFILTER_VALUE(feature_val); 470 } 471 472 feature_id = V4L2_VP9_SEG_LVL_REF_FRAME; 473 if (v4l2_vp9_seg_feat_enabled(seg->feature_enabled, feature_id, segid)) { 474 feature_val = seg->feature_data[segid][feature_id]; 475 val |= RKVDEC_SEGID_REFERINFO_EN(1) | 476 RKVDEC_SEGID_REFERINFO(feature_val); 477 } 478 479 feature_id = V4L2_VP9_SEG_LVL_SKIP; 480 if (v4l2_vp9_seg_feat_enabled(seg->feature_enabled, feature_id, segid)) 481 val |= RKVDEC_SEGID_FRAME_SKIP_EN(1); 482 483 if (!segid && 484 (seg->flags & V4L2_VP9_SEGMENTATION_FLAG_ABS_OR_DELTA_UPDATE)) 485 val |= RKVDEC_SEGID_ABS_DELTA(1); 486 487 writel_relaxed(val, rkvdec->regs + RKVDEC_VP9_SEGID_GRP(segid)); 488} 489 490static void update_dec_buf_info(struct rkvdec_decoded_buffer *buf, 491 const struct v4l2_ctrl_vp9_frame *dec_params) 492{ 493 buf->vp9.width = dec_params->frame_width_minus_1 + 1; 494 buf->vp9.height = dec_params->frame_height_minus_1 + 1; 495 buf->vp9.bit_depth = dec_params->bit_depth; 496} 497 498static void update_ctx_cur_info(struct rkvdec_vp9_ctx *vp9_ctx, 499 struct rkvdec_decoded_buffer *buf, 500 const struct v4l2_ctrl_vp9_frame *dec_params) 501{ 502 vp9_ctx->cur.valid = true; 503 vp9_ctx->cur.reference_mode = dec_params->reference_mode; 504 vp9_ctx->cur.interpolation_filter = dec_params->interpolation_filter; 505 vp9_ctx->cur.flags = dec_params->flags; 506 vp9_ctx->cur.timestamp = buf->base.vb.vb2_buf.timestamp; 507 vp9_ctx->cur.seg = dec_params->seg; 508 vp9_ctx->cur.lf = dec_params->lf; 509} 510 511static void update_ctx_last_info(struct rkvdec_vp9_ctx *vp9_ctx) 512{ 513 vp9_ctx->last = vp9_ctx->cur; 514} 515 516static void config_registers(struct rkvdec_ctx *ctx, 517 const struct rkvdec_vp9_run *run) 518{ 519 unsigned int y_len, uv_len, yuv_len, bit_depth, aligned_height, aligned_pitch, stream_len; 520 const struct v4l2_ctrl_vp9_frame *dec_params; 521 struct rkvdec_decoded_buffer *ref_bufs[3]; 522 struct rkvdec_decoded_buffer *dst, *last, *mv_ref; 523 struct rkvdec_vp9_ctx *vp9_ctx = ctx->priv; 524 u32 val, last_frame_info = 0; 525 const struct v4l2_vp9_segmentation *seg; 526 struct rkvdec_dev *rkvdec = ctx->dev; 527 dma_addr_t addr; 528 bool intra_only; 529 unsigned int i; 530 531 dec_params = run->decode_params; 532 dst = vb2_to_rkvdec_decoded_buf(&run->base.bufs.dst->vb2_buf); 533 ref_bufs[0] = get_ref_buf(ctx, &dst->base.vb, dec_params->last_frame_ts); 534 ref_bufs[1] = get_ref_buf(ctx, &dst->base.vb, dec_params->golden_frame_ts); 535 ref_bufs[2] = get_ref_buf(ctx, &dst->base.vb, dec_params->alt_frame_ts); 536 537 if (vp9_ctx->last.valid) 538 last = get_ref_buf(ctx, &dst->base.vb, vp9_ctx->last.timestamp); 539 else 540 last = dst; 541 542 update_dec_buf_info(dst, dec_params); 543 update_ctx_cur_info(vp9_ctx, dst, dec_params); 544 seg = &dec_params->seg; 545 546 intra_only = !!(dec_params->flags & 547 (V4L2_VP9_FRAME_FLAG_KEY_FRAME | 548 V4L2_VP9_FRAME_FLAG_INTRA_ONLY)); 549 550 writel_relaxed(RKVDEC_MODE(RKVDEC_MODE_VP9), 551 rkvdec->regs + RKVDEC_REG_SYSCTRL); 552 553 bit_depth = dec_params->bit_depth; 554 aligned_height = round_up(ctx->decoded_fmt.fmt.pix_mp.height, 64); 555 556 aligned_pitch = round_up(ctx->decoded_fmt.fmt.pix_mp.width * 557 bit_depth, 558 512) / 8; 559 y_len = aligned_height * aligned_pitch; 560 uv_len = y_len / 2; 561 yuv_len = y_len + uv_len; 562 563 writel_relaxed(RKVDEC_Y_HOR_VIRSTRIDE(aligned_pitch / 16) | 564 RKVDEC_UV_HOR_VIRSTRIDE(aligned_pitch / 16), 565 rkvdec->regs + RKVDEC_REG_PICPAR); 566 writel_relaxed(RKVDEC_Y_VIRSTRIDE(y_len / 16), 567 rkvdec->regs + RKVDEC_REG_Y_VIRSTRIDE); 568 writel_relaxed(RKVDEC_YUV_VIRSTRIDE(yuv_len / 16), 569 rkvdec->regs + RKVDEC_REG_YUV_VIRSTRIDE); 570 571 stream_len = vb2_get_plane_payload(&run->base.bufs.src->vb2_buf, 0); 572 writel_relaxed(RKVDEC_STRM_LEN(stream_len), 573 rkvdec->regs + RKVDEC_REG_STRM_LEN); 574 575 /* 576 * Reset count buffer, because decoder only output intra related syntax 577 * counts when decoding intra frame, but update entropy need to update 578 * all the probabilities. 579 */ 580 if (intra_only) 581 memset(vp9_ctx->count_tbl.cpu, 0, vp9_ctx->count_tbl.size); 582 583 vp9_ctx->cur.segmapid = vp9_ctx->last.segmapid; 584 if (!intra_only && 585 !(dec_params->flags & V4L2_VP9_FRAME_FLAG_ERROR_RESILIENT) && 586 (!(seg->flags & V4L2_VP9_SEGMENTATION_FLAG_ENABLED) || 587 (seg->flags & V4L2_VP9_SEGMENTATION_FLAG_UPDATE_MAP))) 588 vp9_ctx->cur.segmapid++; 589 590 for (i = 0; i < ARRAY_SIZE(ref_bufs); i++) 591 config_ref_registers(ctx, run, ref_bufs[i], &ref_regs[i]); 592 593 for (i = 0; i < 8; i++) 594 config_seg_registers(ctx, i); 595 596 writel_relaxed(RKVDEC_VP9_TX_MODE(vp9_ctx->cur.tx_mode) | 597 RKVDEC_VP9_FRAME_REF_MODE(dec_params->reference_mode), 598 rkvdec->regs + RKVDEC_VP9_CPRHEADER_CONFIG); 599 600 if (!intra_only) { 601 const struct v4l2_vp9_loop_filter *lf; 602 s8 delta; 603 604 if (vp9_ctx->last.valid) 605 lf = &vp9_ctx->last.lf; 606 else 607 lf = &vp9_ctx->cur.lf; 608 609 val = 0; 610 for (i = 0; i < ARRAY_SIZE(lf->ref_deltas); i++) { 611 delta = lf->ref_deltas[i]; 612 val |= RKVDEC_REF_DELTAS_LASTFRAME(i, delta); 613 } 614 615 writel_relaxed(val, 616 rkvdec->regs + RKVDEC_VP9_REF_DELTAS_LASTFRAME); 617 618 for (i = 0; i < ARRAY_SIZE(lf->mode_deltas); i++) { 619 delta = lf->mode_deltas[i]; 620 last_frame_info |= RKVDEC_MODE_DELTAS_LASTFRAME(i, 621 delta); 622 } 623 } 624 625 if (vp9_ctx->last.valid && !intra_only && 626 vp9_ctx->last.seg.flags & V4L2_VP9_SEGMENTATION_FLAG_ENABLED) 627 last_frame_info |= RKVDEC_SEG_EN_LASTFRAME; 628 629 if (vp9_ctx->last.valid && 630 vp9_ctx->last.flags & V4L2_VP9_FRAME_FLAG_SHOW_FRAME) 631 last_frame_info |= RKVDEC_LAST_SHOW_FRAME; 632 633 if (vp9_ctx->last.valid && 634 vp9_ctx->last.flags & 635 (V4L2_VP9_FRAME_FLAG_KEY_FRAME | V4L2_VP9_FRAME_FLAG_INTRA_ONLY)) 636 last_frame_info |= RKVDEC_LAST_INTRA_ONLY; 637 638 if (vp9_ctx->last.valid && 639 last->vp9.width == dst->vp9.width && 640 last->vp9.height == dst->vp9.height) 641 last_frame_info |= RKVDEC_LAST_WIDHHEIGHT_EQCUR; 642 643 writel_relaxed(last_frame_info, 644 rkvdec->regs + RKVDEC_VP9_INFO_LASTFRAME); 645 646 writel_relaxed(stream_len - dec_params->compressed_header_size - 647 dec_params->uncompressed_header_size, 648 rkvdec->regs + RKVDEC_VP9_LASTTILE_SIZE); 649 650 for (i = 0; !intra_only && i < ARRAY_SIZE(ref_bufs); i++) { 651 unsigned int refw = ref_bufs[i]->vp9.width; 652 unsigned int refh = ref_bufs[i]->vp9.height; 653 u32 hscale, vscale; 654 655 hscale = (refw << 14) / dst->vp9.width; 656 vscale = (refh << 14) / dst->vp9.height; 657 writel_relaxed(RKVDEC_VP9_REF_HOR_SCALE(hscale) | 658 RKVDEC_VP9_REF_VER_SCALE(vscale), 659 rkvdec->regs + RKVDEC_VP9_REF_SCALE(i)); 660 } 661 662 addr = vb2_dma_contig_plane_dma_addr(&dst->base.vb.vb2_buf, 0); 663 writel_relaxed(addr, rkvdec->regs + RKVDEC_REG_DECOUT_BASE); 664 addr = vb2_dma_contig_plane_dma_addr(&run->base.bufs.src->vb2_buf, 0); 665 writel_relaxed(addr, rkvdec->regs + RKVDEC_REG_STRM_RLC_BASE); 666 writel_relaxed(vp9_ctx->priv_tbl.dma + 667 offsetof(struct rkvdec_vp9_priv_tbl, probs), 668 rkvdec->regs + RKVDEC_REG_CABACTBL_PROB_BASE); 669 writel_relaxed(vp9_ctx->count_tbl.dma, 670 rkvdec->regs + RKVDEC_REG_VP9COUNT_BASE); 671 672 writel_relaxed(vp9_ctx->priv_tbl.dma + 673 offsetof(struct rkvdec_vp9_priv_tbl, segmap) + 674 (RKVDEC_VP9_MAX_SEGMAP_SIZE * vp9_ctx->cur.segmapid), 675 rkvdec->regs + RKVDEC_REG_VP9_SEGIDCUR_BASE); 676 writel_relaxed(vp9_ctx->priv_tbl.dma + 677 offsetof(struct rkvdec_vp9_priv_tbl, segmap) + 678 (RKVDEC_VP9_MAX_SEGMAP_SIZE * (!vp9_ctx->cur.segmapid)), 679 rkvdec->regs + RKVDEC_REG_VP9_SEGIDLAST_BASE); 680 681 if (!intra_only && 682 !(dec_params->flags & V4L2_VP9_FRAME_FLAG_ERROR_RESILIENT) && 683 vp9_ctx->last.valid) 684 mv_ref = last; 685 else 686 mv_ref = dst; 687 688 writel_relaxed(get_mv_base_addr(mv_ref), 689 rkvdec->regs + RKVDEC_VP9_REF_COLMV_BASE); 690 691 writel_relaxed(ctx->decoded_fmt.fmt.pix_mp.width | 692 (ctx->decoded_fmt.fmt.pix_mp.height << 16), 693 rkvdec->regs + RKVDEC_REG_PERFORMANCE_CYCLE); 694} 695 696static int validate_dec_params(struct rkvdec_ctx *ctx, 697 const struct v4l2_ctrl_vp9_frame *dec_params) 698{ 699 unsigned int aligned_width, aligned_height; 700 701 /* We only support profile 0. */ 702 if (dec_params->profile != 0) { 703 dev_err(ctx->dev->dev, "unsupported profile %d\n", 704 dec_params->profile); 705 return -EINVAL; 706 } 707 708 aligned_width = round_up(dec_params->frame_width_minus_1 + 1, 64); 709 aligned_height = round_up(dec_params->frame_height_minus_1 + 1, 64); 710 711 /* 712 * Userspace should update the capture/decoded format when the 713 * resolution changes. 714 */ 715 if (aligned_width != ctx->decoded_fmt.fmt.pix_mp.width || 716 aligned_height != ctx->decoded_fmt.fmt.pix_mp.height) { 717 dev_err(ctx->dev->dev, 718 "unexpected bitstream resolution %dx%d\n", 719 dec_params->frame_width_minus_1 + 1, 720 dec_params->frame_height_minus_1 + 1); 721 return -EINVAL; 722 } 723 724 return 0; 725} 726 727static int rkvdec_vp9_run_preamble(struct rkvdec_ctx *ctx, 728 struct rkvdec_vp9_run *run) 729{ 730 const struct v4l2_ctrl_vp9_frame *dec_params; 731 const struct v4l2_ctrl_vp9_compressed_hdr *prob_updates; 732 struct rkvdec_vp9_ctx *vp9_ctx = ctx->priv; 733 struct v4l2_ctrl *ctrl; 734 unsigned int fctx_idx; 735 int ret; 736 737 /* v4l2-specific stuff */ 738 rkvdec_run_preamble(ctx, &run->base); 739 740 ctrl = v4l2_ctrl_find(&ctx->ctrl_hdl, 741 V4L2_CID_STATELESS_VP9_FRAME); 742 if (WARN_ON(!ctrl)) 743 return -EINVAL; 744 dec_params = ctrl->p_cur.p; 745 746 ret = validate_dec_params(ctx, dec_params); 747 if (ret) 748 return ret; 749 750 run->decode_params = dec_params; 751 752 ctrl = v4l2_ctrl_find(&ctx->ctrl_hdl, V4L2_CID_STATELESS_VP9_COMPRESSED_HDR); 753 if (WARN_ON(!ctrl)) 754 return -EINVAL; 755 prob_updates = ctrl->p_cur.p; 756 vp9_ctx->cur.tx_mode = prob_updates->tx_mode; 757 758 /* 759 * vp9 stuff 760 * 761 * by this point the userspace has done all parts of 6.2 uncompressed_header() 762 * except this fragment: 763 * if ( FrameIsIntra || error_resilient_mode ) { 764 * setup_past_independence ( ) 765 * if ( frame_type == KEY_FRAME || error_resilient_mode == 1 || 766 * reset_frame_context == 3 ) { 767 * for ( i = 0; i < 4; i ++ ) { 768 * save_probs( i ) 769 * } 770 * } else if ( reset_frame_context == 2 ) { 771 * save_probs( frame_context_idx ) 772 * } 773 * frame_context_idx = 0 774 * } 775 */ 776 fctx_idx = v4l2_vp9_reset_frame_ctx(dec_params, vp9_ctx->frame_context); 777 vp9_ctx->cur.frame_context_idx = fctx_idx; 778 779 /* 6.1 frame(sz): load_probs() and load_probs2() */ 780 vp9_ctx->probability_tables = vp9_ctx->frame_context[fctx_idx]; 781 782 /* 783 * The userspace has also performed 6.3 compressed_header(), but handling the 784 * probs in a special way. All probs which need updating, except MV-related, 785 * have been read from the bitstream and translated through inv_map_table[], 786 * but no 6.3.6 inv_recenter_nonneg(v, m) has been performed. The values passed 787 * by userspace are either translated values (there are no 0 values in 788 * inv_map_table[]), or zero to indicate no update. All MV-related probs which need 789 * updating have been read from the bitstream and (mv_prob << 1) | 1 has been 790 * performed. The values passed by userspace are either new values 791 * to replace old ones (the above mentioned shift and bitwise or never result in 792 * a zero) or zero to indicate no update. 793 * fw_update_probs() performs actual probs updates or leaves probs as-is 794 * for values for which a zero was passed from userspace. 795 */ 796 v4l2_vp9_fw_update_probs(&vp9_ctx->probability_tables, prob_updates, dec_params); 797 798 return 0; 799} 800 801static int rkvdec_vp9_run(struct rkvdec_ctx *ctx) 802{ 803 struct rkvdec_dev *rkvdec = ctx->dev; 804 struct rkvdec_vp9_run run = { }; 805 int ret; 806 807 ret = rkvdec_vp9_run_preamble(ctx, &run); 808 if (ret) { 809 rkvdec_run_postamble(ctx, &run.base); 810 return ret; 811 } 812 813 /* Prepare probs. */ 814 init_probs(ctx, &run); 815 816 /* Configure hardware registers. */ 817 config_registers(ctx, &run); 818 819 rkvdec_run_postamble(ctx, &run.base); 820 821 schedule_delayed_work(&rkvdec->watchdog_work, msecs_to_jiffies(2000)); 822 823 writel(1, rkvdec->regs + RKVDEC_REG_PREF_LUMA_CACHE_COMMAND); 824 writel(1, rkvdec->regs + RKVDEC_REG_PREF_CHR_CACHE_COMMAND); 825 826 writel(0xe, rkvdec->regs + RKVDEC_REG_STRMD_ERR_EN); 827 /* Start decoding! */ 828 writel(RKVDEC_INTERRUPT_DEC_E | RKVDEC_CONFIG_DEC_CLK_GATE_E | 829 RKVDEC_TIMEOUT_E | RKVDEC_BUF_EMPTY_E, 830 rkvdec->regs + RKVDEC_REG_INTERRUPT); 831 832 return 0; 833} 834 835#define copy_tx_and_skip(p1, p2) \ 836do { \ 837 memcpy((p1)->tx8, (p2)->tx8, sizeof((p1)->tx8)); \ 838 memcpy((p1)->tx16, (p2)->tx16, sizeof((p1)->tx16)); \ 839 memcpy((p1)->tx32, (p2)->tx32, sizeof((p1)->tx32)); \ 840 memcpy((p1)->skip, (p2)->skip, sizeof((p1)->skip)); \ 841} while (0) 842 843static void rkvdec_vp9_done(struct rkvdec_ctx *ctx, 844 struct vb2_v4l2_buffer *src_buf, 845 struct vb2_v4l2_buffer *dst_buf, 846 enum vb2_buffer_state result) 847{ 848 struct rkvdec_vp9_ctx *vp9_ctx = ctx->priv; 849 unsigned int fctx_idx; 850 851 /* v4l2-specific stuff */ 852 if (result == VB2_BUF_STATE_ERROR) 853 goto out_update_last; 854 855 /* 856 * vp9 stuff 857 * 858 * 6.1.2 refresh_probs() 859 * 860 * In the spec a complementary condition goes last in 6.1.2 refresh_probs(), 861 * but it makes no sense to perform all the activities from the first "if" 862 * there if we actually are not refreshing the frame context. On top of that, 863 * because of 6.2 uncompressed_header() whenever error_resilient_mode == 1, 864 * refresh_frame_context == 0. Consequently, if we don't jump to out_update_last 865 * it means error_resilient_mode must be 0. 866 */ 867 if (!(vp9_ctx->cur.flags & V4L2_VP9_FRAME_FLAG_REFRESH_FRAME_CTX)) 868 goto out_update_last; 869 870 fctx_idx = vp9_ctx->cur.frame_context_idx; 871 872 if (!(vp9_ctx->cur.flags & V4L2_VP9_FRAME_FLAG_PARALLEL_DEC_MODE)) { 873 /* error_resilient_mode == 0 && frame_parallel_decoding_mode == 0 */ 874 struct v4l2_vp9_frame_context *probs = &vp9_ctx->probability_tables; 875 bool frame_is_intra = vp9_ctx->cur.flags & 876 (V4L2_VP9_FRAME_FLAG_KEY_FRAME | V4L2_VP9_FRAME_FLAG_INTRA_ONLY); 877 struct tx_and_skip { 878 u8 tx8[2][1]; 879 u8 tx16[2][2]; 880 u8 tx32[2][3]; 881 u8 skip[3]; 882 } _tx_skip, *tx_skip = &_tx_skip; 883 struct v4l2_vp9_frame_symbol_counts *counts; 884 885 /* buffer the forward-updated TX and skip probs */ 886 if (frame_is_intra) 887 copy_tx_and_skip(tx_skip, probs); 888 889 /* 6.1.2 refresh_probs(): load_probs() and load_probs2() */ 890 *probs = vp9_ctx->frame_context[fctx_idx]; 891 892 /* if FrameIsIntra then undo the effect of load_probs2() */ 893 if (frame_is_intra) 894 copy_tx_and_skip(probs, tx_skip); 895 896 counts = frame_is_intra ? &vp9_ctx->intra_cnts : &vp9_ctx->inter_cnts; 897 v4l2_vp9_adapt_coef_probs(probs, counts, 898 !vp9_ctx->last.valid || 899 vp9_ctx->last.flags & V4L2_VP9_FRAME_FLAG_KEY_FRAME, 900 frame_is_intra); 901 if (!frame_is_intra) { 902 const struct rkvdec_vp9_inter_frame_symbol_counts *inter_cnts; 903 u32 classes[2][11]; 904 int i; 905 906 inter_cnts = vp9_ctx->count_tbl.cpu; 907 for (i = 0; i < ARRAY_SIZE(classes); ++i) 908 memcpy(classes[i], inter_cnts->classes[i], sizeof(classes[0])); 909 counts->classes = &classes; 910 911 /* load_probs2() already done */ 912 v4l2_vp9_adapt_noncoef_probs(&vp9_ctx->probability_tables, counts, 913 vp9_ctx->cur.reference_mode, 914 vp9_ctx->cur.interpolation_filter, 915 vp9_ctx->cur.tx_mode, vp9_ctx->cur.flags); 916 } 917 } 918 919 /* 6.1.2 refresh_probs(): save_probs(fctx_idx) */ 920 vp9_ctx->frame_context[fctx_idx] = vp9_ctx->probability_tables; 921 922out_update_last: 923 update_ctx_last_info(vp9_ctx); 924} 925 926static void rkvdec_init_v4l2_vp9_count_tbl(struct rkvdec_ctx *ctx) 927{ 928 struct rkvdec_vp9_ctx *vp9_ctx = ctx->priv; 929 struct rkvdec_vp9_intra_frame_symbol_counts *intra_cnts = vp9_ctx->count_tbl.cpu; 930 struct rkvdec_vp9_inter_frame_symbol_counts *inter_cnts = vp9_ctx->count_tbl.cpu; 931 int i, j, k, l, m; 932 933 vp9_ctx->inter_cnts.partition = &inter_cnts->partition; 934 vp9_ctx->inter_cnts.skip = &inter_cnts->skip; 935 vp9_ctx->inter_cnts.intra_inter = &inter_cnts->inter; 936 vp9_ctx->inter_cnts.tx32p = &inter_cnts->tx32p; 937 vp9_ctx->inter_cnts.tx16p = &inter_cnts->tx16p; 938 vp9_ctx->inter_cnts.tx8p = &inter_cnts->tx8p; 939 940 vp9_ctx->intra_cnts.partition = (u32 (*)[16][4])(&intra_cnts->partition); 941 vp9_ctx->intra_cnts.skip = &intra_cnts->skip; 942 vp9_ctx->intra_cnts.intra_inter = &intra_cnts->intra; 943 vp9_ctx->intra_cnts.tx32p = &intra_cnts->tx32p; 944 vp9_ctx->intra_cnts.tx16p = &intra_cnts->tx16p; 945 vp9_ctx->intra_cnts.tx8p = &intra_cnts->tx8p; 946 947 vp9_ctx->inter_cnts.y_mode = &inter_cnts->y_mode; 948 vp9_ctx->inter_cnts.uv_mode = &inter_cnts->uv_mode; 949 vp9_ctx->inter_cnts.comp = &inter_cnts->comp; 950 vp9_ctx->inter_cnts.comp_ref = &inter_cnts->comp_ref; 951 vp9_ctx->inter_cnts.single_ref = &inter_cnts->single_ref; 952 vp9_ctx->inter_cnts.mv_mode = &inter_cnts->mv_mode; 953 vp9_ctx->inter_cnts.filter = &inter_cnts->filter; 954 vp9_ctx->inter_cnts.mv_joint = &inter_cnts->mv_joint; 955 vp9_ctx->inter_cnts.sign = &inter_cnts->sign; 956 /* 957 * rk hardware actually uses "u32 classes[2][11 + 1];" 958 * instead of "u32 classes[2][11];", so this must be explicitly 959 * copied into vp9_ctx->classes when passing the data to the 960 * vp9 library function 961 */ 962 vp9_ctx->inter_cnts.class0 = &inter_cnts->class0; 963 vp9_ctx->inter_cnts.bits = &inter_cnts->bits; 964 vp9_ctx->inter_cnts.class0_fp = &inter_cnts->class0_fp; 965 vp9_ctx->inter_cnts.fp = &inter_cnts->fp; 966 vp9_ctx->inter_cnts.class0_hp = &inter_cnts->class0_hp; 967 vp9_ctx->inter_cnts.hp = &inter_cnts->hp; 968 969#define INNERMOST_LOOP \ 970 do { \ 971 for (m = 0; m < ARRAY_SIZE(vp9_ctx->inter_cnts.coeff[0][0][0][0]); ++m) {\ 972 vp9_ctx->inter_cnts.coeff[i][j][k][l][m] = \ 973 &inter_cnts->ref_cnt[k][i][j][l][m].coeff; \ 974 vp9_ctx->inter_cnts.eob[i][j][k][l][m][0] = \ 975 &inter_cnts->ref_cnt[k][i][j][l][m].eob[0]; \ 976 vp9_ctx->inter_cnts.eob[i][j][k][l][m][1] = \ 977 &inter_cnts->ref_cnt[k][i][j][l][m].eob[1]; \ 978 \ 979 vp9_ctx->intra_cnts.coeff[i][j][k][l][m] = \ 980 &intra_cnts->ref_cnt[k][i][j][l][m].coeff; \ 981 vp9_ctx->intra_cnts.eob[i][j][k][l][m][0] = \ 982 &intra_cnts->ref_cnt[k][i][j][l][m].eob[0]; \ 983 vp9_ctx->intra_cnts.eob[i][j][k][l][m][1] = \ 984 &intra_cnts->ref_cnt[k][i][j][l][m].eob[1]; \ 985 } \ 986 } while (0) 987 988 for (i = 0; i < ARRAY_SIZE(vp9_ctx->inter_cnts.coeff); ++i) 989 for (j = 0; j < ARRAY_SIZE(vp9_ctx->inter_cnts.coeff[0]); ++j) 990 for (k = 0; k < ARRAY_SIZE(vp9_ctx->inter_cnts.coeff[0][0]); ++k) 991 for (l = 0; l < ARRAY_SIZE(vp9_ctx->inter_cnts.coeff[0][0][0]); ++l) 992 INNERMOST_LOOP; 993#undef INNERMOST_LOOP 994} 995 996static int rkvdec_vp9_start(struct rkvdec_ctx *ctx) 997{ 998 struct rkvdec_dev *rkvdec = ctx->dev; 999 struct rkvdec_vp9_priv_tbl *priv_tbl; 1000 struct rkvdec_vp9_ctx *vp9_ctx; 1001 unsigned char *count_tbl; 1002 int ret; 1003 1004 vp9_ctx = kzalloc(sizeof(*vp9_ctx), GFP_KERNEL); 1005 if (!vp9_ctx) 1006 return -ENOMEM; 1007 1008 ctx->priv = vp9_ctx; 1009 1010 BUILD_BUG_ON(sizeof(priv_tbl->probs) % 16); /* ensure probs size is 128-bit aligned */ 1011 priv_tbl = dma_alloc_coherent(rkvdec->dev, sizeof(*priv_tbl), 1012 &vp9_ctx->priv_tbl.dma, GFP_KERNEL); 1013 if (!priv_tbl) { 1014 ret = -ENOMEM; 1015 goto err_free_ctx; 1016 } 1017 1018 vp9_ctx->priv_tbl.size = sizeof(*priv_tbl); 1019 vp9_ctx->priv_tbl.cpu = priv_tbl; 1020 1021 count_tbl = dma_alloc_coherent(rkvdec->dev, RKVDEC_VP9_COUNT_SIZE, 1022 &vp9_ctx->count_tbl.dma, GFP_KERNEL); 1023 if (!count_tbl) { 1024 ret = -ENOMEM; 1025 goto err_free_priv_tbl; 1026 } 1027 1028 vp9_ctx->count_tbl.size = RKVDEC_VP9_COUNT_SIZE; 1029 vp9_ctx->count_tbl.cpu = count_tbl; 1030 rkvdec_init_v4l2_vp9_count_tbl(ctx); 1031 1032 return 0; 1033 1034err_free_priv_tbl: 1035 dma_free_coherent(rkvdec->dev, vp9_ctx->priv_tbl.size, 1036 vp9_ctx->priv_tbl.cpu, vp9_ctx->priv_tbl.dma); 1037 1038err_free_ctx: 1039 kfree(vp9_ctx); 1040 return ret; 1041} 1042 1043static void rkvdec_vp9_stop(struct rkvdec_ctx *ctx) 1044{ 1045 struct rkvdec_vp9_ctx *vp9_ctx = ctx->priv; 1046 struct rkvdec_dev *rkvdec = ctx->dev; 1047 1048 dma_free_coherent(rkvdec->dev, vp9_ctx->count_tbl.size, 1049 vp9_ctx->count_tbl.cpu, vp9_ctx->count_tbl.dma); 1050 dma_free_coherent(rkvdec->dev, vp9_ctx->priv_tbl.size, 1051 vp9_ctx->priv_tbl.cpu, vp9_ctx->priv_tbl.dma); 1052 kfree(vp9_ctx); 1053} 1054 1055static int rkvdec_vp9_adjust_fmt(struct rkvdec_ctx *ctx, 1056 struct v4l2_format *f) 1057{ 1058 struct v4l2_pix_format_mplane *fmt = &f->fmt.pix_mp; 1059 1060 fmt->num_planes = 1; 1061 if (!fmt->plane_fmt[0].sizeimage) 1062 fmt->plane_fmt[0].sizeimage = fmt->width * fmt->height * 2; 1063 return 0; 1064} 1065 1066const struct rkvdec_coded_fmt_ops rkvdec_vp9_fmt_ops = { 1067 .adjust_fmt = rkvdec_vp9_adjust_fmt, 1068 .start = rkvdec_vp9_start, 1069 .stop = rkvdec_vp9_stop, 1070 .run = rkvdec_vp9_run, 1071 .done = rkvdec_vp9_done, 1072}; 1073