/netgear-R7000-V1.0.7.12_1.2.5/ap/gpl/minidlna/ffmpeg-2.3.4/libavcodec/ |
H A D | error_resilience.c | 61 int *linesize = s->cur_pic.f->linesize; 153 int intra_j = IS_INTRA(s->cur_pic.mb_type[mb_index_j]); 166 int intra_j = IS_INTRA(s->cur_pic.mb_type[mb_index_j]); 181 int intra_j = IS_INTRA(s->cur_pic.mb_type[mb_index_j]); 194 int intra_j = IS_INTRA(s->cur_pic.mb_type[mb_index_j]); 211 if (IS_INTER(s->cur_pic.mb_type[mb_index])) 252 int left_intra = IS_INTRA(s->cur_pic.mb_type[( b_x >> is_luma) + (b_y >> is_luma) * s->mb_stride]); 253 int right_intra = IS_INTRA(s->cur_pic.mb_type[((b_x + 1) >> is_luma) + (b_y >> is_luma) * s->mb_stride]); 257 int16_t *left_mv = s->cur_pic.motion_val[0][mvy_stride * b_y + mvx_stride * b_x]; 258 int16_t *right_mv = s->cur_pic [all...] |
H A D | svq3.c | 76 H264Picture *cur_pic; member in struct:__anon3227 322 dest = h->cur_pic.f.data[0] + x + y * h->linesize; 349 dest = h->cur_pic.f.data[i] + (x >> 1) + (y >> 1) * h->uvlinesize; 480 fill_rectangle(h->cur_pic.motion_val[dir][b_xy], 544 h->cur_pic.motion_val[m][b_xy - 1 + i * h->b_stride]); 551 h->cur_pic.motion_val[m][b_xy - h->b_stride], 558 h->cur_pic.motion_val[m][b_xy - h->b_stride + 4]); 566 h->cur_pic.motion_val[m][b_xy - h->b_stride - 1]); 589 memset(h->cur_pic.motion_val[0][b_xy + i * h->b_stride], 597 memset(h->cur_pic [all...] |
H A D | error_resilience.h | 71 ERPicture cur_pic; member in struct:ERContext
|
H A D | h264_mvpred.h | 256 int8_t *ref = h->cur_pic.ref_index[0]; 257 int16_t(*mv)[2] = h->cur_pic.motion_val[0]; 373 const int left_mb_field_flag = IS_INTERLACED(h->cur_pic.mb_type[mb_xy - 1]); 391 topleft_xy += h->mb_stride & (((h->cur_pic.mb_type[top_xy - 1] >> 7) & 1) - 1); 392 topright_xy += h->mb_stride & (((h->cur_pic.mb_type[top_xy + 1] >> 7) & 1) - 1); 393 top_xy += h->mb_stride & (((h->cur_pic.mb_type[top_xy] >> 7) & 1) - 1); 413 h->topleft_type = h->cur_pic.mb_type[topleft_xy]; 414 h->top_type = h->cur_pic.mb_type[top_xy]; 415 h->topright_type = h->cur_pic.mb_type[topright_xy]; 416 h->left_type[LTOP] = h->cur_pic [all...] |
H A D | mpeg_er.c | 50 set_erpic(&er->cur_pic, s->current_picture_ptr);
|
H A D | h264_loopfilter.c | 256 int mb_type = h->cur_pic.mb_type[mb_xy]; 257 int qp = h->cur_pic.qscale_table[mb_xy]; 258 int qp0 = h->cur_pic.qscale_table[mb_xy - 1]; 259 int qp1 = h->cur_pic.qscale_table[h->top_mb_xy]; 497 if (IS_INTRA(mb_type | h->cur_pic.mb_type[mbn_xy])) { 500 if (!CABAC(h) && IS_8x8DCT(h->cur_pic.mb_type[mbn_xy])) { 515 qp = (h->cur_pic.qscale_table[mb_xy] + h->cur_pic.qscale_table[mbn_xy] + 1) >> 1; 519 chroma_qp_avg[0] = (h->chroma_qp[0] + get_chroma_qp(h, 0, h->cur_pic.qscale_table[mbn_xy]) + 1) >> 1; 520 chroma_qp_avg[1] = (h->chroma_qp[1] + get_chroma_qp(h, 1, h->cur_pic [all...] |
H A D | h264_slice.c | 439 dst->cur_pic = src->cur_pic; 581 memset(&h->cur_pic, 0, sizeof(h->cur_pic)); 597 memset(&h->cur_pic, 0, sizeof(h->cur_pic)); 598 av_frame_unref(&h->cur_pic.f); 599 h->cur_pic.tf.f = &h->cur_pic.f; 639 ff_h264_unref_picture(h, &h->cur_pic); [all...] |
H A D | h264_mb_template.c | 48 const int mb_type = h->cur_pic.mb_type[mb_xy]; 60 dest_y = h->cur_pic.f.data[0] + ((mb_x << PIXEL_SHIFT) + mb_y * h->linesize) * 16; 61 dest_cb = h->cur_pic.f.data[1] + (mb_x << PIXEL_SHIFT) * 8 + mb_y * h->uvlinesize * block_h; 62 dest_cr = h->cur_pic.f.data[2] + (mb_x << PIXEL_SHIFT) * 8 + mb_y * h->uvlinesize * block_h; 277 const int mb_type = h->cur_pic.mb_type[mb_xy]; 286 dest[p] = h->cur_pic.f.data[p] +
|
H A D | h264.h | 352 H264Picture cur_pic; member in struct:H264Context 1028 int16_t(*mv_dst)[2] = &h->cur_pic.motion_val[list][b_xy]; 1049 int8_t *ref_index = &h->cur_pic.ref_index[list][b8_xy]; 1067 fill_rectangle(&h->cur_pic.ref_index[0][b8_xy],
|
H A D | h264_cabac.c | 1289 ctx += (h->cur_pic.mb_type[mbb_xy] >> 7) & (h->slice_table[mbb_xy] == h->slice_num); 1333 && MB_FIELD(h) == !!IS_INTERLACED( h->cur_pic.mb_type[mba_xy] ) ) 1339 && IS_INTERLACED( h->cur_pic.mb_type[mbb_xy] ) ) 1349 if( h->slice_table[mba_xy] == h->slice_num && !IS_SKIP(h->cur_pic.mb_type[mba_xy] )) 1351 if( h->slice_table[mbb_xy] == h->slice_num && !IS_SKIP(h->cur_pic.mb_type[mbb_xy] )) 1905 h->cur_pic.mb_type[mb_xy] = MB_TYPE_SKIP; 2024 h->cur_pic.qscale_table[mb_xy] = 0; 2027 h->cur_pic.mb_type[mb_xy] = mb_type; 2328 h->cur_pic.mb_type[mb_xy] = mb_type; 2430 h->cur_pic [all...] |
H A D | rv34.c | 558 Picture *cur_pic = s->current_picture_ptr; local 560 int type = cur_pic->mb_type[mb_pos]; 563 A[0] = cur_pic->motion_val[dir][mv_pos - 1][0]; 564 A[1] = cur_pic->motion_val[dir][mv_pos - 1][1]; 568 B[0] = cur_pic->motion_val[dir][mv_pos - s->b8_stride][0]; 569 B[1] = cur_pic->motion_val[dir][mv_pos - s->b8_stride][1]; 573 C[0] = cur_pic->motion_val[dir][mv_pos - s->b8_stride + 2][0]; 574 C[1] = cur_pic->motion_val[dir][mv_pos - s->b8_stride + 2][1]; 577 C[0] = cur_pic->motion_val[dir][mv_pos - s->b8_stride - 1][0]; 578 C[1] = cur_pic [all...] |
H A D | h264_mb.c | 67 if (ref->tf.progress->data != h->cur_pic.tf.progress->data || 80 if (ref->tf.progress->data != h->cur_pic.tf.progress->data || 98 const int mb_type = h->cur_pic.mb_type[mb_xy]; 812 const int mb_type = h->cur_pic.mb_type[mb_xy];
|
H A D | h264_mc_template.c | 73 const int mb_type = h->cur_pic.mb_type[mb_xy];
|
H A D | h264_picture.c | 200 ff_h264_set_erpic(&h->er.cur_pic, h->cur_pic_ptr);
|
H A D | h264_cavlc.c | 783 h->cur_pic.qscale_table[mb_xy] = 0; 787 h->cur_pic.mb_type[mb_xy] = mb_type; 1087 h->cur_pic.mb_type[mb_xy] = mb_type; 1168 h->cur_pic.qscale_table[mb_xy] = h->qscale;
|
H A D | h264.c | 89 fill_rectangle(&h->cur_pic.ref_index[0][4 * h->mb_xy], 102 AVFrame *cur = &h->cur_pic.f; 480 av_frame_unref(&h->cur_pic.f); 1108 ff_h264_unref_picture(h, &h->cur_pic); 1963 ff_h264_unref_picture(h, &h->cur_pic);
|
/netgear-R7000-V1.0.7.12_1.2.5/ap/gpl/iserver/libav-0.8.8/libavfilter/ |
H A D | vf_scale.c | 266 AVFilterBufferRef *cur_pic = link->cur_buf; local 272 data[0] = cur_pic->data[0] + y * cur_pic->linesize[0]; 274 cur_pic->data[1] : 275 cur_pic->data[1] + (y>>scale->vsub) * cur_pic->linesize[1]; 276 data[2] = cur_pic->data[2] + (y>>scale->vsub) * cur_pic->linesize[2]; 277 data[3] = cur_pic->data[3] + y * cur_pic [all...] |
/netgear-R7000-V1.0.7.12_1.2.5/ap/gpl/iserver/libav-0.8.8/libavcodec/ |
H A D | rv34.c | 561 Picture *cur_pic = s->current_picture_ptr; local 563 int type = cur_pic->f.mb_type[mb_pos]; 569 A[0] = cur_pic->f.motion_val[dir][mv_pos - 1][0]; 570 A[1] = cur_pic->f.motion_val[dir][mv_pos - 1][1]; 574 B[0] = cur_pic->f.motion_val[dir][mv_pos - s->b8_stride][0]; 575 B[1] = cur_pic->f.motion_val[dir][mv_pos - s->b8_stride][1]; 579 C[0] = cur_pic->f.motion_val[dir][mv_pos - s->b8_stride + 2][0]; 580 C[1] = cur_pic->f.motion_val[dir][mv_pos - s->b8_stride + 2][1]; 583 C[0] = cur_pic->f.motion_val[dir][mv_pos - s->b8_stride - 1][0]; 584 C[1] = cur_pic [all...] |
/netgear-R7000-V1.0.7.12_1.2.5/ap/gpl/minidlna/ffmpeg-2.3.4/libavfilter/ |
H A D | vf_scale.c | 409 static int scale_slice(AVFilterLink *link, AVFrame *out_buf, AVFrame *cur_pic, struct SwsContext *sws, int y, int h, int mul, int field) argument 419 in_stride[i] = cur_pic->linesize[i] * mul; 421 in[i] = cur_pic->data[i] + ((y>>vsub)+field) * cur_pic->linesize[i]; 425 in[1] = cur_pic->data[1];
|