/netgear-R7000-V1.0.7.12_1.2.5/ap/gpl/iserver/libav-0.8.8/libavcodec/ |
H A D | h264.h | 821 int16_t (*mv_dst)[2] = &s->current_picture.f.motion_val[list][b_xy]; 841 int8_t *ref_index = &s->current_picture.f.ref_index[list][b8_xy]; 859 fill_rectangle(&s->current_picture.f.ref_index[0][b8_xy],
|
H A D | mpeg12.c | 749 s->current_picture.f.mb_type[s->mb_x + s->mb_y * s->mb_stride] = MB_TYPE_SKIP | MB_TYPE_L0 | MB_TYPE_16x16; 754 mb_type = s->current_picture.f.mb_type[s->mb_x + s->mb_y * s->mb_stride - 1]; 756 mb_type = s->current_picture.f.mb_type[s->mb_width + (s->mb_y - 1) * s->mb_stride - 1]; // FIXME not sure if this is allowed in MPEG at all 759 s->current_picture.f.mb_type[s->mb_x + s->mb_y*s->mb_stride] = 761 // assert(s->current_picture.f.mb_type[s->mb_x + s->mb_y * s->mb_stride - 1] & (MB_TYPE_16x16 | MB_TYPE_16x8)); 1104 s->current_picture.f.mb_type[s->mb_x + s->mb_y * s->mb_stride] = mb_type; 1345 s->current_picture.f.pict_type = s->pict_type; 1346 s->current_picture.f.key_frame = s->pict_type == AV_PICTURE_TYPE_I; 1500 s->current_picture.f.pict_type = s->pict_type; 1501 s->current_picture [all...] |
H A D | h263dec.c | 600 s->current_picture.f.pict_type = s->pict_type; 601 s->current_picture.f.key_frame = s->pict_type == AV_PICTURE_TYPE_I; 645 //the second part of the wmv2 header contains the MB skip bits which are stored in current_picture->mb_type 724 assert(s->current_picture.f.pict_type == s->current_picture_ptr->f.pict_type); 725 assert(s->current_picture.f.pict_type == s->pict_type);
|
H A D | ljpegenc.c | 46 AVFrame * const p= (AVFrame*)&s->current_picture;
|
H A D | ratecontrol.c | 48 s->current_picture.f.quality, s->i_tex_bits, s->p_tex_bits, s->mv_bits, s->misc_bits, 49 s->f_code, s->b_code, s->current_picture.mc_mb_var_sum, s->current_picture.mb_var_sum, s->i_count, s->skip_count, s->header_bits); 544 Picture * const pic= &s->current_picture; 673 Picture * const pic= &s->current_picture;
|
H A D | h264_cavlc.c | 779 s->current_picture.f.qscale_table[mb_xy] = 0; 783 s->current_picture.f.mb_type[mb_xy] = mb_type; 1080 s->current_picture.f.mb_type[mb_xy] = mb_type; 1180 s->current_picture.f.qscale_table[mb_xy] = s->qscale;
|
H A D | ituh263enc.c | 278 int8_t * const qscale_table = s->current_picture.f.qscale_table; 532 motion_x = s->current_picture.f.motion_val[0][s->block_index[i]][0]; 533 motion_y = s->current_picture.f.motion_val[0][s->block_index[i]][1];
|
H A D | snow.h | 114 AVFrame current_picture; member in struct:SnowContext 410 int ref_stride= s->current_picture.linesize[plane_index]; 411 uint8_t *dst8= s->current_picture.data[plane_index];
|
H A D | mpeg4videoenc.c | 129 int8_t * const qscale_table = s->current_picture.f.qscale_table; 206 int8_t * const qscale_table = s->current_picture.f.qscale_table; 750 ff_h263_encode_motion_vector(s, s->current_picture.f.motion_val[0][ s->block_index[i] ][0] - pred_x, 751 s->current_picture.f.motion_val[0][ s->block_index[i] ][1] - pred_y, s->f_code);
|
H A D | svq1dec.c | 694 current = s->current_picture.f.data[i]; 738 *pict = *(AVFrame*)&s->current_picture;
|
H A D | msmpeg4.c | 779 dest= s->current_picture.f.data[0] + (((n >> 1) + 2*s->mb_y) * 8* wrap ) + ((n & 1) + 2*s->mb_x) * 8; 782 dest= s->current_picture.f.data[n - 3] + (s->mb_y * 8 * wrap) + s->mb_x * 8; 1171 uint32_t * const mb_type_ptr = &s->current_picture.f.mb_type[s->mb_x + s->mb_y*s->mb_stride];
|
H A D | mpegvideo.h | 295 Picture current_picture; ///< buffer to store the decompressed current picture member in struct:MpegEncContext
|
H A D | vdpau.c | 256 render = (struct vdpau_render_state *)s->current_picture.f.data[0];
|
H A D | mpeg12enc.c | 204 if (s->current_picture.f.key_frame) {
|
/netgear-R7000-V1.0.7.12_1.2.5/ap/gpl/minidlna/ffmpeg-2.3.4/libavcodec/ |
H A D | mpeg12dec.c | 742 s->current_picture.mb_type[s->mb_x + s->mb_y * s->mb_stride] = 748 mb_type = s->current_picture.mb_type[s->mb_x + s->mb_y * s->mb_stride - 1]; 751 mb_type = s->current_picture.mb_type[s->mb_width + (s->mb_y - 1) * s->mb_stride - 1]; 756 s->current_picture.mb_type[s->mb_x + s->mb_y * s->mb_stride] = 1110 s->current_picture.mb_type[s->mb_x + s->mb_y * s->mb_stride] = mb_type; 1408 s->current_picture.f->pict_type = s->pict_type; 1409 s->current_picture.f->key_frame = s->pict_type == AV_PICTURE_TYPE_I; 1571 s->current_picture.f->pict_type = s->pict_type; 1572 s->current_picture.f->key_frame = s->pict_type == AV_PICTURE_TYPE_I; 1680 s->current_picture [all...] |
H A D | h263dec.c | 542 s->current_picture.f->pict_type = s->pict_type; 543 s->current_picture.f->key_frame = s->pict_type == AV_PICTURE_TYPE_I; 592 * are stored in current_picture->mb_type which is not available before 651 av_assert1(s->current_picture.f->pict_type == s->current_picture_ptr->f->pict_type); 652 av_assert1(s->current_picture.f->pict_type == s->pict_type);
|
H A D | ratecontrol.c | 54 s->current_picture.f->quality, 61 s->current_picture.mc_mb_var_sum, 62 s->current_picture.mb_var_sum, 629 Picture *const pic = &s->current_picture; 759 Picture * const pic = &s->current_picture;
|
H A D | ituh263enc.c | 267 int8_t * const qscale_table = s->current_picture.qscale_table; 521 motion_x = s->current_picture.motion_val[0][s->block_index[i]][0]; 522 motion_y = s->current_picture.motion_val[0][s->block_index[i]][1];
|
H A D | snow.h | 122 AVFrame *current_picture; member in struct:SnowContext 424 int ref_stride= s->current_picture->linesize[plane_index]; 425 uint8_t *dst8= s->current_picture->data[plane_index];
|
H A D | mpegvideo_motion.c | 256 linesize = s->current_picture.f->linesize[0] << field_based; 257 uvlinesize = s->current_picture.f->linesize[1] << field_based; 667 Picture *cur_frame = &s->current_picture;
|
H A D | mpeg4videoenc.c | 136 int8_t *const qscale_table = s->current_picture.qscale_table; 216 int8_t *const qscale_table = s->current_picture.qscale_table; 808 s->current_picture.motion_val[0][s->block_index[i]][0] - pred_x, 809 s->current_picture.motion_val[0][s->block_index[i]][1] - pred_y,
|
H A D | msmpeg4dec.c | 106 uint32_t * const mb_type_ptr = &s->current_picture.mb_type[s->mb_x + s->mb_y*s->mb_stride]; 197 uint32_t * const mb_type_ptr = &s->current_picture.mb_type[s->mb_x + s->mb_y*s->mb_stride];
|
H A D | mpegvideo.h | 319 Picture current_picture; ///< buffer to store the decompressed current picture member in struct:MpegEncContext
|
H A D | mss2.c | 423 f = s->current_picture.f;
|
H A D | vdpau.c | 336 render = (struct vdpau_render_state *)s->current_picture.f->data[0];
|