/netgear-R7000-V1.0.7.12_1.2.5/ap/gpl/minidlna/ffmpeg-2.3.4/libavcodec/ |
H A D | dsputil.c | 116 #define avg2(a, b) ((a + b + 1) >> 1) macro 153 s += abs(pix1[0] - avg2(pix2[0], pix2[1])); 154 s += abs(pix1[1] - avg2(pix2[1], pix2[2])); 155 s += abs(pix1[2] - avg2(pix2[2], pix2[3])); 156 s += abs(pix1[3] - avg2(pix2[3], pix2[4])); 157 s += abs(pix1[4] - avg2(pix2[4], pix2[5])); 158 s += abs(pix1[5] - avg2(pix2[5], pix2[6])); 159 s += abs(pix1[6] - avg2(pix2[6], pix2[7])); 160 s += abs(pix1[7] - avg2(pix2[7], pix2[8])); 161 s += abs(pix1[8] - avg2(pix [all...] |
H A D | vc1dsp.c | 831 #define avg2(a, b) (((a) + (b) + 1) >> 1) macro 845 dst[0] = avg2(dst[0], chroma_mc(0)); 846 dst[1] = avg2(dst[1], chroma_mc(1)); 847 dst[2] = avg2(dst[2], chroma_mc(2)); 848 dst[3] = avg2(dst[3], chroma_mc(3)); 849 dst[4] = avg2(dst[4], chroma_mc(4)); 850 dst[5] = avg2(dst[5], chroma_mc(5)); 851 dst[6] = avg2(dst[6], chroma_mc(6)); 852 dst[7] = avg2(dst[7], chroma_mc(7)); 871 dst[0] = avg2(ds [all...] |
/netgear-R7000-V1.0.7.12_1.2.5/ap/gpl/minidlna/ffmpeg-2.3.4/libavcodec/alpha/ |
H A D | motion_est_alpha.c | 25 static inline uint64_t avg2(uint64_t a, uint64_t b) function 134 p2_l = avg2(l, (l >> 8) | ((uint64_t) r << 56)); 135 p2_r = avg2(r, (r >> 8) | ((uint64_t) pix2[16] << 56)); 157 p2_l = avg2(extql(l, disalign) | extqh(m, disalign), m); 158 p2_r = avg2(extql(m, disalign) | extqh(r, disalign), r); 177 p2_l = avg2(extql(l, disalign) | extqh(m, disalign), 179 p2_r = avg2(extql(m, disalign) | extqh(r, disalign), 213 result += perr(p1_l, avg2(p2_l, np2_l)) 214 + perr(p1_r, avg2(p2_r, np2_r)); 234 result += perr(p1_l, avg2(p2_ [all...] |
H A D | hpeldsp_alpha.c | 32 static inline uint64_t avg2(uint64_t a, uint64_t b) function 145 #define AVG2 avg2
|
/netgear-R7000-V1.0.7.12_1.2.5/ap/gpl/iserver/libav-0.8.8/libavcodec/sh4/ |
H A D | dsputil_align.c | 96 #define OP_C(ofs,sz,avg2) \ 118 #define OP_C0(sz,avg2) \ 132 #define OP_X(ofs,sz,avg2) \ 139 OP(LP(dest+0), avg2(MERGE1(t0,t1,ofs),MERGE2(t0,t1,ofs))); \ 141 OP(LP(dest+4), avg2(MERGE1(t1,t0,ofs),MERGE2(t1,t0,ofs))); \ 144 OP(LP(dest+8), avg2(MERGE1(t0,t1,ofs),MERGE2(t0,t1,ofs))); \ 146 OP(LP(dest+12), avg2(MERGE1(t1,t0,ofs),MERGE2(t1,t0,ofs))); \ 154 #define OP_Y0(sz,avg2) \ 168 OP(LP(dest+0), avg2(t0,t)); t0 = t; \ 170 OP(LP(dest+4), avg2(t [all...] |
H A D | qpel.c | 323 #define avg2(a,b) ((a+b+1)>>1) macro
|
/netgear-R7000-V1.0.7.12_1.2.5/ap/gpl/iserver/libav-0.8.8/libavcodec/alpha/ |
H A D | motion_est_alpha.c | 71 static inline uint64_t avg2(uint64_t a, uint64_t b) function 180 p2_l = avg2(l, (l >> 8) | ((uint64_t) r << 56)); 181 p2_r = avg2(r, (r >> 8) | ((uint64_t) pix2[16] << 56)); 203 p2_l = avg2(extql(l, disalign) | extqh(m, disalign), m); 204 p2_r = avg2(extql(m, disalign) | extqh(r, disalign), r); 223 p2_l = avg2(extql(l, disalign) | extqh(m, disalign), 225 p2_r = avg2(extql(m, disalign) | extqh(r, disalign), 259 result += perr(p1_l, avg2(p2_l, np2_l)) 260 + perr(p1_r, avg2(p2_r, np2_r)); 280 result += perr(p1_l, avg2(p2_ [all...] |
H A D | dsputil_alpha.c | 126 static inline uint64_t avg2(uint64_t a, uint64_t b) function 239 #define AVG2 avg2
|
/netgear-R7000-V1.0.7.12_1.2.5/ap/gpl/iserver/libav-0.8.8/libavcodec/ |
H A D | dsputil.c | 518 #define avg2(a,b) ((a+b+1)>>1) macro 1534 s += abs(pix1[0] - avg2(pix2[0], pix2[1])); 1535 s += abs(pix1[1] - avg2(pix2[1], pix2[2])); 1536 s += abs(pix1[2] - avg2(pix2[2], pix2[3])); 1537 s += abs(pix1[3] - avg2(pix2[3], pix2[4])); 1538 s += abs(pix1[4] - avg2(pix2[4], pix2[5])); 1539 s += abs(pix1[5] - avg2(pix2[5], pix2[6])); 1540 s += abs(pix1[6] - avg2(pix2[6], pix2[7])); 1541 s += abs(pix1[7] - avg2(pix2[7], pix2[8])); 1542 s += abs(pix1[8] - avg2(pix [all...] |
H A D | vc1dsp.c | 698 #define avg2(a,b) ((a+b+1)>>1) macro 710 dst[0] = avg2(dst[0], ((A*src[0] + B*src[1] + C*src[stride+0] + D*src[stride+1] + 32 - 4) >> 6)); 711 dst[1] = avg2(dst[1], ((A*src[1] + B*src[2] + C*src[stride+1] + D*src[stride+2] + 32 - 4) >> 6)); 712 dst[2] = avg2(dst[2], ((A*src[2] + B*src[3] + C*src[stride+2] + D*src[stride+3] + 32 - 4) >> 6)); 713 dst[3] = avg2(dst[3], ((A*src[3] + B*src[4] + C*src[stride+3] + D*src[stride+4] + 32 - 4) >> 6)); 714 dst[4] = avg2(dst[4], ((A*src[4] + B*src[5] + C*src[stride+4] + D*src[stride+5] + 32 - 4) >> 6)); 715 dst[5] = avg2(dst[5], ((A*src[5] + B*src[6] + C*src[stride+5] + D*src[stride+6] + 32 - 4) >> 6)); 716 dst[6] = avg2(dst[6], ((A*src[6] + B*src[7] + C*src[stride+6] + D*src[stride+7] + 32 - 4) >> 6)); 717 dst[7] = avg2(dst[7], ((A*src[7] + B*src[8] + C*src[stride+7] + D*src[stride+8] + 32 - 4) >> 6));
|