1/* 2 * This is optimized for sh, which have post increment addressing (*p++). 3 * Some CPU may be index (p[n]) faster than post increment (*p++). 4 * 5 * copyright (c) 2001-2003 BERO <bero@geocities.co.jp> 6 * 7 * This file is part of FFmpeg. 8 * 9 * FFmpeg is free software; you can redistribute it and/or 10 * modify it under the terms of the GNU Lesser General Public 11 * License as published by the Free Software Foundation; either 12 * version 2.1 of the License, or (at your option) any later version. 13 * 14 * FFmpeg is distributed in the hope that it will be useful, 15 * but WITHOUT ANY WARRANTY; without even the implied warranty of 16 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU 17 * Lesser General Public License for more details. 18 * 19 * You should have received a copy of the GNU Lesser General Public 20 * License along with FFmpeg; if not, write to the Free Software 21 * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA 22 */ 23 24#define PIXOP2(OPNAME, OP) \ 25\ 26static inline void OPNAME ## _pixels4_l2_aligned(uint8_t *dst, const uint8_t *src1, const uint8_t *src2, int dst_stride, int src_stride1, int src_stride2, int h) \ 27{\ 28 do {\ 29 OP(LP(dst ),rnd_avg32(LPC(src1 ),LPC(src2 )) ); \ 30 src1+=src_stride1; \ 31 src2+=src_stride2; \ 32 dst+=dst_stride; \ 33 } while(--h); \ 34}\ 35\ 36static inline void OPNAME ## _pixels4_l2_aligned2(uint8_t *dst, const uint8_t *src1, const uint8_t *src2, int dst_stride, int src_stride1, int src_stride2, int h) \ 37{\ 38 do {\ 39 OP(LP(dst ),rnd_avg32(AV_RN32(src1 ),LPC(src2 )) ); \ 40 src1+=src_stride1; \ 41 src2+=src_stride2; \ 42 dst+=dst_stride; \ 43 } while(--h); \ 44}\ 45\ 46static inline void OPNAME ## _no_rnd_pixels16_l2_aligned2(uint8_t *dst, const uint8_t *src1, const uint8_t *src2, int dst_stride, int src_stride1, int src_stride2, int h) \ 47{\ 48 do {\ 49 OP(LP(dst ),no_rnd_avg32(AV_RN32(src1 ),LPC(src2 )) ); \ 50 OP(LP(dst+4),no_rnd_avg32(AV_RN32(src1+4),LPC(src2+4)) ); \ 51 OP(LP(dst+8),no_rnd_avg32(AV_RN32(src1+8),LPC(src2+8)) ); \ 52 OP(LP(dst+12),no_rnd_avg32(AV_RN32(src1+12),LPC(src2+12)) ); \ 53 src1+=src_stride1; \ 54 src2+=src_stride2; \ 55 dst+=dst_stride; \ 56 } while(--h); \ 57}\ 58\ 59static inline void OPNAME ## _pixels16_l2_aligned2(uint8_t *dst, const uint8_t *src1, const uint8_t *src2, int dst_stride, int src_stride1, int src_stride2, int h) \ 60{\ 61 do {\ 62 OP(LP(dst ),rnd_avg32(AV_RN32(src1 ),LPC(src2 )) ); \ 63 OP(LP(dst+4),rnd_avg32(AV_RN32(src1+4),LPC(src2+4)) ); \ 64 OP(LP(dst+8),rnd_avg32(AV_RN32(src1+8),LPC(src2+8)) ); \ 65 OP(LP(dst+12),rnd_avg32(AV_RN32(src1+12),LPC(src2+12)) ); \ 66 src1+=src_stride1; \ 67 src2+=src_stride2; \ 68 dst+=dst_stride; \ 69 } while(--h); \ 70}\ 71\ 72static inline void OPNAME ## _no_rnd_pixels8_l2_aligned2(uint8_t *dst, const uint8_t *src1, const uint8_t *src2, int dst_stride, int src_stride1, int src_stride2, int h) \ 73{\ 74 do { /* onlye src2 aligned */\ 75 OP(LP(dst ),no_rnd_avg32(AV_RN32(src1 ),LPC(src2 )) ); \ 76 OP(LP(dst+4),no_rnd_avg32(AV_RN32(src1+4),LPC(src2+4)) ); \ 77 src1+=src_stride1; \ 78 src2+=src_stride2; \ 79 dst+=dst_stride; \ 80 } while(--h); \ 81}\ 82\ 83static inline void OPNAME ## _pixels8_l2_aligned2(uint8_t *dst, const uint8_t *src1, const uint8_t *src2, int dst_stride, int src_stride1, int src_stride2, int h) \ 84{\ 85 do {\ 86 OP(LP(dst ),rnd_avg32(AV_RN32(src1 ),LPC(src2 )) ); \ 87 OP(LP(dst+4),rnd_avg32(AV_RN32(src1+4),LPC(src2+4)) ); \ 88 src1+=src_stride1; \ 89 src2+=src_stride2; \ 90 dst+=dst_stride; \ 91 } while(--h); \ 92}\ 93\ 94static inline void OPNAME ## _no_rnd_pixels8_l2_aligned(uint8_t *dst, const uint8_t *src1, const uint8_t *src2, int dst_stride, int src_stride1, int src_stride2, int h) \ 95{\ 96 do {\ 97 OP(LP(dst ),no_rnd_avg32(LPC(src1 ),LPC(src2 )) ); \ 98 OP(LP(dst+4),no_rnd_avg32(LPC(src1+4),LPC(src2+4)) ); \ 99 src1+=src_stride1; \ 100 src2+=src_stride2; \ 101 dst+=dst_stride; \ 102 } while(--h); \ 103}\ 104\ 105static inline void OPNAME ## _pixels8_l2_aligned(uint8_t *dst, const uint8_t *src1, const uint8_t *src2, int dst_stride, int src_stride1, int src_stride2, int h) \ 106{\ 107 do {\ 108 OP(LP(dst ),rnd_avg32(LPC(src1 ),LPC(src2 )) ); \ 109 OP(LP(dst+4),rnd_avg32(LPC(src1+4),LPC(src2+4)) ); \ 110 src1+=src_stride1; \ 111 src2+=src_stride2; \ 112 dst+=dst_stride; \ 113 } while(--h); \ 114}\ 115\ 116static inline void OPNAME ## _no_rnd_pixels16_l2_aligned(uint8_t *dst, const uint8_t *src1, const uint8_t *src2, int dst_stride, int src_stride1, int src_stride2, int h) \ 117{\ 118 do {\ 119 OP(LP(dst ),no_rnd_avg32(LPC(src1 ),LPC(src2 )) ); \ 120 OP(LP(dst+4),no_rnd_avg32(LPC(src1+4),LPC(src2+4)) ); \ 121 OP(LP(dst+8),no_rnd_avg32(LPC(src1+8),LPC(src2+8)) ); \ 122 OP(LP(dst+12),no_rnd_avg32(LPC(src1+12),LPC(src2+12)) ); \ 123 src1+=src_stride1; \ 124 src2+=src_stride2; \ 125 dst+=dst_stride; \ 126 } while(--h); \ 127}\ 128\ 129static inline void OPNAME ## _pixels16_l2_aligned(uint8_t *dst, const uint8_t *src1, const uint8_t *src2, int dst_stride, int src_stride1, int src_stride2, int h) \ 130{\ 131 do {\ 132 OP(LP(dst ),rnd_avg32(LPC(src1 ),LPC(src2 )) ); \ 133 OP(LP(dst+4),rnd_avg32(LPC(src1+4),LPC(src2+4)) ); \ 134 OP(LP(dst+8),rnd_avg32(LPC(src1+8),LPC(src2+8)) ); \ 135 OP(LP(dst+12),rnd_avg32(LPC(src1+12),LPC(src2+12)) ); \ 136 src1+=src_stride1; \ 137 src2+=src_stride2; \ 138 dst+=dst_stride; \ 139 } while(--h); \ 140}\ 141\ 142static inline void OPNAME ## _no_rnd_pixels16_l2_aligned1(uint8_t *dst, const uint8_t *src1, const uint8_t *src2, int dst_stride, int src_stride1, int src_stride2, int h) \ 143{ OPNAME ## _no_rnd_pixels16_l2_aligned2(dst,src2,src1,dst_stride,src_stride2,src_stride1,h); } \ 144\ 145static inline void OPNAME ## _pixels16_l2_aligned1(uint8_t *dst, const uint8_t *src1, const uint8_t *src2, int dst_stride, int src_stride1, int src_stride2, int h) \ 146{ OPNAME ## _pixels16_l2_aligned2(dst,src2,src1,dst_stride,src_stride2,src_stride1,h); } \ 147\ 148static inline void OPNAME ## _no_rnd_pixels8_l2_aligned1(uint8_t *dst, const uint8_t *src1, const uint8_t *src2, int dst_stride, int src_stride1, int src_stride2, int h) \ 149{ OPNAME ## _no_rnd_pixels8_l2_aligned2(dst,src2,src1,dst_stride,src_stride2,src_stride1,h); } \ 150\ 151static inline void OPNAME ## _pixels8_l2_aligned1(uint8_t *dst, const uint8_t *src1, const uint8_t *src2, int dst_stride, int src_stride1, int src_stride2, int h) \ 152{ OPNAME ## _pixels8_l2_aligned2(dst,src2,src1,dst_stride,src_stride2,src_stride1,h); } \ 153\ 154static inline void OPNAME ## _pixels8_l4_aligned(uint8_t *dst, const uint8_t *src1, uint8_t *src2, uint8_t *src3, uint8_t *src4,int dst_stride, int src_stride1, int src_stride2,int src_stride3,int src_stride4, int h){\ 155 do { \ 156 uint32_t a0,a1,a2,a3; \ 157 UNPACK(a0,a1,LPC(src1),LPC(src2)); \ 158 UNPACK(a2,a3,LPC(src3),LPC(src4)); \ 159 OP(LP(dst),rnd_PACK(a0,a1,a2,a3)); \ 160 UNPACK(a0,a1,LPC(src1+4),LPC(src2+4)); \ 161 UNPACK(a2,a3,LPC(src3+4),LPC(src4+4)); \ 162 OP(LP(dst+4),rnd_PACK(a0,a1,a2,a3)); \ 163 src1+=src_stride1;\ 164 src2+=src_stride2;\ 165 src3+=src_stride3;\ 166 src4+=src_stride4;\ 167 dst+=dst_stride;\ 168 } while(--h); \ 169} \ 170\ 171static inline void OPNAME ## _no_rnd_pixels8_l4_aligned(uint8_t *dst, const uint8_t *src1, uint8_t *src2, uint8_t *src3, uint8_t *src4,int dst_stride, int src_stride1, int src_stride2,int src_stride3,int src_stride4, int h){\ 172 do { \ 173 uint32_t a0,a1,a2,a3; \ 174 UNPACK(a0,a1,LPC(src1),LPC(src2)); \ 175 UNPACK(a2,a3,LPC(src3),LPC(src4)); \ 176 OP(LP(dst),no_rnd_PACK(a0,a1,a2,a3)); \ 177 UNPACK(a0,a1,LPC(src1+4),LPC(src2+4)); \ 178 UNPACK(a2,a3,LPC(src3+4),LPC(src4+4)); \ 179 OP(LP(dst+4),no_rnd_PACK(a0,a1,a2,a3)); \ 180 src1+=src_stride1;\ 181 src2+=src_stride2;\ 182 src3+=src_stride3;\ 183 src4+=src_stride4;\ 184 dst+=dst_stride;\ 185 } while(--h); \ 186} \ 187\ 188static inline void OPNAME ## _pixels8_l4_aligned0(uint8_t *dst, const uint8_t *src1, uint8_t *src2, uint8_t *src3, uint8_t *src4,int dst_stride, int src_stride1, int src_stride2,int src_stride3,int src_stride4, int h){\ 189 do { \ 190 uint32_t a0,a1,a2,a3; /* src1 only not aligned */\ 191 UNPACK(a0,a1,AV_RN32(src1),LPC(src2)); \ 192 UNPACK(a2,a3,LPC(src3),LPC(src4)); \ 193 OP(LP(dst),rnd_PACK(a0,a1,a2,a3)); \ 194 UNPACK(a0,a1,AV_RN32(src1+4),LPC(src2+4)); \ 195 UNPACK(a2,a3,LPC(src3+4),LPC(src4+4)); \ 196 OP(LP(dst+4),rnd_PACK(a0,a1,a2,a3)); \ 197 src1+=src_stride1;\ 198 src2+=src_stride2;\ 199 src3+=src_stride3;\ 200 src4+=src_stride4;\ 201 dst+=dst_stride;\ 202 } while(--h); \ 203} \ 204\ 205static inline void OPNAME ## _no_rnd_pixels8_l4_aligned0(uint8_t *dst, const uint8_t *src1, uint8_t *src2, uint8_t *src3, uint8_t *src4,int dst_stride, int src_stride1, int src_stride2,int src_stride3,int src_stride4, int h){\ 206 do { \ 207 uint32_t a0,a1,a2,a3; \ 208 UNPACK(a0,a1,AV_RN32(src1),LPC(src2)); \ 209 UNPACK(a2,a3,LPC(src3),LPC(src4)); \ 210 OP(LP(dst),no_rnd_PACK(a0,a1,a2,a3)); \ 211 UNPACK(a0,a1,AV_RN32(src1+4),LPC(src2+4)); \ 212 UNPACK(a2,a3,LPC(src3+4),LPC(src4+4)); \ 213 OP(LP(dst+4),no_rnd_PACK(a0,a1,a2,a3)); \ 214 src1+=src_stride1;\ 215 src2+=src_stride2;\ 216 src3+=src_stride3;\ 217 src4+=src_stride4;\ 218 dst+=dst_stride;\ 219 } while(--h); \ 220} \ 221\ 222static inline void OPNAME ## _pixels16_l4_aligned(uint8_t *dst, const uint8_t *src1, uint8_t *src2, uint8_t *src3, uint8_t *src4,int dst_stride, int src_stride1, int src_stride2,int src_stride3,int src_stride4, int h){\ 223 do { \ 224 uint32_t a0,a1,a2,a3; \ 225 UNPACK(a0,a1,LPC(src1),LPC(src2)); \ 226 UNPACK(a2,a3,LPC(src3),LPC(src4)); \ 227 OP(LP(dst),rnd_PACK(a0,a1,a2,a3)); \ 228 UNPACK(a0,a1,LPC(src1+4),LPC(src2+4)); \ 229 UNPACK(a2,a3,LPC(src3+4),LPC(src4+4)); \ 230 OP(LP(dst+8),rnd_PACK(a0,a1,a2,a3)); \ 231 UNPACK(a0,a1,LPC(src1+8),LPC(src2+8)); \ 232 UNPACK(a2,a3,LPC(src3+8),LPC(src4+8)); \ 233 OP(LP(dst+8),rnd_PACK(a0,a1,a2,a3)); \ 234 UNPACK(a0,a1,LPC(src1+12),LPC(src2+12)); \ 235 UNPACK(a2,a3,LPC(src3+12),LPC(src4+12)); \ 236 OP(LP(dst+12),rnd_PACK(a0,a1,a2,a3)); \ 237 src1+=src_stride1;\ 238 src2+=src_stride2;\ 239 src3+=src_stride3;\ 240 src4+=src_stride4;\ 241 dst+=dst_stride;\ 242 } while(--h); \ 243} \ 244\ 245static inline void OPNAME ## _no_rnd_pixels16_l4_aligned(uint8_t *dst, const uint8_t *src1, uint8_t *src2, uint8_t *src3, uint8_t *src4,int dst_stride, int src_stride1, int src_stride2,int src_stride3,int src_stride4, int h){\ 246 do { \ 247 uint32_t a0,a1,a2,a3; \ 248 UNPACK(a0,a1,LPC(src1),LPC(src2)); \ 249 UNPACK(a2,a3,LPC(src3),LPC(src4)); \ 250 OP(LP(dst),no_rnd_PACK(a0,a1,a2,a3)); \ 251 UNPACK(a0,a1,LPC(src1+4),LPC(src2+4)); \ 252 UNPACK(a2,a3,LPC(src3+4),LPC(src4+4)); \ 253 OP(LP(dst+4),no_rnd_PACK(a0,a1,a2,a3)); \ 254 UNPACK(a0,a1,LPC(src1+8),LPC(src2+8)); \ 255 UNPACK(a2,a3,LPC(src3+8),LPC(src4+8)); \ 256 OP(LP(dst+8),no_rnd_PACK(a0,a1,a2,a3)); \ 257 UNPACK(a0,a1,LPC(src1+12),LPC(src2+12)); \ 258 UNPACK(a2,a3,LPC(src3+12),LPC(src4+12)); \ 259 OP(LP(dst+12),no_rnd_PACK(a0,a1,a2,a3)); \ 260 src1+=src_stride1;\ 261 src2+=src_stride2;\ 262 src3+=src_stride3;\ 263 src4+=src_stride4;\ 264 dst+=dst_stride;\ 265 } while(--h); \ 266} \ 267\ 268static inline void OPNAME ## _pixels16_l4_aligned0(uint8_t *dst, const uint8_t *src1, uint8_t *src2, uint8_t *src3, uint8_t *src4,int dst_stride, int src_stride1, int src_stride2,int src_stride3,int src_stride4, int h){\ 269 do { /* src1 is unaligned */\ 270 uint32_t a0,a1,a2,a3; \ 271 UNPACK(a0,a1,AV_RN32(src1),LPC(src2)); \ 272 UNPACK(a2,a3,LPC(src3),LPC(src4)); \ 273 OP(LP(dst),rnd_PACK(a0,a1,a2,a3)); \ 274 UNPACK(a0,a1,AV_RN32(src1+4),LPC(src2+4)); \ 275 UNPACK(a2,a3,LPC(src3+4),LPC(src4+4)); \ 276 OP(LP(dst+8),rnd_PACK(a0,a1,a2,a3)); \ 277 UNPACK(a0,a1,AV_RN32(src1+8),LPC(src2+8)); \ 278 UNPACK(a2,a3,LPC(src3+8),LPC(src4+8)); \ 279 OP(LP(dst+8),rnd_PACK(a0,a1,a2,a3)); \ 280 UNPACK(a0,a1,AV_RN32(src1+12),LPC(src2+12)); \ 281 UNPACK(a2,a3,LPC(src3+12),LPC(src4+12)); \ 282 OP(LP(dst+12),rnd_PACK(a0,a1,a2,a3)); \ 283 src1+=src_stride1;\ 284 src2+=src_stride2;\ 285 src3+=src_stride3;\ 286 src4+=src_stride4;\ 287 dst+=dst_stride;\ 288 } while(--h); \ 289} \ 290\ 291static inline void OPNAME ## _no_rnd_pixels16_l4_aligned0(uint8_t *dst, const uint8_t *src1, uint8_t *src2, uint8_t *src3, uint8_t *src4,int dst_stride, int src_stride1, int src_stride2,int src_stride3,int src_stride4, int h){\ 292 do { \ 293 uint32_t a0,a1,a2,a3; \ 294 UNPACK(a0,a1,AV_RN32(src1),LPC(src2)); \ 295 UNPACK(a2,a3,LPC(src3),LPC(src4)); \ 296 OP(LP(dst),no_rnd_PACK(a0,a1,a2,a3)); \ 297 UNPACK(a0,a1,AV_RN32(src1+4),LPC(src2+4)); \ 298 UNPACK(a2,a3,LPC(src3+4),LPC(src4+4)); \ 299 OP(LP(dst+4),no_rnd_PACK(a0,a1,a2,a3)); \ 300 UNPACK(a0,a1,AV_RN32(src1+8),LPC(src2+8)); \ 301 UNPACK(a2,a3,LPC(src3+8),LPC(src4+8)); \ 302 OP(LP(dst+8),no_rnd_PACK(a0,a1,a2,a3)); \ 303 UNPACK(a0,a1,AV_RN32(src1+12),LPC(src2+12)); \ 304 UNPACK(a2,a3,LPC(src3+12),LPC(src4+12)); \ 305 OP(LP(dst+12),no_rnd_PACK(a0,a1,a2,a3)); \ 306 src1+=src_stride1;\ 307 src2+=src_stride2;\ 308 src3+=src_stride3;\ 309 src4+=src_stride4;\ 310 dst+=dst_stride;\ 311 } while(--h); \ 312} \ 313\ 314 315#define op_avg(a, b) a = rnd_avg32(a,b) 316#define op_put(a, b) a = b 317 318PIXOP2(avg, op_avg) 319PIXOP2(put, op_put) 320#undef op_avg 321#undef op_put 322 323#define avg2(a,b) ((a+b+1)>>1) 324#define avg4(a,b,c,d) ((a+b+c+d+2)>>2) 325 326 327static void gmc1_c(uint8_t *dst, uint8_t *src, int stride, int h, int x16, int y16, int rounder) 328{ 329 const int A=(16-x16)*(16-y16); 330 const int B=( x16)*(16-y16); 331 const int C=(16-x16)*( y16); 332 const int D=( x16)*( y16); 333 334 do { 335 int t0,t1,t2,t3; 336 uint8_t *s0 = src; 337 uint8_t *s1 = src+stride; 338 t0 = *s0++; t2 = *s1++; 339 t1 = *s0++; t3 = *s1++; 340 dst[0]= (A*t0 + B*t1 + C*t2 + D*t3 + rounder)>>8; 341 t0 = *s0++; t2 = *s1++; 342 dst[1]= (A*t1 + B*t0 + C*t3 + D*t2 + rounder)>>8; 343 t1 = *s0++; t3 = *s1++; 344 dst[2]= (A*t0 + B*t1 + C*t2 + D*t3 + rounder)>>8; 345 t0 = *s0++; t2 = *s1++; 346 dst[3]= (A*t1 + B*t0 + C*t3 + D*t2 + rounder)>>8; 347 t1 = *s0++; t3 = *s1++; 348 dst[4]= (A*t0 + B*t1 + C*t2 + D*t3 + rounder)>>8; 349 t0 = *s0++; t2 = *s1++; 350 dst[5]= (A*t1 + B*t0 + C*t3 + D*t2 + rounder)>>8; 351 t1 = *s0++; t3 = *s1++; 352 dst[6]= (A*t0 + B*t1 + C*t2 + D*t3 + rounder)>>8; 353 t0 = *s0++; t2 = *s1++; 354 dst[7]= (A*t1 + B*t0 + C*t3 + D*t2 + rounder)>>8; 355 dst+= stride; 356 src+= stride; 357 }while(--h); 358} 359 360static void gmc_c(uint8_t *dst, uint8_t *src, int stride, int h, int ox, int oy, 361 int dxx, int dxy, int dyx, int dyy, int shift, int r, int width, int height) 362{ 363 int y, vx, vy; 364 const int s= 1<<shift; 365 366 width--; 367 height--; 368 369 for(y=0; y<h; y++){ 370 int x; 371 372 vx= ox; 373 vy= oy; 374 for(x=0; x<8; x++){ //XXX FIXME optimize 375 int src_x, src_y, frac_x, frac_y, index; 376 377 src_x= vx>>16; 378 src_y= vy>>16; 379 frac_x= src_x&(s-1); 380 frac_y= src_y&(s-1); 381 src_x>>=shift; 382 src_y>>=shift; 383 384 if((unsigned)src_x < width){ 385 if((unsigned)src_y < height){ 386 index= src_x + src_y*stride; 387 dst[y*stride + x]= ( ( src[index ]*(s-frac_x) 388 + src[index +1]* frac_x )*(s-frac_y) 389 + ( src[index+stride ]*(s-frac_x) 390 + src[index+stride+1]* frac_x )* frac_y 391 + r)>>(shift*2); 392 }else{ 393 index= src_x + av_clip(src_y, 0, height)*stride; 394 dst[y*stride + x]= ( ( src[index ]*(s-frac_x) 395 + src[index +1]* frac_x )*s 396 + r)>>(shift*2); 397 } 398 }else{ 399 if((unsigned)src_y < height){ 400 index= av_clip(src_x, 0, width) + src_y*stride; 401 dst[y*stride + x]= ( ( src[index ]*(s-frac_y) 402 + src[index+stride ]* frac_y )*s 403 + r)>>(shift*2); 404 }else{ 405 index= av_clip(src_x, 0, width) + av_clip(src_y, 0, height)*stride; 406 dst[y*stride + x]= src[index ]; 407 } 408 } 409 410 vx+= dxx; 411 vy+= dyx; 412 } 413 ox += dxy; 414 oy += dyy; 415 } 416} 417#define H264_CHROMA_MC(OPNAME, OP)\ 418static void OPNAME ## h264_chroma_mc2_sh4(uint8_t *dst/*align 8*/, uint8_t *src/*align 1*/, int stride, int h, int x, int y){\ 419 const int A=(8-x)*(8-y);\ 420 const int B=( x)*(8-y);\ 421 const int C=(8-x)*( y);\ 422 const int D=( x)*( y);\ 423 \ 424 assert(x<8 && y<8 && x>=0 && y>=0);\ 425\ 426 do {\ 427 int t0,t1,t2,t3; \ 428 uint8_t *s0 = src; \ 429 uint8_t *s1 = src+stride; \ 430 t0 = *s0++; t2 = *s1++; \ 431 t1 = *s0++; t3 = *s1++; \ 432 OP(dst[0], (A*t0 + B*t1 + C*t2 + D*t3));\ 433 t0 = *s0++; t2 = *s1++; \ 434 OP(dst[1], (A*t1 + B*t0 + C*t3 + D*t2));\ 435 dst+= stride;\ 436 src+= stride;\ 437 }while(--h);\ 438}\ 439\ 440static void OPNAME ## h264_chroma_mc4_sh4(uint8_t *dst/*align 8*/, uint8_t *src/*align 1*/, int stride, int h, int x, int y){\ 441 const int A=(8-x)*(8-y);\ 442 const int B=( x)*(8-y);\ 443 const int C=(8-x)*( y);\ 444 const int D=( x)*( y);\ 445 \ 446 assert(x<8 && y<8 && x>=0 && y>=0);\ 447\ 448 do {\ 449 int t0,t1,t2,t3; \ 450 uint8_t *s0 = src; \ 451 uint8_t *s1 = src+stride; \ 452 t0 = *s0++; t2 = *s1++; \ 453 t1 = *s0++; t3 = *s1++; \ 454 OP(dst[0], (A*t0 + B*t1 + C*t2 + D*t3));\ 455 t0 = *s0++; t2 = *s1++; \ 456 OP(dst[1], (A*t1 + B*t0 + C*t3 + D*t2));\ 457 t1 = *s0++; t3 = *s1++; \ 458 OP(dst[2], (A*t0 + B*t1 + C*t2 + D*t3));\ 459 t0 = *s0++; t2 = *s1++; \ 460 OP(dst[3], (A*t1 + B*t0 + C*t3 + D*t2));\ 461 dst+= stride;\ 462 src+= stride;\ 463 }while(--h);\ 464}\ 465\ 466static void OPNAME ## h264_chroma_mc8_sh4(uint8_t *dst/*align 8*/, uint8_t *src/*align 1*/, int stride, int h, int x, int y){\ 467 const int A=(8-x)*(8-y);\ 468 const int B=( x)*(8-y);\ 469 const int C=(8-x)*( y);\ 470 const int D=( x)*( y);\ 471 \ 472 assert(x<8 && y<8 && x>=0 && y>=0);\ 473\ 474 do {\ 475 int t0,t1,t2,t3; \ 476 uint8_t *s0 = src; \ 477 uint8_t *s1 = src+stride; \ 478 t0 = *s0++; t2 = *s1++; \ 479 t1 = *s0++; t3 = *s1++; \ 480 OP(dst[0], (A*t0 + B*t1 + C*t2 + D*t3));\ 481 t0 = *s0++; t2 = *s1++; \ 482 OP(dst[1], (A*t1 + B*t0 + C*t3 + D*t2));\ 483 t1 = *s0++; t3 = *s1++; \ 484 OP(dst[2], (A*t0 + B*t1 + C*t2 + D*t3));\ 485 t0 = *s0++; t2 = *s1++; \ 486 OP(dst[3], (A*t1 + B*t0 + C*t3 + D*t2));\ 487 t1 = *s0++; t3 = *s1++; \ 488 OP(dst[4], (A*t0 + B*t1 + C*t2 + D*t3));\ 489 t0 = *s0++; t2 = *s1++; \ 490 OP(dst[5], (A*t1 + B*t0 + C*t3 + D*t2));\ 491 t1 = *s0++; t3 = *s1++; \ 492 OP(dst[6], (A*t0 + B*t1 + C*t2 + D*t3));\ 493 t0 = *s0++; t2 = *s1++; \ 494 OP(dst[7], (A*t1 + B*t0 + C*t3 + D*t2));\ 495 dst+= stride;\ 496 src+= stride;\ 497 }while(--h);\ 498} 499 500#define op_avg(a, b) a = (((a)+(((b) + 32)>>6)+1)>>1) 501#define op_put(a, b) a = (((b) + 32)>>6) 502 503H264_CHROMA_MC(put_ , op_put) 504H264_CHROMA_MC(avg_ , op_avg) 505#undef op_avg 506#undef op_put 507 508#define QPEL_MC(r, OPNAME, RND, OP) \ 509static void OPNAME ## mpeg4_qpel8_h_lowpass(uint8_t *dst, uint8_t *src, int dstStride, int srcStride, int h){\ 510 uint8_t *cm = ff_cropTbl + MAX_NEG_CROP;\ 511 do {\ 512 uint8_t *s = src; \ 513 int src0,src1,src2,src3,src4,src5,src6,src7,src8;\ 514 src0= *s++;\ 515 src1= *s++;\ 516 src2= *s++;\ 517 src3= *s++;\ 518 src4= *s++;\ 519 OP(dst[0], (src0+src1)*20 - (src0+src2)*6 + (src1+src3)*3 - (src2+src4));\ 520 src5= *s++;\ 521 OP(dst[1], (src1+src2)*20 - (src0+src3)*6 + (src0+src4)*3 - (src1+src5));\ 522 src6= *s++;\ 523 OP(dst[2], (src2+src3)*20 - (src1+src4)*6 + (src0+src5)*3 - (src0+src6));\ 524 src7= *s++;\ 525 OP(dst[3], (src3+src4)*20 - (src2+src5)*6 + (src1+src6)*3 - (src0+src7));\ 526 src8= *s++;\ 527 OP(dst[4], (src4+src5)*20 - (src3+src6)*6 + (src2+src7)*3 - (src1+src8));\ 528 OP(dst[5], (src5+src6)*20 - (src4+src7)*6 + (src3+src8)*3 - (src2+src8));\ 529 OP(dst[6], (src6+src7)*20 - (src5+src8)*6 + (src4+src8)*3 - (src3+src7));\ 530 OP(dst[7], (src7+src8)*20 - (src6+src8)*6 + (src5+src7)*3 - (src4+src6));\ 531 dst+=dstStride;\ 532 src+=srcStride;\ 533 }while(--h);\ 534}\ 535\ 536static void OPNAME ## mpeg4_qpel8_v_lowpass(uint8_t *dst, uint8_t *src, int dstStride, int srcStride){\ 537 uint8_t *cm = ff_cropTbl + MAX_NEG_CROP;\ 538 int w=8;\ 539 do{\ 540 uint8_t *s = src, *d=dst;\ 541 int src0,src1,src2,src3,src4,src5,src6,src7,src8;\ 542 src0 = *s; s+=srcStride; \ 543 src1 = *s; s+=srcStride; \ 544 src2 = *s; s+=srcStride; \ 545 src3 = *s; s+=srcStride; \ 546 src4 = *s; s+=srcStride; \ 547 OP(*d, (src0+src1)*20 - (src0+src2)*6 + (src1+src3)*3 - (src2+src4));d+=dstStride;\ 548 src5 = *s; s+=srcStride; \ 549 OP(*d, (src1+src2)*20 - (src0+src3)*6 + (src0+src4)*3 - (src1+src5));d+=dstStride;\ 550 src6 = *s; s+=srcStride; \ 551 OP(*d, (src2+src3)*20 - (src1+src4)*6 + (src0+src5)*3 - (src0+src6));d+=dstStride;\ 552 src7 = *s; s+=srcStride; \ 553 OP(*d, (src3+src4)*20 - (src2+src5)*6 + (src1+src6)*3 - (src0+src7));d+=dstStride;\ 554 src8 = *s; \ 555 OP(*d, (src4+src5)*20 - (src3+src6)*6 + (src2+src7)*3 - (src1+src8));d+=dstStride;\ 556 OP(*d, (src5+src6)*20 - (src4+src7)*6 + (src3+src8)*3 - (src2+src8));d+=dstStride;\ 557 OP(*d, (src6+src7)*20 - (src5+src8)*6 + (src4+src8)*3 - (src3+src7));d+=dstStride;\ 558 OP(*d, (src7+src8)*20 - (src6+src8)*6 + (src5+src7)*3 - (src4+src6));\ 559 dst++;\ 560 src++;\ 561 }while(--w);\ 562}\ 563\ 564static void OPNAME ## mpeg4_qpel16_h_lowpass(uint8_t *dst, uint8_t *src, int dstStride, int srcStride, int h){\ 565 uint8_t *cm = ff_cropTbl + MAX_NEG_CROP;\ 566 do {\ 567 uint8_t *s = src;\ 568 int src0,src1,src2,src3,src4,src5,src6,src7,src8;\ 569 int src9,src10,src11,src12,src13,src14,src15,src16;\ 570 src0= *s++;\ 571 src1= *s++;\ 572 src2= *s++;\ 573 src3= *s++;\ 574 src4= *s++;\ 575 OP(dst[ 0], (src0 +src1 )*20 - (src0 +src2 )*6 + (src1 +src3 )*3 - (src2 +src4 ));\ 576 src5= *s++;\ 577 OP(dst[ 1], (src1 +src2 )*20 - (src0 +src3 )*6 + (src0 +src4 )*3 - (src1 +src5 ));\ 578 src6= *s++;\ 579 OP(dst[ 2], (src2 +src3 )*20 - (src1 +src4 )*6 + (src0 +src5 )*3 - (src0 +src6 ));\ 580 src7= *s++;\ 581 OP(dst[ 3], (src3 +src4 )*20 - (src2 +src5 )*6 + (src1 +src6 )*3 - (src0 +src7 ));\ 582 src8= *s++;\ 583 OP(dst[ 4], (src4 +src5 )*20 - (src3 +src6 )*6 + (src2 +src7 )*3 - (src1 +src8 ));\ 584 src9= *s++;\ 585 OP(dst[ 5], (src5 +src6 )*20 - (src4 +src7 )*6 + (src3 +src8 )*3 - (src2 +src9 ));\ 586 src10= *s++;\ 587 OP(dst[ 6], (src6 +src7 )*20 - (src5 +src8 )*6 + (src4 +src9 )*3 - (src3 +src10));\ 588 src11= *s++;\ 589 OP(dst[ 7], (src7 +src8 )*20 - (src6 +src9 )*6 + (src5 +src10)*3 - (src4 +src11));\ 590 src12= *s++;\ 591 OP(dst[ 8], (src8 +src9 )*20 - (src7 +src10)*6 + (src6 +src11)*3 - (src5 +src12));\ 592 src13= *s++;\ 593 OP(dst[ 9], (src9 +src10)*20 - (src8 +src11)*6 + (src7 +src12)*3 - (src6 +src13));\ 594 src14= *s++;\ 595 OP(dst[10], (src10+src11)*20 - (src9 +src12)*6 + (src8 +src13)*3 - (src7 +src14));\ 596 src15= *s++;\ 597 OP(dst[11], (src11+src12)*20 - (src10+src13)*6 + (src9 +src14)*3 - (src8 +src15));\ 598 src16= *s++;\ 599 OP(dst[12], (src12+src13)*20 - (src11+src14)*6 + (src10+src15)*3 - (src9 +src16));\ 600 OP(dst[13], (src13+src14)*20 - (src12+src15)*6 + (src11+src16)*3 - (src10+src16));\ 601 OP(dst[14], (src14+src15)*20 - (src13+src16)*6 + (src12+src16)*3 - (src11+src15));\ 602 OP(dst[15], (src15+src16)*20 - (src14+src16)*6 + (src13+src15)*3 - (src12+src14));\ 603 dst+=dstStride;\ 604 src+=srcStride;\ 605 }while(--h);\ 606}\ 607\ 608static void OPNAME ## mpeg4_qpel16_v_lowpass(uint8_t *dst, uint8_t *src, int dstStride, int srcStride){\ 609 uint8_t *cm = ff_cropTbl + MAX_NEG_CROP;\ 610 int w=16;\ 611 do {\ 612 uint8_t *s = src, *d=dst;\ 613 int src0,src1,src2,src3,src4,src5,src6,src7,src8;\ 614 int src9,src10,src11,src12,src13,src14,src15,src16;\ 615 src0 = *s; s+=srcStride; \ 616 src1 = *s; s+=srcStride; \ 617 src2 = *s; s+=srcStride; \ 618 src3 = *s; s+=srcStride; \ 619 src4 = *s; s+=srcStride; \ 620 OP(*d, (src0 +src1 )*20 - (src0 +src2 )*6 + (src1 +src3 )*3 - (src2 +src4 ));d+=dstStride;\ 621 src5 = *s; s+=srcStride; \ 622 OP(*d, (src1 +src2 )*20 - (src0 +src3 )*6 + (src0 +src4 )*3 - (src1 +src5 ));d+=dstStride;\ 623 src6 = *s; s+=srcStride; \ 624 OP(*d, (src2 +src3 )*20 - (src1 +src4 )*6 + (src0 +src5 )*3 - (src0 +src6 ));d+=dstStride;\ 625 src7 = *s; s+=srcStride; \ 626 OP(*d, (src3 +src4 )*20 - (src2 +src5 )*6 + (src1 +src6 )*3 - (src0 +src7 ));d+=dstStride;\ 627 src8 = *s; s+=srcStride; \ 628 OP(*d, (src4 +src5 )*20 - (src3 +src6 )*6 + (src2 +src7 )*3 - (src1 +src8 ));d+=dstStride;\ 629 src9 = *s; s+=srcStride; \ 630 OP(*d, (src5 +src6 )*20 - (src4 +src7 )*6 + (src3 +src8 )*3 - (src2 +src9 ));d+=dstStride;\ 631 src10 = *s; s+=srcStride; \ 632 OP(*d, (src6 +src7 )*20 - (src5 +src8 )*6 + (src4 +src9 )*3 - (src3 +src10));d+=dstStride;\ 633 src11 = *s; s+=srcStride; \ 634 OP(*d, (src7 +src8 )*20 - (src6 +src9 )*6 + (src5 +src10)*3 - (src4 +src11));d+=dstStride;\ 635 src12 = *s; s+=srcStride; \ 636 OP(*d, (src8 +src9 )*20 - (src7 +src10)*6 + (src6 +src11)*3 - (src5 +src12));d+=dstStride;\ 637 src13 = *s; s+=srcStride; \ 638 OP(*d, (src9 +src10)*20 - (src8 +src11)*6 + (src7 +src12)*3 - (src6 +src13));d+=dstStride;\ 639 src14 = *s; s+=srcStride; \ 640 OP(*d, (src10+src11)*20 - (src9 +src12)*6 + (src8 +src13)*3 - (src7 +src14));d+=dstStride;\ 641 src15 = *s; s+=srcStride; \ 642 OP(*d, (src11+src12)*20 - (src10+src13)*6 + (src9 +src14)*3 - (src8 +src15));d+=dstStride;\ 643 src16 = *s; \ 644 OP(*d, (src12+src13)*20 - (src11+src14)*6 + (src10+src15)*3 - (src9 +src16));d+=dstStride;\ 645 OP(*d, (src13+src14)*20 - (src12+src15)*6 + (src11+src16)*3 - (src10+src16));d+=dstStride;\ 646 OP(*d, (src14+src15)*20 - (src13+src16)*6 + (src12+src16)*3 - (src11+src15));d+=dstStride;\ 647 OP(*d, (src15+src16)*20 - (src14+src16)*6 + (src13+src15)*3 - (src12+src14));\ 648 dst++;\ 649 src++;\ 650 }while(--w);\ 651}\ 652\ 653static void OPNAME ## qpel8_mc00_sh4 (uint8_t *dst, uint8_t *src, int stride){\ 654 OPNAME ## pixels8_c(dst, src, stride, 8);\ 655}\ 656\ 657static void OPNAME ## qpel8_mc10_sh4(uint8_t *dst, uint8_t *src, int stride){\ 658 uint8_t half[64];\ 659 put ## RND ## mpeg4_qpel8_h_lowpass(half, src, 8, stride, 8);\ 660 OPNAME ## pixels8_l2_aligned2(dst, src, half, stride, stride, 8, 8);\ 661}\ 662\ 663static void OPNAME ## qpel8_mc20_sh4(uint8_t *dst, uint8_t *src, int stride){\ 664 OPNAME ## mpeg4_qpel8_h_lowpass(dst, src, stride, stride, 8);\ 665}\ 666\ 667static void OPNAME ## qpel8_mc30_sh4(uint8_t *dst, uint8_t *src, int stride){\ 668 uint8_t half[64];\ 669 put ## RND ## mpeg4_qpel8_h_lowpass(half, src, 8, stride, 8);\ 670 OPNAME ## pixels8_l2_aligned2(dst, src+1, half, stride, stride, 8, 8);\ 671}\ 672\ 673static void OPNAME ## qpel8_mc01_sh4(uint8_t *dst, uint8_t *src, int stride){\ 674 uint8_t full[16*9];\ 675 uint8_t half[64];\ 676 copy_block9(full, src, 16, stride, 9);\ 677 put ## RND ## mpeg4_qpel8_v_lowpass(half, full, 8, 16);\ 678 OPNAME ## pixels8_l2_aligned(dst, full, half, stride, 16, 8, 8);\ 679}\ 680\ 681static void OPNAME ## qpel8_mc02_sh4(uint8_t *dst, uint8_t *src, int stride){\ 682 uint8_t full[16*9];\ 683 copy_block9(full, src, 16, stride, 9);\ 684 OPNAME ## mpeg4_qpel8_v_lowpass(dst, full, stride, 16);\ 685}\ 686\ 687static void OPNAME ## qpel8_mc03_sh4(uint8_t *dst, uint8_t *src, int stride){\ 688 uint8_t full[16*9];\ 689 uint8_t half[64];\ 690 copy_block9(full, src, 16, stride, 9);\ 691 put ## RND ## mpeg4_qpel8_v_lowpass(half, full, 8, 16);\ 692 OPNAME ## pixels8_l2_aligned(dst, full+16, half, stride, 16, 8, 8);\ 693}\ 694static void OPNAME ## qpel8_mc11_sh4(uint8_t *dst, uint8_t *src, int stride){\ 695 uint8_t full[16*9];\ 696 uint8_t halfH[72];\ 697 uint8_t halfHV[64];\ 698 copy_block9(full, src, 16, stride, 9);\ 699 put ## RND ## mpeg4_qpel8_h_lowpass(halfH, full, 8, 16, 9);\ 700 put ## RND ## pixels8_l2_aligned(halfH, halfH, full, 8, 8, 16, 9);\ 701 put ## RND ## mpeg4_qpel8_v_lowpass(halfHV, halfH, 8, 8);\ 702 OPNAME ## pixels8_l2_aligned(dst, halfH, halfHV, stride, 8, 8, 8);\ 703}\ 704static void OPNAME ## qpel8_mc31_sh4(uint8_t *dst, uint8_t *src, int stride){\ 705 uint8_t full[16*9];\ 706 uint8_t halfH[72];\ 707 uint8_t halfHV[64];\ 708 copy_block9(full, src, 16, stride, 9);\ 709 put ## RND ## mpeg4_qpel8_h_lowpass(halfH, full, 8, 16, 9);\ 710 put ## RND ## pixels8_l2_aligned1(halfH, halfH, full+1, 8, 8, 16, 9);\ 711 put ## RND ## mpeg4_qpel8_v_lowpass(halfHV, halfH, 8, 8);\ 712 OPNAME ## pixels8_l2_aligned(dst, halfH, halfHV, stride, 8, 8, 8);\ 713}\ 714static void OPNAME ## qpel8_mc13_sh4(uint8_t *dst, uint8_t *src, int stride){\ 715 uint8_t full[16*9];\ 716 uint8_t halfH[72];\ 717 uint8_t halfHV[64];\ 718 copy_block9(full, src, 16, stride, 9);\ 719 put ## RND ## mpeg4_qpel8_h_lowpass(halfH, full, 8, 16, 9);\ 720 put ## RND ## pixels8_l2_aligned(halfH, halfH, full, 8, 8, 16, 9);\ 721 put ## RND ## mpeg4_qpel8_v_lowpass(halfHV, halfH, 8, 8);\ 722 OPNAME ## pixels8_l2_aligned(dst, halfH+8, halfHV, stride, 8, 8, 8);\ 723}\ 724static void OPNAME ## qpel8_mc33_sh4(uint8_t *dst, uint8_t *src, int stride){\ 725 uint8_t full[16*9];\ 726 uint8_t halfH[72];\ 727 uint8_t halfHV[64];\ 728 copy_block9(full, src, 16, stride, 9);\ 729 put ## RND ## mpeg4_qpel8_h_lowpass(halfH, full, 8, 16, 9);\ 730 put ## RND ## pixels8_l2_aligned1(halfH, halfH, full+1, 8, 8, 16, 9);\ 731 put ## RND ## mpeg4_qpel8_v_lowpass(halfHV, halfH, 8, 8);\ 732 OPNAME ## pixels8_l2_aligned(dst, halfH+8, halfHV, stride, 8, 8, 8);\ 733}\ 734static void OPNAME ## qpel8_mc21_sh4(uint8_t *dst, uint8_t *src, int stride){\ 735 uint8_t halfH[72];\ 736 uint8_t halfHV[64];\ 737 put ## RND ## mpeg4_qpel8_h_lowpass(halfH, src, 8, stride, 9);\ 738 put ## RND ## mpeg4_qpel8_v_lowpass(halfHV, halfH, 8, 8);\ 739 OPNAME ## pixels8_l2_aligned(dst, halfH, halfHV, stride, 8, 8, 8);\ 740}\ 741static void OPNAME ## qpel8_mc23_sh4(uint8_t *dst, uint8_t *src, int stride){\ 742 uint8_t halfH[72];\ 743 uint8_t halfHV[64];\ 744 put ## RND ## mpeg4_qpel8_h_lowpass(halfH, src, 8, stride, 9);\ 745 put ## RND ## mpeg4_qpel8_v_lowpass(halfHV, halfH, 8, 8);\ 746 OPNAME ## pixels8_l2_aligned(dst, halfH+8, halfHV, stride, 8, 8, 8);\ 747}\ 748static void OPNAME ## qpel8_mc12_sh4(uint8_t *dst, uint8_t *src, int stride){\ 749 uint8_t full[16*9];\ 750 uint8_t halfH[72];\ 751 copy_block9(full, src, 16, stride, 9);\ 752 put ## RND ## mpeg4_qpel8_h_lowpass(halfH, full, 8, 16, 9);\ 753 put ## RND ## pixels8_l2_aligned(halfH, halfH, full, 8, 8, 16, 9);\ 754 OPNAME ## mpeg4_qpel8_v_lowpass(dst, halfH, stride, 8);\ 755}\ 756static void OPNAME ## qpel8_mc32_sh4(uint8_t *dst, uint8_t *src, int stride){\ 757 uint8_t full[16*9];\ 758 uint8_t halfH[72];\ 759 copy_block9(full, src, 16, stride, 9);\ 760 put ## RND ## mpeg4_qpel8_h_lowpass(halfH, full, 8, 16, 9);\ 761 put ## RND ## pixels8_l2_aligned1(halfH, halfH, full+1, 8, 8, 16, 9);\ 762 OPNAME ## mpeg4_qpel8_v_lowpass(dst, halfH, stride, 8);\ 763}\ 764static void OPNAME ## qpel8_mc22_sh4(uint8_t *dst, uint8_t *src, int stride){\ 765 uint8_t halfH[72];\ 766 put ## RND ## mpeg4_qpel8_h_lowpass(halfH, src, 8, stride, 9);\ 767 OPNAME ## mpeg4_qpel8_v_lowpass(dst, halfH, stride, 8);\ 768}\ 769static void OPNAME ## qpel16_mc00_sh4 (uint8_t *dst, uint8_t *src, int stride){\ 770 OPNAME ## pixels16_c(dst, src, stride, 16);\ 771}\ 772\ 773static void OPNAME ## qpel16_mc10_sh4(uint8_t *dst, uint8_t *src, int stride){\ 774 uint8_t half[256];\ 775 put ## RND ## mpeg4_qpel16_h_lowpass(half, src, 16, stride, 16);\ 776 OPNAME ## pixels16_l2_aligned2(dst, src, half, stride, stride, 16, 16);\ 777}\ 778\ 779static void OPNAME ## qpel16_mc20_sh4(uint8_t *dst, uint8_t *src, int stride){\ 780 OPNAME ## mpeg4_qpel16_h_lowpass(dst, src, stride, stride, 16);\ 781}\ 782\ 783static void OPNAME ## qpel16_mc30_sh4(uint8_t *dst, uint8_t *src, int stride){\ 784 uint8_t half[256];\ 785 put ## RND ## mpeg4_qpel16_h_lowpass(half, src, 16, stride, 16);\ 786 OPNAME ## pixels16_l2_aligned2(dst, src+1, half, stride, stride, 16, 16);\ 787}\ 788\ 789static void OPNAME ## qpel16_mc01_sh4(uint8_t *dst, uint8_t *src, int stride){\ 790 uint8_t full[24*17];\ 791 uint8_t half[256];\ 792 copy_block17(full, src, 24, stride, 17);\ 793 put ## RND ## mpeg4_qpel16_v_lowpass(half, full, 16, 24);\ 794 OPNAME ## pixels16_l2_aligned(dst, full, half, stride, 24, 16, 16);\ 795}\ 796\ 797static void OPNAME ## qpel16_mc02_sh4(uint8_t *dst, uint8_t *src, int stride){\ 798 uint8_t full[24*17];\ 799 copy_block17(full, src, 24, stride, 17);\ 800 OPNAME ## mpeg4_qpel16_v_lowpass(dst, full, stride, 24);\ 801}\ 802\ 803static void OPNAME ## qpel16_mc03_sh4(uint8_t *dst, uint8_t *src, int stride){\ 804 uint8_t full[24*17];\ 805 uint8_t half[256];\ 806 copy_block17(full, src, 24, stride, 17);\ 807 put ## RND ## mpeg4_qpel16_v_lowpass(half, full, 16, 24);\ 808 OPNAME ## pixels16_l2_aligned(dst, full+24, half, stride, 24, 16, 16);\ 809}\ 810static void OPNAME ## qpel16_mc11_sh4(uint8_t *dst, uint8_t *src, int stride){\ 811 uint8_t full[24*17];\ 812 uint8_t halfH[272];\ 813 uint8_t halfHV[256];\ 814 copy_block17(full, src, 24, stride, 17);\ 815 put ## RND ## mpeg4_qpel16_h_lowpass(halfH, full, 16, 24, 17);\ 816 put ## RND ## pixels16_l2_aligned(halfH, halfH, full, 16, 16, 24, 17);\ 817 put ## RND ## mpeg4_qpel16_v_lowpass(halfHV, halfH, 16, 16);\ 818 OPNAME ## pixels16_l2_aligned(dst, halfH, halfHV, stride, 16, 16, 16);\ 819}\ 820static void OPNAME ## qpel16_mc31_sh4(uint8_t *dst, uint8_t *src, int stride){\ 821 uint8_t full[24*17];\ 822 uint8_t halfH[272];\ 823 uint8_t halfHV[256];\ 824 copy_block17(full, src, 24, stride, 17);\ 825 put ## RND ## mpeg4_qpel16_h_lowpass(halfH, full, 16, 24, 17);\ 826 put ## RND ## pixels16_l2_aligned1(halfH, halfH, full+1, 16, 16, 24, 17);\ 827 put ## RND ## mpeg4_qpel16_v_lowpass(halfHV, halfH, 16, 16);\ 828 OPNAME ## pixels16_l2_aligned(dst, halfH, halfHV, stride, 16, 16, 16);\ 829}\ 830static void OPNAME ## qpel16_mc13_sh4(uint8_t *dst, uint8_t *src, int stride){\ 831 uint8_t full[24*17];\ 832 uint8_t halfH[272];\ 833 uint8_t halfHV[256];\ 834 copy_block17(full, src, 24, stride, 17);\ 835 put ## RND ## mpeg4_qpel16_h_lowpass(halfH, full, 16, 24, 17);\ 836 put ## RND ## pixels16_l2_aligned(halfH, halfH, full, 16, 16, 24, 17);\ 837 put ## RND ## mpeg4_qpel16_v_lowpass(halfHV, halfH, 16, 16);\ 838 OPNAME ## pixels16_l2_aligned(dst, halfH+16, halfHV, stride, 16, 16, 16);\ 839}\ 840static void OPNAME ## qpel16_mc33_sh4(uint8_t *dst, uint8_t *src, int stride){\ 841 uint8_t full[24*17];\ 842 uint8_t halfH[272];\ 843 uint8_t halfHV[256];\ 844 copy_block17(full, src, 24, stride, 17);\ 845 put ## RND ## mpeg4_qpel16_h_lowpass(halfH, full, 16, 24, 17);\ 846 put ## RND ## pixels16_l2_aligned1(halfH, halfH, full+1, 16, 16, 24, 17);\ 847 put ## RND ## mpeg4_qpel16_v_lowpass(halfHV, halfH, 16, 16);\ 848 OPNAME ## pixels16_l2_aligned(dst, halfH+16, halfHV, stride, 16, 16, 16);\ 849}\ 850static void OPNAME ## qpel16_mc21_sh4(uint8_t *dst, uint8_t *src, int stride){\ 851 uint8_t halfH[272];\ 852 uint8_t halfHV[256];\ 853 put ## RND ## mpeg4_qpel16_h_lowpass(halfH, src, 16, stride, 17);\ 854 put ## RND ## mpeg4_qpel16_v_lowpass(halfHV, halfH, 16, 16);\ 855 OPNAME ## pixels16_l2_aligned(dst, halfH, halfHV, stride, 16, 16, 16);\ 856}\ 857static void OPNAME ## qpel16_mc23_sh4(uint8_t *dst, uint8_t *src, int stride){\ 858 uint8_t halfH[272];\ 859 uint8_t halfHV[256];\ 860 put ## RND ## mpeg4_qpel16_h_lowpass(halfH, src, 16, stride, 17);\ 861 put ## RND ## mpeg4_qpel16_v_lowpass(halfHV, halfH, 16, 16);\ 862 OPNAME ## pixels16_l2_aligned(dst, halfH+16, halfHV, stride, 16, 16, 16);\ 863}\ 864static void OPNAME ## qpel16_mc12_sh4(uint8_t *dst, uint8_t *src, int stride){\ 865 uint8_t full[24*17];\ 866 uint8_t halfH[272];\ 867 copy_block17(full, src, 24, stride, 17);\ 868 put ## RND ## mpeg4_qpel16_h_lowpass(halfH, full, 16, 24, 17);\ 869 put ## RND ## pixels16_l2_aligned(halfH, halfH, full, 16, 16, 24, 17);\ 870 OPNAME ## mpeg4_qpel16_v_lowpass(dst, halfH, stride, 16);\ 871}\ 872static void OPNAME ## qpel16_mc32_sh4(uint8_t *dst, uint8_t *src, int stride){\ 873 uint8_t full[24*17];\ 874 uint8_t halfH[272];\ 875 copy_block17(full, src, 24, stride, 17);\ 876 put ## RND ## mpeg4_qpel16_h_lowpass(halfH, full, 16, 24, 17);\ 877 put ## RND ## pixels16_l2_aligned1(halfH, halfH, full+1, 16, 16, 24, 17);\ 878 OPNAME ## mpeg4_qpel16_v_lowpass(dst, halfH, stride, 16);\ 879}\ 880static void OPNAME ## qpel16_mc22_sh4(uint8_t *dst, uint8_t *src, int stride){\ 881 uint8_t halfH[272];\ 882 put ## RND ## mpeg4_qpel16_h_lowpass(halfH, src, 16, stride, 17);\ 883 OPNAME ## mpeg4_qpel16_v_lowpass(dst, halfH, stride, 16);\ 884} 885 886#define op_avg(a, b) a = (((a)+cm[((b) + 16)>>5]+1)>>1) 887#define op_avg_no_rnd(a, b) a = (((a)+cm[((b) + 15)>>5])>>1) 888#define op_put(a, b) a = cm[((b) + 16)>>5] 889#define op_put_no_rnd(a, b) a = cm[((b) + 15)>>5] 890 891QPEL_MC(0, put_ , _ , op_put) 892QPEL_MC(1, put_no_rnd_, _no_rnd_, op_put_no_rnd) 893QPEL_MC(0, avg_ , _ , op_avg) 894//QPEL_MC(1, avg_no_rnd , _ , op_avg) 895#undef op_avg 896#undef op_avg_no_rnd 897#undef op_put 898#undef op_put_no_rnd 899 900#if 1 901#define H264_LOWPASS(OPNAME, OP, OP2) \ 902static inline void OPNAME ## h264_qpel_h_lowpass(uint8_t *dst, uint8_t *src, int dstStride, int srcStride,int w,int h){\ 903 uint8_t *cm = ff_cropTbl + MAX_NEG_CROP;\ 904 do {\ 905 int srcB,srcA,src0,src1,src2,src3,src4,src5,src6;\ 906 uint8_t *s = src-2;\ 907 srcB = *s++;\ 908 srcA = *s++;\ 909 src0 = *s++;\ 910 src1 = *s++;\ 911 src2 = *s++;\ 912 src3 = *s++;\ 913 OP(dst[0], (src0+src1)*20 - (srcA+src2)*5 + (srcB+src3));\ 914 src4 = *s++;\ 915 OP(dst[1], (src1+src2)*20 - (src0+src3)*5 + (srcA+src4));\ 916 src5 = *s++;\ 917 OP(dst[2], (src2+src3)*20 - (src1+src4)*5 + (src0+src5));\ 918 src6 = *s++;\ 919 OP(dst[3], (src3+src4)*20 - (src2+src5)*5 + (src1+src6));\ 920 if (w>4) { /* it optimized */ \ 921 int src7,src8,src9,src10; \ 922 src7 = *s++;\ 923 OP(dst[4], (src4+src5)*20 - (src3+src6)*5 + (src2+src7));\ 924 src8 = *s++;\ 925 OP(dst[5], (src5+src6)*20 - (src4+src7)*5 + (src3+src8));\ 926 src9 = *s++;\ 927 OP(dst[6], (src6+src7)*20 - (src5+src8)*5 + (src4+src9));\ 928 src10 = *s++;\ 929 OP(dst[7], (src7+src8)*20 - (src6+src9)*5 + (src5+src10));\ 930 if (w>8) { \ 931 int src11,src12,src13,src14,src15,src16,src17,src18; \ 932 src11 = *s++;\ 933 OP(dst[8] , (src8 +src9 )*20 - (src7 +src10)*5 + (src6 +src11));\ 934 src12 = *s++;\ 935 OP(dst[9] , (src9 +src10)*20 - (src8 +src11)*5 + (src7 +src12));\ 936 src13 = *s++;\ 937 OP(dst[10], (src10+src11)*20 - (src9 +src12)*5 + (src8 +src13));\ 938 src14 = *s++;\ 939 OP(dst[11], (src11+src12)*20 - (src10+src13)*5 + (src9 +src14));\ 940 src15 = *s++;\ 941 OP(dst[12], (src12+src13)*20 - (src11+src14)*5 + (src10+src15));\ 942 src16 = *s++;\ 943 OP(dst[13], (src13+src14)*20 - (src12+src15)*5 + (src11+src16));\ 944 src17 = *s++;\ 945 OP(dst[14], (src14+src15)*20 - (src13+src16)*5 + (src12+src17));\ 946 src18 = *s++;\ 947 OP(dst[15], (src15+src16)*20 - (src14+src17)*5 + (src13+src18));\ 948 } \ 949 } \ 950 dst+=dstStride;\ 951 src+=srcStride;\ 952 }while(--h);\ 953}\ 954\ 955static inline void OPNAME ## h264_qpel_v_lowpass(uint8_t *dst, uint8_t *src, int dstStride, int srcStride,int w,int h){\ 956 uint8_t *cm = ff_cropTbl + MAX_NEG_CROP;\ 957 do{\ 958 int srcB,srcA,src0,src1,src2,src3,src4,src5,src6;\ 959 uint8_t *s = src-2*srcStride,*d=dst;\ 960 srcB = *s; s+=srcStride;\ 961 srcA = *s; s+=srcStride;\ 962 src0 = *s; s+=srcStride;\ 963 src1 = *s; s+=srcStride;\ 964 src2 = *s; s+=srcStride;\ 965 src3 = *s; s+=srcStride;\ 966 OP(*d, (src0+src1)*20 - (srcA+src2)*5 + (srcB+src3));d+=dstStride;\ 967 src4 = *s; s+=srcStride;\ 968 OP(*d, (src1+src2)*20 - (src0+src3)*5 + (srcA+src4));d+=dstStride;\ 969 src5 = *s; s+=srcStride;\ 970 OP(*d, (src2+src3)*20 - (src1+src4)*5 + (src0+src5));d+=dstStride;\ 971 src6 = *s; s+=srcStride;\ 972 OP(*d, (src3+src4)*20 - (src2+src5)*5 + (src1+src6));d+=dstStride;\ 973 if (h>4) { \ 974 int src7,src8,src9,src10; \ 975 src7 = *s; s+=srcStride;\ 976 OP(*d, (src4+src5)*20 - (src3+src6)*5 + (src2+src7));d+=dstStride;\ 977 src8 = *s; s+=srcStride;\ 978 OP(*d, (src5+src6)*20 - (src4+src7)*5 + (src3+src8));d+=dstStride;\ 979 src9 = *s; s+=srcStride;\ 980 OP(*d, (src6+src7)*20 - (src5+src8)*5 + (src4+src9));d+=dstStride;\ 981 src10 = *s; s+=srcStride;\ 982 OP(*d, (src7+src8)*20 - (src6+src9)*5 + (src5+src10));d+=dstStride;\ 983 if (h>8) { \ 984 int src11,src12,src13,src14,src15,src16,src17,src18; \ 985 src11 = *s; s+=srcStride;\ 986 OP(*d , (src8 +src9 )*20 - (src7 +src10)*5 + (src6 +src11));d+=dstStride;\ 987 src12 = *s; s+=srcStride;\ 988 OP(*d , (src9 +src10)*20 - (src8 +src11)*5 + (src7 +src12));d+=dstStride;\ 989 src13 = *s; s+=srcStride;\ 990 OP(*d, (src10+src11)*20 - (src9 +src12)*5 + (src8 +src13));d+=dstStride;\ 991 src14 = *s; s+=srcStride;\ 992 OP(*d, (src11+src12)*20 - (src10+src13)*5 + (src9 +src14));d+=dstStride;\ 993 src15 = *s; s+=srcStride;\ 994 OP(*d, (src12+src13)*20 - (src11+src14)*5 + (src10+src15));d+=dstStride;\ 995 src16 = *s; s+=srcStride;\ 996 OP(*d, (src13+src14)*20 - (src12+src15)*5 + (src11+src16));d+=dstStride;\ 997 src17 = *s; s+=srcStride;\ 998 OP(*d, (src14+src15)*20 - (src13+src16)*5 + (src12+src17));d+=dstStride;\ 999 src18 = *s; s+=srcStride;\ 1000 OP(*d, (src15+src16)*20 - (src14+src17)*5 + (src13+src18));d+=dstStride;\ 1001 } \ 1002 } \ 1003 dst++;\ 1004 src++;\ 1005 }while(--w);\ 1006}\ 1007\ 1008static inline void OPNAME ## h264_qpel_hv_lowpass(uint8_t *dst, int16_t *tmp, uint8_t *src, int dstStride, int tmpStride, int srcStride,int w,int h){\ 1009 uint8_t *cm = ff_cropTbl + MAX_NEG_CROP;\ 1010 int i;\ 1011 src -= 2*srcStride;\ 1012 i= h+5; \ 1013 do {\ 1014 int srcB,srcA,src0,src1,src2,src3,src4,src5,src6;\ 1015 uint8_t *s = src-2;\ 1016 srcB = *s++;\ 1017 srcA = *s++;\ 1018 src0 = *s++;\ 1019 src1 = *s++;\ 1020 src2 = *s++;\ 1021 src3 = *s++;\ 1022 tmp[0] = ((src0+src1)*20 - (srcA+src2)*5 + (srcB+src3));\ 1023 src4 = *s++;\ 1024 tmp[1] = ((src1+src2)*20 - (src0+src3)*5 + (srcA+src4));\ 1025 src5 = *s++;\ 1026 tmp[2] = ((src2+src3)*20 - (src1+src4)*5 + (src0+src5));\ 1027 src6 = *s++;\ 1028 tmp[3] = ((src3+src4)*20 - (src2+src5)*5 + (src1+src6));\ 1029 if (w>4) { /* it optimized */ \ 1030 int src7,src8,src9,src10; \ 1031 src7 = *s++;\ 1032 tmp[4] = ((src4+src5)*20 - (src3+src6)*5 + (src2+src7));\ 1033 src8 = *s++;\ 1034 tmp[5] = ((src5+src6)*20 - (src4+src7)*5 + (src3+src8));\ 1035 src9 = *s++;\ 1036 tmp[6] = ((src6+src7)*20 - (src5+src8)*5 + (src4+src9));\ 1037 src10 = *s++;\ 1038 tmp[7] = ((src7+src8)*20 - (src6+src9)*5 + (src5+src10));\ 1039 if (w>8) { \ 1040 int src11,src12,src13,src14,src15,src16,src17,src18; \ 1041 src11 = *s++;\ 1042 tmp[8] = ((src8 +src9 )*20 - (src7 +src10)*5 + (src6 +src11));\ 1043 src12 = *s++;\ 1044 tmp[9] = ((src9 +src10)*20 - (src8 +src11)*5 + (src7 +src12));\ 1045 src13 = *s++;\ 1046 tmp[10] = ((src10+src11)*20 - (src9 +src12)*5 + (src8 +src13));\ 1047 src14 = *s++;\ 1048 tmp[11] = ((src11+src12)*20 - (src10+src13)*5 + (src9 +src14));\ 1049 src15 = *s++;\ 1050 tmp[12] = ((src12+src13)*20 - (src11+src14)*5 + (src10+src15));\ 1051 src16 = *s++;\ 1052 tmp[13] = ((src13+src14)*20 - (src12+src15)*5 + (src11+src16));\ 1053 src17 = *s++;\ 1054 tmp[14] = ((src14+src15)*20 - (src13+src16)*5 + (src12+src17));\ 1055 src18 = *s++;\ 1056 tmp[15] = ((src15+src16)*20 - (src14+src17)*5 + (src13+src18));\ 1057 } \ 1058 } \ 1059 tmp+=tmpStride;\ 1060 src+=srcStride;\ 1061 }while(--i);\ 1062 tmp -= tmpStride*(h+5-2);\ 1063 i = w; \ 1064 do {\ 1065 int tmpB,tmpA,tmp0,tmp1,tmp2,tmp3,tmp4,tmp5,tmp6;\ 1066 int16_t *s = tmp-2*tmpStride; \ 1067 uint8_t *d=dst;\ 1068 tmpB = *s; s+=tmpStride;\ 1069 tmpA = *s; s+=tmpStride;\ 1070 tmp0 = *s; s+=tmpStride;\ 1071 tmp1 = *s; s+=tmpStride;\ 1072 tmp2 = *s; s+=tmpStride;\ 1073 tmp3 = *s; s+=tmpStride;\ 1074 OP2(*d, (tmp0+tmp1)*20 - (tmpA+tmp2)*5 + (tmpB+tmp3));d+=dstStride;\ 1075 tmp4 = *s; s+=tmpStride;\ 1076 OP2(*d, (tmp1+tmp2)*20 - (tmp0+tmp3)*5 + (tmpA+tmp4));d+=dstStride;\ 1077 tmp5 = *s; s+=tmpStride;\ 1078 OP2(*d, (tmp2+tmp3)*20 - (tmp1+tmp4)*5 + (tmp0+tmp5));d+=dstStride;\ 1079 tmp6 = *s; s+=tmpStride;\ 1080 OP2(*d, (tmp3+tmp4)*20 - (tmp2+tmp5)*5 + (tmp1+tmp6));d+=dstStride;\ 1081 if (h>4) { \ 1082 int tmp7,tmp8,tmp9,tmp10; \ 1083 tmp7 = *s; s+=tmpStride;\ 1084 OP2(*d, (tmp4+tmp5)*20 - (tmp3+tmp6)*5 + (tmp2+tmp7));d+=dstStride;\ 1085 tmp8 = *s; s+=tmpStride;\ 1086 OP2(*d, (tmp5+tmp6)*20 - (tmp4+tmp7)*5 + (tmp3+tmp8));d+=dstStride;\ 1087 tmp9 = *s; s+=tmpStride;\ 1088 OP2(*d, (tmp6+tmp7)*20 - (tmp5+tmp8)*5 + (tmp4+tmp9));d+=dstStride;\ 1089 tmp10 = *s; s+=tmpStride;\ 1090 OP2(*d, (tmp7+tmp8)*20 - (tmp6+tmp9)*5 + (tmp5+tmp10));d+=dstStride;\ 1091 if (h>8) { \ 1092 int tmp11,tmp12,tmp13,tmp14,tmp15,tmp16,tmp17,tmp18; \ 1093 tmp11 = *s; s+=tmpStride;\ 1094 OP2(*d , (tmp8 +tmp9 )*20 - (tmp7 +tmp10)*5 + (tmp6 +tmp11));d+=dstStride;\ 1095 tmp12 = *s; s+=tmpStride;\ 1096 OP2(*d , (tmp9 +tmp10)*20 - (tmp8 +tmp11)*5 + (tmp7 +tmp12));d+=dstStride;\ 1097 tmp13 = *s; s+=tmpStride;\ 1098 OP2(*d, (tmp10+tmp11)*20 - (tmp9 +tmp12)*5 + (tmp8 +tmp13));d+=dstStride;\ 1099 tmp14 = *s; s+=tmpStride;\ 1100 OP2(*d, (tmp11+tmp12)*20 - (tmp10+tmp13)*5 + (tmp9 +tmp14));d+=dstStride;\ 1101 tmp15 = *s; s+=tmpStride;\ 1102 OP2(*d, (tmp12+tmp13)*20 - (tmp11+tmp14)*5 + (tmp10+tmp15));d+=dstStride;\ 1103 tmp16 = *s; s+=tmpStride;\ 1104 OP2(*d, (tmp13+tmp14)*20 - (tmp12+tmp15)*5 + (tmp11+tmp16));d+=dstStride;\ 1105 tmp17 = *s; s+=tmpStride;\ 1106 OP2(*d, (tmp14+tmp15)*20 - (tmp13+tmp16)*5 + (tmp12+tmp17));d+=dstStride;\ 1107 tmp18 = *s; s+=tmpStride;\ 1108 OP2(*d, (tmp15+tmp16)*20 - (tmp14+tmp17)*5 + (tmp13+tmp18));d+=dstStride;\ 1109 } \ 1110 } \ 1111 dst++;\ 1112 tmp++;\ 1113 }while(--i);\ 1114}\ 1115\ 1116static void OPNAME ## h264_qpel4_h_lowpass(uint8_t *dst, uint8_t *src, int dstStride, int srcStride){\ 1117 OPNAME ## h264_qpel_h_lowpass(dst,src,dstStride,srcStride,4,4); \ 1118}\ 1119static void OPNAME ## h264_qpel8_h_lowpass(uint8_t *dst, uint8_t *src, int dstStride, int srcStride){\ 1120 OPNAME ## h264_qpel_h_lowpass(dst,src,dstStride,srcStride,8,8); \ 1121}\ 1122static void OPNAME ## h264_qpel16_h_lowpass(uint8_t *dst, uint8_t *src, int dstStride, int srcStride){\ 1123 OPNAME ## h264_qpel_h_lowpass(dst,src,dstStride,srcStride,16,16); \ 1124}\ 1125\ 1126static void OPNAME ## h264_qpel4_v_lowpass(uint8_t *dst, uint8_t *src, int dstStride, int srcStride){\ 1127 OPNAME ## h264_qpel_v_lowpass(dst,src,dstStride,srcStride,4,4); \ 1128}\ 1129static void OPNAME ## h264_qpel8_v_lowpass(uint8_t *dst, uint8_t *src, int dstStride, int srcStride){\ 1130 OPNAME ## h264_qpel_v_lowpass(dst,src,dstStride,srcStride,8,8); \ 1131}\ 1132static void OPNAME ## h264_qpel16_v_lowpass(uint8_t *dst, uint8_t *src, int dstStride, int srcStride){\ 1133 OPNAME ## h264_qpel_v_lowpass(dst,src,dstStride,srcStride,16,16); \ 1134}\ 1135static void OPNAME ## h264_qpel4_hv_lowpass(uint8_t *dst, int16_t *tmp, uint8_t *src, int dstStride, int tmpStride, int srcStride){\ 1136 OPNAME ## h264_qpel_hv_lowpass(dst,tmp,src,dstStride,tmpStride,srcStride,4,4); \ 1137}\ 1138static void OPNAME ## h264_qpel8_hv_lowpass(uint8_t *dst, int16_t *tmp, uint8_t *src, int dstStride, int tmpStride, int srcStride){\ 1139 OPNAME ## h264_qpel_hv_lowpass(dst,tmp,src,dstStride,tmpStride,srcStride,8,8); \ 1140}\ 1141static void OPNAME ## h264_qpel16_hv_lowpass(uint8_t *dst, int16_t *tmp, uint8_t *src, int dstStride, int tmpStride, int srcStride){\ 1142 OPNAME ## h264_qpel_hv_lowpass(dst,tmp,src,dstStride,tmpStride,srcStride,16,16); \ 1143}\ 1144 1145#define H264_MC(OPNAME, SIZE) \ 1146static void OPNAME ## h264_qpel ## SIZE ## _mc00_sh4 (uint8_t *dst, uint8_t *src, int stride){\ 1147 OPNAME ## pixels ## SIZE ## _c(dst, src, stride, SIZE);\ 1148}\ 1149\ 1150static void OPNAME ## h264_qpel ## SIZE ## _mc10_sh4(uint8_t *dst, uint8_t *src, int stride){\ 1151 uint8_t half[SIZE*SIZE];\ 1152 put_h264_qpel ## SIZE ## _h_lowpass(half, src, SIZE, stride);\ 1153 OPNAME ## pixels ## SIZE ## _l2_aligned2(dst, src, half, stride, stride, SIZE, SIZE);\ 1154}\ 1155\ 1156static void OPNAME ## h264_qpel ## SIZE ## _mc20_sh4(uint8_t *dst, uint8_t *src, int stride){\ 1157 OPNAME ## h264_qpel ## SIZE ## _h_lowpass(dst, src, stride, stride);\ 1158}\ 1159\ 1160static void OPNAME ## h264_qpel ## SIZE ## _mc30_sh4(uint8_t *dst, uint8_t *src, int stride){\ 1161 uint8_t half[SIZE*SIZE];\ 1162 put_h264_qpel ## SIZE ## _h_lowpass(half, src, SIZE, stride);\ 1163 OPNAME ## pixels ## SIZE ## _l2_aligned2(dst, src+1, half, stride, stride, SIZE, SIZE);\ 1164}\ 1165\ 1166static void OPNAME ## h264_qpel ## SIZE ## _mc01_sh4(uint8_t *dst, uint8_t *src, int stride){\ 1167 uint8_t full[SIZE*(SIZE+5)];\ 1168 uint8_t * const full_mid= full + SIZE*2;\ 1169 uint8_t half[SIZE*SIZE];\ 1170 copy_block ## SIZE (full, src - stride*2, SIZE, stride, SIZE + 5);\ 1171 put_h264_qpel ## SIZE ## _v_lowpass(half, full_mid, SIZE, SIZE);\ 1172 OPNAME ## pixels ## SIZE ## _l2_aligned(dst, full_mid, half, stride, SIZE, SIZE, SIZE);\ 1173}\ 1174\ 1175static void OPNAME ## h264_qpel ## SIZE ## _mc02_sh4(uint8_t *dst, uint8_t *src, int stride){\ 1176 uint8_t full[SIZE*(SIZE+5)];\ 1177 uint8_t * const full_mid= full + SIZE*2;\ 1178 copy_block ## SIZE (full, src - stride*2, SIZE, stride, SIZE + 5);\ 1179 OPNAME ## h264_qpel ## SIZE ## _v_lowpass(dst, full_mid, stride, SIZE);\ 1180}\ 1181\ 1182static void OPNAME ## h264_qpel ## SIZE ## _mc03_sh4(uint8_t *dst, uint8_t *src, int stride){\ 1183 uint8_t full[SIZE*(SIZE+5)];\ 1184 uint8_t * const full_mid= full + SIZE*2;\ 1185 uint8_t half[SIZE*SIZE];\ 1186 copy_block ## SIZE (full, src - stride*2, SIZE, stride, SIZE + 5);\ 1187 put_h264_qpel ## SIZE ## _v_lowpass(half, full_mid, SIZE, SIZE);\ 1188 OPNAME ## pixels ## SIZE ## _l2_aligned(dst, full_mid+SIZE, half, stride, SIZE, SIZE, SIZE);\ 1189}\ 1190\ 1191static void OPNAME ## h264_qpel ## SIZE ## _mc11_sh4(uint8_t *dst, uint8_t *src, int stride){\ 1192 uint8_t full[SIZE*(SIZE+5)];\ 1193 uint8_t * const full_mid= full + SIZE*2;\ 1194 uint8_t halfH[SIZE*SIZE];\ 1195 uint8_t halfV[SIZE*SIZE];\ 1196 put_h264_qpel ## SIZE ## _h_lowpass(halfH, src, SIZE, stride);\ 1197 copy_block ## SIZE (full, src - stride*2, SIZE, stride, SIZE + 5);\ 1198 put_h264_qpel ## SIZE ## _v_lowpass(halfV, full_mid, SIZE, SIZE);\ 1199 OPNAME ## pixels ## SIZE ## _l2_aligned(dst, halfH, halfV, stride, SIZE, SIZE, SIZE);\ 1200}\ 1201\ 1202static void OPNAME ## h264_qpel ## SIZE ## _mc31_sh4(uint8_t *dst, uint8_t *src, int stride){\ 1203 uint8_t full[SIZE*(SIZE+5)];\ 1204 uint8_t * const full_mid= full + SIZE*2;\ 1205 uint8_t halfH[SIZE*SIZE];\ 1206 uint8_t halfV[SIZE*SIZE];\ 1207 put_h264_qpel ## SIZE ## _h_lowpass(halfH, src, SIZE, stride);\ 1208 copy_block ## SIZE (full, src - stride*2 + 1, SIZE, stride, SIZE + 5);\ 1209 put_h264_qpel ## SIZE ## _v_lowpass(halfV, full_mid, SIZE, SIZE);\ 1210 OPNAME ## pixels ## SIZE ## _l2_aligned(dst, halfH, halfV, stride, SIZE, SIZE, SIZE);\ 1211}\ 1212\ 1213static void OPNAME ## h264_qpel ## SIZE ## _mc13_sh4(uint8_t *dst, uint8_t *src, int stride){\ 1214 uint8_t full[SIZE*(SIZE+5)];\ 1215 uint8_t * const full_mid= full + SIZE*2;\ 1216 uint8_t halfH[SIZE*SIZE];\ 1217 uint8_t halfV[SIZE*SIZE];\ 1218 put_h264_qpel ## SIZE ## _h_lowpass(halfH, src + stride, SIZE, stride);\ 1219 copy_block ## SIZE (full, src - stride*2, SIZE, stride, SIZE + 5);\ 1220 put_h264_qpel ## SIZE ## _v_lowpass(halfV, full_mid, SIZE, SIZE);\ 1221 OPNAME ## pixels ## SIZE ## _l2_aligned(dst, halfH, halfV, stride, SIZE, SIZE, SIZE);\ 1222}\ 1223\ 1224static void OPNAME ## h264_qpel ## SIZE ## _mc33_sh4(uint8_t *dst, uint8_t *src, int stride){\ 1225 uint8_t full[SIZE*(SIZE+5)];\ 1226 uint8_t * const full_mid= full + SIZE*2;\ 1227 uint8_t halfH[SIZE*SIZE];\ 1228 uint8_t halfV[SIZE*SIZE];\ 1229 put_h264_qpel ## SIZE ## _h_lowpass(halfH, src + stride, SIZE, stride);\ 1230 copy_block ## SIZE (full, src - stride*2 + 1, SIZE, stride, SIZE + 5);\ 1231 put_h264_qpel ## SIZE ## _v_lowpass(halfV, full_mid, SIZE, SIZE);\ 1232 OPNAME ## pixels ## SIZE ## _l2_aligned(dst, halfH, halfV, stride, SIZE, SIZE, SIZE);\ 1233}\ 1234\ 1235static void OPNAME ## h264_qpel ## SIZE ## _mc22_sh4(uint8_t *dst, uint8_t *src, int stride){\ 1236 int16_t tmp[SIZE*(SIZE+5)];\ 1237 OPNAME ## h264_qpel ## SIZE ## _hv_lowpass(dst, tmp, src, stride, SIZE, stride);\ 1238}\ 1239\ 1240static void OPNAME ## h264_qpel ## SIZE ## _mc21_sh4(uint8_t *dst, uint8_t *src, int stride){\ 1241 int16_t tmp[SIZE*(SIZE+5)];\ 1242 uint8_t halfH[SIZE*SIZE];\ 1243 uint8_t halfHV[SIZE*SIZE];\ 1244 put_h264_qpel ## SIZE ## _h_lowpass(halfH, src, SIZE, stride);\ 1245 put_h264_qpel ## SIZE ## _hv_lowpass(halfHV, tmp, src, SIZE, SIZE, stride);\ 1246 OPNAME ## pixels ## SIZE ## _l2_aligned(dst, halfH, halfHV, stride, SIZE, SIZE, SIZE);\ 1247}\ 1248\ 1249static void OPNAME ## h264_qpel ## SIZE ## _mc23_sh4(uint8_t *dst, uint8_t *src, int stride){\ 1250 int16_t tmp[SIZE*(SIZE+5)];\ 1251 uint8_t halfH[SIZE*SIZE];\ 1252 uint8_t halfHV[SIZE*SIZE];\ 1253 put_h264_qpel ## SIZE ## _h_lowpass(halfH, src + stride, SIZE, stride);\ 1254 put_h264_qpel ## SIZE ## _hv_lowpass(halfHV, tmp, src, SIZE, SIZE, stride);\ 1255 OPNAME ## pixels ## SIZE ## _l2_aligned(dst, halfH, halfHV, stride, SIZE, SIZE, SIZE);\ 1256}\ 1257\ 1258static void OPNAME ## h264_qpel ## SIZE ## _mc12_sh4(uint8_t *dst, uint8_t *src, int stride){\ 1259 uint8_t full[SIZE*(SIZE+5)];\ 1260 uint8_t * const full_mid= full + SIZE*2;\ 1261 int16_t tmp[SIZE*(SIZE+5)];\ 1262 uint8_t halfV[SIZE*SIZE];\ 1263 uint8_t halfHV[SIZE*SIZE];\ 1264 copy_block ## SIZE (full, src - stride*2, SIZE, stride, SIZE + 5);\ 1265 put_h264_qpel ## SIZE ## _v_lowpass(halfV, full_mid, SIZE, SIZE);\ 1266 put_h264_qpel ## SIZE ## _hv_lowpass(halfHV, tmp, src, SIZE, SIZE, stride);\ 1267 OPNAME ## pixels ## SIZE ## _l2_aligned(dst, halfV, halfHV, stride, SIZE, SIZE, SIZE);\ 1268}\ 1269\ 1270static void OPNAME ## h264_qpel ## SIZE ## _mc32_sh4(uint8_t *dst, uint8_t *src, int stride){\ 1271 uint8_t full[SIZE*(SIZE+5)];\ 1272 uint8_t * const full_mid= full + SIZE*2;\ 1273 int16_t tmp[SIZE*(SIZE+5)];\ 1274 uint8_t halfV[SIZE*SIZE];\ 1275 uint8_t halfHV[SIZE*SIZE];\ 1276 copy_block ## SIZE (full, src - stride*2 + 1, SIZE, stride, SIZE + 5);\ 1277 put_h264_qpel ## SIZE ## _v_lowpass(halfV, full_mid, SIZE, SIZE);\ 1278 put_h264_qpel ## SIZE ## _hv_lowpass(halfHV, tmp, src, SIZE, SIZE, stride);\ 1279 OPNAME ## pixels ## SIZE ## _l2_aligned(dst, halfV, halfHV, stride, SIZE, SIZE, SIZE);\ 1280}\ 1281 1282#define op_avg(a, b) a = (((a)+cm[((b) + 16)>>5]+1)>>1) 1283//#define op_avg2(a, b) a = (((a)*w1+cm[((b) + 16)>>5]*w2 + o + 64)>>7) 1284#define op_put(a, b) a = cm[((b) + 16)>>5] 1285#define op2_avg(a, b) a = (((a)+cm[((b) + 512)>>10]+1)>>1) 1286#define op2_put(a, b) a = cm[((b) + 512)>>10] 1287 1288H264_LOWPASS(put_ , op_put, op2_put) 1289H264_LOWPASS(avg_ , op_avg, op2_avg) 1290H264_MC(put_, 4) 1291H264_MC(put_, 8) 1292H264_MC(put_, 16) 1293H264_MC(avg_, 4) 1294H264_MC(avg_, 8) 1295H264_MC(avg_, 16) 1296 1297#undef op_avg 1298#undef op_put 1299#undef op2_avg 1300#undef op2_put 1301#endif 1302 1303static void wmv2_mspel8_h_lowpass(uint8_t *dst, uint8_t *src, int dstStride, int srcStride, int h){ 1304 uint8_t *cm = ff_cropTbl + MAX_NEG_CROP; 1305 1306 do{ 1307 int src_1,src0,src1,src2,src3,src4,src5,src6,src7,src8,src9; 1308 uint8_t *s = src; 1309 src_1 = s[-1]; 1310 src0 = *s++; 1311 src1 = *s++; 1312 src2 = *s++; 1313 dst[0]= cm[(9*(src0 + src1) - (src_1 + src2) + 8)>>4]; 1314 src3 = *s++; 1315 dst[1]= cm[(9*(src1 + src2) - (src0 + src3) + 8)>>4]; 1316 src4 = *s++; 1317 dst[2]= cm[(9*(src2 + src3) - (src1 + src4) + 8)>>4]; 1318 src5 = *s++; 1319 dst[3]= cm[(9*(src3 + src4) - (src2 + src5) + 8)>>4]; 1320 src6 = *s++; 1321 dst[4]= cm[(9*(src4 + src5) - (src3 + src6) + 8)>>4]; 1322 src7 = *s++; 1323 dst[5]= cm[(9*(src5 + src6) - (src4 + src7) + 8)>>4]; 1324 src8 = *s++; 1325 dst[6]= cm[(9*(src6 + src7) - (src5 + src8) + 8)>>4]; 1326 src9 = *s++; 1327 dst[7]= cm[(9*(src7 + src8) - (src6 + src9) + 8)>>4]; 1328 dst+=dstStride; 1329 src+=srcStride; 1330 }while(--h); 1331} 1332 1333static void wmv2_mspel8_v_lowpass(uint8_t *dst, uint8_t *src, int dstStride, int srcStride, int w){ 1334 uint8_t *cm = ff_cropTbl + MAX_NEG_CROP; 1335 1336 do{ 1337 int src_1,src0,src1,src2,src3,src4,src5,src6,src7,src8,src9; 1338 uint8_t *s = src,*d = dst; 1339 src_1 = *(s-srcStride); 1340 src0 = *s; s+=srcStride; 1341 src1 = *s; s+=srcStride; 1342 src2 = *s; s+=srcStride; 1343 *d= cm[(9*(src0 + src1) - (src_1 + src2) + 8)>>4]; d+=dstStride; 1344 src3 = *s; s+=srcStride; 1345 *d= cm[(9*(src1 + src2) - (src0 + src3) + 8)>>4]; d+=dstStride; 1346 src4 = *s; s+=srcStride; 1347 *d= cm[(9*(src2 + src3) - (src1 + src4) + 8)>>4]; d+=dstStride; 1348 src5 = *s; s+=srcStride; 1349 *d= cm[(9*(src3 + src4) - (src2 + src5) + 8)>>4]; d+=dstStride; 1350 src6 = *s; s+=srcStride; 1351 *d= cm[(9*(src4 + src5) - (src3 + src6) + 8)>>4]; d+=dstStride; 1352 src7 = *s; s+=srcStride; 1353 *d= cm[(9*(src5 + src6) - (src4 + src7) + 8)>>4]; d+=dstStride; 1354 src8 = *s; s+=srcStride; 1355 *d= cm[(9*(src6 + src7) - (src5 + src8) + 8)>>4]; d+=dstStride; 1356 src9 = *s; 1357 *d= cm[(9*(src7 + src8) - (src6 + src9) + 8)>>4]; d+=dstStride; 1358 src++; 1359 dst++; 1360 }while(--w); 1361} 1362 1363static void put_mspel8_mc00_sh4 (uint8_t *dst, uint8_t *src, int stride){ 1364 put_pixels8_c(dst, src, stride, 8); 1365} 1366 1367static void put_mspel8_mc10_sh4(uint8_t *dst, uint8_t *src, int stride){ 1368 uint8_t half[64]; 1369 wmv2_mspel8_h_lowpass(half, src, 8, stride, 8); 1370 put_pixels8_l2_aligned2(dst, src, half, stride, stride, 8, 8); 1371} 1372 1373static void put_mspel8_mc20_sh4(uint8_t *dst, uint8_t *src, int stride){ 1374 wmv2_mspel8_h_lowpass(dst, src, stride, stride, 8); 1375} 1376 1377static void put_mspel8_mc30_sh4(uint8_t *dst, uint8_t *src, int stride){ 1378 uint8_t half[64]; 1379 wmv2_mspel8_h_lowpass(half, src, 8, stride, 8); 1380 put_pixels8_l2_aligned2(dst, src+1, half, stride, stride, 8, 8); 1381} 1382 1383static void put_mspel8_mc02_sh4(uint8_t *dst, uint8_t *src, int stride){ 1384 wmv2_mspel8_v_lowpass(dst, src, stride, stride, 8); 1385} 1386 1387static void put_mspel8_mc12_sh4(uint8_t *dst, uint8_t *src, int stride){ 1388 uint8_t halfH[88]; 1389 uint8_t halfV[64]; 1390 uint8_t halfHV[64]; 1391 wmv2_mspel8_h_lowpass(halfH, src-stride, 8, stride, 11); 1392 wmv2_mspel8_v_lowpass(halfV, src, 8, stride, 8); 1393 wmv2_mspel8_v_lowpass(halfHV, halfH+8, 8, 8, 8); 1394 put_pixels8_l2_aligned(dst, halfV, halfHV, stride, 8, 8, 8); 1395} 1396static void put_mspel8_mc32_sh4(uint8_t *dst, uint8_t *src, int stride){ 1397 uint8_t halfH[88]; 1398 uint8_t halfV[64]; 1399 uint8_t halfHV[64]; 1400 wmv2_mspel8_h_lowpass(halfH, src-stride, 8, stride, 11); 1401 wmv2_mspel8_v_lowpass(halfV, src+1, 8, stride, 8); 1402 wmv2_mspel8_v_lowpass(halfHV, halfH+8, 8, 8, 8); 1403 put_pixels8_l2_aligned(dst, halfV, halfHV, stride, 8, 8, 8); 1404} 1405static void put_mspel8_mc22_sh4(uint8_t *dst, uint8_t *src, int stride){ 1406 uint8_t halfH[88]; 1407 wmv2_mspel8_h_lowpass(halfH, src-stride, 8, stride, 11); 1408 wmv2_mspel8_v_lowpass(dst, halfH+8, stride, 8, 8); 1409} 1410