Lines Matching refs:spu_add

47   minus_a = (vec_char16)(spu_add((vec_ushort8)(spu_and(spu_xor(a, 0xFF), 0x7F)), 0x101));
73 minus_a = (vec_char16)spu_add((vec_short8)(spu_xor(a, -1)),
82 minus_a = spu_add(spu_sub(0, a), (vec_short8)(spu_cmpeq(a, ((vec_short8){0x8000,0x8000,0x8000,0x8000,0x8000,0x8000,0x8000,0x8000}))));
90 minus_a = spu_add(spu_sub(0, a), (vec_int4)(spu_cmpeq(a, ((vec_int4){0x80000000,0x80000000,0x80000000,0x80000000}))));
100 return ((vec_uchar16)(spu_sel(spu_add((vec_ushort8)(a), (vec_ushort8)(b)),
101 spu_add(spu_and((vec_ushort8)(a), 0xFF00), spu_and((vec_ushort8)(b), 0xFF00)),
122 return (spu_add(a, b));
127 return (spu_add(a, b));
132 return (spu_add((vec_short8)(a), b));
137 return (spu_add(a, (vec_short8)(b)));
142 return (spu_add(a, b));
147 return (spu_add(a, b));
152 return (spu_add((vec_int4)(a), b));
157 return (spu_add(a, (vec_int4)(b)));
162 return (spu_add(a, b));
177 s1 = (vec_uchar16)(spu_add(spu_rlmask((vec_ushort8)(a), -8), spu_rlmask((vec_ushort8)(b), -8)));
178 s2 = (vec_uchar16)(spu_add(spu_and((vec_ushort8)(a), 0xFF), spu_and((vec_ushort8)(b), 0xFF)));
190 s1 = (vec_uchar16)(spu_add(spu_rlmask((vec_ushort8)(a), -8), spu_rlmask((vec_ushort8)(b), -8)));
191 s2 = (vec_uchar16)(spu_add(spu_and((vec_ushort8)(a), 0xFF), spu_and((vec_ushort8)(b), 0xFF)));
213 s = spu_add(a, b);
222 s = spu_add(a, b);
240 return (spu_or(spu_add(a, b), spu_rlmaska(spu_sl(spu_genc(a, b), 31), -31)));
247 s = spu_add(a, b);
436 return (spu_add(spu_add(spu_rlmask(a, -1), spu_rlmask(b, -1)),
442 return (spu_add(spu_add(spu_rlmaska(a, -1), spu_rlmaska(b, -1)),
448 return (spu_add(spu_add(spu_rlmask(a, -1), spu_rlmask(b, -1)),
454 return (spu_add(spu_add(spu_rlmaska(a, -1), spu_rlmaska(b, -1)),
467 a = spu_add(a, (vec_float4)(spu_and(spu_xor(spu_rlmaska((vec_int4)a, -31), -1), spu_splats((signed int)0x3F7FFFFF))));
471 mask = spu_or(mask, spu_xor((vec_uint4)(spu_rlmaska(spu_add(exp, -1), -31)), -1));
588 ia = spu_convts(spu_add(a, bias), 0);
590 exp = (vec_float4)(spu_sl(spu_add(ia, 127), 23));
609 mask = spu_or(mask, spu_xor((vec_uint4)(spu_rlmaska(spu_add(exp, -1), -31)), -1));
742 exp = spu_add((vec_int4)(spu_and(spu_rlmask((vec_uint4)(a), -23), 0xFF)), -127);
755 return ((vec_uchar16)spu_add((vec_ushort8)(spu_splats((unsigned char)((a + (int)(b)) & 0xF))),
1122 hi = (vec_short8)(spu_sl(spu_add(spu_mule(a, b), round), 1));
1123 lo = (vec_short8)(spu_rlmask(spu_add(spu_mulo(a, b), round), -15));
1142 p1 = spu_add(spu_mulo(a1, b1), spu_mulo(spu_rlqwbyte(a1, -2), spu_rlqwbyte(b1, -2)));
1143 p2 = spu_add(spu_mulo(a2, b2), spu_mulo(spu_rlqwbyte(a2, -2), spu_rlqwbyte(b2, -2)));
1144 return (spu_add(p2, spu_add(p1, c)));
1157 p1 = spu_add(spu_mulo(a1, b1), spu_mulo(spu_rlqwbyte(a1, -2), spu_rlqwbyte(b1, -2)));
1158 p2 = spu_add(spu_mulo(a2, b2), spu_mulo(spu_rlqwbyte(a2, -2), spu_rlqwbyte(b2, -2)));
1159 return (spu_add(p2, spu_add(p1, c)));
1164 return (spu_add(spu_add(spu_mulo(a, b), spu_mulo(spu_rlqwbyte(a, -2), spu_rlqwbyte(b, -2))), c));
1169 return (spu_add(spu_add(spu_mulo(a, b), spu_mulo(spu_rlqwbyte(a, -2), spu_rlqwbyte(b, -2))), c));
1188 return (vec_adds(spu_add(spu_mulo(a, b), spu_mulo(spu_rlqwbyte(a, -2), spu_rlqwbyte(b, -2))), c));
1605 a = spu_add(a, s_half);
1606 s_one = spu_add(s_half, s_half);
1610 mask = spu_or(mask, spu_xor((vec_uint4)(spu_rlmaska(spu_add(exp, -1), -31)), -1));
2163 a01_23 = (vec_uint4)(spu_add(spu_rlmask((vec_ushort8)(a), -8),
2165 a0123 = spu_add(spu_rlmask(a01_23, -16), spu_and(a01_23, 0x1FF));
2173 a01_23 = (vec_int4)(spu_add(spu_rlmaska((vec_short8)(a), -8),
2175 a0123 = spu_add(spu_rlmaska(a01_23, -16), spu_extend((vec_short8)(a01_23)));
2183 a0123 = spu_add(spu_rlmaska((vec_int4)(a), -16), spu_extend(a));
2204 sum_l = spu_add(a, b);
2208 sum_l = spu_add(sum_l, c);
2240 sum_l = spu_add(a, b);
2244 sum_l = spu_add(sum_l, a2);
2248 sum_l = spu_add(sum_l, a1);
2252 sum_l = spu_add(sum_l, a0);
2279 mask = spu_or(mask, spu_xor((vec_uint4)(spu_rlmaska(spu_add(exp, -1), -31)), -1));