Lines Matching refs:__b

65 /// \param __b
70 _mm256_add_pd(__m256d __a, __m256d __b)
72 return (__m256d)((__v4df)__a+(__v4df)__b);
83 /// \param __b
88 _mm256_add_ps(__m256 __a, __m256 __b)
90 return (__m256)((__v8sf)__a+(__v8sf)__b);
101 /// \param __b
106 _mm256_sub_pd(__m256d __a, __m256d __b)
108 return (__m256d)((__v4df)__a-(__v4df)__b);
119 /// \param __b
124 _mm256_sub_ps(__m256 __a, __m256 __b)
126 return (__m256)((__v8sf)__a-(__v8sf)__b);
138 /// \param __b
143 _mm256_addsub_pd(__m256d __a, __m256d __b)
145 return (__m256d)__builtin_ia32_addsubpd256((__v4df)__a, (__v4df)__b);
157 /// \param __b
162 _mm256_addsub_ps(__m256 __a, __m256 __b)
164 return (__m256)__builtin_ia32_addsubps256((__v8sf)__a, (__v8sf)__b);
175 /// \param __b
180 _mm256_div_pd(__m256d __a, __m256d __b)
182 return (__m256d)((__v4df)__a/(__v4df)__b);
193 /// \param __b
198 _mm256_div_ps(__m256 __a, __m256 __b)
200 return (__m256)((__v8sf)__a/(__v8sf)__b);
212 /// \param __b
217 _mm256_max_pd(__m256d __a, __m256d __b)
219 return (__m256d)__builtin_ia32_maxpd256((__v4df)__a, (__v4df)__b);
231 /// \param __b
236 _mm256_max_ps(__m256 __a, __m256 __b)
238 return (__m256)__builtin_ia32_maxps256((__v8sf)__a, (__v8sf)__b);
250 /// \param __b
255 _mm256_min_pd(__m256d __a, __m256d __b)
257 return (__m256d)__builtin_ia32_minpd256((__v4df)__a, (__v4df)__b);
269 /// \param __b
274 _mm256_min_ps(__m256 __a, __m256 __b)
276 return (__m256)__builtin_ia32_minps256((__v8sf)__a, (__v8sf)__b);
287 /// \param __b
292 _mm256_mul_pd(__m256d __a, __m256d __b)
294 return (__m256d)((__v4df)__a * (__v4df)__b);
305 /// \param __b
310 _mm256_mul_ps(__m256 __a, __m256 __b)
312 return (__m256)((__v8sf)__a * (__v8sf)__b);
525 /// \param __b
530 _mm256_and_pd(__m256d __a, __m256d __b)
532 return (__m256d)((__v4du)__a & (__v4du)__b);
543 /// \param __b
548 _mm256_and_ps(__m256 __a, __m256 __b)
550 return (__m256)((__v8su)__a & (__v8su)__b);
563 /// \param __b
569 _mm256_andnot_pd(__m256d __a, __m256d __b)
571 return (__m256d)(~(__v4du)__a & (__v4du)__b);
584 /// \param __b
590 _mm256_andnot_ps(__m256 __a, __m256 __b)
592 return (__m256)(~(__v8su)__a & (__v8su)__b);
603 /// \param __b
608 _mm256_or_pd(__m256d __a, __m256d __b)
610 return (__m256d)((__v4du)__a | (__v4du)__b);
621 /// \param __b
626 _mm256_or_ps(__m256 __a, __m256 __b)
628 return (__m256)((__v8su)__a | (__v8su)__b);
639 /// \param __b
644 _mm256_xor_pd(__m256d __a, __m256d __b)
646 return (__m256d)((__v4du)__a ^ (__v4du)__b);
657 /// \param __b
662 _mm256_xor_ps(__m256 __a, __m256 __b)
664 return (__m256)((__v8su)__a ^ (__v8su)__b);
679 /// \param __b
686 _mm256_hadd_pd(__m256d __a, __m256d __b)
688 return (__m256d)__builtin_ia32_haddpd256((__v4df)__a, (__v4df)__b);
702 /// \param __b
709 _mm256_hadd_ps(__m256 __a, __m256 __b)
711 return (__m256)__builtin_ia32_haddps256((__v8sf)__a, (__v8sf)__b);
725 /// \param __b
732 _mm256_hsub_pd(__m256d __a, __m256d __b)
734 return (__m256d)__builtin_ia32_hsubpd256((__v4df)__a, (__v4df)__b);
748 /// \param __b
755 _mm256_hsub_ps(__m256 __a, __m256 __b)
757 return (__m256)__builtin_ia32_hsubps256((__v8sf)__a, (__v8sf)__b);
1381 /// \param __b
1389 /// 64-bit element in operand \a __b is copied to the same position in the
1393 _mm256_blendv_pd(__m256d __a, __m256d __b, __m256d __c)
1396 (__v4df)__a, (__v4df)__b, (__v4df)__c);
1409 /// \param __b
1417 /// corresponding 32-bit element in operand \a __b is copied to the same
1421 _mm256_blendv_ps(__m256 __a, __m256 __b, __m256 __c)
1424 (__v8sf)__a, (__v8sf)__b, (__v8sf)__c);
2320 __v8si __b = (__v8si)__a;
2321 return __b[0];
2425 /// \param __b
2431 _mm256_unpackhi_pd(__m256d __a, __m256d __b)
2433 return __builtin_shufflevector((__v4df)__a, (__v4df)__b, 1, 5, 1+2, 5+2);
2447 /// \param __b
2453 _mm256_unpacklo_pd(__m256d __a, __m256d __b)
2455 return __builtin_shufflevector((__v4df)__a, (__v4df)__b, 0, 4, 0+2, 4+2);
2472 /// \param __b
2480 _mm256_unpackhi_ps(__m256 __a, __m256 __b)
2482 return __builtin_shufflevector((__v8sf)__a, (__v8sf)__b, 2, 10, 2+1, 10+1, 6, 14, 6+1, 14+1);
2499 /// \param __b
2507 _mm256_unpacklo_ps(__m256 __a, __m256 __b)
2509 return __builtin_shufflevector((__v8sf)__a, (__v8sf)__b, 0, 8, 0+1, 8+1, 4, 12, 4+1, 12+1);
2533 /// \param __b
2537 _mm_testz_pd(__m128d __a, __m128d __b)
2539 return __builtin_ia32_vtestzpd((__v2df)__a, (__v2df)__b);
2562 /// \param __b
2566 _mm_testc_pd(__m128d __a, __m128d __b)
2568 return __builtin_ia32_vtestcpd((__v2df)__a, (__v2df)__b);
2592 /// \param __b
2596 _mm_testnzc_pd(__m128d __a, __m128d __b)
2598 return __builtin_ia32_vtestnzcpd((__v2df)__a, (__v2df)__b);
2621 /// \param __b
2625 _mm_testz_ps(__m128 __a, __m128 __b)
2627 return __builtin_ia32_vtestzps((__v4sf)__a, (__v4sf)__b);
2650 /// \param __b
2654 _mm_testc_ps(__m128 __a, __m128 __b)
2656 return __builtin_ia32_vtestcps((__v4sf)__a, (__v4sf)__b);
2680 /// \param __b
2684 _mm_testnzc_ps(__m128 __a, __m128 __b)
2686 return __builtin_ia32_vtestnzcps((__v4sf)__a, (__v4sf)__b);
2709 /// \param __b
2713 _mm256_testz_pd(__m256d __a, __m256d __b)
2715 return __builtin_ia32_vtestzpd256((__v4df)__a, (__v4df)__b);
2738 /// \param __b
2742 _mm256_testc_pd(__m256d __a, __m256d __b)
2744 return __builtin_ia32_vtestcpd256((__v4df)__a, (__v4df)__b);
2768 /// \param __b
2772 _mm256_testnzc_pd(__m256d __a, __m256d __b)
2774 return __builtin_ia32_vtestnzcpd256((__v4df)__a, (__v4df)__b);
2797 /// \param __b
2801 _mm256_testz_ps(__m256 __a, __m256 __b)
2803 return __builtin_ia32_vtestzps256((__v8sf)__a, (__v8sf)__b);
2826 /// \param __b
2830 _mm256_testc_ps(__m256 __a, __m256 __b)
2832 return __builtin_ia32_vtestcps256((__v8sf)__a, (__v8sf)__b);
2856 /// \param __b
2860 _mm256_testnzc_ps(__m256 __a, __m256 __b)
2862 return __builtin_ia32_vtestnzcps256((__v8sf)__a, (__v8sf)__b);
2882 /// \param __b
2886 _mm256_testz_si256(__m256i __a, __m256i __b)
2888 return __builtin_ia32_ptestz256((__v4di)__a, (__v4di)__b);
2908 /// \param __b
2912 _mm256_testc_si256(__m256i __a, __m256i __b)
2914 return __builtin_ia32_ptestc256((__v4di)__a, (__v4di)__b);
2935 /// \param __b
2939 _mm256_testnzc_si256(__m256i __a, __m256i __b)
2941 return __builtin_ia32_ptestnzc256((__v4di)__a, (__v4di)__b);
3077 __m128d __b = _mm_loadu_pd((const double *)__a);
3078 return (__m256d)__builtin_shufflevector((__v2df)__b, (__v2df)__b,
3097 __m128 __b = _mm_loadu_ps((const float *)__a);
3098 return (__m256)__builtin_shufflevector((__v4sf)__b, (__v4sf)__b,
3554 /// \param __b
3557 _mm256_stream_si256(__m256i *__a, __m256i __b)
3560 __builtin_nontemporal_store((__v4di_aligned)__b, (__v4di_aligned*)__a);
3574 /// \param __b
3577 _mm256_stream_pd(double *__a, __m256d __b)
3580 __builtin_nontemporal_store((__v4df_aligned)__b, (__v4df_aligned*)__a);
3655 /// \param __b
3666 _mm256_set_pd(double __a, double __b, double __c, double __d)
3668 return __extension__ (__m256d){ __d, __c, __b, __a };
3682 /// \param __b
3705 _mm256_set_ps(float __a, float __b, float __c, float __d,
3708 return __extension__ (__m256){ __h, __g, __f, __e, __d, __c, __b, __a };
3895 /// \param __b
3903 _mm256_set_epi64x(long long __a, long long __b, long long __c, long long __d)
3905 return __extension__ (__m256i)(__v4di){ __d, __c, __b, __a };
3921 /// \param __b
3932 _mm256_setr_pd(double __a, double __b, double __c, double __d)
3934 return _mm256_set_pd(__d, __c, __b, __a);
3949 /// \param __b
3972 _mm256_setr_ps(float __a, float __b, float __c, float __d,
3975 return _mm256_set_ps(__h, __g, __f, __e, __d, __c, __b, __a);
4162 /// \param __b
4170 _mm256_setr_epi64x(long long __a, long long __b, long long __c, long long __d)
4172 return _mm256_set_epi64x(__d, __c, __b, __a);
4258 /// \param __b
4263 _mm256_set1_epi8(char __b)
4265 return _mm256_set_epi8(__b, __b, __b, __b, __b, __b, __b, __b,
4266 __b, __b, __b, __b, __b, __b, __b, __b,
4267 __b, __b, __b, __b, __b, __b, __b, __b,
4268 __b, __b, __b, __b, __b, __b, __b, __b);