• Home
  • History
  • Annotate
  • Raw
  • Download
  • only in /freebsd-12-stable/contrib/llvm-project/clang/lib/Headers/

Lines Matching refs:__mmask8

21 static __inline __mmask8 __DEFAULT_FN_ATTRS
22 _knot_mask8(__mmask8 __M)
27 static __inline__ __mmask8 __DEFAULT_FN_ATTRS
28 _kand_mask8(__mmask8 __A, __mmask8 __B)
30 return (__mmask8)__builtin_ia32_kandqi((__mmask8)__A, (__mmask8)__B);
33 static __inline__ __mmask8 __DEFAULT_FN_ATTRS
34 _kandn_mask8(__mmask8 __A, __mmask8 __B)
36 return (__mmask8)__builtin_ia32_kandnqi((__mmask8)__A, (__mmask8)__B);
39 static __inline__ __mmask8 __DEFAULT_FN_ATTRS
40 _kor_mask8(__mmask8 __A, __mmask8 __B)
42 return (__mmask8)__builtin_ia32_korqi((__mmask8)__A, (__mmask8)__B);
45 static __inline__ __mmask8 __DEFAULT_FN_ATTRS
46 _kxnor_mask8(__mmask8 __A, __mmask8 __B)
48 return (__mmask8)__builtin_ia32_kxnorqi((__mmask8)__A, (__mmask8)__B);
51 static __inline__ __mmask8 __DEFAULT_FN_ATTRS
52 _kxor_mask8(__mmask8 __A, __mmask8 __B)
54 return (__mmask8)__builtin_ia32_kxorqi((__mmask8)__A, (__mmask8)__B);
58 _kortestc_mask8_u8(__mmask8 __A, __mmask8 __B)
64 _kortestz_mask8_u8(__mmask8 __A, __mmask8 __B)
70 _kortest_mask8_u8(__mmask8 __A, __mmask8 __B, unsigned char *__C) {
76 _ktestc_mask8_u8(__mmask8 __A, __mmask8 __B)
82 _ktestz_mask8_u8(__mmask8 __A, __mmask8 __B)
88 _ktest_mask8_u8(__mmask8 __A, __mmask8 __B, unsigned char *__C) {
111 static __inline__ __mmask8 __DEFAULT_FN_ATTRS
112 _kadd_mask8(__mmask8 __A, __mmask8 __B)
114 return (__mmask8)__builtin_ia32_kaddqi((__mmask8)__A, (__mmask8)__B);
124 (__mmask8)__builtin_ia32_kshiftliqi((__mmask8)(A), (unsigned int)(I))
127 (__mmask8)__builtin_ia32_kshiftriqi((__mmask8)(A), (unsigned int)(I))
130 _cvtmask8_u32(__mmask8 __A) {
131 return (unsigned int)__builtin_ia32_kmovb((__mmask8)__A);
134 static __inline__ __mmask8 __DEFAULT_FN_ATTRS
136 return (__mmask8)__builtin_ia32_kmovb((__mmask8)__A);
139 static __inline__ __mmask8 __DEFAULT_FN_ATTRS
140 _load_mask8(__mmask8 *__A) {
141 return (__mmask8)__builtin_ia32_kmovb(*(__mmask8 *)__A);
145 _store_mask8(__mmask8 *__A, __mmask8 __B) {
146 *(__mmask8 *)__A = __builtin_ia32_kmovb((__mmask8)__B);
155 _mm512_mask_mullo_epi64(__m512i __W, __mmask8 __U, __m512i __A, __m512i __B) {
156 return (__m512i)__builtin_ia32_selectq_512((__mmask8)__U,
162 _mm512_maskz_mullo_epi64(__mmask8 __U, __m512i __A, __m512i __B) {
163 return (__m512i)__builtin_ia32_selectq_512((__mmask8)__U,
174 _mm512_mask_xor_pd(__m512d __W, __mmask8 __U, __m512d __A, __m512d __B) {
175 return (__m512d)__builtin_ia32_selectpd_512((__mmask8)__U,
181 _mm512_maskz_xor_pd(__mmask8 __U, __m512d __A, __m512d __B) {
182 return (__m512d)__builtin_ia32_selectpd_512((__mmask8)__U,
212 _mm512_mask_or_pd(__m512d __W, __mmask8 __U, __m512d __A, __m512d __B) {
213 return (__m512d)__builtin_ia32_selectpd_512((__mmask8)__U,
219 _mm512_maskz_or_pd(__mmask8 __U, __m512d __A, __m512d __B) {
220 return (__m512d)__builtin_ia32_selectpd_512((__mmask8)__U,
250 _mm512_mask_and_pd(__m512d __W, __mmask8 __U, __m512d __A, __m512d __B) {
251 return (__m512d)__builtin_ia32_selectpd_512((__mmask8)__U,
257 _mm512_maskz_and_pd(__mmask8 __U, __m512d __A, __m512d __B) {
258 return (__m512d)__builtin_ia32_selectpd_512((__mmask8)__U,
288 _mm512_mask_andnot_pd(__m512d __W, __mmask8 __U, __m512d __A, __m512d __B) {
289 return (__m512d)__builtin_ia32_selectpd_512((__mmask8)__U,
295 _mm512_maskz_andnot_pd(__mmask8 __U, __m512d __A, __m512d __B) {
296 return (__m512d)__builtin_ia32_selectpd_512((__mmask8)__U,
324 (__mmask8) -1,
329 _mm512_mask_cvtpd_epi64 (__m512i __W, __mmask8 __U, __m512d __A) {
332 (__mmask8) __U,
337 _mm512_maskz_cvtpd_epi64 (__mmask8 __U, __m512d __A) {
340 (__mmask8) __U,
347 (__mmask8)-1, (int)(R))
352 (__mmask8)(U), (int)(R))
357 (__mmask8)(U), (int)(R))
363 (__mmask8) -1,
368 _mm512_mask_cvtpd_epu64 (__m512i __W, __mmask8 __U, __m512d __A) {
371 (__mmask8) __U,
376 _mm512_maskz_cvtpd_epu64 (__mmask8 __U, __m512d __A) {
379 (__mmask8) __U,
386 (__mmask8)-1, (int)(R))
391 (__mmask8)(U), (int)(R))
396 (__mmask8)(U), (int)(R))
402 (__mmask8) -1,
407 _mm512_mask_cvtps_epi64 (__m512i __W, __mmask8 __U, __m256 __A) {
410 (__mmask8) __U,
415 _mm512_maskz_cvtps_epi64 (__mmask8 __U, __m256 __A) {
418 (__mmask8) __U,
425 (__mmask8)-1, (int)(R))
430 (__mmask8)(U), (int)(R))
435 (__mmask8)(U), (int)(R))
441 (__mmask8) -1,
446 _mm512_mask_cvtps_epu64 (__m512i __W, __mmask8 __U, __m256 __A) {
449 (__mmask8) __U,
454 _mm512_maskz_cvtps_epu64 (__mmask8 __U, __m256 __A) {
457 (__mmask8) __U,
464 (__mmask8)-1, (int)(R))
469 (__mmask8)(U), (int)(R))
474 (__mmask8)(U), (int)(R))
483 _mm512_mask_cvtepi64_pd (__m512d __W, __mmask8 __U, __m512i __A) {
484 return (__m512d)__builtin_ia32_selectpd_512((__mmask8)__U,
490 _mm512_maskz_cvtepi64_pd (__mmask8 __U, __m512i __A) {
491 return (__m512d)__builtin_ia32_selectpd_512((__mmask8)__U,
499 (__mmask8)-1, (int)(R))
504 (__mmask8)(U), (int)(R))
509 (__mmask8)(U), (int)(R))
515 (__mmask8) -1,
520 _mm512_mask_cvtepi64_ps (__m256 __W, __mmask8 __U, __m512i __A) {
523 (__mmask8) __U,
528 _mm512_maskz_cvtepi64_ps (__mmask8 __U, __m512i __A) {
531 (__mmask8) __U,
538 (__mmask8)-1, (int)(R))
542 (__v8sf)(__m256)(W), (__mmask8)(U), \
548 (__mmask8)(U), (int)(R))
555 (__mmask8) -1,
560 _mm512_mask_cvttpd_epi64 (__m512i __W, __mmask8 __U, __m512d __A) {
563 (__mmask8) __U,
568 _mm512_maskz_cvttpd_epi64 (__mmask8 __U, __m512d __A) {
571 (__mmask8) __U,
578 (__mmask8)-1, (int)(R))
583 (__mmask8)(U), (int)(R))
588 (__mmask8)(U), (int)(R))
594 (__mmask8) -1,
599 _mm512_mask_cvttpd_epu64 (__m512i __W, __mmask8 __U, __m512d __A) {
602 (__mmask8) __U,
607 _mm512_maskz_cvttpd_epu64 (__mmask8 __U, __m512d __A) {
610 (__mmask8) __U,
617 (__mmask8)-1, (int)(R))
622 (__mmask8)(U), (int)(R))
627 (__mmask8)(U), (int)(R))
633 (__mmask8) -1,
638 _mm512_mask_cvttps_epi64 (__m512i __W, __mmask8 __U, __m256 __A) {
641 (__mmask8) __U,
646 _mm512_maskz_cvttps_epi64 (__mmask8 __U, __m256 __A) {
649 (__mmask8) __U,
656 (__mmask8)-1, (int)(R))
661 (__mmask8)(U), (int)(R))
666 (__mmask8)(U), (int)(R))
672 (__mmask8) -1,
677 _mm512_mask_cvttps_epu64 (__m512i __W, __mmask8 __U, __m256 __A) {
680 (__mmask8) __U,
685 _mm512_maskz_cvttps_epu64 (__mmask8 __U, __m256 __A) {
688 (__mmask8) __U,
695 (__mmask8)-1, (int)(R))
700 (__mmask8)(U), (int)(R))
705 (__mmask8)(U), (int)(R))
713 _mm512_mask_cvtepu64_pd (__m512d __W, __mmask8 __U, __m512i __A) {
714 return (__m512d)__builtin_ia32_selectpd_512((__mmask8)__U,
720 _mm512_maskz_cvtepu64_pd (__mmask8 __U, __m512i __A) {
721 return (__m512d)__builtin_ia32_selectpd_512((__mmask8)__U,
729 (__mmask8)-1, (int)(R))
734 (__mmask8)(U), (int)(R))
740 (__mmask8)(U), (int)(R))
747 (__mmask8) -1,
752 _mm512_mask_cvtepu64_ps (__m256 __W, __mmask8 __U, __m512i __A) {
755 (__mmask8) __U,
760 _mm512_maskz_cvtepu64_ps (__mmask8 __U, __m512i __A) {
763 (__mmask8) __U,
770 (__mmask8)-1, (int)(R))
774 (__v8sf)(__m256)(W), (__mmask8)(U), \
780 (__mmask8)(U), (int)(R))
786 (__mmask8)-1, \
792 (__v8df)(__m512d)(W), (__mmask8)(U), \
799 (__mmask8)(U), \
806 (__mmask8)-1, (int)(R))
811 (__v8df)(__m512d)(W), (__mmask8)(U), \
818 (__mmask8)(U), (int)(R))
862 (__mmask8) -1, (int)(C),\
871 (__mmask8)(U), (int)(C),\
880 (__mmask8)(U), (int)(C),\
889 (__mmask8) -1, (int)(C),\
898 (__mmask8)(U), (int)(C),\
907 (__mmask8)(U), (int)(C),\
915 (__mmask8)-1, \
921 (__mmask8)(U), \
927 (__mmask8)(U), \
951 (__mmask8)-1, (int)(R))
956 (__mmask8)(U), (int)(R))
961 (__mmask8)(U), (int)(R))
981 (__v4sf)_mm_setzero_ps(), (__mmask8)-1, \
987 (__v4sf)(__m128)(W), (__mmask8)(U), \
994 (__mmask8)(U), (int)(C), \
1000 (__v4sf)_mm_setzero_ps(), (__mmask8)-1, \
1006 (__v4sf)(__m128)(W), (__mmask8)(U), \
1013 (__mmask8)(U), (int)(C), (int)(R))
1019 (__mmask8)-1, (int)(C), \
1025 (__v2df)(__m128d)(W), (__mmask8)(U), \
1032 (__mmask8)(U), (int)(C), \
1039 (__mmask8)-1, (int)(C), (int)(R))
1044 (__v2df)(__m128d)(W), (__mmask8)(U), \
1051 (__mmask8)(U), (int)(C), (int)(R))
1066 _mm512_movm_epi64 (__mmask8 __A)
1071 static __inline__ __mmask8 __DEFAULT_FN_ATTRS512
1074 return (__mmask8) __builtin_ia32_cvtq2mask512 ((__v8di) __A);
1134 _mm512_mask_broadcast_f64x2(__m512d __O, __mmask8 __M, __m128d __A)
1136 return (__m512d)__builtin_ia32_selectpd_512((__mmask8)__M,
1142 _mm512_maskz_broadcast_f64x2(__mmask8 __M, __m128d __A)
1144 return (__m512d)__builtin_ia32_selectpd_512((__mmask8)__M,
1205 _mm512_mask_broadcast_i64x2(__m512i __O, __mmask8 __M, __m128i __A)
1207 return (__m512i)__builtin_ia32_selectq_512((__mmask8)__M,
1213 _mm512_maskz_broadcast_i64x2(__mmask8 __M, __m128i __A)
1215 return (__m512i)__builtin_ia32_selectq_512((__mmask8)__M,
1223 (__mmask8)-1)
1228 (__mmask8)(U))
1233 (__mmask8)(U))
1239 (__mmask8)-1)
1245 (__mmask8)(U))
1251 (__mmask8)(U))
1256 (__mmask8)-1)
1261 (__mmask8)(U))
1266 (__mmask8)(U))
1272 (__mmask8)-1)
1278 (__mmask8)(U))
1284 (__mmask8)(U))
1305 (__m512d)__builtin_ia32_selectpd_512((__mmask8)(U), \
1310 (__m512d)__builtin_ia32_selectpd_512((__mmask8)(U), \
1333 (__m512i)__builtin_ia32_selectq_512((__mmask8)(U), \
1338 (__m512i)__builtin_ia32_selectq_512((__mmask8)(U), \
1351 (__mmask8)__builtin_ia32_fpclasspd512_mask((__v8df)(__m512d)(A), (int)(imm), \
1352 (__mmask8)(U))
1355 (__mmask8)__builtin_ia32_fpclasspd512_mask((__v8df)(__m512d)(A), (int)(imm), \
1356 (__mmask8)-1)
1359 (__mmask8)__builtin_ia32_fpclasssd_mask((__v2df)(__m128d)(A), (int)(imm), \
1360 (__mmask8)-1)
1363 (__mmask8)__builtin_ia32_fpclasssd_mask((__v2df)(__m128d)(A), (int)(imm), \
1364 (__mmask8)(U))
1367 (__mmask8)__builtin_ia32_fpclassss_mask((__v4sf)(__m128)(A), (int)(imm), \
1368 (__mmask8)-1)
1371 (__mmask8)__builtin_ia32_fpclassss_mask((__v4sf)(__m128)(A), (int)(imm), \
1372 (__mmask8)(U))