/external/llvm-project/clang/test/CodeGen/X86/ |
D | avx512vl-builtins-constrained-cmp.c | 5 __mmask8 test_mm256_cmp_ps_mask_eq_oq(__m256 a, __m256 b) { in test_mm256_cmp_ps_mask_eq_oq() 11 __mmask8 test_mm256_cmp_ps_mask_lt_os(__m256 a, __m256 b) { in test_mm256_cmp_ps_mask_lt_os() 17 __mmask8 test_mm256_cmp_ps_mask_le_os(__m256 a, __m256 b) { in test_mm256_cmp_ps_mask_le_os() 23 __mmask8 test_mm256_cmp_ps_mask_unord_q(__m256 a, __m256 b) { in test_mm256_cmp_ps_mask_unord_q() 29 __mmask8 test_mm256_cmp_ps_mask_neq_uq(__m256 a, __m256 b) { in test_mm256_cmp_ps_mask_neq_uq() 35 __mmask8 test_mm256_cmp_ps_mask_nlt_us(__m256 a, __m256 b) { in test_mm256_cmp_ps_mask_nlt_us() 41 __mmask8 test_mm256_cmp_ps_mask_nle_us(__m256 a, __m256 b) { in test_mm256_cmp_ps_mask_nle_us() 47 __mmask8 test_mm256_cmp_ps_mask_ord_q(__m256 a, __m256 b) { in test_mm256_cmp_ps_mask_ord_q() 53 __mmask8 test_mm256_cmp_ps_mask_eq_uq(__m256 a, __m256 b) { in test_mm256_cmp_ps_mask_eq_uq() 59 __mmask8 test_mm256_cmp_ps_mask_nge_us(__m256 a, __m256 b) { in test_mm256_cmp_ps_mask_nge_us() [all …]
|
D | avx512vl-builtins.c | 5 __mmask8 test_mm_cmpeq_epu32_mask(__m128i __a, __m128i __b) { in test_mm_cmpeq_epu32_mask() 9 return (__mmask8)_mm_cmpeq_epu32_mask(__a, __b); in test_mm_cmpeq_epu32_mask() 12 __mmask8 test_mm_mask_cmpeq_epu32_mask(__mmask8 __u, __m128i __a, __m128i __b) { in test_mm_mask_cmpeq_epu32_mask() 16 return (__mmask8)_mm_mask_cmpeq_epu32_mask(__u, __a, __b); in test_mm_mask_cmpeq_epu32_mask() 19 __mmask8 test_mm_cmpeq_epu64_mask(__m128i __a, __m128i __b) { in test_mm_cmpeq_epu64_mask() 23 return (__mmask8)_mm_cmpeq_epu64_mask(__a, __b); in test_mm_cmpeq_epu64_mask() 26 __mmask8 test_mm_mask_cmpeq_epu64_mask(__mmask8 __u, __m128i __a, __m128i __b) { in test_mm_mask_cmpeq_epu64_mask() 30 return (__mmask8)_mm_mask_cmpeq_epu64_mask(__u, __a, __b); in test_mm_mask_cmpeq_epu64_mask() 33 __mmask8 test_mm_cmpge_epi32_mask(__m128i __a, __m128i __b) { in test_mm_cmpge_epi32_mask() 36 return (__mmask8)_mm_cmpge_epi32_mask(__a, __b); in test_mm_cmpge_epi32_mask() [all …]
|
D | avx512vldq-builtins.c | 12 __m256i test_mm256_mask_mullo_epi64 (__m256i __W, __mmask8 __U, __m256i __A, __m256i __B) { in test_mm256_mask_mullo_epi64() 19 __m256i test_mm256_maskz_mullo_epi64 (__mmask8 __U, __m256i __A, __m256i __B) { in test_mm256_maskz_mullo_epi64() 32 __m128i test_mm_mask_mullo_epi64 (__m128i __W, __mmask8 __U, __m128i __A, __m128i __B) { in test_mm_mask_mullo_epi64() 39 __m128i test_mm_maskz_mullo_epi64 (__mmask8 __U, __m128i __A, __m128i __B) { in test_mm_maskz_mullo_epi64() 46 __m256d test_mm256_mask_andnot_pd (__m256d __W, __mmask8 __U, __m256d __A, __m256d __B) { in test_mm256_mask_andnot_pd() 54 __m256d test_mm256_maskz_andnot_pd (__mmask8 __U, __m256d __A, __m256d __B) { in test_mm256_maskz_andnot_pd() 62 __m128d test_mm_mask_andnot_pd (__m128d __W, __mmask8 __U, __m128d __A, __m128d __B) { in test_mm_mask_andnot_pd() 70 __m128d test_mm_maskz_andnot_pd (__mmask8 __U, __m128d __A, __m128d __B) { in test_mm_maskz_andnot_pd() 78 __m256 test_mm256_mask_andnot_ps (__m256 __W, __mmask8 __U, __m256 __A, __m256 __B) { in test_mm256_mask_andnot_ps() 86 __m256 test_mm256_maskz_andnot_ps (__mmask8 __U, __m256 __A, __m256 __B) { in test_mm256_maskz_andnot_ps() [all …]
|
D | avx512f-builtins-constrained-cmp.c | 401 __mmask8 test_mm512_cmp_round_pd_mask(__m512d a, __m512d b) { in test_mm512_cmp_round_pd_mask() 407 __mmask8 test_mm512_mask_cmp_round_pd_mask(__mmask8 m, __m512d a, __m512d b) { in test_mm512_mask_cmp_round_pd_mask() 413 __mmask8 test_mm512_cmp_pd_mask_eq_oq(__m512d a, __m512d b) { in test_mm512_cmp_pd_mask_eq_oq() 419 __mmask8 test_mm512_cmp_pd_mask_lt_os(__m512d a, __m512d b) { in test_mm512_cmp_pd_mask_lt_os() 425 __mmask8 test_mm512_cmp_pd_mask_le_os(__m512d a, __m512d b) { in test_mm512_cmp_pd_mask_le_os() 431 __mmask8 test_mm512_cmp_pd_mask_unord_q(__m512d a, __m512d b) { in test_mm512_cmp_pd_mask_unord_q() 437 __mmask8 test_mm512_cmp_pd_mask_neq_uq(__m512d a, __m512d b) { in test_mm512_cmp_pd_mask_neq_uq() 443 __mmask8 test_mm512_cmp_pd_mask_nlt_us(__m512d a, __m512d b) { in test_mm512_cmp_pd_mask_nlt_us() 449 __mmask8 test_mm512_cmp_pd_mask_nle_us(__m512d a, __m512d b) { in test_mm512_cmp_pd_mask_nle_us() 455 __mmask8 test_mm512_cmp_pd_mask_ord_q(__m512d a, __m512d b) { in test_mm512_cmp_pd_mask_ord_q() [all …]
|
D | avx512dq-builtins.c | 6 __mmask8 test_knot_mask8(__mmask8 a) { in test_knot_mask8() 14 __mmask8 test_kand_mask8(__m512i __A, __m512i __B, __m512i __C, __m512i __D, __m512i __E, __m512i _… in test_kand_mask8() 25 __mmask8 test_kandn_mask8(__m512i __A, __m512i __B, __m512i __C, __m512i __D, __m512i __E, __m512i … in test_kandn_mask8() 37 __mmask8 test_kor_mask8(__m512i __A, __m512i __B, __m512i __C, __m512i __D, __m512i __E, __m512i __… in test_kor_mask8() 48 __mmask8 test_kxnor_mask8(__m512i __A, __m512i __B, __m512i __C, __m512i __D, __m512i __E, __m512i … in test_kxnor_mask8() 60 __mmask8 test_kxor_mask8(__m512i __A, __m512i __B, __m512i __C, __m512i __D, __m512i __E, __m512i _… in test_kxor_mask8() 185 __mmask8 test_kadd_mask8(__m512i __A, __m512i __B, __m512i __C, __m512i __D, __m512i __E, __m512i _… in test_kadd_mask8() 207 __mmask8 test_kshiftli_mask8(__m512i A, __m512i B, __m512i C, __m512i D) { in test_kshiftli_mask8() 215 __mmask8 test_kshiftri_mask8(__m512i A, __m512i B, __m512i C, __m512i D) { in test_kshiftri_mask8() 231 __mmask8 test_cvtu32_mask8(__m512i A, __m512i B, unsigned int C) { in test_cvtu32_mask8() [all …]
|
D | avx512f-builtins.c | 13 __m512d test_mm512_mask_sqrt_pd (__m512d __W, __mmask8 __U, __m512d __A) in test_mm512_mask_sqrt_pd() 22 __m512d test_mm512_maskz_sqrt_pd (__mmask8 __U, __m512d __A) in test_mm512_maskz_sqrt_pd() 31 __m512d test_mm512_mask_sqrt_round_pd(__m512d __W,__mmask8 __U,__m512d __A) in test_mm512_mask_sqrt_round_pd() 40 __m512d test_mm512_maskz_sqrt_round_pd(__mmask8 __U,__m512d __A) in test_mm512_maskz_sqrt_round_pd() 113 __m512d test_mm512_mask_rsqrt14_pd (__m512d __W, __mmask8 __U, __m512d __A) in test_mm512_mask_rsqrt14_pd() 120 __m512d test_mm512_maskz_rsqrt14_pd (__mmask8 __U, __m512d __A) in test_mm512_maskz_rsqrt14_pd() 251 void test_mm512_mask_store_pd(void *p, __m512d a, __mmask8 m) in test_mm512_mask_store_pd() 276 void test_mm512_mask_storeu_epi64(void *__P, __mmask8 __U, __m512i __A) { in test_mm512_mask_storeu_epi64() 317 __m512i test_mm512_mask_loadu_epi64 (__m512i __W, __mmask8 __U, void *__P) in test_mm512_mask_loadu_epi64() 352 __m512d test_mm512_mask_loadu_pd (__m512d __W, __mmask8 __U, void *__P) in test_mm512_mask_loadu_pd() [all …]
|
D | avx512vlcd-builtins.c | 64 __m128i test_mm_mask_conflict_epi64(__m128i __W, __mmask8 __U, __m128i __A) { in test_mm_mask_conflict_epi64() 71 __m128i test_mm_maskz_conflict_epi64(__mmask8 __U, __m128i __A) { in test_mm_maskz_conflict_epi64() 84 __m256i test_mm256_mask_conflict_epi64(__m256i __W, __mmask8 __U, __m256i __A) { in test_mm256_mask_conflict_epi64() 91 __m256i test_mm256_maskz_conflict_epi64(__mmask8 __U, __m256i __A) { in test_mm256_maskz_conflict_epi64() 104 __m128i test_mm_mask_conflict_epi32(__m128i __W, __mmask8 __U, __m128i __A) { in test_mm_mask_conflict_epi32() 111 __m128i test_mm_maskz_conflict_epi32(__mmask8 __U, __m128i __A) { in test_mm_maskz_conflict_epi32() 124 __m256i test_mm256_mask_conflict_epi32(__m256i __W, __mmask8 __U, __m256i __A) { in test_mm256_mask_conflict_epi32() 131 __m256i test_mm256_maskz_conflict_epi32(__mmask8 __U, __m256i __A) { in test_mm256_maskz_conflict_epi32() 144 __m128i test_mm_mask_lzcnt_epi32(__m128i __W, __mmask8 __U, __m128i __A) { in test_mm_mask_lzcnt_epi32() 151 __m128i test_mm_maskz_lzcnt_epi32(__mmask8 __U, __m128i __A) { in test_mm_maskz_lzcnt_epi32() [all …]
|
/external/clang/lib/Headers/ |
D | avx512vlintrin.h | 41 static __inline__ __mmask8 __DEFAULT_FN_ATTRS 43 return (__mmask8)__builtin_ia32_pcmpeqd128_mask((__v4si)__a, (__v4si)__b, in _mm_cmpeq_epi32_mask() 44 (__mmask8)-1); in _mm_cmpeq_epi32_mask() 47 static __inline__ __mmask8 __DEFAULT_FN_ATTRS 48 _mm_mask_cmpeq_epi32_mask(__mmask8 __u, __m128i __a, __m128i __b) { in _mm_mask_cmpeq_epi32_mask() 49 return (__mmask8)__builtin_ia32_pcmpeqd128_mask((__v4si)__a, (__v4si)__b, in _mm_mask_cmpeq_epi32_mask() 53 static __inline__ __mmask8 __DEFAULT_FN_ATTRS 55 return (__mmask8)__builtin_ia32_ucmpd128_mask((__v4si)__a, (__v4si)__b, 0, in _mm_cmpeq_epu32_mask() 56 (__mmask8)-1); in _mm_cmpeq_epu32_mask() 59 static __inline__ __mmask8 __DEFAULT_FN_ATTRS [all …]
|
D | avx512vldqintrin.h | 40 _mm256_mask_mullo_epi64 (__m256i __W, __mmask8 __U, __m256i __A, __m256i __B) { in _mm256_mask_mullo_epi64() 44 (__mmask8) __U); in _mm256_mask_mullo_epi64() 48 _mm256_maskz_mullo_epi64 (__mmask8 __U, __m256i __A, __m256i __B) { in _mm256_maskz_mullo_epi64() 53 (__mmask8) __U); in _mm256_maskz_mullo_epi64() 62 _mm_mask_mullo_epi64 (__m128i __W, __mmask8 __U, __m128i __A, __m128i __B) { in _mm_mask_mullo_epi64() 66 (__mmask8) __U); in _mm_mask_mullo_epi64() 70 _mm_maskz_mullo_epi64 (__mmask8 __U, __m128i __A, __m128i __B) { in _mm_maskz_mullo_epi64() 75 (__mmask8) __U); in _mm_maskz_mullo_epi64() 79 _mm256_mask_andnot_pd (__m256d __W, __mmask8 __U, __m256d __A, __m256d __B) { in _mm256_mask_andnot_pd() 83 (__mmask8) __U); in _mm256_mask_andnot_pd() [all …]
|
D | avx512vlcdintrin.h | 35 _mm_broadcastmb_epi64 (__mmask8 __A) in _mm_broadcastmb_epi64() 41 _mm256_broadcastmb_epi64 (__mmask8 __A) in _mm256_broadcastmb_epi64() 64 (__mmask8) -1); in _mm_conflict_epi64() 68 _mm_mask_conflict_epi64 (__m128i __W, __mmask8 __U, __m128i __A) in _mm_mask_conflict_epi64() 72 (__mmask8) __U); in _mm_mask_conflict_epi64() 76 _mm_maskz_conflict_epi64 (__mmask8 __U, __m128i __A) in _mm_maskz_conflict_epi64() 81 (__mmask8) __U); in _mm_maskz_conflict_epi64() 89 (__mmask8) -1); in _mm256_conflict_epi64() 93 _mm256_mask_conflict_epi64 (__m256i __W, __mmask8 __U, __m256i __A) in _mm256_mask_conflict_epi64() 97 (__mmask8) __U); in _mm256_mask_conflict_epi64() [all …]
|
D | avx512dqintrin.h | 40 _mm512_mask_mullo_epi64 (__m512i __W, __mmask8 __U, __m512i __A, __m512i __B) { in _mm512_mask_mullo_epi64() 44 (__mmask8) __U); in _mm512_mask_mullo_epi64() 48 _mm512_maskz_mullo_epi64 (__mmask8 __U, __m512i __A, __m512i __B) { in _mm512_maskz_mullo_epi64() 53 (__mmask8) __U); in _mm512_maskz_mullo_epi64() 62 _mm512_mask_xor_pd (__m512d __W, __mmask8 __U, __m512d __A, __m512d __B) { in _mm512_mask_xor_pd() 66 (__mmask8) __U); in _mm512_mask_xor_pd() 70 _mm512_maskz_xor_pd (__mmask8 __U, __m512d __A, __m512d __B) { in _mm512_maskz_xor_pd() 75 (__mmask8) __U); in _mm512_maskz_xor_pd() 106 _mm512_mask_or_pd (__m512d __W, __mmask8 __U, __m512d __A, __m512d __B) { in _mm512_mask_or_pd() 110 (__mmask8) __U); in _mm512_mask_or_pd() [all …]
|
D | avx512ifmavlintrin.h | 42 (__mmask8) -1); in _mm_madd52hi_epu64() 46 _mm_mask_madd52hi_epu64 (__m128i __W, __mmask8 __M, __m128i __X, __m128i __Y) in _mm_mask_madd52hi_epu64() 51 (__mmask8) __M); in _mm_mask_madd52hi_epu64() 55 _mm_maskz_madd52hi_epu64 (__mmask8 __M, __m128i __X, __m128i __Y, __m128i __Z) in _mm_maskz_madd52hi_epu64() 60 (__mmask8) __M); in _mm_maskz_madd52hi_epu64() 69 (__mmask8) -1); in _mm256_madd52hi_epu64() 73 _mm256_mask_madd52hi_epu64 (__m256i __W, __mmask8 __M, __m256i __X, in _mm256_mask_madd52hi_epu64() 79 (__mmask8) __M); in _mm256_mask_madd52hi_epu64() 83 _mm256_maskz_madd52hi_epu64 (__mmask8 __M, __m256i __X, __m256i __Y, __m256i __Z) in _mm256_maskz_madd52hi_epu64() 88 (__mmask8) __M); in _mm256_maskz_madd52hi_epu64() [all …]
|
D | avx512fintrin.h | 47 typedef unsigned char __mmask8; typedef 232 _mm512_mask_broadcastq_epi64 (__m512i __O, __mmask8 __M, __m128i __A) in _mm512_mask_broadcastq_epi64() 241 _mm512_maskz_broadcastq_epi64 (__mmask8 __M, __m128i __A) in _mm512_maskz_broadcastq_epi64() 258 _mm512_maskz_set1_epi64(__mmask8 __M, long long __A) in _mm512_maskz_set1_epi64() 535 _mm512_mask_and_epi64(__m512i __src, __mmask8 __k, __m512i __a, __m512i __b) in _mm512_mask_and_epi64() 537 return (__m512i) __builtin_ia32_selectq_512 ((__mmask8) __k, in _mm512_mask_and_epi64() 543 _mm512_maskz_and_epi64(__mmask8 __k, __m512i __a, __m512i __b) in _mm512_maskz_and_epi64() 583 _mm512_mask_andnot_epi64(__m512i __W, __mmask8 __U, __m512i __A, __m512i __B) in _mm512_mask_andnot_epi64() 585 return (__m512i)__builtin_ia32_selectq_512((__mmask8)__U, in _mm512_mask_andnot_epi64() 591 _mm512_maskz_andnot_epi64(__mmask8 __U, __m512i __A, __m512i __B) in _mm512_maskz_andnot_epi64() [all …]
|
/external/clang/test/CodeGen/ |
D | avx512vl-builtins.c | 8 __mmask8 test_mm_cmpeq_epu32_mask(__m128i __a, __m128i __b) { in test_mm_cmpeq_epu32_mask() 12 return (__mmask8)_mm_cmpeq_epu32_mask(__a, __b); in test_mm_cmpeq_epu32_mask() 15 __mmask8 test_mm_mask_cmpeq_epu32_mask(__mmask8 __u, __m128i __a, __m128i __b) { in test_mm_mask_cmpeq_epu32_mask() 19 return (__mmask8)_mm_mask_cmpeq_epu32_mask(__u, __a, __b); in test_mm_mask_cmpeq_epu32_mask() 22 __mmask8 test_mm_cmpeq_epu64_mask(__m128i __a, __m128i __b) { in test_mm_cmpeq_epu64_mask() 26 return (__mmask8)_mm_cmpeq_epu64_mask(__a, __b); in test_mm_cmpeq_epu64_mask() 29 __mmask8 test_mm_mask_cmpeq_epu64_mask(__mmask8 __u, __m128i __a, __m128i __b) { in test_mm_mask_cmpeq_epu64_mask() 33 return (__mmask8)_mm_mask_cmpeq_epu64_mask(__u, __a, __b); in test_mm_mask_cmpeq_epu64_mask() 36 __mmask8 test_mm_cmpge_epi32_mask(__m128i __a, __m128i __b) { in test_mm_cmpge_epi32_mask() 39 return (__mmask8)_mm_cmpge_epi32_mask(__a, __b); in test_mm_cmpge_epi32_mask() [all …]
|
D | avx512vldq-builtins.c | 14 __m256i test_mm256_mask_mullo_epi64 (__m256i __W, __mmask8 __U, __m256i __A, __m256i __B) { in test_mm256_mask_mullo_epi64() 20 __m256i test_mm256_maskz_mullo_epi64 (__mmask8 __U, __m256i __A, __m256i __B) { in test_mm256_maskz_mullo_epi64() 32 __m128i test_mm_mask_mullo_epi64 (__m128i __W, __mmask8 __U, __m128i __A, __m128i __B) { in test_mm_mask_mullo_epi64() 38 __m128i test_mm_maskz_mullo_epi64 (__mmask8 __U, __m128i __A, __m128i __B) { in test_mm_maskz_mullo_epi64() 44 __m256d test_mm256_mask_andnot_pd (__m256d __W, __mmask8 __U, __m256d __A, __m256d __B) { in test_mm256_mask_andnot_pd() 50 __m256d test_mm256_maskz_andnot_pd (__mmask8 __U, __m256d __A, __m256d __B) { in test_mm256_maskz_andnot_pd() 56 __m128d test_mm_mask_andnot_pd (__m128d __W, __mmask8 __U, __m128d __A, __m128d __B) { in test_mm_mask_andnot_pd() 62 __m128d test_mm_maskz_andnot_pd (__mmask8 __U, __m128d __A, __m128d __B) { in test_mm_maskz_andnot_pd() 68 __m256 test_mm256_mask_andnot_ps (__m256 __W, __mmask8 __U, __m256 __A, __m256 __B) { in test_mm256_mask_andnot_ps() 74 __m256 test_mm256_maskz_andnot_ps (__mmask8 __U, __m256 __A, __m256 __B) { in test_mm256_maskz_andnot_ps() [all …]
|
D | avx512dq-builtins.c | 14 __m512i test_mm512_mask_mullo_epi64 (__m512i __W, __mmask8 __U, __m512i __A, __m512i __B) { in test_mm512_mask_mullo_epi64() 20 __m512i test_mm512_maskz_mullo_epi64 (__mmask8 __U, __m512i __A, __m512i __B) { in test_mm512_maskz_mullo_epi64() 32 __m512d test_mm512_mask_xor_pd (__m512d __W, __mmask8 __U, __m512d __A, __m512d __B) { in test_mm512_mask_xor_pd() 38 __m512d test_mm512_maskz_xor_pd (__mmask8 __U, __m512d __A, __m512d __B) { in test_mm512_maskz_xor_pd() 68 __m512d test_mm512_mask_or_pd (__m512d __W, __mmask8 __U, __m512d __A, __m512d __B) { in test_mm512_mask_or_pd() 74 __m512d test_mm512_maskz_or_pd (__mmask8 __U, __m512d __A, __m512d __B) { in test_mm512_maskz_or_pd() 104 __m512d test_mm512_mask_and_pd (__m512d __W, __mmask8 __U, __m512d __A, __m512d __B) { in test_mm512_mask_and_pd() 110 __m512d test_mm512_maskz_and_pd (__mmask8 __U, __m512d __A, __m512d __B) { in test_mm512_maskz_and_pd() 140 __m512d test_mm512_mask_andnot_pd (__m512d __W, __mmask8 __U, __m512d __A, __m512d __B) { in test_mm512_mask_andnot_pd() 146 __m512d test_mm512_maskz_andnot_pd (__mmask8 __U, __m512d __A, __m512d __B) { in test_mm512_maskz_andnot_pd() [all …]
|
D | avx512f-builtins.c | 15 __m512d test_mm512_mask_sqrt_pd (__m512d __W, __mmask8 __U, __m512d __A) in test_mm512_mask_sqrt_pd() 22 __m512d test_mm512_maskz_sqrt_pd (__mmask8 __U, __m512d __A) in test_mm512_maskz_sqrt_pd() 29 __m512d test_mm512_mask_sqrt_round_pd(__m512d __W,__mmask8 __U,__m512d __A) in test_mm512_mask_sqrt_round_pd() 36 __m512d test_mm512_maskz_sqrt_round_pd(__mmask8 __U,__m512d __A) in test_mm512_maskz_sqrt_round_pd() 99 __m512d test_mm512_mask_rsqrt14_pd (__m512d __W, __mmask8 __U, __m512d __A) in test_mm512_mask_rsqrt14_pd() 106 __m512d test_mm512_maskz_rsqrt14_pd (__mmask8 __U, __m512d __A) in test_mm512_maskz_rsqrt14_pd() 237 void test_mm512_mask_store_pd(void *p, __m512d a, __mmask8 m) in test_mm512_mask_store_pd() 250 void test_mm512_mask_storeu_epi64(void *__P, __mmask8 __U, __m512i __A) { in test_mm512_mask_storeu_epi64() 270 __m512i test_mm512_mask_loadu_epi64 (__m512i __W, __mmask8 __U, void *__P) in test_mm512_mask_loadu_epi64() 298 __m512d test_mm512_mask_loadu_pd (__m512d __W, __mmask8 __U, void *__P) in test_mm512_mask_loadu_pd() [all …]
|
D | avx512vlcd-builtins.c | 8 __m128i test_mm_broadcastmb_epi64(__mmask8 __A) { in test_mm_broadcastmb_epi64() 14 __m256i test_mm256_broadcastmb_epi64(__mmask8 __A) { in test_mm256_broadcastmb_epi64() 38 __m128i test_mm_mask_conflict_epi64(__m128i __W, __mmask8 __U, __m128i __A) { in test_mm_mask_conflict_epi64() 44 __m128i test_mm_maskz_conflict_epi64(__mmask8 __U, __m128i __A) { in test_mm_maskz_conflict_epi64() 56 __m256i test_mm256_mask_conflict_epi64(__m256i __W, __mmask8 __U, __m256i __A) { in test_mm256_mask_conflict_epi64() 62 __m256i test_mm256_maskz_conflict_epi64(__mmask8 __U, __m256i __A) { in test_mm256_maskz_conflict_epi64() 74 __m128i test_mm_mask_conflict_epi32(__m128i __W, __mmask8 __U, __m128i __A) { in test_mm_mask_conflict_epi32() 80 __m128i test_mm_maskz_conflict_epi32(__mmask8 __U, __m128i __A) { in test_mm_maskz_conflict_epi32() 92 __m256i test_mm256_mask_conflict_epi32(__m256i __W, __mmask8 __U, __m256i __A) { in test_mm256_mask_conflict_epi32() 98 __m256i test_mm256_maskz_conflict_epi32(__mmask8 __U, __m256i __A) { in test_mm256_maskz_conflict_epi32() [all …]
|
D | avx512vlbw-builtins.c | 48 __mmask8 test_mm_cmpeq_epi16_mask(__m128i __a, __m128i __b) { in test_mm_cmpeq_epi16_mask() 51 return (__mmask8)_mm_cmpeq_epi16_mask(__a, __b); in test_mm_cmpeq_epi16_mask() 54 __mmask8 test_mm_mask_cmpeq_epi16_mask(__mmask8 __u, __m128i __a, __m128i __b) { in test_mm_mask_cmpeq_epi16_mask() 58 return (__mmask8)_mm_mask_cmpeq_epi16_mask(__u, __a, __b); in test_mm_mask_cmpeq_epi16_mask() 100 __mmask8 test_mm_cmpgt_epi16_mask(__m128i __a, __m128i __b) { in test_mm_cmpgt_epi16_mask() 103 return (__mmask8)_mm_cmpgt_epi16_mask(__a, __b); in test_mm_cmpgt_epi16_mask() 106 __mmask8 test_mm_mask_cmpgt_epi16_mask(__mmask8 __u, __m128i __a, __m128i __b) { in test_mm_mask_cmpgt_epi16_mask() 110 return (__mmask8)_mm_mask_cmpgt_epi16_mask(__u, __a, __b); in test_mm_mask_cmpgt_epi16_mask() 126 __mmask8 test_mm_cmpeq_epu16_mask(__m128i __a, __m128i __b) { in test_mm_cmpeq_epu16_mask() 129 return (__mmask8)_mm_cmpeq_epu16_mask(__a, __b); in test_mm_cmpeq_epu16_mask() [all …]
|
/external/llvm-project/clang/lib/Headers/ |
D | avx512vldqintrin.h | 27 _mm256_mask_mullo_epi64(__m256i __W, __mmask8 __U, __m256i __A, __m256i __B) { in _mm256_mask_mullo_epi64() 28 return (__m256i)__builtin_ia32_selectq_256((__mmask8)__U, in _mm256_mask_mullo_epi64() 34 _mm256_maskz_mullo_epi64(__mmask8 __U, __m256i __A, __m256i __B) { in _mm256_maskz_mullo_epi64() 35 return (__m256i)__builtin_ia32_selectq_256((__mmask8)__U, in _mm256_maskz_mullo_epi64() 46 _mm_mask_mullo_epi64(__m128i __W, __mmask8 __U, __m128i __A, __m128i __B) { in _mm_mask_mullo_epi64() 47 return (__m128i)__builtin_ia32_selectq_128((__mmask8)__U, in _mm_mask_mullo_epi64() 53 _mm_maskz_mullo_epi64(__mmask8 __U, __m128i __A, __m128i __B) { in _mm_maskz_mullo_epi64() 54 return (__m128i)__builtin_ia32_selectq_128((__mmask8)__U, in _mm_maskz_mullo_epi64() 60 _mm256_mask_andnot_pd(__m256d __W, __mmask8 __U, __m256d __A, __m256d __B) { in _mm256_mask_andnot_pd() 61 return (__m256d)__builtin_ia32_selectpd_256((__mmask8)__U, in _mm256_mask_andnot_pd() [all …]
|
D | avx512vlintrin.h | 227 _mm256_mask_add_epi32(__m256i __W, __mmask8 __U, __m256i __A, __m256i __B) in _mm256_mask_add_epi32() 229 return (__m256i)__builtin_ia32_selectd_256((__mmask8)__U, in _mm256_mask_add_epi32() 235 _mm256_maskz_add_epi32(__mmask8 __U, __m256i __A, __m256i __B) in _mm256_maskz_add_epi32() 237 return (__m256i)__builtin_ia32_selectd_256((__mmask8)__U, in _mm256_maskz_add_epi32() 243 _mm256_mask_add_epi64(__m256i __W, __mmask8 __U, __m256i __A, __m256i __B) in _mm256_mask_add_epi64() 245 return (__m256i)__builtin_ia32_selectq_256((__mmask8)__U, in _mm256_mask_add_epi64() 251 _mm256_maskz_add_epi64(__mmask8 __U, __m256i __A, __m256i __B) in _mm256_maskz_add_epi64() 253 return (__m256i)__builtin_ia32_selectq_256((__mmask8)__U, in _mm256_maskz_add_epi64() 259 _mm256_mask_sub_epi32(__m256i __W, __mmask8 __U, __m256i __A, __m256i __B) in _mm256_mask_sub_epi32() 261 return (__m256i)__builtin_ia32_selectd_256((__mmask8)__U, in _mm256_mask_sub_epi32() [all …]
|
D | avx512dqintrin.h | 21 static __inline __mmask8 __DEFAULT_FN_ATTRS 22 _knot_mask8(__mmask8 __M) in _knot_mask8() 27 static __inline__ __mmask8 __DEFAULT_FN_ATTRS 28 _kand_mask8(__mmask8 __A, __mmask8 __B) in _kand_mask8() 30 return (__mmask8)__builtin_ia32_kandqi((__mmask8)__A, (__mmask8)__B); in _kand_mask8() 33 static __inline__ __mmask8 __DEFAULT_FN_ATTRS 34 _kandn_mask8(__mmask8 __A, __mmask8 __B) in _kandn_mask8() 36 return (__mmask8)__builtin_ia32_kandnqi((__mmask8)__A, (__mmask8)__B); in _kandn_mask8() 39 static __inline__ __mmask8 __DEFAULT_FN_ATTRS 40 _kor_mask8(__mmask8 __A, __mmask8 __B) in _kor_mask8() [all …]
|
D | avx512vlcdintrin.h | 22 _mm_broadcastmb_epi64 (__mmask8 __A) in _mm_broadcastmb_epi64() 28 _mm256_broadcastmb_epi64 (__mmask8 __A) in _mm256_broadcastmb_epi64() 53 _mm_mask_conflict_epi64 (__m128i __W, __mmask8 __U, __m128i __A) in _mm_mask_conflict_epi64() 55 return (__m128i)__builtin_ia32_selectq_128((__mmask8)__U, in _mm_mask_conflict_epi64() 61 _mm_maskz_conflict_epi64 (__mmask8 __U, __m128i __A) in _mm_maskz_conflict_epi64() 63 return (__m128i)__builtin_ia32_selectq_128((__mmask8)__U, in _mm_maskz_conflict_epi64() 75 _mm256_mask_conflict_epi64 (__m256i __W, __mmask8 __U, __m256i __A) in _mm256_mask_conflict_epi64() 77 return (__m256i)__builtin_ia32_selectq_256((__mmask8)__U, in _mm256_mask_conflict_epi64() 83 _mm256_maskz_conflict_epi64 (__mmask8 __U, __m256i __A) in _mm256_maskz_conflict_epi64() 85 return (__m256i)__builtin_ia32_selectq_256((__mmask8)__U, in _mm256_maskz_conflict_epi64() [all …]
|
D | avx512vlbf16intrin.h | 61 _mm_mask_cvtne2ps_pbh(__m128bh __W, __mmask8 __U, __m128 __A, __m128 __B) { in _mm_mask_cvtne2ps_pbh() 62 return (__m128bh)__builtin_ia32_selectw_128((__mmask8)__U, in _mm_mask_cvtne2ps_pbh() 83 _mm_maskz_cvtne2ps_pbh(__mmask8 __U, __m128 __A, __m128 __B) { in _mm_maskz_cvtne2ps_pbh() 84 return (__m128bh)__builtin_ia32_selectw_128((__mmask8)__U, in _mm_maskz_cvtne2ps_pbh() 167 (__mmask8)-1); in _mm_cvtneps_pbh() 186 _mm_mask_cvtneps_pbh(__m128bh __W, __mmask8 __U, __m128 __A) { in _mm_mask_cvtneps_pbh() 189 (__mmask8)__U); in _mm_mask_cvtneps_pbh() 206 _mm_maskz_cvtneps_pbh(__mmask8 __U, __m128 __A) { in _mm_maskz_cvtneps_pbh() 209 (__mmask8)__U); in _mm_maskz_cvtneps_pbh() 225 (__mmask8)-1); in _mm256_cvtneps_pbh() [all …]
|
D | avx512fintrin.h | 37 typedef unsigned char __mmask8; typedef 235 _mm512_mask_broadcastq_epi64 (__m512i __O, __mmask8 __M, __m128i __A) in _mm512_mask_broadcastq_epi64() 244 _mm512_maskz_broadcastq_epi64 (__mmask8 __M, __m128i __A) in _mm512_maskz_broadcastq_epi64() 327 _mm512_maskz_set1_epi64(__mmask8 __M, long long __A) in _mm512_maskz_set1_epi64() 653 _mm512_mask_and_epi64(__m512i __src, __mmask8 __k, __m512i __a, __m512i __b) in _mm512_mask_and_epi64() 655 return (__m512i) __builtin_ia32_selectq_512 ((__mmask8) __k, in _mm512_mask_and_epi64() 661 _mm512_maskz_and_epi64(__mmask8 __k, __m512i __a, __m512i __b) in _mm512_maskz_and_epi64() 701 _mm512_mask_andnot_epi64(__m512i __W, __mmask8 __U, __m512i __A, __m512i __B) in _mm512_mask_andnot_epi64() 703 return (__m512i)__builtin_ia32_selectq_512((__mmask8)__U, in _mm512_mask_andnot_epi64() 709 _mm512_maskz_andnot_epi64(__mmask8 __U, __m512i __A, __m512i __B) in _mm512_maskz_andnot_epi64() [all …]
|