Home
last modified time | relevance | path

Searched refs:__m512d (Results 1 – 25 of 28) sorted by relevance

12

/external/llvm-project/clang/test/CodeGen/X86/
Davx512f-builtins-constrained-cmp.c401 __mmask8 test_mm512_cmp_round_pd_mask(__m512d a, __m512d b) { in test_mm512_cmp_round_pd_mask()
407 __mmask8 test_mm512_mask_cmp_round_pd_mask(__mmask8 m, __m512d a, __m512d b) { in test_mm512_mask_cmp_round_pd_mask()
413 __mmask8 test_mm512_cmp_pd_mask_eq_oq(__m512d a, __m512d b) { in test_mm512_cmp_pd_mask_eq_oq()
419 __mmask8 test_mm512_cmp_pd_mask_lt_os(__m512d a, __m512d b) { in test_mm512_cmp_pd_mask_lt_os()
425 __mmask8 test_mm512_cmp_pd_mask_le_os(__m512d a, __m512d b) { in test_mm512_cmp_pd_mask_le_os()
431 __mmask8 test_mm512_cmp_pd_mask_unord_q(__m512d a, __m512d b) { in test_mm512_cmp_pd_mask_unord_q()
437 __mmask8 test_mm512_cmp_pd_mask_neq_uq(__m512d a, __m512d b) { in test_mm512_cmp_pd_mask_neq_uq()
443 __mmask8 test_mm512_cmp_pd_mask_nlt_us(__m512d a, __m512d b) { in test_mm512_cmp_pd_mask_nlt_us()
449 __mmask8 test_mm512_cmp_pd_mask_nle_us(__m512d a, __m512d b) { in test_mm512_cmp_pd_mask_nle_us()
455 __mmask8 test_mm512_cmp_pd_mask_ord_q(__m512d a, __m512d b) { in test_mm512_cmp_pd_mask_ord_q()
[all …]
Davx512dq-builtins.c272 __m512d test_mm512_xor_pd (__m512d __A, __m512d __B) { in test_mm512_xor_pd()
275 return (__m512d) _mm512_xor_pd(__A, __B); in test_mm512_xor_pd()
278 __m512d test_mm512_mask_xor_pd (__m512d __W, __mmask8 __U, __m512d __A, __m512d __B) { in test_mm512_mask_xor_pd()
283 return (__m512d) _mm512_mask_xor_pd(__W, __U, __A, __B); in test_mm512_mask_xor_pd()
286 __m512d test_mm512_maskz_xor_pd (__mmask8 __U, __m512d __A, __m512d __B) { in test_mm512_maskz_xor_pd()
291 return (__m512d) _mm512_maskz_xor_pd(__U, __A, __B); in test_mm512_maskz_xor_pd()
316 __m512d test_mm512_or_pd (__m512d __A, __m512d __B) { in test_mm512_or_pd()
319 return (__m512d) _mm512_or_pd(__A, __B); in test_mm512_or_pd()
322 __m512d test_mm512_mask_or_pd (__m512d __W, __mmask8 __U, __m512d __A, __m512d __B) { in test_mm512_mask_or_pd()
327 return (__m512d) _mm512_mask_or_pd(__W, __U, __A, __B); in test_mm512_mask_or_pd()
[all …]
Davx512er-builtins.c6 __m512d test_mm512_rsqrt28_round_pd(__m512d a) { in test_mm512_rsqrt28_round_pd()
12 __m512d test_mm512_mask_rsqrt28_round_pd(__m512d s, __mmask8 m, __m512d a) { in test_mm512_mask_rsqrt28_round_pd()
18 __m512d test_mm512_maskz_rsqrt28_round_pd(__mmask8 m, __m512d a) { in test_mm512_maskz_rsqrt28_round_pd()
24 __m512d test_mm512_rsqrt28_pd(__m512d a) { in test_mm512_rsqrt28_pd()
30 __m512d test_mm512_mask_rsqrt28_pd(__m512d s, __mmask8 m, __m512d a) { in test_mm512_mask_rsqrt28_pd()
36 __m512d test_mm512_maskz_rsqrt28_pd(__mmask8 m, __m512d a) { in test_mm512_maskz_rsqrt28_pd()
132 __m512d test_mm512_rcp28_round_pd(__m512d a) { in test_mm512_rcp28_round_pd()
138 __m512d test_mm512_mask_rcp28_round_pd(__m512d s, __mmask8 m, __m512d a) { in test_mm512_mask_rcp28_round_pd()
144 __m512d test_mm512_maskz_rcp28_round_pd(__mmask8 m, __m512d a) { in test_mm512_maskz_rcp28_round_pd()
150 __m512d test_mm512_rcp28_pd(__m512d a) { in test_mm512_rcp28_pd()
[all …]
Davx512f-builtins.c6 __m512d test_mm512_sqrt_pd(__m512d a) in test_mm512_sqrt_pd()
13 __m512d test_mm512_mask_sqrt_pd (__m512d __W, __mmask8 __U, __m512d __A) in test_mm512_mask_sqrt_pd()
22 __m512d test_mm512_maskz_sqrt_pd (__mmask8 __U, __m512d __A) in test_mm512_maskz_sqrt_pd()
31 __m512d test_mm512_mask_sqrt_round_pd(__m512d __W,__mmask8 __U,__m512d __A) in test_mm512_mask_sqrt_round_pd()
40 __m512d test_mm512_maskz_sqrt_round_pd(__mmask8 __U,__m512d __A) in test_mm512_maskz_sqrt_round_pd()
49 __m512d test_mm512_sqrt_round_pd(__m512d __A) in test_mm512_sqrt_round_pd()
106 __m512d test_mm512_rsqrt14_pd(__m512d a) in test_mm512_rsqrt14_pd()
113 __m512d test_mm512_mask_rsqrt14_pd (__m512d __W, __mmask8 __U, __m512d __A) in test_mm512_mask_rsqrt14_pd()
120 __m512d test_mm512_maskz_rsqrt14_pd (__mmask8 __U, __m512d __A) in test_mm512_maskz_rsqrt14_pd()
155 __m512d test_mm512_add_pd(__m512d a, __m512d b) in test_mm512_add_pd()
[all …]
Davx512f-builtins-constrained.c11 __m512d test_mm512_sqrt_pd(__m512d a) in test_mm512_sqrt_pd()
20 __m512d test_mm512_mask_sqrt_pd (__m512d __W, __mmask8 __U, __m512d __A) in test_mm512_mask_sqrt_pd()
31 __m512d test_mm512_maskz_sqrt_pd (__mmask8 __U, __m512d __A) in test_mm512_maskz_sqrt_pd()
Davx512-reduceMinMaxIntrin.c17 double test_mm512_reduce_max_pd(__m512d __W){ in test_mm512_reduce_max_pd()
43 double test_mm512_reduce_min_pd(__m512d __W){ in test_mm512_reduce_min_pd()
73 double test_mm512_mask_reduce_max_pd(__mmask8 __M, __m512d __W){ in test_mm512_mask_reduce_max_pd()
105 double test_mm512_mask_reduce_min_pd(__mmask8 __M, __m512d __W){ in test_mm512_mask_reduce_min_pd()
Davx512-reduceIntrin.c116 double test_mm512_reduce_add_pd(__m512d __W){ in test_mm512_reduce_add_pd()
130 double test_mm512_reduce_mul_pd(__m512d __W){ in test_mm512_reduce_mul_pd()
182 double test_mm512_mask_reduce_add_pd(__mmask8 __M, __m512d __W){ in test_mm512_mask_reduce_add_pd()
198 double test_mm512_mask_reduce_mul_pd(__mmask8 __M, __m512d __W){ in test_mm512_mask_reduce_mul_pd()
Dx86-nontemporal.c71 void test_mm512_stream_pd(double* A, __m512d B) { in test_mm512_stream_pd()
Dx86-vec-struct-packing.c117 __m512d b;
/external/clang/test/CodeGen/
Davx512dq-builtins.c26 __m512d test_mm512_xor_pd (__m512d __A, __m512d __B) { in test_mm512_xor_pd()
29 return (__m512d) _mm512_xor_pd(__A, __B); in test_mm512_xor_pd()
32 __m512d test_mm512_mask_xor_pd (__m512d __W, __mmask8 __U, __m512d __A, __m512d __B) { in test_mm512_mask_xor_pd()
35 return (__m512d) _mm512_mask_xor_pd(__W, __U, __A, __B); in test_mm512_mask_xor_pd()
38 __m512d test_mm512_maskz_xor_pd (__mmask8 __U, __m512d __A, __m512d __B) { in test_mm512_maskz_xor_pd()
41 return (__m512d) _mm512_maskz_xor_pd(__U, __A, __B); in test_mm512_maskz_xor_pd()
62 __m512d test_mm512_or_pd (__m512d __A, __m512d __B) { in test_mm512_or_pd()
65 return (__m512d) _mm512_or_pd(__A, __B); in test_mm512_or_pd()
68 __m512d test_mm512_mask_or_pd (__m512d __W, __mmask8 __U, __m512d __A, __m512d __B) { in test_mm512_mask_or_pd()
71 return (__m512d) _mm512_mask_or_pd(__W, __U, __A, __B); in test_mm512_mask_or_pd()
[all …]
Davx512er-builtins.c8 __m512d test_mm512_rsqrt28_round_pd(__m512d a) { in test_mm512_rsqrt28_round_pd()
14 __m512d test_mm512_mask_rsqrt28_round_pd(__m512d s, __mmask8 m, __m512d a) { in test_mm512_mask_rsqrt28_round_pd()
20 __m512d test_mm512_maskz_rsqrt28_round_pd(__mmask8 m, __m512d a) { in test_mm512_maskz_rsqrt28_round_pd()
26 __m512d test_mm512_rsqrt28_pd(__m512d a) { in test_mm512_rsqrt28_pd()
32 __m512d test_mm512_mask_rsqrt28_pd(__m512d s, __mmask8 m, __m512d a) { in test_mm512_mask_rsqrt28_pd()
38 __m512d test_mm512_maskz_rsqrt28_pd(__mmask8 m, __m512d a) { in test_mm512_maskz_rsqrt28_pd()
134 __m512d test_mm512_rcp28_round_pd(__m512d a) { in test_mm512_rcp28_round_pd()
140 __m512d test_mm512_mask_rcp28_round_pd(__m512d s, __mmask8 m, __m512d a) { in test_mm512_mask_rcp28_round_pd()
146 __m512d test_mm512_maskz_rcp28_round_pd(__mmask8 m, __m512d a) { in test_mm512_maskz_rcp28_round_pd()
152 __m512d test_mm512_rcp28_pd(__m512d a) { in test_mm512_rcp28_pd()
[all …]
Davx512f-builtins.c8 __m512d test_mm512_sqrt_pd(__m512d a) in test_mm512_sqrt_pd()
15 __m512d test_mm512_mask_sqrt_pd (__m512d __W, __mmask8 __U, __m512d __A) in test_mm512_mask_sqrt_pd()
22 __m512d test_mm512_maskz_sqrt_pd (__mmask8 __U, __m512d __A) in test_mm512_maskz_sqrt_pd()
29 __m512d test_mm512_mask_sqrt_round_pd(__m512d __W,__mmask8 __U,__m512d __A) in test_mm512_mask_sqrt_round_pd()
36 __m512d test_mm512_maskz_sqrt_round_pd(__mmask8 __U,__m512d __A) in test_mm512_maskz_sqrt_round_pd()
43 __m512d test_mm512_sqrt_round_pd(__m512d __A) in test_mm512_sqrt_round_pd()
92 __m512d test_mm512_rsqrt14_pd(__m512d a) in test_mm512_rsqrt14_pd()
99 __m512d test_mm512_mask_rsqrt14_pd (__m512d __W, __mmask8 __U, __m512d __A) in test_mm512_mask_rsqrt14_pd()
106 __m512d test_mm512_maskz_rsqrt14_pd (__mmask8 __U, __m512d __A) in test_mm512_maskz_rsqrt14_pd()
141 __m512d test_mm512_add_pd(__m512d a, __m512d b) in test_mm512_add_pd()
[all …]
/external/clang/lib/Headers/
Davx512dqintrin.h56 static __inline__ __m512d __DEFAULT_FN_ATTRS
57 _mm512_xor_pd (__m512d __A, __m512d __B) { in _mm512_xor_pd()
58 return (__m512d) ((__v8du) __A ^ (__v8du) __B); in _mm512_xor_pd()
61 static __inline__ __m512d __DEFAULT_FN_ATTRS
62 _mm512_mask_xor_pd (__m512d __W, __mmask8 __U, __m512d __A, __m512d __B) { in _mm512_mask_xor_pd()
63 return (__m512d) __builtin_ia32_xorpd512_mask ((__v8df) __A, in _mm512_mask_xor_pd()
69 static __inline__ __m512d __DEFAULT_FN_ATTRS
70 _mm512_maskz_xor_pd (__mmask8 __U, __m512d __A, __m512d __B) { in _mm512_maskz_xor_pd()
71 return (__m512d) __builtin_ia32_xorpd512_mask ((__v8df) __A, in _mm512_maskz_xor_pd()
100 static __inline__ __m512d __DEFAULT_FN_ATTRS
[all …]
Davx512fintrin.h44 typedef double __m512d __attribute__((__vector_size__(64))); typedef
175 static __inline__ __m512d __DEFAULT_FN_ATTRS
178 return (__m512d)__builtin_ia32_undef512(); in _mm512_undefined_pd()
282 static __inline __m512d __DEFAULT_FN_ATTRS
285 return (__m512d){ 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0 }; in _mm512_setzero_pd()
295 static __inline __m512d __DEFAULT_FN_ATTRS
298 return (__m512d){ __w, __w, __w, __w, __w, __w, __w, __w }; in _mm512_set1_pd()
360 static __inline __m512d __DEFAULT_FN_ATTRS
363 return (__m512d) in _mm512_set4_pd()
387 static __inline__ __m512d __DEFAULT_FN_ATTRS
[all …]
Davx512erintrin.h32 (__m512d)__builtin_ia32_exp2pd_mask((__v8df)(__m512d)(A), \
37 (__m512d)__builtin_ia32_exp2pd_mask((__v8df)(__m512d)(A), \
38 (__v8df)(__m512d)(S), (__mmask8)(M), \
42 (__m512d)__builtin_ia32_exp2pd_mask((__v8df)(__m512d)(A), \
81 (__m512d)__builtin_ia32_rsqrt28pd_mask((__v8df)(__m512d)(A), \
86 (__m512d)__builtin_ia32_rsqrt28pd_mask((__v8df)(__m512d)(A), \
87 (__v8df)(__m512d)(S), (__mmask8)(M), \
91 (__m512d)__builtin_ia32_rsqrt28pd_mask((__v8df)(__m512d)(A), \
184 (__m512d)__builtin_ia32_rcp28pd_mask((__v8df)(__m512d)(A), \
189 (__m512d)__builtin_ia32_rcp28pd_mask((__v8df)(__m512d)(A), \
[all …]
/external/llvm-project/clang/lib/Headers/
Davx512dqintrin.h168 static __inline__ __m512d __DEFAULT_FN_ATTRS512
169 _mm512_xor_pd(__m512d __A, __m512d __B) { in _mm512_xor_pd()
170 return (__m512d)((__v8du)__A ^ (__v8du)__B); in _mm512_xor_pd()
173 static __inline__ __m512d __DEFAULT_FN_ATTRS512
174 _mm512_mask_xor_pd(__m512d __W, __mmask8 __U, __m512d __A, __m512d __B) { in _mm512_mask_xor_pd()
175 return (__m512d)__builtin_ia32_selectpd_512((__mmask8)__U, in _mm512_mask_xor_pd()
180 static __inline__ __m512d __DEFAULT_FN_ATTRS512
181 _mm512_maskz_xor_pd(__mmask8 __U, __m512d __A, __m512d __B) { in _mm512_maskz_xor_pd()
182 return (__m512d)__builtin_ia32_selectpd_512((__mmask8)__U, in _mm512_maskz_xor_pd()
206 static __inline__ __m512d __DEFAULT_FN_ATTRS512
[all …]
Davx512fintrin.h30 typedef double __m512d __attribute__((__vector_size__(64), __aligned__(64))); typedef
180 static __inline__ __m512d __DEFAULT_FN_ATTRS512
183 return (__m512d)__builtin_ia32_undef512(); in _mm512_undefined_pd()
261 static __inline __m512d __DEFAULT_FN_ATTRS512
264 return __extension__ (__m512d){ 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0 }; in _mm512_setzero_pd()
274 static __inline __m512d __DEFAULT_FN_ATTRS512
277 return __extension__ (__m512d){ __w, __w, __w, __w, __w, __w, __w, __w }; in _mm512_set1_pd()
357 static __inline __m512d __DEFAULT_FN_ATTRS512
360 return __extension__ (__m512d) in _mm512_set4_pd()
384 static __inline__ __m512d __DEFAULT_FN_ATTRS512
[all …]
Davx512erintrin.h18 (__m512d)__builtin_ia32_exp2pd_mask((__v8df)(__m512d)(A), \
23 (__m512d)__builtin_ia32_exp2pd_mask((__v8df)(__m512d)(A), \
24 (__v8df)(__m512d)(S), (__mmask8)(M), \
28 (__m512d)__builtin_ia32_exp2pd_mask((__v8df)(__m512d)(A), \
67 (__m512d)__builtin_ia32_rsqrt28pd_mask((__v8df)(__m512d)(A), \
72 (__m512d)__builtin_ia32_rsqrt28pd_mask((__v8df)(__m512d)(A), \
73 (__v8df)(__m512d)(S), (__mmask8)(M), \
77 (__m512d)__builtin_ia32_rsqrt28pd_mask((__v8df)(__m512d)(A), \
170 (__m512d)__builtin_ia32_rcp28pd_mask((__v8df)(__m512d)(A), \
175 (__m512d)__builtin_ia32_rcp28pd_mask((__v8df)(__m512d)(A), \
[all …]
/external/clang/test/Sema/
Dbuiltins-x86.c8 typedef double __m512d __attribute__((__vector_size__(64))); typedef
29 __mmask16 test__builtin_ia32_cmpps512_mask(__m512d __a, __m512d __b) { in test__builtin_ia32_cmpps512_mask()
33 __mmask8 test__builtin_ia32_cmppd512_mask(__m512d __a, __m512d __b) { in test__builtin_ia32_cmppd512_mask()
/external/mesa3d/src/gallium/drivers/swr/rasterizer/common/
Dsimdlib_types.hpp327 struct __m512d struct
421 SIMDINLINE Double(__m512d in) : v(in) {} in Double()
428 SIMDINLINE Double& SIMDCALL operator=(__m512d in) in operator =()
444 SIMDINLINE SIMDCALL operator __m512d() const { return v; } in operator __m512d()
446 SIMDALIGN(__m512d, SIMD_ALIGNMENT_BYTES) v;
Dsimdlib_128_avx512.inl41 static SIMDINLINE __m512d __conv(Double r)
53 static SIMDINLINE Double __conv(__m512d r)
Dsimdlib_256_avx512.inl41 static SIMDINLINE __m512d __conv(Double r)
53 static SIMDINLINE Double __conv(__m512d r)
/external/llvm-project/clang/test/Sema/
Dbuiltins-x86.c13 typedef double __m512d __attribute__((__vector_size__(64))); typedef
44 __mmask8 test__builtin_ia32_cmppd512_mask(__m512d __a, __m512d __b) { in test__builtin_ia32_cmppd512_mask()
/external/eigen/Eigen/src/Core/arch/AVX512/
DPacketMath.h33 typedef __m512d Packet8d;
44 struct is_arithmetic<__m512d> {
656 return (__m512d)_mm512_and_si512((__m512i)a,
872 __m512d final_output = _mm512_insertf64x4(final_output, final_0, 0);
1208 __m512d T0 = _mm512_shuffle_pd(kernel.packet[0], kernel.packet[1], 0);
1209 __m512d T1 = _mm512_shuffle_pd(kernel.packet[0], kernel.packet[1], 0xff);
1210 __m512d T2 = _mm512_shuffle_pd(kernel.packet[2], kernel.packet[3], 0);
1211 __m512d T3 = _mm512_shuffle_pd(kernel.packet[2], kernel.packet[3], 0xff);
1240 __m512d T0 = _mm512_unpacklo_pd(kernel.packet[0], kernel.packet[1]);
1241 __m512d T1 = _mm512_unpackhi_pd(kernel.packet[0], kernel.packet[1]);
[all …]
/external/python/pycparser/utils/fake_libc_include/
D_fake_defines.h218 #define __m512d int macro

12