Home
last modified time | relevance | path

Searched refs:__a (Results 1 – 25 of 108) sorted by relevance

12345

/external/clang/lib/Headers/
Daltivec.h39 static vector signed char __ATTRS_o_ai vec_perm(vector signed char __a,
43 static vector unsigned char __ATTRS_o_ai vec_perm(vector unsigned char __a,
47 static vector bool char __ATTRS_o_ai vec_perm(vector bool char __a,
51 static vector short __ATTRS_o_ai vec_perm(vector signed short __a,
55 static vector unsigned short __ATTRS_o_ai vec_perm(vector unsigned short __a,
59 static vector bool short __ATTRS_o_ai vec_perm(vector bool short __a,
63 static vector pixel __ATTRS_o_ai vec_perm(vector pixel __a, vector pixel __b,
66 static vector int __ATTRS_o_ai vec_perm(vector signed int __a,
70 static vector unsigned int __ATTRS_o_ai vec_perm(vector unsigned int __a,
74 static vector bool int __ATTRS_o_ai vec_perm(vector bool int __a,
[all …]
Dvecintrin.h361 vec_perm(vector signed char __a, vector signed char __b, in vec_perm() argument
364 (vector unsigned char)__a, (vector unsigned char)__b, __c); in vec_perm()
368 vec_perm(vector unsigned char __a, vector unsigned char __b, in vec_perm() argument
371 (vector unsigned char)__a, (vector unsigned char)__b, __c); in vec_perm()
375 vec_perm(vector bool char __a, vector bool char __b, in vec_perm() argument
378 (vector unsigned char)__a, (vector unsigned char)__b, __c); in vec_perm()
382 vec_perm(vector signed short __a, vector signed short __b, in vec_perm() argument
385 (vector unsigned char)__a, (vector unsigned char)__b, __c); in vec_perm()
389 vec_perm(vector unsigned short __a, vector unsigned short __b, in vec_perm() argument
392 (vector unsigned char)__a, (vector unsigned char)__b, __c); in vec_perm()
[all …]
Dxmmintrin.h43 _mm_add_ss(__m128 __a, __m128 __b) in _mm_add_ss() argument
45 __a[0] += __b[0]; in _mm_add_ss()
46 return __a; in _mm_add_ss()
50 _mm_add_ps(__m128 __a, __m128 __b) in _mm_add_ps() argument
52 return __a + __b; in _mm_add_ps()
56 _mm_sub_ss(__m128 __a, __m128 __b) in _mm_sub_ss() argument
58 __a[0] -= __b[0]; in _mm_sub_ss()
59 return __a; in _mm_sub_ss()
63 _mm_sub_ps(__m128 __a, __m128 __b) in _mm_sub_ps() argument
65 return __a - __b; in _mm_sub_ps()
[all …]
Demmintrin.h48 _mm_add_sd(__m128d __a, __m128d __b) in _mm_add_sd() argument
50 __a[0] += __b[0]; in _mm_add_sd()
51 return __a; in _mm_add_sd()
55 _mm_add_pd(__m128d __a, __m128d __b) in _mm_add_pd() argument
57 return __a + __b; in _mm_add_pd()
61 _mm_sub_sd(__m128d __a, __m128d __b) in _mm_sub_sd() argument
63 __a[0] -= __b[0]; in _mm_sub_sd()
64 return __a; in _mm_sub_sd()
68 _mm_sub_pd(__m128d __a, __m128d __b) in _mm_sub_pd() argument
70 return __a - __b; in _mm_sub_pd()
[all …]
Davxintrin.h51 _mm256_add_pd(__m256d __a, __m256d __b) in _mm256_add_pd() argument
53 return __a+__b; in _mm256_add_pd()
57 _mm256_add_ps(__m256 __a, __m256 __b) in _mm256_add_ps() argument
59 return __a+__b; in _mm256_add_ps()
63 _mm256_sub_pd(__m256d __a, __m256d __b) in _mm256_sub_pd() argument
65 return __a-__b; in _mm256_sub_pd()
69 _mm256_sub_ps(__m256 __a, __m256 __b) in _mm256_sub_ps() argument
71 return __a-__b; in _mm256_sub_ps()
75 _mm256_addsub_pd(__m256d __a, __m256d __b) in _mm256_addsub_pd() argument
77 return (__m256d)__builtin_ia32_addsubpd256((__v4df)__a, (__v4df)__b); in _mm256_addsub_pd()
[all …]
Dtmmintrin.h33 _mm_abs_pi8(__m64 __a) in _mm_abs_pi8() argument
35 return (__m64)__builtin_ia32_pabsb((__v8qi)__a); in _mm_abs_pi8()
39 _mm_abs_epi8(__m128i __a) in _mm_abs_epi8() argument
41 return (__m128i)__builtin_ia32_pabsb128((__v16qi)__a); in _mm_abs_epi8()
45 _mm_abs_pi16(__m64 __a) in _mm_abs_pi16() argument
47 return (__m64)__builtin_ia32_pabsw((__v4hi)__a); in _mm_abs_pi16()
51 _mm_abs_epi16(__m128i __a) in _mm_abs_epi16() argument
53 return (__m128i)__builtin_ia32_pabsw128((__v8hi)__a); in _mm_abs_epi16()
57 _mm_abs_pi32(__m64 __a) in _mm_abs_pi32() argument
59 return (__m64)__builtin_ia32_pabsd((__v2si)__a); in _mm_abs_pi32()
[all …]
Davx2intrin.h38 _mm256_abs_epi8(__m256i __a) in _mm256_abs_epi8() argument
40 return (__m256i)__builtin_ia32_pabsb256((__v32qi)__a); in _mm256_abs_epi8()
44 _mm256_abs_epi16(__m256i __a) in _mm256_abs_epi16() argument
46 return (__m256i)__builtin_ia32_pabsw256((__v16hi)__a); in _mm256_abs_epi16()
50 _mm256_abs_epi32(__m256i __a) in _mm256_abs_epi32() argument
52 return (__m256i)__builtin_ia32_pabsd256((__v8si)__a); in _mm256_abs_epi32()
56 _mm256_packs_epi16(__m256i __a, __m256i __b) in _mm256_packs_epi16() argument
58 return (__m256i)__builtin_ia32_packsswb256((__v16hi)__a, (__v16hi)__b); in _mm256_packs_epi16()
62 _mm256_packs_epi32(__m256i __a, __m256i __b) in _mm256_packs_epi32() argument
64 return (__m256i)__builtin_ia32_packssdw256((__v8si)__a, (__v8si)__b); in _mm256_packs_epi32()
[all …]
Dpmmintrin.h39 _mm_addsub_ps(__m128 __a, __m128 __b) in _mm_addsub_ps() argument
41 return __builtin_ia32_addsubps(__a, __b); in _mm_addsub_ps()
45 _mm_hadd_ps(__m128 __a, __m128 __b) in _mm_hadd_ps() argument
47 return __builtin_ia32_haddps(__a, __b); in _mm_hadd_ps()
51 _mm_hsub_ps(__m128 __a, __m128 __b) in _mm_hsub_ps() argument
53 return __builtin_ia32_hsubps(__a, __b); in _mm_hsub_ps()
57 _mm_movehdup_ps(__m128 __a) in _mm_movehdup_ps() argument
59 return __builtin_shufflevector(__a, __a, 1, 1, 3, 3); in _mm_movehdup_ps()
63 _mm_moveldup_ps(__m128 __a) in _mm_moveldup_ps() argument
65 return __builtin_shufflevector(__a, __a, 0, 0, 2, 2); in _mm_moveldup_ps()
[all …]
Davx512vlbwintrin.h37 _mm_cmpeq_epi8_mask(__m128i __a, __m128i __b) { in _mm_cmpeq_epi8_mask() argument
38 return (__mmask16)__builtin_ia32_pcmpeqb128_mask((__v16qi)__a, (__v16qi)__b, in _mm_cmpeq_epi8_mask()
43 _mm_mask_cmpeq_epi8_mask(__mmask16 __u, __m128i __a, __m128i __b) { in _mm_mask_cmpeq_epi8_mask() argument
44 return (__mmask16)__builtin_ia32_pcmpeqb128_mask((__v16qi)__a, (__v16qi)__b, in _mm_mask_cmpeq_epi8_mask()
49 _mm_cmpeq_epu8_mask(__m128i __a, __m128i __b) { in _mm_cmpeq_epu8_mask() argument
50 return (__mmask16)__builtin_ia32_ucmpb128_mask((__v16qi)__a, (__v16qi)__b, 0, in _mm_cmpeq_epu8_mask()
55 _mm_mask_cmpeq_epu8_mask(__mmask16 __u, __m128i __a, __m128i __b) { in _mm_mask_cmpeq_epu8_mask() argument
56 return (__mmask16)__builtin_ia32_ucmpb128_mask((__v16qi)__a, (__v16qi)__b, 0, in _mm_mask_cmpeq_epu8_mask()
61 _mm256_cmpeq_epi8_mask(__m256i __a, __m256i __b) { in _mm256_cmpeq_epi8_mask() argument
62 return (__mmask32)__builtin_ia32_pcmpeqb256_mask((__v32qi)__a, (__v32qi)__b, in _mm256_cmpeq_epi8_mask()
[all …]
Davx512bwintrin.h62 _mm512_cmpeq_epi8_mask(__m512i __a, __m512i __b) { in _mm512_cmpeq_epi8_mask() argument
63 return (__mmask64)__builtin_ia32_pcmpeqb512_mask((__v64qi)__a, (__v64qi)__b, in _mm512_cmpeq_epi8_mask()
68 _mm512_mask_cmpeq_epi8_mask(__mmask64 __u, __m512i __a, __m512i __b) { in _mm512_mask_cmpeq_epi8_mask() argument
69 return (__mmask64)__builtin_ia32_pcmpeqb512_mask((__v64qi)__a, (__v64qi)__b, in _mm512_mask_cmpeq_epi8_mask()
74 _mm512_cmpeq_epu8_mask(__m512i __a, __m512i __b) { in _mm512_cmpeq_epu8_mask() argument
75 return (__mmask64)__builtin_ia32_ucmpb512_mask((__v64qi)__a, (__v64qi)__b, 0, in _mm512_cmpeq_epu8_mask()
80 _mm512_mask_cmpeq_epu8_mask(__mmask64 __u, __m512i __a, __m512i __b) { in _mm512_mask_cmpeq_epu8_mask() argument
81 return (__mmask64)__builtin_ia32_ucmpb512_mask((__v64qi)__a, (__v64qi)__b, 0, in _mm512_mask_cmpeq_epu8_mask()
86 _mm512_cmpeq_epi16_mask(__m512i __a, __m512i __b) { in _mm512_cmpeq_epi16_mask() argument
87 return (__mmask32)__builtin_ia32_pcmpeqw512_mask((__v32hi)__a, (__v32hi)__b, in _mm512_cmpeq_epi16_mask()
[all …]
/external/clang/test/CodeGen/
Dsse-builtins.c162 void test_extract_epi16(__m128i __a) { in test_extract_epi16() argument
166 _mm_extract_epi16(__a, 8); in test_extract_epi16()
169 __m128 test_mm_cmpeq_ss(__m128 __a, __m128 __b) { in test_mm_cmpeq_ss() argument
172 return _mm_cmpeq_ss(__a, __b); in test_mm_cmpeq_ss()
175 __m128 test_mm_cmplt_ss(__m128 __a, __m128 __b) { in test_mm_cmplt_ss() argument
178 return _mm_cmplt_ss(__a, __b); in test_mm_cmplt_ss()
181 __m128 test_mm_cmple_ss(__m128 __a, __m128 __b) { in test_mm_cmple_ss() argument
184 return _mm_cmple_ss(__a, __b); in test_mm_cmple_ss()
187 __m128 test_mm_cmpunord_ss(__m128 __a, __m128 __b) { in test_mm_cmpunord_ss() argument
190 return _mm_cmpunord_ss(__a, __b); in test_mm_cmpunord_ss()
[all …]
Davx512vlbw-builtins.c9 __mmask32 test_mm256_cmpeq_epi8_mask(__m256i __a, __m256i __b) { in test_mm256_cmpeq_epi8_mask() argument
12 return (__mmask32)_mm256_cmpeq_epi8_mask(__a, __b); in test_mm256_cmpeq_epi8_mask()
15 __mmask32 test_mm256_mask_cmpeq_epi8_mask(__mmask32 __u, __m256i __a, __m256i __b) { in test_mm256_mask_cmpeq_epi8_mask() argument
18 return (__mmask32)_mm256_mask_cmpeq_epi8_mask(__u, __a, __b); in test_mm256_mask_cmpeq_epi8_mask()
21 __mmask16 test_mm_cmpeq_epi8_mask(__m128i __a, __m128i __b) { in test_mm_cmpeq_epi8_mask() argument
24 return (__mmask16)_mm_cmpeq_epi8_mask(__a, __b); in test_mm_cmpeq_epi8_mask()
27 __mmask16 test_mm_mask_cmpeq_epi8_mask(__mmask16 __u, __m128i __a, __m128i __b) { in test_mm_mask_cmpeq_epi8_mask() argument
30 return (__mmask16)_mm_mask_cmpeq_epi8_mask(__u, __a, __b); in test_mm_mask_cmpeq_epi8_mask()
33 __mmask16 test_mm256_cmpeq_epi16_mask(__m256i __a, __m256i __b) { in test_mm256_cmpeq_epi16_mask() argument
36 return (__mmask16)_mm256_cmpeq_epi16_mask(__a, __b); in test_mm256_cmpeq_epi16_mask()
[all …]
Davx512bw-builtins.c9 __mmask64 test_mm512_cmpeq_epi8_mask(__m512i __a, __m512i __b) { in test_mm512_cmpeq_epi8_mask() argument
12 return (__mmask64)_mm512_cmpeq_epi8_mask(__a, __b); in test_mm512_cmpeq_epi8_mask()
15 __mmask64 test_mm512_mask_cmpeq_epi8_mask(__mmask64 __u, __m512i __a, __m512i __b) { in test_mm512_mask_cmpeq_epi8_mask() argument
18 return (__mmask64)_mm512_mask_cmpeq_epi8_mask(__u, __a, __b); in test_mm512_mask_cmpeq_epi8_mask()
21 __mmask32 test_mm512_cmpeq_epi16_mask(__m512i __a, __m512i __b) { in test_mm512_cmpeq_epi16_mask() argument
24 return (__mmask32)_mm512_cmpeq_epi16_mask(__a, __b); in test_mm512_cmpeq_epi16_mask()
27 __mmask32 test_mm512_mask_cmpeq_epi16_mask(__mmask32 __u, __m512i __a, __m512i __b) { in test_mm512_mask_cmpeq_epi16_mask() argument
30 return (__mmask32)_mm512_mask_cmpeq_epi16_mask(__u, __a, __b); in test_mm512_mask_cmpeq_epi16_mask()
33 __mmask64 test_mm512_cmpgt_epi8_mask(__m512i __a, __m512i __b) { in test_mm512_cmpgt_epi8_mask() argument
36 return (__mmask64)_mm512_cmpgt_epi8_mask(__a, __b); in test_mm512_cmpgt_epi8_mask()
[all …]
Davx-builtins.c97 int test_extract_epi32(__m256i __a) { in test_extract_epi32() argument
101 return _mm256_extract_epi32(__a, 8); in test_extract_epi32()
104 int test_extract_epi16(__m256i __a) { in test_extract_epi16() argument
108 return _mm256_extract_epi16(__a, 16); in test_extract_epi16()
111 int test_extract_epi8(__m256i __a) { in test_extract_epi8() argument
115 return _mm256_extract_epi8(__a, 32); in test_extract_epi8()
118 __m256d test_256_blend_pd(__m256d __a, __m256d __b) { in test_256_blend_pd() argument
121 return _mm256_blend_pd(__a, __b, 0x35); in test_256_blend_pd()
124 __m256 test_256_blend_ps(__m256 __a, __m256 __b) { in test_256_blend_ps() argument
127 return _mm256_blend_ps(__a, __b, 0x35); in test_256_blend_ps()
[all …]
Davx512f-builtins.c686 __mmask16 test_mm512_cmpeq_epi32_mask(__m512i __a, __m512i __b) { in test_mm512_cmpeq_epi32_mask() argument
689 return (__mmask16)_mm512_cmpeq_epi32_mask(__a, __b); in test_mm512_cmpeq_epi32_mask()
692 __mmask16 test_mm512_mask_cmpeq_epi32_mask(__mmask16 __u, __m512i __a, __m512i __b) { in test_mm512_mask_cmpeq_epi32_mask() argument
695 return (__mmask16)_mm512_mask_cmpeq_epi32_mask(__u, __a, __b); in test_mm512_mask_cmpeq_epi32_mask()
698 __mmask8 test_mm512_mask_cmpeq_epi64_mask(__mmask8 __u, __m512i __a, __m512i __b) { in test_mm512_mask_cmpeq_epi64_mask() argument
701 return (__mmask8)_mm512_mask_cmpeq_epi64_mask(__u, __a, __b); in test_mm512_mask_cmpeq_epi64_mask()
704 __mmask8 test_mm512_cmpeq_epi64_mask(__m512i __a, __m512i __b) { in test_mm512_cmpeq_epi64_mask() argument
707 return (__mmask8)_mm512_cmpeq_epi64_mask(__a, __b); in test_mm512_cmpeq_epi64_mask()
710 __mmask16 test_mm512_cmpgt_epi32_mask(__m512i __a, __m512i __b) { in test_mm512_cmpgt_epi32_mask() argument
713 return (__mmask16)_mm512_cmpgt_epi32_mask(__a, __b); in test_mm512_cmpgt_epi32_mask()
[all …]
Davx512vl-builtins.c8 __mmask8 test_mm_cmpeq_epu32_mask(__m128i __a, __m128i __b) { in test_mm_cmpeq_epu32_mask() argument
11 return (__mmask8)_mm_cmpeq_epu32_mask(__a, __b); in test_mm_cmpeq_epu32_mask()
14 __mmask8 test_mm_mask_cmpeq_epu32_mask(__mmask8 __u, __m128i __a, __m128i __b) { in test_mm_mask_cmpeq_epu32_mask() argument
17 return (__mmask8)_mm_mask_cmpeq_epu32_mask(__u, __a, __b); in test_mm_mask_cmpeq_epu32_mask()
20 __mmask8 test_mm_cmpeq_epu64_mask(__m128i __a, __m128i __b) { in test_mm_cmpeq_epu64_mask() argument
23 return (__mmask8)_mm_cmpeq_epu64_mask(__a, __b); in test_mm_cmpeq_epu64_mask()
26 __mmask8 test_mm_mask_cmpeq_epu64_mask(__mmask8 __u, __m128i __a, __m128i __b) { in test_mm_mask_cmpeq_epu64_mask() argument
29 return (__mmask8)_mm_mask_cmpeq_epu64_mask(__u, __a, __b); in test_mm_mask_cmpeq_epu64_mask()
32 __mmask8 test_mm_cmpge_epi32_mask(__m128i __a, __m128i __b) { in test_mm_cmpge_epi32_mask() argument
35 return (__mmask8)_mm_cmpge_epi32_mask(__a, __b); in test_mm_cmpge_epi32_mask()
[all …]
/external/clang/test/Sema/
Dbuiltins-x86.c13 __m128 test__builtin_ia32_cmpps(__m128 __a, __m128 __b) { in test__builtin_ia32_cmpps() argument
14 __builtin_ia32_cmpps(__a, __b, 32); // expected-error {{argument should be a value from 0 to 31}} in test__builtin_ia32_cmpps()
17 __m128d test__builtin_ia32_cmppd(__m128d __a, __m128d __b) { in test__builtin_ia32_cmppd() argument
18 __builtin_ia32_cmppd(__a, __b, 32); // expected-error {{argument should be a value from 0 to 31}} in test__builtin_ia32_cmppd()
21 __m128 test__builtin_ia32_cmpss(__m128 __a, __m128 __b) { in test__builtin_ia32_cmpss() argument
22 __builtin_ia32_cmpss(__a, __b, 32); // expected-error {{argument should be a value from 0 to 31}} in test__builtin_ia32_cmpss()
25 __m128d test__builtin_ia32_cmpsd(__m128d __a, __m128d __b) { in test__builtin_ia32_cmpsd() argument
26 __builtin_ia32_cmpsd(__a, __b, 32); // expected-error {{argument should be a value from 0 to 31}} in test__builtin_ia32_cmpsd()
29 __mmask16 test__builtin_ia32_cmpps512_mask(__m512d __a, __m512d __b) { in test__builtin_ia32_cmpps512_mask() argument
30 …__builtin_ia32_cmpps512_mask(__a, __b, 32, -1, 0); // expected-error {{argument should be a value … in test__builtin_ia32_cmpps512_mask()
[all …]
/external/compiler-rt/lib/builtins/
Dmulsc3.c21 __mulsc3(float __a, float __b, float __c, float __d) in __mulsc3() argument
23 float __ac = __a * __c; in __mulsc3()
25 float __ad = __a * __d; in __mulsc3()
33 if (crt_isinf(__a) || crt_isinf(__b)) in __mulsc3()
35 __a = crt_copysignf(crt_isinf(__a) ? 1 : 0, __a); in __mulsc3()
47 if (crt_isnan(__a)) in __mulsc3()
48 __a = crt_copysignf(0, __a); in __mulsc3()
56 if (crt_isnan(__a)) in __mulsc3()
57 __a = crt_copysignf(0, __a); in __mulsc3()
68 COMPLEX_REAL(z) = CRT_INFINITY * (__a * __c - __b * __d); in __mulsc3()
[all …]
Dmulxc3.c23 __mulxc3(long double __a, long double __b, long double __c, long double __d) in __mulxc3() argument
25 long double __ac = __a * __c; in __mulxc3()
27 long double __ad = __a * __d; in __mulxc3()
35 if (crt_isinf(__a) || crt_isinf(__b)) in __mulxc3()
37 __a = crt_copysignl(crt_isinf(__a) ? 1 : 0, __a); in __mulxc3()
49 if (crt_isnan(__a)) in __mulxc3()
50 __a = crt_copysignl(0, __a); in __mulxc3()
58 if (crt_isnan(__a)) in __mulxc3()
59 __a = crt_copysignl(0, __a); in __mulxc3()
70 COMPLEX_REAL(z) = CRT_INFINITY * (__a * __c - __b * __d); in __mulxc3()
[all …]
Dmuldc3.c21 __muldc3(double __a, double __b, double __c, double __d) in __muldc3() argument
23 double __ac = __a * __c; in __muldc3()
25 double __ad = __a * __d; in __muldc3()
33 if (crt_isinf(__a) || crt_isinf(__b)) in __muldc3()
35 __a = crt_copysign(crt_isinf(__a) ? 1 : 0, __a); in __muldc3()
47 if (crt_isnan(__a)) in __muldc3()
48 __a = crt_copysign(0, __a); in __muldc3()
56 if (crt_isnan(__a)) in __muldc3()
57 __a = crt_copysign(0, __a); in __muldc3()
68 COMPLEX_REAL(z) = CRT_INFINITY * (__a * __c - __b * __d); in __muldc3()
[all …]
Ddivtc3.c21 __divtc3(long double __a, long double __b, long double __c, long double __d) in __divtc3() argument
33 __real__ z = crt_scalbnl((__a * __c + __b * __d) / __denom, -__ilogbw); in __divtc3()
34 __imag__ z = crt_scalbnl((__b * __c - __a * __d) / __denom, -__ilogbw); in __divtc3()
37 if ((__denom == 0.0) && (!crt_isnan(__a) || !crt_isnan(__b))) in __divtc3()
39 __real__ z = crt_copysignl(CRT_INFINITY, __c) * __a; in __divtc3()
42 else if ((crt_isinf(__a) || crt_isinf(__b)) && in __divtc3()
45 __a = crt_copysignl(crt_isinf(__a) ? 1.0 : 0.0, __a); in __divtc3()
47 __real__ z = CRT_INFINITY * (__a * __c + __b * __d); in __divtc3()
48 __imag__ z = CRT_INFINITY * (__b * __c - __a * __d); in __divtc3()
51 crt_isfinite(__a) && crt_isfinite(__b)) in __divtc3()
[all …]
Ddivdc3.c21 __divdc3(double __a, double __b, double __c, double __d) in __divdc3() argument
33 COMPLEX_REAL(z) = crt_scalbn((__a * __c + __b * __d) / __denom, -__ilogbw); in __divdc3()
34 COMPLEX_IMAGINARY(z) = crt_scalbn((__b * __c - __a * __d) / __denom, -__ilogbw); in __divdc3()
37 if ((__denom == 0.0) && (!crt_isnan(__a) || !crt_isnan(__b))) in __divdc3()
39 COMPLEX_REAL(z) = crt_copysign(CRT_INFINITY, __c) * __a; in __divdc3()
42 else if ((crt_isinf(__a) || crt_isinf(__b)) && in __divdc3()
45 __a = crt_copysign(crt_isinf(__a) ? 1.0 : 0.0, __a); in __divdc3()
47 COMPLEX_REAL(z) = CRT_INFINITY * (__a * __c + __b * __d); in __divdc3()
48 COMPLEX_IMAGINARY(z) = CRT_INFINITY * (__b * __c - __a * __d); in __divdc3()
51 crt_isfinite(__a) && crt_isfinite(__b)) in __divdc3()
[all …]
Ddivxc3.c22 __divxc3(long double __a, long double __b, long double __c, long double __d) in __divxc3() argument
34 COMPLEX_REAL(z) = crt_scalbnl((__a * __c + __b * __d) / __denom, -__ilogbw); in __divxc3()
35 COMPLEX_IMAGINARY(z) = crt_scalbnl((__b * __c - __a * __d) / __denom, -__ilogbw); in __divxc3()
38 if ((__denom == 0) && (!crt_isnan(__a) || !crt_isnan(__b))) in __divxc3()
40 COMPLEX_REAL(z) = crt_copysignl(CRT_INFINITY, __c) * __a; in __divxc3()
43 else if ((crt_isinf(__a) || crt_isinf(__b)) && in __divxc3()
46 __a = crt_copysignl(crt_isinf(__a) ? 1 : 0, __a); in __divxc3()
48 COMPLEX_REAL(z) = CRT_INFINITY * (__a * __c + __b * __d); in __divxc3()
49 COMPLEX_IMAGINARY(z) = CRT_INFINITY * (__b * __c - __a * __d); in __divxc3()
52 crt_isfinite(__a) && crt_isfinite(__b)) in __divxc3()
[all …]
Ddivsc3.c21 __divsc3(float __a, float __b, float __c, float __d) in __divsc3() argument
33 COMPLEX_REAL(z) = crt_scalbnf((__a * __c + __b * __d) / __denom, -__ilogbw); in __divsc3()
34 COMPLEX_IMAGINARY(z) = crt_scalbnf((__b * __c - __a * __d) / __denom, -__ilogbw); in __divsc3()
37 if ((__denom == 0) && (!crt_isnan(__a) || !crt_isnan(__b))) in __divsc3()
39 COMPLEX_REAL(z) = crt_copysignf(CRT_INFINITY, __c) * __a; in __divsc3()
42 else if ((crt_isinf(__a) || crt_isinf(__b)) && in __divsc3()
45 __a = crt_copysignf(crt_isinf(__a) ? 1 : 0, __a); in __divsc3()
47 COMPLEX_REAL(z) = CRT_INFINITY * (__a * __c + __b * __d); in __divsc3()
48 COMPLEX_IMAGINARY(z) = CRT_INFINITY * (__b * __c - __a * __d); in __divsc3()
51 crt_isfinite(__a) && crt_isfinite(__b)) in __divsc3()
[all …]
/external/llvm/test/CodeGen/PowerPC/
Dvsx-ldst-builtin-le.ll21 %__a.addr.i31 = alloca i32, align 4
23 %__a.addr.i29 = alloca i32, align 4
25 %__a.addr.i27 = alloca i32, align 4
27 %__a.addr.i25 = alloca i32, align 4
29 %__a.addr.i23 = alloca i32, align 4
31 %__a.addr.i20 = alloca <4 x i32>, align 16
34 %__a.addr.i17 = alloca <4 x i32>, align 16
37 %__a.addr.i14 = alloca <4 x float>, align 16
40 %__a.addr.i11 = alloca <2 x i64>, align 16
43 %__a.addr.i8 = alloca <2 x i64>, align 16
[all …]

12345