/external/clang/test/CodeGen/ |
D | aarch64-neon-tbl.c | 9 int8x8_t test_vtbl1_s8(int8x8_t a, int8x8_t b) { in test_vtbl1_s8() 15 int8x8_t test_vqtbl1_s8(int8x16_t a, int8x8_t b) { in test_vqtbl1_s8() 21 int8x8_t test_vtbl2_s8(int8x8x2_t a, int8x8_t b) { in test_vtbl2_s8() 27 int8x8_t test_vqtbl2_s8(int8x16x2_t a, int8x8_t b) { in test_vqtbl2_s8() 33 int8x8_t test_vtbl3_s8(int8x8x3_t a, int8x8_t b) { in test_vtbl3_s8() 39 int8x8_t test_vqtbl3_s8(int8x16x3_t a, int8x8_t b) { in test_vqtbl3_s8() 45 int8x8_t test_vtbl4_s8(int8x8x4_t a, int8x8_t b) { in test_vtbl4_s8() 51 int8x8_t test_vqtbl4_s8(int8x16x4_t a, int8x8_t b) { in test_vqtbl4_s8() 81 int8x8_t test_vtbx1_s8(int8x8_t a, int8x8_t b, int8x8_t c) { in test_vtbx1_s8() 91 int8x8_t test_vtbx2_s8(int8x8_t a, int8x8x2_t b, int8x8_t c) { in test_vtbx2_s8() [all …]
|
D | aarch64-neon-misc.c | 11 uint8x8_t test_vceqz_s8(int8x8_t a) { in test_vceqz_s8() 167 uint8x8_t test_vcgez_s8(int8x8_t a) { in test_vcgez_s8() 239 uint8x8_t test_vclez_s8(int8x8_t a) { in test_vclez_s8() 311 uint8x8_t test_vcgtz_s8(int8x8_t a) { in test_vcgtz_s8() 383 uint8x8_t test_vcltz_s8(int8x8_t a) { in test_vcltz_s8() 455 int8x8_t test_vrev16_s8(int8x8_t a) { in test_vrev16_s8() 491 int8x8_t test_vrev32_s8(int8x8_t a) { in test_vrev32_s8() 563 int8x8_t test_vrev64_s8(int8x8_t a) { in test_vrev64_s8() 669 int16x4_t test_vpaddl_s8(int8x8_t a) { in test_vpaddl_s8() 741 int16x4_t test_vpadal_s8(int16x4_t a, int8x8_t b) { in test_vpadal_s8() [all …]
|
D | aarch64-neon-shifts.c | 13 int8x8_t test_shift_vshr_smax(int8x8_t a) { in test_shift_vshr_smax() 32 int8x8_t test_shift_vsra_smax(int8x8_t a, int8x8_t b) { in test_shift_vsra_smax()
|
D | arm-neon-shifts.c | 15 int8x8_t test_shift_vshr_smax(int8x8_t a) { in test_shift_vshr_smax() 34 int8x8_t test_shift_vsra_smax(int8x8_t a, int8x8_t b) { in test_shift_vsra_smax()
|
D | aarch64-neon-3v.c | 8 int8x8_t test_vand_s8(int8x8_t a, int8x8_t b) { in test_vand_s8() 104 int8x8_t test_vorr_s8(int8x8_t a, int8x8_t b) { in test_vorr_s8() 200 int8x8_t test_veor_s8(int8x8_t a, int8x8_t b) { in test_veor_s8() 296 int8x8_t test_vbic_s8(int8x8_t a, int8x8_t b) { in test_vbic_s8() 392 int8x8_t test_vorn_s8(int8x8_t a, int8x8_t b) { in test_vorn_s8()
|
D | arm_neon_intrinsics.c | 14 int8x8_t test_vaba_s8(int8x8_t a, int8x8_t b, int8x8_t c) { in test_vaba_s8() 87 int16x8_t test_vabal_s8(int16x8_t a, int8x8_t b, int8x8_t c) { in test_vabal_s8() 124 int8x8_t test_vabd_s8(int8x8_t a, int8x8_t b) { in test_vabd_s8() 209 int16x8_t test_vabdl_s8(int8x8_t a, int8x8_t b) { in test_vabdl_s8() 246 int8x8_t test_vabs_s8(int8x8_t a) { in test_vabs_s8() 295 int8x8_t test_vadd_s8(int8x8_t a, int8x8_t b) { in test_vadd_s8() 404 int8x8_t test_vaddhn_s16(int16x8_t a, int16x8_t b) { in test_vaddhn_s16() 441 int16x8_t test_vaddl_s8(int8x8_t a, int8x8_t b) { in test_vaddl_s8() 478 int16x8_t test_vaddw_s8(int16x8_t a, int8x8_t b) { in test_vaddw_s8() 515 int8x8_t test_vand_s8(int8x8_t a, int8x8_t b) { in test_vand_s8() [all …]
|
D | arm64_vshift.c | 4 int8x8_t test_vqshl_n_s8(int8x8_t in) { in test_vqshl_n_s8() 101 int8x8_t test_vrshr_n_s8(int8x8_t in) { in test_vrshr_n_s8() 198 int8x8_t test_vqshlu_n_s8(int8x8_t in) { in test_vqshlu_n_s8() 247 int8x8_t test_vrsra_n_s8(int8x8_t acc, int8x8_t in) { in test_vrsra_n_s8()
|
D | aarch64-neon-perm.c | 8 int8x8_t test_vuzp1_s8(int8x8_t a, int8x8_t b) { in test_vuzp1_s8() 134 int8x8_t test_vuzp2_s8(int8x8_t a, int8x8_t b) { in test_vuzp2_s8() 260 int8x8_t test_vzip1_s8(int8x8_t a, int8x8_t b) { in test_vzip1_s8() 386 int8x8_t test_vzip2_s8(int8x8_t a, int8x8_t b) { in test_vzip2_s8() 512 int8x8_t test_vtrn1_s8(int8x8_t a, int8x8_t b) { in test_vtrn1_s8() 638 int8x8_t test_vtrn2_s8(int8x8_t a, int8x8_t b) { in test_vtrn2_s8() 764 int8x8x2_t test_vuzp_s8(int8x8_t a, int8x8_t b) { in test_vuzp_s8() 874 int8x8x2_t test_vzip_s8(int8x8_t a, int8x8_t b) { in test_vzip_s8() 984 int8x8x2_t test_vtrn_s8(int8x8_t a, int8x8_t b) { in test_vtrn_s8()
|
D | aarch64-neon-intrinsics.c | 9 int8x8_t test_vadd_s8(int8x8_t v1, int8x8_t v2) { in test_vadd_s8() 123 int8x8_t test_vsub_s8(int8x8_t v1, int8x8_t v2) { in test_vsub_s8() 235 int8x8_t test_vmul_s8(int8x8_t v1, int8x8_t v2) { in test_vmul_s8() 339 int8x8_t test_vmla_s8(int8x8_t v1, int8x8_t v2, int8x8_t v3) { in test_vmla_s8() 345 int8x8_t test_vmla_s16(int16x4_t v1, int16x4_t v2, int16x4_t v3) { in test_vmla_s16() 429 int8x8_t test_vmls_s8(int8x8_t v1, int8x8_t v2, int8x8_t v3) { in test_vmls_s8() 435 int8x8_t test_vmls_s16(int16x4_t v1, int16x4_t v2, int16x4_t v3) { in test_vmls_s16() 569 int8x8_t test_vaba_s8(int8x8_t v1, int8x8_t v2, int8x8_t v3) { in test_vaba_s8() 641 int8x8_t test_vabd_s8(int8x8_t v1, int8x8_t v2) { in test_vabd_s8() 732 int8x8_t test_vbsl_s8(uint8x8_t v1, int8x8_t v2, int8x8_t v3) { in test_vbsl_s8() [all …]
|
D | aarch64-neon-across.c | 9 int16_t test_vaddlv_s8(int8x8_t a) { in test_vaddlv_s8() 69 int8_t test_vmaxv_s8(int8x8_t a) { in test_vmaxv_s8() 129 int8_t test_vminv_s8(int8x8_t a) { in test_vminv_s8() 189 int8_t test_vaddv_s8(int8x8_t a) { in test_vaddv_s8()
|
D | arm64_vsli.c | 9 int8x8_t test_vsli_n_s8(int8x8_t a1, int8x8_t a2) { in test_vsli_n_s8()
|
D | arm64_vsri.c | 10 int8x8_t test_vsri_n_s8(int8x8_t a1, int8x8_t a2) { in test_vsri_n_s8()
|
D | arm64_neon_high_half.c | 246 int8x16_t test_vrshrn_high_n_s16(int8x8_t lowpart, int16x8_t input) { in test_vrshrn_high_n_s16() 276 int8x16_t test_vshrn_high_n_s16(int8x8_t lowpart, int16x8_t input) { in test_vshrn_high_n_s16() 336 int8x16_t test_vqshrn_high_n_s16(int8x8_t lowpart, int16x8_t input) { in test_vqshrn_high_n_s16() 366 int8x16_t test_vqrshrn_high_n_s16(int8x8_t lowpart, int16x8_t input) { in test_vqrshrn_high_n_s16() 396 int8x16_t test_vaddhn_high_s16(int8x8_t lowpart, int16x8_t lhs, int16x8_t rhs) { in test_vaddhn_high_s16() 426 int8x16_t test_vraddhn_high_s16(int8x8_t lowpart, int16x8_t lhs, int16x8_t rhs) { in test_vraddhn_high_s16() 456 int8x16_t test_vmovn_high_s16(int8x8_t lowpart, int16x8_t wide) { in test_vmovn_high_s16() 486 int8x16_t test_vqmovn_high_s16(int8x8_t lowpart, int16x8_t wide) { in test_vqmovn_high_s16()
|
D | aarch64-neon-extract.c | 9 int8x8_t test_vext_s8(int8x8_t a, int8x8_t b) { in test_vext_s8()
|
D | arm64_vadd.c | 19 int8_t test_vaddv_s8(int8x8_t a1) { in test_vaddv_s8() 40 uint8_t test_vaddv_u8(int8x8_t a1) { in test_vaddv_u8()
|
D | aarch64-neon-vcombine.c | 8 int8x16_t test_vcombine_s8(int8x8_t low, int8x8_t high) { in test_vcombine_s8()
|
D | arm-neon-vget.c | 14 int8x8_t low_s8(int8x16_t a) { in low_s8() 70 int8x8_t high_s8(int8x16_t a) { in high_s8()
|
D | aarch64-neon-vget.c | 28 int8_t test_vget_lane_s8(int8x8_t a) { in test_vget_lane_s8() 201 int8x8_t test_vset_lane_s8(int8_t a, int8x8_t b) { in test_vset_lane_s8()
|
D | aarch64-neon-vget-hilo.c | 9 int8x8_t test_vget_high_s8(int8x16_t a) { in test_vget_high_s8() 93 int8x8_t test_vget_low_s8(int8x16_t a) { in test_vget_low_s8()
|
D | arm64-lanes.c | 7 int8_t test_vdupb_lane_s8(int8x8_t src) { in test_vdupb_lane_s8()
|
D | arm64_vqmov.c | 8 int8x16_t test_vqmovn_high_s16(int8x8_t Vdlow, int16x8_t Vn) in test_vqmovn_high_s16()
|
D | aarch64-neon-scalar-copy.c | 41 int8_t test_vdupb_lane_s8(int8x8_t a) { in test_vdupb_lane_s8()
|
/external/clang/test/CodeGenCXX/ |
D | aarch64-mangle-neon-vectors.cpp | 15 typedef __attribute__((neon_vector_type(8))) int8_t int8x8_t; typedef 40 void f1(int8x8_t) {} in f1() argument
|
/external/llvm/test/CodeGen/ARM/ |
D | vmov.ll | 172 %struct.int8x8_t = type { <8 x i8> } 173 define void @vdupn128(%struct.int8x8_t* noalias nocapture sret %agg.result) nounwind { 177 …%0 = getelementptr inbounds %struct.int8x8_t, %struct.int8x8_t* %agg.result, i32 0, i32 0 ; <<8 x … 182 define void @vdupnneg75(%struct.int8x8_t* noalias nocapture sret %agg.result) nounwind { 186 …%0 = getelementptr inbounds %struct.int8x8_t, %struct.int8x8_t* %agg.result, i32 0, i32 0 ; <<8 x …
|
/external/clang/test/Sema/ |
D | aarch64-neon-ranges.c | 6 void test_vext_8bit(int8x8_t small, int8x16_t big) { in test_vext_8bit() 39 void test_ld1st1(int8x8_t small, int8x16_t big, void *addr) { in test_ld1st1()
|