/external/clang/test/CodeGen/ |
D | aarch64-neon-3v.c | 52 int64x1_t test_vand_s64(int64x1_t a, int64x1_t b) { in test_vand_s64() 164 int64x1_t test_vorr_s64(int64x1_t a, int64x1_t b) { in test_vorr_s64() 276 int64x1_t test_veor_s64(int64x1_t a, int64x1_t b) { in test_veor_s64() 395 int64x1_t test_vbic_s64(int64x1_t a, int64x1_t b) { in test_vbic_s64() 523 int64x1_t test_vorn_s64(int64x1_t a, int64x1_t b) { in test_vorn_s64()
|
D | arm_neon_intrinsics.c | 574 int64x1_t test_vadd_s64(int64x1_t a, int64x1_t b) { in test_vadd_s64() 908 int64x1_t test_vand_s64(int64x1_t a, int64x1_t b) { in test_vand_s64() 1025 int64x1_t test_vbic_s64(int64x1_t a, int64x1_t b) { in test_vbic_s64() 1162 int64x1_t test_vbsl_s64(uint64x1_t a, int64x1_t b, int64x1_t c) { in test_vbsl_s64() 2261 int64x2_t test_vcombine_s64(int64x1_t a, int64x1_t b) { in test_vcombine_s64() 2439 int64x1_t test_vcreate_s64(uint64_t a) { in test_vcreate_s64() 2440 int64x1_t tmp = vcreate_s64(a); in test_vcreate_s64() 2748 int64x1_t test_vdup_lane_s64(int64x1_t a) { in test_vdup_lane_s64() 2762 int64x2_t test_vdupq_lane_s64(int64x1_t a) { in test_vdupq_lane_s64() 3038 int64x1_t test_vdup_n_s64(int64_t a) { in test_vdup_n_s64() [all …]
|
D | aarch64-neon-intrinsics.c | 34 int64x1_t test_vadd_s64(int64x1_t v1, int64x1_t v2) { in test_vadd_s64() 165 int64x1_t test_vsub_s64(int64x1_t v1, int64x1_t v2) { in test_vsub_s64() 1850 uint64x1_t test_vtst_s64(int64x1_t a, int64x1_t b) { in test_vtst_s64() 1895 uint64x1_t test_vceq_s64(int64x1_t a, int64x1_t b) { in test_vceq_s64() 2071 uint64x1_t test_vcge_s64(int64x1_t a, int64x1_t b) { in test_vcge_s64() 2234 uint64x1_t test_vcle_s64(int64x1_t a, int64x1_t b) { in test_vcle_s64() 2395 uint64x1_t test_vcgt_s64(int64x1_t a, int64x1_t b) { in test_vcgt_s64() 2560 uint64x1_t test_vclt_s64(int64x1_t a, int64x1_t b) { in test_vclt_s64() 3132 int64x1_t test_vqadd_s64(int64x1_t a, int64x1_t b) { in test_vqadd_s64() 3317 int64x1_t test_vqsub_s64(int64x1_t a, int64x1_t b) { in test_vqsub_s64() [all …]
|
D | aarch64-neon-extract.c | 45 int64x1_t test_vext_s64(int64x1_t a, int64x1_t b) { in test_vext_s64()
|
D | aarch64-neon-vcombine.c | 31 int64x2_t test_vcombine_s64(int64x1_t low, int64x1_t high) { in test_vcombine_s64()
|
D | arm-neon-vget.c | 43 int64x1_t low_s64( int64x2_t a) { in low_s64() 99 int64x1_t high_s64( int64x2_t a) { in high_s64()
|
D | aarch64-neon-vget.c | 198 int64_t test_vget_lane_s64(int64x1_t a) { in test_vget_lane_s64() 429 int64x1_t test_vset_lane_s64(int64_t a, int64x1_t b) { in test_vset_lane_s64()
|
D | aarch64-neon-vget-hilo.c | 32 int64x1_t test_vget_high_s64(int64x2_t a) { in test_vget_high_s64() 130 int64x1_t test_vget_low_s64(int64x2_t a) { in test_vget_low_s64()
|
D | arm64_vdup.c | 6 void test_vdup_lane_s64(int64x1_t a1) { in test_vdup_lane_s64()
|
/external/llvm-project/clang/test/CodeGen/ |
D | aarch64-neon-3v.c | 52 int64x1_t test_vand_s64(int64x1_t a, int64x1_t b) { in test_vand_s64() 164 int64x1_t test_vorr_s64(int64x1_t a, int64x1_t b) { in test_vorr_s64() 276 int64x1_t test_veor_s64(int64x1_t a, int64x1_t b) { in test_veor_s64() 395 int64x1_t test_vbic_s64(int64x1_t a, int64x1_t b) { in test_vbic_s64() 523 int64x1_t test_vorn_s64(int64x1_t a, int64x1_t b) { in test_vorn_s64()
|
D | arm_neon_intrinsics.c | 478 int64x1_t test_vadd_s64(int64x1_t a, int64x1_t b) { in test_vadd_s64() 784 int64x1_t test_vand_s64(int64x1_t a, int64x1_t b) { in test_vand_s64() 900 int64x1_t test_vbic_s64(int64x1_t a, int64x1_t b) { in test_vbic_s64() 1036 int64x1_t test_vbsl_s64(uint64x1_t a, int64x1_t b, int64x1_t c) { in test_vbsl_s64() 2082 int64x2_t test_vcombine_s64(int64x1_t a, int64x1_t b) { in test_vcombine_s64() 2253 int64x1_t test_vcreate_s64(uint64_t a) { in test_vcreate_s64() 2254 int64x1_t tmp = vcreate_s64(a); in test_vcreate_s64() 2569 int64x1_t test_vdup_lane_s64(int64x1_t a) { in test_vdup_lane_s64() 2587 int64x2_t test_vdupq_lane_s64(int64x1_t a) { in test_vdupq_lane_s64() 2864 int64x1_t test_vdup_n_s64(int64_t a) { in test_vdup_n_s64() [all …]
|
D | aarch64-neon-intrinsics.c | 35 int64x1_t test_vadd_s64(int64x1_t v1, int64x1_t v2) { in test_vadd_s64() 168 int64x1_t test_vsub_s64(int64x1_t v1, int64x1_t v2) { in test_vsub_s64() 1007 int64x1_t test_vbsl_s64(uint64x1_t v1, int64x1_t v2, int64x1_t v3) { in test_vbsl_s64() 1670 uint64x1_t test_vtst_s64(int64x1_t a, int64x1_t b) { in test_vtst_s64() 1713 uint64x1_t test_vceq_s64(int64x1_t a, int64x1_t b) { in test_vceq_s64() 1889 uint64x1_t test_vcge_s64(int64x1_t a, int64x1_t b) { in test_vcge_s64() 2052 uint64x1_t test_vcle_s64(int64x1_t a, int64x1_t b) { in test_vcle_s64() 2212 uint64x1_t test_vcgt_s64(int64x1_t a, int64x1_t b) { in test_vcgt_s64() 2375 uint64x1_t test_vclt_s64(int64x1_t a, int64x1_t b) { in test_vclt_s64() 2864 int64x1_t test_vqadd_s64(int64x1_t a, int64x1_t b) { in test_vqadd_s64() [all …]
|
D | aarch64-neon-extract.c | 45 int64x1_t test_vext_s64(int64x1_t a, int64x1_t b) { in test_vext_s64()
|
D | aarch64-neon-vcombine.c | 31 int64x2_t test_vcombine_s64(int64x1_t low, int64x1_t high) { in test_vcombine_s64()
|
D | arm-neon-vget.c | 43 int64x1_t low_s64( int64x2_t a) { in low_s64() 99 int64x1_t high_s64( int64x2_t a) { in high_s64()
|
D | aarch64-neon-vget.c | 168 int64_t test_vget_lane_s64(int64x1_t a) { in test_vget_lane_s64() 363 int64x1_t test_vset_lane_s64(int64_t a, int64x1_t b) { in test_vset_lane_s64()
|
D | aarch64-neon-vget-hilo.c | 32 int64x1_t test_vget_high_s64(int64x2_t a) { in test_vget_high_s64() 130 int64x1_t test_vget_low_s64(int64x2_t a) { in test_vget_low_s64()
|
D | arm-bf16-reinterpret-intrinsics.c | 75 bfloat16x4_t test_vreinterpret_bf16_s64(int64x1_t a) { return vreinterpret_bf16_s64(a); } in test_vreinterpret_bf16_s64() 228 int64x1_t test_vreinterpret_s64_bf16(bfloat16x4_t a) { return vreinterpret_s64_bf16(a); } in test_vreinterpret_s64_bf16()
|
D | arm64_vdup.c | 6 void test_vdup_lane_s64(int64x1_t a1) { in test_vdup_lane_s64()
|
/external/libopus/silk/arm/ |
D | NSQ_neon.c | 62 int64x1_t f = vadd_s64(vget_low_s64(e), vget_high_s64(e)); in silk_noise_shape_quantizer_short_prediction_neon() 102 int64x1_t c = vadd_s64(vget_low_s64(b3), vget_high_s64(b3)); in silk_NSQ_noise_shape_feedback_loop_neon() 103 int64x1_t cS = vrshr_n_s64(c, 15); in silk_NSQ_noise_shape_feedback_loop_neon()
|
/external/neon_2_sse/ |
D | NEON_2_SSE.h | 149 typedef __m64_128 int64x1_t; typedef 219 int64x1_t val[2]; 287 int64x1_t val[4]; 351 int64x1_t val[3]; 448 _NEON2SSESTORAGE int64x1_t vadd_s64(int64x1_t a, int64x1_t b); // VADD.I64 d0,d0,d0 507 _NEON2SSESTORAGE int64x1_t vqadd_s64(int64x1_t a, int64x1_t b); // VQADD.S64 d0,d0,d0 639 _NEON2SSESTORAGE int64x1_t vsub_s64(int64x1_t a, int64x1_t b); // VSUB.I64 d0,d0,d0 672 _NEON2SSESTORAGE int64x1_t vqsub_s64(int64x1_t a, int64x1_t b); // VQSUB.S64 d0,d0,d0 909 _NEON2SSESTORAGE int64x1_t vpaddl_s32(int32x2_t a); // VPADDL.S32 d0,d0 922 _NEON2SSESTORAGE int64x1_t vpadal_s32(int64x1_t a, int32x2_t b); // VPADAL.S32 d0,d0 [all …]
|
/external/libaom/libaom/av1/encoder/arm/neon/ |
D | quantize_neon.c | 107 const int64x1_t v_eobmax_xx32 = in av1_quantize_fp_neon() 111 const int64x1_t v_eobmax_xxx3 = in av1_quantize_fp_neon() 203 const int64x1_t v_eobmax_xx32 = in av1_quantize_lp_neon() 207 const int64x1_t v_eobmax_xxx3 = in av1_quantize_lp_neon()
|
/external/webrtc/common_audio/signal_processing/ |
D | cross_correlation_neon.c | 59 int64x1_t shift = vdup_n_s64(-scaling); in DotProductWithScaleNeon() 60 int64x1_t sum2 = vadd_s64(vget_low_s64(sum0), vget_high_s64(sum0)); in DotProductWithScaleNeon()
|
/external/llvm-project/libc/AOR_v20.02/networking/arm/ |
D | chksum_simd.c | 40 int64x1_t vshiftl = vdup_n_s64(CHAR_BIT * off); in __chksum_arm_simd() 120 int64x1_t vshiftr = vdup_n_s64(-CHAR_BIT * (8 - nbytes)); in __chksum_arm_simd()
|
/external/arm-optimized-routines/networking/arm/ |
D | chksum_simd.c | 39 int64x1_t vshiftl = vdup_n_s64(CHAR_BIT * off); in __chksum_arm_simd() 119 int64x1_t vshiftr = vdup_n_s64(-CHAR_BIT * (8 - nbytes)); in __chksum_arm_simd()
|