Home
last modified time | relevance | path

Searched refs:int64x2_t (Results 1 – 25 of 43) sorted by relevance

12

/external/clang/test/CodeGen/
Daarch64-neon-2velem.c518 int64x2_t test_vmlal_lane_s32(int64x2_t a, int32x2_t b, int32x2_t v) { in test_vmlal_lane_s32()
544 int64x2_t test_vmlal_laneq_s32(int64x2_t a, int32x2_t b, int32x4_t v) { in test_vmlal_laneq_s32()
572 int64x2_t test_vmlal_high_lane_s32(int64x2_t a, int32x4_t b, int32x2_t v) { in test_vmlal_high_lane_s32()
600 int64x2_t test_vmlal_high_laneq_s32(int64x2_t a, int32x4_t b, int32x4_t v) { in test_vmlal_high_laneq_s32()
626 int64x2_t test_vmlsl_lane_s32(int64x2_t a, int32x2_t b, int32x2_t v) { in test_vmlsl_lane_s32()
652 int64x2_t test_vmlsl_laneq_s32(int64x2_t a, int32x2_t b, int32x4_t v) { in test_vmlsl_laneq_s32()
680 int64x2_t test_vmlsl_high_lane_s32(int64x2_t a, int32x4_t b, int32x2_t v) { in test_vmlsl_high_lane_s32()
708 int64x2_t test_vmlsl_high_laneq_s32(int64x2_t a, int32x4_t b, int32x4_t v) { in test_vmlsl_high_laneq_s32()
734 int64x2_t test_vmlal_lane_u32(int64x2_t a, int32x2_t b, int32x2_t v) { in test_vmlal_lane_u32()
760 int64x2_t test_vmlal_laneq_u32(int64x2_t a, int32x2_t b, int32x4_t v) { in test_vmlal_laneq_u32()
[all …]
Darm64-vrnd.c9 int64x2_t rnd5(float64x2_t a) { return vrndq_f64(a); } in rnd5()
17 int64x2_t rnd9(float64x2_t a) { return vrndnq_f64(a); } in rnd9()
19 int64x2_t rnd10(float64x2_t a) { return vrndnq_f64(a); } in rnd10()
26 int64x2_t rnd13(float64x2_t a) { return vrndmq_f64(a); } in rnd13()
28 int64x2_t rnd14(float64x2_t a) { return vrndmq_f64(a); } in rnd14()
35 int64x2_t rnd18(float64x2_t a) { return vrndpq_f64(a); } in rnd18()
42 int64x2_t rnd22(float64x2_t a) { return vrndaq_f64(a); } in rnd22()
49 int64x2_t rnd25(float64x2_t a) { return vrndxq_f64(a); } in rnd25()
Daarch64-neon-3v.c59 int64x2_t test_vandq_s64(int64x2_t a, int64x2_t b) { in test_vandq_s64()
171 int64x2_t test_vorrq_s64(int64x2_t a, int64x2_t b) { in test_vorrq_s64()
283 int64x2_t test_veorq_s64(int64x2_t a, int64x2_t b) { in test_veorq_s64()
403 int64x2_t test_vbicq_s64(int64x2_t a, int64x2_t b) { in test_vbicq_s64()
531 int64x2_t test_vornq_s64(int64x2_t a, int64x2_t b) { in test_vornq_s64()
Daarch64-neon-misc.c101 uint64x2_t test_vceqzq_s64(int64x2_t a) { in test_vceqzq_s64()
329 uint64x2_t test_vcgezq_s64(int64x2_t a) { in test_vcgezq_s64()
445 uint64x2_t test_vclezq_s64(int64x2_t a) { in test_vclezq_s64()
561 uint64x2_t test_vcgtzq_s64(int64x2_t a) { in test_vcgtzq_s64()
677 uint64x2_t test_vcltzq_s64(int64x2_t a) { in test_vcltzq_s64()
1044 int64x2_t test_vpaddlq_s32(int32x4_t a) { in test_vpaddlq_s32()
1171 int64x2_t test_vpadalq_s32(int64x2_t a, int32x4_t b) { in test_vpadalq_s32()
1274 int64x2_t test_vqabsq_s64(int64x2_t a) { in test_vqabsq_s64()
1343 int64x2_t test_vqnegq_s64(int64x2_t a) { in test_vqnegq_s64()
1392 int64x2_t test_vnegq_s64(int64x2_t a) { in test_vnegq_s64()
[all …]
Daarch64-neon-intrinsics.c97 int64x2_t test_vaddq_s64(int64x2_t v1, int64x2_t v2) { in test_vaddq_s64()
228 int64x2_t test_vsubq_s64(int64x2_t v1, int64x2_t v2) { in test_vsubq_s64()
1249 int64x2_t test_vbslq_s64(uint64x2_t v1, int64x2_t v2, int64x2_t v3) { in test_vbslq_s64()
1780 uint64x2_t test_vtstq_s64(int64x2_t v1, int64x2_t v2) { in test_vtstq_s64()
2024 uint64x2_t test_vceqq_s64(int64x2_t v1, int64x2_t v2) { in test_vceqq_s64()
2183 uint64x2_t test_vcgeq_s64(int64x2_t v1, int64x2_t v2) { in test_vcgeq_s64()
2346 uint64x2_t test_vcleq_s64(int64x2_t v1, int64x2_t v2) { in test_vcleq_s64()
2507 uint64x2_t test_vcgtq_s64(int64x2_t v1, int64x2_t v2) { in test_vcgtq_s64()
2672 uint64x2_t test_vcltq_s64(int64x2_t v1, int64x2_t v2) { in test_vcltq_s64()
3224 int64x2_t test_vqaddq_s64(int64x2_t a, int64x2_t b) { in test_vqaddq_s64()
[all …]
Darm_neon_intrinsics.c193 int64x2_t test_vabal_s32(int64x2_t a, int32x2_t b, int32x2_t c) { in test_vabal_s32()
436 int64x2_t test_vabdl_s32(int32x2_t a, int32x2_t b) { in test_vabdl_s32()
637 int64x2_t test_vaddq_s64(int64x2_t a, int64x2_t b) { in test_vaddq_s64()
712 int32x2_t test_vaddhn_s64(int64x2_t a, int64x2_t b) { in test_vaddhn_s64()
787 int64x2_t test_vaddl_s32(int32x2_t a, int32x2_t b) { in test_vaddl_s32()
851 int64x2_t test_vaddw_s32(int64x2_t a, int32x2_t b) { in test_vaddw_s32()
964 int64x2_t test_vandq_s64(int64x2_t a, int64x2_t b) { in test_vandq_s64()
1089 int64x2_t test_vbicq_s64(int64x2_t a, int64x2_t b) { in test_vbicq_s64()
1271 int64x2_t test_vbslq_s64(uint64x2_t a, int64x2_t b, int64x2_t c) { in test_vbslq_s64()
2261 int64x2_t test_vcombine_s64(int64x1_t a, int64x1_t b) { in test_vcombine_s64()
[all …]
Daarch64-neon-perm.c52 int64x2_t test_vuzp1q_s64(int64x2_t a, int64x2_t b) { in test_vuzp1q_s64()
199 int64x2_t test_vuzp2q_s64(int64x2_t a, int64x2_t b) { in test_vuzp2q_s64()
346 int64x2_t test_vzip1q_s64(int64x2_t a, int64x2_t b) { in test_vzip1q_s64()
493 int64x2_t test_vzip2q_s64(int64x2_t a, int64x2_t b) { in test_vzip2q_s64()
640 int64x2_t test_vtrn1q_s64(int64x2_t a, int64x2_t b) { in test_vtrn1q_s64()
787 int64x2_t test_vtrn2q_s64(int64x2_t a, int64x2_t b) { in test_vtrn2q_s64()
Darm64_vcopy.c82 int64x2_t test_vcopyq_laneq_s64(int64x2_t a1, int64x2_t a2) { in test_vcopyq_laneq_s64()
Daarch64-neon-extract.c85 int64x2_t test_vextq_s64(int64x2_t a, int64x2_t b) { in test_vextq_s64()
Darm-asm-diag.c5 typedef __attribute__((neon_vector_type(2))) long long int64x2_t; typedef
7 int64x2_t val[4];
Darm-asm-warn.c19 typedef __attribute__((neon_vector_type(2))) long long int64x2_t; typedef
21 int64x2_t val[4];
Daarch64-neon-vget.c216 int64_t test_vgetq_lane_s64(int64x2_t a) { in test_vgetq_lane_s64()
447 int64x2_t test_vsetq_lane_s64(int64_t a, int64x2_t b) { in test_vsetq_lane_s64()
Darm-neon-vget.c43 int64x1_t low_s64( int64x2_t a) { in low_s64()
99 int64x1_t high_s64( int64x2_t a) { in high_s64()
Daarch64-neon-vget-hilo.c32 int64x1_t test_vget_high_s64(int64x2_t a) { in test_vget_high_s64()
130 int64x1_t test_vget_low_s64(int64x2_t a) { in test_vget_low_s64()
Daarch64-poly128.c93 poly128_t test_vreinterpretq_p128_s64(int64x2_t a) { in test_vreinterpretq_p128_s64()
184 int64x2_t test_vreinterpretq_s64_p128(poly128_t a) { in test_vreinterpretq_s64_p128()
/external/tensorflow/tensorflow/core/kernels/
Dquantization_utils.h430 inline int64x2_t Divide64x2PowRound(const int64x2_t val) { in Divide64x2PowRound()
431 const int64x2_t val_sign = vshrq_n_s64(val, 63); in Divide64x2PowRound()
432 const int64x2_t val_xor = veorq_s64(val, val_sign); in Divide64x2PowRound()
433 const int64x2_t val_pos = vsubq_s64(val_xor, val_sign); in Divide64x2PowRound()
434 const int64x2_t shifted_val_pos = vrshrq_n_s64(val_pos, POW); in Divide64x2PowRound()
435 const int64x2_t shifted_val_pos_xor = veorq_s64(shifted_val_pos, val_sign); in Divide64x2PowRound()
436 const int64x2_t shifted_val = vsubq_s64(shifted_val_pos_xor, val_sign); in Divide64x2PowRound()
457 inline int64x2_t Divide64x2Pow(const int64x2_t val) { in Divide64x2Pow()
459 static const int64x2_t FIRST_BIT = vmovq_n_s64(FIRST_BIT_VAL); in Divide64x2Pow()
460 const int64x2_t val_sign = vshrq_n_s64(val, 63); in Divide64x2Pow()
[all …]
/external/libopus/silk/arm/
DNSQ_neon.c60 int64x2_t e = vpaddlq_s32(d); in silk_noise_shape_quantizer_short_prediction_neon()
97 int64x2_t b0 = vmull_s32(vget_low_s32(a0), vget_low_s32(coef0)); in silk_NSQ_noise_shape_feedback_loop_neon()
98 int64x2_t b1 = vmlal_s32(b0, vget_high_s32(a0), vget_high_s32(coef0)); in silk_NSQ_noise_shape_feedback_loop_neon()
99 int64x2_t b2 = vmlal_s32(b1, vget_low_s32(a1), vget_low_s32(coef1)); in silk_NSQ_noise_shape_feedback_loop_neon()
100 int64x2_t b3 = vmlal_s32(b2, vget_high_s32(a1), vget_high_s32(coef1)); in silk_NSQ_noise_shape_feedback_loop_neon()
/external/webrtc/webrtc/modules/audio_coding/codecs/isac/fix/source/
Dtransform_neon.c406 int64x2_t xr0 = vmull_s32(vget_low_s32(tmpr_0), vget_low_s32(outre1_0)); in PostShiftAndDivideAndDemodulateNeon()
407 int64x2_t xi0 = vmull_s32(vget_low_s32(tmpr_0), vget_low_s32(outre2_0)); in PostShiftAndDivideAndDemodulateNeon()
408 int64x2_t xr2 = vmull_s32(vget_low_s32(tmpr_1), vget_low_s32(outre1_1)); in PostShiftAndDivideAndDemodulateNeon()
409 int64x2_t xi2 = vmull_s32(vget_low_s32(tmpr_1), vget_low_s32(outre2_1)); in PostShiftAndDivideAndDemodulateNeon()
416 int64x2_t xr1 = vmull_high_s32(tmpr_0, outre1_0); in PostShiftAndDivideAndDemodulateNeon()
417 int64x2_t xi1 = vmull_high_s32(tmpr_0, outre2_0); in PostShiftAndDivideAndDemodulateNeon()
418 int64x2_t xr3 = vmull_high_s32(tmpr_1, outre1_1); in PostShiftAndDivideAndDemodulateNeon()
419 int64x2_t xi3 = vmull_high_s32(tmpr_1, outre2_1); in PostShiftAndDivideAndDemodulateNeon()
425 int64x2_t xr1 = vmull_s32(vget_high_s32(tmpr_0), vget_high_s32(outre1_0)); in PostShiftAndDivideAndDemodulateNeon()
426 int64x2_t xi1 = vmull_s32(vget_high_s32(tmpr_0), vget_high_s32(outre2_0)); in PostShiftAndDivideAndDemodulateNeon()
[all …]
Dlattice_neon.c44 int64x2_t tmp2al_low, tmp2al_high, tmp2bl_low, tmp2bl_high; in WebRtcIsacfix_FilterMaLoopNeon()
145 int64x2_t tmp2l_tail; in WebRtcIsacfix_FilterMaLoopNeon()
/external/webrtc/webrtc/modules/video_processing/util/
Ddenoiser_filter_neon.cc19 const int64x2_t b = vpaddlq_s32(a); in HorizontalAddS16x8()
26 const int64x2_t b = vpaddlq_s32(v_32x4); in HorizontalAddS32x4()
123 int64x2_t v_sum_diff_total = vdupq_n_s64(0); in MbDenoise()
179 const int64x2_t fedcba98_76543210 = vpaddlq_s32(fedc_ba98_7654_3210); in MbDenoise()
250 const int64x2_t fedcba98_76543210 = in MbDenoise()
/external/libvpx/libvpx/vp8/encoder/arm/neon/
Ddenoising_neon.c69 int64x2_t v_sum_diff_total = vdupq_n_s64(0); in vp8_denoiser_filter_neon()
132 const int64x2_t fedcba98_76543210 = vpaddlq_s32(fedc_ba98_7654_3210); in vp8_denoiser_filter_neon()
202 const int64x2_t fedcba98_76543210 = in vp8_denoiser_filter_neon()
259 int64x2_t v_sum_diff_total = vdupq_n_s64(0); in vp8_denoiser_filter_uv_neon()
349 const int64x2_t fedcba98_76543210 = vpaddlq_s32(fedc_ba98_7654_3210); in vp8_denoiser_filter_uv_neon()
427 const int64x2_t fedcba98_76543210 = in vp8_denoiser_filter_uv_neon()
/external/libvpx/libvpx/vp9/encoder/arm/neon/
Dvp9_error_neon.c18 int64x2_t error = vdupq_n_s64(0); in vp9_block_error_fp_neon()
33 const int64x2_t err2 = vaddl_s32(vget_low_s32(err1), vget_high_s32(err1)); in vp9_block_error_fp_neon()
/external/libopus/silk/fixed/arm/
Dwarped_autocorrelation_FIX_neon_intr.c41 int64x2_t corr_QC_s64x2[ 2 ], t_s64x2[ 2 ]; in calc_corr()
75 int64x2_t lsh_s64x2; in silk_warped_autocorrelation_FIX_neon()
203 int64x2_t t_s64x2 = vdupq_n_s64( 0 ); in silk_warped_autocorrelation_FIX_neon()
229 int64x2_t corr_QC0_s64x2, corr_QC1_s64x2; in silk_warped_autocorrelation_FIX_neon()
/external/clang/test/CodeGenCXX/
Daarch64-mangle-neon-vectors.cpp21 typedef __attribute__((neon_vector_type(2))) int64_t int64x2_t; typedef
80 void f21(int64x2_t) {} in f21() argument
/external/webrtc/webrtc/common_audio/signal_processing/
Dcross_correlation_neon.c23 int64x2_t sum0 = vdupq_n_s64(0); in DotProductWithScaleNeon()
24 int64x2_t sum1 = vdupq_n_s64(0); in DotProductWithScaleNeon()

12