/external/clang/test/CodeGen/ |
D | neon-crypto.c | 44 uint32x4_t test_vsha1su1q_u32(uint32x4_t w0_3, uint32x4_t w12_15) { in test_vsha1su1q_u32() 50 uint32x4_t test_vsha256su0q_u32(uint32x4_t w0_3, uint32x4_t w4_7) { in test_vsha256su0q_u32() 56 uint32x4_t test_vsha1cq_u32(uint32x4_t hash_abcd, uint32_t hash_e, uint32x4_t wk) { in test_vsha1cq_u32() 62 uint32x4_t test_vsha1pq_u32(uint32x4_t hash_abcd, uint32_t hash_e, uint32x4_t wk) { in test_vsha1pq_u32() 68 uint32x4_t test_vsha1mq_u32(uint32x4_t hash_abcd, uint32_t hash_e, uint32x4_t wk) { in test_vsha1mq_u32() 74 uint32x4_t test_vsha1su0q_u32(uint32x4_t w0_3, uint32x4_t w4_7, uint32x4_t w8_11) { in test_vsha1su0q_u32() 80 uint32x4_t test_vsha256hq_u32(uint32x4_t hash_abcd, uint32x4_t hash_efgh, uint32x4_t wk) { in test_vsha256hq_u32() 86 uint32x4_t test_vsha256h2q_u32(uint32x4_t hash_efgh, uint32x4_t hash_abcd, uint32x4_t wk) { in test_vsha256h2q_u32() 92 uint32x4_t test_vsha256su1q_u32(uint32x4_t w0_3, uint32x4_t w8_11, uint32x4_t w12_15) { in test_vsha256su1q_u32()
|
D | arm64_crypto.c | 30 uint32x4_t test_sha1c(uint32x4_t hash_abcd, uint32_t hash_e, uint32x4_t wk) { in test_sha1c() 37 uint32x4_t test_sha1p(uint32x4_t hash_abcd, uint32_t hash_e, uint32x4_t wk) { in test_sha1p() 44 uint32x4_t test_sha1m(uint32x4_t hash_abcd, uint32_t hash_e, uint32x4_t wk) { in test_sha1m() 59 uint32x4_t test_sha1su0(uint32x4_t wk0_3, uint32x4_t wk4_7, uint32x4_t wk8_11) { in test_sha1su0() 65 uint32x4_t test_sha1su1(uint32x4_t wk0_3, uint32x4_t wk12_15) { in test_sha1su1() 71 uint32x4_t test_sha256h(uint32x4_t hash_abcd, uint32x4_t hash_efgh, uint32x4_t wk) { in test_sha256h() 77 uint32x4_t test_sha256h2(uint32x4_t hash_efgh, uint32x4_t hash_abcd, uint32x4_t wk) { in test_sha256h2() 83 uint32x4_t test_sha256su0(uint32x4_t w0_3, uint32x4_t w4_7) { in test_sha256su0() 89 uint32x4_t test_sha256su1(uint32x4_t w0_3, uint32x4_t w8_11, uint32x4_t w12_15) { in test_sha256su1()
|
D | arm64_neon_high_half.c | 26 uint32x4_t test_vaddw_high_u16(uint32x4_t lhs, uint16x8_t rhs) { in test_vaddw_high_u16() 31 uint64x2_t test_vaddw_high_u32(uint64x2_t lhs, uint32x4_t rhs) { in test_vaddw_high_u32() 56 uint32x4_t test_vsubw_high_u16(uint32x4_t lhs, uint16x8_t rhs) { in test_vsubw_high_u16() 61 uint64x2_t test_vsubw_high_u32(uint64x2_t lhs, uint32x4_t rhs) { in test_vsubw_high_u32() 86 uint32x4_t test_vabdl_high_u16(uint16x8_t lhs, uint16x8_t rhs) { in test_vabdl_high_u16() 91 uint64x2_t test_vabdl_high_u32(uint32x4_t lhs, uint32x4_t rhs) { in test_vabdl_high_u32() 116 uint32x4_t test_vabal_high_u16(uint32x4_t accum, uint16x8_t lhs, uint16x8_t rhs) { in test_vabal_high_u16() 121 uint64x2_t test_vabal_high_u32(uint64x2_t accum, uint32x4_t lhs, uint32x4_t rhs) { in test_vabal_high_u32() 236 uint32x4_t test_vsubl_high_u16(uint16x8_t lhs, uint16x8_t rhs) { in test_vsubl_high_u16() 241 uint64x2_t test_vsubl_high_u32(uint32x4_t lhs, uint32x4_t rhs) { in test_vsubl_high_u32() [all …]
|
D | aarch64-neon-2velem.c | 146 uint32x4_t test_vmulq_lane_u32(uint32x4_t a, uint32x2_t v) { in test_vmulq_lane_u32() 188 uint32x2_t test_vmul_laneq_u32(uint32x2_t a, uint32x4_t v) { in test_vmul_laneq_u32() 194 uint32x4_t test_vmulq_laneq_u32(uint32x4_t a, uint32x4_t v) { in test_vmulq_laneq_u32() 500 uint32x4_t test_vmull_lane_u16(uint16x4_t a, uint16x4_t v) { in test_vmull_lane_u16() 524 uint32x4_t test_vmull_high_lane_u16(uint16x8_t a, uint16x4_t v) { in test_vmull_high_lane_u16() 530 uint64x2_t test_vmull_high_lane_u32(uint32x4_t a, uint32x2_t v) { in test_vmull_high_lane_u32() 548 uint32x4_t test_vmull_laneq_u16(uint16x4_t a, uint16x8_t v) { in test_vmull_laneq_u16() 554 uint64x2_t test_vmull_laneq_u32(uint32x2_t a, uint32x4_t v) { in test_vmull_laneq_u32() 572 uint32x4_t test_vmull_high_laneq_u16(uint16x8_t a, uint16x8_t v) { in test_vmull_high_laneq_u16() 578 uint64x2_t test_vmull_high_laneq_u32(uint32x4_t a, uint32x4_t v) { in test_vmull_high_laneq_u32() [all …]
|
D | aarch64-neon-3v.c | 86 uint32x4_t test_vandq_u32(uint32x4_t a, uint32x4_t b) { in test_vandq_u32() 182 uint32x4_t test_vorrq_u32(uint32x4_t a, uint32x4_t b) { in test_vorrq_u32() 278 uint32x4_t test_veorq_u32(uint32x4_t a, uint32x4_t b) { in test_veorq_u32() 374 uint32x4_t test_vbicq_u32(uint32x4_t a, uint32x4_t b) { in test_vbicq_u32() 470 uint32x4_t test_vornq_u32(uint32x4_t a, uint32x4_t b) { in test_vornq_u32()
|
D | aarch64-neon-misc.c | 59 uint32x4_t test_vceqzq_s32(int32x4_t a) { in test_vceqzq_s32() 101 uint32x4_t test_vceqzq_u32(uint32x4_t a) { in test_vceqzq_u32() 125 uint32x4_t test_vceqzq_f32(float32x4_t a) { in test_vceqzq_f32() 203 uint32x4_t test_vcgezq_s32(int32x4_t a) { in test_vcgezq_s32() 227 uint32x4_t test_vcgezq_f32(float32x4_t a) { in test_vcgezq_f32() 275 uint32x4_t test_vclezq_s32(int32x4_t a) { in test_vclezq_s32() 299 uint32x4_t test_vclezq_f32(float32x4_t a) { in test_vclezq_f32() 347 uint32x4_t test_vcgtzq_s32(int32x4_t a) { in test_vcgtzq_s32() 371 uint32x4_t test_vcgtzq_f32(float32x4_t a) { in test_vcgtzq_f32() 419 uint32x4_t test_vcltzq_s32(int32x4_t a) { in test_vcltzq_s32() [all …]
|
D | arm_neon_intrinsics.c | 80 uint32x4_t test_vabaq_u32(uint32x4_t a, uint32x4_t b, uint32x4_t c) { in test_vabaq_u32() 111 uint32x4_t test_vabal_u16(uint32x4_t a, uint16x4_t b, uint16x4_t c) { in test_vabal_u16() 196 uint32x4_t test_vabdq_u32(uint32x4_t a, uint32x4_t b) { in test_vabdq_u32() 233 uint32x4_t test_vabdl_u16(uint16x4_t a, uint16x4_t b) { in test_vabdl_u16() 391 uint32x4_t test_vaddq_u32(uint32x4_t a, uint32x4_t b) { in test_vaddq_u32() 428 uint16x4_t test_vaddhn_u32(uint32x4_t a, uint32x4_t b) { in test_vaddhn_u32() 465 uint32x4_t test_vaddl_u16(uint16x4_t a, uint16x4_t b) { in test_vaddl_u16() 502 uint32x4_t test_vaddw_u16(uint32x4_t a, uint16x4_t b) { in test_vaddw_u16() 599 uint32x4_t test_vandq_u32(uint32x4_t a, uint32x4_t b) { in test_vandq_u32() 696 uint32x4_t test_vbicq_u32(uint32x4_t a, uint32x4_t b) { in test_vbicq_u32() [all …]
|
D | aarch64-neon-intrinsics.c | 111 uint32x4_t test_vaddq_u32(uint32x4_t v1, uint32x4_t v2) { in test_vaddq_u32() 223 uint32x4_t test_vsubq_u32(uint32x4_t v1, uint32x4_t v2) { in test_vsubq_u32() 308 uint32x4_t test_vmulq_u32(uint32x4_t v1, uint32x4_t v2) { in test_vmulq_u32() 417 uint32x4_t test_vmlaq_u32(uint32x4_t v1, uint32x4_t v2, uint32x4_t v3) { in test_vmlaq_u32() 505 uint32x4_t test_vmlsq_u32(uint32x4_t v1, uint32x4_t v2, uint32x4_t v3) { in test_vmlsq_u32() 635 uint32x4_t test_vabaq_u32(uint32x4_t v1, uint32x4_t v2, uint32x4_t v3) { in test_vabaq_u32() 713 uint32x4_t test_vabdq_u32(uint32x4_t v1, uint32x4_t v2) { in test_vabdq_u32() 816 int32x4_t test_vbslq_s32(uint32x4_t v1, int32x4_t v2, int32x4_t v3) { in test_vbslq_s32() 840 int32x4_t test_vbslq_u32(uint32x4_t v1, int32x4_t v2, int32x4_t v3) { in test_vbslq_u32() 852 float32x4_t test_vbslq_f32(uint32x4_t v1, float32x4_t v2, float32x4_t v3) { in test_vbslq_f32() [all …]
|
D | aarch64-neon-perm.c | 80 uint32x4_t test_vuzp1q_u32(uint32x4_t a, uint32x4_t b) { in test_vuzp1q_u32() 206 uint32x4_t test_vuzp2q_u32(uint32x4_t a, uint32x4_t b) { in test_vuzp2q_u32() 332 uint32x4_t test_vzip1q_u32(uint32x4_t a, uint32x4_t b) { in test_vzip1q_u32() 458 uint32x4_t test_vzip2q_u32(uint32x4_t a, uint32x4_t b) { in test_vzip2q_u32() 584 uint32x4_t test_vtrn1q_u32(uint32x4_t a, uint32x4_t b) { in test_vtrn1q_u32() 710 uint32x4_t test_vtrn2q_u32(uint32x4_t a, uint32x4_t b) { in test_vtrn2q_u32() 849 uint32x4x2_t test_vuzpq_u32(uint32x4_t a, uint32x4_t b) { in test_vuzpq_u32() 959 uint32x4x2_t test_vzipq_u32(uint32x4_t a, uint32x4_t b) { in test_vzipq_u32() 1069 uint32x4x2_t test_vtrnq_u32(uint32x4_t a, uint32x4_t b) { in test_vtrnq_u32()
|
D | arm64_vqmov.c | 32 uint16x8_t test_vqmovn_high_u32(uint16x4_t Vdlow, uint32x4_t Vn) in test_vqmovn_high_u32() 48 uint32x4_t test_vqmovn_high_u64(uint32x2_t Vdlow, uint64x2_t Vn) in test_vqmovn_high_u64() 64 uint16x8_t test_vqmovun_high_s32(uint16x4_t Vdlow, uint32x4_t Vn) in test_vqmovun_high_s32() 72 uint32x4_t test_vqmovun_high_s64(uint32x2_t Vdlow, uint64x2_t Vn) in test_vqmovun_high_s64()
|
D | arm-neon-vcvtX.c | 23 uint32x4_t test_vcvtaq_u32_f32(float32x4_t a) { in test_vcvtaq_u32_f32() 47 uint32x4_t test_vcvtnq_u32_f32(float32x4_t a) { in test_vcvtnq_u32_f32() 71 uint32x4_t test_vcvtpq_u32_f32(float32x4_t a) { in test_vcvtpq_u32_f32() 95 uint32x4_t test_vcvtmq_u32_f32(float32x4_t a) { in test_vcvtmq_u32_f32()
|
D | arm64_vshift.c | 89 uint32x4_t test_vqshlq_n_u32(uint32x4_t in) { in test_vqshlq_n_u32() 186 uint32x4_t test_vrshrq_n_u32(uint32x4_t in) { in test_vrshrq_n_u32() 345 uint32x4_t test_vrsraq_n_u32(uint32x4_t acc, uint32x4_t in) { in test_vrsraq_n_u32()
|
D | arm64_vcopy.c | 40 uint32x4_t test_vcopyq_laneq_u32(uint32x4_t a1, uint32x4_t a2) { in test_vcopyq_laneq_u32()
|
D | aarch64-neon-across.c | 63 uint64_t test_vaddlvq_u32(uint32x4_t a) { in test_vaddlvq_u32() 123 uint32_t test_vmaxvq_u32(uint32x4_t a) { in test_vmaxvq_u32() 183 uint32_t test_vminvq_u32(uint32x4_t a) { in test_vminvq_u32() 243 uint32_t test_vaddvq_u32(uint32x4_t a) { in test_vaddvq_u32()
|
D | arm64_vsli.c | 121 uint32x4_t test_vsliq_n_u32(uint32x4_t a1, uint32x4_t a2) { in test_vsliq_n_u32()
|
D | arm64_vsri.c | 122 uint32x4_t test_vsriq_n_u32(uint32x4_t a1, uint32x4_t a2) { in test_vsriq_n_u32()
|
D | arm64-vrsqrt.c | 18 uint32x4_t test_vrsqrteq_u32(uint32x4_t in) { in test_vrsqrteq_u32()
|
/external/scrypt/lib/crypto/ |
D | crypto_scrypt-neon-salsa208.h | 15 const uint32x4_t abab = {-1,0,-1,0}; in salsa20_8_intrinsic() 26 uint32x4_t x0x5x10x15; in salsa20_8_intrinsic() 27 uint32x4_t x12x1x6x11; in salsa20_8_intrinsic() 28 uint32x4_t x8x13x2x7; in salsa20_8_intrinsic() 29 uint32x4_t x4x9x14x3; in salsa20_8_intrinsic() 31 uint32x4_t x0x1x10x11; in salsa20_8_intrinsic() 32 uint32x4_t x12x13x6x7; in salsa20_8_intrinsic() 33 uint32x4_t x8x9x2x3; in salsa20_8_intrinsic() 34 uint32x4_t x4x5x14x15; in salsa20_8_intrinsic() 36 uint32x4_t x0x1x2x3; in salsa20_8_intrinsic() [all …]
|
/external/webp/src/dsp/ |
D | rescaler_neon.c | 26 #define LOAD_32x4(SRC, DST) const uint32x4_t DST = vld1q_u32((SRC)) 44 static uint32x4_t Interpolate(const rescaler_t* const frow, in Interpolate() 53 const uint32x4_t E = vcombine_u32( in Interpolate() 76 const uint32x4_t B0 = MULT_FIX(A0, fy_scale_half); in RescalerExportRowExpand() 77 const uint32x4_t B1 = MULT_FIX(A1, fy_scale_half); in RescalerExportRowExpand() 93 const uint32x4_t C0 = in RescalerExportRowExpand() 95 const uint32x4_t C1 = in RescalerExportRowExpand() 97 const uint32x4_t D0 = MULT_FIX(C0, fy_scale_half); in RescalerExportRowExpand() 98 const uint32x4_t D1 = MULT_FIX(C1, fy_scale_half); in RescalerExportRowExpand() 124 const uint32x4_t zero = vdupq_n_u32(0); in RescalerExportRowShrink() [all …]
|
/external/scrypt/patches/ |
D | arm-neon.patch | 21 + const uint32x4_t abab = {-1,0,-1,0}; 32 + uint32x4_t x0x5x10x15; 33 + uint32x4_t x12x1x6x11; 34 + uint32x4_t x8x13x2x7; 35 + uint32x4_t x4x9x14x3; 37 + uint32x4_t x0x1x10x11; 38 + uint32x4_t x12x13x6x7; 39 + uint32x4_t x8x9x2x3; 40 + uint32x4_t x4x5x14x15; 42 + uint32x4_t x0x1x2x3; [all …]
|
/external/boringssl/src/crypto/chacha/ |
D | chacha_vec.c | 57 #define ROTV1(x) (vec) vextq_u32((uint32x4_t)x, (uint32x4_t)x, 1) 58 #define ROTV2(x) (vec) vextq_u32((uint32x4_t)x, (uint32x4_t)x, 2) 59 #define ROTV3(x) (vec) vextq_u32((uint32x4_t)x, (uint32x4_t)x, 3) 68 (vec) vsriq_n_u32(vshlq_n_u32((uint32x4_t)x, 7), (uint32x4_t)x, 25) 70 (vec) vsriq_n_u32(vshlq_n_u32((uint32x4_t)x, 8), (uint32x4_t)x, 24) 72 (vec) vsriq_n_u32(vshlq_n_u32((uint32x4_t)x, 12), (uint32x4_t)x, 20)
|
/external/webrtc/webrtc/modules/audio_processing/aec/ |
D | aec_core_neon.c | 109 const uint32x4_t vec_p_inf = vdupq_n_u32(0x7F800000); in vsqrtq_f32() 111 const uint32x4_t div_by_zero = vceqq_u32(vec_p_inf, vreinterpretq_u32_f32(x)); in vsqrtq_f32() 152 const uint32x4_t bigger = vcgtq_f32(absEf, kThresh); in ScaleErrorSignalNEON() 155 uint32x4_t ef_re_if = vreinterpretq_u32_f32(vmulq_f32(ef_re, absEfInv)); in ScaleErrorSignalNEON() 156 uint32x4_t ef_im_if = vreinterpretq_u32_f32(vmulq_f32(ef_im, absEfInv)); in ScaleErrorSignalNEON() 157 uint32x4_t ef_re_u32 = vandq_u32(vmvnq_u32(bigger), in ScaleErrorSignalNEON() 159 uint32x4_t ef_im_u32 = vandq_u32(vmvnq_u32(bigger), in ScaleErrorSignalNEON() 292 const uint32x4_t vec_float_exponent_mask = vdupq_n_u32(0x7F800000); in vpowq_f32() 293 const uint32x4_t vec_eight_biased_exponent = vdupq_n_u32(0x43800000); in vpowq_f32() 294 const uint32x4_t vec_implicit_leading_one = vdupq_n_u32(0x43BF8000); in vpowq_f32() [all …]
|
/external/skia/src/opts/ |
D | Sk4px_NEON.h | 24 vst1q_u32(px, (uint32x4_t)this->fVec); in store4() 30 vst1q_lane_u32(px, (uint32x4_t)this->fVec, 0); in store1() 67 auto as = vshrq_n_u32((uint32x4_t)fVec, SK_A32_SHIFT); // ___3 ___2 ___1 ___0 in alphas() 77 auto a32 = (uint32x4_t)a8; // in Load4Alphas() 85 auto a32 = (uint32x4_t)a8; // in Load2Alphas()
|
/external/libvpx/libvpx/vpx_dsp/arm/ |
D | sad_neon.c | 24 uint32x4_t q1; in vpx_sad8x16_neon() 89 uint32x4_t q1; in vpx_sad16x8_neon() 121 const uint32x4_t vec_l_lo = vaddl_u16(vget_low_u16(vec_lo), in horizontal_long_add_16x8() 123 const uint32x4_t vec_l_hi = vaddl_u16(vget_low_u16(vec_hi), in horizontal_long_add_16x8() 125 const uint32x4_t a = vaddq_u32(vec_l_lo, vec_l_hi); in horizontal_long_add_16x8() 132 const uint32x4_t a = vpaddlq_u16(vec_16x8); in horizontal_add_16x8()
|
/external/webrtc/webrtc/modules/audio_processing/aecm/ |
D | aecm_core_neon.c | 34 static inline void AddLanes(uint32_t* ptr, uint32x4_t v) { in AddLanes() 58 uint32x4_t echo_est_v_low, echo_est_v_high; in WebRtcAecm_CalcLinearEnergiesNeon() 59 uint32x4_t far_energy_v, echo_stored_v, echo_adapt_v; in WebRtcAecm_CalcLinearEnergiesNeon() 150 uint32x4_t echo_est_v_low, echo_est_v_high; in WebRtcAecm_StoreAdaptiveChannelNeon()
|