/external/clang/test/CodeGen/ |
D | aarch64-neon-2velem.c | 26 int32x4_t test_vmlaq_lane_s32(int32x4_t a, int32x4_t b, int32x2_t v) { in test_vmlaq_lane_s32() 44 int32x2_t test_vmla_laneq_s32(int32x2_t a, int32x2_t b, int32x4_t v) { in test_vmla_laneq_s32() 50 int32x4_t test_vmlaq_laneq_s32(int32x4_t a, int32x4_t b, int32x4_t v) { in test_vmlaq_laneq_s32() 74 int32x4_t test_vmlsq_lane_s32(int32x4_t a, int32x4_t b, int32x2_t v) { in test_vmlsq_lane_s32() 92 int32x2_t test_vmls_laneq_s32(int32x2_t a, int32x2_t b, int32x4_t v) { in test_vmls_laneq_s32() 98 int32x4_t test_vmlsq_laneq_s32(int32x4_t a, int32x4_t b, int32x4_t v) { in test_vmlsq_laneq_s32() 122 int32x4_t test_vmulq_lane_s32(int32x4_t a, int32x2_t v) { in test_vmulq_lane_s32() 164 int32x2_t test_vmul_laneq_s32(int32x2_t a, int32x4_t v) { in test_vmul_laneq_s32() 170 int32x4_t test_vmulq_laneq_s32(int32x4_t a, int32x4_t v) { in test_vmulq_laneq_s32() 296 int32x4_t test_vmlal_lane_s16(int32x4_t a, int16x4_t b, int16x4_t v) { in test_vmlal_lane_s16() [all …]
|
D | arm64_neon_high_half.c | 11 int32x4_t test_vaddw_high_s16(int32x4_t lhs, int16x8_t rhs) { in test_vaddw_high_s16() 16 int64x2_t test_vaddw_high_s32(int64x2_t lhs, int32x4_t rhs) { in test_vaddw_high_s32() 41 int32x4_t test_vsubw_high_s16(int32x4_t lhs, int16x8_t rhs) { in test_vsubw_high_s16() 46 int64x2_t test_vsubw_high_s32(int64x2_t lhs, int32x4_t rhs) { in test_vsubw_high_s32() 71 int32x4_t test_vabdl_high_s16(int16x8_t lhs, int16x8_t rhs) { in test_vabdl_high_s16() 76 int64x2_t test_vabdl_high_s32(int32x4_t lhs, int32x4_t rhs) { in test_vabdl_high_s32() 101 int32x4_t test_vabal_high_s16(int32x4_t accum, int16x8_t lhs, int16x8_t rhs) { in test_vabal_high_s16() 106 int64x2_t test_vabal_high_s32(int64x2_t accum, int32x4_t lhs, int32x4_t rhs) { in test_vabal_high_s32() 126 int32x4_t test_vqdmlal_high_s16(int32x4_t accum, int16x8_t lhs, int16x8_t rhs) { in test_vqdmlal_high_s16() 131 int64x2_t test_vqdmlal_high_s32(int64x2_t accum, int32x4_t lhs, int32x4_t rhs) { in test_vqdmlal_high_s32() [all …]
|
D | aarch64-neon-misc.c | 59 uint32x4_t test_vceqzq_s32(int32x4_t a) { in test_vceqzq_s32() 203 uint32x4_t test_vcgezq_s32(int32x4_t a) { in test_vcgezq_s32() 275 uint32x4_t test_vclezq_s32(int32x4_t a) { in test_vclezq_s32() 347 uint32x4_t test_vcgtzq_s32(int32x4_t a) { in test_vcgtzq_s32() 419 uint32x4_t test_vcltzq_s32(int32x4_t a) { in test_vcltzq_s32() 629 int32x4_t test_vrev64q_s32(int32x4_t a) { in test_vrev64q_s32() 711 int32x4_t test_vpaddlq_s16(int16x8_t a) { in test_vpaddlq_s16() 717 int64x2_t test_vpaddlq_s32(int32x4_t a) { in test_vpaddlq_s32() 783 int32x4_t test_vpadalq_s16(int32x4_t a, int16x8_t b) { in test_vpadalq_s16() 789 int64x2_t test_vpadalq_s32(int64x2_t a, int32x4_t b) { in test_vpadalq_s32() [all …]
|
D | arm_neon_intrinsics.c | 62 int32x4_t test_vabaq_s32(int32x4_t a, int32x4_t b, int32x4_t c) { in test_vabaq_s32() 93 int32x4_t test_vabal_s16(int32x4_t a, int16x4_t b, int16x4_t c) { in test_vabal_s16() 178 int32x4_t test_vabdq_s32(int32x4_t a, int32x4_t b) { in test_vabdq_s32() 215 int32x4_t test_vabdl_s16(int16x4_t a, int16x4_t b) { in test_vabdl_s16() 282 int32x4_t test_vabsq_s32(int32x4_t a) { in test_vabsq_s32() 361 int32x4_t test_vaddq_s32(int32x4_t a, int32x4_t b) { in test_vaddq_s32() 410 int16x4_t test_vaddhn_s32(int32x4_t a, int32x4_t b) { in test_vaddhn_s32() 447 int32x4_t test_vaddl_s16(int16x4_t a, int16x4_t b) { in test_vaddl_s16() 484 int32x4_t test_vaddw_s16(int32x4_t a, int16x4_t b) { in test_vaddw_s16() 575 int32x4_t test_vandq_s32(int32x4_t a, int32x4_t b) { in test_vandq_s32() [all …]
|
D | aarch64-neon-intrinsics.c | 75 int32x4_t test_vaddq_s32(int32x4_t v1,int32x4_t v2) { in test_vaddq_s32() 187 int32x4_t test_vsubq_s32(int32x4_t v1,int32x4_t v2) { in test_vsubq_s32() 290 int32x4_t test_vmulq_s32(int32x4_t v1, int32x4_t v2) { in test_vmulq_s32() 393 int32x4_t test_vmlaq_s32(int32x4_t v1, int32x4_t v2, int32x4_t v3) { in test_vmlaq_s32() 482 int32x4_t test_vmlsq_s32(int32x4_t v1, int32x4_t v2, int32x4_t v3) { in test_vmlsq_s32() 617 int32x4_t test_vabaq_s32(int32x4_t v1, int32x4_t v2, int32x4_t v3) { in test_vabaq_s32() 695 int32x4_t test_vabdq_s32(int32x4_t v1, int32x4_t v2) { in test_vabdq_s32() 816 int32x4_t test_vbslq_s32(uint32x4_t v1, int32x4_t v2, int32x4_t v3) { in test_vbslq_s32() 840 int32x4_t test_vbslq_u32(uint32x4_t v1, int32x4_t v2, int32x4_t v3) { in test_vbslq_u32() 1062 uint32x4_t test_vtstq_s32(int32x4_t v1, int32x4_t v2) { in test_vtstq_s32() [all …]
|
D | arm64-vrnd.c | 7 int32x4_t rnd3(float32x4_t a) { return vrndq_f32(a); } in rnd3() 15 int32x4_t rnd8(float32x4_t a) { return vrndnq_f32(a); } in rnd8() 24 int32x4_t rnd12(float32x4_t a) { return vrndmq_f32(a); } in rnd12() 33 int32x4_t rnd16(float32x4_t a) { return vrndpq_f32(a); } in rnd16() 40 int32x4_t rnd20(float32x4_t a) { return vrndaq_f32(a); } in rnd20() 47 int32x4_t rnd24(float32x4_t a) { return vrndxq_f32(a); } in rnd24()
|
D | aarch64-neon-3v.c | 38 int32x4_t test_vandq_s32(int32x4_t a, int32x4_t b) { in test_vandq_s32() 134 int32x4_t test_vorrq_s32(int32x4_t a, int32x4_t b) { in test_vorrq_s32() 230 int32x4_t test_veorq_s32(int32x4_t a, int32x4_t b) { in test_veorq_s32() 326 int32x4_t test_vbicq_s32(int32x4_t a, int32x4_t b) { in test_vbicq_s32() 422 int32x4_t test_vornq_s32(int32x4_t a, int32x4_t b) { in test_vornq_s32()
|
D | arm64_vshift.c | 41 int32x4_t test_vqshlq_n_s32(int32x4_t in) { in test_vqshlq_n_s32() 138 int32x4_t test_vrshrq_n_s32(int32x4_t in) { in test_vrshrq_n_s32() 235 int32x4_t test_vqshluq_n_s32(int32x4_t in) { in test_vqshluq_n_s32() 289 int32x4_t test_vrsraq_n_s32(int32x4_t acc, int32x4_t in) { in test_vrsraq_n_s32()
|
D | arm64-arguments.c | 217 typedef __attribute__((neon_vector_type(4))) int int32x4_t; typedef 218 int32x4_t f36(int i, s36_with_align s1, s36_with_align s2) { in f36() 228 int32x4_t v = vaddq_s32(*(int32x4_t *)&s1, in f36() 229 *(int32x4_t *)&s2); in f36() 239 int32x4_t f37(int i, s37_with_align s1, s37_with_align s2) { in f37() 245 int32x4_t v = vaddq_s32(*(int32x4_t *)&s1, in f37() 246 *(int32x4_t *)&s2); in f37() 250 int32x4_t caller37() { in caller37() 672 int32x4_t a, b; 675 int32x4_t test_hva(int n, ...) { in test_hva() [all …]
|
D | aarch64-neon-perm.c | 38 int32x4_t test_vuzp1q_s32(int32x4_t a, int32x4_t b) { in test_vuzp1q_s32() 164 int32x4_t test_vuzp2q_s32(int32x4_t a, int32x4_t b) { in test_vuzp2q_s32() 290 int32x4_t test_vzip1q_s32(int32x4_t a, int32x4_t b) { in test_vzip1q_s32() 416 int32x4_t test_vzip2q_s32(int32x4_t a, int32x4_t b) { in test_vzip2q_s32() 542 int32x4_t test_vtrn1q_s32(int32x4_t a, int32x4_t b) { in test_vtrn1q_s32() 668 int32x4_t test_vtrn2q_s32(int32x4_t a, int32x4_t b) { in test_vtrn2q_s32() 831 int32x4x2_t test_vuzpq_s32(int32x4_t a, int32x4_t b) { in test_vuzpq_s32() 941 int32x4x2_t test_vzipq_s32(int32x4_t a, int32x4_t b) { in test_vzipq_s32() 1051 int32x4x2_t test_vtrnq_s32(int32x4_t a, int32x4_t b) { in test_vtrnq_s32()
|
D | arm-neon-vcvtX.c | 17 int32x4_t test_vcvtaq_s32_f32(float32x4_t a) { in test_vcvtaq_s32_f32() 41 int32x4_t test_vcvtnq_s32_f32(float32x4_t a) { in test_vcvtnq_s32_f32() 65 int32x4_t test_vcvtpq_s32_f32(float32x4_t a) { in test_vcvtpq_s32_f32() 89 int32x4_t test_vcvtmq_s32_f32(float32x4_t a) { in test_vcvtmq_s32_f32()
|
D | arm64_vcopy.c | 34 int32x4_t test_vcopyq_laneq_s32(int32x4_t a1, int32x4_t a2) { in test_vcopyq_laneq_s32()
|
D | aarch64-neon-across.c | 45 int64_t test_vaddlvq_s32(int32x4_t a) { in test_vaddlvq_s32() 105 int32_t test_vmaxvq_s32(int32x4_t a) { in test_vmaxvq_s32() 165 int32_t test_vminvq_s32(int32x4_t a) { in test_vminvq_s32() 225 int32_t test_vaddvq_s32(int32x4_t a) { in test_vaddvq_s32()
|
D | arm64_vsri.c | 94 int32x4_t test_vsriq_n_s32(int32x4_t a1, int32x4_t a2) { in test_vsriq_n_s32()
|
D | arm64_vsli.c | 93 int32x4_t test_vsliq_n_s32(int32x4_t a1, int32x4_t a2) { in test_vsliq_n_s32()
|
D | aarch64-neon-scalar-x-indexed-elem.c | 151 int64_t test_vqdmulls_laneq_s32(int32_t a, int32x4_t b) { in test_vqdmulls_laneq_s32() 177 int32_t test_vqdmulhs_laneq_s32(int32_t a, int32x4_t b) { in test_vqdmulhs_laneq_s32() 203 int32_t test_vqrdmulhs_laneq_s32(int32_t a, int32x4_t b) { in test_vqrdmulhs_laneq_s32() 227 int64_t test_vqdmlals_laneq_s32(int64_t a, int32_t b, int32x4_t c) { in test_vqdmlals_laneq_s32() 251 int64_t test_vqdmlsls_laneq_s32(int64_t a, int32_t b, int32x4_t c) { in test_vqdmlsls_laneq_s32()
|
D | aarch64-neon-extract.c | 44 int32x4_t test_vextq_s32(int32x4_t a, int32x4_t b) { in test_vextq_s32()
|
/external/skia/src/opts/ |
D | SkBitmapProcState_matrixProcs_neon.cpp | 22 static inline int16x8_t sbpsm_clamp_tile8(int32x4_t low, int32x4_t high, unsigned max) { in sbpsm_clamp_tile8() 36 static inline int32x4_t sbpsm_clamp_tile4(int32x4_t f, unsigned max) { in sbpsm_clamp_tile4() 37 int32x4_t res; in sbpsm_clamp_tile4() 50 static inline int32x4_t sbpsm_clamp_tile4_low_bits(int32x4_t fx) { in sbpsm_clamp_tile4_low_bits() 51 int32x4_t ret; in sbpsm_clamp_tile4_low_bits() 64 static inline int16x8_t sbpsm_repeat_tile8(int32x4_t low, int32x4_t high, unsigned max) { in sbpsm_repeat_tile8() 82 static inline int32x4_t sbpsm_repeat_tile4(int32x4_t f, unsigned max) { in sbpsm_repeat_tile4() 99 static inline int32x4_t sbpsm_repeat_tile4_low_bits(int32x4_t fx, unsigned max) { in sbpsm_repeat_tile4_low_bits() 102 int32x4_t ret; in sbpsm_repeat_tile4_low_bits() 154 int32x4_t vdx8 = vdupq_n_s32(dx8); in decal_nofilter_scale_neon() [all …]
|
D | SkBitmapProcState_arm_neon.cpp | 134 int32x4_t accum = vdupq_n_s32(0); in convolveHorizontally_neon() 154 int32x4_t p0 = vmull_s16(p0_src, coeff0); in convolveHorizontally_neon() 155 int32x4_t p1 = vmull_s16(p1_src, coeff1); in convolveHorizontally_neon() 156 int32x4_t p2 = vmull_s16(p2_src, coeff2); in convolveHorizontally_neon() 157 int32x4_t p3 = vmull_s16(p3_src, coeff3); in convolveHorizontally_neon() 188 int32x4_t p0 = vmull_s16(vget_low_s16(p01_16), coeff0); in convolveHorizontally_neon() 189 int32x4_t p1 = vmull_s16(vget_high_s16(p01_16), coeff1); in convolveHorizontally_neon() 190 int32x4_t p2 = vmull_s16(vget_low_s16(p23_16), coeff2); in convolveHorizontally_neon() 223 int32x4_t accum0, accum1, accum2, accum3; in convolveVertically_neon() 418 int32x4_t accum0 = vdupq_n_s32(0); in convolve4RowsHorizontally_neon() [all …]
|
D | SkBitmapProcState_matrix_neon.h | 69 int32x4_t lbase, hbase; in SCALE_NOFILTER_NAME() 134 int32x4_t xbase, ybase; in AFFINE_NOFILTER_NAME() 135 int32x4_t x2base, y2base; in AFFINE_NOFILTER_NAME() 246 static inline int32x4_t PACK_FILTER_X4_NAME(int32x4_t f, unsigned max, in PACK_FILTER_X4_NAME() 248 int32x4_t ret, res, wide_one; in PACK_FILTER_X4_NAME() 267 static inline int32x4_t PACK_FILTER_Y4_NAME(int32x4_t f, unsigned max, in PACK_FILTER_Y4_NAME() 269 int32x4_t ret, res, wide_one; in PACK_FILTER_Y4_NAME() 324 int32x4_t wide_fx; in SCALE_FILTER_NAME() 332 int32x4_t res; in SCALE_FILTER_NAME() 376 int32x4_t wide_fy, wide_fx; in AFFINE_FILTER_NAME() [all …]
|
/external/libhevc/common/arm/ |
D | ihevc_weighted_pred_neon_intr.c | 118 int32x4_t i4_tmp1_t; in ihevc_weighted_pred_uni_neonintr() 119 int32x4_t i4_tmp2_t; in ihevc_weighted_pred_uni_neonintr() 120 int32x4_t sto_res_tmp1; in ihevc_weighted_pred_uni_neonintr() 124 int32x4_t tmp_lvl_shift_t; in ihevc_weighted_pred_uni_neonintr() 126 int32x4_t tmp_shift_t; in ihevc_weighted_pred_uni_neonintr() 251 int32x4_t i4_tmp1_t; in ihevc_weighted_pred_chroma_uni_neonintr() 252 int32x4_t i4_tmp2_t; in ihevc_weighted_pred_chroma_uni_neonintr() 253 int32x4_t sto_res_tmp1; in ihevc_weighted_pred_chroma_uni_neonintr() 257 int32x4_t tmp_lvl_shift_t_u, tmp_lvl_shift_t_v; in ihevc_weighted_pred_chroma_uni_neonintr() 260 int32x4_t tmp_shift_t; in ihevc_weighted_pred_chroma_uni_neonintr() [all …]
|
/external/webp/src/dsp/ |
D | enc_neon.c | 306 const int32x4_t kCst937 = vdupq_n_s32(937); in FTransform() 307 const int32x4_t kCst1812 = vdupq_n_s32(1812); in FTransform() 315 const int32x4_t a3_2217 = vmull_n_s16(vget_low_s16(a3a2), 2217); in FTransform() 316 const int32x4_t a2_2217 = vmull_n_s16(vget_high_s16(a3a2), 2217); in FTransform() 317 const int32x4_t a2_p_a3 = vmlal_n_s16(a2_2217, vget_low_s16(a3a2), 5352); in FTransform() 318 const int32x4_t a3_m_a2 = vmlsl_n_s16(a3_2217, vget_high_s16(a3a2), 5352); in FTransform() 325 const int32x4_t kCst12000 = vdupq_n_s32(12000 + (1 << 16)); in FTransform() 326 const int32x4_t kCst51000 = vdupq_n_s32(51000); in FTransform() 332 const int32x4_t a3_2217 = vmull_n_s16(vget_low_s16(a3a2), 2217); in FTransform() 333 const int32x4_t a2_2217 = vmull_n_s16(vget_high_s16(a3a2), 2217); in FTransform() [all …]
|
D | upsampling_neon.c | 130 int32x4_t yl = vmull_lane_s16(vget_low_s16(yy), cf16, 0); \ 131 int32x4_t yh = vmull_lane_s16(vget_high_s16(yy), cf16, 0); \ 132 const int32x4_t rl = vmlal_lane_s16(yl, vget_low_s16(vv), cf16, 1);\ 133 const int32x4_t rh = vmlal_lane_s16(yh, vget_high_s16(vv), cf16, 1);\ 134 int32x4_t gl = vmlsl_lane_s16(yl, vget_low_s16(uu), cf16, 2); \ 135 int32x4_t gh = vmlsl_lane_s16(yh, vget_high_s16(uu), cf16, 2); \ 136 const int32x4_t bl = vmovl_s16(vget_low_s16(uu)); \ 137 const int32x4_t bh = vmovl_s16(vget_high_s16(uu)); \
|
/external/libvpx/libvpx/vp9/encoder/arm/neon/ |
D | vp9_dct_neon.c | 26 const int32x4_t a = vpaddlq_s16(sum); in vp9_fdct8x8_1_neon() 62 int32x4_t v_t0_lo = vaddl_s16(vget_low_s16(v_x0), vget_low_s16(v_x1)); in vp9_fdct8x8_neon() 63 int32x4_t v_t0_hi = vaddl_s16(vget_high_s16(v_x0), vget_high_s16(v_x1)); in vp9_fdct8x8_neon() 64 int32x4_t v_t1_lo = vsubl_s16(vget_low_s16(v_x0), vget_low_s16(v_x1)); in vp9_fdct8x8_neon() 65 int32x4_t v_t1_hi = vsubl_s16(vget_high_s16(v_x0), vget_high_s16(v_x1)); in vp9_fdct8x8_neon() 66 int32x4_t v_t2_lo = vmull_n_s16(vget_low_s16(v_x2), (int16_t)cospi_24_64); in vp9_fdct8x8_neon() 67 int32x4_t v_t2_hi = vmull_n_s16(vget_high_s16(v_x2), (int16_t)cospi_24_64); in vp9_fdct8x8_neon() 68 int32x4_t v_t3_lo = vmull_n_s16(vget_low_s16(v_x3), (int16_t)cospi_24_64); in vp9_fdct8x8_neon() 69 int32x4_t v_t3_hi = vmull_n_s16(vget_high_s16(v_x3), (int16_t)cospi_24_64); in vp9_fdct8x8_neon()
|
/external/clang/test/Sema/ |
D | arm-neon-types.c | 24 int32x4_t test4(int32x4_t a, vSInt32 b) { in test4()
|