Home
last modified time | relevance | path

Searched refs:uint16x4_t (Results 1 – 25 of 74) sorted by relevance

123

/external/libvpx/libvpx/vpx_dsp/arm/
Dhighbd_intrapred_neon.c20 static INLINE uint16x4_t dc_sum_4(const uint16_t *ref) { in dc_sum_4()
21 const uint16x4_t ref_u16 = vld1_u16(ref); in dc_sum_4()
22 const uint16x4_t p0 = vpadd_u16(ref_u16, ref_u16); in dc_sum_4()
27 const uint16x4_t dc) { in dc_store_4x4()
28 const uint16x4_t dc_dup = vdup_lane_u16(dc, 0); in dc_store_4x4()
38 const uint16x4_t a = vld1_u16(above); in vpx_highbd_dc_predictor_4x4_neon()
39 const uint16x4_t l = vld1_u16(left); in vpx_highbd_dc_predictor_4x4_neon()
40 uint16x4_t sum; in vpx_highbd_dc_predictor_4x4_neon()
41 uint16x4_t dc; in vpx_highbd_dc_predictor_4x4_neon()
53 const uint16x4_t sum = dc_sum_4(left); in vpx_highbd_dc_left_predictor_4x4_neon()
[all …]
Dsad4d_neon.c37 uint16x4_t a[2]; in sad4x_4d()
77 const uint16x4_t a0 = vadd_u16(vget_low_u16(sum[0]), vget_high_u16(sum[0])); in sad_512_pel_final_neon()
78 const uint16x4_t a1 = vadd_u16(vget_low_u16(sum[1]), vget_high_u16(sum[1])); in sad_512_pel_final_neon()
79 const uint16x4_t a2 = vadd_u16(vget_low_u16(sum[2]), vget_high_u16(sum[2])); in sad_512_pel_final_neon()
80 const uint16x4_t a3 = vadd_u16(vget_low_u16(sum[3]), vget_high_u16(sum[3])); in sad_512_pel_final_neon()
81 const uint16x4_t b0 = vpadd_u16(a0, a1); in sad_512_pel_final_neon()
82 const uint16x4_t b1 = vpadd_u16(a2, a3); in sad_512_pel_final_neon()
90 const uint16x4_t a0 = vpadd_u16(vget_low_u16(sum[0]), vget_high_u16(sum[0])); in sad_1024_pel_final_neon()
91 const uint16x4_t a1 = vpadd_u16(vget_low_u16(sum[1]), vget_high_u16(sum[1])); in sad_1024_pel_final_neon()
92 const uint16x4_t a2 = vpadd_u16(vget_low_u16(sum[2]), vget_high_u16(sum[2])); in sad_1024_pel_final_neon()
[all …]
Dquantize_neon.c154 const uint16x4_t eob_max_0 = in vpx_quantize_b_neon()
156 const uint16x4_t eob_max_1 = vpmax_u16(eob_max_0, eob_max_0); in vpx_quantize_b_neon()
157 const uint16x4_t eob_max_2 = vpmax_u16(eob_max_1, eob_max_1); in vpx_quantize_b_neon()
310 const uint16x4_t eob_max_0 = in vpx_quantize_b_32x32_neon()
312 const uint16x4_t eob_max_1 = vpmax_u16(eob_max_0, eob_max_0); in vpx_quantize_b_32x32_neon()
313 const uint16x4_t eob_max_2 = vpmax_u16(eob_max_1, eob_max_1); in vpx_quantize_b_32x32_neon()
Dintrapred_neon.c20 static INLINE uint16x4_t dc_sum_4(const uint8_t *ref) { in dc_sum_4()
22 const uint16x4_t p0 = vpaddl_u8(ref_u8); in dc_sum_4()
40 uint16x4_t sum; in vpx_dc_predictor_4x4_neon()
50 const uint16x4_t sum = dc_sum_4(left); in vpx_dc_left_predictor_4x4_neon()
58 const uint16x4_t sum = dc_sum_4(above); in vpx_dc_top_predictor_4x4_neon()
75 static INLINE uint16x4_t dc_sum_8(const uint8_t *ref) { in dc_sum_8()
77 uint16x4_t sum = vpaddl_u8(ref_u8); in dc_sum_8()
97 uint16x4_t sum = vadd_u16(vget_low_u16(p0), vget_high_u16(p0)); in vpx_dc_predictor_8x8_neon()
107 const uint16x4_t sum = dc_sum_8(left); in vpx_dc_left_predictor_8x8_neon()
115 const uint16x4_t sum = dc_sum_8(above); in vpx_dc_top_predictor_8x8_neon()
[all …]
/external/clang/test/CodeGen/
Darm_neon_intrinsics.c63 uint16x4_t test_vaba_u16(uint16x4_t a, uint16x4_t b, uint16x4_t c) { in test_vaba_u16()
219 uint32x4_t test_vabal_u16(uint32x4_t a, uint16x4_t b, uint16x4_t c) { in test_vabal_u16()
290 uint16x4_t test_vabd_u16(uint16x4_t a, uint16x4_t b) { in test_vabd_u16()
460 uint32x4_t test_vabdl_u16(uint16x4_t a, uint16x4_t b) { in test_vabdl_u16()
595 uint16x4_t test_vadd_u16(uint16x4_t a, uint16x4_t b) { in test_vadd_u16()
738 uint16x4_t test_vaddhn_u32(uint32x4_t a, uint32x4_t b) { in test_vaddhn_u32()
809 uint32x4_t test_vaddl_u16(uint16x4_t a, uint16x4_t b) { in test_vaddl_u16()
869 uint32x4_t test_vaddw_u16(uint32x4_t a, uint16x4_t b) { in test_vaddw_u16()
922 uint16x4_t test_vand_u16(uint16x4_t a, uint16x4_t b) { in test_vand_u16()
1041 uint16x4_t test_vbic_u16(uint16x4_t a, uint16x4_t b) { in test_vbic_u16()
[all …]
Daarch64-neon-3v.c80 uint16x4_t test_vand_u16(uint16x4_t a, uint16x4_t b) { in test_vand_u16()
192 uint16x4_t test_vorr_u16(uint16x4_t a, uint16x4_t b) { in test_vorr_u16()
304 uint16x4_t test_veor_u16(uint16x4_t a, uint16x4_t b) { in test_veor_u16()
427 uint16x4_t test_vbic_u16(uint16x4_t a, uint16x4_t b) { in test_vbic_u16()
555 uint16x4_t test_vorn_u16(uint16x4_t a, uint16x4_t b) { in test_vorn_u16()
Daarch64-neon-2velem.c187 uint16x4_t test_vmul_lane_u16(uint16x4_t a, uint16x4_t v) { in test_vmul_lane_u16()
195 uint16x8_t test_vmulq_lane_u16(uint16x8_t a, uint16x4_t v) { in test_vmulq_lane_u16()
251 uint16x4_t test_vmul_laneq_u16(uint16x4_t a, uint16x8_t v) { in test_vmul_laneq_u16()
960 uint32x4_t test_vmull_lane_u16(uint16x4_t a, uint16x4_t v) { in test_vmull_lane_u16()
1011 uint32x4_t test_vmull_high_lane_u16(uint16x8_t a, uint16x4_t v) { in test_vmull_high_lane_u16()
1060 uint32x4_t test_vmull_laneq_u16(uint16x4_t a, uint16x8_t v) { in test_vmull_laneq_u16()
1809 uint16x4_t test_vmul_lane_u16_0(uint16x4_t a, uint16x4_t v) { in test_vmul_lane_u16_0()
1817 uint16x8_t test_vmulq_lane_u16_0(uint16x8_t a, uint16x4_t v) { in test_vmulq_lane_u16_0()
1873 uint16x4_t test_vmul_laneq_u16_0(uint16x4_t a, uint16x8_t v) { in test_vmul_laneq_u16_0()
2510 uint32x4_t test_vmull_lane_u16_0(uint16x4_t a, uint16x4_t v) { in test_vmull_lane_u16_0()
[all …]
Daarch64-neon-perm.c73 uint16x4_t test_vuzp1_u16(uint16x4_t a, uint16x4_t b) { in test_vuzp1_u16()
220 uint16x4_t test_vuzp2_u16(uint16x4_t a, uint16x4_t b) { in test_vuzp2_u16()
367 uint16x4_t test_vzip1_u16(uint16x4_t a, uint16x4_t b) { in test_vzip1_u16()
514 uint16x4_t test_vzip2_u16(uint16x4_t a, uint16x4_t b) { in test_vzip2_u16()
661 uint16x4_t test_vtrn1_u16(uint16x4_t a, uint16x4_t b) { in test_vtrn1_u16()
808 uint16x4_t test_vtrn2_u16(uint16x4_t a, uint16x4_t b) { in test_vtrn2_u16()
1014 uint16x4x2_t test_vuzp_u16(uint16x4_t a, uint16x4_t b) { in test_vuzp_u16()
1478 uint16x4x2_t test_vzip_u16(uint16x4_t a, uint16x4_t b) { in test_vzip_u16()
1942 uint16x4x2_t test_vtrn_u16(uint16x4_t a, uint16x4_t b) { in test_vtrn_u16()
Daarch64-neon-misc.c23 uint16x4_t test_vceqz_s16(int16x4_t a) { in test_vceqz_s16()
119 uint16x4_t test_vceqz_u16(uint16x4_t a) { in test_vceqz_u16()
223 uint16x4_t test_vceqz_p16(poly16x4_t a) { in test_vceqz_p16()
271 uint16x4_t test_vcgez_s16(int16x4_t a) { in test_vcgez_s16()
387 uint16x4_t test_vclez_s16(int16x4_t a) { in test_vclez_s16()
503 uint16x4_t test_vcgtz_s16(int16x4_t a) { in test_vcgtz_s16()
619 uint16x4_t test_vcltz_s16(int16x4_t a) { in test_vcltz_s16()
787 uint16x4_t test_vrev32_u16(uint16x4_t a) { in test_vrev32_u16()
878 uint16x4_t test_vrev64_u16(uint16x4_t a) { in test_vrev64_u16()
1001 uint16x4_t test_vpaddl_u8(uint8x8_t a) { in test_vpaddl_u8()
[all …]
Daarch64-neon-intrinsics.c55 uint16x4_t test_vadd_u16(uint16x4_t v1, uint16x4_t v2) { in test_vadd_u16()
186 uint16x4_t test_vsub_u16(uint16x4_t v1, uint16x4_t v2) { in test_vsub_u16()
313 uint16x4_t test_vmul_u16(uint16x4_t v1, uint16x4_t v2) { in test_vmul_u16()
444 uint16x4_t test_vmla_u16(uint16x4_t v1, uint16x4_t v2, uint16x4_t v3) { in test_vmla_u16()
565 uint16x4_t test_vmls_u16(uint16x4_t v1, uint16x4_t v2, uint16x4_t v3) { in test_vmls_u16()
787 uint16x4_t test_vaba_u16(uint16x4_t v1, uint16x4_t v2, uint16x4_t v3) { in test_vaba_u16()
910 uint16x4_t test_vabd_u16(uint16x4_t v1, uint16x4_t v2) { in test_vabd_u16()
1040 int8x8_t test_vbsl_s16(uint16x4_t v1, int16x4_t v2, int16x4_t v3) { in test_vbsl_s16()
1098 uint16x4_t test_vbsl_u16(uint16x4_t v1, uint16x4_t v2, uint16x4_t v3) { in test_vbsl_u16()
1191 poly16x4_t test_vbsl_p16(uint16x4_t v1, poly16x4_t v2, poly16x4_t v3) { in test_vbsl_p16()
[all …]
Daarch64-neon-across.c38 uint32_t test_vaddlv_u16(uint16x4_t a) { in test_vaddlv_u16()
126 uint16_t test_vmaxv_u16(uint16x4_t a) { in test_vmaxv_u16()
216 uint16_t test_vminv_u16(uint16x4_t a) { in test_vminv_u16()
306 uint16_t test_vaddv_u16(uint16x4_t a) { in test_vaddv_u16()
Daarch64-neon-extract.c103 uint16x4_t test_vext_u16(uint16x4_t a, uint16x4_t b) { in test_vext_u16()
Daarch64-neon-vcombine.c45 uint16x8_t test_vcombine_u16(uint16x4_t low, uint16x4_t high) { in test_vcombine_u16()
Darm-neon-vget.c28 uint16x4_t low_u16(uint16x8_t a) { in low_u16()
84 uint16x4_t high_u16(uint16x8_t a) { in high_u16()
/external/libaom/libaom/aom_dsp/arm/
Dintrapred_neon.c31 const uint16x4_t p0 = vpaddl_u8(A); // cascading summation of the top in dc_4x4()
32 const uint16x4_t p1 = vpadd_u16(p0, p0); in dc_4x4()
38 const uint16x4_t p0 = vpaddl_u8(L); // cascading summation of the left in dc_4x4()
39 const uint16x4_t p1 = vpadd_u16(p0, p0); in dc_4x4()
99 const uint16x4_t p0 = vpaddl_u8(A); // cascading summation of the top in dc_8x8()
100 const uint16x4_t p1 = vpadd_u16(p0, p0); in dc_8x8()
101 const uint16x4_t p2 = vpadd_u16(p1, p1); in dc_8x8()
107 const uint16x4_t p0 = vpaddl_u8(L); // cascading summation of the left in dc_8x8()
108 const uint16x4_t p1 = vpadd_u16(p0, p0); in dc_8x8()
109 const uint16x4_t p2 = vpadd_u16(p1, p1); in dc_8x8()
[all …]
/external/libaom/libaom/av1/common/arm/
Dcfl_neon.c33 static INLINE void vsth_u16(uint16_t *ptr, uint16x4_t val) { in vsth_u16()
50 const uint16x4_t top = vpaddl_u8(vldh_dup_u8(input)); in cfl_luma_subsampling_420_lbd_neon()
51 const uint16x4_t sum = vpadal_u8(top, vldh_dup_u8(input + input_stride)); in cfl_luma_subsampling_420_lbd_neon()
54 const uint16x4_t top = vpaddl_u8(vld1_u8(input)); in cfl_luma_subsampling_420_lbd_neon()
55 const uint16x4_t sum = vpadal_u8(top, vld1_u8(input + input_stride)); in cfl_luma_subsampling_420_lbd_neon()
88 const uint16x4_t top = vpaddl_u8(vldh_dup_u8(input)); in cfl_luma_subsampling_422_lbd_neon()
91 const uint16x4_t top = vpaddl_u8(vld1_u8(input)); in cfl_luma_subsampling_422_lbd_neon()
149 const uint16x4_t top = vld1_u16(input); in cfl_luma_subsampling_420_hbd_neon()
150 const uint16x4_t bot = vld1_u16(input + input_stride); in cfl_luma_subsampling_420_hbd_neon()
151 const uint16x4_t sum = vadd_u16(top, bot); in cfl_luma_subsampling_420_hbd_neon()
[all …]
Dblend_a64_vmask_neon.c31 uint16x4_t tmp0_16 = vdup_n_u16(0), tmp1_16 = vdup_n_u16(0); in aom_blend_a64_vmask_neon()
85 const uint16x4_t m1 = vdup_n_u16((uint16_t)mask[i]); in aom_blend_a64_vmask_neon()
86 const uint16x4_t m2 = vdup_n_u16((uint16_t)mask[i + 1]); in aom_blend_a64_vmask_neon()
88 const uint16x4_t max_minus_m1 = vdup_n_u16(64 - (uint16_t)mask[i]); in aom_blend_a64_vmask_neon()
89 const uint16x4_t max_minus_m2 = vdup_n_u16(64 - (uint16_t)mask[i + 1]); in aom_blend_a64_vmask_neon()
Dtranspose_neon.h188 static INLINE void transpose_u16_4x8(uint16x4_t *a0, uint16x4_t *a1, in transpose_u16_4x8()
189 uint16x4_t *a2, uint16x4_t *a3, in transpose_u16_4x8()
190 uint16x4_t *a4, uint16x4_t *a5, in transpose_u16_4x8()
191 uint16x4_t *a6, uint16x4_t *a7, in transpose_u16_4x8()
Dmem_neon.h80 uint16x4_t *const s0, uint16x4_t *const s1, in load_u16_4x4()
81 uint16x4_t *const s2, uint16x4_t *const s3) { in load_u16_4x4()
215 const uint16x4_t s0, const uint16x4_t s1, in store_u16_4x4()
216 const uint16x4_t s2, const uint16x4_t s3) { in store_u16_4x4()
396 uint16x4_t *tu0) { in load_unaligned_u8_2x2()
Dconvolve_neon.h25 uint16x4_t tmp0, tmp1; in wiener_convolve8_vert_4x8()
117 static INLINE uint16x4_t wiener_convolve8_horiz_4x8( in wiener_convolve8_horiz_4x8()
122 uint16x4_t res; in wiener_convolve8_horiz_4x8()
202 static INLINE uint16x4_t convolve8_4x4_s32( in convolve8_4x4_s32()
208 uint16x4_t res; in convolve8_4x4_s32()
/external/webp/src/dsp/
Drescaler_neon.c82 const uint16x4_t C0 = vmovn_u32(B0); in RescalerExportRowExpand_NEON()
83 const uint16x4_t C1 = vmovn_u32(B1); in RescalerExportRowExpand_NEON()
102 const uint16x4_t E0 = vmovn_u32(D0); in RescalerExportRowExpand_NEON()
103 const uint16x4_t E1 = vmovn_u32(D1); in RescalerExportRowExpand_NEON()
142 const uint16x4_t D0 = vmovn_u32(C0); in RescalerExportRowShrink_NEON()
143 const uint16x4_t D1 = vmovn_u32(C1); in RescalerExportRowShrink_NEON()
159 const uint16x4_t B0 = vmovn_u32(A0); in RescalerExportRowShrink_NEON()
160 const uint16x4_t B1 = vmovn_u32(A1); in RescalerExportRowShrink_NEON()
Dyuv_neon.c31 const uint16x4_t r_lo = vget_low_u16(r); in ConvertRGBToY_NEON()
32 const uint16x4_t r_hi = vget_high_u16(r); in ConvertRGBToY_NEON()
33 const uint16x4_t g_lo = vget_low_u16(g); in ConvertRGBToY_NEON()
34 const uint16x4_t g_hi = vget_high_u16(g); in ConvertRGBToY_NEON()
35 const uint16x4_t b_lo = vget_low_u16(b); in ConvertRGBToY_NEON()
36 const uint16x4_t b_hi = vget_high_u16(b); in ConvertRGBToY_NEON()
/external/libhevc/encoder/arm/
Dihevce_coarse_layer_sad_neon.c192 uint16x4_t tmp_a0, tmp_a1; in hme_store_4x4_sads_high_speed_neon()
212 uint16x4_t tmp_a; in hme_store_4x4_sads_high_speed_neon()
345 uint16x4_t tmp_b0, tmp_b1; in hme_store_4x4_sads_high_quality_neon()
374 uint16x4_t tmp_a, tmp_b; in hme_store_4x4_sads_high_quality_neon()
545 uint16x4_t curr = vld1_u16((U16 *)pi2_sads_4x4_current + sad_pos); in hme_combine_4x4_sads_and_compute_cost_high_speed_neon()
546 uint16x4_t south = vld1_u16((U16 *)pi2_sads_4x4_south + sad_pos); in hme_combine_4x4_sads_and_compute_cost_high_speed_neon()
547 uint16x4_t east = vld1_u16((U16 *)pi2_sads_4x4_east + sad_pos); in hme_combine_4x4_sads_and_compute_cost_high_speed_neon()
548 uint16x4_t sad_4x8 = vadd_u16(curr, south); in hme_combine_4x4_sads_and_compute_cost_high_speed_neon()
549 uint16x4_t sad_8x4 = vadd_u16(curr, east); in hme_combine_4x4_sads_and_compute_cost_high_speed_neon()
551 uint16x4_t mv_wt = in hme_combine_4x4_sads_and_compute_cost_high_speed_neon()
[all …]
/external/skqp/third_party/skcms/src/
DTransform_inl.h1046 (uint16x4_t)cast<U16>(to_fixed(r * 65535)), in exec_ops()
1047 (uint16x4_t)cast<U16>(to_fixed(g * 65535)), in exec_ops()
1048 (uint16x4_t)cast<U16>(to_fixed(b * 65535)), in exec_ops()
1065 (uint16x4_t)cast<U16>(to_fixed(r * 65535)), in exec_ops()
1066 (uint16x4_t)cast<U16>(to_fixed(g * 65535)), in exec_ops()
1067 (uint16x4_t)cast<U16>(to_fixed(b * 65535)), in exec_ops()
1068 (uint16x4_t)cast<U16>(to_fixed(a * 65535)), in exec_ops()
1086 (uint16x4_t)swap_endian_16(cast<U16>(to_fixed(r * 65535))), in exec_ops()
1087 (uint16x4_t)swap_endian_16(cast<U16>(to_fixed(g * 65535))), in exec_ops()
1088 (uint16x4_t)swap_endian_16(cast<U16>(to_fixed(b * 65535))), in exec_ops()
[all …]
/external/skia/third_party/skcms/src/
DTransform_inl.h1046 (uint16x4_t)cast<U16>(to_fixed(r * 65535)), in exec_ops()
1047 (uint16x4_t)cast<U16>(to_fixed(g * 65535)), in exec_ops()
1048 (uint16x4_t)cast<U16>(to_fixed(b * 65535)), in exec_ops()
1065 (uint16x4_t)cast<U16>(to_fixed(r * 65535)), in exec_ops()
1066 (uint16x4_t)cast<U16>(to_fixed(g * 65535)), in exec_ops()
1067 (uint16x4_t)cast<U16>(to_fixed(b * 65535)), in exec_ops()
1068 (uint16x4_t)cast<U16>(to_fixed(a * 65535)), in exec_ops()
1086 (uint16x4_t)swap_endian_16(cast<U16>(to_fixed(r * 65535))), in exec_ops()
1087 (uint16x4_t)swap_endian_16(cast<U16>(to_fixed(g * 65535))), in exec_ops()
1088 (uint16x4_t)swap_endian_16(cast<U16>(to_fixed(b * 65535))), in exec_ops()
[all …]

123