/external/libvpx/libvpx/vpx_dsp/arm/ |
D | subtract_neon.c | 32 const uint16x8_t d0 = vsubl_u8(vget_low_u8(s0), vget_low_u8(p0)); in vpx_subtract_block_neon() 33 const uint16x8_t d1 = vsubl_u8(vget_high_u8(s0), vget_high_u8(p0)); in vpx_subtract_block_neon() 34 const uint16x8_t d2 = vsubl_u8(vget_low_u8(s1), vget_low_u8(p1)); in vpx_subtract_block_neon() 35 const uint16x8_t d3 = vsubl_u8(vget_high_u8(s1), vget_high_u8(p1)); in vpx_subtract_block_neon() 49 const uint16x8_t d0 = vsubl_u8(vget_low_u8(s), vget_low_u8(p)); in vpx_subtract_block_neon() 50 const uint16x8_t d1 = vsubl_u8(vget_high_u8(s), vget_high_u8(p)); in vpx_subtract_block_neon() 61 const uint16x8_t v_diff = vsubl_u8(s, p); in vpx_subtract_block_neon() 72 const uint16x8_t d = vsubl_u8(s, p); in vpx_subtract_block_neon()
|
D | variance_neon.c | 45 vsubl_u8(vget_low_u8(a_u8), vget_low_u8(b_u8)); in variance_neon_w4x4() 47 vsubl_u8(vget_high_u8(a_u8), vget_high_u8(b_u8)); in variance_neon_w4x4() 94 vsubl_u8(vget_low_u8(a_u8), vget_low_u8(b_u8)); in variance_neon_w16() 96 vsubl_u8(vget_high_u8(a_u8), vget_high_u8(b_u8)); in variance_neon_w16() 141 const uint16x8_t diff_0_u16 = vsubl_u8(a_0_u8, b_0_u8); in variance_neon_w8x2() 142 const uint16x8_t diff_1_u16 = vsubl_u8(a_1_u8, b_1_u8); in variance_neon_w8x2() 293 q11u16 = vsubl_u8(vget_low_u8(q0u8), vget_low_u8(q2u8)); in vpx_mse16x16_neon() 294 q12u16 = vsubl_u8(vget_high_u8(q0u8), vget_high_u8(q2u8)); in vpx_mse16x16_neon() 295 q13u16 = vsubl_u8(vget_low_u8(q1u8), vget_low_u8(q3u8)); in vpx_mse16x16_neon() 296 q14u16 = vsubl_u8(vget_high_u8(q1u8), vget_high_u8(q3u8)); in vpx_mse16x16_neon() [all …]
|
D | intrapred_neon.c | 887 sub = vreinterpretq_s16_u16(vsubl_u8(above_u8, top_left)); in vpx_tm_predictor_4x4_neon() 922 const int16x8_t sub = vreinterpretq_s16_u16(vsubl_u8(above_u8, top_left)); in vpx_tm_predictor_8x8_neon() 958 vsubl_u8(vget_low_u8(above_u8), vget_low_u8(top_left))); in vpx_tm_predictor_16x16_neon() 960 vsubl_u8(vget_high_u8(above_u8), vget_high_u8(top_left))); in vpx_tm_predictor_16x16_neon() 1015 vsubl_u8(vget_low_u8(above_low), vget_low_u8(top_left))); in vpx_tm_predictor_32x32_neon() 1017 vsubl_u8(vget_high_u8(above_low), vget_high_u8(top_left))); in vpx_tm_predictor_32x32_neon() 1019 vsubl_u8(vget_low_u8(above_high), vget_low_u8(top_left))); in vpx_tm_predictor_32x32_neon() 1021 vsubl_u8(vget_high_u8(above_high), vget_high_u8(top_left))); in vpx_tm_predictor_32x32_neon()
|
D | deblock_neon.c | 353 x = vreinterpretq_s16_u16(vsubl_u8(right_context, left_context)); in vpx_mbpost_proc_across_ip_neon() 450 x = vreinterpretq_s16_u16(vsubl_u8(below_context, above_context[0])); in vpx_mbpost_proc_down_neon()
|
/external/libaom/libaom/aom_dsp/arm/ |
D | subtract_neon.c | 32 vsubl_u8(vget_low_u8(v_src_00), vget_low_u8(v_pred_00)); in aom_subtract_block_neon() 34 vsubl_u8(vget_high_u8(v_src_00), vget_high_u8(v_pred_00)); in aom_subtract_block_neon() 36 vsubl_u8(vget_low_u8(v_src_16), vget_low_u8(v_pred_16)); in aom_subtract_block_neon() 38 vsubl_u8(vget_high_u8(v_src_16), vget_high_u8(v_pred_16)); in aom_subtract_block_neon() 53 vsubl_u8(vget_low_u8(v_src), vget_low_u8(v_pred)); in aom_subtract_block_neon() 55 vsubl_u8(vget_high_u8(v_src), vget_high_u8(v_pred)); in aom_subtract_block_neon() 66 const uint16x8_t v_diff = vsubl_u8(v_src, v_pred); in aom_subtract_block_neon()
|
D | variance_neon.c | 33 const uint16x8_t v_diff = vsubl_u8(v_a, v_b); in variance_neon_w8() 181 q11u16 = vsubl_u8(vget_low_u8(q0u8), vget_low_u8(q2u8)); in aom_variance16x8_neon() 182 q12u16 = vsubl_u8(vget_high_u8(q0u8), vget_high_u8(q2u8)); in aom_variance16x8_neon() 183 q13u16 = vsubl_u8(vget_low_u8(q1u8), vget_low_u8(q3u8)); in aom_variance16x8_neon() 184 q14u16 = vsubl_u8(vget_high_u8(q1u8), vget_high_u8(q3u8)); in aom_variance16x8_neon() 257 q11u16 = vsubl_u8(d0u8, d4u8); in aom_variance8x16_neon() 258 q12u16 = vsubl_u8(d2u8, d6u8); in aom_variance8x16_neon() 315 q11u16 = vsubl_u8(vget_low_u8(q0u8), vget_low_u8(q2u8)); in aom_mse16x16_neon() 316 q12u16 = vsubl_u8(vget_high_u8(q0u8), vget_high_u8(q2u8)); in aom_mse16x16_neon() 317 q13u16 = vsubl_u8(vget_low_u8(q1u8), vget_low_u8(q3u8)); in aom_mse16x16_neon() [all …]
|
/external/XNNPACK/src/qu8-vadd/ |
D | minmax-neon.c | 39 const int16x8_t vxa0 = vreinterpretq_s16_u16(vsubl_u8(vget_low_u8(va01), va_zero_point)); in xnn_qu8_vadd_minmax_ukernel__neon() 40 const int16x8_t vxb0 = vreinterpretq_s16_u16(vsubl_u8(vget_low_u8(vb01), vb_zero_point)); in xnn_qu8_vadd_minmax_ukernel__neon() 41 const int16x8_t vxa1 = vreinterpretq_s16_u16(vsubl_u8(vget_high_u8(va01), va_zero_point)); in xnn_qu8_vadd_minmax_ukernel__neon() 42 const int16x8_t vxb1 = vreinterpretq_s16_u16(vsubl_u8(vget_high_u8(vb01), vb_zero_point)); in xnn_qu8_vadd_minmax_ukernel__neon() 43 const int16x8_t vxa2 = vreinterpretq_s16_u16(vsubl_u8(vget_low_u8(va23), va_zero_point)); in xnn_qu8_vadd_minmax_ukernel__neon() 44 const int16x8_t vxb2 = vreinterpretq_s16_u16(vsubl_u8(vget_low_u8(vb23), vb_zero_point)); in xnn_qu8_vadd_minmax_ukernel__neon() 45 const int16x8_t vxa3 = vreinterpretq_s16_u16(vsubl_u8(vget_high_u8(va23), va_zero_point)); in xnn_qu8_vadd_minmax_ukernel__neon() 46 const int16x8_t vxb3 = vreinterpretq_s16_u16(vsubl_u8(vget_high_u8(vb23), vb_zero_point)); in xnn_qu8_vadd_minmax_ukernel__neon() 109 const int16x8_t vxa0 = vreinterpretq_s16_u16(vsubl_u8(vget_low_u8(va01), va_zero_point)); in xnn_qu8_vadd_minmax_ukernel__neon() 110 const int16x8_t vxb0 = vreinterpretq_s16_u16(vsubl_u8(vget_low_u8(vb01), vb_zero_point)); in xnn_qu8_vadd_minmax_ukernel__neon() [all …]
|
/external/XNNPACK/src/qu8-dwconv/ |
D | up8x9-minmax-neon.c | 82 const int16x8_t vxk0 = vreinterpretq_s16_u16(vsubl_u8(vk0, vkernel_zero_point)); in xnn_qu8_dwconv_minmax_ukernel_up8x9__neon() 89 const int16x8_t vxk1 = vreinterpretq_s16_u16(vsubl_u8(vk1, vkernel_zero_point)); in xnn_qu8_dwconv_minmax_ukernel_up8x9__neon() 96 const int16x8_t vxk2 = vreinterpretq_s16_u16(vsubl_u8(vk2, vkernel_zero_point)); in xnn_qu8_dwconv_minmax_ukernel_up8x9__neon() 103 const int16x8_t vxk3 = vreinterpretq_s16_u16(vsubl_u8(vk3, vkernel_zero_point)); in xnn_qu8_dwconv_minmax_ukernel_up8x9__neon() 110 const int16x8_t vxk4 = vreinterpretq_s16_u16(vsubl_u8(vk4, vkernel_zero_point)); in xnn_qu8_dwconv_minmax_ukernel_up8x9__neon() 117 const int16x8_t vxk5 = vreinterpretq_s16_u16(vsubl_u8(vk5, vkernel_zero_point)); in xnn_qu8_dwconv_minmax_ukernel_up8x9__neon() 124 const int16x8_t vxk6 = vreinterpretq_s16_u16(vsubl_u8(vk6, vkernel_zero_point)); in xnn_qu8_dwconv_minmax_ukernel_up8x9__neon() 131 const int16x8_t vxk7 = vreinterpretq_s16_u16(vsubl_u8(vk7, vkernel_zero_point)); in xnn_qu8_dwconv_minmax_ukernel_up8x9__neon() 138 const int16x8_t vxk8 = vreinterpretq_s16_u16(vsubl_u8(vk8, vkernel_zero_point)); in xnn_qu8_dwconv_minmax_ukernel_up8x9__neon() 173 const int16x8_t vxk0 = vreinterpretq_s16_u16(vsubl_u8(vk0, vkernel_zero_point)); in xnn_qu8_dwconv_minmax_ukernel_up8x9__neon() [all …]
|
/external/libhevc/encoder/arm/ |
D | ihevce_common_utils_neon.c | 541 vsubl_u8(vget_low_u8(src_buf_8x16), vget_low_u8(recon_buf_8x16))); in ihevce_get_luma_eo_sao_params_neon() 543 vsubl_u8(vget_high_u8(src_buf_8x16), vget_high_u8(recon_buf_8x16))); in ihevce_get_luma_eo_sao_params_neon() 547 vsubl_u8(vget_low_u8(recon_buf_8x16), vget_low_u8(recon_buf0_8x16))); in ihevce_get_luma_eo_sao_params_neon() 553 vsubl_u8(vget_low_u8(recon_buf_8x16), vget_low_u8(recon_buf1_8x16))); in ihevce_get_luma_eo_sao_params_neon() 559 vsubl_u8(vget_high_u8(recon_buf_8x16), vget_high_u8(recon_buf0_8x16))); in ihevce_get_luma_eo_sao_params_neon() 565 vsubl_u8(vget_high_u8(recon_buf_8x16), vget_high_u8(recon_buf1_8x16))); in ihevce_get_luma_eo_sao_params_neon() 644 pel_error = vreinterpretq_s16_u16(vsubl_u8(src_buf, recon_buf)); in ihevce_get_luma_eo_sao_params_neon() 646 sign_reg0 = vreinterpretq_s16_u16(vsubl_u8(recon_buf, recon_buf0)); in ihevce_get_luma_eo_sao_params_neon() 651 sign_reg1 = vreinterpretq_s16_u16(vsubl_u8(recon_buf, recon_buf1)); in ihevce_get_luma_eo_sao_params_neon() 726 vsubl_u8(vget_low_u8(src_buf_8x16), vget_low_u8(recon_buf_8x16))); in ihevce_get_luma_eo_sao_params_neon() [all …]
|
D | ihevce_had_compute_neon.c | 67 *r##k = vreinterpretq_s16_u16(vsubl_u8(s##k, p##k)); \ 75 *r##k = vreinterpretq_s16_u16(vsubl_u8(s##k, p##k)); \ 401 res_01 = vreinterpretq_s16_u16(vsubl_u8(vget_low_u8(src_u8), vget_low_u8(pred_u8))); in ihevce_HAD_4x4_8bit_plane_neon() 402 res_23 = vreinterpretq_s16_u16(vsubl_u8(vget_high_u8(src_u8), vget_high_u8(pred_u8))); in ihevce_HAD_4x4_8bit_plane_neon()
|
/external/libjpeg-turbo/simd/arm/ |
D | jquanti-neon.c | 58 vreinterpretq_s16_u16(vsubl_u8(samp_row0, vdup_n_u8(CENTERJSAMPLE))); in jsimd_convsamp_neon() 60 vreinterpretq_s16_u16(vsubl_u8(samp_row1, vdup_n_u8(CENTERJSAMPLE))); in jsimd_convsamp_neon() 62 vreinterpretq_s16_u16(vsubl_u8(samp_row2, vdup_n_u8(CENTERJSAMPLE))); in jsimd_convsamp_neon() 64 vreinterpretq_s16_u16(vsubl_u8(samp_row3, vdup_n_u8(CENTERJSAMPLE))); in jsimd_convsamp_neon() 66 vreinterpretq_s16_u16(vsubl_u8(samp_row4, vdup_n_u8(CENTERJSAMPLE))); in jsimd_convsamp_neon() 68 vreinterpretq_s16_u16(vsubl_u8(samp_row5, vdup_n_u8(CENTERJSAMPLE))); in jsimd_convsamp_neon() 70 vreinterpretq_s16_u16(vsubl_u8(samp_row6, vdup_n_u8(CENTERJSAMPLE))); in jsimd_convsamp_neon() 72 vreinterpretq_s16_u16(vsubl_u8(samp_row7, vdup_n_u8(CENTERJSAMPLE))); in jsimd_convsamp_neon()
|
/external/tensorflow/tensorflow/core/kernels/ |
D | quantized_mul_op.cc | 73 vsubl_u8(full_input_high_8x8, full_input_offset_8x8)); in ScalarMultiply() 75 vsubl_u8(full_input_low_8x8, full_input_offset_8x8)); in ScalarMultiply() 147 vreinterpretq_s16_u16(vsubl_u8(x_high_8x8, offset_x_8x8)); in VectorMultiply() 149 vreinterpretq_s16_u16(vsubl_u8(x_low_8x8, offset_x_8x8)); in VectorMultiply() 151 vreinterpretq_s16_u16(vsubl_u8(y_high_8x8, offset_y_8x8)); in VectorMultiply() 153 vreinterpretq_s16_u16(vsubl_u8(y_low_8x8, offset_y_8x8)); in VectorMultiply() 232 vreinterpretq_s16_u16(vsubl_u8(x_high_8x8, offset_x_8x8)); in VectorTensorMultiply() 234 vreinterpretq_s16_u16(vsubl_u8(x_low_8x8, offset_x_8x8)); in VectorTensorMultiply() 236 vreinterpretq_s16_u16(vsubl_u8(y_high_8x8, offset_y_8x8)); in VectorTensorMultiply() 238 vreinterpretq_s16_u16(vsubl_u8(y_low_8x8, offset_y_8x8)); in VectorTensorMultiply()
|
D | quantization_utils.h | 531 vreinterpretq_s16_u16(vsubl_u8(top_right8x8, top_left8x8)); in ComputeLerp8x8() 536 vreinterpretq_s16_u16(vsubl_u8(bottom_right8x8, bottom_left8x8)); in ComputeLerp8x8()
|
/external/XNNPACK/src/qu8-igemm/ |
D | 4x8-minmax-neon.c | 97 const int16x8_t vxb01234567 = vreinterpretq_s16_u16(vsubl_u8(vb01234567, vb_zero_point)); in xnn_qu8_igemm_minmax_ukernel_4x8__neon() 111 const int16x8_t vxb01234567 = vreinterpretq_s16_u16(vsubl_u8(vb01234567, vb_zero_point)); in xnn_qu8_igemm_minmax_ukernel_4x8__neon() 125 const int16x8_t vxb01234567 = vreinterpretq_s16_u16(vsubl_u8(vb01234567, vb_zero_point)); in xnn_qu8_igemm_minmax_ukernel_4x8__neon() 139 const int16x8_t vxb01234567 = vreinterpretq_s16_u16(vsubl_u8(vb01234567, vb_zero_point)); in xnn_qu8_igemm_minmax_ukernel_4x8__neon() 153 const int16x8_t vxb01234567 = vreinterpretq_s16_u16(vsubl_u8(vb01234567, vb_zero_point)); in xnn_qu8_igemm_minmax_ukernel_4x8__neon() 167 const int16x8_t vxb01234567 = vreinterpretq_s16_u16(vsubl_u8(vb01234567, vb_zero_point)); in xnn_qu8_igemm_minmax_ukernel_4x8__neon() 181 const int16x8_t vxb01234567 = vreinterpretq_s16_u16(vsubl_u8(vb01234567, vb_zero_point)); in xnn_qu8_igemm_minmax_ukernel_4x8__neon() 195 const int16x8_t vxb01234567 = vreinterpretq_s16_u16(vsubl_u8(vb01234567, vb_zero_point)); in xnn_qu8_igemm_minmax_ukernel_4x8__neon() 221 const int16x8_t vxb01234567 = vreinterpretq_s16_u16(vsubl_u8(vb01234567, vb_zero_point)); in xnn_qu8_igemm_minmax_ukernel_4x8__neon() 235 const int16x8_t vxb01234567 = vreinterpretq_s16_u16(vsubl_u8(vb01234567, vb_zero_point)); in xnn_qu8_igemm_minmax_ukernel_4x8__neon() [all …]
|
D | 8x8-minmax-neon.c | 145 const int16x8_t vxb01234567 = vreinterpretq_s16_u16(vsubl_u8(vb01234567, vb_zero_point)); in xnn_qu8_igemm_minmax_ukernel_8x8__neon() 167 const int16x8_t vxb01234567 = vreinterpretq_s16_u16(vsubl_u8(vb01234567, vb_zero_point)); in xnn_qu8_igemm_minmax_ukernel_8x8__neon() 189 const int16x8_t vxb01234567 = vreinterpretq_s16_u16(vsubl_u8(vb01234567, vb_zero_point)); in xnn_qu8_igemm_minmax_ukernel_8x8__neon() 211 const int16x8_t vxb01234567 = vreinterpretq_s16_u16(vsubl_u8(vb01234567, vb_zero_point)); in xnn_qu8_igemm_minmax_ukernel_8x8__neon() 233 const int16x8_t vxb01234567 = vreinterpretq_s16_u16(vsubl_u8(vb01234567, vb_zero_point)); in xnn_qu8_igemm_minmax_ukernel_8x8__neon() 255 const int16x8_t vxb01234567 = vreinterpretq_s16_u16(vsubl_u8(vb01234567, vb_zero_point)); in xnn_qu8_igemm_minmax_ukernel_8x8__neon() 277 const int16x8_t vxb01234567 = vreinterpretq_s16_u16(vsubl_u8(vb01234567, vb_zero_point)); in xnn_qu8_igemm_minmax_ukernel_8x8__neon() 299 const int16x8_t vxb01234567 = vreinterpretq_s16_u16(vsubl_u8(vb01234567, vb_zero_point)); in xnn_qu8_igemm_minmax_ukernel_8x8__neon() 341 const int16x8_t vxb01234567 = vreinterpretq_s16_u16(vsubl_u8(vb01234567, vb_zero_point)); in xnn_qu8_igemm_minmax_ukernel_8x8__neon() 363 const int16x8_t vxb01234567 = vreinterpretq_s16_u16(vsubl_u8(vb01234567, vb_zero_point)); in xnn_qu8_igemm_minmax_ukernel_8x8__neon() [all …]
|
/external/XNNPACK/src/qu8-gemm/ |
D | 4x8-minmax-neon.c | 79 const int16x8_t vxb01234567c0 = vreinterpretq_s16_u16(vsubl_u8(vb01234567c0, vb_zero_point)); in xnn_qu8_gemm_minmax_ukernel_4x8__neon() 91 const int16x8_t vxb01234567c1 = vreinterpretq_s16_u16(vsubl_u8(vb01234567c1, vb_zero_point)); in xnn_qu8_gemm_minmax_ukernel_4x8__neon() 103 const int16x8_t vxb01234567c2 = vreinterpretq_s16_u16(vsubl_u8(vb01234567c2, vb_zero_point)); in xnn_qu8_gemm_minmax_ukernel_4x8__neon() 115 const int16x8_t vxb01234567c3 = vreinterpretq_s16_u16(vsubl_u8(vb01234567c3, vb_zero_point)); in xnn_qu8_gemm_minmax_ukernel_4x8__neon() 127 const int16x8_t vxb01234567c4 = vreinterpretq_s16_u16(vsubl_u8(vb01234567c4, vb_zero_point)); in xnn_qu8_gemm_minmax_ukernel_4x8__neon() 139 const int16x8_t vxb01234567c5 = vreinterpretq_s16_u16(vsubl_u8(vb01234567c5, vb_zero_point)); in xnn_qu8_gemm_minmax_ukernel_4x8__neon() 151 const int16x8_t vxb01234567c6 = vreinterpretq_s16_u16(vsubl_u8(vb01234567c6, vb_zero_point)); in xnn_qu8_gemm_minmax_ukernel_4x8__neon() 163 const int16x8_t vxb01234567c7 = vreinterpretq_s16_u16(vsubl_u8(vb01234567c7, vb_zero_point)); in xnn_qu8_gemm_minmax_ukernel_4x8__neon() 187 const int16x8_t vxb01234567c0 = vreinterpretq_s16_u16(vsubl_u8(vb01234567c0, vb_zero_point)); in xnn_qu8_gemm_minmax_ukernel_4x8__neon() 200 … const int16x8_t vxb01234567c1 = vreinterpretq_s16_u16(vsubl_u8(vb01234567c1, vb_zero_point)); in xnn_qu8_gemm_minmax_ukernel_4x8__neon() [all …]
|
D | 8x8-minmax-neon.c | 119 const int16x8_t vxb01234567c0 = vreinterpretq_s16_u16(vsubl_u8(vb01234567c0, vb_zero_point)); in xnn_qu8_gemm_minmax_ukernel_8x8__neon() 139 const int16x8_t vxb01234567c1 = vreinterpretq_s16_u16(vsubl_u8(vb01234567c1, vb_zero_point)); in xnn_qu8_gemm_minmax_ukernel_8x8__neon() 159 const int16x8_t vxb01234567c2 = vreinterpretq_s16_u16(vsubl_u8(vb01234567c2, vb_zero_point)); in xnn_qu8_gemm_minmax_ukernel_8x8__neon() 179 const int16x8_t vxb01234567c3 = vreinterpretq_s16_u16(vsubl_u8(vb01234567c3, vb_zero_point)); in xnn_qu8_gemm_minmax_ukernel_8x8__neon() 199 const int16x8_t vxb01234567c4 = vreinterpretq_s16_u16(vsubl_u8(vb01234567c4, vb_zero_point)); in xnn_qu8_gemm_minmax_ukernel_8x8__neon() 219 const int16x8_t vxb01234567c5 = vreinterpretq_s16_u16(vsubl_u8(vb01234567c5, vb_zero_point)); in xnn_qu8_gemm_minmax_ukernel_8x8__neon() 239 const int16x8_t vxb01234567c6 = vreinterpretq_s16_u16(vsubl_u8(vb01234567c6, vb_zero_point)); in xnn_qu8_gemm_minmax_ukernel_8x8__neon() 259 const int16x8_t vxb01234567c7 = vreinterpretq_s16_u16(vsubl_u8(vb01234567c7, vb_zero_point)); in xnn_qu8_gemm_minmax_ukernel_8x8__neon() 299 const int16x8_t vxb01234567c0 = vreinterpretq_s16_u16(vsubl_u8(vb01234567c0, vb_zero_point)); in xnn_qu8_gemm_minmax_ukernel_8x8__neon() 320 … const int16x8_t vxb01234567c1 = vreinterpretq_s16_u16(vsubl_u8(vb01234567c1, vb_zero_point)); in xnn_qu8_gemm_minmax_ukernel_8x8__neon() [all …]
|
/external/libhevc/common/arm/ |
D | ihevc_resi_trans_neon_32x32.c | 126 vsubl_u8(vget_low_u8(src_buff), vget_low_u8(pred_buff))); in ihevc_resi_trans_32x32_neon() 128 vsubl_u8(vget_high_u8(src_buff), vget_high_u8(pred_buff))); in ihevc_resi_trans_32x32_neon() 135 vsubl_u8(vget_low_u8(src_buff), vget_low_u8(pred_buff)))); in ihevc_resi_trans_32x32_neon() 139 vsubl_u8(vget_high_u8(src_buff), vget_high_u8(pred_buff)))); in ihevc_resi_trans_32x32_neon() 151 vsubl_u8(vget_low_u8(src_buff), vget_low_u8(pred_buff))); in ihevc_resi_trans_32x32_neon() 153 vsubl_u8(vget_high_u8(src_buff), vget_high_u8(pred_buff))); in ihevc_resi_trans_32x32_neon() 160 vsubl_u8(vget_low_u8(src_buff), vget_low_u8(pred_buff)))); in ihevc_resi_trans_32x32_neon() 164 vsubl_u8(vget_high_u8(src_buff), vget_high_u8(pred_buff)))); in ihevc_resi_trans_32x32_neon()
|
D | ihevc_resi_trans_neon.c | 107 diff_1 = vreinterpretq_s16_u16(vsubl_u8(vget_low_u8(inp_buf), vget_low_u8(pred_buf))); in ihevc_resi_trans_4x4_neon() 108 diff_2 = vreinterpretq_s16_u16(vsubl_u8(vget_high_u8(inp_buf), vget_high_u8(pred_buf))); in ihevc_resi_trans_4x4_neon() 237 vreinterpretq_s16_u16(vsubl_u8(vget_low_u8(src_u8), vget_low_u8(pred_u8))); in ihevc_resi_trans_4x4_ttype1_neon() 239 vreinterpretq_s16_u16(vsubl_u8(vget_high_u8(src_u8), vget_high_u8(pred_u8))); in ihevc_resi_trans_4x4_ttype1_neon() 419 diff_16[k] = vreinterpretq_s16_u16(vsubl_u8(s##k, p##k)); \ in ihevc_resi_trans_8x8_neon() 428 diff_16[k] = vreinterpretq_s16_u16(vsubl_u8(s##k, p##k)); \ in ihevc_resi_trans_8x8_neon() 885 c[i] = vreinterpretq_s16_u16(vsubl_u8(a[i], b[i])); in diff()
|
D | ihevc_intra_pred_filters_neon_intr.c | 1250 sub_val = vsubl_u8(vreinterpret_u8_u32(pu1_ref_val1), dup_sub); in ihevc_intra_pred_luma_horz_neonintr() 1310 sub_res = vsubl_u8(src_tmp, dup_sub); in ihevc_intra_pred_luma_horz_neonintr() 1488 sub_val = vsubl_u8(vreinterpret_u8_u32(src_val1), dup_2_sub); in ihevc_intra_pred_luma_ver_neonintr() 1585 sub_val = vsubl_u8(pu1_src_tmp1, dup_sub); in ihevc_intra_pred_luma_ver_neonintr()
|
/external/webrtc/modules/video_processing/util/ |
D | denoiser_filter_neon.cc | 50 const uint16x8_t v_diff = vsubl_u8(v_a, v_b); in VarianceNeonW8()
|
/external/webp/src/dsp/ |
D | enc_neon.c | 293 return vreinterpretq_s16_u16(vsubl_u8(a, b)); in DiffU8ToS16_NEON() 607 const int16x8_t q_a2 = vreinterpretq_s16_u16(vsubl_u8(q4_in.val[1], in DistoVerticalPass_NEON() 609 const int16x8_t q_a3 = vreinterpretq_s16_u16(vsubl_u8(q4_in.val[0], in DistoVerticalPass_NEON()
|
D | lossless_neon.c | 436 vreinterpretq_s16_u16(vsubl_u8(vget_low_u8(T), vget_low_u8(TL))); in PredictorAdd12_NEON() 438 vreinterpretq_s16_u16(vsubl_u8(vget_high_u8(T), vget_high_u8(TL))); in PredictorAdd12_NEON()
|
D | dec_neon.c | 1306 const int16x8_t d = vreinterpretq_s16_u16(vsubl_u8(T, TL)); // A[c] - A[-1] in TrueMotion_NEON() 1578 const int16x8_t d_lo = vreinterpretq_s16_u16(vsubl_u8(vget_low_u8(T), TL)); in TM16_NEON() 1579 const int16x8_t d_hi = vreinterpretq_s16_u16(vsubl_u8(vget_high_u8(T), TL)); in TM16_NEON()
|
/external/libgav1/libgav1/src/dsp/arm/ |
D | loop_filter_neon.cc | 87 const int16x8_t q0mp0_p1mq1 = vreinterpretq_s16_u16(vsubl_u8(q0p1, p0q1)); in Filter4()
|