/external/libvpx/libvpx/vpx_dsp/arm/ |
D | intrapred_neon.c | 515 uint8x8_t d2u8 = vdup_n_u8(0); in vpx_h_predictor_16x16_neon() local 521 d2u8 = vget_low_u8(q1u8); in vpx_h_predictor_16x16_neon() 522 for (j = 0; j < 2; j++, d2u8 = vget_high_u8(q1u8)) { in vpx_h_predictor_16x16_neon() 523 q0u8 = vdupq_lane_u8(d2u8, 0); in vpx_h_predictor_16x16_neon() 526 q0u8 = vdupq_lane_u8(d2u8, 1); in vpx_h_predictor_16x16_neon() 529 q0u8 = vdupq_lane_u8(d2u8, 2); in vpx_h_predictor_16x16_neon() 532 q0u8 = vdupq_lane_u8(d2u8, 3); in vpx_h_predictor_16x16_neon() 535 q0u8 = vdupq_lane_u8(d2u8, 4); in vpx_h_predictor_16x16_neon() 538 q0u8 = vdupq_lane_u8(d2u8, 5); in vpx_h_predictor_16x16_neon() 541 q0u8 = vdupq_lane_u8(d2u8, 6); in vpx_h_predictor_16x16_neon() [all …]
|
D | idct8x8_1_add_neon.c | 20 uint8x8_t d2u8, d3u8, d30u8, d31u8; in vpx_idct8x8_1_add_neon() local 49 d2u8 = vqmovun_s16(vreinterpretq_s16_u16(q9u16)); in vpx_idct8x8_1_add_neon() 54 vst1_u64((uint64_t *)d2, vreinterpret_u64_u8(d2u8)); in vpx_idct8x8_1_add_neon()
|
D | idct16x16_1_add_neon.c | 20 uint8x8_t d2u8, d3u8, d30u8, d31u8; in vpx_idct16x16_1_add_neon() local 47 d2u8 = vqmovun_s16(vreinterpretq_s16_u16(q9u16)); in vpx_idct16x16_1_add_neon() 52 vst1_u64((uint64_t *)d2, vreinterpret_u64_u8(d2u8)); in vpx_idct16x16_1_add_neon()
|
D | vpx_convolve_copy_neon.c | 27 uint8x8_t d0u8, d2u8; in vpx_convolve_copy_neon() local 78 d2u8 = vld1_u8(src); in vpx_convolve_copy_neon() 83 vst1_u8(dst, d2u8); in vpx_convolve_copy_neon()
|
D | loopfilter_8_neon.c | 34 uint8x8_t d0u8, d1u8, d2u8, d19u8, d20u8, d21u8, d22u8, d23u8, d24u8; in mbloop_filter_neon() local 213 d2u8 = vbsl_u8(d20u8, dthresh, d24u8); in mbloop_filter_neon() 234 *d2ru8 = vbsl_u8(d20u8, d23u8, d2u8); in mbloop_filter_neon() 276 uint8x8_t d0u8, d1u8, d2u8, d3u8, d4u8, d5u8, d6u8, d7u8; in vpx_lpf_horizontal_8_neon() local 308 &d0u8, &d1u8, &d2u8, &d3u8, &d4u8, &d5u8); in vpx_lpf_horizontal_8_neon() 315 vst1_u8(s, d2u8); in vpx_lpf_horizontal_8_neon() 336 uint8x8_t d0u8, d1u8, d2u8, d3u8, d4u8, d5u8, d6u8, d7u8; in vpx_lpf_vertical_8_neon() local 408 &d0u8, &d1u8, &d2u8, &d3u8, &d4u8, &d5u8); in vpx_lpf_vertical_8_neon() 412 d4Result.val[2] = d2u8; in vpx_lpf_vertical_8_neon()
|
D | idct8x8_add_neon.c | 246 uint8x8_t d0u8, d1u8, d2u8, d3u8; in vpx_idct8x8_64_add_neon() local 303 d2u8 = vqmovun_s16(vreinterpretq_s16_u16(q10u16)); in vpx_idct8x8_64_add_neon() 310 vst1_u64((uint64_t *)d2, vreinterpret_u64_u8(d2u8)); in vpx_idct8x8_64_add_neon() 340 d2u8 = vqmovun_s16(vreinterpretq_s16_u16(q10u16)); in vpx_idct8x8_64_add_neon() 347 vst1_u64((uint64_t *)d2, vreinterpret_u64_u8(d2u8)); in vpx_idct8x8_64_add_neon() 359 uint8x8_t d0u8, d1u8, d2u8, d3u8; in vpx_idct8x8_12_add_neon() local 491 d2u8 = vqmovun_s16(vreinterpretq_s16_u16(q10u16)); in vpx_idct8x8_12_add_neon() 498 vst1_u64((uint64_t *)d2, vreinterpret_u64_u8(d2u8)); in vpx_idct8x8_12_add_neon() 528 d2u8 = vqmovun_s16(vreinterpretq_s16_u16(q10u16)); in vpx_idct8x8_12_add_neon() 535 vst1_u64((uint64_t *)d2, vreinterpret_u64_u8(d2u8)); in vpx_idct8x8_12_add_neon()
|
D | vpx_convolve_avg_neon.c | 28 uint8x8_t d0u8, d1u8, d2u8, d3u8; in vpx_convolve_avg_neon() local 111 d2u8 = vld1_u8(d); in vpx_convolve_avg_neon() 117 q1u8 = vcombine_u8(d2u8, d3u8); in vpx_convolve_avg_neon()
|
D | vpx_convolve8_avg_neon.c | 60 uint8x8_t d2u8, d3u8, d24u8, d25u8, d26u8, d27u8, d28u8, d29u8; in vpx_convolve8_avg_horiz_neon() local 195 d2u8 = vqmovn_u16(q1u16); in vpx_convolve8_avg_horiz_neon() 198 d0x2u16 = vtrn_u16(vreinterpret_u16_u8(d2u8), in vpx_convolve8_avg_horiz_neon() 248 uint8x8_t d2u8, d3u8; in vpx_convolve8_avg_vert_neon() local 345 d2u8 = vqmovn_u16(q1u16); in vpx_convolve8_avg_vert_neon() 348 q1u8 = vcombine_u8(d2u8, d3u8); in vpx_convolve8_avg_vert_neon()
|
D | variance_neon.c | 242 uint8x8_t d0u8, d2u8, d4u8, d6u8; in vpx_variance8x16_neon() local 257 d2u8 = vld1_u8(src_ptr); in vpx_variance8x16_neon() 268 q12u16 = vsubl_u8(d2u8, d6u8); in vpx_variance8x16_neon() 373 uint8x8_t d0u8, d1u8, d2u8, d3u8, d4u8, d5u8, d6u8, d7u8; in vpx_get4x4sse_cs_neon() local 386 d2u8 = vld1_u8(src_ptr); in vpx_get4x4sse_cs_neon() 397 q13u16 = vsubl_u8(d2u8, d6u8); in vpx_get4x4sse_cs_neon()
|
D | vpx_convolve8_neon.c | 60 uint8x8_t d2u8, d3u8, d24u8, d25u8, d26u8, d27u8, d28u8, d29u8; in vpx_convolve8_horiz_neon() local 188 d2u8 = vqmovn_u16(q1u16); in vpx_convolve8_horiz_neon() 191 d0x2u16 = vtrn_u16(vreinterpret_u16_u8(d2u8), in vpx_convolve8_horiz_neon()
|
/external/libvpx/libvpx/vp8/common/arm/neon/ |
D | sixtappredict_neon.c | 33 uint8x8_t d0u8, d1u8, d2u8, d3u8, d4u8, d5u8, d18u8, d19u8, d20u8, d21u8; in vp8_sixtap_predict4x4_neon() local 61 d2u8 = vreinterpret_u8_s8(vabs_s8(d2s8)); in vp8_sixtap_predict4x4_neon() 107 q3u16 = vmlal_u8(q3u16, d28u8, d2u8); in vp8_sixtap_predict4x4_neon() 108 q4u16 = vmlal_u8(q4u16, d29u8, d2u8); in vp8_sixtap_predict4x4_neon() 143 d2u8 = vreinterpret_u8_s8(vabs_s8(d2s8)); in vp8_sixtap_predict4x4_neon() 217 q7u16 = vmlal_u8(q7u16, vreinterpret_u8_u32(d0u32x2.val[0]), d2u8); in vp8_sixtap_predict4x4_neon() 218 q8u16 = vmlal_u8(q8u16, vreinterpret_u8_u32(d1u32x2.val[0]), d2u8); in vp8_sixtap_predict4x4_neon() 319 q7u16 = vmlal_u8(q7u16, vreinterpret_u8_u32(d0u32x2.val[0]), d2u8); in vp8_sixtap_predict4x4_neon() 320 q8u16 = vmlal_u8(q8u16, vreinterpret_u8_u32(d1u32x2.val[0]), d2u8); in vp8_sixtap_predict4x4_neon() 321 q12u16 = vmlal_u8(q12u16, d31u8, d2u8); in vp8_sixtap_predict4x4_neon() [all …]
|
D | bilinearpredict_neon.c | 31 uint8x8_t d0u8, d1u8, d2u8, d3u8, d4u8, d5u8, d6u8; in vp8_bilinear_predict4x4_neon() local 58 d2u8 = vld1_u8(src_ptr); src_ptr += src_pixels_per_line; in vp8_bilinear_predict4x4_neon() 64 q1u8 = vcombine_u8(d2u8, d3u8); in vp8_bilinear_predict4x4_neon() 118 d2u8 = vqrshrn_n_u16(q1u16, 7); in vp8_bilinear_predict4x4_neon() 121 vst1_lane_u32((uint32_t *)dst_ptr, vreinterpret_u32_u8(d2u8), 0); in vp8_bilinear_predict4x4_neon() 123 vst1_lane_u32((uint32_t *)dst_ptr, vreinterpret_u32_u8(d2u8), 1); in vp8_bilinear_predict4x4_neon() 139 uint8x8_t d0u8, d1u8, d2u8, d3u8, d4u8, d5u8; in vp8_bilinear_predict8x4_neon() local 206 d2u8 = vqrshrn_n_u16(q1u16, 7); in vp8_bilinear_predict8x4_neon() 211 vst1_u8((uint8_t *)dst_ptr, d2u8); dst_ptr += dst_pitch; in vp8_bilinear_predict8x4_neon() 226 uint8x8_t d0u8, d1u8, d2u8, d3u8, d4u8, d5u8, d6u8, d7u8, d8u8, d9u8, d11u8; in vp8_bilinear_predict8x8_neon() local [all …]
|
D | dc_only_idct_add_neon.c | 22 uint8x8_t d2u8; in vp8_dc_only_idct_add_neon() local 35 d2u8 = vqmovun_s16(vreinterpretq_s16_u16(q1u16)); in vp8_dc_only_idct_add_neon() 37 vst1_lane_u32((uint32_t *)dst_ptr, vreinterpret_u32_u8(d2u8), 0); in vp8_dc_only_idct_add_neon() 39 vst1_lane_u32((uint32_t *)dst_ptr, vreinterpret_u32_u8(d2u8), 1); in vp8_dc_only_idct_add_neon()
|
D | idct_dequant_0_2x_neon.c | 23 uint8x8_t d2u8, d4u8; in idct_dequant_0_2x_neon() local 47 d2u8 = vqmovun_s16(vreinterpretq_s16_u16(q1u16)); in idct_dequant_0_2x_neon() 50 d2s32 = vreinterpret_s32_u8(d2u8); in idct_dequant_0_2x_neon()
|
/external/libvpx/libvpx/vp9/common/arm/neon/ |
D | vp9_iht8x8_add_neon.c | 493 uint8x8_t d0u8, d1u8, d2u8, d3u8; in vp9_iht8x8_64_add_neon() local 611 d2u8 = vqmovun_s16(vreinterpretq_s16_u16(q10u16)); in vp9_iht8x8_64_add_neon() 618 vst1_u64((uint64_t *)d2, vreinterpret_u64_u8(d2u8)); in vp9_iht8x8_64_add_neon()
|