/external/libhevc/common/arm/ |
D | ihevc_inter_pred_chroma_copy.s | 128 vld1.32 {d0[0]},[r0] @vld1_lane_u32((uint32_t *)pu1_src_tmp, src_tmp, 0) 131 vst1.32 {d0[0]},[r1] @vst1_lane_u32((uint32_t *)pu1_dst_tmp, src_tmp, 0) 132 vld1.32 {d0[0]},[r5],r2 @vld1_lane_u32((uint32_t *)pu1_src_tmp, src_tmp, 0) 134 vst1.32 {d0[0]},[r6],r3 @vst1_lane_u32((uint32_t *)pu1_dst_tmp, src_tmp, 0) 135 vld1.32 {d0[0]},[r5],r2 @vld1_lane_u32((uint32_t *)pu1_src_tmp, src_tmp, 0) 137 vst1.32 {d0[0]},[r6],r3 @vst1_lane_u32((uint32_t *)pu1_dst_tmp, src_tmp, 0) 138 vld1.32 {d0[0]},[r5],r2 @vld1_lane_u32((uint32_t *)pu1_src_tmp, src_tmp, 0) 140 vst1.32 {d0[0]},[r6],r3 @vst1_lane_u32((uint32_t *)pu1_dst_tmp, src_tmp, 0) 160 vld1.32 {d0[0]},[r0] @vld1_lane_u32((uint32_t *)pu1_src_tmp, src_tmp, 0) 163 vst1.32 {d0[0]},[r1] @vst1_lane_u32((uint32_t *)pu1_dst_tmp, src_tmp, 0) [all …]
|
D | ihevc_inter_pred_luma_copy.s | 101 vld1.32 {d0[0]},[r0] @vld1_lane_u32((uint32_t *)pu1_src_tmp, src_tmp, 0) 104 vst1.32 {d0[0]},[r1] @vst1_lane_u32((uint32_t *)pu1_dst_tmp, src_tmp, 0) 105 vld1.32 {d0[0]},[r5],r2 @vld1_lane_u32((uint32_t *)pu1_src_tmp, src_tmp, 0) 107 vst1.32 {d0[0]},[r6],r3 @vst1_lane_u32((uint32_t *)pu1_dst_tmp, src_tmp, 0) 108 vld1.32 {d0[0]},[r5],r2 @vld1_lane_u32((uint32_t *)pu1_src_tmp, src_tmp, 0) 110 vst1.32 {d0[0]},[r6],r3 @vst1_lane_u32((uint32_t *)pu1_dst_tmp, src_tmp, 0) 111 vld1.32 {d0[0]},[r5],r2 @vld1_lane_u32((uint32_t *)pu1_src_tmp, src_tmp, 0) 113 vst1.32 {d0[0]},[r6],r3 @vst1_lane_u32((uint32_t *)pu1_dst_tmp, src_tmp, 0)
|
D | ihevc_intra_pred_filters_neon_intr.c | 1303 …uint8x8_t dup_sub, src_tmp, src_tmp_1, round_val, dup_1, dup_2, dup_3, dup_4, dup_5, dup_6, dup_7,… in ihevc_intra_pred_luma_horz_neonintr() local 1315 src_tmp = vld1_u8(pu1_ref_tmp_1); in ihevc_intra_pred_luma_horz_neonintr() 1318 sub_res = vsubl_u8(src_tmp, dup_sub); in ihevc_intra_pred_luma_horz_neonintr()
|
/external/libvpx/libvpx/third_party/libyuv/source/ |
D | scale_neon.cc | 546 const uint8* src_tmp = NULL; in ScaleAddRows_NEON() local 567 : "+r"(src_tmp), // %0 in ScaleAddRows_NEON() 591 const uint8* src_tmp = src_ptr; in ScaleFilterCols_NEON() local 641 "+r"(src_tmp) // %6 in ScaleFilterCols_NEON() 931 const uint8* src_tmp = src_argb; in ScaleARGBCols_NEON() local 954 "+r"(src_tmp) // %6 in ScaleARGBCols_NEON() 975 const uint8* src_tmp = src_argb; in ScaleARGBFilterCols_NEON() local 1023 "+r"(src_tmp) // %6 in ScaleARGBFilterCols_NEON()
|
D | scale_neon64.cc | 550 const uint8* src_tmp = NULL; in ScaleAddRows_NEON() local 570 : "+r"(src_tmp), // %0 in ScaleAddRows_NEON() 594 const uint8* src_tmp = src_ptr; in ScaleFilterCols_NEON() local 646 "+r"(src_tmp) // %6 in ScaleFilterCols_NEON() 930 const uint8* src_tmp = src_argb; in ScaleARGBCols_NEON() local 956 "+r"(src_tmp) // %6 in ScaleARGBCols_NEON() 977 const uint8* src_tmp = src_argb; in ScaleARGBFilterCols_NEON() local 1028 "+r"(src_tmp) // %6 in ScaleARGBFilterCols_NEON()
|
/external/libavc/common/arm/ |
D | ih264_inter_pred_luma_copy_a9q.s | 96 vld1.32 {d0[0]}, [r0] @vld1_lane_u32((uint32_t *)pu1_src_tmp, src_tmp, 0) 99 vst1.32 {d0[0]}, [r1] @vst1_lane_u32((uint32_t *)pu1_dst_tmp, src_tmp, 0) 100 vld1.32 {d0[0]}, [r5], r2 @vld1_lane_u32((uint32_t *)pu1_src_tmp, src_tmp, 0) 102 vst1.32 {d0[0]}, [r6], r3 @vst1_lane_u32((uint32_t *)pu1_dst_tmp, src_tmp, 0) 103 vld1.32 {d0[0]}, [r5], r2 @vld1_lane_u32((uint32_t *)pu1_src_tmp, src_tmp, 0) 105 vst1.32 {d0[0]}, [r6], r3 @vst1_lane_u32((uint32_t *)pu1_dst_tmp, src_tmp, 0) 106 vld1.32 {d0[0]}, [r5], r2 @vld1_lane_u32((uint32_t *)pu1_src_tmp, src_tmp, 0) 108 vst1.32 {d0[0]}, [r6], r3 @vst1_lane_u32((uint32_t *)pu1_dst_tmp, src_tmp, 0)
|
/external/libvpx/libvpx/vpx_dsp/mips/ |
D | vpx_convolve_copy_msa.c | 99 const uint8_t *src_tmp; in copy_16multx8mult_msa() local 104 src_tmp = src; in copy_16multx8mult_msa() 108 LD_UB8(src_tmp, src_stride, in copy_16multx8mult_msa() 110 src_tmp += (8 * src_stride); in copy_16multx8mult_msa()
|
D | vpx_convolve8_vert_msa.c | 202 const uint8_t *src_tmp; in common_vt_8t_16w_mult_msa() local 219 src_tmp = src; in common_vt_8t_16w_mult_msa() 222 LD_SB7(src_tmp, src_stride, src0, src1, src2, src3, src4, src5, src6); in common_vt_8t_16w_mult_msa() 224 src_tmp += (7 * src_stride); in common_vt_8t_16w_mult_msa() 233 LD_SB4(src_tmp, src_stride, src7, src8, src9, src10); in common_vt_8t_16w_mult_msa() 235 src_tmp += (4 * src_stride); in common_vt_8t_16w_mult_msa()
|
D | vpx_convolve8_avg_vert_msa.c | 140 const uint8_t *src_tmp; in common_vt_8t_and_aver_dst_16w_mult_msa() local 157 src_tmp = src; in common_vt_8t_and_aver_dst_16w_mult_msa() 160 LD_SB7(src_tmp, src_stride, src0, src1, src2, src3, src4, src5, src6); in common_vt_8t_and_aver_dst_16w_mult_msa() 162 src_tmp += (7 * src_stride); in common_vt_8t_and_aver_dst_16w_mult_msa() 172 LD_SB4(src_tmp, src_stride, src7, src8, src9, src10); in common_vt_8t_and_aver_dst_16w_mult_msa() 173 src_tmp += (4 * src_stride); in common_vt_8t_and_aver_dst_16w_mult_msa()
|
/external/libvpx/libvpx/vp8/common/arm/neon/ |
D | sixtappredict_neon.c | 1284 unsigned char *src, *src_tmp, *dst, *tmpp; in vp8_sixtap_predict16x16_neon() local 1315 src_tmp = src_ptr - src_pixels_per_line * 2; in vp8_sixtap_predict16x16_neon() 1317 src = src_tmp + i * 8; in vp8_sixtap_predict16x16_neon()
|