/external/libvpx/libvpx/vpx_dsp/mips/ |
D | vpx_convolve_avg_msa.c | 22 LD_UB4(src, src_stride, src0, src1, src2, src3); in avg_width4_msa() 25 LD_UB4(dst, dst_stride, dst0, dst1, dst2, dst3); in avg_width4_msa() 64 LD_UB4(src, src_stride, src0, src1, src2, src3); in avg_width8_msa() 66 LD_UB4(dst, dst_stride, dst0, dst1, dst2, dst3); in avg_width8_msa() 110 LD_UB4(src, src_stride, src0, src2, src4, src6); in avg_width32_msa() 111 LD_UB4(src + 16, src_stride, src1, src3, src5, src7); in avg_width32_msa() 113 LD_UB4(dst_dup, dst_stride, dst0, dst2, dst4, dst6); in avg_width32_msa() 114 LD_UB4(dst_dup + 16, dst_stride, dst1, dst3, dst5, dst7); in avg_width32_msa() 116 LD_UB4(src, src_stride, src8, src10, src12, src14); in avg_width32_msa() 117 LD_UB4(src + 16, src_stride, src9, src11, src13, src15); in avg_width32_msa() [all …]
|
D | vpx_convolve_copy_msa.c | 39 LD_UB4(src, src_stride, src0, src1, src2, src3); in copy_width8_msa() 70 LD_UB4(src, src_stride, src0, src1, src2, src3); in copy_width8_msa() 134 LD_UB4(src, src_stride, src0, src1, src2, src3); in copy_width16_msa() 143 LD_UB4(src, src_stride, src0, src1, src2, src3); in copy_width16_msa() 159 LD_UB4(src, src_stride, src0, src1, src2, src3); in copy_width32_msa() 160 LD_UB4(src + 16, src_stride, src4, src5, src6, src7); in copy_width32_msa() 166 LD_UB4(src, src_stride, src0, src1, src2, src3); in copy_width32_msa() 167 LD_UB4(src + 16, src_stride, src4, src5, src6, src7); in copy_width32_msa() 173 LD_UB4(src, src_stride, src0, src1, src2, src3); in copy_width32_msa() 174 LD_UB4(src + 16, src_stride, src4, src5, src6, src7); in copy_width32_msa() [all …]
|
D | sad_msa.c | 56 LD_UB4(src, src_stride, src0, src1, src2, src3); in sad_8width_msa() 58 LD_UB4(ref, ref_stride, ref0, ref1, ref2, ref3); in sad_8width_msa() 140 LD_UB4(src, 16, src0, src1, src2, src3); in sad_64width_msa() 142 LD_UB4(ref, 16, ref0, ref1, ref2, ref3); in sad_64width_msa() 147 LD_UB4(src, 16, src0, src1, src2, src3); in sad_64width_msa() 149 LD_UB4(ref, 16, ref0, ref1, ref2, ref3); in sad_64width_msa() 178 LD_UB4(ref_ptr, ref_stride, ref0, ref1, ref2, ref3); in sad_4width_x3_msa() 213 LD_UB4(src, src_stride, src0, src1, src2, src3); in sad_8width_x3_msa() 215 LD_UB4(ref, ref_stride, ref00, ref11, ref22, ref33); in sad_8width_x3_msa() 342 LD_UB4(src, 16, src0, src1, src2, src3); in sad_64width_x3_msa() [all …]
|
D | variance_msa.c | 78 LD_UB4(src_ptr, src_stride, src0, src1, src2, src3); in sse_diff_8width_msa() 80 LD_UB4(ref_ptr, ref_stride, ref0, ref1, ref2, ref3); in sse_diff_8width_msa() 236 LD_UB4(src_ptr, 16, src0, src1, src2, src3); in sse_diff_64x32_msa() 238 LD_UB4(ref_ptr, 16, ref0, ref1, ref2, ref3); in sse_diff_64x32_msa() 245 LD_UB4(src_ptr, 16, src0, src1, src2, src3); in sse_diff_64x32_msa() 247 LD_UB4(ref_ptr, 16, ref0, ref1, ref2, ref3); in sse_diff_64x32_msa() 275 LD_UB4(src_ptr, 16, src0, src1, src2, src3); in sse_diff_64x64_msa() 277 LD_UB4(ref_ptr, 16, ref0, ref1, ref2, ref3); in sse_diff_64x64_msa() 284 LD_UB4(src_ptr, 16, src0, src1, src2, src3); in sse_diff_64x64_msa() 286 LD_UB4(ref_ptr, 16, ref0, ref1, ref2, ref3); in sse_diff_64x64_msa() [all …]
|
D | sub_pixel_variance_msa.c | 97 LD_UB4(src_ptr, src_stride, src0, src1, src2, src3); in avg_sse_diff_8width_msa() 99 LD_UB4(ref_ptr, ref_stride, ref0, ref1, ref2, ref3); in avg_sse_diff_8width_msa() 307 LD_UB4(sec_pred, 16, pred0, pred1, pred2, pred3); in avg_sse_diff_64x32_msa() 309 LD_UB4(src_ptr, 16, src0, src1, src2, src3); in avg_sse_diff_64x32_msa() 311 LD_UB4(ref_ptr, 16, ref0, ref1, ref2, ref3); in avg_sse_diff_64x32_msa() 320 LD_UB4(sec_pred, 16, pred0, pred1, pred2, pred3); in avg_sse_diff_64x32_msa() 322 LD_UB4(src_ptr, 16, src0, src1, src2, src3); in avg_sse_diff_64x32_msa() 324 LD_UB4(ref_ptr, 16, ref0, ref1, ref2, ref3); in avg_sse_diff_64x32_msa() 359 LD_UB4(sec_pred, 16, pred0, pred1, pred2, pred3); in avg_sse_diff_64x64_msa() 361 LD_UB4(src_ptr, 16, src0, src1, src2, src3); in avg_sse_diff_64x64_msa() [all …]
|
D | vpx_convolve8_avg_vert_msa.c | 48 LD_UB4(dst, dst_stride, dst0, dst1, dst2, dst3); in common_vt_8t_and_aver_dst_4w_msa() 105 LD_UB4(dst, dst_stride, dst0, dst1, dst2, dst3); in common_vt_8t_and_aver_dst_8w_msa() 175 LD_UB4(dst_tmp, dst_stride, dst0, dst1, dst2, dst3); in common_vt_8t_and_aver_dst_16w_mult_msa() 279 LD_UB4(dst, dst_stride, dst0, dst1, dst2, dst3); in common_vt_2t_and_aver_dst_4x4_msa() 361 LD_UB4(dst, dst_stride, dst0, dst1, dst2, dst3); in common_vt_2t_and_aver_dst_8x4_msa() 451 LD_UB4(src, src_stride, src1, src2, src3, src4); in common_vt_2t_and_aver_dst_16w_msa() 454 LD_UB4(dst, dst_stride, dst0, dst1, dst2, dst3); in common_vt_2t_and_aver_dst_16w_msa() 503 LD_UB4(src, src_stride, src1, src2, src3, src4); in common_vt_2t_and_aver_dst_32w_msa() 504 LD_UB4(dst, dst_stride, dst0, dst1, dst2, dst3); in common_vt_2t_and_aver_dst_32w_msa() 508 LD_UB4(src + 16, src_stride, src6, src7, src8, src9); in common_vt_2t_and_aver_dst_32w_msa() [all …]
|
D | vpx_convolve8_avg_horiz_msa.c | 40 LD_UB4(dst, dst_stride, dst0, dst1, dst2, dst3); in common_hz_8t_and_aver_dst_4x4_msa() 136 LD_UB4(dst, dst_stride, dst0, dst1, dst2, dst3); in common_hz_8t_and_aver_dst_8w_msa() 336 LD_UB4(dst, dst_stride, dst0, dst1, dst2, dst3); in common_hz_2t_and_aver_dst_4x4_msa() 414 LD_UB4(dst, dst_stride, dst0, dst1, dst2, dst3); in common_hz_2t_and_aver_dst_8x4_msa() 442 LD_UB4(dst, dst_stride, dst0, dst1, dst2, dst3); in common_hz_2t_and_aver_dst_8x8mult_msa() 454 LD_UB4(dst, dst_stride, dst0, dst1, dst2, dst3); in common_hz_2t_and_aver_dst_8x8mult_msa() 468 LD_UB4(dst, dst_stride, dst0, dst1, dst2, dst3); in common_hz_2t_and_aver_dst_8x8mult_msa() 479 LD_UB4(dst, dst_stride, dst0, dst1, dst2, dst3); in common_hz_2t_and_aver_dst_8x8mult_msa() 531 LD_UB4(dst, dst_stride, dst0, dst1, dst2, dst3); in common_hz_2t_and_aver_dst_16w_msa() 556 LD_UB4(dst, dst_stride, dst0, dst1, dst2, dst3); in common_hz_2t_and_aver_dst_16w_msa() [all …]
|
D | vpx_convolve8_avg_msa.c | 66 LD_UB4(dst, dst_stride, dst0, dst1, dst2, dst3); in common_hv_8ht_8vt_and_aver_dst_4w_msa() 155 LD_UB4(dst, dst_stride, dst0, dst1, dst2, dst3); in common_hv_8ht_8vt_and_aver_dst_8w_msa() 274 LD_UB4(dst, dst_stride, dst0, dst1, dst2, dst3); in common_hv_2ht_2vt_and_aver_dst_4x4_msa() 374 LD_UB4(dst, dst_stride, dst0, dst1, dst2, dst3); in common_hv_2ht_2vt_and_aver_dst_8x4_msa() 447 LD_UB4(dst, dst_stride, dst0, dst1, dst2, dst3); in common_hv_2ht_2vt_and_aver_dst_8x8mult_msa() 503 LD_UB4(dst, dst_stride, dst0, dst1, dst2, dst3); in common_hv_2ht_2vt_and_aver_dst_16w_msa()
|
D | vpx_convolve8_vert_msa.c | 459 LD_UB4(src, src_stride, src1, src2, src3, src4); in common_vt_2t_16w_msa() 508 LD_UB4(src, src_stride, src1, src2, src3, src4); in common_vt_2t_32w_msa() 512 LD_UB4(src + 16, src_stride, src6, src7, src8, src9); in common_vt_2t_32w_msa() 571 LD_UB4(src, 16, src0, src3, src6, src9); in common_vt_2t_64w_msa()
|
D | loopfilter_16_msa.c | 100 LD_UB4(filter48, 16, p2, p1, p0, q0); in vpx_hz_lpf_t16_16w() 489 LD_UB4((src - 8 * pitch), pitch, p7, p6, p5, p4); in vpx_lpf_horizontal_16_msa() 490 LD_UB4(src + (4 * pitch), pitch, q4, q5, q6, q7); in vpx_lpf_horizontal_16_msa() 835 LD_UB4(filter48, 16, p2, p1, p0, q0); in vpx_vt_lpf_t16_8w() 1159 LD_UB4(filter48, 16, p2, p1, p0, q0); in vpx_vt_lpf_t16_16w()
|
D | inv_txfm_msa.h | 102 LD_UB4(dst_m, dst_stride, dst0_m, dst1_m, dst2_m, dst3_m); \
|
D | idct16x16_msa.c | 277 LD_UB4(dst, dst_stride, dst0, dst1, dst2, dst3); in vpx_idct16x16_1_add_msa()
|
D | macros_msa.h | 313 #define LD_UB4(...) LD_B4(v16u8, __VA_ARGS__) macro
|
/external/libvpx/libvpx/vp8/common/mips/msa/ |
D | idct_msa.c | 252 LD_UB4(dest, dest_stride, dest0, dest1, dest2, dest3); in dequant_idct4x4_addblk_2x_msa() 299 LD_UB4(dest, dest_stride, dest0, dest1, dest2, dest3); in dequant_idct_addconst_2x_msa()
|
D | loopfilter_filters_msa.c | 574 LD_UB4(src - (pitch << 1), pitch, p1, p0, q0, q1); in vp8_loop_filter_simple_horizontal_edge_msa()
|
D | bilinear_filter_msa.c | 438 LD_UB4(src, src_stride, src1, src2, src3, src4); in common_vt_2t_16w_msa()
|
D | vp8_macros_msa.h | 307 #define LD_UB4(...) LD_B4(v16u8, __VA_ARGS__) macro
|