/external/libvpx/libvpx/vpx_dsp/mips/ |
D | vpx_convolve_avg_msa.c | 27 AVER_UB4_UB(src0, dst0, src1, dst1, src2, dst2, src3, dst3, dst0, dst1, in avg_width4_msa() 68 AVER_UB4_UB(src0, dst0, src1, dst1, src2, dst2, src3, dst3, dst0, dst1, in avg_width8_msa() 91 AVER_UB4_UB(src0, dst0, src1, dst1, src2, dst2, src3, dst3, dst0, dst1, in avg_width16_msa() 93 AVER_UB4_UB(src4, dst4, src5, dst5, src6, dst6, src7, dst7, dst4, dst5, in avg_width16_msa() 123 AVER_UB4_UB(src0, dst0, src1, dst1, src2, dst2, src3, dst3, dst0, dst1, in avg_width32_msa() 125 AVER_UB4_UB(src4, dst4, src5, dst5, src6, dst6, src7, dst7, dst4, dst5, in avg_width32_msa() 127 AVER_UB4_UB(src8, dst8, src9, dst9, src10, dst10, src11, dst11, dst8, dst9, in avg_width32_msa() 129 AVER_UB4_UB(src12, dst12, src13, dst13, src14, dst14, src15, dst15, dst12, in avg_width32_msa() 169 AVER_UB4_UB(src0, dst0, src1, dst1, src2, dst2, src3, dst3, dst0, dst1, in avg_width64_msa() 171 AVER_UB4_UB(src4, dst4, src5, dst5, src6, dst6, src7, dst7, dst4, dst5, in avg_width64_msa() [all …]
|
D | sub_pixel_variance_msa.c | 301 AVER_UB4_UB(src0, pred0, src1, pred1, src2, pred2, src3, pred3, src0, src1, in avg_sse_diff_64x32_msa() 314 AVER_UB4_UB(src0, pred0, src1, pred1, src2, pred2, src3, pred3, src0, src1, in avg_sse_diff_64x32_msa() 352 AVER_UB4_UB(src0, pred0, src1, pred1, src2, pred2, src3, pred3, src0, src1, in avg_sse_diff_64x64_msa() 365 AVER_UB4_UB(src0, pred0, src1, pred1, src2, pred2, src3, pred3, src0, src1, in avg_sse_diff_64x64_msa() 1082 AVER_UB4_UB(tmp0, pred0, tmp1, pred1, tmp2, pred2, tmp3, pred3, tmp0, tmp1, in subpel_avg_ssediff_16w_h_msa() 1293 AVER_UB4_UB(out0, pred0, out1, pred1, out2, pred2, out3, pred3, out0, out1, in subpel_avg_ssediff_16w_v_msa() 1539 AVER_UB4_UB(out0, pred0, out1, pred1, out2, pred2, out3, pred3, out0, out1, in subpel_avg_ssediff_16w_hv_msa()
|
D | sad_msa.c | 1225 AVER_UB4_UB(pred0, ref0, pred1, ref1, pred2, ref2, pred3, ref3, comp0, in avgsad_64width_msa() 1236 AVER_UB4_UB(pred0, ref0, pred1, ref1, pred2, ref2, pred3, ref3, comp0, in avgsad_64width_msa() 1247 AVER_UB4_UB(pred0, ref0, pred1, ref1, pred2, ref2, pred3, ref3, comp0, in avgsad_64width_msa() 1258 AVER_UB4_UB(pred0, ref0, pred1, ref1, pred2, ref2, pred3, ref3, comp0, in avgsad_64width_msa()
|
D | vpx_convolve8_avg_msa.c | 302 AVER_UB4_UB(res0, dst0, res1, dst2, res2, dst4, res3, dst6, res0, res1, res2, in common_hv_2ht_2vt_and_aver_dst_4x8_msa()
|
D | vpx_convolve8_avg_vert_msa.c | 196 AVER_UB4_UB(tmp0, dst0, tmp1, dst1, tmp2, dst2, tmp3, dst3, dst0, dst1, in common_vt_8t_and_aver_dst_16w_mult_msa()
|
D | vpx_convolve8_avg_horiz_msa.c | 359 AVER_UB4_UB(res0, dst0, res1, dst2, res2, dst4, res3, dst6, res0, res1, res2, in common_hz_2t_and_aver_dst_4x8_msa()
|
D | macros_msa.h | 620 #define AVER_UB4_UB(...) AVER_UB4(v16u8, __VA_ARGS__) macro
|