Home
last modified time | relevance | path

Searched refs:_mm256_sad_epu8 (Results 1 – 11 of 11) sorted by relevance

/external/libvpx/libvpx/vpx_dsp/x86/
Dsad4d_avx2.c51 r[0] = _mm256_sad_epu8(r[0], s); in vpx_sad32x32x4d_avx2()
52 r[1] = _mm256_sad_epu8(r[1], s); in vpx_sad32x32x4d_avx2()
53 r[2] = _mm256_sad_epu8(r[2], s); in vpx_sad32x32x4d_avx2()
54 r[3] = _mm256_sad_epu8(r[3], s); in vpx_sad32x32x4d_avx2()
103 r_lo[0] = _mm256_sad_epu8(r_lo[0], s_lo); in vpx_sad64x64x4d_avx2()
104 r_lo[1] = _mm256_sad_epu8(r_lo[1], s_lo); in vpx_sad64x64x4d_avx2()
105 r_lo[2] = _mm256_sad_epu8(r_lo[2], s_lo); in vpx_sad64x64x4d_avx2()
106 r_lo[3] = _mm256_sad_epu8(r_lo[3], s_lo); in vpx_sad64x64x4d_avx2()
107 r_hi[0] = _mm256_sad_epu8(r_hi[0], s_hi); in vpx_sad64x64x4d_avx2()
108 r_hi[1] = _mm256_sad_epu8(r_hi[1], s_hi); in vpx_sad64x64x4d_avx2()
[all …]
Dsad_avx2.c25 sad1_reg = _mm256_sad_epu8( \
27 sad2_reg = _mm256_sad_epu8( \
56 sad1_reg = _mm256_sad_epu8( \
58 sad2_reg = _mm256_sad_epu8( \
107 sad1_reg = _mm256_sad_epu8( \
109 sad2_reg = _mm256_sad_epu8( \
144 sad1_reg = _mm256_sad_epu8( \
146 sad2_reg = _mm256_sad_epu8( \
/external/libaom/libaom/aom_dsp/x86/
Dsad4d_avx2.c42 ref0_reg = _mm256_sad_epu8(ref0_reg, src_reg); in aom_sad32x32x4d_avx2()
43 ref1_reg = _mm256_sad_epu8(ref1_reg, src_reg); in aom_sad32x32x4d_avx2()
44 ref2_reg = _mm256_sad_epu8(ref2_reg, src_reg); in aom_sad32x32x4d_avx2()
45 ref3_reg = _mm256_sad_epu8(ref3_reg, src_reg); in aom_sad32x32x4d_avx2()
118 ref0_reg = _mm256_sad_epu8(ref0_reg, src_reg); in aom_sad64x64x4d_avx2()
119 ref1_reg = _mm256_sad_epu8(ref1_reg, src_reg); in aom_sad64x64x4d_avx2()
120 ref2_reg = _mm256_sad_epu8(ref2_reg, src_reg); in aom_sad64x64x4d_avx2()
121 ref3_reg = _mm256_sad_epu8(ref3_reg, src_reg); in aom_sad64x64x4d_avx2()
122 ref0next_reg = _mm256_sad_epu8(ref0next_reg, srcnext_reg); in aom_sad64x64x4d_avx2()
123 ref1next_reg = _mm256_sad_epu8(ref1next_reg, srcnext_reg); in aom_sad64x64x4d_avx2()
[all …]
Dsad_avx2.c28 sad1_reg = _mm256_sad_epu8( \
30 sad2_reg = _mm256_sad_epu8( \
60 sad1_reg = _mm256_sad_epu8( \
62 sad2_reg = _mm256_sad_epu8( \
114 sad1_reg = _mm256_sad_epu8( \
116 sad2_reg = _mm256_sad_epu8( \
152 sad1_reg = _mm256_sad_epu8( \
154 sad2_reg = _mm256_sad_epu8( \
Dsad_impl_avx2.c26 s1 = _mm256_sad_epu8(r1, _mm256_loadu_si256((__m256i const *)src_ptr)); in sad32x32()
27 s2 = _mm256_sad_epu8( in sad32x32()
179 _mm256_sad_epu8(ref1_reg, _mm256_loadu_si256((__m256i const *)src_ptr)); in sad_w64_avg_avx2()
180 sad2_reg = _mm256_sad_epu8( in sad_w64_avg_avx2()
Dmasked_sad_intrin_avx2.c53 res = _mm256_add_epi32(res, _mm256_sad_epu8(pred, src)); in masked_sad32xh_avx2()
107 res = _mm256_add_epi32(res, _mm256_sad_epu8(pred, src)); in masked_sad16xh_avx2()
Dintrapred_avx2.c21 __m256i y0 = _mm256_sad_epu8(x0, zero); in dc_sum_64()
22 __m256i y1 = _mm256_sad_epu8(x1, zero); in dc_sum_64()
33 __m256i y = _mm256_sad_epu8(x, zero); in dc_sum_32()
/external/libaom/libaom/av1/encoder/x86/
Dcorner_match_avx2.c56 sum_vec = _mm256_add_epi16(sum_vec, _mm256_sad_epu8(v, zero)); in compute_cross_correlation_avx2()
/external/libaom/libaom/aom_dsp/simd/
Dv256_intrinsics_x86.h436 v256 t = _mm256_sad_epu8(a, _mm256_setzero_si256()); in v256_hadd_u8()
453 return _mm256_add_epi64(s, _mm256_sad_epu8(a, b)); in v256_sad_u8()
/external/clang/test/CodeGen/
Davx2-builtins.c925 return _mm256_sad_epu8(x, y); in test_mm256_sad_epu8()
/external/clang/lib/Headers/
Davx2intrin.h487 _mm256_sad_epu8(__m256i __a, __m256i __b) in _mm256_sad_epu8() function