Searched refs:_mm256_avg_epu8 (Results 1 – 9 of 9) sorted by relevance
/external/libaom/libaom/aom_dsp/x86/ |
D | variance_impl_avx2.c | 64 src_reg = _mm256_avg_epu8(src_reg, src_next_reg); 185 src_avg = _mm256_avg_epu8(src_avg, src_reg); in aom_sub_pixel_variance32xh_avx2() 250 src_pack = _mm256_avg_epu8(src_pack, src_reg); in aom_sub_pixel_variance32xh_avx2() 314 src_reg = _mm256_avg_epu8(src_reg, sec_reg); in aom_sub_pixel_avg_variance32xh_avx2() 328 src_reg = _mm256_avg_epu8(src_reg, sec_reg); in aom_sub_pixel_avg_variance32xh_avx2() 350 src_reg = _mm256_avg_epu8(src_reg, sec_reg); in aom_sub_pixel_avg_variance32xh_avx2() 366 src_reg = _mm256_avg_epu8(src_reg, sec_reg); in aom_sub_pixel_avg_variance32xh_avx2() 388 src_avg = _mm256_avg_epu8(src_avg, src_reg); in aom_sub_pixel_avg_variance32xh_avx2() 390 src_avg = _mm256_avg_epu8(src_avg, sec_reg); in aom_sub_pixel_avg_variance32xh_avx2() 418 src_avg = _mm256_avg_epu8(src_avg, sec_reg); in aom_sub_pixel_avg_variance32xh_avx2() [all …]
|
D | sad_avx2.c | 110 ref1_reg = _mm256_avg_epu8( \ 112 ref2_reg = _mm256_avg_epu8( \ 148 ref1_reg = _mm256_avg_epu8( \ 150 ref2_reg = _mm256_avg_epu8( \
|
D | sad_impl_avx2.c | 174 ref1_reg = _mm256_avg_epu8( in sad_w64_avg_avx2() 176 ref2_reg = _mm256_avg_epu8( in sad_w64_avg_avx2()
|
D | blend_a64_mask_avx2.c | 269 _mm256_avg_epu8(_mm256_adds_epu8(m_i00, m_i10), zeros); in lowbd_blend_a64_d16_mask_subw0_subh1_w32_avx2() 589 _mm256_avg_epu8(v_rl_b, _mm256_srli_si256(v_rl_b, 1)); in blend_a64_mask_sx_w16_avx2() 620 _mm256_avg_epu8(v_r0_s_b, _mm256_srli_si256(v_r0_s_b, 8)); in blend_a64_mask_sx_w32n_avx2() 622 _mm256_avg_epu8(v_r1_s_b, _mm256_srli_si256(v_r1_s_b, 8)); in blend_a64_mask_sx_w32n_avx2() 729 const __m256i v_m0_b = _mm256_avg_epu8(v_ra_b, v_rb_b); in blend_a64_mask_sy_w32n_avx2()
|
/external/libvpx/libvpx/vpx_dsp/x86/ |
D | variance_avx2.c | 238 const __m256i avg_reg = _mm256_avg_epu8(src_reg, sec_reg); in spv32_x0_y0() 266 const __m256i src_avg = _mm256_avg_epu8(src_0, src_1); in spv32_half_zero() 269 const __m256i avg_reg = _mm256_avg_epu8(src_avg, sec_reg); in spv32_half_zero() 309 __m256i prev_src_avg = _mm256_avg_epu8(src_a, src_b); in spv32_x4_y4() 317 const __m256i src_avg = _mm256_avg_epu8(src_0, src_1); in spv32_x4_y4() 318 const __m256i current_avg = _mm256_avg_epu8(prev_src_avg, src_avg); in spv32_x4_y4() 323 const __m256i avg_reg = _mm256_avg_epu8(current_avg, sec_reg); in spv32_x4_y4() 362 const __m256i avg_reg = _mm256_avg_epu8(exp_src, sec_reg); in spv32_bilin_zero() 402 __m256i prev_src_avg = _mm256_avg_epu8(src_a, src_b); in spv32_x4_yb() 410 const __m256i src_avg = _mm256_avg_epu8(src_0, src_1); in spv32_x4_yb() [all …]
|
D | sad_avx2.c | 103 ref1_reg = _mm256_avg_epu8( \ 105 ref2_reg = _mm256_avg_epu8( \ 140 ref1_reg = _mm256_avg_epu8( \ 142 ref2_reg = _mm256_avg_epu8( \
|
/external/libaom/libaom/aom_dsp/simd/ |
D | v256_intrinsics_x86.h | 526 SIMD_INLINE v256 v256_avg_u8(v256 a, v256 b) { return _mm256_avg_epu8(a, b); } in v256_avg_u8() 530 _mm256_avg_epu8(a, b), in v256_rdavg_u8()
|
/external/clang/test/CodeGen/ |
D | avx2-builtins.c | 105 return _mm256_avg_epu8(a, b); in test_mm256_avg_epu8()
|
/external/clang/lib/Headers/ |
D | avx2intrin.h | 146 _mm256_avg_epu8(__m256i __a, __m256i __b) in _mm256_avg_epu8() function
|