Searched refs:v_sum_d (Results 1 – 6 of 6) sorted by relevance
/external/libaom/libaom/aom_dsp/x86/ |
D | obmc_variance_avx2.c | 33 __m128i v_sum_d = _mm_setzero_si128(); in obmc_variance_w8n() local 66 v_sum_d = _mm_add_epi32(v_sum_d, v_rdiff_d); in obmc_variance_w8n() 67 v_sum_d = _mm_add_epi32(v_sum_d, v_rdiff1_d); in obmc_variance_w8n() 77 v_d = _mm_hadd_epi32(v_sum_d, v_sse_d); in obmc_variance_w8n() 92 __m256i v_sum_d = _mm256_setzero_si256(); in obmc_variance_w16n() local 134 v_sum_d = _mm256_add_epi32(v_sum_d, v_rdiff1_d); in obmc_variance_w16n() 145 v_d = _mm256_hadd_epi32(v_sum_d, v_sse_d); in obmc_variance_w16n()
|
D | obmc_variance_sse4.c | 45 __m128i v_sum_d = _mm_setzero_si128(); in obmc_variance_w8n() local 77 v_sum_d = _mm_add_epi32(v_sum_d, v_rdiff0_d); in obmc_variance_w8n() 78 v_sum_d = _mm_add_epi32(v_sum_d, v_rdiff1_d); in obmc_variance_w8n() 86 *sum = xx_hsum_epi32_si32(v_sum_d); in obmc_variance_w8n() 176 __m128i v_sum_d = _mm_setzero_si128(); in hbd_obmc_variance_w4() local 197 v_sum_d = _mm_add_epi32(v_sum_d, v_rdiff_d); in hbd_obmc_variance_w4() 205 *sum = xx_hsum_epi32_si32(v_sum_d); in hbd_obmc_variance_w4() 216 __m128i v_sum_d = _mm_setzero_si128(); in hbd_obmc_variance_w8n() local 248 v_sum_d = _mm_add_epi32(v_sum_d, v_rdiff0_d); in hbd_obmc_variance_w8n() 249 v_sum_d = _mm_add_epi32(v_sum_d, v_rdiff1_d); in hbd_obmc_variance_w8n() [all …]
|
D | highbd_variance_avx2.c | 24 __m256i v_sum_d = _mm256_setzero_si256(); in aom_highbd_calc8x8var_avx2() local 37 v_sum_d = _mm256_add_epi16(v_sum_d, v_diff); in aom_highbd_calc8x8var_avx2() 42 __m256i v_sum00 = _mm256_cvtepi16_epi32(_mm256_castsi256_si128(v_sum_d)); in aom_highbd_calc8x8var_avx2() 43 __m256i v_sum01 = _mm256_cvtepi16_epi32(_mm256_extracti128_si256(v_sum_d, 1)); in aom_highbd_calc8x8var_avx2() 59 __m256i v_sum_d = _mm256_setzero_si256(); in aom_highbd_calc16x16var_avx2() local 67 v_sum_d = _mm256_add_epi16(v_sum_d, v_diff); in aom_highbd_calc16x16var_avx2() 72 __m256i v_sum0 = _mm256_madd_epi16(v_sum_d, one); in aom_highbd_calc16x16var_avx2()
|
D | sum_squares_sse2.c | 50 __m128i v_sum_d = in aom_sum_squares_2d_i16_4x4_sse2() local 52 v_sum_d = _mm_add_epi32(v_sum_d, _mm_srli_si128(v_sum_d, 8)); in aom_sum_squares_2d_i16_4x4_sse2() 53 return (uint64_t)_mm_cvtsi128_si32(v_sum_d); in aom_sum_squares_2d_i16_4x4_sse2() 180 const __m128i v_sum_d = _mm_add_epi32(v_sum_0123_d, v_sum_4567_d); in aom_sum_squares_i16_64n_sse2() local 182 v_acc0_q = _mm_add_epi64(v_acc0_q, _mm_and_si128(v_sum_d, v_zext_mask_q)); in aom_sum_squares_i16_64n_sse2() 183 v_acc1_q = _mm_add_epi64(v_acc1_q, _mm_srli_epi64(v_sum_d, 32)); in aom_sum_squares_i16_64n_sse2()
|
D | obmc_intrinsic_sse4.h | 25 __m128i v_sum_d = _mm_setzero_si128(); in obmc_variance_w4() local 46 v_sum_d = _mm_add_epi32(v_sum_d, v_rdiff_d); in obmc_variance_w4() 54 *sum = xx_hsum_epi32_si32(v_sum_d); in obmc_variance_w4()
|
D | blend_sse4.h | 146 const __m128i v_sum_d = _mm_madd_epi16(v_s01_w, v_m01_w); in blend_4_b12() local 150 _mm_srli_epi32(v_sum_d, AOM_BLEND_A64_ROUND_BITS - 1); in blend_4_b12()
|