Searched refs:k__DCT_CONST_ROUNDING (Results 1 – 6 of 6) sorted by relevance
/external/libvpx/libvpx/vp9/encoder/x86/ |
D | vp9_dct_sse2.c | 78 const __m128i k__DCT_CONST_ROUNDING = _mm_set1_epi32(DCT_CONST_ROUNDING); in fdct4_sse2() local 92 v[0] = _mm_add_epi32(u[0], k__DCT_CONST_ROUNDING); in fdct4_sse2() 93 v[1] = _mm_add_epi32(u[1], k__DCT_CONST_ROUNDING); in fdct4_sse2() 94 v[2] = _mm_add_epi32(u[2], k__DCT_CONST_ROUNDING); in fdct4_sse2() 95 v[3] = _mm_add_epi32(u[3], k__DCT_CONST_ROUNDING); in fdct4_sse2() 113 const __m128i k__DCT_CONST_ROUNDING = _mm_set1_epi32(DCT_CONST_ROUNDING); in fadst4_sse2() local 139 v[0] = _mm_add_epi32(u[0], k__DCT_CONST_ROUNDING); in fadst4_sse2() 140 v[1] = _mm_add_epi32(u[1], k__DCT_CONST_ROUNDING); in fadst4_sse2() 141 v[2] = _mm_add_epi32(u[2], k__DCT_CONST_ROUNDING); in fadst4_sse2() 142 v[3] = _mm_add_epi32(u[6], k__DCT_CONST_ROUNDING); in fadst4_sse2() [all …]
|
D | vp9_dct_ssse3.c | 48 const __m128i k__DCT_CONST_ROUNDING = _mm_set1_epi32(DCT_CONST_ROUNDING); in vp9_fdct8x8_quant_ssse3() local 125 const __m128i v0 = _mm_add_epi32(u0, k__DCT_CONST_ROUNDING); in vp9_fdct8x8_quant_ssse3() 126 const __m128i v1 = _mm_add_epi32(u1, k__DCT_CONST_ROUNDING); in vp9_fdct8x8_quant_ssse3() 127 const __m128i v2 = _mm_add_epi32(u2, k__DCT_CONST_ROUNDING); in vp9_fdct8x8_quant_ssse3() 128 const __m128i v3 = _mm_add_epi32(u3, k__DCT_CONST_ROUNDING); in vp9_fdct8x8_quant_ssse3() 130 const __m128i v4 = _mm_add_epi32(u4, k__DCT_CONST_ROUNDING); in vp9_fdct8x8_quant_ssse3() 131 const __m128i v5 = _mm_add_epi32(u5, k__DCT_CONST_ROUNDING); in vp9_fdct8x8_quant_ssse3() 132 const __m128i v6 = _mm_add_epi32(u6, k__DCT_CONST_ROUNDING); in vp9_fdct8x8_quant_ssse3() 133 const __m128i v7 = _mm_add_epi32(u7, k__DCT_CONST_ROUNDING); in vp9_fdct8x8_quant_ssse3() 178 const __m128i v0 = _mm_add_epi32(u0, k__DCT_CONST_ROUNDING); in vp9_fdct8x8_quant_ssse3() [all …]
|
/external/libvpx/libvpx/vpx_dsp/x86/ |
D | fwd_dct32x32_impl_avx2.h | 89 const __m256i k__DCT_CONST_ROUNDING = _mm256_set1_epi32(DCT_CONST_ROUNDING); in FDCT32x32_2D_AVX2() local 356 const __m256i s2_20_4 = _mm256_add_epi32(s2_20_2, k__DCT_CONST_ROUNDING); in FDCT32x32_2D_AVX2() 357 const __m256i s2_20_5 = _mm256_add_epi32(s2_20_3, k__DCT_CONST_ROUNDING); in FDCT32x32_2D_AVX2() 358 const __m256i s2_21_4 = _mm256_add_epi32(s2_21_2, k__DCT_CONST_ROUNDING); in FDCT32x32_2D_AVX2() 359 const __m256i s2_21_5 = _mm256_add_epi32(s2_21_3, k__DCT_CONST_ROUNDING); in FDCT32x32_2D_AVX2() 360 const __m256i s2_22_4 = _mm256_add_epi32(s2_22_2, k__DCT_CONST_ROUNDING); in FDCT32x32_2D_AVX2() 361 const __m256i s2_22_5 = _mm256_add_epi32(s2_22_3, k__DCT_CONST_ROUNDING); in FDCT32x32_2D_AVX2() 362 const __m256i s2_23_4 = _mm256_add_epi32(s2_23_2, k__DCT_CONST_ROUNDING); in FDCT32x32_2D_AVX2() 363 const __m256i s2_23_5 = _mm256_add_epi32(s2_23_3, k__DCT_CONST_ROUNDING); in FDCT32x32_2D_AVX2() 364 const __m256i s2_24_4 = _mm256_add_epi32(s2_24_2, k__DCT_CONST_ROUNDING); in FDCT32x32_2D_AVX2() [all …]
|
D | fwd_dct32x32_impl_sse2.h | 106 const __m128i k__DCT_CONST_ROUNDING = _mm_set1_epi32(DCT_CONST_ROUNDING); in FDCT32x32_2D() local 426 const __m128i s2_20_4 = _mm_add_epi32(s2_20_2, k__DCT_CONST_ROUNDING); in FDCT32x32_2D() 427 const __m128i s2_20_5 = _mm_add_epi32(s2_20_3, k__DCT_CONST_ROUNDING); in FDCT32x32_2D() 428 const __m128i s2_21_4 = _mm_add_epi32(s2_21_2, k__DCT_CONST_ROUNDING); in FDCT32x32_2D() 429 const __m128i s2_21_5 = _mm_add_epi32(s2_21_3, k__DCT_CONST_ROUNDING); in FDCT32x32_2D() 430 const __m128i s2_22_4 = _mm_add_epi32(s2_22_2, k__DCT_CONST_ROUNDING); in FDCT32x32_2D() 431 const __m128i s2_22_5 = _mm_add_epi32(s2_22_3, k__DCT_CONST_ROUNDING); in FDCT32x32_2D() 432 const __m128i s2_23_4 = _mm_add_epi32(s2_23_2, k__DCT_CONST_ROUNDING); in FDCT32x32_2D() 433 const __m128i s2_23_5 = _mm_add_epi32(s2_23_3, k__DCT_CONST_ROUNDING); in FDCT32x32_2D() 434 const __m128i s2_24_4 = _mm_add_epi32(s2_24_2, k__DCT_CONST_ROUNDING); in FDCT32x32_2D() [all …]
|
D | fwd_txfm_impl_sse2.h | 79 const __m128i k__DCT_CONST_ROUNDING = _mm_set1_epi32(DCT_CONST_ROUNDING); in FDCT4x4_2D() local 163 const __m128i v0 = _mm_add_epi32(u0, k__DCT_CONST_ROUNDING); in FDCT4x4_2D() 164 const __m128i v1 = _mm_add_epi32(u1, k__DCT_CONST_ROUNDING); in FDCT4x4_2D() 165 const __m128i v2 = _mm_add_epi32(u2, k__DCT_CONST_ROUNDING); in FDCT4x4_2D() 166 const __m128i v3 = _mm_add_epi32(u3, k__DCT_CONST_ROUNDING); in FDCT4x4_2D() 281 const __m128i k__DCT_CONST_ROUNDING = _mm_set1_epi32(DCT_CONST_ROUNDING); in FDCT8x8_2D() local 359 const __m128i v0 = _mm_add_epi32(u0, k__DCT_CONST_ROUNDING); in FDCT8x8_2D() 360 const __m128i v1 = _mm_add_epi32(u1, k__DCT_CONST_ROUNDING); in FDCT8x8_2D() 361 const __m128i v2 = _mm_add_epi32(u2, k__DCT_CONST_ROUNDING); in FDCT8x8_2D() 362 const __m128i v3 = _mm_add_epi32(u3, k__DCT_CONST_ROUNDING); in FDCT8x8_2D() [all …]
|
D | inv_txfm_sse2.c | 186 const __m128i k__DCT_CONST_ROUNDING = _mm_set1_epi32(DCT_CONST_ROUNDING); in idct4_sse2() local 198 u[0] = _mm_add_epi32(v[0], k__DCT_CONST_ROUNDING); in idct4_sse2() 199 u[1] = _mm_add_epi32(v[1], k__DCT_CONST_ROUNDING); in idct4_sse2() 200 u[2] = _mm_add_epi32(v[2], k__DCT_CONST_ROUNDING); in idct4_sse2() 201 u[3] = _mm_add_epi32(v[3], k__DCT_CONST_ROUNDING); in idct4_sse2() 224 const __m128i k__DCT_CONST_ROUNDING = _mm_set1_epi32(DCT_CONST_ROUNDING); in iadst4_sse2() local 252 v[0] = _mm_add_epi32(u[0], k__DCT_CONST_ROUNDING); in iadst4_sse2() 253 v[1] = _mm_add_epi32(u[1], k__DCT_CONST_ROUNDING); in iadst4_sse2() 254 v[2] = _mm_add_epi32(u[2], k__DCT_CONST_ROUNDING); in iadst4_sse2() 255 v[3] = _mm_add_epi32(u[6], k__DCT_CONST_ROUNDING); in iadst4_sse2() [all …]
|