Home
last modified time | relevance | path

Searched refs:_mm256_castsi256_si128 (Results 1 – 9 of 9) sorted by relevance

/external/libvpx/libvpx/vp9/common/x86/
Dvp9_subpixel_8t_intrin_avx2.c191 _mm256_castsi256_si128(srcRegFilt32b1_1)); in vp9_filter_block1d16_h8_avx2()
209 _mm256_castsi256_si128(filt1Reg)); in vp9_filter_block1d16_h8_avx2()
211 _mm256_castsi256_si128(filt4Reg)); in vp9_filter_block1d16_h8_avx2()
215 _mm256_castsi256_si128(firstFilters)); in vp9_filter_block1d16_h8_avx2()
217 _mm256_castsi256_si128(forthFilters)); in vp9_filter_block1d16_h8_avx2()
224 _mm256_castsi256_si128(filt2Reg)); in vp9_filter_block1d16_h8_avx2()
226 _mm256_castsi256_si128(filt3Reg)); in vp9_filter_block1d16_h8_avx2()
230 _mm256_castsi256_si128(secondFilters)); in vp9_filter_block1d16_h8_avx2()
232 _mm256_castsi256_si128(thirdFilters)); in vp9_filter_block1d16_h8_avx2()
248 _mm256_castsi256_si128(filt1Reg)); in vp9_filter_block1d16_h8_avx2()
[all …]
Dvp9_loopfilter_intrin_avx2.c616 flat2_p0 = _mm256_castsi256_si128( in mb_lpf_horizontal_edge_w_avx2_16()
624 flat2_q0 = _mm256_castsi256_si128( in mb_lpf_horizontal_edge_w_avx2_16()
632 flat_p0 = _mm256_castsi256_si128( in mb_lpf_horizontal_edge_w_avx2_16()
640 flat_q0 = _mm256_castsi256_si128( in mb_lpf_horizontal_edge_w_avx2_16()
660 flat2_p1 = _mm256_castsi256_si128( in mb_lpf_horizontal_edge_w_avx2_16()
668 flat2_q1 = _mm256_castsi256_si128( in mb_lpf_horizontal_edge_w_avx2_16()
680 flat_p1 = _mm256_castsi256_si128( in mb_lpf_horizontal_edge_w_avx2_16()
688 flat_q1 = _mm256_castsi256_si128( in mb_lpf_horizontal_edge_w_avx2_16()
708 flat2_p2 = _mm256_castsi256_si128( in mb_lpf_horizontal_edge_w_avx2_16()
716 flat2_q2 = _mm256_castsi256_si128( in mb_lpf_horizontal_edge_w_avx2_16()
[all …]
/external/libvpx/libvpx/vp9/encoder/x86/
Dvp9_variance_impl_intrin_avx2.c78 sum_res = _mm_add_epi16(_mm256_castsi256_si128(sum_ref_src), in vp9_get16x16var_avx2()
81 madd_res = _mm_add_epi32(_mm256_castsi256_si128(madd_ref_src), in vp9_get16x16var_avx2()
85 expand_sum_low = _mm_unpacklo_epi16(_mm256_castsi256_si128(zero_reg), in vp9_get16x16var_avx2()
87 expand_sum_high = _mm_unpackhi_epi16(_mm256_castsi256_si128(zero_reg), in vp9_get16x16var_avx2()
98 _mm256_castsi256_si128(zero_reg)); in vp9_get16x16var_avx2()
100 _mm256_castsi256_si128(zero_reg)); in vp9_get16x16var_avx2()
105 _mm256_castsi256_si128(zero_reg)); in vp9_get16x16var_avx2()
107 _mm256_castsi256_si128(zero_reg)); in vp9_get16x16var_avx2()
207 *((int*)SSE)= _mm_cvtsi128_si32(_mm256_castsi256_si128(madd_ref_src)) + in vp9_get32x32var_avx2()
210 *((int*)Sum)= _mm_cvtsi128_si32(_mm256_castsi256_si128(sum_ref_src)) + in vp9_get32x32var_avx2()
Dvp9_error_intrin_avx2.c61 sse_reg128 = _mm_add_epi64(_mm256_castsi256_si128(sse_reg), in vp9_block_error_avx2()
64 ssz_reg128 = _mm_add_epi64(_mm256_castsi256_si128(ssz_reg), in vp9_block_error_avx2()
Dvp9_sad4d_intrin_avx2.c76 sum = _mm_add_epi32(_mm256_castsi256_si128(sum_mlow), in vp9_sad32x32x4d_avx2()
162 sum = _mm_add_epi32(_mm256_castsi256_si128(sum_mlow), in vp9_sad64x64x4d_avx2()
Dvp9_subpel_variance_impl_intrin_avx2.c113 *((int*)sse)= _mm_cvtsi128_si32(_mm256_castsi256_si128(sse_reg)) + \
117 sum = _mm_cvtsi128_si32(_mm256_castsi256_si128(sum_reg)) + \
Dvp9_dct32x32_avx2.c2686 _mm_storeu_si128((__m128i *)(output_currStep + 0 * 32), _mm256_castsi256_si128(tr2_0)); in FDCT32x32_2D_AVX2()
2687 _mm_storeu_si128((__m128i *)(output_currStep + 1 * 32), _mm256_castsi256_si128(tr2_1)); in FDCT32x32_2D_AVX2()
2688 _mm_storeu_si128((__m128i *)(output_currStep + 2 * 32), _mm256_castsi256_si128(tr2_2)); in FDCT32x32_2D_AVX2()
2689 _mm_storeu_si128((__m128i *)(output_currStep + 3 * 32), _mm256_castsi256_si128(tr2_3)); in FDCT32x32_2D_AVX2()
2690 _mm_storeu_si128((__m128i *)(output_currStep + 4 * 32), _mm256_castsi256_si128(tr2_4)); in FDCT32x32_2D_AVX2()
2691 _mm_storeu_si128((__m128i *)(output_currStep + 5 * 32), _mm256_castsi256_si128(tr2_5)); in FDCT32x32_2D_AVX2()
2692 _mm_storeu_si128((__m128i *)(output_currStep + 6 * 32), _mm256_castsi256_si128(tr2_6)); in FDCT32x32_2D_AVX2()
2693 _mm_storeu_si128((__m128i *)(output_currStep + 7 * 32), _mm256_castsi256_si128(tr2_7)); in FDCT32x32_2D_AVX2()
/external/libhevc/common/x86/
Dihevc_platform_macros.h150 #define _mm256_storeu2_m128i(X,Y,Z) {_mm_storeu_si128 ((Y), _mm256_castsi256_si128((Z)));_mm_storeu…
/external/clang/lib/Headers/
Davxintrin.h1117 _mm256_castsi256_si128(__m256i __a) in _mm256_castsi256_si128() function
1267 __v128 = _mm256_castsi256_si128(__a); in _mm256_storeu2_m128i()