Searched refs:_mm256_castsi256_si128 (Results 1 – 9 of 9) sorted by relevance
/external/libvpx/libvpx/vp9/common/x86/ |
D | vp9_subpixel_8t_intrin_avx2.c | 191 _mm256_castsi256_si128(srcRegFilt32b1_1)); in vp9_filter_block1d16_h8_avx2() 209 _mm256_castsi256_si128(filt1Reg)); in vp9_filter_block1d16_h8_avx2() 211 _mm256_castsi256_si128(filt4Reg)); in vp9_filter_block1d16_h8_avx2() 215 _mm256_castsi256_si128(firstFilters)); in vp9_filter_block1d16_h8_avx2() 217 _mm256_castsi256_si128(forthFilters)); in vp9_filter_block1d16_h8_avx2() 224 _mm256_castsi256_si128(filt2Reg)); in vp9_filter_block1d16_h8_avx2() 226 _mm256_castsi256_si128(filt3Reg)); in vp9_filter_block1d16_h8_avx2() 230 _mm256_castsi256_si128(secondFilters)); in vp9_filter_block1d16_h8_avx2() 232 _mm256_castsi256_si128(thirdFilters)); in vp9_filter_block1d16_h8_avx2() 248 _mm256_castsi256_si128(filt1Reg)); in vp9_filter_block1d16_h8_avx2() [all …]
|
D | vp9_loopfilter_intrin_avx2.c | 616 flat2_p0 = _mm256_castsi256_si128( in mb_lpf_horizontal_edge_w_avx2_16() 624 flat2_q0 = _mm256_castsi256_si128( in mb_lpf_horizontal_edge_w_avx2_16() 632 flat_p0 = _mm256_castsi256_si128( in mb_lpf_horizontal_edge_w_avx2_16() 640 flat_q0 = _mm256_castsi256_si128( in mb_lpf_horizontal_edge_w_avx2_16() 660 flat2_p1 = _mm256_castsi256_si128( in mb_lpf_horizontal_edge_w_avx2_16() 668 flat2_q1 = _mm256_castsi256_si128( in mb_lpf_horizontal_edge_w_avx2_16() 680 flat_p1 = _mm256_castsi256_si128( in mb_lpf_horizontal_edge_w_avx2_16() 688 flat_q1 = _mm256_castsi256_si128( in mb_lpf_horizontal_edge_w_avx2_16() 708 flat2_p2 = _mm256_castsi256_si128( in mb_lpf_horizontal_edge_w_avx2_16() 716 flat2_q2 = _mm256_castsi256_si128( in mb_lpf_horizontal_edge_w_avx2_16() [all …]
|
/external/libvpx/libvpx/vp9/encoder/x86/ |
D | vp9_variance_impl_intrin_avx2.c | 78 sum_res = _mm_add_epi16(_mm256_castsi256_si128(sum_ref_src), in vp9_get16x16var_avx2() 81 madd_res = _mm_add_epi32(_mm256_castsi256_si128(madd_ref_src), in vp9_get16x16var_avx2() 85 expand_sum_low = _mm_unpacklo_epi16(_mm256_castsi256_si128(zero_reg), in vp9_get16x16var_avx2() 87 expand_sum_high = _mm_unpackhi_epi16(_mm256_castsi256_si128(zero_reg), in vp9_get16x16var_avx2() 98 _mm256_castsi256_si128(zero_reg)); in vp9_get16x16var_avx2() 100 _mm256_castsi256_si128(zero_reg)); in vp9_get16x16var_avx2() 105 _mm256_castsi256_si128(zero_reg)); in vp9_get16x16var_avx2() 107 _mm256_castsi256_si128(zero_reg)); in vp9_get16x16var_avx2() 207 *((int*)SSE)= _mm_cvtsi128_si32(_mm256_castsi256_si128(madd_ref_src)) + in vp9_get32x32var_avx2() 210 *((int*)Sum)= _mm_cvtsi128_si32(_mm256_castsi256_si128(sum_ref_src)) + in vp9_get32x32var_avx2()
|
D | vp9_error_intrin_avx2.c | 61 sse_reg128 = _mm_add_epi64(_mm256_castsi256_si128(sse_reg), in vp9_block_error_avx2() 64 ssz_reg128 = _mm_add_epi64(_mm256_castsi256_si128(ssz_reg), in vp9_block_error_avx2()
|
D | vp9_sad4d_intrin_avx2.c | 76 sum = _mm_add_epi32(_mm256_castsi256_si128(sum_mlow), in vp9_sad32x32x4d_avx2() 162 sum = _mm_add_epi32(_mm256_castsi256_si128(sum_mlow), in vp9_sad64x64x4d_avx2()
|
D | vp9_subpel_variance_impl_intrin_avx2.c | 113 *((int*)sse)= _mm_cvtsi128_si32(_mm256_castsi256_si128(sse_reg)) + \ 117 sum = _mm_cvtsi128_si32(_mm256_castsi256_si128(sum_reg)) + \
|
D | vp9_dct32x32_avx2.c | 2686 _mm_storeu_si128((__m128i *)(output_currStep + 0 * 32), _mm256_castsi256_si128(tr2_0)); in FDCT32x32_2D_AVX2() 2687 _mm_storeu_si128((__m128i *)(output_currStep + 1 * 32), _mm256_castsi256_si128(tr2_1)); in FDCT32x32_2D_AVX2() 2688 _mm_storeu_si128((__m128i *)(output_currStep + 2 * 32), _mm256_castsi256_si128(tr2_2)); in FDCT32x32_2D_AVX2() 2689 _mm_storeu_si128((__m128i *)(output_currStep + 3 * 32), _mm256_castsi256_si128(tr2_3)); in FDCT32x32_2D_AVX2() 2690 _mm_storeu_si128((__m128i *)(output_currStep + 4 * 32), _mm256_castsi256_si128(tr2_4)); in FDCT32x32_2D_AVX2() 2691 _mm_storeu_si128((__m128i *)(output_currStep + 5 * 32), _mm256_castsi256_si128(tr2_5)); in FDCT32x32_2D_AVX2() 2692 _mm_storeu_si128((__m128i *)(output_currStep + 6 * 32), _mm256_castsi256_si128(tr2_6)); in FDCT32x32_2D_AVX2() 2693 _mm_storeu_si128((__m128i *)(output_currStep + 7 * 32), _mm256_castsi256_si128(tr2_7)); in FDCT32x32_2D_AVX2()
|
/external/libhevc/common/x86/ |
D | ihevc_platform_macros.h | 150 #define _mm256_storeu2_m128i(X,Y,Z) {_mm_storeu_si128 ((Y), _mm256_castsi256_si128((Z)));_mm_storeu…
|
/external/clang/lib/Headers/ |
D | avxintrin.h | 1117 _mm256_castsi256_si128(__m256i __a) in _mm256_castsi256_si128() function 1267 __v128 = _mm256_castsi256_si128(__a); in _mm256_storeu2_m128i()
|