Home
last modified time | relevance | path

Searched refs:_mm256_unpackhi_epi16 (Results 1 – 25 of 48) sorted by relevance

12

/external/libaom/libaom/aom_dsp/x86/
Dcommon_avx2.h22 __m256i tr0_1 = _mm256_unpackhi_epi16(in[0], in[1]); in mm256_transpose_16x16()
24 __m256i tr0_3 = _mm256_unpackhi_epi16(in[2], in[3]); in mm256_transpose_16x16()
26 __m256i tr0_5 = _mm256_unpackhi_epi16(in[4], in[5]); in mm256_transpose_16x16()
28 __m256i tr0_7 = _mm256_unpackhi_epi16(in[6], in[7]); in mm256_transpose_16x16()
31 __m256i tr0_9 = _mm256_unpackhi_epi16(in[8], in[9]); in mm256_transpose_16x16()
33 __m256i tr0_b = _mm256_unpackhi_epi16(in[10], in[11]); in mm256_transpose_16x16()
35 __m256i tr0_d = _mm256_unpackhi_epi16(in[12], in[13]); in mm256_transpose_16x16()
37 __m256i tr0_f = _mm256_unpackhi_epi16(in[14], in[15]); in mm256_transpose_16x16()
Dconvolve_avx2.h100 s[4] = _mm256_unpackhi_epi16(src_0, src_1); \
101 s[5] = _mm256_unpackhi_epi16(src_2, src_3); \
102 s[6] = _mm256_unpackhi_epi16(src_4, src_5); \
111 s[7] = _mm256_unpackhi_epi16(s6, s7); \
182 s[4] = _mm256_unpackhi_epi16(s0, s1); \
183 s[5] = _mm256_unpackhi_epi16(s2, s3); \
184 s[6] = _mm256_unpackhi_epi16(s4, s5); \
193 s[7] = _mm256_unpackhi_epi16(s6, s7); \
409 const __m256i data_hi = _mm256_unpackhi_epi16(*data_ref_0, *res_unsigned); in comp_avg()
Dtxfm_common_avx2.h32 __m256i t1 = _mm256_unpackhi_epi16(*in0, *in1); in btf_16_w16_avx2()
127 t[2 * i + 1] = _mm256_unpackhi_epi16(in[2 * i], in[2 * i + 1]); in transpose2_8x8_avx2()
204 const __m256i a4 = _mm256_unpackhi_epi16(in[0], in[1]); in transpose_16bit_16x8_avx2()
205 const __m256i a5 = _mm256_unpackhi_epi16(in[2], in[3]); in transpose_16bit_16x8_avx2()
206 const __m256i a6 = _mm256_unpackhi_epi16(in[4], in[5]); in transpose_16bit_16x8_avx2()
207 const __m256i a7 = _mm256_unpackhi_epi16(in[6], in[7]); in transpose_16bit_16x8_avx2()
307 const __m256i a_hi = _mm256_unpackhi_epi16(a, one); in store_rect_16bit_to_32bit_w8_avx2()
Dhighbd_convolve_avx2.c189 s[4] = _mm256_unpackhi_epi16(s01, s12); in av1_highbd_convolve_y_sr_avx2()
190 s[5] = _mm256_unpackhi_epi16(s23, s34); in av1_highbd_convolve_y_sr_avx2()
191 s[6] = _mm256_unpackhi_epi16(s45, s56); in av1_highbd_convolve_y_sr_avx2()
211 s[7] = _mm256_unpackhi_epi16(s67, s78); in av1_highbd_convolve_y_sr_avx2()
880 sig[4] = _mm256_unpackhi_epi16(s0, s1); in pack_8x9_init()
882 sig[5] = _mm256_unpackhi_epi16(s2, s3); in pack_8x9_init()
884 sig[6] = _mm256_unpackhi_epi16(s4, s5); in pack_8x9_init()
899 sig[7] = _mm256_unpackhi_epi16(s2, s3); in pack_8x9_pixels()
959 sig[4] = _mm256_unpackhi_epi16(u0, u2); in pack_16x9_init()
962 sig[12] = _mm256_unpackhi_epi16(u1, u3); in pack_16x9_init()
[all …]
Dblend_a64_mask_avx2.c37 __m256i res0_hi = _mm256_madd_epi16(_mm256_unpackhi_epi16(s0_0, s1_0), in blend_a64_d16_mask_w16_avx2()
38 _mm256_unpackhi_epi16(*m0, max_minus_m0)); in blend_a64_d16_mask_w16_avx2()
61 __m256i res0_hi = _mm256_madd_epi16(_mm256_unpackhi_epi16(s0_0, s1_0), in blend_a64_d16_mask_w32_avx2()
62 _mm256_unpackhi_epi16(*m0, max_minus_m0)); in blend_a64_d16_mask_w32_avx2()
65 __m256i res1_hi = _mm256_madd_epi16(_mm256_unpackhi_epi16(s0_1, s1_1), in blend_a64_d16_mask_w32_avx2()
66 _mm256_unpackhi_epi16(*m1, max_minus_m1)); in blend_a64_d16_mask_w32_avx2()
927 const __m256i mul0h = _mm256_unpackhi_epi16(mul0_lows, mul0_highs); in highbd_blend_a64_d16_mask_w4_avx2()
935 const __m256i mul1h = _mm256_unpackhi_epi16(mul1_lows, mul1_highs); in highbd_blend_a64_d16_mask_w4_avx2()
1045 const __m256i mul0ah = _mm256_unpackhi_epi16(mul0a_lows, mul0a_highs); in highbd_blend_a64_d16_mask_w8_avx2()
1053 const __m256i mul1ah = _mm256_unpackhi_epi16(mul1a_lows, mul1a_highs); in highbd_blend_a64_d16_mask_w8_avx2()
[all …]
Dbitdepth_conversion_avx2.h29 const __m256i a_2 = _mm256_unpackhi_epi16(a_lo, a_hi); in store_tran_low()
Dmasked_sad_intrin_avx2.c232 const __m256i data_r = _mm256_unpackhi_epi16(a, b); in highbd_masked_sad8xh_avx2()
233 const __m256i mask_r = _mm256_unpackhi_epi16(m, m_inv); in highbd_masked_sad8xh_avx2()
289 const __m256i data_r = _mm256_unpackhi_epi16(a, b); in highbd_masked_sad16xh_avx2()
290 const __m256i mask_r = _mm256_unpackhi_epi16(m, m_inv); in highbd_masked_sad16xh_avx2()
Dsum_squares_avx2.c202 __m256i vsrc1 = _mm256_unpackhi_epi16(vsrc[k], vzero); in aom_var_2d_u16_avx2()
221 __m256i vsrc1 = _mm256_unpackhi_epi16(vsrc, vzero); in aom_var_2d_u16_avx2()
Dadaptive_quantize_avx2.c63 coeff[1] = _mm256_unpackhi_epi16(*qcoeff, zero); in update_mask0_avx2()
91 __m256i coeff_vals_hi = _mm256_unpackhi_epi16(coeff_vals, coeff_sign); in store_coefficients_avx2()
Davg_intrin_avx2.c59 a2 = _mm256_unpackhi_epi16(b0, b1); in hadamard_col8x2_avx2()
60 a3 = _mm256_unpackhi_epi16(b2, b3); in hadamard_col8x2_avx2()
63 a6 = _mm256_unpackhi_epi16(b4, b5); in hadamard_col8x2_avx2()
64 a7 = _mm256_unpackhi_epi16(b6, b7); in hadamard_col8x2_avx2()
/external/libaom/libaom/av1/encoder/x86/
Dwedge_utils_avx2.c50 const __m256i v_rd0h_w = _mm256_unpackhi_epi16(v_d0_w, v_r0_w); in av1_wedge_sse_from_residuals_avx2()
54 const __m256i v_m0h_w = _mm256_unpackhi_epi16(v_m0_w, v_mask_max_w); in av1_wedge_sse_from_residuals_avx2()
173 const __m256i v_ab0h_w = _mm256_unpackhi_epi16(v_a0_w, v_b0_w); in av1_wedge_compute_delta_squares_avx2()
175 const __m256i v_ab1h_w = _mm256_unpackhi_epi16(v_a1_w, v_b1_w); in av1_wedge_compute_delta_squares_avx2()
177 const __m256i v_ab2h_w = _mm256_unpackhi_epi16(v_a2_w, v_b2_w); in av1_wedge_compute_delta_squares_avx2()
179 const __m256i v_ab3h_w = _mm256_unpackhi_epi16(v_a3_w, v_b3_w); in av1_wedge_compute_delta_squares_avx2()
/external/libaom/libaom/av1/common/x86/
Dhighbd_jnt_convolve_avx2.c75 const __m256i data_ref_0_hi = _mm256_unpackhi_epi16(data_0, zero); in av1_highbd_dist_wtd_convolve_2d_copy_avx2()
85 const __m256i res_32b_hi = _mm256_unpackhi_epi16(res, zero); in av1_highbd_dist_wtd_convolve_2d_copy_avx2()
179 const __m256i data_ref_0_hi = _mm256_unpackhi_epi16(data_01, zero); in av1_highbd_dist_wtd_convolve_2d_copy_avx2()
189 const __m256i res_32b_hi = _mm256_unpackhi_epi16(res, zero); in av1_highbd_dist_wtd_convolve_2d_copy_avx2()
338 s[4] = _mm256_unpackhi_epi16(s0, s1); in av1_highbd_dist_wtd_convolve_2d_avx2()
339 s[5] = _mm256_unpackhi_epi16(s2, s3); in av1_highbd_dist_wtd_convolve_2d_avx2()
340 s[6] = _mm256_unpackhi_epi16(s4, s5); in av1_highbd_dist_wtd_convolve_2d_avx2()
351 s[7] = _mm256_unpackhi_epi16(s6, s7); in av1_highbd_dist_wtd_convolve_2d_avx2()
416 const __m256i data_ref_0_hi = _mm256_unpackhi_epi16(data_01, zero); in av1_highbd_dist_wtd_convolve_2d_avx2()
597 const __m256i data_ref_0_hi = _mm256_unpackhi_epi16(data_01, zero); in av1_highbd_dist_wtd_convolve_x_avx2()
[all …]
Dwiener_convolve_avx2.c164 s[4] = _mm256_unpackhi_epi16(src_0, src_1); in av1_wiener_convolve_add_src_avx2()
165 s[5] = _mm256_unpackhi_epi16(src_2, src_3); in av1_wiener_convolve_add_src_avx2()
166 s[6] = _mm256_unpackhi_epi16(src_4, src_5); in av1_wiener_convolve_add_src_avx2()
177 s[7] = _mm256_unpackhi_epi16(s6, s7); in av1_wiener_convolve_add_src_avx2()
Dhighbd_wiener_convolve_avx2.c205 const __m256i src_1 = _mm256_unpackhi_epi16(data_0, data_1); in av1_highbd_wiener_convolve_add_src_avx2()
206 const __m256i src_3 = _mm256_unpackhi_epi16(data_2, data_3); in av1_highbd_wiener_convolve_add_src_avx2()
207 const __m256i src_5 = _mm256_unpackhi_epi16(data_4, data_5); in av1_highbd_wiener_convolve_add_src_avx2()
208 const __m256i src_7 = _mm256_unpackhi_epi16(data_6, data_7); in av1_highbd_wiener_convolve_add_src_avx2()
Dwarp_plane_avx2.c643 src[7] = _mm256_unpackhi_epi16(src_6, src_7); in filter_src_pixels_vertical_avx2()
802 src[1] = _mm256_unpackhi_epi16(src_0, src_1); in warp_vertical_filter_avx2()
803 src[3] = _mm256_unpackhi_epi16(src_2, src_3); in warp_vertical_filter_avx2()
804 src[5] = _mm256_unpackhi_epi16(src_4, src_5); in warp_vertical_filter_avx2()
852 src[1] = _mm256_unpackhi_epi16(src_0, src_1); in warp_vertical_filter_gamma0_avx2()
853 src[3] = _mm256_unpackhi_epi16(src_2, src_3); in warp_vertical_filter_gamma0_avx2()
854 src[5] = _mm256_unpackhi_epi16(src_4, src_5); in warp_vertical_filter_gamma0_avx2()
901 src[1] = _mm256_unpackhi_epi16(src_0, src_1); in warp_vertical_filter_delta0_avx2()
902 src[3] = _mm256_unpackhi_epi16(src_2, src_3); in warp_vertical_filter_delta0_avx2()
903 src[5] = _mm256_unpackhi_epi16(src_4, src_5); in warp_vertical_filter_delta0_avx2()
[all …]
Dhighbd_convolve_2d_avx2.c119 s[4] = _mm256_unpackhi_epi16(s0, s1); in av1_highbd_convolve_2d_sr_avx2()
120 s[5] = _mm256_unpackhi_epi16(s2, s3); in av1_highbd_convolve_2d_sr_avx2()
121 s[6] = _mm256_unpackhi_epi16(s4, s5); in av1_highbd_convolve_2d_sr_avx2()
132 s[7] = _mm256_unpackhi_epi16(s6, s7); in av1_highbd_convolve_2d_sr_avx2()
Dconvolve_2d_avx2.c126 s[3] = _mm256_unpackhi_epi16(src_0, src_1); in av1_convolve_2d_sr_avx2()
127 s[4] = _mm256_unpackhi_epi16(src_2, src_3); in av1_convolve_2d_sr_avx2()
138 s[5] = _mm256_unpackhi_epi16(s4, s5); in av1_convolve_2d_sr_avx2()
Djnt_convolve_avx2.c282 const __m256i res_lo_1_32b = _mm256_unpackhi_epi16(res_lo, zero); in av1_dist_wtd_convolve_y_avx2()
340 const __m256i res_hi_1_32b = _mm256_unpackhi_epi16(res_hi, zero); in av1_dist_wtd_convolve_y_avx2()
461 const __m256i res_lo_1_32b = _mm256_unpackhi_epi16(res_lo, zero); in av1_dist_wtd_convolve_y_avx2()
519 const __m256i res_hi_1_32b = _mm256_unpackhi_epi16(res_hi, zero); in av1_dist_wtd_convolve_y_avx2()
695 s[3] = _mm256_unpackhi_epi16(s0, s1); in av1_dist_wtd_convolve_2d_avx2()
696 s[4] = _mm256_unpackhi_epi16(s2, s3); in av1_dist_wtd_convolve_2d_avx2()
707 s[5] = _mm256_unpackhi_epi16(s4, s5); in av1_dist_wtd_convolve_2d_avx2()
/external/libvpx/libvpx/vpx_dsp/x86/
Dfwd_dct32x32_impl_avx2.h358 const __m256i s2_20_1 = _mm256_unpackhi_epi16(step1[27], step1[20]); in FDCT32x32_2D_AVX2()
360 const __m256i s2_21_1 = _mm256_unpackhi_epi16(step1[26], step1[21]); in FDCT32x32_2D_AVX2()
362 const __m256i s2_22_1 = _mm256_unpackhi_epi16(step1[25], step1[22]); in FDCT32x32_2D_AVX2()
364 const __m256i s2_23_1 = _mm256_unpackhi_epi16(step1[24], step1[23]); in FDCT32x32_2D_AVX2()
595 const __m256i s3_10_1 = _mm256_unpackhi_epi16(step2[13], step2[10]); in FDCT32x32_2D_AVX2()
597 const __m256i s3_11_1 = _mm256_unpackhi_epi16(step2[12], step2[11]); in FDCT32x32_2D_AVX2()
673 const __m256i s1_05_1 = _mm256_unpackhi_epi16(step3[6], step3[5]); in FDCT32x32_2D_AVX2()
697 const __m256i s1_18_1 = _mm256_unpackhi_epi16(step3[18], step3[29]); in FDCT32x32_2D_AVX2()
699 const __m256i s1_19_1 = _mm256_unpackhi_epi16(step3[19], step3[28]); in FDCT32x32_2D_AVX2()
701 const __m256i s1_20_1 = _mm256_unpackhi_epi16(step3[20], step3[27]); in FDCT32x32_2D_AVX2()
[all …]
Dhighbd_convolve_avx2.c531 sig[4] = _mm256_unpackhi_epi16(s0, s1); in pack_8x9_init()
533 sig[5] = _mm256_unpackhi_epi16(s2, s3); in pack_8x9_init()
535 sig[6] = _mm256_unpackhi_epi16(s4, s5); in pack_8x9_init()
550 sig[7] = _mm256_unpackhi_epi16(s2, s3); in pack_8x9_pixels()
610 sig[4] = _mm256_unpackhi_epi16(u0, u2); in pack_16x9_init()
613 sig[12] = _mm256_unpackhi_epi16(u1, u3); in pack_16x9_init()
622 sig[5] = _mm256_unpackhi_epi16(u0, u2); in pack_16x9_init()
625 sig[13] = _mm256_unpackhi_epi16(u1, u3); in pack_16x9_init()
634 sig[6] = _mm256_unpackhi_epi16(u0, u2); in pack_16x9_init()
637 sig[14] = _mm256_unpackhi_epi16(u1, u3); in pack_16x9_init()
[all …]
Dbitdepth_conversion_avx2.h37 const __m256i a_2 = _mm256_unpackhi_epi16(a_lo, a_hi); in store_tran_low()
Davg_intrin_avx2.c261 a2 = _mm256_unpackhi_epi16(b0, b1); in hadamard_col8x2_avx2()
262 a3 = _mm256_unpackhi_epi16(b2, b3); in hadamard_col8x2_avx2()
265 a6 = _mm256_unpackhi_epi16(b4, b5); in hadamard_col8x2_avx2()
266 a7 = _mm256_unpackhi_epi16(b6, b7); in hadamard_col8x2_avx2()
/external/libgav1/libgav1/src/dsp/x86/
Dcdef_avx2.cc313 const __m256i v_hsum_5_4 = _mm256_unpackhi_epi16(v_hsum_4_0, v_hsum_5_1); in AddPartial()
314 const __m256i v_hsum_7_6 = _mm256_unpackhi_epi16(v_hsum_6_2, v_hsum_7_3); in AddPartial()
386 const __m256i ab_hi = _mm256_unpackhi_epi16(a, b_reversed); in Cost0Or4_Pair()
421 const __m256i ab_hi = _mm256_unpackhi_epi16(a, b_reversed); in CostOdd_Pair()
Dloop_restoration_10bit_avx2.cc65 const __m256i ss1 = _mm256_unpackhi_epi16(s06, s15); in WienerHorizontalTap7Kernel()
67 const __m256i ss3 = _mm256_unpackhi_epi16(s24, s[3]); in WienerHorizontalTap7Kernel()
86 const __m256i ss1 = _mm256_unpackhi_epi16(s0m, s1m); in WienerHorizontalTap5Kernel()
91 const __m256i s2_hi = _mm256_unpackhi_epi16(s[2], _mm256_setzero_si256()); in WienerHorizontalTap5Kernel()
103 const __m256i ss1 = _mm256_unpackhi_epi16(s02, s[1]); in WienerHorizontalTap3Kernel()
240 b[0] = _mm256_unpackhi_epi16(a[0], a[1]); in WienerVerticalFilter7()
241 b[1] = _mm256_unpackhi_epi16(a[2], a[3]); in WienerVerticalFilter7()
242 b[2] = _mm256_unpackhi_epi16(a[4], a[5]); in WienerVerticalFilter7()
243 b[3] = _mm256_unpackhi_epi16(a[6], round); in WienerVerticalFilter7()
256 b[0] = _mm256_unpackhi_epi16(a[0], a[1]); in WienerVerticalFilter5()
[all …]
/external/gemmlowp/internal/
Dpack_avx.h90 __m256i ymm9 = _mm256_unpackhi_epi16(ymm1, ymm2); in Pack()
91 __m256i ymm10 = _mm256_unpackhi_epi16(ymm3, ymm4); in Pack()

12