/external/libaom/libaom/aom_dsp/x86/ |
D | avg_intrin_avx2.c | 29 __m256i b0 = _mm256_add_epi16(a0, a1); in hadamard_col8x2_avx2() 31 __m256i b2 = _mm256_add_epi16(a2, a3); in hadamard_col8x2_avx2() 33 __m256i b4 = _mm256_add_epi16(a4, a5); in hadamard_col8x2_avx2() 35 __m256i b6 = _mm256_add_epi16(a6, a7); in hadamard_col8x2_avx2() 38 a0 = _mm256_add_epi16(b0, b2); in hadamard_col8x2_avx2() 39 a1 = _mm256_add_epi16(b1, b3); in hadamard_col8x2_avx2() 42 a4 = _mm256_add_epi16(b4, b6); in hadamard_col8x2_avx2() 43 a5 = _mm256_add_epi16(b5, b7); in hadamard_col8x2_avx2() 48 b0 = _mm256_add_epi16(a0, a4); in hadamard_col8x2_avx2() 49 b7 = _mm256_add_epi16(a1, a5); in hadamard_col8x2_avx2() [all …]
|
D | sad_highbd_avx2.c | 68 sum0 = _mm256_add_epi16(u0, u1); in aom_highbd_sad16x8_avx2() 69 sum0 = _mm256_add_epi16(sum0, u2); in aom_highbd_sad16x8_avx2() 70 sum0 = _mm256_add_epi16(sum0, u3); in aom_highbd_sad16x8_avx2() 95 sum1 = _mm256_add_epi16(u0, u1); in aom_highbd_sad16x8_avx2() 96 sum1 = _mm256_add_epi16(sum1, u2); in aom_highbd_sad16x8_avx2() 97 sum1 = _mm256_add_epi16(sum1, u3); in aom_highbd_sad16x8_avx2() 144 sum0 = _mm256_add_epi16(u0, u1); in aom_highbd_sad16x16_avx2() 145 sum0 = _mm256_add_epi16(sum0, u2); in aom_highbd_sad16x16_avx2() 146 sum0 = _mm256_add_epi16(sum0, u3); in aom_highbd_sad16x16_avx2() 198 s0 = _mm256_add_epi16(s0, s1); in sad32x4() [all …]
|
D | convolve_avx2.h | 49 _mm256_sra_epi16(_mm256_add_epi16(res, round_const_h), round_shift_h); \ 58 res = _mm256_sra_epi16(_mm256_add_epi16(res, round_const_h), round_shift_h); \ 139 _mm256_sra_epi16(_mm256_add_epi16(res, round_const_h), round_shift_h); \ 179 const __m256i res_unsigned = _mm256_add_epi16(res_16b, offset_const); \ 207 const __m256i res_unsigned = _mm256_add_epi16(res_16b, offset_const); \ 300 const __m256i res = _mm256_add_epi16(_mm256_add_epi16(res_01, res_45), in convolve_lowbd() 301 _mm256_add_epi16(res_23, res_67)); in convolve_lowbd() 312 const __m256i res = _mm256_add_epi16(res_45, res_23); in convolve_lowbd_4tap() 394 const __m256i wt_res = _mm256_add_epi16(*data_ref_0, *res_unsigned); in comp_avg() 406 _mm256_add_epi16(res_signed, *round_const), round_shift); in convolve_rounding()
|
D | intrapred_avx2.c | 27 return _mm256_add_epi16(y0, u0); in dc_sum_64() 37 return _mm256_add_epi16(y, u); in dc_sum_32() 252 sum_left = _mm256_add_epi16(sum_left, sum_above); in aom_dc_predictor_32x32_avx2() 254 sum_left = _mm256_add_epi16(sum_left, thirtytwo); in aom_dc_predictor_32x32_avx2() 268 sum = _mm256_add_epi16(sum, sixteen); in aom_dc_top_predictor_32x32_avx2() 282 sum = _mm256_add_epi16(sum, sixteen); in aom_dc_left_predictor_32x32_avx2() 392 sum_left = _mm256_add_epi16(sum_left, sum_above); in aom_dc_predictor_32x64_avx2() 404 sum_left = _mm256_add_epi16(sum_left, sum_above); in aom_dc_predictor_64x64_avx2() 416 sum_left = _mm256_add_epi16(sum_left, sum_above); in aom_dc_predictor_64x32_avx2() 428 sum_left = _mm256_add_epi16(sum_left, sum_above); in aom_dc_predictor_64x16_avx2() [all …]
|
D | variance_impl_avx2.c | 45 exp_src_lo = _mm256_add_epi16(exp_src_lo, pw8); \ 46 exp_src_hi = _mm256_add_epi16(exp_src_hi, pw8); \ 78 sum_reg = _mm256_add_epi16(sum_reg, exp_src_lo); \ 80 sum_reg = _mm256_add_epi16(sum_reg, exp_src_hi); \
|
D | highbd_variance_avx2.c | 37 v_sum_d = _mm256_add_epi16(v_sum_d, v_diff); in aom_highbd_calc8x8var_avx2() 67 v_sum_d = _mm256_add_epi16(v_sum_d, v_diff); in aom_highbd_calc16x16var_avx2()
|
D | blend_a64_mask_avx2.c | 137 const __m256i m0 = _mm256_srli_epi16(_mm256_add_epi16(m0_acbd, two_w), 2); in lowbd_blend_a64_d16_mask_subw1_subh1_w16_avx2() 167 const __m256i m0 = _mm256_srli_epi16(_mm256_add_epi16(m0_acbd, two_w), 2); in lowbd_blend_a64_d16_mask_subw1_subh1_w32_avx2() 168 const __m256i m1 = _mm256_srli_epi16(_mm256_add_epi16(m1_acbd, two_w), 2); in lowbd_blend_a64_d16_mask_subw1_subh1_w32_avx2() 456 const __m256i v_rsl_w = _mm256_add_epi16(v_rvsal_w, v_rvsbl_w); in blend_a64_mask_sx_sy_w16_avx2() 494 const __m256i v_rsl_w = _mm256_add_epi16(v_rvsal_w, v_rvsbl_w); in blend_a64_mask_sx_sy_w32n_avx2() 495 const __m256i v_rsh_w = _mm256_add_epi16(v_rvsah_w, v_rvsbh_w); in blend_a64_mask_sx_sy_w32n_avx2() 1008 _mm256_srli_epi16(_mm256_add_epi16(adjacent, two_w), 2); in highbd_blend_a64_d16_mask_subw1_subh1_w4_avx2() 1140 _mm256_srli_epi16(_mm256_add_epi16(m0123, two_w), 2); in highbd_blend_a64_d16_mask_subw1_subh1_w8_avx2() 1148 _mm256_srli_epi16(_mm256_add_epi16(m4567, two_w), 2); in highbd_blend_a64_d16_mask_subw1_subh1_w8_avx2() 1280 const __m256i mask0a = _mm256_srli_epi16(_mm256_add_epi16(m01, two_w), 2); in highbd_blend_a64_d16_mask_subw1_subh1_w16_avx2() [all …]
|
D | variance_avx2.c | 44 *sum = _mm256_add_epi16(*sum, _mm256_add_epi16(diff0, diff1)); in variance_kernel_avx2()
|
/external/libvpx/libvpx/vpx_dsp/x86/ |
D | loopfilter_avx2.c | 600 pixelFilter_p = _mm256_add_epi16(_mm256_add_epi16(p256_6, p256_5), in vpx_lpf_horizontal_16_dual_avx2() 601 _mm256_add_epi16(p256_4, p256_3)); in vpx_lpf_horizontal_16_dual_avx2() 602 pixelFilter_q = _mm256_add_epi16(_mm256_add_epi16(q256_6, q256_5), in vpx_lpf_horizontal_16_dual_avx2() 603 _mm256_add_epi16(q256_4, q256_3)); in vpx_lpf_horizontal_16_dual_avx2() 606 _mm256_add_epi16(p256_0, _mm256_add_epi16(p256_2, p256_1)); in vpx_lpf_horizontal_16_dual_avx2() 607 pixelFilter_p = _mm256_add_epi16(pixelFilter_p, pixetFilter_p2p1p0); in vpx_lpf_horizontal_16_dual_avx2() 610 _mm256_add_epi16(q256_0, _mm256_add_epi16(q256_2, q256_1)); in vpx_lpf_horizontal_16_dual_avx2() 611 pixelFilter_q = _mm256_add_epi16(pixelFilter_q, pixetFilter_q2q1q0); in vpx_lpf_horizontal_16_dual_avx2() 613 pixelFilter_p = _mm256_add_epi16( in vpx_lpf_horizontal_16_dual_avx2() 614 eight, _mm256_add_epi16(pixelFilter_p, pixelFilter_q)); in vpx_lpf_horizontal_16_dual_avx2() [all …]
|
D | avg_intrin_avx2.c | 231 __m256i b0 = _mm256_add_epi16(a0, a1); in hadamard_col8x2_avx2() 233 __m256i b2 = _mm256_add_epi16(a2, a3); in hadamard_col8x2_avx2() 235 __m256i b4 = _mm256_add_epi16(a4, a5); in hadamard_col8x2_avx2() 237 __m256i b6 = _mm256_add_epi16(a6, a7); in hadamard_col8x2_avx2() 240 a0 = _mm256_add_epi16(b0, b2); in hadamard_col8x2_avx2() 241 a1 = _mm256_add_epi16(b1, b3); in hadamard_col8x2_avx2() 244 a4 = _mm256_add_epi16(b4, b6); in hadamard_col8x2_avx2() 245 a5 = _mm256_add_epi16(b5, b7); in hadamard_col8x2_avx2() 250 b0 = _mm256_add_epi16(a0, a4); in hadamard_col8x2_avx2() 251 b7 = _mm256_add_epi16(a1, a5); in hadamard_col8x2_avx2() [all …]
|
D | fwd_dct32x32_impl_avx2.h | 131 step1a[0] = _mm256_add_epi16(ina0, inb0); in FDCT32x32_2D_AVX2() 132 step1a[1] = _mm256_add_epi16(ina1, inb1); in FDCT32x32_2D_AVX2() 133 step1a[2] = _mm256_add_epi16(ina2, inb2); in FDCT32x32_2D_AVX2() 134 step1a[3] = _mm256_add_epi16(ina3, inb3); in FDCT32x32_2D_AVX2() 167 step1a[0] = _mm256_add_epi16(ina0, inb0); in FDCT32x32_2D_AVX2() 168 step1a[1] = _mm256_add_epi16(ina1, inb1); in FDCT32x32_2D_AVX2() 169 step1a[2] = _mm256_add_epi16(ina2, inb2); in FDCT32x32_2D_AVX2() 170 step1a[3] = _mm256_add_epi16(ina3, inb3); in FDCT32x32_2D_AVX2() 203 step1a[0] = _mm256_add_epi16(ina0, inb0); in FDCT32x32_2D_AVX2() 204 step1a[1] = _mm256_add_epi16(ina1, inb1); in FDCT32x32_2D_AVX2() [all …]
|
D | convolve_avx2.h | 66 sum1 = _mm256_add_epi16(x0, x2); in convolve8_16_avx2() 67 sum2 = _mm256_add_epi16(x1, x3); in convolve8_16_avx2() 69 sum1 = _mm256_add_epi16(sum1, k_64); in convolve8_16_avx2()
|
D | variance_avx2.c | 57 *sum = _mm256_add_epi16(*sum, _mm256_add_epi16(diff0, diff1)); in variance_kernel_avx2() 181 exp_src_lo = _mm256_add_epi16(exp_src_lo, pw8); \ 182 exp_src_hi = _mm256_add_epi16(exp_src_hi, pw8); \ 196 *sum_reg = _mm256_add_epi16(*sum_reg, exp_src_lo); \ 198 *sum_reg = _mm256_add_epi16(*sum_reg, exp_src_hi); \
|
/external/libaom/libaom/av1/common/x86/ |
D | convolve_avx2.c | 98 _mm256_add_epi16(res_lo, right_shift_const), right_shift); in av1_convolve_y_sr_avx2() 108 _mm256_add_epi16(res_hi, right_shift_const), right_shift); in av1_convolve_y_sr_avx2() 211 _mm256_add_epi16(res_lo, right_shift_const), right_shift); in av1_convolve_y_sr_avx2() 221 _mm256_add_epi16(res_hi, right_shift_const), right_shift); in av1_convolve_y_sr_avx2() 310 res_16b = _mm256_sra_epi16(_mm256_add_epi16(res_16b, round_0_const), in av1_convolve_x_sr_avx2() 313 res_16b = _mm256_sra_epi16(_mm256_add_epi16(res_16b, round_const), in av1_convolve_x_sr_avx2() 348 res_16b = _mm256_sra_epi16(_mm256_add_epi16(res_16b, round_0_const), in av1_convolve_x_sr_avx2() 351 res_16b = _mm256_sra_epi16(_mm256_add_epi16(res_16b, round_const), in av1_convolve_x_sr_avx2() 383 res_16b = _mm256_sra_epi16(_mm256_add_epi16(res_16b, round_0_const), in av1_convolve_x_sr_avx2() 386 res_16b = _mm256_sra_epi16(_mm256_add_epi16(res_16b, round_const), in av1_convolve_x_sr_avx2() [all …]
|
D | jnt_convolve_avx2.c | 93 res = _mm256_sra_epi16(_mm256_add_epi16(res, round_const), round_shift); in av1_dist_wtd_convolve_x_avx2() 96 const __m256i res_unsigned = _mm256_add_epi16(res, offset_const); in av1_dist_wtd_convolve_x_avx2() 147 res = _mm256_sra_epi16(_mm256_add_epi16(res, round_const), round_shift); in av1_dist_wtd_convolve_x_avx2() 151 const __m256i res_unsigned = _mm256_add_epi16(res, offset_const); in av1_dist_wtd_convolve_x_avx2() 274 res_lo = _mm256_add_epi16(res_lo, offset_const_1); in av1_dist_wtd_convolve_y_avx2() 292 _mm256_add_epi16(res_lo_round, offset_const_2); in av1_dist_wtd_convolve_y_avx2() 332 res_hi = _mm256_add_epi16(res_hi, offset_const_1); in av1_dist_wtd_convolve_y_avx2() 350 _mm256_add_epi16(res_hi_round, offset_const_2); in av1_dist_wtd_convolve_y_avx2() 453 res_lo = _mm256_add_epi16(res_lo, offset_const_1); in av1_dist_wtd_convolve_y_avx2() 471 _mm256_add_epi16(res_lo_round, offset_const_2); in av1_dist_wtd_convolve_y_avx2() [all …]
|
D | cfl_avx2.c | 76 __m256i sum_16x16 = _mm256_add_epi16(top_16x16, bot_16x16); in cfl_luma_subsampling_420_lbd_avx2() 173 __m256i sum = _mm256_add_epi16(top, bot); in cfl_luma_subsampling_420_hbd_avx2() 177 __m256i sum_1 = _mm256_add_epi16(top_1, bot_1); in cfl_luma_subsampling_420_hbd_avx2() 181 hsum = _mm256_add_epi16(hsum, hsum); in cfl_luma_subsampling_420_hbd_avx2() 249 return _mm256_add_epi16(scaled_luma_q0, dc_q0); in predict_unclipped() 419 __m256i l0 = _mm256_add_epi16(_mm256_loadu_si256(src), in subtract_average_avx2() 425 _mm256_add_epi16(_mm256_loadu_si256(src + 1), in subtract_average_avx2()
|
D | reconinter_avx2.c | 26 _mm256_add_epi16(mask_base, _mm256_srli_epi16(diff, 4))); in calc_mask_avx2() 544 _mm256_max_epi16(y0, _mm256_add_epi16(diff, ymask_base)), in av1_build_compound_diffwtd_mask_highbd_avx2() 564 _mm256_max_epi16(y0, _mm256_add_epi16(diff, ymask_base)), in av1_build_compound_diffwtd_mask_highbd_avx2() 586 _mm256_max_epi16(y0, _mm256_add_epi16(diff, ymask_base)), in av1_build_compound_diffwtd_mask_highbd_avx2() 606 _mm256_max_epi16(y0, _mm256_add_epi16(diff, ymask_base)), in av1_build_compound_diffwtd_mask_highbd_avx2()
|
D | convolve_2d_avx2.c | 89 res = _mm256_sra_epi16(_mm256_add_epi16(res, round_const_h), in av1_convolve_2d_sr_avx2() 99 _mm256_sra_epi16(_mm256_add_epi16(res, round_const_h), round_shift_h); in av1_convolve_2d_sr_avx2()
|
/external/libaom/libaom/av1/encoder/x86/ |
D | corner_match_avx2.c | 56 sum_vec = _mm256_add_epi16(sum_vec, _mm256_sad_epu8(v, zero)); in compute_cross_correlation_avx2()
|
D | pickrst_avx2.c | 531 _mm256_add_epi16(_mm256_packs_epi32(vr0, vr1), d0), s0); in av1_lowbd_pixel_proj_error_avx2() 576 _mm256_add_epi16(_mm256_packs_epi32(vr0, vr1), d0), s0); in av1_lowbd_pixel_proj_error_avx2() 694 const __m256i e0 = _mm256_sub_epi16(_mm256_add_epi16(vr, d0), s0); in av1_highbd_pixel_proj_error_avx2() 771 const __m256i e0 = _mm256_sub_epi16(_mm256_add_epi16(vr, d0), s0); in av1_highbd_pixel_proj_error_avx2()
|
/external/clang/test/CodeGen/ |
D | avx2-builtins.c | 38 return _mm256_add_epi16(a, b); in test_mm256_add_epi16()
|
/external/libaom/libaom/aom_dsp/simd/ |
D | v256_intrinsics_x86.h | 91 SIMD_INLINE v256 v256_add_16(v256 a, v256 b) { return _mm256_add_epi16(a, b); } in v256_add_16()
|
/external/clang/lib/Headers/ |
D | avx2intrin.h | 88 _mm256_add_epi16(__m256i __a, __m256i __b) in _mm256_add_epi16() function
|