/external/libaom/libaom/aom_dsp/x86/ |
D | aom_subpixel_8t_intrin_sse2.c | 36 const __m128i tmp_0 = _mm_unpacklo_epi32(filtersReg, filtersReg); in aom_filter_block1d16_h4_sse2() local 40 secondFilters = _mm_unpackhi_epi64(tmp_0, tmp_0); // coeffs 2 3 2 3 2 3 2 3 in aom_filter_block1d16_h4_sse2() 121 __m128i tmp_0, tmp_1; in aom_filter_block1d16_v4_sse2() local 171 tmp_0 = _mm_madd_epi16(resReg23_lo_1, secondFilters); in aom_filter_block1d16_v4_sse2() 173 resReg23_lo = _mm_packs_epi32(tmp_0, tmp_1); in aom_filter_block1d16_v4_sse2() 175 tmp_0 = _mm_madd_epi16(resReg34_lo_1, secondFilters); in aom_filter_block1d16_v4_sse2() 177 resReg34_lo = _mm_packs_epi32(tmp_0, tmp_1); in aom_filter_block1d16_v4_sse2() 181 tmp_0 = _mm_madd_epi16(resReg45_lo_1, thirdFilters); in aom_filter_block1d16_v4_sse2() 183 resReg45_lo = _mm_packs_epi32(tmp_0, tmp_1); in aom_filter_block1d16_v4_sse2() 187 tmp_0 = _mm_madd_epi16(resReg56_lo_1, thirdFilters); in aom_filter_block1d16_v4_sse2() [all …]
|
D | highbd_convolve_sse2.c | 120 const __m128i tmp_0 = _mm_unpacklo_epi32(filtersReg, filtersReg); in aom_highbd_filter_block1d4_h4_sse2() local 124 secondFilters = _mm_unpackhi_epi64(tmp_0, tmp_0); // coeffs 2 3 2 3 2 3 2 3 in aom_highbd_filter_block1d4_h4_sse2() 281 const __m128i tmp_0 = _mm_unpacklo_epi32(filtersReg, filtersReg); in aom_highbd_filter_block1d8_h4_sse2() local 285 secondFilters = _mm_unpackhi_epi64(tmp_0, tmp_0); // coeffs 2 3 2 3 2 3 2 3 in aom_highbd_filter_block1d8_h4_sse2()
|
D | masked_variance_intrin_ssse3.c | 947 const __m128i tmp_0 = _mm_unpacklo_epi32(prod, sign); in highbd_masked_variance() local 949 sum_sq = _mm_add_epi64(sum_sq, _mm_add_epi64(tmp_0, tmp_1)); in highbd_masked_variance()
|
/external/aac/libAACenc/src/ |
D | chaosmeasure.cpp | 138 FIXP_DBL tmp_0 = left_0_div2 + (right_0 >> 1); in FDKaacEnc_FDKaacEnc_CalculateChaosMeasurePeakFast() local 144 if (tmp_0 < center_0) { in FDKaacEnc_FDKaacEnc_CalculateChaosMeasurePeakFast() 146 tmp_0 = schur_div(tmp_0 << leadingBits, center_0 << leadingBits, 8); in FDKaacEnc_FDKaacEnc_CalculateChaosMeasurePeakFast() 147 tmp_0 = fMult(tmp_0, tmp_0); in FDKaacEnc_FDKaacEnc_CalculateChaosMeasurePeakFast() 149 tmp_0 = (FIXP_DBL)MAXVAL_DBL; in FDKaacEnc_FDKaacEnc_CalculateChaosMeasurePeakFast() 151 chaosMeasure[j + 0] = tmp_0; in FDKaacEnc_FDKaacEnc_CalculateChaosMeasurePeakFast()
|
/external/libaom/libaom/av1/common/x86/ |
D | wiener_convolve_sse2.c | 51 const __m128i tmp_0 = _mm_unpacklo_epi32(coeffs_x, coeffs_x); in av1_wiener_convolve_add_src_sse2() local 56 const __m128i coeff_01 = _mm_unpacklo_epi64(tmp_0, tmp_0); in av1_wiener_convolve_add_src_sse2() 58 const __m128i coeff_23 = _mm_unpackhi_epi64(tmp_0, tmp_0); in av1_wiener_convolve_add_src_sse2() 118 const __m128i tmp_0 = _mm_unpacklo_epi32(coeffs_y, coeffs_y); in av1_wiener_convolve_add_src_sse2() local 123 const __m128i coeff_01 = _mm_unpacklo_epi64(tmp_0, tmp_0); in av1_wiener_convolve_add_src_sse2() 125 const __m128i coeff_23 = _mm_unpackhi_epi64(tmp_0, tmp_0); in av1_wiener_convolve_add_src_sse2()
|
D | highbd_wiener_convolve_ssse3.c | 52 const __m128i tmp_0 = _mm_unpacklo_epi32(coeffs_x, coeffs_x); in av1_highbd_wiener_convolve_add_src_ssse3() local 57 const __m128i coeff_01 = _mm_unpacklo_epi64(tmp_0, tmp_0); in av1_highbd_wiener_convolve_add_src_ssse3() 59 const __m128i coeff_23 = _mm_unpackhi_epi64(tmp_0, tmp_0); in av1_highbd_wiener_convolve_add_src_ssse3() 120 const __m128i tmp_0 = _mm_unpacklo_epi32(coeffs_y, coeffs_y); in av1_highbd_wiener_convolve_add_src_ssse3() local 125 const __m128i coeff_01 = _mm_unpacklo_epi64(tmp_0, tmp_0); in av1_highbd_wiener_convolve_add_src_ssse3() 127 const __m128i coeff_23 = _mm_unpackhi_epi64(tmp_0, tmp_0); in av1_highbd_wiener_convolve_add_src_ssse3()
|
D | jnt_convolve_ssse3.c | 62 const __m128i tmp_0 = _mm_unpacklo_epi32(coeffs_x, coeffs_x); in av1_dist_wtd_convolve_2d_ssse3() local 67 const __m128i coeff_01 = _mm_unpacklo_epi64(tmp_0, tmp_0); in av1_dist_wtd_convolve_2d_ssse3() 69 const __m128i coeff_23 = _mm_unpackhi_epi64(tmp_0, tmp_0); in av1_dist_wtd_convolve_2d_ssse3() 130 const __m128i tmp_0 = _mm_unpacklo_epi32(coeffs_y, coeffs_y); in av1_dist_wtd_convolve_2d_ssse3() local 135 const __m128i coeff_01 = _mm_unpacklo_epi64(tmp_0, tmp_0); in av1_dist_wtd_convolve_2d_ssse3() 137 const __m128i coeff_23 = _mm_unpackhi_epi64(tmp_0, tmp_0); in av1_dist_wtd_convolve_2d_ssse3()
|
D | highbd_convolve_2d_sse4.c | 215 const __m128i tmp_0 = _mm_unpacklo_epi32(coeffs_x, coeffs_x); in av1_highbd_dist_wtd_convolve_2d_sse4_1() local 220 const __m128i coeff_01 = _mm_unpacklo_epi64(tmp_0, tmp_0); in av1_highbd_dist_wtd_convolve_2d_sse4_1() 222 const __m128i coeff_23 = _mm_unpackhi_epi64(tmp_0, tmp_0); in av1_highbd_dist_wtd_convolve_2d_sse4_1() 282 const __m128i tmp_0 = _mm_unpacklo_epi32(coeffs_y, coeffs_y); in av1_highbd_dist_wtd_convolve_2d_sse4_1() local 287 const __m128i coeff_01 = _mm_unpacklo_epi64(tmp_0, tmp_0); in av1_highbd_dist_wtd_convolve_2d_sse4_1() 289 const __m128i coeff_23 = _mm_unpackhi_epi64(tmp_0, tmp_0); in av1_highbd_dist_wtd_convolve_2d_sse4_1()
|
D | convolve_2d_sse2.c | 52 const __m128i tmp_0 = _mm_unpacklo_epi32(coeffs_x, coeffs_x); in av1_convolve_2d_sr_sse2() local 57 const __m128i coeff_01 = _mm_unpacklo_epi64(tmp_0, tmp_0); in av1_convolve_2d_sr_sse2() 59 const __m128i coeff_23 = _mm_unpackhi_epi64(tmp_0, tmp_0); in av1_convolve_2d_sr_sse2() 118 const __m128i tmp_0 = _mm_unpacklo_epi32(coeffs_y, coeffs_y); in av1_convolve_2d_sr_sse2() local 123 const __m128i coeff_01 = _mm_unpacklo_epi64(tmp_0, tmp_0); in av1_convolve_2d_sr_sse2() 125 const __m128i coeff_23 = _mm_unpackhi_epi64(tmp_0, tmp_0); in av1_convolve_2d_sr_sse2()
|
D | warp_plane_sse4.c | 286 const __m128i tmp_0 = _mm_loadl_epi64( in prepare_horizontal_filter_coeff() local 304 const __m128i tmp_8 = _mm_unpacklo_epi16(tmp_0, tmp_2); in prepare_horizontal_filter_coeff() 334 const __m128i tmp_0 = in prepare_horizontal_filter_coeff_alpha0() local 339 _mm_shuffle_epi8(tmp_0, _mm_load_si128((__m128i *)shuffle_alpha0_mask01)); in prepare_horizontal_filter_coeff_alpha0() 342 _mm_shuffle_epi8(tmp_0, _mm_load_si128((__m128i *)shuffle_alpha0_mask23)); in prepare_horizontal_filter_coeff_alpha0() 345 _mm_shuffle_epi8(tmp_0, _mm_load_si128((__m128i *)shuffle_alpha0_mask45)); in prepare_horizontal_filter_coeff_alpha0() 348 _mm_shuffle_epi8(tmp_0, _mm_load_si128((__m128i *)shuffle_alpha0_mask67)); in prepare_horizontal_filter_coeff_alpha0() 472 const __m128i tmp_0 = in prepare_vertical_filter_coeffs() local 485 const __m128i tmp_8 = _mm_unpacklo_epi32(tmp_0, tmp_2); in prepare_vertical_filter_coeffs() 487 const __m128i tmp_12 = _mm_unpackhi_epi32(tmp_0, tmp_2); in prepare_vertical_filter_coeffs() [all …]
|
D | highbd_warp_plane_sse4.c | 38 const __m128i tmp_0 = in highbd_prepare_horizontal_filter_coeff() local 52 const __m128i tmp_8 = _mm_unpacklo_epi32(tmp_0, tmp_2); in highbd_prepare_horizontal_filter_coeff() 56 const __m128i tmp_12 = _mm_unpackhi_epi32(tmp_0, tmp_2); in highbd_prepare_horizontal_filter_coeff() 97 const __m128i tmp_0 = _mm_loadu_si128( in highbd_prepare_horizontal_filter_coeff_alpha0() local 101 tmp_0, _mm_loadu_si128((__m128i *)highbd_shuffle_alpha0_mask0)); in highbd_prepare_horizontal_filter_coeff_alpha0() 103 tmp_0, _mm_loadu_si128((__m128i *)highbd_shuffle_alpha0_mask1)); in highbd_prepare_horizontal_filter_coeff_alpha0() 105 tmp_0, _mm_loadu_si128((__m128i *)highbd_shuffle_alpha0_mask2)); in highbd_prepare_horizontal_filter_coeff_alpha0() 107 tmp_0, _mm_loadu_si128((__m128i *)highbd_shuffle_alpha0_mask3)); in highbd_prepare_horizontal_filter_coeff_alpha0() 464 const __m128i tmp_0 = _mm_loadu_si128( in av1_highbd_warp_affine_sse4_1() local 477 const __m128i tmp_8 = _mm_unpacklo_epi32(tmp_0, tmp_2); in av1_highbd_warp_affine_sse4_1() [all …]
|
D | jnt_convolve_sse2.c | 431 const __m128i tmp_0 = _mm_unpacklo_epi32(coeffs_x, coeffs_x); in av1_dist_wtd_convolve_2d_sse2() local 436 const __m128i coeff_01 = _mm_unpacklo_epi64(tmp_0, tmp_0); in av1_dist_wtd_convolve_2d_sse2() 438 const __m128i coeff_23 = _mm_unpackhi_epi64(tmp_0, tmp_0); in av1_dist_wtd_convolve_2d_sse2() 514 const __m128i tmp_0 = _mm_unpacklo_epi32(coeffs_y, coeffs_y); in av1_dist_wtd_convolve_2d_sse2() local 519 const __m128i coeff_01 = _mm_unpacklo_epi64(tmp_0, tmp_0); in av1_dist_wtd_convolve_2d_sse2() 521 const __m128i coeff_23 = _mm_unpackhi_epi64(tmp_0, tmp_0); in av1_dist_wtd_convolve_2d_sse2()
|
D | convolve_sse2.c | 28 const __m128i tmp_0 = _mm_unpacklo_epi32(coeffs_y, coeffs_y); in prepare_coeffs() local 32 coeffs[0] = _mm_unpacklo_epi64(tmp_0, tmp_0); // coeffs 0 1 0 1 0 1 0 1 in prepare_coeffs() 33 coeffs[1] = _mm_unpackhi_epi64(tmp_0, tmp_0); // coeffs 2 3 2 3 2 3 2 3 in prepare_coeffs()
|
D | warp_plane_avx2.c | 102 __m128i tmp_0 = _mm_loadl_epi64( in prepare_horizontal_filter_coeff_avx2() local 128 __m256i tmp0_256 = _mm256_castsi128_si256(tmp_0); in prepare_horizontal_filter_coeff_avx2() 196 __m128i tmp_0 = _mm_loadl_epi64( in prepare_horizontal_filter_coeff_beta0_avx2() local 213 tmp_0 = _mm_unpacklo_epi16(tmp_0, tmp_2); in prepare_horizontal_filter_coeff_beta0_avx2() 218 const __m256i tmp_12 = _mm256_broadcastsi128_si256(tmp_0); in prepare_horizontal_filter_coeff_beta0_avx2() 236 const __m128i tmp_0 = in prepare_horizontal_filter_coeff_alpha0_avx2() local 242 _mm256_inserti128_si256(_mm256_castsi128_si256(tmp_0), tmp_1, 0x1); in prepare_horizontal_filter_coeff_alpha0_avx2() 266 const __m128i tmp_0 = _mm_loadl_epi64( in prepare_horizontal_filter_coeff() local 283 const __m128i tmp_8 = _mm_unpacklo_epi16(tmp_0, tmp_2); in prepare_horizontal_filter_coeff()
|
/external/llvm-project/polly/test/ScopInfo/ |
D | ranged_parameter_wrap_2.ll | 8 ; CHECK: [tmp_0, tmp_1, tmp_2, tmp_3, tmp_4, tmp_5, tmp_6, tmp_7, tmp_8, 10 ; CHECK: -2147483648 <= tmp_0 <= 2147483647 and 26 ; CHECK: ((tmp_0 >= 256 and tmp_1 >= 256 and tmp_2 >= 256) or 27 ; CHECK: (tmp_0 >= 256 and tmp_1 >= 256 and tmp_2 < 0) or 28 ; CHECK: (tmp_0 >= 256 and tmp_1 < 0 and tmp_2 >= 256) or 29 ; CHECK: (tmp_0 >= 256 and tmp_1 < 0 and tmp_2 < 0) or 30 ; CHECK: (tmp_0 < 0 and tmp_1 >= 256 and tmp_2 >= 256) or 31 ; CHECK: (tmp_0 < 0 and tmp_1 >= 256 and tmp_2 < 0) or 32 ; CHECK: (tmp_0 < 0 and tmp_1 < 0 and tmp_2 >= 256) or 33 ; CHECK: (tmp_0 < 0 and tmp_1 < 0 and tmp_2 < 0)) } [all …]
|
/external/libvpx/libvpx/vpx_dsp/x86/ |
D | post_proc_sse2.c | 38 __m128i tmp_0, tmp_1; in vpx_mbpost_proc_down_sse2() local 52 tmp_0 = _mm_mullo_epi16(sum, s); in vpx_mbpost_proc_down_sse2() 55 sumsq_0 = _mm_unpacklo_epi16(tmp_0, tmp_1); in vpx_mbpost_proc_down_sse2() 56 sumsq_1 = _mm_unpackhi_epi16(tmp_0, tmp_1); in vpx_mbpost_proc_down_sse2()
|
D | vpx_subpixel_8t_intrin_avx2.c | 358 __m256i tmp_0, tmp_1; in vpx_filter_block1d16_h4_avx2() local 386 tmp_0 = _mm256_maddubs_epi16(src_reg_shift_0, kernel_reg_23); in vpx_filter_block1d16_h4_avx2() 388 dst_first = _mm256_adds_epi16(tmp_0, tmp_1); in vpx_filter_block1d16_h4_avx2() 397 tmp_0 = _mm256_maddubs_epi16(src_reg_shift_0, kernel_reg_23); in vpx_filter_block1d16_h4_avx2() 399 dst_second = _mm256_adds_epi16(tmp_0, tmp_1); in vpx_filter_block1d16_h4_avx2() 423 tmp_0 = _mm256_maddubs_epi16(src_reg_shift_0, kernel_reg_23); in vpx_filter_block1d16_h4_avx2() 425 dst_first = _mm256_adds_epi16(tmp_0, tmp_1); in vpx_filter_block1d16_h4_avx2() 565 __m256i tmp_0, tmp_1; in vpx_filter_block1d8_h4_avx2() local 591 tmp_0 = _mm256_maddubs_epi16(src_reg_shift_0, kernel_reg_23); in vpx_filter_block1d8_h4_avx2() 593 dst_reg = _mm256_adds_epi16(tmp_0, tmp_1); in vpx_filter_block1d8_h4_avx2() [all …]
|
D | vpx_subpixel_4t_intrin_sse2.c | 440 __m128i tmp_0, tmp_1; in vpx_filter_block1d4_h4_sse2() local 472 tmp_0 = _mm_unpacklo_epi32(src_reg, src_reg_shift_1); in vpx_filter_block1d4_h4_sse2() 476 tmp_0 = _mm_madd_epi16(tmp_0, kernel_reg_23); in vpx_filter_block1d4_h4_sse2() 480 dst_first = _mm_add_epi32(tmp_0, tmp_1); in vpx_filter_block1d4_h4_sse2() 784 __m128i tmp_0, tmp_1; in vpx_highbd_filter_block1d8_h4_sse2() local 809 tmp_0 = _mm_srli_si128(src_reg, 4); in vpx_highbd_filter_block1d8_h4_sse2() 811 src_reg_shift_2 = _mm_unpacklo_epi64(tmp_0, tmp_1); in vpx_highbd_filter_block1d8_h4_sse2() 816 tmp_0 = _mm_srli_si128(src_reg, 2); in vpx_highbd_filter_block1d8_h4_sse2() 818 src_reg_shift_1 = _mm_unpacklo_epi64(tmp_0, tmp_1); in vpx_highbd_filter_block1d8_h4_sse2() 820 tmp_0 = _mm_srli_si128(src_reg, 6); in vpx_highbd_filter_block1d8_h4_sse2() [all …]
|
D | convolve_avx2.h | 155 const __m256i tmp_0 = _mm256_madd_epi16(*src_0, *ker_0); in mm256_madd_add_epi32() local 157 return _mm256_add_epi32(tmp_0, tmp_1); in mm256_madd_add_epi32()
|
D | vpx_subpixel_8t_intrin_ssse3.c | 226 __m128i tmp_0, tmp_1; in vpx_filter_block1d16_h4_ssse3() local 249 tmp_0 = _mm_maddubs_epi16(src_reg_shift_0, kernel_reg_23); in vpx_filter_block1d16_h4_ssse3() 251 dst_first = _mm_adds_epi16(tmp_0, tmp_1); in vpx_filter_block1d16_h4_ssse3() 260 tmp_0 = _mm_maddubs_epi16(src_reg_shift_0, kernel_reg_23); in vpx_filter_block1d16_h4_ssse3() 262 dst_second = _mm_adds_epi16(tmp_0, tmp_1); in vpx_filter_block1d16_h4_ssse3() 411 __m128i tmp_0, tmp_1; in vpx_filter_block1d8_h4_ssse3() local 434 tmp_0 = _mm_maddubs_epi16(src_reg_shift_0, kernel_reg_23); in vpx_filter_block1d8_h4_ssse3() 436 dst_first = _mm_adds_epi16(tmp_0, tmp_1); in vpx_filter_block1d8_h4_ssse3()
|
/external/libaom/libaom/av1/encoder/x86/ |
D | corner_match_sse4.c | 83 __m128i tmp_0 = _mm_unpacklo_epi32(sum1_vec, sum2_vec); in av1_compute_cross_correlation_sse4_1() local 88 __m128i tmp_4 = _mm_unpacklo_epi64(tmp_0, tmp_2); in av1_compute_cross_correlation_sse4_1() 89 __m128i tmp_5 = _mm_unpackhi_epi64(tmp_0, tmp_2); in av1_compute_cross_correlation_sse4_1()
|
/external/webp/src/dsp/ |
D | enc_sse41.c | 76 __m128i tmp_0, tmp_1, tmp_2, tmp_3; in TTransform_SSE41() local 98 tmp_0 = _mm_cvtepu8_epi16(inAB_0); in TTransform_SSE41() 112 const __m128i a0 = _mm_add_epi16(tmp_0, tmp_2); in TTransform_SSE41() 115 const __m128i a3 = _mm_sub_epi16(tmp_0, tmp_2); in TTransform_SSE41() 126 VP8Transpose_2_4x4_16b(&b0, &b1, &b2, &b3, &tmp_0, &tmp_1, &tmp_2, &tmp_3); in TTransform_SSE41() 136 const __m128i a0 = _mm_add_epi16(tmp_0, tmp_2); in TTransform_SSE41() 139 const __m128i a3 = _mm_sub_epi16(tmp_0, tmp_2); in TTransform_SSE41()
|
D | enc_sse2.c | 1100 __m128i tmp_0, tmp_1, tmp_2, tmp_3; in TTransform_SSE2() local 1119 tmp_0 = _mm_unpacklo_epi8(inAB_0, zero); in TTransform_SSE2() 1133 const __m128i a0 = _mm_add_epi16(tmp_0, tmp_2); in TTransform_SSE2() 1136 const __m128i a3 = _mm_sub_epi16(tmp_0, tmp_2); in TTransform_SSE2() 1147 VP8Transpose_2_4x4_16b(&b0, &b1, &b2, &b3, &tmp_0, &tmp_1, &tmp_2, &tmp_3); in TTransform_SSE2() 1157 const __m128i a0 = _mm_add_epi16(tmp_0, tmp_2); in TTransform_SSE2() 1160 const __m128i a3 = _mm_sub_epi16(tmp_0, tmp_2); in TTransform_SSE2()
|
/external/libaom/libaom/av1/common/arm/ |
D | warp_plane_neon.c | 237 uint8x16_t tmp_0 = vandq_u8(src_1, mask); in horizontal_filter_neon() local 242 tmp_2 = vextq_u8(tmp_0, tmp_0, 1); in horizontal_filter_neon() 245 src_1 = vaddq_u8(tmp_0, tmp_2); in horizontal_filter_neon()
|
/external/tensorflow/tensorflow/lite/kernels/internal/optimized/ |
D | neon_tensor_utils.cc | 2665 const float32x4_t tmp_0 = vsubq_f32(input_f32x4_0, mean_f32x4); in NeonMeanStddevNormalization() local 2669 const float32x4_t output_f32x4_0 = vmulq_n_f32(tmp_0, stddev_inv); in NeonMeanStddevNormalization()
|