/external/boringssl/src/crypto/poly1305/ |
D | poly1305_vec.c | 220 st->H[1] = _mm_and_si128(MMASK, _mm_srli_epi64(T5, 26)); in poly1305_first_block() 221 T5 = _mm_or_si128(_mm_srli_epi64(T5, 52), _mm_slli_epi64(T6, 12)); in poly1305_first_block() 223 st->H[3] = _mm_and_si128(MMASK, _mm_srli_epi64(T5, 26)); in poly1305_first_block() 224 st->H[4] = _mm_or_si128(_mm_srli_epi64(T6, 40), HIBIT); in poly1305_first_block() 300 M1 = _mm_and_si128(MMASK, _mm_srli_epi64(T5, 26)); in poly1305_blocks() 301 T5 = _mm_or_si128(_mm_srli_epi64(T5, 52), _mm_slli_epi64(T6, 12)); in poly1305_blocks() 303 M3 = _mm_and_si128(MMASK, _mm_srli_epi64(T5, 26)); in poly1305_blocks() 304 M4 = _mm_or_si128(_mm_srli_epi64(T6, 40), HIBIT); in poly1305_blocks() 364 M1 = _mm_and_si128(MMASK, _mm_srli_epi64(T5, 26)); in poly1305_blocks() 365 T5 = _mm_or_si128(_mm_srli_epi64(T5, 52), _mm_slli_epi64(T6, 12)); in poly1305_blocks() [all …]
|
/external/webp/src/dsp/ |
D | rescaler_sse2.c | 152 const __m128i D0 = _mm_srli_epi64(frac, 32); in RescalerImportRowShrink_SSE2() 179 const __m128i A2 = _mm_srli_epi64(A0, 32); in LoadDispatchAndMult_SSE2() 180 const __m128i A3 = _mm_srli_epi64(A1, 32); in LoadDispatchAndMult_SSE2() 210 const __m128i D0 = _mm_srli_epi64(C0, WEBP_RESCALER_RFIX); in ProcessRow_SSE2() 211 const __m128i D1 = _mm_srli_epi64(C1, WEBP_RESCALER_RFIX); in ProcessRow_SSE2() 269 const __m128i E0 = _mm_srli_epi64(D0, WEBP_RESCALER_RFIX); in RescalerExportRowExpand_SSE2() 270 const __m128i E1 = _mm_srli_epi64(D1, WEBP_RESCALER_RFIX); in RescalerExportRowExpand_SSE2() 271 const __m128i E2 = _mm_srli_epi64(D2, WEBP_RESCALER_RFIX); in RescalerExportRowExpand_SSE2() 272 const __m128i E3 = _mm_srli_epi64(D3, WEBP_RESCALER_RFIX); in RescalerExportRowExpand_SSE2() 305 const __m128i D0 = _mm_srli_epi64(B0, WEBP_RESCALER_RFIX); // = frac in RescalerExportRowShrink_SSE2() [all …]
|
D | lossless_sse2.c | 643 const __m128i b0h = _mm_srli_epi64(a0h, 8); // 000b|gr00|000b|gr00 in ConvertBGRAToBGR_SSE2() 644 const __m128i b4h = _mm_srli_epi64(a4h, 8); // 000b|gr00|000b|gr00 in ConvertBGRAToBGR_SSE2()
|
D | filters_sse2.c | 239 last = _mm_srli_epi64(A7, 56); in HorizontalUnfilter_SSE2()
|
/external/libopus/silk/fixed/x86/ |
D | prefilter_FIX_sse.c | 97 xmm_product1 = _mm_srli_epi64( xmm_product1, 16 ); /* >> 16, zero extending works */ in silk_warped_LPC_analysis_filter_FIX_sse4_1() 98 xmm_product2 = _mm_srli_epi64( xmm_product2, 16 ); in silk_warped_LPC_analysis_filter_FIX_sse4_1() 103 xmm_tempa = _mm_srli_epi64( xmm_tempa, 16 ); in silk_warped_LPC_analysis_filter_FIX_sse4_1() 104 xmm_tempb = _mm_srli_epi64( xmm_tempb, 16 ); in silk_warped_LPC_analysis_filter_FIX_sse4_1()
|
D | burg_modified_FIX_sse4_1.c | 239 xmm_x_ptr_n_k_x2x0 = _mm_srli_epi64( xmm_x_ptr_n_k_x2x0, 16 ); in silk_burg_modified_sse4_1() 241 xmm_x_ptr_sub_x2x0 = _mm_srli_epi64( xmm_x_ptr_sub_x2x0, 16 ); in silk_burg_modified_sse4_1()
|
/external/pdfium/third_party/libopenjpeg20/ |
D | mct.c | 250 lo = _mm_srli_epi64(lo, 13); in opj_mct_encode_real() 260 lo = _mm_srli_epi64(lo, 13); in opj_mct_encode_real() 270 lo = _mm_srli_epi64(lo, 13); in opj_mct_encode_real() 285 lo = _mm_srli_epi64(lo, 13); in opj_mct_encode_real() 295 lo = _mm_srli_epi64(lo, 13); in opj_mct_encode_real() 305 lo = _mm_srli_epi64(lo, 13); in opj_mct_encode_real() 320 lo = _mm_srli_epi64(lo, 13); in opj_mct_encode_real() 330 lo = _mm_srli_epi64(lo, 13); in opj_mct_encode_real() 340 lo = _mm_srli_epi64(lo, 13); in opj_mct_encode_real()
|
/external/libaom/libaom/aom_dsp/x86/ |
D | sum_squares_sse2.c | 51 _mm_add_epi32(v_sum_0123_d, _mm_srli_epi64(v_sum_0123_d, 32)); in aom_sum_squares_2d_i16_4x4_sse2() 67 __m128i v_acc_64 = _mm_add_epi64(_mm_srli_epi64(v_acc_q, 32), in aom_sum_squares_2d_i16_4xn_sse2() 113 v_acc_q = _mm_add_epi64(v_acc_q, _mm_srli_epi64(v_acc_d, 32)); in aom_sum_squares_2d_i16_nxn_sse2() 183 v_acc1_q = _mm_add_epi64(v_acc1_q, _mm_srli_epi64(v_sum_d, 32)); in aom_sum_squares_i16_64n_sse2()
|
D | avg_intrin_sse2.c | 87 maxabsdiff = _mm_max_epi16(maxabsdiff, _mm_srli_epi64(maxabsdiff, 32)); in aom_minmax_8x8_sse2() 88 maxabsdiff = _mm_max_epi16(maxabsdiff, _mm_srli_epi64(maxabsdiff, 16)); in aom_minmax_8x8_sse2() 92 minabsdiff = _mm_min_epi16(minabsdiff, _mm_srli_epi64(minabsdiff, 32)); in aom_minmax_8x8_sse2() 93 minabsdiff = _mm_min_epi16(minabsdiff, _mm_srli_epi64(minabsdiff, 16)); in aom_minmax_8x8_sse2() 118 s0 = _mm_adds_epu16(s0, _mm_srli_epi64(s0, 32)); in aom_avg_8x8_sse2() 119 s0 = _mm_adds_epu16(s0, _mm_srli_epi64(s0, 16)); in aom_avg_8x8_sse2() 137 s0 = _mm_adds_epu16(s0, _mm_srli_epi64(s0, 16)); in aom_avg_4x4_sse2() 403 hi = _mm_srli_epi64(accum, 32); in aom_satd_sse2()
|
D | fwd_txfm_sse2.h | 22 a = _mm_srli_epi64(a, 32); in k_madd_epi32() 23 b = _mm_srli_epi64(b, 32); in k_madd_epi32()
|
/external/libvpx/libvpx/vpx_dsp/x86/ |
D | avg_intrin_sse2.c | 86 maxabsdiff = _mm_max_epi16(maxabsdiff, _mm_srli_epi64(maxabsdiff, 32)); in vpx_minmax_8x8_sse2() 87 maxabsdiff = _mm_max_epi16(maxabsdiff, _mm_srli_epi64(maxabsdiff, 16)); in vpx_minmax_8x8_sse2() 91 minabsdiff = _mm_min_epi16(minabsdiff, _mm_srli_epi64(minabsdiff, 32)); in vpx_minmax_8x8_sse2() 92 minabsdiff = _mm_min_epi16(minabsdiff, _mm_srli_epi64(minabsdiff, 16)); in vpx_minmax_8x8_sse2() 117 s0 = _mm_adds_epu16(s0, _mm_srli_epi64(s0, 32)); in vpx_avg_8x8_sse2() 118 s0 = _mm_adds_epu16(s0, _mm_srli_epi64(s0, 16)); in vpx_avg_8x8_sse2() 136 s0 = _mm_adds_epu16(s0, _mm_srli_epi64(s0, 16)); in vpx_avg_4x4_sse2() 460 hi = _mm_srli_epi64(accum, 32); in vpx_satd_sse2() 564 v0 = _mm_srli_epi64(sum, 32); in vpx_vector_var_sse2() 571 v1 = _mm_srli_epi64(sse, 32); in vpx_vector_var_sse2()
|
D | sum_squares_sse2.c | 30 ss = _mm_add_epi32(ss, _mm_srli_epi64(ss, 32)); in vpx_sum_squares_2d_i16_sse2() 87 v_acc_q = _mm_add_epi64(v_acc_q, _mm_srli_epi64(v_acc_d, 32)); in vpx_sum_squares_2d_i16_sse2()
|
D | fwd_txfm_sse2.h | 24 a = _mm_srli_epi64(a, 32); in k_madd_epi32() 25 b = _mm_srli_epi64(b, 32); in k_madd_epi32()
|
/external/libaom/libaom/av1/encoder/x86/ |
D | av1_highbd_quantize_sse4.c | 38 qcoeff[0] = _mm_srli_epi64(qcoeff[0], shift); in quantize_coeff_phase1() 40 dquan[0] = _mm_srli_epi64(dquan[0], scale); in quantize_coeff_phase1() 55 qcoeff[1] = _mm_srli_epi64(qcoeff[1], shift); in quantize_coeff_phase2() 57 dquan[1] = _mm_srli_epi64(dquan[1], scale); in quantize_coeff_phase2()
|
D | corner_match_avx2.c | 72 low_sumsq = _mm_add_epi32(low_sumsq, _mm_srli_epi64(low_sumsq, 32)); in compute_cross_correlation_avx2()
|
D | wedge_utils_sse2.c | 75 _mm_srli_epi64(v_sq0_d, 32)); in av1_wedge_sse_from_residuals_sse2() 77 _mm_srli_epi64(v_sq1_d, 32)); in av1_wedge_sse_from_residuals_sse2()
|
/external/libopus/silk/x86/ |
D | NSQ_del_dec_sse4_1.c | 444 tmpa = _mm_srli_epi64( tmpa, 16 ); in silk_noise_shape_quantizer_del_dec_sse4_1() 450 psLPC_Q14_tmp = _mm_srli_epi64( psLPC_Q14_tmp, 16 ); in silk_noise_shape_quantizer_del_dec_sse4_1() 457 tmpa = _mm_srli_epi64( tmpa, 16 ); in silk_noise_shape_quantizer_del_dec_sse4_1() 463 psLPC_Q14_tmp = _mm_srli_epi64( psLPC_Q14_tmp, 16 ); in silk_noise_shape_quantizer_del_dec_sse4_1() 472 tmpa = _mm_srli_epi64( tmpa, 16 ); in silk_noise_shape_quantizer_del_dec_sse4_1() 478 psLPC_Q14_tmp = _mm_srli_epi64( psLPC_Q14_tmp, 16 ); in silk_noise_shape_quantizer_del_dec_sse4_1() 485 tmpa = _mm_srli_epi64( tmpa, 16 ); in silk_noise_shape_quantizer_del_dec_sse4_1() 491 psLPC_Q14_tmp = _mm_srli_epi64( psLPC_Q14_tmp, 16 ); in silk_noise_shape_quantizer_del_dec_sse4_1() 775 xmm_x_Q3_x2x0 = _mm_srli_epi64( xmm_x_Q3_x2x0, 16 ); in silk_nsq_del_dec_scale_states_sse4_1() 820 xmm_sLTP_shp_Q14_x2x0 = _mm_srli_epi64( xmm_sLTP_shp_Q14_x2x0, 16 ); in silk_nsq_del_dec_scale_states_sse4_1()
|
D | NSQ_sse4_1.c | 571 xmm_xq_Q14_3210 = _mm_srli_epi64( xmm_xq_Q14_3210, 16 ); in silk_noise_shape_quantizer_10_16_sse4_1() 573 xmm_xq_Q14_7654 = _mm_srli_epi64( xmm_xq_Q14_7654, 16 ); in silk_noise_shape_quantizer_10_16_sse4_1() 646 xmm_x_Q3_x2x0 = _mm_srli_epi64( xmm_x_Q3_x2x0, 16 ); in silk_nsq_scale_states_sse4_1() 690 xmm_sLTP_shp_Q14_x2x0 = _mm_srli_epi64( xmm_sLTP_shp_Q14_x2x0, 16 ); in silk_nsq_scale_states_sse4_1()
|
/external/python/cpython3/Modules/_blake2/impl/ |
D | blake2b-round.h | 35 : (-(c) == 63) ? _mm_xor_si128(_mm_srli_epi64((x), -(c)), _mm_add_epi64((x), (x))) \ 36 : _mm_xor_si128(_mm_srli_epi64((x), -(c)), _mm_slli_epi64((x), 64-(-(c)))) 38 #define _mm_roti_epi64(r, c) _mm_xor_si128(_mm_srli_epi64( (r), -(c) ),_mm_slli_epi64( (r), 64-(-(c…
|
/external/mesa3d/src/gallium/auxiliary/util/ |
D | u_sse.h | 219 __m128i a4 = _mm_srli_epi64(a, 32); /* shift by one dword */ in mm_mullo_epi32() 220 __m128i b4 = _mm_srli_epi64(b, 32); /* shift by one dword */ in mm_mullo_epi32()
|
/external/libhevc/common/x86/ |
D | ihevc_deblk_ssse3_intr.c | 264 temp_pq_str0_16x8b = _mm_srli_epi64(src_row0_8x16b, 16); in ihevc_deblk_luma_vert_ssse3() 265 temp_pq_str1_16x8b = _mm_srli_epi64(src_row1_8x16b, 16); in ihevc_deblk_luma_vert_ssse3() 277 temp_pq1_str0_16x8b = _mm_srli_epi64(src_row0_8x16b, 8); in ihevc_deblk_luma_vert_ssse3() 278 temp_pq1_str1_16x8b = _mm_srli_epi64(src_row1_8x16b, 8); in ihevc_deblk_luma_vert_ssse3() 401 tmp_delta2_8x16b = _mm_srli_epi64(src_row0_8x16b, 16); in ihevc_deblk_luma_vert_ssse3() 402 tmp_delta3_8x16b = _mm_srli_epi64(src_row2_8x16b, 16); in ihevc_deblk_luma_vert_ssse3()
|
/external/epid-sdk/ext/ipp/sources/ippcp/ |
D | pcpsms4ecby8cn.h | 81 __m128i T1 = _mm_and_si128(_mm_srli_epi64(x, 4), M128(lowBits4)); in affine()
|
/external/neven/Embedded/common/src/b_BasicEm/ |
D | MathSSE2.c | 95 m_XMM0 = _mm_srli_epi64( m_XMM0, 32 ); in bbs_dotProduct_64SSE2()
|
/external/boringssl/src/crypto/hrss/ |
D | hrss.c | 152 vec_t next_carry_s = _mm_srli_epi64(a_s[i], 63); in poly3_vec_lshift1() 158 vec_t next_carry_a = _mm_srli_epi64(a_a[i], 63); in poly3_vec_lshift1() 174 a_s[i] = _mm_srli_epi64(a_s[i], 1); in poly3_vec_rshift1() 180 a_a[i] = _mm_srli_epi64(a_a[i], 1); in poly3_vec_rshift1()
|
/external/eigen/Eigen/src/Core/arch/AVX/ |
D | MathFunctions.h | 346 __m128i hi = _mm_slli_epi64(_mm_srli_epi64(emm0, 32), 52);
|