Home
last modified time | relevance | path

Searched refs:_mm_load_si128 (Results 1 – 25 of 35) sorted by relevance

12

/external/libhevc/common/x86/
Dihevc_itrans_recon_32x32_ssse3_intr.c260 m_temp_reg_70 = _mm_load_si128((__m128i *)pi2_tmp_src); in ihevc_itrans_recon_32x32_ssse3()
262 m_temp_reg_71 = _mm_load_si128((__m128i *)pi2_tmp_src); in ihevc_itrans_recon_32x32_ssse3()
264 m_temp_reg_72 = _mm_load_si128((__m128i *)pi2_tmp_src); in ihevc_itrans_recon_32x32_ssse3()
266 m_temp_reg_73 = _mm_load_si128((__m128i *)pi2_tmp_src); in ihevc_itrans_recon_32x32_ssse3()
268 m_temp_reg_74 = _mm_load_si128((__m128i *)pi2_tmp_src); in ihevc_itrans_recon_32x32_ssse3()
270 m_temp_reg_75 = _mm_load_si128((__m128i *)pi2_tmp_src); in ihevc_itrans_recon_32x32_ssse3()
272 m_temp_reg_76 = _mm_load_si128((__m128i *)pi2_tmp_src); in ihevc_itrans_recon_32x32_ssse3()
274 m_temp_reg_77 = _mm_load_si128((__m128i *)pi2_tmp_src); in ihevc_itrans_recon_32x32_ssse3()
277 m_temp_reg_80 = _mm_load_si128((__m128i *)pi2_tmp_src); in ihevc_itrans_recon_32x32_ssse3()
279 m_temp_reg_81 = _mm_load_si128((__m128i *)pi2_tmp_src); in ihevc_itrans_recon_32x32_ssse3()
[all …]
Dihevc_itrans_recon_16x16_ssse3_intr.c210 m_temp_reg_70 = _mm_load_si128((__m128i *)pi2_tmp_src); in ihevc_itrans_recon_16x16_ssse3()
212 m_temp_reg_71 = _mm_load_si128((__m128i *)pi2_tmp_src); in ihevc_itrans_recon_16x16_ssse3()
214 m_temp_reg_72 = _mm_load_si128((__m128i *)pi2_tmp_src); in ihevc_itrans_recon_16x16_ssse3()
216 m_temp_reg_73 = _mm_load_si128((__m128i *)pi2_tmp_src); in ihevc_itrans_recon_16x16_ssse3()
218 m_temp_reg_74 = _mm_load_si128((__m128i *)pi2_tmp_src); in ihevc_itrans_recon_16x16_ssse3()
220 m_temp_reg_75 = _mm_load_si128((__m128i *)pi2_tmp_src); in ihevc_itrans_recon_16x16_ssse3()
222 m_temp_reg_76 = _mm_load_si128((__m128i *)pi2_tmp_src); in ihevc_itrans_recon_16x16_ssse3()
224 m_temp_reg_77 = _mm_load_si128((__m128i *)pi2_tmp_src); in ihevc_itrans_recon_16x16_ssse3()
249 m_coeff1 = _mm_load_si128((__m128i *)&g_ai2_ihevc_trans_16_even[2][0]); //89 75 in ihevc_itrans_recon_16x16_ssse3()
295 m_coeff3 = _mm_load_si128((__m128i *)&g_ai2_ihevc_trans_16_even[3][0]); //75 -18 in ihevc_itrans_recon_16x16_ssse3()
[all …]
Dihevc_itrans_recon_ssse3_intr.c1035 m_temp_reg_70 = _mm_load_si128((__m128i *)pi2_src); in ihevc_itrans_recon_8x8_ssse3()
1037 m_temp_reg_71 = _mm_load_si128((__m128i *)pi2_src); in ihevc_itrans_recon_8x8_ssse3()
1039 m_temp_reg_72 = _mm_load_si128((__m128i *)pi2_src); in ihevc_itrans_recon_8x8_ssse3()
1041 m_temp_reg_73 = _mm_load_si128((__m128i *)pi2_src); in ihevc_itrans_recon_8x8_ssse3()
1044 m_temp_reg_74 = _mm_load_si128((__m128i *)pi2_src); in ihevc_itrans_recon_8x8_ssse3()
1046 m_temp_reg_75 = _mm_load_si128((__m128i *)pi2_src); in ihevc_itrans_recon_8x8_ssse3()
1048 m_temp_reg_76 = _mm_load_si128((__m128i *)pi2_src); in ihevc_itrans_recon_8x8_ssse3()
1050 m_temp_reg_77 = _mm_load_si128((__m128i *)pi2_src); in ihevc_itrans_recon_8x8_ssse3()
1061 m_coeff2 = _mm_load_si128((__m128i *)&g_ai2_ihevc_trans_intr_even_8[3][0]); in ihevc_itrans_recon_8x8_ssse3()
1062 m_coeff1 = _mm_load_si128((__m128i *)&g_ai2_ihevc_trans_intr_even_8[0][0]); in ihevc_itrans_recon_8x8_ssse3()
[all …]
Dihevc_deblk_ssse3_intr.c147 coef_8x16b = _mm_load_si128((__m128i *)(coef_d)); in ihevc_deblk_luma_vert_ssse3()
148 mask_16x8b = _mm_load_si128((__m128i *)(shuffle_d)); in ihevc_deblk_luma_vert_ssse3()
409 coefdelta_0_8x16b = _mm_load_si128((__m128i *)coef_de1); in ihevc_deblk_luma_vert_ssse3()
431 coefdelta_0_8x16b = _mm_load_si128((__m128i *)coef_dep1); in ihevc_deblk_luma_vert_ssse3()
474 tmp3_const_8x16b = _mm_load_si128((__m128i *)(shuffle1)); in ihevc_deblk_luma_vert_ssse3()
508 tmp0_const_8x16b = _mm_load_si128((__m128i *)shuffle2); in ihevc_deblk_luma_vert_ssse3()
509 tmp1_const_8x16b = _mm_load_si128((__m128i *)shuffle3); in ihevc_deblk_luma_vert_ssse3()
606 coef_8x16b = _mm_load_si128((__m128i *)(coef_d)); in ihevc_deblk_luma_horz_ssse3()
607 mask_16x8b = _mm_load_si128((__m128i *)(shuffle_d)); in ihevc_deblk_luma_horz_ssse3()
843 coefdelta_0_8x16b = _mm_load_si128((__m128i *)coef_de1); in ihevc_deblk_luma_horz_ssse3()
[all …]
Dihevc_32x32_itrans_recon_sse42_intr.c3795 m_temp_reg_30 = _mm_load_si128((__m128i *)pi2_src_scratch); in ihevc_itrans_recon_32x32_sse42()
3797 m_temp_reg_31 = _mm_load_si128((__m128i *)pi2_src_scratch); in ihevc_itrans_recon_32x32_sse42()
3799 m_temp_reg_32 = _mm_load_si128((__m128i *)pi2_src_scratch); in ihevc_itrans_recon_32x32_sse42()
3801 m_temp_reg_33 = _mm_load_si128((__m128i *)pi2_src_scratch); in ihevc_itrans_recon_32x32_sse42()
3803 m_temp_reg_34 = _mm_load_si128((__m128i *)pi2_src_scratch); in ihevc_itrans_recon_32x32_sse42()
3805 m_temp_reg_35 = _mm_load_si128((__m128i *)pi2_src_scratch); in ihevc_itrans_recon_32x32_sse42()
3807 m_temp_reg_36 = _mm_load_si128((__m128i *)pi2_src_scratch); in ihevc_itrans_recon_32x32_sse42()
3809 m_temp_reg_37 = _mm_load_si128((__m128i *)pi2_src_scratch); in ihevc_itrans_recon_32x32_sse42()
3812 m_temp_reg_70 = _mm_load_si128((__m128i *)pi2_src_scratch); in ihevc_itrans_recon_32x32_sse42()
3814 m_temp_reg_71 = _mm_load_si128((__m128i *)pi2_src_scratch); in ihevc_itrans_recon_32x32_sse42()
[all …]
Dihevc_inter_pred_filters_ssse3_intr.c5069 s21_8x16b = _mm_load_si128((__m128i *)(pi2_src + (-1 * src_strd))); in ihevc_inter_pred_chroma_vert_w16inp_ssse3()
5072 s22_8x16b = _mm_load_si128((__m128i *)(pi2_src + (0 * src_strd))); in ihevc_inter_pred_chroma_vert_w16inp_ssse3()
5076 s23_8x16b = _mm_load_si128((__m128i *)(pi2_src + (1 * src_strd))); in ihevc_inter_pred_chroma_vert_w16inp_ssse3()
5079 s24_8x16b = _mm_load_si128((__m128i *)(pi2_src + (2 * src_strd))); in ihevc_inter_pred_chroma_vert_w16inp_ssse3()
5135 s25_8x16b = _mm_load_si128((__m128i *)(pi2_src + (3 * src_strd))); in ihevc_inter_pred_chroma_vert_w16inp_ssse3()
5390 s21_8x16b = _mm_load_si128((__m128i *)(pi2_src + (-1 * src_strd))); in ihevc_inter_pred_chroma_vert_w16inp_w16out_ssse3()
5393 s22_8x16b = _mm_load_si128((__m128i *)(pi2_src + (0 * src_strd))); in ihevc_inter_pred_chroma_vert_w16inp_w16out_ssse3()
5397 s23_8x16b = _mm_load_si128((__m128i *)(pi2_src + (1 * src_strd))); in ihevc_inter_pred_chroma_vert_w16inp_w16out_ssse3()
5400 s24_8x16b = _mm_load_si128((__m128i *)(pi2_src + (2 * src_strd))); in ihevc_inter_pred_chroma_vert_w16inp_w16out_ssse3()
Dihevc_sao_ssse3_intr.c170 band_table0_8x16b = _mm_load_si128((__m128i *)(gu2_table_band_idx)); in ihevc_sao_band_offset_luma_ssse3()
171 band_table1_8x16b = _mm_load_si128((__m128i *)(gu2_table_band_idx + 8)); in ihevc_sao_band_offset_luma_ssse3()
172 band_table2_8x16b = _mm_load_si128((__m128i *)(gu2_table_band_idx + 16)); in ihevc_sao_band_offset_luma_ssse3()
173 band_table3_8x16b = _mm_load_si128((__m128i *)(gu2_table_band_idx + 24)); in ihevc_sao_band_offset_luma_ssse3()
493 band_table0_16x8b = _mm_load_si128((__m128i *)(gu2_table_band_idx)); in ihevc_sao_band_offset_chroma_ssse3()
494 band_table1_16x8b = _mm_load_si128((__m128i *)(gu2_table_band_idx + 8)); in ihevc_sao_band_offset_chroma_ssse3()
495 band_table2_16x8b = _mm_load_si128((__m128i *)(gu2_table_band_idx + 16)); in ihevc_sao_band_offset_chroma_ssse3()
496 band_table3_16x8b = _mm_load_si128((__m128i *)(gu2_table_band_idx + 24)); in ihevc_sao_band_offset_chroma_ssse3()
575 temp0_8x16b = _mm_load_si128((__m128i *)(gu2_table_band_idx)); in ihevc_sao_band_offset_chroma_ssse3()
576 band_table1_16x8b = _mm_load_si128((__m128i *)(gu2_table_band_idx + 8)); in ihevc_sao_band_offset_chroma_ssse3()
[all …]
/external/libvpx/libvpx/vp8/encoder/x86/
Dquantize_sse2.c44 __m128i quant_shift0 = _mm_load_si128((__m128i *)(b->quant_shift)); in vp8_regular_quantize_b_sse2()
45 __m128i quant_shift1 = _mm_load_si128((__m128i *)(b->quant_shift + 8)); in vp8_regular_quantize_b_sse2()
46 __m128i z0 = _mm_load_si128((__m128i *)(b->coeff)); in vp8_regular_quantize_b_sse2()
47 __m128i z1 = _mm_load_si128((__m128i *)(b->coeff+8)); in vp8_regular_quantize_b_sse2()
49 __m128i zbin0 = _mm_load_si128((__m128i *)(b->zbin)); in vp8_regular_quantize_b_sse2()
50 __m128i zbin1 = _mm_load_si128((__m128i *)(b->zbin + 8)); in vp8_regular_quantize_b_sse2()
51 __m128i round0 = _mm_load_si128((__m128i *)(b->round)); in vp8_regular_quantize_b_sse2()
52 __m128i round1 = _mm_load_si128((__m128i *)(b->round + 8)); in vp8_regular_quantize_b_sse2()
53 __m128i quant0 = _mm_load_si128((__m128i *)(b->quant)); in vp8_regular_quantize_b_sse2()
54 __m128i quant1 = _mm_load_si128((__m128i *)(b->quant + 8)); in vp8_regular_quantize_b_sse2()
[all …]
Dquantize_sse4.c38 __m128i quant_shift0 = _mm_load_si128((__m128i *)(b->quant_shift)); in vp8_regular_quantize_b_sse4_1()
39 __m128i quant_shift1 = _mm_load_si128((__m128i *)(b->quant_shift + 8)); in vp8_regular_quantize_b_sse4_1()
40 __m128i z0 = _mm_load_si128((__m128i *)(b->coeff)); in vp8_regular_quantize_b_sse4_1()
41 __m128i z1 = _mm_load_si128((__m128i *)(b->coeff+8)); in vp8_regular_quantize_b_sse4_1()
43 __m128i zbin0 = _mm_load_si128((__m128i *)(b->zbin)); in vp8_regular_quantize_b_sse4_1()
44 __m128i zbin1 = _mm_load_si128((__m128i *)(b->zbin + 8)); in vp8_regular_quantize_b_sse4_1()
45 __m128i round0 = _mm_load_si128((__m128i *)(b->round)); in vp8_regular_quantize_b_sse4_1()
46 __m128i round1 = _mm_load_si128((__m128i *)(b->round + 8)); in vp8_regular_quantize_b_sse4_1()
47 __m128i quant0 = _mm_load_si128((__m128i *)(b->quant)); in vp8_regular_quantize_b_sse4_1()
48 __m128i quant1 = _mm_load_si128((__m128i *)(b->quant + 8)); in vp8_regular_quantize_b_sse4_1()
[all …]
Dquantize_ssse3.c45 __m128i z0 = _mm_load_si128((__m128i *)(b->coeff)); in vp8_fast_quantize_b_ssse3()
46 __m128i z1 = _mm_load_si128((__m128i *)(b->coeff + 8)); in vp8_fast_quantize_b_ssse3()
47 __m128i round0 = _mm_load_si128((__m128i *)(b->round)); in vp8_fast_quantize_b_ssse3()
48 __m128i round1 = _mm_load_si128((__m128i *)(b->round + 8)); in vp8_fast_quantize_b_ssse3()
49 __m128i quant_fast0 = _mm_load_si128((__m128i *)(b->quant_fast)); in vp8_fast_quantize_b_ssse3()
50 __m128i quant_fast1 = _mm_load_si128((__m128i *)(b->quant_fast + 8)); in vp8_fast_quantize_b_ssse3()
51 __m128i dequant0 = _mm_load_si128((__m128i *)(d->dequant)); in vp8_fast_quantize_b_ssse3()
52 __m128i dequant1 = _mm_load_si128((__m128i *)(d->dequant + 8)); in vp8_fast_quantize_b_ssse3()
58 __m128i zig_zag = _mm_load_si128((const __m128i *)pshufb_zig_zag_mask); in vp8_fast_quantize_b_ssse3()
/external/libvpx/libvpx/vp9/common/x86/
Dvp9_idct_intrin_sse2.h92 in[0] = _mm_load_si128((const __m128i *)(input + 0 * 16)); in load_buffer_8x16()
93 in[1] = _mm_load_si128((const __m128i *)(input + 1 * 16)); in load_buffer_8x16()
94 in[2] = _mm_load_si128((const __m128i *)(input + 2 * 16)); in load_buffer_8x16()
95 in[3] = _mm_load_si128((const __m128i *)(input + 3 * 16)); in load_buffer_8x16()
96 in[4] = _mm_load_si128((const __m128i *)(input + 4 * 16)); in load_buffer_8x16()
97 in[5] = _mm_load_si128((const __m128i *)(input + 5 * 16)); in load_buffer_8x16()
98 in[6] = _mm_load_si128((const __m128i *)(input + 6 * 16)); in load_buffer_8x16()
99 in[7] = _mm_load_si128((const __m128i *)(input + 7 * 16)); in load_buffer_8x16()
101 in[8] = _mm_load_si128((const __m128i *)(input + 8 * 16)); in load_buffer_8x16()
102 in[9] = _mm_load_si128((const __m128i *)(input + 9 * 16)); in load_buffer_8x16()
[all …]
Dvp9_loopfilter_intrin_sse2.c22 const __m128i blimit = _mm_load_si128((const __m128i *)_blimit); in mb_lpf_horizontal_edge_w_sse2_8()
23 const __m128i limit = _mm_load_si128((const __m128i *)_limit); in mb_lpf_horizontal_edge_w_sse2_8()
24 const __m128i thresh = _mm_load_si128((const __m128i *)_thresh); in mb_lpf_horizontal_edge_w_sse2_8()
383 const __m128i blimit = _mm_load_si128((const __m128i *)_blimit); in mb_lpf_horizontal_edge_w_sse2_16()
384 const __m128i limit = _mm_load_si128((const __m128i *)_limit); in mb_lpf_horizontal_edge_w_sse2_16()
385 const __m128i thresh = _mm_load_si128((const __m128i *)_thresh); in mb_lpf_horizontal_edge_w_sse2_16()
736 work_a = _mm_load_si128((__m128i *)&ap[2 * 16]); in mb_lpf_horizontal_edge_w_sse2_16()
737 p2 = _mm_load_si128((__m128i *)&flat_op[2 * 16]); in mb_lpf_horizontal_edge_w_sse2_16()
743 p1 = _mm_load_si128((__m128i *)&flat_op[1 * 16]); in mb_lpf_horizontal_edge_w_sse2_16()
749 p0 = _mm_load_si128((__m128i *)&flat_op[0]); in mb_lpf_horizontal_edge_w_sse2_16()
[all …]
Dvp9_subpixel_8t_intrin_ssse3.c71 shuffle1 =_mm_load_si128((__m128i const *)filt1_4_h8); in vp9_filter_block1d4_h8_intrin_ssse3()
72 shuffle2 = _mm_load_si128((__m128i const *)filt2_4_h8); in vp9_filter_block1d4_h8_intrin_ssse3()
144 filt1Reg = _mm_load_si128((__m128i const *)filt1_global); in vp9_filter_block1d8_h8_intrin_ssse3()
145 filt2Reg = _mm_load_si128((__m128i const *)filt2_global); in vp9_filter_block1d8_h8_intrin_ssse3()
146 filt3Reg = _mm_load_si128((__m128i const *)filt3_global); in vp9_filter_block1d8_h8_intrin_ssse3()
147 filt4Reg = _mm_load_si128((__m128i const *)filt4_global); in vp9_filter_block1d8_h8_intrin_ssse3()
224 filt1Reg = _mm_load_si128((__m128i const *)filt1_global); in vp9_filter_block1d16_h8_intrin_ssse3()
225 filt2Reg = _mm_load_si128((__m128i const *)filt2_global); in vp9_filter_block1d16_h8_intrin_ssse3()
226 filt3Reg = _mm_load_si128((__m128i const *)filt3_global); in vp9_filter_block1d16_h8_intrin_ssse3()
227 filt4Reg = _mm_load_si128((__m128i const *)filt4_global); in vp9_filter_block1d16_h8_intrin_ssse3()
Dvp9_idct_intrin_sse2.c34 input0 = _mm_load_si128((const __m128i *)input); in vp9_idct4x4_16_add_sse2()
35 input2 = _mm_load_si128((const __m128i *)(input + 8)); in vp9_idct4x4_16_add_sse2()
536 in0 = _mm_load_si128((const __m128i *)input); in vp9_idct8x8_64_add_sse2()
537 in1 = _mm_load_si128((const __m128i *)(input + 8 * 1)); in vp9_idct8x8_64_add_sse2()
538 in2 = _mm_load_si128((const __m128i *)(input + 8 * 2)); in vp9_idct8x8_64_add_sse2()
539 in3 = _mm_load_si128((const __m128i *)(input + 8 * 3)); in vp9_idct8x8_64_add_sse2()
540 in4 = _mm_load_si128((const __m128i *)(input + 8 * 4)); in vp9_idct8x8_64_add_sse2()
541 in5 = _mm_load_si128((const __m128i *)(input + 8 * 5)); in vp9_idct8x8_64_add_sse2()
542 in6 = _mm_load_si128((const __m128i *)(input + 8 * 6)); in vp9_idct8x8_64_add_sse2()
543 in7 = _mm_load_si128((const __m128i *)(input + 8 * 7)); in vp9_idct8x8_64_add_sse2()
[all …]
Dvp9_idct_intrin_ssse3.c693 in[0] = _mm_load_si128((const __m128i *)input); in vp9_idct16x16_10_add_ssse3()
694 in[1] = _mm_load_si128((const __m128i *)(input + 8 * 2)); in vp9_idct16x16_10_add_ssse3()
695 in[2] = _mm_load_si128((const __m128i *)(input + 8 * 4)); in vp9_idct16x16_10_add_ssse3()
696 in[3] = _mm_load_si128((const __m128i *)(input + 8 * 6)); in vp9_idct16x16_10_add_ssse3()
/external/jpeg/
Djidctintelsse.c171 x3 = _mm_load_si128(( __m128i*)(wsptr+24));\ in jpeg_idct_intelsse()
172 x1 = _mm_load_si128(( __m128i*)(wsptr+8));\ in jpeg_idct_intelsse()
200 x0 = _mm_load_si128(( __m128i*)(wsptr));\ in jpeg_idct_intelsse()
201 x4 = _mm_load_si128(( __m128i*)(wsptr+32));\ in jpeg_idct_intelsse()
202 x2 = _mm_load_si128(( __m128i*)(wsptr+16));\ in jpeg_idct_intelsse()
203 x6 = _mm_load_si128(( __m128i*)(wsptr+48));\ in jpeg_idct_intelsse()
281 row0 = _mm_load_si128((__m128i const*)(coef_blockSSE)); in jpeg_idct_intelsse()
282 row2 = _mm_load_si128((__m128i const*)(coef_blockSSE+8*2)); in jpeg_idct_intelsse()
292 row0 = _mm_load_si128((__m128i const*)(coef_blockSSE+8*4)); in jpeg_idct_intelsse()
293 row2 = _mm_load_si128((__m128i const*)(coef_blockSSE+8*6)); in jpeg_idct_intelsse()
[all …]
/external/jemalloc/test/include/test/
DSFMT-sse2.h67 x = _mm_load_si128(a); in mm_recursion()
89 r1 = _mm_load_si128(&ctx->sfmt[N - 2].si); in gen_rand_all()
90 r2 = _mm_load_si128(&ctx->sfmt[N - 1].si); in gen_rand_all()
119 r1 = _mm_load_si128(&ctx->sfmt[N - 2].si); in gen_rand_array()
120 r2 = _mm_load_si128(&ctx->sfmt[N - 1].si); in gen_rand_array()
144 r = _mm_load_si128(&array[j + size - N].si); in gen_rand_array()
/external/libvpx/libvpx/vp9/encoder/x86/
Dvp9_dct_sse2.c410 __m128i in0 = _mm_load_si128((const __m128i *)(input + 0 * stride)); in vp9_fdct8x8_1_sse2()
411 __m128i in1 = _mm_load_si128((const __m128i *)(input + 1 * stride)); in vp9_fdct8x8_1_sse2()
412 __m128i in2 = _mm_load_si128((const __m128i *)(input + 2 * stride)); in vp9_fdct8x8_1_sse2()
413 __m128i in3 = _mm_load_si128((const __m128i *)(input + 3 * stride)); in vp9_fdct8x8_1_sse2()
419 in0 = _mm_load_si128((const __m128i *)(input + 4 * stride)); in vp9_fdct8x8_1_sse2()
420 in1 = _mm_load_si128((const __m128i *)(input + 5 * stride)); in vp9_fdct8x8_1_sse2()
421 in2 = _mm_load_si128((const __m128i *)(input + 6 * stride)); in vp9_fdct8x8_1_sse2()
422 in3 = _mm_load_si128((const __m128i *)(input + 7 * stride)); in vp9_fdct8x8_1_sse2()
465 __m128i in0 = _mm_load_si128((const __m128i *)(input + 0 * stride)); in vp9_fdct8x8_sse2()
466 __m128i in1 = _mm_load_si128((const __m128i *)(input + 1 * stride)); in vp9_fdct8x8_sse2()
[all …]
/external/neven/Embedded/common/src/b_BasicEm/
DMathSSE2.c165 m_XMM0 = _mm_load_si128( (__m128i *)&0[vec1L] ); in bbs_dotProduct_128SSE2()
168 m_XMM2 = _mm_load_si128( (__m128i *)&0[vec2L] ); in bbs_dotProduct_128SSE2()
170 m_XMM6 = _mm_load_si128( (__m128i *)&8[vec1L] ); in bbs_dotProduct_128SSE2()
176 m_XMM3 = _mm_load_si128( (__m128i *)&8[vec2L] ); in bbs_dotProduct_128SSE2()
188 m_XMM0 = _mm_load_si128( (__m128i *)&m_XMM5 ); in bbs_dotProduct_128SSE2()
/external/pdfium/core/src/fxcodec/fx_libopenjpeg/libopenjpeg20/
Dmct.c87 __m128i r = _mm_load_si128((const __m128i *)&(c0[i])); in opj_mct_encode()
88 __m128i g = _mm_load_si128((const __m128i *)&(c1[i])); in opj_mct_encode()
89 __m128i b = _mm_load_si128((const __m128i *)&(c2[i])); in opj_mct_encode()
152 __m128i y = _mm_load_si128((const __m128i *)&(c0[i])); in opj_mct_decode()
153 __m128i u = _mm_load_si128((const __m128i *)&(c1[i])); in opj_mct_decode()
154 __m128i v = _mm_load_si128((const __m128i *)&(c2[i])); in opj_mct_decode()
231 __m128i r = _mm_load_si128((const __m128i *)&(c0[i])); in opj_mct_encode_real()
232 __m128i g = _mm_load_si128((const __m128i *)&(c1[i])); in opj_mct_encode_real()
233 __m128i b = _mm_load_si128((const __m128i *)&(c2[i])); in opj_mct_encode_real()
/external/mesa3d/src/gallium/drivers/llvmpipe/
Dlp_rast_tri.c264 __m128i p0 = _mm_load_si128((__m128i *)&plane[0]); /* c, dcdx, dcdy, eo */ in lp_rast_triangle_3_16()
265 __m128i p1 = _mm_load_si128((__m128i *)&plane[1]); /* c, dcdx, dcdy, eo */ in lp_rast_triangle_3_16()
266 __m128i p2 = _mm_load_si128((__m128i *)&plane[2]); /* c, dcdx, dcdy, eo */ in lp_rast_triangle_3_16()
374 __m128i p0 = _mm_load_si128((__m128i *)&plane[0]); /* c, dcdx, dcdy, eo */ in lp_rast_triangle_3_4()
375 __m128i p1 = _mm_load_si128((__m128i *)&plane[1]); /* c, dcdx, dcdy, eo */ in lp_rast_triangle_3_4()
376 __m128i p2 = _mm_load_si128((__m128i *)&plane[2]); /* c, dcdx, dcdy, eo */ in lp_rast_triangle_3_4()
/external/skia/src/opts/
DSkBlitRow_opts_SSE2.cpp45 __m128i dst_pixel = _mm_load_si128(d); in S32_Blend_BlitRow32_SSE2()
95 __m128i dst_pixel = _mm_load_si128(d); in S32A_Opaque_BlitRow32_SSE2()
215 __m128i dst_pixel = _mm_load_si128(d); in S32A_Blend_BlitRow32_SSE2()
263 __m128i pixels = _mm_load_si128(dst_wide); in Color32A_D565_SSE2()
325 __m128i dst_pixel = _mm_load_si128(d); in SkARGB32_A8_BlitMask_SSE2()
582 __m128i dst_sse = _mm_load_si128(d); in SkBlitLCD16Row_SSE2()
650 __m128i dst_sse = _mm_load_si128(d); in SkBlitLCD16OpaqueRow_SSE2()
786 __m128i dst_pixel = _mm_load_si128(d); in S32A_D565_Opaque_SSE2()
1045 __m128i dst_pixel = _mm_load_si128(d); in S32A_D565_Opaque_Dither_SSE2()
/external/opencv/cv/src/
Dcvstereobm.cpp268 __m128i hsad_l = _mm_load_si128((__m128i*)(hsad + d)); in icvFindStereoCorrespondenceBM_SSE2()
269 __m128i hsad_h = _mm_load_si128((__m128i*)(hsad + d + 8)); in icvFindStereoCorrespondenceBM_SSE2()
270 __m128i cbs = _mm_load_si128((const __m128i*)(cbuf_sub + d)); in icvFindStereoCorrespondenceBM_SSE2()
312 __m128i v0 = _mm_load_si128((__m128i*)(hsad_sub + d)); in icvFindStereoCorrespondenceBM_SSE2()
313 __m128i v1 = _mm_load_si128((__m128i*)(hsad + d)); in icvFindStereoCorrespondenceBM_SSE2()
314 __m128i sad8 = _mm_load_si128((__m128i*)(sad + d)); in icvFindStereoCorrespondenceBM_SSE2()
359 __m128i sad8 = _mm_load_si128((__m128i*)(sad + d)); in icvFindStereoCorrespondenceBM_SSE2()
Dcvsmooth.cpp578 _mm_load_si128((__m128i*) &y[0]), _mm_load_si128((__m128i*) &x[0] ))); in histogram_add()
580 _mm_load_si128((__m128i*) &y[8]), _mm_load_si128((__m128i*) &x[8] ))); in histogram_add()
613 _mm_load_si128((__m128i*) &y[0]), _mm_load_si128((__m128i*) &x[0] ))); in histogram_sub()
615 _mm_load_si128((__m128i*) &y[8]), _mm_load_si128((__m128i*) &x[8] ))); in histogram_sub()
/external/boringssl/src/crypto/poly1305/
Dpoly1305_vec.c189 const xmmi MMASK = _mm_load_si128((xmmi *)poly1305_x64_sse2_message_mask); in poly1305_first_block()
190 const xmmi FIVE = _mm_load_si128((xmmi *)poly1305_x64_sse2_5); in poly1305_first_block()
191 const xmmi HIBIT = _mm_load_si128((xmmi *)poly1305_x64_sse2_1shl128); in poly1305_first_block()
282 const xmmi MMASK = _mm_load_si128((xmmi *)poly1305_x64_sse2_message_mask); in poly1305_blocks()
283 const xmmi FIVE = _mm_load_si128((xmmi *)poly1305_x64_sse2_5); in poly1305_blocks()
284 const xmmi HIBIT = _mm_load_si128((xmmi *)poly1305_x64_sse2_1shl128); in poly1305_blocks()
472 const xmmi MMASK = _mm_load_si128((xmmi *)poly1305_x64_sse2_message_mask); in poly1305_combine()
473 const xmmi HIBIT = _mm_load_si128((xmmi *)poly1305_x64_sse2_1shl128); in poly1305_combine()
474 const xmmi FIVE = _mm_load_si128((xmmi *)poly1305_x64_sse2_5); in poly1305_combine()

12