Home
last modified time | relevance | path

Searched refs:xmm1 (Results 1 – 25 of 305) sorted by relevance

12345678910>>...13

/external/llvm/test/MC/X86/
Dshuffle-comments.s3 palignr $8, %xmm0, %xmm1
4 # CHECK: xmm1 = xmm0[8,9,10,11,12,13,14,15],xmm1[0,1,2,3,4,5,6,7]
5 palignr $8, (%rax), %xmm1
6 # CHECK: xmm1 = mem[8,9,10,11,12,13,14,15],xmm1[0,1,2,3,4,5,6,7]
8 palignr $16, %xmm0, %xmm1
9 # CHECK: xmm1 = xmm1[0,1,2,3,4,5,6,7,8,9,10,11,12,13,14,15]
10 palignr $16, (%rax), %xmm1
11 # CHECK: xmm1 = xmm1[0,1,2,3,4,5,6,7,8,9,10,11,12,13,14,15]
13 palignr $0, %xmm0, %xmm1
14 # CHECK: xmm1 = xmm0[0,1,2,3,4,5,6,7,8,9,10,11,12,13,14,15]
[all …]
Dx86-32-fma3.s5 vfmadd132pd %xmm2, %xmm5, %xmm1
9 vfmadd132pd (%eax), %xmm5, %xmm1
13 vfmadd132ps %xmm2, %xmm5, %xmm1
17 vfmadd132ps (%eax), %xmm5, %xmm1
21 vfmadd213pd %xmm2, %xmm5, %xmm1
25 vfmadd213pd (%eax), %xmm5, %xmm1
29 vfmadd213ps %xmm2, %xmm5, %xmm1
33 vfmadd213ps (%eax), %xmm5, %xmm1
37 vfmadd231pd %xmm2, %xmm5, %xmm1
41 vfmadd231pd (%eax), %xmm5, %xmm1
[all …]
Dx86_64-fma4-encoding.s6 vfmaddss (%rcx), %xmm1, %xmm0, %xmm0
10 vfmaddss %xmm1, (%rcx),%xmm0, %xmm0
14 vfmaddss %xmm2, %xmm1, %xmm0, %xmm0
18 vfmaddsd (%rcx), %xmm1, %xmm0, %xmm0
22 vfmaddsd %xmm1, (%rcx),%xmm0, %xmm0
26 vfmaddsd %xmm2, %xmm1, %xmm0, %xmm0
30 vfmaddsd %xmm10, %xmm1, %xmm0, %xmm0
34 vfmaddps (%rcx), %xmm1, %xmm0, %xmm0
38 vfmaddps %xmm1, (%rcx),%xmm0, %xmm0
42 vfmaddps %xmm2, %xmm1, %xmm0, %xmm0
[all …]
Dx86-32-avx.s276 vunpckhps %xmm1, %xmm2, %xmm4
280 vunpckhpd %xmm1, %xmm2, %xmm4
284 vunpcklps %xmm1, %xmm2, %xmm4
288 vunpcklpd %xmm1, %xmm2, %xmm4
308 vcmpps $0, %xmm0, %xmm6, %xmm1
312 vcmpps $0, (%eax), %xmm6, %xmm1
316 vcmpps $7, %xmm0, %xmm6, %xmm1
320 vcmppd $0, %xmm0, %xmm6, %xmm1
324 vcmppd $0, (%eax), %xmm6, %xmm1
328 vcmppd $7, %xmm0, %xmm6, %xmm1
[all …]
Dx86_64-xop-encoding.s10 vphsubwd (%rcx,%rax), %xmm1
13 vphsubwd %xmm0, %xmm1
18 vphsubdq (%rcx,%rax), %xmm1
21 vphsubdq %xmm0, %xmm1
26 vphsubbw (%rax), %xmm1
29 vphsubbw %xmm2, %xmm1
61 vphadduwd %xmm2, %xmm1
106 vphaddbw (%rcx,%rax), %xmm1
114 vphaddbq (%rcx,%rax), %xmm1
122 vphaddbd (%rcx,%rax), %xmm1
[all …]
/external/llvm/test/CodeGen/X86/
Dvector-shuffle-128-v4.ll307 ; SSE2-NEXT: shufps {{.*#+}} xmm1 = xmm1[0,0],xmm0[2,0]
308 ; SSE2-NEXT: shufps {{.*#+}} xmm0 = xmm0[0,1],xmm1[2,0]
313 ; SSE3-NEXT: shufps {{.*#+}} xmm1 = xmm1[0,0],xmm0[2,0]
314 ; SSE3-NEXT: shufps {{.*#+}} xmm0 = xmm0[0,1],xmm1[2,0]
319 ; SSSE3-NEXT: shufps {{.*#+}} xmm1 = xmm1[0,0],xmm0[2,0]
320 ; SSSE3-NEXT: shufps {{.*#+}} xmm0 = xmm0[0,1],xmm1[2,0]
325 ; SSE41-NEXT: pshufd {{.*#+}} xmm1 = xmm1[0,1,2,0]
326 ; SSE41-NEXT: pblendw {{.*#+}} xmm0 = xmm0[0,1,2,3,4,5],xmm1[6,7]
331 ; AVX1-NEXT: vpshufd {{.*#+}} xmm1 = xmm1[0,1,2,0]
332 ; AVX1-NEXT: vpblendw {{.*#+}} xmm0 = xmm0[0,1,2,3,4,5],xmm1[6,7]
[all …]
Dvector-zext.ll10 ; SSE2-NEXT: movdqa %xmm0, %xmm1
14 ; SSE2-NEXT: punpckhwd {{.*#+}} xmm1 = xmm1[4,4,5,5,6,6,7,7]
15 ; SSE2-NEXT: pand .LCPI0_0(%rip), %xmm1
20 ; SSSE3-NEXT: movdqa %xmm0, %xmm1
24 ; SSSE3-NEXT: punpckhwd {{.*#+}} xmm1 = xmm1[4,4,5,5,6,6,7,7]
25 ; SSSE3-NEXT: pand .LCPI0_0(%rip), %xmm1
30 ; SSE41-NEXT: movdqa %xmm0, %xmm1
31 ; SSE41-NEXT: pmovzxwd {{.*#+}} xmm0 = xmm1[0],zero,xmm1[1],zero,xmm1[2],zero,xmm1[3],zero
32 ; SSE41-NEXT: punpckhwd {{.*#+}} xmm1 = xmm1[4,4,5,5,6,6,7,7]
33 ; SSE41-NEXT: pand .LCPI0_0(%rip), %xmm1
[all …]
Dbswap-vector.ll16 ; CHECK-NOSSSE3-NEXT: pxor %xmm1, %xmm1
18 …m2 = xmm2[8],xmm1[8],xmm2[9],xmm1[9],xmm2[10],xmm1[10],xmm2[11],xmm1[11],xmm2[12],xmm1[12],xmm2[13…
21 …}} xmm0 = xmm0[0],xmm1[0],xmm0[1],xmm1[1],xmm0[2],xmm1[2],xmm0[3],xmm1[3],xmm0[4],xmm1[4],xmm0[5],
49 ; CHECK-NOSSSE3-NEXT: pxor %xmm1, %xmm1
51 …m2 = xmm2[8],xmm1[8],xmm2[9],xmm1[9],xmm2[10],xmm1[10],xmm2[11],xmm1[11],xmm2[12],xmm1[12],xmm2[13…
54 …}} xmm0 = xmm0[0],xmm1[0],xmm0[1],xmm1[1],xmm0[2],xmm1[2],xmm0[3],xmm1[3],xmm0[4],xmm1[4],xmm0[5],
82 ; CHECK-NOSSSE3-NEXT: pxor %xmm1, %xmm1
84 …m2 = xmm2[8],xmm1[8],xmm2[9],xmm1[9],xmm2[10],xmm1[10],xmm2[11],xmm1[11],xmm2[12],xmm1[12],xmm2[13…
88 …}} xmm0 = xmm0[0],xmm1[0],xmm0[1],xmm1[1],xmm0[2],xmm1[2],xmm0[3],xmm1[3],xmm0[4],xmm1[4],xmm0[5],
130 ; CHECK-NOSSSE3-NEXT: movdqa %xmm1, %xmm3
[all …]
Dsse-minmax.ll16 ; CHECK-NEXT: maxsd %xmm1, %xmm0
19 ; UNSAFE-NEXT: maxsd %xmm1, %xmm0
22 ; FINITE-NEXT: maxsd %xmm1, %xmm0
31 ; CHECK-NEXT: minsd %xmm1, %xmm0
34 ; UNSAFE-NEXT: minsd %xmm1, %xmm0
37 ; FINITE-NEXT: minsd %xmm1, %xmm0
46 ; CHECK-NEXT: minsd %xmm0, %xmm1
47 ; CHECK-NEXT: movap{{[sd]}} %xmm1, %xmm0
50 ; UNSAFE-NEXT: minsd %xmm1, %xmm0
53 ; FINITE-NEXT: minsd %xmm0, %xmm1
[all …]
Dcombine-and.ll11 ; CHECK-NEXT: pxor %xmm1, %xmm1
12 ; CHECK-NEXT: pblendw {{.*#+}} xmm0 = xmm0[0,1],xmm1[2,3,4,5,6,7]
21 ; CHECK-NEXT: pxor %xmm1, %xmm1
22 ; CHECK-NEXT: pblendw {{.*#+}} xmm0 = xmm1[0,1],xmm0[2,3],xmm1[4,5,6,7]
31 ; CHECK-NEXT: pxor %xmm1, %xmm1
32 ; CHECK-NEXT: pblendw {{.*#+}} xmm0 = xmm1[0,1,2,3],xmm0[4,5],xmm1[6,7]
41 ; CHECK-NEXT: pxor %xmm1, %xmm1
42 ; CHECK-NEXT: pblendw {{.*#+}} xmm0 = xmm1[0,1,2,3,4,5],xmm0[6,7]
51 ; CHECK-NEXT: pxor %xmm1, %xmm1
52 ; CHECK-NEXT: pblendw {{.*#+}} xmm0 = xmm0[0,1],xmm1[2,3],xmm0[4,5],xmm1[6,7]
[all …]
Dvector-shuffle-combining.ll166 ; SSE-NEXT: pand %xmm1, %xmm0
172 ; AVX-NEXT: vpand %xmm1, %xmm0, %xmm0
184 ; SSE-NEXT: por %xmm1, %xmm0
190 ; AVX-NEXT: vpor %xmm1, %xmm0, %xmm0
202 ; SSE-NEXT: pxor %xmm1, %xmm0
208 ; AVX-NEXT: vpxor %xmm1, %xmm0, %xmm0
220 ; SSE-NEXT: pand %xmm1, %xmm0
226 ; AVX-NEXT: vpand %xmm1, %xmm0, %xmm0
238 ; SSE-NEXT: por %xmm1, %xmm0
244 ; AVX-NEXT: vpor %xmm1, %xmm0, %xmm0
[all …]
Dvector-shuffle-128-v2.ll58 ; SSE-NEXT: pshufd {{.*#+}} xmm0 = xmm1[0,1,0,1]
63 ; AVX1-NEXT: vpshufd {{.*#+}} xmm0 = xmm1[0,1,0,1]
68 ; AVX2-NEXT: vpbroadcastq %xmm1, %xmm0
76 ; SSE-NEXT: pshufd {{.*#+}} xmm0 = xmm1[2,3,0,1]
81 ; AVX-NEXT: vpshufd {{.*#+}} xmm0 = xmm1[2,3,0,1]
89 ; SSE-NEXT: pshufd {{.*#+}} xmm0 = xmm1[2,3,2,3]
94 ; AVX-NEXT: vpshufd {{.*#+}} xmm0 = xmm1[2,3,2,3]
157 ; SSE2-NEXT: movlhps {{.*#+}} xmm1 = xmm1[0,0]
158 ; SSE2-NEXT: movaps %xmm1, %xmm0
163 ; SSE3-NEXT: movddup {{.*#+}} xmm0 = xmm1[0,0]
[all …]
Dvector-shuffle-256-v16.ll26 ; AVX1-NEXT: vpshufhw {{.*#+}} xmm1 = xmm0[0,1,2,3,4,4,4,4]
33 ; AVX2-NEXT: vpbroadcastw %xmm0, %xmm1
44 ; AVX1-NEXT: vpshufb {{.*#+}} xmm1 = xmm0[0,1,0,1,0,1,0,1,0,1,0,1,0,1,0,1]
51 ; AVX2-NEXT: vpbroadcastw %xmm0, %xmm1
62 ; AVX1-NEXT: vpshufb {{.*#+}} xmm1 = xmm0[0,1,0,1,0,1,0,1,0,1,0,1,0,1,0,1]
69 ; AVX2-NEXT: vpbroadcastw %xmm0, %xmm1
80 ; AVX1-NEXT: vpshufb {{.*#+}} xmm1 = xmm0[0,1,0,1,0,1,0,1,0,1,0,1,0,1,0,1]
87 ; AVX2-NEXT: vpbroadcastw %xmm0, %xmm1
98 ; AVX1-NEXT: vpshufb {{.*#+}} xmm1 = xmm0[0,1,0,1,0,1,0,1,0,1,0,1,0,1,0,1]
105 ; AVX2-NEXT: vpbroadcastw %xmm0, %xmm1
[all …]
Dvector-shuffle-128-v16.ll29 ; SSSE3-NEXT: pxor %xmm1, %xmm1
30 ; SSSE3-NEXT: pshufb %xmm1, %xmm0
35 ; SSE41-NEXT: pxor %xmm1, %xmm1
36 ; SSE41-NEXT: pshufb %xmm1, %xmm0
41 ; AVX1-NEXT: vpxor %xmm1, %xmm1, %xmm1
42 ; AVX1-NEXT: vpshufb %xmm1, %xmm0, %xmm0
224 …}} xmm0 = xmm0[0],xmm1[0],xmm0[1],xmm1[1],xmm0[2],xmm1[2],xmm0[3],xmm1[3],xmm0[4],xmm1[4],xmm0[5],
229 …}} xmm0 = xmm0[0],xmm1[0],xmm0[1],xmm1[1],xmm0[2],xmm1[2],xmm0[3],xmm1[3],xmm0[4],xmm1[4],xmm0[5],
238 …m0 = xmm0[8],xmm1[8],xmm0[9],xmm1[9],xmm0[10],xmm1[10],xmm0[11],xmm1[11],xmm0[12],xmm1[12],xmm0[13…
243 …m0 = xmm0[8],xmm1[8],xmm0[9],xmm1[9],xmm0[10],xmm1[10],xmm0[11],xmm1[11],xmm0[12],xmm1[12],xmm0[13…
[all …]
Dvector-sext.ll13 ; SSE2-NEXT: movdqa %xmm0, %xmm1
18 ; SSE2-NEXT: punpckhwd {{.*#+}} xmm1 = xmm1[4,4,5,5,6,6,7,7]
19 ; SSE2-NEXT: pslld $16, %xmm1
20 ; SSE2-NEXT: psrad $16, %xmm1
25 ; SSSE3-NEXT: movdqa %xmm0, %xmm1
30 ; SSSE3-NEXT: punpckhwd {{.*#+}} xmm1 = xmm1[4,4,5,5,6,6,7,7]
31 ; SSSE3-NEXT: pslld $16, %xmm1
32 ; SSSE3-NEXT: psrad $16, %xmm1
37 ; SSE41-NEXT: movdqa %xmm0, %xmm1
38 ; SSE41-NEXT: pmovzxwd %xmm1, %xmm0
[all …]
Dvec_setcc.ll10 ; SSE2: pmaxub %xmm0, %xmm1
11 ; SSE2: pcmpeqb %xmm1, %xmm0
14 ; SSE41: pmaxub %xmm0, %xmm1
15 ; SSE41: pcmpeqb %xmm1, %xmm0
18 ; AVX: vpmaxub %xmm1, %xmm0, %xmm1
19 ; AVX: vpcmpeqb %xmm1, %xmm0, %xmm0
27 ; SSE2: pminub %xmm0, %xmm1
28 ; SSE2: pcmpeqb %xmm1, %xmm0
31 ; SSE41: pminub %xmm0, %xmm1
32 ; SSE41: pcmpeqb %xmm1, %xmm0
[all …]
Dvector-idiv.ll10 ; SSE41-NEXT: movdqa {{.*#+}} xmm1 = [613566757,613566757,613566757,613566757]
11 ; SSE41-NEXT: pshufd {{.*#+}} xmm2 = xmm1[1,1,3,3]
14 ; SSE41-NEXT: pmuludq %xmm0, %xmm1
15 ; SSE41-NEXT: pshufd {{.*#+}} xmm1 = xmm1[1,1,3,3]
16 ; SSE41-NEXT: pblendw {{.*#+}} xmm1 = xmm1[0,1],xmm3[2,3],xmm1[4,5],xmm3[6,7]
17 ; SSE41-NEXT: psubd %xmm1, %xmm0
19 ; SSE41-NEXT: paddd %xmm1, %xmm0
25 ; SSE-NEXT: movdqa {{.*#+}} xmm1 = [613566757,613566757,613566757,613566757]
27 ; SSE-NEXT: pmuludq %xmm1, %xmm2
29 ; SSE-NEXT: pshufd {{.*#+}} xmm1 = xmm1[1,1,3,3]
[all …]
Dvector-blend.ll12 ; SSE2-NEXT: shufps {{.*#+}} xmm0 = xmm0[0,2],xmm1[1,3]
18 ; SSSE3-NEXT: shufps {{.*#+}} xmm0 = xmm0[0,2],xmm1[1,3]
24 ; SSE41-NEXT: blendps {{.*#+}} xmm0 = xmm0[0],xmm1[1],xmm0[2],xmm1[3]
29 ; AVX-NEXT: vblendps {{.*#+}} xmm0 = xmm0[0],xmm1[1],xmm0[2],xmm1[3]
39 ; SSE2-NEXT: movss {{.*#+}} xmm1 = xmm0[0],xmm1[1,2,3]
40 ; SSE2-NEXT: movaps %xmm1, %xmm0
45 ; SSSE3-NEXT: movss {{.*#+}} xmm1 = xmm0[0],xmm1[1,2,3]
46 ; SSSE3-NEXT: movaps %xmm1, %xmm0
51 ; SSE41-NEXT: blendps {{.*#+}} xmm0 = xmm0[0],xmm1[1,2,3]
56 ; AVX-NEXT: vblendps {{.*#+}} xmm0 = xmm0[0],xmm1[1,2,3]
[all …]
Dpalignr.ll21 ; CHECK-NEXT: palignr {{.*#+}} xmm1 = xmm0[4,5,6,7,8,9,10,11,12,13,14,15],xmm1[0,1,2,3]
22 ; CHECK-NEXT: movdqa %xmm1, %xmm0
27 ; CHECK-YONAH-NEXT: shufps {{.*#+}} xmm1 = xmm1[0,0],xmm0[3,0]
28 ; CHECK-YONAH-NEXT: shufps {{.*#+}} xmm0 = xmm0[1,2],xmm1[2,0]
37 ; CHECK-NEXT: palignr {{.*#+}} xmm1 = xmm0[4,5,6,7,8,9,10,11,12,13,14,15],xmm1[0,1,2,3]
38 ; CHECK-NEXT: movdqa %xmm1, %xmm0
43 ; CHECK-YONAH-NEXT: pshufd {{.*#+}} xmm1 = xmm1[0,0,1,1]
45 ; CHECK-YONAH-NEXT: punpcklqdq {{.*#+}} xmm0 = xmm0[0],xmm1[0]
54 ; CHECK-NEXT: palignr {{.*#+}} xmm0 = xmm1[8,9,10,11,12,13,14,15],xmm0[0,1,2,3,4,5,6,7]
59 ; CHECK-YONAH-NEXT: shufpd {{.*#+}} xmm1 = xmm1[1],xmm0[0]
[all …]
/external/boringssl/src/crypto/aes/asm/
Dvpaes-x86_64.pl79 ## Clobbers %xmm1-%xmm5, %r9, %r10, %r11, %rax
89 movdqa %xmm9, %xmm1
91 pandn %xmm0, %xmm1
93 psrld \$4, %xmm1
97 pshufb %xmm1, %xmm0
114 movdqa -0x40(%r11,%r10), %xmm1 # .Lk_mc_forward[]
121 pshufb %xmm1, %xmm0 # 0 = B
127 pshufb %xmm1, %xmm0 # 0 = 2B+C
134 movdqa %xmm9, %xmm1 # 1 : i
136 pandn %xmm0, %xmm1 # 1 = i<<4
[all …]
/external/libyuv/files/source/
Drow_win.cc139 movdqa xmm1, xmm0 in I400ToARGBRow_SSE2()
141 punpckhwd xmm1, xmm1 in I400ToARGBRow_SSE2()
143 por xmm1, xmm5 in I400ToARGBRow_SSE2()
145 movdqa [edx + 16], xmm1 in I400ToARGBRow_SSE2()
250 movdqu xmm1, [eax + 16] in RGB24ToARGBRow_SSSE3()
254 palignr xmm2, xmm1, 8 // xmm2 = { xmm3[0:3] xmm1[8:15]} in RGB24ToARGBRow_SSSE3()
257 palignr xmm1, xmm0, 12 // xmm1 = { xmm3[0:7] xmm0[12:15]} in RGB24ToARGBRow_SSSE3()
261 pshufb xmm1, xmm4 in RGB24ToARGBRow_SSSE3()
263 por xmm1, xmm5 in RGB24ToARGBRow_SSSE3()
266 movdqa [edx + 16], xmm1 in RGB24ToARGBRow_SSSE3()
[all …]
/external/libvpx/libvpx/third_party/libyuv/source/
Dcompare_win.cc32 movdqa xmm1, [eax] in SumSquareError_SSE2()
37 movdqa xmm3, xmm1 // abs trick in SumSquareError_SSE2()
38 psubusb xmm1, xmm2 in SumSquareError_SSE2()
40 por xmm1, xmm2 in SumSquareError_SSE2()
41 movdqa xmm2, xmm1 in SumSquareError_SSE2()
42 punpcklbw xmm1, xmm5 in SumSquareError_SSE2()
44 pmaddwd xmm1, xmm1 in SumSquareError_SSE2()
46 paddd xmm0, xmm1 in SumSquareError_SSE2()
50 pshufd xmm1, xmm0, 0xee in SumSquareError_SSE2()
51 paddd xmm0, xmm1 in SumSquareError_SSE2()
[all …]
/external/libvpx/libvpx/vp8/common/x86/
Dvariance_impl_ssse3.asm68 movdqu xmm1, XMMWORD PTR [rsi+1]
71 punpcklbw xmm0, xmm1
72 punpckhbw xmm2, xmm1
92 movdqu xmm1, XMMWORD PTR [rsi]
94 movdqa xmm3, xmm1
96 punpcklbw xmm1, xmm2
98 pmaddubsw xmm1, [rax]
101 paddw xmm1, [GLOBAL(xmm_bi_rd)]
103 psraw xmm1, xmm_filter_shift
105 packuswb xmm1, xmm3
[all …]
/external/boringssl/mac-x86/crypto/aes/
Daesni-x86.S15 movups 16(%edx),%xmm1
21 movups (%edx),%xmm1
26 pxor %xmm1,%xmm1
41 movups 16(%edx),%xmm1
47 movups (%edx),%xmm1
52 pxor %xmm1,%xmm1
61 movups 16(%edx),%xmm1
71 movups (%edx,%ecx,1),%xmm1
87 movups 16(%edx),%xmm1
97 movups (%edx,%ecx,1),%xmm1
[all …]
/external/boringssl/linux-x86/crypto/aes/
Daesni-x86.S16 movups 16(%edx),%xmm1
22 movups (%edx),%xmm1
27 pxor %xmm1,%xmm1
44 movups 16(%edx),%xmm1
50 movups (%edx),%xmm1
55 pxor %xmm1,%xmm1
66 movups 16(%edx),%xmm1
76 movups (%edx,%ecx,1),%xmm1
94 movups 16(%edx),%xmm1
104 movups (%edx,%ecx,1),%xmm1
[all …]

12345678910>>...13