/external/llvm/test/MC/X86/ |
D | shuffle-comments.s | 3 palignr $8, %xmm0, %xmm1 4 # CHECK: xmm1 = xmm0[8,9,10,11,12,13,14,15],xmm1[0,1,2,3,4,5,6,7] 5 palignr $8, (%rax), %xmm1 6 # CHECK: xmm1 = mem[8,9,10,11,12,13,14,15],xmm1[0,1,2,3,4,5,6,7] 8 palignr $16, %xmm0, %xmm1 9 # CHECK: xmm1 = xmm1[0,1,2,3,4,5,6,7,8,9,10,11,12,13,14,15] 10 palignr $16, (%rax), %xmm1 11 # CHECK: xmm1 = xmm1[0,1,2,3,4,5,6,7,8,9,10,11,12,13,14,15] 13 palignr $0, %xmm0, %xmm1 14 # CHECK: xmm1 = xmm0[0,1,2,3,4,5,6,7,8,9,10,11,12,13,14,15] [all …]
|
D | x86-32-fma3.s | 5 vfmadd132pd %xmm2, %xmm5, %xmm1 9 vfmadd132pd (%eax), %xmm5, %xmm1 13 vfmadd132ps %xmm2, %xmm5, %xmm1 17 vfmadd132ps (%eax), %xmm5, %xmm1 21 vfmadd213pd %xmm2, %xmm5, %xmm1 25 vfmadd213pd (%eax), %xmm5, %xmm1 29 vfmadd213ps %xmm2, %xmm5, %xmm1 33 vfmadd213ps (%eax), %xmm5, %xmm1 37 vfmadd231pd %xmm2, %xmm5, %xmm1 41 vfmadd231pd (%eax), %xmm5, %xmm1 [all …]
|
D | x86_64-fma4-encoding.s | 6 vfmaddss (%rcx), %xmm1, %xmm0, %xmm0 10 vfmaddss %xmm1, (%rcx),%xmm0, %xmm0 14 vfmaddss %xmm2, %xmm1, %xmm0, %xmm0 18 vfmaddsd (%rcx), %xmm1, %xmm0, %xmm0 22 vfmaddsd %xmm1, (%rcx),%xmm0, %xmm0 26 vfmaddsd %xmm2, %xmm1, %xmm0, %xmm0 30 vfmaddsd %xmm10, %xmm1, %xmm0, %xmm0 34 vfmaddps (%rcx), %xmm1, %xmm0, %xmm0 38 vfmaddps %xmm1, (%rcx),%xmm0, %xmm0 42 vfmaddps %xmm2, %xmm1, %xmm0, %xmm0 [all …]
|
D | x86-32-avx.s | 276 vunpckhps %xmm1, %xmm2, %xmm4 280 vunpckhpd %xmm1, %xmm2, %xmm4 284 vunpcklps %xmm1, %xmm2, %xmm4 288 vunpcklpd %xmm1, %xmm2, %xmm4 308 vcmpps $0, %xmm0, %xmm6, %xmm1 312 vcmpps $0, (%eax), %xmm6, %xmm1 316 vcmpps $7, %xmm0, %xmm6, %xmm1 320 vcmppd $0, %xmm0, %xmm6, %xmm1 324 vcmppd $0, (%eax), %xmm6, %xmm1 328 vcmppd $7, %xmm0, %xmm6, %xmm1 [all …]
|
D | x86_64-xop-encoding.s | 10 vphsubwd (%rcx,%rax), %xmm1 13 vphsubwd %xmm0, %xmm1 18 vphsubdq (%rcx,%rax), %xmm1 21 vphsubdq %xmm0, %xmm1 26 vphsubbw (%rax), %xmm1 29 vphsubbw %xmm2, %xmm1 61 vphadduwd %xmm2, %xmm1 106 vphaddbw (%rcx,%rax), %xmm1 114 vphaddbq (%rcx,%rax), %xmm1 122 vphaddbd (%rcx,%rax), %xmm1 [all …]
|
/external/llvm/test/CodeGen/X86/ |
D | vector-shuffle-128-v4.ll | 307 ; SSE2-NEXT: shufps {{.*#+}} xmm1 = xmm1[0,0],xmm0[2,0] 308 ; SSE2-NEXT: shufps {{.*#+}} xmm0 = xmm0[0,1],xmm1[2,0] 313 ; SSE3-NEXT: shufps {{.*#+}} xmm1 = xmm1[0,0],xmm0[2,0] 314 ; SSE3-NEXT: shufps {{.*#+}} xmm0 = xmm0[0,1],xmm1[2,0] 319 ; SSSE3-NEXT: shufps {{.*#+}} xmm1 = xmm1[0,0],xmm0[2,0] 320 ; SSSE3-NEXT: shufps {{.*#+}} xmm0 = xmm0[0,1],xmm1[2,0] 325 ; SSE41-NEXT: pshufd {{.*#+}} xmm1 = xmm1[0,1,2,0] 326 ; SSE41-NEXT: pblendw {{.*#+}} xmm0 = xmm0[0,1,2,3,4,5],xmm1[6,7] 331 ; AVX1-NEXT: vpshufd {{.*#+}} xmm1 = xmm1[0,1,2,0] 332 ; AVX1-NEXT: vpblendw {{.*#+}} xmm0 = xmm0[0,1,2,3,4,5],xmm1[6,7] [all …]
|
D | vector-zext.ll | 10 ; SSE2-NEXT: movdqa %xmm0, %xmm1 14 ; SSE2-NEXT: punpckhwd {{.*#+}} xmm1 = xmm1[4,4,5,5,6,6,7,7] 15 ; SSE2-NEXT: pand .LCPI0_0(%rip), %xmm1 20 ; SSSE3-NEXT: movdqa %xmm0, %xmm1 24 ; SSSE3-NEXT: punpckhwd {{.*#+}} xmm1 = xmm1[4,4,5,5,6,6,7,7] 25 ; SSSE3-NEXT: pand .LCPI0_0(%rip), %xmm1 30 ; SSE41-NEXT: movdqa %xmm0, %xmm1 31 ; SSE41-NEXT: pmovzxwd {{.*#+}} xmm0 = xmm1[0],zero,xmm1[1],zero,xmm1[2],zero,xmm1[3],zero 32 ; SSE41-NEXT: punpckhwd {{.*#+}} xmm1 = xmm1[4,4,5,5,6,6,7,7] 33 ; SSE41-NEXT: pand .LCPI0_0(%rip), %xmm1 [all …]
|
D | bswap-vector.ll | 16 ; CHECK-NOSSSE3-NEXT: pxor %xmm1, %xmm1 18 …m2 = xmm2[8],xmm1[8],xmm2[9],xmm1[9],xmm2[10],xmm1[10],xmm2[11],xmm1[11],xmm2[12],xmm1[12],xmm2[13… 21 …}} xmm0 = xmm0[0],xmm1[0],xmm0[1],xmm1[1],xmm0[2],xmm1[2],xmm0[3],xmm1[3],xmm0[4],xmm1[4],xmm0[5],… 49 ; CHECK-NOSSSE3-NEXT: pxor %xmm1, %xmm1 51 …m2 = xmm2[8],xmm1[8],xmm2[9],xmm1[9],xmm2[10],xmm1[10],xmm2[11],xmm1[11],xmm2[12],xmm1[12],xmm2[13… 54 …}} xmm0 = xmm0[0],xmm1[0],xmm0[1],xmm1[1],xmm0[2],xmm1[2],xmm0[3],xmm1[3],xmm0[4],xmm1[4],xmm0[5],… 82 ; CHECK-NOSSSE3-NEXT: pxor %xmm1, %xmm1 84 …m2 = xmm2[8],xmm1[8],xmm2[9],xmm1[9],xmm2[10],xmm1[10],xmm2[11],xmm1[11],xmm2[12],xmm1[12],xmm2[13… 88 …}} xmm0 = xmm0[0],xmm1[0],xmm0[1],xmm1[1],xmm0[2],xmm1[2],xmm0[3],xmm1[3],xmm0[4],xmm1[4],xmm0[5],… 130 ; CHECK-NOSSSE3-NEXT: movdqa %xmm1, %xmm3 [all …]
|
D | sse-minmax.ll | 16 ; CHECK-NEXT: maxsd %xmm1, %xmm0 19 ; UNSAFE-NEXT: maxsd %xmm1, %xmm0 22 ; FINITE-NEXT: maxsd %xmm1, %xmm0 31 ; CHECK-NEXT: minsd %xmm1, %xmm0 34 ; UNSAFE-NEXT: minsd %xmm1, %xmm0 37 ; FINITE-NEXT: minsd %xmm1, %xmm0 46 ; CHECK-NEXT: minsd %xmm0, %xmm1 47 ; CHECK-NEXT: movap{{[sd]}} %xmm1, %xmm0 50 ; UNSAFE-NEXT: minsd %xmm1, %xmm0 53 ; FINITE-NEXT: minsd %xmm0, %xmm1 [all …]
|
D | combine-and.ll | 11 ; CHECK-NEXT: pxor %xmm1, %xmm1 12 ; CHECK-NEXT: pblendw {{.*#+}} xmm0 = xmm0[0,1],xmm1[2,3,4,5,6,7] 21 ; CHECK-NEXT: pxor %xmm1, %xmm1 22 ; CHECK-NEXT: pblendw {{.*#+}} xmm0 = xmm1[0,1],xmm0[2,3],xmm1[4,5,6,7] 31 ; CHECK-NEXT: pxor %xmm1, %xmm1 32 ; CHECK-NEXT: pblendw {{.*#+}} xmm0 = xmm1[0,1,2,3],xmm0[4,5],xmm1[6,7] 41 ; CHECK-NEXT: pxor %xmm1, %xmm1 42 ; CHECK-NEXT: pblendw {{.*#+}} xmm0 = xmm1[0,1,2,3,4,5],xmm0[6,7] 51 ; CHECK-NEXT: pxor %xmm1, %xmm1 52 ; CHECK-NEXT: pblendw {{.*#+}} xmm0 = xmm0[0,1],xmm1[2,3],xmm0[4,5],xmm1[6,7] [all …]
|
D | vector-shuffle-combining.ll | 166 ; SSE-NEXT: pand %xmm1, %xmm0 172 ; AVX-NEXT: vpand %xmm1, %xmm0, %xmm0 184 ; SSE-NEXT: por %xmm1, %xmm0 190 ; AVX-NEXT: vpor %xmm1, %xmm0, %xmm0 202 ; SSE-NEXT: pxor %xmm1, %xmm0 208 ; AVX-NEXT: vpxor %xmm1, %xmm0, %xmm0 220 ; SSE-NEXT: pand %xmm1, %xmm0 226 ; AVX-NEXT: vpand %xmm1, %xmm0, %xmm0 238 ; SSE-NEXT: por %xmm1, %xmm0 244 ; AVX-NEXT: vpor %xmm1, %xmm0, %xmm0 [all …]
|
D | vector-shuffle-128-v2.ll | 58 ; SSE-NEXT: pshufd {{.*#+}} xmm0 = xmm1[0,1,0,1] 63 ; AVX1-NEXT: vpshufd {{.*#+}} xmm0 = xmm1[0,1,0,1] 68 ; AVX2-NEXT: vpbroadcastq %xmm1, %xmm0 76 ; SSE-NEXT: pshufd {{.*#+}} xmm0 = xmm1[2,3,0,1] 81 ; AVX-NEXT: vpshufd {{.*#+}} xmm0 = xmm1[2,3,0,1] 89 ; SSE-NEXT: pshufd {{.*#+}} xmm0 = xmm1[2,3,2,3] 94 ; AVX-NEXT: vpshufd {{.*#+}} xmm0 = xmm1[2,3,2,3] 157 ; SSE2-NEXT: movlhps {{.*#+}} xmm1 = xmm1[0,0] 158 ; SSE2-NEXT: movaps %xmm1, %xmm0 163 ; SSE3-NEXT: movddup {{.*#+}} xmm0 = xmm1[0,0] [all …]
|
D | vector-shuffle-256-v16.ll | 26 ; AVX1-NEXT: vpshufhw {{.*#+}} xmm1 = xmm0[0,1,2,3,4,4,4,4] 33 ; AVX2-NEXT: vpbroadcastw %xmm0, %xmm1 44 ; AVX1-NEXT: vpshufb {{.*#+}} xmm1 = xmm0[0,1,0,1,0,1,0,1,0,1,0,1,0,1,0,1] 51 ; AVX2-NEXT: vpbroadcastw %xmm0, %xmm1 62 ; AVX1-NEXT: vpshufb {{.*#+}} xmm1 = xmm0[0,1,0,1,0,1,0,1,0,1,0,1,0,1,0,1] 69 ; AVX2-NEXT: vpbroadcastw %xmm0, %xmm1 80 ; AVX1-NEXT: vpshufb {{.*#+}} xmm1 = xmm0[0,1,0,1,0,1,0,1,0,1,0,1,0,1,0,1] 87 ; AVX2-NEXT: vpbroadcastw %xmm0, %xmm1 98 ; AVX1-NEXT: vpshufb {{.*#+}} xmm1 = xmm0[0,1,0,1,0,1,0,1,0,1,0,1,0,1,0,1] 105 ; AVX2-NEXT: vpbroadcastw %xmm0, %xmm1 [all …]
|
D | vector-shuffle-128-v16.ll | 29 ; SSSE3-NEXT: pxor %xmm1, %xmm1 30 ; SSSE3-NEXT: pshufb %xmm1, %xmm0 35 ; SSE41-NEXT: pxor %xmm1, %xmm1 36 ; SSE41-NEXT: pshufb %xmm1, %xmm0 41 ; AVX1-NEXT: vpxor %xmm1, %xmm1, %xmm1 42 ; AVX1-NEXT: vpshufb %xmm1, %xmm0, %xmm0 224 …}} xmm0 = xmm0[0],xmm1[0],xmm0[1],xmm1[1],xmm0[2],xmm1[2],xmm0[3],xmm1[3],xmm0[4],xmm1[4],xmm0[5],… 229 …}} xmm0 = xmm0[0],xmm1[0],xmm0[1],xmm1[1],xmm0[2],xmm1[2],xmm0[3],xmm1[3],xmm0[4],xmm1[4],xmm0[5],… 238 …m0 = xmm0[8],xmm1[8],xmm0[9],xmm1[9],xmm0[10],xmm1[10],xmm0[11],xmm1[11],xmm0[12],xmm1[12],xmm0[13… 243 …m0 = xmm0[8],xmm1[8],xmm0[9],xmm1[9],xmm0[10],xmm1[10],xmm0[11],xmm1[11],xmm0[12],xmm1[12],xmm0[13… [all …]
|
D | vector-sext.ll | 13 ; SSE2-NEXT: movdqa %xmm0, %xmm1 18 ; SSE2-NEXT: punpckhwd {{.*#+}} xmm1 = xmm1[4,4,5,5,6,6,7,7] 19 ; SSE2-NEXT: pslld $16, %xmm1 20 ; SSE2-NEXT: psrad $16, %xmm1 25 ; SSSE3-NEXT: movdqa %xmm0, %xmm1 30 ; SSSE3-NEXT: punpckhwd {{.*#+}} xmm1 = xmm1[4,4,5,5,6,6,7,7] 31 ; SSSE3-NEXT: pslld $16, %xmm1 32 ; SSSE3-NEXT: psrad $16, %xmm1 37 ; SSE41-NEXT: movdqa %xmm0, %xmm1 38 ; SSE41-NEXT: pmovzxwd %xmm1, %xmm0 [all …]
|
D | vec_setcc.ll | 10 ; SSE2: pmaxub %xmm0, %xmm1 11 ; SSE2: pcmpeqb %xmm1, %xmm0 14 ; SSE41: pmaxub %xmm0, %xmm1 15 ; SSE41: pcmpeqb %xmm1, %xmm0 18 ; AVX: vpmaxub %xmm1, %xmm0, %xmm1 19 ; AVX: vpcmpeqb %xmm1, %xmm0, %xmm0 27 ; SSE2: pminub %xmm0, %xmm1 28 ; SSE2: pcmpeqb %xmm1, %xmm0 31 ; SSE41: pminub %xmm0, %xmm1 32 ; SSE41: pcmpeqb %xmm1, %xmm0 [all …]
|
D | vector-idiv.ll | 10 ; SSE41-NEXT: movdqa {{.*#+}} xmm1 = [613566757,613566757,613566757,613566757] 11 ; SSE41-NEXT: pshufd {{.*#+}} xmm2 = xmm1[1,1,3,3] 14 ; SSE41-NEXT: pmuludq %xmm0, %xmm1 15 ; SSE41-NEXT: pshufd {{.*#+}} xmm1 = xmm1[1,1,3,3] 16 ; SSE41-NEXT: pblendw {{.*#+}} xmm1 = xmm1[0,1],xmm3[2,3],xmm1[4,5],xmm3[6,7] 17 ; SSE41-NEXT: psubd %xmm1, %xmm0 19 ; SSE41-NEXT: paddd %xmm1, %xmm0 25 ; SSE-NEXT: movdqa {{.*#+}} xmm1 = [613566757,613566757,613566757,613566757] 27 ; SSE-NEXT: pmuludq %xmm1, %xmm2 29 ; SSE-NEXT: pshufd {{.*#+}} xmm1 = xmm1[1,1,3,3] [all …]
|
D | vector-blend.ll | 12 ; SSE2-NEXT: shufps {{.*#+}} xmm0 = xmm0[0,2],xmm1[1,3] 18 ; SSSE3-NEXT: shufps {{.*#+}} xmm0 = xmm0[0,2],xmm1[1,3] 24 ; SSE41-NEXT: blendps {{.*#+}} xmm0 = xmm0[0],xmm1[1],xmm0[2],xmm1[3] 29 ; AVX-NEXT: vblendps {{.*#+}} xmm0 = xmm0[0],xmm1[1],xmm0[2],xmm1[3] 39 ; SSE2-NEXT: movss {{.*#+}} xmm1 = xmm0[0],xmm1[1,2,3] 40 ; SSE2-NEXT: movaps %xmm1, %xmm0 45 ; SSSE3-NEXT: movss {{.*#+}} xmm1 = xmm0[0],xmm1[1,2,3] 46 ; SSSE3-NEXT: movaps %xmm1, %xmm0 51 ; SSE41-NEXT: blendps {{.*#+}} xmm0 = xmm0[0],xmm1[1,2,3] 56 ; AVX-NEXT: vblendps {{.*#+}} xmm0 = xmm0[0],xmm1[1,2,3] [all …]
|
D | palignr.ll | 21 ; CHECK-NEXT: palignr {{.*#+}} xmm1 = xmm0[4,5,6,7,8,9,10,11,12,13,14,15],xmm1[0,1,2,3] 22 ; CHECK-NEXT: movdqa %xmm1, %xmm0 27 ; CHECK-YONAH-NEXT: shufps {{.*#+}} xmm1 = xmm1[0,0],xmm0[3,0] 28 ; CHECK-YONAH-NEXT: shufps {{.*#+}} xmm0 = xmm0[1,2],xmm1[2,0] 37 ; CHECK-NEXT: palignr {{.*#+}} xmm1 = xmm0[4,5,6,7,8,9,10,11,12,13,14,15],xmm1[0,1,2,3] 38 ; CHECK-NEXT: movdqa %xmm1, %xmm0 43 ; CHECK-YONAH-NEXT: pshufd {{.*#+}} xmm1 = xmm1[0,0,1,1] 45 ; CHECK-YONAH-NEXT: punpcklqdq {{.*#+}} xmm0 = xmm0[0],xmm1[0] 54 ; CHECK-NEXT: palignr {{.*#+}} xmm0 = xmm1[8,9,10,11,12,13,14,15],xmm0[0,1,2,3,4,5,6,7] 59 ; CHECK-YONAH-NEXT: shufpd {{.*#+}} xmm1 = xmm1[1],xmm0[0] [all …]
|
/external/boringssl/src/crypto/aes/asm/ |
D | vpaes-x86_64.pl | 79 ## Clobbers %xmm1-%xmm5, %r9, %r10, %r11, %rax 89 movdqa %xmm9, %xmm1 91 pandn %xmm0, %xmm1 93 psrld \$4, %xmm1 97 pshufb %xmm1, %xmm0 114 movdqa -0x40(%r11,%r10), %xmm1 # .Lk_mc_forward[] 121 pshufb %xmm1, %xmm0 # 0 = B 127 pshufb %xmm1, %xmm0 # 0 = 2B+C 134 movdqa %xmm9, %xmm1 # 1 : i 136 pandn %xmm0, %xmm1 # 1 = i<<4 [all …]
|
/external/libyuv/files/source/ |
D | row_win.cc | 139 movdqa xmm1, xmm0 in I400ToARGBRow_SSE2() 141 punpckhwd xmm1, xmm1 in I400ToARGBRow_SSE2() 143 por xmm1, xmm5 in I400ToARGBRow_SSE2() 145 movdqa [edx + 16], xmm1 in I400ToARGBRow_SSE2() 250 movdqu xmm1, [eax + 16] in RGB24ToARGBRow_SSSE3() 254 palignr xmm2, xmm1, 8 // xmm2 = { xmm3[0:3] xmm1[8:15]} in RGB24ToARGBRow_SSSE3() 257 palignr xmm1, xmm0, 12 // xmm1 = { xmm3[0:7] xmm0[12:15]} in RGB24ToARGBRow_SSSE3() 261 pshufb xmm1, xmm4 in RGB24ToARGBRow_SSSE3() 263 por xmm1, xmm5 in RGB24ToARGBRow_SSSE3() 266 movdqa [edx + 16], xmm1 in RGB24ToARGBRow_SSSE3() [all …]
|
/external/libvpx/libvpx/third_party/libyuv/source/ |
D | compare_win.cc | 32 movdqa xmm1, [eax] in SumSquareError_SSE2() 37 movdqa xmm3, xmm1 // abs trick in SumSquareError_SSE2() 38 psubusb xmm1, xmm2 in SumSquareError_SSE2() 40 por xmm1, xmm2 in SumSquareError_SSE2() 41 movdqa xmm2, xmm1 in SumSquareError_SSE2() 42 punpcklbw xmm1, xmm5 in SumSquareError_SSE2() 44 pmaddwd xmm1, xmm1 in SumSquareError_SSE2() 46 paddd xmm0, xmm1 in SumSquareError_SSE2() 50 pshufd xmm1, xmm0, 0xee in SumSquareError_SSE2() 51 paddd xmm0, xmm1 in SumSquareError_SSE2() [all …]
|
/external/libvpx/libvpx/vp8/common/x86/ |
D | variance_impl_ssse3.asm | 68 movdqu xmm1, XMMWORD PTR [rsi+1] 71 punpcklbw xmm0, xmm1 72 punpckhbw xmm2, xmm1 92 movdqu xmm1, XMMWORD PTR [rsi] 94 movdqa xmm3, xmm1 96 punpcklbw xmm1, xmm2 98 pmaddubsw xmm1, [rax] 101 paddw xmm1, [GLOBAL(xmm_bi_rd)] 103 psraw xmm1, xmm_filter_shift 105 packuswb xmm1, xmm3 [all …]
|
/external/boringssl/mac-x86/crypto/aes/ |
D | aesni-x86.S | 15 movups 16(%edx),%xmm1 21 movups (%edx),%xmm1 26 pxor %xmm1,%xmm1 41 movups 16(%edx),%xmm1 47 movups (%edx),%xmm1 52 pxor %xmm1,%xmm1 61 movups 16(%edx),%xmm1 71 movups (%edx,%ecx,1),%xmm1 87 movups 16(%edx),%xmm1 97 movups (%edx,%ecx,1),%xmm1 [all …]
|
/external/boringssl/linux-x86/crypto/aes/ |
D | aesni-x86.S | 16 movups 16(%edx),%xmm1 22 movups (%edx),%xmm1 27 pxor %xmm1,%xmm1 44 movups 16(%edx),%xmm1 50 movups (%edx),%xmm1 55 pxor %xmm1,%xmm1 66 movups 16(%edx),%xmm1 76 movups (%edx,%ecx,1),%xmm1 94 movups 16(%edx),%xmm1 104 movups (%edx,%ecx,1),%xmm1 [all …]
|