/external/boringssl/mac-x86_64/crypto/aes/ |
D | bsaes-x86_64.S | 26 pxor %xmm8,%xmm6 38 pxor %xmm6,%xmm5 42 pxor %xmm5,%xmm6 67 pxor %xmm6,%xmm4 71 pxor %xmm4,%xmm6 95 pxor %xmm6,%xmm2 99 pxor %xmm2,%xmm6 134 pxor 112(%rax),%xmm6 148 pxor %xmm6,%xmm2 149 pxor %xmm4,%xmm6 [all …]
|
/external/boringssl/linux-x86_64/crypto/aes/ |
D | bsaes-x86_64.S | 28 pxor %xmm8,%xmm6 40 pxor %xmm6,%xmm5 44 pxor %xmm5,%xmm6 69 pxor %xmm6,%xmm4 73 pxor %xmm4,%xmm6 97 pxor %xmm6,%xmm2 101 pxor %xmm2,%xmm6 136 pxor 112(%rax),%xmm6 150 pxor %xmm6,%xmm2 151 pxor %xmm4,%xmm6 [all …]
|
/external/boringssl/win-x86_64/crypto/aes/ |
D | bsaes-x86_64.asm | 30 pxor xmm6,xmm8 42 pxor xmm5,xmm6 46 pxor xmm6,xmm5 71 pxor xmm4,xmm6 75 pxor xmm6,xmm4 99 pxor xmm2,xmm6 103 pxor xmm6,xmm2 138 pxor xmm6,XMMWORD[112+rax] 152 pxor xmm2,xmm6 153 pxor xmm6,xmm4 [all …]
|
/external/boringssl/mac-x86_64/crypto/modes/ |
D | aesni-gcm-x86_64.S | 34 vpclmulqdq $0x01,%xmm3,%xmm7,%xmm6 67 vpxor %xmm5,%xmm6,%xmm6 95 vpxor %xmm1,%xmm6,%xmm6 98 vpxor %xmm2,%xmm6,%xmm6 113 vpxor %xmm2,%xmm6,%xmm6 116 vpxor %xmm3,%xmm6,%xmm6 135 vpxor %xmm3,%xmm6,%xmm6 138 vpxor %xmm5,%xmm6,%xmm6 157 vpxor %xmm5,%xmm6,%xmm6 160 vpxor %xmm1,%xmm6,%xmm6 [all …]
|
/external/boringssl/linux-x86_64/crypto/modes/ |
D | aesni-gcm-x86_64.S | 34 vpclmulqdq $0x01,%xmm3,%xmm7,%xmm6 67 vpxor %xmm5,%xmm6,%xmm6 95 vpxor %xmm1,%xmm6,%xmm6 98 vpxor %xmm2,%xmm6,%xmm6 113 vpxor %xmm2,%xmm6,%xmm6 116 vpxor %xmm3,%xmm6,%xmm6 135 vpxor %xmm3,%xmm6,%xmm6 138 vpxor %xmm5,%xmm6,%xmm6 157 vpxor %xmm5,%xmm6,%xmm6 160 vpxor %xmm1,%xmm6,%xmm6 [all …]
|
/external/libvpx/libvpx/third_party/libyuv/source/ |
D | rotate_win.cc | 57 movq xmm6, qword ptr [eax] in TransposeWx8_SSSE3() 60 punpcklbw xmm6, xmm7 in TransposeWx8_SSSE3() 62 movdqa xmm7, xmm6 in TransposeWx8_SSSE3() 71 punpcklwd xmm4, xmm6 in TransposeWx8_SSSE3() 73 movdqa xmm6, xmm4 in TransposeWx8_SSSE3() 75 palignr xmm6, xmm6, 8 in TransposeWx8_SSSE3() 85 punpckldq xmm2, xmm6 in TransposeWx8_SSSE3() 86 movdqa xmm6, xmm2 in TransposeWx8_SSSE3() 87 palignr xmm6, xmm6, 8 in TransposeWx8_SSSE3() 90 movq qword ptr [edx + esi], xmm6 in TransposeWx8_SSSE3() [all …]
|
/external/libyuv/files/source/ |
D | rotate_win.cc | 59 movq xmm6, qword ptr [eax] in TransposeWx8_SSSE3() 62 punpcklbw xmm6, xmm7 in TransposeWx8_SSSE3() 64 movdqa xmm7, xmm6 in TransposeWx8_SSSE3() 73 punpcklwd xmm4, xmm6 in TransposeWx8_SSSE3() 75 movdqa xmm6, xmm4 in TransposeWx8_SSSE3() 77 palignr xmm6, xmm6, 8 in TransposeWx8_SSSE3() 87 punpckldq xmm2, xmm6 in TransposeWx8_SSSE3() 88 movdqa xmm6, xmm2 in TransposeWx8_SSSE3() 89 palignr xmm6, xmm6, 8 in TransposeWx8_SSSE3() 92 movq qword ptr [edx + esi], xmm6 in TransposeWx8_SSSE3() [all …]
|
/external/boringssl/src/crypto/aes/asm/ |
D | vpaes-x86_64.pl | 81 ## Preserves %xmm6 - %xmm8 so you get some local vectors 345 ## the high bits of %xmm6. 356 movdqa %xmm0, %xmm6 # save short part 358 movhlps %xmm4, %xmm6 # clobber low side with zeros 363 palignr \$8,%xmm6,%xmm0 381 ## %xmm6. The low side's rounds are the same as the 392 movdqa %xmm0, %xmm6 # save cur_lo in xmm6 403 movdqa %xmm6, %xmm7 446 pxor %xmm6, %xmm6 458 ## %xmm6: low side, d c 0 0 [all …]
|
/external/libvpx/libvpx/vpx_dsp/x86/ |
D | highbd_variance_impl_sse2.asm | 66 pxor xmm6, xmm6 ; clear xmm6 for accumulating sse 91 paddd xmm6, xmm1 98 paddd xmm6, xmm3 105 paddd xmm6, xmm1 110 paddd xmm6, xmm3 129 movdqa xmm4, xmm6 130 punpckldq xmm6, xmm0 135 paddd xmm6, xmm4 141 movdqa xmm4, xmm6 147 paddd xmm6, xmm4 [all …]
|
/external/boringssl/win-x86_64/crypto/modes/ |
D | aesni-gcm-x86_64.asm | 38 vpclmulqdq xmm6,xmm7,xmm3,0x01 71 vpxor xmm6,xmm6,xmm5 99 vpxor xmm6,xmm6,xmm1 102 vpxor xmm6,xmm6,xmm2 117 vpxor xmm6,xmm6,xmm2 120 vpxor xmm6,xmm6,xmm3 139 vpxor xmm6,xmm6,xmm3 142 vpxor xmm6,xmm6,xmm5 161 vpxor xmm6,xmm6,xmm5 164 vpxor xmm6,xmm6,xmm1 [all …]
|
/external/boringssl/linux-x86/crypto/chacha/ |
D | chacha-x86.S | 415 pshufd $170,%xmm7,%xmm6 424 movdqa %xmm6,-32(%ebp) 434 pshufd $170,%xmm7,%xmm6 442 movdqa %xmm6,-96(%ebp) 453 movdqa -32(%ebp),%xmm6 459 movdqa %xmm6,-32(%ebx) 465 movdqa 96(%ebp),%xmm6 472 movdqa %xmm6,96(%ebx) 476 movdqa %xmm4,%xmm6 486 pxor %xmm0,%xmm6 [all …]
|
/external/boringssl/mac-x86/crypto/chacha/ |
D | chacha-x86.S | 412 pshufd $170,%xmm7,%xmm6 421 movdqa %xmm6,-32(%ebp) 431 pshufd $170,%xmm7,%xmm6 439 movdqa %xmm6,-96(%ebp) 450 movdqa -32(%ebp),%xmm6 456 movdqa %xmm6,-32(%ebx) 462 movdqa 96(%ebp),%xmm6 469 movdqa %xmm6,96(%ebx) 473 movdqa %xmm4,%xmm6 483 pxor %xmm0,%xmm6 [all …]
|
/external/llvm/test/MC/X86/ |
D | x86_64-xop-encoding.s | 37 vphaddwq %xmm6, %xmm2 50 vphadduwq (%rcx,%rax), %xmm6 69 vphaddudq %xmm6, %xmm2 109 vphaddbw %xmm5, %xmm6 149 vfrczps %xmm6, %xmm5 191 vpshlq %xmm2, %xmm4, %xmm6 197 vpshlq %xmm5, (%rdx,%rcx), %xmm6 241 vpshaq %xmm6, (%rax,%rcx), %xmm5 268 vprotw (%rax), %xmm3, %xmm6 317 vprotd $43, (%rcx), %xmm6 [all …]
|
/external/libvpx/libvpx/vp8/common/x86/ |
D | loopfilter_sse2.asm | 54 movdqa xmm6, xmm1 ; q2 58 psubusb xmm2, xmm6 ; q3-=q2 60 psubusb xmm4, xmm6 ; q1-=q2 61 psubusb xmm6, xmm3 ; q2-=q1 63 por xmm4, xmm6 ; abs(q2-q1) 80 movdqa xmm6, [rsi+2*rax] ; p1 84 movlps xmm6, [rsi + rcx] ; p1 88 movhps xmm6, [rdi + rcx] 91 movdqa [rsp+_p1], xmm6 ; store p1 95 movdqa xmm3, xmm6 ; p1 [all …]
|
D | loopfilter_block_sse2_x86_64.asm | 171 %define scratch2 xmm6 313 %define scratch2 xmm6 390 movdqa xmm6, s14 391 movdqa xmm7, xmm6 392 punpcklbw xmm6, s15 ; e0 f0 396 punpcklwd xmm3, xmm6 ; c0 d0 e0 f0 397 punpckhwd xmm8, xmm6 ; c4 d4 e4 f4 399 movdqa xmm6, xmm5 401 punpckhwd xmm6, xmm7 ; cc dc ec fc 418 punpckldq xmm2, xmm6 ; 8c 9c ac bc cc dc ec fc [all …]
|
D | idctllm_sse2.asm | 188 movdqa xmm6, xmm2 ; a1 196 psubw xmm6, xmm3 ;3 204 punpcklwd xmm4, xmm6 ; 015 011 014 010 013 009 012 008 205 punpckhwd xmm5, xmm6 ; 115 111 114 110 113 109 112 108 212 movdqa xmm6, xmm7 ; 107 103 106 102 105 101 104 100 214 punpckhdq xmm6, xmm5 ; 115 111 107 103 114 110 106 102 222 punpckldq xmm1, xmm6 ; 114 110 014 010 106 102 006 002 223 punpckhdq xmm7, xmm6 ; 115 111 015 011 107 103 007 003 260 movdqa xmm6, xmm2 ; a1 268 psubw xmm6, xmm3 ;3 [all …]
|
D | subpixel_ssse3.asm | 63 movdqa xmm6, XMMWORD PTR [rax+128] ;k1_k3 90 pmaddubsw xmm2, xmm6 118 movdqa xmm6, XMMWORD PTR [rax+128] ;k1_k3 146 pmaddubsw xmm2, xmm6 204 movdqa xmm6, XMMWORD PTR [rax+128] ;k1_k3 229 pmaddubsw xmm2, xmm6 247 pmaddubsw xmm2, xmm6 311 movdqa xmm6, XMMWORD PTR [rax+128] ;k1_k3 334 pmaddubsw xmm2, xmm6 362 movdqa xmm6, XMMWORD PTR [rax+128] ;k1_k3 [all …]
|
/external/boringssl/win-x86/crypto/chacha/ |
D | chacha-x86.asm | 423 pshufd xmm6,xmm7,170 432 movdqa [ebp-32],xmm6 442 pshufd xmm6,xmm7,170 450 movdqa [ebp-96],xmm6 461 movdqa xmm6,[ebp-32] 467 movdqa [ebx-32],xmm6 473 movdqa xmm6,[96+ebp] 480 movdqa [96+ebx],xmm6 484 movdqa xmm6,xmm4 494 pxor xmm6,xmm0 [all …]
|
/external/libjpeg-turbo/simd/ |
D | jccolext-sse2-64.asm | 299 movdqa xmm6,xmm1 301 punpckhwd xmm6,xmm3 303 movdqa xmm4,xmm6 305 pmaddwd xmm6,[rel PW_F0299_F0337] ; xmm6=ROH*FIX(0.299)+GOH*FIX(0.337) 310 movdqa XMMWORD [wk(5)], xmm6 ; wk(5)=ROH*FIX(0.299)+GOH*FIX(0.337) 313 pxor xmm6,xmm6 315 punpckhwd xmm6,xmm5 ; xmm6=BOH 317 psrld xmm6,1 ; xmm6=BOH*FIX(0.500) 322 paddd xmm4,xmm6 331 movdqa xmm6,xmm0 [all …]
|
D | jccolext-sse2.asm | 312 movdqa xmm6,xmm1 314 punpckhwd xmm6,xmm3 316 movdqa xmm4,xmm6 318 pmaddwd xmm6,[GOTOFF(eax,PW_F0299_F0337)] ; xmm6=ROH*FIX(0.299)+GOH*FIX(0.337) 323 movdqa XMMWORD [wk(5)], xmm6 ; wk(5)=ROH*FIX(0.299)+GOH*FIX(0.337) 326 pxor xmm6,xmm6 328 punpckhwd xmm6,xmm5 ; xmm6=BOH 330 psrld xmm6,1 ; xmm6=BOH*FIX(0.500) 335 paddd xmm4,xmm6 344 movdqa xmm6,xmm0 [all …]
|
/external/llvm/test/CodeGen/X86/ |
D | vec_minmax_uint.ll | 84 ; SSE2-NEXT: movdqa %xmm1, %xmm6 85 ; SSE2-NEXT: pxor %xmm4, %xmm6 86 ; SSE2-NEXT: movdqa %xmm6, %xmm7 89 ; SSE2-NEXT: pcmpeqd %xmm5, %xmm6 90 ; SSE2-NEXT: pshufd {{.*#+}} xmm5 = xmm6[1,1,3,3] 92 ; SSE2-NEXT: pshufd {{.*#+}} xmm6 = xmm7[1,1,3,3] 93 ; SSE2-NEXT: por %xmm5, %xmm6 108 ; SSE2-NEXT: pand %xmm6, %xmm1 109 ; SSE2-NEXT: pandn %xmm3, %xmm6 110 ; SSE2-NEXT: por %xmm6, %xmm1 [all …]
|
/external/boringssl/linux-x86/crypto/aes/ |
D | aesni-x86.S | 272 pxor %xmm0,%xmm6 327 pxor %xmm0,%xmm6 398 movdqu 64(%esi),%xmm6 413 movups %xmm6,64(%edi) 414 movdqu 64(%esi),%xmm6 429 movups %xmm6,64(%edi) 445 movups 64(%esi),%xmm6 452 movups %xmm6,64(%edi) 500 movdqu 64(%esi),%xmm6 515 movups %xmm6,64(%edi) [all …]
|
/external/boringssl/mac-x86/crypto/aes/ |
D | aesni-x86.S | 255 pxor %xmm0,%xmm6 308 pxor %xmm0,%xmm6 377 movdqu 64(%esi),%xmm6 392 movups %xmm6,64(%edi) 393 movdqu 64(%esi),%xmm6 408 movups %xmm6,64(%edi) 424 movups 64(%esi),%xmm6 431 movups %xmm6,64(%edi) 479 movdqu 64(%esi),%xmm6 494 movups %xmm6,64(%edi) [all …]
|
/external/libvpx/libvpx/vp8/encoder/x86/ |
D | fwalsh_sse2.asm | 55 pxor xmm6, xmm6 56 movq xmm6, xmm0 58 pcmpeqw xmm7, xmm6 86 pshufd xmm6, xmm1, 0x72 ; d13 d12 a13 a12 92 movdqa xmm1, xmm6 94 punpckhqdq xmm6, xmm7 ; c13 c12 d13 d12 100 paddd xmm1, xmm6 ; b23 b22 a23 a22 101 psubd xmm3, xmm6 ; c23 c22 d23 d22 110 pxor xmm6, xmm6 111 movdqa xmm7, xmm6 [all …]
|
/external/boringssl/win-x86/crypto/aes/ |
D | aesni-x86.asm | 260 pxor xmm6,xmm0 312 pxor xmm6,xmm0 380 movdqu xmm6,[64+esi] 395 movups [64+edi],xmm6 396 movdqu xmm6,[64+esi] 411 movups [64+edi],xmm6 427 movups xmm6,[64+esi] 434 movups [64+edi],xmm6 482 movdqu xmm6,[64+esi] 497 movups [64+edi],xmm6 [all …]
|