/external/llvm/test/CodeGen/X86/ |
D | sad.ll | 169 ; SSE2-NEXT: pshufd {{.*#+}} xmm8 = xmm1[2,3,0,1] 170 …w {{.*#+}} xmm8 = xmm8[0],xmm12[0],xmm8[1],xmm12[1],xmm8[2],xmm12[2],xmm8[3],xmm12[3],xmm8[4],xmm1… 204 ; SSE2-NEXT: movdqa %xmm8, %xmm3 205 ; SSE2-NEXT: punpckhwd {{.*#+}} xmm8 = xmm8[4],xmm12[4],xmm8[5],xmm12[5],xmm8[6],xmm12[6],xmm8[7… 212 ; SSE2-NEXT: psubd %xmm11, %xmm8 220 ; SSE2-NEXT: movdqa %xmm8, %xmm4 222 ; SSE2-NEXT: paddd %xmm4, %xmm8 223 ; SSE2-NEXT: pxor %xmm4, %xmm8 261 ; SSE2-NEXT: paddd %xmm8, %xmm1 402 ; SSE2-NEXT: pxor %xmm8, %xmm8 [all …]
|
D | stack-folding-fp-sse42.ll | 14 …m sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~… 22 …m sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~… 30 …m sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~… 38 …m sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~… 47 …m sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~… 55 …m sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~… 64 …m sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~… 73 …m sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~… 82 …m sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~… 96 …m sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~… [all …]
|
D | stack-folding-fp-avx1.ll | 14 …m sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~… 22 …m sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~… 30 …m sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~… 38 …m sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~… 46 …m sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~… 54 …m sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~… 63 …m sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~… 71 …m sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~… 80 …m sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~… 89 …m sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~… [all …]
|
D | stack-folding-int-avx2.ll | 14 …fect "nop", "=x,~{xmm1},~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~… 25 …fect "nop", "=x,~{xmm1},~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~… 36 …fect "nop", "=x,~{xmm1},~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~… 50 …fect "nop", "=x,~{xmm1},~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~… 57 …m sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~… 67 …m sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~… 76 …fect "nop", "=x,~{xmm1},~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~… 85 …fect "nop", "=x,~{xmm1},~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~… 94 …fect "nop", "=x,~{xmm1},~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~… 103 …m sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~… [all …]
|
D | stack-folding-int-avx1.ll | 14 …m sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~… 23 …m sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~… 32 …m sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~… 41 …m sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~… 50 …fect "nop", "=x,~{xmm1},~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~… 59 …fect "nop", "=x,~{xmm1},~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~… 88 …fect "nop", "=x,~{xmm1},~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~… 108 …m sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~… 117 …fect "nop", "=x,~{xmm1},~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~… 126 …fect "nop", "=x,~{xmm1},~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~… [all …]
|
/external/boringssl/mac-x86_64/crypto/modes/ |
D | aesni-gcm-x86_64.S | 69 vpxor %xmm4,%xmm8,%xmm8 80 vpxor 16+8(%rsp),%xmm8,%xmm8 144 vpxor 112+8(%rsp),%xmm8,%xmm8 158 vpclmulqdq $0x10,%xmm3,%xmm8,%xmm5 161 vpclmulqdq $0x01,%xmm3,%xmm8,%xmm1 165 vpclmulqdq $0x00,%xmm3,%xmm8,%xmm2 168 vpclmulqdq $0x11,%xmm3,%xmm8,%xmm8 183 vpxor %xmm8,%xmm7,%xmm7 274 vpalignr $8,%xmm4,%xmm4,%xmm8 327 vpxor 16+8(%rsp),%xmm8,%xmm8 [all …]
|
D | ghash-x86_64.S | 931 movdqu 0(%rdx),%xmm8 936 pxor %xmm8,%xmm0 940 pshufd $78,%xmm0,%xmm8 941 pxor %xmm0,%xmm8 968 xorps %xmm4,%xmm8 972 pxor %xmm0,%xmm8 974 pxor %xmm1,%xmm8 976 movdqa %xmm8,%xmm9 978 pslldq $8,%xmm8 980 pxor %xmm8,%xmm0 [all …]
|
/external/boringssl/linux-x86_64/crypto/modes/ |
D | aesni-gcm-x86_64.S | 69 vpxor %xmm4,%xmm8,%xmm8 80 vpxor 16+8(%rsp),%xmm8,%xmm8 144 vpxor 112+8(%rsp),%xmm8,%xmm8 158 vpclmulqdq $0x10,%xmm3,%xmm8,%xmm5 161 vpclmulqdq $0x01,%xmm3,%xmm8,%xmm1 165 vpclmulqdq $0x00,%xmm3,%xmm8,%xmm2 168 vpclmulqdq $0x11,%xmm3,%xmm8,%xmm8 183 vpxor %xmm8,%xmm7,%xmm7 274 vpalignr $8,%xmm4,%xmm4,%xmm8 327 vpxor 16+8(%rsp),%xmm8,%xmm8 [all …]
|
D | ghash-x86_64.S | 932 movdqu 0(%rdx),%xmm8 937 pxor %xmm8,%xmm0 941 pshufd $78,%xmm0,%xmm8 942 pxor %xmm0,%xmm8 969 xorps %xmm4,%xmm8 973 pxor %xmm0,%xmm8 975 pxor %xmm1,%xmm8 977 movdqa %xmm8,%xmm9 979 pslldq $8,%xmm8 981 pxor %xmm8,%xmm0 [all …]
|
/external/libvpx/libvpx/vp8/common/x86/ |
D | loopfilter_block_sse2_x86_64.asm | 204 movdqa xmm8, i5 207 LF_FILTER_HEV_MASK xmm0, xmm1, xmm2, xmm3, xmm4, xmm8, xmm9, xmm10 212 movdqa xmm8, i5 213 LF_FILTER xmm1, xmm2, xmm3, xmm8, xmm0, xmm4 219 movdqa i5, xmm8 227 LF_FILTER_HEV_MASK xmm3, xmm8, xmm0, xmm1, xmm2, xmm4, xmm10, xmm11, xmm9 232 movdqa xmm8, i9 233 LF_FILTER xmm0, xmm1, xmm4, xmm8, xmm3, xmm2 239 movdqa i9, xmm8 247 LF_FILTER_HEV_MASK xmm4, xmm8, xmm0, xmm1, xmm2, xmm3, xmm9, xmm11, xmm10 [all …]
|
/external/boringssl/win-x86_64/crypto/modes/ |
D | aesni-gcm-x86_64.asm | 73 vpxor xmm8,xmm8,xmm4 84 vpxor xmm8,xmm8,XMMWORD[((16+8))+rsp] 148 vpxor xmm8,xmm8,XMMWORD[((112+8))+rsp] 162 vpclmulqdq xmm5,xmm8,xmm3,0x10 165 vpclmulqdq xmm1,xmm8,xmm3,0x01 169 vpclmulqdq xmm2,xmm8,xmm3,0x00 172 vpclmulqdq xmm8,xmm8,xmm3,0x11 187 vpxor xmm7,xmm7,xmm8 278 vpalignr xmm8,xmm4,xmm4,8 331 vpxor xmm8,xmm8,XMMWORD[((16+8))+rsp] [all …]
|
D | ghash-x86_64.asm | 973 movdqu xmm8,XMMWORD[r8] 978 pxor xmm0,xmm8 982 pshufd xmm8,xmm0,78 983 pxor xmm8,xmm0 1010 xorps xmm8,xmm4 1014 pxor xmm8,xmm0 1016 pxor xmm8,xmm1 1018 movdqa xmm9,xmm8 1020 pslldq xmm8,8 1022 pxor xmm0,xmm8 [all …]
|
/external/boringssl/mac-x86_64/crypto/aes/ |
D | bsaes-x86_64.S | 12 movdqa (%rax),%xmm8 15 pxor %xmm8,%xmm15 16 pxor %xmm8,%xmm0 17 pxor %xmm8,%xmm1 18 pxor %xmm8,%xmm2 21 pxor %xmm8,%xmm3 22 pxor %xmm8,%xmm4 25 pxor %xmm8,%xmm5 26 pxor %xmm8,%xmm6 33 movdqa 16(%r11),%xmm8 [all …]
|
D | aesni-x86_64.S | 382 pxor %xmm0,%xmm8 446 pxor %xmm0,%xmm8 519 movdqu 96(%rdi),%xmm8 540 movups %xmm8,96(%rsi) 541 movdqu 96(%rdi),%xmm8 561 movups %xmm8,96(%rsi) 583 movdqu 96(%rdi),%xmm8 592 movups %xmm8,96(%rsi) 662 movdqu 96(%rdi),%xmm8 683 movups %xmm8,96(%rsi) [all …]
|
/external/boringssl/linux-x86_64/crypto/aes/ |
D | bsaes-x86_64.S | 14 movdqa (%rax),%xmm8 17 pxor %xmm8,%xmm15 18 pxor %xmm8,%xmm0 19 pxor %xmm8,%xmm1 20 pxor %xmm8,%xmm2 23 pxor %xmm8,%xmm3 24 pxor %xmm8,%xmm4 27 pxor %xmm8,%xmm5 28 pxor %xmm8,%xmm6 35 movdqa 16(%r11),%xmm8 [all …]
|
D | aesni-x86_64.S | 383 pxor %xmm0,%xmm8 447 pxor %xmm0,%xmm8 520 movdqu 96(%rdi),%xmm8 541 movups %xmm8,96(%rsi) 542 movdqu 96(%rdi),%xmm8 562 movups %xmm8,96(%rsi) 584 movdqu 96(%rdi),%xmm8 593 movups %xmm8,96(%rsi) 663 movdqu 96(%rdi),%xmm8 684 movups %xmm8,96(%rsi) [all …]
|
/external/boringssl/win-x86_64/crypto/aes/ |
D | bsaes-x86_64.asm | 16 movdqa xmm8,XMMWORD[rax] 19 pxor xmm15,xmm8 20 pxor xmm0,xmm8 21 pxor xmm1,xmm8 22 pxor xmm2,xmm8 25 pxor xmm3,xmm8 26 pxor xmm4,xmm8 29 pxor xmm5,xmm8 30 pxor xmm6,xmm8 37 movdqa xmm8,XMMWORD[16+r11] [all …]
|
/external/llvm/test/MC/X86/ |
D | intel-syntax-x86-64-avx.s | 21 vgatherdps xmm10, xmmword ptr [r15 + 2*xmm9], xmm8 25 vgatherqps xmm10, qword ptr [r15 + 2*xmm9], xmm8 33 vgatherqps xmm10, xmmword ptr [r15 + 2*ymm9], xmm8 53 vpgatherdd xmm10, xmmword ptr [r15 + 2*xmm9], xmm8 57 vpgatherqd xmm10, qword ptr [r15 + 2*xmm9], xmm8 65 vpgatherqd xmm10, xmmword ptr [r15 + 2*ymm9], xmm8
|
/external/boringssl/mac-x86_64/crypto/sha/ |
D | sha512-x86_64.S | 1846 vpaddq -128(%rbp),%xmm0,%xmm8 1852 vmovdqa %xmm8,0(%rsp) 1853 vpaddq 0(%rbp),%xmm4,%xmm8 1860 vmovdqa %xmm8,64(%rsp) 1873 vpalignr $8,%xmm0,%xmm1,%xmm8 1882 vpsrlq $7,%xmm8,%xmm8 1893 vpxor %xmm9,%xmm8,%xmm8 1901 vpxor %xmm10,%xmm8,%xmm8 1907 vpaddq %xmm8,%xmm0,%xmm0 1942 vpalignr $8,%xmm1,%xmm2,%xmm8 [all …]
|
D | sha1-x86_64.S | 1295 movdqa %xmm3,%xmm8 1303 psrldq $4,%xmm8 1309 pxor %xmm2,%xmm8 1313 pxor %xmm8,%xmm4 1323 movdqa %xmm4,%xmm8 1329 psrld $31,%xmm8 1339 por %xmm8,%xmm4 1380 movdqa %xmm5,%xmm8 1386 pslldq $12,%xmm8 1394 movdqa %xmm8,%xmm10 [all …]
|
/external/boringssl/linux-x86_64/crypto/sha/ |
D | sha512-x86_64.S | 1847 vpaddq -128(%rbp),%xmm0,%xmm8 1853 vmovdqa %xmm8,0(%rsp) 1854 vpaddq 0(%rbp),%xmm4,%xmm8 1861 vmovdqa %xmm8,64(%rsp) 1874 vpalignr $8,%xmm0,%xmm1,%xmm8 1883 vpsrlq $7,%xmm8,%xmm8 1894 vpxor %xmm9,%xmm8,%xmm8 1902 vpxor %xmm10,%xmm8,%xmm8 1908 vpaddq %xmm8,%xmm0,%xmm0 1943 vpalignr $8,%xmm1,%xmm2,%xmm8 [all …]
|
D | sha1-x86_64.S | 1296 movdqa %xmm3,%xmm8 1304 psrldq $4,%xmm8 1310 pxor %xmm2,%xmm8 1314 pxor %xmm8,%xmm4 1324 movdqa %xmm4,%xmm8 1330 psrld $31,%xmm8 1340 por %xmm8,%xmm4 1381 movdqa %xmm5,%xmm8 1387 pslldq $12,%xmm8 1395 movdqa %xmm8,%xmm10 [all …]
|
/external/boringssl/win-x86_64/crypto/sha/ |
D | sha512-x86_64.asm | 1845 movaps XMMWORD[(128+64)+rsp],xmm8 1879 vpaddq xmm8,xmm0,XMMWORD[((-128))+rbp] 1885 vmovdqa XMMWORD[rsp],xmm8 1886 vpaddq xmm8,xmm4,XMMWORD[rbp] 1893 vmovdqa XMMWORD[64+rsp],xmm8 1906 vpalignr xmm8,xmm1,xmm0,8 1915 vpsrlq xmm8,xmm8,7 1926 vpxor xmm8,xmm8,xmm9 1934 vpxor xmm8,xmm8,xmm10 1940 vpaddq xmm0,xmm0,xmm8 [all …]
|
/external/boringssl/linux-x86_64/crypto/cipher/ |
D | chacha20_poly1305_x86_64.S | 261 movdqu 16(%r9),%xmm8 266 movdqa %xmm8,64(%rbp) 273 paddd %xmm12,%xmm8 274 pxor %xmm8,%xmm4 282 paddd %xmm12,%xmm8 283 pxor %xmm8,%xmm4 294 paddd %xmm12,%xmm8 295 pxor %xmm8,%xmm4 303 paddd %xmm12,%xmm8 304 pxor %xmm8,%xmm4 [all …]
|
/external/boringssl/mac-x86_64/crypto/cipher/ |
D | chacha20_poly1305_x86_64.S | 260 movdqu 16(%r9),%xmm8 265 movdqa %xmm8,64(%rbp) 272 paddd %xmm12,%xmm8 273 pxor %xmm8,%xmm4 281 paddd %xmm12,%xmm8 282 pxor %xmm8,%xmm4 293 paddd %xmm12,%xmm8 294 pxor %xmm8,%xmm4 302 paddd %xmm12,%xmm8 303 pxor %xmm8,%xmm4 [all …]
|