/external/boringssl/src/crypto/aes/asm/ |
D | vpaes-x86.pl | 193 &pshufb ("xmm2","xmm0"); 198 &pshufb ("xmm0","xmm1"); 208 &pshufb ("xmm4","xmm2"); # 4 = sb1u 209 &pshufb ("xmm0","xmm3"); # 0 = sb1t 214 &pshufb ("xmm5","xmm2"); # 4 = sb2u 217 &pshufb ("xmm2","xmm3"); # 2 = sb2t 220 &pshufb ("xmm0","xmm1"); # 0 = B 223 &pshufb ("xmm3","xmm4"); # 3 = D 226 &pshufb ("xmm0","xmm1"); # 0 = 2B+C 238 &pshufb ("xmm5","xmm0"); # 2 = a/k [all …]
|
/external/libvpx/libvpx/vpx_dsp/x86/ |
D | intrapred_ssse3.asm | 42 pshufb m2, m0, [GLOBAL(sh_b23456789abcdefff)] 45 pshufb m0, m1 55 pshufb m0, m1 58 pshufb m0, m1 61 pshufb m0, m1 64 pshufb m0, m1 95 pshufb m2, m4, [GLOBAL(sh_b23456789abcdefff)] 100 pshufb m4, m1 117 pshufb m4, m1 122 pshufb m4, m1 [all …]
|
/external/swiftshader/third_party/LLVM/test/CodeGen/X86/ |
D | trunc-ext-ld-st.ll | 6 ;CHECK: pshufb 8 ;CHECK: pshufb 23 ;CHECK: pshufb 25 ;CHECK: pshufb 49 ;CHECK: pshufb 51 ;CHECK: pshufb 63 ;CHECK: pshufb 75 ;CHECK: pshufb
|
D | vec_shuffle-28.ll | 2 ; RUN: grep pshufb %t | count 1 4 ; FIXME: this test has a superfluous punpcklqdq pre-pshufb currently.
|
D | vec_shuffle-36.ll | 4 ; CHECK: pshufb 5 ; CHECK-NOT: pshufb
|
/external/boringssl/src/crypto/modes/asm/ |
D | ghash-x86.pl | 753 &pshufb ($Xi,$T3); 759 &pshufb ($Xi,$T3); 779 &pshufb ($Xi,$T3); 791 &pshufb ($T1,$T3); 792 &pshufb ($Xn,$T3); 830 &pshufb ($Xhn,$T3); 838 &pshufb ($Xn,$T3); 905 &pshufb ($T1,$T3); 912 &pshufb ($Xi,$T3); 998 &pshufb ($Xi,$Xn); [all …]
|
/external/llvm/test/CodeGen/X86/ |
D | pshufb-mask-comments.ll | 4 ; Test that the pshufb mask comment is correct. 9 ; CHECK-NEXT: pshufb {{.*#+}} xmm0 = xmm0[1,0,0,0,0,2,0,0,0,0,3,0,0,0,0,4] 20 ; CHECK-NEXT: pshufb {{.*#+}} xmm0 = xmm0[15,0,0,0,0,0,0,0,0,0,1,0,0,0,0,2] 31 ; CHECK-NEXT: pshufb {{.*#+}} xmm0 = xmm0[1,0,0,15,0,2,0,0],zero,xmm0[0,3,0,0],zero,xmm0[0,4] 62 ; CHECK-NEXT: pshufb %xmm1, %xmm0 73 ; Test for a reused constant that would allow the pshufb to combine to a simpler instruction.
|
D | vector-shuffle-128-v8.ll | 149 ; SSSE3-NEXT: pshufb {{.*#+}} xmm0 = xmm0[8,9,8,9,8,9,8,9,0,1,0,1,0,1,0,1] 154 ; SSE41-NEXT: pshufb {{.*#+}} xmm0 = xmm0[8,9,8,9,8,9,8,9,0,1,0,1,0,1,0,1] 217 ; SSSE3-NEXT: pshufb {{.*#+}} xmm0 = xmm0[14,15,10,11,12,13,8,9,6,7,2,3,4,5,0,1] 222 ; SSE41-NEXT: pshufb {{.*#+}} xmm0 = xmm0[14,15,10,11,12,13,8,9,6,7,2,3,4,5,0,1] 243 ; SSSE3-NEXT: pshufb {{.*#+}} xmm0 = xmm0[2,3,0,1,10,11,8,9,10,11,8,9,2,3,0,1] 248 ; SSE41-NEXT: pshufb {{.*#+}} xmm0 = xmm0[2,3,0,1,10,11,8,9,10,11,8,9,2,3,0,1] 268 ; SSSE3-NEXT: pshufb {{.*#+}} xmm0 = xmm0[10,11,8,9,2,3,0,1,10,11,8,9,2,3,0,1] 273 ; SSE41-NEXT: pshufb {{.*#+}} xmm0 = xmm0[10,11,8,9,2,3,0,1,10,11,8,9,2,3,0,1] 293 ; SSSE3-NEXT: pshufb {{.*#+}} xmm0 = xmm0[10,11,8,9,2,3,0,1,2,3,0,1,10,11,8,9] 298 ; SSE41-NEXT: pshufb {{.*#+}} xmm0 = xmm0[10,11,8,9,2,3,0,1,2,3,0,1,10,11,8,9] [all …]
|
D | trunc-ext-ld-st.ll | 7 ;CHECK: pshufb 48 ;CHECK: pshufb 60 ;CHECK: pshufb 72 ;CHECK: pshufb
|
D | vector-popcnt-128.ll | 59 ; SSSE3-NEXT: pshufb %xmm2, %xmm4 62 ; SSSE3-NEXT: pshufb %xmm0, %xmm3 75 ; SSE41-NEXT: pshufb %xmm2, %xmm4 78 ; SSE41-NEXT: pshufb %xmm0, %xmm3 161 ; SSSE3-NEXT: pshufb %xmm3, %xmm4 164 ; SSSE3-NEXT: pshufb %xmm0, %xmm1 183 ; SSE41-NEXT: pshufb %xmm3, %xmm4 186 ; SSE41-NEXT: pshufb %xmm0, %xmm1 271 ; SSSE3-NEXT: pshufb %xmm2, %xmm4 274 ; SSSE3-NEXT: pshufb %xmm0, %xmm3 [all …]
|
D | vector-shuffle-128-v16.ll | 22 ; SSSE3-NEXT: pshufb %xmm1, %xmm0 28 ; SSE41-NEXT: pshufb %xmm1, %xmm0 55 ; SSSE3-NEXT: pshufb {{.*#+}} xmm0 = xmm0[0,0,0,0,0,0,0,0,1,1,1,1,1,1,1,1] 60 ; SSE41-NEXT: pshufb {{.*#+}} xmm0 = xmm0[0,0,0,0,0,0,0,0,1,1,1,1,1,1,1,1] 83 ; SSSE3-NEXT: pshufb {{.*#+}} xmm0 = xmm0[0,0,0,0,0,0,0,0,8,8,8,8,8,8,8,8] 88 ; SSE41-NEXT: pshufb {{.*#+}} xmm0 = xmm0[0,0,0,0,0,0,0,0,8,8,8,8,8,8,8,8] 144 ; SSSE3-NEXT: pshufb {{.*#+}} xmm0 = xmm0[0,0,0,0,4,4,4,4,8,8,8,8,12,12,12,12] 149 ; SSE41-NEXT: pshufb {{.*#+}} xmm0 = xmm0[0,0,0,0,4,4,4,4,8,8,8,8,12,12,12,12] 284 ; SSSE3-NEXT: pshufb {{.*#+}} xmm0 = xmm0[3,2,1,0,7,6,5,4,11,10,9,8,15,14,13,12] 289 ; SSE41-NEXT: pshufb {{.*#+}} xmm0 = xmm0[3,2,1,0,7,6,5,4,11,10,9,8,15,14,13,12] [all …]
|
D | widen_conv-1.ll | 40 ; X86-NEXT: pshufb {{.*#+}} xmm0 = xmm0[0,4,8,12,u,u,u,u,u,u,u,u,u,u,u,u] 52 ; X64-NEXT: pshufb {{.*#+}} xmm0 = xmm0[0,4,8,12,u,u,u,u,u,u,u,u,u,u,u,u] 79 ; X86-NEXT: pshufb {{.*#+}} xmm0 = xmm0[0,2,4,6,8,10,12,14,u,u,u,u,u,u,u,u] 91 ; X64-NEXT: pshufb {{.*#+}} xmm0 = xmm0[0,2,4,6,8,10,12,14,u,u,u,u,u,u,u,u]
|
D | vector-trunc.ll | 228 ; SSSE3-NEXT: pshufb %xmm2, %xmm1 229 ; SSSE3-NEXT: pshufb %xmm2, %xmm0 236 ; SSE41-NEXT: pshufb %xmm2, %xmm1 237 ; SSE41-NEXT: pshufb %xmm2, %xmm0 284 ; SSSE3-NEXT: pshufb %xmm2, %xmm1 285 ; SSSE3-NEXT: pshufb %xmm2, %xmm0 293 ; SSE41-NEXT: pshufb %xmm2, %xmm1 294 ; SSE41-NEXT: pshufb %xmm2, %xmm0 648 ; SSSE3-NEXT: pshufb %xmm2, %xmm1 649 ; SSSE3-NEXT: pshufb %xmm2, %xmm0 [all …]
|
D | vector-tzcnt-128.ll | 308 ; SSSE3-NEXT: pshufb %xmm4, %xmm5 311 ; SSSE3-NEXT: pshufb %xmm2, %xmm0 333 ; SSE41-NEXT: pshufb %xmm4, %xmm5 336 ; SSE41-NEXT: pshufb %xmm2, %xmm0 444 ; X32-SSE-NEXT: pshufb %xmm4, %xmm5 447 ; X32-SSE-NEXT: pshufb %xmm2, %xmm0 531 ; SSSE3-NEXT: pshufb %xmm4, %xmm5 534 ; SSSE3-NEXT: pshufb %xmm2, %xmm0 556 ; SSE41-NEXT: pshufb %xmm4, %xmm5 559 ; SSE41-NEXT: pshufb %xmm2, %xmm0 [all …]
|
D | vector-shuffle-combining-ssse3.ll | 46 ; SSE-NEXT: pshufb {{.*#+}} xmm0 = xmm0[5,5,5,5,7,7,7,7,5,5,5,5,7,7,7,7] 63 ; SSE-NEXT: pshufb {{.*#+}} xmm0 = xmm0[7,7,7,7,7,7,7,7,3,3,3,3,3,3,3,3] 80 ; SSE-NEXT: pshufb {{.*#+}} xmm0 = xmm0[5,5,5,5,5,5,5,5,1,1,1,1,1,1,1,1] 142 ; SSE-NEXT: pshufb {{.*#+}} xmm0 = zero,zero,zero,zero,zero,zero,zero,zero,zero,zero,xmm0[0,1,2,… 156 ; SSE-NEXT: pshufb {{.*#+}} xmm0 = xmm0[15],zero,zero,zero,zero,zero,zero,zero,zero,zero,zero,ze… 198 ; SSE-NEXT: pshufb {{.*#+}} xmm0 = xmm0[2,3,0,1,6,7,4,5,10,11,8,9,14,15,12,13] 241 ; SSE-NEXT: pshufb {{.*#+}} xmm0 = xmm0[0],zero,zero,zero,xmm0[0],zero,zero,zero,xmm0[0],zero,ze… 256 ; SSE-NEXT: pshufb {{.*#+}} xmm1 = xmm1[0],zero,zero,zero,xmm1[0],zero,zero,zero,xmm1[0],zero,ze…
|
D | vector-bitreverse.ll | 1071 ; SSSE3-NEXT: pshufb %xmm2, %xmm3 1075 ; SSSE3-NEXT: pshufb %xmm0, %xmm1 1156 ; SSSE3-NEXT: pshufb {{.*#+}} xmm0 = xmm0[1,0,3,2,5,4,7,6,9,8,11,10,13,12,15,14] 1161 ; SSSE3-NEXT: pshufb %xmm2, %xmm3 1165 ; SSSE3-NEXT: pshufb %xmm0, %xmm1 1247 ; SSSE3-NEXT: pshufb {{.*#+}} xmm0 = xmm0[3,2,1,0,7,6,5,4,11,10,9,8,15,14,13,12] 1252 ; SSSE3-NEXT: pshufb %xmm2, %xmm3 1256 ; SSSE3-NEXT: pshufb %xmm0, %xmm1 1340 ; SSSE3-NEXT: pshufb {{.*#+}} xmm0 = xmm0[7,6,5,4,3,2,1,0,15,14,13,12,11,10,9,8] 1345 ; SSSE3-NEXT: pshufb %xmm2, %xmm3 [all …]
|
D | x86-fold-pshufb.ll | 19 ; The pshufb from function @pr24562 was wrongly folded into its first operand 22 ; Check that the pshufb is correctly folded to a zero vector.
|
D | insertelement-zero.ll | 476 ; SSSE3-NEXT: pshufb {{.*#+}} xmm0 = zero,xmm0[1,2,3,4,5,6,7,8,9,10,11,12,13,14,15] 480 ; SSSE3-NEXT: pshufb {{.*#+}} xmm2 = xmm2[0],zero,zero,zero,zero,zero,zero,zero,zero,zero,zero,z… 482 ; SSSE3-NEXT: pshufb {{.*#+}} xmm0 = xmm0[0,1,2,3,4,5,6,7,8,9,10,11,12,13,14],zero 483 ; SSSE3-NEXT: pshufb {{.*#+}} xmm1 = zero,zero,zero,zero,zero,zero,zero,zero,zero,zero,zero,zero… 554 ; SSSE3-NEXT: pshufb {{.*#+}} xmm0 = zero,xmm0[1,2,3,4,5,6,7,8,9,10,11,12,13,14,15] 558 ; SSSE3-NEXT: pshufb {{.*#+}} xmm3 = xmm3[0],zero,zero,zero,zero,zero,zero,zero,zero,zero,zero,z… 561 ; SSSE3-NEXT: pshufb %xmm3, %xmm0 563 ; SSSE3-NEXT: pshufb {{.*#+}} xmm4 = zero,zero,zero,zero,zero,zero,zero,zero,zero,zero,zero,zero… 565 ; SSSE3-NEXT: pshufb {{.*#+}} xmm1 = xmm1[0,1,2,3,4,5,6,7,8,9,10,11,12,13],zero,xmm1[15] 566 ; SSSE3-NEXT: pshufb {{.*#+}} xmm2 = zero,zero,zero,zero,zero,zero,zero,zero,zero,zero,zero,zero… [all …]
|
D | bswap-vector.ll | 30 ; CHECK-SSSE3-NEXT: pshufb {{.*#+}} xmm0 = xmm0[1,0,3,2,5,4,7,6,9,8,11,10,13,12,15,14] 63 ; CHECK-SSSE3-NEXT: pshufb {{.*#+}} xmm0 = xmm0[3,2,1,0,7,6,5,4,11,10,9,8,15,14,13,12] 98 ; CHECK-SSSE3-NEXT: pshufb {{.*#+}} xmm0 = xmm0[7,6,5,4,3,2,1,0,15,14,13,12,11,10,9,8] 144 ; CHECK-SSSE3-NEXT: pshufb %xmm2, %xmm0 145 ; CHECK-SSSE3-NEXT: pshufb %xmm2, %xmm1 187 ; CHECK-SSSE3-NEXT: pshufb %xmm2, %xmm0 188 ; CHECK-SSSE3-NEXT: pshufb %xmm2, %xmm1 234 ; CHECK-SSSE3-NEXT: pshufb %xmm2, %xmm0 235 ; CHECK-SSSE3-NEXT: pshufb %xmm2, %xmm1 271 ; CHECK-SSSE3-NEXT: pshufb {{.*#+}} xmm0 = xmm0[3,2,1,0,7,6,5,4,11,10,9,8,15,14,13,12]
|
/external/libyuv/files/source/ |
D | scale_win.cc | 520 pshufb xmm0, xmm3 in ScaleRowDown34_SSSE3() 521 pshufb xmm1, xmm4 in ScaleRowDown34_SSSE3() 522 pshufb xmm2, xmm5 in ScaleRowDown34_SSSE3() 570 pshufb xmm0, xmm2 in ScaleRowDown34_1_Box_SSSE3() 579 pshufb xmm0, xmm3 in ScaleRowDown34_1_Box_SSSE3() 589 pshufb xmm0, xmm4 in ScaleRowDown34_1_Box_SSSE3() 628 pshufb xmm0, xmm2 in ScaleRowDown34_0_Box_SSSE3() 638 pshufb xmm0, xmm3 in ScaleRowDown34_0_Box_SSSE3() 649 pshufb xmm0, xmm4 in ScaleRowDown34_0_Box_SSSE3() 684 pshufb xmm0, xmm4 in ScaleRowDown38_SSSE3() [all …]
|
/external/libvpx/libvpx/third_party/libyuv/source/ |
D | scale_win.cc | 515 pshufb xmm0, xmm3 in ScaleRowDown34_SSSE3() 516 pshufb xmm1, xmm4 in ScaleRowDown34_SSSE3() 517 pshufb xmm2, xmm5 in ScaleRowDown34_SSSE3() 565 pshufb xmm0, xmm2 in ScaleRowDown34_1_Box_SSSE3() 574 pshufb xmm0, xmm3 in ScaleRowDown34_1_Box_SSSE3() 584 pshufb xmm0, xmm4 in ScaleRowDown34_1_Box_SSSE3() 623 pshufb xmm0, xmm2 in ScaleRowDown34_0_Box_SSSE3() 633 pshufb xmm0, xmm3 in ScaleRowDown34_0_Box_SSSE3() 644 pshufb xmm0, xmm4 in ScaleRowDown34_0_Box_SSSE3() 678 pshufb xmm0, xmm4 in ScaleRowDown38_SSSE3() [all …]
|
/external/boringssl/linux-x86/crypto/chacha/ |
D | chacha-x86.S | 487 pshufb (%eax),%xmm6 500 pshufb 16(%eax),%xmm6 511 pshufb (%eax),%xmm7 526 pshufb 16(%eax),%xmm7 537 pshufb (%eax),%xmm6 552 pshufb 16(%eax),%xmm6 562 pshufb (%eax),%xmm7 575 pshufb 16(%eax),%xmm7 585 pshufb (%eax),%xmm6 599 pshufb 16(%eax),%xmm6 [all …]
|
/external/boringssl/mac-x86/crypto/chacha/ |
D | chacha-x86.S | 484 pshufb (%eax),%xmm6 497 pshufb 16(%eax),%xmm6 508 pshufb (%eax),%xmm7 523 pshufb 16(%eax),%xmm7 534 pshufb (%eax),%xmm6 549 pshufb 16(%eax),%xmm6 559 pshufb (%eax),%xmm7 572 pshufb 16(%eax),%xmm7 582 pshufb (%eax),%xmm6 596 pshufb 16(%eax),%xmm6 [all …]
|
/external/boringssl/src/crypto/sha/asm/ |
D | sha1-586.pl | 464 &pshufb (@MSG[0],$BSWAP); 466 &pshufb (@MSG[1],$BSWAP); 467 &pshufb (@MSG[2],$BSWAP); 468 &pshufb (@MSG[3],$BSWAP); 501 &pshufb (@MSG[0],$BSWAP); 507 &pshufb (@MSG[1],$BSWAP); 513 &pshufb (@MSG[2],$BSWAP); 518 &pshufb (@MSG[3],$BSWAP); 631 &pshufb (@X[-4&7],@X[2]); # byte swap 632 &pshufb (@X[-3&7],@X[2]); [all …]
|
/external/boringssl/win-x86/crypto/chacha/ |
D | chacha-x86.asm | 495 pshufb xmm6,[eax] 508 pshufb xmm6,[16+eax] 519 pshufb xmm7,[eax] 534 pshufb xmm7,[16+eax] 545 pshufb xmm6,[eax] 560 pshufb xmm6,[16+eax] 570 pshufb xmm7,[eax] 583 pshufb xmm7,[16+eax] 593 pshufb xmm6,[eax] 607 pshufb xmm6,[16+eax] [all …]
|