/external/libvpx/libvpx/vp8/common/x86/ |
D | loopfilter_block_sse2_x86_64.asm | 18 movdqa scratch1, %2 ; v2 31 movdqa scratch2, %3 ; save p1 68 movdqa scratch2, %6 ; save hev 72 movdqa scratch1, %1 78 movdqa scratch1, %3 85 movdqa scratch2, %5 90 movdqa scratch1, zero 101 movdqa scratch1, zero 113 movdqa scratch1, zero 199 movdqa xmm0, i0 [all …]
|
D | idctllm_sse2.asm | 128 movdqa xmm0, [rax] 129 movdqa xmm2, [rax+16] 130 movdqa xmm1, [rax+32] 131 movdqa xmm3, [rax+48] 134 movdqa [rax], xmm7 135 movdqa [rax+16], xmm7 136 movdqa [rax+32], xmm7 137 movdqa [rax+48], xmm7 147 movdqa xmm4, xmm0 154 movdqa xmm4, xmm2 [all …]
|
D | loopfilter_sse2.asm | 30 movdqa xmm2, [rdi+2*rax] ; q3 31 movdqa xmm1, [rsi+2*rax] ; q2 32 movdqa xmm4, [rsi+rax] ; q1 33 movdqa xmm5, [rsi] ; q0 49 movdqa [rsp+_q2], xmm1 ; store q2 50 movdqa [rsp+_q1], xmm4 ; store q1 52 movdqa xmm7, [rdx] ;limit 54 movdqa xmm6, xmm1 ; q2 55 movdqa xmm3, xmm4 ; q1 66 movdqa xmm0, xmm5 ; q0 [all …]
|
D | recon_sse2.asm | 41 movdqa [rdi], xmm0 44 movdqa [rdi+rcx], xmm1 45 movdqa [rdi+rcx*2],xmm2 56 movdqa [rdi], xmm3 59 movdqa [rdi+rcx], xmm4 60 movdqa [rdi+rcx*2],xmm5 71 movdqa [rdi], xmm0 74 movdqa [rdi+rcx], xmm1 76 movdqa [rdi+rcx*2], xmm2 86 movdqa [rdi], xmm3 [all …]
|
D | subpixel_sse2.asm | 68 movdqa xmm4, xmm1 69 movdqa xmm5, xmm1 71 movdqa xmm6, xmm1 72 movdqa xmm7, xmm1 118 movdqa XMMWORD Ptr [rdi], xmm4 193 movdqa xmm4, xmm1 194 movdqa xmm5, xmm1 196 movdqa xmm6, xmm1 197 movdqa xmm7, xmm1 242 movdqa XMMWORD Ptr [rdi], xmm4 [all …]
|
/external/boringssl/src/crypto/aes/asm/ |
D | vpaes-x86.pl | 162 &movdqa ("xmm7",&QWP($k_inv,$const)); 163 &movdqa ("xmm6",&QWP($k_s0F,$const)); 184 &movdqa ("xmm1","xmm6") 185 &movdqa ("xmm2",&QWP($k_ipt,$const)); 190 &movdqa ("xmm0",&QWP($k_ipt+16,$const)); 202 &movdqa ("xmm4",&QWP($k_sb1,$const)); # 4 : sb1u 203 &movdqa ("xmm0",&QWP($k_sb1+16,$const));# 0 : sb1t 207 &movdqa ("xmm5",&QWP($k_sb2,$const)); # 4 : sb2u 209 &movdqa ("xmm1",&QWP(-0x40,$base,$magic));# .Lk_mc_forward[] 211 &movdqa ("xmm2",&QWP($k_sb2+16,$const));# 2 : sb2t [all …]
|
/external/boringssl/mac-x86_64/crypto/aes/ |
D | vpaes-x86_64.S | 25 movdqa %xmm9,%xmm1 26 movdqa L$k_ipt(%rip),%xmm2 32 movdqa L$k_ipt+16(%rip),%xmm0 43 movdqa %xmm13,%xmm4 44 movdqa %xmm12,%xmm0 48 movdqa %xmm15,%xmm5 50 movdqa -64(%r11,%r10,1),%xmm1 52 movdqa (%r11,%r10,1),%xmm4 53 movdqa %xmm14,%xmm2 55 movdqa %xmm0,%xmm3 [all …]
|
D | bsaes-x86_64.S | 12 movdqa (%rax),%xmm8 14 movdqa 80(%r11),%xmm7 32 movdqa 0(%r11),%xmm7 33 movdqa 16(%r11),%xmm8 34 movdqa %xmm5,%xmm9 36 movdqa %xmm3,%xmm10 48 movdqa %xmm1,%xmm9 50 movdqa %xmm15,%xmm10 62 movdqa 32(%r11),%xmm7 63 movdqa %xmm4,%xmm9 [all …]
|
/external/boringssl/linux-x86_64/crypto/aes/ |
D | vpaes-x86_64.S | 25 movdqa %xmm9,%xmm1 26 movdqa .Lk_ipt(%rip),%xmm2 32 movdqa .Lk_ipt+16(%rip),%xmm0 43 movdqa %xmm13,%xmm4 44 movdqa %xmm12,%xmm0 48 movdqa %xmm15,%xmm5 50 movdqa -64(%r11,%r10,1),%xmm1 52 movdqa (%r11,%r10,1),%xmm4 53 movdqa %xmm14,%xmm2 55 movdqa %xmm0,%xmm3 [all …]
|
D | bsaes-x86_64.S | 14 movdqa (%rax),%xmm8 16 movdqa 80(%r11),%xmm7 34 movdqa 0(%r11),%xmm7 35 movdqa 16(%r11),%xmm8 36 movdqa %xmm5,%xmm9 38 movdqa %xmm3,%xmm10 50 movdqa %xmm1,%xmm9 52 movdqa %xmm15,%xmm10 64 movdqa 32(%r11),%xmm7 65 movdqa %xmm4,%xmm9 [all …]
|
/external/boringssl/mac-x86/crypto/aes/ |
D | vpaes-x86.S | 65 movdqa -48(%ebp),%xmm7 66 movdqa -16(%ebp),%xmm6 73 movdqa %xmm6,%xmm1 74 movdqa (%ebp),%xmm2 79 movdqa 16(%ebp),%xmm0 89 movdqa 32(%ebp),%xmm4 90 movdqa 48(%ebp),%xmm0 94 movdqa 64(%ebp),%xmm5 96 movdqa -64(%ebx,%ecx,1),%xmm1 98 movdqa 80(%ebp),%xmm2 [all …]
|
/external/boringssl/linux-x86/crypto/aes/ |
D | vpaes-x86.S | 66 movdqa -48(%ebp),%xmm7 67 movdqa -16(%ebp),%xmm6 76 movdqa %xmm6,%xmm1 77 movdqa (%ebp),%xmm2 82 movdqa 16(%ebp),%xmm0 92 movdqa 32(%ebp),%xmm4 93 movdqa 48(%ebp),%xmm0 97 movdqa 64(%ebp),%xmm5 99 movdqa -64(%ebx,%ecx,1),%xmm1 101 movdqa 80(%ebp),%xmm2 [all …]
|
/external/boringssl/win-x86/crypto/aes/ |
D | vpaes-x86.asm | 77 movdqa xmm7,[ebp-48] 78 movdqa xmm6,[ebp-16] 84 movdqa xmm1,xmm6 85 movdqa xmm2,[ebp] 90 movdqa xmm0,[16+ebp] 100 movdqa xmm4,[32+ebp] 101 movdqa xmm0,[48+ebp] 105 movdqa xmm5,[64+ebp] 107 movdqa xmm1,[ecx*1+ebx-64] 109 movdqa xmm2,[80+ebp] [all …]
|
/external/llvm/test/MC/X86/ |
D | cfi_def_cfa-crash.s | 22 movdqa %xmm0, -0x80(%rbp) 24 movdqa %xmm1, -0x70(%rbp) 26 movdqa %xmm2, -0x60(%rbp) 28 movdqa %xmm3, -0x50(%rbp) 30 movdqa %xmm4, -0x40(%rbp) 32 movdqa %xmm5, -0x30(%rbp) 34 movdqa %xmm6, -0x20(%rbp) 36 movdqa %xmm7, -0x10(%rbp) 41 movdqa -0x80(%rbp), %xmm0 43 movdqa -0x70(%rbp), %xmm1 [all …]
|
/external/libyuv/files/source/ |
D | row_win.cc | 139 movdqa xmm1, xmm0 in I400ToARGBRow_SSE2() 144 movdqa [edx], xmm0 in I400ToARGBRow_SSE2() 145 movdqa [edx + 16], xmm1 in I400ToARGBRow_SSE2() 159 movdqa xmm5, kShuffleMaskBGRAToARGB in BGRAToARGBRow_SSSE3() 164 movdqa xmm0, [eax] in BGRAToARGBRow_SSSE3() 167 movdqa [eax + edx], xmm0 in BGRAToARGBRow_SSSE3() 180 movdqa xmm5, kShuffleMaskABGRToARGB in ABGRToARGBRow_SSSE3() 185 movdqa xmm0, [eax] in ABGRToARGBRow_SSSE3() 188 movdqa [eax + edx], xmm0 in ABGRToARGBRow_SSSE3() 201 movdqa xmm5, kShuffleMaskRGBAToARGB in RGBAToARGBRow_SSSE3() [all …]
|
D | rotate.cc | 86 movdqa xmm1, xmm0 in TransposeWx8_SSSE3() 91 movdqa xmm3, xmm2 in TransposeWx8_SSSE3() 97 movdqa xmm5, xmm4 in TransposeWx8_SSSE3() 103 movdqa xmm7, xmm6 in TransposeWx8_SSSE3() 108 movdqa xmm2, xmm0 in TransposeWx8_SSSE3() 109 movdqa xmm3, xmm1 in TransposeWx8_SSSE3() 114 movdqa xmm6, xmm4 in TransposeWx8_SSSE3() 115 movdqa xmm7, xmm5 in TransposeWx8_SSSE3() 122 movdqa xmm4, xmm0 in TransposeWx8_SSSE3() 127 movdqa xmm6, xmm2 in TransposeWx8_SSSE3() [all …]
|
D | scale.cc | 214 movdqa xmm0, [eax] in ScaleRowDown2_SSE2() 215 movdqa xmm1, [eax + 16] in ScaleRowDown2_SSE2() 221 movdqa [edx], xmm0 in ScaleRowDown2_SSE2() 244 movdqa xmm0, [eax] in ScaleRowDown2Int_SSE2() 245 movdqa xmm1, [eax + 16] in ScaleRowDown2Int_SSE2() 246 movdqa xmm2, [eax + esi] in ScaleRowDown2Int_SSE2() 247 movdqa xmm3, [eax + esi + 16] in ScaleRowDown2Int_SSE2() 252 movdqa xmm2, xmm0 // average columns (32 to 16 pixels) in ScaleRowDown2Int_SSE2() 254 movdqa xmm3, xmm1 in ScaleRowDown2Int_SSE2() 263 movdqa [edx], xmm0 in ScaleRowDown2Int_SSE2() [all …]
|
/external/libvpx/libvpx/third_party/libyuv/source/ |
D | scale_win.cc | 108 movdqa xmm0, [eax] in ScaleRowDown2_SSE2() 109 movdqa xmm1, [eax + 16] in ScaleRowDown2_SSE2() 115 movdqa [edx], xmm0 in ScaleRowDown2_SSE2() 138 movdqa xmm0, [eax] in ScaleRowDown2Linear_SSE2() 139 movdqa xmm1, [eax + 16] in ScaleRowDown2Linear_SSE2() 142 movdqa xmm2, xmm0 // average columns (32 to 16 pixels) in ScaleRowDown2Linear_SSE2() 144 movdqa xmm3, xmm1 in ScaleRowDown2Linear_SSE2() 153 movdqa [edx], xmm0 in ScaleRowDown2Linear_SSE2() 177 movdqa xmm0, [eax] in ScaleRowDown2Box_SSE2() 178 movdqa xmm1, [eax + 16] in ScaleRowDown2Box_SSE2() [all …]
|
D | row_win.cc | 325 movdqa xmm1, xmm0 in I400ToARGBRow_SSE2() 330 movdqa [edx], xmm0 in I400ToARGBRow_SSE2() 331 movdqa [edx + 16], xmm1 in I400ToARGBRow_SSE2() 354 movdqa xmm1, xmm0 in I400ToARGBRow_Unaligned_SSE2() 376 movdqa xmm4, kShuffleMaskRGB24ToARGB in RGB24ToARGBRow_SSSE3() 384 movdqa xmm2, xmm3 in RGB24ToARGBRow_SSSE3() 390 movdqa [edx + 32], xmm2 in RGB24ToARGBRow_SSSE3() 393 movdqa [edx], xmm0 in RGB24ToARGBRow_SSSE3() 397 movdqa [edx + 16], xmm1 in RGB24ToARGBRow_SSSE3() 400 movdqa [edx + 48], xmm3 in RGB24ToARGBRow_SSSE3() [all …]
|
/external/boringssl/win-x86_64/crypto/aes/ |
D | bsaes-x86_64.asm | 16 movdqa xmm8,XMMWORD[rax] 18 movdqa xmm7,XMMWORD[80+r11] 36 movdqa xmm7,XMMWORD[r11] 37 movdqa xmm8,XMMWORD[16+r11] 38 movdqa xmm9,xmm5 40 movdqa xmm10,xmm3 52 movdqa xmm9,xmm1 54 movdqa xmm10,xmm15 66 movdqa xmm7,XMMWORD[32+r11] 67 movdqa xmm9,xmm4 [all …]
|
D | vpaes-x86_64.asm | 29 movdqa xmm1,xmm9 30 movdqa xmm2,XMMWORD[$L$k_ipt] 36 movdqa xmm0,XMMWORD[(($L$k_ipt+16))] 47 movdqa xmm4,xmm13 48 movdqa xmm0,xmm12 52 movdqa xmm5,xmm15 54 movdqa xmm1,XMMWORD[((-64))+r10*1+r11] 56 movdqa xmm4,XMMWORD[r10*1+r11] 57 movdqa xmm2,xmm14 59 movdqa xmm3,xmm0 [all …]
|
/external/libvpx/libvpx/vp8/encoder/x86/ |
D | dct_sse2.asm | 78 movdqa xmm2, xmm0 81 movdqa xmm1, xmm0 87 movdqa xmm3, xmm0 93 movdqa xmm1, xmm0 96 movdqa xmm4, xmm3 111 movdqa xmm2, xmm0 115 movdqa xmm3, xmm0 118 movdqa xmm2, xmm0 122 movdqa xmm5, XMMWORD PTR[GLOBAL(_7)] 124 movdqa xmm3, xmm0 [all …]
|
D | subtract_sse2.asm | 90 movdqa xmm4, [GLOBAL(t80)] 96 movdqa xmm0, [rsi] ; src 97 movdqa xmm1, [rax] ; pred 99 movdqa xmm2, xmm0 106 movdqa xmm2, xmm0 110 movdqa xmm3, [rsi + rdx] 111 movdqa xmm5, [rax + rbx] 116 movdqa [rdi], xmm0 117 movdqa [rdi +16], xmm2 119 movdqa xmm1, xmm3 [all …]
|
/external/libvpx/libvpx/vp9/common/x86/ |
D | vp9_subpixel_8t_sse2.asm | 21 movdqa xmm7, [rdx] ;load filters 37 movdqa k0k1, xmm0 38 movdqa k2k3, xmm2 39 movdqa k5k4, xmm5 40 movdqa k6k7, xmm6 44 movdqa krd, xmm6 47 movdqa zero, xmm7 67 movdqa xmm1, xmm0 94 movdqa xmm7, [rdx] ;load filters 113 movdqa k0, xmm0 ;store filter factors on stack [all …]
|
D | vp9_subpixel_8t_ssse3.asm | 20 movdqa xmm4, [rdx] ;load filters 33 movdqa k0k1, xmm0 34 movdqa k2k3, xmm1 36 movdqa k4k5, xmm2 37 movdqa k6k7, xmm3 38 movdqa krd, xmm5 73 movdqa xmm1, xmm2 107 movdqa xmm4, [rdx] ;load filters 120 movdqa k0k1, xmm0 121 movdqa k2k3, xmm1 [all …]
|