/external/libvpx/libvpx/vp8/common/x86/ |
D | subpixel_mmx.asm | 56 movq mm4, mm3 ; mm4 = p-2..p5 61 movq mm5, mm4 ; mm5 = p-2..p5 62 punpckhbw mm4, mm0 ; mm5 = p2..p5 63 pmullw mm4, mm7 ; mm5 *= kernel 4 modifiers 64 paddsw mm3, mm4 ; mm3 += mm5 66 movq mm4, mm5 ; mm4 = p-2..p5; 72 movq mm5, mm4 ; mm5 = p-2..p5 73 psrlq mm4, 24 ; mm4 = p1..p5 74 punpcklbw mm4, mm0 ; mm4 = p1..p4 75 pmullw mm4, mm6 ; mm5 *= kernel 3 modifiers [all …]
|
D | dequantize_mmx.asm | 30 pmullw mm1, [rax+0] ; mm4 *= kernel 0 modifiers. 34 pmullw mm1, [rax+8] ; mm4 *= kernel 0 modifiers. 38 pmullw mm1, [rax+16] ; mm4 *= kernel 0 modifiers. 42 pmullw mm1, [rax+24] ; mm4 *= kernel 0 modifiers. 113 movq mm4, mm3 119 paddw mm3, mm4 124 movq mm4, mm0 ; b1 127 paddw mm4, mm7 ;1 133 movq mm3, mm4 ; 23 22 21 20 139 punpckhwd mm4, mm6 ; 33 23 32 22 [all …]
|
D | loopfilter_mmx.asm | 59 movq mm4, [rsi+rax] ; q1 60 movq mm3, mm4 ; q1 61 psubusb mm4, mm6 ; q1-=q2 63 por mm4, mm6 ; abs(q2-q1) 65 psubusb mm4, mm7 66 por mm1, mm4 68 movq mm4, [rsi] ; q0 69 movq mm0, mm4 ; q0 70 psubusb mm4, mm3 ; q0-=q1 72 por mm4, mm3 ; abs(q0-q1) [all …]
|
D | idctllm_mmx.asm | 83 movq mm4, mm3 89 paddw mm3, mm4 94 movq mm4, mm0 ; b1 97 paddw mm4, mm7 ;1 103 movq mm3, mm4 ; 23 22 21 20 109 punpckhwd mm4, mm6 ; 33 23 32 22 117 punpckldq mm2, mm4 ; 32 22 12 02 118 punpckhdq mm5, mm4 ; 33 23 13 03 138 movq mm4, mm3 144 paddw mm3, mm4 [all …]
|
D | postproc_mmx.asm | 126 movq mm4, mm2 129 punpckhwd mm4, mm0 132 paddd mm7, mm4 150 movq mm4, mm5 153 pmulhw mm4, mm4 156 punpcklwd mm1, mm4 157 punpckhwd mm2, mm4 159 movq mm4, mm7 160 pslld mm4, 4 162 psubd mm4, mm7 [all …]
|
D | iwalsh_mmx.asm | 34 movq mm4, mm0 37 paddw mm4, mm3 ;ip[0] + ip[12] aka al 40 movq mm6, mm4 ;temp al 41 paddw mm4, mm5 ;al + bl 56 movq mm3, mm4 ; 03 02 01 00 57 punpcklwd mm4, mm0 ; 11 01 10 00 64 movq mm0, mm4 ; 11 01 10 00 68 punpckhdq mm4, mm6 ; 31 21 11 01 aka ip[4] 74 movq mm5, mm4 87 psubw mm4, mm2 ;ip[4] - ip[8] aka c1 [all …]
|
D | recon_mmx.asm | 53 movq mm4, [rsi+rax] 59 movq [rdi+rcx], mm4 147 movq mm4, [rsi+rax+8] 159 movq [rdi+rcx+8], mm4 171 movq mm4, [rsi+rax+8] 183 movq [rdi+rcx+8], mm4 195 movq mm4, [rsi+rax+8] 207 movq [rdi+rcx+8], mm4 219 movq mm4, [rsi+rax+8] 231 movq [rdi+rcx+8], mm4 [all …]
|
D | sad_mmx.asm | 54 movq mm4, mm0 58 psubusb mm1, mm4 299 movd mm4, DWORD PTR [rsi] 305 punpcklbw mm4, mm6 308 movq mm6, mm4 309 psubusb mm4, mm5 312 por mm4, mm5 314 movq mm5, mm4 315 punpcklbw mm4, mm3 318 paddw mm4, mm5 [all …]
|
D | variance_impl_mmx.asm | 28 pxor mm4, mm4 40 paddd mm4, mm0 41 paddd mm4, mm1 42 paddd mm4, mm2 43 paddd mm4, mm3 48 movq QWORD PTR [rsp], mm4 210 ; movq mm4, [rbx + rdx] 633 movq mm4, mm7 ; 635 psrlq mm4, 32 ; 636 paddd mm4, mm7 ; [all …]
|
/external/boringssl/mac-x86/crypto/sha/ |
D | sha512-586.S | 41 movq 32(%esi),%mm4 76 movq %mm4,%mm1 79 movq %mm4,32(%esp) 80 pand %mm4,%mm5 81 psllq $23,%mm4 87 pxor %mm4,%mm3 88 psllq $23,%mm4 92 pxor %mm4,%mm3 96 psllq $4,%mm4 98 pxor %mm4,%mm3 [all …]
|
/external/boringssl/linux-x86/crypto/sha/ |
D | sha512-586.S | 42 movq 32(%esi),%mm4 77 movq %mm4,%mm1 80 movq %mm4,32(%esp) 81 pand %mm4,%mm5 82 psllq $23,%mm4 88 pxor %mm4,%mm3 89 psllq $23,%mm4 93 pxor %mm4,%mm3 97 psllq $4,%mm4 99 pxor %mm4,%mm3 [all …]
|
/external/boringssl/win-x86/crypto/sha/ |
D | sha512-586.asm | 54 movq mm4,[32+esi] 89 movq mm1,mm4 92 movq [32+esp],mm4 93 pand mm5,mm4 94 psllq mm4,23 100 pxor mm3,mm4 101 psllq mm4,23 105 pxor mm3,mm4 109 psllq mm4,4 111 pxor mm3,mm4 [all …]
|
/external/libvpx/libvpx/vp9/common/x86/ |
D | vp9_loopfilter_mmx.asm | 59 movq mm4, [rsi+rax] ; q1 60 movq mm3, mm4 ; q1 61 psubusb mm4, mm6 ; q1-=q2 63 por mm4, mm6 ; abs(q2-q1) 65 psubusb mm4, mm7 66 por mm1, mm4 68 movq mm4, [rsi] ; q0 69 movq mm0, mm4 ; q0 70 psubusb mm4, mm3 ; q0-=q1 72 por mm4, mm3 ; abs(q0-q1) [all …]
|
/external/libvpx/libvpx/vp8/encoder/x86/ |
D | dct_mmx.asm | 36 movq mm4, [rcx + rax] 45 punpcklwd mm2, mm4 ; 20 30 21 31 46 punpckhwd mm5, mm4 ; 22 32 23 33 65 movq mm4, mm1 70 psubw mm4, mm2 ; c1 = 1 - 2 74 psllw mm4, 3 88 punpcklwd mm1, mm4 ; c1 d1 89 punpckhwd mm5, mm4 ; c1 d1 92 movq mm4, mm5 95 pmaddwd mm4, MMWORD PTR[GLOBAL (_5352_2217)] ; c1*2217 + d1*5352 [all …]
|
D | quantize_mmx.asm | 72 movq mm4, [rsi+8] 77 movq mm7, mm4 78 psraw mm4, 15 80 pxor mm7, mm4 81 psubw mm7, mm4 ; abs 95 pxor mm7, mm4 96 psubw mm7, mm4;gain the sign back 113 movq mm4, [rsi+16] 118 movq mm7, mm4 119 psraw mm4, 15 [all …]
|
D | subtract_mmx.asm | 103 movq mm4, mm3 109 punpckhbw mm4, mm0 112 psubw mm2, mm4 121 movq mm4, mm3 127 punpckhbw mm4, mm0 130 psubw mm2, mm4 177 movq mm4, mm1 181 punpckhbw mm4, mm7 183 psubw mm3, mm4 201 movq mm4, mm1 [all …]
|
D | encodeopt.asm | 79 movq mm4, [rdi] 92 psubw mm3, mm4 100 movq mm4, [rdi+16] 107 psubw mm3, mm4 153 movq mm4, [rdi] 162 psubw mm3, mm4 171 movq mm4, [rdi+16] 178 psubw mm3, mm4 302 movq mm4, [rdi+8] 304 psubw mm3, mm4
|
/external/mesa3d/src/mesa/x86/ |
D | read_rgba_span_x86.S | 126 movq %mm0, %mm4 129 psllq $16, %mm4 131 pand %mm2, %mm4 134 por %mm4, %mm3 227 movq %mm0, %mm4 230 psllq $16, %mm4 232 pand %mm2, %mm4 235 por %mm4, %mm3 260 movq %mm0, %mm4 267 psllq $16, %mm4 [all …]
|
/external/llvm/test/CodeGen/X86/ |
D | stack-folding-3dnow.ll | 6 %1 = tail call x86_mmx asm sideeffect "nop", "=y,~{mm2},~{mm3},~{mm4},~{mm5},~{mm6},~{mm7}"() 15 …%1 = tail call x86_mmx asm sideeffect "nop", "=y,~{mm1},~{mm2},~{mm3},~{mm4},~{mm5},~{mm6},~{mm7}"… 24 …%1 = tail call x86_mmx asm sideeffect "nop", "=y,~{mm1},~{mm2},~{mm3},~{mm4},~{mm5},~{mm6},~{mm7}"… 33 %1 = tail call x86_mmx asm sideeffect "nop", "=y,~{mm2},~{mm3},~{mm4},~{mm5},~{mm6},~{mm7}"() 42 %1 = tail call x86_mmx asm sideeffect "nop", "=y,~{mm2},~{mm3},~{mm4},~{mm5},~{mm6},~{mm7}"() 51 %1 = tail call x86_mmx asm sideeffect "nop", "=y,~{mm2},~{mm3},~{mm4},~{mm5},~{mm6},~{mm7}"() 60 %1 = tail call x86_mmx asm sideeffect "nop", "=y,~{mm2},~{mm3},~{mm4},~{mm5},~{mm6},~{mm7}"() 69 %1 = tail call x86_mmx asm sideeffect "nop", "=y,~{mm2},~{mm3},~{mm4},~{mm5},~{mm6},~{mm7}"() 78 %1 = tail call x86_mmx asm sideeffect "nop", "=y,~{mm2},~{mm3},~{mm4},~{mm5},~{mm6},~{mm7}"() 87 %1 = tail call x86_mmx asm sideeffect "nop", "=y,~{mm2},~{mm3},~{mm4},~{mm5},~{mm6},~{mm7}"() [all …]
|
D | stack-folding-mmx.ll | 15 …%1 = tail call x86_mmx asm sideeffect "nop", "=y,~{mm1},~{mm1},~{mm2},~{mm3},~{mm4},~{mm5},~{mm6},… 24 …%1 = tail call x86_mmx asm sideeffect "nop", "=y,~{mm1},~{mm1},~{mm2},~{mm3},~{mm4},~{mm5},~{mm6},… 65 %1 = tail call x86_mmx asm sideeffect "nop", "=y,~{mm2},~{mm3},~{mm4},~{mm5},~{mm6},~{mm7}"() 74 %1 = tail call x86_mmx asm sideeffect "nop", "=y,~{mm2},~{mm3},~{mm4},~{mm5},~{mm6},~{mm7}"() 83 %1 = tail call x86_mmx asm sideeffect "nop", "=y,~{mm2},~{mm3},~{mm4},~{mm5},~{mm6},~{mm7}"() 92 %1 = tail call x86_mmx asm sideeffect "nop", "=y,~{mm2},~{mm3},~{mm4},~{mm5},~{mm6},~{mm7}"() 101 %1 = tail call x86_mmx asm sideeffect "nop", "=y,~{mm2},~{mm3},~{mm4},~{mm5},~{mm6},~{mm7}"() 110 %1 = tail call x86_mmx asm sideeffect "nop", "=y,~{mm2},~{mm3},~{mm4},~{mm5},~{mm6},~{mm7}"() 119 %1 = tail call x86_mmx asm sideeffect "nop", "=y,~{mm2},~{mm3},~{mm4},~{mm5},~{mm6},~{mm7}"() 128 %1 = tail call x86_mmx asm sideeffect "nop", "=y,~{mm2},~{mm3},~{mm4},~{mm5},~{mm6},~{mm7}"() [all …]
|
D | 2008-02-26-AsmDirectMemOp.ll | 15 …mm4, %mm4 \0A\09.align 1<<4\0A\091: \0A\09movq ($0, $3), …
|
/external/libvpx/libvpx/vpx_scale/win32/ |
D | scaleopt.c | 98 movq mm4, round_values; in horizontal_line_5_4_scale_mmx() local 122 paddw mm1, mm4 in horizontal_line_5_4_scale_mmx() 169 movd mm4, [eax+ecx] in vertical_band_5_4_scale_mmx() local 172 punpcklbw mm4, mm7 in vertical_band_5_4_scale_mmx() local 174 movq mm5, mm4 in vertical_band_5_4_scale_mmx() 175 pmullw mm4, two_fourths in vertical_band_5_4_scale_mmx() local 183 paddw mm3, mm4 in vertical_band_5_4_scale_mmx() 246 movq mm4, round_values; in horizontal_line_5_3_scale_mmx() local 273 paddw mm1, mm4 in horizontal_line_5_3_scale_mmx() 303 paddw mm1, mm4 in horizontal_line_5_3_scale_mmx() [all …]
|
/external/mesa3d/src/mesa/x86-64/ |
D | xform4.S | 260 movq (%rdx), %mm4 /* x1 | x0 */ 265 pfmul %mm0, %mm4 /* x1*m11 | x0*m00 */ 273 pfadd %mm6, %mm4 /* x1*m11+x3*m31 | x0*m00+x3*m30 */ 276 movq %mm4, (%rdi) /* write r0, r1 */ 326 movq (%rdx), %mm4 /* x1 | x0 */ 331 pfmul %mm0, %mm4 /* x1*m11 | x0*m00 */ 339 pfadd %mm4, %mm5 /* x1*m11+x2*m21 | x0*m00+x2*m20 */ 386 movq (%rdx), %mm4 /* x1 | x0 */ 389 pfmul %mm0, %mm4 /* x1*m11 | x0*m00 */ 398 pfadd %mm4, %mm6 /* x1*m11+x3*m31 | x0*m00+x3*m30 */ [all …]
|
/external/libvpx/libvpx/vp9/encoder/x86/ |
D | vp9_sad_mmx.asm | 54 movq mm4, mm0 58 psubusb mm1, mm4 299 movd mm4, DWORD PTR [rsi] 305 punpcklbw mm4, mm6 308 movq mm6, mm4 309 psubusb mm4, mm5 312 por mm4, mm5 314 movq mm5, mm4 315 punpcklbw mm4, mm3 318 paddw mm4, mm5 [all …]
|
/external/flac/libFLAC/ia32/ |
D | fixed_asm.nasm | 101 ; mm4 == last_error_2:last_error_3 117 movq mm4, mm5 ; mm4 = 0:last_error_2 118 psubd mm4, mm2 ; mm4 = 0:last_error_2 - (data[-2] - data[-3]) 119 paddd mm4, mm1 ; mm4 = 0:last_error_2 - (data[-2] - 2 * data[-3]) 120 psubd mm4, mm0 ; mm4 = 0:last_error_3 121 punpckldq mm4, mm5 ; mm4 = last_error_2:last_error_3 142 psubd mm5, mm4 ; mm5 = error_3: 146 psubd mm5, mm4 ; mm5 = :error_4 147 movq mm4, mm6 ; mm4 = error_2:error_3
|