/external/mesa3d/src/mesa/x86/ |
D | sse_normal.S | 76 MOVSS ( M(0), XMM1 ) /* m0 */ 78 UNPCKLPS( XMM2, XMM1 ) /* m5 | m0 */ 81 MULPS ( XMM0, XMM1 ) /* m5*scale | m0*scale */ 87 MULPS ( XMM1, XMM2 ) /* uy*m5*scale | ux*m0*scale */ 139 MOVSS ( M(4), XMM1 ) /* m4 */ 140 UNPCKLPS( XMM1, XMM0 ) /* m4 | m0 */ 146 MOVSS ( M(1), XMM1 ) /* m1 */ 148 UNPCKLPS( XMM2, XMM1 ) /* m5 | m1 */ 149 MULPS ( XMM4, XMM1 ) /* m5*scale | m1*scale */ 167 MULPS ( XMM1, XMM4 ) /* uy*m5 | uy*m1 */ [all …]
|
D | sse_xform2.S | 77 MOVAPS( M(4), XMM1 ) /* m7 | m6 | m5 | m4 */ 87 MULPS( XMM1, XMM4 ) /* oy*m7 | oy*m6 | oy*m5 | oy*m4 */ 192 MOVSS ( M(0), XMM1 ) /* - | - | - | m0 */ 194 UNPCKLPS ( XMM2, XMM1 ) /* - | - | m5 | m0 */ 201 MULPS ( XMM1, XMM0 ) /* - | - | oy*m5 | ox*m0 */ 251 MOVSS ( M(0), XMM1 ) /* - | - | - | m0 */ 253 UNPCKLPS ( XMM2, XMM1 ) /* - | - | m5 | m0 */ 260 MULPS( XMM1, XMM4 ) /* oy*m5 | ox*m0 */ 311 MOVLPS( M(4), XMM1 ) /* m5 | m4 */ 322 MULPS( XMM1, XMM4 ) /* oy*m5 | oy*m4 */ [all …]
|
D | sse_xform3.S | 78 MOVAPS ( REGOFF(16, EDX), XMM1 ) /* m4 | m5 | m6 | m7 */ 93 MULPS ( XMM1, XMM5 ) /* m7*oy | m6*oy | m5*oy | m4*oy */ 204 MOVSS ( M(0), XMM1 ) /* - | - | - | m0 */ 206 UNPCKLPS ( XMM2, XMM1 ) /* - | - | m5 | m0 */ 215 MULPS ( XMM1, XMM0 ) /* - | - | s1*m5 | s0*m0 */ 269 MOVSS ( M(0), XMM1 ) /* - | - | - | m0 */ 271 UNPCKLPS ( XMM2, XMM1 ) /* - | - | m5 | m0 */ 280 MULPS ( XMM1, XMM0 ) /* oy*m5 | ox*m0 */ 343 MOVLPS( M(4), XMM1 ) /* m5 | m4 */ 353 MULPS ( XMM1, XMM4 ) /* oy*m5 | oy*m4 */ [all …]
|
D | sse_xform1.S | 78 MOVAPS( M(12), XMM1 ) /* m15 | m14 | m13 | m12 */ 85 ADDPS( XMM1, XMM2 ) /* + | + | + | + */ 187 MOVSS( M(12), XMM1 ) /* m12 */ 195 ADDSS( XMM1, XMM4 ) /* ox*m0+m12 */ 248 MOVSS( M(0), XMM1 ) /* m0 */ 254 MULSS( XMM1, XMM3 ) /* ox*m0 */ 306 MOVLPS( M(12), XMM1 ) /* m13 | m12 */ 313 ADDPS( XMM1, XMM2 ) /* - | - | ox*m1+m13 | ox*m0+m12 */ 361 MOVSS( M(12), XMM1 ) /* m12 */ 368 ADDSS( XMM1, XMM3 ) /* ox*m0+m12 */ [all …]
|
D | sse_xform4.S | 82 MOVSS( SRC(1), XMM1 ) /* oy */ 83 SHUFPS( CONST(0x0), XMM1, XMM1 ) /* oy | oy | oy | oy */ 84 MULPS( XMM5, XMM1 ) /* oy*m7 | oy*m6 | oy*m5 | oy*m4 */ 94 ADDPS( XMM1, XMM0 ) /* ox*m3+oy*m7 | ... */ 144 MOVAPS( MAT(4), XMM1 ) /* m7 | m6 | m5 | m4 */ 156 MULPS( XMM1, XMM5 ) /* oy*m7 | oy*m6 | oy*m5 | oy*m4 */
|
D | common_x86_asm.S | 204 MOVUPS ( REGIND( ESP ), XMM1 ) 206 DIVPS ( XMM0, XMM1 )
|
/external/swiftshader/third_party/LLVM/lib/Target/X86/ |
D | X86CallingConv.td | 37 // Vector types are returned in XMM0 and XMM1, when they fit. XMM2 and XMM3 41 CCAssignToReg<[XMM0,XMM1,XMM2,XMM3]>>, 43 // 256-bit vectors are returned in YMM0 and XMM1, when they fit. YMM2 and YMM3 65 CCIfType<[f32, f64], CCAssignToReg<[XMM0,XMM1,XMM2]>>>>, 76 CCIfType<[f32], CCIfSubtarget<"hasXMMInt()", CCAssignToReg<[XMM0,XMM1,XMM2]>>>, 77 CCIfType<[f64], CCIfSubtarget<"hasXMMInt()", CCAssignToReg<[XMM0,XMM1,XMM2]>>>, 91 CCIfType<[f32], CCAssignToReg<[XMM0, XMM1]>>, 92 CCIfType<[f64], CCAssignToReg<[XMM0, XMM1]>>, 95 CCIfType<[x86mmx], CCAssignToReg<[XMM0, XMM1]>>, 159 CCAssignToReg<[XMM0, XMM1, XMM2, XMM3, XMM4, XMM5, XMM6, XMM7]>>>, [all …]
|
D | X86GenCallingConv.inc | 176 X86::XMM0, X86::XMM1, X86::XMM2 227 X86::XMM0, X86::XMM1, X86::XMM2, X86::XMM3 329 X86::XMM0, X86::XMM1, X86::XMM2 548 X86::XMM0, X86::XMM1, X86::XMM2, X86::XMM3, X86::XMM4, X86::XMM5, X86::XMM6, X86::XMM7 653 X86::XMM1, X86::XMM2, X86::XMM3, X86::XMM4, X86::XMM5, X86::XMM6 708 X86::XMM0, X86::XMM1, X86::XMM2, X86::XMM3 723 X86::XMM1, X86::XMM2, X86::XMM3 738 X86::XMM0, X86::XMM1, X86::XMM2, X86::XMM3 755 X86::XMM0, X86::XMM1, X86::XMM2, X86::XMM3 854 X86::XMM0, X86::XMM1, X86::XMM2, X86::XMM3 [all …]
|
D | X86InstrControl.td | 143 XMM0, XMM1, XMM2, XMM3, XMM4, XMM5, XMM6, XMM7, 182 XMM0, XMM1, XMM2, XMM3, XMM4, XMM5, XMM6, XMM7, 218 XMM0, XMM1, XMM2, XMM3, XMM4, XMM5, XMM6, XMM7, 250 XMM0, XMM1, XMM2, XMM3, XMM4, XMM5, EFLAGS], 283 XMM0, XMM1, XMM2, XMM3, XMM4, XMM5, EFLAGS],
|
D | X86GenRegisterInfo.inc | 148 XMM1 = 129, 377 const unsigned XMM1_Overlaps[] = { X86::XMM1, X86::YMM1, 0 }; 393 const unsigned YMM1_Overlaps[] = { X86::YMM1, X86::XMM1, 0 }; 476 const unsigned YMM1_SubRegsSet[] = { X86::XMM1, 0 }; 694 { "XMM1", XMM1_Overlaps, XMM1_SubRegsSet, XMM1_SuperRegsSet }, 770 …X86::XMM0, X86::XMM1, X86::XMM2, X86::XMM3, X86::XMM4, X86::XMM5, X86::XMM6, X86::XMM7, X86::XMM8,… 790 …X86::XMM0, X86::XMM1, X86::XMM2, X86::XMM3, X86::XMM4, X86::XMM5, X86::XMM6, X86::XMM7, X86::XMM8,… 810 …X86::XMM0, X86::XMM1, X86::XMM2, X86::XMM3, X86::XMM4, X86::XMM5, X86::XMM6, X86::XMM7, X86::XMM8,… 1215 RI->mapDwarfRegToLLVMReg(18, X86::XMM1, false ); 1258 RI->mapDwarfRegToLLVMReg(22, X86::XMM1, false ); [all …]
|
/external/llvm/lib/Target/X86/ |
D | X86CallingConv.td | 53 // Vector types are returned in XMM0 and XMM1, when they fit. XMM2 and XMM3 57 CCAssignToReg<[XMM0,XMM1,XMM2,XMM3]>>, 59 // 256-bit vectors are returned in YMM0 and XMM1, when they fit. YMM2 and YMM3 87 CCIfType<[f32, f64], CCAssignToReg<[XMM0,XMM1,XMM2]>>>>, 98 CCIfType<[f32], CCIfSubtarget<"hasSSE2()", CCAssignToReg<[XMM0,XMM1,XMM2]>>>, 99 CCIfType<[f64], CCIfSubtarget<"hasSSE2()", CCAssignToReg<[XMM0,XMM1,XMM2]>>>, 112 // Vector types are returned in XMM0,XMM1,XMMM2 and XMM3. 114 CCAssignToReg<[XMM0,XMM1,XMM2,XMM3]>>, 140 // Vector types are returned in XMM0,XMM1,XMMM2 and XMM3. 142 CCAssignToReg<[XMM0,XMM1,XMM2,XMM3]>>, [all …]
|
/external/swiftshader/third_party/LLVM/test/MC/Disassembler/X86/ |
D | enhanced.txt | 7 …:xorps][w: ][2-r:%xmm1=r129][p:,][w: ][0-r:%xmm2=r130] 0:[XMM2/130]=0 1:[XMM2/130]=0 2:[XMM1/129]=0 9 …:andps][w: ][2-r:%xmm1=r129][p:,][w: ][0-r:%xmm2=r130] 0:[XMM2/130]=0 1:[XMM2/130]=0 2:[XMM1/129]=0
|
/external/swiftshader/third_party/LLVM/test/TableGen/ |
D | Slice.td | 42 def XMM1: Register<"xmm1">; 59 [XMM0, XMM1, XMM2, XMM3, XMM4, XMM5, XMM6, XMM7,
|
D | cast.td | 42 def XMM1: Register<"xmm1">; 59 [XMM0, XMM1, XMM2, XMM3, XMM4, XMM5, XMM6, XMM7,
|
D | TargetInstrSpec.td | 43 def XMM1: Register<"xmm1">; 60 [XMM0, XMM1, XMM2, XMM3, XMM4, XMM5, XMM6, XMM7,
|
D | MultiPat.td | 51 def XMM1: Register<"xmm1">; 68 [XMM0, XMM1, XMM2, XMM3, XMM4, XMM5, XMM6, XMM7,
|
/external/llvm/test/TableGen/ |
D | Slice.td | 41 def XMM1: Register<"xmm1">; 58 [XMM0, XMM1, XMM2, XMM3, XMM4, XMM5, XMM6, XMM7,
|
D | TargetInstrSpec.td | 48 def XMM1: Register<"xmm1">; 65 [XMM0, XMM1, XMM2, XMM3, XMM4, XMM5, XMM6, XMM7,
|
D | cast.td | 47 def XMM1: Register<"xmm1">; 64 [XMM0, XMM1, XMM2, XMM3, XMM4, XMM5, XMM6, XMM7,
|
D | MultiPat.td | 51 def XMM1: Register<"xmm1">; 68 [XMM0, XMM1, XMM2, XMM3, XMM4, XMM5, XMM6, XMM7,
|
/external/llvm/test/MC/X86/ |
D | intel-syntax-encoding.s | 65 cmpltps XMM2, XMM1
|
/external/llvm/test/CodeGen/X86/ |
D | avx512bw-mov.ll | 163 ; CHECK-NEXT: ## kill: %XMM1<def> %XMM1<kill> %ZMM1<def> 193 ; CHECK-NEXT: ## kill: %XMM1<def> %XMM1<kill> %ZMM1<def>
|
D | break-false-dep.ll | 75 ; SSE: xorps [[XMM1:%xmm[0-9]+]] 76 ; SSE: , [[XMM1]] 77 ; SSE: cvtsi2ssl %{{.*}}, [[XMM1]]
|
D | ghc-cc64.ll | 16 @f1 = external global float ; assigned to register: XMM1
|
/external/swiftshader/third_party/LLVM/test/CodeGen/X86/ |
D | ghc-cc64.ll | 16 @f1 = external global float ; assigned to register: XMM1
|