/art/runtime/arch/x86/ |
D | portable_entrypoints_x86.S | 71 movl %esp, %ebp 76 movl 12(%ebp), %ecx // %ecx = receiver 77 movl 0(%edx), %eax // %eax = ArtMethod* called 95 movl %esp, %ebp 100 movl 12(%ebp), %ecx // %ecx = receiver 101 movl 0(%edx), %eax // %eax = ArtMethod* called 120 movl %esp, %ebp 125 movl 0(%edx), %eax // %eax = ArtMethod* called
|
D | quick_entrypoints_x86.S | 220 movl %esp, %edx // remember SP 233 movl %edx, %edi // save code pointer in EDI 290 movl LITERAL(0), (%esp) // store NULL for method* 536 movl LOCK_WORD_OFFSET(%eax), %ecx // ecx := lock word 539 movl %fs:THREAD_ID_OFFSET, %edx // edx := thread id 543 movl %eax, %ecx // remember object in case of retry 549 movl %ecx, %eax // restore eax 557 movl %ecx, LOCK_WORD_OFFSET(%eax) // update lockword, cmpxchg not necessary as we hold lock 579 movl LOCK_WORD_OFFSET(%eax), %ecx // ecx := lock word 580 movl %fs:THREAD_ID_OFFSET, %edx // edx := thread id [all …]
|
D | memcmp16_x86.S | 46 movl LEN(%esp), %ecx 51 movl BLK1(%esp), %eax 53 movl BLK2(%esp), %edx 76 movl %eax, %edi 77 movl %edx, %esi
|
/art/compiler/optimizing/ |
D | code_generator_x86.cc | 261 __ movl(Address(ESP, kCurrentMethodStackOffset), EAX); in GenerateFrameEntry() local 273 __ movl(reg, Address(ESP, kCurrentMethodStackOffset)); in LoadCurrentMethod() local 349 __ movl(destination.AsX86().AsCpuRegister(), source.AsX86().AsCpuRegister()); in Move32() local 352 __ movl(destination.AsX86().AsCpuRegister(), Address(ESP, source.GetStackIndex())); in Move32() local 356 __ movl(Address(ESP, destination.GetStackIndex()), source.AsX86().AsCpuRegister()); in Move32() local 371 __ movl(destination.AsX86().AsRegisterPairLow(), source.AsX86().AsRegisterPairLow()); in Move64() local 372 __ movl(destination.AsX86().AsRegisterPairHigh(), source.AsX86().AsRegisterPairHigh()); in Move64() local 376 __ movl(destination.AsX86().AsRegisterPairLow(), in Move64() local 378 __ movl(destination.AsX86().AsRegisterPairHigh(), Address(ESP, in Move64() local 382 __ movl(destination.AsX86().AsRegisterPairLow(), Address(ESP, source.GetStackIndex())); in Move64() local [all …]
|
D | code_generator_x86_64.cc | 232 __ movl(Address(CpuRegister(RSP), kCurrentMethodStackOffset), CpuRegister(RDI)); in GenerateFrameEntry() local 245 __ movl(reg, Address(CpuRegister(RSP), kCurrentMethodStackOffset)); in LoadCurrentMethod() local 282 …__ movl(destination.AsX86_64().AsCpuRegister(), Address(CpuRegister(RSP), source.GetStackIndex())); in Move() local 289 …__ movl(Address(CpuRegister(RSP), destination.GetStackIndex()), source.AsX86_64().AsCpuRegister()); in Move() local 292 __ movl(CpuRegister(TMP), Address(CpuRegister(RSP), source.GetStackIndex())); in Move() local 293 __ movl(Address(CpuRegister(RSP), destination.GetStackIndex()), CpuRegister(TMP)); in Move() local 313 __ movl(location.AsX86_64().AsCpuRegister(), imm); in Move() local 315 __ movl(Address(CpuRegister(RSP), location.GetStackIndex()), imm); in Move() local 566 __ movl(locations->Out().AsX86_64().AsCpuRegister(), Immediate(0)); in VisitCompare() local 570 __ movl(locations->Out().AsX86_64().AsCpuRegister(), Immediate(-1)); in VisitCompare() local [all …]
|
/art/runtime/arch/x86_64/ |
D | quick_entrypoints_x86_64.S | 328 movl FRAME_SIZE_REFS_AND_ARGS_CALLEE_SAVE(%rsp), %edx // pass caller Method* 405 movl (%r11), REG_VAR(gpr_reg32, 1) 456 movl %edx, %r10d 466 movl LITERAL(0), (%rsp) // Store NULL for method* 468 movl %r10d, %ecx // Place size of args in rcx. 477 movl (%r11), %esi // RSI := this pointer 540 movl %edx, %r10d 550 movl LITERAL(0), (%rsp) // Store NULL for method* 552 movl %r10d, %ecx // Place size of args in rcx. 690 movl 8(%rsp), %esi // pass referrer [all …]
|
/art/compiler/utils/x86/ |
D | assembler_x86.cc | 105 void X86Assembler::movl(Register dst, const Immediate& imm) { in movl() function in art::x86::X86Assembler 112 void X86Assembler::movl(Register dst, Register src) { in movl() function in art::x86::X86Assembler 119 void X86Assembler::movl(Register dst, const Address& src) { in movl() function in art::x86::X86Assembler 126 void X86Assembler::movl(const Address& dst, Register src) { in movl() function in art::x86::X86Assembler 133 void X86Assembler::movl(const Address& dst, const Immediate& imm) { in movl() function in art::x86::X86Assembler 140 void X86Assembler::movl(const Address& dst, Label* lbl) { in movl() function in art::x86::X86Assembler 1425 movl(Address(ESP, frame_size + sizeof(StackReference<mirror::ArtMethod>) + in BuildFrame() 1458 movl(Address(ESP, offs), src.AsCpuRegister()); in Store() 1461 movl(Address(ESP, offs), src.AsRegisterPairLow()); in Store() 1462 movl(Address(ESP, FrameOffset(offs.Int32Value()+4)), in Store() [all …]
|
D | assembler_x86.h | 232 void movl(Register dst, const Immediate& src); 233 void movl(Register dst, Register src); 235 void movl(Register dst, const Address& src); 236 void movl(const Address& dst, Register src); 237 void movl(const Address& dst, const Immediate& imm); 238 void movl(const Address& dst, Label* lbl);
|
/art/compiler/utils/x86_64/ |
D | assembler_x86_64_test.cc | 129 GetAssembler()->movl(x86_64::CpuRegister(x86_64::R8), x86_64::CpuRegister(x86_64::R11)); in TEST_F() 130 GetAssembler()->movl(x86_64::CpuRegister(x86_64::RAX), x86_64::CpuRegister(x86_64::R11)); in TEST_F() 131 GetAssembler()->movl(x86_64::CpuRegister(x86_64::RAX), x86_64::Address( in TEST_F() 133 GetAssembler()->movl(x86_64::CpuRegister(x86_64::RAX), x86_64::Address( in TEST_F() 135 GetAssembler()->movl(x86_64::CpuRegister(x86_64::R8), x86_64::Address( in TEST_F()
|
D | assembler_x86_64.cc | 121 void X86_64Assembler::movl(CpuRegister dst, const Immediate& imm) { in movl() function in art::x86_64::X86_64Assembler 138 void X86_64Assembler::movl(CpuRegister dst, CpuRegister src) { in movl() function in art::x86_64::X86_64Assembler 154 void X86_64Assembler::movl(CpuRegister dst, const Address& src) { in movl() function in art::x86_64::X86_64Assembler 170 void X86_64Assembler::movl(const Address& dst, CpuRegister src) { in movl() function in art::x86_64::X86_64Assembler 177 void X86_64Assembler::movl(const Address& dst, const Immediate& imm) { in movl() function in art::x86_64::X86_64Assembler 1748 movl(Address(CpuRegister(RSP), 0), method_reg.AsX86_64().AsCpuRegister()); in BuildFrame() 1758 …movl(Address(CpuRegister(RSP), frame_size + spill.getSpillOffset()), spill.AsX86_64().AsCpuRegiste… in BuildFrame() 1813 movl(Address(CpuRegister(RSP), offs), src.AsCpuRegister()); in Store() 1842 movl(Address(CpuRegister(RSP), dest), src.AsCpuRegister()); in StoreRef() 1853 movl(Address(CpuRegister(RSP), dest), Immediate(imm)); // TODO(64) movq? in StoreImmediateToFrame() [all …]
|
D | assembler_x86_64.h | 274 void movl(CpuRegister dst, const Immediate& src); 276 void movl(CpuRegister dst, CpuRegister src); 279 void movl(CpuRegister dst, const Address& src); 281 void movl(const Address& dst, CpuRegister src); 282 void movl(const Address& dst, const Immediate& imm);
|