/art/compiler/utils/x86_64/ |
D | jni_macro_assembler_x86_64.cc | 73 __ subq(CpuRegister(RSP), Immediate(rest_of_frame)); in BuildFrame() 83 __ movsd(Address(CpuRegister(RSP), offset), spill.AsXmmRegister()); in BuildFrame() 92 __ movq(Address(CpuRegister(RSP), 0), method_reg.AsX86_64().AsCpuRegister()); in BuildFrame() 109 __ movsd(spill.AsXmmRegister(), Address(CpuRegister(RSP), offset)); in RemoveFrame() 119 __ addq(CpuRegister(RSP), Immediate(offset)); in RemoveFrame() 139 __ addq(CpuRegister(RSP), Immediate(-static_cast<int64_t>(adjust))); in IncreaseFrameSize() 147 assembler->addq(CpuRegister(RSP), Immediate(adjust)); in DecreaseFrameSizeImpl() 163 Store(X86_64ManagedRegister::FromCpuRegister(RSP), MemberOffset(offs.Int32Value()), msrc, size); in Store() 201 __ movq(Address(CpuRegister(RSP), dest), src.AsCpuRegister()); in StoreRawPtr() 207 __ movq(reg, CpuRegister(RSP)); in StoreStackPointerToThread() [all …]
|
D | assembler_x86_64.h | 118 return (rm() == RBP || (rm() == RSP && base() == RBP)) ? disp32() : 0; in disp() 220 CHECK_EQ(base_in.AsRegister(), RSP); in Address() 221 Init(CpuRegister(RSP), disp.Int32Value()); in Address() 231 if (base_in.LowBits() == RSP) { in Init() 232 SetSIB(TIMES_1, CpuRegister(RSP), base_in); in Init() 236 if (base_in.LowBits() == RSP) { in Init() 237 SetSIB(TIMES_1, CpuRegister(RSP), base_in); in Init() 242 if (base_in.LowBits() == RSP) { in Init() 243 SetSIB(TIMES_1, CpuRegister(RSP), base_in); in Init() 250 CHECK_NE(index_in.AsRegister(), RSP); // Illegal addressing mode. in Address() [all …]
|
D | assembler_x86_64_test.cc | 184 x86_64::Address(x86_64::CpuRegister(x86_64::RSP), in SetUpHelpers() 187 x86_64::Address(x86_64::CpuRegister(x86_64::RSP), in SetUpHelpers() 190 x86_64::Address(x86_64::CpuRegister(x86_64::RSP), in SetUpHelpers() 193 x86_64::Address(x86_64::CpuRegister(x86_64::RSP), in SetUpHelpers() 195 addresses_.push_back(x86_64::Address(x86_64::CpuRegister(x86_64::RSP), -1)); in SetUpHelpers() 196 addresses_.push_back(x86_64::Address(x86_64::CpuRegister(x86_64::RSP), 0)); in SetUpHelpers() 197 addresses_.push_back(x86_64::Address(x86_64::CpuRegister(x86_64::RSP), 1)); in SetUpHelpers() 198 addresses_.push_back(x86_64::Address(x86_64::CpuRegister(x86_64::RSP), 987654321)); in SetUpHelpers() 212 secondary_register_names_.emplace(x86_64::CpuRegister(x86_64::RSP), "esp"); in SetUpHelpers() 229 tertiary_register_names_.emplace(x86_64::CpuRegister(x86_64::RSP), "sp"); in SetUpHelpers() [all …]
|
D | assembler_x86_64.cc | 42 if (addr.rm() != RSP || addr.cpu_index().AsRegister() == RSP) { in operator <<() 51 if (addr.rm() != RSP || addr.cpu_index().AsRegister() == RSP) { in operator <<() 57 if (addr.rm() != RSP || addr.cpu_index().AsRegister() == RSP) { in operator <<() 5355 movsd(dst, Address(CpuRegister(RSP), 0)); in LoadDoubleConstant() 5356 addq(CpuRegister(RSP), Immediate(2 * sizeof(intptr_t))); in LoadDoubleConstant()
|
/art/runtime/arch/x86_64/ |
D | context_x86_64.cc | 31 gprs_[RSP] = &rsp_; in Reset() 34 rsp_ = kBadGprBase + RSP; in Reset() 118 uintptr_t rsp = gprs[kNumberOfCpuRegisters - RSP - 1] - sizeof(intptr_t); in DoLongJump()
|
D | registers_x86_64.h | 32 RSP = 4, enumerator
|
D | context_x86_64.h | 41 SetGPR(RSP, new_sp); in SetSP()
|
/art/compiler/optimizing/ |
D | code_generator_x86_64.cc | 581 DCHECK_NE(ref_reg, RSP); in EmitNativeCode() 673 DCHECK_NE(ref_reg, RSP); in EmitNativeCode() 1519 __ movq(Address(CpuRegister(RSP), stack_index), CpuRegister(reg_id)); in SaveCoreRegister() 1524 __ movq(CpuRegister(reg_id), Address(CpuRegister(RSP), stack_index)); in RestoreCoreRegister() 1530 __ movups(Address(CpuRegister(RSP), stack_index), XmmRegister(reg_id)); in SaveFloatingPointRegister() 1532 __ movsd(Address(CpuRegister(RSP), stack_index), XmmRegister(reg_id)); in SaveFloatingPointRegister() 1539 __ movups(XmmRegister(reg_id), Address(CpuRegister(RSP), stack_index)); in RestoreFloatingPointRegister() 1541 __ movsd(XmmRegister(reg_id), Address(CpuRegister(RSP), stack_index)); in RestoreFloatingPointRegister() 1643 blocked_core_registers_[RSP] = true; in SetupBlockedRegisters() 1678 __ cmpl(Address(CpuRegister(RSP), codegen_->GetStackOffsetOfShouldDeoptimizeFlag()), in GenerateMethodEntryExitHook() [all …]
|
D | intrinsics_x86_64.cc | 2884 __ popcntq(out, Address(CpuRegister(RSP), src.GetStackIndex())); in GenBitCount() 2887 __ popcntl(out, Address(CpuRegister(RSP), src.GetStackIndex())); in GenBitCount() 2963 __ bsrq(tmp, Address(CpuRegister(RSP), src.GetStackIndex())); in GenOneBit() 2966 __ bsrl(tmp, Address(CpuRegister(RSP), src.GetStackIndex())); in GenOneBit() 2991 __ movq(tmp, Address(CpuRegister(RSP), src.GetStackIndex())); in GenOneBit() 2994 __ movl(tmp, Address(CpuRegister(RSP), src.GetStackIndex())); in GenOneBit() 3078 __ bsrq(out, Address(CpuRegister(RSP), src.GetStackIndex())); in GenLeadingZeros() 3081 __ bsrl(out, Address(CpuRegister(RSP), src.GetStackIndex())); in GenLeadingZeros() 3151 __ bsfq(out, Address(CpuRegister(RSP), src.GetStackIndex())); in GenTrailingZeros() 3154 __ bsfl(out, Address(CpuRegister(RSP), src.GetStackIndex())); in GenTrailingZeros()
|
D | code_generator_x86_64.h | 699 assembler_.lock()->addl(Address(CpuRegister(RSP), 0), Immediate(0));
|