/art/compiler/trampolines/ |
D | trampoline_compiler.cc | 27 #define __ assembler-> 34 std::unique_ptr<ArmAssembler> assembler(static_cast<ArmAssembler*>(Assembler::Create(kThumb2))); in CreateTrampoline() local 49 size_t cs = assembler->CodeSize(); in CreateTrampoline() 52 assembler->FinalizeInstructions(code); in CreateTrampoline() 61 …std::unique_ptr<Arm64Assembler> assembler(static_cast<Arm64Assembler*>(Assembler::Create(kArm64))); in CreateTrampoline() local 85 assembler->EmitSlowPaths(); in CreateTrampoline() 86 size_t cs = assembler->CodeSize(); in CreateTrampoline() 89 assembler->FinalizeInstructions(code); in CreateTrampoline() 98 std::unique_ptr<MipsAssembler> assembler(static_cast<MipsAssembler*>(Assembler::Create(kMips))); in CreateTrampoline() local 115 size_t cs = assembler->CodeSize(); in CreateTrampoline() [all …]
|
/art/compiler/utils/ |
D | assembler_thumb_test.cc | 159 #define __ assembler-> 162 arm::Thumb2Assembler* assembler = static_cast<arm::Thumb2Assembler*>(Assembler::Create(kThumb2)); in TEST() local 175 delete assembler; in TEST() 179 arm::Thumb2Assembler* assembler = static_cast<arm::Thumb2Assembler*>(Assembler::Create(kThumb2)); in TEST() local 180 assembler->Force32Bit(); in TEST() 190 delete assembler; in TEST() 194 arm::Thumb2Assembler* assembler = static_cast<arm::Thumb2Assembler*>(Assembler::Create(kThumb2)); in TEST() local 205 delete assembler; in TEST() 209 arm::Thumb2Assembler* assembler = static_cast<arm::Thumb2Assembler*>(Assembler::Create(kThumb2)); in TEST() local 252 delete assembler; in TEST() [all …]
|
D | assembler_test.h | 52 typedef std::string (*TestFn)(AssemblerTest* assembler_test, Ass* assembler);
|
/art/compiler/linker/arm/ |
D | relative_patcher_thumb2.cc | 80 arm::Thumb2Assembler assembler; in CompileThunkCode() local 81 assembler.LoadFromOffset( in CompileThunkCode() 84 assembler.bkpt(0); in CompileThunkCode() 85 std::vector<uint8_t> thunk_code(assembler.CodeSize()); in CompileThunkCode() 87 assembler.FinalizeInstructions(code); in CompileThunkCode()
|
/art/compiler/utils/x86_64/ |
D | assembler_x86_64_test.cc | 361 std::string shll_fn(AssemblerX86_64Test::Base* assembler_test, x86_64::X86_64Assembler* assembler) { in shll_fn() argument 368 assembler->shll(*reg, shifter); in shll_fn() 384 std::string shlq_fn(AssemblerX86_64Test::Base* assembler_test, x86_64::X86_64Assembler* assembler) { in shlq_fn() argument 391 assembler->shlq(*reg, shifter); in shlq_fn() 407 std::string shrl_fn(AssemblerX86_64Test::Base* assembler_test, x86_64::X86_64Assembler* assembler) { in shrl_fn() argument 414 assembler->shrl(*reg, shifter); in shrl_fn() 430 std::string shrq_fn(AssemblerX86_64Test::Base* assembler_test, x86_64::X86_64Assembler* assembler) { in shrq_fn() argument 437 assembler->shrq(*reg, shifter); in shrq_fn() 453 std::string sarl_fn(AssemblerX86_64Test::Base* assembler_test, x86_64::X86_64Assembler* assembler) { in sarl_fn() argument 460 assembler->sarl(*reg, shifter); in sarl_fn() [all …]
|
/art/compiler/optimizing/ |
D | intrinsics_arm.cc | 138 #define __ assembler-> 156 static void MoveFPToInt(LocationSummary* locations, bool is64bit, ArmAssembler* assembler) { in MoveFPToInt() argument 168 static void MoveIntToFP(LocationSummary* locations, bool is64bit, ArmAssembler* assembler) { in MoveIntToFP() argument 224 static void MathAbsFP(LocationSummary* locations, bool is64bit, ArmAssembler* assembler) { in MathAbsFP() argument 264 ArmAssembler* assembler) { in GenAbsInteger() argument 312 ArmAssembler* assembler) { in GenMinMax() argument 355 ArmAssembler* assembler = GetAssembler(); in VisitMathSqrt() local 365 ArmAssembler* assembler = GetAssembler(); in VisitMemoryPeekByte() local 376 ArmAssembler* assembler = GetAssembler(); in VisitMemoryPeekIntNative() local 387 ArmAssembler* assembler = GetAssembler(); in VisitMemoryPeekLongNative() local [all …]
|
D | intrinsics_x86_64.cc | 159 #define __ assembler-> 177 static void MoveFPToInt(LocationSummary* locations, bool is64bit, X86_64Assembler* assembler) { in MoveFPToInt() argument 183 static void MoveIntToFP(LocationSummary* locations, bool is64bit, X86_64Assembler* assembler) { in MoveIntToFP() argument 227 X86_64Assembler* assembler) { in GenReverseBytes() argument 290 X86_64Assembler* assembler, in MathAbsFP() argument 350 static void GenAbsInteger(LocationSummary* locations, bool is64bit, X86_64Assembler* assembler) { in GenAbsInteger() argument 391 X86_64Assembler* assembler, in GenMinMaxFP() argument 516 X86_64Assembler* assembler) { in GenMinMax() argument 644 X86_64Assembler* assembler, in GenSSE41FPToFPIntrinsic() argument 720 X86_64Assembler* assembler = GetAssembler(); in VisitMathRoundFloat() local [all …]
|
D | intrinsics_x86.cc | 168 #define __ assembler-> 193 static void MoveFPToInt(LocationSummary* locations, bool is64bit, X86Assembler* assembler) { in MoveFPToInt() argument 208 static void MoveIntToFP(LocationSummary* locations, bool is64bit, X86Assembler* assembler) { in MoveIntToFP() argument 278 X86Assembler* assembler) { in GenReverseBytes() argument 317 X86Assembler* assembler = GetAssembler(); in VisitLongReverseBytes() local 348 static void MathAbsFP(LocationSummary* locations, bool is64bit, X86Assembler* assembler) { in MathAbsFP() argument 405 static void GenAbsInteger(LocationSummary* locations, X86Assembler* assembler) { in GenAbsInteger() argument 433 static void GenAbsLong(LocationSummary* locations, X86Assembler* assembler) { in GenAbsLong() argument 476 X86Assembler* assembler) { in GenMinMaxFP() argument 605 X86Assembler* assembler) { in GenMinMax() argument [all …]
|
D | code_generator_x86_64.cc | 4463 X86_64Assembler* assembler = GetAssembler(); in Finalize() local 4464 if (!assembler->IsConstantAreaEmpty()) { in Finalize() 4468 assembler->Align(4, 0); in Finalize() 4469 constant_area_start_ = assembler->CodeSize(); in Finalize() 4470 assembler->AddConstantArea(); in Finalize()
|
/art/compiler/linker/arm64/ |
D | relative_patcher_arm64.cc | 231 arm64::Arm64Assembler assembler; in CompileThunkCode() local 234 assembler.JumpTo(ManagedRegister(arm64::X0), offset, ManagedRegister(arm64::IP0)); in CompileThunkCode() 236 assembler.EmitSlowPaths(); in CompileThunkCode() 237 std::vector<uint8_t> thunk_code(assembler.CodeSize()); in CompileThunkCode() 239 assembler.FinalizeInstructions(code); in CompileThunkCode()
|
/art/compiler/utils/arm/ |
D | assembler_arm.cc | 572 static void EmitLoad(ArmAssembler* assembler, ManagedRegister m_dst, in EmitLoad() argument 579 assembler->LoadFromOffset(kLoadWord, dst.AsCoreRegister(), src_register, src_offset); in EmitLoad() 582 assembler->LoadFromOffset(kLoadWord, dst.AsRegisterPairLow(), src_register, src_offset); in EmitLoad() 583 assembler->LoadFromOffset(kLoadWord, dst.AsRegisterPairHigh(), src_register, src_offset + 4); in EmitLoad() 585 assembler->LoadSFromOffset(dst.AsSRegister(), src_register, src_offset); in EmitLoad() 588 assembler->LoadDFromOffset(dst.AsDRegister(), src_register, src_offset); in EmitLoad()
|
D | assembler_thumb2.h | 522 Branch(const Thumb2Assembler* assembler, Type type, uint32_t location, Condition cond = AL) : 523 assembler_(assembler), type_(type), location_(location), in assembler_() argument 531 Branch(const Thumb2Assembler* assembler, Type type, uint32_t location, Register rn) : in Branch() argument 532 assembler_(assembler), type_(type), location_(location), in Branch() 539 Branch(const Thumb2Assembler* assembler, Type type, uint32_t location, uint32_t target, 541 assembler_(assembler), type_(type), location_(location), in assembler_() argument
|
/art/compiler/ |
D | Android.mk | 145 utils/assembler.cc \
|