/art/compiler/utils/arm/ |
D | assembler_thumb2_test.cc | 94 GetAssembler()->sbfx(arm::R0, arm::R1, 0, 1); in TEST_F() 95 GetAssembler()->sbfx(arm::R0, arm::R1, 0, 8); in TEST_F() 96 GetAssembler()->sbfx(arm::R0, arm::R1, 0, 16); in TEST_F() 97 GetAssembler()->sbfx(arm::R0, arm::R1, 0, 32); in TEST_F() 99 GetAssembler()->sbfx(arm::R0, arm::R1, 8, 1); in TEST_F() 100 GetAssembler()->sbfx(arm::R0, arm::R1, 8, 8); in TEST_F() 101 GetAssembler()->sbfx(arm::R0, arm::R1, 8, 16); in TEST_F() 102 GetAssembler()->sbfx(arm::R0, arm::R1, 8, 24); in TEST_F() 104 GetAssembler()->sbfx(arm::R0, arm::R1, 16, 1); in TEST_F() 105 GetAssembler()->sbfx(arm::R0, arm::R1, 16, 8); in TEST_F() [all …]
|
D | assembler_arm32_test.cc | 402 return std::bind(f, GetAssembler(), _1, _2); in GetBoundFunction2() 407 return std::bind(f, GetAssembler(), _1, _2, _3); in GetBoundFunction3() 413 return std::bind(f, GetAssembler(), _1, _2, _3, _4); in GetBoundFunction4() 419 return std::bind(f, GetAssembler(), _1, _2, _3, _4, _5); in GetBoundFunction5() 693 GetAssembler()->vmstat(); in TEST_F() 701 GetAssembler()->ldrexd(arm::R0, arm::R1, arm::R0); in TEST_F() 702 GetAssembler()->ldrexd(arm::R0, arm::R1, arm::R1); in TEST_F() 703 GetAssembler()->ldrexd(arm::R0, arm::R1, arm::R2); in TEST_F() 713 GetAssembler()->strexd(arm::R9, arm::R0, arm::R1, arm::R0); in TEST_F() 714 GetAssembler()->strexd(arm::R9, arm::R0, arm::R1, arm::R1); in TEST_F() [all …]
|
D | assembler_arm_test.h | 147 (Base::GetAssembler()->*f)(*reg1, *reg2, i, j, c); in RepeatTemplatedRRIIC() 239 (Base::GetAssembler()->*f)(*reg1, *reg2, i, j, c); in RepeatTemplatedRRiiC() 302 (Base::GetAssembler()->*f)(*reg1, *reg2, c); in RepeatTemplatedRRC() 377 (Base::GetAssembler()->*f)(*reg1, *reg2, *reg3, c); in RepeatTemplatedRRRC() 432 (Base::GetAssembler()->*f)(*reg, shift, c); in RepeatTemplatedRSC() 497 (Base::GetAssembler()->*f)(*reg1, *reg2, shift, c); in RepeatTemplatedRRSC()
|
/art/compiler/utils/x86/ |
D | assembler_x86_test.cc | 103 GetAssembler()->movl(x86::EAX, x86::EBX); in TEST_F() 109 GetAssembler()->psrlq(x86::XMM0, CreateImmediate(32)); in TEST_F() 115 GetAssembler()->punpckldq(x86::XMM0, x86::XMM1); in TEST_F() 121 GetAssembler()->LoadLongConstant(x86::XMM0, 51); in TEST_F() 131 GetAssembler()->LockCmpxchgl(x86::Address( in TEST_F() 134 GetAssembler()->LockCmpxchgl(x86::Address( in TEST_F() 137 GetAssembler()->LockCmpxchgl(x86::Address( in TEST_F() 140 GetAssembler()->LockCmpxchgl(x86::Address( in TEST_F() 142 GetAssembler()->LockCmpxchgl(x86::Address( in TEST_F() 156 GetAssembler()->LockCmpxchg8b(x86::Address( in TEST_F() [all …]
|
/art/compiler/optimizing/ |
D | intrinsics_x86_64.cc | 41 X86_64Assembler* IntrinsicCodeGeneratorX86_64::GetAssembler() { in GetAssembler() function in art::x86_64::IntrinsicCodeGeneratorX86_64 42 return reinterpret_cast<X86_64Assembler*>(codegen_->GetAssembler()); in GetAssembler() 55 #define __ reinterpret_cast<X86_64Assembler*>(codegen->GetAssembler())-> 197 MoveFPToInt(invoke->GetLocations(), true, GetAssembler()); in VisitDoubleDoubleToRawLongBits() 200 MoveIntToFP(invoke->GetLocations(), true, GetAssembler()); in VisitDoubleLongBitsToDouble() 211 MoveFPToInt(invoke->GetLocations(), false, GetAssembler()); in VisitFloatFloatToRawIntBits() 214 MoveIntToFP(invoke->GetLocations(), false, GetAssembler()); in VisitFloatIntBitsToFloat() 253 GenReverseBytes(invoke->GetLocations(), Primitive::kPrimInt, GetAssembler()); in VisitIntegerReverseBytes() 261 GenReverseBytes(invoke->GetLocations(), Primitive::kPrimLong, GetAssembler()); in VisitLongReverseBytes() 269 GenReverseBytes(invoke->GetLocations(), Primitive::kPrimShort, GetAssembler()); in VisitShortReverseBytes() [all …]
|
D | intrinsics_x86.cc | 45 X86Assembler* IntrinsicCodeGeneratorX86::GetAssembler() { in GetAssembler() function in art::x86::IntrinsicCodeGeneratorX86 46 return reinterpret_cast<X86Assembler*>(codegen_->GetAssembler()); in GetAssembler() 59 #define __ reinterpret_cast<X86Assembler*>(codegen->GetAssembler())-> 232 MoveFPToInt(invoke->GetLocations(), true, GetAssembler()); in VisitDoubleDoubleToRawLongBits() 235 MoveIntToFP(invoke->GetLocations(), true, GetAssembler()); in VisitDoubleLongBitsToDouble() 246 MoveFPToInt(invoke->GetLocations(), false, GetAssembler()); in VisitFloatFloatToRawIntBits() 249 MoveIntToFP(invoke->GetLocations(), false, GetAssembler()); in VisitFloatIntBitsToFloat() 301 GenReverseBytes(invoke->GetLocations(), Primitive::kPrimInt, GetAssembler()); in VisitIntegerReverseBytes() 317 X86Assembler* assembler = GetAssembler(); in VisitLongReverseBytes() 330 GenReverseBytes(invoke->GetLocations(), Primitive::kPrimShort, GetAssembler()); in VisitShortReverseBytes() [all …]
|
D | intrinsics_arm.cc | 33 ArmAssembler* IntrinsicCodeGeneratorARM::GetAssembler() { in GetAssembler() function in art::arm::IntrinsicCodeGeneratorARM 34 return codegen_->GetAssembler(); in GetAssembler() 41 #define __ codegen->GetAssembler()-> 188 MoveFPToInt(invoke->GetLocations(), true, GetAssembler()); in VisitDoubleDoubleToRawLongBits() 191 MoveIntToFP(invoke->GetLocations(), true, GetAssembler()); in VisitDoubleLongBitsToDouble() 202 MoveFPToInt(invoke->GetLocations(), false, GetAssembler()); in VisitFloatFloatToRawIntBits() 205 MoveIntToFP(invoke->GetLocations(), false, GetAssembler()); in VisitFloatIntBitsToFloat() 241 MathAbsFP(invoke->GetLocations(), true, GetAssembler()); in VisitMathAbsDouble() 249 MathAbsFP(invoke->GetLocations(), false, GetAssembler()); in VisitMathAbsFloat() 298 GenAbsInteger(invoke->GetLocations(), false, GetAssembler()); in VisitMathAbsInt() [all …]
|
D | code_generator_arm64.h | 151 Arm64Assembler* GetAssembler() const { return assembler_; } in GetAssembler() function 152 vixl::MacroAssembler* GetVIXLAssembler() { return GetAssembler()->vixl_masm_; } in GetVIXLAssembler() 211 Arm64Assembler* GetAssembler() const; 213 return GetAssembler()->vixl_masm_; in GetVIXLAssembler() 267 Arm64Assembler* GetAssembler() OVERRIDE { return &assembler_; } in GetAssembler() function 268 vixl::MacroAssembler* GetVIXLAssembler() { return GetAssembler()->vixl_masm_; } in GetVIXLAssembler() 359 inline Arm64Assembler* ParallelMoveResolverARM64::GetAssembler() const { in GetAssembler() function 360 return codegen_->GetAssembler(); in GetAssembler()
|
D | intrinsics_x86.h | 72 X86Assembler* GetAssembler();
|
D | intrinsics_x86_64.h | 72 X86_64Assembler* GetAssembler();
|
D | code_generator_arm.h | 109 ArmAssembler* GetAssembler() const; 170 ArmAssembler* GetAssembler() const { return assembler_; } in FOR_EACH_CONCRETE_INSTRUCTION() 235 ArmAssembler* GetAssembler() OVERRIDE { in GetAssembler() function
|
D | code_generator_mips64.h | 120 Mips64Assembler* GetAssembler() const; 179 Mips64Assembler* GetAssembler() const { return assembler_; } in FOR_EACH_CONCRETE_INSTRUCTION() 229 Mips64Assembler* GetAssembler() OVERRIDE { return &assembler_; } in GetAssembler() function
|
D | code_generator_x86.h | 103 X86Assembler* GetAssembler() const; 153 X86Assembler* GetAssembler() const { return assembler_; } in FOR_EACH_CONCRETE_INSTRUCTION() 228 X86Assembler* GetAssembler() OVERRIDE { in GetAssembler() function
|
D | code_generator_x86_64.h | 112 X86_64Assembler* GetAssembler() const; 163 X86_64Assembler* GetAssembler() const { return assembler_; } in FOR_EACH_CONCRETE_INSTRUCTION() 228 X86_64Assembler* GetAssembler() OVERRIDE { in GetAssembler() function
|
D | intrinsics_arm.h | 76 ArmAssembler* GetAssembler();
|
D | optimizing_cfi_test.cc | 82 Assembler* opt_asm = code_gen->GetAssembler(); in TestImpl()
|
D | optimizing_compiler.cc | 436 ArrayRef<const uint8_t>(*codegen->GetAssembler()->cfi().data()), in CompileOptimized() 473 ArrayRef<const uint8_t>(*codegen->GetAssembler()->cfi().data()), in CompileBaseline() 551 codegen->GetAssembler()->cfi().SetEnabled( in TryCompile()
|
D | intrinsics_arm64.cc | 59 return codegen_->GetAssembler()->vixl_masm_; in GetVIXLAssembler() 66 #define __ codegen->GetAssembler()->vixl_masm_-> 666 vixl::MacroAssembler* masm = codegen->GetAssembler()->vixl_masm_; in GenUnsafeGet() 776 vixl::MacroAssembler* masm = codegen->GetAssembler()->vixl_masm_; in GenUnsafePut() 847 vixl::MacroAssembler* masm = codegen->GetAssembler()->vixl_masm_; in GenCas()
|
D | code_generator.cc | 167 DCHECK_EQ(GetAssembler()->cfi().GetCurrentCFAOffset(), static_cast<int>(frame_size_)); in CompileInternal() 203 size_t code_size = GetAssembler()->CodeSize(); in Finalize() 207 GetAssembler()->FinalizeInstructions(code); in Finalize() 663 pc_info.native_pc = GetAssembler()->CodeSize(); in RecordPcInfo()
|
D | code_generator_arm64.cc | 506 GetAssembler()->cfi().AdjustCFAOffset(frame_size); in GenerateFrameEntry() 507 GetAssembler()->SpillRegisters(GetFramePreservedCoreRegisters(), in GenerateFrameEntry() 509 GetAssembler()->SpillRegisters(GetFramePreservedFPRegisters(), in GenerateFrameEntry() 516 GetAssembler()->cfi().RememberState(); in GenerateFrameExit() 519 GetAssembler()->UnspillRegisters(GetFramePreservedFPRegisters(), in GenerateFrameExit() 521 GetAssembler()->UnspillRegisters(GetFramePreservedCoreRegisters(), in GenerateFrameExit() 524 GetAssembler()->cfi().AdjustCFAOffset(-frame_size); in GenerateFrameExit() 527 GetAssembler()->cfi().RestoreState(); in GenerateFrameExit() 528 GetAssembler()->cfi().DefCFAOffset(GetFrameSize()); in GenerateFrameExit() 1070 assembler_(codegen->GetAssembler()), in InstructionCodeGeneratorARM64()
|
D | code_generator.h | 158 virtual Assembler* GetAssembler() = 0;
|
D | code_generator_arm.cc | 56 #define __ reinterpret_cast<ArmAssembler*>(codegen->GetAssembler())-> 319 #define __ reinterpret_cast<ArmAssembler*>(GetAssembler())-> 493 assembler_(codegen->GetAssembler()), in InstructionCodeGeneratorARM() 972 if (GetAssembler()->ShifterOperandCanHold(R0, left, CMP, value, &operand)) { in GenerateTestAndBranch() 1051 if (GetAssembler()->ShifterOperandCanHold(R0, left, CMP, value, &operand)) { in VisitCondition() 3569 ArmAssembler* ParallelMoveResolverARM::GetAssembler() const { in GetAssembler() function in art::arm::ParallelMoveResolverARM 3570 return codegen_->GetAssembler(); in GetAssembler()
|
D | code_generator_mips64.cc | 106 #define __ down_cast<CodeGeneratorMIPS64*>(codegen)->GetAssembler()-> 417 #define __ down_cast<Mips64Assembler*>(GetAssembler())-> 424 Mips64Assembler* ParallelMoveResolverMIPS64::GetAssembler() const { in GetAssembler() function in art::mips64::ParallelMoveResolverMIPS64 425 return codegen_->GetAssembler(); in GetAssembler() 1006 assembler_(codegen->GetAssembler()), in InstructionCodeGeneratorMIPS64()
|
/art/compiler/utils/x86_64/ |
D | assembler_x86_64_test.cc | 593 GetAssembler()->LockCmpxchgl(x86_64::Address( in TEST_F() 596 GetAssembler()->LockCmpxchgl(x86_64::Address( in TEST_F() 599 GetAssembler()->LockCmpxchgl(x86_64::Address( in TEST_F() 602 GetAssembler()->LockCmpxchgl(x86_64::Address( in TEST_F() 604 GetAssembler()->LockCmpxchgl(x86_64::Address( in TEST_F() 618 GetAssembler()->LockCmpxchgq(x86_64::Address( in TEST_F() 621 GetAssembler()->LockCmpxchgq(x86_64::Address( in TEST_F() 624 GetAssembler()->LockCmpxchgq(x86_64::Address( in TEST_F() 627 GetAssembler()->LockCmpxchgq(x86_64::Address( in TEST_F() 629 GetAssembler()->LockCmpxchgq(x86_64::Address( in TEST_F() [all …]
|
/art/compiler/utils/ |
D | assembler_test.h | 48 Ass* GetAssembler() { in GetAssembler() function
|