/art/compiler/utils/arm/ |
D | jni_macro_assembler_arm_vixl.cc | 37 #define ___ asm_.GetVIXLAssembler()-> 201 UseScratchRegisterScope temps(asm_.GetVIXLAssembler()); in RemoveFrame() 241 UseScratchRegisterScope temps(asm_.GetVIXLAssembler()); in Store() 260 UseScratchRegisterScope temps(asm_.GetVIXLAssembler()); in StoreRef() 267 UseScratchRegisterScope temps(asm_.GetVIXLAssembler()); in StoreRawPtr() 279 UseScratchRegisterScope temps(asm_.GetVIXLAssembler()); in StoreSpanning() 289 UseScratchRegisterScope temps(asm_.GetVIXLAssembler()); in CopyRef() 301 UseScratchRegisterScope temps(asm_.GetVIXLAssembler()); in LoadRef() 325 UseScratchRegisterScope temps(asm_.GetVIXLAssembler()); in StoreImmediateToFrame() 343 UseScratchRegisterScope temps(asm_.GetVIXLAssembler()); in LoadRawPtrFromThread() [all …]
|
D | assembler_arm_vixl.cc | 388 UseScratchRegisterScope temps(GetVIXLAssembler()); in StoreRegisterList() 409 UseScratchRegisterScope temps(GetVIXLAssembler()); in LoadRegisterList()
|
D | assembler_arm_vixl.h | 172 ArmVIXLMacroAssembler* GetVIXLAssembler() { return &vixl_masm_; } in GetVIXLAssembler() function
|
/art/compiler/optimizing/ |
D | intrinsics_arm64.cc | 69 MacroAssembler* IntrinsicCodeGeneratorARM64::GetVIXLAssembler() { in GetVIXLAssembler() function in art::arm64::IntrinsicCodeGeneratorARM64 70 return codegen_->GetVIXLAssembler(); in GetVIXLAssembler() 77 #define __ codegen->GetVIXLAssembler()-> 127 vixl::EmissionCheckScope guard(codegen->GetVIXLAssembler(), kInvokeCodeMarginSizeInBytes); in EmitNativeCode() 276 MoveFPToInt(invoke->GetLocations(), /* is64bit= */ true, GetVIXLAssembler()); in VisitDoubleDoubleToRawLongBits() 279 MoveIntToFP(invoke->GetLocations(), /* is64bit= */ true, GetVIXLAssembler()); in VisitDoubleLongBitsToDouble() 290 MoveFPToInt(invoke->GetLocations(), /* is64bit= */ false, GetVIXLAssembler()); in VisitFloatFloatToRawIntBits() 293 MoveIntToFP(invoke->GetLocations(), /* is64bit= */ false, GetVIXLAssembler()); in VisitFloatIntBitsToFloat() 337 GenReverseBytes(invoke->GetLocations(), DataType::Type::kInt32, GetVIXLAssembler()); in VisitIntegerReverseBytes() 345 GenReverseBytes(invoke->GetLocations(), DataType::Type::kInt64, GetVIXLAssembler()); in VisitLongReverseBytes() [all …]
|
D | code_generator_arm_vixl.cc | 101 #define __ down_cast<CodeGeneratorARMVIXL*>(codegen)->GetVIXLAssembler()-> // NOLINT 200 UseScratchRegisterScope temps(down_cast<CodeGeneratorARMVIXL*>(codegen)->GetVIXLAssembler()); in SaveContiguousSRegisterList() 248 UseScratchRegisterScope temps(down_cast<CodeGeneratorARMVIXL*>(codegen)->GetVIXLAssembler()); in RestoreContiguousSRegisterList() 1172 UseScratchRegisterScope temps(codegen->GetVIXLAssembler()); in GenerateLongDataProc() 1326 UseScratchRegisterScope temps(codegen->GetVIXLAssembler()); in GenerateLongTestConstant() 1378 ExactAssemblyScope guard(codegen->GetVIXLAssembler(), in GenerateLongTestConstant() 1454 ExactAssemblyScope guard(codegen->GetVIXLAssembler(), in GenerateLongTest() 1480 UseScratchRegisterScope temps(codegen->GetVIXLAssembler()); in GenerateLongTest() 1533 ExactAssemblyScope guard(codegen->GetVIXLAssembler(), in GenerateConditionGeneric() 1563 UseScratchRegisterScope temps(codegen->GetVIXLAssembler()); in GenerateEqualLong() [all …]
|
D | intrinsics_arm64.h | 80 vixl::aarch64::MacroAssembler* GetVIXLAssembler();
|
D | code_generator_arm64.cc | 163 #define __ down_cast<CodeGeneratorARM64*>(codegen)->GetVIXLAssembler()-> // NOLINT 182 MacroAssembler* masm = down_cast<CodeGeneratorARM64*>(codegen)->GetVIXLAssembler(); in SaveRestoreLiveRegistersHelper() 586 EmissionCheckScope scope(codegen->GetVIXLAssembler(), in EmitTable() 914 #define __ GetVIXLAssembler()-> 1018 vixl_temps_.Open(GetVIXLAssembler()); in PrepareForEmitNativeCode() 1065 MacroAssembler* masm = GetVIXLAssembler(); in MaybeIncrementHotness() 1124 MacroAssembler* masm = GetVIXLAssembler(); in GenerateFrameEntry() 1136 ExactAssemblyScope eas(GetVIXLAssembler(), in GenerateFrameEntry() 1257 UseScratchRegisterScope temps(GetVIXLAssembler()); in MarkGCCard() 1472 UseScratchRegisterScope temps(GetVIXLAssembler()); in MoveLocation() [all …]
|
D | code_generator_arm64.h | 277 vixl::aarch64::MacroAssembler* GetVIXLAssembler() { return GetAssembler()->GetVIXLAssembler(); } in GetVIXLAssembler() function 404 vixl::aarch64::MacroAssembler* GetVIXLAssembler() const { in GetVIXLAssembler() function 405 return GetAssembler()->GetVIXLAssembler(); in GetVIXLAssembler() 458 vixl::aarch64::MacroAssembler* GetVIXLAssembler() { return GetAssembler()->GetVIXLAssembler(); } in GetVIXLAssembler() function 784 CHECK_EQ(GetVIXLAssembler()->ArePoolsBlocked(), true) in MaybeRecordImplicitNullCheck()
|
D | intrinsics_arm_vixl.cc | 37 #define __ assembler->GetVIXLAssembler()-> 491 ExactAssemblyScope it_scope(assembler->GetVIXLAssembler(), in VisitMathRoundFloat() 670 UseScratchRegisterScope temps(assembler->GetVIXLAssembler()); in GenUnsafeGet() 838 UseScratchRegisterScope temps(assembler->GetVIXLAssembler()); in GenUnsafePut() 1001 UseScratchRegisterScope temps(assembler->GetVIXLAssembler()); in EmitNativeCode() 1020 ExactAssemblyScope aas(assembler->GetVIXLAssembler(), 2 * k16BitT32InstructionSizeInBytes); in EmitNativeCode() 1095 static_cast<vixl32::MacroAssembler*>(assembler->GetVIXLAssembler())-> in GenCas() 1229 ExactAssemblyScope aas(assembler->GetVIXLAssembler(), in VisitStringCompareTo() 1251 ExactAssemblyScope aas(assembler->GetVIXLAssembler(), in VisitStringCompareTo() 1302 UseScratchRegisterScope temps(assembler->GetVIXLAssembler()); in GenerateStringCompareToLoop() [all …]
|
D | optimizing_cfi_test.cc | 194 ->GetAssembler())->GetVIXLAssembler()-> in TEST_ISA()
|
D | code_generator_arm_vixl.h | 319 ArmVIXLMacroAssembler* GetVIXLAssembler() { return GetAssembler()->GetVIXLAssembler(); } in GetVIXLAssembler() function 461 ArmVIXLMacroAssembler* GetVIXLAssembler() { return GetAssembler()->GetVIXLAssembler(); } in GetVIXLAssembler() function 754 CHECK_EQ(GetVIXLAssembler()->ArePoolsBlocked(), true) in MaybeRecordImplicitNullCheck()
|
D | code_generator_vector_arm_vixl.cc | 34 #define __ GetVIXLAssembler()-> 839 UseScratchRegisterScope temps(GetVIXLAssembler()); in VisitVecSADAccumulate() 958 UseScratchRegisterScope temps(GetVIXLAssembler()); in VisitVecLoad() 1009 UseScratchRegisterScope temps(GetVIXLAssembler()); in VisitVecStore()
|
D | codegen_test.cc | 824 vixl::CPUFeatures* features = codegen.GetVIXLAssembler()->GetCPUFeatures(); in TEST_F() 838 vixl::CPUFeatures* features = codegen.GetVIXLAssembler()->GetCPUFeatures(); in TEST_F()
|
D | code_generator_vector_arm64.cc | 39 #define __ GetVIXLAssembler()-> 1439 UseScratchRegisterScope temps(GetVIXLAssembler()); in VisitVecLoad() 1497 UseScratchRegisterScope temps(GetVIXLAssembler()); in VisitVecStore()
|
/art/compiler/trampolines/ |
D | trampoline_compiler.cc | 49 #define ___ assembler.GetVIXLAssembler()-> 64 vixl::aarch32::UseScratchRegisterScope temps(assembler.GetVIXLAssembler()); in CreateTrampoline()
|
/art/compiler/utils/arm64/ |
D | jni_macro_assembler_arm64.cc | 32 #define ___ asm_.GetVIXLAssembler()-> 94 UseScratchRegisterScope temps(asm_.GetVIXLAssembler()); in AddConstant() 185 UseScratchRegisterScope temps(asm_.GetVIXLAssembler()); in StoreStackPointerToThread() 210 UseScratchRegisterScope temps(asm_.GetVIXLAssembler()); in LoadImmediate() 327 UseScratchRegisterScope temps(asm_.GetVIXLAssembler()); in LoadRawPtr() 682 UseScratchRegisterScope temps(asm_.GetVIXLAssembler()); in EmitExceptionPoll() 810 UseScratchRegisterScope temps(asm_.GetVIXLAssembler()); in RemoveFrame()
|
D | assembler_arm64.h | 84 vixl::aarch64::MacroAssembler* GetVIXLAssembler() { return &vixl_masm_; } in GetVIXLAssembler() function
|
/art/compiler/utils/ |
D | assembler_thumb_test.cc | 322 vixl::aarch32::UseScratchRegisterScope temps(assembler.asm_.GetVIXLAssembler()); in TEST_F() 355 vixl::aarch32::UseScratchRegisterScope temps(assembler.asm_.GetVIXLAssembler()); in TEST_F()
|