Home
last modified time | relevance | path

Searched refs:GetVIXLAssembler (Results 1 – 18 of 18) sorted by relevance

/art/compiler/utils/arm/
Djni_macro_assembler_arm_vixl.cc37 #define ___ asm_.GetVIXLAssembler()->
201 UseScratchRegisterScope temps(asm_.GetVIXLAssembler()); in RemoveFrame()
241 UseScratchRegisterScope temps(asm_.GetVIXLAssembler()); in Store()
260 UseScratchRegisterScope temps(asm_.GetVIXLAssembler()); in StoreRef()
267 UseScratchRegisterScope temps(asm_.GetVIXLAssembler()); in StoreRawPtr()
279 UseScratchRegisterScope temps(asm_.GetVIXLAssembler()); in StoreSpanning()
289 UseScratchRegisterScope temps(asm_.GetVIXLAssembler()); in CopyRef()
301 UseScratchRegisterScope temps(asm_.GetVIXLAssembler()); in LoadRef()
325 UseScratchRegisterScope temps(asm_.GetVIXLAssembler()); in StoreImmediateToFrame()
343 UseScratchRegisterScope temps(asm_.GetVIXLAssembler()); in LoadRawPtrFromThread()
[all …]
Dassembler_arm_vixl.cc388 UseScratchRegisterScope temps(GetVIXLAssembler()); in StoreRegisterList()
409 UseScratchRegisterScope temps(GetVIXLAssembler()); in LoadRegisterList()
Dassembler_arm_vixl.h172 ArmVIXLMacroAssembler* GetVIXLAssembler() { return &vixl_masm_; } in GetVIXLAssembler() function
/art/compiler/optimizing/
Dintrinsics_arm64.cc69 MacroAssembler* IntrinsicCodeGeneratorARM64::GetVIXLAssembler() { in GetVIXLAssembler() function in art::arm64::IntrinsicCodeGeneratorARM64
70 return codegen_->GetVIXLAssembler(); in GetVIXLAssembler()
77 #define __ codegen->GetVIXLAssembler()->
127 vixl::EmissionCheckScope guard(codegen->GetVIXLAssembler(), kInvokeCodeMarginSizeInBytes); in EmitNativeCode()
276 MoveFPToInt(invoke->GetLocations(), /* is64bit= */ true, GetVIXLAssembler()); in VisitDoubleDoubleToRawLongBits()
279 MoveIntToFP(invoke->GetLocations(), /* is64bit= */ true, GetVIXLAssembler()); in VisitDoubleLongBitsToDouble()
290 MoveFPToInt(invoke->GetLocations(), /* is64bit= */ false, GetVIXLAssembler()); in VisitFloatFloatToRawIntBits()
293 MoveIntToFP(invoke->GetLocations(), /* is64bit= */ false, GetVIXLAssembler()); in VisitFloatIntBitsToFloat()
337 GenReverseBytes(invoke->GetLocations(), DataType::Type::kInt32, GetVIXLAssembler()); in VisitIntegerReverseBytes()
345 GenReverseBytes(invoke->GetLocations(), DataType::Type::kInt64, GetVIXLAssembler()); in VisitLongReverseBytes()
[all …]
Dcode_generator_arm_vixl.cc101 #define __ down_cast<CodeGeneratorARMVIXL*>(codegen)->GetVIXLAssembler()-> // NOLINT
200 UseScratchRegisterScope temps(down_cast<CodeGeneratorARMVIXL*>(codegen)->GetVIXLAssembler()); in SaveContiguousSRegisterList()
248 UseScratchRegisterScope temps(down_cast<CodeGeneratorARMVIXL*>(codegen)->GetVIXLAssembler()); in RestoreContiguousSRegisterList()
1172 UseScratchRegisterScope temps(codegen->GetVIXLAssembler()); in GenerateLongDataProc()
1326 UseScratchRegisterScope temps(codegen->GetVIXLAssembler()); in GenerateLongTestConstant()
1378 ExactAssemblyScope guard(codegen->GetVIXLAssembler(), in GenerateLongTestConstant()
1454 ExactAssemblyScope guard(codegen->GetVIXLAssembler(), in GenerateLongTest()
1480 UseScratchRegisterScope temps(codegen->GetVIXLAssembler()); in GenerateLongTest()
1533 ExactAssemblyScope guard(codegen->GetVIXLAssembler(), in GenerateConditionGeneric()
1563 UseScratchRegisterScope temps(codegen->GetVIXLAssembler()); in GenerateEqualLong()
[all …]
Dintrinsics_arm64.h80 vixl::aarch64::MacroAssembler* GetVIXLAssembler();
Dcode_generator_arm64.cc163 #define __ down_cast<CodeGeneratorARM64*>(codegen)->GetVIXLAssembler()-> // NOLINT
182 MacroAssembler* masm = down_cast<CodeGeneratorARM64*>(codegen)->GetVIXLAssembler(); in SaveRestoreLiveRegistersHelper()
586 EmissionCheckScope scope(codegen->GetVIXLAssembler(), in EmitTable()
914 #define __ GetVIXLAssembler()->
1018 vixl_temps_.Open(GetVIXLAssembler()); in PrepareForEmitNativeCode()
1065 MacroAssembler* masm = GetVIXLAssembler(); in MaybeIncrementHotness()
1124 MacroAssembler* masm = GetVIXLAssembler(); in GenerateFrameEntry()
1136 ExactAssemblyScope eas(GetVIXLAssembler(), in GenerateFrameEntry()
1257 UseScratchRegisterScope temps(GetVIXLAssembler()); in MarkGCCard()
1472 UseScratchRegisterScope temps(GetVIXLAssembler()); in MoveLocation()
[all …]
Dcode_generator_arm64.h277 vixl::aarch64::MacroAssembler* GetVIXLAssembler() { return GetAssembler()->GetVIXLAssembler(); } in GetVIXLAssembler() function
404 vixl::aarch64::MacroAssembler* GetVIXLAssembler() const { in GetVIXLAssembler() function
405 return GetAssembler()->GetVIXLAssembler(); in GetVIXLAssembler()
458 vixl::aarch64::MacroAssembler* GetVIXLAssembler() { return GetAssembler()->GetVIXLAssembler(); } in GetVIXLAssembler() function
784 CHECK_EQ(GetVIXLAssembler()->ArePoolsBlocked(), true) in MaybeRecordImplicitNullCheck()
Dintrinsics_arm_vixl.cc37 #define __ assembler->GetVIXLAssembler()->
491 ExactAssemblyScope it_scope(assembler->GetVIXLAssembler(), in VisitMathRoundFloat()
670 UseScratchRegisterScope temps(assembler->GetVIXLAssembler()); in GenUnsafeGet()
838 UseScratchRegisterScope temps(assembler->GetVIXLAssembler()); in GenUnsafePut()
1001 UseScratchRegisterScope temps(assembler->GetVIXLAssembler()); in EmitNativeCode()
1020 ExactAssemblyScope aas(assembler->GetVIXLAssembler(), 2 * k16BitT32InstructionSizeInBytes); in EmitNativeCode()
1095 static_cast<vixl32::MacroAssembler*>(assembler->GetVIXLAssembler())-> in GenCas()
1229 ExactAssemblyScope aas(assembler->GetVIXLAssembler(), in VisitStringCompareTo()
1251 ExactAssemblyScope aas(assembler->GetVIXLAssembler(), in VisitStringCompareTo()
1302 UseScratchRegisterScope temps(assembler->GetVIXLAssembler()); in GenerateStringCompareToLoop()
[all …]
Doptimizing_cfi_test.cc194 ->GetAssembler())->GetVIXLAssembler()-> in TEST_ISA()
Dcode_generator_arm_vixl.h319 ArmVIXLMacroAssembler* GetVIXLAssembler() { return GetAssembler()->GetVIXLAssembler(); } in GetVIXLAssembler() function
461 ArmVIXLMacroAssembler* GetVIXLAssembler() { return GetAssembler()->GetVIXLAssembler(); } in GetVIXLAssembler() function
754 CHECK_EQ(GetVIXLAssembler()->ArePoolsBlocked(), true) in MaybeRecordImplicitNullCheck()
Dcode_generator_vector_arm_vixl.cc34 #define __ GetVIXLAssembler()->
839 UseScratchRegisterScope temps(GetVIXLAssembler()); in VisitVecSADAccumulate()
958 UseScratchRegisterScope temps(GetVIXLAssembler()); in VisitVecLoad()
1009 UseScratchRegisterScope temps(GetVIXLAssembler()); in VisitVecStore()
Dcodegen_test.cc824 vixl::CPUFeatures* features = codegen.GetVIXLAssembler()->GetCPUFeatures(); in TEST_F()
838 vixl::CPUFeatures* features = codegen.GetVIXLAssembler()->GetCPUFeatures(); in TEST_F()
Dcode_generator_vector_arm64.cc39 #define __ GetVIXLAssembler()->
1439 UseScratchRegisterScope temps(GetVIXLAssembler()); in VisitVecLoad()
1497 UseScratchRegisterScope temps(GetVIXLAssembler()); in VisitVecStore()
/art/compiler/trampolines/
Dtrampoline_compiler.cc49 #define ___ assembler.GetVIXLAssembler()->
64 vixl::aarch32::UseScratchRegisterScope temps(assembler.GetVIXLAssembler()); in CreateTrampoline()
/art/compiler/utils/arm64/
Djni_macro_assembler_arm64.cc32 #define ___ asm_.GetVIXLAssembler()->
94 UseScratchRegisterScope temps(asm_.GetVIXLAssembler()); in AddConstant()
185 UseScratchRegisterScope temps(asm_.GetVIXLAssembler()); in StoreStackPointerToThread()
210 UseScratchRegisterScope temps(asm_.GetVIXLAssembler()); in LoadImmediate()
327 UseScratchRegisterScope temps(asm_.GetVIXLAssembler()); in LoadRawPtr()
682 UseScratchRegisterScope temps(asm_.GetVIXLAssembler()); in EmitExceptionPoll()
810 UseScratchRegisterScope temps(asm_.GetVIXLAssembler()); in RemoveFrame()
Dassembler_arm64.h84 vixl::aarch64::MacroAssembler* GetVIXLAssembler() { return &vixl_masm_; } in GetVIXLAssembler() function
/art/compiler/utils/
Dassembler_thumb_test.cc322 vixl::aarch32::UseScratchRegisterScope temps(assembler.asm_.GetVIXLAssembler()); in TEST_F()
355 vixl::aarch32::UseScratchRegisterScope temps(assembler.asm_.GetVIXLAssembler()); in TEST_F()