/art/compiler/trampolines/ |
D | trampoline_compiler.cc | 64 ArmVIXLAssembler assembler(allocator); in CreateTrampoline() local 102 Arm64Assembler assembler(allocator); in CreateTrampoline() local 141 MipsAssembler assembler(allocator); in CreateTrampoline() local 173 Mips64Assembler assembler(allocator); in CreateTrampoline() local 205 X86Assembler assembler(allocator); in CreateTrampoline() local 226 x86_64::X86_64Assembler assembler(allocator); in CreateTrampoline() local
|
/art/compiler/optimizing/ |
D | intrinsics_mips64.cc | 151 static void MoveFPToInt(LocationSummary* locations, bool is64bit, Mips64Assembler* assembler) { in MoveFPToInt() 187 static void MoveIntToFP(LocationSummary* locations, bool is64bit, Mips64Assembler* assembler) { in MoveIntToFP() 225 Mips64Assembler* assembler) { in GenReverseBytes() 277 Mips64Assembler* assembler) { in GenNumberOfLeadingZeroes() 308 Mips64Assembler* assembler) { in GenNumberOfTrailingZeroes() 345 Mips64Assembler* assembler) { in GenReverse() 389 Mips64Assembler* assembler) { in GenBitCount() 473 static void MathAbsFP(LocationSummary* locations, bool is64bit, Mips64Assembler* assembler) { in MathAbsFP() 509 static void GenAbsInteger(LocationSummary* locations, bool is64bit, Mips64Assembler* assembler) { in GenAbsInteger() 545 Mips64Assembler* assembler) { in GenMinMaxFP() [all …]
|
D | intrinsics_arm_vixl.cc | 93 ArmVIXLAssembler* assembler = down_cast<ArmVIXLAssembler*>(codegen->GetAssembler()); in EmitNativeCode() local 128 static void GenSystemArrayCopyBaseAddress(ArmVIXLAssembler* assembler, in GenSystemArrayCopyBaseAddress() 151 static void GenSystemArrayCopyEndAddress(ArmVIXLAssembler* assembler, in GenSystemArrayCopyEndAddress() 182 ArmVIXLAssembler* assembler = arm_codegen->GetAssembler(); in EmitNativeCode() local 277 static void MoveFPToInt(LocationSummary* locations, bool is64bit, ArmVIXLAssembler* assembler) { in MoveFPToInt() 287 static void MoveIntToFP(LocationSummary* locations, bool is64bit, ArmVIXLAssembler* assembler) { in MoveIntToFP() 349 ArmVIXLAssembler* assembler = codegen->GetAssembler(); in GenNumberOfLeadingZeros() local 394 ArmVIXLAssembler* assembler = codegen->GetAssembler(); in GenNumberOfTrailingZeros() local 435 static void MathAbsFP(HInvoke* invoke, ArmVIXLAssembler* assembler) { in MathAbsFP() 466 ArmVIXLAssembler* assembler) { in GenAbsInteger() [all …]
|
D | intrinsics_utils.h | 51 Assembler* assembler = codegen->GetAssembler(); in EmitNativeCode() local
|
D | intrinsics_mips.cc | 162 static void MoveFPToInt(LocationSummary* locations, bool is64bit, MipsAssembler* assembler) { in MoveFPToInt() 203 static void MoveIntToFP(LocationSummary* locations, bool is64bit, MipsAssembler* assembler) { in MoveIntToFP() 251 MipsAssembler* assembler) { in GenReverse() 445 MipsAssembler* assembler) { in GenNumberOfLeadingZeroes() 493 MipsAssembler* assembler) { in GenNumberOfTrailingZeroes() 615 MipsAssembler* assembler) { in GenBitCount() 749 MipsAssembler* assembler) { in MathAbsFP() 815 static void GenAbsInteger(LocationSummary* locations, bool is64bit, MipsAssembler* assembler) { in GenAbsInteger() 865 MipsAssembler* assembler) { in GenMinMaxFP() 1096 MipsAssembler* assembler) { in GenMinMax() [all …]
|
D | intrinsics_x86.cc | 199 static void MoveFPToInt(LocationSummary* locations, bool is64bit, X86Assembler* assembler) { in MoveFPToInt() 214 static void MoveIntToFP(LocationSummary* locations, bool is64bit, X86Assembler* assembler) { in MoveIntToFP() 281 X86Assembler* assembler) { in GenReverseBytes() 320 X86Assembler* assembler = GetAssembler(); in VisitLongReverseBytes() local 359 X86Assembler* assembler, in MathAbsFP() 421 static void GenAbsInteger(LocationSummary* locations, X86Assembler* assembler) { in GenAbsInteger() 448 static void GenAbsLong(LocationSummary* locations, X86Assembler* assembler) { in GenAbsLong() 493 X86Assembler* assembler, in GenMinMaxFP() 658 X86Assembler* assembler) { in GenMinMax() 817 X86Assembler* assembler, in GenSSE41FPToFPIntrinsic() [all …]
|
D | intrinsics_x86_64.cc | 145 static void MoveFPToInt(LocationSummary* locations, bool is64bit, X86_64Assembler* assembler) { in MoveFPToInt() 151 static void MoveIntToFP(LocationSummary* locations, bool is64bit, X86_64Assembler* assembler) { in MoveIntToFP() 194 X86_64Assembler* assembler) { in GenReverseBytes() 254 X86_64Assembler* assembler, in MathAbsFP() 297 static void GenAbsInteger(LocationSummary* locations, bool is64bit, X86_64Assembler* assembler) { in GenAbsInteger() 338 X86_64Assembler* assembler, in GenMinMaxFP() 466 X86_64Assembler* assembler) { in GenMinMax() 592 X86_64Assembler* assembler, in GenSSE41FPToFPIntrinsic() 668 X86_64Assembler* assembler = GetAssembler(); in VisitMathRoundFloat() local 712 X86_64Assembler* assembler = GetAssembler(); in VisitMathRoundDouble() local [all …]
|
D | code_generator_x86_64.cc | 7032 X86_64Assembler* assembler = codegen_->GetAssembler(); in CreateJumpTable() local 7061 X86_64Assembler* assembler = GetAssembler(); in Finalize() local
|
D | code_generator_x86.cc | 7651 X86Assembler* assembler = codegen_->GetAssembler(); in CreateJumpTable() local 7681 X86Assembler* assembler = GetAssembler(); in Finalize() local
|
D | code_generator_arm_vixl.cc | 140 EmitAdrCode(ArmVIXLMacroAssembler* assembler, vixl32::Register rd, vixl32::Label* label) in EmitAdrCode() 5648 ArmVIXLAssembler* assembler = codegen->GetAssembler(); in CanEncode32BitConstantAsImmediate() local
|
/art/compiler/linker/arm/ |
D | relative_patcher_thumb2.cc | 183 static void EmitGrayCheckAndFastPath(arm::ArmVIXLAssembler& assembler, in EmitGrayCheckAndFastPath() 206 static void LoadReadBarrierMarkIntrospectionEntrypoint(arm::ArmVIXLAssembler& assembler, in LoadReadBarrierMarkIntrospectionEntrypoint() 223 void Thumb2RelativePatcher::CompileBakerReadBarrierThunk(arm::ArmVIXLAssembler& assembler, in CompileBakerReadBarrierThunk() 360 arm::ArmVIXLAssembler assembler(&allocator); in CompileThunk() local
|
/art/compiler/linker/arm64/ |
D | relative_patcher_arm64.cc | 366 static void EmitGrayCheckAndFastPath(arm64::Arm64Assembler& assembler, in EmitGrayCheckAndFastPath() 394 static void LoadReadBarrierMarkIntrospectionEntrypoint(arm64::Arm64Assembler& assembler, in LoadReadBarrierMarkIntrospectionEntrypoint() 408 void Arm64RelativePatcher::CompileBakerReadBarrierThunk(arm64::Arm64Assembler& assembler, in CompileBakerReadBarrierThunk() 518 arm64::Arm64Assembler assembler(&allocator); in CompileThunk() local
|
/art/compiler/utils/ |
D | assembler_thumb_test.cc | 174 ArmVIXLJNIMacroAssembler assembler; member in art::arm::ArmVIXLAssemblerTest 179 void EmitAndCheck(ArmVIXLJNIMacroAssembler* assembler, const char* testname, in EmitAndCheck() 190 void EmitAndCheck(ArmVIXLJNIMacroAssembler* assembler, const char* testname) { in EmitAndCheck()
|
/art/compiler/utils/x86_64/ |
D | assembler_x86_64_test.cc | 633 std::string shll_fn(AssemblerX86_64Test::Base* assembler_test, x86_64::X86_64Assembler* assembler) { in shll_fn() 654 std::string shlq_fn(AssemblerX86_64Test::Base* assembler_test, x86_64::X86_64Assembler* assembler) { in shlq_fn() 675 std::string shrl_fn(AssemblerX86_64Test::Base* assembler_test, x86_64::X86_64Assembler* assembler) { in shrl_fn() 695 std::string shrq_fn(AssemblerX86_64Test::Base* assembler_test, x86_64::X86_64Assembler* assembler) { in shrq_fn() 715 std::string sarl_fn(AssemblerX86_64Test::Base* assembler_test, x86_64::X86_64Assembler* assembler) { in sarl_fn() 735 std::string sarq_fn(AssemblerX86_64Test::Base* assembler_test, x86_64::X86_64Assembler* assembler) { in sarq_fn() 755 std::string rorl_fn(AssemblerX86_64Test::Base* assembler_test, x86_64::X86_64Assembler* assembler) { in rorl_fn() 775 std::string roll_fn(AssemblerX86_64Test::Base* assembler_test, x86_64::X86_64Assembler* assembler) { in roll_fn() 795 std::string rorq_fn(AssemblerX86_64Test::Base* assembler_test, x86_64::X86_64Assembler* assembler) { in rorq_fn() 815 std::string rolq_fn(AssemblerX86_64Test::Base* assembler_test, x86_64::X86_64Assembler* assembler) { in rolq_fn() [all …]
|
D | jni_macro_assembler_x86_64.cc | 145 static void DecreaseFrameSizeImpl(size_t adjust, X86_64Assembler* assembler) { in DecreaseFrameSizeImpl()
|
/art/compiler/utils/x86/ |
D | assembler_x86_test.cc | 376 std::string rorl_fn(AssemblerX86Test::Base* assembler_test, x86::X86Assembler* assembler) { in rorl_fn() 396 std::string roll_fn(AssemblerX86Test::Base* assembler_test, x86::X86Assembler* assembler) { in roll_fn()
|
D | jni_macro_assembler_x86.cc | 114 static void DecreaseFrameSizeImpl(X86Assembler* assembler, size_t adjust) { in DecreaseFrameSizeImpl()
|