/art/runtime/arch/arm/ |
D | asm_support_arm.S | 40 ldr \rDest, .Lgot_\name\()_\num @ Load offset of the GOT. 41 ldr \rTemp, .Lruntime_instance_\name\()_\num @ Load GOT offset of Runtime::instance_. 44 ldr \rDest, [\rDest, \rTemp] @ Load address of Runtime::instance_. 45 ldr \rDest, [\rDest] @ Load Runtime::instance_.
|
D | quick_entrypoints_arm.S | 46 RUNTIME_CURRENT1 \rTemp1, \rTemp2 @ Load Runtime::Current into rTemp1. 74 RUNTIME_CURRENT2 \rTemp1, \rTemp2 @ Load Runtime::Current into rTemp1. 134 RUNTIME_CURRENT3 \rTemp1, \rTemp2 @ Load Runtime::Current into rTemp1.
|
/art/test/302-float-conversion/ |
D | info.txt | 5 The second test checks the Load hoisting optimization for float pointing conversion.
|
/art/runtime/ |
D | elf_file.h | 49 bool Load(bool executable, std::string* error_msg);
|
D | memory_region.h | 55 ALWAYS_INLINE T Load(uintptr_t offset) const { in Load() function
|
D | elf_file_impl.h | 114 bool Load(bool executable, std::string* error_msg);
|
D | elf_file.cc | 1127 bool ElfFileImpl<ElfTypes>::Load(bool executable, std::string* error_msg) { in Load() function in art::ElfFileImpl 1847 bool ElfFile::Load(bool executable, std::string* error_msg) { in Load() function in art::ElfFile 1848 DELEGATE_TO_IMPL(Load, executable, error_msg); in Load()
|
D | oat_file.cc | 336 bool loaded = elf_file_->Load(executable, error_msg); in ElfFileOpen()
|
/art/compiler/utils/arm64/ |
D | assembler_arm64.h | 114 void Load(ManagedRegister dest, FrameOffset src, size_t size) OVERRIDE; 223 void Load(Arm64ManagedRegister dst, XRegister src, int32_t src_offset, size_t size);
|
D | assembler_arm64.cc | 259 void Arm64Assembler::Load(Arm64ManagedRegister dest, XRegister base, in Load() function in art::arm64::Arm64Assembler 282 void Arm64Assembler::Load(ManagedRegister m_dst, FrameOffset src, size_t size) { in Load() function in art::arm64::Arm64Assembler 283 return Load(m_dst.AsArm64(), SP, src.Int32Value(), size); in Load() 287 return Load(m_dst.AsArm64(), ETR, src.Int32Value(), size); in LoadFromThread64()
|
/art/test/475-regression-inliner-ids/smali/ |
D | TestCase.smali | 22 # The bug is triggered when inlining a method with few Load/StoreLocals but
|
/art/compiler/ |
D | elf_writer_test.cc | 84 CHECK(ef->Load(false, &error_msg)) << error_msg; in TEST_F()
|
/art/compiler/jni/quick/ |
D | jni_compiler.cc | 423 __ Load(out_reg, saved_cookie_offset, 4); in ArtJniCompileMethodInternal() local 461 __ Load(mr_conv->ReturnRegister(), return_save_location, mr_conv->SizeOfReturnValue()); in ArtJniCompileMethodInternal() local 559 __ Load(out_reg, in_off, param_size); in CopyParameter() local
|
/art/compiler/utils/ |
D | assembler.h | 192 template<typename T> T Load(size_t position) { in Load() function 432 virtual void Load(ManagedRegister dest, FrameOffset src, size_t size) = 0;
|
/art/compiler/utils/x86/ |
D | assembler_x86.cc | 1611 int next = buffer_.Load<int32_t>(position); in Bind() 1852 void X86Assembler::Load(ManagedRegister mdest, FrameOffset src, size_t size) { in Load() function in art::x86::X86Assembler 2006 Load(scratch, fr_offs, 4); in CopyRawPtrToThread32() 2015 Load(scratch, src, 4); in Copy() 2017 Load(scratch, FrameOffset(src.Int32Value() + 4), 4); in Copy() 2020 Load(scratch, src, size); in Copy()
|
/art/compiler/optimizing/ |
D | code_generator_arm64.h | 325 void Load(Primitive::Type type, vixl::CPURegister dst, const vixl::MemOperand& src);
|
D | intrinsics_arm64.cc | 654 codegen_->Load(Primitive::kPrimNot, WRegisterFrom(invoke->GetLocations()->Out()), in VisitThreadCurrentThread() 677 codegen->Load(type, trg, mem_op); in GenUnsafeGet() 681 codegen->Load(type, trg, mem_op); in GenUnsafeGet()
|
D | code_generator_arm64.cc | 814 void CodeGeneratorARM64::Load(Primitive::Type type, in Load() function in art::arm64::CodeGeneratorARM64 1149 codegen_->Load(field_info.GetFieldType(), OutputCPURegister(instruction), field); in HandleFieldGet() 1155 codegen_->Load(field_info.GetFieldType(), OutputCPURegister(instruction), field); in HandleFieldGet() 1350 codegen_->Load(type, OutputCPURegister(instruction), source); in VisitArrayGet()
|
/art/compiler/utils/x86_64/ |
D | assembler_x86_64.cc | 2100 int next = buffer_.Load<int32_t>(position); in Bind() 2535 void X86_64Assembler::Load(ManagedRegister mdest, FrameOffset src, size_t size) { in Load() function in art::x86_64::X86_64Assembler 2691 Load(scratch, fr_offs, 8); in CopyRawPtrToThread64() 2699 Load(scratch, src, 4); in Copy() 2701 Load(scratch, FrameOffset(src.Int32Value() + 4), 4); in Copy() 2704 Load(scratch, src, size); in Copy()
|
/art/compiler/utils/mips/ |
D | assembler_mips.h | 188 void Load(ManagedRegister mdest, FrameOffset src, size_t size) OVERRIDE;
|
D | assembler_mips.cc | 154 int32_t next = buffer_.Load<int32_t>(position); in Bind() 685 void MipsAssembler::Load(ManagedRegister mdest, FrameOffset src, size_t size) { in Load() function in art::mips::MipsAssembler
|
/art/runtime/arch/mips/ |
D | quick_entrypoints_mips.S | 1152 # Load parameters from stack into registers 1157 # Load FPRs the same as GPRs. Look at BuildNativeCallFrameStateMachine.
|
/art/runtime/arch/mips64/ |
D | quick_entrypoints_mips64.S | 1417 # Load parameters from stack into registers 1426 # Load FPRs the same as GPRs. Look at BuildNativeCallFrameStateMachine.
|
/art/compiler/utils/mips64/ |
D | assembler_mips64.cc | 801 uint32_t auipc = buffer_.Load<uint32_t>(position); in Bind() 802 uint32_t jic = buffer_.Load<uint32_t>(position + 4); // actually, jic or daddiu in Bind() 1230 void Mips64Assembler::Load(ManagedRegister mdest, FrameOffset src, size_t size) { in Load() function in art::mips64::Mips64Assembler
|
D | assembler_mips64.h | 261 void Load(ManagedRegister mdest, FrameOffset src, size_t size) OVERRIDE;
|