Home
last modified time | relevance | path

Searched refs:masm (Results 1 – 2 of 2) sorted by relevance

/art/compiler/optimizing/
Dintrinsics_arm64.cc167 #define __ masm->
185 static void MoveFPToInt(LocationSummary* locations, bool is64bit, vixl::MacroAssembler* masm) { in MoveFPToInt() argument
192 static void MoveIntToFP(LocationSummary* locations, bool is64bit, vixl::MacroAssembler* masm) { in MoveIntToFP() argument
237 vixl::MacroAssembler* masm) { in GenReverseBytes() argument
291 vixl::MacroAssembler* masm) { in GenNumberOfLeadingZeros() argument
318 vixl::MacroAssembler* masm) { in GenNumberOfTrailingZeros() argument
346 vixl::MacroAssembler* masm) { in GenReverse() argument
371 static void GenBitCount(HInvoke* instr, Primitive::Type type, vixl::MacroAssembler* masm) { in GenBitCount() argument
376 UseScratchRegisterScope temps(masm); in GenBitCount()
412 static void MathAbsFP(LocationSummary* locations, bool is64bit, vixl::MacroAssembler* masm) { in MathAbsFP() argument
[all …]
Dcode_generator_arm64.cc153 MacroAssembler* masm = down_cast<CodeGeneratorARM64*>(codegen)->GetVIXLAssembler(); in SaveRestoreLiveRegistersHelper() local
154 UseScratchRegisterScope temps(masm); in SaveRestoreLiveRegistersHelper()
156 Register base = masm->StackPointer(); in SaveRestoreLiveRegistersHelper()
163 !masm->IsImmLSPair(max_ls_pair_offset, ls_access_size)) { in SaveRestoreLiveRegistersHelper()
171 DCHECK(masm->IsImmLSPair(spill_offset, ls_access_size)); in SaveRestoreLiveRegistersHelper()
172 DCHECK(masm->IsImmLSPair(new_max_ls_pair_offset, ls_access_size)); in SaveRestoreLiveRegistersHelper()
998 MacroAssembler* masm = GetVIXLAssembler(); in GenerateFrameEntry() local
999 BlockPoolsScope block_pools(masm); in GenerateFrameEntry()
1004 UseScratchRegisterScope temps(masm); in GenerateFrameEntry()
1320 MacroAssembler* masm = GetVIXLAssembler(); in LoadAcquire() local
[all …]