Searched refs:masm (Results 1 – 2 of 2) sorted by relevance
/art/compiler/optimizing/ |
D | intrinsics_arm64.cc | 167 #define __ masm-> 185 static void MoveFPToInt(LocationSummary* locations, bool is64bit, vixl::MacroAssembler* masm) { in MoveFPToInt() argument 192 static void MoveIntToFP(LocationSummary* locations, bool is64bit, vixl::MacroAssembler* masm) { in MoveIntToFP() argument 237 vixl::MacroAssembler* masm) { in GenReverseBytes() argument 291 vixl::MacroAssembler* masm) { in GenNumberOfLeadingZeros() argument 318 vixl::MacroAssembler* masm) { in GenNumberOfTrailingZeros() argument 346 vixl::MacroAssembler* masm) { in GenReverse() argument 371 static void GenBitCount(HInvoke* instr, Primitive::Type type, vixl::MacroAssembler* masm) { in GenBitCount() argument 376 UseScratchRegisterScope temps(masm); in GenBitCount() 412 static void MathAbsFP(LocationSummary* locations, bool is64bit, vixl::MacroAssembler* masm) { in MathAbsFP() argument [all …]
|
D | code_generator_arm64.cc | 153 MacroAssembler* masm = down_cast<CodeGeneratorARM64*>(codegen)->GetVIXLAssembler(); in SaveRestoreLiveRegistersHelper() local 154 UseScratchRegisterScope temps(masm); in SaveRestoreLiveRegistersHelper() 156 Register base = masm->StackPointer(); in SaveRestoreLiveRegistersHelper() 163 !masm->IsImmLSPair(max_ls_pair_offset, ls_access_size)) { in SaveRestoreLiveRegistersHelper() 171 DCHECK(masm->IsImmLSPair(spill_offset, ls_access_size)); in SaveRestoreLiveRegistersHelper() 172 DCHECK(masm->IsImmLSPair(new_max_ls_pair_offset, ls_access_size)); in SaveRestoreLiveRegistersHelper() 998 MacroAssembler* masm = GetVIXLAssembler(); in GenerateFrameEntry() local 999 BlockPoolsScope block_pools(masm); in GenerateFrameEntry() 1004 UseScratchRegisterScope temps(masm); in GenerateFrameEntry() 1320 MacroAssembler* masm = GetVIXLAssembler(); in LoadAcquire() local [all …]
|