Searched refs:kArg1 (Results 1 – 11 of 11) sorted by relevance
/art/compiler/dex/quick/ |
D | gen_invoke.cc | 140 LoadConstant(TargetReg(kArg1, kNotWide), arg1); in CallRuntimeHelperImmImm() 149 LoadValueDirectFixed(arg1, TargetReg(kArg1, arg1)); in CallRuntimeHelperImmRegLocation() 151 RegStorage r_tmp = TargetReg(cu_->instruction_set == kMips ? kArg2 : kArg1, kWide); in CallRuntimeHelperImmRegLocation() 164 LoadConstant(TargetReg(kArg1, kNotWide), arg1); in CallRuntimeHelperRegLocationImm() 172 OpRegCopy(TargetReg(kArg1, arg1.GetWideKind()), arg1); in CallRuntimeHelperImmReg() 182 LoadConstant(TargetReg(kArg1, kNotWide), arg1); in CallRuntimeHelperRegImm() 190 LoadCurrMethodDirect(TargetReg(kArg1, kRef)); in CallRuntimeHelperImmMethod() 199 DCHECK(!IsSameReg(TargetReg(kArg1, arg0.GetWideKind()), arg0)); in CallRuntimeHelperRegMethod() 204 LoadCurrMethodDirect(TargetReg(kArg1, kRef)); in CallRuntimeHelperRegMethod() 217 LoadValueDirectFixed(arg1, TargetReg(kArg1, arg1)); in CallRuntimeHelperRegRegLocationMethod() [all …]
|
D | gen_common.cc | 252 RegStorage arg1_32 = m2l_->TargetReg(kArg1, kNotWide); in GenArrayBoundsCheck() 539 FreeTemp(TargetReg(kArg1, kNotWide)); in GenFilledNewArray() 1232 RegStorage r_method = LoadCurrMethodWithHint(TargetReg(kArg1, kRef)); in GenInstanceofCallingHelper() 1269 RegStorage ref_class_reg = TargetReg(kArg1, kRef); // kArg1 will hold the Class* of ref. in GenInstanceofCallingHelper() 1299 branchover = OpCmpBranch(kCondEq, TargetReg(kArg1, kRef), TargetReg(kArg2, kRef), nullptr); in GenInstanceofCallingHelper() 1368 RegStorage method_reg = LoadCurrMethodWithHint(TargetReg(kArg1, kRef)); in GenCheckCast() 1405 m2l_->TargetReg(kArg1, kRef), kNotVolatile); in GenCheckCast() 1408 m2l_->TargetReg(kArg1, kRef), true); in GenCheckCast() 1431 TargetReg(kArg1, kRef), kNotVolatile); in GenCheckCast() 1433 LIR* branch2 = OpCmpBranch(kCondNe, TargetReg(kArg1, kRef), class_reg, nullptr); in GenCheckCast() [all …]
|
D | mir_to_lir.h | 1207 static_assert((kArg1 == kArg0 + 1) && (kArg2 == kArg1 + 1) && (kArg3 == kArg2 + 1) && in TargetReg()
|
/art/compiler/dex/quick/mips/ |
D | target_mips.cc | 224 case kArg1: res_reg = rs_rA1; break; in TargetReg() 251 const SpecialTargetRegister coreArgMappingToPhysicalReg[] = {kArg1, kArg2, kArg3}; in GetNextReg() 268 {kArg1, kArg2, kArg3, kArg4, kArg5, kArg6, kArg7}; in GetNextReg() 649 LockTemp(TargetReg(kArg1)); in LockCallTemps() 663 FreeTemp(TargetReg(kArg1)); in FreeCallTemps() 827 OpRegCopyWide(TargetReg(kArg1, kWide), temp_value); in GenAtomic64Store()
|
D | call_mips.cc | 257 const RegStorage arg1 = TargetReg(kArg1); in GenEntrySequence()
|
/art/compiler/dex/ |
D | compiler_enums.h | 50 kArg1, enumerator
|
/art/compiler/dex/quick/x86/ |
D | call_x86.cc | 160 const RegStorage arg1 = TargetReg32(kArg1); in GenEntrySequence()
|
D | target_x86.cc | 544 LockTemp(TargetReg32(kArg1)); in LockCallTemps() 564 FreeTemp(TargetReg32(kArg1)); in FreeCallTemps() 2363 const SpecialTargetRegister coreArgMappingToPhysicalReg[] = {kArg1, kArg2, kArg3, kArg4, kArg5}; in GetNextReg() 2384 const SpecialTargetRegister coreArgMappingToPhysicalReg[] = {kArg1, kArg2, kArg3}; in GetNextReg()
|
D | int_x86.cc | 1516 if (index_ == m2l_->TargetReg(kArg1, kNotWide)) { in GenArrayBoundsCheck() 1527 x86_m2l->OpRegMem(kOpMov, m2l_->TargetReg(kArg1, kNotWide), array_base_, len_offset_); in GenArrayBoundsCheck() 1529 m2l_->TargetReg(kArg1, kNotWide), true); in GenArrayBoundsCheck() 1563 x86_m2l->OpRegMem(kOpMov, m2l_->TargetReg(kArg1, kNotWide), array_base_, len_offset_); in GenArrayBoundsCheck() 1566 m2l_->TargetReg(kArg1, kNotWide), true); in GenArrayBoundsCheck()
|
/art/compiler/dex/quick/arm64/ |
D | target_arm64.cc | 117 case kArg1: res_reg = rs_w1; break; in TargetReg()
|
/art/compiler/dex/quick/arm/ |
D | target_arm.cc | 106 case kArg1: res_reg = rs_r1; break; in TargetReg()
|