Home
last modified time | relevance | path

Searched refs:kReference (Results 1 – 14 of 14) sorted by relevance

/art/compiler/dex/quick/
Dmir_to_lir.cc196 OpSize op_size = rl_dest.wide ? k64 : (rl_dest.ref ? kReference : k32); in LoadArgDirect()
213 OpSize size = arg.IsRef() ? kReference : in SpillArg()
225 OpSize size = arg.IsRef() ? kReference : in UnspillArg()
259 size = kReference; in GenSpecialIGet()
331 size = kReference; in GenSpecialIPut()
729 GenArrayGet(opt_flags, kReference, rl_src[0], rl_src[1], rl_dest, 2); in CompileDalvikInstruction()
763 GenArrayPut(opt_flags, kReference, rl_src[1], rl_src[2], rl_src[0], 2, !is_null); in CompileDalvikInstruction()
780 GenIGet(mir, opt_flags, kReference, Primitive::kPrimNot, rl_dest, rl_src[0]); in CompileDalvikInstruction()
829 GenIPut(mir, opt_flags, kReference, rl_src[0], rl_src[1]); in CompileDalvikInstruction()
855 GenSget(mir, rl_dest, kReference, Primitive::kPrimNot); in CompileDalvikInstruction()
[all …]
Dmir_to_lir.h569 if (size == kReference) { in RegClassBySize()
986 return LoadBaseDisp(r_base, displacement, r_dest, kReference, is_volatile); in LoadRefDisp()
990 return LoadBaseIndexed(r_base, r_index, r_dest, scale, kReference); in LoadRefIndexed()
1011 return StoreBaseDisp(r_base, displacement, r_src, kReference, is_volatile); in StoreRefDisp()
1015 return StoreBaseIndexed(r_base, r_index, r_src, scale, kReference); in StoreRefIndexed()
1182 return size == kReference; in IsRef()
Dgen_loadstore.cc58 op_size = kReference; in LoadValueDirect()
Dgen_common.cc734 case kReference: in GenSput()
979 case kReference: in GenIPut()
2261 case kReference: in SizeMatchesTypeForEntrypoint()
/art/compiler/dex/quick/mips/
Dutility_mips.cc579 DCHECK((size == k32) || (size == kSingle) || (size == kReference)); in LoadBaseIndexed()
618 case kReference: in LoadBaseIndexed()
651 DCHECK((size == k32) || (size == kSingle) || (size == kReference)); in StoreBaseIndexed()
671 case kReference: in StoreBaseIndexed()
729 case kReference: in LoadBaseDispBody()
906 case kReference: in StoreBaseDispBody()
Dint_mips.cc732 kReference); in GenArrayGet()
/art/compiler/dex/quick/arm64/
Dutility_arm64.cc1064 case kReference: in LoadBaseIndexed()
1149 case kReference: in StoreBaseIndexed()
1213 case kReference: in LoadBaseDispBody()
1259 (size == kReference) ? As64BitReg(r_dest) : r_dest, in LoadBaseDispBody()
1309 case kReference: in StoreBaseDispBody()
1349 (size == kReference) ? As64BitReg(r_src) : r_src, in StoreBaseDispBody()
Dtarget_arm64.cc600 return (size == kReference) ? kRefReg : kCoreReg; in RegClassForFieldLoadStore()
Dint_arm64.cc960 LIR* ldr = LoadBaseDisp(r_dest, 0, r_dest, wide ? k64 : kReference, kNotVolatile); in OpPcRelDexCacheArrayLoad()
/art/compiler/dex/quick/arm/
Dutility_arm.cc758 DCHECK((size == k32) || (size == kSingle) || (size == kReference)); in LoadBaseIndexed()
788 case kReference: in LoadBaseIndexed()
824 DCHECK((size == k32) || (size == kSingle) || (size == kReference)); in StoreBaseIndexed()
855 case kReference: in StoreBaseIndexed()
938 case kReference: in LoadBaseDispBody()
1092 case kReference: in StoreBaseDispBody()
/art/runtime/
Dtransaction.cc326 LogValue(ObjectLog::kReference, offset, reinterpret_cast<uintptr_t>(obj), is_volatile); in LogReferenceValue()
418 case kReference: in UndoFieldWrite()
436 if (field_value.kind == ObjectLog::kReference) { in VisitRoots()
Dtransaction.h133 kReference enumerator
/art/compiler/dex/
Dcompiler_enums.h384 kReference, // Object reference; compressed on 64-bit targets. enumerator
/art/compiler/dex/quick/x86/
Dutility_x86.cc674 case kReference: // TODO: update for reference decompression on 64-bit targets. in LoadBaseIndexedDisp()
828 case kReference: in StoreBaseIndexedDisp()