/art/compiler/dex/quick/ |
D | mir_to_lir.cc | 147 LoadBaseDisp(TargetPtrReg(kSp), offset, reg_arg, wide ? k64 : k32, kNotVolatile); in LoadArg() 196 OpSize op_size = rl_dest.wide ? k64 : (rl_dest.ref ? kReference : k32); in LoadArgDirect() 214 (arg.IsWide() && reg_arg.GetWideKind() == kWide) ? k64 : k32; in SpillArg() 226 (arg.IsWide() && reg_arg.GetWideKind() == kWide) ? k64 : k32; in UnspillArg() 256 size = in_to_reg_storage_mapping_.GetShorty(data.src_arg).IsFP() ? kDouble : k64; in GenSpecialIGet() 328 size = in_to_reg_storage_mapping_.GetShorty(data.src_arg).IsFP() ? kDouble : k64; in GenSpecialIPut() 726 GenArrayGet(opt_flags, rl_dest.fp ? kDouble : k64, rl_src[0], rl_src[1], rl_dest, 3); in CompileDalvikInstruction() 747 … GenArrayPut(opt_flags, rl_src[0].fp ? kDouble : k64, rl_src[1], rl_src[2], rl_src[0], 3, false); in CompileDalvikInstruction() 789 GenIGet(mir, opt_flags, k64, Primitive::kPrimLong, rl_dest, rl_src[0]); in CompileDalvikInstruction() 824 GenIPut(mir, opt_flags, rl_src[0].fp ? kDouble : k64, rl_src[0], rl_src[1]); in CompileDalvikInstruction() [all …]
|
D | gen_loadstore.cc | 60 op_size = k64; in LoadValueDirect() 94 LoadBaseDisp(TargetPtrReg(kSp), SRegOffset(rl_src.s_reg_low), r_dest, k64, kNotVolatile); in LoadValueDirectWide() 271 StoreBaseDisp(TargetPtrReg(kSp), SRegOffset(rl_dest.s_reg_low), rl_dest.reg, k64, kNotVolatile); in StoreValueWide() 335 StoreBaseDisp(TargetPtrReg(kSp), SRegOffset(rl_dest.s_reg_low), rl_dest.reg, k64, kNotVolatile); in StoreFinalValueWide()
|
D | gen_invoke.cc | 471 StoreBaseDisp(TargetPtrReg(kSp), offset, reg, t_loc->wide ? k64 : k32, kNotVolatile); in FlushIns() 481 LoadBaseDisp(TargetPtrReg(kSp), offset, t_loc->reg, t_loc->wide ? k64 : k32, in FlushIns() 706 StoreBaseDisp(TargetPtrReg(kSp), SRegOffset(loc.s_reg_low), loc.reg, k64, kNotVolatile); in GenDalvikArgsFlushPromoted() 805 StoreBaseDisp(TargetPtrReg(kSp), out_offset, rl_arg.reg, k64, kNotVolatile); in GenDalvikArgs() 808 StoreBaseDisp(TargetPtrReg(kSp), out_offset, regWide, k64, kNotVolatile); in GenDalvikArgs() 850 StoreBaseDisp(TargetPtrReg(kSp), out_offset, rl_arg.reg, k64, kNotVolatile); in GenDalvikArgs() 1459 LoadBaseIndexed(rl_object.reg, rl_offset.reg, rl_result.reg, 0, k64); in GenInlinedUnsafeGet() 1463 LoadBaseDisp(rl_temp_offset, 0, rl_result.reg, k64, kNotVolatile); in GenInlinedUnsafeGet() 1507 StoreBaseIndexed(rl_object.reg, rl_offset.reg, rl_value.reg, 0, k64); in GenInlinedUnsafePut() 1511 StoreBaseDisp(rl_temp_offset, 0, rl_value.reg, k64, kNotVolatile); in GenInlinedUnsafePut()
|
D | dex_file_method_inliner.cc | 366 INTRINSIC(JavaLangLong, ReverseBytes, J_J, kIntrinsicReverseBytes, k64), 369 INTRINSIC(JavaLangLong, Reverse, J_J, kIntrinsicReverseBits, k64), 424 INTRINSIC(LibcoreIoMemory, PeekLongNative, J_J, kIntrinsicPeek, k64), 428 INTRINSIC(LibcoreIoMemory, PokeLongNative, JJ_V, kIntrinsicPoke, k64),
|
D | gen_common.cc | 737 case k64: in GenSput() 982 case k64: in GenIPut() 2263 case k64: in SizeMatchesTypeForEntrypoint()
|
D | ralloc_util.cc | 746 StoreBaseDisp(TargetPtrReg(kSp), VRegOffset(v_reg), reg, k64, kNotVolatile); in FlushRegWide() 754 StoreBaseDisp(TargetPtrReg(kSp), VRegOffset(v_reg), reg, k64, kNotVolatile); in FlushRegWide()
|
D | mir_to_lir.h | 1178 return size == k64 || size == kDouble; in IsWide()
|
/art/compiler/dex/quick/x86/ |
D | target_x86.cc | 743 OpSize size = cu_->target64 ? k64 : k32; in SpillCoreRegs() 763 OpSize size = cu_->target64 ? k64 : k32; in UnSpillCoreRegs() 785 StoreBaseDisp(rs_rSP, offset, RegStorage::FloatSolo64(reg), k64, kNotVolatile); in SpillFPRegs() 802 k64, kNotVolatile); in UnSpillFPRegs() 829 if (size == k64 || size == kDouble) { in RegClassForFieldLoadStore() 1729 case k64: in GenMultiplyVector() 1750 case k64: in GenAddVector() 1785 case k64: in GenSubtractVector() 1864 case k64: in GenShiftLeftVector() 1901 case k64: in GenSignedShiftRightVector() [all …]
|
D | fp_x86.cc | 165 StoreBaseDisp(rs_rSP, src_v_reg_offset, rl_src.reg, k64, kNotVolatile); in GenLongToFP() 199 LoadBaseDisp(rs_rSP, dest_v_reg_offset, rl_result.reg, k64, kNotVolatile); in GenLongToFP() 387 StoreBaseDisp(rs_rSP, src1_v_reg_offset, rl_src1.reg, is_double ? k64 : k32, in GenRemFP() 398 StoreBaseDisp(rs_rSP, src2_v_reg_offset, rl_src2.reg, is_double ? k64 : k32, in GenRemFP() 458 LoadBaseDisp(rs_rSP, dest_v_reg_offset, rl_result.reg, k64, kNotVolatile); in GenRemFP()
|
D | utility_x86.cc | 649 bool is64bit = ((size == k64) || (size == kDouble)); in LoadBaseIndexedDisp() 652 case k64: in LoadBaseIndexedDisp() 798 bool is64bit = (size == k64) || (size == kDouble); in StoreBaseIndexedDisp() 803 case k64: in StoreBaseIndexedDisp()
|
D | int_x86.cc | 1032 RegLocation rl_dest = size == k64 ? InlineTargetWide(info) : InlineTarget(info); in GenInlinedPeek() 1048 if (size == k64) { in GenInlinedPeek() 1068 if (size == k64) { in GenInlinedPoke() 1304 RegLocation rl_dest = (size == k64) ? InlineTargetWide(info) : InlineTarget(info); in GenInlinedReverseBits() 1310 RegLocation rl_i = (size == k64) ? LoadValueWide(rl_src_i, kCoreReg) in GenInlinedReverseBits() 1313 if (size == k64) { in GenInlinedReverseBits() 2453 if (size == k64 || size == kDouble) { in GenArrayGet() 2483 if ((size == k64) || (size == kDouble)) { in GenArrayGet() 2500 if (size == k64 || size == kDouble) { in GenArrayPut() 2529 if ((size == k64) || (size == kDouble)) { in GenArrayPut()
|
/art/compiler/dex/quick/mips/ |
D | utility_mips.cc | 607 case k64: in LoadBaseIndexed() 708 case k64: in LoadBaseDispBody() 853 if (UNLIKELY(is_volatile == kVolatile && (size == k64 || size == kDouble)) in LoadBaseDisp() 862 size = cu_->target64 ? k64 : k32; in LoadBaseDisp() 885 case k64: in StoreBaseDispBody() 1012 if (UNLIKELY(is_volatile == kVolatile && (size == k64 || size == kDouble) && in StoreBaseDisp() 1020 size = cu_->target64 ? k64 : k32; in StoreBaseDisp()
|
D | int_mips.cc | 678 if (size == k64 || size == kDouble) { in GenArrayGet() 701 if ((size == k64) || (size == kDouble)) { in GenArrayGet() 755 if (size == k64 || size == kDouble) { in GenArrayPut() 794 if ((size == k64) || (size == kDouble)) { in GenArrayPut()
|
D | target_mips.cc | 892 if (size == k64 || size == kDouble) { in RegClassForFieldLoadStore()
|
/art/compiler/dex/quick/arm64/ |
D | utility_arm64.cc | 1041 DCHECK(size == k64 || size == kDouble); in LoadBaseIndexed() 1059 case k64: in LoadBaseIndexed() 1126 DCHECK(size == k64 || size == kDouble); in StoreBaseIndexed() 1144 case k64: in StoreBaseIndexed() 1201 case k64: in LoadBaseDispBody() 1297 case k64: in StoreBaseDispBody()
|
D | int_arm64.cc | 710 RegLocation rl_dest = (size == k64) ? InlineTargetWide(info) : InlineTarget(info); in GenInlinedPeek() 715 if (size == k64) { in GenInlinedPeek() 730 if (size == k64) { in GenInlinedPoke() 911 LoadBaseIndexed(rs_src, rs_length, rs_tmp, 0, k64); in GenInlinedArrayCopyCharArray() 912 StoreBaseIndexed(rs_dst, rs_length, rs_tmp, 0, k64); in GenInlinedArrayCopyCharArray() 960 LIR* ldr = LoadBaseDisp(r_dest, 0, r_dest, wide ? k64 : kReference, kNotVolatile); in OpPcRelDexCacheArrayLoad() 1266 if (size == k64 || size == kDouble) { in GenArrayPut()
|
D | target_arm64.cc | 788 LoadBaseDisp(rs_xSELF, GetThreadOffset<8>(trampoline).Int32Value(), rs_xLR, k64, kNotVolatile); in LoadHelper()
|
/art/compiler/dex/quick/arm/ |
D | utility_arm.cc | 763 DCHECK((size == k64) || (size == kDouble)); in LoadBaseIndexed() 829 DCHECK((size == k64) || (size == kDouble)); in StoreBaseIndexed() 922 case k64: in LoadBaseDispBody() 1037 if (is_volatile == kVolatile && (size == k64 || size == kDouble) && in LoadBaseDisp() 1072 case k64: in StoreBaseDispBody() 1180 if (is_volatile == kVolatile && (size == k64 || size == kDouble) && in StoreBaseDisp()
|
D | int_arm.cc | 791 if (size == k64) { in GenInlinedPeek() 815 if (size == k64) { in GenInlinedPoke() 1468 if (size == k64 || size == kDouble) { in GenArrayPut()
|
D | target_arm.cc | 569 if (size == k64 || size == kDouble) { in RegClassForFieldLoadStore()
|
/art/compiler/optimizing/ |
D | intrinsics.cc | 56 case k64: in GetType()
|
/art/compiler/dex/ |
D | compiler_enums.h | 383 k64, enumerator
|
D | mir_dataflow.cc | 930 if (type_size == k64 || type_size == kDouble) { in HandleExtended() 1122 if (type_size == k64 || type_size == kDouble) { in DataFlowSSAFormatExtended()
|