/art/test/442-checker-constant-folding/src/ |
D | Main.java | 1379 long imm = 33L; in ReturnInt33() local 1380 return (int) imm; in ReturnInt33() 1396 float imm = 1.0e34f; in ReturnIntMax() local 1397 return (int) imm; in ReturnIntMax() 1413 double imm = Double.NaN; in ReturnInt0() local 1414 return (int) imm; in ReturnInt0() 1430 int imm = 33; in ReturnLong33() local 1431 return (long) imm; in ReturnLong33() 1447 float imm = 34.0f; in ReturnLong34() local 1448 return (long) imm; in ReturnLong34() [all …]
|
/art/compiler/optimizing/ |
D | scheduler_arm64.cc | 94 int64_t imm = Int64FromConstant(instr->GetRight()->AsConstant()); in VisitDiv() local 95 if (imm == 0) { in VisitDiv() 98 } else if (imm == 1 || imm == -1) { in VisitDiv() 101 } else if (IsPowerOfTwo(AbsOrMin(imm))) { in VisitDiv() 105 DCHECK(imm <= -2 || imm >= 2); in VisitDiv() 162 int64_t imm = Int64FromConstant(instruction->GetRight()->AsConstant()); in VisitRem() local 163 if (imm == 0) { in VisitRem() 166 } else if (imm == 1 || imm == -1) { in VisitRem() 169 } else if (IsPowerOfTwo(AbsOrMin(imm))) { in VisitRem() 173 DCHECK(imm <= -2 || imm >= 2); in VisitRem()
|
D | scheduler_arm.cc | 815 void SchedulingLatencyVisitorARM::HandleDivRemConstantIntegralLatencies(int32_t imm) { in HandleDivRemConstantIntegralLatencies() argument 816 if (imm == 0) { in HandleDivRemConstantIntegralLatencies() 819 } else if (imm == 1 || imm == -1) { in HandleDivRemConstantIntegralLatencies() 821 } else if (IsPowerOfTwo(AbsOrMin(imm))) { in HandleDivRemConstantIntegralLatencies() 836 int32_t imm = Int32ConstantFrom(rhs->AsConstant()); in VisitDiv() local 837 HandleDivRemConstantIntegralLatencies(imm); in VisitDiv() 899 int32_t imm = Int32ConstantFrom(rhs->AsConstant()); in VisitRem() local 900 HandleDivRemConstantIntegralLatencies(imm); in VisitRem()
|
D | scheduler_arm.h | 126 void HandleDivRemConstantIntegralLatencies(int32_t imm);
|
D | code_generator_mips.cc | 2114 int32_t imm = CodeGenerator::GetInt32ValueOf(right->AsConstant()); in HandleBinaryOp() local 2116 can_use_imm = IsUint<16>(imm); in HandleBinaryOp() 2120 imm = -imm; in HandleBinaryOp() 2124 int16_t imm_high = High16Bits(imm); in HandleBinaryOp() 2125 int16_t imm_low = Low16Bits(imm); in HandleBinaryOp() 2131 can_use_imm = IsInt<16>(imm); in HandleBinaryOp() 3848 int64_t imm = Int64FromConstant(second.GetConstant()); in DivRemOneOrMinusOne() local 3849 DCHECK(imm == 1 || imm == -1); in DivRemOneOrMinusOne() 3858 if (imm == -1) { in DivRemOneOrMinusOne() 3875 if (imm == -1) { in DivRemOneOrMinusOne() [all …]
|
D | code_generator_mips64.cc | 1960 int64_t imm = CodeGenerator::GetInt64ValueOf(right->AsConstant()); in HandleBinaryOp() local 1962 can_use_imm = IsUint<16>(imm); in HandleBinaryOp() 1967 if (!(type == DataType::Type::kInt32 && imm == INT32_MIN)) { in HandleBinaryOp() 1968 imm = -imm; in HandleBinaryOp() 1972 can_use_imm = IsInt<16>(imm) || (Low16Bits(imm) == 0) || single_use; in HandleBinaryOp() 1974 can_use_imm = IsInt<16>(imm) || (IsInt<32>(imm) && (Low16Bits(imm) == 0)) || single_use; in HandleBinaryOp() 3362 int64_t imm = Int64FromConstant(second.GetConstant()); in DivRemOneOrMinusOne() local 3363 DCHECK(imm == 1 || imm == -1); in DivRemOneOrMinusOne() 3368 if (imm == -1) { in DivRemOneOrMinusOne() 3391 int64_t imm = Int64FromConstant(second.GetConstant()); in DivRemByPowerOfTwo() local [all …]
|
D | code_generator_arm_vixl.cc | 3954 int32_t imm = Int32ConstantFrom(second); in DivRemOneOrMinusOne() local 3955 DCHECK(imm == 1 || imm == -1); in DivRemOneOrMinusOne() 3960 if (imm == 1) { in DivRemOneOrMinusOne() 3979 int32_t imm = Int32ConstantFrom(second); in DivRemByPowerOfTwo() local 3980 uint32_t abs_imm = static_cast<uint32_t>(AbsOrMin(imm)); in DivRemByPowerOfTwo() 3993 if (imm < 0) { in DivRemByPowerOfTwo() 4014 int32_t imm = Int32ConstantFrom(second); in GenerateDivRemWithAnyConstant() local 4018 CalculateMagicAndShiftForDivRem(imm, /* is_long= */ false, &magic, &shift); in GenerateDivRemWithAnyConstant() 4024 if (imm > 0 && magic < 0) { in GenerateDivRemWithAnyConstant() 4026 } else if (imm < 0 && magic > 0) { in GenerateDivRemWithAnyConstant() [all …]
|
/art/compiler/utils/x86/ |
D | assembler_x86.h | 325 void pushl(const Immediate& imm); 335 void movl(const Address& dst, const Immediate& imm); 354 void rorl(Register reg, const Immediate& imm); 356 void roll(Register reg, const Immediate& imm); 365 void movb(const Address& dst, const Immediate& imm); 373 void movw(const Address& dst, const Immediate& imm); 489 void roundsd(XmmRegister dst, XmmRegister src, const Immediate& imm); 490 void roundss(XmmRegister dst, XmmRegister src, const Immediate& imm); 558 void shufpd(XmmRegister dst, XmmRegister src, const Immediate& imm); 559 void shufps(XmmRegister dst, XmmRegister src, const Immediate& imm); [all …]
|
D | jni_macro_assembler_x86.h | 63 void StoreImmediateToFrame(FrameOffset dest, uint32_t imm, ManagedRegister scratch) override;
|
D | jni_macro_assembler_x86.cc | 162 void X86JNIMacroAssembler::StoreImmediateToFrame(FrameOffset dest, uint32_t imm, ManagedRegister) { in StoreImmediateToFrame() argument 163 __ movl(Address(ESP, dest), Immediate(imm)); in StoreImmediateToFrame()
|
/art/compiler/utils/x86_64/ |
D | assembler_x86_64.h | 369 void pushq(const Immediate& imm); 385 void movq(const Address& dst, const Immediate& imm); 387 void movl(const Address& dst, const Immediate& imm); 399 void movb(const Address& dst, const Immediate& imm); 407 void movw(const Address& dst, const Immediate& imm); 529 void roundsd(XmmRegister dst, XmmRegister src, const Immediate& imm); 530 void roundss(XmmRegister dst, XmmRegister src, const Immediate& imm); 597 void shufpd(XmmRegister dst, XmmRegister src, const Immediate& imm); 598 void shufps(XmmRegister dst, XmmRegister src, const Immediate& imm); 599 void pshufd(XmmRegister dst, XmmRegister src, const Immediate& imm); [all …]
|
D | jni_macro_assembler_x86_64.h | 64 void StoreImmediateToFrame(FrameOffset dest, uint32_t imm, ManagedRegister scratch) override;
|
D | jni_macro_assembler_x86_64.cc | 200 uint32_t imm, in StoreImmediateToFrame() argument 202 __ movl(Address(CpuRegister(RSP), dest), Immediate(imm)); // TODO(64) movq? in StoreImmediateToFrame()
|
D | assembler_x86_64_test.cc | 113 x86_64::Immediate imm(value); in TEST() local 114 EXPECT_FALSE(imm.is_int8()); in TEST() 115 EXPECT_FALSE(imm.is_int16()); in TEST() 116 EXPECT_FALSE(imm.is_int32()); in TEST()
|
/art/compiler/utils/arm/ |
D | assembler_arm_vixl.h | 142 void Vmov(vixl32::DRegister rd, double imm) { in Vmov() argument 143 if (vixl::VFP::IsImmFP64(imm)) { in Vmov() 144 MacroAssembler::Vmov(rd, imm); in Vmov() 146 MacroAssembler::Vldr(rd, imm); in Vmov()
|
D | jni_macro_assembler_arm_vixl.h | 70 void StoreImmediateToFrame(FrameOffset dest, uint32_t imm, ManagedRegister scratch) override;
|
D | jni_macro_assembler_arm_vixl.cc | 313 uint32_t imm, in StoreImmediateToFrame() argument 318 asm_.LoadImmediate(scratch, imm); in StoreImmediateToFrame()
|
/art/compiler/utils/ |
D | assembler_test.h | 196 for (int64_t imm : imms) { variable 197 ImmType new_imm = CreateImmediate(imm); 218 sreg << imm * multiplier + bias; 252 for (int64_t imm : imms) { in RepeatTemplatedRegistersImmBits() local 253 ImmType new_imm = CreateImmediate(imm); in RepeatTemplatedRegistersImmBits() 280 sreg << imm + bias; in RepeatTemplatedRegistersImmBits() 313 for (int64_t imm : imms) { in RepeatTemplatedImmBitsRegisters() local 314 ImmType new_imm = CreateImmediate(imm); in RepeatTemplatedImmBitsRegisters() 335 sreg << imm; in RepeatTemplatedImmBitsRegisters() 363 for (int64_t imm : imms) { in RepeatTemplatedRegisterImmBits() local [all …]
|
D | jni_macro_assembler.h | 86 virtual void StoreImmediateToFrame(FrameOffset dest, uint32_t imm, ManagedRegister scratch) = 0;
|
/art/dex2oat/linker/arm/ |
D | relative_patcher_thumb2.cc | 98 uint32_t imm = (diff16 >> 11) & 0x1u; in PatchPcRelativeReference() local 101 insn = (insn & 0xfbf08f00u) | (imm << 26) | (imm4 << 16) | (imm3 << 12) | imm8; in PatchPcRelativeReference()
|
/art/compiler/utils/arm64/ |
D | jni_macro_assembler_arm64.h | 72 void StoreImmediateToFrame(FrameOffset dest, uint32_t imm, ManagedRegister scratch) override;
|
D | jni_macro_assembler_arm64.cc | 158 uint32_t imm, in StoreImmediateToFrame() argument 162 LoadImmediate(scratch.AsXRegister(), imm); in StoreImmediateToFrame()
|
/art/runtime/interpreter/mterp/mips/ |
D | main.S | 675 #define LOAD_IMM(dest, imm) li dest, imm argument
|
/art/compiler/utils/mips/ |
D | assembler_mips.h | 1250 void StoreImmediateToFrame(FrameOffset dest, uint32_t imm, ManagedRegister mscratch) override; 1686 uint32_t EmitI(int opcode, Register rs, Register rt, uint16_t imm); 1690 uint32_t EmitFI(int opcode, int fmt, FRegister rt, uint16_t imm);
|