Lines Matching refs:XLen
2044 static bool CC_RISCVAssign2XLen(unsigned XLen, CCState &State, CCValAssign VA1, in CC_RISCVAssign2XLen() argument
2048 unsigned XLenInBytes = XLen / 8; in CC_RISCVAssign2XLen()
2086 unsigned XLen = DL.getLargestLegalIntTypeSizeInBits(); in CC_RISCV() local
2087 assert(XLen == 32 || XLen == 64); in CC_RISCV()
2088 MVT XLenVT = XLen == 32 ? MVT::i32 : MVT::i64; in CC_RISCV()
2132 } else if (UseGPRForF64 && XLen == 64 && ValVT == MVT::f64) { in CC_RISCV()
2144 unsigned TwoXLenInBytes = (2 * XLen) / 8; in CC_RISCV()
2162 if (UseGPRForF64 && XLen == 32 && ValVT == MVT::f64) { in CC_RISCV()
2206 return CC_RISCVAssign2XLen(XLen, State, VA, AF, ValNo, ValVT, LocVT, in CC_RISCV()
2221 Reg ? 0 : State.AllocateStack(XLen / 8, Align(XLen / 8)); in CC_RISCV()
3473 getIntrinsicForMaskedAtomicRMWBinOp(unsigned XLen, AtomicRMWInst::BinOp BinOp) { in getIntrinsicForMaskedAtomicRMWBinOp() argument
3474 if (XLen == 32) { in getIntrinsicForMaskedAtomicRMWBinOp()
3497 if (XLen == 64) { in getIntrinsicForMaskedAtomicRMWBinOp()
3526 unsigned XLen = Subtarget.getXLen(); in emitMaskedAtomicRMWIntrinsic() local
3528 Builder.getIntN(XLen, static_cast<uint64_t>(AI->getOrdering())); in emitMaskedAtomicRMWIntrinsic()
3532 getIntrinsicForMaskedAtomicRMWBinOp(XLen, AI->getOperation()), Tys); in emitMaskedAtomicRMWIntrinsic()
3534 if (XLen == 64) { in emitMaskedAtomicRMWIntrinsic()
3553 Builder.CreateSub(Builder.getIntN(XLen, XLen - ValWidth), ShiftAmt); in emitMaskedAtomicRMWIntrinsic()
3561 if (XLen == 64) in emitMaskedAtomicRMWIntrinsic()
3578 unsigned XLen = Subtarget.getXLen(); in emitMaskedAtomicCmpXchgIntrinsic() local
3579 Value *Ordering = Builder.getIntN(XLen, static_cast<uint64_t>(Ord)); in emitMaskedAtomicCmpXchgIntrinsic()
3581 if (XLen == 64) { in emitMaskedAtomicCmpXchgIntrinsic()
3592 if (XLen == 64) in emitMaskedAtomicCmpXchgIntrinsic()