Home
last modified time | relevance | path

Searched refs:UXTX (Results 1 – 21 of 21) sorted by relevance

/external/llvm/lib/Target/AArch64/MCTargetDesc/
DAArch64AddressingModes.h43 UXTX, enumerator
63 case AArch64_AM::UXTX: return "uxtx"; in getShiftExtendName()
130 case 3: return AArch64_AM::UXTX; in getExtendType()
157 case AArch64_AM::UXTX: return 3; break; in getExtendEncoding()
/external/vixl/src/aarch64/
Doperands-aarch64.cc322 VIXL_ASSERT(reg.Is64Bits() || ((extend != SXTX) && (extend != UXTX))); in Operand()
335 ((extend_ == UXTX) || (extend_ == SXTX) || in IsPlainRegister()
362 return Operand(reg_, reg_.Is64Bits() ? UXTX : UXTW, shift_amount_); in ToExtendedRegister()
Ddisasm-aarch64.cc165 const char *form = ((mode == UXTX) || (mode == SXTX)) ? "'Rds, 'Rns, 'Xm'Ext" in VisitAddSubExtended()
168 ((mode == UXTX) || (mode == SXTX)) ? "'Rns, 'Xm'Ext" : "'Rns, 'Wm'Ext"; in VisitAddSubExtended()
4744 (instr->GetExtendMode() == UXTX))) { in SubstituteExtendField()
4783 if (!((ext == UXTX) && (shift == 0))) { in SubstituteLSRegOffsetField()
Dmacro-assembler-aarch64.cc855 ((operand.GetExtend() != UXTX) && (operand.GetExtend() != SXTX))); in LogicalMacro()
1747 ((operand.GetExtend() != UXTX) && (operand.GetExtend() != SXTX))); in AddSubWithCarryMacro()
Dconstants-aarch64.h288 UXTX = 3, enumerator
Dsimulator-aarch64.cc415 case UXTX: in ExtendValue()
1287 VIXL_ASSERT((ext == UXTW) || (ext == UXTX) || (ext == SXTW) || (ext == SXTX)); in VisitLoadStoreRegisterOffset()
Dassembler-aarch64.cc4202 case UXTX: in EmitExtendShift()
4286 ext = UXTX; in LoadStoreMemOperand()
/external/v8/src/arm64/
Ddisasm-arm64.cc144 const char *form = ((mode == UXTX) || (mode == SXTX)) ? in VisitAddSubExtended()
146 const char *form_cmp = ((mode == UXTX) || (mode == SXTX)) ? in VisitAddSubExtended()
1648 (instr->ExtendMode() == UXTX))) { in SubstituteExtendField()
1681 if (!((ext == UXTX) && (shift == 0))) { in SubstituteLSRegOffsetField()
Dassembler-arm64-inl.h352 DCHECK(reg.Is64Bits() || ((extend != SXTX) && (extend != UXTX)));
383 return Operand(reg_, reg_.Is64Bits() ? UXTX : UXTW, shift_amount_);
Dconstants-arm64.h342 UXTX = 3, enumerator
Dassembler-arm64.cc2501 case UXTX: in EmitExtendShift()
2576 ext = UXTX; in LoadStore()
Dsimulator-arm64.cc995 case UXTX: in ExtendValue()
1615 DCHECK((ext == UXTW) || (ext == UXTX) || (ext == SXTW) || (ext == SXTX)); in VisitLoadStoreRegisterOffset()
Dmacro-assembler-arm64.cc147 ((operand.extend() != UXTX) && (operand.extend() != SXTX))); in LogicalMacro()
549 ((operand.extend() != UXTX) && (operand.extend() != SXTX))); in AddSubWithCarryMacro()
/external/llvm/lib/Target/AArch64/Utils/
DAArch64BaseInfo.h362 UXTX, enumerator
/external/llvm/lib/Target/AArch64/InstPrinter/
DAArch64InstPrinter.cpp1104 if (ExtType == AArch64_AM::UXTW || ExtType == AArch64_AM::UXTX) { in printArithExtend()
1108 ExtType == AArch64_AM::UXTX) || in printArithExtend()
/external/llvm/lib/Target/AArch64/AsmParser/
DAArch64AsmParser.cpp991 ET == AArch64_AM::UXTX || ET == AArch64_AM::SXTX || in isExtend()
1001 return ET != AArch64_AM::UXTX && ET != AArch64_AM::SXTX; in isExtend64()
1007 return (ET == AArch64_AM::UXTX || ET == AArch64_AM::SXTX || in isExtendLSL64()
1569 if (ET == AArch64_AM::LSL) ET = AArch64_AM::UXTX; in addExtend64Operands()
2392 .Case("uxtx", AArch64_AM::UXTX) in tryParseOptionalShiftExtend()
/external/vixl/test/aarch64/
Dtest-assembler-aarch64.cc638 __ Orr(x9, x0, Operand(x1, UXTX, 3)); in TEST()
732 __ Orn(x9, x0, Operand(x1, UXTX, 3)); in TEST()
799 __ And(x9, x0, Operand(x1, UXTX, 3)); in TEST()
937 __ Bic(x9, x0, Operand(x1, UXTX, 3)); in TEST()
1061 __ Eor(x9, x0, Operand(x1, UXTX, 3)); in TEST()
1128 __ Eon(x9, x0, Operand(x1, UXTX, 3)); in TEST()
8613 __ Adc(x13, x1, Operand(x2, UXTX, 4)); in TEST()
8625 __ Adc(x23, x1, Operand(x2, UXTX, 4)); in TEST()
13313 __ adds(xzr, x0, Operand(x1, UXTX)); in TEST()
13314 __ adds(xzr, x1, Operand(xzr, UXTX)); in TEST()
[all …]
Dtest-disasm-aarch64.cc429 COMPARE(add(x12, x13, Operand(x14, UXTX, 4)), "add x12, x13, x14, uxtx #4"); in TEST()
441 COMPARE(cmn(sp, Operand(xzr, UXTX, 3)), "cmn sp, xzr, lsl #3"); in TEST()
455 COMPARE(sub(x12, x13, Operand(x14, UXTX, 4)), "sub x12, x13, x14, uxtx #4"); in TEST()
467 COMPARE(cmp(sp, Operand(xzr, UXTX, 3)), "cmp sp, xzr, lsl #3"); in TEST()
/external/llvm/lib/Target/AArch64/
DAArch64InstrFormats.td1781 // UXTX and SXTX only.
1808 GPR64sponly, GPR64sp, GPR64, 24>; // UXTX #0
1811 GPR64sp, GPR64sponly, GPR64, 24>; // UXTX #0
1855 // UXTX and SXTX only.
1913 GPR64, GPR64sponly, GPR64, 24>; // UXTX #0
DAArch64InstrInfo.cpp1448 return (ExtType != AArch64_AM::UXTX) || AArch64_AM::getMemDoShift(Val); in isScaledAddr()
DAArch64ISelDAGToDAG.cpp596 assert(Ext != AArch64_AM::UXTX && Ext != AArch64_AM::SXTX); in SelectArithExtendedRegister()