Lines Matching refs:SseReRg
984 i->Ain.SseReRg.op = op; in AMD64Instr_SseReRg()
985 i->Ain.SseReRg.src = re; in AMD64Instr_SseReRg()
986 i->Ain.SseReRg.dst = rg; in AMD64Instr_SseReRg()
1342 vex_printf("%s ", showAMD64SseOp(i->Ain.SseReRg.op)); in ppAMD64Instr()
1343 ppHRegAMD64(i->Ain.SseReRg.src); in ppAMD64Instr()
1345 ppHRegAMD64(i->Ain.SseReRg.dst); in ppAMD64Instr()
1658 if ( (i->Ain.SseReRg.op == Asse_XOR in getRegUsage_AMD64Instr()
1659 || i->Ain.SseReRg.op == Asse_CMPEQ32) in getRegUsage_AMD64Instr()
1660 && sameHReg(i->Ain.SseReRg.src, i->Ain.SseReRg.dst)) { in getRegUsage_AMD64Instr()
1665 addHRegUse(u, HRmWrite, i->Ain.SseReRg.dst); in getRegUsage_AMD64Instr()
1667 addHRegUse(u, HRmRead, i->Ain.SseReRg.src); in getRegUsage_AMD64Instr()
1668 addHRegUse(u, i->Ain.SseReRg.op == Asse_MOV in getRegUsage_AMD64Instr()
1670 i->Ain.SseReRg.dst); in getRegUsage_AMD64Instr()
1879 mapReg(m, &i->Ain.SseReRg.src); in mapRegs_AMD64Instr()
1880 mapReg(m, &i->Ain.SseReRg.dst); in mapRegs_AMD64Instr()
1931 if (i->Ain.SseReRg.op != Asse_MOV) in isMove_AMD64Instr()
1933 *src = i->Ain.SseReRg.src; in isMove_AMD64Instr()
1934 *dst = i->Ain.SseReRg.dst; in isMove_AMD64Instr()
3679 rexAMode_R_enc_enc( vregEnc3210(i->Ain.SseReRg.dst), in emit_AMD64Instr()
3680 vregEnc3210(i->Ain.SseReRg.src) )); in emit_AMD64Instr()
3682 switch (i->Ain.SseReRg.op) { in emit_AMD64Instr()
3740 p = doAMode_R_enc_enc(p, vregEnc3210(i->Ain.SseReRg.dst), in emit_AMD64Instr()
3741 vregEnc3210(i->Ain.SseReRg.src) ); in emit_AMD64Instr()