Lines Matching refs:ops
29 def load_mmx : PatFrag<(ops node:$ptr), (x86mmx (load node:$ptr))>;
30 def load_mvmmx : PatFrag<(ops node:$ptr),
32 def bc_mmx : PatFrag<(ops node:$in), (x86mmx (bitconvert node:$in))>;
600 let MIOperandInfo = (ops ptr_rc, i8imm, ptr_rc_nosp, i32imm, i8imm);
606 let MIOperandInfo = (ops ptr_rc, i8imm, ptr_rc_nosp, i32imm, i8imm);
617 def loadv4f32 : PatFrag<(ops node:$ptr), (v4f32 (load node:$ptr))>;
618 def loadv2f64 : PatFrag<(ops node:$ptr), (v2f64 (load node:$ptr))>;
619 def loadv2i64 : PatFrag<(ops node:$ptr), (v2i64 (load node:$ptr))>;
623 def loadv8f32 : PatFrag<(ops node:$ptr), (v8f32 (load node:$ptr))>;
624 def loadv4f64 : PatFrag<(ops node:$ptr), (v4f64 (load node:$ptr))>;
625 def loadv4i64 : PatFrag<(ops node:$ptr), (v4i64 (load node:$ptr))>;
628 def loadv16f32 : PatFrag<(ops node:$ptr), (v16f32 (load node:$ptr))>;
629 def loadv8f64 : PatFrag<(ops node:$ptr), (v8f64 (load node:$ptr))>;
630 def loadv64i8 : PatFrag<(ops node:$ptr), (v64i8 (load node:$ptr))>;
631 def loadv32i16 : PatFrag<(ops node:$ptr), (v32i16 (load node:$ptr))>;
632 def loadv16i32 : PatFrag<(ops node:$ptr), (v16i32 (load node:$ptr))>;
633 def loadv8i64 : PatFrag<(ops node:$ptr), (v8i64 (load node:$ptr))>;
636 def extloadv2f32 : PatFrag<(ops node:$ptr), (v2f64 (extloadvf32 node:$ptr))>;
637 def extloadv4f32 : PatFrag<(ops node:$ptr), (v4f64 (extloadvf32 node:$ptr))>;
638 def extloadv8f32 : PatFrag<(ops node:$ptr), (v8f64 (extloadvf32 node:$ptr))>;
641 // math instruction such as the FP logical ops: andps, andnps, orps, xorps.
644 def loadf32_128 : PatFrag<(ops node:$ptr),
646 def loadf64_128 : PatFrag<(ops node:$ptr),
650 def alignedstore : PatFrag<(ops node:$val, node:$ptr),
656 def alignedstore256 : PatFrag<(ops node:$val, node:$ptr),
662 def alignedstore512 : PatFrag<(ops node:$val, node:$ptr),
668 def alignedload : PatFrag<(ops node:$ptr), (load node:$ptr), [{
673 def alignedX86vzload : PatFrag<(ops node:$ptr), (X86vzload node:$ptr), [{
678 def alignedload256 : PatFrag<(ops node:$ptr), (load node:$ptr), [{
683 def alignedload512 : PatFrag<(ops node:$ptr), (load node:$ptr), [{
687 def alignedloadfsf32 : PatFrag<(ops node:$ptr),
689 def alignedloadfsf64 : PatFrag<(ops node:$ptr),
694 def alignedloadv4f32 : PatFrag<(ops node:$ptr),
696 def alignedloadv2f64 : PatFrag<(ops node:$ptr),
698 def alignedloadv2i64 : PatFrag<(ops node:$ptr),
703 def alignedloadv8f32 : PatFrag<(ops node:$ptr),
705 def alignedloadv4f64 : PatFrag<(ops node:$ptr),
707 def alignedloadv4i64 : PatFrag<(ops node:$ptr),
711 def alignedloadv16f32 : PatFrag<(ops node:$ptr),
713 def alignedloadv16i32 : PatFrag<(ops node:$ptr),
715 def alignedloadv8f64 : PatFrag<(ops node:$ptr),
717 def alignedloadv8i64 : PatFrag<(ops node:$ptr),
726 def memop : PatFrag<(ops node:$ptr), (load node:$ptr), [{
731 def memopfsf32 : PatFrag<(ops node:$ptr), (f32 (memop node:$ptr))>;
732 def memopfsf64 : PatFrag<(ops node:$ptr), (f64 (memop node:$ptr))>;
736 def memopv4f32 : PatFrag<(ops node:$ptr), (v4f32 (memop node:$ptr))>;
737 def memopv2f64 : PatFrag<(ops node:$ptr), (v2f64 (memop node:$ptr))>;
738 def memopv2i64 : PatFrag<(ops node:$ptr), (v2i64 (memop node:$ptr))>;
741 // math instruction such as the FP logical ops: andps, andnps, orps, xorps.
744 def memopfsf32_128 : PatFrag<(ops node:$ptr),
746 def memopfsf64_128 : PatFrag<(ops node:$ptr),
753 def memop64 : PatFrag<(ops node:$ptr), (load node:$ptr), [{
757 def memopmmx : PatFrag<(ops node:$ptr), (x86mmx (memop64 node:$ptr))>;
759 def mgatherv4i32 : PatFrag<(ops node:$src1, node:$src2, node:$src3),
767 def mgatherv8i32 : PatFrag<(ops node:$src1, node:$src2, node:$src3),
775 def mgatherv2i64 : PatFrag<(ops node:$src1, node:$src2, node:$src3),
782 def mgatherv4i64 : PatFrag<(ops node:$src1, node:$src2, node:$src3),
789 def mgatherv8i64 : PatFrag<(ops node:$src1, node:$src2, node:$src3),
796 def mgatherv16i32 : PatFrag<(ops node:$src1, node:$src2, node:$src3),
804 def mscatterv2i64 : PatFrag<(ops node:$src1, node:$src2, node:$src3),
812 def mscatterv4i32 : PatFrag<(ops node:$src1, node:$src2, node:$src3),
820 def mscatterv4i64 : PatFrag<(ops node:$src1, node:$src2, node:$src3),
828 def mscatterv8i32 : PatFrag<(ops node:$src1, node:$src2, node:$src3),
836 def mscatterv8i64 : PatFrag<(ops node:$src1, node:$src2, node:$src3),
843 def mscatterv16i32 : PatFrag<(ops node:$src1, node:$src2, node:$src3),
852 def bc_v4f32 : PatFrag<(ops node:$in), (v4f32 (bitconvert node:$in))>;
853 def bc_v2f64 : PatFrag<(ops node:$in), (v2f64 (bitconvert node:$in))>;
854 def bc_v16i8 : PatFrag<(ops node:$in), (v16i8 (bitconvert node:$in))>;
855 def bc_v8i16 : PatFrag<(ops node:$in), (v8i16 (bitconvert node:$in))>;
856 def bc_v4i32 : PatFrag<(ops node:$in), (v4i32 (bitconvert node:$in))>;
857 def bc_v2i64 : PatFrag<(ops node:$in), (v2i64 (bitconvert node:$in))>;
860 def bc_v32i8 : PatFrag<(ops node:$in), (v32i8 (bitconvert node:$in))>;
861 def bc_v16i16 : PatFrag<(ops node:$in), (v16i16 (bitconvert node:$in))>;
862 def bc_v8i32 : PatFrag<(ops node:$in), (v8i32 (bitconvert node:$in))>;
863 def bc_v4i64 : PatFrag<(ops node:$in), (v4i64 (bitconvert node:$in))>;
864 def bc_v8f32 : PatFrag<(ops node:$in), (v8f32 (bitconvert node:$in))>;
867 def bc_v16i32 : PatFrag<(ops node:$in), (v16i32 (bitconvert node:$in))>;
868 def bc_v8i64 : PatFrag<(ops node:$in), (v8i64 (bitconvert node:$in))>;
869 def bc_v8f64 : PatFrag<(ops node:$in), (v8f64 (bitconvert node:$in))>;
870 def bc_v16f32 : PatFrag<(ops node:$in), (v16f32 (bitconvert node:$in))>;
872 def vzmovl_v2i64 : PatFrag<(ops node:$src),
875 def vzmovl_v4i32 : PatFrag<(ops node:$src),
879 def vzload_v2i64 : PatFrag<(ops node:$src),
927 def vextract128_extract : PatFrag<(ops node:$bigvec, node:$index),
933 def vinsert128_insert : PatFrag<(ops node:$bigvec, node:$smallvec,
941 def vextract256_extract : PatFrag<(ops node:$bigvec, node:$index),
947 def vinsert256_insert : PatFrag<(ops node:$bigvec, node:$smallvec,
954 def masked_load_aligned128 : PatFrag<(ops node:$src1, node:$src2, node:$src3),
961 def masked_load_aligned256 : PatFrag<(ops node:$src1, node:$src2, node:$src3),
968 def masked_load_aligned512 : PatFrag<(ops node:$src1, node:$src2, node:$src3),
975 def masked_load_unaligned : PatFrag<(ops node:$src1, node:$src2, node:$src3),
983 def X86mstore : PatFrag<(ops node:$src1, node:$src2, node:$src3),
988 def masked_store_aligned128 : PatFrag<(ops node:$src1, node:$src2, node:$src3),
995 def masked_store_aligned256 : PatFrag<(ops node:$src1, node:$src2, node:$src3),
1002 def masked_store_aligned512 : PatFrag<(ops node:$src1, node:$src2, node:$src3),
1009 def masked_store_unaligned : PatFrag<(ops node:$src1, node:$src2, node:$src3),
1017 def X86mtruncstore : PatFrag<(ops node:$src1, node:$src2, node:$src3),
1022 PatFrag<(ops node:$src1, node:$src2, node:$src3),
1027 PatFrag<(ops node:$src1, node:$src2, node:$src3),
1032 PatFrag<(ops node:$src1, node:$src2, node:$src3),