Lines Matching refs:v8f32
344 def : Pat<(v4f32 (extract_subvector (v8f32 VR256:$src), (iPTR 0))),
345 (v4f32 (EXTRACT_SUBREG (v8f32 VR256:$src), sub_xmm))>;
367 (INSERT_SUBREG (v8f32 (IMPLICIT_DEF)), VR128:$src, sub_xmm)>;
377 def : Pat<(v8f32 (scalar_to_vector FR32:$src)),
423 def : Pat<(v4f64 (bitconvert (v8f32 VR256:$src))), (v4f64 VR256:$src)>;
428 def : Pat<(v8f32 (bitconvert (v8i32 VR256:$src))), (v8f32 VR256:$src)>;
429 def : Pat<(v8f32 (bitconvert (v4i64 VR256:$src))), (v8f32 VR256:$src)>;
430 def : Pat<(v8f32 (bitconvert (v4f64 VR256:$src))), (v8f32 VR256:$src)>;
431 def : Pat<(v8f32 (bitconvert (v32i8 VR256:$src))), (v8f32 VR256:$src)>;
432 def : Pat<(v8f32 (bitconvert (v16i16 VR256:$src))), (v8f32 VR256:$src)>;
433 def : Pat<(v4i64 (bitconvert (v8f32 VR256:$src))), (v4i64 VR256:$src)>;
440 def : Pat<(v32i8 (bitconvert (v8f32 VR256:$src))), (v32i8 VR256:$src)>;
445 def : Pat<(v8i32 (bitconvert (v8f32 VR256:$src))), (v8i32 VR256:$src)>;
448 def : Pat<(v16i16 (bitconvert (v8f32 VR256:$src))), (v16i16 VR256:$src)>;
494 [(set VR256:$dst, (v8f32 immAllZerosV))]>;
511 def : Pat<(bc_v32i8 (v8f32 immAllZerosV)),
515 def : Pat<(bc_v16i16 (v8f32 immAllZerosV)),
519 def : Pat<(bc_v8i32 (v8f32 immAllZerosV)),
523 def : Pat<(bc_v4i64 (v8f32 immAllZerosV)),
646 def : Pat<(v8f32 (X86vzmovl (insert_subvector undef,
676 def : Pat<(v8f32 (X86Movss VR256:$src1, VR256:$src2)),
678 (VMOVSSrr (EXTRACT_SUBREG (v8f32 VR256:$src1), sub_xmm),
679 (EXTRACT_SUBREG (v8f32 VR256:$src2), sub_xmm)),
887 [(alignedstore256 (v8f32 VR256:$src), addr:$dst)],
895 [(store (v8f32 VR256:$src), addr:$dst)],
947 def : Pat<(v8f32 (X86vzmovl
1064 (v8f32 VR256:$src), (iPTR 0))), addr:$dst),
1083 (v8f32 VR256:$src), (iPTR 0))), addr:$dst),
2093 def : Pat<(v8f32 (sint_to_fp (v8i32 VR256:$src))),
2095 def : Pat<(v8f32 (sint_to_fp (bc_v8i32 (loadv4i64 addr:$src)))),
2098 def : Pat<(v8i32 (fp_to_sint (v8f32 VR256:$src))),
2553 def : Pat<(v8i32 (X86cmpp (v8f32 VR256:$src1), VR256:$src2, imm:$cc)),
2554 (VCMPPSYrri (v8f32 VR256:$src1), (v8f32 VR256:$src2), imm:$cc)>;
2555 def : Pat<(v8i32 (X86cmpp (v8f32 VR256:$src1), (loadv8f32 addr:$src2), imm:$cc)),
2556 (VCMPPSYrmi (v8f32 VR256:$src1), addr:$src2, imm:$cc)>;
2600 defm VSHUFPSY : sse12_shuffle<VR256, f256mem, v8f32,
2699 defm VUNPCKHPSY: sse12_unpack_interleave<0x15, X86Unpckh, v8f32, loadv8f32,
2705 defm VUNPCKLPSY: sse12_unpack_interleave<0x14, X86Unpckl, v8f32, loadv8f32,
2945 [(set VR256:$dst, (OpNode (bc_v4i64 (v8f32 VR256:$src1)),
3037 OpNode, VR256, v8f32, f256mem, loadv8f32,
3460 [(set VR256:$dst, (v8f32 (OpNode VR256:$src)))],
3629 [(alignednontemporalstore (v8f32 VR256:$src),
5122 v8f32, VR256, loadv8f32, f256mem>, VEX, VEX_L;
5124 v8f32, VR256, loadv8f32, f256mem>, VEX, VEX_L;
5317 def : Pat<(v8f32 (X86Addsub (v8f32 VR256:$lhs), (v8f32 VR256:$rhs))),
5319 def : Pat<(v8f32 (X86Addsub (v8f32 VR256:$lhs), (loadv8f32 addr:$rhs))),
5384 defm VHADDPSY : S3D_Int<0x7C, "vhaddps", v8f32, VR256, f256mem,
5386 defm VHSUBPSY : S3D_Int<0x7D, "vhsubps", v8f32, VR256, f256mem,
5791 def : Pat<(v8f32 (X86PAlignr VR256:$src1, VR256:$src2, (i8 imm:$imm))),
6584 def : Pat<(v8f32 (ffloor VR256:$src)),
6586 def : Pat<(v8f32 (fnearbyint VR256:$src)),
6588 def : Pat<(v8f32 (fceil VR256:$src)),
6590 def : Pat<(v8f32 (frint VR256:$src)),
6592 def : Pat<(v8f32 (ftrunc VR256:$src)),
6712 defm VTESTPSY : avx_bittest<0x0E, "vtestps", VR256, f256mem, loadv8f32, v8f32>,
7014 defm VBLENDPSY : SS41I_binop_rmi<0x0C, "vblendps", X86Blendi, v8f32,
7153 def : Pat<(v8f32 (vselect (v8i32 VR256:$mask), (v8f32 VR256:$src1),
7154 (v8f32 VR256:$src2))),
7189 def : Pat<(v8f32 (X86vzmovl (v8f32 VR256:$src))),
7190 (VBLENDPSYrri (v8f32 (AVX_SET0)), VR256:$src, (i8 1))>;
7197 def : Pat<(v8f32 (X86vzmovl (insert_subvector undef,
7842 f32mem, v8f32, loadf32,
7891 def : Pat<(vinsert128_insert:$ins (v8f32 VR256:$src1), (v4f32 VR128:$src2),
7900 def : Pat<(vinsert128_insert:$ins (v8f32 VR256:$src1), (loadv4f32 addr:$src2),
7968 (v8f32 VR256:$src1),
7975 def : Pat<(store (v4f32 (vextract128_extract:$ext (v8f32 VR256:$src1),
8096 loadv4i64, int_x86_avx_vpermilvar_ps_256, v8f32>, VEX_L;
8106 def : Pat<(v8f32 (X86VPermilpv VR256:$src1, (v8i32 VR256:$src2))),
8108 def : Pat<(v8f32 (X86VPermilpv VR256:$src1, (bc_v8i32 (loadv4i64 addr:$src2)))),
8147 [(set VR256:$dst, (v8f32 (X86VPerm2x128 VR256:$src1, VR256:$src2,
8366 def : Pat<(v8f32 (X86VBroadcast (v4f32 VR128:$src))),
8387 def : Pat<(v8f32 (X86VBroadcast (v8f32 VR256:$src))),
8388 (VBROADCASTSSYrr (v4f32 (EXTRACT_SUBREG (v8f32 VR256:$src),
8399 def : Pat<(v8f32 (X86VBroadcast FR32:$src)),
8455 def : Pat<(v8f32 (X86VBroadcast FR32:$src)),
8456 (VINSERTF128rr (INSERT_SUBREG (v8f32 (IMPLICIT_DEF)),
8505 defm VPERMPS : avx2_perm<0x16, "vpermps", loadv8f32, v8f32, WriteFShuffle256>;
8707 def: Pat<(masked_store addr:$ptr, (v8i32 VR256:$mask), (v8f32 VR256:$src)),
8719 def: Pat<(v8f32 (masked_load addr:$ptr, (v8i32 VR256:$mask), undef)),
8722 def: Pat<(v8f32 (masked_load addr:$ptr, (v8i32 VR256:$mask),
8726 def: Pat<(v8f32 (masked_load addr:$ptr, (v8i32 VR256:$mask), (v8f32 VR256:$src0))),