Lines Matching refs:vmrg
15 ; CHECK-NEXT: vmrg %v0, %v0, %v1, %vm1
17 …%4 = tail call fast <256 x double> @llvm.ve.vl.vmrg.vvvml(<256 x double> %0, <256 x double> %1, <2…
22 declare <256 x double> @llvm.ve.vl.vmrg.vvvml(<256 x double>, <256 x double>, <256 x i1>, i32)
30 ; CHECK-NEXT: vmrg %v2, %v0, %v1, %vm1
35 …%5 = tail call fast <256 x double> @llvm.ve.vl.vmrg.vvvmvl(<256 x double> %0, <256 x double> %1, <…
40 declare <256 x double> @llvm.ve.vl.vmrg.vvvmvl(<256 x double>, <256 x double>, <256 x i1>, <256 x d…
48 ; CHECK-NEXT: vmrg %v0, %s0, %v0, %vm1
50 …%4 = tail call fast <256 x double> @llvm.ve.vl.vmrg.vsvml(i64 %0, <256 x double> %1, <256 x i1> %2…
55 declare <256 x double> @llvm.ve.vl.vmrg.vsvml(i64, <256 x double>, <256 x i1>, i32)
63 ; CHECK-NEXT: vmrg %v1, %s0, %v0, %vm1
68 …%5 = tail call fast <256 x double> @llvm.ve.vl.vmrg.vsvmvl(i64 %0, <256 x double> %1, <256 x i1> %…
73 declare <256 x double> @llvm.ve.vl.vmrg.vsvmvl(i64, <256 x double>, <256 x i1>, <256 x double>, i32)
81 ; CHECK-NEXT: vmrg %v0, 8, %v0, %vm1
83 …%3 = tail call fast <256 x double> @llvm.ve.vl.vmrg.vsvml(i64 8, <256 x double> %0, <256 x i1> %1,…
93 ; CHECK-NEXT: vmrg %v1, 8, %v0, %vm1
98 …%4 = tail call fast <256 x double> @llvm.ve.vl.vmrg.vsvmvl(i64 8, <256 x double> %0, <256 x i1> %1…
108 ; CHECK-NEXT: vmrg.w %v0, %v0, %v1, %vm2
123 ; CHECK-NEXT: vmrg.w %v2, %v0, %v1, %vm2