Lines Matching refs:arg0
68 %arg0 = bitcast <4 x i64> %a0 to <32 x i8>
70 %res = add <32 x i8> %arg0, %arg1
85 %arg0 = bitcast <4 x i64> %a0 to <16 x i16>
87 %res = add <16 x i16> %arg0, %arg1
102 %arg0 = bitcast <4 x i64> %a0 to <8 x i32>
104 %res = add <8 x i32> %arg0, %arg1
133 %arg0 = bitcast <4 x i64> %a0 to <32 x i8>
135 %res = call <32 x i8> @llvm.x86.avx2.padds.b(<32 x i8> %arg0, <32 x i8> %arg1)
151 %arg0 = bitcast <4 x i64> %a0 to <16 x i16>
153 %res = call <16 x i16> @llvm.x86.avx2.padds.w(<16 x i16> %arg0, <16 x i16> %arg1)
169 %arg0 = bitcast <4 x i64> %a0 to <32 x i8>
171 %res = call <32 x i8> @llvm.x86.avx2.paddus.b(<32 x i8> %arg0, <32 x i8> %arg1)
187 %arg0 = bitcast <4 x i64> %a0 to <16 x i16>
189 %res = call <16 x i16> @llvm.x86.avx2.paddus.w(<16 x i16> %arg0, <16 x i16> %arg1)
205 %arg0 = bitcast <4 x i64> %a0 to <32 x i8>
207 …%shuf = shufflevector <32 x i8> %arg0, <32 x i8> %arg1, <32 x i32> <i32 2, i32 3, i32 4, i32 5, i3…
222 %arg0 = bitcast <4 x i64> %a0 to <32 x i8>
224 …%shuf = shufflevector <32 x i8> %arg0, <32 x i8> %arg1, <32 x i32> <i32 1, i32 2, i32 3, i32 4, i3…
272 %arg0 = bitcast <4 x i64> %a0 to <32 x i8>
274 %res = call <32 x i8> @llvm.x86.avx2.pavg.b(<32 x i8> %arg0, <32 x i8> %arg1)
290 %arg0 = bitcast <4 x i64> %a0 to <16 x i16>
292 %res = call <16 x i16> @llvm.x86.avx2.pavg.w(<16 x i16> %arg0, <16 x i16> %arg1)
308 %arg0 = bitcast <4 x i64> %a0 to <16 x i16>
310 …%shuf = shufflevector <16 x i16> %arg0, <16 x i16> %arg1, <16 x i32> <i32 0, i32 17, i32 2, i32 3,…
325 %arg0 = bitcast <2 x i64> %a0 to <4 x i32>
327 %shuf = shufflevector <4 x i32> %arg0, <4 x i32> %arg1, <4 x i32> <i32 4, i32 1, i32 6, i32 3>
342 %arg0 = bitcast <4 x i64> %a0 to <8 x i32>
344 …%shuf = shufflevector <8 x i32> %arg0, <8 x i32> %arg1, <8 x i32> <i32 8, i32 1, i32 10, i32 3, i3…
359 %arg0 = bitcast <4 x i64> %a0 to <32 x i8>
362 %call = call <32 x i8> @llvm.x86.avx2.pblendvb(<32 x i8> %arg0, <32 x i8> %arg1, <32 x i8> %arg2)
378 %arg0 = bitcast <2 x i64> %a0 to <16 x i8>
379 %shuf = shufflevector <16 x i8> %arg0, <16 x i8> undef, <16 x i32> zeroinitializer
394 %arg0 = bitcast <4 x i64> %a0 to <32 x i8>
395 %shuf = shufflevector <32 x i8> %arg0, <32 x i8> undef, <32 x i32> zeroinitializer
410 %arg0 = bitcast <2 x i64> %a0 to <4 x i32>
411 %shuf = shufflevector <4 x i32> %arg0, <4 x i32> undef, <4 x i32> zeroinitializer
426 %arg0 = bitcast <4 x i64> %a0 to <8 x i32>
427 %shuf = shufflevector <8 x i32> %arg0, <8 x i32> undef, <8 x i32> zeroinitializer
540 %arg0 = bitcast <2 x i64> %a0 to <8 x i16>
541 %shuf = shufflevector <8 x i16> %arg0, <8 x i16> undef, <8 x i32> zeroinitializer
556 %arg0 = bitcast <4 x i64> %a0 to <16 x i16>
557 %shuf = shufflevector <16 x i16> %arg0, <16 x i16> undef, <16 x i32> zeroinitializer
572 %arg0 = bitcast <4 x i64> %a0 to <32 x i8>
573 …%shuf = shufflevector <32 x i8> zeroinitializer, <32 x i8> %arg0, <32 x i32> <i32 13, i32 14, i32 …
588 %arg0 = bitcast <4 x i64> %a0 to <32 x i8>
589 …%shuf = shufflevector <32 x i8> %arg0, <32 x i8> zeroinitializer, <32 x i32> <i32 3, i32 4, i32 5,…
604 %arg0 = bitcast <4 x i64> %a0 to <32 x i8>
606 %cmp = icmp eq <32 x i8> %arg0, %arg1
622 %arg0 = bitcast <4 x i64> %a0 to <16 x i16>
624 %cmp = icmp eq <16 x i16> %arg0, %arg1
640 %arg0 = bitcast <4 x i64> %a0 to <8 x i32>
642 %cmp = icmp eq <8 x i32> %arg0, %arg1
673 %arg0 = bitcast <4 x i64> %a0 to <32 x i8>
675 %cmp = icmp sgt <32 x i8> %arg0, %arg1
691 %arg0 = bitcast <4 x i64> %a0 to <16 x i16>
693 %cmp = icmp sgt <16 x i16> %arg0, %arg1
709 %arg0 = bitcast <4 x i64> %a0 to <8 x i32>
711 %cmp = icmp sgt <8 x i32> %arg0, %arg1
742 %arg0 = bitcast <2 x i64> %a0 to <16 x i8>
743 %ext = sext <16 x i8> %arg0 to <16 x i16>
758 %arg0 = bitcast <2 x i64> %a0 to <16 x i8>
759 …%shuf = shufflevector <16 x i8> %arg0, <16 x i8> undef, <8 x i32> <i32 0, i32 1, i32 2, i32 3, i32…
775 %arg0 = bitcast <2 x i64> %a0 to <16 x i8>
776 %shuf = shufflevector <16 x i8> %arg0, <16 x i8> undef, <4 x i32> <i32 0, i32 1, i32 2, i32 3>
791 %arg0 = bitcast <2 x i64> %a0 to <8 x i16>
792 %ext = sext <8 x i16> %arg0 to <8 x i32>
807 %arg0 = bitcast <2 x i64> %a0 to <8 x i16>
808 %shuf = shufflevector <8 x i16> %arg0, <8 x i16> undef, <4 x i32> <i32 0, i32 1, i32 2, i32 3>
823 %arg0 = bitcast <2 x i64> %a0 to <4 x i32>
824 %ext = sext <4 x i32> %arg0 to <4 x i64>
838 %arg0 = bitcast <2 x i64> %a0 to <16 x i8>
839 %ext = zext <16 x i8> %arg0 to <16 x i16>
854 %arg0 = bitcast <2 x i64> %a0 to <16 x i8>
855 …%shuf = shufflevector <16 x i8> %arg0, <16 x i8> undef, <8 x i32> <i32 0, i32 1, i32 2, i32 3, i32…
871 %arg0 = bitcast <2 x i64> %a0 to <16 x i8>
872 %shuf = shufflevector <16 x i8> %arg0, <16 x i8> undef, <4 x i32> <i32 0, i32 1, i32 2, i32 3>
887 %arg0 = bitcast <2 x i64> %a0 to <8 x i16>
888 %ext = zext <8 x i16> %arg0 to <8 x i32>
903 %arg0 = bitcast <2 x i64> %a0 to <8 x i16>
904 %shuf = shufflevector <8 x i16> %arg0, <8 x i16> undef, <4 x i32> <i32 0, i32 1, i32 2, i32 3>
919 %arg0 = bitcast <2 x i64> %a0 to <4 x i32>
920 %ext = zext <4 x i32> %arg0 to <4 x i64>
950 %arg0 = bitcast <4 x i64> %a0 to <16 x i16>
952 %res = call <16 x i16> @llvm.x86.avx2.phadd.w(<16 x i16> %arg0, <16 x i16> %arg1)
968 %arg0 = bitcast <4 x i64> %a0 to <8 x i32>
970 %res = call <8 x i32> @llvm.x86.avx2.phadd.d(<8 x i32> %arg0, <8 x i32> %arg1)
986 %arg0 = bitcast <4 x i64> %a0 to <16 x i16>
988 %res = call <16 x i16> @llvm.x86.avx2.phadd.sw(<16 x i16> %arg0, <16 x i16> %arg1)
1004 %arg0 = bitcast <4 x i64> %a0 to <16 x i16>
1006 %res = call <16 x i16> @llvm.x86.avx2.phsub.w(<16 x i16> %arg0, <16 x i16> %arg1)
1022 %arg0 = bitcast <4 x i64> %a0 to <8 x i32>
1024 %res = call <8 x i32> @llvm.x86.avx2.phsub.d(<8 x i32> %arg0, <8 x i32> %arg1)
1040 %arg0 = bitcast <4 x i64> %a0 to <16 x i16>
1042 %res = call <16 x i16> @llvm.x86.avx2.phsub.sw(<16 x i16> %arg0, <16 x i16> %arg1)
1063 %arg0 = bitcast i32 *%a0 to i8*
1066 …%call = call <4 x i32> @llvm.x86.avx2.gather.d.d(<4 x i32> undef, i8* %arg0, <4 x i32> %arg1, <4 x…
1083 %arg0 = bitcast <2 x i64> %a0 to <4 x i32>
1087 …%call = call <4 x i32> @llvm.x86.avx2.gather.d.d(<4 x i32> %arg0, i8* %arg1, <4 x i32> %arg2, <4 x…
1107 %arg0 = bitcast i32 *%a0 to i8*
1110 …%call = call <8 x i32> @llvm.x86.avx2.gather.d.d.256(<8 x i32> undef, i8* %arg0, <8 x i32> %arg1, …
1127 %arg0 = bitcast <4 x i64> %a0 to <8 x i32>
1131 …%call = call <8 x i32> @llvm.x86.avx2.gather.d.d.256(<8 x i32> %arg0, i8* %arg1, <8 x i32> %arg2, …
1151 %arg0 = bitcast i64 *%a0 to i8*
1153 …%res = call <2 x i64> @llvm.x86.avx2.gather.d.q(<2 x i64> undef, i8* %arg0, <4 x i32> %arg1, <2 x …
1190 %arg0 = bitcast i64 *%a0 to i8*
1192 …%res = call <4 x i64> @llvm.x86.avx2.gather.d.q.256(<4 x i64> undef, i8* %arg0, <4 x i32> %arg1, <…
1229 %arg0 = bitcast double *%a0 to i8*
1234 …%res = call <2 x double> @llvm.x86.avx2.gather.d.pd(<2 x double> undef, i8* %arg0, <4 x i32> %arg1…
1273 %arg0 = bitcast double *%a0 to i8*
1276 …%res = call <4 x double> @llvm.x86.avx2.gather.d.pd.256(<4 x double> undef, i8* %arg0, <4 x i32> %…
1313 %arg0 = bitcast float *%a0 to i8*
1318 …%call = call <4 x float> @llvm.x86.avx2.gather.d.ps(<4 x float> undef, i8* %arg0, <4 x i32> %arg1,…
1357 %arg0 = bitcast float *%a0 to i8*
1360 …%call = call <8 x float> @llvm.x86.avx2.gather.d.ps.256(<8 x float> undef, i8* %arg0, <8 x i32> %a…
1397 %arg0 = bitcast i32 *%a0 to i8*
1399 …%call = call <4 x i32> @llvm.x86.avx2.gather.q.d(<4 x i32> undef, i8* %arg0, <2 x i64> %a1, <4 x i…
1416 %arg0 = bitcast <2 x i64> %a0 to <4 x i32>
1419 …%call = call <4 x i32> @llvm.x86.avx2.gather.q.d(<4 x i32> %arg0, i8* %arg1, <2 x i64> %a2, <4 x i…
1441 %arg0 = bitcast i32 *%a0 to i8*
1443 …%call = call <4 x i32> @llvm.x86.avx2.gather.q.d.256(<4 x i32> undef, i8* %arg0, <4 x i64> %a1, <4…
1462 %arg0 = bitcast <2 x i64> %a0 to <4 x i32>
1465 …%call = call <4 x i32> @llvm.x86.avx2.gather.q.d.256(<4 x i32> %arg0, i8* %arg1, <4 x i64> %a2, <4…
1485 %arg0 = bitcast i64 *%a0 to i8*
1486 …%call = call <2 x i64> @llvm.x86.avx2.gather.q.q(<2 x i64> undef, i8* %arg0, <2 x i64> %a1, <2 x i…
1522 %arg0 = bitcast i64 *%a0 to i8*
1523 …%call = call <4 x i64> @llvm.x86.avx2.gather.q.q.256(<4 x i64> undef, i8* %arg0, <4 x i64> %a1, <4…
1559 %arg0 = bitcast double *%a0 to i8*
1563 …%call = call <2 x double> @llvm.x86.avx2.gather.q.pd(<2 x double> undef, i8* %arg0, <2 x i64> %a1,…
1601 %arg0 = bitcast double *%a0 to i8*
1603 …%call = call <4 x double> @llvm.x86.avx2.gather.q.pd.256(<4 x double> undef, i8* %arg0, <4 x i64> …
1639 %arg0 = bitcast float *%a0 to i8*
1643 …%call = call <4 x float> @llvm.x86.avx2.gather.q.ps(<4 x float> undef, i8* %arg0, <2 x i64> %a1, <…
1681 %arg0 = bitcast float *%a0 to i8*
1685 …%call = call <4 x float> @llvm.x86.avx2.gather.q.ps.256(<4 x float> undef, i8* %arg0, <4 x i64> %a…
1750 %arg0 = bitcast <4 x i64> %a0 to <16 x i16>
1752 %res = call <8 x i32> @llvm.x86.avx2.pmadd.wd(<16 x i16> %arg0, <16 x i16> %arg1)
1768 %arg0 = bitcast <4 x i64> %a0 to <32 x i8>
1770 %res = call <16 x i16> @llvm.x86.avx2.pmadd.ub.sw(<32 x i8> %arg0, <32 x i8> %arg1)
1787 %arg0 = bitcast i32* %a0 to i8*
1789 %call = call <4 x i32> @llvm.x86.avx2.maskload.d(i8* %arg0, <4 x i32> %arg1)
1806 %arg0 = bitcast i32* %a0 to i8*
1808 %call = call <8 x i32> @llvm.x86.avx2.maskload.d.256(i8* %arg0, <8 x i32> %arg1)
1825 %arg0 = bitcast i64* %a0 to i8*
1826 %res = call <2 x i64> @llvm.x86.avx2.maskload.q(i8* %arg0, <2 x i64> %a1)
1842 %arg0 = bitcast i64* %a0 to i8*
1843 %res = call <4 x i64> @llvm.x86.avx2.maskload.q.256(i8* %arg0, <4 x i64> %a1)
1859 %arg0 = bitcast float* %a0 to i8*
1862 call void @llvm.x86.avx2.maskstore.d(i8* %arg0, <4 x i32> %arg1, <4 x i32> %arg2)
1880 %arg0 = bitcast float* %a0 to i8*
1883 call void @llvm.x86.avx2.maskstore.d.256(i8* %arg0, <8 x i32> %arg1, <8 x i32> %arg2)
1899 %arg0 = bitcast i64* %a0 to i8*
1900 call void @llvm.x86.avx2.maskstore.q(i8* %arg0, <2 x i64> %a1, <2 x i64> %a2)
1918 %arg0 = bitcast i64* %a0 to i8*
1919 call void @llvm.x86.avx2.maskstore.q.256(i8* %arg0, <4 x i64> %a1, <4 x i64> %a2)
1934 %arg0 = bitcast <4 x i64> %a0 to <32 x i8>
1936 %cmp = icmp sgt <32 x i8> %arg0, %arg1
1937 %sel = select <32 x i1> %cmp, <32 x i8> %arg0, <32 x i8> %arg1
1952 %arg0 = bitcast <4 x i64> %a0 to <16 x i16>
1954 %cmp = icmp sgt <16 x i16> %arg0, %arg1
1955 %sel = select <16 x i1> %cmp, <16 x i16> %arg0, <16 x i16> %arg1
1970 %arg0 = bitcast <4 x i64> %a0 to <8 x i32>
1972 %cmp = icmp sgt <8 x i32> %arg0, %arg1
1973 %sel = select <8 x i1> %cmp, <8 x i32> %arg0, <8 x i32> %arg1
1988 %arg0 = bitcast <4 x i64> %a0 to <32 x i8>
1990 %cmp = icmp ugt <32 x i8> %arg0, %arg1
1991 %sel = select <32 x i1> %cmp, <32 x i8> %arg0, <32 x i8> %arg1
2006 %arg0 = bitcast <4 x i64> %a0 to <16 x i16>
2008 %cmp = icmp ugt <16 x i16> %arg0, %arg1
2009 %sel = select <16 x i1> %cmp, <16 x i16> %arg0, <16 x i16> %arg1
2024 %arg0 = bitcast <4 x i64> %a0 to <8 x i32>
2026 %cmp = icmp ugt <8 x i32> %arg0, %arg1
2027 %sel = select <8 x i1> %cmp, <8 x i32> %arg0, <8 x i32> %arg1
2042 %arg0 = bitcast <4 x i64> %a0 to <32 x i8>
2044 %cmp = icmp slt <32 x i8> %arg0, %arg1
2045 %sel = select <32 x i1> %cmp, <32 x i8> %arg0, <32 x i8> %arg1
2060 %arg0 = bitcast <4 x i64> %a0 to <16 x i16>
2062 %cmp = icmp slt <16 x i16> %arg0, %arg1
2063 %sel = select <16 x i1> %cmp, <16 x i16> %arg0, <16 x i16> %arg1
2078 %arg0 = bitcast <4 x i64> %a0 to <8 x i32>
2080 %cmp = icmp slt <8 x i32> %arg0, %arg1
2081 %sel = select <8 x i1> %cmp, <8 x i32> %arg0, <8 x i32> %arg1
2096 %arg0 = bitcast <4 x i64> %a0 to <32 x i8>
2098 %cmp = icmp ult <32 x i8> %arg0, %arg1
2099 %sel = select <32 x i1> %cmp, <32 x i8> %arg0, <32 x i8> %arg1
2114 %arg0 = bitcast <4 x i64> %a0 to <16 x i16>
2116 %cmp = icmp ult <16 x i16> %arg0, %arg1
2117 %sel = select <16 x i1> %cmp, <16 x i16> %arg0, <16 x i16> %arg1
2132 %arg0 = bitcast <4 x i64> %a0 to <8 x i32>
2134 %cmp = icmp ult <8 x i32> %arg0, %arg1
2135 %sel = select <8 x i1> %cmp, <8 x i32> %arg0, <8 x i32> %arg1
2152 %arg0 = bitcast <4 x i64> %a0 to <32 x i8>
2153 %res = call i32 @llvm.x86.avx2.pmovmskb(<32 x i8> %arg0)
2168 %arg0 = bitcast <4 x i64> %a0 to <32 x i8>
2170 %call = call <16 x i16> @llvm.x86.avx2.mpsadbw(<32 x i8> %arg0, <32 x i8> %arg1, i8 3)
2186 %arg0 = bitcast <4 x i64> %a0 to <8 x i32>
2188 %res = call <4 x i64> @llvm.x86.avx2.pmul.dq(<8 x i32> %arg0, <8 x i32> %arg1)
2203 %arg0 = bitcast <4 x i64> %a0 to <8 x i32>
2205 %res = call <4 x i64> @llvm.x86.avx2.pmulu.dq(<8 x i32> %arg0, <8 x i32> %arg1)
2220 %arg0 = bitcast <4 x i64> %a0 to <16 x i16>
2222 %res = call <16 x i16> @llvm.x86.avx2.pmulh.w(<16 x i16> %arg0, <16 x i16> %arg1)
2238 %arg0 = bitcast <4 x i64> %a0 to <16 x i16>
2240 %res = call <16 x i16> @llvm.x86.avx2.pmulhu.w(<16 x i16> %arg0, <16 x i16> %arg1)
2256 %arg0 = bitcast <4 x i64> %a0 to <16 x i16>
2258 %res = call <16 x i16> @llvm.x86.avx2.pmul.hr.sw(<16 x i16> %arg0, <16 x i16> %arg1)
2274 %arg0 = bitcast <4 x i64> %a0 to <16 x i16>
2276 %res = mul <16 x i16> %arg0, %arg1
2291 %arg0 = bitcast <4 x i64> %a0 to <8 x i32>
2293 %res = mul <8 x i32> %arg0, %arg1
2322 %arg0 = bitcast <4 x i64> %a0 to <16 x i16>
2324 %call = call <32 x i8> @llvm.x86.avx2.packsswb(<16 x i16> %arg0, <16 x i16> %arg1)
2340 %arg0 = bitcast <4 x i64> %a0 to <8 x i32>
2342 %call = call <16 x i16> @llvm.x86.avx2.packssdw(<8 x i32> %arg0, <8 x i32> %arg1)
2358 %arg0 = bitcast <4 x i64> %a0 to <16 x i16>
2360 %call = call <32 x i8> @llvm.x86.avx2.packuswb(<16 x i16> %arg0, <16 x i16> %arg1)
2376 %arg0 = bitcast <4 x i64> %a0 to <8 x i32>
2378 %call = call <16 x i16> @llvm.x86.avx2.packusdw(<8 x i32> %arg0, <8 x i32> %arg1)
2437 %arg0 = bitcast <4 x i64> %a0 to <8 x i32>
2439 %call = call <8 x i32> @llvm.x86.avx2.permd(<8 x i32> %arg0, <8 x i32> %arg1)
2471 %arg0 = bitcast <4 x i64> %a0 to <32 x i8>
2473 %res = call <4 x i64> @llvm.x86.avx2.psad.bw(<32 x i8> %arg0, <32 x i8> %arg1)
2488 %arg0 = bitcast <4 x i64> %a0 to <8 x i32>
2489 …%shuf = shufflevector <8 x i32> %arg0, <8 x i32> undef, <8 x i32> <i32 3, i32 3, i32 0, i32 0, i32…
2504 %arg0 = bitcast <4 x i64> %a0 to <32 x i8>
2506 %shuf = call <32 x i8> @llvm.x86.avx2.pshuf.b(<32 x i8> %arg0, <32 x i8> %arg1)
2522 %arg0 = bitcast <4 x i64> %a0 to <16 x i16>
2523 …%shuf = shufflevector <16 x i16> %arg0, <16 x i16> undef, <16 x i32> <i32 0, i32 1, i32 2, i32 3, …
2538 %arg0 = bitcast <4 x i64> %a0 to <16 x i16>
2539 …%shuf = shufflevector <16 x i16> %arg0, <16 x i16> undef, <16 x i32> <i32 3, i32 0, i32 1, i32 1, …
2554 %arg0 = bitcast <4 x i64> %a0 to <32 x i8>
2556 %call = call <32 x i8> @llvm.x86.avx2.psign.b(<32 x i8> %arg0, <32 x i8> %arg1)
2572 %arg0 = bitcast <4 x i64> %a0 to <16 x i16>
2574 %call = call <16 x i16> @llvm.x86.avx2.psign.w(<16 x i16> %arg0, <16 x i16> %arg1)
2590 %arg0 = bitcast <4 x i64> %a0 to <8 x i32>
2592 %call = call <8 x i32> @llvm.x86.avx2.psign.d(<8 x i32> %arg0, <8 x i32> %arg1)
2608 %arg0 = bitcast <4 x i64> %a0 to <16 x i16>
2610 %res = call <16 x i16> @llvm.x86.avx2.psll.w(<16 x i16> %arg0, <8 x i16> %arg1)
2626 %arg0 = bitcast <4 x i64> %a0 to <8 x i32>
2628 %res = call <8 x i32> @llvm.x86.avx2.psll.d(<8 x i32> %arg0, <4 x i32> %arg1)
2659 %arg0 = bitcast <4 x i64> %a0 to <16 x i16>
2660 %res = call <16 x i16> @llvm.x86.avx2.pslli.w(<16 x i16> %arg0, i32 3)
2676 %arg0 = bitcast <4 x i64> %a0 to <8 x i32>
2677 %res = call <8 x i32> @llvm.x86.avx2.pslli.d(<8 x i32> %arg0, i32 3)
2708 %arg0 = bitcast <4 x i64> %a0 to <32 x i8>
2709 …%shuf = shufflevector <32 x i8> zeroinitializer, <32 x i8> %arg0, <32 x i32> <i32 13, i32 14, i32 …
2724 %arg0 = bitcast <2 x i64> %a0 to <4 x i32>
2726 %res = call <4 x i32> @llvm.x86.avx2.psllv.d(<4 x i32> %arg0, <4 x i32> %arg1)
2742 %arg0 = bitcast <4 x i64> %a0 to <8 x i32>
2744 %res = call <8 x i32> @llvm.x86.avx2.psllv.d.256(<8 x i32> %arg0, <8 x i32> %arg1)
2790 %arg0 = bitcast <4 x i64> %a0 to <16 x i16>
2792 %res = call <16 x i16> @llvm.x86.avx2.psra.w(<16 x i16> %arg0, <8 x i16> %arg1)
2808 %arg0 = bitcast <4 x i64> %a0 to <8 x i32>
2810 %res = call <8 x i32> @llvm.x86.avx2.psra.d(<8 x i32> %arg0, <4 x i32> %arg1)
2826 %arg0 = bitcast <4 x i64> %a0 to <16 x i16>
2827 %res = call <16 x i16> @llvm.x86.avx2.psrai.w(<16 x i16> %arg0, i32 3)
2843 %arg0 = bitcast <4 x i64> %a0 to <8 x i32>
2844 %res = call <8 x i32> @llvm.x86.avx2.psrai.d(<8 x i32> %arg0, i32 3)
2860 %arg0 = bitcast <2 x i64> %a0 to <4 x i32>
2862 %res = call <4 x i32> @llvm.x86.avx2.psrav.d(<4 x i32> %arg0, <4 x i32> %arg1)
2878 %arg0 = bitcast <4 x i64> %a0 to <8 x i32>
2880 %res = call <8 x i32> @llvm.x86.avx2.psrav.d.256(<8 x i32> %arg0, <8 x i32> %arg1)
2896 %arg0 = bitcast <4 x i64> %a0 to <16 x i16>
2898 %res = call <16 x i16> @llvm.x86.avx2.psrl.w(<16 x i16> %arg0, <8 x i16> %arg1)
2914 %arg0 = bitcast <4 x i64> %a0 to <8 x i32>
2916 %res = call <8 x i32> @llvm.x86.avx2.psrl.d(<8 x i32> %arg0, <4 x i32> %arg1)
2947 %arg0 = bitcast <4 x i64> %a0 to <16 x i16>
2948 %res = call <16 x i16> @llvm.x86.avx2.psrli.w(<16 x i16> %arg0, i32 3)
2964 %arg0 = bitcast <4 x i64> %a0 to <8 x i32>
2965 %res = call <8 x i32> @llvm.x86.avx2.psrli.d(<8 x i32> %arg0, i32 3)
2996 %arg0 = bitcast <4 x i64> %a0 to <32 x i8>
2997 …%shuf = shufflevector <32 x i8> %arg0, <32 x i8> zeroinitializer, <32 x i32> <i32 3, i32 4, i32 5,…
3012 %arg0 = bitcast <2 x i64> %a0 to <4 x i32>
3014 %res = call <4 x i32> @llvm.x86.avx2.psrlv.d(<4 x i32> %arg0, <4 x i32> %arg1)
3030 %arg0 = bitcast <4 x i64> %a0 to <8 x i32>
3032 %res = call <8 x i32> @llvm.x86.avx2.psrlv.d.256(<8 x i32> %arg0, <8 x i32> %arg1)
3079 %arg0 = bitcast <4 x i64> *%a0 to i8*
3080 %res = call <4 x i64> @llvm.x86.avx2.movntdqa(i8* %arg0)
3095 %arg0 = bitcast <4 x i64> %a0 to <32 x i8>
3097 %res = sub <32 x i8> %arg0, %arg1
3112 %arg0 = bitcast <4 x i64> %a0 to <16 x i16>
3114 %res = sub <16 x i16> %arg0, %arg1
3129 %arg0 = bitcast <4 x i64> %a0 to <8 x i32>
3131 %res = sub <8 x i32> %arg0, %arg1
3160 %arg0 = bitcast <4 x i64> %a0 to <32 x i8>
3162 %res = call <32 x i8> @llvm.x86.avx2.psubs.b(<32 x i8> %arg0, <32 x i8> %arg1)
3178 %arg0 = bitcast <4 x i64> %a0 to <16 x i16>
3180 %res = call <16 x i16> @llvm.x86.avx2.psubs.w(<16 x i16> %arg0, <16 x i16> %arg1)
3196 %arg0 = bitcast <4 x i64> %a0 to <32 x i8>
3198 %res = call <32 x i8> @llvm.x86.avx2.psubus.b(<32 x i8> %arg0, <32 x i8> %arg1)
3214 %arg0 = bitcast <4 x i64> %a0 to <16 x i16>
3216 %res = call <16 x i16> @llvm.x86.avx2.psubus.w(<16 x i16> %arg0, <16 x i16> %arg1)
3232 %arg0 = bitcast <4 x i64> %a0 to <32 x i8>
3234 …%res = shufflevector <32 x i8> %arg0, <32 x i8> %arg1, <32 x i32> <i32 8, i32 40, i32 9, i32 41, i…
3249 %arg0 = bitcast <4 x i64> %a0 to <16 x i16>
3251 …%res = shufflevector <16 x i16> %arg0, <16 x i16> %arg1, <16 x i32> <i32 4, i32 20, i32 5, i32 21,…
3266 %arg0 = bitcast <4 x i64> %a0 to <8 x i32>
3268 …%res = shufflevector <8 x i32> %arg0, <8 x i32> %arg1, <8 x i32> <i32 2, i32 10, i32 3, i32 11, i3…
3297 %arg0 = bitcast <4 x i64> %a0 to <32 x i8>
3299 …%res = shufflevector <32 x i8> %arg0, <32 x i8> %arg1, <32 x i32> <i32 0, i32 32, i32 1, i32 33, i…
3314 %arg0 = bitcast <4 x i64> %a0 to <16 x i16>
3316 …%res = shufflevector <16 x i16> %arg0, <16 x i16> %arg1, <16 x i32> <i32 0, i32 16, i32 1, i32 17,…
3331 %arg0 = bitcast <4 x i64> %a0 to <8 x i32>
3333 …%res = shufflevector <8 x i32> %arg0, <8 x i32> %arg1, <8 x i32> <i32 0, i32 8, i32 1, i32 9, i32 …