/external/llvm/test/CodeGen/AMDGPU/ |
D | saddo.ll | 5 declare { i32, i1 } @llvm.sadd.with.overflow.i32(i32, i32) nounwind readnone 6 declare { i64, i1 } @llvm.sadd.with.overflow.i64(i64, i64) nounwind readnone 10 %sadd = call { i64, i1 } @llvm.sadd.with.overflow.i64(i64 %a, i64 %b) nounwind 11 %val = extractvalue { i64, i1 } %sadd, 0 12 %carry = extractvalue { i64, i1 } %sadd, 1 21 %sadd = call { i32, i1 } @llvm.sadd.with.overflow.i32(i32 %a, i32 %b) nounwind 22 %val = extractvalue { i32, i1 } %sadd, 0 23 %carry = extractvalue { i32, i1 } %sadd, 1 33 %sadd = call { i32, i1 } @llvm.sadd.with.overflow.i32(i32 %a, i32 %b) nounwind 34 %val = extractvalue { i32, i1 } %sadd, 0 [all …]
|
/external/llvm-project/llvm/test/Transforms/InstCombine/ |
D | sadd-with-overflow.ll | 4 declare { <2 x i32>, <2 x i1> } @llvm.sadd.with.overflow.v2i32(<2 x i32>, <2 x i32>) 6 declare { i32, i1 } @llvm.sadd.with.overflow.i32(i32, i32) 8 declare { i8, i1 } @llvm.sadd.with.overflow.i8(i8, i8) 12 ; CHECK-NEXT: [[TMP1:%.*]] = call { i32, i1 } @llvm.sadd.with.overflow.i32(i32 [[X:%.*]], i32 20) 16 %b = tail call { i32, i1 } @llvm.sadd.with.overflow.i32(i32 %a, i32 13) 27 %b = tail call { i32, i1 } @llvm.sadd.with.overflow.i32(i32 %a, i32 -7) 33 ; CHECK-NEXT: [[TMP1:%.*]] = call { i8, i1 } @llvm.sadd.with.overflow.i8(i8 [[X:%.*]], i8 127) 37 %b = tail call { i8, i1 } @llvm.sadd.with.overflow.i8(i8 %a, i8 27) 44 ; CHECK-NEXT: [[B:%.*]] = tail call { i8, i1 } @llvm.sadd.with.overflow.i8(i8 [[A]], i8 28) 48 %b = tail call { i8, i1 } @llvm.sadd.with.overflow.i8(i8 %a, i8 28) [all …]
|
/external/llvm-project/llvm/test/Analysis/CostModel/X86/ |
D | arith-ssat.ll | 17 declare i64 @llvm.sadd.sat.i64(i64, i64) 18 declare <2 x i64> @llvm.sadd.sat.v2i64(<2 x i64>, <2 x i64>) 19 declare <4 x i64> @llvm.sadd.sat.v4i64(<4 x i64>, <4 x i64>) 20 declare <8 x i64> @llvm.sadd.sat.v8i64(<8 x i64>, <8 x i64>) 22 declare i32 @llvm.sadd.sat.i32(i32, i32) 23 declare <4 x i32> @llvm.sadd.sat.v4i32(<4 x i32>, <4 x i32>) 24 declare <8 x i32> @llvm.sadd.sat.v8i32(<8 x i32>, <8 x i32>) 25 declare <16 x i32> @llvm.sadd.sat.v16i32(<16 x i32>, <16 x i32>) 27 declare i16 @llvm.sadd.sat.i16(i16, i16) 28 declare <8 x i16> @llvm.sadd.sat.v8i16(<8 x i16>, <8 x i16>) [all …]
|
/external/llvm-project/llvm/test/Analysis/CostModel/ARM/ |
D | arith-ssat.ll | 9 declare i64 @llvm.sadd.sat.i64(i64, i64) 10 declare <2 x i64> @llvm.sadd.sat.v2i64(<2 x i64>, <2 x i64>) 11 declare <4 x i64> @llvm.sadd.sat.v4i64(<4 x i64>, <4 x i64>) 12 declare <8 x i64> @llvm.sadd.sat.v8i64(<8 x i64>, <8 x i64>) 14 declare i32 @llvm.sadd.sat.i32(i32, i32) 15 declare <4 x i32> @llvm.sadd.sat.v4i32(<4 x i32>, <4 x i32>) 16 declare <8 x i32> @llvm.sadd.sat.v8i32(<8 x i32>, <8 x i32>) 17 declare <16 x i32> @llvm.sadd.sat.v16i32(<16 x i32>, <16 x i32>) 19 declare i16 @llvm.sadd.sat.i16(i16, i16) 20 declare <8 x i16> @llvm.sadd.sat.v8i16(<8 x i16>, <8 x i16>) [all …]
|
/external/llvm/test/CodeGen/ARM/ |
D | intrinsics-overflow.ll | 4 %sadd = tail call { i32, i1 } @llvm.uadd.with.overflow.i32(i32 %a, i32 %b) 5 %1 = extractvalue { i32, i1 } %sadd, 1 18 %sadd = tail call { i32, i1 } @llvm.sadd.with.overflow.i32(i32 %a, i32 %b) 19 %1 = extractvalue { i32, i1 } %sadd, 1 31 %sadd = tail call { i32, i1 } @llvm.usub.with.overflow.i32(i32 %a, i32 %b) 32 %1 = extractvalue { i32, i1 } %sadd, 1 43 %sadd = tail call { i32, i1 } @llvm.ssub.with.overflow.i32(i32 %a, i32 %b) 44 %1 = extractvalue { i32, i1 } %sadd, 1 55 declare { i32, i1 } @llvm.sadd.with.overflow.i32(i32, i32) #2
|
/external/llvm-project/llvm/test/CodeGen/AArch64/ |
D | sadd_sat_vec.ll | 4 declare <1 x i8> @llvm.sadd.sat.v1i8(<1 x i8>, <1 x i8>) 5 declare <2 x i8> @llvm.sadd.sat.v2i8(<2 x i8>, <2 x i8>) 6 declare <4 x i8> @llvm.sadd.sat.v4i8(<4 x i8>, <4 x i8>) 7 declare <8 x i8> @llvm.sadd.sat.v8i8(<8 x i8>, <8 x i8>) 8 declare <12 x i8> @llvm.sadd.sat.v12i8(<12 x i8>, <12 x i8>) 9 declare <16 x i8> @llvm.sadd.sat.v16i8(<16 x i8>, <16 x i8>) 10 declare <32 x i8> @llvm.sadd.sat.v32i8(<32 x i8>, <32 x i8>) 11 declare <64 x i8> @llvm.sadd.sat.v64i8(<64 x i8>, <64 x i8>) 13 declare <1 x i16> @llvm.sadd.sat.v1i16(<1 x i16>, <1 x i16>) 14 declare <2 x i16> @llvm.sadd.sat.v2i16(<2 x i16>, <2 x i16>) [all …]
|
D | sadd_sat.ll | 4 declare i4 @llvm.sadd.sat.i4(i4, i4) 5 declare i8 @llvm.sadd.sat.i8(i8, i8) 6 declare i16 @llvm.sadd.sat.i16(i16, i16) 7 declare i32 @llvm.sadd.sat.i32(i32, i32) 8 declare i64 @llvm.sadd.sat.i64(i64, i64) 9 declare <4 x i32> @llvm.sadd.sat.v4i32(<4 x i32>, <4 x i32>) 21 %tmp = call i32 @llvm.sadd.sat.i32(i32 %x, i32 %y); 35 %tmp = call i64 @llvm.sadd.sat.i64(i64 %x, i64 %y); 51 %tmp = call i16 @llvm.sadd.sat.i16(i16 %x, i16 %y); 67 %tmp = call i8 @llvm.sadd.sat.i8(i8 %x, i8 %y); [all …]
|
D | sadd_sat_plus.ll | 4 declare i4 @llvm.sadd.sat.i4(i4, i4) 5 declare i8 @llvm.sadd.sat.i8(i8, i8) 6 declare i16 @llvm.sadd.sat.i16(i16, i16) 7 declare i32 @llvm.sadd.sat.i32(i32, i32) 8 declare i64 @llvm.sadd.sat.i64(i64, i64) 22 %tmp = call i32 @llvm.sadd.sat.i32(i32 %x, i32 %a) 37 %tmp = call i64 @llvm.sadd.sat.i64(i64 %x, i64 %z) 55 %tmp = call i16 @llvm.sadd.sat.i16(i16 %x, i16 %a) 73 %tmp = call i8 @llvm.sadd.sat.i8(i8 %x, i8 %a) 92 %tmp = call i4 @llvm.sadd.sat.i4(i4 %x, i4 %a)
|
/external/llvm-project/llvm/test/CodeGen/SystemZ/ |
D | int-sadd-09.ll | 16 %t = call {i64, i1} @llvm.sadd.with.overflow.i64(i64 %a, i64 1) 32 %t = call {i64, i1} @llvm.sadd.with.overflow.i64(i64 %a, i64 127) 50 %t = call {i64, i1} @llvm.sadd.with.overflow.i64(i64 %a, i64 128) 66 %t = call {i64, i1} @llvm.sadd.with.overflow.i64(i64 %a, i64 -128) 84 %t = call {i64, i1} @llvm.sadd.with.overflow.i64(i64 %a, i64 -129) 101 %t = call {i64, i1} @llvm.sadd.with.overflow.i64(i64 %a, i64 1) 120 %t = call {i64, i1} @llvm.sadd.with.overflow.i64(i64 %a, i64 1) 137 %t = call {i64, i1} @llvm.sadd.with.overflow.i64(i64 %a, i64 1) 156 %t = call {i64, i1} @llvm.sadd.with.overflow.i64(i64 %a, i64 1) 176 %t = call {i64, i1} @llvm.sadd.with.overflow.i64(i64 %a, i64 1) [all …]
|
D | int-sadd-08.ll | 16 %t = call {i32, i1} @llvm.sadd.with.overflow.i32(i32 %a, i32 1) 32 %t = call {i32, i1} @llvm.sadd.with.overflow.i32(i32 %a, i32 127) 50 %t = call {i32, i1} @llvm.sadd.with.overflow.i32(i32 %a, i32 128) 66 %t = call {i32, i1} @llvm.sadd.with.overflow.i32(i32 %a, i32 -128) 84 %t = call {i32, i1} @llvm.sadd.with.overflow.i32(i32 %a, i32 -129) 101 %t = call {i32, i1} @llvm.sadd.with.overflow.i32(i32 %a, i32 1) 120 %t = call {i32, i1} @llvm.sadd.with.overflow.i32(i32 %a, i32 1) 137 %t = call {i32, i1} @llvm.sadd.with.overflow.i32(i32 %a, i32 1) 156 %t = call {i32, i1} @llvm.sadd.with.overflow.i32(i32 %a, i32 1) 176 %t = call {i32, i1} @llvm.sadd.with.overflow.i32(i32 %a, i32 1) [all …]
|
D | int-sadd-03.ll | 16 %t = call {i64, i1} @llvm.sadd.with.overflow.i64(i64 %a, i64 %b) 30 %t = call {i64, i1} @llvm.sadd.with.overflow.i64(i64 %a, i64 %b) 51 %t = call {i64, i1} @llvm.sadd.with.overflow.i64(i64 %a, i64 %b) 75 %t = call {i64, i1} @llvm.sadd.with.overflow.i64(i64 %a, i64 %b) 93 %t = call {i64, i1} @llvm.sadd.with.overflow.i64(i64 %a, i64 %b) 113 %t = call {i64, i1} @llvm.sadd.with.overflow.i64(i64 %a, i64 %b) 131 %t = call {i64, i1} @llvm.sadd.with.overflow.i64(i64 %a, i64 %b) 149 %t = call {i64, i1} @llvm.sadd.with.overflow.i64(i64 %a, i64 %b) 169 %t = call {i64, i1} @llvm.sadd.with.overflow.i64(i64 %a, i64 %b) 189 %t = call {i64, i1} @llvm.sadd.with.overflow.i64(i64 %a, i64 %b) [all …]
|
D | int-sadd-01.ll | 16 %t = call {i32, i1} @llvm.sadd.with.overflow.i32(i32 %a, i32 %b) 30 %t = call {i32, i1} @llvm.sadd.with.overflow.i32(i32 %a, i32 %b) 51 %t = call {i32, i1} @llvm.sadd.with.overflow.i32(i32 %a, i32 %b) 75 %t = call {i32, i1} @llvm.sadd.with.overflow.i32(i32 %a, i32 %b) 93 %t = call {i32, i1} @llvm.sadd.with.overflow.i32(i32 %a, i32 %b) 111 %t = call {i32, i1} @llvm.sadd.with.overflow.i32(i32 %a, i32 %b) 129 %t = call {i32, i1} @llvm.sadd.with.overflow.i32(i32 %a, i32 %b) 149 %t = call {i32, i1} @llvm.sadd.with.overflow.i32(i32 %a, i32 %b) 167 %t = call {i32, i1} @llvm.sadd.with.overflow.i32(i32 %a, i32 %b) 185 %t = call {i32, i1} @llvm.sadd.with.overflow.i32(i32 %a, i32 %b) [all …]
|
D | int-sadd-04.ll | 17 %t = call {i64, i1} @llvm.sadd.with.overflow.i64(i64 %a, i64 %bext) 32 %t = call {i64, i1} @llvm.sadd.with.overflow.i64(i64 %a, i64 %bext) 54 %t = call {i64, i1} @llvm.sadd.with.overflow.i64(i64 %a, i64 %bext) 79 %t = call {i64, i1} @llvm.sadd.with.overflow.i64(i64 %a, i64 %bext) 98 %t = call {i64, i1} @llvm.sadd.with.overflow.i64(i64 %a, i64 %bext) 119 %t = call {i64, i1} @llvm.sadd.with.overflow.i64(i64 %a, i64 %bext) 138 %t = call {i64, i1} @llvm.sadd.with.overflow.i64(i64 %a, i64 %bext) 157 %t = call {i64, i1} @llvm.sadd.with.overflow.i64(i64 %a, i64 %bext) 178 %t = call {i64, i1} @llvm.sadd.with.overflow.i64(i64 %a, i64 %bext) 199 %t = call {i64, i1} @llvm.sadd.with.overflow.i64(i64 %a, i64 %bext) [all …]
|
D | int-sadd-07.ll | 17 %t = call {i64, i1} @llvm.sadd.with.overflow.i64(i64 %a, i64 1) 34 %t = call {i64, i1} @llvm.sadd.with.overflow.i64(i64 %a, i64 32767) 50 %t = call {i64, i1} @llvm.sadd.with.overflow.i64(i64 %a, i64 32768) 66 %t = call {i64, i1} @llvm.sadd.with.overflow.i64(i64 %a, i64 2147483647) 83 %t = call {i64, i1} @llvm.sadd.with.overflow.i64(i64 %a, i64 2147483648) 99 %t = call {i64, i1} @llvm.sadd.with.overflow.i64(i64 %a, i64 -1) 115 %t = call {i64, i1} @llvm.sadd.with.overflow.i64(i64 %a, i64 -32768) 131 %t = call {i64, i1} @llvm.sadd.with.overflow.i64(i64 %a, i64 -32769) 147 %t = call {i64, i1} @llvm.sadd.with.overflow.i64(i64 %a, i64 -2147483648) 164 %t = call {i64, i1} @llvm.sadd.with.overflow.i64(i64 %a, i64 -2147483649) [all …]
|
D | int-sadd-06.ll | 17 %t = call {i32, i1} @llvm.sadd.with.overflow.i32(i32 %a, i32 1) 33 %t = call {i32, i1} @llvm.sadd.with.overflow.i32(i32 %a, i32 32767) 49 %t = call {i32, i1} @llvm.sadd.with.overflow.i32(i32 %a, i32 32768) 65 %t = call {i32, i1} @llvm.sadd.with.overflow.i32(i32 %a, i32 2147483647) 81 %t = call {i32, i1} @llvm.sadd.with.overflow.i32(i32 %a, i32 2147483648) 97 %t = call {i32, i1} @llvm.sadd.with.overflow.i32(i32 %a, i32 -1) 113 %t = call {i32, i1} @llvm.sadd.with.overflow.i32(i32 %a, i32 -32768) 129 %t = call {i32, i1} @llvm.sadd.with.overflow.i32(i32 %a, i32 -32769) 145 %t = call {i32, i1} @llvm.sadd.with.overflow.i32(i32 %a, i32 -2147483648) 161 %t = call {i32, i1} @llvm.sadd.with.overflow.i32(i32 %a, i32 -2147483649) [all …]
|
/external/llvm-project/llvm/test/Transforms/SLPVectorizer/X86/ |
D | arith-add-ssat.ll | 23 declare i64 @llvm.sadd.sat.i64(i64, i64) 24 declare i32 @llvm.sadd.sat.i32(i32, i32) 25 declare i16 @llvm.sadd.sat.i16(i16, i16) 26 declare i8 @llvm.sadd.sat.i8 (i8 , i8 ) 46 ; SSE-NEXT: [[R0:%.*]] = call i64 @llvm.sadd.sat.i64(i64 [[A0]], i64 [[B0]]) 47 ; SSE-NEXT: [[R1:%.*]] = call i64 @llvm.sadd.sat.i64(i64 [[A1]], i64 [[B1]]) 48 ; SSE-NEXT: [[R2:%.*]] = call i64 @llvm.sadd.sat.i64(i64 [[A2]], i64 [[B2]]) 49 ; SSE-NEXT: [[R3:%.*]] = call i64 @llvm.sadd.sat.i64(i64 [[A3]], i64 [[B3]]) 50 ; SSE-NEXT: [[R4:%.*]] = call i64 @llvm.sadd.sat.i64(i64 [[A4]], i64 [[B4]]) 51 ; SSE-NEXT: [[R5:%.*]] = call i64 @llvm.sadd.sat.i64(i64 [[A5]], i64 [[B5]]) [all …]
|
D | arith-add-saddo.ll | 23 declare {i64, i1} @llvm.sadd.with.overflow.i64(i64, i64) 24 declare {i32, i1} @llvm.sadd.with.overflow.i32(i32, i32) 25 declare {i16, i1} @llvm.sadd.with.overflow.i16(i16, i16) 26 declare {i8 , i1} @llvm.sadd.with.overflow.i8 (i8 , i8 ) 46 ; CHECK-NEXT: [[C0:%.*]] = call { i64, i1 } @llvm.sadd.with.overflow.i64(i64 [[A0]], i64 [[B0]]) 47 ; CHECK-NEXT: [[C1:%.*]] = call { i64, i1 } @llvm.sadd.with.overflow.i64(i64 [[A1]], i64 [[B1]]) 48 ; CHECK-NEXT: [[C2:%.*]] = call { i64, i1 } @llvm.sadd.with.overflow.i64(i64 [[A2]], i64 [[B2]]) 49 ; CHECK-NEXT: [[C3:%.*]] = call { i64, i1 } @llvm.sadd.with.overflow.i64(i64 [[A3]], i64 [[B3]]) 50 ; CHECK-NEXT: [[C4:%.*]] = call { i64, i1 } @llvm.sadd.with.overflow.i64(i64 [[A4]], i64 [[B4]]) 51 ; CHECK-NEXT: [[C5:%.*]] = call { i64, i1 } @llvm.sadd.with.overflow.i64(i64 [[A5]], i64 [[B5]]) [all …]
|
/external/llvm-project/llvm/test/CodeGen/Generic/ |
D | overflow.ll | 8 %sadd = tail call { i8, i1 } @llvm.sadd.with.overflow.i8(i8 %a, i8 %b) 9 %cmp = extractvalue { i8, i1 } %sadd, 1 10 %sadd.result = extractvalue { i8, i1 } %sadd, 0 11 %X = select i1 %cmp, i8 %sadd.result, i8 42 15 declare { i8, i1 } @llvm.sadd.with.overflow.i8(i8, i8) nounwind readnone 19 %sadd = tail call { i16, i1 } @llvm.sadd.with.overflow.i16(i16 %a, i16 %b) 20 %cmp = extractvalue { i16, i1 } %sadd, 1 21 %sadd.result = extractvalue { i16, i1 } %sadd, 0 22 %X = select i1 %cmp, i16 %sadd.result, i16 42 26 declare { i16, i1 } @llvm.sadd.with.overflow.i16(i16, i16) nounwind readnone [all …]
|
/external/llvm/test/CodeGen/Generic/ |
D | overflow.ll | 8 %sadd = tail call { i8, i1 } @llvm.sadd.with.overflow.i8(i8 %a, i8 %b) 9 %cmp = extractvalue { i8, i1 } %sadd, 1 10 %sadd.result = extractvalue { i8, i1 } %sadd, 0 11 %X = select i1 %cmp, i8 %sadd.result, i8 42 15 declare { i8, i1 } @llvm.sadd.with.overflow.i8(i8, i8) nounwind readnone 19 %sadd = tail call { i16, i1 } @llvm.sadd.with.overflow.i16(i16 %a, i16 %b) 20 %cmp = extractvalue { i16, i1 } %sadd, 1 21 %sadd.result = extractvalue { i16, i1 } %sadd, 0 22 %X = select i1 %cmp, i16 %sadd.result, i16 42 26 declare { i16, i1 } @llvm.sadd.with.overflow.i16(i16, i16) nounwind readnone [all …]
|
/external/llvm-project/llvm/test/CodeGen/X86/ |
D | combine-add-ssat.ll | 10 declare i32 @llvm.sadd.sat.i32 (i32, i32) 11 declare i64 @llvm.sadd.sat.i64 (i64, i64) 12 declare <8 x i16> @llvm.sadd.sat.v8i16(<8 x i16>, <8 x i16>) 20 %res = call i32 @llvm.sadd.sat.i32(i32 %a0, i32 undef) 34 %res = call <8 x i16> @llvm.sadd.sat.v8i16(<8 x i16> undef, <8 x i16> %a0) 44 %res = call i32 @llvm.sadd.sat.i32(i32 2147483647, i32 100) 58 …%res = call <8 x i16> @llvm.sadd.sat.v8i16(<8 x i16> <i16 0, i16 1, i16 255, i16 65535, i16 -1, i1… 72 …%res = call <8 x i16> @llvm.sadd.sat.v8i16(<8 x i16> <i16 undef, i16 1, i16 undef, i16 65535, i16 … 88 %res = call i32 @llvm.sadd.sat.i32(i32 1, i32 %a0) 102 …%res = call <8 x i16> @llvm.sadd.sat.v8i16(<8 x i16> <i16 1, i16 1, i16 1, i16 1, i16 1, i16 1, i1… [all …]
|
/external/llvm-project/llvm/test/CodeGen/ARM/ |
D | intrinsics-overflow.ll | 6 %sadd = tail call { i32, i1 } @llvm.uadd.with.overflow.i32(i32 %a, i32 %b) 7 %1 = extractvalue { i32, i1 } %sadd, 1 29 %sadd = tail call { i32, i1 } @llvm.sadd.with.overflow.i32(i32 %a, i32 %b) 30 %1 = extractvalue { i32, i1 } %sadd, 1 51 %sadd = tail call { i32, i1 } @llvm.usub.with.overflow.i32(i32 %a, i32 %b) 52 %1 = extractvalue { i32, i1 } %sadd, 1 80 %sadd = tail call { i32, i1 } @llvm.ssub.with.overflow.i32(i32 %a, i32 %b) 81 %1 = extractvalue { i32, i1 } %sadd, 1 102 declare { i32, i1 } @llvm.sadd.with.overflow.i32(i32, i32) #2
|
D | unschedule-first-call.ll | 20 %13 = call { i64, i1 } @llvm.sadd.with.overflow.i64(i64 undef, i64 undef) 22 %15 = call { i64, i1 } @llvm.sadd.with.overflow.i64(i64 %14, i64 1) 29 %22 = call { i64, i1 } @llvm.sadd.with.overflow.i64(i64 %21, i64 0) 33 %26 = call { i64, i1 } @llvm.sadd.with.overflow.i64(i64 %25, i64 0) 41 %34 = call { i64, i1 } @llvm.sadd.with.overflow.i64(i64 0, i64 undef) 43 %36 = call { i64, i1 } @llvm.sadd.with.overflow.i64(i64 %35, i64 1) 50 %43 = call { i64, i1 } @llvm.sadd.with.overflow.i64(i64 %42, i64 0) 55 %48 = call { i64, i1 } @llvm.sadd.with.overflow.i64(i64 %47, i64 0) 73 %66 = call { i64, i1 } @llvm.sadd.with.overflow.i64(i64 %63, i64 %65) 78 %71 = call { i64, i1 } @llvm.sadd.with.overflow.i64(i64 %70, i64 0) [all …]
|
D | vqadd.ll | 8 %tmp3 = call <8 x i8> @llvm.sadd.sat.v8i8(<8 x i8> %tmp1, <8 x i8> %tmp2) 17 %tmp3 = call <4 x i16> @llvm.sadd.sat.v4i16(<4 x i16> %tmp1, <4 x i16> %tmp2) 26 %tmp3 = call <2 x i32> @llvm.sadd.sat.v2i32(<2 x i32> %tmp1, <2 x i32> %tmp2) 35 %tmp3 = call <1 x i64> @llvm.sadd.sat.v1i64(<1 x i64> %tmp1, <1 x i64> %tmp2) 80 %tmp3 = call <16 x i8> @llvm.sadd.sat.v16i8(<16 x i8> %tmp1, <16 x i8> %tmp2) 89 %tmp3 = call <8 x i16> @llvm.sadd.sat.v8i16(<8 x i16> %tmp1, <8 x i16> %tmp2) 98 %tmp3 = call <4 x i32> @llvm.sadd.sat.v4i32(<4 x i32> %tmp1, <4 x i32> %tmp2) 107 %tmp3 = call <2 x i64> @llvm.sadd.sat.v2i64(<2 x i64> %tmp1, <2 x i64> %tmp2) 147 declare <8 x i8> @llvm.sadd.sat.v8i8(<8 x i8>, <8 x i8>) nounwind readnone 148 declare <4 x i16> @llvm.sadd.sat.v4i16(<4 x i16>, <4 x i16>) nounwind readnone [all …]
|
/external/llvm-project/llvm/test/Transforms/NewGVN/ |
D | 2011-07-07-MatchIntrinsicExtract.ll | 47 %sadd = tail call %0 @llvm.sadd.with.overflow.i64(i64 %a, i64 %b) 48 %sadd.0 = extractvalue %0 %sadd, 0 50 %add2 = add i64 %add1, %sadd.0 88 declare %0 @llvm.sadd.with.overflow.i64(i64, i64) nounwind readnone
|
/external/llvm-project/llvm/test/CodeGen/AMDGPU/ |
D | saddo.ll | 7 declare { i32, i1 } @llvm.sadd.with.overflow.i32(i32, i32) nounwind readnone 8 declare { i64, i1 } @llvm.sadd.with.overflow.i64(i64, i64) nounwind readnone 11 declare { <2 x i32>, <2 x i1> } @llvm.sadd.with.overflow.v2i32(<2 x i32>, <2 x i32>) nounwind readn… 77 %sadd = call { i64, i1 } @llvm.sadd.with.overflow.i64(i64 %a, i64 %b) nounwind 78 %val = extractvalue { i64, i1 } %sadd, 0 79 %carry = extractvalue { i64, i1 } %sadd, 1 147 %sadd = call { i32, i1 } @llvm.sadd.with.overflow.i32(i32 %a, i32 %b) nounwind 148 %val = extractvalue { i32, i1 } %sadd, 0 149 %carry = extractvalue { i32, i1 } %sadd, 1 227 %sadd = call { i32, i1 } @llvm.sadd.with.overflow.i32(i32 %a, i32 %b) nounwind [all …]
|