1; NOTE: Assertions have been autogenerated by utils/update_test_checks.py 2; RUN: opt < %s -instsimplify -S | FileCheck %s 3; RUN: opt < %s -passes=instsimplify -S | FileCheck %s 4 5declare {i8, i1} @llvm.uadd.with.overflow.i8(i8 %a, i8 %b) 6declare {i8, i1} @llvm.sadd.with.overflow.i8(i8 %a, i8 %b) 7declare {i8, i1} @llvm.usub.with.overflow.i8(i8 %a, i8 %b) 8declare {i8, i1} @llvm.ssub.with.overflow.i8(i8 %a, i8 %b) 9declare {i8, i1} @llvm.umul.with.overflow.i8(i8 %a, i8 %b) 10declare {i8, i1} @llvm.smul.with.overflow.i8(i8 %a, i8 %b) 11 12define i1 @test_uadd1() { 13; CHECK-LABEL: @test_uadd1( 14; CHECK-NEXT: ret i1 true 15; 16 %x = call {i8, i1} @llvm.uadd.with.overflow.i8(i8 254, i8 3) 17 %overflow = extractvalue {i8, i1} %x, 1 18 ret i1 %overflow 19} 20 21define i8 @test_uadd2() { 22; CHECK-LABEL: @test_uadd2( 23; CHECK-NEXT: ret i8 42 24; 25 %x = call {i8, i1} @llvm.uadd.with.overflow.i8(i8 254, i8 44) 26 %result = extractvalue {i8, i1} %x, 0 27 ret i8 %result 28} 29 30define {i8, i1} @test_uadd3(i8 %v) { 31; CHECK-LABEL: @test_uadd3( 32; CHECK-NEXT: ret { i8, i1 } undef 33; 34 %result = call {i8, i1} @llvm.uadd.with.overflow.i8(i8 %v, i8 undef) 35 ret {i8, i1} %result 36} 37 38define {i8, i1} @test_uadd4(i8 %v) { 39; CHECK-LABEL: @test_uadd4( 40; CHECK-NEXT: ret { i8, i1 } undef 41; 42 %result = call {i8, i1} @llvm.uadd.with.overflow.i8(i8 undef, i8 %v) 43 ret {i8, i1} %result 44} 45 46define i1 @test_sadd1() { 47; CHECK-LABEL: @test_sadd1( 48; CHECK-NEXT: ret i1 true 49; 50 %x = call {i8, i1} @llvm.sadd.with.overflow.i8(i8 126, i8 3) 51 %overflow = extractvalue {i8, i1} %x, 1 52 ret i1 %overflow 53} 54 55define i8 @test_sadd2() { 56; CHECK-LABEL: @test_sadd2( 57; CHECK-NEXT: ret i8 -86 58; 59 %x = call {i8, i1} @llvm.sadd.with.overflow.i8(i8 126, i8 44) 60 %result = extractvalue {i8, i1} %x, 0 61 ret i8 %result 62} 63 64define {i8, i1} @test_sadd3(i8 %v) { 65; CHECK-LABEL: @test_sadd3( 66; CHECK-NEXT: ret { i8, i1 } undef 67; 68 %result = call {i8, i1} @llvm.sadd.with.overflow.i8(i8 %v, i8 undef) 69 ret {i8, i1} %result 70} 71 72define {i8, i1} @test_sadd4(i8 %v) { 73; CHECK-LABEL: @test_sadd4( 74; CHECK-NEXT: ret { i8, i1 } undef 75; 76 %result = call {i8, i1} @llvm.sadd.with.overflow.i8(i8 undef, i8 %v) 77 ret {i8, i1} %result 78} 79 80define {i8, i1} @test_usub1(i8 %V) { 81; CHECK-LABEL: @test_usub1( 82; CHECK-NEXT: ret { i8, i1 } zeroinitializer 83; 84 %x = call {i8, i1} @llvm.usub.with.overflow.i8(i8 %V, i8 %V) 85 ret {i8, i1} %x 86} 87 88define {i8, i1} @test_usub2(i8 %V) { 89; CHECK-LABEL: @test_usub2( 90; CHECK-NEXT: ret { i8, i1 } undef 91; 92 %x = call {i8, i1} @llvm.usub.with.overflow.i8(i8 %V, i8 undef) 93 ret {i8, i1} %x 94} 95 96define {i8, i1} @test_usub3(i8 %V) { 97; CHECK-LABEL: @test_usub3( 98; CHECK-NEXT: ret { i8, i1 } undef 99; 100 %x = call {i8, i1} @llvm.usub.with.overflow.i8(i8 undef, i8 %V) 101 ret {i8, i1} %x 102} 103 104define {i8, i1} @test_ssub1(i8 %V) { 105; CHECK-LABEL: @test_ssub1( 106; CHECK-NEXT: ret { i8, i1 } zeroinitializer 107; 108 %x = call {i8, i1} @llvm.ssub.with.overflow.i8(i8 %V, i8 %V) 109 ret {i8, i1} %x 110} 111 112define {i8, i1} @test_ssub2(i8 %V) { 113; CHECK-LABEL: @test_ssub2( 114; CHECK-NEXT: ret { i8, i1 } undef 115; 116 %x = call {i8, i1} @llvm.ssub.with.overflow.i8(i8 %V, i8 undef) 117 ret {i8, i1} %x 118} 119 120define {i8, i1} @test_ssub3(i8 %V) { 121; CHECK-LABEL: @test_ssub3( 122; CHECK-NEXT: ret { i8, i1 } undef 123; 124 %x = call {i8, i1} @llvm.ssub.with.overflow.i8(i8 undef, i8 %V) 125 ret {i8, i1} %x 126} 127 128define {i8, i1} @test_umul1(i8 %V) { 129; CHECK-LABEL: @test_umul1( 130; CHECK-NEXT: ret { i8, i1 } zeroinitializer 131; 132 %x = call {i8, i1} @llvm.umul.with.overflow.i8(i8 %V, i8 0) 133 ret {i8, i1} %x 134} 135 136define {i8, i1} @test_umul2(i8 %V) { 137; CHECK-LABEL: @test_umul2( 138; CHECK-NEXT: ret { i8, i1 } zeroinitializer 139; 140 %x = call {i8, i1} @llvm.umul.with.overflow.i8(i8 %V, i8 undef) 141 ret {i8, i1} %x 142} 143 144define {i8, i1} @test_umul3(i8 %V) { 145; CHECK-LABEL: @test_umul3( 146; CHECK-NEXT: ret { i8, i1 } zeroinitializer 147; 148 %x = call {i8, i1} @llvm.umul.with.overflow.i8(i8 0, i8 %V) 149 ret {i8, i1} %x 150} 151 152define {i8, i1} @test_umul4(i8 %V) { 153; CHECK-LABEL: @test_umul4( 154; CHECK-NEXT: ret { i8, i1 } zeroinitializer 155; 156 %x = call {i8, i1} @llvm.umul.with.overflow.i8(i8 undef, i8 %V) 157 ret {i8, i1} %x 158} 159 160define {i8, i1} @test_smul1(i8 %V) { 161; CHECK-LABEL: @test_smul1( 162; CHECK-NEXT: ret { i8, i1 } zeroinitializer 163; 164 %x = call {i8, i1} @llvm.smul.with.overflow.i8(i8 %V, i8 0) 165 ret {i8, i1} %x 166} 167 168define {i8, i1} @test_smul2(i8 %V) { 169; CHECK-LABEL: @test_smul2( 170; CHECK-NEXT: ret { i8, i1 } zeroinitializer 171; 172 %x = call {i8, i1} @llvm.smul.with.overflow.i8(i8 %V, i8 undef) 173 ret {i8, i1} %x 174} 175 176define {i8, i1} @test_smul3(i8 %V) { 177; CHECK-LABEL: @test_smul3( 178; CHECK-NEXT: ret { i8, i1 } zeroinitializer 179; 180 %x = call {i8, i1} @llvm.smul.with.overflow.i8(i8 0, i8 %V) 181 ret {i8, i1} %x 182} 183 184define {i8, i1} @test_smul4(i8 %V) { 185; CHECK-LABEL: @test_smul4( 186; CHECK-NEXT: ret { i8, i1 } zeroinitializer 187; 188 %x = call {i8, i1} @llvm.smul.with.overflow.i8(i8 undef, i8 %V) 189 ret {i8, i1} %x 190} 191 192declare i256 @llvm.cttz.i256(i256 %src, i1 %is_zero_undef) 193 194define i256 @test_cttz() { 195; CHECK-LABEL: @test_cttz( 196; CHECK-NEXT: ret i256 1 197; 198 %x = call i256 @llvm.cttz.i256(i256 10, i1 false) 199 ret i256 %x 200} 201 202declare <2 x i256> @llvm.cttz.v2i256(<2 x i256> %src, i1 %is_zero_undef) 203 204define <2 x i256> @test_cttz_vec() { 205; CHECK-LABEL: @test_cttz_vec( 206; CHECK-NEXT: ret <2 x i256> <i256 1, i256 1> 207; 208 %x = call <2 x i256> @llvm.cttz.v2i256(<2 x i256> <i256 10, i256 10>, i1 false) 209 ret <2 x i256> %x 210} 211 212declare i256 @llvm.ctpop.i256(i256 %src) 213 214define i256 @test_ctpop() { 215; CHECK-LABEL: @test_ctpop( 216; CHECK-NEXT: ret i256 2 217; 218 %x = call i256 @llvm.ctpop.i256(i256 10) 219 ret i256 %x 220} 221 222; Test a non-intrinsic that we know about as a library call. 223declare float @fabs(float %x) 224 225define float @test_fabs_libcall() { 226; CHECK-LABEL: @test_fabs_libcall( 227; CHECK-NEXT: [[X:%.*]] = call float @fabs(float -4.200000e+01) 228; CHECK-NEXT: ret float 4.200000e+01 229; 230 231 %x = call float @fabs(float -42.0) 232; This is still a real function call, so instsimplify won't nuke it -- other 233; passes have to do that. 234 235 ret float %x 236} 237 238 239declare float @llvm.fabs.f32(float) nounwind readnone 240declare float @llvm.floor.f32(float) nounwind readnone 241declare float @llvm.ceil.f32(float) nounwind readnone 242declare float @llvm.trunc.f32(float) nounwind readnone 243declare float @llvm.rint.f32(float) nounwind readnone 244declare float @llvm.nearbyint.f32(float) nounwind readnone 245declare float @llvm.canonicalize.f32(float) nounwind readnone 246 247; Test idempotent intrinsics 248define float @test_idempotence(float %a) { 249; CHECK-LABEL: @test_idempotence( 250; CHECK-NEXT: [[A0:%.*]] = call float @llvm.fabs.f32(float [[A:%.*]]) 251; CHECK-NEXT: [[B0:%.*]] = call float @llvm.floor.f32(float [[A]]) 252; CHECK-NEXT: [[C0:%.*]] = call float @llvm.ceil.f32(float [[A]]) 253; CHECK-NEXT: [[D0:%.*]] = call float @llvm.trunc.f32(float [[A]]) 254; CHECK-NEXT: [[E0:%.*]] = call float @llvm.rint.f32(float [[A]]) 255; CHECK-NEXT: [[F0:%.*]] = call float @llvm.nearbyint.f32(float [[A]]) 256; CHECK-NEXT: [[G0:%.*]] = call float @llvm.canonicalize.f32(float [[A]]) 257; CHECK-NEXT: [[R0:%.*]] = fadd float [[A0]], [[B0]] 258; CHECK-NEXT: [[R1:%.*]] = fadd float [[R0]], [[C0]] 259; CHECK-NEXT: [[R2:%.*]] = fadd float [[R1]], [[D0]] 260; CHECK-NEXT: [[R3:%.*]] = fadd float [[R2]], [[E0]] 261; CHECK-NEXT: [[R4:%.*]] = fadd float [[R3]], [[F0]] 262; CHECK-NEXT: [[R5:%.*]] = fadd float [[R4]], [[G0]] 263; CHECK-NEXT: ret float [[R5]] 264; 265 266 %a0 = call float @llvm.fabs.f32(float %a) 267 %a1 = call float @llvm.fabs.f32(float %a0) 268 269 %b0 = call float @llvm.floor.f32(float %a) 270 %b1 = call float @llvm.floor.f32(float %b0) 271 272 %c0 = call float @llvm.ceil.f32(float %a) 273 %c1 = call float @llvm.ceil.f32(float %c0) 274 275 %d0 = call float @llvm.trunc.f32(float %a) 276 %d1 = call float @llvm.trunc.f32(float %d0) 277 278 %e0 = call float @llvm.rint.f32(float %a) 279 %e1 = call float @llvm.rint.f32(float %e0) 280 281 %f0 = call float @llvm.nearbyint.f32(float %a) 282 %f1 = call float @llvm.nearbyint.f32(float %f0) 283 284 %g0 = call float @llvm.canonicalize.f32(float %a) 285 %g1 = call float @llvm.canonicalize.f32(float %g0) 286 287 %r0 = fadd float %a1, %b1 288 %r1 = fadd float %r0, %c1 289 %r2 = fadd float %r1, %d1 290 %r3 = fadd float %r2, %e1 291 %r4 = fadd float %r3, %f1 292 %r5 = fadd float %r4, %g1 293 294 ret float %r5 295} 296 297define i8* @operator_new() { 298; CHECK-LABEL: @operator_new( 299; CHECK-NEXT: entry: 300; CHECK-NEXT: [[CALL:%.*]] = tail call noalias i8* @_Znwm(i64 8) 301; CHECK-NEXT: br i1 false, label [[CAST_END:%.*]], label [[CAST_NOTNULL:%.*]] 302; CHECK: cast.notnull: 303; CHECK-NEXT: [[ADD_PTR:%.*]] = getelementptr inbounds i8, i8* [[CALL]], i64 4 304; CHECK-NEXT: br label [[CAST_END]] 305; CHECK: cast.end: 306; CHECK-NEXT: [[CAST_RESULT:%.*]] = phi i8* [ [[ADD_PTR]], [[CAST_NOTNULL]] ], [ null, [[ENTRY:%.*]] ] 307; CHECK-NEXT: ret i8* [[CAST_RESULT]] 308; 309entry: 310 %call = tail call noalias i8* @_Znwm(i64 8) 311 %cmp = icmp eq i8* %call, null 312 br i1 %cmp, label %cast.end, label %cast.notnull 313 314cast.notnull: ; preds = %entry 315 %add.ptr = getelementptr inbounds i8, i8* %call, i64 4 316 br label %cast.end 317 318cast.end: ; preds = %cast.notnull, %entry 319 %cast.result = phi i8* [ %add.ptr, %cast.notnull ], [ null, %entry ] 320 ret i8* %cast.result 321 322} 323 324declare nonnull noalias i8* @_Znwm(i64) 325 326%"struct.std::nothrow_t" = type { i8 } 327@_ZSt7nothrow = external global %"struct.std::nothrow_t" 328 329define i8* @operator_new_nothrow_t() { 330; CHECK-LABEL: @operator_new_nothrow_t( 331; CHECK-NEXT: entry: 332; CHECK-NEXT: [[CALL:%.*]] = tail call noalias i8* @_ZnamRKSt9nothrow_t(i64 8, %"struct.std::nothrow_t"* @_ZSt7nothrow) 333; CHECK-NEXT: [[CMP:%.*]] = icmp eq i8* [[CALL]], null 334; CHECK-NEXT: br i1 [[CMP]], label [[CAST_END:%.*]], label [[CAST_NOTNULL:%.*]] 335; CHECK: cast.notnull: 336; CHECK-NEXT: [[ADD_PTR:%.*]] = getelementptr inbounds i8, i8* [[CALL]], i64 4 337; CHECK-NEXT: br label [[CAST_END]] 338; CHECK: cast.end: 339; CHECK-NEXT: [[CAST_RESULT:%.*]] = phi i8* [ [[ADD_PTR]], [[CAST_NOTNULL]] ], [ null, [[ENTRY:%.*]] ] 340; CHECK-NEXT: ret i8* [[CAST_RESULT]] 341; 342entry: 343 %call = tail call noalias i8* @_ZnamRKSt9nothrow_t(i64 8, %"struct.std::nothrow_t"* @_ZSt7nothrow) 344 %cmp = icmp eq i8* %call, null 345 br i1 %cmp, label %cast.end, label %cast.notnull 346 347cast.notnull: ; preds = %entry 348 %add.ptr = getelementptr inbounds i8, i8* %call, i64 4 349 br label %cast.end 350 351cast.end: ; preds = %cast.notnull, %entry 352 %cast.result = phi i8* [ %add.ptr, %cast.notnull ], [ null, %entry ] 353 ret i8* %cast.result 354 355} 356 357declare i8* @_ZnamRKSt9nothrow_t(i64, %"struct.std::nothrow_t"*) nounwind 358 359define i8* @malloc_can_return_null() { 360; CHECK-LABEL: @malloc_can_return_null( 361; CHECK-NEXT: entry: 362; CHECK-NEXT: [[CALL:%.*]] = tail call noalias i8* @malloc(i64 8) 363; CHECK-NEXT: [[CMP:%.*]] = icmp eq i8* [[CALL]], null 364; CHECK-NEXT: br i1 [[CMP]], label [[CAST_END:%.*]], label [[CAST_NOTNULL:%.*]] 365; CHECK: cast.notnull: 366; CHECK-NEXT: [[ADD_PTR:%.*]] = getelementptr inbounds i8, i8* [[CALL]], i64 4 367; CHECK-NEXT: br label [[CAST_END]] 368; CHECK: cast.end: 369; CHECK-NEXT: [[CAST_RESULT:%.*]] = phi i8* [ [[ADD_PTR]], [[CAST_NOTNULL]] ], [ null, [[ENTRY:%.*]] ] 370; CHECK-NEXT: ret i8* [[CAST_RESULT]] 371; 372entry: 373 %call = tail call noalias i8* @malloc(i64 8) 374 %cmp = icmp eq i8* %call, null 375 br i1 %cmp, label %cast.end, label %cast.notnull 376 377cast.notnull: ; preds = %entry 378 %add.ptr = getelementptr inbounds i8, i8* %call, i64 4 379 br label %cast.end 380 381cast.end: ; preds = %cast.notnull, %entry 382 %cast.result = phi i8* [ %add.ptr, %cast.notnull ], [ null, %entry ] 383 ret i8* %cast.result 384 385} 386 387define i32 @call_null() { 388; CHECK-LABEL: @call_null( 389; CHECK-NEXT: entry: 390; CHECK-NEXT: [[CALL:%.*]] = call i32 null() 391; CHECK-NEXT: ret i32 undef 392; 393entry: 394 %call = call i32 null() 395 ret i32 %call 396} 397 398define i32 @call_undef() { 399; CHECK-LABEL: @call_undef( 400; CHECK-NEXT: entry: 401; CHECK-NEXT: [[CALL:%.*]] = call i32 undef() 402; CHECK-NEXT: ret i32 undef 403; 404entry: 405 %call = call i32 undef() 406 ret i32 %call 407} 408 409@GV = private constant [8 x i32] [i32 42, i32 43, i32 44, i32 45, i32 46, i32 47, i32 48, i32 49] 410 411define <8 x i32> @partial_masked_load() { 412; CHECK-LABEL: @partial_masked_load( 413; CHECK-NEXT: ret <8 x i32> <i32 undef, i32 undef, i32 42, i32 43, i32 44, i32 45, i32 46, i32 47> 414; 415 %masked.load = call <8 x i32> @llvm.masked.load.v8i32.p0v8i32(<8 x i32>* bitcast (i32* getelementptr ([8 x i32], [8 x i32]* @GV, i64 0, i64 -2) to <8 x i32>*), i32 4, <8 x i1> <i1 false, i1 false, i1 true, i1 true, i1 true, i1 true, i1 true, i1 true>, <8 x i32> undef) 416 ret <8 x i32> %masked.load 417} 418 419define <8 x i32> @masked_load_undef_mask(<8 x i32>* %V) { 420; CHECK-LABEL: @masked_load_undef_mask( 421; CHECK-NEXT: ret <8 x i32> <i32 1, i32 0, i32 1, i32 0, i32 1, i32 0, i32 1, i32 0> 422; 423 %masked.load = call <8 x i32> @llvm.masked.load.v8i32.p0v8i32(<8 x i32>* %V, i32 4, <8 x i1> undef, <8 x i32> <i32 1, i32 0, i32 1, i32 0, i32 1, i32 0, i32 1, i32 0>) 424 ret <8 x i32> %masked.load 425} 426 427declare noalias i8* @malloc(i64) 428 429declare <8 x i32> @llvm.masked.load.v8i32.p0v8i32(<8 x i32>*, i32, <8 x i1>, <8 x i32>) 430 431declare double @llvm.powi.f64(double, i32) 432declare <2 x double> @llvm.powi.v2f64(<2 x double>, i32) 433 434define double @constant_fold_powi() { 435; CHECK-LABEL: @constant_fold_powi( 436; CHECK-NEXT: ret double 9.000000e+00 437; 438 %t0 = call double @llvm.powi.f64(double 3.00000e+00, i32 2) 439 ret double %t0 440} 441 442define <2 x double> @constant_fold_powi_vec() { 443; CHECK-LABEL: @constant_fold_powi_vec( 444; CHECK-NEXT: ret <2 x double> <double 9.000000e+00, double 2.500000e+01> 445; 446 %t0 = call <2 x double> @llvm.powi.v2f64(<2 x double> <double 3.00000e+00, double 5.00000e+00>, i32 2) 447 ret <2 x double> %t0 448} 449 450declare i8 @llvm.fshl.i8(i8, i8, i8) 451declare i9 @llvm.fshr.i9(i9, i9, i9) 452declare <2 x i7> @llvm.fshl.v2i7(<2 x i7>, <2 x i7>, <2 x i7>) 453declare <2 x i8> @llvm.fshr.v2i8(<2 x i8>, <2 x i8>, <2 x i8>) 454 455define i8 @fshl_no_shift(i8 %x, i8 %y) { 456; CHECK-LABEL: @fshl_no_shift( 457; CHECK-NEXT: ret i8 [[X:%.*]] 458; 459 %z = call i8 @llvm.fshl.i8(i8 %x, i8 %y, i8 0) 460 ret i8 %z 461} 462 463define i9 @fshr_no_shift(i9 %x, i9 %y) { 464; CHECK-LABEL: @fshr_no_shift( 465; CHECK-NEXT: ret i9 [[Y:%.*]] 466; 467 %z = call i9 @llvm.fshr.i9(i9 %x, i9 %y, i9 0) 468 ret i9 %z 469} 470 471define i8 @fshl_no_shift_modulo_bitwidth(i8 %x, i8 %y) { 472; CHECK-LABEL: @fshl_no_shift_modulo_bitwidth( 473; CHECK-NEXT: ret i8 [[X:%.*]] 474; 475 %z = call i8 @llvm.fshl.i8(i8 %x, i8 %y, i8 40) 476 ret i8 %z 477} 478 479define i9 @fshr_no_shift_modulo_bitwidth(i9 %x, i9 %y) { 480; CHECK-LABEL: @fshr_no_shift_modulo_bitwidth( 481; CHECK-NEXT: ret i9 [[Y:%.*]] 482; 483 %z = call i9 @llvm.fshr.i9(i9 %x, i9 %y, i9 189) 484 ret i9 %z 485} 486 487define <2 x i7> @fshl_no_shift_modulo_bitwidth_splat(<2 x i7> %x, <2 x i7> %y) { 488; CHECK-LABEL: @fshl_no_shift_modulo_bitwidth_splat( 489; CHECK-NEXT: ret <2 x i7> [[X:%.*]] 490; 491 %z = call <2 x i7> @llvm.fshl.v2i7(<2 x i7> %x, <2 x i7> %y, <2 x i7> <i7 21, i7 21>) 492 ret <2 x i7> %z 493} 494 495define <2 x i8> @fshr_no_shift_modulo_bitwidth_splat(<2 x i8> %x, <2 x i8> %y) { 496; CHECK-LABEL: @fshr_no_shift_modulo_bitwidth_splat( 497; CHECK-NEXT: ret <2 x i8> [[Y:%.*]] 498; 499 %z = call <2 x i8> @llvm.fshr.v2i8(<2 x i8> %x, <2 x i8> %y, <2 x i8> <i8 72, i8 72>) 500 ret <2 x i8> %z 501} 502 503