/external/llvm-project/llvm/test/CodeGen/ARM/ |
D | spill-q.ll | 30 %ld5 = call <4 x float> @llvm.arm.neon.vld1.v4f32.p0i8(i8* undef, i32 1) nounwind 72 %tmp3 = fadd <4 x float> %tmp2, %ld5
|
/external/llvm-project/llvm/test/CodeGen/Thumb2/ |
D | thumb2-spill-q.ll | 30 %ld5 = call <4 x float> @llvm.arm.neon.vld1.v4f32.p0i8(i8* undef, i32 1) nounwind 72 %tmp3 = fadd <4 x float> %tmp2, %ld5
|
/external/llvm-project/llvm/test/Transforms/SLPVectorizer/X86/ |
D | bswap.ll | 124 %ld5 = load i32, i32* getelementptr inbounds ([8 x i32], [8 x i32]* @src32, i32 0, i64 5), align 2 132 %bswap5 = call i32 @llvm.bswap.i32(i32 %ld5) 158 …%ld5 = load i16, i16* getelementptr inbounds ([16 x i16], [16 x i16]* @src16, i16 0, i64 5), align… 166 %bswap5 = call i16 @llvm.bswap.i16(i16 %ld5) 201 …%ld5 = load i16, i16* getelementptr inbounds ([16 x i16], [16 x i16]* @src16, i16 0, i64 5), ali… 217 %bswap5 = call i16 @llvm.bswap.i16(i16 %ld5)
|
D | bitreverse.ll | 126 %ld5 = load i32, i32* getelementptr inbounds ([8 x i32], [8 x i32]* @src32, i32 0, i64 5), align 2 134 %bitreverse5 = call i32 @llvm.bitreverse.i32(i32 %ld5) 160 …%ld5 = load i16, i16* getelementptr inbounds ([16 x i16], [16 x i16]* @src16, i16 0, i64 5), align… 168 %bitreverse5 = call i16 @llvm.bitreverse.i16(i16 %ld5) 209 …%ld5 = load i16, i16* getelementptr inbounds ([16 x i16], [16 x i16]* @src16, i16 0, i64 5), ali… 225 %bitreverse5 = call i16 @llvm.bitreverse.i16(i16 %ld5) 267 %ld5 = load i8, i8* getelementptr inbounds ([32 x i8], [32 x i8]* @src8, i8 0, i64 5), align 1 283 %bitreverse5 = call i8 @llvm.bitreverse.i8(i8 %ld5) 340 %ld5 = load i8, i8* getelementptr inbounds ([32 x i8], [32 x i8]* @src8, i8 0, i64 5), align 1 372 %bitreverse5 = call i8 @llvm.bitreverse.i8(i8 %ld5)
|
D | ctlz.ll | 183 %ld5 = load i32, i32* getelementptr inbounds ([8 x i32], [8 x i32]* @src32, i32 0, i64 5), align 2 191 %ctlz5 = call i32 @llvm.ctlz.i32(i32 %ld5, i1 0) 217 …%ld5 = load i16, i16* getelementptr inbounds ([16 x i16], [16 x i16]* @src16, i16 0, i64 5), align… 225 %ctlz5 = call i16 @llvm.ctlz.i16(i16 %ld5, i1 0) 260 …%ld5 = load i16, i16* getelementptr inbounds ([16 x i16], [16 x i16]* @src16, i16 0, i64 5), ali… 276 %ctlz5 = call i16 @llvm.ctlz.i16(i16 %ld5, i1 0) 318 %ld5 = load i8, i8* getelementptr inbounds ([32 x i8], [32 x i8]* @src8, i8 0, i64 5), align 1 334 %ctlz5 = call i8 @llvm.ctlz.i8(i8 %ld5, i1 0) 385 %ld5 = load i8, i8* getelementptr inbounds ([32 x i8], [32 x i8]* @src8, i8 0, i64 5), align 1 417 %ctlz5 = call i8 @llvm.ctlz.i8(i8 %ld5, i1 0) [all …]
|
D | cttz.ll | 183 %ld5 = load i32, i32* getelementptr inbounds ([8 x i32], [8 x i32]* @src32, i32 0, i64 5), align 2 191 %cttz5 = call i32 @llvm.cttz.i32(i32 %ld5, i1 0) 217 …%ld5 = load i16, i16* getelementptr inbounds ([16 x i16], [16 x i16]* @src16, i16 0, i64 5), align… 225 %cttz5 = call i16 @llvm.cttz.i16(i16 %ld5, i1 0) 260 …%ld5 = load i16, i16* getelementptr inbounds ([16 x i16], [16 x i16]* @src16, i16 0, i64 5), ali… 276 %cttz5 = call i16 @llvm.cttz.i16(i16 %ld5, i1 0) 318 %ld5 = load i8, i8* getelementptr inbounds ([32 x i8], [32 x i8]* @src8, i8 0, i64 5), align 1 334 %cttz5 = call i8 @llvm.cttz.i8(i8 %ld5, i1 0) 385 %ld5 = load i8, i8* getelementptr inbounds ([32 x i8], [32 x i8]* @src8, i8 0, i64 5), align 1 417 %cttz5 = call i8 @llvm.cttz.i8(i8 %ld5, i1 0) [all …]
|
D | uitofp.ll | 106 %ld5 = load i64, i64* getelementptr inbounds ([8 x i64], [8 x i64]* @src64, i32 0, i64 5), align 8 114 %cvt5 = uitofp i64 %ld5 to double 253 …%ld5 = load i32, i32* getelementptr inbounds ([16 x i32], [16 x i32]* @src32, i32 0, i64 5), align… 261 %cvt5 = uitofp i32 %ld5 to double 379 …%ld5 = load i16, i16* getelementptr inbounds ([32 x i16], [32 x i16]* @src16, i32 0, i64 5), align… 387 %cvt5 = uitofp i16 %ld5 to double 526 %ld5 = load i8, i8* getelementptr inbounds ([64 x i8], [64 x i8]* @src8, i32 0, i64 5), align 1 534 %cvt5 = uitofp i8 %ld5 to double 701 %ld5 = load i64, i64* getelementptr inbounds ([8 x i64], [8 x i64]* @src64, i32 0, i64 5), align 8 709 %cvt5 = uitofp i64 %ld5 to float [all …]
|
D | ctpop.ll | 245 %ld5 = load i32, i32* getelementptr inbounds ([8 x i32], [8 x i32]* @src32, i32 0, i64 5), align 2 253 %ctpop5 = call i32 @llvm.ctpop.i32(i32 %ld5) 279 …%ld5 = load i16, i16* getelementptr inbounds ([16 x i16], [16 x i16]* @src16, i16 0, i64 5), align… 287 %ctpop5 = call i16 @llvm.ctpop.i16(i16 %ld5) 322 …%ld5 = load i16, i16* getelementptr inbounds ([16 x i16], [16 x i16]* @src16, i16 0, i64 5), ali… 338 %ctpop5 = call i16 @llvm.ctpop.i16(i16 %ld5) 380 %ld5 = load i8, i8* getelementptr inbounds ([32 x i8], [32 x i8]* @src8, i8 0, i64 5), align 1 396 %ctpop5 = call i8 @llvm.ctpop.i8(i8 %ld5) 447 %ld5 = load i8, i8* getelementptr inbounds ([32 x i8], [32 x i8]* @src8, i8 0, i64 5), align 1 479 %ctpop5 = call i8 @llvm.ctpop.i8(i8 %ld5)
|
D | sitofp.ll | 196 %ld5 = load i64, i64* getelementptr inbounds ([8 x i64], [8 x i64]* @src64, i32 0, i64 5), align 8 204 %cvt5 = sitofp i64 %ld5 to double 322 …%ld5 = load i32, i32* getelementptr inbounds ([16 x i32], [16 x i32]* @src32, i32 0, i64 5), align… 330 %cvt5 = sitofp i32 %ld5 to double 448 …%ld5 = load i16, i16* getelementptr inbounds ([32 x i16], [32 x i16]* @src16, i32 0, i64 5), align… 456 %cvt5 = sitofp i16 %ld5 to double 574 %ld5 = load i8, i8* getelementptr inbounds ([64 x i8], [64 x i8]* @src8, i32 0, i64 5), align 1 582 %cvt5 = sitofp i8 %ld5 to double 749 %ld5 = load i64, i64* getelementptr inbounds ([8 x i64], [8 x i64]* @src64, i32 0, i64 5), align 8 757 %cvt5 = sitofp i64 %ld5 to float [all …]
|
D | fround.ll | 177 …%ld5 = load double, double* getelementptr inbounds ([8 x double], [8 x double]* @src64, i32 0, i64… 185 %ceil5 = call double @llvm.ceil.f64(double %ld5) 348 …%ld5 = load double, double* getelementptr inbounds ([8 x double], [8 x double]* @src64, i32 0, i64… 356 %floor5 = call double @llvm.floor.f64(double %ld5) 519 …%ld5 = load double, double* getelementptr inbounds ([8 x double], [8 x double]* @src64, i32 0, i64… 527 %nearbyint5 = call double @llvm.nearbyint.f64(double %ld5) 690 …%ld5 = load double, double* getelementptr inbounds ([8 x double], [8 x double]* @src64, i32 0, i64… 698 %rint5 = call double @llvm.rint.f64(double %ld5) 861 …%ld5 = load double, double* getelementptr inbounds ([8 x double], [8 x double]* @src64, i32 0, i64… 869 %trunc5 = call double @llvm.trunc.f64(double %ld5) [all …]
|
D | bad-reduction.ll | 241 %ld5 = load i8, i8* %g5, align 1 250 %z5 = zext i8 %ld5 to i64 328 %ld5 = load i8, i8* %g5, align 1 337 %z5 = zext i8 %ld5 to i64
|
/external/llvm/test/Transforms/SLPVectorizer/X86/ |
D | bswap.ll | 124 %ld5 = load i32, i32* getelementptr inbounds ([8 x i32], [8 x i32]* @src32, i32 0, i64 5), align 2 132 %bswap5 = call i32 @llvm.bswap.i32(i32 %ld5) 158 …%ld5 = load i16, i16* getelementptr inbounds ([16 x i16], [16 x i16]* @src16, i16 0, i64 5), align… 166 %bswap5 = call i16 @llvm.bswap.i16(i16 %ld5) 201 …%ld5 = load i16, i16* getelementptr inbounds ([16 x i16], [16 x i16]* @src16, i16 0, i64 5), ali… 217 %bswap5 = call i16 @llvm.bswap.i16(i16 %ld5)
|
D | ctpop.ll | 112 %ld5 = load i32, i32* getelementptr inbounds ([8 x i32], [8 x i32]* @src32, i32 0, i64 5), align 2 120 %ctpop5 = call i32 @llvm.ctpop.i32(i32 %ld5) 146 …%ld5 = load i16, i16* getelementptr inbounds ([16 x i16], [16 x i16]* @src16, i16 0, i64 5), align… 154 %ctpop5 = call i16 @llvm.ctpop.i16(i16 %ld5) 189 …%ld5 = load i16, i16* getelementptr inbounds ([16 x i16], [16 x i16]* @src16, i16 0, i64 5), ali… 205 %ctpop5 = call i16 @llvm.ctpop.i16(i16 %ld5) 247 %ld5 = load i8, i8* getelementptr inbounds ([32 x i8], [32 x i8]* @src8, i8 0, i64 5), align 1 263 %ctpop5 = call i8 @llvm.ctpop.i8(i8 %ld5) 308 %ld5 = load i8, i8* getelementptr inbounds ([32 x i8], [32 x i8]* @src8, i8 0, i64 5), align 1 340 %ctpop5 = call i8 @llvm.ctpop.i8(i8 %ld5)
|
D | cttz.ll | 140 %ld5 = load i32, i32* getelementptr inbounds ([8 x i32], [8 x i32]* @src32, i32 0, i64 5), align 2 148 %cttz5 = call i32 @llvm.cttz.i32(i32 %ld5, i1 0) 195 …%ld5 = load i16, i16* getelementptr inbounds ([16 x i16], [16 x i16]* @src16, i16 0, i64 5), align… 203 %cttz5 = call i16 @llvm.cttz.i16(i16 %ld5, i1 0) 274 …%ld5 = load i16, i16* getelementptr inbounds ([16 x i16], [16 x i16]* @src16, i16 0, i64 5), ali… 290 %cttz5 = call i16 @llvm.cttz.i16(i16 %ld5, i1 0) 377 %ld5 = load i8, i8* getelementptr inbounds ([32 x i8], [32 x i8]* @src8, i8 0, i64 5), align 1 393 %cttz5 = call i8 @llvm.cttz.i8(i8 %ld5, i1 0) 528 %ld5 = load i8, i8* getelementptr inbounds ([32 x i8], [32 x i8]* @src8, i8 0, i64 5), align 1 560 %cttz5 = call i8 @llvm.cttz.i8(i8 %ld5, i1 0) [all …]
|
D | ctlz.ll | 140 %ld5 = load i32, i32* getelementptr inbounds ([8 x i32], [8 x i32]* @src32, i32 0, i64 5), align 2 148 %ctlz5 = call i32 @llvm.ctlz.i32(i32 %ld5, i1 0) 195 …%ld5 = load i16, i16* getelementptr inbounds ([16 x i16], [16 x i16]* @src16, i16 0, i64 5), align… 203 %ctlz5 = call i16 @llvm.ctlz.i16(i16 %ld5, i1 0) 274 …%ld5 = load i16, i16* getelementptr inbounds ([16 x i16], [16 x i16]* @src16, i16 0, i64 5), ali… 290 %ctlz5 = call i16 @llvm.ctlz.i16(i16 %ld5, i1 0) 377 %ld5 = load i8, i8* getelementptr inbounds ([32 x i8], [32 x i8]* @src8, i8 0, i64 5), align 1 393 %ctlz5 = call i8 @llvm.ctlz.i8(i8 %ld5, i1 0) 528 %ld5 = load i8, i8* getelementptr inbounds ([32 x i8], [32 x i8]* @src8, i8 0, i64 5), align 1 560 %ctlz5 = call i8 @llvm.ctlz.i8(i8 %ld5, i1 0) [all …]
|
D | fround.ll | 176 …%ld5 = load double, double* getelementptr inbounds ([8 x double], [8 x double]* @src64, i32 0, i64… 184 %ceil5 = call double @llvm.ceil.f64(double %ld5) 347 …%ld5 = load double, double* getelementptr inbounds ([8 x double], [8 x double]* @src64, i32 0, i64… 355 %floor5 = call double @llvm.floor.f64(double %ld5) 518 …%ld5 = load double, double* getelementptr inbounds ([8 x double], [8 x double]* @src64, i32 0, i64… 526 %nearbyint5 = call double @llvm.nearbyint.f64(double %ld5) 689 …%ld5 = load double, double* getelementptr inbounds ([8 x double], [8 x double]* @src64, i32 0, i64… 697 %rint5 = call double @llvm.rint.f64(double %ld5) 860 …%ld5 = load double, double* getelementptr inbounds ([8 x double], [8 x double]* @src64, i32 0, i64… 868 %trunc5 = call double @llvm.trunc.f64(double %ld5) [all …]
|
D | bitreverse.ll | 204 %ld5 = load i32, i32* getelementptr inbounds ([8 x i32], [8 x i32]* @src32, i32 0, i64 5), align 2 212 %bitreverse5 = call i32 @llvm.bitreverse.i32(i32 %ld5) 271 …%ld5 = load i16, i16* getelementptr inbounds ([16 x i16], [16 x i16]* @src16, i16 0, i64 5), align… 279 %bitreverse5 = call i16 @llvm.bitreverse.i16(i16 %ld5) 362 …%ld5 = load i16, i16* getelementptr inbounds ([16 x i16], [16 x i16]* @src16, i16 0, i64 5), ali… 378 %bitreverse5 = call i16 @llvm.bitreverse.i16(i16 %ld5) 477 %ld5 = load i8, i8* getelementptr inbounds ([32 x i8], [32 x i8]* @src8, i8 0, i64 5), align 1 493 %bitreverse5 = call i8 @llvm.bitreverse.i8(i8 %ld5) 646 %ld5 = load i8, i8* getelementptr inbounds ([32 x i8], [32 x i8]* @src8, i8 0, i64 5), align 1 678 %bitreverse5 = call i8 @llvm.bitreverse.i8(i8 %ld5)
|
/external/llvm/test/CodeGen/ARM/ |
D | spill-q.ll | 30 %ld5 = call <4 x float> @llvm.arm.neon.vld1.v4f32.p0i8(i8* undef, i32 1) nounwind 72 %tmp3 = fadd <4 x float> %tmp2, %ld5
|
/external/llvm/test/CodeGen/Thumb2/ |
D | thumb2-spill-q.ll | 30 %ld5 = call <4 x float> @llvm.arm.neon.vld1.v4f32.p0i8(i8* undef, i32 1) nounwind 72 %tmp3 = fadd <4 x float> %tmp2, %ld5
|
/external/llvm-project/libcxx/test/std/numerics/numbers/ |
D | specialize.pass.cpp | 63 [[maybe_unused]] long double ld5{std::numbers::inv_sqrtpi_v<long double>}; in tests() local
|
D | defined.pass.cpp | 62 [[maybe_unused]] const long double* ld5{&std::numbers::inv_sqrtpi_v<long double>}; in tests() local
|