/external/llvm-project/llvm/test/CodeGen/ARM/ |
D | spill-q.ll | 32 %ld6 = call <4 x float> @llvm.arm.neon.vld1.v4f32.p0i8(i8* undef, i32 1) nounwind 73 %tmp4 = fadd <4 x float> %tmp3, %ld6
|
/external/llvm-project/llvm/test/CodeGen/Thumb2/ |
D | thumb2-spill-q.ll | 32 %ld6 = call <4 x float> @llvm.arm.neon.vld1.v4f32.p0i8(i8* undef, i32 1) nounwind 73 %tmp4 = fadd <4 x float> %tmp3, %ld6
|
/external/llvm-project/llvm/test/Transforms/SLPVectorizer/X86/ |
D | bswap.ll | 125 %ld6 = load i32, i32* getelementptr inbounds ([8 x i32], [8 x i32]* @src32, i32 0, i64 6), align 2 133 %bswap6 = call i32 @llvm.bswap.i32(i32 %ld6) 159 …%ld6 = load i16, i16* getelementptr inbounds ([16 x i16], [16 x i16]* @src16, i16 0, i64 6), align… 167 %bswap6 = call i16 @llvm.bswap.i16(i16 %ld6) 202 …%ld6 = load i16, i16* getelementptr inbounds ([16 x i16], [16 x i16]* @src16, i16 0, i64 6), ali… 218 %bswap6 = call i16 @llvm.bswap.i16(i16 %ld6)
|
D | bitreverse.ll | 127 %ld6 = load i32, i32* getelementptr inbounds ([8 x i32], [8 x i32]* @src32, i32 0, i64 6), align 2 135 %bitreverse6 = call i32 @llvm.bitreverse.i32(i32 %ld6) 161 …%ld6 = load i16, i16* getelementptr inbounds ([16 x i16], [16 x i16]* @src16, i16 0, i64 6), align… 169 %bitreverse6 = call i16 @llvm.bitreverse.i16(i16 %ld6) 210 …%ld6 = load i16, i16* getelementptr inbounds ([16 x i16], [16 x i16]* @src16, i16 0, i64 6), ali… 226 %bitreverse6 = call i16 @llvm.bitreverse.i16(i16 %ld6) 268 %ld6 = load i8, i8* getelementptr inbounds ([32 x i8], [32 x i8]* @src8, i8 0, i64 6), align 1 284 %bitreverse6 = call i8 @llvm.bitreverse.i8(i8 %ld6) 341 %ld6 = load i8, i8* getelementptr inbounds ([32 x i8], [32 x i8]* @src8, i8 0, i64 6), align 1 373 %bitreverse6 = call i8 @llvm.bitreverse.i8(i8 %ld6)
|
D | ctlz.ll | 184 %ld6 = load i32, i32* getelementptr inbounds ([8 x i32], [8 x i32]* @src32, i32 0, i64 6), align 2 192 %ctlz6 = call i32 @llvm.ctlz.i32(i32 %ld6, i1 0) 218 …%ld6 = load i16, i16* getelementptr inbounds ([16 x i16], [16 x i16]* @src16, i16 0, i64 6), align… 226 %ctlz6 = call i16 @llvm.ctlz.i16(i16 %ld6, i1 0) 261 …%ld6 = load i16, i16* getelementptr inbounds ([16 x i16], [16 x i16]* @src16, i16 0, i64 6), ali… 277 %ctlz6 = call i16 @llvm.ctlz.i16(i16 %ld6, i1 0) 319 %ld6 = load i8, i8* getelementptr inbounds ([32 x i8], [32 x i8]* @src8, i8 0, i64 6), align 1 335 %ctlz6 = call i8 @llvm.ctlz.i8(i8 %ld6, i1 0) 386 %ld6 = load i8, i8* getelementptr inbounds ([32 x i8], [32 x i8]* @src8, i8 0, i64 6), align 1 418 %ctlz6 = call i8 @llvm.ctlz.i8(i8 %ld6, i1 0) [all …]
|
D | cttz.ll | 184 %ld6 = load i32, i32* getelementptr inbounds ([8 x i32], [8 x i32]* @src32, i32 0, i64 6), align 2 192 %cttz6 = call i32 @llvm.cttz.i32(i32 %ld6, i1 0) 218 …%ld6 = load i16, i16* getelementptr inbounds ([16 x i16], [16 x i16]* @src16, i16 0, i64 6), align… 226 %cttz6 = call i16 @llvm.cttz.i16(i16 %ld6, i1 0) 261 …%ld6 = load i16, i16* getelementptr inbounds ([16 x i16], [16 x i16]* @src16, i16 0, i64 6), ali… 277 %cttz6 = call i16 @llvm.cttz.i16(i16 %ld6, i1 0) 319 %ld6 = load i8, i8* getelementptr inbounds ([32 x i8], [32 x i8]* @src8, i8 0, i64 6), align 1 335 %cttz6 = call i8 @llvm.cttz.i8(i8 %ld6, i1 0) 386 %ld6 = load i8, i8* getelementptr inbounds ([32 x i8], [32 x i8]* @src8, i8 0, i64 6), align 1 418 %cttz6 = call i8 @llvm.cttz.i8(i8 %ld6, i1 0) [all …]
|
D | uitofp.ll | 107 …%ld6 = load i64, i64* getelementptr inbounds ([8 x i64], [8 x i64]* @src64, i32 0, i64 6), align 16 115 %cvt6 = uitofp i64 %ld6 to double 254 …%ld6 = load i32, i32* getelementptr inbounds ([16 x i32], [16 x i32]* @src32, i32 0, i64 6), align… 262 %cvt6 = uitofp i32 %ld6 to double 380 …%ld6 = load i16, i16* getelementptr inbounds ([32 x i16], [32 x i16]* @src16, i32 0, i64 6), align… 388 %cvt6 = uitofp i16 %ld6 to double 527 %ld6 = load i8, i8* getelementptr inbounds ([64 x i8], [64 x i8]* @src8, i32 0, i64 6), align 2 535 %cvt6 = uitofp i8 %ld6 to double 702 …%ld6 = load i64, i64* getelementptr inbounds ([8 x i64], [8 x i64]* @src64, i32 0, i64 6), align 16 710 %cvt6 = uitofp i64 %ld6 to float [all …]
|
D | ctpop.ll | 246 %ld6 = load i32, i32* getelementptr inbounds ([8 x i32], [8 x i32]* @src32, i32 0, i64 6), align 2 254 %ctpop6 = call i32 @llvm.ctpop.i32(i32 %ld6) 280 …%ld6 = load i16, i16* getelementptr inbounds ([16 x i16], [16 x i16]* @src16, i16 0, i64 6), align… 288 %ctpop6 = call i16 @llvm.ctpop.i16(i16 %ld6) 323 …%ld6 = load i16, i16* getelementptr inbounds ([16 x i16], [16 x i16]* @src16, i16 0, i64 6), ali… 339 %ctpop6 = call i16 @llvm.ctpop.i16(i16 %ld6) 381 %ld6 = load i8, i8* getelementptr inbounds ([32 x i8], [32 x i8]* @src8, i8 0, i64 6), align 1 397 %ctpop6 = call i8 @llvm.ctpop.i8(i8 %ld6) 448 %ld6 = load i8, i8* getelementptr inbounds ([32 x i8], [32 x i8]* @src8, i8 0, i64 6), align 1 480 %ctpop6 = call i8 @llvm.ctpop.i8(i8 %ld6)
|
D | sitofp.ll | 197 …%ld6 = load i64, i64* getelementptr inbounds ([8 x i64], [8 x i64]* @src64, i32 0, i64 6), align 16 205 %cvt6 = sitofp i64 %ld6 to double 323 …%ld6 = load i32, i32* getelementptr inbounds ([16 x i32], [16 x i32]* @src32, i32 0, i64 6), align… 331 %cvt6 = sitofp i32 %ld6 to double 449 …%ld6 = load i16, i16* getelementptr inbounds ([32 x i16], [32 x i16]* @src16, i32 0, i64 6), align… 457 %cvt6 = sitofp i16 %ld6 to double 575 %ld6 = load i8, i8* getelementptr inbounds ([64 x i8], [64 x i8]* @src8, i32 0, i64 6), align 2 583 %cvt6 = sitofp i8 %ld6 to double 750 …%ld6 = load i64, i64* getelementptr inbounds ([8 x i64], [8 x i64]* @src64, i32 0, i64 6), align 16 758 %cvt6 = sitofp i64 %ld6 to float [all …]
|
D | fround.ll | 178 …%ld6 = load double, double* getelementptr inbounds ([8 x double], [8 x double]* @src64, i32 0, i64… 186 %ceil6 = call double @llvm.ceil.f64(double %ld6) 349 …%ld6 = load double, double* getelementptr inbounds ([8 x double], [8 x double]* @src64, i32 0, i64… 357 %floor6 = call double @llvm.floor.f64(double %ld6) 520 …%ld6 = load double, double* getelementptr inbounds ([8 x double], [8 x double]* @src64, i32 0, i64… 528 %nearbyint6 = call double @llvm.nearbyint.f64(double %ld6) 691 …%ld6 = load double, double* getelementptr inbounds ([8 x double], [8 x double]* @src64, i32 0, i64… 699 %rint6 = call double @llvm.rint.f64(double %ld6) 862 …%ld6 = load double, double* getelementptr inbounds ([8 x double], [8 x double]* @src64, i32 0, i64… 870 %trunc6 = call double @llvm.trunc.f64(double %ld6) [all …]
|
D | bad-reduction.ll | 242 %ld6 = load i8, i8* %g6, align 1 251 %z6 = zext i8 %ld6 to i64 329 %ld6 = load i8, i8* %g6, align 1 338 %z6 = zext i8 %ld6 to i64
|
/external/llvm/test/Transforms/SLPVectorizer/X86/ |
D | bswap.ll | 125 %ld6 = load i32, i32* getelementptr inbounds ([8 x i32], [8 x i32]* @src32, i32 0, i64 6), align 2 133 %bswap6 = call i32 @llvm.bswap.i32(i32 %ld6) 159 …%ld6 = load i16, i16* getelementptr inbounds ([16 x i16], [16 x i16]* @src16, i16 0, i64 6), align… 167 %bswap6 = call i16 @llvm.bswap.i16(i16 %ld6) 202 …%ld6 = load i16, i16* getelementptr inbounds ([16 x i16], [16 x i16]* @src16, i16 0, i64 6), ali… 218 %bswap6 = call i16 @llvm.bswap.i16(i16 %ld6)
|
D | ctpop.ll | 113 %ld6 = load i32, i32* getelementptr inbounds ([8 x i32], [8 x i32]* @src32, i32 0, i64 6), align 2 121 %ctpop6 = call i32 @llvm.ctpop.i32(i32 %ld6) 147 …%ld6 = load i16, i16* getelementptr inbounds ([16 x i16], [16 x i16]* @src16, i16 0, i64 6), align… 155 %ctpop6 = call i16 @llvm.ctpop.i16(i16 %ld6) 190 …%ld6 = load i16, i16* getelementptr inbounds ([16 x i16], [16 x i16]* @src16, i16 0, i64 6), ali… 206 %ctpop6 = call i16 @llvm.ctpop.i16(i16 %ld6) 248 %ld6 = load i8, i8* getelementptr inbounds ([32 x i8], [32 x i8]* @src8, i8 0, i64 6), align 1 264 %ctpop6 = call i8 @llvm.ctpop.i8(i8 %ld6) 309 %ld6 = load i8, i8* getelementptr inbounds ([32 x i8], [32 x i8]* @src8, i8 0, i64 6), align 1 341 %ctpop6 = call i8 @llvm.ctpop.i8(i8 %ld6)
|
D | cttz.ll | 141 %ld6 = load i32, i32* getelementptr inbounds ([8 x i32], [8 x i32]* @src32, i32 0, i64 6), align 2 149 %cttz6 = call i32 @llvm.cttz.i32(i32 %ld6, i1 0) 196 …%ld6 = load i16, i16* getelementptr inbounds ([16 x i16], [16 x i16]* @src16, i16 0, i64 6), align… 204 %cttz6 = call i16 @llvm.cttz.i16(i16 %ld6, i1 0) 275 …%ld6 = load i16, i16* getelementptr inbounds ([16 x i16], [16 x i16]* @src16, i16 0, i64 6), ali… 291 %cttz6 = call i16 @llvm.cttz.i16(i16 %ld6, i1 0) 378 %ld6 = load i8, i8* getelementptr inbounds ([32 x i8], [32 x i8]* @src8, i8 0, i64 6), align 1 394 %cttz6 = call i8 @llvm.cttz.i8(i8 %ld6, i1 0) 529 %ld6 = load i8, i8* getelementptr inbounds ([32 x i8], [32 x i8]* @src8, i8 0, i64 6), align 1 561 %cttz6 = call i8 @llvm.cttz.i8(i8 %ld6, i1 0) [all …]
|
D | ctlz.ll | 141 %ld6 = load i32, i32* getelementptr inbounds ([8 x i32], [8 x i32]* @src32, i32 0, i64 6), align 2 149 %ctlz6 = call i32 @llvm.ctlz.i32(i32 %ld6, i1 0) 196 …%ld6 = load i16, i16* getelementptr inbounds ([16 x i16], [16 x i16]* @src16, i16 0, i64 6), align… 204 %ctlz6 = call i16 @llvm.ctlz.i16(i16 %ld6, i1 0) 275 …%ld6 = load i16, i16* getelementptr inbounds ([16 x i16], [16 x i16]* @src16, i16 0, i64 6), ali… 291 %ctlz6 = call i16 @llvm.ctlz.i16(i16 %ld6, i1 0) 378 %ld6 = load i8, i8* getelementptr inbounds ([32 x i8], [32 x i8]* @src8, i8 0, i64 6), align 1 394 %ctlz6 = call i8 @llvm.ctlz.i8(i8 %ld6, i1 0) 529 %ld6 = load i8, i8* getelementptr inbounds ([32 x i8], [32 x i8]* @src8, i8 0, i64 6), align 1 561 %ctlz6 = call i8 @llvm.ctlz.i8(i8 %ld6, i1 0) [all …]
|
D | fround.ll | 177 …%ld6 = load double, double* getelementptr inbounds ([8 x double], [8 x double]* @src64, i32 0, i64… 185 %ceil6 = call double @llvm.ceil.f64(double %ld6) 348 …%ld6 = load double, double* getelementptr inbounds ([8 x double], [8 x double]* @src64, i32 0, i64… 356 %floor6 = call double @llvm.floor.f64(double %ld6) 519 …%ld6 = load double, double* getelementptr inbounds ([8 x double], [8 x double]* @src64, i32 0, i64… 527 %nearbyint6 = call double @llvm.nearbyint.f64(double %ld6) 690 …%ld6 = load double, double* getelementptr inbounds ([8 x double], [8 x double]* @src64, i32 0, i64… 698 %rint6 = call double @llvm.rint.f64(double %ld6) 861 …%ld6 = load double, double* getelementptr inbounds ([8 x double], [8 x double]* @src64, i32 0, i64… 869 %trunc6 = call double @llvm.trunc.f64(double %ld6) [all …]
|
D | bitreverse.ll | 205 %ld6 = load i32, i32* getelementptr inbounds ([8 x i32], [8 x i32]* @src32, i32 0, i64 6), align 2 213 %bitreverse6 = call i32 @llvm.bitreverse.i32(i32 %ld6) 272 …%ld6 = load i16, i16* getelementptr inbounds ([16 x i16], [16 x i16]* @src16, i16 0, i64 6), align… 280 %bitreverse6 = call i16 @llvm.bitreverse.i16(i16 %ld6) 363 …%ld6 = load i16, i16* getelementptr inbounds ([16 x i16], [16 x i16]* @src16, i16 0, i64 6), ali… 379 %bitreverse6 = call i16 @llvm.bitreverse.i16(i16 %ld6) 478 %ld6 = load i8, i8* getelementptr inbounds ([32 x i8], [32 x i8]* @src8, i8 0, i64 6), align 1 494 %bitreverse6 = call i8 @llvm.bitreverse.i8(i8 %ld6) 647 %ld6 = load i8, i8* getelementptr inbounds ([32 x i8], [32 x i8]* @src8, i8 0, i64 6), align 1 679 %bitreverse6 = call i8 @llvm.bitreverse.i8(i8 %ld6)
|
/external/llvm/test/CodeGen/ARM/ |
D | spill-q.ll | 32 %ld6 = call <4 x float> @llvm.arm.neon.vld1.v4f32.p0i8(i8* undef, i32 1) nounwind 73 %tmp4 = fadd <4 x float> %tmp3, %ld6
|
/external/llvm/test/CodeGen/Thumb2/ |
D | thumb2-spill-q.ll | 32 %ld6 = call <4 x float> @llvm.arm.neon.vld1.v4f32.p0i8(i8* undef, i32 1) nounwind 73 %tmp4 = fadd <4 x float> %tmp3, %ld6
|
/external/llvm-project/libcxx/test/std/numerics/numbers/ |
D | specialize.pass.cpp | 64 [[maybe_unused]] long double ld6{std::numbers::ln2_v<long double>}; in tests() local
|
D | defined.pass.cpp | 63 [[maybe_unused]] const long double* ld6{&std::numbers::ln2_v<long double>}; in tests() local
|