/external/llvm/test/CodeGen/AMDGPU/ |
D | use-sgpr-multiple-times.ll | 64 %fma1 = call float @llvm.fma.f32(float %a, float %va1, float %b) #1 66 store volatile float %fma1, float addrspace(1)* %out 162 %fma1 = call float @llvm.fma.f32(float 1024.0, float 1024.0, float %b) #1 164 store volatile float %fma1, float addrspace(1)* %out 190 %fma1 = call float @llvm.fma.f32(float 1024.0, float %b, float 1024.0) #1 192 store volatile float %fma1, float addrspace(1)* %out 218 %fma1 = call float @llvm.fma.f32(float %b, float 1024.0, float 1024.0) #1 220 store volatile float %fma1, float addrspace(1)* %out 238 %fma1 = call float @llvm.fma.f32(float %a, float %b, float 4096.0) #1 240 store volatile float %fma1, float addrspace(1)* %out [all …]
|
D | fma-combine.ll | 60 %fma1 = fadd double %mul, %d 62 store volatile double %fma1, double addrspace(1)* %gep.out.1 141 %fma1 = fsub double %mul, %d 143 store volatile double %fma1, double addrspace(1)* %gep.out.1 198 %fma1 = fsub double %d, %mul 200 store volatile double %fma1, double addrspace(1)* %gep.out.1 257 %fma1 = fsub double %mul.neg, %d 260 store volatile double %fma1, double addrspace(1)* %gep.out.1 291 %fma1 = fsub double %mul, %d 294 store volatile double %fma1, double addrspace(1)* %gep.out.1
|
D | mad-combine.ll | 89 %fma1 = fadd float %mul, %d 92 store volatile float %fma1, float addrspace(1)* %gep.out.1 193 %fma1 = fsub float %mul, %d 195 store volatile float %fma1, float addrspace(1)* %gep.out.1 264 %fma1 = fsub float %d, %mul 266 store volatile float %fma1, float addrspace(1)* %gep.out.1 339 %fma1 = fsub float %mul.neg, %d 342 store volatile float %fma1, float addrspace(1)* %gep.out.1 382 %fma1 = fsub float %mul, %d 385 store volatile float %fma1, float addrspace(1)* %gep.out.1
|
/external/llvm-project/llvm/test/CodeGen/AMDGPU/ |
D | use-sgpr-multiple-times.ll | 64 %fma1 = call float @llvm.fma.f32(float %a, float %va1, float %b) #1 66 store volatile float %fma1, float addrspace(1)* %out 160 %fma1 = call float @llvm.fma.f32(float 1024.0, float 1024.0, float %b) #1 162 store volatile float %fma1, float addrspace(1)* %out 190 %fma1 = call float @llvm.fma.f32(float 1024.0, float %b, float 1024.0) #1 192 store volatile float %fma1, float addrspace(1)* %out 220 %fma1 = call float @llvm.fma.f32(float %b, float 1024.0, float 1024.0) #1 222 store volatile float %fma1, float addrspace(1)* %out 240 %fma1 = call float @llvm.fma.f32(float %a, float %b, float 4096.0) #1 242 store volatile float %fma1, float addrspace(1)* %out [all …]
|
D | mad-combine.ll | 91 %fma1 = fadd float %mul, %d 94 store volatile float %fma1, float addrspace(1)* %gep.out.1 195 %fma1 = fsub float %mul, %d 197 store volatile float %fma1, float addrspace(1)* %gep.out.1 267 %fma1 = fsub float %d, %mul 269 store volatile float %fma1, float addrspace(1)* %gep.out.1 343 %fma1 = fsub float %mul.neg, %d 346 store volatile float %fma1, float addrspace(1)* %gep.out.1 387 %fma1 = fsub float %mul, %d 390 store volatile float %fma1, float addrspace(1)* %gep.out.1
|
D | fma-combine.ll | 68 %fma1 = fadd double %mul, %d 70 store volatile double %fma1, double addrspace(1)* %gep.out.1 149 %fma1 = fsub double %mul, %d 151 store volatile double %fma1, double addrspace(1)* %gep.out.1 206 %fma1 = fsub double %d, %mul 208 store volatile double %fma1, double addrspace(1)* %gep.out.1 266 %fma1 = fsub double %mul.neg, %d 269 store volatile double %fma1, double addrspace(1)* %gep.out.1 301 %fma1 = fsub double %mul, %d 304 store volatile double %fma1, double addrspace(1)* %gep.out.1
|
D | fneg-combines.ll | 2333 %fma1 = call float @llvm.fma.f32(float %fneg.a, float %c, float 2.0) 2336 store volatile float %fma1, float addrspace(1)* %out
|
/external/llvm-project/llvm/test/Transforms/SLPVectorizer/AMDGPU/ |
D | packed-math.ll | 106 %fma1 = call half @llvm.fma.f16(half %i3, half %i4, half %i5) 109 store half %fma1, half addrspace(3)* %arrayidx6, align 2 183 %fma1 = call half @llvm.fma.f16(half %i3.fabs, half %i4, half %i5) 186 store half %fma1, half addrspace(3)* %arrayidx6, align 2 219 %fma1 = call half @llvm.fma.f16(half %i3, half %i4, half %i5) 222 store half %fma1, half addrspace(3)* %arrayidx6, align 2
|
/external/llvm-project/llvm/test/Transforms/SLPVectorizer/X86/ |
D | fma.ll | 56 %fma1 = call double @llvm.fma.f64(double %a1, double %b1, double %c1) 58 …store double %fma1, double* getelementptr inbounds ([8 x double], [8 x double]* @dst64, i32 0, i64… 107 %fma1 = call double @llvm.fma.f64(double %a1, double %b1, double %c1) 111 …store double %fma1, double* getelementptr inbounds ([8 x double], [8 x double]* @dst64, i32 0, i64… 207 %fma1 = call double @llvm.fma.f64(double %a1, double %b1, double %c1) 215 …store double %fma1, double* getelementptr inbounds ([8 x double], [8 x double]* @dst64, i32 0, i64… 270 %fma1 = call float @llvm.fma.f32(float %a1, float %b1, float %c1) 274 …store float %fma1, float* getelementptr inbounds ([16 x float], [16 x float]* @dst32, i32 0, i64 1… 357 %fma1 = call float @llvm.fma.f32(float %a1, float %b1, float %c1) 365 …store float %fma1, float* getelementptr inbounds ([16 x float], [16 x float]* @dst32, i32 0, i64 1… [all …]
|
/external/llvm/test/Transforms/SLPVectorizer/X86/ |
D | fma.ll | 55 %fma1 = call double @llvm.fma.f64(double %a1, double %b1, double %c1) 57 …store double %fma1, double* getelementptr inbounds ([8 x double], [8 x double]* @dst64, i32 0, i64… 106 %fma1 = call double @llvm.fma.f64(double %a1, double %b1, double %c1) 110 …store double %fma1, double* getelementptr inbounds ([8 x double], [8 x double]* @dst64, i32 0, i64… 206 %fma1 = call double @llvm.fma.f64(double %a1, double %b1, double %c1) 214 …store double %fma1, double* getelementptr inbounds ([8 x double], [8 x double]* @dst64, i32 0, i64… 269 %fma1 = call float @llvm.fma.f32(float %a1, float %b1, float %c1) 273 …store float %fma1, float* getelementptr inbounds ([16 x float], [16 x float]* @dst32, i32 0, i64 1… 356 %fma1 = call float @llvm.fma.f32(float %a1, float %b1, float %c1) 364 …store float %fma1, float* getelementptr inbounds ([16 x float], [16 x float]* @dst32, i32 0, i64 1… [all …]
|
/external/llvm-project/llvm/test/CodeGen/PowerPC/ |
D | fma-combine.ll | 239 %fma1 = call reassoc nsz double @llvm.fma.f64(double %fma, double %y, double %add) 240 ret double %fma1
|
/external/mesa3d/src/panfrost/bifrost/ |
D | bifrost_compile.c | 1692 bi_instruction fma1 = { in bi_lower_cube_coord() local 1707 .src = { fma1.dest, cube_ssel.dest, BIR_INDEX_CONSTANT | 0 }, in bi_lower_cube_coord() 1718 .src = { fma1.dest, cube_tsel.dest, BIR_INDEX_CONSTANT | 0 }, in bi_lower_cube_coord() 1728 bi_emit(ctx, fma1); in bi_lower_cube_coord()
|
/external/llvm-project/llvm/test/CodeGen/Thumb2/ |
D | mve-fma-loops.ll | 131 define arm_aapcs_vfpcc void @fma1(float* nocapture readonly %x, float* nocapture readonly %y, float… 132 ; CHECK-LABEL: fma1:
|