/external/llvm/test/Transforms/InstSimplify/ |
D | fold-builtin-fma.ll | 7 declare double @llvm.fma.f64(double, double, double) 11 %1 = call double @llvm.fma.f64(double 7.0, double 8.0, double 0.0) 17 ; Test builtin fma with all finite non-zero constants. 19 %1 = call double @llvm.fma.f64(double 7.0, double 8.0, double 5.0) 25 ; Test builtin fma with a +/-NaN addend. 27 %1 = call double @llvm.fma.f64(double 7.0, double 8.0, double 0x7FF8000000000000) 34 %1 = call double @llvm.fma.f64(double 7.0, double 8.0, double 0xFFF8000000000000) 40 ; Test builtin fma with a +/-Inf addend. 42 %1 = call double @llvm.fma.f64(double 7.0, double 8.0, double 0x7FF0000000000000) 49 %1 = call double @llvm.fma.f64(double 7.0, double 8.0, double 0xFFF0000000000000) [all …]
|
/external/llvm/test/CodeGen/R600/ |
D | use-sgpr-multiple-times.ll | 4 declare float @llvm.fma.f32(float, float, float) #1 24 %fma = call float @llvm.fma.f32(float %a, float %a, float %a) #1 25 store float %fma, float addrspace(1)* %out, align 4 38 %fma = call float @llvm.fma.f32(float %a, float %a, float %b) #1 39 store float %fma, float addrspace(1)* %out, align 4 52 %fma = call float @llvm.fma.f32(float %a, float %b, float %a) #1 53 store float %fma, float addrspace(1)* %out, align 4 66 %fma = call float @llvm.fma.f32(float %b, float %a, float %a) #1 67 store float %fma, float addrspace(1)* %out, align 4 76 %fma = call float @llvm.fma.f32(float %a, float %a, float 2.0) #1 [all …]
|
D | fma.ll | 4 declare float @llvm.fma.f32(float, float, float) nounwind readnone 5 declare <2 x float> @llvm.fma.v2f32(<2 x float>, <2 x float>, <2 x float>) nounwind readnone 6 declare <4 x float> @llvm.fma.v4f32(<4 x float>, <4 x float>, <4 x float>) nounwind readnone 20 %r3 = tail call float @llvm.fma.f32(float %r0, float %r1, float %r2) 37 %r3 = tail call <2 x float> @llvm.fma.v2f32(<2 x float> %r0, <2 x float> %r1, <2 x float> %r2) 58 %r3 = tail call <4 x float> @llvm.fma.v4f32(<4 x float> %r0, <4 x float> %r1, <4 x float> %r2) 74 %fma = call float @llvm.fma.f32(float %a, float 2.0, float %b) 75 store float %fma, float addrspace(1)* %out.gep, align 4 89 %fma = call float @llvm.fma.f32(float %a, float %b, float %c) 90 store float %fma, float addrspace(1)* %out.gep, align 4
|
D | fma-combine.ll | 6 declare double @llvm.fma.f64(double, double, double) #0 7 declare float @llvm.fma.f32(float, float, float) #0 9 ; (fadd (fmul x, y), z) -> (fma x, y, z) 28 %fma = fadd double %mul, %c 29 store double %fma, double addrspace(1)* %gep.out 33 ; (fadd (fmul x, y), z) -> (fma x, y, z) 66 ; (fadd x, (fmul y, z)) -> (fma y, z, x) 85 %fma = fadd double %c, %mul 86 store double %fma, double addrspace(1)* %gep.out 90 ; (fsub (fmul x, y), z) -> (fma x, y, (fneg z)) [all …]
|
D | fma.f64.ll | 4 declare double @llvm.fma.f64(double, double, double) nounwind readnone 5 declare <2 x double> @llvm.fma.v2f64(<2 x double>, <2 x double>, <2 x double>) nounwind readnone 6 declare <4 x double> @llvm.fma.v4f64(<4 x double>, <4 x double>, <4 x double>) nounwind readnone 16 %r3 = tail call double @llvm.fma.f64(double %r0, double %r1, double %r2) 29 … %r3 = tail call <2 x double> @llvm.fma.v2f64(<2 x double> %r0, <2 x double> %r1, <2 x double> %r2) 44 … %r3 = tail call <4 x double> @llvm.fma.v4f64(<4 x double> %r0, <4 x double> %r1, <4 x double> %r2)
|
/external/llvm/test/CodeGen/X86/ |
D | fma-intrinsics-x86_64.ll | 1 ; RUN: llc < %s -mtriple=x86_64-unknown-unknown -march=x86-64 -mcpu=corei7-avx -mattr=+fma | FileCh… 2 ; RUN: llc < %s -mtriple=x86_64-unknown-unknown -march=x86-64 -mcpu=core-avx2 -mattr=+fma,+avx2 | F… 4 ; RUN: llc < %s -mtriple=x86_64-unknown-unknown -mcpu=bdver2 -mattr=+avx,-fma | FileCheck %s --chec… 10 …%res = call < 4 x float > @llvm.x86.fma.vfmadd.ss(< 4 x float > %a0, < 4 x float > %a1, < 4 x floa… 13 declare < 4 x float > @llvm.x86.fma.vfmadd.ss(< 4 x float >, < 4 x float >, < 4 x float >) nounwind… 18 …%res = call < 2 x double > @llvm.x86.fma.vfmadd.sd(< 2 x double > %a0, < 2 x double > %a1, < 2 x d… 21 declare < 2 x double > @llvm.x86.fma.vfmadd.sd(< 2 x double >, < 2 x double >, < 2 x double >) noun… 26 …%res = call < 4 x float > @llvm.x86.fma.vfmadd.ps(< 4 x float > %a0, < 4 x float > %a1, < 4 x floa… 29 declare < 4 x float > @llvm.x86.fma.vfmadd.ps(< 4 x float >, < 4 x float >, < 4 x float >) nounwind… 34 …%res = call < 2 x double > @llvm.x86.fma.vfmadd.pd(< 2 x double > %a0, < 2 x double > %a1, < 2 x d… [all …]
|
D | fma3-intrinsics.ll | 2 ; RUN: llc < %s -mtriple=x86_64-pc-win32 -mattr=+fma,+fma4 | FileCheck %s 7 …%res = call <4 x float> @llvm.x86.fma.vfmadd.ss(<4 x float> %a0, <4 x float> %a1, <4 x float> %a2)… 10 declare <4 x float> @llvm.x86.fma.vfmadd.ss(<4 x float>, <4 x float>, <4 x float>) nounwind readnone 14 …%res = call <4 x float> @llvm.x86.fma.vfmadd.ps(<4 x float> %a0, <4 x float> %a1, <4 x float> %a2)… 17 declare <4 x float> @llvm.x86.fma.vfmadd.ps(<4 x float>, <4 x float>, <4 x float>) nounwind readnone 21 …%res = call <8 x float> @llvm.x86.fma.vfmadd.ps.256(<8 x float> %a0, <8 x float> %a1, <8 x float> … 24 declare <8 x float> @llvm.x86.fma.vfmadd.ps.256(<8 x float>, <8 x float>, <8 x float>) nounwind rea… 28 …%res = call <4 x float> @llvm.x86.fma.vfnmadd.ss(<4 x float> %a0, <4 x float> %a1, <4 x float> %a2… 31 declare <4 x float> @llvm.x86.fma.vfnmadd.ss(<4 x float>, <4 x float>, <4 x float>) nounwind readno… 35 …%res = call <4 x float> @llvm.x86.fma.vfnmadd.ps(<4 x float> %a0, <4 x float> %a1, <4 x float> %a2… [all …]
|
D | fma.ll | 1 ; RUN: llc < %s -mtriple=i386-apple-darwin10 -mattr=+fma,-fma4 | FileCheck %s --check-prefix=CHEC… 2 ; RUN: llc < %s -mtriple=i386-apple-darwin10 -mattr=-fma,-fma4 | FileCheck %s --check-prefix=CHEC… 3 ; RUN: llc < %s -mtriple=x86_64-apple-darwin10 -mattr=+fma,-fma4 | FileCheck %s --check-prefix=CHEC… 4 ; RUN: llc < %s -mtriple=x86_64-apple-darwin10 -mattr=-fma,-fma4 | FileCheck %s --check-prefix=CHE… 6 ; RUN: llc < %s -march=x86 -mcpu=bdver2 -mattr=-fma,-fma4 | FileCheck %s --check-prefix=CHECK-FMA-C… 14 %call = tail call float @llvm.fma.f32(float %a, float %b, float %c) nounwind readnone 20 ; CHECK-FMA-CALL: fma 24 %call = tail call double @llvm.fma.f64(double %a, double %b, double %c) nounwind readnone 33 %call = tail call x86_fp80 @llvm.fma.f80(x86_fp80 %a, x86_fp80 %b, x86_fp80 %c) nounwind readnone 38 ; CHECK-NOT: fma [all …]
|
D | avx512-fma-intrinsics.ll | 3 declare <16 x float> @llvm.x86.fma.mask.vfmadd.ps.512(<16 x float>, <16 x float>, <16 x float>, i16… 4 declare <8 x double> @llvm.x86.fma.mask.vfmadd.pd.512(<8 x double>, <8 x double>, <8 x double>, i8,… 5 declare <16 x float> @llvm.x86.fma.mask.vfmsub.ps.512(<16 x float>, <16 x float>, <16 x float>, i16… 10 …%res = call <8 x double> @llvm.x86.fma.mask.vfmsub.pd.512(<8 x double> %a0, <8 x double> %a1, <8 x… 13 declare <8 x double> @llvm.x86.fma.mask.vfmsub.pd.512(<8 x double>, <8 x double>, <8 x double>, i8,… 18 …%res = call <8 x double> @llvm.x86.fma.mask.vfmsub.pd.512(<8 x double> %a0, <8 x double> %a1, <8 x… 25 …%res = call <16 x float> @llvm.x86.fma.mask.vfnmadd.ps.512(<16 x float> %a0, <16 x float> %a1, <16… 28 declare <16 x float> @llvm.x86.fma.mask.vfnmadd.ps.512(<16 x float>, <16 x float>, <16 x float>, i1… 33 …%res = call <16 x float> @llvm.x86.fma.mask.vfnmadd.ps.512(<16 x float> %a0, <16 x float> %a1, <16… 40 …%res = call <8 x double> @llvm.x86.fma.mask.vfnmadd.pd.512(<8 x double> %a0, <8 x double> %a1, <8 … [all …]
|
D | fma4-intrinsics-x86_64-folded-load.ll | 2 ; RUN: llc < %s -mtriple=x86_64-unknown-unknown -mcpu=bdver2 -mattr=+avx,-fma | FileCheck %s 9 …%res = call < 4 x float > @llvm.x86.fma.vfmadd.ss(< 4 x float > %a0, < 4 x float > %a1, < 4 x floa… 16 …%res = call < 4 x float > @llvm.x86.fma.vfmadd.ss(< 4 x float > %a0, < 4 x float > %y, < 4 x float… 20 declare < 4 x float > @llvm.x86.fma.vfmadd.ss(< 4 x float >, < 4 x float >, < 4 x float >) nounwind… 26 …%res = call < 2 x double > @llvm.x86.fma.vfmadd.sd(< 2 x double > %a0, < 2 x double > %a1, < 2 x d… 33 …%res = call < 2 x double > @llvm.x86.fma.vfmadd.sd(< 2 x double > %a0, < 2 x double > %y, < 2 x do… 36 declare < 2 x double > @llvm.x86.fma.vfmadd.sd(< 2 x double >, < 2 x double >, < 2 x double >) noun… 40 …%res = call < 4 x float > @llvm.x86.fma.vfmadd.ps(< 4 x float > %a0, < 4 x float > %a1, < 4 x floa… 46 …%res = call < 4 x float > @llvm.x86.fma.vfmadd.ps(< 4 x float > %a0, < 4 x float > %x, < 4 x float… 49 declare < 4 x float > @llvm.x86.fma.vfmadd.ps(< 4 x float >, < 4 x float >, < 4 x float >) nounwind… [all …]
|
D | fma-phi-213-to-231.ll | 26 …%0 = call <4 x double> @llvm.x86.fma.vfmaddsub.pd.256(<4 x double> %a, <4 x double> %b, <4 x doubl… 55 …%0 = call <4 x double> @llvm.x86.fma.vfmsubadd.pd.256(<4 x double> %a, <4 x double> %b, <4 x doubl… 84 …%0 = call <4 x double> @llvm.x86.fma.vfmadd.pd.256(<4 x double> %a, <4 x double> %b, <4 x double> … 113 …%0 = call <4 x double> @llvm.x86.fma.vfmsub.pd.256(<4 x double> %a, <4 x double> %b, <4 x double> … 121 declare <4 x double> @llvm.x86.fma.vfmaddsub.pd.256(<4 x double>, <4 x double>, <4 x double>) 122 declare <4 x double> @llvm.x86.fma.vfmsubadd.pd.256(<4 x double>, <4 x double>, <4 x double>) 123 declare <4 x double> @llvm.x86.fma.vfmadd.pd.256(<4 x double>, <4 x double>, <4 x double>) 124 declare <4 x double> @llvm.x86.fma.vfmsub.pd.256(<4 x double>, <4 x double>, <4 x double>) 148 …%0 = call <8 x float> @llvm.x86.fma.vfmaddsub.ps.256(<8 x float> %a, <8 x float> %b, <8 x float> %… 177 …%0 = call <8 x float> @llvm.x86.fma.vfmsubadd.ps.256(<8 x float> %a, <8 x float> %b, <8 x float> %… [all …]
|
/external/llvm/test/CodeGen/AArch64/ |
D | arm64-fmadd.ll | 7 %0 = tail call float @llvm.fma.f32(float %a, float %b, float %c) 15 %0 = tail call float @llvm.fma.f32(float %a, float %b, float %c) 25 %0 = tail call float @llvm.fma.f32(float %a, float %mul, float %c) 34 %0 = tail call float @llvm.fma.f32(float %mul, float %a, float %c) 43 %0 = tail call float @llvm.fma.f32(float %a, float %b, float %mul) 51 %0 = tail call double @llvm.fma.f64(double %a, double %b, double %c) 59 %0 = tail call double @llvm.fma.f64(double %a, double %b, double %c) 69 %0 = tail call double @llvm.fma.f64(double %a, double %mul, double %c) 78 %0 = tail call double @llvm.fma.f64(double %mul, double %a, double %c) 87 %0 = tail call double @llvm.fma.f64(double %a, double %b, double %mul) [all …]
|
D | neon-scalar-by-elem-fma.ll | 3 declare float @llvm.fma.f32(float, float, float) 4 declare double @llvm.fma.f64(double, double, double) 10 %tmp2 = call float @llvm.fma.f32(float %b, float %tmp1, float %a) 18 %tmp2 = call float @llvm.fma.f32(float %tmp1, float %a, float %a) 26 %tmp2 = call float @llvm.fma.f32(float %b, float %tmp1, float %a) 34 %tmp2 = call double @llvm.fma.f64(double %b, double %tmp1, double %a) 42 %tmp2 = call double @llvm.fma.f64(double %b, double %tmp1, double %a) 50 %tmp2 = call double @llvm.fma.f64(double %tmp1, double %b, double %a) 59 %tmp3 = call float @llvm.fma.f32(float %tmp2, float %tmp1, float %a) 68 %tmp3 = call float @llvm.fma.f32(float %tmp1, float %tmp2, float %a) [all …]
|
/external/llvm/test/CodeGen/ARM/ |
D | fusedMAC.ll | 106 %tmp1 = tail call float @llvm.fma.f32(float %a, float %b, float %c) nounwind readnone 114 %tmp1 = tail call double @llvm.fma.f64(double %a, double %b, double %c) nounwind readnone 122 …%tmp1 = tail call <2 x float> @llvm.fma.v2f32(<2 x float> %a, <2 x float> %b, <2 x float> %c) noun… 131 %tmp2 = tail call double @llvm.fma.f64(double %tmp1, double %b, double %c) nounwind readnone 140 %tmp2 = tail call double @llvm.fma.f64(double %a, double %tmp1, double %c) nounwind readnone 149 %tmp3 = tail call float @llvm.fma.f32(float %a, float %b, float %tmp2) nounwind readnone 158 %tmp2 = tail call double @llvm.fma.f64(double %tmp1, double %b, double %c) nounwind readnone 168 %tmp2 = tail call double @llvm.fma.f64(double %a, double %tmp1, double %c) nounwind readnone 177 %tmp1 = tail call double @llvm.fma.f64(double %a, double %b, double %c) nounwind readnone 188 %tmp3 = tail call double @llvm.fma.f64(double %tmp1, double %b, double %tmp2) nounwind readnone [all …]
|
D | fp-fast.ll | 10 %t2 = call float @llvm.fma.f32(float %x, float 2.0, float %t1) 20 %t2 = call float @llvm.fma.f32(float %t1, float 2.0, float %y) 29 %t2 = call float @llvm.fma.f32(float %x, float 1.0, float %y) 38 %t2 = call float @llvm.fma.f32(float %x, float -1.0, float %y) 47 %t2 = call float @llvm.fma.f32(float %x, float 2.0, float %x) 57 %t2 = call float @llvm.fma.f32(float %x, float 5.0, float %t1) 61 declare float @llvm.fma.f32(float, float, float)
|
/external/llvm/test/CodeGen/NVPTX/ |
D | fma-disable.ll | 1 ; RUN: llc < %s -march=nvptx -mcpu=sm_20 -nvptx-fma-level=1 | FileCheck %s -check-prefix=FMA 2 ; RUN: llc < %s -march=nvptx -mcpu=sm_20 -nvptx-fma-level=0 | FileCheck %s -check-prefix=MUL 3 ; RUN: llc < %s -march=nvptx64 -mcpu=sm_20 -nvptx-fma-level=1 | FileCheck %s -check-prefix=FMA 4 ; RUN: llc < %s -march=nvptx64 -mcpu=sm_20 -nvptx-fma-level=0 | FileCheck %s -check-prefix=MUL 8 ; FMA: fma.rn.f32 18 ; FMA: fma.rn.f64
|
D | fma.ll | 7 ; CHECK: fma.rn.f32 %f{{[0-9]+}}, %f{{[0-9]+}}, %f{{[0-9]+}}, %f{{[0-9]+}}; 15 ; CHECK: fma.rn.f32 %f{{[0-9]+}}, %f{{[0-9]+}}, %f{{[0-9]+}}, %f{{[0-9]+}}; 16 ; CHECK: fma.rn.f32 %f{{[0-9]+}}, %f{{[0-9]+}}, %f{{[0-9]+}}, %f{{[0-9]+}}; 26 ; CHECK: fma.rn.f64 %fd{{[0-9]+}}, %fd{{[0-9]+}}, %fd{{[0-9]+}}, %fd{{[0-9]+}}; 34 ; CHECK: fma.rn.f64 %fd{{[0-9]+}}, %fd{{[0-9]+}}, %fd{{[0-9]+}}, %fd{{[0-9]+}}; 35 ; CHECK: fma.rn.f64 %fd{{[0-9]+}}, %fd{{[0-9]+}}, %fd{{[0-9]+}}, %fd{{[0-9]+}};
|
/external/llvm/test/CodeGen/PowerPC/ |
D | vsx-fma-m.ll | 4 ; Also run with -schedule-ppc-vsx-fma-mutation-early as a stress test for the 6 ; RUN: llc < %s -mcpu=pwr7 -mattr=+vsx -schedule-ppc-vsx-fma-mutation-early 12 %0 = tail call double @llvm.fma.f64(double %b, double %c, double %a) 14 %1 = tail call double @llvm.fma.f64(double %b, double %e, double %a) 39 %0 = tail call double @llvm.fma.f64(double %b, double %c, double %a) 41 %1 = tail call double @llvm.fma.f64(double %b, double %e, double %a) 44 %2 = tail call double @llvm.fma.f64(double %b, double %f, double %a) 76 %0 = tail call double @llvm.fma.f64(double %b, double %c, double %a) 78 %1 = tail call double @llvm.fma.f64(double %b, double %e, double %a) 79 %2 = tail call double @llvm.fma.f64(double %b, double %c, double %1) [all …]
|
/external/llvm/test/ExecutionEngine/ |
D | fma3-jit.ll | 10 %fma = tail call double @llvm.fma.f64(double 3.0, double 3.0, double 3.0) nounwind readnone 13 call i32 (i8*,...) @printf(i8* %ptr1, double %fma) 18 declare double @llvm.fma.f64(double, double, double) nounwind readnone
|
/external/llvm/test/Transforms/InstCombine/ |
D | constant-fold-math.ll | 3 declare float @llvm.fma.f32(float, float, float) #0 5 declare <4 x float> @llvm.fma.v4f32(<4 x float>, <4 x float>, <4 x float>) #0 7 declare double @llvm.fma.f64(double, double, double) #0 16 %x = call float @llvm.fma.f32(float 1.0, float 2.0, float 4.0) #0 23 …%x = call <4 x float> @llvm.fma.v4f32(<4 x float> <float 1.0, float 2.0, float 3.0, float 4.0>, <4… 37 %x = call double @llvm.fma.f64(double 1.0, double 2.0, double 4.0) #0
|
/external/llvm/test/CodeGen/SystemZ/ |
D | fp-mul-07.ll | 3 declare double @llvm.fma.f64(double %f1, double %f2, double %f3) 10 %res = call double @llvm.fma.f64 (double %f1, double %f2, double %acc) 20 %res = call double @llvm.fma.f64 (double %f1, double %f2, double %acc) 31 %res = call double @llvm.fma.f64 (double %f1, double %f2, double %acc) 46 %res = call double @llvm.fma.f64 (double %f1, double %f2, double %acc) 61 %res = call double @llvm.fma.f64 (double %f1, double %f2, double %acc) 73 %res = call double @llvm.fma.f64 (double %f1, double %f2, double %acc) 86 %res = call double @llvm.fma.f64 (double %f1, double %f2, double %acc) 100 %res = call double @llvm.fma.f64 (double %f1, double %f2, double %acc)
|
D | fp-mul-06.ll | 3 declare float @llvm.fma.f32(float %f1, float %f2, float %f3) 10 %res = call float @llvm.fma.f32 (float %f1, float %f2, float %acc) 20 %res = call float @llvm.fma.f32 (float %f1, float %f2, float %acc) 31 %res = call float @llvm.fma.f32 (float %f1, float %f2, float %acc) 46 %res = call float @llvm.fma.f32 (float %f1, float %f2, float %acc) 61 %res = call float @llvm.fma.f32 (float %f1, float %f2, float %acc) 73 %res = call float @llvm.fma.f32 (float %f1, float %f2, float %acc) 86 %res = call float @llvm.fma.f32 (float %f1, float %f2, float %acc) 100 %res = call float @llvm.fma.f32 (float %f1, float %f2, float %acc)
|
D | fp-mul-08.ll | 3 declare float @llvm.fma.f32(float %f1, float %f2, float %f3) 11 %res = call float @llvm.fma.f32 (float %f1, float %f2, float %negacc) 22 %res = call float @llvm.fma.f32 (float %f1, float %f2, float %negacc) 34 %res = call float @llvm.fma.f32 (float %f1, float %f2, float %negacc) 50 %res = call float @llvm.fma.f32 (float %f1, float %f2, float %negacc) 66 %res = call float @llvm.fma.f32 (float %f1, float %f2, float %negacc) 79 %res = call float @llvm.fma.f32 (float %f1, float %f2, float %negacc) 93 %res = call float @llvm.fma.f32 (float %f1, float %f2, float %negacc) 108 %res = call float @llvm.fma.f32 (float %f1, float %f2, float %negacc)
|
D | fp-mul-09.ll | 3 declare double @llvm.fma.f64(double %f1, double %f2, double %f3) 11 %res = call double @llvm.fma.f64 (double %f1, double %f2, double %negacc) 22 %res = call double @llvm.fma.f64 (double %f1, double %f2, double %negacc) 34 %res = call double @llvm.fma.f64 (double %f1, double %f2, double %negacc) 50 %res = call double @llvm.fma.f64 (double %f1, double %f2, double %negacc) 66 %res = call double @llvm.fma.f64 (double %f1, double %f2, double %negacc) 79 %res = call double @llvm.fma.f64 (double %f1, double %f2, double %negacc) 93 %res = call double @llvm.fma.f64 (double %f1, double %f2, double %negacc) 108 %res = call double @llvm.fma.f64 (double %f1, double %f2, double %negacc)
|
/external/apache-xml/src/main/java/org/apache/xpath/functions/ |
D | FunctionMultiArgs.java | 215 FunctionMultiArgs fma = (FunctionMultiArgs) expr; in deepEquals() 219 if ((null == fma) || (fma.m_args.length != n)) in deepEquals() 224 if (!m_args[i].deepEquals(fma.m_args[i])) in deepEquals() 229 else if (null != fma.m_args) in deepEquals()
|