/external/python/cpython3/Lib/test/decimaltestdata/ |
D | dqFMA.decTest | 39 dqfma0001 fma 1 1 1 -> 2 40 dqfma0002 fma 1 1 2 -> 3 41 dqfma0003 fma 2 2 3 -> 7 42 dqfma0004 fma 9 9 9 -> 90 43 dqfma0005 fma -1 1 1 -> 0 44 dqfma0006 fma -1 1 2 -> 1 45 dqfma0007 fma -2 2 3 -> -1 46 dqfma0008 fma -9 9 9 -> -72 47 dqfma0011 fma 1 -1 1 -> 0 48 dqfma0012 fma 1 -1 2 -> 1 [all …]
|
D | ddFMA.decTest | 39 ddfma0001 fma 1 1 1 -> 2 40 ddfma0002 fma 1 1 2 -> 3 41 ddfma0003 fma 2 2 3 -> 7 42 ddfma0004 fma 9 9 9 -> 90 43 ddfma0005 fma -1 1 1 -> 0 44 ddfma0006 fma -1 1 2 -> 1 45 ddfma0007 fma -2 2 3 -> -1 46 ddfma0008 fma -9 9 9 -> -72 47 ddfma0011 fma 1 -1 1 -> 0 48 ddfma0012 fma 1 -1 2 -> 1 [all …]
|
D | fma.decTest | 2 -- fma.decTest -- decimal fused multiply add -- 38 fmax0001 fma 1 1 1 -> 2 39 fmax0002 fma 1 1 2 -> 3 40 fmax0003 fma 2 2 3 -> 7 41 fmax0004 fma 9 9 9 -> 90 42 fmax0005 fma -1 1 1 -> 0 43 fmax0006 fma -1 1 2 -> 1 44 fmax0007 fma -2 2 3 -> -1 45 fmax0008 fma -9 9 9 -> -72 46 fmax0011 fma 1 -1 1 -> 0 [all …]
|
/external/python/cpython2/Lib/test/decimaltestdata/ |
D | dqFMA.decTest | 39 dqfma0001 fma 1 1 1 -> 2 40 dqfma0002 fma 1 1 2 -> 3 41 dqfma0003 fma 2 2 3 -> 7 42 dqfma0004 fma 9 9 9 -> 90 43 dqfma0005 fma -1 1 1 -> 0 44 dqfma0006 fma -1 1 2 -> 1 45 dqfma0007 fma -2 2 3 -> -1 46 dqfma0008 fma -9 9 9 -> -72 47 dqfma0011 fma 1 -1 1 -> 0 48 dqfma0012 fma 1 -1 2 -> 1 [all …]
|
D | ddFMA.decTest | 39 ddfma0001 fma 1 1 1 -> 2 40 ddfma0002 fma 1 1 2 -> 3 41 ddfma0003 fma 2 2 3 -> 7 42 ddfma0004 fma 9 9 9 -> 90 43 ddfma0005 fma -1 1 1 -> 0 44 ddfma0006 fma -1 1 2 -> 1 45 ddfma0007 fma -2 2 3 -> -1 46 ddfma0008 fma -9 9 9 -> -72 47 ddfma0011 fma 1 -1 1 -> 0 48 ddfma0012 fma 1 -1 2 -> 1 [all …]
|
D | fma.decTest | 2 -- fma.decTest -- decimal fused multiply add -- 38 fmax0001 fma 1 1 1 -> 2 39 fmax0002 fma 1 1 2 -> 3 40 fmax0003 fma 2 2 3 -> 7 41 fmax0004 fma 9 9 9 -> 90 42 fmax0005 fma -1 1 1 -> 0 43 fmax0006 fma -1 1 2 -> 1 44 fmax0007 fma -2 2 3 -> -1 45 fmax0008 fma -9 9 9 -> -72 46 fmax0011 fma 1 -1 1 -> 0 [all …]
|
/external/swiftshader/third_party/llvm-7.0/llvm/test/Transforms/InstCombine/ |
D | fma.ll | 4 declare float @llvm.fma.f32(float, float, float) #1 5 declare <2 x float> @llvm.fma.v2f32(<2 x float>, <2 x float>, <2 x float>) #1 13 ; CHECK-NEXT: [[FMA:%.*]] = call float @llvm.fma.f32(float [[X:%.*]], float [[Y:%.*]], float [[Z… 18 %fma = call float @llvm.fma.f32(float %x.fneg, float %y.fneg, float %z) 19 ret float %fma 24 ; CHECK-NEXT: [[FMA:%.*]] = call <2 x float> @llvm.fma.v2f32(<2 x float> [[X:%.*]], <2 x float> … 29 %fma = call <2 x float> @llvm.fma.v2f32(<2 x float> %xn, <2 x float> %yn, <2 x float> %z) 30 ret <2 x float> %fma 35 ; CHECK-NEXT: [[FMA:%.*]] = call <2 x float> @llvm.fma.v2f32(<2 x float> [[X:%.*]], <2 x float> … 40 %fma = call <2 x float> @llvm.fma.v2f32(<2 x float> %xn, <2 x float> %yn, <2 x float> %z) [all …]
|
/external/swiftshader/third_party/llvm-7.0/llvm/test/Analysis/CostModel/X86/ |
D | arith-fma.ll | 3 …s-fp-math -cost-model -analyze -mtriple=x86_64-apple-macosx10.8.0 -mattr=+fma,+avx | FileCheck %s… 4 …s-fp-math -cost-model -analyze -mtriple=x86_64-apple-macosx10.8.0 -mattr=+fma,+avx2 | FileCheck %… 5 …s-fp-math -cost-model -analyze -mtriple=x86_64-apple-macosx10.8.0 -mattr=+fma,+avx512f | FileChec… 6 …s-fp-math -cost-model -analyze -mtriple=x86_64-apple-macosx10.8.0 -mattr=+fma,+avx512f,+avx512bw … 8 define i32 @fma(i32 %arg) { 9 ; XOP-LABEL: 'fma' 10 ; XOP-NEXT: Cost Model: Found an estimated cost of 1 for instruction: %F32 = call float @llvm.fma.… 11 …nd an estimated cost of 1 for instruction: %V4F32 = call <4 x float> @llvm.fma.v4f32(<4 x float> u… 12 …nd an estimated cost of 1 for instruction: %V8F32 = call <8 x float> @llvm.fma.v8f32(<8 x float> u… 13 … an estimated cost of 4 for instruction: %V16F32 = call <16 x float> @llvm.fma.v16f32(<16 x float>… [all …]
|
/external/swiftshader/third_party/llvm-7.0/llvm/test/Transforms/InstSimplify/ |
D | fold-builtin-fma.ll | 7 declare double @llvm.fma.f64(double, double, double) 11 %1 = call double @llvm.fma.f64(double 7.0, double 8.0, double 0.0) 17 ; Test builtin fma with all finite non-zero constants. 19 %1 = call double @llvm.fma.f64(double 7.0, double 8.0, double 5.0) 25 ; Test builtin fma with a +/-NaN addend. 27 %1 = call double @llvm.fma.f64(double 7.0, double 8.0, double 0x7FF8000000000000) 34 %1 = call double @llvm.fma.f64(double 7.0, double 8.0, double 0xFFF8000000000000) 40 ; Test builtin fma with a +/-Inf addend. 42 %1 = call double @llvm.fma.f64(double 7.0, double 8.0, double 0x7FF0000000000000) 49 %1 = call double @llvm.fma.f64(double 7.0, double 8.0, double 0xFFF0000000000000) [all …]
|
/external/llvm/test/Transforms/InstSimplify/ |
D | fold-builtin-fma.ll | 7 declare double @llvm.fma.f64(double, double, double) 11 %1 = call double @llvm.fma.f64(double 7.0, double 8.0, double 0.0) 17 ; Test builtin fma with all finite non-zero constants. 19 %1 = call double @llvm.fma.f64(double 7.0, double 8.0, double 5.0) 25 ; Test builtin fma with a +/-NaN addend. 27 %1 = call double @llvm.fma.f64(double 7.0, double 8.0, double 0x7FF8000000000000) 34 %1 = call double @llvm.fma.f64(double 7.0, double 8.0, double 0xFFF8000000000000) 40 ; Test builtin fma with a +/-Inf addend. 42 %1 = call double @llvm.fma.f64(double 7.0, double 8.0, double 0x7FF0000000000000) 49 %1 = call double @llvm.fma.f64(double 7.0, double 8.0, double 0xFFF0000000000000) [all …]
|
/external/swiftshader/third_party/llvm-7.0/llvm/test/CodeGen/AMDGPU/ |
D | use-sgpr-multiple-times.ll | 4 declare float @llvm.fma.f32(float, float, float) #1 5 declare double @llvm.fma.f64(double, double, double) #1 25 %fma = call float @llvm.fma.f32(float %a, float %a, float %a) #1 26 store float %fma, float addrspace(1)* %out, align 4 37 %fma = call float @llvm.fma.f32(float %a, float %a, float %b) #1 38 store float %fma, float addrspace(1)* %out, align 4 63 %fma0 = call float @llvm.fma.f32(float %a, float %va0, float %b) #1 64 %fma1 = call float @llvm.fma.f32(float %a, float %va1, float %b) #1 77 %fma = call float @llvm.fma.f32(float %a, float %b, float %a) #1 78 store float %fma, float addrspace(1)* %out, align 4 [all …]
|
D | coalescer_remat.ll | 3 declare float @llvm.fma.f32(float, float, float) 23 %v0 = phi float [0.0, %entry], [%fma.0, %loop] 24 %v1 = phi float [0.0, %entry], [%fma.1, %loop] 25 %v2 = phi float [0.0, %entry], [%fma.2, %loop] 26 %v3 = phi float [0.0, %entry], [%fma.3, %loop] 37 %fma.0 = call float @llvm.fma.f32(float %v0, float %load.0, float %v0) 38 %fma.1 = call float @llvm.fma.f32(float %v1, float %load.1, float %v1) 39 %fma.2 = call float @llvm.fma.f32(float %v2, float %load.2, float %v2) 40 %fma.3 = call float @llvm.fma.f32(float %v3, float %load.3, float %v3) 47 %ev0 = phi float [0.0, %entry], [%fma.0, %loop] [all …]
|
D | fpext-free.ll | 6 ; fold (fadd (fpext (fmul x, y)), z) -> (fma (fpext x), (fpext y), z) 51 ; fold (fadd x, (fpext (fmul y, z))) -> (fma (fpext y), (fpext z), x) 69 ; fold (fadd (fma x, y, (fpext (fmul u, v))), z) 70 ; -> (fma x, y, (fma (fpext u), (fpext v), z)) 88 %fma = call float @llvm.fmuladd.f32(float %x, float %y, float %mul.ext) 89 %add = fadd float %fma, %z 93 ; fold (fadd x, (fma y, z, (fpext (fmul u, v))) 94 ; -> (fma y, z, (fma (fpext u), (fpext v), x)) 111 %fma = call float @llvm.fmuladd.f32(float %x, float %y, float %mul.ext) 112 %add = fadd float %z, %fma [all …]
|
/external/llvm/test/CodeGen/AMDGPU/ |
D | use-sgpr-multiple-times.ll | 4 declare float @llvm.fma.f32(float, float, float) #1 5 declare double @llvm.fma.f64(double, double, double) #1 25 %fma = call float @llvm.fma.f32(float %a, float %a, float %a) #1 26 store float %fma, float addrspace(1)* %out, align 4 39 %fma = call float @llvm.fma.f32(float %a, float %a, float %b) #1 40 store float %fma, float addrspace(1)* %out, align 4 63 %fma0 = call float @llvm.fma.f32(float %a, float %va0, float %b) #1 64 %fma1 = call float @llvm.fma.f32(float %a, float %va1, float %b) #1 79 %fma = call float @llvm.fma.f32(float %a, float %b, float %a) #1 80 store float %fma, float addrspace(1)* %out, align 4 [all …]
|
D | coalescer_remat.ll | 3 declare float @llvm.fma.f32(float, float, float) 23 %v0 = phi float [0.0, %entry], [%fma.0, %loop] 24 %v1 = phi float [0.0, %entry], [%fma.1, %loop] 25 %v2 = phi float [0.0, %entry], [%fma.2, %loop] 26 %v3 = phi float [0.0, %entry], [%fma.3, %loop] 37 %fma.0 = call float @llvm.fma.f32(float %v0, float %load.0, float %v0) 38 %fma.1 = call float @llvm.fma.f32(float %v1, float %load.1, float %v1) 39 %fma.2 = call float @llvm.fma.f32(float %v2, float %load.2, float %v2) 40 %fma.3 = call float @llvm.fma.f32(float %v3, float %load.3, float %v3) 47 %ev0 = phi float [0.0, %entry], [%fma.0, %loop] [all …]
|
/external/swiftshader/third_party/llvm-7.0/llvm/test/CodeGen/X86/ |
D | fma4-intrinsics-x86-upgrade.ll | 2 ; RUN: llc < %s -mtriple=x86_64-unknown-unknown -mattr=+fma4,-fma -show-mc-encoding | FileCheck %s … 3 ; RUN: llc < %s -mtriple=x86_64-unknown-unknown -mattr=+fma4,+fma -show-mc-encoding | FileCheck %s … 10 %res = call <4 x float> @llvm.x86.fma.vfmadd.ps(<4 x float> %a0, <4 x float> %a1, <4 x float> %a2) 13 declare <4 x float> @llvm.x86.fma.vfmadd.ps(<4 x float>, <4 x float>, <4 x float>) 20 …%res = call <2 x double> @llvm.x86.fma.vfmadd.pd(<2 x double> %a0, <2 x double> %a1, <2 x double> … 23 declare <2 x double> @llvm.x86.fma.vfmadd.pd(<2 x double>, <2 x double>, <2 x double>) 30 …%res = call <8 x float> @llvm.x86.fma.vfmadd.ps.256(<8 x float> %a0, <8 x float> %a1, <8 x float> … 33 declare <8 x float> @llvm.x86.fma.vfmadd.ps.256(<8 x float>, <8 x float>, <8 x float>) 40 …%res = call <4 x double> @llvm.x86.fma.vfmadd.pd.256(<4 x double> %a0, <4 x double> %a1, <4 x doub… 43 declare <4 x double> @llvm.x86.fma.vfmadd.pd.256(<4 x double>, <4 x double>, <4 x double>) [all …]
|
D | fma-commute-x86.ll | 3 ; RUN: llc < %s -mtriple=x86_64-pc-win32 -mattr=+fma | FileCheck %s --check-prefix=FMA 8 declare <4 x float> @llvm.x86.fma.vfmadd.ss(<4 x float>, <4 x float>, <4 x float>) nounwind readnone 16 …%res = call <4 x float> @llvm.x86.fma.vfmadd.ss(<4 x float> %b, <4 x float> %a, <4 x float> %a) no… 26 …%res = call <4 x float> @llvm.x86.fma.vfmadd.ss(<4 x float> %a, <4 x float> %b, <4 x float> %a) no… 36 …%res = call <4 x float> @llvm.x86.fma.vfmadd.ss(<4 x float> %b, <4 x float> %b, <4 x float> %a) no… 40 declare <4 x float> @llvm.x86.fma.vfmadd.ps(<4 x float>, <4 x float>, <4 x float>) nounwind readnone 47 …%res = call <4 x float> @llvm.x86.fma.vfmadd.ps(<4 x float> %b, <4 x float> %a, <4 x float> %a) no… 57 …%res = call <4 x float> @llvm.x86.fma.vfmadd.ps(<4 x float> %a, <4 x float> %b, <4 x float> %a) no… 67 …%res = call <4 x float> @llvm.x86.fma.vfmadd.ps(<4 x float> %b, <4 x float> %b, <4 x float> %a) no… 71 declare <8 x float> @llvm.x86.fma.vfmadd.ps.256(<8 x float>, <8 x float>, <8 x float>) nounwind rea… [all …]
|
D | fma4-commute-x86.ll | 37 declare <4 x float> @llvm.x86.fma.vfmadd.ps(<4 x float>, <4 x float>, <4 x float>) nounwind readnone 44 …%res = call <4 x float> @llvm.x86.fma.vfmadd.ps(<4 x float> %b, <4 x float> %a, <4 x float> %a) no… 54 …%res = call <4 x float> @llvm.x86.fma.vfmadd.ps(<4 x float> %a, <4 x float> %b, <4 x float> %a) no… 64 …%res = call <4 x float> @llvm.x86.fma.vfmadd.ps(<4 x float> %b, <4 x float> %b, <4 x float> %a) no… 68 declare <8 x float> @llvm.x86.fma.vfmadd.ps.256(<8 x float>, <8 x float>, <8 x float>) nounwind rea… 75 …%res = call <8 x float> @llvm.x86.fma.vfmadd.ps.256(<8 x float> %b, <8 x float> %a, <8 x float> %a… 85 …%res = call <8 x float> @llvm.x86.fma.vfmadd.ps.256(<8 x float> %a, <8 x float> %b, <8 x float> %a… 95 …%res = call <8 x float> @llvm.x86.fma.vfmadd.ps.256(<8 x float> %b, <8 x float> %b, <8 x float> %a… 130 declare <2 x double> @llvm.x86.fma.vfmadd.pd(<2 x double>, <2 x double>, <2 x double>) nounwind rea… 137 …%res = call <2 x double> @llvm.x86.fma.vfmadd.pd(<2 x double> %b, <2 x double> %a, <2 x double> %a… [all …]
|
D | fma-intrinsics-fast-isel.ll | 2 ; RUN: llc < %s -fast-isel -mtriple=x86_64-unknown-unknown -mattr=+fma,-fma4 | FileCheck %s --check… 10 %0 = tail call <4 x float> @llvm.fma.v4f32(<4 x float> %a, <4 x float> %b, <4 x float> %c) #2 20 %0 = tail call <2 x double> @llvm.fma.v2f64(<2 x double> %a, <2 x double> %b, <2 x double> %c) #2 33 %3 = tail call float @llvm.fma.f32(float %0, float %1, float %2) #2 47 %3 = tail call double @llvm.fma.f64(double %0, double %1, double %2) #2 60 %0 = tail call <4 x float> @llvm.fma.v4f32(<4 x float> %a, <4 x float> %b, <4 x float> %sub.i) #2 72 …%0 = tail call <2 x double> @llvm.fma.v2f64(<2 x double> %a, <2 x double> %b, <2 x double> %sub.i)… 86 %3 = tail call float @llvm.fma.f32(float %0, float %1, float %2) #2 101 %3 = tail call double @llvm.fma.f64(double %0, double %1, double %2) #2 114 %0 = tail call <4 x float> @llvm.fma.v4f32(<4 x float> %sub.i, <4 x float> %b, <4 x float> %c) #2 [all …]
|
/external/llvm/test/CodeGen/X86/ |
D | fma.ll | 1 ; RUN: llc < %s -mtriple=i386-apple-darwin10 -mattr=+fma,-fma4 | FileCheck %s --check-prefix=CHEC… 2 ; RUN: llc < %s -mtriple=i386-apple-darwin10 -mattr=-fma,-fma4 | FileCheck %s --check-prefix=CHEC… 3 ; RUN: llc < %s -mtriple=x86_64-apple-darwin10 -mattr=+fma,-fma4 | FileCheck %s --check-prefix=CHEC… 4 ; RUN: llc < %s -mtriple=x86_64-apple-darwin10 -mattr=-fma,-fma4 | FileCheck %s --check-prefix=CHE… 5 ; RUN: llc < %s -mtriple=x86_64-apple-darwin10 -mattr=+avx512f,-fma,-fma4 | FileCheck %s --check-p… 7 ; RUN: llc < %s -march=x86 -mcpu=bdver2 -mattr=-fma,-fma4 | FileCheck %s --check-prefix=CHECK --che… 14 %call = call float @llvm.fma.f32(float %a, float %b, float %c) 20 ; CHECK-FMA-CALL: fma 23 %call = call double @llvm.fma.f64(double %a, double %b, double %c) 31 %call = call x86_fp80 @llvm.fma.f80(x86_fp80 %a, x86_fp80 %b, x86_fp80 %c) [all …]
|
D | fma-commute-x86.ll | 2 ; RUN: llc < %s -mtriple=x86_64-pc-win32 -mattr=+fma | FileCheck %s 7 declare <4 x float> @llvm.x86.fma.vfmadd.ss(<4 x float>, <4 x float>, <4 x float>) nounwind readnone 15 …%res = call <4 x float> @llvm.x86.fma.vfmadd.ss(<4 x float> %b, <4 x float> %a, <4 x float> %a) no… 25 …%res = call <4 x float> @llvm.x86.fma.vfmadd.ss(<4 x float> %a, <4 x float> %b, <4 x float> %a) no… 35 …%res = call <4 x float> @llvm.x86.fma.vfmadd.ss(<4 x float> %b, <4 x float> %b, <4 x float> %a) no… 39 declare <4 x float> @llvm.x86.fma.vfmadd.ps(<4 x float>, <4 x float>, <4 x float>) nounwind readnone 46 …%res = call <4 x float> @llvm.x86.fma.vfmadd.ps(<4 x float> %b, <4 x float> %a, <4 x float> %a) no… 56 …%res = call <4 x float> @llvm.x86.fma.vfmadd.ps(<4 x float> %a, <4 x float> %b, <4 x float> %a) no… 66 …%res = call <4 x float> @llvm.x86.fma.vfmadd.ps(<4 x float> %b, <4 x float> %b, <4 x float> %a) no… 70 declare <8 x float> @llvm.x86.fma.vfmadd.ps.256(<8 x float>, <8 x float>, <8 x float>) nounwind rea… [all …]
|
D | fma4-intrinsics-x86_64-folded-load.ll | 2 ; RUN: llc < %s -mtriple=x86_64-unknown-unknown -mcpu=bdver2 -mattr=+avx,-fma | FileCheck %s 9 …%res = call < 4 x float > @llvm.x86.fma.vfmadd.ss(< 4 x float > %a0, < 4 x float > %a1, < 4 x floa… 16 …%res = call < 4 x float > @llvm.x86.fma.vfmadd.ss(< 4 x float > %a0, < 4 x float > %y, < 4 x float… 20 declare < 4 x float > @llvm.x86.fma.vfmadd.ss(< 4 x float >, < 4 x float >, < 4 x float >) nounwind… 26 …%res = call < 2 x double > @llvm.x86.fma.vfmadd.sd(< 2 x double > %a0, < 2 x double > %a1, < 2 x d… 33 …%res = call < 2 x double > @llvm.x86.fma.vfmadd.sd(< 2 x double > %a0, < 2 x double > %y, < 2 x do… 36 declare < 2 x double > @llvm.x86.fma.vfmadd.sd(< 2 x double >, < 2 x double >, < 2 x double >) noun… 40 …%res = call < 4 x float > @llvm.x86.fma.vfmadd.ps(< 4 x float > %a0, < 4 x float > %a1, < 4 x floa… 46 …%res = call < 4 x float > @llvm.x86.fma.vfmadd.ps(< 4 x float > %a0, < 4 x float > %x, < 4 x float… 49 declare < 4 x float > @llvm.x86.fma.vfmadd.ps(< 4 x float >, < 4 x float >, < 4 x float >) nounwind… [all …]
|
D | fma-intrinsics-x86.ll | 1 ; RUN: llc < %s -mtriple=x86_64-unknown-unknown -march=x86-64 -mcpu=corei7-avx -mattr=+fma | FileCh… 2 ; RUN: llc < %s -mtriple=x86_64-unknown-unknown -march=x86-64 -mcpu=core-avx2 -mattr=+fma,+avx2 | F… 3 ; RUN: llc < %s -mtriple=x86_64-pc-windows -march=x86-64 -mcpu=core-avx2 -mattr=+fma,+avx2 | FileCh… 5 ; RUN: llc < %s -mtriple=x86_64-unknown-unknown -mcpu=bdver2 -mattr=+avx,-fma | FileCheck %s --chec… 22 %res = call <4 x float> @llvm.x86.fma.vfmadd.ss(<4 x float> %a0, <4 x float> %a1, <4 x float> %a2) 39 %res = call <4 x float> @llvm.x86.fma.vfmadd.ss(<4 x float> %a1, <4 x float> %a0, <4 x float> %a2) 42 declare <4 x float> @llvm.x86.fma.vfmadd.ss(<4 x float>, <4 x float>, <4 x float>) 57 …%res = call <2 x double> @llvm.x86.fma.vfmadd.sd(<2 x double> %a0, <2 x double> %a1, <2 x double> … 75 …%res = call <2 x double> @llvm.x86.fma.vfmadd.sd(<2 x double> %a1, <2 x double> %a0, <2 x double> … 78 declare <2 x double> @llvm.x86.fma.vfmadd.sd(<2 x double>, <2 x double>, <2 x double>) [all …]
|
/external/llvm/test/CodeGen/AArch64/ |
D | arm64-fmadd.ll | 7 %0 = tail call float @llvm.fma.f32(float %a, float %b, float %c) 15 %0 = tail call float @llvm.fma.f32(float %a, float %b, float %c) 25 %0 = tail call float @llvm.fma.f32(float %a, float %mul, float %c) 34 %0 = tail call float @llvm.fma.f32(float %mul, float %a, float %c) 43 %0 = tail call float @llvm.fma.f32(float %a, float %b, float %mul) 51 %0 = tail call double @llvm.fma.f64(double %a, double %b, double %c) 59 %0 = tail call double @llvm.fma.f64(double %a, double %b, double %c) 69 %0 = tail call double @llvm.fma.f64(double %a, double %mul, double %c) 78 %0 = tail call double @llvm.fma.f64(double %mul, double %a, double %c) 87 %0 = tail call double @llvm.fma.f64(double %a, double %b, double %mul) [all …]
|
/external/swiftshader/third_party/llvm-7.0/llvm/test/CodeGen/AArch64/ |
D | arm64-fmadd.ll | 7 %0 = tail call float @llvm.fma.f32(float %a, float %b, float %c) 15 %0 = tail call float @llvm.fma.f32(float %a, float %b, float %c) 25 %0 = tail call float @llvm.fma.f32(float %a, float %mul, float %c) 34 %0 = tail call float @llvm.fma.f32(float %mul, float %a, float %c) 43 %0 = tail call float @llvm.fma.f32(float %a, float %b, float %mul) 51 %0 = tail call double @llvm.fma.f64(double %a, double %b, double %c) 59 %0 = tail call double @llvm.fma.f64(double %a, double %b, double %c) 69 %0 = tail call double @llvm.fma.f64(double %a, double %mul, double %c) 78 %0 = tail call double @llvm.fma.f64(double %mul, double %a, double %c) 87 %0 = tail call double @llvm.fma.f64(double %a, double %b, double %mul) [all …]
|