Home
last modified time | relevance | path

Searched refs:vmla_f32 (Results 1 – 25 of 30) sorted by relevance

12

/external/XNNPACK/src/f32-gemm/gen/
D4x2-minmax-neon-lane-ld64.c94 vacc0x01 = vmla_f32(vacc0x01, va0, vb01); in xnn_f32_gemm_minmax_ukernel_4x2__neon_lane_ld64()
95 vacc1x01 = vmla_f32(vacc1x01, va1, vb01); in xnn_f32_gemm_minmax_ukernel_4x2__neon_lane_ld64()
96 vacc2x01 = vmla_f32(vacc2x01, va2, vb01); in xnn_f32_gemm_minmax_ukernel_4x2__neon_lane_ld64()
97 vacc3x01 = vmla_f32(vacc3x01, va3, vb01); in xnn_f32_gemm_minmax_ukernel_4x2__neon_lane_ld64()
/external/XNNPACK/src/f32-ibilinear-chw/gen/
Dneon-p4.c117 const float32x2_t vl = vmla_f32(vtl, vld, valphav); in xnn_f32_ibilinear_chw_ukernel__neon_p4()
118 const float32x2_t vr = vmla_f32(vtr, vrd, valphav); in xnn_f32_ibilinear_chw_ukernel__neon_p4()
121 const float32x2_t vo = vmla_f32(vl, vd, valphah); in xnn_f32_ibilinear_chw_ukernel__neon_p4()
155 const float32x2_t vlr = vmla_f32(vtltr, vldrd, valphav); in xnn_f32_ibilinear_chw_ukernel__neon_p4()
Dneon-p8.c205 const float32x2_t vl = vmla_f32(vtl, vld, valphav); in xnn_f32_ibilinear_chw_ukernel__neon_p8()
206 const float32x2_t vr = vmla_f32(vtr, vrd, valphav); in xnn_f32_ibilinear_chw_ukernel__neon_p8()
209 const float32x2_t vo = vmla_f32(vl, vd, valphah); in xnn_f32_ibilinear_chw_ukernel__neon_p8()
243 const float32x2_t vlr = vmla_f32(vtltr, vldrd, valphav); in xnn_f32_ibilinear_chw_ukernel__neon_p8()
/external/XNNPACK/src/f32-igemm/gen/
D4x2-minmax-neon-lane-ld64.c116 vacc0x01 = vmla_f32(vacc0x01, va0, vb01); in xnn_f32_igemm_minmax_ukernel_4x2__neon_lane_ld64()
117 vacc1x01 = vmla_f32(vacc1x01, va1, vb01); in xnn_f32_igemm_minmax_ukernel_4x2__neon_lane_ld64()
118 vacc2x01 = vmla_f32(vacc2x01, va2, vb01); in xnn_f32_igemm_minmax_ukernel_4x2__neon_lane_ld64()
119 vacc3x01 = vmla_f32(vacc3x01, va3, vb01); in xnn_f32_igemm_minmax_ukernel_4x2__neon_lane_ld64()
/external/XNNPACK/src/f32-spmm/gen/
D4x1-minmax-neon.c79 vacc01 = vmla_f32(vacc01, vi01, vw); in xnn_f32_spmm_minmax_ukernel_4x1__neon()
105 vacc0 = vmla_f32(vacc0, vi0, vw); in xnn_f32_spmm_minmax_ukernel_4x1__neon()
D4x1-minmax-neon-pipelined.c85 vacc01 = vmla_f32(vacc01, vi01, vb); in xnn_f32_spmm_minmax_ukernel_4x1__neon_pipelined()
113 vacc0 = vmla_f32(vacc0, vi0, vb); in xnn_f32_spmm_minmax_ukernel_4x1__neon_pipelined()
D4x1-minmax-neon-x2.c99 vacc01 = vmla_f32(vacc01, vi01, vw); in xnn_f32_spmm_minmax_ukernel_4x1__neon_x2()
125 vacc0 = vmla_f32(vacc0, vi0, vw); in xnn_f32_spmm_minmax_ukernel_4x1__neon_x2()
D8x1-minmax-neon.c111 vacc01 = vmla_f32(vacc01, vi01, vw); in xnn_f32_spmm_minmax_ukernel_8x1__neon()
137 vacc0 = vmla_f32(vacc0, vi0, vw); in xnn_f32_spmm_minmax_ukernel_8x1__neon()
D8x1-minmax-neon-pipelined.c120 vacc01 = vmla_f32(vacc01, vi01, vb); in xnn_f32_spmm_minmax_ukernel_8x1__neon_pipelined()
148 vacc0 = vmla_f32(vacc0, vi0, vb); in xnn_f32_spmm_minmax_ukernel_8x1__neon_pipelined()
D8x1-minmax-neon-x2.c138 vacc01 = vmla_f32(vacc01, vi01, vw); in xnn_f32_spmm_minmax_ukernel_8x1__neon_x2()
164 vacc0 = vmla_f32(vacc0, vi0, vw); in xnn_f32_spmm_minmax_ukernel_8x1__neon_x2()
D12x1-minmax-neon.c149 vacc01 = vmla_f32(vacc01, vi01, vw); in xnn_f32_spmm_minmax_ukernel_12x1__neon()
175 vacc0 = vmla_f32(vacc0, vi0, vw); in xnn_f32_spmm_minmax_ukernel_12x1__neon()
D16x1-minmax-neon.c155 vacc01 = vmla_f32(vacc01, vi01, vw); in xnn_f32_spmm_minmax_ukernel_16x1__neon()
181 vacc0 = vmla_f32(vacc0, vi0, vw); in xnn_f32_spmm_minmax_ukernel_16x1__neon()
D16x1-minmax-neon-pipelined.c168 vacc01 = vmla_f32(vacc01, vi01, vb); in xnn_f32_spmm_minmax_ukernel_16x1__neon_pipelined()
196 vacc0 = vmla_f32(vacc0, vi0, vb); in xnn_f32_spmm_minmax_ukernel_16x1__neon_pipelined()
D16x1-minmax-neon-x2.c196 vacc01 = vmla_f32(vacc01, vi01, vw); in xnn_f32_spmm_minmax_ukernel_16x1__neon_x2()
222 vacc0 = vmla_f32(vacc0, vi0, vw); in xnn_f32_spmm_minmax_ukernel_16x1__neon_x2()
D32x1-minmax-neon.c224 vacc01 = vmla_f32(vacc01, vi01, vw); in xnn_f32_spmm_minmax_ukernel_32x1__neon()
250 vacc0 = vmla_f32(vacc0, vi0, vw); in xnn_f32_spmm_minmax_ukernel_32x1__neon()
D32x1-minmax-neon-pipelined.c247 vacc01 = vmla_f32(vacc01, vi01, vb); in xnn_f32_spmm_minmax_ukernel_32x1__neon_pipelined()
276 vacc0 = vmla_f32(vacc0, vi0, vb); in xnn_f32_spmm_minmax_ukernel_32x1__neon_pipelined()
D32x1-minmax-neon-x2.c295 vacc01 = vmla_f32(vacc01, vi01, vw); in xnn_f32_spmm_minmax_ukernel_32x1__neon_x2()
321 vacc0 = vmla_f32(vacc0, vi0, vw); in xnn_f32_spmm_minmax_ukernel_32x1__neon_x2()
/external/XNNPACK/src/f32-igemm/
DMRx2-neon-ld64.c.in104 vacc${M}x01 = vmla_f32(vacc${M}x01, va${M}, vb01);
/external/XNNPACK/src/f32-gemm/
DMRx2-neon-ld64.c.in102 vacc${M}x01 = vmla_f32(vacc${M}x01, va${M}, vb01);
/external/XNNPACK/src/f32-spmm/
Dneon-pipelined.c.in8 $VMULADD_F32 = "vfma_f32" if FMA else "vmla_f32"
Dneon.c.in8 $VMULADD_F32 = "vfma_f32" if FMA else "vmla_f32"
/external/llvm-project/clang/test/CodeGen/
Daarch64-neon-intrinsics-constrained.c110 return vmla_f32(v1, v2, v3); in test_vmla_f32()
/external/XNNPACK/src/f32-ibilinear-chw/
Dneon.c.in10 $VMULADD_F32 = "vfma_f32" if FMA else "vmla_f32"
/external/tensorflow/tensorflow/lite/kernels/internal/optimized/
Ddepthwiseconv_float.h168 acc = vmla_f32(acc, input, filters);
728 acc = vmla_f32(acc, input, filter);
/external/tensorflow/tensorflow/core/kernels/neon/
Ddepthwiseconv_float.h172 acc = vmla_f32(acc, input, filters);

12