Home
last modified time | relevance | path

Searched refs:vcgeq_f32 (Results 1 – 14 of 14) sorted by relevance

/external/XNNPACK/src/f32-vrnd/gen/
Dvrndu-neon-x8.c49 uint32x4_t vadjmask0123 = vcgeq_f32(vrndx0123, vx0123); in xnn_f32_vrndu_ukernel__neon_x8()
50 uint32x4_t vadjmask4567 = vcgeq_f32(vrndx4567, vx4567); in xnn_f32_vrndu_ukernel__neon_x8()
71 uint32x4_t vadjmask = vcgeq_f32(vrndx, vx); in xnn_f32_vrndu_ukernel__neon_x8()
84 uint32x4_t vadjmask = vcgeq_f32(vrndx, vx); in xnn_f32_vrndu_ukernel__neon_x8()
Dvrndu-neon-x4.c43 uint32x4_t vadjmask0123 = vcgeq_f32(vrndx0123, vx0123); in xnn_f32_vrndu_ukernel__neon_x4()
60 uint32x4_t vadjmask = vcgeq_f32(vrndx, vx); in xnn_f32_vrndu_ukernel__neon_x4()
/external/XNNPACK/src/f32-vrnd/
Dvrndu-neon.c.in49 uint32x4_t vadjmask${ABC[N:N+4]} = vcgeq_f32(vrndx${ABC[N:N+4]}, vx${ABC[N:N+4]});
71 uint32x4_t vadjmask = vcgeq_f32(vrndx, vx);
84 uint32x4_t vadjmask = vcgeq_f32(vrndx, vx);
/external/XNNPACK/src/math/
Droundu-neon-cvt.c52 const uint32x4_t vadjmask = vorrq_u32(vcgeq_f32(vrndx, vx), vsign_mask); in xnn_math_f32_roundu__neon_cvt()
/external/llvm-project/clang/test/CodeGen/
Daarch64-neon-intrinsics-constrained.c347 return vcgeq_f32(v1, v2); in test_vcgeq_f32()
Daarch64-neon-intrinsics.c1970 return vcgeq_f32(v1, v2); in test_vcgeq_f32()
Darm_neon_intrinsics.c1503 return vcgeq_f32(a, b); in test_vcgeq_f32()
/external/tensorflow/tensorflow/lite/kernels/internal/optimized/
Doptimized_ops.h8103 const uint32x4_t mask1 = vcgeq_f32(input1, zero_dup); in PReluScalarBroadcast()
8104 const uint32x4_t mask2 = vcgeq_f32(input2, zero_dup); in PReluScalarBroadcast()
8105 const uint32x4_t mask3 = vcgeq_f32(input3, zero_dup); in PReluScalarBroadcast()
8106 const uint32x4_t mask4 = vcgeq_f32(input4, zero_dup); in PReluScalarBroadcast()
8121 const uint32x4_t mask = vcgeq_f32(input, zero_dup); in PReluScalarBroadcast()
8155 const uint32x4_t mask1 = vcgeq_f32(input1, zero_dup); in PReluElementWise()
8156 const uint32x4_t mask2 = vcgeq_f32(input2, zero_dup); in PReluElementWise()
8157 const uint32x4_t mask3 = vcgeq_f32(input3, zero_dup); in PReluElementWise()
8158 const uint32x4_t mask4 = vcgeq_f32(input4, zero_dup); in PReluElementWise()
8175 const uint32x4_t mask = vcgeq_f32(input, zero_dup); in PReluElementWise()
/external/skia/include/private/
DSkNx_neon.h208 AI SkNx operator>=(const SkNx& o) const {return vreinterpretq_f32_u32(vcgeq_f32(fVec, o.fVec));}
/external/skqp/include/private/
DSkNx_neon.h225 AI SkNx operator>=(const SkNx& o) const {return vreinterpretq_f32_u32(vcgeq_f32(fVec, o.fVec));}
/external/pffft/
Dsse2neon.h3855 vcgeq_f32(vreinterpretq_f32_m128(a), vreinterpretq_f32_m128(b))); in _mm_cmpge_ps()
4232 vcgeq_f32(vreinterpretq_f32_m128(a), vreinterpretq_f32_m128(b)); in _mm_comige_ss()
/external/clang/test/CodeGen/
Daarch64-neon-intrinsics.c2152 return vcgeq_f32(v1, v2); in test_vcgeq_f32()
Darm_neon_intrinsics.c1651 return vcgeq_f32(a, b); in test_vcgeq_f32()
/external/neon_2_sse/
DNEON_2_SSE.h741 _NEON2SSESTORAGE uint32x4_t vcgeq_f32(float32x4_t a, float32x4_t b); // VCGE.F32 q0, q0, q0
5200 _NEON2SSESTORAGE uint32x4_t vcgeq_f32(float32x4_t a, float32x4_t b); // VCGE.F32 q0, q0, q0
5201 _NEON2SSE_INLINE uint32x4_t vcgeq_f32(float32x4_t a, float32x4_t b) in vcgeq_f32() function