Searched refs:vcgeq_f32 (Results 1 – 14 of 14) sorted by relevance
/external/XNNPACK/src/f32-vrnd/gen/ |
D | vrndu-neon-x8.c | 49 uint32x4_t vadjmask0123 = vcgeq_f32(vrndx0123, vx0123); in xnn_f32_vrndu_ukernel__neon_x8() 50 uint32x4_t vadjmask4567 = vcgeq_f32(vrndx4567, vx4567); in xnn_f32_vrndu_ukernel__neon_x8() 71 uint32x4_t vadjmask = vcgeq_f32(vrndx, vx); in xnn_f32_vrndu_ukernel__neon_x8() 84 uint32x4_t vadjmask = vcgeq_f32(vrndx, vx); in xnn_f32_vrndu_ukernel__neon_x8()
|
D | vrndu-neon-x4.c | 43 uint32x4_t vadjmask0123 = vcgeq_f32(vrndx0123, vx0123); in xnn_f32_vrndu_ukernel__neon_x4() 60 uint32x4_t vadjmask = vcgeq_f32(vrndx, vx); in xnn_f32_vrndu_ukernel__neon_x4()
|
/external/XNNPACK/src/f32-vrnd/ |
D | vrndu-neon.c.in | 49 uint32x4_t vadjmask${ABC[N:N+4]} = vcgeq_f32(vrndx${ABC[N:N+4]}, vx${ABC[N:N+4]}); 71 uint32x4_t vadjmask = vcgeq_f32(vrndx, vx); 84 uint32x4_t vadjmask = vcgeq_f32(vrndx, vx);
|
/external/XNNPACK/src/math/ |
D | roundu-neon-cvt.c | 52 const uint32x4_t vadjmask = vorrq_u32(vcgeq_f32(vrndx, vx), vsign_mask); in xnn_math_f32_roundu__neon_cvt()
|
/external/llvm-project/clang/test/CodeGen/ |
D | aarch64-neon-intrinsics-constrained.c | 347 return vcgeq_f32(v1, v2); in test_vcgeq_f32()
|
D | aarch64-neon-intrinsics.c | 1970 return vcgeq_f32(v1, v2); in test_vcgeq_f32()
|
D | arm_neon_intrinsics.c | 1503 return vcgeq_f32(a, b); in test_vcgeq_f32()
|
/external/tensorflow/tensorflow/lite/kernels/internal/optimized/ |
D | optimized_ops.h | 8103 const uint32x4_t mask1 = vcgeq_f32(input1, zero_dup); in PReluScalarBroadcast() 8104 const uint32x4_t mask2 = vcgeq_f32(input2, zero_dup); in PReluScalarBroadcast() 8105 const uint32x4_t mask3 = vcgeq_f32(input3, zero_dup); in PReluScalarBroadcast() 8106 const uint32x4_t mask4 = vcgeq_f32(input4, zero_dup); in PReluScalarBroadcast() 8121 const uint32x4_t mask = vcgeq_f32(input, zero_dup); in PReluScalarBroadcast() 8155 const uint32x4_t mask1 = vcgeq_f32(input1, zero_dup); in PReluElementWise() 8156 const uint32x4_t mask2 = vcgeq_f32(input2, zero_dup); in PReluElementWise() 8157 const uint32x4_t mask3 = vcgeq_f32(input3, zero_dup); in PReluElementWise() 8158 const uint32x4_t mask4 = vcgeq_f32(input4, zero_dup); in PReluElementWise() 8175 const uint32x4_t mask = vcgeq_f32(input, zero_dup); in PReluElementWise()
|
/external/skia/include/private/ |
D | SkNx_neon.h | 208 AI SkNx operator>=(const SkNx& o) const {return vreinterpretq_f32_u32(vcgeq_f32(fVec, o.fVec));}
|
/external/skqp/include/private/ |
D | SkNx_neon.h | 225 AI SkNx operator>=(const SkNx& o) const {return vreinterpretq_f32_u32(vcgeq_f32(fVec, o.fVec));}
|
/external/pffft/ |
D | sse2neon.h | 3855 vcgeq_f32(vreinterpretq_f32_m128(a), vreinterpretq_f32_m128(b))); in _mm_cmpge_ps() 4232 vcgeq_f32(vreinterpretq_f32_m128(a), vreinterpretq_f32_m128(b)); in _mm_comige_ss()
|
/external/clang/test/CodeGen/ |
D | aarch64-neon-intrinsics.c | 2152 return vcgeq_f32(v1, v2); in test_vcgeq_f32()
|
D | arm_neon_intrinsics.c | 1651 return vcgeq_f32(a, b); in test_vcgeq_f32()
|
/external/neon_2_sse/ |
D | NEON_2_SSE.h | 741 _NEON2SSESTORAGE uint32x4_t vcgeq_f32(float32x4_t a, float32x4_t b); // VCGE.F32 q0, q0, q0 5200 _NEON2SSESTORAGE uint32x4_t vcgeq_f32(float32x4_t a, float32x4_t b); // VCGE.F32 q0, q0, q0 5201 _NEON2SSE_INLINE uint32x4_t vcgeq_f32(float32x4_t a, float32x4_t b) in vcgeq_f32() function
|