Home
last modified time | relevance | path

Searched refs:vcaltq_f32 (Results 1 – 24 of 24) sorted by relevance

/external/XNNPACK/src/f32-vrnd/gen/
Dvrndz-neon-x8.c36 uint32x4_t vrndmask0123 = vcaltq_f32(vx0123, vintegral_threshold); in xnn_f32_vrndz_ukernel__neon_x8()
37 uint32x4_t vrndmask4567 = vcaltq_f32(vx4567, vintegral_threshold); in xnn_f32_vrndz_ukernel__neon_x8()
54 uint32x4_t vrndmask = vcaltq_f32(vx, vintegral_threshold); in xnn_f32_vrndz_ukernel__neon_x8()
63 uint32x4_t vrndmask = vcaltq_f32(vx, vintegral_threshold); in xnn_f32_vrndz_ukernel__neon_x8()
Dvrndne-neon-x8.c34 uint32x4_t vrndmask0123 = vcaltq_f32(vmagic_number, vx0123); in xnn_f32_vrndne_ukernel__neon_x8()
36 uint32x4_t vrndmask4567 = vcaltq_f32(vmagic_number, vx4567); in xnn_f32_vrndne_ukernel__neon_x8()
56 uint32x4_t vrndmask = vcaltq_f32(vmagic_number, vx); in xnn_f32_vrndne_ukernel__neon_x8()
66 uint32x4_t vrndmask = vcaltq_f32(vmagic_number, vx); in xnn_f32_vrndne_ukernel__neon_x8()
Dvrndd-neon-x8.c37 uint32x4_t vrndmask0123 = vcaltq_f32(vx0123, vintegral_threshold); in xnn_f32_vrndd_ukernel__neon_x8()
38 uint32x4_t vrndmask4567 = vcaltq_f32(vx4567, vintegral_threshold); in xnn_f32_vrndd_ukernel__neon_x8()
64 uint32x4_t vrndmask = vcaltq_f32(vx, vintegral_threshold); in xnn_f32_vrndd_ukernel__neon_x8()
76 uint32x4_t vrndmask = vcaltq_f32(vx, vintegral_threshold); in xnn_f32_vrndd_ukernel__neon_x8()
Dvrndu-neon-x8.c37 uint32x4_t vrndmask0123 = vcaltq_f32(vx0123, vintegral_threshold); in xnn_f32_vrndu_ukernel__neon_x8()
38 uint32x4_t vrndmask4567 = vcaltq_f32(vx4567, vintegral_threshold); in xnn_f32_vrndu_ukernel__neon_x8()
67 uint32x4_t vrndmask = vcaltq_f32(vx, vintegral_threshold); in xnn_f32_vrndu_ukernel__neon_x8()
81 uint32x4_t vrndmask = vcaltq_f32(vx, vintegral_threshold); in xnn_f32_vrndu_ukernel__neon_x8()
Dvrndz-neon-x4.c34 uint32x4_t vrndmask0123 = vcaltq_f32(vx0123, vintegral_threshold); in xnn_f32_vrndz_ukernel__neon_x4()
47 uint32x4_t vrndmask = vcaltq_f32(vx, vintegral_threshold); in xnn_f32_vrndz_ukernel__neon_x4()
Dvrndne-neon-x4.c33 uint32x4_t vrndmask0123 = vcaltq_f32(vmagic_number, vx0123); in xnn_f32_vrndne_ukernel__neon_x4()
48 uint32x4_t vrndmask = vcaltq_f32(vmagic_number, vx); in xnn_f32_vrndne_ukernel__neon_x4()
Dvrndd-neon-x4.c35 uint32x4_t vrndmask0123 = vcaltq_f32(vx0123, vintegral_threshold); in xnn_f32_vrndd_ukernel__neon_x4()
54 uint32x4_t vrndmask = vcaltq_f32(vx, vintegral_threshold); in xnn_f32_vrndd_ukernel__neon_x4()
Dvrndu-neon-x4.c35 uint32x4_t vrndmask0123 = vcaltq_f32(vx0123, vintegral_threshold); in xnn_f32_vrndu_ukernel__neon_x4()
57 uint32x4_t vrndmask = vcaltq_f32(vx, vintegral_threshold); in xnn_f32_vrndu_ukernel__neon_x4()
/external/XNNPACK/src/f32-vrnd/
Dvrndz-neon.c.in36 uint32x4_t vrndmask${ABC[N:N+4]} = vcaltq_f32(vx${ABC[N:N+4]}, vintegral_threshold);
54 uint32x4_t vrndmask = vcaltq_f32(vx, vintegral_threshold);
63 uint32x4_t vrndmask = vcaltq_f32(vx, vintegral_threshold);
Dvrndne-neon.c.in34 uint32x4_t vrndmask${ABC[N:N+4]} = vcaltq_f32(vmagic_number, vx${ABC[N:N+4]});
55 uint32x4_t vrndmask = vcaltq_f32(vmagic_number, vx);
65 uint32x4_t vrndmask = vcaltq_f32(vmagic_number, vx);
Dvrndu-neon.c.in37 uint32x4_t vrndmask${ABC[N:N+4]} = vcaltq_f32(vx${ABC[N:N+4]}, vintegral_threshold);
67 uint32x4_t vrndmask = vcaltq_f32(vx, vintegral_threshold);
81 uint32x4_t vrndmask = vcaltq_f32(vx, vintegral_threshold);
Dvrndd-neon.c.in37 uint32x4_t vrndmask${ABC[N:N+4]} = vcaltq_f32(vx${ABC[N:N+4]}, vintegral_threshold);
64 uint32x4_t vrndmask = vcaltq_f32(vx, vintegral_threshold);
76 uint32x4_t vrndmask = vcaltq_f32(vx, vintegral_threshold);
/external/XNNPACK/src/math/
Droundz-neon-cvt.c38 const uint32x4_t vrndmask = vbicq_u32(vcaltq_f32(vx, vintegral_threshold), vsign_mask); in xnn_math_f32_roundz__neon_cvt()
Droundne-neon-addsub.c40 const uint32x4_t vrndmask = vorrq_u32(vcaltq_f32(vmagic_number, vx), vsign_mask); in xnn_math_f32_roundne__neon_addsub()
Droundd-neon-cvt.c40 const uint32x4_t vrndmask = vbicq_u32(vcaltq_f32(vx, vintegral_threshold), vsign_mask); in xnn_math_f32_roundd__neon_cvt()
Droundd-neon-addsub.c42 const uint32x4_t vrndmask = vorrq_u32(vcaltq_f32(vmagic_number, vx), vsign_mask); in xnn_math_f32_roundd__neon_addsub()
Droundz-neon-addsub.c43 const uint32x4_t vrndmask = vorrq_u32(vcaltq_f32(vmagic_number, vx), vsign_mask); in xnn_math_f32_roundz__neon_addsub()
Droundu-neon-cvt.c40 const uint32x4_t vrndmask = vbicq_u32(vcaltq_f32(vx, vintegral_threshold), vsign_mask); in xnn_math_f32_roundu__neon_cvt()
Droundu-neon-addsub.c42 const uint32x4_t vrndmask = vorrq_u32(vcaltq_f32(vmagic_number, vx), vsign_mask); in xnn_math_f32_roundu__neon_addsub()
/external/llvm-project/clang/test/CodeGen/
Daarch64-neon-intrinsics.c1463 return vcaltq_f32(v1, v2); in test_vcaltq_f32()
Darm_neon_intrinsics.c1287 return vcaltq_f32(a, b); in test_vcaltq_f32()
/external/clang/test/CodeGen/
Daarch64-neon-intrinsics.c1615 return vcaltq_f32(v1, v2); in test_vcaltq_f32()
Darm_neon_intrinsics.c1433 return vcaltq_f32(a, b); in test_vcaltq_f32()
/external/neon_2_sse/
DNEON_2_SSE.h801 _NEON2SSESTORAGE uint32x4_t vcaltq_f32(float32x4_t a, float32x4_t b); // VACGT.F32 q0, q0, q0
5631 _NEON2SSESTORAGE uint32x4_t vcaltq_f32(float32x4_t a, float32x4_t b); // VACGT.F32 q0, q0, q0
5632 _NEON2SSE_INLINE uint32x4_t vcaltq_f32(float32x4_t a, float32x4_t b) // VACGT.F32 q0, q0, q0 in vcaltq_f32() function