/external/XNNPACK/src/f32-vrnd/gen/ |
D | vrndz-neon-x8.c | 36 uint32x4_t vrndmask0123 = vcaltq_f32(vx0123, vintegral_threshold); in xnn_f32_vrndz_ukernel__neon_x8() 37 uint32x4_t vrndmask4567 = vcaltq_f32(vx4567, vintegral_threshold); in xnn_f32_vrndz_ukernel__neon_x8() 54 uint32x4_t vrndmask = vcaltq_f32(vx, vintegral_threshold); in xnn_f32_vrndz_ukernel__neon_x8() 63 uint32x4_t vrndmask = vcaltq_f32(vx, vintegral_threshold); in xnn_f32_vrndz_ukernel__neon_x8()
|
D | vrndne-neon-x8.c | 34 uint32x4_t vrndmask0123 = vcaltq_f32(vmagic_number, vx0123); in xnn_f32_vrndne_ukernel__neon_x8() 36 uint32x4_t vrndmask4567 = vcaltq_f32(vmagic_number, vx4567); in xnn_f32_vrndne_ukernel__neon_x8() 56 uint32x4_t vrndmask = vcaltq_f32(vmagic_number, vx); in xnn_f32_vrndne_ukernel__neon_x8() 66 uint32x4_t vrndmask = vcaltq_f32(vmagic_number, vx); in xnn_f32_vrndne_ukernel__neon_x8()
|
D | vrndd-neon-x8.c | 37 uint32x4_t vrndmask0123 = vcaltq_f32(vx0123, vintegral_threshold); in xnn_f32_vrndd_ukernel__neon_x8() 38 uint32x4_t vrndmask4567 = vcaltq_f32(vx4567, vintegral_threshold); in xnn_f32_vrndd_ukernel__neon_x8() 64 uint32x4_t vrndmask = vcaltq_f32(vx, vintegral_threshold); in xnn_f32_vrndd_ukernel__neon_x8() 76 uint32x4_t vrndmask = vcaltq_f32(vx, vintegral_threshold); in xnn_f32_vrndd_ukernel__neon_x8()
|
D | vrndu-neon-x8.c | 37 uint32x4_t vrndmask0123 = vcaltq_f32(vx0123, vintegral_threshold); in xnn_f32_vrndu_ukernel__neon_x8() 38 uint32x4_t vrndmask4567 = vcaltq_f32(vx4567, vintegral_threshold); in xnn_f32_vrndu_ukernel__neon_x8() 67 uint32x4_t vrndmask = vcaltq_f32(vx, vintegral_threshold); in xnn_f32_vrndu_ukernel__neon_x8() 81 uint32x4_t vrndmask = vcaltq_f32(vx, vintegral_threshold); in xnn_f32_vrndu_ukernel__neon_x8()
|
D | vrndz-neon-x4.c | 34 uint32x4_t vrndmask0123 = vcaltq_f32(vx0123, vintegral_threshold); in xnn_f32_vrndz_ukernel__neon_x4() 47 uint32x4_t vrndmask = vcaltq_f32(vx, vintegral_threshold); in xnn_f32_vrndz_ukernel__neon_x4()
|
D | vrndne-neon-x4.c | 33 uint32x4_t vrndmask0123 = vcaltq_f32(vmagic_number, vx0123); in xnn_f32_vrndne_ukernel__neon_x4() 48 uint32x4_t vrndmask = vcaltq_f32(vmagic_number, vx); in xnn_f32_vrndne_ukernel__neon_x4()
|
D | vrndd-neon-x4.c | 35 uint32x4_t vrndmask0123 = vcaltq_f32(vx0123, vintegral_threshold); in xnn_f32_vrndd_ukernel__neon_x4() 54 uint32x4_t vrndmask = vcaltq_f32(vx, vintegral_threshold); in xnn_f32_vrndd_ukernel__neon_x4()
|
D | vrndu-neon-x4.c | 35 uint32x4_t vrndmask0123 = vcaltq_f32(vx0123, vintegral_threshold); in xnn_f32_vrndu_ukernel__neon_x4() 57 uint32x4_t vrndmask = vcaltq_f32(vx, vintegral_threshold); in xnn_f32_vrndu_ukernel__neon_x4()
|
/external/XNNPACK/src/f32-vrnd/ |
D | vrndz-neon.c.in | 36 uint32x4_t vrndmask${ABC[N:N+4]} = vcaltq_f32(vx${ABC[N:N+4]}, vintegral_threshold); 54 uint32x4_t vrndmask = vcaltq_f32(vx, vintegral_threshold); 63 uint32x4_t vrndmask = vcaltq_f32(vx, vintegral_threshold);
|
D | vrndne-neon.c.in | 34 uint32x4_t vrndmask${ABC[N:N+4]} = vcaltq_f32(vmagic_number, vx${ABC[N:N+4]}); 55 uint32x4_t vrndmask = vcaltq_f32(vmagic_number, vx); 65 uint32x4_t vrndmask = vcaltq_f32(vmagic_number, vx);
|
D | vrndu-neon.c.in | 37 uint32x4_t vrndmask${ABC[N:N+4]} = vcaltq_f32(vx${ABC[N:N+4]}, vintegral_threshold); 67 uint32x4_t vrndmask = vcaltq_f32(vx, vintegral_threshold); 81 uint32x4_t vrndmask = vcaltq_f32(vx, vintegral_threshold);
|
D | vrndd-neon.c.in | 37 uint32x4_t vrndmask${ABC[N:N+4]} = vcaltq_f32(vx${ABC[N:N+4]}, vintegral_threshold); 64 uint32x4_t vrndmask = vcaltq_f32(vx, vintegral_threshold); 76 uint32x4_t vrndmask = vcaltq_f32(vx, vintegral_threshold);
|
/external/XNNPACK/src/math/ |
D | roundz-neon-cvt.c | 38 const uint32x4_t vrndmask = vbicq_u32(vcaltq_f32(vx, vintegral_threshold), vsign_mask); in xnn_math_f32_roundz__neon_cvt()
|
D | roundne-neon-addsub.c | 40 const uint32x4_t vrndmask = vorrq_u32(vcaltq_f32(vmagic_number, vx), vsign_mask); in xnn_math_f32_roundne__neon_addsub()
|
D | roundd-neon-cvt.c | 40 const uint32x4_t vrndmask = vbicq_u32(vcaltq_f32(vx, vintegral_threshold), vsign_mask); in xnn_math_f32_roundd__neon_cvt()
|
D | roundd-neon-addsub.c | 42 const uint32x4_t vrndmask = vorrq_u32(vcaltq_f32(vmagic_number, vx), vsign_mask); in xnn_math_f32_roundd__neon_addsub()
|
D | roundz-neon-addsub.c | 43 const uint32x4_t vrndmask = vorrq_u32(vcaltq_f32(vmagic_number, vx), vsign_mask); in xnn_math_f32_roundz__neon_addsub()
|
D | roundu-neon-cvt.c | 40 const uint32x4_t vrndmask = vbicq_u32(vcaltq_f32(vx, vintegral_threshold), vsign_mask); in xnn_math_f32_roundu__neon_cvt()
|
D | roundu-neon-addsub.c | 42 const uint32x4_t vrndmask = vorrq_u32(vcaltq_f32(vmagic_number, vx), vsign_mask); in xnn_math_f32_roundu__neon_addsub()
|
/external/llvm-project/clang/test/CodeGen/ |
D | aarch64-neon-intrinsics.c | 1463 return vcaltq_f32(v1, v2); in test_vcaltq_f32()
|
D | arm_neon_intrinsics.c | 1287 return vcaltq_f32(a, b); in test_vcaltq_f32()
|
/external/clang/test/CodeGen/ |
D | aarch64-neon-intrinsics.c | 1615 return vcaltq_f32(v1, v2); in test_vcaltq_f32()
|
D | arm_neon_intrinsics.c | 1433 return vcaltq_f32(a, b); in test_vcaltq_f32()
|
/external/neon_2_sse/ |
D | NEON_2_SSE.h | 801 _NEON2SSESTORAGE uint32x4_t vcaltq_f32(float32x4_t a, float32x4_t b); // VACGT.F32 q0, q0, q0 5631 _NEON2SSESTORAGE uint32x4_t vcaltq_f32(float32x4_t a, float32x4_t b); // VACGT.F32 q0, q0, q0 5632 _NEON2SSE_INLINE uint32x4_t vcaltq_f32(float32x4_t a, float32x4_t b) // VACGT.F32 q0, q0, q0 in vcaltq_f32() function
|