/external/XNNPACK/src/qu8-requantization/ |
D | precise-neon.c | 110 …const int32x4_t x_scaled = vuzp1q_s32(vreinterpretq_s32_s64(x01_scaled), vreinterpretq_s32_s64(x23… in xnn_qu8_requantize_precise__neon() local 119 const int32x4_t x_scaled = vcombine_s32(vmovn_s64(x01_scaled), vmovn_s64(x23_scaled)); in xnn_qu8_requantize_precise__neon() local
|
D | fp32-sse2.c | 48 const __m128 x_scaled = _mm_mul_ps(_mm_cvtepi32_ps(x), vscale); in xnn_qu8_requantize_fp32__sse2() local
|
D | fp32-scalar-magic.c | 43 const float x_scaled = (float) x * scale; in xnn_qu8_requantize_fp32__scalar_magic() local
|
D | fp32-scalar-lrintf.c | 41 const float x_scaled = (float) x * scale; in xnn_qu8_requantize_fp32__scalar_lrintf() local
|
D | fp32-wasmsimd.c | 46 const v128_t x_scaled = wasm_f32x4_mul(wasm_f32x4_convert_i32x4(x), vscale); in xnn_qu8_requantize_fp32__wasmsimd() local
|
D | precise-scalar-signed64.c | 68 const int32_t x_scaled = (int32_t) asr_s64(x_adjusted_product + rounding, shift); in xnn_qu8_requantize_precise__scalar_signed64() local
|
D | q31-neon.c | 77 const int32x4_t x_scaled = vrshlq_s32(x_adjusted_product, vshift); in xnn_qu8_requantize_q31__neon() local
|
D | fp32-neon.c | 55 const float32x4_t x_scaled = vmulq_f32(vcvtq_f32_s32(x), vscale); in xnn_qu8_requantize_fp32__neon() local
|
D | precise-scalar-unsigned64.c | 75 const int32_t x_scaled = (int32_t)(x >= 0 ? x_abs_scaled : -x_abs_scaled); in xnn_qu8_requantize_precise__scalar_unsigned64() local
|
D | q31-scalar.c | 103 const int32_t x_scaled = asr_s32(x_q31product, shift) + (int32_t)(x_remainder > threshold); in xnn_qu8_requantize_q31__scalar() local
|
D | precise-scalar-unsigned32.c | 103 const int32_t x_scaled = (int32_t)(x >= 0 ? x_abs_scaled : -x_abs_scaled); in xnn_qu8_requantize_precise__scalar_unsigned32() local
|
D | precise-sse4.c | 88 const __m128i x_scaled = _mm_sign_epi32(x_abs_scaled, x); in xnn_qu8_requantize_precise__sse4() local
|
/external/XNNPACK/src/qs8-requantization/ |
D | precise-neon.c | 110 …const int32x4_t x_scaled = vuzp1q_s32(vreinterpretq_s32_s64(x01_scaled), vreinterpretq_s32_s64(x23… in xnn_qs8_requantize_precise__neon() local 119 const int32x4_t x_scaled = vcombine_s32(vmovn_s64(x01_scaled), vmovn_s64(x23_scaled)); in xnn_qs8_requantize_precise__neon() local
|
D | fp32-scalar-magic.c | 43 const float x_scaled = (float) x * scale; in xnn_qs8_requantize_fp32__scalar_magic() local
|
D | fp32-sse4.c | 48 const __m128 x_scaled = _mm_mul_ps(_mm_cvtepi32_ps(x), vscale); in xnn_qs8_requantize_fp32__sse4() local
|
D | fp32-sse2.c | 48 const __m128 x_scaled = _mm_mul_ps(_mm_cvtepi32_ps(x), vscale); in xnn_qs8_requantize_fp32__sse2() local
|
D | fp32-scalar-lrintf.c | 41 const float x_scaled = (float) x * scale; in xnn_qs8_requantize_fp32__scalar_lrintf() local
|
D | fp32-wasmsimd.c | 46 const v128_t x_scaled = wasm_f32x4_mul(wasm_f32x4_convert_i32x4(x), vscale); in xnn_qs8_requantize_fp32__wasmsimd() local
|
D | precise-scalar-signed64.c | 68 const int32_t x_scaled = (int32_t) asr_s64(x_adjusted_product + rounding, shift); in xnn_qs8_requantize_precise__scalar_signed64() local
|
D | q31-neon.c | 77 const int32x4_t x_scaled = vrshlq_s32(x_adjusted_product, vshift); in xnn_qs8_requantize_q31__neon() local
|
D | precise-scalar-unsigned64.c | 75 const int32_t x_scaled = (int32_t) (x >= 0 ? x_abs_scaled : -x_abs_scaled); in xnn_qs8_requantize_precise__scalar_unsigned64() local
|
D | fp32-neon.c | 55 const float32x4_t x_scaled = vmulq_f32(vcvtq_f32_s32(x), vscale); in xnn_qs8_requantize_fp32__neon() local
|
D | q31-scalar.c | 103 const int32_t x_scaled = asr_s32(x_q31product, shift) + (int32_t) (x_remainder > threshold); in xnn_qs8_requantize_q31__scalar() local
|
D | precise-scalar-unsigned32.c | 103 const int32_t x_scaled = (int32_t) (x >= 0 ? x_abs_scaled : -x_abs_scaled); in xnn_qs8_requantize_precise__scalar_unsigned32() local
|
D | precise-sse4.c | 88 const __m128i x_scaled = _mm_sign_epi32(x_abs_scaled, x); in xnn_qs8_requantize_precise__sse4() local
|