Home
last modified time | relevance | path

Searched refs:va1c0 (Results 1 – 25 of 76) sorted by relevance

1234

/external/XNNPACK/src/f32-gemm/gen-inc/
D4x8inc-minmax-neonfma-dup-ld64.c83 const float32x4_t va1c0 = vdupq_lane_f32(va1, 0); in xnn_f32_gemminc_minmax_ukernel_4x8__neonfma_dup_ld64() local
87 vacc1x0123 = vfmaq_f32(vacc1x0123, va1c0, vb0123c0); in xnn_f32_gemminc_minmax_ukernel_4x8__neonfma_dup_ld64()
91 vacc1x4567 = vfmaq_f32(vacc1x4567, va1c0, vb4567c0); in xnn_f32_gemminc_minmax_ukernel_4x8__neonfma_dup_ld64()
D3x8inc-minmax-wasmsimd-arm-splat.c76 const v128_t va1c0 = wasm_v32x4_shuffle(va1, va1, 0, 0, 0, 0); in xnn_f32_gemminc_minmax_ukernel_3x8__wasmsimd_arm_splat() local
83 vacc1x0123 = wasm_f32x4_add(vacc1x0123, wasm_f32x4_mul(va1c0, vb0123c0)); in xnn_f32_gemminc_minmax_ukernel_3x8__wasmsimd_arm_splat()
86 vacc1x4567 = wasm_f32x4_add(vacc1x4567, wasm_f32x4_mul(va1c0, vb4567c0)); in xnn_f32_gemminc_minmax_ukernel_3x8__wasmsimd_arm_splat()
D4x8inc-minmax-neon-dup-ld64.c83 const float32x4_t va1c0 = vdupq_lane_f32(va1, 0); in xnn_f32_gemminc_minmax_ukernel_4x8__neon_dup_ld64() local
87 vacc1x0123 = vmlaq_f32(vacc1x0123, va1c0, vb0123c0); in xnn_f32_gemminc_minmax_ukernel_4x8__neon_dup_ld64()
91 vacc1x4567 = vmlaq_f32(vacc1x4567, va1c0, vb4567c0); in xnn_f32_gemminc_minmax_ukernel_4x8__neon_dup_ld64()
D3x8inc-minmax-wasmsimd-x86-splat.c74 const v128_t va1c0 = wasm_v32x4_shuffle(va1, va1, 0, 0, 0, 0); in xnn_f32_gemminc_minmax_ukernel_3x8__wasmsimd_x86_splat() local
81 vacc1x0123 = wasm_f32x4_add(vacc1x0123, wasm_f32x4_mul(va1c0, vb0123c0)); in xnn_f32_gemminc_minmax_ukernel_3x8__wasmsimd_x86_splat()
84 vacc1x4567 = wasm_f32x4_add(vacc1x4567, wasm_f32x4_mul(va1c0, vb4567c0)); in xnn_f32_gemminc_minmax_ukernel_3x8__wasmsimd_x86_splat()
D4x8inc-minmax-neon-dup-ld128.c84 const float32x4_t va1c0 = vdupq_lane_f32(vget_low_f32(va1), 0); in xnn_f32_gemminc_minmax_ukernel_4x8__neon_dup_ld128() local
88 vacc1x0123 = vmlaq_f32(vacc1x0123, va1c0, vb0123c0); in xnn_f32_gemminc_minmax_ukernel_4x8__neon_dup_ld128()
92 vacc1x4567 = vmlaq_f32(vacc1x4567, va1c0, vb4567c0); in xnn_f32_gemminc_minmax_ukernel_4x8__neon_dup_ld128()
D4x8inc-minmax-neonfma-dup-ld128.c84 const float32x4_t va1c0 = vdupq_lane_f32(vget_low_f32(va1), 0); in xnn_f32_gemminc_minmax_ukernel_4x8__neonfma_dup_ld128() local
88 vacc1x0123 = vfmaq_f32(vacc1x0123, va1c0, vb0123c0); in xnn_f32_gemminc_minmax_ukernel_4x8__neonfma_dup_ld128()
92 vacc1x4567 = vfmaq_f32(vacc1x4567, va1c0, vb4567c0); in xnn_f32_gemminc_minmax_ukernel_4x8__neonfma_dup_ld128()
D4x8inc-minmax-wasmsimd-arm-splat.c86 const v128_t va1c0 = wasm_v32x4_shuffle(va1, va1, 0, 0, 0, 0); in xnn_f32_gemminc_minmax_ukernel_4x8__wasmsimd_arm_splat() local
94 vacc1x0123 = wasm_f32x4_add(vacc1x0123, wasm_f32x4_mul(va1c0, vb0123c0)); in xnn_f32_gemminc_minmax_ukernel_4x8__wasmsimd_arm_splat()
98 vacc1x4567 = wasm_f32x4_add(vacc1x4567, wasm_f32x4_mul(va1c0, vb4567c0)); in xnn_f32_gemminc_minmax_ukernel_4x8__wasmsimd_arm_splat()
D4x8inc-minmax-wasmsimd-x86-splat.c84 const v128_t va1c0 = wasm_v32x4_shuffle(va1, va1, 0, 0, 0, 0); in xnn_f32_gemminc_minmax_ukernel_4x8__wasmsimd_x86_splat() local
92 vacc1x0123 = wasm_f32x4_add(vacc1x0123, wasm_f32x4_mul(va1c0, vb0123c0)); in xnn_f32_gemminc_minmax_ukernel_4x8__wasmsimd_x86_splat()
96 vacc1x4567 = wasm_f32x4_add(vacc1x4567, wasm_f32x4_mul(va1c0, vb4567c0)); in xnn_f32_gemminc_minmax_ukernel_4x8__wasmsimd_x86_splat()
/external/XNNPACK/src/f32-igemm/gen/
D4x8-minmax-neon-dup-ld64.c103 const float32x4_t va1c0 = vdupq_lane_f32(va1, 0); in xnn_f32_igemm_minmax_ukernel_4x8__neon_dup_ld64() local
107 vacc1x0123 = vmlaq_f32(vacc1x0123, va1c0, vb0123c0); in xnn_f32_igemm_minmax_ukernel_4x8__neon_dup_ld64()
111 vacc1x4567 = vmlaq_f32(vacc1x4567, va1c0, vb4567c0); in xnn_f32_igemm_minmax_ukernel_4x8__neon_dup_ld64()
D4x8-minmax-neonfma-dup-ld64.c103 const float32x4_t va1c0 = vdupq_lane_f32(va1, 0); in xnn_f32_igemm_minmax_ukernel_4x8__neonfma_dup_ld64() local
107 vacc1x0123 = vfmaq_f32(vacc1x0123, va1c0, vb0123c0); in xnn_f32_igemm_minmax_ukernel_4x8__neonfma_dup_ld64()
111 vacc1x4567 = vfmaq_f32(vacc1x4567, va1c0, vb4567c0); in xnn_f32_igemm_minmax_ukernel_4x8__neonfma_dup_ld64()
D3x8-minmax-wasmsimd-arm-splat.c93 const v128_t va1c0 = wasm_v32x4_shuffle(va1, va1, 0, 0, 0, 0); in xnn_f32_igemm_minmax_ukernel_3x8__wasmsimd_arm_splat() local
100 vacc1x0123 = wasm_f32x4_add(vacc1x0123, wasm_f32x4_mul(va1c0, vb0123c0)); in xnn_f32_igemm_minmax_ukernel_3x8__wasmsimd_arm_splat()
103 vacc1x4567 = wasm_f32x4_add(vacc1x4567, wasm_f32x4_mul(va1c0, vb4567c0)); in xnn_f32_igemm_minmax_ukernel_3x8__wasmsimd_arm_splat()
D3x8-minmax-wasmsimd-x86-splat.c91 const v128_t va1c0 = wasm_v32x4_shuffle(va1, va1, 0, 0, 0, 0); in xnn_f32_igemm_minmax_ukernel_3x8__wasmsimd_x86_splat() local
98 vacc1x0123 = wasm_f32x4_add(vacc1x0123, wasm_f32x4_mul(va1c0, vb0123c0)); in xnn_f32_igemm_minmax_ukernel_3x8__wasmsimd_x86_splat()
101 vacc1x4567 = wasm_f32x4_add(vacc1x4567, wasm_f32x4_mul(va1c0, vb4567c0)); in xnn_f32_igemm_minmax_ukernel_3x8__wasmsimd_x86_splat()
D4x8-relu-wasmsimd-splat.c104 const v128_t va1c0 = wasm_v32x4_shuffle(va1, va1, 0, 0, 0, 0); in xnn_f32_igemm_relu_ukernel_4x8__wasmsimd_splat() local
112 vacc1x0123 = wasm_f32x4_add(vacc1x0123, wasm_f32x4_mul(va1c0, vb0123c0)); in xnn_f32_igemm_relu_ukernel_4x8__wasmsimd_splat()
116 vacc1x4567 = wasm_f32x4_add(vacc1x4567, wasm_f32x4_mul(va1c0, vb4567c0)); in xnn_f32_igemm_relu_ukernel_4x8__wasmsimd_splat()
D4x8-wasmsimd-splat.c104 const v128_t va1c0 = wasm_v32x4_shuffle(va1, va1, 0, 0, 0, 0); in xnn_f32_igemm_ukernel_4x8__wasmsimd_splat() local
112 vacc1x0123 = wasm_f32x4_add(vacc1x0123, wasm_f32x4_mul(va1c0, vb0123c0)); in xnn_f32_igemm_ukernel_4x8__wasmsimd_splat()
116 vacc1x4567 = wasm_f32x4_add(vacc1x4567, wasm_f32x4_mul(va1c0, vb4567c0)); in xnn_f32_igemm_ukernel_4x8__wasmsimd_splat()
D4x8-minmax-neon-dup-ld128.c104 const float32x4_t va1c0 = vdupq_lane_f32(vget_low_f32(va1), 0); in xnn_f32_igemm_minmax_ukernel_4x8__neon_dup_ld128() local
108 vacc1x0123 = vmlaq_f32(vacc1x0123, va1c0, vb0123c0); in xnn_f32_igemm_minmax_ukernel_4x8__neon_dup_ld128()
112 vacc1x4567 = vmlaq_f32(vacc1x4567, va1c0, vb4567c0); in xnn_f32_igemm_minmax_ukernel_4x8__neon_dup_ld128()
D4x8-minmax-neonfma-dup-ld128.c104 const float32x4_t va1c0 = vdupq_lane_f32(vget_low_f32(va1), 0); in xnn_f32_igemm_minmax_ukernel_4x8__neonfma_dup_ld128() local
108 vacc1x0123 = vfmaq_f32(vacc1x0123, va1c0, vb0123c0); in xnn_f32_igemm_minmax_ukernel_4x8__neonfma_dup_ld128()
112 vacc1x4567 = vfmaq_f32(vacc1x4567, va1c0, vb4567c0); in xnn_f32_igemm_minmax_ukernel_4x8__neonfma_dup_ld128()
/external/XNNPACK/src/f32-gemm/gen/
D4x8-minmax-neonfma-dup-ld64.c81 const float32x4_t va1c0 = vdupq_lane_f32(va1, 0); in xnn_f32_gemm_minmax_ukernel_4x8__neonfma_dup_ld64() local
85 vacc1x0123 = vfmaq_f32(vacc1x0123, va1c0, vb0123c0); in xnn_f32_gemm_minmax_ukernel_4x8__neonfma_dup_ld64()
89 vacc1x4567 = vfmaq_f32(vacc1x4567, va1c0, vb4567c0); in xnn_f32_gemm_minmax_ukernel_4x8__neonfma_dup_ld64()
D4x8-minmax-neon-dup-ld64.c81 const float32x4_t va1c0 = vdupq_lane_f32(va1, 0); in xnn_f32_gemm_minmax_ukernel_4x8__neon_dup_ld64() local
85 vacc1x0123 = vmlaq_f32(vacc1x0123, va1c0, vb0123c0); in xnn_f32_gemm_minmax_ukernel_4x8__neon_dup_ld64()
89 vacc1x4567 = vmlaq_f32(vacc1x4567, va1c0, vb4567c0); in xnn_f32_gemm_minmax_ukernel_4x8__neon_dup_ld64()
D3x8-minmax-wasmsimd-arm-splat.c74 const v128_t va1c0 = wasm_v32x4_shuffle(va1, va1, 0, 0, 0, 0); in xnn_f32_gemm_minmax_ukernel_3x8__wasmsimd_arm_splat() local
81 vacc1x0123 = wasm_f32x4_add(vacc1x0123, wasm_f32x4_mul(va1c0, vb0123c0)); in xnn_f32_gemm_minmax_ukernel_3x8__wasmsimd_arm_splat()
84 vacc1x4567 = wasm_f32x4_add(vacc1x4567, wasm_f32x4_mul(va1c0, vb4567c0)); in xnn_f32_gemm_minmax_ukernel_3x8__wasmsimd_arm_splat()
D3x8-minmax-wasmsimd-x86-splat.c72 const v128_t va1c0 = wasm_v32x4_shuffle(va1, va1, 0, 0, 0, 0); in xnn_f32_gemm_minmax_ukernel_3x8__wasmsimd_x86_splat() local
79 vacc1x0123 = wasm_f32x4_add(vacc1x0123, wasm_f32x4_mul(va1c0, vb0123c0)); in xnn_f32_gemm_minmax_ukernel_3x8__wasmsimd_x86_splat()
82 vacc1x4567 = wasm_f32x4_add(vacc1x4567, wasm_f32x4_mul(va1c0, vb4567c0)); in xnn_f32_gemm_minmax_ukernel_3x8__wasmsimd_x86_splat()
D4x8-wasmsimd-splat.c82 const v128_t va1c0 = wasm_v32x4_shuffle(va1, va1, 0, 0, 0, 0); in xnn_f32_gemm_ukernel_4x8__wasmsimd_splat() local
90 vacc1x0123 = wasm_f32x4_add(vacc1x0123, wasm_f32x4_mul(va1c0, vb0123c0)); in xnn_f32_gemm_ukernel_4x8__wasmsimd_splat()
94 vacc1x4567 = wasm_f32x4_add(vacc1x4567, wasm_f32x4_mul(va1c0, vb4567c0)); in xnn_f32_gemm_ukernel_4x8__wasmsimd_splat()
D4x8-minmax-neon-dup-ld128.c82 const float32x4_t va1c0 = vdupq_lane_f32(vget_low_f32(va1), 0); in xnn_f32_gemm_minmax_ukernel_4x8__neon_dup_ld128() local
86 vacc1x0123 = vmlaq_f32(vacc1x0123, va1c0, vb0123c0); in xnn_f32_gemm_minmax_ukernel_4x8__neon_dup_ld128()
90 vacc1x4567 = vmlaq_f32(vacc1x4567, va1c0, vb4567c0); in xnn_f32_gemm_minmax_ukernel_4x8__neon_dup_ld128()
D4x8-relu-wasmsimd-splat.c82 const v128_t va1c0 = wasm_v32x4_shuffle(va1, va1, 0, 0, 0, 0); in xnn_f32_gemm_relu_ukernel_4x8__wasmsimd_splat() local
90 vacc1x0123 = wasm_f32x4_add(vacc1x0123, wasm_f32x4_mul(va1c0, vb0123c0)); in xnn_f32_gemm_relu_ukernel_4x8__wasmsimd_splat()
94 vacc1x4567 = wasm_f32x4_add(vacc1x4567, wasm_f32x4_mul(va1c0, vb4567c0)); in xnn_f32_gemm_relu_ukernel_4x8__wasmsimd_splat()
D4x8-minmax-neonfma-dup-ld128.c82 const float32x4_t va1c0 = vdupq_lane_f32(vget_low_f32(va1), 0); in xnn_f32_gemm_minmax_ukernel_4x8__neonfma_dup_ld128() local
86 vacc1x0123 = vfmaq_f32(vacc1x0123, va1c0, vb0123c0); in xnn_f32_gemm_minmax_ukernel_4x8__neonfma_dup_ld128()
90 vacc1x4567 = vfmaq_f32(vacc1x4567, va1c0, vb4567c0); in xnn_f32_gemm_minmax_ukernel_4x8__neonfma_dup_ld128()
D4x2-minmax-neonfma-lane-ld64.c82 const float32x2_t va1c0 = vdup_lane_f32(va1, 0); in xnn_f32_gemm_minmax_ukernel_4x2__neonfma_lane_ld64() local
86 vacc1x01 = vfma_f32(vacc1x01, va1c0, vb01c0); in xnn_f32_gemm_minmax_ukernel_4x2__neonfma_lane_ld64()

1234