Home
last modified time | relevance | path

Searched refs:vacc20 (Results 1 – 25 of 26) sorted by relevance

12

/external/XNNPACK/src/f32-gemm/gen/
D4x2-minmax-scalar.c66 float vacc20 = vacc00; in xnn_f32_gemm_minmax_ukernel_4x2__scalar() local
86 vacc20 += va2 * vb0; in xnn_f32_gemm_minmax_ukernel_4x2__scalar()
98 vacc20 = math_max_f32(vacc20, vmin); in xnn_f32_gemm_minmax_ukernel_4x2__scalar()
107 vacc20 = math_min_f32(vacc20, vmax); in xnn_f32_gemm_minmax_ukernel_4x2__scalar()
116 c2[0] = vacc20; in xnn_f32_gemm_minmax_ukernel_4x2__scalar()
135 c2[0] = vacc20; in xnn_f32_gemm_minmax_ukernel_4x2__scalar()
D4x2-minmax-wasm.c66 float vacc20 = vacc00; in xnn_f32_gemm_minmax_ukernel_4x2__wasm() local
86 vacc20 += va2 * vb0; in xnn_f32_gemm_minmax_ukernel_4x2__wasm()
98 vacc20 = __builtin_wasm_max_f32(vacc20, vmin); in xnn_f32_gemm_minmax_ukernel_4x2__wasm()
107 vacc20 = __builtin_wasm_min_f32(vacc20, vmax); in xnn_f32_gemm_minmax_ukernel_4x2__wasm()
116 c2[0] = vacc20; in xnn_f32_gemm_minmax_ukernel_4x2__wasm()
135 c2[0] = vacc20; in xnn_f32_gemm_minmax_ukernel_4x2__wasm()
D4x4-minmax-scalar.c70 float vacc20 = vacc00; in xnn_f32_gemm_minmax_ukernel_4x4__scalar() local
100 vacc20 += va2 * vb0; in xnn_f32_gemm_minmax_ukernel_4x4__scalar()
120 vacc20 = math_max_f32(vacc20, vmin); in xnn_f32_gemm_minmax_ukernel_4x4__scalar()
137 vacc20 = math_min_f32(vacc20, vmax); in xnn_f32_gemm_minmax_ukernel_4x4__scalar()
152 c2[0] = vacc20; in xnn_f32_gemm_minmax_ukernel_4x4__scalar()
180 c2[0] = vacc20; in xnn_f32_gemm_minmax_ukernel_4x4__scalar()
182 vacc20 = vacc22; in xnn_f32_gemm_minmax_ukernel_4x4__scalar()
195 c2[0] = vacc20; in xnn_f32_gemm_minmax_ukernel_4x4__scalar()
D4x4-minmax-wasm.c70 float vacc20 = vacc00; in xnn_f32_gemm_minmax_ukernel_4x4__wasm() local
100 vacc20 += va2 * vb0; in xnn_f32_gemm_minmax_ukernel_4x4__wasm()
120 vacc20 = __builtin_wasm_max_f32(vacc20, vmin); in xnn_f32_gemm_minmax_ukernel_4x4__wasm()
137 vacc20 = __builtin_wasm_min_f32(vacc20, vmax); in xnn_f32_gemm_minmax_ukernel_4x4__wasm()
152 c2[0] = vacc20; in xnn_f32_gemm_minmax_ukernel_4x4__wasm()
180 c2[0] = vacc20; in xnn_f32_gemm_minmax_ukernel_4x4__wasm()
182 vacc20 = vacc22; in xnn_f32_gemm_minmax_ukernel_4x4__wasm()
195 c2[0] = vacc20; in xnn_f32_gemm_minmax_ukernel_4x4__wasm()
D4x4-relu-wasm.c68 float vacc20 = vacc00; in xnn_f32_gemm_relu_ukernel_4x4__wasm() local
98 vacc20 += va2 * vb0; in xnn_f32_gemm_relu_ukernel_4x4__wasm()
118 vacc20 = __builtin_wasm_max_f32(vacc20, 0.0f); in xnn_f32_gemm_relu_ukernel_4x4__wasm()
133 c2[0] = vacc20; in xnn_f32_gemm_relu_ukernel_4x4__wasm()
161 c2[0] = vacc20; in xnn_f32_gemm_relu_ukernel_4x4__wasm()
163 vacc20 = vacc22; in xnn_f32_gemm_relu_ukernel_4x4__wasm()
176 c2[0] = vacc20; in xnn_f32_gemm_relu_ukernel_4x4__wasm()
D4x4-relu-scalar.c68 float vacc20 = vacc00; in xnn_f32_gemm_relu_ukernel_4x4__scalar() local
98 vacc20 += va2 * vb0; in xnn_f32_gemm_relu_ukernel_4x4__scalar()
118 vacc20 = math_max_f32(vacc20, 0.0f); in xnn_f32_gemm_relu_ukernel_4x4__scalar()
133 c2[0] = vacc20; in xnn_f32_gemm_relu_ukernel_4x4__scalar()
161 c2[0] = vacc20; in xnn_f32_gemm_relu_ukernel_4x4__scalar()
163 vacc20 = vacc22; in xnn_f32_gemm_relu_ukernel_4x4__scalar()
176 c2[0] = vacc20; in xnn_f32_gemm_relu_ukernel_4x4__scalar()
D4x2-relu-scalar.c64 float vacc20 = vacc00; in xnn_f32_gemm_relu_ukernel_4x2__scalar() local
84 vacc20 += va2 * vb0; in xnn_f32_gemm_relu_ukernel_4x2__scalar()
96 vacc20 = math_max_f32(vacc20, 0.0f); in xnn_f32_gemm_relu_ukernel_4x2__scalar()
105 c2[0] = vacc20; in xnn_f32_gemm_relu_ukernel_4x2__scalar()
124 c2[0] = vacc20; in xnn_f32_gemm_relu_ukernel_4x2__scalar()
D4x2-relu-wasm.c64 float vacc20 = vacc00; in xnn_f32_gemm_relu_ukernel_4x2__wasm() local
84 vacc20 += va2 * vb0; in xnn_f32_gemm_relu_ukernel_4x2__wasm()
96 vacc20 = __builtin_wasm_max_f32(vacc20, 0.0f); in xnn_f32_gemm_relu_ukernel_4x2__wasm()
105 c2[0] = vacc20; in xnn_f32_gemm_relu_ukernel_4x2__wasm()
124 c2[0] = vacc20; in xnn_f32_gemm_relu_ukernel_4x2__wasm()
D4x4-wasm.c68 float vacc20 = vacc00; in xnn_f32_gemm_ukernel_4x4__wasm() local
98 vacc20 += va2 * vb0; in xnn_f32_gemm_ukernel_4x4__wasm()
117 c2[0] = vacc20; in xnn_f32_gemm_ukernel_4x4__wasm()
145 c2[0] = vacc20; in xnn_f32_gemm_ukernel_4x4__wasm()
147 vacc20 = vacc22; in xnn_f32_gemm_ukernel_4x4__wasm()
160 c2[0] = vacc20; in xnn_f32_gemm_ukernel_4x4__wasm()
D4x4-scalar.c68 float vacc20 = vacc00; in xnn_f32_gemm_ukernel_4x4__scalar() local
98 vacc20 += va2 * vb0; in xnn_f32_gemm_ukernel_4x4__scalar()
117 c2[0] = vacc20; in xnn_f32_gemm_ukernel_4x4__scalar()
145 c2[0] = vacc20; in xnn_f32_gemm_ukernel_4x4__scalar()
147 vacc20 = vacc22; in xnn_f32_gemm_ukernel_4x4__scalar()
160 c2[0] = vacc20; in xnn_f32_gemm_ukernel_4x4__scalar()
D4x2-wasm.c64 float vacc20 = vacc00; in xnn_f32_gemm_ukernel_4x2__wasm() local
84 vacc20 += va2 * vb0; in xnn_f32_gemm_ukernel_4x2__wasm()
97 c2[0] = vacc20; in xnn_f32_gemm_ukernel_4x2__wasm()
116 c2[0] = vacc20; in xnn_f32_gemm_ukernel_4x2__wasm()
D4x2-scalar.c64 float vacc20 = vacc00; in xnn_f32_gemm_ukernel_4x2__scalar() local
84 vacc20 += va2 * vb0; in xnn_f32_gemm_ukernel_4x2__scalar()
97 c2[0] = vacc20; in xnn_f32_gemm_ukernel_4x2__scalar()
116 c2[0] = vacc20; in xnn_f32_gemm_ukernel_4x2__scalar()
/external/XNNPACK/src/f32-igemm/gen/
D4x2-minmax-wasm.c63 float vacc20 = vacc00; in xnn_f32_igemm_minmax_ukernel_4x2__wasm() local
108 vacc20 += va2 * vb0; in xnn_f32_igemm_minmax_ukernel_4x2__wasm()
122 vacc20 = __builtin_wasm_max_f32(vacc20, vmin); in xnn_f32_igemm_minmax_ukernel_4x2__wasm()
131 vacc20 = __builtin_wasm_min_f32(vacc20, vmax); in xnn_f32_igemm_minmax_ukernel_4x2__wasm()
140 c2[0] = vacc20; in xnn_f32_igemm_minmax_ukernel_4x2__wasm()
155 c2[0] = vacc20; in xnn_f32_igemm_minmax_ukernel_4x2__wasm()
D4x2-minmax-scalar.c63 float vacc20 = vacc00; in xnn_f32_igemm_minmax_ukernel_4x2__scalar() local
108 vacc20 += va2 * vb0; in xnn_f32_igemm_minmax_ukernel_4x2__scalar()
122 vacc20 = math_max_f32(vacc20, vmin); in xnn_f32_igemm_minmax_ukernel_4x2__scalar()
131 vacc20 = math_min_f32(vacc20, vmax); in xnn_f32_igemm_minmax_ukernel_4x2__scalar()
140 c2[0] = vacc20; in xnn_f32_igemm_minmax_ukernel_4x2__scalar()
155 c2[0] = vacc20; in xnn_f32_igemm_minmax_ukernel_4x2__scalar()
D4x4-minmax-scalar.c67 float vacc20 = vacc00; in xnn_f32_igemm_minmax_ukernel_4x4__scalar() local
122 vacc20 += va2 * vb0; in xnn_f32_igemm_minmax_ukernel_4x4__scalar()
144 vacc20 = math_max_f32(vacc20, vmin); in xnn_f32_igemm_minmax_ukernel_4x4__scalar()
161 vacc20 = math_min_f32(vacc20, vmax); in xnn_f32_igemm_minmax_ukernel_4x4__scalar()
176 c2[0] = vacc20; in xnn_f32_igemm_minmax_ukernel_4x4__scalar()
200 c2[0] = vacc20; in xnn_f32_igemm_minmax_ukernel_4x4__scalar()
202 vacc20 = vacc22; in xnn_f32_igemm_minmax_ukernel_4x4__scalar()
215 c2[0] = vacc20; in xnn_f32_igemm_minmax_ukernel_4x4__scalar()
D4x4-minmax-wasm.c67 float vacc20 = vacc00; in xnn_f32_igemm_minmax_ukernel_4x4__wasm() local
122 vacc20 += va2 * vb0; in xnn_f32_igemm_minmax_ukernel_4x4__wasm()
144 vacc20 = __builtin_wasm_max_f32(vacc20, vmin); in xnn_f32_igemm_minmax_ukernel_4x4__wasm()
161 vacc20 = __builtin_wasm_min_f32(vacc20, vmax); in xnn_f32_igemm_minmax_ukernel_4x4__wasm()
176 c2[0] = vacc20; in xnn_f32_igemm_minmax_ukernel_4x4__wasm()
200 c2[0] = vacc20; in xnn_f32_igemm_minmax_ukernel_4x4__wasm()
202 vacc20 = vacc22; in xnn_f32_igemm_minmax_ukernel_4x4__wasm()
215 c2[0] = vacc20; in xnn_f32_igemm_minmax_ukernel_4x4__wasm()
D4x4-relu-scalar.c65 float vacc20 = vacc00; in xnn_f32_igemm_relu_ukernel_4x4__scalar() local
120 vacc20 += va2 * vb0; in xnn_f32_igemm_relu_ukernel_4x4__scalar()
142 vacc20 = math_max_f32(vacc20, 0.0f); in xnn_f32_igemm_relu_ukernel_4x4__scalar()
157 c2[0] = vacc20; in xnn_f32_igemm_relu_ukernel_4x4__scalar()
181 c2[0] = vacc20; in xnn_f32_igemm_relu_ukernel_4x4__scalar()
183 vacc20 = vacc22; in xnn_f32_igemm_relu_ukernel_4x4__scalar()
196 c2[0] = vacc20; in xnn_f32_igemm_relu_ukernel_4x4__scalar()
D4x4-relu-wasm.c65 float vacc20 = vacc00; in xnn_f32_igemm_relu_ukernel_4x4__wasm() local
120 vacc20 += va2 * vb0; in xnn_f32_igemm_relu_ukernel_4x4__wasm()
142 vacc20 = __builtin_wasm_max_f32(vacc20, 0.0f); in xnn_f32_igemm_relu_ukernel_4x4__wasm()
157 c2[0] = vacc20; in xnn_f32_igemm_relu_ukernel_4x4__wasm()
181 c2[0] = vacc20; in xnn_f32_igemm_relu_ukernel_4x4__wasm()
183 vacc20 = vacc22; in xnn_f32_igemm_relu_ukernel_4x4__wasm()
196 c2[0] = vacc20; in xnn_f32_igemm_relu_ukernel_4x4__wasm()
D4x2-relu-scalar.c61 float vacc20 = vacc00; in xnn_f32_igemm_relu_ukernel_4x2__scalar() local
106 vacc20 += va2 * vb0; in xnn_f32_igemm_relu_ukernel_4x2__scalar()
120 vacc20 = math_max_f32(vacc20, 0.0f); in xnn_f32_igemm_relu_ukernel_4x2__scalar()
129 c2[0] = vacc20; in xnn_f32_igemm_relu_ukernel_4x2__scalar()
144 c2[0] = vacc20; in xnn_f32_igemm_relu_ukernel_4x2__scalar()
D4x2-relu-wasm.c61 float vacc20 = vacc00; in xnn_f32_igemm_relu_ukernel_4x2__wasm() local
106 vacc20 += va2 * vb0; in xnn_f32_igemm_relu_ukernel_4x2__wasm()
120 vacc20 = __builtin_wasm_max_f32(vacc20, 0.0f); in xnn_f32_igemm_relu_ukernel_4x2__wasm()
129 c2[0] = vacc20; in xnn_f32_igemm_relu_ukernel_4x2__wasm()
144 c2[0] = vacc20; in xnn_f32_igemm_relu_ukernel_4x2__wasm()
D4x4-wasm.c65 float vacc20 = vacc00; in xnn_f32_igemm_ukernel_4x4__wasm() local
120 vacc20 += va2 * vb0; in xnn_f32_igemm_ukernel_4x4__wasm()
141 c2[0] = vacc20; in xnn_f32_igemm_ukernel_4x4__wasm()
165 c2[0] = vacc20; in xnn_f32_igemm_ukernel_4x4__wasm()
167 vacc20 = vacc22; in xnn_f32_igemm_ukernel_4x4__wasm()
180 c2[0] = vacc20; in xnn_f32_igemm_ukernel_4x4__wasm()
D4x4-scalar.c65 float vacc20 = vacc00; in xnn_f32_igemm_ukernel_4x4__scalar() local
120 vacc20 += va2 * vb0; in xnn_f32_igemm_ukernel_4x4__scalar()
141 c2[0] = vacc20; in xnn_f32_igemm_ukernel_4x4__scalar()
165 c2[0] = vacc20; in xnn_f32_igemm_ukernel_4x4__scalar()
167 vacc20 = vacc22; in xnn_f32_igemm_ukernel_4x4__scalar()
180 c2[0] = vacc20; in xnn_f32_igemm_ukernel_4x4__scalar()
D4x2-scalar.c61 float vacc20 = vacc00; in xnn_f32_igemm_ukernel_4x2__scalar() local
106 vacc20 += va2 * vb0; in xnn_f32_igemm_ukernel_4x2__scalar()
121 c2[0] = vacc20; in xnn_f32_igemm_ukernel_4x2__scalar()
136 c2[0] = vacc20; in xnn_f32_igemm_ukernel_4x2__scalar()
/external/XNNPACK/src/f32-gemm/gen-inc/
D4x4inc-minmax-scalar.c71 float vacc20 = acc[8]; in xnn_f32_gemminc_minmax_ukernel_4x4__scalar() local
102 vacc20 += va2 * vb0; in xnn_f32_gemminc_minmax_ukernel_4x4__scalar()
122 vacc20 = math_max_f32(vacc20, vmin); in xnn_f32_gemminc_minmax_ukernel_4x4__scalar()
139 vacc20 = math_min_f32(vacc20, vmax); in xnn_f32_gemminc_minmax_ukernel_4x4__scalar()
154 c2[0] = vacc20; in xnn_f32_gemminc_minmax_ukernel_4x4__scalar()
182 c2[0] = vacc20; in xnn_f32_gemminc_minmax_ukernel_4x4__scalar()
184 vacc20 = vacc22; in xnn_f32_gemminc_minmax_ukernel_4x4__scalar()
197 c2[0] = vacc20; in xnn_f32_gemminc_minmax_ukernel_4x4__scalar()
D4x4inc-minmax-wasm.c71 float vacc20 = acc[8]; in xnn_f32_gemminc_minmax_ukernel_4x4__wasm() local
102 vacc20 += va2 * vb0; in xnn_f32_gemminc_minmax_ukernel_4x4__wasm()
122 vacc20 = __builtin_wasm_max_f32(vacc20, vmin); in xnn_f32_gemminc_minmax_ukernel_4x4__wasm()
139 vacc20 = __builtin_wasm_min_f32(vacc20, vmax); in xnn_f32_gemminc_minmax_ukernel_4x4__wasm()
154 c2[0] = vacc20; in xnn_f32_gemminc_minmax_ukernel_4x4__wasm()
182 c2[0] = vacc20; in xnn_f32_gemminc_minmax_ukernel_4x4__wasm()
184 vacc20 = vacc22; in xnn_f32_gemminc_minmax_ukernel_4x4__wasm()
197 c2[0] = vacc20; in xnn_f32_gemminc_minmax_ukernel_4x4__wasm()

12