/external/XNNPACK/src/f32-gemm/gen/ |
D | 4x2-minmax-scalar.c | 66 float vacc20 = vacc00; in xnn_f32_gemm_minmax_ukernel_4x2__scalar() local 86 vacc20 += va2 * vb0; in xnn_f32_gemm_minmax_ukernel_4x2__scalar() 98 vacc20 = math_max_f32(vacc20, vmin); in xnn_f32_gemm_minmax_ukernel_4x2__scalar() 107 vacc20 = math_min_f32(vacc20, vmax); in xnn_f32_gemm_minmax_ukernel_4x2__scalar() 116 c2[0] = vacc20; in xnn_f32_gemm_minmax_ukernel_4x2__scalar() 135 c2[0] = vacc20; in xnn_f32_gemm_minmax_ukernel_4x2__scalar()
|
D | 4x2-minmax-wasm.c | 66 float vacc20 = vacc00; in xnn_f32_gemm_minmax_ukernel_4x2__wasm() local 86 vacc20 += va2 * vb0; in xnn_f32_gemm_minmax_ukernel_4x2__wasm() 98 vacc20 = __builtin_wasm_max_f32(vacc20, vmin); in xnn_f32_gemm_minmax_ukernel_4x2__wasm() 107 vacc20 = __builtin_wasm_min_f32(vacc20, vmax); in xnn_f32_gemm_minmax_ukernel_4x2__wasm() 116 c2[0] = vacc20; in xnn_f32_gemm_minmax_ukernel_4x2__wasm() 135 c2[0] = vacc20; in xnn_f32_gemm_minmax_ukernel_4x2__wasm()
|
D | 4x4-minmax-scalar.c | 70 float vacc20 = vacc00; in xnn_f32_gemm_minmax_ukernel_4x4__scalar() local 100 vacc20 += va2 * vb0; in xnn_f32_gemm_minmax_ukernel_4x4__scalar() 120 vacc20 = math_max_f32(vacc20, vmin); in xnn_f32_gemm_minmax_ukernel_4x4__scalar() 137 vacc20 = math_min_f32(vacc20, vmax); in xnn_f32_gemm_minmax_ukernel_4x4__scalar() 152 c2[0] = vacc20; in xnn_f32_gemm_minmax_ukernel_4x4__scalar() 180 c2[0] = vacc20; in xnn_f32_gemm_minmax_ukernel_4x4__scalar() 182 vacc20 = vacc22; in xnn_f32_gemm_minmax_ukernel_4x4__scalar() 195 c2[0] = vacc20; in xnn_f32_gemm_minmax_ukernel_4x4__scalar()
|
D | 4x4-minmax-wasm.c | 70 float vacc20 = vacc00; in xnn_f32_gemm_minmax_ukernel_4x4__wasm() local 100 vacc20 += va2 * vb0; in xnn_f32_gemm_minmax_ukernel_4x4__wasm() 120 vacc20 = __builtin_wasm_max_f32(vacc20, vmin); in xnn_f32_gemm_minmax_ukernel_4x4__wasm() 137 vacc20 = __builtin_wasm_min_f32(vacc20, vmax); in xnn_f32_gemm_minmax_ukernel_4x4__wasm() 152 c2[0] = vacc20; in xnn_f32_gemm_minmax_ukernel_4x4__wasm() 180 c2[0] = vacc20; in xnn_f32_gemm_minmax_ukernel_4x4__wasm() 182 vacc20 = vacc22; in xnn_f32_gemm_minmax_ukernel_4x4__wasm() 195 c2[0] = vacc20; in xnn_f32_gemm_minmax_ukernel_4x4__wasm()
|
D | 4x4-relu-wasm.c | 68 float vacc20 = vacc00; in xnn_f32_gemm_relu_ukernel_4x4__wasm() local 98 vacc20 += va2 * vb0; in xnn_f32_gemm_relu_ukernel_4x4__wasm() 118 vacc20 = __builtin_wasm_max_f32(vacc20, 0.0f); in xnn_f32_gemm_relu_ukernel_4x4__wasm() 133 c2[0] = vacc20; in xnn_f32_gemm_relu_ukernel_4x4__wasm() 161 c2[0] = vacc20; in xnn_f32_gemm_relu_ukernel_4x4__wasm() 163 vacc20 = vacc22; in xnn_f32_gemm_relu_ukernel_4x4__wasm() 176 c2[0] = vacc20; in xnn_f32_gemm_relu_ukernel_4x4__wasm()
|
D | 4x4-relu-scalar.c | 68 float vacc20 = vacc00; in xnn_f32_gemm_relu_ukernel_4x4__scalar() local 98 vacc20 += va2 * vb0; in xnn_f32_gemm_relu_ukernel_4x4__scalar() 118 vacc20 = math_max_f32(vacc20, 0.0f); in xnn_f32_gemm_relu_ukernel_4x4__scalar() 133 c2[0] = vacc20; in xnn_f32_gemm_relu_ukernel_4x4__scalar() 161 c2[0] = vacc20; in xnn_f32_gemm_relu_ukernel_4x4__scalar() 163 vacc20 = vacc22; in xnn_f32_gemm_relu_ukernel_4x4__scalar() 176 c2[0] = vacc20; in xnn_f32_gemm_relu_ukernel_4x4__scalar()
|
D | 4x2-relu-scalar.c | 64 float vacc20 = vacc00; in xnn_f32_gemm_relu_ukernel_4x2__scalar() local 84 vacc20 += va2 * vb0; in xnn_f32_gemm_relu_ukernel_4x2__scalar() 96 vacc20 = math_max_f32(vacc20, 0.0f); in xnn_f32_gemm_relu_ukernel_4x2__scalar() 105 c2[0] = vacc20; in xnn_f32_gemm_relu_ukernel_4x2__scalar() 124 c2[0] = vacc20; in xnn_f32_gemm_relu_ukernel_4x2__scalar()
|
D | 4x2-relu-wasm.c | 64 float vacc20 = vacc00; in xnn_f32_gemm_relu_ukernel_4x2__wasm() local 84 vacc20 += va2 * vb0; in xnn_f32_gemm_relu_ukernel_4x2__wasm() 96 vacc20 = __builtin_wasm_max_f32(vacc20, 0.0f); in xnn_f32_gemm_relu_ukernel_4x2__wasm() 105 c2[0] = vacc20; in xnn_f32_gemm_relu_ukernel_4x2__wasm() 124 c2[0] = vacc20; in xnn_f32_gemm_relu_ukernel_4x2__wasm()
|
D | 4x4-wasm.c | 68 float vacc20 = vacc00; in xnn_f32_gemm_ukernel_4x4__wasm() local 98 vacc20 += va2 * vb0; in xnn_f32_gemm_ukernel_4x4__wasm() 117 c2[0] = vacc20; in xnn_f32_gemm_ukernel_4x4__wasm() 145 c2[0] = vacc20; in xnn_f32_gemm_ukernel_4x4__wasm() 147 vacc20 = vacc22; in xnn_f32_gemm_ukernel_4x4__wasm() 160 c2[0] = vacc20; in xnn_f32_gemm_ukernel_4x4__wasm()
|
D | 4x4-scalar.c | 68 float vacc20 = vacc00; in xnn_f32_gemm_ukernel_4x4__scalar() local 98 vacc20 += va2 * vb0; in xnn_f32_gemm_ukernel_4x4__scalar() 117 c2[0] = vacc20; in xnn_f32_gemm_ukernel_4x4__scalar() 145 c2[0] = vacc20; in xnn_f32_gemm_ukernel_4x4__scalar() 147 vacc20 = vacc22; in xnn_f32_gemm_ukernel_4x4__scalar() 160 c2[0] = vacc20; in xnn_f32_gemm_ukernel_4x4__scalar()
|
D | 4x2-wasm.c | 64 float vacc20 = vacc00; in xnn_f32_gemm_ukernel_4x2__wasm() local 84 vacc20 += va2 * vb0; in xnn_f32_gemm_ukernel_4x2__wasm() 97 c2[0] = vacc20; in xnn_f32_gemm_ukernel_4x2__wasm() 116 c2[0] = vacc20; in xnn_f32_gemm_ukernel_4x2__wasm()
|
D | 4x2-scalar.c | 64 float vacc20 = vacc00; in xnn_f32_gemm_ukernel_4x2__scalar() local 84 vacc20 += va2 * vb0; in xnn_f32_gemm_ukernel_4x2__scalar() 97 c2[0] = vacc20; in xnn_f32_gemm_ukernel_4x2__scalar() 116 c2[0] = vacc20; in xnn_f32_gemm_ukernel_4x2__scalar()
|
/external/XNNPACK/src/f32-igemm/gen/ |
D | 4x2-minmax-wasm.c | 63 float vacc20 = vacc00; in xnn_f32_igemm_minmax_ukernel_4x2__wasm() local 108 vacc20 += va2 * vb0; in xnn_f32_igemm_minmax_ukernel_4x2__wasm() 122 vacc20 = __builtin_wasm_max_f32(vacc20, vmin); in xnn_f32_igemm_minmax_ukernel_4x2__wasm() 131 vacc20 = __builtin_wasm_min_f32(vacc20, vmax); in xnn_f32_igemm_minmax_ukernel_4x2__wasm() 140 c2[0] = vacc20; in xnn_f32_igemm_minmax_ukernel_4x2__wasm() 155 c2[0] = vacc20; in xnn_f32_igemm_minmax_ukernel_4x2__wasm()
|
D | 4x2-minmax-scalar.c | 63 float vacc20 = vacc00; in xnn_f32_igemm_minmax_ukernel_4x2__scalar() local 108 vacc20 += va2 * vb0; in xnn_f32_igemm_minmax_ukernel_4x2__scalar() 122 vacc20 = math_max_f32(vacc20, vmin); in xnn_f32_igemm_minmax_ukernel_4x2__scalar() 131 vacc20 = math_min_f32(vacc20, vmax); in xnn_f32_igemm_minmax_ukernel_4x2__scalar() 140 c2[0] = vacc20; in xnn_f32_igemm_minmax_ukernel_4x2__scalar() 155 c2[0] = vacc20; in xnn_f32_igemm_minmax_ukernel_4x2__scalar()
|
D | 4x4-minmax-scalar.c | 67 float vacc20 = vacc00; in xnn_f32_igemm_minmax_ukernel_4x4__scalar() local 122 vacc20 += va2 * vb0; in xnn_f32_igemm_minmax_ukernel_4x4__scalar() 144 vacc20 = math_max_f32(vacc20, vmin); in xnn_f32_igemm_minmax_ukernel_4x4__scalar() 161 vacc20 = math_min_f32(vacc20, vmax); in xnn_f32_igemm_minmax_ukernel_4x4__scalar() 176 c2[0] = vacc20; in xnn_f32_igemm_minmax_ukernel_4x4__scalar() 200 c2[0] = vacc20; in xnn_f32_igemm_minmax_ukernel_4x4__scalar() 202 vacc20 = vacc22; in xnn_f32_igemm_minmax_ukernel_4x4__scalar() 215 c2[0] = vacc20; in xnn_f32_igemm_minmax_ukernel_4x4__scalar()
|
D | 4x4-minmax-wasm.c | 67 float vacc20 = vacc00; in xnn_f32_igemm_minmax_ukernel_4x4__wasm() local 122 vacc20 += va2 * vb0; in xnn_f32_igemm_minmax_ukernel_4x4__wasm() 144 vacc20 = __builtin_wasm_max_f32(vacc20, vmin); in xnn_f32_igemm_minmax_ukernel_4x4__wasm() 161 vacc20 = __builtin_wasm_min_f32(vacc20, vmax); in xnn_f32_igemm_minmax_ukernel_4x4__wasm() 176 c2[0] = vacc20; in xnn_f32_igemm_minmax_ukernel_4x4__wasm() 200 c2[0] = vacc20; in xnn_f32_igemm_minmax_ukernel_4x4__wasm() 202 vacc20 = vacc22; in xnn_f32_igemm_minmax_ukernel_4x4__wasm() 215 c2[0] = vacc20; in xnn_f32_igemm_minmax_ukernel_4x4__wasm()
|
D | 4x4-relu-scalar.c | 65 float vacc20 = vacc00; in xnn_f32_igemm_relu_ukernel_4x4__scalar() local 120 vacc20 += va2 * vb0; in xnn_f32_igemm_relu_ukernel_4x4__scalar() 142 vacc20 = math_max_f32(vacc20, 0.0f); in xnn_f32_igemm_relu_ukernel_4x4__scalar() 157 c2[0] = vacc20; in xnn_f32_igemm_relu_ukernel_4x4__scalar() 181 c2[0] = vacc20; in xnn_f32_igemm_relu_ukernel_4x4__scalar() 183 vacc20 = vacc22; in xnn_f32_igemm_relu_ukernel_4x4__scalar() 196 c2[0] = vacc20; in xnn_f32_igemm_relu_ukernel_4x4__scalar()
|
D | 4x4-relu-wasm.c | 65 float vacc20 = vacc00; in xnn_f32_igemm_relu_ukernel_4x4__wasm() local 120 vacc20 += va2 * vb0; in xnn_f32_igemm_relu_ukernel_4x4__wasm() 142 vacc20 = __builtin_wasm_max_f32(vacc20, 0.0f); in xnn_f32_igemm_relu_ukernel_4x4__wasm() 157 c2[0] = vacc20; in xnn_f32_igemm_relu_ukernel_4x4__wasm() 181 c2[0] = vacc20; in xnn_f32_igemm_relu_ukernel_4x4__wasm() 183 vacc20 = vacc22; in xnn_f32_igemm_relu_ukernel_4x4__wasm() 196 c2[0] = vacc20; in xnn_f32_igemm_relu_ukernel_4x4__wasm()
|
D | 4x2-relu-scalar.c | 61 float vacc20 = vacc00; in xnn_f32_igemm_relu_ukernel_4x2__scalar() local 106 vacc20 += va2 * vb0; in xnn_f32_igemm_relu_ukernel_4x2__scalar() 120 vacc20 = math_max_f32(vacc20, 0.0f); in xnn_f32_igemm_relu_ukernel_4x2__scalar() 129 c2[0] = vacc20; in xnn_f32_igemm_relu_ukernel_4x2__scalar() 144 c2[0] = vacc20; in xnn_f32_igemm_relu_ukernel_4x2__scalar()
|
D | 4x2-relu-wasm.c | 61 float vacc20 = vacc00; in xnn_f32_igemm_relu_ukernel_4x2__wasm() local 106 vacc20 += va2 * vb0; in xnn_f32_igemm_relu_ukernel_4x2__wasm() 120 vacc20 = __builtin_wasm_max_f32(vacc20, 0.0f); in xnn_f32_igemm_relu_ukernel_4x2__wasm() 129 c2[0] = vacc20; in xnn_f32_igemm_relu_ukernel_4x2__wasm() 144 c2[0] = vacc20; in xnn_f32_igemm_relu_ukernel_4x2__wasm()
|
D | 4x4-wasm.c | 65 float vacc20 = vacc00; in xnn_f32_igemm_ukernel_4x4__wasm() local 120 vacc20 += va2 * vb0; in xnn_f32_igemm_ukernel_4x4__wasm() 141 c2[0] = vacc20; in xnn_f32_igemm_ukernel_4x4__wasm() 165 c2[0] = vacc20; in xnn_f32_igemm_ukernel_4x4__wasm() 167 vacc20 = vacc22; in xnn_f32_igemm_ukernel_4x4__wasm() 180 c2[0] = vacc20; in xnn_f32_igemm_ukernel_4x4__wasm()
|
D | 4x4-scalar.c | 65 float vacc20 = vacc00; in xnn_f32_igemm_ukernel_4x4__scalar() local 120 vacc20 += va2 * vb0; in xnn_f32_igemm_ukernel_4x4__scalar() 141 c2[0] = vacc20; in xnn_f32_igemm_ukernel_4x4__scalar() 165 c2[0] = vacc20; in xnn_f32_igemm_ukernel_4x4__scalar() 167 vacc20 = vacc22; in xnn_f32_igemm_ukernel_4x4__scalar() 180 c2[0] = vacc20; in xnn_f32_igemm_ukernel_4x4__scalar()
|
D | 4x2-scalar.c | 61 float vacc20 = vacc00; in xnn_f32_igemm_ukernel_4x2__scalar() local 106 vacc20 += va2 * vb0; in xnn_f32_igemm_ukernel_4x2__scalar() 121 c2[0] = vacc20; in xnn_f32_igemm_ukernel_4x2__scalar() 136 c2[0] = vacc20; in xnn_f32_igemm_ukernel_4x2__scalar()
|
/external/XNNPACK/src/f32-gemm/gen-inc/ |
D | 4x4inc-minmax-scalar.c | 71 float vacc20 = acc[8]; in xnn_f32_gemminc_minmax_ukernel_4x4__scalar() local 102 vacc20 += va2 * vb0; in xnn_f32_gemminc_minmax_ukernel_4x4__scalar() 122 vacc20 = math_max_f32(vacc20, vmin); in xnn_f32_gemminc_minmax_ukernel_4x4__scalar() 139 vacc20 = math_min_f32(vacc20, vmax); in xnn_f32_gemminc_minmax_ukernel_4x4__scalar() 154 c2[0] = vacc20; in xnn_f32_gemminc_minmax_ukernel_4x4__scalar() 182 c2[0] = vacc20; in xnn_f32_gemminc_minmax_ukernel_4x4__scalar() 184 vacc20 = vacc22; in xnn_f32_gemminc_minmax_ukernel_4x4__scalar() 197 c2[0] = vacc20; in xnn_f32_gemminc_minmax_ukernel_4x4__scalar()
|
D | 4x4inc-minmax-wasm.c | 71 float vacc20 = acc[8]; in xnn_f32_gemminc_minmax_ukernel_4x4__wasm() local 102 vacc20 += va2 * vb0; in xnn_f32_gemminc_minmax_ukernel_4x4__wasm() 122 vacc20 = __builtin_wasm_max_f32(vacc20, vmin); in xnn_f32_gemminc_minmax_ukernel_4x4__wasm() 139 vacc20 = __builtin_wasm_min_f32(vacc20, vmax); in xnn_f32_gemminc_minmax_ukernel_4x4__wasm() 154 c2[0] = vacc20; in xnn_f32_gemminc_minmax_ukernel_4x4__wasm() 182 c2[0] = vacc20; in xnn_f32_gemminc_minmax_ukernel_4x4__wasm() 184 vacc20 = vacc22; in xnn_f32_gemminc_minmax_ukernel_4x4__wasm() 197 c2[0] = vacc20; in xnn_f32_gemminc_minmax_ukernel_4x4__wasm()
|