/external/XNNPACK/src/f32-vbinary/gen/ |
D | vsubc-minmax-scalar-x8.c | 49 float vy4 = va4 - vb; in xnn_f32_vsubc_minmax_ukernel__scalar_x8() local 59 vy4 = math_max_f32(vy4, vy_min); in xnn_f32_vsubc_minmax_ukernel__scalar_x8() 68 vy4 = math_min_f32(vy4, vy_max); in xnn_f32_vsubc_minmax_ukernel__scalar_x8() 77 y[4] = vy4; in xnn_f32_vsubc_minmax_ukernel__scalar_x8()
|
D | vsubc-minmax-wasm-x8.c | 49 float vy4 = va4 - vb; in xnn_f32_vsubc_minmax_ukernel__wasm_x8() local 59 vy4 = __builtin_wasm_max_f32(vy4, vy_min); in xnn_f32_vsubc_minmax_ukernel__wasm_x8() 68 vy4 = __builtin_wasm_min_f32(vy4, vy_max); in xnn_f32_vsubc_minmax_ukernel__wasm_x8() 77 y[4] = vy4; in xnn_f32_vsubc_minmax_ukernel__wasm_x8()
|
D | vrdivc-minmax-wasm-x8.c | 49 float vy4 = vb / va4; in xnn_f32_vrdivc_minmax_ukernel__wasm_x8() local 59 vy4 = __builtin_wasm_max_f32(vy4, vy_min); in xnn_f32_vrdivc_minmax_ukernel__wasm_x8() 68 vy4 = __builtin_wasm_min_f32(vy4, vy_max); in xnn_f32_vrdivc_minmax_ukernel__wasm_x8() 77 y[4] = vy4; in xnn_f32_vrdivc_minmax_ukernel__wasm_x8()
|
D | vaddc-minmax-wasm-x8.c | 49 float vy4 = va4 + vb; in xnn_f32_vaddc_minmax_ukernel__wasm_x8() local 59 vy4 = __builtin_wasm_max_f32(vy4, vy_min); in xnn_f32_vaddc_minmax_ukernel__wasm_x8() 68 vy4 = __builtin_wasm_min_f32(vy4, vy_max); in xnn_f32_vaddc_minmax_ukernel__wasm_x8() 77 y[4] = vy4; in xnn_f32_vaddc_minmax_ukernel__wasm_x8()
|
D | vdivc-minmax-scalar-x8.c | 49 float vy4 = va4 / vb; in xnn_f32_vdivc_minmax_ukernel__scalar_x8() local 59 vy4 = math_max_f32(vy4, vy_min); in xnn_f32_vdivc_minmax_ukernel__scalar_x8() 68 vy4 = math_min_f32(vy4, vy_max); in xnn_f32_vdivc_minmax_ukernel__scalar_x8() 77 y[4] = vy4; in xnn_f32_vdivc_minmax_ukernel__scalar_x8()
|
D | vdivc-minmax-wasm-x8.c | 49 float vy4 = va4 / vb; in xnn_f32_vdivc_minmax_ukernel__wasm_x8() local 59 vy4 = __builtin_wasm_max_f32(vy4, vy_min); in xnn_f32_vdivc_minmax_ukernel__wasm_x8() 68 vy4 = __builtin_wasm_min_f32(vy4, vy_max); in xnn_f32_vdivc_minmax_ukernel__wasm_x8() 77 y[4] = vy4; in xnn_f32_vdivc_minmax_ukernel__wasm_x8()
|
D | vrdivc-minmax-scalar-x8.c | 49 float vy4 = vb / va4; in xnn_f32_vrdivc_minmax_ukernel__scalar_x8() local 59 vy4 = math_max_f32(vy4, vy_min); in xnn_f32_vrdivc_minmax_ukernel__scalar_x8() 68 vy4 = math_min_f32(vy4, vy_max); in xnn_f32_vrdivc_minmax_ukernel__scalar_x8() 77 y[4] = vy4; in xnn_f32_vrdivc_minmax_ukernel__scalar_x8()
|
D | vmulc-minmax-scalar-x8.c | 49 float vy4 = va4 * vb; in xnn_f32_vmulc_minmax_ukernel__scalar_x8() local 59 vy4 = math_max_f32(vy4, vy_min); in xnn_f32_vmulc_minmax_ukernel__scalar_x8() 68 vy4 = math_min_f32(vy4, vy_max); in xnn_f32_vmulc_minmax_ukernel__scalar_x8() 77 y[4] = vy4; in xnn_f32_vmulc_minmax_ukernel__scalar_x8()
|
D | vmulc-minmax-wasm-x8.c | 49 float vy4 = va4 * vb; in xnn_f32_vmulc_minmax_ukernel__wasm_x8() local 59 vy4 = __builtin_wasm_max_f32(vy4, vy_min); in xnn_f32_vmulc_minmax_ukernel__wasm_x8() 68 vy4 = __builtin_wasm_min_f32(vy4, vy_max); in xnn_f32_vmulc_minmax_ukernel__wasm_x8() 77 y[4] = vy4; in xnn_f32_vmulc_minmax_ukernel__wasm_x8()
|
D | vrsubc-minmax-scalar-x8.c | 49 float vy4 = vb - va4; in xnn_f32_vrsubc_minmax_ukernel__scalar_x8() local 59 vy4 = math_max_f32(vy4, vy_min); in xnn_f32_vrsubc_minmax_ukernel__scalar_x8() 68 vy4 = math_min_f32(vy4, vy_max); in xnn_f32_vrsubc_minmax_ukernel__scalar_x8() 77 y[4] = vy4; in xnn_f32_vrsubc_minmax_ukernel__scalar_x8()
|
D | vrsubc-minmax-wasm-x8.c | 49 float vy4 = vb - va4; in xnn_f32_vrsubc_minmax_ukernel__wasm_x8() local 59 vy4 = __builtin_wasm_max_f32(vy4, vy_min); in xnn_f32_vrsubc_minmax_ukernel__wasm_x8() 68 vy4 = __builtin_wasm_min_f32(vy4, vy_max); in xnn_f32_vrsubc_minmax_ukernel__wasm_x8() 77 y[4] = vy4; in xnn_f32_vrsubc_minmax_ukernel__wasm_x8()
|
D | vaddc-minmax-scalar-x8.c | 49 float vy4 = va4 + vb; in xnn_f32_vaddc_minmax_ukernel__scalar_x8() local 59 vy4 = math_max_f32(vy4, vy_min); in xnn_f32_vaddc_minmax_ukernel__scalar_x8() 68 vy4 = math_min_f32(vy4, vy_max); in xnn_f32_vaddc_minmax_ukernel__scalar_x8() 77 y[4] = vy4; in xnn_f32_vaddc_minmax_ukernel__scalar_x8()
|
D | vsub-minmax-scalar-x8.c | 58 float vy4 = va4 - vb4; in xnn_f32_vsub_minmax_ukernel__scalar_x8() local 68 vy4 = math_max_f32(vy4, vy_min); in xnn_f32_vsub_minmax_ukernel__scalar_x8() 77 vy4 = math_min_f32(vy4, vy_max); in xnn_f32_vsub_minmax_ukernel__scalar_x8() 86 y[4] = vy4; in xnn_f32_vsub_minmax_ukernel__scalar_x8()
|
D | vmul-minmax-scalar-x8.c | 58 float vy4 = va4 * vb4; in xnn_f32_vmul_minmax_ukernel__scalar_x8() local 68 vy4 = math_max_f32(vy4, vy_min); in xnn_f32_vmul_minmax_ukernel__scalar_x8() 77 vy4 = math_min_f32(vy4, vy_max); in xnn_f32_vmul_minmax_ukernel__scalar_x8() 86 y[4] = vy4; in xnn_f32_vmul_minmax_ukernel__scalar_x8()
|
D | vmul-minmax-wasm-x8.c | 58 float vy4 = va4 * vb4; in xnn_f32_vmul_minmax_ukernel__wasm_x8() local 68 vy4 = __builtin_wasm_max_f32(vy4, vy_min); in xnn_f32_vmul_minmax_ukernel__wasm_x8() 77 vy4 = __builtin_wasm_min_f32(vy4, vy_max); in xnn_f32_vmul_minmax_ukernel__wasm_x8() 86 y[4] = vy4; in xnn_f32_vmul_minmax_ukernel__wasm_x8()
|
D | vdiv-minmax-scalar-x8.c | 58 float vy4 = va4 / vb4; in xnn_f32_vdiv_minmax_ukernel__scalar_x8() local 68 vy4 = math_max_f32(vy4, vy_min); in xnn_f32_vdiv_minmax_ukernel__scalar_x8() 77 vy4 = math_min_f32(vy4, vy_max); in xnn_f32_vdiv_minmax_ukernel__scalar_x8() 86 y[4] = vy4; in xnn_f32_vdiv_minmax_ukernel__scalar_x8()
|
D | vdiv-minmax-wasm-x8.c | 58 float vy4 = va4 / vb4; in xnn_f32_vdiv_minmax_ukernel__wasm_x8() local 68 vy4 = __builtin_wasm_max_f32(vy4, vy_min); in xnn_f32_vdiv_minmax_ukernel__wasm_x8() 77 vy4 = __builtin_wasm_min_f32(vy4, vy_max); in xnn_f32_vdiv_minmax_ukernel__wasm_x8() 86 y[4] = vy4; in xnn_f32_vdiv_minmax_ukernel__wasm_x8()
|
D | vadd-minmax-wasm-x8.c | 58 float vy4 = va4 + vb4; in xnn_f32_vadd_minmax_ukernel__wasm_x8() local 68 vy4 = __builtin_wasm_max_f32(vy4, vy_min); in xnn_f32_vadd_minmax_ukernel__wasm_x8() 77 vy4 = __builtin_wasm_min_f32(vy4, vy_max); in xnn_f32_vadd_minmax_ukernel__wasm_x8() 86 y[4] = vy4; in xnn_f32_vadd_minmax_ukernel__wasm_x8()
|
D | vsub-minmax-wasm-x8.c | 58 float vy4 = va4 - vb4; in xnn_f32_vsub_minmax_ukernel__wasm_x8() local 68 vy4 = __builtin_wasm_max_f32(vy4, vy_min); in xnn_f32_vsub_minmax_ukernel__wasm_x8() 77 vy4 = __builtin_wasm_min_f32(vy4, vy_max); in xnn_f32_vsub_minmax_ukernel__wasm_x8() 86 y[4] = vy4; in xnn_f32_vsub_minmax_ukernel__wasm_x8()
|
D | vadd-minmax-scalar-x8.c | 58 float vy4 = va4 + vb4; in xnn_f32_vadd_minmax_ukernel__scalar_x8() local 68 vy4 = math_max_f32(vy4, vy_min); in xnn_f32_vadd_minmax_ukernel__scalar_x8() 77 vy4 = math_min_f32(vy4, vy_max); in xnn_f32_vadd_minmax_ukernel__scalar_x8() 86 y[4] = vy4; in xnn_f32_vadd_minmax_ukernel__scalar_x8()
|
D | vsqrdiffc-scalar-x8.c | 47 float vy4 = va4 - vb; in xnn_f32_vsqrdiffc_ukernel__scalar_x8() local 56 vy4 = vy4 * vy4; in xnn_f32_vsqrdiffc_ukernel__scalar_x8() 66 y[4] = vy4; in xnn_f32_vsqrdiffc_ukernel__scalar_x8()
|
D | vsqrdiff-scalar-x8.c | 56 float vy4 = va4 - vb4; in xnn_f32_vsqrdiff_ukernel__scalar_x8() local 65 vy4 = vy4 * vy4; in xnn_f32_vsqrdiff_ukernel__scalar_x8() 75 y[4] = vy4; in xnn_f32_vsqrdiff_ukernel__scalar_x8()
|
D | vrdivc-relu-scalar-x8.c | 47 float vy4 = vb / va4; in xnn_f32_vrdivc_relu_ukernel__scalar_x8() local 57 vy4 = math_max_f32(vy4, 0.0f); in xnn_f32_vrdivc_relu_ukernel__scalar_x8() 66 y[4] = vy4; in xnn_f32_vrdivc_relu_ukernel__scalar_x8()
|
D | vrsubc-relu-wasm-x8.c | 47 float vy4 = vb - va4; in xnn_f32_vrsubc_relu_ukernel__wasm_x8() local 57 vy4 = __builtin_wasm_max_f32(vy4, 0.0f); in xnn_f32_vrsubc_relu_ukernel__wasm_x8() 66 y[4] = vy4; in xnn_f32_vrsubc_relu_ukernel__wasm_x8()
|
D | vaddc-relu-scalar-x8.c | 47 float vy4 = va4 + vb; in xnn_f32_vaddc_relu_ukernel__scalar_x8() local 57 vy4 = math_max_f32(vy4, 0.0f); in xnn_f32_vaddc_relu_ukernel__scalar_x8() 66 y[4] = vy4; in xnn_f32_vaddc_relu_ukernel__scalar_x8()
|