/external/XNNPACK/src/f32-vbinary/gen/ |
D | vsubc-minmax-scalar-x8.c | 52 float vy7 = va7 - vb; in xnn_f32_vsubc_minmax_ukernel__scalar_x8() local 62 vy7 = math_max_f32(vy7, vy_min); in xnn_f32_vsubc_minmax_ukernel__scalar_x8() 71 vy7 = math_min_f32(vy7, vy_max); in xnn_f32_vsubc_minmax_ukernel__scalar_x8() 80 y[7] = vy7; in xnn_f32_vsubc_minmax_ukernel__scalar_x8()
|
D | vsubc-minmax-wasm-x8.c | 52 float vy7 = va7 - vb; in xnn_f32_vsubc_minmax_ukernel__wasm_x8() local 62 vy7 = __builtin_wasm_max_f32(vy7, vy_min); in xnn_f32_vsubc_minmax_ukernel__wasm_x8() 71 vy7 = __builtin_wasm_min_f32(vy7, vy_max); in xnn_f32_vsubc_minmax_ukernel__wasm_x8() 80 y[7] = vy7; in xnn_f32_vsubc_minmax_ukernel__wasm_x8()
|
D | vrdivc-minmax-wasm-x8.c | 52 float vy7 = vb / va7; in xnn_f32_vrdivc_minmax_ukernel__wasm_x8() local 62 vy7 = __builtin_wasm_max_f32(vy7, vy_min); in xnn_f32_vrdivc_minmax_ukernel__wasm_x8() 71 vy7 = __builtin_wasm_min_f32(vy7, vy_max); in xnn_f32_vrdivc_minmax_ukernel__wasm_x8() 80 y[7] = vy7; in xnn_f32_vrdivc_minmax_ukernel__wasm_x8()
|
D | vaddc-minmax-wasm-x8.c | 52 float vy7 = va7 + vb; in xnn_f32_vaddc_minmax_ukernel__wasm_x8() local 62 vy7 = __builtin_wasm_max_f32(vy7, vy_min); in xnn_f32_vaddc_minmax_ukernel__wasm_x8() 71 vy7 = __builtin_wasm_min_f32(vy7, vy_max); in xnn_f32_vaddc_minmax_ukernel__wasm_x8() 80 y[7] = vy7; in xnn_f32_vaddc_minmax_ukernel__wasm_x8()
|
D | vdivc-minmax-scalar-x8.c | 52 float vy7 = va7 / vb; in xnn_f32_vdivc_minmax_ukernel__scalar_x8() local 62 vy7 = math_max_f32(vy7, vy_min); in xnn_f32_vdivc_minmax_ukernel__scalar_x8() 71 vy7 = math_min_f32(vy7, vy_max); in xnn_f32_vdivc_minmax_ukernel__scalar_x8() 80 y[7] = vy7; in xnn_f32_vdivc_minmax_ukernel__scalar_x8()
|
D | vdivc-minmax-wasm-x8.c | 52 float vy7 = va7 / vb; in xnn_f32_vdivc_minmax_ukernel__wasm_x8() local 62 vy7 = __builtin_wasm_max_f32(vy7, vy_min); in xnn_f32_vdivc_minmax_ukernel__wasm_x8() 71 vy7 = __builtin_wasm_min_f32(vy7, vy_max); in xnn_f32_vdivc_minmax_ukernel__wasm_x8() 80 y[7] = vy7; in xnn_f32_vdivc_minmax_ukernel__wasm_x8()
|
D | vrdivc-minmax-scalar-x8.c | 52 float vy7 = vb / va7; in xnn_f32_vrdivc_minmax_ukernel__scalar_x8() local 62 vy7 = math_max_f32(vy7, vy_min); in xnn_f32_vrdivc_minmax_ukernel__scalar_x8() 71 vy7 = math_min_f32(vy7, vy_max); in xnn_f32_vrdivc_minmax_ukernel__scalar_x8() 80 y[7] = vy7; in xnn_f32_vrdivc_minmax_ukernel__scalar_x8()
|
D | vmulc-minmax-scalar-x8.c | 52 float vy7 = va7 * vb; in xnn_f32_vmulc_minmax_ukernel__scalar_x8() local 62 vy7 = math_max_f32(vy7, vy_min); in xnn_f32_vmulc_minmax_ukernel__scalar_x8() 71 vy7 = math_min_f32(vy7, vy_max); in xnn_f32_vmulc_minmax_ukernel__scalar_x8() 80 y[7] = vy7; in xnn_f32_vmulc_minmax_ukernel__scalar_x8()
|
D | vmulc-minmax-wasm-x8.c | 52 float vy7 = va7 * vb; in xnn_f32_vmulc_minmax_ukernel__wasm_x8() local 62 vy7 = __builtin_wasm_max_f32(vy7, vy_min); in xnn_f32_vmulc_minmax_ukernel__wasm_x8() 71 vy7 = __builtin_wasm_min_f32(vy7, vy_max); in xnn_f32_vmulc_minmax_ukernel__wasm_x8() 80 y[7] = vy7; in xnn_f32_vmulc_minmax_ukernel__wasm_x8()
|
D | vrsubc-minmax-scalar-x8.c | 52 float vy7 = vb - va7; in xnn_f32_vrsubc_minmax_ukernel__scalar_x8() local 62 vy7 = math_max_f32(vy7, vy_min); in xnn_f32_vrsubc_minmax_ukernel__scalar_x8() 71 vy7 = math_min_f32(vy7, vy_max); in xnn_f32_vrsubc_minmax_ukernel__scalar_x8() 80 y[7] = vy7; in xnn_f32_vrsubc_minmax_ukernel__scalar_x8()
|
D | vrsubc-minmax-wasm-x8.c | 52 float vy7 = vb - va7; in xnn_f32_vrsubc_minmax_ukernel__wasm_x8() local 62 vy7 = __builtin_wasm_max_f32(vy7, vy_min); in xnn_f32_vrsubc_minmax_ukernel__wasm_x8() 71 vy7 = __builtin_wasm_min_f32(vy7, vy_max); in xnn_f32_vrsubc_minmax_ukernel__wasm_x8() 80 y[7] = vy7; in xnn_f32_vrsubc_minmax_ukernel__wasm_x8()
|
D | vaddc-minmax-scalar-x8.c | 52 float vy7 = va7 + vb; in xnn_f32_vaddc_minmax_ukernel__scalar_x8() local 62 vy7 = math_max_f32(vy7, vy_min); in xnn_f32_vaddc_minmax_ukernel__scalar_x8() 71 vy7 = math_min_f32(vy7, vy_max); in xnn_f32_vaddc_minmax_ukernel__scalar_x8() 80 y[7] = vy7; in xnn_f32_vaddc_minmax_ukernel__scalar_x8()
|
D | vsub-minmax-scalar-x8.c | 61 float vy7 = va7 - vb7; in xnn_f32_vsub_minmax_ukernel__scalar_x8() local 71 vy7 = math_max_f32(vy7, vy_min); in xnn_f32_vsub_minmax_ukernel__scalar_x8() 80 vy7 = math_min_f32(vy7, vy_max); in xnn_f32_vsub_minmax_ukernel__scalar_x8() 89 y[7] = vy7; in xnn_f32_vsub_minmax_ukernel__scalar_x8()
|
D | vmul-minmax-scalar-x8.c | 61 float vy7 = va7 * vb7; in xnn_f32_vmul_minmax_ukernel__scalar_x8() local 71 vy7 = math_max_f32(vy7, vy_min); in xnn_f32_vmul_minmax_ukernel__scalar_x8() 80 vy7 = math_min_f32(vy7, vy_max); in xnn_f32_vmul_minmax_ukernel__scalar_x8() 89 y[7] = vy7; in xnn_f32_vmul_minmax_ukernel__scalar_x8()
|
D | vmul-minmax-wasm-x8.c | 61 float vy7 = va7 * vb7; in xnn_f32_vmul_minmax_ukernel__wasm_x8() local 71 vy7 = __builtin_wasm_max_f32(vy7, vy_min); in xnn_f32_vmul_minmax_ukernel__wasm_x8() 80 vy7 = __builtin_wasm_min_f32(vy7, vy_max); in xnn_f32_vmul_minmax_ukernel__wasm_x8() 89 y[7] = vy7; in xnn_f32_vmul_minmax_ukernel__wasm_x8()
|
D | vdiv-minmax-scalar-x8.c | 61 float vy7 = va7 / vb7; in xnn_f32_vdiv_minmax_ukernel__scalar_x8() local 71 vy7 = math_max_f32(vy7, vy_min); in xnn_f32_vdiv_minmax_ukernel__scalar_x8() 80 vy7 = math_min_f32(vy7, vy_max); in xnn_f32_vdiv_minmax_ukernel__scalar_x8() 89 y[7] = vy7; in xnn_f32_vdiv_minmax_ukernel__scalar_x8()
|
D | vdiv-minmax-wasm-x8.c | 61 float vy7 = va7 / vb7; in xnn_f32_vdiv_minmax_ukernel__wasm_x8() local 71 vy7 = __builtin_wasm_max_f32(vy7, vy_min); in xnn_f32_vdiv_minmax_ukernel__wasm_x8() 80 vy7 = __builtin_wasm_min_f32(vy7, vy_max); in xnn_f32_vdiv_minmax_ukernel__wasm_x8() 89 y[7] = vy7; in xnn_f32_vdiv_minmax_ukernel__wasm_x8()
|
D | vadd-minmax-wasm-x8.c | 61 float vy7 = va7 + vb7; in xnn_f32_vadd_minmax_ukernel__wasm_x8() local 71 vy7 = __builtin_wasm_max_f32(vy7, vy_min); in xnn_f32_vadd_minmax_ukernel__wasm_x8() 80 vy7 = __builtin_wasm_min_f32(vy7, vy_max); in xnn_f32_vadd_minmax_ukernel__wasm_x8() 89 y[7] = vy7; in xnn_f32_vadd_minmax_ukernel__wasm_x8()
|
D | vsub-minmax-wasm-x8.c | 61 float vy7 = va7 - vb7; in xnn_f32_vsub_minmax_ukernel__wasm_x8() local 71 vy7 = __builtin_wasm_max_f32(vy7, vy_min); in xnn_f32_vsub_minmax_ukernel__wasm_x8() 80 vy7 = __builtin_wasm_min_f32(vy7, vy_max); in xnn_f32_vsub_minmax_ukernel__wasm_x8() 89 y[7] = vy7; in xnn_f32_vsub_minmax_ukernel__wasm_x8()
|
D | vadd-minmax-scalar-x8.c | 61 float vy7 = va7 + vb7; in xnn_f32_vadd_minmax_ukernel__scalar_x8() local 71 vy7 = math_max_f32(vy7, vy_min); in xnn_f32_vadd_minmax_ukernel__scalar_x8() 80 vy7 = math_min_f32(vy7, vy_max); in xnn_f32_vadd_minmax_ukernel__scalar_x8() 89 y[7] = vy7; in xnn_f32_vadd_minmax_ukernel__scalar_x8()
|
D | vsqrdiffc-scalar-x8.c | 50 float vy7 = va7 - vb; in xnn_f32_vsqrdiffc_ukernel__scalar_x8() local 59 vy7 = vy7 * vy7; in xnn_f32_vsqrdiffc_ukernel__scalar_x8() 69 y[7] = vy7; in xnn_f32_vsqrdiffc_ukernel__scalar_x8()
|
D | vsqrdiff-scalar-x8.c | 59 float vy7 = va7 - vb7; in xnn_f32_vsqrdiff_ukernel__scalar_x8() local 68 vy7 = vy7 * vy7; in xnn_f32_vsqrdiff_ukernel__scalar_x8() 78 y[7] = vy7; in xnn_f32_vsqrdiff_ukernel__scalar_x8()
|
D | vrdivc-relu-scalar-x8.c | 50 float vy7 = vb / va7; in xnn_f32_vrdivc_relu_ukernel__scalar_x8() local 60 vy7 = math_max_f32(vy7, 0.0f); in xnn_f32_vrdivc_relu_ukernel__scalar_x8() 69 y[7] = vy7; in xnn_f32_vrdivc_relu_ukernel__scalar_x8()
|
D | vrsubc-relu-wasm-x8.c | 50 float vy7 = vb - va7; in xnn_f32_vrsubc_relu_ukernel__wasm_x8() local 60 vy7 = __builtin_wasm_max_f32(vy7, 0.0f); in xnn_f32_vrsubc_relu_ukernel__wasm_x8() 69 y[7] = vy7; in xnn_f32_vrsubc_relu_ukernel__wasm_x8()
|
D | vaddc-relu-scalar-x8.c | 50 float vy7 = va7 + vb; in xnn_f32_vaddc_relu_ukernel__scalar_x8() local 60 vy7 = math_max_f32(vy7, 0.0f); in xnn_f32_vaddc_relu_ukernel__scalar_x8() 69 y[7] = vy7; in xnn_f32_vaddc_relu_ukernel__scalar_x8()
|