/external/XNNPACK/src/qs8-gemm/gen/ |
D | 1x4c8-minmax-wasmsimd-ld128.c | 96 const v128_t vprod0x01 = wasm_i64x2_add(wasm_i64x2_mul(vacc0x01, vmultiplier), vrounding); in xnn_qs8_gemm_minmax_ukernel_1x4c8__wasmsimd_ld128() local 101 const v128_t vq31prod0x0123 = wasm_v32x4_shuffle(vprod0x01, vprod0x23, 1, 3, 5, 7); in xnn_qs8_gemm_minmax_ukernel_1x4c8__wasmsimd_ld128()
|
D | 1x4c8-minmax-wasmsimd-ld64.c | 92 const v128_t vprod0x01 = wasm_i64x2_add(wasm_i64x2_mul(vacc0x01, vmultiplier), vrounding); in xnn_qs8_gemm_minmax_ukernel_1x4c8__wasmsimd_ld64() local 97 const v128_t vq31prod0x0123 = wasm_v32x4_shuffle(vprod0x01, vprod0x23, 1, 3, 5, 7); in xnn_qs8_gemm_minmax_ukernel_1x4c8__wasmsimd_ld64()
|
D | 1x4c8-xw-minmax-wasmsimd.c | 92 const v128_t vprod0x01 = wasm_i64x2_add(wasm_i64x2_mul(vacc0x01, vmultiplier), vrounding); in xnn_qs8_gemm_xw_minmax_ukernel_1x4c8__wasmsimd() local 97 const v128_t vq31prod0x0123 = wasm_v32x4_shuffle(vprod0x01, vprod0x23, 1, 3, 5, 7); in xnn_qs8_gemm_xw_minmax_ukernel_1x4c8__wasmsimd()
|
D | 2x4c8-minmax-wasmsimd-ld128.c | 125 const v128_t vprod0x01 = wasm_i64x2_add(wasm_i64x2_mul(vacc0x01, vmultiplier), vrounding); in xnn_qs8_gemm_minmax_ukernel_2x4c8__wasmsimd_ld128() local 133 const v128_t vq31prod0x0123 = wasm_v32x4_shuffle(vprod0x01, vprod0x23, 1, 3, 5, 7); in xnn_qs8_gemm_minmax_ukernel_2x4c8__wasmsimd_ld128()
|
D | 2x4c8-xw-minmax-wasmsimd.c | 121 const v128_t vprod0x01 = wasm_i64x2_add(wasm_i64x2_mul(vacc0x01, vmultiplier), vrounding); in xnn_qs8_gemm_xw_minmax_ukernel_2x4c8__wasmsimd() local 129 const v128_t vq31prod0x0123 = wasm_v32x4_shuffle(vprod0x01, vprod0x23, 1, 3, 5, 7); in xnn_qs8_gemm_xw_minmax_ukernel_2x4c8__wasmsimd()
|
D | 2x4c8-minmax-wasmsimd-ld64.c | 121 const v128_t vprod0x01 = wasm_i64x2_add(wasm_i64x2_mul(vacc0x01, vmultiplier), vrounding); in xnn_qs8_gemm_minmax_ukernel_2x4c8__wasmsimd_ld64() local 129 const v128_t vq31prod0x0123 = wasm_v32x4_shuffle(vprod0x01, vprod0x23, 1, 3, 5, 7); in xnn_qs8_gemm_minmax_ukernel_2x4c8__wasmsimd_ld64()
|
D | 3x4c8-minmax-wasmsimd-ld128.c | 154 const v128_t vprod0x01 = wasm_i64x2_add(wasm_i64x2_mul(vacc0x01, vmultiplier), vrounding); in xnn_qs8_gemm_minmax_ukernel_3x4c8__wasmsimd_ld128() local 165 const v128_t vq31prod0x0123 = wasm_v32x4_shuffle(vprod0x01, vprod0x23, 1, 3, 5, 7); in xnn_qs8_gemm_minmax_ukernel_3x4c8__wasmsimd_ld128()
|
D | 3x4c8-minmax-wasmsimd-ld64.c | 150 const v128_t vprod0x01 = wasm_i64x2_add(wasm_i64x2_mul(vacc0x01, vmultiplier), vrounding); in xnn_qs8_gemm_minmax_ukernel_3x4c8__wasmsimd_ld64() local 161 const v128_t vq31prod0x0123 = wasm_v32x4_shuffle(vprod0x01, vprod0x23, 1, 3, 5, 7); in xnn_qs8_gemm_minmax_ukernel_3x4c8__wasmsimd_ld64()
|
D | 3x4c8-xw-minmax-wasmsimd.c | 150 const v128_t vprod0x01 = wasm_i64x2_add(wasm_i64x2_mul(vacc0x01, vmultiplier), vrounding); in xnn_qs8_gemm_xw_minmax_ukernel_3x4c8__wasmsimd() local 161 const v128_t vq31prod0x0123 = wasm_v32x4_shuffle(vprod0x01, vprod0x23, 1, 3, 5, 7); in xnn_qs8_gemm_xw_minmax_ukernel_3x4c8__wasmsimd()
|
/external/XNNPACK/src/qs8-igemm/gen/ |
D | 1x4c8-minmax-wasmsimd-ld64.c | 105 const v128_t vprod0x01 = wasm_i64x2_add(wasm_i64x2_mul(vacc0x01, vmultiplier), vrounding); in xnn_qs8_igemm_minmax_ukernel_1x4c8__wasmsimd_ld64() local 110 const v128_t vq31prod0x0123 = wasm_v32x4_shuffle(vprod0x01, vprod0x23, 1, 3, 5, 7); in xnn_qs8_igemm_minmax_ukernel_1x4c8__wasmsimd_ld64()
|
D | 1x4c8-minmax-wasmsimd-ld128.c | 109 const v128_t vprod0x01 = wasm_i64x2_add(wasm_i64x2_mul(vacc0x01, vmultiplier), vrounding); in xnn_qs8_igemm_minmax_ukernel_1x4c8__wasmsimd_ld128() local 114 const v128_t vq31prod0x0123 = wasm_v32x4_shuffle(vprod0x01, vprod0x23, 1, 3, 5, 7); in xnn_qs8_igemm_minmax_ukernel_1x4c8__wasmsimd_ld128()
|
D | 2x4c8-minmax-wasmsimd-ld64.c | 136 const v128_t vprod0x01 = wasm_i64x2_add(wasm_i64x2_mul(vacc0x01, vmultiplier), vrounding); in xnn_qs8_igemm_minmax_ukernel_2x4c8__wasmsimd_ld64() local 144 const v128_t vq31prod0x0123 = wasm_v32x4_shuffle(vprod0x01, vprod0x23, 1, 3, 5, 7); in xnn_qs8_igemm_minmax_ukernel_2x4c8__wasmsimd_ld64()
|
D | 2x4c8-minmax-wasmsimd-ld128.c | 140 const v128_t vprod0x01 = wasm_i64x2_add(wasm_i64x2_mul(vacc0x01, vmultiplier), vrounding); in xnn_qs8_igemm_minmax_ukernel_2x4c8__wasmsimd_ld128() local 148 const v128_t vq31prod0x0123 = wasm_v32x4_shuffle(vprod0x01, vprod0x23, 1, 3, 5, 7); in xnn_qs8_igemm_minmax_ukernel_2x4c8__wasmsimd_ld128()
|
D | 3x4c8-minmax-wasmsimd-ld64.c | 167 const v128_t vprod0x01 = wasm_i64x2_add(wasm_i64x2_mul(vacc0x01, vmultiplier), vrounding); in xnn_qs8_igemm_minmax_ukernel_3x4c8__wasmsimd_ld64() local 178 const v128_t vq31prod0x0123 = wasm_v32x4_shuffle(vprod0x01, vprod0x23, 1, 3, 5, 7); in xnn_qs8_igemm_minmax_ukernel_3x4c8__wasmsimd_ld64()
|
D | 3x4c8-minmax-wasmsimd-ld128.c | 171 const v128_t vprod0x01 = wasm_i64x2_add(wasm_i64x2_mul(vacc0x01, vmultiplier), vrounding); in xnn_qs8_igemm_minmax_ukernel_3x4c8__wasmsimd_ld128() local 182 const v128_t vq31prod0x0123 = wasm_v32x4_shuffle(vprod0x01, vprod0x23, 1, 3, 5, 7); in xnn_qs8_igemm_minmax_ukernel_3x4c8__wasmsimd_ld128()
|