Home
last modified time | relevance | path

Searched refs:vb8x0 (Results 1 – 8 of 8) sorted by relevance

/external/XNNPACK/src/qs8-gemm/gen/
D3x16c8-minmax-neon-mlal-padal.c123 const int8x8_t vb8x0 = vld1_s8(w); w = (const void*) ((uintptr_t) w + 8 * sizeof( int8_t)); in xnn_qs8_gemm_minmax_ukernel_3x16c8__neon_mlal_padal() local
213 int16x8_t vprod0x8 = vmull_s8(vb8x0, va0x0); in xnn_qs8_gemm_minmax_ukernel_3x16c8__neon_mlal_padal()
214 int16x8_t vprod1x8 = vmull_s8(vb8x0, va1x0); in xnn_qs8_gemm_minmax_ukernel_3x16c8__neon_mlal_padal()
215 int16x8_t vprod2x8 = vmull_s8(vb8x0, va2x0); in xnn_qs8_gemm_minmax_ukernel_3x16c8__neon_mlal_padal()
D2x16c8-minmax-neon-mlal-padal.c99 const int8x8_t vb8x0 = vld1_s8(w); w = (const void*) ((uintptr_t) w + 8 * sizeof( int8_t)); in xnn_qs8_gemm_minmax_ukernel_2x16c8__neon_mlal_padal() local
165 int16x8_t vprod0x8 = vmull_s8(vb8x0, va0x0); in xnn_qs8_gemm_minmax_ukernel_2x16c8__neon_mlal_padal()
166 int16x8_t vprod1x8 = vmull_s8(vb8x0, va1x0); in xnn_qs8_gemm_minmax_ukernel_2x16c8__neon_mlal_padal()
D1x16c8-minmax-neon-mlal-padal.c75 const int8x8_t vb8x0 = vld1_s8(w); w = (const void*) ((uintptr_t) w + 8 * sizeof( int8_t)); in xnn_qs8_gemm_minmax_ukernel_1x16c8__neon_mlal_padal() local
117 int16x8_t vprod0x8 = vmull_s8(vb8x0, va0x0); in xnn_qs8_gemm_minmax_ukernel_1x16c8__neon_mlal_padal()
D4x16c8-minmax-neon-mlal-padal.c147 const int8x8_t vb8x0 = vld1_s8(w); w = (const void*) ((uintptr_t) w + 8 * sizeof( int8_t)); in xnn_qs8_gemm_minmax_ukernel_4x16c8__neon_mlal_padal() local
261 int16x8_t vprod0x8 = vmull_s8(vb8x0, va0x0); in xnn_qs8_gemm_minmax_ukernel_4x16c8__neon_mlal_padal()
262 int16x8_t vprod1x8 = vmull_s8(vb8x0, va1x0); in xnn_qs8_gemm_minmax_ukernel_4x16c8__neon_mlal_padal()
263 int16x8_t vprod2x8 = vmull_s8(vb8x0, va2x0); in xnn_qs8_gemm_minmax_ukernel_4x16c8__neon_mlal_padal()
264 int16x8_t vprod3x8 = vmull_s8(vb8x0, va3x0); in xnn_qs8_gemm_minmax_ukernel_4x16c8__neon_mlal_padal()
/external/XNNPACK/src/qs8-igemm/gen/
D2x16c8-minmax-neon-mlal-padal.c112 const int8x8_t vb8x0 = vld1_s8(w); w = (const void*) ((uintptr_t) w + 8 * sizeof( int8_t)); in xnn_qs8_igemm_minmax_ukernel_2x16c8__neon_mlal_padal() local
178 int16x8_t vprod0x8 = vmull_s8(vb8x0, va0x0); in xnn_qs8_igemm_minmax_ukernel_2x16c8__neon_mlal_padal()
179 int16x8_t vprod1x8 = vmull_s8(vb8x0, va1x0); in xnn_qs8_igemm_minmax_ukernel_2x16c8__neon_mlal_padal()
D1x16c8-minmax-neon-mlal-padal.c86 const int8x8_t vb8x0 = vld1_s8(w); w = (const void*) ((uintptr_t) w + 8 * sizeof( int8_t)); in xnn_qs8_igemm_minmax_ukernel_1x16c8__neon_mlal_padal() local
128 int16x8_t vprod0x8 = vmull_s8(vb8x0, va0x0); in xnn_qs8_igemm_minmax_ukernel_1x16c8__neon_mlal_padal()
D4x16c8-minmax-neon-mlal-padal.c164 const int8x8_t vb8x0 = vld1_s8(w); w = (const void*) ((uintptr_t) w + 8 * sizeof( int8_t)); in xnn_qs8_igemm_minmax_ukernel_4x16c8__neon_mlal_padal() local
278 int16x8_t vprod0x8 = vmull_s8(vb8x0, va0x0); in xnn_qs8_igemm_minmax_ukernel_4x16c8__neon_mlal_padal()
279 int16x8_t vprod1x8 = vmull_s8(vb8x0, va1x0); in xnn_qs8_igemm_minmax_ukernel_4x16c8__neon_mlal_padal()
280 int16x8_t vprod2x8 = vmull_s8(vb8x0, va2x0); in xnn_qs8_igemm_minmax_ukernel_4x16c8__neon_mlal_padal()
281 int16x8_t vprod3x8 = vmull_s8(vb8x0, va3x0); in xnn_qs8_igemm_minmax_ukernel_4x16c8__neon_mlal_padal()
D3x16c8-minmax-neon-mlal-padal.c138 const int8x8_t vb8x0 = vld1_s8(w); w = (const void*) ((uintptr_t) w + 8 * sizeof( int8_t)); in xnn_qs8_igemm_minmax_ukernel_3x16c8__neon_mlal_padal() local
228 int16x8_t vprod0x8 = vmull_s8(vb8x0, va0x0); in xnn_qs8_igemm_minmax_ukernel_3x16c8__neon_mlal_padal()
229 int16x8_t vprod1x8 = vmull_s8(vb8x0, va1x0); in xnn_qs8_igemm_minmax_ukernel_3x16c8__neon_mlal_padal()
230 int16x8_t vprod2x8 = vmull_s8(vb8x0, va2x0); in xnn_qs8_igemm_minmax_ukernel_3x16c8__neon_mlal_padal()