/external/XNNPACK/src/qs8-igemm/gen/ |
D | 3x8c8-minmax-avx2.c | 175 const __m256i vq31prod0x01234567 = _mm256_blend_epi16(vq31prod0x0246, vq31prod0x1357, 0xCC); in xnn_qs8_igemm_minmax_ukernel_3x8c8__avx2() 176 const __m256i vq31prod1x01234567 = _mm256_blend_epi16(vq31prod1x0246, vq31prod1x1357, 0xCC); in xnn_qs8_igemm_minmax_ukernel_3x8c8__avx2() 177 const __m256i vq31prod2x01234567 = _mm256_blend_epi16(vq31prod2x0246, vq31prod2x1357, 0xCC); in xnn_qs8_igemm_minmax_ukernel_3x8c8__avx2()
|
D | 2x8c8-minmax-avx2.c | 147 const __m256i vq31prod0x01234567 = _mm256_blend_epi16(vq31prod0x0246, vq31prod0x1357, 0xCC); in xnn_qs8_igemm_minmax_ukernel_2x8c8__avx2() 148 const __m256i vq31prod1x01234567 = _mm256_blend_epi16(vq31prod1x0246, vq31prod1x1357, 0xCC); in xnn_qs8_igemm_minmax_ukernel_2x8c8__avx2()
|
D | 1x8c8-minmax-avx2.c | 119 const __m256i vq31prod0x01234567 = _mm256_blend_epi16(vq31prod0x0246, vq31prod0x1357, 0xCC); in xnn_qs8_igemm_minmax_ukernel_1x8c8__avx2()
|
/external/XNNPACK/src/qs8-gemm/gen/ |
D | 3x8c8-minmax-avx2.c | 158 const __m256i vq31prod0x01234567 = _mm256_blend_epi16(vq31prod0x0246, vq31prod0x1357, 0xCC); in xnn_qs8_gemm_minmax_ukernel_3x8c8__avx2() 159 const __m256i vq31prod1x01234567 = _mm256_blend_epi16(vq31prod1x0246, vq31prod1x1357, 0xCC); in xnn_qs8_gemm_minmax_ukernel_3x8c8__avx2() 160 const __m256i vq31prod2x01234567 = _mm256_blend_epi16(vq31prod2x0246, vq31prod2x1357, 0xCC); in xnn_qs8_gemm_minmax_ukernel_3x8c8__avx2()
|
D | 3x8c8-xw-minmax-avx2.c | 154 const __m256i vq31prod0x01234567 = _mm256_blend_epi16(vq31prod0x0246, vq31prod0x1357, 0xCC); in xnn_qs8_gemm_xw_minmax_ukernel_3x8c8__avx2() 155 const __m256i vq31prod1x01234567 = _mm256_blend_epi16(vq31prod1x0246, vq31prod1x1357, 0xCC); in xnn_qs8_gemm_xw_minmax_ukernel_3x8c8__avx2() 156 const __m256i vq31prod2x01234567 = _mm256_blend_epi16(vq31prod2x0246, vq31prod2x1357, 0xCC); in xnn_qs8_gemm_xw_minmax_ukernel_3x8c8__avx2()
|
D | 2x8c8-minmax-avx2.c | 132 const __m256i vq31prod0x01234567 = _mm256_blend_epi16(vq31prod0x0246, vq31prod0x1357, 0xCC); in xnn_qs8_gemm_minmax_ukernel_2x8c8__avx2() 133 const __m256i vq31prod1x01234567 = _mm256_blend_epi16(vq31prod1x0246, vq31prod1x1357, 0xCC); in xnn_qs8_gemm_minmax_ukernel_2x8c8__avx2()
|
D | 2x8c8-xw-minmax-avx2.c | 128 const __m256i vq31prod0x01234567 = _mm256_blend_epi16(vq31prod0x0246, vq31prod0x1357, 0xCC); in xnn_qs8_gemm_xw_minmax_ukernel_2x8c8__avx2() 129 const __m256i vq31prod1x01234567 = _mm256_blend_epi16(vq31prod1x0246, vq31prod1x1357, 0xCC); in xnn_qs8_gemm_xw_minmax_ukernel_2x8c8__avx2()
|
D | 1x8c8-minmax-avx2.c | 106 const __m256i vq31prod0x01234567 = _mm256_blend_epi16(vq31prod0x0246, vq31prod0x1357, 0xCC); in xnn_qs8_gemm_minmax_ukernel_1x8c8__avx2()
|
D | 1x8c8-xw-minmax-avx2.c | 102 const __m256i vq31prod0x01234567 = _mm256_blend_epi16(vq31prod0x0246, vq31prod0x1357, 0xCC); in xnn_qs8_gemm_xw_minmax_ukernel_1x8c8__avx2()
|
/external/XNNPACK/src/qs8-dwconv/gen/ |
D | up32x9-minmax-avx2-mul16.c | 252 const __m256i vq31prod01234567 = _mm256_blend_epi16(vq31prod0246, vq31prod1357, 0xCC); in xnn_qs8_dwconv_minmax_ukernel_up32x9__avx2_mul16() 253 const __m256i vq31prod89ABCDEF = _mm256_blend_epi16(vq31prod8ACE, vq31prod9BDF, 0xCC); in xnn_qs8_dwconv_minmax_ukernel_up32x9__avx2_mul16() 254 const __m256i vq31prodGHIJKLMN = _mm256_blend_epi16(vq31prodGIKM, vq31prodHJLN, 0xCC); in xnn_qs8_dwconv_minmax_ukernel_up32x9__avx2_mul16() 255 const __m256i vq31prodOPQRSTUV = _mm256_blend_epi16(vq31prodOQSU, vq31prodPRTV, 0xCC); in xnn_qs8_dwconv_minmax_ukernel_up32x9__avx2_mul16() 401 const __m256i vq31prod01234567 = _mm256_blend_epi16(vq31prod0246, vq31prod1357, 0xCC); in xnn_qs8_dwconv_minmax_ukernel_up32x9__avx2_mul16() 402 const __m256i vq31prod89ABCDEF = _mm256_blend_epi16(vq31prod8ACE, vq31prod9BDF, 0xCC); in xnn_qs8_dwconv_minmax_ukernel_up32x9__avx2_mul16()
|
D | up16x9-minmax-avx2-mul16.c | 186 const __m256i vq31prod01234567 = _mm256_blend_epi16(vq31prod0246, vq31prod1357, 0xCC); in xnn_qs8_dwconv_minmax_ukernel_up16x9__avx2_mul16() 187 const __m256i vq31prod89ABCDEF = _mm256_blend_epi16(vq31prod8ACE, vq31prod9BDF, 0xCC); in xnn_qs8_dwconv_minmax_ukernel_up16x9__avx2_mul16() 309 const __m256i vq31prod01234567 = _mm256_blend_epi16(vq31prod0246, vq31prod1357, 0xCC); in xnn_qs8_dwconv_minmax_ukernel_up16x9__avx2_mul16() 310 const __m256i vq31prod89ABCDEF = _mm256_blend_epi16(vq31prod8ACE, vq31prod9BDF, 0xCC); in xnn_qs8_dwconv_minmax_ukernel_up16x9__avx2_mul16()
|
D | up32x9-minmax-avx2-mul32.c | 252 const __m256i vq31prod01234567 = _mm256_blend_epi16(vq31prod0246, vq31prod1357, 0xCC); in xnn_qs8_dwconv_minmax_ukernel_up32x9__avx2_mul32() 253 const __m256i vq31prod89ABCDEF = _mm256_blend_epi16(vq31prod8ACE, vq31prod9BDF, 0xCC); in xnn_qs8_dwconv_minmax_ukernel_up32x9__avx2_mul32() 254 const __m256i vq31prodGHIJKLMN = _mm256_blend_epi16(vq31prodGIKM, vq31prodHJLN, 0xCC); in xnn_qs8_dwconv_minmax_ukernel_up32x9__avx2_mul32() 255 const __m256i vq31prodOPQRSTUV = _mm256_blend_epi16(vq31prodOQSU, vq31prodPRTV, 0xCC); in xnn_qs8_dwconv_minmax_ukernel_up32x9__avx2_mul32() 368 const __m256i vq31prod01234567 = _mm256_blend_epi16(vq31prod0246, vq31prod1357, 0xCC); in xnn_qs8_dwconv_minmax_ukernel_up32x9__avx2_mul32()
|
D | up24x9-minmax-avx2-mul32.c | 219 const __m256i vq31prod01234567 = _mm256_blend_epi16(vq31prod0246, vq31prod1357, 0xCC); in xnn_qs8_dwconv_minmax_ukernel_up24x9__avx2_mul32() 220 const __m256i vq31prod89ABCDEF = _mm256_blend_epi16(vq31prod8ACE, vq31prod9BDF, 0xCC); in xnn_qs8_dwconv_minmax_ukernel_up24x9__avx2_mul32() 221 const __m256i vq31prodGHIJKLMN = _mm256_blend_epi16(vq31prodGIKM, vq31prodHJLN, 0xCC); in xnn_qs8_dwconv_minmax_ukernel_up24x9__avx2_mul32() 330 const __m256i vq31prod01234567 = _mm256_blend_epi16(vq31prod0246, vq31prod1357, 0xCC); in xnn_qs8_dwconv_minmax_ukernel_up24x9__avx2_mul32()
|
D | up16x9-minmax-avx2-mul32.c | 186 const __m256i vq31prod01234567 = _mm256_blend_epi16(vq31prod0246, vq31prod1357, 0xCC); in xnn_qs8_dwconv_minmax_ukernel_up16x9__avx2_mul32() 187 const __m256i vq31prod89ABCDEF = _mm256_blend_epi16(vq31prod8ACE, vq31prod9BDF, 0xCC); in xnn_qs8_dwconv_minmax_ukernel_up16x9__avx2_mul32() 288 const __m256i vq31prod01234567 = _mm256_blend_epi16(vq31prod0246, vq31prod1357, 0xCC); in xnn_qs8_dwconv_minmax_ukernel_up16x9__avx2_mul32()
|
D | up8x9-minmax-avx2-mul32.c | 153 const __m256i vq31prod01234567 = _mm256_blend_epi16(vq31prod0246, vq31prod1357, 0xCC); in xnn_qs8_dwconv_minmax_ukernel_up8x9__avx2_mul32() 238 const __m256i vq31prod01234567 = _mm256_blend_epi16(vq31prod0246, vq31prod1357, 0xCC); in xnn_qs8_dwconv_minmax_ukernel_up8x9__avx2_mul32()
|
/external/XNNPACK/src/qs8-dwconv/ |
D | unipass-avx2-mul16.c.in | 81 …const __m256i vq31prod${ABC[C:C+8]} = _mm256_blend_epi16(vq31prod${ABC[C:C+8:2]}, vq31prod${ABC[C+… 156 …const __m256i vq31prod${ABC[0:8]} = _mm256_blend_epi16(vq31prod${ABC[0:8:2]}, vq31prod${ABC[1:8:2]… 157 …const __m256i vq31prod${ABC[8:16]} = _mm256_blend_epi16(vq31prod${ABC[8:16:2]}, vq31prod${ABC[9:16…
|
D | unipass-avx2-mul32.c.in | 78 …const __m256i vq31prod${ABC[C:C+8]} = _mm256_blend_epi16(vq31prod${ABC[C:C+8:2]}, vq31prod${ABC[C+… 170 …const __m256i vq31prod${ABC[0:8]} = _mm256_blend_epi16(vq31prod${ABC[0:8:2]}, vq31prod${ABC[1:8:2]…
|
/external/gemmlowp/fixedpoint/ |
D | fixedpoint_avx.h | 347 result = _mm256_blend_epi16(_mm256_srli_si256(a0b0_a2b2_rounded_2x, 4),
|
/external/XNNPACK/src/qs8-igemm/ |
D | MRx8c8-avx2.c.in | 131 …const __m256i vq31prod${M}x01234567 = _mm256_blend_epi16(vq31prod${M}x0246, vq31prod${M}x1357, 0xC…
|
/external/XNNPACK/src/qs8-gemm/ |
D | MRx8c8-avx2.c.in | 132 …const __m256i vq31prod${M}x01234567 = _mm256_blend_epi16(vq31prod${M}x0246, vq31prod${M}x1357, 0xC…
|
/external/clang/test/CodeGen/ |
D | avx2-builtins.c | 121 return _mm256_blend_epi16(a, b, 2); in test_mm256_blend_epi16()
|
/external/llvm-project/clang/test/CodeGen/X86/ |
D | avx2-builtins.c | 121 return _mm256_blend_epi16(a, b, 2); in test_mm256_blend_epi16()
|
/external/llvm-project/clang/lib/Headers/ |
D | avx2intrin.h | 151 #define _mm256_blend_epi16(V1, V2, M) \ macro
|
/external/clang/lib/Headers/ |
D | avx2intrin.h | 164 #define _mm256_blend_epi16(V1, V2, M) __extension__ ({ \ macro
|