/external/XNNPACK/src/qu8-gavgpool/ |
D | 7x-minmax-neon-c8.c | 87 const int32x4_t vneg_mask_hi = vreinterpretq_s32_u32(vcltq_s32(vacc_hi, vmovq_n_s32(0))); in xnn_qu8_gavgpool_minmax_ukernel_7x__neon_c8() local 97 const int64x2_t vadjusted_product45 = vaddw_s32(vproduct45, vget_low_s32(vneg_mask_hi)); in xnn_qu8_gavgpool_minmax_ukernel_7x__neon_c8() 98 const int64x2_t vadjusted_product67 = vaddw_high_s32(vproduct67, vneg_mask_hi); in xnn_qu8_gavgpool_minmax_ukernel_7x__neon_c8() 107 const int64x2_t vadjusted_product45 = vaddw_s32(vproduct45, vget_low_s32(vneg_mask_hi)); in xnn_qu8_gavgpool_minmax_ukernel_7x__neon_c8() 108 const int64x2_t vadjusted_product67 = vaddw_s32(vproduct67, vget_high_s32(vneg_mask_hi)); in xnn_qu8_gavgpool_minmax_ukernel_7x__neon_c8() 157 const int32x4_t vneg_mask_hi = vreinterpretq_s32_u32(vcltq_s32(vacc_hi, vmovq_n_s32(0))); in xnn_qu8_gavgpool_minmax_ukernel_7x__neon_c8() local 167 const int64x2_t vadjusted_product45 = vaddw_s32(vproduct45, vget_low_s32(vneg_mask_hi)); in xnn_qu8_gavgpool_minmax_ukernel_7x__neon_c8() 168 const int64x2_t vadjusted_product67 = vaddw_high_s32(vproduct67, vneg_mask_hi); in xnn_qu8_gavgpool_minmax_ukernel_7x__neon_c8() 177 const int64x2_t vadjusted_product45 = vaddw_s32(vproduct45, vget_low_s32(vneg_mask_hi)); in xnn_qu8_gavgpool_minmax_ukernel_7x__neon_c8() 178 const int64x2_t vadjusted_product67 = vaddw_s32(vproduct67, vget_high_s32(vneg_mask_hi)); in xnn_qu8_gavgpool_minmax_ukernel_7x__neon_c8()
|
D | 7x-minmax-sse2-c8.c | 90 const __m128i vneg_mask_hi = _mm_cmpgt_epi32(_mm_setzero_si128(), vacc_hi); in xnn_qu8_gavgpool_minmax_ukernel_7x__sse2_c8() local 93 const __m128i vabs_hi0123 = _mm_sub_epi32(_mm_xor_si128(vacc_hi, vneg_mask_hi), vneg_mask_hi); in xnn_qu8_gavgpool_minmax_ukernel_7x__sse2_c8() 118 …onst __m128i vscaled_hi = _mm_sub_epi32(_mm_xor_si128(vabs_scaled_hi, vneg_mask_hi), vneg_mask_hi); in xnn_qu8_gavgpool_minmax_ukernel_7x__sse2_c8() 159 const __m128i vneg_mask_hi = _mm_cmpgt_epi32(_mm_setzero_si128(), vacc_hi); in xnn_qu8_gavgpool_minmax_ukernel_7x__sse2_c8() local 162 const __m128i vabs_hi0123 = _mm_sub_epi32(_mm_xor_si128(vacc_hi, vneg_mask_hi), vneg_mask_hi); in xnn_qu8_gavgpool_minmax_ukernel_7x__sse2_c8() 187 …onst __m128i vscaled_hi = _mm_sub_epi32(_mm_xor_si128(vabs_scaled_hi, vneg_mask_hi), vneg_mask_hi); in xnn_qu8_gavgpool_minmax_ukernel_7x__sse2_c8()
|
D | 7p7x-minmax-neon-c8.c | 163 const int32x4_t vneg_mask_hi = vreinterpretq_s32_u32(vcltq_s32(vacc_hi, vmovq_n_s32(0))); in xnn_qu8_gavgpool_minmax_ukernel_7p7x__neon_c8() local 173 const int64x2_t vadjusted_product45 = vaddw_s32(vproduct45, vget_low_s32(vneg_mask_hi)); in xnn_qu8_gavgpool_minmax_ukernel_7p7x__neon_c8() 174 const int64x2_t vadjusted_product67 = vaddw_high_s32(vproduct67, vneg_mask_hi); in xnn_qu8_gavgpool_minmax_ukernel_7p7x__neon_c8() 183 const int64x2_t vadjusted_product45 = vaddw_s32(vproduct45, vget_low_s32(vneg_mask_hi)); in xnn_qu8_gavgpool_minmax_ukernel_7p7x__neon_c8() 184 const int64x2_t vadjusted_product67 = vaddw_s32(vproduct67, vget_high_s32(vneg_mask_hi)); in xnn_qu8_gavgpool_minmax_ukernel_7p7x__neon_c8() 235 const int32x4_t vneg_mask_hi = vreinterpretq_s32_u32(vcltq_s32(vacc_hi, vmovq_n_s32(0))); in xnn_qu8_gavgpool_minmax_ukernel_7p7x__neon_c8() local 245 const int64x2_t vadjusted_product45 = vaddw_s32(vproduct45, vget_low_s32(vneg_mask_hi)); in xnn_qu8_gavgpool_minmax_ukernel_7p7x__neon_c8() 246 const int64x2_t vadjusted_product67 = vaddw_high_s32(vproduct67, vneg_mask_hi); in xnn_qu8_gavgpool_minmax_ukernel_7p7x__neon_c8() 255 const int64x2_t vadjusted_product45 = vaddw_s32(vproduct45, vget_low_s32(vneg_mask_hi)); in xnn_qu8_gavgpool_minmax_ukernel_7p7x__neon_c8() 256 const int64x2_t vadjusted_product67 = vaddw_s32(vproduct67, vget_high_s32(vneg_mask_hi)); in xnn_qu8_gavgpool_minmax_ukernel_7p7x__neon_c8()
|
D | 7p7x-minmax-sse2-c8.c | 184 const __m128i vneg_mask_hi = _mm_cmpgt_epi32(_mm_setzero_si128(), vacc_hi); in xnn_qu8_gavgpool_minmax_ukernel_7p7x__sse2_c8() local 187 const __m128i vabs_hi0123 = _mm_sub_epi32(_mm_xor_si128(vacc_hi, vneg_mask_hi), vneg_mask_hi); in xnn_qu8_gavgpool_minmax_ukernel_7p7x__sse2_c8() 212 …onst __m128i vscaled_hi = _mm_sub_epi32(_mm_xor_si128(vabs_scaled_hi, vneg_mask_hi), vneg_mask_hi); in xnn_qu8_gavgpool_minmax_ukernel_7p7x__sse2_c8() 255 const __m128i vneg_mask_hi = _mm_cmpgt_epi32(_mm_setzero_si128(), vacc_hi); in xnn_qu8_gavgpool_minmax_ukernel_7p7x__sse2_c8() local 258 const __m128i vabs_hi0123 = _mm_sub_epi32(_mm_xor_si128(vacc_hi, vneg_mask_hi), vneg_mask_hi); in xnn_qu8_gavgpool_minmax_ukernel_7p7x__sse2_c8() 283 …onst __m128i vscaled_hi = _mm_sub_epi32(_mm_xor_si128(vabs_scaled_hi, vneg_mask_hi), vneg_mask_hi); in xnn_qu8_gavgpool_minmax_ukernel_7p7x__sse2_c8()
|
/external/XNNPACK/src/qu8-avgpool/ |
D | 9x-minmax-neon-c8.c | 142 const int32x4_t vneg_mask_hi = vreinterpretq_s32_u32(vcltq_s32(vacc_hi, vmovq_n_s32(0))); in xnn_qu8_avgpool_minmax_ukernel_9x__neon_c8() local 152 const int64x2_t vadjusted_product45 = vaddw_s32(vproduct45, vget_low_s32(vneg_mask_hi)); in xnn_qu8_avgpool_minmax_ukernel_9x__neon_c8() 153 const int64x2_t vadjusted_product67 = vaddw_high_s32(vproduct67, vneg_mask_hi); in xnn_qu8_avgpool_minmax_ukernel_9x__neon_c8() 162 const int64x2_t vadjusted_product45 = vaddw_s32(vproduct45, vget_low_s32(vneg_mask_hi)); in xnn_qu8_avgpool_minmax_ukernel_9x__neon_c8() 163 const int64x2_t vadjusted_product67 = vaddw_s32(vproduct67, vget_high_s32(vneg_mask_hi)); in xnn_qu8_avgpool_minmax_ukernel_9x__neon_c8() 215 const int32x4_t vneg_mask_hi = vreinterpretq_s32_u32(vcltq_s32(vacc_hi, vmovq_n_s32(0))); in xnn_qu8_avgpool_minmax_ukernel_9x__neon_c8() local 225 const int64x2_t vadjusted_product45 = vaddw_s32(vproduct45, vget_low_s32(vneg_mask_hi)); in xnn_qu8_avgpool_minmax_ukernel_9x__neon_c8() 226 const int64x2_t vadjusted_product67 = vaddw_high_s32(vproduct67, vneg_mask_hi); in xnn_qu8_avgpool_minmax_ukernel_9x__neon_c8() 235 const int64x2_t vadjusted_product45 = vaddw_s32(vproduct45, vget_low_s32(vneg_mask_hi)); in xnn_qu8_avgpool_minmax_ukernel_9x__neon_c8() 236 const int64x2_t vadjusted_product67 = vaddw_s32(vproduct67, vget_high_s32(vneg_mask_hi)); in xnn_qu8_avgpool_minmax_ukernel_9x__neon_c8()
|
D | 9x-minmax-sse2-c8.c | 146 const __m128i vneg_mask_hi = _mm_cmpgt_epi32(_mm_setzero_si128(), vacc_hi); in xnn_qu8_avgpool_minmax_ukernel_9x__sse2_c8() local 149 const __m128i vabs_hi0123 = _mm_sub_epi32(_mm_xor_si128(vacc_hi, vneg_mask_hi), vneg_mask_hi); in xnn_qu8_avgpool_minmax_ukernel_9x__sse2_c8() 174 …onst __m128i vscaled_hi = _mm_sub_epi32(_mm_xor_si128(vabs_scaled_hi, vneg_mask_hi), vneg_mask_hi); in xnn_qu8_avgpool_minmax_ukernel_9x__sse2_c8() 221 const __m128i vneg_mask_hi = _mm_cmpgt_epi32(_mm_setzero_si128(), vacc_hi); in xnn_qu8_avgpool_minmax_ukernel_9x__sse2_c8() local 224 const __m128i vabs_hi0123 = _mm_sub_epi32(_mm_xor_si128(vacc_hi, vneg_mask_hi), vneg_mask_hi); in xnn_qu8_avgpool_minmax_ukernel_9x__sse2_c8() 249 …onst __m128i vscaled_hi = _mm_sub_epi32(_mm_xor_si128(vabs_scaled_hi, vneg_mask_hi), vneg_mask_hi); in xnn_qu8_avgpool_minmax_ukernel_9x__sse2_c8()
|
D | 9p8x-minmax-neon-c8.c | 286 const int32x4_t vneg_mask_hi = vreinterpretq_s32_u32(vcltq_s32(vacc_hi, vmovq_n_s32(0))); in xnn_qu8_avgpool_minmax_ukernel_9p8x__neon_c8() local 296 const int64x2_t vadjusted_product45 = vaddw_s32(vproduct45, vget_low_s32(vneg_mask_hi)); in xnn_qu8_avgpool_minmax_ukernel_9p8x__neon_c8() 297 const int64x2_t vadjusted_product67 = vaddw_high_s32(vproduct67, vneg_mask_hi); in xnn_qu8_avgpool_minmax_ukernel_9p8x__neon_c8() 306 const int64x2_t vadjusted_product45 = vaddw_s32(vproduct45, vget_low_s32(vneg_mask_hi)); in xnn_qu8_avgpool_minmax_ukernel_9p8x__neon_c8() 307 const int64x2_t vadjusted_product67 = vaddw_s32(vproduct67, vget_high_s32(vneg_mask_hi)); in xnn_qu8_avgpool_minmax_ukernel_9p8x__neon_c8() 360 const int32x4_t vneg_mask_hi = vreinterpretq_s32_u32(vcltq_s32(vacc_hi, vmovq_n_s32(0))); in xnn_qu8_avgpool_minmax_ukernel_9p8x__neon_c8() local 370 const int64x2_t vadjusted_product45 = vaddw_s32(vproduct45, vget_low_s32(vneg_mask_hi)); in xnn_qu8_avgpool_minmax_ukernel_9p8x__neon_c8() 371 const int64x2_t vadjusted_product67 = vaddw_high_s32(vproduct67, vneg_mask_hi); in xnn_qu8_avgpool_minmax_ukernel_9p8x__neon_c8() 380 const int64x2_t vadjusted_product45 = vaddw_s32(vproduct45, vget_low_s32(vneg_mask_hi)); in xnn_qu8_avgpool_minmax_ukernel_9p8x__neon_c8() 381 const int64x2_t vadjusted_product67 = vaddw_s32(vproduct67, vget_high_s32(vneg_mask_hi)); in xnn_qu8_avgpool_minmax_ukernel_9p8x__neon_c8()
|
D | 9p8x-minmax-sse2-c8.c | 311 const __m128i vneg_mask_hi = _mm_cmpgt_epi32(_mm_setzero_si128(), vacc_hi); in xnn_qu8_avgpool_minmax_ukernel_9p8x__sse2_c8() local 314 … const __m128i vabs_hi0123 = _mm_sub_epi32(_mm_xor_si128(vacc_hi, vneg_mask_hi), vneg_mask_hi); in xnn_qu8_avgpool_minmax_ukernel_9p8x__sse2_c8() 339 …onst __m128i vscaled_hi = _mm_sub_epi32(_mm_xor_si128(vabs_scaled_hi, vneg_mask_hi), vneg_mask_hi); in xnn_qu8_avgpool_minmax_ukernel_9p8x__sse2_c8() 386 const __m128i vneg_mask_hi = _mm_cmpgt_epi32(_mm_setzero_si128(), vacc_hi); in xnn_qu8_avgpool_minmax_ukernel_9p8x__sse2_c8() local 389 … const __m128i vabs_hi0123 = _mm_sub_epi32(_mm_xor_si128(vacc_hi, vneg_mask_hi), vneg_mask_hi); in xnn_qu8_avgpool_minmax_ukernel_9p8x__sse2_c8() 414 …onst __m128i vscaled_hi = _mm_sub_epi32(_mm_xor_si128(vabs_scaled_hi, vneg_mask_hi), vneg_mask_hi); in xnn_qu8_avgpool_minmax_ukernel_9p8x__sse2_c8()
|