Home
last modified time | relevance | path

Searched refs:vneg_mask_hi (Results 1 – 8 of 8) sorted by relevance

/external/XNNPACK/src/qu8-gavgpool/
D7x-minmax-neon-c8.c87 const int32x4_t vneg_mask_hi = vreinterpretq_s32_u32(vcltq_s32(vacc_hi, vmovq_n_s32(0))); in xnn_qu8_gavgpool_minmax_ukernel_7x__neon_c8() local
97 const int64x2_t vadjusted_product45 = vaddw_s32(vproduct45, vget_low_s32(vneg_mask_hi)); in xnn_qu8_gavgpool_minmax_ukernel_7x__neon_c8()
98 const int64x2_t vadjusted_product67 = vaddw_high_s32(vproduct67, vneg_mask_hi); in xnn_qu8_gavgpool_minmax_ukernel_7x__neon_c8()
107 const int64x2_t vadjusted_product45 = vaddw_s32(vproduct45, vget_low_s32(vneg_mask_hi)); in xnn_qu8_gavgpool_minmax_ukernel_7x__neon_c8()
108 const int64x2_t vadjusted_product67 = vaddw_s32(vproduct67, vget_high_s32(vneg_mask_hi)); in xnn_qu8_gavgpool_minmax_ukernel_7x__neon_c8()
157 const int32x4_t vneg_mask_hi = vreinterpretq_s32_u32(vcltq_s32(vacc_hi, vmovq_n_s32(0))); in xnn_qu8_gavgpool_minmax_ukernel_7x__neon_c8() local
167 const int64x2_t vadjusted_product45 = vaddw_s32(vproduct45, vget_low_s32(vneg_mask_hi)); in xnn_qu8_gavgpool_minmax_ukernel_7x__neon_c8()
168 const int64x2_t vadjusted_product67 = vaddw_high_s32(vproduct67, vneg_mask_hi); in xnn_qu8_gavgpool_minmax_ukernel_7x__neon_c8()
177 const int64x2_t vadjusted_product45 = vaddw_s32(vproduct45, vget_low_s32(vneg_mask_hi)); in xnn_qu8_gavgpool_minmax_ukernel_7x__neon_c8()
178 const int64x2_t vadjusted_product67 = vaddw_s32(vproduct67, vget_high_s32(vneg_mask_hi)); in xnn_qu8_gavgpool_minmax_ukernel_7x__neon_c8()
D7x-minmax-sse2-c8.c90 const __m128i vneg_mask_hi = _mm_cmpgt_epi32(_mm_setzero_si128(), vacc_hi); in xnn_qu8_gavgpool_minmax_ukernel_7x__sse2_c8() local
93 const __m128i vabs_hi0123 = _mm_sub_epi32(_mm_xor_si128(vacc_hi, vneg_mask_hi), vneg_mask_hi); in xnn_qu8_gavgpool_minmax_ukernel_7x__sse2_c8()
118 …onst __m128i vscaled_hi = _mm_sub_epi32(_mm_xor_si128(vabs_scaled_hi, vneg_mask_hi), vneg_mask_hi); in xnn_qu8_gavgpool_minmax_ukernel_7x__sse2_c8()
159 const __m128i vneg_mask_hi = _mm_cmpgt_epi32(_mm_setzero_si128(), vacc_hi); in xnn_qu8_gavgpool_minmax_ukernel_7x__sse2_c8() local
162 const __m128i vabs_hi0123 = _mm_sub_epi32(_mm_xor_si128(vacc_hi, vneg_mask_hi), vneg_mask_hi); in xnn_qu8_gavgpool_minmax_ukernel_7x__sse2_c8()
187 …onst __m128i vscaled_hi = _mm_sub_epi32(_mm_xor_si128(vabs_scaled_hi, vneg_mask_hi), vneg_mask_hi); in xnn_qu8_gavgpool_minmax_ukernel_7x__sse2_c8()
D7p7x-minmax-neon-c8.c163 const int32x4_t vneg_mask_hi = vreinterpretq_s32_u32(vcltq_s32(vacc_hi, vmovq_n_s32(0))); in xnn_qu8_gavgpool_minmax_ukernel_7p7x__neon_c8() local
173 const int64x2_t vadjusted_product45 = vaddw_s32(vproduct45, vget_low_s32(vneg_mask_hi)); in xnn_qu8_gavgpool_minmax_ukernel_7p7x__neon_c8()
174 const int64x2_t vadjusted_product67 = vaddw_high_s32(vproduct67, vneg_mask_hi); in xnn_qu8_gavgpool_minmax_ukernel_7p7x__neon_c8()
183 const int64x2_t vadjusted_product45 = vaddw_s32(vproduct45, vget_low_s32(vneg_mask_hi)); in xnn_qu8_gavgpool_minmax_ukernel_7p7x__neon_c8()
184 const int64x2_t vadjusted_product67 = vaddw_s32(vproduct67, vget_high_s32(vneg_mask_hi)); in xnn_qu8_gavgpool_minmax_ukernel_7p7x__neon_c8()
235 const int32x4_t vneg_mask_hi = vreinterpretq_s32_u32(vcltq_s32(vacc_hi, vmovq_n_s32(0))); in xnn_qu8_gavgpool_minmax_ukernel_7p7x__neon_c8() local
245 const int64x2_t vadjusted_product45 = vaddw_s32(vproduct45, vget_low_s32(vneg_mask_hi)); in xnn_qu8_gavgpool_minmax_ukernel_7p7x__neon_c8()
246 const int64x2_t vadjusted_product67 = vaddw_high_s32(vproduct67, vneg_mask_hi); in xnn_qu8_gavgpool_minmax_ukernel_7p7x__neon_c8()
255 const int64x2_t vadjusted_product45 = vaddw_s32(vproduct45, vget_low_s32(vneg_mask_hi)); in xnn_qu8_gavgpool_minmax_ukernel_7p7x__neon_c8()
256 const int64x2_t vadjusted_product67 = vaddw_s32(vproduct67, vget_high_s32(vneg_mask_hi)); in xnn_qu8_gavgpool_minmax_ukernel_7p7x__neon_c8()
D7p7x-minmax-sse2-c8.c184 const __m128i vneg_mask_hi = _mm_cmpgt_epi32(_mm_setzero_si128(), vacc_hi); in xnn_qu8_gavgpool_minmax_ukernel_7p7x__sse2_c8() local
187 const __m128i vabs_hi0123 = _mm_sub_epi32(_mm_xor_si128(vacc_hi, vneg_mask_hi), vneg_mask_hi); in xnn_qu8_gavgpool_minmax_ukernel_7p7x__sse2_c8()
212 …onst __m128i vscaled_hi = _mm_sub_epi32(_mm_xor_si128(vabs_scaled_hi, vneg_mask_hi), vneg_mask_hi); in xnn_qu8_gavgpool_minmax_ukernel_7p7x__sse2_c8()
255 const __m128i vneg_mask_hi = _mm_cmpgt_epi32(_mm_setzero_si128(), vacc_hi); in xnn_qu8_gavgpool_minmax_ukernel_7p7x__sse2_c8() local
258 const __m128i vabs_hi0123 = _mm_sub_epi32(_mm_xor_si128(vacc_hi, vneg_mask_hi), vneg_mask_hi); in xnn_qu8_gavgpool_minmax_ukernel_7p7x__sse2_c8()
283 …onst __m128i vscaled_hi = _mm_sub_epi32(_mm_xor_si128(vabs_scaled_hi, vneg_mask_hi), vneg_mask_hi); in xnn_qu8_gavgpool_minmax_ukernel_7p7x__sse2_c8()
/external/XNNPACK/src/qu8-avgpool/
D9x-minmax-neon-c8.c142 const int32x4_t vneg_mask_hi = vreinterpretq_s32_u32(vcltq_s32(vacc_hi, vmovq_n_s32(0))); in xnn_qu8_avgpool_minmax_ukernel_9x__neon_c8() local
152 const int64x2_t vadjusted_product45 = vaddw_s32(vproduct45, vget_low_s32(vneg_mask_hi)); in xnn_qu8_avgpool_minmax_ukernel_9x__neon_c8()
153 const int64x2_t vadjusted_product67 = vaddw_high_s32(vproduct67, vneg_mask_hi); in xnn_qu8_avgpool_minmax_ukernel_9x__neon_c8()
162 const int64x2_t vadjusted_product45 = vaddw_s32(vproduct45, vget_low_s32(vneg_mask_hi)); in xnn_qu8_avgpool_minmax_ukernel_9x__neon_c8()
163 const int64x2_t vadjusted_product67 = vaddw_s32(vproduct67, vget_high_s32(vneg_mask_hi)); in xnn_qu8_avgpool_minmax_ukernel_9x__neon_c8()
215 const int32x4_t vneg_mask_hi = vreinterpretq_s32_u32(vcltq_s32(vacc_hi, vmovq_n_s32(0))); in xnn_qu8_avgpool_minmax_ukernel_9x__neon_c8() local
225 const int64x2_t vadjusted_product45 = vaddw_s32(vproduct45, vget_low_s32(vneg_mask_hi)); in xnn_qu8_avgpool_minmax_ukernel_9x__neon_c8()
226 const int64x2_t vadjusted_product67 = vaddw_high_s32(vproduct67, vneg_mask_hi); in xnn_qu8_avgpool_minmax_ukernel_9x__neon_c8()
235 const int64x2_t vadjusted_product45 = vaddw_s32(vproduct45, vget_low_s32(vneg_mask_hi)); in xnn_qu8_avgpool_minmax_ukernel_9x__neon_c8()
236 const int64x2_t vadjusted_product67 = vaddw_s32(vproduct67, vget_high_s32(vneg_mask_hi)); in xnn_qu8_avgpool_minmax_ukernel_9x__neon_c8()
D9x-minmax-sse2-c8.c146 const __m128i vneg_mask_hi = _mm_cmpgt_epi32(_mm_setzero_si128(), vacc_hi); in xnn_qu8_avgpool_minmax_ukernel_9x__sse2_c8() local
149 const __m128i vabs_hi0123 = _mm_sub_epi32(_mm_xor_si128(vacc_hi, vneg_mask_hi), vneg_mask_hi); in xnn_qu8_avgpool_minmax_ukernel_9x__sse2_c8()
174 …onst __m128i vscaled_hi = _mm_sub_epi32(_mm_xor_si128(vabs_scaled_hi, vneg_mask_hi), vneg_mask_hi); in xnn_qu8_avgpool_minmax_ukernel_9x__sse2_c8()
221 const __m128i vneg_mask_hi = _mm_cmpgt_epi32(_mm_setzero_si128(), vacc_hi); in xnn_qu8_avgpool_minmax_ukernel_9x__sse2_c8() local
224 const __m128i vabs_hi0123 = _mm_sub_epi32(_mm_xor_si128(vacc_hi, vneg_mask_hi), vneg_mask_hi); in xnn_qu8_avgpool_minmax_ukernel_9x__sse2_c8()
249 …onst __m128i vscaled_hi = _mm_sub_epi32(_mm_xor_si128(vabs_scaled_hi, vneg_mask_hi), vneg_mask_hi); in xnn_qu8_avgpool_minmax_ukernel_9x__sse2_c8()
D9p8x-minmax-neon-c8.c286 const int32x4_t vneg_mask_hi = vreinterpretq_s32_u32(vcltq_s32(vacc_hi, vmovq_n_s32(0))); in xnn_qu8_avgpool_minmax_ukernel_9p8x__neon_c8() local
296 const int64x2_t vadjusted_product45 = vaddw_s32(vproduct45, vget_low_s32(vneg_mask_hi)); in xnn_qu8_avgpool_minmax_ukernel_9p8x__neon_c8()
297 const int64x2_t vadjusted_product67 = vaddw_high_s32(vproduct67, vneg_mask_hi); in xnn_qu8_avgpool_minmax_ukernel_9p8x__neon_c8()
306 const int64x2_t vadjusted_product45 = vaddw_s32(vproduct45, vget_low_s32(vneg_mask_hi)); in xnn_qu8_avgpool_minmax_ukernel_9p8x__neon_c8()
307 const int64x2_t vadjusted_product67 = vaddw_s32(vproduct67, vget_high_s32(vneg_mask_hi)); in xnn_qu8_avgpool_minmax_ukernel_9p8x__neon_c8()
360 const int32x4_t vneg_mask_hi = vreinterpretq_s32_u32(vcltq_s32(vacc_hi, vmovq_n_s32(0))); in xnn_qu8_avgpool_minmax_ukernel_9p8x__neon_c8() local
370 const int64x2_t vadjusted_product45 = vaddw_s32(vproduct45, vget_low_s32(vneg_mask_hi)); in xnn_qu8_avgpool_minmax_ukernel_9p8x__neon_c8()
371 const int64x2_t vadjusted_product67 = vaddw_high_s32(vproduct67, vneg_mask_hi); in xnn_qu8_avgpool_minmax_ukernel_9p8x__neon_c8()
380 const int64x2_t vadjusted_product45 = vaddw_s32(vproduct45, vget_low_s32(vneg_mask_hi)); in xnn_qu8_avgpool_minmax_ukernel_9p8x__neon_c8()
381 const int64x2_t vadjusted_product67 = vaddw_s32(vproduct67, vget_high_s32(vneg_mask_hi)); in xnn_qu8_avgpool_minmax_ukernel_9p8x__neon_c8()
D9p8x-minmax-sse2-c8.c311 const __m128i vneg_mask_hi = _mm_cmpgt_epi32(_mm_setzero_si128(), vacc_hi); in xnn_qu8_avgpool_minmax_ukernel_9p8x__sse2_c8() local
314 … const __m128i vabs_hi0123 = _mm_sub_epi32(_mm_xor_si128(vacc_hi, vneg_mask_hi), vneg_mask_hi); in xnn_qu8_avgpool_minmax_ukernel_9p8x__sse2_c8()
339 …onst __m128i vscaled_hi = _mm_sub_epi32(_mm_xor_si128(vabs_scaled_hi, vneg_mask_hi), vneg_mask_hi); in xnn_qu8_avgpool_minmax_ukernel_9p8x__sse2_c8()
386 const __m128i vneg_mask_hi = _mm_cmpgt_epi32(_mm_setzero_si128(), vacc_hi); in xnn_qu8_avgpool_minmax_ukernel_9p8x__sse2_c8() local
389 … const __m128i vabs_hi0123 = _mm_sub_epi32(_mm_xor_si128(vacc_hi, vneg_mask_hi), vneg_mask_hi); in xnn_qu8_avgpool_minmax_ukernel_9p8x__sse2_c8()
414 …onst __m128i vscaled_hi = _mm_sub_epi32(_mm_xor_si128(vabs_scaled_hi, vneg_mask_hi), vneg_mask_hi); in xnn_qu8_avgpool_minmax_ukernel_9p8x__sse2_c8()