Lines Matching refs:int64x2_t

62   const int64x2_t vleft_shift = vld1q_dup_s64(&params->neon.left_shift);  in xnn_qs8_gavgpool_minmax_ukernel_7x__neon_c32_acc2()
143 const int64x2_t vprod01 = vmull_s32(vget_low_s32(vacc0123), vget_low_s32(vmultiplier)); in xnn_qs8_gavgpool_minmax_ukernel_7x__neon_c32_acc2()
144 const int64x2_t vprod23 = vmull_high_s32(vacc0123, vmultiplier); in xnn_qs8_gavgpool_minmax_ukernel_7x__neon_c32_acc2()
145 const int64x2_t vprod45 = vmull_s32(vget_low_s32(vacc4567), vget_low_s32(vmultiplier)); in xnn_qs8_gavgpool_minmax_ukernel_7x__neon_c32_acc2()
146 const int64x2_t vprod67 = vmull_high_s32(vacc4567, vmultiplier); in xnn_qs8_gavgpool_minmax_ukernel_7x__neon_c32_acc2()
147 const int64x2_t vprod89 = vmull_s32(vget_low_s32(vacc89AB), vget_low_s32(vmultiplier)); in xnn_qs8_gavgpool_minmax_ukernel_7x__neon_c32_acc2()
148 const int64x2_t vprodAB = vmull_high_s32(vacc89AB, vmultiplier); in xnn_qs8_gavgpool_minmax_ukernel_7x__neon_c32_acc2()
149 const int64x2_t vprodCD = vmull_s32(vget_low_s32(vaccCDEF), vget_low_s32(vmultiplier)); in xnn_qs8_gavgpool_minmax_ukernel_7x__neon_c32_acc2()
150 const int64x2_t vprodEF = vmull_high_s32(vaccCDEF, vmultiplier); in xnn_qs8_gavgpool_minmax_ukernel_7x__neon_c32_acc2()
151 const int64x2_t vprodGH = vmull_s32(vget_low_s32(vaccGHIJ), vget_low_s32(vmultiplier)); in xnn_qs8_gavgpool_minmax_ukernel_7x__neon_c32_acc2()
152 const int64x2_t vprodIJ = vmull_high_s32(vaccGHIJ, vmultiplier); in xnn_qs8_gavgpool_minmax_ukernel_7x__neon_c32_acc2()
153 const int64x2_t vprodKL = vmull_s32(vget_low_s32(vaccKLMN), vget_low_s32(vmultiplier)); in xnn_qs8_gavgpool_minmax_ukernel_7x__neon_c32_acc2()
154 const int64x2_t vprodMN = vmull_high_s32(vaccKLMN, vmultiplier); in xnn_qs8_gavgpool_minmax_ukernel_7x__neon_c32_acc2()
155 const int64x2_t vprodOP = vmull_s32(vget_low_s32(vaccOPQR), vget_low_s32(vmultiplier)); in xnn_qs8_gavgpool_minmax_ukernel_7x__neon_c32_acc2()
156 const int64x2_t vprodQR = vmull_high_s32(vaccOPQR, vmultiplier); in xnn_qs8_gavgpool_minmax_ukernel_7x__neon_c32_acc2()
157 const int64x2_t vprodST = vmull_s32(vget_low_s32(vaccSTUV), vget_low_s32(vmultiplier)); in xnn_qs8_gavgpool_minmax_ukernel_7x__neon_c32_acc2()
158 const int64x2_t vprodUV = vmull_high_s32(vaccSTUV, vmultiplier); in xnn_qs8_gavgpool_minmax_ukernel_7x__neon_c32_acc2()
160 const int64x2_t vadjprod01 = vaddw_s32(vprod01, vget_low_s32(vsgnacc0123)); in xnn_qs8_gavgpool_minmax_ukernel_7x__neon_c32_acc2()
161 const int64x2_t vadjprod23 = vaddw_high_s32(vprod23, vsgnacc0123); in xnn_qs8_gavgpool_minmax_ukernel_7x__neon_c32_acc2()
162 const int64x2_t vadjprod45 = vaddw_s32(vprod45, vget_low_s32(vsgnacc4567)); in xnn_qs8_gavgpool_minmax_ukernel_7x__neon_c32_acc2()
163 const int64x2_t vadjprod67 = vaddw_high_s32(vprod67, vsgnacc4567); in xnn_qs8_gavgpool_minmax_ukernel_7x__neon_c32_acc2()
164 const int64x2_t vadjprod89 = vaddw_s32(vprod89, vget_low_s32(vsgnacc89AB)); in xnn_qs8_gavgpool_minmax_ukernel_7x__neon_c32_acc2()
165 const int64x2_t vadjprodAB = vaddw_high_s32(vprodAB, vsgnacc89AB); in xnn_qs8_gavgpool_minmax_ukernel_7x__neon_c32_acc2()
166 const int64x2_t vadjprodCD = vaddw_s32(vprodCD, vget_low_s32(vsgnaccCDEF)); in xnn_qs8_gavgpool_minmax_ukernel_7x__neon_c32_acc2()
167 const int64x2_t vadjprodEF = vaddw_high_s32(vprodEF, vsgnaccCDEF); in xnn_qs8_gavgpool_minmax_ukernel_7x__neon_c32_acc2()
168 const int64x2_t vadjprodGH = vaddw_s32(vprodGH, vget_low_s32(vsgnaccGHIJ)); in xnn_qs8_gavgpool_minmax_ukernel_7x__neon_c32_acc2()
169 const int64x2_t vadjprodIJ = vaddw_high_s32(vprodIJ, vsgnaccGHIJ); in xnn_qs8_gavgpool_minmax_ukernel_7x__neon_c32_acc2()
170 const int64x2_t vadjprodKL = vaddw_s32(vprodKL, vget_low_s32(vsgnaccKLMN)); in xnn_qs8_gavgpool_minmax_ukernel_7x__neon_c32_acc2()
171 const int64x2_t vadjprodMN = vaddw_high_s32(vprodMN, vsgnaccKLMN); in xnn_qs8_gavgpool_minmax_ukernel_7x__neon_c32_acc2()
172 const int64x2_t vadjprodOP = vaddw_s32(vprodOP, vget_low_s32(vsgnaccOPQR)); in xnn_qs8_gavgpool_minmax_ukernel_7x__neon_c32_acc2()
173 const int64x2_t vadjprodQR = vaddw_high_s32(vprodQR, vsgnaccOPQR); in xnn_qs8_gavgpool_minmax_ukernel_7x__neon_c32_acc2()
174 const int64x2_t vadjprodST = vaddw_s32(vprodST, vget_low_s32(vsgnaccSTUV)); in xnn_qs8_gavgpool_minmax_ukernel_7x__neon_c32_acc2()
175 const int64x2_t vadjprodUV = vaddw_high_s32(vprodUV, vsgnaccSTUV); in xnn_qs8_gavgpool_minmax_ukernel_7x__neon_c32_acc2()
177 const int64x2_t vprod01 = vmull_s32(vget_low_s32(vacc0123), vmultiplier); in xnn_qs8_gavgpool_minmax_ukernel_7x__neon_c32_acc2()
178 const int64x2_t vprod23 = vmull_s32(vget_high_s32(vacc0123), vmultiplier); in xnn_qs8_gavgpool_minmax_ukernel_7x__neon_c32_acc2()
179 const int64x2_t vprod45 = vmull_s32(vget_low_s32(vacc4567), vmultiplier); in xnn_qs8_gavgpool_minmax_ukernel_7x__neon_c32_acc2()
180 const int64x2_t vprod67 = vmull_s32(vget_high_s32(vacc4567), vmultiplier); in xnn_qs8_gavgpool_minmax_ukernel_7x__neon_c32_acc2()
181 const int64x2_t vprod89 = vmull_s32(vget_low_s32(vacc89AB), vmultiplier); in xnn_qs8_gavgpool_minmax_ukernel_7x__neon_c32_acc2()
182 const int64x2_t vprodAB = vmull_s32(vget_high_s32(vacc89AB), vmultiplier); in xnn_qs8_gavgpool_minmax_ukernel_7x__neon_c32_acc2()
183 const int64x2_t vprodCD = vmull_s32(vget_low_s32(vaccCDEF), vmultiplier); in xnn_qs8_gavgpool_minmax_ukernel_7x__neon_c32_acc2()
184 const int64x2_t vprodEF = vmull_s32(vget_high_s32(vaccCDEF), vmultiplier); in xnn_qs8_gavgpool_minmax_ukernel_7x__neon_c32_acc2()
185 const int64x2_t vprodGH = vmull_s32(vget_low_s32(vaccGHIJ), vmultiplier); in xnn_qs8_gavgpool_minmax_ukernel_7x__neon_c32_acc2()
186 const int64x2_t vprodIJ = vmull_s32(vget_high_s32(vaccGHIJ), vmultiplier); in xnn_qs8_gavgpool_minmax_ukernel_7x__neon_c32_acc2()
187 const int64x2_t vprodKL = vmull_s32(vget_low_s32(vaccKLMN), vmultiplier); in xnn_qs8_gavgpool_minmax_ukernel_7x__neon_c32_acc2()
188 const int64x2_t vprodMN = vmull_s32(vget_high_s32(vaccKLMN), vmultiplier); in xnn_qs8_gavgpool_minmax_ukernel_7x__neon_c32_acc2()
189 const int64x2_t vprodOP = vmull_s32(vget_low_s32(vaccOPQR), vmultiplier); in xnn_qs8_gavgpool_minmax_ukernel_7x__neon_c32_acc2()
190 const int64x2_t vprodQR = vmull_s32(vget_high_s32(vaccOPQR), vmultiplier); in xnn_qs8_gavgpool_minmax_ukernel_7x__neon_c32_acc2()
191 const int64x2_t vprodST = vmull_s32(vget_low_s32(vaccSTUV), vmultiplier); in xnn_qs8_gavgpool_minmax_ukernel_7x__neon_c32_acc2()
192 const int64x2_t vprodUV = vmull_s32(vget_high_s32(vaccSTUV), vmultiplier); in xnn_qs8_gavgpool_minmax_ukernel_7x__neon_c32_acc2()
194 const int64x2_t vadjprod01 = vaddw_s32(vprod01, vget_low_s32(vsgnacc0123)); in xnn_qs8_gavgpool_minmax_ukernel_7x__neon_c32_acc2()
195 const int64x2_t vadjprod23 = vaddw_s32(vprod23, vget_high_s32(vsgnacc0123)); in xnn_qs8_gavgpool_minmax_ukernel_7x__neon_c32_acc2()
196 const int64x2_t vadjprod45 = vaddw_s32(vprod45, vget_low_s32(vsgnacc4567)); in xnn_qs8_gavgpool_minmax_ukernel_7x__neon_c32_acc2()
197 const int64x2_t vadjprod67 = vaddw_s32(vprod67, vget_high_s32(vsgnacc4567)); in xnn_qs8_gavgpool_minmax_ukernel_7x__neon_c32_acc2()
198 const int64x2_t vadjprod89 = vaddw_s32(vprod89, vget_low_s32(vsgnacc89AB)); in xnn_qs8_gavgpool_minmax_ukernel_7x__neon_c32_acc2()
199 const int64x2_t vadjprodAB = vaddw_s32(vprodAB, vget_high_s32(vsgnacc89AB)); in xnn_qs8_gavgpool_minmax_ukernel_7x__neon_c32_acc2()
200 const int64x2_t vadjprodCD = vaddw_s32(vprodCD, vget_low_s32(vsgnaccCDEF)); in xnn_qs8_gavgpool_minmax_ukernel_7x__neon_c32_acc2()
201 const int64x2_t vadjprodEF = vaddw_s32(vprodEF, vget_high_s32(vsgnaccCDEF)); in xnn_qs8_gavgpool_minmax_ukernel_7x__neon_c32_acc2()
202 const int64x2_t vadjprodGH = vaddw_s32(vprodGH, vget_low_s32(vsgnaccGHIJ)); in xnn_qs8_gavgpool_minmax_ukernel_7x__neon_c32_acc2()
203 const int64x2_t vadjprodIJ = vaddw_s32(vprodIJ, vget_high_s32(vsgnaccGHIJ)); in xnn_qs8_gavgpool_minmax_ukernel_7x__neon_c32_acc2()
204 const int64x2_t vadjprodKL = vaddw_s32(vprodKL, vget_low_s32(vsgnaccKLMN)); in xnn_qs8_gavgpool_minmax_ukernel_7x__neon_c32_acc2()
205 const int64x2_t vadjprodMN = vaddw_s32(vprodMN, vget_high_s32(vsgnaccKLMN)); in xnn_qs8_gavgpool_minmax_ukernel_7x__neon_c32_acc2()
206 const int64x2_t vadjprodOP = vaddw_s32(vprodOP, vget_low_s32(vsgnaccOPQR)); in xnn_qs8_gavgpool_minmax_ukernel_7x__neon_c32_acc2()
207 const int64x2_t vadjprodQR = vaddw_s32(vprodQR, vget_high_s32(vsgnaccOPQR)); in xnn_qs8_gavgpool_minmax_ukernel_7x__neon_c32_acc2()
208 const int64x2_t vadjprodST = vaddw_s32(vprodST, vget_low_s32(vsgnaccSTUV)); in xnn_qs8_gavgpool_minmax_ukernel_7x__neon_c32_acc2()
209 const int64x2_t vadjprodUV = vaddw_s32(vprodUV, vget_high_s32(vsgnaccSTUV)); in xnn_qs8_gavgpool_minmax_ukernel_7x__neon_c32_acc2()
212 const int64x2_t vacc01 = vrshlq_s64(vadjprod01, vleft_shift); in xnn_qs8_gavgpool_minmax_ukernel_7x__neon_c32_acc2()
213 const int64x2_t vacc23 = vrshlq_s64(vadjprod23, vleft_shift); in xnn_qs8_gavgpool_minmax_ukernel_7x__neon_c32_acc2()
214 const int64x2_t vacc45 = vrshlq_s64(vadjprod45, vleft_shift); in xnn_qs8_gavgpool_minmax_ukernel_7x__neon_c32_acc2()
215 const int64x2_t vacc67 = vrshlq_s64(vadjprod67, vleft_shift); in xnn_qs8_gavgpool_minmax_ukernel_7x__neon_c32_acc2()
216 const int64x2_t vacc89 = vrshlq_s64(vadjprod89, vleft_shift); in xnn_qs8_gavgpool_minmax_ukernel_7x__neon_c32_acc2()
217 const int64x2_t vaccAB = vrshlq_s64(vadjprodAB, vleft_shift); in xnn_qs8_gavgpool_minmax_ukernel_7x__neon_c32_acc2()
218 const int64x2_t vaccCD = vrshlq_s64(vadjprodCD, vleft_shift); in xnn_qs8_gavgpool_minmax_ukernel_7x__neon_c32_acc2()
219 const int64x2_t vaccEF = vrshlq_s64(vadjprodEF, vleft_shift); in xnn_qs8_gavgpool_minmax_ukernel_7x__neon_c32_acc2()
220 const int64x2_t vaccGH = vrshlq_s64(vadjprodGH, vleft_shift); in xnn_qs8_gavgpool_minmax_ukernel_7x__neon_c32_acc2()
221 const int64x2_t vaccIJ = vrshlq_s64(vadjprodIJ, vleft_shift); in xnn_qs8_gavgpool_minmax_ukernel_7x__neon_c32_acc2()
222 const int64x2_t vaccKL = vrshlq_s64(vadjprodKL, vleft_shift); in xnn_qs8_gavgpool_minmax_ukernel_7x__neon_c32_acc2()
223 const int64x2_t vaccMN = vrshlq_s64(vadjprodMN, vleft_shift); in xnn_qs8_gavgpool_minmax_ukernel_7x__neon_c32_acc2()
224 const int64x2_t vaccOP = vrshlq_s64(vadjprodOP, vleft_shift); in xnn_qs8_gavgpool_minmax_ukernel_7x__neon_c32_acc2()
225 const int64x2_t vaccQR = vrshlq_s64(vadjprodQR, vleft_shift); in xnn_qs8_gavgpool_minmax_ukernel_7x__neon_c32_acc2()
226 const int64x2_t vaccST = vrshlq_s64(vadjprodST, vleft_shift); in xnn_qs8_gavgpool_minmax_ukernel_7x__neon_c32_acc2()
227 const int64x2_t vaccUV = vrshlq_s64(vadjprodUV, vleft_shift); in xnn_qs8_gavgpool_minmax_ukernel_7x__neon_c32_acc2()
303 const int64x2_t vprod01 = vmull_s32(vget_low_s32(vacc0123), vget_low_s32(vmultiplier)); in xnn_qs8_gavgpool_minmax_ukernel_7x__neon_c32_acc2()
304 const int64x2_t vprod23 = vmull_high_s32(vacc0123, vmultiplier); in xnn_qs8_gavgpool_minmax_ukernel_7x__neon_c32_acc2()
305 const int64x2_t vprod45 = vmull_s32(vget_low_s32(vacc4567), vget_low_s32(vmultiplier)); in xnn_qs8_gavgpool_minmax_ukernel_7x__neon_c32_acc2()
306 const int64x2_t vprod67 = vmull_high_s32(vacc4567, vmultiplier); in xnn_qs8_gavgpool_minmax_ukernel_7x__neon_c32_acc2()
308 const int64x2_t vadjprod01 = vaddw_s32(vprod01, vget_low_s32(vsgnacc0123)); in xnn_qs8_gavgpool_minmax_ukernel_7x__neon_c32_acc2()
309 const int64x2_t vadjprod23 = vaddw_high_s32(vprod23, vsgnacc0123); in xnn_qs8_gavgpool_minmax_ukernel_7x__neon_c32_acc2()
310 const int64x2_t vadjprod45 = vaddw_s32(vprod45, vget_low_s32(vsgnacc4567)); in xnn_qs8_gavgpool_minmax_ukernel_7x__neon_c32_acc2()
311 const int64x2_t vadjprod67 = vaddw_high_s32(vprod67, vsgnacc4567); in xnn_qs8_gavgpool_minmax_ukernel_7x__neon_c32_acc2()
313 const int64x2_t vprod01 = vmull_s32(vget_low_s32(vacc0123), vmultiplier); in xnn_qs8_gavgpool_minmax_ukernel_7x__neon_c32_acc2()
314 const int64x2_t vprod23 = vmull_s32(vget_high_s32(vacc0123), vmultiplier); in xnn_qs8_gavgpool_minmax_ukernel_7x__neon_c32_acc2()
315 const int64x2_t vprod45 = vmull_s32(vget_low_s32(vacc4567), vmultiplier); in xnn_qs8_gavgpool_minmax_ukernel_7x__neon_c32_acc2()
316 const int64x2_t vprod67 = vmull_s32(vget_high_s32(vacc4567), vmultiplier); in xnn_qs8_gavgpool_minmax_ukernel_7x__neon_c32_acc2()
318 const int64x2_t vadjprod01 = vaddw_s32(vprod01, vget_low_s32(vsgnacc0123)); in xnn_qs8_gavgpool_minmax_ukernel_7x__neon_c32_acc2()
319 const int64x2_t vadjprod23 = vaddw_s32(vprod23, vget_high_s32(vsgnacc0123)); in xnn_qs8_gavgpool_minmax_ukernel_7x__neon_c32_acc2()
320 const int64x2_t vadjprod45 = vaddw_s32(vprod45, vget_low_s32(vsgnacc4567)); in xnn_qs8_gavgpool_minmax_ukernel_7x__neon_c32_acc2()
321 const int64x2_t vadjprod67 = vaddw_s32(vprod67, vget_high_s32(vsgnacc4567)); in xnn_qs8_gavgpool_minmax_ukernel_7x__neon_c32_acc2()
324 const int64x2_t vacc01 = vrshlq_s64(vadjprod01, vleft_shift); in xnn_qs8_gavgpool_minmax_ukernel_7x__neon_c32_acc2()
325 const int64x2_t vacc23 = vrshlq_s64(vadjprod23, vleft_shift); in xnn_qs8_gavgpool_minmax_ukernel_7x__neon_c32_acc2()
326 const int64x2_t vacc45 = vrshlq_s64(vadjprod45, vleft_shift); in xnn_qs8_gavgpool_minmax_ukernel_7x__neon_c32_acc2()
327 const int64x2_t vacc67 = vrshlq_s64(vadjprod67, vleft_shift); in xnn_qs8_gavgpool_minmax_ukernel_7x__neon_c32_acc2()