/external/XNNPACK/src/f32-vbinary/gen/ |
D | vrdivc-minmax-avx512f-x32.c | 41 __m512 vy0123456789ABCDEF = _mm512_div_ps(vb, va0123456789ABCDEF); in xnn_f32_vrdivc_minmax_ukernel__avx512f_x32() 42 __m512 vyGHIJKLMNOPQRSTUV = _mm512_div_ps(vb, vaGHIJKLMNOPQRSTUV); in xnn_f32_vrdivc_minmax_ukernel__avx512f_x32() 59 __m512 vy = _mm512_div_ps(vb, va); in xnn_f32_vrdivc_minmax_ukernel__avx512f_x32() 74 __m512 vy = _mm512_div_ps(vb, va); in xnn_f32_vrdivc_minmax_ukernel__avx512f_x32()
|
D | vdivc-minmax-avx512f-x32.c | 41 __m512 vy0123456789ABCDEF = _mm512_div_ps(va0123456789ABCDEF, vb); in xnn_f32_vdivc_minmax_ukernel__avx512f_x32() 42 __m512 vyGHIJKLMNOPQRSTUV = _mm512_div_ps(vaGHIJKLMNOPQRSTUV, vb); in xnn_f32_vdivc_minmax_ukernel__avx512f_x32() 59 __m512 vy = _mm512_div_ps(va, vb); in xnn_f32_vdivc_minmax_ukernel__avx512f_x32() 74 __m512 vy = _mm512_div_ps(va, vb); in xnn_f32_vdivc_minmax_ukernel__avx512f_x32()
|
D | vdiv-minmax-avx512f-x32.c | 44 __m512 vy0123456789ABCDEF = _mm512_div_ps(va0123456789ABCDEF, vb0123456789ABCDEF); in xnn_f32_vdiv_minmax_ukernel__avx512f_x32() 45 __m512 vyGHIJKLMNOPQRSTUV = _mm512_div_ps(vaGHIJKLMNOPQRSTUV, vbGHIJKLMNOPQRSTUV); in xnn_f32_vdiv_minmax_ukernel__avx512f_x32() 65 __m512 vy = _mm512_div_ps(va, vb); in xnn_f32_vdiv_minmax_ukernel__avx512f_x32() 81 __m512 vy = _mm512_div_ps(va, vb); in xnn_f32_vdiv_minmax_ukernel__avx512f_x32()
|
D | vdivc-minmax-avx512f-x16.c | 40 __m512 vy0123456789ABCDEF = _mm512_div_ps(va0123456789ABCDEF, vb); in xnn_f32_vdivc_minmax_ukernel__avx512f_x16() 59 __m512 vy = _mm512_div_ps(va, vb); in xnn_f32_vdivc_minmax_ukernel__avx512f_x16()
|
/external/XNNPACK/src/f32-sigmoid/gen/ |
D | avx512f-rr1-p5-scalef-div-x128.c | 147 __m512 vf0 = _mm512_div_ps(ve0, vd0); in xnn_f32_sigmoid_ukernel__avx512f_rr1_p5_scalef_div_x128() 148 __m512 vf1 = _mm512_div_ps(ve1, vd1); in xnn_f32_sigmoid_ukernel__avx512f_rr1_p5_scalef_div_x128() 149 __m512 vf2 = _mm512_div_ps(ve2, vd2); in xnn_f32_sigmoid_ukernel__avx512f_rr1_p5_scalef_div_x128() 150 __m512 vf3 = _mm512_div_ps(ve3, vd3); in xnn_f32_sigmoid_ukernel__avx512f_rr1_p5_scalef_div_x128() 151 __m512 vf4 = _mm512_div_ps(ve4, vd4); in xnn_f32_sigmoid_ukernel__avx512f_rr1_p5_scalef_div_x128() 152 __m512 vf5 = _mm512_div_ps(ve5, vd5); in xnn_f32_sigmoid_ukernel__avx512f_rr1_p5_scalef_div_x128() 153 __m512 vf6 = _mm512_div_ps(ve6, vd6); in xnn_f32_sigmoid_ukernel__avx512f_rr1_p5_scalef_div_x128() 154 __m512 vf7 = _mm512_div_ps(ve7, vd7); in xnn_f32_sigmoid_ukernel__avx512f_rr1_p5_scalef_div_x128() 194 __m512 vf = _mm512_div_ps(ve, vd); in xnn_f32_sigmoid_ukernel__avx512f_rr1_p5_scalef_div_x128() 225 __m512 vf = _mm512_div_ps(ve, vd); in xnn_f32_sigmoid_ukernel__avx512f_rr1_p5_scalef_div_x128()
|
D | avx512f-rr1-p5-scalef-div-x112.c | 135 __m512 vf0 = _mm512_div_ps(ve0, vd0); in xnn_f32_sigmoid_ukernel__avx512f_rr1_p5_scalef_div_x112() 136 __m512 vf1 = _mm512_div_ps(ve1, vd1); in xnn_f32_sigmoid_ukernel__avx512f_rr1_p5_scalef_div_x112() 137 __m512 vf2 = _mm512_div_ps(ve2, vd2); in xnn_f32_sigmoid_ukernel__avx512f_rr1_p5_scalef_div_x112() 138 __m512 vf3 = _mm512_div_ps(ve3, vd3); in xnn_f32_sigmoid_ukernel__avx512f_rr1_p5_scalef_div_x112() 139 __m512 vf4 = _mm512_div_ps(ve4, vd4); in xnn_f32_sigmoid_ukernel__avx512f_rr1_p5_scalef_div_x112() 140 __m512 vf5 = _mm512_div_ps(ve5, vd5); in xnn_f32_sigmoid_ukernel__avx512f_rr1_p5_scalef_div_x112() 141 __m512 vf6 = _mm512_div_ps(ve6, vd6); in xnn_f32_sigmoid_ukernel__avx512f_rr1_p5_scalef_div_x112() 179 __m512 vf = _mm512_div_ps(ve, vd); in xnn_f32_sigmoid_ukernel__avx512f_rr1_p5_scalef_div_x112() 210 __m512 vf = _mm512_div_ps(ve, vd); in xnn_f32_sigmoid_ukernel__avx512f_rr1_p5_scalef_div_x112()
|
D | avx512f-rr1-lut16-p3-perm-scalef-div-x112.c | 138 __m512 vf0 = _mm512_div_ps(ve0, vd0); in xnn_f32_sigmoid_ukernel__avx512f_rr1_lut16_p3_perm_scalef_div_x112() 139 __m512 vf1 = _mm512_div_ps(ve1, vd1); in xnn_f32_sigmoid_ukernel__avx512f_rr1_lut16_p3_perm_scalef_div_x112() 140 __m512 vf2 = _mm512_div_ps(ve2, vd2); in xnn_f32_sigmoid_ukernel__avx512f_rr1_lut16_p3_perm_scalef_div_x112() 141 __m512 vf3 = _mm512_div_ps(ve3, vd3); in xnn_f32_sigmoid_ukernel__avx512f_rr1_lut16_p3_perm_scalef_div_x112() 142 __m512 vf4 = _mm512_div_ps(ve4, vd4); in xnn_f32_sigmoid_ukernel__avx512f_rr1_lut16_p3_perm_scalef_div_x112() 143 __m512 vf5 = _mm512_div_ps(ve5, vd5); in xnn_f32_sigmoid_ukernel__avx512f_rr1_lut16_p3_perm_scalef_div_x112() 144 __m512 vf6 = _mm512_div_ps(ve6, vd6); in xnn_f32_sigmoid_ukernel__avx512f_rr1_lut16_p3_perm_scalef_div_x112() 183 __m512 vf = _mm512_div_ps(ve, vd); in xnn_f32_sigmoid_ukernel__avx512f_rr1_lut16_p3_perm_scalef_div_x112() 215 __m512 vf = _mm512_div_ps(ve, vd); in xnn_f32_sigmoid_ukernel__avx512f_rr1_lut16_p3_perm_scalef_div_x112()
|
D | avx512f-rr2-lut32-p2-perm2-scalef-div-x128.c | 156 __m512 vf0 = _mm512_div_ps(ve0, vd0); in xnn_f32_sigmoid_ukernel__avx512f_rr2_lut32_p2_perm2_scalef_div_x128() 157 __m512 vf1 = _mm512_div_ps(ve1, vd1); in xnn_f32_sigmoid_ukernel__avx512f_rr2_lut32_p2_perm2_scalef_div_x128() 158 __m512 vf2 = _mm512_div_ps(ve2, vd2); in xnn_f32_sigmoid_ukernel__avx512f_rr2_lut32_p2_perm2_scalef_div_x128() 159 __m512 vf3 = _mm512_div_ps(ve3, vd3); in xnn_f32_sigmoid_ukernel__avx512f_rr2_lut32_p2_perm2_scalef_div_x128() 160 __m512 vf4 = _mm512_div_ps(ve4, vd4); in xnn_f32_sigmoid_ukernel__avx512f_rr2_lut32_p2_perm2_scalef_div_x128() 161 __m512 vf5 = _mm512_div_ps(ve5, vd5); in xnn_f32_sigmoid_ukernel__avx512f_rr2_lut32_p2_perm2_scalef_div_x128() 162 __m512 vf6 = _mm512_div_ps(ve6, vd6); in xnn_f32_sigmoid_ukernel__avx512f_rr2_lut32_p2_perm2_scalef_div_x128() 163 __m512 vf7 = _mm512_div_ps(ve7, vd7); in xnn_f32_sigmoid_ukernel__avx512f_rr2_lut32_p2_perm2_scalef_div_x128() 204 __m512 vf = _mm512_div_ps(ve, vd); in xnn_f32_sigmoid_ukernel__avx512f_rr2_lut32_p2_perm2_scalef_div_x128() 236 __m512 vf = _mm512_div_ps(ve, vd); in xnn_f32_sigmoid_ukernel__avx512f_rr2_lut32_p2_perm2_scalef_div_x128()
|
D | avx512f-rr1-lut16-p3-perm-scalef-div-x128.c | 150 __m512 vf0 = _mm512_div_ps(ve0, vd0); in xnn_f32_sigmoid_ukernel__avx512f_rr1_lut16_p3_perm_scalef_div_x128() 151 __m512 vf1 = _mm512_div_ps(ve1, vd1); in xnn_f32_sigmoid_ukernel__avx512f_rr1_lut16_p3_perm_scalef_div_x128() 152 __m512 vf2 = _mm512_div_ps(ve2, vd2); in xnn_f32_sigmoid_ukernel__avx512f_rr1_lut16_p3_perm_scalef_div_x128() 153 __m512 vf3 = _mm512_div_ps(ve3, vd3); in xnn_f32_sigmoid_ukernel__avx512f_rr1_lut16_p3_perm_scalef_div_x128() 154 __m512 vf4 = _mm512_div_ps(ve4, vd4); in xnn_f32_sigmoid_ukernel__avx512f_rr1_lut16_p3_perm_scalef_div_x128() 155 __m512 vf5 = _mm512_div_ps(ve5, vd5); in xnn_f32_sigmoid_ukernel__avx512f_rr1_lut16_p3_perm_scalef_div_x128() 156 __m512 vf6 = _mm512_div_ps(ve6, vd6); in xnn_f32_sigmoid_ukernel__avx512f_rr1_lut16_p3_perm_scalef_div_x128() 157 __m512 vf7 = _mm512_div_ps(ve7, vd7); in xnn_f32_sigmoid_ukernel__avx512f_rr1_lut16_p3_perm_scalef_div_x128() 198 __m512 vf = _mm512_div_ps(ve, vd); in xnn_f32_sigmoid_ukernel__avx512f_rr1_lut16_p3_perm_scalef_div_x128() 230 __m512 vf = _mm512_div_ps(ve, vd); in xnn_f32_sigmoid_ukernel__avx512f_rr1_lut16_p3_perm_scalef_div_x128()
|
D | avx512f-rr2-lut32-p2-perm2-scalef-div-x96.c | 132 __m512 vf0 = _mm512_div_ps(ve0, vd0); in xnn_f32_sigmoid_ukernel__avx512f_rr2_lut32_p2_perm2_scalef_div_x96() 133 __m512 vf1 = _mm512_div_ps(ve1, vd1); in xnn_f32_sigmoid_ukernel__avx512f_rr2_lut32_p2_perm2_scalef_div_x96() 134 __m512 vf2 = _mm512_div_ps(ve2, vd2); in xnn_f32_sigmoid_ukernel__avx512f_rr2_lut32_p2_perm2_scalef_div_x96() 135 __m512 vf3 = _mm512_div_ps(ve3, vd3); in xnn_f32_sigmoid_ukernel__avx512f_rr2_lut32_p2_perm2_scalef_div_x96() 136 __m512 vf4 = _mm512_div_ps(ve4, vd4); in xnn_f32_sigmoid_ukernel__avx512f_rr2_lut32_p2_perm2_scalef_div_x96() 137 __m512 vf5 = _mm512_div_ps(ve5, vd5); in xnn_f32_sigmoid_ukernel__avx512f_rr2_lut32_p2_perm2_scalef_div_x96() 174 __m512 vf = _mm512_div_ps(ve, vd); in xnn_f32_sigmoid_ukernel__avx512f_rr2_lut32_p2_perm2_scalef_div_x96() 206 __m512 vf = _mm512_div_ps(ve, vd); in xnn_f32_sigmoid_ukernel__avx512f_rr2_lut32_p2_perm2_scalef_div_x96()
|
D | avx512f-rr1-p5-scalef-div-x96.c | 123 __m512 vf0 = _mm512_div_ps(ve0, vd0); in xnn_f32_sigmoid_ukernel__avx512f_rr1_p5_scalef_div_x96() 124 __m512 vf1 = _mm512_div_ps(ve1, vd1); in xnn_f32_sigmoid_ukernel__avx512f_rr1_p5_scalef_div_x96() 125 __m512 vf2 = _mm512_div_ps(ve2, vd2); in xnn_f32_sigmoid_ukernel__avx512f_rr1_p5_scalef_div_x96() 126 __m512 vf3 = _mm512_div_ps(ve3, vd3); in xnn_f32_sigmoid_ukernel__avx512f_rr1_p5_scalef_div_x96() 127 __m512 vf4 = _mm512_div_ps(ve4, vd4); in xnn_f32_sigmoid_ukernel__avx512f_rr1_p5_scalef_div_x96() 128 __m512 vf5 = _mm512_div_ps(ve5, vd5); in xnn_f32_sigmoid_ukernel__avx512f_rr1_p5_scalef_div_x96() 164 __m512 vf = _mm512_div_ps(ve, vd); in xnn_f32_sigmoid_ukernel__avx512f_rr1_p5_scalef_div_x96() 195 __m512 vf = _mm512_div_ps(ve, vd); in xnn_f32_sigmoid_ukernel__avx512f_rr1_p5_scalef_div_x96()
|
D | avx512f-rr1-lut16-p3-perm-scalef-div-x96.c | 126 __m512 vf0 = _mm512_div_ps(ve0, vd0); in xnn_f32_sigmoid_ukernel__avx512f_rr1_lut16_p3_perm_scalef_div_x96() 127 __m512 vf1 = _mm512_div_ps(ve1, vd1); in xnn_f32_sigmoid_ukernel__avx512f_rr1_lut16_p3_perm_scalef_div_x96() 128 __m512 vf2 = _mm512_div_ps(ve2, vd2); in xnn_f32_sigmoid_ukernel__avx512f_rr1_lut16_p3_perm_scalef_div_x96() 129 __m512 vf3 = _mm512_div_ps(ve3, vd3); in xnn_f32_sigmoid_ukernel__avx512f_rr1_lut16_p3_perm_scalef_div_x96() 130 __m512 vf4 = _mm512_div_ps(ve4, vd4); in xnn_f32_sigmoid_ukernel__avx512f_rr1_lut16_p3_perm_scalef_div_x96() 131 __m512 vf5 = _mm512_div_ps(ve5, vd5); in xnn_f32_sigmoid_ukernel__avx512f_rr1_lut16_p3_perm_scalef_div_x96() 168 __m512 vf = _mm512_div_ps(ve, vd); in xnn_f32_sigmoid_ukernel__avx512f_rr1_lut16_p3_perm_scalef_div_x96() 200 __m512 vf = _mm512_div_ps(ve, vd); in xnn_f32_sigmoid_ukernel__avx512f_rr1_lut16_p3_perm_scalef_div_x96()
|
D | avx512f-rr2-lut32-p2-perm2-scalef-div-x112.c | 144 __m512 vf0 = _mm512_div_ps(ve0, vd0); in xnn_f32_sigmoid_ukernel__avx512f_rr2_lut32_p2_perm2_scalef_div_x112() 145 __m512 vf1 = _mm512_div_ps(ve1, vd1); in xnn_f32_sigmoid_ukernel__avx512f_rr2_lut32_p2_perm2_scalef_div_x112() 146 __m512 vf2 = _mm512_div_ps(ve2, vd2); in xnn_f32_sigmoid_ukernel__avx512f_rr2_lut32_p2_perm2_scalef_div_x112() 147 __m512 vf3 = _mm512_div_ps(ve3, vd3); in xnn_f32_sigmoid_ukernel__avx512f_rr2_lut32_p2_perm2_scalef_div_x112() 148 __m512 vf4 = _mm512_div_ps(ve4, vd4); in xnn_f32_sigmoid_ukernel__avx512f_rr2_lut32_p2_perm2_scalef_div_x112() 149 __m512 vf5 = _mm512_div_ps(ve5, vd5); in xnn_f32_sigmoid_ukernel__avx512f_rr2_lut32_p2_perm2_scalef_div_x112() 150 __m512 vf6 = _mm512_div_ps(ve6, vd6); in xnn_f32_sigmoid_ukernel__avx512f_rr2_lut32_p2_perm2_scalef_div_x112() 189 __m512 vf = _mm512_div_ps(ve, vd); in xnn_f32_sigmoid_ukernel__avx512f_rr2_lut32_p2_perm2_scalef_div_x112() 221 __m512 vf = _mm512_div_ps(ve, vd); in xnn_f32_sigmoid_ukernel__avx512f_rr2_lut32_p2_perm2_scalef_div_x112()
|
D | avx512f-rr2-lut32-p2-perm2-scalef-div-x80.c | 120 __m512 vf0 = _mm512_div_ps(ve0, vd0); in xnn_f32_sigmoid_ukernel__avx512f_rr2_lut32_p2_perm2_scalef_div_x80() 121 __m512 vf1 = _mm512_div_ps(ve1, vd1); in xnn_f32_sigmoid_ukernel__avx512f_rr2_lut32_p2_perm2_scalef_div_x80() 122 __m512 vf2 = _mm512_div_ps(ve2, vd2); in xnn_f32_sigmoid_ukernel__avx512f_rr2_lut32_p2_perm2_scalef_div_x80() 123 __m512 vf3 = _mm512_div_ps(ve3, vd3); in xnn_f32_sigmoid_ukernel__avx512f_rr2_lut32_p2_perm2_scalef_div_x80() 124 __m512 vf4 = _mm512_div_ps(ve4, vd4); in xnn_f32_sigmoid_ukernel__avx512f_rr2_lut32_p2_perm2_scalef_div_x80() 159 __m512 vf = _mm512_div_ps(ve, vd); in xnn_f32_sigmoid_ukernel__avx512f_rr2_lut32_p2_perm2_scalef_div_x80() 191 __m512 vf = _mm512_div_ps(ve, vd); in xnn_f32_sigmoid_ukernel__avx512f_rr2_lut32_p2_perm2_scalef_div_x80()
|
D | avx512f-rr1-p5-scalef-div-x80.c | 111 __m512 vf0 = _mm512_div_ps(ve0, vd0); in xnn_f32_sigmoid_ukernel__avx512f_rr1_p5_scalef_div_x80() 112 __m512 vf1 = _mm512_div_ps(ve1, vd1); in xnn_f32_sigmoid_ukernel__avx512f_rr1_p5_scalef_div_x80() 113 __m512 vf2 = _mm512_div_ps(ve2, vd2); in xnn_f32_sigmoid_ukernel__avx512f_rr1_p5_scalef_div_x80() 114 __m512 vf3 = _mm512_div_ps(ve3, vd3); in xnn_f32_sigmoid_ukernel__avx512f_rr1_p5_scalef_div_x80() 115 __m512 vf4 = _mm512_div_ps(ve4, vd4); in xnn_f32_sigmoid_ukernel__avx512f_rr1_p5_scalef_div_x80() 149 __m512 vf = _mm512_div_ps(ve, vd); in xnn_f32_sigmoid_ukernel__avx512f_rr1_p5_scalef_div_x80() 180 __m512 vf = _mm512_div_ps(ve, vd); in xnn_f32_sigmoid_ukernel__avx512f_rr1_p5_scalef_div_x80()
|
D | avx512f-rr1-lut16-p3-perm-scalef-div-x80.c | 114 __m512 vf0 = _mm512_div_ps(ve0, vd0); in xnn_f32_sigmoid_ukernel__avx512f_rr1_lut16_p3_perm_scalef_div_x80() 115 __m512 vf1 = _mm512_div_ps(ve1, vd1); in xnn_f32_sigmoid_ukernel__avx512f_rr1_lut16_p3_perm_scalef_div_x80() 116 __m512 vf2 = _mm512_div_ps(ve2, vd2); in xnn_f32_sigmoid_ukernel__avx512f_rr1_lut16_p3_perm_scalef_div_x80() 117 __m512 vf3 = _mm512_div_ps(ve3, vd3); in xnn_f32_sigmoid_ukernel__avx512f_rr1_lut16_p3_perm_scalef_div_x80() 118 __m512 vf4 = _mm512_div_ps(ve4, vd4); in xnn_f32_sigmoid_ukernel__avx512f_rr1_lut16_p3_perm_scalef_div_x80() 153 __m512 vf = _mm512_div_ps(ve, vd); in xnn_f32_sigmoid_ukernel__avx512f_rr1_lut16_p3_perm_scalef_div_x80() 185 __m512 vf = _mm512_div_ps(ve, vd); in xnn_f32_sigmoid_ukernel__avx512f_rr1_lut16_p3_perm_scalef_div_x80()
|
D | avx512f-rr1-p5-scalef-div-x64.c | 99 __m512 vf0 = _mm512_div_ps(ve0, vd0); in xnn_f32_sigmoid_ukernel__avx512f_rr1_p5_scalef_div_x64() 100 __m512 vf1 = _mm512_div_ps(ve1, vd1); in xnn_f32_sigmoid_ukernel__avx512f_rr1_p5_scalef_div_x64() 101 __m512 vf2 = _mm512_div_ps(ve2, vd2); in xnn_f32_sigmoid_ukernel__avx512f_rr1_p5_scalef_div_x64() 102 __m512 vf3 = _mm512_div_ps(ve3, vd3); in xnn_f32_sigmoid_ukernel__avx512f_rr1_p5_scalef_div_x64() 134 __m512 vf = _mm512_div_ps(ve, vd); in xnn_f32_sigmoid_ukernel__avx512f_rr1_p5_scalef_div_x64() 165 __m512 vf = _mm512_div_ps(ve, vd); in xnn_f32_sigmoid_ukernel__avx512f_rr1_p5_scalef_div_x64()
|
D | avx512f-rr1-lut16-p3-perm-scalef-div-x64.c | 102 __m512 vf0 = _mm512_div_ps(ve0, vd0); in xnn_f32_sigmoid_ukernel__avx512f_rr1_lut16_p3_perm_scalef_div_x64() 103 __m512 vf1 = _mm512_div_ps(ve1, vd1); in xnn_f32_sigmoid_ukernel__avx512f_rr1_lut16_p3_perm_scalef_div_x64() 104 __m512 vf2 = _mm512_div_ps(ve2, vd2); in xnn_f32_sigmoid_ukernel__avx512f_rr1_lut16_p3_perm_scalef_div_x64() 105 __m512 vf3 = _mm512_div_ps(ve3, vd3); in xnn_f32_sigmoid_ukernel__avx512f_rr1_lut16_p3_perm_scalef_div_x64() 138 __m512 vf = _mm512_div_ps(ve, vd); in xnn_f32_sigmoid_ukernel__avx512f_rr1_lut16_p3_perm_scalef_div_x64() 170 __m512 vf = _mm512_div_ps(ve, vd); in xnn_f32_sigmoid_ukernel__avx512f_rr1_lut16_p3_perm_scalef_div_x64()
|
D | avx512f-rr2-lut32-p2-perm2-scalef-div-x64.c | 108 __m512 vf0 = _mm512_div_ps(ve0, vd0); in xnn_f32_sigmoid_ukernel__avx512f_rr2_lut32_p2_perm2_scalef_div_x64() 109 __m512 vf1 = _mm512_div_ps(ve1, vd1); in xnn_f32_sigmoid_ukernel__avx512f_rr2_lut32_p2_perm2_scalef_div_x64() 110 __m512 vf2 = _mm512_div_ps(ve2, vd2); in xnn_f32_sigmoid_ukernel__avx512f_rr2_lut32_p2_perm2_scalef_div_x64() 111 __m512 vf3 = _mm512_div_ps(ve3, vd3); in xnn_f32_sigmoid_ukernel__avx512f_rr2_lut32_p2_perm2_scalef_div_x64() 144 __m512 vf = _mm512_div_ps(ve, vd); in xnn_f32_sigmoid_ukernel__avx512f_rr2_lut32_p2_perm2_scalef_div_x64() 176 __m512 vf = _mm512_div_ps(ve, vd); in xnn_f32_sigmoid_ukernel__avx512f_rr2_lut32_p2_perm2_scalef_div_x64()
|
D | avx512f-rr1-lut16-p3-perm-scalef-div-x48.c | 90 __m512 vf0 = _mm512_div_ps(ve0, vd0); in xnn_f32_sigmoid_ukernel__avx512f_rr1_lut16_p3_perm_scalef_div_x48() 91 __m512 vf1 = _mm512_div_ps(ve1, vd1); in xnn_f32_sigmoid_ukernel__avx512f_rr1_lut16_p3_perm_scalef_div_x48() 92 __m512 vf2 = _mm512_div_ps(ve2, vd2); in xnn_f32_sigmoid_ukernel__avx512f_rr1_lut16_p3_perm_scalef_div_x48() 123 __m512 vf = _mm512_div_ps(ve, vd); in xnn_f32_sigmoid_ukernel__avx512f_rr1_lut16_p3_perm_scalef_div_x48() 155 __m512 vf = _mm512_div_ps(ve, vd); in xnn_f32_sigmoid_ukernel__avx512f_rr1_lut16_p3_perm_scalef_div_x48()
|
D | avx512f-rr1-p5-scalef-div-x48.c | 87 __m512 vf0 = _mm512_div_ps(ve0, vd0); in xnn_f32_sigmoid_ukernel__avx512f_rr1_p5_scalef_div_x48() 88 __m512 vf1 = _mm512_div_ps(ve1, vd1); in xnn_f32_sigmoid_ukernel__avx512f_rr1_p5_scalef_div_x48() 89 __m512 vf2 = _mm512_div_ps(ve2, vd2); in xnn_f32_sigmoid_ukernel__avx512f_rr1_p5_scalef_div_x48() 119 __m512 vf = _mm512_div_ps(ve, vd); in xnn_f32_sigmoid_ukernel__avx512f_rr1_p5_scalef_div_x48() 150 __m512 vf = _mm512_div_ps(ve, vd); in xnn_f32_sigmoid_ukernel__avx512f_rr1_p5_scalef_div_x48()
|
D | avx512f-rr2-lut32-p2-perm2-scalef-div-x48.c | 96 __m512 vf0 = _mm512_div_ps(ve0, vd0); in xnn_f32_sigmoid_ukernel__avx512f_rr2_lut32_p2_perm2_scalef_div_x48() 97 __m512 vf1 = _mm512_div_ps(ve1, vd1); in xnn_f32_sigmoid_ukernel__avx512f_rr2_lut32_p2_perm2_scalef_div_x48() 98 __m512 vf2 = _mm512_div_ps(ve2, vd2); in xnn_f32_sigmoid_ukernel__avx512f_rr2_lut32_p2_perm2_scalef_div_x48() 129 __m512 vf = _mm512_div_ps(ve, vd); in xnn_f32_sigmoid_ukernel__avx512f_rr2_lut32_p2_perm2_scalef_div_x48() 161 __m512 vf = _mm512_div_ps(ve, vd); in xnn_f32_sigmoid_ukernel__avx512f_rr2_lut32_p2_perm2_scalef_div_x48()
|
D | avx512f-rr1-p5-scalef-div-x32.c | 75 __m512 vf0 = _mm512_div_ps(ve0, vd0); in xnn_f32_sigmoid_ukernel__avx512f_rr1_p5_scalef_div_x32() 76 __m512 vf1 = _mm512_div_ps(ve1, vd1); in xnn_f32_sigmoid_ukernel__avx512f_rr1_p5_scalef_div_x32() 104 __m512 vf = _mm512_div_ps(ve, vd); in xnn_f32_sigmoid_ukernel__avx512f_rr1_p5_scalef_div_x32() 135 __m512 vf = _mm512_div_ps(ve, vd); in xnn_f32_sigmoid_ukernel__avx512f_rr1_p5_scalef_div_x32()
|
D | avx512f-rr2-lut32-p2-perm2-scalef-div-x32.c | 84 __m512 vf0 = _mm512_div_ps(ve0, vd0); in xnn_f32_sigmoid_ukernel__avx512f_rr2_lut32_p2_perm2_scalef_div_x32() 85 __m512 vf1 = _mm512_div_ps(ve1, vd1); in xnn_f32_sigmoid_ukernel__avx512f_rr2_lut32_p2_perm2_scalef_div_x32() 114 __m512 vf = _mm512_div_ps(ve, vd); in xnn_f32_sigmoid_ukernel__avx512f_rr2_lut32_p2_perm2_scalef_div_x32() 146 __m512 vf = _mm512_div_ps(ve, vd); in xnn_f32_sigmoid_ukernel__avx512f_rr2_lut32_p2_perm2_scalef_div_x32()
|
D | avx512f-rr1-lut16-p3-perm-scalef-div-x32.c | 78 __m512 vf0 = _mm512_div_ps(ve0, vd0); in xnn_f32_sigmoid_ukernel__avx512f_rr1_lut16_p3_perm_scalef_div_x32() 79 __m512 vf1 = _mm512_div_ps(ve1, vd1); in xnn_f32_sigmoid_ukernel__avx512f_rr1_lut16_p3_perm_scalef_div_x32() 108 __m512 vf = _mm512_div_ps(ve, vd); in xnn_f32_sigmoid_ukernel__avx512f_rr1_lut16_p3_perm_scalef_div_x32() 140 __m512 vf = _mm512_div_ps(ve, vd); in xnn_f32_sigmoid_ukernel__avx512f_rr1_lut16_p3_perm_scalef_div_x32()
|