Home
last modified time | relevance | path

Searched refs:_mm512_fmadd_ps (Results 1 – 25 of 222) sorted by relevance

123456789

/external/XNNPACK/src/f32-raddstoreexpminusmax/gen/
Davx512f-p5-scalef-x192.c87 __m512 vt0 = _mm512_fmadd_ps(vn0, vminus_ln2_hi, vx0); in xnn_f32_raddstoreexpminusmax_ukernel__avx512f_p5_scalef_x192()
88 __m512 vt1 = _mm512_fmadd_ps(vn1, vminus_ln2_hi, vx1); in xnn_f32_raddstoreexpminusmax_ukernel__avx512f_p5_scalef_x192()
89 __m512 vt2 = _mm512_fmadd_ps(vn2, vminus_ln2_hi, vx2); in xnn_f32_raddstoreexpminusmax_ukernel__avx512f_p5_scalef_x192()
90 __m512 vt3 = _mm512_fmadd_ps(vn3, vminus_ln2_hi, vx3); in xnn_f32_raddstoreexpminusmax_ukernel__avx512f_p5_scalef_x192()
91 __m512 vt4 = _mm512_fmadd_ps(vn4, vminus_ln2_hi, vx4); in xnn_f32_raddstoreexpminusmax_ukernel__avx512f_p5_scalef_x192()
92 __m512 vt5 = _mm512_fmadd_ps(vn5, vminus_ln2_hi, vx5); in xnn_f32_raddstoreexpminusmax_ukernel__avx512f_p5_scalef_x192()
93 __m512 vt6 = _mm512_fmadd_ps(vn6, vminus_ln2_hi, vx6); in xnn_f32_raddstoreexpminusmax_ukernel__avx512f_p5_scalef_x192()
94 __m512 vt7 = _mm512_fmadd_ps(vn7, vminus_ln2_hi, vx7); in xnn_f32_raddstoreexpminusmax_ukernel__avx512f_p5_scalef_x192()
95 __m512 vt8 = _mm512_fmadd_ps(vn8, vminus_ln2_hi, vx8); in xnn_f32_raddstoreexpminusmax_ukernel__avx512f_p5_scalef_x192()
96 __m512 vt9 = _mm512_fmadd_ps(vn9, vminus_ln2_hi, vx9); in xnn_f32_raddstoreexpminusmax_ukernel__avx512f_p5_scalef_x192()
[all …]
Davx512f-p5-scalef-x192-acc2.c88 __m512 vt0 = _mm512_fmadd_ps(vn0, vminus_ln2_hi, vx0); in xnn_f32_raddstoreexpminusmax_ukernel__avx512f_p5_scalef_x192_acc2()
89 __m512 vt1 = _mm512_fmadd_ps(vn1, vminus_ln2_hi, vx1); in xnn_f32_raddstoreexpminusmax_ukernel__avx512f_p5_scalef_x192_acc2()
90 __m512 vt2 = _mm512_fmadd_ps(vn2, vminus_ln2_hi, vx2); in xnn_f32_raddstoreexpminusmax_ukernel__avx512f_p5_scalef_x192_acc2()
91 __m512 vt3 = _mm512_fmadd_ps(vn3, vminus_ln2_hi, vx3); in xnn_f32_raddstoreexpminusmax_ukernel__avx512f_p5_scalef_x192_acc2()
92 __m512 vt4 = _mm512_fmadd_ps(vn4, vminus_ln2_hi, vx4); in xnn_f32_raddstoreexpminusmax_ukernel__avx512f_p5_scalef_x192_acc2()
93 __m512 vt5 = _mm512_fmadd_ps(vn5, vminus_ln2_hi, vx5); in xnn_f32_raddstoreexpminusmax_ukernel__avx512f_p5_scalef_x192_acc2()
94 __m512 vt6 = _mm512_fmadd_ps(vn6, vminus_ln2_hi, vx6); in xnn_f32_raddstoreexpminusmax_ukernel__avx512f_p5_scalef_x192_acc2()
95 __m512 vt7 = _mm512_fmadd_ps(vn7, vminus_ln2_hi, vx7); in xnn_f32_raddstoreexpminusmax_ukernel__avx512f_p5_scalef_x192_acc2()
96 __m512 vt8 = _mm512_fmadd_ps(vn8, vminus_ln2_hi, vx8); in xnn_f32_raddstoreexpminusmax_ukernel__avx512f_p5_scalef_x192_acc2()
97 __m512 vt9 = _mm512_fmadd_ps(vn9, vminus_ln2_hi, vx9); in xnn_f32_raddstoreexpminusmax_ukernel__avx512f_p5_scalef_x192_acc2()
[all …]
Davx512f-p5-scalef-x192-acc3.c89 __m512 vt0 = _mm512_fmadd_ps(vn0, vminus_ln2_hi, vx0); in xnn_f32_raddstoreexpminusmax_ukernel__avx512f_p5_scalef_x192_acc3()
90 __m512 vt1 = _mm512_fmadd_ps(vn1, vminus_ln2_hi, vx1); in xnn_f32_raddstoreexpminusmax_ukernel__avx512f_p5_scalef_x192_acc3()
91 __m512 vt2 = _mm512_fmadd_ps(vn2, vminus_ln2_hi, vx2); in xnn_f32_raddstoreexpminusmax_ukernel__avx512f_p5_scalef_x192_acc3()
92 __m512 vt3 = _mm512_fmadd_ps(vn3, vminus_ln2_hi, vx3); in xnn_f32_raddstoreexpminusmax_ukernel__avx512f_p5_scalef_x192_acc3()
93 __m512 vt4 = _mm512_fmadd_ps(vn4, vminus_ln2_hi, vx4); in xnn_f32_raddstoreexpminusmax_ukernel__avx512f_p5_scalef_x192_acc3()
94 __m512 vt5 = _mm512_fmadd_ps(vn5, vminus_ln2_hi, vx5); in xnn_f32_raddstoreexpminusmax_ukernel__avx512f_p5_scalef_x192_acc3()
95 __m512 vt6 = _mm512_fmadd_ps(vn6, vminus_ln2_hi, vx6); in xnn_f32_raddstoreexpminusmax_ukernel__avx512f_p5_scalef_x192_acc3()
96 __m512 vt7 = _mm512_fmadd_ps(vn7, vminus_ln2_hi, vx7); in xnn_f32_raddstoreexpminusmax_ukernel__avx512f_p5_scalef_x192_acc3()
97 __m512 vt8 = _mm512_fmadd_ps(vn8, vminus_ln2_hi, vx8); in xnn_f32_raddstoreexpminusmax_ukernel__avx512f_p5_scalef_x192_acc3()
98 __m512 vt9 = _mm512_fmadd_ps(vn9, vminus_ln2_hi, vx9); in xnn_f32_raddstoreexpminusmax_ukernel__avx512f_p5_scalef_x192_acc3()
[all …]
Davx512f-p5-scalef-x160.c81 __m512 vt0 = _mm512_fmadd_ps(vn0, vminus_ln2_hi, vx0); in xnn_f32_raddstoreexpminusmax_ukernel__avx512f_p5_scalef_x160()
82 __m512 vt1 = _mm512_fmadd_ps(vn1, vminus_ln2_hi, vx1); in xnn_f32_raddstoreexpminusmax_ukernel__avx512f_p5_scalef_x160()
83 __m512 vt2 = _mm512_fmadd_ps(vn2, vminus_ln2_hi, vx2); in xnn_f32_raddstoreexpminusmax_ukernel__avx512f_p5_scalef_x160()
84 __m512 vt3 = _mm512_fmadd_ps(vn3, vminus_ln2_hi, vx3); in xnn_f32_raddstoreexpminusmax_ukernel__avx512f_p5_scalef_x160()
85 __m512 vt4 = _mm512_fmadd_ps(vn4, vminus_ln2_hi, vx4); in xnn_f32_raddstoreexpminusmax_ukernel__avx512f_p5_scalef_x160()
86 __m512 vt5 = _mm512_fmadd_ps(vn5, vminus_ln2_hi, vx5); in xnn_f32_raddstoreexpminusmax_ukernel__avx512f_p5_scalef_x160()
87 __m512 vt6 = _mm512_fmadd_ps(vn6, vminus_ln2_hi, vx6); in xnn_f32_raddstoreexpminusmax_ukernel__avx512f_p5_scalef_x160()
88 __m512 vt7 = _mm512_fmadd_ps(vn7, vminus_ln2_hi, vx7); in xnn_f32_raddstoreexpminusmax_ukernel__avx512f_p5_scalef_x160()
89 __m512 vt8 = _mm512_fmadd_ps(vn8, vminus_ln2_hi, vx8); in xnn_f32_raddstoreexpminusmax_ukernel__avx512f_p5_scalef_x160()
90 __m512 vt9 = _mm512_fmadd_ps(vn9, vminus_ln2_hi, vx9); in xnn_f32_raddstoreexpminusmax_ukernel__avx512f_p5_scalef_x160()
[all …]
Davx512f-p5-scalef-x160-acc2.c82 __m512 vt0 = _mm512_fmadd_ps(vn0, vminus_ln2_hi, vx0); in xnn_f32_raddstoreexpminusmax_ukernel__avx512f_p5_scalef_x160_acc2()
83 __m512 vt1 = _mm512_fmadd_ps(vn1, vminus_ln2_hi, vx1); in xnn_f32_raddstoreexpminusmax_ukernel__avx512f_p5_scalef_x160_acc2()
84 __m512 vt2 = _mm512_fmadd_ps(vn2, vminus_ln2_hi, vx2); in xnn_f32_raddstoreexpminusmax_ukernel__avx512f_p5_scalef_x160_acc2()
85 __m512 vt3 = _mm512_fmadd_ps(vn3, vminus_ln2_hi, vx3); in xnn_f32_raddstoreexpminusmax_ukernel__avx512f_p5_scalef_x160_acc2()
86 __m512 vt4 = _mm512_fmadd_ps(vn4, vminus_ln2_hi, vx4); in xnn_f32_raddstoreexpminusmax_ukernel__avx512f_p5_scalef_x160_acc2()
87 __m512 vt5 = _mm512_fmadd_ps(vn5, vminus_ln2_hi, vx5); in xnn_f32_raddstoreexpminusmax_ukernel__avx512f_p5_scalef_x160_acc2()
88 __m512 vt6 = _mm512_fmadd_ps(vn6, vminus_ln2_hi, vx6); in xnn_f32_raddstoreexpminusmax_ukernel__avx512f_p5_scalef_x160_acc2()
89 __m512 vt7 = _mm512_fmadd_ps(vn7, vminus_ln2_hi, vx7); in xnn_f32_raddstoreexpminusmax_ukernel__avx512f_p5_scalef_x160_acc2()
90 __m512 vt8 = _mm512_fmadd_ps(vn8, vminus_ln2_hi, vx8); in xnn_f32_raddstoreexpminusmax_ukernel__avx512f_p5_scalef_x160_acc2()
91 __m512 vt9 = _mm512_fmadd_ps(vn9, vminus_ln2_hi, vx9); in xnn_f32_raddstoreexpminusmax_ukernel__avx512f_p5_scalef_x160_acc2()
[all …]
Davx512f-p5-scalef-x192-acc6.c92 __m512 vt0 = _mm512_fmadd_ps(vn0, vminus_ln2_hi, vx0); in xnn_f32_raddstoreexpminusmax_ukernel__avx512f_p5_scalef_x192_acc6()
93 __m512 vt1 = _mm512_fmadd_ps(vn1, vminus_ln2_hi, vx1); in xnn_f32_raddstoreexpminusmax_ukernel__avx512f_p5_scalef_x192_acc6()
94 __m512 vt2 = _mm512_fmadd_ps(vn2, vminus_ln2_hi, vx2); in xnn_f32_raddstoreexpminusmax_ukernel__avx512f_p5_scalef_x192_acc6()
95 __m512 vt3 = _mm512_fmadd_ps(vn3, vminus_ln2_hi, vx3); in xnn_f32_raddstoreexpminusmax_ukernel__avx512f_p5_scalef_x192_acc6()
96 __m512 vt4 = _mm512_fmadd_ps(vn4, vminus_ln2_hi, vx4); in xnn_f32_raddstoreexpminusmax_ukernel__avx512f_p5_scalef_x192_acc6()
97 __m512 vt5 = _mm512_fmadd_ps(vn5, vminus_ln2_hi, vx5); in xnn_f32_raddstoreexpminusmax_ukernel__avx512f_p5_scalef_x192_acc6()
98 __m512 vt6 = _mm512_fmadd_ps(vn6, vminus_ln2_hi, vx6); in xnn_f32_raddstoreexpminusmax_ukernel__avx512f_p5_scalef_x192_acc6()
99 __m512 vt7 = _mm512_fmadd_ps(vn7, vminus_ln2_hi, vx7); in xnn_f32_raddstoreexpminusmax_ukernel__avx512f_p5_scalef_x192_acc6()
100 __m512 vt8 = _mm512_fmadd_ps(vn8, vminus_ln2_hi, vx8); in xnn_f32_raddstoreexpminusmax_ukernel__avx512f_p5_scalef_x192_acc6()
101 __m512 vt9 = _mm512_fmadd_ps(vn9, vminus_ln2_hi, vx9); in xnn_f32_raddstoreexpminusmax_ukernel__avx512f_p5_scalef_x192_acc6()
[all …]
Davx512f-p5-scalef-x144-acc3.c80 __m512 vt0 = _mm512_fmadd_ps(vn0, vminus_ln2_hi, vx0); in xnn_f32_raddstoreexpminusmax_ukernel__avx512f_p5_scalef_x144_acc3()
81 __m512 vt1 = _mm512_fmadd_ps(vn1, vminus_ln2_hi, vx1); in xnn_f32_raddstoreexpminusmax_ukernel__avx512f_p5_scalef_x144_acc3()
82 __m512 vt2 = _mm512_fmadd_ps(vn2, vminus_ln2_hi, vx2); in xnn_f32_raddstoreexpminusmax_ukernel__avx512f_p5_scalef_x144_acc3()
83 __m512 vt3 = _mm512_fmadd_ps(vn3, vminus_ln2_hi, vx3); in xnn_f32_raddstoreexpminusmax_ukernel__avx512f_p5_scalef_x144_acc3()
84 __m512 vt4 = _mm512_fmadd_ps(vn4, vminus_ln2_hi, vx4); in xnn_f32_raddstoreexpminusmax_ukernel__avx512f_p5_scalef_x144_acc3()
85 __m512 vt5 = _mm512_fmadd_ps(vn5, vminus_ln2_hi, vx5); in xnn_f32_raddstoreexpminusmax_ukernel__avx512f_p5_scalef_x144_acc3()
86 __m512 vt6 = _mm512_fmadd_ps(vn6, vminus_ln2_hi, vx6); in xnn_f32_raddstoreexpminusmax_ukernel__avx512f_p5_scalef_x144_acc3()
87 __m512 vt7 = _mm512_fmadd_ps(vn7, vminus_ln2_hi, vx7); in xnn_f32_raddstoreexpminusmax_ukernel__avx512f_p5_scalef_x144_acc3()
88 __m512 vt8 = _mm512_fmadd_ps(vn8, vminus_ln2_hi, vx8); in xnn_f32_raddstoreexpminusmax_ukernel__avx512f_p5_scalef_x144_acc3()
90 vt0 = _mm512_fmadd_ps(vn0, vminus_ln2_lo, vt0); in xnn_f32_raddstoreexpminusmax_ukernel__avx512f_p5_scalef_x144_acc3()
[all …]
Davx512f-p5-scalef-x144.c78 __m512 vt0 = _mm512_fmadd_ps(vn0, vminus_ln2_hi, vx0); in xnn_f32_raddstoreexpminusmax_ukernel__avx512f_p5_scalef_x144()
79 __m512 vt1 = _mm512_fmadd_ps(vn1, vminus_ln2_hi, vx1); in xnn_f32_raddstoreexpminusmax_ukernel__avx512f_p5_scalef_x144()
80 __m512 vt2 = _mm512_fmadd_ps(vn2, vminus_ln2_hi, vx2); in xnn_f32_raddstoreexpminusmax_ukernel__avx512f_p5_scalef_x144()
81 __m512 vt3 = _mm512_fmadd_ps(vn3, vminus_ln2_hi, vx3); in xnn_f32_raddstoreexpminusmax_ukernel__avx512f_p5_scalef_x144()
82 __m512 vt4 = _mm512_fmadd_ps(vn4, vminus_ln2_hi, vx4); in xnn_f32_raddstoreexpminusmax_ukernel__avx512f_p5_scalef_x144()
83 __m512 vt5 = _mm512_fmadd_ps(vn5, vminus_ln2_hi, vx5); in xnn_f32_raddstoreexpminusmax_ukernel__avx512f_p5_scalef_x144()
84 __m512 vt6 = _mm512_fmadd_ps(vn6, vminus_ln2_hi, vx6); in xnn_f32_raddstoreexpminusmax_ukernel__avx512f_p5_scalef_x144()
85 __m512 vt7 = _mm512_fmadd_ps(vn7, vminus_ln2_hi, vx7); in xnn_f32_raddstoreexpminusmax_ukernel__avx512f_p5_scalef_x144()
86 __m512 vt8 = _mm512_fmadd_ps(vn8, vminus_ln2_hi, vx8); in xnn_f32_raddstoreexpminusmax_ukernel__avx512f_p5_scalef_x144()
88 vt0 = _mm512_fmadd_ps(vn0, vminus_ln2_lo, vt0); in xnn_f32_raddstoreexpminusmax_ukernel__avx512f_p5_scalef_x144()
[all …]
/external/XNNPACK/src/f32-raddexpminusmax/gen/
Davx512f-p5-scalef-x192.c86 __m512 vt0 = _mm512_fmadd_ps(vn0, vminus_ln2_hi, vx0); in xnn_f32_raddexpminusmax_ukernel__avx512f_p5_scalef_x192()
87 __m512 vt1 = _mm512_fmadd_ps(vn1, vminus_ln2_hi, vx1); in xnn_f32_raddexpminusmax_ukernel__avx512f_p5_scalef_x192()
88 __m512 vt2 = _mm512_fmadd_ps(vn2, vminus_ln2_hi, vx2); in xnn_f32_raddexpminusmax_ukernel__avx512f_p5_scalef_x192()
89 __m512 vt3 = _mm512_fmadd_ps(vn3, vminus_ln2_hi, vx3); in xnn_f32_raddexpminusmax_ukernel__avx512f_p5_scalef_x192()
90 __m512 vt4 = _mm512_fmadd_ps(vn4, vminus_ln2_hi, vx4); in xnn_f32_raddexpminusmax_ukernel__avx512f_p5_scalef_x192()
91 __m512 vt5 = _mm512_fmadd_ps(vn5, vminus_ln2_hi, vx5); in xnn_f32_raddexpminusmax_ukernel__avx512f_p5_scalef_x192()
92 __m512 vt6 = _mm512_fmadd_ps(vn6, vminus_ln2_hi, vx6); in xnn_f32_raddexpminusmax_ukernel__avx512f_p5_scalef_x192()
93 __m512 vt7 = _mm512_fmadd_ps(vn7, vminus_ln2_hi, vx7); in xnn_f32_raddexpminusmax_ukernel__avx512f_p5_scalef_x192()
94 __m512 vt8 = _mm512_fmadd_ps(vn8, vminus_ln2_hi, vx8); in xnn_f32_raddexpminusmax_ukernel__avx512f_p5_scalef_x192()
95 __m512 vt9 = _mm512_fmadd_ps(vn9, vminus_ln2_hi, vx9); in xnn_f32_raddexpminusmax_ukernel__avx512f_p5_scalef_x192()
[all …]
Davx512f-p5-scalef-x192-acc3.c88 __m512 vt0 = _mm512_fmadd_ps(vn0, vminus_ln2_hi, vx0); in xnn_f32_raddexpminusmax_ukernel__avx512f_p5_scalef_x192_acc3()
89 __m512 vt1 = _mm512_fmadd_ps(vn1, vminus_ln2_hi, vx1); in xnn_f32_raddexpminusmax_ukernel__avx512f_p5_scalef_x192_acc3()
90 __m512 vt2 = _mm512_fmadd_ps(vn2, vminus_ln2_hi, vx2); in xnn_f32_raddexpminusmax_ukernel__avx512f_p5_scalef_x192_acc3()
91 __m512 vt3 = _mm512_fmadd_ps(vn3, vminus_ln2_hi, vx3); in xnn_f32_raddexpminusmax_ukernel__avx512f_p5_scalef_x192_acc3()
92 __m512 vt4 = _mm512_fmadd_ps(vn4, vminus_ln2_hi, vx4); in xnn_f32_raddexpminusmax_ukernel__avx512f_p5_scalef_x192_acc3()
93 __m512 vt5 = _mm512_fmadd_ps(vn5, vminus_ln2_hi, vx5); in xnn_f32_raddexpminusmax_ukernel__avx512f_p5_scalef_x192_acc3()
94 __m512 vt6 = _mm512_fmadd_ps(vn6, vminus_ln2_hi, vx6); in xnn_f32_raddexpminusmax_ukernel__avx512f_p5_scalef_x192_acc3()
95 __m512 vt7 = _mm512_fmadd_ps(vn7, vminus_ln2_hi, vx7); in xnn_f32_raddexpminusmax_ukernel__avx512f_p5_scalef_x192_acc3()
96 __m512 vt8 = _mm512_fmadd_ps(vn8, vminus_ln2_hi, vx8); in xnn_f32_raddexpminusmax_ukernel__avx512f_p5_scalef_x192_acc3()
97 __m512 vt9 = _mm512_fmadd_ps(vn9, vminus_ln2_hi, vx9); in xnn_f32_raddexpminusmax_ukernel__avx512f_p5_scalef_x192_acc3()
[all …]
Davx512f-p5-scalef-x192-acc6.c91 __m512 vt0 = _mm512_fmadd_ps(vn0, vminus_ln2_hi, vx0); in xnn_f32_raddexpminusmax_ukernel__avx512f_p5_scalef_x192_acc6()
92 __m512 vt1 = _mm512_fmadd_ps(vn1, vminus_ln2_hi, vx1); in xnn_f32_raddexpminusmax_ukernel__avx512f_p5_scalef_x192_acc6()
93 __m512 vt2 = _mm512_fmadd_ps(vn2, vminus_ln2_hi, vx2); in xnn_f32_raddexpminusmax_ukernel__avx512f_p5_scalef_x192_acc6()
94 __m512 vt3 = _mm512_fmadd_ps(vn3, vminus_ln2_hi, vx3); in xnn_f32_raddexpminusmax_ukernel__avx512f_p5_scalef_x192_acc6()
95 __m512 vt4 = _mm512_fmadd_ps(vn4, vminus_ln2_hi, vx4); in xnn_f32_raddexpminusmax_ukernel__avx512f_p5_scalef_x192_acc6()
96 __m512 vt5 = _mm512_fmadd_ps(vn5, vminus_ln2_hi, vx5); in xnn_f32_raddexpminusmax_ukernel__avx512f_p5_scalef_x192_acc6()
97 __m512 vt6 = _mm512_fmadd_ps(vn6, vminus_ln2_hi, vx6); in xnn_f32_raddexpminusmax_ukernel__avx512f_p5_scalef_x192_acc6()
98 __m512 vt7 = _mm512_fmadd_ps(vn7, vminus_ln2_hi, vx7); in xnn_f32_raddexpminusmax_ukernel__avx512f_p5_scalef_x192_acc6()
99 __m512 vt8 = _mm512_fmadd_ps(vn8, vminus_ln2_hi, vx8); in xnn_f32_raddexpminusmax_ukernel__avx512f_p5_scalef_x192_acc6()
100 __m512 vt9 = _mm512_fmadd_ps(vn9, vminus_ln2_hi, vx9); in xnn_f32_raddexpminusmax_ukernel__avx512f_p5_scalef_x192_acc6()
[all …]
Davx512f-p5-scalef-x192-acc2.c87 __m512 vt0 = _mm512_fmadd_ps(vn0, vminus_ln2_hi, vx0); in xnn_f32_raddexpminusmax_ukernel__avx512f_p5_scalef_x192_acc2()
88 __m512 vt1 = _mm512_fmadd_ps(vn1, vminus_ln2_hi, vx1); in xnn_f32_raddexpminusmax_ukernel__avx512f_p5_scalef_x192_acc2()
89 __m512 vt2 = _mm512_fmadd_ps(vn2, vminus_ln2_hi, vx2); in xnn_f32_raddexpminusmax_ukernel__avx512f_p5_scalef_x192_acc2()
90 __m512 vt3 = _mm512_fmadd_ps(vn3, vminus_ln2_hi, vx3); in xnn_f32_raddexpminusmax_ukernel__avx512f_p5_scalef_x192_acc2()
91 __m512 vt4 = _mm512_fmadd_ps(vn4, vminus_ln2_hi, vx4); in xnn_f32_raddexpminusmax_ukernel__avx512f_p5_scalef_x192_acc2()
92 __m512 vt5 = _mm512_fmadd_ps(vn5, vminus_ln2_hi, vx5); in xnn_f32_raddexpminusmax_ukernel__avx512f_p5_scalef_x192_acc2()
93 __m512 vt6 = _mm512_fmadd_ps(vn6, vminus_ln2_hi, vx6); in xnn_f32_raddexpminusmax_ukernel__avx512f_p5_scalef_x192_acc2()
94 __m512 vt7 = _mm512_fmadd_ps(vn7, vminus_ln2_hi, vx7); in xnn_f32_raddexpminusmax_ukernel__avx512f_p5_scalef_x192_acc2()
95 __m512 vt8 = _mm512_fmadd_ps(vn8, vminus_ln2_hi, vx8); in xnn_f32_raddexpminusmax_ukernel__avx512f_p5_scalef_x192_acc2()
96 __m512 vt9 = _mm512_fmadd_ps(vn9, vminus_ln2_hi, vx9); in xnn_f32_raddexpminusmax_ukernel__avx512f_p5_scalef_x192_acc2()
[all …]
Davx512f-p5-scalef-x160.c80 __m512 vt0 = _mm512_fmadd_ps(vn0, vminus_ln2_hi, vx0); in xnn_f32_raddexpminusmax_ukernel__avx512f_p5_scalef_x160()
81 __m512 vt1 = _mm512_fmadd_ps(vn1, vminus_ln2_hi, vx1); in xnn_f32_raddexpminusmax_ukernel__avx512f_p5_scalef_x160()
82 __m512 vt2 = _mm512_fmadd_ps(vn2, vminus_ln2_hi, vx2); in xnn_f32_raddexpminusmax_ukernel__avx512f_p5_scalef_x160()
83 __m512 vt3 = _mm512_fmadd_ps(vn3, vminus_ln2_hi, vx3); in xnn_f32_raddexpminusmax_ukernel__avx512f_p5_scalef_x160()
84 __m512 vt4 = _mm512_fmadd_ps(vn4, vminus_ln2_hi, vx4); in xnn_f32_raddexpminusmax_ukernel__avx512f_p5_scalef_x160()
85 __m512 vt5 = _mm512_fmadd_ps(vn5, vminus_ln2_hi, vx5); in xnn_f32_raddexpminusmax_ukernel__avx512f_p5_scalef_x160()
86 __m512 vt6 = _mm512_fmadd_ps(vn6, vminus_ln2_hi, vx6); in xnn_f32_raddexpminusmax_ukernel__avx512f_p5_scalef_x160()
87 __m512 vt7 = _mm512_fmadd_ps(vn7, vminus_ln2_hi, vx7); in xnn_f32_raddexpminusmax_ukernel__avx512f_p5_scalef_x160()
88 __m512 vt8 = _mm512_fmadd_ps(vn8, vminus_ln2_hi, vx8); in xnn_f32_raddexpminusmax_ukernel__avx512f_p5_scalef_x160()
89 __m512 vt9 = _mm512_fmadd_ps(vn9, vminus_ln2_hi, vx9); in xnn_f32_raddexpminusmax_ukernel__avx512f_p5_scalef_x160()
[all …]
Davx512f-p5-scalef-x160-acc2.c81 __m512 vt0 = _mm512_fmadd_ps(vn0, vminus_ln2_hi, vx0); in xnn_f32_raddexpminusmax_ukernel__avx512f_p5_scalef_x160_acc2()
82 __m512 vt1 = _mm512_fmadd_ps(vn1, vminus_ln2_hi, vx1); in xnn_f32_raddexpminusmax_ukernel__avx512f_p5_scalef_x160_acc2()
83 __m512 vt2 = _mm512_fmadd_ps(vn2, vminus_ln2_hi, vx2); in xnn_f32_raddexpminusmax_ukernel__avx512f_p5_scalef_x160_acc2()
84 __m512 vt3 = _mm512_fmadd_ps(vn3, vminus_ln2_hi, vx3); in xnn_f32_raddexpminusmax_ukernel__avx512f_p5_scalef_x160_acc2()
85 __m512 vt4 = _mm512_fmadd_ps(vn4, vminus_ln2_hi, vx4); in xnn_f32_raddexpminusmax_ukernel__avx512f_p5_scalef_x160_acc2()
86 __m512 vt5 = _mm512_fmadd_ps(vn5, vminus_ln2_hi, vx5); in xnn_f32_raddexpminusmax_ukernel__avx512f_p5_scalef_x160_acc2()
87 __m512 vt6 = _mm512_fmadd_ps(vn6, vminus_ln2_hi, vx6); in xnn_f32_raddexpminusmax_ukernel__avx512f_p5_scalef_x160_acc2()
88 __m512 vt7 = _mm512_fmadd_ps(vn7, vminus_ln2_hi, vx7); in xnn_f32_raddexpminusmax_ukernel__avx512f_p5_scalef_x160_acc2()
89 __m512 vt8 = _mm512_fmadd_ps(vn8, vminus_ln2_hi, vx8); in xnn_f32_raddexpminusmax_ukernel__avx512f_p5_scalef_x160_acc2()
90 __m512 vt9 = _mm512_fmadd_ps(vn9, vminus_ln2_hi, vx9); in xnn_f32_raddexpminusmax_ukernel__avx512f_p5_scalef_x160_acc2()
[all …]
Davx512f-p5-scalef-x160-acc5.c84 __m512 vt0 = _mm512_fmadd_ps(vn0, vminus_ln2_hi, vx0); in xnn_f32_raddexpminusmax_ukernel__avx512f_p5_scalef_x160_acc5()
85 __m512 vt1 = _mm512_fmadd_ps(vn1, vminus_ln2_hi, vx1); in xnn_f32_raddexpminusmax_ukernel__avx512f_p5_scalef_x160_acc5()
86 __m512 vt2 = _mm512_fmadd_ps(vn2, vminus_ln2_hi, vx2); in xnn_f32_raddexpminusmax_ukernel__avx512f_p5_scalef_x160_acc5()
87 __m512 vt3 = _mm512_fmadd_ps(vn3, vminus_ln2_hi, vx3); in xnn_f32_raddexpminusmax_ukernel__avx512f_p5_scalef_x160_acc5()
88 __m512 vt4 = _mm512_fmadd_ps(vn4, vminus_ln2_hi, vx4); in xnn_f32_raddexpminusmax_ukernel__avx512f_p5_scalef_x160_acc5()
89 __m512 vt5 = _mm512_fmadd_ps(vn5, vminus_ln2_hi, vx5); in xnn_f32_raddexpminusmax_ukernel__avx512f_p5_scalef_x160_acc5()
90 __m512 vt6 = _mm512_fmadd_ps(vn6, vminus_ln2_hi, vx6); in xnn_f32_raddexpminusmax_ukernel__avx512f_p5_scalef_x160_acc5()
91 __m512 vt7 = _mm512_fmadd_ps(vn7, vminus_ln2_hi, vx7); in xnn_f32_raddexpminusmax_ukernel__avx512f_p5_scalef_x160_acc5()
92 __m512 vt8 = _mm512_fmadd_ps(vn8, vminus_ln2_hi, vx8); in xnn_f32_raddexpminusmax_ukernel__avx512f_p5_scalef_x160_acc5()
93 __m512 vt9 = _mm512_fmadd_ps(vn9, vminus_ln2_hi, vx9); in xnn_f32_raddexpminusmax_ukernel__avx512f_p5_scalef_x160_acc5()
[all …]
Davx512f-p5-scalef-x144.c77 __m512 vt0 = _mm512_fmadd_ps(vn0, vminus_ln2_hi, vx0); in xnn_f32_raddexpminusmax_ukernel__avx512f_p5_scalef_x144()
78 __m512 vt1 = _mm512_fmadd_ps(vn1, vminus_ln2_hi, vx1); in xnn_f32_raddexpminusmax_ukernel__avx512f_p5_scalef_x144()
79 __m512 vt2 = _mm512_fmadd_ps(vn2, vminus_ln2_hi, vx2); in xnn_f32_raddexpminusmax_ukernel__avx512f_p5_scalef_x144()
80 __m512 vt3 = _mm512_fmadd_ps(vn3, vminus_ln2_hi, vx3); in xnn_f32_raddexpminusmax_ukernel__avx512f_p5_scalef_x144()
81 __m512 vt4 = _mm512_fmadd_ps(vn4, vminus_ln2_hi, vx4); in xnn_f32_raddexpminusmax_ukernel__avx512f_p5_scalef_x144()
82 __m512 vt5 = _mm512_fmadd_ps(vn5, vminus_ln2_hi, vx5); in xnn_f32_raddexpminusmax_ukernel__avx512f_p5_scalef_x144()
83 __m512 vt6 = _mm512_fmadd_ps(vn6, vminus_ln2_hi, vx6); in xnn_f32_raddexpminusmax_ukernel__avx512f_p5_scalef_x144()
84 __m512 vt7 = _mm512_fmadd_ps(vn7, vminus_ln2_hi, vx7); in xnn_f32_raddexpminusmax_ukernel__avx512f_p5_scalef_x144()
85 __m512 vt8 = _mm512_fmadd_ps(vn8, vminus_ln2_hi, vx8); in xnn_f32_raddexpminusmax_ukernel__avx512f_p5_scalef_x144()
87 vt0 = _mm512_fmadd_ps(vn0, vminus_ln2_lo, vt0); in xnn_f32_raddexpminusmax_ukernel__avx512f_p5_scalef_x144()
[all …]
/external/XNNPACK/src/f32-vscaleexpminusmax/gen/
Davx512f-p5-scalef-x192.c87 __m512 vt0 = _mm512_fmadd_ps(vn0, vminus_ln2_hi, vx0); in xnn_f32_vscaleexpminusmax_ukernel__avx512f_p5_scalef_x192()
88 __m512 vt1 = _mm512_fmadd_ps(vn1, vminus_ln2_hi, vx1); in xnn_f32_vscaleexpminusmax_ukernel__avx512f_p5_scalef_x192()
89 __m512 vt2 = _mm512_fmadd_ps(vn2, vminus_ln2_hi, vx2); in xnn_f32_vscaleexpminusmax_ukernel__avx512f_p5_scalef_x192()
90 __m512 vt3 = _mm512_fmadd_ps(vn3, vminus_ln2_hi, vx3); in xnn_f32_vscaleexpminusmax_ukernel__avx512f_p5_scalef_x192()
91 __m512 vt4 = _mm512_fmadd_ps(vn4, vminus_ln2_hi, vx4); in xnn_f32_vscaleexpminusmax_ukernel__avx512f_p5_scalef_x192()
92 __m512 vt5 = _mm512_fmadd_ps(vn5, vminus_ln2_hi, vx5); in xnn_f32_vscaleexpminusmax_ukernel__avx512f_p5_scalef_x192()
93 __m512 vt6 = _mm512_fmadd_ps(vn6, vminus_ln2_hi, vx6); in xnn_f32_vscaleexpminusmax_ukernel__avx512f_p5_scalef_x192()
94 __m512 vt7 = _mm512_fmadd_ps(vn7, vminus_ln2_hi, vx7); in xnn_f32_vscaleexpminusmax_ukernel__avx512f_p5_scalef_x192()
95 __m512 vt8 = _mm512_fmadd_ps(vn8, vminus_ln2_hi, vx8); in xnn_f32_vscaleexpminusmax_ukernel__avx512f_p5_scalef_x192()
96 __m512 vt9 = _mm512_fmadd_ps(vn9, vminus_ln2_hi, vx9); in xnn_f32_vscaleexpminusmax_ukernel__avx512f_p5_scalef_x192()
[all …]
Davx512f-p5-scalef-x176.c84 __m512 vt0 = _mm512_fmadd_ps(vn0, vminus_ln2_hi, vx0); in xnn_f32_vscaleexpminusmax_ukernel__avx512f_p5_scalef_x176()
85 __m512 vt1 = _mm512_fmadd_ps(vn1, vminus_ln2_hi, vx1); in xnn_f32_vscaleexpminusmax_ukernel__avx512f_p5_scalef_x176()
86 __m512 vt2 = _mm512_fmadd_ps(vn2, vminus_ln2_hi, vx2); in xnn_f32_vscaleexpminusmax_ukernel__avx512f_p5_scalef_x176()
87 __m512 vt3 = _mm512_fmadd_ps(vn3, vminus_ln2_hi, vx3); in xnn_f32_vscaleexpminusmax_ukernel__avx512f_p5_scalef_x176()
88 __m512 vt4 = _mm512_fmadd_ps(vn4, vminus_ln2_hi, vx4); in xnn_f32_vscaleexpminusmax_ukernel__avx512f_p5_scalef_x176()
89 __m512 vt5 = _mm512_fmadd_ps(vn5, vminus_ln2_hi, vx5); in xnn_f32_vscaleexpminusmax_ukernel__avx512f_p5_scalef_x176()
90 __m512 vt6 = _mm512_fmadd_ps(vn6, vminus_ln2_hi, vx6); in xnn_f32_vscaleexpminusmax_ukernel__avx512f_p5_scalef_x176()
91 __m512 vt7 = _mm512_fmadd_ps(vn7, vminus_ln2_hi, vx7); in xnn_f32_vscaleexpminusmax_ukernel__avx512f_p5_scalef_x176()
92 __m512 vt8 = _mm512_fmadd_ps(vn8, vminus_ln2_hi, vx8); in xnn_f32_vscaleexpminusmax_ukernel__avx512f_p5_scalef_x176()
93 __m512 vt9 = _mm512_fmadd_ps(vn9, vminus_ln2_hi, vx9); in xnn_f32_vscaleexpminusmax_ukernel__avx512f_p5_scalef_x176()
[all …]
Davx512f-p5-scalef-x160.c81 __m512 vt0 = _mm512_fmadd_ps(vn0, vminus_ln2_hi, vx0); in xnn_f32_vscaleexpminusmax_ukernel__avx512f_p5_scalef_x160()
82 __m512 vt1 = _mm512_fmadd_ps(vn1, vminus_ln2_hi, vx1); in xnn_f32_vscaleexpminusmax_ukernel__avx512f_p5_scalef_x160()
83 __m512 vt2 = _mm512_fmadd_ps(vn2, vminus_ln2_hi, vx2); in xnn_f32_vscaleexpminusmax_ukernel__avx512f_p5_scalef_x160()
84 __m512 vt3 = _mm512_fmadd_ps(vn3, vminus_ln2_hi, vx3); in xnn_f32_vscaleexpminusmax_ukernel__avx512f_p5_scalef_x160()
85 __m512 vt4 = _mm512_fmadd_ps(vn4, vminus_ln2_hi, vx4); in xnn_f32_vscaleexpminusmax_ukernel__avx512f_p5_scalef_x160()
86 __m512 vt5 = _mm512_fmadd_ps(vn5, vminus_ln2_hi, vx5); in xnn_f32_vscaleexpminusmax_ukernel__avx512f_p5_scalef_x160()
87 __m512 vt6 = _mm512_fmadd_ps(vn6, vminus_ln2_hi, vx6); in xnn_f32_vscaleexpminusmax_ukernel__avx512f_p5_scalef_x160()
88 __m512 vt7 = _mm512_fmadd_ps(vn7, vminus_ln2_hi, vx7); in xnn_f32_vscaleexpminusmax_ukernel__avx512f_p5_scalef_x160()
89 __m512 vt8 = _mm512_fmadd_ps(vn8, vminus_ln2_hi, vx8); in xnn_f32_vscaleexpminusmax_ukernel__avx512f_p5_scalef_x160()
90 __m512 vt9 = _mm512_fmadd_ps(vn9, vminus_ln2_hi, vx9); in xnn_f32_vscaleexpminusmax_ukernel__avx512f_p5_scalef_x160()
[all …]
/external/XNNPACK/src/f32-vscaleextexp/gen/
Davx512f-p5-scalef-x192.c74 __m512 vt0 = _mm512_fmadd_ps(vn0, vminus_ln2_hi, vx0); in xnn_f32_vscaleextexp_ukernel__avx512f_p5_scalef_x192()
75 __m512 vt1 = _mm512_fmadd_ps(vn1, vminus_ln2_hi, vx1); in xnn_f32_vscaleextexp_ukernel__avx512f_p5_scalef_x192()
76 __m512 vt2 = _mm512_fmadd_ps(vn2, vminus_ln2_hi, vx2); in xnn_f32_vscaleextexp_ukernel__avx512f_p5_scalef_x192()
77 __m512 vt3 = _mm512_fmadd_ps(vn3, vminus_ln2_hi, vx3); in xnn_f32_vscaleextexp_ukernel__avx512f_p5_scalef_x192()
78 __m512 vt4 = _mm512_fmadd_ps(vn4, vminus_ln2_hi, vx4); in xnn_f32_vscaleextexp_ukernel__avx512f_p5_scalef_x192()
79 __m512 vt5 = _mm512_fmadd_ps(vn5, vminus_ln2_hi, vx5); in xnn_f32_vscaleextexp_ukernel__avx512f_p5_scalef_x192()
80 __m512 vt6 = _mm512_fmadd_ps(vn6, vminus_ln2_hi, vx6); in xnn_f32_vscaleextexp_ukernel__avx512f_p5_scalef_x192()
81 __m512 vt7 = _mm512_fmadd_ps(vn7, vminus_ln2_hi, vx7); in xnn_f32_vscaleextexp_ukernel__avx512f_p5_scalef_x192()
82 __m512 vt8 = _mm512_fmadd_ps(vn8, vminus_ln2_hi, vx8); in xnn_f32_vscaleextexp_ukernel__avx512f_p5_scalef_x192()
83 __m512 vt9 = _mm512_fmadd_ps(vn9, vminus_ln2_hi, vx9); in xnn_f32_vscaleextexp_ukernel__avx512f_p5_scalef_x192()
[all …]
Davx512f-p5-scalef-x176.c72 __m512 vt0 = _mm512_fmadd_ps(vn0, vminus_ln2_hi, vx0); in xnn_f32_vscaleextexp_ukernel__avx512f_p5_scalef_x176()
73 __m512 vt1 = _mm512_fmadd_ps(vn1, vminus_ln2_hi, vx1); in xnn_f32_vscaleextexp_ukernel__avx512f_p5_scalef_x176()
74 __m512 vt2 = _mm512_fmadd_ps(vn2, vminus_ln2_hi, vx2); in xnn_f32_vscaleextexp_ukernel__avx512f_p5_scalef_x176()
75 __m512 vt3 = _mm512_fmadd_ps(vn3, vminus_ln2_hi, vx3); in xnn_f32_vscaleextexp_ukernel__avx512f_p5_scalef_x176()
76 __m512 vt4 = _mm512_fmadd_ps(vn4, vminus_ln2_hi, vx4); in xnn_f32_vscaleextexp_ukernel__avx512f_p5_scalef_x176()
77 __m512 vt5 = _mm512_fmadd_ps(vn5, vminus_ln2_hi, vx5); in xnn_f32_vscaleextexp_ukernel__avx512f_p5_scalef_x176()
78 __m512 vt6 = _mm512_fmadd_ps(vn6, vminus_ln2_hi, vx6); in xnn_f32_vscaleextexp_ukernel__avx512f_p5_scalef_x176()
79 __m512 vt7 = _mm512_fmadd_ps(vn7, vminus_ln2_hi, vx7); in xnn_f32_vscaleextexp_ukernel__avx512f_p5_scalef_x176()
80 __m512 vt8 = _mm512_fmadd_ps(vn8, vminus_ln2_hi, vx8); in xnn_f32_vscaleextexp_ukernel__avx512f_p5_scalef_x176()
81 __m512 vt9 = _mm512_fmadd_ps(vn9, vminus_ln2_hi, vx9); in xnn_f32_vscaleextexp_ukernel__avx512f_p5_scalef_x176()
[all …]
Davx512f-p5-scalef-x160.c70 __m512 vt0 = _mm512_fmadd_ps(vn0, vminus_ln2_hi, vx0); in xnn_f32_vscaleextexp_ukernel__avx512f_p5_scalef_x160()
71 __m512 vt1 = _mm512_fmadd_ps(vn1, vminus_ln2_hi, vx1); in xnn_f32_vscaleextexp_ukernel__avx512f_p5_scalef_x160()
72 __m512 vt2 = _mm512_fmadd_ps(vn2, vminus_ln2_hi, vx2); in xnn_f32_vscaleextexp_ukernel__avx512f_p5_scalef_x160()
73 __m512 vt3 = _mm512_fmadd_ps(vn3, vminus_ln2_hi, vx3); in xnn_f32_vscaleextexp_ukernel__avx512f_p5_scalef_x160()
74 __m512 vt4 = _mm512_fmadd_ps(vn4, vminus_ln2_hi, vx4); in xnn_f32_vscaleextexp_ukernel__avx512f_p5_scalef_x160()
75 __m512 vt5 = _mm512_fmadd_ps(vn5, vminus_ln2_hi, vx5); in xnn_f32_vscaleextexp_ukernel__avx512f_p5_scalef_x160()
76 __m512 vt6 = _mm512_fmadd_ps(vn6, vminus_ln2_hi, vx6); in xnn_f32_vscaleextexp_ukernel__avx512f_p5_scalef_x160()
77 __m512 vt7 = _mm512_fmadd_ps(vn7, vminus_ln2_hi, vx7); in xnn_f32_vscaleextexp_ukernel__avx512f_p5_scalef_x160()
78 __m512 vt8 = _mm512_fmadd_ps(vn8, vminus_ln2_hi, vx8); in xnn_f32_vscaleextexp_ukernel__avx512f_p5_scalef_x160()
79 __m512 vt9 = _mm512_fmadd_ps(vn9, vminus_ln2_hi, vx9); in xnn_f32_vscaleextexp_ukernel__avx512f_p5_scalef_x160()
[all …]
Davx512f-p5-scalef-x144.c68 __m512 vt0 = _mm512_fmadd_ps(vn0, vminus_ln2_hi, vx0); in xnn_f32_vscaleextexp_ukernel__avx512f_p5_scalef_x144()
69 __m512 vt1 = _mm512_fmadd_ps(vn1, vminus_ln2_hi, vx1); in xnn_f32_vscaleextexp_ukernel__avx512f_p5_scalef_x144()
70 __m512 vt2 = _mm512_fmadd_ps(vn2, vminus_ln2_hi, vx2); in xnn_f32_vscaleextexp_ukernel__avx512f_p5_scalef_x144()
71 __m512 vt3 = _mm512_fmadd_ps(vn3, vminus_ln2_hi, vx3); in xnn_f32_vscaleextexp_ukernel__avx512f_p5_scalef_x144()
72 __m512 vt4 = _mm512_fmadd_ps(vn4, vminus_ln2_hi, vx4); in xnn_f32_vscaleextexp_ukernel__avx512f_p5_scalef_x144()
73 __m512 vt5 = _mm512_fmadd_ps(vn5, vminus_ln2_hi, vx5); in xnn_f32_vscaleextexp_ukernel__avx512f_p5_scalef_x144()
74 __m512 vt6 = _mm512_fmadd_ps(vn6, vminus_ln2_hi, vx6); in xnn_f32_vscaleextexp_ukernel__avx512f_p5_scalef_x144()
75 __m512 vt7 = _mm512_fmadd_ps(vn7, vminus_ln2_hi, vx7); in xnn_f32_vscaleextexp_ukernel__avx512f_p5_scalef_x144()
76 __m512 vt8 = _mm512_fmadd_ps(vn8, vminus_ln2_hi, vx8); in xnn_f32_vscaleextexp_ukernel__avx512f_p5_scalef_x144()
78 vt0 = _mm512_fmadd_ps(vn0, vminus_ln2_lo, vt0); in xnn_f32_vscaleextexp_ukernel__avx512f_p5_scalef_x144()
[all …]
/external/XNNPACK/src/f32-raddextexp/gen/
Davx512f-p5-scalef-x192.c74 __m512 vt0 = _mm512_fmadd_ps(vn0, vminus_ln2_hi, vx0); in xnn_f32_raddextexp_ukernel__avx512f_p5_scalef_x192()
75 __m512 vt1 = _mm512_fmadd_ps(vn1, vminus_ln2_hi, vx1); in xnn_f32_raddextexp_ukernel__avx512f_p5_scalef_x192()
76 __m512 vt2 = _mm512_fmadd_ps(vn2, vminus_ln2_hi, vx2); in xnn_f32_raddextexp_ukernel__avx512f_p5_scalef_x192()
77 __m512 vt3 = _mm512_fmadd_ps(vn3, vminus_ln2_hi, vx3); in xnn_f32_raddextexp_ukernel__avx512f_p5_scalef_x192()
78 __m512 vt4 = _mm512_fmadd_ps(vn4, vminus_ln2_hi, vx4); in xnn_f32_raddextexp_ukernel__avx512f_p5_scalef_x192()
79 __m512 vt5 = _mm512_fmadd_ps(vn5, vminus_ln2_hi, vx5); in xnn_f32_raddextexp_ukernel__avx512f_p5_scalef_x192()
80 __m512 vt6 = _mm512_fmadd_ps(vn6, vminus_ln2_hi, vx6); in xnn_f32_raddextexp_ukernel__avx512f_p5_scalef_x192()
81 __m512 vt7 = _mm512_fmadd_ps(vn7, vminus_ln2_hi, vx7); in xnn_f32_raddextexp_ukernel__avx512f_p5_scalef_x192()
82 __m512 vt8 = _mm512_fmadd_ps(vn8, vminus_ln2_hi, vx8); in xnn_f32_raddextexp_ukernel__avx512f_p5_scalef_x192()
83 __m512 vt9 = _mm512_fmadd_ps(vn9, vminus_ln2_hi, vx9); in xnn_f32_raddextexp_ukernel__avx512f_p5_scalef_x192()
[all …]
Davx512f-p5-scalef-x192-acc2.c76 __m512 vt0 = _mm512_fmadd_ps(vn0, vminus_ln2_hi, vx0); in xnn_f32_raddextexp_ukernel__avx512f_p5_scalef_x192_acc2()
77 __m512 vt1 = _mm512_fmadd_ps(vn1, vminus_ln2_hi, vx1); in xnn_f32_raddextexp_ukernel__avx512f_p5_scalef_x192_acc2()
78 __m512 vt2 = _mm512_fmadd_ps(vn2, vminus_ln2_hi, vx2); in xnn_f32_raddextexp_ukernel__avx512f_p5_scalef_x192_acc2()
79 __m512 vt3 = _mm512_fmadd_ps(vn3, vminus_ln2_hi, vx3); in xnn_f32_raddextexp_ukernel__avx512f_p5_scalef_x192_acc2()
80 __m512 vt4 = _mm512_fmadd_ps(vn4, vminus_ln2_hi, vx4); in xnn_f32_raddextexp_ukernel__avx512f_p5_scalef_x192_acc2()
81 __m512 vt5 = _mm512_fmadd_ps(vn5, vminus_ln2_hi, vx5); in xnn_f32_raddextexp_ukernel__avx512f_p5_scalef_x192_acc2()
82 __m512 vt6 = _mm512_fmadd_ps(vn6, vminus_ln2_hi, vx6); in xnn_f32_raddextexp_ukernel__avx512f_p5_scalef_x192_acc2()
83 __m512 vt7 = _mm512_fmadd_ps(vn7, vminus_ln2_hi, vx7); in xnn_f32_raddextexp_ukernel__avx512f_p5_scalef_x192_acc2()
84 __m512 vt8 = _mm512_fmadd_ps(vn8, vminus_ln2_hi, vx8); in xnn_f32_raddextexp_ukernel__avx512f_p5_scalef_x192_acc2()
85 __m512 vt9 = _mm512_fmadd_ps(vn9, vminus_ln2_hi, vx9); in xnn_f32_raddextexp_ukernel__avx512f_p5_scalef_x192_acc2()
[all …]

123456789