Lines Matching refs:__v16sf

33 typedef float __v16sf __attribute__((__vector_size__(64)));  typedef
706 return (__m512)((__v16sf)__a + (__v16sf)__b); in _mm512_add_ps()
718 return (__m512)((__v16sf)__a * (__v16sf)__b); in _mm512_mul_ps()
730 return (__m512)((__v16sf)__a - (__v16sf)__b); in _mm512_sub_ps()
884 (__m512)__builtin_ia32_maxps512_mask((__v16sf)(__m512)(A), \
885 (__v16sf)(__m512)(B), \
886 (__v16sf)(__m512)(W), (__mmask16)(U), \
890 (__m512)__builtin_ia32_maxps512_mask((__v16sf)(__m512)(A), \
891 (__v16sf)(__m512)(B), \
892 (__v16sf)_mm512_setzero_ps(), \
896 (__m512)__builtin_ia32_maxps512_mask((__v16sf)(__m512)(A), \
897 (__v16sf)(__m512)(B), \
898 (__v16sf)_mm512_undefined_ps(), \
904 return (__m512) __builtin_ia32_maxps512_mask ((__v16sf) __A, in _mm512_max_ps()
905 (__v16sf) __B, in _mm512_max_ps()
906 (__v16sf) in _mm512_max_ps()
915 return (__m512) __builtin_ia32_maxps512_mask ((__v16sf) __A, in _mm512_mask_max_ps()
916 (__v16sf) __B, in _mm512_mask_max_ps()
917 (__v16sf) __W, in _mm512_mask_max_ps()
925 return (__m512) __builtin_ia32_maxps512_mask ((__v16sf) __A, in _mm512_maskz_max_ps()
926 (__v16sf) __B, in _mm512_maskz_max_ps()
927 (__v16sf) in _mm512_maskz_max_ps()
1158 (__m512)__builtin_ia32_minps512_mask((__v16sf)(__m512)(A), \
1159 (__v16sf)(__m512)(B), \
1160 (__v16sf)(__m512)(W), (__mmask16)(U), \
1164 (__m512)__builtin_ia32_minps512_mask((__v16sf)(__m512)(A), \
1165 (__v16sf)(__m512)(B), \
1166 (__v16sf)_mm512_setzero_ps(), \
1170 (__m512)__builtin_ia32_minps512_mask((__v16sf)(__m512)(A), \
1171 (__v16sf)(__m512)(B), \
1172 (__v16sf)_mm512_undefined_ps(), \
1189 return (__m512) __builtin_ia32_minps512_mask ((__v16sf) __A, in _mm512_min_ps()
1190 (__v16sf) __B, in _mm512_min_ps()
1191 (__v16sf) in _mm512_min_ps()
1200 return (__m512) __builtin_ia32_minps512_mask ((__v16sf) __A, in _mm512_mask_min_ps()
1201 (__v16sf) __B, in _mm512_mask_min_ps()
1202 (__v16sf) __W, in _mm512_mask_min_ps()
1210 return (__m512) __builtin_ia32_minps512_mask ((__v16sf) __A, in _mm512_maskz_min_ps()
1211 (__v16sf) __B, in _mm512_maskz_min_ps()
1212 (__v16sf) in _mm512_maskz_min_ps()
1527 (__m512)__builtin_ia32_sqrtps512_mask((__v16sf)(__m512)(A), \
1528 (__v16sf)(__m512)(W), (__mmask16)(U), \
1532 (__m512)__builtin_ia32_sqrtps512_mask((__v16sf)(__m512)(A), \
1533 (__v16sf)_mm512_setzero_ps(), \
1537 (__m512)__builtin_ia32_sqrtps512_mask((__v16sf)(__m512)(A), \
1538 (__v16sf)_mm512_undefined_ps(), \
1544 return (__m512)__builtin_ia32_sqrtps512_mask((__v16sf)__a, in _mm512_sqrt_ps()
1545 (__v16sf) _mm512_setzero_ps (), in _mm512_sqrt_ps()
1553 return (__m512)__builtin_ia32_sqrtps512_mask((__v16sf)__A, in _mm512_mask_sqrt_ps()
1554 (__v16sf) __W, in _mm512_mask_sqrt_ps()
1562 return (__m512)__builtin_ia32_sqrtps512_mask((__v16sf)__A, in _mm512_maskz_sqrt_ps()
1563 (__v16sf) _mm512_setzero_ps (), in _mm512_maskz_sqrt_ps()
1596 return (__m512) __builtin_ia32_rsqrt14ps512_mask ((__v16sf) __A, in _mm512_rsqrt14_ps()
1597 (__v16sf) in _mm512_rsqrt14_ps()
1605 return (__m512) __builtin_ia32_rsqrt14ps512_mask ((__v16sf) __A, in _mm512_mask_rsqrt14_ps()
1606 (__v16sf) __W, in _mm512_mask_rsqrt14_ps()
1613 return (__m512) __builtin_ia32_rsqrt14ps512_mask ((__v16sf) __A, in _mm512_maskz_rsqrt14_ps()
1614 (__v16sf) in _mm512_maskz_rsqrt14_ps()
1704 return (__m512) __builtin_ia32_rcp14ps512_mask ((__v16sf) __A, in _mm512_rcp14_ps()
1705 (__v16sf) in _mm512_rcp14_ps()
1713 return (__m512) __builtin_ia32_rcp14ps512_mask ((__v16sf) __A, in _mm512_mask_rcp14_ps()
1714 (__v16sf) __W, in _mm512_mask_rcp14_ps()
1721 return (__m512) __builtin_ia32_rcp14ps512_mask ((__v16sf) __A, in _mm512_maskz_rcp14_ps()
1722 (__v16sf) in _mm512_maskz_rcp14_ps()
1786 return (__m512) __builtin_ia32_rndscaleps_mask ((__v16sf) __A, in _mm512_floor_ps()
1788 (__v16sf) __A, -1, in _mm512_floor_ps()
1795 return (__m512) __builtin_ia32_rndscaleps_mask ((__v16sf) __A, in _mm512_mask_floor_ps()
1797 (__v16sf) __W, __U, in _mm512_mask_floor_ps()
1822 return (__m512) __builtin_ia32_rndscaleps_mask ((__v16sf) __A, in _mm512_mask_ceil_ps()
1824 (__v16sf) __W, __U, in _mm512_mask_ceil_ps()
1831 return (__m512) __builtin_ia32_rndscaleps_mask ((__v16sf) __A, in _mm512_ceil_ps()
1833 (__v16sf) __A, -1, in _mm512_ceil_ps()
1998 return (__m512) __builtin_ia32_addps512_mask ((__v16sf) __A, in _mm512_mask_add_ps()
1999 (__v16sf) __B, in _mm512_mask_add_ps()
2000 (__v16sf) __W, in _mm512_mask_add_ps()
2007 return (__m512) __builtin_ia32_addps512_mask ((__v16sf) __A, in _mm512_maskz_add_ps()
2008 (__v16sf) __B, in _mm512_maskz_add_ps()
2009 (__v16sf) _mm512_setzero_ps (), in _mm512_maskz_add_ps()
2033 (__m512)__builtin_ia32_addps512_mask((__v16sf)(__m512)(A), \
2034 (__v16sf)(__m512)(B), \
2035 (__v16sf)_mm512_setzero_ps(), \
2039 (__m512)__builtin_ia32_addps512_mask((__v16sf)(__m512)(A), \
2040 (__v16sf)(__m512)(B), \
2041 (__v16sf)(__m512)(W), (__mmask16)(U), \
2045 (__m512)__builtin_ia32_addps512_mask((__v16sf)(__m512)(A), \
2046 (__v16sf)(__m512)(B), \
2047 (__v16sf)_mm512_setzero_ps(), \
2142 return (__m512) __builtin_ia32_subps512_mask ((__v16sf) __A, in _mm512_mask_sub_ps()
2143 (__v16sf) __B, in _mm512_mask_sub_ps()
2144 (__v16sf) __W, in _mm512_mask_sub_ps()
2151 return (__m512) __builtin_ia32_subps512_mask ((__v16sf) __A, in _mm512_maskz_sub_ps()
2152 (__v16sf) __B, in _mm512_maskz_sub_ps()
2153 (__v16sf) in _mm512_maskz_sub_ps()
2178 (__m512)__builtin_ia32_subps512_mask((__v16sf)(__m512)(A), \
2179 (__v16sf)(__m512)(B), \
2180 (__v16sf)_mm512_setzero_ps(), \
2184 (__m512)__builtin_ia32_subps512_mask((__v16sf)(__m512)(A), \
2185 (__v16sf)(__m512)(B), \
2186 (__v16sf)(__m512)(W), (__mmask16)(U), \
2190 (__m512)__builtin_ia32_subps512_mask((__v16sf)(__m512)(A), \
2191 (__v16sf)(__m512)(B), \
2192 (__v16sf)_mm512_setzero_ps(), \
2287 return (__m512) __builtin_ia32_mulps512_mask ((__v16sf) __A, in _mm512_mask_mul_ps()
2288 (__v16sf) __B, in _mm512_mask_mul_ps()
2289 (__v16sf) __W, in _mm512_mask_mul_ps()
2296 return (__m512) __builtin_ia32_mulps512_mask ((__v16sf) __A, in _mm512_maskz_mul_ps()
2297 (__v16sf) __B, in _mm512_maskz_mul_ps()
2298 (__v16sf) in _mm512_maskz_mul_ps()
2323 (__m512)__builtin_ia32_mulps512_mask((__v16sf)(__m512)(A), \
2324 (__v16sf)(__m512)(B), \
2325 (__v16sf)_mm512_setzero_ps(), \
2329 (__m512)__builtin_ia32_mulps512_mask((__v16sf)(__m512)(A), \
2330 (__v16sf)(__m512)(B), \
2331 (__v16sf)(__m512)(W), (__mmask16)(U), \
2335 (__m512)__builtin_ia32_mulps512_mask((__v16sf)(__m512)(A), \
2336 (__v16sf)(__m512)(B), \
2337 (__v16sf)_mm512_setzero_ps(), \
2440 return (__m512)((__v16sf)__a/(__v16sf)__b); in _mm512_div_ps()
2445 return (__m512) __builtin_ia32_divps512_mask ((__v16sf) __A, in _mm512_mask_div_ps()
2446 (__v16sf) __B, in _mm512_mask_div_ps()
2447 (__v16sf) __W, in _mm512_mask_div_ps()
2454 return (__m512) __builtin_ia32_divps512_mask ((__v16sf) __A, in _mm512_maskz_div_ps()
2455 (__v16sf) __B, in _mm512_maskz_div_ps()
2456 (__v16sf) in _mm512_maskz_div_ps()
2481 (__m512)__builtin_ia32_divps512_mask((__v16sf)(__m512)(A), \
2482 (__v16sf)(__m512)(B), \
2483 (__v16sf)_mm512_setzero_ps(), \
2487 (__m512)__builtin_ia32_divps512_mask((__v16sf)(__m512)(A), \
2488 (__v16sf)(__m512)(B), \
2489 (__v16sf)(__m512)(W), (__mmask16)(U), \
2493 (__m512)__builtin_ia32_divps512_mask((__v16sf)(__m512)(A), \
2494 (__v16sf)(__m512)(B), \
2495 (__v16sf)_mm512_setzero_ps(), \
2499 (__m512)__builtin_ia32_rndscaleps_mask((__v16sf)(__m512)(A), (int)(B), \
2500 (__v16sf)(__m512)(A), (__mmask16)-1, \
2504 (__m512)__builtin_ia32_rndscaleps_mask((__v16sf)(__m512)(C), (int)(imm), \
2505 (__v16sf)(__m512)(A), (__mmask16)(B), \
2509 (__m512)__builtin_ia32_rndscaleps_mask((__v16sf)(__m512)(B), (int)(imm), \
2510 (__v16sf)_mm512_setzero_ps(), \
2515 (__m512)__builtin_ia32_rndscaleps_mask((__v16sf)(__m512)(C), (int)(imm), \
2516 (__v16sf)(__m512)(A), (__mmask16)(B), \
2520 (__m512)__builtin_ia32_rndscaleps_mask((__v16sf)(__m512)(B), (int)(imm), \
2521 (__v16sf)_mm512_setzero_ps(), \
2525 (__m512)__builtin_ia32_rndscaleps_mask((__v16sf)(__m512)(A), (int)(imm), \
2526 (__v16sf)_mm512_undefined_ps(), \
2765 (__m512)__builtin_ia32_vfmaddps512_mask((__v16sf)(__m512)(A), \
2766 (__v16sf)(__m512)(B), \
2767 (__v16sf)(__m512)(C), (__mmask16)-1, \
2772 (__m512)__builtin_ia32_vfmaddps512_mask((__v16sf)(__m512)(A), \
2773 (__v16sf)(__m512)(B), \
2774 (__v16sf)(__m512)(C), \
2779 (__m512)__builtin_ia32_vfmaddps512_mask3((__v16sf)(__m512)(A), \
2780 (__v16sf)(__m512)(B), \
2781 (__v16sf)(__m512)(C), \
2786 (__m512)__builtin_ia32_vfmaddps512_maskz((__v16sf)(__m512)(A), \
2787 (__v16sf)(__m512)(B), \
2788 (__v16sf)(__m512)(C), \
2793 (__m512)__builtin_ia32_vfmaddps512_mask((__v16sf)(__m512)(A), \
2794 (__v16sf)(__m512)(B), \
2795 -(__v16sf)(__m512)(C), \
2800 (__m512)__builtin_ia32_vfmaddps512_mask((__v16sf)(__m512)(A), \
2801 (__v16sf)(__m512)(B), \
2802 -(__v16sf)(__m512)(C), \
2807 (__m512)__builtin_ia32_vfmaddps512_maskz((__v16sf)(__m512)(A), \
2808 (__v16sf)(__m512)(B), \
2809 -(__v16sf)(__m512)(C), \
2814 (__m512)__builtin_ia32_vfmaddps512_mask(-(__v16sf)(__m512)(A), \
2815 (__v16sf)(__m512)(B), \
2816 (__v16sf)(__m512)(C), (__mmask16)-1, \
2821 (__m512)__builtin_ia32_vfmaddps512_mask3(-(__v16sf)(__m512)(A), \
2822 (__v16sf)(__m512)(B), \
2823 (__v16sf)(__m512)(C), \
2828 (__m512)__builtin_ia32_vfmaddps512_maskz(-(__v16sf)(__m512)(A), \
2829 (__v16sf)(__m512)(B), \
2830 (__v16sf)(__m512)(C), \
2835 (__m512)__builtin_ia32_vfmaddps512_mask(-(__v16sf)(__m512)(A), \
2836 (__v16sf)(__m512)(B), \
2837 -(__v16sf)(__m512)(C), \
2842 (__m512)__builtin_ia32_vfmaddps512_maskz(-(__v16sf)(__m512)(A), \
2843 (__v16sf)(__m512)(B), \
2844 -(__v16sf)(__m512)(C), \
2851 return (__m512) __builtin_ia32_vfmaddps512_mask ((__v16sf) __A, in _mm512_fmadd_ps()
2852 (__v16sf) __B, in _mm512_fmadd_ps()
2853 (__v16sf) __C, in _mm512_fmadd_ps()
2861 return (__m512) __builtin_ia32_vfmaddps512_mask ((__v16sf) __A, in _mm512_mask_fmadd_ps()
2862 (__v16sf) __B, in _mm512_mask_fmadd_ps()
2863 (__v16sf) __C, in _mm512_mask_fmadd_ps()
2871 return (__m512) __builtin_ia32_vfmaddps512_mask3 ((__v16sf) __A, in _mm512_mask3_fmadd_ps()
2872 (__v16sf) __B, in _mm512_mask3_fmadd_ps()
2873 (__v16sf) __C, in _mm512_mask3_fmadd_ps()
2881 return (__m512) __builtin_ia32_vfmaddps512_maskz ((__v16sf) __A, in _mm512_maskz_fmadd_ps()
2882 (__v16sf) __B, in _mm512_maskz_fmadd_ps()
2883 (__v16sf) __C, in _mm512_maskz_fmadd_ps()
2891 return (__m512) __builtin_ia32_vfmaddps512_mask ((__v16sf) __A, in _mm512_fmsub_ps()
2892 (__v16sf) __B, in _mm512_fmsub_ps()
2893 -(__v16sf) __C, in _mm512_fmsub_ps()
2901 return (__m512) __builtin_ia32_vfmaddps512_mask ((__v16sf) __A, in _mm512_mask_fmsub_ps()
2902 (__v16sf) __B, in _mm512_mask_fmsub_ps()
2903 -(__v16sf) __C, in _mm512_mask_fmsub_ps()
2911 return (__m512) __builtin_ia32_vfmaddps512_maskz ((__v16sf) __A, in _mm512_maskz_fmsub_ps()
2912 (__v16sf) __B, in _mm512_maskz_fmsub_ps()
2913 -(__v16sf) __C, in _mm512_maskz_fmsub_ps()
2921 return (__m512) __builtin_ia32_vfmaddps512_mask (-(__v16sf) __A, in _mm512_fnmadd_ps()
2922 (__v16sf) __B, in _mm512_fnmadd_ps()
2923 (__v16sf) __C, in _mm512_fnmadd_ps()
2931 return (__m512) __builtin_ia32_vfmaddps512_mask3 (-(__v16sf) __A, in _mm512_mask3_fnmadd_ps()
2932 (__v16sf) __B, in _mm512_mask3_fnmadd_ps()
2933 (__v16sf) __C, in _mm512_mask3_fnmadd_ps()
2941 return (__m512) __builtin_ia32_vfmaddps512_maskz (-(__v16sf) __A, in _mm512_maskz_fnmadd_ps()
2942 (__v16sf) __B, in _mm512_maskz_fnmadd_ps()
2943 (__v16sf) __C, in _mm512_maskz_fnmadd_ps()
2951 return (__m512) __builtin_ia32_vfmaddps512_mask (-(__v16sf) __A, in _mm512_fnmsub_ps()
2952 (__v16sf) __B, in _mm512_fnmsub_ps()
2953 -(__v16sf) __C, in _mm512_fnmsub_ps()
2961 return (__m512) __builtin_ia32_vfmaddps512_maskz (-(__v16sf) __A, in _mm512_maskz_fnmsub_ps()
2962 (__v16sf) __B, in _mm512_maskz_fnmsub_ps()
2963 -(__v16sf) __C, in _mm512_maskz_fnmsub_ps()
3088 (__m512)__builtin_ia32_vfmaddsubps512_mask((__v16sf)(__m512)(A), \
3089 (__v16sf)(__m512)(B), \
3090 (__v16sf)(__m512)(C), \
3095 (__m512)__builtin_ia32_vfmaddsubps512_mask((__v16sf)(__m512)(A), \
3096 (__v16sf)(__m512)(B), \
3097 (__v16sf)(__m512)(C), \
3102 (__m512)__builtin_ia32_vfmaddsubps512_mask3((__v16sf)(__m512)(A), \
3103 (__v16sf)(__m512)(B), \
3104 (__v16sf)(__m512)(C), \
3109 (__m512)__builtin_ia32_vfmaddsubps512_maskz((__v16sf)(__m512)(A), \
3110 (__v16sf)(__m512)(B), \
3111 (__v16sf)(__m512)(C), \
3116 (__m512)__builtin_ia32_vfmaddsubps512_mask((__v16sf)(__m512)(A), \
3117 (__v16sf)(__m512)(B), \
3118 -(__v16sf)(__m512)(C), \
3123 (__m512)__builtin_ia32_vfmaddsubps512_mask((__v16sf)(__m512)(A), \
3124 (__v16sf)(__m512)(B), \
3125 -(__v16sf)(__m512)(C), \
3130 (__m512)__builtin_ia32_vfmaddsubps512_maskz((__v16sf)(__m512)(A), \
3131 (__v16sf)(__m512)(B), \
3132 -(__v16sf)(__m512)(C), \
3139 return (__m512) __builtin_ia32_vfmaddsubps512_mask ((__v16sf) __A, in _mm512_fmaddsub_ps()
3140 (__v16sf) __B, in _mm512_fmaddsub_ps()
3141 (__v16sf) __C, in _mm512_fmaddsub_ps()
3149 return (__m512) __builtin_ia32_vfmaddsubps512_mask ((__v16sf) __A, in _mm512_mask_fmaddsub_ps()
3150 (__v16sf) __B, in _mm512_mask_fmaddsub_ps()
3151 (__v16sf) __C, in _mm512_mask_fmaddsub_ps()
3159 return (__m512) __builtin_ia32_vfmaddsubps512_mask3 ((__v16sf) __A, in _mm512_mask3_fmaddsub_ps()
3160 (__v16sf) __B, in _mm512_mask3_fmaddsub_ps()
3161 (__v16sf) __C, in _mm512_mask3_fmaddsub_ps()
3169 return (__m512) __builtin_ia32_vfmaddsubps512_maskz ((__v16sf) __A, in _mm512_maskz_fmaddsub_ps()
3170 (__v16sf) __B, in _mm512_maskz_fmaddsub_ps()
3171 (__v16sf) __C, in _mm512_maskz_fmaddsub_ps()
3179 return (__m512) __builtin_ia32_vfmaddsubps512_mask ((__v16sf) __A, in _mm512_fmsubadd_ps()
3180 (__v16sf) __B, in _mm512_fmsubadd_ps()
3181 -(__v16sf) __C, in _mm512_fmsubadd_ps()
3189 return (__m512) __builtin_ia32_vfmaddsubps512_mask ((__v16sf) __A, in _mm512_mask_fmsubadd_ps()
3190 (__v16sf) __B, in _mm512_mask_fmsubadd_ps()
3191 -(__v16sf) __C, in _mm512_mask_fmsubadd_ps()
3199 return (__m512) __builtin_ia32_vfmaddsubps512_maskz ((__v16sf) __A, in _mm512_maskz_fmsubadd_ps()
3200 (__v16sf) __B, in _mm512_maskz_fmsubadd_ps()
3201 -(__v16sf) __C, in _mm512_maskz_fmsubadd_ps()
3224 (__m512)__builtin_ia32_vfmsubps512_mask3((__v16sf)(__m512)(A), \
3225 (__v16sf)(__m512)(B), \
3226 (__v16sf)(__m512)(C), \
3233 return (__m512) __builtin_ia32_vfmsubps512_mask3 ((__v16sf) __A, in _mm512_mask3_fmsub_ps()
3234 (__v16sf) __B, in _mm512_mask3_fmsub_ps()
3235 (__v16sf) __C, in _mm512_mask3_fmsub_ps()
3258 (__m512)__builtin_ia32_vfmsubaddps512_mask3((__v16sf)(__m512)(A), \
3259 (__v16sf)(__m512)(B), \
3260 (__v16sf)(__m512)(C), \
3267 return (__m512) __builtin_ia32_vfmsubaddps512_mask3 ((__v16sf) __A, in _mm512_mask3_fmsubadd_ps()
3268 (__v16sf) __B, in _mm512_mask3_fmsubadd_ps()
3269 (__v16sf) __C, in _mm512_mask3_fmsubadd_ps()
3292 (__m512)__builtin_ia32_vfnmaddps512_mask((__v16sf)(__m512)(A), \
3293 (__v16sf)(__m512)(B), \
3294 (__v16sf)(__m512)(C), \
3301 return (__m512) __builtin_ia32_vfnmaddps512_mask ((__v16sf) __A, in _mm512_mask_fnmadd_ps()
3302 (__v16sf) __B, in _mm512_mask_fnmadd_ps()
3303 (__v16sf) __C, in _mm512_mask_fnmadd_ps()
3343 (__m512)__builtin_ia32_vfnmsubps512_mask((__v16sf)(__m512)(A), \
3344 (__v16sf)(__m512)(B), \
3345 (__v16sf)(__m512)(C), \
3350 (__m512)__builtin_ia32_vfnmsubps512_mask3((__v16sf)(__m512)(A), \
3351 (__v16sf)(__m512)(B), \
3352 (__v16sf)(__m512)(C), \
3359 return (__m512) __builtin_ia32_vfnmsubps512_mask ((__v16sf) __A, in _mm512_mask_fnmsub_ps()
3360 (__v16sf) __B, in _mm512_mask_fnmsub_ps()
3361 (__v16sf) __C, in _mm512_mask_fnmsub_ps()
3369 return (__m512) __builtin_ia32_vfnmsubps512_mask3 ((__v16sf) __A, in _mm512_mask3_fnmsub_ps()
3370 (__v16sf) __B, in _mm512_mask3_fnmsub_ps()
3371 (__v16sf) __C, in _mm512_mask3_fnmsub_ps()
3498 (__m128)__builtin_ia32_extractf32x4_mask((__v16sf)(__m512)(A), (int)(I), \
3503 (__m128)__builtin_ia32_extractf32x4_mask((__v16sf)(__m512)(A), (int)(imm), \
3508 (__m128)__builtin_ia32_extractf32x4_mask((__v16sf)(__m512)(A), (int)(imm), \
3525 (__v16sf) __W, in _mm512_mask_blend_ps()
3526 (__v16sf) __A); in _mm512_mask_blend_ps()
3548 (__mmask16)__builtin_ia32_cmpps512_mask((__v16sf)(__m512)(A), \
3549 (__v16sf)(__m512)(B), (int)(P), \
3553 (__mmask16)__builtin_ia32_cmpps512_mask((__v16sf)(__m512)(A), \
3554 (__v16sf)(__m512)(B), (int)(P), \
3582 (__m512i)__builtin_ia32_cvttps2udq512_mask((__v16sf)(__m512)(A), \
3587 (__m512i)__builtin_ia32_cvttps2udq512_mask((__v16sf)(__m512)(A), \
3592 (__m512i)__builtin_ia32_cvttps2udq512_mask((__v16sf)(__m512)(A), \
3600 return (__m512i) __builtin_ia32_cvttps2udq512_mask ((__v16sf) __A, in _mm512_cvttps_epu32()
3610 return (__m512i) __builtin_ia32_cvttps2udq512_mask ((__v16sf) __A, in _mm512_mask_cvttps_epu32()
3619 return (__m512i) __builtin_ia32_cvttps2udq512_mask ((__v16sf) __A, in _mm512_maskz_cvttps_epu32()
3627 (__v16sf)_mm512_setzero_ps(), \
3632 (__v16sf)(__m512)(W), \
3637 (__v16sf)_mm512_setzero_ps(), \
3642 (__v16sf)_mm512_setzero_ps(), \
3647 (__v16sf)(__m512)(W), \
3652 (__v16sf)_mm512_setzero_ps(), \
3659 (__v16sf) _mm512_undefined_ps (), in _mm512_cvtepu32_ps()
3668 (__v16sf) __W, in _mm512_mask_cvtepu32_ps()
3677 (__v16sf) _mm512_setzero_ps (), in _mm512_maskz_cvtepu32_ps()
3711 (__v16sf) _mm512_undefined_ps (), in _mm512_cvtepi32_ps()
3720 (__v16sf) __W, in _mm512_mask_cvtepi32_ps()
3729 (__v16sf) _mm512_setzero_ps (), in _mm512_maskz_cvtepi32_ps()
3802 (__m256i)__builtin_ia32_vcvtps2ph512_mask((__v16sf)(__m512)(A), (int)(I), \
3807 (__m256i)__builtin_ia32_vcvtps2ph512_mask((__v16sf)(__m512)(A), (int)(I), \
3812 (__m256i)__builtin_ia32_vcvtps2ph512_mask((__v16sf)(__m512)(A), (int)(I), \
3817 (__m256i)__builtin_ia32_vcvtps2ph512_mask((__v16sf)(__m512)(A), (int)(I), \
3822 (__m256i)__builtin_ia32_vcvtps2ph512_mask((__v16sf)(__m512)(A), (int)(I), \
3827 (__m256i)__builtin_ia32_vcvtps2ph512_mask((__v16sf)(__m512)(A), (int)(I), \
3833 (__v16sf)_mm512_undefined_ps(), \
3838 (__v16sf)(__m512)(W), \
3843 (__v16sf)_mm512_setzero_ps(), \
3851 (__v16sf) in _mm512_cvtph_ps()
3861 (__v16sf) __W, in _mm512_mask_cvtph_ps()
3870 (__v16sf) _mm512_setzero_ps (), in _mm512_maskz_cvtph_ps()
3918 (__m512i)__builtin_ia32_cvttps2dq512_mask((__v16sf)(__m512)(A), \
3923 (__m512i)__builtin_ia32_cvttps2dq512_mask((__v16sf)(__m512)(A), \
3928 (__m512i)__builtin_ia32_cvttps2dq512_mask((__v16sf)(__m512)(A), \
3936 __builtin_ia32_cvttps2dq512_mask((__v16sf) __a, in _mm512_cvttps_epi32()
3944 return (__m512i) __builtin_ia32_cvttps2dq512_mask ((__v16sf) __A, in _mm512_mask_cvttps_epi32()
3953 return (__m512i) __builtin_ia32_cvttps2dq512_mask ((__v16sf) __A, in _mm512_maskz_cvttps_epi32()
3960 (__m512i)__builtin_ia32_cvtps2dq512_mask((__v16sf)(__m512)(A), \
3965 (__m512i)__builtin_ia32_cvtps2dq512_mask((__v16sf)(__m512)(A), \
3970 (__m512i)__builtin_ia32_cvtps2dq512_mask((__v16sf)(__m512)(A), \
3977 return (__m512i) __builtin_ia32_cvtps2dq512_mask ((__v16sf) __A, in _mm512_cvtps_epi32()
3986 return (__m512i) __builtin_ia32_cvtps2dq512_mask ((__v16sf) __A, in _mm512_mask_cvtps_epi32()
3995 return (__m512i) __builtin_ia32_cvtps2dq512_mask ((__v16sf) __A, in _mm512_maskz_cvtps_epi32()
4047 (__m512i)__builtin_ia32_cvtps2udq512_mask((__v16sf)(__m512)(A), \
4052 (__m512i)__builtin_ia32_cvtps2udq512_mask((__v16sf)(__m512)(A), \
4057 (__m512i)__builtin_ia32_cvtps2udq512_mask((__v16sf)(__m512)(A), \
4064 return (__m512i) __builtin_ia32_cvtps2udq512_mask ((__v16sf) __A,\ in _mm512_cvtps_epu32()
4074 return (__m512i) __builtin_ia32_cvtps2udq512_mask ((__v16sf) __A, in _mm512_mask_cvtps_epu32()
4083 return (__m512i) __builtin_ia32_cvtps2udq512_mask ((__v16sf) __A, in _mm512_maskz_cvtps_epu32()
4185 return (__m512)__builtin_shufflevector((__v16sf)__a, (__v16sf)__b, in _mm512_unpackhi_ps()
4196 (__v16sf)_mm512_unpackhi_ps(__A, __B), in _mm512_mask_unpackhi_ps()
4197 (__v16sf)__W); in _mm512_mask_unpackhi_ps()
4204 (__v16sf)_mm512_unpackhi_ps(__A, __B), in _mm512_maskz_unpackhi_ps()
4205 (__v16sf)_mm512_setzero_ps()); in _mm512_maskz_unpackhi_ps()
4211 return (__m512)__builtin_shufflevector((__v16sf)__a, (__v16sf)__b, in _mm512_unpacklo_ps()
4222 (__v16sf)_mm512_unpacklo_ps(__A, __B), in _mm512_mask_unpacklo_ps()
4223 (__v16sf)__W); in _mm512_mask_unpacklo_ps()
4230 (__v16sf)_mm512_unpacklo_ps(__A, __B), in _mm512_maskz_unpacklo_ps()
4231 (__v16sf)_mm512_setzero_ps()); in _mm512_maskz_unpacklo_ps()
4414 (__v16sf) __W, in _mm512_mask_loadu_ps()
4422 (__v16sf) in _mm512_maskz_loadu_ps()
4465 return (__m512) __builtin_ia32_loadaps512_mask ((const __v16sf *)__p, in _mm512_load_ps()
4466 (__v16sf) in _mm512_load_ps()
4474 return (__m512) __builtin_ia32_loadaps512_mask ((const __v16sf *) __P, in _mm512_mask_load_ps()
4475 (__v16sf) __W, in _mm512_mask_load_ps()
4482 return (__m512) __builtin_ia32_loadaps512_mask ((const __v16sf *)__P, in _mm512_maskz_load_ps()
4483 (__v16sf) in _mm512_maskz_load_ps()
4570 __builtin_ia32_storeups512_mask ((float *)__P, (__v16sf) __A, in _mm512_mask_storeu_ps()
4577 __builtin_ia32_storeups512_mask((float *)__P, (__v16sf)__A, (__mmask16)-1); in _mm512_storeu_ps()
4595 __builtin_ia32_storeaps512_mask ((__v16sf *)__P, (__v16sf) __A, in _mm512_mask_store_ps()
5602 (__m512)__builtin_ia32_fixupimmps512_mask((__v16sf)(__m512)(A), \
5603 (__v16sf)(__m512)(B), \
5608 (__m512)__builtin_ia32_fixupimmps512_mask((__v16sf)(__m512)(A), \
5609 (__v16sf)(__m512)(B), \
5614 (__m512)__builtin_ia32_fixupimmps512_mask((__v16sf)(__m512)(A), \
5615 (__v16sf)(__m512)(B), \
5621 (__m512)__builtin_ia32_fixupimmps512_mask((__v16sf)(__m512)(A), \
5622 (__v16sf)(__m512)(B), \
5628 (__m512)__builtin_ia32_fixupimmps512_maskz((__v16sf)(__m512)(A), \
5629 (__v16sf)(__m512)(B), \
5635 (__m512)__builtin_ia32_fixupimmps512_maskz((__v16sf)(__m512)(A), \
5636 (__v16sf)(__m512)(B), \
6488 return (__m512) __builtin_ia32_vpermi2varps512_mask ((__v16sf) __A, in _mm512_mask2_permutex2var_ps()
6491 (__v16sf) __B, in _mm512_mask2_permutex2var_ps()
6529 (__m512)__builtin_shufflevector((__v16sf)(__m512)(X), \
6530 (__v16sf)_mm512_undefined_ps(), \
6550 (__v16sf)_mm512_permute_ps((X), (C)), \
6551 (__v16sf)(__m512)(W)); })
6555 (__v16sf)_mm512_permute_ps((X), (C)), \
6556 (__v16sf)_mm512_setzero_ps()); })
6590 return (__m512) __builtin_ia32_vpermilvarps512_mask ((__v16sf) __A, in _mm512_permutevar_ps()
6592 (__v16sf) in _mm512_permutevar_ps()
6600 return (__m512) __builtin_ia32_vpermilvarps512_mask ((__v16sf) __A, in _mm512_mask_permutevar_ps()
6602 (__v16sf) __W, in _mm512_mask_permutevar_ps()
6609 return (__m512) __builtin_ia32_vpermilvarps512_mask ((__v16sf) __A, in _mm512_maskz_permutevar_ps()
6611 (__v16sf) in _mm512_maskz_permutevar_ps()
6652 (__v16sf) __A, in _mm512_permutex2var_ps()
6653 (__v16sf) __B, in _mm512_permutex2var_ps()
6662 (__v16sf) __A, in _mm512_mask_permutex2var_ps()
6663 (__v16sf) __B, in _mm512_mask_permutex2var_ps()
6673 (__v16sf) __A, in _mm512_maskz_permutex2var_ps()
6674 (__v16sf) __B, in _mm512_maskz_permutex2var_ps()
6887 (__m512)__builtin_ia32_scalefps512_mask((__v16sf)(__m512)(A), \
6888 (__v16sf)(__m512)(B), \
6889 (__v16sf)_mm512_undefined_ps(), \
6893 (__m512)__builtin_ia32_scalefps512_mask((__v16sf)(__m512)(A), \
6894 (__v16sf)(__m512)(B), \
6895 (__v16sf)(__m512)(W), \
6899 (__m512)__builtin_ia32_scalefps512_mask((__v16sf)(__m512)(A), \
6900 (__v16sf)(__m512)(B), \
6901 (__v16sf)_mm512_setzero_ps(), \
6907 return (__m512) __builtin_ia32_scalefps512_mask ((__v16sf) __A, in _mm512_scalef_ps()
6908 (__v16sf) __B, in _mm512_scalef_ps()
6909 (__v16sf) in _mm512_scalef_ps()
6918 return (__m512) __builtin_ia32_scalefps512_mask ((__v16sf) __A, in _mm512_mask_scalef_ps()
6919 (__v16sf) __B, in _mm512_mask_scalef_ps()
6920 (__v16sf) __W, in _mm512_mask_scalef_ps()
6928 return (__m512) __builtin_ia32_scalefps512_mask ((__v16sf) __A, in _mm512_maskz_scalef_ps()
6929 (__v16sf) __B, in _mm512_maskz_scalef_ps()
6930 (__v16sf) in _mm512_maskz_scalef_ps()
7062 (__m512)__builtin_ia32_shuf_f32x4_mask((__v16sf)(__m512)(A), \
7063 (__v16sf)(__m512)(B), (int)(imm), \
7064 (__v16sf)_mm512_undefined_ps(), \
7068 (__m512)__builtin_ia32_shuf_f32x4_mask((__v16sf)(__m512)(A), \
7069 (__v16sf)(__m512)(B), (int)(imm), \
7070 (__v16sf)(__m512)(W), \
7074 (__m512)__builtin_ia32_shuf_f32x4_mask((__v16sf)(__m512)(A), \
7075 (__v16sf)(__m512)(B), (int)(imm), \
7076 (__v16sf)_mm512_setzero_ps(), \
7156 (__m512d)__builtin_shufflevector((__v16sf)(__m512)(A), \
7157 (__v16sf)(__m512)(B), \
7177 (__v16sf)_mm512_shuffle_ps((A), (B), (M)), \
7178 (__v16sf)(__m512)(W)); })
7182 (__v16sf)_mm512_shuffle_ps((A), (B), (M)), \
7183 (__v16sf)_mm512_setzero_ps()); })
7265 (__v16sf) in _mm512_broadcast_f32x4()
7274 (__v16sf) __O, in _mm512_mask_broadcast_f32x4()
7282 (__v16sf) in _mm512_maskz_broadcast_f32x4()
7385 (__v16sf) _mm512_broadcastss_ps(__A), in _mm512_mask_broadcastss_ps()
7386 (__v16sf) __O); in _mm512_mask_broadcastss_ps()
7393 (__v16sf) _mm512_broadcastss_ps(__A), in _mm512_maskz_broadcastss_ps()
7394 (__v16sf) _mm512_setzero_ps()); in _mm512_maskz_broadcastss_ps()
7903 (__m512)__builtin_ia32_insertf32x4_mask((__v16sf)(__m512)(A), \
7905 (__v16sf)_mm512_undefined_ps(), \
7909 (__m512)__builtin_ia32_insertf32x4_mask((__v16sf)(__m512)(A), \
7911 (__v16sf)(__m512)(W), \
7915 (__m512)__builtin_ia32_insertf32x4_mask((__v16sf)(__m512)(A), \
7917 (__v16sf)_mm512_setzero_ps(), \
7978 (__m512)__builtin_ia32_getmantps512_mask((__v16sf)(__m512)(A), \
7980 (__v16sf)_mm512_undefined_ps(), \
7984 (__m512)__builtin_ia32_getmantps512_mask((__v16sf)(__m512)(A), \
7986 (__v16sf)(__m512)(W), \
7990 (__m512)__builtin_ia32_getmantps512_mask((__v16sf)(__m512)(A), \
7992 (__v16sf)_mm512_setzero_ps(), \
7996 (__m512)__builtin_ia32_getmantps512_mask((__v16sf)(__m512)(A), \
7998 (__v16sf)_mm512_undefined_ps(), \
8003 (__m512)__builtin_ia32_getmantps512_mask((__v16sf)(__m512)(A), \
8005 (__v16sf)(__m512)(W), \
8010 (__m512)__builtin_ia32_getmantps512_mask((__v16sf)(__m512)(A), \
8012 (__v16sf)_mm512_setzero_ps(), \
8059 (__m512)__builtin_ia32_getexpps512_mask((__v16sf)(__m512)(A), \
8060 (__v16sf)_mm512_undefined_ps(), \
8064 (__m512)__builtin_ia32_getexpps512_mask((__v16sf)(__m512)(A), \
8065 (__v16sf)(__m512)(W), \
8069 (__m512)__builtin_ia32_getexpps512_mask((__v16sf)(__m512)(A), \
8070 (__v16sf)_mm512_setzero_ps(), \
8076 return (__m512) __builtin_ia32_getexpps512_mask ((__v16sf) __A, in _mm512_getexp_ps()
8077 (__v16sf) _mm512_undefined_ps (), in _mm512_getexp_ps()
8085 return (__m512) __builtin_ia32_getexpps512_mask ((__v16sf) __A, in _mm512_mask_getexp_ps()
8086 (__v16sf) __W, in _mm512_mask_getexp_ps()
8094 return (__m512) __builtin_ia32_getexpps512_mask ((__v16sf) __A, in _mm512_maskz_getexp_ps()
8095 (__v16sf) _mm512_setzero_ps (), in _mm512_maskz_getexp_ps()
8149 (__m512)__builtin_ia32_gathersiv16sf((__v16sf)_mm512_undefined_ps(), \
8151 (__v16sf)(__m512)(index), \
8155 (__m512)__builtin_ia32_gathersiv16sf((__v16sf)(__m512)(v1_old), \
8157 (__v16sf)(__m512)(index), \
8239 (__v16sf)(__m512)(v1), (int)(scale)); })
8244 (__v16sf)(__m512)(v1), (int)(scale)); })
8763 return (__m512) __builtin_ia32_permvarsf512_mask ((__v16sf) __Y, in _mm512_permutexvar_ps()
8765 (__v16sf) _mm512_undefined_ps (), in _mm512_permutexvar_ps()
8772 return (__m512) __builtin_ia32_permvarsf512_mask ((__v16sf) __Y, in _mm512_mask_permutexvar_ps()
8774 (__v16sf) __W, in _mm512_mask_permutexvar_ps()
8781 return (__m512) __builtin_ia32_permvarsf512_mask ((__v16sf) __Y, in _mm512_maskz_permutexvar_ps()
8783 (__v16sf) _mm512_setzero_ps (), in _mm512_maskz_permutexvar_ps()
8884 __builtin_nontemporal_store((__v16sf)__A, (__v16sf*)__P); in _mm512_stream_ps()
8924 return (__m512) __builtin_ia32_compresssf512_mask ((__v16sf) __A, in _mm512_mask_compress_ps()
8925 (__v16sf) __W, in _mm512_mask_compress_ps()
8932 return (__m512) __builtin_ia32_compresssf512_mask ((__v16sf) __A, in _mm512_maskz_compress_ps()
8933 (__v16sf) in _mm512_maskz_compress_ps()
9002 return (__m512)__builtin_shufflevector((__v16sf)__A, (__v16sf)__A, in _mm512_movehdup_ps()
9010 (__v16sf)_mm512_movehdup_ps(__A), in _mm512_mask_movehdup_ps()
9011 (__v16sf)__W); in _mm512_mask_movehdup_ps()
9018 (__v16sf)_mm512_movehdup_ps(__A), in _mm512_maskz_movehdup_ps()
9019 (__v16sf)_mm512_setzero_ps()); in _mm512_maskz_movehdup_ps()
9025 return (__m512)__builtin_shufflevector((__v16sf)__A, (__v16sf)__A, in _mm512_moveldup_ps()
9033 (__v16sf)_mm512_moveldup_ps(__A), in _mm512_mask_moveldup_ps()
9034 (__v16sf)__W); in _mm512_mask_moveldup_ps()
9041 (__v16sf)_mm512_moveldup_ps(__A), in _mm512_maskz_moveldup_ps()
9042 (__v16sf)_mm512_setzero_ps()); in _mm512_maskz_moveldup_ps()
9142 return (__m512) __builtin_ia32_expandloadsf512_mask ((const __v16sf *)__P, in _mm512_mask_expandloadu_ps()
9143 (__v16sf) __W, in _mm512_mask_expandloadu_ps()
9150 return (__m512) __builtin_ia32_expandloadsf512_mask ((const __v16sf *)__P, in _mm512_maskz_expandloadu_ps()
9151 (__v16sf) _mm512_setzero_ps(), in _mm512_maskz_expandloadu_ps()
9174 return (__m512) __builtin_ia32_expandsf512_mask ((__v16sf) __A, in _mm512_mask_expand_ps()
9175 (__v16sf) __W, in _mm512_mask_expand_ps()
9182 return (__m512) __builtin_ia32_expandsf512_mask ((__v16sf) __A, in _mm512_maskz_expand_ps()
9183 (__v16sf) _mm512_setzero_ps(), in _mm512_maskz_expand_ps()
9267 (__v16sf) __A, in _mm512_mask_mov_ps()
9268 (__v16sf) __W); in _mm512_mask_mov_ps()
9275 (__v16sf) __A, in _mm512_maskz_mov_ps()
9276 (__v16sf) _mm512_setzero_ps ()); in _mm512_maskz_mov_ps()
9296 __builtin_ia32_compressstoresf512_mask ((__v16sf *) __P, (__v16sf) __A, in _mm512_mask_compressstoreu_ps()