Searched refs:ARCH_X86_64 (Results 1 – 25 of 88) sorted by relevance
1234
46 #if ARCH_X86_6462 #if ARCH_X86_6482 #if ARCH_X86_64256 #if ARCH_X86_64 in x86_readtsc()274 #if ARCH_X86_64 in x86_readtsc64()296 #if ARCH_X86_64 in x86_readtscp()322 #if ARCH_X86_64347 #elif ARCH_X86_64
23 ifeq ($(ARCH_X86_64),yes)28 ifeq ($(ARCH_X86_64),yes)32 ifeq ($(ARCH_X86)$(ARCH_X86_64),yes)
20 #if (ARCH_X86 || ARCH_X86_64) && HAVE_MMX
47 #if ARCH_X86_6463 #if ARCH_X86_6483 #if ARCH_X86_64255 #if ARCH_X86_64 in x86_readtsc()273 #if ARCH_X86_64 in x86_readtsc64()295 #if ARCH_X86_64 in x86_readtscp()321 #if ARCH_X86_64346 #elif ARCH_X86_64
17 #if ARCH_X86 || ARCH_X86_64
15 #if ARCH_X86 || ARCH_X86_6429 #if ARCH_X86 || ARCH_X86_6442 #if ARCH_X86 || ARCH_X86_64 in main()
31 #if defined(_WIN64) && ARCH_X86_64143 #if ARCH_X86 || ARCH_X86_64
20 #if ARCH_X86 || ARCH_X86_6429 #if ARCH_X86 || ARCH_X86_6452 #if ARCH_X86 || ARCH_X86_64 in main()
16 #if ARCH_X86 || ARCH_X86_6425 #if ARCH_X86 || ARCH_X86_64 in ClearSystemState()
34 #if defined(_WIN64) && ARCH_X86_64104 #if ARCH_X86 || ARCH_X86_64
25 #if HAVE_SSE2 && ARCH_X86_6471 #if ARCH_X86_64 in vp8_loop_filter_bh_sse2()104 #if ARCH_X86_64 in vp8_loop_filter_bv_sse2()
15 #elif ARCH_X86 || ARCH_X86_6497 #elif ARCH_X86 || ARCH_X86_64 in vp8_machine_specific_config()
117 #if ARCH_X86_64 in vp9_diamond_search_sad_avx()141 #if ARCH_X86_64 in vp9_diamond_search_sad_avx()178 #if ARCH_X86_64 // sizeof(intptr_t) == 8 in vp9_diamond_search_sad_avx()297 #if ARCH_X86_64 in vp9_diamond_search_sad_avx()
61 %if ARCH_X86_64108 %if ARCH_X86_64
48 %if ARCH_X86_6457 %if ARCH_X86_64168 %if ARCH_X86_64 == 0263 %elif ARCH_X86_64 ; memory284 %if ARCH_X86_64 == 0321 %if ARCH_X86_64610 %elif ARCH_X86_64 ; *nix x64 ;=============================================951 %if ARCH_X86_64 || cpuflag(sse2)1008 %if ARCH_X86_641029 %if ARCH_X86_64
33 ; TODO(slavarnway): using xmm registers for these on ARCH_X86_64 +55 %if ARCH_X86_6475 %if ARCH_X86_6486 %if ARCH_X86_64386 %if ARCH_X86_64 && X86_SUBPIX_VFILTER_PREFER_SLOW_CELERON408 %if ARCH_X86_64531 ; ARCH_X86_64631 %endif ; ARCH_X86_64647 %if ARCH_X86_64727 ; ARCH_X86_64[all …]
97 %if ARCH_X86_64274 %if ARCH_X86_64278 %if ARCH_X86_64 && mmsize == 16501 %if ARCH_X86_64505 %if ARCH_X86_64 && mmsize == 16623 %if ARCH_X86_64627 %if ARCH_X86_64 && mmsize == 16722 %if ARCH_X86_64726 %if ARCH_X86_64 && mmsize == 16849 %if ARCH_X86_64[all …]
94 %if ARCH_X86_64271 %if ARCH_X86_64275 %if ARCH_X86_64 && mmsize == 16498 %if ARCH_X86_64502 %if ARCH_X86_64 && mmsize == 16620 %if ARCH_X86_64624 %if ARCH_X86_64 && mmsize == 16719 %if ARCH_X86_64723 %if ARCH_X86_64 && mmsize == 16846 %if ARCH_X86_64[all …]
29 ; TODO(slavarnway): using xmm registers for these on ARCH_X86_64 +51 %if ARCH_X86_6471 %if ARCH_X86_6482 %if ARCH_X86_64342 %if ARCH_X86_64 && X86_SUBPIX_VFILTER_PREFER_SLOW_CELERON364 %if ARCH_X86_64470 ; ARCH_X86_64570 %endif ; ARCH_X86_64586 %if ARCH_X86_64657 ; ARCH_X86_64[all …]
67 %if ARCH_X86_64200 %if ARCH_X86_64280 %if ARCH_X86_64361 %if ARCH_X86_64442 %if ARCH_X86_64
73 VP8_COMMON_SRCS-$(ARCH_X86)$(ARCH_X86_64) += common/x86/vp8_asm_stubs.c74 VP8_COMMON_SRCS-$(ARCH_X86)$(ARCH_X86_64) += common/x86/loopfilter_x86.c98 ifeq ($(ARCH_X86_64),yes)
43 %if ARCH_X86_6486 %elif ARCH_X86_64 == 0142 %elif ARCH_X86_64 ; memory163 %if ARCH_X86_64 == 0200 %if ARCH_X86_64
91 DSP_SRCS-$(ARCH_X86)$(ARCH_X86_64) += x86/convolve.h170 DSP_SRCS-$(ARCH_X86)$(ARCH_X86_64) += x86/loopfilter_sse2.c209 ifeq ($(ARCH_X86_64),yes)316 ifeq ($(ARCH_X86_64),yes)387 ifeq ($(ARCH_X86_64),yes)
5 %define ARCH_X86_64 0