/external/llvm/test/CodeGen/X86/ |
D | fma-scalar-memfold.ll | 17 ; CHECK: vmovss (%rcx), %[[XMM:xmm[0-9]+]] 19 ; CHECK-NEXT: vmovss %[[XMM]], (%rcx) 42 ; CHECK: vmovss (%rcx), %[[XMM:xmm[0-9]+]] 44 ; CHECK-NEXT: vmovss %[[XMM]], (%rcx) 67 ; CHECK: vmovss (%rcx), %[[XMM:xmm[0-9]+]] 69 ; CHECK-NEXT: vmovss %[[XMM]], (%rcx) 92 ; CHECK: vmovss (%rcx), %[[XMM:xmm[0-9]+]] 94 ; CHECK-NEXT: vmovss %[[XMM]], (%rcx) 117 ; CHECK: vmovss (%rcx), %[[XMM:xmm[0-9]+]] 119 ; CHECK-NEXT: vmovss %[[XMM]], (%rcx) [all …]
|
D | function-subtarget-features.ll | 5 ; In this case avx has a vmovss instruction and otherwise we should be using movss 15 ; CHECK: vmovss 39 ; CHECK: vmovss
|
D | lower-vec-shift.ll | 21 ; AVX-NEXT: vmovss 24 ; AVX2-NEXT: vmovss 55 ; AVX-NEXT: vmovss 107 ; AVX-NEXT: vmovss
|
D | f16c-intrinsics-fast-isel.ll | 14 ; X32-NEXT: vmovss %xmm0, (%esp) 41 ; X32-NEXT: vmovss {{.*#+}} xmm0 = mem[0],zero,zero,zero 43 ; X32-NEXT: vmovss {{.*#+}} xmm0 = xmm0[0],xmm1[1,2,3] 52 ; X64-NEXT: vmovss {{.*#+}} xmm0 = xmm0[0],xmm1[1,2,3]
|
D | avx-load-store.ll | 28 ;; + ins_subvec+ zext into only a single vmovss or vmovsd or vinsertps from memory 33 ; CHECK: vmovss (% 94 ; CHECK-O0: vmovss LCPI 96 ; CHECK-O0: vmovss %xmm
|
D | avx512-build-vector.ll | 19 ; CHECK-NEXT: vmovss {{.*#+}} xmm0 = xmm0[0],xmm2[1,2,3] 20 ; CHECK-NEXT: vmovss {{.*#+}} xmm1 = xmm1[0],xmm2[1,2,3]
|
D | load-slice.ll | 21 ; STRESS: vmovss 64([[BASE:[^(]+]]), [[OUT_Real:%xmm[0-9]+]] 25 ; STRESS-NEXT: vmovss 68([[BASE]]), [[OUT_Imm:%xmm[0-9]+]] 36 ; REGULAR: vmovss 64([[BASE:[^)]+]]), [[OUT_Real:%xmm[0-9]+]] 40 ; REGULAR-NEXT: vmovss 68([[BASE]]), [[OUT_Imm:%xmm[0-9]+]]
|
D | fold-load-unops.ll | 16 ; AVX-NEXT: vmovss (%rdi), %xmm0 35 ; AVX-NEXT: vmovss (%rdi), %xmm0 54 ; AVX-NEXT: vmovss (%rdi), %xmm0
|
D | fma-do-not-commute.ll | 12 ; CHECK: vmovss (%rsi), [[ADDR2:%xmm[0-9]+]]
|
D | fastmath-float-half-conversion.ll | 21 ; F16C-NEXT: vmovss
|
D | fast-isel-fptrunc-fpext.ll | 46 ; AVX: vmovss (%rdi), %xmm0
|
D | logical-load-fold.ll | 44 ; AVX-NEXT: vmovss {{.*#+}} xmm1 = mem[0],zero,zero,zero
|
D | vector-shuffle-variable-256.ll | 223 ; AVX1-NEXT: vmovss {{.*#+}} xmm0 = mem[0],zero,zero,zero 224 ; AVX1-NEXT: vmovss {{.*#+}} xmm1 = mem[0],zero,zero,zero 225 ; AVX1-NEXT: vmovss {{.*#+}} xmm2 = mem[0],zero,zero,zero 229 ; AVX1-NEXT: vmovss {{.*#+}} xmm3 = mem[0],zero,zero,zero 295 ; ALL-NEXT: vmovss {{.*#+}} xmm0 = mem[0],zero,zero,zero 296 ; ALL-NEXT: vmovss {{.*#+}} xmm1 = mem[0],zero,zero,zero 297 ; ALL-NEXT: vmovss {{.*#+}} xmm2 = mem[0],zero,zero,zero 301 ; ALL-NEXT: vmovss {{.*#+}} xmm3 = mem[0],zero,zero,zero
|
D | avx512vl-vbroadcast.ll | 12 ; CHECK-NEXT: vmovss %xmm0, {{[0-9]+}}(%rsp) # 4-byte Folded Spill 31 ; CHECK-NEXT: vmovss %xmm0, {{[0-9]+}}(%rsp) # 4-byte Folded Spill
|
D | fp-trunc.ll | 22 ; AVX-NEXT: vmovss %xmm0, (%esp)
|
D | fp-load-trunc.ll | 22 ; AVX-NEXT: vmovss %xmm0, (%esp)
|
D | avx512-unsafe-fp-math.ll | 81 ; CHECK-NEXT: vmovss {{.*#+}} xmm1 = mem[0],zero,zero,zero
|
D | avx-basic.ll | 130 ; CHECK-NEXT: vmovss {{.*#+}} xmm0 = mem[0],zero,zero,zero
|
D | merge-consecutive-loads-128.ll | 100 ; AVX-NEXT: vmovss {{.*#+}} xmm0 = mem[0],zero,zero,zero 717 ; AVX-NEXT: vmovss {{.*#+}} xmm0 = mem[0],zero,zero,zero 761 ; AVX-NEXT: vmovss {{.*#+}} xmm0 = mem[0],zero,zero,zero 762 ; AVX-NEXT: vmovss {{.*#+}} xmm1 = mem[0],zero,zero,zero
|
D | avx-intrinsics-fast-isel.ll | 2271 ; X32-NEXT: vmovss {{.*#+}} xmm0 = mem[0],zero,zero,zero 2272 ; X32-NEXT: vmovss {{.*#+}} xmm1 = mem[0],zero,zero,zero 2273 ; X32-NEXT: vmovss {{.*#+}} xmm2 = mem[0],zero,zero,zero 2274 ; X32-NEXT: vmovss {{.*#+}} xmm3 = mem[0],zero,zero,zero 2275 ; X32-NEXT: vmovss {{.*#+}} xmm4 = mem[0],zero,zero,zero 2276 ; X32-NEXT: vmovss {{.*#+}} xmm5 = mem[0],zero,zero,zero 2277 ; X32-NEXT: vmovss {{.*#+}} xmm6 = mem[0],zero,zero,zero 2278 ; X32-NEXT: vmovss {{.*#+}} xmm7 = mem[0],zero,zero,zero 2474 ; X32-NEXT: vmovss {{.*#+}} xmm0 = mem[0],zero,zero,zero 2908 ; X32-NEXT: vmovss {{.*#+}} xmm0 = mem[0],zero,zero,zero [all …]
|
D | avx512-select.ll | 42 ; CHECK-NEXT: vmovss %xmm2, %xmm0, %xmm1 {%k1}
|
D | avx512-cmp.ll | 77 ; ALL-NEXT: vmovss {{.*#+}} xmm0 = mem[0],zero,zero,zero
|
D | vec_fptrunc.ll | 23 ; X32-AVX-NEXT: vmovss %xmm0, (%eax)
|
/external/swiftshader/third_party/LLVM/test/CodeGen/X86/ |
D | avx-load-store.ll | 28 ;; + ins_subvec+ zext into only a single vmovss or vmovsd 30 ; CHECK: vmovss (% 85 ; CHECK-O0: vmovss LCPI 87 ; CHECK-O0: vmovss %xmm
|
/external/llvm/test/MC/X86/ |
D | x86_64-avx-encoding.s | 261 vmovss -4(%rbx,%rcx,8), %xmm10 265 vmovss %xmm14, %xmm10, %xmm15 4227 vmovss %xmm0, %xmm0, %xmm8 4231 vmovss %xmm0, %xmm8, %xmm0 4235 vmovss %xmm8, %xmm0, %xmm0
|