/art/runtime/arch/x86/ |
D | memcmp16_x86.S | 143 movdqa (%esi), %xmm0 147 movdqa 16(%esi), %xmm2 153 movdqa %xmm0, %xmm1 154 movdqa 32(%esi), %xmm0 155 movdqa 48(%esi), %xmm2 173 movdqa %xmm0, %xmm1 193 movdqa 16(%esi), %xmm1 194 movdqa %xmm1, %xmm2 198 movdqa 32(%esi), %xmm3 219 movdqa 16(%esi), %xmm0 [all …]
|
/art/runtime/arch/x86_64/ |
D | memcmp16_x86_64.S | 319 movdqa %xmm2, %xmm1 353 movdqa %xmm2, %xmm1 389 movdqa (%rdi), %xmm2 394 movdqa 16(%rdi), %xmm2 399 movdqa 32(%rdi), %xmm2 404 movdqa 48(%rdi), %xmm2 411 movdqa 64(%rdi), %xmm2 416 movdqa 80(%rdi), %xmm2 439 movdqa (%rdi), %xmm2 444 movdqa 16(%rdi), %xmm2 [all …]
|
/art/compiler/utils/x86/ |
D | assembler_x86_test.cc | 711 DriverStr(RepeatFF(&x86::X86Assembler::movdqa, "movdqa %{reg2}, %{reg1}"), "movdqa"); in TEST_F() 719 DriverStr(RepeatFF(&x86::X86Assembler::movdqa, "vmovdqa %{reg2}, %{reg1}"), "avx_movdqa"); in TEST_F() 723 DriverStr(RepeatFA(&x86::X86Assembler::movdqa, "movdqa {mem}, %{reg}"), "movdqa_load"); in TEST_F() 731 DriverStr(RepeatFA(&x86::X86Assembler::movdqa, "vmovdqa {mem}, %{reg}"), "avx_movdqa_load"); in TEST_F() 735 DriverStr(RepeatAF(&x86::X86Assembler::movdqa, "movdqa %{reg}, {mem}"), "movdqa_store"); in TEST_F() 743 DriverStr(RepeatAF(&x86::X86Assembler::movdqa, "vmovdqa %{reg}, {mem}"), "avx_movdqa_store"); in TEST_F()
|
D | assembler_x86.h | 531 void movdqa(XmmRegister dst, XmmRegister src); // move 532 void movdqa(XmmRegister dst, const Address& src); // load aligned 534 void movdqa(const Address& dst, XmmRegister src); // store aligned
|
D | assembler_x86.cc | 1221 void X86Assembler::movdqa(XmmRegister dst, XmmRegister src) { in movdqa() function in art::x86::X86Assembler 1252 void X86Assembler::movdqa(XmmRegister dst, const Address& src) { in movdqa() function in art::x86::X86Assembler 1314 void X86Assembler::movdqa(const Address& dst, XmmRegister src) { in movdqa() function in art::x86::X86Assembler
|
/art/compiler/utils/x86_64/ |
D | assembler_x86_64_test.cc | 1393 DriverStr(RepeatFF(&x86_64::X86_64Assembler::movdqa, "movdqa %{reg2}, %{reg1}"), "movdqa"); in TEST_F() 1401 DriverStr(RepeatFF(&x86_64::X86_64Assembler::movdqa, "vmovdqa %{reg2}, %{reg1}"), "avx_movdqa"); in TEST_F() 1405 DriverStr(RepeatAF(&x86_64::X86_64Assembler::movdqa, "movdqa %{reg}, {mem}"), "movdqa_s"); in TEST_F() 1413 DriverStr(RepeatAF(&x86_64::X86_64Assembler::movdqa, "vmovdqa %{reg}, {mem}"), "avx_movdqa_s"); in TEST_F() 1417 DriverStr(RepeatFA(&x86_64::X86_64Assembler::movdqa, "movdqa {mem}, %{reg}"), "movdqa_l"); in TEST_F() 1425 DriverStr(RepeatFA(&x86_64::X86_64Assembler::movdqa, "vmovdqa {mem}, %{reg}"), "avx_movdqa_l"); in TEST_F()
|
D | assembler_x86_64.h | 563 void movdqa(XmmRegister dst, XmmRegister src); // move 564 void movdqa(XmmRegister dst, const Address& src); // load aligned 566 void movdqa(const Address& dst, XmmRegister src); // store aligned
|
D | assembler_x86_64.cc | 1595 void X86_64Assembler::movdqa(XmmRegister dst, XmmRegister src) { in movdqa() function in art::x86_64::X86_64Assembler 1656 void X86_64Assembler::movdqa(XmmRegister dst, const Address& src) { in movdqa() function in art::x86_64::X86_64Assembler 1765 void X86_64Assembler::movdqa(const Address& dst, XmmRegister src) { in movdqa() function in art::x86_64::X86_64Assembler
|
/art/compiler/optimizing/ |
D | code_generator_vector_x86_64.cc | 1293 is_aligned16 ? __ movdqa(reg, address) : __ movdqu(reg, address); in VisitVecLoad() 1305 is_aligned16 ? __ movdqa(reg, address) : __ movdqu(reg, address); in VisitVecLoad() 1341 is_aligned16 ? __ movdqa(address, reg) : __ movdqu(address, reg); in VisitVecStore()
|
D | code_generator_vector_x86.cc | 1320 is_aligned16 ? __ movdqa(reg, address) : __ movdqu(reg, address); in VisitVecLoad() 1332 is_aligned16 ? __ movdqa(reg, address) : __ movdqu(reg, address); in VisitVecLoad() 1368 is_aligned16 ? __ movdqa(address, reg) : __ movdqu(address, reg); in VisitVecStore()
|