/art/runtime/arch/x86/ |
H A D | memcmp16_x86.S | 139 movdqa (%esi), %xmm0 143 movdqa 16(%esi), %xmm2 149 movdqa %xmm0, %xmm1 150 movdqa 32(%esi), %xmm0 151 movdqa 48(%esi), %xmm2 169 movdqa %xmm0, %xmm1 189 movdqa 16(%esi), %xmm1 190 movdqa %xmm1, %xmm2 194 movdqa 32(%esi), %xmm3 215 movdqa 1 [all...] |
/art/runtime/arch/x86_64/ |
H A D | memcmp16_x86_64.S | 319 movdqa %xmm2, %xmm1 353 movdqa %xmm2, %xmm1 389 movdqa (%rdi), %xmm2 394 movdqa 16(%rdi), %xmm2 399 movdqa 32(%rdi), %xmm2 404 movdqa 48(%rdi), %xmm2 411 movdqa 64(%rdi), %xmm2 416 movdqa 80(%rdi), %xmm2 439 movdqa (%rdi), %xmm2 444 movdqa 1 [all...] |
/art/compiler/utils/x86/ |
H A D | assembler_x86_test.cc | 512 DriverStr(RepeatFF(&x86::X86Assembler::movdqa, "movdqa %{reg2}, %{reg1}"), "movdqa"); 516 DriverStr(RepeatFA(&x86::X86Assembler::movdqa, "movdqa {mem}, %{reg}"), "movdqa_load"); 520 DriverStr(RepeatAF(&x86::X86Assembler::movdqa, "movdqa %{reg}, {mem}"), "movdqa_store");
|
H A D | assembler_x86.h | 432 void movdqa(XmmRegister dst, XmmRegister src); // move 433 void movdqa(XmmRegister dst, const Address& src); // load aligned 435 void movdqa(const Address& dst, XmmRegister src); // store aligned
|
H A D | assembler_x86.cc | 780 void X86Assembler::movdqa(XmmRegister dst, XmmRegister src) { function in class:art::x86::X86Assembler 789 void X86Assembler::movdqa(XmmRegister dst, const Address& src) { function in class:art::x86::X86Assembler 807 void X86Assembler::movdqa(const Address& dst, XmmRegister src) { function in class:art::x86::X86Assembler
|
/art/compiler/utils/x86_64/ |
H A D | assembler_x86_64.h | 468 void movdqa(XmmRegister dst, XmmRegister src); // move 469 void movdqa(XmmRegister dst, const Address& src); // load aligned 471 void movdqa(const Address& dst, XmmRegister src); // store aligned
|
H A D | assembler_x86_64_test.cc | 1154 DriverStr(RepeatFF(&x86_64::X86_64Assembler::movdqa, "movdqa %{reg2}, %{reg1}"), "movdqa"); 1158 DriverStr(RepeatAF(&x86_64::X86_64Assembler::movdqa, "movdqa %{reg}, {mem}"), "movdqa_s"); 1162 DriverStr(RepeatFA(&x86_64::X86_64Assembler::movdqa, "movdqa {mem}, %{reg}"), "movdqa_l");
|
H A D | assembler_x86_64.cc | 863 void X86_64Assembler::movdqa(XmmRegister dst, XmmRegister src) { function in class:art::x86_64::X86_64Assembler 873 void X86_64Assembler::movdqa(XmmRegister dst, const Address& src) { function in class:art::x86_64::X86_64Assembler 893 void X86_64Assembler::movdqa(const Address& dst, XmmRegister src) { function in class:art::x86_64::X86_64Assembler
|
/art/compiler/optimizing/ |
H A D | code_generator_vector_x86.cc | 1163 is_aligned16 ? __ movdqa(reg, address) : __ movdqu(reg, address); 1176 is_aligned16 ? __ movdqa(reg, address) : __ movdqu(reg, address); 1212 is_aligned16 ? __ movdqa(address, reg) : __ movdqu(address, reg);
|
H A D | code_generator_vector_x86_64.cc | 1136 is_aligned16 ? __ movdqa(reg, address) : __ movdqu(reg, address); 1149 is_aligned16 ? __ movdqa(reg, address) : __ movdqu(reg, address); 1185 is_aligned16 ? __ movdqa(address, reg) : __ movdqu(address, reg);
|